[ https://issues.apache.org/jira/browse/CARBONDATA-3451?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16880140#comment-16880140 ] dhatchayani commented on CARBONDATA-3451: ----------------------------------------- Please check this again. It is already fixed in [CARBONDATA-3441|https://issues.apache.org/jira/browse/CARBONDATA-3441] > Select aggregation query with filter fails on hive table with decimal type using CarbonHiveSerDe in Spark 2.1 > ------------------------------------------------------------------------------------------------------------- > > Key: CARBONDATA-3451 > URL: https://issues.apache.org/jira/browse/CARBONDATA-3451 > Project: CarbonData > Issue Type: Bug > Components: data-query > Affects Versions: 1.6.0 > Environment: Spark 2.1 > Reporter: Chetan Bhat > Priority: Minor > > Test steps : > In Spark 2.1 beeline user creates a carbon table and loads data. > create table Test_Boundary (c1_int int,c2_Bigint Bigint,c3_Decimal Decimal(38,38),c4_double double,c5_string string,c6_Timestamp Timestamp,c7_Datatype_Desc string) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES('inverted_index'='c1_int,c2_Bigint,c5_string,c6_Timestamp','sort_columns'='c1_int,c2_Bigint,c5_string,c6_Timestamp'); > LOAD DATA INPATH 'hdfs://hacluster/chetan/Test_Data1.csv' INTO table Test_Boundary OPTIONS('DELIMITER'=',','QUOTECHAR'='','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'=''); > From hive beeline user creates a hive table from the already created carbon table using CarbonHiveSerDe. > CREATE TABLE IF NOT EXISTS Test_Boundary1 (c1_int int,c2_Bigint Bigint,c3_Decimal Decimal(38,38),c4_double double,c5_string string,c6_Timestamp Timestamp,c7_Datatype_Desc string) ROW FORMAT SERDE 'org.apache.carbondata.hive.CarbonHiveSerDe' WITH SERDEPROPERTIES ('mapreduce.input.carboninputformat.databaseName'='default','mapreduce.input.carboninputformat.tableName'='Test_Boundary') STORED AS INPUTFORMAT 'org.apache.carbondata.hive.MapredCarbonInputFormat' OUTPUTFORMAT 'org.apache.carbondata.hive.MapredCarbonOutputFormat' LOCATION 'hdfs://hacluster//user/hive/warehouse/carbon.store/default/test_boundary'; > User executes below select aggregation query on the hive table. > select min(c3_Decimal),max(c3_Decimal),sum(c3_Decimal),avg(c3_Decimal) , count(c3_Decimal), variance(c3_Decimal) from test_boundary1 where exp(c1_int)=0.0 or exp(c1_int)=1.0; > select min(c3_Decimal),max(c3_Decimal),sum(c3_Decimal),avg(c3_Decimal) , count(c3_Decimal), variance(c3_Decimal) from test_boundary1 where log(c1_int,1)=0.0 or log(c1_int,1) IS NULL; > select min(c3_Decimal),max(c3_Decimal),sum(c3_Decimal),avg(c3_Decimal) , count(c3_Decimal), variance(c3_Decimal) from test_boundary1 where pmod(c1_int,1)=0 or pmod(c1_int,1)IS NULL; > > Actual Result : Select aggregation query with filter fails on hive table with decimal type using CarbonHiveSerDe in Spark 2.1 > Expected Result : Select aggregation query with filter should be success on hive table with decimal type using CarbonHiveSerDe in Spark 2.1 > -- This message was sent by Atlassian JIRA (v7.6.3#76005) |
Free forum by Nabble | Edit this page |