Vandana Yadav created CARBONDATA-1846:
----------------------------------------- Summary: Incorrect output on presto CLI while executing IN operator with multiple load Key: CARBONDATA-1846 URL: https://issues.apache.org/jira/browse/CARBONDATA-1846 Project: CarbonData Issue Type: Bug Components: presto-integration Affects Versions: 1.3.0 Environment: spark 2.1 Reporter: Vandana Yadav Priority: Minor Attachments: 2000_UniqData.csv, 3000_UniqData.csv, 4000_UniqData.csv, 5000_UniqData.csv, 6000_UniqData.csv, 7000_UniqData.csv Incorrect output on presto CLI while executing IN operator with multiple load Steps to reproduce: On beeline: 1) Create a table: CREATE TABLE uniqdata_time (CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION string, DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1 int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES ("TABLE_BLOCKSIZE"= "256 MB") 2)Load Data with multiple segments in it: a) LOAD DATA INPATH 'hdfs://localhost:54310/Data/uniqdata/2000_UniqData.csv' into table uniqdata_time OPTIONS('DELIMITER'=',' , 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1') b) LOAD DATA INPATH 'hdfs://localhost:54310/Data/uniqdata/3000_UniqData.csv' into table uniqdata_time OPTIONS('DELIMITER'=',' , 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1') c) LOAD DATA INPATH 'hdfs://localhost:54310/Data/uniqdata/4000_UniqData.csv' into table uniqdata_time OPTIONS('DELIMITER'=',' , 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1') d) LOAD DATA INPATH 'hdfs://localhost:54310/Data/uniqdata/5000_UniqData.csv' into table uniqdata_time OPTIONS('DELIMITER'=',' , 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1') e) LOAD DATA INPATH 'hdfs://localhost:54310/Data/uniqdata/6000_UniqData.csv' into table uniqdata_time OPTIONS('DELIMITER'=',' , 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1') f) LOAD DATA INPATH 'hdfs://localhost:54310/Data/uniqdata/7000_UniqData.csv' into table uniqdata_time OPTIONS('DELIMITER'=',' , 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1') 3) Execute Query: select DOB from UNIQDATA_time where DOB in (cast('2015-10-04 01:00:03' as timestamp),cast('2015-10-07' as timestamp),cast('2015-10-07 01:00:03' as timestamp)); Output on Beeline: +------------------------+--+ | DOB | +------------------------+--+ | 2015-10-04 01:00:03.0 | | 2015-10-07 01:00:03.0 | +------------------------+--+ 2 rows selected (0.351 seconds) 4) Output on Presto CLI: a) Expected Output: similar to the result on Beeline. b) Actual Output : DOB ----- (0 rows) -- This message was sent by Atlassian JIRA (v6.4.14#64029) |
Free forum by Nabble | Edit this page |