[ https://issues.apache.org/jira/browse/CARBONDATA-1846?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] anubhav tarar reassigned CARBONDATA-1846: ----------------------------------------- Assignee: anubhav tarar > Incorrect output on presto CLI while executing IN operator with multiple load > ----------------------------------------------------------------------------- > > Key: CARBONDATA-1846 > URL: https://issues.apache.org/jira/browse/CARBONDATA-1846 > Project: CarbonData > Issue Type: Bug > Components: presto-integration > Affects Versions: 1.3.0 > Environment: spark 2.1 > Reporter: Vandana Yadav > Assignee: anubhav tarar > Priority: Minor > Attachments: 2000_UniqData.csv, 3000_UniqData.csv, 4000_UniqData.csv, 5000_UniqData.csv, 6000_UniqData.csv, 7000_UniqData.csv > > Time Spent: 1h 10m > Remaining Estimate: 0h > > Incorrect output on presto CLI while executing IN operator with multiple load > Steps to reproduce: > On beeline: > 1) Create a table: > CREATE TABLE uniqdata_time (CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION string, DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1 int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES ("TABLE_BLOCKSIZE"= "256 MB") > 2)Load Data with multiple segments in it: > a) LOAD DATA INPATH 'hdfs://localhost:54310/Data/uniqdata/2000_UniqData.csv' into table uniqdata_time OPTIONS('DELIMITER'=',' , 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1') > b) LOAD DATA INPATH 'hdfs://localhost:54310/Data/uniqdata/3000_UniqData.csv' into table uniqdata_time OPTIONS('DELIMITER'=',' , 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1') > c) LOAD DATA INPATH 'hdfs://localhost:54310/Data/uniqdata/4000_UniqData.csv' into table uniqdata_time OPTIONS('DELIMITER'=',' , 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1') > d) LOAD DATA INPATH 'hdfs://localhost:54310/Data/uniqdata/5000_UniqData.csv' into table uniqdata_time OPTIONS('DELIMITER'=',' , 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1') > e) LOAD DATA INPATH 'hdfs://localhost:54310/Data/uniqdata/6000_UniqData.csv' into table uniqdata_time OPTIONS('DELIMITER'=',' , 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1') > f) LOAD DATA INPATH 'hdfs://localhost:54310/Data/uniqdata/7000_UniqData.csv' into table uniqdata_time OPTIONS('DELIMITER'=',' , 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1') > 3) Execute Query: > select DOB from UNIQDATA_time where DOB in (cast('2015-10-04 01:00:03' as timestamp),cast('2015-10-07' as timestamp),cast('2015-10-07 01:00:03' as timestamp)); > Output on Beeline: > +------------------------+--+ > | DOB | > +------------------------+--+ > | 2015-10-04 01:00:03.0 | > | 2015-10-07 01:00:03.0 | > +------------------------+--+ > 2 rows selected (0.351 seconds) > 4) Output on Presto CLI: > a) Expected Output: similar to the result on Beeline. > b) Actual Output : > DOB > ----- > (0 rows) -- This message was sent by Atlassian JIRA (v6.4.14#64029) |
Free forum by Nabble | Edit this page |