[ https://issues.apache.org/jira/browse/CARBONDATA-1982?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Geetika Gupta updated CARBONDATA-1982: -------------------------------------- Description: I created a partitioned table using: CREATE TABLE uniqdata_int_dec(CUST_NAME String,ACTIVE_EMUI_VERSION string, DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 bigint, DECIMAL_COLUMN2 decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double, INTEGER_COLUMN1 int) Partitioned by (cust_id int, decimal_column1 decimal(30,10)) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES ("TABLE_BLOCKSIZE"= "256 MB") Load data command: LOAD DATA INPATH 'hdfs://localhost:54311/2000_UniqData.csv' into table uniqdata_int_dec partition(cust_id123='1', abc='12345678901.1234') OPTIONS ('FILEHEADER'='CUST_ID,CUST_NAME ,ACTIVE_EMUI_VERSION,DOB,DOJ, BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1, Double_COLUMN2,INTEGER_COLUMN1','BAD_RECORDS_ACTION'='FORCE'); OUTPUT: 0: jdbc:hive2://localhost:10000> LOAD DATA INPATH 'hdfs://localhost:54311/2000_UniqData.csv' into table uniqdata_int_dec partition(cust_id123='1', decimal_column1='12345678901.1234') OPTIONS ('FILEHEADER'='CUST_ID,CUST_NAME ,ACTIVE_EMUI_VERSION,DOB,DOJ, BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1, Double_COLUMN2,INTEGER_COLUMN1','BAD_RECORDS_ACTION'='FORCE'); Error: java.lang.IndexOutOfBoundsException: Index: 1, Size: 1 (state=,code=0) The above command throws java.lang.IndexOutOfBoundsException: Index: 1, Size: 1 whereas it should throw a proper exception like invalid column expression for partition load command. was: I created a partitioned table using: CREATE TABLE uniqdata_int_dec(CUST_NAME String,ACTIVE_EMUI_VERSION string, DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 bigint, DECIMAL_COLUMN2 decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double, INTEGER_COLUMN1 int) Partitioned by (cust_id int, decimal_column1 decimal(30,10)) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES ("TABLE_BLOCKSIZE"= "256 MB") Load data command: LOAD DATA INPATH 'hdfs://localhost:54311/2000_UniqData.csv' into table uniqdata_int_dec partition(cust_id123='1', abc='12345678901.1234') OPTIONS ('FILEHEADER'='CUST_ID,CUST_NAME ,ACTIVE_EMUI_VERSION,DOB,DOJ, BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1, Double_COLUMN2,INTEGER_COLUMN1','BAD_RECORDS_ACTION'='FORCE'); OUTPUT: 0: jdbc:hive2://localhost:10000> LOAD DATA INPATH 'hdfs://localhost:54311/Files/2000_UniqData.csv' into table uniqdata_int_dec partition(cust_id123='1', decimal_column1='12345678901.1234') OPTIONS ('FILEHEADER'='CUST_ID,CUST_NAME ,ACTIVE_EMUI_VERSION,DOB,DOJ, BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1, Double_COLUMN2,INTEGER_COLUMN1','BAD_RECORDS_ACTION'='FORCE'); Error: java.lang.IndexOutOfBoundsException: Index: 1, Size: 1 (state=,code=0) The above command throws java.lang.IndexOutOfBoundsException: Index: 1, Size: 1 whereas it should throw a proper exception like invalid column expression for partition load command. > Loading data into partition table with invalid partition column should throw proper exception > --------------------------------------------------------------------------------------------- > > Key: CARBONDATA-1982 > URL: https://issues.apache.org/jira/browse/CARBONDATA-1982 > Project: CarbonData > Issue Type: Bug > Components: data-query > Affects Versions: 1.3.0 > Environment: spark2.1 > Reporter: Geetika Gupta > Assignee: anubhav tarar > Priority: Minor > Fix For: 1.3.0 > > Attachments: 2000_UniqData.csv > > > I created a partitioned table using: > CREATE TABLE uniqdata_int_dec(CUST_NAME String,ACTIVE_EMUI_VERSION string, DOB timestamp, > DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 bigint, > DECIMAL_COLUMN2 decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double, > INTEGER_COLUMN1 int) Partitioned by (cust_id int, decimal_column1 decimal(30,10)) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES ("TABLE_BLOCKSIZE"= "256 MB") > Load data command: > LOAD DATA INPATH 'hdfs://localhost:54311/2000_UniqData.csv' into table uniqdata_int_dec partition(cust_id123='1', abc='12345678901.1234') OPTIONS ('FILEHEADER'='CUST_ID,CUST_NAME ,ACTIVE_EMUI_VERSION,DOB,DOJ, BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1, Double_COLUMN2,INTEGER_COLUMN1','BAD_RECORDS_ACTION'='FORCE'); > OUTPUT: > 0: jdbc:hive2://localhost:10000> LOAD DATA INPATH 'hdfs://localhost:54311/2000_UniqData.csv' into table uniqdata_int_dec partition(cust_id123='1', decimal_column1='12345678901.1234') OPTIONS ('FILEHEADER'='CUST_ID,CUST_NAME ,ACTIVE_EMUI_VERSION,DOB,DOJ, BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1, Double_COLUMN2,INTEGER_COLUMN1','BAD_RECORDS_ACTION'='FORCE'); > Error: java.lang.IndexOutOfBoundsException: Index: 1, Size: 1 (state=,code=0) > The above command throws java.lang.IndexOutOfBoundsException: Index: 1, Size: 1 whereas it should throw a proper exception like invalid column expression for partition load command. -- This message was sent by Atlassian JIRA (v7.6.3#76005) |
Free forum by Nabble | Edit this page |