[ https://issues.apache.org/jira/browse/CARBONDATA-1421?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16146837#comment-16146837 ] Zhichao Zhang commented on CARBONDATA-1421: -------------------------------------------- Hi Pallavi Singh, I run below code and it worked with pr-1266, My config: carbon.enable.auto.load.merge=true carbon.compaction.level.threshold=2,4 My code: {code:java} import spark.implicits._ val df1 = spark.sparkContext.parallelize(0 to 50) .map(x => ("a", x.toString(), (x % 2).toString(), x, x.toLong, x * 2)) .toDF("stringField1", "stringField2", "stringField3", "intField", "longField", "int2Field") val df2 = spark.sparkContext.parallelize(51 to 100) .map(x => ("b", x.toString(), (x % 2).toString(), x, x.toLong, x * 2)) .toDF("stringField1", "stringField2", "stringField3", "intField", "longField", "int2Field") val df3 = df1.union(df2) spark.sql("DROP TABLE IF EXISTS study_carbondata ").show() spark.sql(""" | CREATE TABLE IF NOT EXISTS study_carbondata ( | stringField1 string, | stringField2 string, | stringField3 string, | intField int, | longField bigint, | int2Field int | ) | STORED BY 'carbondata' | TBLPROPERTIES( | 'TABLE_BLOCKSIZE'='8' | ) """.stripMargin) df3.write .format("carbondata") .option("tableName", "study_carbondata") .option("compress", "true") // just valid when tempCSV is true .option("tempCSV", "false") .option("single_pass", "true") .option("sort_scope", "LOCAL_SORT") //GLOBAL_SORT LOCAL_SORT .mode(SaveMode.Append) .save() df3.write .format("carbondata") .option("tableName", "study_carbondata") .option("compress", "true") // just valid when tempCSV is true .option("tempCSV", "false") .option("single_pass", "true") .option("sort_scope", "LOCAL_SORT") //GLOBAL_SORT LOCAL_SORT .mode(SaveMode.Append) .save() {code} > Auto Compaction Failing in CarbonData Loading > --------------------------------------------- > > Key: CARBONDATA-1421 > URL: https://issues.apache.org/jira/browse/CARBONDATA-1421 > Project: CarbonData > Issue Type: Bug > Components: data-load > Affects Versions: 1.2.0 > Reporter: Pallavi Singh > Fix For: 1.2.0 > > > I ran the create query followed by multiple load queries and the auto-compaction is failing. > 0: jdbc:hive2://localhost:10000> LOAD DATA inpath 'hdfs://localhost:54310/data/4000_UniqData.csv' INTO table uniqdata options('DELIMITER'=',', 'FILEHEADER'='CUST_ID, CUST_NAME, ACTIVE_EMUI_VERSION, DOB, DOJ, BIGINT_COLUMN1, BIGINT_COLUMN2, DECIMAL_COLUMN1, DECIMAL_COLUMN2, Double_COLUMN1, Double_COLUMN2, INTEGER_COLUMN1'); > +---------+--+ > | Result | > +---------+--+ > +---------+--+ > No rows selected (1.183 seconds) > 0: jdbc:hive2://localhost:10000> LOAD DATA inpath 'hdfs://localhost:54310/data/5000_UniqData.csv' INTO table uniqdata options('DELIMITER'=',', 'FILEHEADER'='CUST_ID, CUST_NAME, ACTIVE_EMUI_VERSION, DOB, DOJ, BIGINT_COLUMN1, BIGINT_COLUMN2, DECIMAL_COLUMN1, DECIMAL_COLUMN2, Double_COLUMN1, Double_COLUMN2, INTEGER_COLUMN1'); > Error: java.lang.Exception: Dataload is success. Auto-Compaction has failed. Please check logs. (state=,code=0) > 0: jdbc:hive2://localhost:10000> LOAD DATA inpath 'hdfs://localhost:54310/data/7000_UniqData.csv' INTO table uniqdata options('DELIMITER'=',', 'FILEHEADER'='CUST_ID, CUST_NAME, ACTIVE_EMUI_VERSION, DOB, DOJ, BIGINT_COLUMN1, BIGINT_COLUMN2, DECIMAL_COLUMN1, DECIMAL_COLUMN2, Double_COLUMN1, Double_COLUMN2, INTEGER_COLUMN1'); > Error: java.lang.Exception: Dataload is success. Auto-Compaction has failed. Please check logs. (state=,code=0) > 0: jdbc:hive2://localhost:10000> > 0: jdbc:hive2://localhost:10000> > 0: jdbc:hive2://localhost:10000> > 0: jdbc:hive2://localhost:10000> > 0: jdbc:hive2://localhost:10000> > 0: jdbc:hive2://localhost:10000> > 0: jdbc:hive2://localhost:10000> > 0: jdbc:hive2://localhost:10000> show segments for table uniqdata; > +--------------------+----------+--------------------------+--------------------------+--+ > | SegmentSequenceId | Status | Load Start Time | Load End Time | > +--------------------+----------+--------------------------+--------------------------+--+ > | 4 | Success | 2017-08-29 10:37:13.053 | 2017-08-29 10:37:13.888 | > | 3 | Success | 2017-08-29 10:36:57.851 | 2017-08-29 10:36:59.08 | > | 2 | Success | 2017-08-29 10:36:49.439 | 2017-08-29 10:36:50.373 | > | 1 | Success | 2017-08-29 10:36:37.365 | 2017-08-29 10:36:38.768 | > | 0 | Success | 2017-08-29 10:36:21.011 | 2017-08-29 10:36:26.1 | > +--------------------+----------+--------------------------+--------------------------+--+ > 5 rows selected (0.099 seconds) > 0: jdbc:hive2://localhost:10000> LOAD DATA inpath 'hdfs://localhost:54310/data/7000_UniqData.csv' INTO table uniqdata options('DELIMITER'=',', 'FILEHEADER'='CUST_ID, CUST_NAME, ACTIVE_EMUI_VERSION, DOB, DOJ, BIGINT_COLUMN1, BIGINT_COLUMN2, DECIMAL_COLUMN1, DECIMAL_COLUMN2, Double_COLUMN1, Double_COLUMN2, INTEGER_COLUMN1'); > Error: java.lang.Exception: Dataload is success. Auto-Compaction has failed. Please check logs. (state=,code=0) > 0: jdbc:hive2://localhost:10000> show segments for table uniqdata; > +--------------------+----------+--------------------------+--------------------------+--+ > | SegmentSequenceId | Status | Load Start Time | Load End Time | > +--------------------+----------+--------------------------+--------------------------+--+ > | 5 | Success | 2017-08-29 10:38:15.727 | 2017-08-29 10:38:16.548 | > | 4 | Success | 2017-08-29 10:37:13.053 | 2017-08-29 10:37:13.888 | > | 3 | Success | 2017-08-29 10:36:57.851 | 2017-08-29 10:36:59.08 | > | 2 | Success | 2017-08-29 10:36:49.439 | 2017-08-29 10:36:50.373 | > | 1 | Success | 2017-08-29 10:36:37.365 | 2017-08-29 10:36:38.768 | > | 0 | Success | 2017-08-29 10:36:21.011 | 2017-08-29 10:36:26.1 | > +--------------------+----------+--------------------------+--------------------------+--+ -- This message was sent by Atlassian JIRA (v6.4.14#64029) |
Free forum by Nabble | Edit this page |