[ https://issues.apache.org/jira/browse/CARBONDATA-1422?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16146975#comment-16146975 ] Pallavi Singh commented on CARBONDATA-1422: ------------------------------------------- Hi Zhichao Zhang, Here is the stack-trace of the error: 17/08/30 15:12:30 INFO CatalystSqlParser: Parsing command: array<string> 17/08/30 15:12:30 AUDIT CarbonDataRDDFactory$: [pallavi][hduser][Thread-136]Compaction request received for table 30aug_5.uniqdata 17/08/30 15:12:30 INFO HdfsFileLock: pool-24-thread-7 HDFS lock path:hdfs://localhost:54310/user/hive/warehouse/carbon.store/30aug_5/uniqdata/compaction.lock 17/08/30 15:12:30 INFO CarbonDataRDDFactory$: pool-24-thread-7 Acquired the compaction lock for table 30aug_5.uniqdata 17/08/30 15:12:30 INFO DataManagementFunc$: pool-24-thread-7 loads identified for merge is 0 17/08/30 15:12:30 INFO DataManagementFunc$: pool-24-thread-7 loads identified for merge is 1 17/08/30 15:12:30 INFO DataManagementFunc$: pool-24-thread-7 loads identified for merge is 2 17/08/30 15:12:30 INFO DataManagementFunc$: pool-24-thread-7 loads identified for merge is 3 17/08/30 15:12:30 INFO DataManagementFunc$: pool-24-thread-7 loads identified for merge is 4 17/08/30 15:12:30 INFO DataManagementFunc$: pool-24-thread-7 loads identified for merge is 5 17/08/30 15:12:30 INFO DataManagementFunc$: pool-24-thread-7 loads identified for merge is 6 17/08/30 15:12:30 INFO DataManagementFunc$: pool-24-thread-7 loads identified for merge is 7 17/08/30 15:12:30 INFO TableInfo: pool-28-thread-1 Table block size not specified for 30aug_5_uniqdata. Therefore considering the default value 1024 MB 17/08/30 15:12:30 INFO DistributionUtil$: pool-28-thread-1 Requesting total executors: 1 17/08/30 15:12:30 INFO CarbonMergerRDD: No.of Executors required=1 , spark.executor.instances=1, no.of.nodes where data present=1 17/08/30 15:12:30 INFO CarbonMergerRDD: Time taken to wait for executor allocation is =0millis 17/08/30 15:12:30 INFO CarbonMergerRDD: for the node pallavi 17/08/30 15:12:30 INFO CarbonMergerRDD: Task ID is 0no. of blocks is 8 17/08/30 15:12:30 INFO CarbonMergerRDD: Identified no.of.Blocks: 8,parallelism: 4 , no.of.nodes: 1, no.of.tasks: 1 17/08/30 15:12:30 INFO CarbonMergerRDD: Time taken to identify Blocks to scan : 177 17/08/30 15:12:30 INFO CarbonMergerRDD: Node: pallavi, No.Of Blocks: 8 17/08/30 15:12:30 INFO SparkContext: Starting job: collect at Compactor.scala:100 17/08/30 15:12:30 INFO DAGScheduler: Got job 4 (collect at Compactor.scala:100) with 1 output partitions 17/08/30 15:12:30 INFO DAGScheduler: Final stage: ResultStage 4 (collect at Compactor.scala:100) 17/08/30 15:12:30 INFO DAGScheduler: Parents of final stage: List() 17/08/30 15:12:30 INFO DAGScheduler: Missing parents: List() 17/08/30 15:12:30 INFO DAGScheduler: Submitting ResultStage 4 (CarbonMergerRDD[11] at RDD at CarbonRDD.scala:35), which has no missing parents 17/08/30 15:12:30 INFO MemoryStore: Block broadcast_6 stored as values in memory (estimated size 14.2 KB, free 366.2 MB) 17/08/30 15:12:30 INFO MemoryStore: Block broadcast_6_piece0 stored as bytes in memory (estimated size 6.7 KB, free 366.2 MB) 17/08/30 15:12:30 INFO BlockManagerInfo: Added broadcast_6_piece0 in memory on 192.168.2.188:41587 (size: 6.7 KB, free: 366.3 MB) 17/08/30 15:12:30 INFO SparkContext: Created broadcast 6 from broadcast at DAGScheduler.scala:996 17/08/30 15:12:30 INFO DAGScheduler: Submitting 1 missing tasks from ResultStage 4 (CarbonMergerRDD[11] at RDD at CarbonRDD.scala:35) 17/08/30 15:12:30 INFO TaskSchedulerImpl: Adding task set 4.0 with 1 tasks 17/08/30 15:12:30 INFO TaskSetManager: Starting task 0.0 in stage 4.0 (TID 4, 192.168.2.188, executor 0, partition 0, ANY, 9296 bytes) 17/08/30 15:12:30 INFO BlockManagerInfo: Added broadcast_6_piece0 in memory on 192.168.2.188:38477 (size: 6.7 KB, free: 366.3 MB) 17/08/30 15:12:31 INFO TaskSetManager: Finished task 0.0 in stage 4.0 (TID 4) in 266 ms on 192.168.2.188 (executor 0) (1/1) 17/08/30 15:12:31 INFO TaskSchedulerImpl: Removed TaskSet 4.0, whose tasks have all completed, from pool 17/08/30 15:12:31 INFO DAGScheduler: ResultStage 4 (collect at Compactor.scala:100) finished in 0.267 s 17/08/30 15:12:31 INFO DAGScheduler: Job 4 finished: collect at Compactor.scala:100, took 0.276120 s 17/08/30 15:12:31 AUDIT Compactor$: [pallavi][hduser][Thread-143]Compaction request failed for table 30aug_5.uniqdata 17/08/30 15:12:31 ERROR Compactor$: pool-28-thread-1 Compaction request failed for table 30aug_5.uniqdata 17/08/30 15:12:31 ERROR DataManagementFunc$: pool-24-thread-7 Exception in compaction thread java.lang.Exception: Compaction Failure in Merger Rdd. java.util.concurrent.ExecutionException: java.lang.Exception: Compaction Failure in Merger Rdd. at java.util.concurrent.FutureTask.report(FutureTask.java:122) at java.util.concurrent.FutureTask.get(FutureTask.java:192) at org.apache.carbondata.spark.rdd.DataManagementFunc$$anonfun$executeCompaction$1.apply(DataManagementFunc.scala:193) at org.apache.carbondata.spark.rdd.DataManagementFunc$$anonfun$executeCompaction$1.apply(DataManagementFunc.scala:192) at scala.collection.Iterator$class.foreach(Iterator.scala:893) at scala.collection.AbstractIterator.foreach(Iterator.scala:1336) at scala.collection.IterableLike$class.foreach(IterableLike.scala:72) at scala.collection.AbstractIterable.foreach(Iterable.scala:54) at org.apache.carbondata.spark.rdd.DataManagementFunc$.executeCompaction(DataManagementFunc.scala:192) at org.apache.carbondata.spark.rdd.CarbonDataRDDFactory$$anon$2.run(CarbonDataRDDFactory.scala:273) at org.apache.carbondata.spark.rdd.CarbonDataRDDFactory$.startCompactionThreads(CarbonDataRDDFactory.scala:364) at org.apache.carbondata.spark.rdd.CarbonDataRDDFactory$.alterTableForCompaction(CarbonDataRDDFactory.scala:157) at org.apache.spark.sql.execution.command.AlterTableCompaction.processData(carbonTableSchema.scala:175) at org.apache.spark.sql.execution.command.AlterTableCompaction.run(carbonTableSchema.scala:139) at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:58) at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:56) at org.apache.spark.sql.execution.command.ExecutedCommandExec.doExecute(commands.scala:74) at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:114) at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:114) at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:135) at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:132) at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:113) at org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:87) at org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:87) at org.apache.spark.sql.Dataset.<init>(Dataset.scala:185) at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:64) at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:592) at org.apache.spark.sql.SQLContext.sql(SQLContext.scala:699) at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.org$apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute(SparkExecuteStatementOperation.scala:220) at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:163) at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:160) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1698) at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1.run(SparkExecuteStatementOperation.scala:173) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:748) Caused by: java.lang.Exception: Compaction Failure in Merger Rdd. at org.apache.carbondata.spark.rdd.Compactor$.triggerCompaction(Compactor.scala:143) at org.apache.carbondata.spark.rdd.Compactor.triggerCompaction(Compactor.scala) at org.apache.carbondata.spark.compaction.CompactionCallable.call(CompactionCallable.java:40) at org.apache.carbondata.spark.compaction.CompactionCallable.call(CompactionCallable.java:29) ... 4 more 17/08/30 15:12:31 ERROR CarbonDataRDDFactory$: pool-24-thread-7 Exception in compaction thread java.lang.Exception: Compaction Failure in Merger Rdd. 17/08/30 15:12:31 INFO HdfsFileLock: pool-24-thread-7 Deleted the lock file hdfs://localhost:54310/user/hive/warehouse/carbon.store/30aug_5/uniqdata/compaction.lock 17/08/30 15:12:31 ERROR CarbonDataRDDFactory$: pool-24-thread-7 Exception in start compaction thread. Exception in compaction java.lang.Exception: Compaction Failure in Merger Rdd. 17/08/30 15:12:31 ERROR AbstractDFSCarbonFile: pool-24-thread-7 Exception occurred:File does not exist: hdfs://localhost:54310/user/hive/warehouse/carbon.store/30aug_5/uniqdata/compaction.lock 17/08/30 15:12:31 ERROR HdfsFileLock: pool-24-thread-7 Not able to delete the lock file because it is not existed in location hdfs://localhost:54310/user/hive/warehouse/carbon.store/30aug_5/uniqdata/compaction.lock 17/08/30 15:12:31 ERROR SparkExecuteStatementOperation: Error executing query, currentState RUNNING, java.lang.RuntimeException: Compaction failed. Please check logs for more info. Exception in compaction java.lang.Exception: Compaction Failure in Merger Rdd. at scala.sys.package$.error(package.scala:27) at org.apache.spark.sql.execution.command.AlterTableCompaction.processData(carbonTableSchema.scala:184) at org.apache.spark.sql.execution.command.AlterTableCompaction.run(carbonTableSchema.scala:139) at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:58) at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:56) at org.apache.spark.sql.execution.command.ExecutedCommandExec.doExecute(commands.scala:74) at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:114) at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:114) at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:135) at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:132) at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:113) at org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:87) at org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:87) at org.apache.spark.sql.Dataset.<init>(Dataset.scala:185) at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:64) at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:592) at org.apache.spark.sql.SQLContext.sql(SQLContext.scala:699) at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.org$apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute(SparkExecuteStatementOperation.scala:220) at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:163) at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:160) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1698) at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1.run(SparkExecuteStatementOperation.scala:173) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:748) 17/08/30 15:12:31 ERROR SparkExecuteStatementOperation: Error running hive query: org.apache.hive.service.cli.HiveSQLException: java.lang.RuntimeException: Compaction failed. Please check logs for more info. Exception in compaction java.lang.Exception: Compaction Failure in Merger Rdd. at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.org$apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute(SparkExecuteStatementOperation.scala:258) at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:163) at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:160) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1698) at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1.run(SparkExecuteStatementOperation.scala:173) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:748) > Major and Minor Compaction Failing > ---------------------------------- > > Key: CARBONDATA-1422 > URL: https://issues.apache.org/jira/browse/CARBONDATA-1422 > Project: CarbonData > Issue Type: Bug > Affects Versions: 1.2.0 > Reporter: Pallavi Singh > Fix For: 1.2.0 > > > The major and minor compaction is failing. > The Compaction is failing in the default scenario where the table property dictionary_include is not specified which is the default behaviour. > Please find the error logs below : > 0: jdbc:hive2://localhost:10000> show segments for table uniqdata; > +--------------------+----------+--------------------------+--------------------------+--+ > | SegmentSequenceId | Status | Load Start Time | Load End Time | > +--------------------+----------+--------------------------+--------------------------+--+ > | 9 | Success | 2017-08-29 11:17:29.927 | 2017-08-29 11:17:30.555 | > | 8 | Success | 2017-08-29 11:17:27.572 | 2017-08-29 11:17:28.363 | > | 7 | Success | 2017-08-29 11:17:23.583 | 2017-08-29 11:17:25.112 | > | 6 | Success | 2017-08-29 11:17:07.966 | 2017-08-29 11:17:09.322 | > | 5 | Success | 2017-08-29 10:38:15.727 | 2017-08-29 10:38:16.548 | > | 4 | Success | 2017-08-29 10:37:13.053 | 2017-08-29 10:37:13.888 | > | 3 | Success | 2017-08-29 10:36:57.851 | 2017-08-29 10:36:59.08 | > | 2 | Success | 2017-08-29 10:36:49.439 | 2017-08-29 10:36:50.373 | > | 1 | Success | 2017-08-29 10:36:37.365 | 2017-08-29 10:36:38.768 | > | 0 | Success | 2017-08-29 10:36:21.011 | 2017-08-29 10:36:26.1 | > +--------------------+----------+--------------------------+--------------------------+--+ > 10 rows selected (0.081 seconds) > 0: jdbc:hive2://localhost:10000> ALTER TABLE uniqdata COMPACT 'MINOR'; > Error: java.lang.RuntimeException: Compaction failed. Please check logs for more info. Exception in compaction java.lang.Exception: Compaction Failure in Merger Rdd. (state=,code=0) > 0: jdbc:hive2://localhost:10000> ALTER TABLE uniqdata COMPACT 'MAJOR'; > Error: java.lang.RuntimeException: Compaction failed. Please check logs for more info. Exception in compaction java.lang.Exception: Compaction Failure in Merger Rdd. (state=,code=0) > 0: jdbc:hive2://localhost:10000> -- This message was sent by Atlassian JIRA (v6.4.14#64029) |
Free forum by Nabble | Edit this page |