[jira] [Created] (CARBONDATA-1540) Memory issue while executing complex data type queries on cluster

classic Classic list List threaded Threaded
1 message Options
Reply | Threaded
Open this post in threaded view
|

[jira] [Created] (CARBONDATA-1540) Memory issue while executing complex data type queries on cluster

Akash R Nilugal (Jira)
Vandana Yadav created CARBONDATA-1540:
-----------------------------------------

             Summary: Memory issue while executing complex data type queries on cluster
                 Key: CARBONDATA-1540
                 URL: https://issues.apache.org/jira/browse/CARBONDATA-1540
             Project: CarbonData
          Issue Type: Bug
          Components: data-query
    Affects Versions: 1.2.0
         Environment: spark 2.1
            Reporter: Vandana Yadav
         Attachments: Array.csv

Memory issue while executing complex data type queries on the cluster:

Steps to reproduce:

1) Create COmplex data type table:
create table Array_com (CUST_ID string, YEAR int, MONTH int, AGE int, GENDER string, EDUCATED string, IS_MARRIED string, ARRAY_INT array<int>,ARRAY_STRING array<string>,ARRAY_DATE array<timestamp>,CARD_COUNT int,DEBIT_COUNT int, CREDIT_COUNT int, DEPOSIT double, HQ_DEPOSIT double) STORED BY 'org.apache.carbondata.format';

2) Load Data into the table:
LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/complex/Array.csv' INTO table Array_com  options ('DELIMITER'=',', 'QUOTECHAR'='"', 'FILEHEADER'='CUST_ID,YEAR,MONTH,AGE,GENDER,EDUCATED,IS_MARRIED,ARRAY_INT,ARRAY_STRING,ARRAY_DATE,CARD_COUNT,DEBIT_COUNT,CREDIT_COUNT,DEPOSIT,HQ_DEPOSIT','COMPLEX_DELIMITER_LEVEL_1'='$');

3) Execute the Select Query:
select array_int[0], array_int[0]+ 10 as a  from array_com

Expected Result: select query should display the correct result.

Actual Result: Error: org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 3.0 failed 4 times, most recent failure: Lost task 0.3 in stage 3.0 (TID 7, 148.251.7.173, executor 3): ExecutorLostFailure (executor 3 exited caused by one of the running tasks) Reason: Remote RPC client disassociated. Likely due to containers exceeding thresholds, or network issues. Check driver logs for WARN messages.
Driver stacktrace: (state=,code=0)


Thrift server Log:
17/10/09 11:03:33 INFO SparkExecuteStatementOperation: Running query 'select array_int[0], array_int[0]+ 10 as a  from array_com' with 5e2f2e3e-e737-496e-bb6f-31269aaed2be
17/10/09 11:03:33 INFO CarbonSparkSqlParser: Parsing command: select array_int[0], array_int[0]+ 10 as a  from array_com
17/10/09 11:03:33 INFO HiveMetaStore: 7: get_table : db=default tbl=array_com
17/10/09 11:03:33 INFO audit: ugi=root ip=unknown-ip-addr cmd=get_table : db=default tbl=array_com
17/10/09 11:03:33 INFO HiveMetaStore: 7: Opening raw store with implemenation class:org.apache.hadoop.hive.metastore.ObjectStore
17/10/09 11:03:33 INFO ObjectStore: ObjectStore, initialize called
17/10/09 11:03:33 INFO Query: Reading in results for query "org.datanucleus.store.rdbms.query.SQLQuery@0" since the connection used is closing
17/10/09 11:03:33 INFO MetaStoreDirectSql: Using direct SQL, underlying DB is DERBY
17/10/09 11:03:33 INFO ObjectStore: Initialized ObjectStore
17/10/09 11:03:33 INFO CatalystSqlParser: Parsing command: array<string>
17/10/09 11:03:33 INFO CarbonLateDecodeRule: pool-24-thread-6 Starting to optimize plan
17/10/09 11:03:33 STATISTIC QueryStatisticsRecorderImpl: Time taken for Carbon Optimizer to optimize: 15
17/10/09 11:03:33 INFO CarbonLateDecodeRule: pool-24-thread-6 Skip CarbonOptimizer
17/10/09 11:03:33 INFO CodeGenerator: Code generated in 15.017295 ms
17/10/09 11:03:33 INFO BlockManagerInfo: Removed broadcast_0_piece0 on 46.4.88.233:44387 in memory (size: 23.5 KB, free: 2004.6 MB)
17/10/09 11:03:33 INFO BlockManagerInfo: Removed broadcast_0_piece0 on 176.9.29.112:42871 in memory (size: 23.5 KB, free: 3.6 GB)
17/10/09 11:03:33 INFO ContextCleaner: Cleaned accumulator 0
17/10/09 11:03:33 INFO ContextCleaner: Cleaned accumulator 1
17/10/09 11:03:33 INFO ContextCleaner: Cleaned shuffle 0
17/10/09 11:03:33 INFO BlockManagerInfo: Removed broadcast_1_piece0 on 46.4.88.233:44387 in memory (size: 10.2 KB, free: 2004.6 MB)
17/10/09 11:03:33 INFO BlockManagerInfo: Removed broadcast_1_piece0 on 176.9.29.112:42871 in memory (size: 10.2 KB, free: 3.6 GB)
17/10/09 11:03:33 INFO TableInfo: pool-24-thread-6 Table block size not specified for default_array_com. Therefore considering the default value 1024 MB
17/10/09 11:03:33 INFO BlockManagerInfo: Removed broadcast_2_piece0 on 46.4.88.233:44387 in memory (size: 3.8 KB, free: 2004.6 MB)
17/10/09 11:03:33 INFO BlockManagerInfo: Removed broadcast_2_piece0 on 176.9.29.112:42871 in memory (size: 3.8 KB, free: 3.6 GB)
17/10/09 11:03:33 INFO BlockManagerInfo: Removed broadcast_3_piece0 on 46.4.88.233:44387 in memory (size: 23.2 KB, free: 2004.6 MB)
17/10/09 11:03:33 INFO BlockManagerInfo: Removed broadcast_3_piece0 on 46.4.88.233:36341 in memory (size: 23.2 KB, free: 3.6 GB)
17/10/09 11:03:33 INFO BlockManagerInfo: Removed broadcast_4_piece0 on 46.4.88.233:44387 in memory (size: 7.3 KB, free: 2004.6 MB)
17/10/09 11:03:33 INFO BlockManagerInfo: Removed broadcast_4_piece0 on 46.4.88.233:36341 in memory (size: 7.3 KB, free: 3.6 GB)
17/10/09 11:03:33 INFO UnsafeMemoryManager: pool-24-thread-6 Working Memory manager is created with size 536870912 with org.apache.carbondata.core.memory.UnsafeMemoryAllocator@750f4066
17/10/09 11:03:33 INFO UnsafeMemoryManager: pool-24-thread-6 Memory block (org.apache.carbondata.core.memory.MemoryBlock@2c5c1fce) is created with size 8388608. Total memory used 8388608Bytes, left 528482304Bytes
17/10/09 11:03:33 INFO UnsafeMemoryManager: pool-24-thread-6 Memory block (org.apache.carbondata.core.memory.MemoryBlock@55d7a9fc) is created with size 511. Total memory used 8389119Bytes, left 528481793Bytes
17/10/09 11:03:33 INFO UnsafeMemoryManager: pool-24-thread-6 Freeing memory of size: 8388608available memory:  536870401
17/10/09 11:03:33 STATISTIC DriverQueryStatisticsRecorderImpl: Print query statistic for query id: 431867348397223
+--------+--------------------+---------------------+------------------------+
|  Module|      Operation Step|     Total Query Cost|              Query Cost|
+--------+--------------------+---------------------+------------------------+
|  Driver|  Load blocks driver|                     |                     87 |
|        +--------------------+                     +------------------------+
|    Part|    Block allocation|                  88 |                      0 |
|        +--------------------+                     +------------------------+
|        |Block identification|                     |                      1 |
+--------+--------------------+---------------------+------------------------+

17/10/09 11:03:33 INFO CarbonScanRDD:
 Identified no.of.blocks: 1,
 no.of.tasks: 1,
 no.of.nodes: 0,
 parallelism: 24
       
17/10/09 11:03:33 INFO SparkContext: Starting job: run at AccessController.java:0
17/10/09 11:03:33 INFO DAGScheduler: Got job 2 (run at AccessController.java:0) with 1 output partitions
17/10/09 11:03:33 INFO DAGScheduler: Final stage: ResultStage 3 (run at AccessController.java:0)
17/10/09 11:03:33 INFO DAGScheduler: Parents of final stage: List()
17/10/09 11:03:33 INFO DAGScheduler: Missing parents: List()
17/10/09 11:03:33 INFO DAGScheduler: Submitting ResultStage 3 (MapPartitionsRDD[19] at run at AccessController.java:0), which has no missing parents
17/10/09 11:03:33 INFO MemoryStore: Block broadcast_5 stored as values in memory (estimated size 10.3 KB, free 2004.6 MB)
17/10/09 11:03:33 INFO MemoryStore: Block broadcast_5_piece0 stored as bytes in memory (estimated size 5.3 KB, free 2004.6 MB)
17/10/09 11:03:33 INFO BlockManagerInfo: Added broadcast_5_piece0 in memory on 46.4.88.233:44387 (size: 5.3 KB, free: 2004.6 MB)
17/10/09 11:03:33 INFO SparkContext: Created broadcast 5 from broadcast at DAGScheduler.scala:996
17/10/09 11:03:33 INFO DAGScheduler: Submitting 1 missing tasks from ResultStage 3 (MapPartitionsRDD[19] at run at AccessController.java:0)
17/10/09 11:03:33 INFO TaskSchedulerImpl: Adding task set 3.0 with 1 tasks
17/10/09 11:03:33 INFO TaskSetManager: Starting task 0.0 in stage 3.0 (TID 4, 46.4.88.233, executor 1, partition 0, ANY, 6807 bytes)
17/10/09 11:03:33 INFO BlockManagerInfo: Added broadcast_5_piece0 in memory on 46.4.88.233:36341 (size: 5.3 KB, free: 3.6 GB)
17/10/09 11:05:34 ERROR TaskSchedulerImpl: Lost executor 1 on 46.4.88.233: Remote RPC client disassociated. Likely due to containers exceeding thresholds, or network issues. Check driver logs for WARN messages.
17/10/09 11:05:34 INFO StandaloneAppClient$ClientEndpoint: Executor updated: app-20171009104414-0003/1 is now EXITED (Command exited with code 134)
17/10/09 11:05:34 INFO StandaloneSchedulerBackend: Executor app-20171009104414-0003/1 removed: Command exited with code 134
17/10/09 11:05:34 WARN TaskSetManager: Lost task 0.0 in stage 3.0 (TID 4, 46.4.88.233, executor 1): ExecutorLostFailure (executor 1 exited caused by one of the running tasks) Reason: Remote RPC client disassociated. Likely due to containers exceeding thresholds, or network issues. Check driver logs for WARN messages.
17/10/09 11:05:34 INFO DAGScheduler: Executor lost: 1 (epoch 1)
17/10/09 11:05:34 INFO BlockManagerMaster: Removal of executor 1 requested
17/10/09 11:05:34 INFO BlockManagerMasterEndpoint: Trying to remove executor 1 from BlockManagerMaster.
17/10/09 11:05:34 INFO CoarseGrainedSchedulerBackend$DriverEndpoint: Asked to remove non-existent executor 1
17/10/09 11:05:34 INFO BlockManagerMasterEndpoint: Removing block manager BlockManagerId(1, 46.4.88.233, 36341, None)
17/10/09 11:05:34 INFO BlockManagerMasterEndpoint: Trying to remove executor 1 from BlockManagerMaster.
17/10/09 11:05:34 INFO TaskSetManager: Starting task 0.1 in stage 3.0 (TID 5, 148.251.7.173, executor 0, partition 0, ANY, 6807 bytes)
17/10/09 11:05:34 INFO BlockManagerMaster: Removed 1 successfully in removeExecutor
17/10/09 11:05:34 INFO DAGScheduler: Shuffle files lost for executor: 1 (epoch 1)
17/10/09 11:05:34 INFO BlockManagerInfo: Added broadcast_5_piece0 in memory on 148.251.7.173:37951 (size: 5.3 KB, free: 3.6 GB)
17/10/09 11:07:35 ERROR TaskSchedulerImpl: Lost executor 0 on 148.251.7.173: Remote RPC client disassociated. Likely due to containers exceeding thresholds, or network issues. Check driver logs for WARN messages.
17/10/09 11:07:35 WARN TaskSetManager: Lost task 0.1 in stage 3.0 (TID 5, 148.251.7.173, executor 0): ExecutorLostFailure (executor 0 exited caused by one of the running tasks) Reason: Remote RPC client disassociated. Likely due to containers exceeding thresholds, or network issues. Check driver logs for WARN messages.
17/10/09 11:07:35 INFO DAGScheduler: Executor lost: 0 (epoch 2)
17/10/09 11:07:35 INFO BlockManagerMasterEndpoint: Trying to remove executor 0 from BlockManagerMaster.
17/10/09 11:07:35 INFO BlockManagerMasterEndpoint: Removing block manager BlockManagerId(0, 148.251.7.173, 37951, None)
17/10/09 11:07:35 INFO BlockManagerMaster: Removed 0 successfully in removeExecutor
17/10/09 11:07:35 INFO DAGScheduler: Shuffle files lost for executor: 0 (epoch 2)
17/10/09 11:07:35 INFO TaskSetManager: Starting task 0.2 in stage 3.0 (TID 6, 176.9.29.112, executor 2, partition 0, ANY, 6807 bytes)
17/10/09 11:07:35 INFO StandaloneAppClient$ClientEndpoint: Executor updated: app-20171009104414-0003/0 is now EXITED (Command exited with code 134)
17/10/09 11:07:35 INFO StandaloneSchedulerBackend: Executor app-20171009104414-0003/0 removed: Command exited with code 134
17/10/09 11:07:35 INFO BlockManagerMasterEndpoint: Trying to remove executor 0 from BlockManagerMaster.
17/10/09 11:07:35 INFO BlockManagerMaster: Removal of executor 0 requested
17/10/09 11:07:35 INFO CoarseGrainedSchedulerBackend$DriverEndpoint: Asked to remove non-existent executor 0
17/10/09 11:07:35 INFO BlockManagerInfo: Added broadcast_5_piece0 in memory on 176.9.29.112:42871 (size: 5.3 KB, free: 3.6 GB)
17/10/09 11:09:36 ERROR TaskSchedulerImpl: Lost executor 2 on 176.9.29.112: Remote RPC client disassociated. Likely due to containers exceeding thresholds, or network issues. Check driver logs for WARN messages.
17/10/09 11:09:36 WARN TaskSetManager: Lost task 0.2 in stage 3.0 (TID 6, 176.9.29.112, executor 2): ExecutorLostFailure (executor 2 exited caused by one of the running tasks) Reason: Remote RPC client disassociated. Likely due to containers exceeding thresholds, or network issues. Check driver logs for WARN messages.
17/10/09 11:09:36 INFO DAGScheduler: Executor lost: 2 (epoch 3)
17/10/09 11:09:36 INFO BlockManagerMasterEndpoint: Trying to remove executor 2 from BlockManagerMaster.
17/10/09 11:09:36 INFO BlockManagerMasterEndpoint: Removing block manager BlockManagerId(2, 176.9.29.112, 42871, None)
17/10/09 11:09:36 INFO BlockManagerMaster: Removed 2 successfully in removeExecutor
17/10/09 11:09:36 INFO DAGScheduler: Shuffle files lost for executor: 2 (epoch 3)
17/10/09 11:09:36 INFO StandaloneAppClient$ClientEndpoint: Executor updated: app-20171009104414-0003/2 is now EXITED (Command exited with code 134)
17/10/09 11:09:36 INFO StandaloneSchedulerBackend: Executor app-20171009104414-0003/2 removed: Command exited with code 134
17/10/09 11:09:36 INFO BlockManagerMaster: Removal of executor 2 requested
17/10/09 11:09:36 INFO BlockManagerMasterEndpoint: Trying to remove executor 2 from BlockManagerMaster.
17/10/09 11:09:36 INFO CoarseGrainedSchedulerBackend$DriverEndpoint: Asked to remove non-existent executor 2
17/10/09 11:09:36 INFO StandaloneAppClient$ClientEndpoint: Executor added: app-20171009104414-0003/3 on worker-20171009095843-148.251.7.173-33995 (148.251.7.173:33995) with 8 cores
17/10/09 11:09:36 INFO StandaloneSchedulerBackend: Granted executor ID app-20171009104414-0003/3 on hostPort 148.251.7.173:33995 with 8 cores, 7.0 GB RAM
17/10/09 11:09:36 INFO StandaloneAppClient$ClientEndpoint: Executor updated: app-20171009104414-0003/3 is now RUNNING
17/10/09 11:09:37 INFO CoarseGrainedSchedulerBackend$DriverEndpoint: Registered executor NettyRpcEndpointRef(null) (148.251.7.173:56362) with ID 3
17/10/09 11:09:37 INFO TaskSetManager: Starting task 0.3 in stage 3.0 (TID 7, 148.251.7.173, executor 3, partition 0, ANY, 6807 bytes)
17/10/09 11:09:37 INFO BlockManagerMasterEndpoint: Registering block manager 148.251.7.173:40003 with 3.6 GB RAM, BlockManagerId(3, 148.251.7.173, 40003, None)
17/10/09 11:09:38 INFO BlockManagerInfo: Added broadcast_5_piece0 in memory on 148.251.7.173:40003 (size: 5.3 KB, free: 3.6 GB)
17/10/09 11:11:39 ERROR TaskSchedulerImpl: Lost executor 3 on 148.251.7.173: Remote RPC client disassociated. Likely due to containers exceeding thresholds, or network issues. Check driver logs for WARN messages.
17/10/09 11:11:39 WARN TaskSetManager: Lost task 0.3 in stage 3.0 (TID 7, 148.251.7.173, executor 3): ExecutorLostFailure (executor 3 exited caused by one of the running tasks) Reason: Remote RPC client disassociated. Likely due to containers exceeding thresholds, or network issues. Check driver logs for WARN messages.
17/10/09 11:11:39 ERROR TaskSetManager: Task 0 in stage 3.0 failed 4 times; aborting job
17/10/09 11:11:39 INFO TaskSchedulerImpl: Removed TaskSet 3.0, whose tasks have all completed, from pool
17/10/09 11:11:39 INFO StandaloneAppClient$ClientEndpoint: Executor updated: app-20171009104414-0003/3 is now EXITED (Command exited with code 134)
17/10/09 11:11:39 INFO StandaloneSchedulerBackend: Executor app-20171009104414-0003/3 removed: Command exited with code 134
17/10/09 11:11:39 INFO BlockManagerMaster: Removal of executor 3 requested
17/10/09 11:11:39 INFO BlockManagerMasterEndpoint: Trying to remove executor 3 from BlockManagerMaster.
17/10/09 11:11:39 INFO CoarseGrainedSchedulerBackend$DriverEndpoint: Asked to remove non-existent executor 3
17/10/09 11:11:39 INFO BlockManagerMasterEndpoint: Removing block manager BlockManagerId(3, 148.251.7.173, 40003, None)
17/10/09 11:11:39 INFO StandaloneAppClient$ClientEndpoint: Executor added: app-20171009104414-0003/4 on worker-20171009095843-148.251.7.173-33995 (148.251.7.173:33995) with 8 cores
17/10/09 11:11:39 INFO StandaloneSchedulerBackend: Granted executor ID app-20171009104414-0003/4 on hostPort 148.251.7.173:33995 with 8 cores, 7.0 GB RAM
17/10/09 11:11:39 INFO TaskSchedulerImpl: Cancelling stage 3
17/10/09 11:11:39 INFO DAGScheduler: ResultStage 3 (run at AccessController.java:0) failed in 485.691 s due to Job aborted due to stage failure: Task 0 in stage 3.0 failed 4 times, most recent failure: Lost task 0.3 in stage 3.0 (TID 7, 148.251.7.173, executor 3): ExecutorLostFailure (executor 3 exited caused by one of the running tasks) Reason: Remote RPC client disassociated. Likely due to containers exceeding thresholds, or network issues. Check driver logs for WARN messages.
Driver stacktrace:
17/10/09 11:11:39 INFO DAGScheduler: Job 2 failed: run at AccessController.java:0, took 485.699162 s
17/10/09 11:11:39 INFO DAGScheduler: Executor lost: 3 (epoch 4)
17/10/09 11:11:39 INFO BlockManagerMasterEndpoint: Trying to remove executor 3 from BlockManagerMaster.
17/10/09 11:11:39 INFO BlockManagerMaster: Removed 3 successfully in removeExecutor
17/10/09 11:11:39 INFO DAGScheduler: Shuffle files lost for executor: 3 (epoch 4)
17/10/09 11:11:39 ERROR SparkExecuteStatementOperation: Error executing query, currentState RUNNING,
org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 3.0 failed 4 times, most recent failure: Lost task 0.3 in stage 3.0 (TID 7, 148.251.7.173, executor 3): ExecutorLostFailure (executor 3 exited caused by one of the running tasks) Reason: Remote RPC client disassociated. Likely due to containers exceeding thresholds, or network issues. Check driver logs for WARN messages.
Driver stacktrace:
        at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1435)
        at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1423)
        at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1422)
        at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
        at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48)
        at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1422)
        at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:802)
        at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:802)
        at scala.Option.foreach(Option.scala:257)
        at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:802)
        at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1650)
        at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1605)
        at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1594)
        at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)
        at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:628)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:1918)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:1931)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:1944)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:1958)
        at org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.scala:935)
        at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
        at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
        at org.apache.spark.rdd.RDD.withScope(RDD.scala:362)
        at org.apache.spark.rdd.RDD.collect(RDD.scala:934)
        at org.apache.spark.sql.execution.SparkPlan.executeCollect(SparkPlan.scala:275)
        at org.apache.spark.sql.Dataset$$anonfun$org$apache$spark$sql$Dataset$$execute$1$1.apply(Dataset.scala:2371)
        at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:57)
        at org.apache.spark.sql.Dataset.withNewExecutionId(Dataset.scala:2765)
        at org.apache.spark.sql.Dataset.org$apache$spark$sql$Dataset$$execute$1(Dataset.scala:2370)
        at org.apache.spark.sql.Dataset$$anonfun$org$apache$spark$sql$Dataset$$collect$1.apply(Dataset.scala:2375)
        at org.apache.spark.sql.Dataset$$anonfun$org$apache$spark$sql$Dataset$$collect$1.apply(Dataset.scala:2375)
        at org.apache.spark.sql.Dataset.withCallback(Dataset.scala:2778)
        at org.apache.spark.sql.Dataset.org$apache$spark$sql$Dataset$$collect(Dataset.scala:2375)
        at org.apache.spark.sql.Dataset.collect(Dataset.scala:2351)
        at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.org$apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute(SparkExecuteStatementOperation.scala:235)
        at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:163)
        at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:160)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:422)
        at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1698)
        at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1.run(SparkExecuteStatementOperation.scala:173)
        at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
        at java.util.concurrent.FutureTask.run(FutureTask.java:266)
        at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
        at java.lang.Thread.run(Thread.java:748)
17/10/09 11:11:39 INFO StandaloneAppClient$ClientEndpoint: Executor updated: app-20171009104414-0003/4 is now RUNNING
17/10/09 11:11:39 ERROR SparkExecuteStatementOperation: Error running hive query:
org.apache.hive.service.cli.HiveSQLException: org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 3.0 failed 4 times, most recent failure: Lost task 0.3 in stage 3.0 (TID 7, 148.251.7.173, executor 3): ExecutorLostFailure (executor 3 exited caused by one of the running tasks) Reason: Remote RPC client disassociated. Likely due to containers exceeding thresholds, or network issues. Check driver logs for WARN messages.
Driver stacktrace:
        at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.org$apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute(SparkExecuteStatementOperation.scala:258)
        at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:163)
        at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:160)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:422)
        at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1698)
        at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1.run(SparkExecuteStatementOperation.scala:173)
        at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
        at java.util.concurrent.FutureTask.run(FutureTask.java:266)
        at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
        at java.lang.Thread.run(Thread.java:748)
17/10/09 11:11:41 INFO CoarseGrainedSchedulerBackend$DriverEndpoint: Registered executor NettyRpcEndpointRef(null) (148.251.7.173:56378) with ID 4
17/10/09 11:11:41 INFO BlockManagerMasterEndpoint: Registering block manager 148.251.7.173:40125 with 3.6 GB RAM, BlockManagerId(4, 148.251.7.173, 40125, None)
17/10/09 11:14:14 INFO BlockManagerInfo: Removed broadcast_5_piece0 on 46.4.88.233:44387 in memory (size: 5.3 KB, free: 2004.6 MB)
17/10/09 11:14:14 INFO ContextCleaner: Cleaned accumulator 171
17/10/09 11:14:14 INFO ContextCleaner: Cleaned accumulator 170



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)