carbondata-0.2 load data failed in yarn molde

classic Classic list List threaded Threaded
7 messages Options
Reply | Threaded
Open this post in threaded view
|

carbondata-0.2 load data failed in yarn molde

geda
This post was updated on .
hello:
i test  data in spark locak model ,then load data inpath to table ,works well.
but when i use yarn-client modle,  with 1w rows , size :940k ,but error happend ,there is no lock find in  tmp dir ,i don't know how to debug,help.thanks.
spark1.6 hadoop 2.7|2.6 carbondata 0.2
local mode: run ok
$SPARK_HOME/bin/spark-shell --master local[4]  --jars /usr/local/spark/lib/carbondata_2.10-0.2.0-incubating-shade-hadoop2.7.1.jar


yarn command : run bad
 $SPARK_HOME/bin/spark-shell --verbose  --master yarn-client --driver-memory 1G --driver-cores 1   --executor-memory 4G --num-executors 5 --executor-cores 1 --conf "spark.executor.extraJavaOptions=-XX:NewRatio=2 -XX:PermSize=512m -XX:MaxPermSize=512m -XX:SurvivorRatio=6  -verbose:gc -XX:-PrintGCDetails -XX:+PrintGCTimeStamps " --conf "spark.driver.extraJavaOptions=-XX:MaxPermSize=512m -XX:PermSize=512m"  --conf spark.yarn.driver.memoryOverhead=1024 --conf spark.yarn.executor.memoryOverhead=3096    --jars /usr/local/spark/lib/carbondata_2.10-0.2.0-incubating-shade-hadoop2.7.1.jar

import java.io._
import org.apache.hadoop.hive.conf.HiveConf
import org.apache.spark.sql.CarbonContext
val storePath = "hdfs://test:8020/usr/carbondata/store"
val cc = new CarbonContext(sc, storePath)
cc.setConf(HiveConf.ConfVars.HIVECHECKFILEFORMAT.varname, "false")
cc.setConf("carbon.kettle.home","/usr/local/spark/carbondata/carbonplugins")
cc.sql("CREATE TABLE `LINEORDER3` (   LO_ORDERKEY       bigint,   LO_LINENUMBER     int,   LO_CUSTKEY        bigint,   LO_PARTKEY        bigint,   LO_SUPPKEY        bigint,   LO_ORDERDATE      int,   LO_ORDERPRIOTITY  string,   LO_SHIPPRIOTITY   int,   LO_QUANTITY       int,   LO_EXTENDEDPRICE  int,   LO_ORDTOTALPRICE  int,   LO_DISCOUNT       int,   LO_REVENUE        int,   LO_SUPPLYCOST     int,   LO_TAX            int,   LO_COMMITDATE     int,   LO_SHIPMODE       string ) STORED BY 'carbondata'")
cc.sql(s"load data local inpath 'hdfs://test:8020/tmp/lineorder_1w.tbl'  into table lineorder3 options('DELIMITER'='|', 'FILEHEADER'='LO_ORDERKEY, LO_LINENUMBER, LO_CUSTKEY, LO_PARTKEY , LO_SUPPKEY , LO_ORDERDATE ,   LO_ORDERPRIOTITY ,   LO_SHIPPRIOTITY , LO_QUANTITY ,LO_EXTENDEDPRICE , LO_ORDTOTALPRICE ,LO_DISCOUNT , LO_REVENUE  ,   LO_SUPPLYCOST,   LO_TAX, LO_COMMITDATE,   LO_SHIPMODE')")

org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 2.0 failed 4 times, most recent failure: Lost task 0.3 in stage 2.0 (TID 8, datanode03-bi-dev): java.lang.RuntimeException: Dictionary file lo_orderpriotity is locked for updation. Please try after some time
        at scala.sys.package$.error(package.scala:27)
        at org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
        at org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(CarbonGlobalDictionaryRDD.scala:293)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
        at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
        at org.apache.spark.scheduler.Task.run(Task.scala:89)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
        at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
        at java.lang.Thread.run(Thread.java:745)

Driver stacktrace:
        at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1431)
        at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1419)
        at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1418)
        at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
        at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
        at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1418)
        at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
        at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
        at scala.Option.foreach(Option.scala:236)
        at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:799)
        at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1640)
        at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1599)
        at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1588)
        at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)
        at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:620)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:1832)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:1845)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:1858)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:1929)
        at org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.scala:927)
        at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
        at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111)
        at org.apache.spark.rdd.RDD.withScope(RDD.scala:316)
        at org.apache.spark.rdd.RDD.collect(RDD.scala:926)
        at org.apache.carbondata.spark.util.GlobalDictionaryUtil$.generateGlobalDictionary(GlobalDictionaryUtil.scala:800)
        at org.apache.spark.sql.execution.command.LoadTableUsingKettle.run(carbonTableSchema.scala:1197)
        at org.apache.spark.sql.execution.command.LoadTable.run(carbonTableSchema.scala:1036)
        at org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult$lzycompute(commands.scala:58)
        at org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult(commands.scala:56)
        at org.apache.spark.sql.execution.ExecutedCommand.doExecute(commands.scala:70)
        at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:132)
        at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:130)
        at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
        at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:130)
        at org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:55)
        at org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:55)
        at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:145)
        at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:130)
        at org.apache.carbondata.spark.rdd.CarbonDataFrameRDD.<init>(CarbonDataFrameRDD.scala:23)
        at org.apache.spark.sql.CarbonContext.sql(CarbonContext.scala:137)
        at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:37)
        at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:42)
        at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:44)
        at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:46)
        at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:48)
        at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:50)
        at $line32.$read$$iwC$$iwC$$iwC$$iwC.<init>(<console>:52)
        at $line32.$read$$iwC$$iwC$$iwC.<init>(<console>:54)
        at $line32.$read$$iwC$$iwC.<init>(<console>:56)
        at $line32.$read$$iwC.<init>(<console>:58)
        at $line32.$read.<init>(<console>:60)
        at $line32.$read$.<init>(<console>:64)
        at $line32.$read$.<clinit>(<console>)
        at $line32.$eval$.<init>(<console>:7)
        at $line32.$eval$.<clinit>(<console>)
        at $line32.$eval.$print(<console>)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
        at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:606)
        at org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(SparkIMain.scala:1065)
        at org.apache.spark.repl.SparkIMain$Request.loadAndRun(SparkIMain.scala:1346)
        at org.apache.spark.repl.SparkIMain.loadAndRunReq$1(SparkIMain.scala:840)
        at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:871)
        at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:819)
        at org.apache.spark.repl.SparkILoop.reallyInterpret$1(SparkILoop.scala:857)
        at org.apache.spark.repl.SparkILoop.interpretStartingWith(SparkILoop.scala:902)
        at org.apache.spark.repl.SparkILoop.command(SparkILoop.scala:814)
        at org.apache.spark.repl.SparkILoop.processLine$1(SparkILoop.scala:657)
        at org.apache.spark.repl.SparkILoop.innerLoop$1(SparkILoop.scala:665)
        at org.apache.spark.repl.SparkILoop.org$apache$spark$repl$SparkILoop$$loop(SparkILoop.scala:670)
        at org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl$SparkILoop$$process$1.apply$mcZ$sp(SparkILoop.scala:997)
        at org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945)
        at org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945)
        at scala.tools.nsc.util.ScalaClassLoader$.savingContextLoader(ScalaClassLoader.scala:135)
        at org.apache.spark.repl.SparkILoop.org$apache$spark$repl$SparkILoop$$process(SparkILoop.scala:945)
        at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:1059)
        at org.apache.spark.repl.Main$.main(Main.scala:31)
        at org.apache.spark.repl.Main.main(Main.scala)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
        at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:606)
        at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
        at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181)
        at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206)
        at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
        at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
Caused by: java.lang.RuntimeException: Dictionary file lo_orderpriotity is locked for updation. Please try after some time
        at scala.sys.package$.error(package.scala:27)
Reply | Threaded
Open this post in threaded view
|

Re: carbondata-0.2 load data failed in yarn molde

Liang Chen
Administrator
Hi

Have you solved this issue after applying new configurations?

Regards
Liang
geda wrote
hello:
i test  data in spark locak model ,then load data inpath to table ,works well.
but when i use yarn-client modle,  with 1w rows , size :940k ,but error happend ,there is no lock find in  tmp dir ,i don't know how to debug,help.thanks.
spark1.6 hadoop 2.7|2.6 carbondata 0.2
local mode: run ok
$SPARK_HOME/bin/spark-shell --master local[4]  --jars /usr/local/spark/lib/carbondata_2.10-0.2.0-incubating-shade-hadoop2.7.1.jar


yarn command : run bad
 $SPARK_HOME/bin/spark-shell --verbose  --master yarn-client --driver-memory 1G --driver-cores 1   --executor-memory 4G --num-executors 5 --executor-cores 1 --conf "spark.executor.extraJavaOptions=-XX:NewRatio=2 -XX:PermSize=512m -XX:MaxPermSize=512m -XX:SurvivorRatio=6  -verbose:gc -XX:-PrintGCDetails -XX:+PrintGCTimeStamps " --conf "spark.driver.extraJavaOptions=-XX:MaxPermSize=512m -XX:PermSize=512m"  --conf spark.yarn.driver.memoryOverhead=1024 --conf spark.yarn.executor.memoryOverhead=3096    --jars /usr/local/spark/lib/carbondata_2.10-0.2.0-incubating-shade-hadoop2.7.1.jar

import java.io._
import org.apache.hadoop.hive.conf.HiveConf
import org.apache.spark.sql.CarbonContext
val storePath = "hdfs://test:8020/usr/carbondata/store"
val cc = new CarbonContext(sc, storePath)
cc.setConf(HiveConf.ConfVars.HIVECHECKFILEFORMAT.varname, "false")
cc.setConf("carbon.kettle.home","/usr/local/spark/carbondata/carbonplugins")
cc.sql("CREATE TABLE `LINEORDER3` (   LO_ORDERKEY       bigint,   LO_LINENUMBER     int,   LO_CUSTKEY        bigint,   LO_PARTKEY        bigint,   LO_SUPPKEY        bigint,   LO_ORDERDATE      int,   LO_ORDERPRIOTITY  string,   LO_SHIPPRIOTITY   int,   LO_QUANTITY       int,   LO_EXTENDEDPRICE  int,   LO_ORDTOTALPRICE  int,   LO_DISCOUNT       int,   LO_REVENUE        int,   LO_SUPPLYCOST     int,   LO_TAX            int,   LO_COMMITDATE     int,   LO_SHIPMODE       string ) STORED BY 'carbondata'")
cc.sql(s"load data local inpath 'hdfs://test:8020/tmp/lineorder_1w.tbl'  into table lineorder3 options('DELIMITER'='|', 'FILEHEADER'='LO_ORDERKEY, LO_LINENUMBER, LO_CUSTKEY, LO_PARTKEY , LO_SUPPKEY , LO_ORDERDATE ,   LO_ORDERPRIOTITY ,   LO_SHIPPRIOTITY , LO_QUANTITY ,LO_EXTENDEDPRICE , LO_ORDTOTALPRICE ,LO_DISCOUNT , LO_REVENUE  ,   LO_SUPPLYCOST,   LO_TAX, LO_COMMITDATE,   LO_SHIPMODE')")

org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 2.0 failed 4 times, most recent failure: Lost task 0.3 in stage 2.0 (TID 8, datanode03-bi-dev): java.lang.RuntimeException: Dictionary file lo_orderpriotity is locked for updation. Please try after some time
        at scala.sys.package$.error(package.scala:27)
        at org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
        at org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(CarbonGlobalDictionaryRDD.scala:293)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
        at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
        at org.apache.spark.scheduler.Task.run(Task.scala:89)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
        at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
        at java.lang.Thread.run(Thread.java:745)

Driver stacktrace:
        at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1431)
        at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1419)
        at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1418)
        at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
        at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
        at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1418)
        at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
        at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
        at scala.Option.foreach(Option.scala:236)
        at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:799)
        at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1640)
        at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1599)
        at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1588)
        at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)
        at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:620)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:1832)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:1845)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:1858)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:1929)
        at org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.scala:927)
        at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
        at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111)
        at org.apache.spark.rdd.RDD.withScope(RDD.scala:316)
        at org.apache.spark.rdd.RDD.collect(RDD.scala:926)
        at org.apache.carbondata.spark.util.GlobalDictionaryUtil$.generateGlobalDictionary(GlobalDictionaryUtil.scala:800)
        at org.apache.spark.sql.execution.command.LoadTableUsingKettle.run(carbonTableSchema.scala:1197)
        at org.apache.spark.sql.execution.command.LoadTable.run(carbonTableSchema.scala:1036)
        at org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult$lzycompute(commands.scala:58)
        at org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult(commands.scala:56)
        at org.apache.spark.sql.execution.ExecutedCommand.doExecute(commands.scala:70)
        at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:132)
        at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:130)
        at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
        at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:130)
        at org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:55)
        at org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:55)
        at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:145)
        at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:130)
        at org.apache.carbondata.spark.rdd.CarbonDataFrameRDD.<init>(CarbonDataFrameRDD.scala:23)
        at org.apache.spark.sql.CarbonContext.sql(CarbonContext.scala:137)
        at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:37)
        at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:42)
        at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:44)
        at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:46)
        at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:48)
        at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:50)
        at $line32.$read$$iwC$$iwC$$iwC$$iwC.<init>(<console>:52)
        at $line32.$read$$iwC$$iwC$$iwC.<init>(<console>:54)
        at $line32.$read$$iwC$$iwC.<init>(<console>:56)
        at $line32.$read$$iwC.<init>(<console>:58)
        at $line32.$read.<init>(<console>:60)
        at $line32.$read$.<init>(<console>:64)
        at $line32.$read$.<clinit>(<console>)
        at $line32.$eval$.<init>(<console>:7)
        at $line32.$eval$.<clinit>(<console>)
        at $line32.$eval.$print(<console>)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
        at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:606)
        at org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(SparkIMain.scala:1065)
        at org.apache.spark.repl.SparkIMain$Request.loadAndRun(SparkIMain.scala:1346)
        at org.apache.spark.repl.SparkIMain.loadAndRunReq$1(SparkIMain.scala:840)
        at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:871)
        at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:819)
        at org.apache.spark.repl.SparkILoop.reallyInterpret$1(SparkILoop.scala:857)
        at org.apache.spark.repl.SparkILoop.interpretStartingWith(SparkILoop.scala:902)
        at org.apache.spark.repl.SparkILoop.command(SparkILoop.scala:814)
        at org.apache.spark.repl.SparkILoop.processLine$1(SparkILoop.scala:657)
        at org.apache.spark.repl.SparkILoop.innerLoop$1(SparkILoop.scala:665)
        at org.apache.spark.repl.SparkILoop.org$apache$spark$repl$SparkILoop$$loop(SparkILoop.scala:670)
        at org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl$SparkILoop$$process$1.apply$mcZ$sp(SparkILoop.scala:997)
        at org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945)
        at org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945)
        at scala.tools.nsc.util.ScalaClassLoader$.savingContextLoader(ScalaClassLoader.scala:135)
        at org.apache.spark.repl.SparkILoop.org$apache$spark$repl$SparkILoop$$process(SparkILoop.scala:945)
        at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:1059)
        at org.apache.spark.repl.Main$.main(Main.scala:31)
        at org.apache.spark.repl.Main.main(Main.scala)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
        at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:606)
        at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
        at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181)
        at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206)
        at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
        at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
Caused by: java.lang.RuntimeException: Dictionary file lo_orderpriotity is locked for updation. Please try after some time
        at scala.sys.package$.error(package.scala:27)
Reply | Threaded
Open this post in threaded view
|

Re: carbondata-0.2 load data failed in yarn molde

geda
Yes,it work thanks。
follow wiki quick start,use spark-default.conf , should config  carbondata.properties
but  i  use spark-shell , not include carbon.properties in it.



2016-12-09 11:48 GMT+08:00 Liang Chen [via Apache CarbonData Mailing List archive] <[hidden email]>:
Hi

Have you solved this issue after applying new configurations?

Regards
Liang
geda wrote
hello:
i test  data in spark locak model ,then load data inpath to table ,works well.
but when i use yarn-client modle,  with 1w rows , size :940k ,but error happend ,there is no lock find in  tmp dir ,i don't know how to debug,help.thanks.
spark1.6 hadoop 2.7|2.6 carbondata 0.2
local mode: run ok
$SPARK_HOME/bin/spark-shell --master local[4]  --jars /usr/local/spark/lib/carbondata_2.10-0.2.0-incubating-shade-hadoop2.7.1.jar


yarn command : run bad
 $SPARK_HOME/bin/spark-shell --verbose  --master yarn-client --driver-memory 1G --driver-cores 1   --executor-memory 4G --num-executors 5 --executor-cores 1 --conf "spark.executor.extraJavaOptions=-XX:NewRatio=2 -XX:PermSize=512m -XX:MaxPermSize=512m -XX:SurvivorRatio=6  -verbose:gc -XX:-PrintGCDetails -XX:+PrintGCTimeStamps " --conf "spark.driver.extraJavaOptions=-XX:MaxPermSize=512m -XX:PermSize=512m"  --conf spark.yarn.driver.memoryOverhead=1024 --conf spark.yarn.executor.memoryOverhead=3096    --jars /usr/local/spark/lib/carbondata_2.10-0.2.0-incubating-shade-hadoop2.7.1.jar

import java.io._
import org.apache.hadoop.hive.conf.HiveConf
import org.apache.spark.sql.CarbonContext
val storePath = "hdfs://test:8020/usr/carbondata/store"
val cc = new CarbonContext(sc, storePath)
cc.setConf(HiveConf.ConfVars.HIVECHECKFILEFORMAT.varname, "false")
cc.setConf("carbon.kettle.home","/usr/local/spark/carbondata/carbonplugins")
cc.sql("CREATE TABLE `LINEORDER3` (   LO_ORDERKEY       bigint,   LO_LINENUMBER     int,   LO_CUSTKEY        bigint,   LO_PARTKEY        bigint,   LO_SUPPKEY        bigint,   LO_ORDERDATE      int,   LO_ORDERPRIOTITY  string,   LO_SHIPPRIOTITY   int,   LO_QUANTITY       int,   LO_EXTENDEDPRICE  int,   LO_ORDTOTALPRICE  int,   LO_DISCOUNT       int,   LO_REVENUE        int,   LO_SUPPLYCOST     int,   LO_TAX            int,   LO_COMMITDATE     int,   LO_SHIPMODE       string ) STORED BY 'carbondata'")
cc.sql(s"load data local inpath 'hdfs://test:8020/tmp/lineorder_1w.tbl'  into table lineorder3 options('DELIMITER'='|', 'FILEHEADER'='LO_ORDERKEY, LO_LINENUMBER, LO_CUSTKEY, LO_PARTKEY , LO_SUPPKEY , LO_ORDERDATE ,   LO_ORDERPRIOTITY ,   LO_SHIPPRIOTITY , LO_QUANTITY ,LO_EXTENDEDPRICE , LO_ORDTOTALPRICE ,LO_DISCOUNT , LO_REVENUE  ,   LO_SUPPLYCOST,   LO_TAX, LO_COMMITDATE,   LO_SHIPMODE')")

org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 2.0 failed 4 times, most recent failure: Lost task 0.3 in stage 2.0 (TID 8, datanode03-bi-dev): java.lang.RuntimeException: Dictionary file lo_orderpriotity is locked for updation. Please try after some time
        at scala.sys.package$.error(package.scala:27)
        at org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
        at org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(CarbonGlobalDictionaryRDD.scala:293)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
        at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
        at org.apache.spark.scheduler.Task.run(Task.scala:89)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
        at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
        at java.lang.Thread.run(Thread.java:745)

Driver stacktrace:
        at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1431)
        at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1419)
        at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1418)
        at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
        at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
        at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1418)
        at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
        at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
        at scala.Option.foreach(Option.scala:236)
        at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:799)
        at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1640)
        at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1599)
        at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1588)
        at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)
        at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:620)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:1832)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:1845)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:1858)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:1929)
        at org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.scala:927)
        at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
        at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111)
        at org.apache.spark.rdd.RDD.withScope(RDD.scala:316)
        at org.apache.spark.rdd.RDD.collect(RDD.scala:926)
        at org.apache.carbondata.spark.util.GlobalDictionaryUtil$.generateGlobalDictionary(GlobalDictionaryUtil.scala:800)
        at org.apache.spark.sql.execution.command.LoadTableUsingKettle.run(carbonTableSchema.scala:1197)
        at org.apache.spark.sql.execution.command.LoadTable.run(carbonTableSchema.scala:1036)
        at org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult$lzycompute(commands.scala:58)
        at org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult(commands.scala:56)
        at org.apache.spark.sql.execution.ExecutedCommand.doExecute(commands.scala:70)
        at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:132)
        at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:130)
        at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
        at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:130)
        at org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:55)
        at org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:55)
        at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:145)
        at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:130)
        at org.apache.carbondata.spark.rdd.CarbonDataFrameRDD.<init>(CarbonDataFrameRDD.scala:23)
        at org.apache.spark.sql.CarbonContext.sql(CarbonContext.scala:137)
        at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:37)
        at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:42)
        at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:44)
        at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:46)
        at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:48)
        at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:50)
        at $line32.$read$$iwC$$iwC$$iwC$$iwC.<init>(<console>:52)
        at $line32.$read$$iwC$$iwC$$iwC.<init>(<console>:54)
        at $line32.$read$$iwC$$iwC.<init>(<console>:56)
        at $line32.$read$$iwC.<init>(<console>:58)
        at $line32.$read.<init>(<console>:60)
        at $line32.$read$.<init>(<console>:64)
        at $line32.$read$.<clinit>(<console>)
        at $line32.$eval$.<init>(<console>:7)
        at $line32.$eval$.<clinit>(<console>)
        at $line32.$eval.$print(<console>)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
        at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:606)
        at org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(SparkIMain.scala:1065)
        at org.apache.spark.repl.SparkIMain$Request.loadAndRun(SparkIMain.scala:1346)
        at org.apache.spark.repl.SparkIMain.loadAndRunReq$1(SparkIMain.scala:840)
        at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:871)
        at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:819)
        at org.apache.spark.repl.SparkILoop.reallyInterpret$1(SparkILoop.scala:857)
        at org.apache.spark.repl.SparkILoop.interpretStartingWith(SparkILoop.scala:902)
        at org.apache.spark.repl.SparkILoop.command(SparkILoop.scala:814)
        at org.apache.spark.repl.SparkILoop.processLine$1(SparkILoop.scala:657)
        at org.apache.spark.repl.SparkILoop.innerLoop$1(SparkILoop.scala:665)
        at org.apache.spark.repl.SparkILoop.org$apache$spark$repl$SparkILoop$$loop(SparkILoop.scala:670)
        at org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl$SparkILoop$$process$1.apply$mcZ$sp(SparkILoop.scala:997)
        at org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945)
        at org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945)
        at scala.tools.nsc.util.ScalaClassLoader$.savingContextLoader(ScalaClassLoader.scala:135)
        at org.apache.spark.repl.SparkILoop.org$apache$spark$repl$SparkILoop$$process(SparkILoop.scala:945)
        at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:1059)
        at org.apache.spark.repl.Main$.main(Main.scala:31)
        at org.apache.spark.repl.Main.main(Main.scala)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
        at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:606)
        at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
        at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181)
        at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206)
        at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
        at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
Caused by: java.lang.RuntimeException: Dictionary file lo_orderpriotity is locked for updation. Please try after some time
        at scala.sys.package$.error(package.scala:27)



To unsubscribe from carbondata-0.2 load data failed in yarn molde, click here.
NAML

Reply | Threaded
Open this post in threaded view
|

Re: carbondata-0.2 load data failed in yarn molde

lionel061201
Hi team,
Looks like I've met the same problem about dictionary file is locked. Could
you share what changes you made about the configuration?

ERROR 23-12 09:55:26,222 - Executor task launch worker-0
java.lang.RuntimeException: Dictionary file vehsyspwrmod is locked for
updation. Please try after some time
at scala.sys.package$.error(package.scala:27)
at
org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
at
org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(CarbonGlobalDictionaryRDD.scala:293)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
at org.apache.spark.scheduler.Task.run(Task.scala:89)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
ERROR 23-12 09:55:26,223 - Executor task launch worker-7
java.lang.RuntimeException: Dictionary file vehindlightleft is locked for
updation. Please try after some time
at scala.sys.package$.error(package.scala:27)
at
org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
at
org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(CarbonGlobalDictionaryRDD.scala:293)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
at org.apache.spark.scheduler.Task.run(Task.scala:89)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
ERROR 23-12 09:55:26,223 - Executor task launch worker-4
java.lang.RuntimeException: Dictionary file vehwindowrearleft is locked for
updation. Please try after some time
at scala.sys.package$.error(package.scala:27)
at
org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
at
org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(CarbonGlobalDictionaryRDD.scala:293)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
at org.apache.spark.scheduler.Task.run(Task.scala:89)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
ERROR 23-12 09:55:26,226 - Exception in task 5.0 in stage 5.0 (TID 9096)
java.lang.RuntimeException: Dictionary file vehsyspwrmod is locked for
updation. Please try after some time
at scala.sys.package$.error(package.scala:27)
at
org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
at
org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(CarbonGlobalDictionaryRDD.scala:293)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
at org.apache.spark.scheduler.Task.run(Task.scala:89)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
ERROR 23-12 09:55:26,226 - Exception in task 13.0 in stage 5.0 (TID 9104)
java.lang.RuntimeException: Dictionary file vehwindowrearleft is locked for
updation. Please try after some time
at scala.sys.package$.error(package.scala:27)
at
org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
at
org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(CarbonGlobalDictionaryRDD.scala:293)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
at org.apache.spark.scheduler.Task.run(Task.scala:89)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
ERROR 23-12 09:55:26,226 - Exception in task 21.0 in stage 5.0 (TID 9112)
java.lang.RuntimeException: Dictionary file vehindlightleft is locked for
updation. Please try after some time
at scala.sys.package$.error(package.scala:27)
at
org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
at
org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(CarbonGlobalDictionaryRDD.scala:293)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
at org.apache.spark.scheduler.Task.run(Task.scala:89)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
ERROR 23-12 09:55:26,231 - Executor task launch worker-5
java.lang.RuntimeException: Dictionary file vehraindetected is locked for
updation. Please try after some time
at scala.sys.package$.error(package.scala:27)
at
org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
at
org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(CarbonGlobalDictionaryRDD.scala:293)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
at org.apache.spark.scheduler.Task.run(Task.scala:89)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
ERROR 23-12 09:55:26,231 - Exception in task 29.0 in stage 5.0 (TID 9120)
java.lang.RuntimeException: Dictionary file vehraindetected is locked for
updation. Please try after some time
at scala.sys.package$.error(package.scala:27)
at
org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
at
org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(CarbonGlobalDictionaryRDD.scala:293)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
at org.apache.spark.scheduler.Task.run(Task.scala:89)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)


failed at stage
collect at GlobalDictionaryUtil.scala:800


Thanks & Best Regards,
Lionel

On Fri, Dec 9, 2016 at 9:09 PM, geda <[hidden email]> wrote:

> Yes,it work thanks。
> follow wiki quick start,use spark-default.conf , should config
>  carbondata.properties
> but  i  use spark-shell , not include carbon.properties in it.
>
>
>
> 2016-12-09 11:48 GMT+08:00 Liang Chen [via Apache CarbonData Mailing List
> archive] <[hidden email]>:
>
> > Hi
> >
> > Have you solved this issue after applying new configurations?
> >
> > Regards
> > Liang
> >
> > geda wrote
> > hello:
> > i test  data in spark locak model ,then load data inpath to table ,works
> > well.
> > but when i use yarn-client modle,  with 1w rows , size :940k ,but error
> > happend ,there is no lock find in  tmp dir ,i don't know how to
> > debug,help.thanks.
> > spark1.6 hadoop 2.7|2.6 carbondata 0.2
> > local mode: run ok
> > $SPARK_HOME/bin/spark-shell --master local[4]  --jars
> /usr/local/spark/lib/
> > carbondata_2.10-0.2.0-incubating-shade-hadoop2.7.1.jar
> >
> >
> > yarn command : run bad
> >  $SPARK_HOME/bin/spark-shell --verbose  --master yarn-client
> > --driver-memory 1G --driver-cores 1   --executor-memory 4G
> --num-executors
> > 5 --executor-cores 1 --conf "spark.executor.
> extraJavaOptions=-XX:NewRatio=2
> > -XX:PermSize=512m -XX:MaxPermSize=512m -XX:SurvivorRatio=6  -verbose:gc
> > -XX:-PrintGCDetails -XX:+PrintGCTimeStamps " --conf "spark.driver.
> > extraJavaOptions=-XX:MaxPermSize=512m -XX:PermSize=512m"  --conf
> > spark.yarn.driver.memoryOverhead=1024 --conf spark.yarn.executor.
> memoryOverhead=3096
> >    --jars /usr/local/spark/lib/carbondata_2.10-0.2.0-
> > incubating-shade-hadoop2.7.1.jar
> >
> > import java.io._
> > import org.apache.hadoop.hive.conf.HiveConf
> > import org.apache.spark.sql.CarbonContext
> > val storePath = "hdfs://test:8020/usr/carbondata/store"
> > val cc = new CarbonContext(sc, storePath)
> > cc.setConf(HiveConf.ConfVars.HIVECHECKFILEFORMAT.varname, "false")
> > cc.setConf("carbon.kettle.home","/usr/local/spark/
> carbondata/carbonplugins")
> >
> > cc.sql("CREATE TABLE `LINEORDER3` (   LO_ORDERKEY       bigint,
> > LO_LINENUMBER     int,   LO_CUSTKEY        bigint,   LO_PARTKEY
> >  bigint,   LO_SUPPKEY        bigint,   LO_ORDERDATE      int,
> > LO_ORDERPRIOTITY  string,   LO_SHIPPRIOTITY   int,   LO_QUANTITY
>  int,
> >   LO_EXTENDEDPRICE  int,   LO_ORDTOTALPRICE  int,   LO_DISCOUNT
>  int,
> >   LO_REVENUE        int,   LO_SUPPLYCOST     int,   LO_TAX
> int,
> >   LO_COMMITDATE     int,   LO_SHIPMODE       string ) STORED BY
> > 'carbondata'")
> > cc.sql(s"load data local inpath 'hdfs://test:8020/tmp/lineorder_1w.tbl'
> >  into table lineorder3 options('DELIMITER'='|',
> 'FILEHEADER'='LO_ORDERKEY,
> > LO_LINENUMBER, LO_CUSTKEY, LO_PARTKEY , LO_SUPPKEY , LO_ORDERDATE ,
> > LO_ORDERPRIOTITY ,   LO_SHIPPRIOTITY , LO_QUANTITY ,LO_EXTENDEDPRICE ,
> > LO_ORDTOTALPRICE ,LO_DISCOUNT , LO_REVENUE  ,   LO_SUPPLYCOST,   LO_TAX,
> > LO_COMMITDATE,   LO_SHIPMODE')")
> >
> > org.apache.spark.SparkException: Job aborted due to stage failure: Task
> 0
> > in stage 2.0 failed 4 times, most recent failure: Lost task 0.3 in stage
> > 2.0 (TID 8, datanode03-bi-dev): java.lang.RuntimeException: Dictionary
> file
> > lo_orderpriotity is locked for updation. Please try after some time
> >         at scala.sys.package$.error(package.scala:27)
> >         at org.apache.carbondata.spark.rdd.
> CarbonGlobalDictionaryGenerate
> > RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> >         at org.apache.carbondata.spark.rdd.
> CarbonGlobalDictionaryGenerate
> > RDD.compute(CarbonGlobalDictionaryRDD.scala:293)
> >         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.
> scala:306)
> >
> >         at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> >         at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.
> scala:66)
> >
> >         at org.apache.spark.scheduler.Task.run(Task.scala:89)
> >         at org.apache.spark.executor.Executor$TaskRunner.run(
> Executor.scala:214)
> >
> >         at java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1145)
> >
> >         at java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:615)
> >
> >         at java.lang.Thread.run(Thread.java:745)
> >
> > Driver stacktrace:
> >         at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$
> > scheduler$DAGScheduler$$failJobAndIndependentStages(
> DAGScheduler.scala:1431)
> >
> >         at org.apache.spark.scheduler.DAGScheduler$$anonfun$
> > abortStage$1.apply(DAGScheduler.scala:1419)
> >         at org.apache.spark.scheduler.DAGScheduler$$anonfun$
> > abortStage$1.apply(DAGScheduler.scala:1418)
> >         at scala.collection.mutable.ResizableArray$class.foreach(
> ResizableArray.scala:59)
> >
> >         at scala.collection.mutable.ArrayBuffer.foreach(
> ArrayBuffer.scala:47)
> >
> >         at org.apache.spark.scheduler.DAGScheduler.abortStage(
> DAGScheduler.scala:1418)
> >
> >         at org.apache.spark.scheduler.DAGScheduler$$anonfun$
> > handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
> >         at org.apache.spark.scheduler.DAGScheduler$$anonfun$
> > handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
> >         at scala.Option.foreach(Option.scala:236)
> >         at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(
> DAGScheduler.scala:799)
> >
> >         at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.
> > doOnReceive(DAGScheduler.scala:1640)
> >         at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.
> > onReceive(DAGScheduler.scala:1599)
> >         at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.
> > onReceive(DAGScheduler.scala:1588)
> >         at org.apache.spark.util.EventLoop$$anon$1.run(
> EventLoop.scala:48)
> >
> >         at org.apache.spark.scheduler.DAGScheduler.runJob(
> DAGScheduler.scala:620)
> >
> >         at org.apache.spark.SparkContext.runJob(SparkContext.scala:1832)
> >         at org.apache.spark.SparkContext.runJob(SparkContext.scala:1845)
> >         at org.apache.spark.SparkContext.runJob(SparkContext.scala:1858)
> >         at org.apache.spark.SparkContext.runJob(SparkContext.scala:1929)
> >         at org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.
> scala:927)
> >
> >         at org.apache.spark.rdd.RDDOperationScope$.withScope(
> RDDOperationScope.scala:150)
> >
> >         at org.apache.spark.rdd.RDDOperationScope$.withScope(
> RDDOperationScope.scala:111)
> >
> >         at org.apache.spark.rdd.RDD.withScope(RDD.scala:316)
> >         at org.apache.spark.rdd.RDD.collect(RDD.scala:926)
> >         at org.apache.carbondata.spark.util.GlobalDictionaryUtil$.
> > generateGlobalDictionary(GlobalDictionaryUtil.scala:800)
> >         at org.apache.spark.sql.execution.command.
> > LoadTableUsingKettle.run(carbonTableSchema.scala:1197)
> >         at org.apache.spark.sql.execution.command.LoadTable.
> > run(carbonTableSchema.scala:1036)
> >         at org.apache.spark.sql.execution.ExecutedCommand.
> > sideEffectResult$lzycompute(commands.scala:58)
> >         at org.apache.spark.sql.execution.ExecutedCommand.
> > sideEffectResult(commands.scala:56)
> >         at org.apache.spark.sql.execution.ExecutedCommand.
> doExecute(commands.scala:70)
> >
> >         at org.apache.spark.sql.execution.SparkPlan$$anonfun$
> > execute$5.apply(SparkPlan.scala:132)
> >         at org.apache.spark.sql.execution.SparkPlan$$anonfun$
> > execute$5.apply(SparkPlan.scala:130)
> >         at org.apache.spark.rdd.RDDOperationScope$.withScope(
> RDDOperationScope.scala:150)
> >
> >         at org.apache.spark.sql.execution.SparkPlan.execute(
> SparkPlan.scala:130)
> >
> >         at org.apache.spark.sql.execution.QueryExecution.
> toRdd$lzycompute(QueryExecution.scala:55)
> >
> >         at org.apache.spark.sql.execution.QueryExecution.
> toRdd(QueryExecution.scala:55)
> >
> >         at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:145)
> >         at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:130)
> >         at org.apache.carbondata.spark.rdd.CarbonDataFrameRDD.<init>(
> CarbonDataFrameRDD.scala:23)
> >
> >         at org.apache.spark.sql.CarbonContext.sql(
> CarbonContext.scala:137)
> >
> >         at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$
> iwC.<init>(<console>:37)
> >
> >         at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<
> init>(<console>:42)
> >
> >         at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>
> (<console>:44)
> >
> >         at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<
> console>:46)
> >
> >         at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>
> :48)
> >
> >         at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:50)
> >         at $line32.$read$$iwC$$iwC$$iwC$$iwC.<init>(<console>:52)
> >         at $line32.$read$$iwC$$iwC$$iwC.<init>(<console>:54)
> >         at $line32.$read$$iwC$$iwC.<init>(<console>:56)
> >         at $line32.$read$$iwC.<init>(<console>:58)
> >         at $line32.$read.<init>(<console>:60)
> >         at $line32.$read$.<init>(<console>:64)
> >         at $line32.$read$.<clinit>(<console>)
> >         at $line32.$eval$.<init>(<console>:7)
> >         at $line32.$eval$.<clinit>(<console>)
> >         at $line32.$eval.$print(<console>)
> >         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> >         at sun.reflect.NativeMethodAccessorImpl.invoke(
> > NativeMethodAccessorImpl.java:57)
> >         at sun.reflect.DelegatingMethodAccessorImpl.invoke(
> > DelegatingMethodAccessorImpl.java:43)
> >         at java.lang.reflect.Method.invoke(Method.java:606)
> >         at org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(
> SparkIMain.scala:1065)
> >
> >         at org.apache.spark.repl.SparkIMain$Request.loadAndRun(
> SparkIMain.scala:1346)
> >
> >         at org.apache.spark.repl.SparkIMain.loadAndRunReq$1(
> SparkIMain.scala:840)
> >
> >         at org.apache.spark.repl.SparkIMain.interpret(
> SparkIMain.scala:871)
> >
> >         at org.apache.spark.repl.SparkIMain.interpret(
> SparkIMain.scala:819)
> >
> >         at org.apache.spark.repl.SparkILoop.reallyInterpret$1(
> SparkILoop.scala:857)
> >
> >         at org.apache.spark.repl.SparkILoop.interpretStartingWith(
> SparkILoop.scala:902)
> >
> >         at org.apache.spark.repl.SparkILoop.command(SparkILoop.
> scala:814)
> >         at org.apache.spark.repl.SparkILoop.processLine$1(
> SparkILoop.scala:657)
> >
> >         at org.apache.spark.repl.SparkILoop.innerLoop$1(
> SparkILoop.scala:665)
> >
> >         at org.apache.spark.repl.SparkILoop.org$apache$spark$
> > repl$SparkILoop$$loop(SparkILoop.scala:670)
> >         at org.apache.spark.repl.SparkILoop$$anonfun$org$
> > apache$spark$repl$SparkILoop$$process$1.apply$mcZ$sp(
> SparkILoop.scala:997)
> >
> >         at org.apache.spark.repl.SparkILoop$$anonfun$org$
> > apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945)
> >         at org.apache.spark.repl.SparkILoop$$anonfun$org$
> > apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945)
> >         at scala.tools.nsc.util.ScalaClassLoader$.savingContextLoader(
> ScalaClassLoader.scala:135)
> >
> >         at org.apache.spark.repl.SparkILoop.org$apache$spark$
> > repl$SparkILoop$$process(SparkILoop.scala:945)
> >         at org.apache.spark.repl.SparkILoop.process(SparkILoop.
> scala:1059)
> >
> >         at org.apache.spark.repl.Main$.main(Main.scala:31)
> >         at org.apache.spark.repl.Main.main(Main.scala)
> >         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> >         at sun.reflect.NativeMethodAccessorImpl.invoke(
> > NativeMethodAccessorImpl.java:57)
> >         at sun.reflect.DelegatingMethodAccessorImpl.invoke(
> > DelegatingMethodAccessorImpl.java:43)
> >         at java.lang.reflect.Method.invoke(Method.java:606)
> >         at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$
> > deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
> >         at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(
> SparkSubmit.scala:181)
> >
> >         at org.apache.spark.deploy.SparkSubmit$.submit(
> SparkSubmit.scala:206)
> >
> >         at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.
> scala:121)
> >
> >         at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
> > Caused by: java.lang.RuntimeException: Dictionary file lo_orderpriotity
> is
> > locked for updation. Please try after some time
> >         at scala.sys.package$.error(package.scala:27)
> >
> >
> >
> > ------------------------------
> > If you reply to this email, your message will be added to the discussion
> > below:
> > http://apache-carbondata-mailing-list-archive.1130556.
> > n5.nabble.com/carbondata-0-2-load-data-failed-in-yarn-
> > molde-tp3908p3973.html
> > To unsubscribe from carbondata-0.2 load data failed in yarn molde, click
> > here
> > <http://apache-carbondata-mailing-list-archive.1130556.
> n5.nabble.com/template/NamlServlet.jtp?macro=
> unsubscribe_by_code&node=3908&code=YmVpZG91NzdAZ21haWwuY29tfDM5MD
> h8MTU5NTU5NDExOQ==>
> > .
> > NAML
> > <http://apache-carbondata-mailing-list-archive.1130556.
> n5.nabble.com/template/NamlServlet.jtp?macro=macro_viewer&id=instant_html%
> 21nabble%3Aemail.naml&base=nabble.naml.namespaces.
> BasicNamespace-nabble.view.web.template.NabbleNamespace-
> nabble.view.web.template.NodeNamespace&breadcrumbs=
> notify_subscribers%21nabble%3Aemail.naml-instant_emails%
> 21nabble%3Aemail.naml-send_instant_email%21nabble%3Aemail.naml>
> >
>
>
>
>
> --
> View this message in context: http://apache-carbondata-
> mailing-list-archive.1130556.n5.nabble.com/carbondata-0-2-
> load-data-failed-in-yarn-molde-tp3908p4012.html
> Sent from the Apache CarbonData Mailing List archive mailing list archive
> at Nabble.com.
>
Reply | Threaded
Open this post in threaded view
|

Re: carbondata-0.2 load data failed in yarn molde

David CaiQiang
I think the root cause is metadata lock type.
Please add "carbon.lock.type" configuration to carbon.properties as following.
#Local mode
carbon.lock.type=LOCALLOCK

#Cluster mode
carbon.lock.type=HDFSLOCK
Best Regards
David Cai
Reply | Threaded
Open this post in threaded view
|

Re: carbondata-0.2 load data failed in yarn molde

manishgupta88
In reply to this post by lionel061201
Hi Lu Cao,

The problem you are facing "Dictionary file is locked for updation" can
also come when the path formation is incorrect for the dictionary files.

You have to set carbon.properties file path both in driver and executor
side. In spark application master executor logs you will find the path
printed for dictionary files. Just validate that path with the one you have
configured in carbon.properties file.

Regards
Manish Gupta

On Fri, Dec 23, 2016 at 7:43 AM, Lu Cao <[hidden email]> wrote:

> Hi team,
> Looks like I've met the same problem about dictionary file is locked. Could
> you share what changes you made about the configuration?
>
> ERROR 23-12 09:55:26,222 - Executor task launch worker-0
> java.lang.RuntimeException: Dictionary file vehsyspwrmod is locked for
> updation. Please try after some time
> at scala.sys.package$.error(package.scala:27)
> at
> org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> at
> org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(
> CarbonGlobalDictionaryRDD.scala:293)
> at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> at org.apache.spark.scheduler.Task.run(Task.scala:89)
> at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1145)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:745)
> ERROR 23-12 09:55:26,223 - Executor task launch worker-7
> java.lang.RuntimeException: Dictionary file vehindlightleft is locked for
> updation. Please try after some time
> at scala.sys.package$.error(package.scala:27)
> at
> org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> at
> org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(
> CarbonGlobalDictionaryRDD.scala:293)
> at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> at org.apache.spark.scheduler.Task.run(Task.scala:89)
> at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1145)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:745)
> ERROR 23-12 09:55:26,223 - Executor task launch worker-4
> java.lang.RuntimeException: Dictionary file vehwindowrearleft is locked for
> updation. Please try after some time
> at scala.sys.package$.error(package.scala:27)
> at
> org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> at
> org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(
> CarbonGlobalDictionaryRDD.scala:293)
> at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> at org.apache.spark.scheduler.Task.run(Task.scala:89)
> at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1145)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:745)
> ERROR 23-12 09:55:26,226 - Exception in task 5.0 in stage 5.0 (TID 9096)
> java.lang.RuntimeException: Dictionary file vehsyspwrmod is locked for
> updation. Please try after some time
> at scala.sys.package$.error(package.scala:27)
> at
> org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> at
> org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(
> CarbonGlobalDictionaryRDD.scala:293)
> at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> at org.apache.spark.scheduler.Task.run(Task.scala:89)
> at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1145)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:745)
> ERROR 23-12 09:55:26,226 - Exception in task 13.0 in stage 5.0 (TID 9104)
> java.lang.RuntimeException: Dictionary file vehwindowrearleft is locked for
> updation. Please try after some time
> at scala.sys.package$.error(package.scala:27)
> at
> org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> at
> org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(
> CarbonGlobalDictionaryRDD.scala:293)
> at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> at org.apache.spark.scheduler.Task.run(Task.scala:89)
> at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1145)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:745)
> ERROR 23-12 09:55:26,226 - Exception in task 21.0 in stage 5.0 (TID 9112)
> java.lang.RuntimeException: Dictionary file vehindlightleft is locked for
> updation. Please try after some time
> at scala.sys.package$.error(package.scala:27)
> at
> org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> at
> org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(
> CarbonGlobalDictionaryRDD.scala:293)
> at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> at org.apache.spark.scheduler.Task.run(Task.scala:89)
> at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1145)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:745)
> ERROR 23-12 09:55:26,231 - Executor task launch worker-5
> java.lang.RuntimeException: Dictionary file vehraindetected is locked for
> updation. Please try after some time
> at scala.sys.package$.error(package.scala:27)
> at
> org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> at
> org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(
> CarbonGlobalDictionaryRDD.scala:293)
> at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> at org.apache.spark.scheduler.Task.run(Task.scala:89)
> at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1145)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:745)
> ERROR 23-12 09:55:26,231 - Exception in task 29.0 in stage 5.0 (TID 9120)
> java.lang.RuntimeException: Dictionary file vehraindetected is locked for
> updation. Please try after some time
> at scala.sys.package$.error(package.scala:27)
> at
> org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> at
> org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(
> CarbonGlobalDictionaryRDD.scala:293)
> at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> at org.apache.spark.scheduler.Task.run(Task.scala:89)
> at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1145)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:745)
>
>
> failed at stage
> collect at GlobalDictionaryUtil.scala:800
>
>
> Thanks & Best Regards,
> Lionel
>
> On Fri, Dec 9, 2016 at 9:09 PM, geda <[hidden email]> wrote:
>
> > Yes,it work thanks。
> > follow wiki quick start,use spark-default.conf , should config
> >  carbondata.properties
> > but  i  use spark-shell , not include carbon.properties in it.
> >
> >
> >
> > 2016-12-09 11:48 GMT+08:00 Liang Chen [via Apache CarbonData Mailing List
> > archive] <[hidden email]>:
> >
> > > Hi
> > >
> > > Have you solved this issue after applying new configurations?
> > >
> > > Regards
> > > Liang
> > >
> > > geda wrote
> > > hello:
> > > i test  data in spark locak model ,then load data inpath to table
> ,works
> > > well.
> > > but when i use yarn-client modle,  with 1w rows , size :940k ,but error
> > > happend ,there is no lock find in  tmp dir ,i don't know how to
> > > debug,help.thanks.
> > > spark1.6 hadoop 2.7|2.6 carbondata 0.2
> > > local mode: run ok
> > > $SPARK_HOME/bin/spark-shell --master local[4]  --jars
> > /usr/local/spark/lib/
> > > carbondata_2.10-0.2.0-incubating-shade-hadoop2.7.1.jar
> > >
> > >
> > > yarn command : run bad
> > >  $SPARK_HOME/bin/spark-shell --verbose  --master yarn-client
> > > --driver-memory 1G --driver-cores 1   --executor-memory 4G
> > --num-executors
> > > 5 --executor-cores 1 --conf "spark.executor.
> > extraJavaOptions=-XX:NewRatio=2
> > > -XX:PermSize=512m -XX:MaxPermSize=512m -XX:SurvivorRatio=6  -verbose:gc
> > > -XX:-PrintGCDetails -XX:+PrintGCTimeStamps " --conf "spark.driver.
> > > extraJavaOptions=-XX:MaxPermSize=512m -XX:PermSize=512m"  --conf
> > > spark.yarn.driver.memoryOverhead=1024 --conf spark.yarn.executor.
> > memoryOverhead=3096
> > >    --jars /usr/local/spark/lib/carbondata_2.10-0.2.0-
> > > incubating-shade-hadoop2.7.1.jar
> > >
> > > import java.io._
> > > import org.apache.hadoop.hive.conf.HiveConf
> > > import org.apache.spark.sql.CarbonContext
> > > val storePath = "hdfs://test:8020/usr/carbondata/store"
> > > val cc = new CarbonContext(sc, storePath)
> > > cc.setConf(HiveConf.ConfVars.HIVECHECKFILEFORMAT.varname, "false")
> > > cc.setConf("carbon.kettle.home","/usr/local/spark/
> > carbondata/carbonplugins")
> > >
> > > cc.sql("CREATE TABLE `LINEORDER3` (   LO_ORDERKEY       bigint,
> > > LO_LINENUMBER     int,   LO_CUSTKEY        bigint,   LO_PARTKEY
> > >  bigint,   LO_SUPPKEY        bigint,   LO_ORDERDATE      int,
> > > LO_ORDERPRIOTITY  string,   LO_SHIPPRIOTITY   int,   LO_QUANTITY
> >  int,
> > >   LO_EXTENDEDPRICE  int,   LO_ORDTOTALPRICE  int,   LO_DISCOUNT
> >  int,
> > >   LO_REVENUE        int,   LO_SUPPLYCOST     int,   LO_TAX
> > int,
> > >   LO_COMMITDATE     int,   LO_SHIPMODE       string ) STORED BY
> > > 'carbondata'")
> > > cc.sql(s"load data local inpath 'hdfs://test:8020/tmp/
> lineorder_1w.tbl'
> > >  into table lineorder3 options('DELIMITER'='|',
> > 'FILEHEADER'='LO_ORDERKEY,
> > > LO_LINENUMBER, LO_CUSTKEY, LO_PARTKEY , LO_SUPPKEY , LO_ORDERDATE ,
> > > LO_ORDERPRIOTITY ,   LO_SHIPPRIOTITY , LO_QUANTITY ,LO_EXTENDEDPRICE ,
> > > LO_ORDTOTALPRICE ,LO_DISCOUNT , LO_REVENUE  ,   LO_SUPPLYCOST,
>  LO_TAX,
> > > LO_COMMITDATE,   LO_SHIPMODE')")
> > >
> > > org.apache.spark.SparkException: Job aborted due to stage failure:
> Task
> > 0
> > > in stage 2.0 failed 4 times, most recent failure: Lost task 0.3 in
> stage
> > > 2.0 (TID 8, datanode03-bi-dev): java.lang.RuntimeException: Dictionary
> > file
> > > lo_orderpriotity is locked for updation. Please try after some time
> > >         at scala.sys.package$.error(package.scala:27)
> > >         at org.apache.carbondata.spark.rdd.
> > CarbonGlobalDictionaryGenerate
> > > RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> > >         at org.apache.carbondata.spark.rdd.
> > CarbonGlobalDictionaryGenerate
> > > RDD.compute(CarbonGlobalDictionaryRDD.scala:293)
> > >         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.
> > scala:306)
> > >
> > >         at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> > >         at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.
> > scala:66)
> > >
> > >         at org.apache.spark.scheduler.Task.run(Task.scala:89)
> > >         at org.apache.spark.executor.Executor$TaskRunner.run(
> > Executor.scala:214)
> > >
> > >         at java.util.concurrent.ThreadPoolExecutor.runWorker(
> > ThreadPoolExecutor.java:1145)
> > >
> > >         at java.util.concurrent.ThreadPoolExecutor$Worker.run(
> > ThreadPoolExecutor.java:615)
> > >
> > >         at java.lang.Thread.run(Thread.java:745)
> > >
> > > Driver stacktrace:
> > >         at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$
> > > scheduler$DAGScheduler$$failJobAndIndependentStages(
> > DAGScheduler.scala:1431)
> > >
> > >         at org.apache.spark.scheduler.DAGScheduler$$anonfun$
> > > abortStage$1.apply(DAGScheduler.scala:1419)
> > >         at org.apache.spark.scheduler.DAGScheduler$$anonfun$
> > > abortStage$1.apply(DAGScheduler.scala:1418)
> > >         at scala.collection.mutable.ResizableArray$class.foreach(
> > ResizableArray.scala:59)
> > >
> > >         at scala.collection.mutable.ArrayBuffer.foreach(
> > ArrayBuffer.scala:47)
> > >
> > >         at org.apache.spark.scheduler.DAGScheduler.abortStage(
> > DAGScheduler.scala:1418)
> > >
> > >         at org.apache.spark.scheduler.DAGScheduler$$anonfun$
> > > handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
> > >         at org.apache.spark.scheduler.DAGScheduler$$anonfun$
> > > handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
> > >         at scala.Option.foreach(Option.scala:236)
> > >         at org.apache.spark.scheduler.DAGScheduler.
> handleTaskSetFailed(
> > DAGScheduler.scala:799)
> > >
> > >         at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.
> > > doOnReceive(DAGScheduler.scala:1640)
> > >         at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.
> > > onReceive(DAGScheduler.scala:1599)
> > >         at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.
> > > onReceive(DAGScheduler.scala:1588)
> > >         at org.apache.spark.util.EventLoop$$anon$1.run(
> > EventLoop.scala:48)
> > >
> > >         at org.apache.spark.scheduler.DAGScheduler.runJob(
> > DAGScheduler.scala:620)
> > >
> > >         at org.apache.spark.SparkContext.runJob(SparkContext.scala:
> 1832)
> > >         at org.apache.spark.SparkContext.runJob(SparkContext.scala:
> 1845)
> > >         at org.apache.spark.SparkContext.runJob(SparkContext.scala:
> 1858)
> > >         at org.apache.spark.SparkContext.runJob(SparkContext.scala:
> 1929)
> > >         at org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.
> > scala:927)
> > >
> > >         at org.apache.spark.rdd.RDDOperationScope$.withScope(
> > RDDOperationScope.scala:150)
> > >
> > >         at org.apache.spark.rdd.RDDOperationScope$.withScope(
> > RDDOperationScope.scala:111)
> > >
> > >         at org.apache.spark.rdd.RDD.withScope(RDD.scala:316)
> > >         at org.apache.spark.rdd.RDD.collect(RDD.scala:926)
> > >         at org.apache.carbondata.spark.util.GlobalDictionaryUtil$.
> > > generateGlobalDictionary(GlobalDictionaryUtil.scala:800)
> > >         at org.apache.spark.sql.execution.command.
> > > LoadTableUsingKettle.run(carbonTableSchema.scala:1197)
> > >         at org.apache.spark.sql.execution.command.LoadTable.
> > > run(carbonTableSchema.scala:1036)
> > >         at org.apache.spark.sql.execution.ExecutedCommand.
> > > sideEffectResult$lzycompute(commands.scala:58)
> > >         at org.apache.spark.sql.execution.ExecutedCommand.
> > > sideEffectResult(commands.scala:56)
> > >         at org.apache.spark.sql.execution.ExecutedCommand.
> > doExecute(commands.scala:70)
> > >
> > >         at org.apache.spark.sql.execution.SparkPlan$$anonfun$
> > > execute$5.apply(SparkPlan.scala:132)
> > >         at org.apache.spark.sql.execution.SparkPlan$$anonfun$
> > > execute$5.apply(SparkPlan.scala:130)
> > >         at org.apache.spark.rdd.RDDOperationScope$.withScope(
> > RDDOperationScope.scala:150)
> > >
> > >         at org.apache.spark.sql.execution.SparkPlan.execute(
> > SparkPlan.scala:130)
> > >
> > >         at org.apache.spark.sql.execution.QueryExecution.
> > toRdd$lzycompute(QueryExecution.scala:55)
> > >
> > >         at org.apache.spark.sql.execution.QueryExecution.
> > toRdd(QueryExecution.scala:55)
> > >
> > >         at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:145)
> > >         at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:130)
> > >         at org.apache.carbondata.spark.rdd.CarbonDataFrameRDD.<init>(
> > CarbonDataFrameRDD.scala:23)
> > >
> > >         at org.apache.spark.sql.CarbonContext.sql(
> > CarbonContext.scala:137)
> > >
> > >         at $line32.$read$$iwC$$iwC$$iwC$$
> iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$
> > iwC.<init>(<console>:37)
> > >
> > >         at $line32.$read$$iwC$$iwC$$iwC$$
> iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<
> > init>(<console>:42)
> > >
> > >         at $line32.$read$$iwC$$iwC$$iwC$$
> iwC$$iwC$$iwC$$iwC$$iwC.<init>
> > (<console>:44)
> > >
> > >         at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<
> > console>:46)
> > >
> > >         at $line32.$read$$iwC$$iwC$$iwC$$
> iwC$$iwC$$iwC.<init>(<console>
> > :48)
> > >
> > >         at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:50)
> > >         at $line32.$read$$iwC$$iwC$$iwC$$iwC.<init>(<console>:52)
> > >         at $line32.$read$$iwC$$iwC$$iwC.<init>(<console>:54)
> > >         at $line32.$read$$iwC$$iwC.<init>(<console>:56)
> > >         at $line32.$read$$iwC.<init>(<console>:58)
> > >         at $line32.$read.<init>(<console>:60)
> > >         at $line32.$read$.<init>(<console>:64)
> > >         at $line32.$read$.<clinit>(<console>)
> > >         at $line32.$eval$.<init>(<console>:7)
> > >         at $line32.$eval$.<clinit>(<console>)
> > >         at $line32.$eval.$print(<console>)
> > >         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> > >         at sun.reflect.NativeMethodAccessorImpl.invoke(
> > > NativeMethodAccessorImpl.java:57)
> > >         at sun.reflect.DelegatingMethodAccessorImpl.invoke(
> > > DelegatingMethodAccessorImpl.java:43)
> > >         at java.lang.reflect.Method.invoke(Method.java:606)
> > >         at org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(
> > SparkIMain.scala:1065)
> > >
> > >         at org.apache.spark.repl.SparkIMain$Request.loadAndRun(
> > SparkIMain.scala:1346)
> > >
> > >         at org.apache.spark.repl.SparkIMain.loadAndRunReq$1(
> > SparkIMain.scala:840)
> > >
> > >         at org.apache.spark.repl.SparkIMain.interpret(
> > SparkIMain.scala:871)
> > >
> > >         at org.apache.spark.repl.SparkIMain.interpret(
> > SparkIMain.scala:819)
> > >
> > >         at org.apache.spark.repl.SparkILoop.reallyInterpret$1(
> > SparkILoop.scala:857)
> > >
> > >         at org.apache.spark.repl.SparkILoop.interpretStartingWith(
> > SparkILoop.scala:902)
> > >
> > >         at org.apache.spark.repl.SparkILoop.command(SparkILoop.
> > scala:814)
> > >         at org.apache.spark.repl.SparkILoop.processLine$1(
> > SparkILoop.scala:657)
> > >
> > >         at org.apache.spark.repl.SparkILoop.innerLoop$1(
> > SparkILoop.scala:665)
> > >
> > >         at org.apache.spark.repl.SparkILoop.org$apache$spark$
> > > repl$SparkILoop$$loop(SparkILoop.scala:670)
> > >         at org.apache.spark.repl.SparkILoop$$anonfun$org$
> > > apache$spark$repl$SparkILoop$$process$1.apply$mcZ$sp(
> > SparkILoop.scala:997)
> > >
> > >         at org.apache.spark.repl.SparkILoop$$anonfun$org$
> > > apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945)
> > >         at org.apache.spark.repl.SparkILoop$$anonfun$org$
> > > apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945)
> > >         at scala.tools.nsc.util.ScalaClassLoader$.savingContextLoader(
> > ScalaClassLoader.scala:135)
> > >
> > >         at org.apache.spark.repl.SparkILoop.org$apache$spark$
> > > repl$SparkILoop$$process(SparkILoop.scala:945)
> > >         at org.apache.spark.repl.SparkILoop.process(SparkILoop.
> > scala:1059)
> > >
> > >         at org.apache.spark.repl.Main$.main(Main.scala:31)
> > >         at org.apache.spark.repl.Main.main(Main.scala)
> > >         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> > >         at sun.reflect.NativeMethodAccessorImpl.invoke(
> > > NativeMethodAccessorImpl.java:57)
> > >         at sun.reflect.DelegatingMethodAccessorImpl.invoke(
> > > DelegatingMethodAccessorImpl.java:43)
> > >         at java.lang.reflect.Method.invoke(Method.java:606)
> > >         at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$
> > > deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
> > >         at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(
> > SparkSubmit.scala:181)
> > >
> > >         at org.apache.spark.deploy.SparkSubmit$.submit(
> > SparkSubmit.scala:206)
> > >
> > >         at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.
> > scala:121)
> > >
> > >         at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
> > > Caused by: java.lang.RuntimeException: Dictionary file lo_orderpriotity
> > is
> > > locked for updation. Please try after some time
> > >         at scala.sys.package$.error(package.scala:27)
> > >
> > >
> > >
> > > ------------------------------
> > > If you reply to this email, your message will be added to the
> discussion
> > > below:
> > > http://apache-carbondata-mailing-list-archive.1130556.
> > > n5.nabble.com/carbondata-0-2-load-data-failed-in-yarn-
> > > molde-tp3908p3973.html
> > > To unsubscribe from carbondata-0.2 load data failed in yarn molde,
> click
> > > here
> > > <http://apache-carbondata-mailing-list-archive.1130556.
> > n5.nabble.com/template/NamlServlet.jtp?macro=
> > unsubscribe_by_code&node=3908&code=YmVpZG91NzdAZ21haWwuY29tfDM5MD
> > h8MTU5NTU5NDExOQ==>
> > > .
> > > NAML
> > > <http://apache-carbondata-mailing-list-archive.1130556.
> > n5.nabble.com/template/NamlServlet.jtp?macro=macro_
> viewer&id=instant_html%
> > 21nabble%3Aemail.naml&base=nabble.naml.namespaces.
> > BasicNamespace-nabble.view.web.template.NabbleNamespace-
> > nabble.view.web.template.NodeNamespace&breadcrumbs=
> > notify_subscribers%21nabble%3Aemail.naml-instant_emails%
> > 21nabble%3Aemail.naml-send_instant_email%21nabble%3Aemail.naml>
> > >
> >
> >
> >
> >
> > --
> > View this message in context: http://apache-carbondata-
> > mailing-list-archive.1130556.n5.nabble.com/carbondata-0-2-
> > load-data-failed-in-yarn-molde-tp3908p4012.html
> > Sent from the Apache CarbonData Mailing List archive mailing list archive
> > at Nabble.com.
> >
>
Reply | Threaded
Open this post in threaded view
|

Re: carbondata-0.2 load data failed in yarn molde

lionel061201
Hi,
Thank you all, I have found the root cause is the $SPARK_HOME/conf path was
reset by CDM spark service restart.
U r correct, put the carbon.properties file in both driver and executor can
solve this problem.

Thanks,
Lionel

On Fri, Dec 23, 2016 at 2:40 PM, manish gupta <[hidden email]>
wrote:

> Hi Lu Cao,
>
> The problem you are facing "Dictionary file is locked for updation" can
> also come when the path formation is incorrect for the dictionary files.
>
> You have to set carbon.properties file path both in driver and executor
> side. In spark application master executor logs you will find the path
> printed for dictionary files. Just validate that path with the one you have
> configured in carbon.properties file.
>
> Regards
> Manish Gupta
>
> On Fri, Dec 23, 2016 at 7:43 AM, Lu Cao <[hidden email]> wrote:
>
> > Hi team,
> > Looks like I've met the same problem about dictionary file is locked.
> Could
> > you share what changes you made about the configuration?
> >
> > ERROR 23-12 09:55:26,222 - Executor task launch worker-0
> > java.lang.RuntimeException: Dictionary file vehsyspwrmod is locked for
> > updation. Please try after some time
> > at scala.sys.package$.error(package.scala:27)
> > at
> > org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> > RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> > at
> > org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> RDD.compute(
> > CarbonGlobalDictionaryRDD.scala:293)
> > at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> > at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> > at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> > at org.apache.spark.scheduler.Task.run(Task.scala:89)
> > at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> > at
> > java.util.concurrent.ThreadPoolExecutor.runWorker(
> > ThreadPoolExecutor.java:1145)
> > at
> > java.util.concurrent.ThreadPoolExecutor$Worker.run(
> > ThreadPoolExecutor.java:615)
> > at java.lang.Thread.run(Thread.java:745)
> > ERROR 23-12 09:55:26,223 - Executor task launch worker-7
> > java.lang.RuntimeException: Dictionary file vehindlightleft is locked for
> > updation. Please try after some time
> > at scala.sys.package$.error(package.scala:27)
> > at
> > org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> > RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> > at
> > org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> RDD.compute(
> > CarbonGlobalDictionaryRDD.scala:293)
> > at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> > at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> > at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> > at org.apache.spark.scheduler.Task.run(Task.scala:89)
> > at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> > at
> > java.util.concurrent.ThreadPoolExecutor.runWorker(
> > ThreadPoolExecutor.java:1145)
> > at
> > java.util.concurrent.ThreadPoolExecutor$Worker.run(
> > ThreadPoolExecutor.java:615)
> > at java.lang.Thread.run(Thread.java:745)
> > ERROR 23-12 09:55:26,223 - Executor task launch worker-4
> > java.lang.RuntimeException: Dictionary file vehwindowrearleft is locked
> for
> > updation. Please try after some time
> > at scala.sys.package$.error(package.scala:27)
> > at
> > org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> > RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> > at
> > org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> RDD.compute(
> > CarbonGlobalDictionaryRDD.scala:293)
> > at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> > at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> > at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> > at org.apache.spark.scheduler.Task.run(Task.scala:89)
> > at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> > at
> > java.util.concurrent.ThreadPoolExecutor.runWorker(
> > ThreadPoolExecutor.java:1145)
> > at
> > java.util.concurrent.ThreadPoolExecutor$Worker.run(
> > ThreadPoolExecutor.java:615)
> > at java.lang.Thread.run(Thread.java:745)
> > ERROR 23-12 09:55:26,226 - Exception in task 5.0 in stage 5.0 (TID 9096)
> > java.lang.RuntimeException: Dictionary file vehsyspwrmod is locked for
> > updation. Please try after some time
> > at scala.sys.package$.error(package.scala:27)
> > at
> > org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> > RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> > at
> > org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> RDD.compute(
> > CarbonGlobalDictionaryRDD.scala:293)
> > at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> > at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> > at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> > at org.apache.spark.scheduler.Task.run(Task.scala:89)
> > at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> > at
> > java.util.concurrent.ThreadPoolExecutor.runWorker(
> > ThreadPoolExecutor.java:1145)
> > at
> > java.util.concurrent.ThreadPoolExecutor$Worker.run(
> > ThreadPoolExecutor.java:615)
> > at java.lang.Thread.run(Thread.java:745)
> > ERROR 23-12 09:55:26,226 - Exception in task 13.0 in stage 5.0 (TID 9104)
> > java.lang.RuntimeException: Dictionary file vehwindowrearleft is locked
> for
> > updation. Please try after some time
> > at scala.sys.package$.error(package.scala:27)
> > at
> > org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> > RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> > at
> > org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> RDD.compute(
> > CarbonGlobalDictionaryRDD.scala:293)
> > at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> > at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> > at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> > at org.apache.spark.scheduler.Task.run(Task.scala:89)
> > at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> > at
> > java.util.concurrent.ThreadPoolExecutor.runWorker(
> > ThreadPoolExecutor.java:1145)
> > at
> > java.util.concurrent.ThreadPoolExecutor$Worker.run(
> > ThreadPoolExecutor.java:615)
> > at java.lang.Thread.run(Thread.java:745)
> > ERROR 23-12 09:55:26,226 - Exception in task 21.0 in stage 5.0 (TID 9112)
> > java.lang.RuntimeException: Dictionary file vehindlightleft is locked for
> > updation. Please try after some time
> > at scala.sys.package$.error(package.scala:27)
> > at
> > org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> > RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> > at
> > org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> RDD.compute(
> > CarbonGlobalDictionaryRDD.scala:293)
> > at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> > at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> > at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> > at org.apache.spark.scheduler.Task.run(Task.scala:89)
> > at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> > at
> > java.util.concurrent.ThreadPoolExecutor.runWorker(
> > ThreadPoolExecutor.java:1145)
> > at
> > java.util.concurrent.ThreadPoolExecutor$Worker.run(
> > ThreadPoolExecutor.java:615)
> > at java.lang.Thread.run(Thread.java:745)
> > ERROR 23-12 09:55:26,231 - Executor task launch worker-5
> > java.lang.RuntimeException: Dictionary file vehraindetected is locked for
> > updation. Please try after some time
> > at scala.sys.package$.error(package.scala:27)
> > at
> > org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> > RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> > at
> > org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> RDD.compute(
> > CarbonGlobalDictionaryRDD.scala:293)
> > at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> > at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> > at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> > at org.apache.spark.scheduler.Task.run(Task.scala:89)
> > at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> > at
> > java.util.concurrent.ThreadPoolExecutor.runWorker(
> > ThreadPoolExecutor.java:1145)
> > at
> > java.util.concurrent.ThreadPoolExecutor$Worker.run(
> > ThreadPoolExecutor.java:615)
> > at java.lang.Thread.run(Thread.java:745)
> > ERROR 23-12 09:55:26,231 - Exception in task 29.0 in stage 5.0 (TID 9120)
> > java.lang.RuntimeException: Dictionary file vehraindetected is locked for
> > updation. Please try after some time
> > at scala.sys.package$.error(package.scala:27)
> > at
> > org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> > RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> > at
> > org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> RDD.compute(
> > CarbonGlobalDictionaryRDD.scala:293)
> > at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> > at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> > at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> > at org.apache.spark.scheduler.Task.run(Task.scala:89)
> > at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> > at
> > java.util.concurrent.ThreadPoolExecutor.runWorker(
> > ThreadPoolExecutor.java:1145)
> > at
> > java.util.concurrent.ThreadPoolExecutor$Worker.run(
> > ThreadPoolExecutor.java:615)
> > at java.lang.Thread.run(Thread.java:745)
> >
> >
> > failed at stage
> > collect at GlobalDictionaryUtil.scala:800
> >
> >
> > Thanks & Best Regards,
> > Lionel
> >
> > On Fri, Dec 9, 2016 at 9:09 PM, geda <[hidden email]> wrote:
> >
> > > Yes,it work thanks。
> > > follow wiki quick start,use spark-default.conf , should config
> > >  carbondata.properties
> > > but  i  use spark-shell , not include carbon.properties in it.
> > >
> > >
> > >
> > > 2016-12-09 11:48 GMT+08:00 Liang Chen [via Apache CarbonData Mailing
> List
> > > archive] <[hidden email]>:
> > >
> > > > Hi
> > > >
> > > > Have you solved this issue after applying new configurations?
> > > >
> > > > Regards
> > > > Liang
> > > >
> > > > geda wrote
> > > > hello:
> > > > i test  data in spark locak model ,then load data inpath to table
> > ,works
> > > > well.
> > > > but when i use yarn-client modle,  with 1w rows , size :940k ,but
> error
> > > > happend ,there is no lock find in  tmp dir ,i don't know how to
> > > > debug,help.thanks.
> > > > spark1.6 hadoop 2.7|2.6 carbondata 0.2
> > > > local mode: run ok
> > > > $SPARK_HOME/bin/spark-shell --master local[4]  --jars
> > > /usr/local/spark/lib/
> > > > carbondata_2.10-0.2.0-incubating-shade-hadoop2.7.1.jar
> > > >
> > > >
> > > > yarn command : run bad
> > > >  $SPARK_HOME/bin/spark-shell --verbose  --master yarn-client
> > > > --driver-memory 1G --driver-cores 1   --executor-memory 4G
> > > --num-executors
> > > > 5 --executor-cores 1 --conf "spark.executor.
> > > extraJavaOptions=-XX:NewRatio=2
> > > > -XX:PermSize=512m -XX:MaxPermSize=512m -XX:SurvivorRatio=6
> -verbose:gc
> > > > -XX:-PrintGCDetails -XX:+PrintGCTimeStamps " --conf "spark.driver.
> > > > extraJavaOptions=-XX:MaxPermSize=512m -XX:PermSize=512m"  --conf
> > > > spark.yarn.driver.memoryOverhead=1024 --conf spark.yarn.executor.
> > > memoryOverhead=3096
> > > >    --jars /usr/local/spark/lib/carbondata_2.10-0.2.0-
> > > > incubating-shade-hadoop2.7.1.jar
> > > >
> > > > import java.io._
> > > > import org.apache.hadoop.hive.conf.HiveConf
> > > > import org.apache.spark.sql.CarbonContext
> > > > val storePath = "hdfs://test:8020/usr/carbondata/store"
> > > > val cc = new CarbonContext(sc, storePath)
> > > > cc.setConf(HiveConf.ConfVars.HIVECHECKFILEFORMAT.varname, "false")
> > > > cc.setConf("carbon.kettle.home","/usr/local/spark/
> > > carbondata/carbonplugins")
> > > >
> > > > cc.sql("CREATE TABLE `LINEORDER3` (   LO_ORDERKEY       bigint,
> > > > LO_LINENUMBER     int,   LO_CUSTKEY        bigint,   LO_PARTKEY
> > > >  bigint,   LO_SUPPKEY        bigint,   LO_ORDERDATE      int,
> > > > LO_ORDERPRIOTITY  string,   LO_SHIPPRIOTITY   int,   LO_QUANTITY
> > >  int,
> > > >   LO_EXTENDEDPRICE  int,   LO_ORDTOTALPRICE  int,   LO_DISCOUNT
> > >  int,
> > > >   LO_REVENUE        int,   LO_SUPPLYCOST     int,   LO_TAX
> > > int,
> > > >   LO_COMMITDATE     int,   LO_SHIPMODE       string ) STORED BY
> > > > 'carbondata'")
> > > > cc.sql(s"load data local inpath 'hdfs://test:8020/tmp/
> > lineorder_1w.tbl'
> > > >  into table lineorder3 options('DELIMITER'='|',
> > > 'FILEHEADER'='LO_ORDERKEY,
> > > > LO_LINENUMBER, LO_CUSTKEY, LO_PARTKEY , LO_SUPPKEY , LO_ORDERDATE ,
> > > > LO_ORDERPRIOTITY ,   LO_SHIPPRIOTITY , LO_QUANTITY ,LO_EXTENDEDPRICE
> ,
> > > > LO_ORDTOTALPRICE ,LO_DISCOUNT , LO_REVENUE  ,   LO_SUPPLYCOST,
> >  LO_TAX,
> > > > LO_COMMITDATE,   LO_SHIPMODE')")
> > > >
> > > > org.apache.spark.SparkException: Job aborted due to stage failure:
> > Task
> > > 0
> > > > in stage 2.0 failed 4 times, most recent failure: Lost task 0.3 in
> > stage
> > > > 2.0 (TID 8, datanode03-bi-dev): java.lang.RuntimeException:
> Dictionary
> > > file
> > > > lo_orderpriotity is locked for updation. Please try after some time
> > > >         at scala.sys.package$.error(package.scala:27)
> > > >         at org.apache.carbondata.spark.rdd.
> > > CarbonGlobalDictionaryGenerate
> > > > RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> > > >         at org.apache.carbondata.spark.rdd.
> > > CarbonGlobalDictionaryGenerate
> > > > RDD.compute(CarbonGlobalDictionaryRDD.scala:293)
> > > >         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.
> > > scala:306)
> > > >
> > > >         at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> > > >         at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.
> > > scala:66)
> > > >
> > > >         at org.apache.spark.scheduler.Task.run(Task.scala:89)
> > > >         at org.apache.spark.executor.Executor$TaskRunner.run(
> > > Executor.scala:214)
> > > >
> > > >         at java.util.concurrent.ThreadPoolExecutor.runWorker(
> > > ThreadPoolExecutor.java:1145)
> > > >
> > > >         at java.util.concurrent.ThreadPoolExecutor$Worker.run(
> > > ThreadPoolExecutor.java:615)
> > > >
> > > >         at java.lang.Thread.run(Thread.java:745)
> > > >
> > > > Driver stacktrace:
> > > >         at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$
> > > > scheduler$DAGScheduler$$failJobAndIndependentStages(
> > > DAGScheduler.scala:1431)
> > > >
> > > >         at org.apache.spark.scheduler.DAGScheduler$$anonfun$
> > > > abortStage$1.apply(DAGScheduler.scala:1419)
> > > >         at org.apache.spark.scheduler.DAGScheduler$$anonfun$
> > > > abortStage$1.apply(DAGScheduler.scala:1418)
> > > >         at scala.collection.mutable.ResizableArray$class.foreach(
> > > ResizableArray.scala:59)
> > > >
> > > >         at scala.collection.mutable.ArrayBuffer.foreach(
> > > ArrayBuffer.scala:47)
> > > >
> > > >         at org.apache.spark.scheduler.DAGScheduler.abortStage(
> > > DAGScheduler.scala:1418)
> > > >
> > > >         at org.apache.spark.scheduler.DAGScheduler$$anonfun$
> > > > handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
> > > >         at org.apache.spark.scheduler.DAGScheduler$$anonfun$
> > > > handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
> > > >         at scala.Option.foreach(Option.scala:236)
> > > >         at org.apache.spark.scheduler.DAGScheduler.
> > handleTaskSetFailed(
> > > DAGScheduler.scala:799)
> > > >
> > > >         at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.
> > > > doOnReceive(DAGScheduler.scala:1640)
> > > >         at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.
> > > > onReceive(DAGScheduler.scala:1599)
> > > >         at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.
> > > > onReceive(DAGScheduler.scala:1588)
> > > >         at org.apache.spark.util.EventLoop$$anon$1.run(
> > > EventLoop.scala:48)
> > > >
> > > >         at org.apache.spark.scheduler.DAGScheduler.runJob(
> > > DAGScheduler.scala:620)
> > > >
> > > >         at org.apache.spark.SparkContext.runJob(SparkContext.scala:
> > 1832)
> > > >         at org.apache.spark.SparkContext.runJob(SparkContext.scala:
> > 1845)
> > > >         at org.apache.spark.SparkContext.runJob(SparkContext.scala:
> > 1858)
> > > >         at org.apache.spark.SparkContext.runJob(SparkContext.scala:
> > 1929)
> > > >         at org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.
> > > scala:927)
> > > >
> > > >         at org.apache.spark.rdd.RDDOperationScope$.withScope(
> > > RDDOperationScope.scala:150)
> > > >
> > > >         at org.apache.spark.rdd.RDDOperationScope$.withScope(
> > > RDDOperationScope.scala:111)
> > > >
> > > >         at org.apache.spark.rdd.RDD.withScope(RDD.scala:316)
> > > >         at org.apache.spark.rdd.RDD.collect(RDD.scala:926)
> > > >         at org.apache.carbondata.spark.util.GlobalDictionaryUtil$.
> > > > generateGlobalDictionary(GlobalDictionaryUtil.scala:800)
> > > >         at org.apache.spark.sql.execution.command.
> > > > LoadTableUsingKettle.run(carbonTableSchema.scala:1197)
> > > >         at org.apache.spark.sql.execution.command.LoadTable.
> > > > run(carbonTableSchema.scala:1036)
> > > >         at org.apache.spark.sql.execution.ExecutedCommand.
> > > > sideEffectResult$lzycompute(commands.scala:58)
> > > >         at org.apache.spark.sql.execution.ExecutedCommand.
> > > > sideEffectResult(commands.scala:56)
> > > >         at org.apache.spark.sql.execution.ExecutedCommand.
> > > doExecute(commands.scala:70)
> > > >
> > > >         at org.apache.spark.sql.execution.SparkPlan$$anonfun$
> > > > execute$5.apply(SparkPlan.scala:132)
> > > >         at org.apache.spark.sql.execution.SparkPlan$$anonfun$
> > > > execute$5.apply(SparkPlan.scala:130)
> > > >         at org.apache.spark.rdd.RDDOperationScope$.withScope(
> > > RDDOperationScope.scala:150)
> > > >
> > > >         at org.apache.spark.sql.execution.SparkPlan.execute(
> > > SparkPlan.scala:130)
> > > >
> > > >         at org.apache.spark.sql.execution.QueryExecution.
> > > toRdd$lzycompute(QueryExecution.scala:55)
> > > >
> > > >         at org.apache.spark.sql.execution.QueryExecution.
> > > toRdd(QueryExecution.scala:55)
> > > >
> > > >         at org.apache.spark.sql.DataFrame.<init>(DataFrame.
> scala:145)
> > > >         at org.apache.spark.sql.DataFrame.<init>(DataFrame.
> scala:130)
> > > >         at org.apache.carbondata.spark.
> rdd.CarbonDataFrameRDD.<init>(
> > > CarbonDataFrameRDD.scala:23)
> > > >
> > > >         at org.apache.spark.sql.CarbonContext.sql(
> > > CarbonContext.scala:137)
> > > >
> > > >         at $line32.$read$$iwC$$iwC$$iwC$$
> > iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$
> > > iwC.<init>(<console>:37)
> > > >
> > > >         at $line32.$read$$iwC$$iwC$$iwC$$
> > iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<
> > > init>(<console>:42)
> > > >
> > > >         at $line32.$read$$iwC$$iwC$$iwC$$
> > iwC$$iwC$$iwC$$iwC$$iwC.<init>
> > > (<console>:44)
> > > >
> > > >         at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<
> > > console>:46)
> > > >
> > > >         at $line32.$read$$iwC$$iwC$$iwC$$
> > iwC$$iwC$$iwC.<init>(<console>
> > > :48)
> > > >
> > > >         at $line32.$read$$iwC$$iwC$$iwC$$
> iwC$$iwC.<init>(<console>:50)
> > > >         at $line32.$read$$iwC$$iwC$$iwC$$iwC.<init>(<console>:52)
> > > >         at $line32.$read$$iwC$$iwC$$iwC.<init>(<console>:54)
> > > >         at $line32.$read$$iwC$$iwC.<init>(<console>:56)
> > > >         at $line32.$read$$iwC.<init>(<console>:58)
> > > >         at $line32.$read.<init>(<console>:60)
> > > >         at $line32.$read$.<init>(<console>:64)
> > > >         at $line32.$read$.<clinit>(<console>)
> > > >         at $line32.$eval$.<init>(<console>:7)
> > > >         at $line32.$eval$.<clinit>(<console>)
> > > >         at $line32.$eval.$print(<console>)
> > > >         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
> Method)
> > > >         at sun.reflect.NativeMethodAccessorImpl.invoke(
> > > > NativeMethodAccessorImpl.java:57)
> > > >         at sun.reflect.DelegatingMethodAccessorImpl.invoke(
> > > > DelegatingMethodAccessorImpl.java:43)
> > > >         at java.lang.reflect.Method.invoke(Method.java:606)
> > > >         at org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(
> > > SparkIMain.scala:1065)
> > > >
> > > >         at org.apache.spark.repl.SparkIMain$Request.loadAndRun(
> > > SparkIMain.scala:1346)
> > > >
> > > >         at org.apache.spark.repl.SparkIMain.loadAndRunReq$1(
> > > SparkIMain.scala:840)
> > > >
> > > >         at org.apache.spark.repl.SparkIMain.interpret(
> > > SparkIMain.scala:871)
> > > >
> > > >         at org.apache.spark.repl.SparkIMain.interpret(
> > > SparkIMain.scala:819)
> > > >
> > > >         at org.apache.spark.repl.SparkILoop.reallyInterpret$1(
> > > SparkILoop.scala:857)
> > > >
> > > >         at org.apache.spark.repl.SparkILoop.interpretStartingWith(
> > > SparkILoop.scala:902)
> > > >
> > > >         at org.apache.spark.repl.SparkILoop.command(SparkILoop.
> > > scala:814)
> > > >         at org.apache.spark.repl.SparkILoop.processLine$1(
> > > SparkILoop.scala:657)
> > > >
> > > >         at org.apache.spark.repl.SparkILoop.innerLoop$1(
> > > SparkILoop.scala:665)
> > > >
> > > >         at org.apache.spark.repl.SparkILoop.org$apache$spark$
> > > > repl$SparkILoop$$loop(SparkILoop.scala:670)
> > > >         at org.apache.spark.repl.SparkILoop$$anonfun$org$
> > > > apache$spark$repl$SparkILoop$$process$1.apply$mcZ$sp(
> > > SparkILoop.scala:997)
> > > >
> > > >         at org.apache.spark.repl.SparkILoop$$anonfun$org$
> > > > apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945)
> > > >         at org.apache.spark.repl.SparkILoop$$anonfun$org$
> > > > apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945)
> > > >         at scala.tools.nsc.util.ScalaClassLoader$.
> savingContextLoader(
> > > ScalaClassLoader.scala:135)
> > > >
> > > >         at org.apache.spark.repl.SparkILoop.org$apache$spark$
> > > > repl$SparkILoop$$process(SparkILoop.scala:945)
> > > >         at org.apache.spark.repl.SparkILoop.process(SparkILoop.
> > > scala:1059)
> > > >
> > > >         at org.apache.spark.repl.Main$.main(Main.scala:31)
> > > >         at org.apache.spark.repl.Main.main(Main.scala)
> > > >         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
> Method)
> > > >         at sun.reflect.NativeMethodAccessorImpl.invoke(
> > > > NativeMethodAccessorImpl.java:57)
> > > >         at sun.reflect.DelegatingMethodAccessorImpl.invoke(
> > > > DelegatingMethodAccessorImpl.java:43)
> > > >         at java.lang.reflect.Method.invoke(Method.java:606)
> > > >         at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$
> > > > deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
> > > >         at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(
> > > SparkSubmit.scala:181)
> > > >
> > > >         at org.apache.spark.deploy.SparkSubmit$.submit(
> > > SparkSubmit.scala:206)
> > > >
> > > >         at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.
> > > scala:121)
> > > >
> > > >         at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.
> scala)
> > > > Caused by: java.lang.RuntimeException: Dictionary file
> lo_orderpriotity
> > > is
> > > > locked for updation. Please try after some time
> > > >         at scala.sys.package$.error(package.scala:27)
> > > >
> > > >
> > > >
> > > > ------------------------------
> > > > If you reply to this email, your message will be added to the
> > discussion
> > > > below:
> > > > http://apache-carbondata-mailing-list-archive.1130556.
> > > > n5.nabble.com/carbondata-0-2-load-data-failed-in-yarn-
> > > > molde-tp3908p3973.html
> > > > To unsubscribe from carbondata-0.2 load data failed in yarn molde,
> > click
> > > > here
> > > > <http://apache-carbondata-mailing-list-archive.1130556.
> > > n5.nabble.com/template/NamlServlet.jtp?macro=
> > > unsubscribe_by_code&node=3908&code=YmVpZG91NzdAZ21haWwuY29tfDM5MD
> > > h8MTU5NTU5NDExOQ==>
> > > > .
> > > > NAML
> > > > <http://apache-carbondata-mailing-list-archive.1130556.
> > > n5.nabble.com/template/NamlServlet.jtp?macro=macro_
> > viewer&id=instant_html%
> > > 21nabble%3Aemail.naml&base=nabble.naml.namespaces.
> > > BasicNamespace-nabble.view.web.template.NabbleNamespace-
> > > nabble.view.web.template.NodeNamespace&breadcrumbs=
> > > notify_subscribers%21nabble%3Aemail.naml-instant_emails%
> > > 21nabble%3Aemail.naml-send_instant_email%21nabble%3Aemail.naml>
> > > >
> > >
> > >
> > >
> > >
> > > --
> > > View this message in context: http://apache-carbondata-
> > > mailing-list-archive.1130556.n5.nabble.com/carbondata-0-2-
> > > load-data-failed-in-yarn-molde-tp3908p4012.html
> > > Sent from the Apache CarbonData Mailing List archive mailing list
> archive
> > > at Nabble.com.
> > >
> >
>