You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@carbondata.apache.org by geda <be...@gmail.com> on 2016/12/07 03:48:44 UTC
carbondata-0.2 load data failed in yarn molde
hello:
i test data in spark locak model ,then load data inpath to table ,works
well.
but when i use yarn-client modle, with 1w rows ,error happend ,there is no
lock find in tmp dir ,i don't know how to debug,help.thanks.
local mode: run ok
$SPARK_HOME/bin/spark-shell --master local[4] --jars
/usr/local/spark/lib/carbondata_2.10-0.2.0-incubating-shade-hadoop2.7.1.jar
yarn command : run bad
$SPARK_HOME/bin/spark-shell --verbose --master yarn-client --driver-memory
1G --driver-cores 1 --executor-memory 4G --num-executors 5
--executor-cores 1 --conf "spark.executor.extraJavaOptions=-XX:NewRatio=2
-XX:PermSize=512m -XX:MaxPermSize=512m -XX:SurvivorRatio=6 -verbose:gc
-XX:-PrintGCDetails -XX:+PrintGCTimeStamps " --conf
"spark.driver.extraJavaOptions=-XX:MaxPermSize=512m -XX:PermSize=512m"
--conf spark.yarn.driver.memoryOverhead=1024 --conf
spark.yarn.executor.memoryOverhead=3096 --jars
/usr/local/spark/lib/carbondata_2.10-0.2.0-incubating-shade-hadoop2.7.1.jar
org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in
stage 2.0 failed 4 times, most recent failure: Lost task 0.3 in stage 2.0
(TID 8, datanode03-bi-dev): java.lang.RuntimeException: Dictionary file
lo_orderpriotity is locked for updation. Please try after some time
at scala.sys.package$.error(package.scala:27)
at
org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
at
org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(CarbonGlobalDictionaryRDD.scala:293)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
at
org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
at org.apache.spark.scheduler.Task.run(Task.scala:89)
at
org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
Driver stacktrace:
at
org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1431)
at
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1419)
at
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1418)
at
scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
at
scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
at
org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1418)
at
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
at
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
at scala.Option.foreach(Option.scala:236)
at
org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:799)
at
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1640)
at
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1599)
at
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1588)
at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)
at
org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:620)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1832)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1845)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1858)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1929)
at org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.scala:927)
at
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
at
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111)
at org.apache.spark.rdd.RDD.withScope(RDD.scala:316)
at org.apache.spark.rdd.RDD.collect(RDD.scala:926)
at
org.apache.carbondata.spark.util.GlobalDictionaryUtil$.generateGlobalDictionary(GlobalDictionaryUtil.scala:800)
at
org.apache.spark.sql.execution.command.LoadTableUsingKettle.run(carbonTableSchema.scala:1197)
at
org.apache.spark.sql.execution.command.LoadTable.run(carbonTableSchema.scala:1036)
at
org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult$lzycompute(commands.scala:58)
at
org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult(commands.scala:56)
at
org.apache.spark.sql.execution.ExecutedCommand.doExecute(commands.scala:70)
at
org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:132)
at
org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:130)
at
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
at
org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:130)
at
org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:55)
at
org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:55)
at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:145)
at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:130)
at
org.apache.carbondata.spark.rdd.CarbonDataFrameRDD.<init>(CarbonDataFrameRDD.scala:23)
at org.apache.spark.sql.CarbonContext.sql(CarbonContext.scala:137)
at
$line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:37)
at
$line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:42)
at
$line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:44)
at
$line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:46)
at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:48)
at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:50)
at $line32.$read$$iwC$$iwC$$iwC$$iwC.<init>(<console>:52)
at $line32.$read$$iwC$$iwC$$iwC.<init>(<console>:54)
at $line32.$read$$iwC$$iwC.<init>(<console>:56)
at $line32.$read$$iwC.<init>(<console>:58)
at $line32.$read.<init>(<console>:60)
at $line32.$read$.<init>(<console>:64)
at $line32.$read$.<clinit>(<console>)
at $line32.$eval$.<init>(<console>:7)
at $line32.$eval$.<clinit>(<console>)
at $line32.$eval.$print(<console>)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at
org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(SparkIMain.scala:1065)
at
org.apache.spark.repl.SparkIMain$Request.loadAndRun(SparkIMain.scala:1346)
at
org.apache.spark.repl.SparkIMain.loadAndRunReq$1(SparkIMain.scala:840)
at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:871)
at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:819)
at
org.apache.spark.repl.SparkILoop.reallyInterpret$1(SparkILoop.scala:857)
at
org.apache.spark.repl.SparkILoop.interpretStartingWith(SparkILoop.scala:902)
at org.apache.spark.repl.SparkILoop.command(SparkILoop.scala:814)
at
org.apache.spark.repl.SparkILoop.processLine$1(SparkILoop.scala:657)
at
org.apache.spark.repl.SparkILoop.innerLoop$1(SparkILoop.scala:665)
at
org.apache.spark.repl.SparkILoop.org$apache$spark$repl$SparkILoop$$loop(SparkILoop.scala:670)
at
org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl$SparkILoop$$process$1.apply$mcZ$sp(SparkILoop.scala:997)
at
org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945)
at
org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945)
at
scala.tools.nsc.util.ScalaClassLoader$.savingContextLoader(ScalaClassLoader.scala:135)
at
org.apache.spark.repl.SparkILoop.org$apache$spark$repl$SparkILoop$$process(SparkILoop.scala:945)
at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:1059)
at org.apache.spark.repl.Main$.main(Main.scala:31)
at org.apache.spark.repl.Main.main(Main.scala)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at
org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
at
org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181)
at
org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206)
at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
Caused by: java.lang.RuntimeException: Dictionary file lo_orderpriotity is
locked for updation. Please try after some time
at scala.sys.package$.error(package.scala:27)
--
View this message in context: http://apache-carbondata-mailing-list-archive.1130556.n5.nabble.com/carbondata-0-2-load-data-failed-in-yarn-molde-tp3908.html
Sent from the Apache CarbonData Mailing List archive mailing list archive at Nabble.com.
Re: carbondata-0.2 load data failed in yarn molde
Posted by Lu Cao <wh...@gmail.com>.
Hi,
Thank you all, I have found the root cause is the $SPARK_HOME/conf path was
reset by CDM spark service restart.
U r correct, put the carbon.properties file in both driver and executor can
solve this problem.
Thanks,
Lionel
On Fri, Dec 23, 2016 at 2:40 PM, manish gupta <to...@gmail.com>
wrote:
> Hi Lu Cao,
>
> The problem you are facing "Dictionary file is locked for updation" can
> also come when the path formation is incorrect for the dictionary files.
>
> You have to set carbon.properties file path both in driver and executor
> side. In spark application master executor logs you will find the path
> printed for dictionary files. Just validate that path with the one you have
> configured in carbon.properties file.
>
> Regards
> Manish Gupta
>
> On Fri, Dec 23, 2016 at 7:43 AM, Lu Cao <wh...@gmail.com> wrote:
>
> > Hi team,
> > Looks like I've met the same problem about dictionary file is locked.
> Could
> > you share what changes you made about the configuration?
> >
> > ERROR 23-12 09:55:26,222 - Executor task launch worker-0
> > java.lang.RuntimeException: Dictionary file vehsyspwrmod is locked for
> > updation. Please try after some time
> > at scala.sys.package$.error(package.scala:27)
> > at
> > org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> > RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> > at
> > org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> RDD.compute(
> > CarbonGlobalDictionaryRDD.scala:293)
> > at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> > at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> > at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> > at org.apache.spark.scheduler.Task.run(Task.scala:89)
> > at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> > at
> > java.util.concurrent.ThreadPoolExecutor.runWorker(
> > ThreadPoolExecutor.java:1145)
> > at
> > java.util.concurrent.ThreadPoolExecutor$Worker.run(
> > ThreadPoolExecutor.java:615)
> > at java.lang.Thread.run(Thread.java:745)
> > ERROR 23-12 09:55:26,223 - Executor task launch worker-7
> > java.lang.RuntimeException: Dictionary file vehindlightleft is locked for
> > updation. Please try after some time
> > at scala.sys.package$.error(package.scala:27)
> > at
> > org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> > RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> > at
> > org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> RDD.compute(
> > CarbonGlobalDictionaryRDD.scala:293)
> > at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> > at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> > at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> > at org.apache.spark.scheduler.Task.run(Task.scala:89)
> > at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> > at
> > java.util.concurrent.ThreadPoolExecutor.runWorker(
> > ThreadPoolExecutor.java:1145)
> > at
> > java.util.concurrent.ThreadPoolExecutor$Worker.run(
> > ThreadPoolExecutor.java:615)
> > at java.lang.Thread.run(Thread.java:745)
> > ERROR 23-12 09:55:26,223 - Executor task launch worker-4
> > java.lang.RuntimeException: Dictionary file vehwindowrearleft is locked
> for
> > updation. Please try after some time
> > at scala.sys.package$.error(package.scala:27)
> > at
> > org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> > RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> > at
> > org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> RDD.compute(
> > CarbonGlobalDictionaryRDD.scala:293)
> > at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> > at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> > at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> > at org.apache.spark.scheduler.Task.run(Task.scala:89)
> > at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> > at
> > java.util.concurrent.ThreadPoolExecutor.runWorker(
> > ThreadPoolExecutor.java:1145)
> > at
> > java.util.concurrent.ThreadPoolExecutor$Worker.run(
> > ThreadPoolExecutor.java:615)
> > at java.lang.Thread.run(Thread.java:745)
> > ERROR 23-12 09:55:26,226 - Exception in task 5.0 in stage 5.0 (TID 9096)
> > java.lang.RuntimeException: Dictionary file vehsyspwrmod is locked for
> > updation. Please try after some time
> > at scala.sys.package$.error(package.scala:27)
> > at
> > org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> > RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> > at
> > org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> RDD.compute(
> > CarbonGlobalDictionaryRDD.scala:293)
> > at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> > at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> > at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> > at org.apache.spark.scheduler.Task.run(Task.scala:89)
> > at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> > at
> > java.util.concurrent.ThreadPoolExecutor.runWorker(
> > ThreadPoolExecutor.java:1145)
> > at
> > java.util.concurrent.ThreadPoolExecutor$Worker.run(
> > ThreadPoolExecutor.java:615)
> > at java.lang.Thread.run(Thread.java:745)
> > ERROR 23-12 09:55:26,226 - Exception in task 13.0 in stage 5.0 (TID 9104)
> > java.lang.RuntimeException: Dictionary file vehwindowrearleft is locked
> for
> > updation. Please try after some time
> > at scala.sys.package$.error(package.scala:27)
> > at
> > org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> > RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> > at
> > org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> RDD.compute(
> > CarbonGlobalDictionaryRDD.scala:293)
> > at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> > at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> > at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> > at org.apache.spark.scheduler.Task.run(Task.scala:89)
> > at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> > at
> > java.util.concurrent.ThreadPoolExecutor.runWorker(
> > ThreadPoolExecutor.java:1145)
> > at
> > java.util.concurrent.ThreadPoolExecutor$Worker.run(
> > ThreadPoolExecutor.java:615)
> > at java.lang.Thread.run(Thread.java:745)
> > ERROR 23-12 09:55:26,226 - Exception in task 21.0 in stage 5.0 (TID 9112)
> > java.lang.RuntimeException: Dictionary file vehindlightleft is locked for
> > updation. Please try after some time
> > at scala.sys.package$.error(package.scala:27)
> > at
> > org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> > RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> > at
> > org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> RDD.compute(
> > CarbonGlobalDictionaryRDD.scala:293)
> > at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> > at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> > at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> > at org.apache.spark.scheduler.Task.run(Task.scala:89)
> > at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> > at
> > java.util.concurrent.ThreadPoolExecutor.runWorker(
> > ThreadPoolExecutor.java:1145)
> > at
> > java.util.concurrent.ThreadPoolExecutor$Worker.run(
> > ThreadPoolExecutor.java:615)
> > at java.lang.Thread.run(Thread.java:745)
> > ERROR 23-12 09:55:26,231 - Executor task launch worker-5
> > java.lang.RuntimeException: Dictionary file vehraindetected is locked for
> > updation. Please try after some time
> > at scala.sys.package$.error(package.scala:27)
> > at
> > org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> > RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> > at
> > org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> RDD.compute(
> > CarbonGlobalDictionaryRDD.scala:293)
> > at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> > at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> > at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> > at org.apache.spark.scheduler.Task.run(Task.scala:89)
> > at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> > at
> > java.util.concurrent.ThreadPoolExecutor.runWorker(
> > ThreadPoolExecutor.java:1145)
> > at
> > java.util.concurrent.ThreadPoolExecutor$Worker.run(
> > ThreadPoolExecutor.java:615)
> > at java.lang.Thread.run(Thread.java:745)
> > ERROR 23-12 09:55:26,231 - Exception in task 29.0 in stage 5.0 (TID 9120)
> > java.lang.RuntimeException: Dictionary file vehraindetected is locked for
> > updation. Please try after some time
> > at scala.sys.package$.error(package.scala:27)
> > at
> > org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> > RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> > at
> > org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> RDD.compute(
> > CarbonGlobalDictionaryRDD.scala:293)
> > at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> > at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> > at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> > at org.apache.spark.scheduler.Task.run(Task.scala:89)
> > at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> > at
> > java.util.concurrent.ThreadPoolExecutor.runWorker(
> > ThreadPoolExecutor.java:1145)
> > at
> > java.util.concurrent.ThreadPoolExecutor$Worker.run(
> > ThreadPoolExecutor.java:615)
> > at java.lang.Thread.run(Thread.java:745)
> >
> >
> > failed at stage
> > collect at GlobalDictionaryUtil.scala:800
> >
> >
> > Thanks & Best Regards,
> > Lionel
> >
> > On Fri, Dec 9, 2016 at 9:09 PM, geda <be...@gmail.com> wrote:
> >
> > > Yes,it work thanks。
> > > follow wiki quick start,use spark-default.conf , should config
> > > carbondata.properties
> > > but i use spark-shell , not include carbon.properties in it.
> > >
> > >
> > >
> > > 2016-12-09 11:48 GMT+08:00 Liang Chen [via Apache CarbonData Mailing
> List
> > > archive] <ml...@n5.nabble.com>:
> > >
> > > > Hi
> > > >
> > > > Have you solved this issue after applying new configurations?
> > > >
> > > > Regards
> > > > Liang
> > > >
> > > > geda wrote
> > > > hello:
> > > > i test data in spark locak model ,then load data inpath to table
> > ,works
> > > > well.
> > > > but when i use yarn-client modle, with 1w rows , size :940k ,but
> error
> > > > happend ,there is no lock find in tmp dir ,i don't know how to
> > > > debug,help.thanks.
> > > > spark1.6 hadoop 2.7|2.6 carbondata 0.2
> > > > local mode: run ok
> > > > $SPARK_HOME/bin/spark-shell --master local[4] --jars
> > > /usr/local/spark/lib/
> > > > carbondata_2.10-0.2.0-incubating-shade-hadoop2.7.1.jar
> > > >
> > > >
> > > > yarn command : run bad
> > > > $SPARK_HOME/bin/spark-shell --verbose --master yarn-client
> > > > --driver-memory 1G --driver-cores 1 --executor-memory 4G
> > > --num-executors
> > > > 5 --executor-cores 1 --conf "spark.executor.
> > > extraJavaOptions=-XX:NewRatio=2
> > > > -XX:PermSize=512m -XX:MaxPermSize=512m -XX:SurvivorRatio=6
> -verbose:gc
> > > > -XX:-PrintGCDetails -XX:+PrintGCTimeStamps " --conf "spark.driver.
> > > > extraJavaOptions=-XX:MaxPermSize=512m -XX:PermSize=512m" --conf
> > > > spark.yarn.driver.memoryOverhead=1024 --conf spark.yarn.executor.
> > > memoryOverhead=3096
> > > > --jars /usr/local/spark/lib/carbondata_2.10-0.2.0-
> > > > incubating-shade-hadoop2.7.1.jar
> > > >
> > > > import java.io._
> > > > import org.apache.hadoop.hive.conf.HiveConf
> > > > import org.apache.spark.sql.CarbonContext
> > > > val storePath = "hdfs://test:8020/usr/carbondata/store"
> > > > val cc = new CarbonContext(sc, storePath)
> > > > cc.setConf(HiveConf.ConfVars.HIVECHECKFILEFORMAT.varname, "false")
> > > > cc.setConf("carbon.kettle.home","/usr/local/spark/
> > > carbondata/carbonplugins")
> > > >
> > > > cc.sql("CREATE TABLE `LINEORDER3` ( LO_ORDERKEY bigint,
> > > > LO_LINENUMBER int, LO_CUSTKEY bigint, LO_PARTKEY
> > > > bigint, LO_SUPPKEY bigint, LO_ORDERDATE int,
> > > > LO_ORDERPRIOTITY string, LO_SHIPPRIOTITY int, LO_QUANTITY
> > > int,
> > > > LO_EXTENDEDPRICE int, LO_ORDTOTALPRICE int, LO_DISCOUNT
> > > int,
> > > > LO_REVENUE int, LO_SUPPLYCOST int, LO_TAX
> > > int,
> > > > LO_COMMITDATE int, LO_SHIPMODE string ) STORED BY
> > > > 'carbondata'")
> > > > cc.sql(s"load data local inpath 'hdfs://test:8020/tmp/
> > lineorder_1w.tbl'
> > > > into table lineorder3 options('DELIMITER'='|',
> > > 'FILEHEADER'='LO_ORDERKEY,
> > > > LO_LINENUMBER, LO_CUSTKEY, LO_PARTKEY , LO_SUPPKEY , LO_ORDERDATE ,
> > > > LO_ORDERPRIOTITY , LO_SHIPPRIOTITY , LO_QUANTITY ,LO_EXTENDEDPRICE
> ,
> > > > LO_ORDTOTALPRICE ,LO_DISCOUNT , LO_REVENUE , LO_SUPPLYCOST,
> > LO_TAX,
> > > > LO_COMMITDATE, LO_SHIPMODE')")
> > > >
> > > > org.apache.spark.SparkException: Job aborted due to stage failure:
> > Task
> > > 0
> > > > in stage 2.0 failed 4 times, most recent failure: Lost task 0.3 in
> > stage
> > > > 2.0 (TID 8, datanode03-bi-dev): java.lang.RuntimeException:
> Dictionary
> > > file
> > > > lo_orderpriotity is locked for updation. Please try after some time
> > > > at scala.sys.package$.error(package.scala:27)
> > > > at org.apache.carbondata.spark.rdd.
> > > CarbonGlobalDictionaryGenerate
> > > > RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> > > > at org.apache.carbondata.spark.rdd.
> > > CarbonGlobalDictionaryGenerate
> > > > RDD.compute(CarbonGlobalDictionaryRDD.scala:293)
> > > > at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.
> > > scala:306)
> > > >
> > > > at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> > > > at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.
> > > scala:66)
> > > >
> > > > at org.apache.spark.scheduler.Task.run(Task.scala:89)
> > > > at org.apache.spark.executor.Executor$TaskRunner.run(
> > > Executor.scala:214)
> > > >
> > > > at java.util.concurrent.ThreadPoolExecutor.runWorker(
> > > ThreadPoolExecutor.java:1145)
> > > >
> > > > at java.util.concurrent.ThreadPoolExecutor$Worker.run(
> > > ThreadPoolExecutor.java:615)
> > > >
> > > > at java.lang.Thread.run(Thread.java:745)
> > > >
> > > > Driver stacktrace:
> > > > at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$
> > > > scheduler$DAGScheduler$$failJobAndIndependentStages(
> > > DAGScheduler.scala:1431)
> > > >
> > > > at org.apache.spark.scheduler.DAGScheduler$$anonfun$
> > > > abortStage$1.apply(DAGScheduler.scala:1419)
> > > > at org.apache.spark.scheduler.DAGScheduler$$anonfun$
> > > > abortStage$1.apply(DAGScheduler.scala:1418)
> > > > at scala.collection.mutable.ResizableArray$class.foreach(
> > > ResizableArray.scala:59)
> > > >
> > > > at scala.collection.mutable.ArrayBuffer.foreach(
> > > ArrayBuffer.scala:47)
> > > >
> > > > at org.apache.spark.scheduler.DAGScheduler.abortStage(
> > > DAGScheduler.scala:1418)
> > > >
> > > > at org.apache.spark.scheduler.DAGScheduler$$anonfun$
> > > > handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
> > > > at org.apache.spark.scheduler.DAGScheduler$$anonfun$
> > > > handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
> > > > at scala.Option.foreach(Option.scala:236)
> > > > at org.apache.spark.scheduler.DAGScheduler.
> > handleTaskSetFailed(
> > > DAGScheduler.scala:799)
> > > >
> > > > at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.
> > > > doOnReceive(DAGScheduler.scala:1640)
> > > > at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.
> > > > onReceive(DAGScheduler.scala:1599)
> > > > at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.
> > > > onReceive(DAGScheduler.scala:1588)
> > > > at org.apache.spark.util.EventLoop$$anon$1.run(
> > > EventLoop.scala:48)
> > > >
> > > > at org.apache.spark.scheduler.DAGScheduler.runJob(
> > > DAGScheduler.scala:620)
> > > >
> > > > at org.apache.spark.SparkContext.runJob(SparkContext.scala:
> > 1832)
> > > > at org.apache.spark.SparkContext.runJob(SparkContext.scala:
> > 1845)
> > > > at org.apache.spark.SparkContext.runJob(SparkContext.scala:
> > 1858)
> > > > at org.apache.spark.SparkContext.runJob(SparkContext.scala:
> > 1929)
> > > > at org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.
> > > scala:927)
> > > >
> > > > at org.apache.spark.rdd.RDDOperationScope$.withScope(
> > > RDDOperationScope.scala:150)
> > > >
> > > > at org.apache.spark.rdd.RDDOperationScope$.withScope(
> > > RDDOperationScope.scala:111)
> > > >
> > > > at org.apache.spark.rdd.RDD.withScope(RDD.scala:316)
> > > > at org.apache.spark.rdd.RDD.collect(RDD.scala:926)
> > > > at org.apache.carbondata.spark.util.GlobalDictionaryUtil$.
> > > > generateGlobalDictionary(GlobalDictionaryUtil.scala:800)
> > > > at org.apache.spark.sql.execution.command.
> > > > LoadTableUsingKettle.run(carbonTableSchema.scala:1197)
> > > > at org.apache.spark.sql.execution.command.LoadTable.
> > > > run(carbonTableSchema.scala:1036)
> > > > at org.apache.spark.sql.execution.ExecutedCommand.
> > > > sideEffectResult$lzycompute(commands.scala:58)
> > > > at org.apache.spark.sql.execution.ExecutedCommand.
> > > > sideEffectResult(commands.scala:56)
> > > > at org.apache.spark.sql.execution.ExecutedCommand.
> > > doExecute(commands.scala:70)
> > > >
> > > > at org.apache.spark.sql.execution.SparkPlan$$anonfun$
> > > > execute$5.apply(SparkPlan.scala:132)
> > > > at org.apache.spark.sql.execution.SparkPlan$$anonfun$
> > > > execute$5.apply(SparkPlan.scala:130)
> > > > at org.apache.spark.rdd.RDDOperationScope$.withScope(
> > > RDDOperationScope.scala:150)
> > > >
> > > > at org.apache.spark.sql.execution.SparkPlan.execute(
> > > SparkPlan.scala:130)
> > > >
> > > > at org.apache.spark.sql.execution.QueryExecution.
> > > toRdd$lzycompute(QueryExecution.scala:55)
> > > >
> > > > at org.apache.spark.sql.execution.QueryExecution.
> > > toRdd(QueryExecution.scala:55)
> > > >
> > > > at org.apache.spark.sql.DataFrame.<init>(DataFrame.
> scala:145)
> > > > at org.apache.spark.sql.DataFrame.<init>(DataFrame.
> scala:130)
> > > > at org.apache.carbondata.spark.
> rdd.CarbonDataFrameRDD.<init>(
> > > CarbonDataFrameRDD.scala:23)
> > > >
> > > > at org.apache.spark.sql.CarbonContext.sql(
> > > CarbonContext.scala:137)
> > > >
> > > > at $line32.$read$$iwC$$iwC$$iwC$$
> > iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$
> > > iwC.<init>(<console>:37)
> > > >
> > > > at $line32.$read$$iwC$$iwC$$iwC$$
> > iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<
> > > init>(<console>:42)
> > > >
> > > > at $line32.$read$$iwC$$iwC$$iwC$$
> > iwC$$iwC$$iwC$$iwC$$iwC.<init>
> > > (<console>:44)
> > > >
> > > > at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<
> > > console>:46)
> > > >
> > > > at $line32.$read$$iwC$$iwC$$iwC$$
> > iwC$$iwC$$iwC.<init>(<console>
> > > :48)
> > > >
> > > > at $line32.$read$$iwC$$iwC$$iwC$$
> iwC$$iwC.<init>(<console>:50)
> > > > at $line32.$read$$iwC$$iwC$$iwC$$iwC.<init>(<console>:52)
> > > > at $line32.$read$$iwC$$iwC$$iwC.<init>(<console>:54)
> > > > at $line32.$read$$iwC$$iwC.<init>(<console>:56)
> > > > at $line32.$read$$iwC.<init>(<console>:58)
> > > > at $line32.$read.<init>(<console>:60)
> > > > at $line32.$read$.<init>(<console>:64)
> > > > at $line32.$read$.<clinit>(<console>)
> > > > at $line32.$eval$.<init>(<console>:7)
> > > > at $line32.$eval$.<clinit>(<console>)
> > > > at $line32.$eval.$print(<console>)
> > > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
> Method)
> > > > at sun.reflect.NativeMethodAccessorImpl.invoke(
> > > > NativeMethodAccessorImpl.java:57)
> > > > at sun.reflect.DelegatingMethodAccessorImpl.invoke(
> > > > DelegatingMethodAccessorImpl.java:43)
> > > > at java.lang.reflect.Method.invoke(Method.java:606)
> > > > at org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(
> > > SparkIMain.scala:1065)
> > > >
> > > > at org.apache.spark.repl.SparkIMain$Request.loadAndRun(
> > > SparkIMain.scala:1346)
> > > >
> > > > at org.apache.spark.repl.SparkIMain.loadAndRunReq$1(
> > > SparkIMain.scala:840)
> > > >
> > > > at org.apache.spark.repl.SparkIMain.interpret(
> > > SparkIMain.scala:871)
> > > >
> > > > at org.apache.spark.repl.SparkIMain.interpret(
> > > SparkIMain.scala:819)
> > > >
> > > > at org.apache.spark.repl.SparkILoop.reallyInterpret$1(
> > > SparkILoop.scala:857)
> > > >
> > > > at org.apache.spark.repl.SparkILoop.interpretStartingWith(
> > > SparkILoop.scala:902)
> > > >
> > > > at org.apache.spark.repl.SparkILoop.command(SparkILoop.
> > > scala:814)
> > > > at org.apache.spark.repl.SparkILoop.processLine$1(
> > > SparkILoop.scala:657)
> > > >
> > > > at org.apache.spark.repl.SparkILoop.innerLoop$1(
> > > SparkILoop.scala:665)
> > > >
> > > > at org.apache.spark.repl.SparkILoop.org$apache$spark$
> > > > repl$SparkILoop$$loop(SparkILoop.scala:670)
> > > > at org.apache.spark.repl.SparkILoop$$anonfun$org$
> > > > apache$spark$repl$SparkILoop$$process$1.apply$mcZ$sp(
> > > SparkILoop.scala:997)
> > > >
> > > > at org.apache.spark.repl.SparkILoop$$anonfun$org$
> > > > apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945)
> > > > at org.apache.spark.repl.SparkILoop$$anonfun$org$
> > > > apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945)
> > > > at scala.tools.nsc.util.ScalaClassLoader$.
> savingContextLoader(
> > > ScalaClassLoader.scala:135)
> > > >
> > > > at org.apache.spark.repl.SparkILoop.org$apache$spark$
> > > > repl$SparkILoop$$process(SparkILoop.scala:945)
> > > > at org.apache.spark.repl.SparkILoop.process(SparkILoop.
> > > scala:1059)
> > > >
> > > > at org.apache.spark.repl.Main$.main(Main.scala:31)
> > > > at org.apache.spark.repl.Main.main(Main.scala)
> > > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
> Method)
> > > > at sun.reflect.NativeMethodAccessorImpl.invoke(
> > > > NativeMethodAccessorImpl.java:57)
> > > > at sun.reflect.DelegatingMethodAccessorImpl.invoke(
> > > > DelegatingMethodAccessorImpl.java:43)
> > > > at java.lang.reflect.Method.invoke(Method.java:606)
> > > > at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$
> > > > deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
> > > > at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(
> > > SparkSubmit.scala:181)
> > > >
> > > > at org.apache.spark.deploy.SparkSubmit$.submit(
> > > SparkSubmit.scala:206)
> > > >
> > > > at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.
> > > scala:121)
> > > >
> > > > at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.
> scala)
> > > > Caused by: java.lang.RuntimeException: Dictionary file
> lo_orderpriotity
> > > is
> > > > locked for updation. Please try after some time
> > > > at scala.sys.package$.error(package.scala:27)
> > > >
> > > >
> > > >
> > > > ------------------------------
> > > > If you reply to this email, your message will be added to the
> > discussion
> > > > below:
> > > > http://apache-carbondata-mailing-list-archive.1130556.
> > > > n5.nabble.com/carbondata-0-2-load-data-failed-in-yarn-
> > > > molde-tp3908p3973.html
> > > > To unsubscribe from carbondata-0.2 load data failed in yarn molde,
> > click
> > > > here
> > > > <http://apache-carbondata-mailing-list-archive.1130556.
> > > n5.nabble.com/template/NamlServlet.jtp?macro=
> > > unsubscribe_by_code&node=3908&code=YmVpZG91NzdAZ21haWwuY29tfDM5MD
> > > h8MTU5NTU5NDExOQ==>
> > > > .
> > > > NAML
> > > > <http://apache-carbondata-mailing-list-archive.1130556.
> > > n5.nabble.com/template/NamlServlet.jtp?macro=macro_
> > viewer&id=instant_html%
> > > 21nabble%3Aemail.naml&base=nabble.naml.namespaces.
> > > BasicNamespace-nabble.view.web.template.NabbleNamespace-
> > > nabble.view.web.template.NodeNamespace&breadcrumbs=
> > > notify_subscribers%21nabble%3Aemail.naml-instant_emails%
> > > 21nabble%3Aemail.naml-send_instant_email%21nabble%3Aemail.naml>
> > > >
> > >
> > >
> > >
> > >
> > > --
> > > View this message in context: http://apache-carbondata-
> > > mailing-list-archive.1130556.n5.nabble.com/carbondata-0-2-
> > > load-data-failed-in-yarn-molde-tp3908p4012.html
> > > Sent from the Apache CarbonData Mailing List archive mailing list
> archive
> > > at Nabble.com.
> > >
> >
>
Re: carbondata-0.2 load data failed in yarn molde
Posted by manish gupta <to...@gmail.com>.
Hi Lu Cao,
The problem you are facing "Dictionary file is locked for updation" can
also come when the path formation is incorrect for the dictionary files.
You have to set carbon.properties file path both in driver and executor
side. In spark application master executor logs you will find the path
printed for dictionary files. Just validate that path with the one you have
configured in carbon.properties file.
Regards
Manish Gupta
On Fri, Dec 23, 2016 at 7:43 AM, Lu Cao <wh...@gmail.com> wrote:
> Hi team,
> Looks like I've met the same problem about dictionary file is locked. Could
> you share what changes you made about the configuration?
>
> ERROR 23-12 09:55:26,222 - Executor task launch worker-0
> java.lang.RuntimeException: Dictionary file vehsyspwrmod is locked for
> updation. Please try after some time
> at scala.sys.package$.error(package.scala:27)
> at
> org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> at
> org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(
> CarbonGlobalDictionaryRDD.scala:293)
> at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> at org.apache.spark.scheduler.Task.run(Task.scala:89)
> at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1145)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:745)
> ERROR 23-12 09:55:26,223 - Executor task launch worker-7
> java.lang.RuntimeException: Dictionary file vehindlightleft is locked for
> updation. Please try after some time
> at scala.sys.package$.error(package.scala:27)
> at
> org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> at
> org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(
> CarbonGlobalDictionaryRDD.scala:293)
> at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> at org.apache.spark.scheduler.Task.run(Task.scala:89)
> at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1145)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:745)
> ERROR 23-12 09:55:26,223 - Executor task launch worker-4
> java.lang.RuntimeException: Dictionary file vehwindowrearleft is locked for
> updation. Please try after some time
> at scala.sys.package$.error(package.scala:27)
> at
> org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> at
> org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(
> CarbonGlobalDictionaryRDD.scala:293)
> at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> at org.apache.spark.scheduler.Task.run(Task.scala:89)
> at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1145)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:745)
> ERROR 23-12 09:55:26,226 - Exception in task 5.0 in stage 5.0 (TID 9096)
> java.lang.RuntimeException: Dictionary file vehsyspwrmod is locked for
> updation. Please try after some time
> at scala.sys.package$.error(package.scala:27)
> at
> org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> at
> org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(
> CarbonGlobalDictionaryRDD.scala:293)
> at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> at org.apache.spark.scheduler.Task.run(Task.scala:89)
> at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1145)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:745)
> ERROR 23-12 09:55:26,226 - Exception in task 13.0 in stage 5.0 (TID 9104)
> java.lang.RuntimeException: Dictionary file vehwindowrearleft is locked for
> updation. Please try after some time
> at scala.sys.package$.error(package.scala:27)
> at
> org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> at
> org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(
> CarbonGlobalDictionaryRDD.scala:293)
> at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> at org.apache.spark.scheduler.Task.run(Task.scala:89)
> at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1145)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:745)
> ERROR 23-12 09:55:26,226 - Exception in task 21.0 in stage 5.0 (TID 9112)
> java.lang.RuntimeException: Dictionary file vehindlightleft is locked for
> updation. Please try after some time
> at scala.sys.package$.error(package.scala:27)
> at
> org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> at
> org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(
> CarbonGlobalDictionaryRDD.scala:293)
> at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> at org.apache.spark.scheduler.Task.run(Task.scala:89)
> at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1145)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:745)
> ERROR 23-12 09:55:26,231 - Executor task launch worker-5
> java.lang.RuntimeException: Dictionary file vehraindetected is locked for
> updation. Please try after some time
> at scala.sys.package$.error(package.scala:27)
> at
> org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> at
> org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(
> CarbonGlobalDictionaryRDD.scala:293)
> at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> at org.apache.spark.scheduler.Task.run(Task.scala:89)
> at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1145)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:745)
> ERROR 23-12 09:55:26,231 - Exception in task 29.0 in stage 5.0 (TID 9120)
> java.lang.RuntimeException: Dictionary file vehraindetected is locked for
> updation. Please try after some time
> at scala.sys.package$.error(package.scala:27)
> at
> org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> at
> org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(
> CarbonGlobalDictionaryRDD.scala:293)
> at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> at org.apache.spark.scheduler.Task.run(Task.scala:89)
> at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1145)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:745)
>
>
> failed at stage
> collect at GlobalDictionaryUtil.scala:800
>
>
> Thanks & Best Regards,
> Lionel
>
> On Fri, Dec 9, 2016 at 9:09 PM, geda <be...@gmail.com> wrote:
>
> > Yes,it work thanks。
> > follow wiki quick start,use spark-default.conf , should config
> > carbondata.properties
> > but i use spark-shell , not include carbon.properties in it.
> >
> >
> >
> > 2016-12-09 11:48 GMT+08:00 Liang Chen [via Apache CarbonData Mailing List
> > archive] <ml...@n5.nabble.com>:
> >
> > > Hi
> > >
> > > Have you solved this issue after applying new configurations?
> > >
> > > Regards
> > > Liang
> > >
> > > geda wrote
> > > hello:
> > > i test data in spark locak model ,then load data inpath to table
> ,works
> > > well.
> > > but when i use yarn-client modle, with 1w rows , size :940k ,but error
> > > happend ,there is no lock find in tmp dir ,i don't know how to
> > > debug,help.thanks.
> > > spark1.6 hadoop 2.7|2.6 carbondata 0.2
> > > local mode: run ok
> > > $SPARK_HOME/bin/spark-shell --master local[4] --jars
> > /usr/local/spark/lib/
> > > carbondata_2.10-0.2.0-incubating-shade-hadoop2.7.1.jar
> > >
> > >
> > > yarn command : run bad
> > > $SPARK_HOME/bin/spark-shell --verbose --master yarn-client
> > > --driver-memory 1G --driver-cores 1 --executor-memory 4G
> > --num-executors
> > > 5 --executor-cores 1 --conf "spark.executor.
> > extraJavaOptions=-XX:NewRatio=2
> > > -XX:PermSize=512m -XX:MaxPermSize=512m -XX:SurvivorRatio=6 -verbose:gc
> > > -XX:-PrintGCDetails -XX:+PrintGCTimeStamps " --conf "spark.driver.
> > > extraJavaOptions=-XX:MaxPermSize=512m -XX:PermSize=512m" --conf
> > > spark.yarn.driver.memoryOverhead=1024 --conf spark.yarn.executor.
> > memoryOverhead=3096
> > > --jars /usr/local/spark/lib/carbondata_2.10-0.2.0-
> > > incubating-shade-hadoop2.7.1.jar
> > >
> > > import java.io._
> > > import org.apache.hadoop.hive.conf.HiveConf
> > > import org.apache.spark.sql.CarbonContext
> > > val storePath = "hdfs://test:8020/usr/carbondata/store"
> > > val cc = new CarbonContext(sc, storePath)
> > > cc.setConf(HiveConf.ConfVars.HIVECHECKFILEFORMAT.varname, "false")
> > > cc.setConf("carbon.kettle.home","/usr/local/spark/
> > carbondata/carbonplugins")
> > >
> > > cc.sql("CREATE TABLE `LINEORDER3` ( LO_ORDERKEY bigint,
> > > LO_LINENUMBER int, LO_CUSTKEY bigint, LO_PARTKEY
> > > bigint, LO_SUPPKEY bigint, LO_ORDERDATE int,
> > > LO_ORDERPRIOTITY string, LO_SHIPPRIOTITY int, LO_QUANTITY
> > int,
> > > LO_EXTENDEDPRICE int, LO_ORDTOTALPRICE int, LO_DISCOUNT
> > int,
> > > LO_REVENUE int, LO_SUPPLYCOST int, LO_TAX
> > int,
> > > LO_COMMITDATE int, LO_SHIPMODE string ) STORED BY
> > > 'carbondata'")
> > > cc.sql(s"load data local inpath 'hdfs://test:8020/tmp/
> lineorder_1w.tbl'
> > > into table lineorder3 options('DELIMITER'='|',
> > 'FILEHEADER'='LO_ORDERKEY,
> > > LO_LINENUMBER, LO_CUSTKEY, LO_PARTKEY , LO_SUPPKEY , LO_ORDERDATE ,
> > > LO_ORDERPRIOTITY , LO_SHIPPRIOTITY , LO_QUANTITY ,LO_EXTENDEDPRICE ,
> > > LO_ORDTOTALPRICE ,LO_DISCOUNT , LO_REVENUE , LO_SUPPLYCOST,
> LO_TAX,
> > > LO_COMMITDATE, LO_SHIPMODE')")
> > >
> > > org.apache.spark.SparkException: Job aborted due to stage failure:
> Task
> > 0
> > > in stage 2.0 failed 4 times, most recent failure: Lost task 0.3 in
> stage
> > > 2.0 (TID 8, datanode03-bi-dev): java.lang.RuntimeException: Dictionary
> > file
> > > lo_orderpriotity is locked for updation. Please try after some time
> > > at scala.sys.package$.error(package.scala:27)
> > > at org.apache.carbondata.spark.rdd.
> > CarbonGlobalDictionaryGenerate
> > > RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> > > at org.apache.carbondata.spark.rdd.
> > CarbonGlobalDictionaryGenerate
> > > RDD.compute(CarbonGlobalDictionaryRDD.scala:293)
> > > at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.
> > scala:306)
> > >
> > > at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> > > at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.
> > scala:66)
> > >
> > > at org.apache.spark.scheduler.Task.run(Task.scala:89)
> > > at org.apache.spark.executor.Executor$TaskRunner.run(
> > Executor.scala:214)
> > >
> > > at java.util.concurrent.ThreadPoolExecutor.runWorker(
> > ThreadPoolExecutor.java:1145)
> > >
> > > at java.util.concurrent.ThreadPoolExecutor$Worker.run(
> > ThreadPoolExecutor.java:615)
> > >
> > > at java.lang.Thread.run(Thread.java:745)
> > >
> > > Driver stacktrace:
> > > at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$
> > > scheduler$DAGScheduler$$failJobAndIndependentStages(
> > DAGScheduler.scala:1431)
> > >
> > > at org.apache.spark.scheduler.DAGScheduler$$anonfun$
> > > abortStage$1.apply(DAGScheduler.scala:1419)
> > > at org.apache.spark.scheduler.DAGScheduler$$anonfun$
> > > abortStage$1.apply(DAGScheduler.scala:1418)
> > > at scala.collection.mutable.ResizableArray$class.foreach(
> > ResizableArray.scala:59)
> > >
> > > at scala.collection.mutable.ArrayBuffer.foreach(
> > ArrayBuffer.scala:47)
> > >
> > > at org.apache.spark.scheduler.DAGScheduler.abortStage(
> > DAGScheduler.scala:1418)
> > >
> > > at org.apache.spark.scheduler.DAGScheduler$$anonfun$
> > > handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
> > > at org.apache.spark.scheduler.DAGScheduler$$anonfun$
> > > handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
> > > at scala.Option.foreach(Option.scala:236)
> > > at org.apache.spark.scheduler.DAGScheduler.
> handleTaskSetFailed(
> > DAGScheduler.scala:799)
> > >
> > > at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.
> > > doOnReceive(DAGScheduler.scala:1640)
> > > at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.
> > > onReceive(DAGScheduler.scala:1599)
> > > at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.
> > > onReceive(DAGScheduler.scala:1588)
> > > at org.apache.spark.util.EventLoop$$anon$1.run(
> > EventLoop.scala:48)
> > >
> > > at org.apache.spark.scheduler.DAGScheduler.runJob(
> > DAGScheduler.scala:620)
> > >
> > > at org.apache.spark.SparkContext.runJob(SparkContext.scala:
> 1832)
> > > at org.apache.spark.SparkContext.runJob(SparkContext.scala:
> 1845)
> > > at org.apache.spark.SparkContext.runJob(SparkContext.scala:
> 1858)
> > > at org.apache.spark.SparkContext.runJob(SparkContext.scala:
> 1929)
> > > at org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.
> > scala:927)
> > >
> > > at org.apache.spark.rdd.RDDOperationScope$.withScope(
> > RDDOperationScope.scala:150)
> > >
> > > at org.apache.spark.rdd.RDDOperationScope$.withScope(
> > RDDOperationScope.scala:111)
> > >
> > > at org.apache.spark.rdd.RDD.withScope(RDD.scala:316)
> > > at org.apache.spark.rdd.RDD.collect(RDD.scala:926)
> > > at org.apache.carbondata.spark.util.GlobalDictionaryUtil$.
> > > generateGlobalDictionary(GlobalDictionaryUtil.scala:800)
> > > at org.apache.spark.sql.execution.command.
> > > LoadTableUsingKettle.run(carbonTableSchema.scala:1197)
> > > at org.apache.spark.sql.execution.command.LoadTable.
> > > run(carbonTableSchema.scala:1036)
> > > at org.apache.spark.sql.execution.ExecutedCommand.
> > > sideEffectResult$lzycompute(commands.scala:58)
> > > at org.apache.spark.sql.execution.ExecutedCommand.
> > > sideEffectResult(commands.scala:56)
> > > at org.apache.spark.sql.execution.ExecutedCommand.
> > doExecute(commands.scala:70)
> > >
> > > at org.apache.spark.sql.execution.SparkPlan$$anonfun$
> > > execute$5.apply(SparkPlan.scala:132)
> > > at org.apache.spark.sql.execution.SparkPlan$$anonfun$
> > > execute$5.apply(SparkPlan.scala:130)
> > > at org.apache.spark.rdd.RDDOperationScope$.withScope(
> > RDDOperationScope.scala:150)
> > >
> > > at org.apache.spark.sql.execution.SparkPlan.execute(
> > SparkPlan.scala:130)
> > >
> > > at org.apache.spark.sql.execution.QueryExecution.
> > toRdd$lzycompute(QueryExecution.scala:55)
> > >
> > > at org.apache.spark.sql.execution.QueryExecution.
> > toRdd(QueryExecution.scala:55)
> > >
> > > at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:145)
> > > at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:130)
> > > at org.apache.carbondata.spark.rdd.CarbonDataFrameRDD.<init>(
> > CarbonDataFrameRDD.scala:23)
> > >
> > > at org.apache.spark.sql.CarbonContext.sql(
> > CarbonContext.scala:137)
> > >
> > > at $line32.$read$$iwC$$iwC$$iwC$$
> iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$
> > iwC.<init>(<console>:37)
> > >
> > > at $line32.$read$$iwC$$iwC$$iwC$$
> iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<
> > init>(<console>:42)
> > >
> > > at $line32.$read$$iwC$$iwC$$iwC$$
> iwC$$iwC$$iwC$$iwC$$iwC.<init>
> > (<console>:44)
> > >
> > > at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<
> > console>:46)
> > >
> > > at $line32.$read$$iwC$$iwC$$iwC$$
> iwC$$iwC$$iwC.<init>(<console>
> > :48)
> > >
> > > at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:50)
> > > at $line32.$read$$iwC$$iwC$$iwC$$iwC.<init>(<console>:52)
> > > at $line32.$read$$iwC$$iwC$$iwC.<init>(<console>:54)
> > > at $line32.$read$$iwC$$iwC.<init>(<console>:56)
> > > at $line32.$read$$iwC.<init>(<console>:58)
> > > at $line32.$read.<init>(<console>:60)
> > > at $line32.$read$.<init>(<console>:64)
> > > at $line32.$read$.<clinit>(<console>)
> > > at $line32.$eval$.<init>(<console>:7)
> > > at $line32.$eval$.<clinit>(<console>)
> > > at $line32.$eval.$print(<console>)
> > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> > > at sun.reflect.NativeMethodAccessorImpl.invoke(
> > > NativeMethodAccessorImpl.java:57)
> > > at sun.reflect.DelegatingMethodAccessorImpl.invoke(
> > > DelegatingMethodAccessorImpl.java:43)
> > > at java.lang.reflect.Method.invoke(Method.java:606)
> > > at org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(
> > SparkIMain.scala:1065)
> > >
> > > at org.apache.spark.repl.SparkIMain$Request.loadAndRun(
> > SparkIMain.scala:1346)
> > >
> > > at org.apache.spark.repl.SparkIMain.loadAndRunReq$1(
> > SparkIMain.scala:840)
> > >
> > > at org.apache.spark.repl.SparkIMain.interpret(
> > SparkIMain.scala:871)
> > >
> > > at org.apache.spark.repl.SparkIMain.interpret(
> > SparkIMain.scala:819)
> > >
> > > at org.apache.spark.repl.SparkILoop.reallyInterpret$1(
> > SparkILoop.scala:857)
> > >
> > > at org.apache.spark.repl.SparkILoop.interpretStartingWith(
> > SparkILoop.scala:902)
> > >
> > > at org.apache.spark.repl.SparkILoop.command(SparkILoop.
> > scala:814)
> > > at org.apache.spark.repl.SparkILoop.processLine$1(
> > SparkILoop.scala:657)
> > >
> > > at org.apache.spark.repl.SparkILoop.innerLoop$1(
> > SparkILoop.scala:665)
> > >
> > > at org.apache.spark.repl.SparkILoop.org$apache$spark$
> > > repl$SparkILoop$$loop(SparkILoop.scala:670)
> > > at org.apache.spark.repl.SparkILoop$$anonfun$org$
> > > apache$spark$repl$SparkILoop$$process$1.apply$mcZ$sp(
> > SparkILoop.scala:997)
> > >
> > > at org.apache.spark.repl.SparkILoop$$anonfun$org$
> > > apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945)
> > > at org.apache.spark.repl.SparkILoop$$anonfun$org$
> > > apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945)
> > > at scala.tools.nsc.util.ScalaClassLoader$.savingContextLoader(
> > ScalaClassLoader.scala:135)
> > >
> > > at org.apache.spark.repl.SparkILoop.org$apache$spark$
> > > repl$SparkILoop$$process(SparkILoop.scala:945)
> > > at org.apache.spark.repl.SparkILoop.process(SparkILoop.
> > scala:1059)
> > >
> > > at org.apache.spark.repl.Main$.main(Main.scala:31)
> > > at org.apache.spark.repl.Main.main(Main.scala)
> > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> > > at sun.reflect.NativeMethodAccessorImpl.invoke(
> > > NativeMethodAccessorImpl.java:57)
> > > at sun.reflect.DelegatingMethodAccessorImpl.invoke(
> > > DelegatingMethodAccessorImpl.java:43)
> > > at java.lang.reflect.Method.invoke(Method.java:606)
> > > at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$
> > > deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
> > > at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(
> > SparkSubmit.scala:181)
> > >
> > > at org.apache.spark.deploy.SparkSubmit$.submit(
> > SparkSubmit.scala:206)
> > >
> > > at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.
> > scala:121)
> > >
> > > at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
> > > Caused by: java.lang.RuntimeException: Dictionary file lo_orderpriotity
> > is
> > > locked for updation. Please try after some time
> > > at scala.sys.package$.error(package.scala:27)
> > >
> > >
> > >
> > > ------------------------------
> > > If you reply to this email, your message will be added to the
> discussion
> > > below:
> > > http://apache-carbondata-mailing-list-archive.1130556.
> > > n5.nabble.com/carbondata-0-2-load-data-failed-in-yarn-
> > > molde-tp3908p3973.html
> > > To unsubscribe from carbondata-0.2 load data failed in yarn molde,
> click
> > > here
> > > <http://apache-carbondata-mailing-list-archive.1130556.
> > n5.nabble.com/template/NamlServlet.jtp?macro=
> > unsubscribe_by_code&node=3908&code=YmVpZG91NzdAZ21haWwuY29tfDM5MD
> > h8MTU5NTU5NDExOQ==>
> > > .
> > > NAML
> > > <http://apache-carbondata-mailing-list-archive.1130556.
> > n5.nabble.com/template/NamlServlet.jtp?macro=macro_
> viewer&id=instant_html%
> > 21nabble%3Aemail.naml&base=nabble.naml.namespaces.
> > BasicNamespace-nabble.view.web.template.NabbleNamespace-
> > nabble.view.web.template.NodeNamespace&breadcrumbs=
> > notify_subscribers%21nabble%3Aemail.naml-instant_emails%
> > 21nabble%3Aemail.naml-send_instant_email%21nabble%3Aemail.naml>
> > >
> >
> >
> >
> >
> > --
> > View this message in context: http://apache-carbondata-
> > mailing-list-archive.1130556.n5.nabble.com/carbondata-0-2-
> > load-data-failed-in-yarn-molde-tp3908p4012.html
> > Sent from the Apache CarbonData Mailing List archive mailing list archive
> > at Nabble.com.
> >
>
Re: carbondata-0.2 load data failed in yarn molde
Posted by QiangCai <qi...@qq.com>.
I think the root cause is metadata lock type.
Please add "carbon.lock.type" configuration to carbon.properties as
following.
#Local mode
carbon.lock.type=LOCALLOCK
#Cluster mode
carbon.lock.type=HDFSLOCK
--
View this message in context: http://apache-carbondata-mailing-list-archive.1130556.n5.nabble.com/carbondata-0-2-load-data-failed-in-yarn-molde-tp3908p4887.html
Sent from the Apache CarbonData Mailing List archive mailing list archive at Nabble.com.
Re: carbondata-0.2 load data failed in yarn molde
Posted by Lu Cao <wh...@gmail.com>.
Hi team,
Looks like I've met the same problem about dictionary file is locked. Could
you share what changes you made about the configuration?
ERROR 23-12 09:55:26,222 - Executor task launch worker-0
java.lang.RuntimeException: Dictionary file vehsyspwrmod is locked for
updation. Please try after some time
at scala.sys.package$.error(package.scala:27)
at
org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
at
org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(CarbonGlobalDictionaryRDD.scala:293)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
at org.apache.spark.scheduler.Task.run(Task.scala:89)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
ERROR 23-12 09:55:26,223 - Executor task launch worker-7
java.lang.RuntimeException: Dictionary file vehindlightleft is locked for
updation. Please try after some time
at scala.sys.package$.error(package.scala:27)
at
org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
at
org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(CarbonGlobalDictionaryRDD.scala:293)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
at org.apache.spark.scheduler.Task.run(Task.scala:89)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
ERROR 23-12 09:55:26,223 - Executor task launch worker-4
java.lang.RuntimeException: Dictionary file vehwindowrearleft is locked for
updation. Please try after some time
at scala.sys.package$.error(package.scala:27)
at
org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
at
org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(CarbonGlobalDictionaryRDD.scala:293)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
at org.apache.spark.scheduler.Task.run(Task.scala:89)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
ERROR 23-12 09:55:26,226 - Exception in task 5.0 in stage 5.0 (TID 9096)
java.lang.RuntimeException: Dictionary file vehsyspwrmod is locked for
updation. Please try after some time
at scala.sys.package$.error(package.scala:27)
at
org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
at
org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(CarbonGlobalDictionaryRDD.scala:293)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
at org.apache.spark.scheduler.Task.run(Task.scala:89)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
ERROR 23-12 09:55:26,226 - Exception in task 13.0 in stage 5.0 (TID 9104)
java.lang.RuntimeException: Dictionary file vehwindowrearleft is locked for
updation. Please try after some time
at scala.sys.package$.error(package.scala:27)
at
org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
at
org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(CarbonGlobalDictionaryRDD.scala:293)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
at org.apache.spark.scheduler.Task.run(Task.scala:89)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
ERROR 23-12 09:55:26,226 - Exception in task 21.0 in stage 5.0 (TID 9112)
java.lang.RuntimeException: Dictionary file vehindlightleft is locked for
updation. Please try after some time
at scala.sys.package$.error(package.scala:27)
at
org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
at
org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(CarbonGlobalDictionaryRDD.scala:293)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
at org.apache.spark.scheduler.Task.run(Task.scala:89)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
ERROR 23-12 09:55:26,231 - Executor task launch worker-5
java.lang.RuntimeException: Dictionary file vehraindetected is locked for
updation. Please try after some time
at scala.sys.package$.error(package.scala:27)
at
org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
at
org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(CarbonGlobalDictionaryRDD.scala:293)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
at org.apache.spark.scheduler.Task.run(Task.scala:89)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
ERROR 23-12 09:55:26,231 - Exception in task 29.0 in stage 5.0 (TID 9120)
java.lang.RuntimeException: Dictionary file vehraindetected is locked for
updation. Please try after some time
at scala.sys.package$.error(package.scala:27)
at
org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
at
org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(CarbonGlobalDictionaryRDD.scala:293)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
at org.apache.spark.scheduler.Task.run(Task.scala:89)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
failed at stage
collect at GlobalDictionaryUtil.scala:800
Thanks & Best Regards,
Lionel
On Fri, Dec 9, 2016 at 9:09 PM, geda <be...@gmail.com> wrote:
> Yes,it work thanks。
> follow wiki quick start,use spark-default.conf , should config
> carbondata.properties
> but i use spark-shell , not include carbon.properties in it.
>
>
>
> 2016-12-09 11:48 GMT+08:00 Liang Chen [via Apache CarbonData Mailing List
> archive] <ml...@n5.nabble.com>:
>
> > Hi
> >
> > Have you solved this issue after applying new configurations?
> >
> > Regards
> > Liang
> >
> > geda wrote
> > hello:
> > i test data in spark locak model ,then load data inpath to table ,works
> > well.
> > but when i use yarn-client modle, with 1w rows , size :940k ,but error
> > happend ,there is no lock find in tmp dir ,i don't know how to
> > debug,help.thanks.
> > spark1.6 hadoop 2.7|2.6 carbondata 0.2
> > local mode: run ok
> > $SPARK_HOME/bin/spark-shell --master local[4] --jars
> /usr/local/spark/lib/
> > carbondata_2.10-0.2.0-incubating-shade-hadoop2.7.1.jar
> >
> >
> > yarn command : run bad
> > $SPARK_HOME/bin/spark-shell --verbose --master yarn-client
> > --driver-memory 1G --driver-cores 1 --executor-memory 4G
> --num-executors
> > 5 --executor-cores 1 --conf "spark.executor.
> extraJavaOptions=-XX:NewRatio=2
> > -XX:PermSize=512m -XX:MaxPermSize=512m -XX:SurvivorRatio=6 -verbose:gc
> > -XX:-PrintGCDetails -XX:+PrintGCTimeStamps " --conf "spark.driver.
> > extraJavaOptions=-XX:MaxPermSize=512m -XX:PermSize=512m" --conf
> > spark.yarn.driver.memoryOverhead=1024 --conf spark.yarn.executor.
> memoryOverhead=3096
> > --jars /usr/local/spark/lib/carbondata_2.10-0.2.0-
> > incubating-shade-hadoop2.7.1.jar
> >
> > import java.io._
> > import org.apache.hadoop.hive.conf.HiveConf
> > import org.apache.spark.sql.CarbonContext
> > val storePath = "hdfs://test:8020/usr/carbondata/store"
> > val cc = new CarbonContext(sc, storePath)
> > cc.setConf(HiveConf.ConfVars.HIVECHECKFILEFORMAT.varname, "false")
> > cc.setConf("carbon.kettle.home","/usr/local/spark/
> carbondata/carbonplugins")
> >
> > cc.sql("CREATE TABLE `LINEORDER3` ( LO_ORDERKEY bigint,
> > LO_LINENUMBER int, LO_CUSTKEY bigint, LO_PARTKEY
> > bigint, LO_SUPPKEY bigint, LO_ORDERDATE int,
> > LO_ORDERPRIOTITY string, LO_SHIPPRIOTITY int, LO_QUANTITY
> int,
> > LO_EXTENDEDPRICE int, LO_ORDTOTALPRICE int, LO_DISCOUNT
> int,
> > LO_REVENUE int, LO_SUPPLYCOST int, LO_TAX
> int,
> > LO_COMMITDATE int, LO_SHIPMODE string ) STORED BY
> > 'carbondata'")
> > cc.sql(s"load data local inpath 'hdfs://test:8020/tmp/lineorder_1w.tbl'
> > into table lineorder3 options('DELIMITER'='|',
> 'FILEHEADER'='LO_ORDERKEY,
> > LO_LINENUMBER, LO_CUSTKEY, LO_PARTKEY , LO_SUPPKEY , LO_ORDERDATE ,
> > LO_ORDERPRIOTITY , LO_SHIPPRIOTITY , LO_QUANTITY ,LO_EXTENDEDPRICE ,
> > LO_ORDTOTALPRICE ,LO_DISCOUNT , LO_REVENUE , LO_SUPPLYCOST, LO_TAX,
> > LO_COMMITDATE, LO_SHIPMODE')")
> >
> > org.apache.spark.SparkException: Job aborted due to stage failure: Task
> 0
> > in stage 2.0 failed 4 times, most recent failure: Lost task 0.3 in stage
> > 2.0 (TID 8, datanode03-bi-dev): java.lang.RuntimeException: Dictionary
> file
> > lo_orderpriotity is locked for updation. Please try after some time
> > at scala.sys.package$.error(package.scala:27)
> > at org.apache.carbondata.spark.rdd.
> CarbonGlobalDictionaryGenerate
> > RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> > at org.apache.carbondata.spark.rdd.
> CarbonGlobalDictionaryGenerate
> > RDD.compute(CarbonGlobalDictionaryRDD.scala:293)
> > at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.
> scala:306)
> >
> > at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> > at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.
> scala:66)
> >
> > at org.apache.spark.scheduler.Task.run(Task.scala:89)
> > at org.apache.spark.executor.Executor$TaskRunner.run(
> Executor.scala:214)
> >
> > at java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1145)
> >
> > at java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:615)
> >
> > at java.lang.Thread.run(Thread.java:745)
> >
> > Driver stacktrace:
> > at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$
> > scheduler$DAGScheduler$$failJobAndIndependentStages(
> DAGScheduler.scala:1431)
> >
> > at org.apache.spark.scheduler.DAGScheduler$$anonfun$
> > abortStage$1.apply(DAGScheduler.scala:1419)
> > at org.apache.spark.scheduler.DAGScheduler$$anonfun$
> > abortStage$1.apply(DAGScheduler.scala:1418)
> > at scala.collection.mutable.ResizableArray$class.foreach(
> ResizableArray.scala:59)
> >
> > at scala.collection.mutable.ArrayBuffer.foreach(
> ArrayBuffer.scala:47)
> >
> > at org.apache.spark.scheduler.DAGScheduler.abortStage(
> DAGScheduler.scala:1418)
> >
> > at org.apache.spark.scheduler.DAGScheduler$$anonfun$
> > handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
> > at org.apache.spark.scheduler.DAGScheduler$$anonfun$
> > handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
> > at scala.Option.foreach(Option.scala:236)
> > at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(
> DAGScheduler.scala:799)
> >
> > at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.
> > doOnReceive(DAGScheduler.scala:1640)
> > at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.
> > onReceive(DAGScheduler.scala:1599)
> > at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.
> > onReceive(DAGScheduler.scala:1588)
> > at org.apache.spark.util.EventLoop$$anon$1.run(
> EventLoop.scala:48)
> >
> > at org.apache.spark.scheduler.DAGScheduler.runJob(
> DAGScheduler.scala:620)
> >
> > at org.apache.spark.SparkContext.runJob(SparkContext.scala:1832)
> > at org.apache.spark.SparkContext.runJob(SparkContext.scala:1845)
> > at org.apache.spark.SparkContext.runJob(SparkContext.scala:1858)
> > at org.apache.spark.SparkContext.runJob(SparkContext.scala:1929)
> > at org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.
> scala:927)
> >
> > at org.apache.spark.rdd.RDDOperationScope$.withScope(
> RDDOperationScope.scala:150)
> >
> > at org.apache.spark.rdd.RDDOperationScope$.withScope(
> RDDOperationScope.scala:111)
> >
> > at org.apache.spark.rdd.RDD.withScope(RDD.scala:316)
> > at org.apache.spark.rdd.RDD.collect(RDD.scala:926)
> > at org.apache.carbondata.spark.util.GlobalDictionaryUtil$.
> > generateGlobalDictionary(GlobalDictionaryUtil.scala:800)
> > at org.apache.spark.sql.execution.command.
> > LoadTableUsingKettle.run(carbonTableSchema.scala:1197)
> > at org.apache.spark.sql.execution.command.LoadTable.
> > run(carbonTableSchema.scala:1036)
> > at org.apache.spark.sql.execution.ExecutedCommand.
> > sideEffectResult$lzycompute(commands.scala:58)
> > at org.apache.spark.sql.execution.ExecutedCommand.
> > sideEffectResult(commands.scala:56)
> > at org.apache.spark.sql.execution.ExecutedCommand.
> doExecute(commands.scala:70)
> >
> > at org.apache.spark.sql.execution.SparkPlan$$anonfun$
> > execute$5.apply(SparkPlan.scala:132)
> > at org.apache.spark.sql.execution.SparkPlan$$anonfun$
> > execute$5.apply(SparkPlan.scala:130)
> > at org.apache.spark.rdd.RDDOperationScope$.withScope(
> RDDOperationScope.scala:150)
> >
> > at org.apache.spark.sql.execution.SparkPlan.execute(
> SparkPlan.scala:130)
> >
> > at org.apache.spark.sql.execution.QueryExecution.
> toRdd$lzycompute(QueryExecution.scala:55)
> >
> > at org.apache.spark.sql.execution.QueryExecution.
> toRdd(QueryExecution.scala:55)
> >
> > at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:145)
> > at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:130)
> > at org.apache.carbondata.spark.rdd.CarbonDataFrameRDD.<init>(
> CarbonDataFrameRDD.scala:23)
> >
> > at org.apache.spark.sql.CarbonContext.sql(
> CarbonContext.scala:137)
> >
> > at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$
> iwC.<init>(<console>:37)
> >
> > at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<
> init>(<console>:42)
> >
> > at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>
> (<console>:44)
> >
> > at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<
> console>:46)
> >
> > at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>
> :48)
> >
> > at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:50)
> > at $line32.$read$$iwC$$iwC$$iwC$$iwC.<init>(<console>:52)
> > at $line32.$read$$iwC$$iwC$$iwC.<init>(<console>:54)
> > at $line32.$read$$iwC$$iwC.<init>(<console>:56)
> > at $line32.$read$$iwC.<init>(<console>:58)
> > at $line32.$read.<init>(<console>:60)
> > at $line32.$read$.<init>(<console>:64)
> > at $line32.$read$.<clinit>(<console>)
> > at $line32.$eval$.<init>(<console>:7)
> > at $line32.$eval$.<clinit>(<console>)
> > at $line32.$eval.$print(<console>)
> > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> > at sun.reflect.NativeMethodAccessorImpl.invoke(
> > NativeMethodAccessorImpl.java:57)
> > at sun.reflect.DelegatingMethodAccessorImpl.invoke(
> > DelegatingMethodAccessorImpl.java:43)
> > at java.lang.reflect.Method.invoke(Method.java:606)
> > at org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(
> SparkIMain.scala:1065)
> >
> > at org.apache.spark.repl.SparkIMain$Request.loadAndRun(
> SparkIMain.scala:1346)
> >
> > at org.apache.spark.repl.SparkIMain.loadAndRunReq$1(
> SparkIMain.scala:840)
> >
> > at org.apache.spark.repl.SparkIMain.interpret(
> SparkIMain.scala:871)
> >
> > at org.apache.spark.repl.SparkIMain.interpret(
> SparkIMain.scala:819)
> >
> > at org.apache.spark.repl.SparkILoop.reallyInterpret$1(
> SparkILoop.scala:857)
> >
> > at org.apache.spark.repl.SparkILoop.interpretStartingWith(
> SparkILoop.scala:902)
> >
> > at org.apache.spark.repl.SparkILoop.command(SparkILoop.
> scala:814)
> > at org.apache.spark.repl.SparkILoop.processLine$1(
> SparkILoop.scala:657)
> >
> > at org.apache.spark.repl.SparkILoop.innerLoop$1(
> SparkILoop.scala:665)
> >
> > at org.apache.spark.repl.SparkILoop.org$apache$spark$
> > repl$SparkILoop$$loop(SparkILoop.scala:670)
> > at org.apache.spark.repl.SparkILoop$$anonfun$org$
> > apache$spark$repl$SparkILoop$$process$1.apply$mcZ$sp(
> SparkILoop.scala:997)
> >
> > at org.apache.spark.repl.SparkILoop$$anonfun$org$
> > apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945)
> > at org.apache.spark.repl.SparkILoop$$anonfun$org$
> > apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945)
> > at scala.tools.nsc.util.ScalaClassLoader$.savingContextLoader(
> ScalaClassLoader.scala:135)
> >
> > at org.apache.spark.repl.SparkILoop.org$apache$spark$
> > repl$SparkILoop$$process(SparkILoop.scala:945)
> > at org.apache.spark.repl.SparkILoop.process(SparkILoop.
> scala:1059)
> >
> > at org.apache.spark.repl.Main$.main(Main.scala:31)
> > at org.apache.spark.repl.Main.main(Main.scala)
> > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> > at sun.reflect.NativeMethodAccessorImpl.invoke(
> > NativeMethodAccessorImpl.java:57)
> > at sun.reflect.DelegatingMethodAccessorImpl.invoke(
> > DelegatingMethodAccessorImpl.java:43)
> > at java.lang.reflect.Method.invoke(Method.java:606)
> > at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$
> > deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
> > at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(
> SparkSubmit.scala:181)
> >
> > at org.apache.spark.deploy.SparkSubmit$.submit(
> SparkSubmit.scala:206)
> >
> > at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.
> scala:121)
> >
> > at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
> > Caused by: java.lang.RuntimeException: Dictionary file lo_orderpriotity
> is
> > locked for updation. Please try after some time
> > at scala.sys.package$.error(package.scala:27)
> >
> >
> >
> > ------------------------------
> > If you reply to this email, your message will be added to the discussion
> > below:
> > http://apache-carbondata-mailing-list-archive.1130556.
> > n5.nabble.com/carbondata-0-2-load-data-failed-in-yarn-
> > molde-tp3908p3973.html
> > To unsubscribe from carbondata-0.2 load data failed in yarn molde, click
> > here
> > <http://apache-carbondata-mailing-list-archive.1130556.
> n5.nabble.com/template/NamlServlet.jtp?macro=
> unsubscribe_by_code&node=3908&code=YmVpZG91NzdAZ21haWwuY29tfDM5MD
> h8MTU5NTU5NDExOQ==>
> > .
> > NAML
> > <http://apache-carbondata-mailing-list-archive.1130556.
> n5.nabble.com/template/NamlServlet.jtp?macro=macro_viewer&id=instant_html%
> 21nabble%3Aemail.naml&base=nabble.naml.namespaces.
> BasicNamespace-nabble.view.web.template.NabbleNamespace-
> nabble.view.web.template.NodeNamespace&breadcrumbs=
> notify_subscribers%21nabble%3Aemail.naml-instant_emails%
> 21nabble%3Aemail.naml-send_instant_email%21nabble%3Aemail.naml>
> >
>
>
>
>
> --
> View this message in context: http://apache-carbondata-
> mailing-list-archive.1130556.n5.nabble.com/carbondata-0-2-
> load-data-failed-in-yarn-molde-tp3908p4012.html
> Sent from the Apache CarbonData Mailing List archive mailing list archive
> at Nabble.com.
>
Re: carbondata-0.2 load data failed in yarn molde
Posted by geda <be...@gmail.com>.
Yes,it work thanks。
follow wiki quick start,use spark-default.conf , should config
carbondata.properties
but i use spark-shell , not include carbon.properties in it.
2016-12-09 11:48 GMT+08:00 Liang Chen [via Apache CarbonData Mailing List
archive] <ml...@n5.nabble.com>:
> Hi
>
> Have you solved this issue after applying new configurations?
>
> Regards
> Liang
>
> geda wrote
> hello:
> i test data in spark locak model ,then load data inpath to table ,works
> well.
> but when i use yarn-client modle, with 1w rows , size :940k ,but error
> happend ,there is no lock find in tmp dir ,i don't know how to
> debug,help.thanks.
> spark1.6 hadoop 2.7|2.6 carbondata 0.2
> local mode: run ok
> $SPARK_HOME/bin/spark-shell --master local[4] --jars /usr/local/spark/lib/
> carbondata_2.10-0.2.0-incubating-shade-hadoop2.7.1.jar
>
>
> yarn command : run bad
> $SPARK_HOME/bin/spark-shell --verbose --master yarn-client
> --driver-memory 1G --driver-cores 1 --executor-memory 4G --num-executors
> 5 --executor-cores 1 --conf "spark.executor.extraJavaOptions=-XX:NewRatio=2
> -XX:PermSize=512m -XX:MaxPermSize=512m -XX:SurvivorRatio=6 -verbose:gc
> -XX:-PrintGCDetails -XX:+PrintGCTimeStamps " --conf "spark.driver.
> extraJavaOptions=-XX:MaxPermSize=512m -XX:PermSize=512m" --conf
> spark.yarn.driver.memoryOverhead=1024 --conf spark.yarn.executor.memoryOverhead=3096
> --jars /usr/local/spark/lib/carbondata_2.10-0.2.0-
> incubating-shade-hadoop2.7.1.jar
>
> import java.io._
> import org.apache.hadoop.hive.conf.HiveConf
> import org.apache.spark.sql.CarbonContext
> val storePath = "hdfs://test:8020/usr/carbondata/store"
> val cc = new CarbonContext(sc, storePath)
> cc.setConf(HiveConf.ConfVars.HIVECHECKFILEFORMAT.varname, "false")
> cc.setConf("carbon.kettle.home","/usr/local/spark/carbondata/carbonplugins")
>
> cc.sql("CREATE TABLE `LINEORDER3` ( LO_ORDERKEY bigint,
> LO_LINENUMBER int, LO_CUSTKEY bigint, LO_PARTKEY
> bigint, LO_SUPPKEY bigint, LO_ORDERDATE int,
> LO_ORDERPRIOTITY string, LO_SHIPPRIOTITY int, LO_QUANTITY int,
> LO_EXTENDEDPRICE int, LO_ORDTOTALPRICE int, LO_DISCOUNT int,
> LO_REVENUE int, LO_SUPPLYCOST int, LO_TAX int,
> LO_COMMITDATE int, LO_SHIPMODE string ) STORED BY
> 'carbondata'")
> cc.sql(s"load data local inpath 'hdfs://test:8020/tmp/lineorder_1w.tbl'
> into table lineorder3 options('DELIMITER'='|', 'FILEHEADER'='LO_ORDERKEY,
> LO_LINENUMBER, LO_CUSTKEY, LO_PARTKEY , LO_SUPPKEY , LO_ORDERDATE ,
> LO_ORDERPRIOTITY , LO_SHIPPRIOTITY , LO_QUANTITY ,LO_EXTENDEDPRICE ,
> LO_ORDTOTALPRICE ,LO_DISCOUNT , LO_REVENUE , LO_SUPPLYCOST, LO_TAX,
> LO_COMMITDATE, LO_SHIPMODE')")
>
> org.apache.spark.SparkException: Job aborted due to stage failure: Task 0
> in stage 2.0 failed 4 times, most recent failure: Lost task 0.3 in stage
> 2.0 (TID 8, datanode03-bi-dev): java.lang.RuntimeException: Dictionary file
> lo_orderpriotity is locked for updation. Please try after some time
> at scala.sys.package$.error(package.scala:27)
> at org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> RDD$$anon$1.<init>(CarbonGlobalDictionaryRDD.scala:353)
> at org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerate
> RDD.compute(CarbonGlobalDictionaryRDD.scala:293)
> at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
>
> at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
>
> at org.apache.spark.scheduler.Task.run(Task.scala:89)
> at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
>
> at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>
> at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>
> at java.lang.Thread.run(Thread.java:745)
>
> Driver stacktrace:
> at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$
> scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1431)
>
> at org.apache.spark.scheduler.DAGScheduler$$anonfun$
> abortStage$1.apply(DAGScheduler.scala:1419)
> at org.apache.spark.scheduler.DAGScheduler$$anonfun$
> abortStage$1.apply(DAGScheduler.scala:1418)
> at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
>
> at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
>
> at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1418)
>
> at org.apache.spark.scheduler.DAGScheduler$$anonfun$
> handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
> at org.apache.spark.scheduler.DAGScheduler$$anonfun$
> handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
> at scala.Option.foreach(Option.scala:236)
> at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:799)
>
> at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.
> doOnReceive(DAGScheduler.scala:1640)
> at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.
> onReceive(DAGScheduler.scala:1599)
> at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.
> onReceive(DAGScheduler.scala:1588)
> at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)
>
> at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:620)
>
> at org.apache.spark.SparkContext.runJob(SparkContext.scala:1832)
> at org.apache.spark.SparkContext.runJob(SparkContext.scala:1845)
> at org.apache.spark.SparkContext.runJob(SparkContext.scala:1858)
> at org.apache.spark.SparkContext.runJob(SparkContext.scala:1929)
> at org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.scala:927)
>
> at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
>
> at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111)
>
> at org.apache.spark.rdd.RDD.withScope(RDD.scala:316)
> at org.apache.spark.rdd.RDD.collect(RDD.scala:926)
> at org.apache.carbondata.spark.util.GlobalDictionaryUtil$.
> generateGlobalDictionary(GlobalDictionaryUtil.scala:800)
> at org.apache.spark.sql.execution.command.
> LoadTableUsingKettle.run(carbonTableSchema.scala:1197)
> at org.apache.spark.sql.execution.command.LoadTable.
> run(carbonTableSchema.scala:1036)
> at org.apache.spark.sql.execution.ExecutedCommand.
> sideEffectResult$lzycompute(commands.scala:58)
> at org.apache.spark.sql.execution.ExecutedCommand.
> sideEffectResult(commands.scala:56)
> at org.apache.spark.sql.execution.ExecutedCommand.doExecute(commands.scala:70)
>
> at org.apache.spark.sql.execution.SparkPlan$$anonfun$
> execute$5.apply(SparkPlan.scala:132)
> at org.apache.spark.sql.execution.SparkPlan$$anonfun$
> execute$5.apply(SparkPlan.scala:130)
> at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
>
> at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:130)
>
> at org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:55)
>
> at org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:55)
>
> at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:145)
> at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:130)
> at org.apache.carbondata.spark.rdd.CarbonDataFrameRDD.<init>(CarbonDataFrameRDD.scala:23)
>
> at org.apache.spark.sql.CarbonContext.sql(CarbonContext.scala:137)
>
> at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:37)
>
> at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:42)
>
> at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:44)
>
> at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:46)
>
> at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:48)
>
> at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:50)
> at $line32.$read$$iwC$$iwC$$iwC$$iwC.<init>(<console>:52)
> at $line32.$read$$iwC$$iwC$$iwC.<init>(<console>:54)
> at $line32.$read$$iwC$$iwC.<init>(<console>:56)
> at $line32.$read$$iwC.<init>(<console>:58)
> at $line32.$read.<init>(<console>:60)
> at $line32.$read$.<init>(<console>:64)
> at $line32.$read$.<clinit>(<console>)
> at $line32.$eval$.<init>(<console>:7)
> at $line32.$eval$.<clinit>(<console>)
> at $line32.$eval.$print(<console>)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at sun.reflect.NativeMethodAccessorImpl.invoke(
> NativeMethodAccessorImpl.java:57)
> at sun.reflect.DelegatingMethodAccessorImpl.invoke(
> DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:606)
> at org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(SparkIMain.scala:1065)
>
> at org.apache.spark.repl.SparkIMain$Request.loadAndRun(SparkIMain.scala:1346)
>
> at org.apache.spark.repl.SparkIMain.loadAndRunReq$1(SparkIMain.scala:840)
>
> at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:871)
>
> at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:819)
>
> at org.apache.spark.repl.SparkILoop.reallyInterpret$1(SparkILoop.scala:857)
>
> at org.apache.spark.repl.SparkILoop.interpretStartingWith(SparkILoop.scala:902)
>
> at org.apache.spark.repl.SparkILoop.command(SparkILoop.scala:814)
> at org.apache.spark.repl.SparkILoop.processLine$1(SparkILoop.scala:657)
>
> at org.apache.spark.repl.SparkILoop.innerLoop$1(SparkILoop.scala:665)
>
> at org.apache.spark.repl.SparkILoop.org$apache$spark$
> repl$SparkILoop$$loop(SparkILoop.scala:670)
> at org.apache.spark.repl.SparkILoop$$anonfun$org$
> apache$spark$repl$SparkILoop$$process$1.apply$mcZ$sp(SparkILoop.scala:997)
>
> at org.apache.spark.repl.SparkILoop$$anonfun$org$
> apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945)
> at org.apache.spark.repl.SparkILoop$$anonfun$org$
> apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945)
> at scala.tools.nsc.util.ScalaClassLoader$.savingContextLoader(ScalaClassLoader.scala:135)
>
> at org.apache.spark.repl.SparkILoop.org$apache$spark$
> repl$SparkILoop$$process(SparkILoop.scala:945)
> at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:1059)
>
> at org.apache.spark.repl.Main$.main(Main.scala:31)
> at org.apache.spark.repl.Main.main(Main.scala)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at sun.reflect.NativeMethodAccessorImpl.invoke(
> NativeMethodAccessorImpl.java:57)
> at sun.reflect.DelegatingMethodAccessorImpl.invoke(
> DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:606)
> at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$
> deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
> at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181)
>
> at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206)
>
> at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
>
> at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
> Caused by: java.lang.RuntimeException: Dictionary file lo_orderpriotity is
> locked for updation. Please try after some time
> at scala.sys.package$.error(package.scala:27)
>
>
>
> ------------------------------
> If you reply to this email, your message will be added to the discussion
> below:
> http://apache-carbondata-mailing-list-archive.1130556.
> n5.nabble.com/carbondata-0-2-load-data-failed-in-yarn-
> molde-tp3908p3973.html
> To unsubscribe from carbondata-0.2 load data failed in yarn molde, click
> here
> <http://apache-carbondata-mailing-list-archive.1130556.n5.nabble.com/template/NamlServlet.jtp?macro=unsubscribe_by_code&node=3908&code=YmVpZG91NzdAZ21haWwuY29tfDM5MDh8MTU5NTU5NDExOQ==>
> .
> NAML
> <http://apache-carbondata-mailing-list-archive.1130556.n5.nabble.com/template/NamlServlet.jtp?macro=macro_viewer&id=instant_html%21nabble%3Aemail.naml&base=nabble.naml.namespaces.BasicNamespace-nabble.view.web.template.NabbleNamespace-nabble.view.web.template.NodeNamespace&breadcrumbs=notify_subscribers%21nabble%3Aemail.naml-instant_emails%21nabble%3Aemail.naml-send_instant_email%21nabble%3Aemail.naml>
>
--
View this message in context: http://apache-carbondata-mailing-list-archive.1130556.n5.nabble.com/carbondata-0-2-load-data-failed-in-yarn-molde-tp3908p4012.html
Sent from the Apache CarbonData Mailing List archive mailing list archive at Nabble.com.
Re: carbondata-0.2 load data failed in yarn molde
Posted by Liang Chen <ch...@gmail.com>.
Hi
Have you solved this issue after applying new configurations?
Regards
Liang
geda wrote
> hello:
> i test data in spark locak model ,then load data inpath to table ,works
> well.
> but when i use yarn-client modle, with 1w rows , size :940k ,but error
> happend ,there is no lock find in tmp dir ,i don't know how to
> debug,help.thanks.
> spark1.6 hadoop 2.7|2.6 carbondata 0.2
> local mode: run ok
> $SPARK_HOME/bin/spark-shell --master local[4] --jars
> /usr/local/spark/lib/carbondata_2.10-0.2.0-incubating-shade-hadoop2.7.1.jar
>
>
> yarn command : run bad
> $SPARK_HOME/bin/spark-shell --verbose --master yarn-client
> --driver-memory 1G --driver-cores 1 --executor-memory 4G --num-executors
> 5 --executor-cores 1 --conf
> "spark.executor.extraJavaOptions=-XX:NewRatio=2 -XX:PermSize=512m
> -XX:MaxPermSize=512m -XX:SurvivorRatio=6 -verbose:gc -XX:-PrintGCDetails
> -XX:+PrintGCTimeStamps " --conf
> "spark.driver.extraJavaOptions=-XX:MaxPermSize=512m -XX:PermSize=512m"
> --conf spark.yarn.driver.memoryOverhead=1024 --conf
> spark.yarn.executor.memoryOverhead=3096 --jars
> /usr/local/spark/lib/carbondata_2.10-0.2.0-incubating-shade-hadoop2.7.1.jar
>
> import java.io._
> import org.apache.hadoop.hive.conf.HiveConf
> import org.apache.spark.sql.CarbonContext
> val storePath = "hdfs://test:8020/usr/carbondata/store"
> val cc = new CarbonContext(sc, storePath)
> cc.setConf(HiveConf.ConfVars.HIVECHECKFILEFORMAT.varname, "false")
> cc.setConf("carbon.kettle.home","/usr/local/spark/carbondata/carbonplugins")
> cc.sql("CREATE TABLE `LINEORDER3` ( LO_ORDERKEY bigint,
> LO_LINENUMBER int, LO_CUSTKEY bigint, LO_PARTKEY
> bigint, LO_SUPPKEY bigint, LO_ORDERDATE int,
> LO_ORDERPRIOTITY string, LO_SHIPPRIOTITY int, LO_QUANTITY
> int, LO_EXTENDEDPRICE int, LO_ORDTOTALPRICE int, LO_DISCOUNT
> int, LO_REVENUE int, LO_SUPPLYCOST int, LO_TAX
> int, LO_COMMITDATE int, LO_SHIPMODE string ) STORED BY
> 'carbondata'")
> cc.sql(s"load data local inpath 'hdfs://test:8020/tmp/lineorder_1w.tbl'
> into table lineorder3 options('DELIMITER'='|', 'FILEHEADER'='LO_ORDERKEY,
> LO_LINENUMBER, LO_CUSTKEY, LO_PARTKEY , LO_SUPPKEY , LO_ORDERDATE ,
> LO_ORDERPRIOTITY , LO_SHIPPRIOTITY , LO_QUANTITY ,LO_EXTENDEDPRICE ,
> LO_ORDTOTALPRICE ,LO_DISCOUNT , LO_REVENUE , LO_SUPPLYCOST, LO_TAX,
> LO_COMMITDATE, LO_SHIPMODE')")
>
> org.apache.spark.SparkException: Job aborted due to stage failure: Task 0
> in stage 2.0 failed 4 times, most recent failure: Lost task 0.3 in stage
> 2.0 (TID 8, datanode03-bi-dev): java.lang.RuntimeException: Dictionary
> file lo_orderpriotity is locked for updation. Please try after some time
> at scala.sys.package$.error(package.scala:27)
> at
> org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD$$anon$1.
> <init>
> (CarbonGlobalDictionaryRDD.scala:353)
> at
> org.apache.carbondata.spark.rdd.CarbonGlobalDictionaryGenerateRDD.compute(CarbonGlobalDictionaryRDD.scala:293)
> at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> at
> org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> at org.apache.spark.scheduler.Task.run(Task.scala:89)
> at
> org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:745)
>
> Driver stacktrace:
> at
> org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1431)
> at
> org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1419)
> at
> org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1418)
> at
> scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
> at
> scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
> at
> org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1418)
> at
> org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
> at
> org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
> at scala.Option.foreach(Option.scala:236)
> at
> org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:799)
> at
> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1640)
> at
> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1599)
> at
> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1588)
> at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)
> at
> org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:620)
> at org.apache.spark.SparkContext.runJob(SparkContext.scala:1832)
> at org.apache.spark.SparkContext.runJob(SparkContext.scala:1845)
> at org.apache.spark.SparkContext.runJob(SparkContext.scala:1858)
> at org.apache.spark.SparkContext.runJob(SparkContext.scala:1929)
> at
> org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.scala:927)
> at
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
> at
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111)
> at org.apache.spark.rdd.RDD.withScope(RDD.scala:316)
> at org.apache.spark.rdd.RDD.collect(RDD.scala:926)
> at
> org.apache.carbondata.spark.util.GlobalDictionaryUtil$.generateGlobalDictionary(GlobalDictionaryUtil.scala:800)
> at
> org.apache.spark.sql.execution.command.LoadTableUsingKettle.run(carbonTableSchema.scala:1197)
> at
> org.apache.spark.sql.execution.command.LoadTable.run(carbonTableSchema.scala:1036)
> at
> org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult$lzycompute(commands.scala:58)
> at
> org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult(commands.scala:56)
> at
> org.apache.spark.sql.execution.ExecutedCommand.doExecute(commands.scala:70)
> at
> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:132)
> at
> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:130)
> at
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
> at
> org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:130)
> at
> org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:55)
> at
> org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:55)
> at org.apache.spark.sql.DataFrame.
> <init>
> (DataFrame.scala:145)
> at org.apache.spark.sql.DataFrame.
> <init>
> (DataFrame.scala:130)
> at org.apache.carbondata.spark.rdd.CarbonDataFrameRDD.
> <init>
> (CarbonDataFrameRDD.scala:23)
> at org.apache.spark.sql.CarbonContext.sql(CarbonContext.scala:137)
> at
> $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.
> <init>
> (
> <console>
> :37)
> at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.
> <init>
> (
> <console>
> :42)
> at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.
> <init>
> (
> <console>
> :44)
> at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.
> <init>
> (
> <console>
> :46)
> at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.
> <init>
> (
> <console>
> :48)
> at $line32.$read$$iwC$$iwC$$iwC$$iwC$$iwC.
> <init>
> (
> <console>
> :50)
> at $line32.$read$$iwC$$iwC$$iwC$$iwC.
> <init>
> (
> <console>
> :52)
> at $line32.$read$$iwC$$iwC$$iwC.
> <init>
> (
> <console>
> :54)
> at $line32.$read$$iwC$$iwC.
> <init>
> (
> <console>
> :56)
> at $line32.$read$$iwC.
> <init>
> (
> <console>
> :58)
> at $line32.$read.
> <init>
> (
> <console>
> :60)
> at $line32.$read$.
> <init>
> (
> <console>
> :64)
> at $line32.$read$.
> <clinit>
> (
> <console>
> )
> at $line32.$eval$.
> <init>
> (
> <console>
> :7)
> at $line32.$eval$.
> <clinit>
> (
> <console>
> )
> at $line32.$eval.$print(
> <console>
> )
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:606)
> at
> org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(SparkIMain.scala:1065)
> at
> org.apache.spark.repl.SparkIMain$Request.loadAndRun(SparkIMain.scala:1346)
> at
> org.apache.spark.repl.SparkIMain.loadAndRunReq$1(SparkIMain.scala:840)
> at
> org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:871)
> at
> org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:819)
> at
> org.apache.spark.repl.SparkILoop.reallyInterpret$1(SparkILoop.scala:857)
> at
> org.apache.spark.repl.SparkILoop.interpretStartingWith(SparkILoop.scala:902)
> at org.apache.spark.repl.SparkILoop.command(SparkILoop.scala:814)
> at
> org.apache.spark.repl.SparkILoop.processLine$1(SparkILoop.scala:657)
> at
> org.apache.spark.repl.SparkILoop.innerLoop$1(SparkILoop.scala:665)
> at
> org.apache.spark.repl.SparkILoop.org$apache$spark$repl$SparkILoop$$loop(SparkILoop.scala:670)
> at
> org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl$SparkILoop$$process$1.apply$mcZ$sp(SparkILoop.scala:997)
> at
> org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945)
> at
> org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945)
> at
> scala.tools.nsc.util.ScalaClassLoader$.savingContextLoader(ScalaClassLoader.scala:135)
> at
> org.apache.spark.repl.SparkILoop.org$apache$spark$repl$SparkILoop$$process(SparkILoop.scala:945)
> at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:1059)
> at org.apache.spark.repl.Main$.main(Main.scala:31)
> at org.apache.spark.repl.Main.main(Main.scala)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:606)
> at
> org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
> at
> org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181)
> at
> org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206)
> at
> org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
> at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
> Caused by: java.lang.RuntimeException: Dictionary file lo_orderpriotity is
> locked for updation. Please try after some time
> at scala.sys.package$.error(package.scala:27)
--
View this message in context: http://apache-carbondata-mailing-list-archive.1130556.n5.nabble.com/carbondata-0-2-load-data-failed-in-yarn-molde-tp3908p3973.html
Sent from the Apache CarbonData Mailing List archive mailing list archive at Nabble.com.