You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@kylin.apache.org by 仇同心 <qi...@jd.com> on 2017/07/04 01:51:29 UTC
Detailed query of "RAW"
Hi,all
There is a business scenario is the detailed query. There are 45 metric for design to “RAW”, one dimension and high base,but in “Rowkey” of “Advanced Setting”,the Encoding is fixed._length,not dict.
But when build the cube, there are some errors in “#4 Step Name: Build Dimension Dictionary”:
com.google.common.util.concurrent.ExecutionError: java.lang.OutOfMemoryError: Java heap space
at com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2254)
at com.google.common.cache.LocalCache.get(LocalCache.java:3985)
at com.google.common.cache.LocalCache.getOrLoad(LocalCache.java:3989)
at com.google.common.cache.LocalCache$LocalLoadingCache.get(LocalCache.java:4873)
at org.apache.kylin.dict.DictionaryManager.getDictionaryInfo(DictionaryManager.java:119)
at org.apache.kylin.dict.DictionaryManager.checkDupByContent(DictionaryManager.java:191)
at org.apache.kylin.dict.DictionaryManager.trySaveNewDict(DictionaryManager.java:169)
at org.apache.kylin.dict.DictionaryManager.saveDictionary(DictionaryManager.java:324)
at org.apache.kylin.cube.CubeManager.saveDictionary(CubeManager.java:234)
at org.apache.kylin.cube.cli.DictionaryGeneratorCLI.processSegment(DictionaryGeneratorCLI.java:68)
at org.apache.kylin.cube.cli.DictionaryGeneratorCLI.processSegment(DictionaryGeneratorCLI.java:54)
at org.apache.kylin.engine.mr.steps.CreateDictionaryJob.run(CreateDictionaryJob.java:66)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
at org.apache.kylin.engine.mr.common.HadoopShellExecutable.doWork(HadoopShellExecutable.java:63)
at org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:124)
at org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:64)
at org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:124)
at org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:142)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.OutOfMemoryError: Java heap space
at org.apache.kylin.dict.CacheDictionary.enableCache(CacheDictionary.java:95)
at org.apache.kylin.dict.TrieDictionaryForest.initForestCache(TrieDictionaryForest.java:394)
at org.apache.kylin.dict.TrieDictionaryForest.init(TrieDictionaryForest.java:77)
at org.apache.kylin.dict.TrieDictionaryForest.readFields(TrieDictionaryForest.java:237)
at org.apache.kylin.dict.DictionaryInfoSerializer.deserialize(DictionaryInfoSerializer.java:74)
at org.apache.kylin.dict.DictionaryInfoSerializer.deserialize(DictionaryInfoSerializer.java:34)
at org.apache.kylin.common.persistence.ResourceStore.getResource(ResourceStore.java:154)
at org.apache.kylin.dict.DictionaryManager.load(DictionaryManager.java:445)
at org.apache.kylin.dict.DictionaryManager$1.load(DictionaryManager.java:102)
at org.apache.kylin.dict.DictionaryManager$1.load(DictionaryManager.java:99)
at com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3584)
at com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2372)
at com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2335)
at com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2250)
at com.google.common.cache.LocalCache.get(LocalCache.java:3985)
at com.google.common.cache.LocalCache.getOrLoad(LocalCache.java:3989)
at com.google.common.cache.LocalCache$LocalLoadingCache.get(LocalCache.java:4873)
at org.apache.kylin.dict.DictionaryManager.getDictionaryInfo(DictionaryManager.java:119)
at org.apache.kylin.dict.DictionaryManager.checkDupByContent(DictionaryManager.java:191)
at org.apache.kylin.dict.DictionaryManager.trySaveNewDict(DictionaryManager.java:169)
at org.apache.kylin.dict.DictionaryManager.saveDictionary(DictionaryManager.java:324)
at org.apache.kylin.cube.CubeManager.saveDictionary(CubeManager.java:234)
at org.apache.kylin.cube.cli.DictionaryGeneratorCLI.processSegment(DictionaryGeneratorCLI.java:68)
at org.apache.kylin.cube.cli.DictionaryGeneratorCLI.processSegment(DictionaryGeneratorCLI.java:54)
at org.apache.kylin.engine.mr.steps.CreateDictionaryJob.run(CreateDictionaryJob.java:66)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
at org.apache.kylin.engine.mr.common.HadoopShellExecutable.doWork(HadoopShellExecutable.java:63)
at org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:124)
at org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:64)
at org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:124)
at org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:142)
Sometimes this is wrong:
java.lang.NegativeArraySizeException
at org.apache.hadoop.io.BytesWritable.setCapacity(BytesWritable.java:144)
at org.apache.hadoop.io.BytesWritable.setSize(BytesWritable.java:123)
at org.apache.hadoop.io.BytesWritable.readFields(BytesWritable.java:179)
at org.apache.hadoop.io.SequenceFile$Reader.getCurrentValue(SequenceFile.java:2178)
at org.apache.hadoop.io.SequenceFile$Reader.next(SequenceFile.java:2306)
at org.apache.kylin.engine.mr.steps.CreateDictionaryJob$2.getDictionary(CreateDictionaryJob.java:87)
at org.apache.kylin.cube.cli.DictionaryGeneratorCLI.processSegment(DictionaryGeneratorCLI.java:65)
at org.apache.kylin.cube.cli.DictionaryGeneratorCLI.processSegment(DictionaryGeneratorCLI.java:54)
at org.apache.kylin.engine.mr.steps.CreateDictionaryJob.run(CreateDictionaryJob.java:66)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
at org.apache.kylin.engine.mr.common.HadoopShellExecutable.doWork(HadoopShellExecutable.java:63)
at org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:124)
at org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:64)
at org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:124)
at org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:142)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
result code:2
Now I use kylin 2.0,can you help me ?
Tkanks!
Re: Detailed query of "RAW"
Posted by Billy Liu <bi...@apache.org>.
The NegativeArraySizeException error means the dictionary entry has been
"OutOfRange". Kylin is not designed for detailed query, the current "RAW"
measure still has many limitations.
2017-07-04 9:51 GMT+08:00 仇同心 <qi...@jd.com>:
> Hi,all
> There is a business scenario is the detailed query. There are 45 metric
> for design to “RAW”, one dimension and high base,but in “Rowkey” of
> “Advanced Setting”,the Encoding is fixed._length,not dict.
> But when build the cube, there are some errors in “#4 Step Name: Build
> Dimension Dictionary”:
>
> com.google.common.util.concurrent.ExecutionError:
> java.lang.OutOfMemoryError: Java heap space
> at com.google.common.cache.LocalCache$Segment.get(
> LocalCache.java:2254)
> at com.google.common.cache.LocalCache.get(LocalCache.java:3985)
> at com.google.common.cache.LocalCache.getOrLoad(
> LocalCache.java:3989)
> at com.google.common.cache.LocalCache$LocalLoadingCache.
> get(LocalCache.java:4873)
> at org.apache.kylin.dict.DictionaryManager.getDictionaryInfo(
> DictionaryManager.java:119)
> at org.apache.kylin.dict.DictionaryManager.checkDupByContent(
> DictionaryManager.java:191)
> at org.apache.kylin.dict.DictionaryManager.trySaveNewDict(
> DictionaryManager.java:169)
> at org.apache.kylin.dict.DictionaryManager.saveDictionary(
> DictionaryManager.java:324)
> at org.apache.kylin.cube.CubeManager.saveDictionary(
> CubeManager.java:234)
> at org.apache.kylin.cube.cli.DictionaryGeneratorCLI.
> processSegment(DictionaryGeneratorCLI.java:68)
> at org.apache.kylin.cube.cli.DictionaryGeneratorCLI.
> processSegment(DictionaryGeneratorCLI.java:54)
> at org.apache.kylin.engine.mr.steps.CreateDictionaryJob.run(
> CreateDictionaryJob.java:66)
> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
> at org.apache.kylin.engine.mr.common.HadoopShellExecutable.
> doWork(HadoopShellExecutable.java:63)
> at org.apache.kylin.job.execution.AbstractExecutable.
> execute(AbstractExecutable.java:124)
> at org.apache.kylin.job.execution.DefaultChainedExecutable.
> doWork(DefaultChainedExecutable.java:64)
> at org.apache.kylin.job.execution.AbstractExecutable.
> execute(AbstractExecutable.java:124)
> at org.apache.kylin.job.impl.threadpool.DefaultScheduler$
> JobRunner.run(DefaultScheduler.java:142)
> at java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1145)
> at java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:745)
> Caused by: java.lang.OutOfMemoryError: Java heap space
> at org.apache.kylin.dict.CacheDictionary.enableCache(
> CacheDictionary.java:95)
> at org.apache.kylin.dict.TrieDictionaryForest.initForestCache(
> TrieDictionaryForest.java:394)
> at org.apache.kylin.dict.TrieDictionaryForest.init(
> TrieDictionaryForest.java:77)
> at org.apache.kylin.dict.TrieDictionaryForest.readFields(
> TrieDictionaryForest.java:237)
> at org.apache.kylin.dict.DictionaryInfoSerializer.deserialize(
> DictionaryInfoSerializer.java:74)
> at org.apache.kylin.dict.DictionaryInfoSerializer.deserialize(
> DictionaryInfoSerializer.java:34)
> at org.apache.kylin.common.persistence.ResourceStore.
> getResource(ResourceStore.java:154)
> at org.apache.kylin.dict.DictionaryManager.load(
> DictionaryManager.java:445)
> at org.apache.kylin.dict.DictionaryManager$1.load(
> DictionaryManager.java:102)
> at org.apache.kylin.dict.DictionaryManager$1.load(
> DictionaryManager.java:99)
> at com.google.common.cache.LocalCache$LoadingValueReference.
> loadFuture(LocalCache.java:3584)
> at com.google.common.cache.LocalCache$Segment.loadSync(
> LocalCache.java:2372)
> at com.google.common.cache.LocalCache$Segment.
> lockedGetOrLoad(LocalCache.java:2335)
> at com.google.common.cache.LocalCache$Segment.get(
> LocalCache.java:2250)
> at com.google.common.cache.LocalCache.get(LocalCache.java:3985)
> at com.google.common.cache.LocalCache.getOrLoad(
> LocalCache.java:3989)
> at com.google.common.cache.LocalCache$LocalLoadingCache.
> get(LocalCache.java:4873)
> at org.apache.kylin.dict.DictionaryManager.getDictionaryInfo(
> DictionaryManager.java:119)
> at org.apache.kylin.dict.DictionaryManager.checkDupByContent(
> DictionaryManager.java:191)
> at org.apache.kylin.dict.DictionaryManager.trySaveNewDict(
> DictionaryManager.java:169)
> at org.apache.kylin.dict.DictionaryManager.saveDictionary(
> DictionaryManager.java:324)
> at org.apache.kylin.cube.CubeManager.saveDictionary(
> CubeManager.java:234)
> at org.apache.kylin.cube.cli.DictionaryGeneratorCLI.
> processSegment(DictionaryGeneratorCLI.java:68)
> at org.apache.kylin.cube.cli.DictionaryGeneratorCLI.
> processSegment(DictionaryGeneratorCLI.java:54)
> at org.apache.kylin.engine.mr.steps.CreateDictionaryJob.run(
> CreateDictionaryJob.java:66)
> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
> at org.apache.kylin.engine.mr.common.HadoopShellExecutable.
> doWork(HadoopShellExecutable.java:63)
> at org.apache.kylin.job.execution.AbstractExecutable.
> execute(AbstractExecutable.java:124)
> at org.apache.kylin.job.execution.DefaultChainedExecutable.
> doWork(DefaultChainedExecutable.java:64)
> at org.apache.kylin.job.execution.AbstractExecutable.
> execute(AbstractExecutable.java:124)
> at org.apache.kylin.job.impl.threadpool.DefaultScheduler$
> JobRunner.run(DefaultScheduler.java:142)
>
>
> Sometimes this is wrong:
>
> java.lang.NegativeArraySizeException
> at org.apache.hadoop.io.BytesWritable.setCapacity(
> BytesWritable.java:144)
> at org.apache.hadoop.io.BytesWritable.setSize(
> BytesWritable.java:123)
> at org.apache.hadoop.io.BytesWritable.readFields(
> BytesWritable.java:179)
> at org.apache.hadoop.io.SequenceFile$Reader.
> getCurrentValue(SequenceFile.java:2178)
> at org.apache.hadoop.io.SequenceFile$Reader.next(
> SequenceFile.java:2306)
> at org.apache.kylin.engine.mr.steps.CreateDictionaryJob$2.
> getDictionary(CreateDictionaryJob.java:87)
> at org.apache.kylin.cube.cli.DictionaryGeneratorCLI.
> processSegment(DictionaryGeneratorCLI.java:65)
> at org.apache.kylin.cube.cli.DictionaryGeneratorCLI.
> processSegment(DictionaryGeneratorCLI.java:54)
> at org.apache.kylin.engine.mr.steps.CreateDictionaryJob.run(
> CreateDictionaryJob.java:66)
> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
> at org.apache.kylin.engine.mr.common.HadoopShellExecutable.
> doWork(HadoopShellExecutable.java:63)
> at org.apache.kylin.job.execution.AbstractExecutable.
> execute(AbstractExecutable.java:124)
> at org.apache.kylin.job.execution.DefaultChainedExecutable.
> doWork(DefaultChainedExecutable.java:64)
> at org.apache.kylin.job.execution.AbstractExecutable.
> execute(AbstractExecutable.java:124)
> at org.apache.kylin.job.impl.threadpool.DefaultScheduler$
> JobRunner.run(DefaultScheduler.java:142)
> at java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1145)
> at java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:745)
>
> result code:2
>
>
> Now I use kylin 2.0,can you help me ?
>
>
> Tkanks!
>
Re: Detailed query of "RAW"
Posted by Billy Liu <bi...@apache.org>.
The NegativeArraySizeException error means the dictionary entry has been
"OutOfRange". Kylin is not designed for detailed query, the current "RAW"
measure still has many limitations.
2017-07-04 9:51 GMT+08:00 仇同心 <qi...@jd.com>:
> Hi,all
> There is a business scenario is the detailed query. There are 45 metric
> for design to “RAW”, one dimension and high base,but in “Rowkey” of
> “Advanced Setting”,the Encoding is fixed._length,not dict.
> But when build the cube, there are some errors in “#4 Step Name: Build
> Dimension Dictionary”:
>
> com.google.common.util.concurrent.ExecutionError:
> java.lang.OutOfMemoryError: Java heap space
> at com.google.common.cache.LocalCache$Segment.get(
> LocalCache.java:2254)
> at com.google.common.cache.LocalCache.get(LocalCache.java:3985)
> at com.google.common.cache.LocalCache.getOrLoad(
> LocalCache.java:3989)
> at com.google.common.cache.LocalCache$LocalLoadingCache.
> get(LocalCache.java:4873)
> at org.apache.kylin.dict.DictionaryManager.getDictionaryInfo(
> DictionaryManager.java:119)
> at org.apache.kylin.dict.DictionaryManager.checkDupByContent(
> DictionaryManager.java:191)
> at org.apache.kylin.dict.DictionaryManager.trySaveNewDict(
> DictionaryManager.java:169)
> at org.apache.kylin.dict.DictionaryManager.saveDictionary(
> DictionaryManager.java:324)
> at org.apache.kylin.cube.CubeManager.saveDictionary(
> CubeManager.java:234)
> at org.apache.kylin.cube.cli.DictionaryGeneratorCLI.
> processSegment(DictionaryGeneratorCLI.java:68)
> at org.apache.kylin.cube.cli.DictionaryGeneratorCLI.
> processSegment(DictionaryGeneratorCLI.java:54)
> at org.apache.kylin.engine.mr.steps.CreateDictionaryJob.run(
> CreateDictionaryJob.java:66)
> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
> at org.apache.kylin.engine.mr.common.HadoopShellExecutable.
> doWork(HadoopShellExecutable.java:63)
> at org.apache.kylin.job.execution.AbstractExecutable.
> execute(AbstractExecutable.java:124)
> at org.apache.kylin.job.execution.DefaultChainedExecutable.
> doWork(DefaultChainedExecutable.java:64)
> at org.apache.kylin.job.execution.AbstractExecutable.
> execute(AbstractExecutable.java:124)
> at org.apache.kylin.job.impl.threadpool.DefaultScheduler$
> JobRunner.run(DefaultScheduler.java:142)
> at java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1145)
> at java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:745)
> Caused by: java.lang.OutOfMemoryError: Java heap space
> at org.apache.kylin.dict.CacheDictionary.enableCache(
> CacheDictionary.java:95)
> at org.apache.kylin.dict.TrieDictionaryForest.initForestCache(
> TrieDictionaryForest.java:394)
> at org.apache.kylin.dict.TrieDictionaryForest.init(
> TrieDictionaryForest.java:77)
> at org.apache.kylin.dict.TrieDictionaryForest.readFields(
> TrieDictionaryForest.java:237)
> at org.apache.kylin.dict.DictionaryInfoSerializer.deserialize(
> DictionaryInfoSerializer.java:74)
> at org.apache.kylin.dict.DictionaryInfoSerializer.deserialize(
> DictionaryInfoSerializer.java:34)
> at org.apache.kylin.common.persistence.ResourceStore.
> getResource(ResourceStore.java:154)
> at org.apache.kylin.dict.DictionaryManager.load(
> DictionaryManager.java:445)
> at org.apache.kylin.dict.DictionaryManager$1.load(
> DictionaryManager.java:102)
> at org.apache.kylin.dict.DictionaryManager$1.load(
> DictionaryManager.java:99)
> at com.google.common.cache.LocalCache$LoadingValueReference.
> loadFuture(LocalCache.java:3584)
> at com.google.common.cache.LocalCache$Segment.loadSync(
> LocalCache.java:2372)
> at com.google.common.cache.LocalCache$Segment.
> lockedGetOrLoad(LocalCache.java:2335)
> at com.google.common.cache.LocalCache$Segment.get(
> LocalCache.java:2250)
> at com.google.common.cache.LocalCache.get(LocalCache.java:3985)
> at com.google.common.cache.LocalCache.getOrLoad(
> LocalCache.java:3989)
> at com.google.common.cache.LocalCache$LocalLoadingCache.
> get(LocalCache.java:4873)
> at org.apache.kylin.dict.DictionaryManager.getDictionaryInfo(
> DictionaryManager.java:119)
> at org.apache.kylin.dict.DictionaryManager.checkDupByContent(
> DictionaryManager.java:191)
> at org.apache.kylin.dict.DictionaryManager.trySaveNewDict(
> DictionaryManager.java:169)
> at org.apache.kylin.dict.DictionaryManager.saveDictionary(
> DictionaryManager.java:324)
> at org.apache.kylin.cube.CubeManager.saveDictionary(
> CubeManager.java:234)
> at org.apache.kylin.cube.cli.DictionaryGeneratorCLI.
> processSegment(DictionaryGeneratorCLI.java:68)
> at org.apache.kylin.cube.cli.DictionaryGeneratorCLI.
> processSegment(DictionaryGeneratorCLI.java:54)
> at org.apache.kylin.engine.mr.steps.CreateDictionaryJob.run(
> CreateDictionaryJob.java:66)
> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
> at org.apache.kylin.engine.mr.common.HadoopShellExecutable.
> doWork(HadoopShellExecutable.java:63)
> at org.apache.kylin.job.execution.AbstractExecutable.
> execute(AbstractExecutable.java:124)
> at org.apache.kylin.job.execution.DefaultChainedExecutable.
> doWork(DefaultChainedExecutable.java:64)
> at org.apache.kylin.job.execution.AbstractExecutable.
> execute(AbstractExecutable.java:124)
> at org.apache.kylin.job.impl.threadpool.DefaultScheduler$
> JobRunner.run(DefaultScheduler.java:142)
>
>
> Sometimes this is wrong:
>
> java.lang.NegativeArraySizeException
> at org.apache.hadoop.io.BytesWritable.setCapacity(
> BytesWritable.java:144)
> at org.apache.hadoop.io.BytesWritable.setSize(
> BytesWritable.java:123)
> at org.apache.hadoop.io.BytesWritable.readFields(
> BytesWritable.java:179)
> at org.apache.hadoop.io.SequenceFile$Reader.
> getCurrentValue(SequenceFile.java:2178)
> at org.apache.hadoop.io.SequenceFile$Reader.next(
> SequenceFile.java:2306)
> at org.apache.kylin.engine.mr.steps.CreateDictionaryJob$2.
> getDictionary(CreateDictionaryJob.java:87)
> at org.apache.kylin.cube.cli.DictionaryGeneratorCLI.
> processSegment(DictionaryGeneratorCLI.java:65)
> at org.apache.kylin.cube.cli.DictionaryGeneratorCLI.
> processSegment(DictionaryGeneratorCLI.java:54)
> at org.apache.kylin.engine.mr.steps.CreateDictionaryJob.run(
> CreateDictionaryJob.java:66)
> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
> at org.apache.kylin.engine.mr.common.HadoopShellExecutable.
> doWork(HadoopShellExecutable.java:63)
> at org.apache.kylin.job.execution.AbstractExecutable.
> execute(AbstractExecutable.java:124)
> at org.apache.kylin.job.execution.DefaultChainedExecutable.
> doWork(DefaultChainedExecutable.java:64)
> at org.apache.kylin.job.execution.AbstractExecutable.
> execute(AbstractExecutable.java:124)
> at org.apache.kylin.job.impl.threadpool.DefaultScheduler$
> JobRunner.run(DefaultScheduler.java:142)
> at java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1145)
> at java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:745)
>
> result code:2
>
>
> Now I use kylin 2.0,can you help me ?
>
>
> Tkanks!
>