You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@kylin.apache.org by Mohit Bhatnagar <mo...@adobe.com> on 2015/06/01 13:36:18 UTC

Building cube issue: IncompatibleClassChangeError error

Hi,

I m hitting the below IncompatibleClassChangeError  error. We have the following versions already installed in our Prod cluster and I have also changed the kylin/pom.xml file to have the same versions.
Here are the details:
  <!-- Hadoop versions -->
        <hadoop2.version>2.5.0</hadoop2.version>
        <yarn.version>2.5.0</yarn.version>
        <zookeeper.version>3.4.5</zookeeper.version>
        <hive.version>0.13.1</hive.version>
        <hive-hcatalog.version>0.13.1</hive-hcatalog.version>
        <hbase-hadoop2.version>0.98.6-hadoop2</hbase-hadoop2.version>

>From Hadoop 2.0 onwards JobContext is an interface. Don't know why it is expecting a class:
Here is the log:


[pool-6-thread-1]:[2015-06-01 16:56:53,644][INFO][org.apache.kylin.job.impl.threadpool.DefaultScheduler$FetcherRunner.run(DefaultScheduler.java:117)] - Job Fetcher: 0 running, 1 actual running, 1 ready, 44 others
[pool-7-thread-2]:[2015-06-01 16:57:48,749][INFO][org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:441)] - Cleaning up the staging area /user/mobhatna/.staging/job_1432706588203_13062
[pool-7-thread-2]:[2015-06-01 16:57:49,031][ERROR][org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:109)] - error running Executable
java.lang.IncompatibleClassChangeError: Found interface org.apache.hadoop.mapreduce.JobContext, but class was expected
       at org.apache.hive.hcatalog.mapreduce.HCatBaseInputFormat.getSplits(HCatBaseInputFormat.java:102)
       at org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:493)
       at org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:510)
       at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:394)
       at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1285)
       at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1282)
       at java.security.AccessController.doPrivileged(Native Method)
       at javax.security.auth.Subject.doAs(Subject.java:415)
       at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1614)
       at org.apache.hadoop.mapreduce.Job.submit(Job.java:1282)
       at org.apache.kylin.job.hadoop.AbstractHadoopJob.waitForCompletion(AbstractHadoopJob.java:123)
       at org.apache.kylin.job.hadoop.cube.FactDistinctColumnsJob.run(FactDistinctColumnsJob.java:80)
       at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
       at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
       at org.apache.kylin.job.common.MapReduceExecutable.doWork(MapReduceExecutable.java:112)
       at org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
       at org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:50)
       at org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
       at org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:132)
       at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
       at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
       at java.lang.Thread.run(Thread.java:745)
[pool-7-thread-2]:[2015-06-01 16:57:52,178][DEBUG][org.apache.kylin.common.persistence.ResourceStore.putResource(ResourceStore.java:171)] - Saving resource /execute_output/81a626bd-60a7-4993-9d3e-ca3013e76fb9-01 (Store kylin_metadata@hbase)
[pool-7-thread-2]:[2015-06-01 16:57:54,016][DEBUG][org.apache.kylin.common.persistence.ResourceStore.putResource(ResourceStore.java:171)] - Saving resource /execute_output/81a626bd-60a7-4993-9d3e-ca3013e76fb9-01 (Store kylin_metadata@hbase)
[pool-7-thread-2]:[2015-06-01 16:57:54,282][INFO][org.apache.kylin.job.manager.ExecutableManager.updateJobOutput(ExecutableManager.java:222)] - job id:81a626bd-60a7-4993-9d3e-ca3013e76fb9-01 from RUNNING to ERROR
[pool-7-thread-2]:[2015-06-01 16:57:54,282][ERROR][org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:109)] - error running Executable
org.apache.kylin.job.exception.ExecuteException: java.lang.IncompatibleClassChangeError: Found interface org.apache.hadoop.mapreduce.JobContext, but class was expected
       at org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:111)
       at org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:50)
       at org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
       at org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:132)
       at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
       at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
       at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.IncompatibleClassChangeError: Found interface org.apache.hadoop.mapreduce.JobContext, but class was expected
       at org.apache.hive.hcatalog.mapreduce.HCatBaseInputFormat.getSplits(HCatBaseInputFormat.java:102)
       at org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:493)
       at org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:510)
       at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:394)
       at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1285)
       at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1282)
       at java.security.AccessController.doPrivileged(Native Method)
       at javax.security.auth.Subject.doAs(Subject.java:415)
       at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1614)
       at org.apache.hadoop.mapreduce.Job.submit(Job.java:1282)
       at org.apache.kylin.job.hadoop.AbstractHadoopJob.waitForCompletion(AbstractHadoopJob.java:123)
       at org.apache.kylin.job.hadoop.cube.FactDistinctColumnsJob.run(FactDistinctColumnsJob.java:80)
       at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
       at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
       at org.apache.kylin.job.common.MapReduceExecutable.doWork(MapReduceExecutable.java:112)
       at org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)

Re: RE: Building cube issue: IncompatibleClassChangeError error

Posted by Li Yang <li...@apache.org>.
Dev env... with the default pom, we use hdp 2.2.x as sandbox. That works.

On Thu, Jun 4, 2015 at 4:10 PM, Mohit Bhatnagar <mo...@adobe.com> wrote:

> Hi Yang,
>
> While installing I followed "Off Hadoop CLI Installation" tutorial to set
> up the development  environment. So I have Kylin server on my Windows
> machine pointing to my Dev hadoop cluster.
> I cross checked the versions which are installed on my hadoop cluster and
> made the same changes to Kylin/pom.xml. Even though it downloads the same
> version of Hadoop hive hbase jars in my local repo as mentioned below, I am
> still hitting the issue. I am not sure from where it is referencing the old
> Hadoop versions.
>
> Please note that when I build Kylin sourcecode on my local machine (which
> has kylin server running and pointing to Hadoop cluster), it downloads
> Hadoop hive hbase jars on my local repo.
>
> Thanks,
> Mohit
>
>
>
>
> -----Original Message-----
> From: Li Yang [mailto:liyang@apache.org]
> Sent: Thursday, June 04, 2015 12:16 PM
> To: dev@kylin.incubator.apache.org
> Subject: Re: RE: Building cube issue: IncompatibleClassChangeError error
>
> Aiming to be compatible with all Hadoop 2.x versions, Kylin does not ship
> any Hadoop/Hive jar. There must be multiple versions of Hadoop/Hive in your
> cluster.
>
> From the stack trace, my guess is the Hive version of the Kylin host and
> the Hive server might be different... not sure however.
>
> On Tue, Jun 2, 2015 at 2:29 PM, 250635732@qq.com <25...@qq.com> wrote:
>
> > Hi, there,
> >
> > I think the root cause is you got some old hadoop or yarn version jars
> > in your environment.
> >
> > Which step does you go when you issue cube build job? You may want to
> > check the yarn history server
> >
> > for more log information.
> >
> > Thanks,
> > Sun.
> >
> >
> > 250635732@qq.com
> >
> > From: Mohit Bhatnagar
> > Date: 2015-06-02 14:17
> > To: dev@kylin.incubator.apache.org
> > CC: Sakshi Gupta
> > Subject: RE: Building cube issue: IncompatibleClassChangeError error
> > Can anyone help me out with this?
> >
> > -----Original Message-----
> > From: Mohit Bhatnagar [mailto:mobhatna@adobe.com]
> > Sent: Monday, June 01, 2015 5:06 PM
> > To: dev@kylin.incubator.apache.org
> > Cc: Sakshi Gupta
> > Subject: Building cube issue: IncompatibleClassChangeError error
> >
> > Hi,
> >
> > I m hitting the below IncompatibleClassChangeError  error. We have the
> > following versions already installed in our Prod cluster and I have
> > also changed the kylin/pom.xml file to have the same versions.
> > Here are the details:
> >   <!-- Hadoop versions -->
> >         <hadoop2.version>2.5.0</hadoop2.version>
> >         <yarn.version>2.5.0</yarn.version>
> >         <zookeeper.version>3.4.5</zookeeper.version>
> >         <hive.version>0.13.1</hive.version>
> >         <hive-hcatalog.version>0.13.1</hive-hcatalog.version>
> >         <hbase-hadoop2.version>0.98.6-hadoop2</hbase-hadoop2.version>
> >
> > From Hadoop 2.0 onwards JobContext is an interface. Don't know why it
> > is expecting a class:
> > Here is the log:
> >
> >
> > [pool-6-thread-1]:[2015-06-01
> > 16:56:53,644][INFO][org.apache.kylin.job.impl.threadpool.DefaultSchedu
> > ler$FetcherRunner.run(DefaultScheduler.java:117)]
> > - Job Fetcher: 0 running, 1 actual running, 1 ready, 44 others
> > [pool-7-thread-2]:[2015-06-01
> > 16:57:48,749][INFO][org.apache.hadoop.mapreduce.JobSubmitter.submitJob
> > Internal(JobSubmitter.java:441)]
> > - Cleaning up the staging area
> > /user/mobhatna/.staging/job_1432706588203_13062
> > [pool-7-thread-2]:[2015-06-01
> > 16:57:49,031][ERROR][org.apache.kylin.job.execution.AbstractExecutable
> > .execute(AbstractExecutable.java:109)]
> > - error running Executable
> > java.lang.IncompatibleClassChangeError: Found interface
> > org.apache.hadoop.mapreduce.JobContext, but class was expected
> >        at
> >
> org.apache.hive.hcatalog.mapreduce.HCatBaseInputFormat.getSplits(HCatBaseInputFormat.java:102)
> >        at
> >
> org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:493)
> >        at
> >
> org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:510)
> >        at
> >
> org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:394)
> >        at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1285)
> >        at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1282)
> >        at java.security.AccessController.doPrivileged(Native Method)
> >        at javax.security.auth.Subject.doAs(Subject.java:415)
> >        at
> >
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1614)
> >        at org.apache.hadoop.mapreduce.Job.submit(Job.java:1282)
> >        at
> >
> org.apache.kylin.job.hadoop.AbstractHadoopJob.waitForCompletion(AbstractHadoopJob.java:123)
> >        at
> >
> org.apache.kylin.job.hadoop.cube.FactDistinctColumnsJob.run(FactDistinctColumnsJob.java:80)
> >        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
> >        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
> >        at
> >
> org.apache.kylin.job.common.MapReduceExecutable.doWork(MapReduceExecutable.java:112)
> >        at
> >
> org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
> >        at
> >
> org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:50)
> >        at
> >
> org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
> >        at
> >
> org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:132)
> >        at
> >
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> >        at
> >
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> >        at java.lang.Thread.run(Thread.java:745)
> > [pool-7-thread-2]:[2015-06-01
> > 16:57:52,178][DEBUG][org.apache.kylin.common.persistence.ResourceStore
> > .putResource(ResourceStore.java:171)]
> > - Saving resource
> > /execute_output/81a626bd-60a7-4993-9d3e-ca3013e76fb9-01
> > (Store kylin_metadata@hbase)
> > [pool-7-thread-2]:[2015-06-01
> > 16:57:54,016][DEBUG][org.apache.kylin.common.persistence.ResourceStore
> > .putResource(ResourceStore.java:171)]
> > - Saving resource
> > /execute_output/81a626bd-60a7-4993-9d3e-ca3013e76fb9-01
> > (Store kylin_metadata@hbase)
> > [pool-7-thread-2]:[2015-06-01
> > 16:57:54,282][INFO][org.apache.kylin.job.manager.ExecutableManager.upd
> > ateJobOutput(ExecutableManager.java:222)]
> > - job id:81a626bd-60a7-4993-9d3e-ca3013e76fb9-01 from RUNNING to ERROR
> > [pool-7-thread-2]:[2015-06-01
> > 16:57:54,282][ERROR][org.apache.kylin.job.execution.AbstractExecutable
> > .execute(AbstractExecutable.java:109)]
> > - error running Executable
> > org.apache.kylin.job.exception.ExecuteException:
> > java.lang.IncompatibleClassChangeError: Found interface
> > org.apache.hadoop.mapreduce.JobContext, but class was expected
> >        at
> >
> org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:111)
> >        at
> >
> org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:50)
> >        at
> >
> org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
> >        at
> >
> org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:132)
> >        at
> >
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> >        at
> >
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> >        at java.lang.Thread.run(Thread.java:745)
> > Caused by: java.lang.IncompatibleClassChangeError: Found interface
> > org.apache.hadoop.mapreduce.JobContext, but class was expected
> >        at
> >
> org.apache.hive.hcatalog.mapreduce.HCatBaseInputFormat.getSplits(HCatBaseInputFormat.java:102)
> >        at
> >
> org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:493)
> >        at
> >
> org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:510)
> >        at
> >
> org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:394)
> >        at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1285)
> >        at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1282)
> >        at java.security.AccessController.doPrivileged(Native Method)
> >        at javax.security.auth.Subject.doAs(Subject.java:415)
> >        at
> >
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1614)
> >        at org.apache.hadoop.mapreduce.Job.submit(Job.java:1282)
> >        at
> >
> org.apache.kylin.job.hadoop.AbstractHadoopJob.waitForCompletion(AbstractHadoopJob.java:123)
> >        at
> >
> org.apache.kylin.job.hadoop.cube.FactDistinctColumnsJob.run(FactDistinctColumnsJob.java:80)
> >        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
> >        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
> >        at
> >
> org.apache.kylin.job.common.MapReduceExecutable.doWork(MapReduceExecutable.java:112)
> >        at
> > org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExec
> > utable.java:107)
> >
>

RE: RE: Building cube issue: IncompatibleClassChangeError error

Posted by Mohit Bhatnagar <mo...@adobe.com>.
Hi Yang,

While installing I followed "Off Hadoop CLI Installation" tutorial to set up the development  environment. So I have Kylin server on my Windows machine pointing to my Dev hadoop cluster. 
I cross checked the versions which are installed on my hadoop cluster and made the same changes to Kylin/pom.xml. Even though it downloads the same version of Hadoop hive hbase jars in my local repo as mentioned below, I am still hitting the issue. I am not sure from where it is referencing the old Hadoop versions.

Please note that when I build Kylin sourcecode on my local machine (which has kylin server running and pointing to Hadoop cluster), it downloads Hadoop hive hbase jars on my local repo.

Thanks,
Mohit 




-----Original Message-----
From: Li Yang [mailto:liyang@apache.org] 
Sent: Thursday, June 04, 2015 12:16 PM
To: dev@kylin.incubator.apache.org
Subject: Re: RE: Building cube issue: IncompatibleClassChangeError error

Aiming to be compatible with all Hadoop 2.x versions, Kylin does not ship any Hadoop/Hive jar. There must be multiple versions of Hadoop/Hive in your cluster.

From the stack trace, my guess is the Hive version of the Kylin host and the Hive server might be different... not sure however.

On Tue, Jun 2, 2015 at 2:29 PM, 250635732@qq.com <25...@qq.com> wrote:

> Hi, there,
>
> I think the root cause is you got some old hadoop or yarn version jars 
> in your environment.
>
> Which step does you go when you issue cube build job? You may want to 
> check the yarn history server
>
> for more log information.
>
> Thanks,
> Sun.
>
>
> 250635732@qq.com
>
> From: Mohit Bhatnagar
> Date: 2015-06-02 14:17
> To: dev@kylin.incubator.apache.org
> CC: Sakshi Gupta
> Subject: RE: Building cube issue: IncompatibleClassChangeError error 
> Can anyone help me out with this?
>
> -----Original Message-----
> From: Mohit Bhatnagar [mailto:mobhatna@adobe.com]
> Sent: Monday, June 01, 2015 5:06 PM
> To: dev@kylin.incubator.apache.org
> Cc: Sakshi Gupta
> Subject: Building cube issue: IncompatibleClassChangeError error
>
> Hi,
>
> I m hitting the below IncompatibleClassChangeError  error. We have the 
> following versions already installed in our Prod cluster and I have 
> also changed the kylin/pom.xml file to have the same versions.
> Here are the details:
>   <!-- Hadoop versions -->
>         <hadoop2.version>2.5.0</hadoop2.version>
>         <yarn.version>2.5.0</yarn.version>
>         <zookeeper.version>3.4.5</zookeeper.version>
>         <hive.version>0.13.1</hive.version>
>         <hive-hcatalog.version>0.13.1</hive-hcatalog.version>
>         <hbase-hadoop2.version>0.98.6-hadoop2</hbase-hadoop2.version>
>
> From Hadoop 2.0 onwards JobContext is an interface. Don't know why it 
> is expecting a class:
> Here is the log:
>
>
> [pool-6-thread-1]:[2015-06-01
> 16:56:53,644][INFO][org.apache.kylin.job.impl.threadpool.DefaultSchedu
> ler$FetcherRunner.run(DefaultScheduler.java:117)]
> - Job Fetcher: 0 running, 1 actual running, 1 ready, 44 others
> [pool-7-thread-2]:[2015-06-01
> 16:57:48,749][INFO][org.apache.hadoop.mapreduce.JobSubmitter.submitJob
> Internal(JobSubmitter.java:441)]
> - Cleaning up the staging area
> /user/mobhatna/.staging/job_1432706588203_13062
> [pool-7-thread-2]:[2015-06-01
> 16:57:49,031][ERROR][org.apache.kylin.job.execution.AbstractExecutable
> .execute(AbstractExecutable.java:109)]
> - error running Executable
> java.lang.IncompatibleClassChangeError: Found interface 
> org.apache.hadoop.mapreduce.JobContext, but class was expected
>        at
> org.apache.hive.hcatalog.mapreduce.HCatBaseInputFormat.getSplits(HCatBaseInputFormat.java:102)
>        at
> org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:493)
>        at
> org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:510)
>        at
> org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:394)
>        at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1285)
>        at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1282)
>        at java.security.AccessController.doPrivileged(Native Method)
>        at javax.security.auth.Subject.doAs(Subject.java:415)
>        at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1614)
>        at org.apache.hadoop.mapreduce.Job.submit(Job.java:1282)
>        at
> org.apache.kylin.job.hadoop.AbstractHadoopJob.waitForCompletion(AbstractHadoopJob.java:123)
>        at
> org.apache.kylin.job.hadoop.cube.FactDistinctColumnsJob.run(FactDistinctColumnsJob.java:80)
>        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
>        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
>        at
> org.apache.kylin.job.common.MapReduceExecutable.doWork(MapReduceExecutable.java:112)
>        at
> org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
>        at
> org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:50)
>        at
> org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
>        at
> org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:132)
>        at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>        at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>        at java.lang.Thread.run(Thread.java:745)
> [pool-7-thread-2]:[2015-06-01
> 16:57:52,178][DEBUG][org.apache.kylin.common.persistence.ResourceStore
> .putResource(ResourceStore.java:171)]
> - Saving resource 
> /execute_output/81a626bd-60a7-4993-9d3e-ca3013e76fb9-01
> (Store kylin_metadata@hbase)
> [pool-7-thread-2]:[2015-06-01
> 16:57:54,016][DEBUG][org.apache.kylin.common.persistence.ResourceStore
> .putResource(ResourceStore.java:171)]
> - Saving resource 
> /execute_output/81a626bd-60a7-4993-9d3e-ca3013e76fb9-01
> (Store kylin_metadata@hbase)
> [pool-7-thread-2]:[2015-06-01
> 16:57:54,282][INFO][org.apache.kylin.job.manager.ExecutableManager.upd
> ateJobOutput(ExecutableManager.java:222)]
> - job id:81a626bd-60a7-4993-9d3e-ca3013e76fb9-01 from RUNNING to ERROR
> [pool-7-thread-2]:[2015-06-01
> 16:57:54,282][ERROR][org.apache.kylin.job.execution.AbstractExecutable
> .execute(AbstractExecutable.java:109)]
> - error running Executable
> org.apache.kylin.job.exception.ExecuteException:
> java.lang.IncompatibleClassChangeError: Found interface 
> org.apache.hadoop.mapreduce.JobContext, but class was expected
>        at
> org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:111)
>        at
> org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:50)
>        at
> org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
>        at
> org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:132)
>        at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>        at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>        at java.lang.Thread.run(Thread.java:745)
> Caused by: java.lang.IncompatibleClassChangeError: Found interface 
> org.apache.hadoop.mapreduce.JobContext, but class was expected
>        at
> org.apache.hive.hcatalog.mapreduce.HCatBaseInputFormat.getSplits(HCatBaseInputFormat.java:102)
>        at
> org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:493)
>        at
> org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:510)
>        at
> org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:394)
>        at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1285)
>        at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1282)
>        at java.security.AccessController.doPrivileged(Native Method)
>        at javax.security.auth.Subject.doAs(Subject.java:415)
>        at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1614)
>        at org.apache.hadoop.mapreduce.Job.submit(Job.java:1282)
>        at
> org.apache.kylin.job.hadoop.AbstractHadoopJob.waitForCompletion(AbstractHadoopJob.java:123)
>        at
> org.apache.kylin.job.hadoop.cube.FactDistinctColumnsJob.run(FactDistinctColumnsJob.java:80)
>        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
>        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
>        at
> org.apache.kylin.job.common.MapReduceExecutable.doWork(MapReduceExecutable.java:112)
>        at
> org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExec
> utable.java:107)
>

Re: RE: Building cube issue: IncompatibleClassChangeError error

Posted by Li Yang <li...@apache.org>.
Aiming to be compatible with all Hadoop 2.x versions, Kylin does not ship
any Hadoop/Hive jar. There must be multiple versions of Hadoop/Hive in your
cluster.

>From the stack trace, my guess is the Hive version of the Kylin host and
the Hive server might be different... not sure however.

On Tue, Jun 2, 2015 at 2:29 PM, 250635732@qq.com <25...@qq.com> wrote:

> Hi, there,
>
> I think the root cause is you got some old hadoop or yarn version jars in
> your environment.
>
> Which step does you go when you issue cube build job? You may want to
> check the yarn history server
>
> for more log information.
>
> Thanks,
> Sun.
>
>
> 250635732@qq.com
>
> From: Mohit Bhatnagar
> Date: 2015-06-02 14:17
> To: dev@kylin.incubator.apache.org
> CC: Sakshi Gupta
> Subject: RE: Building cube issue: IncompatibleClassChangeError error
> Can anyone help me out with this?
>
> -----Original Message-----
> From: Mohit Bhatnagar [mailto:mobhatna@adobe.com]
> Sent: Monday, June 01, 2015 5:06 PM
> To: dev@kylin.incubator.apache.org
> Cc: Sakshi Gupta
> Subject: Building cube issue: IncompatibleClassChangeError error
>
> Hi,
>
> I m hitting the below IncompatibleClassChangeError  error. We have the
> following versions already installed in our Prod cluster and I have also
> changed the kylin/pom.xml file to have the same versions.
> Here are the details:
>   <!-- Hadoop versions -->
>         <hadoop2.version>2.5.0</hadoop2.version>
>         <yarn.version>2.5.0</yarn.version>
>         <zookeeper.version>3.4.5</zookeeper.version>
>         <hive.version>0.13.1</hive.version>
>         <hive-hcatalog.version>0.13.1</hive-hcatalog.version>
>         <hbase-hadoop2.version>0.98.6-hadoop2</hbase-hadoop2.version>
>
> From Hadoop 2.0 onwards JobContext is an interface. Don't know why it is
> expecting a class:
> Here is the log:
>
>
> [pool-6-thread-1]:[2015-06-01
> 16:56:53,644][INFO][org.apache.kylin.job.impl.threadpool.DefaultScheduler$FetcherRunner.run(DefaultScheduler.java:117)]
> - Job Fetcher: 0 running, 1 actual running, 1 ready, 44 others
> [pool-7-thread-2]:[2015-06-01
> 16:57:48,749][INFO][org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:441)]
> - Cleaning up the staging area
> /user/mobhatna/.staging/job_1432706588203_13062
> [pool-7-thread-2]:[2015-06-01
> 16:57:49,031][ERROR][org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:109)]
> - error running Executable
> java.lang.IncompatibleClassChangeError: Found interface
> org.apache.hadoop.mapreduce.JobContext, but class was expected
>        at
> org.apache.hive.hcatalog.mapreduce.HCatBaseInputFormat.getSplits(HCatBaseInputFormat.java:102)
>        at
> org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:493)
>        at
> org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:510)
>        at
> org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:394)
>        at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1285)
>        at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1282)
>        at java.security.AccessController.doPrivileged(Native Method)
>        at javax.security.auth.Subject.doAs(Subject.java:415)
>        at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1614)
>        at org.apache.hadoop.mapreduce.Job.submit(Job.java:1282)
>        at
> org.apache.kylin.job.hadoop.AbstractHadoopJob.waitForCompletion(AbstractHadoopJob.java:123)
>        at
> org.apache.kylin.job.hadoop.cube.FactDistinctColumnsJob.run(FactDistinctColumnsJob.java:80)
>        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
>        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
>        at
> org.apache.kylin.job.common.MapReduceExecutable.doWork(MapReduceExecutable.java:112)
>        at
> org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
>        at
> org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:50)
>        at
> org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
>        at
> org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:132)
>        at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>        at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>        at java.lang.Thread.run(Thread.java:745)
> [pool-7-thread-2]:[2015-06-01
> 16:57:52,178][DEBUG][org.apache.kylin.common.persistence.ResourceStore.putResource(ResourceStore.java:171)]
> - Saving resource /execute_output/81a626bd-60a7-4993-9d3e-ca3013e76fb9-01
> (Store kylin_metadata@hbase)
> [pool-7-thread-2]:[2015-06-01
> 16:57:54,016][DEBUG][org.apache.kylin.common.persistence.ResourceStore.putResource(ResourceStore.java:171)]
> - Saving resource /execute_output/81a626bd-60a7-4993-9d3e-ca3013e76fb9-01
> (Store kylin_metadata@hbase)
> [pool-7-thread-2]:[2015-06-01
> 16:57:54,282][INFO][org.apache.kylin.job.manager.ExecutableManager.updateJobOutput(ExecutableManager.java:222)]
> - job id:81a626bd-60a7-4993-9d3e-ca3013e76fb9-01 from RUNNING to ERROR
> [pool-7-thread-2]:[2015-06-01
> 16:57:54,282][ERROR][org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:109)]
> - error running Executable
> org.apache.kylin.job.exception.ExecuteException:
> java.lang.IncompatibleClassChangeError: Found interface
> org.apache.hadoop.mapreduce.JobContext, but class was expected
>        at
> org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:111)
>        at
> org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:50)
>        at
> org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
>        at
> org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:132)
>        at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>        at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>        at java.lang.Thread.run(Thread.java:745)
> Caused by: java.lang.IncompatibleClassChangeError: Found interface
> org.apache.hadoop.mapreduce.JobContext, but class was expected
>        at
> org.apache.hive.hcatalog.mapreduce.HCatBaseInputFormat.getSplits(HCatBaseInputFormat.java:102)
>        at
> org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:493)
>        at
> org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:510)
>        at
> org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:394)
>        at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1285)
>        at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1282)
>        at java.security.AccessController.doPrivileged(Native Method)
>        at javax.security.auth.Subject.doAs(Subject.java:415)
>        at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1614)
>        at org.apache.hadoop.mapreduce.Job.submit(Job.java:1282)
>        at
> org.apache.kylin.job.hadoop.AbstractHadoopJob.waitForCompletion(AbstractHadoopJob.java:123)
>        at
> org.apache.kylin.job.hadoop.cube.FactDistinctColumnsJob.run(FactDistinctColumnsJob.java:80)
>        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
>        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
>        at
> org.apache.kylin.job.common.MapReduceExecutable.doWork(MapReduceExecutable.java:112)
>        at
> org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
>

Re: RE: Building cube issue: IncompatibleClassChangeError error

Posted by "250635732@qq.com" <25...@qq.com>.
Hi, there, 

I think the root cause is you got some old hadoop or yarn version jars in your environment. 

Which step does you go when you issue cube build job? You may want to check the yarn history server

for more log information. 

Thanks,
Sun.


250635732@qq.com
 
From: Mohit Bhatnagar
Date: 2015-06-02 14:17
To: dev@kylin.incubator.apache.org
CC: Sakshi Gupta
Subject: RE: Building cube issue: IncompatibleClassChangeError error
Can anyone help me out with this?  
 
-----Original Message-----
From: Mohit Bhatnagar [mailto:mobhatna@adobe.com] 
Sent: Monday, June 01, 2015 5:06 PM
To: dev@kylin.incubator.apache.org
Cc: Sakshi Gupta
Subject: Building cube issue: IncompatibleClassChangeError error
 
Hi,
 
I m hitting the below IncompatibleClassChangeError  error. We have the following versions already installed in our Prod cluster and I have also changed the kylin/pom.xml file to have the same versions.
Here are the details:
  <!-- Hadoop versions -->
        <hadoop2.version>2.5.0</hadoop2.version>
        <yarn.version>2.5.0</yarn.version>
        <zookeeper.version>3.4.5</zookeeper.version>
        <hive.version>0.13.1</hive.version>
        <hive-hcatalog.version>0.13.1</hive-hcatalog.version>
        <hbase-hadoop2.version>0.98.6-hadoop2</hbase-hadoop2.version>
 
From Hadoop 2.0 onwards JobContext is an interface. Don't know why it is expecting a class:
Here is the log:
 
 
[pool-6-thread-1]:[2015-06-01 16:56:53,644][INFO][org.apache.kylin.job.impl.threadpool.DefaultScheduler$FetcherRunner.run(DefaultScheduler.java:117)] - Job Fetcher: 0 running, 1 actual running, 1 ready, 44 others
[pool-7-thread-2]:[2015-06-01 16:57:48,749][INFO][org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:441)] - Cleaning up the staging area /user/mobhatna/.staging/job_1432706588203_13062
[pool-7-thread-2]:[2015-06-01 16:57:49,031][ERROR][org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:109)] - error running Executable
java.lang.IncompatibleClassChangeError: Found interface org.apache.hadoop.mapreduce.JobContext, but class was expected
       at org.apache.hive.hcatalog.mapreduce.HCatBaseInputFormat.getSplits(HCatBaseInputFormat.java:102)
       at org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:493)
       at org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:510)
       at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:394)
       at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1285)
       at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1282)
       at java.security.AccessController.doPrivileged(Native Method)
       at javax.security.auth.Subject.doAs(Subject.java:415)
       at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1614)
       at org.apache.hadoop.mapreduce.Job.submit(Job.java:1282)
       at org.apache.kylin.job.hadoop.AbstractHadoopJob.waitForCompletion(AbstractHadoopJob.java:123)
       at org.apache.kylin.job.hadoop.cube.FactDistinctColumnsJob.run(FactDistinctColumnsJob.java:80)
       at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
       at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
       at org.apache.kylin.job.common.MapReduceExecutable.doWork(MapReduceExecutable.java:112)
       at org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
       at org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:50)
       at org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
       at org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:132)
       at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
       at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
       at java.lang.Thread.run(Thread.java:745)
[pool-7-thread-2]:[2015-06-01 16:57:52,178][DEBUG][org.apache.kylin.common.persistence.ResourceStore.putResource(ResourceStore.java:171)] - Saving resource /execute_output/81a626bd-60a7-4993-9d3e-ca3013e76fb9-01 (Store kylin_metadata@hbase)
[pool-7-thread-2]:[2015-06-01 16:57:54,016][DEBUG][org.apache.kylin.common.persistence.ResourceStore.putResource(ResourceStore.java:171)] - Saving resource /execute_output/81a626bd-60a7-4993-9d3e-ca3013e76fb9-01 (Store kylin_metadata@hbase)
[pool-7-thread-2]:[2015-06-01 16:57:54,282][INFO][org.apache.kylin.job.manager.ExecutableManager.updateJobOutput(ExecutableManager.java:222)] - job id:81a626bd-60a7-4993-9d3e-ca3013e76fb9-01 from RUNNING to ERROR
[pool-7-thread-2]:[2015-06-01 16:57:54,282][ERROR][org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:109)] - error running Executable
org.apache.kylin.job.exception.ExecuteException: java.lang.IncompatibleClassChangeError: Found interface org.apache.hadoop.mapreduce.JobContext, but class was expected
       at org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:111)
       at org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:50)
       at org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
       at org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:132)
       at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
       at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
       at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.IncompatibleClassChangeError: Found interface org.apache.hadoop.mapreduce.JobContext, but class was expected
       at org.apache.hive.hcatalog.mapreduce.HCatBaseInputFormat.getSplits(HCatBaseInputFormat.java:102)
       at org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:493)
       at org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:510)
       at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:394)
       at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1285)
       at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1282)
       at java.security.AccessController.doPrivileged(Native Method)
       at javax.security.auth.Subject.doAs(Subject.java:415)
       at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1614)
       at org.apache.hadoop.mapreduce.Job.submit(Job.java:1282)
       at org.apache.kylin.job.hadoop.AbstractHadoopJob.waitForCompletion(AbstractHadoopJob.java:123)
       at org.apache.kylin.job.hadoop.cube.FactDistinctColumnsJob.run(FactDistinctColumnsJob.java:80)
       at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
       at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
       at org.apache.kylin.job.common.MapReduceExecutable.doWork(MapReduceExecutable.java:112)
       at org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)

RE: Building cube issue: IncompatibleClassChangeError error

Posted by Mohit Bhatnagar <mo...@adobe.com>.
Can anyone help me out with this?  

-----Original Message-----
From: Mohit Bhatnagar [mailto:mobhatna@adobe.com] 
Sent: Monday, June 01, 2015 5:06 PM
To: dev@kylin.incubator.apache.org
Cc: Sakshi Gupta
Subject: Building cube issue: IncompatibleClassChangeError error

Hi,

I m hitting the below IncompatibleClassChangeError  error. We have the following versions already installed in our Prod cluster and I have also changed the kylin/pom.xml file to have the same versions.
Here are the details:
  <!-- Hadoop versions -->
        <hadoop2.version>2.5.0</hadoop2.version>
        <yarn.version>2.5.0</yarn.version>
        <zookeeper.version>3.4.5</zookeeper.version>
        <hive.version>0.13.1</hive.version>
        <hive-hcatalog.version>0.13.1</hive-hcatalog.version>
        <hbase-hadoop2.version>0.98.6-hadoop2</hbase-hadoop2.version>

>From Hadoop 2.0 onwards JobContext is an interface. Don't know why it is expecting a class:
Here is the log:


[pool-6-thread-1]:[2015-06-01 16:56:53,644][INFO][org.apache.kylin.job.impl.threadpool.DefaultScheduler$FetcherRunner.run(DefaultScheduler.java:117)] - Job Fetcher: 0 running, 1 actual running, 1 ready, 44 others
[pool-7-thread-2]:[2015-06-01 16:57:48,749][INFO][org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:441)] - Cleaning up the staging area /user/mobhatna/.staging/job_1432706588203_13062
[pool-7-thread-2]:[2015-06-01 16:57:49,031][ERROR][org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:109)] - error running Executable
java.lang.IncompatibleClassChangeError: Found interface org.apache.hadoop.mapreduce.JobContext, but class was expected
       at org.apache.hive.hcatalog.mapreduce.HCatBaseInputFormat.getSplits(HCatBaseInputFormat.java:102)
       at org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:493)
       at org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:510)
       at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:394)
       at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1285)
       at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1282)
       at java.security.AccessController.doPrivileged(Native Method)
       at javax.security.auth.Subject.doAs(Subject.java:415)
       at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1614)
       at org.apache.hadoop.mapreduce.Job.submit(Job.java:1282)
       at org.apache.kylin.job.hadoop.AbstractHadoopJob.waitForCompletion(AbstractHadoopJob.java:123)
       at org.apache.kylin.job.hadoop.cube.FactDistinctColumnsJob.run(FactDistinctColumnsJob.java:80)
       at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
       at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
       at org.apache.kylin.job.common.MapReduceExecutable.doWork(MapReduceExecutable.java:112)
       at org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
       at org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:50)
       at org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
       at org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:132)
       at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
       at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
       at java.lang.Thread.run(Thread.java:745)
[pool-7-thread-2]:[2015-06-01 16:57:52,178][DEBUG][org.apache.kylin.common.persistence.ResourceStore.putResource(ResourceStore.java:171)] - Saving resource /execute_output/81a626bd-60a7-4993-9d3e-ca3013e76fb9-01 (Store kylin_metadata@hbase)
[pool-7-thread-2]:[2015-06-01 16:57:54,016][DEBUG][org.apache.kylin.common.persistence.ResourceStore.putResource(ResourceStore.java:171)] - Saving resource /execute_output/81a626bd-60a7-4993-9d3e-ca3013e76fb9-01 (Store kylin_metadata@hbase)
[pool-7-thread-2]:[2015-06-01 16:57:54,282][INFO][org.apache.kylin.job.manager.ExecutableManager.updateJobOutput(ExecutableManager.java:222)] - job id:81a626bd-60a7-4993-9d3e-ca3013e76fb9-01 from RUNNING to ERROR
[pool-7-thread-2]:[2015-06-01 16:57:54,282][ERROR][org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:109)] - error running Executable
org.apache.kylin.job.exception.ExecuteException: java.lang.IncompatibleClassChangeError: Found interface org.apache.hadoop.mapreduce.JobContext, but class was expected
       at org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:111)
       at org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:50)
       at org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
       at org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:132)
       at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
       at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
       at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.IncompatibleClassChangeError: Found interface org.apache.hadoop.mapreduce.JobContext, but class was expected
       at org.apache.hive.hcatalog.mapreduce.HCatBaseInputFormat.getSplits(HCatBaseInputFormat.java:102)
       at org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:493)
       at org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:510)
       at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:394)
       at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1285)
       at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1282)
       at java.security.AccessController.doPrivileged(Native Method)
       at javax.security.auth.Subject.doAs(Subject.java:415)
       at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1614)
       at org.apache.hadoop.mapreduce.Job.submit(Job.java:1282)
       at org.apache.kylin.job.hadoop.AbstractHadoopJob.waitForCompletion(AbstractHadoopJob.java:123)
       at org.apache.kylin.job.hadoop.cube.FactDistinctColumnsJob.run(FactDistinctColumnsJob.java:80)
       at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
       at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
       at org.apache.kylin.job.common.MapReduceExecutable.doWork(MapReduceExecutable.java:112)
       at org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)