You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@pig.apache.org by Claudio Romo Otto <cl...@namictech.com> on 2013/10/24 14:37:58 UTC

Support for Hadoop 2.2

Does Pig support Hadoop 2.2? When I try Pig 0.12 and Hadoop 2.2 I get an 
error even with simple operations like

data = LOAD 'cql://keyspace1/testcf?' USING CqlStorage();
dump data;

I only got a warning first and then and exception:

2013-10-24 09:35:19,300 [main] WARN 
org.apache.pig.backend.hadoop20.PigJobControl - falling back to default 
JobControl (not using hadoop 0.20 ?)
java.lang.NoSuchFieldException: runnerState
     at java.lang.Class.getDeclaredField(Class.java:1938)
     at 
org.apache.pig.backend.hadoop20.PigJobControl.<clinit>(PigJobControl.java:51)
     at 
org.apache.pig.backend.hadoop.executionengine.shims.HadoopShims.newJobControl(HadoopShims.java:97)
     at 
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler.compile(JobControlCompiler.java:285)
     at 
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher.launchPig(MapReduceLauncher.java:177)
     at org.apache.pig.PigServer.launchPlan(PigServer.java:1264)
     at 
org.apache.pig.PigServer.executeCompiledLogicalPlan(PigServer.java:1249)
     at org.apache.pig.PigServer.storeEx(PigServer.java:931)
     at org.apache.pig.PigServer.store(PigServer.java:898)
     at org.apache.pig.PigServer.openIterator(PigServer.java:811)
     at 
org.apache.pig.tools.grunt.GruntParser.processDump(GruntParser.java:696)
     at 
org.apache.pig.tools.pigscript.parser.PigScriptParser.parse(PigScriptParser.java:320)
     at 
org.apache.pig.tools.grunt.GruntParser.parseStopOnError(GruntParser.java:194)
     at 
org.apache.pig.tools.grunt.GruntParser.parseStopOnError(GruntParser.java:170)
     at org.apache.pig.tools.grunt.Grunt.run(Grunt.java:69)
     at org.apache.pig.Main.run(Main.java:538)
     at org.apache.pig.Main.main(Main.java:157)
     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
     at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
     at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
     at java.lang.reflect.Method.invoke(Method.java:606)
     at org.apache.hadoop.util.RunJar.main(RunJar.java:212)

--------------------------------------

Backend error message during job submission
-------------------------------------------
Unexpected System Error Occured: java.lang.IncompatibleClassChangeError: 
Found interface org.apache.hadoop.mapreduce.JobContext, but class was 
expected
         at 
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigOutputFormat.setupUdfEnvAndStores(PigOutputFormat.java:225)
         at 
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigOutputFormat.checkOutputSpecs(PigOutputFormat.java:186)
         at 
org.apache.hadoop.mapreduce.JobSubmitter.checkSpecs(JobSubmitter.java:456)
         at 
org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:342)
         at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1268)
         at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1265)
         at java.security.AccessController.doPrivileged(Native Method)
         at javax.security.auth.Subject.doAs(Subject.java:415)
         at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1491)
         at org.apache.hadoop.mapreduce.Job.submit(Job.java:1265)
         at 
org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob.submit(ControlledJob.java:335)
         at 
org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl.run(JobControl.java:240)
         at 
org.apache.pig.backend.hadoop20.PigJobControl.run(PigJobControl.java:121)
         at java.lang.Thread.run(Thread.java:724)
         at 
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher$1.run(MapReduceLauncher.java:257)

Pig Stack Trace
---------------
ERROR 1066: Unable to open iterator for alias data

org.apache.pig.impl.logicalLayer.FrontendException: ERROR 1066: Unable 
to open iterator for alias data
         at org.apache.pig.PigServer.openIterator(PigServer.java:836)
         at 
org.apache.pig.tools.grunt.GruntParser.processDump(GruntParser.java:696)
         at 
org.apache.pig.tools.pigscript.parser.PigScriptParser.parse(PigScriptParser.java:320)
         at 
org.apache.pig.tools.grunt.GruntParser.parseStopOnError(GruntParser.java:194)
         at 
org.apache.pig.tools.grunt.GruntParser.parseStopOnError(GruntParser.java:170)
         at org.apache.pig.tools.grunt.Grunt.run(Grunt.java:69)
         at org.apache.pig.Main.run(Main.java:538)
         at org.apache.pig.Main.main(Main.java:157)
         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
         at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
         at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
         at java.lang.reflect.Method.invoke(Method.java:606)
         at org.apache.hadoop.util.RunJar.main(RunJar.java:212)
Caused by: java.io.IOException: Job terminated with anomalous status FAILED
         at org.apache.pig.PigServer.openIterator(PigServer.java:828)
         ... 12 more



Re: Support for Hadoop 2.2

Posted by Uttam Kumar <uk...@gmail.com>.
Hi Folks,

I am fairly new to Hadoop MR2 environment.  I have Hadoop0.23.1 (single
node) on Linux box and can successfully execute YARN test job, but when I
execute PIG (Ver. 12) to access and process files at HDFS, generate error
msg as follows.  I recompiled PIG with Hadoop23, but still no luck.

ant clean jar-withouthadoop -Dhadoopversion=23

Can someone please help and advise what I am missing here !!

Note: PIG in local mode works without any issue.

================

*grunt> A = load 'NYSE_dividends' as (exch, symb, dt, div);*
2013-11-26 13:20:23,151 [main] WARN  org.apache.hadoop.conf.
Configuration - fs.default.name is deprecated. Instead, use fs.defaultFS
2013-11-26 13:20:23,151 [main] WARN  org.apache.hadoop.conf.Configuration -
mapred.job.tracker is deprecated. Instead, use mapreduce.jobtracker.address
2013-11-26 13:20:23,323 [main] WARN  org.apache.hadoop.conf.Configuration -
fs.default.name is deprecated. Instead, use fs.defaultFS
2013-11-26 13:20:23,323 [main] WARN  org.apache.hadoop.conf.Configuration -
mapred.job.tracker is deprecated. Instead, use mapreduce.jobtracker.address
*grunt> dump A ;*
2013-11-26 13:20:31,805 [main] WARN  org.apache.hadoop.conf.Configuration -
fs.default.name is deprecated. Instead, use fs.defaultFS
2013-11-26 13:20:31,806 [main] WARN  org.apache.hadoop.conf.Configuration -
mapred.job.tracker is deprecated. Instead, use mapreduce.jobtracker.address
2013-11-26 13:20:31,841 [main] INFO
org.apache.pig.tools.pigstats.ScriptState - Pig features used in the
script: UNKNOWN
2013-11-26 13:20:31,913 [main] INFO
org.apache.pig.newplan.logical.optimizer.LogicalPlanOptimizer -
{RULES_ENABLED=[AddForEach, ColumnMapKeyPrune,
DuplicateForEachColumnRewrite, GroupByConstParallelSetter,
ImplicitSplitInserter, LimitOptimizer, LoadTypeCastInserter, MergeFilter,
MergeForEach, NewPartitionFilterOptimizer, PartitionFilterOptimizer,
PushDownForEachFlatten, PushUpFilter, SplitFilter, StreamTypeCastInserter],
RULES_DISABLED=[FilterLogicExpressionSimplifier]}
2013-11-26 13:20:32,159 [main] WARN  org.apache.hadoop.conf.Configuration -
fs.default.name is deprecated. Instead, use fs.defaultFS
2013-11-26 13:20:32,160 [main] WARN  org.apache.hadoop.conf.Configuration -
mapred.job.tracker is deprecated. Instead, use mapreduce.jobtracker.address
2013-11-26 13:20:32,266 [main] INFO
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MRCompiler -
File concatenation threshold: 100 optimistic? false
2013-11-26 13:20:32,304 [main] WARN  org.apache.hadoop.conf.Configuration -
fs.default.name is deprecated. Instead, use fs.defaultFS
2013-11-26 13:20:32,304 [main] WARN  org.apache.hadoop.conf.Configuration -
mapred.job.tracker is deprecated. Instead, use mapreduce.jobtracker.address
2013-11-26 13:20:32,318 [main] INFO
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MultiQueryOptimizer
- MR plan size before optimization: 1
2013-11-26 13:20:32,319 [main] INFO
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MultiQueryOptimizer
- MR plan size after optimization: 1
2013-11-26 13:20:32,322 [main] WARN  org.apache.hadoop.conf.Configuration -
fs.default.name is deprecated. Instead, use fs.defaultFS
2013-11-26 13:20:32,322 [main] WARN  org.apache.hadoop.conf.Configuration -
mapred.job.tracker is deprecated. Instead, use mapreduce.jobtracker.address
2013-11-26 13:20:32,498 [main] INFO
org.apache.pig.tools.pigstats.ScriptState - Pig script settings are added
to the job
2013-11-26 13:20:32,504 [main] WARN  org.apache.hadoop.conf.Configuration -
mapred.job.reduce.markreset.buffer.percent is deprecated. Instead, use
mapreduce.reduce.markreset.buffer.percent
2013-11-26 13:20:32,504 [main] INFO
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler
- mapred.job.reduce.markreset.buffer.percent is not set, set to default 0.3
2013-11-26 13:20:32,504 [main] WARN  org.apache.hadoop.conf.Configuration -
mapred.job.reduce.markreset.buffer.percent is deprecated. Instead, use
mapreduce.reduce.markreset.buffer.percent
2013-11-26 13:20:32,504 [main] WARN  org.apache.hadoop.conf.Configuration -
mapred.output.compress is deprecated. Instead, use
mapreduce.output.fileoutputformat.compress
2013-11-26 13:20:32,506 [main] INFO
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler
- creating jar file Job1088179330840878220.jar
2013-11-26 13:20:35,473 [main] INFO
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler
- jar file Job1088179330840878220.jar created
2013-11-26 13:20:35,473 [main] WARN  org.apache.hadoop.conf.Configuration -
mapred.jar is deprecated. Instead, use mapreduce.job.jar
2013-11-26 13:20:35,488 [main] INFO
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler
- Setting up single store job
2013-11-26 13:20:35,492 [main] INFO
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler
- Map only job, skipping reducer estimation
2013-11-26 13:20:35,513 [main] INFO
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
- 1 map-reduce job(s) waiting for submission.
2013-11-26 13:20:35,514 [main] WARN  org.apache.hadoop.conf.Configuration -
mapred.job.tracker.http.address is deprecated. Instead, use
mapreduce.jobtracker.http.address
2013-11-26 13:20:35,954 [JobControl] INFO
org.apache.hadoop.mapreduce.lib.input.FileInputFormat - Total input paths
to process : 1
2013-11-26 13:20:35,954 [JobControl] INFO
org.apache.pig.backend.hadoop.executionengine.util.MapRedUtil - Total input
paths to process : 1
2013-11-26 13:20:35,971 [JobControl] WARN
org.apache.hadoop.util.NativeCodeLoader - Unable to load native-hadoop
library for your platform... using builtin-java classes where applicable
2013-11-26 13:20:35,971 [JobControl] WARN
org.apache.hadoop.io.compress.snappy.LoadSnappy - Snappy native library not
loaded
2013-11-26 13:20:35,973 [JobControl] INFO
org.apache.hadoop.mapreduce.JobSubmitter - Cleaning up the staging area
null/hadoop/.staging/job_1385489868206_0001

*2013-11-26 13:20:35,975 [JobControl] ERROR
org.apache.pig.backend.hadoop23.PigJobControl - Error while trying to run
jobs.java.lang.RuntimeException:
java.lang.reflect.InvocationTargetException*
        at
org.apache.pig.backend.hadoop23.PigJobControl.submit(PigJobControl.java:130)
        at
org.apache.pig.backend.hadoop23.PigJobControl.run(PigJobControl.java:191)
        at java.lang.Thread.run(Thread.java:619)
        at
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher$1.run(MapReduceLauncher.java:270)


*Caused by: java.lang.reflect.InvocationTargetException        at
sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)        at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)*
        at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
        at java.lang.reflect.Method.invoke(Method.java:597)
        at
org.apache.pig.backend.hadoop23.PigJobControl.submit(PigJobControl.java:128)
        ... 3 more
Caused by: java.lang.NoSuchMethodError:
org.apache.hadoop.fs.FileSystem.getDefaultBlockSize(Lorg/apache/hadoop/fs/Path;)J
        at
org.apache.pig.backend.hadoop.executionengine.shims.HadoopShims.getDefaultBlockSize(HadoopShims.java:108)
        at
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigInputFormat.getSplits(PigInputFormat.java:277)
        at
org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:451)
        at
org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:468)
        at
org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:360)
        at org.apache.hadoop.mapreduce.Job$11.run(Job.java:1221)
        at org.apache.hadoop.mapreduce.Job$11.run(Job.java:1218)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:396)
        at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1177)
        at org.apache.hadoop.mapreduce.Job.submit(Job.java:1218)
        at
org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob.submit(ControlledJob.java:336)
        ... 8 more
2013-11-26 13:20:35,976 [main] INFO
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
- HadoopJobId: job_1385489868206_0001
2013-11-26 13:20:35,976 [main] INFO
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
- Processing aliases A
2013-11-26 13:20:35,976 [main] INFO
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
- detailed locations: M: A[1,4] C:  R:
2013-11-26 13:20:35,979 [main] INFO
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
- 0% complete
2013-11-26 13:20:35,983 [main] WARN
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
- Ooops! Some job has failed! Specify -stop_on_failure if you want Pig to
stop immediately on failure.
2013-11-26 13:20:35,983 [main] INFO
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
- job job_1385489868206_0001 has failed! Stop running all dependent jobs
2013-11-26 13:20:35,983 [main] INFO
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
- 100% complete
2013-11-26 13:20:36,104 [main] WARN
org.apache.hadoop.mapred.ClientServiceDelegate - Exception thrown by remote
end.
RemoteTrace:
 at LocalTrace:

org.apache.hadoop.yarn.exceptions.impl.pb.YarnRemoteExceptionPBImpl: Trying
to get information for an absent application application_1385489868206_0001
        at
org.apache.hadoop.yarn.ipc.ProtoOverHadoopRpcEngine$Invoker.invoke(ProtoOverHadoopRpcEngine.java:150)
        at $Proxy7.getApplicationReport(Unknown Source)
        at
org.apache.hadoop.yarn.api.impl.pb.client.ClientRMProtocolPBClientImpl.getApplicationReport(ClientRMProtocolPBClientImpl.java:116)
        at
org.apache.hadoop.mapred.ResourceMgrDelegate.getApplicationReport(ResourceMgrDelegate.java:343)
        at
org.apache.hadoop.mapred.ClientServiceDelegate.getProxy(ClientServiceDelegate.java:142)
        at
org.apache.hadoop.mapred.ClientServiceDelegate.invoke(ClientServiceDelegate.java:288)
        at
org.apache.hadoop.mapred.ClientServiceDelegate.getJobStatus(ClientServiceDelegate.java:365)
        at
org.apache.hadoop.mapred.YARNRunner.getJobStatus(YARNRunner.java:483)
        at org.apache.hadoop.mapreduce.Cluster.getJob(Cluster.java:186)
        at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:627)
        at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:625)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:396)
        at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1177)
        at
org.apache.hadoop.mapred.JobClient.getJobUsingCluster(JobClient.java:625)
        at
org.apache.hadoop.mapred.JobClient.getTaskReports(JobClient.java:679)
        at
org.apache.hadoop.mapred.JobClient.getMapTaskReports(JobClient.java:673)
        at
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.Launcher.getStats(Launcher.java:150)
        at
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher.launchPig(MapReduceLauncher.java:428)
        at org.apache.pig.PigServer.launchPlan(PigServer.java:1322)
        at
org.apache.pig.PigServer.executeCompiledLogicalPlan(PigServer.java:1307)
        at org.apache.pig.PigServer.storeEx(PigServer.java:978)
        at org.apache.pig.PigServer.store(PigServer.java:942)
        at org.apache.pig.PigServer.openIterator(PigServer.java:855)
        at
org.apache.pig.tools.grunt.GruntParser.processDump(GruntParser.java:774)
        at
org.apache.pig.tools.pigscript.parser.PigScriptParser.parse(PigScriptParser.java:372)
        at
org.apache.pig.tools.grunt.GruntParser.parseStopOnError(GruntParser.java:198)
        at
org.apache.pig.tools.grunt.GruntParser.parseStopOnError(GruntParser.java:173)
        at org.apache.pig.tools.grunt.Grunt.run(Grunt.java:69)
        at org.apache.pig.Main.run(Main.java:541)
        at org.apache.pig.Main.main(Main.java:156)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
        at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
        at java.lang.reflect.Method.invoke(Method.java:597)
        at org.apache.hadoop.util.RunJar.main(RunJar.java:200)
2013-11-26 13:20:36,105 [main] ERROR
org.apache.hadoop.security.UserGroupInformation -
PriviledgedActionException as:hadoop (auth:SIMPLE)
cause:org.apache.hadoop.yarn.exceptions.impl.pb.YarnRemoteExceptionPBImpl:
Trying to get information for an absent application
application_1385489868206_0001
2013-11-26 13:20:36,105 [main] ERROR
org.apache.pig.tools.pigstats.SimplePigStats - ERROR: Trying to get
information for an absent application application_1385489868206_0001
2013-11-26 13:20:36,105 [main] ERROR
org.apache.pig.tools.pigstats.PigStatsUtil - 1 map reduce job(s) failed!
2013-11-26 13:20:36,107 [main] INFO
org.apache.pig.tools.pigstats.SimplePigStats - Script Statistics:

HadoopVersion   PigVersion      UserId  StartedAt       FinishedAt
Features
0.23.1  0.12.1-SNAPSHOT hadoop  2013-11-26 13:20:32     2013-11-26
13:20:36     UNKNOWN

Failed!

Failed Jobs:
JobId   Alias   Feature Message Outputs
job_1385489868206_0001  A       MAP_ONLY        Message: Unexpected System
Error Occured: java.lang.RuntimeException:
java.lang.reflect.InvocationTargetException
        at
org.apache.pig.backend.hadoop23.PigJobControl.submit(PigJobControl.java:130)
        at
org.apache.pig.backend.hadoop23.PigJobControl.run(PigJobControl.java:191)
        at java.lang.Thread.run(Thread.java:619)
        at
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher$1.run(MapReduceLauncher.java:270)
Caused by: java.lang.reflect.InvocationTargetException
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
        at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
        at java.lang.reflect.Method.invoke(Method.java:597)
        at
org.apache.pig.backend.hadoop23.PigJobControl.submit(PigJobControl.java:128)
        ... 3 more
Caused by: java.lang.NoSuchMethodError:
org.apache.hadoop.fs.FileSystem.getDefaultBlockSize(Lorg/apache/hadoop/fs/Path;)J
        at
org.apache.pig.backend.hadoop.executionengine.shims.HadoopShims.getDefaultBlockSize(HadoopShims.java:108)
        at
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigInputFormat.getSplits(PigInputFormat.java:277)
        at
org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:451)
        at
org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:468)
        at
org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:360)
        at org.apache.hadoop.mapreduce.Job$11.run(Job.java:1221)
        at org.apache.hadoop.mapreduce.Job$11.run(Job.java:1218)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:396)
        at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1177)
        at org.apache.hadoop.mapreduce.Job.submit(Job.java:1218)
        at
org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob.submit(ControlledJob.java:336)
        ... 8 more
        hdfs://sasserver05:54310/tmp/temp601167191/tmp-1605231029,

Input(s):
Failed to read data from
"hdfs://sasserver05:54310/user/hadoop/NYSE_dividends"

Output(s):
Failed to produce result in
"hdfs://sasserver05:54310/tmp/temp601167191/tmp-1605231029"

Counters:
Total records written : 0
Total bytes written : 0
Spillable Memory Manager spill count : 0
Total bags proactively spilled: 0
Total records proactively spilled: 0

Job DAG:
job_1385489868206_0001


2013-11-26 13:20:36,107 [main] INFO
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
- Failed!
2013-11-26 13:20:36,112 [main] ERROR org.apache.pig.tools.grunt.Grunt -
ERROR 1066: Unable to open iterator for alias A. Backend error : Trying to
get information for an absent application application_1385489868206_0001
Details at logfile: /home/hadoop/pig_1385489876211.log
grunt>

=============



On Tue, Nov 26, 2013 at 3:32 PM, Claudio Romo Otto <
claudio.romo@namictech.com> wrote:

> Hi Juan,
>
>     In a nutshell, you must pay attention to memory settings inside
> mapred-site.xml, yarn-site.xml, hadoop-env.sh and yarn-env.sh, so you have
> to design a memory distribution strategy according to your performance
> requirements. In this way you will have, among other things, enough memory
> for the Scheduler.
>
> Remember to reserve at least 600 - 800 mb for the operative system to
> avoid OOM errors.
>
> Best regards
> El 26/11/13 16:07, Juan Martin Pampliega escribió:
>
>> Hi Claudio,
>>
>> It would be nice to know which were the settings that you had to tune to
>> get this. I am having a similar issue with some jobs that I am running.
>> Thanks,
>> Juan.
>>
>>
>> On Wed, Oct 30, 2013 at 7:40 PM, Claudio Romo Otto <
>> claudio.romo@namictech.com> wrote:
>>
>>  Jarcec, finally I got solved this problem by learning more on hadoop 2
>>> (lot of reading), and then tuning some settings to let the work move from
>>> the SCHEDULED state. With this said, the last problem was only concerning
>>> on hadoop.
>>>
>>> Thanks for your support!
>>>
>>> El 30/10/13 18:03, Jarek Jarcec Cecho escribió:
>>>
>>>   Hi Claudio,
>>>
>>>> it's hard to guess from the limited information. I would suggest to take
>>>> a look into logs to see what is happening.
>>>>
>>>> One guess though - You've mentioned that the task was "running" for 30
>>>> minutes, but it still seems to be in SCHEDULED time - are your node
>>>> managers correctly running?
>>>>
>>>> Jarcec
>>>>
>>>> On Fri, Oct 25, 2013 at 04:10:12PM -0300, Claudio Romo Otto wrote:
>>>>
>>>>  You got it!
>>>>>
>>>>> The solution was to compile with  -Dhadoopversion=23 option. After
>>>>> your message I tried another test removing Cassandra from the chain
>>>>> and Pig sent successfully the job to hadoop.
>>>>>
>>>>> BUT! the problem changed, now the Map task remains forever stuck on
>>>>> Hadoop (30 minutes waiting, no other jobs running):
>>>>>
>>>>> Task
>>>>>
>>>>> Progress
>>>>>
>>>>> State
>>>>>
>>>>> Start Time
>>>>>
>>>>> Finish Time
>>>>>
>>>>> Elapsed Time
>>>>> task_1382631533263_0012_m_000000 <http://topgps-test-3.
>>>>> dnsalias.com:8088/proxy/application_1382631533263_
>>>>> 0012/mapreduce/task/task_1382631533263_0012_m_000000>
>>>>>
>>>>>          SCHEDULED       Fri, 25 Oct 2013 18:18:32 GMT   N/A     0sec
>>>>>
>>>>>
>>>>>
>>>>> Attempt
>>>>>
>>>>> Progress
>>>>>
>>>>> State
>>>>>
>>>>> Node
>>>>>
>>>>> Logs
>>>>>
>>>>> Started
>>>>>
>>>>> Finished
>>>>>
>>>>> Elapsed
>>>>>
>>>>> Note
>>>>> attempt_1382631533263_0012_m_000000_0   0,00    STARTING        N/A
>>>>> N/A     N/A
>>>>> N/A     0sec
>>>>>
>>>>>
>>>>> Don't know if this is a Hadoop problem or Pig, what do you think?
>>>>>
>>>>>
>>>>> El 25/10/13 13:11, Jarek Jarcec Cecho escribió:
>>>>>
>>>>>  It seems that Pig was correctly compiled against Hadoop 23, but the
>>>>>> Cassandra piece was not, check out the where the exception is coming
>>>>>> from:
>>>>>>
>>>>>>   Caused by: java.lang.IncompatibleClassChangeError: Found interface
>>>>>>
>>>>>>> org.apache.hadoop.mapreduce.JobContext, but class was expected
>>>>>>>       at org.apache.cassandra.hadoop.AbstractColumnFamilyInputForma
>>>>>>> t.getSplits(AbstractColumnFamilyInputFormat.java:113)
>>>>>>>
>>>>>>>  So, I would say that you also need to get Hadoop 2 compatible
>>>>>> Cassandra
>>>>>> connector first.
>>>>>>
>>>>>> Jarcec
>>>>>>
>>>>>> On Thu, Oct 24, 2013 at 10:34:49PM -0300, Claudio Romo Otto wrote:
>>>>>>
>>>>>>  After change from hadoop20 to hadoop23 the warning dissapeared but I
>>>>>>> got the same exception (Found interface
>>>>>>> org.apache.hadoop.mapreduce.JobContext, but class was expected)
>>>>>>>
>>>>>>> I have tried over a fresh install: hadoop 2.2.0 and pig 0.12.1
>>>>>>> compiled by me, no other product nor configuration, just two
>>>>>>> servers, one master with ResourceManager and NameNode, one slave
>>>>>>> with DataNode and NodeManager.
>>>>>>>
>>>>>>> I can't understand why over this fresh cluster Pig 0.12 fails. Here
>>>>>>> is the new trace:
>>>>>>>
>>>>>>> 2013-10-24 16:10:52,351 [JobControl] ERROR
>>>>>>> org.apache.pig.backend.hadoop23.PigJobControl - Error while trying
>>>>>>> to run jobs.
>>>>>>> java.lang.RuntimeException: java.lang.reflect.
>>>>>>> InvocationTargetException
>>>>>>>       at org.apache.pig.backend.hadoop23.PigJobControl.submit(
>>>>>>> PigJobControl.java:130)
>>>>>>>       at org.apache.pig.backend.hadoop23.PigJobControl.run(
>>>>>>> PigJobControl.java:191)
>>>>>>>       at java.lang.Thread.run(Thread.java:724)
>>>>>>>       at org.apache.pig.backend.hadoop.executionengine.
>>>>>>> mapReduceLayer.
>>>>>>> MapReduceLauncher$1.run(MapReduceLauncher.java:257)
>>>>>>> Caused by: java.lang.reflect.InvocationTargetException
>>>>>>>       at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>>>>>       at sun.reflect.NativeMethodAccessorImpl.invoke(
>>>>>>> NativeMethodAccessorImpl.java:57)
>>>>>>>       at sun.reflect.DelegatingMethodAccessorImpl.invoke(
>>>>>>> DelegatingMethodAccessorImpl.java:43)
>>>>>>>       at java.lang.reflect.Method.invoke(Method.java:606)
>>>>>>>       at org.apache.pig.backend.hadoop23.PigJobControl.submit(
>>>>>>> PigJobControl.java:128)
>>>>>>>       ... 3 more
>>>>>>> Caused by: java.lang.IncompatibleClassChangeError: Found interface
>>>>>>> org.apache.hadoop.mapreduce.JobContext, but class was expected
>>>>>>>       at org.apache.cassandra.hadoop.AbstractColumnFamilyInputForma
>>>>>>> t.getSplits(AbstractColumnFamilyInputFormat.java:113)
>>>>>>>       at org.apache.pig.backend.hadoop.executionengine.
>>>>>>> mapReduceLayer.PigInputFormat.getSplits(PigInputFormat.java:274)
>>>>>>>       at org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(
>>>>>>> JobSubmitter.java:491)
>>>>>>>       at org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(
>>>>>>> JobSubmitter.java:508)
>>>>>>>       at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(
>>>>>>> JobSubmitter.java:392)
>>>>>>>       at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1268)
>>>>>>>       at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1265)
>>>>>>>       at java.security.AccessController.doPrivileged(Native Method)
>>>>>>>       at javax.security.auth.Subject.doAs(Subject.java:415)
>>>>>>>       at org.apache.hadoop.security.UserGroupInformation.doAs(
>>>>>>> UserGroupInformation.java:1491)
>>>>>>>       at org.apache.hadoop.mapreduce.Job.submit(Job.java:1265)
>>>>>>>       at org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob.
>>>>>>> submit(ControlledJob.java:335)
>>>>>>>       ... 8 more
>>>>>>>
>>>>>>>
>>>>>>> El 24/10/13 21:33, Prashant Kommireddi escribió:
>>>>>>>
>>>>>>>  Yes it does. You need to recompile Pig for hadoop 2
>>>>>>>>
>>>>>>>> ant clean jar-withouthadoop -Dhadoopversion=23
>>>>>>>>
>>>>>>>>
>>>>>>>> On Thu, Oct 24, 2013 at 5:37 AM, Claudio Romo Otto <
>>>>>>>> claudio.romo@namictech.com> wrote:
>>>>>>>>
>>>>>>>>   Does Pig support Hadoop 2.2? When I try Pig 0.12 and Hadoop 2.2 I
>>>>>>>>
>>>>>>>>> get an
>>>>>>>>> error even with simple operations like
>>>>>>>>>
>>>>>>>>> data = LOAD 'cql://keyspace1/testcf?' USING CqlStorage();
>>>>>>>>> dump data;
>>>>>>>>>
>>>>>>>>> I only got a warning first and then and exception:
>>>>>>>>>
>>>>>>>>> 2013-10-24 09:35:19,300 [main] WARN org.apache.pig.backend.**
>>>>>>>>> hadoop20.PigJobControl
>>>>>>>>> - falling back to default JobControl (not using hadoop 0.20 ?)
>>>>>>>>> java.lang.**NoSuchFieldException: runnerState
>>>>>>>>>       at java.lang.Class.**getDeclaredField(Class.java:**1938)
>>>>>>>>>       at org.apache.pig.backend.**hadoop20.PigJobControl.<**
>>>>>>>>> clinit>(PigJobControl.java:51)
>>>>>>>>>       at org.apache.pig.backend.hadoop.**executionengine.shims.**
>>>>>>>>> HadoopShims.newJobControl(**HadoopShims.java:97)
>>>>>>>>>       at org.apache.pig.backend.hadoop.**executionengine.**
>>>>>>>>> mapReduceLayer.**
>>>>>>>>> JobControlCompiler.compile(**JobControlCompiler.java:285)
>>>>>>>>>       at org.apache.pig.backend.hadoop.**executionengine.**
>>>>>>>>> mapReduceLayer.**
>>>>>>>>> MapReduceLauncher.launchPig(**MapReduceLauncher.java:177)
>>>>>>>>>       at org.apache.pig.PigServer.**launchPlan(PigServer.java:**
>>>>>>>>> 1264)
>>>>>>>>>       at org.apache.pig.PigServer.**executeCompiledLogicalPlan(**
>>>>>>>>> PigServer.java:1249)
>>>>>>>>>       at org.apache.pig.PigServer.**storeEx(PigServer.java:931)
>>>>>>>>>       at org.apache.pig.PigServer.**store(PigServer.java:898)
>>>>>>>>>       at org.apache.pig.PigServer.**openIterator(PigServer.java:**
>>>>>>>>> 811)
>>>>>>>>>       at org.apache.pig.tools.grunt.**GruntParser.processDump(**
>>>>>>>>> GruntParser.java:696)
>>>>>>>>>       at org.apache.pig.tools.**pigscript.parser.**
>>>>>>>>> PigScriptParser.parse(**
>>>>>>>>> PigScriptParser.java:320)
>>>>>>>>>       at org.apache.pig.tools.grunt.**
>>>>>>>>> GruntParser.parseStopOnError(**
>>>>>>>>> GruntParser.java:194)
>>>>>>>>>       at org.apache.pig.tools.grunt.**
>>>>>>>>> GruntParser.parseStopOnError(**
>>>>>>>>> GruntParser.java:170)
>>>>>>>>>       at org.apache.pig.tools.grunt.**Grunt.run(Grunt.java:69)
>>>>>>>>>       at org.apache.pig.Main.run(Main.**java:538)
>>>>>>>>>       at org.apache.pig.Main.main(Main.**java:157)
>>>>>>>>>       at sun.reflect.**NativeMethodAccessorImpl.**invoke0(Native
>>>>>>>>> Method)
>>>>>>>>>       at sun.reflect.**NativeMethodAccessorImpl.**invoke(**
>>>>>>>>> NativeMethodAccessorImpl.java:**57)
>>>>>>>>>       at sun.reflect.**DelegatingMethodAccessorImpl.**invoke(**
>>>>>>>>> DelegatingMethodAccessorImpl.**java:43)
>>>>>>>>>       at java.lang.reflect.Method.**invoke(Method.java:606)
>>>>>>>>>       at org.apache.hadoop.util.RunJar.**main(RunJar.java:212)
>>>>>>>>>
>>>>>>>>> ------------------------------**--------
>>>>>>>>>
>>>>>>>>> Backend error message during job submission
>>>>>>>>> ------------------------------**-------------
>>>>>>>>> Unexpected System Error Occured: java.lang.**
>>>>>>>>> IncompatibleClassChangeError:
>>>>>>>>> Found interface org.apache.hadoop.mapreduce.**JobContext, but
>>>>>>>>> class
>>>>>>>>> was
>>>>>>>>> expected
>>>>>>>>>           at org.apache.pig.backend.hadoop.**executionengine.**
>>>>>>>>> mapReduceLayer.**PigOutputFormat.**setupUdfEnvAndStores(**
>>>>>>>>> PigOutputFormat.java:225)
>>>>>>>>>           at org.apache.pig.backend.hadoop.**executionengine.**
>>>>>>>>> mapReduceLayer.**PigOutputFormat.**checkOutputSpecs(**
>>>>>>>>> PigOutputFormat.java:186)
>>>>>>>>>           at org.apache.hadoop.mapreduce.**
>>>>>>>>> JobSubmitter.checkSpecs(**
>>>>>>>>> JobSubmitter.java:456)
>>>>>>>>>           at org.apache.hadoop.mapreduce.**JobSubmitter.**
>>>>>>>>> submitJobInternal(
>>>>>>>>> **JobSubmitter.java:342)
>>>>>>>>>           at org.apache.hadoop.mapreduce.**
>>>>>>>>> Job$10.run(Job.java:1268)
>>>>>>>>>           at org.apache.hadoop.mapreduce.**
>>>>>>>>> Job$10.run(Job.java:1265)
>>>>>>>>>           at java.security.**AccessController.doPrivileged(
>>>>>>>>> **Native
>>>>>>>>> Method)
>>>>>>>>>           at javax.security.auth.Subject.**doAs(Subject.java:415)
>>>>>>>>>           at org.apache.hadoop.security.**
>>>>>>>>> UserGroupInformation.doAs(**
>>>>>>>>> UserGroupInformation.java:**1491)
>>>>>>>>>           at org.apache.hadoop.mapreduce.**
>>>>>>>>> Job.submit(Job.java:1265)
>>>>>>>>>           at org.apache.hadoop.mapreduce.**
>>>>>>>>> lib.jobcontrol.ControlledJob.**
>>>>>>>>> submit(ControlledJob.java:335)
>>>>>>>>>           at org.apache.hadoop.mapreduce.**
>>>>>>>>> lib.jobcontrol.JobControl.run(**
>>>>>>>>> JobControl.java:240)
>>>>>>>>>           at org.apache.pig.backend.**
>>>>>>>>> hadoop20.PigJobControl.run(**
>>>>>>>>> PigJobControl.java:121)
>>>>>>>>>           at java.lang.Thread.run(Thread.**java:724)
>>>>>>>>>           at org.apache.pig.backend.hadoop.**executionengine.**
>>>>>>>>> mapReduceLayer.**MapReduceLauncher$1.run(**
>>>>>>>>> MapReduceLauncher.java:257)
>>>>>>>>>
>>>>>>>>> Pig Stack Trace
>>>>>>>>> ---------------
>>>>>>>>> ERROR 1066: Unable to open iterator for alias data
>>>>>>>>>
>>>>>>>>> org.apache.pig.impl.**logicalLayer.**FrontendException: ERROR
>>>>>>>>> 1066:
>>>>>>>>> Unable to open iterator for alias data
>>>>>>>>>           at org.apache.pig.PigServer.**
>>>>>>>>> openIterator(PigServer.java:**836)
>>>>>>>>>           at org.apache.pig.tools.grunt.**
>>>>>>>>> GruntParser.processDump(**
>>>>>>>>> GruntParser.java:696)
>>>>>>>>>           at org.apache.pig.tools.**pigscript.parser.**
>>>>>>>>> PigScriptParser.parse(**PigScriptParser.java:320)
>>>>>>>>>           at org.apache.pig.tools.grunt.**
>>>>>>>>> GruntParser.parseStopOnError(**
>>>>>>>>> GruntParser.java:194)
>>>>>>>>>           at org.apache.pig.tools.grunt.**
>>>>>>>>> GruntParser.parseStopOnError(**
>>>>>>>>> GruntParser.java:170)
>>>>>>>>>           at org.apache.pig.tools.grunt.**Grunt.run(Grunt.java:69)
>>>>>>>>>           at org.apache.pig.Main.run(Main.**java:538)
>>>>>>>>>           at org.apache.pig.Main.main(Main.**java:157)
>>>>>>>>>           at sun.reflect.**NativeMethodAccessorImpl.**
>>>>>>>>> invoke0(Native
>>>>>>>>> Method)
>>>>>>>>>           at sun.reflect.**NativeMethodAccessorImpl.**invoke(**
>>>>>>>>> NativeMethodAccessorImpl.java:**57)
>>>>>>>>>           at sun.reflect.**DelegatingMethodAccessorImpl.*
>>>>>>>>> *invoke(**
>>>>>>>>> DelegatingMethodAccessorImpl.**java:43)
>>>>>>>>>           at java.lang.reflect.Method.**invoke(Method.java:606)
>>>>>>>>>           at org.apache.hadoop.util.RunJar.**main(RunJar.java:212)
>>>>>>>>> Caused by: java.io.IOException: Job terminated with anomalous
>>>>>>>>> status
>>>>>>>>> FAILED
>>>>>>>>>           at org.apache.pig.PigServer.**
>>>>>>>>> openIterator(PigServer.java:**828)
>>>>>>>>>           ... 12 more
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>


-- 
Regards,
Uttam Kumar

Re: Support for Hadoop 2.2

Posted by Claudio Romo Otto <cl...@namictech.com>.
Hi Juan,

     In a nutshell, you must pay attention to memory settings inside 
mapred-site.xml, yarn-site.xml, hadoop-env.sh and yarn-env.sh, so you 
have to design a memory distribution strategy according to your 
performance requirements. In this way you will have, among other things, 
enough memory for the Scheduler.

Remember to reserve at least 600 - 800 mb for the operative system to 
avoid OOM errors.

Best regards
El 26/11/13 16:07, Juan Martin Pampliega escribió:
> Hi Claudio,
>
> It would be nice to know which were the settings that you had to tune to
> get this. I am having a similar issue with some jobs that I am running.
> Thanks,
> Juan.
>
>
> On Wed, Oct 30, 2013 at 7:40 PM, Claudio Romo Otto <
> claudio.romo@namictech.com> wrote:
>
>> Jarcec, finally I got solved this problem by learning more on hadoop 2
>> (lot of reading), and then tuning some settings to let the work move from
>> the SCHEDULED state. With this said, the last problem was only concerning
>> on hadoop.
>>
>> Thanks for your support!
>>
>> El 30/10/13 18:03, Jarek Jarcec Cecho escribió:
>>
>>   Hi Claudio,
>>> it's hard to guess from the limited information. I would suggest to take
>>> a look into logs to see what is happening.
>>>
>>> One guess though - You've mentioned that the task was "running" for 30
>>> minutes, but it still seems to be in SCHEDULED time - are your node
>>> managers correctly running?
>>>
>>> Jarcec
>>>
>>> On Fri, Oct 25, 2013 at 04:10:12PM -0300, Claudio Romo Otto wrote:
>>>
>>>> You got it!
>>>>
>>>> The solution was to compile with  -Dhadoopversion=23 option. After
>>>> your message I tried another test removing Cassandra from the chain
>>>> and Pig sent successfully the job to hadoop.
>>>>
>>>> BUT! the problem changed, now the Map task remains forever stuck on
>>>> Hadoop (30 minutes waiting, no other jobs running):
>>>>
>>>> Task
>>>>
>>>> Progress
>>>>
>>>> State
>>>>
>>>> Start Time
>>>>
>>>> Finish Time
>>>>
>>>> Elapsed Time
>>>> task_1382631533263_0012_m_000000 <http://topgps-test-3.
>>>> dnsalias.com:8088/proxy/application_1382631533263_
>>>> 0012/mapreduce/task/task_1382631533263_0012_m_000000>
>>>>
>>>>          SCHEDULED       Fri, 25 Oct 2013 18:18:32 GMT   N/A     0sec
>>>>
>>>>
>>>>
>>>> Attempt
>>>>
>>>> Progress
>>>>
>>>> State
>>>>
>>>> Node
>>>>
>>>> Logs
>>>>
>>>> Started
>>>>
>>>> Finished
>>>>
>>>> Elapsed
>>>>
>>>> Note
>>>> attempt_1382631533263_0012_m_000000_0   0,00    STARTING        N/A
>>>> N/A     N/A
>>>> N/A     0sec
>>>>
>>>>
>>>> Don't know if this is a Hadoop problem or Pig, what do you think?
>>>>
>>>>
>>>> El 25/10/13 13:11, Jarek Jarcec Cecho escribió:
>>>>
>>>>> It seems that Pig was correctly compiled against Hadoop 23, but the
>>>>> Cassandra piece was not, check out the where the exception is coming from:
>>>>>
>>>>>   Caused by: java.lang.IncompatibleClassChangeError: Found interface
>>>>>> org.apache.hadoop.mapreduce.JobContext, but class was expected
>>>>>>       at org.apache.cassandra.hadoop.AbstractColumnFamilyInputForma
>>>>>> t.getSplits(AbstractColumnFamilyInputFormat.java:113)
>>>>>>
>>>>> So, I would say that you also need to get Hadoop 2 compatible Cassandra
>>>>> connector first.
>>>>>
>>>>> Jarcec
>>>>>
>>>>> On Thu, Oct 24, 2013 at 10:34:49PM -0300, Claudio Romo Otto wrote:
>>>>>
>>>>>> After change from hadoop20 to hadoop23 the warning dissapeared but I
>>>>>> got the same exception (Found interface
>>>>>> org.apache.hadoop.mapreduce.JobContext, but class was expected)
>>>>>>
>>>>>> I have tried over a fresh install: hadoop 2.2.0 and pig 0.12.1
>>>>>> compiled by me, no other product nor configuration, just two
>>>>>> servers, one master with ResourceManager and NameNode, one slave
>>>>>> with DataNode and NodeManager.
>>>>>>
>>>>>> I can't understand why over this fresh cluster Pig 0.12 fails. Here
>>>>>> is the new trace:
>>>>>>
>>>>>> 2013-10-24 16:10:52,351 [JobControl] ERROR
>>>>>> org.apache.pig.backend.hadoop23.PigJobControl - Error while trying
>>>>>> to run jobs.
>>>>>> java.lang.RuntimeException: java.lang.reflect.
>>>>>> InvocationTargetException
>>>>>>       at org.apache.pig.backend.hadoop23.PigJobControl.submit(
>>>>>> PigJobControl.java:130)
>>>>>>       at org.apache.pig.backend.hadoop23.PigJobControl.run(
>>>>>> PigJobControl.java:191)
>>>>>>       at java.lang.Thread.run(Thread.java:724)
>>>>>>       at org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.
>>>>>> MapReduceLauncher$1.run(MapReduceLauncher.java:257)
>>>>>> Caused by: java.lang.reflect.InvocationTargetException
>>>>>>       at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>>>>       at sun.reflect.NativeMethodAccessorImpl.invoke(
>>>>>> NativeMethodAccessorImpl.java:57)
>>>>>>       at sun.reflect.DelegatingMethodAccessorImpl.invoke(
>>>>>> DelegatingMethodAccessorImpl.java:43)
>>>>>>       at java.lang.reflect.Method.invoke(Method.java:606)
>>>>>>       at org.apache.pig.backend.hadoop23.PigJobControl.submit(
>>>>>> PigJobControl.java:128)
>>>>>>       ... 3 more
>>>>>> Caused by: java.lang.IncompatibleClassChangeError: Found interface
>>>>>> org.apache.hadoop.mapreduce.JobContext, but class was expected
>>>>>>       at org.apache.cassandra.hadoop.AbstractColumnFamilyInputForma
>>>>>> t.getSplits(AbstractColumnFamilyInputFormat.java:113)
>>>>>>       at org.apache.pig.backend.hadoop.executionengine.
>>>>>> mapReduceLayer.PigInputFormat.getSplits(PigInputFormat.java:274)
>>>>>>       at org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(
>>>>>> JobSubmitter.java:491)
>>>>>>       at org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(
>>>>>> JobSubmitter.java:508)
>>>>>>       at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(
>>>>>> JobSubmitter.java:392)
>>>>>>       at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1268)
>>>>>>       at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1265)
>>>>>>       at java.security.AccessController.doPrivileged(Native Method)
>>>>>>       at javax.security.auth.Subject.doAs(Subject.java:415)
>>>>>>       at org.apache.hadoop.security.UserGroupInformation.doAs(
>>>>>> UserGroupInformation.java:1491)
>>>>>>       at org.apache.hadoop.mapreduce.Job.submit(Job.java:1265)
>>>>>>       at org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob.
>>>>>> submit(ControlledJob.java:335)
>>>>>>       ... 8 more
>>>>>>
>>>>>>
>>>>>> El 24/10/13 21:33, Prashant Kommireddi escribió:
>>>>>>
>>>>>>> Yes it does. You need to recompile Pig for hadoop 2
>>>>>>>
>>>>>>> ant clean jar-withouthadoop -Dhadoopversion=23
>>>>>>>
>>>>>>>
>>>>>>> On Thu, Oct 24, 2013 at 5:37 AM, Claudio Romo Otto <
>>>>>>> claudio.romo@namictech.com> wrote:
>>>>>>>
>>>>>>>   Does Pig support Hadoop 2.2? When I try Pig 0.12 and Hadoop 2.2 I
>>>>>>>> get an
>>>>>>>> error even with simple operations like
>>>>>>>>
>>>>>>>> data = LOAD 'cql://keyspace1/testcf?' USING CqlStorage();
>>>>>>>> dump data;
>>>>>>>>
>>>>>>>> I only got a warning first and then and exception:
>>>>>>>>
>>>>>>>> 2013-10-24 09:35:19,300 [main] WARN org.apache.pig.backend.**
>>>>>>>> hadoop20.PigJobControl
>>>>>>>> - falling back to default JobControl (not using hadoop 0.20 ?)
>>>>>>>> java.lang.**NoSuchFieldException: runnerState
>>>>>>>>       at java.lang.Class.**getDeclaredField(Class.java:**1938)
>>>>>>>>       at org.apache.pig.backend.**hadoop20.PigJobControl.<**
>>>>>>>> clinit>(PigJobControl.java:51)
>>>>>>>>       at org.apache.pig.backend.hadoop.**executionengine.shims.**
>>>>>>>> HadoopShims.newJobControl(**HadoopShims.java:97)
>>>>>>>>       at org.apache.pig.backend.hadoop.**executionengine.**
>>>>>>>> mapReduceLayer.**
>>>>>>>> JobControlCompiler.compile(**JobControlCompiler.java:285)
>>>>>>>>       at org.apache.pig.backend.hadoop.**executionengine.**
>>>>>>>> mapReduceLayer.**
>>>>>>>> MapReduceLauncher.launchPig(**MapReduceLauncher.java:177)
>>>>>>>>       at org.apache.pig.PigServer.**launchPlan(PigServer.java:**1264)
>>>>>>>>       at org.apache.pig.PigServer.**executeCompiledLogicalPlan(**
>>>>>>>> PigServer.java:1249)
>>>>>>>>       at org.apache.pig.PigServer.**storeEx(PigServer.java:931)
>>>>>>>>       at org.apache.pig.PigServer.**store(PigServer.java:898)
>>>>>>>>       at org.apache.pig.PigServer.**openIterator(PigServer.java:**
>>>>>>>> 811)
>>>>>>>>       at org.apache.pig.tools.grunt.**GruntParser.processDump(**
>>>>>>>> GruntParser.java:696)
>>>>>>>>       at org.apache.pig.tools.**pigscript.parser.**
>>>>>>>> PigScriptParser.parse(**
>>>>>>>> PigScriptParser.java:320)
>>>>>>>>       at org.apache.pig.tools.grunt.**GruntParser.parseStopOnError(**
>>>>>>>> GruntParser.java:194)
>>>>>>>>       at org.apache.pig.tools.grunt.**GruntParser.parseStopOnError(**
>>>>>>>> GruntParser.java:170)
>>>>>>>>       at org.apache.pig.tools.grunt.**Grunt.run(Grunt.java:69)
>>>>>>>>       at org.apache.pig.Main.run(Main.**java:538)
>>>>>>>>       at org.apache.pig.Main.main(Main.**java:157)
>>>>>>>>       at sun.reflect.**NativeMethodAccessorImpl.**invoke0(Native
>>>>>>>> Method)
>>>>>>>>       at sun.reflect.**NativeMethodAccessorImpl.**invoke(**
>>>>>>>> NativeMethodAccessorImpl.java:**57)
>>>>>>>>       at sun.reflect.**DelegatingMethodAccessorImpl.**invoke(**
>>>>>>>> DelegatingMethodAccessorImpl.**java:43)
>>>>>>>>       at java.lang.reflect.Method.**invoke(Method.java:606)
>>>>>>>>       at org.apache.hadoop.util.RunJar.**main(RunJar.java:212)
>>>>>>>>
>>>>>>>> ------------------------------**--------
>>>>>>>>
>>>>>>>> Backend error message during job submission
>>>>>>>> ------------------------------**-------------
>>>>>>>> Unexpected System Error Occured: java.lang.**
>>>>>>>> IncompatibleClassChangeError:
>>>>>>>> Found interface org.apache.hadoop.mapreduce.**JobContext, but class
>>>>>>>> was
>>>>>>>> expected
>>>>>>>>           at org.apache.pig.backend.hadoop.**executionengine.**
>>>>>>>> mapReduceLayer.**PigOutputFormat.**setupUdfEnvAndStores(**
>>>>>>>> PigOutputFormat.java:225)
>>>>>>>>           at org.apache.pig.backend.hadoop.**executionengine.**
>>>>>>>> mapReduceLayer.**PigOutputFormat.**checkOutputSpecs(**
>>>>>>>> PigOutputFormat.java:186)
>>>>>>>>           at org.apache.hadoop.mapreduce.**JobSubmitter.checkSpecs(**
>>>>>>>> JobSubmitter.java:456)
>>>>>>>>           at org.apache.hadoop.mapreduce.**JobSubmitter.**
>>>>>>>> submitJobInternal(
>>>>>>>> **JobSubmitter.java:342)
>>>>>>>>           at org.apache.hadoop.mapreduce.**Job$10.run(Job.java:1268)
>>>>>>>>           at org.apache.hadoop.mapreduce.**Job$10.run(Job.java:1265)
>>>>>>>>           at java.security.**AccessController.doPrivileged(**Native
>>>>>>>> Method)
>>>>>>>>           at javax.security.auth.Subject.**doAs(Subject.java:415)
>>>>>>>>           at org.apache.hadoop.security.**
>>>>>>>> UserGroupInformation.doAs(**
>>>>>>>> UserGroupInformation.java:**1491)
>>>>>>>>           at org.apache.hadoop.mapreduce.**Job.submit(Job.java:1265)
>>>>>>>>           at org.apache.hadoop.mapreduce.**
>>>>>>>> lib.jobcontrol.ControlledJob.**
>>>>>>>> submit(ControlledJob.java:335)
>>>>>>>>           at org.apache.hadoop.mapreduce.**
>>>>>>>> lib.jobcontrol.JobControl.run(**
>>>>>>>> JobControl.java:240)
>>>>>>>>           at org.apache.pig.backend.**hadoop20.PigJobControl.run(**
>>>>>>>> PigJobControl.java:121)
>>>>>>>>           at java.lang.Thread.run(Thread.**java:724)
>>>>>>>>           at org.apache.pig.backend.hadoop.**executionengine.**
>>>>>>>> mapReduceLayer.**MapReduceLauncher$1.run(**
>>>>>>>> MapReduceLauncher.java:257)
>>>>>>>>
>>>>>>>> Pig Stack Trace
>>>>>>>> ---------------
>>>>>>>> ERROR 1066: Unable to open iterator for alias data
>>>>>>>>
>>>>>>>> org.apache.pig.impl.**logicalLayer.**FrontendException: ERROR 1066:
>>>>>>>> Unable to open iterator for alias data
>>>>>>>>           at org.apache.pig.PigServer.**
>>>>>>>> openIterator(PigServer.java:**836)
>>>>>>>>           at org.apache.pig.tools.grunt.**GruntParser.processDump(**
>>>>>>>> GruntParser.java:696)
>>>>>>>>           at org.apache.pig.tools.**pigscript.parser.**
>>>>>>>> PigScriptParser.parse(**PigScriptParser.java:320)
>>>>>>>>           at org.apache.pig.tools.grunt.**
>>>>>>>> GruntParser.parseStopOnError(**
>>>>>>>> GruntParser.java:194)
>>>>>>>>           at org.apache.pig.tools.grunt.**
>>>>>>>> GruntParser.parseStopOnError(**
>>>>>>>> GruntParser.java:170)
>>>>>>>>           at org.apache.pig.tools.grunt.**Grunt.run(Grunt.java:69)
>>>>>>>>           at org.apache.pig.Main.run(Main.**java:538)
>>>>>>>>           at org.apache.pig.Main.main(Main.**java:157)
>>>>>>>>           at sun.reflect.**NativeMethodAccessorImpl.**invoke0(Native
>>>>>>>> Method)
>>>>>>>>           at sun.reflect.**NativeMethodAccessorImpl.**invoke(**
>>>>>>>> NativeMethodAccessorImpl.java:**57)
>>>>>>>>           at sun.reflect.**DelegatingMethodAccessorImpl.**invoke(**
>>>>>>>> DelegatingMethodAccessorImpl.**java:43)
>>>>>>>>           at java.lang.reflect.Method.**invoke(Method.java:606)
>>>>>>>>           at org.apache.hadoop.util.RunJar.**main(RunJar.java:212)
>>>>>>>> Caused by: java.io.IOException: Job terminated with anomalous status
>>>>>>>> FAILED
>>>>>>>>           at org.apache.pig.PigServer.**
>>>>>>>> openIterator(PigServer.java:**828)
>>>>>>>>           ... 12 more
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>


Re: Support for Hadoop 2.2

Posted by Juan Martin Pampliega <jp...@gmail.com>.
Hi Claudio,

It would be nice to know which were the settings that you had to tune to
get this. I am having a similar issue with some jobs that I am running.
Thanks,
Juan.


On Wed, Oct 30, 2013 at 7:40 PM, Claudio Romo Otto <
claudio.romo@namictech.com> wrote:

> Jarcec, finally I got solved this problem by learning more on hadoop 2
> (lot of reading), and then tuning some settings to let the work move from
> the SCHEDULED state. With this said, the last problem was only concerning
> on hadoop.
>
> Thanks for your support!
>
> El 30/10/13 18:03, Jarek Jarcec Cecho escribió:
>
>  Hi Claudio,
>> it's hard to guess from the limited information. I would suggest to take
>> a look into logs to see what is happening.
>>
>> One guess though - You've mentioned that the task was "running" for 30
>> minutes, but it still seems to be in SCHEDULED time - are your node
>> managers correctly running?
>>
>> Jarcec
>>
>> On Fri, Oct 25, 2013 at 04:10:12PM -0300, Claudio Romo Otto wrote:
>>
>>> You got it!
>>>
>>> The solution was to compile with  -Dhadoopversion=23 option. After
>>> your message I tried another test removing Cassandra from the chain
>>> and Pig sent successfully the job to hadoop.
>>>
>>> BUT! the problem changed, now the Map task remains forever stuck on
>>> Hadoop (30 minutes waiting, no other jobs running):
>>>
>>> Task
>>>
>>> Progress
>>>
>>> State
>>>
>>> Start Time
>>>
>>> Finish Time
>>>
>>> Elapsed Time
>>> task_1382631533263_0012_m_000000 <http://topgps-test-3.
>>> dnsalias.com:8088/proxy/application_1382631533263_
>>> 0012/mapreduce/task/task_1382631533263_0012_m_000000>
>>>
>>>         SCHEDULED       Fri, 25 Oct 2013 18:18:32 GMT   N/A     0sec
>>>
>>>
>>>
>>> Attempt
>>>
>>> Progress
>>>
>>> State
>>>
>>> Node
>>>
>>> Logs
>>>
>>> Started
>>>
>>> Finished
>>>
>>> Elapsed
>>>
>>> Note
>>> attempt_1382631533263_0012_m_000000_0   0,00    STARTING        N/A
>>> N/A     N/A
>>> N/A     0sec
>>>
>>>
>>> Don't know if this is a Hadoop problem or Pig, what do you think?
>>>
>>>
>>> El 25/10/13 13:11, Jarek Jarcec Cecho escribió:
>>>
>>>> It seems that Pig was correctly compiled against Hadoop 23, but the
>>>> Cassandra piece was not, check out the where the exception is coming from:
>>>>
>>>>  Caused by: java.lang.IncompatibleClassChangeError: Found interface
>>>>> org.apache.hadoop.mapreduce.JobContext, but class was expected
>>>>>      at org.apache.cassandra.hadoop.AbstractColumnFamilyInputForma
>>>>> t.getSplits(AbstractColumnFamilyInputFormat.java:113)
>>>>>
>>>> So, I would say that you also need to get Hadoop 2 compatible Cassandra
>>>> connector first.
>>>>
>>>> Jarcec
>>>>
>>>> On Thu, Oct 24, 2013 at 10:34:49PM -0300, Claudio Romo Otto wrote:
>>>>
>>>>> After change from hadoop20 to hadoop23 the warning dissapeared but I
>>>>> got the same exception (Found interface
>>>>> org.apache.hadoop.mapreduce.JobContext, but class was expected)
>>>>>
>>>>> I have tried over a fresh install: hadoop 2.2.0 and pig 0.12.1
>>>>> compiled by me, no other product nor configuration, just two
>>>>> servers, one master with ResourceManager and NameNode, one slave
>>>>> with DataNode and NodeManager.
>>>>>
>>>>> I can't understand why over this fresh cluster Pig 0.12 fails. Here
>>>>> is the new trace:
>>>>>
>>>>> 2013-10-24 16:10:52,351 [JobControl] ERROR
>>>>> org.apache.pig.backend.hadoop23.PigJobControl - Error while trying
>>>>> to run jobs.
>>>>> java.lang.RuntimeException: java.lang.reflect.
>>>>> InvocationTargetException
>>>>>      at org.apache.pig.backend.hadoop23.PigJobControl.submit(
>>>>> PigJobControl.java:130)
>>>>>      at org.apache.pig.backend.hadoop23.PigJobControl.run(
>>>>> PigJobControl.java:191)
>>>>>      at java.lang.Thread.run(Thread.java:724)
>>>>>      at org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.
>>>>> MapReduceLauncher$1.run(MapReduceLauncher.java:257)
>>>>> Caused by: java.lang.reflect.InvocationTargetException
>>>>>      at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>>>      at sun.reflect.NativeMethodAccessorImpl.invoke(
>>>>> NativeMethodAccessorImpl.java:57)
>>>>>      at sun.reflect.DelegatingMethodAccessorImpl.invoke(
>>>>> DelegatingMethodAccessorImpl.java:43)
>>>>>      at java.lang.reflect.Method.invoke(Method.java:606)
>>>>>      at org.apache.pig.backend.hadoop23.PigJobControl.submit(
>>>>> PigJobControl.java:128)
>>>>>      ... 3 more
>>>>> Caused by: java.lang.IncompatibleClassChangeError: Found interface
>>>>> org.apache.hadoop.mapreduce.JobContext, but class was expected
>>>>>      at org.apache.cassandra.hadoop.AbstractColumnFamilyInputForma
>>>>> t.getSplits(AbstractColumnFamilyInputFormat.java:113)
>>>>>      at org.apache.pig.backend.hadoop.executionengine.
>>>>> mapReduceLayer.PigInputFormat.getSplits(PigInputFormat.java:274)
>>>>>      at org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(
>>>>> JobSubmitter.java:491)
>>>>>      at org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(
>>>>> JobSubmitter.java:508)
>>>>>      at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(
>>>>> JobSubmitter.java:392)
>>>>>      at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1268)
>>>>>      at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1265)
>>>>>      at java.security.AccessController.doPrivileged(Native Method)
>>>>>      at javax.security.auth.Subject.doAs(Subject.java:415)
>>>>>      at org.apache.hadoop.security.UserGroupInformation.doAs(
>>>>> UserGroupInformation.java:1491)
>>>>>      at org.apache.hadoop.mapreduce.Job.submit(Job.java:1265)
>>>>>      at org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob.
>>>>> submit(ControlledJob.java:335)
>>>>>      ... 8 more
>>>>>
>>>>>
>>>>> El 24/10/13 21:33, Prashant Kommireddi escribió:
>>>>>
>>>>>> Yes it does. You need to recompile Pig for hadoop 2
>>>>>>
>>>>>> ant clean jar-withouthadoop -Dhadoopversion=23
>>>>>>
>>>>>>
>>>>>> On Thu, Oct 24, 2013 at 5:37 AM, Claudio Romo Otto <
>>>>>> claudio.romo@namictech.com> wrote:
>>>>>>
>>>>>>  Does Pig support Hadoop 2.2? When I try Pig 0.12 and Hadoop 2.2 I
>>>>>>> get an
>>>>>>> error even with simple operations like
>>>>>>>
>>>>>>> data = LOAD 'cql://keyspace1/testcf?' USING CqlStorage();
>>>>>>> dump data;
>>>>>>>
>>>>>>> I only got a warning first and then and exception:
>>>>>>>
>>>>>>> 2013-10-24 09:35:19,300 [main] WARN org.apache.pig.backend.**
>>>>>>> hadoop20.PigJobControl
>>>>>>> - falling back to default JobControl (not using hadoop 0.20 ?)
>>>>>>> java.lang.**NoSuchFieldException: runnerState
>>>>>>>      at java.lang.Class.**getDeclaredField(Class.java:**1938)
>>>>>>>      at org.apache.pig.backend.**hadoop20.PigJobControl.<**
>>>>>>> clinit>(PigJobControl.java:51)
>>>>>>>      at org.apache.pig.backend.hadoop.**executionengine.shims.**
>>>>>>> HadoopShims.newJobControl(**HadoopShims.java:97)
>>>>>>>      at org.apache.pig.backend.hadoop.**executionengine.**
>>>>>>> mapReduceLayer.**
>>>>>>> JobControlCompiler.compile(**JobControlCompiler.java:285)
>>>>>>>      at org.apache.pig.backend.hadoop.**executionengine.**
>>>>>>> mapReduceLayer.**
>>>>>>> MapReduceLauncher.launchPig(**MapReduceLauncher.java:177)
>>>>>>>      at org.apache.pig.PigServer.**launchPlan(PigServer.java:**1264)
>>>>>>>      at org.apache.pig.PigServer.**executeCompiledLogicalPlan(**
>>>>>>> PigServer.java:1249)
>>>>>>>      at org.apache.pig.PigServer.**storeEx(PigServer.java:931)
>>>>>>>      at org.apache.pig.PigServer.**store(PigServer.java:898)
>>>>>>>      at org.apache.pig.PigServer.**openIterator(PigServer.java:**
>>>>>>> 811)
>>>>>>>      at org.apache.pig.tools.grunt.**GruntParser.processDump(**
>>>>>>> GruntParser.java:696)
>>>>>>>      at org.apache.pig.tools.**pigscript.parser.**
>>>>>>> PigScriptParser.parse(**
>>>>>>> PigScriptParser.java:320)
>>>>>>>      at org.apache.pig.tools.grunt.**GruntParser.parseStopOnError(**
>>>>>>> GruntParser.java:194)
>>>>>>>      at org.apache.pig.tools.grunt.**GruntParser.parseStopOnError(**
>>>>>>> GruntParser.java:170)
>>>>>>>      at org.apache.pig.tools.grunt.**Grunt.run(Grunt.java:69)
>>>>>>>      at org.apache.pig.Main.run(Main.**java:538)
>>>>>>>      at org.apache.pig.Main.main(Main.**java:157)
>>>>>>>      at sun.reflect.**NativeMethodAccessorImpl.**invoke0(Native
>>>>>>> Method)
>>>>>>>      at sun.reflect.**NativeMethodAccessorImpl.**invoke(**
>>>>>>> NativeMethodAccessorImpl.java:**57)
>>>>>>>      at sun.reflect.**DelegatingMethodAccessorImpl.**invoke(**
>>>>>>> DelegatingMethodAccessorImpl.**java:43)
>>>>>>>      at java.lang.reflect.Method.**invoke(Method.java:606)
>>>>>>>      at org.apache.hadoop.util.RunJar.**main(RunJar.java:212)
>>>>>>>
>>>>>>> ------------------------------**--------
>>>>>>>
>>>>>>> Backend error message during job submission
>>>>>>> ------------------------------**-------------
>>>>>>> Unexpected System Error Occured: java.lang.**
>>>>>>> IncompatibleClassChangeError:
>>>>>>> Found interface org.apache.hadoop.mapreduce.**JobContext, but class
>>>>>>> was
>>>>>>> expected
>>>>>>>          at org.apache.pig.backend.hadoop.**executionengine.**
>>>>>>> mapReduceLayer.**PigOutputFormat.**setupUdfEnvAndStores(**
>>>>>>> PigOutputFormat.java:225)
>>>>>>>          at org.apache.pig.backend.hadoop.**executionengine.**
>>>>>>> mapReduceLayer.**PigOutputFormat.**checkOutputSpecs(**
>>>>>>> PigOutputFormat.java:186)
>>>>>>>          at org.apache.hadoop.mapreduce.**JobSubmitter.checkSpecs(**
>>>>>>> JobSubmitter.java:456)
>>>>>>>          at org.apache.hadoop.mapreduce.**JobSubmitter.**
>>>>>>> submitJobInternal(
>>>>>>> **JobSubmitter.java:342)
>>>>>>>          at org.apache.hadoop.mapreduce.**Job$10.run(Job.java:1268)
>>>>>>>          at org.apache.hadoop.mapreduce.**Job$10.run(Job.java:1265)
>>>>>>>          at java.security.**AccessController.doPrivileged(**Native
>>>>>>> Method)
>>>>>>>          at javax.security.auth.Subject.**doAs(Subject.java:415)
>>>>>>>          at org.apache.hadoop.security.**
>>>>>>> UserGroupInformation.doAs(**
>>>>>>> UserGroupInformation.java:**1491)
>>>>>>>          at org.apache.hadoop.mapreduce.**Job.submit(Job.java:1265)
>>>>>>>          at org.apache.hadoop.mapreduce.**
>>>>>>> lib.jobcontrol.ControlledJob.**
>>>>>>> submit(ControlledJob.java:335)
>>>>>>>          at org.apache.hadoop.mapreduce.**
>>>>>>> lib.jobcontrol.JobControl.run(**
>>>>>>> JobControl.java:240)
>>>>>>>          at org.apache.pig.backend.**hadoop20.PigJobControl.run(**
>>>>>>> PigJobControl.java:121)
>>>>>>>          at java.lang.Thread.run(Thread.**java:724)
>>>>>>>          at org.apache.pig.backend.hadoop.**executionengine.**
>>>>>>> mapReduceLayer.**MapReduceLauncher$1.run(**
>>>>>>> MapReduceLauncher.java:257)
>>>>>>>
>>>>>>> Pig Stack Trace
>>>>>>> ---------------
>>>>>>> ERROR 1066: Unable to open iterator for alias data
>>>>>>>
>>>>>>> org.apache.pig.impl.**logicalLayer.**FrontendException: ERROR 1066:
>>>>>>> Unable to open iterator for alias data
>>>>>>>          at org.apache.pig.PigServer.**
>>>>>>> openIterator(PigServer.java:**836)
>>>>>>>          at org.apache.pig.tools.grunt.**GruntParser.processDump(**
>>>>>>> GruntParser.java:696)
>>>>>>>          at org.apache.pig.tools.**pigscript.parser.**
>>>>>>> PigScriptParser.parse(**PigScriptParser.java:320)
>>>>>>>          at org.apache.pig.tools.grunt.**
>>>>>>> GruntParser.parseStopOnError(**
>>>>>>> GruntParser.java:194)
>>>>>>>          at org.apache.pig.tools.grunt.**
>>>>>>> GruntParser.parseStopOnError(**
>>>>>>> GruntParser.java:170)
>>>>>>>          at org.apache.pig.tools.grunt.**Grunt.run(Grunt.java:69)
>>>>>>>          at org.apache.pig.Main.run(Main.**java:538)
>>>>>>>          at org.apache.pig.Main.main(Main.**java:157)
>>>>>>>          at sun.reflect.**NativeMethodAccessorImpl.**invoke0(Native
>>>>>>> Method)
>>>>>>>          at sun.reflect.**NativeMethodAccessorImpl.**invoke(**
>>>>>>> NativeMethodAccessorImpl.java:**57)
>>>>>>>          at sun.reflect.**DelegatingMethodAccessorImpl.**invoke(**
>>>>>>> DelegatingMethodAccessorImpl.**java:43)
>>>>>>>          at java.lang.reflect.Method.**invoke(Method.java:606)
>>>>>>>          at org.apache.hadoop.util.RunJar.**main(RunJar.java:212)
>>>>>>> Caused by: java.io.IOException: Job terminated with anomalous status
>>>>>>> FAILED
>>>>>>>          at org.apache.pig.PigServer.**
>>>>>>> openIterator(PigServer.java:**828)
>>>>>>>          ... 12 more
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>

Re: Support for Hadoop 2.2

Posted by Claudio Romo Otto <cl...@namictech.com>.
Jarcec, finally I got solved this problem by learning more on hadoop 2 
(lot of reading), and then tuning some settings to let the work move 
from the SCHEDULED state. With this said, the last problem was only 
concerning on hadoop.

Thanks for your support!

El 30/10/13 18:03, Jarek Jarcec Cecho escribió:
> Hi Claudio,
> it's hard to guess from the limited information. I would suggest to take a look into logs to see what is happening.
>
> One guess though - You've mentioned that the task was "running" for 30 minutes, but it still seems to be in SCHEDULED time - are your node managers correctly running?
>
> Jarcec
>
> On Fri, Oct 25, 2013 at 04:10:12PM -0300, Claudio Romo Otto wrote:
>> You got it!
>>
>> The solution was to compile with  -Dhadoopversion=23 option. After
>> your message I tried another test removing Cassandra from the chain
>> and Pig sent successfully the job to hadoop.
>>
>> BUT! the problem changed, now the Map task remains forever stuck on
>> Hadoop (30 minutes waiting, no other jobs running):
>>
>> Task
>> 	
>> Progress
>> 	
>> State
>> 	
>> Start Time
>> 	
>> Finish Time
>> 	
>> Elapsed Time
>> task_1382631533263_0012_m_000000 <http://topgps-test-3.dnsalias.com:8088/proxy/application_1382631533263_0012/mapreduce/task/task_1382631533263_0012_m_000000>
>> 	
>> 	SCHEDULED 	Fri, 25 Oct 2013 18:18:32 GMT 	N/A 	0sec
>>
>>
>>
>> Attempt
>> 	
>> Progress
>> 	
>> State
>> 	
>> Node
>> 	
>> Logs
>> 	
>> Started
>> 	
>> Finished
>> 	
>> Elapsed
>> 	
>> Note
>> attempt_1382631533263_0012_m_000000_0 	0,00 	STARTING 	N/A 	N/A 	N/A
>> N/A 	0sec
>>
>>
>> Don't know if this is a Hadoop problem or Pig, what do you think?
>>
>>
>> El 25/10/13 13:11, Jarek Jarcec Cecho escribió:
>>> It seems that Pig was correctly compiled against Hadoop 23, but the Cassandra piece was not, check out the where the exception is coming from:
>>>
>>>> Caused by: java.lang.IncompatibleClassChangeError: Found interface
>>>> org.apache.hadoop.mapreduce.JobContext, but class was expected
>>>>      at org.apache.cassandra.hadoop.AbstractColumnFamilyInputFormat.getSplits(AbstractColumnFamilyInputFormat.java:113)
>>> So, I would say that you also need to get Hadoop 2 compatible Cassandra connector first.
>>>
>>> Jarcec
>>>
>>> On Thu, Oct 24, 2013 at 10:34:49PM -0300, Claudio Romo Otto wrote:
>>>> After change from hadoop20 to hadoop23 the warning dissapeared but I
>>>> got the same exception (Found interface
>>>> org.apache.hadoop.mapreduce.JobContext, but class was expected)
>>>>
>>>> I have tried over a fresh install: hadoop 2.2.0 and pig 0.12.1
>>>> compiled by me, no other product nor configuration, just two
>>>> servers, one master with ResourceManager and NameNode, one slave
>>>> with DataNode and NodeManager.
>>>>
>>>> I can't understand why over this fresh cluster Pig 0.12 fails. Here
>>>> is the new trace:
>>>>
>>>> 2013-10-24 16:10:52,351 [JobControl] ERROR
>>>> org.apache.pig.backend.hadoop23.PigJobControl - Error while trying
>>>> to run jobs.
>>>> java.lang.RuntimeException: java.lang.reflect.InvocationTargetException
>>>>      at org.apache.pig.backend.hadoop23.PigJobControl.submit(PigJobControl.java:130)
>>>>      at org.apache.pig.backend.hadoop23.PigJobControl.run(PigJobControl.java:191)
>>>>      at java.lang.Thread.run(Thread.java:724)
>>>>      at org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher$1.run(MapReduceLauncher.java:257)
>>>> Caused by: java.lang.reflect.InvocationTargetException
>>>>      at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>>      at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>>>      at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>>      at java.lang.reflect.Method.invoke(Method.java:606)
>>>>      at org.apache.pig.backend.hadoop23.PigJobControl.submit(PigJobControl.java:128)
>>>>      ... 3 more
>>>> Caused by: java.lang.IncompatibleClassChangeError: Found interface
>>>> org.apache.hadoop.mapreduce.JobContext, but class was expected
>>>>      at org.apache.cassandra.hadoop.AbstractColumnFamilyInputFormat.getSplits(AbstractColumnFamilyInputFormat.java:113)
>>>>      at org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigInputFormat.getSplits(PigInputFormat.java:274)
>>>>      at org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:491)
>>>>      at org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:508)
>>>>      at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:392)
>>>>      at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1268)
>>>>      at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1265)
>>>>      at java.security.AccessController.doPrivileged(Native Method)
>>>>      at javax.security.auth.Subject.doAs(Subject.java:415)
>>>>      at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1491)
>>>>      at org.apache.hadoop.mapreduce.Job.submit(Job.java:1265)
>>>>      at org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob.submit(ControlledJob.java:335)
>>>>      ... 8 more
>>>>
>>>>
>>>> El 24/10/13 21:33, Prashant Kommireddi escribió:
>>>>> Yes it does. You need to recompile Pig for hadoop 2
>>>>>
>>>>> ant clean jar-withouthadoop -Dhadoopversion=23
>>>>>
>>>>>
>>>>> On Thu, Oct 24, 2013 at 5:37 AM, Claudio Romo Otto <
>>>>> claudio.romo@namictech.com> wrote:
>>>>>
>>>>>> Does Pig support Hadoop 2.2? When I try Pig 0.12 and Hadoop 2.2 I get an
>>>>>> error even with simple operations like
>>>>>>
>>>>>> data = LOAD 'cql://keyspace1/testcf?' USING CqlStorage();
>>>>>> dump data;
>>>>>>
>>>>>> I only got a warning first and then and exception:
>>>>>>
>>>>>> 2013-10-24 09:35:19,300 [main] WARN org.apache.pig.backend.**hadoop20.PigJobControl
>>>>>> - falling back to default JobControl (not using hadoop 0.20 ?)
>>>>>> java.lang.**NoSuchFieldException: runnerState
>>>>>>      at java.lang.Class.**getDeclaredField(Class.java:**1938)
>>>>>>      at org.apache.pig.backend.**hadoop20.PigJobControl.<**
>>>>>> clinit>(PigJobControl.java:51)
>>>>>>      at org.apache.pig.backend.hadoop.**executionengine.shims.**
>>>>>> HadoopShims.newJobControl(**HadoopShims.java:97)
>>>>>>      at org.apache.pig.backend.hadoop.**executionengine.**mapReduceLayer.**
>>>>>> JobControlCompiler.compile(**JobControlCompiler.java:285)
>>>>>>      at org.apache.pig.backend.hadoop.**executionengine.**mapReduceLayer.**
>>>>>> MapReduceLauncher.launchPig(**MapReduceLauncher.java:177)
>>>>>>      at org.apache.pig.PigServer.**launchPlan(PigServer.java:**1264)
>>>>>>      at org.apache.pig.PigServer.**executeCompiledLogicalPlan(**
>>>>>> PigServer.java:1249)
>>>>>>      at org.apache.pig.PigServer.**storeEx(PigServer.java:931)
>>>>>>      at org.apache.pig.PigServer.**store(PigServer.java:898)
>>>>>>      at org.apache.pig.PigServer.**openIterator(PigServer.java:**811)
>>>>>>      at org.apache.pig.tools.grunt.**GruntParser.processDump(**
>>>>>> GruntParser.java:696)
>>>>>>      at org.apache.pig.tools.**pigscript.parser.**PigScriptParser.parse(**
>>>>>> PigScriptParser.java:320)
>>>>>>      at org.apache.pig.tools.grunt.**GruntParser.parseStopOnError(**
>>>>>> GruntParser.java:194)
>>>>>>      at org.apache.pig.tools.grunt.**GruntParser.parseStopOnError(**
>>>>>> GruntParser.java:170)
>>>>>>      at org.apache.pig.tools.grunt.**Grunt.run(Grunt.java:69)
>>>>>>      at org.apache.pig.Main.run(Main.**java:538)
>>>>>>      at org.apache.pig.Main.main(Main.**java:157)
>>>>>>      at sun.reflect.**NativeMethodAccessorImpl.**invoke0(Native Method)
>>>>>>      at sun.reflect.**NativeMethodAccessorImpl.**invoke(**
>>>>>> NativeMethodAccessorImpl.java:**57)
>>>>>>      at sun.reflect.**DelegatingMethodAccessorImpl.**invoke(**
>>>>>> DelegatingMethodAccessorImpl.**java:43)
>>>>>>      at java.lang.reflect.Method.**invoke(Method.java:606)
>>>>>>      at org.apache.hadoop.util.RunJar.**main(RunJar.java:212)
>>>>>>
>>>>>> ------------------------------**--------
>>>>>>
>>>>>> Backend error message during job submission
>>>>>> ------------------------------**-------------
>>>>>> Unexpected System Error Occured: java.lang.**IncompatibleClassChangeError:
>>>>>> Found interface org.apache.hadoop.mapreduce.**JobContext, but class was
>>>>>> expected
>>>>>>          at org.apache.pig.backend.hadoop.**executionengine.**
>>>>>> mapReduceLayer.**PigOutputFormat.**setupUdfEnvAndStores(**
>>>>>> PigOutputFormat.java:225)
>>>>>>          at org.apache.pig.backend.hadoop.**executionengine.**
>>>>>> mapReduceLayer.**PigOutputFormat.**checkOutputSpecs(**
>>>>>> PigOutputFormat.java:186)
>>>>>>          at org.apache.hadoop.mapreduce.**JobSubmitter.checkSpecs(**
>>>>>> JobSubmitter.java:456)
>>>>>>          at org.apache.hadoop.mapreduce.**JobSubmitter.**submitJobInternal(
>>>>>> **JobSubmitter.java:342)
>>>>>>          at org.apache.hadoop.mapreduce.**Job$10.run(Job.java:1268)
>>>>>>          at org.apache.hadoop.mapreduce.**Job$10.run(Job.java:1265)
>>>>>>          at java.security.**AccessController.doPrivileged(**Native Method)
>>>>>>          at javax.security.auth.Subject.**doAs(Subject.java:415)
>>>>>>          at org.apache.hadoop.security.**UserGroupInformation.doAs(**
>>>>>> UserGroupInformation.java:**1491)
>>>>>>          at org.apache.hadoop.mapreduce.**Job.submit(Job.java:1265)
>>>>>>          at org.apache.hadoop.mapreduce.**lib.jobcontrol.ControlledJob.**
>>>>>> submit(ControlledJob.java:335)
>>>>>>          at org.apache.hadoop.mapreduce.**lib.jobcontrol.JobControl.run(**
>>>>>> JobControl.java:240)
>>>>>>          at org.apache.pig.backend.**hadoop20.PigJobControl.run(**
>>>>>> PigJobControl.java:121)
>>>>>>          at java.lang.Thread.run(Thread.**java:724)
>>>>>>          at org.apache.pig.backend.hadoop.**executionengine.**
>>>>>> mapReduceLayer.**MapReduceLauncher$1.run(**MapReduceLauncher.java:257)
>>>>>>
>>>>>> Pig Stack Trace
>>>>>> ---------------
>>>>>> ERROR 1066: Unable to open iterator for alias data
>>>>>>
>>>>>> org.apache.pig.impl.**logicalLayer.**FrontendException: ERROR 1066:
>>>>>> Unable to open iterator for alias data
>>>>>>          at org.apache.pig.PigServer.**openIterator(PigServer.java:**836)
>>>>>>          at org.apache.pig.tools.grunt.**GruntParser.processDump(**
>>>>>> GruntParser.java:696)
>>>>>>          at org.apache.pig.tools.**pigscript.parser.**
>>>>>> PigScriptParser.parse(**PigScriptParser.java:320)
>>>>>>          at org.apache.pig.tools.grunt.**GruntParser.parseStopOnError(**
>>>>>> GruntParser.java:194)
>>>>>>          at org.apache.pig.tools.grunt.**GruntParser.parseStopOnError(**
>>>>>> GruntParser.java:170)
>>>>>>          at org.apache.pig.tools.grunt.**Grunt.run(Grunt.java:69)
>>>>>>          at org.apache.pig.Main.run(Main.**java:538)
>>>>>>          at org.apache.pig.Main.main(Main.**java:157)
>>>>>>          at sun.reflect.**NativeMethodAccessorImpl.**invoke0(Native Method)
>>>>>>          at sun.reflect.**NativeMethodAccessorImpl.**invoke(**
>>>>>> NativeMethodAccessorImpl.java:**57)
>>>>>>          at sun.reflect.**DelegatingMethodAccessorImpl.**invoke(**
>>>>>> DelegatingMethodAccessorImpl.**java:43)
>>>>>>          at java.lang.reflect.Method.**invoke(Method.java:606)
>>>>>>          at org.apache.hadoop.util.RunJar.**main(RunJar.java:212)
>>>>>> Caused by: java.io.IOException: Job terminated with anomalous status FAILED
>>>>>>          at org.apache.pig.PigServer.**openIterator(PigServer.java:**828)
>>>>>>          ... 12 more
>>>>>>
>>>>>>
>>>>>>


Re: Support for Hadoop 2.2

Posted by Jarek Jarcec Cecho <ja...@apache.org>.
Hi Claudio,
it's hard to guess from the limited information. I would suggest to take a look into logs to see what is happening. 

One guess though - You've mentioned that the task was "running" for 30 minutes, but it still seems to be in SCHEDULED time - are your node managers correctly running?

Jarcec

On Fri, Oct 25, 2013 at 04:10:12PM -0300, Claudio Romo Otto wrote:
> You got it!
> 
> The solution was to compile with  -Dhadoopversion=23 option. After
> your message I tried another test removing Cassandra from the chain
> and Pig sent successfully the job to hadoop.
> 
> BUT! the problem changed, now the Map task remains forever stuck on
> Hadoop (30 minutes waiting, no other jobs running):
> 
> Task
> 	
> Progress
> 	
> State
> 	
> Start Time
> 	
> Finish Time
> 	
> Elapsed Time
> task_1382631533263_0012_m_000000 <http://topgps-test-3.dnsalias.com:8088/proxy/application_1382631533263_0012/mapreduce/task/task_1382631533263_0012_m_000000>
> 	
> 	SCHEDULED 	Fri, 25 Oct 2013 18:18:32 GMT 	N/A 	0sec
> 
> 
> 
> Attempt
> 	
> Progress
> 	
> State
> 	
> Node
> 	
> Logs
> 	
> Started
> 	
> Finished
> 	
> Elapsed
> 	
> Note
> attempt_1382631533263_0012_m_000000_0 	0,00 	STARTING 	N/A 	N/A 	N/A
> N/A 	0sec
> 
> 
> Don't know if this is a Hadoop problem or Pig, what do you think?
> 
> 
> El 25/10/13 13:11, Jarek Jarcec Cecho escribió:
> >It seems that Pig was correctly compiled against Hadoop 23, but the Cassandra piece was not, check out the where the exception is coming from:
> >
> >>Caused by: java.lang.IncompatibleClassChangeError: Found interface
> >>org.apache.hadoop.mapreduce.JobContext, but class was expected
> >>     at org.apache.cassandra.hadoop.AbstractColumnFamilyInputFormat.getSplits(AbstractColumnFamilyInputFormat.java:113)
> >So, I would say that you also need to get Hadoop 2 compatible Cassandra connector first.
> >
> >Jarcec
> >
> >On Thu, Oct 24, 2013 at 10:34:49PM -0300, Claudio Romo Otto wrote:
> >>After change from hadoop20 to hadoop23 the warning dissapeared but I
> >>got the same exception (Found interface
> >>org.apache.hadoop.mapreduce.JobContext, but class was expected)
> >>
> >>I have tried over a fresh install: hadoop 2.2.0 and pig 0.12.1
> >>compiled by me, no other product nor configuration, just two
> >>servers, one master with ResourceManager and NameNode, one slave
> >>with DataNode and NodeManager.
> >>
> >>I can't understand why over this fresh cluster Pig 0.12 fails. Here
> >>is the new trace:
> >>
> >>2013-10-24 16:10:52,351 [JobControl] ERROR
> >>org.apache.pig.backend.hadoop23.PigJobControl - Error while trying
> >>to run jobs.
> >>java.lang.RuntimeException: java.lang.reflect.InvocationTargetException
> >>     at org.apache.pig.backend.hadoop23.PigJobControl.submit(PigJobControl.java:130)
> >>     at org.apache.pig.backend.hadoop23.PigJobControl.run(PigJobControl.java:191)
> >>     at java.lang.Thread.run(Thread.java:724)
> >>     at org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher$1.run(MapReduceLauncher.java:257)
> >>Caused by: java.lang.reflect.InvocationTargetException
> >>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> >>     at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
> >>     at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> >>     at java.lang.reflect.Method.invoke(Method.java:606)
> >>     at org.apache.pig.backend.hadoop23.PigJobControl.submit(PigJobControl.java:128)
> >>     ... 3 more
> >>Caused by: java.lang.IncompatibleClassChangeError: Found interface
> >>org.apache.hadoop.mapreduce.JobContext, but class was expected
> >>     at org.apache.cassandra.hadoop.AbstractColumnFamilyInputFormat.getSplits(AbstractColumnFamilyInputFormat.java:113)
> >>     at org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigInputFormat.getSplits(PigInputFormat.java:274)
> >>     at org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:491)
> >>     at org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:508)
> >>     at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:392)
> >>     at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1268)
> >>     at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1265)
> >>     at java.security.AccessController.doPrivileged(Native Method)
> >>     at javax.security.auth.Subject.doAs(Subject.java:415)
> >>     at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1491)
> >>     at org.apache.hadoop.mapreduce.Job.submit(Job.java:1265)
> >>     at org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob.submit(ControlledJob.java:335)
> >>     ... 8 more
> >>
> >>
> >>El 24/10/13 21:33, Prashant Kommireddi escribió:
> >>>Yes it does. You need to recompile Pig for hadoop 2
> >>>
> >>>ant clean jar-withouthadoop -Dhadoopversion=23
> >>>
> >>>
> >>>On Thu, Oct 24, 2013 at 5:37 AM, Claudio Romo Otto <
> >>>claudio.romo@namictech.com> wrote:
> >>>
> >>>>Does Pig support Hadoop 2.2? When I try Pig 0.12 and Hadoop 2.2 I get an
> >>>>error even with simple operations like
> >>>>
> >>>>data = LOAD 'cql://keyspace1/testcf?' USING CqlStorage();
> >>>>dump data;
> >>>>
> >>>>I only got a warning first and then and exception:
> >>>>
> >>>>2013-10-24 09:35:19,300 [main] WARN org.apache.pig.backend.**hadoop20.PigJobControl
> >>>>- falling back to default JobControl (not using hadoop 0.20 ?)
> >>>>java.lang.**NoSuchFieldException: runnerState
> >>>>     at java.lang.Class.**getDeclaredField(Class.java:**1938)
> >>>>     at org.apache.pig.backend.**hadoop20.PigJobControl.<**
> >>>>clinit>(PigJobControl.java:51)
> >>>>     at org.apache.pig.backend.hadoop.**executionengine.shims.**
> >>>>HadoopShims.newJobControl(**HadoopShims.java:97)
> >>>>     at org.apache.pig.backend.hadoop.**executionengine.**mapReduceLayer.**
> >>>>JobControlCompiler.compile(**JobControlCompiler.java:285)
> >>>>     at org.apache.pig.backend.hadoop.**executionengine.**mapReduceLayer.**
> >>>>MapReduceLauncher.launchPig(**MapReduceLauncher.java:177)
> >>>>     at org.apache.pig.PigServer.**launchPlan(PigServer.java:**1264)
> >>>>     at org.apache.pig.PigServer.**executeCompiledLogicalPlan(**
> >>>>PigServer.java:1249)
> >>>>     at org.apache.pig.PigServer.**storeEx(PigServer.java:931)
> >>>>     at org.apache.pig.PigServer.**store(PigServer.java:898)
> >>>>     at org.apache.pig.PigServer.**openIterator(PigServer.java:**811)
> >>>>     at org.apache.pig.tools.grunt.**GruntParser.processDump(**
> >>>>GruntParser.java:696)
> >>>>     at org.apache.pig.tools.**pigscript.parser.**PigScriptParser.parse(**
> >>>>PigScriptParser.java:320)
> >>>>     at org.apache.pig.tools.grunt.**GruntParser.parseStopOnError(**
> >>>>GruntParser.java:194)
> >>>>     at org.apache.pig.tools.grunt.**GruntParser.parseStopOnError(**
> >>>>GruntParser.java:170)
> >>>>     at org.apache.pig.tools.grunt.**Grunt.run(Grunt.java:69)
> >>>>     at org.apache.pig.Main.run(Main.**java:538)
> >>>>     at org.apache.pig.Main.main(Main.**java:157)
> >>>>     at sun.reflect.**NativeMethodAccessorImpl.**invoke0(Native Method)
> >>>>     at sun.reflect.**NativeMethodAccessorImpl.**invoke(**
> >>>>NativeMethodAccessorImpl.java:**57)
> >>>>     at sun.reflect.**DelegatingMethodAccessorImpl.**invoke(**
> >>>>DelegatingMethodAccessorImpl.**java:43)
> >>>>     at java.lang.reflect.Method.**invoke(Method.java:606)
> >>>>     at org.apache.hadoop.util.RunJar.**main(RunJar.java:212)
> >>>>
> >>>>------------------------------**--------
> >>>>
> >>>>Backend error message during job submission
> >>>>------------------------------**-------------
> >>>>Unexpected System Error Occured: java.lang.**IncompatibleClassChangeError:
> >>>>Found interface org.apache.hadoop.mapreduce.**JobContext, but class was
> >>>>expected
> >>>>         at org.apache.pig.backend.hadoop.**executionengine.**
> >>>>mapReduceLayer.**PigOutputFormat.**setupUdfEnvAndStores(**
> >>>>PigOutputFormat.java:225)
> >>>>         at org.apache.pig.backend.hadoop.**executionengine.**
> >>>>mapReduceLayer.**PigOutputFormat.**checkOutputSpecs(**
> >>>>PigOutputFormat.java:186)
> >>>>         at org.apache.hadoop.mapreduce.**JobSubmitter.checkSpecs(**
> >>>>JobSubmitter.java:456)
> >>>>         at org.apache.hadoop.mapreduce.**JobSubmitter.**submitJobInternal(
> >>>>**JobSubmitter.java:342)
> >>>>         at org.apache.hadoop.mapreduce.**Job$10.run(Job.java:1268)
> >>>>         at org.apache.hadoop.mapreduce.**Job$10.run(Job.java:1265)
> >>>>         at java.security.**AccessController.doPrivileged(**Native Method)
> >>>>         at javax.security.auth.Subject.**doAs(Subject.java:415)
> >>>>         at org.apache.hadoop.security.**UserGroupInformation.doAs(**
> >>>>UserGroupInformation.java:**1491)
> >>>>         at org.apache.hadoop.mapreduce.**Job.submit(Job.java:1265)
> >>>>         at org.apache.hadoop.mapreduce.**lib.jobcontrol.ControlledJob.**
> >>>>submit(ControlledJob.java:335)
> >>>>         at org.apache.hadoop.mapreduce.**lib.jobcontrol.JobControl.run(**
> >>>>JobControl.java:240)
> >>>>         at org.apache.pig.backend.**hadoop20.PigJobControl.run(**
> >>>>PigJobControl.java:121)
> >>>>         at java.lang.Thread.run(Thread.**java:724)
> >>>>         at org.apache.pig.backend.hadoop.**executionengine.**
> >>>>mapReduceLayer.**MapReduceLauncher$1.run(**MapReduceLauncher.java:257)
> >>>>
> >>>>Pig Stack Trace
> >>>>---------------
> >>>>ERROR 1066: Unable to open iterator for alias data
> >>>>
> >>>>org.apache.pig.impl.**logicalLayer.**FrontendException: ERROR 1066:
> >>>>Unable to open iterator for alias data
> >>>>         at org.apache.pig.PigServer.**openIterator(PigServer.java:**836)
> >>>>         at org.apache.pig.tools.grunt.**GruntParser.processDump(**
> >>>>GruntParser.java:696)
> >>>>         at org.apache.pig.tools.**pigscript.parser.**
> >>>>PigScriptParser.parse(**PigScriptParser.java:320)
> >>>>         at org.apache.pig.tools.grunt.**GruntParser.parseStopOnError(**
> >>>>GruntParser.java:194)
> >>>>         at org.apache.pig.tools.grunt.**GruntParser.parseStopOnError(**
> >>>>GruntParser.java:170)
> >>>>         at org.apache.pig.tools.grunt.**Grunt.run(Grunt.java:69)
> >>>>         at org.apache.pig.Main.run(Main.**java:538)
> >>>>         at org.apache.pig.Main.main(Main.**java:157)
> >>>>         at sun.reflect.**NativeMethodAccessorImpl.**invoke0(Native Method)
> >>>>         at sun.reflect.**NativeMethodAccessorImpl.**invoke(**
> >>>>NativeMethodAccessorImpl.java:**57)
> >>>>         at sun.reflect.**DelegatingMethodAccessorImpl.**invoke(**
> >>>>DelegatingMethodAccessorImpl.**java:43)
> >>>>         at java.lang.reflect.Method.**invoke(Method.java:606)
> >>>>         at org.apache.hadoop.util.RunJar.**main(RunJar.java:212)
> >>>>Caused by: java.io.IOException: Job terminated with anomalous status FAILED
> >>>>         at org.apache.pig.PigServer.**openIterator(PigServer.java:**828)
> >>>>         ... 12 more
> >>>>
> >>>>
> >>>>
> 

Re: Support for Hadoop 2.2

Posted by Claudio Romo Otto <cl...@namictech.com>.
You got it!

The solution was to compile with  -Dhadoopversion=23 option. After your 
message I tried another test removing Cassandra from the chain and Pig 
sent successfully the job to hadoop.

BUT! the problem changed, now the Map task remains forever stuck on 
Hadoop (30 minutes waiting, no other jobs running):

Task
	
Progress
	
State
	
Start Time
	
Finish Time
	
Elapsed Time
task_1382631533263_0012_m_000000 
<http://topgps-test-3.dnsalias.com:8088/proxy/application_1382631533263_0012/mapreduce/task/task_1382631533263_0012_m_000000> 
	
	SCHEDULED 	Fri, 25 Oct 2013 18:18:32 GMT 	N/A 	0sec



Attempt
	
Progress
	
State
	
Node
	
Logs
	
Started
	
Finished
	
Elapsed
	
Note
attempt_1382631533263_0012_m_000000_0 	0,00 	STARTING 	N/A 	N/A 	N/A 
N/A 	0sec


Don't know if this is a Hadoop problem or Pig, what do you think?


El 25/10/13 13:11, Jarek Jarcec Cecho escribió:
> It seems that Pig was correctly compiled against Hadoop 23, but the Cassandra piece was not, check out the where the exception is coming from:
>
>> Caused by: java.lang.IncompatibleClassChangeError: Found interface
>> org.apache.hadoop.mapreduce.JobContext, but class was expected
>>      at org.apache.cassandra.hadoop.AbstractColumnFamilyInputFormat.getSplits(AbstractColumnFamilyInputFormat.java:113)
> So, I would say that you also need to get Hadoop 2 compatible Cassandra connector first.
>
> Jarcec
>
> On Thu, Oct 24, 2013 at 10:34:49PM -0300, Claudio Romo Otto wrote:
>> After change from hadoop20 to hadoop23 the warning dissapeared but I
>> got the same exception (Found interface
>> org.apache.hadoop.mapreduce.JobContext, but class was expected)
>>
>> I have tried over a fresh install: hadoop 2.2.0 and pig 0.12.1
>> compiled by me, no other product nor configuration, just two
>> servers, one master with ResourceManager and NameNode, one slave
>> with DataNode and NodeManager.
>>
>> I can't understand why over this fresh cluster Pig 0.12 fails. Here
>> is the new trace:
>>
>> 2013-10-24 16:10:52,351 [JobControl] ERROR
>> org.apache.pig.backend.hadoop23.PigJobControl - Error while trying
>> to run jobs.
>> java.lang.RuntimeException: java.lang.reflect.InvocationTargetException
>>      at org.apache.pig.backend.hadoop23.PigJobControl.submit(PigJobControl.java:130)
>>      at org.apache.pig.backend.hadoop23.PigJobControl.run(PigJobControl.java:191)
>>      at java.lang.Thread.run(Thread.java:724)
>>      at org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher$1.run(MapReduceLauncher.java:257)
>> Caused by: java.lang.reflect.InvocationTargetException
>>      at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>      at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>      at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>      at java.lang.reflect.Method.invoke(Method.java:606)
>>      at org.apache.pig.backend.hadoop23.PigJobControl.submit(PigJobControl.java:128)
>>      ... 3 more
>> Caused by: java.lang.IncompatibleClassChangeError: Found interface
>> org.apache.hadoop.mapreduce.JobContext, but class was expected
>>      at org.apache.cassandra.hadoop.AbstractColumnFamilyInputFormat.getSplits(AbstractColumnFamilyInputFormat.java:113)
>>      at org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigInputFormat.getSplits(PigInputFormat.java:274)
>>      at org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:491)
>>      at org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:508)
>>      at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:392)
>>      at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1268)
>>      at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1265)
>>      at java.security.AccessController.doPrivileged(Native Method)
>>      at javax.security.auth.Subject.doAs(Subject.java:415)
>>      at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1491)
>>      at org.apache.hadoop.mapreduce.Job.submit(Job.java:1265)
>>      at org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob.submit(ControlledJob.java:335)
>>      ... 8 more
>>
>>
>> El 24/10/13 21:33, Prashant Kommireddi escribió:
>>> Yes it does. You need to recompile Pig for hadoop 2
>>>
>>> ant clean jar-withouthadoop -Dhadoopversion=23
>>>
>>>
>>> On Thu, Oct 24, 2013 at 5:37 AM, Claudio Romo Otto <
>>> claudio.romo@namictech.com> wrote:
>>>
>>>> Does Pig support Hadoop 2.2? When I try Pig 0.12 and Hadoop 2.2 I get an
>>>> error even with simple operations like
>>>>
>>>> data = LOAD 'cql://keyspace1/testcf?' USING CqlStorage();
>>>> dump data;
>>>>
>>>> I only got a warning first and then and exception:
>>>>
>>>> 2013-10-24 09:35:19,300 [main] WARN org.apache.pig.backend.**hadoop20.PigJobControl
>>>> - falling back to default JobControl (not using hadoop 0.20 ?)
>>>> java.lang.**NoSuchFieldException: runnerState
>>>>      at java.lang.Class.**getDeclaredField(Class.java:**1938)
>>>>      at org.apache.pig.backend.**hadoop20.PigJobControl.<**
>>>> clinit>(PigJobControl.java:51)
>>>>      at org.apache.pig.backend.hadoop.**executionengine.shims.**
>>>> HadoopShims.newJobControl(**HadoopShims.java:97)
>>>>      at org.apache.pig.backend.hadoop.**executionengine.**mapReduceLayer.**
>>>> JobControlCompiler.compile(**JobControlCompiler.java:285)
>>>>      at org.apache.pig.backend.hadoop.**executionengine.**mapReduceLayer.**
>>>> MapReduceLauncher.launchPig(**MapReduceLauncher.java:177)
>>>>      at org.apache.pig.PigServer.**launchPlan(PigServer.java:**1264)
>>>>      at org.apache.pig.PigServer.**executeCompiledLogicalPlan(**
>>>> PigServer.java:1249)
>>>>      at org.apache.pig.PigServer.**storeEx(PigServer.java:931)
>>>>      at org.apache.pig.PigServer.**store(PigServer.java:898)
>>>>      at org.apache.pig.PigServer.**openIterator(PigServer.java:**811)
>>>>      at org.apache.pig.tools.grunt.**GruntParser.processDump(**
>>>> GruntParser.java:696)
>>>>      at org.apache.pig.tools.**pigscript.parser.**PigScriptParser.parse(**
>>>> PigScriptParser.java:320)
>>>>      at org.apache.pig.tools.grunt.**GruntParser.parseStopOnError(**
>>>> GruntParser.java:194)
>>>>      at org.apache.pig.tools.grunt.**GruntParser.parseStopOnError(**
>>>> GruntParser.java:170)
>>>>      at org.apache.pig.tools.grunt.**Grunt.run(Grunt.java:69)
>>>>      at org.apache.pig.Main.run(Main.**java:538)
>>>>      at org.apache.pig.Main.main(Main.**java:157)
>>>>      at sun.reflect.**NativeMethodAccessorImpl.**invoke0(Native Method)
>>>>      at sun.reflect.**NativeMethodAccessorImpl.**invoke(**
>>>> NativeMethodAccessorImpl.java:**57)
>>>>      at sun.reflect.**DelegatingMethodAccessorImpl.**invoke(**
>>>> DelegatingMethodAccessorImpl.**java:43)
>>>>      at java.lang.reflect.Method.**invoke(Method.java:606)
>>>>      at org.apache.hadoop.util.RunJar.**main(RunJar.java:212)
>>>>
>>>> ------------------------------**--------
>>>>
>>>> Backend error message during job submission
>>>> ------------------------------**-------------
>>>> Unexpected System Error Occured: java.lang.**IncompatibleClassChangeError:
>>>> Found interface org.apache.hadoop.mapreduce.**JobContext, but class was
>>>> expected
>>>>          at org.apache.pig.backend.hadoop.**executionengine.**
>>>> mapReduceLayer.**PigOutputFormat.**setupUdfEnvAndStores(**
>>>> PigOutputFormat.java:225)
>>>>          at org.apache.pig.backend.hadoop.**executionengine.**
>>>> mapReduceLayer.**PigOutputFormat.**checkOutputSpecs(**
>>>> PigOutputFormat.java:186)
>>>>          at org.apache.hadoop.mapreduce.**JobSubmitter.checkSpecs(**
>>>> JobSubmitter.java:456)
>>>>          at org.apache.hadoop.mapreduce.**JobSubmitter.**submitJobInternal(
>>>> **JobSubmitter.java:342)
>>>>          at org.apache.hadoop.mapreduce.**Job$10.run(Job.java:1268)
>>>>          at org.apache.hadoop.mapreduce.**Job$10.run(Job.java:1265)
>>>>          at java.security.**AccessController.doPrivileged(**Native Method)
>>>>          at javax.security.auth.Subject.**doAs(Subject.java:415)
>>>>          at org.apache.hadoop.security.**UserGroupInformation.doAs(**
>>>> UserGroupInformation.java:**1491)
>>>>          at org.apache.hadoop.mapreduce.**Job.submit(Job.java:1265)
>>>>          at org.apache.hadoop.mapreduce.**lib.jobcontrol.ControlledJob.**
>>>> submit(ControlledJob.java:335)
>>>>          at org.apache.hadoop.mapreduce.**lib.jobcontrol.JobControl.run(**
>>>> JobControl.java:240)
>>>>          at org.apache.pig.backend.**hadoop20.PigJobControl.run(**
>>>> PigJobControl.java:121)
>>>>          at java.lang.Thread.run(Thread.**java:724)
>>>>          at org.apache.pig.backend.hadoop.**executionengine.**
>>>> mapReduceLayer.**MapReduceLauncher$1.run(**MapReduceLauncher.java:257)
>>>>
>>>> Pig Stack Trace
>>>> ---------------
>>>> ERROR 1066: Unable to open iterator for alias data
>>>>
>>>> org.apache.pig.impl.**logicalLayer.**FrontendException: ERROR 1066:
>>>> Unable to open iterator for alias data
>>>>          at org.apache.pig.PigServer.**openIterator(PigServer.java:**836)
>>>>          at org.apache.pig.tools.grunt.**GruntParser.processDump(**
>>>> GruntParser.java:696)
>>>>          at org.apache.pig.tools.**pigscript.parser.**
>>>> PigScriptParser.parse(**PigScriptParser.java:320)
>>>>          at org.apache.pig.tools.grunt.**GruntParser.parseStopOnError(**
>>>> GruntParser.java:194)
>>>>          at org.apache.pig.tools.grunt.**GruntParser.parseStopOnError(**
>>>> GruntParser.java:170)
>>>>          at org.apache.pig.tools.grunt.**Grunt.run(Grunt.java:69)
>>>>          at org.apache.pig.Main.run(Main.**java:538)
>>>>          at org.apache.pig.Main.main(Main.**java:157)
>>>>          at sun.reflect.**NativeMethodAccessorImpl.**invoke0(Native Method)
>>>>          at sun.reflect.**NativeMethodAccessorImpl.**invoke(**
>>>> NativeMethodAccessorImpl.java:**57)
>>>>          at sun.reflect.**DelegatingMethodAccessorImpl.**invoke(**
>>>> DelegatingMethodAccessorImpl.**java:43)
>>>>          at java.lang.reflect.Method.**invoke(Method.java:606)
>>>>          at org.apache.hadoop.util.RunJar.**main(RunJar.java:212)
>>>> Caused by: java.io.IOException: Job terminated with anomalous status FAILED
>>>>          at org.apache.pig.PigServer.**openIterator(PigServer.java:**828)
>>>>          ... 12 more
>>>>
>>>>
>>>>


Re: Support for Hadoop 2.2

Posted by Jarek Jarcec Cecho <ja...@apache.org>.
It seems that Pig was correctly compiled against Hadoop 23, but the Cassandra piece was not, check out the where the exception is coming from:

> Caused by: java.lang.IncompatibleClassChangeError: Found interface
> org.apache.hadoop.mapreduce.JobContext, but class was expected
>     at org.apache.cassandra.hadoop.AbstractColumnFamilyInputFormat.getSplits(AbstractColumnFamilyInputFormat.java:113)

So, I would say that you also need to get Hadoop 2 compatible Cassandra connector first.

Jarcec

On Thu, Oct 24, 2013 at 10:34:49PM -0300, Claudio Romo Otto wrote:
> After change from hadoop20 to hadoop23 the warning dissapeared but I
> got the same exception (Found interface
> org.apache.hadoop.mapreduce.JobContext, but class was expected)
> 
> I have tried over a fresh install: hadoop 2.2.0 and pig 0.12.1
> compiled by me, no other product nor configuration, just two
> servers, one master with ResourceManager and NameNode, one slave
> with DataNode and NodeManager.
> 
> I can't understand why over this fresh cluster Pig 0.12 fails. Here
> is the new trace:
> 
> 2013-10-24 16:10:52,351 [JobControl] ERROR
> org.apache.pig.backend.hadoop23.PigJobControl - Error while trying
> to run jobs.
> java.lang.RuntimeException: java.lang.reflect.InvocationTargetException
>     at org.apache.pig.backend.hadoop23.PigJobControl.submit(PigJobControl.java:130)
>     at org.apache.pig.backend.hadoop23.PigJobControl.run(PigJobControl.java:191)
>     at java.lang.Thread.run(Thread.java:724)
>     at org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher$1.run(MapReduceLauncher.java:257)
> Caused by: java.lang.reflect.InvocationTargetException
>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>     at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>     at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>     at java.lang.reflect.Method.invoke(Method.java:606)
>     at org.apache.pig.backend.hadoop23.PigJobControl.submit(PigJobControl.java:128)
>     ... 3 more
> Caused by: java.lang.IncompatibleClassChangeError: Found interface
> org.apache.hadoop.mapreduce.JobContext, but class was expected
>     at org.apache.cassandra.hadoop.AbstractColumnFamilyInputFormat.getSplits(AbstractColumnFamilyInputFormat.java:113)
>     at org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigInputFormat.getSplits(PigInputFormat.java:274)
>     at org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:491)
>     at org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:508)
>     at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:392)
>     at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1268)
>     at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1265)
>     at java.security.AccessController.doPrivileged(Native Method)
>     at javax.security.auth.Subject.doAs(Subject.java:415)
>     at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1491)
>     at org.apache.hadoop.mapreduce.Job.submit(Job.java:1265)
>     at org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob.submit(ControlledJob.java:335)
>     ... 8 more
> 
> 
> El 24/10/13 21:33, Prashant Kommireddi escribió:
> >Yes it does. You need to recompile Pig for hadoop 2
> >
> >ant clean jar-withouthadoop -Dhadoopversion=23
> >
> >
> >On Thu, Oct 24, 2013 at 5:37 AM, Claudio Romo Otto <
> >claudio.romo@namictech.com> wrote:
> >
> >>Does Pig support Hadoop 2.2? When I try Pig 0.12 and Hadoop 2.2 I get an
> >>error even with simple operations like
> >>
> >>data = LOAD 'cql://keyspace1/testcf?' USING CqlStorage();
> >>dump data;
> >>
> >>I only got a warning first and then and exception:
> >>
> >>2013-10-24 09:35:19,300 [main] WARN org.apache.pig.backend.**hadoop20.PigJobControl
> >>- falling back to default JobControl (not using hadoop 0.20 ?)
> >>java.lang.**NoSuchFieldException: runnerState
> >>     at java.lang.Class.**getDeclaredField(Class.java:**1938)
> >>     at org.apache.pig.backend.**hadoop20.PigJobControl.<**
> >>clinit>(PigJobControl.java:51)
> >>     at org.apache.pig.backend.hadoop.**executionengine.shims.**
> >>HadoopShims.newJobControl(**HadoopShims.java:97)
> >>     at org.apache.pig.backend.hadoop.**executionengine.**mapReduceLayer.**
> >>JobControlCompiler.compile(**JobControlCompiler.java:285)
> >>     at org.apache.pig.backend.hadoop.**executionengine.**mapReduceLayer.**
> >>MapReduceLauncher.launchPig(**MapReduceLauncher.java:177)
> >>     at org.apache.pig.PigServer.**launchPlan(PigServer.java:**1264)
> >>     at org.apache.pig.PigServer.**executeCompiledLogicalPlan(**
> >>PigServer.java:1249)
> >>     at org.apache.pig.PigServer.**storeEx(PigServer.java:931)
> >>     at org.apache.pig.PigServer.**store(PigServer.java:898)
> >>     at org.apache.pig.PigServer.**openIterator(PigServer.java:**811)
> >>     at org.apache.pig.tools.grunt.**GruntParser.processDump(**
> >>GruntParser.java:696)
> >>     at org.apache.pig.tools.**pigscript.parser.**PigScriptParser.parse(**
> >>PigScriptParser.java:320)
> >>     at org.apache.pig.tools.grunt.**GruntParser.parseStopOnError(**
> >>GruntParser.java:194)
> >>     at org.apache.pig.tools.grunt.**GruntParser.parseStopOnError(**
> >>GruntParser.java:170)
> >>     at org.apache.pig.tools.grunt.**Grunt.run(Grunt.java:69)
> >>     at org.apache.pig.Main.run(Main.**java:538)
> >>     at org.apache.pig.Main.main(Main.**java:157)
> >>     at sun.reflect.**NativeMethodAccessorImpl.**invoke0(Native Method)
> >>     at sun.reflect.**NativeMethodAccessorImpl.**invoke(**
> >>NativeMethodAccessorImpl.java:**57)
> >>     at sun.reflect.**DelegatingMethodAccessorImpl.**invoke(**
> >>DelegatingMethodAccessorImpl.**java:43)
> >>     at java.lang.reflect.Method.**invoke(Method.java:606)
> >>     at org.apache.hadoop.util.RunJar.**main(RunJar.java:212)
> >>
> >>------------------------------**--------
> >>
> >>Backend error message during job submission
> >>------------------------------**-------------
> >>Unexpected System Error Occured: java.lang.**IncompatibleClassChangeError:
> >>Found interface org.apache.hadoop.mapreduce.**JobContext, but class was
> >>expected
> >>         at org.apache.pig.backend.hadoop.**executionengine.**
> >>mapReduceLayer.**PigOutputFormat.**setupUdfEnvAndStores(**
> >>PigOutputFormat.java:225)
> >>         at org.apache.pig.backend.hadoop.**executionengine.**
> >>mapReduceLayer.**PigOutputFormat.**checkOutputSpecs(**
> >>PigOutputFormat.java:186)
> >>         at org.apache.hadoop.mapreduce.**JobSubmitter.checkSpecs(**
> >>JobSubmitter.java:456)
> >>         at org.apache.hadoop.mapreduce.**JobSubmitter.**submitJobInternal(
> >>**JobSubmitter.java:342)
> >>         at org.apache.hadoop.mapreduce.**Job$10.run(Job.java:1268)
> >>         at org.apache.hadoop.mapreduce.**Job$10.run(Job.java:1265)
> >>         at java.security.**AccessController.doPrivileged(**Native Method)
> >>         at javax.security.auth.Subject.**doAs(Subject.java:415)
> >>         at org.apache.hadoop.security.**UserGroupInformation.doAs(**
> >>UserGroupInformation.java:**1491)
> >>         at org.apache.hadoop.mapreduce.**Job.submit(Job.java:1265)
> >>         at org.apache.hadoop.mapreduce.**lib.jobcontrol.ControlledJob.**
> >>submit(ControlledJob.java:335)
> >>         at org.apache.hadoop.mapreduce.**lib.jobcontrol.JobControl.run(**
> >>JobControl.java:240)
> >>         at org.apache.pig.backend.**hadoop20.PigJobControl.run(**
> >>PigJobControl.java:121)
> >>         at java.lang.Thread.run(Thread.**java:724)
> >>         at org.apache.pig.backend.hadoop.**executionengine.**
> >>mapReduceLayer.**MapReduceLauncher$1.run(**MapReduceLauncher.java:257)
> >>
> >>Pig Stack Trace
> >>---------------
> >>ERROR 1066: Unable to open iterator for alias data
> >>
> >>org.apache.pig.impl.**logicalLayer.**FrontendException: ERROR 1066:
> >>Unable to open iterator for alias data
> >>         at org.apache.pig.PigServer.**openIterator(PigServer.java:**836)
> >>         at org.apache.pig.tools.grunt.**GruntParser.processDump(**
> >>GruntParser.java:696)
> >>         at org.apache.pig.tools.**pigscript.parser.**
> >>PigScriptParser.parse(**PigScriptParser.java:320)
> >>         at org.apache.pig.tools.grunt.**GruntParser.parseStopOnError(**
> >>GruntParser.java:194)
> >>         at org.apache.pig.tools.grunt.**GruntParser.parseStopOnError(**
> >>GruntParser.java:170)
> >>         at org.apache.pig.tools.grunt.**Grunt.run(Grunt.java:69)
> >>         at org.apache.pig.Main.run(Main.**java:538)
> >>         at org.apache.pig.Main.main(Main.**java:157)
> >>         at sun.reflect.**NativeMethodAccessorImpl.**invoke0(Native Method)
> >>         at sun.reflect.**NativeMethodAccessorImpl.**invoke(**
> >>NativeMethodAccessorImpl.java:**57)
> >>         at sun.reflect.**DelegatingMethodAccessorImpl.**invoke(**
> >>DelegatingMethodAccessorImpl.**java:43)
> >>         at java.lang.reflect.Method.**invoke(Method.java:606)
> >>         at org.apache.hadoop.util.RunJar.**main(RunJar.java:212)
> >>Caused by: java.io.IOException: Job terminated with anomalous status FAILED
> >>         at org.apache.pig.PigServer.**openIterator(PigServer.java:**828)
> >>         ... 12 more
> >>
> >>
> >>
> 

Re: Support for Hadoop 2.2

Posted by Claudio Romo Otto <cl...@namictech.com>.
After change from hadoop20 to hadoop23 the warning dissapeared but I got 
the same exception (Found interface 
org.apache.hadoop.mapreduce.JobContext, but class was expected)

I have tried over a fresh install: hadoop 2.2.0 and pig 0.12.1 compiled 
by me, no other product nor configuration, just two servers, one master 
with ResourceManager and NameNode, one slave with DataNode and NodeManager.

I can't understand why over this fresh cluster Pig 0.12 fails. Here is 
the new trace:

2013-10-24 16:10:52,351 [JobControl] ERROR 
org.apache.pig.backend.hadoop23.PigJobControl - Error while trying to 
run jobs.
java.lang.RuntimeException: java.lang.reflect.InvocationTargetException
     at 
org.apache.pig.backend.hadoop23.PigJobControl.submit(PigJobControl.java:130)
     at 
org.apache.pig.backend.hadoop23.PigJobControl.run(PigJobControl.java:191)
     at java.lang.Thread.run(Thread.java:724)
     at 
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher$1.run(MapReduceLauncher.java:257)
Caused by: java.lang.reflect.InvocationTargetException
     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
     at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
     at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
     at java.lang.reflect.Method.invoke(Method.java:606)
     at 
org.apache.pig.backend.hadoop23.PigJobControl.submit(PigJobControl.java:128)
     ... 3 more
Caused by: java.lang.IncompatibleClassChangeError: Found interface 
org.apache.hadoop.mapreduce.JobContext, but class was expected
     at 
org.apache.cassandra.hadoop.AbstractColumnFamilyInputFormat.getSplits(AbstractColumnFamilyInputFormat.java:113)
     at 
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigInputFormat.getSplits(PigInputFormat.java:274)
     at 
org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:491)
     at 
org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:508)
     at 
org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:392)
     at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1268)
     at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1265)
     at java.security.AccessController.doPrivileged(Native Method)
     at javax.security.auth.Subject.doAs(Subject.java:415)
     at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1491)
     at org.apache.hadoop.mapreduce.Job.submit(Job.java:1265)
     at 
org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob.submit(ControlledJob.java:335)
     ... 8 more


El 24/10/13 21:33, Prashant Kommireddi escribió:
> Yes it does. You need to recompile Pig for hadoop 2
>
> ant clean jar-withouthadoop -Dhadoopversion=23
>
>
> On Thu, Oct 24, 2013 at 5:37 AM, Claudio Romo Otto <
> claudio.romo@namictech.com> wrote:
>
>> Does Pig support Hadoop 2.2? When I try Pig 0.12 and Hadoop 2.2 I get an
>> error even with simple operations like
>>
>> data = LOAD 'cql://keyspace1/testcf?' USING CqlStorage();
>> dump data;
>>
>> I only got a warning first and then and exception:
>>
>> 2013-10-24 09:35:19,300 [main] WARN org.apache.pig.backend.**hadoop20.PigJobControl
>> - falling back to default JobControl (not using hadoop 0.20 ?)
>> java.lang.**NoSuchFieldException: runnerState
>>      at java.lang.Class.**getDeclaredField(Class.java:**1938)
>>      at org.apache.pig.backend.**hadoop20.PigJobControl.<**
>> clinit>(PigJobControl.java:51)
>>      at org.apache.pig.backend.hadoop.**executionengine.shims.**
>> HadoopShims.newJobControl(**HadoopShims.java:97)
>>      at org.apache.pig.backend.hadoop.**executionengine.**mapReduceLayer.**
>> JobControlCompiler.compile(**JobControlCompiler.java:285)
>>      at org.apache.pig.backend.hadoop.**executionengine.**mapReduceLayer.**
>> MapReduceLauncher.launchPig(**MapReduceLauncher.java:177)
>>      at org.apache.pig.PigServer.**launchPlan(PigServer.java:**1264)
>>      at org.apache.pig.PigServer.**executeCompiledLogicalPlan(**
>> PigServer.java:1249)
>>      at org.apache.pig.PigServer.**storeEx(PigServer.java:931)
>>      at org.apache.pig.PigServer.**store(PigServer.java:898)
>>      at org.apache.pig.PigServer.**openIterator(PigServer.java:**811)
>>      at org.apache.pig.tools.grunt.**GruntParser.processDump(**
>> GruntParser.java:696)
>>      at org.apache.pig.tools.**pigscript.parser.**PigScriptParser.parse(**
>> PigScriptParser.java:320)
>>      at org.apache.pig.tools.grunt.**GruntParser.parseStopOnError(**
>> GruntParser.java:194)
>>      at org.apache.pig.tools.grunt.**GruntParser.parseStopOnError(**
>> GruntParser.java:170)
>>      at org.apache.pig.tools.grunt.**Grunt.run(Grunt.java:69)
>>      at org.apache.pig.Main.run(Main.**java:538)
>>      at org.apache.pig.Main.main(Main.**java:157)
>>      at sun.reflect.**NativeMethodAccessorImpl.**invoke0(Native Method)
>>      at sun.reflect.**NativeMethodAccessorImpl.**invoke(**
>> NativeMethodAccessorImpl.java:**57)
>>      at sun.reflect.**DelegatingMethodAccessorImpl.**invoke(**
>> DelegatingMethodAccessorImpl.**java:43)
>>      at java.lang.reflect.Method.**invoke(Method.java:606)
>>      at org.apache.hadoop.util.RunJar.**main(RunJar.java:212)
>>
>> ------------------------------**--------
>>
>> Backend error message during job submission
>> ------------------------------**-------------
>> Unexpected System Error Occured: java.lang.**IncompatibleClassChangeError:
>> Found interface org.apache.hadoop.mapreduce.**JobContext, but class was
>> expected
>>          at org.apache.pig.backend.hadoop.**executionengine.**
>> mapReduceLayer.**PigOutputFormat.**setupUdfEnvAndStores(**
>> PigOutputFormat.java:225)
>>          at org.apache.pig.backend.hadoop.**executionengine.**
>> mapReduceLayer.**PigOutputFormat.**checkOutputSpecs(**
>> PigOutputFormat.java:186)
>>          at org.apache.hadoop.mapreduce.**JobSubmitter.checkSpecs(**
>> JobSubmitter.java:456)
>>          at org.apache.hadoop.mapreduce.**JobSubmitter.**submitJobInternal(
>> **JobSubmitter.java:342)
>>          at org.apache.hadoop.mapreduce.**Job$10.run(Job.java:1268)
>>          at org.apache.hadoop.mapreduce.**Job$10.run(Job.java:1265)
>>          at java.security.**AccessController.doPrivileged(**Native Method)
>>          at javax.security.auth.Subject.**doAs(Subject.java:415)
>>          at org.apache.hadoop.security.**UserGroupInformation.doAs(**
>> UserGroupInformation.java:**1491)
>>          at org.apache.hadoop.mapreduce.**Job.submit(Job.java:1265)
>>          at org.apache.hadoop.mapreduce.**lib.jobcontrol.ControlledJob.**
>> submit(ControlledJob.java:335)
>>          at org.apache.hadoop.mapreduce.**lib.jobcontrol.JobControl.run(**
>> JobControl.java:240)
>>          at org.apache.pig.backend.**hadoop20.PigJobControl.run(**
>> PigJobControl.java:121)
>>          at java.lang.Thread.run(Thread.**java:724)
>>          at org.apache.pig.backend.hadoop.**executionengine.**
>> mapReduceLayer.**MapReduceLauncher$1.run(**MapReduceLauncher.java:257)
>>
>> Pig Stack Trace
>> ---------------
>> ERROR 1066: Unable to open iterator for alias data
>>
>> org.apache.pig.impl.**logicalLayer.**FrontendException: ERROR 1066:
>> Unable to open iterator for alias data
>>          at org.apache.pig.PigServer.**openIterator(PigServer.java:**836)
>>          at org.apache.pig.tools.grunt.**GruntParser.processDump(**
>> GruntParser.java:696)
>>          at org.apache.pig.tools.**pigscript.parser.**
>> PigScriptParser.parse(**PigScriptParser.java:320)
>>          at org.apache.pig.tools.grunt.**GruntParser.parseStopOnError(**
>> GruntParser.java:194)
>>          at org.apache.pig.tools.grunt.**GruntParser.parseStopOnError(**
>> GruntParser.java:170)
>>          at org.apache.pig.tools.grunt.**Grunt.run(Grunt.java:69)
>>          at org.apache.pig.Main.run(Main.**java:538)
>>          at org.apache.pig.Main.main(Main.**java:157)
>>          at sun.reflect.**NativeMethodAccessorImpl.**invoke0(Native Method)
>>          at sun.reflect.**NativeMethodAccessorImpl.**invoke(**
>> NativeMethodAccessorImpl.java:**57)
>>          at sun.reflect.**DelegatingMethodAccessorImpl.**invoke(**
>> DelegatingMethodAccessorImpl.**java:43)
>>          at java.lang.reflect.Method.**invoke(Method.java:606)
>>          at org.apache.hadoop.util.RunJar.**main(RunJar.java:212)
>> Caused by: java.io.IOException: Job terminated with anomalous status FAILED
>>          at org.apache.pig.PigServer.**openIterator(PigServer.java:**828)
>>          ... 12 more
>>
>>
>>


Re: Support for Hadoop 2.2

Posted by Prashant Kommireddi <pr...@gmail.com>.
Yes it does. You need to recompile Pig for hadoop 2

ant clean jar-withouthadoop -Dhadoopversion=23


On Thu, Oct 24, 2013 at 5:37 AM, Claudio Romo Otto <
claudio.romo@namictech.com> wrote:

> Does Pig support Hadoop 2.2? When I try Pig 0.12 and Hadoop 2.2 I get an
> error even with simple operations like
>
> data = LOAD 'cql://keyspace1/testcf?' USING CqlStorage();
> dump data;
>
> I only got a warning first and then and exception:
>
> 2013-10-24 09:35:19,300 [main] WARN org.apache.pig.backend.**hadoop20.PigJobControl
> - falling back to default JobControl (not using hadoop 0.20 ?)
> java.lang.**NoSuchFieldException: runnerState
>     at java.lang.Class.**getDeclaredField(Class.java:**1938)
>     at org.apache.pig.backend.**hadoop20.PigJobControl.<**
> clinit>(PigJobControl.java:51)
>     at org.apache.pig.backend.hadoop.**executionengine.shims.**
> HadoopShims.newJobControl(**HadoopShims.java:97)
>     at org.apache.pig.backend.hadoop.**executionengine.**mapReduceLayer.**
> JobControlCompiler.compile(**JobControlCompiler.java:285)
>     at org.apache.pig.backend.hadoop.**executionengine.**mapReduceLayer.**
> MapReduceLauncher.launchPig(**MapReduceLauncher.java:177)
>     at org.apache.pig.PigServer.**launchPlan(PigServer.java:**1264)
>     at org.apache.pig.PigServer.**executeCompiledLogicalPlan(**
> PigServer.java:1249)
>     at org.apache.pig.PigServer.**storeEx(PigServer.java:931)
>     at org.apache.pig.PigServer.**store(PigServer.java:898)
>     at org.apache.pig.PigServer.**openIterator(PigServer.java:**811)
>     at org.apache.pig.tools.grunt.**GruntParser.processDump(**
> GruntParser.java:696)
>     at org.apache.pig.tools.**pigscript.parser.**PigScriptParser.parse(**
> PigScriptParser.java:320)
>     at org.apache.pig.tools.grunt.**GruntParser.parseStopOnError(**
> GruntParser.java:194)
>     at org.apache.pig.tools.grunt.**GruntParser.parseStopOnError(**
> GruntParser.java:170)
>     at org.apache.pig.tools.grunt.**Grunt.run(Grunt.java:69)
>     at org.apache.pig.Main.run(Main.**java:538)
>     at org.apache.pig.Main.main(Main.**java:157)
>     at sun.reflect.**NativeMethodAccessorImpl.**invoke0(Native Method)
>     at sun.reflect.**NativeMethodAccessorImpl.**invoke(**
> NativeMethodAccessorImpl.java:**57)
>     at sun.reflect.**DelegatingMethodAccessorImpl.**invoke(**
> DelegatingMethodAccessorImpl.**java:43)
>     at java.lang.reflect.Method.**invoke(Method.java:606)
>     at org.apache.hadoop.util.RunJar.**main(RunJar.java:212)
>
> ------------------------------**--------
>
> Backend error message during job submission
> ------------------------------**-------------
> Unexpected System Error Occured: java.lang.**IncompatibleClassChangeError:
> Found interface org.apache.hadoop.mapreduce.**JobContext, but class was
> expected
>         at org.apache.pig.backend.hadoop.**executionengine.**
> mapReduceLayer.**PigOutputFormat.**setupUdfEnvAndStores(**
> PigOutputFormat.java:225)
>         at org.apache.pig.backend.hadoop.**executionengine.**
> mapReduceLayer.**PigOutputFormat.**checkOutputSpecs(**
> PigOutputFormat.java:186)
>         at org.apache.hadoop.mapreduce.**JobSubmitter.checkSpecs(**
> JobSubmitter.java:456)
>         at org.apache.hadoop.mapreduce.**JobSubmitter.**submitJobInternal(
> **JobSubmitter.java:342)
>         at org.apache.hadoop.mapreduce.**Job$10.run(Job.java:1268)
>         at org.apache.hadoop.mapreduce.**Job$10.run(Job.java:1265)
>         at java.security.**AccessController.doPrivileged(**Native Method)
>         at javax.security.auth.Subject.**doAs(Subject.java:415)
>         at org.apache.hadoop.security.**UserGroupInformation.doAs(**
> UserGroupInformation.java:**1491)
>         at org.apache.hadoop.mapreduce.**Job.submit(Job.java:1265)
>         at org.apache.hadoop.mapreduce.**lib.jobcontrol.ControlledJob.**
> submit(ControlledJob.java:335)
>         at org.apache.hadoop.mapreduce.**lib.jobcontrol.JobControl.run(**
> JobControl.java:240)
>         at org.apache.pig.backend.**hadoop20.PigJobControl.run(**
> PigJobControl.java:121)
>         at java.lang.Thread.run(Thread.**java:724)
>         at org.apache.pig.backend.hadoop.**executionengine.**
> mapReduceLayer.**MapReduceLauncher$1.run(**MapReduceLauncher.java:257)
>
> Pig Stack Trace
> ---------------
> ERROR 1066: Unable to open iterator for alias data
>
> org.apache.pig.impl.**logicalLayer.**FrontendException: ERROR 1066:
> Unable to open iterator for alias data
>         at org.apache.pig.PigServer.**openIterator(PigServer.java:**836)
>         at org.apache.pig.tools.grunt.**GruntParser.processDump(**
> GruntParser.java:696)
>         at org.apache.pig.tools.**pigscript.parser.**
> PigScriptParser.parse(**PigScriptParser.java:320)
>         at org.apache.pig.tools.grunt.**GruntParser.parseStopOnError(**
> GruntParser.java:194)
>         at org.apache.pig.tools.grunt.**GruntParser.parseStopOnError(**
> GruntParser.java:170)
>         at org.apache.pig.tools.grunt.**Grunt.run(Grunt.java:69)
>         at org.apache.pig.Main.run(Main.**java:538)
>         at org.apache.pig.Main.main(Main.**java:157)
>         at sun.reflect.**NativeMethodAccessorImpl.**invoke0(Native Method)
>         at sun.reflect.**NativeMethodAccessorImpl.**invoke(**
> NativeMethodAccessorImpl.java:**57)
>         at sun.reflect.**DelegatingMethodAccessorImpl.**invoke(**
> DelegatingMethodAccessorImpl.**java:43)
>         at java.lang.reflect.Method.**invoke(Method.java:606)
>         at org.apache.hadoop.util.RunJar.**main(RunJar.java:212)
> Caused by: java.io.IOException: Job terminated with anomalous status FAILED
>         at org.apache.pig.PigServer.**openIterator(PigServer.java:**828)
>         ... 12 more
>
>
>