You are viewing a plain text version of this content. The canonical link for it is here.
Posted to mapreduce-user@hadoop.apache.org by arun k <ar...@gmail.com> on 2011/09/16 09:04:52 UTC

(Unknown)

Hi !

Thanks Harsh !
The problem was that i have set up queue info in mapred-site.xml instead of
capacity-scheduler.xml .

Arun


On Fri, Sep 16, 2011 at 10:52 AM, Harsh J <ha...@cloudera.com> wrote:

> Arun,
>
> Please do not cross-post to multiple lists. Lets continue this on
> mapreduce-user@ alone.
>
> Your problem isn't the job submission here, but your Capacity
> Scheduler configuration. For every queue you configure, you need to
> add in capacities: Please see the queue properties documentation at
>
> http://hadoop.apache.org/common/docs/current/capacity_scheduler.html#Queue+properties
> for the vital configs required in additional to mapred.queue.names.
> Once done, you should have a fully functional JobTracker!
>
> On Fri, Sep 16, 2011 at 10:17 AM, arun k <ar...@gmail.com> wrote:
> > Hi all !
> >
> > Harsh ! Namenode appears to be out of safe mode :
> > In http://nn-host:50070 i see in time
> >
> > T1>Safe mode is ON. The ratio of reported blocks 0.0000 has not reached
> the
> > threshold 0.9990. Safe mode will be turned off automatically.
> > 7 files and directories, 1 blocks = 8 total. Heap Size is 15.06 MB /
> 966.69
> > MB (1%)
> >
> > T2>Safe mode is ON. The ratio of reported blocks 1.0000 has reached the
> > threshold 0.9990. Safe mode will be turned off automatically in 17
> seconds.
> > 7 files and directories, 1 blocks = 8 total. Heap Size is 15.06 MB /
> 966.69
> > MB (1%)
> >
> > T3>9 files and directories, 3 blocks = 12 total. Heap Size is 15.06 MB /
> > 966.69 MB (1%)
> >
> > Added properties :
> >
> >  mapred.jobtracker.taskScheduler org.apache.hadoopertiep.mapred.CTS
> >
> >  mapred.queue.names                          myqueue1,myqueue2
> >  mapred.capacity-scheduler.queue.myqueue1.capacity               25
> >  mapred.capacity-scheduler.queue.myqueue1.capacity               75
> > ${HADOOP_HOME}$ bin/hadoop jar hadoop*examples*.jar wordcount
> > -Dmapred.job.queue.name=
> > myqueue1 /user/hduser/wcinput /user/hduser/wcoutput
> >
> > I get the error:
> > java.io.IOException: Call to localhost/127.0.0.1:54311 failed on local
> > exception: java.io.IOException: Connection reset by peer
> >     at org.apache.hadoop.ipc.Client.wrapException(Client.java:1065)
> >     at org.apache.hadoop.ipc.Client.call(Client.java:1033)
> >     at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:224)
> >         ...................
> >
> > When i give
> > $jps
> > 32463 NameNode
> > 32763 SecondaryNameNode
> > 32611 DataNode
> > 931 Jps
> >
> >
> > The jobracker log gives info
> > ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
> > 2011-09-16 00:21:42,012 INFO org.apache.hadoop.mapred.JobTracker:
> Cleaning
> > up the system directory
> > 2011-09-16 00:21:42,014 INFO org.apache.hadoop.mapred.JobTracker: problem
> > cleaning system directory:
> > hdfs://localhost:54310/app203/hadoop203/tmp/mapred/system
> > org.apache.hadoop.ipc.RemoteException:
> > org.apache.hadoop.hdfs.server.namenode.SafeModeException: Cannot delete
> > /app203/hadoop203/tmp/mapred/system. Name node is in safe mode.
> > The ratio of reported blocks 1.0000 has reached the threshold 0.9990.
> Safe
> > mode will be turned off automatically in 6 seconds.
> >     at
> >
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.deleteInternal(FSNamesystem.java:1851)
> >     at
> >
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.delete(FSNamesystem.java:1831)
> > 2011-09-16 00:21:52,321 FATAL org.apache.hadoop.mapred.JobTracker:
> > java.io.IOException: Queue 'myqueue1' doesn't have configured capacity!
> >     at
> >
> org.apache.hadoop.mapred.CapacityTaskScheduler.parseQueues(CapacityTaskScheduler.java:905)
> >     at
> >
> org.apache.hadoop.mapred.CapacityTaskScheduler.start(CapacityTaskScheduler.java:822)
> >     at
> > org.apache.hadoop.mapred.JobTracker.offerService(JobTracker.java:2563)
> >     at org.apache.hadoop.mapred.JobTracker.main(JobTracker.java:4957)
> >
> > 2011-09-16 00:21:52,322 INFO org.apache.hadoop.mapred.JobTracker:
> > SHUTDOWN_MSG:
> > ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
> >
> > Even if i submit the job to "myqueue2" i see the same error of "myqueue1"
> > 2011-09-16 00:21:52,321 FATAL org.apache.hadoop.mapred.JobTracker:
> > java.io.IOException: Queue 'myqueue1' doesn't have configured capacity!
> >
> > Thanks,
> > Arun
> >
> >
> >
> > On Thu, Sep 15, 2011 at 5:23 PM, Harsh J <ha...@cloudera.com> wrote:
> >>
> >> Hello Arun,
> >>
> >> To me it looks like your HDFS isn't setup properly, in this case. Can
> >> you ensure all DNs are properly up? Your NN appears to have gotten
> >> stuck somehow into a safemode. Check out your http://nn-host:50070
> >> page for more details on why.
> >>
> >> Your JT won't come up until the NN is properly up and out of safemode
> >> (for which it needs the DNs). And once it comes up, I think you should
> >> be good to go, keeping in mind the changes Thomas mentioned earlier.
> >>
> >> On Thu, Sep 15, 2011 at 3:58 PM, arun k <ar...@gmail.com> wrote:
> >> > Hi all !
> >> >
> >> > Thanks Thomas ! it's working in terminal.
> >> > I saw the queues in web UI of JT.
> >> > when i try to run normally again (default) i get this error :
> >> > i tried formatting namenode and making safemode off and restart but
> >> > didn't
> >> > work.
> >> >
> >> > hduser@arun-Presario-C500-RU914PA-ACJ:/usr/local/hadoop$ bin/hadoop
> jar
> >> > hadoop*examples*.jar wordcount  /user/hduser/wcinput
> >> > /user/hduser/wcoutput6
> >> > java.io.IOException: Call to localhost/127.0.0.1:54311 failed on
> local
> >> > exception: java.io.IOException: Connection reset by peer
> >> >
> >> > The log of JobTracker shows :
> >> > 2011-09-15 12:46:13,346 INFO org.apache.hadoop.mapred.JobTracker:
> >> > JobTracker
> >> > up at: 54311
> >> > 2011-09-15 12:46:13,347 INFO org.apache.hadoop.mapred.JobTracker:
> >> > JobTracker
> >> > webserver: 50030
> >> > 2011-09-15 12:46:13,634 INFO org.apache.hadoop.mapred.JobTracker:
> >> > Cleaning
> >> > up the system directory
> >> > 2011-09-15 12:46:13,646 INFO org.apache.hadoop.mapred.JobTracker:
> >> > problem
> >> > cleaning system directory:
> >> > hdfs://localhost:54310/app/hadoop/tmp/mapred/system
> >> > org.apache.hadoop.ipc.RemoteException:
> >> > org.apache.hadoop.hdfs.server.namenode.SafeModeException: Cannot
> delete
> >> > /app/hadoop/tmp/mapred/system. Name node is in safe mode.
> >> >
> >> > Thanks,
> >> > Arun
> >> >
> >> >
> >> >
> >> >
> >> >
> >> >
> >> > On Wed, Sep 14, 2011 at 7:46 PM, Thomas Graves <tgraves@yahoo-inc.com
> >
> >> > wrote:
> >> >>
> >> >> I believe it defaults to submit a job to the default queue if you
> don't
> >> >> specify it.  You don't have the default queue defined in your list of
> >> >> mapred.queue.names.  So add -Dmapred.job.queue.name=myqueue1 (or
> >> >> another
> >> >> queue you have defined) to the wordcount command like:
> >> >>
> >> >> bin/hadoop jar
> >> >> > hadoop*examples*.jar wordcount -Dmapred.job.queue.name=myqueue1
> >> >> /user/hduser/wcinput /user/hduser/wcoutput5
> >> >>
> >> >> Tom
> >> >>
> >> >>
> >> >> On 9/14/11 5:57 AM, "arun k" <ar...@gmail.com> wrote:
> >> >>
> >> >> > Hi !
> >> >> >
> >> >> > I have set up single-node cluster using
> >> >> >
> >> >> >
> >> >> >
> http://www.google.co.in/url?sa=t&source=web&cd=1&ved=0CB0QFjAA&url=http%3A%2F%
> >> >> >
> >> >> >
> >> >> > 2Fwww.michael-noll.com
> %2Ftutorials%2Frunning-hadoop-on-ubuntu-linux-single-nod
> >> >> >
> >> >> >
> >> >> >
> e-cluster%2F&rct=j&q=michael%20noll%20single%20node&ei=b4ZwTvrCLsOrrAei-N32Bg&
> >> >> >
> >> >> >
> >> >> >
> usg=AFQjCNGhuvv0tNdvPj4u23bbj-qXJDlixg&sig2=7ij8Dy7aQZUkBwhTnS1rLw&cad=rja
> >> >> > and could run wordcount example application.
> >> >> > I was trying to run this application using capacity scheduler.
> >> >> > As per
> >> >> >
> http://hadoop.apache.org/common/docs/current/capacity_scheduler.htmli
> >> >> > have done :
> >> >> > 1.Copied the hadoop-capacity-scheduler-*.jar from *
> >> >> > contrib/capacity-scheduler* directory to HADOOP_HOME/lib
> >> >> > 2.Set mapred.jobtracker.taskScheduler
> >> >> > 3.Set *mapred.queue.names to myqueue1,myqueue2.
> >> >> > 4.Set *mapred.capacity-scheduler.queue.<queue-name>.capacity to 30
> >> >> > and
> >> >> > 70
> >> >> > for two queues.
> >> >> >
> >> >> > When i run i get the error :
> >> >> > hduser@arun-Presario-C500-RU914PA-ACJ:/usr/local/hadoop$
> bin/hadoop
> >> >> > jar
> >> >> > hadoop*examples*.jar wordcount /user/hduser/wcinput
> >> >> > /user/hduser/wcoutput5
> >> >> > 11/09/14 16:00:56 INFO input.FileInputFormat: Total input paths to
> >> >> > process :
> >> >> > 4
> >> >> > org.apache.hadoop.ipc.RemoteException: java.io.IOException: Queue
> >> >> > "default"
> >> >> > does not exist
> >> >> >     at
> >> >> > org.apache.hadoop.mapred.JobTracker.submitJob(JobTracker.java:2998)
> >> >> >     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> >> >> >     at
> >> >> >
> >> >> >
> >> >> >
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
> >> >> >     at
> >> >> >
> >> >> >
> >> >> >
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.j
> >> >> > ava:25)
> >> >> >     at java.lang.reflect.Method.invoke(Method.java:597)
> >> >> >     at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508)
> >> >> >     at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:959)
> >> >> >     at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:955)
> >> >> >     at java.security.AccessController.doPrivileged(Native Method)
> >> >> >     at javax.security.auth.Subject.doAs(Subject.java:396)
> >> >> >     at org.apache.hadoop.ipc.Server$Handler.run(Server.java:953)
> >> >> >
> >> >> >     at org.apache.hadoop.ipc.Client.call(Client.java:740)
> >> >> >     at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220)
> >> >> >     at org.apache.hadoop.mapred.$Proxy0.submitJob(Unknown Source)
> >> >> >     at
> >> >> >
> >> >> >
> org.apache.hadoop.mapred.JobClient.submitJobInternal(JobClient.java:800)
> >> >> >     at org.apache.hadoop.mapreduce.Job.submit(Job.java:432)
> >> >> >     at
> >> >> > org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:447)
> >> >> >     at org.apache.hadoop.examples.WordCount.main(WordCount.java:67)
> >> >> >     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> >> >> >     at
> >> >> >
> >> >> >
> >> >> >
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
> >> >> >     at
> >> >> >
> >> >> >
> >> >> >
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.j
> >> >> > ava:25)
> >> >> >     at java.lang.reflect.Method.invoke(Method.java:597)
> >> >> >     at
> >> >> >
> >> >> >
> >> >> >
> org.apache.hadoop.util.ProgramDriver$ProgramDescription.invoke(ProgramDriver.j
> >> >> > ava:68)
> >> >> >     at
> >> >> > org.apache.hadoop.util.ProgramDriver.driver(ProgramDriver.java:139)
> >> >> >     at
> >> >> >
> org.apache.hadoop.examples.ExampleDriver.main(ExampleDriver.java:64)
> >> >> >     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> >> >> >     at
> >> >> >
> >> >> >
> >> >> >
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
> >> >> >     at
> >> >> >
> >> >> >
> >> >> >
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.j
> >> >> > ava:25)
> >> >> >     at java.lang.reflect.Method.invoke(Method.java:597)
> >> >> >     at org.apache.hadoop.util.RunJar.main(RunJar.java:156)
> >> >> >
> >> >> > I didn't submit jobs to a particular queue as such. Do i need to do
> >> >> > it ?
> >> >> > How
> >> >> > can i do it ?
> >> >> > Any help ?
> >> >> >
> >> >> > Thanks,
> >> >> > Arun
> >> >> >
> >> >> >
> >> >> > *
> >> >> > *
> >> >>
> >> >
> >> >
> >>
> >>
> >>
> >> --
> >> Harsh J
> >
> >
>
>
>
> --
> Harsh J
>