You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-dev@hadoop.apache.org by "Hemanth Yamijala (JIRA)" <ji...@apache.org> on 2009/06/15 05:56:07 UTC
[jira] Commented: (HADOOP-6035) jobtracker stops when namenode goes
out of safemode runing capacit scheduler
[ https://issues.apache.org/jira/browse/HADOOP-6035?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12719383#action_12719383 ]
Hemanth Yamijala commented on HADOOP-6035:
------------------------------------------
Hi, can you tell us what is the configuration for the capacity scheduler ? Have you changed any of the defaults ? Have you added any queues ?
> jobtracker stops when namenode goes out of safemode runing capacit scheduler
> ----------------------------------------------------------------------------
>
> Key: HADOOP-6035
> URL: https://issues.apache.org/jira/browse/HADOOP-6035
> Project: Hadoop Core
> Issue Type: Bug
> Affects Versions: 0.20.0
> Environment: Fedora 10
> Reporter: Anjali M
> Priority: Minor
>
> I am facing a problem running the capacity scheduler in hadoop-0.20.0.
> The jobtracker is listing the queues when namenode is in the safemode.
> Once the namenode goes out of the safemode the jt stops working. On
> accessing jobqueue details it shows the following error.
> HTTP ERROR: 500
> INTERNAL_SERVER_ERROR
> RequestURI=/jobqueue_details.jsp
> Caused by:
> java.lang.NullPointerException
> at org.apache.hadoop.mapred.JobQueuesManager.getRunningJobQueue(JobQueuesManager.java:156)
> at org.apache.hadoop.mapred.CapacityTaskScheduler.getJobs(CapacityTaskScheduler.java:1495)
> at org.apache.hadoop.mapred.jobqueue_005fdetails_jsp._jspService(jobqueue_005fdetails_jsp.java:64)
> at org.apache.jasper.runtime.HttpJspBase.service(HttpJspBase.java:97)
> at javax.servlet.http.HttpServlet.service(HttpServlet.java:820)
> at org.mortbay.jetty.servlet.ServletHolder.handle(ServletHolder.java:502)
> at org.mortbay.jetty.servlet.ServletHandler.handle(ServletHandler.java:363)
> at org.mortbay.jetty.security.SecurityHandler.handle(SecurityHandler.java:216)
> at org.mortbay.jetty.servlet.SessionHandler.handle(SessionHandler.java:181)
> at org.mortbay.jetty.handler.ContextHandler.handle(ContextHandler.java:766)
> at org.mortbay.jetty.webapp.WebAppContext.handle(WebAppContext.java:417)
> at org.mortbay.jetty.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:230)
> at org.mortbay.jetty.handler.HandlerWrapper.handle(HandlerWrapper.java:152)
> at org.mortbay.jetty.Server.handle(Server.java:324)
> at org.mortbay.jetty.HttpConnection.handleRequest(HttpConnection.java:534)
> at org.mortbay.jetty.HttpConnection$RequestHandler.headerComplete(HttpConnection.java:864)
> at org.mortbay.jetty.HttpParser.parseNext(HttpParser.java:533)
> at org.mortbay.jetty.HttpParser.parseAvailable(HttpParser.java:207)
> at org.mortbay.jetty.HttpConnection.handle(HttpConnection.java:403)
> at org.mortbay.io.nio.SelectChannelEndPoint.run(SelectChannelEndPoint.java:409)
> at org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:522)
> Is it because any of the configuration in capacity-scheduler.xml is incorrect?
> I tried forcing the namenode out of the safemode in bin/hadoop
> dfsadmin, but still it does not work.
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.