You are viewing a plain text version of this content. The canonical link for it is here.
Posted to mapreduce-user@hadoop.apache.org by Zhudacai <zh...@hisilicon.com> on 2015/07/24 08:34:59 UTC

Container exited with a non-zero exit code

Hi.all,

I've just done a fresh install of Hadoop with three nodes, one master (NameNode, SecondNameNode, ResourceTracker) and two slaves (DataNode). The HDFS are successfully formatted, all services are up. When I run the examples, e.g. teragen, terasort, I occasionally got this exception:

/15/07/23 19:55:34 INFO mapreduce.Job: map 0% reduce 0%//
//15/07/23 19:55:40 INFO mapreduce.Job: Task Id : attempt_1437652487249_0001_m_000000_0, Status : FAILED//
//Exception from container-launch.//
//Container id: container_1437652487249_0001_01_000002//
//Exit code: 134//
//Exception message: /bin/bash: line 1: 21736 Aborted //
///usr/openjdk-1.8.0-internal/bin/java -Djava.net.preferIPv4Stack=true -Dhadoop.metrics.log.level=WARN -Xmx200m //
//-Djava.io.tmpdir=/home/hadoop3/tmp/nm-local-dir/usercache/root/appcache/application_1437652487249_0001/container_1437652487249_0001_01_000002/tmp //
//-Dlog4j.configuration=container-log4j.properties //
//-Dyarn.app.container.log.dir=/home/hadoop3/hadoop-2.6.0/logs/userlogs/application_1437652487249_0001/container_1437652487249_0001_01_000002 //
//-Dyarn.app.container.log.filesize=0 //
//-Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild 192.168.1.9 39868 attempt_1437652487249_0001_m_000000_0 2 > //
///home/hadoop3/hadoop-2.6.0/logs/userlogs/application_1437652487249_0001/container_1437652487249_0001_01_000002/stdout 2> /home/hadoop3/hadoop-2.6.0/logs/userlogs/application_1437652487249_0001/container_1437652487249_0001_01_000002/stderr//
//
//Stack trace: ExitCodeException exitCode=134: /bin/bash: line 1: 21736 Aborted //
///usr/openjdk-1.8.0-internal/bin/java -Djava.net.preferIPv4Stack=true //
//-Dhadoop.metrics.log.level=WARN -Xmx200m -Djava.io.tmpdir=/home/hadoop3/tmp/nm-local-dir/usercache/root/appcache/application_1437652487249_0001/container_1437652487249_0001_01_000002/tmp //
//-Dlog4j.configuration=container-log4j.properties //
//-Dyarn.app.container.log.dir=/home/hadoop3/hadoop-2.6.0/logs/userlogs/application_1437652487249_0001/container_1437652487249_0001_01_000002 -Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild 192.168.1.9 39868 attempt_1437652487249_0001_m_000000_0 2 > /home/hadoop3/hadoop-2.6.0/logs/userlogs/application_1437652487249_0001/container_1437652487249_0001_01_000002/stdout 2> /home/hadoop3/hadoop-2.6.0/logs/userlogs/application_1437652487249_0001/container_1437652487249_0001_01_000002/stderr//
//
//at org.apache.hadoop.util.Shell.runCommand(Shell.java:538)//
//at org.apache.hadoop.util.Shell.run(Shell.java:455)//
//at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:715)//
//at org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java:211)//
//at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:302)//
//at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:82)//
//at java.util.concurrent.FutureTask.run(FutureTask.java:266)//
//at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)//
//at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)//
//at java.lang.Thread.run(Thread.java:745)//
//
//
//Container exited with a non-zero exit code 134//
//
//15/07/23 19:55:46 INFO mapreduce.Job: map 3% reduce 0%/

But the job could succfully completed.
Here're the configurations:

core-site.xml
/<configuration>//
//<property>//
//<name>fs.defaultFS</name>//
//<value>hdfs://master:9000</value>//
//<final>true</final>//
//</property>//
//<property>//
//<name>hadoop.tmp.dir</name>//
//<value>file:/home/hadoop3/tmp</value>//
//</property>//
//</configuration>//
//
//hdfs-site.xml//
//<configuration>//
//<property>//
//<name>dfs.namenode.secondary.http-address</name>//
//<value>master:50090</value>//
//</property>//
//<property>//
//<name>dfs.namenode.name.dir</name>//
//<value>file:/home/hadoop3/tmp/dfs/name</value>//
//</property>//
//<property>//
//<name>dfs.datanode.data.dir</name>//
//<value>file:/home/hadoop3/tmp/dfs/data</value>//
//</property>//
//<property>//
//<name>dfs.replication</name>//
//<value>1</value>//
//</property>//
//</configuration>//
/
yarn-site.xml
/<configuration>//
//<property>//
//<name>yarn.resourcemanager.hostname</name>//
//<value>master</value>//
//</property>//
//<property>//
//<name>yarn.nodemanager.aux-services</name>//
//<value>mapreduce_shuffle</value>//
//</property>//
//<property>//
//<name>yarn.nodemanager.aux-services.mapreduce_shuffle.class</name>//
//<value>org.apache.hadoop.mapred.ShuffleHandler</value>//
//</property>//
//<property>//
//<name>yarn.resourcemanager.address</name>//
//<value>master:8032</value>//
//</property>//
//<property>//
//<name>yarn.resourcemanager.scheduler.address</name>//
//<value>master:8030</value>//
//</property>//
//<property>//
//<name>yarn.resourcemanager.resource-tracker.address</name>//
//<value>master:8035</value>//
//</property>//
//<property>//
//<name>yarn.resourcemanager.admin.address</name>//
//<value>master:8033</value>//
//</property>//
//<property>//
//<name>yarn.resourcemanager.webapp.address</name>//
//<value>master:8088</value>//
//</property>//
//<property>//
//<name>yarn.nodemanager.resource.cpu-vcores</name>//
//<value>16</value>//
//</property>//
//</configuration>/

mapred-site.xml
/<configuration>//
//<property>//
//<name>mapreduce.framework.name</name>//
//<value>yarn</value>//
//</property>//
//</configuration>/

I also noticed if hadoop was installed on single node, the exception would never show up.
The number of the exception increases with the values of -Dmapred.map.tasks and -Dmapred.reduce.tasks.

I'm using hadoop 2.6.0, OpenJDK 1.8, runing on the arm64 platform.

Best Regards

Jared