You are viewing a plain text version of this content. The canonical link for it is here.
Posted to hdfs-user@hadoop.apache.org by Sisu Xi <xi...@gmail.com> on 2014/07/13 07:58:53 UTC

hadoop multinode, only master node doing the work

Hi, all:

I am new to hadoop. I followed the tutorial on
http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-multi-node-cluster/

and installed hadoop 2.2.0 on two 4-core Ubuntu 12.04 machines.

I can start the pi program, however, only the master node is doing the work
(I checked the top on each machine).
Seems the two nodes are configured correctly, because I can start  the
program in the slave node as well, and still only the master node is doing
the actual work.
I have tried different number of mappers for the pi program, and the
results is the same.

Is there anything else I can check?

In the end is my configure file on each host.

Thanks very much!

Sisu

---------yarn-site.xml-------

<property>
  <name>yarn.nodemanager.aux-services</name>
  <value>mapreduce_shuffle</value>
</property>


<property>
  <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name>
  <value>org.apache.hadoop.mapred.ShuffleHandler</value>
</property>


<property>
  <name>yarn.resourcemanager.address</name>
  <value>master:8032</value>
</property>

---------------hdfs-site.xml--------------------

<property>
  <name>dfs.replication</name>
  <value>2</value>
</property>


<property>
  <name>dfs.namenode.name.dir</name>
  <value>file:/home/xisisu/mydata/hdfs/namenode</value>
</property>


<property>
  <name>dfs.datanode.data.dir</name>
  <value>file:/home/xisisu/mydata/hdfs/datanode</value>
</property>


-------------core-site.xml-------------

<property>
<name>fs.default.name</name>
<value>hdfs://master:9000</value>
</property>

------------------mapred-site.xml-----------------

<property>
  <name>mapreduce.framework.name</name>
  <value>yarn</value>
</property>

<property>
  <name>mapred.job.tracker</name>
  <value>master:54311</value>
  <description>The host and port that the MapReduce job tracker runs
  at.  If "local", then jobs are run in-process as a single map
  and reduce task.
  </description>
</property>




-- 


*Sisu Xi, PhD Candidate*

http://www.cse.wustl.edu/~xis/
Department of Computer Science and Engineering
Campus Box 1045
Washington University in St. Louis
One Brookings Drive
St. Louis, MO 63130

Re: hadoop multinode, only master node doing the work

Posted by Sisu Xi <xi...@gmail.com>.
Hi, Ravi:

Thanks very much!!!

I added:
<property>
  <name>yarn.resourcemanager.hostname</name>
  <value>master</value>
</property>
in the yarn-site.xml on both machines, and both machine can do the actual
work now!
And the web UI is showing two machines on it!  :)

Thanks again for your help!

Sisu


On Sun, Jul 13, 2014 at 2:53 PM, Ravi Mutyala <ra...@hortonworks.com> wrote:

> Yes. As vinod pointed in that link, check yarn.resourcemanager.hostname
> on slave node yarn-site.xml and put your master node hostname there.
>
>
> On Sun, Jul 13, 2014 at 2:47 PM, Sisu Xi <xi...@gmail.com> wrote:
>
>> Hi, Ravi:
>>
>> Thanks again for your help!
>>
>> The website only shows one node. a screenshot is attached.
>>
>> I restart the service, on master, it says:
>> xisisu@master:/usr/local/hadoop$ jps
>> 15862 Jps
>> 15570 NodeManager
>> 15461 ResourceManager
>> 15149 DataNode
>> 15025 NameNode
>>
>>  on slave, it says:
>> xisisu@slave-01:/usr/local/hadoop$ jps
>> 19752 DataNode
>> 19859 NodeManager
>> 20003 Jps
>>
>> I also checked the yarn log on the slave node, seems it cannot connect to
>> the server,
>> xisisu@slave-01:/usr/local/hadoop$ tail
>> logs/yarn-xisisu-nodemanager-slave-01.log
>> 2014-07-13 14:41:35,947 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 0 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>> 2014-07-13 14:41:36,947 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 1 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>> 2014-07-13 14:41:37,961 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 2 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>> 2014-07-13 14:41:38,962 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 3 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>> 2014-07-13 14:41:39,962 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 4 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>> 2014-07-13 14:41:40,963 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 5 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>> 2014-07-13 14:41:41,963 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 6 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>> 2014-07-13 14:41:42,963 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 7 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>> 2014-07-13 14:41:43,964 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 8 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>> 2014-07-13 14:41:44,964 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 9 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>>
>>
>> I googled and found this:
>>
>> http://hortonworks.com/community/forums/topic/retrying-connect-to-server-0-0-0-00-0-0-08030/
>>
>> Seems it is the configuration error?
>>
>> Thanks very much!
>>
>> Sisu
>>
>>
>>
>>
>>
>>
>> On Sun, Jul 13, 2014 at 2:33 PM, Ravi Mutyala <ra...@hortonworks.com>
>> wrote:
>>
>>> http://localhost:8088/cluster/nodes
>>
>>
>>
>>
>>
>> --
>>
>>
>> *Sisu Xi, PhD Candidate*
>>
>> http://www.cse.wustl.edu/~xis/
>> Department of Computer Science and Engineering
>> Campus Box 1045
>> Washington University in St. Louis
>> One Brookings Drive
>> St. Louis, MO 63130
>>
>
>
> CONFIDENTIALITY NOTICE
> NOTICE: This message is intended for the use of the individual or entity
> to which it is addressed and may contain information that is confidential,
> privileged and exempt from disclosure under applicable law. If the reader
> of this message is not the intended recipient, you are hereby notified that
> any printing, copying, dissemination, distribution, disclosure or
> forwarding of this communication is strictly prohibited. If you have
> received this communication in error, please contact the sender immediately
> and delete it from your system. Thank You.
>



-- 


*Sisu Xi, PhD Candidate*

http://www.cse.wustl.edu/~xis/
Department of Computer Science and Engineering
Campus Box 1045
Washington University in St. Louis
One Brookings Drive
St. Louis, MO 63130

Re: hadoop multinode, only master node doing the work

Posted by Sisu Xi <xi...@gmail.com>.
Hi, Ravi:

Thanks very much!!!

I added:
<property>
  <name>yarn.resourcemanager.hostname</name>
  <value>master</value>
</property>
in the yarn-site.xml on both machines, and both machine can do the actual
work now!
And the web UI is showing two machines on it!  :)

Thanks again for your help!

Sisu


On Sun, Jul 13, 2014 at 2:53 PM, Ravi Mutyala <ra...@hortonworks.com> wrote:

> Yes. As vinod pointed in that link, check yarn.resourcemanager.hostname
> on slave node yarn-site.xml and put your master node hostname there.
>
>
> On Sun, Jul 13, 2014 at 2:47 PM, Sisu Xi <xi...@gmail.com> wrote:
>
>> Hi, Ravi:
>>
>> Thanks again for your help!
>>
>> The website only shows one node. a screenshot is attached.
>>
>> I restart the service, on master, it says:
>> xisisu@master:/usr/local/hadoop$ jps
>> 15862 Jps
>> 15570 NodeManager
>> 15461 ResourceManager
>> 15149 DataNode
>> 15025 NameNode
>>
>>  on slave, it says:
>> xisisu@slave-01:/usr/local/hadoop$ jps
>> 19752 DataNode
>> 19859 NodeManager
>> 20003 Jps
>>
>> I also checked the yarn log on the slave node, seems it cannot connect to
>> the server,
>> xisisu@slave-01:/usr/local/hadoop$ tail
>> logs/yarn-xisisu-nodemanager-slave-01.log
>> 2014-07-13 14:41:35,947 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 0 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>> 2014-07-13 14:41:36,947 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 1 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>> 2014-07-13 14:41:37,961 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 2 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>> 2014-07-13 14:41:38,962 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 3 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>> 2014-07-13 14:41:39,962 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 4 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>> 2014-07-13 14:41:40,963 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 5 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>> 2014-07-13 14:41:41,963 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 6 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>> 2014-07-13 14:41:42,963 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 7 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>> 2014-07-13 14:41:43,964 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 8 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>> 2014-07-13 14:41:44,964 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 9 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>>
>>
>> I googled and found this:
>>
>> http://hortonworks.com/community/forums/topic/retrying-connect-to-server-0-0-0-00-0-0-08030/
>>
>> Seems it is the configuration error?
>>
>> Thanks very much!
>>
>> Sisu
>>
>>
>>
>>
>>
>>
>> On Sun, Jul 13, 2014 at 2:33 PM, Ravi Mutyala <ra...@hortonworks.com>
>> wrote:
>>
>>> http://localhost:8088/cluster/nodes
>>
>>
>>
>>
>>
>> --
>>
>>
>> *Sisu Xi, PhD Candidate*
>>
>> http://www.cse.wustl.edu/~xis/
>> Department of Computer Science and Engineering
>> Campus Box 1045
>> Washington University in St. Louis
>> One Brookings Drive
>> St. Louis, MO 63130
>>
>
>
> CONFIDENTIALITY NOTICE
> NOTICE: This message is intended for the use of the individual or entity
> to which it is addressed and may contain information that is confidential,
> privileged and exempt from disclosure under applicable law. If the reader
> of this message is not the intended recipient, you are hereby notified that
> any printing, copying, dissemination, distribution, disclosure or
> forwarding of this communication is strictly prohibited. If you have
> received this communication in error, please contact the sender immediately
> and delete it from your system. Thank You.
>



-- 


*Sisu Xi, PhD Candidate*

http://www.cse.wustl.edu/~xis/
Department of Computer Science and Engineering
Campus Box 1045
Washington University in St. Louis
One Brookings Drive
St. Louis, MO 63130

Re: hadoop multinode, only master node doing the work

Posted by Sisu Xi <xi...@gmail.com>.
Hi, Ravi:

Thanks very much!!!

I added:
<property>
  <name>yarn.resourcemanager.hostname</name>
  <value>master</value>
</property>
in the yarn-site.xml on both machines, and both machine can do the actual
work now!
And the web UI is showing two machines on it!  :)

Thanks again for your help!

Sisu


On Sun, Jul 13, 2014 at 2:53 PM, Ravi Mutyala <ra...@hortonworks.com> wrote:

> Yes. As vinod pointed in that link, check yarn.resourcemanager.hostname
> on slave node yarn-site.xml and put your master node hostname there.
>
>
> On Sun, Jul 13, 2014 at 2:47 PM, Sisu Xi <xi...@gmail.com> wrote:
>
>> Hi, Ravi:
>>
>> Thanks again for your help!
>>
>> The website only shows one node. a screenshot is attached.
>>
>> I restart the service, on master, it says:
>> xisisu@master:/usr/local/hadoop$ jps
>> 15862 Jps
>> 15570 NodeManager
>> 15461 ResourceManager
>> 15149 DataNode
>> 15025 NameNode
>>
>>  on slave, it says:
>> xisisu@slave-01:/usr/local/hadoop$ jps
>> 19752 DataNode
>> 19859 NodeManager
>> 20003 Jps
>>
>> I also checked the yarn log on the slave node, seems it cannot connect to
>> the server,
>> xisisu@slave-01:/usr/local/hadoop$ tail
>> logs/yarn-xisisu-nodemanager-slave-01.log
>> 2014-07-13 14:41:35,947 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 0 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>> 2014-07-13 14:41:36,947 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 1 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>> 2014-07-13 14:41:37,961 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 2 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>> 2014-07-13 14:41:38,962 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 3 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>> 2014-07-13 14:41:39,962 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 4 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>> 2014-07-13 14:41:40,963 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 5 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>> 2014-07-13 14:41:41,963 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 6 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>> 2014-07-13 14:41:42,963 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 7 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>> 2014-07-13 14:41:43,964 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 8 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>> 2014-07-13 14:41:44,964 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 9 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>>
>>
>> I googled and found this:
>>
>> http://hortonworks.com/community/forums/topic/retrying-connect-to-server-0-0-0-00-0-0-08030/
>>
>> Seems it is the configuration error?
>>
>> Thanks very much!
>>
>> Sisu
>>
>>
>>
>>
>>
>>
>> On Sun, Jul 13, 2014 at 2:33 PM, Ravi Mutyala <ra...@hortonworks.com>
>> wrote:
>>
>>> http://localhost:8088/cluster/nodes
>>
>>
>>
>>
>>
>> --
>>
>>
>> *Sisu Xi, PhD Candidate*
>>
>> http://www.cse.wustl.edu/~xis/
>> Department of Computer Science and Engineering
>> Campus Box 1045
>> Washington University in St. Louis
>> One Brookings Drive
>> St. Louis, MO 63130
>>
>
>
> CONFIDENTIALITY NOTICE
> NOTICE: This message is intended for the use of the individual or entity
> to which it is addressed and may contain information that is confidential,
> privileged and exempt from disclosure under applicable law. If the reader
> of this message is not the intended recipient, you are hereby notified that
> any printing, copying, dissemination, distribution, disclosure or
> forwarding of this communication is strictly prohibited. If you have
> received this communication in error, please contact the sender immediately
> and delete it from your system. Thank You.
>



-- 


*Sisu Xi, PhD Candidate*

http://www.cse.wustl.edu/~xis/
Department of Computer Science and Engineering
Campus Box 1045
Washington University in St. Louis
One Brookings Drive
St. Louis, MO 63130

Re: hadoop multinode, only master node doing the work

Posted by Sisu Xi <xi...@gmail.com>.
Hi, Ravi:

Thanks very much!!!

I added:
<property>
  <name>yarn.resourcemanager.hostname</name>
  <value>master</value>
</property>
in the yarn-site.xml on both machines, and both machine can do the actual
work now!
And the web UI is showing two machines on it!  :)

Thanks again for your help!

Sisu


On Sun, Jul 13, 2014 at 2:53 PM, Ravi Mutyala <ra...@hortonworks.com> wrote:

> Yes. As vinod pointed in that link, check yarn.resourcemanager.hostname
> on slave node yarn-site.xml and put your master node hostname there.
>
>
> On Sun, Jul 13, 2014 at 2:47 PM, Sisu Xi <xi...@gmail.com> wrote:
>
>> Hi, Ravi:
>>
>> Thanks again for your help!
>>
>> The website only shows one node. a screenshot is attached.
>>
>> I restart the service, on master, it says:
>> xisisu@master:/usr/local/hadoop$ jps
>> 15862 Jps
>> 15570 NodeManager
>> 15461 ResourceManager
>> 15149 DataNode
>> 15025 NameNode
>>
>>  on slave, it says:
>> xisisu@slave-01:/usr/local/hadoop$ jps
>> 19752 DataNode
>> 19859 NodeManager
>> 20003 Jps
>>
>> I also checked the yarn log on the slave node, seems it cannot connect to
>> the server,
>> xisisu@slave-01:/usr/local/hadoop$ tail
>> logs/yarn-xisisu-nodemanager-slave-01.log
>> 2014-07-13 14:41:35,947 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 0 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>> 2014-07-13 14:41:36,947 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 1 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>> 2014-07-13 14:41:37,961 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 2 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>> 2014-07-13 14:41:38,962 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 3 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>> 2014-07-13 14:41:39,962 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 4 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>> 2014-07-13 14:41:40,963 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 5 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>> 2014-07-13 14:41:41,963 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 6 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>> 2014-07-13 14:41:42,963 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 7 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>> 2014-07-13 14:41:43,964 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 8 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>> 2014-07-13 14:41:44,964 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 9 time(s); retry
>> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
>> SECONDS)
>>
>>
>> I googled and found this:
>>
>> http://hortonworks.com/community/forums/topic/retrying-connect-to-server-0-0-0-00-0-0-08030/
>>
>> Seems it is the configuration error?
>>
>> Thanks very much!
>>
>> Sisu
>>
>>
>>
>>
>>
>>
>> On Sun, Jul 13, 2014 at 2:33 PM, Ravi Mutyala <ra...@hortonworks.com>
>> wrote:
>>
>>> http://localhost:8088/cluster/nodes
>>
>>
>>
>>
>>
>> --
>>
>>
>> *Sisu Xi, PhD Candidate*
>>
>> http://www.cse.wustl.edu/~xis/
>> Department of Computer Science and Engineering
>> Campus Box 1045
>> Washington University in St. Louis
>> One Brookings Drive
>> St. Louis, MO 63130
>>
>
>
> CONFIDENTIALITY NOTICE
> NOTICE: This message is intended for the use of the individual or entity
> to which it is addressed and may contain information that is confidential,
> privileged and exempt from disclosure under applicable law. If the reader
> of this message is not the intended recipient, you are hereby notified that
> any printing, copying, dissemination, distribution, disclosure or
> forwarding of this communication is strictly prohibited. If you have
> received this communication in error, please contact the sender immediately
> and delete it from your system. Thank You.
>



-- 


*Sisu Xi, PhD Candidate*

http://www.cse.wustl.edu/~xis/
Department of Computer Science and Engineering
Campus Box 1045
Washington University in St. Louis
One Brookings Drive
St. Louis, MO 63130

Re: hadoop multinode, only master node doing the work

Posted by Ravi Mutyala <ra...@hortonworks.com>.
Yes. As vinod pointed in that link, check yarn.resourcemanager.hostname on
slave node yarn-site.xml and put your master node hostname there.


On Sun, Jul 13, 2014 at 2:47 PM, Sisu Xi <xi...@gmail.com> wrote:

> Hi, Ravi:
>
> Thanks again for your help!
>
> The website only shows one node. a screenshot is attached.
>
> I restart the service, on master, it says:
> xisisu@master:/usr/local/hadoop$ jps
> 15862 Jps
> 15570 NodeManager
> 15461 ResourceManager
> 15149 DataNode
> 15025 NameNode
>
>  on slave, it says:
> xisisu@slave-01:/usr/local/hadoop$ jps
> 19752 DataNode
> 19859 NodeManager
> 20003 Jps
>
> I also checked the yarn log on the slave node, seems it cannot connect to
> the server,
> xisisu@slave-01:/usr/local/hadoop$ tail
> logs/yarn-xisisu-nodemanager-slave-01.log
> 2014-07-13 14:41:35,947 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 0 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
> 2014-07-13 14:41:36,947 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 1 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
> 2014-07-13 14:41:37,961 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 2 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
> 2014-07-13 14:41:38,962 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 3 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
> 2014-07-13 14:41:39,962 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 4 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
> 2014-07-13 14:41:40,963 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 5 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
> 2014-07-13 14:41:41,963 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 6 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
> 2014-07-13 14:41:42,963 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 7 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
> 2014-07-13 14:41:43,964 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 8 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
> 2014-07-13 14:41:44,964 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 9 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
>
>
> I googled and found this:
>
> http://hortonworks.com/community/forums/topic/retrying-connect-to-server-0-0-0-00-0-0-08030/
>
> Seems it is the configuration error?
>
> Thanks very much!
>
> Sisu
>
>
>
>
>
>
> On Sun, Jul 13, 2014 at 2:33 PM, Ravi Mutyala <ra...@hortonworks.com>
> wrote:
>
>> http://localhost:8088/cluster/nodes
>
>
>
>
>
> --
>
>
> *Sisu Xi, PhD Candidate*
>
> http://www.cse.wustl.edu/~xis/
> Department of Computer Science and Engineering
> Campus Box 1045
> Washington University in St. Louis
> One Brookings Drive
> St. Louis, MO 63130
>

-- 
CONFIDENTIALITY NOTICE
NOTICE: This message is intended for the use of the individual or entity to 
which it is addressed and may contain information that is confidential, 
privileged and exempt from disclosure under applicable law. If the reader 
of this message is not the intended recipient, you are hereby notified that 
any printing, copying, dissemination, distribution, disclosure or 
forwarding of this communication is strictly prohibited. If you have 
received this communication in error, please contact the sender immediately 
and delete it from your system. Thank You.

Re: hadoop multinode, only master node doing the work

Posted by Ravi Mutyala <ra...@hortonworks.com>.
Yes. As vinod pointed in that link, check yarn.resourcemanager.hostname on
slave node yarn-site.xml and put your master node hostname there.


On Sun, Jul 13, 2014 at 2:47 PM, Sisu Xi <xi...@gmail.com> wrote:

> Hi, Ravi:
>
> Thanks again for your help!
>
> The website only shows one node. a screenshot is attached.
>
> I restart the service, on master, it says:
> xisisu@master:/usr/local/hadoop$ jps
> 15862 Jps
> 15570 NodeManager
> 15461 ResourceManager
> 15149 DataNode
> 15025 NameNode
>
>  on slave, it says:
> xisisu@slave-01:/usr/local/hadoop$ jps
> 19752 DataNode
> 19859 NodeManager
> 20003 Jps
>
> I also checked the yarn log on the slave node, seems it cannot connect to
> the server,
> xisisu@slave-01:/usr/local/hadoop$ tail
> logs/yarn-xisisu-nodemanager-slave-01.log
> 2014-07-13 14:41:35,947 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 0 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
> 2014-07-13 14:41:36,947 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 1 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
> 2014-07-13 14:41:37,961 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 2 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
> 2014-07-13 14:41:38,962 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 3 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
> 2014-07-13 14:41:39,962 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 4 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
> 2014-07-13 14:41:40,963 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 5 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
> 2014-07-13 14:41:41,963 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 6 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
> 2014-07-13 14:41:42,963 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 7 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
> 2014-07-13 14:41:43,964 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 8 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
> 2014-07-13 14:41:44,964 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 9 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
>
>
> I googled and found this:
>
> http://hortonworks.com/community/forums/topic/retrying-connect-to-server-0-0-0-00-0-0-08030/
>
> Seems it is the configuration error?
>
> Thanks very much!
>
> Sisu
>
>
>
>
>
>
> On Sun, Jul 13, 2014 at 2:33 PM, Ravi Mutyala <ra...@hortonworks.com>
> wrote:
>
>> http://localhost:8088/cluster/nodes
>
>
>
>
>
> --
>
>
> *Sisu Xi, PhD Candidate*
>
> http://www.cse.wustl.edu/~xis/
> Department of Computer Science and Engineering
> Campus Box 1045
> Washington University in St. Louis
> One Brookings Drive
> St. Louis, MO 63130
>

-- 
CONFIDENTIALITY NOTICE
NOTICE: This message is intended for the use of the individual or entity to 
which it is addressed and may contain information that is confidential, 
privileged and exempt from disclosure under applicable law. If the reader 
of this message is not the intended recipient, you are hereby notified that 
any printing, copying, dissemination, distribution, disclosure or 
forwarding of this communication is strictly prohibited. If you have 
received this communication in error, please contact the sender immediately 
and delete it from your system. Thank You.

Re: hadoop multinode, only master node doing the work

Posted by Ravi Mutyala <ra...@hortonworks.com>.
Yes. As vinod pointed in that link, check yarn.resourcemanager.hostname on
slave node yarn-site.xml and put your master node hostname there.


On Sun, Jul 13, 2014 at 2:47 PM, Sisu Xi <xi...@gmail.com> wrote:

> Hi, Ravi:
>
> Thanks again for your help!
>
> The website only shows one node. a screenshot is attached.
>
> I restart the service, on master, it says:
> xisisu@master:/usr/local/hadoop$ jps
> 15862 Jps
> 15570 NodeManager
> 15461 ResourceManager
> 15149 DataNode
> 15025 NameNode
>
>  on slave, it says:
> xisisu@slave-01:/usr/local/hadoop$ jps
> 19752 DataNode
> 19859 NodeManager
> 20003 Jps
>
> I also checked the yarn log on the slave node, seems it cannot connect to
> the server,
> xisisu@slave-01:/usr/local/hadoop$ tail
> logs/yarn-xisisu-nodemanager-slave-01.log
> 2014-07-13 14:41:35,947 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 0 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
> 2014-07-13 14:41:36,947 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 1 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
> 2014-07-13 14:41:37,961 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 2 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
> 2014-07-13 14:41:38,962 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 3 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
> 2014-07-13 14:41:39,962 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 4 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
> 2014-07-13 14:41:40,963 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 5 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
> 2014-07-13 14:41:41,963 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 6 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
> 2014-07-13 14:41:42,963 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 7 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
> 2014-07-13 14:41:43,964 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 8 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
> 2014-07-13 14:41:44,964 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 9 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
>
>
> I googled and found this:
>
> http://hortonworks.com/community/forums/topic/retrying-connect-to-server-0-0-0-00-0-0-08030/
>
> Seems it is the configuration error?
>
> Thanks very much!
>
> Sisu
>
>
>
>
>
>
> On Sun, Jul 13, 2014 at 2:33 PM, Ravi Mutyala <ra...@hortonworks.com>
> wrote:
>
>> http://localhost:8088/cluster/nodes
>
>
>
>
>
> --
>
>
> *Sisu Xi, PhD Candidate*
>
> http://www.cse.wustl.edu/~xis/
> Department of Computer Science and Engineering
> Campus Box 1045
> Washington University in St. Louis
> One Brookings Drive
> St. Louis, MO 63130
>

-- 
CONFIDENTIALITY NOTICE
NOTICE: This message is intended for the use of the individual or entity to 
which it is addressed and may contain information that is confidential, 
privileged and exempt from disclosure under applicable law. If the reader 
of this message is not the intended recipient, you are hereby notified that 
any printing, copying, dissemination, distribution, disclosure or 
forwarding of this communication is strictly prohibited. If you have 
received this communication in error, please contact the sender immediately 
and delete it from your system. Thank You.

Re: hadoop multinode, only master node doing the work

Posted by Ravi Mutyala <ra...@hortonworks.com>.
Yes. As vinod pointed in that link, check yarn.resourcemanager.hostname on
slave node yarn-site.xml and put your master node hostname there.


On Sun, Jul 13, 2014 at 2:47 PM, Sisu Xi <xi...@gmail.com> wrote:

> Hi, Ravi:
>
> Thanks again for your help!
>
> The website only shows one node. a screenshot is attached.
>
> I restart the service, on master, it says:
> xisisu@master:/usr/local/hadoop$ jps
> 15862 Jps
> 15570 NodeManager
> 15461 ResourceManager
> 15149 DataNode
> 15025 NameNode
>
>  on slave, it says:
> xisisu@slave-01:/usr/local/hadoop$ jps
> 19752 DataNode
> 19859 NodeManager
> 20003 Jps
>
> I also checked the yarn log on the slave node, seems it cannot connect to
> the server,
> xisisu@slave-01:/usr/local/hadoop$ tail
> logs/yarn-xisisu-nodemanager-slave-01.log
> 2014-07-13 14:41:35,947 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 0 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
> 2014-07-13 14:41:36,947 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 1 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
> 2014-07-13 14:41:37,961 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 2 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
> 2014-07-13 14:41:38,962 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 3 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
> 2014-07-13 14:41:39,962 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 4 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
> 2014-07-13 14:41:40,963 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 5 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
> 2014-07-13 14:41:41,963 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 6 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
> 2014-07-13 14:41:42,963 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 7 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
> 2014-07-13 14:41:43,964 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 8 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
> 2014-07-13 14:41:44,964 INFO org.apache.hadoop.ipc.Client: Retrying
> connect to server: 0.0.0.0/0.0.0.0:8031. Already tried 9 time(s); retry
> policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1
> SECONDS)
>
>
> I googled and found this:
>
> http://hortonworks.com/community/forums/topic/retrying-connect-to-server-0-0-0-00-0-0-08030/
>
> Seems it is the configuration error?
>
> Thanks very much!
>
> Sisu
>
>
>
>
>
>
> On Sun, Jul 13, 2014 at 2:33 PM, Ravi Mutyala <ra...@hortonworks.com>
> wrote:
>
>> http://localhost:8088/cluster/nodes
>
>
>
>
>
> --
>
>
> *Sisu Xi, PhD Candidate*
>
> http://www.cse.wustl.edu/~xis/
> Department of Computer Science and Engineering
> Campus Box 1045
> Washington University in St. Louis
> One Brookings Drive
> St. Louis, MO 63130
>

-- 
CONFIDENTIALITY NOTICE
NOTICE: This message is intended for the use of the individual or entity to 
which it is addressed and may contain information that is confidential, 
privileged and exempt from disclosure under applicable law. If the reader 
of this message is not the intended recipient, you are hereby notified that 
any printing, copying, dissemination, distribution, disclosure or 
forwarding of this communication is strictly prohibited. If you have 
received this communication in error, please contact the sender immediately 
and delete it from your system. Thank You.

Re: hadoop multinode, only master node doing the work

Posted by Sisu Xi <xi...@gmail.com>.
Hi, Ravi:

Thanks again for your help!

The website only shows one node. a screenshot is attached.

I restart the service, on master, it says:
xisisu@master:/usr/local/hadoop$ jps
15862 Jps
15570 NodeManager
15461 ResourceManager
15149 DataNode
15025 NameNode

on slave, it says:
xisisu@slave-01:/usr/local/hadoop$ jps
19752 DataNode
19859 NodeManager
20003 Jps

I also checked the yarn log on the slave node, seems it cannot connect to
the server,
xisisu@slave-01:/usr/local/hadoop$ tail
logs/yarn-xisisu-nodemanager-slave-01.log
2014-07-13 14:41:35,947 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 0 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
2014-07-13 14:41:36,947 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 1 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
2014-07-13 14:41:37,961 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 2 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
2014-07-13 14:41:38,962 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 3 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
2014-07-13 14:41:39,962 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 4 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
2014-07-13 14:41:40,963 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 5 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
2014-07-13 14:41:41,963 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 6 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
2014-07-13 14:41:42,963 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 7 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
2014-07-13 14:41:43,964 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 8 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
2014-07-13 14:41:44,964 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 9 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)


I googled and found this:
http://hortonworks.com/community/forums/topic/retrying-connect-to-server-0-0-0-00-0-0-08030/

Seems it is the configuration error?

Thanks very much!

Sisu






On Sun, Jul 13, 2014 at 2:33 PM, Ravi Mutyala <ra...@hortonworks.com> wrote:

> http://localhost:8088/cluster/nodes





-- 


*Sisu Xi, PhD Candidate*

http://www.cse.wustl.edu/~xis/
Department of Computer Science and Engineering
Campus Box 1045
Washington University in St. Louis
One Brookings Drive
St. Louis, MO 63130

Re: hadoop multinode, only master node doing the work

Posted by Sisu Xi <xi...@gmail.com>.
Hi, Ravi:

Thanks again for your help!

The website only shows one node. a screenshot is attached.

I restart the service, on master, it says:
xisisu@master:/usr/local/hadoop$ jps
15862 Jps
15570 NodeManager
15461 ResourceManager
15149 DataNode
15025 NameNode

on slave, it says:
xisisu@slave-01:/usr/local/hadoop$ jps
19752 DataNode
19859 NodeManager
20003 Jps

I also checked the yarn log on the slave node, seems it cannot connect to
the server,
xisisu@slave-01:/usr/local/hadoop$ tail
logs/yarn-xisisu-nodemanager-slave-01.log
2014-07-13 14:41:35,947 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 0 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
2014-07-13 14:41:36,947 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 1 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
2014-07-13 14:41:37,961 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 2 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
2014-07-13 14:41:38,962 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 3 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
2014-07-13 14:41:39,962 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 4 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
2014-07-13 14:41:40,963 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 5 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
2014-07-13 14:41:41,963 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 6 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
2014-07-13 14:41:42,963 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 7 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
2014-07-13 14:41:43,964 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 8 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
2014-07-13 14:41:44,964 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 9 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)


I googled and found this:
http://hortonworks.com/community/forums/topic/retrying-connect-to-server-0-0-0-00-0-0-08030/

Seems it is the configuration error?

Thanks very much!

Sisu






On Sun, Jul 13, 2014 at 2:33 PM, Ravi Mutyala <ra...@hortonworks.com> wrote:

> http://localhost:8088/cluster/nodes





-- 


*Sisu Xi, PhD Candidate*

http://www.cse.wustl.edu/~xis/
Department of Computer Science and Engineering
Campus Box 1045
Washington University in St. Louis
One Brookings Drive
St. Louis, MO 63130

Re: hadoop multinode, only master node doing the work

Posted by Sisu Xi <xi...@gmail.com>.
Hi, Ravi:

Thanks again for your help!

The website only shows one node. a screenshot is attached.

I restart the service, on master, it says:
xisisu@master:/usr/local/hadoop$ jps
15862 Jps
15570 NodeManager
15461 ResourceManager
15149 DataNode
15025 NameNode

on slave, it says:
xisisu@slave-01:/usr/local/hadoop$ jps
19752 DataNode
19859 NodeManager
20003 Jps

I also checked the yarn log on the slave node, seems it cannot connect to
the server,
xisisu@slave-01:/usr/local/hadoop$ tail
logs/yarn-xisisu-nodemanager-slave-01.log
2014-07-13 14:41:35,947 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 0 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
2014-07-13 14:41:36,947 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 1 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
2014-07-13 14:41:37,961 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 2 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
2014-07-13 14:41:38,962 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 3 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
2014-07-13 14:41:39,962 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 4 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
2014-07-13 14:41:40,963 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 5 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
2014-07-13 14:41:41,963 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 6 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
2014-07-13 14:41:42,963 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 7 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
2014-07-13 14:41:43,964 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 8 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
2014-07-13 14:41:44,964 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 9 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)


I googled and found this:
http://hortonworks.com/community/forums/topic/retrying-connect-to-server-0-0-0-00-0-0-08030/

Seems it is the configuration error?

Thanks very much!

Sisu






On Sun, Jul 13, 2014 at 2:33 PM, Ravi Mutyala <ra...@hortonworks.com> wrote:

> http://localhost:8088/cluster/nodes





-- 


*Sisu Xi, PhD Candidate*

http://www.cse.wustl.edu/~xis/
Department of Computer Science and Engineering
Campus Box 1045
Washington University in St. Louis
One Brookings Drive
St. Louis, MO 63130

Re: hadoop multinode, only master node doing the work

Posted by Sisu Xi <xi...@gmail.com>.
Hi, Ravi:

Thanks again for your help!

The website only shows one node. a screenshot is attached.

I restart the service, on master, it says:
xisisu@master:/usr/local/hadoop$ jps
15862 Jps
15570 NodeManager
15461 ResourceManager
15149 DataNode
15025 NameNode

on slave, it says:
xisisu@slave-01:/usr/local/hadoop$ jps
19752 DataNode
19859 NodeManager
20003 Jps

I also checked the yarn log on the slave node, seems it cannot connect to
the server,
xisisu@slave-01:/usr/local/hadoop$ tail
logs/yarn-xisisu-nodemanager-slave-01.log
2014-07-13 14:41:35,947 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 0 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
2014-07-13 14:41:36,947 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 1 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
2014-07-13 14:41:37,961 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 2 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
2014-07-13 14:41:38,962 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 3 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
2014-07-13 14:41:39,962 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 4 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
2014-07-13 14:41:40,963 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 5 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
2014-07-13 14:41:41,963 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 6 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
2014-07-13 14:41:42,963 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 7 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
2014-07-13 14:41:43,964 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 8 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
2014-07-13 14:41:44,964 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: 0.0.0.0/0.0.0.0:8031. Already tried 9 time(s); retry policy is
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)


I googled and found this:
http://hortonworks.com/community/forums/topic/retrying-connect-to-server-0-0-0-00-0-0-08030/

Seems it is the configuration error?

Thanks very much!

Sisu






On Sun, Jul 13, 2014 at 2:33 PM, Ravi Mutyala <ra...@hortonworks.com> wrote:

> http://localhost:8088/cluster/nodes





-- 


*Sisu Xi, PhD Candidate*

http://www.cse.wustl.edu/~xis/
Department of Computer Science and Engineering
Campus Box 1045
Washington University in St. Louis
One Brookings Drive
St. Louis, MO 63130

Re: hadoop multinode, only master node doing the work

Posted by Ravi Mutyala <ra...@hortonworks.com>.
http://localhost:8088/cluster/nodes and see if you have nodemanagers on all
nodes (or just the master). If nodemanager on slaves is not running, try
starting it.

And look in /var/log/hadoop-yarn/yarn/yarn-yarn-nodemanager*.log (or
wherever your yarn logs are configured) to see if there are any errors
starting nodemanager.

-- 
CONFIDENTIALITY NOTICE
NOTICE: This message is intended for the use of the individual or entity to 
which it is addressed and may contain information that is confidential, 
privileged and exempt from disclosure under applicable law. If the reader 
of this message is not the intended recipient, you are hereby notified that 
any printing, copying, dissemination, distribution, disclosure or 
forwarding of this communication is strictly prohibited. If you have 
received this communication in error, please contact the sender immediately 
and delete it from your system. Thank You.

Re: hadoop multinode, only master node doing the work

Posted by Ravi Mutyala <ra...@hortonworks.com>.
http://localhost:8088/cluster/nodes and see if you have nodemanagers on all
nodes (or just the master). If nodemanager on slaves is not running, try
starting it.

And look in /var/log/hadoop-yarn/yarn/yarn-yarn-nodemanager*.log (or
wherever your yarn logs are configured) to see if there are any errors
starting nodemanager.

-- 
CONFIDENTIALITY NOTICE
NOTICE: This message is intended for the use of the individual or entity to 
which it is addressed and may contain information that is confidential, 
privileged and exempt from disclosure under applicable law. If the reader 
of this message is not the intended recipient, you are hereby notified that 
any printing, copying, dissemination, distribution, disclosure or 
forwarding of this communication is strictly prohibited. If you have 
received this communication in error, please contact the sender immediately 
and delete it from your system. Thank You.

Re: hadoop multinode, only master node doing the work

Posted by Ravi Mutyala <ra...@hortonworks.com>.
http://localhost:8088/cluster/nodes and see if you have nodemanagers on all
nodes (or just the master). If nodemanager on slaves is not running, try
starting it.

And look in /var/log/hadoop-yarn/yarn/yarn-yarn-nodemanager*.log (or
wherever your yarn logs are configured) to see if there are any errors
starting nodemanager.

-- 
CONFIDENTIALITY NOTICE
NOTICE: This message is intended for the use of the individual or entity to 
which it is addressed and may contain information that is confidential, 
privileged and exempt from disclosure under applicable law. If the reader 
of this message is not the intended recipient, you are hereby notified that 
any printing, copying, dissemination, distribution, disclosure or 
forwarding of this communication is strictly prohibited. If you have 
received this communication in error, please contact the sender immediately 
and delete it from your system. Thank You.

Re: hadoop multinode, only master node doing the work

Posted by Ravi Mutyala <ra...@hortonworks.com>.
http://localhost:8088/cluster/nodes and see if you have nodemanagers on all
nodes (or just the master). If nodemanager on slaves is not running, try
starting it.

And look in /var/log/hadoop-yarn/yarn/yarn-yarn-nodemanager*.log (or
wherever your yarn logs are configured) to see if there are any errors
starting nodemanager.

-- 
CONFIDENTIALITY NOTICE
NOTICE: This message is intended for the use of the individual or entity to 
which it is addressed and may contain information that is confidential, 
privileged and exempt from disclosure under applicable law. If the reader 
of this message is not the intended recipient, you are hereby notified that 
any printing, copying, dissemination, distribution, disclosure or 
forwarding of this communication is strictly prohibited. If you have 
received this communication in error, please contact the sender immediately 
and delete it from your system. Thank You.

Re: hadoop multinode, only master node doing the work

Posted by Sisu Xi <xi...@gmail.com>.
Hi, Ravi:

Yes! you are right! I did not realize there is a data dir on the slave node
as well. Thanks very much!
After deleting it and restart, now I can see the slave node.

However, when I ran the pi program (even from the slave node), still only
the master node is doing the work. The command I am using is:
xisisu@slave-01:/usr/local/hadoop$ hadoop jar
share/hadoop/mapreduce/hadoop-mapreduce-examples-2.2.0.jar pi 100 10000000

Attached is the output for hadoop dfsadmin -report

xisisu@slave-01:/usr/local/hadoop$ hadoop dfsadmin -report
DEPRECATED: Use of this script to execute hdfs command is deprecated.
Instead use the hdfs command for it.

OpenJDK 64-Bit Server VM warning: You have loaded library
/usr/local/hadoop/lib/native/libhadoop.so.1.0.0 which might have disabled
stack guard. The VM will try to fix the stack guard now.
It's highly recommended that you fix the library with 'execstack -c
<libfile>', or link it with '-z noexecstack'.
14/07/13 14:10:34 WARN util.NativeCodeLoader: Unable to load native-hadoop
library for your platform... using builtin-java classes where applicable
Configured Capacity: 28137644032 (26.21 GB)
Present Capacity: 16888445354 (15.73 GB)
DFS Remaining: 16887668736 (15.73 GB)
DFS Used: 776618 (758.42 KB)
DFS Used%: 0.00%
Under replicated blocks: 0
Blocks with corrupt replicas: 0
Missing blocks: 0

-------------------------------------------------
Datanodes available: 2 (2 total, 0 dead)

Live datanodes:
Name: 172.16.20.72:50010 (slave-01)
Hostname: slave-01
Decommission Status : Normal
Configured Capacity: 14068822016 (13.10 GB)
DFS Used: 388309 (379.21 KB)
Non DFS Used: 5623800619 (5.24 GB)
DFS Remaining: 8444633088 (7.86 GB)
DFS Used%: 0.00%
DFS Remaining%: 60.02%
Last contact: Sun Jul 13 14:10:34 CDT 2014


Name: 172.16.20.170:50010 (master)
Hostname: master
Decommission Status : Normal
Configured Capacity: 14068822016 (13.10 GB)
DFS Used: 388309 (379.21 KB)
Non DFS Used: 5625398059 (5.24 GB)
DFS Remaining: 8443035648 (7.86 GB)
DFS Used%: 0.00%
DFS Remaining%: 60.01%
Last contact: Sun Jul 13 14:10:34 CDT 2014



Sisu




On Sun, Jul 13, 2014 at 2:03 PM, Ravi Mutyala <ra...@hortonworks.com> wrote:

> Try cleaning up /home/xisisu/mydata/hdfs/datanode on the slave nodes
> andvstart  'datanode' on that again. That should work. If it doesn't, send
> the log message.
>
> Also see if nodemanagers are running on slaves nodes. If everything is
> running on master, then its not just HDFS issue, its also yarn nodemanagers
> not running.
>
> CONFIDENTIALITY NOTICE
> NOTICE: This message is intended for the use of the individual or entity
> to which it is addressed and may contain information that is confidential,
> privileged and exempt from disclosure under applicable law. If the reader
> of this message is not the intended recipient, you are hereby notified that
> any printing, copying, dissemination, distribution, disclosure or
> forwarding of this communication is strictly prohibited. If you have
> received this communication in error, please contact the sender immediately
> and delete it from your system. Thank You.




-- 


*Sisu Xi, PhD Candidate*

http://www.cse.wustl.edu/~xis/
Department of Computer Science and Engineering
Campus Box 1045
Washington University in St. Louis
One Brookings Drive
St. Louis, MO 63130

Re: hadoop multinode, only master node doing the work

Posted by Sisu Xi <xi...@gmail.com>.
Hi, Ravi:

Yes! you are right! I did not realize there is a data dir on the slave node
as well. Thanks very much!
After deleting it and restart, now I can see the slave node.

However, when I ran the pi program (even from the slave node), still only
the master node is doing the work. The command I am using is:
xisisu@slave-01:/usr/local/hadoop$ hadoop jar
share/hadoop/mapreduce/hadoop-mapreduce-examples-2.2.0.jar pi 100 10000000

Attached is the output for hadoop dfsadmin -report

xisisu@slave-01:/usr/local/hadoop$ hadoop dfsadmin -report
DEPRECATED: Use of this script to execute hdfs command is deprecated.
Instead use the hdfs command for it.

OpenJDK 64-Bit Server VM warning: You have loaded library
/usr/local/hadoop/lib/native/libhadoop.so.1.0.0 which might have disabled
stack guard. The VM will try to fix the stack guard now.
It's highly recommended that you fix the library with 'execstack -c
<libfile>', or link it with '-z noexecstack'.
14/07/13 14:10:34 WARN util.NativeCodeLoader: Unable to load native-hadoop
library for your platform... using builtin-java classes where applicable
Configured Capacity: 28137644032 (26.21 GB)
Present Capacity: 16888445354 (15.73 GB)
DFS Remaining: 16887668736 (15.73 GB)
DFS Used: 776618 (758.42 KB)
DFS Used%: 0.00%
Under replicated blocks: 0
Blocks with corrupt replicas: 0
Missing blocks: 0

-------------------------------------------------
Datanodes available: 2 (2 total, 0 dead)

Live datanodes:
Name: 172.16.20.72:50010 (slave-01)
Hostname: slave-01
Decommission Status : Normal
Configured Capacity: 14068822016 (13.10 GB)
DFS Used: 388309 (379.21 KB)
Non DFS Used: 5623800619 (5.24 GB)
DFS Remaining: 8444633088 (7.86 GB)
DFS Used%: 0.00%
DFS Remaining%: 60.02%
Last contact: Sun Jul 13 14:10:34 CDT 2014


Name: 172.16.20.170:50010 (master)
Hostname: master
Decommission Status : Normal
Configured Capacity: 14068822016 (13.10 GB)
DFS Used: 388309 (379.21 KB)
Non DFS Used: 5625398059 (5.24 GB)
DFS Remaining: 8443035648 (7.86 GB)
DFS Used%: 0.00%
DFS Remaining%: 60.01%
Last contact: Sun Jul 13 14:10:34 CDT 2014



Sisu




On Sun, Jul 13, 2014 at 2:03 PM, Ravi Mutyala <ra...@hortonworks.com> wrote:

> Try cleaning up /home/xisisu/mydata/hdfs/datanode on the slave nodes
> andvstart  'datanode' on that again. That should work. If it doesn't, send
> the log message.
>
> Also see if nodemanagers are running on slaves nodes. If everything is
> running on master, then its not just HDFS issue, its also yarn nodemanagers
> not running.
>
> CONFIDENTIALITY NOTICE
> NOTICE: This message is intended for the use of the individual or entity
> to which it is addressed and may contain information that is confidential,
> privileged and exempt from disclosure under applicable law. If the reader
> of this message is not the intended recipient, you are hereby notified that
> any printing, copying, dissemination, distribution, disclosure or
> forwarding of this communication is strictly prohibited. If you have
> received this communication in error, please contact the sender immediately
> and delete it from your system. Thank You.




-- 


*Sisu Xi, PhD Candidate*

http://www.cse.wustl.edu/~xis/
Department of Computer Science and Engineering
Campus Box 1045
Washington University in St. Louis
One Brookings Drive
St. Louis, MO 63130

Re: hadoop multinode, only master node doing the work

Posted by Sisu Xi <xi...@gmail.com>.
Hi, Ravi:

Yes! you are right! I did not realize there is a data dir on the slave node
as well. Thanks very much!
After deleting it and restart, now I can see the slave node.

However, when I ran the pi program (even from the slave node), still only
the master node is doing the work. The command I am using is:
xisisu@slave-01:/usr/local/hadoop$ hadoop jar
share/hadoop/mapreduce/hadoop-mapreduce-examples-2.2.0.jar pi 100 10000000

Attached is the output for hadoop dfsadmin -report

xisisu@slave-01:/usr/local/hadoop$ hadoop dfsadmin -report
DEPRECATED: Use of this script to execute hdfs command is deprecated.
Instead use the hdfs command for it.

OpenJDK 64-Bit Server VM warning: You have loaded library
/usr/local/hadoop/lib/native/libhadoop.so.1.0.0 which might have disabled
stack guard. The VM will try to fix the stack guard now.
It's highly recommended that you fix the library with 'execstack -c
<libfile>', or link it with '-z noexecstack'.
14/07/13 14:10:34 WARN util.NativeCodeLoader: Unable to load native-hadoop
library for your platform... using builtin-java classes where applicable
Configured Capacity: 28137644032 (26.21 GB)
Present Capacity: 16888445354 (15.73 GB)
DFS Remaining: 16887668736 (15.73 GB)
DFS Used: 776618 (758.42 KB)
DFS Used%: 0.00%
Under replicated blocks: 0
Blocks with corrupt replicas: 0
Missing blocks: 0

-------------------------------------------------
Datanodes available: 2 (2 total, 0 dead)

Live datanodes:
Name: 172.16.20.72:50010 (slave-01)
Hostname: slave-01
Decommission Status : Normal
Configured Capacity: 14068822016 (13.10 GB)
DFS Used: 388309 (379.21 KB)
Non DFS Used: 5623800619 (5.24 GB)
DFS Remaining: 8444633088 (7.86 GB)
DFS Used%: 0.00%
DFS Remaining%: 60.02%
Last contact: Sun Jul 13 14:10:34 CDT 2014


Name: 172.16.20.170:50010 (master)
Hostname: master
Decommission Status : Normal
Configured Capacity: 14068822016 (13.10 GB)
DFS Used: 388309 (379.21 KB)
Non DFS Used: 5625398059 (5.24 GB)
DFS Remaining: 8443035648 (7.86 GB)
DFS Used%: 0.00%
DFS Remaining%: 60.01%
Last contact: Sun Jul 13 14:10:34 CDT 2014



Sisu




On Sun, Jul 13, 2014 at 2:03 PM, Ravi Mutyala <ra...@hortonworks.com> wrote:

> Try cleaning up /home/xisisu/mydata/hdfs/datanode on the slave nodes
> andvstart  'datanode' on that again. That should work. If it doesn't, send
> the log message.
>
> Also see if nodemanagers are running on slaves nodes. If everything is
> running on master, then its not just HDFS issue, its also yarn nodemanagers
> not running.
>
> CONFIDENTIALITY NOTICE
> NOTICE: This message is intended for the use of the individual or entity
> to which it is addressed and may contain information that is confidential,
> privileged and exempt from disclosure under applicable law. If the reader
> of this message is not the intended recipient, you are hereby notified that
> any printing, copying, dissemination, distribution, disclosure or
> forwarding of this communication is strictly prohibited. If you have
> received this communication in error, please contact the sender immediately
> and delete it from your system. Thank You.




-- 


*Sisu Xi, PhD Candidate*

http://www.cse.wustl.edu/~xis/
Department of Computer Science and Engineering
Campus Box 1045
Washington University in St. Louis
One Brookings Drive
St. Louis, MO 63130

Re: hadoop multinode, only master node doing the work

Posted by Sisu Xi <xi...@gmail.com>.
Hi, Ravi:

Yes! you are right! I did not realize there is a data dir on the slave node
as well. Thanks very much!
After deleting it and restart, now I can see the slave node.

However, when I ran the pi program (even from the slave node), still only
the master node is doing the work. The command I am using is:
xisisu@slave-01:/usr/local/hadoop$ hadoop jar
share/hadoop/mapreduce/hadoop-mapreduce-examples-2.2.0.jar pi 100 10000000

Attached is the output for hadoop dfsadmin -report

xisisu@slave-01:/usr/local/hadoop$ hadoop dfsadmin -report
DEPRECATED: Use of this script to execute hdfs command is deprecated.
Instead use the hdfs command for it.

OpenJDK 64-Bit Server VM warning: You have loaded library
/usr/local/hadoop/lib/native/libhadoop.so.1.0.0 which might have disabled
stack guard. The VM will try to fix the stack guard now.
It's highly recommended that you fix the library with 'execstack -c
<libfile>', or link it with '-z noexecstack'.
14/07/13 14:10:34 WARN util.NativeCodeLoader: Unable to load native-hadoop
library for your platform... using builtin-java classes where applicable
Configured Capacity: 28137644032 (26.21 GB)
Present Capacity: 16888445354 (15.73 GB)
DFS Remaining: 16887668736 (15.73 GB)
DFS Used: 776618 (758.42 KB)
DFS Used%: 0.00%
Under replicated blocks: 0
Blocks with corrupt replicas: 0
Missing blocks: 0

-------------------------------------------------
Datanodes available: 2 (2 total, 0 dead)

Live datanodes:
Name: 172.16.20.72:50010 (slave-01)
Hostname: slave-01
Decommission Status : Normal
Configured Capacity: 14068822016 (13.10 GB)
DFS Used: 388309 (379.21 KB)
Non DFS Used: 5623800619 (5.24 GB)
DFS Remaining: 8444633088 (7.86 GB)
DFS Used%: 0.00%
DFS Remaining%: 60.02%
Last contact: Sun Jul 13 14:10:34 CDT 2014


Name: 172.16.20.170:50010 (master)
Hostname: master
Decommission Status : Normal
Configured Capacity: 14068822016 (13.10 GB)
DFS Used: 388309 (379.21 KB)
Non DFS Used: 5625398059 (5.24 GB)
DFS Remaining: 8443035648 (7.86 GB)
DFS Used%: 0.00%
DFS Remaining%: 60.01%
Last contact: Sun Jul 13 14:10:34 CDT 2014



Sisu




On Sun, Jul 13, 2014 at 2:03 PM, Ravi Mutyala <ra...@hortonworks.com> wrote:

> Try cleaning up /home/xisisu/mydata/hdfs/datanode on the slave nodes
> andvstart  'datanode' on that again. That should work. If it doesn't, send
> the log message.
>
> Also see if nodemanagers are running on slaves nodes. If everything is
> running on master, then its not just HDFS issue, its also yarn nodemanagers
> not running.
>
> CONFIDENTIALITY NOTICE
> NOTICE: This message is intended for the use of the individual or entity
> to which it is addressed and may contain information that is confidential,
> privileged and exempt from disclosure under applicable law. If the reader
> of this message is not the intended recipient, you are hereby notified that
> any printing, copying, dissemination, distribution, disclosure or
> forwarding of this communication is strictly prohibited. If you have
> received this communication in error, please contact the sender immediately
> and delete it from your system. Thank You.




-- 


*Sisu Xi, PhD Candidate*

http://www.cse.wustl.edu/~xis/
Department of Computer Science and Engineering
Campus Box 1045
Washington University in St. Louis
One Brookings Drive
St. Louis, MO 63130

Re: hadoop multinode, only master node doing the work

Posted by Ravi Mutyala <ra...@hortonworks.com>.
Try cleaning up /home/xisisu/mydata/hdfs/datanode on the slave nodes
andvstart  'datanode' on that again. That should work. If it doesn't, send
the log message.

Also see if nodemanagers are running on slaves nodes. If everything is
running on master, then its not just HDFS issue, its also yarn nodemanagers
not running.

-- 
CONFIDENTIALITY NOTICE
NOTICE: This message is intended for the use of the individual or entity to 
which it is addressed and may contain information that is confidential, 
privileged and exempt from disclosure under applicable law. If the reader 
of this message is not the intended recipient, you are hereby notified that 
any printing, copying, dissemination, distribution, disclosure or 
forwarding of this communication is strictly prohibited. If you have 
received this communication in error, please contact the sender immediately 
and delete it from your system. Thank You.

Re: hadoop multinode, only master node doing the work

Posted by Ravi Mutyala <ra...@hortonworks.com>.
Try cleaning up /home/xisisu/mydata/hdfs/datanode on the slave nodes
andvstart  'datanode' on that again. That should work. If it doesn't, send
the log message.

Also see if nodemanagers are running on slaves nodes. If everything is
running on master, then its not just HDFS issue, its also yarn nodemanagers
not running.

-- 
CONFIDENTIALITY NOTICE
NOTICE: This message is intended for the use of the individual or entity to 
which it is addressed and may contain information that is confidential, 
privileged and exempt from disclosure under applicable law. If the reader 
of this message is not the intended recipient, you are hereby notified that 
any printing, copying, dissemination, distribution, disclosure or 
forwarding of this communication is strictly prohibited. If you have 
received this communication in error, please contact the sender immediately 
and delete it from your system. Thank You.

Re: hadoop multinode, only master node doing the work

Posted by Ravi Mutyala <ra...@hortonworks.com>.
Try cleaning up /home/xisisu/mydata/hdfs/datanode on the slave nodes
andvstart  'datanode' on that again. That should work. If it doesn't, send
the log message.

Also see if nodemanagers are running on slaves nodes. If everything is
running on master, then its not just HDFS issue, its also yarn nodemanagers
not running.

-- 
CONFIDENTIALITY NOTICE
NOTICE: This message is intended for the use of the individual or entity to 
which it is addressed and may contain information that is confidential, 
privileged and exempt from disclosure under applicable law. If the reader 
of this message is not the intended recipient, you are hereby notified that 
any printing, copying, dissemination, distribution, disclosure or 
forwarding of this communication is strictly prohibited. If you have 
received this communication in error, please contact the sender immediately 
and delete it from your system. Thank You.

Re: hadoop multinode, only master node doing the work

Posted by Ravi Mutyala <ra...@hortonworks.com>.
Try cleaning up /home/xisisu/mydata/hdfs/datanode on the slave nodes
andvstart  'datanode' on that again. That should work. If it doesn't, send
the log message.

Also see if nodemanagers are running on slaves nodes. If everything is
running on master, then its not just HDFS issue, its also yarn nodemanagers
not running.

-- 
CONFIDENTIALITY NOTICE
NOTICE: This message is intended for the use of the individual or entity to 
which it is addressed and may contain information that is confidential, 
privileged and exempt from disclosure under applicable law. If the reader 
of this message is not the intended recipient, you are hereby notified that 
any printing, copying, dissemination, distribution, disclosure or 
forwarding of this communication is strictly prohibited. If you have 
received this communication in error, please contact the sender immediately 
and delete it from your system. Thank You.

Re: hadoop multinode, only master node doing the work

Posted by Sisu Xi <xi...@gmail.com>.
I checked the log on the slave node,

it says:

2014-07-13 13:35:37,290 FATAL
org.apache.hadoop.hdfs.server.datanode.DataNode: Initialization failed for
block pool Block pool BP-1408390155-172.16.20.170-1405276491305 (storage id
DS-111686822-127.0.1.1-50010-1405200560127) service to master/
172.16.20.170:54310
java.io.IOException: Incompatible clusterIDs in
/home/xisisu/mydata/hdfs/datanode: namenode clusterID =
CID-5fe42d16-b0cc-4104-b04c-8b4561e907dd; datanode clusterID =
CID-95916f4d-89f1-47e4-a92e-7b88ce97ed8b
        at
org.apache.hadoop.hdfs.server.datanode.DataStorage.doTransition(DataStorage.java:391)
        at
org.apache.hadoop.hdfs.server.datanode.DataStorage.recoverTransitionRead(DataStorage.java:191)
        at
org.apache.hadoop.hdfs.server.datanode.DataStorage.recoverTransitionRead(DataStorage.java:219)
        at
org.apache.hadoop.hdfs.server.datanode.DataNode.initStorage(DataNode.java:837)
        at
org.apache.hadoop.hdfs.server.datanode.DataNode.initBlockPool(DataNode.java:808)
        at
org.apache.hadoop.hdfs.server.datanode.BPOfferService.verifyAndSetNamespaceInfo(BPOfferService.java:280)
        at
org.apache.hadoop.hdfs.server.datanode.BPServiceActor.connectToNNAndHandshake(BPServiceActor.java:222)
        at
org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:664)
        at java.lang.Thread.run(Thread.java:744)
2014-07-13 13:35:37,292 WARN
org.apache.hadoop.hdfs.server.datanode.DataNode: Ending block pool service
for: Block pool BP-1408390155-172.16.20.170-1405276491305 (storage id
DS-111686822-127.0.1.1-50010-1405200560127) service to master/
172.16.20.170:54310
2014-07-13 13:35:37,392 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Removed Block pool
BP-1408390155-172.16.20.170-1405276491305 (storage id
DS-111686822-127.0.1.1-50010-1405200560127)
2014-07-13 13:35:39,392 WARN
org.apache.hadoop.hdfs.server.datanode.DataNode: Exiting Datanode
2014-07-13 13:35:39,394 INFO org.apache.hadoop.util.ExitUtil: Exiting with
status 0
2014-07-13 13:35:39,395 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: SHUTDOWN_MSG:
/************************************************************
SHUTDOWN_MSG: Shutting down DataNode at slave-01/127.0.1.1
************************************************************/


I did some google search, it says the problem is with the data dir on the
master node.
I already removed the whole data dir, formated the namenode, and restarted
the hadoop.
But still the same problem.

Sisu




On Sun, Jul 13, 2014 at 12:53 PM, Sisu Xi <xi...@gmail.com> wrote:

> Hi, Sam:
>
> Thanks for your help! You are right, seems only one node is running.
> Here is the output:
>
> xisisu@slave-01:/usr/local/hadoop$ hadoop dfsadmin -report
> DEPRECATED: Use of this script to execute hdfs command is deprecated.
> Instead use the hdfs command for it.
>
> OpenJDK 64-Bit Server VM warning: You have loaded library
> /usr/local/hadoop/lib/native/libhadoop.so.1.0.0 which might have disabled
> stack guard. The VM will try to fix the stack guard now.
> It's highly recommended that you fix the library with 'execstack -c
> <libfile>', or link it with '-z noexecstack'.
> 14/07/13 12:50:14 WARN util.NativeCodeLoader: Unable to load native-hadoop
> library for your platform... using builtin-java classes where applicable
> Configured Capacity: 14068822016 (13.10 GB)
> Present Capacity: 8425232425 (7.85 GB)
> DFS Remaining: 8423534592 (7.85 GB)
> DFS Used: 1697833 (1.62 MB)
> DFS Used%: 0.02%
> Under replicated blocks: 21
> Blocks with corrupt replicas: 0
> Missing blocks: 0
>
> -------------------------------------------------
> Datanodes available: 1 (1 total, 0 dead)
>
> Live datanodes:
> Name: 172.16.20.170:50010 (master)
> Hostname: master
> Decommission Status : Normal
> Configured Capacity: 14068822016 (13.10 GB)
> DFS Used: 1697833 (1.62 MB)
> Non DFS Used: 5643589591 (5.26 GB)
> DFS Remaining: 8423534592 (7.85 GB)
> DFS Used%: 0.01%
> DFS Remaining%: 59.87%
> Last contact: Sun Jul 13 12:50:13 CDT 2014
>
>
> I get the same output when I run the command on the slave node.
>
>
> Is there anything I am missing in the config file?
>
> Thanks very much!
>
> Sisu
>
>
>
> On Sun, Jul 13, 2014 at 1:36 AM, Kilaru, Sambaiah <
> Sambaiah_Kilaru@intuit.com> wrote:
>
>>  HI Sisu Xi,
>>
>>  On the master node can you check
>> Hadoop dfsadmin –report
>> And listing all the slave nodes or you can check master URL and it should
>> all datanodes listed as slave nodes.
>> Check for RM UI and slave nodes listed there also.
>>
>>  Thanks,
>> Sam
>>
>>   From: Sisu Xi <xi...@gmail.com>
>> Reply-To: "user@hadoop.apache.org" <us...@hadoop.apache.org>
>> Date: Sunday, July 13, 2014 at 11:28 AM
>> To: "user@hadoop.apache.org" <us...@hadoop.apache.org>
>> Subject: hadoop multinode, only master node doing the work
>>
>>   Hi, all:
>>
>>  I am new to hadoop. I followed the tutorial on
>>
>> http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-multi-node-cluster/
>>
>>  and installed hadoop 2.2.0 on two 4-core Ubuntu 12.04 machines.
>>
>>  I can start the pi program, however, only the master node is doing the
>> work (I checked the top on each machine).
>> Seems the two nodes are configured correctly, because I can start  the
>> program in the slave node as well, and still only the master node is doing
>> the actual work.
>> I have tried different number of mappers for the pi program, and the
>> results is the same.
>>
>>  Is there anything else I can check?
>>
>>  In the end is my configure file on each host.
>>
>>  Thanks very much!
>>
>>  Sisu
>>
>>  ---------yarn-site.xml-------
>>
>>  <property>
>>   <name>yarn.nodemanager.aux-services</name>
>>   <value>mapreduce_shuffle</value>
>> </property>
>>
>>
>>  <property>
>>   <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name>
>>   <value>org.apache.hadoop.mapred.ShuffleHandler</value>
>> </property>
>>
>>
>>  <property>
>>   <name>yarn.resourcemanager.address</name>
>>   <value>master:8032</value>
>> </property>
>>
>>  ---------------hdfs-site.xml--------------------
>>
>>  <property>
>>   <name>dfs.replication</name>
>>   <value>2</value>
>> </property>
>>
>>
>>  <property>
>>   <name>dfs.namenode.name.dir</name>
>>   <value>file:/home/xisisu/mydata/hdfs/namenode</value>
>> </property>
>>
>>
>>  <property>
>>   <name>dfs.datanode.data.dir</name>
>>   <value>file:/home/xisisu/mydata/hdfs/datanode</value>
>> </property>
>>
>>
>>  -------------core-site.xml-------------
>>
>>  <property>
>>  <name>fs.default.name</name>
>> <value>hdfs://master:9000</value>
>> </property>
>>
>>  ------------------mapred-site.xml-----------------
>>
>>  <property>
>>   <name>mapreduce.framework.name</name>
>>   <value>yarn</value>
>> </property>
>>
>>  <property>
>>   <name>mapred.job.tracker</name>
>>   <value>master:54311</value>
>>   <description>The host and port that the MapReduce job tracker runs
>>   at.  If "local", then jobs are run in-process as a single map
>>   and reduce task.
>>   </description>
>> </property>
>>
>>
>>
>>
>>  --
>>
>>
>>  *Sisu Xi, PhD Candidate*
>>
>> http://www.cse.wustl.edu/~xis/
>> Department of Computer Science and Engineering
>> Campus Box 1045
>> Washington University in St. Louis
>> One Brookings Drive
>> St. Louis, MO 63130
>>
>
>
>
> --
>
>
> *Sisu Xi, PhD Candidate*
>
> http://www.cse.wustl.edu/~xis/
> Department of Computer Science and Engineering
> Campus Box 1045
> Washington University in St. Louis
> One Brookings Drive
> St. Louis, MO 63130
>



-- 


*Sisu Xi, PhD Candidate*

http://www.cse.wustl.edu/~xis/
Department of Computer Science and Engineering
Campus Box 1045
Washington University in St. Louis
One Brookings Drive
St. Louis, MO 63130

Re: hadoop multinode, only master node doing the work

Posted by Sisu Xi <xi...@gmail.com>.
I checked the log on the slave node,

it says:

2014-07-13 13:35:37,290 FATAL
org.apache.hadoop.hdfs.server.datanode.DataNode: Initialization failed for
block pool Block pool BP-1408390155-172.16.20.170-1405276491305 (storage id
DS-111686822-127.0.1.1-50010-1405200560127) service to master/
172.16.20.170:54310
java.io.IOException: Incompatible clusterIDs in
/home/xisisu/mydata/hdfs/datanode: namenode clusterID =
CID-5fe42d16-b0cc-4104-b04c-8b4561e907dd; datanode clusterID =
CID-95916f4d-89f1-47e4-a92e-7b88ce97ed8b
        at
org.apache.hadoop.hdfs.server.datanode.DataStorage.doTransition(DataStorage.java:391)
        at
org.apache.hadoop.hdfs.server.datanode.DataStorage.recoverTransitionRead(DataStorage.java:191)
        at
org.apache.hadoop.hdfs.server.datanode.DataStorage.recoverTransitionRead(DataStorage.java:219)
        at
org.apache.hadoop.hdfs.server.datanode.DataNode.initStorage(DataNode.java:837)
        at
org.apache.hadoop.hdfs.server.datanode.DataNode.initBlockPool(DataNode.java:808)
        at
org.apache.hadoop.hdfs.server.datanode.BPOfferService.verifyAndSetNamespaceInfo(BPOfferService.java:280)
        at
org.apache.hadoop.hdfs.server.datanode.BPServiceActor.connectToNNAndHandshake(BPServiceActor.java:222)
        at
org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:664)
        at java.lang.Thread.run(Thread.java:744)
2014-07-13 13:35:37,292 WARN
org.apache.hadoop.hdfs.server.datanode.DataNode: Ending block pool service
for: Block pool BP-1408390155-172.16.20.170-1405276491305 (storage id
DS-111686822-127.0.1.1-50010-1405200560127) service to master/
172.16.20.170:54310
2014-07-13 13:35:37,392 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Removed Block pool
BP-1408390155-172.16.20.170-1405276491305 (storage id
DS-111686822-127.0.1.1-50010-1405200560127)
2014-07-13 13:35:39,392 WARN
org.apache.hadoop.hdfs.server.datanode.DataNode: Exiting Datanode
2014-07-13 13:35:39,394 INFO org.apache.hadoop.util.ExitUtil: Exiting with
status 0
2014-07-13 13:35:39,395 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: SHUTDOWN_MSG:
/************************************************************
SHUTDOWN_MSG: Shutting down DataNode at slave-01/127.0.1.1
************************************************************/


I did some google search, it says the problem is with the data dir on the
master node.
I already removed the whole data dir, formated the namenode, and restarted
the hadoop.
But still the same problem.

Sisu




On Sun, Jul 13, 2014 at 12:53 PM, Sisu Xi <xi...@gmail.com> wrote:

> Hi, Sam:
>
> Thanks for your help! You are right, seems only one node is running.
> Here is the output:
>
> xisisu@slave-01:/usr/local/hadoop$ hadoop dfsadmin -report
> DEPRECATED: Use of this script to execute hdfs command is deprecated.
> Instead use the hdfs command for it.
>
> OpenJDK 64-Bit Server VM warning: You have loaded library
> /usr/local/hadoop/lib/native/libhadoop.so.1.0.0 which might have disabled
> stack guard. The VM will try to fix the stack guard now.
> It's highly recommended that you fix the library with 'execstack -c
> <libfile>', or link it with '-z noexecstack'.
> 14/07/13 12:50:14 WARN util.NativeCodeLoader: Unable to load native-hadoop
> library for your platform... using builtin-java classes where applicable
> Configured Capacity: 14068822016 (13.10 GB)
> Present Capacity: 8425232425 (7.85 GB)
> DFS Remaining: 8423534592 (7.85 GB)
> DFS Used: 1697833 (1.62 MB)
> DFS Used%: 0.02%
> Under replicated blocks: 21
> Blocks with corrupt replicas: 0
> Missing blocks: 0
>
> -------------------------------------------------
> Datanodes available: 1 (1 total, 0 dead)
>
> Live datanodes:
> Name: 172.16.20.170:50010 (master)
> Hostname: master
> Decommission Status : Normal
> Configured Capacity: 14068822016 (13.10 GB)
> DFS Used: 1697833 (1.62 MB)
> Non DFS Used: 5643589591 (5.26 GB)
> DFS Remaining: 8423534592 (7.85 GB)
> DFS Used%: 0.01%
> DFS Remaining%: 59.87%
> Last contact: Sun Jul 13 12:50:13 CDT 2014
>
>
> I get the same output when I run the command on the slave node.
>
>
> Is there anything I am missing in the config file?
>
> Thanks very much!
>
> Sisu
>
>
>
> On Sun, Jul 13, 2014 at 1:36 AM, Kilaru, Sambaiah <
> Sambaiah_Kilaru@intuit.com> wrote:
>
>>  HI Sisu Xi,
>>
>>  On the master node can you check
>> Hadoop dfsadmin –report
>> And listing all the slave nodes or you can check master URL and it should
>> all datanodes listed as slave nodes.
>> Check for RM UI and slave nodes listed there also.
>>
>>  Thanks,
>> Sam
>>
>>   From: Sisu Xi <xi...@gmail.com>
>> Reply-To: "user@hadoop.apache.org" <us...@hadoop.apache.org>
>> Date: Sunday, July 13, 2014 at 11:28 AM
>> To: "user@hadoop.apache.org" <us...@hadoop.apache.org>
>> Subject: hadoop multinode, only master node doing the work
>>
>>   Hi, all:
>>
>>  I am new to hadoop. I followed the tutorial on
>>
>> http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-multi-node-cluster/
>>
>>  and installed hadoop 2.2.0 on two 4-core Ubuntu 12.04 machines.
>>
>>  I can start the pi program, however, only the master node is doing the
>> work (I checked the top on each machine).
>> Seems the two nodes are configured correctly, because I can start  the
>> program in the slave node as well, and still only the master node is doing
>> the actual work.
>> I have tried different number of mappers for the pi program, and the
>> results is the same.
>>
>>  Is there anything else I can check?
>>
>>  In the end is my configure file on each host.
>>
>>  Thanks very much!
>>
>>  Sisu
>>
>>  ---------yarn-site.xml-------
>>
>>  <property>
>>   <name>yarn.nodemanager.aux-services</name>
>>   <value>mapreduce_shuffle</value>
>> </property>
>>
>>
>>  <property>
>>   <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name>
>>   <value>org.apache.hadoop.mapred.ShuffleHandler</value>
>> </property>
>>
>>
>>  <property>
>>   <name>yarn.resourcemanager.address</name>
>>   <value>master:8032</value>
>> </property>
>>
>>  ---------------hdfs-site.xml--------------------
>>
>>  <property>
>>   <name>dfs.replication</name>
>>   <value>2</value>
>> </property>
>>
>>
>>  <property>
>>   <name>dfs.namenode.name.dir</name>
>>   <value>file:/home/xisisu/mydata/hdfs/namenode</value>
>> </property>
>>
>>
>>  <property>
>>   <name>dfs.datanode.data.dir</name>
>>   <value>file:/home/xisisu/mydata/hdfs/datanode</value>
>> </property>
>>
>>
>>  -------------core-site.xml-------------
>>
>>  <property>
>>  <name>fs.default.name</name>
>> <value>hdfs://master:9000</value>
>> </property>
>>
>>  ------------------mapred-site.xml-----------------
>>
>>  <property>
>>   <name>mapreduce.framework.name</name>
>>   <value>yarn</value>
>> </property>
>>
>>  <property>
>>   <name>mapred.job.tracker</name>
>>   <value>master:54311</value>
>>   <description>The host and port that the MapReduce job tracker runs
>>   at.  If "local", then jobs are run in-process as a single map
>>   and reduce task.
>>   </description>
>> </property>
>>
>>
>>
>>
>>  --
>>
>>
>>  *Sisu Xi, PhD Candidate*
>>
>> http://www.cse.wustl.edu/~xis/
>> Department of Computer Science and Engineering
>> Campus Box 1045
>> Washington University in St. Louis
>> One Brookings Drive
>> St. Louis, MO 63130
>>
>
>
>
> --
>
>
> *Sisu Xi, PhD Candidate*
>
> http://www.cse.wustl.edu/~xis/
> Department of Computer Science and Engineering
> Campus Box 1045
> Washington University in St. Louis
> One Brookings Drive
> St. Louis, MO 63130
>



-- 


*Sisu Xi, PhD Candidate*

http://www.cse.wustl.edu/~xis/
Department of Computer Science and Engineering
Campus Box 1045
Washington University in St. Louis
One Brookings Drive
St. Louis, MO 63130

Re: hadoop multinode, only master node doing the work

Posted by Sisu Xi <xi...@gmail.com>.
I checked the log on the slave node,

it says:

2014-07-13 13:35:37,290 FATAL
org.apache.hadoop.hdfs.server.datanode.DataNode: Initialization failed for
block pool Block pool BP-1408390155-172.16.20.170-1405276491305 (storage id
DS-111686822-127.0.1.1-50010-1405200560127) service to master/
172.16.20.170:54310
java.io.IOException: Incompatible clusterIDs in
/home/xisisu/mydata/hdfs/datanode: namenode clusterID =
CID-5fe42d16-b0cc-4104-b04c-8b4561e907dd; datanode clusterID =
CID-95916f4d-89f1-47e4-a92e-7b88ce97ed8b
        at
org.apache.hadoop.hdfs.server.datanode.DataStorage.doTransition(DataStorage.java:391)
        at
org.apache.hadoop.hdfs.server.datanode.DataStorage.recoverTransitionRead(DataStorage.java:191)
        at
org.apache.hadoop.hdfs.server.datanode.DataStorage.recoverTransitionRead(DataStorage.java:219)
        at
org.apache.hadoop.hdfs.server.datanode.DataNode.initStorage(DataNode.java:837)
        at
org.apache.hadoop.hdfs.server.datanode.DataNode.initBlockPool(DataNode.java:808)
        at
org.apache.hadoop.hdfs.server.datanode.BPOfferService.verifyAndSetNamespaceInfo(BPOfferService.java:280)
        at
org.apache.hadoop.hdfs.server.datanode.BPServiceActor.connectToNNAndHandshake(BPServiceActor.java:222)
        at
org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:664)
        at java.lang.Thread.run(Thread.java:744)
2014-07-13 13:35:37,292 WARN
org.apache.hadoop.hdfs.server.datanode.DataNode: Ending block pool service
for: Block pool BP-1408390155-172.16.20.170-1405276491305 (storage id
DS-111686822-127.0.1.1-50010-1405200560127) service to master/
172.16.20.170:54310
2014-07-13 13:35:37,392 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Removed Block pool
BP-1408390155-172.16.20.170-1405276491305 (storage id
DS-111686822-127.0.1.1-50010-1405200560127)
2014-07-13 13:35:39,392 WARN
org.apache.hadoop.hdfs.server.datanode.DataNode: Exiting Datanode
2014-07-13 13:35:39,394 INFO org.apache.hadoop.util.ExitUtil: Exiting with
status 0
2014-07-13 13:35:39,395 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: SHUTDOWN_MSG:
/************************************************************
SHUTDOWN_MSG: Shutting down DataNode at slave-01/127.0.1.1
************************************************************/


I did some google search, it says the problem is with the data dir on the
master node.
I already removed the whole data dir, formated the namenode, and restarted
the hadoop.
But still the same problem.

Sisu




On Sun, Jul 13, 2014 at 12:53 PM, Sisu Xi <xi...@gmail.com> wrote:

> Hi, Sam:
>
> Thanks for your help! You are right, seems only one node is running.
> Here is the output:
>
> xisisu@slave-01:/usr/local/hadoop$ hadoop dfsadmin -report
> DEPRECATED: Use of this script to execute hdfs command is deprecated.
> Instead use the hdfs command for it.
>
> OpenJDK 64-Bit Server VM warning: You have loaded library
> /usr/local/hadoop/lib/native/libhadoop.so.1.0.0 which might have disabled
> stack guard. The VM will try to fix the stack guard now.
> It's highly recommended that you fix the library with 'execstack -c
> <libfile>', or link it with '-z noexecstack'.
> 14/07/13 12:50:14 WARN util.NativeCodeLoader: Unable to load native-hadoop
> library for your platform... using builtin-java classes where applicable
> Configured Capacity: 14068822016 (13.10 GB)
> Present Capacity: 8425232425 (7.85 GB)
> DFS Remaining: 8423534592 (7.85 GB)
> DFS Used: 1697833 (1.62 MB)
> DFS Used%: 0.02%
> Under replicated blocks: 21
> Blocks with corrupt replicas: 0
> Missing blocks: 0
>
> -------------------------------------------------
> Datanodes available: 1 (1 total, 0 dead)
>
> Live datanodes:
> Name: 172.16.20.170:50010 (master)
> Hostname: master
> Decommission Status : Normal
> Configured Capacity: 14068822016 (13.10 GB)
> DFS Used: 1697833 (1.62 MB)
> Non DFS Used: 5643589591 (5.26 GB)
> DFS Remaining: 8423534592 (7.85 GB)
> DFS Used%: 0.01%
> DFS Remaining%: 59.87%
> Last contact: Sun Jul 13 12:50:13 CDT 2014
>
>
> I get the same output when I run the command on the slave node.
>
>
> Is there anything I am missing in the config file?
>
> Thanks very much!
>
> Sisu
>
>
>
> On Sun, Jul 13, 2014 at 1:36 AM, Kilaru, Sambaiah <
> Sambaiah_Kilaru@intuit.com> wrote:
>
>>  HI Sisu Xi,
>>
>>  On the master node can you check
>> Hadoop dfsadmin –report
>> And listing all the slave nodes or you can check master URL and it should
>> all datanodes listed as slave nodes.
>> Check for RM UI and slave nodes listed there also.
>>
>>  Thanks,
>> Sam
>>
>>   From: Sisu Xi <xi...@gmail.com>
>> Reply-To: "user@hadoop.apache.org" <us...@hadoop.apache.org>
>> Date: Sunday, July 13, 2014 at 11:28 AM
>> To: "user@hadoop.apache.org" <us...@hadoop.apache.org>
>> Subject: hadoop multinode, only master node doing the work
>>
>>   Hi, all:
>>
>>  I am new to hadoop. I followed the tutorial on
>>
>> http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-multi-node-cluster/
>>
>>  and installed hadoop 2.2.0 on two 4-core Ubuntu 12.04 machines.
>>
>>  I can start the pi program, however, only the master node is doing the
>> work (I checked the top on each machine).
>> Seems the two nodes are configured correctly, because I can start  the
>> program in the slave node as well, and still only the master node is doing
>> the actual work.
>> I have tried different number of mappers for the pi program, and the
>> results is the same.
>>
>>  Is there anything else I can check?
>>
>>  In the end is my configure file on each host.
>>
>>  Thanks very much!
>>
>>  Sisu
>>
>>  ---------yarn-site.xml-------
>>
>>  <property>
>>   <name>yarn.nodemanager.aux-services</name>
>>   <value>mapreduce_shuffle</value>
>> </property>
>>
>>
>>  <property>
>>   <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name>
>>   <value>org.apache.hadoop.mapred.ShuffleHandler</value>
>> </property>
>>
>>
>>  <property>
>>   <name>yarn.resourcemanager.address</name>
>>   <value>master:8032</value>
>> </property>
>>
>>  ---------------hdfs-site.xml--------------------
>>
>>  <property>
>>   <name>dfs.replication</name>
>>   <value>2</value>
>> </property>
>>
>>
>>  <property>
>>   <name>dfs.namenode.name.dir</name>
>>   <value>file:/home/xisisu/mydata/hdfs/namenode</value>
>> </property>
>>
>>
>>  <property>
>>   <name>dfs.datanode.data.dir</name>
>>   <value>file:/home/xisisu/mydata/hdfs/datanode</value>
>> </property>
>>
>>
>>  -------------core-site.xml-------------
>>
>>  <property>
>>  <name>fs.default.name</name>
>> <value>hdfs://master:9000</value>
>> </property>
>>
>>  ------------------mapred-site.xml-----------------
>>
>>  <property>
>>   <name>mapreduce.framework.name</name>
>>   <value>yarn</value>
>> </property>
>>
>>  <property>
>>   <name>mapred.job.tracker</name>
>>   <value>master:54311</value>
>>   <description>The host and port that the MapReduce job tracker runs
>>   at.  If "local", then jobs are run in-process as a single map
>>   and reduce task.
>>   </description>
>> </property>
>>
>>
>>
>>
>>  --
>>
>>
>>  *Sisu Xi, PhD Candidate*
>>
>> http://www.cse.wustl.edu/~xis/
>> Department of Computer Science and Engineering
>> Campus Box 1045
>> Washington University in St. Louis
>> One Brookings Drive
>> St. Louis, MO 63130
>>
>
>
>
> --
>
>
> *Sisu Xi, PhD Candidate*
>
> http://www.cse.wustl.edu/~xis/
> Department of Computer Science and Engineering
> Campus Box 1045
> Washington University in St. Louis
> One Brookings Drive
> St. Louis, MO 63130
>



-- 


*Sisu Xi, PhD Candidate*

http://www.cse.wustl.edu/~xis/
Department of Computer Science and Engineering
Campus Box 1045
Washington University in St. Louis
One Brookings Drive
St. Louis, MO 63130

Re: hadoop multinode, only master node doing the work

Posted by Sisu Xi <xi...@gmail.com>.
I checked the log on the slave node,

it says:

2014-07-13 13:35:37,290 FATAL
org.apache.hadoop.hdfs.server.datanode.DataNode: Initialization failed for
block pool Block pool BP-1408390155-172.16.20.170-1405276491305 (storage id
DS-111686822-127.0.1.1-50010-1405200560127) service to master/
172.16.20.170:54310
java.io.IOException: Incompatible clusterIDs in
/home/xisisu/mydata/hdfs/datanode: namenode clusterID =
CID-5fe42d16-b0cc-4104-b04c-8b4561e907dd; datanode clusterID =
CID-95916f4d-89f1-47e4-a92e-7b88ce97ed8b
        at
org.apache.hadoop.hdfs.server.datanode.DataStorage.doTransition(DataStorage.java:391)
        at
org.apache.hadoop.hdfs.server.datanode.DataStorage.recoverTransitionRead(DataStorage.java:191)
        at
org.apache.hadoop.hdfs.server.datanode.DataStorage.recoverTransitionRead(DataStorage.java:219)
        at
org.apache.hadoop.hdfs.server.datanode.DataNode.initStorage(DataNode.java:837)
        at
org.apache.hadoop.hdfs.server.datanode.DataNode.initBlockPool(DataNode.java:808)
        at
org.apache.hadoop.hdfs.server.datanode.BPOfferService.verifyAndSetNamespaceInfo(BPOfferService.java:280)
        at
org.apache.hadoop.hdfs.server.datanode.BPServiceActor.connectToNNAndHandshake(BPServiceActor.java:222)
        at
org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:664)
        at java.lang.Thread.run(Thread.java:744)
2014-07-13 13:35:37,292 WARN
org.apache.hadoop.hdfs.server.datanode.DataNode: Ending block pool service
for: Block pool BP-1408390155-172.16.20.170-1405276491305 (storage id
DS-111686822-127.0.1.1-50010-1405200560127) service to master/
172.16.20.170:54310
2014-07-13 13:35:37,392 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Removed Block pool
BP-1408390155-172.16.20.170-1405276491305 (storage id
DS-111686822-127.0.1.1-50010-1405200560127)
2014-07-13 13:35:39,392 WARN
org.apache.hadoop.hdfs.server.datanode.DataNode: Exiting Datanode
2014-07-13 13:35:39,394 INFO org.apache.hadoop.util.ExitUtil: Exiting with
status 0
2014-07-13 13:35:39,395 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: SHUTDOWN_MSG:
/************************************************************
SHUTDOWN_MSG: Shutting down DataNode at slave-01/127.0.1.1
************************************************************/


I did some google search, it says the problem is with the data dir on the
master node.
I already removed the whole data dir, formated the namenode, and restarted
the hadoop.
But still the same problem.

Sisu




On Sun, Jul 13, 2014 at 12:53 PM, Sisu Xi <xi...@gmail.com> wrote:

> Hi, Sam:
>
> Thanks for your help! You are right, seems only one node is running.
> Here is the output:
>
> xisisu@slave-01:/usr/local/hadoop$ hadoop dfsadmin -report
> DEPRECATED: Use of this script to execute hdfs command is deprecated.
> Instead use the hdfs command for it.
>
> OpenJDK 64-Bit Server VM warning: You have loaded library
> /usr/local/hadoop/lib/native/libhadoop.so.1.0.0 which might have disabled
> stack guard. The VM will try to fix the stack guard now.
> It's highly recommended that you fix the library with 'execstack -c
> <libfile>', or link it with '-z noexecstack'.
> 14/07/13 12:50:14 WARN util.NativeCodeLoader: Unable to load native-hadoop
> library for your platform... using builtin-java classes where applicable
> Configured Capacity: 14068822016 (13.10 GB)
> Present Capacity: 8425232425 (7.85 GB)
> DFS Remaining: 8423534592 (7.85 GB)
> DFS Used: 1697833 (1.62 MB)
> DFS Used%: 0.02%
> Under replicated blocks: 21
> Blocks with corrupt replicas: 0
> Missing blocks: 0
>
> -------------------------------------------------
> Datanodes available: 1 (1 total, 0 dead)
>
> Live datanodes:
> Name: 172.16.20.170:50010 (master)
> Hostname: master
> Decommission Status : Normal
> Configured Capacity: 14068822016 (13.10 GB)
> DFS Used: 1697833 (1.62 MB)
> Non DFS Used: 5643589591 (5.26 GB)
> DFS Remaining: 8423534592 (7.85 GB)
> DFS Used%: 0.01%
> DFS Remaining%: 59.87%
> Last contact: Sun Jul 13 12:50:13 CDT 2014
>
>
> I get the same output when I run the command on the slave node.
>
>
> Is there anything I am missing in the config file?
>
> Thanks very much!
>
> Sisu
>
>
>
> On Sun, Jul 13, 2014 at 1:36 AM, Kilaru, Sambaiah <
> Sambaiah_Kilaru@intuit.com> wrote:
>
>>  HI Sisu Xi,
>>
>>  On the master node can you check
>> Hadoop dfsadmin –report
>> And listing all the slave nodes or you can check master URL and it should
>> all datanodes listed as slave nodes.
>> Check for RM UI and slave nodes listed there also.
>>
>>  Thanks,
>> Sam
>>
>>   From: Sisu Xi <xi...@gmail.com>
>> Reply-To: "user@hadoop.apache.org" <us...@hadoop.apache.org>
>> Date: Sunday, July 13, 2014 at 11:28 AM
>> To: "user@hadoop.apache.org" <us...@hadoop.apache.org>
>> Subject: hadoop multinode, only master node doing the work
>>
>>   Hi, all:
>>
>>  I am new to hadoop. I followed the tutorial on
>>
>> http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-multi-node-cluster/
>>
>>  and installed hadoop 2.2.0 on two 4-core Ubuntu 12.04 machines.
>>
>>  I can start the pi program, however, only the master node is doing the
>> work (I checked the top on each machine).
>> Seems the two nodes are configured correctly, because I can start  the
>> program in the slave node as well, and still only the master node is doing
>> the actual work.
>> I have tried different number of mappers for the pi program, and the
>> results is the same.
>>
>>  Is there anything else I can check?
>>
>>  In the end is my configure file on each host.
>>
>>  Thanks very much!
>>
>>  Sisu
>>
>>  ---------yarn-site.xml-------
>>
>>  <property>
>>   <name>yarn.nodemanager.aux-services</name>
>>   <value>mapreduce_shuffle</value>
>> </property>
>>
>>
>>  <property>
>>   <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name>
>>   <value>org.apache.hadoop.mapred.ShuffleHandler</value>
>> </property>
>>
>>
>>  <property>
>>   <name>yarn.resourcemanager.address</name>
>>   <value>master:8032</value>
>> </property>
>>
>>  ---------------hdfs-site.xml--------------------
>>
>>  <property>
>>   <name>dfs.replication</name>
>>   <value>2</value>
>> </property>
>>
>>
>>  <property>
>>   <name>dfs.namenode.name.dir</name>
>>   <value>file:/home/xisisu/mydata/hdfs/namenode</value>
>> </property>
>>
>>
>>  <property>
>>   <name>dfs.datanode.data.dir</name>
>>   <value>file:/home/xisisu/mydata/hdfs/datanode</value>
>> </property>
>>
>>
>>  -------------core-site.xml-------------
>>
>>  <property>
>>  <name>fs.default.name</name>
>> <value>hdfs://master:9000</value>
>> </property>
>>
>>  ------------------mapred-site.xml-----------------
>>
>>  <property>
>>   <name>mapreduce.framework.name</name>
>>   <value>yarn</value>
>> </property>
>>
>>  <property>
>>   <name>mapred.job.tracker</name>
>>   <value>master:54311</value>
>>   <description>The host and port that the MapReduce job tracker runs
>>   at.  If "local", then jobs are run in-process as a single map
>>   and reduce task.
>>   </description>
>> </property>
>>
>>
>>
>>
>>  --
>>
>>
>>  *Sisu Xi, PhD Candidate*
>>
>> http://www.cse.wustl.edu/~xis/
>> Department of Computer Science and Engineering
>> Campus Box 1045
>> Washington University in St. Louis
>> One Brookings Drive
>> St. Louis, MO 63130
>>
>
>
>
> --
>
>
> *Sisu Xi, PhD Candidate*
>
> http://www.cse.wustl.edu/~xis/
> Department of Computer Science and Engineering
> Campus Box 1045
> Washington University in St. Louis
> One Brookings Drive
> St. Louis, MO 63130
>



-- 


*Sisu Xi, PhD Candidate*

http://www.cse.wustl.edu/~xis/
Department of Computer Science and Engineering
Campus Box 1045
Washington University in St. Louis
One Brookings Drive
St. Louis, MO 63130

Re: hadoop multinode, only master node doing the work

Posted by Sisu Xi <xi...@gmail.com>.
Hi, Sam:

Thanks for your help! You are right, seems only one node is running.
Here is the output:

xisisu@slave-01:/usr/local/hadoop$ hadoop dfsadmin -report
DEPRECATED: Use of this script to execute hdfs command is deprecated.
Instead use the hdfs command for it.

OpenJDK 64-Bit Server VM warning: You have loaded library
/usr/local/hadoop/lib/native/libhadoop.so.1.0.0 which might have disabled
stack guard. The VM will try to fix the stack guard now.
It's highly recommended that you fix the library with 'execstack -c
<libfile>', or link it with '-z noexecstack'.
14/07/13 12:50:14 WARN util.NativeCodeLoader: Unable to load native-hadoop
library for your platform... using builtin-java classes where applicable
Configured Capacity: 14068822016 (13.10 GB)
Present Capacity: 8425232425 (7.85 GB)
DFS Remaining: 8423534592 (7.85 GB)
DFS Used: 1697833 (1.62 MB)
DFS Used%: 0.02%
Under replicated blocks: 21
Blocks with corrupt replicas: 0
Missing blocks: 0

-------------------------------------------------
Datanodes available: 1 (1 total, 0 dead)

Live datanodes:
Name: 172.16.20.170:50010 (master)
Hostname: master
Decommission Status : Normal
Configured Capacity: 14068822016 (13.10 GB)
DFS Used: 1697833 (1.62 MB)
Non DFS Used: 5643589591 (5.26 GB)
DFS Remaining: 8423534592 (7.85 GB)
DFS Used%: 0.01%
DFS Remaining%: 59.87%
Last contact: Sun Jul 13 12:50:13 CDT 2014


I get the same output when I run the command on the slave node.


Is there anything I am missing in the config file?

Thanks very much!

Sisu



On Sun, Jul 13, 2014 at 1:36 AM, Kilaru, Sambaiah <
Sambaiah_Kilaru@intuit.com> wrote:

>  HI Sisu Xi,
>
>  On the master node can you check
> Hadoop dfsadmin –report
> And listing all the slave nodes or you can check master URL and it should
> all datanodes listed as slave nodes.
> Check for RM UI and slave nodes listed there also.
>
>  Thanks,
> Sam
>
>   From: Sisu Xi <xi...@gmail.com>
> Reply-To: "user@hadoop.apache.org" <us...@hadoop.apache.org>
> Date: Sunday, July 13, 2014 at 11:28 AM
> To: "user@hadoop.apache.org" <us...@hadoop.apache.org>
> Subject: hadoop multinode, only master node doing the work
>
>   Hi, all:
>
>  I am new to hadoop. I followed the tutorial on
>
> http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-multi-node-cluster/
>
>  and installed hadoop 2.2.0 on two 4-core Ubuntu 12.04 machines.
>
>  I can start the pi program, however, only the master node is doing the
> work (I checked the top on each machine).
> Seems the two nodes are configured correctly, because I can start  the
> program in the slave node as well, and still only the master node is doing
> the actual work.
> I have tried different number of mappers for the pi program, and the
> results is the same.
>
>  Is there anything else I can check?
>
>  In the end is my configure file on each host.
>
>  Thanks very much!
>
>  Sisu
>
>  ---------yarn-site.xml-------
>
>  <property>
>   <name>yarn.nodemanager.aux-services</name>
>   <value>mapreduce_shuffle</value>
> </property>
>
>
>  <property>
>   <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name>
>   <value>org.apache.hadoop.mapred.ShuffleHandler</value>
> </property>
>
>
>  <property>
>   <name>yarn.resourcemanager.address</name>
>   <value>master:8032</value>
> </property>
>
>  ---------------hdfs-site.xml--------------------
>
>  <property>
>   <name>dfs.replication</name>
>   <value>2</value>
> </property>
>
>
>  <property>
>   <name>dfs.namenode.name.dir</name>
>   <value>file:/home/xisisu/mydata/hdfs/namenode</value>
> </property>
>
>
>  <property>
>   <name>dfs.datanode.data.dir</name>
>   <value>file:/home/xisisu/mydata/hdfs/datanode</value>
> </property>
>
>
>  -------------core-site.xml-------------
>
>  <property>
>  <name>fs.default.name</name>
> <value>hdfs://master:9000</value>
> </property>
>
>  ------------------mapred-site.xml-----------------
>
>  <property>
>   <name>mapreduce.framework.name</name>
>   <value>yarn</value>
> </property>
>
>  <property>
>   <name>mapred.job.tracker</name>
>   <value>master:54311</value>
>   <description>The host and port that the MapReduce job tracker runs
>   at.  If "local", then jobs are run in-process as a single map
>   and reduce task.
>   </description>
> </property>
>
>
>
>
>  --
>
>
>  *Sisu Xi, PhD Candidate*
>
> http://www.cse.wustl.edu/~xis/
> Department of Computer Science and Engineering
> Campus Box 1045
> Washington University in St. Louis
> One Brookings Drive
> St. Louis, MO 63130
>



-- 


*Sisu Xi, PhD Candidate*

http://www.cse.wustl.edu/~xis/
Department of Computer Science and Engineering
Campus Box 1045
Washington University in St. Louis
One Brookings Drive
St. Louis, MO 63130

Re: hadoop multinode, only master node doing the work

Posted by Sisu Xi <xi...@gmail.com>.
Hi, Sam:

Thanks for your help! You are right, seems only one node is running.
Here is the output:

xisisu@slave-01:/usr/local/hadoop$ hadoop dfsadmin -report
DEPRECATED: Use of this script to execute hdfs command is deprecated.
Instead use the hdfs command for it.

OpenJDK 64-Bit Server VM warning: You have loaded library
/usr/local/hadoop/lib/native/libhadoop.so.1.0.0 which might have disabled
stack guard. The VM will try to fix the stack guard now.
It's highly recommended that you fix the library with 'execstack -c
<libfile>', or link it with '-z noexecstack'.
14/07/13 12:50:14 WARN util.NativeCodeLoader: Unable to load native-hadoop
library for your platform... using builtin-java classes where applicable
Configured Capacity: 14068822016 (13.10 GB)
Present Capacity: 8425232425 (7.85 GB)
DFS Remaining: 8423534592 (7.85 GB)
DFS Used: 1697833 (1.62 MB)
DFS Used%: 0.02%
Under replicated blocks: 21
Blocks with corrupt replicas: 0
Missing blocks: 0

-------------------------------------------------
Datanodes available: 1 (1 total, 0 dead)

Live datanodes:
Name: 172.16.20.170:50010 (master)
Hostname: master
Decommission Status : Normal
Configured Capacity: 14068822016 (13.10 GB)
DFS Used: 1697833 (1.62 MB)
Non DFS Used: 5643589591 (5.26 GB)
DFS Remaining: 8423534592 (7.85 GB)
DFS Used%: 0.01%
DFS Remaining%: 59.87%
Last contact: Sun Jul 13 12:50:13 CDT 2014


I get the same output when I run the command on the slave node.


Is there anything I am missing in the config file?

Thanks very much!

Sisu



On Sun, Jul 13, 2014 at 1:36 AM, Kilaru, Sambaiah <
Sambaiah_Kilaru@intuit.com> wrote:

>  HI Sisu Xi,
>
>  On the master node can you check
> Hadoop dfsadmin –report
> And listing all the slave nodes or you can check master URL and it should
> all datanodes listed as slave nodes.
> Check for RM UI and slave nodes listed there also.
>
>  Thanks,
> Sam
>
>   From: Sisu Xi <xi...@gmail.com>
> Reply-To: "user@hadoop.apache.org" <us...@hadoop.apache.org>
> Date: Sunday, July 13, 2014 at 11:28 AM
> To: "user@hadoop.apache.org" <us...@hadoop.apache.org>
> Subject: hadoop multinode, only master node doing the work
>
>   Hi, all:
>
>  I am new to hadoop. I followed the tutorial on
>
> http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-multi-node-cluster/
>
>  and installed hadoop 2.2.0 on two 4-core Ubuntu 12.04 machines.
>
>  I can start the pi program, however, only the master node is doing the
> work (I checked the top on each machine).
> Seems the two nodes are configured correctly, because I can start  the
> program in the slave node as well, and still only the master node is doing
> the actual work.
> I have tried different number of mappers for the pi program, and the
> results is the same.
>
>  Is there anything else I can check?
>
>  In the end is my configure file on each host.
>
>  Thanks very much!
>
>  Sisu
>
>  ---------yarn-site.xml-------
>
>  <property>
>   <name>yarn.nodemanager.aux-services</name>
>   <value>mapreduce_shuffle</value>
> </property>
>
>
>  <property>
>   <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name>
>   <value>org.apache.hadoop.mapred.ShuffleHandler</value>
> </property>
>
>
>  <property>
>   <name>yarn.resourcemanager.address</name>
>   <value>master:8032</value>
> </property>
>
>  ---------------hdfs-site.xml--------------------
>
>  <property>
>   <name>dfs.replication</name>
>   <value>2</value>
> </property>
>
>
>  <property>
>   <name>dfs.namenode.name.dir</name>
>   <value>file:/home/xisisu/mydata/hdfs/namenode</value>
> </property>
>
>
>  <property>
>   <name>dfs.datanode.data.dir</name>
>   <value>file:/home/xisisu/mydata/hdfs/datanode</value>
> </property>
>
>
>  -------------core-site.xml-------------
>
>  <property>
>  <name>fs.default.name</name>
> <value>hdfs://master:9000</value>
> </property>
>
>  ------------------mapred-site.xml-----------------
>
>  <property>
>   <name>mapreduce.framework.name</name>
>   <value>yarn</value>
> </property>
>
>  <property>
>   <name>mapred.job.tracker</name>
>   <value>master:54311</value>
>   <description>The host and port that the MapReduce job tracker runs
>   at.  If "local", then jobs are run in-process as a single map
>   and reduce task.
>   </description>
> </property>
>
>
>
>
>  --
>
>
>  *Sisu Xi, PhD Candidate*
>
> http://www.cse.wustl.edu/~xis/
> Department of Computer Science and Engineering
> Campus Box 1045
> Washington University in St. Louis
> One Brookings Drive
> St. Louis, MO 63130
>



-- 


*Sisu Xi, PhD Candidate*

http://www.cse.wustl.edu/~xis/
Department of Computer Science and Engineering
Campus Box 1045
Washington University in St. Louis
One Brookings Drive
St. Louis, MO 63130

Re: hadoop multinode, only master node doing the work

Posted by Sisu Xi <xi...@gmail.com>.
Hi, Sam:

Thanks for your help! You are right, seems only one node is running.
Here is the output:

xisisu@slave-01:/usr/local/hadoop$ hadoop dfsadmin -report
DEPRECATED: Use of this script to execute hdfs command is deprecated.
Instead use the hdfs command for it.

OpenJDK 64-Bit Server VM warning: You have loaded library
/usr/local/hadoop/lib/native/libhadoop.so.1.0.0 which might have disabled
stack guard. The VM will try to fix the stack guard now.
It's highly recommended that you fix the library with 'execstack -c
<libfile>', or link it with '-z noexecstack'.
14/07/13 12:50:14 WARN util.NativeCodeLoader: Unable to load native-hadoop
library for your platform... using builtin-java classes where applicable
Configured Capacity: 14068822016 (13.10 GB)
Present Capacity: 8425232425 (7.85 GB)
DFS Remaining: 8423534592 (7.85 GB)
DFS Used: 1697833 (1.62 MB)
DFS Used%: 0.02%
Under replicated blocks: 21
Blocks with corrupt replicas: 0
Missing blocks: 0

-------------------------------------------------
Datanodes available: 1 (1 total, 0 dead)

Live datanodes:
Name: 172.16.20.170:50010 (master)
Hostname: master
Decommission Status : Normal
Configured Capacity: 14068822016 (13.10 GB)
DFS Used: 1697833 (1.62 MB)
Non DFS Used: 5643589591 (5.26 GB)
DFS Remaining: 8423534592 (7.85 GB)
DFS Used%: 0.01%
DFS Remaining%: 59.87%
Last contact: Sun Jul 13 12:50:13 CDT 2014


I get the same output when I run the command on the slave node.


Is there anything I am missing in the config file?

Thanks very much!

Sisu



On Sun, Jul 13, 2014 at 1:36 AM, Kilaru, Sambaiah <
Sambaiah_Kilaru@intuit.com> wrote:

>  HI Sisu Xi,
>
>  On the master node can you check
> Hadoop dfsadmin –report
> And listing all the slave nodes or you can check master URL and it should
> all datanodes listed as slave nodes.
> Check for RM UI and slave nodes listed there also.
>
>  Thanks,
> Sam
>
>   From: Sisu Xi <xi...@gmail.com>
> Reply-To: "user@hadoop.apache.org" <us...@hadoop.apache.org>
> Date: Sunday, July 13, 2014 at 11:28 AM
> To: "user@hadoop.apache.org" <us...@hadoop.apache.org>
> Subject: hadoop multinode, only master node doing the work
>
>   Hi, all:
>
>  I am new to hadoop. I followed the tutorial on
>
> http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-multi-node-cluster/
>
>  and installed hadoop 2.2.0 on two 4-core Ubuntu 12.04 machines.
>
>  I can start the pi program, however, only the master node is doing the
> work (I checked the top on each machine).
> Seems the two nodes are configured correctly, because I can start  the
> program in the slave node as well, and still only the master node is doing
> the actual work.
> I have tried different number of mappers for the pi program, and the
> results is the same.
>
>  Is there anything else I can check?
>
>  In the end is my configure file on each host.
>
>  Thanks very much!
>
>  Sisu
>
>  ---------yarn-site.xml-------
>
>  <property>
>   <name>yarn.nodemanager.aux-services</name>
>   <value>mapreduce_shuffle</value>
> </property>
>
>
>  <property>
>   <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name>
>   <value>org.apache.hadoop.mapred.ShuffleHandler</value>
> </property>
>
>
>  <property>
>   <name>yarn.resourcemanager.address</name>
>   <value>master:8032</value>
> </property>
>
>  ---------------hdfs-site.xml--------------------
>
>  <property>
>   <name>dfs.replication</name>
>   <value>2</value>
> </property>
>
>
>  <property>
>   <name>dfs.namenode.name.dir</name>
>   <value>file:/home/xisisu/mydata/hdfs/namenode</value>
> </property>
>
>
>  <property>
>   <name>dfs.datanode.data.dir</name>
>   <value>file:/home/xisisu/mydata/hdfs/datanode</value>
> </property>
>
>
>  -------------core-site.xml-------------
>
>  <property>
>  <name>fs.default.name</name>
> <value>hdfs://master:9000</value>
> </property>
>
>  ------------------mapred-site.xml-----------------
>
>  <property>
>   <name>mapreduce.framework.name</name>
>   <value>yarn</value>
> </property>
>
>  <property>
>   <name>mapred.job.tracker</name>
>   <value>master:54311</value>
>   <description>The host and port that the MapReduce job tracker runs
>   at.  If "local", then jobs are run in-process as a single map
>   and reduce task.
>   </description>
> </property>
>
>
>
>
>  --
>
>
>  *Sisu Xi, PhD Candidate*
>
> http://www.cse.wustl.edu/~xis/
> Department of Computer Science and Engineering
> Campus Box 1045
> Washington University in St. Louis
> One Brookings Drive
> St. Louis, MO 63130
>



-- 


*Sisu Xi, PhD Candidate*

http://www.cse.wustl.edu/~xis/
Department of Computer Science and Engineering
Campus Box 1045
Washington University in St. Louis
One Brookings Drive
St. Louis, MO 63130

Re: hadoop multinode, only master node doing the work

Posted by Sisu Xi <xi...@gmail.com>.
Hi, Sam:

Thanks for your help! You are right, seems only one node is running.
Here is the output:

xisisu@slave-01:/usr/local/hadoop$ hadoop dfsadmin -report
DEPRECATED: Use of this script to execute hdfs command is deprecated.
Instead use the hdfs command for it.

OpenJDK 64-Bit Server VM warning: You have loaded library
/usr/local/hadoop/lib/native/libhadoop.so.1.0.0 which might have disabled
stack guard. The VM will try to fix the stack guard now.
It's highly recommended that you fix the library with 'execstack -c
<libfile>', or link it with '-z noexecstack'.
14/07/13 12:50:14 WARN util.NativeCodeLoader: Unable to load native-hadoop
library for your platform... using builtin-java classes where applicable
Configured Capacity: 14068822016 (13.10 GB)
Present Capacity: 8425232425 (7.85 GB)
DFS Remaining: 8423534592 (7.85 GB)
DFS Used: 1697833 (1.62 MB)
DFS Used%: 0.02%
Under replicated blocks: 21
Blocks with corrupt replicas: 0
Missing blocks: 0

-------------------------------------------------
Datanodes available: 1 (1 total, 0 dead)

Live datanodes:
Name: 172.16.20.170:50010 (master)
Hostname: master
Decommission Status : Normal
Configured Capacity: 14068822016 (13.10 GB)
DFS Used: 1697833 (1.62 MB)
Non DFS Used: 5643589591 (5.26 GB)
DFS Remaining: 8423534592 (7.85 GB)
DFS Used%: 0.01%
DFS Remaining%: 59.87%
Last contact: Sun Jul 13 12:50:13 CDT 2014


I get the same output when I run the command on the slave node.


Is there anything I am missing in the config file?

Thanks very much!

Sisu



On Sun, Jul 13, 2014 at 1:36 AM, Kilaru, Sambaiah <
Sambaiah_Kilaru@intuit.com> wrote:

>  HI Sisu Xi,
>
>  On the master node can you check
> Hadoop dfsadmin –report
> And listing all the slave nodes or you can check master URL and it should
> all datanodes listed as slave nodes.
> Check for RM UI and slave nodes listed there also.
>
>  Thanks,
> Sam
>
>   From: Sisu Xi <xi...@gmail.com>
> Reply-To: "user@hadoop.apache.org" <us...@hadoop.apache.org>
> Date: Sunday, July 13, 2014 at 11:28 AM
> To: "user@hadoop.apache.org" <us...@hadoop.apache.org>
> Subject: hadoop multinode, only master node doing the work
>
>   Hi, all:
>
>  I am new to hadoop. I followed the tutorial on
>
> http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-multi-node-cluster/
>
>  and installed hadoop 2.2.0 on two 4-core Ubuntu 12.04 machines.
>
>  I can start the pi program, however, only the master node is doing the
> work (I checked the top on each machine).
> Seems the two nodes are configured correctly, because I can start  the
> program in the slave node as well, and still only the master node is doing
> the actual work.
> I have tried different number of mappers for the pi program, and the
> results is the same.
>
>  Is there anything else I can check?
>
>  In the end is my configure file on each host.
>
>  Thanks very much!
>
>  Sisu
>
>  ---------yarn-site.xml-------
>
>  <property>
>   <name>yarn.nodemanager.aux-services</name>
>   <value>mapreduce_shuffle</value>
> </property>
>
>
>  <property>
>   <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name>
>   <value>org.apache.hadoop.mapred.ShuffleHandler</value>
> </property>
>
>
>  <property>
>   <name>yarn.resourcemanager.address</name>
>   <value>master:8032</value>
> </property>
>
>  ---------------hdfs-site.xml--------------------
>
>  <property>
>   <name>dfs.replication</name>
>   <value>2</value>
> </property>
>
>
>  <property>
>   <name>dfs.namenode.name.dir</name>
>   <value>file:/home/xisisu/mydata/hdfs/namenode</value>
> </property>
>
>
>  <property>
>   <name>dfs.datanode.data.dir</name>
>   <value>file:/home/xisisu/mydata/hdfs/datanode</value>
> </property>
>
>
>  -------------core-site.xml-------------
>
>  <property>
>  <name>fs.default.name</name>
> <value>hdfs://master:9000</value>
> </property>
>
>  ------------------mapred-site.xml-----------------
>
>  <property>
>   <name>mapreduce.framework.name</name>
>   <value>yarn</value>
> </property>
>
>  <property>
>   <name>mapred.job.tracker</name>
>   <value>master:54311</value>
>   <description>The host and port that the MapReduce job tracker runs
>   at.  If "local", then jobs are run in-process as a single map
>   and reduce task.
>   </description>
> </property>
>
>
>
>
>  --
>
>
>  *Sisu Xi, PhD Candidate*
>
> http://www.cse.wustl.edu/~xis/
> Department of Computer Science and Engineering
> Campus Box 1045
> Washington University in St. Louis
> One Brookings Drive
> St. Louis, MO 63130
>



-- 


*Sisu Xi, PhD Candidate*

http://www.cse.wustl.edu/~xis/
Department of Computer Science and Engineering
Campus Box 1045
Washington University in St. Louis
One Brookings Drive
St. Louis, MO 63130

Re: hadoop multinode, only master node doing the work

Posted by "Kilaru, Sambaiah" <Sa...@intuit.com>.
HI Sisu Xi,

On the master node can you check
Hadoop dfsadmin –report
And listing all the slave nodes or you can check master URL and it should all datanodes listed as slave nodes.
Check for RM UI and slave nodes listed there also.

Thanks,
Sam

From: Sisu Xi <xi...@gmail.com>>
Reply-To: "user@hadoop.apache.org<ma...@hadoop.apache.org>" <us...@hadoop.apache.org>>
Date: Sunday, July 13, 2014 at 11:28 AM
To: "user@hadoop.apache.org<ma...@hadoop.apache.org>" <us...@hadoop.apache.org>>
Subject: hadoop multinode, only master node doing the work

Hi, all:

I am new to hadoop. I followed the tutorial on
http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-multi-node-cluster/

and installed hadoop 2.2.0 on two 4-core Ubuntu 12.04 machines.

I can start the pi program, however, only the master node is doing the work (I checked the top on each machine).
Seems the two nodes are configured correctly, because I can start  the program in the slave node as well, and still only the master node is doing the actual work.
I have tried different number of mappers for the pi program, and the results is the same.

Is there anything else I can check?

In the end is my configure file on each host.

Thanks very much!

Sisu

---------yarn-site.xml-------

<property>
  <name>yarn.nodemanager.aux-services</name>
  <value>mapreduce_shuffle</value>
</property>


<property>
  <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name>
  <value>org.apache.hadoop.mapred.ShuffleHandler</value>
</property>


<property>
  <name>yarn.resourcemanager.address</name>
  <value>master:8032</value>
</property>

---------------hdfs-site.xml--------------------

<property>
  <name>dfs.replication</name>
  <value>2</value>
</property>


<property>
  <name>dfs.namenode.name.dir</name>
  <value>file:/home/xisisu/mydata/hdfs/namenode</value>
</property>


<property>
  <name>dfs.datanode.data.dir</name>
  <value>file:/home/xisisu/mydata/hdfs/datanode</value>
</property>


-------------core-site.xml-------------

<property>
<name>fs.default.name<http://fs.default.name></name>
<value>hdfs://master:9000</value>
</property>

------------------mapred-site.xml-----------------

<property>
  <name>mapreduce.framework.name<http://mapreduce.framework.name></name>
  <value>yarn</value>
</property>

<property>
  <name>mapred.job.tracker</name>
  <value>master:54311</value>
  <description>The host and port that the MapReduce job tracker runs
  at.  If "local", then jobs are run in-process as a single map
  and reduce task.
  </description>
</property>




--
[https://dl.dropboxusercontent.com/u/268069/Sisu_Contact.jpeg]

Sisu Xi, PhD Candidate

http://www.cse.wustl.edu/~xis/
Department of Computer Science and Engineering
Campus Box 1045
Washington University in St. Louis
One Brookings Drive
St. Louis, MO 63130

Re: hadoop multinode, only master node doing the work

Posted by "Kilaru, Sambaiah" <Sa...@intuit.com>.
HI Sisu Xi,

On the master node can you check
Hadoop dfsadmin –report
And listing all the slave nodes or you can check master URL and it should all datanodes listed as slave nodes.
Check for RM UI and slave nodes listed there also.

Thanks,
Sam

From: Sisu Xi <xi...@gmail.com>>
Reply-To: "user@hadoop.apache.org<ma...@hadoop.apache.org>" <us...@hadoop.apache.org>>
Date: Sunday, July 13, 2014 at 11:28 AM
To: "user@hadoop.apache.org<ma...@hadoop.apache.org>" <us...@hadoop.apache.org>>
Subject: hadoop multinode, only master node doing the work

Hi, all:

I am new to hadoop. I followed the tutorial on
http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-multi-node-cluster/

and installed hadoop 2.2.0 on two 4-core Ubuntu 12.04 machines.

I can start the pi program, however, only the master node is doing the work (I checked the top on each machine).
Seems the two nodes are configured correctly, because I can start  the program in the slave node as well, and still only the master node is doing the actual work.
I have tried different number of mappers for the pi program, and the results is the same.

Is there anything else I can check?

In the end is my configure file on each host.

Thanks very much!

Sisu

---------yarn-site.xml-------

<property>
  <name>yarn.nodemanager.aux-services</name>
  <value>mapreduce_shuffle</value>
</property>


<property>
  <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name>
  <value>org.apache.hadoop.mapred.ShuffleHandler</value>
</property>


<property>
  <name>yarn.resourcemanager.address</name>
  <value>master:8032</value>
</property>

---------------hdfs-site.xml--------------------

<property>
  <name>dfs.replication</name>
  <value>2</value>
</property>


<property>
  <name>dfs.namenode.name.dir</name>
  <value>file:/home/xisisu/mydata/hdfs/namenode</value>
</property>


<property>
  <name>dfs.datanode.data.dir</name>
  <value>file:/home/xisisu/mydata/hdfs/datanode</value>
</property>


-------------core-site.xml-------------

<property>
<name>fs.default.name<http://fs.default.name></name>
<value>hdfs://master:9000</value>
</property>

------------------mapred-site.xml-----------------

<property>
  <name>mapreduce.framework.name<http://mapreduce.framework.name></name>
  <value>yarn</value>
</property>

<property>
  <name>mapred.job.tracker</name>
  <value>master:54311</value>
  <description>The host and port that the MapReduce job tracker runs
  at.  If "local", then jobs are run in-process as a single map
  and reduce task.
  </description>
</property>




--
[https://dl.dropboxusercontent.com/u/268069/Sisu_Contact.jpeg]

Sisu Xi, PhD Candidate

http://www.cse.wustl.edu/~xis/
Department of Computer Science and Engineering
Campus Box 1045
Washington University in St. Louis
One Brookings Drive
St. Louis, MO 63130

Re: hadoop multinode, only master node doing the work

Posted by "Kilaru, Sambaiah" <Sa...@intuit.com>.
HI Sisu Xi,

On the master node can you check
Hadoop dfsadmin –report
And listing all the slave nodes or you can check master URL and it should all datanodes listed as slave nodes.
Check for RM UI and slave nodes listed there also.

Thanks,
Sam

From: Sisu Xi <xi...@gmail.com>>
Reply-To: "user@hadoop.apache.org<ma...@hadoop.apache.org>" <us...@hadoop.apache.org>>
Date: Sunday, July 13, 2014 at 11:28 AM
To: "user@hadoop.apache.org<ma...@hadoop.apache.org>" <us...@hadoop.apache.org>>
Subject: hadoop multinode, only master node doing the work

Hi, all:

I am new to hadoop. I followed the tutorial on
http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-multi-node-cluster/

and installed hadoop 2.2.0 on two 4-core Ubuntu 12.04 machines.

I can start the pi program, however, only the master node is doing the work (I checked the top on each machine).
Seems the two nodes are configured correctly, because I can start  the program in the slave node as well, and still only the master node is doing the actual work.
I have tried different number of mappers for the pi program, and the results is the same.

Is there anything else I can check?

In the end is my configure file on each host.

Thanks very much!

Sisu

---------yarn-site.xml-------

<property>
  <name>yarn.nodemanager.aux-services</name>
  <value>mapreduce_shuffle</value>
</property>


<property>
  <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name>
  <value>org.apache.hadoop.mapred.ShuffleHandler</value>
</property>


<property>
  <name>yarn.resourcemanager.address</name>
  <value>master:8032</value>
</property>

---------------hdfs-site.xml--------------------

<property>
  <name>dfs.replication</name>
  <value>2</value>
</property>


<property>
  <name>dfs.namenode.name.dir</name>
  <value>file:/home/xisisu/mydata/hdfs/namenode</value>
</property>


<property>
  <name>dfs.datanode.data.dir</name>
  <value>file:/home/xisisu/mydata/hdfs/datanode</value>
</property>


-------------core-site.xml-------------

<property>
<name>fs.default.name<http://fs.default.name></name>
<value>hdfs://master:9000</value>
</property>

------------------mapred-site.xml-----------------

<property>
  <name>mapreduce.framework.name<http://mapreduce.framework.name></name>
  <value>yarn</value>
</property>

<property>
  <name>mapred.job.tracker</name>
  <value>master:54311</value>
  <description>The host and port that the MapReduce job tracker runs
  at.  If "local", then jobs are run in-process as a single map
  and reduce task.
  </description>
</property>




--
[https://dl.dropboxusercontent.com/u/268069/Sisu_Contact.jpeg]

Sisu Xi, PhD Candidate

http://www.cse.wustl.edu/~xis/
Department of Computer Science and Engineering
Campus Box 1045
Washington University in St. Louis
One Brookings Drive
St. Louis, MO 63130

Re: hadoop multinode, only master node doing the work

Posted by "Kilaru, Sambaiah" <Sa...@intuit.com>.
HI Sisu Xi,

On the master node can you check
Hadoop dfsadmin –report
And listing all the slave nodes or you can check master URL and it should all datanodes listed as slave nodes.
Check for RM UI and slave nodes listed there also.

Thanks,
Sam

From: Sisu Xi <xi...@gmail.com>>
Reply-To: "user@hadoop.apache.org<ma...@hadoop.apache.org>" <us...@hadoop.apache.org>>
Date: Sunday, July 13, 2014 at 11:28 AM
To: "user@hadoop.apache.org<ma...@hadoop.apache.org>" <us...@hadoop.apache.org>>
Subject: hadoop multinode, only master node doing the work

Hi, all:

I am new to hadoop. I followed the tutorial on
http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-multi-node-cluster/

and installed hadoop 2.2.0 on two 4-core Ubuntu 12.04 machines.

I can start the pi program, however, only the master node is doing the work (I checked the top on each machine).
Seems the two nodes are configured correctly, because I can start  the program in the slave node as well, and still only the master node is doing the actual work.
I have tried different number of mappers for the pi program, and the results is the same.

Is there anything else I can check?

In the end is my configure file on each host.

Thanks very much!

Sisu

---------yarn-site.xml-------

<property>
  <name>yarn.nodemanager.aux-services</name>
  <value>mapreduce_shuffle</value>
</property>


<property>
  <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name>
  <value>org.apache.hadoop.mapred.ShuffleHandler</value>
</property>


<property>
  <name>yarn.resourcemanager.address</name>
  <value>master:8032</value>
</property>

---------------hdfs-site.xml--------------------

<property>
  <name>dfs.replication</name>
  <value>2</value>
</property>


<property>
  <name>dfs.namenode.name.dir</name>
  <value>file:/home/xisisu/mydata/hdfs/namenode</value>
</property>


<property>
  <name>dfs.datanode.data.dir</name>
  <value>file:/home/xisisu/mydata/hdfs/datanode</value>
</property>


-------------core-site.xml-------------

<property>
<name>fs.default.name<http://fs.default.name></name>
<value>hdfs://master:9000</value>
</property>

------------------mapred-site.xml-----------------

<property>
  <name>mapreduce.framework.name<http://mapreduce.framework.name></name>
  <value>yarn</value>
</property>

<property>
  <name>mapred.job.tracker</name>
  <value>master:54311</value>
  <description>The host and port that the MapReduce job tracker runs
  at.  If "local", then jobs are run in-process as a single map
  and reduce task.
  </description>
</property>




--
[https://dl.dropboxusercontent.com/u/268069/Sisu_Contact.jpeg]

Sisu Xi, PhD Candidate

http://www.cse.wustl.edu/~xis/
Department of Computer Science and Engineering
Campus Box 1045
Washington University in St. Louis
One Brookings Drive
St. Louis, MO 63130