You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-dev@hadoop.apache.org by maddy gulati <ac...@gmail.com> on 2012/04/28 22:19:21 UTC

Problem trying to set up single node cluster on Ubuntu

Folks,

I am trying to set up a single node cluster for Hadoop ( version 1.0.1) on
Ubuntu 11.04. After configuring everything, I could not get all the
components to start.

My installation directory for hadoop is /usr/local/hadoop
I followed the following tutorial for the setup :
http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-single-node-cluster/
Below is the snapshot of my log file:

mandeep@mandeep-System-Product-Name:/var/log/hadoop/mandeep$ cat
hadoop-mandeep-datanode-mandeep-System-Product-Name.log
2012-04-29 01:29:37,240 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: STARTUP_MSG:
/************************************************************
STARTUP_MSG: Starting DataNode
STARTUP_MSG:   host = mandeep-System-Product-Name/127.0.1.1
STARTUP_MSG:   args = []
STARTUP_MSG:   version = 1.0.1
STARTUP_MSG:   build =
https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0 -r
1243785; compiled by 'hortonfo' on Tue Feb 14 08:15:38 UTC 2012
************************************************************/
2012-04-29 01:29:37,398 INFO org.apache.hadoop.metrics2.impl.MetricsConfig:
loaded properties from hadoop-metrics2.properties
2012-04-29 01:29:37,446 INFO
org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
MetricsSystem,sub=Stats registered.
2012-04-29 01:29:37,449 INFO
org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot
period at 10 second(s).
2012-04-29 01:29:37,449 INFO
org.apache.hadoop.metrics2.impl.MetricsSystemImpl: DataNode metrics system
started
2012-04-29 01:29:37,583 INFO
org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source ugi
registered.
2012-04-29 01:29:37,586 WARN
org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Source name ugi already
exists!
2012-04-29 01:29:37,615 INFO org.apache.hadoop.util.NativeCodeLoader:
Loaded the native-hadoop library
2012-04-29 01:29:37,701 ERROR
org.apache.hadoop.hdfs.server.datanode.DataNode:
java.lang.IllegalArgumentException: Does not contain a valid host:port
authority: file:///
 at org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:162)
at
org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:198)
 at
org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:228)
at
org.apache.hadoop.hdfs.server.namenode.NameNode.getServiceAddress(NameNode.java:222)
 at
org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:337)
at org.apache.hadoop.hdfs.server.datanode.DataNode.<init>(DataNode.java:299)
 at
org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1582)
at
org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1521)
 at
org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1539)
at
org.apache.hadoop.hdfs.server.datanode.DataNode.secureMain(DataNode.java:1665)
 at org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:1682)

2012-04-29 01:29:37,718 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: SHUTDOWN_MSG:
/************************************************************
SHUTDOWN_MSG: Shutting down DataNode at mandeep-System-Product-Name/
127.0.1.1
************************************************************/
2012-04-29 01:33:26,907 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: STARTUP_MSG:
/************************************************************


-- 
Mandeep Singh Gulati
System Analyst | WorldWide Risk and Information Management
American Express India Pvt. Ltd.
Gurgaon, India

Re: Problem trying to set up single node cluster on Ubuntu

Posted by maddy gulati <ac...@gmail.com>.
Just to add, I had earlier installed Hadoop using debian package manager.
And then uninstalled it and reinstalled by building the source code
manually. Can that lead to problems. Because I can see that as per the old
installation ;

echo $HADOOP_CONF_DIR
/etc/hadoop

Even after uninstalling, the output is the same

On Sun, Apr 29, 2012 at 1:49 AM, maddy gulati <ac...@gmail.com>wrote:

> Folks,
>
> I am trying to set up a single node cluster for Hadoop ( version 1.0.1) on
> Ubuntu 11.04. After configuring everything, I could not get all the
> components to start.
>
> My installation directory for hadoop is /usr/local/hadoop
> I followed the following tutorial for the setup :
> http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-single-node-cluster/
> Below is the snapshot of my log file:
>
> mandeep@mandeep-System-Product-Name:/var/log/hadoop/mandeep$ cat
> hadoop-mandeep-datanode-mandeep-System-Product-Name.log
> 2012-04-29 01:29:37,240 INFO
> org.apache.hadoop.hdfs.server.datanode.DataNode: STARTUP_MSG:
> /************************************************************
> STARTUP_MSG: Starting DataNode
> STARTUP_MSG:   host = mandeep-System-Product-Name/127.0.1.1
> STARTUP_MSG:   args = []
> STARTUP_MSG:   version = 1.0.1
> STARTUP_MSG:   build =
> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0 -r
> 1243785; compiled by 'hortonfo' on Tue Feb 14 08:15:38 UTC 2012
> ************************************************************/
> 2012-04-29 01:29:37,398 INFO
> org.apache.hadoop.metrics2.impl.MetricsConfig: loaded properties from
> hadoop-metrics2.properties
> 2012-04-29 01:29:37,446 INFO
> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
> MetricsSystem,sub=Stats registered.
> 2012-04-29 01:29:37,449 INFO
> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot
> period at 10 second(s).
> 2012-04-29 01:29:37,449 INFO
> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: DataNode metrics system
> started
> 2012-04-29 01:29:37,583 INFO
> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source ugi
> registered.
> 2012-04-29 01:29:37,586 WARN
> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Source name ugi already
> exists!
> 2012-04-29 01:29:37,615 INFO org.apache.hadoop.util.NativeCodeLoader:
> Loaded the native-hadoop library
> 2012-04-29 01:29:37,701 ERROR
> org.apache.hadoop.hdfs.server.datanode.DataNode:
> java.lang.IllegalArgumentException: Does not contain a valid host:port
> authority: file:///
>  at org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:162)
> at
> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:198)
>  at
> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:228)
> at
> org.apache.hadoop.hdfs.server.namenode.NameNode.getServiceAddress(NameNode.java:222)
>  at
> org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:337)
> at
> org.apache.hadoop.hdfs.server.datanode.DataNode.<init>(DataNode.java:299)
>  at
> org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1582)
> at
> org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1521)
>  at
> org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1539)
> at
> org.apache.hadoop.hdfs.server.datanode.DataNode.secureMain(DataNode.java:1665)
>  at
> org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:1682)
>
> 2012-04-29 01:29:37,718 INFO
> org.apache.hadoop.hdfs.server.datanode.DataNode: SHUTDOWN_MSG:
> /************************************************************
> SHUTDOWN_MSG: Shutting down DataNode at mandeep-System-Product-Name/
> 127.0.1.1
> ************************************************************/
> 2012-04-29 01:33:26,907 INFO
> org.apache.hadoop.hdfs.server.datanode.DataNode: STARTUP_MSG:
> /************************************************************
>
>
> --
> Mandeep Singh Gulati
> System Analyst | WorldWide Risk and Information Management
> American Express India Pvt. Ltd.
> Gurgaon, India
>
>


-- 
Mandeep Singh Gulati
System Analyst | WorldWide Risk and Information Management
American Express India Pvt. Ltd.
Gurgaon, India

Re: Problem trying to set up single node cluster on Ubuntu

Posted by shashwat shriparv <dw...@gmail.com>.
use sudo with chown and then chmod to 755

On Sun, Apr 29, 2012 at 11:02 PM, maddy gulati <ac...@gmail.com>wrote:

> I have confirgured the xml configuration files corectly. I tried to repeat
> the installation process with a different user id . Now getting following
> when I execute start-all.sh :
>
>
>
> Warning: $HADOOP_HOME is deprecated.
>
> chown: changing ownership of `/usr/libexec/../logs': Operation not
> permitted
> starting namenode, logging to
> /usr/libexec/../logs/hadoop-maddy-namenode-mandeep-System-Product-Name.out
> /usr/sbin/hadoop-daemon.sh: line 135:
> /usr/libexec/../logs/hadoop-maddy-namenode-mandeep-System-Product-Name.out:
> Permission denied
> head: cannot open
>
> `/usr/libexec/../logs/hadoop-maddy-namenode-mandeep-System-Product-Name.out'
> for reading: No such file or directory
> localhost: chown: changing ownership of `/usr/libexec/../logs': Operation
> not permitted
> localhost: starting datanode, logging to
> /usr/libexec/../logs/hadoop-maddy-datanode-mandeep-System-Product-Name.out
> localhost: /usr/sbin/hadoop-daemon.sh: line 135:
> /usr/libexec/../logs/hadoop-maddy-datanode-mandeep-System-Product-Name.out:
> Permission denied
> localhost: head: cannot open
>
> `/usr/libexec/../logs/hadoop-maddy-datanode-mandeep-System-Product-Name.out'
> for reading: No such file or directory
> localhost: chown: changing ownership of `/usr/libexec/../logs': Operation
> not permitted
> localhost: starting secondarynamenode, logging to
>
> /usr/libexec/../logs/hadoop-maddy-secondarynamenode-mandeep-System-Product-Name.out
> localhost: /usr/sbin/hadoop-daemon.sh: line 135:
>
> /usr/libexec/../logs/hadoop-maddy-secondarynamenode-mandeep-System-Product-Name.out:
> Permission denied
> localhost: head: cannot open
>
> `/usr/libexec/../logs/hadoop-maddy-secondarynamenode-mandeep-System-Product-Name.out'
> for reading: No such file or directory
> chown: changing ownership of `/usr/libexec/../logs': Operation not
> permitted
> starting jobtracker, logging to
>
> /usr/libexec/../logs/hadoop-maddy-jobtracker-mandeep-System-Product-Name.out
> /usr/sbin/hadoop-daemon.sh: line 135:
>
> /usr/libexec/../logs/hadoop-maddy-jobtracker-mandeep-System-Product-Name.out:
> Permission denied
> head: cannot open
>
> `/usr/libexec/../logs/hadoop-maddy-jobtracker-mandeep-System-Product-Name.out'
> for reading: No such file or directory
> localhost: chown: changing ownership of `/usr/libexec/../logs': Operation
> not permitted
> localhost: starting tasktracker, logging to
>
> /usr/libexec/../logs/hadoop-maddy-tasktracker-mandeep-System-Product-Name.out
> localhost: /usr/sbin/hadoop-daemon.sh: line 135:
>
> /usr/libexec/../logs/hadoop-maddy-tasktracker-mandeep-System-Product-Name.out:
> Permission denied
> localhost: head: cannot open
>
> `/usr/libexec/../logs/hadoop-maddy-tasktracker-mandeep-System-Product-Name.out'
> for reading: No such file or directory
>
>
>
>
> Below are the contents of my configuration files
>
>
> Core Site:
> <?xml version="1.0"?>
> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
>
> <!-- Put site-specific property overrides in this file. -->
>
> <configuration>
>
> <property>
>  <name>hadoop.tmp.dir</name>
>  <value>/app/hadoop/tmp</value>
>  <description>A base for other temporary directories.</description>
> </property>
>
> <property>
>  <name>fs.default.name</name>
>  <value>hdfs://localhost:54310</value>
>  <description>The name of the default file system.  A URI whose
>  scheme and authority determine the FileSystem implementation.  The
>  uri's scheme determines the config property (fs.SCHEME.impl) naming
>  the FileSystem implementation class.  The uri's authority is used to
>  determine the host, port, etc. for a filesystem.</description>
> </property>
>
>
> </configuration>
>
>
>
>
>
>
> mapred-site.xml
>
> <?xml version="1.0"?>
> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
>
> <!-- Put site-specific property overrides in this file. -->
>
> <configuration>
>
> <property>
>  <name>mapred.job.tracker</name>
>  <value>localhost:54311</value>
>  <description>The host and port that the MapReduce job tracker runs
>  at.  If "local", then jobs are run in-process as a single map
>  and reduce task.
>  </description>
> </property>
>
> </configuration>
>
>
>
>
>
>
>
> hdfs-site.xml
>
> <?xml version="1.0"?>
> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
>
> <!-- Put site-specific property overrides in this file. -->
>
> <configuration>
>
> <property>
>  <name>dfs.replication</name>
>  <value>1</value>
>  <description>Default block replication.
>  The actual number of replications can be specified when the file is
> created.
>  The default is used if replication is not specified in create time.
>  </description>
> </property>
>
>
>
> </configuration>
>
>
>
>
>
>
>
>
>
>
>
>
> On Sun, Apr 29, 2012 at 4:05 AM, Marcos Ortiz <ml...@uci.cu> wrote:
>
> > Look here
> > http://search-hadoop.com/m/**NRMV72pWYVM1/ERROR+org.apache.**
> > hadoop.hdfs.server.datanode.**DataNode%5C%3A+java.lang.**
> > IllegalArgumentException%5C%**3A+Does+not+contain+a+valid+**
> > host%5C%3Aport+authority%5C%**3A+/v=threaded<
> http://search-hadoop.com/m/NRMV72pWYVM1/ERROR+org.apache.hadoop.hdfs.server.datanode.DataNode%5C%3A+java.lang.IllegalArgumentException%5C%3A+Does+not+contain+a+valid+host%5C%3Aport+authority%5C%3A+/v=threaded
> >
> >
> > To solve this, you can check your config files: core-site.xml,
> > mapred-site.xml and hdfs-site.xml
> >
> > http://www.michael-noll.com/**tutorials/running-hadoop-on-**
> > ubuntu-linux-single-node-**cluster/#conf-site-xml<
> http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-single-node-cluster/#conf-site-xml
> >
> >
> > Of course, yo can check is the official documentation for 1.0.1:
> >
> > http://hadoop.apache.org/**common/docs/r1.0.1/single_**node_setup.html<
> http://hadoop.apache.org/common/docs/r1.0.1/single_node_setup.html>
> >
> > Regards
> >
> >
> > On 4/28/2012 3:19 PM, maddy gulati wrote:
> >
> >> Folks,
> >>
> >> I am trying to set up a single node cluster for Hadoop ( version 1.0.1)
> on
> >> Ubuntu 11.04. After configuring everything, I could not get all the
> >> components to start.
> >>
> >> My installation directory for hadoop is /usr/local/hadoop
> >> I followed the following tutorial for the setup :
> >> http://www.michael-noll.com/**tutorials/running-hadoop-on-**
> >> ubuntu-linux-single-node-**cluster/<
> http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-single-node-cluster/
> >
> >> Below is the snapshot of my log file:
> >>
> >> mandeep@mandeep-System-**Product-Name:/var/log/hadoop/**mandeep$ cat
> >> hadoop-mandeep-datanode-**mandeep-System-Product-Name.**log
> >> 2012-04-29 01:29:37,240 INFO
> >> org.apache.hadoop.hdfs.server.**datanode.DataNode: STARTUP_MSG:
> >> /****************************************************************
> >> STARTUP_MSG: Starting DataNode
> >> STARTUP_MSG:   host = mandeep-System-Product-Name/12**7.0.1.1<
> http://127.0.1.1>
> >> STARTUP_MSG:   args = []
> >> STARTUP_MSG:   version = 1.0.1
> >> STARTUP_MSG:   build =
> >> https://svn.apache.org/repos/**asf/hadoop/common/branches/**branch-1.0<
> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0>-r
> >> 1243785; compiled by 'hortonfo' on Tue Feb 14 08:15:38 UTC 2012
> >> ****************************************************************/
> >> 2012-04-29 01:29:37,398 INFO org.apache.hadoop.metrics2.**
> >> impl.MetricsConfig:
> >> loaded properties from hadoop-metrics2.properties
> >> 2012-04-29 01:29:37,446 INFO
> >> org.apache.hadoop.metrics2.**impl.MetricsSourceAdapter: MBean for source
> >> MetricsSystem,sub=Stats registered.
> >> 2012-04-29 01:29:37,449 INFO
> >> org.apache.hadoop.metrics2.**impl.MetricsSystemImpl: Scheduled snapshot
> >> period at 10 second(s).
> >> 2012-04-29 01:29:37,449 INFO
> >> org.apache.hadoop.metrics2.**impl.MetricsSystemImpl: DataNode metrics
> >> system
> >> started
> >> 2012-04-29 01:29:37,583 INFO
> >> org.apache.hadoop.metrics2.**impl.MetricsSourceAdapter: MBean for source
> >> ugi
> >> registered.
> >> 2012-04-29 01:29:37,586 WARN
> >> org.apache.hadoop.metrics2.**impl.MetricsSystemImpl: Source name ugi
> >> already
> >> exists!
> >> 2012-04-29 01:29:37,615 INFO org.apache.hadoop.util.**NativeCodeLoader:
> >> Loaded the native-hadoop library
> >> 2012-04-29 01:29:37,701 ERROR
> >> org.apache.hadoop.hdfs.server.**datanode.DataNode:
> >> java.lang.**IllegalArgumentException: Does not contain a valid host:port
> >> authority: file:///
> >>  at org.apache.hadoop.net.**NetUtils.createSocketAddr(**
> >> NetUtils.java:162)
> >> at
> >> org.apache.hadoop.hdfs.server.**namenode.NameNode.getAddress(**
> >> NameNode.java:198)
> >>  at
> >> org.apache.hadoop.hdfs.server.**namenode.NameNode.getAddress(**
> >> NameNode.java:228)
> >> at
> >> org.apache.hadoop.hdfs.server.**namenode.NameNode.**
> >> getServiceAddress(NameNode.**java:222)
> >>  at
> >> org.apache.hadoop.hdfs.server.**datanode.DataNode.**
> >> startDataNode(DataNode.java:**337)
> >> at org.apache.hadoop.hdfs.server.**datanode.DataNode.<init>(**
> >> DataNode.java:299)
> >>  at
> >> org.apache.hadoop.hdfs.server.**datanode.DataNode.**
> >> makeInstance(DataNode.java:**1582)
> >> at
> >> org.apache.hadoop.hdfs.server.**datanode.DataNode.**
> >> instantiateDataNode(DataNode.**java:1521)
> >>  at
> >> org.apache.hadoop.hdfs.server.**datanode.DataNode.**
> >> createDataNode(DataNode.java:**1539)
> >> at
> >> org.apache.hadoop.hdfs.server.**datanode.DataNode.secureMain(**
> >> DataNode.java:1665)
> >>  at org.apache.hadoop.hdfs.server.**datanode.DataNode.main(**
> >> DataNode.java:1682)
> >>
> >> 2012-04-29 01:29:37,718 INFO
> >> org.apache.hadoop.hdfs.server.**datanode.DataNode: SHUTDOWN_MSG:
> >> /****************************************************************
> >> SHUTDOWN_MSG: Shutting down DataNode at mandeep-System-Product-Name/
> >> 127.0.1.1
> >> ****************************************************************/
> >> 2012-04-29 01:33:26,907 INFO
> >> org.apache.hadoop.hdfs.server.**datanode.DataNode: STARTUP_MSG:
> >> /****************************************************************
> >>
> >>
> >>
> > --
> > Marcos Luis Ortíz Valmaseda (@marcosluis2186)
> >  Data Engineer at UCI
> >  http://marcosluis2186.**posterous.com<
> http://marcosluis2186.posterous.com>
> >
> > 10mo. ANIVERSARIO DE LA CREACION DE LA UNIVERSIDAD DE LAS CIENCIAS
> > INFORMATICAS...
> > CONECTADOS AL FUTURO, CONECTADOS A LA REVOLUCION
> >
> > http://www.uci.cu
> > http://www.facebook.com/**universidad.uci<
> http://www.facebook.com/universidad.uci>
> > http://www.flickr.com/photos/**universidad_uci<
> http://www.flickr.com/photos/universidad_uci>
> >
>
>
>
> --
> Mandeep Singh Gulati
> System Analyst | WorldWide Risk and Information Management
> American Express India Pvt. Ltd.
> Gurgaon, India
>



-- 


∞
Shashwat Shriparv

Re: Problem trying to set up single node cluster on Ubuntu

Posted by maddy gulati <ac...@gmail.com>.
I have confirgured the xml configuration files corectly. I tried to repeat
the installation process with a different user id . Now getting following
when I execute start-all.sh :



Warning: $HADOOP_HOME is deprecated.

chown: changing ownership of `/usr/libexec/../logs': Operation not permitted
starting namenode, logging to
/usr/libexec/../logs/hadoop-maddy-namenode-mandeep-System-Product-Name.out
/usr/sbin/hadoop-daemon.sh: line 135:
/usr/libexec/../logs/hadoop-maddy-namenode-mandeep-System-Product-Name.out:
Permission denied
head: cannot open
`/usr/libexec/../logs/hadoop-maddy-namenode-mandeep-System-Product-Name.out'
for reading: No such file or directory
localhost: chown: changing ownership of `/usr/libexec/../logs': Operation
not permitted
localhost: starting datanode, logging to
/usr/libexec/../logs/hadoop-maddy-datanode-mandeep-System-Product-Name.out
localhost: /usr/sbin/hadoop-daemon.sh: line 135:
/usr/libexec/../logs/hadoop-maddy-datanode-mandeep-System-Product-Name.out:
Permission denied
localhost: head: cannot open
`/usr/libexec/../logs/hadoop-maddy-datanode-mandeep-System-Product-Name.out'
for reading: No such file or directory
localhost: chown: changing ownership of `/usr/libexec/../logs': Operation
not permitted
localhost: starting secondarynamenode, logging to
/usr/libexec/../logs/hadoop-maddy-secondarynamenode-mandeep-System-Product-Name.out
localhost: /usr/sbin/hadoop-daemon.sh: line 135:
/usr/libexec/../logs/hadoop-maddy-secondarynamenode-mandeep-System-Product-Name.out:
Permission denied
localhost: head: cannot open
`/usr/libexec/../logs/hadoop-maddy-secondarynamenode-mandeep-System-Product-Name.out'
for reading: No such file or directory
chown: changing ownership of `/usr/libexec/../logs': Operation not permitted
starting jobtracker, logging to
/usr/libexec/../logs/hadoop-maddy-jobtracker-mandeep-System-Product-Name.out
/usr/sbin/hadoop-daemon.sh: line 135:
/usr/libexec/../logs/hadoop-maddy-jobtracker-mandeep-System-Product-Name.out:
Permission denied
head: cannot open
`/usr/libexec/../logs/hadoop-maddy-jobtracker-mandeep-System-Product-Name.out'
for reading: No such file or directory
localhost: chown: changing ownership of `/usr/libexec/../logs': Operation
not permitted
localhost: starting tasktracker, logging to
/usr/libexec/../logs/hadoop-maddy-tasktracker-mandeep-System-Product-Name.out
localhost: /usr/sbin/hadoop-daemon.sh: line 135:
/usr/libexec/../logs/hadoop-maddy-tasktracker-mandeep-System-Product-Name.out:
Permission denied
localhost: head: cannot open
`/usr/libexec/../logs/hadoop-maddy-tasktracker-mandeep-System-Product-Name.out'
for reading: No such file or directory




Below are the contents of my configuration files


Core Site:
<?xml version="1.0"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>

<!-- Put site-specific property overrides in this file. -->

<configuration>

<property>
  <name>hadoop.tmp.dir</name>
  <value>/app/hadoop/tmp</value>
  <description>A base for other temporary directories.</description>
</property>

<property>
  <name>fs.default.name</name>
  <value>hdfs://localhost:54310</value>
  <description>The name of the default file system.  A URI whose
  scheme and authority determine the FileSystem implementation.  The
  uri's scheme determines the config property (fs.SCHEME.impl) naming
  the FileSystem implementation class.  The uri's authority is used to
  determine the host, port, etc. for a filesystem.</description>
</property>


</configuration>






mapred-site.xml

<?xml version="1.0"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>

<!-- Put site-specific property overrides in this file. -->

<configuration>

<property>
  <name>mapred.job.tracker</name>
  <value>localhost:54311</value>
  <description>The host and port that the MapReduce job tracker runs
  at.  If "local", then jobs are run in-process as a single map
  and reduce task.
  </description>
</property>

</configuration>







hdfs-site.xml

<?xml version="1.0"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>

<!-- Put site-specific property overrides in this file. -->

<configuration>

<property>
  <name>dfs.replication</name>
  <value>1</value>
  <description>Default block replication.
  The actual number of replications can be specified when the file is
created.
  The default is used if replication is not specified in create time.
  </description>
</property>



</configuration>












On Sun, Apr 29, 2012 at 4:05 AM, Marcos Ortiz <ml...@uci.cu> wrote:

> Look here
> http://search-hadoop.com/m/**NRMV72pWYVM1/ERROR+org.apache.**
> hadoop.hdfs.server.datanode.**DataNode%5C%3A+java.lang.**
> IllegalArgumentException%5C%**3A+Does+not+contain+a+valid+**
> host%5C%3Aport+authority%5C%**3A+/v=threaded<http://search-hadoop.com/m/NRMV72pWYVM1/ERROR+org.apache.hadoop.hdfs.server.datanode.DataNode%5C%3A+java.lang.IllegalArgumentException%5C%3A+Does+not+contain+a+valid+host%5C%3Aport+authority%5C%3A+/v=threaded>
>
> To solve this, you can check your config files: core-site.xml,
> mapred-site.xml and hdfs-site.xml
>
> http://www.michael-noll.com/**tutorials/running-hadoop-on-**
> ubuntu-linux-single-node-**cluster/#conf-site-xml<http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-single-node-cluster/#conf-site-xml>
>
> Of course, yo can check is the official documentation for 1.0.1:
>
> http://hadoop.apache.org/**common/docs/r1.0.1/single_**node_setup.html<http://hadoop.apache.org/common/docs/r1.0.1/single_node_setup.html>
>
> Regards
>
>
> On 4/28/2012 3:19 PM, maddy gulati wrote:
>
>> Folks,
>>
>> I am trying to set up a single node cluster for Hadoop ( version 1.0.1) on
>> Ubuntu 11.04. After configuring everything, I could not get all the
>> components to start.
>>
>> My installation directory for hadoop is /usr/local/hadoop
>> I followed the following tutorial for the setup :
>> http://www.michael-noll.com/**tutorials/running-hadoop-on-**
>> ubuntu-linux-single-node-**cluster/<http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-single-node-cluster/>
>> Below is the snapshot of my log file:
>>
>> mandeep@mandeep-System-**Product-Name:/var/log/hadoop/**mandeep$ cat
>> hadoop-mandeep-datanode-**mandeep-System-Product-Name.**log
>> 2012-04-29 01:29:37,240 INFO
>> org.apache.hadoop.hdfs.server.**datanode.DataNode: STARTUP_MSG:
>> /****************************************************************
>> STARTUP_MSG: Starting DataNode
>> STARTUP_MSG:   host = mandeep-System-Product-Name/12**7.0.1.1<http://127.0.1.1>
>> STARTUP_MSG:   args = []
>> STARTUP_MSG:   version = 1.0.1
>> STARTUP_MSG:   build =
>> https://svn.apache.org/repos/**asf/hadoop/common/branches/**branch-1.0<https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0>-r
>> 1243785; compiled by 'hortonfo' on Tue Feb 14 08:15:38 UTC 2012
>> ****************************************************************/
>> 2012-04-29 01:29:37,398 INFO org.apache.hadoop.metrics2.**
>> impl.MetricsConfig:
>> loaded properties from hadoop-metrics2.properties
>> 2012-04-29 01:29:37,446 INFO
>> org.apache.hadoop.metrics2.**impl.MetricsSourceAdapter: MBean for source
>> MetricsSystem,sub=Stats registered.
>> 2012-04-29 01:29:37,449 INFO
>> org.apache.hadoop.metrics2.**impl.MetricsSystemImpl: Scheduled snapshot
>> period at 10 second(s).
>> 2012-04-29 01:29:37,449 INFO
>> org.apache.hadoop.metrics2.**impl.MetricsSystemImpl: DataNode metrics
>> system
>> started
>> 2012-04-29 01:29:37,583 INFO
>> org.apache.hadoop.metrics2.**impl.MetricsSourceAdapter: MBean for source
>> ugi
>> registered.
>> 2012-04-29 01:29:37,586 WARN
>> org.apache.hadoop.metrics2.**impl.MetricsSystemImpl: Source name ugi
>> already
>> exists!
>> 2012-04-29 01:29:37,615 INFO org.apache.hadoop.util.**NativeCodeLoader:
>> Loaded the native-hadoop library
>> 2012-04-29 01:29:37,701 ERROR
>> org.apache.hadoop.hdfs.server.**datanode.DataNode:
>> java.lang.**IllegalArgumentException: Does not contain a valid host:port
>> authority: file:///
>>  at org.apache.hadoop.net.**NetUtils.createSocketAddr(**
>> NetUtils.java:162)
>> at
>> org.apache.hadoop.hdfs.server.**namenode.NameNode.getAddress(**
>> NameNode.java:198)
>>  at
>> org.apache.hadoop.hdfs.server.**namenode.NameNode.getAddress(**
>> NameNode.java:228)
>> at
>> org.apache.hadoop.hdfs.server.**namenode.NameNode.**
>> getServiceAddress(NameNode.**java:222)
>>  at
>> org.apache.hadoop.hdfs.server.**datanode.DataNode.**
>> startDataNode(DataNode.java:**337)
>> at org.apache.hadoop.hdfs.server.**datanode.DataNode.<init>(**
>> DataNode.java:299)
>>  at
>> org.apache.hadoop.hdfs.server.**datanode.DataNode.**
>> makeInstance(DataNode.java:**1582)
>> at
>> org.apache.hadoop.hdfs.server.**datanode.DataNode.**
>> instantiateDataNode(DataNode.**java:1521)
>>  at
>> org.apache.hadoop.hdfs.server.**datanode.DataNode.**
>> createDataNode(DataNode.java:**1539)
>> at
>> org.apache.hadoop.hdfs.server.**datanode.DataNode.secureMain(**
>> DataNode.java:1665)
>>  at org.apache.hadoop.hdfs.server.**datanode.DataNode.main(**
>> DataNode.java:1682)
>>
>> 2012-04-29 01:29:37,718 INFO
>> org.apache.hadoop.hdfs.server.**datanode.DataNode: SHUTDOWN_MSG:
>> /****************************************************************
>> SHUTDOWN_MSG: Shutting down DataNode at mandeep-System-Product-Name/
>> 127.0.1.1
>> ****************************************************************/
>> 2012-04-29 01:33:26,907 INFO
>> org.apache.hadoop.hdfs.server.**datanode.DataNode: STARTUP_MSG:
>> /****************************************************************
>>
>>
>>
> --
> Marcos Luis Ortíz Valmaseda (@marcosluis2186)
>  Data Engineer at UCI
>  http://marcosluis2186.**posterous.com<http://marcosluis2186.posterous.com>
>
> 10mo. ANIVERSARIO DE LA CREACION DE LA UNIVERSIDAD DE LAS CIENCIAS
> INFORMATICAS...
> CONECTADOS AL FUTURO, CONECTADOS A LA REVOLUCION
>
> http://www.uci.cu
> http://www.facebook.com/**universidad.uci<http://www.facebook.com/universidad.uci>
> http://www.flickr.com/photos/**universidad_uci<http://www.flickr.com/photos/universidad_uci>
>



-- 
Mandeep Singh Gulati
System Analyst | WorldWide Risk and Information Management
American Express India Pvt. Ltd.
Gurgaon, India

Re: Problem trying to set up single node cluster on Ubuntu

Posted by Marcos Ortiz <ml...@uci.cu>.
Look here
http://search-hadoop.com/m/NRMV72pWYVM1/ERROR+org.apache.hadoop.hdfs.server.datanode.DataNode%5C%3A+java.lang.IllegalArgumentException%5C%3A+Does+not+contain+a+valid+host%5C%3Aport+authority%5C%3A+/v=threaded

To solve this, you can check your config files: core-site.xml, 
mapred-site.xml and hdfs-site.xml

http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-single-node-cluster/#conf-site-xml

Of course, yo can check is the official documentation for 1.0.1:

http://hadoop.apache.org/common/docs/r1.0.1/single_node_setup.html

Regards

On 4/28/2012 3:19 PM, maddy gulati wrote:
> Folks,
>
> I am trying to set up a single node cluster for Hadoop ( version 1.0.1) on
> Ubuntu 11.04. After configuring everything, I could not get all the
> components to start.
>
> My installation directory for hadoop is /usr/local/hadoop
> I followed the following tutorial for the setup :
> http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-single-node-cluster/
> Below is the snapshot of my log file:
>
> mandeep@mandeep-System-Product-Name:/var/log/hadoop/mandeep$ cat
> hadoop-mandeep-datanode-mandeep-System-Product-Name.log
> 2012-04-29 01:29:37,240 INFO
> org.apache.hadoop.hdfs.server.datanode.DataNode: STARTUP_MSG:
> /************************************************************
> STARTUP_MSG: Starting DataNode
> STARTUP_MSG:   host = mandeep-System-Product-Name/127.0.1.1
> STARTUP_MSG:   args = []
> STARTUP_MSG:   version = 1.0.1
> STARTUP_MSG:   build =
> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0 -r
> 1243785; compiled by 'hortonfo' on Tue Feb 14 08:15:38 UTC 2012
> ************************************************************/
> 2012-04-29 01:29:37,398 INFO org.apache.hadoop.metrics2.impl.MetricsConfig:
> loaded properties from hadoop-metrics2.properties
> 2012-04-29 01:29:37,446 INFO
> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
> MetricsSystem,sub=Stats registered.
> 2012-04-29 01:29:37,449 INFO
> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot
> period at 10 second(s).
> 2012-04-29 01:29:37,449 INFO
> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: DataNode metrics system
> started
> 2012-04-29 01:29:37,583 INFO
> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source ugi
> registered.
> 2012-04-29 01:29:37,586 WARN
> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Source name ugi already
> exists!
> 2012-04-29 01:29:37,615 INFO org.apache.hadoop.util.NativeCodeLoader:
> Loaded the native-hadoop library
> 2012-04-29 01:29:37,701 ERROR
> org.apache.hadoop.hdfs.server.datanode.DataNode:
> java.lang.IllegalArgumentException: Does not contain a valid host:port
> authority: file:///
>   at org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:162)
> at
> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:198)
>   at
> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:228)
> at
> org.apache.hadoop.hdfs.server.namenode.NameNode.getServiceAddress(NameNode.java:222)
>   at
> org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:337)
> at org.apache.hadoop.hdfs.server.datanode.DataNode.<init>(DataNode.java:299)
>   at
> org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1582)
> at
> org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1521)
>   at
> org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1539)
> at
> org.apache.hadoop.hdfs.server.datanode.DataNode.secureMain(DataNode.java:1665)
>   at org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:1682)
>
> 2012-04-29 01:29:37,718 INFO
> org.apache.hadoop.hdfs.server.datanode.DataNode: SHUTDOWN_MSG:
> /************************************************************
> SHUTDOWN_MSG: Shutting down DataNode at mandeep-System-Product-Name/
> 127.0.1.1
> ************************************************************/
> 2012-04-29 01:33:26,907 INFO
> org.apache.hadoop.hdfs.server.datanode.DataNode: STARTUP_MSG:
> /************************************************************
>
>

-- 
Marcos Luis Ortíz Valmaseda (@marcosluis2186)
  Data Engineer at UCI
  http://marcosluis2186.posterous.com

10mo. ANIVERSARIO DE LA CREACION DE LA UNIVERSIDAD DE LAS CIENCIAS INFORMATICAS...
CONECTADOS AL FUTURO, CONECTADOS A LA REVOLUCION

http://www.uci.cu
http://www.facebook.com/universidad.uci
http://www.flickr.com/photos/universidad_uci