You are viewing a plain text version of this content. The canonical link for it is here.
Posted to hdfs-dev@hadoop.apache.org by "xyzzy (JIRA)" <ji...@apache.org> on 2014/04/14 02:52:14 UTC

[jira] [Created] (HDFS-6239) start-dfs.sh does not start remote DataNode due to escape characters

xyzzy created HDFS-6239:
---------------------------

             Summary: start-dfs.sh does not start remote DataNode due to escape characters
                 Key: HDFS-6239
                 URL: https://issues.apache.org/jira/browse/HDFS-6239
             Project: Hadoop HDFS
          Issue Type: Bug
          Components: scripts
    Affects Versions: 1.2.1
         Environment: GNU bash, version 4.1.2(1)-release (x86_64-redhat-linux-gnu)

Linux foo 2.6.32-431.3.1.el6.x86_64 #1 SMP Fri Dec 13 06:58:20 EST 2013 x86_64 x86_64 x86_64 GNU/Linux

            Reporter: xyzzy


start-dfs.sh fails to start remote data nodes and task nodes, though it is possible to start them manually through hadoop-daemon.sh.

I've been able to debug and find the root cause the bug, and I thought it was a trivial fix, but I do not know how to do it.

This is the issue when I debug using bash -x:

hadoop-daemons.sh calls slave.sh:

exec "$bin/slaves.sh" --config $HADOOP_CONF_DIR cd "$HADOOP_HOME" \; "$bin/hadoop-daemon.sh" --config $HADOOP_CONF_DIR "$@"

In slaves.sh, it becomes:

ssh foo cd /foo/h/libexec/.. ';' /foo/hcain/bin/hadoop-daemon.sh --config /foo/h/libexec/../conf start datanode

The problem is ';' . Because the semi-colon is surrounded by quotes, it doesn't execute the code after that. I manually ran the above command, and as expected the data node did not start. When I removed the quotes around the semi-colon, everything works. Please note that you can see the issue only when you do bash -x. If you echo the statement, the quotes around the semi-colon are not visible.

This issue is always reproducible for me, and because of it, I have to manually start daemons on each machine. 



--
This message was sent by Atlassian JIRA
(v6.2#6252)