You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-user@hadoop.apache.org by Edson Ramiro <er...@gmail.com> on 2010/03/29 21:01:02 UTC

java.io.IOException: Function not implemented

Hi all,

I'm trying to install Hadoop on a cluster, but I'm getting this error.

I'm using java version "1.6.0_17" and hadoop-0.20.1+169.56.tar.gz from
Cloudera.

Its running in a NFS home shared between the nodes and masters.

The NameNode works well, but all nodes try to connect and fail.

Any Idea ?

Thanks in Advance.

==> logs/hadoop-ramiro-datanode-a05.log <==
2010-03-29 15:56:00,168 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: lcpad/192.168.1.51:9000. Already tried 0 time(s).
2010-03-29 15:56:01,172 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: lcpad/192.168.1.51:9000. Already tried 1 time(s).
2010-03-29 15:56:02,176 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: lcpad/192.168.1.51:9000. Already tried 2 time(s).
2010-03-29 15:56:03,180 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: lcpad/192.168.1.51:9000. Already tried 3 time(s).
2010-03-29 15:56:04,184 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: lcpad/192.168.1.51:9000. Already tried 4 time(s).
2010-03-29 15:56:05,188 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: lcpad/192.168.1.51:9000. Already tried 5 time(s).
2010-03-29 15:56:06,192 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: lcpad/192.168.1.51:9000. Already tried 6 time(s).
2010-03-29 15:56:07,196 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: lcpad/192.168.1.51:9000. Already tried 7 time(s).
2010-03-29 15:56:08,200 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: lcpad/192.168.1.51:9000. Already tried 8 time(s).
2010-03-29 15:56:09,204 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: lcpad/192.168.1.51:9000. Already tried 9 time(s).
2010-03-29 15:56:09,204 ERROR
org.apache.hadoop.hdfs.server.datanode.DataNode: java.io.IOException: Call
to lcpad/192.168.1.51:9000 failed on local exception: java.io.IOException:
Function not implemented
        at org.apache.hadoop.ipc.Client.wrapException(Client.java:775)
        at org.apache.hadoop.ipc.Client.call(Client.java:743)
        at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220)
        at $Proxy4.getProtocolVersion(Unknown Source)
        at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:359)
        at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:346)
        at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:383)
        at org.apache.hadoop.ipc.RPC.waitForProxy(RPC.java:314)
        at org.apache.hadoop.ipc.RPC.waitForProxy(RPC.java:291)
        at
org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:278)
        at
org.apache.hadoop.hdfs.server.datanode.DataNode.<init>(DataNode.java:225)
        at
org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1309)
        at
org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1264)
        at
org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1272)
        at
org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:1394)
Caused by: java.io.IOException: Function not implemented
        at sun.nio.ch.EPollArrayWrapper.epollCreate(Native Method)
        at sun.nio.ch.EPollArrayWrapper.<init>(EPollArrayWrapper.java:68)
        at sun.nio.ch.EPollSelectorImpl.<init>(EPollSelectorImpl.java:52)
        at
sun.nio.ch.EPollSelectorProvider.openSelector(EPollSelectorProvider.java:18)
        at
org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.get(SocketIOWithTimeout.java:407)
        at
org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:322)
        at
org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:203)
        at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:407)
        at
org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:304)
        at
org.apache.hadoop.ipc.Client$Connection.access$1700(Client.java:176)
        at org.apache.hadoop.ipc.Client.getConnection(Client.java:860)
        at org.apache.hadoop.ipc.Client.call(Client.java:720)
        ... 13 more

Edson Ramiro

Re: java.io.IOException: Function not implemented

Posted by Todd Lipcon <to...@cloudera.com>.
Hey Edson,

Unfortunately I'm not sure what's going on here - for whatever reason, the
kernel isn't allowing Java NIO to use epoll, and thus the IPC framework from
Hadoop isn't working correctly. I don't think this is a hadoop specific bug.

Does this issue occur on all of the nodes?

-Todd

On Mon, Mar 29, 2010 at 2:26 PM, Edson Ramiro <er...@gmail.com> wrote:

> I'm not involved with Debian community :(
>
> ramiro@h02:~/hadoop$ cat /proc/sys/fs/epoll/max_user_watches
> 3373957
>
> and the Java is not the OpenSDK.
> The version is:
>
> ramiro@lcpad:/usr/lib/jvm/java-6-sun$ java -version
> java version "1.6.0_17"
> Java(TM) SE Runtime Environment (build 1.6.0_17-b04)
> Java HotSpot(TM) 64-Bit Server VM (build 14.3-b01, mixed mode)
>
> Edson Ramiro
>
>
> On 29 March 2010 17:14, Todd Lipcon <to...@cloudera.com> wrote:
>
> > Hi Edson,
> >
> > It looks like for some reason your kernel does not have epoll enabled.
> It's
> > very strange, since your kernel is very recent (in fact, bleeding edge!)
> >
> > Can you check the contents of /proc/sys/fs/epoll/max_user_watches
> >
> > Are you involved with the Debian community? This sounds like a general
> Java
> > bug. Can you also please verify that you're using the Sun JVM and not
> > OpenJDK (the debian folks like OpenJDK but it has subtle issues with
> > Hadoop)
> > You'll have to add a non-free repository and install sun-java6-jdk
> >
> > -Todd
> >
> > On Mon, Mar 29, 2010 at 1:05 PM, Edson Ramiro <er...@gmail.com>
> wrote:
> >
> > > I'm using
> > >
> > > Linux h02 2.6.32.9 #2 SMP Sat Mar 6 19:09:13 BRT 2010 x86_64 GNU/Linux
> > >
> > > ramiro@h02:~/hadoop$ cat /etc/debian_version
> > > squeeze/sid
> > >
> > > Thanks for reply
> > >
> > > Edson Ramiro
> > >
> > >
> > > On 29 March 2010 16:56, Todd Lipcon <to...@cloudera.com> wrote:
> > >
> > > > Hi Edson,
> > > >
> > > > What operating system are you on? What kernel version?
> > > >
> > > > Thanks
> > > > -Todd
> > > >
> > > > On Mon, Mar 29, 2010 at 12:01 PM, Edson Ramiro <er...@gmail.com>
> > > wrote:
> > > >
> > > > > Hi all,
> > > > >
> > > > > I'm trying to install Hadoop on a cluster, but I'm getting this
> > error.
> > > > >
> > > > > I'm using java version "1.6.0_17" and hadoop-0.20.1+169.56.tar.gz
> > from
> > > > > Cloudera.
> > > > >
> > > > > Its running in a NFS home shared between the nodes and masters.
> > > > >
> > > > > The NameNode works well, but all nodes try to connect and fail.
> > > > >
> > > > > Any Idea ?
> > > > >
> > > > > Thanks in Advance.
> > > > >
> > > > > ==> logs/hadoop-ramiro-datanode-a05.log <==
> > > > > 2010-03-29 15:56:00,168 INFO org.apache.hadoop.ipc.Client: Retrying
> > > > connect
> > > > > to server: lcpad/192.168.1.51:9000. Already tried 0 time(s).
> > > > > 2010-03-29 15:56:01,172 INFO org.apache.hadoop.ipc.Client: Retrying
> > > > connect
> > > > > to server: lcpad/192.168.1.51:9000. Already tried 1 time(s).
> > > > > 2010-03-29 15:56:02,176 INFO org.apache.hadoop.ipc.Client: Retrying
> > > > connect
> > > > > to server: lcpad/192.168.1.51:9000. Already tried 2 time(s).
> > > > > 2010-03-29 15:56:03,180 INFO org.apache.hadoop.ipc.Client: Retrying
> > > > connect
> > > > > to server: lcpad/192.168.1.51:9000. Already tried 3 time(s).
> > > > > 2010-03-29 15:56:04,184 INFO org.apache.hadoop.ipc.Client: Retrying
> > > > connect
> > > > > to server: lcpad/192.168.1.51:9000. Already tried 4 time(s).
> > > > > 2010-03-29 15:56:05,188 INFO org.apache.hadoop.ipc.Client: Retrying
> > > > connect
> > > > > to server: lcpad/192.168.1.51:9000. Already tried 5 time(s).
> > > > > 2010-03-29 15:56:06,192 INFO org.apache.hadoop.ipc.Client: Retrying
> > > > connect
> > > > > to server: lcpad/192.168.1.51:9000. Already tried 6 time(s).
> > > > > 2010-03-29 15:56:07,196 INFO org.apache.hadoop.ipc.Client: Retrying
> > > > connect
> > > > > to server: lcpad/192.168.1.51:9000. Already tried 7 time(s).
> > > > > 2010-03-29 15:56:08,200 INFO org.apache.hadoop.ipc.Client: Retrying
> > > > connect
> > > > > to server: lcpad/192.168.1.51:9000. Already tried 8 time(s).
> > > > > 2010-03-29 15:56:09,204 INFO org.apache.hadoop.ipc.Client: Retrying
> > > > connect
> > > > > to server: lcpad/192.168.1.51:9000. Already tried 9 time(s).
> > > > > 2010-03-29 15:56:09,204 ERROR
> > > > > org.apache.hadoop.hdfs.server.datanode.DataNode:
> java.io.IOException:
> > > > Call
> > > > > to lcpad/192.168.1.51:9000 failed on local exception:
> > > > java.io.IOException:
> > > > > Function not implemented
> > > > >        at
> org.apache.hadoop.ipc.Client.wrapException(Client.java:775)
> > > > >        at org.apache.hadoop.ipc.Client.call(Client.java:743)
> > > > >        at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220)
> > > > >        at $Proxy4.getProtocolVersion(Unknown Source)
> > > > >        at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:359)
> > > > >        at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:346)
> > > > >        at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:383)
> > > > >        at org.apache.hadoop.ipc.RPC.waitForProxy(RPC.java:314)
> > > > >        at org.apache.hadoop.ipc.RPC.waitForProxy(RPC.java:291)
> > > > >        at
> > > > >
> > > > >
> > > >
> > >
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:278)
> > > > >        at
> > > > >
> > >
> org.apache.hadoop.hdfs.server.datanode.DataNode.<init>(DataNode.java:225)
> > > > >        at
> > > > >
> > > > >
> > > >
> > >
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1309)
> > > > >        at
> > > > >
> > > > >
> > > >
> > >
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1264)
> > > > >        at
> > > > >
> > > > >
> > > >
> > >
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1272)
> > > > >        at
> > > > >
> > >
> org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:1394)
> > > > > Caused by: java.io.IOException: Function not implemented
> > > > >        at sun.nio.ch.EPollArrayWrapper.epollCreate(Native Method)
> > > > >        at
> > > sun.nio.ch.EPollArrayWrapper.<init>(EPollArrayWrapper.java:68)
> > > > >        at
> > > sun.nio.ch.EPollSelectorImpl.<init>(EPollSelectorImpl.java:52)
> > > > >        at
> > > > >
> > > > >
> > > >
> > >
> >
> sun.nio.ch.EPollSelectorProvider.openSelector(EPollSelectorProvider.java:18)
> > > > >        at
> > > > >
> > > > >
> > > >
> > >
> >
> org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.get(SocketIOWithTimeout.java:407)
> > > > >        at
> > > > >
> > > > >
> > > >
> > >
> >
> org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:322)
> > > > >        at
> > > > >
> > > > >
> > > >
> > >
> >
> org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:203)
> > > > >        at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:407)
> > > > >        at
> > > > >
> > org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:304)
> > > > >        at
> > > > >
> org.apache.hadoop.ipc.Client$Connection.access$1700(Client.java:176)
> > > > >        at
> org.apache.hadoop.ipc.Client.getConnection(Client.java:860)
> > > > >        at org.apache.hadoop.ipc.Client.call(Client.java:720)
> > > > >        ... 13 more
> > > > >
> > > > > Edson Ramiro
> > > > >
> > > >
> > > >
> > > >
> > > > --
> > > > Todd Lipcon
> > > > Software Engineer, Cloudera
> > > >
> > >
> >
> >
> >
> > --
> > Todd Lipcon
> > Software Engineer, Cloudera
> >
>



-- 
Todd Lipcon
Software Engineer, Cloudera

Re: java.io.IOException: Function not implemented

Posted by Edson Ramiro <er...@gmail.com>.
Hi,

The admin enable the epoll, so the Hadoop is not more showing the error
message 'Function not implement'.

Thanks all,

Edson Ramiro


On 31 March 2010 07:52, Steve Loughran <st...@apache.org> wrote:

> Edson Ramiro wrote:
>
>> May be it's a bug.
>>
>> I'm not the admin. : (
>>
>> so, I'll talk to him and may be he install a 2.6.32.9 in another node to
>> test  : )
>>
>> Thanks
>>
>> Edson Ramiro
>>
>>
>> On 30 March 2010 20:00, Todd Lipcon <to...@cloudera.com> wrote:
>>
>>  Hi Edson,
>>>
>>> I noticed that only the h01 nodes are running 2.6.32.9, the other broken
>>> DNs
>>> are 2.6.32.10.
>>>
>>> Is there some reason you are running a kernel that is literally 2 weeks
>>> old?
>>> I wouldn't be at all surprised if there were a bug here, or some issue
>>> with
>>> your Debian "unstable" distribution...
>>>
>>>
> If you are running the SCM trunk of the OS, you are part of the dev team.
> They will be grateful for the bugs you find and fix, but you get to find and
> fix them. In Ant one bugrep was <touch> stopped setting dates in the past,
> turned out that on the debian nightly builds, you couldn't touch any file
> into the past...
>
> -steve
>
>
>
>

Re: java.io.IOException: Function not implemented

Posted by Steve Loughran <st...@apache.org>.
Edson Ramiro wrote:
> May be it's a bug.
> 
> I'm not the admin. : (
> 
> so, I'll talk to him and may be he install a 2.6.32.9 in another node to
> test  : )
> 
> Thanks
> 
> Edson Ramiro
> 
> 
> On 30 March 2010 20:00, Todd Lipcon <to...@cloudera.com> wrote:
> 
>> Hi Edson,
>>
>> I noticed that only the h01 nodes are running 2.6.32.9, the other broken
>> DNs
>> are 2.6.32.10.
>>
>> Is there some reason you are running a kernel that is literally 2 weeks
>> old?
>> I wouldn't be at all surprised if there were a bug here, or some issue with
>> your Debian "unstable" distribution...
>>

If you are running the SCM trunk of the OS, you are part of the dev 
team. They will be grateful for the bugs you find and fix, but you get 
to find and fix them. In Ant one bugrep was <touch> stopped setting 
dates in the past, turned out that on the debian nightly builds, you 
couldn't touch any file into the past...

-steve




Re: java.io.IOException: Function not implemented

Posted by Edson Ramiro <er...@gmail.com>.
May be it's a bug.

I'm not the admin. : (

so, I'll talk to him and may be he install a 2.6.32.9 in another node to
test  : )

Thanks

Edson Ramiro


On 30 March 2010 20:00, Todd Lipcon <to...@cloudera.com> wrote:

> Hi Edson,
>
> I noticed that only the h01 nodes are running 2.6.32.9, the other broken
> DNs
> are 2.6.32.10.
>
> Is there some reason you are running a kernel that is literally 2 weeks
> old?
> I wouldn't be at all surprised if there were a bug here, or some issue with
> your Debian "unstable" distribution...
>
> -Todd
>
> On Tue, Mar 30, 2010 at 3:54 PM, Edson Ramiro <er...@gmail.com> wrote:
>
> > Hi all,
> >
> > Thanks for help Todd and Steve,
> >
> > I configured Hadoop (0.20.2) again and I'm getting the same error
> (Function
> > not implemented).
> >
> > Do you think it's a Hadoop bug?
> >
> > This is the situation:
> >
> > I've 28 nodes where just four are running the datanode.
> >
> > In all other nodes the tasktracker in running ok.
> >
> > The NN and JT are running ok.
> >
> > The configuration of the machines is the same, its a nfs shared home.
> >
> > In all machines the Java version is "1.6.0_17".
> >
> > This is the kernel version of the nodes, note that are two versions and
> in
> > both the
> > datanode doesn't work. Just in the h0* machines.
> >
> > ramiro@lcpad:~/hadoop-0.20.2$ ./bin/slaves.sh uname -a  | sort
> > a01: Linux a01 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64
> GNU/Linux
> > a02: Linux a02 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64
> GNU/Linux
> > a03: Linux a03 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64
> GNU/Linux
> > a04: Linux a04 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64
> GNU/Linux
> > a05: Linux a05 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64
> GNU/Linux
> > a06: Linux a06 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64
> GNU/Linux
> > a07: Linux a07 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64
> GNU/Linux
> > a09: Linux a09 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64
> GNU/Linux
> > a10: Linux a10 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64
> GNU/Linux
> > ag06: Linux ag06 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
> > GNU/Linux
> > ag07: Linux ag07 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
> > GNU/Linux
> > bl02: Linux bl02 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
> > GNU/Linux
> > bl03: Linux bl03 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
> > GNU/Linux
> > bl04: Linux bl04 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
> > GNU/Linux
> > bl06: Linux bl06 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
> > GNU/Linux
> > bl07: Linux bl07 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
> > GNU/Linux
> > ct02: Linux ct02 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
> > GNU/Linux
> > ct03: Linux ct03 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
> > GNU/Linux
> > ct04: Linux ct04 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
> > GNU/Linux
> > ct06: Linux ct06 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
> > GNU/Linux
> > h01: Linux h01 2.6.32.9 #2 SMP Sat Mar 6 19:09:13 BRT 2010 x86_64
> GNU/Linux
> > h02: Linux h02 2.6.32.9 #2 SMP Sat Mar 6 19:09:13 BRT 2010 x86_64
> GNU/Linux
> > h03: Linux h03 2.6.32.9 #2 SMP Sat Mar 6 19:09:13 BRT 2010 x86_64
> GNU/Linux
> > h04: Linux h04 2.6.32.9 #2 SMP Sat Mar 6 19:09:13 BRT 2010 x86_64
> GNU/Linux
> > sd02: Linux sd02 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
> > GNU/Linux
> > sd05: Linux sd05 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
> > GNU/Linux
> > sd06: Linux sd06 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
> > GNU/Linux
> > sd07: Linux sd07 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
> > GNU/Linux
> >
> >
> > These are the java processes running on each clients.
> > Jjust the h0* machines are running ok.
> >
> > ramiro@lcpad:~/hadoop-0.20.2$ ./bin/slaves.sh pgrep -lc java | sort
> > a01: 1
> > a02: 1
> > a03: 1
> > a04: 1
> > a05: 1
> > a06: 1
> > a07: 1
> > a09: 1
> > a10: 1
> > ag06: 1
> > ag07: 1
> > bl02: 1
> > bl03: 1
> > bl04: 1
> > bl06: 1
> > bl07: 1
> > ct02: 1
> > ct03: 1
> > ct04: 1
> > ct06: 1
> > h01: 2
> > h02: 2
> > h03: 2
> > h04: 2
> > sd02: 1
> > sd05: 1
> > sd06: 1
> > sd07: 1
> >
> > This is my configuration:
> >
> > ramiro@lcpad:~/hadoop-0.20.2$ cat conf/*site*
> > <?xml version="1.0"?>
> > <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
> >
> > <!-- Put site-specific property overrides in this file. -->
> >
> > <configuration>
> > <property>
> > <name>fs.default.name</name>
> > <value>hdfs://lcpad:9000</value>
> > </property>
> > </configuration>
> > <?xml version="1.0"?>
> > <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
> >
> > <!-- Put site-specific property overrides in this file. -->
> >
> > <configuration>
> > <property>
> > <name>dfs.replication</name>
> > <value>1</value>
> > </property>
> > </configuration>
> > <?xml version="1.0"?>
> > <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
> >
> > <!-- Put site-specific property overrides in this file. -->
> >
> > <configuration>
> >  <property>
> >    <name>mapred.job.tracker</name>
> >    <value>lcpad:9001</value>
> >  </property>
> > </configuration>
> >
> > Thanks in Advance,
> >
> > Edson Ramiro
> >
> >
> > On 30 March 2010 05:58, Steve Loughran <st...@apache.org> wrote:
> >
> > > Edson Ramiro wrote:
> > >
> > >> I'm not involved with Debian community :(
> > >>
> > >
> > > I think you are now...
> > >
> >
>
>
>
> --
> Todd Lipcon
> Software Engineer, Cloudera
>

Re: java.io.IOException: Function not implemented

Posted by Todd Lipcon <to...@cloudera.com>.
Hi Edson,

I noticed that only the h01 nodes are running 2.6.32.9, the other broken DNs
are 2.6.32.10.

Is there some reason you are running a kernel that is literally 2 weeks old?
I wouldn't be at all surprised if there were a bug here, or some issue with
your Debian "unstable" distribution...

-Todd

On Tue, Mar 30, 2010 at 3:54 PM, Edson Ramiro <er...@gmail.com> wrote:

> Hi all,
>
> Thanks for help Todd and Steve,
>
> I configured Hadoop (0.20.2) again and I'm getting the same error (Function
> not implemented).
>
> Do you think it's a Hadoop bug?
>
> This is the situation:
>
> I've 28 nodes where just four are running the datanode.
>
> In all other nodes the tasktracker in running ok.
>
> The NN and JT are running ok.
>
> The configuration of the machines is the same, its a nfs shared home.
>
> In all machines the Java version is "1.6.0_17".
>
> This is the kernel version of the nodes, note that are two versions and in
> both the
> datanode doesn't work. Just in the h0* machines.
>
> ramiro@lcpad:~/hadoop-0.20.2$ ./bin/slaves.sh uname -a  | sort
> a01: Linux a01 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64 GNU/Linux
> a02: Linux a02 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64 GNU/Linux
> a03: Linux a03 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64 GNU/Linux
> a04: Linux a04 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64 GNU/Linux
> a05: Linux a05 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64 GNU/Linux
> a06: Linux a06 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64 GNU/Linux
> a07: Linux a07 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64 GNU/Linux
> a09: Linux a09 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64 GNU/Linux
> a10: Linux a10 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64 GNU/Linux
> ag06: Linux ag06 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
> GNU/Linux
> ag07: Linux ag07 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
> GNU/Linux
> bl02: Linux bl02 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
> GNU/Linux
> bl03: Linux bl03 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
> GNU/Linux
> bl04: Linux bl04 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
> GNU/Linux
> bl06: Linux bl06 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
> GNU/Linux
> bl07: Linux bl07 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
> GNU/Linux
> ct02: Linux ct02 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
> GNU/Linux
> ct03: Linux ct03 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
> GNU/Linux
> ct04: Linux ct04 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
> GNU/Linux
> ct06: Linux ct06 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
> GNU/Linux
> h01: Linux h01 2.6.32.9 #2 SMP Sat Mar 6 19:09:13 BRT 2010 x86_64 GNU/Linux
> h02: Linux h02 2.6.32.9 #2 SMP Sat Mar 6 19:09:13 BRT 2010 x86_64 GNU/Linux
> h03: Linux h03 2.6.32.9 #2 SMP Sat Mar 6 19:09:13 BRT 2010 x86_64 GNU/Linux
> h04: Linux h04 2.6.32.9 #2 SMP Sat Mar 6 19:09:13 BRT 2010 x86_64 GNU/Linux
> sd02: Linux sd02 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
> GNU/Linux
> sd05: Linux sd05 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
> GNU/Linux
> sd06: Linux sd06 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
> GNU/Linux
> sd07: Linux sd07 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
> GNU/Linux
>
>
> These are the java processes running on each clients.
> Jjust the h0* machines are running ok.
>
> ramiro@lcpad:~/hadoop-0.20.2$ ./bin/slaves.sh pgrep -lc java | sort
> a01: 1
> a02: 1
> a03: 1
> a04: 1
> a05: 1
> a06: 1
> a07: 1
> a09: 1
> a10: 1
> ag06: 1
> ag07: 1
> bl02: 1
> bl03: 1
> bl04: 1
> bl06: 1
> bl07: 1
> ct02: 1
> ct03: 1
> ct04: 1
> ct06: 1
> h01: 2
> h02: 2
> h03: 2
> h04: 2
> sd02: 1
> sd05: 1
> sd06: 1
> sd07: 1
>
> This is my configuration:
>
> ramiro@lcpad:~/hadoop-0.20.2$ cat conf/*site*
> <?xml version="1.0"?>
> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
>
> <!-- Put site-specific property overrides in this file. -->
>
> <configuration>
> <property>
> <name>fs.default.name</name>
> <value>hdfs://lcpad:9000</value>
> </property>
> </configuration>
> <?xml version="1.0"?>
> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
>
> <!-- Put site-specific property overrides in this file. -->
>
> <configuration>
> <property>
> <name>dfs.replication</name>
> <value>1</value>
> </property>
> </configuration>
> <?xml version="1.0"?>
> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
>
> <!-- Put site-specific property overrides in this file. -->
>
> <configuration>
>  <property>
>    <name>mapred.job.tracker</name>
>    <value>lcpad:9001</value>
>  </property>
> </configuration>
>
> Thanks in Advance,
>
> Edson Ramiro
>
>
> On 30 March 2010 05:58, Steve Loughran <st...@apache.org> wrote:
>
> > Edson Ramiro wrote:
> >
> >> I'm not involved with Debian community :(
> >>
> >
> > I think you are now...
> >
>



-- 
Todd Lipcon
Software Engineer, Cloudera

Re: java.io.IOException: Function not implemented

Posted by Edson Ramiro <er...@gmail.com>.
Hi all,

Thanks for help Todd and Steve,

I configured Hadoop (0.20.2) again and I'm getting the same error (Function
not implemented).

Do you think it's a Hadoop bug?

This is the situation:

I've 28 nodes where just four are running the datanode.

In all other nodes the tasktracker in running ok.

The NN and JT are running ok.

The configuration of the machines is the same, its a nfs shared home.

In all machines the Java version is "1.6.0_17".

This is the kernel version of the nodes, note that are two versions and in
both the
datanode doesn't work. Just in the h0* machines.

ramiro@lcpad:~/hadoop-0.20.2$ ./bin/slaves.sh uname -a  | sort
a01: Linux a01 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64 GNU/Linux
a02: Linux a02 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64 GNU/Linux
a03: Linux a03 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64 GNU/Linux
a04: Linux a04 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64 GNU/Linux
a05: Linux a05 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64 GNU/Linux
a06: Linux a06 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64 GNU/Linux
a07: Linux a07 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64 GNU/Linux
a09: Linux a09 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64 GNU/Linux
a10: Linux a10 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64 GNU/Linux
ag06: Linux ag06 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
GNU/Linux
ag07: Linux ag07 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
GNU/Linux
bl02: Linux bl02 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
GNU/Linux
bl03: Linux bl03 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
GNU/Linux
bl04: Linux bl04 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
GNU/Linux
bl06: Linux bl06 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
GNU/Linux
bl07: Linux bl07 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
GNU/Linux
ct02: Linux ct02 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
GNU/Linux
ct03: Linux ct03 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
GNU/Linux
ct04: Linux ct04 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
GNU/Linux
ct06: Linux ct06 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
GNU/Linux
h01: Linux h01 2.6.32.9 #2 SMP Sat Mar 6 19:09:13 BRT 2010 x86_64 GNU/Linux
h02: Linux h02 2.6.32.9 #2 SMP Sat Mar 6 19:09:13 BRT 2010 x86_64 GNU/Linux
h03: Linux h03 2.6.32.9 #2 SMP Sat Mar 6 19:09:13 BRT 2010 x86_64 GNU/Linux
h04: Linux h04 2.6.32.9 #2 SMP Sat Mar 6 19:09:13 BRT 2010 x86_64 GNU/Linux
sd02: Linux sd02 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
GNU/Linux
sd05: Linux sd05 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
GNU/Linux
sd06: Linux sd06 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
GNU/Linux
sd07: Linux sd07 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
GNU/Linux


These are the java processes running on each clients.
Jjust the h0* machines are running ok.

ramiro@lcpad:~/hadoop-0.20.2$ ./bin/slaves.sh pgrep -lc java | sort
a01: 1
a02: 1
a03: 1
a04: 1
a05: 1
a06: 1
a07: 1
a09: 1
a10: 1
ag06: 1
ag07: 1
bl02: 1
bl03: 1
bl04: 1
bl06: 1
bl07: 1
ct02: 1
ct03: 1
ct04: 1
ct06: 1
h01: 2
h02: 2
h03: 2
h04: 2
sd02: 1
sd05: 1
sd06: 1
sd07: 1

This is my configuration:

ramiro@lcpad:~/hadoop-0.20.2$ cat conf/*site*
<?xml version="1.0"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>

<!-- Put site-specific property overrides in this file. -->

<configuration>
<property>
<name>fs.default.name</name>
<value>hdfs://lcpad:9000</value>
</property>
</configuration>
<?xml version="1.0"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>

<!-- Put site-specific property overrides in this file. -->

<configuration>
<property>
<name>dfs.replication</name>
<value>1</value>
</property>
</configuration>
<?xml version="1.0"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>

<!-- Put site-specific property overrides in this file. -->

<configuration>
  <property>
    <name>mapred.job.tracker</name>
    <value>lcpad:9001</value>
  </property>
</configuration>

Thanks in Advance,

Edson Ramiro


On 30 March 2010 05:58, Steve Loughran <st...@apache.org> wrote:

> Edson Ramiro wrote:
>
>> I'm not involved with Debian community :(
>>
>
> I think you are now...
>

Re: java.io.IOException: Function not implemented

Posted by Steve Loughran <st...@apache.org>.
Edson Ramiro wrote:
> I'm not involved with Debian community :(

I think you are now...

Re: java.io.IOException: Function not implemented

Posted by Edson Ramiro <er...@gmail.com>.
I'm not involved with Debian community :(

ramiro@h02:~/hadoop$ cat /proc/sys/fs/epoll/max_user_watches
3373957

and the Java is not the OpenSDK.
The version is:

ramiro@lcpad:/usr/lib/jvm/java-6-sun$ java -version
java version "1.6.0_17"
Java(TM) SE Runtime Environment (build 1.6.0_17-b04)
Java HotSpot(TM) 64-Bit Server VM (build 14.3-b01, mixed mode)

Edson Ramiro


On 29 March 2010 17:14, Todd Lipcon <to...@cloudera.com> wrote:

> Hi Edson,
>
> It looks like for some reason your kernel does not have epoll enabled. It's
> very strange, since your kernel is very recent (in fact, bleeding edge!)
>
> Can you check the contents of /proc/sys/fs/epoll/max_user_watches
>
> Are you involved with the Debian community? This sounds like a general Java
> bug. Can you also please verify that you're using the Sun JVM and not
> OpenJDK (the debian folks like OpenJDK but it has subtle issues with
> Hadoop)
> You'll have to add a non-free repository and install sun-java6-jdk
>
> -Todd
>
> On Mon, Mar 29, 2010 at 1:05 PM, Edson Ramiro <er...@gmail.com> wrote:
>
> > I'm using
> >
> > Linux h02 2.6.32.9 #2 SMP Sat Mar 6 19:09:13 BRT 2010 x86_64 GNU/Linux
> >
> > ramiro@h02:~/hadoop$ cat /etc/debian_version
> > squeeze/sid
> >
> > Thanks for reply
> >
> > Edson Ramiro
> >
> >
> > On 29 March 2010 16:56, Todd Lipcon <to...@cloudera.com> wrote:
> >
> > > Hi Edson,
> > >
> > > What operating system are you on? What kernel version?
> > >
> > > Thanks
> > > -Todd
> > >
> > > On Mon, Mar 29, 2010 at 12:01 PM, Edson Ramiro <er...@gmail.com>
> > wrote:
> > >
> > > > Hi all,
> > > >
> > > > I'm trying to install Hadoop on a cluster, but I'm getting this
> error.
> > > >
> > > > I'm using java version "1.6.0_17" and hadoop-0.20.1+169.56.tar.gz
> from
> > > > Cloudera.
> > > >
> > > > Its running in a NFS home shared between the nodes and masters.
> > > >
> > > > The NameNode works well, but all nodes try to connect and fail.
> > > >
> > > > Any Idea ?
> > > >
> > > > Thanks in Advance.
> > > >
> > > > ==> logs/hadoop-ramiro-datanode-a05.log <==
> > > > 2010-03-29 15:56:00,168 INFO org.apache.hadoop.ipc.Client: Retrying
> > > connect
> > > > to server: lcpad/192.168.1.51:9000. Already tried 0 time(s).
> > > > 2010-03-29 15:56:01,172 INFO org.apache.hadoop.ipc.Client: Retrying
> > > connect
> > > > to server: lcpad/192.168.1.51:9000. Already tried 1 time(s).
> > > > 2010-03-29 15:56:02,176 INFO org.apache.hadoop.ipc.Client: Retrying
> > > connect
> > > > to server: lcpad/192.168.1.51:9000. Already tried 2 time(s).
> > > > 2010-03-29 15:56:03,180 INFO org.apache.hadoop.ipc.Client: Retrying
> > > connect
> > > > to server: lcpad/192.168.1.51:9000. Already tried 3 time(s).
> > > > 2010-03-29 15:56:04,184 INFO org.apache.hadoop.ipc.Client: Retrying
> > > connect
> > > > to server: lcpad/192.168.1.51:9000. Already tried 4 time(s).
> > > > 2010-03-29 15:56:05,188 INFO org.apache.hadoop.ipc.Client: Retrying
> > > connect
> > > > to server: lcpad/192.168.1.51:9000. Already tried 5 time(s).
> > > > 2010-03-29 15:56:06,192 INFO org.apache.hadoop.ipc.Client: Retrying
> > > connect
> > > > to server: lcpad/192.168.1.51:9000. Already tried 6 time(s).
> > > > 2010-03-29 15:56:07,196 INFO org.apache.hadoop.ipc.Client: Retrying
> > > connect
> > > > to server: lcpad/192.168.1.51:9000. Already tried 7 time(s).
> > > > 2010-03-29 15:56:08,200 INFO org.apache.hadoop.ipc.Client: Retrying
> > > connect
> > > > to server: lcpad/192.168.1.51:9000. Already tried 8 time(s).
> > > > 2010-03-29 15:56:09,204 INFO org.apache.hadoop.ipc.Client: Retrying
> > > connect
> > > > to server: lcpad/192.168.1.51:9000. Already tried 9 time(s).
> > > > 2010-03-29 15:56:09,204 ERROR
> > > > org.apache.hadoop.hdfs.server.datanode.DataNode: java.io.IOException:
> > > Call
> > > > to lcpad/192.168.1.51:9000 failed on local exception:
> > > java.io.IOException:
> > > > Function not implemented
> > > >        at org.apache.hadoop.ipc.Client.wrapException(Client.java:775)
> > > >        at org.apache.hadoop.ipc.Client.call(Client.java:743)
> > > >        at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220)
> > > >        at $Proxy4.getProtocolVersion(Unknown Source)
> > > >        at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:359)
> > > >        at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:346)
> > > >        at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:383)
> > > >        at org.apache.hadoop.ipc.RPC.waitForProxy(RPC.java:314)
> > > >        at org.apache.hadoop.ipc.RPC.waitForProxy(RPC.java:291)
> > > >        at
> > > >
> > > >
> > >
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:278)
> > > >        at
> > > >
> > org.apache.hadoop.hdfs.server.datanode.DataNode.<init>(DataNode.java:225)
> > > >        at
> > > >
> > > >
> > >
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1309)
> > > >        at
> > > >
> > > >
> > >
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1264)
> > > >        at
> > > >
> > > >
> > >
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1272)
> > > >        at
> > > >
> > org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:1394)
> > > > Caused by: java.io.IOException: Function not implemented
> > > >        at sun.nio.ch.EPollArrayWrapper.epollCreate(Native Method)
> > > >        at
> > sun.nio.ch.EPollArrayWrapper.<init>(EPollArrayWrapper.java:68)
> > > >        at
> > sun.nio.ch.EPollSelectorImpl.<init>(EPollSelectorImpl.java:52)
> > > >        at
> > > >
> > > >
> > >
> >
> sun.nio.ch.EPollSelectorProvider.openSelector(EPollSelectorProvider.java:18)
> > > >        at
> > > >
> > > >
> > >
> >
> org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.get(SocketIOWithTimeout.java:407)
> > > >        at
> > > >
> > > >
> > >
> >
> org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:322)
> > > >        at
> > > >
> > > >
> > >
> >
> org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:203)
> > > >        at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:407)
> > > >        at
> > > >
> org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:304)
> > > >        at
> > > > org.apache.hadoop.ipc.Client$Connection.access$1700(Client.java:176)
> > > >        at org.apache.hadoop.ipc.Client.getConnection(Client.java:860)
> > > >        at org.apache.hadoop.ipc.Client.call(Client.java:720)
> > > >        ... 13 more
> > > >
> > > > Edson Ramiro
> > > >
> > >
> > >
> > >
> > > --
> > > Todd Lipcon
> > > Software Engineer, Cloudera
> > >
> >
>
>
>
> --
> Todd Lipcon
> Software Engineer, Cloudera
>

Re: java.io.IOException: Function not implemented

Posted by Todd Lipcon <to...@cloudera.com>.
Hi Edson,

It looks like for some reason your kernel does not have epoll enabled. It's
very strange, since your kernel is very recent (in fact, bleeding edge!)

Can you check the contents of /proc/sys/fs/epoll/max_user_watches

Are you involved with the Debian community? This sounds like a general Java
bug. Can you also please verify that you're using the Sun JVM and not
OpenJDK (the debian folks like OpenJDK but it has subtle issues with Hadoop)
You'll have to add a non-free repository and install sun-java6-jdk

-Todd

On Mon, Mar 29, 2010 at 1:05 PM, Edson Ramiro <er...@gmail.com> wrote:

> I'm using
>
> Linux h02 2.6.32.9 #2 SMP Sat Mar 6 19:09:13 BRT 2010 x86_64 GNU/Linux
>
> ramiro@h02:~/hadoop$ cat /etc/debian_version
> squeeze/sid
>
> Thanks for reply
>
> Edson Ramiro
>
>
> On 29 March 2010 16:56, Todd Lipcon <to...@cloudera.com> wrote:
>
> > Hi Edson,
> >
> > What operating system are you on? What kernel version?
> >
> > Thanks
> > -Todd
> >
> > On Mon, Mar 29, 2010 at 12:01 PM, Edson Ramiro <er...@gmail.com>
> wrote:
> >
> > > Hi all,
> > >
> > > I'm trying to install Hadoop on a cluster, but I'm getting this error.
> > >
> > > I'm using java version "1.6.0_17" and hadoop-0.20.1+169.56.tar.gz from
> > > Cloudera.
> > >
> > > Its running in a NFS home shared between the nodes and masters.
> > >
> > > The NameNode works well, but all nodes try to connect and fail.
> > >
> > > Any Idea ?
> > >
> > > Thanks in Advance.
> > >
> > > ==> logs/hadoop-ramiro-datanode-a05.log <==
> > > 2010-03-29 15:56:00,168 INFO org.apache.hadoop.ipc.Client: Retrying
> > connect
> > > to server: lcpad/192.168.1.51:9000. Already tried 0 time(s).
> > > 2010-03-29 15:56:01,172 INFO org.apache.hadoop.ipc.Client: Retrying
> > connect
> > > to server: lcpad/192.168.1.51:9000. Already tried 1 time(s).
> > > 2010-03-29 15:56:02,176 INFO org.apache.hadoop.ipc.Client: Retrying
> > connect
> > > to server: lcpad/192.168.1.51:9000. Already tried 2 time(s).
> > > 2010-03-29 15:56:03,180 INFO org.apache.hadoop.ipc.Client: Retrying
> > connect
> > > to server: lcpad/192.168.1.51:9000. Already tried 3 time(s).
> > > 2010-03-29 15:56:04,184 INFO org.apache.hadoop.ipc.Client: Retrying
> > connect
> > > to server: lcpad/192.168.1.51:9000. Already tried 4 time(s).
> > > 2010-03-29 15:56:05,188 INFO org.apache.hadoop.ipc.Client: Retrying
> > connect
> > > to server: lcpad/192.168.1.51:9000. Already tried 5 time(s).
> > > 2010-03-29 15:56:06,192 INFO org.apache.hadoop.ipc.Client: Retrying
> > connect
> > > to server: lcpad/192.168.1.51:9000. Already tried 6 time(s).
> > > 2010-03-29 15:56:07,196 INFO org.apache.hadoop.ipc.Client: Retrying
> > connect
> > > to server: lcpad/192.168.1.51:9000. Already tried 7 time(s).
> > > 2010-03-29 15:56:08,200 INFO org.apache.hadoop.ipc.Client: Retrying
> > connect
> > > to server: lcpad/192.168.1.51:9000. Already tried 8 time(s).
> > > 2010-03-29 15:56:09,204 INFO org.apache.hadoop.ipc.Client: Retrying
> > connect
> > > to server: lcpad/192.168.1.51:9000. Already tried 9 time(s).
> > > 2010-03-29 15:56:09,204 ERROR
> > > org.apache.hadoop.hdfs.server.datanode.DataNode: java.io.IOException:
> > Call
> > > to lcpad/192.168.1.51:9000 failed on local exception:
> > java.io.IOException:
> > > Function not implemented
> > >        at org.apache.hadoop.ipc.Client.wrapException(Client.java:775)
> > >        at org.apache.hadoop.ipc.Client.call(Client.java:743)
> > >        at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220)
> > >        at $Proxy4.getProtocolVersion(Unknown Source)
> > >        at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:359)
> > >        at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:346)
> > >        at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:383)
> > >        at org.apache.hadoop.ipc.RPC.waitForProxy(RPC.java:314)
> > >        at org.apache.hadoop.ipc.RPC.waitForProxy(RPC.java:291)
> > >        at
> > >
> > >
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:278)
> > >        at
> > >
> org.apache.hadoop.hdfs.server.datanode.DataNode.<init>(DataNode.java:225)
> > >        at
> > >
> > >
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1309)
> > >        at
> > >
> > >
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1264)
> > >        at
> > >
> > >
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1272)
> > >        at
> > >
> org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:1394)
> > > Caused by: java.io.IOException: Function not implemented
> > >        at sun.nio.ch.EPollArrayWrapper.epollCreate(Native Method)
> > >        at
> sun.nio.ch.EPollArrayWrapper.<init>(EPollArrayWrapper.java:68)
> > >        at
> sun.nio.ch.EPollSelectorImpl.<init>(EPollSelectorImpl.java:52)
> > >        at
> > >
> > >
> >
> sun.nio.ch.EPollSelectorProvider.openSelector(EPollSelectorProvider.java:18)
> > >        at
> > >
> > >
> >
> org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.get(SocketIOWithTimeout.java:407)
> > >        at
> > >
> > >
> >
> org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:322)
> > >        at
> > >
> > >
> >
> org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:203)
> > >        at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:407)
> > >        at
> > > org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:304)
> > >        at
> > > org.apache.hadoop.ipc.Client$Connection.access$1700(Client.java:176)
> > >        at org.apache.hadoop.ipc.Client.getConnection(Client.java:860)
> > >        at org.apache.hadoop.ipc.Client.call(Client.java:720)
> > >        ... 13 more
> > >
> > > Edson Ramiro
> > >
> >
> >
> >
> > --
> > Todd Lipcon
> > Software Engineer, Cloudera
> >
>



-- 
Todd Lipcon
Software Engineer, Cloudera

Re: java.io.IOException: Function not implemented

Posted by Edson Ramiro <er...@gmail.com>.
I'm using

Linux h02 2.6.32.9 #2 SMP Sat Mar 6 19:09:13 BRT 2010 x86_64 GNU/Linux

ramiro@h02:~/hadoop$ cat /etc/debian_version
squeeze/sid

Thanks for reply

Edson Ramiro


On 29 March 2010 16:56, Todd Lipcon <to...@cloudera.com> wrote:

> Hi Edson,
>
> What operating system are you on? What kernel version?
>
> Thanks
> -Todd
>
> On Mon, Mar 29, 2010 at 12:01 PM, Edson Ramiro <er...@gmail.com> wrote:
>
> > Hi all,
> >
> > I'm trying to install Hadoop on a cluster, but I'm getting this error.
> >
> > I'm using java version "1.6.0_17" and hadoop-0.20.1+169.56.tar.gz from
> > Cloudera.
> >
> > Its running in a NFS home shared between the nodes and masters.
> >
> > The NameNode works well, but all nodes try to connect and fail.
> >
> > Any Idea ?
> >
> > Thanks in Advance.
> >
> > ==> logs/hadoop-ramiro-datanode-a05.log <==
> > 2010-03-29 15:56:00,168 INFO org.apache.hadoop.ipc.Client: Retrying
> connect
> > to server: lcpad/192.168.1.51:9000. Already tried 0 time(s).
> > 2010-03-29 15:56:01,172 INFO org.apache.hadoop.ipc.Client: Retrying
> connect
> > to server: lcpad/192.168.1.51:9000. Already tried 1 time(s).
> > 2010-03-29 15:56:02,176 INFO org.apache.hadoop.ipc.Client: Retrying
> connect
> > to server: lcpad/192.168.1.51:9000. Already tried 2 time(s).
> > 2010-03-29 15:56:03,180 INFO org.apache.hadoop.ipc.Client: Retrying
> connect
> > to server: lcpad/192.168.1.51:9000. Already tried 3 time(s).
> > 2010-03-29 15:56:04,184 INFO org.apache.hadoop.ipc.Client: Retrying
> connect
> > to server: lcpad/192.168.1.51:9000. Already tried 4 time(s).
> > 2010-03-29 15:56:05,188 INFO org.apache.hadoop.ipc.Client: Retrying
> connect
> > to server: lcpad/192.168.1.51:9000. Already tried 5 time(s).
> > 2010-03-29 15:56:06,192 INFO org.apache.hadoop.ipc.Client: Retrying
> connect
> > to server: lcpad/192.168.1.51:9000. Already tried 6 time(s).
> > 2010-03-29 15:56:07,196 INFO org.apache.hadoop.ipc.Client: Retrying
> connect
> > to server: lcpad/192.168.1.51:9000. Already tried 7 time(s).
> > 2010-03-29 15:56:08,200 INFO org.apache.hadoop.ipc.Client: Retrying
> connect
> > to server: lcpad/192.168.1.51:9000. Already tried 8 time(s).
> > 2010-03-29 15:56:09,204 INFO org.apache.hadoop.ipc.Client: Retrying
> connect
> > to server: lcpad/192.168.1.51:9000. Already tried 9 time(s).
> > 2010-03-29 15:56:09,204 ERROR
> > org.apache.hadoop.hdfs.server.datanode.DataNode: java.io.IOException:
> Call
> > to lcpad/192.168.1.51:9000 failed on local exception:
> java.io.IOException:
> > Function not implemented
> >        at org.apache.hadoop.ipc.Client.wrapException(Client.java:775)
> >        at org.apache.hadoop.ipc.Client.call(Client.java:743)
> >        at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220)
> >        at $Proxy4.getProtocolVersion(Unknown Source)
> >        at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:359)
> >        at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:346)
> >        at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:383)
> >        at org.apache.hadoop.ipc.RPC.waitForProxy(RPC.java:314)
> >        at org.apache.hadoop.ipc.RPC.waitForProxy(RPC.java:291)
> >        at
> >
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:278)
> >        at
> > org.apache.hadoop.hdfs.server.datanode.DataNode.<init>(DataNode.java:225)
> >        at
> >
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1309)
> >        at
> >
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1264)
> >        at
> >
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1272)
> >        at
> > org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:1394)
> > Caused by: java.io.IOException: Function not implemented
> >        at sun.nio.ch.EPollArrayWrapper.epollCreate(Native Method)
> >        at sun.nio.ch.EPollArrayWrapper.<init>(EPollArrayWrapper.java:68)
> >        at sun.nio.ch.EPollSelectorImpl.<init>(EPollSelectorImpl.java:52)
> >        at
> >
> >
> sun.nio.ch.EPollSelectorProvider.openSelector(EPollSelectorProvider.java:18)
> >        at
> >
> >
> org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.get(SocketIOWithTimeout.java:407)
> >        at
> >
> >
> org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:322)
> >        at
> >
> >
> org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:203)
> >        at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:407)
> >        at
> > org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:304)
> >        at
> > org.apache.hadoop.ipc.Client$Connection.access$1700(Client.java:176)
> >        at org.apache.hadoop.ipc.Client.getConnection(Client.java:860)
> >        at org.apache.hadoop.ipc.Client.call(Client.java:720)
> >        ... 13 more
> >
> > Edson Ramiro
> >
>
>
>
> --
> Todd Lipcon
> Software Engineer, Cloudera
>

Re: java.io.IOException: Function not implemented

Posted by Todd Lipcon <to...@cloudera.com>.
Hi Edson,

What operating system are you on? What kernel version?

Thanks
-Todd

On Mon, Mar 29, 2010 at 12:01 PM, Edson Ramiro <er...@gmail.com> wrote:

> Hi all,
>
> I'm trying to install Hadoop on a cluster, but I'm getting this error.
>
> I'm using java version "1.6.0_17" and hadoop-0.20.1+169.56.tar.gz from
> Cloudera.
>
> Its running in a NFS home shared between the nodes and masters.
>
> The NameNode works well, but all nodes try to connect and fail.
>
> Any Idea ?
>
> Thanks in Advance.
>
> ==> logs/hadoop-ramiro-datanode-a05.log <==
> 2010-03-29 15:56:00,168 INFO org.apache.hadoop.ipc.Client: Retrying connect
> to server: lcpad/192.168.1.51:9000. Already tried 0 time(s).
> 2010-03-29 15:56:01,172 INFO org.apache.hadoop.ipc.Client: Retrying connect
> to server: lcpad/192.168.1.51:9000. Already tried 1 time(s).
> 2010-03-29 15:56:02,176 INFO org.apache.hadoop.ipc.Client: Retrying connect
> to server: lcpad/192.168.1.51:9000. Already tried 2 time(s).
> 2010-03-29 15:56:03,180 INFO org.apache.hadoop.ipc.Client: Retrying connect
> to server: lcpad/192.168.1.51:9000. Already tried 3 time(s).
> 2010-03-29 15:56:04,184 INFO org.apache.hadoop.ipc.Client: Retrying connect
> to server: lcpad/192.168.1.51:9000. Already tried 4 time(s).
> 2010-03-29 15:56:05,188 INFO org.apache.hadoop.ipc.Client: Retrying connect
> to server: lcpad/192.168.1.51:9000. Already tried 5 time(s).
> 2010-03-29 15:56:06,192 INFO org.apache.hadoop.ipc.Client: Retrying connect
> to server: lcpad/192.168.1.51:9000. Already tried 6 time(s).
> 2010-03-29 15:56:07,196 INFO org.apache.hadoop.ipc.Client: Retrying connect
> to server: lcpad/192.168.1.51:9000. Already tried 7 time(s).
> 2010-03-29 15:56:08,200 INFO org.apache.hadoop.ipc.Client: Retrying connect
> to server: lcpad/192.168.1.51:9000. Already tried 8 time(s).
> 2010-03-29 15:56:09,204 INFO org.apache.hadoop.ipc.Client: Retrying connect
> to server: lcpad/192.168.1.51:9000. Already tried 9 time(s).
> 2010-03-29 15:56:09,204 ERROR
> org.apache.hadoop.hdfs.server.datanode.DataNode: java.io.IOException: Call
> to lcpad/192.168.1.51:9000 failed on local exception: java.io.IOException:
> Function not implemented
>        at org.apache.hadoop.ipc.Client.wrapException(Client.java:775)
>        at org.apache.hadoop.ipc.Client.call(Client.java:743)
>        at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220)
>        at $Proxy4.getProtocolVersion(Unknown Source)
>        at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:359)
>        at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:346)
>        at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:383)
>        at org.apache.hadoop.ipc.RPC.waitForProxy(RPC.java:314)
>        at org.apache.hadoop.ipc.RPC.waitForProxy(RPC.java:291)
>        at
>
> org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:278)
>        at
> org.apache.hadoop.hdfs.server.datanode.DataNode.<init>(DataNode.java:225)
>        at
>
> org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1309)
>        at
>
> org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1264)
>        at
>
> org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1272)
>        at
> org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:1394)
> Caused by: java.io.IOException: Function not implemented
>        at sun.nio.ch.EPollArrayWrapper.epollCreate(Native Method)
>        at sun.nio.ch.EPollArrayWrapper.<init>(EPollArrayWrapper.java:68)
>        at sun.nio.ch.EPollSelectorImpl.<init>(EPollSelectorImpl.java:52)
>        at
>
> sun.nio.ch.EPollSelectorProvider.openSelector(EPollSelectorProvider.java:18)
>        at
>
> org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.get(SocketIOWithTimeout.java:407)
>        at
>
> org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:322)
>        at
>
> org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:203)
>        at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:407)
>        at
> org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:304)
>        at
> org.apache.hadoop.ipc.Client$Connection.access$1700(Client.java:176)
>        at org.apache.hadoop.ipc.Client.getConnection(Client.java:860)
>        at org.apache.hadoop.ipc.Client.call(Client.java:720)
>        ... 13 more
>
> Edson Ramiro
>



-- 
Todd Lipcon
Software Engineer, Cloudera