You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@nutch.apache.org by Matt Zytaruk <ma...@wavefire.com> on 2005/12/13 19:46:33 UTC
Map Reduce Errors
Hello all, I've been trying to parse a segment of data (probably around
500k pages) I previously fetched, and everytime I try, I get an error.
Below is the error given by the slaves. The master gives a similar
error. This usually happens late in the reduce phase, but has also
happened during the map phase once. Any ideas what might be going on
here? Network issues? bugs in the tracker?
Thanks for any help you might be able to give.
-matt zytaruk
Slaves:
060102 200647 task_m_bvkze5 Child Error
java.io.IOException: Task process exit with nonzero status.
at org.apache.nutch.mapred.TaskRunner.runChild(TaskRunner.java:139)
at org.apache.nutch.mapred.TaskRunner.run(TaskRunner.java:92)
060102 200833 task_m_bvkze5 done; removing files.
060102 200855 Client connection to 64.141.15.126:8050: closing
java.lang.reflect.UndeclaredThrowableException
at $Proxy0.pollForClosedTask(Unknown Source)
at
org.apache.nutch.mapred.TaskTracker.offerService(TaskTracker.java:241)
at org.apache.nutch.mapred.TaskTracker.run(TaskTracker.java:268)
at org.apache.nutch.mapred.TaskTracker.main(TaskTracker.java:633)
Caused by: java.io.IOException: timed out waiting for response
at org.apache.nutch.ipc.Client.call(Client.java:296)
at org.apache.nutch.ipc.RPC$Invoker.invoke(RPC.java:127)
... 4 more
060102 201229 Lost connection to JobTracker
[crawler-d-01.internal.wavefire.ca/64.141.15.126:8050]. Retrying...
Master:
Exception in thread "main" java.lang.reflect.UndeclaredThrowableException
at $Proxy0.getJobStatus(Unknown Source)
at org.apache.nutch.mapred.JobClient.getJob(JobClient.java:272)
at org.apache.nutch.mapred.JobClient.runJob(JobClient.java:295)
at org.apache.nutch.crawl.ParseSegment.parse(ParseSegment.java:91)
at org.apache.nutch.crawl.ParseSegment.main(ParseSegment.java:110)
Caused by: java.io.IOException: timed out waiting for response
at org.apache.nutch.ipc.Client.call(Client.java:296)
at org.apache.nutch.ipc.RPC$Invoker.invoke(RPC.java:127)
Re: Map Reduce Errors
Posted by Stefan Groschupf <sg...@media-style.com>.
Matt,
I can reproduce the error at least on my servers.
Timed out.java.io.IOException: Task process exit with nonzero status.
at org.apache.nutch.mapred.TaskRunner.runChild(TaskRunner.java:139)
at org.apache.nutch.mapred.TaskRunner.run(TaskRunner.java:92)
I will try to reproduce the problem in a debugging environment today.
Stefan
Am 14.12.2005 um 00:24 schrieb Matt Zytaruk:
> Well, I tried it with the two machines that should be working fine,
> and it still crapped out, though this time with a different exception.
>
> 060104 001346 task_m_y6k6jq Child Error
> java.lang.NullPointerException
> at java.io.BufferedInputStream.read(BufferedInputStream.java:
> 279)
> at sun.nio.cs.StreamDecoder$CharsetSD.readBytes
> (StreamDecoder.java:408)
> at sun.nio.cs.StreamDecoder$CharsetSD.implRead
> (StreamDecoder.java:450)
> at sun.nio.cs.StreamDecoder.read(StreamDecoder.java:182)
> at java.io.InputStreamReader.read(InputStreamReader.java:167)
> at java.io.BufferedReader.fill(BufferedReader.java:136)
> at java.io.BufferedReader.readLine(BufferedReader.java:299)
> at java.io.BufferedReader.readLine(BufferedReader.java:362)
> at org.apache.nutch.mapred.TaskRunner.logStream
> (TaskRunner.java:164)
> at org.apache.nutch.mapred.TaskRunner.runChild
> (TaskRunner.java:136)
> at org.apache.nutch.mapred.TaskRunner.run(TaskRunner.java:92)
> 060104 001407 Server connection on port 49627 from 127.0.0.2: exiting
>
> -Matt Zytaruk
>
>
> Stefan Groschupf wrote:
>
>>> It do see stuff about crc's being ignored sometimes at the end
>>> of an operation. Is there a setting for this?
>>>
>>
>> Isn't it in the nutch-default.xml?
>>
>>> I have also just learned that the box i've been using as the job
>>> tracker and ndfs name node has a wonky system timer. So maybe
>>> that is the problem. I'm currently getting a test running using
>>> only the other two machines.
>>
>> Well this could may be the problem, I'm not sure how java handles
>> the system current millis behind the sense. Since you got a time
>> out exception may this could be a problem.
>> Stefan
>>
>>>
>>> -Matt
>>>
>>>
>>>
>>> Stefan Groschupf wrote:
>>>
>>>> hmm, sounds strange, but I'm interested to dig to find the
>>>> problem source, since I'm very much interested to get 0.8
>>>> stabile asap.
>>>> However find such a problem source is a pain in the neck.
>>>> Do you use the latest sources from svn?
>>>> Do you ignore crc errors? Doug mentioned that he notice often
>>>> problems with this.
>>>>
>>>>
>>>> Stefan
>>>>
>>>>
>>>> Am 13.12.2005 um 20:28 schrieb Matt Zytaruk:
>>>>
>>>>> I dont think the network settings are the problem, as I have
>>>>> been able to parse other segments using map reduce no
>>>>> problem. If it was the network configuration, wouldn't it
>>>>> never work? However, things do not seem to be stable, as
>>>>> some operations in ndfs will error, and then I do the same
>>>>> thing 5 minutes later and it works fine. Same with other
>>>>> things, some crawls work fine, others throw exceptions and
>>>>> crash (I actually had a crawl crash with the same problem as
>>>>> below). This is using 3 Opteron boxes running Suse Linux.
>>>>>
>>>>> -Matt Zytaruk
>>>>>
>>>>> Stefan Groschupf wrote:
>>>>>
>>>>>> Looks like a problem with the tcp ip communication.
>>>>>> Any firewalls running on the boxes? May any ports closed?
>>>>>> Are the dns names correct configured?
>>>>>>
>>>>>> Is your job tracker running stable?
>>>>>>
>>>>>> Stefan
>>>>>>
>>>>>> Am 13.12.2005 um 19:46 schrieb Matt Zytaruk:
>>>>>>
>>>>>>> Hello all, I've been trying to parse a segment of data
>>>>>>> (probably around 500k pages) I previously fetched, and
>>>>>>> everytime I try, I get an error. Below is the error given
>>>>>>> by the slaves. The master gives a similar error. This
>>>>>>> usually happens late in the reduce phase, but has also
>>>>>>> happened during the map phase once. Any ideas what might
>>>>>>> be going on here? Network issues? bugs in the tracker?
>>>>>>>
>>>>>>> Thanks for any help you might be able to give.
>>>>>>> -matt zytaruk
>>>>>>>
>>>>>>> Slaves:
>>>>>>>
>>>>>>> 060102 200647 task_m_bvkze5 Child Error
>>>>>>> java.io.IOException: Task process exit with nonzero status.
>>>>>>> at org.apache.nutch.mapred.TaskRunner.runChild
>>>>>>> (TaskRunner.java:139)
>>>>>>> at org.apache.nutch.mapred.TaskRunner.run
>>>>>>> (TaskRunner.java:92)
>>>>>>> 060102 200833 task_m_bvkze5 done; removing files.
>>>>>>> 060102 200855 Client connection to 64.141.15.126:8050: closing
>>>>>>> java.lang.reflect.UndeclaredThrowableException
>>>>>>> at $Proxy0.pollForClosedTask(Unknown Source)
>>>>>>> at org.apache.nutch.mapred.TaskTracker.offerService
>>>>>>> (TaskTracker.java:241)
>>>>>>> at org.apache.nutch.mapred.TaskTracker.run
>>>>>>> (TaskTracker.java: 268)
>>>>>>> at org.apache.nutch.mapred.TaskTracker.main
>>>>>>> (TaskTracker.java: 633)
>>>>>>> Caused by: java.io.IOException: timed out waiting for response
>>>>>>> at org.apache.nutch.ipc.Client.call(Client.java:296)
>>>>>>> at org.apache.nutch.ipc.RPC$Invoker.invoke(RPC.java:127)
>>>>>>> ... 4 more
>>>>>>> 060102 201229 Lost connection to JobTracker [crawler-
>>>>>>> d-01.internal.wavefire.ca/64.141.15.126:8050]. Retrying...
>>>>>>>
>>>>>>> Master:
>>>>>>> Exception in thread "main"
>>>>>>> java.lang.reflect.UndeclaredThrowableException
>>>>>>> at $Proxy0.getJobStatus(Unknown Source)
>>>>>>> at org.apache.nutch.mapred.JobClient.getJob
>>>>>>> (JobClient.java: 272)
>>>>>>> at org.apache.nutch.mapred.JobClient.runJob
>>>>>>> (JobClient.java: 295)
>>>>>>> at org.apache.nutch.crawl.ParseSegment.parse
>>>>>>> (ParseSegment.java:91)
>>>>>>> at org.apache.nutch.crawl.ParseSegment.main
>>>>>>> (ParseSegment.java:110)
>>>>>>> Caused by: java.io.IOException: timed out waiting for response
>>>>>>> at org.apache.nutch.ipc.Client.call(Client.java:296)
>>>>>>> at org.apache.nutch.ipc.RPC$Invoker.invoke(RPC.java:127)
>>>>>>>
>>>>>>>
>>>>>>
>>>>>> ---------------------------------------------------------------
>>>>>> company: http://www.media-style.com
>>>>>> forum: http://www.text-mining.org
>>>>>> blog: http://www.find23.net
>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>>
>>>>
>>>>
>>>>
>>>>
>>>
>>>
>>
>> ---------------------------------------------------------------
>> company: http://www.media-style.com
>> forum: http://www.text-mining.org
>> blog: http://www.find23.net
>>
>>
>>
>
>
---------------------------------------------------------------
company: http://www.media-style.com
forum: http://www.text-mining.org
blog: http://www.find23.net
Re: Map Reduce Errors
Posted by Matt Zytaruk <ma...@wavefire.com>.
Well, I tried it with the two machines that should be working fine, and
it still crapped out, though this time with a different exception.
060104 001346 task_m_y6k6jq Child Error
java.lang.NullPointerException
at java.io.BufferedInputStream.read(BufferedInputStream.java:279)
at
sun.nio.cs.StreamDecoder$CharsetSD.readBytes(StreamDecoder.java:408)
at
sun.nio.cs.StreamDecoder$CharsetSD.implRead(StreamDecoder.java:450)
at sun.nio.cs.StreamDecoder.read(StreamDecoder.java:182)
at java.io.InputStreamReader.read(InputStreamReader.java:167)
at java.io.BufferedReader.fill(BufferedReader.java:136)
at java.io.BufferedReader.readLine(BufferedReader.java:299)
at java.io.BufferedReader.readLine(BufferedReader.java:362)
at org.apache.nutch.mapred.TaskRunner.logStream(TaskRunner.java:164)
at org.apache.nutch.mapred.TaskRunner.runChild(TaskRunner.java:136)
at org.apache.nutch.mapred.TaskRunner.run(TaskRunner.java:92)
060104 001407 Server connection on port 49627 from 127.0.0.2: exiting
-Matt Zytaruk
Stefan Groschupf wrote:
>> It do see stuff about crc's being ignored sometimes at the end of an
>> operation. Is there a setting for this?
>>
>
> Isn't it in the nutch-default.xml?
>
>> I have also just learned that the box i've been using as the job
>> tracker and ndfs name node has a wonky system timer. So maybe that
>> is the problem. I'm currently getting a test running using only the
>> other two machines.
>
> Well this could may be the problem, I'm not sure how java handles the
> system current millis behind the sense. Since you got a time out
> exception may this could be a problem.
> Stefan
>
>>
>> -Matt
>>
>>
>>
>> Stefan Groschupf wrote:
>>
>>> hmm, sounds strange, but I'm interested to dig to find the problem
>>> source, since I'm very much interested to get 0.8 stabile asap.
>>> However find such a problem source is a pain in the neck.
>>> Do you use the latest sources from svn?
>>> Do you ignore crc errors? Doug mentioned that he notice often
>>> problems with this.
>>>
>>>
>>> Stefan
>>>
>>>
>>> Am 13.12.2005 um 20:28 schrieb Matt Zytaruk:
>>>
>>>> I dont think the network settings are the problem, as I have been
>>>> able to parse other segments using map reduce no problem. If it
>>>> was the network configuration, wouldn't it never work? However,
>>>> things do not seem to be stable, as some operations in ndfs will
>>>> error, and then I do the same thing 5 minutes later and it works
>>>> fine. Same with other things, some crawls work fine, others
>>>> throw exceptions and crash (I actually had a crawl crash with the
>>>> same problem as below). This is using 3 Opteron boxes running
>>>> Suse Linux.
>>>>
>>>> -Matt Zytaruk
>>>>
>>>> Stefan Groschupf wrote:
>>>>
>>>>> Looks like a problem with the tcp ip communication.
>>>>> Any firewalls running on the boxes? May any ports closed?
>>>>> Are the dns names correct configured?
>>>>>
>>>>> Is your job tracker running stable?
>>>>>
>>>>> Stefan
>>>>>
>>>>> Am 13.12.2005 um 19:46 schrieb Matt Zytaruk:
>>>>>
>>>>>> Hello all, I've been trying to parse a segment of data
>>>>>> (probably around 500k pages) I previously fetched, and
>>>>>> everytime I try, I get an error. Below is the error given by
>>>>>> the slaves. The master gives a similar error. This usually
>>>>>> happens late in the reduce phase, but has also happened during
>>>>>> the map phase once. Any ideas what might be going on here?
>>>>>> Network issues? bugs in the tracker?
>>>>>>
>>>>>> Thanks for any help you might be able to give.
>>>>>> -matt zytaruk
>>>>>>
>>>>>> Slaves:
>>>>>>
>>>>>> 060102 200647 task_m_bvkze5 Child Error
>>>>>> java.io.IOException: Task process exit with nonzero status.
>>>>>> at org.apache.nutch.mapred.TaskRunner.runChild
>>>>>> (TaskRunner.java:139)
>>>>>> at org.apache.nutch.mapred.TaskRunner.run
>>>>>> (TaskRunner.java:92)
>>>>>> 060102 200833 task_m_bvkze5 done; removing files.
>>>>>> 060102 200855 Client connection to 64.141.15.126:8050: closing
>>>>>> java.lang.reflect.UndeclaredThrowableException
>>>>>> at $Proxy0.pollForClosedTask(Unknown Source)
>>>>>> at org.apache.nutch.mapred.TaskTracker.offerService
>>>>>> (TaskTracker.java:241)
>>>>>> at org.apache.nutch.mapred.TaskTracker.run
>>>>>> (TaskTracker.java: 268)
>>>>>> at org.apache.nutch.mapred.TaskTracker.main
>>>>>> (TaskTracker.java: 633)
>>>>>> Caused by: java.io.IOException: timed out waiting for response
>>>>>> at org.apache.nutch.ipc.Client.call(Client.java:296)
>>>>>> at org.apache.nutch.ipc.RPC$Invoker.invoke(RPC.java:127)
>>>>>> ... 4 more
>>>>>> 060102 201229 Lost connection to JobTracker [crawler-
>>>>>> d-01.internal.wavefire.ca/64.141.15.126:8050]. Retrying...
>>>>>>
>>>>>> Master:
>>>>>> Exception in thread "main"
>>>>>> java.lang.reflect.UndeclaredThrowableException
>>>>>> at $Proxy0.getJobStatus(Unknown Source)
>>>>>> at org.apache.nutch.mapred.JobClient.getJob
>>>>>> (JobClient.java: 272)
>>>>>> at org.apache.nutch.mapred.JobClient.runJob
>>>>>> (JobClient.java: 295)
>>>>>> at org.apache.nutch.crawl.ParseSegment.parse
>>>>>> (ParseSegment.java:91)
>>>>>> at org.apache.nutch.crawl.ParseSegment.main
>>>>>> (ParseSegment.java:110)
>>>>>> Caused by: java.io.IOException: timed out waiting for response
>>>>>> at org.apache.nutch.ipc.Client.call(Client.java:296)
>>>>>> at org.apache.nutch.ipc.RPC$Invoker.invoke(RPC.java:127)
>>>>>>
>>>>>>
>>>>>
>>>>> ---------------------------------------------------------------
>>>>> company: http://www.media-style.com
>>>>> forum: http://www.text-mining.org
>>>>> blog: http://www.find23.net
>>>>>
>>>>>
>>>>>
>>>>
>>>>
>>>
>>>
>>>
>>>
>>
>>
>
> ---------------------------------------------------------------
> company: http://www.media-style.com
> forum: http://www.text-mining.org
> blog: http://www.find23.net
>
>
>
Re: Map Reduce Errors
Posted by Stefan Groschupf <sg...@media-style.com>.
> It do see stuff about crc's being ignored sometimes at the end of
> an operation. Is there a setting for this?
>
Isn't it in the nutch-default.xml?
> I have also just learned that the box i've been using as the job
> tracker and ndfs name node has a wonky system timer. So maybe that
> is the problem. I'm currently getting a test running using only the
> other two machines.
Well this could may be the problem, I'm not sure how java handles the
system current millis behind the sense. Since you got a time out
exception may this could be a problem.
Stefan
>
> -Matt
>
>
>
> Stefan Groschupf wrote:
>
>> hmm, sounds strange, but I'm interested to dig to find the
>> problem source, since I'm very much interested to get 0.8 stabile
>> asap.
>> However find such a problem source is a pain in the neck.
>> Do you use the latest sources from svn?
>> Do you ignore crc errors? Doug mentioned that he notice often
>> problems with this.
>>
>>
>> Stefan
>>
>>
>> Am 13.12.2005 um 20:28 schrieb Matt Zytaruk:
>>
>>> I dont think the network settings are the problem, as I have
>>> been able to parse other segments using map reduce no problem.
>>> If it was the network configuration, wouldn't it never work?
>>> However, things do not seem to be stable, as some operations in
>>> ndfs will error, and then I do the same thing 5 minutes later
>>> and it works fine. Same with other things, some crawls work
>>> fine, others throw exceptions and crash (I actually had a crawl
>>> crash with the same problem as below). This is using 3 Opteron
>>> boxes running Suse Linux.
>>>
>>> -Matt Zytaruk
>>>
>>> Stefan Groschupf wrote:
>>>
>>>> Looks like a problem with the tcp ip communication.
>>>> Any firewalls running on the boxes? May any ports closed?
>>>> Are the dns names correct configured?
>>>>
>>>> Is your job tracker running stable?
>>>>
>>>> Stefan
>>>>
>>>> Am 13.12.2005 um 19:46 schrieb Matt Zytaruk:
>>>>
>>>>> Hello all, I've been trying to parse a segment of data
>>>>> (probably around 500k pages) I previously fetched, and
>>>>> everytime I try, I get an error. Below is the error given by
>>>>> the slaves. The master gives a similar error. This usually
>>>>> happens late in the reduce phase, but has also happened
>>>>> during the map phase once. Any ideas what might be going on
>>>>> here? Network issues? bugs in the tracker?
>>>>>
>>>>> Thanks for any help you might be able to give.
>>>>> -matt zytaruk
>>>>>
>>>>> Slaves:
>>>>>
>>>>> 060102 200647 task_m_bvkze5 Child Error
>>>>> java.io.IOException: Task process exit with nonzero status.
>>>>> at org.apache.nutch.mapred.TaskRunner.runChild
>>>>> (TaskRunner.java:139)
>>>>> at org.apache.nutch.mapred.TaskRunner.run
>>>>> (TaskRunner.java:92)
>>>>> 060102 200833 task_m_bvkze5 done; removing files.
>>>>> 060102 200855 Client connection to 64.141.15.126:8050: closing
>>>>> java.lang.reflect.UndeclaredThrowableException
>>>>> at $Proxy0.pollForClosedTask(Unknown Source)
>>>>> at org.apache.nutch.mapred.TaskTracker.offerService
>>>>> (TaskTracker.java:241)
>>>>> at org.apache.nutch.mapred.TaskTracker.run
>>>>> (TaskTracker.java: 268)
>>>>> at org.apache.nutch.mapred.TaskTracker.main
>>>>> (TaskTracker.java: 633)
>>>>> Caused by: java.io.IOException: timed out waiting for response
>>>>> at org.apache.nutch.ipc.Client.call(Client.java:296)
>>>>> at org.apache.nutch.ipc.RPC$Invoker.invoke(RPC.java:127)
>>>>> ... 4 more
>>>>> 060102 201229 Lost connection to JobTracker [crawler-
>>>>> d-01.internal.wavefire.ca/64.141.15.126:8050]. Retrying...
>>>>>
>>>>> Master:
>>>>> Exception in thread "main"
>>>>> java.lang.reflect.UndeclaredThrowableException
>>>>> at $Proxy0.getJobStatus(Unknown Source)
>>>>> at org.apache.nutch.mapred.JobClient.getJob
>>>>> (JobClient.java: 272)
>>>>> at org.apache.nutch.mapred.JobClient.runJob
>>>>> (JobClient.java: 295)
>>>>> at org.apache.nutch.crawl.ParseSegment.parse
>>>>> (ParseSegment.java:91)
>>>>> at org.apache.nutch.crawl.ParseSegment.main
>>>>> (ParseSegment.java:110)
>>>>> Caused by: java.io.IOException: timed out waiting for response
>>>>> at org.apache.nutch.ipc.Client.call(Client.java:296)
>>>>> at org.apache.nutch.ipc.RPC$Invoker.invoke(RPC.java:127)
>>>>>
>>>>>
>>>>
>>>> ---------------------------------------------------------------
>>>> company: http://www.media-style.com
>>>> forum: http://www.text-mining.org
>>>> blog: http://www.find23.net
>>>>
>>>>
>>>>
>>>
>>>
>>
>>
>>
>>
>
>
---------------------------------------------------------------
company: http://www.media-style.com
forum: http://www.text-mining.org
blog: http://www.find23.net
Re: Map Reduce Errors
Posted by Matt Zytaruk <ma...@wavefire.com>.
The source I'm using is about 2 weeks or so old.
It do see stuff about crc's being ignored sometimes at the end of an
operation. Is there a setting for this?
I have also just learned that the box i've been using as the job tracker
and ndfs name node has a wonky system timer. So maybe that is the
problem. I'm currently getting a test running using only the other two
machines.
-Matt
Stefan Groschupf wrote:
> hmm, sounds strange, but I'm interested to dig to find the problem
> source, since I'm very much interested to get 0.8 stabile asap.
> However find such a problem source is a pain in the neck.
> Do you use the latest sources from svn?
> Do you ignore crc errors? Doug mentioned that he notice often
> problems with this.
>
>
> Stefan
>
>
> Am 13.12.2005 um 20:28 schrieb Matt Zytaruk:
>
>> I dont think the network settings are the problem, as I have been
>> able to parse other segments using map reduce no problem. If it was
>> the network configuration, wouldn't it never work? However, things
>> do not seem to be stable, as some operations in ndfs will error, and
>> then I do the same thing 5 minutes later and it works fine. Same
>> with other things, some crawls work fine, others throw exceptions
>> and crash (I actually had a crawl crash with the same problem as
>> below). This is using 3 Opteron boxes running Suse Linux.
>>
>> -Matt Zytaruk
>>
>> Stefan Groschupf wrote:
>>
>>> Looks like a problem with the tcp ip communication.
>>> Any firewalls running on the boxes? May any ports closed?
>>> Are the dns names correct configured?
>>>
>>> Is your job tracker running stable?
>>>
>>> Stefan
>>>
>>> Am 13.12.2005 um 19:46 schrieb Matt Zytaruk:
>>>
>>>> Hello all, I've been trying to parse a segment of data (probably
>>>> around 500k pages) I previously fetched, and everytime I try, I
>>>> get an error. Below is the error given by the slaves. The master
>>>> gives a similar error. This usually happens late in the reduce
>>>> phase, but has also happened during the map phase once. Any ideas
>>>> what might be going on here? Network issues? bugs in the tracker?
>>>>
>>>> Thanks for any help you might be able to give.
>>>> -matt zytaruk
>>>>
>>>> Slaves:
>>>>
>>>> 060102 200647 task_m_bvkze5 Child Error
>>>> java.io.IOException: Task process exit with nonzero status.
>>>> at org.apache.nutch.mapred.TaskRunner.runChild
>>>> (TaskRunner.java:139)
>>>> at org.apache.nutch.mapred.TaskRunner.run(TaskRunner.java:92)
>>>> 060102 200833 task_m_bvkze5 done; removing files.
>>>> 060102 200855 Client connection to 64.141.15.126:8050: closing
>>>> java.lang.reflect.UndeclaredThrowableException
>>>> at $Proxy0.pollForClosedTask(Unknown Source)
>>>> at org.apache.nutch.mapred.TaskTracker.offerService
>>>> (TaskTracker.java:241)
>>>> at org.apache.nutch.mapred.TaskTracker.run
>>>> (TaskTracker.java: 268)
>>>> at org.apache.nutch.mapred.TaskTracker.main
>>>> (TaskTracker.java: 633)
>>>> Caused by: java.io.IOException: timed out waiting for response
>>>> at org.apache.nutch.ipc.Client.call(Client.java:296)
>>>> at org.apache.nutch.ipc.RPC$Invoker.invoke(RPC.java:127)
>>>> ... 4 more
>>>> 060102 201229 Lost connection to JobTracker [crawler-
>>>> d-01.internal.wavefire.ca/64.141.15.126:8050]. Retrying...
>>>>
>>>> Master:
>>>> Exception in thread "main"
>>>> java.lang.reflect.UndeclaredThrowableException
>>>> at $Proxy0.getJobStatus(Unknown Source)
>>>> at org.apache.nutch.mapred.JobClient.getJob(JobClient.java:
>>>> 272)
>>>> at org.apache.nutch.mapred.JobClient.runJob(JobClient.java:
>>>> 295)
>>>> at org.apache.nutch.crawl.ParseSegment.parse
>>>> (ParseSegment.java:91)
>>>> at org.apache.nutch.crawl.ParseSegment.main
>>>> (ParseSegment.java:110)
>>>> Caused by: java.io.IOException: timed out waiting for response
>>>> at org.apache.nutch.ipc.Client.call(Client.java:296)
>>>> at org.apache.nutch.ipc.RPC$Invoker.invoke(RPC.java:127)
>>>>
>>>>
>>>
>>> ---------------------------------------------------------------
>>> company: http://www.media-style.com
>>> forum: http://www.text-mining.org
>>> blog: http://www.find23.net
>>>
>>>
>>>
>>
>>
>
>
>
>
Re: Map Reduce Errors
Posted by Stefan Groschupf <sg...@media-style.com>.
hmm, sounds strange, but I'm interested to dig to find the problem
source, since I'm very much interested to get 0.8 stabile asap.
However find such a problem source is a pain in the neck.
Do you use the latest sources from svn?
Do you ignore crc errors? Doug mentioned that he notice often
problems with this.
Stefan
Am 13.12.2005 um 20:28 schrieb Matt Zytaruk:
> I dont think the network settings are the problem, as I have been
> able to parse other segments using map reduce no problem. If it was
> the network configuration, wouldn't it never work? However, things
> do not seem to be stable, as some operations in ndfs will error,
> and then I do the same thing 5 minutes later and it works fine.
> Same with other things, some crawls work fine, others throw
> exceptions and crash (I actually had a crawl crash with the same
> problem as below). This is using 3 Opteron boxes running Suse Linux.
>
> -Matt Zytaruk
>
> Stefan Groschupf wrote:
>
>> Looks like a problem with the tcp ip communication.
>> Any firewalls running on the boxes? May any ports closed?
>> Are the dns names correct configured?
>>
>> Is your job tracker running stable?
>>
>> Stefan
>>
>> Am 13.12.2005 um 19:46 schrieb Matt Zytaruk:
>>
>>> Hello all, I've been trying to parse a segment of data (probably
>>> around 500k pages) I previously fetched, and everytime I try, I
>>> get an error. Below is the error given by the slaves. The master
>>> gives a similar error. This usually happens late in the reduce
>>> phase, but has also happened during the map phase once. Any
>>> ideas what might be going on here? Network issues? bugs in the
>>> tracker?
>>>
>>> Thanks for any help you might be able to give.
>>> -matt zytaruk
>>>
>>> Slaves:
>>>
>>> 060102 200647 task_m_bvkze5 Child Error
>>> java.io.IOException: Task process exit with nonzero status.
>>> at org.apache.nutch.mapred.TaskRunner.runChild
>>> (TaskRunner.java:139)
>>> at org.apache.nutch.mapred.TaskRunner.run(TaskRunner.java:92)
>>> 060102 200833 task_m_bvkze5 done; removing files.
>>> 060102 200855 Client connection to 64.141.15.126:8050: closing
>>> java.lang.reflect.UndeclaredThrowableException
>>> at $Proxy0.pollForClosedTask(Unknown Source)
>>> at org.apache.nutch.mapred.TaskTracker.offerService
>>> (TaskTracker.java:241)
>>> at org.apache.nutch.mapred.TaskTracker.run
>>> (TaskTracker.java: 268)
>>> at org.apache.nutch.mapred.TaskTracker.main
>>> (TaskTracker.java: 633)
>>> Caused by: java.io.IOException: timed out waiting for response
>>> at org.apache.nutch.ipc.Client.call(Client.java:296)
>>> at org.apache.nutch.ipc.RPC$Invoker.invoke(RPC.java:127)
>>> ... 4 more
>>> 060102 201229 Lost connection to JobTracker [crawler-
>>> d-01.internal.wavefire.ca/64.141.15.126:8050]. Retrying...
>>>
>>> Master:
>>> Exception in thread "main"
>>> java.lang.reflect.UndeclaredThrowableException
>>> at $Proxy0.getJobStatus(Unknown Source)
>>> at org.apache.nutch.mapred.JobClient.getJob(JobClient.java:
>>> 272)
>>> at org.apache.nutch.mapred.JobClient.runJob(JobClient.java:
>>> 295)
>>> at org.apache.nutch.crawl.ParseSegment.parse
>>> (ParseSegment.java:91)
>>> at org.apache.nutch.crawl.ParseSegment.main
>>> (ParseSegment.java:110)
>>> Caused by: java.io.IOException: timed out waiting for response
>>> at org.apache.nutch.ipc.Client.call(Client.java:296)
>>> at org.apache.nutch.ipc.RPC$Invoker.invoke(RPC.java:127)
>>>
>>>
>>
>> ---------------------------------------------------------------
>> company: http://www.media-style.com
>> forum: http://www.text-mining.org
>> blog: http://www.find23.net
>>
>>
>>
>
>
Re: Map Reduce Errors
Posted by Matt Zytaruk <ma...@wavefire.com>.
I dont think the network settings are the problem, as I have been able
to parse other segments using map reduce no problem. If it was the
network configuration, wouldn't it never work? However, things do not
seem to be stable, as some operations in ndfs will error, and then I do
the same thing 5 minutes later and it works fine. Same with other
things, some crawls work fine, others throw exceptions and crash (I
actually had a crawl crash with the same problem as below). This is
using 3 Opteron boxes running Suse Linux.
-Matt Zytaruk
Stefan Groschupf wrote:
> Looks like a problem with the tcp ip communication.
> Any firewalls running on the boxes? May any ports closed?
> Are the dns names correct configured?
>
> Is your job tracker running stable?
>
> Stefan
>
> Am 13.12.2005 um 19:46 schrieb Matt Zytaruk:
>
>> Hello all, I've been trying to parse a segment of data (probably
>> around 500k pages) I previously fetched, and everytime I try, I get
>> an error. Below is the error given by the slaves. The master gives a
>> similar error. This usually happens late in the reduce phase, but
>> has also happened during the map phase once. Any ideas what might be
>> going on here? Network issues? bugs in the tracker?
>>
>> Thanks for any help you might be able to give.
>> -matt zytaruk
>>
>> Slaves:
>>
>> 060102 200647 task_m_bvkze5 Child Error
>> java.io.IOException: Task process exit with nonzero status.
>> at org.apache.nutch.mapred.TaskRunner.runChild
>> (TaskRunner.java:139)
>> at org.apache.nutch.mapred.TaskRunner.run(TaskRunner.java:92)
>> 060102 200833 task_m_bvkze5 done; removing files.
>> 060102 200855 Client connection to 64.141.15.126:8050: closing
>> java.lang.reflect.UndeclaredThrowableException
>> at $Proxy0.pollForClosedTask(Unknown Source)
>> at org.apache.nutch.mapred.TaskTracker.offerService
>> (TaskTracker.java:241)
>> at org.apache.nutch.mapred.TaskTracker.run(TaskTracker.java: 268)
>> at org.apache.nutch.mapred.TaskTracker.main(TaskTracker.java:
>> 633)
>> Caused by: java.io.IOException: timed out waiting for response
>> at org.apache.nutch.ipc.Client.call(Client.java:296)
>> at org.apache.nutch.ipc.RPC$Invoker.invoke(RPC.java:127)
>> ... 4 more
>> 060102 201229 Lost connection to JobTracker [crawler-
>> d-01.internal.wavefire.ca/64.141.15.126:8050]. Retrying...
>>
>> Master:
>> Exception in thread "main"
>> java.lang.reflect.UndeclaredThrowableException
>> at $Proxy0.getJobStatus(Unknown Source)
>> at org.apache.nutch.mapred.JobClient.getJob(JobClient.java:272)
>> at org.apache.nutch.mapred.JobClient.runJob(JobClient.java:295)
>> at org.apache.nutch.crawl.ParseSegment.parse
>> (ParseSegment.java:91)
>> at org.apache.nutch.crawl.ParseSegment.main
>> (ParseSegment.java:110)
>> Caused by: java.io.IOException: timed out waiting for response
>> at org.apache.nutch.ipc.Client.call(Client.java:296)
>> at org.apache.nutch.ipc.RPC$Invoker.invoke(RPC.java:127)
>>
>>
>
> ---------------------------------------------------------------
> company: http://www.media-style.com
> forum: http://www.text-mining.org
> blog: http://www.find23.net
>
>
>
Re: Map Reduce Errors
Posted by Stefan Groschupf <sg...@media-style.com>.
Looks like a problem with the tcp ip communication.
Any firewalls running on the boxes? May any ports closed?
Are the dns names correct configured?
Is your job tracker running stable?
Stefan
Am 13.12.2005 um 19:46 schrieb Matt Zytaruk:
> Hello all, I've been trying to parse a segment of data (probably
> around 500k pages) I previously fetched, and everytime I try, I get
> an error. Below is the error given by the slaves. The master gives
> a similar error. This usually happens late in the reduce phase,
> but has also happened during the map phase once. Any ideas what
> might be going on here? Network issues? bugs in the tracker?
>
> Thanks for any help you might be able to give.
> -matt zytaruk
>
> Slaves:
>
> 060102 200647 task_m_bvkze5 Child Error
> java.io.IOException: Task process exit with nonzero status.
> at org.apache.nutch.mapred.TaskRunner.runChild
> (TaskRunner.java:139)
> at org.apache.nutch.mapred.TaskRunner.run(TaskRunner.java:92)
> 060102 200833 task_m_bvkze5 done; removing files.
> 060102 200855 Client connection to 64.141.15.126:8050: closing
> java.lang.reflect.UndeclaredThrowableException
> at $Proxy0.pollForClosedTask(Unknown Source)
> at org.apache.nutch.mapred.TaskTracker.offerService
> (TaskTracker.java:241)
> at org.apache.nutch.mapred.TaskTracker.run(TaskTracker.java:
> 268)
> at org.apache.nutch.mapred.TaskTracker.main(TaskTracker.java:
> 633)
> Caused by: java.io.IOException: timed out waiting for response
> at org.apache.nutch.ipc.Client.call(Client.java:296)
> at org.apache.nutch.ipc.RPC$Invoker.invoke(RPC.java:127)
> ... 4 more
> 060102 201229 Lost connection to JobTracker [crawler-
> d-01.internal.wavefire.ca/64.141.15.126:8050]. Retrying...
>
> Master:
> Exception in thread "main"
> java.lang.reflect.UndeclaredThrowableException
> at $Proxy0.getJobStatus(Unknown Source)
> at org.apache.nutch.mapred.JobClient.getJob(JobClient.java:272)
> at org.apache.nutch.mapred.JobClient.runJob(JobClient.java:295)
> at org.apache.nutch.crawl.ParseSegment.parse
> (ParseSegment.java:91)
> at org.apache.nutch.crawl.ParseSegment.main
> (ParseSegment.java:110)
> Caused by: java.io.IOException: timed out waiting for response
> at org.apache.nutch.ipc.Client.call(Client.java:296)
> at org.apache.nutch.ipc.RPC$Invoker.invoke(RPC.java:127)
>
>
---------------------------------------------------------------
company: http://www.media-style.com
forum: http://www.text-mining.org
blog: http://www.find23.net