You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-dev@hadoop.apache.org by "P.ILAYARAJA" <il...@rediff.co.in> on 2008/11/03 15:57:09 UTC
Nutch/Hadoop: Crawl is crashing
Hi,
I started an internet crawl of 30 million pages in a single segment.
The crawl was crashing with the following exception:
java.lang.ArrayIndexOutOfBoundsException: 17
at org.apache.hadoop.mapred.MapTask$MapOutputBuffer.mergeParts(MapTask.java:540)
at org.apache.hadoop.mapred.MapTask$MapOutputBuffer.flush(MapTask.java:607)
at org.apache.hadoop.mapred.MapTask.run(MapTask.java:193)
at org.apache.hadoop.mapred.TaskTracker$Child.main(TaskTracker.java:1760)
Any idea on why is it happenning and what would be the soln.
... am using hadoop 0.15.3 and nutch 1.0 versions.
Regards,
Ilay