You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@nutch.apache.org by "Sami Siren (JIRA)" <ji...@apache.org> on 2007/01/24 19:01:49 UTC

[jira] Commented: (NUTCH-433) java.io.EOFException in newer nightlies in mergesegs or indexing from hadoop.io.DataOutputBuffer

    [ https://issues.apache.org/jira/browse/NUTCH-433?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12467101 ] 

Sami Siren commented on NUTCH-433:
----------------------------------

I am working on this and will probably submit a patch today.

> java.io.EOFException in newer nightlies in mergesegs or indexing from hadoop.io.DataOutputBuffer
> ------------------------------------------------------------------------------------------------
>
>                 Key: NUTCH-433
>                 URL: https://issues.apache.org/jira/browse/NUTCH-433
>             Project: Nutch
>          Issue Type: Bug
>          Components: generator, indexer
>    Affects Versions: 0.9.0
>         Environment: Both Linux/i686 and Mac OS X PPC/Intel, but platform independent
>            Reporter: Brian Whitman
>            Priority: Critical
>
> The nightly builds have not been working at all for the past couple of weeks. Sami Siren has narrowed it down to HADOOP-331.
> To replicate: download the nightly, then:
> bin/nutch inject crawl/crawldb urls/  # a single URL is in urls/urls -- http://apache.org
> bin/nutch generate crawl/crawldb crawl/segments
> bin/nutch fetch crawl/segments/2007...
> bin/nutch updatedb crawl/crawldb crawl/segments/2007...
> # generate a new segment with 5 URIs
> bin/nutch generate crawl/crawldb crawl/segments -topN 5
> bin/nutch fetch crawl/segments/2007... # new segment
> bin/nutch updatedb crawl/crawldb crawl/segments/2007... # new segment
> # merge the segments and index
> bin/nutch mergesegs crawl/merged -dir crawl/segments
> ..
> We get a crash in the mergesegs. This crash, with the exact same script and start URI, configuration and plugins, does not happen on a nightly from early January.
> 2007-01-18 14:57:11,411 INFO  segment.SegmentMerger - Merging 2 segments to crawl/merged_07_01_18_14_56_22/20070118145711
> 2007-01-18 14:57:11,482 INFO  segment.SegmentMerger - SegmentMerger:   adding crawl/segments/20070118145628
> 2007-01-18 14:57:11,489 INFO  segment.SegmentMerger - SegmentMerger:   adding crawl/segments/20070118145641
> 2007-01-18 14:57:11,495 INFO  segment.SegmentMerger - SegmentMerger: using segment data from: content crawl_generate crawl_fetch crawl_parse parse_data parse_text
> 2007-01-18 14:57:11,594 INFO  mapred.InputFormatBase - Total input paths to process : 12
> 2007-01-18 14:57:11,819 INFO  mapred.JobClient - Running job: job_5ug2ip
> 2007-01-18 14:57:12,073 WARN  mapred.LocalJobRunner - job_5ug2ip
> java.io.EOFException
>         at java.io.DataInputStream.readFully(DataInputStream.java:178)
>         at org.apache.hadoop.io.DataOutputBuffer$Buffer.write(DataOutputBuffer.java:57)
>         at org.apache.hadoop.io.DataOutputBuffer.write(DataOutputBuffer.java:91)
>         at org.apache.hadoop.io.UTF8.readChars(UTF8.java:212)
>         at org.apache.hadoop.io.UTF8.readString(UTF8.java:204)
>         at org.apache.hadoop.io.ObjectWritable.readObject(ObjectWritable.java:173)
>         at org.apache.hadoop.io.ObjectWritable.readFields(ObjectWritable.java:61)
>         at org.apache.nutch.metadata.MetaWrapper.readFields(MetaWrapper.java:100)
>         at org.apache.hadoop.mapred.MapTask$MapOutputBuffer.spill(MapTask.java:427)
>         at org.apache.hadoop.mapred.MapTask$MapOutputBuffer.sortAndSpillToDisk(MapTask.java:385)
>         at org.apache.hadoop.mapred.MapTask$MapOutputBuffer.access$200(MapTask.java:239)
>         at org.apache.hadoop.mapred.MapTask.run(MapTask.java:188)
>         at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:109)

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.