You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@nutch.apache.org by qi wu <ch...@gmail.com> on 2007/03/14 15:30:36 UTC
Any hints for debuging errors like "java.io.exception: read 95 bytes, should read 159" ?
Hi,
I am trying to modify the Fetcher code in Nutch.81 , but always get the exceptions below in the hadoop.log.
java.lang.RuntimeException: java.io.IOException: Version: 4
Status: 5 (fetch_success)
Fetch time: Wed Mar 14 22:15:58 CST 2007
Modified time: Thu Jan 01 08:00:00 CST 1970
Retries since fetch: 0
Fetch interval: 0.5 days
Score: 1.0
Signature: f4b0490606f416542246fb2bf625d0cb
Metadata: null
read 95 bytes, should read 159
at org.apache.hadoop.mapred.ReduceTask$ValuesIterator.next(ReduceTask.java:129)
at org.apache.nutch.crawl.CrawlDbReducer.reduce(CrawlDbReducer.java:56)
at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:247)
at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:112)
Why "159" bytes should be read here? Any hints for me to find the answer?
Thanks
-Qi
Re: Any hints for debuging errors like "java.io.exception: read 95 bytes, should read 159" ?
Posted by qi wu <ch...@gmail.com>.
Thank you Dennis!
Yes, the modified Fetcher.java of mine added new values to metadata,but were not retrieved out during the subsequent process!
----- Original Message -----
From: "Dennis Kubes" <nu...@dragonflymc.com>
To: <nu...@lucene.apache.org>
Sent: Wednesday, March 14, 2007 11:40 PM
Subject: Re: Any hints for debuging errors like "java.io.exception: read 95 bytes, should read 159" ?
> Did you change the Parse metadata writable object. This error looks
> like ones that show up when a change to a writeable source code has been
> made to write out something (i.e. like adding meta tags to the parsedata
> metadata) but the code to read it in has not been made in the writeable
> source.
>
> The errors is basically stating that you wrote something out but haven't
> read it back in.
>
> Dennis Kubes
>
> qi wu wrote:
>> Hi,
>>
>> I am trying to modify the Fetcher code in Nutch.81 , but always get the exceptions below in the hadoop.log.
>>
>> java.lang.RuntimeException: java.io.IOException: Version: 4
>> Status: 5 (fetch_success)
>> Fetch time: Wed Mar 14 22:15:58 CST 2007
>> Modified time: Thu Jan 01 08:00:00 CST 1970
>> Retries since fetch: 0
>> Fetch interval: 0.5 days
>> Score: 1.0
>> Signature: f4b0490606f416542246fb2bf625d0cb
>> Metadata: null
>> read 95 bytes, should read 159
>> at org.apache.hadoop.mapred.ReduceTask$ValuesIterator.next(ReduceTask.java:129)
>> at org.apache.nutch.crawl.CrawlDbReducer.reduce(CrawlDbReducer.java:56)
>> at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:247)
>> at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:112)
>>
>> Why "159" bytes should be read here? Any hints for me to find the answer?
>>
>> Thanks
>> -Qi
>>
>
Re: Any hints for debuging errors like "java.io.exception: read
95 bytes, should read 159" ?
Posted by Dennis Kubes <nu...@dragonflymc.com>.
Did you change the Parse metadata writable object. This error looks
like ones that show up when a change to a writeable source code has been
made to write out something (i.e. like adding meta tags to the parsedata
metadata) but the code to read it in has not been made in the writeable
source.
The errors is basically stating that you wrote something out but haven't
read it back in.
Dennis Kubes
qi wu wrote:
> Hi,
>
> I am trying to modify the Fetcher code in Nutch.81 , but always get the exceptions below in the hadoop.log.
>
> java.lang.RuntimeException: java.io.IOException: Version: 4
> Status: 5 (fetch_success)
> Fetch time: Wed Mar 14 22:15:58 CST 2007
> Modified time: Thu Jan 01 08:00:00 CST 1970
> Retries since fetch: 0
> Fetch interval: 0.5 days
> Score: 1.0
> Signature: f4b0490606f416542246fb2bf625d0cb
> Metadata: null
> read 95 bytes, should read 159
> at org.apache.hadoop.mapred.ReduceTask$ValuesIterator.next(ReduceTask.java:129)
> at org.apache.nutch.crawl.CrawlDbReducer.reduce(CrawlDbReducer.java:56)
> at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:247)
> at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:112)
>
> Why "159" bytes should be read here? Any hints for me to find the answer?
>
> Thanks
> -Qi
>