You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@nutch.apache.org by qi wu <ch...@gmail.com> on 2007/03/14 15:30:36 UTC

Any hints for debuging errors like "java.io.exception: read 95 bytes, should read 159" ?

Hi,
 
I am trying to modify the Fetcher code in Nutch.81 , but always get the exceptions below in the hadoop.log.

java.lang.RuntimeException: java.io.IOException: Version: 4
Status: 5 (fetch_success)
Fetch time: Wed Mar 14 22:15:58 CST 2007
Modified time: Thu Jan 01 08:00:00 CST 1970
Retries since fetch: 0
Fetch interval: 0.5 days
Score: 1.0
Signature: f4b0490606f416542246fb2bf625d0cb
Metadata: null
 read 95 bytes, should read 159
 at org.apache.hadoop.mapred.ReduceTask$ValuesIterator.next(ReduceTask.java:129)
 at org.apache.nutch.crawl.CrawlDbReducer.reduce(CrawlDbReducer.java:56)
 at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:247)
 at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:112) 

Why "159" bytes should be read here? Any hints for me to find the answer?

Thanks
-Qi
 

Re: Any hints for debuging errors like "java.io.exception: read 95 bytes, should read 159" ?

Posted by qi wu <ch...@gmail.com>.
Thank you Dennis!
Yes, the modified Fetcher.java of mine added new  values to metadata,but were not retrieved out during the subsequent process!

----- Original Message ----- 
From: "Dennis Kubes" <nu...@dragonflymc.com>
To: <nu...@lucene.apache.org>
Sent: Wednesday, March 14, 2007 11:40 PM
Subject: Re: Any hints for debuging errors like "java.io.exception: read 95 bytes, should read 159" ?


> Did you change the Parse metadata writable object.  This error looks 
> like ones that show up when a change to a writeable source code has been 
> made to write out something (i.e. like adding meta tags to the parsedata 
> metadata) but the code to read it in has not been made in the writeable 
> source.
> 
> The errors is basically stating that you wrote something out but haven't 
> read it back in.
> 
> Dennis Kubes
> 
> qi wu wrote:
>> Hi,
>>  
>> I am trying to modify the Fetcher code in Nutch.81 , but always get the exceptions below in the hadoop.log.
>> 
>> java.lang.RuntimeException: java.io.IOException: Version: 4
>> Status: 5 (fetch_success)
>> Fetch time: Wed Mar 14 22:15:58 CST 2007
>> Modified time: Thu Jan 01 08:00:00 CST 1970
>> Retries since fetch: 0
>> Fetch interval: 0.5 days
>> Score: 1.0
>> Signature: f4b0490606f416542246fb2bf625d0cb
>> Metadata: null
>>  read 95 bytes, should read 159
>>  at org.apache.hadoop.mapred.ReduceTask$ValuesIterator.next(ReduceTask.java:129)
>>  at org.apache.nutch.crawl.CrawlDbReducer.reduce(CrawlDbReducer.java:56)
>>  at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:247)
>>  at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:112) 
>> 
>> Why "159" bytes should be read here? Any hints for me to find the answer?
>> 
>> Thanks
>> -Qi
>>  
>

Re: Any hints for debuging errors like "java.io.exception: read 95 bytes, should read 159" ?

Posted by Dennis Kubes <nu...@dragonflymc.com>.
Did you change the Parse metadata writable object.  This error looks 
like ones that show up when a change to a writeable source code has been 
made to write out something (i.e. like adding meta tags to the parsedata 
metadata) but the code to read it in has not been made in the writeable 
source.

The errors is basically stating that you wrote something out but haven't 
read it back in.

Dennis Kubes

qi wu wrote:
> Hi,
>  
> I am trying to modify the Fetcher code in Nutch.81 , but always get the exceptions below in the hadoop.log.
> 
> java.lang.RuntimeException: java.io.IOException: Version: 4
> Status: 5 (fetch_success)
> Fetch time: Wed Mar 14 22:15:58 CST 2007
> Modified time: Thu Jan 01 08:00:00 CST 1970
> Retries since fetch: 0
> Fetch interval: 0.5 days
> Score: 1.0
> Signature: f4b0490606f416542246fb2bf625d0cb
> Metadata: null
>  read 95 bytes, should read 159
>  at org.apache.hadoop.mapred.ReduceTask$ValuesIterator.next(ReduceTask.java:129)
>  at org.apache.nutch.crawl.CrawlDbReducer.reduce(CrawlDbReducer.java:56)
>  at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:247)
>  at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:112) 
> 
> Why "159" bytes should be read here? Any hints for me to find the answer?
> 
> Thanks
> -Qi
>