You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@nutch.apache.org by blackwater dev <bl...@gmail.com> on 2009/01/01 04:51:55 UTC

craw failing

I just grabbed nutch and did some crawls and noticed when I did a search, I
was getting an error so looked into the log and see this:

Indexer: starting
Indexer: linkdb: crawl.test/linkdb
Indexer: adding segment: crawl.test/segments/20081224211810
Indexer: adding segment: crawl.test/segments/20081224211837
Indexer: adding segment: crawl.test/segments/20081224211858
Optimizing index.
Indexer: done
Dedup: starting
Dedup: adding indexes in: crawl.test/indexes
Exception in thread "main" java.io.IOException: Job failed!
        at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:604)
        at
org.apache.nutch.indexer.DeleteDuplicates.dedup(DeleteDuplicates.java:439)
        at org.apache.nutch.crawl.Crawl.main(Crawl.java:135)

How can I fix this?


Thanks!