You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@nutch.apache.org by Sebastian Nagel <wa...@googlemail.com> on 2011/09/30 22:53:17 UTC

Re: 1.4 release - newer hadoop jars

> can you package 1.4 with updated hadoop jars? i have problems with running nutch in local mode. If i
> run multiple tasks at once, they delete each other temporary files. Its worth a try if newer hadoop
> libs will fix that.

Hi Radim,

I don't know whether current versions of hadoop fix this problem.
It's a bitter experience made by many Nutch users that running tasks
simultaneously in local mode may fail accidentally from time to time,
see http://www.mail-archive.com/user@nutch.apache.org/msg03239.html

Set hadoop.tmp.dir for each job so that it points to an unique directory.
Instead of setting it in the xml config file which is uncomfortable
you can set it via java command line options or the NUTCH_OPTS environment
variable:
  NUTCH_OPTS=-Dhadoop.tmp.dir=/tmp/uniq_dir $NUTCH_HOME/bin/nutch ...

Sebastian