You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@nutch.apache.org by Wojciech Ciesielski <wo...@softwaremind.pl> on 2006/03/28 13:24:17 UTC
problem with starting injection...
Hi there,
I can not initialize crawldb with set of URLs... After downloading nutch
from SVN, builiding it (successfully) with ant in cygwin environment, we
encounter something like this:
bin/nutch inject crawl/crawldb dmoz
060328 132017 Injector: starting
060328 132017 Injector: crawlDb: crawl\crawldb
060328 132017 Injector: urlDir: dmoz
060328 132018 parsing
jar:file:/T:/_data/nutch_svn/nutch/lib/hadoop-0.1-dev.jar!/hadoop-default.xml
060328 132018 parsing file:/T:/_data/nutch_svn/nutch/conf/nutch-default.xml
060328 132018 parsing file:/T:/_data/nutch_svn/nutch/conf/nutch-site.xml
060328 132018 parsing file:/T:/_data/nutch_svn/nutch/conf/hadoop-site.xml
060328 132018 Injector: Converting injected urls to crawl db entries.
060328 132018 parsing
jar:file:/T:/_data/nutch_svn/nutch/lib/hadoop-0.1-dev.jar!/hadoop-default.xml
060328 132018 parsing file:/T:/_data/nutch_svn/nutch/conf/nutch-default.xml
060328 132018 parsing
jar:file:/T:/_data/nutch_svn/nutch/lib/hadoop-0.1-dev.jar!/mapred-default.xml
060328 132018 parsing file:/T:/_data/nutch_svn/nutch/conf/nutch-site.xml
060328 132018 parsing file:/T:/_data/nutch_svn/nutch/conf/hadoop-site.xml
060328 132021 parsing
jar:file:/T:/_data/nutch_svn/nutch/lib/hadoop-0.1-dev.jar!/hadoop-default.xml
060328 132021 parsing file:/T:/_data/nutch_svn/nutch/conf/nutch-default.xml
060328 132021 parsing
jar:file:/T:/_data/nutch_svn/nutch/lib/hadoop-0.1-dev.jar!/mapred-default.xml
060328 132021 parsing
jar:file:/T:/_data/nutch_svn/nutch/lib/hadoop-0.1-dev.jar!/mapred-default.xml
060328 132021 parsing file:/T:/_data/nutch_svn/nutch/conf/nutch-site.xml
060328 132021 parsing file:/T:/_data/nutch_svn/nutch/conf/hadoop-site.xml
060328 132021 parsing
jar:file:/T:/_data/nutch_svn/nutch/lib/hadoop-0.1-dev.jar!/hadoop-default.xml
060328 132021 parsing
jar:file:/T:/_data/nutch_svn/nutch/lib/hadoop-0.1-dev.jar!/mapred-default.xml
060328 132021 SEVERE error parsing conf file:
java.lang.RuntimeException: \tmp\hadoop\mapred\local\job_s3hcq6.xml\localR
unner not found
Exception in thread "main" java.lang.RuntimeException:
java.lang.RuntimeException: \tmp\hadoop\mapred\local\job_s3hcq6.x
ml\localRunner not found
at
org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:405)
at
org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:345)
at
org.apache.hadoop.conf.Configuration.getProps(Configuration.java:333)
at org.apache.hadoop.conf.Configuration.get(Configuration.java:122)
at org.apache.hadoop.mapred.JobConf.getUser(JobConf.java:151)
at
org.apache.hadoop.mapred.LocalJobRunner$Job.<init>(LocalJobRunner.java:58)
at
org.apache.hadoop.mapred.LocalJobRunner.submitJob(LocalJobRunner.java:181)
at org.apache.hadoop.mapred.JobClient.submitJob(JobClient.java:273)
at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:306)
at org.apache.nutch.crawl.Injector.inject(Injector.java:115)
at org.apache.nutch.crawl.Injector.main(Injector.java:139)
Caused by: java.lang.RuntimeException:
\tmp\hadoop\mapred\local\job_s3hcq6.xml\localRunner not found
at
org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:372)
... 10 more
Any assistance would be appreciated, thanks in advance.
Wojtek
Re: problem with starting injection...
Posted by Andrzej Bialecki <ab...@getopt.org>.
Wojciech Ciesielski wrote:
> Hi there,
>
> I can not initialize crawldb with set of URLs... After downloading
> nutch from SVN, builiding it (successfully) with ant in cygwin
> environment, we encounter something like this:
Running on Cygwin is broken at the moment, please use a *nix environment
if it's an option, or revert to revision
388299 .
--
Best regards,
Andrzej Bialecki <><
___. ___ ___ ___ _ _ __________________________________
[__ || __|__/|__||\/| Information Retrieval, Semantic Web
___|||__|| \| || | Embedded Unix, System Integration
http://www.sigram.com Contact: info at sigram dot com