You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@nutch.apache.org by lei wang <nu...@gmail.com> on 2009/07/02 13:58:55 UTC

Re: How torunning nutch on 2G memory tasknode

yes,i also occurred this problem,when i export HADOOP_HEAPSIZE=2000
,4000,6000,my Memory is 6GB

On Wed, Jun 24, 2009 at 4:59 PM, SunGod <su...@cheemer.org> wrote:

> Error occurred in  "crawldb TestDB/crawldb" reduce phase
>
> i get error msg --- java.lang.OutOfMemoryError: Java heap space
>
> my command
>  bin/nutch crawl url -dir TestDB -depth 4 -threads 3
>
>  single fetchlist around in 200000
>
> my settings on the memory
>
> hadoop-env.sh
> export HADOOP_HEAPSIZE=800
>
> hadoop-site.xml
> <property>
>  <name>mapred.tasktracker.map.tasks.maximum</name>
>  <value>4</value>
> </property>
> <property>
>  <name>mapred.tasktracker.reduce.tasks.maximum</name>
>  <value>4</value>
> </property>
> <property>
>  <name>mapred.map.tasks</name>
>  <value>2</value>
> </property>
> <property>
>  <name>mapred.reduce.tasks</name>
>  <value>2</value>
> </property>
> <property>
>  <name>mapred.map.max.attempts</name>
>  <value>4</value>
> </property>
> <property>
>  <name>mapred.reduce.max.attempts</name>
>  <value>4</value>
> </property>
> <property>
>  <name>mapred.child.java.opts</name>
>  <value>-Xmx250m</value>
> </property>
>