You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@nutch.apache.org by "Doğacan Güney (JIRA)" <ji...@apache.org> on 2007/06/18 10:57:26 UTC
[jira] Closed: (NUTCH-492) java.lang.OutOfMemoryError while
indexing.
[ https://issues.apache.org/jira/browse/NUTCH-492?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Doğacan Güney closed NUTCH-492.
-------------------------------
Resolution: Duplicate
Assignee: Doğacan Güney
Duplicate of NUTCH-356.
> java.lang.OutOfMemoryError while indexing.
> ------------------------------------------
>
> Key: NUTCH-492
> URL: https://issues.apache.org/jira/browse/NUTCH-492
> Project: Nutch
> Issue Type: Bug
> Components: indexer
> Affects Versions: 0.9.0
> Reporter: Nicolás Lichtmaier
> Assignee: Doğacan Güney
>
> I'm getting this:
> java.lang.OutOfMemoryError: Java heap space
> at java.util.Arrays.copyOf(Arrays.java:2786)
> at java.io.ByteArrayOutputStream.write(ByteArrayOutputStream.java:94)
> at java.io.DataOutputStream.write(DataOutputStream.java:90)
> at org.apache.hadoop.io.Text.writeString(Text.java:399)
> at org.apache.nutch.metadata.Metadata.write(Metadata.java:225)
> at org.apache.nutch.parse.ParseData.write(ParseData.java:165)
> at org.apache.hadoop.io.ObjectWritable.writeObject(ObjectWritable.java:154)
> at org.apache.hadoop.io.ObjectWritable.write(ObjectWritable.java:65)
> at org.apache.hadoop.mapred.MapTask$MapOutputBuffer.collect(MapTask.java:315)
> at org.apache.nutch.indexer.Indexer.map(Indexer.java:306)
> at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:48)
> at org.apache.hadoop.mapred.MapTask.run(MapTask.java:175)
> at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:126)
> 2007-05-26 11:07:22,517 FATAL indexer.Indexer - Indexer: java.io.IOException: Job failed!
> at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:604)
> at org.apache.nutch.indexer.Indexer.index(Indexer.java:273)
> at org.apache.nutch.indexer.Indexer.run(Indexer.java:295)
> at org.apache.hadoop.util.ToolBase.doMain(ToolBase.java:189)
> at org.apache.nutch.indexer.Indexer.main(Indexer.java:278)
> Something weird I'm seeing in hadoop.log is that the plugins are loaded again and again. I've created a custom plugin (if that can be causing something). According to the code a nre plugin repository is created for each "configuration object". I'm sure I'm not modifying the configuration object in any part of my code (I've checked).
> Why are the plugins loaded again and again and again until the heap is full?
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.