You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Amit Gupta (JIRA)" <ji...@apache.org> on 2015/05/04 13:21:06 UTC

[jira] [Reopened] (SPARK-7337) FPGrowth algo throwing OutOfMemoryError

     [ https://issues.apache.org/jira/browse/SPARK-7337?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Amit Gupta reopened SPARK-7337:
-------------------------------

I am running it in "local" mode and using Java API. It should spill over hard-disk. I can clearly see that 500 tasks are not created by next stage hence OutOfMemoryError is coming.

Please let me know if you need further information. I tried to run same logic with custom code with 50 partitions and it worked. When I tried FPGrowth with same data with >50 partitions it failed. Perhaps you need to check why 500 tasks are not created for "collect at FPGrowth.scala:131" (source of error).

> FPGrowth algo throwing OutOfMemoryError
> ---------------------------------------
>
>                 Key: SPARK-7337
>                 URL: https://issues.apache.org/jira/browse/SPARK-7337
>             Project: Spark
>          Issue Type: Bug
>          Components: MLlib
>    Affects Versions: 1.3.1
>         Environment: Ubuntu
>            Reporter: Amit Gupta
>         Attachments: FPGrowthBug.png
>
>
> When running FPGrowth algo with huge data in GBs and with numPartitions=500 then after some time it throws OutOfMemoryError.
> Algo runs correctly upto "collect at FPGrowth.scala:131" where it creates 500 tasks. It fails at next stage "flatMap at FPGrowth.scala:150" where it fails to create 500 tasks and create some internal calculated 17 tasks.
> Please refer to attachment - print screen.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org