You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Patrick Wendell (JIRA)" <ji...@apache.org> on 2014/08/16 07:58:18 UTC

[jira] [Resolved] (SPARK-2916) [MLlib] While running regression tests with dense vectors of length greater than 1000, the treeAggregate blows up after several iterations

     [ https://issues.apache.org/jira/browse/SPARK-2916?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Patrick Wendell resolved SPARK-2916.
------------------------------------

    Resolution: Fixed

Fixed by virtue of SPARK-3015

> [MLlib] While running regression tests with dense vectors of length greater than 1000, the treeAggregate blows up after several iterations
> ------------------------------------------------------------------------------------------------------------------------------------------
>
>                 Key: SPARK-2916
>                 URL: https://issues.apache.org/jira/browse/SPARK-2916
>             Project: Spark
>          Issue Type: Bug
>          Components: MLlib, Spark Core
>            Reporter: Burak Yavuz
>            Priority: Blocker
>
> While running any of the regression algorithms with gradient descent, the treeAggregate blows up after several iterations.
> Observed on EC2 cluster with 16 nodes, matrix dimensions of 1,000,000 x 5,000
> In order to replicate the problem, use aggregate multiple times, maybe over 50-60 times.
> Testing lead to the possible workaround:
> setting 
> `spark.cleaner.referenceTracking false`
> seems to help. So the problem is most probably related to the cleanup.



--
This message was sent by Atlassian JIRA
(v6.2#6252)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org