You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "George Papa (Jira)" <ji...@apache.org> on 2019/09/11 13:14:00 UTC
[jira] [Updated] (SPARK-29055) Memory leak in Spark Driver
[ https://issues.apache.org/jira/browse/SPARK-29055?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
George Papa updated SPARK-29055:
--------------------------------
Attachment: (was: image-2019-09-11-16-13-32-650.png)
> Memory leak in Spark Driver
> ---------------------------
>
> Key: SPARK-29055
> URL: https://issues.apache.org/jira/browse/SPARK-29055
> Project: Spark
> Issue Type: Bug
> Components: Block Manager, Spark Core
> Affects Versions: 2.3.3, 2.4.0, 2.4.1, 2.4.2, 2.4.3, 2.4.4
> Reporter: George Papa
> Priority: Major
> Attachments: image-2019-09-11-16-13-20-588.png
>
>
> In Spark 2.3.3+ the driver memory is increasing continuously. I don't have this issue with Spark 2.1.1.
> In Spark 2.1.1 I see the ContextCleaner runs and cleans the driver and BlockManager removes the broadcast blocks from the memory, as you can see in the following screenshot:
> !image-2019-09-11-16-13-32-650.png|width=685,height=89!
> But in Spark 2.3.3+ I don't see this cleaning and the driver storage increases!!
> *NOTE:* After few hours of use I have application interruption with the following error :
> {color:#FF0000}java.lang.OutOfMemoryError: GC overhead limit exceeded{color}
>
--
This message was sent by Atlassian Jira
(v8.3.2#803003)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org