You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Shixiong Zhu (JIRA)" <ji...@apache.org> on 2017/11/10 22:16:01 UTC
[jira] [Resolved] (SPARK-19644) Memory leak in Spark Streaming
(Encoder/Scala Reflection)
[ https://issues.apache.org/jira/browse/SPARK-19644?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Shixiong Zhu resolved SPARK-19644.
----------------------------------
Resolution: Fixed
Assignee: Shixiong Zhu
Fix Version/s: 2.3.0
2.2.1
> Memory leak in Spark Streaming (Encoder/Scala Reflection)
> ---------------------------------------------------------
>
> Key: SPARK-19644
> URL: https://issues.apache.org/jira/browse/SPARK-19644
> Project: Spark
> Issue Type: Bug
> Components: DStreams, SQL, Structured Streaming
> Affects Versions: 2.0.2
> Environment: 3 AWS EC2 c3.xLarge
> Number of cores - 3
> Number of executors 3
> Memory to each executor 2GB
> Reporter: Deenbandhu Agarwal
> Assignee: Shixiong Zhu
> Labels: memory_leak, performance
> Fix For: 2.2.1, 2.3.0
>
> Attachments: Dominator_tree.png, Path2GCRoot.png, heapdump.png
>
>
> I am using streaming on the production for some aggregation and fetching data from cassandra and saving data back to cassandra.
> I see a gradual increase in old generation heap capacity from 1161216 Bytes to 1397760 Bytes over a period of six hours.
> After 50 hours of processing instances of class scala.collection.immutable.$colon$colon incresed to 12,811,793 which is a huge number.
> I think this is a clear case of memory leak
> Updated: The root cause is when creating an encoder object, it leaks several Scala internal objects due to a Scala memory leak issue: https://github.com/scala/bug/issues/8302
--
This message was sent by Atlassian JIRA
(v6.4.14#64029)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org