You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Wenchen Fan (Jira)" <ji...@apache.org> on 2023/02/10 15:58:00 UTC
[jira] [Resolved] (SPARK-42162) Memory usage on executors increased drastically for a complex query with large number of addition operations
[ https://issues.apache.org/jira/browse/SPARK-42162?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Wenchen Fan resolved SPARK-42162.
---------------------------------
Fix Version/s: 3.4.0
Resolution: Fixed
Issue resolved by pull request 39722
[https://github.com/apache/spark/pull/39722]
> Memory usage on executors increased drastically for a complex query with large number of addition operations
> ------------------------------------------------------------------------------------------------------------
>
> Key: SPARK-42162
> URL: https://issues.apache.org/jira/browse/SPARK-42162
> Project: Spark
> Issue Type: Bug
> Components: SQL
> Affects Versions: 3.3.0
> Reporter: Supun Nakandala
> Priority: Major
> Fix For: 3.4.0
>
>
> With the [recent changes|https://github.com/apache/spark/pull/37851] in the expression canonicalization, a complex query with a large number of Add operations ends up consuming 10x more memory on the executors.
> The reason for this issue is that with the new changes the canonicalization process ends up generating lot of intermediate objects, especially for complex queries with a large number of commutative operators. In this specific case, a heap histogram analysis shows that a large number of Add objects use the extra memory.
> This issue does not happen before PR [#37851.|https://github.com/apache/spark/pull/37851]
> The high memory usage causes the executors to lose heartbeat signals and results in task failures.
--
This message was sent by Atlassian Jira
(v8.20.10#820010)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org