You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Apache Spark (Jira)" <ji...@apache.org> on 2022/04/20 02:16:00 UTC

[jira] [Assigned] (SPARK-38960) Spark should fail fast if initial memory too large(set by "spark.executor.extraJavaOptions") for executor to start

     [ https://issues.apache.org/jira/browse/SPARK-38960?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Apache Spark reassigned SPARK-38960:
------------------------------------

    Assignee:     (was: Apache Spark)

> Spark should fail fast if initial memory too large(set by "spark.executor.extraJavaOptions") for executor to start
> ------------------------------------------------------------------------------------------------------------------
>
>                 Key: SPARK-38960
>                 URL: https://issues.apache.org/jira/browse/SPARK-38960
>             Project: Spark
>          Issue Type: Improvement
>          Components: Spark Core, Spark Submit, YARN
>    Affects Versions: 3.4.0
>            Reporter: panbingkun
>            Priority: Minor
>             Fix For: 3.4.0
>
>
> if you set initial memory(set by "spark.executor.extraJavaOptions=-Xms\{XXX}G" ) larger than maximum memory(set by "spark.executor.memory")
> Eg.
>      *spark.executor.memory=1G*
>      *spark.executor.extraJavaOptions=-Xms2G*
>  
> from the driver process you just see executor failures with no warning, since the more meaningful errors are buried in the executor logs. 
> Eg., on Yarn, you see:
> {noformat}
> Error occurred during initialization of VM
> Initial heap size set to a larger value than the maximum heap size{noformat}
> Instead we should just fail fast with a clear error message in the driver logs.



--
This message was sent by Atlassian Jira
(v8.20.7#820007)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org