You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "harish chandra (JIRA)" <ji...@apache.org> on 2017/06/03 08:39:04 UTC
[jira] [Updated] (SPARK-20975) Excutors are no released if
speculation + dynamic allocation enabled
[ https://issues.apache.org/jira/browse/SPARK-20975?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
harish chandra updated SPARK-20975:
-----------------------------------
Description:
Whenever user enable spark speculation + dynamic allocation then after running few jobs in spark context, few executors keeps running like forever and never get free.
*Configuration*
{code:java}
- "spark.master=yarn-client"
- "spark.yarn.am.extraJavaOptions=-Dhdp.version=2.5.3.0-37"
- "spark.sql.sources.maxConcurrentWrites=1"
- "parquet.memory.pool.ratio=0.1"
- "hive.map.aggr=true"
- "spark.sql.shuffle.partitions=1200"
- "spark.scheduler.mode=FAIR"
- "spark.scheduler.allocation.file=/etc/spark/conf/fairscheduler.xml.template"
- "spark.speculation=true"
- "spark.dynamicAllocation.enabled=true"
- "spark.shuffle.service.enabled=true"
- "spark.dynamicAllocation.executorIdleTimeout=15s"
- "spark.dynamicAllocation.cachedExecutorIdleTimeout=15s"
- "spark.dynamicAllocation.initialExecutors=1"
- "spark.dynamicAllocation.maxExecutors=900"
- "spark.dynamicAllocation.minExecutors=1"
- "spark.yarn.max.executor.failures=10000"
- "spark.executor.cores=2"
- "spark.executor.memory=8G"
- "spark.sql.codegen=true"
- "spark.sql.codegen.wholeStage=true"
- "spark.sql.shuffle.partitions=75"
{code}
was:
Whenever user enable spark speculation + dynamic allocation then after running few jobs in spark context, few executors keeps running like forever and never get free.
*Configuration*
{code:java}
- "spark.master=yarn-client"
- "spark.yarn.am.extraJavaOptions=-Dhdp.version=2.5.3.0-37"
- "spark.sql.sources.maxConcurrentWrites=1"
- "parquet.memory.pool.ratio=0.1"
- "hive.map.aggr=true"
- "spark.sql.shuffle.partitions=1200"
- "spark.scheduler.mode=FAIR"
- "spark.scheduler.allocation.file=/etc/spark/conf/fairscheduler.xml.template"
- "spark.speculation=true"
- "spark.dynamicAllocation.enabled=true"
- "spark.shuffle.service.enabled=true"
- "spark.dynamicAllocation.executorIdleTimeout=15s"
- "spark.dynamicAllocation.cachedExecutorIdleTimeout=15s"
- "spark.dynamicAllocation.initialExecutors=1"
- "spark.dynamicAllocation.maxExecutors=900"
- "spark.dynamicAllocation.minExecutors=1"
- "spark.yarn.max.executor.failures=10000"
- "spark.yarn.queue=raven"
- "spark.executor.cores=2"
- "spark.executor.memory=8G"
- "spark.sql.codegen=true"
- "spark.sql.codegen.wholeStage=true"
- "spark.sql.shuffle.partitions=75"
{code}
> Excutors are no released if speculation + dynamic allocation enabled
> ---------------------------------------------------------------------
>
> Key: SPARK-20975
> URL: https://issues.apache.org/jira/browse/SPARK-20975
> Project: Spark
> Issue Type: Bug
> Components: Spark Core
> Affects Versions: 1.6.2
> Reporter: harish chandra
>
> Whenever user enable spark speculation + dynamic allocation then after running few jobs in spark context, few executors keeps running like forever and never get free.
> *Configuration*
> {code:java}
> - "spark.master=yarn-client"
> - "spark.yarn.am.extraJavaOptions=-Dhdp.version=2.5.3.0-37"
> - "spark.sql.sources.maxConcurrentWrites=1"
> - "parquet.memory.pool.ratio=0.1"
> - "hive.map.aggr=true"
> - "spark.sql.shuffle.partitions=1200"
> - "spark.scheduler.mode=FAIR"
> - "spark.scheduler.allocation.file=/etc/spark/conf/fairscheduler.xml.template"
> - "spark.speculation=true"
> - "spark.dynamicAllocation.enabled=true"
> - "spark.shuffle.service.enabled=true"
> - "spark.dynamicAllocation.executorIdleTimeout=15s"
> - "spark.dynamicAllocation.cachedExecutorIdleTimeout=15s"
> - "spark.dynamicAllocation.initialExecutors=1"
> - "spark.dynamicAllocation.maxExecutors=900"
> - "spark.dynamicAllocation.minExecutors=1"
> - "spark.yarn.max.executor.failures=10000"
> - "spark.executor.cores=2"
> - "spark.executor.memory=8G"
> - "spark.sql.codegen=true"
> - "spark.sql.codegen.wholeStage=true"
> - "spark.sql.shuffle.partitions=75"
> {code}
--
This message was sent by Atlassian JIRA
(v6.3.15#6346)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org