You are viewing a plain text version of this content. The canonical link for it is here.
Posted to reviews@spark.apache.org by ryan-williams <gi...@git.apache.org> on 2015/10/16 06:57:29 UTC

[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

GitHub user ryan-williams opened a pull request:

    https://github.com/apache/spark/pull/9147

    [SPARK-11120] Allow sane default number of executor failures when dynamically allocating in YARN

    I also added some information to container-failure error msgs about what host they failed on, which would have helped me identify the problem that lead me to this JIRA and PR sooner.

You can merge this pull request into a Git repository by running:

    $ git pull https://github.com/ryan-williams/spark dyn-exec-failures

Alternatively you can review and apply these changes as the patch at:

    https://github.com/apache/spark/pull/9147.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

    This closes #9147
    
----
commit 5b91832ce3d7b813af62825ac6e01bebc5ed57c5
Author: Ryan Williams <ry...@gmail.com>
Date:   2015-10-15T00:40:03Z

    allow 2*maxExecutors failures when dyn-alloc'ing

commit 46633e74d0f732a293e8ca4cbfc6a5d683d6cc99
Author: Ryan Williams <ry...@gmail.com>
Date:   2015-10-15T02:05:16Z

    add host info to container-failure log msgs

----


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by andrewor14 <gi...@git.apache.org>.
Github user andrewor14 commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-149376554
  
    Actually, I'm just going to merge this since all the YARN / dynamic allocation tests already passed. The one that failed is a known flaky test. This is going into master. Thanks @ryan-williams.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148841298
  
    Merged build finished. Test PASSed.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148866814
  
    Merged build started.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by andrewor14 <gi...@git.apache.org>.
Github user andrewor14 commented on a diff in the pull request:

    https://github.com/apache/spark/pull/9147#discussion_r42301083
  
    --- Diff: yarn/src/main/scala/org/apache/spark/deploy/yarn/ApplicationMaster.scala ---
    @@ -62,10 +62,23 @@ private[spark] class ApplicationMaster(
         .asInstanceOf[YarnConfiguration]
       private val isClusterMode = args.userClass != null
     
    -  // Default to numExecutors * 2, with minimum of 3
    -  private val maxNumExecutorFailures = sparkConf.getInt("spark.yarn.max.executor.failures",
    -    sparkConf.getInt("spark.yarn.max.worker.failures",
    -      math.max(sparkConf.getInt("spark.executor.instances", 0) *  2, 3)))
    +  // Default to numExecutors * 2 (maxExecutors in the case that we are
    +  // dynamically allocating executors), with minimum of 3.
    +  private val maxNumExecutorFailures = {
    +    val effectiveNumExecutors =
    +      sparkConf.getInt(
    +        if (Utils.isDynamicAllocationEnabled(sparkConf)) {
    +          "spark.dynamicAllocation.maxExecutors"
    +        } else {
    +          "spark.executor.instances"
    +        },
    +        0
    +      )
    +
    +    val defaultMaxNumExecutorFailures = math.max(3, 2 * effectiveNumExecutors)
    +
    --- End diff --
    
    can you rewrite this so it's more readable:
    ```
    val defaultKey =
      if (Utils.isDynamicAllocationEnabled(sparkConf)) {
        "spark.dynamicAllocation.maxExecutors"
      } else {
        "spark.executor.instances"
      }
    val effectiveNumExecutors = sparkConf.getInt(defaultKey, 0)
    val defaultMaxNumExeuctorFailures = math.max(3, 2 * effectiveNumExecutors)
    sparkConf.getInt(..., defaultMaxNumExecutorFailures)
    ```


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by SparkQA <gi...@git.apache.org>.
Github user SparkQA commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148609868
  
      [Test build #43831 has finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/43831/console) for   PR 9147 at commit [`46633e7`](https://github.com/apache/spark/commit/46633e74d0f732a293e8ca4cbfc6a5d683d6cc99).
     * This patch **fails Scala style tests**.
     * This patch merges cleanly.
     * This patch adds no public classes.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148866225
  
    Merged build started.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148845281
  
    Merged build finished. Test FAILed.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148822214
  
    Merged build started.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by SparkQA <gi...@git.apache.org>.
Github user SparkQA commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148841106
  
      [Test build #43841 has finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/43841/console) for   PR 9147 at commit [`ba53ed5`](https://github.com/apache/spark/commit/ba53ed5dc5fd9b9b7f9383843368ea00f04c7a55).
     * This patch **passes all tests**.
     * This patch merges cleanly.
     * This patch adds no public classes.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-149376261
  
    Test FAILed.
    Refer to this link for build results (access rights to CI server needed): 
    https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/43949/
    Test FAILed.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148867278
  
    Merged build started.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by SparkQA <gi...@git.apache.org>.
Github user SparkQA commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148823093
  
      [Test build #43849 has started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/43849/consoleFull) for   PR 9147 at commit [`aee9396`](https://github.com/apache/spark/commit/aee9396c791da134a3dbbe87d72af1fbc8f37099).


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148867267
  
     Merged build triggered.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148841300
  
    Test PASSed.
    Refer to this link for build results (access rights to CI server needed): 
    https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/43841/
    Test PASSed.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by vanzin <gi...@git.apache.org>.
Github user vanzin commented on a diff in the pull request:

    https://github.com/apache/spark/pull/9147#discussion_r42280906
  
    --- Diff: yarn/src/main/scala/org/apache/spark/deploy/yarn/ApplicationMaster.scala ---
    @@ -62,10 +62,24 @@ private[spark] class ApplicationMaster(
         .asInstanceOf[YarnConfiguration]
       private val isClusterMode = args.userClass != null
     
    -  // Default to numExecutors * 2, with minimum of 3
    -  private val maxNumExecutorFailures = sparkConf.getInt("spark.yarn.max.executor.failures",
    -    sparkConf.getInt("spark.yarn.max.worker.failures",
    -      math.max(sparkConf.getInt("spark.executor.instances", 0) *  2, 3)))
    +  // Default to numExecutors * 2 (maxExecutors in the case that we are
    +  // dynamically allocating executors), with minimum of 3.
    +  private val maxNumExecutorFailures =
    +    sparkConf.getInt("spark.yarn.max.executor.failures",
    +      sparkConf.getInt("spark.yarn.max.worker.failures",
    --- End diff --
    
    This is not your fault, but since you're touching it... this should be added to `SparkConf.configsWithAlternatives` instead.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by ryan-williams <gi...@git.apache.org>.
Github user ryan-williams commented on a diff in the pull request:

    https://github.com/apache/spark/pull/9147#discussion_r42284143
  
    --- Diff: yarn/src/main/scala/org/apache/spark/deploy/yarn/ApplicationMaster.scala ---
    @@ -62,10 +62,24 @@ private[spark] class ApplicationMaster(
         .asInstanceOf[YarnConfiguration]
       private val isClusterMode = args.userClass != null
     
    -  // Default to numExecutors * 2, with minimum of 3
    -  private val maxNumExecutorFailures = sparkConf.getInt("spark.yarn.max.executor.failures",
    -    sparkConf.getInt("spark.yarn.max.worker.failures",
    -      math.max(sparkConf.getInt("spark.executor.instances", 0) *  2, 3)))
    +  // Default to numExecutors * 2 (maxExecutors in the case that we are
    +  // dynamically allocating executors), with minimum of 3.
    +  private val maxNumExecutorFailures =
    +    sparkConf.getInt("spark.yarn.max.executor.failures",
    +      sparkConf.getInt("spark.yarn.max.worker.failures",
    --- End diff --
    
    ah, hadn't seen that, done


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by ryan-williams <gi...@git.apache.org>.
Github user ryan-williams commented on a diff in the pull request:

    https://github.com/apache/spark/pull/9147#discussion_r42271853
  
    --- Diff: yarn/src/main/scala/org/apache/spark/deploy/yarn/ApplicationMaster.scala ---
    @@ -62,10 +62,23 @@ private[spark] class ApplicationMaster(
         .asInstanceOf[YarnConfiguration]
       private val isClusterMode = args.userClass != null
     
    -  // Default to numExecutors * 2, with minimum of 3
    -  private val maxNumExecutorFailures = sparkConf.getInt("spark.yarn.max.executor.failures",
    -    sparkConf.getInt("spark.yarn.max.worker.failures",
    -      math.max(sparkConf.getInt("spark.executor.instances", 0) *  2, 3)))
    +  // Default to numExecutors * 2 (maxExecutors in the case that we are
    +  // dynamically allocating executors), with minimum of 3.
    +  private val maxNumExecutorFailures =
    +    sparkConf.getInt("spark.yarn.max.executor.failures",
    +      sparkConf.getInt("spark.yarn.max.worker.failures",
    +        math.max(
    +          3,
    +          2 * sparkConf.getInt(
    +            if (Utils.isDynamicAllocationEnabled(sparkConf))
    +              "spark.dynamicAllocation.maxExecutors"
    --- End diff --
    
    To be clear, this change does not place any additional requirements on a user to set `maxExecutors` to get sane dynamic allocation (DA) default behavior.
    
    It merely alleviates one class of "gotcha" that caused me some trouble this week: when setting standard DA params, the `val maxNumExecutorFailures` here becomes `3` by default, which does not seem sensible for apps that are going up to many 100s of executors.
    
    It seems to me that the extant `math.max(sparkConf.getInt("spark.executor.instances", 0) *  2, 3)` expression is not _intentionally_ making DA apps have a limit of `3` failures, but that it simply wasn't taking into account the fact that `spark.executor.instances` is not set in DA mode.
    
    It's true that we could also "resolve" this by declaring `spark.yarn.max.worker.failures` to be yet another configuration param that must be set to a non-default value in order to get sane DA behavior.
    
    Off the top of my head, there is already one param (`spark.shuffle.service.enabled=true`) that is not named in a way that suggests that it is important for DA apps to set, and we could make `spark.yarn.max.worker.failures` a second.
    
    My belief is that it would be better to not require yet another parameter (especially one that is not named in a way that makes it obvious that it is or could be important for DA to not fail in unexpected ways) for sane DA behavior, but to just fix the clearly-inadvertently-missed setting of a good default value here.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by SparkQA <gi...@git.apache.org>.
Github user SparkQA commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148793946
  
      [Test build #43841 has started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/43841/consoleFull) for   PR 9147 at commit [`ba53ed5`](https://github.com/apache/spark/commit/ba53ed5dc5fd9b9b7f9383843368ea00f04c7a55).


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148790093
  
     Merged build triggered.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by SparkQA <gi...@git.apache.org>.
Github user SparkQA commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148875385
  
      [Test build #43872 has finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/43872/console) for   PR 9147 at commit [`aa5d349`](https://github.com/apache/spark/commit/aa5d3491f15828c2b9ae93a16ab7a4503c8a4341).
     * This patch **fails Spark unit tests**.
     * This patch merges cleanly.
     * This patch adds no public classes.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by vanzin <gi...@git.apache.org>.
Github user vanzin commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148815611
  
    Change seems fine to me, I just think the code could be cleaned up a bit to help readability.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by vanzin <gi...@git.apache.org>.
Github user vanzin commented on a diff in the pull request:

    https://github.com/apache/spark/pull/9147#discussion_r42288525
  
    --- Diff: yarn/src/main/scala/org/apache/spark/deploy/yarn/ApplicationMaster.scala ---
    @@ -62,10 +62,26 @@ private[spark] class ApplicationMaster(
         .asInstanceOf[YarnConfiguration]
       private val isClusterMode = args.userClass != null
     
    -  // Default to numExecutors * 2, with minimum of 3
    -  private val maxNumExecutorFailures = sparkConf.getInt("spark.yarn.max.executor.failures",
    -    sparkConf.getInt("spark.yarn.max.worker.failures",
    -      math.max(sparkConf.getInt("spark.executor.instances", 0) *  2, 3)))
    +  // Default to numExecutors * 2 (maxExecutors in the case that we are
    +  // dynamically allocating executors), with minimum of 3.
    +  private val maxNumExecutorFailures = {
    +    val effectiveNumExecutors =
    +      sparkConf.getInt(
    +        if (Utils.isDynamicAllocationEnabled(sparkConf)) {
    +          "spark.dynamicAllocation.maxExecutors"
    +        } else {
    +          "spark.executor.instances"
    +        },
    +        0
    +      )
    +
    +    val defaultMaxNumExecutorFailures = math.max(3, 2 * effectiveNumExecutors)
    +
    +    sparkConf.getInt(
    --- End diff --
    
    minor nit, but this fits in one line now.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148845286
  
    Test FAILed.
    Refer to this link for build results (access rights to CI server needed): 
    https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/43859/
    Test FAILed.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148852142
  
    Merged build finished. Test FAILed.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by SparkQA <gi...@git.apache.org>.
Github user SparkQA commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148867661
  
      [Test build #43872 has started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/43872/consoleFull) for   PR 9147 at commit [`aa5d349`](https://github.com/apache/spark/commit/aa5d3491f15828c2b9ae93a16ab7a4503c8a4341).


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by SparkQA <gi...@git.apache.org>.
Github user SparkQA commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148856541
  
      [Test build #43848 has finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/43848/console) for   PR 9147 at commit [`32b208a`](https://github.com/apache/spark/commit/32b208a2feb047aa4663fcb8d8697e60e8b50829).
     * This patch **passes all tests**.
     * This patch merges cleanly.
     * This patch adds no public classes.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148790122
  
    Merged build started.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by andrewor14 <gi...@git.apache.org>.
Github user andrewor14 commented on a diff in the pull request:

    https://github.com/apache/spark/pull/9147#discussion_r42301236
  
    --- Diff: yarn/src/main/scala/org/apache/spark/deploy/yarn/ApplicationMaster.scala ---
    @@ -62,10 +62,19 @@ private[spark] class ApplicationMaster(
         .asInstanceOf[YarnConfiguration]
       private val isClusterMode = args.userClass != null
     
    -  // Default to numExecutors * 2, with minimum of 3
    -  private val maxNumExecutorFailures = sparkConf.getInt("spark.yarn.max.executor.failures",
    -    sparkConf.getInt("spark.yarn.max.worker.failures",
    -      math.max(sparkConf.getInt("spark.executor.instances", 0) *  2, 3)))
    +  // Default to twice the number of executors (twice the maximum number of executors if dynamic
    +  // allocation is enabled), with a minimum of 3.
    +  val defaultKey =
    +    if (Utils.isDynamicAllocationEnabled(sparkConf)) {
    +      "spark.dynamicAllocation.maxExecutors"
    +    } else {
    +      "spark.executor.instances"
    +    }
    +  val effectiveNumExecutors = sparkConf.getInt(defaultKey, 0)
    +  val defaultMaxNumExecutorFailures = math.max(3, 2 * effectiveNumExecutors)
    --- End diff --
    
    you addressed my comment quickly, but these should not be public or outside the scope of `maxNumExecutorFailures`. What I meant is more like:
    ```
    // Default to twice ...
    private val maxNumExecutorFailures: Int = {
      val defaultKey = ...
      sparkConf.getInt(...)
    }
    ```


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by SparkQA <gi...@git.apache.org>.
Github user SparkQA commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148822901
  
      [Test build #43848 has started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/43848/consoleFull) for   PR 9147 at commit [`32b208a`](https://github.com/apache/spark/commit/32b208a2feb047aa4663fcb8d8697e60e8b50829).


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-149373843
  
    Merged build started.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148867752
  
    Merged build finished. Test FAILed.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by ryan-williams <gi...@git.apache.org>.
Github user ryan-williams commented on a diff in the pull request:

    https://github.com/apache/spark/pull/9147#discussion_r42301346
  
    --- Diff: yarn/src/main/scala/org/apache/spark/deploy/yarn/ApplicationMaster.scala ---
    @@ -62,10 +62,19 @@ private[spark] class ApplicationMaster(
         .asInstanceOf[YarnConfiguration]
       private val isClusterMode = args.userClass != null
     
    -  // Default to numExecutors * 2, with minimum of 3
    -  private val maxNumExecutorFailures = sparkConf.getInt("spark.yarn.max.executor.failures",
    -    sparkConf.getInt("spark.yarn.max.worker.failures",
    -      math.max(sparkConf.getInt("spark.executor.instances", 0) *  2, 3)))
    +  // Default to twice the number of executors (twice the maximum number of executors if dynamic
    +  // allocation is enabled), with a minimum of 3.
    +  val defaultKey =
    +    if (Utils.isDynamicAllocationEnabled(sparkConf)) {
    +      "spark.dynamicAllocation.maxExecutors"
    +    } else {
    +      "spark.executor.instances"
    +    }
    +  val effectiveNumExecutors = sparkConf.getInt(defaultKey, 0)
    +  val defaultMaxNumExecutorFailures = math.max(3, 2 * effectiveNumExecutors)
    --- End diff --
    
    woops, didn't notice that you still wanted that scoping, fixed


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148822192
  
     Merged build triggered.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by ryan-williams <gi...@git.apache.org>.
Github user ryan-williams commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148848610
  
    Test failure above looks like a jenkinsāŸ·git issue?


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148875423
  
    Test FAILed.
    Refer to this link for build results (access rights to CI server needed): 
    https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/43872/
    Test FAILed.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by ryan-williams <gi...@git.apache.org>.
Github user ryan-williams commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148867212
  
    not super excited to write a test for this, tbh. what kind of test are you imagining? `ApplicationMaster*Suite` doesn't even exist atm...


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by andrewor14 <gi...@git.apache.org>.
Github user andrewor14 commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-149373168
  
    retest this please


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148607737
  
     Merged build triggered.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148875422
  
    Merged build finished. Test FAILed.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by jerryshao <gi...@git.apache.org>.
Github user jerryshao commented on a diff in the pull request:

    https://github.com/apache/spark/pull/9147#discussion_r42218995
  
    --- Diff: yarn/src/main/scala/org/apache/spark/deploy/yarn/ApplicationMaster.scala ---
    @@ -62,10 +62,23 @@ private[spark] class ApplicationMaster(
         .asInstanceOf[YarnConfiguration]
       private val isClusterMode = args.userClass != null
     
    -  // Default to numExecutors * 2, with minimum of 3
    -  private val maxNumExecutorFailures = sparkConf.getInt("spark.yarn.max.executor.failures",
    -    sparkConf.getInt("spark.yarn.max.worker.failures",
    -      math.max(sparkConf.getInt("spark.executor.instances", 0) *  2, 3)))
    +  // Default to numExecutors * 2 (maxExecutors in the case that we are
    +  // dynamically allocating executors), with minimum of 3.
    +  private val maxNumExecutorFailures =
    +    sparkConf.getInt("spark.yarn.max.executor.failures",
    +      sparkConf.getInt("spark.yarn.max.worker.failures",
    +        math.max(
    +          3,
    +          2 * sparkConf.getInt(
    +            if (Utils.isDynamicAllocationEnabled(sparkConf))
    +              "spark.dynamicAllocation.maxExecutors"
    --- End diff --
    
    If "spark.dynamicAllocation.maxExecutors" is not set, it is still 3 as minimum. So from your change, user have to configure "spark.dynamicAllocation.maxExecutors" to increase max failure attempts if dynamic allocation is enabled. So in that way why not directly configure "spark.yarn.max.worker.failures", that will be more straightforward?


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148866805
  
     Merged build triggered.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by vanzin <gi...@git.apache.org>.
Github user vanzin commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148831817
  
    LGTM pending tests.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by asfgit <gi...@git.apache.org>.
Github user asfgit closed the pull request at:

    https://github.com/apache/spark/pull/9147


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by ryan-williams <gi...@git.apache.org>.
Github user ryan-williams commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-149254257
  
    Seems like this did not attempt to retest, @andrewor14?


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148607750
  
    Merged build started.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by SparkQA <gi...@git.apache.org>.
Github user SparkQA commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-149256879
  
    **[Test build #1929 has started](https://amplab.cs.berkeley.edu/jenkins/job/NewSparkPullRequestBuilder/1929/consoleFull)** for PR 9147 at commit [`aa5d349`](https://github.com/apache/spark/commit/aa5d3491f15828c2b9ae93a16ab7a4503c8a4341).


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by ryan-williams <gi...@git.apache.org>.
Github user ryan-williams commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-149301882
  
    not sure how [this test failure](https://amplab.cs.berkeley.edu/jenkins/job/NewSparkPullRequestBuilder/1929/testReport/org.apache.spark.launcher/LauncherBackendSuite/standalone_client__launcher_handle/) could be related to this change


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by vanzin <gi...@git.apache.org>.
Github user vanzin commented on a diff in the pull request:

    https://github.com/apache/spark/pull/9147#discussion_r42281493
  
    --- Diff: yarn/src/main/scala/org/apache/spark/deploy/yarn/ApplicationMaster.scala ---
    @@ -62,10 +62,24 @@ private[spark] class ApplicationMaster(
         .asInstanceOf[YarnConfiguration]
       private val isClusterMode = args.userClass != null
     
    -  // Default to numExecutors * 2, with minimum of 3
    -  private val maxNumExecutorFailures = sparkConf.getInt("spark.yarn.max.executor.failures",
    -    sparkConf.getInt("spark.yarn.max.worker.failures",
    -      math.max(sparkConf.getInt("spark.executor.instances", 0) *  2, 3)))
    +  // Default to numExecutors * 2 (maxExecutors in the case that we are
    +  // dynamically allocating executors), with minimum of 3.
    +  private val maxNumExecutorFailures =
    +    sparkConf.getInt("spark.yarn.max.executor.failures",
    +      sparkConf.getInt("spark.yarn.max.worker.failures",
    +        math.max(
    +          3,
    +          2 * sparkConf.getInt(
    --- End diff --
    
    I think it would be easier to read if this were pulled into another variable.
    
        private val maxNumExecutorFailures = {
          val executorCount = sparkConf.getInt(...)
          val defaultFailures = math.max(3, executorCount)
          sparkConf.getInt(...)
        }



---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148821314
  
    Merged build started.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148609872
  
    Test FAILed.
    Refer to this link for build results (access rights to CI server needed): 
    https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/43831/
    Test FAILed.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148866216
  
     Merged build triggered.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by andrewor14 <gi...@git.apache.org>.
Github user andrewor14 commented on a diff in the pull request:

    https://github.com/apache/spark/pull/9147#discussion_r42300922
  
    --- Diff: yarn/src/main/scala/org/apache/spark/deploy/yarn/ApplicationMaster.scala ---
    @@ -62,10 +62,23 @@ private[spark] class ApplicationMaster(
         .asInstanceOf[YarnConfiguration]
       private val isClusterMode = args.userClass != null
     
    -  // Default to numExecutors * 2, with minimum of 3
    -  private val maxNumExecutorFailures = sparkConf.getInt("spark.yarn.max.executor.failures",
    -    sparkConf.getInt("spark.yarn.max.worker.failures",
    -      math.max(sparkConf.getInt("spark.executor.instances", 0) *  2, 3)))
    +  // Default to numExecutors * 2 (maxExecutors in the case that we are
    +  // dynamically allocating executors), with minimum of 3.
    --- End diff --
    
    this is hard to read. Can you rephrase:
    ```
    // If dynamic allocation is enabled, default to the upper limit.
    // Otherwise, default to 2 * numExecutors with a minimum of 3.
    ```


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by SparkQA <gi...@git.apache.org>.
Github user SparkQA commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-149299142
  
    **[Test build #1929 has finished](https://amplab.cs.berkeley.edu/jenkins/job/NewSparkPullRequestBuilder/1929/consoleFull)** for PR 9147 at commit [`aa5d349`](https://github.com/apache/spark/commit/aa5d3491f15828c2b9ae93a16ab7a4503c8a4341).
     * This patch **fails Spark unit tests**.
     * This patch merges cleanly.
     * This patch adds no public classes.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by ryan-williams <gi...@git.apache.org>.
Github user ryan-williams commented on a diff in the pull request:

    https://github.com/apache/spark/pull/9147#discussion_r42387794
  
    --- Diff: yarn/src/main/scala/org/apache/spark/deploy/yarn/YarnAllocator.scala ---
    @@ -430,17 +430,20 @@ private[yarn] class YarnAllocator(
         for (completedContainer <- completedContainers) {
           val containerId = completedContainer.getContainerId
           val alreadyReleased = releasedContainers.remove(containerId)
    +      val hostOpt = allocatedContainerToHostMap.get(containerId)
    +      val onHostStr = hostOpt.map(host => s" on host: $host").getOrElse("")
           val exitReason = if (!alreadyReleased) {
             // Decrement the number of executors running. The next iteration of
             // the ApplicationMaster's reporting thread will take care of allocating.
             numExecutorsRunning -= 1
    -        logInfo("Completed container %s (state: %s, exit status: %s)".format(
    +        logInfo("Completed container %s%s (state: %s, exit status: %s)".format(
    --- End diff --
    
    Yea I originally did some more aggressive converting to interpolated-strings here because that seems like a better way to do things in general, but such strings often make for absurdly long lines, and I don't have a solution I like to that (e.g. breaking onto multiple lines with `+`s seems gross), so I think I reverted that here and went for the more minimal change.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-149373822
  
     Merged build triggered.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by ryan-williams <gi...@git.apache.org>.
Github user ryan-williams commented on a diff in the pull request:

    https://github.com/apache/spark/pull/9147#discussion_r42284385
  
    --- Diff: yarn/src/main/scala/org/apache/spark/deploy/yarn/ApplicationMaster.scala ---
    @@ -62,10 +62,24 @@ private[spark] class ApplicationMaster(
         .asInstanceOf[YarnConfiguration]
       private val isClusterMode = args.userClass != null
     
    -  // Default to numExecutors * 2, with minimum of 3
    -  private val maxNumExecutorFailures = sparkConf.getInt("spark.yarn.max.executor.failures",
    -    sparkConf.getInt("spark.yarn.max.worker.failures",
    -      math.max(sparkConf.getInt("spark.executor.instances", 0) *  2, 3)))
    +  // Default to numExecutors * 2 (maxExecutors in the case that we are
    +  // dynamically allocating executors), with minimum of 3.
    +  private val maxNumExecutorFailures =
    +    sparkConf.getInt("spark.yarn.max.executor.failures",
    +      sparkConf.getInt("spark.yarn.max.worker.failures",
    +        math.max(
    +          3,
    +          2 * sparkConf.getInt(
    --- End diff --
    
    cool, done with some slightly more verbose `val`-names, lmk if you want me to change


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148867753
  
    Test FAILed.
    Refer to this link for build results (access rights to CI server needed): 
    https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/43869/
    Test FAILed.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by srowen <gi...@git.apache.org>.
Github user srowen commented on a diff in the pull request:

    https://github.com/apache/spark/pull/9147#discussion_r42388772
  
    --- Diff: yarn/src/main/scala/org/apache/spark/deploy/yarn/YarnAllocator.scala ---
    @@ -430,17 +430,20 @@ private[yarn] class YarnAllocator(
         for (completedContainer <- completedContainers) {
           val containerId = completedContainer.getContainerId
           val alreadyReleased = releasedContainers.remove(containerId)
    +      val hostOpt = allocatedContainerToHostMap.get(containerId)
    +      val onHostStr = hostOpt.map(host => s" on host: $host").getOrElse("")
           val exitReason = if (!alreadyReleased) {
             // Decrement the number of executors running. The next iteration of
             // the ApplicationMaster's reporting thread will take care of allocating.
             numExecutorsRunning -= 1
    -        logInfo("Completed container %s (state: %s, exit status: %s)".format(
    +        logInfo("Completed container %s%s (state: %s, exit status: %s)".format(
    --- End diff --
    
    (FWIW I don't mind `s"..." + s"..."` myself.) I'll trigger a test.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148868116
  
    Test FAILed.
    Refer to this link for build results (access rights to CI server needed): 
    https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/43871/
    Test FAILed.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148857069
  
    Test PASSed.
    Refer to this link for build results (access rights to CI server needed): 
    https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/43848/
    Test PASSed.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by andrewor14 <gi...@git.apache.org>.
Github user andrewor14 commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148946076
  
    I was thinking just checking the value of `maxNumExecutorFailures` with and without dynamic allocation being set and asserting that it's as we expect. Though this change is pretty straightforward so we can add the test separately if you prefer.
    
    retest this please


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by ryan-williams <gi...@git.apache.org>.
Github user ryan-williams commented on a diff in the pull request:

    https://github.com/apache/spark/pull/9147#discussion_r42292614
  
    --- Diff: yarn/src/main/scala/org/apache/spark/deploy/yarn/ApplicationMaster.scala ---
    @@ -62,10 +62,26 @@ private[spark] class ApplicationMaster(
         .asInstanceOf[YarnConfiguration]
       private val isClusterMode = args.userClass != null
     
    -  // Default to numExecutors * 2, with minimum of 3
    -  private val maxNumExecutorFailures = sparkConf.getInt("spark.yarn.max.executor.failures",
    -    sparkConf.getInt("spark.yarn.max.worker.failures",
    -      math.max(sparkConf.getInt("spark.executor.instances", 0) *  2, 3)))
    +  // Default to numExecutors * 2 (maxExecutors in the case that we are
    +  // dynamically allocating executors), with minimum of 3.
    +  private val maxNumExecutorFailures = {
    +    val effectiveNumExecutors =
    +      sparkConf.getInt(
    +        if (Utils.isDynamicAllocationEnabled(sparkConf)) {
    +          "spark.dynamicAllocation.maxExecutors"
    +        } else {
    +          "spark.executor.instances"
    +        },
    +        0
    +      )
    +
    +    val defaultMaxNumExecutorFailures = math.max(3, 2 * effectiveNumExecutors)
    +
    +    sparkConf.getInt(
    --- End diff --
    
    OK I addressed this so the tests will run one more time


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148868115
  
    Merged build finished. Test FAILed.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by ryan-williams <gi...@git.apache.org>.
Github user ryan-williams commented on a diff in the pull request:

    https://github.com/apache/spark/pull/9147#discussion_r42301156
  
    --- Diff: yarn/src/main/scala/org/apache/spark/deploy/yarn/ApplicationMaster.scala ---
    @@ -62,10 +62,23 @@ private[spark] class ApplicationMaster(
         .asInstanceOf[YarnConfiguration]
       private val isClusterMode = args.userClass != null
     
    -  // Default to numExecutors * 2, with minimum of 3
    -  private val maxNumExecutorFailures = sparkConf.getInt("spark.yarn.max.executor.failures",
    -    sparkConf.getInt("spark.yarn.max.worker.failures",
    -      math.max(sparkConf.getInt("spark.executor.instances", 0) *  2, 3)))
    +  // Default to numExecutors * 2 (maxExecutors in the case that we are
    +  // dynamically allocating executors), with minimum of 3.
    +  private val maxNumExecutorFailures = {
    +    val effectiveNumExecutors =
    +      sparkConf.getInt(
    +        if (Utils.isDynamicAllocationEnabled(sparkConf)) {
    +          "spark.dynamicAllocation.maxExecutors"
    +        } else {
    +          "spark.executor.instances"
    +        },
    +        0
    +      )
    +
    +    val defaultMaxNumExecutorFailures = math.max(3, 2 * effectiveNumExecutors)
    +
    --- End diff --
    
    done


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148609871
  
    Merged build finished. Test FAILed.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-149376260
  
    Merged build finished. Test FAILed.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by andrewor14 <gi...@git.apache.org>.
Github user andrewor14 commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148866742
  
    Looks good. Can you add a short test?


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148821293
  
     Merged build triggered.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148857064
  
    Merged build finished. Test PASSed.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148852144
  
    Test FAILed.
    Refer to this link for build results (access rights to CI server needed): 
    https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/43849/
    Test FAILed.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by SparkQA <gi...@git.apache.org>.
Github user SparkQA commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148851684
  
      [Test build #43849 has finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/43849/console) for   PR 9147 at commit [`aee9396`](https://github.com/apache/spark/commit/aee9396c791da134a3dbbe87d72af1fbc8f37099).
     * This patch **fails Spark unit tests**.
     * This patch merges cleanly.
     * This patch adds no public classes.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by SparkQA <gi...@git.apache.org>.
Github user SparkQA commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148609664
  
      [Test build #43831 has started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/43831/consoleFull) for   PR 9147 at commit [`46633e7`](https://github.com/apache/spark/commit/46633e74d0f732a293e8ca4cbfc6a5d683d6cc99).


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by srowen <gi...@git.apache.org>.
Github user srowen commented on a diff in the pull request:

    https://github.com/apache/spark/pull/9147#discussion_r42320955
  
    --- Diff: yarn/src/main/scala/org/apache/spark/deploy/yarn/YarnAllocator.scala ---
    @@ -430,17 +430,20 @@ private[yarn] class YarnAllocator(
         for (completedContainer <- completedContainers) {
           val containerId = completedContainer.getContainerId
           val alreadyReleased = releasedContainers.remove(containerId)
    +      val hostOpt = allocatedContainerToHostMap.get(containerId)
    +      val onHostStr = hostOpt.map(host => s" on host: $host").getOrElse("")
           val exitReason = if (!alreadyReleased) {
             // Decrement the number of executors running. The next iteration of
             // the ApplicationMaster's reporting thread will take care of allocating.
             numExecutorsRunning -= 1
    -        logInfo("Completed container %s (state: %s, exit status: %s)".format(
    +        logInfo("Completed container %s%s (state: %s, exit status: %s)".format(
    --- End diff --
    
    You're welcome to convert this to string interpolation now. (Just if you find you need to make other changes; not worth it only for this.)


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by ryan-williams <gi...@git.apache.org>.
Github user ryan-williams commented on a diff in the pull request:

    https://github.com/apache/spark/pull/9147#discussion_r42301088
  
    --- Diff: yarn/src/main/scala/org/apache/spark/deploy/yarn/ApplicationMaster.scala ---
    @@ -62,10 +62,23 @@ private[spark] class ApplicationMaster(
         .asInstanceOf[YarnConfiguration]
       private val isClusterMode = args.userClass != null
     
    -  // Default to numExecutors * 2, with minimum of 3
    -  private val maxNumExecutorFailures = sparkConf.getInt("spark.yarn.max.executor.failures",
    -    sparkConf.getInt("spark.yarn.max.worker.failures",
    -      math.max(sparkConf.getInt("spark.executor.instances", 0) *  2, 3)))
    +  // Default to numExecutors * 2 (maxExecutors in the case that we are
    +  // dynamically allocating executors), with minimum of 3.
    --- End diff --
    
    done (a version of that anyway)


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by vanzin <gi...@git.apache.org>.
Github user vanzin commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148849042
  
    yeah. don't worry about it, no need to run the tests again.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148842488
  
    Merged build started.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request: [SPARK-11120] Allow sane default number of exe...

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the pull request:

    https://github.com/apache/spark/pull/9147#issuecomment-148842465
  
     Merged build triggered.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org