You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@spark.apache.org by do...@apache.org on 2019/02/09 03:51:21 UTC
[spark] branch branch-2.3 updated: Revert
"[SPARK-26082][MESOS][FOLLOWUP] Add UT on fetcher cache option on
MesosClusterScheduler"
This is an automated email from the ASF dual-hosted git repository.
dongjoon pushed a commit to branch branch-2.3
in repository https://gitbox.apache.org/repos/asf/spark.git
The following commit(s) were added to refs/heads/branch-2.3 by this push:
new 97f8ed4 Revert "[SPARK-26082][MESOS][FOLLOWUP] Add UT on fetcher cache option on MesosClusterScheduler"
97f8ed4 is described below
commit 97f8ed484969f19083c20786185751162baccde1
Author: Dongjoon Hyun <dh...@apple.com>
AuthorDate: Fri Feb 8 19:50:56 2019 -0800
Revert "[SPARK-26082][MESOS][FOLLOWUP] Add UT on fetcher cache option on MesosClusterScheduler"
This reverts commit 3abf45d6f3568d801460f558cb66f31260413a12.
---
.../cluster/mesos/MesosClusterSchedulerSuite.scala | 48 ----------------------
1 file changed, 48 deletions(-)
diff --git a/resource-managers/mesos/src/test/scala/org/apache/spark/scheduler/cluster/mesos/MesosClusterSchedulerSuite.scala b/resource-managers/mesos/src/test/scala/org/apache/spark/scheduler/cluster/mesos/MesosClusterSchedulerSuite.scala
index bc2e6d4..e534b9d 100644
--- a/resource-managers/mesos/src/test/scala/org/apache/spark/scheduler/cluster/mesos/MesosClusterSchedulerSuite.scala
+++ b/resource-managers/mesos/src/test/scala/org/apache/spark/scheduler/cluster/mesos/MesosClusterSchedulerSuite.scala
@@ -254,54 +254,6 @@ class MesosClusterSchedulerSuite extends SparkFunSuite with LocalSparkContext wi
assert(networkInfos.get(0).getLabels.getLabels(1).getValue == "val2")
}
- test("supports setting fetcher cache") {
- setScheduler(Map("spark.mesos.fetcherCache.enable" -> "true"))
-
- val mem = 1000
- val cpu = 1
-
- val response = scheduler.submitDriver(
- new MesosDriverDescription("d1", "jar", mem, cpu, true,
- command,
- Map("spark.mesos.executor.home" -> "test",
- "spark.app.name" -> "test"),
- "s1",
- new Date()))
-
- assert(response.success)
-
- val offer = Utils.createOffer("o1", "s1", mem, cpu)
- scheduler.resourceOffers(driver, List(offer).asJava)
-
- val launchedTasks = Utils.verifyTaskLaunched(driver, "o1")
- val uris = launchedTasks.head.getCommand.getUrisList
- assert(uris.stream().allMatch(_.getCache))
- }
-
- test("supports disabling fetcher cache") {
- setScheduler(Map("spark.mesos.fetcherCache.enable" -> "false"))
-
- val mem = 1000
- val cpu = 1
-
- val response = scheduler.submitDriver(
- new MesosDriverDescription("d1", "jar", mem, cpu, true,
- command,
- Map("spark.mesos.executor.home" -> "test",
- "spark.app.name" -> "test"),
- "s1",
- new Date()))
-
- assert(response.success)
-
- val offer = Utils.createOffer("o1", "s1", mem, cpu)
- scheduler.resourceOffers(driver, List(offer).asJava)
-
- val launchedTasks = Utils.verifyTaskLaunched(driver, "o1")
- val uris = launchedTasks.head.getCommand.getUrisList
- assert(uris.stream().allMatch(!_.getCache))
- }
-
test("accept/decline offers with driver constraints") {
setScheduler()
---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@spark.apache.org
For additional commands, e-mail: commits-help@spark.apache.org