You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@spark.apache.org by pw...@apache.org on 2013/12/25 01:35:50 UTC
[19/20] git commit: Responded to Reynold's style comments
Responded to Reynold's style comments
Project: http://git-wip-us.apache.org/repos/asf/incubator-spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/incubator-spark/commit/1efe3adf
Tree: http://git-wip-us.apache.org/repos/asf/incubator-spark/tree/1efe3adf
Diff: http://git-wip-us.apache.org/repos/asf/incubator-spark/diff/1efe3adf
Branch: refs/heads/master
Commit: 1efe3adf560d207f9106ffd4e15934e422adb636
Parents: b7bfae1
Author: Kay Ousterhout <ka...@gmail.com>
Authored: Tue Dec 24 14:18:39 2013 -0800
Committer: Kay Ousterhout <ka...@gmail.com>
Committed: Tue Dec 24 14:18:39 2013 -0800
----------------------------------------------------------------------
.../scala/org/apache/spark/scheduler/TaskSchedulerImpl.scala | 7 ++++---
.../scala/org/apache/spark/scheduler/TaskSetManager.scala | 4 ++--
.../scala/org/apache/spark/scheduler/local/LocalBackend.scala | 2 +-
3 files changed, 7 insertions(+), 6 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/incubator-spark/blob/1efe3adf/core/src/main/scala/org/apache/spark/scheduler/TaskSchedulerImpl.scala
----------------------------------------------------------------------
diff --git a/core/src/main/scala/org/apache/spark/scheduler/TaskSchedulerImpl.scala b/core/src/main/scala/org/apache/spark/scheduler/TaskSchedulerImpl.scala
index 7409168..dbac6b9 100644
--- a/core/src/main/scala/org/apache/spark/scheduler/TaskSchedulerImpl.scala
+++ b/core/src/main/scala/org/apache/spark/scheduler/TaskSchedulerImpl.scala
@@ -46,9 +46,10 @@ import org.apache.spark.scheduler.SchedulingMode.SchedulingMode
* we are holding a lock on ourselves.
*/
private[spark] class TaskSchedulerImpl(
- val sc: SparkContext,
- val maxTaskFailures : Int = System.getProperty("spark.task.maxFailures", "4").toInt,
- isLocal: Boolean = false) extends TaskScheduler with Logging {
+ val sc: SparkContext,
+ val maxTaskFailures : Int = System.getProperty("spark.task.maxFailures", "4").toInt,
+ isLocal: Boolean = false)
+ extends TaskScheduler with Logging {
// How often to check for speculative tasks
val SPECULATION_INTERVAL = System.getProperty("spark.speculation.interval", "100").toLong
http://git-wip-us.apache.org/repos/asf/incubator-spark/blob/1efe3adf/core/src/main/scala/org/apache/spark/scheduler/TaskSetManager.scala
----------------------------------------------------------------------
diff --git a/core/src/main/scala/org/apache/spark/scheduler/TaskSetManager.scala b/core/src/main/scala/org/apache/spark/scheduler/TaskSetManager.scala
index aa3fb0b..c676e73 100644
--- a/core/src/main/scala/org/apache/spark/scheduler/TaskSetManager.scala
+++ b/core/src/main/scala/org/apache/spark/scheduler/TaskSetManager.scala
@@ -17,6 +17,7 @@
package org.apache.spark.scheduler
+import java.io.NotSerializableException
import java.util.Arrays
import scala.collection.mutable.ArrayBuffer
@@ -28,8 +29,7 @@ import scala.math.min
import org.apache.spark.{ExceptionFailure, FetchFailed, Logging, Resubmitted, SparkEnv,
Success, TaskEndReason, TaskKilled, TaskResultLost, TaskState}
import org.apache.spark.TaskState.TaskState
-import org.apache.spark.util.{SystemClock, Clock}
-import java.io.NotSerializableException
+import org.apache.spark.util.{Clock, SystemClock}
/**
http://git-wip-us.apache.org/repos/asf/incubator-spark/blob/1efe3adf/core/src/main/scala/org/apache/spark/scheduler/local/LocalBackend.scala
----------------------------------------------------------------------
diff --git a/core/src/main/scala/org/apache/spark/scheduler/local/LocalBackend.scala b/core/src/main/scala/org/apache/spark/scheduler/local/LocalBackend.scala
index 69c1c04..4edc6a0 100644
--- a/core/src/main/scala/org/apache/spark/scheduler/local/LocalBackend.scala
+++ b/core/src/main/scala/org/apache/spark/scheduler/local/LocalBackend.scala
@@ -93,7 +93,7 @@ private[spark] class LocalBackend(scheduler: TaskSchedulerImpl, val totalCores:
}
override def reviveOffers() {
- localActor ! ReviveOffers
+ localActor ! ReviveOffers
}
override def defaultParallelism() = totalCores