You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@spark.apache.org by pw...@apache.org on 2013/12/25 01:35:50 UTC

[19/20] git commit: Responded to Reynold's style comments

Responded to Reynold's style comments


Project: http://git-wip-us.apache.org/repos/asf/incubator-spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/incubator-spark/commit/1efe3adf
Tree: http://git-wip-us.apache.org/repos/asf/incubator-spark/tree/1efe3adf
Diff: http://git-wip-us.apache.org/repos/asf/incubator-spark/diff/1efe3adf

Branch: refs/heads/master
Commit: 1efe3adf560d207f9106ffd4e15934e422adb636
Parents: b7bfae1
Author: Kay Ousterhout <ka...@gmail.com>
Authored: Tue Dec 24 14:18:39 2013 -0800
Committer: Kay Ousterhout <ka...@gmail.com>
Committed: Tue Dec 24 14:18:39 2013 -0800

----------------------------------------------------------------------
 .../scala/org/apache/spark/scheduler/TaskSchedulerImpl.scala  | 7 ++++---
 .../scala/org/apache/spark/scheduler/TaskSetManager.scala     | 4 ++--
 .../scala/org/apache/spark/scheduler/local/LocalBackend.scala | 2 +-
 3 files changed, 7 insertions(+), 6 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/incubator-spark/blob/1efe3adf/core/src/main/scala/org/apache/spark/scheduler/TaskSchedulerImpl.scala
----------------------------------------------------------------------
diff --git a/core/src/main/scala/org/apache/spark/scheduler/TaskSchedulerImpl.scala b/core/src/main/scala/org/apache/spark/scheduler/TaskSchedulerImpl.scala
index 7409168..dbac6b9 100644
--- a/core/src/main/scala/org/apache/spark/scheduler/TaskSchedulerImpl.scala
+++ b/core/src/main/scala/org/apache/spark/scheduler/TaskSchedulerImpl.scala
@@ -46,9 +46,10 @@ import org.apache.spark.scheduler.SchedulingMode.SchedulingMode
  * we are holding a lock on ourselves.
  */
 private[spark] class TaskSchedulerImpl(
-  val sc: SparkContext,
-  val maxTaskFailures : Int = System.getProperty("spark.task.maxFailures", "4").toInt,
-  isLocal: Boolean = false) extends TaskScheduler with Logging {
+    val sc: SparkContext,
+    val maxTaskFailures : Int = System.getProperty("spark.task.maxFailures", "4").toInt,
+    isLocal: Boolean = false)
+  extends TaskScheduler with Logging {
 
   // How often to check for speculative tasks
   val SPECULATION_INTERVAL = System.getProperty("spark.speculation.interval", "100").toLong

http://git-wip-us.apache.org/repos/asf/incubator-spark/blob/1efe3adf/core/src/main/scala/org/apache/spark/scheduler/TaskSetManager.scala
----------------------------------------------------------------------
diff --git a/core/src/main/scala/org/apache/spark/scheduler/TaskSetManager.scala b/core/src/main/scala/org/apache/spark/scheduler/TaskSetManager.scala
index aa3fb0b..c676e73 100644
--- a/core/src/main/scala/org/apache/spark/scheduler/TaskSetManager.scala
+++ b/core/src/main/scala/org/apache/spark/scheduler/TaskSetManager.scala
@@ -17,6 +17,7 @@
 
 package org.apache.spark.scheduler
 
+import java.io.NotSerializableException
 import java.util.Arrays
 
 import scala.collection.mutable.ArrayBuffer
@@ -28,8 +29,7 @@ import scala.math.min
 import org.apache.spark.{ExceptionFailure, FetchFailed, Logging, Resubmitted, SparkEnv,
   Success, TaskEndReason, TaskKilled, TaskResultLost, TaskState}
 import org.apache.spark.TaskState.TaskState
-import org.apache.spark.util.{SystemClock, Clock}
-import java.io.NotSerializableException
+import org.apache.spark.util.{Clock, SystemClock}
 
 
 /**

http://git-wip-us.apache.org/repos/asf/incubator-spark/blob/1efe3adf/core/src/main/scala/org/apache/spark/scheduler/local/LocalBackend.scala
----------------------------------------------------------------------
diff --git a/core/src/main/scala/org/apache/spark/scheduler/local/LocalBackend.scala b/core/src/main/scala/org/apache/spark/scheduler/local/LocalBackend.scala
index 69c1c04..4edc6a0 100644
--- a/core/src/main/scala/org/apache/spark/scheduler/local/LocalBackend.scala
+++ b/core/src/main/scala/org/apache/spark/scheduler/local/LocalBackend.scala
@@ -93,7 +93,7 @@ private[spark] class LocalBackend(scheduler: TaskSchedulerImpl, val totalCores:
   }
 
   override def reviveOffers() {
-   localActor ! ReviveOffers
+    localActor ! ReviveOffers
   }
 
   override def defaultParallelism() = totalCores