You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@spark.apache.org by ad...@apache.org on 2014/07/15 18:13:47 UTC

git commit: Reformat multi-line closure argument.

Repository: spark
Updated Branches:
  refs/heads/master 04b01bb10 -> cb09e93c1


Reformat multi-line closure argument.

Author: William Benton <wi...@redhat.com>

Closes #1419 from willb/reformat-2486 and squashes the following commits:

2676231 [William Benton] Reformat multi-line closure argument.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/cb09e93c
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/cb09e93c
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/cb09e93c

Branch: refs/heads/master
Commit: cb09e93c1d7ef9c8f0a1abe4e659783c74993a4e
Parents: 04b01bb
Author: William Benton <wi...@redhat.com>
Authored: Tue Jul 15 09:13:39 2014 -0700
Committer: Aaron Davidson <aa...@databricks.com>
Committed: Tue Jul 15 09:13:39 2014 -0700

----------------------------------------------------------------------
 core/src/main/scala/org/apache/spark/util/Utils.scala | 5 +++--
 1 file changed, 3 insertions(+), 2 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/spark/blob/cb09e93c/core/src/main/scala/org/apache/spark/util/Utils.scala
----------------------------------------------------------------------
diff --git a/core/src/main/scala/org/apache/spark/util/Utils.scala b/core/src/main/scala/org/apache/spark/util/Utils.scala
index d72c97b..10c33d6 100644
--- a/core/src/main/scala/org/apache/spark/util/Utils.scala
+++ b/core/src/main/scala/org/apache/spark/util/Utils.scala
@@ -809,11 +809,12 @@ private[spark] object Utils extends Logging {
    */
   def getCallSite: CallSite = {
     val trace = Thread.currentThread.getStackTrace()
-      .filterNot((ste:StackTraceElement) => 
+      .filterNot { ste:StackTraceElement => 
         // When running under some profilers, the current stack trace might contain some bogus 
         // frames. This is intended to ensure that we don't crash in these situations by
         // ignoring any frames that we can't examine.
-        (ste == null || ste.getMethodName == null || ste.getMethodName.contains("getStackTrace")))
+        (ste == null || ste.getMethodName == null || ste.getMethodName.contains("getStackTrace"))
+      }
 
     // Keep crawling up the stack trace until we find the first function not inside of the spark
     // package. We track the last (shallowest) contiguous Spark method. This might be an RDD