You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@spark.apache.org by ad...@apache.org on 2014/07/15 18:13:47 UTC
git commit: Reformat multi-line closure argument.
Repository: spark
Updated Branches:
refs/heads/master 04b01bb10 -> cb09e93c1
Reformat multi-line closure argument.
Author: William Benton <wi...@redhat.com>
Closes #1419 from willb/reformat-2486 and squashes the following commits:
2676231 [William Benton] Reformat multi-line closure argument.
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/cb09e93c
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/cb09e93c
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/cb09e93c
Branch: refs/heads/master
Commit: cb09e93c1d7ef9c8f0a1abe4e659783c74993a4e
Parents: 04b01bb
Author: William Benton <wi...@redhat.com>
Authored: Tue Jul 15 09:13:39 2014 -0700
Committer: Aaron Davidson <aa...@databricks.com>
Committed: Tue Jul 15 09:13:39 2014 -0700
----------------------------------------------------------------------
core/src/main/scala/org/apache/spark/util/Utils.scala | 5 +++--
1 file changed, 3 insertions(+), 2 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/spark/blob/cb09e93c/core/src/main/scala/org/apache/spark/util/Utils.scala
----------------------------------------------------------------------
diff --git a/core/src/main/scala/org/apache/spark/util/Utils.scala b/core/src/main/scala/org/apache/spark/util/Utils.scala
index d72c97b..10c33d6 100644
--- a/core/src/main/scala/org/apache/spark/util/Utils.scala
+++ b/core/src/main/scala/org/apache/spark/util/Utils.scala
@@ -809,11 +809,12 @@ private[spark] object Utils extends Logging {
*/
def getCallSite: CallSite = {
val trace = Thread.currentThread.getStackTrace()
- .filterNot((ste:StackTraceElement) =>
+ .filterNot { ste:StackTraceElement =>
// When running under some profilers, the current stack trace might contain some bogus
// frames. This is intended to ensure that we don't crash in these situations by
// ignoring any frames that we can't examine.
- (ste == null || ste.getMethodName == null || ste.getMethodName.contains("getStackTrace")))
+ (ste == null || ste.getMethodName == null || ste.getMethodName.contains("getStackTrace"))
+ }
// Keep crawling up the stack trace until we find the first function not inside of the spark
// package. We track the last (shallowest) contiguous Spark method. This might be an RDD