You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@spark.apache.org by rx...@apache.org on 2016/04/02 09:00:36 UTC

spark git commit: [HOTFIX] Fix compilation break.

Repository: spark
Updated Branches:
  refs/heads/master d7982a3a9 -> 67d753516


[HOTFIX] Fix compilation break.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/67d75351
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/67d75351
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/67d75351

Branch: refs/heads/master
Commit: 67d753516da9b6318cd4001bb7ae91703aaf098d
Parents: d7982a3
Author: Reynold Xin <rx...@databricks.com>
Authored: Sat Apr 2 00:00:19 2016 -0700
Committer: Reynold Xin <rx...@databricks.com>
Committed: Sat Apr 2 00:00:19 2016 -0700

----------------------------------------------------------------------
 .../spark/sql/execution/streaming/HDFSMetadataLogSuite.scala | 1 +
 .../spark/sql/streaming/StreamingAggregationSuite.scala      | 8 +++-----
 2 files changed, 4 insertions(+), 5 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/spark/blob/67d75351/sql/core/src/test/scala/org/apache/spark/sql/execution/streaming/HDFSMetadataLogSuite.scala
----------------------------------------------------------------------
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/streaming/HDFSMetadataLogSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/streaming/HDFSMetadataLogSuite.scala
index d5db9db..1328142 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/streaming/HDFSMetadataLogSuite.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/streaming/HDFSMetadataLogSuite.scala
@@ -21,6 +21,7 @@ import java.io.{File, FileNotFoundException, IOException}
 import java.net.URI
 import java.util.ConcurrentModificationException
 
+import scala.language.implicitConversions
 import scala.util.Random
 
 import org.apache.hadoop.conf.Configuration

http://git-wip-us.apache.org/repos/asf/spark/blob/67d75351/sql/core/src/test/scala/org/apache/spark/sql/streaming/StreamingAggregationSuite.scala
----------------------------------------------------------------------
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/streaming/StreamingAggregationSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/streaming/StreamingAggregationSuite.scala
index b63ce89..3af7c01 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/streaming/StreamingAggregationSuite.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/streaming/StreamingAggregationSuite.scala
@@ -18,8 +18,9 @@
 package org.apache.spark.sql.streaming
 
 import org.apache.spark.SparkException
-import org.apache.spark.sql.{Encoder, StreamTest, SumOf, TypedColumn}
+import org.apache.spark.sql.StreamTest
 import org.apache.spark.sql.execution.streaming._
+import org.apache.spark.sql.expressions.scala.typed
 import org.apache.spark.sql.functions._
 import org.apache.spark.sql.test.SharedSQLContext
 
@@ -118,11 +119,8 @@ class StreamingAggregationSuite extends StreamTest with SharedSQLContext {
   }
 
   test("typed aggregators") {
-    def sum[I, N : Numeric : Encoder](f: I => N): TypedColumn[I, N] =
-      new SumOf(f).toColumn
-
     val inputData = MemoryStream[(String, Int)]
-    val aggregated = inputData.toDS().groupByKey(_._1).agg(sum(_._2))
+    val aggregated = inputData.toDS().groupByKey(_._1).agg(typed.sumLong(_._2))
 
     testStream(aggregated)(
       AddData(inputData, ("a", 10), ("a", 20), ("b", 1), ("b", 2), ("c", 1)),


---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@spark.apache.org
For additional commands, e-mail: commits-help@spark.apache.org