You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@spark.apache.org by rx...@apache.org on 2016/04/02 09:00:36 UTC
spark git commit: [HOTFIX] Fix compilation break.
Repository: spark
Updated Branches:
refs/heads/master d7982a3a9 -> 67d753516
[HOTFIX] Fix compilation break.
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/67d75351
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/67d75351
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/67d75351
Branch: refs/heads/master
Commit: 67d753516da9b6318cd4001bb7ae91703aaf098d
Parents: d7982a3
Author: Reynold Xin <rx...@databricks.com>
Authored: Sat Apr 2 00:00:19 2016 -0700
Committer: Reynold Xin <rx...@databricks.com>
Committed: Sat Apr 2 00:00:19 2016 -0700
----------------------------------------------------------------------
.../spark/sql/execution/streaming/HDFSMetadataLogSuite.scala | 1 +
.../spark/sql/streaming/StreamingAggregationSuite.scala | 8 +++-----
2 files changed, 4 insertions(+), 5 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/spark/blob/67d75351/sql/core/src/test/scala/org/apache/spark/sql/execution/streaming/HDFSMetadataLogSuite.scala
----------------------------------------------------------------------
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/streaming/HDFSMetadataLogSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/streaming/HDFSMetadataLogSuite.scala
index d5db9db..1328142 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/streaming/HDFSMetadataLogSuite.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/streaming/HDFSMetadataLogSuite.scala
@@ -21,6 +21,7 @@ import java.io.{File, FileNotFoundException, IOException}
import java.net.URI
import java.util.ConcurrentModificationException
+import scala.language.implicitConversions
import scala.util.Random
import org.apache.hadoop.conf.Configuration
http://git-wip-us.apache.org/repos/asf/spark/blob/67d75351/sql/core/src/test/scala/org/apache/spark/sql/streaming/StreamingAggregationSuite.scala
----------------------------------------------------------------------
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/streaming/StreamingAggregationSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/streaming/StreamingAggregationSuite.scala
index b63ce89..3af7c01 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/streaming/StreamingAggregationSuite.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/streaming/StreamingAggregationSuite.scala
@@ -18,8 +18,9 @@
package org.apache.spark.sql.streaming
import org.apache.spark.SparkException
-import org.apache.spark.sql.{Encoder, StreamTest, SumOf, TypedColumn}
+import org.apache.spark.sql.StreamTest
import org.apache.spark.sql.execution.streaming._
+import org.apache.spark.sql.expressions.scala.typed
import org.apache.spark.sql.functions._
import org.apache.spark.sql.test.SharedSQLContext
@@ -118,11 +119,8 @@ class StreamingAggregationSuite extends StreamTest with SharedSQLContext {
}
test("typed aggregators") {
- def sum[I, N : Numeric : Encoder](f: I => N): TypedColumn[I, N] =
- new SumOf(f).toColumn
-
val inputData = MemoryStream[(String, Int)]
- val aggregated = inputData.toDS().groupByKey(_._1).agg(sum(_._2))
+ val aggregated = inputData.toDS().groupByKey(_._1).agg(typed.sumLong(_._2))
testStream(aggregated)(
AddData(inputData, ("a", 10), ("a", 20), ("b", 1), ("b", 2), ("c", 1)),
---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@spark.apache.org
For additional commands, e-mail: commits-help@spark.apache.org