You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@spark.apache.org by rx...@apache.org on 2016/11/05 06:34:47 UTC
spark git commit: [SPARK-18189] [SQL] [Followup] Move test from
ReplSuite to prevent java.lang.ClassCircularityError
Repository: spark
Updated Branches:
refs/heads/master 0e3312ee7 -> 0f7c9e84e
[SPARK-18189] [SQL] [Followup] Move test from ReplSuite to prevent java.lang.ClassCircularityError
closes #15774
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/0f7c9e84
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/0f7c9e84
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/0f7c9e84
Branch: refs/heads/master
Commit: 0f7c9e84e0d00813bf56712097677add5657f19f
Parents: 0e3312e
Author: Reynold Xin <rx...@databricks.com>
Authored: Fri Nov 4 23:34:29 2016 -0700
Committer: Reynold Xin <rx...@databricks.com>
Committed: Fri Nov 4 23:34:29 2016 -0700
----------------------------------------------------------------------
.../scala/org/apache/spark/repl/ReplSuite.scala | 17 -----------------
.../scala/org/apache/spark/sql/DatasetSuite.scala | 12 ++++++++++++
2 files changed, 12 insertions(+), 17 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/spark/blob/0f7c9e84/repl/scala-2.11/src/test/scala/org/apache/spark/repl/ReplSuite.scala
----------------------------------------------------------------------
diff --git a/repl/scala-2.11/src/test/scala/org/apache/spark/repl/ReplSuite.scala b/repl/scala-2.11/src/test/scala/org/apache/spark/repl/ReplSuite.scala
index 96d2dfc..9262e93 100644
--- a/repl/scala-2.11/src/test/scala/org/apache/spark/repl/ReplSuite.scala
+++ b/repl/scala-2.11/src/test/scala/org/apache/spark/repl/ReplSuite.scala
@@ -473,21 +473,4 @@ class ReplSuite extends SparkFunSuite {
assertDoesNotContain("AssertionError", output)
assertDoesNotContain("Exception", output)
}
-
- test("SPARK-18189: Fix serialization issue in KeyValueGroupedDataset") {
- val resultValue = 12345
- val output = runInterpreter("local",
- s"""
- |val keyValueGrouped = Seq((1, 2), (3, 4)).toDS().groupByKey(_._1)
- |val mapGroups = keyValueGrouped.mapGroups((k, v) => (k, 1))
- |val broadcasted = sc.broadcast($resultValue)
- |
- |// Using broadcast triggers serialization issue in KeyValueGroupedDataset
- |val dataset = mapGroups.map(_ => broadcasted.value)
- |dataset.collect()
- """.stripMargin)
- assertDoesNotContain("error:", output)
- assertDoesNotContain("Exception", output)
- assertContains(s": Array[Int] = Array($resultValue, $resultValue)", output)
- }
}
http://git-wip-us.apache.org/repos/asf/spark/blob/0f7c9e84/sql/core/src/test/scala/org/apache/spark/sql/DatasetSuite.scala
----------------------------------------------------------------------
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/DatasetSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/DatasetSuite.scala
index 55f0487..6fa7b04 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/DatasetSuite.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/DatasetSuite.scala
@@ -923,6 +923,18 @@ class DatasetSuite extends QueryTest with SharedSQLContext {
.groupByKey(_.a).flatMapGroups { case (x, iter) => List[Int]() })
}
+ test("SPARK-18189: Fix serialization issue in KeyValueGroupedDataset") {
+ val resultValue = 12345
+ val keyValueGrouped = Seq((1, 2), (3, 4)).toDS().groupByKey(_._1)
+ val mapGroups = keyValueGrouped.mapGroups((k, v) => (k, 1))
+ val broadcasted = spark.sparkContext.broadcast(resultValue)
+
+ // Using broadcast triggers serialization issue in KeyValueGroupedDataset
+ val dataset = mapGroups.map(_ => broadcasted.value)
+
+ assert(dataset.collect() sameElements Array(resultValue, resultValue))
+ }
+
Seq(true, false).foreach { eager =>
def testCheckpointing(testName: String)(f: => Unit): Unit = {
test(s"Dataset.checkpoint() - $testName (eager = $eager)") {
---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@spark.apache.org
For additional commands, e-mail: commits-help@spark.apache.org