You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Jungtaek Lim (Jira)" <ji...@apache.org> on 2019/09/20 03:03:00 UTC

[jira] [Resolved] (SPARK-29132) Flaky test: org.apache.spark.sql.execution.SQLJsonProtocolSuite.SparkListenerSQLExecutionEnd backward compatibility

     [ https://issues.apache.org/jira/browse/SPARK-29132?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Jungtaek Lim resolved SPARK-29132.
----------------------------------
    Resolution: Invalid

Just affected by problematic commit (had been reverted) [https://github.com/apache/spark/commit/850833fa177ec1f265e143fc383e40ec2c8341a6]

See more: [https://github.com/apache/spark/pull/25790#issuecomment-531517800]

 

> Flaky test: org.apache.spark.sql.execution.SQLJsonProtocolSuite.SparkListenerSQLExecutionEnd backward compatibility
> -------------------------------------------------------------------------------------------------------------------
>
>                 Key: SPARK-29132
>                 URL: https://issues.apache.org/jira/browse/SPARK-29132
>             Project: Spark
>          Issue Type: Bug
>          Components: SQL, Tests
>    Affects Versions: 3.0.0
>            Reporter: Jungtaek Lim
>            Priority: Major
>
> [https://amplab.cs.berkeley.edu/jenkins/view/Spark%20QA%20Test/job/spark-master-test-maven-hadoop-2.7-jdk-11-ubuntu-testing/1549/testReport/]
> The stack trace is somewhat odd, so it could be innocent - affected by other problematic test.
> {code:java}
> java.lang.IllegalStateException: Shutdown hooks cannot be modified during shutdown.
>       at org.apache.spark.util.SparkShutdownHookManager.add(ShutdownHookManager.scala:195)
>       at org.apache.spark.util.ShutdownHookManager$.addShutdownHook(ShutdownHookManager.scala:153)
>       at org.apache.spark.storage.DiskBlockManager.addShutdownHook(DiskBlockManager.scala:157)
>       at org.apache.spark.storage.DiskBlockManager.<init>(DiskBlockManager.scala:54)
>       at org.apache.spark.storage.BlockManager.<init>(BlockManager.scala:148)
>       at org.apache.spark.SparkEnv$.create(SparkEnv.scala:370)
>       at org.apache.spark.SparkEnv$.createDriverEnv(SparkEnv.scala:185)
>       at org.apache.spark.SparkContext.createSparkEnv(SparkContext.scala:273)
>       at org.apache.spark.SparkContext.<init>(SparkContext.scala:456)
>       at org.apache.spark.SparkContext.<init>(SparkContext.scala:129)
>       at org.apache.spark.sql.test.TestSparkSession.<init>(TestSQLContext.scala:30)
>       at org.apache.spark.sql.test.TestSparkSession.<init>(TestSQLContext.scala:34)
>       at org.apache.spark.sql.execution.SQLJsonProtocolSuite.$anonfun$new$2(SQLJsonProtocolSuite.scala:56)
>       at org.scalatest.OutcomeOf.outcomeOf(OutcomeOf.scala:85)
>       at org.scalatest.OutcomeOf.outcomeOf$(OutcomeOf.scala:83)
>       at org.scalatest.OutcomeOf$.outcomeOf(OutcomeOf.scala:104)
>       at org.scalatest.Transformer.apply(Transformer.scala:22)
>       at org.scalatest.Transformer.apply(Transformer.scala:20)
>       at org.scalatest.FunSuiteLike$$anon$1.apply(FunSuiteLike.scala:186)
>       at org.apache.spark.SparkFunSuite.withFixture(SparkFunSuite.scala:149)
>       at org.scalatest.FunSuiteLike.invokeWithFixture$1(FunSuiteLike.scala:184)
>       at org.scalatest.FunSuiteLike.$anonfun$runTest$1(FunSuiteLike.scala:196)
>       at org.scalatest.SuperEngine.runTestImpl(Engine.scala:289)
>       at org.scalatest.FunSuiteLike.runTest(FunSuiteLike.scala:196)
>       at org.scalatest.FunSuiteLike.runTest$(FunSuiteLike.scala:178)
>       at org.apache.spark.SparkFunSuite.org$scalatest$BeforeAndAfterEach$$super$runTest(SparkFunSuite.scala:56)
>       at org.scalatest.BeforeAndAfterEach.runTest(BeforeAndAfterEach.scala:221)
>       at org.scalatest.BeforeAndAfterEach.runTest$(BeforeAndAfterEach.scala:214)
>       at org.apache.spark.SparkFunSuite.runTest(SparkFunSuite.scala:56)
>       at org.scalatest.FunSuiteLike.$anonfun$runTests$1(FunSuiteLike.scala:229)
>       at org.scalatest.SuperEngine.$anonfun$runTestsInBranch$1(Engine.scala:396)
>       at scala.collection.immutable.List.foreach(List.scala:392)
>       at org.scalatest.SuperEngine.traverseSubNodes$1(Engine.scala:384)
>       at org.scalatest.SuperEngine.runTestsInBranch(Engine.scala:379)
>       at org.scalatest.SuperEngine.runTestsImpl(Engine.scala:461)
>       at org.scalatest.FunSuiteLike.runTests(FunSuiteLike.scala:229)
>       at org.scalatest.FunSuiteLike.runTests$(FunSuiteLike.scala:228)
>       at org.scalatest.FunSuite.runTests(FunSuite.scala:1560)
>       at org.scalatest.Suite.run(Suite.scala:1147)
>       at org.scalatest.Suite.run$(Suite.scala:1129)
>       at org.scalatest.FunSuite.org$scalatest$FunSuiteLike$$super$run(FunSuite.scala:1560)
>       at org.scalatest.FunSuiteLike.$anonfun$run$1(FunSuiteLike.scala:233)
>       at org.scalatest.SuperEngine.runImpl(Engine.scala:521)
>       at org.scalatest.FunSuiteLike.run(FunSuiteLike.scala:233)
>       at org.scalatest.FunSuiteLike.run$(FunSuiteLike.scala:232)
>       at org.apache.spark.SparkFunSuite.org$scalatest$BeforeAndAfterAll$$super$run(SparkFunSuite.scala:56)
>       at org.scalatest.BeforeAndAfterAll.liftedTree1$1(BeforeAndAfterAll.scala:213)
>       at org.scalatest.BeforeAndAfterAll.run(BeforeAndAfterAll.scala:210)
>       at org.scalatest.BeforeAndAfterAll.run$(BeforeAndAfterAll.scala:208)
>       at org.apache.spark.SparkFunSuite.run(SparkFunSuite.scala:56)
>       at org.scalatest.Suite.callExecuteOnSuite$1(Suite.scala:1210)
>       at org.scalatest.Suite.$anonfun$runNestedSuites$1(Suite.scala:1257)
>       at scala.collection.IndexedSeqOptimized.foreach(IndexedSeqOptimized.scala:36)
>       at scala.collection.IndexedSeqOptimized.foreach$(IndexedSeqOptimized.scala:33)
>       at scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:198)
>       at org.scalatest.Suite.runNestedSuites(Suite.scala:1255)
>       at org.scalatest.Suite.runNestedSuites$(Suite.scala:1189)
>       at org.scalatest.tools.DiscoverySuite.runNestedSuites(DiscoverySuite.scala:30)
>       at org.scalatest.Suite.run(Suite.scala:1144)
>       at org.scalatest.Suite.run$(Suite.scala:1129)
>       at org.scalatest.tools.DiscoverySuite.run(DiscoverySuite.scala:30)
>       at org.scalatest.tools.SuiteRunner.run(SuiteRunner.scala:45)
>       at org.scalatest.tools.Runner$.$anonfun$doRunRunRunDaDoRunRun$13(Runner.scala:1346)
>       at org.scalatest.tools.Runner$.$anonfun$doRunRunRunDaDoRunRun$13$adapted(Runner.scala:1340)
>       at scala.collection.immutable.List.foreach(List.scala:392)
>       at org.scalatest.tools.Runner$.doRunRunRunDaDoRunRun(Runner.scala:1340)
>       at org.scalatest.tools.Runner$.$anonfun$runOptionallyWithPassFailReporter$24(Runner.scala:1031)
>       at org.scalatest.tools.Runner$.$anonfun$runOptionallyWithPassFailReporter$24$adapted(Runner.scala:1010)
>       at org.scalatest.tools.Runner$.withClassLoaderAndDispatchReporter(Runner.scala:1506)
>       at org.scalatest.tools.Runner$.runOptionallyWithPassFailReporter(Runner.scala:1010)
>       at org.scalatest.tools.Runner$.main(Runner.scala:827)
>       at org.scalatest.tools.Runner.main(Runner.scala) {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org