You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Jungtaek Lim (Jira)" <ji...@apache.org> on 2019/09/20 03:10:00 UTC
[jira] [Resolved] (SPARK-29134) Flaky test:
org.apache.spark.sql.execution.SQLExecutionSuite.concurrent query execution
with fork-join pool (SPARK-13747)
[ https://issues.apache.org/jira/browse/SPARK-29134?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Jungtaek Lim resolved SPARK-29134.
----------------------------------
Resolution: Invalid
Just affected by problematic commit (had been reverted) [[https://github.com/apache/spark/commit/850833fa177ec1f265e143fc383e40ec2c8341a6]]
See more: [[https://github.com/apache/spark/pull/25790#issuecomment-531517800]]
> Flaky test: org.apache.spark.sql.execution.SQLExecutionSuite.concurrent query execution with fork-join pool (SPARK-13747)
> -------------------------------------------------------------------------------------------------------------------------
>
> Key: SPARK-29134
> URL: https://issues.apache.org/jira/browse/SPARK-29134
> Project: Spark
> Issue Type: Bug
> Components: SQL, Tests
> Affects Versions: 3.0.0
> Reporter: Jungtaek Lim
> Priority: Major
>
> [https://amplab.cs.berkeley.edu/jenkins/view/Spark%20QA%20Test/job/spark-master-test-maven-hadoop-2.7-jdk-11-ubuntu-testing/1544/testReport/]
> {code:java}
> scala.ScalaReflectionException: type T1 is not a class
> at scala.reflect.api.Symbols$SymbolApi.asClass(Symbols.scala:284)
> at scala.reflect.api.Symbols$SymbolApi.asClass$(Symbols.scala:284)
> at scala.reflect.internal.Symbols$SymbolContextApiImpl.asClass(Symbols.scala:106)
> at org.apache.spark.sql.catalyst.ScalaReflection$.getClassFromType(ScalaReflection.scala:617)
> at org.apache.spark.sql.catalyst.ScalaReflection$.$anonfun$dataTypeFor$1(ScalaReflection.scala:105)
> at scala.reflect.internal.tpe.TypeConstraints$UndoLog.undo(TypeConstraints.scala:68)
> at org.apache.spark.sql.catalyst.ScalaReflection.cleanUpReflectionObjects(ScalaReflection.scala:848)
> at org.apache.spark.sql.catalyst.ScalaReflection.cleanUpReflectionObjects$(ScalaReflection.scala:847)
> at org.apache.spark.sql.catalyst.ScalaReflection$.cleanUpReflectionObjects(ScalaReflection.scala:47)
> at org.apache.spark.sql.catalyst.ScalaReflection$.dataTypeFor(ScalaReflection.scala:86)
> at org.apache.spark.sql.catalyst.ScalaReflection$.$anonfun$serializerFor$6(ScalaReflection.scala:551)
> at scala.collection.TraversableLike.$anonfun$map$1(TraversableLike.scala:237)
> at scala.collection.immutable.List.foreach(List.scala:392)
> at scala.collection.TraversableLike.map(TraversableLike.scala:237)
> at scala.collection.TraversableLike.map$(TraversableLike.scala:230)
> at scala.collection.immutable.List.map(List.scala:298)
> at org.apache.spark.sql.catalyst.ScalaReflection$.$anonfun$serializerFor$1(ScalaReflection.scala:541)
> at scala.reflect.internal.tpe.TypeConstraints$UndoLog.undo(TypeConstraints.scala:68)
> at org.apache.spark.sql.catalyst.ScalaReflection.cleanUpReflectionObjects(ScalaReflection.scala:848)
> at org.apache.spark.sql.catalyst.ScalaReflection.cleanUpReflectionObjects$(ScalaReflection.scala:847)
> at org.apache.spark.sql.catalyst.ScalaReflection$.cleanUpReflectionObjects(ScalaReflection.scala:47)
> at org.apache.spark.sql.catalyst.ScalaReflection$.serializerFor(ScalaReflection.scala:410)
> at org.apache.spark.sql.catalyst.ScalaReflection$.$anonfun$serializerForType$1(ScalaReflection.scala:399)
> at scala.reflect.internal.tpe.TypeConstraints$UndoLog.undo(TypeConstraints.scala:68)
> at org.apache.spark.sql.catalyst.ScalaReflection.cleanUpReflectionObjects(ScalaReflection.scala:848)
> at org.apache.spark.sql.catalyst.ScalaReflection.cleanUpReflectionObjects$(ScalaReflection.scala:847)
> at org.apache.spark.sql.catalyst.ScalaReflection$.cleanUpReflectionObjects(ScalaReflection.scala:47)
> at org.apache.spark.sql.catalyst.ScalaReflection$.serializerForType(ScalaReflection.scala:391)
> at org.apache.spark.sql.catalyst.encoders.ExpressionEncoder$.apply(ExpressionEncoder.scala:54)
> at org.apache.spark.sql.Encoders$.product(Encoders.scala:285)
> at org.apache.spark.sql.LowPrioritySQLImplicits.newProductEncoder(SQLImplicits.scala:251)
> at org.apache.spark.sql.LowPrioritySQLImplicits.newProductEncoder$(SQLImplicits.scala:251)
> at org.apache.spark.sql.SQLImplicits.newProductEncoder(SQLImplicits.scala:32)
> at org.apache.spark.sql.execution.SQLExecutionSuite.$anonfun$new$3(SQLExecutionSuite.scala:50)
> at scala.runtime.java8.JFunction1$mcJI$sp.apply(JFunction1$mcJI$sp.java:23)
> at scala.Function1.apply$mcVI$sp(Function1.scala:41)
> at scala.collection.parallel.immutable.ParRange$ParRangeIterator.foreach(ParRange.scala:94)
> at scala.collection.parallel.ParIterableLike$Foreach.leaf(ParIterableLike.scala:974)
> at scala.collection.parallel.Task.$anonfun$tryLeaf$1(Tasks.scala:53)
> at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
> at scala.util.control.Breaks$$anon$1.catchBreak(Breaks.scala:67)
> at scala.collection.parallel.Task.tryLeaf(Tasks.scala:56)
> at scala.collection.parallel.Task.tryLeaf$(Tasks.scala:50)
> at scala.collection.parallel.ParIterableLike$Foreach.tryLeaf(ParIterableLike.scala:971)
> at scala.collection.parallel.AdaptiveWorkStealingTasks$WrappedTask.internal(Tasks.scala:160)
> at scala.collection.parallel.AdaptiveWorkStealingTasks$WrappedTask.internal$(Tasks.scala:157)
> at scala.collection.parallel.AdaptiveWorkStealingForkJoinTasks$WrappedTask.internal(Tasks.scala:440)
> at scala.collection.parallel.AdaptiveWorkStealingTasks$WrappedTask.compute(Tasks.scala:150)
> at scala.collection.parallel.AdaptiveWorkStealingTasks$WrappedTask.compute$(Tasks.scala:149)
> at scala.collection.parallel.AdaptiveWorkStealingForkJoinTasks$WrappedTask.compute(Tasks.scala:440)
> at java.base/java.util.concurrent.RecursiveAction.exec(RecursiveAction.java:189)
> at java.base/java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:290)
> at java.base/java.util.concurrent.ForkJoinPool$WorkQueue.topLevelExec(ForkJoinPool.java:1020)
> at java.base/java.util.concurrent.ForkJoinPool.scan(ForkJoinPool.java:1656)
> at java.base/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1594)
> at java.base/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:177) {code}
--
This message was sent by Atlassian Jira
(v8.3.4#803005)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org