You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Josh Rosen (JIRA)" <ji...@apache.org> on 2016/09/12 22:38:20 UTC
[jira] [Updated] (SPARK-13406) NPE in LazilyGeneratedOrdering
[ https://issues.apache.org/jira/browse/SPARK-13406?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Josh Rosen updated SPARK-13406:
-------------------------------
Assignee: (was: Josh Rosen)
> NPE in LazilyGeneratedOrdering
> ------------------------------
>
> Key: SPARK-13406
> URL: https://issues.apache.org/jira/browse/SPARK-13406
> Project: Spark
> Issue Type: Improvement
> Components: SQL
> Reporter: Davies Liu
>
> {code}
> File "/Users/davies/work/spark/python/lib/pyspark.zip/pyspark/sql/dataframe.py", line ?, in pyspark.sql.dataframe.DataFrameStatFunctions.sampleBy
> Failed example:
> sampled.groupBy("key").count().orderBy("key").show()
> Exception raised:
> Traceback (most recent call last):
> File "//anaconda/lib/python2.7/doctest.py", line 1315, in __run
> compileflags, 1) in test.globs
> File "<doctest pyspark.sql.dataframe.DataFrameStatFunctions.sampleBy[3]>", line 1, in <module>
> sampled.groupBy("key").count().orderBy("key").show()
> File "/Users/davies/work/spark/python/lib/pyspark.zip/pyspark/sql/dataframe.py", line 217, in show
> print(self._jdf.showString(n, truncate))
> File "/Users/davies/work/spark/python/lib/py4j-0.9.1-src.zip/py4j/java_gateway.py", line 835, in __call__
> answer, self.gateway_client, self.target_id, self.name)
> File "/Users/davies/work/spark/python/lib/pyspark.zip/pyspark/sql/utils.py", line 45, in deco
> return f(*a, **kw)
> File "/Users/davies/work/spark/python/lib/py4j-0.9.1-src.zip/py4j/protocol.py", line 310, in get_return_value
> format(target_id, ".", name), value)
> Py4JJavaError: An error occurred while calling o681.showString.
> : org.apache.spark.SparkDriverExecutionException: Execution error
> at org.apache.spark.scheduler.DAGScheduler.handleTaskCompletion(DAGScheduler.scala:1189)
> at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1658)
> at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1620)
> at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1609)
> at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)
> at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:623)
> at org.apache.spark.SparkContext.runJob(SparkContext.scala:1782)
> at org.apache.spark.SparkContext.runJob(SparkContext.scala:1845)
> at org.apache.spark.rdd.RDD$$anonfun$reduce$1.apply(RDD.scala:937)
> at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
> at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111)
> at org.apache.spark.rdd.RDD.withScope(RDD.scala:323)
> at org.apache.spark.rdd.RDD.reduce(RDD.scala:919)
> at org.apache.spark.rdd.RDD$$anonfun$takeOrdered$1.apply(RDD.scala:1318)
> at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
> at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111)
> at org.apache.spark.rdd.RDD.withScope(RDD.scala:323)
> at org.apache.spark.rdd.RDD.takeOrdered(RDD.scala:1305)
> at org.apache.spark.sql.execution.TakeOrderedAndProject.executeCollect(limit.scala:94)
> at org.apache.spark.sql.execution.SparkPlan.executeCollectPublic(SparkPlan.scala:157)
> at org.apache.spark.sql.DataFrame$$anonfun$org$apache$spark$sql$DataFrame$$execute$1$1.apply(DataFrame.scala:1520)
> at org.apache.spark.sql.DataFrame$$anonfun$org$apache$spark$sql$DataFrame$$execute$1$1.apply(DataFrame.scala:1520)
> at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:53)
> at org.apache.spark.sql.DataFrame.withNewExecutionId(DataFrame.scala:1769)
> at org.apache.spark.sql.DataFrame.org$apache$spark$sql$DataFrame$$execute$1(DataFrame.scala:1519)
> at org.apache.spark.sql.DataFrame.org$apache$spark$sql$DataFrame$$collect(DataFrame.scala:1526)
> at org.apache.spark.sql.DataFrame$$anonfun$head$1.apply(DataFrame.scala:1396)
> at org.apache.spark.sql.DataFrame$$anonfun$head$1.apply(DataFrame.scala:1395)
> at org.apache.spark.sql.DataFrame.withCallback(DataFrame.scala:1782)
> at org.apache.spark.sql.DataFrame.head(DataFrame.scala:1395)
> at org.apache.spark.sql.DataFrame.take(DataFrame.scala:1477)
> at org.apache.spark.sql.DataFrame.showString(DataFrame.scala:167)
> at sun.reflect.GeneratedMethodAccessor63.invoke(Unknown Source)
> at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:606)
> at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:231)
> at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:381)
> at py4j.Gateway.invoke(Gateway.java:290)
> at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:133)
> at py4j.commands.CallCommand.execute(CallCommand.java:79)
> at py4j.GatewayConnection.run(GatewayConnection.java:209)
> at java.lang.Thread.run(Thread.java:745)
> Caused by: java.lang.NullPointerException
> at org.apache.spark.sql.catalyst.expressions.codegen.LazilyGeneratedOrdering.compare(GenerateOrdering.scala:157)
> at org.apache.spark.sql.catalyst.expressions.codegen.LazilyGeneratedOrdering.compare(GenerateOrdering.scala:148)
> at scala.math.Ordering$$anon$4.compare(Ordering.scala:111)
> at java.util.PriorityQueue.siftUpUsingComparator(PriorityQueue.java:649)
> at java.util.PriorityQueue.siftUp(PriorityQueue.java:627)
> at java.util.PriorityQueue.offer(PriorityQueue.java:329)
> at org.apache.spark.util.BoundedPriorityQueue.$plus$eq(BoundedPriorityQueue.scala:47)
> at org.apache.spark.util.BoundedPriorityQueue$$anonfun$$plus$plus$eq$1.apply(BoundedPriorityQueue.scala:41)
> at org.apache.spark.util.BoundedPriorityQueue$$anonfun$$plus$plus$eq$1.apply(BoundedPriorityQueue.scala:41)
> at scala.collection.Iterator$class.foreach(Iterator.scala:742)
> at scala.collection.AbstractIterator.foreach(Iterator.scala:1194)
> at scala.collection.IterableLike$class.foreach(IterableLike.scala:72)
> at org.apache.spark.util.BoundedPriorityQueue.foreach(BoundedPriorityQueue.scala:31)
> at org.apache.spark.util.BoundedPriorityQueue.$plus$plus$eq(BoundedPriorityQueue.scala:41)
> at org.apache.spark.rdd.RDD$$anonfun$takeOrdered$1$$anonfun$apply$46.apply(RDD.scala:1319)
> at org.apache.spark.rdd.RDD$$anonfun$takeOrdered$1$$anonfun$apply$46.apply(RDD.scala:1318)
> at org.apache.spark.rdd.RDD$$anonfun$reduce$1$$anonfun$15.apply(RDD.scala:932)
> at org.apache.spark.rdd.RDD$$anonfun$reduce$1$$anonfun$15.apply(RDD.scala:929)
> at org.apache.spark.scheduler.JobWaiter.taskSucceeded(JobWaiter.scala:57)
> at org.apache.spark.scheduler.DAGScheduler.handleTaskCompletion(DAGScheduler.scala:1185)
> at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1658)
> at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1620)
> at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1609)
> at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)
> {code}
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org