You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@spark.apache.org by Andrew Ash <an...@andrewash.com> on 2014/02/14 12:38:29 UTC
RDD.top() stacktrace
Spark 0.9.0
Hi Spark devs,
I'm pretty sure this stacktrace is a bug in the way Spark is using the type
system but I don't quite know what it is. Something to do with type bounds
judging from my Googling.
Can someone with more Scala-foo than me please take a look? In the
meantime I'll be avoiding top() for a bit.
Thanks!
Andrew
This stracktrace came about when I called
val myRDD: RDD[(String,Int)] = ...
myRDD.reduceByKey(_+_).top(100)
But my toy example doesn't trigger the repro:
sc.parallelize(Seq( ("A",10), ("B",5), ("A",4), ("C", 15)
)).reduceByKey(_+_).top(2)
14/02/14 03:15:07 ERROR OneForOneStrategy:
scala.collection.immutable.$colon$colon cannot be cast to
org.apache.spark.util.BoundedPriorityQueue
java.lang.ClassCastException: scala.collection.immutable.$colon$colon
cannot be cast to org.apache.spark.util.BoundedPriorityQueue
at org.apache.spark.rdd.RDD$$anonfun$top$2.apply(RDD.scala:873)
at org.apache.spark.rdd.RDD$$anonfun$6.apply(RDD.scala:671)
at org.apache.spark.rdd.RDD$$anonfun$6.apply(RDD.scala:668)
at
org.apache.spark.scheduler.JobWaiter.taskSucceeded(JobWaiter.scala:56)
at
org.apache.spark.scheduler.DAGScheduler.handleTaskCompletion(DAGScheduler.scala:859)
at
org.apache.spark.scheduler.DAGScheduler.processEvent(DAGScheduler.scala:616)
at
org.apache.spark.scheduler.DAGScheduler$$anonfun$start$1$$anon$2$$anonfun$receive$1.applyOrElse(DAGScheduler.scala:207)
at akka.actor.ActorCell.receiveMessage(ActorCell.scala:498)
at akka.actor.ActorCell.invoke(ActorCell.scala:456)
at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:237)
at akka.dispatch.Mailbox.run(Mailbox.scala:219)
at
akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:386)
at
scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
at
scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
at
scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
at
scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
Re: RDD.top() stacktrace
Posted by Andrew Ash <an...@andrewash.com>.
I'm sorry I do not yet have a consistent repro that doesn't use my internal
dataset. That example is the repro I attempted to create but it does not
trigger the issue. Hopefully I'll be able to get a consistent repro to
share.
On Fri, Feb 14, 2014 at 3:58 AM, Xuefeng Wu <be...@gmail.com> wrote:
> Hi Andrew,
>
> Sorry, I can not reproduce the issue by:
>
> scala> import org.apache.spark.rdd.RDD
>
> scala> val myRDD: RDD[(String,Int)] = sc.parallelize(Seq( ("A",10),
> ("B",5), ("A",4), ("C", 15)))
>
> scala> myRDD.reduceByKey(_+_).top(2)
>
> Any different compare with your example ?
>
>
>
> On Fri, Feb 14, 2014 at 7:38 PM, Andrew Ash <an...@andrewash.com> wrote:
>
> > Spark 0.9.0
> >
> >
> > Hi Spark devs,
> >
> > I'm pretty sure this stacktrace is a bug in the way Spark is using the
> type
> > system but I don't quite know what it is. Something to do with type
> bounds
> > judging from my Googling.
> >
> > Can someone with more Scala-foo than me please take a look? In the
> > meantime I'll be avoiding top() for a bit.
> >
> > Thanks!
> > Andrew
> >
> >
> >
> > This stracktrace came about when I called
> > val myRDD: RDD[(String,Int)] = ...
> > myRDD.reduceByKey(_+_).top(100)
> >
> > But my toy example doesn't trigger the repro:
> > sc.parallelize(Seq( ("A",10), ("B",5), ("A",4), ("C", 15)
> > )).reduceByKey(_+_).top(2)
> >
> >
> >
> > 14/02/14 03:15:07 ERROR OneForOneStrategy:
> > scala.collection.immutable.$colon$colon cannot be cast to
> > org.apache.spark.util.BoundedPriorityQueue
> > java.lang.ClassCastException: scala.collection.immutable.$colon$colon
> > cannot be cast to org.apache.spark.util.BoundedPriorityQueue
> > at org.apache.spark.rdd.RDD$$anonfun$top$2.apply(RDD.scala:873)
> > at org.apache.spark.rdd.RDD$$anonfun$6.apply(RDD.scala:671)
> > at org.apache.spark.rdd.RDD$$anonfun$6.apply(RDD.scala:668)
> > at
> > org.apache.spark.scheduler.JobWaiter.taskSucceeded(JobWaiter.scala:56)
> > at
> >
> >
> org.apache.spark.scheduler.DAGScheduler.handleTaskCompletion(DAGScheduler.scala:859)
> > at
> >
> >
> org.apache.spark.scheduler.DAGScheduler.processEvent(DAGScheduler.scala:616)
> > at
> >
> >
> org.apache.spark.scheduler.DAGScheduler$$anonfun$start$1$$anon$2$$anonfun$receive$1.applyOrElse(DAGScheduler.scala:207)
> > at akka.actor.ActorCell.receiveMessage(ActorCell.scala:498)
> > at akka.actor.ActorCell.invoke(ActorCell.scala:456)
> > at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:237)
> > at akka.dispatch.Mailbox.run(Mailbox.scala:219)
> > at
> >
> >
> akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:386)
> > at
> > scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
> > at
> >
> >
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
> > at
> > scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
> > at
> >
> >
> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
> >
>
>
>
> --
>
> ~Yours, Xuefeng Wu/吴雪峰 敬上
>
Re: RDD.top() stacktrace
Posted by Xuefeng Wu <be...@gmail.com>.
Hi Andrew,
Sorry, I can not reproduce the issue by:
scala> import org.apache.spark.rdd.RDD
scala> val myRDD: RDD[(String,Int)] = sc.parallelize(Seq( ("A",10),
("B",5), ("A",4), ("C", 15)))
scala> myRDD.reduceByKey(_+_).top(2)
Any different compare with your example ?
On Fri, Feb 14, 2014 at 7:38 PM, Andrew Ash <an...@andrewash.com> wrote:
> Spark 0.9.0
>
>
> Hi Spark devs,
>
> I'm pretty sure this stacktrace is a bug in the way Spark is using the type
> system but I don't quite know what it is. Something to do with type bounds
> judging from my Googling.
>
> Can someone with more Scala-foo than me please take a look? In the
> meantime I'll be avoiding top() for a bit.
>
> Thanks!
> Andrew
>
>
>
> This stracktrace came about when I called
> val myRDD: RDD[(String,Int)] = ...
> myRDD.reduceByKey(_+_).top(100)
>
> But my toy example doesn't trigger the repro:
> sc.parallelize(Seq( ("A",10), ("B",5), ("A",4), ("C", 15)
> )).reduceByKey(_+_).top(2)
>
>
>
> 14/02/14 03:15:07 ERROR OneForOneStrategy:
> scala.collection.immutable.$colon$colon cannot be cast to
> org.apache.spark.util.BoundedPriorityQueue
> java.lang.ClassCastException: scala.collection.immutable.$colon$colon
> cannot be cast to org.apache.spark.util.BoundedPriorityQueue
> at org.apache.spark.rdd.RDD$$anonfun$top$2.apply(RDD.scala:873)
> at org.apache.spark.rdd.RDD$$anonfun$6.apply(RDD.scala:671)
> at org.apache.spark.rdd.RDD$$anonfun$6.apply(RDD.scala:668)
> at
> org.apache.spark.scheduler.JobWaiter.taskSucceeded(JobWaiter.scala:56)
> at
>
> org.apache.spark.scheduler.DAGScheduler.handleTaskCompletion(DAGScheduler.scala:859)
> at
>
> org.apache.spark.scheduler.DAGScheduler.processEvent(DAGScheduler.scala:616)
> at
>
> org.apache.spark.scheduler.DAGScheduler$$anonfun$start$1$$anon$2$$anonfun$receive$1.applyOrElse(DAGScheduler.scala:207)
> at akka.actor.ActorCell.receiveMessage(ActorCell.scala:498)
> at akka.actor.ActorCell.invoke(ActorCell.scala:456)
> at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:237)
> at akka.dispatch.Mailbox.run(Mailbox.scala:219)
> at
>
> akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:386)
> at
> scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
> at
>
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
> at
> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
> at
>
> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
>
--
~Yours, Xuefeng Wu/吴雪峰 敬上