You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Brett Stime (JIRA)" <ji...@apache.org> on 2016/01/22 17:58:39 UTC

[jira] [Commented] (SPARK-12831) akka.remote.OversizedPayloadException on DirectTaskResult

    [ https://issues.apache.org/jira/browse/SPARK-12831?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15112676#comment-15112676 ] 

Brett Stime commented on SPARK-12831:
-------------------------------------

Actually, even with more conservative timeouts, the jobs get stalled indefinitely (for hours). Smaller frame sizes haven't helped.

I'm going to try a custom build of Spark core with the following line in Executor.scala:

private val akkaReservedSizeBytes = conf.getInt("spark.akka.reserved.bytes", AkkaUtils.reservedSizeBytes)

...and I'll use the new val in the place of the direct reference to AkkaUtils.reservedSizeBytes .

That should at least allow working around the issue if one knows where to look. If it's helpful, I could file e.g., a pull request on GitHub or whatever would be preferred. Ideally, a more thorough fix would detect and recover from the issue instead of hanging (in addition to making the issue less likely by increasing the size of the reservation).

> akka.remote.OversizedPayloadException on DirectTaskResult
> ---------------------------------------------------------
>
>                 Key: SPARK-12831
>                 URL: https://issues.apache.org/jira/browse/SPARK-12831
>             Project: Spark
>          Issue Type: Bug
>          Components: Spark Core
>            Reporter: Brett Stime
>
> Getting the following error in my executor logs:
> ERROR akka.ErrorMonitor: Transient association error (association remains live)
> akka.remote.OversizedPayloadException: Discarding oversized payload sent to Actor[akka.tcp://sparkDriver@172.21.25.199:51562/user/CoarseGrainedScheduler#-2039547722]: max allowed size 134217728 bytes, actual size of encoded class org.apache.spark.rpc.akka.AkkaMessage was 134419636 bytes.
> Seems like the quick fix would be to make AkkaUtils.reservedSizeBytes a little bigger--maybe proportional to spark.akka.frameSize and/or user configurable.
> A more robust solution might be to catch OversizedPayloadException and retry using the BlockManager.
> I should also mention that this has the effect of stalling the entire job (my use case also requires fairly liberal timeouts). For now, I'll see if setting spark.akka.frameSize a little smaller gives me more proportional overhead.
> Thanks.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org