You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@beam.apache.org by "Reinier Kip (JIRA)" <ji...@apache.org> on 2017/12/12 15:22:03 UTC

[jira] [Comment Edited] (BEAM-2831) Pipeline crashes due to Beam encoder breaking Flink memory management

    [ https://issues.apache.org/jira/browse/BEAM-2831?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16287734#comment-16287734 ] 

Reinier Kip edited comment on BEAM-2831 at 12/12/17 3:21 PM:
-------------------------------------------------------------

Hey guys and gals, is there intent to address this issue in Beam? I encounter someone about every month that struggles with this issue; it is not very straightforward to solve. If type rigidity is an issue, a hint added to the exception message to use another encoder would already be of great help.

Whatever the solution, I can put in the time to submit the actual change.


was (Author: rjkip):
Hey guys and gals, is there intent to address this issue in Beam? I encounter someone about every month that struggles with this issue; it is not very straightforward to solve. If type rigidity is an issue, a hint added to the exception message to use another encoder would already be of great help.

> Pipeline crashes due to Beam encoder breaking Flink memory management
> ---------------------------------------------------------------------
>
>                 Key: BEAM-2831
>                 URL: https://issues.apache.org/jira/browse/BEAM-2831
>             Project: Beam
>          Issue Type: Bug
>          Components: runner-flink
>    Affects Versions: 2.0.0, 2.1.0
>         Environment: Flink 1.2.1 and 1.3.0, Java HotSpot and OpenJDK 8, macOS 10.12.6 and unknown Linux
>            Reporter: Reinier Kip
>            Assignee: Aljoscha Krettek
>
> I’ve been running a Beam pipeline on Flink. Depending on the dataset size and the heap memory configuration of the jobmanager and taskmanager, I may run into an EOFException, which causes the job to fail.
> As [discussed on Flink's mailinglist|http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/EOFException-related-to-memory-segments-during-run-of-Beam-pipeline-on-Flink-td15255.html] (stacktrace enclosed), Flink catches these EOFExceptions and activates disk spillover. Because Beam wraps these exceptions, this mechanism fails, the exception travels up the stack, and the job aborts.
> Hopefully this is enough information and this is something that can be adjusted for in Beam. I'd be glad to provide more information where needed.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)