You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@beam.apache.org by "Reinier Kip (JIRA)" <ji...@apache.org> on 2017/12/12 15:13:00 UTC

[jira] [Updated] (BEAM-2831) Pipeline crashes due to Beam encoder breaking Flink memory management

     [ https://issues.apache.org/jira/browse/BEAM-2831?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Reinier Kip updated BEAM-2831:
------------------------------
    Summary: Pipeline crashes due to Beam encoder breaking Flink memory management  (was: Possible bug in Beam+Flink memory management, disk spillover)

> Pipeline crashes due to Beam encoder breaking Flink memory management
> ---------------------------------------------------------------------
>
>                 Key: BEAM-2831
>                 URL: https://issues.apache.org/jira/browse/BEAM-2831
>             Project: Beam
>          Issue Type: Bug
>          Components: runner-flink
>    Affects Versions: 2.0.0, 2.1.0
>         Environment: Flink 1.2.1 and 1.3.0, Java HotSpot and OpenJDK 8, macOS 10.12.6 and unknown Linux
>            Reporter: Reinier Kip
>            Assignee: Aljoscha Krettek
>
> I’ve been running a Beam pipeline on Flink. Depending on the dataset size and the heap memory configuration of the jobmanager and taskmanager, I may run into an EOFException, which causes the job to fail.
> As [discussed on Flink's mailinglist|http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/EOFException-related-to-memory-segments-during-run-of-Beam-pipeline-on-Flink-td15255.html] (stacktrace enclosed), Flink catches these EOFExceptions and activates disk spillover. Because Beam wraps these exceptions, this mechanism fails, the exception travels up the stack, and the job aborts.
> Hopefully this is enough information and this is something that can be adjusted for in Beam. I'd be glad to provide more information where needed.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)