You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@flink.apache.org by "Andrey Zagrebin (JIRA)" <ji...@apache.org> on 2019/01/18 13:01:00 UTC

[jira] [Assigned] (FLINK-10653) Introduce Pluggable Shuffle Manager Architecture

     [ https://issues.apache.org/jira/browse/FLINK-10653?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Andrey Zagrebin reassigned FLINK-10653:
---------------------------------------

    Assignee: zhijiang  (was: Andrey Zagrebin)

> Introduce Pluggable Shuffle Manager Architecture
> ------------------------------------------------
>
>                 Key: FLINK-10653
>                 URL: https://issues.apache.org/jira/browse/FLINK-10653
>             Project: Flink
>          Issue Type: New Feature
>          Components: Network
>            Reporter: zhijiang
>            Assignee: zhijiang
>            Priority: Major
>             Fix For: 1.8.0
>
>
> This is the umbrella issue for improving shuffle architecture.
> Shuffle is the process of data transfer between stages, which involves in writing outputs on sender side and reading data on receiver side. In flink implementation, it covers three parts of writer, transport layer and reader separately which are uniformed for both streaming and batch jobs.
> In detail, the current ResultPartitionWriter interface on upstream side only supports in-memory outputs for streaming job and local persistent file outputs for batch job. If we extend to implement another writer such as DfsWriter, RdmaWriter, SortMergeWriter, etc based on ResultPartitionWriter interface, it has not theĀ unified mechanism to extend the reader side accordingly. 
> In order to make the shuffle architecture more flexible and support more scenarios especially for batch jobs, a high level shuffle architecture is necessary to manage and extend both writer and reader sides together.
> Refer to the design doc for more details.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)