You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@flink.apache.org by "Robert Metzger (Jira)" <ji...@apache.org> on 2019/12/12 16:03:00 UTC

[jira] [Commented] (FLINK-14868) Provides the ability for multiple sinks to write data serially

    [ https://issues.apache.org/jira/browse/FLINK-14868?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16994812#comment-16994812 ] 

Robert Metzger commented on FLINK-14868:
----------------------------------------

Thanks a lot for filing this ticket. If you want to write data to the sinks in a way that the records are written "together" into a number of sinks, I believe you can solve this yourself by implementing a custom sink, wrapping the sinks you want to write in. Check out the {{SinkFunction}} interface for that.

 

> Provides the ability for multiple sinks to write data serially
> --------------------------------------------------------------
>
>                 Key: FLINK-14868
>                 URL: https://issues.apache.org/jira/browse/FLINK-14868
>             Project: Flink
>          Issue Type: Wish
>          Components: API / DataStream, Table SQL / Runtime
>    Affects Versions: 1.9.1
>            Reporter: ShenDa
>            Priority: Major
>
> At present, Flink can use multiple sinks to write data into different data source such as HBase,Kafka,Elasticsearch,etc.And this process is concurrent ,in other words, one record will be written into data sources simultaneously.
> But there is no approach that can sinking data serially.We really wish Flink can providing this kind of ability that a sink can write data into target database only after the previous sink transfers data successfully.And if the previous sink encounters any exception, the next sink will not work.
> h1.  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)