You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@flink.apache.org by Flavio Pompermaier <po...@okkam.it> on 2015/12/14 11:32:22 UTC

Streaming to db question

Hi flinkers,
I was going to evaluate if Flink streaming could fit a use case we have,
where data comes into the system, gets transformed and then added to a db
(a very common problem..).
In such use case you have to manage the merge of existing records as new
data come in. How can you ensure that only one row/entity of the db is
updated at a time with Flink?
Is there any example?

Best,
Flavio

Re: Streaming to db question

Posted by Flavio Pompermaier <po...@okkam.it>.
I was thinking to something more like
http://www.infoq.com/articles/key-lessons-learned-from-transition-to-nosql
that basically implement what you call Out-of-core state at
https://cwiki.apache.org/confluence/display/FLINK/Stateful+Stream+Processing.
Riak provide
some feature to handle the eventually consistent nature of that use
case...or are you more likely to go with the current proposed soluion (the
one in the Flink wiki...)?

On Mon, Dec 14, 2015 at 8:18 PM, Stephan Ewen <se...@apache.org> wrote:

> Hi!
>
> If the sink that writes to the Database executes partitioned by the
> primary key, then this should naturally prevent row conflicts.
>
> Greetings,
> Stephan
>
>
> On Mon, Dec 14, 2015 at 11:32 AM, Flavio Pompermaier <pompermaier@okkam.it
> > wrote:
>
>> Hi flinkers,
>> I was going to evaluate if Flink streaming could fit a use case we have,
>> where data comes into the system, gets transformed and then added to a db
>> (a very common problem..).
>> In such use case you have to manage the merge of existing records as new
>> data come in. How can you ensure that only one row/entity of the db is
>> updated at a time with Flink?
>> Is there any example?
>>
>> Best,
>> Flavio
>>
>

Re: Streaming to db question

Posted by Stephan Ewen <se...@apache.org>.
Hi!

If the sink that writes to the Database executes partitioned by the primary
key, then this should naturally prevent row conflicts.

Greetings,
Stephan


On Mon, Dec 14, 2015 at 11:32 AM, Flavio Pompermaier <po...@okkam.it>
wrote:

> Hi flinkers,
> I was going to evaluate if Flink streaming could fit a use case we have,
> where data comes into the system, gets transformed and then added to a db
> (a very common problem..).
> In such use case you have to manage the merge of existing records as new
> data come in. How can you ensure that only one row/entity of the db is
> updated at a time with Flink?
> Is there any example?
>
> Best,
> Flavio
>