You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@flink.apache.org by Marco Villalobos <mv...@kineteque.com> on 2020/08/07 21:51:42 UTC

State Processor API to boot strap keyed state for Stream Application.

I have read the documentation and various blogs that state that it is
possible to load data into a data-set and use that data to bootstrap a
stream application.

The documentation literally says this, "...you can read a batch of data
from any store, preprocess it, and write the result to a savepoint that you
use to bootstrap the state of a streaming application." (source:
https://ci.apache.org/projects/flink/flink-docs-master/dev/libs/state_processor_api.html
).

Another blog states, "You can create both Batch and Stream environment in a
single job." (source:
https://www.kharekartik.dev/2019/12/14/bootstrap-your-flink-jobs/

I want to try this approach, but I cannot find any real examples online.

I have failed on numerous attempts.

I have a few questions:

1) is there an example that demonstrate this feature?
2) how can you launch batch and stream environment from a single job?
3) does this require two jobs?

Anybody, please help.

Re: State Processor API to boot strap keyed state for Stream Application.

Posted by Arvid Heise <ar...@ververica.com>.
For future readers: this thread has been resolved in "Please help, I need
to bootstrap keyed state into a stream" on the user mailing list asked by
Marco.

On Fri, Aug 7, 2020 at 11:52 PM Marco Villalobos <mv...@kineteque.com>
wrote:

> I have read the documentation and various blogs that state that it is
> possible to load data into a data-set and use that data to bootstrap a
> stream application.
>
> The documentation literally says this, "...you can read a batch of data
> from any store, preprocess it, and write the result to a savepoint that you
> use to bootstrap the state of a streaming application." (source:
> https://ci.apache.org/projects/flink/flink-docs-master/dev/libs/state_processor_api.html
> ).
>
> Another blog states, "You can create both Batch and Stream environment in
> a single job." (source:
> https://www.kharekartik.dev/2019/12/14/bootstrap-your-flink-jobs/
>
> I want to try this approach, but I cannot find any real examples online.
>
> I have failed on numerous attempts.
>
> I have a few questions:
>
> 1) is there an example that demonstrate this feature?
> 2) how can you launch batch and stream environment from a single job?
> 3) does this require two jobs?
>
> Anybody, please help.
>
>

-- 

Arvid Heise | Senior Java Developer

<https://www.ververica.com/>

Follow us @VervericaData

--

Join Flink Forward <https://flink-forward.org/> - The Apache Flink
Conference

Stream Processing | Event Driven | Real Time

--

Ververica GmbH | Invalidenstrasse 115, 10115 Berlin, Germany

--
Ververica GmbH
Registered at Amtsgericht Charlottenburg: HRB 158244 B
Managing Directors: Timothy Alexander Steinert, Yip Park Tung Jason, Ji
(Toni) Cheng