Hi Robert,

you basically just (re)write your application with DataStream API, use the
new KafkaSource, and then let the automatic batch detection mode work [1].
The most important part is that all your sources need to be bounded.
Assuming that you just have a Kafka source, you need to use setBounded with
the appropriate end offset/timestamp.

Note that the rewritten Kafka source still has a couple of issues that
should be addressed by the first bugfix release of 1.12 in this month. So
while it's safe to use for development, I'd wait for 1.12.1 to roll it out
on production.

If you have specific questions on the migration from DataSet and
DataStream, please let me know.

[1]
https://ci.apache.org/projects/flink/flink-docs-release-1.12/dev/datastream_execution_mode.html

On Mon, Jan 4, 2021 at 7:34 PM Robert Cullen <cinquate...@gmail.com> wrote:

> I have a Kafka source that I would like to run a batch job on.  Since
> Version 1.12.0 is now soft deprecating the DataSet API in favor of the
> DataStream API, can someone show me an example of this? (Using DataStream)
>
> thanks
> --
> Robert Cullen
> 240-475-4490
>


-- 

Arvid Heise | Senior Java Developer

<https://www.ververica.com/>

Follow us @VervericaData

--

Join Flink Forward <https://flink-forward.org/> - The Apache Flink
Conference

Stream Processing | Event Driven | Real Time

--

Ververica GmbH | Invalidenstrasse 115, 10115 Berlin, Germany

--
Ververica GmbH
Registered at Amtsgericht Charlottenburg: HRB 158244 B
Managing Directors: Timothy Alexander Steinert, Yip Park Tung Jason, Ji
(Toni) Cheng

Reply via email to