Thank you for the clarification. On Fri, Sep 20, 2019 at 6:59 AM Fabian Hueske <fhue...@gmail.com> wrote:
> Hi, > > It depends. > > There are many things that can be changed. A savepoint in Flink contains > only the state of the application and not the configuration of the system. > So an application can be migrated to another cluster that runs with a > different configuration. > There are some exceptions like the configuration of the default state > backend (in case it is not configured in the application itself) and the > checkpoint techniques. > > If it is about the configuration of the application itself (and not the > system), you can do a lot of things in Flink. > You can even implement the application in a way that it reconfigures > itself while it is running. > > Since the last release (Flink 1.9), Flink features the Savepoint Processor > API which allows to create or modify savepoints with a batch program. > This can be used to adjust or bootstrap savepoints. > > Best, Fabian > > > Am Mi., 18. Sept. 2019 um 18:56 Uhr schrieb Abrar Sheikh < > abrar200...@gmail.com>: > >> Hey all, >> >> One of the known things with Spark Stateful Streaming application is that >> we cannot alter Spark Configurations or Kafka Configurations after the >> first run of the stateful streaming application, this has been explained >> well in >> https://www.linkedin.com/pulse/upgrading-running-spark-streaming-application-code-changes-prakash/ >> >> Is this also something Stateful Flink Application share in common with >> Spark? >> >> Thanks, >> >> -- >> Abrar Sheikh >> > -- Abrar Sheikh