Hi,

also +1.
As vivo said, SavePoint is not compatible.
I have heard from a lot of users just said "my previous program does not
work any more!".
If these utilities provide such migration functions, it would be perfect!

Best,
tison.


vino yang <yanghua1...@gmail.com> 于2018年8月18日周六 上午11:04写道:

> Hi,
>
> +1, from my side. Considering that Savepoint is not compatible with every
> version of Flink.
> It would be very useful if it could do some conversion between different
> versions of Flink Savepoint.
>
> Thanks, vino.
>
> bupt_ljy <bupt_...@163.com> 于2018年8月18日周六 上午2:55写道:
>
> > Hi,
> > +1, I think it will be a very great tool for Flink, especially the
> > creating new state part. On production, we’re really worried about the
> > availability of the savepoints, because the generating logic is inside
> > Flink and we don’t have a good way to validate it. But with this tool, we
> > can construct a new state for our programs very soon even if the
> savepoints
> > data is broken.
> > It’s great, thanks!
> >
> >
> > Original Message
> > Sender:Jamie grierjgr...@lyft.com
> > Recipient:dev...@flink.apache.org
> > Date:Saturday, Aug 18, 2018 02:32
> > Subject:Re: [Proposal] Utilities for reading, transforming and
> > creatingStreaming savepoints
> >
> >
> > This is great, Gyula! A colleague here at Lyft has also done some work
> > around bootstrapping DataStream programs and we've also talked a bit
> about
> > doing this by running DataSet programs. On Fri, Aug 17, 2018 at 3:28 AM,
> > Gyula Fóra gyula.f...@gmail.com wrote:  Hi All!   I want to share with
> > you a little project we have been working on at King  (with some help
> from
> > some dataArtisans folks). I think this would be a  valuable addition to
> > Flink and solve a bunch of outstanding production  use-cases and
> headaches
> > around state bootstrapping and state analytics.   We have built a quick
> and
> > dirty POC implementation on top of Flink 1.6,  please check the README
> for
> > some nice examples to get a quick idea:   https://github.com/king/bravo
> >  *Short story*  Bravo is a convenient state reader and writer library
> > leveraging the  Flink’s batch processing capabilities. It supports
> > processing and writing  Flink streaming savepoints. At the moment it only
> > supports processing  RocksDB savepoints but this can be extended in the
> > future for other state  backends and checkpoint types.   Our goal is to
> > cover a few basic features:   - Converting keyed states to Flink DataSets
> > for processing and analytics  - Reading/Writing non-keyed operators
> states
> > - Bootstrap keyed states from Flink DataSets and create new valid
> > savepoints  - Transform existing savepoints by replacing/changing some
> > states    Some example use-cases:   - Point-in-time state analytics
> across
> > all operators and keys  - Bootstrap state of a streaming job from
> external
> > resources such as  reading from database/filesystem  - Validate and
> > potentially repair corrupted state of a streaming job  - Change max
> > parallelism of a job    Our main goal is to start working together with
> > other Flink production  users and make this something useful that can be
> > part of Flink. So if you  have use-cases please talk to us :)  I have
> also
> > started a google doc which contains a little bit more info than  the
> readme
> > and could be a starting place for discussions:
> > https://docs.google.com/document/d/103k6wPX20kMu5H3SOOXSg5PZIaYpw
> > dhqBMr-ppkFL5E/edit?usp=sharing   I know there are a bunch of rough edges
> > and bugs (and no tests) but our  motto is: If you are not embarrassed,
> you
> > released too late :)   Please let me know what you think!   Cheers,
> Gyula
>

Reply via email to