Hi,

you could use Zookeeper if you want to dynamically change the DB name /
credentials at runtime.

The GlobalJobParameters are immutable at runtime, so you can not pass
updates through it to the cluster.
They are intended for parameters for all operators/the entire job and the
web interface.

Regards,
Robert


On Thu, Aug 25, 2016 at 7:03 AM, Baswaraj Kasture <kbaswar...@gmail.com>
wrote:

> Thanks to all for your inputs.
> Yeah, I could put all these common configurations/rules in DB and workers
> can pick it up dynamically at run time.
> In this case DB configuration/connection details need to be hard coded  ?
> Is there any way worker can pickup  DB name/credentials etc at run time
> dynamically ?
>
> I am going through the feature/API documentation, but how about using
> function closer  and setGlobalJobParameters/getGlobalJobParameters ?
>
> +Baswaraj
>
> On Wed, Aug 24, 2016 at 5:17 PM, Maximilian Michels <m...@apache.org>
> wrote:
>
>> Hi!
>>
>> 1. The community is working on adding side inputs to the DataStream
>> API. That will allow you to easily distribute data to all of your
>> workers.
>>
>> 2. In the meantime, you could use `.broadcast()` on a DataSet to
>> broadcast data to all workers. You still have to join that data with
>> another stream though.
>>
>> 3. The easiest method of all is to simply load your file in the
>> RichMapFunction's open() method. The file can reside in a distributed
>> file system which is accessible by all workers.
>>
>> Cheers,
>> Max
>>
>> On Wed, Aug 24, 2016 at 6:45 AM, Jark Wu <wuchong...@alibaba-inc.com>
>> wrote:
>> > Hi,
>> >
>> > I think what Bswaraj want is excatly something like Storm Distributed
>> Cache
>> > API[1] (if I’m not misunderstanding).
>> >
>> > The distributed cache feature in storm is used to efficiently distribute
>> > files (or blobs, which is the equivalent terminology for a file in the
>> > distributed cache and is used interchangeably in this document) that are
>> > large and can change during the lifetime of a topology, such as
>> geo-location
>> > data, dictionaries, etc. Typical use cases include phrase recognition,
>> > entity extraction, document classification, URL re-writing,
>> location/address
>> > detection and so forth. Such files may be several KB to several GB in
>> size.
>> > For small datasets that don't need dynamic updates, including them in
>> the
>> > topology jar could be fine. But for large files, the startup times could
>> > become very large. In these cases, the distributed cache feature can
>> provide
>> > fast topology startup, especially if the files were previously
>> downloaded
>> > for the same submitter and are still in the cache. This is useful with
>> > frequent deployments, sometimes few times a day with updated jars,
>> because
>> > the large cached files will remain available without changes. The large
>> > cached blobs that do not change frequently will remain available in the
>> > distributed cache.
>> >
>> >
>> > We can look into this whether it is a common use case and how to
>> implement
>> > it in Flink.
>> >
>> > [1] http://storm.apache.org/releases/2.0.0-SNAPSHOT/distcache-
>> blobstore.html
>> >
>> >
>> > - Jark Wu
>> >
>> > 在 2016年8月23日,下午9:45,Lohith Samaga M <lohith.sam...@mphasis.com> 写道:
>> >
>> > Hi
>> > May be you could use Cassandra to store and fetch all such reference
>> data.
>> > This way the reference data can be updated without restarting your
>> > application.
>> >
>> > Lohith
>> >
>> > Sent from my Sony Xperia™ smartphone
>> >
>> >
>> >
>> > ---- Baswaraj Kasture wrote ----
>> >
>> > Thanks Kostas !
>> > I am using DataStream API.
>> >
>> > I have few config/property files (key vale text file) and also have
>> business
>> > rule files (json).
>> > These rules and configurations are needed when we process incoming
>> event.
>> > Is there any way to share them to task nodes from driver program ?
>> > I think this is very common use case and am sure other users may face
>> > similar issues.
>> >
>> > +Baswaraj
>> >
>> > On Mon, Aug 22, 2016 at 4:56 PM, Kostas Kloudas
>> > <k.klou...@data-artisans.com> wrote:
>> >>
>> >> Hello Baswaraj,
>> >>
>> >> Are you using the DataSet (batch) or the DataStream API?
>> >>
>> >> If you are in the first, you can use a broadcast variable for your
>> task.
>> >> If you are using the DataStream one, then there is no proper support
>> for
>> >> that.
>> >>
>> >> Thanks,
>> >> Kostas
>> >>
>> >> On Aug 20, 2016, at 12:33 PM, Baswaraj Kasture <kbaswar...@gmail.com>
>> >> wrote:
>> >>
>> >> Am running Flink standalone cluster.
>> >>
>> >> I have text file that need to be shared across tasks when i submit my
>> >> application.
>> >> in other words , put this text file in class path of running tasks.
>> >>
>> >> How can we achieve this with flink ?
>> >>
>> >> In spark, spark-submit has --jars option that puts all the files
>> specified
>> >> in class path of executors (executors run in separate JVM and spawned
>> >> dynamically, so it is possible).
>> >>
>> >> Flink's task managers run tasks in separate thread under taskmanager
>> JVM
>> >> (?) , how can we make this text file to be accessible on all tasks
>> spawned
>> >> by current application ?
>> >>
>> >> Using HDFS, NFS or including file in program jar is one way that i
>> know,
>> >> but am looking for solution that can allows me to provide text file at
>> run
>> >> time and still accessible in all tasks.
>> >> Thanks.
>> >>
>> >>
>> >
>> >
>> > Information transmitted by this e-mail is proprietary to Mphasis, its
>> > associated companies and/ or its customers and is intended
>> > for use only by the individual or entity to which it is addressed, and
>> may
>> > contain information that is privileged, confidential or
>> > exempt from disclosure under applicable law. If you are not the intended
>> > recipient or it appears that this mail has been forwarded
>> > to you without proper authority, you are notified that any use or
>> > dissemination of this information in any manner is strictly
>> > prohibited. In such cases, please notify us immediately at
>> > mailmas...@mphasis.com and delete this mail from your records.
>> >
>> >
>>
>
>

Reply via email to