Hello Baswaraj, Are you using the DataSet (batch) or the DataStream API?
If you are in the first, you can use a broadcast variable <https://ci.apache.org/projects/flink/flink-docs-master/apis/batch/index.html#broadcast-variables> for your task. If you are using the DataStream one, then there is no proper support for that. Thanks, Kostas > On Aug 20, 2016, at 12:33 PM, Baswaraj Kasture <kbaswar...@gmail.com> wrote: > > Am running Flink standalone cluster. > > I have text file that need to be shared across tasks when i submit my > application. > in other words , put this text file in class path of running tasks. > > How can we achieve this with flink ? > > In spark, spark-submit has --jars option that puts all the files specified in > class path of executors (executors run in separate JVM and spawned > dynamically, so it is possible). > > Flink's task managers run tasks in separate thread under taskmanager JVM (?) > , how can we make this text file to be accessible on all tasks spawned by > current application ? > > Using HDFS, NFS or including file in program jar is one way that i know, but > am looking for solution that can allows me to provide text file at run time > and still accessible in all tasks. > Thanks.