Hi Alex, welcome to the Flink community! Right now, there is no way to specify a Flink program without writing code (Java, Scala, Python(beta)).
In principle it is possible to put such functionality on top of the DataStream or DataSet APIs. This has been done before for other programming APIs (Flink's own libraries Table API, Gelly, FlinkML, and externals Apache Beam / Google DataFlow, Mahout, Cascading, ...). However, all of these are again programming APIs, some specialized for certain use-cases. Specifying Flink programs by config files (or graphically) would require a data model, a DataStream/DataSet program generator and probably a code generation component. Best, Fabian 2016-04-22 18:41 GMT+02:00 Alexander Smirnov <alexander.smirn...@gmail.com>: > Hi guys! > > I’m new to Flink, and actually to this mailing list as well :) this is my > first message. > I’m still reading the documentation and I would say Flink is an amazing > system!! Thanks everybody who participated in the development! > > The information I didn’t find in the documentation - if it is possible to > describe data(stream) transformation without any code (Java/Scala). > I mean if it is possible to describe datasource functions, all of the > operators, connections between them, and sinks in a plain text > configuration file and then feed it to Flink. > In this case it would be possible to change data flow without > recompilation/redeployment. > > Is there a similar functionality in Flink? May be some third party plugin? > > Thank you, > Alex