Hi Alexander, since I was looking for something similar some days ago here is what I know about this argument: during the Stratosphere project there was Meteor and Supremo allowing that [1] but then it was dismissed in favour of Pig integration that I don't wheter it was ever completed. Yiu could give a try to Piglet project[2] that allows to use PIG with Spark and Flink but I don't know how well it works (Flink integration is also very recent and not documented anywhere).
Best, Flavio [1] http://stratosphere.eu/assets/papers/Sopremo_Meteor%20BigData.pdf [2] https://github.com/ksattler/piglet On 23 Apr 2016 07:48, "Aljoscha Krettek" <aljos...@apache.org> wrote: Hi, I think if the Table API/SQL API evolves enough it should be able to supply a Flink program as just an SQL query with source/sink definitions. Hopefully, in the future. :-) Cheers, Aljoscha On Fri, 22 Apr 2016 at 23:10 Fabian Hueske <fhue...@gmail.com> wrote: > Hi Alex, > > welcome to the Flink community! > Right now, there is no way to specify a Flink program without writing code > (Java, Scala, Python(beta)). > > In principle it is possible to put such functionality on top of the > DataStream or DataSet APIs. > This has been done before for other programming APIs (Flink's own > libraries Table API, Gelly, FlinkML, and externals Apache Beam / Google > DataFlow, Mahout, Cascading, ...). However, all of these are again > programming APIs, some specialized for certain use-cases. > > Specifying Flink programs by config files (or graphically) would require a > data model, a DataStream/DataSet program generator and probably a code > generation component. > > Best, Fabian > > 2016-04-22 18:41 GMT+02:00 Alexander Smirnov <alexander.smirn...@gmail.com > >: > >> Hi guys! >> >> I’m new to Flink, and actually to this mailing list as well :) this is my >> first message. >> I’m still reading the documentation and I would say Flink is an amazing >> system!! Thanks everybody who participated in the development! >> >> The information I didn’t find in the documentation - if it is possible to >> describe data(stream) transformation without any code (Java/Scala). >> I mean if it is possible to describe datasource functions, all of the >> operators, connections between them, and sinks in a plain text >> configuration file and then feed it to Flink. >> In this case it would be possible to change data flow without >> recompilation/redeployment. >> >> Is there a similar functionality in Flink? May be some third party plugin? >> >> Thank you, >> Alex > > >