Yes. I think the argument is fairly valid - we can always adjust the API in the future, in fact most of the APIs are labeled publicEvolving at this moment. I was only trying to provide the info, that the interfaces in flink-ml-api might change in the near future, for others when voting.
In fact, I am actually always +1 on moving flink-ml-api to /opt :-) Regarding the Python ML API. sorry for not noticing it earlier as I haven't given it a deep look yet. will do very soon! -- Rong On Sun, Feb 9, 2020 at 7:33 PM Hequn Cheng <he...@apache.org> wrote: > Hi Rong, > > Thanks a lot for joining the discussion! > > It would be great if we can have a long term plan. My intention is to > provide a way for users to add dependencies of Flink ML, either through the > opt or download page. This would be more and more critical along with the > improvement of the Flink ML, as you said there are multiple PRs under > review and I'm also going to support Python Pipeline API recently[1]. > > Meanwhile, it also makes sense to include the API into the opt, so it > would probably not break the long term plan. > However, even find something wrong in the future, we can revisit this > easily instead of blocking the improvement for users. What do you think? > > Best, > Hequn > > [1] > http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/DISCUSS-Support-Python-ML-Pipeline-API-td37291.html > > On Sat, Feb 8, 2020 at 1:57 AM Rong Rong <walter...@gmail.com> wrote: > >> CC @Xu Yang <xuyang1...@gmail.com> >> >> Thanks for starting the discussion @Hequn Cheng <chenghe...@gmail.com> and >> sorry for joining the discussion late. >> >> I've mainly helped merging the code in flink-ml-api and flink-ml-lib in >> the past several months. >> IMO the flink-ml-api are an extension on top of the table API and agree >> that it should be treated as a part of the "core" core. >> >> However, I think given the fact that there are multiple PRs still under >> review [1], is it a better idea to come up with a long term plan first >> before make the decision to moving it to /opt now? >> >> >> -- >> Rong >> >> [1] >> https://github.com/apache/flink/pulls?utf8=%E2%9C%93&q=is%3Apr+is%3Aopen+label%3Acomponent%3DLibrary%2FMachineLearning+ >> >> On Fri, Feb 7, 2020 at 5:54 AM Hequn Cheng <he...@apache.org> wrote: >> >>> Hi, >>> >>> @Till Rohrmann <trohrm...@apache.org> Thanks for the great inputs. I >>> agree >>> with you that we should have a long term plan for this. It definitely >>> deserves another discussion. >>> @Jeff Zhang <zjf...@gmail.com> Thanks for your reports and ideas. It's a >>> good idea to improve the error messages. Do we have any JIRAs for it or >>> maybe we can create one for it. >>> >>> Thank you again for your feedback and suggestions. I will go on with the >>> PR. Thanks! >>> >>> Best, >>> Hequn >>> >>> On Thu, Feb 6, 2020 at 11:51 PM Jeff Zhang <zjf...@gmail.com> wrote: >>> >>> > I have another concern which may not be closely related to this thread. >>> > Since flink doesn't include all the necessary jars, I think it is >>> critical >>> > for flink to display meaningful error message when any class is >>> missing. >>> > e.g. Here's the error message when I use kafka but miss >>> > including flink-json. To be honest, the kind of error message is hard >>> to >>> > understand for new users. >>> > >>> > >>> > Reason: No factory implements >>> > 'org.apache.flink.table.factories.DeserializationSchemaFactory'. The >>> > following properties are requested: >>> > connector.properties.bootstrap.servers=localhost:9092 >>> > connector.properties.group.id=testGroup >>> > connector.properties.zookeeper.connect=localhost:2181 >>> > connector.startup-mode=earliest-offset connector.topic=generated.events >>> > connector.type=kafka connector.version=universal format.type=json >>> > schema.0.data-type=VARCHAR(2147483647) schema.0.name=status >>> > schema.1.data-type=VARCHAR(2147483647) schema.1.name=direction >>> > schema.2.data-type=BIGINT schema.2.name=event_ts update-mode=append >>> The >>> > following factories have been considered: >>> > org.apache.flink.table.catalog.hive.factories.HiveCatalogFactory >>> > org.apache.flink.table.module.hive.HiveModuleFactory >>> > org.apache.flink.table.module.CoreModuleFactory >>> > org.apache.flink.table.catalog.GenericInMemoryCatalogFactory >>> > org.apache.flink.table.sources.CsvBatchTableSourceFactory >>> > org.apache.flink.table.sources.CsvAppendTableSourceFactory >>> > org.apache.flink.table.sinks.CsvBatchTableSinkFactory >>> > org.apache.flink.table.sinks.CsvAppendTableSinkFactory >>> > org.apache.flink.table.planner.delegation.BlinkPlannerFactory >>> > org.apache.flink.table.planner.delegation.BlinkExecutorFactory >>> > org.apache.flink.table.planner.StreamPlannerFactory >>> > org.apache.flink.table.executor.StreamExecutorFactory >>> > >>> org.apache.flink.streaming.connectors.kafka.KafkaTableSourceSinkFactory at >>> > >>> > >>> org.apache.flink.table.factories.TableFactoryService.filterByFactoryClass(TableFactoryService.java:238) >>> > at >>> > >>> > >>> org.apache.flink.table.factories.TableFactoryService.filter(TableFactoryService.java:185) >>> > at >>> > >>> > >>> org.apache.flink.table.factories.TableFactoryService.findSingleInternal(TableFactoryService.java:143) >>> > at >>> > >>> > >>> org.apache.flink.table.factories.TableFactoryService.find(TableFactoryService.java:113) >>> > at >>> > >>> > >>> org.apache.flink.streaming.connectors.kafka.KafkaTableSourceSinkFactoryBase.getDeserializationSchema(KafkaTableSourceSinkFactoryBase.java:277) >>> > at >>> > >>> > >>> org.apache.flink.streaming.connectors.kafka.KafkaTableSourceSinkFactoryBase.createStreamTableSource(KafkaTableSourceSinkFactoryBase.java:161) >>> > at >>> > >>> > >>> org.apache.flink.table.factories.StreamTableSourceFactory.createTableSource(StreamTableSourceFactory.java:49) >>> > at >>> > >>> > >>> org.apache.flink.table.factories.TableFactoryUtil.findAndCreateTableSource(TableFactoryUtil.java:53) >>> > ... 36 more >>> > >>> > >>> > >>> > Till Rohrmann <trohrm...@apache.org> 于2020年2月6日周四 下午11:30写道: >>> > >>> > > I would not object given that it is rather small at the moment. >>> However, >>> > I >>> > > also think that we should have a plan how to handle the ever growing >>> > Flink >>> > > ecosystem and how to make it easily accessible to our users. E.g. >>> one far >>> > > fetched idea could be something like a configuration script which >>> > downloads >>> > > the required components for the user. But this deserves definitely a >>> > > separate discussion and does not really belong here. >>> > > >>> > > Cheers, >>> > > Till >>> > > >>> > > On Thu, Feb 6, 2020 at 3:35 PM Hequn Cheng <he...@apache.org> wrote: >>> > > >>> > > > >>> > > > Hi everyone, >>> > > > >>> > > > Thank you all for the great inputs! >>> > > > >>> > > > I think probably what we all agree on is we should try to make a >>> leaner >>> > > > flink-dist. However, we may also need to do some compromises >>> > considering >>> > > > the user experience that users don't need to download the >>> dependencies >>> > > from >>> > > > different places. Otherwise, we can move all the jars in the >>> current >>> > opt >>> > > > folder to the download page. >>> > > > >>> > > > The missing of clear rules for guiding such compromises makes >>> things >>> > more >>> > > > complicated now. I would agree that the decisive factor for what >>> goes >>> > > into >>> > > > Flink's binary distribution should be how core it is to Flink. >>> > Meanwhile, >>> > > > it's better to treat Flink API as a (core) core to Flink. Not only >>> it >>> > is >>> > > a >>> > > > very clear rule that easy to be followed but also in most cases, >>> API is >>> > > > very significant and deserved to be included in the dist. >>> > > > >>> > > > Given this, it might make sense to put flink-ml-api and >>> flink-ml-lib >>> > into >>> > > > the opt. >>> > > > What do you think? >>> > > > >>> > > > Best, >>> > > > Hequn >>> > > > >>> > > > On Wed, Feb 5, 2020 at 12:39 AM Chesnay Schepler < >>> ches...@apache.org> >>> > > > wrote: >>> > > > >>> > > >> Around a year ago I started a discussion >>> > > >> < >>> > > >>> > >>> http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/DISCUSS-Towards-a-leaner-flink-dist-tp25615.html >>> > > > >>> > > >> on reducing the amount of jars we ship with the distribution. >>> > > >> >>> > > >> While there was no definitive conclusion there was a shared >>> sentiment >>> > > >> that APIs should be shipped with the distribution. >>> > > >> >>> > > >> On 04/02/2020 17:25, Till Rohrmann wrote: >>> > > >> >>> > > >> I think there is no such rule that APIs go automatically into >>> opt/ and >>> > > >> "libraries" not. The contents of opt/ have mainly grown over time >>> w/o >>> > > >> following a strict rule. >>> > > >> >>> > > >> I think the decisive factor for what goes into Flink's binary >>> > > distribution >>> > > >> should be how core it is to Flink. Of course another important >>> > > >> consideration is which use cases Flink should promote "out of the >>> box" >>> > > (not >>> > > >> sure whether this is actual true for content shipped in opt/ >>> because >>> > you >>> > > >> also have to move it to lib). >>> > > >> >>> > > >> For example, Gelly would be an example which I would rather see >>> as an >>> > > >> optional component than shipping it with every Flink binary >>> > > distribution. >>> > > >> >>> > > >> Cheers, >>> > > >> Till >>> > > >> >>> > > >> On Tue, Feb 4, 2020 at 11:24 AM Becket Qin <becket....@gmail.com> >>> < >>> > > becket....@gmail.com> wrote: >>> > > >> >>> > > >> >>> > > >> Thanks for the suggestion, Till. >>> > > >> >>> > > >> I am curious about how do we usually decide when to put the jars >>> into >>> > > the >>> > > >> opt folder? >>> > > >> >>> > > >> Technically speaking, it seems that `flink-ml-api` should be put >>> into >>> > > the >>> > > >> opt directory because they are actually API instead of libraries, >>> just >>> > > like >>> > > >> CEP and Table. >>> > > >> >>> > > >> `flink-ml-lib` seems to be on the border. On one hand, it is a >>> > library. >>> > > On >>> > > >> the other hand, unlike SQL formats and Hadoop whose major code are >>> > > outside >>> > > >> of Flink, the algorithm codes are in Flink. So `flink-ml-lib` is >>> more >>> > > like >>> > > >> those of built-in SQL UDFs. So it seems fine to either put it in >>> the >>> > opt >>> > > >> folder or in the downloads page. >>> > > >> >>> > > >> From the user experience perspective, it might be better to have >>> both >>> > > >> `flink-ml-lib` and `flink-ml-api` in opt folder so users needn't >>> go to >>> > > two >>> > > >> places for the required dependencies. >>> > > >> >>> > > >> Thanks, >>> > > >> >>> > > >> Jiangjie (Becket) Qin >>> > > >> >>> > > >> On Tue, Feb 4, 2020 at 2:32 PM Hequn Cheng <he...@apache.org> < >>> > > he...@apache.org> wrote: >>> > > >> >>> > > >> >>> > > >> Hi Till, >>> > > >> >>> > > >> Thanks a lot for your suggestion. It's a good idea to offer the >>> > flink-ml >>> > > >> libraries as optional dependencies on the download page which can >>> make >>> > > >> >>> > > >> the >>> > > >> >>> > > >> dist smaller. >>> > > >> >>> > > >> But I also have some concerns for it, e.g., the download page now >>> only >>> > > >> includes the latest 3 releases. We may need to find ways to >>> support >>> > more >>> > > >> versions. >>> > > >> On the other hand, the size of the flink-ml libraries now is very >>> > > >> small(about 246K), so it would not bring much impact on the size >>> of >>> > > dist. >>> > > >> >>> > > >> What do you think? >>> > > >> >>> > > >> Best, >>> > > >> Hequn >>> > > >> >>> > > >> On Mon, Feb 3, 2020 at 6:24 PM Till Rohrmann < >>> trohrm...@apache.org> < >>> > > trohrm...@apache.org> >>> > > >> >>> > > >> wrote: >>> > > >> >>> > > >> An alternative solution would be to offer the flink-ml libraries >>> as >>> > > >> optional dependencies on the download page. Similar to how we >>> offer >>> > the >>> > > >> different SQL formats and Hadoop releases [1]. >>> > > >> >>> > > >> [1] https://flink.apache.org/downloads.html >>> > > >> >>> > > >> Cheers, >>> > > >> Till >>> > > >> >>> > > >> On Mon, Feb 3, 2020 at 10:19 AM Hequn Cheng <he...@apache.org> < >>> > > he...@apache.org> wrote: >>> > > >> >>> > > >> >>> > > >> Thank you all for your feedback and suggestions! >>> > > >> >>> > > >> Best, Hequn >>> > > >> >>> > > >> On Mon, Feb 3, 2020 at 5:07 PM Becket Qin <becket....@gmail.com> >>> < >>> > > becket....@gmail.com> >>> > > >> >>> > > >> wrote: >>> > > >> >>> > > >> Thanks for bringing up the discussion, Hequn. >>> > > >> >>> > > >> +1 on adding `flink-ml-api` and `flink-ml-lib` into opt. This >>> would >>> > > >> >>> > > >> make >>> > > >> >>> > > >> it much easier for the users to try out some simple ml tasks. >>> > > >> >>> > > >> Thanks, >>> > > >> >>> > > >> Jiangjie (Becket) Qin >>> > > >> >>> > > >> On Mon, Feb 3, 2020 at 4:34 PM jincheng sun < >>> > > >> >>> > > >> sunjincheng...@gmail.com >>> > > >> >>> > > >> wrote: >>> > > >> >>> > > >> >>> > > >> Thank you for pushing forward @Hequn Cheng <he...@apache.org> < >>> > > he...@apache.org> ! >>> > > >> >>> > > >> Hi @Becket Qin <becket....@gmail.com> <becket....@gmail.com> , >>> Do >>> > you >>> > > have any concerns >>> > > >> >>> > > >> on >>> > > >> >>> > > >> this ? >>> > > >> >>> > > >> Best, >>> > > >> Jincheng >>> > > >> >>> > > >> Hequn Cheng <he...@apache.org> <he...@apache.org> 于2020年2月3日周一 >>> > > 下午2:09写道: >>> > > >> >>> > > >> >>> > > >> Hi everyone, >>> > > >> >>> > > >> Thanks for the feedback. As there are no objections, I've opened a >>> > > >> >>> > > >> JIRA >>> > > >> >>> > > >> issue(FLINK-15847[1]) to address this issue. >>> > > >> The implementation details can be discussed in the issue or in the >>> > > >> following PR. >>> > > >> >>> > > >> Best, >>> > > >> Hequn >>> > > >> >>> > > >> [1] https://issues.apache.org/jira/browse/FLINK-15847 >>> > > >> >>> > > >> On Wed, Jan 8, 2020 at 9:15 PM Hequn Cheng <chenghe...@gmail.com> >>> < >>> > > chenghe...@gmail.com> >>> > > >> >>> > > >> wrote: >>> > > >> >>> > > >> Hi Jincheng, >>> > > >> >>> > > >> Thanks a lot for your feedback! >>> > > >> Yes, I agree with you. There are cases that multi jars need to >>> > > >> >>> > > >> be >>> > > >> >>> > > >> uploaded. I will prepare another discussion later. Maybe with a >>> > > >> >>> > > >> simple >>> > > >> >>> > > >> design doc. >>> > > >> >>> > > >> Best, Hequn >>> > > >> >>> > > >> On Wed, Jan 8, 2020 at 3:06 PM jincheng sun < >>> > > >> >>> > > >> sunjincheng...@gmail.com> >>> > > >> >>> > > >> wrote: >>> > > >> >>> > > >> >>> > > >> Thanks for bring up this discussion Hequn! >>> > > >> >>> > > >> +1 for include `flink-ml-api` and `flink-ml-lib` in opt. >>> > > >> >>> > > >> BTW: I think would be great if bring up a discussion for upload >>> > > >> >>> > > >> multiple >>> > > >> >>> > > >> Jars at the same time. as PyFlink JOB also can have the benefit >>> > > >> >>> > > >> if >>> > > >> >>> > > >> we >>> > > >> >>> > > >> do >>> > > >> >>> > > >> that improvement. >>> > > >> >>> > > >> Best, >>> > > >> Jincheng >>> > > >> >>> > > >> >>> > > >> Hequn Cheng <chenghe...@gmail.com> <chenghe...@gmail.com> >>> > 于2020年1月8日周三 >>> > > 上午11:50写道: >>> > > >> >>> > > >> >>> > > >> Hi everyone, >>> > > >> >>> > > >> FLIP-39[1] rebuilds Flink ML pipeline on top of TableAPI >>> > > >> >>> > > >> which >>> > > >> >>> > > >> moves >>> > > >> >>> > > >> Flink >>> > > >> >>> > > >> ML a step further. Base on it, users can develop their ML >>> > > >> >>> > > >> jobs >>> > > >> >>> > > >> and >>> > > >> >>> > > >> more >>> > > >> >>> > > >> and >>> > > >> >>> > > >> more machine learning platforms are providing ML services. >>> > > >> >>> > > >> However, the problem now is the jars of flink-ml-api and >>> > > >> >>> > > >> flink-ml-lib >>> > > >> >>> > > >> are >>> > > >> >>> > > >> only exist on maven repo. Whenever users want to submit ML >>> > > >> >>> > > >> jobs, >>> > > >> >>> > > >> they >>> > > >> >>> > > >> can >>> > > >> >>> > > >> only depend on the ml modules and package a fat jar. This >>> > > >> >>> > > >> would be >>> > > >> >>> > > >> inconvenient especially for the machine learning platforms on >>> > > >> >>> > > >> which >>> > > >> >>> > > >> nearly >>> > > >> >>> > > >> all jobs depend on Flink ML modules and have to package a fat >>> > > >> >>> > > >> jar. >>> > > >> >>> > > >> Given this, it would be better to include jars of >>> > > >> >>> > > >> flink-ml-api >>> > > >> >>> > > >> and >>> > > >> >>> > > >> flink-ml-lib in the `opt` folder, so that users can directly >>> > > >> >>> > > >> use >>> > > >> >>> > > >> the >>> > > >> >>> > > >> jars >>> > > >> >>> > > >> with the binary release. For example, users can move the jars >>> > > >> >>> > > >> into >>> > > >> >>> > > >> the >>> > > >> >>> > > >> `lib` folder or use -j to upload the jars. (Currently, -j >>> > > >> >>> > > >> only >>> > > >> >>> > > >> support >>> > > >> >>> > > >> upload one jar. Supporting multi jars for -j can be discussed >>> > > >> >>> > > >> in >>> > > >> >>> > > >> another >>> > > >> >>> > > >> discussion.) >>> > > >> >>> > > >> Putting the jars in the `opt` folder instead of the `lib` >>> > > >> >>> > > >> folder >>> > > >> >>> > > >> is >>> > > >> >>> > > >> because >>> > > >> >>> > > >> currently, the ml jars are still optional for the Flink >>> > > >> >>> > > >> project by >>> > > >> >>> > > >> default. >>> > > >> >>> > > >> What do you think? Welcome any feedback! >>> > > >> >>> > > >> Best, >>> > > >> >>> > > >> Hequn >>> > > >> >>> > > >> [1] >>> > > >> >>> > > >> >>> > > >> >>> > > >> >>> > > >>> > >>> https://cwiki.apache.org/confluence/display/FLINK/FLIP-39+Flink+ML+pipeline+and+ML+libs >>> > > >> >>> > > >> >>> > > >> >>> > > >>> > >>> > >>> > -- >>> > Best Regards >>> > >>> > Jeff Zhang >>> > >>> >>