+1 to make blink planner as the default planner for SQL Client, hence we can give the blink planner a bit more exposure.
Best, Hequn On Fri, Jan 3, 2020 at 6:32 PM Jark Wu <imj...@gmail.com> wrote: > Hi Benoît, > > Thanks for the reminder. I will look into the issue and hopefully we can > target it into 1.9.2 and 1.10. > > Cheers, > Jark > > On Fri, 3 Jan 2020 at 18:21, Benoît Paris < > benoit.pa...@centraliens-lille.org> wrote: > >> > If anyone finds that blink planner has any significant defects and >> has a larger regression than the old planner, please let us know. >> >> Overall, the Blink-exclusive features are must (TopN, deduplicate, >> LAST_VALUE, plan reuse, etc)! But all use cases of the Legacy planner in >> production are not covered: >> An edge case of Temporal Table Functions does not allow computed Tables >> (as opposed to TableSources) to be used on the query side in Blink ( >> https://issues.apache.org/jira/browse/FLINK-14200) >> >> Cheers >> Ben >> >> >> On Fri, Jan 3, 2020 at 10:00 AM Jeff Zhang <zjf...@gmail.com> wrote: >> >>> +1, I have already made blink as the default planner of flink >>> interpreter in Zeppelin >>> >>> >>> Jingsong Li <jingsongl...@gmail.com> 于2020年1月3日周五 下午4:37写道: >>> >>>> Hi Jark, >>>> >>>> +1 for default blink planner in SQL-CLI. >>>> I believe this new planner can be put into practice in production. >>>> We've worked hard for nearly a year, but the old planner didn't move on. >>>> >>>> And I'd like to cc to user@flink.apache.org. >>>> If anyone finds that blink planner has any significant defects and has >>>> a larger regression than the old planner, please let us know. We will be >>>> very grateful. >>>> >>>> Best, >>>> Jingsong Lee >>>> >>>> On Fri, Jan 3, 2020 at 4:14 PM Leonard Xu <xbjt...@gmail.com> wrote: >>>> >>>>> +1 for this. >>>>> We bring many SQL/API features and enhance stability in 1.10 release, >>>>> and almost all of them happens in Blink planner. >>>>> SQL CLI is the most convenient entrypoint for me, I believe many users >>>>> will have a better experience If we set Blink planner as default planner. >>>>> >>>>> Best, >>>>> Leonard >>>>> >>>>> > 在 2020年1月3日,15:16,Terry Wang <zjuwa...@gmail.com> 写道: >>>>> > >>>>> > Since what blink planner can do is a superset of flink planner, big >>>>> +1 for changing the default planner to Blink planner from my side. >>>>> > >>>>> > Best, >>>>> > Terry Wang >>>>> > >>>>> > >>>>> > >>>>> >> 2020年1月3日 15:00,Jark Wu <imj...@gmail.com> 写道: >>>>> >> >>>>> >> Hi everyone, >>>>> >> >>>>> >> In 1.10 release, Flink SQL supports many awesome features and >>>>> improvements, >>>>> >> including: >>>>> >> - support watermark statement and computed column in DDL >>>>> >> - fully support all data types in Hive >>>>> >> - Batch SQL performance improvements (TPC-DS 7x than Hive MR) >>>>> >> - support INSERT OVERWRITE and INSERT PARTITION >>>>> >> >>>>> >> However, all the features and improvements are only avaiable in >>>>> Blink >>>>> >> planner, not in Old planner. >>>>> >> There are also some other features are limited in Blink planner, >>>>> e.g. >>>>> >> Dimension Table Join [1], >>>>> >> TopN [2], Deduplicate [3], streaming aggregates optimization [4], >>>>> and so on. >>>>> >> >>>>> >> But Old planner is still the default planner in Table API & SQL. It >>>>> is >>>>> >> frustrating for users to set >>>>> >> to blink planner manually when every time start a SQL CLI. And it's >>>>> >> surprising to see unsupported >>>>> >> exception if they trying out the new features but not switch >>>>> planner. >>>>> >> >>>>> >> SQL CLI is a very important entrypoint for trying out new feautures >>>>> and >>>>> >> prototyping for users. >>>>> >> In order to give new planner more exposures, I would like to >>>>> suggest to set >>>>> >> default planner >>>>> >> for SQL Client to Blink planner before 1.10 release. >>>>> >> >>>>> >> The approach is just changing the default SQL CLI yaml >>>>> configuration[5]. In >>>>> >> this way, the existing >>>>> >> environment is still compatible and unaffected. >>>>> >> >>>>> >> Changing the default planner for the whole Table API & SQL is >>>>> another topic >>>>> >> and is out of scope of this discussion. >>>>> >> >>>>> >> What do you think? >>>>> >> >>>>> >> Best, >>>>> >> Jark >>>>> >> >>>>> >> [1]: >>>>> >> >>>>> https://ci.apache.org/projects/flink/flink-docs-master/dev/table/streaming/joins.html#join-with-a-temporal-table >>>>> >> [2]: >>>>> >> >>>>> https://ci.apache.org/projects/flink/flink-docs-master/dev/table/sql/queries.html#top-n >>>>> >> [3]: >>>>> >> >>>>> https://ci.apache.org/projects/flink/flink-docs-master/dev/table/sql/queries.html#deduplication >>>>> >> [4]: >>>>> >> >>>>> https://ci.apache.org/projects/flink/flink-docs-master/dev/table/tuning/streaming_aggregation_optimization.html >>>>> >> [5]: >>>>> >> >>>>> https://github.com/apache/flink/blob/master/flink-table/flink-sql-client/conf/sql-client-defaults.yaml#L100 >>>>> > >>>>> >>>>> >>>> >>>> -- >>>> Best, Jingsong Lee >>>> >>> >>> >>> -- >>> Best Regards >>> >>> Jeff Zhang >>> >> >> >> -- >> Benoît Paris >> Ingénieur Machine Learning Explicable >> Tél : +33 6 60 74 23 00 >> http://benoit.paris >> http://explicable.ml >> >