Hi everyone, SQL DDL support has been a long-time ask from the community. Current Flink SQL support only DML (e.g. SELECT and INSERT statements). In its current form, Flink SQL users still need to define/create table sources and sinks programmatically in Java/Scala. Also, in SQL Client, without DDL support, the current implementation does not allow dynamical creation of table, type or functions with SQL, this adds friction for its adoption.
I drafted a design doc [1] with a few other community members that proposes the design and implementation for adding DDL support in Flink. The initial design considers DDL for table, view, type, library and function. It will be great to get feedback on the design from the community, and align with latest effort in unified SQL connector API [2] and Flink Hive integration [3]. Any feedback is highly appreciated. Thanks Shuyi Chen [1] https://docs.google.com/document/d/1TTP-GCC8wSsibJaSUyFZ_5NBAHYEB1FVmPpP7RgDGBA/edit?usp=sharing [2] https://docs.google.com/document/d/1Yaxp1UJUFW-peGLt8EIidwKIZEWrrA-pznWLuvaH39Y/edit?usp=sharing [3] https://docs.google.com/document/d/1SkppRD_rE3uOKSN-LuZCqn4f7dz0zW5aa6T_hBZq5_o/edit?usp=sharing -- "So you have to trust that the dots will somehow connect in your future."