Hi Fabian,

The priority seems reasonable to me. I skimmed through your TPCH branch and
find data type and related functions would be quite important to enable
most of TPC-H queries.

I would try out your TPC-H branch and pick up some issues you filed these
two days for the 1.1.0 release.

Hope this helps :)

Best,
Yijie

On Thu, May 19, 2016 at 11:59 PM, Fabian Hueske <fhue...@gmail.com> wrote:

> Hi everybody,
>
> I'd like to start a discussion about blocking issues and outstanding
> features of the Table API and SQL for the 1.1.0 release. As you probably
> know, the Table API was completely reworked and ported to Apache Calcite.
> Moreover, we added initial support for SQL on batch and streaming tables.
>
> We have come quite far but there are still a couple of issue that need to
> be resolved before we can release a new version of Flink. I would like to
> start collecting and prioritizing issues such that we can work towards a
> feature set that we would like to be included in the next release. In order
> to prepare this list, I tried to execute the TPC-H query set using the
> currently supported SQL feature set. Only one (Q18) out  of the 22 queries
> could be executed. The others failed due to unsupported features or bugs.
>
> In the following, I list issues ordered by priority that I think need be
> resolved for the release.
>
>     - FLINK-3728:  Detect unsupported operators and improve error messages.
> While we can effectively prevent unsupported operations in the Table API,
> this is not easily possible with SQL queries. At the moment, unsupported
> operations are either not detected and translated into invalid plans or
> throw a hard to understand exceptions.
>     - FLINK-3859: Add support for DECIMAL. Without this feature, it is not
> possible to use floating point literals in SQL queries.
>     - FLINK-3152 / FLINK-3580: Add support for date types and date
> functions.
>     - FLINK-3586: Prevent AVG(LONG) overflow by using BigInteger as
> intermediate data type.
>     - FLINK-2971: Add support for outer joins (a PR for this issue exists
> #1981)
>     - FLINK-3936 : Add MIN / MAX aggregation function for BOOLEAN types
>     - FLINK-3916: Add support for generic types which are handled by
>     - FLINK-3723: This is an proposal to split the Table API select()
> method into select() for projection and aggregate() for aggregations. At
> the moment, both are handled by select() (such as in SQL) and internally
> separated by the Table API. We should decide for Flink 1.1.0 whether to
> implement the proposal or not.
>     - FLINK-3871 / FLINK-3873: Add Table Source and TableSink for Avro
> encoded Kafka sources
>     - FLINK-3872 / FLINK-3874 : Add TableSource and TableSink for JSON
> encoded Kafka sources
>     - More TableSource / TableSinks
>
> Please review this list, add issues that you think should go in as well,
> and discuss the priorities of the features.
> Also if you would like to get involved with improving the Table API / SQL,
> drop a mail to the mailing list or a comment to a JIRA issue.
>
> I think it would be good if somebody would coordinate these efforts. I
> would be happy to do it. However, I will leave in one month for a
> two-months parental leave and I don't know how much I can contribute in
> that time. So if somebody would like to step up and help coordinating,
> please let me and the others know.
>
> Cheers, Fabian
>

Reply via email to