Leonard Xu created FLINK-14549: ---------------------------------- Summary: Bring more detail by using logicalType rather than conversionClass in exception msg Key: FLINK-14549 URL: https://issues.apache.org/jira/browse/FLINK-14549 Project: Flink Issue Type: Bug Components: Table SQL / Planner Affects Versions: 1.9.1 Reporter: Leonard Xu Fix For: 1.10.0
We use DataType‘s conversionClass name in validating the query result's field type and sink table schema which is no precise when the DataType has variable parameters like DECIMAL(p,s)、TIMESTAMP(p). Exception in thread "main" org.apache.flink.table.api.ValidationException: Field types of query result and registered TableSink `default_catalog`.`default_database`.`q2_sinkTable` do not match.Exception in thread "main" org.apache.flink.table.api.ValidationException: Field types of query result and registered TableSink `default_catalog`.`default_database`.`q2_sinkTable` do not match.Query result schema: [d_week_seq1: Long, EXPR$1: BigDecimal, EXPR$2: BigDecimal, EXPR$3: BigDecimal]TableSink schema: [d_week_seq1: Long, EXPR$1: BigDecimal, EXPR$2: BigDecimal, EXPR$3: BigDecimal] at org.apache.flink.table.planner.sinks.TableSinkUtils$.validateSink(TableSinkUtils.scala:68) at org.apache.flink.table.planner.delegation.PlannerBase$$anonfun$2.apply(PlannerBase.scala:179) at org.apache.flink.table.planner.delegation.PlannerBase$$anonfun$2.apply(PlannerBase.scala:178) -- This message was sent by Atlassian Jira (v8.3.4#803005)