DuBin created FLINK-18414: ----------------------------- Summary: Kafka Json connector in Table API support more option Key: FLINK-18414 URL: https://issues.apache.org/jira/browse/FLINK-18414 Project: Flink Issue Type: Improvement Components: Connectors / Kafka, Formats (JSON, Avro, Parquet, ORC, SequenceFile), Table SQL / Ecosystem Affects Versions: 1.10.1 Reporter: DuBin
Currently, the Flink use a 'org.apache.flink.formats.json.JsonRowDeserializationSchema' to deserialize the record into Row if we define the Kafka Json Table Source. But the parser is hard-coded in the class : private final ObjectMapper objectMapper = new ObjectMapper(); Imagine that the Json data source contains data like this: {"a":NaN,"b":1.2} or it contains some dirty data, it will throw exception in the deserialize function all the time, because Kafka do not have a schema validation on Json format. So can we add more options in the 'org.apache.flink.formats.json.JsonRowFormatFactory' , in the 'org.apache.flink.formats.json.JsonRowFormatFactory#createDeserializationSchema'? e.g. add more option for the objectMapper, some dirty data handler(just return an empty row, defined by the user) -- This message was sent by Atlassian Jira (v8.3.4#803005)