Hi all, We are using flink-1.17.0 table API and RocksDB as backend to provide a service to our users to run sql queries. The tables are created using the avro schema and when the schema is changed in a compatible manner i.e adding a field with default, we are unable to recover the job from the savepoint. This is mentioned in the flink doc on evolution [1] as well.
Are there any plans to support schema evolution in the table API? Our current approach involves rebuilding the entire state by discarding the output and then utilizing that state in the actual job. This is already done for table-store [2] [1] https://nightlies.apache.org/flink/flink-docs-release-1.17/docs/dev/table/concepts/overview/#stateful-upgrades-and-evolution [2] https://cwiki.apache.org/confluence/display/FLINK/FLIP-226%3A+Introduce+Schema+Evolution+on+Table+Store