Dear Flink Community,
I am currently in the process of upgrading our Flink cluster from *version
1.18.0 to 1.20.1*. The cluster itself is functioning correctly
post-upgrade, and I am able to deploy Flink jobs successfully. However, I
have encountered an issue when attempting to restore a job using a *savepoint
or state taken from Flink 1.18.0*.
*Issue Description*
-
When deploying the Flink job to the *Flink 1.20.1 cluster* using a
*savepoint
from Flink 1.18.0*, the job is assigned *only one Kafka partition
(partition 0)*. As a result, messages from the other partitions are not
being consumed.
-
However, if I deploy the same job *without a savepoint*, the job
correctly assigns all three partitions (*0, 1, 2*) and consumes messages
as expected.
I have researched this issue extensively but have not found a clear
explanation. I would appreciate any guidance on the following queries:
1.
*Is this issue related to the compatibility of savepoint restoration
between Flink 1.18.0 and Flink 1.20.1?*
2.
*Is this behavior a known bug or an expected outcome?*
3.
*If this is a bug, what are the recommended steps to resolve it?*
-
Are there any configuration changes required to properly restore
partitions?
-
Would fixing this require modifications in the application code?
Your insights and assistance on this matter would be highly appreciated.
Thanks & Regards
Jasvendra Kumar