It looks like you were trying to resume from a checkpoint taken with the FsStateBackend into a revised version of the job that uses the RocksDbStateBackend. Switching state backends in this way is not supported: checkpoints and savepoints are written in a state-backend-specific format, and can only be read by the same backend that wrote them.
It is possible, however, to migrate between state backends using the State Processor API [1]. [1] https://ci.apache.org/projects/flink/flink-docs-stable/dev/libs/state_processor_api.html Best, David On Fri, Oct 2, 2020 at 4:07 PM 大森林 <appleyu...@foxmail.com> wrote: > *I want to do an experiment of"incremental checkpoint"* > > my code is: > > https://paste.ubuntu.com/p/DpTyQKq6Vk/ > > > > pom.xml is: > > <?xml version="1.0" encoding="UTF-8"?> > <project xmlns="http://maven.apache.org/POM/4.0.0" > xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" > xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 > http://maven.apache.org/xsd/maven-4.0.0.xsd"> > <modelVersion>4.0.0</modelVersion> > > <groupId>example</groupId> > <artifactId>datastream_api</artifactId> > <version>1.0-SNAPSHOT</version> > <build> > <plugins> > <plugin> > <groupId>org.apache.maven.plugins</groupId> > <artifactId>maven-compiler-plugin</artifactId> > <version>3.1</version> > <configuration> > <source>1.8</source> > <target>1.8</target> > </configuration> > </plugin> > > <plugin> > <groupId>org.scala-tools</groupId> > <artifactId>maven-scala-plugin</artifactId> > <version>2.15.2</version> > <executions> > <execution> > <goals> > <goal>compile</goal> > <goal>testCompile</goal> > </goals> > </execution> > </executions> > </plugin> > > > > </plugins> > </build> > > <dependencies> > > <!-- > https://mvnrepository.com/artifact/org.apache.flink/flink-streaming-scala > --> > <dependency> > <groupId>org.apache.flink</groupId> > <artifactId>flink-streaming-scala_2.11</artifactId> > <version>1.11.1</version> > <!-<scope>provided</scope>-> > </dependency> > > <!-<dependency>-> > <!-<groupId>org.apache.flink</groupId>-> > <!-<artifactId>flink-streaming-java_2.12</artifactId>-> > <!-<version>1.11.1</version>-> > <!-<!–<scope>compile</scope>–>-> > <!-</dependency>-> > > <dependency> > <groupId>org.apache.flink</groupId> > <artifactId>flink-clients_2.11</artifactId> > <version>1.11.1</version> > </dependency> > > > > <dependency> > <groupId>org.apache.flink</groupId> > <artifactId>flink-statebackend-rocksdb_2.11</artifactId> > <version>1.11.2</version> > <!-<scope>test</scope>-> > </dependency> > > <dependency> > <groupId>org.apache.hadoop</groupId> > <artifactId>hadoop-client</artifactId> > <version>3.3.0</version> > </dependency> > > <dependency> > <groupId>org.apache.flink</groupId> > <artifactId>flink-core</artifactId> > <version>1.11.1</version> > </dependency> > > <!-<dependency>-> > <!-<groupId>org.slf4j</groupId>-> > <!-<artifactId>slf4j-simple</artifactId>-> > <!-<version>1.7.25</version>-> > <!-<scope>compile</scope>-> > <!-</dependency>-> > > <!-- https://mvnrepository.com/artifact/org.apache.flink/flink-cep --> > <dependency> > <groupId>org.apache.flink</groupId> > <artifactId>flink-cep_2.11</artifactId> > <version>1.11.1</version> > </dependency> > > <dependency> > <groupId>org.apache.flink</groupId> > <artifactId>flink-cep-scala_2.11</artifactId> > <version>1.11.1</version> > </dependency> > > <dependency> > <groupId>org.apache.flink</groupId> > <artifactId>flink-scala_2.11</artifactId> > <version>1.11.1</version> > </dependency> > > > > <dependency> > <groupId>org.projectlombok</groupId> > <artifactId>lombok</artifactId> > <version>1.18.4</version> > <!-<scope>provided</scope>-> > </dependency> > > </dependencies> > </project> > > > > the error I got is: > > https://paste.ubuntu.com/p/49HRYXFzR2/ > > > > *some of the above error is:* > > *Caused by: java.lang.IllegalStateException: Unexpected state handle type, > expected: class org.apache.flink.runtime.state.KeyGroupsStateHandle, but > found: class > org.apache.flink.runtime.state.IncrementalRemoteKeyedStateHandle* > > > > > > The steps are: > > 1.mvn clean scala:compile compile package > > 2.nc -lk 9999 > > 3.flink run -c wordcount_increstate datastream_api-1.0-SNAPSHOT.jar > Job has been submitted with JobID df6d62a43620f258155b8538ef0ddf1b > > 4.input the following conents in nc -lk 9999 > > before > error > error > error > error > > 5. > > flink run -s > hdfs://Desktop:9000/tmp/flinkck/df6d62a43620f258155b8538ef0ddf1b/chk-22 -c > StateWordCount datastream_api-1.0-SNAPSHOT.jar > > Then the above error happens. > > > > Please help,Thanks~! > > > I have tried to subscried to user@flink.apache.org; > > but no replies.If possible ,send to appleyu...@foxmail.com with your > valuable replies,thanks. > > >