It looks like you were trying to resume from a checkpoint taken with the
FsStateBackend into a revised version of the job that uses the
RocksDbStateBackend. Switching state backends in this way is not supported:
checkpoints and savepoints are written in a state-backend-specific format,
and can only be read by the same backend that wrote them.

It is possible, however, to migrate between state backends using the State
Processor API [1].

[1]
https://ci.apache.org/projects/flink/flink-docs-stable/dev/libs/state_processor_api.html

Best,
David

On Fri, Oct 2, 2020 at 4:07 PM 大森林 <appleyu...@foxmail.com> wrote:

> *I want to do an experiment of"incremental checkpoint"*
>
> my code is:
>
> https://paste.ubuntu.com/p/DpTyQKq6Vk/
>
>
>
> pom.xml is:
>
> <?xml version="1.0" encoding="UTF-8"?>
> <project xmlns="http://maven.apache.org/POM/4.0.0";
> xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance";
> xsi:schemaLocation="http://maven.apache.org/POM/4.0.0
> http://maven.apache.org/xsd/maven-4.0.0.xsd";>
> <modelVersion>4.0.0</modelVersion>
>
> <groupId>example</groupId>
> <artifactId>datastream_api</artifactId>
> <version>1.0-SNAPSHOT</version>
> <build>
> <plugins>
> <plugin>
> <groupId>org.apache.maven.plugins</groupId>
> <artifactId>maven-compiler-plugin</artifactId>
> <version>3.1</version>
> <configuration>
> <source>1.8</source>
> <target>1.8</target>
> </configuration>
> </plugin>
>
> <plugin>
> <groupId>org.scala-tools</groupId>
> <artifactId>maven-scala-plugin</artifactId>
> <version>2.15.2</version>
> <executions>
> <execution>
> <goals>
> <goal>compile</goal>
> <goal>testCompile</goal>
> </goals>
> </execution>
> </executions>
> </plugin>
>
>
>
> </plugins>
> </build>
>
> <dependencies>
>
> <!--
> https://mvnrepository.com/artifact/org.apache.flink/flink-streaming-scala
> -->
> <dependency>
> <groupId>org.apache.flink</groupId>
> <artifactId>flink-streaming-scala_2.11</artifactId>
> <version>1.11.1</version>
> <!-<scope>provided</scope>->
> </dependency>
>
> <!-<dependency>->
> <!-<groupId>org.apache.flink</groupId>->
> <!-<artifactId>flink-streaming-java_2.12</artifactId>->
> <!-<version>1.11.1</version>->
> <!-<!–<scope>compile</scope>–>->
> <!-</dependency>->
>
> <dependency>
> <groupId>org.apache.flink</groupId>
> <artifactId>flink-clients_2.11</artifactId>
> <version>1.11.1</version>
> </dependency>
>
>
>
> <dependency>
> <groupId>org.apache.flink</groupId>
> <artifactId>flink-statebackend-rocksdb_2.11</artifactId>
> <version>1.11.2</version>
> <!-<scope>test</scope>->
> </dependency>
>
> <dependency>
> <groupId>org.apache.hadoop</groupId>
> <artifactId>hadoop-client</artifactId>
> <version>3.3.0</version>
> </dependency>
>
> <dependency>
> <groupId>org.apache.flink</groupId>
> <artifactId>flink-core</artifactId>
> <version>1.11.1</version>
> </dependency>
>
> <!-<dependency>->
> <!-<groupId>org.slf4j</groupId>->
> <!-<artifactId>slf4j-simple</artifactId>->
> <!-<version>1.7.25</version>->
> <!-<scope>compile</scope>->
> <!-</dependency>->
>
> <!-- https://mvnrepository.com/artifact/org.apache.flink/flink-cep -->
> <dependency>
> <groupId>org.apache.flink</groupId>
> <artifactId>flink-cep_2.11</artifactId>
> <version>1.11.1</version>
> </dependency>
>
> <dependency>
> <groupId>org.apache.flink</groupId>
> <artifactId>flink-cep-scala_2.11</artifactId>
> <version>1.11.1</version>
> </dependency>
>
> <dependency>
> <groupId>org.apache.flink</groupId>
> <artifactId>flink-scala_2.11</artifactId>
> <version>1.11.1</version>
> </dependency>
>
>
>
> <dependency>
> <groupId>org.projectlombok</groupId>
> <artifactId>lombok</artifactId>
> <version>1.18.4</version>
> <!-<scope>provided</scope>->
> </dependency>
>
> </dependencies>
> </project>
>
>
>
> the error I got is:
>
> https://paste.ubuntu.com/p/49HRYXFzR2/
>
>
>
> *some of the above error is:*
>
> *Caused by: java.lang.IllegalStateException: Unexpected state handle type,
> expected: class org.apache.flink.runtime.state.KeyGroupsStateHandle, but
> found: class
> org.apache.flink.runtime.state.IncrementalRemoteKeyedStateHandle*
>
>
>
>
>
> The steps are:
>
> 1.mvn clean scala:compile compile package
>
> 2.nc -lk 9999
>
> 3.flink run -c wordcount_increstate  datastream_api-1.0-SNAPSHOT.jar
> Job has been submitted with JobID df6d62a43620f258155b8538ef0ddf1b
>
> 4.input the following conents in nc -lk 9999
>
> before
> error
> error
> error
> error
>
> 5.
>
> flink run -s
> hdfs://Desktop:9000/tmp/flinkck/df6d62a43620f258155b8538ef0ddf1b/chk-22 -c
> StateWordCount datastream_api-1.0-SNAPSHOT.jar
>
> Then the above error happens.
>
>
>
> Please help,Thanks~!
>
>
> I have tried to subscried to user@flink.apache.org;
>
> but no replies.If possible ,send to appleyu...@foxmail.com with your
> valuable replies,thanks.
>
>
>

Reply via email to