Does the jar file you are trying to submit contain the org/apache/kafka/common/serialization/ByteArrayDeserializer class?
On Thu, Sep 9, 2021 at 2:10 PM Harshvardhan Shinde < harshvardhan.shi...@oyorooms.com> wrote: > Here's the complete stack trace: > > Server Response:org.apache.flink.runtime.rest.handler.RestHandlerException: > Could not execute application. at > org.apache.flink.runtime.webmonitor.handlers.JarRunHandler.lambda$handleRequest$1(JarRunHandler.java:108) > at > java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:836) > at > java.util.concurrent.CompletableFuture$UniHandle.tryFire(CompletableFuture.java:811) > at > java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488) > at > java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1609) > at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) > at java.util.concurrent.FutureTask.run(FutureTask.java:266) at > java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) > at > java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) > at java.lang.Thread.run(Thread.java:748) Caused by: > java.util.concurrent.CompletionException: java.lang.NoClassDefFoundError: > org/apache/kafka/common/serialization/ByteArrayDeserializer at > java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:273) > at > java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:280) > at > java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1606) > ... 7 more Caused by: java.lang.NoClassDefFoundError: > org/apache/kafka/common/serialization/ByteArrayDeserializer at > org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer.setDeserializer(FlinkKafkaConsumer.java:322) > at > org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer.<init>(FlinkKafkaConsumer.java:223) > at > org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer.<init>(FlinkKafkaConsumer.java:154) > at > org.apache.flink.streaming.connectors.kafka.table.KafkaDynamicSource.createKafkaConsumer(KafkaDynamicSource.java:383) > at > org.apache.flink.streaming.connectors.kafka.table.KafkaDynamicSource.getScanRuntimeProvider(KafkaDynamicSource.java:205) > at > org.apache.flink.table.planner.connectors.DynamicSourceUtils.validateScanSource(DynamicSourceUtils.java:453) > at > org.apache.flink.table.planner.connectors.DynamicSourceUtils.prepareDynamicSource(DynamicSourceUtils.java:161) > at > org.apache.flink.table.planner.connectors.DynamicSourceUtils.convertSourceToRel(DynamicSourceUtils.java:119) > at > org.apache.flink.table.planner.plan.schema.CatalogSourceTable.toRel(CatalogSourceTable.java:85) > at > org.apache.calcite.sql2rel.SqlToRelConverter.toRel(SqlToRelConverter.java:3585) > at > org.apache.calcite.sql2rel.SqlToRelConverter.convertIdentifier(SqlToRelConverter.java:2507) > at > org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2144) > at > org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2093) > at > org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2050) > at > org.apache.calcite.sql2rel.SqlToRelConverter.convertSelectImpl(SqlToRelConverter.java:663) > at > org.apache.calcite.sql2rel.SqlToRelConverter.convertSelect(SqlToRelConverter.java:644) > at > org.apache.calcite.sql2rel.SqlToRelConverter.convertQueryRecursive(SqlToRelConverter.java:3438) > at > org.apache.calcite.sql2rel.SqlToRelConverter.convertQuery(SqlToRelConverter.java:570) > at > org.apache.flink.table.planner.calcite.FlinkPlannerImpl.org$apache$flink$table$planner$calcite$FlinkPlannerImpl$$rel(FlinkPlannerImpl.scala:169) > at > org.apache.flink.table.planner.calcite.FlinkPlannerImpl.rel(FlinkPlannerImpl.scala:161) > at > org.apache.flink.table.planner.operations.SqlToOperationConverter.toQueryOperation(SqlToOperationConverter.java:989) > at > org.apache.flink.table.planner.operations.SqlToOperationConverter.convertSqlQuery(SqlToOperationConverter.java:958) > at > org.apache.flink.table.planner.operations.SqlToOperationConverter.convert(SqlToOperationConverter.java:283) > at > org.apache.flink.table.planner.delegation.ParserImpl.parse(ParserImpl.java:101) > at > org.apache.flink.table.api.internal.TableEnvironmentImpl.sqlQuery(TableEnvironmentImpl.java:704) > at com.harsh.test.StreamingJob.main(StreamingJob.java:106) at > sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) at > org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:355) > at > org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:222) > at org.apache.flink.client.ClientUtils.executeProgram(ClientUtils.java:114) > at > org.apache.flink.client.deployment.application.DetachedApplicationRunner.tryExecuteJobs(DetachedApplicationRunner.java:84) > at > org.apache.flink.client.deployment.application.DetachedApplicationRunner.run(DetachedApplicationRunner.java:70) > at > org.apache.flink.runtime.webmonitor.handlers.JarRunHandler.lambda$handleRequest$0(JarRunHandler.java:102) > at > java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1604) > ... 7 more > > On Thu, Sep 9, 2021 at 5:36 PM Robert Metzger <rmetz...@apache.org> wrote: > >> Can you share the full stack trace, not just a part of it? >> >> On Thu, Sep 9, 2021 at 1:43 PM Harshvardhan Shinde < >> harshvardhan.shi...@oyorooms.com> wrote: >> >>> Hi, >>> >>> I added the dependencies while trying to resolve the same issue, thought >>> I was missing them. >>> >>> Thanks >>> >>> On Thu, Sep 9, 2021 at 4:26 PM Robert Metzger <rmetz...@apache.org> >>> wrote: >>> >>>> Hey, >>>> >>>> Why do you have these dependencies in your pom? >>>> >>>> <!-- >>>> https://mvnrepository.com/artifact/org.apache.kafka/kafka-clients --> >>>> <dependency> >>>> <groupId>org.apache.kafka</groupId> >>>> <artifactId>kafka-clients</artifactId> >>>> <version>2.8.0</version> >>>> </dependency> >>>> >>>> <dependency> >>>> <groupId>org.apache.kafka</groupId> >>>> <artifactId>kafka_2.12</artifactId> >>>> <version>2.8.0</version> >>>> </dependency> >>>> >>>> >>>> They are not needed for using the Kafka connector of Flink (the flink >>>> kafka connector dependencies pulls the required dependencies) >>>> >>>> >>>> On Thu, Sep 9, 2021 at 12:02 PM Harshvardhan Shinde < >>>> harshvardhan.shi...@oyorooms.com> wrote: >>>> >>>>> Hi, >>>>> >>>>> I'm trying a simple flink job that reads data from a kafka topic and >>>>> creates a Hive table. >>>>> >>>>> I'm following the steps from here >>>>> <https://ci.apache.org/projects/flink/flink-docs-release-1.13/docs/connectors/table/hive/overview/#connecting-to-hive> >>>>> . >>>>> >>>>> Here's my code: >>>>> >>>>> import org.apache.flink.table.api.EnvironmentSettings; >>>>> import org.apache.flink.table.api.Table; >>>>> import org.apache.flink.table.api.TableEnvironment; >>>>> import org.apache.flink.table.catalog.hive.HiveCatalog; >>>>> >>>>> EnvironmentSettings settings = >>>>> EnvironmentSettings.newInstance().useBlinkPlanner().inStreamingMode().build(); >>>>> TableEnvironment tableEnv = TableEnvironment.create(settings); >>>>> >>>>> String name = "myhive"; >>>>> String defaultDatabase = "harsh_test"; >>>>> String hiveConfDir = "/etc/hive/conf"; >>>>> >>>>> HiveCatalog hive = new HiveCatalog(name, defaultDatabase, hiveConfDir); >>>>> tableEnv.registerCatalog(name, hive); >>>>> >>>>> // set the HiveCatalog as the current catalog of the session >>>>> tableEnv.useCatalog(name); >>>>> >>>>> tableEnv.executeSql("CREATE TABLE IF NOT EXISTS transactions (\n" + >>>>> " `created_at` TIMESTAMP(3) METADATA FROM 'timestamp',\n" + >>>>> " `partition` BIGINT METADATA VIRTUAL,\n" + >>>>> " `offset` BIGINT METADATA VIRTUAL,\n" + >>>>> " account_id BIGINT,\n" + >>>>> " amount BIGINT,\n" + >>>>> " transaction_time TIMESTAMP(3),\n" + >>>>> " WATERMARK FOR transaction_time AS transaction_time - INTERVAL >>>>> '5' SECOND\n" + >>>>> ") WITH (\n" + >>>>> " 'connector' = 'kafka',\n" + >>>>> " 'topic' = 'flink-stream-table',\n" + >>>>> " 'properties.bootstrap.servers' = '<BROKER_ADDRESS>:9092',\n" + >>>>> " 'scan.startup.mode' = 'earliest-offset',\n" + >>>>> " 'format' = 'csv'\n" + >>>>> ")"); >>>>> >>>>> Table table = tableEnv.sqlQuery("Select * from transactions"); >>>>> table.execute().print(); >>>>> >>>>> The code builds successfully, but I'm getting the following runtime >>>>> error: >>>>> >>>>> Caused by: java.util.concurrent.CompletionException: >>>>> java.lang.NoClassDefFoundError: >>>>> org/apache/kafka/common/serialization/ByteArrayDeserializer at >>>>> java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:273) >>>>> at >>>>> java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:280) >>>>> at >>>>> java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1606) >>>>> .. >>>>> >>>>> Here are my pom.xml file contents: >>>>> >>>>> <!-- >>>>> Licensed to the Apache Software Foundation (ASF) under one >>>>> or more contributor license agreements. See the NOTICE file >>>>> distributed with this work for additional information >>>>> regarding copyright ownership. The ASF licenses this file >>>>> to you under the Apache License, Version 2.0 (the >>>>> "License"); you may not use this file except in compliance >>>>> with the License. You may obtain a copy of the License at >>>>> >>>>> http://www.apache.org/licenses/LICENSE-2.0 >>>>> >>>>> Unless required by applicable law or agreed to in writing, >>>>> software distributed under the License is distributed on an >>>>> "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY >>>>> KIND, either express or implied. See the License for the >>>>> specific language governing permissions and limitations >>>>> under the License. >>>>> --> >>>>> <project xmlns="http://maven.apache.org/POM/4.0.0" >>>>> xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" >>>>> xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 >>>>> http://maven.apache.org/xsd/maven-4.0.0.xsd"> >>>>> <modelVersion>4.0.0</modelVersion> >>>>> >>>>> <groupId>com.harsh.test</groupId> >>>>> <artifactId>harsh-flink-test</artifactId> >>>>> <version>1.0-SNAPSHOT</version> >>>>> <packaging>jar</packaging> >>>>> >>>>> <name>Flink Quickstart Job</name> >>>>> <url>http://www.myorganization.org</url> >>>>> >>>>> <properties> >>>>> <project.build.sourceEncoding>UTF-8</project.build.sourceEncoding> >>>>> <flink.version>1.13.2</flink.version> >>>>> <java.version>1.8</java.version> >>>>> <hive.version>2.3.6</hive.version> >>>>> <scala.binary.version>2.12</scala.binary.version> >>>>> <maven.compiler.source>${java.version}</maven.compiler.source> >>>>> <maven.compiler.target>${java.version}</maven.compiler.target> >>>>> </properties> >>>>> >>>>> <repositories> >>>>> <repository> >>>>> <id>apache.snapshots</id> >>>>> <name>Apache Development Snapshot Repository</name> >>>>> >>>>> <url>https://repository.apache.org/content/repositories/snapshots/</url> >>>>> <releases> >>>>> <enabled>false</enabled> >>>>> </releases> >>>>> <snapshots> >>>>> <enabled>true</enabled> >>>>> </snapshots> >>>>> </repository> >>>>> </repositories> >>>>> >>>>> <dependencies> >>>>> <!-- Apache Flink dependencies --> >>>>> <!-- These dependencies are provided, because they should not be >>>>> packaged into the JAR file. --> >>>>> <dependency> >>>>> <groupId>org.apache.flink</groupId> >>>>> <artifactId>flink-java</artifactId> >>>>> <version>${flink.version}</version> >>>>> </dependency> >>>>> <dependency> >>>>> <groupId>org.apache.flink</groupId> >>>>> >>>>> <artifactId>flink-streaming-java_${scala.binary.version}</artifactId> >>>>> <version>${flink.version}</version> >>>>> </dependency> >>>>> >>>>> <!-- Add connector dependencies here. They must be in the default >>>>> scope (compile). --> >>>>> >>>>> <!-- Example: >>>>> >>>>> <dependency> >>>>> <groupId>org.apache.flink</groupId> >>>>> >>>>> <artifactId>flink-connector-kafka-0.10_${scala.binary.version}</artifactId> >>>>> <version>${flink.version}</version> >>>>> </dependency> >>>>> --> >>>>> >>>>> <!-- >>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-connector-kafka >>>>> --> >>>>> <dependency> >>>>> <groupId>org.apache.flink</groupId> >>>>> >>>>> <artifactId>flink-connector-kafka_${scala.binary.version}</artifactId> >>>>> <version>${flink.version}</version> >>>>> </dependency> >>>>> >>>>> <dependency> >>>>> <groupId>org.apache.flink</groupId> >>>>> >>>>> <artifactId>flink-table-api-java-bridge_${scala.binary.version}</artifactId> >>>>> <version>${flink.version}</version> >>>>> </dependency> >>>>> >>>>> <dependency> >>>>> <groupId>org.apache.flink</groupId> >>>>> >>>>> <artifactId>flink-table-planner-blink_${scala.binary.version}</artifactId> >>>>> <version>${flink.version}</version> >>>>> </dependency> >>>>> >>>>> <!-- >>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-table-planner >>>>> --> >>>>> <dependency> >>>>> <groupId>org.apache.flink</groupId> >>>>> <artifactId>flink-table-planner_2.12</artifactId> >>>>> <version>1.13.2</version> >>>>> </dependency> >>>>> >>>>> >>>>> <!-- Flink Dependency --> >>>>> <dependency> >>>>> <groupId>org.apache.flink</groupId> >>>>> >>>>> <artifactId>flink-connector-hive_${scala.binary.version}</artifactId> >>>>> <version>${flink.version}</version> >>>>> </dependency> >>>>> >>>>> <!-- Hive Dependency --> >>>>> <dependency> >>>>> <groupId>org.apache.hive</groupId> >>>>> <artifactId>hive-exec</artifactId> >>>>> <version>${hive.version}</version> >>>>> </dependency> >>>>> >>>>> <!-- >>>>> https://mvnrepository.com/artifact/javax.servlet/javax.servlet-api --> >>>>> <dependency> >>>>> <groupId>javax.servlet</groupId> >>>>> <artifactId>javax.servlet-api</artifactId> >>>>> <version>3.1.0</version> >>>>> <scope>provided</scope> >>>>> </dependency> >>>>> >>>>> <!-- >>>>> https://mvnrepository.com/artifact/org.apache.htrace/htrace-core4 --> >>>>> <dependency> >>>>> <groupId>org.apache.htrace</groupId> >>>>> <artifactId>htrace-core4</artifactId> >>>>> <version>4.0.1-incubating</version> >>>>> </dependency> >>>>> >>>>> <!-- >>>>> https://mvnrepository.com/artifact/commons-configuration/commons-configuration >>>>> --> >>>>> <dependency> >>>>> <groupId>commons-configuration</groupId> >>>>> <artifactId>commons-configuration</artifactId> >>>>> <version>1.10</version> >>>>> </dependency> >>>>> >>>>> <!-- >>>>> https://mvnrepository.com/artifact/commons-logging/commons-logging --> >>>>> <dependency> >>>>> <groupId>commons-logging</groupId> >>>>> <artifactId>commons-logging</artifactId> >>>>> <version>1.2</version> >>>>> </dependency> >>>>> >>>>> <!-- >>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-shaded-hadoop-2 >>>>> --> >>>>> <dependency> >>>>> <groupId>org.apache.flink</groupId> >>>>> <artifactId>flink-shaded-hadoop-2</artifactId> >>>>> <version>2.8.3-10.0</version> >>>>> </dependency> >>>>> >>>>> >>>>> <!-- >>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-hadoop-compatibility >>>>> --> >>>>> <dependency> >>>>> <groupId>org.apache.flink</groupId> >>>>> <artifactId>flink-hadoop-compatibility_2.12</artifactId> >>>>> <version>1.13.2</version> >>>>> </dependency> >>>>> >>>>> <!-- >>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-hadoop-fs --> >>>>> <dependency> >>>>> <groupId>org.apache.flink</groupId> >>>>> <artifactId>flink-hadoop-fs</artifactId> >>>>> <version>1.13.2</version> >>>>> </dependency> >>>>> >>>>> <!-- >>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-csv --> >>>>> <dependency> >>>>> <groupId>org.apache.flink</groupId> >>>>> <artifactId>flink-csv</artifactId> >>>>> <version>1.13.2</version> >>>>> </dependency> >>>>> >>>>> <!-- >>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-json --> >>>>> <dependency> >>>>> <groupId>org.apache.flink</groupId> >>>>> <artifactId>flink-json</artifactId> >>>>> <version>1.13.2</version> >>>>> </dependency> >>>>> >>>>> <!-- >>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-sql-connector-hive-1.2.2 >>>>> --> >>>>> <dependency> >>>>> <groupId>org.apache.flink</groupId> >>>>> <artifactId>flink-sql-connector-hive-2.3.6_2.12</artifactId> >>>>> <version>1.13.2</version> >>>>> </dependency> >>>>> >>>>> <!-- >>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-core --> >>>>> <dependency> >>>>> <groupId>org.apache.flink</groupId> >>>>> <artifactId>flink-core</artifactId> >>>>> <version>1.13.2</version> >>>>> </dependency> >>>>> >>>>> <!-- >>>>> https://mvnrepository.com/artifact/org.apache.kafka/kafka-clients --> >>>>> <dependency> >>>>> <groupId>org.apache.kafka</groupId> >>>>> <artifactId>kafka-clients</artifactId> >>>>> <version>2.8.0</version> >>>>> </dependency> >>>>> >>>>> <dependency> >>>>> <groupId>org.apache.kafka</groupId> >>>>> <artifactId>kafka_2.12</artifactId> >>>>> <version>2.8.0</version> >>>>> </dependency> >>>>> >>>>> >>>>> <!-- Add logging framework, to produce console output when >>>>> running in the IDE. --> >>>>> <!-- These dependencies are excluded from the application JAR by >>>>> default. --> >>>>> <dependency> >>>>> <groupId>org.slf4j</groupId> >>>>> <artifactId>slf4j-log4j12</artifactId> >>>>> <version>1.7.7</version> >>>>> <scope>runtime</scope> >>>>> </dependency> >>>>> <dependency> >>>>> <groupId>log4j</groupId> >>>>> <artifactId>log4j</artifactId> >>>>> <version>1.2.17</version> >>>>> <scope>runtime</scope> >>>>> </dependency> >>>>> </dependencies> >>>>> >>>>> <build> >>>>> <plugins> >>>>> >>>>> <!-- Java Compiler --> >>>>> <plugin> >>>>> <groupId>org.apache.maven.plugins</groupId> >>>>> <artifactId>maven-compiler-plugin</artifactId> >>>>> <version>3.1</version> >>>>> <configuration> >>>>> <source>${java.version}</source> >>>>> <target>${java.version}</target> >>>>> </configuration> >>>>> </plugin> >>>>> >>>>> <!-- We use the maven-shade plugin to create a fat jar that >>>>> contains all necessary dependencies. --> >>>>> <!-- Change the value of <mainClass>...</mainClass> if your >>>>> program entry point changes. --> >>>>> <plugin> >>>>> <groupId>org.apache.maven.plugins</groupId> >>>>> <artifactId>maven-shade-plugin</artifactId> >>>>> <version>3.0.0</version> >>>>> <executions> >>>>> <!-- Run shade goal on package phase --> >>>>> <execution> >>>>> <phase>package</phase> >>>>> <goals> >>>>> <goal>shade</goal> >>>>> </goals> >>>>> <configuration> >>>>> <artifactSet> >>>>> <excludes> >>>>> >>>>> <exclude>org.apache.flink:force-shading</exclude> >>>>> >>>>> <exclude>com.google.code.findbugs:jsr305</exclude> >>>>> <exclude>org.slf4j:*</exclude> >>>>> <exclude>log4j:*</exclude> >>>>> </excludes> >>>>> </artifactSet> >>>>> <filters> >>>>> <filter> >>>>> <!-- Do not copy the signatures in >>>>> the META-INF folder. >>>>> Otherwise, this might cause >>>>> SecurityExceptions when using the JAR. --> >>>>> <artifact>*:*</artifact> >>>>> <excludes> >>>>> <exclude>META-INF/*.SF</exclude> >>>>> <exclude>META-INF/*.DSA</exclude> >>>>> <exclude>META-INF/*.RSA</exclude> >>>>> </excludes> >>>>> </filter> >>>>> </filters> >>>>> <transformers> >>>>> <transformer >>>>> implementation="org.apache.maven.plugins.shade.resource.ManifestResourceTransformer"> >>>>> >>>>> <mainClass>com.harsh.test.StreamingJob</mainClass> >>>>> </transformer> >>>>> </transformers> >>>>> </configuration> >>>>> </execution> >>>>> </executions> >>>>> </plugin> >>>>> </plugins> >>>>> >>>>> <pluginManagement> >>>>> <plugins> >>>>> >>>>> <!-- This improves the out-of-the-box experience in >>>>> Eclipse by resolving some warnings. --> >>>>> <plugin> >>>>> <groupId>org.eclipse.m2e</groupId> >>>>> <artifactId>lifecycle-mapping</artifactId> >>>>> <version>1.0.0</version> >>>>> <configuration> >>>>> <lifecycleMappingMetadata> >>>>> <pluginExecutions> >>>>> <pluginExecution> >>>>> <pluginExecutionFilter> >>>>> >>>>> <groupId>org.apache.maven.plugins</groupId> >>>>> >>>>> <artifactId>maven-shade-plugin</artifactId> >>>>> >>>>> <versionRange>[3.0.0,)</versionRange> >>>>> <goals> >>>>> <goal>shade</goal> >>>>> </goals> >>>>> </pluginExecutionFilter> >>>>> <action> >>>>> <ignore/> >>>>> </action> >>>>> </pluginExecution> >>>>> <pluginExecution> >>>>> <pluginExecutionFilter> >>>>> >>>>> <groupId>org.apache.maven.plugins</groupId> >>>>> >>>>> <artifactId>maven-compiler-plugin</artifactId> >>>>> >>>>> <versionRange>[3.1,)</versionRange> >>>>> <goals> >>>>> <goal>testCompile</goal> >>>>> <goal>compile</goal> >>>>> </goals> >>>>> </pluginExecutionFilter> >>>>> <action> >>>>> <ignore/> >>>>> </action> >>>>> </pluginExecution> >>>>> </pluginExecutions> >>>>> </lifecycleMappingMetadata> >>>>> </configuration> >>>>> </plugin> >>>>> </plugins> >>>>> </pluginManagement> >>>>> </build> >>>>> >>>>> <!-- This profile helps to make things run out of the box in IntelliJ >>>>> --> >>>>> <!-- Its adds Flink's core classes to the runtime class path. --> >>>>> <!-- Otherwise they are missing in IntelliJ, because the dependency >>>>> is 'provided' --> >>>>> <profiles> >>>>> <profile> >>>>> <id>add-dependencies-for-IDEA</id> >>>>> >>>>> <activation> >>>>> <property> >>>>> <name>idea.version</name> >>>>> </property> >>>>> </activation> >>>>> >>>>> <dependencies> >>>>> <dependency> >>>>> <groupId>org.apache.flink</groupId> >>>>> <artifactId>flink-java</artifactId> >>>>> <version>${flink.version}</version> >>>>> <scope>compile</scope> >>>>> </dependency> >>>>> <dependency> >>>>> <groupId>org.apache.flink</groupId> >>>>> >>>>> <artifactId>flink-streaming-java_${scala.binary.version}</artifactId> >>>>> <version>${flink.version}</version> >>>>> <scope>compile</scope> >>>>> </dependency> >>>>> </dependencies> >>>>> </profile> >>>>> </profiles> >>>>> >>>>> </project> >>>>> >>>>> >>>>> Please help me resolve the issue. >>>>> >>>>> Thanks >>>>> >>>>> >>>>> >>>>> >>> >>> -- >>> Thanks and Regards, >>> Harshvardhan >>> Data Platform >>> >> > > -- > Thanks and Regards, > Harshvardhan > Data Platform >