Does the jar file you are trying to submit contain
the org/apache/kafka/common/serialization/ByteArrayDeserializer class?

On Thu, Sep 9, 2021 at 2:10 PM Harshvardhan Shinde <
harshvardhan.shi...@oyorooms.com> wrote:

> Here's the complete stack trace:
>
> Server Response:org.apache.flink.runtime.rest.handler.RestHandlerException:
> Could not execute application. at
> org.apache.flink.runtime.webmonitor.handlers.JarRunHandler.lambda$handleRequest$1(JarRunHandler.java:108)
> at
> java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:836)
> at
> java.util.concurrent.CompletableFuture$UniHandle.tryFire(CompletableFuture.java:811)
> at
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488)
> at
> java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1609)
> at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> at java.util.concurrent.FutureTask.run(FutureTask.java:266) at
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
> at
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> at java.lang.Thread.run(Thread.java:748) Caused by:
> java.util.concurrent.CompletionException: java.lang.NoClassDefFoundError:
> org/apache/kafka/common/serialization/ByteArrayDeserializer at
> java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:273)
> at
> java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:280)
> at
> java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1606)
> ... 7 more Caused by: java.lang.NoClassDefFoundError:
> org/apache/kafka/common/serialization/ByteArrayDeserializer at
> org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer.setDeserializer(FlinkKafkaConsumer.java:322)
> at
> org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer.<init>(FlinkKafkaConsumer.java:223)
> at
> org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer.<init>(FlinkKafkaConsumer.java:154)
> at
> org.apache.flink.streaming.connectors.kafka.table.KafkaDynamicSource.createKafkaConsumer(KafkaDynamicSource.java:383)
> at
> org.apache.flink.streaming.connectors.kafka.table.KafkaDynamicSource.getScanRuntimeProvider(KafkaDynamicSource.java:205)
> at
> org.apache.flink.table.planner.connectors.DynamicSourceUtils.validateScanSource(DynamicSourceUtils.java:453)
> at
> org.apache.flink.table.planner.connectors.DynamicSourceUtils.prepareDynamicSource(DynamicSourceUtils.java:161)
> at
> org.apache.flink.table.planner.connectors.DynamicSourceUtils.convertSourceToRel(DynamicSourceUtils.java:119)
> at
> org.apache.flink.table.planner.plan.schema.CatalogSourceTable.toRel(CatalogSourceTable.java:85)
> at
> org.apache.calcite.sql2rel.SqlToRelConverter.toRel(SqlToRelConverter.java:3585)
> at
> org.apache.calcite.sql2rel.SqlToRelConverter.convertIdentifier(SqlToRelConverter.java:2507)
> at
> org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2144)
> at
> org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2093)
> at
> org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2050)
> at
> org.apache.calcite.sql2rel.SqlToRelConverter.convertSelectImpl(SqlToRelConverter.java:663)
> at
> org.apache.calcite.sql2rel.SqlToRelConverter.convertSelect(SqlToRelConverter.java:644)
> at
> org.apache.calcite.sql2rel.SqlToRelConverter.convertQueryRecursive(SqlToRelConverter.java:3438)
> at
> org.apache.calcite.sql2rel.SqlToRelConverter.convertQuery(SqlToRelConverter.java:570)
> at 
> org.apache.flink.table.planner.calcite.FlinkPlannerImpl.org$apache$flink$table$planner$calcite$FlinkPlannerImpl$$rel(FlinkPlannerImpl.scala:169)
> at
> org.apache.flink.table.planner.calcite.FlinkPlannerImpl.rel(FlinkPlannerImpl.scala:161)
> at
> org.apache.flink.table.planner.operations.SqlToOperationConverter.toQueryOperation(SqlToOperationConverter.java:989)
> at
> org.apache.flink.table.planner.operations.SqlToOperationConverter.convertSqlQuery(SqlToOperationConverter.java:958)
> at
> org.apache.flink.table.planner.operations.SqlToOperationConverter.convert(SqlToOperationConverter.java:283)
> at
> org.apache.flink.table.planner.delegation.ParserImpl.parse(ParserImpl.java:101)
> at
> org.apache.flink.table.api.internal.TableEnvironmentImpl.sqlQuery(TableEnvironmentImpl.java:704)
> at com.harsh.test.StreamingJob.main(StreamingJob.java:106) at
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:498) at
> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:355)
> at
> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:222)
> at org.apache.flink.client.ClientUtils.executeProgram(ClientUtils.java:114)
> at
> org.apache.flink.client.deployment.application.DetachedApplicationRunner.tryExecuteJobs(DetachedApplicationRunner.java:84)
> at
> org.apache.flink.client.deployment.application.DetachedApplicationRunner.run(DetachedApplicationRunner.java:70)
> at
> org.apache.flink.runtime.webmonitor.handlers.JarRunHandler.lambda$handleRequest$0(JarRunHandler.java:102)
> at
> java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1604)
> ... 7 more
>
> On Thu, Sep 9, 2021 at 5:36 PM Robert Metzger <rmetz...@apache.org> wrote:
>
>> Can you share the full stack trace, not just a part of it?
>>
>> On Thu, Sep 9, 2021 at 1:43 PM Harshvardhan Shinde <
>> harshvardhan.shi...@oyorooms.com> wrote:
>>
>>> Hi,
>>>
>>> I added the dependencies while trying to resolve the same issue, thought
>>> I was missing them.
>>>
>>> Thanks
>>>
>>> On Thu, Sep 9, 2021 at 4:26 PM Robert Metzger <rmetz...@apache.org>
>>> wrote:
>>>
>>>> Hey,
>>>>
>>>> Why do you have these dependencies in your pom?
>>>>
>>>>         <!-- 
>>>> https://mvnrepository.com/artifact/org.apache.kafka/kafka-clients -->
>>>>         <dependency>
>>>>             <groupId>org.apache.kafka</groupId>
>>>>             <artifactId>kafka-clients</artifactId>
>>>>             <version>2.8.0</version>
>>>>         </dependency>
>>>>
>>>>         <dependency>
>>>>             <groupId>org.apache.kafka</groupId>
>>>>             <artifactId>kafka_2.12</artifactId>
>>>>             <version>2.8.0</version>
>>>>         </dependency>
>>>>
>>>>
>>>> They are not needed for using the Kafka connector of Flink (the flink
>>>> kafka connector dependencies pulls the required dependencies)
>>>>
>>>>
>>>> On Thu, Sep 9, 2021 at 12:02 PM Harshvardhan Shinde <
>>>> harshvardhan.shi...@oyorooms.com> wrote:
>>>>
>>>>> Hi,
>>>>>
>>>>> I'm trying a simple flink job that reads data from a kafka topic and
>>>>> creates a Hive table.
>>>>>
>>>>> I'm following the steps from here
>>>>> <https://ci.apache.org/projects/flink/flink-docs-release-1.13/docs/connectors/table/hive/overview/#connecting-to-hive>
>>>>> .
>>>>>
>>>>> Here's my code:
>>>>>
>>>>> import org.apache.flink.table.api.EnvironmentSettings;
>>>>> import org.apache.flink.table.api.Table;
>>>>> import org.apache.flink.table.api.TableEnvironment;
>>>>> import org.apache.flink.table.catalog.hive.HiveCatalog;
>>>>>
>>>>> EnvironmentSettings settings = 
>>>>> EnvironmentSettings.newInstance().useBlinkPlanner().inStreamingMode().build();
>>>>> TableEnvironment tableEnv = TableEnvironment.create(settings);
>>>>>
>>>>> String name            = "myhive";
>>>>> String defaultDatabase = "harsh_test";
>>>>> String hiveConfDir     = "/etc/hive/conf";
>>>>>
>>>>> HiveCatalog hive = new HiveCatalog(name, defaultDatabase, hiveConfDir);
>>>>> tableEnv.registerCatalog(name, hive);
>>>>>
>>>>> // set the HiveCatalog as the current catalog of the session
>>>>> tableEnv.useCatalog(name);
>>>>>
>>>>> tableEnv.executeSql("CREATE TABLE IF NOT EXISTS transactions (\n" +
>>>>>       "  `created_at` TIMESTAMP(3) METADATA FROM 'timestamp',\n" +
>>>>>       "      `partition` BIGINT METADATA VIRTUAL,\n" +
>>>>>       "      `offset` BIGINT METADATA VIRTUAL,\n" +
>>>>>       "    account_id  BIGINT,\n" +
>>>>>       "    amount      BIGINT,\n" +
>>>>>       "    transaction_time TIMESTAMP(3),\n" +
>>>>>       "    WATERMARK FOR transaction_time AS transaction_time - INTERVAL 
>>>>> '5' SECOND\n" +
>>>>>       ") WITH (\n" +
>>>>>       "    'connector' = 'kafka',\n" +
>>>>>       "    'topic'     = 'flink-stream-table',\n" +
>>>>>       "    'properties.bootstrap.servers' = '<BROKER_ADDRESS>:9092',\n" +
>>>>>       "   'scan.startup.mode' = 'earliest-offset',\n" +
>>>>>       "    'format'    = 'csv'\n" +
>>>>>       ")");
>>>>>
>>>>> Table table = tableEnv.sqlQuery("Select * from transactions");
>>>>> table.execute().print();
>>>>>
>>>>> The code builds successfully, but I'm getting the following runtime
>>>>> error:
>>>>>
>>>>> Caused by: java.util.concurrent.CompletionException:
>>>>> java.lang.NoClassDefFoundError:
>>>>> org/apache/kafka/common/serialization/ByteArrayDeserializer at
>>>>> java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:273)
>>>>> at
>>>>> java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:280)
>>>>> at
>>>>> java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1606)
>>>>> ..
>>>>>
>>>>> Here are my pom.xml file contents:
>>>>>
>>>>> <!--
>>>>> Licensed to the Apache Software Foundation (ASF) under one
>>>>> or more contributor license agreements.  See the NOTICE file
>>>>> distributed with this work for additional information
>>>>> regarding copyright ownership.  The ASF licenses this file
>>>>> to you under the Apache License, Version 2.0 (the
>>>>> "License"); you may not use this file except in compliance
>>>>> with the License.  You may obtain a copy of the License at
>>>>>
>>>>>   http://www.apache.org/licenses/LICENSE-2.0
>>>>>
>>>>> Unless required by applicable law or agreed to in writing,
>>>>> software distributed under the License is distributed on an
>>>>> "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
>>>>> KIND, either express or implied.  See the License for the
>>>>> specific language governing permissions and limitations
>>>>> under the License.
>>>>> -->
>>>>> <project xmlns="http://maven.apache.org/POM/4.0.0"; 
>>>>> xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance";
>>>>>          xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 
>>>>> http://maven.apache.org/xsd/maven-4.0.0.xsd";>
>>>>>     <modelVersion>4.0.0</modelVersion>
>>>>>
>>>>>     <groupId>com.harsh.test</groupId>
>>>>>     <artifactId>harsh-flink-test</artifactId>
>>>>>     <version>1.0-SNAPSHOT</version>
>>>>>     <packaging>jar</packaging>
>>>>>
>>>>>     <name>Flink Quickstart Job</name>
>>>>>     <url>http://www.myorganization.org</url>
>>>>>
>>>>>     <properties>
>>>>>         <project.build.sourceEncoding>UTF-8</project.build.sourceEncoding>
>>>>>         <flink.version>1.13.2</flink.version>
>>>>>         <java.version>1.8</java.version>
>>>>>         <hive.version>2.3.6</hive.version>
>>>>>         <scala.binary.version>2.12</scala.binary.version>
>>>>>         <maven.compiler.source>${java.version}</maven.compiler.source>
>>>>>         <maven.compiler.target>${java.version}</maven.compiler.target>
>>>>>     </properties>
>>>>>
>>>>>     <repositories>
>>>>>         <repository>
>>>>>             <id>apache.snapshots</id>
>>>>>             <name>Apache Development Snapshot Repository</name>
>>>>>             
>>>>> <url>https://repository.apache.org/content/repositories/snapshots/</url>
>>>>>             <releases>
>>>>>                 <enabled>false</enabled>
>>>>>             </releases>
>>>>>             <snapshots>
>>>>>                 <enabled>true</enabled>
>>>>>             </snapshots>
>>>>>         </repository>
>>>>>     </repositories>
>>>>>
>>>>>     <dependencies>
>>>>>         <!-- Apache Flink dependencies -->
>>>>>         <!-- These dependencies are provided, because they should not be 
>>>>> packaged into the JAR file. -->
>>>>>         <dependency>
>>>>>             <groupId>org.apache.flink</groupId>
>>>>>             <artifactId>flink-java</artifactId>
>>>>>             <version>${flink.version}</version>
>>>>>         </dependency>
>>>>>         <dependency>
>>>>>             <groupId>org.apache.flink</groupId>
>>>>>             
>>>>> <artifactId>flink-streaming-java_${scala.binary.version}</artifactId>
>>>>>             <version>${flink.version}</version>
>>>>>         </dependency>
>>>>>
>>>>>         <!-- Add connector dependencies here. They must be in the default 
>>>>> scope (compile). -->
>>>>>
>>>>>         <!-- Example:
>>>>>
>>>>>         <dependency>
>>>>>             <groupId>org.apache.flink</groupId>
>>>>>             
>>>>> <artifactId>flink-connector-kafka-0.10_${scala.binary.version}</artifactId>
>>>>>             <version>${flink.version}</version>
>>>>>         </dependency>
>>>>>         -->
>>>>>
>>>>>         <!-- 
>>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-connector-kafka 
>>>>> -->
>>>>>         <dependency>
>>>>>             <groupId>org.apache.flink</groupId>
>>>>>             
>>>>> <artifactId>flink-connector-kafka_${scala.binary.version}</artifactId>
>>>>>             <version>${flink.version}</version>
>>>>>         </dependency>
>>>>>
>>>>>         <dependency>
>>>>>             <groupId>org.apache.flink</groupId>
>>>>>             
>>>>> <artifactId>flink-table-api-java-bridge_${scala.binary.version}</artifactId>
>>>>>             <version>${flink.version}</version>
>>>>>         </dependency>
>>>>>
>>>>>         <dependency>
>>>>>             <groupId>org.apache.flink</groupId>
>>>>>             
>>>>> <artifactId>flink-table-planner-blink_${scala.binary.version}</artifactId>
>>>>>             <version>${flink.version}</version>
>>>>>         </dependency>
>>>>>
>>>>>         <!-- 
>>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-table-planner 
>>>>> -->
>>>>>         <dependency>
>>>>>             <groupId>org.apache.flink</groupId>
>>>>>             <artifactId>flink-table-planner_2.12</artifactId>
>>>>>             <version>1.13.2</version>
>>>>>         </dependency>
>>>>>
>>>>>
>>>>>         <!-- Flink Dependency -->
>>>>>         <dependency>
>>>>>             <groupId>org.apache.flink</groupId>
>>>>>             
>>>>> <artifactId>flink-connector-hive_${scala.binary.version}</artifactId>
>>>>>             <version>${flink.version}</version>
>>>>>         </dependency>
>>>>>
>>>>>         <!-- Hive Dependency -->
>>>>>         <dependency>
>>>>>             <groupId>org.apache.hive</groupId>
>>>>>             <artifactId>hive-exec</artifactId>
>>>>>             <version>${hive.version}</version>
>>>>>         </dependency>
>>>>>
>>>>>         <!-- 
>>>>> https://mvnrepository.com/artifact/javax.servlet/javax.servlet-api -->
>>>>>         <dependency>
>>>>>             <groupId>javax.servlet</groupId>
>>>>>             <artifactId>javax.servlet-api</artifactId>
>>>>>             <version>3.1.0</version>
>>>>>             <scope>provided</scope>
>>>>>         </dependency>
>>>>>
>>>>>         <!-- 
>>>>> https://mvnrepository.com/artifact/org.apache.htrace/htrace-core4 -->
>>>>>         <dependency>
>>>>>             <groupId>org.apache.htrace</groupId>
>>>>>             <artifactId>htrace-core4</artifactId>
>>>>>             <version>4.0.1-incubating</version>
>>>>>         </dependency>
>>>>>
>>>>>         <!-- 
>>>>> https://mvnrepository.com/artifact/commons-configuration/commons-configuration
>>>>>  -->
>>>>>         <dependency>
>>>>>             <groupId>commons-configuration</groupId>
>>>>>             <artifactId>commons-configuration</artifactId>
>>>>>             <version>1.10</version>
>>>>>         </dependency>
>>>>>
>>>>>         <!-- 
>>>>> https://mvnrepository.com/artifact/commons-logging/commons-logging -->
>>>>>         <dependency>
>>>>>             <groupId>commons-logging</groupId>
>>>>>             <artifactId>commons-logging</artifactId>
>>>>>             <version>1.2</version>
>>>>>         </dependency>
>>>>>
>>>>>         <!-- 
>>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-shaded-hadoop-2 
>>>>> -->
>>>>>         <dependency>
>>>>>             <groupId>org.apache.flink</groupId>
>>>>>             <artifactId>flink-shaded-hadoop-2</artifactId>
>>>>>             <version>2.8.3-10.0</version>
>>>>>         </dependency>
>>>>>
>>>>>
>>>>>         <!-- 
>>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-hadoop-compatibility
>>>>>  -->
>>>>>         <dependency>
>>>>>             <groupId>org.apache.flink</groupId>
>>>>>             <artifactId>flink-hadoop-compatibility_2.12</artifactId>
>>>>>             <version>1.13.2</version>
>>>>>         </dependency>
>>>>>
>>>>>         <!-- 
>>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-hadoop-fs -->
>>>>>         <dependency>
>>>>>             <groupId>org.apache.flink</groupId>
>>>>>             <artifactId>flink-hadoop-fs</artifactId>
>>>>>             <version>1.13.2</version>
>>>>>         </dependency>
>>>>>
>>>>>         <!-- 
>>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-csv -->
>>>>>         <dependency>
>>>>>             <groupId>org.apache.flink</groupId>
>>>>>             <artifactId>flink-csv</artifactId>
>>>>>             <version>1.13.2</version>
>>>>>         </dependency>
>>>>>
>>>>>         <!-- 
>>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-json -->
>>>>>         <dependency>
>>>>>             <groupId>org.apache.flink</groupId>
>>>>>             <artifactId>flink-json</artifactId>
>>>>>             <version>1.13.2</version>
>>>>>         </dependency>
>>>>>
>>>>>         <!-- 
>>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-sql-connector-hive-1.2.2
>>>>>  -->
>>>>>         <dependency>
>>>>>             <groupId>org.apache.flink</groupId>
>>>>>             <artifactId>flink-sql-connector-hive-2.3.6_2.12</artifactId>
>>>>>             <version>1.13.2</version>
>>>>>         </dependency>
>>>>>
>>>>>         <!-- 
>>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-core -->
>>>>>         <dependency>
>>>>>             <groupId>org.apache.flink</groupId>
>>>>>             <artifactId>flink-core</artifactId>
>>>>>             <version>1.13.2</version>
>>>>>         </dependency>
>>>>>
>>>>>         <!-- 
>>>>> https://mvnrepository.com/artifact/org.apache.kafka/kafka-clients -->
>>>>>         <dependency>
>>>>>             <groupId>org.apache.kafka</groupId>
>>>>>             <artifactId>kafka-clients</artifactId>
>>>>>             <version>2.8.0</version>
>>>>>         </dependency>
>>>>>
>>>>>         <dependency>
>>>>>             <groupId>org.apache.kafka</groupId>
>>>>>             <artifactId>kafka_2.12</artifactId>
>>>>>             <version>2.8.0</version>
>>>>>         </dependency>
>>>>>
>>>>>
>>>>>         <!-- Add logging framework, to produce console output when 
>>>>> running in the IDE. -->
>>>>>         <!-- These dependencies are excluded from the application JAR by 
>>>>> default. -->
>>>>>         <dependency>
>>>>>             <groupId>org.slf4j</groupId>
>>>>>             <artifactId>slf4j-log4j12</artifactId>
>>>>>             <version>1.7.7</version>
>>>>>             <scope>runtime</scope>
>>>>>         </dependency>
>>>>>         <dependency>
>>>>>             <groupId>log4j</groupId>
>>>>>             <artifactId>log4j</artifactId>
>>>>>             <version>1.2.17</version>
>>>>>             <scope>runtime</scope>
>>>>>         </dependency>
>>>>>     </dependencies>
>>>>>
>>>>>     <build>
>>>>>         <plugins>
>>>>>
>>>>>             <!-- Java Compiler -->
>>>>>             <plugin>
>>>>>                 <groupId>org.apache.maven.plugins</groupId>
>>>>>                 <artifactId>maven-compiler-plugin</artifactId>
>>>>>                 <version>3.1</version>
>>>>>                 <configuration>
>>>>>                     <source>${java.version}</source>
>>>>>                     <target>${java.version}</target>
>>>>>                 </configuration>
>>>>>             </plugin>
>>>>>
>>>>>             <!-- We use the maven-shade plugin to create a fat jar that 
>>>>> contains all necessary dependencies. -->
>>>>>             <!-- Change the value of <mainClass>...</mainClass> if your 
>>>>> program entry point changes. -->
>>>>>             <plugin>
>>>>>                 <groupId>org.apache.maven.plugins</groupId>
>>>>>                 <artifactId>maven-shade-plugin</artifactId>
>>>>>                 <version>3.0.0</version>
>>>>>                 <executions>
>>>>>                     <!-- Run shade goal on package phase -->
>>>>>                     <execution>
>>>>>                         <phase>package</phase>
>>>>>                         <goals>
>>>>>                             <goal>shade</goal>
>>>>>                         </goals>
>>>>>                         <configuration>
>>>>>                             <artifactSet>
>>>>>                                 <excludes>
>>>>>                                     
>>>>> <exclude>org.apache.flink:force-shading</exclude>
>>>>>                                     
>>>>> <exclude>com.google.code.findbugs:jsr305</exclude>
>>>>>                                     <exclude>org.slf4j:*</exclude>
>>>>>                                     <exclude>log4j:*</exclude>
>>>>>                                 </excludes>
>>>>>                             </artifactSet>
>>>>>                             <filters>
>>>>>                                 <filter>
>>>>>                                     <!-- Do not copy the signatures in 
>>>>> the META-INF folder.
>>>>>                                     Otherwise, this might cause 
>>>>> SecurityExceptions when using the JAR. -->
>>>>>                                     <artifact>*:*</artifact>
>>>>>                                     <excludes>
>>>>>                                         <exclude>META-INF/*.SF</exclude>
>>>>>                                         <exclude>META-INF/*.DSA</exclude>
>>>>>                                         <exclude>META-INF/*.RSA</exclude>
>>>>>                                     </excludes>
>>>>>                                 </filter>
>>>>>                             </filters>
>>>>>                             <transformers>
>>>>>                                 <transformer 
>>>>> implementation="org.apache.maven.plugins.shade.resource.ManifestResourceTransformer">
>>>>>                                     
>>>>> <mainClass>com.harsh.test.StreamingJob</mainClass>
>>>>>                                 </transformer>
>>>>>                             </transformers>
>>>>>                         </configuration>
>>>>>                     </execution>
>>>>>                 </executions>
>>>>>             </plugin>
>>>>>         </plugins>
>>>>>
>>>>>         <pluginManagement>
>>>>>             <plugins>
>>>>>
>>>>>                 <!-- This improves the out-of-the-box experience in 
>>>>> Eclipse by resolving some warnings. -->
>>>>>                 <plugin>
>>>>>                     <groupId>org.eclipse.m2e</groupId>
>>>>>                     <artifactId>lifecycle-mapping</artifactId>
>>>>>                     <version>1.0.0</version>
>>>>>                     <configuration>
>>>>>                         <lifecycleMappingMetadata>
>>>>>                             <pluginExecutions>
>>>>>                                 <pluginExecution>
>>>>>                                     <pluginExecutionFilter>
>>>>>                                         
>>>>> <groupId>org.apache.maven.plugins</groupId>
>>>>>                                         
>>>>> <artifactId>maven-shade-plugin</artifactId>
>>>>>                                         
>>>>> <versionRange>[3.0.0,)</versionRange>
>>>>>                                         <goals>
>>>>>                                             <goal>shade</goal>
>>>>>                                         </goals>
>>>>>                                     </pluginExecutionFilter>
>>>>>                                     <action>
>>>>>                                         <ignore/>
>>>>>                                     </action>
>>>>>                                 </pluginExecution>
>>>>>                                 <pluginExecution>
>>>>>                                     <pluginExecutionFilter>
>>>>>                                         
>>>>> <groupId>org.apache.maven.plugins</groupId>
>>>>>                                         
>>>>> <artifactId>maven-compiler-plugin</artifactId>
>>>>>                                         
>>>>> <versionRange>[3.1,)</versionRange>
>>>>>                                         <goals>
>>>>>                                             <goal>testCompile</goal>
>>>>>                                             <goal>compile</goal>
>>>>>                                         </goals>
>>>>>                                     </pluginExecutionFilter>
>>>>>                                     <action>
>>>>>                                         <ignore/>
>>>>>                                     </action>
>>>>>                                 </pluginExecution>
>>>>>                             </pluginExecutions>
>>>>>                         </lifecycleMappingMetadata>
>>>>>                     </configuration>
>>>>>                 </plugin>
>>>>>             </plugins>
>>>>>         </pluginManagement>
>>>>>     </build>
>>>>>
>>>>>     <!-- This profile helps to make things run out of the box in IntelliJ 
>>>>> -->
>>>>>     <!-- Its adds Flink's core classes to the runtime class path. -->
>>>>>     <!-- Otherwise they are missing in IntelliJ, because the dependency 
>>>>> is 'provided' -->
>>>>>     <profiles>
>>>>>         <profile>
>>>>>             <id>add-dependencies-for-IDEA</id>
>>>>>
>>>>>             <activation>
>>>>>                 <property>
>>>>>                     <name>idea.version</name>
>>>>>                 </property>
>>>>>             </activation>
>>>>>
>>>>>             <dependencies>
>>>>>                 <dependency>
>>>>>                     <groupId>org.apache.flink</groupId>
>>>>>                     <artifactId>flink-java</artifactId>
>>>>>                     <version>${flink.version}</version>
>>>>>                     <scope>compile</scope>
>>>>>                 </dependency>
>>>>>                 <dependency>
>>>>>                     <groupId>org.apache.flink</groupId>
>>>>>                     
>>>>> <artifactId>flink-streaming-java_${scala.binary.version}</artifactId>
>>>>>                     <version>${flink.version}</version>
>>>>>                     <scope>compile</scope>
>>>>>                 </dependency>
>>>>>             </dependencies>
>>>>>         </profile>
>>>>>     </profiles>
>>>>>
>>>>> </project>
>>>>>
>>>>>
>>>>> Please help me resolve the issue.
>>>>>
>>>>> Thanks
>>>>>
>>>>>
>>>>>
>>>>>
>>>
>>> --
>>> Thanks and Regards,
>>> Harshvardhan
>>> Data Platform
>>>
>>
>
> --
> Thanks and Regards,
> Harshvardhan
> Data Platform
>

Reply via email to