Hi,
I checked for the dependency using the stackoverflow link you sent, the
ByteArrayDeserializer class is present in my jar but when I try to run it,
I'm getting the same error message.
I also removed the kafka and kafka-client dependencies from my pom.xml file
and added the jar from the link you provided.

On Fri, Sep 10, 2021 at 4:46 PM Timo Walther <twal...@apache.org> wrote:

> It seems that your Kafka clients dependency is not in your JAR file.
>
>
> ByteArrayDeserializer is a symptom that seems to occur often. At least,
> I can find a similar question on Stackoverflow:
>
>
> https://stackoverflow.com/questions/58408494/flink-1-3-class-not-found-org-apache-kafka-common-bytearraydeserializer
>
> Maybe this helps.
>
> Btw for Table API and SQL, you could also try out the fat JAR that we
> provide for the SQL Client:
>
>
> https://ci.apache.org/projects/flink/flink-docs-release-1.13/docs/connectors/table/kafka/
>
> links to
>
>
> https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-kafka_2.11/1.13.2/flink-sql-connector-kafka_2.11-1.13.2.jar
>
> Here we try to do the dependency management for users.
>
> Regards,
> Timo
>
> On 10.09.21 11:20, Harshvardhan Shinde wrote:
> > I'm unable to figure out which dependency to add in order for the
> > ByteArrayDeserializer class to get included in the jar. I have added all
> > the dependencies as per the documentation still unable to figure out the
> > cause.
> >
> > On Fri, Sep 10, 2021 at 12:17 AM Robert Metzger <rmetz...@apache.org
> > <mailto:rmetz...@apache.org>> wrote:
> >
> >     Does the jar file you are trying to submit contain
> >     the org/apache/kafka/common/serialization/ByteArrayDeserializer
> class?
> >
> >     On Thu, Sep 9, 2021 at 2:10 PM Harshvardhan Shinde
> >     <harshvardhan.shi...@oyorooms.com
> >     <mailto:harshvardhan.shi...@oyorooms.com>> wrote:
> >
> >         Here's the complete stack trace:
> >
> >         Server
> >
>  Response:org.apache.flink.runtime.rest.handler.RestHandlerException:
> >         Could not execute application. at
> >
>  
> org.apache.flink.runtime.webmonitor.handlers.JarRunHandler.lambda$handleRequest$1(JarRunHandler.java:108)
> >         at
> >
>  java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:836)
> >         at
> >
>  
> java.util.concurrent.CompletableFuture$UniHandle.tryFire(CompletableFuture.java:811)
> >         at
> >
>  
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488)
> >         at
> >
>  
> java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1609)
> >         at
> >
>  java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> >         at java.util.concurrent.FutureTask.run(FutureTask.java:266) at
> >
>  
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
> >         at
> >
>  
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
> >         at
> >
>  
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> >         at
> >
>  
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> >         at java.lang.Thread.run(Thread.java:748) Caused by:
> >         java.util.concurrent.CompletionException:
> >         java.lang.NoClassDefFoundError:
> >         org/apache/kafka/common/serialization/ByteArrayDeserializer at
> >
>  
> java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:273)
> >         at
> >
>  
> java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:280)
> >         at
> >
>  
> java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1606)
> >         ... 7 more Caused by: java.lang.NoClassDefFoundError:
> >         org/apache/kafka/common/serialization/ByteArrayDeserializer at
> >
>  
> org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer.setDeserializer(FlinkKafkaConsumer.java:322)
> >         at
> >
>  
> org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer.<init>(FlinkKafkaConsumer.java:223)
> >         at
> >
>  
> org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer.<init>(FlinkKafkaConsumer.java:154)
> >         at
> >
>  
> org.apache.flink.streaming.connectors.kafka.table.KafkaDynamicSource.createKafkaConsumer(KafkaDynamicSource.java:383)
> >         at
> >
>  
> org.apache.flink.streaming.connectors.kafka.table.KafkaDynamicSource.getScanRuntimeProvider(KafkaDynamicSource.java:205)
> >         at
> >
>  
> org.apache.flink.table.planner.connectors.DynamicSourceUtils.validateScanSource(DynamicSourceUtils.java:453)
> >         at
> >
>  
> org.apache.flink.table.planner.connectors.DynamicSourceUtils.prepareDynamicSource(DynamicSourceUtils.java:161)
> >         at
> >
>  
> org.apache.flink.table.planner.connectors.DynamicSourceUtils.convertSourceToRel(DynamicSourceUtils.java:119)
> >         at
> >
>  
> org.apache.flink.table.planner.plan.schema.CatalogSourceTable.toRel(CatalogSourceTable.java:85)
> >         at
> >
>  
> org.apache.calcite.sql2rel.SqlToRelConverter.toRel(SqlToRelConverter.java:3585)
> >         at
> >
>  
> org.apache.calcite.sql2rel.SqlToRelConverter.convertIdentifier(SqlToRelConverter.java:2507)
> >         at
> >
>  
> org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2144)
> >         at
> >
>  
> org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2093)
> >         at
> >
>  
> org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2050)
> >         at
> >
>  
> org.apache.calcite.sql2rel.SqlToRelConverter.convertSelectImpl(SqlToRelConverter.java:663)
> >         at
> >
>  
> org.apache.calcite.sql2rel.SqlToRelConverter.convertSelect(SqlToRelConverter.java:644)
> >         at
> >
>  
> org.apache.calcite.sql2rel.SqlToRelConverter.convertQueryRecursive(SqlToRelConverter.java:3438)
> >         at
> >
>  
> org.apache.calcite.sql2rel.SqlToRelConverter.convertQuery(SqlToRelConverter.java:570)
> >         at org.apache.flink.table.planner.calcite.FlinkPlannerImpl.org
> >         <
> http://org.apache.flink.table.planner.calcite.FlinkPlannerImpl.org
> >$apache$flink$table$planner$calcite$FlinkPlannerImpl$$rel(FlinkPlannerImpl.scala:169)
> >         at
> >
>  
> org.apache.flink.table.planner.calcite.FlinkPlannerImpl.rel(FlinkPlannerImpl.scala:161)
> >         at
> >
>  
> org.apache.flink.table.planner.operations.SqlToOperationConverter.toQueryOperation(SqlToOperationConverter.java:989)
> >         at
> >
>  
> org.apache.flink.table.planner.operations.SqlToOperationConverter.convertSqlQuery(SqlToOperationConverter.java:958)
> >         at
> >
>  
> org.apache.flink.table.planner.operations.SqlToOperationConverter.convert(SqlToOperationConverter.java:283)
> >         at
> >
>  
> org.apache.flink.table.planner.delegation.ParserImpl.parse(ParserImpl.java:101)
> >         at
> >
>  
> org.apache.flink.table.api.internal.TableEnvironmentImpl.sqlQuery(TableEnvironmentImpl.java:704)
> >         at com.harsh.test.StreamingJob.main(StreamingJob.java:106) at
> >         sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at
> >
>  sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> >         at
> >
>  
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> >         at java.lang.reflect.Method.invoke(Method.java:498) at
> >
>  
> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:355)
> >         at
> >
>  
> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:222)
> >         at
> >
>  org.apache.flink.client.ClientUtils.executeProgram(ClientUtils.java:114)
> >         at
> >
>  
> org.apache.flink.client.deployment.application.DetachedApplicationRunner.tryExecuteJobs(DetachedApplicationRunner.java:84)
> >         at
> >
>  
> org.apache.flink.client.deployment.application.DetachedApplicationRunner.run(DetachedApplicationRunner.java:70)
> >         at
> >
>  
> org.apache.flink.runtime.webmonitor.handlers.JarRunHandler.lambda$handleRequest$0(JarRunHandler.java:102)
> >         at
> >
>  
> java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1604)
> >         ... 7 more
> >
> >         On Thu, Sep 9, 2021 at 5:36 PM Robert Metzger
> >         <rmetz...@apache.org <mailto:rmetz...@apache.org>> wrote:
> >
> >             Can you share the full stack trace, not just a part of it?
> >
> >             On Thu, Sep 9, 2021 at 1:43 PM Harshvardhan Shinde
> >             <harshvardhan.shi...@oyorooms.com
> >             <mailto:harshvardhan.shi...@oyorooms.com>> wrote:
> >
> >                 Hi,
> >
> >                 I added the dependencies while trying to resolve the
> >                 same issue, thought I was missing them.
> >
> >                 Thanks
> >
> >                 On Thu, Sep 9, 2021 at 4:26 PM Robert Metzger
> >                 <rmetz...@apache.org <mailto:rmetz...@apache.org>>
> wrote:
> >
> >                     Hey,
> >
> >                     Why do you have these dependencies in your pom?
> >
> >                     <!--
> >
> https://mvnrepository.com/artifact/org.apache.kafka/kafka-clients
> >                     <
> https://mvnrepository.com/artifact/org.apache.kafka/kafka-clients>
> >                     -->
> >                     <dependency>
> >                     <groupId>org.apache.kafka</groupId>
> >                     <artifactId>kafka-clients</artifactId>
> >                     <version>2.8.0</version>
> >                     </dependency>
> >
> >                     <dependency>
> >                     <groupId>org.apache.kafka</groupId>
> >                     <artifactId>kafka_2.12</artifactId>
> >                     <version>2.8.0</version>
> >                     </dependency>
> >
> >
> >                     They are not needed for using the Kafka connector of
> >                     Flink (the flink kafka connector dependencies pulls
> >                     the required dependencies)
> >
> >
> >                     On Thu, Sep 9, 2021 at 12:02 PM Harshvardhan Shinde
> >                     <harshvardhan.shi...@oyorooms.com
> >                     <mailto:harshvardhan.shi...@oyorooms.com>> wrote:
> >
> >                         Hi,
> >
> >                         I'm trying a simple flink job that reads data
> >                         from a kafka topic and creates a Hive table.
> >
> >                         I'm following the steps from here
> >                         <
> https://ci.apache.org/projects/flink/flink-docs-release-1.13/docs/connectors/table/hive/overview/#connecting-to-hive
> >.
> >
> >                         Here's my code:
> >
> >                         import
> org.apache.flink.table.api.EnvironmentSettings;
> >                         import org.apache.flink.table.api.Table;
> >                         import
> org.apache.flink.table.api.TableEnvironment;
> >                         import
> org.apache.flink.table.catalog.hive.HiveCatalog;
> >
> >                         EnvironmentSettings settings =
> EnvironmentSettings.newInstance().useBlinkPlanner().inStreamingMode().build();
> >                         TableEnvironment tableEnv =
> TableEnvironment.create(settings);
> >
> >                         String name            ="myhive";
> >                         String defaultDatabase ="harsh_test";
> >                         String hiveConfDir     ="/etc/hive/conf";
> >
> >                         HiveCatalog hive =new HiveCatalog(name,
> defaultDatabase, hiveConfDir);
> >                         tableEnv.registerCatalog(name, hive);
> >
> >                         // set the HiveCatalog as the current catalog of
> >                         the session
> >                         tableEnv.useCatalog(name);
> >
> >                         tableEnv.executeSql("CREATE TABLE IF NOT EXISTS
> transactions (\n" +
> >                                " `created_at` TIMESTAMP(3) METADATA FROM
> >                         'timestamp',\n" +
> >                                " `partition` BIGINT METADATA VIRTUAL,\n"
> +
> >                                " `offset` BIGINT METADATA VIRTUAL,\n" +
> >                                " account_id BIGINT,\n" +
> >                                " amount BIGINT,\n" +
> >                                " transaction_time TIMESTAMP(3),\n" +
> >                                " WATERMARK FOR transaction_time AS
> >                         transaction_time - INTERVAL '5' SECOND\n" +
> >                                ") WITH (\n" +
> >                                " 'connector' = 'kafka',\n" +
> >                                " 'topic' = 'flink-stream-table',\n" +
> >                                " 'properties.bootstrap.servers' =
> >                         '<BROKER_ADDRESS>:9092',\n" +
> >                                " 'scan.startup.mode' =
> 'earliest-offset',\n" +
> >                                " 'format' = 'csv'\n" +
> >                                ")");
> >
> >                         Table table = tableEnv.sqlQuery("Select * from
> transactions");
> >                         table.execute().print();
> >
> >                         The code builds successfully, but I'm getting
> >                         the following runtime error:
> >
> >                         Caused by:
> >                         java.util.concurrent.CompletionException:
> >                         java.lang.NoClassDefFoundError:
> >
>  org/apache/kafka/common/serialization/ByteArrayDeserializer
> >                         at
> >
>  
> java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:273)
> >                         at
> >
>  
> java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:280)
> >                         at
> >
>  
> java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1606)
> >                         ..
> >
> >                         Here are my pom.xml file contents:
> >
> >                         <!--
> >                         Licensed to the Apache Software Foundation (ASF)
> >                         under one
> >                         or more contributor license agreements. See the
> >                         NOTICE file
> >                         distributed with this work for additional
> >                         information
> >                         regarding copyright ownership. The ASF licenses
> >                         this file
> >                         to you under the Apache License, Version 2.0 (the
> >                         "License"); you may not use this file except in
> >                         compliance
> >                         with the License. You may obtain a copy of the
> >                         License at
> >
> >                         http://www.apache.org/licenses/LICENSE-2.0
> >                         <http://www.apache.org/licenses/LICENSE-2.0>
> >
> >                         Unless required by applicable law or agreed to
> >                         in writing,
> >                         software distributed under the License is
> >                         distributed on an
> >                         "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS
> >                         OF ANY
> >                         KIND, either express or implied. See the License
> >                         for the
> >                         specific language governing permissions and
> >                         limitations
> >                         under the License.
> >                         -->
> >                         <project
> >                         xmlns="http://maven.apache.org/POM/4.0.0
> >                         <http://maven.apache.org/POM/4.0.0>"
> >                         xmlns:xsi="
> http://www.w3.org/2001/XMLSchema-instance
> >                         <http://www.w3.org/2001/XMLSchema-instance>"
> >                         xsi:schemaLocation="
> http://maven.apache.org/POM/4.0.0
> >                         <http://maven.apache.org/POM/4.0.0>
> >                         http://maven.apache.org/xsd/maven-4.0.0.xsd
> >                         <http://maven.apache.org/xsd/maven-4.0.0.xsd>">
> >                         <modelVersion>4.0.0</modelVersion>
> >
> >                         <groupId>com.harsh.test</groupId>
> >                         <artifactId>harsh-flink-test</artifactId>
> >                         <version>1.0-SNAPSHOT</version>
> >                         <packaging>jar</packaging>
> >
> >                         <name>Flink Quickstart Job</name>
> >                         <url>http://www.myorganization.org  <
> http://www.myorganization.org></url>
> >
> >                         <properties>
> >
>  <project.build.sourceEncoding>UTF-8</project.build.sourceEncoding>
> >                         <flink.version>1.13.2</flink.version>
> >                         <java.version>1.8</java.version>
> >                         <hive.version>2.3.6</hive.version>
> >                         <scala.binary.version>2.12</scala.binary.version>
> >
>  <maven.compiler.source>${java.version}</maven.compiler.source>
> >
>  <maven.compiler.target>${java.version}</maven.compiler.target>
> >                         </properties>
> >
> >                         <repositories>
> >                         <repository>
> >                         <id>apache.snapshots</id>
> >                         <name>Apache Development Snapshot
> Repository</name>
> >                         <url>
> https://repository.apache.org/content/repositories/snapshots/  <
> https://repository.apache.org/content/repositories/snapshots/></url>
> >                         <releases>
> >                         <enabled>false</enabled>
> >                         </releases>
> >                         <snapshots>
> >                         <enabled>true</enabled>
> >                         </snapshots>
> >                         </repository>
> >                         </repositories>
> >
> >                         <dependencies>
> >                         <!-- Apache Flink dependencies -->
> >                         <!-- These dependencies are provided, because
> >                         they should not be packaged into the JAR file.
> -->
> >                         <dependency>
> >                         <groupId>org.apache.flink</groupId>
> >                         <artifactId>flink-java</artifactId>
> >                         <version>${flink.version}</version>
> >                         </dependency>
> >                         <dependency>
> >                         <groupId>org.apache.flink</groupId>
> >
>  <artifactId>flink-streaming-java_${scala.binary.version}</artifactId>
> >                         <version>${flink.version}</version>
> >                         </dependency>
> >
> >                         <!-- Add connector dependencies here. They must
> >                         be in the default scope (compile). -->
> >
> >                         <!-- Example:
> >
> >                         <dependency>
> >                         <groupId>org.apache.flink</groupId>
> >
>  <artifactId>flink-connector-kafka-0.10_${scala.binary.version}</artifactId>
> >                         <version>${flink.version}</version>
> >                         </dependency>
> >                         -->
> >
> >                         <!--
> >
> https://mvnrepository.com/artifact/org.apache.flink/flink-connector-kafka
> >                         <
> https://mvnrepository.com/artifact/org.apache.flink/flink-connector-kafka>
> >                         -->
> >                         <dependency>
> >                         <groupId>org.apache.flink</groupId>
> >
>  <artifactId>flink-connector-kafka_${scala.binary.version}</artifactId>
> >                         <version>${flink.version}</version>
> >                         </dependency>
> >
> >                         <dependency>
> >                         <groupId>org.apache.flink</groupId>
> >
>  <artifactId>flink-table-api-java-bridge_${scala.binary.version}</artifactId>
> >                         <version>${flink.version}</version>
> >                         </dependency>
> >
> >                         <dependency>
> >                         <groupId>org.apache.flink</groupId>
> >
>  <artifactId>flink-table-planner-blink_${scala.binary.version}</artifactId>
> >                         <version>${flink.version}</version>
> >                         </dependency>
> >
> >                         <!--
> >
> https://mvnrepository.com/artifact/org.apache.flink/flink-table-planner
> >                         <
> https://mvnrepository.com/artifact/org.apache.flink/flink-table-planner>
> >                         -->
> >                         <dependency>
> >                         <groupId>org.apache.flink</groupId>
> >                         <artifactId>flink-table-planner_2.12</artifactId>
> >                         <version>1.13.2</version>
> >                         </dependency>
> >
> >
> >                         <!-- Flink Dependency -->
> >                         <dependency>
> >                         <groupId>org.apache.flink</groupId>
> >
>  <artifactId>flink-connector-hive_${scala.binary.version}</artifactId>
> >                         <version>${flink.version}</version>
> >                         </dependency>
> >
> >                         <!-- Hive Dependency -->
> >                         <dependency>
> >                         <groupId>org.apache.hive</groupId>
> >                         <artifactId>hive-exec</artifactId>
> >                         <version>${hive.version}</version>
> >                         </dependency>
> >
> >                         <!--
> >
> https://mvnrepository.com/artifact/javax.servlet/javax.servlet-api
> >                         <
> https://mvnrepository.com/artifact/javax.servlet/javax.servlet-api>
> >                         -->
> >                         <dependency>
> >                         <groupId>javax.servlet</groupId>
> >                         <artifactId>javax.servlet-api</artifactId>
> >                         <version>3.1.0</version>
> >                         <scope>provided</scope>
> >                         </dependency>
> >
> >                         <!--
> >
> https://mvnrepository.com/artifact/org.apache.htrace/htrace-core4
> >                         <
> https://mvnrepository.com/artifact/org.apache.htrace/htrace-core4>
> >                         -->
> >                         <dependency>
> >                         <groupId>org.apache.htrace</groupId>
> >                         <artifactId>htrace-core4</artifactId>
> >                         <version>4.0.1-incubating</version>
> >                         </dependency>
> >
> >                         <!--
> >
> https://mvnrepository.com/artifact/commons-configuration/commons-configuration
> >                         <
> https://mvnrepository.com/artifact/commons-configuration/commons-configuration
> >
> >                         -->
> >                         <dependency>
> >                         <groupId>commons-configuration</groupId>
> >                         <artifactId>commons-configuration</artifactId>
> >                         <version>1.10</version>
> >                         </dependency>
> >
> >                         <!--
> >
> https://mvnrepository.com/artifact/commons-logging/commons-logging
> >                         <
> https://mvnrepository.com/artifact/commons-logging/commons-logging>
> >                         -->
> >                         <dependency>
> >                         <groupId>commons-logging</groupId>
> >                         <artifactId>commons-logging</artifactId>
> >                         <version>1.2</version>
> >                         </dependency>
> >
> >                         <!--
> >
> https://mvnrepository.com/artifact/org.apache.flink/flink-shaded-hadoop-2
> >                         <
> https://mvnrepository.com/artifact/org.apache.flink/flink-shaded-hadoop-2>
> >                         -->
> >                         <dependency>
> >                         <groupId>org.apache.flink</groupId>
> >                         <artifactId>flink-shaded-hadoop-2</artifactId>
> >                         <version>2.8.3-10.0</version>
> >                         </dependency>
> >
> >
> >                         <!--
> >
> https://mvnrepository.com/artifact/org.apache.flink/flink-hadoop-compatibility
> >                         <
> https://mvnrepository.com/artifact/org.apache.flink/flink-hadoop-compatibility
> >
> >                         -->
> >                         <dependency>
> >                         <groupId>org.apache.flink</groupId>
> >
>  <artifactId>flink-hadoop-compatibility_2.12</artifactId>
> >                         <version>1.13.2</version>
> >                         </dependency>
> >
> >                         <!--
> >
> https://mvnrepository.com/artifact/org.apache.flink/flink-hadoop-fs
> >                         <
> https://mvnrepository.com/artifact/org.apache.flink/flink-hadoop-fs>
> >                         -->
> >                         <dependency>
> >                         <groupId>org.apache.flink</groupId>
> >                         <artifactId>flink-hadoop-fs</artifactId>
> >                         <version>1.13.2</version>
> >                         </dependency>
> >
> >                         <!--
> >
> https://mvnrepository.com/artifact/org.apache.flink/flink-csv
> >                         <
> https://mvnrepository.com/artifact/org.apache.flink/flink-csv>
> >                         -->
> >                         <dependency>
> >                         <groupId>org.apache.flink</groupId>
> >                         <artifactId>flink-csv</artifactId>
> >                         <version>1.13.2</version>
> >                         </dependency>
> >
> >                         <!--
> >
> https://mvnrepository.com/artifact/org.apache.flink/flink-json
> >                         <
> https://mvnrepository.com/artifact/org.apache.flink/flink-json>
> >                         -->
> >                         <dependency>
> >                         <groupId>org.apache.flink</groupId>
> >                         <artifactId>flink-json</artifactId>
> >                         <version>1.13.2</version>
> >                         </dependency>
> >
> >                         <!--
> >
> https://mvnrepository.com/artifact/org.apache.flink/flink-sql-connector-hive-1.2.2
> >                         <
> https://mvnrepository.com/artifact/org.apache.flink/flink-sql-connector-hive-1.2.2
> >
> >                         -->
> >                         <dependency>
> >                         <groupId>org.apache.flink</groupId>
> >
>  <artifactId>flink-sql-connector-hive-2.3.6_2.12</artifactId>
> >                         <version>1.13.2</version>
> >                         </dependency>
> >
> >                         <!--
> >
> https://mvnrepository.com/artifact/org.apache.flink/flink-core
> >                         <
> https://mvnrepository.com/artifact/org.apache.flink/flink-core>
> >                         -->
> >                         <dependency>
> >                         <groupId>org.apache.flink</groupId>
> >                         <artifactId>flink-core</artifactId>
> >                         <version>1.13.2</version>
> >                         </dependency>
> >
> >                         <!--
> >
> https://mvnrepository.com/artifact/org.apache.kafka/kafka-clients
> >                         <
> https://mvnrepository.com/artifact/org.apache.kafka/kafka-clients>
> >                         -->
> >                         <dependency>
> >                         <groupId>org.apache.kafka</groupId>
> >                         <artifactId>kafka-clients</artifactId>
> >                         <version>2.8.0</version>
> >                         </dependency>
> >
> >                         <dependency>
> >                         <groupId>org.apache.kafka</groupId>
> >                         <artifactId>kafka_2.12</artifactId>
> >                         <version>2.8.0</version>
> >                         </dependency>
> >
> >
> >                         <!-- Add logging framework, to produce console
> >                         output when running in the IDE. -->
> >                         <!-- These dependencies are excluded from the
> >                         application JAR by default. -->
> >                         <dependency>
> >                         <groupId>org.slf4j</groupId>
> >                         <artifactId>slf4j-log4j12</artifactId>
> >                         <version>1.7.7</version>
> >                         <scope>runtime</scope>
> >                         </dependency>
> >                         <dependency>
> >                         <groupId>log4j</groupId>
> >                         <artifactId>log4j</artifactId>
> >                         <version>1.2.17</version>
> >                         <scope>runtime</scope>
> >                         </dependency>
> >                         </dependencies>
> >
> >                         <build>
> >                         <plugins>
> >
> >                         <!-- Java Compiler -->
> >                         <plugin>
> >                         <groupId>org.apache.maven.plugins</groupId>
> >                         <artifactId>maven-compiler-plugin</artifactId>
> >                         <version>3.1</version>
> >                         <configuration>
> >                         <source>${java.version}</source>
> >                         <target>${java.version}</target>
> >                         </configuration>
> >                         </plugin>
> >
> >                         <!-- We use the maven-shade plugin to create a
> >                         fat jar that contains all necessary
> >                         dependencies. -->
> >                         <!-- Change the value of
> >                         <mainClass>...</mainClass> if your program entry
> >                         point changes. -->
> >                         <plugin>
> >                         <groupId>org.apache.maven.plugins</groupId>
> >                         <artifactId>maven-shade-plugin</artifactId>
> >                         <version>3.0.0</version>
> >                         <executions>
> >                         <!-- Run shade goal on package phase -->
> >                         <execution>
> >                         <phase>package</phase>
> >                         <goals>
> >                         <goal>shade</goal>
> >                         </goals>
> >                         <configuration>
> >                         <artifactSet>
> >                         <excludes>
> >                         <exclude>org.apache.flink:force-shading</exclude>
> >
>  <exclude>com.google.code.findbugs:jsr305</exclude>
> >                         <exclude>org.slf4j:*</exclude>
> >                         <exclude>log4j:*</exclude>
> >                         </excludes>
> >                         </artifactSet>
> >                         <filters>
> >                         <filter>
> >                         <!-- Do not copy the signatures in the META-INF
> >                         folder.
> >                         Otherwise, this might cause SecurityExceptions
> >                         when using the JAR. -->
> >                         <artifact>*:*</artifact>
> >                         <excludes>
> >                         <exclude>META-INF/*.SF</exclude>
> >                         <exclude>META-INF/*.DSA</exclude>
> >                         <exclude>META-INF/*.RSA</exclude>
> >                         </excludes>
> >                         </filter>
> >                         </filters>
> >                         <transformers>
> >                         <transformer
> >
>  
> implementation="org.apache.maven.plugins.shade.resource.ManifestResourceTransformer">
> >
>  <mainClass>com.harsh.test.StreamingJob</mainClass>
> >                         </transformer>
> >                         </transformers>
> >                         </configuration>
> >                         </execution>
> >                         </executions>
> >                         </plugin>
> >                         </plugins>
> >
> >                         <pluginManagement>
> >                         <plugins>
> >
> >                         <!-- This improves the out-of-the-box experience
> >                         in Eclipse by resolving some warnings. -->
> >                         <plugin>
> >                         <groupId>org.eclipse.m2e</groupId>
> >                         <artifactId>lifecycle-mapping</artifactId>
> >                         <version>1.0.0</version>
> >                         <configuration>
> >                         <lifecycleMappingMetadata>
> >                         <pluginExecutions>
> >                         <pluginExecution>
> >                         <pluginExecutionFilter>
> >                         <groupId>org.apache.maven.plugins</groupId>
> >                         <artifactId>maven-shade-plugin</artifactId>
> >                         <versionRange>[3.0.0,)</versionRange>
> >                         <goals>
> >                         <goal>shade</goal>
> >                         </goals>
> >                         </pluginExecutionFilter>
> >                         <action>
> >                         <ignore/>
> >                         </action>
> >                         </pluginExecution>
> >                         <pluginExecution>
> >                         <pluginExecutionFilter>
> >                         <groupId>org.apache.maven.plugins</groupId>
> >                         <artifactId>maven-compiler-plugin</artifactId>
> >                         <versionRange>[3.1,)</versionRange>
> >                         <goals>
> >                         <goal>testCompile</goal>
> >                         <goal>compile</goal>
> >                         </goals>
> >                         </pluginExecutionFilter>
> >                         <action>
> >                         <ignore/>
> >                         </action>
> >                         </pluginExecution>
> >                         </pluginExecutions>
> >                         </lifecycleMappingMetadata>
> >                         </configuration>
> >                         </plugin>
> >                         </plugins>
> >                         </pluginManagement>
> >                         </build>
> >
> >                         <!-- This profile helps to make things run out
> >                         of the box in IntelliJ -->
> >                         <!-- Its adds Flink's core classes to the
> >                         runtime class path. -->
> >                         <!-- Otherwise they are missing in IntelliJ,
> >                         because the dependency is 'provided' -->
> >                         <profiles>
> >                         <profile>
> >                         <id>add-dependencies-for-IDEA</id>
> >
> >                         <activation>
> >                         <property>
> >                         <name>idea.version</name>
> >                         </property>
> >                         </activation>
> >
> >                         <dependencies>
> >                         <dependency>
> >                         <groupId>org.apache.flink</groupId>
> >                         <artifactId>flink-java</artifactId>
> >                         <version>${flink.version}</version>
> >                         <scope>compile</scope>
> >                         </dependency>
> >                         <dependency>
> >                         <groupId>org.apache.flink</groupId>
> >
>  <artifactId>flink-streaming-java_${scala.binary.version}</artifactId>
> >                         <version>${flink.version}</version>
> >                         <scope>compile</scope>
> >                         </dependency>
> >                         </dependencies>
> >                         </profile>
> >                         </profiles>
> >
> >                         </project>
> >
> >
> >                         Please help me resolve the issue.
> >
> >                         Thanks
> >
> >
> >
> >
> >
> >                 --
> >                 Thanks and Regards,
> >                 Harshvardhan
> >                 Data Platform
> >
> >
> >
> >         --
> >         Thanks and Regards,
> >         Harshvardhan
> >         Data Platform
> >
> >
> >
> > --
> > Thanks and Regards,
> > Harshvardhan
> > Data Platform
>
>

-- 
Thanks and Regards,
Harshvardhan
Data Platform

Reply via email to