[ 
https://issues.apache.org/jira/browse/FLINK-25253?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chesnay Schepler updated FLINK-25253:
-------------------------------------
    Component/s: Connectors / Hive

> An ClassNotFoundException of missing Hadoop class occurred when submitting to 
> yarn
> ----------------------------------------------------------------------------------
>
>                 Key: FLINK-25253
>                 URL: https://issues.apache.org/jira/browse/FLINK-25253
>             Project: Flink
>          Issue Type: Bug
>          Components: Connectors / Hive, Deployment / YARN
>    Affects Versions: 1.14.0
>         Environment: Environment version:
> Hadoop 3.1.1
> Hive 3.1.1
> Flink 1.4.0
> kafka 2.6.1
>  
>            Reporter: ghost synth
>            Priority: Blocker
>         Attachments: FlinkPlaySubmit.scala, flink lib.png, 
> original-TropicalaLink-1.0-SNAPSHOT.jar, pom.xml, submit_log.log
>
>   Original Estimate: 96h
>  Remaining Estimate: 96h
>
> I use the Hive Table connector to write hive from Kafka and submit to yarn 
> successfully, but it will always report during execution
> *Caused by: java.lang.ClassNotFoundException: 
> org.apache.hadoop.mapred.JobConf*
> Before submitting, I executed "export HADOOP_CLASSPATH=`hadoop classpath`" to 
> import hadoop dependencies.
> I found in the JM log that the classpath already contains Hadoop 
> dependencies, but an exception still occurs
> The original jar that I submitted only contains code and does not contain 
> dependencies,The program loads dependencies from the hadoop path and lib 
> under the flink directory
> The attachment contains flink lib, code, jar and JM log
> Thanks



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

Reply via email to