Well, this just won't work when you are running Spark on Hadoop...

On 12/10/20 9:14 PM, lec ssmi wrote:
If you can not assembly the jdbc driver jar in your application jar package, you can put the jdbc driver jar in the spark classpath, generally, $SPARK_HOME/jars  or $SPARK_HOME/lib.


Artemis User <arte...@dtechspace.com <mailto:arte...@dtechspace.com>> 于2020年12月11日周五 上午5:21写道:

    What happened was that you made the mysql jar file only available
    to the spark driver, not the executors.  Use the --jars parameter
    instead of driver-class-path to specify your third-party jar
    files, or copy the third-party jar files to the jars directory for
    Spark in your HDFS, and specify the path of HDFS using --archives
    in spark-submit.

    -- ND

    On 12/10/20 10:02 AM, ismail elhammoud wrote:
    Hello,

    Guys I have an issue with mysql connector java, even if I
    declared it in sbt file It couldn't work if I don't give the
    whole path

    spark-submit --master yarn --driver-class-path
    /home/node2/Téléchargements/mysql-connector-java-5.1.24-bin.jar
    ./Sdatahub-assembly-0.1.jar


    Regards,
    Isma

Reply via email to