Hello spent a lot of time to find what I did wrong , but not found.
I have a minikube WIndows based cluster ( Hyper V as hypervisor ) and try
to run examples against Spark 2.3. Tried several  docker images builds:
* several  builds that I build myself
* andrusha/spark-k8s:2.3.0-hadoop2.7 from docker  hub
But when I try to submit job driver log returns  class not found exception
org.apache.spark.examples.SparkPi

spark-submit --master k8s://https://ip:8443  --deploy-mode cluster  --name
spark-pi --class org.apache.spark.examples.SparkPi --conf
spark.executor.instances=1 --executor-memory 1G --conf spark.kubernete
s.container.image=andrusha/spark-k8s:2.3.0-hadoop2.7
local:///opt/spark/examples/spark-examples_2.11-2.3.0.jar

I tried to use https://github.com/apache-spark-on-k8s/spark fork and it is
works without problems, more complex examples work also.

Reply via email to