Hi,

I'm trying to finish up a PR (https://github.com/apache/spark/pull/10089)
which is currently failing PySpark tests. The instructions to run the test
suite seem a little dated. I was able to find these:
https://cwiki.apache.org/confluence/display/SPARK/PySpark+Internals
http://spark.apache.org/docs/latest/building-spark.html

I've tried running `python/run-tests`, but it fails hard at the ORC tests. I
suspect it has to do with the external libraries not being compiled or put
in the right location.
I've tried running `SPARK_TESTING=1 ./bin/pyspark
python/pyspark/streaming/tests.py` as suggested, but this doesn't work on
Spark 2.0.
I've tried running `SPARK_TESTING=1 ./bin/spark-submit
python/pyspark/streaming/tests.py`and that worked a little better, but it
failed at `pyspark.streaming.tests.KafkaStreamTests`, with
`java.lang.ClassNotFoundException:
org.apache.spark.streaming.kafka.KafkaTestUtils`. I suspect the same issue
with external libraries.

I've compiling Spark with `build/mvn -Pyarn -Phadoop-2.4
-Dhadoop.version=2.4.0 -DskipTests clean package` with no trouble.

Is there any better documentation somewhere about how to run the PySpark
tests?



--
View this message in context: 
http://apache-spark-developers-list.1001551.n3.nabble.com/How-to-run-PySpark-tests-tp16357.html
Sent from the Apache Spark Developers List mailing list archive at Nabble.com.

---------------------------------------------------------------------
To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
For additional commands, e-mail: dev-h...@spark.apache.org

Reply via email to