There's informal way to test specific tests. For instance:

SPARK_TESTING=1 ../bin/pyspark pyspark.sql.tests VectorizedUDFTests

I have a partial fix for our testing script to support this way in my local
but couldn't have enough time to make a PR for it yet.


2018년 8월 20일 (월) 오전 11:08, Imran Rashid <iras...@cloudera.com.invalid>님이 작성:

> Hi,
>
> I haven't spent a lot of time working on the python side of spark before
> so apologize if this is a basic question, but I'm trying to figure out the
> best way to run a small subset of python tests in a tight loop while
> developing.  The closer I can get to sbt's "~test-only *FooSuite -- -z
> test-blah" the better.
>
> I'm familiar with the "--modules" in python/run-tests, but even running
> one module takes a long time when I want to just run one teeny test
> repeatedly.  Is there a way to run just one file?  And a way to run only
> one test within a file?
>
> So far, I know I can assembly my own command line like run-tests does,
> with all the env vars like PYSPARK_SUBMIT_ARGS etc. and just pass in one
> test file.  Seems tedious.  Would it be helpful to add a "--single-test"
> option (or something) to run-tests.py?
>
> And for running one test within a file, I know for the unit test files
> (like tests.py), I could modify the "main" section to have it run just one
> test, but would be nice to be able to do that from the command line.
> (maybe there is something similar for doctests, not sure.)  Again, could
> add a command line option to run-tests for that, though would be more work
> to plumb it through to each suite.
>
> thanks,
> Imran
>

Reply via email to