Hi, Please check the relevant section of the developer tools docs:
https://spark.apache.org/developer-tools.html#running-individual-tests On 1/25/22 00:44, Fangjia Shen wrote: > Hello all, > > How do you run Spark's test suites when you want to test the correctness > of your code? Is there a way to run a specific test suite for Spark? For > example, running test suite XXXSuite alone, instead of every class under > the test/ directories. > > Here's some background info about what I want to do: I'm a graduate > student trying to study Spark's design and find ways to improve Spark's > performance by doing Software/Hardware co-design. I'm relatively new to > Maven and so far struggling to find to a way to properly run Spark's own > test suites. > > Let's say I did some modifications to a XXXExec node which belongs to > the org.apache.spark.sql package. I want to see if my design passes the > test cases. What should I do? > > > What command should I use: > > *<spark_root>/build/mvn test * or *<spark_root>/dev/run-tests* ? > > And where should I run that command: > > *<spark_root>* or *<package_dir>* ? - where <package_dir> is where > the modified scala file is located, e.g. "<spark_root>/sql/core/". > > > I tried adding -Dtest=XXXSuite to *mvn test *but still get to run tens > of thousands of tests. This is taking way too much time and unbearable > if I'm just modifying a few file in a specific module. > > I would really appreciate any suggestion or comment. > > > Best regards, > > Fangjia Shen > > Purdue University > > > -- Best regards, Maciej Szymkiewicz Web: https://zero323.net PGP: A30CEF0C31A501EC
OpenPGP_signature
Description: OpenPGP digital signature