Yes! This is very helpful! On Wed, Dec 5, 2018 at 9:21 PM Wenchen Fan <cloud0...@gmail.com> wrote:
> great job! thanks a lot! > > On Thu, Dec 6, 2018 at 9:39 AM Hyukjin Kwon <gurwls...@gmail.com> wrote: > >> It's merged now and in developer tools page - >> http://spark.apache.org/developer-tools.html#individual-tests >> Have some func with PySpark testing! >> >> 2018년 12월 5일 (수) 오후 4:30, Hyukjin Kwon <gurwls...@gmail.com>님이 작성: >> >>> Hey all, I kind of met the goal with a minimised fix with keeping >>> available framework and options. See >>> >>> https://github.com/apache/spark/pull/23203 >>> https://github.com/apache/spark-website/pull/161 >>> >>> I know it's not perfect and other Python testing framework provide many >>> good other features but should be good enough for now. >>> Thanks! >>> >>> >>> 2017년 8월 17일 (목) 오전 2:38, Nicholas Chammas <nicholas.cham...@gmail.com>님이 >>> 작성: >>> >>>> Looks like it doesn’t take too much work to get pytest working on our >>>> code base, since it knows how to run unittest tests. >>>> >>>> https://github.com/apache/spark/compare/master…nchammas:pytest >>>> <https://github.com/apache/spark/compare/master...nchammas:pytest> >>>> >>>> For example I was able to do this from that branch and it did the right >>>> thing, running only the tests with string in their name: >>>> >>>> python [pytest *]$ ../bin/spark-submit ./pytest-run-tests.py >>>> ./pyspark/sql/tests.py -v -k string >>>> >>>> However, looking more closely at the whole test setup, I’m hesitant to >>>> work any further on this. >>>> >>>> My intention was to see if we could leverage pytest, tox, and other >>>> test tools that are standard in the Python ecosystem to replace some of the >>>> homegrown stuff we have. We have our own test dependency tracking code, our >>>> own breakdown of tests into module-scoped chunks, and our own machinery to >>>> parallelize test execution. It seems like it would be a lot of work to reap >>>> the benefits of using the standard tools while ensuring that we don’t lose >>>> any of the benefits our current test setup provides. >>>> >>>> Nick >>>> >>>> On Tue, Aug 15, 2017 at 3:26 PM Bryan Cutler cutl...@gmail.com >>>> <http://mailto:cutl...@gmail.com> wrote: >>>> >>>> This generally works for me to just run tests within a class or even a >>>>> single test. Not as flexible as pytest -k, which would be nice.. >>>>> >>>>> $ SPARK_TESTING=1 bin/pyspark pyspark.sql.tests ArrowTests >>>>> On Tue, Aug 15, 2017 at 5:49 AM, Nicholas Chammas < >>>>> nicholas.cham...@gmail.com> wrote: >>>>> >>>>>> Pytest does support unittest-based tests >>>>>> <https://docs.pytest.org/en/latest/unittest.html>, allowing for >>>>>> incremental adoption. I'll see how convenient it is to use with our >>>>>> current >>>>>> test layout. >>>>>> >>>>>> On Tue, Aug 15, 2017 at 1:03 AM Hyukjin Kwon <gurwls...@gmail.com> >>>>>> wrote: >>>>>> >>>>>>> For me, I would like this if this can be done with relatively small >>>>>>> changes. >>>>>>> How about adding more granular options, for example, specifying or >>>>>>> filtering smaller set of test goals in the run-tests.py script? >>>>>>> I think it'd be quite small change and we could roughly reach this >>>>>>> goal if I understood correctly. >>>>>>> >>>>>>> >>>>>>> 2017-08-15 3:06 GMT+09:00 Nicholas Chammas < >>>>>>> nicholas.cham...@gmail.com>: >>>>>>> >>>>>>>> Say you’re working on something and you want to rerun the PySpark >>>>>>>> tests, focusing on a specific test or group of tests. Is there a way >>>>>>>> to do >>>>>>>> that? >>>>>>>> >>>>>>>> I know that you can test entire modules with this: >>>>>>>> >>>>>>>> ./python/run-tests --modules pyspark-sql >>>>>>>> >>>>>>>> But I’m looking for something more granular, like pytest’s -k >>>>>>>> option. >>>>>>>> >>>>>>>> On that note, does anyone else think it would be valuable to use a >>>>>>>> test runner like pytest to run our Python tests? The biggest benefits >>>>>>>> would >>>>>>>> be the use of fixtures >>>>>>>> <https://docs.pytest.org/en/latest/fixture.html>, and more >>>>>>>> flexibility on test running and reporting. Just wondering if we’ve >>>>>>>> already >>>>>>>> considered this. >>>>>>>> >>>>>>>> Nick >>>>>>>> >>>>>>>> >>>>>>> >>>>>>> >>>> >>> -- [image: Spark+AI Summit North America 2019] <http://t.sidekickopen24.com/s1t/c/5/f18dQhb0S7lM8dDMPbW2n0x6l2B9nMJN7t5X-FfhMynN2z8MDjQsyTKW56dzQQ1-_gV6102?t=https%3A%2F%2Fdatabricks.com%2Fsparkaisummit%2Fnorth-america&si=undefined&pi=406b8c9a-b648-4923-9ed1-9a51ffe213fa>