great job! thanks a lot!

On Thu, Dec 6, 2018 at 9:39 AM Hyukjin Kwon <gurwls...@gmail.com> wrote:

> It's merged now and in developer tools page -
> http://spark.apache.org/developer-tools.html#individual-tests
> Have some func with PySpark testing!
>
> 2018년 12월 5일 (수) 오후 4:30, Hyukjin Kwon <gurwls...@gmail.com>님이 작성:
>
>> Hey all, I kind of met the goal with a minimised fix with keeping
>> available framework and options. See
>>
>> https://github.com/apache/spark/pull/23203
>> https://github.com/apache/spark-website/pull/161
>>
>> I know it's not perfect and other Python testing framework provide many
>> good other features but should be good enough for now.
>> Thanks!
>>
>>
>> 2017년 8월 17일 (목) 오전 2:38, Nicholas Chammas <nicholas.cham...@gmail.com>님이
>> 작성:
>>
>>> Looks like it doesn’t take too much work to get pytest working on our
>>> code base, since it knows how to run unittest tests.
>>>
>>> https://github.com/apache/spark/compare/master…nchammas:pytest
>>> <https://github.com/apache/spark/compare/master...nchammas:pytest>
>>>
>>> For example I was able to do this from that branch and it did the right
>>> thing, running only the tests with string in their name:
>>>
>>> python [pytest *]$ ../bin/spark-submit ./pytest-run-tests.py 
>>> ./pyspark/sql/tests.py -v -k string
>>>
>>> However, looking more closely at the whole test setup, I’m hesitant to
>>> work any further on this.
>>>
>>> My intention was to see if we could leverage pytest, tox, and other test
>>> tools that are standard in the Python ecosystem to replace some of the
>>> homegrown stuff we have. We have our own test dependency tracking code, our
>>> own breakdown of tests into module-scoped chunks, and our own machinery to
>>> parallelize test execution. It seems like it would be a lot of work to reap
>>> the benefits of using the standard tools while ensuring that we don’t lose
>>> any of the benefits our current test setup provides.
>>>
>>> Nick
>>>
>>> On Tue, Aug 15, 2017 at 3:26 PM Bryan Cutler cutl...@gmail.com
>>> <http://mailto:cutl...@gmail.com> wrote:
>>>
>>> This generally works for me to just run tests within a class or even a
>>>> single test.  Not as flexible as pytest -k, which would be nice..
>>>>
>>>> $ SPARK_TESTING=1 bin/pyspark pyspark.sql.tests ArrowTests
>>>> On Tue, Aug 15, 2017 at 5:49 AM, Nicholas Chammas <
>>>> nicholas.cham...@gmail.com> wrote:
>>>>
>>>>> Pytest does support unittest-based tests
>>>>> <https://docs.pytest.org/en/latest/unittest.html>, allowing for
>>>>> incremental adoption. I'll see how convenient it is to use with our 
>>>>> current
>>>>> test layout.
>>>>>
>>>>> On Tue, Aug 15, 2017 at 1:03 AM Hyukjin Kwon <gurwls...@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> For me, I would like this if this can be done with relatively small
>>>>>> changes.
>>>>>> How about adding more granular options, for example, specifying or
>>>>>> filtering smaller set of test goals in the run-tests.py script?
>>>>>> I think it'd be quite small change and we could roughly reach this
>>>>>> goal if I understood correctly.
>>>>>>
>>>>>>
>>>>>> 2017-08-15 3:06 GMT+09:00 Nicholas Chammas <
>>>>>> nicholas.cham...@gmail.com>:
>>>>>>
>>>>>>> Say you’re working on something and you want to rerun the PySpark
>>>>>>> tests, focusing on a specific test or group of tests. Is there a way to 
>>>>>>> do
>>>>>>> that?
>>>>>>>
>>>>>>> I know that you can test entire modules with this:
>>>>>>>
>>>>>>> ./python/run-tests --modules pyspark-sql
>>>>>>>
>>>>>>> But I’m looking for something more granular, like pytest’s -k
>>>>>>> option.
>>>>>>>
>>>>>>> On that note, does anyone else think it would be valuable to use a
>>>>>>> test runner like pytest to run our Python tests? The biggest benefits 
>>>>>>> would
>>>>>>> be the use of fixtures
>>>>>>> <https://docs.pytest.org/en/latest/fixture.html>, and more
>>>>>>> flexibility on test running and reporting. Just wondering if we’ve 
>>>>>>> already
>>>>>>> considered this.
>>>>>>>
>>>>>>> Nick
>>>>>>> ​
>>>>>>>
>>>>>>
>>>>>> ​
>>>
>>

Reply via email to