For a shorter reproducer ...

df <- createDataFrame(list(list(1L, 1, "1", 0.1)), c("a", "b", "c", "d"))
collect(gapply(df, "a", function(key, x) { x }, schema(df)))

And running the below multiple times (5~7):

collect(gapply(df, "a", function(key, x) { x }, schema(df)))

looks occasionally throwing an error.


I will leave here and probably explain more information if a JIRA is open.
This does not look a regression anyway.



2017-06-14 16:22 GMT+09:00 Hyukjin Kwon <gurwls...@gmail.com>:

>
> Per https://github.com/apache/spark/tree/v2.1.1,
>
> 1. CentOS 7.2.1511 / R 3.3.3 - this test hangs.
>
> I messed it up a bit while downgrading the R to 3.3.3 (It was an actual
> machine not a VM) so it took me a while to re-try this.
> I re-built this again and checked the R version is 3.3.3 at least. I hope
> this one could double checked.
>
> Here is the self-reproducer:
>
> irisDF <- suppressWarnings(createDataFrame (iris))
> schema <-  structType(structField("Sepal_Length", "double"),
> structField("Avg", "double"))
> df4 <- gapply(
>   cols = "Sepal_Length",
>   irisDF,
>   function(key, x) {
>     y <- data.frame(key, mean(x$Sepal_Width), stringsAsFactors = FALSE)
>   },
>   schema)
> collect(df4)
>
>
>
> 2017-06-14 16:07 GMT+09:00 Felix Cheung <felixcheun...@hotmail.com>:
>
>> Thanks! Will try to setup RHEL/CentOS to test it out
>>
>> _____________________________
>> From: Nick Pentreath <nick.pentre...@gmail.com>
>> Sent: Tuesday, June 13, 2017 11:38 PM
>> Subject: Re: [VOTE] Apache Spark 2.2.0 (RC4)
>> To: Felix Cheung <felixcheun...@hotmail.com>, Hyukjin Kwon <
>> gurwls...@gmail.com>, dev <dev@spark.apache.org>
>>
>> Cc: Sean Owen <so...@cloudera.com>
>>
>>
>> Hi yeah sorry for slow response - I was RHEL and OpenJDK but will have to
>> report back later with the versions as am AFK.
>>
>> R version not totally sure but again will revert asap
>> On Wed, 14 Jun 2017 at 05:09, Felix Cheung <felixcheun...@hotmail.com>
>> wrote:
>>
>>> Thanks
>>> This was with an external package and unrelated
>>>
>>>   >> macOS Sierra 10.12.3 / R 3.2.3 - passed with a warning (
>>> https://gist.github.com/HyukjinKwon/85cbcfb245825852df20ed6a9ecfd845)
>>>
>>> As for CentOS - would it be possible to test against R older than 3.4.0?
>>> This is the same error reported by Nick below.
>>>
>>> _____________________________
>>> From: Hyukjin Kwon <gurwls...@gmail.com>
>>> Sent: Tuesday, June 13, 2017 8:02 PM
>>>
>>> Subject: Re: [VOTE] Apache Spark 2.2.0 (RC4)
>>> To: dev <dev@spark.apache.org>
>>> Cc: Sean Owen <so...@cloudera.com>, Nick Pentreath <
>>> nick.pentre...@gmail.com>, Felix Cheung <felixcheun...@hotmail.com>
>>>
>>>
>>>
>>> For the test failure on R, I checked:
>>>
>>>
>>> Per https://github.com/apache/spark/tree/v2.2.0-rc4,
>>>
>>> 1. Windows Server 2012 R2 / R 3.3.1 - passed (
>>> https://ci.appveyor.com/project/spark-test/spark/build/755-
>>> r-test-v2.2.0-rc4)
>>> 2. macOS Sierra 10.12.3 / R 3.4.0 - passed
>>> 3. macOS Sierra 10.12.3 / R 3.2.3 - passed with a warning (
>>> https://gist.github.com/HyukjinKwon/85cbcfb245825852df20ed6a9ecfd845)
>>> 4. CentOS 7.2.1511 / R 3.4.0 - reproduced (https://gist.github.com/Hyukj
>>> inKwon/2a736b9f80318618cc147ac2bb1a987d)
>>>
>>>
>>> Per https://github.com/apache/spark/tree/v2.1.1,
>>>
>>> 1. CentOS 7.2.1511 / R 3.4.0 - reproduced (https://gist.github.com/Hyukj
>>> inKwon/6064b0d10bab8fc1dc6212452d83b301)
>>>
>>>
>>> This looks being failed only in CentOS 7.2.1511 / R 3.4.0 given my tests
>>> and observations.
>>>
>>> This is failed in Spark 2.1.1. So, it sounds not a regression although
>>> it is a bug that should be fixed (whether in Spark or R).
>>>
>>>
>>> 2017-06-14 8:28 GMT+09:00 Xiao Li <gatorsm...@gmail.com>:
>>>
>>>> -1
>>>>
>>>> Spark 2.2 is unable to read the partitioned table created by Spark 2.1
>>>> or earlier.
>>>>
>>>> Opened a JIRA https://issues.apache.org/jira/browse/SPARK-21085
>>>>
>>>> Will fix it soon.
>>>>
>>>> Thanks,
>>>>
>>>> Xiao Li
>>>>
>>>>
>>>>
>>>> 2017-06-13 9:39 GMT-07:00 Joseph Bradley <jos...@databricks.com>:
>>>>
>>>>> Re: the QA JIRAs:
>>>>> Thanks for discussing them.  I still feel they are very helpful; I
>>>>> particularly notice not having to spend a solid 2-3 weeks of time QAing
>>>>> (unlike in earlier Spark releases).  One other point not mentioned above: 
>>>>> I
>>>>> think they serve as a very helpful reminder/training for the community for
>>>>> rigor in development.  Since we instituted QA JIRAs, contributors have 
>>>>> been
>>>>> a lot better about adding in docs early, rather than waiting until the end
>>>>> of the cycle (though I know this is drawing conclusions from 
>>>>> correlations).
>>>>>
>>>>> I would vote in favor of the RC...but I'll wait to see about the
>>>>> reported failures.
>>>>>
>>>>> On Fri, Jun 9, 2017 at 3:30 PM, Sean Owen <so...@cloudera.com> wrote:
>>>>>
>>>>>> Different errors as in https://issues.apache.org/j
>>>>>> ira/browse/SPARK-20520 but that's also reporting R test failures.
>>>>>>
>>>>>> I went back and tried to run the R tests and they passed, at least on
>>>>>> Ubuntu 17 / R 3.3.
>>>>>>
>>>>>>
>>>>>> On Fri, Jun 9, 2017 at 9:12 AM Nick Pentreath <
>>>>>> nick.pentre...@gmail.com> wrote:
>>>>>>
>>>>>>> All Scala, Python tests pass. ML QA and doc issues are resolved (as
>>>>>>> well as R it seems).
>>>>>>>
>>>>>>> However, I'm seeing the following test failure on R consistently:
>>>>>>> https://gist.github.com/MLnick/5f26152f97ae8473f807c6895817cf72
>>>>>>>
>>>>>>>
>>>>>>> On Thu, 8 Jun 2017 at 08:48 Denny Lee <denny.g....@gmail.com> wrote:
>>>>>>>
>>>>>>>> +1 non-binding
>>>>>>>>
>>>>>>>> Tested on macOS Sierra, Ubuntu 16.04
>>>>>>>> test suite includes various test cases including Spark SQL, ML,
>>>>>>>> GraphFrames, Structured Streaming
>>>>>>>>
>>>>>>>>
>>>>>>>> On Wed, Jun 7, 2017 at 9:40 PM vaquar khan <vaquar.k...@gmail.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> +1 non-binding
>>>>>>>>>
>>>>>>>>> Regards,
>>>>>>>>> vaquar khan
>>>>>>>>>
>>>>>>>>> On Jun 7, 2017 4:32 PM, "Ricardo Almeida" <
>>>>>>>>> ricardo.alme...@actnowib.com> wrote:
>>>>>>>>>
>>>>>>>>> +1 (non-binding)
>>>>>>>>>
>>>>>>>>> Built and tested with -Phadoop-2.7 -Dhadoop.version=2.7.3 -Pyarn
>>>>>>>>> -Phive -Phive-thriftserver -Pscala-2.11 on
>>>>>>>>>
>>>>>>>>>    - Ubuntu 17.04, Java 8 (OpenJDK 1.8.0_111)
>>>>>>>>>    - macOS 10.12.5 Java 8 (build 1.8.0_131)
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On 5 June 2017 at 21:14, Michael Armbrust <mich...@databricks.com>
>>>>>>>>> wrote:
>>>>>>>>>
>>>>>>>>>> Please vote on releasing the following candidate as Apache Spark
>>>>>>>>>> version 2.2.0. The vote is open until Thurs, June 8th, 2017 at
>>>>>>>>>> 12:00 PST and passes if a majority of at least 3 +1 PMC votes are
>>>>>>>>>> cast.
>>>>>>>>>>
>>>>>>>>>> [ ] +1 Release this package as Apache Spark 2.2.0
>>>>>>>>>> [ ] -1 Do not release this package because ...
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> To learn more about Apache Spark, please see
>>>>>>>>>> http://spark.apache.org/
>>>>>>>>>>
>>>>>>>>>> The tag to be voted on is v2.2.0-rc4
>>>>>>>>>> <https://github.com/apache/spark/tree/v2.2.0-rc4> (
>>>>>>>>>> 377cfa8ac7ff7a8a6a6d273182e18ea7dc25ce7e)
>>>>>>>>>>
>>>>>>>>>> List of JIRA tickets resolved can be found with this filter
>>>>>>>>>> <https://issues.apache.org/jira/browse/SPARK-20134?jql=project%20%3D%20SPARK%20AND%20fixVersion%20%3D%202.2.0>
>>>>>>>>>> .
>>>>>>>>>>
>>>>>>>>>> The release files, including signatures, digests, etc. can be
>>>>>>>>>> found at:
>>>>>>>>>> http://home.apache.org/~pwendell/spark-releases/spark-2.2.0-
>>>>>>>>>> rc4-bin/
>>>>>>>>>>
>>>>>>>>>> Release artifacts are signed with the following key:
>>>>>>>>>> https://people.apache.org/keys/committer/pwendell.asc
>>>>>>>>>>
>>>>>>>>>> The staging repository for this release can be found at:
>>>>>>>>>> https://repository.apache.org/content/repositories/orgapache
>>>>>>>>>> spark-1241/
>>>>>>>>>>
>>>>>>>>>> The documentation corresponding to this release can be found at:
>>>>>>>>>> http://people.apache.org/~pwendell/spark-releases/spark-2.2.
>>>>>>>>>> 0-rc4-docs/
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> *FAQ*
>>>>>>>>>>
>>>>>>>>>> *How can I help test this release?*
>>>>>>>>>>
>>>>>>>>>> If you are a Spark user, you can help us test this release by
>>>>>>>>>> taking an existing Spark workload and running on this release 
>>>>>>>>>> candidate,
>>>>>>>>>> then reporting any regressions.
>>>>>>>>>>
>>>>>>>>>> *What should happen to JIRA tickets still targeting 2.2.0?*
>>>>>>>>>>
>>>>>>>>>> Committers should look at those and triage. Extremely important
>>>>>>>>>> bug fixes, documentation, and API tweaks that impact compatibility 
>>>>>>>>>> should
>>>>>>>>>> be worked on immediately. Everything else please retarget to 2.3.0 
>>>>>>>>>> or 2.2.1.
>>>>>>>>>>
>>>>>>>>>> *But my bug isn't fixed!??!*
>>>>>>>>>>
>>>>>>>>>> In order to make timely releases, we will typically not hold the
>>>>>>>>>> release unless the bug in question is a regression from 2.1.1.
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>
>>>>>
>>>>> --
>>>>>
>>>>> Joseph Bradley
>>>>>
>>>>> Software Engineer - Machine Learning
>>>>>
>>>>> Databricks, Inc.
>>>>>
>>>>> [image: http://databricks.com] <http://databricks.com/>
>>>>>
>>>>
>>>>
>>>
>>>
>>>
>>
>>
>

Reply via email to