For the test failure on R, I checked:
Per https://github.com/apache/spark/tree/v2.2.0-rc4, 1. Windows Server 2012 R2 / R 3.3.1 - passed ( https://ci.appveyor.com/project/spark-test/spark/build/755-r-test-v2.2.0-rc4 ) 2. macOS Sierra 10.12.3 / R 3.4.0 - passed 3. macOS Sierra 10.12.3 / R 3.2.3 - passed with a warning ( https://gist.github.com/HyukjinKwon/85cbcfb245825852df20ed6a9ecfd845) 4. CentOS 7.2.1511 / R 3.4.0 - reproduced ( https://gist.github.com/HyukjinKwon/2a736b9f80318618cc147ac2bb1a987d) Per https://github.com/apache/spark/tree/v2.1.1, 1. CentOS 7.2.1511 / R 3.4.0 - reproduced ( https://gist.github.com/HyukjinKwon/6064b0d10bab8fc1dc6212452d83b301) This looks being failed only in CentOS 7.2.1511 / R 3.4.0 given my tests and observations. This is failed in Spark 2.1.1. So, it sounds not a regression although it is a bug that should be fixed (whether in Spark or R). 2017-06-14 8:28 GMT+09:00 Xiao Li <gatorsm...@gmail.com>: > -1 > > Spark 2.2 is unable to read the partitioned table created by Spark 2.1 or > earlier. > > Opened a JIRA https://issues.apache.org/jira/browse/SPARK-21085 > > Will fix it soon. > > Thanks, > > Xiao Li > > > > 2017-06-13 9:39 GMT-07:00 Joseph Bradley <jos...@databricks.com>: > >> Re: the QA JIRAs: >> Thanks for discussing them. I still feel they are very helpful; I >> particularly notice not having to spend a solid 2-3 weeks of time QAing >> (unlike in earlier Spark releases). One other point not mentioned above: I >> think they serve as a very helpful reminder/training for the community for >> rigor in development. Since we instituted QA JIRAs, contributors have been >> a lot better about adding in docs early, rather than waiting until the end >> of the cycle (though I know this is drawing conclusions from correlations). >> >> I would vote in favor of the RC...but I'll wait to see about the reported >> failures. >> >> On Fri, Jun 9, 2017 at 3:30 PM, Sean Owen <so...@cloudera.com> wrote: >> >>> Different errors as in https://issues.apache.org/jira/browse/SPARK-20520 but >>> that's also reporting R test failures. >>> >>> I went back and tried to run the R tests and they passed, at least on >>> Ubuntu 17 / R 3.3. >>> >>> >>> On Fri, Jun 9, 2017 at 9:12 AM Nick Pentreath <nick.pentre...@gmail.com> >>> wrote: >>> >>>> All Scala, Python tests pass. ML QA and doc issues are resolved (as >>>> well as R it seems). >>>> >>>> However, I'm seeing the following test failure on R consistently: >>>> https://gist.github.com/MLnick/5f26152f97ae8473f807c6895817cf72 >>>> >>>> >>>> On Thu, 8 Jun 2017 at 08:48 Denny Lee <denny.g....@gmail.com> wrote: >>>> >>>>> +1 non-binding >>>>> >>>>> Tested on macOS Sierra, Ubuntu 16.04 >>>>> test suite includes various test cases including Spark SQL, ML, >>>>> GraphFrames, Structured Streaming >>>>> >>>>> >>>>> On Wed, Jun 7, 2017 at 9:40 PM vaquar khan <vaquar.k...@gmail.com> >>>>> wrote: >>>>> >>>>>> +1 non-binding >>>>>> >>>>>> Regards, >>>>>> vaquar khan >>>>>> >>>>>> On Jun 7, 2017 4:32 PM, "Ricardo Almeida" < >>>>>> ricardo.alme...@actnowib.com> wrote: >>>>>> >>>>>> +1 (non-binding) >>>>>> >>>>>> Built and tested with -Phadoop-2.7 -Dhadoop.version=2.7.3 -Pyarn >>>>>> -Phive -Phive-thriftserver -Pscala-2.11 on >>>>>> >>>>>> - Ubuntu 17.04, Java 8 (OpenJDK 1.8.0_111) >>>>>> - macOS 10.12.5 Java 8 (build 1.8.0_131) >>>>>> >>>>>> >>>>>> On 5 June 2017 at 21:14, Michael Armbrust <mich...@databricks.com> >>>>>> wrote: >>>>>> >>>>>>> Please vote on releasing the following candidate as Apache Spark >>>>>>> version 2.2.0. The vote is open until Thurs, June 8th, 2017 at >>>>>>> 12:00 PST and passes if a majority of at least 3 +1 PMC votes are >>>>>>> cast. >>>>>>> >>>>>>> [ ] +1 Release this package as Apache Spark 2.2.0 >>>>>>> [ ] -1 Do not release this package because ... >>>>>>> >>>>>>> >>>>>>> To learn more about Apache Spark, please see >>>>>>> http://spark.apache.org/ >>>>>>> >>>>>>> The tag to be voted on is v2.2.0-rc4 >>>>>>> <https://github.com/apache/spark/tree/v2.2.0-rc4> (377cfa8ac7ff7a8 >>>>>>> a6a6d273182e18ea7dc25ce7e) >>>>>>> >>>>>>> List of JIRA tickets resolved can be found with this filter >>>>>>> <https://issues.apache.org/jira/browse/SPARK-20134?jql=project%20%3D%20SPARK%20AND%20fixVersion%20%3D%202.2.0> >>>>>>> . >>>>>>> >>>>>>> The release files, including signatures, digests, etc. can be found >>>>>>> at: >>>>>>> http://home.apache.org/~pwendell/spark-releases/spark-2.2.0-rc4-bin/ >>>>>>> >>>>>>> Release artifacts are signed with the following key: >>>>>>> https://people.apache.org/keys/committer/pwendell.asc >>>>>>> >>>>>>> The staging repository for this release can be found at: >>>>>>> https://repository.apache.org/content/repositories/orgapache >>>>>>> spark-1241/ >>>>>>> >>>>>>> The documentation corresponding to this release can be found at: >>>>>>> http://people.apache.org/~pwendell/spark-releases/spark-2.2. >>>>>>> 0-rc4-docs/ >>>>>>> >>>>>>> >>>>>>> *FAQ* >>>>>>> >>>>>>> *How can I help test this release?* >>>>>>> >>>>>>> If you are a Spark user, you can help us test this release by taking >>>>>>> an existing Spark workload and running on this release candidate, then >>>>>>> reporting any regressions. >>>>>>> >>>>>>> *What should happen to JIRA tickets still targeting 2.2.0?* >>>>>>> >>>>>>> Committers should look at those and triage. Extremely important bug >>>>>>> fixes, documentation, and API tweaks that impact compatibility should be >>>>>>> worked on immediately. Everything else please retarget to 2.3.0 or >>>>>>> 2.2.1. >>>>>>> >>>>>>> *But my bug isn't fixed!??!* >>>>>>> >>>>>>> In order to make timely releases, we will typically not hold the >>>>>>> release unless the bug in question is a regression from 2.1.1. >>>>>>> >>>>>> >>>>>> >>>>>> >> >> >> -- >> >> Joseph Bradley >> >> Software Engineer - Machine Learning >> >> Databricks, Inc. >> >> [image: http://databricks.com] <http://databricks.com/> >> > >