+1 (non-binding) Tested GraphX, build infrastructure, & core test suite on OSX 10.9 w/ Java 1.7/1.8
On 4/6/15, 5:21 AM, "Sean Owen" <so...@cloudera.com> wrote: >SPARK-6673 is not, in the end, relevant for 1.3.x I believe; we just >resolved it for 1.4 anyway. False alarm there. > >I back-ported SPARK-6205 into the 1.3 branch for next time. We'll pick >it up if there's another RC, but by itself is not something that needs >a new RC. (I will give the same treatment to branch 1.2 if needed in >light of the 1.2.2 release.) > >I applied the simple change in SPARK-6205 in order to continue >executing tests and all was well. I still see a few failures in Hive >tests: > >- show_create_table_serde *** FAILED *** >- show_tblproperties *** FAILED *** >- udf_std *** FAILED *** >- udf_stddev *** FAILED *** > >with ... > >mvn -Phadoop-2.4 -Pyarn -Phive -Phive-0.13.1 -Dhadoop.version=2.6.0 >-DskipTests clean package; mvn -Phadoop-2.4 -Pyarn -Phive >-Phive-0.13.1 -Dhadoop.version=2.6.0 test > >... but these are not regressions from 1.3.0. > >+1 from me at this point on the current artifacts. > >On Sun, Apr 5, 2015 at 9:24 AM, Sean Owen <so...@cloudera.com> wrote: >> Signatures and hashes are good. >> LICENSE, NOTICE still check out. >> Compiles for a Hadoop 2.6 + YARN + Hive profile. >> >> I still see the UISeleniumSuite test failure observed in 1.3.0, which >> is minor and already fixed. I don't know why I didn't back-port it: >> https://issues.apache.org/jira/browse/SPARK-6205 >> >> If we roll another, let's get this easy fix in, but it is only an >> issue with tests. >> >> >> On JIRA, I checked open issues with Fix Version = 1.3.0 or 1.3.1 and >> all look legitimate (e.g. reopened or in progress) >> >> >> There is 1 open Blocker for 1.3.1 per Andrew: >> https://issues.apache.org/jira/browse/SPARK-6673 spark-shell.cmd can't >> start even when spark was built in Windows >> >> I believe this can be resolved quickly but as a matter of hygiene >> should be fixed or demoted before release. >> >> >> FYI there are 16 Critical issues marked for 1.3.0 / 1.3.1; worth >> examining before release to see how critical they are: >> >> SPARK-6701,Flaky test: o.a.s.deploy.yarn.YarnClusterSuite Python >> application,,Open,4/3/15 >> SPARK-6484,"Ganglia metrics xml reporter doesn't escape >> correctly",Josh Rosen,Open,3/24/15 >> SPARK-6270,Standalone Master hangs when streaming job >>completes,,Open,3/11/15 >> SPARK-6209,ExecutorClassLoader can leak connections after failing to >> load classes from the REPL class server,Josh Rosen,In Progress,4/2/15 >> SPARK-5113,Audit and document use of hostnames and IP addresses in >> Spark,,Open,3/24/15 >> SPARK-5098,Number of running tasks become negative after tasks >> lost,,Open,1/14/15 >> SPARK-4925,Publish Spark SQL hive-thriftserver maven artifact,Patrick >> Wendell,Reopened,3/23/15 >> SPARK-4922,Support dynamic allocation for coarse-grained >>Mesos,,Open,3/31/15 >> SPARK-4888,"Spark EC2 doesn't mount local disks for i2.8xlarge >> instances",,Open,1/27/15 >> SPARK-4879,Missing output partitions after job completes with >> speculative execution,Josh Rosen,Open,3/5/15 >> SPARK-4751,Support dynamic allocation for standalone mode,Andrew >> Or,Open,12/22/14 >> SPARK-4454,Race condition in DAGScheduler,Josh Rosen,Reopened,2/18/15 >> SPARK-4452,Shuffle data structures can starve others on the same >> thread for memory,Tianshuo Deng,Open,1/24/15 >> SPARK-4352,Incorporate locality preferences in dynamic allocation >> requests,,Open,1/26/15 >> SPARK-4227,Document external shuffle service,,Open,3/23/15 >> SPARK-3650,Triangle Count handles reverse edges >>incorrectly,,Open,2/23/15 >> >> On Sun, Apr 5, 2015 at 1:09 AM, Patrick Wendell <pwend...@gmail.com> >>wrote: >>> Please vote on releasing the following candidate as Apache Spark >>>version 1.3.1! >>> >>> The tag to be voted on is v1.3.1-rc1 (commit 0dcb5d9f): >>> >>>https://git-wip-us.apache.org/repos/asf?p=spark.git;a=commit;h=0dcb5d9f3 >>>1b713ed90bcec63ebc4e530cbb69851 >>> >>> The list of fixes present in this release can be found at: >>> http://bit.ly/1C2nVPY >>> >>> The release files, including signatures, digests, etc. can be found at: >>> http://people.apache.org/~pwendell/spark-1.3.1-rc1/ >>> >>> Release artifacts are signed with the following key: >>> https://people.apache.org/keys/committer/pwendell.asc >>> >>> The staging repository for this release can be found at: >>> https://repository.apache.org/content/repositories/orgapachespark-1080 >>> >>> The documentation corresponding to this release can be found at: >>> http://people.apache.org/~pwendell/spark-1.3.1-rc1-docs/ >>> >>> Please vote on releasing this package as Apache Spark 1.3.1! >>> >>> The vote is open until Wednesday, April 08, at 01:10 UTC and passes >>> if a majority of at least 3 +1 PMC votes are cast. >>> >>> [ ] +1 Release this package as Apache Spark 1.3.1 >>> [ ] -1 Do not release this package because ... >>> >>> To learn more about Apache Spark, please see >>> http://spark.apache.org/ >>> >>> - Patrick >>> >>> --------------------------------------------------------------------- >>> To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org >>> For additional commands, e-mail: dev-h...@spark.apache.org >>> > >--------------------------------------------------------------------- >To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org >For additional commands, e-mail: dev-h...@spark.apache.org > ________________________________________________________ The information contained in this e-mail is confidential and/or proprietary to Capital One and/or its affiliates. The information transmitted herewith is intended only for use by the individual or entity to which it is addressed. If the reader of this message is not the intended recipient, you are hereby notified that any review, retransmission, dissemination, distribution, copying or other use of, or taking of any action in reliance upon this information is strictly prohibited. If you have received this communication in error, please contact the sender and delete the material from your computer. --------------------------------------------------------------------- To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org For additional commands, e-mail: dev-h...@spark.apache.org