+1 I tested the source and Hadoop 2.4 release. Checksums and signatures are OK. Compiles fine with Java 8 on OS X. Tests... don't fail any more than usual.
FWIW I've also been using the 1.1.0-SNAPSHOT for some time in another project and have encountered no problems. I notice that the 1.1.0 release removes the CDH4-specific build, but adds two MapR-specific builds. Compare with https://dist.apache.org/repos/dist/release/spark/spark-1.0.2/ I commented on the commit: https://github.com/apache/spark/commit/ceb19830b88486faa87ff41e18d03ede713a73cc I'm in favor of removing all vendor-specific builds. This change *looks* a bit funny as there was no JIRA (?) and appears to swap one vendor for another. Of course there's nothing untoward going on, but what was the reasoning? It's best avoided, and MapR already distributes Spark just fine, no? This is a gray area with ASF projects. I mention it as well because it came up with Apache Flink recently (http://mail-archives.eu.apache.org/mod_mbox/incubator-flink-dev/201408.mbox/%3CCANC1h_u%3DN0YKFu3pDaEVYz5ZcQtjQnXEjQA2ReKmoS%2Bye7%3Do%3DA%40mail.gmail.com%3E) Another vendor rightly noted this could look like favoritism. They changed to remove vendor releases. On Fri, Aug 29, 2014 at 3:14 AM, Patrick Wendell <pwend...@gmail.com> wrote: > Please vote on releasing the following candidate as Apache Spark version > 1.1.0! > > The tag to be voted on is v1.1.0-rc2 (commit 711aebb3): > https://git-wip-us.apache.org/repos/asf?p=spark.git;a=commit;h=711aebb329ca28046396af1e34395a0df92b5327 > > The release files, including signatures, digests, etc. can be found at: > http://people.apache.org/~pwendell/spark-1.1.0-rc2/ > > Release artifacts are signed with the following key: > https://people.apache.org/keys/committer/pwendell.asc > > The staging repository for this release can be found at: > https://repository.apache.org/content/repositories/orgapachespark-1029/ > > The documentation corresponding to this release can be found at: > http://people.apache.org/~pwendell/spark-1.1.0-rc2-docs/ > > Please vote on releasing this package as Apache Spark 1.1.0! > > The vote is open until Monday, September 01, at 03:11 UTC and passes if > a majority of at least 3 +1 PMC votes are cast. > > [ ] +1 Release this package as Apache Spark 1.1.0 > [ ] -1 Do not release this package because ... > > To learn more about Apache Spark, please see > http://spark.apache.org/ > > == Regressions fixed since RC1 == > LZ4 compression issue: https://issues.apache.org/jira/browse/SPARK-3277 > > == What justifies a -1 vote for this release? == > This vote is happening very late into the QA period compared with > previous votes, so -1 votes should only occur for significant > regressions from 1.0.2. Bugs already present in 1.0.X will not block > this release. > > == What default changes should I be aware of? == > 1. The default value of "spark.io.compression.codec" is now "snappy" > --> Old behavior can be restored by switching to "lzf" > > 2. PySpark now performs external spilling during aggregations. > --> Old behavior can be restored by setting "spark.shuffle.spill" to "false". > > --------------------------------------------------------------------- > To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org > For additional commands, e-mail: dev-h...@spark.apache.org > --------------------------------------------------------------------- To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org For additional commands, e-mail: dev-h...@spark.apache.org