On 14 Nov 2017, at 00:10, Arun Suresh <asur...@apache.org<mailto:asur...@apache.org>> wrote:
Hi Folks, Apache Hadoop 2.9.0 is the first release of Hadoop 2.9 line and will be the starting release for Apache Hadoop 2.9.x line - it includes 30 New Features with 500+ subtasks, 407 Improvements, 790 Bug fixes new fixed issues since 2.8.2. More information about the 2.9.0 release plan can be found here: *https://cwiki.apache.org/confluence/display/HADOOP/Roadmap#Roadmap-Version2.9 <https://cwiki.apache.org/confluence/display/HADOOP/Roadmap#Roadmap-Version2.9>* New RC is available at: *https://home.apache.org/~asuresh/hadoop-2.9.0-RC3/ <https://home.apache.org/~asuresh/hadoop-2.9.0-RC3/>* The RC tag in git is: release-2.9.0-RC3, and the latest commit id is: 756ebc8394e473ac25feac05fa493f6d612e6c50. The maven artifacts are available via repository.apache.org<http://repository.apache.org> at: <https://www.google.com/url?q=https%3A%2F%2Frepository.apache.org%2Fcontent%2Frepositories%2Forgapachehadoop-1066&sa=D&sntz=1&usg=AFQjCNFcern4uingMV_sEreko_zeLlgdlg>*https://repository.apache.org/content/repositories/orgapachehadoop-1068/ <https://repository.apache.org/content/repositories/orgapachehadoop-1068/>* We are carrying over the votes from the previous RC given that the delta is the license fix. Given the above - we are also going to stick with the original deadline for the vote : ending on Friday 17th November 2017 2pm PT time. Thanks, -Arun/Subru I hadn't finished my testing yet; I'd been assuming that it was 5 days from thw new RC. I believe every RC should (must?) still have that 5 day test period. I have done all my core tests, but haven't done the final tests of the S3Guard CLI tests. I did have to actually put in an afternoon making my spark cloud integration tests work against branch-2, so it wasn't like I could just D/L and test in half an hour As it was, I had filed one minor bug, but didnt' consider that an issue as it was only with the new FileSystem.create(path)" Builder. I was going to go vote +! unless those CLI tests were completly broken. What I do want is release notes to highlight which things we think unstable/experimental features to use with caution 1. Filesystem.create(Path) 2. S3Guard 3. AlyiunOSS (it hasn't been out long enough to be trusted) We ourselves know what's still stabilising, other should too Download: D/L tar file, & .asc, check signature pg2 --verify hadoop-2.9.0.tar.gz.asc gpg: assuming signed data in 'hadoop-2.9.0.tar.gz' gpg: Signature made Mon 13 Nov 23:45:49 2017 GMT gpg: using RSA key 0x7ECDEEEA64ECB6E6 gpg: Good signature from "Arun Suresh <asur...@apache.org<mailto:asur...@apache.org>>" [unknown] gpg: WARNING: This key is not certified with a trusted signature! gpg: There is no indication that the signature belongs to the owner. Primary key fingerprint: 412B BFB1 27CB 48DA 6BA2 E3EE 7ECD EEEA 64EC B6E6 This is a valid signature; Arun is listed in KEYS. He's not trusted though: at the time of checking his public key hadn't been signed by anyone I trust though it shows we need more cross-trust between them. (Since then Andrew Wang has authenticated him, so he is now transitively trusted). Downstream build Clean build of Spark through the staged artifacts (must remember to rm the local copies when the release ships). This verifies artifacts and source code compatibility. mvn clean install -Pyarn,hadoop-2.7,hadoop-cloud,snapshots-and-staging -Dhadoop.version=2.9.0 Everything compiled fine. I did not do a full test run; I didn't have the time and without knowing whether everything worked locally on the 2.8.1 release, could have been a distraction on a failure. Downstream tests: Ran all my spark cloud integration tests: s3 (S3 ireland + S3Guard) , azure (Azure ireland(, swift (RAX US); https://github.com/hortonworks-spark/cloud-integration/tree/branch-2 there were 3 failures of SparkSQL & ORC in test setup, but I think that's related to the latest spark code & my injected-into-org.apache.spark-namespace tests; they're a bit brittle to changes in spark-master. All the work used the "legacy" FileOutputCommitter; with s3guard turned on. This delivers the consistency needed for Job and task commit to be correct, though without the performance needed for it to be usable in production. That's dependent on HADOOP-13786, which I don't propose to backport. (it loves its Java 8 & targets 3.1) hadoop cloud module object store tests Checked out the relevant git commit & did a local run of hadoop-aws, hadoop-azure and hadoop-openstack hadoop-amazon * s3a happy, including s3guard * All the s3n & s3:// tests skipped/failed because I didn't have the test bindings for those set up: untested. Azure wasb: happy; one transient failure which ADLs: Not set up test this OSS: not set up to test this Swift: Found one test failure due to swift not supporting FileSystem.createNonRecursive(); this means the new builder based FileSystem.create(Path) mechanism doesn't work for it. Nothing that serious, given its an experimental API, but it does show that we are undertesting this stuff. https://issues.apache.org/jira/browse/HADOOP-15050 Note: the swift tests I ran downstream with spark did work. This is only the new API call which fails.