+1

I verified the checksum & signature and ran some of the example programs. I also ran RAT to check that sources are licensed correctly.

I only examined the merged artifact.  Do we need to release the others?

The out-of-box experience isn't great. There's no documentation at the top-level telling how to get started. In a subsequent release we might add links to the various documents.

Also, 'bin/start-all.sh' warns that one should use start-dfs.sh and start-mapred.sh instead, but those scripts, out-of-the-box, complain that that they can't find Common.

There are some empty java source files in the release, namely:

./common/src/java/org/apache/hadoop/fs/ftp/FTPFileSystemConfigKeys.java

./hdfs/src/java/org/apache/hadoop/hdfs/server/namenode/metrics/NameNodeStatisticsMBean.java
./hdfs/src/java/org/apache/hadoop/hdfs/server/namenode/metrics/NameNodeStatistics.java
./hdfs/src/java/org/apache/hadoop/hdfs/server/datanode/metrics/DataNodeStatistics.java
./hdfs/src/java/org/apache/hadoop/hdfs/server/datanode/metrics/DataNodeStatisticsMBean.java

These issues should not prevent this release, in my opinion.

Doug

On 08/16/2010 10:29 PM, Tom White wrote:
I have created a new candidate build for Hadoop 0.21.0. This fixes
MAPREDUCE-2012 and MAPREDUCE-2014 which were found for the previous
release candidate (1).

This release is being classified as a minor release, which means that
it is API compatible with 0.20.2.

*** This release candidate has not been tested extensively, so it
should not be considered stable.
*** Please download, test and vote before the vote closes on Thursday 19 August.

http://people.apache.org/~tomwhite/hadoop-0.21.0-candidate-2/

The hadoop-0.21.0.tar.gz file is an old-style combined release which
includes Common, HDFS, and MapReduce.

Thanks,
Tom

Reply via email to