It's also the key Andrew has in the project's KEYS file: http://www.apache.org/dist/hadoop/common/KEYS
On Tue, Aug 30, 2016 at 4:12 PM, Andrew Wang <andrew.w...@cloudera.com> wrote: > Hi Eric, thanks for trying this out, > > I tried this gpg command to get my key, seemed to work: > > # gpg --keyserver pgp.mit.edu --recv-keys 7501105C > gpg: requesting key 7501105C from hkp server pgp.mit.edu > gpg: /root/.gnupg/trustdb.gpg: trustdb created > gpg: key 7501105C: public key "Andrew Wang (CODE SIGNING KEY) < > andrew.w...@cloudera.com>" imported > gpg: no ultimately trusted keys found > gpg: Total number processed: 1 > gpg: imported: 1 (RSA: 1) > > Also found via search: > http://pgp.mit.edu/pks/lookup?search=wang%40apache.org&op=index > > > On Tue, Aug 30, 2016 at 2:06 PM, Eric Badger <ebad...@yahoo-inc.com> wrote: > >> I don't know why my email client keeps getting rid of all of my spacing. >> Resending the same email so that it is actually legible... >> >> All on OSX 10.11.6: >> - Verified the hashes. However, Andrew, I don't know where to find your >> public key, so I wasn't able to verify that they were signed by you. >> - Built from source >> - Deployed a pseudo-distributed clusterRan a few sample jobs >> - Poked around the RM UI >> - Poked around the attached website locally via the tarball >> >> >> I did find one odd thing, though. It could be a misconfiguration on my >> system, but I've never had this problem before with other releases (though >> I deal almost exclusively in 2.x and so I imagine things might be >> different). When I run a sleep job, I do not see any >> diagnostics/logs/counters printed out by the client. Initially I ran the >> job like I would on 2.7 and it failed (because I had not set >> yarn.app.mapreduce.am.env and mapreduce.admin.user.env), but I didn't see >> anything until I looked at the RM UI. There I was able to see all of the >> logs for the failed job and diagnose the issue. Then, once I fixed my >> parameters and ran the job again, I still didn't see any >> diagnostics/logs/counters. >> >> >> ebadger@foo: env | grep HADOOP >> HADOOP_HOME=/Users/ebadger/Downloads/hadoop-3.0.0-alpha1- >> src/hadoop-dist/target/hadoop-3.0.0-alpha1/ >> HADOOP_CONF_DIR=/Users/ebadger/conf >> ebadger@foo: $HADOOP_HOME/bin/hadoop jar $HADOOP_HOME/share/hadoop/ >> mapreduce/hadoop-mapreduce-client-jobclient-3.0.0-alpha1-tests.jar sleep >> -Dyarn.app.mapreduce.am.env="HADOOP_MAPRED_HOME=$HADOOP_HOME" >> -Dmapreduce.admin.user.env="HADOOP_MAPRED_HOME=$HADOOP_HOME" -mt 1 -rt 1 >> -m 1 -r 1 >> WARNING: log4j.properties is not found. HADOOP_CONF_DIR may be incomplete. >> ebadger@foo: >> >> >> After running the above command, the RM UI showed a successful job, but as >> you can see, I did not have anything printed onto the command line. >> Hopefully this is just a misconfiguration on my part, but I figured that I >> would point it out just in case. >> >> >> Thanks, >> >> >> Eric >> >> >> >> On Tuesday, August 30, 2016 4:00 PM, Eric Badger >> <ebad...@yahoo-inc.com.INVALID> wrote: >> >> >> >> All on OSX 10.11.6: >> Verified the hashes. However, Andrew, I don't know where to find your >> public key, so I wasn't able to verify that they were signed by you.Built >> from sourceDeployed a pseudo-distributed clusterRan a few sample jobsPoked >> around the RM UIPoked around the attached website locally via the tarball >> I did find one odd thing, though. It could be a misconfiguration on my >> system, but I've never had this problem before with other releases (though >> I deal almost exclusively in 2.x and so I imagine things might be >> different). When I run a sleep job, I do not see any >> diagnostics/logs/counters printed out by the client. Initially I ran the >> job like I would on 2.7 and it failed (because I had not set >> yarn.app.mapreduce.am.env and mapreduce.admin.user.env), but I didn't see >> anything until I looked at the RM UI. There I was able to see all of the >> logs for the failed job and diagnose the issue. Then, once I fixed my >> parameters and ran the job again, I still didn't see any >> diagnostics/logs/counters. >> ebadger@foo: env | grep HADOOPHADOOP_HOME=/Users/ >> ebadger/Downloads/hadoop-3.0.0-alpha1-src/hadoop-dist/ >> target/hadoop-3.0.0-alpha1/HADOOP_CONF_DIR=/Users/ebadger/confebadger@foo: >> $HADOOP_HOME/bin/hadoop jar $HADOOP_HOME/share/hadoop/ >> mapreduce/hadoop-mapreduce-client-jobclient-3.0.0-alpha1-tests.jar sleep >> -Dyarn.app.mapreduce.am.env="HADOOP_MAPRED_HOME=$HADOOP_HOME" >> -Dmapreduce.admin.user.env="HADOOP_MAPRED_HOME=$HADOOP_HOME" -mt 1 -rt 1 >> -m 1 -r 1WARNING: log4j.properties is not found. HADOOP_CONF_DIR may be >> incomplete.ebadger@foo: >> After running the above command, the RM UI showed a successful job, but as >> you can see, I did not have anything printed onto the command line. >> Hopefully this is just a misconfiguration on my part, but I figured that I >> would point it out just in case. >> Thanks, >> Eric >> >> >> >> On Tuesday, August 30, 2016 12:58 PM, Andrew Wang < >> andrew.w...@cloudera.com> wrote: >> >> >> I'll put my own +1 on it: >> >> * Built from source >> * Started pseudo cluster and ran Pi job successfully >> >> On Tue, Aug 30, 2016 at 10:17 AM, Zhe Zhang <z...@apache.org> wrote: >> >> > >> > Thanks Andrew for the great work! It's really exciting to finally see a >> > Hadoop 3 RC. >> > >> > I noticed CHANGES and RELEASENOTES markdown files which were not in >> > previous RCs like 2.7.3. What are good tools to verify them? I tried >> > reading them on IntelliJ but format looks odd. >> > >> > I'm still testing the RC: >> > - Downloaded and verified checksum >> > - Built from source >> > - Will start small cluster and test simple programs, focusing on EC >> > functionalities >> > >> > -- Zhe >> > >> > On Tue, Aug 30, 2016 at 8:51 AM Andrew Wang <andrew.w...@cloudera.com> >> > wrote: >> > >> >> Hi all, >> >> >> >> Thanks to the combined work of many, many contributors, here's an RC0 >> for >> >> 3.0.0-alpha1: >> >> >> >> http://home.apache.org/~wang/3.0.0-alpha1-RC0/ >> >> >> >> alpha1 is the first in a series of planned alpha releases leading up to >> >> GA. >> >> The objective is to get an artifact out to downstreams for testing and >> to >> >> iterate quickly based on their feedback. So, please keep that in mind >> when >> >> voting; hopefully most issues can be addressed by future alphas rather >> >> than >> >> future RCs. >> >> >> >> Sorry for getting this out on a Tuesday, but I'd still like this vote to >> >> run the normal 5 days, thus ending Saturday (9/3) at 9AM PDT. I'll >> extend >> >> if we lack the votes. >> >> >> >> Please try it out and let me know what you think. >> >> >> >> Best, >> >> Andrew >> >> >> > >> -- busbey --------------------------------------------------------------------- To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org