Thanks Josh. I downloaded spark-1.6.1-bin-hadoop2.3.tgz and spark-1.6.1-bin-hadoop2.4.tgz which expand without error.
On Sat, Apr 16, 2016 at 4:54 PM, Josh Rosen <joshro...@databricks.com> wrote: > Using a different machine / toolchain, I've downloaded and re-uploaded all > of the 1.6.1 artifacts to that S3 bucket, so hopefully everything should be > working now. Let me know if you still encounter any problems with > unarchiving. > > On Sat, Apr 16, 2016 at 3:10 PM Ted Yu <yuzhih...@gmail.com> wrote: > >> Pardon me - there is no tarball for hadoop 2.7 >> >> I downloaded >> https://s3.amazonaws.com/spark-related-packages/spark-1.6.1-bin-hadoop2.6.tgz >> and successfully expanded it. >> >> FYI >> >> On Sat, Apr 16, 2016 at 2:52 PM, Jon Gregg <jonrgr...@gmail.com> wrote: >> >>> That link points to hadoop2.6.tgz. I tried changing the URL to >>> https://s3.amazonaws.com/spark-related-packages/spark-1.6.1-bin-hadoop2.7.tgz >>> and I get a NoSuchKey error. >>> >>> Should I just go with it even though it says hadoop2.6? >>> >>> On Sat, Apr 16, 2016 at 5:37 PM, Ted Yu <yuzhih...@gmail.com> wrote: >>> >>>> BTW this was the original thread: >>>> >>>> http://search-hadoop.com/m/q3RTt0Oxul0W6Ak >>>> >>>> The link for spark-1.6.1-bin-hadoop2.7 is >>>> https://s3.amazonaws.com/spark-related-packages/spark-1.6.1-bin-hadoop2.7.tgz >>>> <https://s3.amazonaws.com/spark-related-packages/spark-1.6.1-bin-hadoop2.6.tgz> >>>> >>>> On Sat, Apr 16, 2016 at 2:14 PM, Ted Yu <yuzhih...@gmail.com> wrote: >>>> >>>>> From the output you posted: >>>>> ------- >>>>> Unpacking Spark >>>>> >>>>> gzip: stdin: not in gzip format >>>>> tar: Child returned status 1 >>>>> tar: Error is not recoverable: exiting now >>>>> ------- >>>>> >>>>> The artifact for spark-1.6.1-bin-hadoop2.6 is corrupt. >>>>> >>>>> This problem has been reported in other threads. >>>>> >>>>> Try spark-1.6.1-bin-hadoop2.7 - the artifact should be good. >>>>> >>>>> On Sat, Apr 16, 2016 at 2:09 PM, YaoPau <jonrgr...@gmail.com> wrote: >>>>> >>>>>> I launched a cluster with: "./spark-ec2 --key-pair my_pem >>>>>> --identity-file >>>>>> ../../ssh/my_pem.pem launch jg_spark2" and I got the "Spark standalone >>>>>> cluster started at >>>>>> http://ec2-54-88-249-255.compute-1.amazonaws.com:8080" >>>>>> and "Done!" success confirmations at the end. I confirmed on EC2 >>>>>> that 1 >>>>>> Master and 1 Slave were both launched and passed their status checks. >>>>>> >>>>>> But none of the Spark commands seem to work (spark-shell, pyspark, >>>>>> etc), and >>>>>> port 8080 isn't being used. The full output from launching the >>>>>> cluster is >>>>>> below. Any ideas what the issue is? >>>>>> >>>>>> >>>>>>>>>>>>>>>>>>>>>> >>>>>> launch >>>>>> >>>>>> jg_spark2/Users/jg/dev/spark-1.6.1-bin-hadoop2.6/ec2/lib/boto-2.34.0/boto/plugin.py:40: >>>>>> PendingDeprecationWarning: the imp module is deprecated in favour of >>>>>> importlib; see the module's documentation for alternative uses >>>>>> import imp >>>>>> >>>>>> /Users/jg/dev/spark-1.6.1-bin-hadoop2.6/ec2/lib/boto-2.34.0/boto/provider.py:197: >>>>>> ResourceWarning: unclosed file <_io.TextIOWrapper >>>>>> name='/Users/jg/.aws/credentials' mode='r' encoding='UTF-8'> >>>>>> self.shared_credentials.load_from_path(shared_path) >>>>>> Setting up security groups... >>>>>> Creating security group jg_spark2-master >>>>>> Creating security group jg_spark2-slaves >>>>>> Searching for existing cluster jg_spark2 in region us-east-1... >>>>>> Spark AMI: ami-5bb18832 >>>>>> Launching instances... >>>>>> Launched 1 slave in us-east-1a, regid = r-e7d97944 >>>>>> Launched master in us-east-1a, regid = r-d3d87870 >>>>>> Waiting for AWS to propagate instance metadata... >>>>>> Waiting for cluster to enter 'ssh-ready' state............ >>>>>> >>>>>> Warning: SSH connection error. (This could be temporary.) >>>>>> Host: ec2-54-88-249-255.compute-1.amazonaws.com >>>>>> SSH return code: 255 >>>>>> SSH output: b'ssh: connect to host >>>>>> ec2-54-88-249-255.compute-1.amazonaws.com >>>>>> port 22: Connection refused' >>>>>> >>>>>> >>>>>> ./Users/jg/dev/spark-1.6.1-bin-hadoop2.6/ec2/lib/boto-2.34.0/boto/connection.py:190: >>>>>> ResourceWarning: unclosed <ssl.SSLSocket fd=4, >>>>>> family=AddressFamily.AF_INET, >>>>>> type=SocketKind.SOCK_STREAM, proto=6, laddr=('192.168.1.66', 55580), >>>>>> raddr=('54.239.20.1', 443)> >>>>>> self.queue.pop(0) >>>>>> >>>>>> >>>>>> Warning: SSH connection error. (This could be temporary.) >>>>>> Host: ec2-54-88-249-255.compute-1.amazonaws.com >>>>>> SSH return code: 255 >>>>>> SSH output: b'ssh: connect to host >>>>>> ec2-54-88-249-255.compute-1.amazonaws.com >>>>>> port 22: Connection refused' >>>>>> >>>>>> >>>>>> ./Users/jg/dev/spark-1.6.1-bin-hadoop2.6/ec2/lib/boto-2.34.0/boto/connection.py:190: >>>>>> ResourceWarning: unclosed <ssl.SSLSocket fd=4, >>>>>> family=AddressFamily.AF_INET, >>>>>> type=SocketKind.SOCK_STREAM, proto=6, laddr=('192.168.1.66', 55760), >>>>>> raddr=('54.239.26.182', 443)> >>>>>> self.queue.pop(0) >>>>>> >>>>>> >>>>>> Warning: SSH connection error. (This could be temporary.) >>>>>> Host: ec2-54-88-249-255.compute-1.amazonaws.com >>>>>> SSH return code: 255 >>>>>> SSH output: b'ssh: connect to host >>>>>> ec2-54-88-249-255.compute-1.amazonaws.com >>>>>> port 22: Connection refused' >>>>>> >>>>>> >>>>>> ./Users/jg/dev/spark-1.6.1-bin-hadoop2.6/ec2/lib/boto-2.34.0/boto/connection.py:190: >>>>>> ResourceWarning: unclosed <ssl.SSLSocket fd=4, >>>>>> family=AddressFamily.AF_INET, >>>>>> type=SocketKind.SOCK_STREAM, proto=6, laddr=('192.168.1.66', 55827), >>>>>> raddr=('54.239.20.1', 443)> >>>>>> self.queue.pop(0) >>>>>> >>>>>> >>>>>> Warning: SSH connection error. (This could be temporary.) >>>>>> Host: ec2-54-88-249-255.compute-1.amazonaws.com >>>>>> SSH return code: 255 >>>>>> SSH output: b'ssh: connect to host >>>>>> ec2-54-88-249-255.compute-1.amazonaws.com >>>>>> port 22: Connection refused' >>>>>> >>>>>> >>>>>> ./Users/jg/dev/spark-1.6.1-bin-hadoop2.6/ec2/lib/boto-2.34.0/boto/connection.py:190: >>>>>> ResourceWarning: unclosed <ssl.SSLSocket fd=4, >>>>>> family=AddressFamily.AF_INET, >>>>>> type=SocketKind.SOCK_STREAM, proto=6, laddr=('192.168.1.66', 55925), >>>>>> raddr=('207.171.162.181', 443)> >>>>>> self.queue.pop(0) >>>>>> >>>>>> Cluster is now in 'ssh-ready' state. Waited 612 seconds. >>>>>> Generating cluster's SSH key on master... >>>>>> Warning: Permanently added >>>>>> 'ec2-54-88-249-255.compute-1.amazonaws.com,54.88.249.255' (ECDSA) to >>>>>> the >>>>>> list of known hosts. >>>>>> Connection to ec2-54-88-249-255.compute-1.amazonaws.com closed. >>>>>> Warning: Permanently added >>>>>> 'ec2-54-88-249-255.compute-1.amazonaws.com,54.88.249.255' (ECDSA) to >>>>>> the >>>>>> list of known hosts. >>>>>> Transferring cluster's SSH key to slaves... >>>>>> ec2-54-209-124-74.compute-1.amazonaws.com >>>>>> Warning: Permanently added >>>>>> 'ec2-54-209-124-74.compute-1.amazonaws.com,54.209.124.74' (ECDSA) to >>>>>> the >>>>>> list of known hosts. >>>>>> Cloning spark-ec2 scripts from >>>>>> https://github.com/amplab/spark-ec2/tree/branch-1.5 on master... >>>>>> Warning: Permanently added >>>>>> 'ec2-54-88-249-255.compute-1.amazonaws.com,54.88.249.255' (ECDSA) to >>>>>> the >>>>>> list of known hosts. >>>>>> Cloning into 'spark-ec2'... >>>>>> remote: Counting objects: 2072, done. >>>>>> remote: Total 2072 (delta 0), reused 0 (delta 0), pack-reused 2072 >>>>>> Receiving objects: 100% (2072/2072), 355.67 KiB | 0 bytes/s, done. >>>>>> Resolving deltas: 100% (792/792), done. >>>>>> Checking connectivity... done. >>>>>> Connection to ec2-54-88-249-255.compute-1.amazonaws.com closed. >>>>>> Deploying files to master... >>>>>> Warning: Permanently added >>>>>> 'ec2-54-88-249-255.compute-1.amazonaws.com,54.88.249.255' (ECDSA) to >>>>>> the >>>>>> list of known hosts. >>>>>> building file list ... done >>>>>> root/spark-ec2/ec2-variables.sh >>>>>> >>>>>> sent 1658 bytes received 42 bytes 1133.33 bytes/sec >>>>>> total size is 1517 speedup is 0.89 >>>>>> Running setup on master... >>>>>> Warning: Permanently added >>>>>> 'ec2-54-88-249-255.compute-1.amazonaws.com,54.88.249.255' (ECDSA) to >>>>>> the >>>>>> list of known hosts. >>>>>> Connection to ec2-54-88-249-255.compute-1.amazonaws.com closed. >>>>>> Warning: Permanently added >>>>>> 'ec2-54-88-249-255.compute-1.amazonaws.com,54.88.249.255' (ECDSA) to >>>>>> the >>>>>> list of known hosts. >>>>>> Setting up Spark on ip-172-31-55-237.ec2.internal... >>>>>> Setting executable permissions on scripts... >>>>>> RSYNC'ing /root/spark-ec2 to other cluster nodes... >>>>>> ec2-54-209-124-74.compute-1.amazonaws.com >>>>>> Warning: Permanently added >>>>>> 'ec2-54-209-124-74.compute-1.amazonaws.com,172.31.59.121' (ECDSA) to >>>>>> the >>>>>> list of known hosts. >>>>>> id_rsa >>>>>> 100% 1679 1.6KB/s 00:00 >>>>>> [timing] rsync /root/spark-ec2: 00h 00m 00s >>>>>> Running setup-slave on all cluster nodes to mount filesystems, etc... >>>>>> [1] 20:31:41 [SUCCESS] ec2-54-88-249-255.compute-1.amazonaws.com >>>>>> checking/fixing resolution of hostname >>>>>> Setting up slave on ip-172-31-55-237.ec2.internal... of type m1.large >>>>>> 1024+0 records in >>>>>> 1024+0 records out >>>>>> 1073741824 bytes (1.1 GB) copied, 1.86735 s, 575 MB/s >>>>>> mkswap: /mnt/swap: warning: don't erase bootbits sectors >>>>>> on whole disk. Use -f to force. >>>>>> Setting up swapspace version 1, size = 1048572 KiB >>>>>> no label, UUID=87b6ce4e-5dad-4ecb-8098-7bed955b1392 >>>>>> Added 1024 MB swap file /mnt/swap >>>>>> Stderr: Warning: Permanently added >>>>>> 'ec2-54-88-249-255.compute-1.amazonaws.com,172.31.55.237' (ECDSA) to >>>>>> the >>>>>> list of known hosts. >>>>>> Connection to ec2-54-88-249-255.compute-1.amazonaws.com closed. >>>>>> [2] 20:31:42 [SUCCESS] ec2-54-209-124-74.compute-1.amazonaws.com >>>>>> checking/fixing resolution of hostname >>>>>> Setting up slave on ip-172-31-59-121.ec2.internal... of type m1.large >>>>>> 1024+0 records in >>>>>> 1024+0 records out >>>>>> 1073741824 bytes (1.1 GB) copied, 1.91229 s, 561 MB/s >>>>>> mkswap: /mnt/swap: warning: don't erase bootbits sectors >>>>>> on whole disk. Use -f to force. >>>>>> Setting up swapspace version 1, size = 1048572 KiB >>>>>> no label, UUID=33ce4520-38e3-4918-8c8f-be98034e904d >>>>>> Added 1024 MB swap file /mnt/swap >>>>>> Stderr: Connection to ec2-54-209-124-74.compute-1.amazonaws.com >>>>>> closed. >>>>>> [timing] setup-slave: 00h 00m 16s >>>>>> Initializing scala >>>>>> Unpacking Scala >>>>>> --2016-04-16 20:31:42-- >>>>>> http://s3.amazonaws.com/spark-related-packages/scala-2.10.3.tgz >>>>>> Resolving s3.amazonaws.com (s3.amazonaws.com)... 54.231.114.178 >>>>>> Connecting to s3.amazonaws.com (s3.amazonaws.com >>>>>> )|54.231.114.178|:80... >>>>>> connected. >>>>>> HTTP request sent, awaiting response... 200 OK >>>>>> Length: 30531249 (29M) [application/x-compressed] >>>>>> Saving to: ‘scala-2.10.3.tgz’ >>>>>> >>>>>> >>>>>> 100%[===============================================================================================>] >>>>>> 30,531,249 2.32MB/s in 15s >>>>>> >>>>>> 2016-04-16 20:31:57 (1.93 MB/s) - ‘scala-2.10.3.tgz’ saved >>>>>> [30531249/30531249] >>>>>> >>>>>> [timing] scala init: 00h 00m 16s >>>>>> Initializing spark >>>>>> --2016-04-16 20:31:58-- >>>>>> >>>>>> http://s3.amazonaws.com/spark-related-packages/spark-1.6.1-bin-hadoop1.tgz >>>>>> Resolving s3.amazonaws.com (s3.amazonaws.com)... 54.231.32.114 >>>>>> Connecting to s3.amazonaws.com (s3.amazonaws.com >>>>>> )|54.231.32.114|:80... >>>>>> connected. >>>>>> HTTP request sent, awaiting response... 200 OK >>>>>> Length: 277258240 (264M) [application/x-compressed] >>>>>> Saving to: ‘spark-1.6.1-bin-hadoop1.tgz’ >>>>>> >>>>>> >>>>>> 100%[===============================================================================================>] >>>>>> 277,258,240 44.5MB/s in 6.3s >>>>>> >>>>>> 2016-04-16 20:32:04 (41.8 MB/s) - ‘spark-1.6.1-bin-hadoop1.tgz’ saved >>>>>> [277258240/277258240] >>>>>> >>>>>> Unpacking Spark >>>>>> >>>>>> gzip: stdin: not in gzip format >>>>>> tar: Child returned status 1 >>>>>> tar: Error is not recoverable: exiting now >>>>>> mv: missing destination file operand after `spark' >>>>>> Try `mv --help' for more information. >>>>>> [timing] spark init: 00h 00m 06s >>>>>> Initializing ephemeral-hdfs >>>>>> --2016-04-16 20:32:04-- >>>>>> http://s3.amazonaws.com/spark-related-packages/hadoop-1.0.4.tar.gz >>>>>> Resolving s3.amazonaws.com (s3.amazonaws.com)... 54.231.13.48 >>>>>> Connecting to s3.amazonaws.com (s3.amazonaws.com)|54.231.13.48|:80... >>>>>> connected. >>>>>> HTTP request sent, awaiting response... 200 OK >>>>>> Length: 62793050 (60M) [application/x-gzip] >>>>>> Saving to: ‘hadoop-1.0.4.tar.gz’ >>>>>> >>>>>> >>>>>> 100%[===============================================================================================>] >>>>>> 62,793,050 45.3MB/s in 1.3s >>>>>> >>>>>> 2016-04-16 20:32:06 (45.3 MB/s) - ‘hadoop-1.0.4.tar.gz’ saved >>>>>> [62793050/62793050] >>>>>> >>>>>> Unpacking Hadoop >>>>>> RSYNC'ing /root/ephemeral-hdfs to slaves... >>>>>> ec2-54-209-124-74.compute-1.amazonaws.com >>>>>> Warning: Permanently added >>>>>> 'ec2-54-209-124-74.compute-1.amazonaws.com,172.31.59.121' (ECDSA) to >>>>>> the >>>>>> list of known hosts. >>>>>> [timing] ephemeral-hdfs init: 00h 00m 15s >>>>>> Initializing persistent-hdfs >>>>>> --2016-04-16 20:32:19-- >>>>>> http://s3.amazonaws.com/spark-related-packages/hadoop-1.0.4.tar.gz >>>>>> Resolving s3.amazonaws.com (s3.amazonaws.com)... 54.231.115.26 >>>>>> Connecting to s3.amazonaws.com (s3.amazonaws.com >>>>>> )|54.231.115.26|:80... >>>>>> connected. >>>>>> HTTP request sent, awaiting response... 200 OK >>>>>> Length: 62793050 (60M) [application/x-gzip] >>>>>> Saving to: ‘hadoop-1.0.4.tar.gz’ >>>>>> >>>>>> >>>>>> 100%[===============================================================================================>] >>>>>> 62,793,050 58.1MB/s in 1.0s >>>>>> >>>>>> 2016-04-16 20:32:20 (58.1 MB/s) - ‘hadoop-1.0.4.tar.gz’ saved >>>>>> [62793050/62793050] >>>>>> >>>>>> Unpacking Hadoop >>>>>> RSYNC'ing /root/persistent-hdfs to slaves... >>>>>> ec2-54-209-124-74.compute-1.amazonaws.com >>>>>> [timing] persistent-hdfs init: 00h 00m 14s >>>>>> Initializing spark-standalone >>>>>> [timing] spark-standalone init: 00h 00m 00s >>>>>> Initializing tachyon >>>>>> --2016-04-16 20:32:33-- >>>>>> https://s3.amazonaws.com/Tachyon/tachyon-0.8.2-bin.tar.gz >>>>>> Resolving s3.amazonaws.com (s3.amazonaws.com)... 54.231.80.67 >>>>>> Connecting to s3.amazonaws.com (s3.amazonaws.com >>>>>> )|54.231.80.67|:443... >>>>>> connected. >>>>>> HTTP request sent, awaiting response... 200 OK >>>>>> Length: 72834292 (69M) [application/x-tar] >>>>>> Saving to: ‘tachyon-0.8.2-bin.tar.gz’ >>>>>> >>>>>> >>>>>> 100%[===============================================================================================>] >>>>>> 72,834,292 35.9MB/s in 1.9s >>>>>> >>>>>> 2016-04-16 20:32:35 (35.9 MB/s) - ‘tachyon-0.8.2-bin.tar.gz’ saved >>>>>> [72834292/72834292] >>>>>> >>>>>> Unpacking Tachyon >>>>>> [timing] tachyon init: 00h 00m 04s >>>>>> Initializing rstudio >>>>>> --2016-04-16 20:32:37-- >>>>>> http://download2.rstudio.org/rstudio-server-rhel-0.99.446-x86_64.rpm >>>>>> Resolving download2.rstudio.org (download2.rstudio.org)... >>>>>> 52.85.140.113, >>>>>> 52.85.140.193, 52.85.140.208, ... >>>>>> Connecting to download2.rstudio.org >>>>>> (download2.rstudio.org)|52.85.140.113|:80... connected. >>>>>> HTTP request sent, awaiting response... 200 OK >>>>>> Length: 35035164 (33M) [application/x-redhat-package-manager] >>>>>> Saving to: ‘rstudio-server-rhel-0.99.446-x86_64.rpm’ >>>>>> >>>>>> >>>>>> 100%[===============================================================================================>] >>>>>> 35,035,164 66.8MB/s in 0.5s >>>>>> >>>>>> 2016-04-16 20:32:38 (66.8 MB/s) - >>>>>> ‘rstudio-server-rhel-0.99.446-x86_64.rpm’ >>>>>> saved [35035164/35035164] >>>>>> >>>>>> Loaded plugins: priorities, update-motd, upgrade-helper >>>>>> Examining rstudio-server-rhel-0.99.446-x86_64.rpm: >>>>>> rstudio-server-0.99.446-1.x86_64 >>>>>> Marking rstudio-server-rhel-0.99.446-x86_64.rpm to be installed >>>>>> Resolving Dependencies >>>>>> --> Running transaction check >>>>>> ---> Package rstudio-server.x86_64 0:0.99.446-1 will be installed >>>>>> --> Finished Dependency Resolution >>>>>> >>>>>> Dependencies Resolved >>>>>> >>>>>> >>>>>> ========================================================================================================================================= >>>>>> Package Arch Version >>>>>> Repository Size >>>>>> >>>>>> ========================================================================================================================================= >>>>>> Installing: >>>>>> rstudio-server x86_64 0.99.446-1 >>>>>> /rstudio-server-rhel-0.99.446-x86_64 252 M >>>>>> >>>>>> Transaction Summary >>>>>> >>>>>> ========================================================================================================================================= >>>>>> Install 1 Package >>>>>> >>>>>> Total size: 252 M >>>>>> Installed size: 252 M >>>>>> Downloading packages: >>>>>> Running transaction check >>>>>> Running transaction test >>>>>> Transaction test succeeded >>>>>> Running transaction >>>>>> Installing : rstudio-server-0.99.446-1.x86_64 >>>>>> 1/1 >>>>>> groupadd: group 'rstudio-server' already exists >>>>>> rsession: no process killed >>>>>> rstudio-server start/running, process 4817 >>>>>> Verifying : rstudio-server-0.99.446-1.x86_64 >>>>>> 1/1 >>>>>> >>>>>> Installed: >>>>>> rstudio-server.x86_64 0:0.99.446-1 >>>>>> >>>>>> Complete! >>>>>> rstudio-server start/running, process 4854 >>>>>> [timing] rstudio init: 00h 01m 25s >>>>>> Initializing ganglia >>>>>> Connection to ec2-54-209-124-74.compute-1.amazonaws.com closed. >>>>>> [timing] ganglia init: 00h 00m 01s >>>>>> Creating local config files... >>>>>> Connection to ec2-54-209-124-74.compute-1.amazonaws.com closed. >>>>>> Connection to ec2-54-209-124-74.compute-1.amazonaws.com closed. >>>>>> Configuring /etc/ganglia/gmond.conf >>>>>> Configuring /etc/ganglia/gmetad.conf >>>>>> Configuring /etc/httpd/conf.d/ganglia.conf >>>>>> Configuring /etc/httpd/conf/httpd.conf >>>>>> Configuring /root/mapreduce/hadoop.version >>>>>> Configuring /root/mapreduce/conf/core-site.xml >>>>>> Configuring /root/mapreduce/conf/slaves >>>>>> Configuring /root/mapreduce/conf/mapred-site.xml >>>>>> Configuring /root/mapreduce/conf/hdfs-site.xml >>>>>> Configuring /root/mapreduce/conf/hadoop-env.sh >>>>>> Configuring /root/mapreduce/conf/masters >>>>>> Configuring /root/persistent-hdfs/conf/core-site.xml >>>>>> Configuring /root/persistent-hdfs/conf/slaves >>>>>> Configuring /root/persistent-hdfs/conf/mapred-site.xml >>>>>> Configuring /root/persistent-hdfs/conf/hdfs-site.xml >>>>>> Configuring /root/persistent-hdfs/conf/hadoop-env.sh >>>>>> Configuring /root/persistent-hdfs/conf/masters >>>>>> Configuring /root/ephemeral-hdfs/conf/core-site.xml >>>>>> Configuring /root/ephemeral-hdfs/conf/yarn-site.xml >>>>>> Configuring /root/ephemeral-hdfs/conf/slaves >>>>>> Configuring /root/ephemeral-hdfs/conf/mapred-site.xml >>>>>> Configuring /root/ephemeral-hdfs/conf/hadoop-metrics2.properties >>>>>> Configuring /root/ephemeral-hdfs/conf/capacity-scheduler.xml >>>>>> Configuring /root/ephemeral-hdfs/conf/yarn-env.sh >>>>>> Configuring /root/ephemeral-hdfs/conf/hdfs-site.xml >>>>>> Configuring /root/ephemeral-hdfs/conf/hadoop-env.sh >>>>>> Configuring /root/ephemeral-hdfs/conf/masters >>>>>> Configuring /root/spark/conf/core-site.xml >>>>>> Configuring /root/spark/conf/spark-defaults.conf >>>>>> Configuring /root/spark/conf/spark-env.sh >>>>>> Configuring /root/tachyon/conf/slaves >>>>>> Configuring /root/tachyon/conf/workers >>>>>> Configuring /root/tachyon/conf/tachyon-env.sh >>>>>> Deploying Spark config files... >>>>>> RSYNC'ing /root/spark/conf to slaves... >>>>>> ec2-54-209-124-74.compute-1.amazonaws.com >>>>>> Setting up scala >>>>>> RSYNC'ing /root/scala to slaves... >>>>>> ec2-54-209-124-74.compute-1.amazonaws.com >>>>>> [timing] scala setup: 00h 00m 02s >>>>>> Setting up spark >>>>>> RSYNC'ing /root/spark to slaves... >>>>>> ec2-54-209-124-74.compute-1.amazonaws.com >>>>>> [timing] spark setup: 00h 00m 01s >>>>>> Setting up ephemeral-hdfs >>>>>> ec2-54-209-124-74.compute-1.amazonaws.com >>>>>> Connection to ec2-54-209-124-74.compute-1.amazonaws.com closed. >>>>>> RSYNC'ing /root/ephemeral-hdfs/conf to slaves... >>>>>> ec2-54-209-124-74.compute-1.amazonaws.com >>>>>> Formatting ephemeral HDFS namenode... >>>>>> Warning: $HADOOP_HOME is deprecated. >>>>>> >>>>>> 16/04/16 20:34:09 INFO namenode.NameNode: STARTUP_MSG: >>>>>> /************************************************************ >>>>>> STARTUP_MSG: Starting NameNode >>>>>> STARTUP_MSG: host = ip-172-31-55-237.ec2.internal/172.31.55.237 >>>>>> STARTUP_MSG: args = [-format] >>>>>> STARTUP_MSG: version = 1.0.4 >>>>>> STARTUP_MSG: build = >>>>>> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0 -r >>>>>> 1393290; compiled by 'hortonfo' on Wed Oct 3 05:13:58 UTC 2012 >>>>>> ************************************************************/ >>>>>> 16/04/16 20:34:09 INFO util.GSet: VM type = 64-bit >>>>>> 16/04/16 20:34:09 INFO util.GSet: 2% max memory = 17.78 MB >>>>>> 16/04/16 20:34:09 INFO util.GSet: capacity = 2^21 = 2097152 >>>>>> entries >>>>>> 16/04/16 20:34:09 INFO util.GSet: recommended=2097152, actual=2097152 >>>>>> 16/04/16 20:34:09 INFO namenode.FSNamesystem: fsOwner=root >>>>>> 16/04/16 20:34:09 INFO namenode.FSNamesystem: supergroup=supergroup >>>>>> 16/04/16 20:34:09 INFO namenode.FSNamesystem: >>>>>> isPermissionEnabled=false >>>>>> 16/04/16 20:34:09 INFO namenode.FSNamesystem: >>>>>> dfs.block.invalidate.limit=100 >>>>>> 16/04/16 20:34:09 INFO namenode.FSNamesystem: >>>>>> isAccessTokenEnabled=false >>>>>> accessKeyUpdateInterval=0 min(s), accessTokenLifetime=0 min(s) >>>>>> 16/04/16 20:34:09 INFO namenode.NameNode: Caching file names occuring >>>>>> more >>>>>> than 10 times >>>>>> 16/04/16 20:34:09 INFO common.Storage: Image file of size 110 saved >>>>>> in 0 >>>>>> seconds. >>>>>> 16/04/16 20:34:09 INFO common.Storage: Storage directory >>>>>> /mnt/ephemeral-hdfs/dfs/name has been successfully formatted. >>>>>> 16/04/16 20:34:09 INFO namenode.NameNode: SHUTDOWN_MSG: >>>>>> /************************************************************ >>>>>> SHUTDOWN_MSG: Shutting down NameNode at >>>>>> ip-172-31-55-237.ec2.internal/172.31.55.237 >>>>>> ************************************************************/ >>>>>> Starting ephemeral HDFS... >>>>>> Warning: $HADOOP_HOME is deprecated. >>>>>> >>>>>> starting namenode, logging to >>>>>> >>>>>> /mnt/ephemeral-hdfs/logs/hadoop-root-namenode-ip-172-31-55-237.ec2.internal.out >>>>>> ec2-54-209-124-74.compute-1.amazonaws.com: Warning: $HADOOP_HOME is >>>>>> deprecated. >>>>>> ec2-54-209-124-74.compute-1.amazonaws.com: >>>>>> ec2-54-209-124-74.compute-1.amazonaws.com: starting datanode, >>>>>> logging to >>>>>> >>>>>> /mnt/ephemeral-hdfs/logs/hadoop-root-datanode-ip-172-31-59-121.ec2.internal.out >>>>>> ec2-54-88-249-255.compute-1.amazonaws.com: Warning: Permanently added >>>>>> 'ec2-54-88-249-255.compute-1.amazonaws.com,172.31.55.237' (ECDSA) to >>>>>> the >>>>>> list of known hosts. >>>>>> ec2-54-88-249-255.compute-1.amazonaws.com: Warning: $HADOOP_HOME is >>>>>> deprecated. >>>>>> ec2-54-88-249-255.compute-1.amazonaws.com: >>>>>> ec2-54-88-249-255.compute-1.amazonaws.com: starting >>>>>> secondarynamenode, >>>>>> logging to >>>>>> >>>>>> /mnt/ephemeral-hdfs/logs/hadoop-root-secondarynamenode-ip-172-31-55-237.ec2.internal.out >>>>>> [timing] ephemeral-hdfs setup: 00h 00m 07s >>>>>> Setting up persistent-hdfs >>>>>> Pseudo-terminal will not be allocated because stdin is not a terminal. >>>>>> RSYNC'ing /root/persistent-hdfs/conf to slaves... >>>>>> ec2-54-209-124-74.compute-1.amazonaws.com >>>>>> Formatting persistent HDFS namenode... >>>>>> Warning: $HADOOP_HOME is deprecated. >>>>>> >>>>>> 16/04/16 20:34:16 INFO namenode.NameNode: STARTUP_MSG: >>>>>> /************************************************************ >>>>>> STARTUP_MSG: Starting NameNode >>>>>> STARTUP_MSG: host = ip-172-31-55-237.ec2.internal/172.31.55.237 >>>>>> STARTUP_MSG: args = [-format] >>>>>> STARTUP_MSG: version = 1.0.4 >>>>>> STARTUP_MSG: build = >>>>>> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0 -r >>>>>> 1393290; compiled by 'hortonfo' on Wed Oct 3 05:13:58 UTC 2012 >>>>>> ************************************************************/ >>>>>> 16/04/16 20:34:16 INFO util.GSet: VM type = 64-bit >>>>>> 16/04/16 20:34:16 INFO util.GSet: 2% max memory = 17.78 MB >>>>>> 16/04/16 20:34:16 INFO util.GSet: capacity = 2^21 = 2097152 >>>>>> entries >>>>>> 16/04/16 20:34:16 INFO util.GSet: recommended=2097152, actual=2097152 >>>>>> 16/04/16 20:34:16 INFO namenode.FSNamesystem: fsOwner=root >>>>>> 16/04/16 20:34:16 INFO namenode.FSNamesystem: supergroup=supergroup >>>>>> 16/04/16 20:34:16 INFO namenode.FSNamesystem: >>>>>> isPermissionEnabled=false >>>>>> 16/04/16 20:34:17 INFO namenode.FSNamesystem: >>>>>> dfs.block.invalidate.limit=100 >>>>>> 16/04/16 20:34:17 INFO namenode.FSNamesystem: >>>>>> isAccessTokenEnabled=false >>>>>> accessKeyUpdateInterval=0 min(s), accessTokenLifetime=0 min(s) >>>>>> 16/04/16 20:34:17 INFO namenode.NameNode: Caching file names occuring >>>>>> more >>>>>> than 10 times >>>>>> 16/04/16 20:34:17 INFO common.Storage: Image file of size 110 saved >>>>>> in 0 >>>>>> seconds. >>>>>> 16/04/16 20:34:17 INFO common.Storage: Storage directory >>>>>> /vol/persistent-hdfs/dfs/name has been successfully formatted. >>>>>> 16/04/16 20:34:17 INFO namenode.NameNode: SHUTDOWN_MSG: >>>>>> /************************************************************ >>>>>> SHUTDOWN_MSG: Shutting down NameNode at >>>>>> ip-172-31-55-237.ec2.internal/172.31.55.237 >>>>>> ************************************************************/ >>>>>> Persistent HDFS installed, won't start by default... >>>>>> [timing] persistent-hdfs setup: 00h 00m 03s >>>>>> Setting up spark-standalone >>>>>> RSYNC'ing /root/spark/conf to slaves... >>>>>> ec2-54-209-124-74.compute-1.amazonaws.com >>>>>> RSYNC'ing /root/spark-ec2 to slaves... >>>>>> ec2-54-209-124-74.compute-1.amazonaws.com >>>>>> ./spark-standalone/setup.sh: line 22: /root/spark/sbin/stop-all.sh: >>>>>> No such >>>>>> file or directory >>>>>> ./spark-standalone/setup.sh: line 27: >>>>>> /root/spark/sbin/start-master.sh: No >>>>>> such file or directory >>>>>> ./spark-standalone/setup.sh: line 33: >>>>>> /root/spark/sbin/start-slaves.sh: No >>>>>> such file or directory >>>>>> [timing] spark-standalone setup: 00h 00m 25s >>>>>> Setting up tachyon >>>>>> RSYNC'ing /root/tachyon to slaves... >>>>>> ec2-54-209-124-74.compute-1.amazonaws.com >>>>>> Connecting to ec2-54-209-124-74.compute-1.amazonaws.com as root... >>>>>> Pseudo-terminal will not be allocated because stdin is not a terminal. >>>>>> Formatting Tachyon Worker @ ip-172-31-59-121.ec2.internal >>>>>> Formatting Tachyon Master @ ec2-54-88-249-255.compute-1.amazonaws.com >>>>>> Killed 0 processes on ip-172-31-55-237.ec2.internal >>>>>> Killed 0 processes on ip-172-31-55-237.ec2.internal >>>>>> Connecting to ec2-54-209-124-74.compute-1.amazonaws.com as root... >>>>>> Pseudo-terminal will not be allocated because stdin is not a terminal. >>>>>> Killed 0 processes on ip-172-31-59-121.ec2.internal >>>>>> Starting master @ ec2-54-88-249-255.compute-1.amazonaws.com >>>>>> Connecting to ec2-54-209-124-74.compute-1.amazonaws.com as root... >>>>>> Pseudo-terminal will not be allocated because stdin is not a terminal. >>>>>> Formatting RamFS: /mnt/ramdisk (6154mb) >>>>>> Starting worker @ ip-172-31-59-121.ec2.internal >>>>>> [timing] tachyon setup: 00h 00m 17s >>>>>> Setting up rstudio >>>>>> spark-ec2/setup.sh: line 110: ./rstudio/setup.sh: No such file or >>>>>> directory >>>>>> [timing] rstudio setup: 00h 00m 00s >>>>>> Setting up ganglia >>>>>> RSYNC'ing /etc/ganglia to slaves... >>>>>> ec2-54-209-124-74.compute-1.amazonaws.com >>>>>> Shutting down GANGLIA gmond: [FAILED] >>>>>> Starting GANGLIA gmond: [ OK ] >>>>>> Shutting down GANGLIA gmond: [FAILED] >>>>>> Starting GANGLIA gmond: [ OK ] >>>>>> Connection to ec2-54-209-124-74.compute-1.amazonaws.com closed. >>>>>> Shutting down GANGLIA gmetad: [FAILED] >>>>>> Starting GANGLIA gmetad: [ OK ] >>>>>> Stopping httpd: [FAILED] >>>>>> Starting httpd: httpd: Syntax error on line 154 of >>>>>> /etc/httpd/conf/httpd.conf: Cannot load >>>>>> /etc/httpd/modules/mod_authz_core.so >>>>>> into server: /etc/httpd/modules/mod_authz_core.so: cannot open shared >>>>>> object >>>>>> file: No such file or directory >>>>>> [FAILED] >>>>>> [timing] ganglia setup: 00h 00m 02s >>>>>> Connection to ec2-54-88-249-255.compute-1.amazonaws.com closed. >>>>>> Spark standalone cluster started at >>>>>> http://ec2-54-88-249-255.compute-1.amazonaws.com:8080 >>>>>> Ganglia started at >>>>>> http://ec2-54-88-249-255.compute-1.amazonaws.com:5080/ganglia >>>>>> Done! >>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> View this message in context: >>>>>> http://apache-spark-user-list.1001560.n3.nabble.com/A-number-of-issues-when-running-spark-ec2-tp26791.html >>>>>> Sent from the Apache Spark User List mailing list archive at >>>>>> Nabble.com. >>>>>> >>>>>> --------------------------------------------------------------------- >>>>>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org >>>>>> For additional commands, e-mail: user-h...@spark.apache.org >>>>>> >>>>>> >>>>> >>>> >>> >>