use sudo with chown and then chmod to 755

On Sun, Apr 29, 2012 at 11:02 PM, maddy gulati <achilles....@gmail.com>wrote:

> I have confirgured the xml configuration files corectly. I tried to repeat
> the installation process with a different user id . Now getting following
> when I execute start-all.sh :
>
>
>
> Warning: $HADOOP_HOME is deprecated.
>
> chown: changing ownership of `/usr/libexec/../logs': Operation not
> permitted
> starting namenode, logging to
> /usr/libexec/../logs/hadoop-maddy-namenode-mandeep-System-Product-Name.out
> /usr/sbin/hadoop-daemon.sh: line 135:
> /usr/libexec/../logs/hadoop-maddy-namenode-mandeep-System-Product-Name.out:
> Permission denied
> head: cannot open
>
> `/usr/libexec/../logs/hadoop-maddy-namenode-mandeep-System-Product-Name.out'
> for reading: No such file or directory
> localhost: chown: changing ownership of `/usr/libexec/../logs': Operation
> not permitted
> localhost: starting datanode, logging to
> /usr/libexec/../logs/hadoop-maddy-datanode-mandeep-System-Product-Name.out
> localhost: /usr/sbin/hadoop-daemon.sh: line 135:
> /usr/libexec/../logs/hadoop-maddy-datanode-mandeep-System-Product-Name.out:
> Permission denied
> localhost: head: cannot open
>
> `/usr/libexec/../logs/hadoop-maddy-datanode-mandeep-System-Product-Name.out'
> for reading: No such file or directory
> localhost: chown: changing ownership of `/usr/libexec/../logs': Operation
> not permitted
> localhost: starting secondarynamenode, logging to
>
> /usr/libexec/../logs/hadoop-maddy-secondarynamenode-mandeep-System-Product-Name.out
> localhost: /usr/sbin/hadoop-daemon.sh: line 135:
>
> /usr/libexec/../logs/hadoop-maddy-secondarynamenode-mandeep-System-Product-Name.out:
> Permission denied
> localhost: head: cannot open
>
> `/usr/libexec/../logs/hadoop-maddy-secondarynamenode-mandeep-System-Product-Name.out'
> for reading: No such file or directory
> chown: changing ownership of `/usr/libexec/../logs': Operation not
> permitted
> starting jobtracker, logging to
>
> /usr/libexec/../logs/hadoop-maddy-jobtracker-mandeep-System-Product-Name.out
> /usr/sbin/hadoop-daemon.sh: line 135:
>
> /usr/libexec/../logs/hadoop-maddy-jobtracker-mandeep-System-Product-Name.out:
> Permission denied
> head: cannot open
>
> `/usr/libexec/../logs/hadoop-maddy-jobtracker-mandeep-System-Product-Name.out'
> for reading: No such file or directory
> localhost: chown: changing ownership of `/usr/libexec/../logs': Operation
> not permitted
> localhost: starting tasktracker, logging to
>
> /usr/libexec/../logs/hadoop-maddy-tasktracker-mandeep-System-Product-Name.out
> localhost: /usr/sbin/hadoop-daemon.sh: line 135:
>
> /usr/libexec/../logs/hadoop-maddy-tasktracker-mandeep-System-Product-Name.out:
> Permission denied
> localhost: head: cannot open
>
> `/usr/libexec/../logs/hadoop-maddy-tasktracker-mandeep-System-Product-Name.out'
> for reading: No such file or directory
>
>
>
>
> Below are the contents of my configuration files
>
>
> Core Site:
> <?xml version="1.0"?>
> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
>
> <!-- Put site-specific property overrides in this file. -->
>
> <configuration>
>
> <property>
>  <name>hadoop.tmp.dir</name>
>  <value>/app/hadoop/tmp</value>
>  <description>A base for other temporary directories.</description>
> </property>
>
> <property>
>  <name>fs.default.name</name>
>  <value>hdfs://localhost:54310</value>
>  <description>The name of the default file system.  A URI whose
>  scheme and authority determine the FileSystem implementation.  The
>  uri's scheme determines the config property (fs.SCHEME.impl) naming
>  the FileSystem implementation class.  The uri's authority is used to
>  determine the host, port, etc. for a filesystem.</description>
> </property>
>
>
> </configuration>
>
>
>
>
>
>
> mapred-site.xml
>
> <?xml version="1.0"?>
> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
>
> <!-- Put site-specific property overrides in this file. -->
>
> <configuration>
>
> <property>
>  <name>mapred.job.tracker</name>
>  <value>localhost:54311</value>
>  <description>The host and port that the MapReduce job tracker runs
>  at.  If "local", then jobs are run in-process as a single map
>  and reduce task.
>  </description>
> </property>
>
> </configuration>
>
>
>
>
>
>
>
> hdfs-site.xml
>
> <?xml version="1.0"?>
> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
>
> <!-- Put site-specific property overrides in this file. -->
>
> <configuration>
>
> <property>
>  <name>dfs.replication</name>
>  <value>1</value>
>  <description>Default block replication.
>  The actual number of replications can be specified when the file is
> created.
>  The default is used if replication is not specified in create time.
>  </description>
> </property>
>
>
>
> </configuration>
>
>
>
>
>
>
>
>
>
>
>
>
> On Sun, Apr 29, 2012 at 4:05 AM, Marcos Ortiz <mlor...@uci.cu> wrote:
>
> > Look here
> > http://search-hadoop.com/m/**NRMV72pWYVM1/ERROR+org.apache.**
> > hadoop.hdfs.server.datanode.**DataNode%5C%3A+java.lang.**
> > IllegalArgumentException%5C%**3A+Does+not+contain+a+valid+**
> > host%5C%3Aport+authority%5C%**3A+/v=threaded<
> http://search-hadoop.com/m/NRMV72pWYVM1/ERROR+org.apache.hadoop.hdfs.server.datanode.DataNode%5C%3A+java.lang.IllegalArgumentException%5C%3A+Does+not+contain+a+valid+host%5C%3Aport+authority%5C%3A+/v=threaded
> >
> >
> > To solve this, you can check your config files: core-site.xml,
> > mapred-site.xml and hdfs-site.xml
> >
> > http://www.michael-noll.com/**tutorials/running-hadoop-on-**
> > ubuntu-linux-single-node-**cluster/#conf-site-xml<
> http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-single-node-cluster/#conf-site-xml
> >
> >
> > Of course, yo can check is the official documentation for 1.0.1:
> >
> > http://hadoop.apache.org/**common/docs/r1.0.1/single_**node_setup.html<
> http://hadoop.apache.org/common/docs/r1.0.1/single_node_setup.html>
> >
> > Regards
> >
> >
> > On 4/28/2012 3:19 PM, maddy gulati wrote:
> >
> >> Folks,
> >>
> >> I am trying to set up a single node cluster for Hadoop ( version 1.0.1)
> on
> >> Ubuntu 11.04. After configuring everything, I could not get all the
> >> components to start.
> >>
> >> My installation directory for hadoop is /usr/local/hadoop
> >> I followed the following tutorial for the setup :
> >> http://www.michael-noll.com/**tutorials/running-hadoop-on-**
> >> ubuntu-linux-single-node-**cluster/<
> http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-single-node-cluster/
> >
> >> Below is the snapshot of my log file:
> >>
> >> mandeep@mandeep-System-**Product-Name:/var/log/hadoop/**mandeep$ cat
> >> hadoop-mandeep-datanode-**mandeep-System-Product-Name.**log
> >> 2012-04-29 01:29:37,240 INFO
> >> org.apache.hadoop.hdfs.server.**datanode.DataNode: STARTUP_MSG:
> >> /****************************************************************
> >> STARTUP_MSG: Starting DataNode
> >> STARTUP_MSG:   host = mandeep-System-Product-Name/12**7.0.1.1<
> http://127.0.1.1>
> >> STARTUP_MSG:   args = []
> >> STARTUP_MSG:   version = 1.0.1
> >> STARTUP_MSG:   build =
> >> https://svn.apache.org/repos/**asf/hadoop/common/branches/**branch-1.0<
> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0>-r
> >> 1243785; compiled by 'hortonfo' on Tue Feb 14 08:15:38 UTC 2012
> >> ****************************************************************/
> >> 2012-04-29 01:29:37,398 INFO org.apache.hadoop.metrics2.**
> >> impl.MetricsConfig:
> >> loaded properties from hadoop-metrics2.properties
> >> 2012-04-29 01:29:37,446 INFO
> >> org.apache.hadoop.metrics2.**impl.MetricsSourceAdapter: MBean for source
> >> MetricsSystem,sub=Stats registered.
> >> 2012-04-29 01:29:37,449 INFO
> >> org.apache.hadoop.metrics2.**impl.MetricsSystemImpl: Scheduled snapshot
> >> period at 10 second(s).
> >> 2012-04-29 01:29:37,449 INFO
> >> org.apache.hadoop.metrics2.**impl.MetricsSystemImpl: DataNode metrics
> >> system
> >> started
> >> 2012-04-29 01:29:37,583 INFO
> >> org.apache.hadoop.metrics2.**impl.MetricsSourceAdapter: MBean for source
> >> ugi
> >> registered.
> >> 2012-04-29 01:29:37,586 WARN
> >> org.apache.hadoop.metrics2.**impl.MetricsSystemImpl: Source name ugi
> >> already
> >> exists!
> >> 2012-04-29 01:29:37,615 INFO org.apache.hadoop.util.**NativeCodeLoader:
> >> Loaded the native-hadoop library
> >> 2012-04-29 01:29:37,701 ERROR
> >> org.apache.hadoop.hdfs.server.**datanode.DataNode:
> >> java.lang.**IllegalArgumentException: Does not contain a valid host:port
> >> authority: file:///
> >>  at org.apache.hadoop.net.**NetUtils.createSocketAddr(**
> >> NetUtils.java:162)
> >> at
> >> org.apache.hadoop.hdfs.server.**namenode.NameNode.getAddress(**
> >> NameNode.java:198)
> >>  at
> >> org.apache.hadoop.hdfs.server.**namenode.NameNode.getAddress(**
> >> NameNode.java:228)
> >> at
> >> org.apache.hadoop.hdfs.server.**namenode.NameNode.**
> >> getServiceAddress(NameNode.**java:222)
> >>  at
> >> org.apache.hadoop.hdfs.server.**datanode.DataNode.**
> >> startDataNode(DataNode.java:**337)
> >> at org.apache.hadoop.hdfs.server.**datanode.DataNode.<init>(**
> >> DataNode.java:299)
> >>  at
> >> org.apache.hadoop.hdfs.server.**datanode.DataNode.**
> >> makeInstance(DataNode.java:**1582)
> >> at
> >> org.apache.hadoop.hdfs.server.**datanode.DataNode.**
> >> instantiateDataNode(DataNode.**java:1521)
> >>  at
> >> org.apache.hadoop.hdfs.server.**datanode.DataNode.**
> >> createDataNode(DataNode.java:**1539)
> >> at
> >> org.apache.hadoop.hdfs.server.**datanode.DataNode.secureMain(**
> >> DataNode.java:1665)
> >>  at org.apache.hadoop.hdfs.server.**datanode.DataNode.main(**
> >> DataNode.java:1682)
> >>
> >> 2012-04-29 01:29:37,718 INFO
> >> org.apache.hadoop.hdfs.server.**datanode.DataNode: SHUTDOWN_MSG:
> >> /****************************************************************
> >> SHUTDOWN_MSG: Shutting down DataNode at mandeep-System-Product-Name/
> >> 127.0.1.1
> >> ****************************************************************/
> >> 2012-04-29 01:33:26,907 INFO
> >> org.apache.hadoop.hdfs.server.**datanode.DataNode: STARTUP_MSG:
> >> /****************************************************************
> >>
> >>
> >>
> > --
> > Marcos Luis Ortíz Valmaseda (@marcosluis2186)
> >  Data Engineer at UCI
> >  http://marcosluis2186.**posterous.com<
> http://marcosluis2186.posterous.com>
> >
> > 10mo. ANIVERSARIO DE LA CREACION DE LA UNIVERSIDAD DE LAS CIENCIAS
> > INFORMATICAS...
> > CONECTADOS AL FUTURO, CONECTADOS A LA REVOLUCION
> >
> > http://www.uci.cu
> > http://www.facebook.com/**universidad.uci<
> http://www.facebook.com/universidad.uci>
> > http://www.flickr.com/photos/**universidad_uci<
> http://www.flickr.com/photos/universidad_uci>
> >
>
>
>
> --
> Mandeep Singh Gulati
> System Analyst | WorldWide Risk and Information Management
> American Express India Pvt. Ltd.
> Gurgaon, India
>



-- 


∞
Shashwat Shriparv

Reply via email to