thank's Mohammad with this command:
babak@ubuntu:~/Downloads/hadoop/bin$ hadoop namenode -format this is my output: 12/06/06 20:05:20 INFO namenode.NameNode: STARTUP_MSG: /************************************************************ STARTUP_MSG: Starting NameNode STARTUP_MSG: host = ubuntu/127.0.1.1 STARTUP_MSG: args = [-format] STARTUP_MSG: version = 0.20.2 STARTUP_MSG: build = https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20 -r 911707; compiled by 'chrisdo' on Fri Feb 19 08:07:34 UTC 2010 ************************************************************/ 12/06/06 20:05:20 INFO namenode.FSNamesystem: fsOwner=babak,babak,adm,dialout,cdrom,plugdev,lpadmin,admin,sambashare 12/06/06 20:05:20 INFO namenode.FSNamesystem: supergroup=supergroup 12/06/06 20:05:20 INFO namenode.FSNamesystem: isPermissionEnabled=true 12/06/06 20:05:20 INFO common.Storage: Image file of size 95 saved in 0 seconds. 12/06/06 20:05:20 INFO common.Storage: Storage directory /tmp/hadoop-babak/dfs/name has been successfully formatted. 12/06/06 20:05:20 INFO namenode.NameNode: SHUTDOWN_MSG: /************************************************************ SHUTDOWN_MSG: Shutting down NameNode at ubuntu/127.0.1.1 ************************************************************/ by this command: babak@ubuntu:~/Downloads/hadoop/bin$ start-dfs.sh this is the out put mkdir: kann Verzeichnis „/home/babak/Downloads/hadoop/bin/../logs“ nicht anlegen: Keine Berechtigung this out put(it's in german and it means no right to make this folder) On Wed, Jun 6, 2012 at 7:59 PM, Mohammad Tariq <donta...@gmail.com> wrote: > once we are done with the configuration, we need to format the file > system..use this command to do that- > bin/hadoop namenode -format > > after this, hadoop daemon processes should be started using following > commands - > bin/start-dfs.sh (it'll start NN & DN) > bin/start-mapred.sh (it'll start JT & TT) > > after this use jps to check if everything is alright or point your > browser to localhost:50070..if you further find any problem provide us > with the error logs..:) > > Regards, > Mohammad Tariq > > > On Wed, Jun 6, 2012 at 11:22 PM, Babak Bastan <babak...@gmail.com> wrote: > > were you able to format hdfs properly??? > > I did'nt get your question,Do you mean HADOOP_HOME? or where did I > install > > Hadoop? > > > > On Wed, Jun 6, 2012 at 7:49 PM, Mohammad Tariq <donta...@gmail.com> > wrote: > >> > >> if you are getting only this, it means your hadoop is not > >> running..were you able to format hdfs properly??? > >> > >> Regards, > >> Mohammad Tariq > >> > >> > >> On Wed, Jun 6, 2012 at 11:17 PM, Babak Bastan <babak...@gmail.com> > wrote: > >> > Hi MohammadmI irun jps in my shel I can see this result: > >> > 2213 Jps > >> > > >> > > >> > On Wed, Jun 6, 2012 at 7:44 PM, Mohammad Tariq <donta...@gmail.com> > >> > wrote: > >> >> > >> >> you can also use "jps" command at your shell to see whether Hadoop > >> >> processes are running or not. > >> >> > >> >> Regards, > >> >> Mohammad Tariq > >> >> > >> >> > >> >> On Wed, Jun 6, 2012 at 11:12 PM, Mohammad Tariq <donta...@gmail.com> > >> >> wrote: > >> >> > Hi Babak, > >> >> > > >> >> > You have to type it in you web browser..Hadoop provides us a web > GUI > >> >> > that not only allows us to browse through the file system, but to > >> >> > download the files as well..Apart from that it also provides a web > >> >> > GUI > >> >> > that can be used to see the status of Jobtracker and > >> >> > Tasktracker..When > >> >> > you run a Hive or Pig job or a Mapreduce job, you can point your > >> >> > browser to http://localhost:50030 to see the status and logs of > your > >> >> > job. > >> >> > > >> >> > Regards, > >> >> > Mohammad Tariq > >> >> > > >> >> > > >> >> > On Wed, Jun 6, 2012 at 8:28 PM, Babak Bastan <babak...@gmail.com> > >> >> > wrote: > >> >> >> Thank you shashwat for the answer, > >> >> >> where should I type http://localhost:50070? > >> >> >> I typed here: hive>http://localhost:50070 but nothing as result > >> >> >> > >> >> >> > >> >> >> On Wed, Jun 6, 2012 at 3:32 PM, shashwat shriparv > >> >> >> <dwivedishash...@gmail.com> wrote: > >> >> >>> > >> >> >>> first type http://localhost:50070 whether this is opening or not > >> >> >>> and > >> >> >>> check > >> >> >>> how many nodes are available, check some of the hadoop shell > >> >> >>> commands > >> >> >>> from > http://hadoop.apache.org/common/docs/r0.18.3/hdfs_shell.html > >> >> >>> run > >> >> >>> example mapreduce task on hadoop take example from here > >> >> >>> > >> >> >>> > >> >> >>> : > http://www.michael-noll.com/blog/2011/04/09/benchmarking-and-stress-testing-an-hadoop-cluster-with-terasort-testdfsio-nnbench-mrbench/ > >> >> >>> > >> >> >>> if all the above you can do sucessfully means hadoop is > configured > >> >> >>> correctly > >> >> >>> > >> >> >>> Regards > >> >> >>> Shashwat > >> >> >>> > >> >> >>> > >> >> >>> On Wed, Jun 6, 2012 at 1:30 AM, Babak Bastan <babak...@gmail.com > > > >> >> >>> wrote: > >> >> >>>> > >> >> >>>> no I'm not working on CDH.Is there a way to test if my Hadoop > >> >> >>>> works > >> >> >>>> fine > >> >> >>>> or not? > >> >> >>>> > >> >> >>>> > >> >> >>>> On Tue, Jun 5, 2012 at 9:55 PM, Bejoy KS <bejoy...@yahoo.com> > >> >> >>>> wrote: > >> >> >>>>> > >> >> >>>>> Hi Babak > >> >> >>>>> > >> >> >>>>> You gotta follow those instructions in the apace site to set up > >> >> >>>>> hadoop > >> >> >>>>> from scratch and ensure that hdfs is working first. You should > be > >> >> >>>>> able to > >> >> >>>>> read and write files to hdfs before you do your next steps. > >> >> >>>>> > >> >> >>>>> Are you on CDH or apache distribution of hadoop? If it is CDH > >> >> >>>>> there > >> >> >>>>> are > >> >> >>>>> detailed instructions on Cloudera web site. > >> >> >>>>> > >> >> >>>>> Regards > >> >> >>>>> Bejoy KS > >> >> >>>>> > >> >> >>>>> Sent from handheld, please excuse typos. > >> >> >>>>> ________________________________ > >> >> >>>>> From: Babak Bastan <babak...@gmail.com> > >> >> >>>>> Date: Tue, 5 Jun 2012 21:30:22 +0200 > >> >> >>>>> To: <user@hive.apache.org> > >> >> >>>>> ReplyTo: user@hive.apache.org > >> >> >>>>> Subject: Re: Error while Creating Table in Hive > >> >> >>>>> > >> >> >>>>> @Bejoy: I set the fs.default.name in the core-site.xml and I > did > >> >> >>>>> all > >> >> >>>>> of > >> >> >>>>> thing that was mentioned in the reference but no effect > >> >> >>>>> > >> >> >>>>> On Tue, Jun 5, 2012 at 8:43 PM, Babak Bastan < > babak...@gmail.com> > >> >> >>>>> wrote: > >> >> >>>>>> > >> >> >>>>>> Ok sorry but that was my Mistake .I thought it works but no. > >> >> >>>>>> I wrote the command without ; and then I think It works but > with > >> >> >>>>>> ; > >> >> >>>>>> at > >> >> >>>>>> the end of command > >> >> >>>>>> > >> >> >>>>>> CREATE TABLE pokes (foo INT, bar STRING); > >> >> >>>>>> > >> >> >>>>>> does'nt work > >> >> >>>>>> > >> >> >>>>>> > >> >> >>>>>> On Tue, Jun 5, 2012 at 8:34 PM, shashwat shriparv > >> >> >>>>>> <dwivedishash...@gmail.com> wrote: > >> >> >>>>>>> > >> >> >>>>>>> inside configuration. all properties will be inside the > >> >> >>>>>>> configuration > >> >> >>>>>>> tags > >> >> >>>>>>> > >> >> >>>>>>> > >> >> >>>>>>> On Tue, Jun 5, 2012 at 11:53 PM, Babak Bastan > >> >> >>>>>>> <babak...@gmail.com> > >> >> >>>>>>> wrote: > >> >> >>>>>>>> > >> >> >>>>>>>> Thank you so much my friend your idee works fine(no error) > you > >> >> >>>>>>>> are > >> >> >>>>>>>> the best :) > >> >> >>>>>>>> > >> >> >>>>>>>> > >> >> >>>>>>>> On Tue, Jun 5, 2012 at 8:20 PM, Babak Bastan > >> >> >>>>>>>> <babak...@gmail.com> > >> >> >>>>>>>> wrote: > >> >> >>>>>>>>> > >> >> >>>>>>>>> It must be inside the <configuration></configuration> or > >> >> >>>>>>>>> outside > >> >> >>>>>>>>> this? > >> >> >>>>>>>>> > >> >> >>>>>>>>> > >> >> >>>>>>>>> On Tue, Jun 5, 2012 at 8:15 PM, shashwat shriparv > >> >> >>>>>>>>> <dwivedishash...@gmail.com> wrote: > >> >> >>>>>>>>>> > >> >> >>>>>>>>>> It will be inside hive/conf > >> >> >>>>>>>>>> > >> >> >>>>>>>>>> > >> >> >>>>>>>>>> On Tue, Jun 5, 2012 at 11:43 PM, Babak Bastan > >> >> >>>>>>>>>> <babak...@gmail.com> > >> >> >>>>>>>>>> wrote: > >> >> >>>>>>>>>>> > >> >> >>>>>>>>>>> Thanks sShashwat, and where is this hive-site.xml > >> >> >>>>>>>>>>> > >> >> >>>>>>>>>>> On Tue, Jun 5, 2012 at 8:02 PM, shashwat shriparv > >> >> >>>>>>>>>>> <dwivedishash...@gmail.com> wrote: > >> >> >>>>>>>>>>>> > >> >> >>>>>>>>>>>> set > >> >> >>>>>>>>>>>> > >> >> >>>>>>>>>>>> hive.metastore.warehouse.dir in hive-site.xml > >> >> >>>>>>>>>>>> > >> >> >>>>>>>>>>>> <property> > >> >> >>>>>>>>>>>> <name>hive.metastore.local</name> > >> >> >>>>>>>>>>>> <value>true</value> > >> >> >>>>>>>>>>>> </property> > >> >> >>>>>>>>>>>> > >> >> >>>>>>>>>>>> <name>hive.metastore.warehouse.dir</name> > >> >> >>>>>>>>>>>> <value>/home/<your > >> >> >>>>>>>>>>>> username>/hivefolder</value> > >> >> >>>>>>>>>>>> <description>location of default database > >> >> >>>>>>>>>>>> for > >> >> >>>>>>>>>>>> the > >> >> >>>>>>>>>>>> warehouse</description> > >> >> >>>>>>>>>>>> </property> > >> >> >>>>>>>>>>>> > >> >> >>>>>>>>>>>> > >> >> >>>>>>>>>>>> On Tue, Jun 5, 2012 at 10:43 PM, Babak Bastan > >> >> >>>>>>>>>>>> <babak...@gmail.com> wrote: > >> >> >>>>>>>>>>>>> > >> >> >>>>>>>>>>>>> Hello Experts , > >> >> >>>>>>>>>>>>> > >> >> >>>>>>>>>>>>> I'm new in Hive .When try to create a test Table in > Hive > >> >> >>>>>>>>>>>>> I > >> >> >>>>>>>>>>>>> get > >> >> >>>>>>>>>>>>> an error.I want to run this command: > >> >> >>>>>>>>>>>>> CREATE TABLE Test (DateT STRING, Url STRING, Content > >> >> >>>>>>>>>>>>> STRING); > >> >> >>>>>>>>>>>>> but this error occured: > >> >> >>>>>>>>>>>>> FAILED: Error in metadata: MetaException(message:Got > >> >> >>>>>>>>>>>>> exception: > >> >> >>>>>>>>>>>>> java.io.FileNotFoundException File > >> >> >>>>>>>>>>>>> file:/user/hive/warehouse/test does not > >> >> >>>>>>>>>>>>> exist.) > >> >> >>>>>>>>>>>>> FAILED: Execution Error, return code 1 from > >> >> >>>>>>>>>>>>> org.apache.hadoop.hive.ql.exec.DDLTask > >> >> >>>>>>>>>>>>> How can I solve this Problem? > >> >> >>>>>>>>>>>>> Thank you so much > >> >> >>>>>>>>>>>> > >> >> >>>>>>>>>>>> > >> >> >>>>>>>>>>>> > >> >> >>>>>>>>>>>> > >> >> >>>>>>>>>>>> -- > >> >> >>>>>>>>>>>> > >> >> >>>>>>>>>>>> > >> >> >>>>>>>>>>>> ∞ > >> >> >>>>>>>>>>>> > >> >> >>>>>>>>>>>> Shashwat Shriparv > >> >> >>>>>>>>>>>> > >> >> >>>>>>>>>>>> > >> >> >>>>>>>>>>> > >> >> >>>>>>>>>> > >> >> >>>>>>>>>> > >> >> >>>>>>>>>> > >> >> >>>>>>>>>> -- > >> >> >>>>>>>>>> > >> >> >>>>>>>>>> > >> >> >>>>>>>>>> ∞ > >> >> >>>>>>>>>> > >> >> >>>>>>>>>> Shashwat Shriparv > >> >> >>>>>>>>>> > >> >> >>>>>>>>>> > >> >> >>>>>>>>> > >> >> >>>>>>>> > >> >> >>>>>>> > >> >> >>>>>>> > >> >> >>>>>>> > >> >> >>>>>>> -- > >> >> >>>>>>> > >> >> >>>>>>> > >> >> >>>>>>> ∞ > >> >> >>>>>>> > >> >> >>>>>>> Shashwat Shriparv > >> >> >>>>>>> > >> >> >>>>>>> > >> >> >>>>>> > >> >> >>>>> > >> >> >>>> > >> >> >>> > >> >> >>> > >> >> >>> > >> >> >>> -- > >> >> >>> > >> >> >>> > >> >> >>> ∞ > >> >> >>> > >> >> >>> Shashwat Shriparv > >> >> >>> > >> >> >>> > >> >> >> > >> > > >> > > > > > >