thank's Mohammad

with this command:

babak@ubuntu:~/Downloads/hadoop/bin$ hadoop namenode -format

this is my output:

12/06/06 20:05:20 INFO namenode.NameNode: STARTUP_MSG:
/************************************************************
STARTUP_MSG: Starting NameNode
STARTUP_MSG:   host = ubuntu/127.0.1.1
STARTUP_MSG:   args = [-format]
STARTUP_MSG:   version = 0.20.2
STARTUP_MSG:   build =
https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20 -r
911707; compiled by 'chrisdo' on Fri Feb 19 08:07:34 UTC 2010
************************************************************/
12/06/06 20:05:20 INFO namenode.FSNamesystem:
fsOwner=babak,babak,adm,dialout,cdrom,plugdev,lpadmin,admin,sambashare
12/06/06 20:05:20 INFO namenode.FSNamesystem: supergroup=supergroup
12/06/06 20:05:20 INFO namenode.FSNamesystem: isPermissionEnabled=true
12/06/06 20:05:20 INFO common.Storage: Image file of size 95 saved in 0 seconds.
12/06/06 20:05:20 INFO common.Storage: Storage directory
/tmp/hadoop-babak/dfs/name has been successfully formatted.
12/06/06 20:05:20 INFO namenode.NameNode: SHUTDOWN_MSG:
/************************************************************
SHUTDOWN_MSG: Shutting down NameNode at ubuntu/127.0.1.1
************************************************************/

by this command:

babak@ubuntu:~/Downloads/hadoop/bin$ start-dfs.sh

this is the out put

mkdir: kann Verzeichnis „/home/babak/Downloads/hadoop/bin/../logs“
nicht anlegen: Keine Berechtigung

this out put(it's in german and it means no right to make this folder)

On Wed, Jun 6, 2012 at 7:59 PM, Mohammad Tariq <donta...@gmail.com> wrote:

> once we are done with the configuration, we need to format the file
> system..use this command to do that-
> bin/hadoop namenode -format
>
> after this, hadoop daemon processes should be started using following
> commands -
> bin/start-dfs.sh (it'll start NN & DN)
> bin/start-mapred.sh (it'll start JT & TT)
>
> after this use jps to check if everything is alright or point your
> browser to localhost:50070..if you further find any problem provide us
> with the error logs..:)
>
> Regards,
>     Mohammad Tariq
>
>
> On Wed, Jun 6, 2012 at 11:22 PM, Babak Bastan <babak...@gmail.com> wrote:
> > were you able to format hdfs properly???
> > I did'nt get your question,Do you mean HADOOP_HOME? or where did I
> install
> > Hadoop?
> >
> > On Wed, Jun 6, 2012 at 7:49 PM, Mohammad Tariq <donta...@gmail.com>
> wrote:
> >>
> >> if you are getting only this, it means your hadoop is not
> >> running..were you able to format hdfs properly???
> >>
> >> Regards,
> >>     Mohammad Tariq
> >>
> >>
> >> On Wed, Jun 6, 2012 at 11:17 PM, Babak Bastan <babak...@gmail.com>
> wrote:
> >> > Hi MohammadmI irun jps in my shel I can see this result:
> >> > 2213 Jps
> >> >
> >> >
> >> > On Wed, Jun 6, 2012 at 7:44 PM, Mohammad Tariq <donta...@gmail.com>
> >> > wrote:
> >> >>
> >> >> you can also use "jps" command at your shell to see whether Hadoop
> >> >> processes are running or not.
> >> >>
> >> >> Regards,
> >> >>     Mohammad Tariq
> >> >>
> >> >>
> >> >> On Wed, Jun 6, 2012 at 11:12 PM, Mohammad Tariq <donta...@gmail.com>
> >> >> wrote:
> >> >> > Hi Babak,
> >> >> >
> >> >> >  You have to type it in you web browser..Hadoop provides us a web
> GUI
> >> >> > that not only allows us to browse through the file system, but to
> >> >> > download the files as well..Apart from that it also provides a web
> >> >> > GUI
> >> >> > that can be used to see the status of Jobtracker and
> >> >> > Tasktracker..When
> >> >> > you run a Hive or Pig job or a Mapreduce job, you can point your
> >> >> > browser to http://localhost:50030 to see the status and logs of
> your
> >> >> > job.
> >> >> >
> >> >> > Regards,
> >> >> >     Mohammad Tariq
> >> >> >
> >> >> >
> >> >> > On Wed, Jun 6, 2012 at 8:28 PM, Babak Bastan <babak...@gmail.com>
> >> >> > wrote:
> >> >> >> Thank you shashwat for the answer,
> >> >> >> where should I type http://localhost:50070?
> >> >> >> I typed here: hive>http://localhost:50070 but nothing as result
> >> >> >>
> >> >> >>
> >> >> >> On Wed, Jun 6, 2012 at 3:32 PM, shashwat shriparv
> >> >> >> <dwivedishash...@gmail.com> wrote:
> >> >> >>>
> >> >> >>> first type http://localhost:50070 whether this is opening or not
> >> >> >>> and
> >> >> >>> check
> >> >> >>> how many nodes are available, check some of the hadoop shell
> >> >> >>> commands
> >> >> >>> from
> http://hadoop.apache.org/common/docs/r0.18.3/hdfs_shell.html
> >> >> >>> run
> >> >> >>> example mapreduce task on hadoop take example from here
> >> >> >>>
> >> >> >>>
> >> >> >>> :
> http://www.michael-noll.com/blog/2011/04/09/benchmarking-and-stress-testing-an-hadoop-cluster-with-terasort-testdfsio-nnbench-mrbench/
> >> >> >>>
> >> >> >>> if all the above you can do sucessfully means hadoop is
> configured
> >> >> >>> correctly
> >> >> >>>
> >> >> >>> Regards
> >> >> >>> Shashwat
> >> >> >>>
> >> >> >>>
> >> >> >>> On Wed, Jun 6, 2012 at 1:30 AM, Babak Bastan <babak...@gmail.com
> >
> >> >> >>> wrote:
> >> >> >>>>
> >> >> >>>> no I'm not working on CDH.Is there a way to test if my Hadoop
> >> >> >>>> works
> >> >> >>>> fine
> >> >> >>>> or not?
> >> >> >>>>
> >> >> >>>>
> >> >> >>>> On Tue, Jun 5, 2012 at 9:55 PM, Bejoy KS <bejoy...@yahoo.com>
> >> >> >>>> wrote:
> >> >> >>>>>
> >> >> >>>>> Hi Babak
> >> >> >>>>>
> >> >> >>>>> You gotta follow those instructions in the apace site to set up
> >> >> >>>>> hadoop
> >> >> >>>>> from scratch and ensure that hdfs is working first. You should
> be
> >> >> >>>>> able to
> >> >> >>>>> read and write files to hdfs before you do your next steps.
> >> >> >>>>>
> >> >> >>>>> Are you on CDH or apache distribution of hadoop? If it is CDH
> >> >> >>>>> there
> >> >> >>>>> are
> >> >> >>>>> detailed instructions on Cloudera web site.
> >> >> >>>>>
> >> >> >>>>> Regards
> >> >> >>>>> Bejoy KS
> >> >> >>>>>
> >> >> >>>>> Sent from handheld, please excuse typos.
> >> >> >>>>> ________________________________
> >> >> >>>>> From: Babak Bastan <babak...@gmail.com>
> >> >> >>>>> Date: Tue, 5 Jun 2012 21:30:22 +0200
> >> >> >>>>> To: <user@hive.apache.org>
> >> >> >>>>> ReplyTo: user@hive.apache.org
> >> >> >>>>> Subject: Re: Error while Creating Table in Hive
> >> >> >>>>>
> >> >> >>>>> @Bejoy: I set the fs.default.name in the core-site.xml and I
> did
> >> >> >>>>> all
> >> >> >>>>> of
> >> >> >>>>> thing that was mentioned in the reference but no effect
> >> >> >>>>>
> >> >> >>>>> On Tue, Jun 5, 2012 at 8:43 PM, Babak Bastan <
> babak...@gmail.com>
> >> >> >>>>> wrote:
> >> >> >>>>>>
> >> >> >>>>>> Ok sorry but that was my Mistake .I thought it works but no.
> >> >> >>>>>> I wrote the command without ; and then I think It works but
> with
> >> >> >>>>>> ;
> >> >> >>>>>> at
> >> >> >>>>>> the end of command
> >> >> >>>>>>
> >> >> >>>>>> CREATE TABLE pokes (foo INT, bar STRING);
> >> >> >>>>>>
> >> >> >>>>>> does'nt work
> >> >> >>>>>>
> >> >> >>>>>>
> >> >> >>>>>> On Tue, Jun 5, 2012 at 8:34 PM, shashwat shriparv
> >> >> >>>>>> <dwivedishash...@gmail.com> wrote:
> >> >> >>>>>>>
> >> >> >>>>>>> inside configuration. all properties will be inside the
> >> >> >>>>>>> configuration
> >> >> >>>>>>> tags
> >> >> >>>>>>>
> >> >> >>>>>>>
> >> >> >>>>>>> On Tue, Jun 5, 2012 at 11:53 PM, Babak Bastan
> >> >> >>>>>>> <babak...@gmail.com>
> >> >> >>>>>>> wrote:
> >> >> >>>>>>>>
> >> >> >>>>>>>> Thank you so much my friend your idee works fine(no error)
> you
> >> >> >>>>>>>> are
> >> >> >>>>>>>> the best :)
> >> >> >>>>>>>>
> >> >> >>>>>>>>
> >> >> >>>>>>>> On Tue, Jun 5, 2012 at 8:20 PM, Babak Bastan
> >> >> >>>>>>>> <babak...@gmail.com>
> >> >> >>>>>>>> wrote:
> >> >> >>>>>>>>>
> >> >> >>>>>>>>> It must be inside the <configuration></configuration> or
> >> >> >>>>>>>>> outside
> >> >> >>>>>>>>> this?
> >> >> >>>>>>>>>
> >> >> >>>>>>>>>
> >> >> >>>>>>>>> On Tue, Jun 5, 2012 at 8:15 PM, shashwat shriparv
> >> >> >>>>>>>>> <dwivedishash...@gmail.com> wrote:
> >> >> >>>>>>>>>>
> >> >> >>>>>>>>>> It will be inside hive/conf
> >> >> >>>>>>>>>>
> >> >> >>>>>>>>>>
> >> >> >>>>>>>>>> On Tue, Jun 5, 2012 at 11:43 PM, Babak Bastan
> >> >> >>>>>>>>>> <babak...@gmail.com>
> >> >> >>>>>>>>>> wrote:
> >> >> >>>>>>>>>>>
> >> >> >>>>>>>>>>> Thanks sShashwat, and where is this hive-site.xml
> >> >> >>>>>>>>>>>
> >> >> >>>>>>>>>>> On Tue, Jun 5, 2012 at 8:02 PM, shashwat shriparv
> >> >> >>>>>>>>>>> <dwivedishash...@gmail.com> wrote:
> >> >> >>>>>>>>>>>>
> >> >> >>>>>>>>>>>> set
> >> >> >>>>>>>>>>>>
> >> >> >>>>>>>>>>>> hive.metastore.warehouse.dir in hive-site.xml
> >> >> >>>>>>>>>>>>
> >> >> >>>>>>>>>>>> <property>
> >> >> >>>>>>>>>>>>   <name>hive.metastore.local</name>
> >> >> >>>>>>>>>>>>   <value>true</value>
> >> >> >>>>>>>>>>>> </property>
> >> >> >>>>>>>>>>>>
> >> >> >>>>>>>>>>>> <name>hive.metastore.warehouse.dir</name>
> >> >> >>>>>>>>>>>>                <value>/home/<your
> >> >> >>>>>>>>>>>> username>/hivefolder</value>
> >> >> >>>>>>>>>>>>                <description>location of default database
> >> >> >>>>>>>>>>>> for
> >> >> >>>>>>>>>>>> the
> >> >> >>>>>>>>>>>> warehouse</description>
> >> >> >>>>>>>>>>>>        </property>
> >> >> >>>>>>>>>>>>
> >> >> >>>>>>>>>>>>
> >> >> >>>>>>>>>>>> On Tue, Jun 5, 2012 at 10:43 PM, Babak Bastan
> >> >> >>>>>>>>>>>> <babak...@gmail.com> wrote:
> >> >> >>>>>>>>>>>>>
> >> >> >>>>>>>>>>>>> Hello Experts ,
> >> >> >>>>>>>>>>>>>
> >> >> >>>>>>>>>>>>> I'm new in Hive .When try to create a test Table in
> Hive
> >> >> >>>>>>>>>>>>> I
> >> >> >>>>>>>>>>>>> get
> >> >> >>>>>>>>>>>>> an error.I want to run this command:
> >> >> >>>>>>>>>>>>> CREATE TABLE Test (DateT STRING, Url STRING, Content
> >> >> >>>>>>>>>>>>> STRING);
> >> >> >>>>>>>>>>>>> but this error occured:
> >> >> >>>>>>>>>>>>> FAILED: Error in metadata: MetaException(message:Got
> >> >> >>>>>>>>>>>>> exception:
> >> >> >>>>>>>>>>>>> java.io.FileNotFoundException File
> >> >> >>>>>>>>>>>>> file:/user/hive/warehouse/test does not
> >> >> >>>>>>>>>>>>> exist.)
> >> >> >>>>>>>>>>>>> FAILED: Execution Error, return code 1 from
> >> >> >>>>>>>>>>>>> org.apache.hadoop.hive.ql.exec.DDLTask
> >> >> >>>>>>>>>>>>> How can I solve this Problem?
> >> >> >>>>>>>>>>>>> Thank you so much
> >> >> >>>>>>>>>>>>
> >> >> >>>>>>>>>>>>
> >> >> >>>>>>>>>>>>
> >> >> >>>>>>>>>>>>
> >> >> >>>>>>>>>>>> --
> >> >> >>>>>>>>>>>>
> >> >> >>>>>>>>>>>>
> >> >> >>>>>>>>>>>> ∞
> >> >> >>>>>>>>>>>>
> >> >> >>>>>>>>>>>> Shashwat Shriparv
> >> >> >>>>>>>>>>>>
> >> >> >>>>>>>>>>>>
> >> >> >>>>>>>>>>>
> >> >> >>>>>>>>>>
> >> >> >>>>>>>>>>
> >> >> >>>>>>>>>>
> >> >> >>>>>>>>>> --
> >> >> >>>>>>>>>>
> >> >> >>>>>>>>>>
> >> >> >>>>>>>>>> ∞
> >> >> >>>>>>>>>>
> >> >> >>>>>>>>>> Shashwat Shriparv
> >> >> >>>>>>>>>>
> >> >> >>>>>>>>>>
> >> >> >>>>>>>>>
> >> >> >>>>>>>>
> >> >> >>>>>>>
> >> >> >>>>>>>
> >> >> >>>>>>>
> >> >> >>>>>>> --
> >> >> >>>>>>>
> >> >> >>>>>>>
> >> >> >>>>>>> ∞
> >> >> >>>>>>>
> >> >> >>>>>>> Shashwat Shriparv
> >> >> >>>>>>>
> >> >> >>>>>>>
> >> >> >>>>>>
> >> >> >>>>>
> >> >> >>>>
> >> >> >>>
> >> >> >>>
> >> >> >>>
> >> >> >>> --
> >> >> >>>
> >> >> >>>
> >> >> >>> ∞
> >> >> >>>
> >> >> >>> Shashwat Shriparv
> >> >> >>>
> >> >> >>>
> >> >> >>
> >> >
> >> >
> >
> >
>

Reply via email to