could you post your logs???that would help me in understanding the
problem properly.

Regards,
    Mohammad Tariq


On Thu, Jun 7, 2012 at 1:02 AM, Babak Bastan <babak...@gmail.com> wrote:
> Thank you very much mohamad for your attention.I followed the steps but the
> error is the same as the last time.
> and there is my hosts file:
>
> 127.0.0.1       localhost
> #127.0.0.1      ubuntu.ubuntu-domain    ubuntu
>
>
> # The following lines are desirable for IPv6 capable hosts
>
> #::1     ip6-localhost ip6-loopback
> #fe00::0 ip6-localnet
> #ff00::0 ip6-mcastprefix
> #ff02::1 ip6-allnodes
> #ff02::2 ip6-allrouters
>
> but no effect :(
>
> On Wed, Jun 6, 2012 at 8:25 PM, Mohammad Tariq <donta...@gmail.com> wrote:
>>
>> also change the permissions of these directories to 777.
>>
>> Regards,
>>     Mohammad Tariq
>>
>>
>> On Wed, Jun 6, 2012 at 11:54 PM, Mohammad Tariq <donta...@gmail.com>
>> wrote:
>> > create a directory "/home/username/hdfs" (or at some place of your
>> > choice)..inside this hdfs directory create three sub directories -
>> > name, data, and temp, then follow these steps :
>> >
>> > add following properties in your core-site.xml -
>> >
>> > <property>
>> >          <name>fs.default.name</name>
>> >          <value>hdfs://localhost:9000/</value>
>> >        </property>
>> >
>> >        <property>
>> >          <name>hadoop.tmp.dir</name>
>> >          <value>/home/mohammad/hdfs/temp</value>
>> >        </property>
>> >
>> > then add following two properties in your hdfs-site.xml -
>> >
>> > <property>
>> >                <name>dfs.replication</name>
>> >                <value>1</value>
>> >        </property>
>> >
>> >        <property>
>> >                <name>dfs.name.dir</name>
>> >                <value>/home/mohammad/hdfs/name</value>
>> >        </property>
>> >
>> >        <property>
>> >                <name>dfs.data.dir</name>
>> >                <value>/home/mohammad/hdfs/data</value>
>> >        </property>
>> >
>> > finally add this property in your mapred-site.xml -
>> >
>> >       <property>
>> >          <name>mapred.job.tracker</name>
>> >          <value>hdfs://localhost:9001</value>
>> >        </property>
>> >
>> > NOTE: you can give any name to these directories of your choice, just
>> > keep in mind you have to give same names as values of
>> >           above specified properties in your configuration files.
>> > (give full path of these directories, not just the name of the
>> > directory)
>> >
>> > After this  follow the steps provided in the previous reply.
>> >
>> > Regards,
>> >     Mohammad Tariq
>> >
>> >
>> > On Wed, Jun 6, 2012 at 11:42 PM, Babak Bastan <babak...@gmail.com>
>> > wrote:
>> >> thank's Mohammad
>> >>
>> >> with this command:
>> >>
>> >> babak@ubuntu:~/Downloads/hadoop/bin$ hadoop namenode -format
>> >>
>> >> this is my output:
>> >>
>> >> 12/06/06 20:05:20 INFO namenode.NameNode: STARTUP_MSG:
>> >> /************************************************************
>> >> STARTUP_MSG: Starting NameNode
>> >> STARTUP_MSG:   host = ubuntu/127.0.1.1
>> >> STARTUP_MSG:   args = [-format]
>> >> STARTUP_MSG:   version = 0.20.2
>> >> STARTUP_MSG:   build =
>> >> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20 -r
>> >> 911707; compiled by 'chrisdo' on Fri Feb 19 08:07:34 UTC 2010
>> >> ************************************************************/
>> >> 12/06/06 20:05:20 INFO namenode.FSNamesystem:
>> >> fsOwner=babak,babak,adm,dialout,cdrom,plugdev,lpadmin,admin,sambashare
>> >> 12/06/06 20:05:20 INFO namenode.FSNamesystem: supergroup=supergroup
>> >> 12/06/06 20:05:20 INFO namenode.FSNamesystem: isPermissionEnabled=true
>> >> 12/06/06 20:05:20 INFO common.Storage: Image file of size 95 saved in 0
>> >> seconds.
>> >> 12/06/06 20:05:20 INFO common.Storage: Storage directory
>> >> /tmp/hadoop-babak/dfs/name has been successfully formatted.
>> >> 12/06/06 20:05:20 INFO namenode.NameNode: SHUTDOWN_MSG:
>> >> /************************************************************
>> >> SHUTDOWN_MSG: Shutting down NameNode at ubuntu/127.0.1.1
>> >> ************************************************************/
>> >>
>> >> by this command:
>> >>
>> >> babak@ubuntu:~/Downloads/hadoop/bin$ start-dfs.sh
>> >>
>> >> this is the out put
>> >>
>> >> mkdir: kann Verzeichnis „/home/babak/Downloads/hadoop/bin/../logs“
>> >> nicht
>> >> anlegen: Keine Berechtigung
>> >>
>> >> this out put(it's in german and it means no right to make this folder)
>> >>
>> >>
>> >> On Wed, Jun 6, 2012 at 7:59 PM, Mohammad Tariq <donta...@gmail.com>
>> >> wrote:
>> >>>
>> >>> once we are done with the configuration, we need to format the file
>> >>> system..use this command to do that-
>> >>> bin/hadoop namenode -format
>> >>>
>> >>> after this, hadoop daemon processes should be started using following
>> >>> commands -
>> >>> bin/start-dfs.sh (it'll start NN & DN)
>> >>> bin/start-mapred.sh (it'll start JT & TT)
>> >>>
>> >>> after this use jps to check if everything is alright or point your
>> >>> browser to localhost:50070..if you further find any problem provide us
>> >>> with the error logs..:)
>> >>>
>> >>> Regards,
>> >>>     Mohammad Tariq
>> >>>
>> >>>
>> >>> On Wed, Jun 6, 2012 at 11:22 PM, Babak Bastan <babak...@gmail.com>
>> >>> wrote:
>> >>> > were you able to format hdfs properly???
>> >>> > I did'nt get your question,Do you mean HADOOP_HOME? or where did I
>> >>> > install
>> >>> > Hadoop?
>> >>> >
>> >>> > On Wed, Jun 6, 2012 at 7:49 PM, Mohammad Tariq <donta...@gmail.com>
>> >>> > wrote:
>> >>> >>
>> >>> >> if you are getting only this, it means your hadoop is not
>> >>> >> running..were you able to format hdfs properly???
>> >>> >>
>> >>> >> Regards,
>> >>> >>     Mohammad Tariq
>> >>> >>
>> >>> >>
>> >>> >> On Wed, Jun 6, 2012 at 11:17 PM, Babak Bastan <babak...@gmail.com>
>> >>> >> wrote:
>> >>> >> > Hi MohammadmI irun jps in my shel I can see this result:
>> >>> >> > 2213 Jps
>> >>> >> >
>> >>> >> >
>> >>> >> > On Wed, Jun 6, 2012 at 7:44 PM, Mohammad Tariq
>> >>> >> > <donta...@gmail.com>
>> >>> >> > wrote:
>> >>> >> >>
>> >>> >> >> you can also use "jps" command at your shell to see whether
>> >>> >> >> Hadoop
>> >>> >> >> processes are running or not.
>> >>> >> >>
>> >>> >> >> Regards,
>> >>> >> >>     Mohammad Tariq
>> >>> >> >>
>> >>> >> >>
>> >>> >> >> On Wed, Jun 6, 2012 at 11:12 PM, Mohammad Tariq
>> >>> >> >> <donta...@gmail.com>
>> >>> >> >> wrote:
>> >>> >> >> > Hi Babak,
>> >>> >> >> >
>> >>> >> >> >  You have to type it in you web browser..Hadoop provides us a
>> >>> >> >> > web
>> >>> >> >> > GUI
>> >>> >> >> > that not only allows us to browse through the file system, but
>> >>> >> >> > to
>> >>> >> >> > download the files as well..Apart from that it also provides a
>> >>> >> >> > web
>> >>> >> >> > GUI
>> >>> >> >> > that can be used to see the status of Jobtracker and
>> >>> >> >> > Tasktracker..When
>> >>> >> >> > you run a Hive or Pig job or a Mapreduce job, you can point
>> >>> >> >> > your
>> >>> >> >> > browser to http://localhost:50030 to see the status and logs
>> >>> >> >> > of
>> >>> >> >> > your
>> >>> >> >> > job.
>> >>> >> >> >
>> >>> >> >> > Regards,
>> >>> >> >> >     Mohammad Tariq
>> >>> >> >> >
>> >>> >> >> >
>> >>> >> >> > On Wed, Jun 6, 2012 at 8:28 PM, Babak Bastan
>> >>> >> >> > <babak...@gmail.com>
>> >>> >> >> > wrote:
>> >>> >> >> >> Thank you shashwat for the answer,
>> >>> >> >> >> where should I type http://localhost:50070?
>> >>> >> >> >> I typed here: hive>http://localhost:50070 but nothing as
>> >>> >> >> >> result
>> >>> >> >> >>
>> >>> >> >> >>
>> >>> >> >> >> On Wed, Jun 6, 2012 at 3:32 PM, shashwat shriparv
>> >>> >> >> >> <dwivedishash...@gmail.com> wrote:
>> >>> >> >> >>>
>> >>> >> >> >>> first type http://localhost:50070 whether this is opening or
>> >>> >> >> >>> not
>> >>> >> >> >>> and
>> >>> >> >> >>> check
>> >>> >> >> >>> how many nodes are available, check some of the hadoop shell
>> >>> >> >> >>> commands
>> >>> >> >> >>>
>> >>> >> >> >>>
>> >>> >> >> >>> from http://hadoop.apache.org/common/docs/r0.18.3/hdfs_shell.html
>> >>> >> >> >>> run
>> >>> >> >> >>> example mapreduce task on hadoop take example from here
>> >>> >> >> >>>
>> >>> >> >> >>>
>> >>> >> >> >>>
>> >>> >> >> >>>
>> >>> >> >> >>> : http://www.michael-noll.com/blog/2011/04/09/benchmarking-and-stress-testing-an-hadoop-cluster-with-terasort-testdfsio-nnbench-mrbench/
>> >>> >> >> >>>
>> >>> >> >> >>> if all the above you can do sucessfully means hadoop is
>> >>> >> >> >>> configured
>> >>> >> >> >>> correctly
>> >>> >> >> >>>
>> >>> >> >> >>> Regards
>> >>> >> >> >>> Shashwat
>> >>> >> >> >>>
>> >>> >> >> >>>
>> >>> >> >> >>> On Wed, Jun 6, 2012 at 1:30 AM, Babak Bastan
>> >>> >> >> >>> <babak...@gmail.com>
>> >>> >> >> >>> wrote:
>> >>> >> >> >>>>
>> >>> >> >> >>>> no I'm not working on CDH.Is there a way to test if my
>> >>> >> >> >>>> Hadoop
>> >>> >> >> >>>> works
>> >>> >> >> >>>> fine
>> >>> >> >> >>>> or not?
>> >>> >> >> >>>>
>> >>> >> >> >>>>
>> >>> >> >> >>>> On Tue, Jun 5, 2012 at 9:55 PM, Bejoy KS
>> >>> >> >> >>>> <bejoy...@yahoo.com>
>> >>> >> >> >>>> wrote:
>> >>> >> >> >>>>>
>> >>> >> >> >>>>> Hi Babak
>> >>> >> >> >>>>>
>> >>> >> >> >>>>> You gotta follow those instructions in the apace site to
>> >>> >> >> >>>>> set
>> >>> >> >> >>>>> up
>> >>> >> >> >>>>> hadoop
>> >>> >> >> >>>>> from scratch and ensure that hdfs is working first. You
>> >>> >> >> >>>>> should
>> >>> >> >> >>>>> be
>> >>> >> >> >>>>> able to
>> >>> >> >> >>>>> read and write files to hdfs before you do your next
>> >>> >> >> >>>>> steps.
>> >>> >> >> >>>>>
>> >>> >> >> >>>>> Are you on CDH or apache distribution of hadoop? If it is
>> >>> >> >> >>>>> CDH
>> >>> >> >> >>>>> there
>> >>> >> >> >>>>> are
>> >>> >> >> >>>>> detailed instructions on Cloudera web site.
>> >>> >> >> >>>>>
>> >>> >> >> >>>>> Regards
>> >>> >> >> >>>>> Bejoy KS
>> >>> >> >> >>>>>
>> >>> >> >> >>>>> Sent from handheld, please excuse typos.
>> >>> >> >> >>>>> ________________________________
>> >>> >> >> >>>>> From: Babak Bastan <babak...@gmail.com>
>> >>> >> >> >>>>> Date: Tue, 5 Jun 2012 21:30:22 +0200
>> >>> >> >> >>>>> To: <user@hive.apache.org>
>> >>> >> >> >>>>> ReplyTo: user@hive.apache.org
>> >>> >> >> >>>>> Subject: Re: Error while Creating Table in Hive
>> >>> >> >> >>>>>
>> >>> >> >> >>>>> @Bejoy: I set the fs.default.name in the core-site.xml and
>> >>> >> >> >>>>> I
>> >>> >> >> >>>>> did
>> >>> >> >> >>>>> all
>> >>> >> >> >>>>> of
>> >>> >> >> >>>>> thing that was mentioned in the reference but no effect
>> >>> >> >> >>>>>
>> >>> >> >> >>>>> On Tue, Jun 5, 2012 at 8:43 PM, Babak Bastan
>> >>> >> >> >>>>> <babak...@gmail.com>
>> >>> >> >> >>>>> wrote:
>> >>> >> >> >>>>>>
>> >>> >> >> >>>>>> Ok sorry but that was my Mistake .I thought it works but
>> >>> >> >> >>>>>> no.
>> >>> >> >> >>>>>> I wrote the command without ; and then I think It works
>> >>> >> >> >>>>>> but
>> >>> >> >> >>>>>> with
>> >>> >> >> >>>>>> ;
>> >>> >> >> >>>>>> at
>> >>> >> >> >>>>>> the end of command
>> >>> >> >> >>>>>>
>> >>> >> >> >>>>>> CREATE TABLE pokes (foo INT, bar STRING);
>> >>> >> >> >>>>>>
>> >>> >> >> >>>>>> does'nt work
>> >>> >> >> >>>>>>
>> >>> >> >> >>>>>>
>> >>> >> >> >>>>>> On Tue, Jun 5, 2012 at 8:34 PM, shashwat shriparv
>> >>> >> >> >>>>>> <dwivedishash...@gmail.com> wrote:
>> >>> >> >> >>>>>>>
>> >>> >> >> >>>>>>> inside configuration. all properties will be inside the
>> >>> >> >> >>>>>>> configuration
>> >>> >> >> >>>>>>> tags
>> >>> >> >> >>>>>>>
>> >>> >> >> >>>>>>>
>> >>> >> >> >>>>>>> On Tue, Jun 5, 2012 at 11:53 PM, Babak Bastan
>> >>> >> >> >>>>>>> <babak...@gmail.com>
>> >>> >> >> >>>>>>> wrote:
>> >>> >> >> >>>>>>>>
>> >>> >> >> >>>>>>>> Thank you so much my friend your idee works fine(no
>> >>> >> >> >>>>>>>> error)
>> >>> >> >> >>>>>>>> you
>> >>> >> >> >>>>>>>> are
>> >>> >> >> >>>>>>>> the best :)
>> >>> >> >> >>>>>>>>
>> >>> >> >> >>>>>>>>
>> >>> >> >> >>>>>>>> On Tue, Jun 5, 2012 at 8:20 PM, Babak Bastan
>> >>> >> >> >>>>>>>> <babak...@gmail.com>
>> >>> >> >> >>>>>>>> wrote:
>> >>> >> >> >>>>>>>>>
>> >>> >> >> >>>>>>>>> It must be inside the <configuration></configuration>
>> >>> >> >> >>>>>>>>> or
>> >>> >> >> >>>>>>>>> outside
>> >>> >> >> >>>>>>>>> this?
>> >>> >> >> >>>>>>>>>
>> >>> >> >> >>>>>>>>>
>> >>> >> >> >>>>>>>>> On Tue, Jun 5, 2012 at 8:15 PM, shashwat shriparv
>> >>> >> >> >>>>>>>>> <dwivedishash...@gmail.com> wrote:
>> >>> >> >> >>>>>>>>>>
>> >>> >> >> >>>>>>>>>> It will be inside hive/conf
>> >>> >> >> >>>>>>>>>>
>> >>> >> >> >>>>>>>>>>
>> >>> >> >> >>>>>>>>>> On Tue, Jun 5, 2012 at 11:43 PM, Babak Bastan
>> >>> >> >> >>>>>>>>>> <babak...@gmail.com>
>> >>> >> >> >>>>>>>>>> wrote:
>> >>> >> >> >>>>>>>>>>>
>> >>> >> >> >>>>>>>>>>> Thanks sShashwat, and where is this hive-site.xml
>> >>> >> >> >>>>>>>>>>>
>> >>> >> >> >>>>>>>>>>> On Tue, Jun 5, 2012 at 8:02 PM, shashwat shriparv
>> >>> >> >> >>>>>>>>>>> <dwivedishash...@gmail.com> wrote:
>> >>> >> >> >>>>>>>>>>>>
>> >>> >> >> >>>>>>>>>>>> set
>> >>> >> >> >>>>>>>>>>>>
>> >>> >> >> >>>>>>>>>>>> hive.metastore.warehouse.dir in hive-site.xml
>> >>> >> >> >>>>>>>>>>>>
>> >>> >> >> >>>>>>>>>>>> <property>
>> >>> >> >> >>>>>>>>>>>>   <name>hive.metastore.local</name>
>> >>> >> >> >>>>>>>>>>>>   <value>true</value>
>> >>> >> >> >>>>>>>>>>>> </property>
>> >>> >> >> >>>>>>>>>>>>
>> >>> >> >> >>>>>>>>>>>> <name>hive.metastore.warehouse.dir</name>
>> >>> >> >> >>>>>>>>>>>>                <value>/home/<your
>> >>> >> >> >>>>>>>>>>>> username>/hivefolder</value>
>> >>> >> >> >>>>>>>>>>>>                <description>location of default
>> >>> >> >> >>>>>>>>>>>> database
>> >>> >> >> >>>>>>>>>>>> for
>> >>> >> >> >>>>>>>>>>>> the
>> >>> >> >> >>>>>>>>>>>> warehouse</description>
>> >>> >> >> >>>>>>>>>>>>        </property>
>> >>> >> >> >>>>>>>>>>>>
>> >>> >> >> >>>>>>>>>>>>
>> >>> >> >> >>>>>>>>>>>> On Tue, Jun 5, 2012 at 10:43 PM, Babak Bastan
>> >>> >> >> >>>>>>>>>>>> <babak...@gmail.com> wrote:
>> >>> >> >> >>>>>>>>>>>>>
>> >>> >> >> >>>>>>>>>>>>> Hello Experts ,
>> >>> >> >> >>>>>>>>>>>>>
>> >>> >> >> >>>>>>>>>>>>> I'm new in Hive .When try to create a test Table
>> >>> >> >> >>>>>>>>>>>>> in
>> >>> >> >> >>>>>>>>>>>>> Hive
>> >>> >> >> >>>>>>>>>>>>> I
>> >>> >> >> >>>>>>>>>>>>> get
>> >>> >> >> >>>>>>>>>>>>> an error.I want to run this command:
>> >>> >> >> >>>>>>>>>>>>> CREATE TABLE Test (DateT STRING, Url STRING,
>> >>> >> >> >>>>>>>>>>>>> Content
>> >>> >> >> >>>>>>>>>>>>> STRING);
>> >>> >> >> >>>>>>>>>>>>> but this error occured:
>> >>> >> >> >>>>>>>>>>>>> FAILED: Error in metadata:
>> >>> >> >> >>>>>>>>>>>>> MetaException(message:Got
>> >>> >> >> >>>>>>>>>>>>> exception:
>> >>> >> >> >>>>>>>>>>>>> java.io.FileNotFoundException File
>> >>> >> >> >>>>>>>>>>>>> file:/user/hive/warehouse/test does not
>> >>> >> >> >>>>>>>>>>>>> exist.)
>> >>> >> >> >>>>>>>>>>>>> FAILED: Execution Error, return code 1 from
>> >>> >> >> >>>>>>>>>>>>> org.apache.hadoop.hive.ql.exec.DDLTask
>> >>> >> >> >>>>>>>>>>>>> How can I solve this Problem?
>> >>> >> >> >>>>>>>>>>>>> Thank you so much
>> >>> >> >> >>>>>>>>>>>>
>> >>> >> >> >>>>>>>>>>>>
>> >>> >> >> >>>>>>>>>>>>
>> >>> >> >> >>>>>>>>>>>>
>> >>> >> >> >>>>>>>>>>>> --
>> >>> >> >> >>>>>>>>>>>>
>> >>> >> >> >>>>>>>>>>>>
>> >>> >> >> >>>>>>>>>>>> ∞
>> >>> >> >> >>>>>>>>>>>>
>> >>> >> >> >>>>>>>>>>>> Shashwat Shriparv
>> >>> >> >> >>>>>>>>>>>>
>> >>> >> >> >>>>>>>>>>>>
>> >>> >> >> >>>>>>>>>>>
>> >>> >> >> >>>>>>>>>>
>> >>> >> >> >>>>>>>>>>
>> >>> >> >> >>>>>>>>>>
>> >>> >> >> >>>>>>>>>> --
>> >>> >> >> >>>>>>>>>>
>> >>> >> >> >>>>>>>>>>
>> >>> >> >> >>>>>>>>>> ∞
>> >>> >> >> >>>>>>>>>>
>> >>> >> >> >>>>>>>>>> Shashwat Shriparv
>> >>> >> >> >>>>>>>>>>
>> >>> >> >> >>>>>>>>>>
>> >>> >> >> >>>>>>>>>
>> >>> >> >> >>>>>>>>
>> >>> >> >> >>>>>>>
>> >>> >> >> >>>>>>>
>> >>> >> >> >>>>>>>
>> >>> >> >> >>>>>>> --
>> >>> >> >> >>>>>>>
>> >>> >> >> >>>>>>>
>> >>> >> >> >>>>>>> ∞
>> >>> >> >> >>>>>>>
>> >>> >> >> >>>>>>> Shashwat Shriparv
>> >>> >> >> >>>>>>>
>> >>> >> >> >>>>>>>
>> >>> >> >> >>>>>>
>> >>> >> >> >>>>>
>> >>> >> >> >>>>
>> >>> >> >> >>>
>> >>> >> >> >>>
>> >>> >> >> >>>
>> >>> >> >> >>> --
>> >>> >> >> >>>
>> >>> >> >> >>>
>> >>> >> >> >>> ∞
>> >>> >> >> >>>
>> >>> >> >> >>> Shashwat Shriparv
>> >>> >> >> >>>
>> >>> >> >> >>>
>> >>> >> >> >>
>> >>> >> >
>> >>> >> >
>> >>> >
>> >>> >
>> >>
>> >>
>
>

Reply via email to