actually this blog post explains how to install cloudera's hadoop
distribution...if you have followed this post and installed cloudera's
distribution then your logs should ideally be inside
/usr/lib/hadoop/logs (if everything was fine)..anyway try the steps I
have given and let me know.

Regards,
    Mohammad Tariq


On Thu, Jun 7, 2012 at 1:52 AM, Babak Bastan <babak...@gmail.com> wrote:
> by the way,you are a very nice man my friend:Thank you so much :)
>
> what do you mean aboat this post in stackoverflow?
>
> I am assuming that is your first installation of hadoop.
>
> At the beginning please check if your daemons are working. To do that use
> (in terminal):
>
> jps
>
> If only jps appears that means all daemons are down. Please check the log
> files. Especially the namenode. Log folder is probably somewhere there
> /usr/lib/hadoop/logs
>
> If you have some permission problems. Use this guide during the
> installation.
>
> Good installation guide
>
> I am shooting with this explanations but these are most common problems.
>
>
> On Wed, Jun 6, 2012 at 10:15 PM, Babak Bastan <babak...@gmail.com> wrote:
>>
>> I checked it but no hadoop folder :(
>> yes you are right.I'm a student and I want to make a very very simple
>> programm hive but untill now hmmmmmmmmm
>>
>>
>> On Wed, Jun 6, 2012 at 10:12 PM, Babak Bastan <babak...@gmail.com> wrote:
>>>
>>> no one error:
>>> i.e if I run this one
>>>
>>> hostname --fqdn
>>>
>>>  with the condition that I send to you :
>>>
>>> 127.0.0.1       localhost
>>> #127.0.0.1      ubuntu.ubuntu-domain    ubuntu
>>> # The following lines are desirable for IPv6 capable hosts
>>> #::1     ip6-localhost ip6-loopback
>>> #fe00::0 ip6-localnet
>>> #ff00::0 ip6-mcastprefix
>>> #ff02::1 ip6-allnodes
>>> #ff02::2 ip6-allrouters
>>>
>>> I get this error:
>>>
>>> hostname: Name or service not known
>>>
>>> Or in the second step by this command:
>>>
>>> babak@ubuntu:~/Downloads/hadoop/bin$ start-hdfs.sh
>>>
>>> these lines of error:
>>>
>>>
>>> mkdir: kann Verzeichnis „/home/babak/Downloads/hadoop/bin/../logs“ nicht
>>> anlegen: Keine Berechtigung
>>> starting namenode, logging to
>>> /home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-namenode-ubuntu.out
>>> /home/babak/Downloads/hadoop/bin/hadoop-daemon.sh: Zeile 117:
>>> /home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-namenode-ubuntu.out:
>>> Datei oder Verzeichnis nicht gefunden
>>> head:
>>> „/home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-namenode-ubuntu.out“
>>> kann nicht zum Lesen geöffnet werden: Datei oder Verzeichnis nicht gefunden
>>> localhost: mkdir: kann Verzeichnis
>>> „/home/babak/Downloads/hadoop/bin/../logs“ nicht anlegen: Keine Berechtigung
>>> localhost: starting datanode, logging to
>>> /home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-datanode-ubuntu.out
>>> localhost: /home/babak/Downloads/hadoop/bin/hadoop-daemon.sh: Zeile 117:
>>> /home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-datanode-ubuntu.out:
>>> Datei oder Verzeichnis nicht gefunden
>>> localhost: head:
>>> „/home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-datanode-ubuntu.out“
>>> kann nicht zum Lesen geöffnet werden: Datei oder Verzeichnis nicht gefunden
>>> localhost: mkdir: kann Verzeichnis
>>> „/home/babak/Downloads/hadoop/bin/../logs“ nicht anlegen: Keine Berechtigung
>>> localhost: starting secondarynamenode, logging to
>>> /home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-secondarynamenode-ubuntu.out
>>> localhost: /home/babak/Downloads/hadoop/bin/hadoop-daemon.sh: Zeile 117:
>>> /home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-secondarynamenode-ubuntu.out:
>>> Datei oder Verzeichnis nicht gefunden
>>> localhost: head:
>>> „/home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-secondarynamenode-ubuntu.out“
>>> kann nicht zum Lesen geöffnet werden: Datei oder Verzeichnis nicht gefunden
>>>
>>> they said no permision to make logs in this
>>> path:/home/babak/Downloads/hadoop/bin/../logs
>>>
>>>  and generally I cant create a table in hive and get this one:
>>>
>>> FAILED: Error in metadata: MetaException(message:Got exception:
>>> java.io.FileNotFoundException File file:/user/hive/warehouse/test does not
>>> exist.)
>>> FAILED: Execution Error, return code 1 from
>>> org.apache.hadoop.hive.ql.exec.DDLTask
>>>
>>> On Wed, Jun 6, 2012 at 10:02 PM, shashwat shriparv
>>> <dwivedishash...@gmail.com> wrote:
>>>>
>>>> whats the error babak ???
>>>>
>>>>
>>>> On Thu, Jun 7, 2012 at 1:25 AM, Babak Bastan <babak...@gmail.com> wrote:
>>>>>
>>>>> What the hell is that?I see no log folder there
>>>>>
>>>>>
>>>>> On Wed, Jun 6, 2012 at 9:41 PM, Mohammad Tariq <donta...@gmail.com>
>>>>> wrote:
>>>>>>
>>>>>> go to your HADOOP_HOME i.e your hadoop directory(that includes bin,
>>>>>> conf etc)..you can find logs directory there..
>>>>>>
>>>>>> Regards,
>>>>>>     Mohammad Tariq
>>>>>>
>>>>>>
>>>>>> On Thu, Jun 7, 2012 at 1:09 AM, Babak Bastan <babak...@gmail.com>
>>>>>> wrote:
>>>>>> > hoe can I get my log mohammad?
>>>>>> >
>>>>>> >
>>>>>> > On Wed, Jun 6, 2012 at 9:36 PM, Mohammad Tariq <donta...@gmail.com>
>>>>>> > wrote:
>>>>>> >>
>>>>>> >> could you post your logs???that would help me in understanding the
>>>>>> >> problem properly.
>>>>>> >>
>>>>>> >> Regards,
>>>>>> >>     Mohammad Tariq
>>>>>> >>
>>>>>> >>
>>>>>> >> On Thu, Jun 7, 2012 at 1:02 AM, Babak Bastan <babak...@gmail.com>
>>>>>> >> wrote:
>>>>>> >> > Thank you very much mohamad for your attention.I followed the
>>>>>> >> > steps but
>>>>>> >> > the
>>>>>> >> > error is the same as the last time.
>>>>>> >> > and there is my hosts file:
>>>>>> >> >
>>>>>> >> > 127.0.0.1       localhost
>>>>>> >> > #127.0.0.1      ubuntu.ubuntu-domain    ubuntu
>>>>>> >> >
>>>>>> >> >
>>>>>> >> > # The following lines are desirable for IPv6 capable hosts
>>>>>> >> >
>>>>>> >> > #::1     ip6-localhost ip6-loopback
>>>>>> >> > #fe00::0 ip6-localnet
>>>>>> >> > #ff00::0 ip6-mcastprefix
>>>>>> >> > #ff02::1 ip6-allnodes
>>>>>> >> > #ff02::2 ip6-allrouters
>>>>>> >> >
>>>>>> >> > but no effect :(
>>>>>> >> >
>>>>>> >> > On Wed, Jun 6, 2012 at 8:25 PM, Mohammad Tariq
>>>>>> >> > <donta...@gmail.com>
>>>>>> >> > wrote:
>>>>>> >> >>
>>>>>> >> >> also change the permissions of these directories to 777.
>>>>>> >> >>
>>>>>> >> >> Regards,
>>>>>> >> >>     Mohammad Tariq
>>>>>> >> >>
>>>>>> >> >>
>>>>>> >> >> On Wed, Jun 6, 2012 at 11:54 PM, Mohammad Tariq
>>>>>> >> >> <donta...@gmail.com>
>>>>>> >> >> wrote:
>>>>>> >> >> > create a directory "/home/username/hdfs" (or at some place of
>>>>>> >> >> > your
>>>>>> >> >> > choice)..inside this hdfs directory create three sub
>>>>>> >> >> > directories -
>>>>>> >> >> > name, data, and temp, then follow these steps :
>>>>>> >> >> >
>>>>>> >> >> > add following properties in your core-site.xml -
>>>>>> >> >> >
>>>>>> >> >> > <property>
>>>>>> >> >> >          <name>fs.default.name</name>
>>>>>> >> >> >          <value>hdfs://localhost:9000/</value>
>>>>>> >> >> >        </property>
>>>>>> >> >> >
>>>>>> >> >> >        <property>
>>>>>> >> >> >          <name>hadoop.tmp.dir</name>
>>>>>> >> >> >          <value>/home/mohammad/hdfs/temp</value>
>>>>>> >> >> >        </property>
>>>>>> >> >> >
>>>>>> >> >> > then add following two properties in your hdfs-site.xml -
>>>>>> >> >> >
>>>>>> >> >> > <property>
>>>>>> >> >> >                <name>dfs.replication</name>
>>>>>> >> >> >                <value>1</value>
>>>>>> >> >> >        </property>
>>>>>> >> >> >
>>>>>> >> >> >        <property>
>>>>>> >> >> >                <name>dfs.name.dir</name>
>>>>>> >> >> >                <value>/home/mohammad/hdfs/name</value>
>>>>>> >> >> >        </property>
>>>>>> >> >> >
>>>>>> >> >> >        <property>
>>>>>> >> >> >                <name>dfs.data.dir</name>
>>>>>> >> >> >                <value>/home/mohammad/hdfs/data</value>
>>>>>> >> >> >        </property>
>>>>>> >> >> >
>>>>>> >> >> > finally add this property in your mapred-site.xml -
>>>>>> >> >> >
>>>>>> >> >> >       <property>
>>>>>> >> >> >          <name>mapred.job.tracker</name>
>>>>>> >> >> >          <value>hdfs://localhost:9001</value>
>>>>>> >> >> >        </property>
>>>>>> >> >> >
>>>>>> >> >> > NOTE: you can give any name to these directories of your
>>>>>> >> >> > choice, just
>>>>>> >> >> > keep in mind you have to give same names as values of
>>>>>> >> >> >           above specified properties in your configuration
>>>>>> >> >> > files.
>>>>>> >> >> > (give full path of these directories, not just the name of the
>>>>>> >> >> > directory)
>>>>>> >> >> >
>>>>>> >> >> > After this  follow the steps provided in the previous reply.
>>>>>> >> >> >
>>>>>> >> >> > Regards,
>>>>>> >> >> >     Mohammad Tariq
>>>>>> >> >> >
>>>>>> >> >> >
>>>>>> >> >> > On Wed, Jun 6, 2012 at 11:42 PM, Babak Bastan
>>>>>> >> >> > <babak...@gmail.com>
>>>>>> >> >> > wrote:
>>>>>> >> >> >> thank's Mohammad
>>>>>> >> >> >>
>>>>>> >> >> >> with this command:
>>>>>> >> >> >>
>>>>>> >> >> >> babak@ubuntu:~/Downloads/hadoop/bin$ hadoop namenode -format
>>>>>> >> >> >>
>>>>>> >> >> >> this is my output:
>>>>>> >> >> >>
>>>>>> >> >> >> 12/06/06 20:05:20 INFO namenode.NameNode: STARTUP_MSG:
>>>>>> >> >> >> /************************************************************
>>>>>> >> >> >> STARTUP_MSG: Starting NameNode
>>>>>> >> >> >> STARTUP_MSG:   host = ubuntu/127.0.1.1
>>>>>> >> >> >> STARTUP_MSG:   args = [-format]
>>>>>> >> >> >> STARTUP_MSG:   version = 0.20.2
>>>>>> >> >> >> STARTUP_MSG:   build =
>>>>>> >> >> >>
>>>>>> >> >> >> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20
>>>>>> >> >> >> -r
>>>>>> >> >> >> 911707; compiled by 'chrisdo' on Fri Feb 19 08:07:34 UTC 2010
>>>>>> >> >> >> ************************************************************/
>>>>>> >> >> >> 12/06/06 20:05:20 INFO namenode.FSNamesystem:
>>>>>> >> >> >>
>>>>>> >> >> >>
>>>>>> >> >> >> fsOwner=babak,babak,adm,dialout,cdrom,plugdev,lpadmin,admin,sambashare
>>>>>> >> >> >> 12/06/06 20:05:20 INFO namenode.FSNamesystem:
>>>>>> >> >> >> supergroup=supergroup
>>>>>> >> >> >> 12/06/06 20:05:20 INFO namenode.FSNamesystem:
>>>>>> >> >> >> isPermissionEnabled=true
>>>>>> >> >> >> 12/06/06 20:05:20 INFO common.Storage: Image file of size 95
>>>>>> >> >> >> saved
>>>>>> >> >> >> in 0
>>>>>> >> >> >> seconds.
>>>>>> >> >> >> 12/06/06 20:05:20 INFO common.Storage: Storage directory
>>>>>> >> >> >> /tmp/hadoop-babak/dfs/name has been successfully formatted.
>>>>>> >> >> >> 12/06/06 20:05:20 INFO namenode.NameNode: SHUTDOWN_MSG:
>>>>>> >> >> >> /************************************************************
>>>>>> >> >> >> SHUTDOWN_MSG: Shutting down NameNode at ubuntu/127.0.1.1
>>>>>> >> >> >> ************************************************************/
>>>>>> >> >> >>
>>>>>> >> >> >> by this command:
>>>>>> >> >> >>
>>>>>> >> >> >> babak@ubuntu:~/Downloads/hadoop/bin$ start-dfs.sh
>>>>>> >> >> >>
>>>>>> >> >> >> this is the out put
>>>>>> >> >> >>
>>>>>> >> >> >> mkdir: kann Verzeichnis
>>>>>> >> >> >> „/home/babak/Downloads/hadoop/bin/../logs“
>>>>>> >> >> >> nicht
>>>>>> >> >> >> anlegen: Keine Berechtigung
>>>>>> >> >> >>
>>>>>> >> >> >> this out put(it's in german and it means no right to make
>>>>>> >> >> >> this
>>>>>> >> >> >> folder)
>>>>>> >> >> >>
>>>>>> >> >> >>
>>>>>> >> >> >> On Wed, Jun 6, 2012 at 7:59 PM, Mohammad Tariq
>>>>>> >> >> >> <donta...@gmail.com>
>>>>>> >> >> >> wrote:
>>>>>> >> >> >>>
>>>>>> >> >> >>> once we are done with the configuration, we need to format
>>>>>> >> >> >>> the file
>>>>>> >> >> >>> system..use this command to do that-
>>>>>> >> >> >>> bin/hadoop namenode -format
>>>>>> >> >> >>>
>>>>>> >> >> >>> after this, hadoop daemon processes should be started using
>>>>>> >> >> >>> following
>>>>>> >> >> >>> commands -
>>>>>> >> >> >>> bin/start-dfs.sh (it'll start NN & DN)
>>>>>> >> >> >>> bin/start-mapred.sh (it'll start JT & TT)
>>>>>> >> >> >>>
>>>>>> >> >> >>> after this use jps to check if everything is alright or
>>>>>> >> >> >>> point your
>>>>>> >> >> >>> browser to localhost:50070..if you further find any problem
>>>>>> >> >> >>> provide
>>>>>> >> >> >>> us
>>>>>> >> >> >>> with the error logs..:)
>>>>>> >> >> >>>
>>>>>> >> >> >>> Regards,
>>>>>> >> >> >>>     Mohammad Tariq
>>>>>> >> >> >>>
>>>>>> >> >> >>>
>>>>>> >> >> >>> On Wed, Jun 6, 2012 at 11:22 PM, Babak Bastan
>>>>>> >> >> >>> <babak...@gmail.com>
>>>>>> >> >> >>> wrote:
>>>>>> >> >> >>> > were you able to format hdfs properly???
>>>>>> >> >> >>> > I did'nt get your question,Do you mean HADOOP_HOME? or
>>>>>> >> >> >>> > where did
>>>>>> >> >> >>> > I
>>>>>> >> >> >>> > install
>>>>>> >> >> >>> > Hadoop?
>>>>>> >> >> >>> >
>>>>>> >> >> >>> > On Wed, Jun 6, 2012 at 7:49 PM, Mohammad Tariq
>>>>>> >> >> >>> > <donta...@gmail.com>
>>>>>> >> >> >>> > wrote:
>>>>>> >> >> >>> >>
>>>>>> >> >> >>> >> if you are getting only this, it means your hadoop is not
>>>>>> >> >> >>> >> running..were you able to format hdfs properly???
>>>>>> >> >> >>> >>
>>>>>> >> >> >>> >> Regards,
>>>>>> >> >> >>> >>     Mohammad Tariq
>>>>>> >> >> >>> >>
>>>>>> >> >> >>> >>
>>>>>> >> >> >>> >> On Wed, Jun 6, 2012 at 11:17 PM, Babak Bastan
>>>>>> >> >> >>> >> <babak...@gmail.com>
>>>>>> >> >> >>> >> wrote:
>>>>>> >> >> >>> >> > Hi MohammadmI irun jps in my shel I can see this
>>>>>> >> >> >>> >> > result:
>>>>>> >> >> >>> >> > 2213 Jps
>>>>>> >> >> >>> >> >
>>>>>> >> >> >>> >> >
>>>>>> >> >> >>> >> > On Wed, Jun 6, 2012 at 7:44 PM, Mohammad Tariq
>>>>>> >> >> >>> >> > <donta...@gmail.com>
>>>>>> >> >> >>> >> > wrote:
>>>>>> >> >> >>> >> >>
>>>>>> >> >> >>> >> >> you can also use "jps" command at your shell to see
>>>>>> >> >> >>> >> >> whether
>>>>>> >> >> >>> >> >> Hadoop
>>>>>> >> >> >>> >> >> processes are running or not.
>>>>>> >> >> >>> >> >>
>>>>>> >> >> >>> >> >> Regards,
>>>>>> >> >> >>> >> >>     Mohammad Tariq
>>>>>> >> >> >>> >> >>
>>>>>> >> >> >>> >> >>
>>>>>> >> >> >>> >> >> On Wed, Jun 6, 2012 at 11:12 PM, Mohammad Tariq
>>>>>> >> >> >>> >> >> <donta...@gmail.com>
>>>>>> >> >> >>> >> >> wrote:
>>>>>> >> >> >>> >> >> > Hi Babak,
>>>>>> >> >> >>> >> >> >
>>>>>> >> >> >>> >> >> >  You have to type it in you web browser..Hadoop
>>>>>> >> >> >>> >> >> > provides us
>>>>>> >> >> >>> >> >> > a
>>>>>> >> >> >>> >> >> > web
>>>>>> >> >> >>> >> >> > GUI
>>>>>> >> >> >>> >> >> > that not only allows us to browse through the file
>>>>>> >> >> >>> >> >> > system,
>>>>>> >> >> >>> >> >> > but
>>>>>> >> >> >>> >> >> > to
>>>>>> >> >> >>> >> >> > download the files as well..Apart from that it also
>>>>>> >> >> >>> >> >> > provides a
>>>>>> >> >> >>> >> >> > web
>>>>>> >> >> >>> >> >> > GUI
>>>>>> >> >> >>> >> >> > that can be used to see the status of Jobtracker and
>>>>>> >> >> >>> >> >> > Tasktracker..When
>>>>>> >> >> >>> >> >> > you run a Hive or Pig job or a Mapreduce job, you
>>>>>> >> >> >>> >> >> > can point
>>>>>> >> >> >>> >> >> > your
>>>>>> >> >> >>> >> >> > browser to http://localhost:50030 to see the status
>>>>>> >> >> >>> >> >> > and
>>>>>> >> >> >>> >> >> > logs
>>>>>> >> >> >>> >> >> > of
>>>>>> >> >> >>> >> >> > your
>>>>>> >> >> >>> >> >> > job.
>>>>>> >> >> >>> >> >> >
>>>>>> >> >> >>> >> >> > Regards,
>>>>>> >> >> >>> >> >> >     Mohammad Tariq
>>>>>> >> >> >>> >> >> >
>>>>>> >> >> >>> >> >> >
>>>>>> >> >> >>> >> >> > On Wed, Jun 6, 2012 at 8:28 PM, Babak Bastan
>>>>>> >> >> >>> >> >> > <babak...@gmail.com>
>>>>>> >> >> >>> >> >> > wrote:
>>>>>> >> >> >>> >> >> >> Thank you shashwat for the answer,
>>>>>> >> >> >>> >> >> >> where should I type http://localhost:50070?
>>>>>> >> >> >>> >> >> >> I typed here: hive>http://localhost:50070 but
>>>>>> >> >> >>> >> >> >> nothing as
>>>>>> >> >> >>> >> >> >> result
>>>>>> >> >> >>> >> >> >>
>>>>>> >> >> >>> >> >> >>
>>>>>> >> >> >>> >> >> >> On Wed, Jun 6, 2012 at 3:32 PM, shashwat shriparv
>>>>>> >> >> >>> >> >> >> <dwivedishash...@gmail.com> wrote:
>>>>>> >> >> >>> >> >> >>>
>>>>>> >> >> >>> >> >> >>> first type http://localhost:50070 whether this is
>>>>>> >> >> >>> >> >> >>> opening
>>>>>> >> >> >>> >> >> >>> or
>>>>>> >> >> >>> >> >> >>> not
>>>>>> >> >> >>> >> >> >>> and
>>>>>> >> >> >>> >> >> >>> check
>>>>>> >> >> >>> >> >> >>> how many nodes are available, check some of the
>>>>>> >> >> >>> >> >> >>> hadoop
>>>>>> >> >> >>> >> >> >>> shell
>>>>>> >> >> >>> >> >> >>> commands
>>>>>> >> >> >>> >> >> >>>
>>>>>> >> >> >>> >> >> >>>
>>>>>> >> >> >>> >> >> >>>
>>>>>> >> >> >>> >> >> >>>
>>>>>> >> >> >>> >> >> >>> from http://hadoop.apache.org/common/docs/r0.18.3/hdfs_shell.html
>>>>>> >> >> >>> >> >> >>> run
>>>>>> >> >> >>> >> >> >>> example mapreduce task on hadoop take example from
>>>>>> >> >> >>> >> >> >>> here
>>>>>> >> >> >>> >> >> >>>
>>>>>> >> >> >>> >> >> >>>
>>>>>> >> >> >>> >> >> >>>
>>>>>> >> >> >>> >> >> >>>
>>>>>> >> >> >>> >> >> >>>
>>>>>> >> >> >>> >> >> >>>
>>>>>> >> >> >>> >> >> >>> : http://www.michael-noll.com/blog/2011/04/09/benchmarking-and-stress-testing-an-hadoop-cluster-with-terasort-testdfsio-nnbench-mrbench/
>>>>>> >> >> >>> >> >> >>>
>>>>>> >> >> >>> >> >> >>> if all the above you can do sucessfully means
>>>>>> >> >> >>> >> >> >>> hadoop is
>>>>>> >> >> >>> >> >> >>> configured
>>>>>> >> >> >>> >> >> >>> correctly
>>>>>> >> >> >>> >> >> >>>
>>>>>> >> >> >>> >> >> >>> Regards
>>>>>> >> >> >>> >> >> >>> Shashwat
>>>>>> >> >> >>> >> >> >>>
>>>>>> >> >> >>> >> >> >>>
>>>>>> >> >> >>> >> >> >>> On Wed, Jun 6, 2012 at 1:30 AM, Babak Bastan
>>>>>> >> >> >>> >> >> >>> <babak...@gmail.com>
>>>>>> >> >> >>> >> >> >>> wrote:
>>>>>> >> >> >>> >> >> >>>>
>>>>>> >> >> >>> >> >> >>>> no I'm not working on CDH.Is there a way to test
>>>>>> >> >> >>> >> >> >>>> if my
>>>>>> >> >> >>> >> >> >>>> Hadoop
>>>>>> >> >> >>> >> >> >>>> works
>>>>>> >> >> >>> >> >> >>>> fine
>>>>>> >> >> >>> >> >> >>>> or not?
>>>>>> >> >> >>> >> >> >>>>
>>>>>> >> >> >>> >> >> >>>>
>>>>>> >> >> >>> >> >> >>>> On Tue, Jun 5, 2012 at 9:55 PM, Bejoy KS
>>>>>> >> >> >>> >> >> >>>> <bejoy...@yahoo.com>
>>>>>> >> >> >>> >> >> >>>> wrote:
>>>>>> >> >> >>> >> >> >>>>>
>>>>>> >> >> >>> >> >> >>>>> Hi Babak
>>>>>> >> >> >>> >> >> >>>>>
>>>>>> >> >> >>> >> >> >>>>> You gotta follow those instructions in the apace
>>>>>> >> >> >>> >> >> >>>>> site
>>>>>> >> >> >>> >> >> >>>>> to
>>>>>> >> >> >>> >> >> >>>>> set
>>>>>> >> >> >>> >> >> >>>>> up
>>>>>> >> >> >>> >> >> >>>>> hadoop
>>>>>> >> >> >>> >> >> >>>>> from scratch and ensure that hdfs is working
>>>>>> >> >> >>> >> >> >>>>> first. You
>>>>>> >> >> >>> >> >> >>>>> should
>>>>>> >> >> >>> >> >> >>>>> be
>>>>>> >> >> >>> >> >> >>>>> able to
>>>>>> >> >> >>> >> >> >>>>> read and write files to hdfs before you do your
>>>>>> >> >> >>> >> >> >>>>> next
>>>>>> >> >> >>> >> >> >>>>> steps.
>>>>>> >> >> >>> >> >> >>>>>
>>>>>> >> >> >>> >> >> >>>>> Are you on CDH or apache distribution of hadoop?
>>>>>> >> >> >>> >> >> >>>>> If it
>>>>>> >> >> >>> >> >> >>>>> is
>>>>>> >> >> >>> >> >> >>>>> CDH
>>>>>> >> >> >>> >> >> >>>>> there
>>>>>> >> >> >>> >> >> >>>>> are
>>>>>> >> >> >>> >> >> >>>>> detailed instructions on Cloudera web site.
>>>>>> >> >> >>> >> >> >>>>>
>>>>>> >> >> >>> >> >> >>>>> Regards
>>>>>> >> >> >>> >> >> >>>>> Bejoy KS
>>>>>> >> >> >>> >> >> >>>>>
>>>>>> >> >> >>> >> >> >>>>> Sent from handheld, please excuse typos.
>>>>>> >> >> >>> >> >> >>>>> ________________________________
>>>>>> >> >> >>> >> >> >>>>> From: Babak Bastan <babak...@gmail.com>
>>>>>> >> >> >>> >> >> >>>>> Date: Tue, 5 Jun 2012 21:30:22 +0200
>>>>>> >> >> >>> >> >> >>>>> To: <user@hive.apache.org>
>>>>>> >> >> >>> >> >> >>>>> ReplyTo: user@hive.apache.org
>>>>>> >> >> >>> >> >> >>>>> Subject: Re: Error while Creating Table in Hive
>>>>>> >> >> >>> >> >> >>>>>
>>>>>> >> >> >>> >> >> >>>>> @Bejoy: I set the fs.default.name in the
>>>>>> >> >> >>> >> >> >>>>> core-site.xml
>>>>>> >> >> >>> >> >> >>>>> and
>>>>>> >> >> >>> >> >> >>>>> I
>>>>>> >> >> >>> >> >> >>>>> did
>>>>>> >> >> >>> >> >> >>>>> all
>>>>>> >> >> >>> >> >> >>>>> of
>>>>>> >> >> >>> >> >> >>>>> thing that was mentioned in the reference but no
>>>>>> >> >> >>> >> >> >>>>> effect
>>>>>> >> >> >>> >> >> >>>>>
>>>>>> >> >> >>> >> >> >>>>> On Tue, Jun 5, 2012 at 8:43 PM, Babak Bastan
>>>>>> >> >> >>> >> >> >>>>> <babak...@gmail.com>
>>>>>> >> >> >>> >> >> >>>>> wrote:
>>>>>> >> >> >>> >> >> >>>>>>
>>>>>> >> >> >>> >> >> >>>>>> Ok sorry but that was my Mistake .I thought it
>>>>>> >> >> >>> >> >> >>>>>> works
>>>>>> >> >> >>> >> >> >>>>>> but
>>>>>> >> >> >>> >> >> >>>>>> no.
>>>>>> >> >> >>> >> >> >>>>>> I wrote the command without ; and then I think
>>>>>> >> >> >>> >> >> >>>>>> It
>>>>>> >> >> >>> >> >> >>>>>> works
>>>>>> >> >> >>> >> >> >>>>>> but
>>>>>> >> >> >>> >> >> >>>>>> with
>>>>>> >> >> >>> >> >> >>>>>> ;
>>>>>> >> >> >>> >> >> >>>>>> at
>>>>>> >> >> >>> >> >> >>>>>> the end of command
>>>>>> >> >> >>> >> >> >>>>>>
>>>>>> >> >> >>> >> >> >>>>>> CREATE TABLE pokes (foo INT, bar STRING);
>>>>>> >> >> >>> >> >> >>>>>>
>>>>>> >> >> >>> >> >> >>>>>> does'nt work
>>>>>> >> >> >>> >> >> >>>>>>
>>>>>> >> >> >>> >> >> >>>>>>
>>>>>> >> >> >>> >> >> >>>>>> On Tue, Jun 5, 2012 at 8:34 PM, shashwat
>>>>>> >> >> >>> >> >> >>>>>> shriparv
>>>>>> >> >> >>> >> >> >>>>>> <dwivedishash...@gmail.com> wrote:
>>>>>> >> >> >>> >> >> >>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>> inside configuration. all properties will be
>>>>>> >> >> >>> >> >> >>>>>>> inside
>>>>>> >> >> >>> >> >> >>>>>>> the
>>>>>> >> >> >>> >> >> >>>>>>> configuration
>>>>>> >> >> >>> >> >> >>>>>>> tags
>>>>>> >> >> >>> >> >> >>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>> On Tue, Jun 5, 2012 at 11:53 PM, Babak Bastan
>>>>>> >> >> >>> >> >> >>>>>>> <babak...@gmail.com>
>>>>>> >> >> >>> >> >> >>>>>>> wrote:
>>>>>> >> >> >>> >> >> >>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>> Thank you so much my friend your idee works
>>>>>> >> >> >>> >> >> >>>>>>>> fine(no
>>>>>> >> >> >>> >> >> >>>>>>>> error)
>>>>>> >> >> >>> >> >> >>>>>>>> you
>>>>>> >> >> >>> >> >> >>>>>>>> are
>>>>>> >> >> >>> >> >> >>>>>>>> the best :)
>>>>>> >> >> >>> >> >> >>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>> On Tue, Jun 5, 2012 at 8:20 PM, Babak Bastan
>>>>>> >> >> >>> >> >> >>>>>>>> <babak...@gmail.com>
>>>>>> >> >> >>> >> >> >>>>>>>> wrote:
>>>>>> >> >> >>> >> >> >>>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>>> It must be inside the
>>>>>> >> >> >>> >> >> >>>>>>>>> <configuration></configuration>
>>>>>> >> >> >>> >> >> >>>>>>>>> or
>>>>>> >> >> >>> >> >> >>>>>>>>> outside
>>>>>> >> >> >>> >> >> >>>>>>>>> this?
>>>>>> >> >> >>> >> >> >>>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>>> On Tue, Jun 5, 2012 at 8:15 PM, shashwat
>>>>>> >> >> >>> >> >> >>>>>>>>> shriparv
>>>>>> >> >> >>> >> >> >>>>>>>>> <dwivedishash...@gmail.com> wrote:
>>>>>> >> >> >>> >> >> >>>>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>>>> It will be inside hive/conf
>>>>>> >> >> >>> >> >> >>>>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>>>> On Tue, Jun 5, 2012 at 11:43 PM, Babak
>>>>>> >> >> >>> >> >> >>>>>>>>>> Bastan
>>>>>> >> >> >>> >> >> >>>>>>>>>> <babak...@gmail.com>
>>>>>> >> >> >>> >> >> >>>>>>>>>> wrote:
>>>>>> >> >> >>> >> >> >>>>>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>>>>> Thanks sShashwat, and where is this
>>>>>> >> >> >>> >> >> >>>>>>>>>>> hive-site.xml
>>>>>> >> >> >>> >> >> >>>>>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>>>>> On Tue, Jun 5, 2012 at 8:02 PM, shashwat
>>>>>> >> >> >>> >> >> >>>>>>>>>>> shriparv
>>>>>> >> >> >>> >> >> >>>>>>>>>>> <dwivedishash...@gmail.com> wrote:
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>>>>>> set
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>>>>>> hive.metastore.warehouse.dir in
>>>>>> >> >> >>> >> >> >>>>>>>>>>>> hive-site.xml
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>>>>>> <property>
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>   <name>hive.metastore.local</name>
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>   <value>true</value>
>>>>>> >> >> >>> >> >> >>>>>>>>>>>> </property>
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>>>>>> <name>hive.metastore.warehouse.dir</name>
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>                <value>/home/<your
>>>>>> >> >> >>> >> >> >>>>>>>>>>>> username>/hivefolder</value>
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>                <description>location of
>>>>>> >> >> >>> >> >> >>>>>>>>>>>> default
>>>>>> >> >> >>> >> >> >>>>>>>>>>>> database
>>>>>> >> >> >>> >> >> >>>>>>>>>>>> for
>>>>>> >> >> >>> >> >> >>>>>>>>>>>> the
>>>>>> >> >> >>> >> >> >>>>>>>>>>>> warehouse</description>
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>        </property>
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>>>>>> On Tue, Jun 5, 2012 at 10:43 PM, Babak
>>>>>> >> >> >>> >> >> >>>>>>>>>>>> Bastan
>>>>>> >> >> >>> >> >> >>>>>>>>>>>> <babak...@gmail.com> wrote:
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>> Hello Experts ,
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>> I'm new in Hive .When try to create a
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>> test
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>> Table
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>> in
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>> Hive
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>> I
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>> get
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>> an error.I want to run this command:
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>> CREATE TABLE Test (DateT STRING, Url
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>> STRING,
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>> Content
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>> STRING);
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>> but this error occured:
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>> FAILED: Error in metadata:
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>> MetaException(message:Got
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>> exception:
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>> java.io.FileNotFoundException File
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>> file:/user/hive/warehouse/test does not
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>> exist.)
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>> FAILED: Execution Error, return code 1
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>> from
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>> org.apache.hadoop.hive.ql.exec.DDLTask
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>> How can I solve this Problem?
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>> Thank you so much
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>>>>>> --
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>>>>>> ∞
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>>>>>> Shashwat Shriparv
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>>>> --
>>>>>> >> >> >>> >> >> >>>>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>>>> ∞
>>>>>> >> >> >>> >> >> >>>>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>>>> Shashwat Shriparv
>>>>>> >> >> >>> >> >> >>>>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>> --
>>>>>> >> >> >>> >> >> >>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>> ∞
>>>>>> >> >> >>> >> >> >>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>> Shashwat Shriparv
>>>>>> >> >> >>> >> >> >>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>>
>>>>>> >> >> >>> >> >> >>>>>>
>>>>>> >> >> >>> >> >> >>>>>
>>>>>> >> >> >>> >> >> >>>>
>>>>>> >> >> >>> >> >> >>>
>>>>>> >> >> >>> >> >> >>>
>>>>>> >> >> >>> >> >> >>>
>>>>>> >> >> >>> >> >> >>> --
>>>>>> >> >> >>> >> >> >>>
>>>>>> >> >> >>> >> >> >>>
>>>>>> >> >> >>> >> >> >>> ∞
>>>>>> >> >> >>> >> >> >>>
>>>>>> >> >> >>> >> >> >>> Shashwat Shriparv
>>>>>> >> >> >>> >> >> >>>
>>>>>> >> >> >>> >> >> >>>
>>>>>> >> >> >>> >> >> >>
>>>>>> >> >> >>> >> >
>>>>>> >> >> >>> >> >
>>>>>> >> >> >>> >
>>>>>> >> >> >>> >
>>>>>> >> >> >>
>>>>>> >> >> >>
>>>>>> >> >
>>>>>> >> >
>>>>>> >
>>>>>> >
>>>>>
>>>>>
>>>>
>>>>
>>>>
>>>> --
>>>>
>>>>
>>>> ∞
>>>>
>>>> Shashwat Shriparv
>>>>
>>>>
>>>
>>
>

Reply via email to