actually this blog post explains how to install cloudera's hadoop distribution...if you have followed this post and installed cloudera's distribution then your logs should ideally be inside /usr/lib/hadoop/logs (if everything was fine)..anyway try the steps I have given and let me know.
Regards, Mohammad Tariq On Thu, Jun 7, 2012 at 1:52 AM, Babak Bastan <babak...@gmail.com> wrote: > by the way,you are a very nice man my friend:Thank you so much :) > > what do you mean aboat this post in stackoverflow? > > I am assuming that is your first installation of hadoop. > > At the beginning please check if your daemons are working. To do that use > (in terminal): > > jps > > If only jps appears that means all daemons are down. Please check the log > files. Especially the namenode. Log folder is probably somewhere there > /usr/lib/hadoop/logs > > If you have some permission problems. Use this guide during the > installation. > > Good installation guide > > I am shooting with this explanations but these are most common problems. > > > On Wed, Jun 6, 2012 at 10:15 PM, Babak Bastan <babak...@gmail.com> wrote: >> >> I checked it but no hadoop folder :( >> yes you are right.I'm a student and I want to make a very very simple >> programm hive but untill now hmmmmmmmmm >> >> >> On Wed, Jun 6, 2012 at 10:12 PM, Babak Bastan <babak...@gmail.com> wrote: >>> >>> no one error: >>> i.e if I run this one >>> >>> hostname --fqdn >>> >>> with the condition that I send to you : >>> >>> 127.0.0.1 localhost >>> #127.0.0.1 ubuntu.ubuntu-domain ubuntu >>> # The following lines are desirable for IPv6 capable hosts >>> #::1 ip6-localhost ip6-loopback >>> #fe00::0 ip6-localnet >>> #ff00::0 ip6-mcastprefix >>> #ff02::1 ip6-allnodes >>> #ff02::2 ip6-allrouters >>> >>> I get this error: >>> >>> hostname: Name or service not known >>> >>> Or in the second step by this command: >>> >>> babak@ubuntu:~/Downloads/hadoop/bin$ start-hdfs.sh >>> >>> these lines of error: >>> >>> >>> mkdir: kann Verzeichnis „/home/babak/Downloads/hadoop/bin/../logs“ nicht >>> anlegen: Keine Berechtigung >>> starting namenode, logging to >>> /home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-namenode-ubuntu.out >>> /home/babak/Downloads/hadoop/bin/hadoop-daemon.sh: Zeile 117: >>> /home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-namenode-ubuntu.out: >>> Datei oder Verzeichnis nicht gefunden >>> head: >>> „/home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-namenode-ubuntu.out“ >>> kann nicht zum Lesen geöffnet werden: Datei oder Verzeichnis nicht gefunden >>> localhost: mkdir: kann Verzeichnis >>> „/home/babak/Downloads/hadoop/bin/../logs“ nicht anlegen: Keine Berechtigung >>> localhost: starting datanode, logging to >>> /home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-datanode-ubuntu.out >>> localhost: /home/babak/Downloads/hadoop/bin/hadoop-daemon.sh: Zeile 117: >>> /home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-datanode-ubuntu.out: >>> Datei oder Verzeichnis nicht gefunden >>> localhost: head: >>> „/home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-datanode-ubuntu.out“ >>> kann nicht zum Lesen geöffnet werden: Datei oder Verzeichnis nicht gefunden >>> localhost: mkdir: kann Verzeichnis >>> „/home/babak/Downloads/hadoop/bin/../logs“ nicht anlegen: Keine Berechtigung >>> localhost: starting secondarynamenode, logging to >>> /home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-secondarynamenode-ubuntu.out >>> localhost: /home/babak/Downloads/hadoop/bin/hadoop-daemon.sh: Zeile 117: >>> /home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-secondarynamenode-ubuntu.out: >>> Datei oder Verzeichnis nicht gefunden >>> localhost: head: >>> „/home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-secondarynamenode-ubuntu.out“ >>> kann nicht zum Lesen geöffnet werden: Datei oder Verzeichnis nicht gefunden >>> >>> they said no permision to make logs in this >>> path:/home/babak/Downloads/hadoop/bin/../logs >>> >>> and generally I cant create a table in hive and get this one: >>> >>> FAILED: Error in metadata: MetaException(message:Got exception: >>> java.io.FileNotFoundException File file:/user/hive/warehouse/test does not >>> exist.) >>> FAILED: Execution Error, return code 1 from >>> org.apache.hadoop.hive.ql.exec.DDLTask >>> >>> On Wed, Jun 6, 2012 at 10:02 PM, shashwat shriparv >>> <dwivedishash...@gmail.com> wrote: >>>> >>>> whats the error babak ??? >>>> >>>> >>>> On Thu, Jun 7, 2012 at 1:25 AM, Babak Bastan <babak...@gmail.com> wrote: >>>>> >>>>> What the hell is that?I see no log folder there >>>>> >>>>> >>>>> On Wed, Jun 6, 2012 at 9:41 PM, Mohammad Tariq <donta...@gmail.com> >>>>> wrote: >>>>>> >>>>>> go to your HADOOP_HOME i.e your hadoop directory(that includes bin, >>>>>> conf etc)..you can find logs directory there.. >>>>>> >>>>>> Regards, >>>>>> Mohammad Tariq >>>>>> >>>>>> >>>>>> On Thu, Jun 7, 2012 at 1:09 AM, Babak Bastan <babak...@gmail.com> >>>>>> wrote: >>>>>> > hoe can I get my log mohammad? >>>>>> > >>>>>> > >>>>>> > On Wed, Jun 6, 2012 at 9:36 PM, Mohammad Tariq <donta...@gmail.com> >>>>>> > wrote: >>>>>> >> >>>>>> >> could you post your logs???that would help me in understanding the >>>>>> >> problem properly. >>>>>> >> >>>>>> >> Regards, >>>>>> >> Mohammad Tariq >>>>>> >> >>>>>> >> >>>>>> >> On Thu, Jun 7, 2012 at 1:02 AM, Babak Bastan <babak...@gmail.com> >>>>>> >> wrote: >>>>>> >> > Thank you very much mohamad for your attention.I followed the >>>>>> >> > steps but >>>>>> >> > the >>>>>> >> > error is the same as the last time. >>>>>> >> > and there is my hosts file: >>>>>> >> > >>>>>> >> > 127.0.0.1 localhost >>>>>> >> > #127.0.0.1 ubuntu.ubuntu-domain ubuntu >>>>>> >> > >>>>>> >> > >>>>>> >> > # The following lines are desirable for IPv6 capable hosts >>>>>> >> > >>>>>> >> > #::1 ip6-localhost ip6-loopback >>>>>> >> > #fe00::0 ip6-localnet >>>>>> >> > #ff00::0 ip6-mcastprefix >>>>>> >> > #ff02::1 ip6-allnodes >>>>>> >> > #ff02::2 ip6-allrouters >>>>>> >> > >>>>>> >> > but no effect :( >>>>>> >> > >>>>>> >> > On Wed, Jun 6, 2012 at 8:25 PM, Mohammad Tariq >>>>>> >> > <donta...@gmail.com> >>>>>> >> > wrote: >>>>>> >> >> >>>>>> >> >> also change the permissions of these directories to 777. >>>>>> >> >> >>>>>> >> >> Regards, >>>>>> >> >> Mohammad Tariq >>>>>> >> >> >>>>>> >> >> >>>>>> >> >> On Wed, Jun 6, 2012 at 11:54 PM, Mohammad Tariq >>>>>> >> >> <donta...@gmail.com> >>>>>> >> >> wrote: >>>>>> >> >> > create a directory "/home/username/hdfs" (or at some place of >>>>>> >> >> > your >>>>>> >> >> > choice)..inside this hdfs directory create three sub >>>>>> >> >> > directories - >>>>>> >> >> > name, data, and temp, then follow these steps : >>>>>> >> >> > >>>>>> >> >> > add following properties in your core-site.xml - >>>>>> >> >> > >>>>>> >> >> > <property> >>>>>> >> >> > <name>fs.default.name</name> >>>>>> >> >> > <value>hdfs://localhost:9000/</value> >>>>>> >> >> > </property> >>>>>> >> >> > >>>>>> >> >> > <property> >>>>>> >> >> > <name>hadoop.tmp.dir</name> >>>>>> >> >> > <value>/home/mohammad/hdfs/temp</value> >>>>>> >> >> > </property> >>>>>> >> >> > >>>>>> >> >> > then add following two properties in your hdfs-site.xml - >>>>>> >> >> > >>>>>> >> >> > <property> >>>>>> >> >> > <name>dfs.replication</name> >>>>>> >> >> > <value>1</value> >>>>>> >> >> > </property> >>>>>> >> >> > >>>>>> >> >> > <property> >>>>>> >> >> > <name>dfs.name.dir</name> >>>>>> >> >> > <value>/home/mohammad/hdfs/name</value> >>>>>> >> >> > </property> >>>>>> >> >> > >>>>>> >> >> > <property> >>>>>> >> >> > <name>dfs.data.dir</name> >>>>>> >> >> > <value>/home/mohammad/hdfs/data</value> >>>>>> >> >> > </property> >>>>>> >> >> > >>>>>> >> >> > finally add this property in your mapred-site.xml - >>>>>> >> >> > >>>>>> >> >> > <property> >>>>>> >> >> > <name>mapred.job.tracker</name> >>>>>> >> >> > <value>hdfs://localhost:9001</value> >>>>>> >> >> > </property> >>>>>> >> >> > >>>>>> >> >> > NOTE: you can give any name to these directories of your >>>>>> >> >> > choice, just >>>>>> >> >> > keep in mind you have to give same names as values of >>>>>> >> >> > above specified properties in your configuration >>>>>> >> >> > files. >>>>>> >> >> > (give full path of these directories, not just the name of the >>>>>> >> >> > directory) >>>>>> >> >> > >>>>>> >> >> > After this follow the steps provided in the previous reply. >>>>>> >> >> > >>>>>> >> >> > Regards, >>>>>> >> >> > Mohammad Tariq >>>>>> >> >> > >>>>>> >> >> > >>>>>> >> >> > On Wed, Jun 6, 2012 at 11:42 PM, Babak Bastan >>>>>> >> >> > <babak...@gmail.com> >>>>>> >> >> > wrote: >>>>>> >> >> >> thank's Mohammad >>>>>> >> >> >> >>>>>> >> >> >> with this command: >>>>>> >> >> >> >>>>>> >> >> >> babak@ubuntu:~/Downloads/hadoop/bin$ hadoop namenode -format >>>>>> >> >> >> >>>>>> >> >> >> this is my output: >>>>>> >> >> >> >>>>>> >> >> >> 12/06/06 20:05:20 INFO namenode.NameNode: STARTUP_MSG: >>>>>> >> >> >> /************************************************************ >>>>>> >> >> >> STARTUP_MSG: Starting NameNode >>>>>> >> >> >> STARTUP_MSG: host = ubuntu/127.0.1.1 >>>>>> >> >> >> STARTUP_MSG: args = [-format] >>>>>> >> >> >> STARTUP_MSG: version = 0.20.2 >>>>>> >> >> >> STARTUP_MSG: build = >>>>>> >> >> >> >>>>>> >> >> >> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20 >>>>>> >> >> >> -r >>>>>> >> >> >> 911707; compiled by 'chrisdo' on Fri Feb 19 08:07:34 UTC 2010 >>>>>> >> >> >> ************************************************************/ >>>>>> >> >> >> 12/06/06 20:05:20 INFO namenode.FSNamesystem: >>>>>> >> >> >> >>>>>> >> >> >> >>>>>> >> >> >> fsOwner=babak,babak,adm,dialout,cdrom,plugdev,lpadmin,admin,sambashare >>>>>> >> >> >> 12/06/06 20:05:20 INFO namenode.FSNamesystem: >>>>>> >> >> >> supergroup=supergroup >>>>>> >> >> >> 12/06/06 20:05:20 INFO namenode.FSNamesystem: >>>>>> >> >> >> isPermissionEnabled=true >>>>>> >> >> >> 12/06/06 20:05:20 INFO common.Storage: Image file of size 95 >>>>>> >> >> >> saved >>>>>> >> >> >> in 0 >>>>>> >> >> >> seconds. >>>>>> >> >> >> 12/06/06 20:05:20 INFO common.Storage: Storage directory >>>>>> >> >> >> /tmp/hadoop-babak/dfs/name has been successfully formatted. >>>>>> >> >> >> 12/06/06 20:05:20 INFO namenode.NameNode: SHUTDOWN_MSG: >>>>>> >> >> >> /************************************************************ >>>>>> >> >> >> SHUTDOWN_MSG: Shutting down NameNode at ubuntu/127.0.1.1 >>>>>> >> >> >> ************************************************************/ >>>>>> >> >> >> >>>>>> >> >> >> by this command: >>>>>> >> >> >> >>>>>> >> >> >> babak@ubuntu:~/Downloads/hadoop/bin$ start-dfs.sh >>>>>> >> >> >> >>>>>> >> >> >> this is the out put >>>>>> >> >> >> >>>>>> >> >> >> mkdir: kann Verzeichnis >>>>>> >> >> >> „/home/babak/Downloads/hadoop/bin/../logs“ >>>>>> >> >> >> nicht >>>>>> >> >> >> anlegen: Keine Berechtigung >>>>>> >> >> >> >>>>>> >> >> >> this out put(it's in german and it means no right to make >>>>>> >> >> >> this >>>>>> >> >> >> folder) >>>>>> >> >> >> >>>>>> >> >> >> >>>>>> >> >> >> On Wed, Jun 6, 2012 at 7:59 PM, Mohammad Tariq >>>>>> >> >> >> <donta...@gmail.com> >>>>>> >> >> >> wrote: >>>>>> >> >> >>> >>>>>> >> >> >>> once we are done with the configuration, we need to format >>>>>> >> >> >>> the file >>>>>> >> >> >>> system..use this command to do that- >>>>>> >> >> >>> bin/hadoop namenode -format >>>>>> >> >> >>> >>>>>> >> >> >>> after this, hadoop daemon processes should be started using >>>>>> >> >> >>> following >>>>>> >> >> >>> commands - >>>>>> >> >> >>> bin/start-dfs.sh (it'll start NN & DN) >>>>>> >> >> >>> bin/start-mapred.sh (it'll start JT & TT) >>>>>> >> >> >>> >>>>>> >> >> >>> after this use jps to check if everything is alright or >>>>>> >> >> >>> point your >>>>>> >> >> >>> browser to localhost:50070..if you further find any problem >>>>>> >> >> >>> provide >>>>>> >> >> >>> us >>>>>> >> >> >>> with the error logs..:) >>>>>> >> >> >>> >>>>>> >> >> >>> Regards, >>>>>> >> >> >>> Mohammad Tariq >>>>>> >> >> >>> >>>>>> >> >> >>> >>>>>> >> >> >>> On Wed, Jun 6, 2012 at 11:22 PM, Babak Bastan >>>>>> >> >> >>> <babak...@gmail.com> >>>>>> >> >> >>> wrote: >>>>>> >> >> >>> > were you able to format hdfs properly??? >>>>>> >> >> >>> > I did'nt get your question,Do you mean HADOOP_HOME? or >>>>>> >> >> >>> > where did >>>>>> >> >> >>> > I >>>>>> >> >> >>> > install >>>>>> >> >> >>> > Hadoop? >>>>>> >> >> >>> > >>>>>> >> >> >>> > On Wed, Jun 6, 2012 at 7:49 PM, Mohammad Tariq >>>>>> >> >> >>> > <donta...@gmail.com> >>>>>> >> >> >>> > wrote: >>>>>> >> >> >>> >> >>>>>> >> >> >>> >> if you are getting only this, it means your hadoop is not >>>>>> >> >> >>> >> running..were you able to format hdfs properly??? >>>>>> >> >> >>> >> >>>>>> >> >> >>> >> Regards, >>>>>> >> >> >>> >> Mohammad Tariq >>>>>> >> >> >>> >> >>>>>> >> >> >>> >> >>>>>> >> >> >>> >> On Wed, Jun 6, 2012 at 11:17 PM, Babak Bastan >>>>>> >> >> >>> >> <babak...@gmail.com> >>>>>> >> >> >>> >> wrote: >>>>>> >> >> >>> >> > Hi MohammadmI irun jps in my shel I can see this >>>>>> >> >> >>> >> > result: >>>>>> >> >> >>> >> > 2213 Jps >>>>>> >> >> >>> >> > >>>>>> >> >> >>> >> > >>>>>> >> >> >>> >> > On Wed, Jun 6, 2012 at 7:44 PM, Mohammad Tariq >>>>>> >> >> >>> >> > <donta...@gmail.com> >>>>>> >> >> >>> >> > wrote: >>>>>> >> >> >>> >> >> >>>>>> >> >> >>> >> >> you can also use "jps" command at your shell to see >>>>>> >> >> >>> >> >> whether >>>>>> >> >> >>> >> >> Hadoop >>>>>> >> >> >>> >> >> processes are running or not. >>>>>> >> >> >>> >> >> >>>>>> >> >> >>> >> >> Regards, >>>>>> >> >> >>> >> >> Mohammad Tariq >>>>>> >> >> >>> >> >> >>>>>> >> >> >>> >> >> >>>>>> >> >> >>> >> >> On Wed, Jun 6, 2012 at 11:12 PM, Mohammad Tariq >>>>>> >> >> >>> >> >> <donta...@gmail.com> >>>>>> >> >> >>> >> >> wrote: >>>>>> >> >> >>> >> >> > Hi Babak, >>>>>> >> >> >>> >> >> > >>>>>> >> >> >>> >> >> > You have to type it in you web browser..Hadoop >>>>>> >> >> >>> >> >> > provides us >>>>>> >> >> >>> >> >> > a >>>>>> >> >> >>> >> >> > web >>>>>> >> >> >>> >> >> > GUI >>>>>> >> >> >>> >> >> > that not only allows us to browse through the file >>>>>> >> >> >>> >> >> > system, >>>>>> >> >> >>> >> >> > but >>>>>> >> >> >>> >> >> > to >>>>>> >> >> >>> >> >> > download the files as well..Apart from that it also >>>>>> >> >> >>> >> >> > provides a >>>>>> >> >> >>> >> >> > web >>>>>> >> >> >>> >> >> > GUI >>>>>> >> >> >>> >> >> > that can be used to see the status of Jobtracker and >>>>>> >> >> >>> >> >> > Tasktracker..When >>>>>> >> >> >>> >> >> > you run a Hive or Pig job or a Mapreduce job, you >>>>>> >> >> >>> >> >> > can point >>>>>> >> >> >>> >> >> > your >>>>>> >> >> >>> >> >> > browser to http://localhost:50030 to see the status >>>>>> >> >> >>> >> >> > and >>>>>> >> >> >>> >> >> > logs >>>>>> >> >> >>> >> >> > of >>>>>> >> >> >>> >> >> > your >>>>>> >> >> >>> >> >> > job. >>>>>> >> >> >>> >> >> > >>>>>> >> >> >>> >> >> > Regards, >>>>>> >> >> >>> >> >> > Mohammad Tariq >>>>>> >> >> >>> >> >> > >>>>>> >> >> >>> >> >> > >>>>>> >> >> >>> >> >> > On Wed, Jun 6, 2012 at 8:28 PM, Babak Bastan >>>>>> >> >> >>> >> >> > <babak...@gmail.com> >>>>>> >> >> >>> >> >> > wrote: >>>>>> >> >> >>> >> >> >> Thank you shashwat for the answer, >>>>>> >> >> >>> >> >> >> where should I type http://localhost:50070? >>>>>> >> >> >>> >> >> >> I typed here: hive>http://localhost:50070 but >>>>>> >> >> >>> >> >> >> nothing as >>>>>> >> >> >>> >> >> >> result >>>>>> >> >> >>> >> >> >> >>>>>> >> >> >>> >> >> >> >>>>>> >> >> >>> >> >> >> On Wed, Jun 6, 2012 at 3:32 PM, shashwat shriparv >>>>>> >> >> >>> >> >> >> <dwivedishash...@gmail.com> wrote: >>>>>> >> >> >>> >> >> >>> >>>>>> >> >> >>> >> >> >>> first type http://localhost:50070 whether this is >>>>>> >> >> >>> >> >> >>> opening >>>>>> >> >> >>> >> >> >>> or >>>>>> >> >> >>> >> >> >>> not >>>>>> >> >> >>> >> >> >>> and >>>>>> >> >> >>> >> >> >>> check >>>>>> >> >> >>> >> >> >>> how many nodes are available, check some of the >>>>>> >> >> >>> >> >> >>> hadoop >>>>>> >> >> >>> >> >> >>> shell >>>>>> >> >> >>> >> >> >>> commands >>>>>> >> >> >>> >> >> >>> >>>>>> >> >> >>> >> >> >>> >>>>>> >> >> >>> >> >> >>> >>>>>> >> >> >>> >> >> >>> >>>>>> >> >> >>> >> >> >>> from http://hadoop.apache.org/common/docs/r0.18.3/hdfs_shell.html >>>>>> >> >> >>> >> >> >>> run >>>>>> >> >> >>> >> >> >>> example mapreduce task on hadoop take example from >>>>>> >> >> >>> >> >> >>> here >>>>>> >> >> >>> >> >> >>> >>>>>> >> >> >>> >> >> >>> >>>>>> >> >> >>> >> >> >>> >>>>>> >> >> >>> >> >> >>> >>>>>> >> >> >>> >> >> >>> >>>>>> >> >> >>> >> >> >>> >>>>>> >> >> >>> >> >> >>> : http://www.michael-noll.com/blog/2011/04/09/benchmarking-and-stress-testing-an-hadoop-cluster-with-terasort-testdfsio-nnbench-mrbench/ >>>>>> >> >> >>> >> >> >>> >>>>>> >> >> >>> >> >> >>> if all the above you can do sucessfully means >>>>>> >> >> >>> >> >> >>> hadoop is >>>>>> >> >> >>> >> >> >>> configured >>>>>> >> >> >>> >> >> >>> correctly >>>>>> >> >> >>> >> >> >>> >>>>>> >> >> >>> >> >> >>> Regards >>>>>> >> >> >>> >> >> >>> Shashwat >>>>>> >> >> >>> >> >> >>> >>>>>> >> >> >>> >> >> >>> >>>>>> >> >> >>> >> >> >>> On Wed, Jun 6, 2012 at 1:30 AM, Babak Bastan >>>>>> >> >> >>> >> >> >>> <babak...@gmail.com> >>>>>> >> >> >>> >> >> >>> wrote: >>>>>> >> >> >>> >> >> >>>> >>>>>> >> >> >>> >> >> >>>> no I'm not working on CDH.Is there a way to test >>>>>> >> >> >>> >> >> >>>> if my >>>>>> >> >> >>> >> >> >>>> Hadoop >>>>>> >> >> >>> >> >> >>>> works >>>>>> >> >> >>> >> >> >>>> fine >>>>>> >> >> >>> >> >> >>>> or not? >>>>>> >> >> >>> >> >> >>>> >>>>>> >> >> >>> >> >> >>>> >>>>>> >> >> >>> >> >> >>>> On Tue, Jun 5, 2012 at 9:55 PM, Bejoy KS >>>>>> >> >> >>> >> >> >>>> <bejoy...@yahoo.com> >>>>>> >> >> >>> >> >> >>>> wrote: >>>>>> >> >> >>> >> >> >>>>> >>>>>> >> >> >>> >> >> >>>>> Hi Babak >>>>>> >> >> >>> >> >> >>>>> >>>>>> >> >> >>> >> >> >>>>> You gotta follow those instructions in the apace >>>>>> >> >> >>> >> >> >>>>> site >>>>>> >> >> >>> >> >> >>>>> to >>>>>> >> >> >>> >> >> >>>>> set >>>>>> >> >> >>> >> >> >>>>> up >>>>>> >> >> >>> >> >> >>>>> hadoop >>>>>> >> >> >>> >> >> >>>>> from scratch and ensure that hdfs is working >>>>>> >> >> >>> >> >> >>>>> first. You >>>>>> >> >> >>> >> >> >>>>> should >>>>>> >> >> >>> >> >> >>>>> be >>>>>> >> >> >>> >> >> >>>>> able to >>>>>> >> >> >>> >> >> >>>>> read and write files to hdfs before you do your >>>>>> >> >> >>> >> >> >>>>> next >>>>>> >> >> >>> >> >> >>>>> steps. >>>>>> >> >> >>> >> >> >>>>> >>>>>> >> >> >>> >> >> >>>>> Are you on CDH or apache distribution of hadoop? >>>>>> >> >> >>> >> >> >>>>> If it >>>>>> >> >> >>> >> >> >>>>> is >>>>>> >> >> >>> >> >> >>>>> CDH >>>>>> >> >> >>> >> >> >>>>> there >>>>>> >> >> >>> >> >> >>>>> are >>>>>> >> >> >>> >> >> >>>>> detailed instructions on Cloudera web site. >>>>>> >> >> >>> >> >> >>>>> >>>>>> >> >> >>> >> >> >>>>> Regards >>>>>> >> >> >>> >> >> >>>>> Bejoy KS >>>>>> >> >> >>> >> >> >>>>> >>>>>> >> >> >>> >> >> >>>>> Sent from handheld, please excuse typos. >>>>>> >> >> >>> >> >> >>>>> ________________________________ >>>>>> >> >> >>> >> >> >>>>> From: Babak Bastan <babak...@gmail.com> >>>>>> >> >> >>> >> >> >>>>> Date: Tue, 5 Jun 2012 21:30:22 +0200 >>>>>> >> >> >>> >> >> >>>>> To: <user@hive.apache.org> >>>>>> >> >> >>> >> >> >>>>> ReplyTo: user@hive.apache.org >>>>>> >> >> >>> >> >> >>>>> Subject: Re: Error while Creating Table in Hive >>>>>> >> >> >>> >> >> >>>>> >>>>>> >> >> >>> >> >> >>>>> @Bejoy: I set the fs.default.name in the >>>>>> >> >> >>> >> >> >>>>> core-site.xml >>>>>> >> >> >>> >> >> >>>>> and >>>>>> >> >> >>> >> >> >>>>> I >>>>>> >> >> >>> >> >> >>>>> did >>>>>> >> >> >>> >> >> >>>>> all >>>>>> >> >> >>> >> >> >>>>> of >>>>>> >> >> >>> >> >> >>>>> thing that was mentioned in the reference but no >>>>>> >> >> >>> >> >> >>>>> effect >>>>>> >> >> >>> >> >> >>>>> >>>>>> >> >> >>> >> >> >>>>> On Tue, Jun 5, 2012 at 8:43 PM, Babak Bastan >>>>>> >> >> >>> >> >> >>>>> <babak...@gmail.com> >>>>>> >> >> >>> >> >> >>>>> wrote: >>>>>> >> >> >>> >> >> >>>>>> >>>>>> >> >> >>> >> >> >>>>>> Ok sorry but that was my Mistake .I thought it >>>>>> >> >> >>> >> >> >>>>>> works >>>>>> >> >> >>> >> >> >>>>>> but >>>>>> >> >> >>> >> >> >>>>>> no. >>>>>> >> >> >>> >> >> >>>>>> I wrote the command without ; and then I think >>>>>> >> >> >>> >> >> >>>>>> It >>>>>> >> >> >>> >> >> >>>>>> works >>>>>> >> >> >>> >> >> >>>>>> but >>>>>> >> >> >>> >> >> >>>>>> with >>>>>> >> >> >>> >> >> >>>>>> ; >>>>>> >> >> >>> >> >> >>>>>> at >>>>>> >> >> >>> >> >> >>>>>> the end of command >>>>>> >> >> >>> >> >> >>>>>> >>>>>> >> >> >>> >> >> >>>>>> CREATE TABLE pokes (foo INT, bar STRING); >>>>>> >> >> >>> >> >> >>>>>> >>>>>> >> >> >>> >> >> >>>>>> does'nt work >>>>>> >> >> >>> >> >> >>>>>> >>>>>> >> >> >>> >> >> >>>>>> >>>>>> >> >> >>> >> >> >>>>>> On Tue, Jun 5, 2012 at 8:34 PM, shashwat >>>>>> >> >> >>> >> >> >>>>>> shriparv >>>>>> >> >> >>> >> >> >>>>>> <dwivedishash...@gmail.com> wrote: >>>>>> >> >> >>> >> >> >>>>>>> >>>>>> >> >> >>> >> >> >>>>>>> inside configuration. all properties will be >>>>>> >> >> >>> >> >> >>>>>>> inside >>>>>> >> >> >>> >> >> >>>>>>> the >>>>>> >> >> >>> >> >> >>>>>>> configuration >>>>>> >> >> >>> >> >> >>>>>>> tags >>>>>> >> >> >>> >> >> >>>>>>> >>>>>> >> >> >>> >> >> >>>>>>> >>>>>> >> >> >>> >> >> >>>>>>> On Tue, Jun 5, 2012 at 11:53 PM, Babak Bastan >>>>>> >> >> >>> >> >> >>>>>>> <babak...@gmail.com> >>>>>> >> >> >>> >> >> >>>>>>> wrote: >>>>>> >> >> >>> >> >> >>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>> Thank you so much my friend your idee works >>>>>> >> >> >>> >> >> >>>>>>>> fine(no >>>>>> >> >> >>> >> >> >>>>>>>> error) >>>>>> >> >> >>> >> >> >>>>>>>> you >>>>>> >> >> >>> >> >> >>>>>>>> are >>>>>> >> >> >>> >> >> >>>>>>>> the best :) >>>>>> >> >> >>> >> >> >>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>> On Tue, Jun 5, 2012 at 8:20 PM, Babak Bastan >>>>>> >> >> >>> >> >> >>>>>>>> <babak...@gmail.com> >>>>>> >> >> >>> >> >> >>>>>>>> wrote: >>>>>> >> >> >>> >> >> >>>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>>> It must be inside the >>>>>> >> >> >>> >> >> >>>>>>>>> <configuration></configuration> >>>>>> >> >> >>> >> >> >>>>>>>>> or >>>>>> >> >> >>> >> >> >>>>>>>>> outside >>>>>> >> >> >>> >> >> >>>>>>>>> this? >>>>>> >> >> >>> >> >> >>>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>>> On Tue, Jun 5, 2012 at 8:15 PM, shashwat >>>>>> >> >> >>> >> >> >>>>>>>>> shriparv >>>>>> >> >> >>> >> >> >>>>>>>>> <dwivedishash...@gmail.com> wrote: >>>>>> >> >> >>> >> >> >>>>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>>>> It will be inside hive/conf >>>>>> >> >> >>> >> >> >>>>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>>>> On Tue, Jun 5, 2012 at 11:43 PM, Babak >>>>>> >> >> >>> >> >> >>>>>>>>>> Bastan >>>>>> >> >> >>> >> >> >>>>>>>>>> <babak...@gmail.com> >>>>>> >> >> >>> >> >> >>>>>>>>>> wrote: >>>>>> >> >> >>> >> >> >>>>>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>>>>> Thanks sShashwat, and where is this >>>>>> >> >> >>> >> >> >>>>>>>>>>> hive-site.xml >>>>>> >> >> >>> >> >> >>>>>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>>>>> On Tue, Jun 5, 2012 at 8:02 PM, shashwat >>>>>> >> >> >>> >> >> >>>>>>>>>>> shriparv >>>>>> >> >> >>> >> >> >>>>>>>>>>> <dwivedishash...@gmail.com> wrote: >>>>>> >> >> >>> >> >> >>>>>>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> set >>>>>> >> >> >>> >> >> >>>>>>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> hive.metastore.warehouse.dir in >>>>>> >> >> >>> >> >> >>>>>>>>>>>> hive-site.xml >>>>>> >> >> >>> >> >> >>>>>>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> <property> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> <name>hive.metastore.local</name> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> <value>true</value> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> </property> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> <name>hive.metastore.warehouse.dir</name> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> <value>/home/<your >>>>>> >> >> >>> >> >> >>>>>>>>>>>> username>/hivefolder</value> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> <description>location of >>>>>> >> >> >>> >> >> >>>>>>>>>>>> default >>>>>> >> >> >>> >> >> >>>>>>>>>>>> database >>>>>> >> >> >>> >> >> >>>>>>>>>>>> for >>>>>> >> >> >>> >> >> >>>>>>>>>>>> the >>>>>> >> >> >>> >> >> >>>>>>>>>>>> warehouse</description> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> </property> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> On Tue, Jun 5, 2012 at 10:43 PM, Babak >>>>>> >> >> >>> >> >> >>>>>>>>>>>> Bastan >>>>>> >> >> >>> >> >> >>>>>>>>>>>> <babak...@gmail.com> wrote: >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> Hello Experts , >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> I'm new in Hive .When try to create a >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> test >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> Table >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> in >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> Hive >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> I >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> get >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> an error.I want to run this command: >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> CREATE TABLE Test (DateT STRING, Url >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> STRING, >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> Content >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> STRING); >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> but this error occured: >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> FAILED: Error in metadata: >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> MetaException(message:Got >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> exception: >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> java.io.FileNotFoundException File >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> file:/user/hive/warehouse/test does not >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> exist.) >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> FAILED: Execution Error, return code 1 >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> from >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> org.apache.hadoop.hive.ql.exec.DDLTask >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> How can I solve this Problem? >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> Thank you so much >>>>>> >> >> >>> >> >> >>>>>>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> -- >>>>>> >> >> >>> >> >> >>>>>>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> ∞ >>>>>> >> >> >>> >> >> >>>>>>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> Shashwat Shriparv >>>>>> >> >> >>> >> >> >>>>>>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>>>> -- >>>>>> >> >> >>> >> >> >>>>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>>>> ∞ >>>>>> >> >> >>> >> >> >>>>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>>>> Shashwat Shriparv >>>>>> >> >> >>> >> >> >>>>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>>> >>>>>> >> >> >>> >> >> >>>>>>> >>>>>> >> >> >>> >> >> >>>>>>> >>>>>> >> >> >>> >> >> >>>>>>> >>>>>> >> >> >>> >> >> >>>>>>> -- >>>>>> >> >> >>> >> >> >>>>>>> >>>>>> >> >> >>> >> >> >>>>>>> >>>>>> >> >> >>> >> >> >>>>>>> ∞ >>>>>> >> >> >>> >> >> >>>>>>> >>>>>> >> >> >>> >> >> >>>>>>> Shashwat Shriparv >>>>>> >> >> >>> >> >> >>>>>>> >>>>>> >> >> >>> >> >> >>>>>>> >>>>>> >> >> >>> >> >> >>>>>> >>>>>> >> >> >>> >> >> >>>>> >>>>>> >> >> >>> >> >> >>>> >>>>>> >> >> >>> >> >> >>> >>>>>> >> >> >>> >> >> >>> >>>>>> >> >> >>> >> >> >>> >>>>>> >> >> >>> >> >> >>> -- >>>>>> >> >> >>> >> >> >>> >>>>>> >> >> >>> >> >> >>> >>>>>> >> >> >>> >> >> >>> ∞ >>>>>> >> >> >>> >> >> >>> >>>>>> >> >> >>> >> >> >>> Shashwat Shriparv >>>>>> >> >> >>> >> >> >>> >>>>>> >> >> >>> >> >> >>> >>>>>> >> >> >>> >> >> >> >>>>>> >> >> >>> >> > >>>>>> >> >> >>> >> > >>>>>> >> >> >>> > >>>>>> >> >> >>> > >>>>>> >> >> >> >>>>>> >> >> >> >>>>>> >> > >>>>>> >> > >>>>>> > >>>>>> > >>>>> >>>>> >>>> >>>> >>>> >>>> -- >>>> >>>> >>>> ∞ >>>> >>>> Shashwat Shriparv >>>> >>>> >>> >> >