follow this link - http://hadoop.apache.org/common/docs/r0.20.203.0/single_node_setup.html..it worked for most of us without any prooblem.
do all the things required to configure hadoop on linux in pseudo distributed mode as given in this link..start with a simple setup as shown there..then we'll add more properties if you need detailed help you can also visit - http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-single-node-cluster/ Regards, Mohammad Tariq On Thu, Jun 7, 2012 at 3:04 AM, Babak Bastan <babak...@gmail.com> wrote: > I try to install another one in blog,most of them was performed without > problem but in this step > > udo mkdir /hadoop && sudo chown hdfs:hdfs /hadoop && sudo chmod 777 /hadoop > > I get this error: > > Error: 0: couldn't open source file </hadoop.ui> > > and in this step: > mkdir /usr/lib/hadoop-0.20/.ssh > this error: > mkdir: kann Verzeichnis „/usr/lib/hadoop-0.20/.ssh“ nicht anlegen: Keine > Berechtigung > ---> no permission to make a directory > > On Wed, Jun 6, 2012 at 11:21 PM, Mohammad Tariq <donta...@gmail.com> wrote: >> >> ok..we'll give it a final shot..then i'll email configured hadoop to >> your email address..delete the hdfs directory which contains tmp, data >> and name..recreate it..format hdfs again and then start the processes. >> >> Regards, >> Mohammad Tariq >> >> >> On Thu, Jun 7, 2012 at 2:22 AM, Babak Bastan <babak...@gmail.com> wrote: >> > I 've performed the steps but the same error in this step as before: >> > bin/start-dfs.sh >> > It is about my permission to make directory >> > >> > On Wed, Jun 6, 2012 at 10:33 PM, Mohammad Tariq <donta...@gmail.com> >> > wrote: >> >> >> >> actually this blog post explains how to install cloudera's hadoop >> >> distribution...if you have followed this post and installed cloudera's >> >> distribution then your logs should ideally be inside >> >> /usr/lib/hadoop/logs (if everything was fine)..anyway try the steps I >> >> have given and let me know. >> >> >> >> Regards, >> >> Mohammad Tariq >> >> >> >> >> >> On Thu, Jun 7, 2012 at 1:52 AM, Babak Bastan <babak...@gmail.com> >> >> wrote: >> >> > by the way,you are a very nice man my friend:Thank you so much :) >> >> > >> >> > what do you mean aboat this post in stackoverflow? >> >> > >> >> > I am assuming that is your first installation of hadoop. >> >> > >> >> > At the beginning please check if your daemons are working. To do that >> >> > use >> >> > (in terminal): >> >> > >> >> > jps >> >> > >> >> > If only jps appears that means all daemons are down. Please check the >> >> > log >> >> > files. Especially the namenode. Log folder is probably somewhere >> >> > there >> >> > /usr/lib/hadoop/logs >> >> > >> >> > If you have some permission problems. Use this guide during the >> >> > installation. >> >> > >> >> > Good installation guide >> >> > >> >> > I am shooting with this explanations but these are most common >> >> > problems. >> >> > >> >> > >> >> > On Wed, Jun 6, 2012 at 10:15 PM, Babak Bastan <babak...@gmail.com> >> >> > wrote: >> >> >> >> >> >> I checked it but no hadoop folder :( >> >> >> yes you are right.I'm a student and I want to make a very very >> >> >> simple >> >> >> programm hive but untill now hmmmmmmmmm >> >> >> >> >> >> >> >> >> On Wed, Jun 6, 2012 at 10:12 PM, Babak Bastan <babak...@gmail.com> >> >> >> wrote: >> >> >>> >> >> >>> no one error: >> >> >>> i.e if I run this one >> >> >>> >> >> >>> hostname --fqdn >> >> >>> >> >> >>> with the condition that I send to you : >> >> >>> >> >> >>> 127.0.0.1 localhost >> >> >>> #127.0.0.1 ubuntu.ubuntu-domain ubuntu >> >> >>> # The following lines are desirable for IPv6 capable hosts >> >> >>> #::1 ip6-localhost ip6-loopback >> >> >>> #fe00::0 ip6-localnet >> >> >>> #ff00::0 ip6-mcastprefix >> >> >>> #ff02::1 ip6-allnodes >> >> >>> #ff02::2 ip6-allrouters >> >> >>> >> >> >>> I get this error: >> >> >>> >> >> >>> hostname: Name or service not known >> >> >>> >> >> >>> Or in the second step by this command: >> >> >>> >> >> >>> babak@ubuntu:~/Downloads/hadoop/bin$ start-hdfs.sh >> >> >>> >> >> >>> these lines of error: >> >> >>> >> >> >>> >> >> >>> mkdir: kann Verzeichnis „/home/babak/Downloads/hadoop/bin/../logs“ >> >> >>> nicht >> >> >>> anlegen: Keine Berechtigung >> >> >>> starting namenode, logging to >> >> >>> >> >> >>> >> >> >>> /home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-namenode-ubuntu.out >> >> >>> /home/babak/Downloads/hadoop/bin/hadoop-daemon.sh: Zeile 117: >> >> >>> >> >> >>> >> >> >>> /home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-namenode-ubuntu.out: >> >> >>> Datei oder Verzeichnis nicht gefunden >> >> >>> head: >> >> >>> >> >> >>> >> >> >>> „/home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-namenode-ubuntu.out“ >> >> >>> kann nicht zum Lesen geöffnet werden: Datei oder Verzeichnis nicht >> >> >>> gefunden >> >> >>> localhost: mkdir: kann Verzeichnis >> >> >>> „/home/babak/Downloads/hadoop/bin/../logs“ nicht anlegen: Keine >> >> >>> Berechtigung >> >> >>> localhost: starting datanode, logging to >> >> >>> >> >> >>> >> >> >>> /home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-datanode-ubuntu.out >> >> >>> localhost: /home/babak/Downloads/hadoop/bin/hadoop-daemon.sh: Zeile >> >> >>> 117: >> >> >>> >> >> >>> >> >> >>> /home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-datanode-ubuntu.out: >> >> >>> Datei oder Verzeichnis nicht gefunden >> >> >>> localhost: head: >> >> >>> >> >> >>> >> >> >>> „/home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-datanode-ubuntu.out“ >> >> >>> kann nicht zum Lesen geöffnet werden: Datei oder Verzeichnis nicht >> >> >>> gefunden >> >> >>> localhost: mkdir: kann Verzeichnis >> >> >>> „/home/babak/Downloads/hadoop/bin/../logs“ nicht anlegen: Keine >> >> >>> Berechtigung >> >> >>> localhost: starting secondarynamenode, logging to >> >> >>> >> >> >>> >> >> >>> /home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-secondarynamenode-ubuntu.out >> >> >>> localhost: /home/babak/Downloads/hadoop/bin/hadoop-daemon.sh: Zeile >> >> >>> 117: >> >> >>> >> >> >>> >> >> >>> /home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-secondarynamenode-ubuntu.out: >> >> >>> Datei oder Verzeichnis nicht gefunden >> >> >>> localhost: head: >> >> >>> >> >> >>> >> >> >>> „/home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-secondarynamenode-ubuntu.out“ >> >> >>> kann nicht zum Lesen geöffnet werden: Datei oder Verzeichnis nicht >> >> >>> gefunden >> >> >>> >> >> >>> they said no permision to make logs in this >> >> >>> path:/home/babak/Downloads/hadoop/bin/../logs >> >> >>> >> >> >>> and generally I cant create a table in hive and get this one: >> >> >>> >> >> >>> FAILED: Error in metadata: MetaException(message:Got exception: >> >> >>> java.io.FileNotFoundException File file:/user/hive/warehouse/test >> >> >>> does >> >> >>> not >> >> >>> exist.) >> >> >>> FAILED: Execution Error, return code 1 from >> >> >>> org.apache.hadoop.hive.ql.exec.DDLTask >> >> >>> >> >> >>> On Wed, Jun 6, 2012 at 10:02 PM, shashwat shriparv >> >> >>> <dwivedishash...@gmail.com> wrote: >> >> >>>> >> >> >>>> whats the error babak ??? >> >> >>>> >> >> >>>> >> >> >>>> On Thu, Jun 7, 2012 at 1:25 AM, Babak Bastan <babak...@gmail.com> >> >> >>>> wrote: >> >> >>>>> >> >> >>>>> What the hell is that?I see no log folder there >> >> >>>>> >> >> >>>>> >> >> >>>>> On Wed, Jun 6, 2012 at 9:41 PM, Mohammad Tariq >> >> >>>>> <donta...@gmail.com> >> >> >>>>> wrote: >> >> >>>>>> >> >> >>>>>> go to your HADOOP_HOME i.e your hadoop directory(that includes >> >> >>>>>> bin, >> >> >>>>>> conf etc)..you can find logs directory there.. >> >> >>>>>> >> >> >>>>>> Regards, >> >> >>>>>> Mohammad Tariq >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> On Thu, Jun 7, 2012 at 1:09 AM, Babak Bastan >> >> >>>>>> <babak...@gmail.com> >> >> >>>>>> wrote: >> >> >>>>>> > hoe can I get my log mohammad? >> >> >>>>>> > >> >> >>>>>> > >> >> >>>>>> > On Wed, Jun 6, 2012 at 9:36 PM, Mohammad Tariq >> >> >>>>>> > <donta...@gmail.com> >> >> >>>>>> > wrote: >> >> >>>>>> >> >> >> >>>>>> >> could you post your logs???that would help me in >> >> >>>>>> >> understanding >> >> >>>>>> >> the >> >> >>>>>> >> problem properly. >> >> >>>>>> >> >> >> >>>>>> >> Regards, >> >> >>>>>> >> Mohammad Tariq >> >> >>>>>> >> >> >> >>>>>> >> >> >> >>>>>> >> On Thu, Jun 7, 2012 at 1:02 AM, Babak Bastan >> >> >>>>>> >> <babak...@gmail.com> >> >> >>>>>> >> wrote: >> >> >>>>>> >> > Thank you very much mohamad for your attention.I followed >> >> >>>>>> >> > the >> >> >>>>>> >> > steps but >> >> >>>>>> >> > the >> >> >>>>>> >> > error is the same as the last time. >> >> >>>>>> >> > and there is my hosts file: >> >> >>>>>> >> > >> >> >>>>>> >> > 127.0.0.1 localhost >> >> >>>>>> >> > #127.0.0.1 ubuntu.ubuntu-domain ubuntu >> >> >>>>>> >> > >> >> >>>>>> >> > >> >> >>>>>> >> > # The following lines are desirable for IPv6 capable hosts >> >> >>>>>> >> > >> >> >>>>>> >> > #::1 ip6-localhost ip6-loopback >> >> >>>>>> >> > #fe00::0 ip6-localnet >> >> >>>>>> >> > #ff00::0 ip6-mcastprefix >> >> >>>>>> >> > #ff02::1 ip6-allnodes >> >> >>>>>> >> > #ff02::2 ip6-allrouters >> >> >>>>>> >> > >> >> >>>>>> >> > but no effect :( >> >> >>>>>> >> > >> >> >>>>>> >> > On Wed, Jun 6, 2012 at 8:25 PM, Mohammad Tariq >> >> >>>>>> >> > <donta...@gmail.com> >> >> >>>>>> >> > wrote: >> >> >>>>>> >> >> >> >> >>>>>> >> >> also change the permissions of these directories to 777. >> >> >>>>>> >> >> >> >> >>>>>> >> >> Regards, >> >> >>>>>> >> >> Mohammad Tariq >> >> >>>>>> >> >> >> >> >>>>>> >> >> >> >> >>>>>> >> >> On Wed, Jun 6, 2012 at 11:54 PM, Mohammad Tariq >> >> >>>>>> >> >> <donta...@gmail.com> >> >> >>>>>> >> >> wrote: >> >> >>>>>> >> >> > create a directory "/home/username/hdfs" (or at some >> >> >>>>>> >> >> > place >> >> >>>>>> >> >> > of >> >> >>>>>> >> >> > your >> >> >>>>>> >> >> > choice)..inside this hdfs directory create three sub >> >> >>>>>> >> >> > directories - >> >> >>>>>> >> >> > name, data, and temp, then follow these steps : >> >> >>>>>> >> >> > >> >> >>>>>> >> >> > add following properties in your core-site.xml - >> >> >>>>>> >> >> > >> >> >>>>>> >> >> > <property> >> >> >>>>>> >> >> > <name>fs.default.name</name> >> >> >>>>>> >> >> > <value>hdfs://localhost:9000/</value> >> >> >>>>>> >> >> > </property> >> >> >>>>>> >> >> > >> >> >>>>>> >> >> > <property> >> >> >>>>>> >> >> > <name>hadoop.tmp.dir</name> >> >> >>>>>> >> >> > <value>/home/mohammad/hdfs/temp</value> >> >> >>>>>> >> >> > </property> >> >> >>>>>> >> >> > >> >> >>>>>> >> >> > then add following two properties in your hdfs-site.xml >> >> >>>>>> >> >> > - >> >> >>>>>> >> >> > >> >> >>>>>> >> >> > <property> >> >> >>>>>> >> >> > <name>dfs.replication</name> >> >> >>>>>> >> >> > <value>1</value> >> >> >>>>>> >> >> > </property> >> >> >>>>>> >> >> > >> >> >>>>>> >> >> > <property> >> >> >>>>>> >> >> > <name>dfs.name.dir</name> >> >> >>>>>> >> >> > <value>/home/mohammad/hdfs/name</value> >> >> >>>>>> >> >> > </property> >> >> >>>>>> >> >> > >> >> >>>>>> >> >> > <property> >> >> >>>>>> >> >> > <name>dfs.data.dir</name> >> >> >>>>>> >> >> > <value>/home/mohammad/hdfs/data</value> >> >> >>>>>> >> >> > </property> >> >> >>>>>> >> >> > >> >> >>>>>> >> >> > finally add this property in your mapred-site.xml - >> >> >>>>>> >> >> > >> >> >>>>>> >> >> > <property> >> >> >>>>>> >> >> > <name>mapred.job.tracker</name> >> >> >>>>>> >> >> > <value>hdfs://localhost:9001</value> >> >> >>>>>> >> >> > </property> >> >> >>>>>> >> >> > >> >> >>>>>> >> >> > NOTE: you can give any name to these directories of your >> >> >>>>>> >> >> > choice, just >> >> >>>>>> >> >> > keep in mind you have to give same names as values of >> >> >>>>>> >> >> > above specified properties in your >> >> >>>>>> >> >> > configuration >> >> >>>>>> >> >> > files. >> >> >>>>>> >> >> > (give full path of these directories, not just the name >> >> >>>>>> >> >> > of >> >> >>>>>> >> >> > the >> >> >>>>>> >> >> > directory) >> >> >>>>>> >> >> > >> >> >>>>>> >> >> > After this follow the steps provided in the previous >> >> >>>>>> >> >> > reply. >> >> >>>>>> >> >> > >> >> >>>>>> >> >> > Regards, >> >> >>>>>> >> >> > Mohammad Tariq >> >> >>>>>> >> >> > >> >> >>>>>> >> >> > >> >> >>>>>> >> >> > On Wed, Jun 6, 2012 at 11:42 PM, Babak Bastan >> >> >>>>>> >> >> > <babak...@gmail.com> >> >> >>>>>> >> >> > wrote: >> >> >>>>>> >> >> >> thank's Mohammad >> >> >>>>>> >> >> >> >> >> >>>>>> >> >> >> with this command: >> >> >>>>>> >> >> >> >> >> >>>>>> >> >> >> babak@ubuntu:~/Downloads/hadoop/bin$ hadoop namenode >> >> >>>>>> >> >> >> -format >> >> >>>>>> >> >> >> >> >> >>>>>> >> >> >> this is my output: >> >> >>>>>> >> >> >> >> >> >>>>>> >> >> >> 12/06/06 20:05:20 INFO namenode.NameNode: STARTUP_MSG: >> >> >>>>>> >> >> >> >> >> >>>>>> >> >> >> >> >> >>>>>> >> >> >> /************************************************************ >> >> >>>>>> >> >> >> STARTUP_MSG: Starting NameNode >> >> >>>>>> >> >> >> STARTUP_MSG: host = ubuntu/127.0.1.1 >> >> >>>>>> >> >> >> STARTUP_MSG: args = [-format] >> >> >>>>>> >> >> >> STARTUP_MSG: version = 0.20.2 >> >> >>>>>> >> >> >> STARTUP_MSG: build = >> >> >>>>>> >> >> >> >> >> >>>>>> >> >> >> >> >> >>>>>> >> >> >> >> >> >>>>>> >> >> >> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20 >> >> >>>>>> >> >> >> -r >> >> >>>>>> >> >> >> 911707; compiled by 'chrisdo' on Fri Feb 19 08:07:34 >> >> >>>>>> >> >> >> UTC >> >> >>>>>> >> >> >> 2010 >> >> >>>>>> >> >> >> >> >> >>>>>> >> >> >> >> >> >>>>>> >> >> >> ************************************************************/ >> >> >>>>>> >> >> >> 12/06/06 20:05:20 INFO namenode.FSNamesystem: >> >> >>>>>> >> >> >> >> >> >>>>>> >> >> >> >> >> >>>>>> >> >> >> >> >> >>>>>> >> >> >> >> >> >>>>>> >> >> >> fsOwner=babak,babak,adm,dialout,cdrom,plugdev,lpadmin,admin,sambashare >> >> >>>>>> >> >> >> 12/06/06 20:05:20 INFO namenode.FSNamesystem: >> >> >>>>>> >> >> >> supergroup=supergroup >> >> >>>>>> >> >> >> 12/06/06 20:05:20 INFO namenode.FSNamesystem: >> >> >>>>>> >> >> >> isPermissionEnabled=true >> >> >>>>>> >> >> >> 12/06/06 20:05:20 INFO common.Storage: Image file of >> >> >>>>>> >> >> >> size >> >> >>>>>> >> >> >> 95 >> >> >>>>>> >> >> >> saved >> >> >>>>>> >> >> >> in 0 >> >> >>>>>> >> >> >> seconds. >> >> >>>>>> >> >> >> 12/06/06 20:05:20 INFO common.Storage: Storage >> >> >>>>>> >> >> >> directory >> >> >>>>>> >> >> >> /tmp/hadoop-babak/dfs/name has been successfully >> >> >>>>>> >> >> >> formatted. >> >> >>>>>> >> >> >> 12/06/06 20:05:20 INFO namenode.NameNode: SHUTDOWN_MSG: >> >> >>>>>> >> >> >> >> >> >>>>>> >> >> >> >> >> >>>>>> >> >> >> /************************************************************ >> >> >>>>>> >> >> >> SHUTDOWN_MSG: Shutting down NameNode at >> >> >>>>>> >> >> >> ubuntu/127.0.1.1 >> >> >>>>>> >> >> >> >> >> >>>>>> >> >> >> >> >> >>>>>> >> >> >> ************************************************************/ >> >> >>>>>> >> >> >> >> >> >>>>>> >> >> >> by this command: >> >> >>>>>> >> >> >> >> >> >>>>>> >> >> >> babak@ubuntu:~/Downloads/hadoop/bin$ start-dfs.sh >> >> >>>>>> >> >> >> >> >> >>>>>> >> >> >> this is the out put >> >> >>>>>> >> >> >> >> >> >>>>>> >> >> >> mkdir: kann Verzeichnis >> >> >>>>>> >> >> >> „/home/babak/Downloads/hadoop/bin/../logs“ >> >> >>>>>> >> >> >> nicht >> >> >>>>>> >> >> >> anlegen: Keine Berechtigung >> >> >>>>>> >> >> >> >> >> >>>>>> >> >> >> this out put(it's in german and it means no right to >> >> >>>>>> >> >> >> make >> >> >>>>>> >> >> >> this >> >> >>>>>> >> >> >> folder) >> >> >>>>>> >> >> >> >> >> >>>>>> >> >> >> >> >> >>>>>> >> >> >> On Wed, Jun 6, 2012 at 7:59 PM, Mohammad Tariq >> >> >>>>>> >> >> >> <donta...@gmail.com> >> >> >>>>>> >> >> >> wrote: >> >> >>>>>> >> >> >>> >> >> >>>>>> >> >> >>> once we are done with the configuration, we need to >> >> >>>>>> >> >> >>> format >> >> >>>>>> >> >> >>> the file >> >> >>>>>> >> >> >>> system..use this command to do that- >> >> >>>>>> >> >> >>> bin/hadoop namenode -format >> >> >>>>>> >> >> >>> >> >> >>>>>> >> >> >>> after this, hadoop daemon processes should be started >> >> >>>>>> >> >> >>> using >> >> >>>>>> >> >> >>> following >> >> >>>>>> >> >> >>> commands - >> >> >>>>>> >> >> >>> bin/start-dfs.sh (it'll start NN & DN) >> >> >>>>>> >> >> >>> bin/start-mapred.sh (it'll start JT & TT) >> >> >>>>>> >> >> >>> >> >> >>>>>> >> >> >>> after this use jps to check if everything is alright >> >> >>>>>> >> >> >>> or >> >> >>>>>> >> >> >>> point your >> >> >>>>>> >> >> >>> browser to localhost:50070..if you further find any >> >> >>>>>> >> >> >>> problem >> >> >>>>>> >> >> >>> provide >> >> >>>>>> >> >> >>> us >> >> >>>>>> >> >> >>> with the error logs..:) >> >> >>>>>> >> >> >>> >> >> >>>>>> >> >> >>> Regards, >> >> >>>>>> >> >> >>> Mohammad Tariq >> >> >>>>>> >> >> >>> >> >> >>>>>> >> >> >>> >> >> >>>>>> >> >> >>> On Wed, Jun 6, 2012 at 11:22 PM, Babak Bastan >> >> >>>>>> >> >> >>> <babak...@gmail.com> >> >> >>>>>> >> >> >>> wrote: >> >> >>>>>> >> >> >>> > were you able to format hdfs properly??? >> >> >>>>>> >> >> >>> > I did'nt get your question,Do you mean HADOOP_HOME? >> >> >>>>>> >> >> >>> > or >> >> >>>>>> >> >> >>> > where did >> >> >>>>>> >> >> >>> > I >> >> >>>>>> >> >> >>> > install >> >> >>>>>> >> >> >>> > Hadoop? >> >> >>>>>> >> >> >>> > >> >> >>>>>> >> >> >>> > On Wed, Jun 6, 2012 at 7:49 PM, Mohammad Tariq >> >> >>>>>> >> >> >>> > <donta...@gmail.com> >> >> >>>>>> >> >> >>> > wrote: >> >> >>>>>> >> >> >>> >> >> >> >>>>>> >> >> >>> >> if you are getting only this, it means your hadoop >> >> >>>>>> >> >> >>> >> is >> >> >>>>>> >> >> >>> >> not >> >> >>>>>> >> >> >>> >> running..were you able to format hdfs properly??? >> >> >>>>>> >> >> >>> >> >> >> >>>>>> >> >> >>> >> Regards, >> >> >>>>>> >> >> >>> >> Mohammad Tariq >> >> >>>>>> >> >> >>> >> >> >> >>>>>> >> >> >>> >> >> >> >>>>>> >> >> >>> >> On Wed, Jun 6, 2012 at 11:17 PM, Babak Bastan >> >> >>>>>> >> >> >>> >> <babak...@gmail.com> >> >> >>>>>> >> >> >>> >> wrote: >> >> >>>>>> >> >> >>> >> > Hi MohammadmI irun jps in my shel I can see this >> >> >>>>>> >> >> >>> >> > result: >> >> >>>>>> >> >> >>> >> > 2213 Jps >> >> >>>>>> >> >> >>> >> > >> >> >>>>>> >> >> >>> >> > >> >> >>>>>> >> >> >>> >> > On Wed, Jun 6, 2012 at 7:44 PM, Mohammad Tariq >> >> >>>>>> >> >> >>> >> > <donta...@gmail.com> >> >> >>>>>> >> >> >>> >> > wrote: >> >> >>>>>> >> >> >>> >> >> >> >> >>>>>> >> >> >>> >> >> you can also use "jps" command at your shell to >> >> >>>>>> >> >> >>> >> >> see >> >> >>>>>> >> >> >>> >> >> whether >> >> >>>>>> >> >> >>> >> >> Hadoop >> >> >>>>>> >> >> >>> >> >> processes are running or not. >> >> >>>>>> >> >> >>> >> >> >> >> >>>>>> >> >> >>> >> >> Regards, >> >> >>>>>> >> >> >>> >> >> Mohammad Tariq >> >> >>>>>> >> >> >>> >> >> >> >> >>>>>> >> >> >>> >> >> >> >> >>>>>> >> >> >>> >> >> On Wed, Jun 6, 2012 at 11:12 PM, Mohammad Tariq >> >> >>>>>> >> >> >>> >> >> <donta...@gmail.com> >> >> >>>>>> >> >> >>> >> >> wrote: >> >> >>>>>> >> >> >>> >> >> > Hi Babak, >> >> >>>>>> >> >> >>> >> >> > >> >> >>>>>> >> >> >>> >> >> > You have to type it in you web >> >> >>>>>> >> >> >>> >> >> > browser..Hadoop >> >> >>>>>> >> >> >>> >> >> > provides us >> >> >>>>>> >> >> >>> >> >> > a >> >> >>>>>> >> >> >>> >> >> > web >> >> >>>>>> >> >> >>> >> >> > GUI >> >> >>>>>> >> >> >>> >> >> > that not only allows us to browse through the >> >> >>>>>> >> >> >>> >> >> > file >> >> >>>>>> >> >> >>> >> >> > system, >> >> >>>>>> >> >> >>> >> >> > but >> >> >>>>>> >> >> >>> >> >> > to >> >> >>>>>> >> >> >>> >> >> > download the files as well..Apart from that it >> >> >>>>>> >> >> >>> >> >> > also >> >> >>>>>> >> >> >>> >> >> > provides a >> >> >>>>>> >> >> >>> >> >> > web >> >> >>>>>> >> >> >>> >> >> > GUI >> >> >>>>>> >> >> >>> >> >> > that can be used to see the status of >> >> >>>>>> >> >> >>> >> >> > Jobtracker >> >> >>>>>> >> >> >>> >> >> > and >> >> >>>>>> >> >> >>> >> >> > Tasktracker..When >> >> >>>>>> >> >> >>> >> >> > you run a Hive or Pig job or a Mapreduce job, >> >> >>>>>> >> >> >>> >> >> > you >> >> >>>>>> >> >> >>> >> >> > can point >> >> >>>>>> >> >> >>> >> >> > your >> >> >>>>>> >> >> >>> >> >> > browser to http://localhost:50030 to see the >> >> >>>>>> >> >> >>> >> >> > status >> >> >>>>>> >> >> >>> >> >> > and >> >> >>>>>> >> >> >>> >> >> > logs >> >> >>>>>> >> >> >>> >> >> > of >> >> >>>>>> >> >> >>> >> >> > your >> >> >>>>>> >> >> >>> >> >> > job. >> >> >>>>>> >> >> >>> >> >> > >> >> >>>>>> >> >> >>> >> >> > Regards, >> >> >>>>>> >> >> >>> >> >> > Mohammad Tariq >> >> >>>>>> >> >> >>> >> >> > >> >> >>>>>> >> >> >>> >> >> > >> >> >>>>>> >> >> >>> >> >> > On Wed, Jun 6, 2012 at 8:28 PM, Babak Bastan >> >> >>>>>> >> >> >>> >> >> > <babak...@gmail.com> >> >> >>>>>> >> >> >>> >> >> > wrote: >> >> >>>>>> >> >> >>> >> >> >> Thank you shashwat for the answer, >> >> >>>>>> >> >> >>> >> >> >> where should I type http://localhost:50070? >> >> >>>>>> >> >> >>> >> >> >> I typed here: hive>http://localhost:50070 but >> >> >>>>>> >> >> >>> >> >> >> nothing as >> >> >>>>>> >> >> >>> >> >> >> result >> >> >>>>>> >> >> >>> >> >> >> >> >> >>>>>> >> >> >>> >> >> >> >> >> >>>>>> >> >> >>> >> >> >> On Wed, Jun 6, 2012 at 3:32 PM, shashwat >> >> >>>>>> >> >> >>> >> >> >> shriparv >> >> >>>>>> >> >> >>> >> >> >> <dwivedishash...@gmail.com> wrote: >> >> >>>>>> >> >> >>> >> >> >>> >> >> >>>>>> >> >> >>> >> >> >>> first type http://localhost:50070 whether >> >> >>>>>> >> >> >>> >> >> >>> this >> >> >>>>>> >> >> >>> >> >> >>> is >> >> >>>>>> >> >> >>> >> >> >>> opening >> >> >>>>>> >> >> >>> >> >> >>> or >> >> >>>>>> >> >> >>> >> >> >>> not >> >> >>>>>> >> >> >>> >> >> >>> and >> >> >>>>>> >> >> >>> >> >> >>> check >> >> >>>>>> >> >> >>> >> >> >>> how many nodes are available, check some of >> >> >>>>>> >> >> >>> >> >> >>> the >> >> >>>>>> >> >> >>> >> >> >>> hadoop >> >> >>>>>> >> >> >>> >> >> >>> shell >> >> >>>>>> >> >> >>> >> >> >>> commands >> >> >>>>>> >> >> >>> >> >> >>> >> >> >>>>>> >> >> >>> >> >> >>> >> >> >>>>>> >> >> >>> >> >> >>> >> >> >>>>>> >> >> >>> >> >> >>> >> >> >>>>>> >> >> >>> >> >> >>> >> >> >>>>>> >> >> >>> >> >> >>> >> >> >>>>>> >> >> >>> >> >> >>> from http://hadoop.apache.org/common/docs/r0.18.3/hdfs_shell.html >> >> >>>>>> >> >> >>> >> >> >>> run >> >> >>>>>> >> >> >>> >> >> >>> example mapreduce task on hadoop take >> >> >>>>>> >> >> >>> >> >> >>> example >> >> >>>>>> >> >> >>> >> >> >>> from >> >> >>>>>> >> >> >>> >> >> >>> here >> >> >>>>>> >> >> >>> >> >> >>> >> >> >>>>>> >> >> >>> >> >> >>> >> >> >>>>>> >> >> >>> >> >> >>> >> >> >>>>>> >> >> >>> >> >> >>> >> >> >>>>>> >> >> >>> >> >> >>> >> >> >>>>>> >> >> >>> >> >> >>> >> >> >>>>>> >> >> >>> >> >> >>> >> >> >>>>>> >> >> >>> >> >> >>> >> >> >>>>>> >> >> >>> >> >> >>> : http://www.michael-noll.com/blog/2011/04/09/benchmarking-and-stress-testing-an-hadoop-cluster-with-terasort-testdfsio-nnbench-mrbench/ >> >> >>>>>> >> >> >>> >> >> >>> >> >> >>>>>> >> >> >>> >> >> >>> if all the above you can do sucessfully >> >> >>>>>> >> >> >>> >> >> >>> means >> >> >>>>>> >> >> >>> >> >> >>> hadoop is >> >> >>>>>> >> >> >>> >> >> >>> configured >> >> >>>>>> >> >> >>> >> >> >>> correctly >> >> >>>>>> >> >> >>> >> >> >>> >> >> >>>>>> >> >> >>> >> >> >>> Regards >> >> >>>>>> >> >> >>> >> >> >>> Shashwat >> >> >>>>>> >> >> >>> >> >> >>> >> >> >>>>>> >> >> >>> >> >> >>> >> >> >>>>>> >> >> >>> >> >> >>> On Wed, Jun 6, 2012 at 1:30 AM, Babak Bastan >> >> >>>>>> >> >> >>> >> >> >>> <babak...@gmail.com> >> >> >>>>>> >> >> >>> >> >> >>> wrote: >> >> >>>>>> >> >> >>> >> >> >>>> >> >> >>>>>> >> >> >>> >> >> >>>> no I'm not working on CDH.Is there a way to >> >> >>>>>> >> >> >>> >> >> >>>> test >> >> >>>>>> >> >> >>> >> >> >>>> if my >> >> >>>>>> >> >> >>> >> >> >>>> Hadoop >> >> >>>>>> >> >> >>> >> >> >>>> works >> >> >>>>>> >> >> >>> >> >> >>>> fine >> >> >>>>>> >> >> >>> >> >> >>>> or not? >> >> >>>>>> >> >> >>> >> >> >>>> >> >> >>>>>> >> >> >>> >> >> >>>> >> >> >>>>>> >> >> >>> >> >> >>>> On Tue, Jun 5, 2012 at 9:55 PM, Bejoy KS >> >> >>>>>> >> >> >>> >> >> >>>> <bejoy...@yahoo.com> >> >> >>>>>> >> >> >>> >> >> >>>> wrote: >> >> >>>>>> >> >> >>> >> >> >>>>> >> >> >>>>>> >> >> >>> >> >> >>>>> Hi Babak >> >> >>>>>> >> >> >>> >> >> >>>>> >> >> >>>>>> >> >> >>> >> >> >>>>> You gotta follow those instructions in the >> >> >>>>>> >> >> >>> >> >> >>>>> apace >> >> >>>>>> >> >> >>> >> >> >>>>> site >> >> >>>>>> >> >> >>> >> >> >>>>> to >> >> >>>>>> >> >> >>> >> >> >>>>> set >> >> >>>>>> >> >> >>> >> >> >>>>> up >> >> >>>>>> >> >> >>> >> >> >>>>> hadoop >> >> >>>>>> >> >> >>> >> >> >>>>> from scratch and ensure that hdfs is >> >> >>>>>> >> >> >>> >> >> >>>>> working >> >> >>>>>> >> >> >>> >> >> >>>>> first. You >> >> >>>>>> >> >> >>> >> >> >>>>> should >> >> >>>>>> >> >> >>> >> >> >>>>> be >> >> >>>>>> >> >> >>> >> >> >>>>> able to >> >> >>>>>> >> >> >>> >> >> >>>>> read and write files to hdfs before you do >> >> >>>>>> >> >> >>> >> >> >>>>> your >> >> >>>>>> >> >> >>> >> >> >>>>> next >> >> >>>>>> >> >> >>> >> >> >>>>> steps. >> >> >>>>>> >> >> >>> >> >> >>>>> >> >> >>>>>> >> >> >>> >> >> >>>>> Are you on CDH or apache distribution of >> >> >>>>>> >> >> >>> >> >> >>>>> hadoop? >> >> >>>>>> >> >> >>> >> >> >>>>> If it >> >> >>>>>> >> >> >>> >> >> >>>>> is >> >> >>>>>> >> >> >>> >> >> >>>>> CDH >> >> >>>>>> >> >> >>> >> >> >>>>> there >> >> >>>>>> >> >> >>> >> >> >>>>> are >> >> >>>>>> >> >> >>> >> >> >>>>> detailed instructions on Cloudera web >> >> >>>>>> >> >> >>> >> >> >>>>> site. >> >> >>>>>> >> >> >>> >> >> >>>>> >> >> >>>>>> >> >> >>> >> >> >>>>> Regards >> >> >>>>>> >> >> >>> >> >> >>>>> Bejoy KS >> >> >>>>>> >> >> >>> >> >> >>>>> >> >> >>>>>> >> >> >>> >> >> >>>>> Sent from handheld, please excuse typos. >> >> >>>>>> >> >> >>> >> >> >>>>> ________________________________ >> >> >>>>>> >> >> >>> >> >> >>>>> From: Babak Bastan <babak...@gmail.com> >> >> >>>>>> >> >> >>> >> >> >>>>> Date: Tue, 5 Jun 2012 21:30:22 +0200 >> >> >>>>>> >> >> >>> >> >> >>>>> To: <user@hive.apache.org> >> >> >>>>>> >> >> >>> >> >> >>>>> ReplyTo: user@hive.apache.org >> >> >>>>>> >> >> >>> >> >> >>>>> Subject: Re: Error while Creating Table in >> >> >>>>>> >> >> >>> >> >> >>>>> Hive >> >> >>>>>> >> >> >>> >> >> >>>>> >> >> >>>>>> >> >> >>> >> >> >>>>> @Bejoy: I set the fs.default.name in the >> >> >>>>>> >> >> >>> >> >> >>>>> core-site.xml >> >> >>>>>> >> >> >>> >> >> >>>>> and >> >> >>>>>> >> >> >>> >> >> >>>>> I >> >> >>>>>> >> >> >>> >> >> >>>>> did >> >> >>>>>> >> >> >>> >> >> >>>>> all >> >> >>>>>> >> >> >>> >> >> >>>>> of >> >> >>>>>> >> >> >>> >> >> >>>>> thing that was mentioned in the reference >> >> >>>>>> >> >> >>> >> >> >>>>> but >> >> >>>>>> >> >> >>> >> >> >>>>> no >> >> >>>>>> >> >> >>> >> >> >>>>> effect >> >> >>>>>> >> >> >>> >> >> >>>>> >> >> >>>>>> >> >> >>> >> >> >>>>> On Tue, Jun 5, 2012 at 8:43 PM, Babak >> >> >>>>>> >> >> >>> >> >> >>>>> Bastan >> >> >>>>>> >> >> >>> >> >> >>>>> <babak...@gmail.com> >> >> >>>>>> >> >> >>> >> >> >>>>> wrote: >> >> >>>>>> >> >> >>> >> >> >>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>> Ok sorry but that was my Mistake .I >> >> >>>>>> >> >> >>> >> >> >>>>>> thought >> >> >>>>>> >> >> >>> >> >> >>>>>> it >> >> >>>>>> >> >> >>> >> >> >>>>>> works >> >> >>>>>> >> >> >>> >> >> >>>>>> but >> >> >>>>>> >> >> >>> >> >> >>>>>> no. >> >> >>>>>> >> >> >>> >> >> >>>>>> I wrote the command without ; and then I >> >> >>>>>> >> >> >>> >> >> >>>>>> think >> >> >>>>>> >> >> >>> >> >> >>>>>> It >> >> >>>>>> >> >> >>> >> >> >>>>>> works >> >> >>>>>> >> >> >>> >> >> >>>>>> but >> >> >>>>>> >> >> >>> >> >> >>>>>> with >> >> >>>>>> >> >> >>> >> >> >>>>>> ; >> >> >>>>>> >> >> >>> >> >> >>>>>> at >> >> >>>>>> >> >> >>> >> >> >>>>>> the end of command >> >> >>>>>> >> >> >>> >> >> >>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>> CREATE TABLE pokes (foo INT, bar STRING); >> >> >>>>>> >> >> >>> >> >> >>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>> does'nt work >> >> >>>>>> >> >> >>> >> >> >>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>> On Tue, Jun 5, 2012 at 8:34 PM, shashwat >> >> >>>>>> >> >> >>> >> >> >>>>>> shriparv >> >> >>>>>> >> >> >>> >> >> >>>>>> <dwivedishash...@gmail.com> wrote: >> >> >>>>>> >> >> >>> >> >> >>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>> inside configuration. all properties >> >> >>>>>> >> >> >>> >> >> >>>>>>> will >> >> >>>>>> >> >> >>> >> >> >>>>>>> be >> >> >>>>>> >> >> >>> >> >> >>>>>>> inside >> >> >>>>>> >> >> >>> >> >> >>>>>>> the >> >> >>>>>> >> >> >>> >> >> >>>>>>> configuration >> >> >>>>>> >> >> >>> >> >> >>>>>>> tags >> >> >>>>>> >> >> >>> >> >> >>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>> On Tue, Jun 5, 2012 at 11:53 PM, Babak >> >> >>>>>> >> >> >>> >> >> >>>>>>> Bastan >> >> >>>>>> >> >> >>> >> >> >>>>>>> <babak...@gmail.com> >> >> >>>>>> >> >> >>> >> >> >>>>>>> wrote: >> >> >>>>>> >> >> >>> >> >> >>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>> Thank you so much my friend your idee >> >> >>>>>> >> >> >>> >> >> >>>>>>>> works >> >> >>>>>> >> >> >>> >> >> >>>>>>>> fine(no >> >> >>>>>> >> >> >>> >> >> >>>>>>>> error) >> >> >>>>>> >> >> >>> >> >> >>>>>>>> you >> >> >>>>>> >> >> >>> >> >> >>>>>>>> are >> >> >>>>>> >> >> >>> >> >> >>>>>>>> the best :) >> >> >>>>>> >> >> >>> >> >> >>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>> On Tue, Jun 5, 2012 at 8:20 PM, Babak >> >> >>>>>> >> >> >>> >> >> >>>>>>>> Bastan >> >> >>>>>> >> >> >>> >> >> >>>>>>>> <babak...@gmail.com> >> >> >>>>>> >> >> >>> >> >> >>>>>>>> wrote: >> >> >>>>>> >> >> >>> >> >> >>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>> It must be inside the >> >> >>>>>> >> >> >>> >> >> >>>>>>>>> <configuration></configuration> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>> or >> >> >>>>>> >> >> >>> >> >> >>>>>>>>> outside >> >> >>>>>> >> >> >>> >> >> >>>>>>>>> this? >> >> >>>>>> >> >> >>> >> >> >>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>> On Tue, Jun 5, 2012 at 8:15 PM, >> >> >>>>>> >> >> >>> >> >> >>>>>>>>> shashwat >> >> >>>>>> >> >> >>> >> >> >>>>>>>>> shriparv >> >> >>>>>> >> >> >>> >> >> >>>>>>>>> <dwivedishash...@gmail.com> wrote: >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>> It will be inside hive/conf >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>> On Tue, Jun 5, 2012 at 11:43 PM, >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>> Babak >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>> Bastan >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>> <babak...@gmail.com> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>> wrote: >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>> Thanks sShashwat, and where is this >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>> hive-site.xml >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>> On Tue, Jun 5, 2012 at 8:02 PM, >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>> shashwat >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>> shriparv >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>> <dwivedishash...@gmail.com> wrote: >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> set >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> hive.metastore.warehouse.dir in >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> hive-site.xml >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> <property> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> <name>hive.metastore.local</name> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> <value>true</value> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> </property> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> <name>hive.metastore.warehouse.dir</name> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> <value>/home/<your >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> username>/hivefolder</value> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> <description>location >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> of >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> default >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> database >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> for >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> the >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> warehouse</description> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> </property> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> On Tue, Jun 5, 2012 at 10:43 PM, >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> Babak >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> Bastan >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> <babak...@gmail.com> wrote: >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> Hello Experts , >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> I'm new in Hive .When try to >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> create a >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> test >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> Table >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> in >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> Hive >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> I >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> get >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> an error.I want to run this >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> command: >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> CREATE TABLE Test (DateT STRING, >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> Url >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> STRING, >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> Content >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> STRING); >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> but this error occured: >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> FAILED: Error in metadata: >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> MetaException(message:Got >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> exception: >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> java.io.FileNotFoundException File >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> file:/user/hive/warehouse/test >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> does >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> not >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> exist.) >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> FAILED: Execution Error, return >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> code >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> 1 >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> from >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> org.apache.hadoop.hive.ql.exec.DDLTask >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> How can I solve this Problem? >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>>> Thank you so much >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> -- >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> ∞ >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> Shashwat Shriparv >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>> -- >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>> ∞ >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>> Shashwat Shriparv >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>> -- >> >> >>>>>> >> >> >>> >> >> >>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>> ∞ >> >> >>>>>> >> >> >>> >> >> >>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>> Shashwat Shriparv >> >> >>>>>> >> >> >>> >> >> >>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>>> >> >> >>>>>> >> >> >>> >> >> >>>>> >> >> >>>>>> >> >> >>> >> >> >>>> >> >> >>>>>> >> >> >>> >> >> >>> >> >> >>>>>> >> >> >>> >> >> >>> >> >> >>>>>> >> >> >>> >> >> >>> >> >> >>>>>> >> >> >>> >> >> >>> -- >> >> >>>>>> >> >> >>> >> >> >>> >> >> >>>>>> >> >> >>> >> >> >>> >> >> >>>>>> >> >> >>> >> >> >>> ∞ >> >> >>>>>> >> >> >>> >> >> >>> >> >> >>>>>> >> >> >>> >> >> >>> Shashwat Shriparv >> >> >>>>>> >> >> >>> >> >> >>> >> >> >>>>>> >> >> >>> >> >> >>> >> >> >>>>>> >> >> >>> >> >> >> >> >> >>>>>> >> >> >>> >> > >> >> >>>>>> >> >> >>> >> > >> >> >>>>>> >> >> >>> > >> >> >>>>>> >> >> >>> > >> >> >>>>>> >> >> >> >> >> >>>>>> >> >> >> >> >> >>>>>> >> > >> >> >>>>>> >> > >> >> >>>>>> > >> >> >>>>>> > >> >> >>>>> >> >> >>>>> >> >> >>>> >> >> >>>> >> >> >>>> >> >> >>>> -- >> >> >>>> >> >> >>>> >> >> >>>> ∞ >> >> >>>> >> >> >>>> Shashwat Shriparv >> >> >>>> >> >> >>>> >> >> >>> >> >> >> >> >> > >> > >> > > >