also change the permissions of these directories to 777. Regards, Mohammad Tariq
On Wed, Jun 6, 2012 at 11:54 PM, Mohammad Tariq <donta...@gmail.com> wrote: > create a directory "/home/username/hdfs" (or at some place of your > choice)..inside this hdfs directory create three sub directories - > name, data, and temp, then follow these steps : > > add following properties in your core-site.xml - > > <property> > <name>fs.default.name</name> > <value>hdfs://localhost:9000/</value> > </property> > > <property> > <name>hadoop.tmp.dir</name> > <value>/home/mohammad/hdfs/temp</value> > </property> > > then add following two properties in your hdfs-site.xml - > > <property> > <name>dfs.replication</name> > <value>1</value> > </property> > > <property> > <name>dfs.name.dir</name> > <value>/home/mohammad/hdfs/name</value> > </property> > > <property> > <name>dfs.data.dir</name> > <value>/home/mohammad/hdfs/data</value> > </property> > > finally add this property in your mapred-site.xml - > > <property> > <name>mapred.job.tracker</name> > <value>hdfs://localhost:9001</value> > </property> > > NOTE: you can give any name to these directories of your choice, just > keep in mind you have to give same names as values of > above specified properties in your configuration files. > (give full path of these directories, not just the name of the > directory) > > After this follow the steps provided in the previous reply. > > Regards, > Mohammad Tariq > > > On Wed, Jun 6, 2012 at 11:42 PM, Babak Bastan <babak...@gmail.com> wrote: >> thank's Mohammad >> >> with this command: >> >> babak@ubuntu:~/Downloads/hadoop/bin$ hadoop namenode -format >> >> this is my output: >> >> 12/06/06 20:05:20 INFO namenode.NameNode: STARTUP_MSG: >> /************************************************************ >> STARTUP_MSG: Starting NameNode >> STARTUP_MSG: host = ubuntu/127.0.1.1 >> STARTUP_MSG: args = [-format] >> STARTUP_MSG: version = 0.20.2 >> STARTUP_MSG: build = >> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20 -r >> 911707; compiled by 'chrisdo' on Fri Feb 19 08:07:34 UTC 2010 >> ************************************************************/ >> 12/06/06 20:05:20 INFO namenode.FSNamesystem: >> fsOwner=babak,babak,adm,dialout,cdrom,plugdev,lpadmin,admin,sambashare >> 12/06/06 20:05:20 INFO namenode.FSNamesystem: supergroup=supergroup >> 12/06/06 20:05:20 INFO namenode.FSNamesystem: isPermissionEnabled=true >> 12/06/06 20:05:20 INFO common.Storage: Image file of size 95 saved in 0 >> seconds. >> 12/06/06 20:05:20 INFO common.Storage: Storage directory >> /tmp/hadoop-babak/dfs/name has been successfully formatted. >> 12/06/06 20:05:20 INFO namenode.NameNode: SHUTDOWN_MSG: >> /************************************************************ >> SHUTDOWN_MSG: Shutting down NameNode at ubuntu/127.0.1.1 >> ************************************************************/ >> >> by this command: >> >> babak@ubuntu:~/Downloads/hadoop/bin$ start-dfs.sh >> >> this is the out put >> >> mkdir: kann Verzeichnis „/home/babak/Downloads/hadoop/bin/../logs“ nicht >> anlegen: Keine Berechtigung >> >> this out put(it's in german and it means no right to make this folder) >> >> >> On Wed, Jun 6, 2012 at 7:59 PM, Mohammad Tariq <donta...@gmail.com> wrote: >>> >>> once we are done with the configuration, we need to format the file >>> system..use this command to do that- >>> bin/hadoop namenode -format >>> >>> after this, hadoop daemon processes should be started using following >>> commands - >>> bin/start-dfs.sh (it'll start NN & DN) >>> bin/start-mapred.sh (it'll start JT & TT) >>> >>> after this use jps to check if everything is alright or point your >>> browser to localhost:50070..if you further find any problem provide us >>> with the error logs..:) >>> >>> Regards, >>> Mohammad Tariq >>> >>> >>> On Wed, Jun 6, 2012 at 11:22 PM, Babak Bastan <babak...@gmail.com> wrote: >>> > were you able to format hdfs properly??? >>> > I did'nt get your question,Do you mean HADOOP_HOME? or where did I >>> > install >>> > Hadoop? >>> > >>> > On Wed, Jun 6, 2012 at 7:49 PM, Mohammad Tariq <donta...@gmail.com> >>> > wrote: >>> >> >>> >> if you are getting only this, it means your hadoop is not >>> >> running..were you able to format hdfs properly??? >>> >> >>> >> Regards, >>> >> Mohammad Tariq >>> >> >>> >> >>> >> On Wed, Jun 6, 2012 at 11:17 PM, Babak Bastan <babak...@gmail.com> >>> >> wrote: >>> >> > Hi MohammadmI irun jps in my shel I can see this result: >>> >> > 2213 Jps >>> >> > >>> >> > >>> >> > On Wed, Jun 6, 2012 at 7:44 PM, Mohammad Tariq <donta...@gmail.com> >>> >> > wrote: >>> >> >> >>> >> >> you can also use "jps" command at your shell to see whether Hadoop >>> >> >> processes are running or not. >>> >> >> >>> >> >> Regards, >>> >> >> Mohammad Tariq >>> >> >> >>> >> >> >>> >> >> On Wed, Jun 6, 2012 at 11:12 PM, Mohammad Tariq <donta...@gmail.com> >>> >> >> wrote: >>> >> >> > Hi Babak, >>> >> >> > >>> >> >> > You have to type it in you web browser..Hadoop provides us a web >>> >> >> > GUI >>> >> >> > that not only allows us to browse through the file system, but to >>> >> >> > download the files as well..Apart from that it also provides a web >>> >> >> > GUI >>> >> >> > that can be used to see the status of Jobtracker and >>> >> >> > Tasktracker..When >>> >> >> > you run a Hive or Pig job or a Mapreduce job, you can point your >>> >> >> > browser to http://localhost:50030 to see the status and logs of >>> >> >> > your >>> >> >> > job. >>> >> >> > >>> >> >> > Regards, >>> >> >> > Mohammad Tariq >>> >> >> > >>> >> >> > >>> >> >> > On Wed, Jun 6, 2012 at 8:28 PM, Babak Bastan <babak...@gmail.com> >>> >> >> > wrote: >>> >> >> >> Thank you shashwat for the answer, >>> >> >> >> where should I type http://localhost:50070? >>> >> >> >> I typed here: hive>http://localhost:50070 but nothing as result >>> >> >> >> >>> >> >> >> >>> >> >> >> On Wed, Jun 6, 2012 at 3:32 PM, shashwat shriparv >>> >> >> >> <dwivedishash...@gmail.com> wrote: >>> >> >> >>> >>> >> >> >>> first type http://localhost:50070 whether this is opening or not >>> >> >> >>> and >>> >> >> >>> check >>> >> >> >>> how many nodes are available, check some of the hadoop shell >>> >> >> >>> commands >>> >> >> >>> >>> >> >> >>> from http://hadoop.apache.org/common/docs/r0.18.3/hdfs_shell.html >>> >> >> >>> run >>> >> >> >>> example mapreduce task on hadoop take example from here >>> >> >> >>> >>> >> >> >>> >>> >> >> >>> >>> >> >> >>> : http://www.michael-noll.com/blog/2011/04/09/benchmarking-and-stress-testing-an-hadoop-cluster-with-terasort-testdfsio-nnbench-mrbench/ >>> >> >> >>> >>> >> >> >>> if all the above you can do sucessfully means hadoop is >>> >> >> >>> configured >>> >> >> >>> correctly >>> >> >> >>> >>> >> >> >>> Regards >>> >> >> >>> Shashwat >>> >> >> >>> >>> >> >> >>> >>> >> >> >>> On Wed, Jun 6, 2012 at 1:30 AM, Babak Bastan >>> >> >> >>> <babak...@gmail.com> >>> >> >> >>> wrote: >>> >> >> >>>> >>> >> >> >>>> no I'm not working on CDH.Is there a way to test if my Hadoop >>> >> >> >>>> works >>> >> >> >>>> fine >>> >> >> >>>> or not? >>> >> >> >>>> >>> >> >> >>>> >>> >> >> >>>> On Tue, Jun 5, 2012 at 9:55 PM, Bejoy KS <bejoy...@yahoo.com> >>> >> >> >>>> wrote: >>> >> >> >>>>> >>> >> >> >>>>> Hi Babak >>> >> >> >>>>> >>> >> >> >>>>> You gotta follow those instructions in the apace site to set >>> >> >> >>>>> up >>> >> >> >>>>> hadoop >>> >> >> >>>>> from scratch and ensure that hdfs is working first. You should >>> >> >> >>>>> be >>> >> >> >>>>> able to >>> >> >> >>>>> read and write files to hdfs before you do your next steps. >>> >> >> >>>>> >>> >> >> >>>>> Are you on CDH or apache distribution of hadoop? If it is CDH >>> >> >> >>>>> there >>> >> >> >>>>> are >>> >> >> >>>>> detailed instructions on Cloudera web site. >>> >> >> >>>>> >>> >> >> >>>>> Regards >>> >> >> >>>>> Bejoy KS >>> >> >> >>>>> >>> >> >> >>>>> Sent from handheld, please excuse typos. >>> >> >> >>>>> ________________________________ >>> >> >> >>>>> From: Babak Bastan <babak...@gmail.com> >>> >> >> >>>>> Date: Tue, 5 Jun 2012 21:30:22 +0200 >>> >> >> >>>>> To: <user@hive.apache.org> >>> >> >> >>>>> ReplyTo: user@hive.apache.org >>> >> >> >>>>> Subject: Re: Error while Creating Table in Hive >>> >> >> >>>>> >>> >> >> >>>>> @Bejoy: I set the fs.default.name in the core-site.xml and I >>> >> >> >>>>> did >>> >> >> >>>>> all >>> >> >> >>>>> of >>> >> >> >>>>> thing that was mentioned in the reference but no effect >>> >> >> >>>>> >>> >> >> >>>>> On Tue, Jun 5, 2012 at 8:43 PM, Babak Bastan >>> >> >> >>>>> <babak...@gmail.com> >>> >> >> >>>>> wrote: >>> >> >> >>>>>> >>> >> >> >>>>>> Ok sorry but that was my Mistake .I thought it works but no. >>> >> >> >>>>>> I wrote the command without ; and then I think It works but >>> >> >> >>>>>> with >>> >> >> >>>>>> ; >>> >> >> >>>>>> at >>> >> >> >>>>>> the end of command >>> >> >> >>>>>> >>> >> >> >>>>>> CREATE TABLE pokes (foo INT, bar STRING); >>> >> >> >>>>>> >>> >> >> >>>>>> does'nt work >>> >> >> >>>>>> >>> >> >> >>>>>> >>> >> >> >>>>>> On Tue, Jun 5, 2012 at 8:34 PM, shashwat shriparv >>> >> >> >>>>>> <dwivedishash...@gmail.com> wrote: >>> >> >> >>>>>>> >>> >> >> >>>>>>> inside configuration. all properties will be inside the >>> >> >> >>>>>>> configuration >>> >> >> >>>>>>> tags >>> >> >> >>>>>>> >>> >> >> >>>>>>> >>> >> >> >>>>>>> On Tue, Jun 5, 2012 at 11:53 PM, Babak Bastan >>> >> >> >>>>>>> <babak...@gmail.com> >>> >> >> >>>>>>> wrote: >>> >> >> >>>>>>>> >>> >> >> >>>>>>>> Thank you so much my friend your idee works fine(no error) >>> >> >> >>>>>>>> you >>> >> >> >>>>>>>> are >>> >> >> >>>>>>>> the best :) >>> >> >> >>>>>>>> >>> >> >> >>>>>>>> >>> >> >> >>>>>>>> On Tue, Jun 5, 2012 at 8:20 PM, Babak Bastan >>> >> >> >>>>>>>> <babak...@gmail.com> >>> >> >> >>>>>>>> wrote: >>> >> >> >>>>>>>>> >>> >> >> >>>>>>>>> It must be inside the <configuration></configuration> or >>> >> >> >>>>>>>>> outside >>> >> >> >>>>>>>>> this? >>> >> >> >>>>>>>>> >>> >> >> >>>>>>>>> >>> >> >> >>>>>>>>> On Tue, Jun 5, 2012 at 8:15 PM, shashwat shriparv >>> >> >> >>>>>>>>> <dwivedishash...@gmail.com> wrote: >>> >> >> >>>>>>>>>> >>> >> >> >>>>>>>>>> It will be inside hive/conf >>> >> >> >>>>>>>>>> >>> >> >> >>>>>>>>>> >>> >> >> >>>>>>>>>> On Tue, Jun 5, 2012 at 11:43 PM, Babak Bastan >>> >> >> >>>>>>>>>> <babak...@gmail.com> >>> >> >> >>>>>>>>>> wrote: >>> >> >> >>>>>>>>>>> >>> >> >> >>>>>>>>>>> Thanks sShashwat, and where is this hive-site.xml >>> >> >> >>>>>>>>>>> >>> >> >> >>>>>>>>>>> On Tue, Jun 5, 2012 at 8:02 PM, shashwat shriparv >>> >> >> >>>>>>>>>>> <dwivedishash...@gmail.com> wrote: >>> >> >> >>>>>>>>>>>> >>> >> >> >>>>>>>>>>>> set >>> >> >> >>>>>>>>>>>> >>> >> >> >>>>>>>>>>>> hive.metastore.warehouse.dir in hive-site.xml >>> >> >> >>>>>>>>>>>> >>> >> >> >>>>>>>>>>>> <property> >>> >> >> >>>>>>>>>>>> <name>hive.metastore.local</name> >>> >> >> >>>>>>>>>>>> <value>true</value> >>> >> >> >>>>>>>>>>>> </property> >>> >> >> >>>>>>>>>>>> >>> >> >> >>>>>>>>>>>> <name>hive.metastore.warehouse.dir</name> >>> >> >> >>>>>>>>>>>> <value>/home/<your >>> >> >> >>>>>>>>>>>> username>/hivefolder</value> >>> >> >> >>>>>>>>>>>> <description>location of default >>> >> >> >>>>>>>>>>>> database >>> >> >> >>>>>>>>>>>> for >>> >> >> >>>>>>>>>>>> the >>> >> >> >>>>>>>>>>>> warehouse</description> >>> >> >> >>>>>>>>>>>> </property> >>> >> >> >>>>>>>>>>>> >>> >> >> >>>>>>>>>>>> >>> >> >> >>>>>>>>>>>> On Tue, Jun 5, 2012 at 10:43 PM, Babak Bastan >>> >> >> >>>>>>>>>>>> <babak...@gmail.com> wrote: >>> >> >> >>>>>>>>>>>>> >>> >> >> >>>>>>>>>>>>> Hello Experts , >>> >> >> >>>>>>>>>>>>> >>> >> >> >>>>>>>>>>>>> I'm new in Hive .When try to create a test Table in >>> >> >> >>>>>>>>>>>>> Hive >>> >> >> >>>>>>>>>>>>> I >>> >> >> >>>>>>>>>>>>> get >>> >> >> >>>>>>>>>>>>> an error.I want to run this command: >>> >> >> >>>>>>>>>>>>> CREATE TABLE Test (DateT STRING, Url STRING, Content >>> >> >> >>>>>>>>>>>>> STRING); >>> >> >> >>>>>>>>>>>>> but this error occured: >>> >> >> >>>>>>>>>>>>> FAILED: Error in metadata: MetaException(message:Got >>> >> >> >>>>>>>>>>>>> exception: >>> >> >> >>>>>>>>>>>>> java.io.FileNotFoundException File >>> >> >> >>>>>>>>>>>>> file:/user/hive/warehouse/test does not >>> >> >> >>>>>>>>>>>>> exist.) >>> >> >> >>>>>>>>>>>>> FAILED: Execution Error, return code 1 from >>> >> >> >>>>>>>>>>>>> org.apache.hadoop.hive.ql.exec.DDLTask >>> >> >> >>>>>>>>>>>>> How can I solve this Problem? >>> >> >> >>>>>>>>>>>>> Thank you so much >>> >> >> >>>>>>>>>>>> >>> >> >> >>>>>>>>>>>> >>> >> >> >>>>>>>>>>>> >>> >> >> >>>>>>>>>>>> >>> >> >> >>>>>>>>>>>> -- >>> >> >> >>>>>>>>>>>> >>> >> >> >>>>>>>>>>>> >>> >> >> >>>>>>>>>>>> ∞ >>> >> >> >>>>>>>>>>>> >>> >> >> >>>>>>>>>>>> Shashwat Shriparv >>> >> >> >>>>>>>>>>>> >>> >> >> >>>>>>>>>>>> >>> >> >> >>>>>>>>>>> >>> >> >> >>>>>>>>>> >>> >> >> >>>>>>>>>> >>> >> >> >>>>>>>>>> >>> >> >> >>>>>>>>>> -- >>> >> >> >>>>>>>>>> >>> >> >> >>>>>>>>>> >>> >> >> >>>>>>>>>> ∞ >>> >> >> >>>>>>>>>> >>> >> >> >>>>>>>>>> Shashwat Shriparv >>> >> >> >>>>>>>>>> >>> >> >> >>>>>>>>>> >>> >> >> >>>>>>>>> >>> >> >> >>>>>>>> >>> >> >> >>>>>>> >>> >> >> >>>>>>> >>> >> >> >>>>>>> >>> >> >> >>>>>>> -- >>> >> >> >>>>>>> >>> >> >> >>>>>>> >>> >> >> >>>>>>> ∞ >>> >> >> >>>>>>> >>> >> >> >>>>>>> Shashwat Shriparv >>> >> >> >>>>>>> >>> >> >> >>>>>>> >>> >> >> >>>>>> >>> >> >> >>>>> >>> >> >> >>>> >>> >> >> >>> >>> >> >> >>> >>> >> >> >>> >>> >> >> >>> -- >>> >> >> >>> >>> >> >> >>> >>> >> >> >>> ∞ >>> >> >> >>> >>> >> >> >>> Shashwat Shriparv >>> >> >> >>> >>> >> >> >>> >>> >> >> >> >>> >> > >>> >> > >>> > >>> > >> >>