hoe can I get my log mohammad? On Wed, Jun 6, 2012 at 9:36 PM, Mohammad Tariq <donta...@gmail.com> wrote:
> could you post your logs???that would help me in understanding the > problem properly. > > Regards, > Mohammad Tariq > > > On Thu, Jun 7, 2012 at 1:02 AM, Babak Bastan <babak...@gmail.com> wrote: > > Thank you very much mohamad for your attention.I followed the steps but > the > > error is the same as the last time. > > and there is my hosts file: > > > > 127.0.0.1 localhost > > #127.0.0.1 ubuntu.ubuntu-domain ubuntu > > > > > > # The following lines are desirable for IPv6 capable hosts > > > > #::1 ip6-localhost ip6-loopback > > #fe00::0 ip6-localnet > > #ff00::0 ip6-mcastprefix > > #ff02::1 ip6-allnodes > > #ff02::2 ip6-allrouters > > > > but no effect :( > > > > On Wed, Jun 6, 2012 at 8:25 PM, Mohammad Tariq <donta...@gmail.com> > wrote: > >> > >> also change the permissions of these directories to 777. > >> > >> Regards, > >> Mohammad Tariq > >> > >> > >> On Wed, Jun 6, 2012 at 11:54 PM, Mohammad Tariq <donta...@gmail.com> > >> wrote: > >> > create a directory "/home/username/hdfs" (or at some place of your > >> > choice)..inside this hdfs directory create three sub directories - > >> > name, data, and temp, then follow these steps : > >> > > >> > add following properties in your core-site.xml - > >> > > >> > <property> > >> > <name>fs.default.name</name> > >> > <value>hdfs://localhost:9000/</value> > >> > </property> > >> > > >> > <property> > >> > <name>hadoop.tmp.dir</name> > >> > <value>/home/mohammad/hdfs/temp</value> > >> > </property> > >> > > >> > then add following two properties in your hdfs-site.xml - > >> > > >> > <property> > >> > <name>dfs.replication</name> > >> > <value>1</value> > >> > </property> > >> > > >> > <property> > >> > <name>dfs.name.dir</name> > >> > <value>/home/mohammad/hdfs/name</value> > >> > </property> > >> > > >> > <property> > >> > <name>dfs.data.dir</name> > >> > <value>/home/mohammad/hdfs/data</value> > >> > </property> > >> > > >> > finally add this property in your mapred-site.xml - > >> > > >> > <property> > >> > <name>mapred.job.tracker</name> > >> > <value>hdfs://localhost:9001</value> > >> > </property> > >> > > >> > NOTE: you can give any name to these directories of your choice, just > >> > keep in mind you have to give same names as values of > >> > above specified properties in your configuration files. > >> > (give full path of these directories, not just the name of the > >> > directory) > >> > > >> > After this follow the steps provided in the previous reply. > >> > > >> > Regards, > >> > Mohammad Tariq > >> > > >> > > >> > On Wed, Jun 6, 2012 at 11:42 PM, Babak Bastan <babak...@gmail.com> > >> > wrote: > >> >> thank's Mohammad > >> >> > >> >> with this command: > >> >> > >> >> babak@ubuntu:~/Downloads/hadoop/bin$ hadoop namenode -format > >> >> > >> >> this is my output: > >> >> > >> >> 12/06/06 20:05:20 INFO namenode.NameNode: STARTUP_MSG: > >> >> /************************************************************ > >> >> STARTUP_MSG: Starting NameNode > >> >> STARTUP_MSG: host = ubuntu/127.0.1.1 > >> >> STARTUP_MSG: args = [-format] > >> >> STARTUP_MSG: version = 0.20.2 > >> >> STARTUP_MSG: build = > >> >> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20-r > >> >> 911707; compiled by 'chrisdo' on Fri Feb 19 08:07:34 UTC 2010 > >> >> ************************************************************/ > >> >> 12/06/06 20:05:20 INFO namenode.FSNamesystem: > >> >> > fsOwner=babak,babak,adm,dialout,cdrom,plugdev,lpadmin,admin,sambashare > >> >> 12/06/06 20:05:20 INFO namenode.FSNamesystem: supergroup=supergroup > >> >> 12/06/06 20:05:20 INFO namenode.FSNamesystem: > isPermissionEnabled=true > >> >> 12/06/06 20:05:20 INFO common.Storage: Image file of size 95 saved > in 0 > >> >> seconds. > >> >> 12/06/06 20:05:20 INFO common.Storage: Storage directory > >> >> /tmp/hadoop-babak/dfs/name has been successfully formatted. > >> >> 12/06/06 20:05:20 INFO namenode.NameNode: SHUTDOWN_MSG: > >> >> /************************************************************ > >> >> SHUTDOWN_MSG: Shutting down NameNode at ubuntu/127.0.1.1 > >> >> ************************************************************/ > >> >> > >> >> by this command: > >> >> > >> >> babak@ubuntu:~/Downloads/hadoop/bin$ start-dfs.sh > >> >> > >> >> this is the out put > >> >> > >> >> mkdir: kann Verzeichnis „/home/babak/Downloads/hadoop/bin/../logs“ > >> >> nicht > >> >> anlegen: Keine Berechtigung > >> >> > >> >> this out put(it's in german and it means no right to make this > folder) > >> >> > >> >> > >> >> On Wed, Jun 6, 2012 at 7:59 PM, Mohammad Tariq <donta...@gmail.com> > >> >> wrote: > >> >>> > >> >>> once we are done with the configuration, we need to format the file > >> >>> system..use this command to do that- > >> >>> bin/hadoop namenode -format > >> >>> > >> >>> after this, hadoop daemon processes should be started using > following > >> >>> commands - > >> >>> bin/start-dfs.sh (it'll start NN & DN) > >> >>> bin/start-mapred.sh (it'll start JT & TT) > >> >>> > >> >>> after this use jps to check if everything is alright or point your > >> >>> browser to localhost:50070..if you further find any problem provide > us > >> >>> with the error logs..:) > >> >>> > >> >>> Regards, > >> >>> Mohammad Tariq > >> >>> > >> >>> > >> >>> On Wed, Jun 6, 2012 at 11:22 PM, Babak Bastan <babak...@gmail.com> > >> >>> wrote: > >> >>> > were you able to format hdfs properly??? > >> >>> > I did'nt get your question,Do you mean HADOOP_HOME? or where did I > >> >>> > install > >> >>> > Hadoop? > >> >>> > > >> >>> > On Wed, Jun 6, 2012 at 7:49 PM, Mohammad Tariq < > donta...@gmail.com> > >> >>> > wrote: > >> >>> >> > >> >>> >> if you are getting only this, it means your hadoop is not > >> >>> >> running..were you able to format hdfs properly??? > >> >>> >> > >> >>> >> Regards, > >> >>> >> Mohammad Tariq > >> >>> >> > >> >>> >> > >> >>> >> On Wed, Jun 6, 2012 at 11:17 PM, Babak Bastan < > babak...@gmail.com> > >> >>> >> wrote: > >> >>> >> > Hi MohammadmI irun jps in my shel I can see this result: > >> >>> >> > 2213 Jps > >> >>> >> > > >> >>> >> > > >> >>> >> > On Wed, Jun 6, 2012 at 7:44 PM, Mohammad Tariq > >> >>> >> > <donta...@gmail.com> > >> >>> >> > wrote: > >> >>> >> >> > >> >>> >> >> you can also use "jps" command at your shell to see whether > >> >>> >> >> Hadoop > >> >>> >> >> processes are running or not. > >> >>> >> >> > >> >>> >> >> Regards, > >> >>> >> >> Mohammad Tariq > >> >>> >> >> > >> >>> >> >> > >> >>> >> >> On Wed, Jun 6, 2012 at 11:12 PM, Mohammad Tariq > >> >>> >> >> <donta...@gmail.com> > >> >>> >> >> wrote: > >> >>> >> >> > Hi Babak, > >> >>> >> >> > > >> >>> >> >> > You have to type it in you web browser..Hadoop provides us > a > >> >>> >> >> > web > >> >>> >> >> > GUI > >> >>> >> >> > that not only allows us to browse through the file system, > but > >> >>> >> >> > to > >> >>> >> >> > download the files as well..Apart from that it also > provides a > >> >>> >> >> > web > >> >>> >> >> > GUI > >> >>> >> >> > that can be used to see the status of Jobtracker and > >> >>> >> >> > Tasktracker..When > >> >>> >> >> > you run a Hive or Pig job or a Mapreduce job, you can point > >> >>> >> >> > your > >> >>> >> >> > browser to http://localhost:50030 to see the status and > logs > >> >>> >> >> > of > >> >>> >> >> > your > >> >>> >> >> > job. > >> >>> >> >> > > >> >>> >> >> > Regards, > >> >>> >> >> > Mohammad Tariq > >> >>> >> >> > > >> >>> >> >> > > >> >>> >> >> > On Wed, Jun 6, 2012 at 8:28 PM, Babak Bastan > >> >>> >> >> > <babak...@gmail.com> > >> >>> >> >> > wrote: > >> >>> >> >> >> Thank you shashwat for the answer, > >> >>> >> >> >> where should I type http://localhost:50070? > >> >>> >> >> >> I typed here: hive>http://localhost:50070 but nothing as > >> >>> >> >> >> result > >> >>> >> >> >> > >> >>> >> >> >> > >> >>> >> >> >> On Wed, Jun 6, 2012 at 3:32 PM, shashwat shriparv > >> >>> >> >> >> <dwivedishash...@gmail.com> wrote: > >> >>> >> >> >>> > >> >>> >> >> >>> first type http://localhost:50070 whether this is > opening or > >> >>> >> >> >>> not > >> >>> >> >> >>> and > >> >>> >> >> >>> check > >> >>> >> >> >>> how many nodes are available, check some of the hadoop > shell > >> >>> >> >> >>> commands > >> >>> >> >> >>> > >> >>> >> >> >>> > >> >>> >> >> >>> from > http://hadoop.apache.org/common/docs/r0.18.3/hdfs_shell.html > >> >>> >> >> >>> run > >> >>> >> >> >>> example mapreduce task on hadoop take example from here > >> >>> >> >> >>> > >> >>> >> >> >>> > >> >>> >> >> >>> > >> >>> >> >> >>> > >> >>> >> >> >>> : > http://www.michael-noll.com/blog/2011/04/09/benchmarking-and-stress-testing-an-hadoop-cluster-with-terasort-testdfsio-nnbench-mrbench/ > >> >>> >> >> >>> > >> >>> >> >> >>> if all the above you can do sucessfully means hadoop is > >> >>> >> >> >>> configured > >> >>> >> >> >>> correctly > >> >>> >> >> >>> > >> >>> >> >> >>> Regards > >> >>> >> >> >>> Shashwat > >> >>> >> >> >>> > >> >>> >> >> >>> > >> >>> >> >> >>> On Wed, Jun 6, 2012 at 1:30 AM, Babak Bastan > >> >>> >> >> >>> <babak...@gmail.com> > >> >>> >> >> >>> wrote: > >> >>> >> >> >>>> > >> >>> >> >> >>>> no I'm not working on CDH.Is there a way to test if my > >> >>> >> >> >>>> Hadoop > >> >>> >> >> >>>> works > >> >>> >> >> >>>> fine > >> >>> >> >> >>>> or not? > >> >>> >> >> >>>> > >> >>> >> >> >>>> > >> >>> >> >> >>>> On Tue, Jun 5, 2012 at 9:55 PM, Bejoy KS > >> >>> >> >> >>>> <bejoy...@yahoo.com> > >> >>> >> >> >>>> wrote: > >> >>> >> >> >>>>> > >> >>> >> >> >>>>> Hi Babak > >> >>> >> >> >>>>> > >> >>> >> >> >>>>> You gotta follow those instructions in the apace site to > >> >>> >> >> >>>>> set > >> >>> >> >> >>>>> up > >> >>> >> >> >>>>> hadoop > >> >>> >> >> >>>>> from scratch and ensure that hdfs is working first. You > >> >>> >> >> >>>>> should > >> >>> >> >> >>>>> be > >> >>> >> >> >>>>> able to > >> >>> >> >> >>>>> read and write files to hdfs before you do your next > >> >>> >> >> >>>>> steps. > >> >>> >> >> >>>>> > >> >>> >> >> >>>>> Are you on CDH or apache distribution of hadoop? If it > is > >> >>> >> >> >>>>> CDH > >> >>> >> >> >>>>> there > >> >>> >> >> >>>>> are > >> >>> >> >> >>>>> detailed instructions on Cloudera web site. > >> >>> >> >> >>>>> > >> >>> >> >> >>>>> Regards > >> >>> >> >> >>>>> Bejoy KS > >> >>> >> >> >>>>> > >> >>> >> >> >>>>> Sent from handheld, please excuse typos. > >> >>> >> >> >>>>> ________________________________ > >> >>> >> >> >>>>> From: Babak Bastan <babak...@gmail.com> > >> >>> >> >> >>>>> Date: Tue, 5 Jun 2012 21:30:22 +0200 > >> >>> >> >> >>>>> To: <user@hive.apache.org> > >> >>> >> >> >>>>> ReplyTo: user@hive.apache.org > >> >>> >> >> >>>>> Subject: Re: Error while Creating Table in Hive > >> >>> >> >> >>>>> > >> >>> >> >> >>>>> @Bejoy: I set the fs.default.name in the core-site.xml > and > >> >>> >> >> >>>>> I > >> >>> >> >> >>>>> did > >> >>> >> >> >>>>> all > >> >>> >> >> >>>>> of > >> >>> >> >> >>>>> thing that was mentioned in the reference but no effect > >> >>> >> >> >>>>> > >> >>> >> >> >>>>> On Tue, Jun 5, 2012 at 8:43 PM, Babak Bastan > >> >>> >> >> >>>>> <babak...@gmail.com> > >> >>> >> >> >>>>> wrote: > >> >>> >> >> >>>>>> > >> >>> >> >> >>>>>> Ok sorry but that was my Mistake .I thought it works > but > >> >>> >> >> >>>>>> no. > >> >>> >> >> >>>>>> I wrote the command without ; and then I think It works > >> >>> >> >> >>>>>> but > >> >>> >> >> >>>>>> with > >> >>> >> >> >>>>>> ; > >> >>> >> >> >>>>>> at > >> >>> >> >> >>>>>> the end of command > >> >>> >> >> >>>>>> > >> >>> >> >> >>>>>> CREATE TABLE pokes (foo INT, bar STRING); > >> >>> >> >> >>>>>> > >> >>> >> >> >>>>>> does'nt work > >> >>> >> >> >>>>>> > >> >>> >> >> >>>>>> > >> >>> >> >> >>>>>> On Tue, Jun 5, 2012 at 8:34 PM, shashwat shriparv > >> >>> >> >> >>>>>> <dwivedishash...@gmail.com> wrote: > >> >>> >> >> >>>>>>> > >> >>> >> >> >>>>>>> inside configuration. all properties will be inside > the > >> >>> >> >> >>>>>>> configuration > >> >>> >> >> >>>>>>> tags > >> >>> >> >> >>>>>>> > >> >>> >> >> >>>>>>> > >> >>> >> >> >>>>>>> On Tue, Jun 5, 2012 at 11:53 PM, Babak Bastan > >> >>> >> >> >>>>>>> <babak...@gmail.com> > >> >>> >> >> >>>>>>> wrote: > >> >>> >> >> >>>>>>>> > >> >>> >> >> >>>>>>>> Thank you so much my friend your idee works fine(no > >> >>> >> >> >>>>>>>> error) > >> >>> >> >> >>>>>>>> you > >> >>> >> >> >>>>>>>> are > >> >>> >> >> >>>>>>>> the best :) > >> >>> >> >> >>>>>>>> > >> >>> >> >> >>>>>>>> > >> >>> >> >> >>>>>>>> On Tue, Jun 5, 2012 at 8:20 PM, Babak Bastan > >> >>> >> >> >>>>>>>> <babak...@gmail.com> > >> >>> >> >> >>>>>>>> wrote: > >> >>> >> >> >>>>>>>>> > >> >>> >> >> >>>>>>>>> It must be inside the > <configuration></configuration> > >> >>> >> >> >>>>>>>>> or > >> >>> >> >> >>>>>>>>> outside > >> >>> >> >> >>>>>>>>> this? > >> >>> >> >> >>>>>>>>> > >> >>> >> >> >>>>>>>>> > >> >>> >> >> >>>>>>>>> On Tue, Jun 5, 2012 at 8:15 PM, shashwat shriparv > >> >>> >> >> >>>>>>>>> <dwivedishash...@gmail.com> wrote: > >> >>> >> >> >>>>>>>>>> > >> >>> >> >> >>>>>>>>>> It will be inside hive/conf > >> >>> >> >> >>>>>>>>>> > >> >>> >> >> >>>>>>>>>> > >> >>> >> >> >>>>>>>>>> On Tue, Jun 5, 2012 at 11:43 PM, Babak Bastan > >> >>> >> >> >>>>>>>>>> <babak...@gmail.com> > >> >>> >> >> >>>>>>>>>> wrote: > >> >>> >> >> >>>>>>>>>>> > >> >>> >> >> >>>>>>>>>>> Thanks sShashwat, and where is this hive-site.xml > >> >>> >> >> >>>>>>>>>>> > >> >>> >> >> >>>>>>>>>>> On Tue, Jun 5, 2012 at 8:02 PM, shashwat shriparv > >> >>> >> >> >>>>>>>>>>> <dwivedishash...@gmail.com> wrote: > >> >>> >> >> >>>>>>>>>>>> > >> >>> >> >> >>>>>>>>>>>> set > >> >>> >> >> >>>>>>>>>>>> > >> >>> >> >> >>>>>>>>>>>> hive.metastore.warehouse.dir in hive-site.xml > >> >>> >> >> >>>>>>>>>>>> > >> >>> >> >> >>>>>>>>>>>> <property> > >> >>> >> >> >>>>>>>>>>>> <name>hive.metastore.local</name> > >> >>> >> >> >>>>>>>>>>>> <value>true</value> > >> >>> >> >> >>>>>>>>>>>> </property> > >> >>> >> >> >>>>>>>>>>>> > >> >>> >> >> >>>>>>>>>>>> <name>hive.metastore.warehouse.dir</name> > >> >>> >> >> >>>>>>>>>>>> <value>/home/<your > >> >>> >> >> >>>>>>>>>>>> username>/hivefolder</value> > >> >>> >> >> >>>>>>>>>>>> <description>location of default > >> >>> >> >> >>>>>>>>>>>> database > >> >>> >> >> >>>>>>>>>>>> for > >> >>> >> >> >>>>>>>>>>>> the > >> >>> >> >> >>>>>>>>>>>> warehouse</description> > >> >>> >> >> >>>>>>>>>>>> </property> > >> >>> >> >> >>>>>>>>>>>> > >> >>> >> >> >>>>>>>>>>>> > >> >>> >> >> >>>>>>>>>>>> On Tue, Jun 5, 2012 at 10:43 PM, Babak Bastan > >> >>> >> >> >>>>>>>>>>>> <babak...@gmail.com> wrote: > >> >>> >> >> >>>>>>>>>>>>> > >> >>> >> >> >>>>>>>>>>>>> Hello Experts , > >> >>> >> >> >>>>>>>>>>>>> > >> >>> >> >> >>>>>>>>>>>>> I'm new in Hive .When try to create a test Table > >> >>> >> >> >>>>>>>>>>>>> in > >> >>> >> >> >>>>>>>>>>>>> Hive > >> >>> >> >> >>>>>>>>>>>>> I > >> >>> >> >> >>>>>>>>>>>>> get > >> >>> >> >> >>>>>>>>>>>>> an error.I want to run this command: > >> >>> >> >> >>>>>>>>>>>>> CREATE TABLE Test (DateT STRING, Url STRING, > >> >>> >> >> >>>>>>>>>>>>> Content > >> >>> >> >> >>>>>>>>>>>>> STRING); > >> >>> >> >> >>>>>>>>>>>>> but this error occured: > >> >>> >> >> >>>>>>>>>>>>> FAILED: Error in metadata: > >> >>> >> >> >>>>>>>>>>>>> MetaException(message:Got > >> >>> >> >> >>>>>>>>>>>>> exception: > >> >>> >> >> >>>>>>>>>>>>> java.io.FileNotFoundException File > >> >>> >> >> >>>>>>>>>>>>> file:/user/hive/warehouse/test does not > >> >>> >> >> >>>>>>>>>>>>> exist.) > >> >>> >> >> >>>>>>>>>>>>> FAILED: Execution Error, return code 1 from > >> >>> >> >> >>>>>>>>>>>>> org.apache.hadoop.hive.ql.exec.DDLTask > >> >>> >> >> >>>>>>>>>>>>> How can I solve this Problem? > >> >>> >> >> >>>>>>>>>>>>> Thank you so much > >> >>> >> >> >>>>>>>>>>>> > >> >>> >> >> >>>>>>>>>>>> > >> >>> >> >> >>>>>>>>>>>> > >> >>> >> >> >>>>>>>>>>>> > >> >>> >> >> >>>>>>>>>>>> -- > >> >>> >> >> >>>>>>>>>>>> > >> >>> >> >> >>>>>>>>>>>> > >> >>> >> >> >>>>>>>>>>>> ∞ > >> >>> >> >> >>>>>>>>>>>> > >> >>> >> >> >>>>>>>>>>>> Shashwat Shriparv > >> >>> >> >> >>>>>>>>>>>> > >> >>> >> >> >>>>>>>>>>>> > >> >>> >> >> >>>>>>>>>>> > >> >>> >> >> >>>>>>>>>> > >> >>> >> >> >>>>>>>>>> > >> >>> >> >> >>>>>>>>>> > >> >>> >> >> >>>>>>>>>> -- > >> >>> >> >> >>>>>>>>>> > >> >>> >> >> >>>>>>>>>> > >> >>> >> >> >>>>>>>>>> ∞ > >> >>> >> >> >>>>>>>>>> > >> >>> >> >> >>>>>>>>>> Shashwat Shriparv > >> >>> >> >> >>>>>>>>>> > >> >>> >> >> >>>>>>>>>> > >> >>> >> >> >>>>>>>>> > >> >>> >> >> >>>>>>>> > >> >>> >> >> >>>>>>> > >> >>> >> >> >>>>>>> > >> >>> >> >> >>>>>>> > >> >>> >> >> >>>>>>> -- > >> >>> >> >> >>>>>>> > >> >>> >> >> >>>>>>> > >> >>> >> >> >>>>>>> ∞ > >> >>> >> >> >>>>>>> > >> >>> >> >> >>>>>>> Shashwat Shriparv > >> >>> >> >> >>>>>>> > >> >>> >> >> >>>>>>> > >> >>> >> >> >>>>>> > >> >>> >> >> >>>>> > >> >>> >> >> >>>> > >> >>> >> >> >>> > >> >>> >> >> >>> > >> >>> >> >> >>> > >> >>> >> >> >>> -- > >> >>> >> >> >>> > >> >>> >> >> >>> > >> >>> >> >> >>> ∞ > >> >>> >> >> >>> > >> >>> >> >> >>> Shashwat Shriparv > >> >>> >> >> >>> > >> >>> >> >> >>> > >> >>> >> >> >> > >> >>> >> > > >> >>> >> > > >> >>> > > >> >>> > > >> >> > >> >> > > > > >