thanks for concern but it gets resolved actually a problem was.. */var/log/hadoop/userlogs
maximum size due to which task tracker is not be able to write any log on any server. * On Tue, Jul 19, 2011 at 10:53 PM, Viral Bajaria <[email protected]>wrote: > Vikas, > > I don't think it's the ping from name-node that is an issue here, you > should run a ping command from data-node to all data-nodes/name-node. > > Thanks, > Viral > > On Tue, Jul 19, 2011 at 6:50 AM, Edward Capriolo <[email protected]>wrote: > >> >> >> On Tue, Jul 19, 2011 at 9:46 AM, Vikas Srivastava < >> [email protected]> wrote: >> >>> Hey Edward, >>> >>> thanks for responding but i try to ping all the *data-node* from * >>> name-node* and they all are responding.. >>> >>> i won't be able to figure it out where the problem persist. >>> >>> query is running fine when i dont use any map reduce.. but while using >>> and map tasks...its get stuck into that.. >>> >>> Regards >>> Vikas Srivastava >>> 9560885900 >>> >>> >>> >>> >>> >>> On Tue, Jul 19, 2011 at 7:03 PM, Edward Capriolo >>> <[email protected]>wrote: >>> >>>> It must be a hostname or DNS problem. Use dig and ping to find out what >>>> is wrong. >>>> >>>> On Tue, Jul 19, 2011 at 9:05 AM, Vikas Srivastava < >>>> [email protected]> wrote: >>>> >>>>> >>>>> >>>>> On Tue, Jul 19, 2011 at 6:29 PM, Vikas Srivastava < >>>>> [email protected]> wrote: >>>>> >>>>>> >>>>>> HI Team, >>>>>>> >>>>>>> >>>>>>> we are using 1 namenode with 11 Datanode each of (16GB ram and 1.4 tb >>>>>>> hdd) >>>>>>> >>>>>>> i m getting this error while running any query , simple its not >>>>>>> working when we use any map tasks. >>>>>>> >>>>>>> and we are using hive on hadoop. >>>>>>> >>>>>>> Total MapReduce jobs = 1 >>>>>>> Launching Job 1 out of 1 >>>>>>> Number of reduce tasks not specified. Estimated from input data size: >>>>>>> 120 >>>>>>> In order to change the average load for a reducer (in bytes): >>>>>>> set hive.exec.reducers.bytes.per.reducer=<number> >>>>>>> In order to limit the maximum number of reducers: >>>>>>> set hive.exec.reducers.max=<number> >>>>>>> In order to set a constant number of reducers: >>>>>>> set mapred.reduce.tasks=<number> >>>>>>> Starting Job = job_201107191711_0013, Tracking URL = >>>>>>> http://hadoopname:50030/jobdetails.jsp?jobid=job_201107191711_0013 >>>>>>> Kill Command = /home/hadoop/hadoop/bin/../bin/hadoop job >>>>>>> -Dmapred.job.tracker=10.0.3.28:9001 -kill job_201107191711_0013 >>>>>>> 2011-07-19 18:06:34,973 Stage-1 map = 100%, reduce = 100% >>>>>>> Ended Job = job_201107191711_0013 with errors >>>>>>> java.lang.RuntimeException: Error while reading from task log url >>>>>>> at >>>>>>> org.apache.hadoop.hive.ql.exec.errors.TaskLogProcessor.getErrors(TaskLogProcessor.java:130) >>>>>>> at >>>>>>> org.apache.hadoop.hive.ql.exec.ExecDriver.showJobFailDebugInfo(ExecDriver.java:889) >>>>>>> at >>>>>>> org.apache.hadoop.hive.ql.exec.ExecDriver.execute(ExecDriver.java:680) >>>>>>> at >>>>>>> org.apache.hadoop.hive.ql.exec.MapRedTask.execute(MapRedTask.java:123) >>>>>>> at >>>>>>> org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:130) >>>>>>> at >>>>>>> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:57) >>>>>>> at >>>>>>> org.apache.hadoop.hive.ql.exec.TaskRunner.run(TaskRunner.java:47) >>>>>>> Caused by: java.net.UnknownHostException: hadoopdata3 >>>>>>> at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:177) >>>>>>> at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:366) >>>>>>> at java.net.Socket.connect(Socket.java:519) >>>>>>> at java.net.Socket.connect(Socket.java:469) >>>>>>> at sun.net.NetworkClient.doConnect(NetworkClient.java:163) >>>>>>> at >>>>>>> sun.net.www.http.HttpClient.openServer(HttpClient.java:394) >>>>>>> at >>>>>>> sun.net.www.http.HttpClient.openServer(HttpClient.java:529) >>>>>>> at sun.net.www.http.HttpClient.<init>(HttpClient.java:233) >>>>>>> at sun.net.www.http.HttpClient.New(HttpClient.java:306) >>>>>>> at sun.net.www.http.HttpClient.New(HttpClient.java:323) >>>>>>> at >>>>>>> sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:837) >>>>>>> at >>>>>>> sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:778) >>>>>>> at >>>>>>> sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.java:703) >>>>>>> at >>>>>>> sun.net.www.protocol.http.HttpURLConnection.getInputStream(HttpURLConnection.java:1026) >>>>>>> at java.net.URL.openStream(URL.java:1009) >>>>>>> at >>>>>>> org.apache.hadoop.hive.ql.exec.errors.TaskLogProcessor.getErrors(TaskLogProcessor.java:120) >>>>>>> ... 6 more >>>>>>> Ended Job = job_201107191711_0013 with exception >>>>>>> 'java.lang.RuntimeException(Error while reading from task log url)' >>>>>>> FAILED: Execution Error, return code 1 from >>>>>>> org.apache.hadoop.hive.ql.exec.MapRedTask >>>>>>> >>>>>>> -- >>>>>>> With Regards >>>>>>> Vikas Srivastava >>>>>>> >>>>>>> DWH & Analytics Team >>>>>>> Mob:+91 9560885900 >>>>>>> One97 | Let's get talking ! >>>>>>> >>>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> With Regards >>>>>> Vikas Srivastava >>>>>> >>>>>> DWH & Analytics Team >>>>>> Mob:+91 9560885900 >>>>>> One97 | Let's get talking ! >>>>>> >>>>>> >>>>> >>>>> >>>>> -- >>>>> With Regards >>>>> Vikas Srivastava >>>>> >>>>> DWH & Analytics Team >>>>> Mob:+91 9560885900 >>>>> One97 | Let's get talking ! >>>>> >>>>> >>>> >>> >>> >>> -- >>> With Regards >>> Vikas Srivastava >>> >>> DWH & Analytics Team >>> Mob:+91 9560885900 >>> One97 | Let's get talking ! >>> >>> >> Try again. >> >> >> Caused by: java.net.UnknownHostException: hadoopdata3 >> >> This clearly indicates that some of your nodes are not able to reach each >> other. Check your DNS, check your systems hostname and make sure it matches >> DNS. Check your hostname, check your host file, check your resolver settings >> including your search domain. One machine is trying to contact hadoopdata3 >> and is not finding it in a DNS lookup. > > > -- With Regards Vikas Srivastava DWH & Analytics Team Mob:+91 9560885900 One97 | Let's get talking !
