when you ran the query, did the VM shutdown ?
On Wed, Jan 23, 2013 at 12:57 PM, venkatramanan < venkatraman...@smartek21.com> wrote: > Hi, > > I got the following error while executing the "select count(1) from > tweettrend;" > > Below are the exact log msg from the jobtracker Web Interface > > *Hive Cli Error:* > > Exception in thread "Thread-21" java.lang.RuntimeException: Error while > reading from task log url > at > org.apache.hadoop.hive.ql.exec.errors.TaskLogProcessor.getStackTraces(TaskLogProcessor.java:240) > at > org.apache.hadoop.hive.ql.exec.JobDebugger.showJobFailDebugInfo(JobDebugger.java:227) > at org.apache.hadoop.hive.ql.exec.JobDebugger.run(JobDebugger.java:92) > at java.lang.Thread.run(Thread.java:722) > Caused by: java.net.UnknownHostException: savitha-VirtualBox > at > java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:178) > at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:391) > at java.net.Socket.connect(Socket.java:579) > at java.net.Socket.connect(Socket.java:528) > at sun.net.NetworkClient.doConnect(NetworkClient.java:180) > at sun.net.www.http.HttpClient.openServer(HttpClient.java:378) > at sun.net.www.http.HttpClient.openServer(HttpClient.java:473) > at sun.net.www.http.HttpClient.<init>(HttpClient.java:203) > at sun.net.www.http.HttpClient.New(HttpClient.java:290) > at sun.net.www.http.HttpClient.New(HttpClient.java:306) > at > sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:995) > at > sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:931) > at > sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.java:849) > at > sun.net.www.protocol.http.HttpURLConnection.getInputStream(HttpURLConnection.java:1299) > at java.net.URL.openStream(URL.java:1037) > at > org.apache.hadoop.hive.ql.exec.errors.TaskLogProcessor.getStackTraces(TaskLogProcessor.java:192) > ... 3 more > FAILED: Execution Error, return code 2 from > org.apache.hadoop.hive.ql.exec.MapRedTask > MapReduce Jobs Launched: > Job 0: Map: 2 Reduce: 1 Cumulative CPU: 9.0 sec HDFS Read: 408671053 > HDFS Write: 0 FAIL > Total MapReduce CPU Time Spent: 9 seconds 0 msec > > *syslog logs* > > utCopier.copyOutput(ReduceTask.java:1394) > at > org.apache.hadoop.mapred.ReduceTask$ReduceCopier$MapOutputCopier.run(ReduceTask.java:1326) > > 2013-01-23 12:15:44,884 INFO org.apache.hadoop.mapred.ReduceTask: Task > attempt_201301231151_0002_r_000000_0: Failed fetch #10 from > attempt_201301231151_0002_m_000001_0 > 2013-01-23 12:15:44,884 INFO org.apache.hadoop.mapred.ReduceTask: Failed to > fetch map-output from attempt_201301231151_0002_m_000001_0 even after > MAX_FETCH_RETRIES_PER_MAP retries... or it is a read error, reporting to > the JobTracker > 2013-01-23 12:15:44,885 FATAL org.apache.hadoop.mapred.ReduceTask: Shuffle > failed with too many fetch failures and insufficient progress!Killing task > attempt_201301231151_0002_r_000000_0. > 2013-01-23 12:15:44,889 WARN org.apache.hadoop.mapred.ReduceTask: > attempt_201301231151_0002_r_000000_0 adding host savitha-VirtualBox to > penalty box, next contact in 137 seconds > 2013-01-23 12:15:44,889 INFO org.apache.hadoop.mapred.ReduceTask: > attempt_201301231151_0002_r_000000_0: Got 1 map-outputs from previous failures > 2013-01-23 12:15:45,218 FATAL org.apache.hadoop.mapred.Task: > attempt_201301231151_0002_r_000000_0 GetMapEventsThread Ignoring exception : > org.apache.hadoop.ipc.RemoteException: java.io.IOException: JvmValidate > Failed. Ignoring request from task: attempt_201301231151_0002_r_000000_0, > with JvmId: jvm_201301231151_0002_r_1079250852 > at > org.apache.hadoop.mapred.TaskTracker.validateJVM(TaskTracker.java:3278) > at > org.apache.hadoop.mapred.TaskTracker.getMapCompletionEvents(TaskTracker.java:3537) > at sun.reflect.GeneratedMethodAccessor3.invoke(Unknown Source) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:601) > at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:563) > at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1388) > at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1384) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:415) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121) > at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1382) > > at org.apache.hadoop.ipc.Client.call(Client.java:1070) > at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:225) > at $Proxy1.getMapCompletionEvents(Unknown Source) > at > org.apache.hadoop.mapred.ReduceTask$ReduceCopier$GetMapEventsThread.getMapCompletionEvents(ReduceTask.java:2846) > at > org.apache.hadoop.mapred.ReduceTask$ReduceCopier$GetMapEventsThread.run(ReduceTask.java:2810) > > 2013-01-23 12:15:45,220 FATAL org.apache.hadoop.mapred.Task: Failed to > contact the tasktracker > org.apache.hadoop.ipc.RemoteException: java.io.IOException: JvmValidate > Failed. Ignoring request from task: attempt_201301231151_0002_r_000000_0, > with JvmId: jvm_201301231151_0002_r_1079250852 > at > org.apache.hadoop.mapred.TaskTracker.validateJVM(TaskTracker.java:3278) > at > org.apache.hadoop.mapred.TaskTracker.fatalError(TaskTracker.java:3520) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:601) > at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:563) > at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1388) > at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1384) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:415) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121) > at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1382) > > at org.apache.hadoop.ipc.Client.call(Client.java:1070) > at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:225) > at $Proxy1.fatalError(Unknown Source) > at org.apache.hadoop.mapred.Task.reportFatalError(Task.java:298) > at > org.apache.hadoop.mapred.ReduceTask$ReduceCopier$GetMapEventsThread.run(ReduceTask.java:2829) > > thanks, > Venkat > > -------- Original Message -------- Subject: Re: Configure Hive in Cluster > Date: > Thu, 17 Jan 2013 17:23:03 +0530 From: venkatramanan > <venkatraman...@smartek21.com> <venkatraman...@smartek21.com> Reply-To: > <user@hive.apache.org> <user@hive.apache.org> To: > <user@hive.apache.org><user@hive.apache.org> > > > Can you suggest me the mandatory hive parameters and clustering > configuration steps > > On Thursday 17 January 2013 12:56 PM, Nitin Pawar wrote: > > looks like a very small cluster with very limited memory to run mapreduce > jobs also number of map/reduce slots on nodes are less so at a time only > one map is running. > > but still 15 min is a lot of time for 600MB memory > > > On Thu, Jan 17, 2013 at 12:47 PM, venkatramanan < > venkatraman...@smartek21.com> wrote: > >> Below details are the cluster configuration >> >> Configured Capacity : 82.8 GB >> DFS Used : 1.16 GB >> Non DFS Used : 31.95 GB >> DFS Remaining : 49.69 GB >> DFS Used% : 1.4 % >> DFS Remaining% : 60.01 % >> Live Nodes <http://localhost:50070/dfsnodelist.jsp?whatNodes=LIVE> >> : 2 >> Dead Nodes <http://localhost:50070/dfsnodelist.jsp?whatNodes=DEAD> >> : 0 >> Decommissioning >> Nodes<http://localhost:50070/dfsnodelist.jsp?whatNodes=DECOMMISSIONING>: 0 >> Number of Under-Replicated Blocks : 0 >> >> My Select Query is: >> >> "select * from tweet where Id = 810;" >> >> This query takes 15 min to complete >> >> >> >> On Thursday 17 January 2013 12:29 PM, Nitin Pawar wrote: >> >> how many number of nodes you have for select query? >> whats your select query? >> >> if its just a select * from table then it does not run any mapreduce job >> so its just taking time to show data on your screen if you are using >> that query >> >> >> On Thu, Jan 17, 2013 at 12:24 PM, venkatramanan < >> venkatraman...@smartek21.com> wrote: >> >>> I didnt set any hive parameters and my total table size is 610 MB only >>> >>> >>> >>> On Thursday 17 January 2013 12:11 PM, Nitin Pawar wrote: >>> >>> a bit more details on size of table and select query will help >>> also did you set any hive parameters ? >>> >>> >>> On Thu, Jan 17, 2013 at 12:12 PM, venkatramanan < >>> venkatraman...@smartek21.com> wrote: >>> >>>> Hi All, >>>> >>>> Am Newbie in apache hive. I have create a table and thats points to the >>>> HDFS Folder path and its takes 15 min to execute the simple "*select*" >>>> stmt, Can anyone suggest me for a best practices and performance >>>> improvement on hive. >>>> >>>> Thanks in Advance >>>> >>>> Venkat >>>> >>> >>> >>> >>> -- >>> Nitin Pawar >>> >>> >>> >> >> >> -- >> Nitin Pawar >> >> >> > > > -- > Nitin Pawar > > > > > -- Nitin Pawar