Sorry for all of the messages! Stupid gmail client.

It turns out that I was able to resolve the failures by setting
"hive.merge.mapfiles=false" (by default it is true).

I think this is because I'm running Hadoop version 0.20.1, r810220.

This thread is what lead me to try this:
http://www.mail-archive.com/hive-u...@hadoop.apache.org/msg03955.html


On Tue, Dec 7, 2010 at 12:26 PM, Ryan LeCompte <lecom...@gmail.com> wrote:

> Digging even further, here's what I see:
>
> NOTE: We have a table in Hive called "test_table" but this seems to look
> for "default.test_table" ? )
>
> 2010-12-07 00:52:24,600 ERROR metadata.Hive (Hive.java:getTable(357)) -
> NoSuchObjectException(message:default.test_table table not found)
>         at
> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.get_table(HiveMetaStore.java:399)
>         at
> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getTable(HiveMetaStoreClient.java:432)
>         at org.apache.hadoop.hive.ql.metadata.Hive.getTable(Hive.java:354)
>         at org.apache.hadoop.hive.ql.metadata.Hive.getTable(Hive.java:333)
>         at
> org.apache.hadoop.hive.ql.exec.DDLTask.dropTable(DDLTask.java:1088)
>         at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:129)
>         at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:99)
>         at
> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:64)
>         at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:582)
>         at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:462)
>         at org.apache.hadoop.hive.ql.Driver.runCommand(Driver.java:324)
>         at org.apache.hadoop.hive.ql.Driver.run(Driver.java:312)
>         at
> org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:123)
>
>
>
> On Tue, Dec 7, 2010 at 12:19 PM, Ryan LeCompte <lecom...@gmail.com> wrote:
>
>> I just put the Hive log4j config file on DEBUG, and here is the error that
>> I'm seeing:
>>
>> 2010-12-07 12:16:50,281 WARN  mapred.JobClient
>> (JobClient.java:configureCommandLineOptions(539)) - Use GenericOptionsParser
>> for parsing the arguments. Applications should implement Tool for the same.
>> 2010-12-07 12:17:26,399 DEBUG conf.Configuration
>> (Configuration.java:<init>(210)) - java.io.IOException: config()
>> at org.apache.hadoop.conf.Configuration.<init>(Configuration.java:210)
>>  at org.apache.hadoop.conf.Configuration.<init>(Configuration.java:197)
>> at org.apache.hadoop.hive.conf.HiveConf.<init>(HiveConf.java:429)
>>  at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:230)
>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>  at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>> at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>  at java.lang.reflect.Method.invoke(Method.java:597)
>> at org.apache.hadoop.util.RunJar.main(RunJar.java:156)
>>
>> 2010-12-07 12:17:26,405 DEBUG conf.Configuration
>> (Configuration.java:<init>(210)) - java.io.IOException: config()
>> at org.apache.hadoop.conf.Configuration.<init>(Configuration.java:210)
>>  at org.apache.hadoop.conf.Configuration.<init>(Configuration.java:197)
>> at org.apache.hadoop.mapred.JobConf.<init>(JobConf.java:172)
>>  at org.apache.hadoop.hive.conf.HiveConf.initialize(HiveConf.java:449)
>> at org.apache.hadoop.hive.conf.HiveConf.<init>(HiveConf.java:430)
>>  at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:230)
>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>  at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>> at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>  at java.lang.reflect.Method.invoke(Method.java:597)
>> at org.apache.hadoop.util.RunJar.main(RunJar.java:156)
>>
>>
>>
>> On Tue, Dec 7, 2010 at 12:12 PM, Ryan LeCompte <lecom...@gmail.com>wrote:
>>
>>> Hello all,
>>>
>>> This morning we upgraded our cluster to use Hive 0.6.0 from Hive 0.5.0. I
>>> ran the Derby DB migration script before upgrading.
>>>
>>> All of our queries are running fine, however our INSERT OVERWRITE queries
>>> are failing immediately. I am not sure what the problem is.
>>>
>>> When we submit an INSERT OVERWRITE query, the query fails with the maps
>>> in "Pending". Here is the failure from the web interface:
>>>
>>> HTTP ERROR 500
>>>
>>>  java.lang.ArrayIndexOutOfBoundsException: 0
>>>     at 
>>> org.apache.hadoop.mapred.JobInProgress.getTaskInProgress(JobInProgress.java:2523)
>>>     at 
>>> org.apache.hadoop.mapred.taskdetails_jsp._jspService(taskdetails_jsp.java:115)
>>>     at org.apache.jasper.runtime.HttpJspBase.service(HttpJspBase.java:97)
>>>     at javax.servlet.http.HttpServlet.service(HttpServlet.java:820)
>>>     at 
>>> org.mortbay.jetty.servlet.ServletHolder.handle(ServletHolder.java:502)
>>>     at 
>>> org.mortbay.jetty.servlet.ServletHandler.handle(ServletHandler.java:363)
>>>     at 
>>> org.mortbay.jetty.security.SecurityHandler.handle(SecurityHandler.java:216)
>>>     at 
>>> org.mortbay.jetty.servlet.SessionHandler.handle(SessionHandler.java:181)
>>>     at 
>>> org.mortbay.jetty.handler.ContextHandler.handle(ContextHandler.java:766)
>>>     at org.mortbay.jetty.webapp.WebAppContext.handle(WebAppContext.java:417)
>>>     at 
>>> org.mortbay.jetty.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:230)
>>>     at 
>>> org.mortbay.jetty.handler.HandlerWrapper.handle(HandlerWrapper.java:152)
>>>     at org.mortbay.jetty.Server.handle(Server.java:324)
>>>     at 
>>> org.mortbay.jetty.HttpConnection.handleRequest(HttpConnection.java:534)
>>>     at 
>>> org.mortbay.jetty.HttpConnection$RequestHandler.headerComplete(HttpConnection.java:864)
>>>     at org.mortbay.jetty.HttpParser.parseNext(HttpParser.java:533)
>>>     at org.mortbay.jetty.HttpParser.parseAvailable(HttpParser.java:207)
>>>     at org.mortbay.jetty.HttpConnection.handle(HttpConnection.java:403)
>>>     at 
>>> org.mortbay.io.nio.SelectChannelEndPoint.run(SelectChannelEndPoint.java:409)
>>>     at 
>>> org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:522)
>>>
>>>
>>> Does anyone have any ideas?
>>>
>>>
>>> Thanks,
>>>
>>> Ryan
>>>
>>>
>>>
>>
>

Reply via email to