hmm .. there's a possibility that the execute and close for a given statement are not handled by same thrift thread. Can you please verify the testcase with a single worker thread on the server side ? You can run the server using following : hive --service hiveserver --maxWorkerThreads 1 --minWorkerThreads 1
thanks Prasad On Wed, Feb 6, 2013 at 3:44 AM, Gabor Makrai <makrai.l...@gmail.com> wrote: > Hi guys, > > Bad news for me. I checked out and compiled the Hive trunk and got the > same problem. > I attached to output of command lsof before and after my test program with > 100 "SHOW TABLES" iterations. Is there any explanation why my JDBC server > process doesn't release those files? > > Thanks, > Gabor > > > On Tue, Feb 5, 2013 at 6:20 AM, 王锋 <wfeng1...@163.com> wrote: > >> >> >> I got it. pls see https://issues.apache.org/jira/browse/THRIFT-1205 >> >> I upgrade the thrift to libthrift-0.9.0. >> >> thanks >> >> >> >> >> At 2013-02-05 13:06:05,"王锋" <wfeng1...@163.com> wrote: >> >> >> When I was using hiveserver ,the exception was thrown: >> >> 2060198 Hive history >> file=/tmp/hdfs/hive_job_log_hdfs_201302010032_1918750748.txt >> 2060199 Exception in thread "pool-1-thread-95" >> java.lang.OutOfMemoryError: Java heap space >> 2060200 at >> org.apache.thrift.protocol.TBinaryProtocol.readStringBody(TBinaryProtocol.java:353) >> 2060201 at >> org.apache.thrift.protocol.TBinaryProtocol.readMessageBegin(TBinaryProtocol.java:215) >> 2060202 at >> org.apache.hadoop.hive.service.ThriftHive$Processor.process(ThriftHive.java:730) >> 2060203 at >> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:253) >> 2060204 at >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1110) >> 2060205 at >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603) >> 2060206 at java.lang.Thread.run(Thread.java:722) >> >> I using Hive-0.7.1-cdh3u1 with thrift-0.5.0.jar and >> thrift-fb303-0.5.0.jar. >> how can it be fixed? how about hive-0.7-1 using thrift -0.9.0? thanks. >> >> >> >> >> At 2013-02-04 19:19:16,"Bennie Schut" <bsc...@ebuddy.com> wrote: >> >> Looking at the versions you might be hitting >> https://issues.apache.org/jira/browse/HIVE-3481 which is fixed in 0.10*** >> * >> >> ** ** >> >> On my dev machine the test runs with success :Running time: 298.952409914 >> **** >> >> This includes this patch so it’s worth looking at.**** >> >> ** ** >> >> *From:* Gabor Makrai [mailto:makrai.l...@gmail.com] >> *Sent:* Monday, February 04, 2013 11:58 AM >> *To:* user@hive.apache.org >> *Subject:* Re: Problem with Hive JDBC server**** >> >> ** ** >> >> Yes, of course! I attached the code!**** >> >> ** ** >> >> On Mon, Feb 4, 2013 at 11:57 AM, Gabor Makrai <makrai.l...@gmail.com> >> wrote:**** >> >> Yes, of course! :) I attached the code!**** >> >> ** ** >> >> On Mon, Feb 4, 2013 at 11:53 AM, Bennie Schut <bsc...@ebuddy.com> wrote:* >> *** >> >> Since it’s small can you post the code?**** >> >> **** >> >> *From:* Gabor Makrai [mailto:makrai.l...@gmail.com] >> *Sent:* Monday, February 04, 2013 11:45 AM >> *To:* user@hive.apache.org >> *Subject:* Problem with Hive JDBC server**** >> >> **** >> >> Hi guys,**** >> >> **** >> >> I'm writing you because I experienced a very strange problem which >> probably affects all Hive distribution.**** >> >> I made a small "only main function" Java program where I'm only >> connecting to my Hive JDBC, and getting the list of the database tables >> (LIST TABLES) and closing the ResultSet, the Statement and the Connection >> and doing this a 1000 times. The problem is that the running Hive JDBC >> server does not release files and with time it will throw Exception >> because, it will get "Too many open files" IOException from the JVM.**** >> >> **** >> >> I tested with Hive 0.9, 0.8.1, and the patched Hive 0.9 installed in >> CDH4.1.1.**** >> >> **** >> >> If it is a know issue, than could you tell me the solution for it? If it >> is not, than I can create a new ticket in Jira, and with a little help, I >> probably can fix the problem and contribute the solution for it.**** >> >> **** >> >> Thanks,**** >> >> Gabor**** >> >> ** ** >> >> ** ** >> >> >> >> >> >> >