Hi,

I was running on a 0.7 trunk build from February 2011 until last Friday and
upgraded to trunk again then.

Things  works ok except memory usage when doing queries with large number of
partitions is quite dramatically up.
I could query 12 months of data in one table with ~72k of partitions with 4
buckets each without trouble before with a HEAP of 768MB, I have tried with
a HEAP of up to 4GB with the new 0.7 build but still I fail.

I don't see a lot of other people complain, so I wonder if I just have
unusually many partitions or if there is a bug here or I have done something
stupid in the upgrade.

Exception in thread "main" java.lang.OutOfMemoryError: Java heap space
        at java.util.ArrayList.<init>(ArrayList.java:112)
        at java.util.ArrayList.<init>(ArrayList.java:119)
        at org.datanucleus.store.rdbms.sql.SQLText.append(SQLText.java:158)
        at
org.datanucleus.store.rdbms.sql.SQLStatement.getSelectStatement(SQLStatement.java:1282)
        at
org.datanucleus.store.rdbms.scostore.RDBMSJoinListStore.listIterator(RDBMSJoinListStore.java:116)
        at
org.datanucleus.store.mapped.scostore.AbstractListStore.listIterator(AbstractListStore.java:84)
        at
org.datanucleus.store.mapped.scostore.AbstractListStore.iterator(AbstractListStore.java:74)
        at org.datanucleus.sco.backed.List.loadFromStore(List.java:241)
        at org.datanucleus.sco.backed.List.iterator(List.java:507)
        at
org.apache.hadoop.hive.metastore.ObjectStore.convertToFieldSchemas(ObjectStore.java:866)
        at
org.apache.hadoop.hive.metastore.ObjectStore.convertToStorageDescriptor(ObjectStore.java:920)
        at
org.apache.hadoop.hive.metastore.ObjectStore.convertToStorageDescriptor(ObjectStore.java:929)
        at
org.apache.hadoop.hive.metastore.ObjectStore.convertToPart(ObjectStore.java:1063)
        at
org.apache.hadoop.hive.metastore.ObjectStore.getPartitionsWithAuth(ObjectStore.java:1138)
        at
org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler$25.run(HiveMetaStore.java:1595)
        at
org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler$25.run(HiveMetaStore.java:1592)
        at
org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.executeWithRetry(HiveMetaStore.java:319)
        at
org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.get_partitions_with_auth(HiveMetaStore.java:1592)
        at
org.apache.hadoop.hive.metastore.HiveMetaStoreClient.listPartitionsWithAuthInfo(HiveMetaStoreClient.java:593)
        at
org.apache.hadoop.hive.ql.metadata.Hive.getPartitions(Hive.java:1456)
        at
org.apache.hadoop.hive.ql.optimizer.ppr.PartitionPruner.prune(PartitionPruner.java:207)
        at
org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genMapRedTasks(SemanticAnalyzer.java:6583)
        at
org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:7113)
        at
org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:238)
        at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:376)
        at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:334)
        at org.apache.hadoop.hive.ql.Driver.run(Driver.java:843)
        at
org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:218)
        at
org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:302)
        at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:531)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)


Regards,
Terje

Reply via email to