[ https://issues.apache.org/jira/browse/HIVE-16719?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16017655#comment-16017655 ]
Zsolt Fekete commented on HIVE-16719: ------------------------------------- Note that during testing my refactor and modification I found that it is likely that updating TABLE_PARAMS, SD_PARAMS, SERDE_PARAMS did not work previously, because these calls in ObjectStore: {{mSerde.getParameters().put(serdeProp, newSchemaLoc);}} does not have any affect. One should create a new map and call {{mSerde.setParameters(..);}} with that. So I also fixed this issue. > HiveMetaTool fails when the data does not fit in memory > ------------------------------------------------------- > > Key: HIVE-16719 > URL: https://issues.apache.org/jira/browse/HIVE-16719 > Project: Hive > Issue Type: Bug > Components: Metastore > Reporter: Zsolt Fekete > Assignee: Zsolt Fekete > Attachments: HIVE-16719.1.patch > > > Currently HiveMetaTool reads full tables (as DataNucleus entities) into > memory by calling PersistenceManager's retrieveAll(). > See these methods of ObjectStore: updateMDatabaseURI, updateTblPropURI, > updateMStorageDescriptorTblPropURI, updateMStorageDescriptorTblURI, > updateSerdeURI. > This might cause failure when the affected tables (SDS, DBS, TABLE_PARAMS, > SD_PARAMS, SERDES, SERDE_PARAMS) are too big. -- This message was sent by Atlassian JIRA (v6.3.15#6346)