you can try hive 0.5, after create the metadata, use upgrade sql file in hive 0.7.1 to upgrade to 0.7.1
On Sat, Aug 20, 2011 at 2:20 PM, Xiaobo Gu <guxiaobo1...@gmail.com> wrote: > Hi, > I have just set up a PostgreSQL 9.0.2 server for hive 0.7.1 metastore, > and I am using the postgresql-9.0-801.jdbc4.jar jdbc driver, when I > test the following HQL, > > > CREATE TABLE records (year STRING, temperature INT, quality INT) > ROW FORMAT DELIMITED > FIELDS TERMINATED BY '\t'; > > > the following error messages are thrown in metastore services standard error: > > 11/08/20 10:12:33 INFO metastore.HiveMetaStore: 1: create_table: > db=default tbl=records > 11/08/20 10:12:33 INFO HiveMetaStore.audit: ugi=gpadmin > ip=/192.168.72.6 cmd=create_table: db=default tbl=records > 11/08/20 10:12:33 INFO metastore.HiveMetaStore: 1: Opening raw store > with implemenation class:org.apache.hadoop.hive.metastore.ObjectStore > 11/08/20 10:12:33 INFO metastore.ObjectStore: ObjectStore, initialize called > 11/08/20 10:12:33 INFO metastore.ObjectStore: Initialized ObjectStore > 11/08/20 10:12:33 INFO DataNucleus.Datastore: The class > "org.apache.hadoop.hive.metastore.model.MFieldSchema" is tagged as > "embedded-only" so does not have its own data > store table. > 11/08/20 10:12:33 INFO DataNucleus.Persistence: Managing Persistence > of Class : org.apache.hadoop.hive.metastore.model.MSerDeInfo [Table : > "SERDES", InheritanceStrategy > : new-table] > 11/08/20 10:12:33 INFO DataNucleus.Datastore: The class > "org.apache.hadoop.hive.metastore.model.MOrder" is tagged as > "embedded-only" so does not have its own datastore > table. > 11/08/20 10:12:33 INFO DataNucleus.Persistence: Managing Persistence > of Class : org.apache.hadoop.hive.metastore.model.MStorageDescriptor > [Table : "SDS", InheritanceStr > ategy : new-table] > 11/08/20 10:12:33 INFO DataNucleus.Persistence: Managing Persistence > of Class : org.apache.hadoop.hive.metastore.model.MTable [Table : > "TBLS", InheritanceStrategy : new > -table] > 11/08/20 10:12:33 INFO DataNucleus.Persistence: Managing Persistence > of Field : org.apache.hadoop.hive.metastore.model.MSerDeInfo.parameters > [Table : "SERDE_PARAMS"] > 11/08/20 10:12:33 INFO DataNucleus.Persistence: Managing Persistence > of Field : org.apache.hadoop.hive.metastore.model.MTable.parameters > [Table : "TABLE_PARAMS"] > 11/08/20 10:12:33 INFO DataNucleus.Persistence: Managing Persistence > of Field : org.apache.hadoop.hive.metastore.model.MTable.partitionKeys > [Table : "PARTITION_KEYS"] > 11/08/20 10:12:33 INFO DataNucleus.Persistence: Managing Persistence > of Field : > org.apache.hadoop.hive.metastore.model.MStorageDescriptor.bucketCols > [Table : "BUCKETING > _COLS"] > 11/08/20 10:12:33 INFO DataNucleus.Persistence: Managing Persistence > of Field : org.apache.hadoop.hive.metastore.model.MStorageDescriptor.cols > [Table : "COLUMNS"] > 11/08/20 10:12:33 INFO DataNucleus.Persistence: Managing Persistence > of Field : > org.apache.hadoop.hive.metastore.model.MStorageDescriptor.parameters > [Table : "SD_PARAMS > "] > 11/08/20 10:12:33 INFO DataNucleus.Persistence: Managing Persistence > of Field : org.apache.hadoop.hive.metastore.model.MStorageDescriptor.sortCols > [Table : "SORT_COLS"] > 11/08/20 10:12:33 INFO Datastore.Schema: Validating 1 unique key(s) > for table "SERDES" > 11/08/20 10:12:33 INFO Datastore.Schema: Validating 0 foreign key(s) > for table "SERDES" > 11/08/20 10:12:33 INFO Datastore.Schema: Validating 1 index(es) for > table "SERDES" > 11/08/20 10:12:33 INFO Datastore.Schema: Validating 2 unique key(s) > for table "TBLS" > 11/08/20 10:12:33 INFO Datastore.Schema: Validating 2 foreign key(s) > for table "TBLS" > 11/08/20 10:12:33 INFO Datastore.Schema: Validating 4 index(es) for table > "TBLS" > 11/08/20 10:12:33 INFO Datastore.Schema: Validating 1 unique key(s) > for table "SDS" > 11/08/20 10:12:33 INFO Datastore.Schema: Validating 1 foreign key(s) > for table "SDS" > 11/08/20 10:12:33 INFO Datastore.Schema: Validating 2 index(es) for table > "SDS" > 11/08/20 10:12:33 INFO Datastore.Schema: Validating 1 unique key(s) > for table "SORT_COLS" > 11/08/20 10:12:33 INFO Datastore.Schema: Validating 1 foreign key(s) > for table "SORT_COLS" > 11/08/20 10:12:33 INFO Datastore.Schema: Validating 2 index(es) for > table "SORT_COLS" > 11/08/20 10:12:33 INFO Datastore.Schema: Validating 1 unique key(s) > for table "TABLE_PARAMS" > 11/08/20 10:12:33 INFO Datastore.Schema: Validating 1 foreign key(s) > for table "TABLE_PARAMS" > 11/08/20 10:12:33 INFO Datastore.Schema: Validating 2 index(es) for > table "TABLE_PARAMS" > 11/08/20 10:12:33 INFO Datastore.Schema: Validating 1 unique key(s) > for table "SD_PARAMS" > 11/08/20 10:12:33 INFO Datastore.Schema: Validating 1 foreign key(s) > for table "SD_PARAMS" > 11/08/20 10:12:33 INFO Datastore.Schema: Validating 2 index(es) for > table "SD_PARAMS" > 11/08/20 10:12:33 INFO Datastore.Schema: Validating 1 unique key(s) > for table "SERDE_PARAMS" > 11/08/20 10:12:33 INFO Datastore.Schema: Validating 1 foreign key(s) > for table "SERDE_PARAMS" > 11/08/20 10:12:33 INFO Datastore.Schema: Validating 2 index(es) for > table "SERDE_PARAMS" > 11/08/20 10:12:33 INFO Datastore.Schema: Validating 1 unique key(s) > for table "PARTITION_KEYS" > 11/08/20 10:12:33 INFO Datastore.Schema: Validating 1 foreign key(s) > for table "PARTITION_KEYS" > 11/08/20 10:12:33 INFO Datastore.Schema: Validating 2 index(es) for > table "PARTITION_KEYS" > 11/08/20 10:12:33 INFO Datastore.Schema: Validating 1 unique key(s) > for table "COLUMNS" > 11/08/20 10:12:33 INFO Datastore.Schema: Validating 1 foreign key(s) > for table "COLUMNS" > 11/08/20 10:12:33 INFO Datastore.Schema: Validating 2 index(es) for > table "COLUMNS" > 11/08/20 10:12:33 INFO Datastore.Schema: Validating 1 unique key(s) > for table "BUCKETING_COLS" > 11/08/20 10:12:33 INFO Datastore.Schema: Validating 1 foreign key(s) > for table "BUCKETING_COLS" > 11/08/20 10:12:33 INFO Datastore.Schema: Validating 2 index(es) for > table "BUCKETING_COLS" > 11/08/20 10:12:34 INFO DataNucleus.MetaData: Listener found > initialisation for persistable class > org.apache.hadoop.hive.metastore.model.MSerDeInfo > 11/08/20 10:12:34 INFO DataNucleus.MetaData: Listener found > initialisation for persistable class > org.apache.hadoop.hive.metastore.model.MStorageDescriptor > 11/08/20 10:12:34 INFO DataNucleus.MetaData: Listener found > initialisation for persistable class > org.apache.hadoop.hive.metastore.model.MTable > 11/08/20 10:12:34 INFO DataNucleus.MetaData: Listener found > initialisation for persistable class > org.apache.hadoop.hive.metastore.model.MFieldSchema > 11/08/20 10:12:34 WARN Datastore.Persist: Insert of object > "org.apache.hadoop.hive.metastore.model.MStorageDescriptor@46bb05de" > using statement "INSERT INTO "SDS" ("SD_ > ID","NUM_BUCKETS","SERDE_ID","IS_COMPRESSED","OUTPUT_FORMAT","LOCATION","INPUT_FORMAT") > VALUES (?,?,?,?,?,?,?)" failed : ERROR: column "IS_COMPRESSED" is of > type bit bu > t expression is of type boolean > Hint: You will need to rewrite or cast the expression. > Position: 129 > 11/08/20 10:12:34 INFO metastore.hivemetastoressimpl: deleting > hdfs://linuxsvr2/user/hive/warehouse/records > 11/08/20 10:12:34 INFO metastore.hivemetastoressimpl: Deleted the > diretory hdfs://linuxsvr2/user/hive/warehouse/records > 11/08/20 10:12:34 ERROR metastore.HiveMetaStore: JDO datastore error. > Retrying metastore command after 1000 ms (attempt 1 of 1) > 11/08/20 10:12:35 WARN metastore.HiveMetaStore: Location: > hdfs://linuxsvr2/user/hive/warehouse/records specified for > non-external table:records > 11/08/20 10:12:35 WARN Datastore.Persist: Insert of object > "org.apache.hadoop.hive.metastore.model.MStorageDescriptor@7d1c19e6" > using statement "INSERT INTO "SDS" ("SD_ > ID","NUM_BUCKETS","SERDE_ID","IS_COMPRESSED","OUTPUT_FORMAT","LOCATION","INPUT_FORMAT") > VALUES (?,?,?,?,?,?,?)" failed : ERROR: column "IS_COMPRESSED" is of > type bit bu > t expression is of type boolean > Hint: You will need to rewrite or cast the expression. > Position: 129 > 11/08/20 10:12:35 INFO metastore.hivemetastoressimpl: deleting > hdfs://linuxsvr2/user/hive/warehouse/records > 11/08/20 10:12:35 INFO metastore.hivemetastoressimpl: Deleted the > diretory hdfs://linuxsvr2/user/hive/warehouse/records > 11/08/20 10:12:35 ERROR api.ThriftHiveMetastore$Processor: Internal > error processing create_table > javax.jdo.JDODataStoreException: Insert of object > "org.apache.hadoop.hive.metastore.model.MStorageDescriptor@7d1c19e6" > using statement "INSERT INTO "SDS" ("SD_ID","NUM_ > BUCKETS","SERDE_ID","IS_COMPRESSED","OUTPUT_FORMAT","LOCATION","INPUT_FORMAT") > VALUES (?,?,?,?,?,?,?)" failed : ERROR: column "IS_COMPRESSED" is of > type bit but express > ion is of type boolean > Hint: You will need to rewrite or cast the expression. > Position: 129 > at > org.datanucleus.jdo.NucleusJDOHelper.getJDOExceptionForNucleusException(NucleusJDOHelper.java:313) > at > org.datanucleus.jdo.JDOPersistenceManager.jdoMakePersistent(JDOPersistenceManager.java:660) > at > org.datanucleus.jdo.JDOPersistenceManager.makePersistent(JDOPersistenceManager.java:680) > at > org.apache.hadoop.hive.metastore.ObjectStore.createTable(ObjectStore.java:606) > at > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.create_table_core(HiveMetaStore.java:924) > at > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.access$600(HiveMetaStore.java:109) > at > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler$15.run(HiveMetaStore.java:945) > at > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler$15.run(HiveMetaStore.java:942) > at > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.executeWithRetry(HiveMetaStore.java:307) > at > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.create_table(HiveMetaStore.java:942) > at > org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Processor$create_table.process(ThriftHiveMetastore.java:5297) > at > org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Processor.process(ThriftHiveMetastore.java:4789) > at > org.apache.hadoop.hive.metastore.HiveMetaStore$TLoggingProcessor.process(HiveMetaStore.java:3167) > at > org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:253) > at > java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) > at java.lang.Thread.run(Thread.java:662) > NestedThrowablesStackTrace: > org.postgresql.util.PSQLException: ERROR: column "IS_COMPRESSED" is of > type bit but expression is of type boolean > Hint: You will need to rewrite or cast the expression. > Position: 129 > at > org.postgresql.core.v3.QueryExecutorImpl.receiveErrorResponse(QueryExecutorImpl.java:2102) > at > org.postgresql.core.v3.QueryExecutorImpl.processResults(QueryExecutorImpl.java:1835) > at > org.postgresql.core.v3.QueryExecutorImpl.execute(QueryExecutorImpl.java:257) > at > org.postgresql.jdbc2.AbstractJdbc2Statement.execute(AbstractJdbc2Statement.java:500) > at > org.postgresql.jdbc2.AbstractJdbc2Statement.executeWithFlags(AbstractJdbc2Statement.java:388) > at > org.postgresql.jdbc2.AbstractJdbc2Statement.executeUpdate(AbstractJdbc2Statement.java:334) > at > org.apache.commons.dbcp.DelegatingPreparedStatement.executeUpdate(DelegatingPreparedStatement.java:105) > at > org.apache.commons.dbcp.DelegatingPreparedStatement.executeUpdate(DelegatingPreparedStatement.java:105) > at > org.datanucleus.store.rdbms.SQLController.executeStatementUpdate(SQLController.java:396) > at > org.datanucleus.store.rdbms.request.InsertRequest.execute(InsertRequest.java:406) > at > org.datanucleus.store.rdbms.RDBMSPersistenceHandler.insertTable(RDBMSPersistenceHandler.java:146) > at > org.datanucleus.store.rdbms.RDBMSPersistenceHandler.insertObject(RDBMSPersistenceHandler.java:121) > at > org.datanucleus.jdo.state.JDOStateManagerImpl.internalMakePersistent(JDOStateManagerImpl.java:3275) > at > org.datanucleus.jdo.state.JDOStateManagerImpl.makePersistent(JDOStateManagerImpl.java:3249) > at > org.datanucleus.ObjectManagerImpl.persistObjectInternal(ObjectManagerImpl.java:1428) > at > org.datanucleus.store.mapped.mapping.PersistableMapping.setObjectAsValue(PersistableMapping.java:664) > at > org.datanucleus.store.mapped.mapping.PersistableMapping.setObject(PersistableMapping.java:423) > at > org.datanucleus.store.rdbms.fieldmanager.ParameterSetter.storeObjectField(ParameterSetter.java:197) > at > org.datanucleus.state.AbstractStateManager.providedObjectField(AbstractStateManager.java:1023) > at > org.apache.hadoop.hive.metastore.model.MTable.jdoProvideField(MTable.java) > at > org.apache.hadoop.hive.metastore.model.MTable.jdoProvideFields(MTable.java) > at > org.datanucleus.jdo.state.JDOStateManagerImpl.provideFields(JDOStateManagerImpl.java:2803) > at > org.datanucleus.store.rdbms.request.InsertRequest.execute(InsertRequest.java:294) > at > org.datanucleus.store.rdbms.RDBMSPersistenceHandler.insertTable(RDBMSPersistenceHandler.java:146) > at > org.datanucleus.store.rdbms.RDBMSPersistenceHandler.insertObject(RDBMSPersistenceHandler.java:121) > at > org.datanucleus.jdo.state.JDOStateManagerImpl.internalMakePersistent(JDOStateManagerImpl.java:3275) > at > org.datanucleus.jdo.state.JDOStateManagerImpl.makePersistent(JDOStateManagerImpl.java:3249) > at > org.datanucleus.ObjectManagerImpl.persistObjectInternal(ObjectManagerImpl.java:1428) > at > org.datanucleus.ObjectManagerImpl.persistObject(ObjectManagerImpl.java:1241) > at > org.datanucleus.jdo.JDOPersistenceManager.jdoMakePersistent(JDOPersistenceManager.java:655) > at > org.datanucleus.jdo.JDOPersistenceManager.makePersistent(JDOPersistenceManager.java:680) > at > org.apache.hadoop.hive.metastore.ObjectStore.createTable(ObjectStore.java:606) > at > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.create_table_core(HiveMetaStore.java:924) > at > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.access$600(HiveMetaStore.java:109) > at > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler$15.run(HiveMetaStore.java:945) > at > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler$15.run(HiveMetaStore.java:942) > at > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.executeWithRetry(HiveMetaStore.java:307) > at > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.create_table(HiveMetaStore.java:942) > at > org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Processor$create_table.process(ThriftHiveMetastore.java:5297) > at > org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Processor.process(ThriftHiveMetastore.java:4789) > at > org.apache.hadoop.hive.metastore.HiveMetaStore$TLoggingProcessor.process(HiveMetaStore.java:3167) > at > org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:253) > at > java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) > at java.lang.Thread.run(Thread.java:662) > [gpadmin@linuxsvr2 hive]$ >