you can try hive 0.5, after create the metadata, use upgrade sql file
in hive 0.7.1 to upgrade to 0.7.1

On Sat, Aug 20, 2011 at 2:20 PM, Xiaobo Gu <guxiaobo1...@gmail.com> wrote:
> Hi,
> I have just set up a PostgreSQL 9.0.2 server for hive 0.7.1 metastore,
> and I am using the postgresql-9.0-801.jdbc4.jar jdbc driver, when I
> test the following HQL,
>
>
> CREATE TABLE records (year STRING, temperature INT, quality INT)
> ROW FORMAT DELIMITED
> FIELDS TERMINATED BY '\t';
>
>
> the following error messages are thrown in metastore services standard error:
>
> 11/08/20 10:12:33 INFO metastore.HiveMetaStore: 1: create_table:
> db=default tbl=records
> 11/08/20 10:12:33 INFO HiveMetaStore.audit: ugi=gpadmin
> ip=/192.168.72.6        cmd=create_table: db=default tbl=records
> 11/08/20 10:12:33 INFO metastore.HiveMetaStore: 1: Opening raw store
> with implemenation class:org.apache.hadoop.hive.metastore.ObjectStore
> 11/08/20 10:12:33 INFO metastore.ObjectStore: ObjectStore, initialize called
> 11/08/20 10:12:33 INFO metastore.ObjectStore: Initialized ObjectStore
> 11/08/20 10:12:33 INFO DataNucleus.Datastore: The class
> "org.apache.hadoop.hive.metastore.model.MFieldSchema" is tagged as
> "embedded-only" so does not have its own data
> store table.
> 11/08/20 10:12:33 INFO DataNucleus.Persistence: Managing Persistence
> of Class : org.apache.hadoop.hive.metastore.model.MSerDeInfo [Table :
> "SERDES", InheritanceStrategy
>  : new-table]
> 11/08/20 10:12:33 INFO DataNucleus.Datastore: The class
> "org.apache.hadoop.hive.metastore.model.MOrder" is tagged as
> "embedded-only" so does not have its own datastore
> table.
> 11/08/20 10:12:33 INFO DataNucleus.Persistence: Managing Persistence
> of Class : org.apache.hadoop.hive.metastore.model.MStorageDescriptor
> [Table : "SDS", InheritanceStr
> ategy : new-table]
> 11/08/20 10:12:33 INFO DataNucleus.Persistence: Managing Persistence
> of Class : org.apache.hadoop.hive.metastore.model.MTable [Table :
> "TBLS", InheritanceStrategy : new
> -table]
> 11/08/20 10:12:33 INFO DataNucleus.Persistence: Managing Persistence
> of Field : org.apache.hadoop.hive.metastore.model.MSerDeInfo.parameters
> [Table : "SERDE_PARAMS"]
> 11/08/20 10:12:33 INFO DataNucleus.Persistence: Managing Persistence
> of Field : org.apache.hadoop.hive.metastore.model.MTable.parameters
> [Table : "TABLE_PARAMS"]
> 11/08/20 10:12:33 INFO DataNucleus.Persistence: Managing Persistence
> of Field : org.apache.hadoop.hive.metastore.model.MTable.partitionKeys
> [Table : "PARTITION_KEYS"]
> 11/08/20 10:12:33 INFO DataNucleus.Persistence: Managing Persistence
> of Field : 
> org.apache.hadoop.hive.metastore.model.MStorageDescriptor.bucketCols
> [Table : "BUCKETING
> _COLS"]
> 11/08/20 10:12:33 INFO DataNucleus.Persistence: Managing Persistence
> of Field : org.apache.hadoop.hive.metastore.model.MStorageDescriptor.cols
> [Table : "COLUMNS"]
> 11/08/20 10:12:33 INFO DataNucleus.Persistence: Managing Persistence
> of Field : 
> org.apache.hadoop.hive.metastore.model.MStorageDescriptor.parameters
> [Table : "SD_PARAMS
> "]
> 11/08/20 10:12:33 INFO DataNucleus.Persistence: Managing Persistence
> of Field : org.apache.hadoop.hive.metastore.model.MStorageDescriptor.sortCols
> [Table : "SORT_COLS"]
> 11/08/20 10:12:33 INFO Datastore.Schema: Validating 1 unique key(s)
> for table "SERDES"
> 11/08/20 10:12:33 INFO Datastore.Schema: Validating 0 foreign key(s)
> for table "SERDES"
> 11/08/20 10:12:33 INFO Datastore.Schema: Validating 1 index(es) for
> table "SERDES"
> 11/08/20 10:12:33 INFO Datastore.Schema: Validating 2 unique key(s)
> for table "TBLS"
> 11/08/20 10:12:33 INFO Datastore.Schema: Validating 2 foreign key(s)
> for table "TBLS"
> 11/08/20 10:12:33 INFO Datastore.Schema: Validating 4 index(es) for table 
> "TBLS"
> 11/08/20 10:12:33 INFO Datastore.Schema: Validating 1 unique key(s)
> for table "SDS"
> 11/08/20 10:12:33 INFO Datastore.Schema: Validating 1 foreign key(s)
> for table "SDS"
> 11/08/20 10:12:33 INFO Datastore.Schema: Validating 2 index(es) for table 
> "SDS"
> 11/08/20 10:12:33 INFO Datastore.Schema: Validating 1 unique key(s)
> for table "SORT_COLS"
> 11/08/20 10:12:33 INFO Datastore.Schema: Validating 1 foreign key(s)
> for table "SORT_COLS"
> 11/08/20 10:12:33 INFO Datastore.Schema: Validating 2 index(es) for
> table "SORT_COLS"
> 11/08/20 10:12:33 INFO Datastore.Schema: Validating 1 unique key(s)
> for table "TABLE_PARAMS"
> 11/08/20 10:12:33 INFO Datastore.Schema: Validating 1 foreign key(s)
> for table "TABLE_PARAMS"
> 11/08/20 10:12:33 INFO Datastore.Schema: Validating 2 index(es) for
> table "TABLE_PARAMS"
> 11/08/20 10:12:33 INFO Datastore.Schema: Validating 1 unique key(s)
> for table "SD_PARAMS"
> 11/08/20 10:12:33 INFO Datastore.Schema: Validating 1 foreign key(s)
> for table "SD_PARAMS"
> 11/08/20 10:12:33 INFO Datastore.Schema: Validating 2 index(es) for
> table "SD_PARAMS"
> 11/08/20 10:12:33 INFO Datastore.Schema: Validating 1 unique key(s)
> for table "SERDE_PARAMS"
> 11/08/20 10:12:33 INFO Datastore.Schema: Validating 1 foreign key(s)
> for table "SERDE_PARAMS"
> 11/08/20 10:12:33 INFO Datastore.Schema: Validating 2 index(es) for
> table "SERDE_PARAMS"
> 11/08/20 10:12:33 INFO Datastore.Schema: Validating 1 unique key(s)
> for table "PARTITION_KEYS"
> 11/08/20 10:12:33 INFO Datastore.Schema: Validating 1 foreign key(s)
> for table "PARTITION_KEYS"
> 11/08/20 10:12:33 INFO Datastore.Schema: Validating 2 index(es) for
> table "PARTITION_KEYS"
> 11/08/20 10:12:33 INFO Datastore.Schema: Validating 1 unique key(s)
> for table "COLUMNS"
> 11/08/20 10:12:33 INFO Datastore.Schema: Validating 1 foreign key(s)
> for table "COLUMNS"
> 11/08/20 10:12:33 INFO Datastore.Schema: Validating 2 index(es) for
> table "COLUMNS"
> 11/08/20 10:12:33 INFO Datastore.Schema: Validating 1 unique key(s)
> for table "BUCKETING_COLS"
> 11/08/20 10:12:33 INFO Datastore.Schema: Validating 1 foreign key(s)
> for table "BUCKETING_COLS"
> 11/08/20 10:12:33 INFO Datastore.Schema: Validating 2 index(es) for
> table "BUCKETING_COLS"
> 11/08/20 10:12:34 INFO DataNucleus.MetaData: Listener found
> initialisation for persistable class
> org.apache.hadoop.hive.metastore.model.MSerDeInfo
> 11/08/20 10:12:34 INFO DataNucleus.MetaData: Listener found
> initialisation for persistable class
> org.apache.hadoop.hive.metastore.model.MStorageDescriptor
> 11/08/20 10:12:34 INFO DataNucleus.MetaData: Listener found
> initialisation for persistable class
> org.apache.hadoop.hive.metastore.model.MTable
> 11/08/20 10:12:34 INFO DataNucleus.MetaData: Listener found
> initialisation for persistable class
> org.apache.hadoop.hive.metastore.model.MFieldSchema
> 11/08/20 10:12:34 WARN Datastore.Persist: Insert of object
> "org.apache.hadoop.hive.metastore.model.MStorageDescriptor@46bb05de"
> using statement "INSERT INTO "SDS" ("SD_
> ID","NUM_BUCKETS","SERDE_ID","IS_COMPRESSED","OUTPUT_FORMAT","LOCATION","INPUT_FORMAT")
> VALUES (?,?,?,?,?,?,?)" failed : ERROR: column "IS_COMPRESSED" is of
> type bit bu
> t expression is of type boolean
>  Hint: You will need to rewrite or cast the expression.
>  Position: 129
> 11/08/20 10:12:34 INFO metastore.hivemetastoressimpl: deleting
> hdfs://linuxsvr2/user/hive/warehouse/records
> 11/08/20 10:12:34 INFO metastore.hivemetastoressimpl: Deleted the
> diretory hdfs://linuxsvr2/user/hive/warehouse/records
> 11/08/20 10:12:34 ERROR metastore.HiveMetaStore: JDO datastore error.
> Retrying metastore command after 1000 ms (attempt 1 of 1)
> 11/08/20 10:12:35 WARN metastore.HiveMetaStore: Location:
> hdfs://linuxsvr2/user/hive/warehouse/records specified for
> non-external table:records
> 11/08/20 10:12:35 WARN Datastore.Persist: Insert of object
> "org.apache.hadoop.hive.metastore.model.MStorageDescriptor@7d1c19e6"
> using statement "INSERT INTO "SDS" ("SD_
> ID","NUM_BUCKETS","SERDE_ID","IS_COMPRESSED","OUTPUT_FORMAT","LOCATION","INPUT_FORMAT")
> VALUES (?,?,?,?,?,?,?)" failed : ERROR: column "IS_COMPRESSED" is of
> type bit bu
> t expression is of type boolean
>  Hint: You will need to rewrite or cast the expression.
>  Position: 129
> 11/08/20 10:12:35 INFO metastore.hivemetastoressimpl: deleting
> hdfs://linuxsvr2/user/hive/warehouse/records
> 11/08/20 10:12:35 INFO metastore.hivemetastoressimpl: Deleted the
> diretory hdfs://linuxsvr2/user/hive/warehouse/records
> 11/08/20 10:12:35 ERROR api.ThriftHiveMetastore$Processor: Internal
> error processing create_table
> javax.jdo.JDODataStoreException: Insert of object
> "org.apache.hadoop.hive.metastore.model.MStorageDescriptor@7d1c19e6"
> using statement "INSERT INTO "SDS" ("SD_ID","NUM_
> BUCKETS","SERDE_ID","IS_COMPRESSED","OUTPUT_FORMAT","LOCATION","INPUT_FORMAT")
> VALUES (?,?,?,?,?,?,?)" failed : ERROR: column "IS_COMPRESSED" is of
> type bit but express
> ion is of type boolean
>  Hint: You will need to rewrite or cast the expression.
>  Position: 129
>        at 
> org.datanucleus.jdo.NucleusJDOHelper.getJDOExceptionForNucleusException(NucleusJDOHelper.java:313)
>        at 
> org.datanucleus.jdo.JDOPersistenceManager.jdoMakePersistent(JDOPersistenceManager.java:660)
>        at 
> org.datanucleus.jdo.JDOPersistenceManager.makePersistent(JDOPersistenceManager.java:680)
>        at 
> org.apache.hadoop.hive.metastore.ObjectStore.createTable(ObjectStore.java:606)
>        at 
> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.create_table_core(HiveMetaStore.java:924)
>        at 
> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.access$600(HiveMetaStore.java:109)
>        at 
> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler$15.run(HiveMetaStore.java:945)
>        at 
> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler$15.run(HiveMetaStore.java:942)
>        at 
> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.executeWithRetry(HiveMetaStore.java:307)
>        at 
> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.create_table(HiveMetaStore.java:942)
>        at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Processor$create_table.process(ThriftHiveMetastore.java:5297)
>        at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Processor.process(ThriftHiveMetastore.java:4789)
>        at 
> org.apache.hadoop.hive.metastore.HiveMetaStore$TLoggingProcessor.process(HiveMetaStore.java:3167)
>        at 
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:253)
>        at 
> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
>        at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
>        at java.lang.Thread.run(Thread.java:662)
> NestedThrowablesStackTrace:
> org.postgresql.util.PSQLException: ERROR: column "IS_COMPRESSED" is of
> type bit but expression is of type boolean
>  Hint: You will need to rewrite or cast the expression.
>  Position: 129
>        at 
> org.postgresql.core.v3.QueryExecutorImpl.receiveErrorResponse(QueryExecutorImpl.java:2102)
>        at 
> org.postgresql.core.v3.QueryExecutorImpl.processResults(QueryExecutorImpl.java:1835)
>        at 
> org.postgresql.core.v3.QueryExecutorImpl.execute(QueryExecutorImpl.java:257)
>        at 
> org.postgresql.jdbc2.AbstractJdbc2Statement.execute(AbstractJdbc2Statement.java:500)
>        at 
> org.postgresql.jdbc2.AbstractJdbc2Statement.executeWithFlags(AbstractJdbc2Statement.java:388)
>        at 
> org.postgresql.jdbc2.AbstractJdbc2Statement.executeUpdate(AbstractJdbc2Statement.java:334)
>        at 
> org.apache.commons.dbcp.DelegatingPreparedStatement.executeUpdate(DelegatingPreparedStatement.java:105)
>        at 
> org.apache.commons.dbcp.DelegatingPreparedStatement.executeUpdate(DelegatingPreparedStatement.java:105)
>        at 
> org.datanucleus.store.rdbms.SQLController.executeStatementUpdate(SQLController.java:396)
>        at 
> org.datanucleus.store.rdbms.request.InsertRequest.execute(InsertRequest.java:406)
>        at 
> org.datanucleus.store.rdbms.RDBMSPersistenceHandler.insertTable(RDBMSPersistenceHandler.java:146)
>        at 
> org.datanucleus.store.rdbms.RDBMSPersistenceHandler.insertObject(RDBMSPersistenceHandler.java:121)
>        at 
> org.datanucleus.jdo.state.JDOStateManagerImpl.internalMakePersistent(JDOStateManagerImpl.java:3275)
>        at 
> org.datanucleus.jdo.state.JDOStateManagerImpl.makePersistent(JDOStateManagerImpl.java:3249)
>        at 
> org.datanucleus.ObjectManagerImpl.persistObjectInternal(ObjectManagerImpl.java:1428)
>        at 
> org.datanucleus.store.mapped.mapping.PersistableMapping.setObjectAsValue(PersistableMapping.java:664)
>        at 
> org.datanucleus.store.mapped.mapping.PersistableMapping.setObject(PersistableMapping.java:423)
>        at 
> org.datanucleus.store.rdbms.fieldmanager.ParameterSetter.storeObjectField(ParameterSetter.java:197)
>        at 
> org.datanucleus.state.AbstractStateManager.providedObjectField(AbstractStateManager.java:1023)
>        at 
> org.apache.hadoop.hive.metastore.model.MTable.jdoProvideField(MTable.java)
>        at 
> org.apache.hadoop.hive.metastore.model.MTable.jdoProvideFields(MTable.java)
>        at 
> org.datanucleus.jdo.state.JDOStateManagerImpl.provideFields(JDOStateManagerImpl.java:2803)
>        at 
> org.datanucleus.store.rdbms.request.InsertRequest.execute(InsertRequest.java:294)
>        at 
> org.datanucleus.store.rdbms.RDBMSPersistenceHandler.insertTable(RDBMSPersistenceHandler.java:146)
>        at 
> org.datanucleus.store.rdbms.RDBMSPersistenceHandler.insertObject(RDBMSPersistenceHandler.java:121)
>        at 
> org.datanucleus.jdo.state.JDOStateManagerImpl.internalMakePersistent(JDOStateManagerImpl.java:3275)
>        at 
> org.datanucleus.jdo.state.JDOStateManagerImpl.makePersistent(JDOStateManagerImpl.java:3249)
>        at 
> org.datanucleus.ObjectManagerImpl.persistObjectInternal(ObjectManagerImpl.java:1428)
>        at 
> org.datanucleus.ObjectManagerImpl.persistObject(ObjectManagerImpl.java:1241)
>        at 
> org.datanucleus.jdo.JDOPersistenceManager.jdoMakePersistent(JDOPersistenceManager.java:655)
>        at 
> org.datanucleus.jdo.JDOPersistenceManager.makePersistent(JDOPersistenceManager.java:680)
>        at 
> org.apache.hadoop.hive.metastore.ObjectStore.createTable(ObjectStore.java:606)
>        at 
> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.create_table_core(HiveMetaStore.java:924)
>        at 
> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.access$600(HiveMetaStore.java:109)
>        at 
> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler$15.run(HiveMetaStore.java:945)
>        at 
> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler$15.run(HiveMetaStore.java:942)
>        at 
> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.executeWithRetry(HiveMetaStore.java:307)
>        at 
> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.create_table(HiveMetaStore.java:942)
>        at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Processor$create_table.process(ThriftHiveMetastore.java:5297)
>        at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Processor.process(ThriftHiveMetastore.java:4789)
>        at 
> org.apache.hadoop.hive.metastore.HiveMetaStore$TLoggingProcessor.process(HiveMetaStore.java:3167)
>        at 
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:253)
>        at 
> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
>        at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
>        at java.lang.Thread.run(Thread.java:662)
> [gpadmin@linuxsvr2 hive]$
>

Reply via email to