We have a hadoop/hive cluster which is using cloudera's distribution. Metastore is stored in mysql and all the relevant drivers are in classpath and in conf files. While running queries on hive I am getting these errors
$ hive -hiveconf hive.root.logger=INFO,console Hive history file=/tmp/vipul/hive_job_log_vipul_201104271409_1310845475.txt 11/04/27 14:09:01 INFO exec.HiveHistory: Hive history file=/tmp/vipul/hive_job_log_vipul_201104271409_1310845475.txt hive> select time from requests_stat_min; 11/04/27 14:15:38 INFO parse.ParseDriver: Parsing command: select time from requests_stat_min 11/04/27 14:15:38 INFO parse.ParseDriver: Parse Completed 11/04/27 14:15:39 INFO parse.SemanticAnalyzer: Starting Semantic Analysis 11/04/27 14:15:39 INFO parse.SemanticAnalyzer: Completed phase 1 of Semantic Analysis 11/04/27 14:15:39 INFO parse.SemanticAnalyzer: Get metadata for source tables 11/04/27 14:15:39 INFO metastore.HiveMetaStore: 0: Opening raw store with implemenation class:org.apache.hadoop.hive.metastore.ObjectStore 11/04/27 14:15:39 INFO metastore.ObjectStore: ObjectStore, initialize called 11/04/27 14:15:39 ERROR DataNucleus.Plugin: Bundle "org.eclipse.jdt.core" requires "org.eclipse.core.resources" but it cannot be resolved. 11/04/27 14:15:39 ERROR DataNucleus.Plugin: Bundle "org.eclipse.jdt.core" requires "org.eclipse.core.runtime" but it cannot be resolved. 11/04/27 14:15:39 ERROR DataNucleus.Plugin: Bundle "org.eclipse.jdt.core" requires "org.eclipse.text" but it cannot be resolved. 11/04/27 14:15:40 INFO metastore.ObjectStore: Initialized ObjectStore 11/04/27 14:15:40 INFO metastore.HiveMetaStore: 0: get_table : db=default tbl=requests_stat_min 11/04/27 14:15:41 INFO hive.log: DDL: struct requests_stat_min { string time, i64 request_count, i64 get_count, i64 post_count, i64 unique_orders, i64 unique_events, i64 start_orders, i32 max_response_time, i32 min_response_time, i32 avg_response_time, i32 max_response_size, i32 min_response_size, i32 avg_response_size, i64 unique_referrers} 11/04/27 14:15:41 INFO parse.SemanticAnalyzer: Get metadata for subqueries 11/04/27 14:15:41 INFO parse.SemanticAnalyzer: Get metadata for destination tables 11/04/27 14:15:41 INFO parse.SemanticAnalyzer: Completed getting MetaData in Semantic Analysis 11/04/27 14:15:41 INFO ppd.OpProcFactory: Processing for FS(2) 11/04/27 14:15:41 INFO ppd.OpProcFactory: Processing for SEL(1) 11/04/27 14:15:41 INFO ppd.OpProcFactory: Processing for TS(0) 11/04/27 14:15:41 INFO metastore.HiveMetaStore: 0: get_partition_names : db=default tbl=requests_stat_min 11/04/27 14:15:41 INFO metastore.HiveMetaStore: 0: get_partition : db=default tbl=requests_stat_min 11/04/27 14:15:41 INFO metastore.HiveMetaStore: 0: get_partition : db=default tbl=requests_stat_min 11/04/27 14:15:42 INFO metastore.HiveMetaStore: 0: get_partition : db=default tbl=requests_stat_min 11/04/27 14:15:42 INFO metastore.HiveMetaStore: 0: get_partition : db=default tbl=requests_stat_min 11/04/27 14:15:42 INFO metastore.HiveMetaStore: 0: get_partition : db=default tbl=requests_stat_min 11/04/27 14:15:42 INFO metastore.HiveMetaStore: 0: get_partition : db=default tbl=requests_stat_min 11/04/27 14:15:42 INFO metastore.HiveMetaStore: 0: get_partition : db=default tbl=requests_stat_min 11/04/27 14:15:42 INFO metastore.HiveMetaStore: 0: get_partition : db=default tbl=requests_stat_min 11/04/27 14:15:42 INFO metastore.HiveMetaStore: 0: get_partition : db=default tbl=requests_stat_min 11/04/27 14:15:42 INFO metastore.HiveMetaStore: 0: get_partition : db=default tbl=requests_stat_min 11/04/27 14:15:42 INFO hive.log: DDL: struct requests_stat_min { string time, i64 request_count, i64 get_count, i64 post_count, i64 unique_orders, i64 unique_events, i64 start_orders, i32 max_response_time, i32 min_response_time, i32 avg_response_time, i32 max_response_size, i32 min_response_size, i32 avg_response_size, i64 unique_referrers} 11/04/27 14:15:42 INFO hive.log: DDL: struct requests_stat_min { string time, i64 request_count, i64 get_count, i64 post_count, i64 unique_orders, i64 unique_events, i64 start_orders, i32 max_response_time, i32 min_response_time, i32 11/04/27 14:15:42 INFO parse.SemanticAnalyzer: Completed plan generation 11/04/27 14:15:42 INFO ql.Driver: Semantic Analysis Completed 11/04/27 14:15:42 INFO ql.Driver: Starting command: select time from requests_stat_min Total MapReduce jobs = 1 11/04/27 14:15:42 INFO ql.Driver: Total MapReduce jobs = 1 Launching Job 1 out of 1 11/04/27 14:15:42 INFO ql.Driver: Launching Job 1 out of 1 Number of reduce tasks is set to 0 since there's no reduce operator 11/04/27 14:15:42 INFO exec.ExecDriver: Number of reduce tasks is set to 0 since there's no reduce operator FAILED: Unknown exception : null 11/04/27 14:15:42 ERROR ql.Driver: FAILED: Unknown exception : null java.lang.NullPointerException at java.util.Hashtable.put(Hashtable.java:394) at java.util.Properties.setProperty(Properties.java:143) at org.apache.hadoop.conf.Configuration.set(Configuration.java:460) at org.apache.hadoop.hive.conf.HiveConf.setVar(HiveConf.java:293) at org.apache.hadoop.hive.ql.exec.ExecDriver.execute(ExecDriver.java:505) at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:100) at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:64) at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:572) at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:452) at org.apache.hadoop.hive.ql.Driver.runCommand(Driver.java:314) at org.apache.hadoop.hive.ql.Driver.run(Driver.java:302) at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:123) at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:181) at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:287) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:597) at org.apache.hadoop.util.RunJar.main(RunJar.java:186) I am assuming this has to do with metastore but I cant figure out whats wrong. The way we run query is via a remote client. Any help is greatly appreciated! -- Vipul Sharma sharmavipul AT gmail DOT com