Hi, I am trying to connect the hive shell running on my laptop to a remote hadoop / hbase cluster and test out the HBase/Hive integration. I manage to connect and create the table in hbase from remote Hive shell. I am also passing the auxpath parameter to the shell (specifying the Hive/HBase integration related jars). In addition I have copied over these files to HDFS as well (I am using the user name hadoop - so the jars are stored in HDFS under /user/hadoop).
However when I fire a query on the HBase table - select * from h1 where key=12; - the map reduce job launches but the map task fails with the following error: ---- java.io.IOException: Cannot create an instance of InputSplit class = org.apache.hadoop.hive.hbase.HBaseSplit:org.apache.hadoop.hive.hbase.HBaseSplit at org.apache.hadoop.hive.ql.io.HiveInputFormat$HiveInputSplit.readFields(HiveInputFormat.java:143) at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:333) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:307) at org.apache.hadoop.mapred.Child.main(Child.java:170) ---- This basically indicates that the Mapper task is unable to locate the Hive/HBase storage handler that it requires when running. This happens even though this has been specified in the auxpath and uploaded to HDFS. Any ideas/pointers/debug options on what I might be doing wrong? Any help is much appreciated. p.s. the exploded jars do get copied too under the taskTracker directory on the cluster node Thanks