Do you have hbase-site.xml in classpath ?

On 17 March 2016 at 17:08, songj songj <songjun...@gmail.com> wrote:

> <property>
>    <name>zookeeper.znode.parent</name>
>    <value>/hbase</value>
> </property>
>
> and I found it that ,bind any ip which the hive can access to
> 'hbase-cluster' ,they are all ok!
>
>
>
> 2016-03-17 16:46 GMT+08:00 Divya Gehlot <divya.htco...@gmail.com>:
>
>> Hi,
>> Please check your zookeeper.znode.parent property
>> where is it pointing to ?
>>
>> On 17 March 2016 at 15:21, songj songj <songjun...@gmail.com> wrote:
>>
>>> hi all:
>>> I have 2 cluster,one is hive cluster(2.0.0),another is hbase
>>> cluster(1.1.1),
>>> this two clusters have dependent hdfs:
>>>
>>> hive cluster:
>>> <property>
>>>    <name>fs.defaultFS</name>
>>>    <value>hdfs://*hive-cluster*</value>
>>> </property>
>>>
>>> hbase cluster:
>>> <property>
>>>    <name>fs.defaultFS</name>
>>>    <value>hdfs://*hbase-cluster*</value>
>>> </property>
>>>
>>> *1)*but when I use hive shell to access hbase cluster
>>> >set hbase.zookeeper.quorum=10.24.19.88;
>>> >CREATE EXTERNAL TABLE IF NOT EXISTS pagecounts_hbase (rowkey STRING,
>>> pageviews STRING, bytes STRING) STORED BY
>>> 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' WITH SERDEPROPERTIES
>>> ('hbase.columns.mapping' = ':key,cf:c1,cf:c2') TBLPROPERTIES ('
>>> hbase.table.name' = 'test');
>>>
>>> *2)*then I got exceptions:
>>>
>>> FAILED: Execution Error, return code 1 from
>>> org.apache.hadoop.hive.ql.exec.DDLTask.
>>> MetaException(message:MetaException(message:java.io.IOException:
>>> java.lang.reflect.InvocationTargetException
>>>      at
>>> org.apache.hadoop.hbase.client.ConnectionFactory.createConnection(ConnectionFactory.java:240)
>>>      at
>>> org.apache.hadoop.hbase.client.ConnectionManager.createConnection(ConnectionManager.java:420)
>>>      at
>>> org.apache.hadoop.hbase.client.ConnectionManager.createConnection(ConnectionManager.java:413)
>>>      at
>>> org.apache.hadoop.hbase.client.ConnectionManager.getConnectionInternal(ConnectionManager.java:291)
>>>      at
>>> org.apache.hadoop.hbase.client.HBaseAdmin.<init>(HBaseAdmin.java:222)
>>>      at
>>> org.apache.hadoop.hive.hbase.HBaseStorageHandler.getHBaseAdmin(HBaseStorageHandler.java:102)
>>>      at
>>> org.apache.hadoop.hive.hbase.HBaseStorageHandler.preCreateTable(HBaseStorageHandler.java:182)
>>>      at
>>> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createTable(HiveMetaStoreClient.java:608)
>>>      at
>>> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createTable(HiveMetaStoreClient.java:601)
>>>      at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>      at
>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>>      at
>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>      at java.lang.reflect.Method.invoke(Method.java:606)
>>>      at
>>> org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:90)
>>>      at com.sun.proxy.$Proxy15.createTable(Unknown Source)
>>>      at
>>> org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:671)
>>>      at
>>> org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:3973)
>>>      at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:295)
>>>      at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:160)
>>>      at
>>> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:85)
>>>      at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1604)
>>>      at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1364)
>>>      at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1177)
>>>      at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1004)
>>>      at org.apache.hadoop.hive.ql.Driver.run(Driver.java:994)
>>>      at
>>> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:201)
>>>      at
>>> org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:153)
>>>      at
>>> org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:364)
>>>      at
>>> org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:712)
>>>      at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:631)
>>>      at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:570)
>>>      at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>      at
>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>>      at
>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>      at java.lang.reflect.Method.invoke(Method.java:606)
>>>      at org.apache.hadoop.util.RunJar.run(RunJar.java:221)
>>>      at org.apache.hadoop.util.RunJar.main(RunJar.java:136)
>>> Caused by: java.lang.reflect.InvocationTargetException
>>>      at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
>>> Method)
>>>      at
>>> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
>>>      at
>>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
>>>      at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
>>>      at
>>> org.apache.hadoop.hbase.client.ConnectionFactory.createConnection(ConnectionFactory.java:238)
>>>      ... 36 more
>>> Caused by: java.lang.ExceptionInInitializerError
>>>      at org.apache.hadoop.hbase.ClusterId.parseFrom(ClusterId.java:64)
>>>      at
>>> org.apache.hadoop.hbase.zookeeper.ZKClusterId.readClusterIdZNode(ZKClusterId.java:75)
>>>      at
>>> org.apache.hadoop.hbase.client.ZooKeeperRegistry.getClusterId(ZooKeeperRegistry.java:105)
>>>      at
>>> org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.retrieveClusterId(ConnectionManager.java:879)
>>>      at
>>> org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.<init>(ConnectionManager.java:635)
>>>      ... 41 more
>>> Caused by: java.lang.IllegalArgumentException:
>>> java.net.UnknownHostException: hbase-cluster
>>>      at
>>> org.apache.hadoop.security.SecurityUtil.buildTokenService(SecurityUtil.java:374)
>>>      at
>>> org.apache.hadoop.hdfs.NameNodeProxies.createNonHAProxy(NameNodeProxies.java:312)
>>>      at
>>> org.apache.hadoop.hdfs.NameNodeProxies.createProxy(NameNodeProxies.java:178)
>>>      at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:665)
>>>      at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:601)
>>>      at
>>> org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:148)
>>>      at
>>> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2596)
>>>      at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:91)
>>>      at
>>> org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2630)
>>>      at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2612)
>>>      at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:370)
>>>      at org.apache.hadoop.fs.Path.getFileSystem(Path.java:296)
>>>      at
>>> org.apache.hadoop.hbase.util.DynamicClassLoader.<init>(DynamicClassLoader.java:104)
>>>      at
>>> org.apache.hadoop.hbase.protobuf.ProtobufUtil.<clinit>(ProtobufUtil.java:238)
>>>      ... 46 more
>>> *Caused by: java.net.UnknownHostException: hbase-cluster*
>>>      ... 60 more
>>> )
>>>      at
>>> org.apache.hadoop.hive.hbase.HBaseStorageHandler.getHBaseAdmin(HBaseStorageHandler.java:106)
>>>      at
>>> org.apache.hadoop.hive.hbase.HBaseStorageHandler.preCreateTable(HBaseStorageHandler.java:182)
>>>      at
>>> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createTable(HiveMetaStoreClient.java:608)
>>>      at
>>> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createTable(HiveMetaStoreClient.java:601)
>>>      at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>      at
>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>>      at
>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>      at java.lang.reflect.Method.invoke(Method.java:606)
>>>      at
>>> org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:90)
>>>      at com.sun.proxy.$Proxy15.createTable(Unknown Source)
>>>      at
>>> org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:671)
>>>      at
>>> org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:3973)
>>>      at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:295)
>>>      at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:160)
>>>      at
>>> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:85)
>>>      at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1604)
>>>      at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1364)
>>>      at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1177)
>>>      at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1004)
>>>      at org.apache.hadoop.hive.ql.Driver.run(Driver.java:994)
>>>      at
>>> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:201)
>>>      at
>>> org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:153)
>>>      at
>>> org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:364)
>>>      at
>>> org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:712)
>>>      at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:631)
>>>      at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:570)
>>>      at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>      at
>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>>      at
>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>      at java.lang.reflect.Method.invoke(Method.java:606)
>>>      at org.apache.hadoop.util.RunJar.run(RunJar.java:221)
>>>      at org.apache.hadoop.util.RunJar.main(RunJar.java:136)
>>>
>>> *3) *then I modify hosts file on hive (*the ip is hive machine*)
>>> cat '10.24.19.32 hbase-cluster' >>/etc/hosts
>>>
>>> then everything is ok!
>>>
>>> *or *cat '10.24.19.88 hbase-cluster' >>/etc/hosts (*the ip is hbase
>>> machine*)
>>>
>>> it is also ok!
>>>
>>> *So, why hive cluster should resolve the host 'hbase-cluster' *
>>> *which is the hdfs url of the hbase cluster?*
>>> *and why I use the hive/hbase ip to bind host 'hbase-cluster', they are
>>> all ok?*
>>>
>>>
>>>
>>
>

Reply via email to