<property>
   <name>zookeeper.znode.parent</name>
   <value>/hbase</value>
</property>

and I found it that ,bind any ip which the hive can access to
'hbase-cluster' ,they are all ok!



2016-03-17 16:46 GMT+08:00 Divya Gehlot <divya.htco...@gmail.com>:

> Hi,
> Please check your zookeeper.znode.parent property
> where is it pointing to ?
>
> On 17 March 2016 at 15:21, songj songj <songjun...@gmail.com> wrote:
>
>> hi all:
>> I have 2 cluster,one is hive cluster(2.0.0),another is hbase
>> cluster(1.1.1),
>> this two clusters have dependent hdfs:
>>
>> hive cluster:
>> <property>
>>    <name>fs.defaultFS</name>
>>    <value>hdfs://*hive-cluster*</value>
>> </property>
>>
>> hbase cluster:
>> <property>
>>    <name>fs.defaultFS</name>
>>    <value>hdfs://*hbase-cluster*</value>
>> </property>
>>
>> *1)*but when I use hive shell to access hbase cluster
>> >set hbase.zookeeper.quorum=10.24.19.88;
>> >CREATE EXTERNAL TABLE IF NOT EXISTS pagecounts_hbase (rowkey STRING,
>> pageviews STRING, bytes STRING) STORED BY
>> 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' WITH SERDEPROPERTIES
>> ('hbase.columns.mapping' = ':key,cf:c1,cf:c2') TBLPROPERTIES ('
>> hbase.table.name' = 'test');
>>
>> *2)*then I got exceptions:
>>
>> FAILED: Execution Error, return code 1 from
>> org.apache.hadoop.hive.ql.exec.DDLTask.
>> MetaException(message:MetaException(message:java.io.IOException:
>> java.lang.reflect.InvocationTargetException
>>      at
>> org.apache.hadoop.hbase.client.ConnectionFactory.createConnection(ConnectionFactory.java:240)
>>      at
>> org.apache.hadoop.hbase.client.ConnectionManager.createConnection(ConnectionManager.java:420)
>>      at
>> org.apache.hadoop.hbase.client.ConnectionManager.createConnection(ConnectionManager.java:413)
>>      at
>> org.apache.hadoop.hbase.client.ConnectionManager.getConnectionInternal(ConnectionManager.java:291)
>>      at
>> org.apache.hadoop.hbase.client.HBaseAdmin.<init>(HBaseAdmin.java:222)
>>      at
>> org.apache.hadoop.hive.hbase.HBaseStorageHandler.getHBaseAdmin(HBaseStorageHandler.java:102)
>>      at
>> org.apache.hadoop.hive.hbase.HBaseStorageHandler.preCreateTable(HBaseStorageHandler.java:182)
>>      at
>> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createTable(HiveMetaStoreClient.java:608)
>>      at
>> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createTable(HiveMetaStoreClient.java:601)
>>      at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>      at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>      at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>      at java.lang.reflect.Method.invoke(Method.java:606)
>>      at
>> org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:90)
>>      at com.sun.proxy.$Proxy15.createTable(Unknown Source)
>>      at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:671)
>>      at
>> org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:3973)
>>      at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:295)
>>      at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:160)
>>      at
>> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:85)
>>      at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1604)
>>      at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1364)
>>      at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1177)
>>      at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1004)
>>      at org.apache.hadoop.hive.ql.Driver.run(Driver.java:994)
>>      at
>> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:201)
>>      at
>> org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:153)
>>      at
>> org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:364)
>>      at
>> org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:712)
>>      at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:631)
>>      at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:570)
>>      at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>      at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>      at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>      at java.lang.reflect.Method.invoke(Method.java:606)
>>      at org.apache.hadoop.util.RunJar.run(RunJar.java:221)
>>      at org.apache.hadoop.util.RunJar.main(RunJar.java:136)
>> Caused by: java.lang.reflect.InvocationTargetException
>>      at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
>> Method)
>>      at
>> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
>>      at
>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
>>      at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
>>      at
>> org.apache.hadoop.hbase.client.ConnectionFactory.createConnection(ConnectionFactory.java:238)
>>      ... 36 more
>> Caused by: java.lang.ExceptionInInitializerError
>>      at org.apache.hadoop.hbase.ClusterId.parseFrom(ClusterId.java:64)
>>      at
>> org.apache.hadoop.hbase.zookeeper.ZKClusterId.readClusterIdZNode(ZKClusterId.java:75)
>>      at
>> org.apache.hadoop.hbase.client.ZooKeeperRegistry.getClusterId(ZooKeeperRegistry.java:105)
>>      at
>> org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.retrieveClusterId(ConnectionManager.java:879)
>>      at
>> org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.<init>(ConnectionManager.java:635)
>>      ... 41 more
>> Caused by: java.lang.IllegalArgumentException:
>> java.net.UnknownHostException: hbase-cluster
>>      at
>> org.apache.hadoop.security.SecurityUtil.buildTokenService(SecurityUtil.java:374)
>>      at
>> org.apache.hadoop.hdfs.NameNodeProxies.createNonHAProxy(NameNodeProxies.java:312)
>>      at
>> org.apache.hadoop.hdfs.NameNodeProxies.createProxy(NameNodeProxies.java:178)
>>      at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:665)
>>      at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:601)
>>      at
>> org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:148)
>>      at
>> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2596)
>>      at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:91)
>>      at
>> org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2630)
>>      at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2612)
>>      at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:370)
>>      at org.apache.hadoop.fs.Path.getFileSystem(Path.java:296)
>>      at
>> org.apache.hadoop.hbase.util.DynamicClassLoader.<init>(DynamicClassLoader.java:104)
>>      at
>> org.apache.hadoop.hbase.protobuf.ProtobufUtil.<clinit>(ProtobufUtil.java:238)
>>      ... 46 more
>> *Caused by: java.net.UnknownHostException: hbase-cluster*
>>      ... 60 more
>> )
>>      at
>> org.apache.hadoop.hive.hbase.HBaseStorageHandler.getHBaseAdmin(HBaseStorageHandler.java:106)
>>      at
>> org.apache.hadoop.hive.hbase.HBaseStorageHandler.preCreateTable(HBaseStorageHandler.java:182)
>>      at
>> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createTable(HiveMetaStoreClient.java:608)
>>      at
>> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createTable(HiveMetaStoreClient.java:601)
>>      at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>      at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>      at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>      at java.lang.reflect.Method.invoke(Method.java:606)
>>      at
>> org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:90)
>>      at com.sun.proxy.$Proxy15.createTable(Unknown Source)
>>      at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:671)
>>      at
>> org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:3973)
>>      at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:295)
>>      at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:160)
>>      at
>> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:85)
>>      at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1604)
>>      at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1364)
>>      at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1177)
>>      at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1004)
>>      at org.apache.hadoop.hive.ql.Driver.run(Driver.java:994)
>>      at
>> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:201)
>>      at
>> org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:153)
>>      at
>> org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:364)
>>      at
>> org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:712)
>>      at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:631)
>>      at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:570)
>>      at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>      at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>      at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>      at java.lang.reflect.Method.invoke(Method.java:606)
>>      at org.apache.hadoop.util.RunJar.run(RunJar.java:221)
>>      at org.apache.hadoop.util.RunJar.main(RunJar.java:136)
>>
>> *3) *then I modify hosts file on hive (*the ip is hive machine*)
>> cat '10.24.19.32 hbase-cluster' >>/etc/hosts
>>
>> then everything is ok!
>>
>> *or *cat '10.24.19.88 hbase-cluster' >>/etc/hosts (*the ip is hbase
>> machine*)
>>
>> it is also ok!
>>
>> *So, why hive cluster should resolve the host 'hbase-cluster' *
>> *which is the hdfs url of the hbase cluster?*
>> *and why I use the hive/hbase ip to bind host 'hbase-cluster', they are
>> all ok?*
>>
>>
>>
>

Reply via email to