I am really trying to get HBase to work... Is there maybe a tutorial for all the config files? Best regards, Lydia
> Am 23.09.2015 um 17:40 schrieb Maximilian Michels <m...@apache.org>: > > In the issue, it states that it should be sufficient to append the > hbase-protocol.jar file to the Hadoop classpath. Flink respects the Hadoop > classpath and will append it to its own classpath upon launching a cluster. > > To do that, you need to modify the classpath with one of the commands below. > Note that this has to be performed on all cluster nodes. > > export HADOOP_CLASSPATH="${HADOOP_CLASSPATH}:/path/to/hbase-protocol.jar" > export HADOOP_CLASSPATH="${HADOOP_CLASSPATH}:$(hbase mapredcp)" > export HADOOP_CLASSPATH="${HADOOP_CLASSPATH}:$(hbase classpath)" > > Alternatively, you can build a fat jar from your project with the missing > dependency. Flink will then automatically distribute the jar file upon job > submission. Just add this Maven dependency to your fat-jar pom: > > <dependency> > <groupId>org.apache.hbase</groupId> > <artifactId>hbase-protocol</artifactId> > <version>1.1.2</version> > </dependency> > > Let me know if any of the two approaches work for you. After all, this is a > workaround because of an HBase optimzation.. > > Cheers, > Max > > > >> On Wed, Sep 23, 2015 at 11:16 AM, Aljoscha Krettek <aljos...@apache.org> >> wrote: >> It might me that this is causing the problem: >> https://issues.apache.org/jira/browse/HBASE-10304 >> >> In your log I see the same exception. Anyone has any idea what we could do >> about this? >> >> >>> On Tue, 22 Sep 2015 at 22:40 Lydia Ickler <ickle...@googlemail.com> wrote: >>> Hi, >>> >>> I am trying to get the HBaseReadExample to run. I have filled a table with >>> the HBaseWriteExample and purposely split it over 3 regions. >>> Now when I try to read from it the first split seems to be scanned (170 >>> rows) fine and after that the Connections of Zookeeper and RCP are suddenly >>> closed down. >>> >>> Does anyone has an idea why this is happening? >>> >>> Best regards, >>> Lydia >>> >>> >>> 22:28:10,178 DEBUG org.apache.flink.runtime.operators.DataSourceTask - >>> Opening input split Locatable Split (2) at [grips5:60020]: DataSource (at >>> createInput(ExecutionEnvironment.java:502) >>> (org.apache.flink.HBaseReadExample$1)) (1/1) >>> 22:28:10,178 INFO org.apache.flink.addons.hbase.TableInputFormat >>> - opening split [2|[grips5:60020]|aaaaaaaa|-] >>> 22:28:10,189 DEBUG org.apache.zookeeper.ClientCnxn >>> - Reading reply sessionid:0x24ff6a96ecd000a, packet:: clientPath:null >>> serverPath:null finished:false header:: 3,4 replyHeader:: 3,51539607639,0 >>> request:: '/hbase/meta-region-server,F response:: >>> #ffffffff0001a726567696f6e7365727665723a363030$ >>> 22:28:10,202 DEBUG org.apache.zookeeper.ClientCnxn >>> - Reading reply sessionid:0x24ff6a96ecd000a, packet:: clientPath:null >>> serverPath:null finished:false header:: 4,4 replyHeader:: 4,51539607639,0 >>> request:: '/hbase/meta-region-server,F response:: >>> #ffffffff0001a726567696f6e7365727665723a363030$ >>> 22:28:10,211 DEBUG LocalActorRefProvider(akka://flink) >>> - resolve of path sequence [/temp/$b] failed >>> 22:28:10,233 DEBUG org.apache.hadoop.hbase.util.ByteStringer >>> - Failed to classload HBaseZeroCopyByteString: >>> java.lang.IllegalAccessError: class >>> com.google.protobuf.HBaseZeroCopyByteString cannot access its superclass >>> com.google.protobuf.LiteralByteString >>> 22:28:10,358 DEBUG org.apache.hadoop.ipc.RpcClient >>> - Use SIMPLE authentication for service ClientService, sasl=false >>> 22:28:10,370 DEBUG org.apache.hadoop.ipc.RpcClient >>> - Connecting to grips1/130.73.20.14:60020 >>> 22:28:10,380 DEBUG org.apache.hadoop.ipc.RpcClient >>> - IPC Client (2145423150) connection to grips1/130.73.20.14:60020 from >>> hduser: starting, connections 1 >>> 22:28:10,394 DEBUG org.apache.hadoop.ipc.RpcClient >>> - IPC Client (2145423150) connection to grips1/130.73.20.14:60020 from >>> hduser: got response header call_id: 0, totalSize: 469 bytes >>> 22:28:10,397 DEBUG org.apache.hadoop.ipc.RpcClient >>> - IPC Client (2145423150) connection to grips1/130.73.20.14:60020 from >>> hduser: wrote request header call_id: 0 method_name: "Get" request_param: >>> true >>> 22:28:10,413 DEBUG org.apache.zookeeper.ClientCnxn >>> - Reading reply sessionid:0x24ff6a96ecd000a, packet:: clientPath:null >>> serverPath:null finished:false header:: 5,4 replyHeader:: 5,51539607639,0 >>> request:: '/hbase/meta-region-server,F response:: >>> #ffffffff0001a726567696f6e7365727665723a363030$ >>> 22:28:10,424 DEBUG org.apache.hadoop.ipc.RpcClient >>> - IPC Client (2145423150) connection to grips1/130.73.20.14:60020 from >>> hduser: wrote request header call_id: 1 method_name: "Scan" request_param: >>> true priority: 100 >>> 22:28:10,426 DEBUG org.apache.hadoop.ipc.RpcClient >>> - IPC Client (2145423150) connection to grips1/130.73.20.14:60020 from >>> hduser: got response header call_id: 1 cell_block_meta { length: 480 }, >>> totalSize: 497 bytes >>> 22:28:10,432 DEBUG org.apache.hadoop.hbase.client.ClientSmallScanner - >>> Finished with small scan at {ENCODED => 1588230740, NAME => >>> 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} >>> 22:28:10,434 DEBUG org.apache.hadoop.ipc.RpcClient >>> - Use SIMPLE authentication for service ClientService, sasl=false >>> 22:28:10,434 DEBUG org.apache.hadoop.ipc.RpcClient >>> - Connecting to grips5/130.73.20.16:60020 >>> 22:28:10,435 DEBUG org.apache.hadoop.ipc.RpcClient >>> - IPC Client (2145423150) connection to grips5/130.73.20.16:60020 from >>> hduser: wrote request header call_id: 2 method_name: "Scan" request_param: >>> true >>> 22:28:10,436 DEBUG org.apache.hadoop.ipc.RpcClient >>> - IPC Client (2145423150) connection to grips5/130.73.20.16:60020 from >>> hduser: starting, connections 2 >>> 22:28:10,437 DEBUG org.apache.hadoop.ipc.RpcClient >>> - IPC Client (2145423150) connection to grips5/130.73.20.16:60020 from >>> hduser: got response header call_id: 2, totalSize: 12 bytes >>> 22:28:10,438 DEBUG org.apache.flink.runtime.operators.DataSourceTask - >>> Starting to read input from split Locatable Split (2) at [grips5:60020]: >>> DataSource (at createInput(ExecutionEnvironment.java:502) >>> (org.apache.flink.HBaseReadExample$1)) (1/1) >>> 22:28:10,438 DEBUG org.apache.hadoop.ipc.RpcClient >>> - IPC Client (2145423150) connection to grips5/130.73.20.16:60020 from >>> hduser: wrote request header call_id: 3 method_name: "Scan" request_param: >>> true >>> 22:28:10,457 DEBUG org.apache.hadoop.ipc.RpcClient >>> - IPC Client (2145423150) connection to grips5/130.73.20.16:60020 from >>> hduser: got response header call_id: 3 cell_block_meta { length: 4679 }, >>> totalSize: 4899 bytes >>> 22:28:10,476 DEBUG org.apache.hadoop.ipc.RpcClient >>> - IPC Client (2145423150) connection to grips5/130.73.20.16:60020 from >>> hduser: wrote request header call_id: 4 method_name: "Scan" request_param: >>> true >>> 22:28:10,480 DEBUG org.apache.hadoop.ipc.RpcClient >>> - IPC Client (2145423150) connection to grips5/130.73.20.16:60020 from >>> hduser: got response header call_id: 4 cell_block_meta { length: 3306 }, >>> totalSize: 3466 bytes >>> 22:28:10,482 DEBUG org.apache.hadoop.ipc.RpcClient >>> - IPC Client (2145423150) connection to grips5/130.73.20.16:60020 from >>> hduser: got response header call_id: 5, totalSize: 8 bytes >>> 22:28:10,482 DEBUG org.apache.hadoop.ipc.RpcClient >>> - IPC Client (2145423150) connection to grips5/130.73.20.16:60020 from >>> hduser: wrote request header call_id: 5 method_name: "Scan" request_param: >>> true >>> 22:28:10,487 DEBUG org.apache.flink.runtime.operators.DataSourceTask - >>> Closing input split Locatable Split (2) at [grips5:60020]: DataSource (at >>> createInput(ExecutionEnvironment.java:502) >>> (org.apache.flink.HBaseReadExample$1)) (1/1) >>> 22:28:10,489 INFO >>> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation >>> - Closing zookeeper sessionid=0x24ff6a96ecd000a >>> 22:28:10,489 DEBUG org.apache.zookeeper.ZooKeeper >>> - Closing session: 0x24ff6a96ecd000a >>> 22:28:10,489 DEBUG org.apache.zookeeper.ClientCnxn >>> - Closing client for session: 0x24ff6a96ecd000a >>> 22:28:10,499 DEBUG org.apache.zookeeper.ClientCnxn >>> - Reading reply sessionid:0x24ff6a96ecd000a, packet:: clientPath:null >>> serverPath:null finished:false header:: 6,-11 replyHeader:: >>> 6,51539607640,0 request:: null response:: null >>> 22:28:10,499 DEBUG org.apache.zookeeper.ClientCnxn >>> - Disconnecting client for session: 0x24ff6a96ecd000a >>> 22:28:10,499 INFO org.apache.zookeeper.ClientCnxn >>> - EventThread shut down >>> 22:28:10,499 INFO org.apache.zookeeper.ZooKeeper >>> - Session: 0x24ff6a96ecd000a closed >>> 22:28:10,499 DEBUG org.apache.hadoop.ipc.RpcClient >>> - Stopping rpc client >>> 22:28:10,501 DEBUG org.apache.hadoop.ipc.RpcClient >>> - IPC Client (2145423150) connection to grips1/130.73.20.14:60020 from >>> hduser: closed >>> 22:28:10,502 DEBUG org.apache.hadoop.ipc.RpcClient >>> - IPC Client (2145423150) connection to grips1/130.73.20.14:60020 from >>> hduser: stopped, connections 0 >>> 22:28:10,502 DEBUG org.apache.hadoop.ipc.RpcClient >>> - IPC Client (2145423150) connection to grips5/130.73.20.16:60020 from >>> hduser: closed >>> 22:28:10,502 DEBUG org.apache.hadoop.ipc.RpcClient >>> - IPC Client (2145423150) connection to grips5/130.73.20.16:60020 from >>> hduser: stopped, connections 0 >>> 22:28:10,502 INFO org.apache.flink.addons.hbase.TableInputFormat >>> - Closing split (scanned 170 rows) >>> 22:28:10,508 DEBUG org.apache.flink.runtime.operators.DataSourceTask - >>> Opening input split Locatable Split (1) at [grips4:60020]: DataSource (at >>> createInput(ExecutionEnvironment.java:502) >>> (org.apache.flink.HBaseReadExample$1)) (1/1) >>> 22:28:10,509 INFO org.apache.flink.addons.hbase.TableInputFormat >>> - opening split [1|[grips4:60020]|55555555|aaaaaaaa] >>> 22:28:11,380 DEBUG org.apache.flink.runtime.taskmanager.TaskManager >>> - Received message SendHeartbeat at akka://flink/user/taskmanager from >>> Actor[akka://flink/deadLetters]. >>> 22:28:11,380 DEBUG org.apache.flink.runtime.taskmanager.TaskManager >>> - Sending heartbeat to JobManager >