Do you have the hbase-site.xml available in the classpath? On 10 May 2016 23:10, "Palle" <pa...@sport.dk> wrote:
> HBase write problem > > Hi all. > > I have a problem writing to HBase. > > I am using a slightly modified example of this class to proof the concept: > > https://github.com/apache/flink/blob/master/flink-batch-connectors/flink-hbase/src/test/java/org/apache/flink/addons/hbase/example/HBaseWriteExample.java > > However all the HBase-specific stuff is exactly the same as in the > HBaseWriteExample. > > The problem I see is that the job never completes (been running for more > than an hour now) and it is only 13 key/value pairs that is to be written > to HBase :-) > I have tested the map/reduce stuff works if I replace the HBase connection > stuff with just a write to a text file - works OK. I have also tested that > I can insert data in HBase from a similar Hadoop MapReduce job. > > Here is the part of the code where I guess the problem is: > > @Override > public Tuple2<Text, Mutation> map(Tuple2<String, Integer> t) throws > Exception { > LOG.info("Tuple2 map() called"); > reuse.f0 = new Text(t.f0); > Put put = new Put(t.f0.getBytes()); > put.add(MasterConstants.CF_SOME, MasterConstants.COUNT, > Bytes.toBytes(t.f1)); > reuse.f1 = put; > return reuse; > } > }).output(new HadoopOutputFormat<Text, Mutation>(new > TableOutputFormat<Text>(), job)); > > env.execute("Flink HBase Event Count Hello World Test"); > > This code matches the code in the HBaseWriteExample.java I should think. > > The "Tuple2" log line I see exactly the 13 times I expect, and the last > log line I see is this: > 2016-05-10 21:48:42,715 INFO > org.apache.hadoop.hbase.mapreduce.TableOutputFormat - Created > table instance for event_type_count > > Any suggestions to what the problem could be? > > Thanks, > Palle >