Hey,

Thanks all for the help!

Best,
João

> On 23 Apr 2015, at 23:54, Dejan Menges <[email protected]> wrote:
> 
> Hi,
> 
> Yes, that's the one.
> 
> FYI - we are using Hortonworks, so it's even part of patchset for 2.1
> 
> Cheers,
> Dejan
> 
> On Thu, Apr 23, 2015, 22:25 Enis Söztutar <[email protected]> wrote:
> 
>> In case this is HBASE-11234, HDP-2.2 releases contain the fix.
>> 
>> Enis
>> 
>> On Thu, Apr 23, 2015 at 12:06 PM, Ted Yu <[email protected]> wrote:
>> 
>>> I think Dejan was referring to HBASE-11234
>>> 
>>> Cheers
>>> 
>>> On Thu, Apr 23, 2015 at 8:28 AM, Dejan Menges <[email protected]>
>>> wrote:
>>> 
>>>> Hi,
>>>> 
>>>> This is a known bug, there's fix already. We had it as well.
>>>> 
>>>> Cheers,
>>>> Dejan
>>>> 
>>>> On Thu, Apr 23, 2015 at 5:19 PM João Alves <[email protected]> wrote:
>>>> 
>>>>> Hi all,
>>>>> 
>>>>> I have a cluster with HDP 2.1 stack running HBase 0.98.0.2. I have
>> one
>>>>> HBase table where there is at least one row that is impossible to get
>>>> using
>>>>> either the java API or the hbase shell. I was unable to find online
>> any
>>>>> examples that encompass this particular situation, maybe you guys can
>>>> help
>>>>> me. The output error is the following:
>>>>> 
>>>>> 
>>>>> ERROR: java.io.IOException
>>>>>        at
>>>> org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2046)
>>>>>        at
>>> org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:92)
>>>>>        at
>>>>> 
>>>> 
>>> 
>> org.apache.hadoop.hbase.ipc.SimpleRpcScheduler.consumerLoop(SimpleRpcScheduler.java:160)
>>>>>        at
>>>>> 
>>>> 
>>> 
>> org.apache.hadoop.hbase.ipc.SimpleRpcScheduler.access$000(SimpleRpcScheduler.java:38)
>>>>>        at
>>>>> 
>>>> 
>>> 
>> org.apache.hadoop.hbase.ipc.SimpleRpcScheduler$1.run(SimpleRpcScheduler.java:110)
>>>>>        at java.lang.Thread.run(Thread.java:744)
>>>>> Caused by: java.lang.IndexOutOfBoundsException
>>>>>        at java.nio.Buffer.checkBounds(Buffer.java:559)
>>>>>        at java.nio.HeapByteBuffer.get(HeapByteBuffer.java:143)
>>>>>        at org.apache.hadoop.hbase.io
>>>>> 
>> .encoding.FastDiffDeltaEncoder$1.decode(FastDiffDeltaEncoder.java:489)
>>>>>        at org.apache.hadoop.hbase.io
>>>>> 
>>>> 
>>> 
>> .encoding.FastDiffDeltaEncoder$1.decodeNext(FastDiffDeltaEncoder.java:540)
>>>>>        at org.apache.hadoop.hbase.io
>>>>> 
>>>> 
>>> 
>> .encoding.BufferedDataBlockEncoder$BufferedEncodedSeeker.seekToKeyInBlock(BufferedDataBlockEncoder.java:336)
>>>>>        at org.apache.hadoop.hbase.io
>>>>> 
>>>> 
>>> 
>> .hfile.HFileReaderV2$EncodedScannerV2.loadBlockAndSeekToKey(HFileReaderV2.java:1134)
>>>>>        at org.apache.hadoop.hbase.io
>>>>> .hfile.HFileReaderV2$AbstractScannerV2.seekTo(HFileReaderV2.java:501)
>>>>>        at org.apache.hadoop.hbase.io
>>>>> .hfile.HFileReaderV2$AbstractScannerV2.seekTo(HFileReaderV2.java:515)
>>>>>        at
>>>>> 
>>>> 
>>> 
>> org.apache.hadoop.hbase.regionserver.StoreFileScanner.seekAtOrAfter(StoreFileScanner.java:238)
>>>>>        at
>>>>> 
>>>> 
>>> 
>> org.apache.hadoop.hbase.regionserver.StoreFileScanner.seek(StoreFileScanner.java:153)
>>>>>        at
>>>>> 
>>>> 
>>> 
>> org.apache.hadoop.hbase.regionserver.StoreScanner.seekScanners(StoreScanner.java:317)
>>>>>        at
>>>>> 
>>>> 
>>> 
>> org.apache.hadoop.hbase.regionserver.StoreScanner.<init>(StoreScanner.java:176)
>>>>>        at
>>>>> 
>>> org.apache.hadoop.hbase.regionserver.HStore.getScanner(HStore.java:1847)
>>>>>        at
>>>>> 
>>>> 
>>> 
>> org.apache.hadoop.hbase.regionserver.HRegion$RegionScannerImpl.<init>(HRegion.java:3716)
>>>>>        at
>>>>> 
>>>> 
>>> 
>> org.apache.hadoop.hbase.regionserver.HRegion.instantiateRegionScanner(HRegion.java:1890)
>>>>>        at
>>>>> 
>>>> 
>>> 
>> org.apache.hadoop.hbase.regionserver.HRegion.getScanner(HRegion.java:1876)
>>>>>        at
>>>>> 
>>>> 
>>> 
>> org.apache.hadoop.hbase.regionserver.HRegion.getScanner(HRegion.java:1853)
>>>>>        at
>>>>> org.apache.hadoop.hbase.regionserver.HRegion.get(HRegion.java:4738)
>>>>>        at
>>>>> org.apache.hadoop.hbase.regionserver.HRegion.get(HRegion.java:4712)
>>>>>        at
>>>>> 
>>>> 
>>> 
>> org.apache.hadoop.hbase.regionserver.HRegionServer.get(HRegionServer.java:2847)
>>>>>        at
>>>>> 
>>>> 
>>> 
>> org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:28857)
>>>>>        at
>>>> org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2008)
>>>>>        ... 5 more
>>>>> 
>>>>> 
>>>>> The description of the table is:
>>>>> 
>>>>> {NAME => 'd', DATA_BLOCK_ENCODING => 'FAST_DIFF', BLOOMFILTER =>
>> 'ROW',
>>>>> REPLICATION_SCOPE => '0', COMPRESSION => ' true
>>>>> SNAPPY', VERSIONS => '1', TTL => '2147483647', MIN_VERSIONS => '0',
>>>>> KEEP_DELETED_CELLS => 'false', BLOCKSIZE => '65536', IN_
>>>>> MEMORY => 'false', BLOCKCACHE => 'true'}, {NAME => 'm',
>>>>> DATA_BLOCK_ENCODING => 'FAST_DIFF', BLOOMFILTER => 'ROW', REPLICATIO
>>>>> N_SCOPE => '0', VERSIONS => '1', COMPRESSION => 'SNAPPY',
>> MIN_VERSIONS
>>>> =>
>>>>> '0', TTL => '2147483647', KEEP_DELETED_CELLS => 'f
>>>>> alse', BLOCKSIZE => '65536', IN_MEMORY => 'false', BLOCKCACHE =>
>>> 'true'}
>>>>> 
>>>>> Thanks for the help,
>>>>> João
>>>>> 
>>>>> 
>>>> 
>>> 
>> 

Reply via email to