Ouch :/

Not to be a pedant, but you're sure HDFS is configured against the 2TB of
space (and not, say, the root partition)? You're sure the the temporary
file is growing to 2TB (hadoop fs -ls output)? Are you using any
BlockEncoding or Compression with this column family? Any other store/table
configuration? This happens repeatably? Can you provide jstack of the RS
process along with log lines while this file is growing excessively?

Thanks,
Nick

On Tue, Jan 13, 2015 at 9:47 AM, sathyafmt <[email protected]> wrote:

> Thanks Esteban. We do have lots of space ~2TB. The compaction starts on
> around a 300MB column and dies after consuming all the 2TB of space.
>
> sathya
>
>
>
> --
> View this message in context:
> http://apache-hbase.679495.n3.nabble.com/HBase-with-opentsdb-creates-huge-tmp-file-runs-out-of-hdfs-space-tp4067577p4067583.html
> Sent from the HBase User mailing list archive at Nabble.com.
>

Reply via email to