It's also good to note that only the Data files are compressed already.
Depending on your data the Index and other files may be a significant
percent of total on disk data.
On 05/02/2014 01:14 PM, tommaso barbugli wrote:
In my tests compressing with lzop sstables (with cassandra compression
t
In my tests compressing with lzop sstables (with cassandra compression
turned on) resulted in approx. 50% smaller files.
Thats probably because the chunks of data compressed by lzop are way bigger
than the average size of writes performed on Cassandra (not sure how data
is compressed but I guess it
On Fri, May 2, 2014 at 2:07 AM, tommaso barbugli wrote:
> If you are thinking about using Amazon S3 storage I wrote a tool that
> performs snapshots and backups on multiple nodes.
> Backups are stored compressed on S3.
> https://github.com/tbarbugli/cassandra_snapshotter
>
https://github.com/Jere
Artur,
Replies inline.
On Fri, May 2, 2014 at 10:42 AM, Artur Kronenberg <
artur.kronenb...@openmarket.com> wrote:
> we are running a 7 node cluster with an RF of 5. Each node holds about 70%
> of the data and we are now wondering about the backup process.
>
What are you using for a backup proc
If you are thinking about using Amazon S3 storage I wrote a tool that
performs snapshots and backups on multiple nodes.
Backups are stored compressed on S3.
https://github.com/tbarbugli/cassandra_snapshotter
Cheers,
Tommaso
2014-05-02 10:42 GMT+02:00 Artur Kronenberg :
> Hi,
>
> we are running
Hi,
we are running a 7 node cluster with an RF of 5. Each node holds about
70% of the data and we are now wondering about the backup process.
1. Is there a best practice procedure or a tool that we can use to have
one backup that holds 100 % of the data or is it necessary for us to
take mult