Hi All,

Thanks for the help but after yet another day of investigation I think I
might be running into this
https://issues.apache.org/jira/browse/CASSANDRA-8061 issue where tmplink
files aren't removed until Cassandra is restarted.

Thanks again for all the suggestions!

Nate

--
*Nathanael Yoder*
Principal Engineer & Data Scientist, Whistle
415-944-7344 // n...@whistle.com

On Tue, Dec 9, 2014 at 10:18 AM, Nate Yoder <n...@whistle.com> wrote:

> Hi Reynald,
>
> Good idea but I have incremental backups turned off and other than *.db
> files nothing else appears to be in the data directory for that table.
>
> Is there any other output that would be helpful in helping you all help me?
>
> Thanks,
> Nate
>
> --
> *Nathanael Yoder*
> Principal Engineer & Data Scientist, Whistle
> 415-944-7344 // n...@whistle.com
>
> On Tue, Dec 9, 2014 at 9:27 AM, Reynald Bourtembourg <
> reynald.bourtembo...@esrf.fr> wrote:
>
>>  Hi Nate,
>>
>> Are you using incremental backups?
>>
>> Extract from the documentation (
>> http://www.datastax.com/documentation/cassandra/2.1/cassandra/operations/ops_backup_incremental_t.html
>> ):
>>
>> *When incremental backups are enabled (disabled by default), Cassandra
>> hard-links each flushed SSTable to a backups directory under the keyspace
>> data directory. This allows storing backups offsite without transferring
>> entire snapshots. Also, incremental backups combine with snapshots to
>> provide a dependable, up-to-date backup mechanism.*
>>
>> *As with snapshots, Cassandra does not automatically clear incremental
>> backup files. DataStax recommends setting up a process to clear incremental
>> backup hard-links each time a new snapshot is created.*
>>  These backups are stored in directories named "backups" at the same
>> level as the "snapshots' directories.
>>
>> Reynald
>>
>>
>> On 09/12/2014 18:13, Nate Yoder wrote:
>>
>> Thanks for the advice.  Totally makes sense.  Once I figure out how to
>> make my data stop taking up more than 2x more space without being useful
>> I'll definitely make the change :)
>>
>>  Nate
>>
>>
>>
>>   --
>> *Nathanael Yoder*
>> Principal Engineer & Data Scientist, Whistle
>> 415-944-7344 // n...@whistle.com
>>
>> On Tue, Dec 9, 2014 at 9:02 AM, Jonathan Haddad <j...@jonhaddad.com>
>> wrote:
>>
>>> Well, I personally don't like RF=2.  It means if you're using CL=QUORUM
>>> and a node goes down, you're going to have a bad time. (downtime) If you're
>>> using CL=ONE then you'd be ok.  However, I am not wild about losing a node
>>> and having only 1 copy of my data available in prod.
>>>
>>>
>>> On Tue Dec 09 2014 at 8:40:37 AM Nate Yoder <n...@whistle.com> wrote:
>>>
>>>> Thanks Jonathan.  So there is nothing too idiotic about my current
>>>> set-up with 6 boxes each with 256 vnodes each and a RF of 2?
>>>>
>>>>  I appreciate the help,
>>>> Nate
>>>>
>>>>
>>>>
>>>>   --
>>>> *Nathanael Yoder*
>>>> Principal Engineer & Data Scientist, Whistle
>>>> 415-944-7344 // n...@whistle.com
>>>>
>>>>  On Tue, Dec 9, 2014 at 8:31 AM, Jonathan Haddad <j...@jonhaddad.com>
>>>> wrote:
>>>>
>>>>> You don't need a prime number of nodes in your ring, but it's not a
>>>>> bad idea to it be a multiple of your RF when your cluster is small.
>>>>>
>>>>>
>>>>> On Tue Dec 09 2014 at 8:29:35 AM Nate Yoder <n...@whistle.com> wrote:
>>>>>
>>>>>> Hi Ian,
>>>>>>
>>>>>>  Thanks for the suggestion but I had actually already done that
>>>>>> prior to the scenario I described (to get myself some free space) and 
>>>>>> when
>>>>>> I ran nodetool cfstats it listed 0 snapshots as expected, so 
>>>>>> unfortunately
>>>>>> I don't think that is where my space went.
>>>>>>
>>>>>>  One additional piece of information I forgot to point out is that
>>>>>> when I ran nodetool status on the node it included all 6 nodes.
>>>>>>
>>>>>>  I have also heard it mentioned that I may want to have a prime
>>>>>> number of nodes which may help protect against split-brain.  Is this 
>>>>>> true?
>>>>>> If so does it still apply when I am using vnodes?
>>>>>>
>>>>>>  Thanks again,
>>>>>> Nate
>>>>>>
>>>>>>   --
>>>>>> *Nathanael Yoder*
>>>>>> Principal Engineer & Data Scientist, Whistle
>>>>>> 415-944-7344 // n...@whistle.com
>>>>>>
>>>>>> On Tue, Dec 9, 2014 at 7:42 AM, Ian Rose <ianr...@fullstory.com>
>>>>>> wrote:
>>>>>>
>>>>>>> Try `nodetool clearsnapshot` which will delete any snapshots you
>>>>>>> have.  I have never taken a snapshot with nodetool yet I found several
>>>>>>> snapshots on my disk recently (which can take a lot of space).  So 
>>>>>>> perhaps
>>>>>>> they are automatically generated by some operation?  No idea.  
>>>>>>> Regardless,
>>>>>>> nuking those freed up a ton of space for me.
>>>>>>>
>>>>>>>  - Ian
>>>>>>>
>>>>>>>
>>>>>>> On Mon, Dec 8, 2014 at 8:12 PM, Nate Yoder <n...@whistle.com> wrote:
>>>>>>>
>>>>>>>> Hi All,
>>>>>>>>
>>>>>>>>  I am new to Cassandra so I apologise in advance if I have missed
>>>>>>>> anything obvious but this one currently has me stumped.
>>>>>>>>
>>>>>>>>  I am currently running a 6 node Cassandra 2.1.1 cluster on EC2
>>>>>>>> using C3.2XLarge nodes which overall is working very well for us.  
>>>>>>>> However,
>>>>>>>> after letting it run for a while I seem to get into a situation where 
>>>>>>>> the
>>>>>>>> amount of disk space used far exceeds the total amount of data on each 
>>>>>>>> node
>>>>>>>> and I haven't been able to get the size to go back down except by 
>>>>>>>> stopping
>>>>>>>> and restarting the node.
>>>>>>>>
>>>>>>>>  For example, in my data I have almost all of my data in one
>>>>>>>> table.  On one of my nodes right now the total space used (as reported 
>>>>>>>> by
>>>>>>>> nodetool cfstats) is 57.2 GB and there are no snapshots. However, when 
>>>>>>>> I
>>>>>>>> look at the size of the data files (using du) the data file for that 
>>>>>>>> table
>>>>>>>> is 107GB.  Because the C3.2XLarge only have 160 GB of SSD you can see 
>>>>>>>> why
>>>>>>>> this quickly becomes a problem.
>>>>>>>>
>>>>>>>>  Running nodetool compact didn't reduce the size and neither does
>>>>>>>> running nodetool repair -pr on the node.  I also tried nodetool flush 
>>>>>>>> and
>>>>>>>> nodetool cleanup (even though I have not added or removed any nodes
>>>>>>>> recently) but it didn't change anything either.  In order to keep my
>>>>>>>> cluster up I then stopped and started that node and the size of the 
>>>>>>>> data
>>>>>>>> file dropped to 54GB while the total column family size (as reported by
>>>>>>>> nodetool) stayed about the same.
>>>>>>>>
>>>>>>>>  Any suggestions as to what I could be doing wrong?
>>>>>>>>
>>>>>>>>  Thanks,
>>>>>>>> Nate
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>
>>>>
>>
>>
>

Reply via email to