Do you have a repro that you can share with us? If so, please file a jira and 
we'll take a look.

> On Mar 18, 2022, at 12:15 PM, James Brown <jbr...@easypost.com> wrote:
> 
> This in 4.0.3 after running nodetool snapshot that we're seeing sstables 
> change, yes.
> 
> James Brown
> Infrastructure Architect @ easypost.com <http://easypost.com/>
> 
> On 2022-03-18 at 12:06:00, Jeff Jirsa <jji...@gmail.com 
> <mailto:jji...@gmail.com>> wrote:
>> This is nodetool snapshot yes? 3.11 or 4.0?
>> 
>> In versions prior to 3.0, sstables would be written with -tmp- in the name, 
>> then renamed when complete, so an sstable definitely never changed once it 
>> had the final file name. With the new transaction log mechanism, we use one 
>> name and a transaction log to note what's in flight and what's not, so if 
>> the snapshot system is including sstables being written (from flush, from 
>> compaction, or from streaming), those aren't final and should be skipped.
>> 
>> 
>> 
>> 
>> On Fri, Mar 18, 2022 at 11:46 AM James Brown <jbr...@easypost.com 
>> <mailto:jbr...@easypost.com>> wrote:
>> We use the boring combo of cassandra snapshots + tar to backup our cassandra 
>> nodes; every once in a while, we'll notice tar failing with the following:
>> 
>> tar: 
>> data/addresses/addresses-eb0196100b7d11ec852b1541747d640a/snapshots/backup20220318183708/nb-167-big-Data.db:
>>  file changed as we read it
>> 
>> I find this a bit perplexing; what would cause an sstable inside a snapshot 
>> to change? The only thing I can think of is an incremental repair changing 
>> the "repaired_at" flag on the sstable, but it seems like that should 
>> "un-share" the hardlinked sstable rather than running the risk of mutating a 
>> snapshot.
>> 
>> 
>> James Brown
>> Cassandra admin @ easypost.com <http://easypost.com/>

Reply via email to