On Thursday 5 September 2024 19:55:56 BST Frank Steinmetzger wrote:
> Am Thu, Sep 05, 2024 at 06:30:54AM -0500 schrieb Dale:
> > > Use rsync with:
> > >  --checksum
> > > 
> > > and
> > > 
> > >  --dry-run
> 
> I suggest calculating a checksum file from your active files. Then you don’t
> have to read the files over and over for each backup iteration you compare
> it against.
> 
> > > You can also run find to identify which files were changed during the
> > > period you were running with the dodgy RAM.  Thankfully you didn't run
> > > for too long before you spotted it.
> 
> This. No need to check everything you ever stored. Just the most recent
> stuff, or at maximum, since you got the new PC.
> 
> > I have just shy of 45,000 files in 780 directories or so.  Almost 6,000
> > in another.  Some files are small, some are several GBs or so.  Thing
> > is, backups go from a single parent directory if you will.  Plus, I'd
> > want to compare them all anyway.  Just to be sure.
> 
> I aqcuired the habit of writing checksum files in all my media directories
> such as music albums, tv series and such, whenever I create one such
> directory. That way even years later I can still check whether the files are
> intact. I actually experienced broken music files from time to time (mostly
> on the MicroSD card in my tablet). So with checksum files, I can verify
> which file is bad and which (on another machine) is still good.

There is also dm-verity for a more involved solution.  I think for Dale 
something like this should work:

find path-to-directory/ -type f | xargs md5sum > digest.log

then to compare with a backup of the same directory you could run:

md5sum -c digest.log | grep FAILED

Someone more knowledgeable should be able to knock out some clever python 
script to do the same at speed.

Attachment: signature.asc
Description: This is a digitally signed message part.

Reply via email to