On Thursday 5 September 2024 19:55:56 BST Frank Steinmetzger wrote: > Am Thu, Sep 05, 2024 at 06:30:54AM -0500 schrieb Dale: > > > Use rsync with: > > > --checksum > > > > > > and > > > > > > --dry-run > > I suggest calculating a checksum file from your active files. Then you don’t > have to read the files over and over for each backup iteration you compare > it against. > > > > You can also run find to identify which files were changed during the > > > period you were running with the dodgy RAM. Thankfully you didn't run > > > for too long before you spotted it. > > This. No need to check everything you ever stored. Just the most recent > stuff, or at maximum, since you got the new PC. > > > I have just shy of 45,000 files in 780 directories or so. Almost 6,000 > > in another. Some files are small, some are several GBs or so. Thing > > is, backups go from a single parent directory if you will. Plus, I'd > > want to compare them all anyway. Just to be sure. > > I aqcuired the habit of writing checksum files in all my media directories > such as music albums, tv series and such, whenever I create one such > directory. That way even years later I can still check whether the files are > intact. I actually experienced broken music files from time to time (mostly > on the MicroSD card in my tablet). So with checksum files, I can verify > which file is bad and which (on another machine) is still good.
There is also dm-verity for a more involved solution. I think for Dale something like this should work: find path-to-directory/ -type f | xargs md5sum > digest.log then to compare with a backup of the same directory you could run: md5sum -c digest.log | grep FAILED Someone more knowledgeable should be able to knock out some clever python script to do the same at speed.
signature.asc
Description: This is a digitally signed message part.