David Howe schreef:
Jamie Lokier wrote:
There are methods to perform efficient updates of large numbers of
files and a large amount of data, across simultaneous renames, copies
and edits. But that is the realm of "similarity detection indexing",
which is beyond the scope of rsync. At least with the present design,
which doesn't use persistent indexing of any kind. (This is a feature!)
I can see that being a lot of work, yes - but in the less complex case
(where you have "deleted" directories and "new" directories), it might
be worthwhile to look for exact matches (file+stamp) in the sacrificial
directories for a mv before they are deleted, which would catch the
majority of files in a worst-case scenario of the root directories being
renamed or moved.
I am less worried about individual file renames and/or "missing" the
opportunity to diff a large file that has been both moved and updated,
than having to resync multiple gigs of stuff over a slow link, because
some user renamed a directory.
Agreed. My solution is written in Perl, the lbackup is in Python. I am
looking for some unity.
Regards, Nico
--
Behandeld door / Handled by: N.J. van der Horn (Nico)
---
ICT Support Vanderhorn IT-works, www.vanderhorn.nl,
Kamer van Koophandel / Chambre of Commerce 24228233,
Voorstraat 55, 3135 HW Vlaardingen, The Netherlands,
Tel +31 10 2486060, Fax +31 10 2486061
--
Please use reply-all for most replies to avoid omitting the mailing list.
To unsubscribe or change options: https://lists.samba.org/mailman/listinfo/rsync
Before posting, read: http://www.catb.org/~esr/faqs/smart-questions.html