Oh, that. There was a lot of talk about it, but it hasn't happened within rsync. I ended up writing my own client-server model utility in perl. We had a master copy of a distribution of EDI tools and views - 170GB or so in a couple million files, as I recall, and we had to keep up around 20 identical copies of it all around the world. I didn't dream of trying to implement the rsync algorithm. It worked strictly on timestamps, sizes, and filetypes, comparing the list from the master with the list from the replica. It made a very lightweight process for each replica, and the generation of the list for the master was done only once per sync. If your changes are mostly new files instead of small changes in large files, it might be what you need. If Tim Renwick is still monitoring this list, maybe he could tar it up and pass you a copy. It'd definitely require some porting for a new environment, unless you're replicating a Maxtor MaxAttach 4000 to others like it, using a Solaris box to handle the master replication tasks. Fortunately, it's commented out the wazoo, so to speak(which made it relatively painless for Philips to lay me off).
> It would be nice to have it read the data once, and then sync it to all > of the destinations once. IIRC, there was a move to do this at some > point. Am I right? Tim Conway Unix System Administration Contractor - IBM Global Services [EMAIL PROTECTED] -- To unsubscribe or change options: http://lists.samba.org/mailman/listinfo/rsync Before posting, read: http://www.catb.org/~esr/faqs/smart-questions.html