Hello, 

i`m using rsync to sync large virtual machine files from one esx server to 
another. 

rsync is running inside the so called "esx console" which is basically a 
specially crafted linux vm with some restrictions.

the speed is "reasonable", but i guess it`s not the optimum - at least i donŽt 
know where the bottleneck is.

i`m not using ssh as transport but run rsync in deamon mode on the target. so 
this speeds things up if large amounts of data go over the wire.

i read that rsync would be not very efficient with ultra-large files (i`m 
syncing files with up to 80gb size)

regarding the bottleneck:  neither cpu, network or disk is at their limits - 
neither on the source nor on the destination system.
i don`t see 100% cpu, i don`t see 100% network or 100% disk i/o usage

furthermore, i wonder:
isn`t rsync just too intelligent for such file transfers, as the position of 
data inside that files (containing harddisk-images) won`t really change?
i.e. we don`t need to check for data relocation, we just need to know if some 
blocks changed inside a block of size "x" and if there was a change, we could 
transfer that whole block again. so i wonder if we need a "rolling checksum" at 
all to handle this. wouldn`t checksums over fixed block size be sufficient for 
this task? 

regards
roland

______________________________________________________
GRATIS für alle WEB.DE-Nutzer: Die maxdome Movie-FLAT!
Jetzt freischalten unter http://movieflat.web.de

-- 
Please use reply-all for most replies to avoid omitting the mailing list.
To unsubscribe or change options: https://lists.samba.org/mailman/listinfo/rsync
Before posting, read: http://www.catb.org/~esr/faqs/smart-questions.html

Reply via email to