I have been reading the architecture section of ceph document. One thing has 
not been clear to me is how the data HA works when we encounter OSD or server 
failure. Does the Crush algorithm recalculate based on the new cluster map and 
point the data to the 2nd or 3rd replica for existing data block read or write? 
Given the 2nd or 3rd replica's location (OSDs) were calculated by the primary 
OSD instead of the client, this is not clearly to me if and how this is being 
done.
A related question to the data HA mechanism, if client (librados) does 
recalculate the primary OSD location and point to the 2nd OSD, how much latency 
or how long the IO hang will client (e.g., VM) experience on an average load 
scenario? In our traditional commercial hypervisor environment, we experienced 
SCSI time out and Linux guest OS file system turns to readonly mode due to NFS 
Datastore/Network hiccups.
Thanks. --weiguo
                                          
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to