Hi John, 
please find my answers below - just to clarify, you are talking about the 
storage for the instances right? 



Le 3 oct. 2013 à 11:04, John Ashford <logica...@hotmail.com> a écrit :

> 1 – Glusterfs V Ceph
> Im reading a lot of different opinions about which of these is the best 
> storage backend. My need is for a fully stable product that has fault 
> tolerance built in. It needs to support maybe 400 low traffic web sites and a 
> few very high traffic. I saw a Redhat diag suggesting throughput on a Gbit 
> nic with 2 storage servers (Glusterfs) would be around 200Mbps. I can put 
> quad nics in the 2 or 3 storage machines to give extra breathing room. 
> Gluster is of course a mature product and has Redhat pushing it forward but 
> some people complain of speed issues. Any real life experiences of throughput 
> using Ceph? I know Ceph is new but it seems there is considerable weight 
> behind its development so while some say its not production ready I wonder if 
> anyone has the experience to refute/concur?

Based on my experience, and other folks around' experience, we all noticed that 
GlusterFS doesn't seem to cope well as a shared backend for the instances . 
With my setup, around 9 compute nodes, all connected through Gbit, and Sata 
drives, I didn't had the 200 Mpbs/sec. What was their protocol? Measuring 
access speed, latency, and read-time from the shared directory AND from the 
instance itself should give you a good idea of what to expect with your 
infrastructure.
Regarding CephFS, it's getting better, but last time I checked, everytime I was 
running either iozone or vdbench on that, the server kept crashing, and I had 
to manually rebooting it. Next to that, the MDS are not active/active (not sure 
about the Cuttlefish release though), which means that solely one is accessed 
when an object is requested. I always found Ceph somehow *better* than 
GlusterFS, more realiable (based on my testing), even if it's more complicated 
to manage - you have a real overwiew of what's going on within your cluster. 
Everytime I had to debug GlusterFS, it was a real pain ; not to mention I had 
data corrupted and even missing sometime more than one. 
Regarding the load, glusterfs deamons sucked out all the cpu constantly for
 replicating objects.

So I say, deploy both, test, and pick one (iozone, vdbench, dd are three good 
tools to bench your storage)
>  
> 2 – vm instances on clustered storage
You meant on local storage ?
> Im reading how if you run your vm instances on gluster/ceph you benefit from 
> live migration and faster access times since disk access is usually to the 
> local disk. I just need to clarify this – and this may fully expose my 
> ignorance -  but surely the instance runs on the compute node, not storage 
> node so I don’t get how people are claiming its faster running vm instance on 
> the storage cluster unless they are actually running compute on the storage 
> cluster in which case you don’t have proper separation of compute/storage. 
> Also, you would have the networking overhead unless running a compute node on 
> storage cluster? What am I missing?!
When you go with local storage, you lose some benefits, but gain others : 

Pros for local : 
        - More performance
        - Easier to manage (not any skill is necessary to manage the disk)
        - Easier to repair (mkfs_adm check, fsck, etc…)
        - Doesn't rely on network availability

Cons for local :        
        - Live migration takes much much longer since you will need to copy the 
disk over
        - Lose a node, lose everything : instances running on it, storage
        - if you lose the disk, without backup you are screwed. WIth backup, 
it'll take much time to restore
        - If you remove a file, you are screwed (lets say "terminate" instead 
of rebooting)

Pros for shared : 
        - Easier to balance the workload in minutes
        - More reliable since the file is copied in multiple locations
        - If you lose a node or a disk, you can retrieve the file 
        - MooseFS only feature : it has a trash, if you delete a file (let's 
say a terminated instance), you can browse the MooseFS trash (or 
"quanrantaine"). That's a sweet feature

Cons for shared : 
        - More costly in terms of configuration, network ports and knowledge
        - Some corruption can happen if the algorithm doesn't cope well with 
split-brains situations
        - Huge network-load (induced by the replicas, check, recycling, etc..) 
that requires a good architecting (dedicated network)

>  
>  Thanks
> John

Regards,
Razique

Attachment: signature.asc
Description: Message signed with OpenPGP using GPGMail

_______________________________________________
Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
Post to     : openstack@lists.openstack.org
Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack

Reply via email to