From my experience you’ll be better off planning exactly how many OSD’s and 
nodes you’re going to have and if possible equip them from the start.

By just adding a new drive to the same pool ceph will start to rearrange data 
across the whole cluster which might lead to less client IO depending on what 
you’re comfortable with. In a worst case scenario, your clients won’t have 
enough IO and your services might be ”down” until it’s healthy again.

Rebuilding 60 x 6TB drives will take quite some time. Each SATA drive has about 
75MB-125MB throughput at best, so a rebuild of once such drive would take 
approx. 16-17 hours. Usually it takes some x2 or x3 times longer in a normal 
case and if your controllers or network is limited.

// david


> 21 mars 2016 kl. 13:13 skrev Bastian Rosner <b...@d00m.org>:
> 
> Yes, rebuild in case of a whole chassis failure is indeed an issue. That 
> depends on how the failure domain looks like.
> 
> I'm currently thinking of initially not running fully equipped nodes.
> Let's say four of these machines with 60x 6TB drives each, so only loaded 2/3.
> That's raw 1440TB distributed over eight OSD nodes.
> Each individual OSD-node would therefore host "only" 30 OSDs but still allow 
> for fast expansion.
> 
> Usually delivery and installation of a bunch of HDDs is much faster than 
> servers.
> 
> I really wonder how easy it is to add additional disks and whether chance for 
> node- or even chassis-failure increases.
> 
> Cheers, Bastian
> 
> Am 2016-03-21 10:33, schrieb David:
>> Sounds like you’ll have a field day waiting for rebuild in case of a
>> node failure or an upgrade of the crush map ;)
>> David
>>> 21 mars 2016 kl. 09:55 skrev Bastian Rosner <b...@d00m.org>:
>>> Hi,
>>> any chance that somebody here already got hands on Dell DSS 7000 machines?
>>> 4U chassis containing 90x 3.5" drives and 2x dual-socket server sleds 
>>> (DSS7500). Sounds ideal for high capacity and density clusters, since each 
>>> of the server-sleds would run 45 drives, which I believe is a suitable 
>>> number of OSDs per node.
>>> When searching for this model there's not much detailed information out 
>>> there.
>>> Sadly I could not find a review from somebody who actually owns a bunch of 
>>> them and runs a decent PB-size cluster with it.
> 
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to