Hi,

October 27 2014 5:07 PM, "Wido den Hollander" <w...@42on.com> wrote: 
> On 10/27/2014 04:30 PM, Mike wrote:
> 
>> Hello,
>> My company is plaining to build a big Ceph cluster for achieving and
>> storing data.
>> By requirements from customer - 70% of capacity is SATA, 30% SSD.
>> First day data is storing in SSD storage, on next day moving SATA storage.
> 
> How are you planning on moving this data? Do you expect Ceph to do this?
> 
> What kind of access to Ceph are you planning on using? RBD? Raw RADOS?
> The RADOS Gateway (S3/Swift)?
> 
>> By now we decide use a SuperMicro's SKU with 72 bays for HDD = 22 SSD +
>> 50 SATA drives.
> 
> That are some serious machines. It will require a LOT of CPU power in
> those machines to run 72 OSDs. Probably 4 CPUs per machine.
> 
>> Our racks can hold 10 this servers and 50 this racks in ceph cluster =
>> 36000 OSD's,
> 
> 36.000 OSDs shouldn't really be the problem, but you are thinking really
> big scale here.
> 

AFAIK, the OSDs should scale, since they only peer with ~100 others regardless 
of the cluster size. I wonder about the mon's though -- 36,000 OSDs will send a 
lot of pg_stats updates so the mon's will have some work to keep up. But the 
main issue I foresee is on the clients: don't be surprised when you see that 
each client needs close to 100k threads when connected to this cluster. A 
hypervisor with 10 VMs running would approach 1 million threads -- I have no 
idea if that will present any problems. There were discussions about limiting 
the number of client threads, but I don't know if there was any progress on 
that yet.

Anyway, it would be good to know if there are any current installations even 
close to this size (even in test). We are in the early days of planning a 10k 
OSD test, but haven't exceed ~1200 yet.

Cheers, Dan


>> With 4tb SATA drives and replica = 2 and nerfull ratio = 0.8 we have 40
>> Petabyte of useful capacity.
>> 
>> It's too big or normal use case for ceph?
> 
> No, it's not to big for Ceph. This is what it was designed for. But a
> setup like this shouldn't be taken lightly.
> 
> Think about the network connectivity required to connect all these
> machines and other decisions to be made.
> 
> _______________________________________________ 
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> 
> --
> Wido den Hollander
> 42on B.V.
> Ceph trainer and consultant
> 
> Phone: +31 (0)20 700 9902
> Skype: contact42on
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to