On 10/27/2014 05:32 PM, Dan van der Ster wrote:
> Hi,
> 
> October 27 2014 5:07 PM, "Wido den Hollander" <w...@42on.com> wrote: 
>> On 10/27/2014 04:30 PM, Mike wrote:
>>
>>> Hello,
>>> My company is plaining to build a big Ceph cluster for achieving and
>>> storing data.
>>> By requirements from customer - 70% of capacity is SATA, 30% SSD.
>>> First day data is storing in SSD storage, on next day moving SATA storage.
>>
>> How are you planning on moving this data? Do you expect Ceph to do this?
>>
>> What kind of access to Ceph are you planning on using? RBD? Raw RADOS?
>> The RADOS Gateway (S3/Swift)?
>>
>>> By now we decide use a SuperMicro's SKU with 72 bays for HDD = 22 SSD +
>>> 50 SATA drives.
>>
>> That are some serious machines. It will require a LOT of CPU power in
>> those machines to run 72 OSDs. Probably 4 CPUs per machine.
>>
>>> Our racks can hold 10 this servers and 50 this racks in ceph cluster =
>>> 36000 OSD's,
>>
>> 36.000 OSDs shouldn't really be the problem, but you are thinking really
>> big scale here.
>>
> 
> AFAIK, the OSDs should scale, since they only peer with ~100 others 
> regardless of the cluster size. I wonder about the mon's though -- 36,000 
> OSDs will send a lot of pg_stats updates so the mon's will have some work to 
> keep up. But the main issue I foresee is on the clients: don't be surprised 
> when you see that each client needs close to 100k threads when connected to 
> this cluster. A hypervisor with 10 VMs running would approach 1 million 
> threads -- I have no idea if that will present any problems. There were 
> discussions about limiting the number of client threads, but I don't know if 
> there was any progress on that yet.
> 

True about the mons. 3 monitors will not cut it here. You need 9 MONs at
least I think, on dedicated resources.

> Anyway, it would be good to know if there are any current installations even 
> close to this size (even in test). We are in the early days of planning a 10k 
> OSD test, but haven't exceed ~1200 yet.
> 
> Cheers, Dan
> 
> 
>>> With 4tb SATA drives and replica = 2 and nerfull ratio = 0.8 we have 40
>>> Petabyte of useful capacity.
>>>
>>> It's too big or normal use case for ceph?
>>
>> No, it's not to big for Ceph. This is what it was designed for. But a
>> setup like this shouldn't be taken lightly.
>>
>> Think about the network connectivity required to connect all these
>> machines and other decisions to be made.
>>
>> _______________________________________________ 
>>> ceph-users mailing list
>>> ceph-users@lists.ceph.com
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>> --
>> Wido den Hollander
>> 42on B.V.
>> Ceph trainer and consultant
>>
>> Phone: +31 (0)20 700 9902
>> Skype: contact42on
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


-- 
Wido den Hollander
42on B.V.
Ceph trainer and consultant

Phone: +31 (0)20 700 9902
Skype: contact42on
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to