Wonderful, just as I expected. Do folks normally have several RGW running
on individual machines with a load balancer at larger scales?

On Wed, Dec 21, 2016 at 8:22 AM, LOPEZ Jean-Charles <jelo...@redhat.com>
wrote:

> Hi Gerald,
>
> for the s3 and swift case, the clients are not accessing the ceph cluster.
> They are s3 and swift clients and only discuss with the RGW over HTTP. The
> RGW is the ceph client that does all the interaction with the ceph cluster.
>
> Best
> JC
>
> On Dec 21, 2016, at 07:27, Gerald Spencer <ger.spenc...@gmail.com> wrote:
>
> I was under the impression that when a client talks to the cluster, it
> grabs the osd map and computes the crush algorithm to determine where it
> stores the object. Does the rgw server do this for clients? If I had 12
> clients all talking through one gateway, would that server have to pass all
> of the objects from the clients to the cluster?
>
>
> And 48 osd nodes, each with 12 x 6TB drives and a PCIe write journal. That
> would be 576 osds in the cluster, with about 3.4PB raw...
>
>
> On Tue, Dec 20, 2016 at 1:12 AM Wido den Hollander <w...@42on.com> wrote:
>
>>
>>
>> > Op 20 december 2016 om 3:24 schreef Gerald Spencer <
>> ger.spenc...@gmail.com>:
>>
>> >
>>
>> >
>>
>> > Hello all,
>>
>> >
>>
>> > We're currently waiting on a delivery of equipment for a small 50TB
>> proof
>>
>> > of concept cluster, and I've been lurking/learning a ton from you.
>> Thanks
>>
>> > for how active everyone is.
>>
>> >
>>
>> > Question(s):
>>
>> > How does the raids gateway work exactly?
>>
>>
>>
>> The RGW doesn't do any RAID. It chunks up larger objects into smaller
>> RADOS chunks. The first chunk is always 512k (IIRC) and then it chunks up
>> into 4MB RADOS objects.
>>
>>
>>
>> > Does it introduce a single point of failure?
>>
>>
>>
>> It does if you deploy only one RGW. Always deploy multiple with
>> loadbalancing in front.
>>
>>
>>
>> > Does all of the traffic go through the host running the rgw server?
>>
>>
>>
>> Yes it does.
>>
>>
>>
>> >
>>
>> > I just don't fully understand that side of things. As for architecture
>> our
>>
>> > poc will have:
>>
>> > - 1 monitor
>>
>> > - 4 OSDs with 12 x 6TB drives, 1 x 800 PCIe journal
>>
>> >
>>
>>
>>
>> Underscaled machines, go for less disks per machine but more machines.
>> More smaller machines works a lot better with Ceph then a few big machines.
>>
>>
>>
>> > I'd all goes as planned, this will scale up to:
>>
>> > - 3 monitors
>>
>>
>>
>> Always run with 3 MONs. Otherwise it is a serious SPOF.
>>
>>
>>
>> > - 48 osds
>>
>> >
>>
>> > This should give us enough storage (~1.2PB) wth enough throughput to
>> handle
>>
>> > the data requirements of our machines to saturate our 100Gb link...
>>
>> >
>>
>>
>>
>> That won't happen with just 4 machines. Replica 3x taken into account is
>> well. You will need a lot more machines to get the 100Gb link fully
>> utilized.
>>
>>
>>
>> Wido
>>
>>
>>
>> >
>>
>> >
>>
>> >
>>
>> >
>>
>> > Cheers,
>>
>> > G
>>
>> > _______________________________________________
>>
>> > ceph-users mailing list
>>
>> > ceph-users@lists.ceph.com
>>
>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to