Most likely, they'd has well across the OSDs - but if I'm correct in my own
research (still a bit of a Ceph noob, myself) you'd still want more PGs.

An OSD can process several things at once - and should, for peak
throughput.  I wouldn't be surprised if some operations serialize within a
PG; if so, one PG per OSD isn't going to be able to max out the potential
of your OSDs.  I suspect there's reasoning along this line behind Ceph's
recommendation for PGs per OSD - I believe best practice is somewhere
between 10-30.



--
Mike Shuey

On Thu, May 5, 2016 at 7:58 PM, Roland Mechler <rmech...@opendns.com> wrote:

> Thanks for your response. So... if I configured 3 PGs for the pool, would
> they necessarily each have their primary on a different OSD, thus spreading
> the load? Or, would it be better to have more PGs to ensure an even
> distribution?
>
> I was also wondering about the pros and cons performance wise of having a
> pool size of 3 vs 2. It seems there would be a benefit for reads (1.5 times
> the bandwidth) but a penalty for writes because the primary has to forward
> to 2 nodes instead of 1. Does that make sense?
>
> -Roland
>
> On Thu, May 5, 2016 at 4:13 PM, Michael Shuey <sh...@fmepnet.org> wrote:
>
>> Reads will be limited to 1/3 of the total bandwidth.  A set of PGs has
>> a "primary" - that's the first one (and only one, if it's up & in)
>> consulted on a read.  The other PGs will still exist, but they'll only
>> take writes (and only after the primary PG forwards along data).  If
>> you have multiple PGs, reads (and write-mastering duties) will be
>> spread across all 3 servers.
>>
>>
>> --
>> Mike Shuey
>>
>>
>> On Thu, May 5, 2016 at 5:36 PM, Roland Mechler <rmech...@opendns.com>
>> wrote:
>> > Let's say I have a small cluster (3 nodes) with 1 OSD per node. If I
>> create
>> > a pool with size 3, such that each object in the pool will be
>> replicated to
>> > each OSD/node, is there any reason to create the pool with more than 1
>> PG?
>> > It seems that increasing the number of PGs beyond 1 would not provide
>> any
>> > additional benefit in terms of data balancing or durability, and would
>> have
>> > a cost in terms of resource usage. But when I try this, I get a "pool
>> <pool>
>> > has many more objects per pg than average (too few pgs?)" warning from
>> ceph
>> > health. Is there a cost to having a large number of objects per PG?
>> >
>> > -Roland
>> >
>> > _______________________________________________
>> > ceph-users mailing list
>> > ceph-users@lists.ceph.com
>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> >
>>
>
>
>
> --
> Roland Mechler
> Site Reliability Engineer
> [image: OpenDNS] <http://opendns.com/>
> Mobile:
> 604-727-5257
> Email:
> rmech...@cisco.com
> *OpenDNS Vancouver* <http://opendns.com/>
> 675 West Hastings St, Suite 500
> Vancouver, BC V6B 1N2
> Canada
> <http://maps.google.com/maps?q=675+West+Hastings+St,+Suite+200%2CVancouver%2C+BC%2C+Canada&hl=en>
>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to