Hi John,

1. How can I see/determine the number of OSDs a pool can access?
2. For ".rgw", ".rgw.buckets", and ".rgw.buckets.index" how should I plan
out the PG number for these?  We're only doing object storage, so
.rgw.buckets will get the most objects, and .rgw.buckets.index will (I
assume) get a smaller amount by some ratio.
 >> About 1 and 2 , go visit http://ceph.com/pgcalc/ and caculate the pg by
pools/OSDs numbers.

3. What are .rgw, .rgw.buckets.extra, and .rgw.control used for?
Not very sure, but I remember .rgw used for domain root.
                      .rgw.control for logs.
                      .rgw > unknown.




Best wishes,
Mika


2016-02-13 9:45 GMT+08:00 John Hogenmiller (yt) <j...@yourtech.us>:

> I started a cluster with 9 OSD across 3 nodes. Then I expanded it to 419
> OSDs across 7 nodes.  Along the way, I increased the pg_num/pgp_num in the
> rbd pool. Thanks to help earlier on this list, I was able to do that.
>
> Tonight I started to do some perf testing and quickly realized that I
> never updated pg_num/pgp_num on the .rgw.buckets and rgw.buckets.index
> pools.   When I went to do so, I was told the pg_num was too high with "new
> PGs on ~8 OSDs exceed per-OSD max of 32".  I did a lot of reading up on
> osds, pg's, and crush map to try and figure out why .rgw.buckets was only
> seeing 8 OSDs, when it should see 419.
>
> Through trial and error, I did increase it to the max I could, and let
> everything rebalance as far as it could. I then found that the number of
> OSDs for the pool increased. The only way I know currently to see the
> number of OSDs a pool can see is to attempt to set pg_num to some
> ridiculously high value and get the error.
>
> ```
> root@ljb01:# ceph osd pool set .rgw.buckets.index pg_num 9999999999
>
>
> 2016-02-13 01:27:06.913291 7f6a1c9cc700  0 -- :/3369124606 >>
> 172.29.4.154:6789/0 pipe(0x7f6a20064550 sd=3 :0 s=1 pgs=0 cs=0 l=1
> c=0x7f6a2005e220).fault
> Error E2BIG: specified pg_num 9999999999 is too large (creating 9999999743
> new PGs on ~256 OSDs exceeds per-OSD max of 32)
> ```
>
> My questions:
>
> 1. How can I see/determine the number of OSDs a pool can access?
> 2. For ".rgw", ".rgw.buckets", and ".rgw.buckets.index" how should I plan
> out the PG number for these?  We're only doing object storage, so
> .rgw.buckets will get the most objects, and .rgw.buckets.index will (I
> assume) get a smaller amount by some ratio.
> 3. What are .rgw, .rgw.buckets.extra, and .rgw.control used for?
>
> Just in general, I'm looking for more information on how to plan for
> configuring for receiving a lot of objects across a lot of OSDs.
>
> pool name                 KB      objects       clones     degraded
>  unfound           rd        rd KB           wr        wr KB
> .log                       0          127            0            0
>      0      5076804      5076677      3384550            0
> .rgw                    1607         8676            0            0
>      0       135382        94455        62465        17676
> .rgw.buckets        61241052        34145            0            0
>      0       790047     36795213       328715     64702857
> .rgw.buckets.extra            0          330            0            0
>        0      6910047      4607382         1723            0
> .rgw.buckets.index            0         8083            0            0
>        0     26064639     26084041     18862332            0
> .rgw.control               0            8            0            0
>      0            0            0            0            0
> .rgw.gc                    0           32            0            0
>      0       240948       249931       192926            0
> .rgw.root                  1            3            0            0
>      0          864          576            3            3
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to