Thanks Janne!! I deleted all the pools. A few default rgw pools got auto-created, and the rest I created manually. Now Ceph looks happy.
On Thu, Oct 31, 2019 at 11:18 PM Janne Johansson <icepic...@gmail.com> wrote: > > > Den tors 31 okt. 2019 kl 04:22 skrev soumya tr <soumya....@gmail.com>: > >> Thanks 潘东元 for the response. >> >> The creation of a new pool works, and all the PGs corresponding to that >> pool have active+clean state. >> >> When I initially set ceph 3 node cluster using juju charms (replication >> count per object was set to 3), there were issues with ceph-osd services. >> So I had to delete the units and readd them (I did all of them together, >> which must have created issues with rebalancing). I assume that the PGs in >> the inactive state points to the 3 old OSDs which were deleted. >> >> I assume I will have to create all the pools again. But my concern is >> about the default pools. >> >> ------------------------------- >> pool 1 'default.rgw.buckets.data' replicated size 3 min_size 2 crush_rule >> 0 object_hash rjenkins pg_num 16 pgp_num 16 last_change 15 flags hashpspool >> stripe_width 0 application rgw >> pool 2 'default.rgw.control' replicated size 3 min_size 2 crush_rule 0 >> object_hash rjenkins pg_num 2 pgp_num 2 last_change 19 flags hashpspool >> stripe_width 0 application rgw >> pool 3 'default.rgw.data.root' replicated size 3 min_size 2 crush_rule 0 >> object_hash rjenkins pg_num 2 pgp_num 2 last_change 23 flags hashpspool >> stripe_width 0 application rgw >> pool 4 'default.rgw.gc' replicated size 3 min_size 2 crush_rule 0 >> object_hash rjenkins pg_num 2 pgp_num 2 last_change 27 flags hashpspool >> stripe_width 0 application rgw >> pool 5 'default.rgw.log' replicated size 3 min_size 2 crush_rule 0 >> object_hash rjenkins pg_num 2 pgp_num 2 last_change 31 flags hashpspool >> stripe_width 0 application rgw >> pool 6 'default.rgw.intent-log' replicated size 3 min_size 2 crush_rule 0 >> object_hash rjenkins pg_num 2 pgp_num 2 last_change 35 flags hashpspool >> stripe_width 0 application rgw >> pool 7 'default.rgw.meta' replicated size 3 min_size 2 crush_rule 0 >> object_hash rjenkins pg_num 2 pgp_num 2 last_change 39 flags hashpspool >> stripe_width 0 application rgw >> pool 8 'default.rgw.usage' replicated size 3 min_size 2 crush_rule 0 >> object_hash rjenkins pg_num 2 pgp_num 2 last_change 43 flags hashpspool >> stripe_width 0 application rgw >> pool 9 'default.rgw.users.keys' replicated size 3 min_size 2 crush_rule 0 >> object_hash rjenkins pg_num 2 pgp_num 2 last_change 47 flags hashpspool >> stripe_width 0 application rgw >> pool 10 'default.rgw.users.email' replicated size 3 min_size 2 crush_rule >> 0 object_hash rjenkins pg_num 2 pgp_num 2 last_change 51 flags hashpspool >> stripe_width 0 application rgw >> pool 11 'default.rgw.users.swift' replicated size 3 min_size 2 crush_rule >> 0 object_hash rjenkins pg_num 2 pgp_num 2 last_change 55 flags hashpspool >> stripe_width 0 application rgw >> pool 12 'default.rgw.users.uid' replicated size 3 min_size 2 crush_rule 0 >> object_hash rjenkins pg_num 2 pgp_num 2 last_change 59 flags hashpspool >> stripe_width 0 application rgw >> pool 13 'default.rgw.buckets.extra' replicated size 3 min_size 2 >> crush_rule 0 object_hash rjenkins pg_num 2 pgp_num 2 last_change 63 flags >> hashpspool stripe_width 0 application rgw >> pool 14 'default.rgw.buckets.index' replicated size 3 min_size 2 >> crush_rule 0 object_hash rjenkins pg_num 4 pgp_num 4 last_change 67 flags >> hashpspool stripe_width 0 application rgw >> ------------------------------- >> >> Can you please update if recreating them using rados cli will break >> anything? >> >> > Those pools belong to radosgw, and if they are missing, they will be > created on demand the next time radosgw starts up. > the "defaul" is the name of the radosgw zone, which defaults to... > "default". They are not needed by any other part of ceph. > > -- > May the most significant bit of your life be positive. > -- Regards, Soumya Linux Sytem Administrator DirectI <http://poornam.com> "I like the dreams of the future better than the history of the past."
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com