I do not think so ..or maybe I did not understand what are you saying There is no key listed on mgr config
ceph config-key list [ "config-history/1/", "config-history/2/", "config-history/2/+mgr/mgr/dashboard/server_addr", "config-history/3/", "config-history/3/+mgr/mgr/prometheus/server_addr", "config-history/4/", "config-history/4/+mgr/mgr/dashboard/username", "config-history/5/", "config-history/5/+mgr/mgr/dashboard/password", "config-history/6/", "config-history/6/+mgr/mgr/balancer/mode", "config-history/7/", "config-history/7/+mgr/mgr/balancer/active", "config-history/8/", "config-history/8/+mgr/mgr/dashboard/password", "config-history/8/-mgr/mgr/dashboard/password", "config/mgr/mgr/balancer/active", "config/mgr/mgr/balancer/mode", "config/mgr/mgr/dashboard/password", "config/mgr/mgr/dashboard/server_addr", "config/mgr/mgr/dashboard/username", "config/mgr/mgr/prometheus/server_addr", "mgr/dashboard/crt", "mgr/dashboard/key" On Wed, 31 Oct 2018 at 09:59, <c...@jack.fr.eu.org> wrote: > Isn't this a mgr variable ? > > On 10/31/2018 02:49 PM, Steven Vacaroaia wrote: > > Hi, > > > > Any idea why different value for mon_max_pg_per_osd is not "recognized" > ? > > I am using mimic 13.2.2 > > > > Here is what I have in /etc/ceph/ceph.conf > > > > > > [mon] > > mon_allow_pool_delete = true > > mon_osd_min_down_reporters = 1 > > mon_max_pg_per_osd = 400 > > > > checking the value with > > ceph daemon osd.6 config show| grep mon_max_pg_per_osd still shows the > > default ( 250) > > > > > > Injecting a different value appears to works > > ceph tell osd.* injectargs '--mon_max_pg_per_osd 500' > > > > ceph daemon osd.6 config show| grep mon_max_pg_per_osd > > "mon_max_pg_per_osd": "500", > > > > BUT > > > > cluster is still complaining TOO_MANY_PGS too many PGs per OSD (262 > > > max 250) > > > > I have restarted ceph.target services on monitor/manager server > > What else has to be done to have the cluster using the new value ? > > > > Steven > > > > > > > > _______________________________________________ > > ceph-users mailing list > > ceph-users@lists.ceph.com > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com