Hi , I 'm in the process of upgrading my ceph cluster from Firefly to Hammer.
The ceph cluster has 12 OSD spread across 4 nodes. Mons have been upgraded to hammer, since I have created pools with value 512 and 256 , so am bit confused with the warning message. --snip-- ceph -s cluster a7160e16-0aaf-4e78-9e7c-7fbec08642f0 health HEALTH_WARN too many PGs per OSD (480 > max 300) monmap e1: 3 mons at {mon01= 172.16.10.5:6789/0,mon02=172.16.10.6:6789/0,mon03=172.16.10.7:6789/0} election epoch 116, quorum 0,1,2 mon01,mon02,mon03 osdmap e6814: 12 osds: 12 up, 12 in pgmap v2961763: 1920 pgs, 4 pools, 230 GB data, 29600 objects 692 GB used, 21652 GB / 22345 GB avail 1920 active+clean --snip-- ## Conf and ceph output --snip-- [global] fsid = a7160e16-0aaf-4e78-9e7c-7fbec08642f0 public_network = 172.16.10.0/24 cluster_network = 172.16.10.0/24 mon_initial_members = mon01, mon02, mon03 mon_host = 172.16.10.5,172.16.10.6,172.16.10.7 auth_cluster_required = cephx auth_service_required = cephx auth_client_required = cephx filestore_xattr_use_omap = true mon_clock_drift_allowed = .15 mon_clock_drift_warn_backoff = 30 mon_osd_down_out_interval = 300 mon_osd_report_timeout = 300 mon_osd_full_ratio = .85 mon_osd_nearfull_ratio = .75 osd_backfill_full_ratio = .75 osd_pool_default_size = 3 osd_pool_default_min_size = 2 osd_pool_default_pg_num = 512 osd_pool_default_pgp_num = 512 --snip-- ceph df POOLS: NAME ID USED %USED MAX AVAIL OBJECTS images 3 216G 0.97 7179G 27793 vms 4 14181M 0.06 7179G 1804 volumes 5 0 0 7179G 1 backups 6 0 0 7179G 0 ceph osd pool get poolname pg_num images: 256 backup: 512 vms: 512 volumes: 512 --snip-- Since it is a warning .. can I upgrade the OSDs without destroying the data. or Should I roll back. Thanks in advance
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com