output from "ceph osd pool ls detail" pool 1 'cephfs_data' replicated size 3 min_size 2 crush_rule 0 object_hash rjenkins pg_num 32 pgp_num 32 last_change 74 lfor 0/64 flags hashpspool stripe_width 0 application cephfs
pool 2 'cephfs_metadata' replicated size 3 min_size 2 crush_rule 0 object_hash rjenkins pg_num 32 pgp_num 32 last_change 75 lfor 0/67 flags hashpspool stripe_width 0 application cephfs On Thu, Aug 29, 2019 at 6:13 PM Heðin Ejdesgaard Møller <h...@synack.fo> wrote: > What's the output of > ceph osd pool ls detail > > > On hós, 2019-08-29 at 18:06 +0530, Amudhan P wrote: > > output from "ceph -s " > > > > cluster: > > id: 7c138e13-7b98-4309-b591-d4091a1742b4 > > health: HEALTH_WARN > > Degraded data redundancy: 1141587/7723191 objects > > degraded (14.781%), 15 pgs degraded, 16 pgs undersized > > > > services: > > mon: 1 daemons, quorum mon01 > > mgr: mon01(active) > > mds: cephfs-tst-1/1/1 up {0=mon01=up:active} > > osd: 3 osds: 3 up, 3 in; 16 remapped pgs > > > > data: > > pools: 2 pools, 64 pgs > > objects: 2.57 M objects, 59 GiB > > usage: 190 GiB used, 5.3 TiB / 5.5 TiB avail > > pgs: 1141587/7723191 objects degraded (14.781%) > > 48 active+clean > > 15 active+undersized+degraded+remapped+backfill_wait > > 1 active+undersized+remapped+backfilling > > > > io: > > recovery: 0 B/s, 10 objects/s > > > > output from "ceph osd tree" > > ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF > > -1 5.45819 root default > > -3 1.81940 host test-node1 > > 0 hdd 1.81940 osd.0 up 1.00000 1.00000 > > -5 1.81940 host test-node2 > > 1 hdd 1.81940 osd.1 up 1.00000 1.00000 > > -7 1.81940 host test-node3 > > 2 hdd 1.81940 osd.2 up 1.00000 1.00000 > > > > failure domain not configured yet, setup is 3 OSD node each with a > > single disk, 1 node with mon&mds&mgr running. > > the cluster was healthy until I run a script for creating multiple > > folders. > > > > regards > > Amudhan > > > > On Thu, Aug 29, 2019 at 5:33 PM Heðin Ejdesgaard Møller < > > h...@synack.fo> wrote: > > > In adition to ceph -s, could you provide the output of > > > ceph osd tree > > > and specify what your failure domain is ? > > > > > > /Heðin > > > > > > > > > On hós, 2019-08-29 at 13:55 +0200, Janne Johansson wrote: > > > > > > > > > > > > Den tors 29 aug. 2019 kl 13:50 skrev Amudhan P < > > > amudha...@gmail.com>: > > > > > Hi, > > > > > > > > > > I am using ceph version 13.2.6 (mimic) on test setup trying > > > with > > > > > cephfs. > > > > > my ceph health status showing warning . > > > > > > > > > > "ceph health" > > > > > HEALTH_WARN Degraded data redundancy: 1197023/7723191 objects > > > > > degraded (15.499%) > > > > > > > > > > "ceph health detail" > > > > > HEALTH_WARN Degraded data redundancy: 1197128/7723191 objects > > > > > degraded (15.500%) > > > > > PG_DEGRADED Degraded data redundancy: 1197128/7723191 objects > > > > > degraded (15.500%) > > > > > pg 2.0 is stuck undersized for 1076.454929, current state > > > > > active+undersized+ > > > > > pg 2.2 is stuck undersized for 1076.456639, current state > > > > > active+undersized+ > > > > > > > > > > > > > How does "ceph -s" look? > > > > It should have more info on what else is wrong. > > > > > > > > -- > > > > May the most significant bit of your life be positive. > > > > _______________________________________________ > > > > ceph-users mailing list -- ceph-users@ceph.io > > > > To unsubscribe send an email to ceph-users-le...@ceph.io > > > > >
_______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io