Hi Greg, Thanks for the info and hope this will be solved in the upcoming minor updates of kraken. Regarding k+1 , I will take your feedback to our architect team and to increase this to k+2 and revert back the pool to normal state.
Thanks, Muthu On 1 February 2017 at 02:01, Shinobu Kinjo <ski...@redhat.com> wrote: > On Wed, Feb 1, 2017 at 3:38 AM, Gregory Farnum <gfar...@redhat.com> wrote: > > On Tue, Jan 31, 2017 at 9:06 AM, Muthusamy Muthiah > > <muthiah.muthus...@gmail.com> wrote: > >> Hi Greg, > >> > >> the problem is in kraken, when a pool is created with EC profile , > min_size > >> equals erasure size. > >> > >> For 3+1 profile , following is the pool status , > >> pool 2 'cdvr_ec' erasure size 4 min_size 4 crush_ruleset 1 object_hash > >> rjenkins pg_num 1024 pgp_num 1024 last_change 234 flags hashpspool > >> stripe_width 4128 > >> > >> For 4+1 profile: > >> pool 5 'cdvr_ec' erasure size 5 min_size 5 crush_ruleset 1 object_hash > >> rjenkins pg_num 4096 pgp_num 4096 > >> > >> For 3+2 profile : > >> pool 3 'cdvr_ec' erasure size 5 min_size 4 crush_ruleset 1 object_hash > >> rjenkins pg_num 1024 pgp_num 1024 last_change 412 flags hashpspool > >> stripe_width 4128 > >> > >> Where as on Jewel release for EC 4+1: > >> pool 30 'cdvr_ec' erasure size 5 min_size 4 crush_ruleset 1 object_hash > >> rjenkins pg_num 4096 pgp_num 4096 > >> > >> Trying to modify min_size and verify the status. > >> > >> Is there any reason behind this change in ceph kraken or a bug. > > > > The change was made on purpose because running with k replicas on a > > k+m pool is a bad idea. However, it definitely should have recovered > > the missing shard and then gone active, which doesn't appear to have > > Yeah, that might be true. > > > happened in this case. > > > > It looks like we just screwed up and don't let EC pools do recovery on > > min size. You can restore the old behavior by setting min_size equal > > to k and we'll be fixing this for the next release. (In general, k+1 > > If it's true, we should stop users to set up this ratio of data / > coding chunks. Or there should be any reasonable warning. > > Should it be feature request? > > > pools are not a good idea, which is why we didn't catch this in > > testing.) > > -Greg > > > >> > >> Thanks, > >> Muthu > >> > >> > >> > >> > >> On 31 January 2017 at 18:17, Muthusamy Muthiah < > muthiah.muthus...@gmail.com> > >> wrote: > >>> > >>> Hi Greg, > >>> > >>> Following are the test outcomes on EC profile ( n = k + m) > >>> > >>> > >>> > >>> 1. Kraken filestore and bluetore with m=1 , recovery does not > start > >>> . > >>> > >>> 2. Jewel filestore and bluestore with m=1 , recovery happens . > >>> > >>> 3. Kraken bluestore all default configuration and m=1, no > recovery. > >>> > >>> 4. Kraken bluestore with m=2 , recovery happens when one OSD is > down > >>> and for 2 OSD fails. > >>> > >>> > >>> > >>> So, the issue seems to be on ceph-kraken release. Your views… > >>> > >>> > >>> > >>> Thanks, > >>> > >>> Muthu > >>> > >>> > >>> > >>> > >>> On 31 January 2017 at 14:18, Muthusamy Muthiah > >>> <muthiah.muthus...@gmail.com> wrote: > >>>> > >>>> Hi Greg, > >>>> > >>>> Now we could see the same problem exists for kraken-filestore also. > >>>> Attached the requested osdmap and crushmap. > >>>> > >>>> OSD.1 was stopped in this following procedure and OSD map for a PG is > >>>> displayed. > >>>> > >>>> ceph osd dump | grep cdvr_ec > >>>> 2017-01-31 08:39:44.827079 7f323d66c700 -1 WARNING: the following > >>>> dangerous and experimental features are enabled: bluestore,rocksdb > >>>> 2017-01-31 08:39:44.848901 7f323d66c700 -1 WARNING: the following > >>>> dangerous and experimental features are enabled: bluestore,rocksdb > >>>> pool 2 'cdvr_ec' erasure size 4 min_size 4 crush_ruleset 1 object_hash > >>>> rjenkins pg_num 1024 pgp_num 1024 last_change 234 flags hashpspool > >>>> stripe_width 4128 > >>>> > >>>> [root@ca-cn2 ~]# ceph osd getmap -o /tmp/osdmap > >>>> > >>>> > >>>> [root@ca-cn2 ~]# osdmaptool --pool 2 --test-map-object object1 > >>>> /tmp/osdmap > >>>> osdmaptool: osdmap file '/tmp/osdmap' > >>>> object 'object1' -> 2.2bc -> [20,47,1,36] > >>>> > >>>> [root@ca-cn2 ~]# ceph osd map cdvr_ec object1 > >>>> osdmap e402 pool 'cdvr_ec' (2) object 'object1' -> pg 2.bac5debc > (2.2bc) > >>>> -> up ([20,47,1,36], p20) acting ([20,47,1,36], p20) > >>>> > >>>> [root@ca-cn2 ~]# systemctl stop ceph-osd@1.service > >>>> > >>>> [root@ca-cn2 ~]# ceph osd getmap -o /tmp/osdmap1 > >>>> > >>>> > >>>> [root@ca-cn2 ~]# osdmaptool --pool 2 --test-map-object object1 > >>>> /tmp/osdmap1 > >>>> osdmaptool: osdmap file '/tmp/osdmap1' > >>>> object 'object1' -> 2.2bc -> [20,47,2147483647,36] > >>>> > >>>> > >>>> [root@ca-cn2 ~]# ceph osd map cdvr_ec object1 > >>>> osdmap e406 pool 'cdvr_ec' (2) object 'object1' -> pg 2.bac5debc > (2.2bc) > >>>> -> up ([20,47,39,36], p20) acting ([20,47,NONE,36], p20) > >>>> > >>>> > >>>> [root@ca-cn2 ~]# ceph osd tree > >>>> 2017-01-31 08:42:19.606876 7f4ed856a700 -1 WARNING: the following > >>>> dangerous and experimental features are enabled: bluestore,rocksdb > >>>> 2017-01-31 08:42:19.628358 7f4ed856a700 -1 WARNING: the following > >>>> dangerous and experimental features are enabled: bluestore,rocksdb > >>>> ID WEIGHT TYPE NAME UP/DOWN REWEIGHT PRIMARY-AFFINITY > >>>> -1 327.47314 root default > >>>> -2 65.49463 host ca-cn4 > >>>> 3 5.45789 osd.3 up 1.00000 1.00000 > >>>> 5 5.45789 osd.5 up 1.00000 1.00000 > >>>> 10 5.45789 osd.10 up 1.00000 1.00000 > >>>> 16 5.45789 osd.16 up 1.00000 1.00000 > >>>> 21 5.45789 osd.21 up 1.00000 1.00000 > >>>> 27 5.45789 osd.27 up 1.00000 1.00000 > >>>> 30 5.45789 osd.30 up 1.00000 1.00000 > >>>> 35 5.45789 osd.35 up 1.00000 1.00000 > >>>> 42 5.45789 osd.42 up 1.00000 1.00000 > >>>> 47 5.45789 osd.47 up 1.00000 1.00000 > >>>> 51 5.45789 osd.51 up 1.00000 1.00000 > >>>> 53 5.45789 osd.53 up 1.00000 1.00000 > >>>> -3 65.49463 host ca-cn3 > >>>> 2 5.45789 osd.2 up 1.00000 1.00000 > >>>> 6 5.45789 osd.6 up 1.00000 1.00000 > >>>> 11 5.45789 osd.11 up 1.00000 1.00000 > >>>> 15 5.45789 osd.15 up 1.00000 1.00000 > >>>> 20 5.45789 osd.20 up 1.00000 1.00000 > >>>> 25 5.45789 osd.25 up 1.00000 1.00000 > >>>> 29 5.45789 osd.29 up 1.00000 1.00000 > >>>> 33 5.45789 osd.33 up 1.00000 1.00000 > >>>> 38 5.45789 osd.38 up 1.00000 1.00000 > >>>> 40 5.45789 osd.40 up 1.00000 1.00000 > >>>> 45 5.45789 osd.45 up 1.00000 1.00000 > >>>> 49 5.45789 osd.49 up 1.00000 1.00000 > >>>> -4 65.49463 host ca-cn5 > >>>> 0 5.45789 osd.0 up 1.00000 1.00000 > >>>> 7 5.45789 osd.7 up 1.00000 1.00000 > >>>> 12 5.45789 osd.12 up 1.00000 1.00000 > >>>> 17 5.45789 osd.17 up 1.00000 1.00000 > >>>> 23 5.45789 osd.23 up 1.00000 1.00000 > >>>> 26 5.45789 osd.26 up 1.00000 1.00000 > >>>> 32 5.45789 osd.32 up 1.00000 1.00000 > >>>> 34 5.45789 osd.34 up 1.00000 1.00000 > >>>> 41 5.45789 osd.41 up 1.00000 1.00000 > >>>> 46 5.45789 osd.46 up 1.00000 1.00000 > >>>> 52 5.45789 osd.52 up 1.00000 1.00000 > >>>> 56 5.45789 osd.56 up 1.00000 1.00000 > >>>> -5 65.49463 host ca-cn1 > >>>> 4 5.45789 osd.4 up 1.00000 1.00000 > >>>> 9 5.45789 osd.9 up 1.00000 1.00000 > >>>> 14 5.45789 osd.14 up 1.00000 1.00000 > >>>> 19 5.45789 osd.19 up 1.00000 1.00000 > >>>> 24 5.45789 osd.24 up 1.00000 1.00000 > >>>> 36 5.45789 osd.36 up 1.00000 1.00000 > >>>> 43 5.45789 osd.43 up 1.00000 1.00000 > >>>> 50 5.45789 osd.50 up 1.00000 1.00000 > >>>> 55 5.45789 osd.55 up 1.00000 1.00000 > >>>> 57 5.45789 osd.57 up 1.00000 1.00000 > >>>> 58 5.45789 osd.58 up 1.00000 1.00000 > >>>> 59 5.45789 osd.59 up 1.00000 1.00000 > >>>> -6 65.49463 host ca-cn2 > >>>> 1 5.45789 osd.1 down 0 1.00000 > >>>> 8 5.45789 osd.8 up 1.00000 1.00000 > >>>> 13 5.45789 osd.13 up 1.00000 1.00000 > >>>> 18 5.45789 osd.18 up 1.00000 1.00000 > >>>> 22 5.45789 osd.22 up 1.00000 1.00000 > >>>> 28 5.45789 osd.28 up 1.00000 1.00000 > >>>> 31 5.45789 osd.31 up 1.00000 1.00000 > >>>> 37 5.45789 osd.37 up 1.00000 1.00000 > >>>> 39 5.45789 osd.39 up 1.00000 1.00000 > >>>> 44 5.45789 osd.44 up 1.00000 1.00000 > >>>> 48 5.45789 osd.48 up 1.00000 1.00000 > >>>> 54 5.45789 osd.54 up 1.00000 1.00000 > >>>> > >>>> health HEALTH_ERR > >>>> 69 pgs are stuck inactive for more than 300 seconds > >>>> 69 pgs incomplete > >>>> 69 pgs stuck inactive > >>>> 69 pgs stuck unclean > >>>> 512 requests are blocked > 32 sec > >>>> monmap e2: 5 mons at > >>>> {ca-cn1=10.50.5.117:6789/0,ca-cn2=10.50.5.118:6789/0,ca-cn3= > 10.50.5.119:6789/0,ca-cn4=10.50.5.120:6789/0,ca-cn5=10.50.5.121:6789/0} > >>>> election epoch 8, quorum 0,1,2,3,4 > >>>> ca-cn1,ca-cn2,ca-cn3,ca-cn4,ca-cn5 > >>>> mgr active: ca-cn4 standbys: ca-cn2, ca-cn5, ca-cn3, ca-cn1 > >>>> osdmap e406: 60 osds: 59 up, 59 in; 69 remapped pgs > >>>> flags sortbitwise,require_jewel_osds,require_kraken_osds > >>>> pgmap v23018: 1024 pgs, 1 pools, 3892 GB data, 7910 kobjects > >>>> 6074 GB used, 316 TB / 322 TB avail > >>>> 955 active+clean > >>>> 69 remapped+incomplete > >>>> > >>>> Thanks, > >>>> Muthu > >>>> > >>>> > >>>> On 31 January 2017 at 02:54, Gregory Farnum <gfar...@redhat.com> > wrote: > >>>>> > >>>>> You might also check out "ceph osd tree" and crush dump and make sure > >>>>> they look the way you expect. > >>>>> > >>>>> On Mon, Jan 30, 2017 at 1:23 PM, Gregory Farnum <gfar...@redhat.com> > >>>>> wrote: > >>>>> > On Sun, Jan 29, 2017 at 6:40 AM, Muthusamy Muthiah > >>>>> > <muthiah.muthus...@gmail.com> wrote: > >>>>> >> Hi All, > >>>>> >> > >>>>> >> Also tried EC profile 3+1 on 5 node cluster with bluestore > enabled . > >>>>> >> When > >>>>> >> an OSD is down the cluster goes to ERROR state even when the > cluster > >>>>> >> is n+1 > >>>>> >> . No recovery happening. > >>>>> >> > >>>>> >> health HEALTH_ERR > >>>>> >> 75 pgs are stuck inactive for more than 300 seconds > >>>>> >> 75 pgs incomplete > >>>>> >> 75 pgs stuck inactive > >>>>> >> 75 pgs stuck unclean > >>>>> >> monmap e2: 5 mons at > >>>>> >> > >>>>> >> {ca-cn1=10.50.5.117:6789/0,ca-cn2=10.50.5.118:6789/0,ca-cn3= > 10.50.5.119:6789/0,ca-cn4=10.50.5.120:6789/0,ca-cn5=10.50.5.121:6789/0} > >>>>> >> election epoch 10, quorum 0,1,2,3,4 > >>>>> >> ca-cn1,ca-cn2,ca-cn3,ca-cn4,ca-cn5 > >>>>> >> mgr active: ca-cn1 standbys: ca-cn4, ca-cn3, ca-cn5, > ca-cn2 > >>>>> >> osdmap e264: 60 osds: 59 up, 59 in; 75 remapped pgs > >>>>> >> flags sortbitwise,require_jewel_ > osds,require_kraken_osds > >>>>> >> pgmap v119402: 1024 pgs, 1 pools, 28519 GB data, 21548 > kobjects > >>>>> >> 39976 GB used, 282 TB / 322 TB avail > >>>>> >> 941 active+clean > >>>>> >> 75 remapped+incomplete > >>>>> >> 8 active+clean+scrubbing > >>>>> >> > >>>>> >> this seems to be an issue with bluestore , recovery not happening > >>>>> >> properly > >>>>> >> with EC . > >>>>> > > >>>>> > It's possible but it seems a lot more likely this is some kind of > >>>>> > config issue. Can you share your osd map ("ceph osd getmap")? > >>>>> > -Greg > >>>>> > > >>>>> >> > >>>>> >> Thanks, > >>>>> >> Muthu > >>>>> >> > >>>>> >> On 24 January 2017 at 12:57, Muthusamy Muthiah > >>>>> >> <muthiah.muthus...@gmail.com> > >>>>> >> wrote: > >>>>> >>> > >>>>> >>> Hi Greg, > >>>>> >>> > >>>>> >>> We use EC:4+1 on 5 node cluster in production deployments with > >>>>> >>> filestore > >>>>> >>> and it does recovery and peering when one OSD goes down. After > few > >>>>> >>> mins , > >>>>> >>> other OSD from a node where the fault OSD exists will take over > the > >>>>> >>> PGs > >>>>> >>> temporarily and all PGs goes to active + clean state . Cluster > also > >>>>> >>> does not > >>>>> >>> goes down during this recovery process. > >>>>> >>> > >>>>> >>> Only on bluestore we see cluster going to error state when one > OSD > >>>>> >>> is > >>>>> >>> down. > >>>>> >>> We are still validating this and let you know additional > findings. > >>>>> >>> > >>>>> >>> Thanks, > >>>>> >>> Muthu > >>>>> >>> > >>>>> >>> On 21 January 2017 at 02:06, Shinobu Kinjo <ski...@redhat.com> > >>>>> >>> wrote: > >>>>> >>>> > >>>>> >>>> `ceph pg dump` should show you something like: > >>>>> >>>> > >>>>> >>>> * active+undersized+degraded ... [NONE,3,2,4,1] 3 > >>>>> >>>> [NONE,3,2,4,1] > >>>>> >>>> > >>>>> >>>> Sam, > >>>>> >>>> > >>>>> >>>> Am I wrong? Or is it up to something else? > >>>>> >>>> > >>>>> >>>> > >>>>> >>>> On Sat, Jan 21, 2017 at 4:22 AM, Gregory Farnum > >>>>> >>>> <gfar...@redhat.com> > >>>>> >>>> wrote: > >>>>> >>>> > I'm pretty sure the default configs won't let an EC PG go > active > >>>>> >>>> > with > >>>>> >>>> > only "k" OSDs in its PG; it needs at least k+1 (or possibly > more? > >>>>> >>>> > Not > >>>>> >>>> > certain). Running an "n+1" EC config is just not a good idea. > >>>>> >>>> > For testing you could probably adjust this with the > equivalent of > >>>>> >>>> > min_size for EC pools, but I don't know the parameters off the > >>>>> >>>> > top of > >>>>> >>>> > my head. > >>>>> >>>> > -Greg > >>>>> >>>> > > >>>>> >>>> > On Fri, Jan 20, 2017 at 2:15 AM, Muthusamy Muthiah > >>>>> >>>> > <muthiah.muthus...@gmail.com> wrote: > >>>>> >>>> >> Hi , > >>>>> >>>> >> > >>>>> >>>> >> We are validating kraken 11.2.0 with bluestore on 5 node > >>>>> >>>> >> cluster with > >>>>> >>>> >> EC > >>>>> >>>> >> 4+1. > >>>>> >>>> >> > >>>>> >>>> >> When an OSD is down , the peering is not happening and ceph > >>>>> >>>> >> health > >>>>> >>>> >> status > >>>>> >>>> >> moved to ERR state after few mins. This was working in > previous > >>>>> >>>> >> development > >>>>> >>>> >> releases. Any additional configuration required in v11.2.0 > >>>>> >>>> >> > >>>>> >>>> >> Following is our ceph configuration: > >>>>> >>>> >> > >>>>> >>>> >> mon_osd_down_out_interval = 30 > >>>>> >>>> >> mon_osd_report_timeout = 30 > >>>>> >>>> >> mon_osd_down_out_subtree_limit = host > >>>>> >>>> >> mon_osd_reporter_subtree_level = host > >>>>> >>>> >> > >>>>> >>>> >> and the recovery parameters set to default. > >>>>> >>>> >> > >>>>> >>>> >> [root@ca-cn1 ceph]# ceph osd crush show-tunables > >>>>> >>>> >> > >>>>> >>>> >> { > >>>>> >>>> >> "choose_local_tries": 0, > >>>>> >>>> >> "choose_local_fallback_tries": 0, > >>>>> >>>> >> "choose_total_tries": 50, > >>>>> >>>> >> "chooseleaf_descend_once": 1, > >>>>> >>>> >> "chooseleaf_vary_r": 1, > >>>>> >>>> >> "chooseleaf_stable": 1, > >>>>> >>>> >> "straw_calc_version": 1, > >>>>> >>>> >> "allowed_bucket_algs": 54, > >>>>> >>>> >> "profile": "jewel", > >>>>> >>>> >> "optimal_tunables": 1, > >>>>> >>>> >> "legacy_tunables": 0, > >>>>> >>>> >> "minimum_required_version": "jewel", > >>>>> >>>> >> "require_feature_tunables": 1, > >>>>> >>>> >> "require_feature_tunables2": 1, > >>>>> >>>> >> "has_v2_rules": 1, > >>>>> >>>> >> "require_feature_tunables3": 1, > >>>>> >>>> >> "has_v3_rules": 0, > >>>>> >>>> >> "has_v4_buckets": 0, > >>>>> >>>> >> "require_feature_tunables5": 1, > >>>>> >>>> >> "has_v5_rules": 0 > >>>>> >>>> >> } > >>>>> >>>> >> > >>>>> >>>> >> ceph status: > >>>>> >>>> >> > >>>>> >>>> >> health HEALTH_ERR > >>>>> >>>> >> 173 pgs are stuck inactive for more than 300 > seconds > >>>>> >>>> >> 173 pgs incomplete > >>>>> >>>> >> 173 pgs stuck inactive > >>>>> >>>> >> 173 pgs stuck unclean > >>>>> >>>> >> monmap e2: 5 mons at > >>>>> >>>> >> > >>>>> >>>> >> > >>>>> >>>> >> {ca-cn1=10.50.5.117:6789/0,ca-cn2=10.50.5.118:6789/0,ca-cn3= > 10.50.5.119:6789/0,ca-cn4=10.50.5.120:6789/0,ca-cn5=10.50.5.121:6789/0} > >>>>> >>>> >> election epoch 106, quorum 0,1,2,3,4 > >>>>> >>>> >> ca-cn1,ca-cn2,ca-cn3,ca-cn4,ca-cn5 > >>>>> >>>> >> mgr active: ca-cn1 standbys: ca-cn2, ca-cn4, ca-cn5, > >>>>> >>>> >> ca-cn3 > >>>>> >>>> >> osdmap e1128: 60 osds: 59 up, 59 in; 173 remapped pgs > >>>>> >>>> >> flags > >>>>> >>>> >> sortbitwise,require_jewel_osds,require_kraken_osds > >>>>> >>>> >> pgmap v782747: 2048 pgs, 1 pools, 63133 GB data, 46293 > >>>>> >>>> >> kobjects > >>>>> >>>> >> 85199 GB used, 238 TB / 322 TB avail > >>>>> >>>> >> 1868 active+clean > >>>>> >>>> >> 173 remapped+incomplete > >>>>> >>>> >> 7 active+clean+scrubbing > >>>>> >>>> >> > >>>>> >>>> >> MON log: > >>>>> >>>> >> > >>>>> >>>> >> 2017-01-20 09:25:54.715684 7f55bcafb700 0 > log_channel(cluster) > >>>>> >>>> >> log > >>>>> >>>> >> [INF] : > >>>>> >>>> >> osd.54 out (down for 31.703786) > >>>>> >>>> >> 2017-01-20 09:25:54.725688 7f55bf4d5700 0 > >>>>> >>>> >> mon.ca-cn1@0(leader).osd > >>>>> >>>> >> e1120 > >>>>> >>>> >> crush map has features 288250512065953792, adjusting msgr > >>>>> >>>> >> requires > >>>>> >>>> >> 2017-01-20 09:25:54.729019 7f55bf4d5700 0 > log_channel(cluster) > >>>>> >>>> >> log > >>>>> >>>> >> [INF] : > >>>>> >>>> >> osdmap e1120: 60 osds: 59 up, 59 in > >>>>> >>>> >> 2017-01-20 09:25:54.735987 7f55bf4d5700 0 > log_channel(cluster) > >>>>> >>>> >> log > >>>>> >>>> >> [INF] : > >>>>> >>>> >> pgmap v781993: 2048 pgs: 1869 active+clean, 173 incomplete, 6 > >>>>> >>>> >> active+clean+scrubbing; 63159 GB data, 85201 GB used, 238 TB > / > >>>>> >>>> >> 322 TB > >>>>> >>>> >> avail; > >>>>> >>>> >> 21825 B/s rd, 163 MB/s wr, 2046 op/s > >>>>> >>>> >> 2017-01-20 09:25:55.737749 7f55bf4d5700 0 > >>>>> >>>> >> mon.ca-cn1@0(leader).osd > >>>>> >>>> >> e1121 > >>>>> >>>> >> crush map has features 288250512065953792, adjusting msgr > >>>>> >>>> >> requires > >>>>> >>>> >> 2017-01-20 09:25:55.744338 7f55bf4d5700 0 > log_channel(cluster) > >>>>> >>>> >> log > >>>>> >>>> >> [INF] : > >>>>> >>>> >> osdmap e1121: 60 osds: 59 up, 59 in > >>>>> >>>> >> 2017-01-20 09:25:55.749616 7f55bf4d5700 0 > log_channel(cluster) > >>>>> >>>> >> log > >>>>> >>>> >> [INF] : > >>>>> >>>> >> pgmap v781994: 2048 pgs: 29 remapped+incomplete, 1869 > >>>>> >>>> >> active+clean, > >>>>> >>>> >> 144 > >>>>> >>>> >> incomplete, 6 active+clean+scrubbing; 63159 GB data, 85201 GB > >>>>> >>>> >> used, > >>>>> >>>> >> 238 TB / > >>>>> >>>> >> 322 TB avail; 44503 B/s rd, 45681 kB/s wr, 518 op/s > >>>>> >>>> >> 2017-01-20 09:25:56.768721 7f55bf4d5700 0 > log_channel(cluster) > >>>>> >>>> >> log > >>>>> >>>> >> [INF] : > >>>>> >>>> >> pgmap v781995: 2048 pgs: 47 remapped+incomplete, 1869 > >>>>> >>>> >> active+clean, > >>>>> >>>> >> 126 > >>>>> >>>> >> incomplete, 6 active+clean+scrubbing; 63159 GB data, 85201 GB > >>>>> >>>> >> used, > >>>>> >>>> >> 238 TB / > >>>>> >>>> >> 322 TB avail; 20275 B/s rd, 72742 kB/s wr, 665 op/s > >>>>> >>>> >> > >>>>> >>>> >> Thanks, > >>>>> >>>> >> Muthu > >>>>> >>>> >> > >>>>> >>>> >> > >>>>> >>>> >> _______________________________________________ > >>>>> >>>> >> ceph-users mailing list > >>>>> >>>> >> ceph-users@lists.ceph.com > >>>>> >>>> >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >>>>> >>>> >> > >>>>> >>>> > _______________________________________________ > >>>>> >>>> > ceph-users mailing list > >>>>> >>>> > ceph-users@lists.ceph.com > >>>>> >>>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >>>>> >>> > >>>>> >>> > >>>>> >> > >>>>> >> > >>>>> >> _______________________________________________ > >>>>> >> ceph-users mailing list > >>>>> >> ceph-users@lists.ceph.com > >>>>> >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >>>>> >> > >>>> > >>>> > >>> > >> >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com