Just loosing one disk doesn’t automagically delete it from CRUSH, but in the 
output you had 10 disks listed, so there must be something else going - did you 
delete the disk from the crush map as well?

Ceph waits by default 300 secs AFAIK to mark an OSD out after it will start to 
recover.


> On 29 Jun 2016, at 10:42, Mario Giammarco <mgiamma...@gmail.com> wrote:
> 
> I thank you for your reply so I can add my experience:
> 
> 1) the other time this thing happened to me I had a cluster with min_size=2 
> and size=3 and the problem was the same. That time I put min_size=1 to 
> recover the pool but it did not help. So I do not understand where is the 
> advantage to put three copies when ceph can decide to discard all three.
> 2) I started with 11 hdds. The hard disk failed. Ceph waited forever for hard 
> disk coming back. But hard disk is really completelly broken so I have 
> followed the procedure to really delete from cluster. Anyway ceph did not 
> recover.
> 3) I have 307 pgs more than 300 but it is due to the fact that I had 11 hdds 
> now only 10. I will add more hdds after I repair the pool
> 4) I have reduced the monitors to 3
> 
> 
> 
> Il giorno mer 29 giu 2016 alle ore 10:25 Christian Balzer <ch...@gol.com 
> <mailto:ch...@gol.com>> ha scritto:
> 
> Hello,
> 
> On Wed, 29 Jun 2016 06:02:59 +0000 Mario Giammarco wrote:
> 
> > pool 0 'rbd' replicated size 2 min_size 1 crush_ruleset 0 object_hash
>                                ^
> And that's the root cause of all your woes.
> The default replication size is 3 for a reason and while I do run pools
> with replication of 2 they are either HDD RAIDs or extremely trustworthy
> and well monitored SSD.
> 
> That said, something more than a single HDD failure must have happened
> here, you should check the logs and backtrace all the step you did after
> that OSD failed.
> 
> You said there were 11 HDDs and your first ceph -s output showed:
> ---
>      osdmap e10182: 10 osds: 10 up, 10 in
> ----
> And your crush map states the same.
> 
> So how and WHEN did you remove that OSD?
> My suspicion would be it was removed before recovery was complete.
> 
> Also, as I think was mentioned before, 7 mons are overkill 3-5 would be a
> saner number.
> 
> Christian
> 
> > rjenkins pg_num 512 pgp_num 512 last_change 9313 flags hashpspool
> > stripe_width 0
> >        removed_snaps [1~3]
> > pool 1 'rbd2' replicated size 2 min_size 1 crush_ruleset 0 object_hash
> > rjenkins pg_num 512 pgp_num 512 last_change 9314 flags hashpspool
> > stripe_width 0
> >        removed_snaps [1~3]
> > pool 2 'rbd3' replicated size 2 min_size 1 crush_ruleset 0 object_hash
> > rjenkins pg_num 512 pgp_num 512 last_change 10537 flags hashpspool
> > stripe_width 0
> >        removed_snaps [1~3]
> >
> >
> > ID WEIGHT  REWEIGHT SIZE   USE   AVAIL %USE  VAR
> > 5 1.81000  1.00000  1857G  984G  872G 53.00 0.86
> > 6 1.81000  1.00000  1857G 1202G  655G 64.73 1.05
> > 2 1.81000  1.00000  1857G 1158G  698G 62.38 1.01
> > 3 1.35999  1.00000  1391G  906G  485G 65.12 1.06
> > 4 0.89999  1.00000   926G  702G  223G 75.88 1.23
> > 7 1.81000  1.00000  1857G 1063G  793G 57.27 0.93
> > 8 1.81000  1.00000  1857G 1011G  846G 54.44 0.88
> > 9 0.89999  1.00000   926G  573G  352G 61.91 1.01
> > 0 1.81000  1.00000  1857G 1227G  629G 66.10 1.07
> > 13 0.45000  1.00000   460G  307G  153G 66.74 1.08
> >              TOTAL 14846G 9136G 5710G 61.54
> > MIN/MAX VAR: 0.86/1.23  STDDEV: 6.47
> >
> >
> >
> > ceph version 0.94.7 (d56bdf93ced6b80b07397d57e3fa68fe68304432)
> >
> > http://pastebin.com/SvGfcSHb <http://pastebin.com/SvGfcSHb>
> > http://pastebin.com/gYFatsNS <http://pastebin.com/gYFatsNS>
> > http://pastebin.com/VZD7j2vN <http://pastebin.com/VZD7j2vN>
> >
> > I do not understand why I/O on ENTIRE cluster is blocked when only few
> > pgs are incomplete.
> >
> > Many thanks,
> > Mario
> >
> >
> > Il giorno mar 28 giu 2016 alle ore 19:34 Stefan Priebe - Profihost AG <
> > s.pri...@profihost.ag <mailto:s.pri...@profihost.ag>> ha scritto:
> >
> > > And ceph health detail
> > >
> > > Stefan
> > >
> > > Excuse my typo sent from my mobile phone.
> > >
> > > Am 28.06.2016 um 19:28 schrieb Oliver Dzombic <i...@ip-interactive.de 
> > > <mailto:i...@ip-interactive.de>>:
> > >
> > > Hi Mario,
> > >
> > > please give some more details:
> > >
> > > Please the output of:
> > >
> > > ceph osd pool ls detail
> > > ceph osd df
> > > ceph --version
> > >
> > > ceph -w for 10 seconds ( use http://pastebin.com/ <http://pastebin.com/> 
> > > please )
> > >
> > > ceph osd crush dump ( also pastebin pls )
> > >
> > > --
> > > Mit freundlichen Gruessen / Best regards
> > >
> > > Oliver Dzombic
> > > IP-Interactive
> > >
> > > mailto:i...@ip-interactive.de <mailto:i...@ip-interactive.de> 
> > > <i...@ip-interactive.de <mailto:i...@ip-interactive.de>>
> > >
> > > Anschrift:
> > >
> > > IP Interactive UG ( haftungsbeschraenkt )
> > > Zum Sonnenberg 1-3
> > > 63571 Gelnhausen
> > >
> > > HRB 93402 beim Amtsgericht Hanau
> > > Geschäftsführung: Oliver Dzombic
> > >
> > > Steuer Nr.: 35 236 3622 1
> > > UST ID: DE274086107
> > >
> > >
> > > Am 28.06.2016 um 18:59 schrieb Mario Giammarco:
> > >
> > > Hello,
> > >
> > > this is the second time that happens to me, I hope that someone can
> > >
> > > explain what I can do.
> > >
> > > Proxmox ceph cluster with 8 servers, 11 hdd. Min_size=1, size=2.
> > >
> > >
> > > One hdd goes down due to bad sectors.
> > >
> > > Ceph recovers but it ends with:
> > >
> > >
> > > cluster f2a8dd7d-949a-4a29-acab-11d4900249f4
> > >
> > >     health HEALTH_WARN
> > >
> > >            3 pgs down
> > >
> > >            19 pgs incomplete
> > >
> > >            19 pgs stuck inactive
> > >
> > >            19 pgs stuck unclean
> > >
> > >            7 requests are blocked > 32 sec
> > >
> > >     monmap e11: 7 mons at
> > >
> > > {0=192.168.0.204:6789/0,1=192.168.0.201:6789/0 
> > > <http://192.168.0.204:6789/0,1=192.168.0.201:6789/0>,
> > >
> > > 2=192.168.0.203:6789/0,3=192.168.0.205:6789/0,4=192.168.0.202 
> > > <http://192.168.0.203:6789/0,3=192.168.0.205:6789/0,4=192.168.0.202>:
> > >
> > > 6789/0,5=192.168.0.206:6789/0,6=192.168.0.207:6789/0 
> > > <http://192.168.0.206:6789/0,6=192.168.0.207:6789/0>}
> > >
> > >            election epoch 722, quorum
> > >
> > > 0,1,2,3,4,5,6 1,4,2,0,3,5,6
> > >
> > >     osdmap e10182: 10 osds: 10 up, 10 in
> > >
> > >      pgmap v3295880: 1024 pgs, 2 pools, 4563 GB data, 1143 kobjects
> > >
> > >            9136 GB used, 5710 GB / 14846 GB avail
> > >
> > >                1005 active+clean
> > >
> > >                  16 incomplete
> > >
> > >                   3 down+incomplete
> > >
> > >
> > > Unfortunately "7 requests blocked" means no virtual machine can boot
> > >
> > > because ceph has stopped i/o.
> > >
> > >
> > > I can accept to lose some data, but not ALL data!
> > >
> > > Can you help me please?
> > >
> > > Thanks,
> > >
> > > Mario
> > >
> > >
> > > _______________________________________________
> > >
> > > ceph-users mailing list
> > >
> > > ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>
> > >
> > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 
> > > <http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com>
> > >
> > >
> > > _______________________________________________
> > > ceph-users mailing list
> > > ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>
> > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 
> > > <http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com>
> > >
> > > _______________________________________________
> > > ceph-users mailing list
> > > ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>
> > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 
> > > <http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com>
> > >
> 
> 
> --
> Christian Balzer        Network/Systems Engineer
> ch...@gol.com <mailto:ch...@gol.com>           Global OnLine Japan/Rakuten 
> Communications
> http://www.gol.com/ <http://www.gol.com/>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to