Try running "ceph health detail" on each of the monitors. Your disk space thresholds probably aren't configured correctly or something. -Greg
Software Engineer #42 @ http://inktank.com | http://ceph.com On Tue, Jun 17, 2014 at 2:09 AM, Andrija Panic <andrija.pa...@gmail.com> wrote: > Hi, > > thanks for that, but is not space issue: > > OSD drives are only 12% full. > and /var drive on which MON lives is over 70% only on CS3 server, but I > have increased alert treshold in ceph.conf (mon data avail warn = 15, mon > data avail crit = 5), and since I increased them those alerts are gone > (anyway, these alerts for /var full over 70% can be normally seen in logs > and in ceph -w output). > > Here I get no normal/visible warning in eather logs or ceph -w output... > > Thanks, > Andrija > > > > > On 17 June 2014 11:00, Stanislav Yanchev <s.yanc...@maxtelecom.bg> wrote: > >> Try grep in cs1 and cs3 could be a disk space issue. >> >> >> >> >> >> Regards, >> >> *Stanislav Yanchev* >> Core System Administrator >> >> [image: MAX TELECOM] >> >> Mobile: +359 882 549 441 >> s.yanc...@maxtelecom.bg >> www.maxtelecom.bg >> >> >> *From:* ceph-users [mailto:ceph-users-boun...@lists.ceph.com] *On Behalf >> Of *Andrija Panic >> *Sent:* Tuesday, June 17, 2014 11:57 AM >> *To:* Christian Balzer >> *Cc:* ceph-users@lists.ceph.com >> *Subject:* Re: [ceph-users] Cluster status reported wrongly as >> HEALTH_WARN >> >> >> >> Hi Christian, >> >> >> >> that seems true, thanks. >> >> >> >> But again, there are only occurence in GZ logs files (that were >> logrotated, not in current log files): >> >> Example: >> >> >> >> [root@cs2 ~]# grep -ir "WRN" /var/log/ceph/ >> >> Binary file /var/log/ceph/ceph-mon.cs2.log-20140612.gz matches >> >> Binary file /var/log/ceph/ceph.log-20140614.gz matches >> >> Binary file /var/log/ceph/ceph.log-20140611.gz matches >> >> Binary file /var/log/ceph/ceph.log-20140612.gz matches >> >> Binary file /var/log/ceph/ceph.log-20140613.gz matches >> >> >> >> Thanks, >> >> Andrija >> >> >> >> On 17 June 2014 10:48, Christian Balzer <ch...@gol.com> wrote: >> >> >> Hello, >> >> >> On Tue, 17 Jun 2014 10:30:44 +0200 Andrija Panic wrote: >> >> > Hi, >> > >> > I have 3 node (2 OSD per node) CEPH cluster, running fine, not much >> data, >> > network also fine: >> > Ceph ceph-0.72.2. >> > >> > When I issue "ceph status" command, I get randomly HEALTH_OK, and >> > imidiately after that when repeating command, I get HEALTH_WARN >> > >> > Examle given down - these commands were issues within less than 1 sec >> > between them >> > There are NO occuring of word "warn" in the logs (grep -ir "warn" >> > /var/log/ceph) on any of the servers... >> > I get false alerts with my status monitoring script, for this reason... >> > >> >> If I recall correctly, the logs will show INF, WRN and ERR, so grep for >> WRN. >> >> Regards, >> >> Christian >> >> >> > Any help would be greatly appriciated. >> > >> > Thanks, >> > >> > [root@cs3 ~]# ceph status >> > cluster cab20370-bf6a-4589-8010-8d5fc8682eab >> > health HEALTH_OK >> > monmap e2: 3 mons at >> > >> {cs1=10.44.xxx.10:6789/0,cs2=10.44.xxx.11:6789/0,cs3=10.44.xxx.12:6789/0}, >> > election epoch 122, quorum 0,1,2 cs1,cs2,cs3 >> > osdmap e890: 6 osds: 6 up, 6 in >> > pgmap v2379904: 448 pgs, 4 pools, 862 GB data, 217 kobjects >> > 2576 GB used, 19732 GB / 22309 GB avail >> > 448 active+clean >> > client io 17331 kB/s rd, 113 kB/s wr, 176 op/s >> > >> > [root@cs3 ~]# ceph status >> > cluster cab20370-bf6a-4589-8010-8d5fc8682eab >> > health HEALTH_WARN >> > monmap e2: 3 mons at >> > >> {cs1=10.44.xxx.10:6789/0,cs2=10.44.xxx.11:6789/0,cs3=10.44.xxx.12:6789/0}, >> > election epoch 122, quorum 0,1,2 cs1,cs2,cs3 >> > osdmap e890: 6 osds: 6 up, 6 in >> > pgmap v2379905: 448 pgs, 4 pools, 862 GB data, 217 kobjects >> > 2576 GB used, 19732 GB / 22309 GB avail >> > 448 active+clean >> > client io 28383 kB/s rd, 566 kB/s wr, 321 op/s >> > >> > [root@cs3 ~]# ceph status >> > cluster cab20370-bf6a-4589-8010-8d5fc8682eab >> > health HEALTH_OK >> > monmap e2: 3 mons at >> > >> {cs1=10.44.xxx.10:6789/0,cs2=10.44.xxx.11:6789/0,cs3=10.44.xxx.12:6789/0}, >> > election epoch 122, quorum 0,1,2 cs1,cs2,cs3 >> > osdmap e890: 6 osds: 6 up, 6 in >> > pgmap v2379913: 448 pgs, 4 pools, 862 GB data, 217 kobjects >> > 2576 GB used, 19732 GB / 22309 GB avail >> > 448 active+clean >> > client io 21632 kB/s rd, 49354 B/s wr, 283 op/s >> > >> >> >> -- >> >> Christian Balzer Network/Systems Engineer >> ch...@gol.com Global OnLine Japan/Fusion Communications >> http://www.gol.com/ >> >> >> >> >> >> -- >> >> >> >> Andrija Panić >> >> -------------------------------------- >> >> http://admintweets.com >> >> -------------------------------------- >> >> <http://gfidisc.maxtelecom.bg> >> >> *Confidentiality notice* >> ------------------------------ >> >> >> >> The information contained in this message (including any attachments) is >> confidential and may be legally privileged or otherwise protected from >> disclosure. This message is intended solely for the addressee(s). If you >> are not the intended recipient, please notify the sender by return e-mail >> and delete this message from your system. Any unauthorised use, >> reproduction, or dissemination of this message is strictly prohibited. Any >> liability arising from any third party acting, or refraining from acting, >> on any information contained in this e-mail is hereby excluded. Please note >> that e-mails are susceptible to change. Max Telecom shall not be liable for >> the improper or incomplete transmission of the information contained in >> this communication, nor shall it be liable for any delay in its receipt. >> >> <http://gfidisc.maxtelecom.bg> >> > > > > -- > > Andrija Panić > -------------------------------------- > http://admintweets.com > -------------------------------------- > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com