`ceph pg stat` might be cleaner to watch than the `ceph status | grep
pgs`.  I also like watching `ceph osd pool stats` which breaks down all IO
by pool.  You also have the option of the dashboard mgr service which has a
lot of useful information including the pool IO breakdown.

On Thu, Mar 1, 2018 at 7:22 AM Georgios Dimitrakakis <gior...@acmac.uoc.gr>
wrote:

>
>  Excellent! Good to know that the behavior is intentional!
>
>  Thanks a lot John for the feedback!
>
>  Best regards,
>
>  G.
>
> > On Thu, Mar 1, 2018 at 12:03 PM, Georgios Dimitrakakis
> > <gior...@acmac.uoc.gr> wrote:
> >> I have recently updated to Luminous (12.2.4) and I have noticed that
> >> using
> >> "ceph -w" only produces an initial output like the one below but
> >> never gets
> >> updated afterwards. Is this a feature because I was used to the old
> >> way that
> >> was constantly
> >> producing info.
> >
> > It's intentional.  "ceph -w" is the command that follows the Ceph
> > cluster log.  The monitor used to dump the pg status into the cluster
> > log every 5 seconds, which was useful sometimes, but also made the
> > log
> > pretty unreadable for anything else, because other output was quickly
> > swamped with the pg status spam.
> >
> > To replicate the de-facto old behaviour (print the pg status every 5
> > seconds), you can always do something like `watch -n1 "ceph status |
> > grep pgs"`
> >
> > There's work ongoing to create a nice replacement that does a status
> > stream without spamming the cluster log to accomplish it here:
> > https://github.com/ceph/ceph/pull/20100
> >
> > Cheers,
> > John
> >
> >>
> >> Here is what I get as initial output which is not updated:
> >>
> >> $ ceph -w
> >>   cluster:
> >>     id:     d357a551-5b7a-4501-8d8f-009c63b2c972
> >>     health: HEALTH_OK
> >>
> >>   services:
> >>     mon: 1 daemons, quorum node1
> >>     mgr: node1(active)
> >>     osd: 2 osds: 2 up, 2 in
> >>     rgw: 1 daemon active
> >>
> >>   data:
> >>     pools:   11 pools, 152 pgs
> >>     objects: 9786 objects, 33754 MB
> >>     usage:   67494 MB used, 3648 GB / 3714 GB avail
> >>     pgs:     152 active+clean
> >>
> >>
> >>
> >> Even if I create a new volume in my Openstack installation, assign
> >> it to a
> >> VM, mount it and format it, I have to stop and re-execute the "ceph
> >> -w"
> >> command to see the following line:
> >>
> >>
> >>   io:
> >>     client:   767 B/s rd, 511 B/s wr, 0 op/s rd, 0 op/s wr
> >>
> >> which also pauses after the first display.
> >>
> >>
> >> Kind regards,
> >>
> >>
> >> G.
> >> _______________________________________________
> >> ceph-users mailing list
> >> ceph-users@lists.ceph.com
> >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to