On 7/05/2022 10:04 am, Alan Somers wrote:
The branch main has been updated by asomers:

URL: 
https://cgit.FreeBSD.org/src/commit/?id=1d2421ad8b6d508ef155752bdfc5948f7373bac3

commit 1d2421ad8b6d508ef155752bdfc5948f7373bac3
Author:     Alan Somers <asom...@freebsd.org>
AuthorDate: 2022-05-05 21:35:23 +0000
Commit:     Alan Somers <asom...@freebsd.org>
CommitDate: 2022-05-06 23:25:43 +0000

     Correctly measure system load averages > 1024
The old fixed-point arithmetic used for calculating load averages had an
     overflow at 1024.  So on systems with extremely high load, the observed
     load average would actually fall back to 0 and shoot up again, creating
     a kind of sawtooth graph.
Fix this by using 64-bit math internally, while still reporting the load
     average to userspace as a 32-bit number.
Sponsored by: Axcient
     Reviewed by:    imp
     Differential Revision: https://reviews.freebsd.org/D35134

Can MFC?

---
  sys/kern/kern_synch.c | 9 +++++----
  sys/kern/tty_info.c   | 2 +-
  sys/sys/param.h       | 8 ++++----
  3 files changed, 10 insertions(+), 9 deletions(-)

diff --git a/sys/kern/kern_synch.c b/sys/kern/kern_synch.c
index e78878987b57..381d6315044c 100644
--- a/sys/kern/kern_synch.c
+++ b/sys/kern/kern_synch.c
@@ -87,7 +87,7 @@ struct loadavg averunnable =
   * Constants for averages over 1, 5, and 15 minutes
   * when sampling at 5 second intervals.
   */
-static fixpt_t cexp[3] = {
+static uint64_t cexp[3] = {
        0.9200444146293232 * FSCALE,    /* exp(-1/12) */
        0.9834714538216174 * FSCALE,    /* exp(-1/60) */
        0.9944598480048967 * FSCALE,    /* exp(-1/180) */
@@ -611,14 +611,15 @@ setrunnable(struct thread *td, int srqflags)
  static void
  loadav(void *arg)
  {
-       int i, nrun;
+       int i;
+       uint64_t nrun;
        struct loadavg *avg;
- nrun = sched_load();
+       nrun = (uint64_t)sched_load();
        avg = &averunnable;
for (i = 0; i < 3; i++)
-               avg->ldavg[i] = (cexp[i] * avg->ldavg[i] +
+               avg->ldavg[i] = (cexp[i] * (uint64_t)avg->ldavg[i] +
                    nrun * FSCALE * (FSCALE - cexp[i])) >> FSHIFT;
/*
diff --git a/sys/kern/tty_info.c b/sys/kern/tty_info.c
index 60675557e4ed..237aa47a18da 100644
--- a/sys/kern/tty_info.c
+++ b/sys/kern/tty_info.c
@@ -302,7 +302,7 @@ tty_info(struct tty *tp)
        sbuf_set_drain(&sb, sbuf_tty_drain, tp);
/* Print load average. */
-       load = (averunnable.ldavg[0] * 100 + FSCALE / 2) >> FSHIFT;
+       load = ((int64_t)averunnable.ldavg[0] * 100 + FSCALE / 2) >> FSHIFT;
        sbuf_printf(&sb, "%sload: %d.%02d ", tp->t_column == 0 ? "" : "\n",
            load / 100, load % 100);
diff --git a/sys/sys/param.h b/sys/sys/param.h
index 2d463b9ac7a2..b0b53f1a7776 100644
--- a/sys/sys/param.h
+++ b/sys/sys/param.h
@@ -361,12 +361,12 @@ __END_DECLS
   * Scale factor for scaled integers used to count %cpu time and load avgs.
   *
   * The number of CPU `tick's that map to a unique `%age' can be expressed
- * by the formula (1 / (2 ^ (FSHIFT - 11))).  The maximum load average that
- * can be calculated (assuming 32 bits) can be closely approximated using
- * the formula (2 ^ (2 * (16 - FSHIFT))) for (FSHIFT < 15).
+ * by the formula (1 / (2 ^ (FSHIFT - 11))).  Since the intermediate
+ * calculation is done with 64-bit precision, the maximum load average that can
+ * be calculated is approximately 2^32 / FSCALE.
   *
   * For the scheduler to maintain a 1:1 mapping of CPU `tick' to `%age',
- * FSHIFT must be at least 11; this gives us a maximum load avg of ~1024.
+ * FSHIFT must be at least 11.  This gives a maximum load avg of 2 million.
   */
  #define       FSHIFT  11              /* bits to right of fixed binary point 
*/
  #define FSCALE        (1<<FSHIFT)




Reply via email to