Sorry the `sysctl kern.timecounter` was before your patch.
Here's the one after your patch:
$ sysctl kern.timecounter
kern.timecounter.tick=1
kern.timecounter.timestepwarnings=0
kern.timecounter.hardware=i8254
kern.timecounter.choice=i8254(0) tsc(-1000) acpihpet0(1000) acpitimer0(1000)
On Tue, Jul 05, 2022 at 10:31:33PM -0400, Mohamed Aslan wrote:
> Hello,
>
> I just tested your patch on my lenovo e495 laptop, unfortunately
> still no tsc.
>
> $ dmesg | grep 'tsc:'
> tsc: cpu0/cpu1 sync round 1: 20468 regressions
> tsc: cpu0/cpu1 sync round 1: cpu0 lags cpu1 by 5351060292 cycles
> tsc: cpu0/cpu1 sync round 1: cpu1 lags cpu0 by 0 cycles
> tsc: cpu0/cpu2 sync round 1: 10272 regressions
> tsc: cpu0/cpu2 sync round 1: cpu0 lags cpu2 by 5351060271 cycles
> tsc: cpu0/cpu2 sync round 1: cpu2 lags cpu0 by 0 cycles
> tsc: cpu0/cpu3 sync round 1: 9709 regressions
> tsc: cpu0/cpu3 sync round 1: cpu0 lags cpu3 by 5351060271 cycles
> tsc: cpu0/cpu3 sync round 1: cpu3 lags cpu0 by 0 cycles
> tsc: cpu0/cpu4 sync round 1: 10386 regressions
> tsc: cpu0/cpu4 sync round 1: cpu0 lags cpu4 by 5351060271 cycles
> tsc: cpu0/cpu4 sync round 1: cpu4 lags cpu0 by 0 cycles
> tsc: cpu0/cpu5 sync round 1: 10408 regressions
> tsc: cpu0/cpu5 sync round 1: cpu0 lags cpu5 by 5351060271 cycles
> tsc: cpu0/cpu5 sync round 1: cpu5 lags cpu0 by 0 cycles
> tsc: cpu0/cpu6 sync round 1: 10102 regressions
> tsc: cpu0/cpu6 sync round 1: cpu0 lags cpu6 by 5351060271 cycles
> tsc: cpu0/cpu6 sync round 1: cpu6 lags cpu0 by 0 cycles
> tsc: cpu0/cpu7 sync round 1: 9336 regressions
> tsc: cpu0/cpu7 sync round 1: cpu0 lags cpu7 by 5351060271 cycles
> tsc: cpu0/cpu7 sync round 1: cpu7 lags cpu0 by 0 cycles
>
> $ sysctl kern.timecounter
> kern.timecounter.tick=1
> kern.timecounter.timestepwarnings=0
> kern.timecounter.hardware=tsc
> kern.timecounter.choice=i8254(0) tsc(2000) acpihpet0(1000) acpitimer0(1000)
>
> $ sysctl hw
> hw.model=AMD Ryzen 5 3500U with Radeon Vega Mobile Gfx
> hw.ncpu=8
> hw.vendor=LENOVO
> hw.product=20NE0002US
> hw.version=ThinkPad E495
> hw.ncpufound=8
> hw.smt=0
> hw.ncpuonline=4
>
>
> Best,
> M.
>
>
> On Mon, Jul 04, 2022 at 09:06:55PM -0500, Scott Cheloha wrote:
> > Hi,
> >
> > Once again, I am trying to change our approach to TSC sync testing to
> > eliminate false positive results. Instead of trying to repair the TSC
> > by measuring skew, we just spin in a lockless loop looking for skew
> > and mark the TSC as broken if we detect any.
> >
> > This is motivated in part by some multisocket machines that do not use
> > the TSC as a timecounter because the current sync test confuses NUMA
> > latency for TSC skew.
> >
> > The only difference between this version and the prior version (v2) is
> > that we check whether we have the IA32_TSC_ADJUST register by hand in
> > tsc_reset_adjust(). If someone wants to help me rearrange cpu_hatch()
> > so we do CPU identification before TSC sync testing we can remove the
> > workaround later.
> >
> > If you have the IA32_TSC_ADJUST register and it is non-zero going into
> > the test, you will see something on the console like this:
> >
> > tsc: cpu5: IA32_TSC_ADJUST: -150 -> 0
> >
> > This does *not* mean you failed the test. It just means you probably
> > have a bug in your BIOS (or some other firmware) and you should report
> > it to your vendor.
> >
> > If you fail the test you will see something like this:
> >
> > tsc: cpu0/cpu2: sync test round 1/2 failed
> > tsc: cpu0/cpu2: cpu2: 13043 lags 438 cycles
> >
> > A printout like this would mean that the sync test for cpu2 failed.
> > In particular, cpu2's TSC trails cpu0's TSC by at least 438 cycles.
> > If this happens for *any* CPU we mark the TSC timecounter as
> > defective.
> >
> > --
> >
> > Please test! Send your dmesg, pass or fail.
> >
> > I am especially interested in:
> >
> > 1. A test from dv@. Your dual-socket machine has the IA32_TSC_ADJUST
> > register but it failed the test running patch v2. Maybe it will pass
> > with this version?
> >
> > 2. Other multisocket machines.
> >
> > 3. There were reports of TSC issues with OpenBSD VMs running on ESXi.
> > What happens when you run with this patch?
> >
> > 4. OpenBSD VMs on other hypervisors.
> >
> > 5. Non-Lenovo machines, non-Intel machines.
> >
> > -Scott
> >
> > Index: amd64/tsc.c
> > ===================================================================
> > RCS file: /cvs/src/sys/arch/amd64/amd64/tsc.c,v
> > retrieving revision 1.24
> > diff -u -p -r1.24 tsc.c
> > --- amd64/tsc.c 31 Aug 2021 15:11:54 -0000 1.24
> > +++ amd64/tsc.c 5 Jul 2022 01:52:10 -0000
> > @@ -36,13 +36,6 @@ int tsc_recalibrate;
> > uint64_t tsc_frequency;
> > int tsc_is_invariant;
> >
> > -#define TSC_DRIFT_MAX 250
> > -#define TSC_SKEW_MAX 100
> > -int64_t tsc_drift_observed;
> > -
> > -volatile int64_t tsc_sync_val;
> > -volatile struct cpu_info *tsc_sync_cpu;
> > -
> > u_int tsc_get_timecount(struct timecounter *tc);
> > void tsc_delay(int usecs);
> >
> > @@ -236,22 +229,12 @@ cpu_recalibrate_tsc(struct timecounter *
> > u_int
> > tsc_get_timecount(struct timecounter *tc)
> > {
> > - return rdtsc_lfence() + curcpu()->ci_tsc_skew;
> > + return rdtsc_lfence();
> > }
> >
> > void
> > tsc_timecounter_init(struct cpu_info *ci, uint64_t cpufreq)
> > {
> > -#ifdef TSC_DEBUG
> > - printf("%s: TSC skew=%lld observed drift=%lld\n", ci->ci_dev->dv_xname,
> > - (long long)ci->ci_tsc_skew, (long long)tsc_drift_observed);
> > -#endif
> > - if (ci->ci_tsc_skew < -TSC_SKEW_MAX || ci->ci_tsc_skew > TSC_SKEW_MAX) {
> > - printf("%s: disabling user TSC (skew=%lld)\n",
> > - ci->ci_dev->dv_xname, (long long)ci->ci_tsc_skew);
> > - tsc_timecounter.tc_user = 0;
> > - }
> > -
> > if (!(ci->ci_flags & CPUF_PRIMARY) ||
> > !(ci->ci_flags & CPUF_CONST_TSC) ||
> > !(ci->ci_flags & CPUF_INVAR_TSC))
> > @@ -268,111 +251,267 @@ tsc_timecounter_init(struct cpu_info *ci
> > calibrate_tsc_freq();
> > }
> >
> > - if (tsc_drift_observed > TSC_DRIFT_MAX) {
> > - printf("ERROR: %lld cycle TSC drift observed\n",
> > - (long long)tsc_drift_observed);
> > - tsc_timecounter.tc_quality = -1000;
> > - tsc_timecounter.tc_user = 0;
> > - tsc_is_invariant = 0;
> > - }
> > -
> > tc_init(&tsc_timecounter);
> > }
> >
> > -/*
> > - * Record drift (in clock cycles). Called during AP startup.
> > - */
> > void
> > -tsc_sync_drift(int64_t drift)
> > +tsc_delay(int usecs)
> > {
> > - if (drift < 0)
> > - drift = -drift;
> > - if (drift > tsc_drift_observed)
> > - tsc_drift_observed = drift;
> > + uint64_t interval, start;
> > +
> > + interval = (uint64_t)usecs * tsc_frequency / 1000000;
> > + start = rdtsc_lfence();
> > + while (rdtsc_lfence() - start < interval)
> > + CPU_BUSY_CYCLE();
> > }
> >
> > +#ifdef MULTIPROCESSOR
> > +
> > +#define TSC_DEBUG 1
> > +
> > /*
> > - * Called during startup of APs, by the boot processor. Interrupts
> > - * are disabled on entry.
> > + * Protections for global variables in this code:
> > + *
> > + * a Modified atomically
> > + * b Protected by a barrier
> > + * p Only modified by the primary CPU
> > */
> > +
> > +#define TSC_TEST_MS 1 /* Test round duration */
> > +#define TSC_TEST_ROUNDS 2 /* Number of test rounds */
> > +
> > +struct tsc_test_status {
> > + volatile uint64_t val __aligned(64); /* [b] CPU's latest TSC value */
> > + uint64_t lag_count; /* [b] # of lags seen by CPU */
> > + uint64_t lag_max; /* [b] Biggest lag seen */
> > + int64_t adj_val; /* [b] initial TSC_ADJUST val */
> > +};
> > +struct tsc_test_status tsc_ap_status; /* [b] Test results from AP */
> > +struct tsc_test_status tsc_bp_status; /* [p] Test results from BP */
> > +uint64_t tsc_test_cycles; /* [p] TSC cycles per test round */
> > +const char *tsc_ap_name; /* [b] Name of AP */
> > +volatile u_int tsc_egress_barrier; /* [a] Test end barrier */
> > +volatile u_int tsc_ingress_barrier; /* [a] Test start barrier */
> > +volatile u_int tsc_test_rounds; /* [p] Remaining test rounds */
> > +int tsc_is_synchronized = 1; /* [p] TSC sync'd across all
> > CPUs? */
> > +
> > +void tsc_report_test_results(void);
> > +void tsc_reset_adjust(struct tsc_test_status *);
> > +void tsc_test_ap(void);
> > +void tsc_test_bp(void);
> > +
> > void
> > -tsc_read_bp(struct cpu_info *ci, uint64_t *bptscp, uint64_t *aptscp)
> > +tsc_test_sync_bp(struct cpu_info *ci)
> > {
> > - uint64_t bptsc;
> > + /* TSC must be constant and invariant to use it as a timecounter. */
> > +#ifndef TSC_DEBUG
> > + if (!tsc_is_invariant)
> > + return;
> > +#endif
> >
> > - if (atomic_swap_ptr(&tsc_sync_cpu, ci) != NULL)
> > - panic("tsc_sync_bp: 1");
> > + /* Reset IA32_TSC_ADJUST if it exists. */
> > + tsc_reset_adjust(&tsc_bp_status);
> >
> > - /* Flag it and read our TSC. */
> > - atomic_setbits_int(&ci->ci_flags, CPUF_SYNCTSC);
> > - bptsc = (rdtsc_lfence() >> 1);
> > + /* Reset the test cycle limit and round count. */
> > + tsc_test_cycles = TSC_TEST_MS * tsc_frequency / 1000;
> > + tsc_test_rounds = TSC_TEST_ROUNDS;
> > +
> > + do {
> > + /*
> > + * Pass through the ingress barrier, run the test,
> > + * then wait for the AP to reach the egress barrier.
> > + */
> > + atomic_inc_int(&tsc_ingress_barrier);
> > + while (tsc_ingress_barrier != 2)
> > + CPU_BUSY_CYCLE();
> > + tsc_test_bp();
> > + while (tsc_egress_barrier != 1)
> > + CPU_BUSY_CYCLE();
> > +
> > + /*
> > + * Report what happened. Adjust the timecounter quality
> > + * if we failed the test.
> > + *
> > + * XXX We need a tc_detach() function to cleanly
> > + * disable the timecounter. Lowering the quality
> > + * like this is a nasty hack.
> > + */
> > + tsc_report_test_results();
> > + if (tsc_ap_status.lag_count || tsc_bp_status.lag_count) {
> > + tsc_timecounter.tc_quality = -1000;
> > + tsc_is_synchronized = 0;
> > + tsc_test_rounds = 0;
> > + } else
> > + tsc_test_rounds--;
> > +
> > + /*
> > + * Clean up for the next round.
> > + *
> > + * It is safe to reset the ingress barrier because
> > + * at this point we know the AP has reached the egress
> > + * barrier.
> > + */
> > + memset(&tsc_ap_status, 0, sizeof tsc_ap_status);
> > + memset(&tsc_bp_status, 0, sizeof tsc_bp_status);
> > + tsc_ingress_barrier = 0;
> > + if (tsc_test_rounds == 0)
> > + tsc_ap_name = NULL;
> > +
> > + /*
> > + * Pass through the egress barrier and release the AP.
> > + * The AP is responsible for resetting the egress barrier.
> > + */
> > + if (atomic_inc_int_nv(&tsc_egress_barrier) != 2)
> > + panic("%s: unexpected egress count", __func__);
> > + } while (tsc_test_rounds > 0);
> > +}
> > +
> > +void
> > +tsc_test_sync_ap(struct cpu_info *ci)
> > +{
> > +#ifndef TSC_DEBUG
> > + if (!tsc_is_invariant)
> > + return;
> > +#endif
> >
> > - /* Wait for remote to complete, and read ours again. */
> > - while ((ci->ci_flags & CPUF_SYNCTSC) != 0)
> > - membar_consumer();
> > - bptsc += (rdtsc_lfence() >> 1);
> > + /* The BP needs our name in order to report any problems. */
> > + tsc_ap_name = ci->ci_dev->dv_xname;
> >
> > - /* Wait for the results to come in. */
> > - while (tsc_sync_cpu == ci)
> > - CPU_BUSY_CYCLE();
> > - if (tsc_sync_cpu != NULL)
> > - panic("tsc_sync_bp: 2");
> > + tsc_reset_adjust(&tsc_ap_status);
> >
> > - *bptscp = bptsc;
> > - *aptscp = tsc_sync_val;
> > + /*
> > + * The AP is only responsible for running the test and
> > + * resetting the egress barrier. The BP sets up and tears
> > + * down everything else.
> > + */
> > + do {
> > + atomic_inc_int(&tsc_ingress_barrier);
> > + while (tsc_ingress_barrier != 2)
> > + CPU_BUSY_CYCLE();
> > + tsc_test_ap();
> > + atomic_inc_int(&tsc_egress_barrier);
> > + while (atomic_cas_uint(&tsc_egress_barrier, 2, 0) != 2)
> > + CPU_BUSY_CYCLE();
> > + } while (tsc_test_rounds > 0);
> > }
> >
> > void
> > -tsc_sync_bp(struct cpu_info *ci)
> > +tsc_report_test_results(void)
> > {
> > - uint64_t bptsc, aptsc;
> > + u_int round = TSC_TEST_ROUNDS - tsc_test_rounds + 1;
> >
> > - tsc_read_bp(ci, &bptsc, &aptsc); /* discarded - cache effects */
> > - tsc_read_bp(ci, &bptsc, &aptsc);
> > -
> > - /* Compute final value to adjust for skew. */
> > - ci->ci_tsc_skew = bptsc - aptsc;
> > + if (tsc_bp_status.adj_val != 0) {
> > + printf("tsc: cpu0: IA32_TSC_ADJUST: %lld -> 0\n",
> > + tsc_bp_status.adj_val);
> > + }
> > + if (tsc_ap_status.adj_val != 0) {
> > + printf("tsc: %s: IA32_TSC_ADJUST: %lld -> 0\n",
> > + tsc_ap_name, tsc_ap_status.adj_val);
> > + }
> > + if (tsc_ap_status.lag_count > 0 || tsc_bp_status.lag_count > 0) {
> > + printf("tsc: cpu0/%s: sync test round %u/%u failed\n",
> > + tsc_ap_name, round, TSC_TEST_ROUNDS);
> > + }
> > + if (tsc_bp_status.lag_count > 0) {
> > + printf("tsc: cpu0/%s: cpu0: %llu lags %llu cycles\n",
> > + tsc_ap_name, tsc_bp_status.lag_count,
> > + tsc_bp_status.lag_max);
> > + }
> > + if (tsc_ap_status.lag_count > 0) {
> > + printf("tsc: cpu0/%s: %s: %llu lags %llu cycles\n",
> > + tsc_ap_name, tsc_ap_name, tsc_ap_status.lag_count,
> > + tsc_ap_status.lag_max);
> > + }
> > }
> >
> > /*
> > - * Called during startup of AP, by the AP itself. Interrupts are
> > - * disabled on entry.
> > + * Reset IA32_TSC_ADJUST if we have it.
> > + *
> > + * XXX We should rearrange cpu_hatch() so that the feature flags
> > + * are already set before we get here. Do CPUID() and CPUID_LEAF()
> > + * here as a workaround until then.
> > */
> > void
> > -tsc_post_ap(struct cpu_info *ci)
> > +tsc_reset_adjust(struct tsc_test_status *tts)
> > {
> > - uint64_t tsc;
> > -
> > - /* Wait for go-ahead from primary. */
> > - while ((ci->ci_flags & CPUF_SYNCTSC) == 0)
> > - membar_consumer();
> > - tsc = (rdtsc_lfence() >> 1);
> > + uint32_t eax, ebx, ecx, edx;
> >
> > - /* Instruct primary to read its counter. */
> > - atomic_clearbits_int(&ci->ci_flags, CPUF_SYNCTSC);
> > - tsc += (rdtsc_lfence() >> 1);
> > -
> > - /* Post result. Ensure the whole value goes out atomically. */
> > - (void)atomic_swap_64(&tsc_sync_val, tsc);
> > -
> > - if (atomic_swap_ptr(&tsc_sync_cpu, NULL) != ci)
> > - panic("tsc_sync_ap");
> > + CPUID(0, eax, ebx, ecx, edx);
> > + if (eax >= 7) {
> > + CPUID_LEAF(7, 0, eax, ebx, ecx, edx);
> > + if (ISSET(ebx, SEFF0EBX_TSC_ADJUST)) {
> > + tts->adj_val = rdmsr(MSR_TSC_ADJUST);
> > + if (tts->adj_val != 0)
> > + wrmsr(MSR_TSC_ADJUST, 0);
> > + }
> > + }
> > }
> >
> > void
> > -tsc_sync_ap(struct cpu_info *ci)
> > +tsc_test_ap(void)
> > {
> > - tsc_post_ap(ci);
> > - tsc_post_ap(ci);
> > + uint64_t ap_val, bp_val, end, lag;
> > + u_int i = 0;
> > +
> > + ap_val = rdtsc_lfence();
> > + end = ap_val + tsc_test_cycles;
> > + while (ap_val < end) {
> > + /*
> > + * The LFENCE ensures bp_val predates ap_val. If ap_val
> > + * is smaller than bp_val then the AP's TSC must lag that
> > + * of the BP, i.e. the two cannot be synchronized.
> > + */
> > + bp_val = tsc_bp_status.val;
> > + ap_val = rdtsc_lfence();
> > + tsc_ap_status.val = ap_val;
> > +
> > + /*
> > + * Ensure the other CPU has a chance to run. This is
> > + * crucial if the other CPU is our SMT sibling. SMT
> > + * starvation can prevent this test from detecting
> > + * relatively large lags. Eight is an arbitrary value,
> > + * but it seems to work in practice without impacting
> > + * the sensitivity of the test for non-sibling threads.
> > + */
> > + if (++i % 8 == 0)
> > + CPU_BUSY_CYCLE();
> > +
> > + /*
> > + * Record the magnitude of the problem if our TSC lags
> > + * the other.
> > + */
> > + if (__predict_false(ap_val < bp_val)) {
> > + tsc_ap_status.lag_count++;
> > + lag = bp_val - ap_val;
> > + if (tsc_ap_status.lag_max < lag)
> > + tsc_ap_status.lag_max = lag;
> > + }
> > + }
> > }
> >
> > void
> > -tsc_delay(int usecs)
> > +tsc_test_bp(void)
> > {
> > - uint64_t interval, start;
> > + uint64_t ap_val, bp_val, end, lag;
> > + u_int i = 0;
> >
> > - interval = (uint64_t)usecs * tsc_frequency / 1000000;
> > - start = rdtsc_lfence();
> > - while (rdtsc_lfence() - start < interval)
> > - CPU_BUSY_CYCLE();
> > + bp_val = rdtsc_lfence();
> > + end = bp_val + tsc_test_cycles;
> > + while (bp_val < end) {
> > + ap_val = tsc_ap_status.val;
> > + bp_val = rdtsc_lfence();
> > + tsc_bp_status.val = bp_val;
> > +
> > + if (++i % 8 == 0)
> > + CPU_BUSY_CYCLE();
> > +
> > + if (__predict_false(bp_val < ap_val)) {
> > + tsc_bp_status.lag_count++;
> > + lag = ap_val - bp_val;
> > + if (tsc_bp_status.lag_max < lag)
> > + tsc_bp_status.lag_max = lag;
> > + }
> > + }
> > }
> > +
> > +#endif /* MULTIPROCESSOR */
> > Index: amd64/cpu.c
> > ===================================================================
> > RCS file: /cvs/src/sys/arch/amd64/amd64/cpu.c,v
> > retrieving revision 1.156
> > diff -u -p -r1.156 cpu.c
> > --- amd64/cpu.c 26 Apr 2022 08:35:30 -0000 1.156
> > +++ amd64/cpu.c 5 Jul 2022 01:52:11 -0000
> > @@ -772,9 +772,9 @@ cpu_init(struct cpu_info *ci)
> > lcr4(cr4 & ~CR4_PGE);
> > lcr4(cr4);
> >
> > - /* Synchronize TSC */
> > + /* Check if TSC is synchronized. */
> > if (cold && !CPU_IS_PRIMARY(ci))
> > - tsc_sync_ap(ci);
> > + tsc_test_sync_ap(ci);
> > #endif
> > }
> >
> > @@ -854,18 +854,14 @@ cpu_start_secondary(struct cpu_info *ci)
> > #endif
> > } else {
> > /*
> > - * Synchronize time stamp counters. Invalidate cache and
> > - * synchronize twice (in tsc_sync_bp) to minimize possible
> > - * cache effects. Disable interrupts to try and rule out any
> > - * external interference.
> > + * Test if TSCs are synchronized. Invalidate cache to
> > + * minimize possible cache effects. Disable interrupts to
> > + * rule out external interference.
> > */
> > s = intr_disable();
> > wbinvd();
> > - tsc_sync_bp(ci);
> > + tsc_test_sync_bp(ci);
> > intr_restore(s);
> > -#ifdef TSC_DEBUG
> > - printf("TSC skew=%lld\n", (long long)ci->ci_tsc_skew);
> > -#endif
> > }
> >
> > if ((ci->ci_flags & CPUF_IDENTIFIED) == 0) {
> > @@ -890,7 +886,6 @@ void
> > cpu_boot_secondary(struct cpu_info *ci)
> > {
> > int i;
> > - int64_t drift;
> > u_long s;
> >
> > atomic_setbits_int(&ci->ci_flags, CPUF_GO);
> > @@ -905,18 +900,11 @@ cpu_boot_secondary(struct cpu_info *ci)
> > db_enter();
> > #endif
> > } else if (cold) {
> > - /* Synchronize TSC again, check for drift. */
> > - drift = ci->ci_tsc_skew;
> > + /* Test if TSCs are synchronized again. */
> > s = intr_disable();
> > wbinvd();
> > - tsc_sync_bp(ci);
> > + tsc_test_sync_bp(ci);
> > intr_restore(s);
> > - drift -= ci->ci_tsc_skew;
> > -#ifdef TSC_DEBUG
> > - printf("TSC skew=%lld drift=%lld\n",
> > - (long long)ci->ci_tsc_skew, (long long)drift);
> > -#endif
> > - tsc_sync_drift(drift);
> > }
> > }
> >
> > @@ -942,13 +930,12 @@ cpu_hatch(void *v)
> > #endif
> >
> > /*
> > - * Synchronize the TSC for the first time. Note that interrupts are
> > - * off at this point.
> > + * Test if our TSC is synchronized for the first time.
> > + * Note that interrupts are off at this point.
> > */
> > wbinvd();
> > ci->ci_flags |= CPUF_PRESENT;
> > - ci->ci_tsc_skew = 0; /* reset on resume */
> > - tsc_sync_ap(ci);
> > + tsc_test_sync_ap(ci);
> >
> > lapic_enable();
> > lapic_startclock();
> > Index: include/cpu.h
> > ===================================================================
> > RCS file: /cvs/src/sys/arch/amd64/include/cpu.h,v
> > retrieving revision 1.144
> > diff -u -p -r1.144 cpu.h
> > --- include/cpu.h 28 Jun 2022 12:11:41 -0000 1.144
> > +++ include/cpu.h 5 Jul 2022 01:52:11 -0000
> > @@ -209,8 +209,6 @@ struct cpu_info {
> > paddr_t ci_vmxon_region_pa;
> > struct vmxon_region *ci_vmxon_region;
> >
> > - int64_t ci_tsc_skew; /* counter skew vs cpu0 */
> > -
> > char ci_panicbuf[512];
> >
> > paddr_t ci_vmcs_pa;
> > @@ -230,7 +228,6 @@ struct cpu_info {
> > #define CPUF_INVAR_TSC 0x0100 /* CPU has invariant TSC */
> > #define CPUF_USERXSTATE 0x0200 /* CPU has curproc's xsave
> > state */
> >
> > -#define CPUF_SYNCTSC 0x0800 /* Synchronize TSC */
> > #define CPUF_PRESENT 0x1000 /* CPU is present */
> > #define CPUF_RUNNING 0x2000 /* CPU is running */
> > #define CPUF_PAUSE 0x4000 /* CPU is paused in DDB */
> > Index: include/cpuvar.h
> > ===================================================================
> > RCS file: /cvs/src/sys/arch/amd64/include/cpuvar.h,v
> > retrieving revision 1.11
> > diff -u -p -r1.11 cpuvar.h
> > --- include/cpuvar.h 16 May 2021 04:33:05 -0000 1.11
> > +++ include/cpuvar.h 5 Jul 2022 01:52:11 -0000
> > @@ -97,8 +97,7 @@ void identifycpu(struct cpu_info *);
> > void cpu_init(struct cpu_info *);
> > void cpu_init_first(void);
> >
> > -void tsc_sync_drift(int64_t);
> > -void tsc_sync_bp(struct cpu_info *);
> > -void tsc_sync_ap(struct cpu_info *);
> > +void tsc_test_sync_bp(struct cpu_info *);
> > +void tsc_test_sync_ap(struct cpu_info *);
> >
> > #endif
> >
>
> --
> Mohamed 'Aslan' Abdelsalam, Ph.D.
> http://www.sce.carleton.ca/~maslan/
--
Mohamed 'Aslan' Abdelsalam, Ph.D.
http://www.sce.carleton.ca/~maslan/