On Tue, Aug 15, 2017 at 17:14 +0200, Mike Belopuhov wrote:
> Hi,
>
> I've just triggered an assert in hfsc_deferred (a callout) on an
> MP kernel running on an SP virtual machine:
>
> panic: kernel diagnostic assertion "HFSC_ENABLED(ifq)" failed: file
> "/home/mike/src/openbsd/sys/net/hfsc.c", line 950
> Stopped at db_enter+0x9: leave
> TID PID UID PRFLAGS PFLAGS CPU COMMAND
> *247463 28420 0 0x3 0 0 pfctl
> db_enter() at db_enter+0x9
>
> panic(ffffffff817f78f0,4,ffffffff81a3ffc0,ffffffff8110c140,ffff8000000c2060,fff
> fffff81598b1c) at panic+0x102
> __assert(ffffffff81769d93,ffffffff817d7350,3b6,ffffffff817d72bd) at
> __assert+0x
> 35
> hfsc_deferred(ffff8000000c2060) at hfsc_deferred+0x9e
> timeout_run(ffff80000004adc8) at timeout_run+0x4c
> softclock(0) at softclock+0x146
> softintr_dispatch(0) at softintr_dispatch+0x9f
> Xsoftclock() at Xsoftclock+0x1f
> --- interrupt ---
> end of kernel
> end trace frame: 0x728d481974c08548, count: 7
> 0x2cfe9c031c90000:
> https://www.openbsd.org/ddb.html describes the minimum info required in bug
> reports. Insufficient info makes it difficult to find and fix bugs.
> ddb{0}> ps
> PID TID PPID UID S FLAGS WAIT COMMAND
> *28420 247463 5000 0 7 0x3 pfctl
>
>
> pfctl runs in the loop reloading the ruleset. So at some point we
> disable HFSC on the interface but lose a race with hfsc_deferred
> before re-enabling it.
>
While talking to visa@, I've added this piece that explains where
there race is happening:
When we *disable* hfsc, we put back the prio ifq ops:
203 ifq_attach(struct ifqueue *ifq, const struct ifq_ops *newops, void *opsarg)
204 {
205 struct mbuf_list ml = MBUF_LIST_INITIALIZER();
206 struct mbuf_list free_ml = MBUF_LIST_INITIALIZER();
207 struct mbuf *m;
208 const struct ifq_ops *oldops;
209 void *newq, *oldq;
210
211 newq = newops->ifqop_alloc(ifq->ifq_idx, opsarg);
212
213 mtx_enter(&ifq->ifq_mtx);
214 ifq->ifq_ops->ifqop_purge(ifq, &ml);
215 ifq->ifq_len = 0;
216
217 oldops = ifq->ifq_ops;
218 oldq = ifq->ifq_q;
219
220 ifq->ifq_ops = newops;
221 ifq->ifq_q = newq;
222
223 while ((m = ml_dequeue(&ml)) != NULL) {
224 m = ifq->ifq_ops->ifqop_enq(ifq, m);
225 if (m != NULL) {
226 ifq->ifq_qdrops++;
227 ml_enqueue(&free_ml, m);
228 } else
229 ifq->ifq_len++;
230 }
231 mtx_leave(&ifq->ifq_mtx);
232
233 oldops->ifqop_free(ifq->ifq_idx, oldq);
234
235 ml_purge(&free_ml);
236 }
Line 214 calls hfsc_purge, on line 231 we release the IPL_NET mutex
protecting us from interrupts and then finally on line 233 we call
hfsc_free that does timeout_del. This opens a window of opportunity
after we release the mutex for a networking interrupt to fire and
call the softclock softintr before returning control to ifq_attach.
And visa@ has pointed out another potential race after ifqop_alloc
(aka hfsc_alloc that does timeout_add) and line 213 where we grab
the IPL_NET mutex to set ifq_ops on line 220. Since NET_LOCK doesn't
have an interrupt protection from softclock softintrs anymore an
IPL_NET interrupt can fire and if lucky trigger the hfsc_deferred
before we set ifq_ops. To avoid this race I'm proposing to move the
timeout_add from hfsc_alloc to hfsc_enq_begin, i.e. add the timeout
when there are packets to deal with.
> IFQ has a mechanism to lock the underlying object and I believe this
> is the right tool for this job. Any other ideas?
>
> I don't think it's a good idea to hold the mutex (ifq_q_enter and
> ifq_q_leave effectively lock and unlock it) during the ifq_start,
> so we have to make a concession and run the ifq_start before knowing
> whether or not HFSC is attached. IMO, it's a small price to pay to
> avoide clutter. Kernel lock assertion is pointless at this point.
>
> OK?
>
A new diff combining all three modifications:
diff --git sys/net/hfsc.c sys/net/hfsc.c
index 410bea733c6..b81afd43531 100644
--- sys/net/hfsc.c
+++ sys/net/hfsc.c
@@ -584,14 +584,13 @@ hfsc_idx(unsigned int nqueues, const struct mbuf *m)
void *
hfsc_alloc(unsigned int idx, void *q)
{
struct hfsc_if *hif = q;
+
KASSERT(idx == 0); /* when hfsc is enabled we only use the first ifq */
KASSERT(hif != NULL);
-
- timeout_add(&hif->hif_defer, 1);
return (hif);
}
void
hfsc_free(unsigned int idx, void *q)
@@ -825,12 +824,15 @@ hfsc_enq(struct ifqueue *ifq, struct mbuf *m)
}
dm = hfsc_class_enqueue(cl, m);
/* successfully queued. */
- if (dm != m && hfsc_class_qlength(cl) == 1)
+ if (dm != m && hfsc_class_qlength(cl) == 1) {
hfsc_set_active(hif, cl, m->m_pkthdr.len);
+ if (!timeout_pending(&hif->hif_defer))
+ timeout_add(&hif->hif_defer, 1);
+ }
/* drop occurred. */
if (dm != NULL)
PKTCNTR_INC(&cl->cl_stats.drop_cnt, dm->m_pkthdr.len);
@@ -944,20 +946,22 @@ hfsc_deferred(void *arg)
{
struct ifnet *ifp = arg;
struct ifqueue *ifq = &ifp->if_snd;
struct hfsc_if *hif;
- KERNEL_ASSERT_LOCKED();
- KASSERT(HFSC_ENABLED(ifq));
+ if (!HFSC_ENABLED(ifq))
+ return;
if (!ifq_empty(ifq))
ifq_start(ifq);
- hif = ifq->ifq_q;
-
+ hif = ifq_q_enter(&ifp->if_snd, ifq_hfsc_ops);
+ if (hif == NULL)
+ return;
/* XXX HRTIMER nearest virtual/fit time is likely less than 1/HZ. */
timeout_add(&hif->hif_defer, 1);
+ ifq_q_leave(&ifp->if_snd, hif);
}
void
hfsc_cl_purge(struct hfsc_if *hif, struct hfsc_class *cl, struct mbuf_list *ml)
{