On Thu, Sep 04, 2025 at 05:44:53PM +0800, Herbert Xu wrote: > On Mon, Sep 01, 2025 at 08:00:15AM -0700, Paul E. McKenney wrote: > > > > However, another option for the the above rcu_dereference_check() to > > become something like this: > > > > rcu_dereference_check(p, lockdep_rht_mutex_is_held(ht) || > > rcu_read_lock_any_held()); > > > > This would be happy with any RCU reader, including rcu_read_lock(), > > preempt_disable(), local_irq_disable(), local_bh_disable(), and various > > handler contexts. One downside is that this would *always* be happy in > > a kernel built with CONFIG_PREEMPT_{NONE,VOLUNTARY}=y. > > Why is that a problem? Assuming that it is always safe to perform > an rcu_dereference_all() on such a kernel, not warning would appear > to be the right thing to do. > > > If this is happening often enough, it would be easy for me to create an > > rcu_dereference_all_check() that allows all forms of vanilla RCU readers > > (but not, for example, SRCU readers), but with only two use cases, > > it is not clear to me that this is an overall win. > > > > Or am I missing a turn in here somewhere? > > As I said rhashtable is just a middle-man, while it does some > RCU actions internally, in the end the user is free to choose > the RCU variant (vanilla, bh, sched). So I'm faced with the > choice of using only rcu_dereference_all in rhashtable code, > or splitting the rhashtable API into three variants just like > RCU. I'd rather not do this especially because it appears RCU > itself has essentially morphed into a single variant because > call_rcu etc. always waits for all three variants. > > So how about this patch? > > ---8<--- > Add rcu_dereference_all and rcu_dereference_all_check so that > library code such as rhashtable can be used with any RCU variant. > > As it stands rcu_dereference is used within rashtable, which > creates false-positive warnings if the user calls it from another > RCU variant context, such as preempt_disable(). > > Use the rcu_dereference_all and rcu_dereference_all_check calls > in rhashtable to suppress these warnings. > > Also replace the rcu_dereference_raw calls in the list iterators > with rcu_dereference_all to uncover buggy calls. > > Reported-by: Menglong Dong <dong...@chinatelecom.cn> > Signed-off-by: Herbert Xu <herb...@gondor.apana.org.au>
I am guessing that you want to send this up via the rhashtable path. > diff --git a/include/linux/rcupdate.h b/include/linux/rcupdate.h > index 120536f4c6eb..b4261a0498f0 100644 > --- a/include/linux/rcupdate.h > +++ b/include/linux/rcupdate.h > @@ -713,6 +713,23 @@ do { > \ > (c) || rcu_read_lock_sched_held(), \ > __rcu) > > +/** > + * rcu_dereference_all_check() - rcu_dereference_all with debug checking > + * @p: The pointer to read, prior to dereferencing > + * @c: The conditions under which the dereference will take place > + * > + * This is the RCU-all counterpart to rcu_dereference_check(). > + * However, please note that starting in v5.0 kernels, vanilla RCU grace > + * periods wait for preempt_disable() regions of code in addition to > + * regions of code demarked by rcu_read_lock() and rcu_read_unlock(). > + * This means that synchronize_rcu(), call_rcu, and friends all take not > + * only rcu_read_lock() but also rcu_read_lock_*() into account. How about something like this? * This is similar to rcu_dereference_check(), but allows protection * by all forms of vanilla RCU readers, including preemption disabled, * bh-disabled, and interrupt-disabled regions of code. Note that "vanilla * RCU" excludes SRCU and the various Tasks RCU flavors. Please note * that this macro should not be backported to any Linux-kernel version * preceding v5.0 due to changes in synchronize_rcu() semantics prior * to that version. The "should not" vs. "can not" accounts for the possibility of people using synchronize_rcu_mult(), but someone wanting to do that best know what they are doing. ;-) > + */ > +#define rcu_dereference_all_check(p, c) \ > + __rcu_dereference_check((p), __UNIQUE_ID(rcu), \ > + (c) || rcu_read_lock_any_held(), \ > + __rcu) > + > /* > * The tracing infrastructure traces RCU (we want that), but unfortunately > * some of the RCU checks causes tracing to lock up the system. > @@ -767,6 +784,14 @@ do { > \ > */ > #define rcu_dereference_sched(p) rcu_dereference_sched_check(p, 0) > > +/** > + * rcu_dereference_all() - fetch RCU-all-protected pointer for dereferencing > + * @p: The pointer to read, prior to dereferencing > + * > + * Makes rcu_dereference_check() do the dirty work. > + */ > +#define rcu_dereference_all(p) rcu_dereference_all_check(p, 0) > + With some comment-header change similar to the above: Reviewed-by: Paul E. McKenney <paul...@kernel.org> > /** > * rcu_pointer_handoff() - Hand off a pointer from RCU to other mechanism > * @p: The pointer to hand off > diff --git a/include/linux/rhashtable.h b/include/linux/rhashtable.h > index 6c85b28ea30b..e9b5ac26b42d 100644 > --- a/include/linux/rhashtable.h > +++ b/include/linux/rhashtable.h > @@ -272,13 +272,13 @@ struct rhash_lock_head __rcu **rht_bucket_nested_insert( > rcu_dereference_protected(p, lockdep_rht_mutex_is_held(ht)) > > #define rht_dereference_rcu(p, ht) \ > - rcu_dereference_check(p, lockdep_rht_mutex_is_held(ht)) > + rcu_dereference_all_check(p, lockdep_rht_mutex_is_held(ht)) > > #define rht_dereference_bucket(p, tbl, hash) \ > rcu_dereference_protected(p, lockdep_rht_bucket_is_held(tbl, hash)) > > #define rht_dereference_bucket_rcu(p, tbl, hash) \ > - rcu_dereference_check(p, lockdep_rht_bucket_is_held(tbl, hash)) > + rcu_dereference_all_check(p, lockdep_rht_bucket_is_held(tbl, hash)) > > #define rht_entry(tpos, pos, member) \ > ({ tpos = container_of(pos, typeof(*tpos), member); 1; }) > @@ -373,7 +373,7 @@ static inline struct rhash_head *__rht_ptr( > static inline struct rhash_head *rht_ptr_rcu( > struct rhash_lock_head __rcu *const *bkt) > { > - return __rht_ptr(rcu_dereference(*bkt), bkt); > + return __rht_ptr(rcu_dereference_all(*bkt), bkt); > } > > static inline struct rhash_head *rht_ptr( > @@ -497,7 +497,7 @@ static inline void rht_assign_unlock(struct bucket_table > *tbl, > for (({barrier(); }), \ > pos = head; \ > !rht_is_a_nulls(pos); \ > - pos = rcu_dereference_raw(pos->next)) > + pos = rcu_dereference_all(pos->next)) > > /** > * rht_for_each_rcu - iterate over rcu hash chain > @@ -513,7 +513,7 @@ static inline void rht_assign_unlock(struct bucket_table > *tbl, > for (({barrier(); }), \ > pos = rht_ptr_rcu(rht_bucket(tbl, hash)); \ > !rht_is_a_nulls(pos); \ > - pos = rcu_dereference_raw(pos->next)) > + pos = rcu_dereference_all(pos->next)) > > /** > * rht_for_each_entry_rcu_from - iterated over rcu hash chain from given head > @@ -560,7 +560,7 @@ static inline void rht_assign_unlock(struct bucket_table > *tbl, > * list returned by rhltable_lookup. > */ > #define rhl_for_each_rcu(pos, list) \ > - for (pos = list; pos; pos = rcu_dereference_raw(pos->next)) > + for (pos = list; pos; pos = rcu_dereference_all(pos->next)) > > /** > * rhl_for_each_entry_rcu - iterate over rcu hash table list of given type > @@ -574,7 +574,7 @@ static inline void rht_assign_unlock(struct bucket_table > *tbl, > */ > #define rhl_for_each_entry_rcu(tpos, pos, list, member) > \ > for (pos = list; pos && rht_entry(tpos, pos, member); \ > - pos = rcu_dereference_raw(pos->next)) > + pos = rcu_dereference_all(pos->next)) > > static inline int rhashtable_compare(struct rhashtable_compare_arg *arg, > const void *obj) > > Cheers, > -- > Email: Herbert Xu <herb...@gondor.apana.org.au> > Home Page: http://gondor.apana.org.au/~herbert/ > PGP Key: http://gondor.apana.org.au/~herbert/pubkey.txt