Hi, Jakub,

On 22.12.2017 00:01, Jakub Kicinski wrote:
> We don't need the RTNL lock for all operations on offload state.
> We only need to hold it around ndo calls.  The device offload
> initialization doesn't require it.  The soon-to-come querying
> of the offload info will only need it partially.  We will also
> be able to remove the waitqueue in following patches.
> 
> Use struct rw_semaphore because map offload will require sleeping
> with the semaphore held for read.
> 
> Suggested-by: Kirill Tkhai <ktk...@virtuozzo.com>
> Signed-off-by: Jakub Kicinski <jakub.kicin...@netronome.com>
> Reviewed-by: Quentin Monnet <quentin.mon...@netronome.com>
> ---
> v2:
>  - use dev_get_by_index_rcu() instead of implicit lock dependencies;
>  - use DECLARE_RWSEM() instead of init_rwsem() (Kirill).
> ---
>  kernel/bpf/offload.c | 33 ++++++++++++++++++++++++++-------
>  1 file changed, 26 insertions(+), 7 deletions(-)
> 
> diff --git a/kernel/bpf/offload.c b/kernel/bpf/offload.c
> index 8455b89d1bbf..f049073a37e6 100644
> --- a/kernel/bpf/offload.c
> +++ b/kernel/bpf/offload.c
> @@ -20,8 +20,12 @@
>  #include <linux/netdevice.h>
>  #include <linux/printk.h>
>  #include <linux/rtnetlink.h>
> +#include <linux/rwsem.h>
>  
> -/* protected by RTNL */
> +/* Protects bpf_prog_offload_devs and offload members of all progs.
> + * RTNL lock cannot be taken when holding this lock.
> + */
> +static DECLARE_RWSEM(bpf_devs_lock);
>  static LIST_HEAD(bpf_prog_offload_devs);
>  
>  int bpf_prog_offload_init(struct bpf_prog *prog, union bpf_attr *attr)
> @@ -43,19 +47,30 @@ int bpf_prog_offload_init(struct bpf_prog *prog, union 
> bpf_attr *attr)
>       offload->prog = prog;
>       init_waitqueue_head(&offload->verifier_done);
>  
> -     rtnl_lock();
> -     offload->netdev = __dev_get_by_index(net, attr->prog_ifindex);
> +     rcu_read_lock();
> +     offload->netdev = dev_get_by_index_rcu(net, attr->prog_ifindex);
>       if (!offload->netdev) {
> -             rtnl_unlock();
> -             kfree(offload);
> -             return -EINVAL;
> +             rcu_read_unlock();
> +             goto err_free;
>       }
> +     dev_hold(offload->netdev);
> +     rcu_read_unlock();

Small remark about this. There is already dev_get_by_index() in net/core/dev.c
with the functionality above. I haven't found in next patches the reason
we have to use directly rcu_read_lock() here.

So, it seems we may replace the above block with simple dev_get_by_index().
Everything else looks good for me.

>  
> +     down_write(&bpf_devs_lock);
> +     if (offload->netdev->reg_state != NETREG_REGISTERED)
> +             goto err_unlock;
>       prog->aux->offload = offload;
>       list_add_tail(&offload->offloads, &bpf_prog_offload_devs);
> -     rtnl_unlock();
> +     dev_put(offload->netdev);
> +     up_write(&bpf_devs_lock);
>  
>       return 0;
> +err_unlock:
> +     up_write(&bpf_devs_lock);
> +     dev_put(offload->netdev);
> +err_free:
> +     kfree(offload);
> +     return -EINVAL;
>  }
>  
>  static int __bpf_offload_ndo(struct bpf_prog *prog, enum bpf_netdev_command 
> cmd,
> @@ -126,7 +141,9 @@ void bpf_prog_offload_destroy(struct bpf_prog *prog)
>       wake_up(&offload->verifier_done);
>  
>       rtnl_lock();
> +     down_write(&bpf_devs_lock);
>       __bpf_prog_offload_destroy(prog);
> +     up_write(&bpf_devs_lock);
>       rtnl_unlock();
>  
>       kfree(offload);
> @@ -181,11 +198,13 @@ static int bpf_offload_notification(struct 
> notifier_block *notifier,
>               if (netdev->reg_state != NETREG_UNREGISTERING)
>                       break;
>  
> +             down_write(&bpf_devs_lock);
>               list_for_each_entry_safe(offload, tmp, &bpf_prog_offload_devs,
>                                        offloads) {
>                       if (offload->netdev == netdev)
>                               __bpf_prog_offload_destroy(offload->prog);
>               }
> +             up_write(&bpf_devs_lock);
>               break;
>       default:
>               break;

Kirill
 

Reply via email to