> On Nov 6, 2020, at 2:07 PM, Martin KaFai Lau <ka...@fb.com> wrote:
> 
> sk_storage_charge() is the only user of omem_charge().
> This patch simplifies it by folding omem_charge() into
> sk_storage_charge().
> 
> Signed-off-by: Martin KaFai Lau <ka...@fb.com>

Acked-by: Song Liu <songliubrav...@fb.com>

> ---
> net/core/bpf_sk_storage.c | 23 ++++++++++-------------
> 1 file changed, 10 insertions(+), 13 deletions(-)
> 
> diff --git a/net/core/bpf_sk_storage.c b/net/core/bpf_sk_storage.c
> index c907f0dc7f87..001eac65e40f 100644
> --- a/net/core/bpf_sk_storage.c
> +++ b/net/core/bpf_sk_storage.c
> @@ -15,18 +15,6 @@
> 
> DEFINE_BPF_STORAGE_CACHE(sk_cache);
> 
> -static int omem_charge(struct sock *sk, unsigned int size)
> -{
> -     /* same check as in sock_kmalloc() */
> -     if (size <= sysctl_optmem_max &&
> -         atomic_read(&sk->sk_omem_alloc) + size < sysctl_optmem_max) {
> -             atomic_add(size, &sk->sk_omem_alloc);
> -             return 0;
> -     }
> -
> -     return -ENOMEM;
> -}
> -
> static struct bpf_local_storage_data *
> sk_storage_lookup(struct sock *sk, struct bpf_map *map, bool cacheit_lockit)
> {
> @@ -316,7 +304,16 @@ BPF_CALL_2(bpf_sk_storage_delete, struct bpf_map *, map, 
> struct sock *, sk)
> static int sk_storage_charge(struct bpf_local_storage_map *smap,
>                            void *owner, u32 size)
> {
> -     return omem_charge(owner, size);
> +     struct sock *sk = (struct sock *)owner;
> +
> +     /* same check as in sock_kmalloc() */
> +     if (size <= sysctl_optmem_max &&
> +         atomic_read(&sk->sk_omem_alloc) + size < sysctl_optmem_max) {
> +             atomic_add(size, &sk->sk_omem_alloc);
> +             return 0;
> +     }
> +
> +     return -ENOMEM;
> }
> 
> static void sk_storage_uncharge(struct bpf_local_storage_map *smap,
> -- 
> 2.24.1
> 

Reply via email to