sk_storage_charge() is the only user of omem_charge().
This patch simplifies it by folding omem_charge() into
sk_storage_charge().

Signed-off-by: Martin KaFai Lau <ka...@fb.com>
---
 net/core/bpf_sk_storage.c | 23 ++++++++++-------------
 1 file changed, 10 insertions(+), 13 deletions(-)

diff --git a/net/core/bpf_sk_storage.c b/net/core/bpf_sk_storage.c
index c907f0dc7f87..001eac65e40f 100644
--- a/net/core/bpf_sk_storage.c
+++ b/net/core/bpf_sk_storage.c
@@ -15,18 +15,6 @@
 
 DEFINE_BPF_STORAGE_CACHE(sk_cache);
 
-static int omem_charge(struct sock *sk, unsigned int size)
-{
-       /* same check as in sock_kmalloc() */
-       if (size <= sysctl_optmem_max &&
-           atomic_read(&sk->sk_omem_alloc) + size < sysctl_optmem_max) {
-               atomic_add(size, &sk->sk_omem_alloc);
-               return 0;
-       }
-
-       return -ENOMEM;
-}
-
 static struct bpf_local_storage_data *
 sk_storage_lookup(struct sock *sk, struct bpf_map *map, bool cacheit_lockit)
 {
@@ -316,7 +304,16 @@ BPF_CALL_2(bpf_sk_storage_delete, struct bpf_map *, map, 
struct sock *, sk)
 static int sk_storage_charge(struct bpf_local_storage_map *smap,
                             void *owner, u32 size)
 {
-       return omem_charge(owner, size);
+       struct sock *sk = (struct sock *)owner;
+
+       /* same check as in sock_kmalloc() */
+       if (size <= sysctl_optmem_max &&
+           atomic_read(&sk->sk_omem_alloc) + size < sysctl_optmem_max) {
+               atomic_add(size, &sk->sk_omem_alloc);
+               return 0;
+       }
+
+       return -ENOMEM;
 }
 
 static void sk_storage_uncharge(struct bpf_local_storage_map *smap,
-- 
2.24.1

Reply via email to