Hello Yuchung,

On 18/10/17 - 11:22:51, Yuchung Cheng wrote:
> New socket option TCP_FASTOPEN_KEY to allow different keys per
> listener.  The listener by default uses the global key until the
> socket option is set.  The key is a 16 bytes long binary data. This
> option has no effect on regular non-listener TCP sockets.

can you explain what the use-case is to have per-listener TFO keys?


Thanks,
Christoph

> 
> Signed-off-by: Yuchung Cheng <ych...@google.com>
> Reviewed-by: Eric Dumazet <eduma...@google.com>
> ---
>  include/net/request_sock.h |  2 ++
>  include/net/tcp.h          |  5 +++--
>  include/uapi/linux/tcp.h   |  1 +
>  net/ipv4/sysctl_net_ipv4.c |  3 ++-
>  net/ipv4/tcp.c             | 33 +++++++++++++++++++++++++++
>  net/ipv4/tcp_fastopen.c    | 56 
> +++++++++++++++++++++++++++++++++-------------
>  net/ipv4/tcp_ipv4.c        |  1 +
>  7 files changed, 82 insertions(+), 19 deletions(-)
> 
> diff --git a/include/net/request_sock.h b/include/net/request_sock.h
> index 23e22054aa60..347015515a7d 100644
> --- a/include/net/request_sock.h
> +++ b/include/net/request_sock.h
> @@ -150,6 +150,8 @@ struct fastopen_queue {
>       spinlock_t      lock;
>       int             qlen;           /* # of pending (TCP_SYN_RECV) reqs */
>       int             max_qlen;       /* != 0 iff TFO is currently enabled */
> +
> +     struct tcp_fastopen_context __rcu *ctx; /* cipher context for cookie */
>  };
>  
>  /** struct request_sock_queue - queue of request_socks
> diff --git a/include/net/tcp.h b/include/net/tcp.h
> index 3b3b9b968e2d..1efe8365cb28 100644
> --- a/include/net/tcp.h
> +++ b/include/net/tcp.h
> @@ -1555,9 +1555,10 @@ struct tcp_fastopen_request {
>       int                             copied; /* queued in tcp_connect() */
>  };
>  void tcp_free_fastopen_req(struct tcp_sock *tp);
> -
> +void tcp_fastopen_destroy_cipher(struct sock *sk);
>  void tcp_fastopen_ctx_destroy(struct net *net);
> -int tcp_fastopen_reset_cipher(struct net *net, void *key, unsigned int len);
> +int tcp_fastopen_reset_cipher(struct net *net, struct sock *sk,
> +                           void *key, unsigned int len);
>  void tcp_fastopen_add_skb(struct sock *sk, struct sk_buff *skb);
>  struct sock *tcp_try_fastopen(struct sock *sk, struct sk_buff *skb,
>                             struct request_sock *req,
> diff --git a/include/uapi/linux/tcp.h b/include/uapi/linux/tcp.h
> index 15c25eccab2b..69c7493e42f8 100644
> --- a/include/uapi/linux/tcp.h
> +++ b/include/uapi/linux/tcp.h
> @@ -119,6 +119,7 @@ enum {
>  #define TCP_FASTOPEN_CONNECT 30      /* Attempt FastOpen with connect */
>  #define TCP_ULP                      31      /* Attach a ULP to a TCP 
> connection */
>  #define TCP_MD5SIG_EXT               32      /* TCP MD5 Signature with 
> extensions */
> +#define TCP_FASTOPEN_KEY     33      /* Set the key for Fast Open (cookie) */
>  
>  struct tcp_repair_opt {
>       __u32   opt_code;
> diff --git a/net/ipv4/sysctl_net_ipv4.c b/net/ipv4/sysctl_net_ipv4.c
> index cac8dd309f39..81d218346cf7 100644
> --- a/net/ipv4/sysctl_net_ipv4.c
> +++ b/net/ipv4/sysctl_net_ipv4.c
> @@ -284,7 +284,8 @@ static int proc_tcp_fastopen_key(struct ctl_table *table, 
> int write,
>                       ret = -EINVAL;
>                       goto bad_key;
>               }
> -             tcp_fastopen_reset_cipher(net, user_key, 
> TCP_FASTOPEN_KEY_LENGTH);
> +             tcp_fastopen_reset_cipher(net, NULL, user_key,
> +                                       TCP_FASTOPEN_KEY_LENGTH);
>       }
>  
>  bad_key:
> diff --git a/net/ipv4/tcp.c b/net/ipv4/tcp.c
> index 3b34850d361f..8b1fa4dd4538 100644
> --- a/net/ipv4/tcp.c
> +++ b/net/ipv4/tcp.c
> @@ -2571,6 +2571,17 @@ static int do_tcp_setsockopt(struct sock *sk, int 
> level,
>               release_sock(sk);
>               return err;
>       }
> +     case TCP_FASTOPEN_KEY: {
> +             __u8 key[TCP_FASTOPEN_KEY_LENGTH];
> +
> +             if (optlen != sizeof(key))
> +                     return -EINVAL;
> +
> +             if (copy_from_user(key, optval, optlen))
> +                     return -EFAULT;
> +
> +             return tcp_fastopen_reset_cipher(net, sk, key, sizeof(key));
> +     }
>       default:
>               /* fallthru */
>               break;
> @@ -3157,6 +3168,28 @@ static int do_tcp_getsockopt(struct sock *sk, int 
> level,
>                       return -EFAULT;
>               return 0;
>  
> +     case TCP_FASTOPEN_KEY: {
> +             __u8 key[TCP_FASTOPEN_KEY_LENGTH];
> +             struct tcp_fastopen_context *ctx;
> +
> +             if (get_user(len, optlen))
> +                     return -EFAULT;
> +
> +             rcu_read_lock();
> +             ctx = rcu_dereference(icsk->icsk_accept_queue.fastopenq.ctx);
> +             if (ctx)
> +                     memcpy(key, ctx->key, sizeof(key));
> +             else
> +                     len = 0;
> +             rcu_read_unlock();
> +
> +             len = min_t(unsigned int, len, sizeof(key));
> +             if (put_user(len, optlen))
> +                     return -EFAULT;
> +             if (copy_to_user(optval, key, len))
> +                     return -EFAULT;
> +             return 0;
> +     }
>       case TCP_THIN_LINEAR_TIMEOUTS:
>               val = tp->thin_lto;
>               break;
> diff --git a/net/ipv4/tcp_fastopen.c b/net/ipv4/tcp_fastopen.c
> index 7ee4aadcdd71..21075ce19cb6 100644
> --- a/net/ipv4/tcp_fastopen.c
> +++ b/net/ipv4/tcp_fastopen.c
> @@ -29,7 +29,7 @@ void tcp_fastopen_init_key_once(struct net *net)
>        * for a valid cookie, so this is an acceptable risk.
>        */
>       get_random_bytes(key, sizeof(key));
> -     tcp_fastopen_reset_cipher(net, key, sizeof(key));
> +     tcp_fastopen_reset_cipher(net, NULL, key, sizeof(key));
>  }
>  
>  static void tcp_fastopen_ctx_free(struct rcu_head *head)
> @@ -40,6 +40,16 @@ static void tcp_fastopen_ctx_free(struct rcu_head *head)
>       kfree(ctx);
>  }
>  
> +void tcp_fastopen_destroy_cipher(struct sock *sk)
> +{
> +     struct tcp_fastopen_context *ctx;
> +
> +     ctx = rcu_dereference_protected(
> +                     inet_csk(sk)->icsk_accept_queue.fastopenq.ctx, 1);
> +     if (ctx)
> +             call_rcu(&ctx->rcu, tcp_fastopen_ctx_free);
> +}
> +
>  void tcp_fastopen_ctx_destroy(struct net *net)
>  {
>       struct tcp_fastopen_context *ctxt;
> @@ -55,10 +65,12 @@ void tcp_fastopen_ctx_destroy(struct net *net)
>               call_rcu(&ctxt->rcu, tcp_fastopen_ctx_free);
>  }
>  
> -int tcp_fastopen_reset_cipher(struct net *net, void *key, unsigned int len)
> +int tcp_fastopen_reset_cipher(struct net *net, struct sock *sk,
> +                           void *key, unsigned int len)
>  {
> -     int err;
>       struct tcp_fastopen_context *ctx, *octx;
> +     struct fastopen_queue *q;
> +     int err;
>  
>       ctx = kmalloc(sizeof(*ctx), GFP_KERNEL);
>       if (!ctx)
> @@ -79,27 +91,39 @@ error:            kfree(ctx);
>       }
>       memcpy(ctx->key, key, len);
>  
> -     spin_lock(&net->ipv4.tcp_fastopen_ctx_lock);
>  
> -     octx = rcu_dereference_protected(net->ipv4.tcp_fastopen_ctx,
> -                             
> lockdep_is_held(&net->ipv4.tcp_fastopen_ctx_lock));
> -     rcu_assign_pointer(net->ipv4.tcp_fastopen_ctx, ctx);
> -     spin_unlock(&net->ipv4.tcp_fastopen_ctx_lock);
> +     if (sk) {
> +             q = &inet_csk(sk)->icsk_accept_queue.fastopenq;
> +             spin_lock_bh(&q->lock);
> +             octx = rcu_dereference_protected(q->ctx,
> +                                              lockdep_is_held(&q->lock));
> +             rcu_assign_pointer(q->ctx, ctx);
> +             spin_unlock_bh(&q->lock);
> +     } else {
> +             spin_lock(&net->ipv4.tcp_fastopen_ctx_lock);
> +             octx = rcu_dereference_protected(net->ipv4.tcp_fastopen_ctx,
> +                     lockdep_is_held(&net->ipv4.tcp_fastopen_ctx_lock));
> +             rcu_assign_pointer(net->ipv4.tcp_fastopen_ctx, ctx);
> +             spin_unlock(&net->ipv4.tcp_fastopen_ctx_lock);
> +     }
>  
>       if (octx)
>               call_rcu(&octx->rcu, tcp_fastopen_ctx_free);
>       return err;
>  }
>  
> -static bool __tcp_fastopen_cookie_gen(struct net *net,
> -                                   const void *path,
> +static bool __tcp_fastopen_cookie_gen(struct sock *sk, const void *path,
>                                     struct tcp_fastopen_cookie *foc)
>  {
>       struct tcp_fastopen_context *ctx;
>       bool ok = false;
>  
>       rcu_read_lock();
> -     ctx = rcu_dereference(net->ipv4.tcp_fastopen_ctx);
> +
> +     ctx = rcu_dereference(inet_csk(sk)->icsk_accept_queue.fastopenq.ctx);
> +     if (!ctx)
> +             ctx = rcu_dereference(sock_net(sk)->ipv4.tcp_fastopen_ctx);
> +
>       if (ctx) {
>               crypto_cipher_encrypt_one(ctx->tfm, foc->val, path);
>               foc->len = TCP_FASTOPEN_COOKIE_SIZE;
> @@ -115,7 +139,7 @@ static bool __tcp_fastopen_cookie_gen(struct net *net,
>   *
>   * XXX (TFO) - refactor when TCP_FASTOPEN_COOKIE_SIZE != AES_BLOCK_SIZE.
>   */
> -static bool tcp_fastopen_cookie_gen(struct net *net,
> +static bool tcp_fastopen_cookie_gen(struct sock *sk,
>                                   struct request_sock *req,
>                                   struct sk_buff *syn,
>                                   struct tcp_fastopen_cookie *foc)
> @@ -124,7 +148,7 @@ static bool tcp_fastopen_cookie_gen(struct net *net,
>               const struct iphdr *iph = ip_hdr(syn);
>  
>               __be32 path[4] = { iph->saddr, iph->daddr, 0, 0 };
> -             return __tcp_fastopen_cookie_gen(net, path, foc);
> +             return __tcp_fastopen_cookie_gen(sk, path, foc);
>       }
>  
>  #if IS_ENABLED(CONFIG_IPV6)
> @@ -132,13 +156,13 @@ static bool tcp_fastopen_cookie_gen(struct net *net,
>               const struct ipv6hdr *ip6h = ipv6_hdr(syn);
>               struct tcp_fastopen_cookie tmp;
>  
> -             if (__tcp_fastopen_cookie_gen(net, &ip6h->saddr, &tmp)) {
> +             if (__tcp_fastopen_cookie_gen(sk, &ip6h->saddr, &tmp)) {
>                       struct in6_addr *buf = &tmp.addr;
>                       int i;
>  
>                       for (i = 0; i < 4; i++)
>                               buf->s6_addr32[i] ^= ip6h->daddr.s6_addr32[i];
> -                     return __tcp_fastopen_cookie_gen(net, buf, foc);
> +                     return __tcp_fastopen_cookie_gen(sk, buf, foc);
>               }
>       }
>  #endif
> @@ -313,7 +337,7 @@ struct sock *tcp_try_fastopen(struct sock *sk, struct 
> sk_buff *skb,
>               goto fastopen;
>  
>       if (foc->len >= 0 &&  /* Client presents or requests a cookie */
> -         tcp_fastopen_cookie_gen(sock_net(sk), req, skb, &valid_foc) &&
> +         tcp_fastopen_cookie_gen(sk, req, skb, &valid_foc) &&
>           foc->len == TCP_FASTOPEN_COOKIE_SIZE &&
>           foc->len == valid_foc.len &&
>           !memcmp(foc->val, valid_foc.val, foc->len)) {
> diff --git a/net/ipv4/tcp_ipv4.c b/net/ipv4/tcp_ipv4.c
> index 5418ecf03b78..d80e1313200a 100644
> --- a/net/ipv4/tcp_ipv4.c
> +++ b/net/ipv4/tcp_ipv4.c
> @@ -1892,6 +1892,7 @@ void tcp_v4_destroy_sock(struct sock *sk)
>  
>       /* If socket is aborted during connect operation */
>       tcp_free_fastopen_req(tp);
> +     tcp_fastopen_destroy_cipher(sk);
>       tcp_saved_syn_free(tp);
>  
>       sk_sockets_allocated_dec(sk);
> -- 
> 2.15.0.rc1.287.g2b38de12cc-goog
> 

Reply via email to