Commit 4f661542 authored by Eric Dumazet's avatar Eric Dumazet Committed by David S. Miller

tcp: fix zerocopy and notsent_lowat issues

My recent patch had at least three problems :

1) TX zerocopy wants notification when skb is acknowledged,
   thus we need to call skb_zcopy_clear() if the skb is
   cached into sk->sk_tx_skb_cache

2) Some applications might expect precise EPOLLOUT
   notifications, so we need to update sk->sk_wmem_queued
   and call sk_mem_uncharge() from sk_wmem_free_skb()
   in all cases. The SOCK_QUEUE_SHRUNK flag must also be set.

3) Reuse of saved skb should have used skb_cloned() instead
  of simply checking if the fast clone has been freed.

Fixes: 472c2e07 ("tcp: add one skb cache for tx")
Signed-off-by: default avatarEric Dumazet <edumazet@google.com>
Cc: Willem de Bruijn <willemb@google.com>
Cc: Soheil Hassas Yeganeh <soheil@google.com>
Acked-by: default avatarSoheil Hassas Yeganeh <soheil@google.com>
Tested-by: default avatarHolger Hoffstätte <holger@applied-asynchrony.com>
Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
parent 4d5ec89f
...@@ -1465,13 +1465,14 @@ static inline void sk_mem_uncharge(struct sock *sk, int size) ...@@ -1465,13 +1465,14 @@ static inline void sk_mem_uncharge(struct sock *sk, int size)
static inline void sk_wmem_free_skb(struct sock *sk, struct sk_buff *skb) static inline void sk_wmem_free_skb(struct sock *sk, struct sk_buff *skb)
{ {
sock_set_flag(sk, SOCK_QUEUE_SHRUNK);
sk->sk_wmem_queued -= skb->truesize;
sk_mem_uncharge(sk, skb->truesize);
if (!sk->sk_tx_skb_cache) { if (!sk->sk_tx_skb_cache) {
skb_zcopy_clear(skb, true);
sk->sk_tx_skb_cache = skb; sk->sk_tx_skb_cache = skb;
return; return;
} }
sock_set_flag(sk, SOCK_QUEUE_SHRUNK);
sk->sk_wmem_queued -= skb->truesize;
sk_mem_uncharge(sk, skb->truesize);
__kfree_skb(skb); __kfree_skb(skb);
} }
......
...@@ -865,14 +865,9 @@ struct sk_buff *sk_stream_alloc_skb(struct sock *sk, int size, gfp_t gfp, ...@@ -865,14 +865,9 @@ struct sk_buff *sk_stream_alloc_skb(struct sock *sk, int size, gfp_t gfp,
{ {
struct sk_buff *skb; struct sk_buff *skb;
skb = sk->sk_tx_skb_cache; if (likely(!size)) {
if (skb && !size) { skb = sk->sk_tx_skb_cache;
const struct sk_buff_fclones *fclones; if (skb && !skb_cloned(skb)) {
fclones = container_of(skb, struct sk_buff_fclones, skb1);
if (refcount_read(&fclones->fclone_ref) == 1) {
sk->sk_wmem_queued -= skb->truesize;
sk_mem_uncharge(sk, skb->truesize);
skb->truesize -= skb->data_len; skb->truesize -= skb->data_len;
sk->sk_tx_skb_cache = NULL; sk->sk_tx_skb_cache = NULL;
pskb_trim(skb, 0); pskb_trim(skb, 0);
...@@ -2543,8 +2538,6 @@ void tcp_write_queue_purge(struct sock *sk) ...@@ -2543,8 +2538,6 @@ void tcp_write_queue_purge(struct sock *sk)
tcp_rtx_queue_purge(sk); tcp_rtx_queue_purge(sk);
skb = sk->sk_tx_skb_cache; skb = sk->sk_tx_skb_cache;
if (skb) { if (skb) {
sk->sk_wmem_queued -= skb->truesize;
sk_mem_uncharge(sk, skb->truesize);
__kfree_skb(skb); __kfree_skb(skb);
sk->sk_tx_skb_cache = NULL; sk->sk_tx_skb_cache = NULL;
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment