|
|
|
@ -114,7 +114,7 @@ Signed-off-by: David S. Miller <davem@davemloft.net>
|
|
|
|
|
if (nval != oval)
|
|
|
|
|
continue;
|
|
|
|
|
|
|
|
|
|
@@ -2149,7 +2149,7 @@ static bool tcp_small_queue_check(struct
|
|
|
|
|
@@ -2150,7 +2150,7 @@ static bool tcp_small_queue_check(struct
|
|
|
|
|
skb->prev == sk->sk_write_queue.next)
|
|
|
|
|
return false;
|
|
|
|
|
|
|
|
|
@ -123,7 +123,7 @@ Signed-off-by: David S. Miller <davem@davemloft.net>
|
|
|
|
|
/* It is possible TX completion already happened
|
|
|
|
|
* before we set TSQ_THROTTLED, so we must
|
|
|
|
|
* test again the condition.
|
|
|
|
|
@@ -2247,8 +2247,8 @@ static bool tcp_write_xmit(struct sock *
|
|
|
|
|
@@ -2248,8 +2248,8 @@ static bool tcp_write_xmit(struct sock *
|
|
|
|
|
unlikely(tso_fragment(sk, skb, limit, mss_now, gfp)))
|
|
|
|
|
break;
|
|
|
|
|
|
|
|
|
@ -134,7 +134,7 @@ Signed-off-by: David S. Miller <davem@davemloft.net>
|
|
|
|
|
if (tcp_small_queue_check(sk, skb, 0))
|
|
|
|
|
break;
|
|
|
|
|
|
|
|
|
|
@@ -3559,8 +3559,6 @@ void __tcp_send_ack(struct sock *sk, u32
|
|
|
|
|
@@ -3560,8 +3560,6 @@ void __tcp_send_ack(struct sock *sk, u32
|
|
|
|
|
/* We do not want pure acks influencing TCP Small Queues or fq/pacing
|
|
|
|
|
* too much.
|
|
|
|
|
* SKB_TRUESIZE(max(1 .. 66, MAX_TCP_HEADER)) is unfortunately ~784
|
|
|
|
|