diff options
author | Kevin Darbyshire-Bryant <ldir@darbyshire-bryant.me.uk> | 2017-10-18 13:19:10 +0100 |
---|---|---|
committer | Stijn Tintel <stijn@linux-ipv6.be> | 2017-10-18 19:44:09 +0300 |
commit | 886d66abcd92a6f2001865d8319504d72cd0d53e (patch) | |
tree | 65414d052b80dc1d631eff44f48544a0dc4fec00 /target/linux/generic/backport-4.9 | |
parent | 060e37567e90665d2cb26b9eaa5384ec60b19478 (diff) | |
download | upstream-886d66abcd92a6f2001865d8319504d72cd0d53e.tar.gz upstream-886d66abcd92a6f2001865d8319504d72cd0d53e.tar.bz2 upstream-886d66abcd92a6f2001865d8319504d72cd0d53e.zip |
kernel: bump 4.9 to 4.9.57
Refresh patches.
Compile-tested for ar71xx - Archer C7 v2
Runtime-tested on ar71xx - Archer C7 v2
Fixes the following CVEs:
- CVE-2017-7518
- CVE-2017-0786
- CVE-2017-1000255
- CVE-2017-12188
- CVE-2017-15265
Signed-off-by: Kevin Darbyshire-Bryant <ldir@darbyshire-bryant.me.uk>
Diffstat (limited to 'target/linux/generic/backport-4.9')
4 files changed, 6 insertions, 6 deletions
diff --git a/target/linux/generic/backport-4.9/024-3-tcp-tsq-add-shortcut-in-tcp_tasklet_func.patch b/target/linux/generic/backport-4.9/024-3-tcp-tsq-add-shortcut-in-tcp_tasklet_func.patch index 8f45c23e3f..d3db74252a 100644 --- a/target/linux/generic/backport-4.9/024-3-tcp-tsq-add-shortcut-in-tcp_tasklet_func.patch +++ b/target/linux/generic/backport-4.9/024-3-tcp-tsq-add-shortcut-in-tcp_tasklet_func.patch @@ -60,7 +60,7 @@ Signed-off-by: David S. Miller <davem@davemloft.net> nval = cmpxchg(&tp->tsq_flags, oval, nval); if (nval != oval) continue; -@@ -2179,6 +2179,8 @@ static bool tcp_write_xmit(struct sock * +@@ -2182,6 +2182,8 @@ static bool tcp_write_xmit(struct sock * unlikely(tso_fragment(sk, skb, limit, mss_now, gfp))) break; diff --git a/target/linux/generic/backport-4.9/024-5-tcp-tsq-add-a-shortcut-in-tcp_small_queue_check.patch b/target/linux/generic/backport-4.9/024-5-tcp-tsq-add-a-shortcut-in-tcp_small_queue_check.patch index dd1a37eb21..925d2dcead 100644 --- a/target/linux/generic/backport-4.9/024-5-tcp-tsq-add-a-shortcut-in-tcp_small_queue_check.patch +++ b/target/linux/generic/backport-4.9/024-5-tcp-tsq-add-a-shortcut-in-tcp_small_queue_check.patch @@ -19,7 +19,7 @@ Signed-off-by: David S. Miller <davem@davemloft.net> --- a/net/ipv4/tcp_output.c +++ b/net/ipv4/tcp_output.c -@@ -2084,6 +2084,15 @@ static bool tcp_small_queue_check(struct +@@ -2087,6 +2087,15 @@ static bool tcp_small_queue_check(struct limit <<= factor; if (atomic_read(&sk->sk_wmem_alloc) > limit) { diff --git a/target/linux/generic/backport-4.9/024-6-tcp-tcp_mtu_probe-is-likely-to-exit-early.patch b/target/linux/generic/backport-4.9/024-6-tcp-tcp_mtu_probe-is-likely-to-exit-early.patch index f0ae2a938e..c5bb42d448 100644 --- a/target/linux/generic/backport-4.9/024-6-tcp-tcp_mtu_probe-is-likely-to-exit-early.patch +++ b/target/linux/generic/backport-4.9/024-6-tcp-tcp_mtu_probe-is-likely-to-exit-early.patch @@ -17,7 +17,7 @@ Signed-off-by: David S. Miller <davem@davemloft.net> --- a/net/ipv4/tcp_output.c +++ b/net/ipv4/tcp_output.c -@@ -1925,26 +1925,26 @@ static inline void tcp_mtu_check_reprobe +@@ -1928,26 +1928,26 @@ static inline void tcp_mtu_check_reprobe */ static int tcp_mtu_probe(struct sock *sk) { diff --git a/target/linux/generic/backport-4.9/024-8-tcp-tsq-move-tsq_flags-close-to-sk_wmem_alloc.patch b/target/linux/generic/backport-4.9/024-8-tcp-tsq-move-tsq_flags-close-to-sk_wmem_alloc.patch index 545fe60c1c..6604a20fca 100644 --- a/target/linux/generic/backport-4.9/024-8-tcp-tsq-move-tsq_flags-close-to-sk_wmem_alloc.patch +++ b/target/linux/generic/backport-4.9/024-8-tcp-tsq-move-tsq_flags-close-to-sk_wmem_alloc.patch @@ -114,7 +114,7 @@ Signed-off-by: David S. Miller <davem@davemloft.net> if (nval != oval) continue; -@@ -2093,7 +2093,7 @@ static bool tcp_small_queue_check(struct +@@ -2096,7 +2096,7 @@ static bool tcp_small_queue_check(struct skb->prev == sk->sk_write_queue.next) return false; @@ -123,7 +123,7 @@ Signed-off-by: David S. Miller <davem@davemloft.net> /* It is possible TX completion already happened * before we set TSQ_THROTTLED, so we must * test again the condition. -@@ -2191,8 +2191,8 @@ static bool tcp_write_xmit(struct sock * +@@ -2194,8 +2194,8 @@ static bool tcp_write_xmit(struct sock * unlikely(tso_fragment(sk, skb, limit, mss_now, gfp))) break; @@ -134,7 +134,7 @@ Signed-off-by: David S. Miller <davem@davemloft.net> if (tcp_small_queue_check(sk, skb, 0)) break; -@@ -3495,8 +3495,6 @@ void tcp_send_ack(struct sock *sk) +@@ -3508,8 +3508,6 @@ void tcp_send_ack(struct sock *sk) /* We do not want pure acks influencing TCP Small Queues or fq/pacing * too much. * SKB_TRUESIZE(max(1 .. 66, MAX_TCP_HEADER)) is unfortunately ~784 |