|
@@ -1288,13 +1288,13 @@ static u8 tcp_sacktag_one(struct sock *sk,
|
|
|
/* Shift newly-SACKed bytes from this skb to the immediately previous
|
|
|
* already-SACKed sk_buff. Mark the newly-SACKed bytes as such.
|
|
|
*/
|
|
|
-static bool tcp_shifted_skb(struct sock *sk, struct sk_buff *skb,
|
|
|
+static bool tcp_shifted_skb(struct sock *sk, struct sk_buff *prev,
|
|
|
+ struct sk_buff *skb,
|
|
|
struct tcp_sacktag_state *state,
|
|
|
unsigned int pcount, int shifted, int mss,
|
|
|
bool dup_sack)
|
|
|
{
|
|
|
struct tcp_sock *tp = tcp_sk(sk);
|
|
|
- struct sk_buff *prev = tcp_write_queue_prev(sk, skb);
|
|
|
u32 start_seq = TCP_SKB_CB(skb)->seq; /* start of newly-SACKed */
|
|
|
u32 end_seq = start_seq + shifted; /* end of newly-SACKed */
|
|
|
|
|
@@ -1495,7 +1495,7 @@ static struct sk_buff *tcp_shift_skb_data(struct sock *sk, struct sk_buff *skb,
|
|
|
|
|
|
if (!skb_shift(prev, skb, len))
|
|
|
goto fallback;
|
|
|
- if (!tcp_shifted_skb(sk, skb, state, pcount, len, mss, dup_sack))
|
|
|
+ if (!tcp_shifted_skb(sk, prev, skb, state, pcount, len, mss, dup_sack))
|
|
|
goto out;
|
|
|
|
|
|
/* Hole filled allows collapsing with the next as well, this is very
|
|
@@ -1514,7 +1514,8 @@ static struct sk_buff *tcp_shift_skb_data(struct sock *sk, struct sk_buff *skb,
|
|
|
len = skb->len;
|
|
|
if (skb_shift(prev, skb, len)) {
|
|
|
pcount += tcp_skb_pcount(skb);
|
|
|
- tcp_shifted_skb(sk, skb, state, tcp_skb_pcount(skb), len, mss, 0);
|
|
|
+ tcp_shifted_skb(sk, prev, skb, state, tcp_skb_pcount(skb),
|
|
|
+ len, mss, 0);
|
|
|
}
|
|
|
|
|
|
out:
|