|
@@ -1137,31 +1137,6 @@ static inline int tcp_full_space(const struct sock *sk)
|
|
|
return tcp_win_from_space(sk->sk_rcvbuf);
|
|
|
}
|
|
|
|
|
|
-static inline void tcp_openreq_init(struct request_sock *req,
|
|
|
- struct tcp_options_received *rx_opt,
|
|
|
- struct sk_buff *skb, struct sock *sk)
|
|
|
-{
|
|
|
- struct inet_request_sock *ireq = inet_rsk(req);
|
|
|
-
|
|
|
- req->rcv_wnd = 0; /* So that tcp_send_synack() knows! */
|
|
|
- req->cookie_ts = 0;
|
|
|
- tcp_rsk(req)->rcv_isn = TCP_SKB_CB(skb)->seq;
|
|
|
- tcp_rsk(req)->rcv_nxt = TCP_SKB_CB(skb)->seq + 1;
|
|
|
- tcp_rsk(req)->snt_synack = tcp_time_stamp;
|
|
|
- tcp_rsk(req)->last_oow_ack_time = 0;
|
|
|
- req->mss = rx_opt->mss_clamp;
|
|
|
- req->ts_recent = rx_opt->saw_tstamp ? rx_opt->rcv_tsval : 0;
|
|
|
- ireq->tstamp_ok = rx_opt->tstamp_ok;
|
|
|
- ireq->sack_ok = rx_opt->sack_ok;
|
|
|
- ireq->snd_wscale = rx_opt->snd_wscale;
|
|
|
- ireq->wscale_ok = rx_opt->wscale_ok;
|
|
|
- ireq->acked = 0;
|
|
|
- ireq->ecn_ok = 0;
|
|
|
- ireq->ir_rmt_port = tcp_hdr(skb)->source;
|
|
|
- ireq->ir_num = ntohs(tcp_hdr(skb)->dest);
|
|
|
- ireq->ir_mark = inet_request_mark(sk, skb);
|
|
|
-}
|
|
|
-
|
|
|
extern void tcp_openreq_init_rwin(struct request_sock *req,
|
|
|
struct sock *sk, struct dst_entry *dst);
|
|
|
|
|
@@ -1241,36 +1216,8 @@ static inline bool tcp_paws_reject(const struct tcp_options_received *rx_opt,
|
|
|
return true;
|
|
|
}
|
|
|
|
|
|
-/* Return true if we're currently rate-limiting out-of-window ACKs and
|
|
|
- * thus shouldn't send a dupack right now. We rate-limit dupacks in
|
|
|
- * response to out-of-window SYNs or ACKs to mitigate ACK loops or DoS
|
|
|
- * attacks that send repeated SYNs or ACKs for the same connection. To
|
|
|
- * do this, we do not send a duplicate SYNACK or ACK if the remote
|
|
|
- * endpoint is sending out-of-window SYNs or pure ACKs at a high rate.
|
|
|
- */
|
|
|
-static inline bool tcp_oow_rate_limited(struct net *net,
|
|
|
- const struct sk_buff *skb,
|
|
|
- int mib_idx, u32 *last_oow_ack_time)
|
|
|
-{
|
|
|
- /* Data packets without SYNs are not likely part of an ACK loop. */
|
|
|
- if ((TCP_SKB_CB(skb)->seq != TCP_SKB_CB(skb)->end_seq) &&
|
|
|
- !tcp_hdr(skb)->syn)
|
|
|
- goto not_rate_limited;
|
|
|
-
|
|
|
- if (*last_oow_ack_time) {
|
|
|
- s32 elapsed = (s32)(tcp_time_stamp - *last_oow_ack_time);
|
|
|
-
|
|
|
- if (0 <= elapsed && elapsed < sysctl_tcp_invalid_ratelimit) {
|
|
|
- NET_INC_STATS_BH(net, mib_idx);
|
|
|
- return true; /* rate-limited: don't send yet! */
|
|
|
- }
|
|
|
- }
|
|
|
-
|
|
|
- *last_oow_ack_time = tcp_time_stamp;
|
|
|
-
|
|
|
-not_rate_limited:
|
|
|
- return false; /* not rate-limited: go ahead, send dupack now! */
|
|
|
-}
|
|
|
+bool tcp_oow_rate_limited(struct net *net, const struct sk_buff *skb,
|
|
|
+ int mib_idx, u32 *last_oow_ack_time);
|
|
|
|
|
|
static inline void tcp_mib_init(struct net *net)
|
|
|
{
|