提交 72659ecc 编写于 作者: O Octavian Purdila 提交者: David S. Miller

tcp: account SYN-ACK timeouts & retransmissions

Currently we don't increment SYN-ACK timeouts & retransmissions
although we do increment the same stats for SYN. We seem to have lost
the SYN-ACK accounting with the introduction of tcp_syn_recv_timer
(commit 2248761e in the netdev-vger-cvs tree).

This patch fixes this issue. In the process we also rename the v4/v6
syn/ack retransmit functions for clarity. We also add a new
request_socket operations (syn_ack_timeout) so we can keep code in
inet_connection_sock.c protocol agnostic.
Signed-off-by: NOctavian Purdila <opurdila@ixiacom.com>
Signed-off-by: NDavid S. Miller <davem@davemloft.net>
上级 0ec00f03
...@@ -45,6 +45,8 @@ struct request_sock_ops { ...@@ -45,6 +45,8 @@ struct request_sock_ops {
void (*send_reset)(struct sock *sk, void (*send_reset)(struct sock *sk,
struct sk_buff *skb); struct sk_buff *skb);
void (*destructor)(struct request_sock *req); void (*destructor)(struct request_sock *req);
void (*syn_ack_timeout)(struct sock *sk,
struct request_sock *req);
}; };
/* struct request_sock - mini sock to represent a connection request /* struct request_sock - mini sock to represent a connection request
......
...@@ -400,6 +400,8 @@ extern int compat_tcp_setsockopt(struct sock *sk, ...@@ -400,6 +400,8 @@ extern int compat_tcp_setsockopt(struct sock *sk,
int level, int optname, int level, int optname,
char __user *optval, unsigned int optlen); char __user *optval, unsigned int optlen);
extern void tcp_set_keepalive(struct sock *sk, int val); extern void tcp_set_keepalive(struct sock *sk, int val);
extern void tcp_syn_ack_timeout(struct sock *sk,
struct request_sock *req);
extern int tcp_recvmsg(struct kiocb *iocb, struct sock *sk, extern int tcp_recvmsg(struct kiocb *iocb, struct sock *sk,
struct msghdr *msg, struct msghdr *msg,
size_t len, int nonblock, size_t len, int nonblock,
......
...@@ -529,6 +529,8 @@ void inet_csk_reqsk_queue_prune(struct sock *parent, ...@@ -529,6 +529,8 @@ void inet_csk_reqsk_queue_prune(struct sock *parent,
syn_ack_recalc(req, thresh, max_retries, syn_ack_recalc(req, thresh, max_retries,
queue->rskq_defer_accept, queue->rskq_defer_accept,
&expire, &resend); &expire, &resend);
if (req->rsk_ops->syn_ack_timeout)
req->rsk_ops->syn_ack_timeout(parent, req);
if (!expire && if (!expire &&
(!resend || (!resend ||
!req->rsk_ops->rtx_syn_ack(parent, req, NULL) || !req->rsk_ops->rtx_syn_ack(parent, req, NULL) ||
......
...@@ -742,9 +742,9 @@ static void tcp_v4_reqsk_send_ack(struct sock *sk, struct sk_buff *skb, ...@@ -742,9 +742,9 @@ static void tcp_v4_reqsk_send_ack(struct sock *sk, struct sk_buff *skb,
* This still operates on a request_sock only, not on a big * This still operates on a request_sock only, not on a big
* socket. * socket.
*/ */
static int __tcp_v4_send_synack(struct sock *sk, struct dst_entry *dst, static int tcp_v4_send_synack(struct sock *sk, struct dst_entry *dst,
struct request_sock *req, struct request_sock *req,
struct request_values *rvp) struct request_values *rvp)
{ {
const struct inet_request_sock *ireq = inet_rsk(req); const struct inet_request_sock *ireq = inet_rsk(req);
int err = -1; int err = -1;
...@@ -775,10 +775,11 @@ static int __tcp_v4_send_synack(struct sock *sk, struct dst_entry *dst, ...@@ -775,10 +775,11 @@ static int __tcp_v4_send_synack(struct sock *sk, struct dst_entry *dst,
return err; return err;
} }
static int tcp_v4_send_synack(struct sock *sk, struct request_sock *req, static int tcp_v4_rtx_synack(struct sock *sk, struct request_sock *req,
struct request_values *rvp) struct request_values *rvp)
{ {
return __tcp_v4_send_synack(sk, NULL, req, rvp); TCP_INC_STATS_BH(sock_net(sk), TCP_MIB_RETRANSSEGS);
return tcp_v4_send_synack(sk, NULL, req, rvp);
} }
/* /*
...@@ -1192,10 +1193,11 @@ static int tcp_v4_inbound_md5_hash(struct sock *sk, struct sk_buff *skb) ...@@ -1192,10 +1193,11 @@ static int tcp_v4_inbound_md5_hash(struct sock *sk, struct sk_buff *skb)
struct request_sock_ops tcp_request_sock_ops __read_mostly = { struct request_sock_ops tcp_request_sock_ops __read_mostly = {
.family = PF_INET, .family = PF_INET,
.obj_size = sizeof(struct tcp_request_sock), .obj_size = sizeof(struct tcp_request_sock),
.rtx_syn_ack = tcp_v4_send_synack, .rtx_syn_ack = tcp_v4_rtx_synack,
.send_ack = tcp_v4_reqsk_send_ack, .send_ack = tcp_v4_reqsk_send_ack,
.destructor = tcp_v4_reqsk_destructor, .destructor = tcp_v4_reqsk_destructor,
.send_reset = tcp_v4_send_reset, .send_reset = tcp_v4_send_reset,
.syn_ack_timeout = tcp_syn_ack_timeout,
}; };
#ifdef CONFIG_TCP_MD5SIG #ifdef CONFIG_TCP_MD5SIG
...@@ -1373,8 +1375,8 @@ int tcp_v4_conn_request(struct sock *sk, struct sk_buff *skb) ...@@ -1373,8 +1375,8 @@ int tcp_v4_conn_request(struct sock *sk, struct sk_buff *skb)
} }
tcp_rsk(req)->snt_isn = isn; tcp_rsk(req)->snt_isn = isn;
if (__tcp_v4_send_synack(sk, dst, req, if (tcp_v4_send_synack(sk, dst, req,
(struct request_values *)&tmp_ext) || (struct request_values *)&tmp_ext) ||
want_cookie) want_cookie)
goto drop_and_free; goto drop_and_free;
......
...@@ -474,6 +474,12 @@ static void tcp_synack_timer(struct sock *sk) ...@@ -474,6 +474,12 @@ static void tcp_synack_timer(struct sock *sk)
TCP_TIMEOUT_INIT, TCP_RTO_MAX); TCP_TIMEOUT_INIT, TCP_RTO_MAX);
} }
void tcp_syn_ack_timeout(struct sock *sk, struct request_sock *req)
{
NET_INC_STATS_BH(sock_net(sk), LINUX_MIB_TCPTIMEOUTS);
}
EXPORT_SYMBOL(tcp_syn_ack_timeout);
void tcp_set_keepalive(struct sock *sk, int val) void tcp_set_keepalive(struct sock *sk, int val)
{ {
if ((1 << sk->sk_state) & (TCPF_CLOSE | TCPF_LISTEN)) if ((1 << sk->sk_state) & (TCPF_CLOSE | TCPF_LISTEN))
......
...@@ -520,6 +520,13 @@ static int tcp_v6_send_synack(struct sock *sk, struct request_sock *req, ...@@ -520,6 +520,13 @@ static int tcp_v6_send_synack(struct sock *sk, struct request_sock *req,
return err; return err;
} }
static int tcp_v6_rtx_synack(struct sock *sk, struct request_sock *req,
struct request_values *rvp)
{
TCP_INC_STATS_BH(sock_net(sk), TCP_MIB_RETRANSSEGS);
return tcp_v6_send_synack(sk, req, rvp);
}
static inline void syn_flood_warning(struct sk_buff *skb) static inline void syn_flood_warning(struct sk_buff *skb)
{ {
#ifdef CONFIG_SYN_COOKIES #ifdef CONFIG_SYN_COOKIES
...@@ -890,10 +897,11 @@ static int tcp_v6_inbound_md5_hash (struct sock *sk, struct sk_buff *skb) ...@@ -890,10 +897,11 @@ static int tcp_v6_inbound_md5_hash (struct sock *sk, struct sk_buff *skb)
struct request_sock_ops tcp6_request_sock_ops __read_mostly = { struct request_sock_ops tcp6_request_sock_ops __read_mostly = {
.family = AF_INET6, .family = AF_INET6,
.obj_size = sizeof(struct tcp6_request_sock), .obj_size = sizeof(struct tcp6_request_sock),
.rtx_syn_ack = tcp_v6_send_synack, .rtx_syn_ack = tcp_v6_rtx_synack,
.send_ack = tcp_v6_reqsk_send_ack, .send_ack = tcp_v6_reqsk_send_ack,
.destructor = tcp_v6_reqsk_destructor, .destructor = tcp_v6_reqsk_destructor,
.send_reset = tcp_v6_send_reset .send_reset = tcp_v6_send_reset,
.syn_ack_timeout = tcp_syn_ack_timeout,
}; };
#ifdef CONFIG_TCP_MD5SIG #ifdef CONFIG_TCP_MD5SIG
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册