inet: Fix __inet_inherit_port() to correctly increment bsockets and num_owners
[linux-2.6.git] / net / ipv4 / tcp_timer.c
index 74c54b3..74a6aa0 100644 (file)
@@ -140,10 +140,10 @@ static void tcp_mtu_probing(struct inet_connection_sock *icsk, struct sock *sk)
  */
 static bool retransmits_timed_out(struct sock *sk,
                                  unsigned int boundary,
+                                 unsigned int timeout,
                                  bool syn_set)
 {
-       unsigned int timeout, linear_backoff_thresh;
-       unsigned int start_ts;
+       unsigned int linear_backoff_thresh, start_ts;
        unsigned int rto_base = syn_set ? TCP_TIMEOUT_INIT : TCP_RTO_MIN;
 
        if (!inet_csk(sk)->icsk_retransmits)
@@ -154,14 +154,15 @@ static bool retransmits_timed_out(struct sock *sk,
        else
                start_ts = tcp_sk(sk)->retrans_stamp;
 
-       linear_backoff_thresh = ilog2(TCP_RTO_MAX/rto_base);
-
-       if (boundary <= linear_backoff_thresh)
-               timeout = ((2 << boundary) - 1) * rto_base;
-       else
-               timeout = ((2 << linear_backoff_thresh) - 1) * rto_base +
-                         (boundary - linear_backoff_thresh) * TCP_RTO_MAX;
+       if (likely(timeout == 0)) {
+               linear_backoff_thresh = ilog2(TCP_RTO_MAX/rto_base);
 
+               if (boundary <= linear_backoff_thresh)
+                       timeout = ((2 << boundary) - 1) * rto_base;
+               else
+                       timeout = ((2 << linear_backoff_thresh) - 1) * rto_base +
+                               (boundary - linear_backoff_thresh) * TCP_RTO_MAX;
+       }
        return (tcp_time_stamp - start_ts) >= timeout;
 }
 
@@ -178,7 +179,7 @@ static int tcp_write_timeout(struct sock *sk)
                retry_until = icsk->icsk_syn_retries ? : sysctl_tcp_syn_retries;
                syn_set = 1;
        } else {
-               if (retransmits_timed_out(sk, sysctl_tcp_retries1, 0)) {
+               if (retransmits_timed_out(sk, sysctl_tcp_retries1, 0, 0)) {
                        /* Black hole detection */
                        tcp_mtu_probing(icsk, sk);
 
@@ -191,14 +192,15 @@ static int tcp_write_timeout(struct sock *sk)
 
                        retry_until = tcp_orphan_retries(sk, alive);
                        do_reset = alive ||
-                                  !retransmits_timed_out(sk, retry_until, 0);
+                               !retransmits_timed_out(sk, retry_until, 0, 0);
 
                        if (tcp_out_of_resources(sk, do_reset))
                                return 1;
                }
        }
 
-       if (retransmits_timed_out(sk, retry_until, syn_set)) {
+       if (retransmits_timed_out(sk, retry_until,
+                                 syn_set ? 0 : icsk->icsk_user_timeout, syn_set)) {
                /* Has it gone just too far? */
                tcp_write_err(sk);
                return 1;
@@ -365,18 +367,19 @@ void tcp_retransmit_timer(struct sock *sk)
        if (icsk->icsk_retransmits == 0) {
                int mib_idx;
 
-               if (icsk->icsk_ca_state == TCP_CA_Disorder) {
-                       if (tcp_is_sack(tp))
-                               mib_idx = LINUX_MIB_TCPSACKFAILURES;
-                       else
-                               mib_idx = LINUX_MIB_TCPRENOFAILURES;
-               } else if (icsk->icsk_ca_state == TCP_CA_Recovery) {
+               if (icsk->icsk_ca_state == TCP_CA_Recovery) {
                        if (tcp_is_sack(tp))
                                mib_idx = LINUX_MIB_TCPSACKRECOVERYFAIL;
                        else
                                mib_idx = LINUX_MIB_TCPRENORECOVERYFAIL;
                } else if (icsk->icsk_ca_state == TCP_CA_Loss) {
                        mib_idx = LINUX_MIB_TCPLOSSFAILURES;
+               } else if ((icsk->icsk_ca_state == TCP_CA_Disorder) ||
+                          tp->sacked_out) {
+                       if (tcp_is_sack(tp))
+                               mib_idx = LINUX_MIB_TCPSACKFAILURES;
+                       else
+                               mib_idx = LINUX_MIB_TCPRENOFAILURES;
                } else {
                        mib_idx = LINUX_MIB_TCPTIMEOUTS;
                }
@@ -440,7 +443,7 @@ out_reset_timer:
                icsk->icsk_rto = min(icsk->icsk_rto << 1, TCP_RTO_MAX);
        }
        inet_csk_reset_xmit_timer(sk, ICSK_TIME_RETRANS, icsk->icsk_rto, TCP_RTO_MAX);
-       if (retransmits_timed_out(sk, sysctl_tcp_retries1 + 1, 0))
+       if (retransmits_timed_out(sk, sysctl_tcp_retries1 + 1, 0, 0))
                __sk_dst_reset(sk);
 
 out:;
@@ -560,7 +563,14 @@ static void tcp_keepalive_timer (unsigned long data)
        elapsed = keepalive_time_elapsed(tp);
 
        if (elapsed >= keepalive_time_when(tp)) {
-               if (icsk->icsk_probes_out >= keepalive_probes(tp)) {
+               /* If the TCP_USER_TIMEOUT option is enabled, use that
+                * to determine when to timeout instead.
+                */
+               if ((icsk->icsk_user_timeout != 0 &&
+                   elapsed >= icsk->icsk_user_timeout &&
+                   icsk->icsk_probes_out > 0) ||
+                   (icsk->icsk_user_timeout == 0 &&
+                   icsk->icsk_probes_out >= keepalive_probes(tp))) {
                        tcp_send_active_reset(sk, GFP_ATOMIC);
                        tcp_write_err(sk);
                        goto out;