| Vlad Yasevich | 60c778b | 2008-01-11 09:57:09 -0500 | [diff] [blame] | 1 | /* SCTP kernel implementation | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 2 |  * Copyright (c) 1999-2000 Cisco, Inc. | 
 | 3 |  * Copyright (c) 1999-2001 Motorola, Inc. | 
 | 4 |  * Copyright (c) 2001-2003 International Business Machines Corp. | 
 | 5 |  * Copyright (c) 2001 Intel Corp. | 
 | 6 |  * Copyright (c) 2001 La Monte H.P. Yarroll | 
 | 7 |  * | 
| Vlad Yasevich | 60c778b | 2008-01-11 09:57:09 -0500 | [diff] [blame] | 8 |  * This file is part of the SCTP kernel implementation | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 9 |  * | 
 | 10 |  * This module provides the abstraction for an SCTP tranport representing | 
 | 11 |  * a remote transport address.  For local transport addresses, we just use | 
 | 12 |  * union sctp_addr. | 
 | 13 |  * | 
| Vlad Yasevich | 60c778b | 2008-01-11 09:57:09 -0500 | [diff] [blame] | 14 |  * This SCTP implementation is free software; | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 15 |  * you can redistribute it and/or modify it under the terms of | 
 | 16 |  * the GNU General Public License as published by | 
 | 17 |  * the Free Software Foundation; either version 2, or (at your option) | 
 | 18 |  * any later version. | 
 | 19 |  * | 
| Vlad Yasevich | 60c778b | 2008-01-11 09:57:09 -0500 | [diff] [blame] | 20 |  * This SCTP implementation is distributed in the hope that it | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 21 |  * will be useful, but WITHOUT ANY WARRANTY; without even the implied | 
 | 22 |  *                 ************************ | 
 | 23 |  * warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. | 
 | 24 |  * See the GNU General Public License for more details. | 
 | 25 |  * | 
 | 26 |  * You should have received a copy of the GNU General Public License | 
 | 27 |  * along with GNU CC; see the file COPYING.  If not, write to | 
 | 28 |  * the Free Software Foundation, 59 Temple Place - Suite 330, | 
 | 29 |  * Boston, MA 02111-1307, USA. | 
 | 30 |  * | 
 | 31 |  * Please send any bug reports or fixes you make to the | 
 | 32 |  * email address(es): | 
 | 33 |  *    lksctp developers <lksctp-developers@lists.sourceforge.net> | 
 | 34 |  * | 
 | 35 |  * Or submit a bug report through the following website: | 
 | 36 |  *    http://www.sf.net/projects/lksctp | 
 | 37 |  * | 
 | 38 |  * Written or modified by: | 
 | 39 |  *    La Monte H.P. Yarroll <piggy@acm.org> | 
 | 40 |  *    Karl Knutson          <karl@athena.chicago.il.us> | 
 | 41 |  *    Jon Grimm             <jgrimm@us.ibm.com> | 
 | 42 |  *    Xingang Guo           <xingang.guo@intel.com> | 
 | 43 |  *    Hui Huang             <hui.huang@nokia.com> | 
 | 44 |  *    Sridhar Samudrala	    <sri@us.ibm.com> | 
 | 45 |  *    Ardelle Fan	    <ardelle.fan@intel.com> | 
 | 46 |  * | 
 | 47 |  * Any bugs reported given to us we will try to fix... any fixes shared will | 
 | 48 |  * be incorporated into the next SCTP release. | 
 | 49 |  */ | 
 | 50 |  | 
| Joe Perches | 145ce50 | 2010-08-24 13:21:08 +0000 | [diff] [blame] | 51 | #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt | 
 | 52 |  | 
| Tejun Heo | 5a0e3ad | 2010-03-24 17:04:11 +0900 | [diff] [blame] | 53 | #include <linux/slab.h> | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 54 | #include <linux/types.h> | 
| Sridhar Samudrala | ad8fec1 | 2006-07-21 14:48:50 -0700 | [diff] [blame] | 55 | #include <linux/random.h> | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 56 | #include <net/sctp/sctp.h> | 
 | 57 | #include <net/sctp/sm.h> | 
 | 58 |  | 
 | 59 | /* 1st Level Abstractions.  */ | 
 | 60 |  | 
 | 61 | /* Initialize a new transport from provided memory.  */ | 
| Eric W. Biederman | 89bf345 | 2012-08-07 07:26:14 +0000 | [diff] [blame] | 62 | static struct sctp_transport *sctp_transport_init(struct net *net, | 
 | 63 | 						  struct sctp_transport *peer, | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 64 | 						  const union sctp_addr *addr, | 
| Al Viro | dd0fc66 | 2005-10-07 07:46:04 +0100 | [diff] [blame] | 65 | 						  gfp_t gfp) | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 66 | { | 
 | 67 | 	/* Copy in the address.  */ | 
 | 68 | 	peer->ipaddr = *addr; | 
 | 69 | 	peer->af_specific = sctp_get_af_specific(addr->sa.sa_family); | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 70 | 	memset(&peer->saddr, 0, sizeof(union sctp_addr)); | 
 | 71 |  | 
| Neil Horman | 4244854 | 2012-06-30 03:04:26 +0000 | [diff] [blame] | 72 | 	peer->sack_generation = 0; | 
 | 73 |  | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 74 | 	/* From 6.3.1 RTO Calculation: | 
 | 75 | 	 * | 
 | 76 | 	 * C1) Until an RTT measurement has been made for a packet sent to the | 
 | 77 | 	 * given destination transport address, set RTO to the protocol | 
 | 78 | 	 * parameter 'RTO.Initial'. | 
 | 79 | 	 */ | 
| Eric W. Biederman | e1fc3b1 | 2012-08-07 07:29:57 +0000 | [diff] [blame] | 80 | 	peer->rto = msecs_to_jiffies(net->sctp.rto_initial); | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 81 |  | 
 | 82 | 	peer->last_time_heard = jiffies; | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 83 | 	peer->last_time_ecne_reduced = jiffies; | 
 | 84 |  | 
| Frank Filz | 52ccb8e | 2005-12-22 11:36:46 -0800 | [diff] [blame] | 85 | 	peer->param_flags = SPP_HB_DISABLE | | 
 | 86 | 			    SPP_PMTUD_ENABLE | | 
 | 87 | 			    SPP_SACKDELAY_ENABLE; | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 88 |  | 
 | 89 | 	/* Initialize the default path max_retrans.  */ | 
| Eric W. Biederman | e1fc3b1 | 2012-08-07 07:29:57 +0000 | [diff] [blame] | 90 | 	peer->pathmaxrxt  = net->sctp.max_retrans_path; | 
 | 91 | 	peer->pf_retrans  = net->sctp.pf_retrans; | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 92 |  | 
 | 93 | 	INIT_LIST_HEAD(&peer->transmitted); | 
 | 94 | 	INIT_LIST_HEAD(&peer->send_ready); | 
 | 95 | 	INIT_LIST_HEAD(&peer->transports); | 
 | 96 |  | 
| Pavel Emelyanov | b24b8a2 | 2008-01-23 21:20:07 -0800 | [diff] [blame] | 97 | 	setup_timer(&peer->T3_rtx_timer, sctp_generate_t3_rtx_event, | 
 | 98 | 			(unsigned long)peer); | 
 | 99 | 	setup_timer(&peer->hb_timer, sctp_generate_heartbeat_event, | 
 | 100 | 			(unsigned long)peer); | 
| Vlad Yasevich | 50b5d6a | 2010-05-06 00:56:07 -0700 | [diff] [blame] | 101 | 	setup_timer(&peer->proto_unreach_timer, | 
 | 102 | 		    sctp_generate_proto_unreach_event, (unsigned long)peer); | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 103 |  | 
| Sridhar Samudrala | ad8fec1 | 2006-07-21 14:48:50 -0700 | [diff] [blame] | 104 | 	/* Initialize the 64-bit random nonce sent with heartbeat. */ | 
 | 105 | 	get_random_bytes(&peer->hb_nonce, sizeof(peer->hb_nonce)); | 
 | 106 |  | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 107 | 	atomic_set(&peer->refcnt, 1); | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 108 |  | 
 | 109 | 	return peer; | 
 | 110 | } | 
 | 111 |  | 
 | 112 | /* Allocate and initialize a new transport.  */ | 
| Eric W. Biederman | 89bf345 | 2012-08-07 07:26:14 +0000 | [diff] [blame] | 113 | struct sctp_transport *sctp_transport_new(struct net *net, | 
 | 114 | 					  const union sctp_addr *addr, | 
| Al Viro | dd0fc66 | 2005-10-07 07:46:04 +0100 | [diff] [blame] | 115 | 					  gfp_t gfp) | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 116 | { | 
| YOSHIFUJI Hideaki | d808ad9 | 2007-02-09 23:25:18 +0900 | [diff] [blame] | 117 | 	struct sctp_transport *transport; | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 118 |  | 
| YOSHIFUJI Hideaki | d808ad9 | 2007-02-09 23:25:18 +0900 | [diff] [blame] | 119 | 	transport = t_new(struct sctp_transport, gfp); | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 120 | 	if (!transport) | 
 | 121 | 		goto fail; | 
 | 122 |  | 
| Eric W. Biederman | 89bf345 | 2012-08-07 07:26:14 +0000 | [diff] [blame] | 123 | 	if (!sctp_transport_init(net, transport, addr, gfp)) | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 124 | 		goto fail_init; | 
 | 125 |  | 
 | 126 | 	transport->malloced = 1; | 
 | 127 | 	SCTP_DBG_OBJCNT_INC(transport); | 
 | 128 |  | 
 | 129 | 	return transport; | 
 | 130 |  | 
 | 131 | fail_init: | 
 | 132 | 	kfree(transport); | 
 | 133 |  | 
 | 134 | fail: | 
 | 135 | 	return NULL; | 
 | 136 | } | 
 | 137 |  | 
 | 138 | /* This transport is no longer needed.  Free up if possible, or | 
 | 139 |  * delay until it last reference count. | 
 | 140 |  */ | 
 | 141 | void sctp_transport_free(struct sctp_transport *transport) | 
 | 142 | { | 
 | 143 | 	transport->dead = 1; | 
 | 144 |  | 
 | 145 | 	/* Try to delete the heartbeat timer.  */ | 
 | 146 | 	if (del_timer(&transport->hb_timer)) | 
 | 147 | 		sctp_transport_put(transport); | 
 | 148 |  | 
 | 149 | 	/* Delete the T3_rtx timer if it's active. | 
 | 150 | 	 * There is no point in not doing this now and letting | 
 | 151 | 	 * structure hang around in memory since we know | 
 | 152 | 	 * the tranport is going away. | 
 | 153 | 	 */ | 
| Ying Xue | 25cc4ae | 2013-02-03 20:32:57 +0000 | [diff] [blame] | 154 | 	if (del_timer(&transport->T3_rtx_timer)) | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 155 | 		sctp_transport_put(transport); | 
 | 156 |  | 
| Wei Yongjun | 55fa0cf | 2010-05-09 16:56:07 +0000 | [diff] [blame] | 157 | 	/* Delete the ICMP proto unreachable timer if it's active. */ | 
| Ying Xue | 25cc4ae | 2013-02-03 20:32:57 +0000 | [diff] [blame] | 158 | 	if (del_timer(&transport->proto_unreach_timer)) | 
| Wei Yongjun | 55fa0cf | 2010-05-09 16:56:07 +0000 | [diff] [blame] | 159 | 		sctp_association_put(transport->asoc); | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 160 |  | 
 | 161 | 	sctp_transport_put(transport); | 
 | 162 | } | 
 | 163 |  | 
| Thomas Graf | 45122ca26 | 2012-12-06 09:25:05 +0000 | [diff] [blame] | 164 | static void sctp_transport_destroy_rcu(struct rcu_head *head) | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 165 | { | 
| Thomas Graf | 45122ca26 | 2012-12-06 09:25:05 +0000 | [diff] [blame] | 166 | 	struct sctp_transport *transport; | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 167 |  | 
| Thomas Graf | 45122ca26 | 2012-12-06 09:25:05 +0000 | [diff] [blame] | 168 | 	transport = container_of(head, struct sctp_transport, rcu); | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 169 |  | 
 | 170 | 	dst_release(transport->dst); | 
 | 171 | 	kfree(transport); | 
 | 172 | 	SCTP_DBG_OBJCNT_DEC(transport); | 
 | 173 | } | 
 | 174 |  | 
| Thomas Graf | 45122ca26 | 2012-12-06 09:25:05 +0000 | [diff] [blame] | 175 | /* Destroy the transport data structure. | 
 | 176 |  * Assumes there are no more users of this structure. | 
 | 177 |  */ | 
 | 178 | static void sctp_transport_destroy(struct sctp_transport *transport) | 
 | 179 | { | 
 | 180 | 	SCTP_ASSERT(transport->dead, "Transport is not dead", return); | 
 | 181 |  | 
 | 182 | 	call_rcu(&transport->rcu, sctp_transport_destroy_rcu); | 
| Daniel Borkmann | 8c98653 | 2013-02-01 04:37:43 +0000 | [diff] [blame] | 183 |  | 
 | 184 | 	sctp_packet_free(&transport->packet); | 
 | 185 |  | 
 | 186 | 	if (transport->asoc) | 
 | 187 | 		sctp_association_put(transport->asoc); | 
| Thomas Graf | 45122ca26 | 2012-12-06 09:25:05 +0000 | [diff] [blame] | 188 | } | 
 | 189 |  | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 190 | /* Start T3_rtx timer if it is not already running and update the heartbeat | 
 | 191 |  * timer.  This routine is called every time a DATA chunk is sent. | 
 | 192 |  */ | 
| Vlad Yasevich | d9efc22 | 2010-04-30 22:41:09 -0400 | [diff] [blame] | 193 | void sctp_transport_reset_timers(struct sctp_transport *transport) | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 194 | { | 
 | 195 | 	/* RFC 2960 6.3.2 Retransmission Timer Rules | 
 | 196 | 	 * | 
 | 197 | 	 * R1) Every time a DATA chunk is sent to any address(including a | 
 | 198 | 	 * retransmission), if the T3-rtx timer of that address is not running | 
 | 199 | 	 * start it running so that it will expire after the RTO of that | 
 | 200 | 	 * address. | 
 | 201 | 	 */ | 
 | 202 |  | 
| Vlad Yasevich | d9efc22 | 2010-04-30 22:41:09 -0400 | [diff] [blame] | 203 | 	if (!timer_pending(&transport->T3_rtx_timer)) | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 204 | 		if (!mod_timer(&transport->T3_rtx_timer, | 
 | 205 | 			       jiffies + transport->rto)) | 
 | 206 | 			sctp_transport_hold(transport); | 
 | 207 |  | 
 | 208 | 	/* When a data chunk is sent, reset the heartbeat interval.  */ | 
 | 209 | 	if (!mod_timer(&transport->hb_timer, | 
 | 210 | 		       sctp_transport_timeout(transport))) | 
 | 211 | 	    sctp_transport_hold(transport); | 
 | 212 | } | 
 | 213 |  | 
 | 214 | /* This transport has been assigned to an association. | 
 | 215 |  * Initialize fields from the association or from the sock itself. | 
 | 216 |  * Register the reference count in the association. | 
 | 217 |  */ | 
 | 218 | void sctp_transport_set_owner(struct sctp_transport *transport, | 
 | 219 | 			      struct sctp_association *asoc) | 
 | 220 | { | 
 | 221 | 	transport->asoc = asoc; | 
 | 222 | 	sctp_association_hold(asoc); | 
 | 223 | } | 
 | 224 |  | 
 | 225 | /* Initialize the pmtu of a transport. */ | 
| Vlad Yasevich | 9914ae3 | 2011-04-26 21:51:31 +0000 | [diff] [blame] | 226 | void sctp_transport_pmtu(struct sctp_transport *transport, struct sock *sk) | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 227 | { | 
| Vlad Yasevich | da0420b | 2011-04-26 21:54:17 +0000 | [diff] [blame] | 228 | 	/* If we don't have a fresh route, look one up */ | 
| David S. Miller | f5b0a87 | 2012-07-19 12:31:33 -0700 | [diff] [blame] | 229 | 	if (!transport->dst || transport->dst->obsolete) { | 
| Vlad Yasevich | da0420b | 2011-04-26 21:54:17 +0000 | [diff] [blame] | 230 | 		dst_release(transport->dst); | 
 | 231 | 		transport->af_specific->get_dst(transport, &transport->saddr, | 
| David S. Miller | 8663c93 | 2011-05-06 16:32:47 -0700 | [diff] [blame] | 232 | 						&transport->fl, sk); | 
| Vlad Yasevich | da0420b | 2011-04-26 21:54:17 +0000 | [diff] [blame] | 233 | 	} | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 234 |  | 
| Vlad Yasevich | da0420b | 2011-04-26 21:54:17 +0000 | [diff] [blame] | 235 | 	if (transport->dst) { | 
 | 236 | 		transport->pathmtu = dst_mtu(transport->dst); | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 237 | 	} else | 
| Frank Filz | 52ccb8e | 2005-12-22 11:36:46 -0800 | [diff] [blame] | 238 | 		transport->pathmtu = SCTP_DEFAULT_MAXSEGMENT; | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 239 | } | 
 | 240 |  | 
| David S. Miller | 02f3d4c | 2012-07-16 03:57:14 -0700 | [diff] [blame] | 241 | void sctp_transport_update_pmtu(struct sock *sk, struct sctp_transport *t, u32 pmtu) | 
| Vlad Yasevich | c910b47 | 2007-06-07 13:47:03 -0400 | [diff] [blame] | 242 | { | 
 | 243 | 	struct dst_entry *dst; | 
 | 244 |  | 
 | 245 | 	if (unlikely(pmtu < SCTP_DEFAULT_MINSEGMENT)) { | 
| Joe Perches | 145ce50 | 2010-08-24 13:21:08 +0000 | [diff] [blame] | 246 | 		pr_warn("%s: Reported pmtu %d too low, using default minimum of %d\n", | 
 | 247 | 			__func__, pmtu, | 
 | 248 | 			SCTP_DEFAULT_MINSEGMENT); | 
| Vlad Yasevich | c910b47 | 2007-06-07 13:47:03 -0400 | [diff] [blame] | 249 | 		/* Use default minimum segment size and disable | 
 | 250 | 		 * pmtu discovery on this transport. | 
 | 251 | 		 */ | 
 | 252 | 		t->pathmtu = SCTP_DEFAULT_MINSEGMENT; | 
| Vlad Yasevich | c910b47 | 2007-06-07 13:47:03 -0400 | [diff] [blame] | 253 | 	} else { | 
 | 254 | 		t->pathmtu = pmtu; | 
 | 255 | 	} | 
 | 256 |  | 
 | 257 | 	dst = sctp_transport_dst_check(t); | 
| David S. Miller | 02f3d4c | 2012-07-16 03:57:14 -0700 | [diff] [blame] | 258 | 	if (!dst) | 
 | 259 | 		t->af_specific->get_dst(t, &t->saddr, &t->fl, sk); | 
 | 260 |  | 
 | 261 | 	if (dst) { | 
| David S. Miller | 6700c27 | 2012-07-17 03:29:28 -0700 | [diff] [blame] | 262 | 		dst->ops->update_pmtu(dst, sk, NULL, pmtu); | 
| David S. Miller | 02f3d4c | 2012-07-16 03:57:14 -0700 | [diff] [blame] | 263 |  | 
 | 264 | 		dst = sctp_transport_dst_check(t); | 
 | 265 | 		if (!dst) | 
 | 266 | 			t->af_specific->get_dst(t, &t->saddr, &t->fl, sk); | 
 | 267 | 	} | 
| Vlad Yasevich | c910b47 | 2007-06-07 13:47:03 -0400 | [diff] [blame] | 268 | } | 
 | 269 |  | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 270 | /* Caches the dst entry and source address for a transport's destination | 
 | 271 |  * address. | 
 | 272 |  */ | 
 | 273 | void sctp_transport_route(struct sctp_transport *transport, | 
 | 274 | 			  union sctp_addr *saddr, struct sctp_sock *opt) | 
 | 275 | { | 
 | 276 | 	struct sctp_association *asoc = transport->asoc; | 
 | 277 | 	struct sctp_af *af = transport->af_specific; | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 278 |  | 
| David S. Miller | 8663c93 | 2011-05-06 16:32:47 -0700 | [diff] [blame] | 279 | 	af->get_dst(transport, saddr, &transport->fl, sctp_opt2sk(opt)); | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 280 |  | 
 | 281 | 	if (saddr) | 
 | 282 | 		memcpy(&transport->saddr, saddr, sizeof(union sctp_addr)); | 
 | 283 | 	else | 
| David S. Miller | 8663c93 | 2011-05-06 16:32:47 -0700 | [diff] [blame] | 284 | 		af->get_saddr(opt, transport, &transport->fl); | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 285 |  | 
| Frank Filz | 52ccb8e | 2005-12-22 11:36:46 -0800 | [diff] [blame] | 286 | 	if ((transport->param_flags & SPP_PMTUD_DISABLE) && transport->pathmtu) { | 
 | 287 | 		return; | 
 | 288 | 	} | 
| Vlad Yasevich | da0420b | 2011-04-26 21:54:17 +0000 | [diff] [blame] | 289 | 	if (transport->dst) { | 
 | 290 | 		transport->pathmtu = dst_mtu(transport->dst); | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 291 |  | 
 | 292 | 		/* Initialize sk->sk_rcv_saddr, if the transport is the | 
 | 293 | 		 * association's active path for getsockname(). | 
| YOSHIFUJI Hideaki | d808ad9 | 2007-02-09 23:25:18 +0900 | [diff] [blame] | 294 | 		 */ | 
| Vlad Yasevich | a78102e | 2009-11-11 11:54:37 +0000 | [diff] [blame] | 295 | 		if (asoc && (!asoc->peer.primary_path || | 
 | 296 | 				(transport == asoc->peer.active_path))) | 
| Neil Horman | bf031ff | 2005-12-02 20:32:29 -0800 | [diff] [blame] | 297 | 			opt->pf->af->to_sk_saddr(&transport->saddr, | 
 | 298 | 						 asoc->base.sk); | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 299 | 	} else | 
| Frank Filz | 52ccb8e | 2005-12-22 11:36:46 -0800 | [diff] [blame] | 300 | 		transport->pathmtu = SCTP_DEFAULT_MAXSEGMENT; | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 301 | } | 
 | 302 |  | 
 | 303 | /* Hold a reference to a transport.  */ | 
 | 304 | void sctp_transport_hold(struct sctp_transport *transport) | 
 | 305 | { | 
 | 306 | 	atomic_inc(&transport->refcnt); | 
 | 307 | } | 
 | 308 |  | 
 | 309 | /* Release a reference to a transport and clean up | 
 | 310 |  * if there are no more references. | 
 | 311 |  */ | 
 | 312 | void sctp_transport_put(struct sctp_transport *transport) | 
 | 313 | { | 
 | 314 | 	if (atomic_dec_and_test(&transport->refcnt)) | 
 | 315 | 		sctp_transport_destroy(transport); | 
 | 316 | } | 
 | 317 |  | 
 | 318 | /* Update transport's RTO based on the newly calculated RTT. */ | 
 | 319 | void sctp_transport_update_rto(struct sctp_transport *tp, __u32 rtt) | 
 | 320 | { | 
 | 321 | 	/* Check for valid transport.  */ | 
 | 322 | 	SCTP_ASSERT(tp, "NULL transport", return); | 
 | 323 |  | 
 | 324 | 	/* We should not be doing any RTO updates unless rto_pending is set.  */ | 
 | 325 | 	SCTP_ASSERT(tp->rto_pending, "rto_pending not set", return); | 
 | 326 |  | 
 | 327 | 	if (tp->rttvar || tp->srtt) { | 
| Eric W. Biederman | e1fc3b1 | 2012-08-07 07:29:57 +0000 | [diff] [blame] | 328 | 		struct net *net = sock_net(tp->asoc->base.sk); | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 329 | 		/* 6.3.1 C3) When a new RTT measurement R' is made, set | 
 | 330 | 		 * RTTVAR <- (1 - RTO.Beta) * RTTVAR + RTO.Beta * |SRTT - R'| | 
 | 331 | 		 * SRTT <- (1 - RTO.Alpha) * SRTT + RTO.Alpha * R' | 
 | 332 | 		 */ | 
 | 333 |  | 
 | 334 | 		/* Note:  The above algorithm has been rewritten to | 
 | 335 | 		 * express rto_beta and rto_alpha as inverse powers | 
 | 336 | 		 * of two. | 
 | 337 | 		 * For example, assuming the default value of RTO.Alpha of | 
 | 338 | 		 * 1/8, rto_alpha would be expressed as 3. | 
 | 339 | 		 */ | 
| Eric W. Biederman | e1fc3b1 | 2012-08-07 07:29:57 +0000 | [diff] [blame] | 340 | 		tp->rttvar = tp->rttvar - (tp->rttvar >> net->sctp.rto_beta) | 
| Schoch Christian | 92d64c2 | 2012-11-28 05:18:29 +0000 | [diff] [blame] | 341 | 			+ (((__u32)abs64((__s64)tp->srtt - (__s64)rtt)) >> net->sctp.rto_beta); | 
| Eric W. Biederman | e1fc3b1 | 2012-08-07 07:29:57 +0000 | [diff] [blame] | 342 | 		tp->srtt = tp->srtt - (tp->srtt >> net->sctp.rto_alpha) | 
 | 343 | 			+ (rtt >> net->sctp.rto_alpha); | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 344 | 	} else { | 
 | 345 | 		/* 6.3.1 C2) When the first RTT measurement R is made, set | 
 | 346 | 		 * SRTT <- R, RTTVAR <- R/2. | 
 | 347 | 		 */ | 
 | 348 | 		tp->srtt = rtt; | 
 | 349 | 		tp->rttvar = rtt >> 1; | 
 | 350 | 	} | 
 | 351 |  | 
 | 352 | 	/* 6.3.1 G1) Whenever RTTVAR is computed, if RTTVAR = 0, then | 
 | 353 | 	 * adjust RTTVAR <- G, where G is the CLOCK GRANULARITY. | 
 | 354 | 	 */ | 
 | 355 | 	if (tp->rttvar == 0) | 
 | 356 | 		tp->rttvar = SCTP_CLOCK_GRANULARITY; | 
 | 357 |  | 
 | 358 | 	/* 6.3.1 C3) After the computation, update RTO <- SRTT + 4 * RTTVAR. */ | 
 | 359 | 	tp->rto = tp->srtt + (tp->rttvar << 2); | 
 | 360 |  | 
 | 361 | 	/* 6.3.1 C6) Whenever RTO is computed, if it is less than RTO.Min | 
 | 362 | 	 * seconds then it is rounded up to RTO.Min seconds. | 
 | 363 | 	 */ | 
 | 364 | 	if (tp->rto < tp->asoc->rto_min) | 
 | 365 | 		tp->rto = tp->asoc->rto_min; | 
 | 366 |  | 
 | 367 | 	/* 6.3.1 C7) A maximum value may be placed on RTO provided it is | 
 | 368 | 	 * at least RTO.max seconds. | 
 | 369 | 	 */ | 
 | 370 | 	if (tp->rto > tp->asoc->rto_max) | 
 | 371 | 		tp->rto = tp->asoc->rto_max; | 
 | 372 |  | 
| Michele Baldessari | 196d675 | 2012-12-01 04:49:42 +0000 | [diff] [blame] | 373 | 	sctp_max_rto(tp->asoc, tp); | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 374 | 	tp->rtt = rtt; | 
 | 375 |  | 
 | 376 | 	/* Reset rto_pending so that a new RTT measurement is started when a | 
 | 377 | 	 * new data chunk is sent. | 
 | 378 | 	 */ | 
 | 379 | 	tp->rto_pending = 0; | 
 | 380 |  | 
 | 381 | 	SCTP_DEBUG_PRINTK("%s: transport: %p, rtt: %d, srtt: %d " | 
| Harvey Harrison | 0dc4787 | 2008-03-05 20:47:47 -0800 | [diff] [blame] | 382 | 			  "rttvar: %d, rto: %ld\n", __func__, | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 383 | 			  tp, rtt, tp->srtt, tp->rttvar, tp->rto); | 
 | 384 | } | 
 | 385 |  | 
 | 386 | /* This routine updates the transport's cwnd and partial_bytes_acked | 
 | 387 |  * parameters based on the bytes acked in the received SACK. | 
 | 388 |  */ | 
 | 389 | void sctp_transport_raise_cwnd(struct sctp_transport *transport, | 
 | 390 | 			       __u32 sack_ctsn, __u32 bytes_acked) | 
 | 391 | { | 
| Vlad Yasevich | cf9b481 | 2010-04-30 22:41:10 -0400 | [diff] [blame] | 392 | 	struct sctp_association *asoc = transport->asoc; | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 393 | 	__u32 cwnd, ssthresh, flight_size, pba, pmtu; | 
 | 394 |  | 
 | 395 | 	cwnd = transport->cwnd; | 
 | 396 | 	flight_size = transport->flight_size; | 
 | 397 |  | 
| Vlad Yasevich | a646523 | 2008-06-04 12:38:43 -0700 | [diff] [blame] | 398 | 	/* See if we need to exit Fast Recovery first */ | 
| Vlad Yasevich | cf9b481 | 2010-04-30 22:41:10 -0400 | [diff] [blame] | 399 | 	if (asoc->fast_recovery && | 
 | 400 | 	    TSN_lte(asoc->fast_recovery_exit, sack_ctsn)) | 
 | 401 | 		asoc->fast_recovery = 0; | 
| Vlad Yasevich | a646523 | 2008-06-04 12:38:43 -0700 | [diff] [blame] | 402 |  | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 403 | 	/* The appropriate cwnd increase algorithm is performed if, and only | 
| Vlad Yasevich | a646523 | 2008-06-04 12:38:43 -0700 | [diff] [blame] | 404 | 	 * if the cumulative TSN whould advanced and the congestion window is | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 405 | 	 * being fully utilized. | 
 | 406 | 	 */ | 
| Vlad Yasevich | a646523 | 2008-06-04 12:38:43 -0700 | [diff] [blame] | 407 | 	if (TSN_lte(sack_ctsn, transport->asoc->ctsn_ack_point) || | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 408 | 	    (flight_size < cwnd)) | 
 | 409 | 		return; | 
 | 410 |  | 
 | 411 | 	ssthresh = transport->ssthresh; | 
 | 412 | 	pba = transport->partial_bytes_acked; | 
| Frank Filz | 52ccb8e | 2005-12-22 11:36:46 -0800 | [diff] [blame] | 413 | 	pmtu = transport->asoc->pathmtu; | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 414 |  | 
 | 415 | 	if (cwnd <= ssthresh) { | 
| Vlad Yasevich | a646523 | 2008-06-04 12:38:43 -0700 | [diff] [blame] | 416 | 		/* RFC 4960 7.2.1 | 
 | 417 | 		 * o  When cwnd is less than or equal to ssthresh, an SCTP | 
 | 418 | 		 *    endpoint MUST use the slow-start algorithm to increase | 
 | 419 | 		 *    cwnd only if the current congestion window is being fully | 
 | 420 | 		 *    utilized, an incoming SACK advances the Cumulative TSN | 
 | 421 | 		 *    Ack Point, and the data sender is not in Fast Recovery. | 
 | 422 | 		 *    Only when these three conditions are met can the cwnd be | 
 | 423 | 		 *    increased; otherwise, the cwnd MUST not be increased. | 
 | 424 | 		 *    If these conditions are met, then cwnd MUST be increased | 
 | 425 | 		 *    by, at most, the lesser of 1) the total size of the | 
 | 426 | 		 *    previously outstanding DATA chunk(s) acknowledged, and | 
 | 427 | 		 *    2) the destination's path MTU.  This upper bound protects | 
 | 428 | 		 *    against the ACK-Splitting attack outlined in [SAVAGE99]. | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 429 | 		 */ | 
| Vlad Yasevich | cf9b481 | 2010-04-30 22:41:10 -0400 | [diff] [blame] | 430 | 		if (asoc->fast_recovery) | 
| Vlad Yasevich | a646523 | 2008-06-04 12:38:43 -0700 | [diff] [blame] | 431 | 			return; | 
 | 432 |  | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 433 | 		if (bytes_acked > pmtu) | 
 | 434 | 			cwnd += pmtu; | 
 | 435 | 		else | 
 | 436 | 			cwnd += bytes_acked; | 
 | 437 | 		SCTP_DEBUG_PRINTK("%s: SLOW START: transport: %p, " | 
 | 438 | 				  "bytes_acked: %d, cwnd: %d, ssthresh: %d, " | 
 | 439 | 				  "flight_size: %d, pba: %d\n", | 
| Harvey Harrison | 0dc4787 | 2008-03-05 20:47:47 -0800 | [diff] [blame] | 440 | 				  __func__, | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 441 | 				  transport, bytes_acked, cwnd, | 
 | 442 | 				  ssthresh, flight_size, pba); | 
 | 443 | 	} else { | 
 | 444 | 		/* RFC 2960 7.2.2 Whenever cwnd is greater than ssthresh, | 
 | 445 | 		 * upon each SACK arrival that advances the Cumulative TSN Ack | 
 | 446 | 		 * Point, increase partial_bytes_acked by the total number of | 
 | 447 | 		 * bytes of all new chunks acknowledged in that SACK including | 
 | 448 | 		 * chunks acknowledged by the new Cumulative TSN Ack and by | 
 | 449 | 		 * Gap Ack Blocks. | 
 | 450 | 		 * | 
 | 451 | 		 * When partial_bytes_acked is equal to or greater than cwnd | 
 | 452 | 		 * and before the arrival of the SACK the sender had cwnd or | 
 | 453 | 		 * more bytes of data outstanding (i.e., before arrival of the | 
 | 454 | 		 * SACK, flightsize was greater than or equal to cwnd), | 
 | 455 | 		 * increase cwnd by MTU, and reset partial_bytes_acked to | 
 | 456 | 		 * (partial_bytes_acked - cwnd). | 
 | 457 | 		 */ | 
 | 458 | 		pba += bytes_acked; | 
 | 459 | 		if (pba >= cwnd) { | 
 | 460 | 			cwnd += pmtu; | 
 | 461 | 			pba = ((cwnd < pba) ? (pba - cwnd) : 0); | 
 | 462 | 		} | 
 | 463 | 		SCTP_DEBUG_PRINTK("%s: CONGESTION AVOIDANCE: " | 
 | 464 | 				  "transport: %p, bytes_acked: %d, cwnd: %d, " | 
 | 465 | 				  "ssthresh: %d, flight_size: %d, pba: %d\n", | 
| Harvey Harrison | 0dc4787 | 2008-03-05 20:47:47 -0800 | [diff] [blame] | 466 | 				  __func__, | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 467 | 				  transport, bytes_acked, cwnd, | 
 | 468 | 				  ssthresh, flight_size, pba); | 
 | 469 | 	} | 
 | 470 |  | 
 | 471 | 	transport->cwnd = cwnd; | 
 | 472 | 	transport->partial_bytes_acked = pba; | 
 | 473 | } | 
 | 474 |  | 
 | 475 | /* This routine is used to lower the transport's cwnd when congestion is | 
 | 476 |  * detected. | 
 | 477 |  */ | 
 | 478 | void sctp_transport_lower_cwnd(struct sctp_transport *transport, | 
 | 479 | 			       sctp_lower_cwnd_t reason) | 
 | 480 | { | 
| Vlad Yasevich | cf9b481 | 2010-04-30 22:41:10 -0400 | [diff] [blame] | 481 | 	struct sctp_association *asoc = transport->asoc; | 
 | 482 |  | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 483 | 	switch (reason) { | 
 | 484 | 	case SCTP_LOWER_CWND_T3_RTX: | 
 | 485 | 		/* RFC 2960 Section 7.2.3, sctpimpguide | 
 | 486 | 		 * When the T3-rtx timer expires on an address, SCTP should | 
 | 487 | 		 * perform slow start by: | 
 | 488 | 		 *      ssthresh = max(cwnd/2, 4*MTU) | 
 | 489 | 		 *      cwnd = 1*MTU | 
 | 490 | 		 *      partial_bytes_acked = 0 | 
 | 491 | 		 */ | 
 | 492 | 		transport->ssthresh = max(transport->cwnd/2, | 
| Vlad Yasevich | cf9b481 | 2010-04-30 22:41:10 -0400 | [diff] [blame] | 493 | 					  4*asoc->pathmtu); | 
 | 494 | 		transport->cwnd = asoc->pathmtu; | 
| Vlad Yasevich | 33ce828 | 2009-09-04 18:20:58 -0400 | [diff] [blame] | 495 |  | 
| Vlad Yasevich | cf9b481 | 2010-04-30 22:41:10 -0400 | [diff] [blame] | 496 | 		/* T3-rtx also clears fast recovery */ | 
 | 497 | 		asoc->fast_recovery = 0; | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 498 | 		break; | 
 | 499 |  | 
 | 500 | 	case SCTP_LOWER_CWND_FAST_RTX: | 
 | 501 | 		/* RFC 2960 7.2.4 Adjust the ssthresh and cwnd of the | 
 | 502 | 		 * destination address(es) to which the missing DATA chunks | 
 | 503 | 		 * were last sent, according to the formula described in | 
 | 504 | 		 * Section 7.2.3. | 
| YOSHIFUJI Hideaki | d808ad9 | 2007-02-09 23:25:18 +0900 | [diff] [blame] | 505 | 		 * | 
 | 506 | 		 * RFC 2960 7.2.3, sctpimpguide Upon detection of packet | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 507 | 		 * losses from SACK (see Section 7.2.4), An endpoint | 
 | 508 | 		 * should do the following: | 
 | 509 | 		 *      ssthresh = max(cwnd/2, 4*MTU) | 
 | 510 | 		 *      cwnd = ssthresh | 
 | 511 | 		 *      partial_bytes_acked = 0 | 
 | 512 | 		 */ | 
| Vlad Yasevich | cf9b481 | 2010-04-30 22:41:10 -0400 | [diff] [blame] | 513 | 		if (asoc->fast_recovery) | 
| Vlad Yasevich | a646523 | 2008-06-04 12:38:43 -0700 | [diff] [blame] | 514 | 			return; | 
 | 515 |  | 
 | 516 | 		/* Mark Fast recovery */ | 
| Vlad Yasevich | cf9b481 | 2010-04-30 22:41:10 -0400 | [diff] [blame] | 517 | 		asoc->fast_recovery = 1; | 
 | 518 | 		asoc->fast_recovery_exit = asoc->next_tsn - 1; | 
| Vlad Yasevich | a646523 | 2008-06-04 12:38:43 -0700 | [diff] [blame] | 519 |  | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 520 | 		transport->ssthresh = max(transport->cwnd/2, | 
| Vlad Yasevich | cf9b481 | 2010-04-30 22:41:10 -0400 | [diff] [blame] | 521 | 					  4*asoc->pathmtu); | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 522 | 		transport->cwnd = transport->ssthresh; | 
 | 523 | 		break; | 
 | 524 |  | 
 | 525 | 	case SCTP_LOWER_CWND_ECNE: | 
 | 526 | 		/* RFC 2481 Section 6.1.2. | 
 | 527 | 		 * If the sender receives an ECN-Echo ACK packet | 
 | 528 | 		 * then the sender knows that congestion was encountered in the | 
 | 529 | 		 * network on the path from the sender to the receiver. The | 
 | 530 | 		 * indication of congestion should be treated just as a | 
 | 531 | 		 * congestion loss in non-ECN Capable TCP. That is, the TCP | 
 | 532 | 		 * source halves the congestion window "cwnd" and reduces the | 
 | 533 | 		 * slow start threshold "ssthresh". | 
 | 534 | 		 * A critical condition is that TCP does not react to | 
 | 535 | 		 * congestion indications more than once every window of | 
 | 536 | 		 * data (or more loosely more than once every round-trip time). | 
 | 537 | 		 */ | 
| Wei Yongjun | f61f6f8 | 2009-03-02 09:46:13 +0000 | [diff] [blame] | 538 | 		if (time_after(jiffies, transport->last_time_ecne_reduced + | 
 | 539 | 					transport->rtt)) { | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 540 | 			transport->ssthresh = max(transport->cwnd/2, | 
| Vlad Yasevich | cf9b481 | 2010-04-30 22:41:10 -0400 | [diff] [blame] | 541 | 						  4*asoc->pathmtu); | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 542 | 			transport->cwnd = transport->ssthresh; | 
 | 543 | 			transport->last_time_ecne_reduced = jiffies; | 
 | 544 | 		} | 
 | 545 | 		break; | 
 | 546 |  | 
 | 547 | 	case SCTP_LOWER_CWND_INACTIVE: | 
 | 548 | 		/* RFC 2960 Section 7.2.1, sctpimpguide | 
 | 549 | 		 * When the endpoint does not transmit data on a given | 
 | 550 | 		 * transport address, the cwnd of the transport address | 
 | 551 | 		 * should be adjusted to max(cwnd/2, 4*MTU) per RTO. | 
 | 552 | 		 * NOTE: Although the draft recommends that this check needs | 
 | 553 | 		 * to be done every RTO interval, we do it every hearbeat | 
 | 554 | 		 * interval. | 
 | 555 | 		 */ | 
| Vlad Yasevich | 245cba7 | 2009-11-23 15:53:58 -0500 | [diff] [blame] | 556 | 		transport->cwnd = max(transport->cwnd/2, | 
| Vlad Yasevich | cf9b481 | 2010-04-30 22:41:10 -0400 | [diff] [blame] | 557 | 					 4*asoc->pathmtu); | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 558 | 		break; | 
| Stephen Hemminger | 3ff50b7 | 2007-04-20 17:09:22 -0700 | [diff] [blame] | 559 | 	} | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 560 |  | 
 | 561 | 	transport->partial_bytes_acked = 0; | 
 | 562 | 	SCTP_DEBUG_PRINTK("%s: transport: %p reason: %d cwnd: " | 
| Harvey Harrison | 0dc4787 | 2008-03-05 20:47:47 -0800 | [diff] [blame] | 563 | 			  "%d ssthresh: %d\n", __func__, | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 564 | 			  transport, reason, | 
 | 565 | 			  transport->cwnd, transport->ssthresh); | 
 | 566 | } | 
 | 567 |  | 
| Vlad Yasevich | 46d5a80 | 2009-11-23 15:54:00 -0500 | [diff] [blame] | 568 | /* Apply Max.Burst limit to the congestion window: | 
 | 569 |  * sctpimpguide-05 2.14.2 | 
 | 570 |  * D) When the time comes for the sender to | 
 | 571 |  * transmit new DATA chunks, the protocol parameter Max.Burst MUST | 
 | 572 |  * first be applied to limit how many new DATA chunks may be sent. | 
 | 573 |  * The limit is applied by adjusting cwnd as follows: | 
 | 574 |  * 	if ((flightsize+ Max.Burst * MTU) < cwnd) | 
 | 575 |  * 		cwnd = flightsize + Max.Burst * MTU | 
 | 576 |  */ | 
 | 577 |  | 
 | 578 | void sctp_transport_burst_limited(struct sctp_transport *t) | 
 | 579 | { | 
 | 580 | 	struct sctp_association *asoc = t->asoc; | 
 | 581 | 	u32 old_cwnd = t->cwnd; | 
 | 582 | 	u32 max_burst_bytes; | 
 | 583 |  | 
 | 584 | 	if (t->burst_limited) | 
 | 585 | 		return; | 
 | 586 |  | 
 | 587 | 	max_burst_bytes = t->flight_size + (asoc->max_burst * asoc->pathmtu); | 
 | 588 | 	if (max_burst_bytes < old_cwnd) { | 
 | 589 | 		t->cwnd = max_burst_bytes; | 
 | 590 | 		t->burst_limited = old_cwnd; | 
 | 591 | 	} | 
 | 592 | } | 
 | 593 |  | 
 | 594 | /* Restore the old cwnd congestion window, after the burst had it's | 
 | 595 |  * desired effect. | 
 | 596 |  */ | 
 | 597 | void sctp_transport_burst_reset(struct sctp_transport *t) | 
 | 598 | { | 
 | 599 | 	if (t->burst_limited) { | 
 | 600 | 		t->cwnd = t->burst_limited; | 
 | 601 | 		t->burst_limited = 0; | 
 | 602 | 	} | 
 | 603 | } | 
 | 604 |  | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 605 | /* What is the next timeout value for this transport? */ | 
 | 606 | unsigned long sctp_transport_timeout(struct sctp_transport *t) | 
 | 607 | { | 
 | 608 | 	unsigned long timeout; | 
| Sridhar Samudrala | ad8fec1 | 2006-07-21 14:48:50 -0700 | [diff] [blame] | 609 | 	timeout = t->rto + sctp_jitter(t->rto); | 
| Neil Horman | 5aa93bc | 2012-07-21 07:56:07 +0000 | [diff] [blame] | 610 | 	if ((t->state != SCTP_UNCONFIRMED) && | 
 | 611 | 	    (t->state != SCTP_PF)) | 
| Sridhar Samudrala | ad8fec1 | 2006-07-21 14:48:50 -0700 | [diff] [blame] | 612 | 		timeout += t->hbinterval; | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 613 | 	timeout += jiffies; | 
 | 614 | 	return timeout; | 
 | 615 | } | 
| Vlad Yasevich | 749bf92 | 2007-03-19 17:02:30 -0700 | [diff] [blame] | 616 |  | 
 | 617 | /* Reset transport variables to their initial values */ | 
 | 618 | void sctp_transport_reset(struct sctp_transport *t) | 
 | 619 | { | 
 | 620 | 	struct sctp_association *asoc = t->asoc; | 
 | 621 |  | 
 | 622 | 	/* RFC 2960 (bis), Section 5.2.4 | 
 | 623 | 	 * All the congestion control parameters (e.g., cwnd, ssthresh) | 
 | 624 | 	 * related to this peer MUST be reset to their initial values | 
 | 625 | 	 * (see Section 6.2.1) | 
 | 626 | 	 */ | 
 | 627 | 	t->cwnd = min(4*asoc->pathmtu, max_t(__u32, 2*asoc->pathmtu, 4380)); | 
| Vlad Yasevich | 46d5a80 | 2009-11-23 15:54:00 -0500 | [diff] [blame] | 628 | 	t->burst_limited = 0; | 
| Vlad Yasevich | 289f424 | 2007-03-22 12:26:25 -0700 | [diff] [blame] | 629 | 	t->ssthresh = asoc->peer.i.a_rwnd; | 
| Andrei Pelinescu-Onciul | 5fdd4ba | 2009-11-29 00:14:02 -0800 | [diff] [blame] | 630 | 	t->rto = asoc->rto_initial; | 
| Michele Baldessari | 196d675 | 2012-12-01 04:49:42 +0000 | [diff] [blame] | 631 | 	sctp_max_rto(asoc, t); | 
| Vlad Yasevich | 749bf92 | 2007-03-19 17:02:30 -0700 | [diff] [blame] | 632 | 	t->rtt = 0; | 
 | 633 | 	t->srtt = 0; | 
 | 634 | 	t->rttvar = 0; | 
 | 635 |  | 
 | 636 | 	/* Reset these additional varibles so that we have a clean | 
 | 637 | 	 * slate. | 
 | 638 | 	 */ | 
 | 639 | 	t->partial_bytes_acked = 0; | 
 | 640 | 	t->flight_size = 0; | 
 | 641 | 	t->error_count = 0; | 
 | 642 | 	t->rto_pending = 0; | 
| Vlad Yasevich | faee47c | 2009-02-13 08:33:43 +0000 | [diff] [blame] | 643 | 	t->hb_sent = 0; | 
| Vlad Yasevich | 749bf92 | 2007-03-19 17:02:30 -0700 | [diff] [blame] | 644 |  | 
 | 645 | 	/* Initialize the state information for SFR-CACC */ | 
 | 646 | 	t->cacc.changeover_active = 0; | 
 | 647 | 	t->cacc.cycling_changeover = 0; | 
 | 648 | 	t->cacc.next_tsn_at_change = 0; | 
 | 649 | 	t->cacc.cacc_saw_newack = 0; | 
 | 650 | } | 
| Michio Honda | ddc4bbe | 2011-06-17 11:03:23 +0900 | [diff] [blame] | 651 |  | 
 | 652 | /* Schedule retransmission on the given transport */ | 
 | 653 | void sctp_transport_immediate_rtx(struct sctp_transport *t) | 
 | 654 | { | 
 | 655 | 	/* Stop pending T3_rtx_timer */ | 
| Ying Xue | 25cc4ae | 2013-02-03 20:32:57 +0000 | [diff] [blame] | 656 | 	if (del_timer(&t->T3_rtx_timer)) | 
| Michio Honda | ddc4bbe | 2011-06-17 11:03:23 +0900 | [diff] [blame] | 657 | 		sctp_transport_put(t); | 
| Ying Xue | 25cc4ae | 2013-02-03 20:32:57 +0000 | [diff] [blame] | 658 |  | 
| Michio Honda | ddc4bbe | 2011-06-17 11:03:23 +0900 | [diff] [blame] | 659 | 	sctp_retransmit(&t->asoc->outqueue, t, SCTP_RTXR_T3_RTX); | 
 | 660 | 	if (!timer_pending(&t->T3_rtx_timer)) { | 
 | 661 | 		if (!mod_timer(&t->T3_rtx_timer, jiffies + t->rto)) | 
 | 662 | 			sctp_transport_hold(t); | 
 | 663 | 	} | 
 | 664 | 	return; | 
 | 665 | } |