| xj | b04a402 | 2021-11-25 15:01:52 +0800 | [diff] [blame] | 1 | /* | 
|  | 2 | * net/tipc/link.c: TIPC link code | 
|  | 3 | * | 
|  | 4 | * Copyright (c) 1996-2007, 2012-2016, Ericsson AB | 
|  | 5 | * Copyright (c) 2004-2007, 2010-2013, Wind River Systems | 
|  | 6 | * All rights reserved. | 
|  | 7 | * | 
|  | 8 | * Redistribution and use in source and binary forms, with or without | 
|  | 9 | * modification, are permitted provided that the following conditions are met: | 
|  | 10 | * | 
|  | 11 | * 1. Redistributions of source code must retain the above copyright | 
|  | 12 | *    notice, this list of conditions and the following disclaimer. | 
|  | 13 | * 2. Redistributions in binary form must reproduce the above copyright | 
|  | 14 | *    notice, this list of conditions and the following disclaimer in the | 
|  | 15 | *    documentation and/or other materials provided with the distribution. | 
|  | 16 | * 3. Neither the names of the copyright holders nor the names of its | 
|  | 17 | *    contributors may be used to endorse or promote products derived from | 
|  | 18 | *    this software without specific prior written permission. | 
|  | 19 | * | 
|  | 20 | * Alternatively, this software may be distributed under the terms of the | 
|  | 21 | * GNU General Public License ("GPL") version 2 as published by the Free | 
|  | 22 | * Software Foundation. | 
|  | 23 | * | 
|  | 24 | * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" | 
|  | 25 | * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE | 
|  | 26 | * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE | 
|  | 27 | * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE | 
|  | 28 | * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR | 
|  | 29 | * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF | 
|  | 30 | * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS | 
|  | 31 | * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN | 
|  | 32 | * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) | 
|  | 33 | * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE | 
|  | 34 | * POSSIBILITY OF SUCH DAMAGE. | 
|  | 35 | */ | 
|  | 36 |  | 
|  | 37 | #include "core.h" | 
|  | 38 | #include "subscr.h" | 
|  | 39 | #include "link.h" | 
|  | 40 | #include "bcast.h" | 
|  | 41 | #include "socket.h" | 
|  | 42 | #include "name_distr.h" | 
|  | 43 | #include "discover.h" | 
|  | 44 | #include "netlink.h" | 
|  | 45 | #include "monitor.h" | 
|  | 46 |  | 
|  | 47 | #include <linux/pkt_sched.h> | 
|  | 48 |  | 
|  | 49 | struct tipc_stats { | 
|  | 50 | u32 sent_pkts; | 
|  | 51 | u32 recv_pkts; | 
|  | 52 | u32 sent_states; | 
|  | 53 | u32 recv_states; | 
|  | 54 | u32 sent_probes; | 
|  | 55 | u32 recv_probes; | 
|  | 56 | u32 sent_nacks; | 
|  | 57 | u32 recv_nacks; | 
|  | 58 | u32 sent_acks; | 
|  | 59 | u32 sent_bundled; | 
|  | 60 | u32 sent_bundles; | 
|  | 61 | u32 recv_bundled; | 
|  | 62 | u32 recv_bundles; | 
|  | 63 | u32 retransmitted; | 
|  | 64 | u32 sent_fragmented; | 
|  | 65 | u32 sent_fragments; | 
|  | 66 | u32 recv_fragmented; | 
|  | 67 | u32 recv_fragments; | 
|  | 68 | u32 link_congs;		/* # port sends blocked by congestion */ | 
|  | 69 | u32 deferred_recv; | 
|  | 70 | u32 duplicates; | 
|  | 71 | u32 max_queue_sz;	/* send queue size high water mark */ | 
|  | 72 | u32 accu_queue_sz;	/* used for send queue size profiling */ | 
|  | 73 | u32 queue_sz_counts;	/* used for send queue size profiling */ | 
|  | 74 | u32 msg_length_counts;	/* used for message length profiling */ | 
|  | 75 | u32 msg_lengths_total;	/* used for message length profiling */ | 
|  | 76 | u32 msg_length_profile[7]; /* used for msg. length profiling */ | 
|  | 77 | }; | 
|  | 78 |  | 
|  | 79 | /** | 
|  | 80 | * struct tipc_link - TIPC link data structure | 
|  | 81 | * @addr: network address of link's peer node | 
|  | 82 | * @name: link name character string | 
|  | 83 | * @media_addr: media address to use when sending messages over link | 
|  | 84 | * @timer: link timer | 
|  | 85 | * @net: pointer to namespace struct | 
|  | 86 | * @refcnt: reference counter for permanent references (owner node & timer) | 
|  | 87 | * @peer_session: link session # being used by peer end of link | 
|  | 88 | * @peer_bearer_id: bearer id used by link's peer endpoint | 
|  | 89 | * @bearer_id: local bearer id used by link | 
|  | 90 | * @tolerance: minimum link continuity loss needed to reset link [in ms] | 
|  | 91 | * @abort_limit: # of unacknowledged continuity probes needed to reset link | 
|  | 92 | * @state: current state of link FSM | 
|  | 93 | * @peer_caps: bitmap describing capabilities of peer node | 
|  | 94 | * @silent_intv_cnt: # of timer intervals without any reception from peer | 
|  | 95 | * @proto_msg: template for control messages generated by link | 
|  | 96 | * @pmsg: convenience pointer to "proto_msg" field | 
|  | 97 | * @priority: current link priority | 
|  | 98 | * @net_plane: current link network plane ('A' through 'H') | 
|  | 99 | * @mon_state: cookie with information needed by link monitor | 
|  | 100 | * @backlog_limit: backlog queue congestion thresholds (indexed by importance) | 
|  | 101 | * @exp_msg_count: # of tunnelled messages expected during link changeover | 
|  | 102 | * @reset_rcv_checkpt: seq # of last acknowledged message at time of link reset | 
|  | 103 | * @mtu: current maximum packet size for this link | 
|  | 104 | * @advertised_mtu: advertised own mtu when link is being established | 
|  | 105 | * @transmitq: queue for sent, non-acked messages | 
|  | 106 | * @backlogq: queue for messages waiting to be sent | 
|  | 107 | * @snt_nxt: next sequence number to use for outbound messages | 
|  | 108 | * @last_retransmitted: sequence number of most recently retransmitted message | 
|  | 109 | * @stale_cnt: counter for number of identical retransmit attempts | 
|  | 110 | * @stale_limit: time when repeated identical retransmits must force link reset | 
|  | 111 | * @ackers: # of peers that needs to ack each packet before it can be released | 
|  | 112 | * @acked: # last packet acked by a certain peer. Used for broadcast. | 
|  | 113 | * @rcv_nxt: next sequence number to expect for inbound messages | 
|  | 114 | * @deferred_queue: deferred queue saved OOS b'cast message received from node | 
|  | 115 | * @unacked_window: # of inbound messages rx'd without ack'ing back to peer | 
|  | 116 | * @inputq: buffer queue for messages to be delivered upwards | 
|  | 117 | * @namedq: buffer queue for name table messages to be delivered upwards | 
|  | 118 | * @next_out: ptr to first unsent outbound message in queue | 
|  | 119 | * @wakeupq: linked list of wakeup msgs waiting for link congestion to abate | 
|  | 120 | * @long_msg_seq_no: next identifier to use for outbound fragmented messages | 
|  | 121 | * @reasm_buf: head of partially reassembled inbound message fragments | 
|  | 122 | * @bc_rcvr: marks that this is a broadcast receiver link | 
|  | 123 | * @stats: collects statistics regarding link activity | 
|  | 124 | */ | 
|  | 125 | struct tipc_link { | 
|  | 126 | u32 addr; | 
|  | 127 | char name[TIPC_MAX_LINK_NAME]; | 
|  | 128 | struct net *net; | 
|  | 129 |  | 
|  | 130 | /* Management and link supervision data */ | 
|  | 131 | u16 peer_session; | 
|  | 132 | u16 session; | 
|  | 133 | u16 snd_nxt_state; | 
|  | 134 | u16 rcv_nxt_state; | 
|  | 135 | u32 peer_bearer_id; | 
|  | 136 | u32 bearer_id; | 
|  | 137 | u32 tolerance; | 
|  | 138 | u32 abort_limit; | 
|  | 139 | u32 state; | 
|  | 140 | u16 peer_caps; | 
|  | 141 | bool in_session; | 
|  | 142 | bool active; | 
|  | 143 | u32 silent_intv_cnt; | 
|  | 144 | char if_name[TIPC_MAX_IF_NAME]; | 
|  | 145 | u32 priority; | 
|  | 146 | char net_plane; | 
|  | 147 | struct tipc_mon_state mon_state; | 
|  | 148 | u16 rst_cnt; | 
|  | 149 |  | 
|  | 150 | /* Failover/synch */ | 
|  | 151 | u16 drop_point; | 
|  | 152 | struct sk_buff *failover_reasm_skb; | 
|  | 153 |  | 
|  | 154 | /* Max packet negotiation */ | 
|  | 155 | u16 mtu; | 
|  | 156 | u16 advertised_mtu; | 
|  | 157 |  | 
|  | 158 | /* Sending */ | 
|  | 159 | struct sk_buff_head transmq; | 
|  | 160 | struct sk_buff_head backlogq; | 
|  | 161 | struct { | 
|  | 162 | u16 len; | 
|  | 163 | u16 limit; | 
|  | 164 | struct sk_buff *target_bskb; | 
|  | 165 | } backlog[5]; | 
|  | 166 | u16 snd_nxt; | 
|  | 167 | u16 last_retransm; | 
|  | 168 | u16 window; | 
|  | 169 | u16 stale_cnt; | 
|  | 170 | unsigned long stale_limit; | 
|  | 171 |  | 
|  | 172 | /* Reception */ | 
|  | 173 | u16 rcv_nxt; | 
|  | 174 | u32 rcv_unacked; | 
|  | 175 | struct sk_buff_head deferdq; | 
|  | 176 | struct sk_buff_head *inputq; | 
|  | 177 | struct sk_buff_head *namedq; | 
|  | 178 |  | 
|  | 179 | /* Congestion handling */ | 
|  | 180 | struct sk_buff_head wakeupq; | 
|  | 181 |  | 
|  | 182 | /* Fragmentation/reassembly */ | 
|  | 183 | struct sk_buff *reasm_buf; | 
|  | 184 |  | 
|  | 185 | /* Broadcast */ | 
|  | 186 | u16 ackers; | 
|  | 187 | u16 acked; | 
|  | 188 | struct tipc_link *bc_rcvlink; | 
|  | 189 | struct tipc_link *bc_sndlink; | 
|  | 190 | unsigned long prev_retr; | 
|  | 191 | u16 prev_from; | 
|  | 192 | u16 prev_to; | 
|  | 193 | u8 nack_state; | 
|  | 194 | bool bc_peer_is_up; | 
|  | 195 |  | 
|  | 196 | /* Statistics */ | 
|  | 197 | struct tipc_stats stats; | 
|  | 198 | }; | 
|  | 199 |  | 
|  | 200 | /* | 
|  | 201 | * Error message prefixes | 
|  | 202 | */ | 
|  | 203 | static const char *link_co_err = "Link tunneling error, "; | 
|  | 204 | static const char *link_rst_msg = "Resetting link "; | 
|  | 205 |  | 
|  | 206 | /* Send states for broadcast NACKs | 
|  | 207 | */ | 
|  | 208 | enum { | 
|  | 209 | BC_NACK_SND_CONDITIONAL, | 
|  | 210 | BC_NACK_SND_UNCONDITIONAL, | 
|  | 211 | BC_NACK_SND_SUPPRESS, | 
|  | 212 | }; | 
|  | 213 |  | 
|  | 214 | #define TIPC_BC_RETR_LIMIT 10   /* [ms] */ | 
|  | 215 |  | 
|  | 216 | /* | 
|  | 217 | * Interval between NACKs when packets arrive out of order | 
|  | 218 | */ | 
|  | 219 | #define TIPC_NACK_INTV (TIPC_MIN_LINK_WIN * 2) | 
|  | 220 |  | 
|  | 221 | /* Link FSM states: | 
|  | 222 | */ | 
|  | 223 | enum { | 
|  | 224 | LINK_ESTABLISHED     = 0xe, | 
|  | 225 | LINK_ESTABLISHING    = 0xe  << 4, | 
|  | 226 | LINK_RESET           = 0x1  << 8, | 
|  | 227 | LINK_RESETTING       = 0x2  << 12, | 
|  | 228 | LINK_PEER_RESET      = 0xd  << 16, | 
|  | 229 | LINK_FAILINGOVER     = 0xf  << 20, | 
|  | 230 | LINK_SYNCHING        = 0xc  << 24 | 
|  | 231 | }; | 
|  | 232 |  | 
|  | 233 | /* Link FSM state checking routines | 
|  | 234 | */ | 
|  | 235 | static int link_is_up(struct tipc_link *l) | 
|  | 236 | { | 
|  | 237 | return l->state & (LINK_ESTABLISHED | LINK_SYNCHING); | 
|  | 238 | } | 
|  | 239 |  | 
|  | 240 | static int tipc_link_proto_rcv(struct tipc_link *l, struct sk_buff *skb, | 
|  | 241 | struct sk_buff_head *xmitq); | 
|  | 242 | static void tipc_link_build_proto_msg(struct tipc_link *l, int mtyp, bool probe, | 
|  | 243 | bool probe_reply, u16 rcvgap, | 
|  | 244 | int tolerance, int priority, | 
|  | 245 | struct sk_buff_head *xmitq); | 
|  | 246 | static void link_print(struct tipc_link *l, const char *str); | 
|  | 247 | static int tipc_link_build_nack_msg(struct tipc_link *l, | 
|  | 248 | struct sk_buff_head *xmitq); | 
|  | 249 | static void tipc_link_build_bc_init_msg(struct tipc_link *l, | 
|  | 250 | struct sk_buff_head *xmitq); | 
|  | 251 | static bool tipc_link_release_pkts(struct tipc_link *l, u16 to); | 
|  | 252 |  | 
|  | 253 | /* | 
|  | 254 | *  Simple non-static link routines (i.e. referenced outside this file) | 
|  | 255 | */ | 
|  | 256 | bool tipc_link_is_up(struct tipc_link *l) | 
|  | 257 | { | 
|  | 258 | return link_is_up(l); | 
|  | 259 | } | 
|  | 260 |  | 
|  | 261 | bool tipc_link_peer_is_down(struct tipc_link *l) | 
|  | 262 | { | 
|  | 263 | return l->state == LINK_PEER_RESET; | 
|  | 264 | } | 
|  | 265 |  | 
|  | 266 | bool tipc_link_is_reset(struct tipc_link *l) | 
|  | 267 | { | 
|  | 268 | return l->state & (LINK_RESET | LINK_FAILINGOVER | LINK_ESTABLISHING); | 
|  | 269 | } | 
|  | 270 |  | 
|  | 271 | bool tipc_link_is_establishing(struct tipc_link *l) | 
|  | 272 | { | 
|  | 273 | return l->state == LINK_ESTABLISHING; | 
|  | 274 | } | 
|  | 275 |  | 
|  | 276 | bool tipc_link_is_synching(struct tipc_link *l) | 
|  | 277 | { | 
|  | 278 | return l->state == LINK_SYNCHING; | 
|  | 279 | } | 
|  | 280 |  | 
|  | 281 | bool tipc_link_is_failingover(struct tipc_link *l) | 
|  | 282 | { | 
|  | 283 | return l->state == LINK_FAILINGOVER; | 
|  | 284 | } | 
|  | 285 |  | 
|  | 286 | bool tipc_link_is_blocked(struct tipc_link *l) | 
|  | 287 | { | 
|  | 288 | return l->state & (LINK_RESETTING | LINK_PEER_RESET | LINK_FAILINGOVER); | 
|  | 289 | } | 
|  | 290 |  | 
|  | 291 | static bool link_is_bc_sndlink(struct tipc_link *l) | 
|  | 292 | { | 
|  | 293 | return !l->bc_sndlink; | 
|  | 294 | } | 
|  | 295 |  | 
|  | 296 | static bool link_is_bc_rcvlink(struct tipc_link *l) | 
|  | 297 | { | 
|  | 298 | return ((l->bc_rcvlink == l) && !link_is_bc_sndlink(l)); | 
|  | 299 | } | 
|  | 300 |  | 
|  | 301 | void tipc_link_set_active(struct tipc_link *l, bool active) | 
|  | 302 | { | 
|  | 303 | l->active = active; | 
|  | 304 | } | 
|  | 305 |  | 
|  | 306 | u32 tipc_link_id(struct tipc_link *l) | 
|  | 307 | { | 
|  | 308 | return l->peer_bearer_id << 16 | l->bearer_id; | 
|  | 309 | } | 
|  | 310 |  | 
|  | 311 | int tipc_link_window(struct tipc_link *l) | 
|  | 312 | { | 
|  | 313 | return l->window; | 
|  | 314 | } | 
|  | 315 |  | 
|  | 316 | int tipc_link_prio(struct tipc_link *l) | 
|  | 317 | { | 
|  | 318 | return l->priority; | 
|  | 319 | } | 
|  | 320 |  | 
|  | 321 | unsigned long tipc_link_tolerance(struct tipc_link *l) | 
|  | 322 | { | 
|  | 323 | return l->tolerance; | 
|  | 324 | } | 
|  | 325 |  | 
|  | 326 | struct sk_buff_head *tipc_link_inputq(struct tipc_link *l) | 
|  | 327 | { | 
|  | 328 | return l->inputq; | 
|  | 329 | } | 
|  | 330 |  | 
|  | 331 | char tipc_link_plane(struct tipc_link *l) | 
|  | 332 | { | 
|  | 333 | return l->net_plane; | 
|  | 334 | } | 
|  | 335 |  | 
|  | 336 | void tipc_link_update_caps(struct tipc_link *l, u16 capabilities) | 
|  | 337 | { | 
|  | 338 | l->peer_caps = capabilities; | 
|  | 339 | } | 
|  | 340 |  | 
|  | 341 | void tipc_link_add_bc_peer(struct tipc_link *snd_l, | 
|  | 342 | struct tipc_link *uc_l, | 
|  | 343 | struct sk_buff_head *xmitq) | 
|  | 344 | { | 
|  | 345 | struct tipc_link *rcv_l = uc_l->bc_rcvlink; | 
|  | 346 |  | 
|  | 347 | snd_l->ackers++; | 
|  | 348 | rcv_l->acked = snd_l->snd_nxt - 1; | 
|  | 349 | snd_l->state = LINK_ESTABLISHED; | 
|  | 350 | tipc_link_build_bc_init_msg(uc_l, xmitq); | 
|  | 351 | } | 
|  | 352 |  | 
|  | 353 | void tipc_link_remove_bc_peer(struct tipc_link *snd_l, | 
|  | 354 | struct tipc_link *rcv_l, | 
|  | 355 | struct sk_buff_head *xmitq) | 
|  | 356 | { | 
|  | 357 | u16 ack = snd_l->snd_nxt - 1; | 
|  | 358 |  | 
|  | 359 | snd_l->ackers--; | 
|  | 360 | rcv_l->bc_peer_is_up = true; | 
|  | 361 | rcv_l->state = LINK_ESTABLISHED; | 
|  | 362 | tipc_link_bc_ack_rcv(rcv_l, ack, xmitq); | 
|  | 363 | tipc_link_reset(rcv_l); | 
|  | 364 | rcv_l->state = LINK_RESET; | 
|  | 365 | if (!snd_l->ackers) { | 
|  | 366 | tipc_link_reset(snd_l); | 
|  | 367 | snd_l->state = LINK_RESET; | 
|  | 368 | __skb_queue_purge(xmitq); | 
|  | 369 | } | 
|  | 370 | } | 
|  | 371 |  | 
|  | 372 | int tipc_link_bc_peers(struct tipc_link *l) | 
|  | 373 | { | 
|  | 374 | return l->ackers; | 
|  | 375 | } | 
|  | 376 |  | 
|  | 377 | static u16 link_bc_rcv_gap(struct tipc_link *l) | 
|  | 378 | { | 
|  | 379 | struct sk_buff *skb = skb_peek(&l->deferdq); | 
|  | 380 | u16 gap = 0; | 
|  | 381 |  | 
|  | 382 | if (more(l->snd_nxt, l->rcv_nxt)) | 
|  | 383 | gap = l->snd_nxt - l->rcv_nxt; | 
|  | 384 | if (skb) | 
|  | 385 | gap = buf_seqno(skb) - l->rcv_nxt; | 
|  | 386 | return gap; | 
|  | 387 | } | 
|  | 388 |  | 
|  | 389 | void tipc_link_set_mtu(struct tipc_link *l, int mtu) | 
|  | 390 | { | 
|  | 391 | l->mtu = mtu; | 
|  | 392 | } | 
|  | 393 |  | 
|  | 394 | int tipc_link_mtu(struct tipc_link *l) | 
|  | 395 | { | 
|  | 396 | return l->mtu; | 
|  | 397 | } | 
|  | 398 |  | 
|  | 399 | u16 tipc_link_rcv_nxt(struct tipc_link *l) | 
|  | 400 | { | 
|  | 401 | return l->rcv_nxt; | 
|  | 402 | } | 
|  | 403 |  | 
|  | 404 | u16 tipc_link_acked(struct tipc_link *l) | 
|  | 405 | { | 
|  | 406 | return l->acked; | 
|  | 407 | } | 
|  | 408 |  | 
|  | 409 | char *tipc_link_name(struct tipc_link *l) | 
|  | 410 | { | 
|  | 411 | return l->name; | 
|  | 412 | } | 
|  | 413 |  | 
|  | 414 | u32 tipc_link_state(struct tipc_link *l) | 
|  | 415 | { | 
|  | 416 | return l->state; | 
|  | 417 | } | 
|  | 418 |  | 
|  | 419 | /** | 
|  | 420 | * tipc_link_create - create a new link | 
|  | 421 | * @n: pointer to associated node | 
|  | 422 | * @if_name: associated interface name | 
|  | 423 | * @bearer_id: id (index) of associated bearer | 
|  | 424 | * @tolerance: link tolerance to be used by link | 
|  | 425 | * @net_plane: network plane (A,B,c..) this link belongs to | 
|  | 426 | * @mtu: mtu to be advertised by link | 
|  | 427 | * @priority: priority to be used by link | 
|  | 428 | * @window: send window to be used by link | 
|  | 429 | * @session: session to be used by link | 
|  | 430 | * @ownnode: identity of own node | 
|  | 431 | * @peer: node id of peer node | 
|  | 432 | * @peer_caps: bitmap describing peer node capabilities | 
|  | 433 | * @bc_sndlink: the namespace global link used for broadcast sending | 
|  | 434 | * @bc_rcvlink: the peer specific link used for broadcast reception | 
|  | 435 | * @inputq: queue to put messages ready for delivery | 
|  | 436 | * @namedq: queue to put binding table update messages ready for delivery | 
|  | 437 | * @link: return value, pointer to put the created link | 
|  | 438 | * | 
|  | 439 | * Returns true if link was created, otherwise false | 
|  | 440 | */ | 
|  | 441 | bool tipc_link_create(struct net *net, char *if_name, int bearer_id, | 
|  | 442 | int tolerance, char net_plane, u32 mtu, int priority, | 
|  | 443 | int window, u32 session, u32 self, | 
|  | 444 | u32 peer, u8 *peer_id, u16 peer_caps, | 
|  | 445 | struct tipc_link *bc_sndlink, | 
|  | 446 | struct tipc_link *bc_rcvlink, | 
|  | 447 | struct sk_buff_head *inputq, | 
|  | 448 | struct sk_buff_head *namedq, | 
|  | 449 | struct tipc_link **link) | 
|  | 450 | { | 
|  | 451 | char peer_str[NODE_ID_STR_LEN] = {0,}; | 
|  | 452 | char self_str[NODE_ID_STR_LEN] = {0,}; | 
|  | 453 | struct tipc_link *l; | 
|  | 454 |  | 
|  | 455 | l = kzalloc(sizeof(*l), GFP_ATOMIC); | 
|  | 456 | if (!l) | 
|  | 457 | return false; | 
|  | 458 | *link = l; | 
|  | 459 | l->session = session; | 
|  | 460 |  | 
|  | 461 | /* Set link name for unicast links only */ | 
|  | 462 | if (peer_id) { | 
|  | 463 | tipc_nodeid2string(self_str, tipc_own_id(net)); | 
|  | 464 | if (strlen(self_str) > 16) | 
|  | 465 | sprintf(self_str, "%x", self); | 
|  | 466 | tipc_nodeid2string(peer_str, peer_id); | 
|  | 467 | if (strlen(peer_str) > 16) | 
|  | 468 | sprintf(peer_str, "%x", peer); | 
|  | 469 | } | 
|  | 470 | /* Peer i/f name will be completed by reset/activate message */ | 
|  | 471 | snprintf(l->name, sizeof(l->name), "%s:%s-%s:unknown", | 
|  | 472 | self_str, if_name, peer_str); | 
|  | 473 |  | 
|  | 474 | strcpy(l->if_name, if_name); | 
|  | 475 | l->addr = peer; | 
|  | 476 | l->peer_caps = peer_caps; | 
|  | 477 | l->net = net; | 
|  | 478 | l->in_session = false; | 
|  | 479 | l->bearer_id = bearer_id; | 
|  | 480 | l->tolerance = tolerance; | 
|  | 481 | if (bc_rcvlink) | 
|  | 482 | bc_rcvlink->tolerance = tolerance; | 
|  | 483 | l->net_plane = net_plane; | 
|  | 484 | l->advertised_mtu = mtu; | 
|  | 485 | l->mtu = mtu; | 
|  | 486 | l->priority = priority; | 
|  | 487 | tipc_link_set_queue_limits(l, window); | 
|  | 488 | l->ackers = 1; | 
|  | 489 | l->bc_sndlink = bc_sndlink; | 
|  | 490 | l->bc_rcvlink = bc_rcvlink; | 
|  | 491 | l->inputq = inputq; | 
|  | 492 | l->namedq = namedq; | 
|  | 493 | l->state = LINK_RESETTING; | 
|  | 494 | __skb_queue_head_init(&l->transmq); | 
|  | 495 | __skb_queue_head_init(&l->backlogq); | 
|  | 496 | __skb_queue_head_init(&l->deferdq); | 
|  | 497 | skb_queue_head_init(&l->wakeupq); | 
|  | 498 | skb_queue_head_init(l->inputq); | 
|  | 499 | return true; | 
|  | 500 | } | 
|  | 501 |  | 
|  | 502 | /** | 
|  | 503 | * tipc_link_bc_create - create new link to be used for broadcast | 
|  | 504 | * @n: pointer to associated node | 
|  | 505 | * @mtu: mtu to be used initially if no peers | 
|  | 506 | * @window: send window to be used | 
|  | 507 | * @inputq: queue to put messages ready for delivery | 
|  | 508 | * @namedq: queue to put binding table update messages ready for delivery | 
|  | 509 | * @link: return value, pointer to put the created link | 
|  | 510 | * | 
|  | 511 | * Returns true if link was created, otherwise false | 
|  | 512 | */ | 
|  | 513 | bool tipc_link_bc_create(struct net *net, u32 ownnode, u32 peer, | 
|  | 514 | int mtu, int window, u16 peer_caps, | 
|  | 515 | struct sk_buff_head *inputq, | 
|  | 516 | struct sk_buff_head *namedq, | 
|  | 517 | struct tipc_link *bc_sndlink, | 
|  | 518 | struct tipc_link **link) | 
|  | 519 | { | 
|  | 520 | struct tipc_link *l; | 
|  | 521 |  | 
|  | 522 | if (!tipc_link_create(net, "", MAX_BEARERS, 0, 'Z', mtu, 0, window, | 
|  | 523 | 0, ownnode, peer, NULL, peer_caps, bc_sndlink, | 
|  | 524 | NULL, inputq, namedq, link)) | 
|  | 525 | return false; | 
|  | 526 |  | 
|  | 527 | l = *link; | 
|  | 528 | strcpy(l->name, tipc_bclink_name); | 
|  | 529 | tipc_link_reset(l); | 
|  | 530 | l->state = LINK_RESET; | 
|  | 531 | l->ackers = 0; | 
|  | 532 | l->bc_rcvlink = l; | 
|  | 533 |  | 
|  | 534 | /* Broadcast send link is always up */ | 
|  | 535 | if (link_is_bc_sndlink(l)) | 
|  | 536 | l->state = LINK_ESTABLISHED; | 
|  | 537 |  | 
|  | 538 | /* Disable replicast if even a single peer doesn't support it */ | 
|  | 539 | if (link_is_bc_rcvlink(l) && !(peer_caps & TIPC_BCAST_RCAST)) | 
|  | 540 | tipc_bcast_disable_rcast(net); | 
|  | 541 |  | 
|  | 542 | return true; | 
|  | 543 | } | 
|  | 544 |  | 
|  | 545 | /** | 
|  | 546 | * tipc_link_fsm_evt - link finite state machine | 
|  | 547 | * @l: pointer to link | 
|  | 548 | * @evt: state machine event to be processed | 
|  | 549 | */ | 
|  | 550 | int tipc_link_fsm_evt(struct tipc_link *l, int evt) | 
|  | 551 | { | 
|  | 552 | int rc = 0; | 
|  | 553 |  | 
|  | 554 | switch (l->state) { | 
|  | 555 | case LINK_RESETTING: | 
|  | 556 | switch (evt) { | 
|  | 557 | case LINK_PEER_RESET_EVT: | 
|  | 558 | l->state = LINK_PEER_RESET; | 
|  | 559 | break; | 
|  | 560 | case LINK_RESET_EVT: | 
|  | 561 | l->state = LINK_RESET; | 
|  | 562 | break; | 
|  | 563 | case LINK_FAILURE_EVT: | 
|  | 564 | case LINK_FAILOVER_BEGIN_EVT: | 
|  | 565 | case LINK_ESTABLISH_EVT: | 
|  | 566 | case LINK_FAILOVER_END_EVT: | 
|  | 567 | case LINK_SYNCH_BEGIN_EVT: | 
|  | 568 | case LINK_SYNCH_END_EVT: | 
|  | 569 | default: | 
|  | 570 | goto illegal_evt; | 
|  | 571 | } | 
|  | 572 | break; | 
|  | 573 | case LINK_RESET: | 
|  | 574 | switch (evt) { | 
|  | 575 | case LINK_PEER_RESET_EVT: | 
|  | 576 | l->state = LINK_ESTABLISHING; | 
|  | 577 | break; | 
|  | 578 | case LINK_FAILOVER_BEGIN_EVT: | 
|  | 579 | l->state = LINK_FAILINGOVER; | 
|  | 580 | case LINK_FAILURE_EVT: | 
|  | 581 | case LINK_RESET_EVT: | 
|  | 582 | case LINK_ESTABLISH_EVT: | 
|  | 583 | case LINK_FAILOVER_END_EVT: | 
|  | 584 | break; | 
|  | 585 | case LINK_SYNCH_BEGIN_EVT: | 
|  | 586 | case LINK_SYNCH_END_EVT: | 
|  | 587 | default: | 
|  | 588 | goto illegal_evt; | 
|  | 589 | } | 
|  | 590 | break; | 
|  | 591 | case LINK_PEER_RESET: | 
|  | 592 | switch (evt) { | 
|  | 593 | case LINK_RESET_EVT: | 
|  | 594 | l->state = LINK_ESTABLISHING; | 
|  | 595 | break; | 
|  | 596 | case LINK_PEER_RESET_EVT: | 
|  | 597 | case LINK_ESTABLISH_EVT: | 
|  | 598 | case LINK_FAILURE_EVT: | 
|  | 599 | break; | 
|  | 600 | case LINK_SYNCH_BEGIN_EVT: | 
|  | 601 | case LINK_SYNCH_END_EVT: | 
|  | 602 | case LINK_FAILOVER_BEGIN_EVT: | 
|  | 603 | case LINK_FAILOVER_END_EVT: | 
|  | 604 | default: | 
|  | 605 | goto illegal_evt; | 
|  | 606 | } | 
|  | 607 | break; | 
|  | 608 | case LINK_FAILINGOVER: | 
|  | 609 | switch (evt) { | 
|  | 610 | case LINK_FAILOVER_END_EVT: | 
|  | 611 | l->state = LINK_RESET; | 
|  | 612 | break; | 
|  | 613 | case LINK_PEER_RESET_EVT: | 
|  | 614 | case LINK_RESET_EVT: | 
|  | 615 | case LINK_ESTABLISH_EVT: | 
|  | 616 | case LINK_FAILURE_EVT: | 
|  | 617 | break; | 
|  | 618 | case LINK_FAILOVER_BEGIN_EVT: | 
|  | 619 | case LINK_SYNCH_BEGIN_EVT: | 
|  | 620 | case LINK_SYNCH_END_EVT: | 
|  | 621 | default: | 
|  | 622 | goto illegal_evt; | 
|  | 623 | } | 
|  | 624 | break; | 
|  | 625 | case LINK_ESTABLISHING: | 
|  | 626 | switch (evt) { | 
|  | 627 | case LINK_ESTABLISH_EVT: | 
|  | 628 | l->state = LINK_ESTABLISHED; | 
|  | 629 | break; | 
|  | 630 | case LINK_FAILOVER_BEGIN_EVT: | 
|  | 631 | l->state = LINK_FAILINGOVER; | 
|  | 632 | break; | 
|  | 633 | case LINK_RESET_EVT: | 
|  | 634 | l->state = LINK_RESET; | 
|  | 635 | break; | 
|  | 636 | case LINK_FAILURE_EVT: | 
|  | 637 | case LINK_PEER_RESET_EVT: | 
|  | 638 | case LINK_SYNCH_BEGIN_EVT: | 
|  | 639 | case LINK_FAILOVER_END_EVT: | 
|  | 640 | break; | 
|  | 641 | case LINK_SYNCH_END_EVT: | 
|  | 642 | default: | 
|  | 643 | goto illegal_evt; | 
|  | 644 | } | 
|  | 645 | break; | 
|  | 646 | case LINK_ESTABLISHED: | 
|  | 647 | switch (evt) { | 
|  | 648 | case LINK_PEER_RESET_EVT: | 
|  | 649 | l->state = LINK_PEER_RESET; | 
|  | 650 | rc |= TIPC_LINK_DOWN_EVT; | 
|  | 651 | break; | 
|  | 652 | case LINK_FAILURE_EVT: | 
|  | 653 | l->state = LINK_RESETTING; | 
|  | 654 | rc |= TIPC_LINK_DOWN_EVT; | 
|  | 655 | break; | 
|  | 656 | case LINK_RESET_EVT: | 
|  | 657 | l->state = LINK_RESET; | 
|  | 658 | break; | 
|  | 659 | case LINK_ESTABLISH_EVT: | 
|  | 660 | case LINK_SYNCH_END_EVT: | 
|  | 661 | break; | 
|  | 662 | case LINK_SYNCH_BEGIN_EVT: | 
|  | 663 | l->state = LINK_SYNCHING; | 
|  | 664 | break; | 
|  | 665 | case LINK_FAILOVER_BEGIN_EVT: | 
|  | 666 | case LINK_FAILOVER_END_EVT: | 
|  | 667 | default: | 
|  | 668 | goto illegal_evt; | 
|  | 669 | } | 
|  | 670 | break; | 
|  | 671 | case LINK_SYNCHING: | 
|  | 672 | switch (evt) { | 
|  | 673 | case LINK_PEER_RESET_EVT: | 
|  | 674 | l->state = LINK_PEER_RESET; | 
|  | 675 | rc |= TIPC_LINK_DOWN_EVT; | 
|  | 676 | break; | 
|  | 677 | case LINK_FAILURE_EVT: | 
|  | 678 | l->state = LINK_RESETTING; | 
|  | 679 | rc |= TIPC_LINK_DOWN_EVT; | 
|  | 680 | break; | 
|  | 681 | case LINK_RESET_EVT: | 
|  | 682 | l->state = LINK_RESET; | 
|  | 683 | break; | 
|  | 684 | case LINK_ESTABLISH_EVT: | 
|  | 685 | case LINK_SYNCH_BEGIN_EVT: | 
|  | 686 | break; | 
|  | 687 | case LINK_SYNCH_END_EVT: | 
|  | 688 | l->state = LINK_ESTABLISHED; | 
|  | 689 | break; | 
|  | 690 | case LINK_FAILOVER_BEGIN_EVT: | 
|  | 691 | case LINK_FAILOVER_END_EVT: | 
|  | 692 | default: | 
|  | 693 | goto illegal_evt; | 
|  | 694 | } | 
|  | 695 | break; | 
|  | 696 | default: | 
|  | 697 | pr_err("Unknown FSM state %x in %s\n", l->state, l->name); | 
|  | 698 | } | 
|  | 699 | return rc; | 
|  | 700 | illegal_evt: | 
|  | 701 | pr_err("Illegal FSM event %x in state %x on link %s\n", | 
|  | 702 | evt, l->state, l->name); | 
|  | 703 | return rc; | 
|  | 704 | } | 
|  | 705 |  | 
|  | 706 | /* link_profile_stats - update statistical profiling of traffic | 
|  | 707 | */ | 
|  | 708 | static void link_profile_stats(struct tipc_link *l) | 
|  | 709 | { | 
|  | 710 | struct sk_buff *skb; | 
|  | 711 | struct tipc_msg *msg; | 
|  | 712 | int length; | 
|  | 713 |  | 
|  | 714 | /* Update counters used in statistical profiling of send traffic */ | 
|  | 715 | l->stats.accu_queue_sz += skb_queue_len(&l->transmq); | 
|  | 716 | l->stats.queue_sz_counts++; | 
|  | 717 |  | 
|  | 718 | skb = skb_peek(&l->transmq); | 
|  | 719 | if (!skb) | 
|  | 720 | return; | 
|  | 721 | msg = buf_msg(skb); | 
|  | 722 | length = msg_size(msg); | 
|  | 723 |  | 
|  | 724 | if (msg_user(msg) == MSG_FRAGMENTER) { | 
|  | 725 | if (msg_type(msg) != FIRST_FRAGMENT) | 
|  | 726 | return; | 
|  | 727 | length = msg_size(msg_get_wrapped(msg)); | 
|  | 728 | } | 
|  | 729 | l->stats.msg_lengths_total += length; | 
|  | 730 | l->stats.msg_length_counts++; | 
|  | 731 | if (length <= 64) | 
|  | 732 | l->stats.msg_length_profile[0]++; | 
|  | 733 | else if (length <= 256) | 
|  | 734 | l->stats.msg_length_profile[1]++; | 
|  | 735 | else if (length <= 1024) | 
|  | 736 | l->stats.msg_length_profile[2]++; | 
|  | 737 | else if (length <= 4096) | 
|  | 738 | l->stats.msg_length_profile[3]++; | 
|  | 739 | else if (length <= 16384) | 
|  | 740 | l->stats.msg_length_profile[4]++; | 
|  | 741 | else if (length <= 32768) | 
|  | 742 | l->stats.msg_length_profile[5]++; | 
|  | 743 | else | 
|  | 744 | l->stats.msg_length_profile[6]++; | 
|  | 745 | } | 
|  | 746 |  | 
|  | 747 | /* tipc_link_timeout - perform periodic task as instructed from node timeout | 
|  | 748 | */ | 
|  | 749 | int tipc_link_timeout(struct tipc_link *l, struct sk_buff_head *xmitq) | 
|  | 750 | { | 
|  | 751 | int mtyp = 0; | 
|  | 752 | int rc = 0; | 
|  | 753 | bool state = false; | 
|  | 754 | bool probe = false; | 
|  | 755 | bool setup = false; | 
|  | 756 | u16 bc_snt = l->bc_sndlink->snd_nxt - 1; | 
|  | 757 | u16 bc_acked = l->bc_rcvlink->acked; | 
|  | 758 | struct tipc_mon_state *mstate = &l->mon_state; | 
|  | 759 |  | 
|  | 760 | switch (l->state) { | 
|  | 761 | case LINK_ESTABLISHED: | 
|  | 762 | case LINK_SYNCHING: | 
|  | 763 | mtyp = STATE_MSG; | 
|  | 764 | link_profile_stats(l); | 
|  | 765 | tipc_mon_get_state(l->net, l->addr, mstate, l->bearer_id); | 
|  | 766 | if (mstate->reset || (l->silent_intv_cnt > l->abort_limit)) | 
|  | 767 | return tipc_link_fsm_evt(l, LINK_FAILURE_EVT); | 
|  | 768 | state = bc_acked != bc_snt; | 
|  | 769 | state |= l->bc_rcvlink->rcv_unacked; | 
|  | 770 | state |= l->rcv_unacked; | 
|  | 771 | state |= !skb_queue_empty(&l->transmq); | 
|  | 772 | state |= !skb_queue_empty(&l->deferdq); | 
|  | 773 | probe = mstate->probing; | 
|  | 774 | probe |= l->silent_intv_cnt; | 
|  | 775 | if (probe || mstate->monitoring) | 
|  | 776 | l->silent_intv_cnt++; | 
|  | 777 | break; | 
|  | 778 | case LINK_RESET: | 
|  | 779 | setup = l->rst_cnt++ <= 4; | 
|  | 780 | setup |= !(l->rst_cnt % 16); | 
|  | 781 | mtyp = RESET_MSG; | 
|  | 782 | break; | 
|  | 783 | case LINK_ESTABLISHING: | 
|  | 784 | setup = true; | 
|  | 785 | mtyp = ACTIVATE_MSG; | 
|  | 786 | break; | 
|  | 787 | case LINK_PEER_RESET: | 
|  | 788 | case LINK_RESETTING: | 
|  | 789 | case LINK_FAILINGOVER: | 
|  | 790 | break; | 
|  | 791 | default: | 
|  | 792 | break; | 
|  | 793 | } | 
|  | 794 |  | 
|  | 795 | if (state || probe || setup) | 
|  | 796 | tipc_link_build_proto_msg(l, mtyp, probe, 0, 0, 0, 0, xmitq); | 
|  | 797 |  | 
|  | 798 | return rc; | 
|  | 799 | } | 
|  | 800 |  | 
|  | 801 | /** | 
|  | 802 | * link_schedule_user - schedule a message sender for wakeup after congestion | 
|  | 803 | * @l: congested link | 
|  | 804 | * @hdr: header of message that is being sent | 
|  | 805 | * Create pseudo msg to send back to user when congestion abates | 
|  | 806 | */ | 
|  | 807 | static int link_schedule_user(struct tipc_link *l, struct tipc_msg *hdr) | 
|  | 808 | { | 
|  | 809 | u32 dnode = tipc_own_addr(l->net); | 
|  | 810 | u32 dport = msg_origport(hdr); | 
|  | 811 | struct sk_buff *skb; | 
|  | 812 |  | 
|  | 813 | /* Create and schedule wakeup pseudo message */ | 
|  | 814 | skb = tipc_msg_create(SOCK_WAKEUP, 0, INT_H_SIZE, 0, | 
|  | 815 | dnode, l->addr, dport, 0, 0); | 
|  | 816 | if (!skb) | 
|  | 817 | return -ENOBUFS; | 
|  | 818 | msg_set_dest_droppable(buf_msg(skb), true); | 
|  | 819 | TIPC_SKB_CB(skb)->chain_imp = msg_importance(hdr); | 
|  | 820 | skb_queue_tail(&l->wakeupq, skb); | 
|  | 821 | l->stats.link_congs++; | 
|  | 822 | return -ELINKCONG; | 
|  | 823 | } | 
|  | 824 |  | 
|  | 825 | /** | 
|  | 826 | * link_prepare_wakeup - prepare users for wakeup after congestion | 
|  | 827 | * @l: congested link | 
|  | 828 | * Wake up a number of waiting users, as permitted by available space | 
|  | 829 | * in the send queue | 
|  | 830 | */ | 
|  | 831 | static void link_prepare_wakeup(struct tipc_link *l) | 
|  | 832 | { | 
|  | 833 | struct sk_buff *skb, *tmp; | 
|  | 834 | int imp, i = 0; | 
|  | 835 |  | 
|  | 836 | skb_queue_walk_safe(&l->wakeupq, skb, tmp) { | 
|  | 837 | imp = TIPC_SKB_CB(skb)->chain_imp; | 
|  | 838 | if (l->backlog[imp].len < l->backlog[imp].limit) { | 
|  | 839 | skb_unlink(skb, &l->wakeupq); | 
|  | 840 | skb_queue_tail(l->inputq, skb); | 
|  | 841 | } else if (i++ > 10) { | 
|  | 842 | break; | 
|  | 843 | } | 
|  | 844 | } | 
|  | 845 | } | 
|  | 846 |  | 
|  | 847 | void tipc_link_reset(struct tipc_link *l) | 
|  | 848 | { | 
|  | 849 | struct sk_buff_head list; | 
|  | 850 | u32 imp; | 
|  | 851 |  | 
|  | 852 | __skb_queue_head_init(&list); | 
|  | 853 |  | 
|  | 854 | l->in_session = false; | 
|  | 855 | l->session++; | 
|  | 856 | l->mtu = l->advertised_mtu; | 
|  | 857 |  | 
|  | 858 | spin_lock_bh(&l->wakeupq.lock); | 
|  | 859 | skb_queue_splice_init(&l->wakeupq, &list); | 
|  | 860 | spin_unlock_bh(&l->wakeupq.lock); | 
|  | 861 |  | 
|  | 862 | spin_lock_bh(&l->inputq->lock); | 
|  | 863 | skb_queue_splice_init(&list, l->inputq); | 
|  | 864 | spin_unlock_bh(&l->inputq->lock); | 
|  | 865 |  | 
|  | 866 | __skb_queue_purge(&l->transmq); | 
|  | 867 | __skb_queue_purge(&l->deferdq); | 
|  | 868 | __skb_queue_purge(&l->backlogq); | 
|  | 869 | for (imp = 0; imp <= TIPC_SYSTEM_IMPORTANCE; imp++) { | 
|  | 870 | l->backlog[imp].len = 0; | 
|  | 871 | l->backlog[imp].target_bskb = NULL; | 
|  | 872 | } | 
|  | 873 | kfree_skb(l->reasm_buf); | 
|  | 874 | kfree_skb(l->failover_reasm_skb); | 
|  | 875 | l->reasm_buf = NULL; | 
|  | 876 | l->failover_reasm_skb = NULL; | 
|  | 877 | l->rcv_unacked = 0; | 
|  | 878 | l->snd_nxt = 1; | 
|  | 879 | l->rcv_nxt = 1; | 
|  | 880 | l->snd_nxt_state = 1; | 
|  | 881 | l->rcv_nxt_state = 1; | 
|  | 882 | l->acked = 0; | 
|  | 883 | l->silent_intv_cnt = 0; | 
|  | 884 | l->rst_cnt = 0; | 
|  | 885 | l->stale_cnt = 0; | 
|  | 886 | l->bc_peer_is_up = false; | 
|  | 887 | memset(&l->mon_state, 0, sizeof(l->mon_state)); | 
|  | 888 | tipc_link_reset_stats(l); | 
|  | 889 | } | 
|  | 890 |  | 
|  | 891 | /** | 
|  | 892 | * tipc_link_xmit(): enqueue buffer list according to queue situation | 
|  | 893 | * @link: link to use | 
|  | 894 | * @list: chain of buffers containing message | 
|  | 895 | * @xmitq: returned list of packets to be sent by caller | 
|  | 896 | * | 
|  | 897 | * Consumes the buffer chain. | 
|  | 898 | * Returns 0 if success, or errno: -ELINKCONG, -EMSGSIZE or -ENOBUFS | 
|  | 899 | * Messages at TIPC_SYSTEM_IMPORTANCE are always accepted | 
|  | 900 | */ | 
|  | 901 | int tipc_link_xmit(struct tipc_link *l, struct sk_buff_head *list, | 
|  | 902 | struct sk_buff_head *xmitq) | 
|  | 903 | { | 
|  | 904 | struct tipc_msg *hdr = buf_msg(skb_peek(list)); | 
|  | 905 | unsigned int maxwin = l->window; | 
|  | 906 | int imp = msg_importance(hdr); | 
|  | 907 | unsigned int mtu = l->mtu; | 
|  | 908 | u16 ack = l->rcv_nxt - 1; | 
|  | 909 | u16 seqno = l->snd_nxt; | 
|  | 910 | u16 bc_ack = l->bc_rcvlink->rcv_nxt - 1; | 
|  | 911 | struct sk_buff_head *transmq = &l->transmq; | 
|  | 912 | struct sk_buff_head *backlogq = &l->backlogq; | 
|  | 913 | struct sk_buff *skb, *_skb, **tskb; | 
|  | 914 | int pkt_cnt = skb_queue_len(list); | 
|  | 915 | int rc = 0; | 
|  | 916 |  | 
|  | 917 | if (unlikely(msg_size(hdr) > mtu)) { | 
|  | 918 | skb_queue_purge(list); | 
|  | 919 | return -EMSGSIZE; | 
|  | 920 | } | 
|  | 921 |  | 
|  | 922 | /* Allow oversubscription of one data msg per source at congestion */ | 
|  | 923 | if (unlikely(l->backlog[imp].len >= l->backlog[imp].limit)) { | 
|  | 924 | if (imp == TIPC_SYSTEM_IMPORTANCE) { | 
|  | 925 | pr_warn("%s<%s>, link overflow", link_rst_msg, l->name); | 
|  | 926 | return -ENOBUFS; | 
|  | 927 | } | 
|  | 928 | rc = link_schedule_user(l, hdr); | 
|  | 929 | } | 
|  | 930 |  | 
|  | 931 | if (pkt_cnt > 1) { | 
|  | 932 | l->stats.sent_fragmented++; | 
|  | 933 | l->stats.sent_fragments += pkt_cnt; | 
|  | 934 | } | 
|  | 935 |  | 
|  | 936 | /* Prepare each packet for sending, and add to relevant queue: */ | 
|  | 937 | while (skb_queue_len(list)) { | 
|  | 938 | skb = skb_peek(list); | 
|  | 939 | hdr = buf_msg(skb); | 
|  | 940 | msg_set_seqno(hdr, seqno); | 
|  | 941 | msg_set_ack(hdr, ack); | 
|  | 942 | msg_set_bcast_ack(hdr, bc_ack); | 
|  | 943 |  | 
|  | 944 | if (likely(skb_queue_len(transmq) < maxwin)) { | 
|  | 945 | _skb = skb_clone(skb, GFP_ATOMIC); | 
|  | 946 | if (!_skb) { | 
|  | 947 | skb_queue_purge(list); | 
|  | 948 | return -ENOBUFS; | 
|  | 949 | } | 
|  | 950 | __skb_dequeue(list); | 
|  | 951 | __skb_queue_tail(transmq, skb); | 
|  | 952 | __skb_queue_tail(xmitq, _skb); | 
|  | 953 | TIPC_SKB_CB(skb)->ackers = l->ackers; | 
|  | 954 | l->rcv_unacked = 0; | 
|  | 955 | l->stats.sent_pkts++; | 
|  | 956 | seqno++; | 
|  | 957 | continue; | 
|  | 958 | } | 
|  | 959 | tskb = &l->backlog[imp].target_bskb; | 
|  | 960 | if (tipc_msg_bundle(*tskb, hdr, mtu)) { | 
|  | 961 | kfree_skb(__skb_dequeue(list)); | 
|  | 962 | l->stats.sent_bundled++; | 
|  | 963 | continue; | 
|  | 964 | } | 
|  | 965 | if (tipc_msg_make_bundle(tskb, hdr, mtu, l->addr)) { | 
|  | 966 | kfree_skb(__skb_dequeue(list)); | 
|  | 967 | __skb_queue_tail(backlogq, *tskb); | 
|  | 968 | l->backlog[imp].len++; | 
|  | 969 | l->stats.sent_bundled++; | 
|  | 970 | l->stats.sent_bundles++; | 
|  | 971 | continue; | 
|  | 972 | } | 
|  | 973 | l->backlog[imp].target_bskb = NULL; | 
|  | 974 | l->backlog[imp].len += skb_queue_len(list); | 
|  | 975 | skb_queue_splice_tail_init(list, backlogq); | 
|  | 976 | } | 
|  | 977 | l->snd_nxt = seqno; | 
|  | 978 | return rc; | 
|  | 979 | } | 
|  | 980 |  | 
|  | 981 | static void tipc_link_advance_backlog(struct tipc_link *l, | 
|  | 982 | struct sk_buff_head *xmitq) | 
|  | 983 | { | 
|  | 984 | struct sk_buff *skb, *_skb; | 
|  | 985 | struct tipc_msg *hdr; | 
|  | 986 | u16 seqno = l->snd_nxt; | 
|  | 987 | u16 ack = l->rcv_nxt - 1; | 
|  | 988 | u16 bc_ack = l->bc_rcvlink->rcv_nxt - 1; | 
|  | 989 | u32 imp; | 
|  | 990 |  | 
|  | 991 | while (skb_queue_len(&l->transmq) < l->window) { | 
|  | 992 | skb = skb_peek(&l->backlogq); | 
|  | 993 | if (!skb) | 
|  | 994 | break; | 
|  | 995 | _skb = skb_clone(skb, GFP_ATOMIC); | 
|  | 996 | if (!_skb) | 
|  | 997 | break; | 
|  | 998 | __skb_dequeue(&l->backlogq); | 
|  | 999 | hdr = buf_msg(skb); | 
|  | 1000 | imp = msg_importance(hdr); | 
|  | 1001 | l->backlog[imp].len--; | 
|  | 1002 | if (unlikely(skb == l->backlog[imp].target_bskb)) | 
|  | 1003 | l->backlog[imp].target_bskb = NULL; | 
|  | 1004 | __skb_queue_tail(&l->transmq, skb); | 
|  | 1005 | __skb_queue_tail(xmitq, _skb); | 
|  | 1006 | TIPC_SKB_CB(skb)->ackers = l->ackers; | 
|  | 1007 | msg_set_seqno(hdr, seqno); | 
|  | 1008 | msg_set_ack(hdr, ack); | 
|  | 1009 | msg_set_bcast_ack(hdr, bc_ack); | 
|  | 1010 | l->rcv_unacked = 0; | 
|  | 1011 | l->stats.sent_pkts++; | 
|  | 1012 | seqno++; | 
|  | 1013 | } | 
|  | 1014 | l->snd_nxt = seqno; | 
|  | 1015 | } | 
|  | 1016 |  | 
|  | 1017 | static void link_retransmit_failure(struct tipc_link *l, struct sk_buff *skb) | 
|  | 1018 | { | 
|  | 1019 | struct tipc_msg *hdr = buf_msg(skb); | 
|  | 1020 |  | 
|  | 1021 | pr_warn("Retransmission failure on link <%s>\n", l->name); | 
|  | 1022 | link_print(l, "State of link "); | 
|  | 1023 | pr_info("Failed msg: usr %u, typ %u, len %u, err %u\n", | 
|  | 1024 | msg_user(hdr), msg_type(hdr), msg_size(hdr), msg_errcode(hdr)); | 
|  | 1025 | pr_info("sqno %u, prev: %x, src: %x\n", | 
|  | 1026 | msg_seqno(hdr), msg_prevnode(hdr), msg_orignode(hdr)); | 
|  | 1027 | } | 
|  | 1028 |  | 
|  | 1029 | /* tipc_link_retrans() - retransmit one or more packets | 
|  | 1030 | * @l: the link to transmit on | 
|  | 1031 | * @r: the receiving link ordering the retransmit. Same as l if unicast | 
|  | 1032 | * @from: retransmit from (inclusive) this sequence number | 
|  | 1033 | * @to: retransmit to (inclusive) this sequence number | 
|  | 1034 | * xmitq: queue for accumulating the retransmitted packets | 
|  | 1035 | */ | 
|  | 1036 | static int tipc_link_retrans(struct tipc_link *l, struct tipc_link *r, | 
|  | 1037 | u16 from, u16 to, struct sk_buff_head *xmitq) | 
|  | 1038 | { | 
|  | 1039 | struct sk_buff *_skb, *skb = skb_peek(&l->transmq); | 
|  | 1040 | u16 bc_ack = l->bc_rcvlink->rcv_nxt - 1; | 
|  | 1041 | u16 ack = l->rcv_nxt - 1; | 
|  | 1042 | struct tipc_msg *hdr; | 
|  | 1043 |  | 
|  | 1044 | if (!skb) | 
|  | 1045 | return 0; | 
|  | 1046 |  | 
|  | 1047 | /* Detect repeated retransmit failures on same packet */ | 
|  | 1048 | if (r->last_retransm != buf_seqno(skb)) { | 
|  | 1049 | r->last_retransm = buf_seqno(skb); | 
|  | 1050 | r->stale_limit = jiffies + msecs_to_jiffies(r->tolerance); | 
|  | 1051 | r->stale_cnt = 0; | 
|  | 1052 | } else if (++r->stale_cnt > 99 && time_after(jiffies, r->stale_limit)) { | 
|  | 1053 | link_retransmit_failure(l, skb); | 
|  | 1054 | if (link_is_bc_sndlink(l)) | 
|  | 1055 | return TIPC_LINK_DOWN_EVT; | 
|  | 1056 | return tipc_link_fsm_evt(l, LINK_FAILURE_EVT); | 
|  | 1057 | } | 
|  | 1058 |  | 
|  | 1059 | skb_queue_walk(&l->transmq, skb) { | 
|  | 1060 | hdr = buf_msg(skb); | 
|  | 1061 | if (less(msg_seqno(hdr), from)) | 
|  | 1062 | continue; | 
|  | 1063 | if (more(msg_seqno(hdr), to)) | 
|  | 1064 | break; | 
|  | 1065 | _skb = __pskb_copy(skb, MIN_H_SIZE, GFP_ATOMIC); | 
|  | 1066 | if (!_skb) | 
|  | 1067 | return 0; | 
|  | 1068 | hdr = buf_msg(_skb); | 
|  | 1069 | msg_set_ack(hdr, ack); | 
|  | 1070 | msg_set_bcast_ack(hdr, bc_ack); | 
|  | 1071 | _skb->priority = TC_PRIO_CONTROL; | 
|  | 1072 | __skb_queue_tail(xmitq, _skb); | 
|  | 1073 | l->stats.retransmitted++; | 
|  | 1074 | } | 
|  | 1075 | return 0; | 
|  | 1076 | } | 
|  | 1077 |  | 
|  | 1078 | /* tipc_data_input - deliver data and name distr msgs to upper layer | 
|  | 1079 | * | 
|  | 1080 | * Consumes buffer if message is of right type | 
|  | 1081 | * Node lock must be held | 
|  | 1082 | */ | 
|  | 1083 | static bool tipc_data_input(struct tipc_link *l, struct sk_buff *skb, | 
|  | 1084 | struct sk_buff_head *inputq) | 
|  | 1085 | { | 
|  | 1086 | struct sk_buff_head *mc_inputq = l->bc_rcvlink->inputq; | 
|  | 1087 | struct tipc_msg *hdr = buf_msg(skb); | 
|  | 1088 |  | 
|  | 1089 | switch (msg_user(hdr)) { | 
|  | 1090 | case TIPC_LOW_IMPORTANCE: | 
|  | 1091 | case TIPC_MEDIUM_IMPORTANCE: | 
|  | 1092 | case TIPC_HIGH_IMPORTANCE: | 
|  | 1093 | case TIPC_CRITICAL_IMPORTANCE: | 
|  | 1094 | if (unlikely(msg_in_group(hdr) || msg_mcast(hdr))) { | 
|  | 1095 | skb_queue_tail(mc_inputq, skb); | 
|  | 1096 | return true; | 
|  | 1097 | } | 
|  | 1098 | /* else: fall through */ | 
|  | 1099 | case CONN_MANAGER: | 
|  | 1100 | skb_queue_tail(inputq, skb); | 
|  | 1101 | return true; | 
|  | 1102 | case GROUP_PROTOCOL: | 
|  | 1103 | skb_queue_tail(mc_inputq, skb); | 
|  | 1104 | return true; | 
|  | 1105 | case NAME_DISTRIBUTOR: | 
|  | 1106 | l->bc_rcvlink->state = LINK_ESTABLISHED; | 
|  | 1107 | skb_queue_tail(l->namedq, skb); | 
|  | 1108 | return true; | 
|  | 1109 | case MSG_BUNDLER: | 
|  | 1110 | case TUNNEL_PROTOCOL: | 
|  | 1111 | case MSG_FRAGMENTER: | 
|  | 1112 | case BCAST_PROTOCOL: | 
|  | 1113 | return false; | 
|  | 1114 | default: | 
|  | 1115 | pr_warn("Dropping received illegal msg type\n"); | 
|  | 1116 | kfree_skb(skb); | 
|  | 1117 | return true; | 
|  | 1118 | }; | 
|  | 1119 | } | 
|  | 1120 |  | 
|  | 1121 | /* tipc_link_input - process packet that has passed link protocol check | 
|  | 1122 | * | 
|  | 1123 | * Consumes buffer | 
|  | 1124 | */ | 
|  | 1125 | static int tipc_link_input(struct tipc_link *l, struct sk_buff *skb, | 
|  | 1126 | struct sk_buff_head *inputq) | 
|  | 1127 | { | 
|  | 1128 | struct tipc_msg *hdr = buf_msg(skb); | 
|  | 1129 | struct sk_buff **reasm_skb = &l->reasm_buf; | 
|  | 1130 | struct sk_buff *iskb; | 
|  | 1131 | struct sk_buff_head tmpq; | 
|  | 1132 | int usr = msg_user(hdr); | 
|  | 1133 | int rc = 0; | 
|  | 1134 | int pos = 0; | 
|  | 1135 | int ipos = 0; | 
|  | 1136 |  | 
|  | 1137 | if (unlikely(usr == TUNNEL_PROTOCOL)) { | 
|  | 1138 | if (msg_type(hdr) == SYNCH_MSG) { | 
|  | 1139 | __skb_queue_purge(&l->deferdq); | 
|  | 1140 | goto drop; | 
|  | 1141 | } | 
|  | 1142 | if (!tipc_msg_extract(skb, &iskb, &ipos)) | 
|  | 1143 | return rc; | 
|  | 1144 | kfree_skb(skb); | 
|  | 1145 | skb = iskb; | 
|  | 1146 | hdr = buf_msg(skb); | 
|  | 1147 | if (less(msg_seqno(hdr), l->drop_point)) | 
|  | 1148 | goto drop; | 
|  | 1149 | if (tipc_data_input(l, skb, inputq)) | 
|  | 1150 | return rc; | 
|  | 1151 | usr = msg_user(hdr); | 
|  | 1152 | reasm_skb = &l->failover_reasm_skb; | 
|  | 1153 | } | 
|  | 1154 |  | 
|  | 1155 | if (usr == MSG_BUNDLER) { | 
|  | 1156 | skb_queue_head_init(&tmpq); | 
|  | 1157 | l->stats.recv_bundles++; | 
|  | 1158 | l->stats.recv_bundled += msg_msgcnt(hdr); | 
|  | 1159 | while (tipc_msg_extract(skb, &iskb, &pos)) | 
|  | 1160 | tipc_data_input(l, iskb, &tmpq); | 
|  | 1161 | tipc_skb_queue_splice_tail(&tmpq, inputq); | 
|  | 1162 | return 0; | 
|  | 1163 | } else if (usr == MSG_FRAGMENTER) { | 
|  | 1164 | l->stats.recv_fragments++; | 
|  | 1165 | if (tipc_buf_append(reasm_skb, &skb)) { | 
|  | 1166 | l->stats.recv_fragmented++; | 
|  | 1167 | tipc_data_input(l, skb, inputq); | 
|  | 1168 | } else if (!*reasm_skb && !link_is_bc_rcvlink(l)) { | 
|  | 1169 | pr_warn_ratelimited("Unable to build fragment list\n"); | 
|  | 1170 | return tipc_link_fsm_evt(l, LINK_FAILURE_EVT); | 
|  | 1171 | } | 
|  | 1172 | return 0; | 
|  | 1173 | } else if (usr == BCAST_PROTOCOL) { | 
|  | 1174 | tipc_bcast_lock(l->net); | 
|  | 1175 | tipc_link_bc_init_rcv(l->bc_rcvlink, hdr); | 
|  | 1176 | tipc_bcast_unlock(l->net); | 
|  | 1177 | } | 
|  | 1178 | drop: | 
|  | 1179 | kfree_skb(skb); | 
|  | 1180 | return 0; | 
|  | 1181 | } | 
|  | 1182 |  | 
|  | 1183 | static bool tipc_link_release_pkts(struct tipc_link *l, u16 acked) | 
|  | 1184 | { | 
|  | 1185 | bool released = false; | 
|  | 1186 | struct sk_buff *skb, *tmp; | 
|  | 1187 |  | 
|  | 1188 | skb_queue_walk_safe(&l->transmq, skb, tmp) { | 
|  | 1189 | if (more(buf_seqno(skb), acked)) | 
|  | 1190 | break; | 
|  | 1191 | __skb_unlink(skb, &l->transmq); | 
|  | 1192 | kfree_skb(skb); | 
|  | 1193 | released = true; | 
|  | 1194 | } | 
|  | 1195 | return released; | 
|  | 1196 | } | 
|  | 1197 |  | 
|  | 1198 | /* tipc_link_build_state_msg: prepare link state message for transmission | 
|  | 1199 | * | 
|  | 1200 | * Note that sending of broadcast ack is coordinated among nodes, to reduce | 
|  | 1201 | * risk of ack storms towards the sender | 
|  | 1202 | */ | 
|  | 1203 | int tipc_link_build_state_msg(struct tipc_link *l, struct sk_buff_head *xmitq) | 
|  | 1204 | { | 
|  | 1205 | if (!l) | 
|  | 1206 | return 0; | 
|  | 1207 |  | 
|  | 1208 | /* Broadcast ACK must be sent via a unicast link => defer to caller */ | 
|  | 1209 | if (link_is_bc_rcvlink(l)) { | 
|  | 1210 | if (((l->rcv_nxt ^ tipc_own_addr(l->net)) & 0xf) != 0xf) | 
|  | 1211 | return 0; | 
|  | 1212 | l->rcv_unacked = 0; | 
|  | 1213 |  | 
|  | 1214 | /* Use snd_nxt to store peer's snd_nxt in broadcast rcv link */ | 
|  | 1215 | l->snd_nxt = l->rcv_nxt; | 
|  | 1216 | return TIPC_LINK_SND_STATE; | 
|  | 1217 | } | 
|  | 1218 |  | 
|  | 1219 | /* Unicast ACK */ | 
|  | 1220 | l->rcv_unacked = 0; | 
|  | 1221 | l->stats.sent_acks++; | 
|  | 1222 | tipc_link_build_proto_msg(l, STATE_MSG, 0, 0, 0, 0, 0, xmitq); | 
|  | 1223 | return 0; | 
|  | 1224 | } | 
|  | 1225 |  | 
|  | 1226 | /* tipc_link_build_reset_msg: prepare link RESET or ACTIVATE message | 
|  | 1227 | */ | 
|  | 1228 | void tipc_link_build_reset_msg(struct tipc_link *l, struct sk_buff_head *xmitq) | 
|  | 1229 | { | 
|  | 1230 | int mtyp = RESET_MSG; | 
|  | 1231 | struct sk_buff *skb; | 
|  | 1232 |  | 
|  | 1233 | if (l->state == LINK_ESTABLISHING) | 
|  | 1234 | mtyp = ACTIVATE_MSG; | 
|  | 1235 |  | 
|  | 1236 | tipc_link_build_proto_msg(l, mtyp, 0, 0, 0, 0, 0, xmitq); | 
|  | 1237 |  | 
|  | 1238 | /* Inform peer that this endpoint is going down if applicable */ | 
|  | 1239 | skb = skb_peek_tail(xmitq); | 
|  | 1240 | if (skb && (l->state == LINK_RESET)) | 
|  | 1241 | msg_set_peer_stopping(buf_msg(skb), 1); | 
|  | 1242 | } | 
|  | 1243 |  | 
|  | 1244 | /* tipc_link_build_nack_msg: prepare link nack message for transmission | 
|  | 1245 | * Note that sending of broadcast NACK is coordinated among nodes, to | 
|  | 1246 | * reduce the risk of NACK storms towards the sender | 
|  | 1247 | */ | 
|  | 1248 | static int tipc_link_build_nack_msg(struct tipc_link *l, | 
|  | 1249 | struct sk_buff_head *xmitq) | 
|  | 1250 | { | 
|  | 1251 | u32 def_cnt = ++l->stats.deferred_recv; | 
|  | 1252 | int match1, match2; | 
|  | 1253 |  | 
|  | 1254 | if (link_is_bc_rcvlink(l)) { | 
|  | 1255 | match1 = def_cnt & 0xf; | 
|  | 1256 | match2 = tipc_own_addr(l->net) & 0xf; | 
|  | 1257 | if (match1 == match2) | 
|  | 1258 | return TIPC_LINK_SND_STATE; | 
|  | 1259 | return 0; | 
|  | 1260 | } | 
|  | 1261 |  | 
|  | 1262 | if ((skb_queue_len(&l->deferdq) == 1) || !(def_cnt % TIPC_NACK_INTV)) | 
|  | 1263 | tipc_link_build_proto_msg(l, STATE_MSG, 0, 0, 0, 0, 0, xmitq); | 
|  | 1264 | return 0; | 
|  | 1265 | } | 
|  | 1266 |  | 
|  | 1267 | /* tipc_link_rcv - process TIPC packets/messages arriving from off-node | 
|  | 1268 | * @l: the link that should handle the message | 
|  | 1269 | * @skb: TIPC packet | 
|  | 1270 | * @xmitq: queue to place packets to be sent after this call | 
|  | 1271 | */ | 
|  | 1272 | int tipc_link_rcv(struct tipc_link *l, struct sk_buff *skb, | 
|  | 1273 | struct sk_buff_head *xmitq) | 
|  | 1274 | { | 
|  | 1275 | struct sk_buff_head *defq = &l->deferdq; | 
|  | 1276 | struct tipc_msg *hdr; | 
|  | 1277 | u16 seqno, rcv_nxt, win_lim; | 
|  | 1278 | int rc = 0; | 
|  | 1279 |  | 
|  | 1280 | do { | 
|  | 1281 | hdr = buf_msg(skb); | 
|  | 1282 | seqno = msg_seqno(hdr); | 
|  | 1283 | rcv_nxt = l->rcv_nxt; | 
|  | 1284 | win_lim = rcv_nxt + TIPC_MAX_LINK_WIN; | 
|  | 1285 |  | 
|  | 1286 | /* Verify and update link state */ | 
|  | 1287 | if (unlikely(msg_user(hdr) == LINK_PROTOCOL)) | 
|  | 1288 | return tipc_link_proto_rcv(l, skb, xmitq); | 
|  | 1289 |  | 
|  | 1290 | if (unlikely(!link_is_up(l))) { | 
|  | 1291 | if (l->state == LINK_ESTABLISHING) | 
|  | 1292 | rc = TIPC_LINK_UP_EVT; | 
|  | 1293 | goto drop; | 
|  | 1294 | } | 
|  | 1295 |  | 
|  | 1296 | /* Don't send probe at next timeout expiration */ | 
|  | 1297 | l->silent_intv_cnt = 0; | 
|  | 1298 |  | 
|  | 1299 | /* Drop if outside receive window */ | 
|  | 1300 | if (unlikely(less(seqno, rcv_nxt) || more(seqno, win_lim))) { | 
|  | 1301 | l->stats.duplicates++; | 
|  | 1302 | goto drop; | 
|  | 1303 | } | 
|  | 1304 |  | 
|  | 1305 | /* Forward queues and wake up waiting users */ | 
|  | 1306 | if (likely(tipc_link_release_pkts(l, msg_ack(hdr)))) { | 
|  | 1307 | l->stale_cnt = 0; | 
|  | 1308 | tipc_link_advance_backlog(l, xmitq); | 
|  | 1309 | if (unlikely(!skb_queue_empty(&l->wakeupq))) | 
|  | 1310 | link_prepare_wakeup(l); | 
|  | 1311 | } | 
|  | 1312 |  | 
|  | 1313 | /* Defer delivery if sequence gap */ | 
|  | 1314 | if (unlikely(seqno != rcv_nxt)) { | 
|  | 1315 | __tipc_skb_queue_sorted(defq, seqno, skb); | 
|  | 1316 | rc |= tipc_link_build_nack_msg(l, xmitq); | 
|  | 1317 | break; | 
|  | 1318 | } | 
|  | 1319 |  | 
|  | 1320 | /* Deliver packet */ | 
|  | 1321 | l->rcv_nxt++; | 
|  | 1322 | l->stats.recv_pkts++; | 
|  | 1323 | if (!tipc_data_input(l, skb, l->inputq)) | 
|  | 1324 | rc |= tipc_link_input(l, skb, l->inputq); | 
|  | 1325 | if (unlikely(++l->rcv_unacked >= TIPC_MIN_LINK_WIN)) | 
|  | 1326 | rc |= tipc_link_build_state_msg(l, xmitq); | 
|  | 1327 | if (unlikely(rc & ~TIPC_LINK_SND_STATE)) | 
|  | 1328 | break; | 
|  | 1329 | } while ((skb = __skb_dequeue(defq))); | 
|  | 1330 |  | 
|  | 1331 | return rc; | 
|  | 1332 | drop: | 
|  | 1333 | kfree_skb(skb); | 
|  | 1334 | return rc; | 
|  | 1335 | } | 
|  | 1336 |  | 
|  | 1337 | static void tipc_link_build_proto_msg(struct tipc_link *l, int mtyp, bool probe, | 
|  | 1338 | bool probe_reply, u16 rcvgap, | 
|  | 1339 | int tolerance, int priority, | 
|  | 1340 | struct sk_buff_head *xmitq) | 
|  | 1341 | { | 
|  | 1342 | struct tipc_link *bcl = l->bc_rcvlink; | 
|  | 1343 | struct sk_buff *skb; | 
|  | 1344 | struct tipc_msg *hdr; | 
|  | 1345 | struct sk_buff_head *dfq = &l->deferdq; | 
|  | 1346 | bool node_up = link_is_up(bcl); | 
|  | 1347 | struct tipc_mon_state *mstate = &l->mon_state; | 
|  | 1348 | int dlen = 0; | 
|  | 1349 | void *data; | 
|  | 1350 |  | 
|  | 1351 | /* Don't send protocol message during reset or link failover */ | 
|  | 1352 | if (tipc_link_is_blocked(l)) | 
|  | 1353 | return; | 
|  | 1354 |  | 
|  | 1355 | if (!tipc_link_is_up(l) && (mtyp == STATE_MSG)) | 
|  | 1356 | return; | 
|  | 1357 |  | 
|  | 1358 | if (!skb_queue_empty(dfq)) | 
|  | 1359 | rcvgap = buf_seqno(skb_peek(dfq)) - l->rcv_nxt; | 
|  | 1360 |  | 
|  | 1361 | skb = tipc_msg_create(LINK_PROTOCOL, mtyp, INT_H_SIZE, | 
|  | 1362 | tipc_max_domain_size, l->addr, | 
|  | 1363 | tipc_own_addr(l->net), 0, 0, 0); | 
|  | 1364 | if (!skb) | 
|  | 1365 | return; | 
|  | 1366 |  | 
|  | 1367 | hdr = buf_msg(skb); | 
|  | 1368 | data = msg_data(hdr); | 
|  | 1369 | msg_set_session(hdr, l->session); | 
|  | 1370 | msg_set_bearer_id(hdr, l->bearer_id); | 
|  | 1371 | msg_set_net_plane(hdr, l->net_plane); | 
|  | 1372 | msg_set_next_sent(hdr, l->snd_nxt); | 
|  | 1373 | msg_set_ack(hdr, l->rcv_nxt - 1); | 
|  | 1374 | msg_set_bcast_ack(hdr, bcl->rcv_nxt - 1); | 
|  | 1375 | msg_set_bc_ack_invalid(hdr, !node_up); | 
|  | 1376 | msg_set_last_bcast(hdr, l->bc_sndlink->snd_nxt - 1); | 
|  | 1377 | msg_set_link_tolerance(hdr, tolerance); | 
|  | 1378 | msg_set_linkprio(hdr, priority); | 
|  | 1379 | msg_set_redundant_link(hdr, node_up); | 
|  | 1380 | msg_set_seq_gap(hdr, 0); | 
|  | 1381 | msg_set_seqno(hdr, l->snd_nxt + U16_MAX / 2); | 
|  | 1382 |  | 
|  | 1383 | if (mtyp == STATE_MSG) { | 
|  | 1384 | if (l->peer_caps & TIPC_LINK_PROTO_SEQNO) | 
|  | 1385 | msg_set_seqno(hdr, l->snd_nxt_state++); | 
|  | 1386 | msg_set_seq_gap(hdr, rcvgap); | 
|  | 1387 | msg_set_bc_gap(hdr, link_bc_rcv_gap(bcl)); | 
|  | 1388 | msg_set_probe(hdr, probe); | 
|  | 1389 | msg_set_is_keepalive(hdr, probe || probe_reply); | 
|  | 1390 | tipc_mon_prep(l->net, data, &dlen, mstate, l->bearer_id); | 
|  | 1391 | msg_set_size(hdr, INT_H_SIZE + dlen); | 
|  | 1392 | skb_trim(skb, INT_H_SIZE + dlen); | 
|  | 1393 | l->stats.sent_states++; | 
|  | 1394 | l->rcv_unacked = 0; | 
|  | 1395 | } else { | 
|  | 1396 | /* RESET_MSG or ACTIVATE_MSG */ | 
|  | 1397 | msg_set_max_pkt(hdr, l->advertised_mtu); | 
|  | 1398 | strcpy(data, l->if_name); | 
|  | 1399 | msg_set_size(hdr, INT_H_SIZE + TIPC_MAX_IF_NAME); | 
|  | 1400 | skb_trim(skb, INT_H_SIZE + TIPC_MAX_IF_NAME); | 
|  | 1401 | } | 
|  | 1402 | if (probe) | 
|  | 1403 | l->stats.sent_probes++; | 
|  | 1404 | if (rcvgap) | 
|  | 1405 | l->stats.sent_nacks++; | 
|  | 1406 | skb->priority = TC_PRIO_CONTROL; | 
|  | 1407 | __skb_queue_tail(xmitq, skb); | 
|  | 1408 | } | 
|  | 1409 |  | 
|  | 1410 | void tipc_link_create_dummy_tnl_msg(struct tipc_link *l, | 
|  | 1411 | struct sk_buff_head *xmitq) | 
|  | 1412 | { | 
|  | 1413 | u32 onode = tipc_own_addr(l->net); | 
|  | 1414 | struct tipc_msg *hdr, *ihdr; | 
|  | 1415 | struct sk_buff_head tnlq; | 
|  | 1416 | struct sk_buff *skb; | 
|  | 1417 | u32 dnode = l->addr; | 
|  | 1418 |  | 
|  | 1419 | skb_queue_head_init(&tnlq); | 
|  | 1420 | skb = tipc_msg_create(TUNNEL_PROTOCOL, FAILOVER_MSG, | 
|  | 1421 | INT_H_SIZE, BASIC_H_SIZE, | 
|  | 1422 | dnode, onode, 0, 0, 0); | 
|  | 1423 | if (!skb) { | 
|  | 1424 | pr_warn("%sunable to create tunnel packet\n", link_co_err); | 
|  | 1425 | return; | 
|  | 1426 | } | 
|  | 1427 |  | 
|  | 1428 | hdr = buf_msg(skb); | 
|  | 1429 | msg_set_msgcnt(hdr, 1); | 
|  | 1430 | msg_set_bearer_id(hdr, l->peer_bearer_id); | 
|  | 1431 |  | 
|  | 1432 | ihdr = (struct tipc_msg *)msg_data(hdr); | 
|  | 1433 | tipc_msg_init(onode, ihdr, TIPC_LOW_IMPORTANCE, TIPC_DIRECT_MSG, | 
|  | 1434 | BASIC_H_SIZE, dnode); | 
|  | 1435 | msg_set_errcode(ihdr, TIPC_ERR_NO_PORT); | 
|  | 1436 | __skb_queue_tail(&tnlq, skb); | 
|  | 1437 | tipc_link_xmit(l, &tnlq, xmitq); | 
|  | 1438 | } | 
|  | 1439 |  | 
|  | 1440 | /* tipc_link_tnl_prepare(): prepare and return a list of tunnel packets | 
|  | 1441 | * with contents of the link's transmit and backlog queues. | 
|  | 1442 | */ | 
|  | 1443 | void tipc_link_tnl_prepare(struct tipc_link *l, struct tipc_link *tnl, | 
|  | 1444 | int mtyp, struct sk_buff_head *xmitq) | 
|  | 1445 | { | 
|  | 1446 | struct sk_buff *skb, *tnlskb; | 
|  | 1447 | struct tipc_msg *hdr, tnlhdr; | 
|  | 1448 | struct sk_buff_head *queue = &l->transmq; | 
|  | 1449 | struct sk_buff_head tmpxq, tnlq; | 
|  | 1450 | u16 pktlen, pktcnt, seqno = l->snd_nxt; | 
|  | 1451 |  | 
|  | 1452 | if (!tnl) | 
|  | 1453 | return; | 
|  | 1454 |  | 
|  | 1455 | skb_queue_head_init(&tnlq); | 
|  | 1456 | skb_queue_head_init(&tmpxq); | 
|  | 1457 |  | 
|  | 1458 | /* At least one packet required for safe algorithm => add dummy */ | 
|  | 1459 | skb = tipc_msg_create(TIPC_LOW_IMPORTANCE, TIPC_DIRECT_MSG, | 
|  | 1460 | BASIC_H_SIZE, 0, l->addr, tipc_own_addr(l->net), | 
|  | 1461 | 0, 0, TIPC_ERR_NO_PORT); | 
|  | 1462 | if (!skb) { | 
|  | 1463 | pr_warn("%sunable to create tunnel packet\n", link_co_err); | 
|  | 1464 | return; | 
|  | 1465 | } | 
|  | 1466 | skb_queue_tail(&tnlq, skb); | 
|  | 1467 | tipc_link_xmit(l, &tnlq, &tmpxq); | 
|  | 1468 | __skb_queue_purge(&tmpxq); | 
|  | 1469 |  | 
|  | 1470 | /* Initialize reusable tunnel packet header */ | 
|  | 1471 | tipc_msg_init(tipc_own_addr(l->net), &tnlhdr, TUNNEL_PROTOCOL, | 
|  | 1472 | mtyp, INT_H_SIZE, l->addr); | 
|  | 1473 | pktcnt = skb_queue_len(&l->transmq) + skb_queue_len(&l->backlogq); | 
|  | 1474 | msg_set_msgcnt(&tnlhdr, pktcnt); | 
|  | 1475 | msg_set_bearer_id(&tnlhdr, l->peer_bearer_id); | 
|  | 1476 | tnl: | 
|  | 1477 | /* Wrap each packet into a tunnel packet */ | 
|  | 1478 | skb_queue_walk(queue, skb) { | 
|  | 1479 | hdr = buf_msg(skb); | 
|  | 1480 | if (queue == &l->backlogq) | 
|  | 1481 | msg_set_seqno(hdr, seqno++); | 
|  | 1482 | pktlen = msg_size(hdr); | 
|  | 1483 | msg_set_size(&tnlhdr, pktlen + INT_H_SIZE); | 
|  | 1484 | tnlskb = tipc_buf_acquire(pktlen + INT_H_SIZE, GFP_ATOMIC); | 
|  | 1485 | if (!tnlskb) { | 
|  | 1486 | pr_warn("%sunable to send packet\n", link_co_err); | 
|  | 1487 | return; | 
|  | 1488 | } | 
|  | 1489 | skb_copy_to_linear_data(tnlskb, &tnlhdr, INT_H_SIZE); | 
|  | 1490 | skb_copy_to_linear_data_offset(tnlskb, INT_H_SIZE, hdr, pktlen); | 
|  | 1491 | __skb_queue_tail(&tnlq, tnlskb); | 
|  | 1492 | } | 
|  | 1493 | if (queue != &l->backlogq) { | 
|  | 1494 | queue = &l->backlogq; | 
|  | 1495 | goto tnl; | 
|  | 1496 | } | 
|  | 1497 |  | 
|  | 1498 | tipc_link_xmit(tnl, &tnlq, xmitq); | 
|  | 1499 |  | 
|  | 1500 | if (mtyp == FAILOVER_MSG) { | 
|  | 1501 | tnl->drop_point = l->rcv_nxt; | 
|  | 1502 | tnl->failover_reasm_skb = l->reasm_buf; | 
|  | 1503 | l->reasm_buf = NULL; | 
|  | 1504 | } | 
|  | 1505 | } | 
|  | 1506 |  | 
|  | 1507 | /* tipc_link_validate_msg(): validate message against current link state | 
|  | 1508 | * Returns true if message should be accepted, otherwise false | 
|  | 1509 | */ | 
|  | 1510 | bool tipc_link_validate_msg(struct tipc_link *l, struct tipc_msg *hdr) | 
|  | 1511 | { | 
|  | 1512 | u16 curr_session = l->peer_session; | 
|  | 1513 | u16 session = msg_session(hdr); | 
|  | 1514 | int mtyp = msg_type(hdr); | 
|  | 1515 |  | 
|  | 1516 | if (msg_user(hdr) != LINK_PROTOCOL) | 
|  | 1517 | return true; | 
|  | 1518 |  | 
|  | 1519 | switch (mtyp) { | 
|  | 1520 | case RESET_MSG: | 
|  | 1521 | if (!l->in_session) | 
|  | 1522 | return true; | 
|  | 1523 | /* Accept only RESET with new session number */ | 
|  | 1524 | return more(session, curr_session); | 
|  | 1525 | case ACTIVATE_MSG: | 
|  | 1526 | if (!l->in_session) | 
|  | 1527 | return true; | 
|  | 1528 | /* Accept only ACTIVATE with new or current session number */ | 
|  | 1529 | return !less(session, curr_session); | 
|  | 1530 | case STATE_MSG: | 
|  | 1531 | /* Accept only STATE with current session number */ | 
|  | 1532 | if (!l->in_session) | 
|  | 1533 | return false; | 
|  | 1534 | if (session != curr_session) | 
|  | 1535 | return false; | 
|  | 1536 | /* Extra sanity check */ | 
|  | 1537 | if (!link_is_up(l) && msg_ack(hdr)) | 
|  | 1538 | return false; | 
|  | 1539 | if (!(l->peer_caps & TIPC_LINK_PROTO_SEQNO)) | 
|  | 1540 | return true; | 
|  | 1541 | /* Accept only STATE with new sequence number */ | 
|  | 1542 | return !less(msg_seqno(hdr), l->rcv_nxt_state); | 
|  | 1543 | default: | 
|  | 1544 | return false; | 
|  | 1545 | } | 
|  | 1546 | } | 
|  | 1547 |  | 
|  | 1548 | /* tipc_link_proto_rcv(): receive link level protocol message : | 
|  | 1549 | * Note that network plane id propagates through the network, and may | 
|  | 1550 | * change at any time. The node with lowest numerical id determines | 
|  | 1551 | * network plane | 
|  | 1552 | */ | 
|  | 1553 | static int tipc_link_proto_rcv(struct tipc_link *l, struct sk_buff *skb, | 
|  | 1554 | struct sk_buff_head *xmitq) | 
|  | 1555 | { | 
|  | 1556 | struct tipc_msg *hdr = buf_msg(skb); | 
|  | 1557 | u16 rcvgap = 0; | 
|  | 1558 | u16 ack = msg_ack(hdr); | 
|  | 1559 | u16 gap = msg_seq_gap(hdr); | 
|  | 1560 | u16 peers_snd_nxt =  msg_next_sent(hdr); | 
|  | 1561 | u16 peers_tol = msg_link_tolerance(hdr); | 
|  | 1562 | u16 peers_prio = msg_linkprio(hdr); | 
|  | 1563 | u16 rcv_nxt = l->rcv_nxt; | 
|  | 1564 | u16 dlen = msg_data_sz(hdr); | 
|  | 1565 | int mtyp = msg_type(hdr); | 
|  | 1566 | bool reply = msg_probe(hdr); | 
|  | 1567 | void *data; | 
|  | 1568 | char *if_name; | 
|  | 1569 | int rc = 0; | 
|  | 1570 |  | 
|  | 1571 | if (tipc_link_is_blocked(l) || !xmitq) | 
|  | 1572 | goto exit; | 
|  | 1573 |  | 
|  | 1574 | if (tipc_own_addr(l->net) > msg_prevnode(hdr)) | 
|  | 1575 | l->net_plane = msg_net_plane(hdr); | 
|  | 1576 |  | 
|  | 1577 | skb_linearize(skb); | 
|  | 1578 | hdr = buf_msg(skb); | 
|  | 1579 | data = msg_data(hdr); | 
|  | 1580 |  | 
|  | 1581 | if (!tipc_link_validate_msg(l, hdr)) | 
|  | 1582 | goto exit; | 
|  | 1583 |  | 
|  | 1584 | switch (mtyp) { | 
|  | 1585 | case RESET_MSG: | 
|  | 1586 | case ACTIVATE_MSG: | 
|  | 1587 | /* Complete own link name with peer's interface name */ | 
|  | 1588 | if_name =  strrchr(l->name, ':') + 1; | 
|  | 1589 | if (sizeof(l->name) - (if_name - l->name) <= TIPC_MAX_IF_NAME) | 
|  | 1590 | break; | 
|  | 1591 | if (msg_data_sz(hdr) < TIPC_MAX_IF_NAME) | 
|  | 1592 | break; | 
|  | 1593 | strncpy(if_name, data, TIPC_MAX_IF_NAME); | 
|  | 1594 |  | 
|  | 1595 | /* Update own tolerance if peer indicates a non-zero value */ | 
|  | 1596 | if (in_range(peers_tol, TIPC_MIN_LINK_TOL, TIPC_MAX_LINK_TOL)) { | 
|  | 1597 | l->tolerance = peers_tol; | 
|  | 1598 | l->bc_rcvlink->tolerance = peers_tol; | 
|  | 1599 | } | 
|  | 1600 | /* Update own priority if peer's priority is higher */ | 
|  | 1601 | if (in_range(peers_prio, l->priority + 1, TIPC_MAX_LINK_PRI)) | 
|  | 1602 | l->priority = peers_prio; | 
|  | 1603 |  | 
|  | 1604 | /* If peer is going down we want full re-establish cycle */ | 
|  | 1605 | if (msg_peer_stopping(hdr)) { | 
|  | 1606 | rc = tipc_link_fsm_evt(l, LINK_FAILURE_EVT); | 
|  | 1607 | break; | 
|  | 1608 | } | 
|  | 1609 | /* ACTIVATE_MSG serves as PEER_RESET if link is already down */ | 
|  | 1610 | if (mtyp == RESET_MSG || !link_is_up(l)) | 
|  | 1611 | rc = tipc_link_fsm_evt(l, LINK_PEER_RESET_EVT); | 
|  | 1612 |  | 
|  | 1613 | /* ACTIVATE_MSG takes up link if it was already locally reset */ | 
|  | 1614 | if (mtyp == ACTIVATE_MSG && l->state == LINK_ESTABLISHING) | 
|  | 1615 | rc = TIPC_LINK_UP_EVT; | 
|  | 1616 |  | 
|  | 1617 | l->peer_session = msg_session(hdr); | 
|  | 1618 | l->in_session = true; | 
|  | 1619 | l->peer_bearer_id = msg_bearer_id(hdr); | 
|  | 1620 | if (l->mtu > msg_max_pkt(hdr)) | 
|  | 1621 | l->mtu = msg_max_pkt(hdr); | 
|  | 1622 | break; | 
|  | 1623 |  | 
|  | 1624 | case STATE_MSG: | 
|  | 1625 | l->rcv_nxt_state = msg_seqno(hdr) + 1; | 
|  | 1626 |  | 
|  | 1627 | /* Update own tolerance if peer indicates a non-zero value */ | 
|  | 1628 | if (in_range(peers_tol, TIPC_MIN_LINK_TOL, TIPC_MAX_LINK_TOL)) { | 
|  | 1629 | l->tolerance = peers_tol; | 
|  | 1630 | l->bc_rcvlink->tolerance = peers_tol; | 
|  | 1631 | } | 
|  | 1632 | /* Update own prio if peer indicates a different value */ | 
|  | 1633 | if ((peers_prio != l->priority) && | 
|  | 1634 | in_range(peers_prio, 1, TIPC_MAX_LINK_PRI)) { | 
|  | 1635 | l->priority = peers_prio; | 
|  | 1636 | rc = tipc_link_fsm_evt(l, LINK_FAILURE_EVT); | 
|  | 1637 | } | 
|  | 1638 |  | 
|  | 1639 | l->silent_intv_cnt = 0; | 
|  | 1640 | l->stats.recv_states++; | 
|  | 1641 | if (msg_probe(hdr)) | 
|  | 1642 | l->stats.recv_probes++; | 
|  | 1643 |  | 
|  | 1644 | if (!link_is_up(l)) { | 
|  | 1645 | if (l->state == LINK_ESTABLISHING) | 
|  | 1646 | rc = TIPC_LINK_UP_EVT; | 
|  | 1647 | break; | 
|  | 1648 | } | 
|  | 1649 | tipc_mon_rcv(l->net, data, dlen, l->addr, | 
|  | 1650 | &l->mon_state, l->bearer_id); | 
|  | 1651 |  | 
|  | 1652 | /* Send NACK if peer has sent pkts we haven't received yet */ | 
|  | 1653 | if (more(peers_snd_nxt, rcv_nxt) && !tipc_link_is_synching(l)) | 
|  | 1654 | rcvgap = peers_snd_nxt - l->rcv_nxt; | 
|  | 1655 | if (rcvgap || reply) | 
|  | 1656 | tipc_link_build_proto_msg(l, STATE_MSG, 0, reply, | 
|  | 1657 | rcvgap, 0, 0, xmitq); | 
|  | 1658 | tipc_link_release_pkts(l, ack); | 
|  | 1659 |  | 
|  | 1660 | /* If NACK, retransmit will now start at right position */ | 
|  | 1661 | if (gap) { | 
|  | 1662 | rc = tipc_link_retrans(l, l, ack + 1, ack + gap, xmitq); | 
|  | 1663 | l->stats.recv_nacks++; | 
|  | 1664 | } | 
|  | 1665 |  | 
|  | 1666 | tipc_link_advance_backlog(l, xmitq); | 
|  | 1667 | if (unlikely(!skb_queue_empty(&l->wakeupq))) | 
|  | 1668 | link_prepare_wakeup(l); | 
|  | 1669 | } | 
|  | 1670 | exit: | 
|  | 1671 | kfree_skb(skb); | 
|  | 1672 | return rc; | 
|  | 1673 | } | 
|  | 1674 |  | 
|  | 1675 | /* tipc_link_build_bc_proto_msg() - create broadcast protocol message | 
|  | 1676 | */ | 
|  | 1677 | static bool tipc_link_build_bc_proto_msg(struct tipc_link *l, bool bcast, | 
|  | 1678 | u16 peers_snd_nxt, | 
|  | 1679 | struct sk_buff_head *xmitq) | 
|  | 1680 | { | 
|  | 1681 | struct sk_buff *skb; | 
|  | 1682 | struct tipc_msg *hdr; | 
|  | 1683 | struct sk_buff *dfrd_skb = skb_peek(&l->deferdq); | 
|  | 1684 | u16 ack = l->rcv_nxt - 1; | 
|  | 1685 | u16 gap_to = peers_snd_nxt - 1; | 
|  | 1686 |  | 
|  | 1687 | skb = tipc_msg_create(BCAST_PROTOCOL, STATE_MSG, INT_H_SIZE, | 
|  | 1688 | 0, l->addr, tipc_own_addr(l->net), 0, 0, 0); | 
|  | 1689 | if (!skb) | 
|  | 1690 | return false; | 
|  | 1691 | hdr = buf_msg(skb); | 
|  | 1692 | msg_set_last_bcast(hdr, l->bc_sndlink->snd_nxt - 1); | 
|  | 1693 | msg_set_bcast_ack(hdr, ack); | 
|  | 1694 | msg_set_bcgap_after(hdr, ack); | 
|  | 1695 | if (dfrd_skb) | 
|  | 1696 | gap_to = buf_seqno(dfrd_skb) - 1; | 
|  | 1697 | msg_set_bcgap_to(hdr, gap_to); | 
|  | 1698 | msg_set_non_seq(hdr, bcast); | 
|  | 1699 | __skb_queue_tail(xmitq, skb); | 
|  | 1700 | return true; | 
|  | 1701 | } | 
|  | 1702 |  | 
|  | 1703 | /* tipc_link_build_bc_init_msg() - synchronize broadcast link endpoints. | 
|  | 1704 | * | 
|  | 1705 | * Give a newly added peer node the sequence number where it should | 
|  | 1706 | * start receiving and acking broadcast packets. | 
|  | 1707 | */ | 
|  | 1708 | static void tipc_link_build_bc_init_msg(struct tipc_link *l, | 
|  | 1709 | struct sk_buff_head *xmitq) | 
|  | 1710 | { | 
|  | 1711 | struct sk_buff_head list; | 
|  | 1712 |  | 
|  | 1713 | __skb_queue_head_init(&list); | 
|  | 1714 | if (!tipc_link_build_bc_proto_msg(l->bc_rcvlink, false, 0, &list)) | 
|  | 1715 | return; | 
|  | 1716 | msg_set_bc_ack_invalid(buf_msg(skb_peek(&list)), true); | 
|  | 1717 | tipc_link_xmit(l, &list, xmitq); | 
|  | 1718 | } | 
|  | 1719 |  | 
|  | 1720 | /* tipc_link_bc_init_rcv - receive initial broadcast synch data from peer | 
|  | 1721 | */ | 
|  | 1722 | void tipc_link_bc_init_rcv(struct tipc_link *l, struct tipc_msg *hdr) | 
|  | 1723 | { | 
|  | 1724 | int mtyp = msg_type(hdr); | 
|  | 1725 | u16 peers_snd_nxt = msg_bc_snd_nxt(hdr); | 
|  | 1726 |  | 
|  | 1727 | if (link_is_up(l)) | 
|  | 1728 | return; | 
|  | 1729 |  | 
|  | 1730 | if (msg_user(hdr) == BCAST_PROTOCOL) { | 
|  | 1731 | l->rcv_nxt = peers_snd_nxt; | 
|  | 1732 | l->state = LINK_ESTABLISHED; | 
|  | 1733 | return; | 
|  | 1734 | } | 
|  | 1735 |  | 
|  | 1736 | if (l->peer_caps & TIPC_BCAST_SYNCH) | 
|  | 1737 | return; | 
|  | 1738 |  | 
|  | 1739 | if (msg_peer_node_is_up(hdr)) | 
|  | 1740 | return; | 
|  | 1741 |  | 
|  | 1742 | /* Compatibility: accept older, less safe initial synch data */ | 
|  | 1743 | if ((mtyp == RESET_MSG) || (mtyp == ACTIVATE_MSG)) | 
|  | 1744 | l->rcv_nxt = peers_snd_nxt; | 
|  | 1745 | } | 
|  | 1746 |  | 
|  | 1747 | /* link_bc_retr eval()- check if the indicated range can be retransmitted now | 
|  | 1748 | * - Adjust permitted range if there is overlap with previous retransmission | 
|  | 1749 | */ | 
|  | 1750 | static bool link_bc_retr_eval(struct tipc_link *l, u16 *from, u16 *to) | 
|  | 1751 | { | 
|  | 1752 | unsigned long elapsed = jiffies_to_msecs(jiffies - l->prev_retr); | 
|  | 1753 |  | 
|  | 1754 | if (less(*to, *from)) | 
|  | 1755 | return false; | 
|  | 1756 |  | 
|  | 1757 | /* New retransmission request */ | 
|  | 1758 | if ((elapsed > TIPC_BC_RETR_LIMIT) || | 
|  | 1759 | less(*to, l->prev_from) || more(*from, l->prev_to)) { | 
|  | 1760 | l->prev_from = *from; | 
|  | 1761 | l->prev_to = *to; | 
|  | 1762 | l->prev_retr = jiffies; | 
|  | 1763 | return true; | 
|  | 1764 | } | 
|  | 1765 |  | 
|  | 1766 | /* Inside range of previous retransmit */ | 
|  | 1767 | if (!less(*from, l->prev_from) && !more(*to, l->prev_to)) | 
|  | 1768 | return false; | 
|  | 1769 |  | 
|  | 1770 | /* Fully or partially outside previous range => exclude overlap */ | 
|  | 1771 | if (less(*from, l->prev_from)) { | 
|  | 1772 | *to = l->prev_from - 1; | 
|  | 1773 | l->prev_from = *from; | 
|  | 1774 | } | 
|  | 1775 | if (more(*to, l->prev_to)) { | 
|  | 1776 | *from = l->prev_to + 1; | 
|  | 1777 | l->prev_to = *to; | 
|  | 1778 | } | 
|  | 1779 | l->prev_retr = jiffies; | 
|  | 1780 | return true; | 
|  | 1781 | } | 
|  | 1782 |  | 
|  | 1783 | /* tipc_link_bc_sync_rcv - update rcv link according to peer's send state | 
|  | 1784 | */ | 
|  | 1785 | int tipc_link_bc_sync_rcv(struct tipc_link *l, struct tipc_msg *hdr, | 
|  | 1786 | struct sk_buff_head *xmitq) | 
|  | 1787 | { | 
|  | 1788 | struct tipc_link *snd_l = l->bc_sndlink; | 
|  | 1789 | u16 peers_snd_nxt = msg_bc_snd_nxt(hdr); | 
|  | 1790 | u16 from = msg_bcast_ack(hdr) + 1; | 
|  | 1791 | u16 to = from + msg_bc_gap(hdr) - 1; | 
|  | 1792 | int rc = 0; | 
|  | 1793 |  | 
|  | 1794 | if (!link_is_up(l)) | 
|  | 1795 | return rc; | 
|  | 1796 |  | 
|  | 1797 | if (!msg_peer_node_is_up(hdr)) | 
|  | 1798 | return rc; | 
|  | 1799 |  | 
|  | 1800 | /* Open when peer ackowledges our bcast init msg (pkt #1) */ | 
|  | 1801 | if (msg_ack(hdr)) | 
|  | 1802 | l->bc_peer_is_up = true; | 
|  | 1803 |  | 
|  | 1804 | if (!l->bc_peer_is_up) | 
|  | 1805 | return rc; | 
|  | 1806 |  | 
|  | 1807 | l->stats.recv_nacks++; | 
|  | 1808 |  | 
|  | 1809 | /* Ignore if peers_snd_nxt goes beyond receive window */ | 
|  | 1810 | if (more(peers_snd_nxt, l->rcv_nxt + l->window)) | 
|  | 1811 | return rc; | 
|  | 1812 |  | 
|  | 1813 | if (link_bc_retr_eval(snd_l, &from, &to)) | 
|  | 1814 | rc = tipc_link_retrans(snd_l, l, from, to, xmitq); | 
|  | 1815 |  | 
|  | 1816 | l->snd_nxt = peers_snd_nxt; | 
|  | 1817 | if (link_bc_rcv_gap(l)) | 
|  | 1818 | rc |= TIPC_LINK_SND_STATE; | 
|  | 1819 |  | 
|  | 1820 | /* Return now if sender supports nack via STATE messages */ | 
|  | 1821 | if (l->peer_caps & TIPC_BCAST_STATE_NACK) | 
|  | 1822 | return rc; | 
|  | 1823 |  | 
|  | 1824 | /* Otherwise, be backwards compatible */ | 
|  | 1825 |  | 
|  | 1826 | if (!more(peers_snd_nxt, l->rcv_nxt)) { | 
|  | 1827 | l->nack_state = BC_NACK_SND_CONDITIONAL; | 
|  | 1828 | return 0; | 
|  | 1829 | } | 
|  | 1830 |  | 
|  | 1831 | /* Don't NACK if one was recently sent or peeked */ | 
|  | 1832 | if (l->nack_state == BC_NACK_SND_SUPPRESS) { | 
|  | 1833 | l->nack_state = BC_NACK_SND_UNCONDITIONAL; | 
|  | 1834 | return 0; | 
|  | 1835 | } | 
|  | 1836 |  | 
|  | 1837 | /* Conditionally delay NACK sending until next synch rcv */ | 
|  | 1838 | if (l->nack_state == BC_NACK_SND_CONDITIONAL) { | 
|  | 1839 | l->nack_state = BC_NACK_SND_UNCONDITIONAL; | 
|  | 1840 | if ((peers_snd_nxt - l->rcv_nxt) < TIPC_MIN_LINK_WIN) | 
|  | 1841 | return 0; | 
|  | 1842 | } | 
|  | 1843 |  | 
|  | 1844 | /* Send NACK now but suppress next one */ | 
|  | 1845 | tipc_link_build_bc_proto_msg(l, true, peers_snd_nxt, xmitq); | 
|  | 1846 | l->nack_state = BC_NACK_SND_SUPPRESS; | 
|  | 1847 | return 0; | 
|  | 1848 | } | 
|  | 1849 |  | 
|  | 1850 | void tipc_link_bc_ack_rcv(struct tipc_link *l, u16 acked, | 
|  | 1851 | struct sk_buff_head *xmitq) | 
|  | 1852 | { | 
|  | 1853 | struct sk_buff *skb, *tmp; | 
|  | 1854 | struct tipc_link *snd_l = l->bc_sndlink; | 
|  | 1855 |  | 
|  | 1856 | if (!link_is_up(l) || !l->bc_peer_is_up) | 
|  | 1857 | return; | 
|  | 1858 |  | 
|  | 1859 | if (!more(acked, l->acked)) | 
|  | 1860 | return; | 
|  | 1861 |  | 
|  | 1862 | /* Skip over packets peer has already acked */ | 
|  | 1863 | skb_queue_walk(&snd_l->transmq, skb) { | 
|  | 1864 | if (more(buf_seqno(skb), l->acked)) | 
|  | 1865 | break; | 
|  | 1866 | } | 
|  | 1867 |  | 
|  | 1868 | /* Update/release the packets peer is acking now */ | 
|  | 1869 | skb_queue_walk_from_safe(&snd_l->transmq, skb, tmp) { | 
|  | 1870 | if (more(buf_seqno(skb), acked)) | 
|  | 1871 | break; | 
|  | 1872 | if (!--TIPC_SKB_CB(skb)->ackers) { | 
|  | 1873 | __skb_unlink(skb, &snd_l->transmq); | 
|  | 1874 | kfree_skb(skb); | 
|  | 1875 | } | 
|  | 1876 | } | 
|  | 1877 | l->acked = acked; | 
|  | 1878 | tipc_link_advance_backlog(snd_l, xmitq); | 
|  | 1879 | if (unlikely(!skb_queue_empty(&snd_l->wakeupq))) | 
|  | 1880 | link_prepare_wakeup(snd_l); | 
|  | 1881 | } | 
|  | 1882 |  | 
|  | 1883 | /* tipc_link_bc_nack_rcv(): receive broadcast nack message | 
|  | 1884 | * This function is here for backwards compatibility, since | 
|  | 1885 | * no BCAST_PROTOCOL/STATE messages occur from TIPC v2.5. | 
|  | 1886 | */ | 
|  | 1887 | int tipc_link_bc_nack_rcv(struct tipc_link *l, struct sk_buff *skb, | 
|  | 1888 | struct sk_buff_head *xmitq) | 
|  | 1889 | { | 
|  | 1890 | struct tipc_msg *hdr = buf_msg(skb); | 
|  | 1891 | u32 dnode = msg_destnode(hdr); | 
|  | 1892 | int mtyp = msg_type(hdr); | 
|  | 1893 | u16 acked = msg_bcast_ack(hdr); | 
|  | 1894 | u16 from = acked + 1; | 
|  | 1895 | u16 to = msg_bcgap_to(hdr); | 
|  | 1896 | u16 peers_snd_nxt = to + 1; | 
|  | 1897 | int rc = 0; | 
|  | 1898 |  | 
|  | 1899 | kfree_skb(skb); | 
|  | 1900 |  | 
|  | 1901 | if (!tipc_link_is_up(l) || !l->bc_peer_is_up) | 
|  | 1902 | return 0; | 
|  | 1903 |  | 
|  | 1904 | if (mtyp != STATE_MSG) | 
|  | 1905 | return 0; | 
|  | 1906 |  | 
|  | 1907 | if (dnode == tipc_own_addr(l->net)) { | 
|  | 1908 | tipc_link_bc_ack_rcv(l, acked, xmitq); | 
|  | 1909 | rc = tipc_link_retrans(l->bc_sndlink, l, from, to, xmitq); | 
|  | 1910 | l->stats.recv_nacks++; | 
|  | 1911 | return rc; | 
|  | 1912 | } | 
|  | 1913 |  | 
|  | 1914 | /* Msg for other node => suppress own NACK at next sync if applicable */ | 
|  | 1915 | if (more(peers_snd_nxt, l->rcv_nxt) && !less(l->rcv_nxt, from)) | 
|  | 1916 | l->nack_state = BC_NACK_SND_SUPPRESS; | 
|  | 1917 |  | 
|  | 1918 | return 0; | 
|  | 1919 | } | 
|  | 1920 |  | 
|  | 1921 | void tipc_link_set_queue_limits(struct tipc_link *l, u32 win) | 
|  | 1922 | { | 
|  | 1923 | int max_bulk = TIPC_MAX_PUBL / (l->mtu / ITEM_SIZE); | 
|  | 1924 |  | 
|  | 1925 | l->window = win; | 
|  | 1926 | l->backlog[TIPC_LOW_IMPORTANCE].limit      = max_t(u16, 50, win); | 
|  | 1927 | l->backlog[TIPC_MEDIUM_IMPORTANCE].limit   = max_t(u16, 100, win * 2); | 
|  | 1928 | l->backlog[TIPC_HIGH_IMPORTANCE].limit     = max_t(u16, 150, win * 3); | 
|  | 1929 | l->backlog[TIPC_CRITICAL_IMPORTANCE].limit = max_t(u16, 200, win * 4); | 
|  | 1930 | l->backlog[TIPC_SYSTEM_IMPORTANCE].limit   = max_bulk; | 
|  | 1931 | } | 
|  | 1932 |  | 
|  | 1933 | /** | 
|  | 1934 | * link_reset_stats - reset link statistics | 
|  | 1935 | * @l: pointer to link | 
|  | 1936 | */ | 
|  | 1937 | void tipc_link_reset_stats(struct tipc_link *l) | 
|  | 1938 | { | 
|  | 1939 | memset(&l->stats, 0, sizeof(l->stats)); | 
|  | 1940 | } | 
|  | 1941 |  | 
|  | 1942 | static void link_print(struct tipc_link *l, const char *str) | 
|  | 1943 | { | 
|  | 1944 | struct sk_buff *hskb = skb_peek(&l->transmq); | 
|  | 1945 | u16 head = hskb ? msg_seqno(buf_msg(hskb)) : l->snd_nxt - 1; | 
|  | 1946 | u16 tail = l->snd_nxt - 1; | 
|  | 1947 |  | 
|  | 1948 | pr_info("%s Link <%s> state %x\n", str, l->name, l->state); | 
|  | 1949 | pr_info("XMTQ: %u [%u-%u], BKLGQ: %u, SNDNX: %u, RCVNX: %u\n", | 
|  | 1950 | skb_queue_len(&l->transmq), head, tail, | 
|  | 1951 | skb_queue_len(&l->backlogq), l->snd_nxt, l->rcv_nxt); | 
|  | 1952 | } | 
|  | 1953 |  | 
|  | 1954 | /* Parse and validate nested (link) properties valid for media, bearer and link | 
|  | 1955 | */ | 
|  | 1956 | int tipc_nl_parse_link_prop(struct nlattr *prop, struct nlattr *props[]) | 
|  | 1957 | { | 
|  | 1958 | int err; | 
|  | 1959 |  | 
|  | 1960 | err = nla_parse_nested(props, TIPC_NLA_PROP_MAX, prop, | 
|  | 1961 | tipc_nl_prop_policy, NULL); | 
|  | 1962 | if (err) | 
|  | 1963 | return err; | 
|  | 1964 |  | 
|  | 1965 | if (props[TIPC_NLA_PROP_PRIO]) { | 
|  | 1966 | u32 prio; | 
|  | 1967 |  | 
|  | 1968 | prio = nla_get_u32(props[TIPC_NLA_PROP_PRIO]); | 
|  | 1969 | if (prio > TIPC_MAX_LINK_PRI) | 
|  | 1970 | return -EINVAL; | 
|  | 1971 | } | 
|  | 1972 |  | 
|  | 1973 | if (props[TIPC_NLA_PROP_TOL]) { | 
|  | 1974 | u32 tol; | 
|  | 1975 |  | 
|  | 1976 | tol = nla_get_u32(props[TIPC_NLA_PROP_TOL]); | 
|  | 1977 | if ((tol < TIPC_MIN_LINK_TOL) || (tol > TIPC_MAX_LINK_TOL)) | 
|  | 1978 | return -EINVAL; | 
|  | 1979 | } | 
|  | 1980 |  | 
|  | 1981 | if (props[TIPC_NLA_PROP_WIN]) { | 
|  | 1982 | u32 win; | 
|  | 1983 |  | 
|  | 1984 | win = nla_get_u32(props[TIPC_NLA_PROP_WIN]); | 
|  | 1985 | if ((win < TIPC_MIN_LINK_WIN) || (win > TIPC_MAX_LINK_WIN)) | 
|  | 1986 | return -EINVAL; | 
|  | 1987 | } | 
|  | 1988 |  | 
|  | 1989 | return 0; | 
|  | 1990 | } | 
|  | 1991 |  | 
|  | 1992 | static int __tipc_nl_add_stats(struct sk_buff *skb, struct tipc_stats *s) | 
|  | 1993 | { | 
|  | 1994 | int i; | 
|  | 1995 | struct nlattr *stats; | 
|  | 1996 |  | 
|  | 1997 | struct nla_map { | 
|  | 1998 | u32 key; | 
|  | 1999 | u32 val; | 
|  | 2000 | }; | 
|  | 2001 |  | 
|  | 2002 | struct nla_map map[] = { | 
|  | 2003 | {TIPC_NLA_STATS_RX_INFO, 0}, | 
|  | 2004 | {TIPC_NLA_STATS_RX_FRAGMENTS, s->recv_fragments}, | 
|  | 2005 | {TIPC_NLA_STATS_RX_FRAGMENTED, s->recv_fragmented}, | 
|  | 2006 | {TIPC_NLA_STATS_RX_BUNDLES, s->recv_bundles}, | 
|  | 2007 | {TIPC_NLA_STATS_RX_BUNDLED, s->recv_bundled}, | 
|  | 2008 | {TIPC_NLA_STATS_TX_INFO, 0}, | 
|  | 2009 | {TIPC_NLA_STATS_TX_FRAGMENTS, s->sent_fragments}, | 
|  | 2010 | {TIPC_NLA_STATS_TX_FRAGMENTED, s->sent_fragmented}, | 
|  | 2011 | {TIPC_NLA_STATS_TX_BUNDLES, s->sent_bundles}, | 
|  | 2012 | {TIPC_NLA_STATS_TX_BUNDLED, s->sent_bundled}, | 
|  | 2013 | {TIPC_NLA_STATS_MSG_PROF_TOT, (s->msg_length_counts) ? | 
|  | 2014 | s->msg_length_counts : 1}, | 
|  | 2015 | {TIPC_NLA_STATS_MSG_LEN_CNT, s->msg_length_counts}, | 
|  | 2016 | {TIPC_NLA_STATS_MSG_LEN_TOT, s->msg_lengths_total}, | 
|  | 2017 | {TIPC_NLA_STATS_MSG_LEN_P0, s->msg_length_profile[0]}, | 
|  | 2018 | {TIPC_NLA_STATS_MSG_LEN_P1, s->msg_length_profile[1]}, | 
|  | 2019 | {TIPC_NLA_STATS_MSG_LEN_P2, s->msg_length_profile[2]}, | 
|  | 2020 | {TIPC_NLA_STATS_MSG_LEN_P3, s->msg_length_profile[3]}, | 
|  | 2021 | {TIPC_NLA_STATS_MSG_LEN_P4, s->msg_length_profile[4]}, | 
|  | 2022 | {TIPC_NLA_STATS_MSG_LEN_P5, s->msg_length_profile[5]}, | 
|  | 2023 | {TIPC_NLA_STATS_MSG_LEN_P6, s->msg_length_profile[6]}, | 
|  | 2024 | {TIPC_NLA_STATS_RX_STATES, s->recv_states}, | 
|  | 2025 | {TIPC_NLA_STATS_RX_PROBES, s->recv_probes}, | 
|  | 2026 | {TIPC_NLA_STATS_RX_NACKS, s->recv_nacks}, | 
|  | 2027 | {TIPC_NLA_STATS_RX_DEFERRED, s->deferred_recv}, | 
|  | 2028 | {TIPC_NLA_STATS_TX_STATES, s->sent_states}, | 
|  | 2029 | {TIPC_NLA_STATS_TX_PROBES, s->sent_probes}, | 
|  | 2030 | {TIPC_NLA_STATS_TX_NACKS, s->sent_nacks}, | 
|  | 2031 | {TIPC_NLA_STATS_TX_ACKS, s->sent_acks}, | 
|  | 2032 | {TIPC_NLA_STATS_RETRANSMITTED, s->retransmitted}, | 
|  | 2033 | {TIPC_NLA_STATS_DUPLICATES, s->duplicates}, | 
|  | 2034 | {TIPC_NLA_STATS_LINK_CONGS, s->link_congs}, | 
|  | 2035 | {TIPC_NLA_STATS_MAX_QUEUE, s->max_queue_sz}, | 
|  | 2036 | {TIPC_NLA_STATS_AVG_QUEUE, s->queue_sz_counts ? | 
|  | 2037 | (s->accu_queue_sz / s->queue_sz_counts) : 0} | 
|  | 2038 | }; | 
|  | 2039 |  | 
|  | 2040 | stats = nla_nest_start(skb, TIPC_NLA_LINK_STATS); | 
|  | 2041 | if (!stats) | 
|  | 2042 | return -EMSGSIZE; | 
|  | 2043 |  | 
|  | 2044 | for (i = 0; i <  ARRAY_SIZE(map); i++) | 
|  | 2045 | if (nla_put_u32(skb, map[i].key, map[i].val)) | 
|  | 2046 | goto msg_full; | 
|  | 2047 |  | 
|  | 2048 | nla_nest_end(skb, stats); | 
|  | 2049 |  | 
|  | 2050 | return 0; | 
|  | 2051 | msg_full: | 
|  | 2052 | nla_nest_cancel(skb, stats); | 
|  | 2053 |  | 
|  | 2054 | return -EMSGSIZE; | 
|  | 2055 | } | 
|  | 2056 |  | 
|  | 2057 | /* Caller should hold appropriate locks to protect the link */ | 
|  | 2058 | int __tipc_nl_add_link(struct net *net, struct tipc_nl_msg *msg, | 
|  | 2059 | struct tipc_link *link, int nlflags) | 
|  | 2060 | { | 
|  | 2061 | u32 self = tipc_own_addr(net); | 
|  | 2062 | struct nlattr *attrs; | 
|  | 2063 | struct nlattr *prop; | 
|  | 2064 | void *hdr; | 
|  | 2065 | int err; | 
|  | 2066 |  | 
|  | 2067 | hdr = genlmsg_put(msg->skb, msg->portid, msg->seq, &tipc_genl_family, | 
|  | 2068 | nlflags, TIPC_NL_LINK_GET); | 
|  | 2069 | if (!hdr) | 
|  | 2070 | return -EMSGSIZE; | 
|  | 2071 |  | 
|  | 2072 | attrs = nla_nest_start(msg->skb, TIPC_NLA_LINK); | 
|  | 2073 | if (!attrs) | 
|  | 2074 | goto msg_full; | 
|  | 2075 |  | 
|  | 2076 | if (nla_put_string(msg->skb, TIPC_NLA_LINK_NAME, link->name)) | 
|  | 2077 | goto attr_msg_full; | 
|  | 2078 | if (nla_put_u32(msg->skb, TIPC_NLA_LINK_DEST, tipc_cluster_mask(self))) | 
|  | 2079 | goto attr_msg_full; | 
|  | 2080 | if (nla_put_u32(msg->skb, TIPC_NLA_LINK_MTU, link->mtu)) | 
|  | 2081 | goto attr_msg_full; | 
|  | 2082 | if (nla_put_u32(msg->skb, TIPC_NLA_LINK_RX, link->stats.recv_pkts)) | 
|  | 2083 | goto attr_msg_full; | 
|  | 2084 | if (nla_put_u32(msg->skb, TIPC_NLA_LINK_TX, link->stats.sent_pkts)) | 
|  | 2085 | goto attr_msg_full; | 
|  | 2086 |  | 
|  | 2087 | if (tipc_link_is_up(link)) | 
|  | 2088 | if (nla_put_flag(msg->skb, TIPC_NLA_LINK_UP)) | 
|  | 2089 | goto attr_msg_full; | 
|  | 2090 | if (link->active) | 
|  | 2091 | if (nla_put_flag(msg->skb, TIPC_NLA_LINK_ACTIVE)) | 
|  | 2092 | goto attr_msg_full; | 
|  | 2093 |  | 
|  | 2094 | prop = nla_nest_start(msg->skb, TIPC_NLA_LINK_PROP); | 
|  | 2095 | if (!prop) | 
|  | 2096 | goto attr_msg_full; | 
|  | 2097 | if (nla_put_u32(msg->skb, TIPC_NLA_PROP_PRIO, link->priority)) | 
|  | 2098 | goto prop_msg_full; | 
|  | 2099 | if (nla_put_u32(msg->skb, TIPC_NLA_PROP_TOL, link->tolerance)) | 
|  | 2100 | goto prop_msg_full; | 
|  | 2101 | if (nla_put_u32(msg->skb, TIPC_NLA_PROP_WIN, | 
|  | 2102 | link->window)) | 
|  | 2103 | goto prop_msg_full; | 
|  | 2104 | if (nla_put_u32(msg->skb, TIPC_NLA_PROP_PRIO, link->priority)) | 
|  | 2105 | goto prop_msg_full; | 
|  | 2106 | nla_nest_end(msg->skb, prop); | 
|  | 2107 |  | 
|  | 2108 | err = __tipc_nl_add_stats(msg->skb, &link->stats); | 
|  | 2109 | if (err) | 
|  | 2110 | goto attr_msg_full; | 
|  | 2111 |  | 
|  | 2112 | nla_nest_end(msg->skb, attrs); | 
|  | 2113 | genlmsg_end(msg->skb, hdr); | 
|  | 2114 |  | 
|  | 2115 | return 0; | 
|  | 2116 |  | 
|  | 2117 | prop_msg_full: | 
|  | 2118 | nla_nest_cancel(msg->skb, prop); | 
|  | 2119 | attr_msg_full: | 
|  | 2120 | nla_nest_cancel(msg->skb, attrs); | 
|  | 2121 | msg_full: | 
|  | 2122 | genlmsg_cancel(msg->skb, hdr); | 
|  | 2123 |  | 
|  | 2124 | return -EMSGSIZE; | 
|  | 2125 | } | 
|  | 2126 |  | 
|  | 2127 | static int __tipc_nl_add_bc_link_stat(struct sk_buff *skb, | 
|  | 2128 | struct tipc_stats *stats) | 
|  | 2129 | { | 
|  | 2130 | int i; | 
|  | 2131 | struct nlattr *nest; | 
|  | 2132 |  | 
|  | 2133 | struct nla_map { | 
|  | 2134 | __u32 key; | 
|  | 2135 | __u32 val; | 
|  | 2136 | }; | 
|  | 2137 |  | 
|  | 2138 | struct nla_map map[] = { | 
|  | 2139 | {TIPC_NLA_STATS_RX_INFO, stats->recv_pkts}, | 
|  | 2140 | {TIPC_NLA_STATS_RX_FRAGMENTS, stats->recv_fragments}, | 
|  | 2141 | {TIPC_NLA_STATS_RX_FRAGMENTED, stats->recv_fragmented}, | 
|  | 2142 | {TIPC_NLA_STATS_RX_BUNDLES, stats->recv_bundles}, | 
|  | 2143 | {TIPC_NLA_STATS_RX_BUNDLED, stats->recv_bundled}, | 
|  | 2144 | {TIPC_NLA_STATS_TX_INFO, stats->sent_pkts}, | 
|  | 2145 | {TIPC_NLA_STATS_TX_FRAGMENTS, stats->sent_fragments}, | 
|  | 2146 | {TIPC_NLA_STATS_TX_FRAGMENTED, stats->sent_fragmented}, | 
|  | 2147 | {TIPC_NLA_STATS_TX_BUNDLES, stats->sent_bundles}, | 
|  | 2148 | {TIPC_NLA_STATS_TX_BUNDLED, stats->sent_bundled}, | 
|  | 2149 | {TIPC_NLA_STATS_RX_NACKS, stats->recv_nacks}, | 
|  | 2150 | {TIPC_NLA_STATS_RX_DEFERRED, stats->deferred_recv}, | 
|  | 2151 | {TIPC_NLA_STATS_TX_NACKS, stats->sent_nacks}, | 
|  | 2152 | {TIPC_NLA_STATS_TX_ACKS, stats->sent_acks}, | 
|  | 2153 | {TIPC_NLA_STATS_RETRANSMITTED, stats->retransmitted}, | 
|  | 2154 | {TIPC_NLA_STATS_DUPLICATES, stats->duplicates}, | 
|  | 2155 | {TIPC_NLA_STATS_LINK_CONGS, stats->link_congs}, | 
|  | 2156 | {TIPC_NLA_STATS_MAX_QUEUE, stats->max_queue_sz}, | 
|  | 2157 | {TIPC_NLA_STATS_AVG_QUEUE, stats->queue_sz_counts ? | 
|  | 2158 | (stats->accu_queue_sz / stats->queue_sz_counts) : 0} | 
|  | 2159 | }; | 
|  | 2160 |  | 
|  | 2161 | nest = nla_nest_start(skb, TIPC_NLA_LINK_STATS); | 
|  | 2162 | if (!nest) | 
|  | 2163 | return -EMSGSIZE; | 
|  | 2164 |  | 
|  | 2165 | for (i = 0; i <  ARRAY_SIZE(map); i++) | 
|  | 2166 | if (nla_put_u32(skb, map[i].key, map[i].val)) | 
|  | 2167 | goto msg_full; | 
|  | 2168 |  | 
|  | 2169 | nla_nest_end(skb, nest); | 
|  | 2170 |  | 
|  | 2171 | return 0; | 
|  | 2172 | msg_full: | 
|  | 2173 | nla_nest_cancel(skb, nest); | 
|  | 2174 |  | 
|  | 2175 | return -EMSGSIZE; | 
|  | 2176 | } | 
|  | 2177 |  | 
|  | 2178 | int tipc_nl_add_bc_link(struct net *net, struct tipc_nl_msg *msg) | 
|  | 2179 | { | 
|  | 2180 | int err; | 
|  | 2181 | void *hdr; | 
|  | 2182 | struct nlattr *attrs; | 
|  | 2183 | struct nlattr *prop; | 
|  | 2184 | struct tipc_net *tn = net_generic(net, tipc_net_id); | 
|  | 2185 | struct tipc_link *bcl = tn->bcl; | 
|  | 2186 |  | 
|  | 2187 | if (!bcl) | 
|  | 2188 | return 0; | 
|  | 2189 |  | 
|  | 2190 | tipc_bcast_lock(net); | 
|  | 2191 |  | 
|  | 2192 | hdr = genlmsg_put(msg->skb, msg->portid, msg->seq, &tipc_genl_family, | 
|  | 2193 | NLM_F_MULTI, TIPC_NL_LINK_GET); | 
|  | 2194 | if (!hdr) { | 
|  | 2195 | tipc_bcast_unlock(net); | 
|  | 2196 | return -EMSGSIZE; | 
|  | 2197 | } | 
|  | 2198 |  | 
|  | 2199 | attrs = nla_nest_start(msg->skb, TIPC_NLA_LINK); | 
|  | 2200 | if (!attrs) | 
|  | 2201 | goto msg_full; | 
|  | 2202 |  | 
|  | 2203 | /* The broadcast link is always up */ | 
|  | 2204 | if (nla_put_flag(msg->skb, TIPC_NLA_LINK_UP)) | 
|  | 2205 | goto attr_msg_full; | 
|  | 2206 |  | 
|  | 2207 | if (nla_put_flag(msg->skb, TIPC_NLA_LINK_BROADCAST)) | 
|  | 2208 | goto attr_msg_full; | 
|  | 2209 | if (nla_put_string(msg->skb, TIPC_NLA_LINK_NAME, bcl->name)) | 
|  | 2210 | goto attr_msg_full; | 
|  | 2211 | if (nla_put_u32(msg->skb, TIPC_NLA_LINK_RX, 0)) | 
|  | 2212 | goto attr_msg_full; | 
|  | 2213 | if (nla_put_u32(msg->skb, TIPC_NLA_LINK_TX, 0)) | 
|  | 2214 | goto attr_msg_full; | 
|  | 2215 |  | 
|  | 2216 | prop = nla_nest_start(msg->skb, TIPC_NLA_LINK_PROP); | 
|  | 2217 | if (!prop) | 
|  | 2218 | goto attr_msg_full; | 
|  | 2219 | if (nla_put_u32(msg->skb, TIPC_NLA_PROP_WIN, bcl->window)) | 
|  | 2220 | goto prop_msg_full; | 
|  | 2221 | nla_nest_end(msg->skb, prop); | 
|  | 2222 |  | 
|  | 2223 | err = __tipc_nl_add_bc_link_stat(msg->skb, &bcl->stats); | 
|  | 2224 | if (err) | 
|  | 2225 | goto attr_msg_full; | 
|  | 2226 |  | 
|  | 2227 | tipc_bcast_unlock(net); | 
|  | 2228 | nla_nest_end(msg->skb, attrs); | 
|  | 2229 | genlmsg_end(msg->skb, hdr); | 
|  | 2230 |  | 
|  | 2231 | return 0; | 
|  | 2232 |  | 
|  | 2233 | prop_msg_full: | 
|  | 2234 | nla_nest_cancel(msg->skb, prop); | 
|  | 2235 | attr_msg_full: | 
|  | 2236 | nla_nest_cancel(msg->skb, attrs); | 
|  | 2237 | msg_full: | 
|  | 2238 | tipc_bcast_unlock(net); | 
|  | 2239 | genlmsg_cancel(msg->skb, hdr); | 
|  | 2240 |  | 
|  | 2241 | return -EMSGSIZE; | 
|  | 2242 | } | 
|  | 2243 |  | 
|  | 2244 | void tipc_link_set_tolerance(struct tipc_link *l, u32 tol, | 
|  | 2245 | struct sk_buff_head *xmitq) | 
|  | 2246 | { | 
|  | 2247 | l->tolerance = tol; | 
|  | 2248 | if (l->bc_rcvlink) | 
|  | 2249 | l->bc_rcvlink->tolerance = tol; | 
|  | 2250 | if (link_is_up(l)) | 
|  | 2251 | tipc_link_build_proto_msg(l, STATE_MSG, 0, 0, 0, tol, 0, xmitq); | 
|  | 2252 | } | 
|  | 2253 |  | 
|  | 2254 | void tipc_link_set_prio(struct tipc_link *l, u32 prio, | 
|  | 2255 | struct sk_buff_head *xmitq) | 
|  | 2256 | { | 
|  | 2257 | l->priority = prio; | 
|  | 2258 | tipc_link_build_proto_msg(l, STATE_MSG, 0, 0, 0, 0, prio, xmitq); | 
|  | 2259 | } | 
|  | 2260 |  | 
|  | 2261 | void tipc_link_set_abort_limit(struct tipc_link *l, u32 limit) | 
|  | 2262 | { | 
|  | 2263 | l->abort_limit = limit; | 
|  | 2264 | } |