bclink            118 net/tipc/bcast.c static struct bclink *bclink = NULL;
bclink            154 net/tipc/bcast.c 	struct sk_buff *buf = n_ptr->bclink.deferred_head;
bclink            156 net/tipc/bcast.c 	n_ptr->bclink.gap_after = n_ptr->bclink.gap_to =
bclink            157 net/tipc/bcast.c 		mod(n_ptr->bclink.last_in);
bclink            159 net/tipc/bcast.c 		n_ptr->bclink.gap_to = mod(buf_seqno(buf) - 1);
bclink            211 net/tipc/bcast.c 	if (less_eq(acked, n_ptr->bclink.acked))
bclink            219 net/tipc/bcast.c 	while (crs && less_eq(buf_seqno(crs), n_ptr->bclink.acked)) {
bclink            236 net/tipc/bcast.c 	n_ptr->bclink.acked = acked;
bclink            272 net/tipc/bcast.c 	if (!less(n_ptr->bclink.gap_after, n_ptr->bclink.gap_to))
bclink            281 net/tipc/bcast.c 		msg_set_bcast_ack(msg, mod(n_ptr->bclink.last_in));
bclink            282 net/tipc/bcast.c 		msg_set_bcgap_after(msg, n_ptr->bclink.gap_after);
bclink            283 net/tipc/bcast.c 		msg_set_bcgap_to(msg, n_ptr->bclink.gap_to);
bclink            301 net/tipc/bcast.c 		n_ptr->bclink.nack_sync = tipc_own_tag;
bclink            313 net/tipc/bcast.c 	if (!n_ptr->bclink.supported ||
bclink            314 net/tipc/bcast.c 	    less_eq(last_sent, mod(n_ptr->bclink.last_in)))
bclink            318 net/tipc/bcast.c 	if (n_ptr->bclink.gap_after == n_ptr->bclink.gap_to)
bclink            319 net/tipc/bcast.c 		n_ptr->bclink.gap_to = last_sent;
bclink            340 net/tipc/bcast.c 	my_after = n_ptr->bclink.gap_after;
bclink            341 net/tipc/bcast.c 	my_to = n_ptr->bclink.gap_to;
bclink            345 net/tipc/bcast.c 			n_ptr->bclink.gap_after = gap_to;
bclink            347 net/tipc/bcast.c 			n_ptr->bclink.gap_to = n_ptr->bclink.gap_after;
bclink            350 net/tipc/bcast.c 			n_ptr->bclink.gap_to = gap_after;
bclink            355 net/tipc/bcast.c 		struct sk_buff *buf = n_ptr->bclink.deferred_head;
bclink            356 net/tipc/bcast.c 		u32 prev = n_ptr->bclink.gap_to;
bclink            370 net/tipc/bcast.c 			n_ptr->bclink.gap_to = gap_after;
bclink            376 net/tipc/bcast.c 		if (n_ptr->bclink.gap_to != n_ptr->bclink.gap_after) {
bclink            428 net/tipc/bcast.c 	if (unlikely(!node || !tipc_node_is_up(node) || !node->bclink.supported ||
bclink            467 net/tipc/bcast.c 	deferred = node->bclink.deferred_head;
bclink            468 net/tipc/bcast.c 	next_in = mod(node->bclink.last_in + 1);
bclink            473 net/tipc/bcast.c 		node->bclink.last_in++;
bclink            489 net/tipc/bcast.c 			if (tipc_link_recv_fragment(&node->bclink.defragm,
bclink            502 net/tipc/bcast.c 			node->bclink.deferred_head = deferred->next;
bclink            507 net/tipc/bcast.c 		u32 gap_after = node->bclink.gap_after;
bclink            508 net/tipc/bcast.c 		u32 gap_to = node->bclink.gap_to;
bclink            510 net/tipc/bcast.c 		if (tipc_link_defer_pkt(&node->bclink.deferred_head,
bclink            511 net/tipc/bcast.c 					&node->bclink.deferred_tail,
bclink            513 net/tipc/bcast.c 			node->bclink.nack_sync++;
bclink            516 net/tipc/bcast.c 				node->bclink.gap_after = seqno;
bclink            518 net/tipc/bcast.c 				node->bclink.gap_to = seqno;
bclink            520 net/tipc/bcast.c 		if (bclink_ack_allowed(node->bclink.nack_sync)) {
bclink            543 net/tipc/bcast.c 	return (n_ptr->bclink.supported &&
bclink            544 net/tipc/bcast.c 		(tipc_bclink_get_last_sent() != n_ptr->bclink.acked));
bclink            778 net/tipc/bcast.c 	bclink = kzalloc(sizeof(*bclink), GFP_ATOMIC);
bclink            779 net/tipc/bcast.c 	if (!bcbearer || !bclink) {
bclink            784 net/tipc/bcast.c 		kfree(bclink);
bclink            785 net/tipc/bcast.c 		bclink = NULL;
bclink            794 net/tipc/bcast.c 	bcl = &bclink->link;
bclink            797 net/tipc/bcast.c 	spin_lock_init(&bclink->node.lock);
bclink            798 net/tipc/bcast.c 	bcl->owner = &bclink->node;
bclink            824 net/tipc/bcast.c 		kfree(bclink);
bclink            825 net/tipc/bcast.c 		bclink = NULL;
bclink           1043 net/tipc/link.c 	msg_set_bcast_ack(msg, l_ptr->owner->bclink.last_in);
bclink           1558 net/tipc/link.c 		msg_set_bcast_ack(buf_msg(buf), l_ptr->owner->bclink.last_in);
bclink           1577 net/tipc/link.c 		msg_set_bcast_ack(buf_msg(buf),l_ptr->owner->bclink.last_in);
bclink           1601 net/tipc/link.c 			msg_set_bcast_ack(msg, l_ptr->owner->bclink.last_in);
bclink           1697 net/tipc/link.c 		tipc_printf(TIPC_OUTPUT, "Supported: %d,  ", n_ptr->bclink.supported);
bclink           1698 net/tipc/link.c 		tipc_printf(TIPC_OUTPUT, "Acked: %u\n", n_ptr->bclink.acked);
bclink           1699 net/tipc/link.c 		tipc_printf(TIPC_OUTPUT, "Last in: %u,  ", n_ptr->bclink.last_in);
bclink           1700 net/tipc/link.c 		tipc_printf(TIPC_OUTPUT, "Gap after: %u,  ", n_ptr->bclink.gap_after);
bclink           1701 net/tipc/link.c 		tipc_printf(TIPC_OUTPUT, "Gap to: %u\n", n_ptr->bclink.gap_to);
bclink           1702 net/tipc/link.c 		tipc_printf(TIPC_OUTPUT, "Nack sync: %u\n\n", n_ptr->bclink.nack_sync);
bclink           1751 net/tipc/link.c 		msg_set_bcast_ack(msg, l_ptr->owner->bclink.last_in);
bclink           1905 net/tipc/link.c 		if (less(n_ptr->bclink.acked, msg_bcast_ack(msg))) {
bclink           1906 net/tipc/link.c 			if (tipc_node_is_up(n_ptr) && n_ptr->bclink.supported)
bclink           2136 net/tipc/link.c 	msg_set_bcast_ack(msg, mod(l_ptr->owner->bclink.last_in));
bclink           2297 net/tipc/link.c 		l_ptr->owner->bclink.supported = (max_pkt_info != 0);
bclink           2306 net/tipc/link.c 			l_ptr->owner->bclink.last_in = mod(msg_last_bcast(msg));
bclink           2496 net/tipc/link.c 		msg_set_bcast_ack(msg, l_ptr->owner->bclink.last_in);
bclink            351 net/tipc/node.c 	n_ptr->bclink.acked = tipc_bclink_get_last_sent();
bclink            374 net/tipc/node.c 	if (n_ptr->bclink.supported) {
bclink            395 net/tipc/node.c 	n_ptr->bclink.gap_after = n_ptr->bclink.gap_to = 0;
bclink            396 net/tipc/node.c 	while (n_ptr->bclink.deferred_head) {
bclink            397 net/tipc/node.c 		struct sk_buff* buf = n_ptr->bclink.deferred_head;
bclink            398 net/tipc/node.c 		n_ptr->bclink.deferred_head = buf->next;
bclink            401 net/tipc/node.c 	if (n_ptr->bclink.defragm) {
bclink            402 net/tipc/node.c 		buf_discard(n_ptr->bclink.defragm);
bclink            403 net/tipc/node.c 		n_ptr->bclink.defragm = NULL;
bclink            405 net/tipc/node.c 	if (in_own_cluster(n_ptr->addr) && n_ptr->bclink.supported) {
bclink            406 net/tipc/node.c 		tipc_bclink_acknowledge(n_ptr, mod(n_ptr->bclink.acked + 10000));
bclink            425 net/tipc/node.c 				if (n_ptr->bclink.supported) {
bclink             94 net/tipc/node.h 	} bclink;