iucv_sk 103 net/iucv/af_iucv.c if (!memcmp(&iucv_sk(sk)->src_name, nm, 8)) iucv_sk 145 net/iucv/af_iucv.c struct iucv_sock *iucv = iucv_sk(sk); iucv_sk 217 net/iucv/af_iucv.c INIT_LIST_HEAD(&iucv_sk(sk)->accept_q); iucv_sk 218 net/iucv/af_iucv.c spin_lock_init(&iucv_sk(sk)->accept_q_lock); iucv_sk 219 net/iucv/af_iucv.c skb_queue_head_init(&iucv_sk(sk)->send_skb_q); iucv_sk 220 net/iucv/af_iucv.c INIT_LIST_HEAD(&iucv_sk(sk)->message_q.list); iucv_sk 221 net/iucv/af_iucv.c spin_lock_init(&iucv_sk(sk)->message_q.lock); iucv_sk 222 net/iucv/af_iucv.c skb_queue_head_init(&iucv_sk(sk)->backlog_skb_q); iucv_sk 223 net/iucv/af_iucv.c iucv_sk(sk)->send_tag = 0; iucv_sk 277 net/iucv/af_iucv.c struct iucv_sock *par = iucv_sk(parent); iucv_sk 281 net/iucv/af_iucv.c list_add_tail(&iucv_sk(sk)->accept_q, &par->accept_q); iucv_sk 283 net/iucv/af_iucv.c iucv_sk(sk)->parent = parent; iucv_sk 290 net/iucv/af_iucv.c struct iucv_sock *par = iucv_sk(iucv_sk(sk)->parent); iucv_sk 293 net/iucv/af_iucv.c list_del_init(&iucv_sk(sk)->accept_q); iucv_sk 295 net/iucv/af_iucv.c iucv_sk(sk)->parent->sk_ack_backlog--; iucv_sk 296 net/iucv/af_iucv.c iucv_sk(sk)->parent = NULL; iucv_sk 305 net/iucv/af_iucv.c list_for_each_entry_safe(isk, n, &iucv_sk(parent)->accept_q, accept_q) { iucv_sk 388 net/iucv/af_iucv.c iucv = iucv_sk(sk); iucv_sk 417 net/iucv/af_iucv.c struct iucv_sock *iucv = iucv_sk(sk); iucv_sk 463 net/iucv/af_iucv.c iucv = iucv_sk(sk); iucv_sk 474 net/iucv/af_iucv.c memcpy(iucv_sk(sk)->dst_user_id, sa->siucv_user_id, 8); iucv_sk 475 net/iucv/af_iucv.c memcpy(iucv_sk(sk)->dst_name, sa->siucv_name, 8); iucv_sk 478 net/iucv/af_iucv.c low_nmcpy(user_data, iucv_sk(sk)->src_name); iucv_sk 481 net/iucv/af_iucv.c iucv = iucv_sk(sk); iucv_sk 599 net/iucv/af_iucv.c memcpy(siucv->siucv_user_id, iucv_sk(sk)->dst_user_id, 8); iucv_sk 600 net/iucv/af_iucv.c memcpy(siucv->siucv_name, &iucv_sk(sk)->dst_name, 8); iucv_sk 602 net/iucv/af_iucv.c memcpy(siucv->siucv_user_id, iucv_sk(sk)->src_user_id, 8); iucv_sk 603 net/iucv/af_iucv.c memcpy(siucv->siucv_name, iucv_sk(sk)->src_name, 8); iucv_sk 616 net/iucv/af_iucv.c struct iucv_sock *iucv = iucv_sk(sk); iucv_sk 700 net/iucv/af_iucv.c skb_queue_tail(&iucv_sk(sk)->backlog_skb_q, nskb); iucv_sk 730 net/iucv/af_iucv.c skb = skb_dequeue(&iucv_sk(sk)->backlog_skb_q); iucv_sk 739 net/iucv/af_iucv.c skb_queue_head(&iucv_sk(sk)->backlog_skb_q, skb); iucv_sk 744 net/iucv/af_iucv.c struct iucv_sock *iucv = iucv_sk(sk); iucv_sk 765 net/iucv/af_iucv.c struct iucv_sock *iucv = iucv_sk(sk); iucv_sk 841 net/iucv/af_iucv.c list_for_each_entry_safe(isk, n, &iucv_sk(parent)->accept_q, accept_q) { iucv_sk 892 net/iucv/af_iucv.c struct iucv_sock *iucv = iucv_sk(sk); iucv_sk 934 net/iucv/af_iucv.c err = iucv_path_quiesce(iucv_sk(sk)->path, NULL); iucv_sk 960 net/iucv/af_iucv.c if (iucv_sk(sk)->path) { iucv_sk 961 net/iucv/af_iucv.c iucv_path_sever(iucv_sk(sk)->path, NULL); iucv_sk 962 net/iucv/af_iucv.c iucv_path_free(iucv_sk(sk)->path); iucv_sk 963 net/iucv/af_iucv.c iucv_sk(sk)->path = NULL; iucv_sk 991 net/iucv/af_iucv.c !memcmp(&iucv_sk(sk)->src_name, src_name, 8)) { iucv_sk 996 net/iucv/af_iucv.c iucv = iucv_sk(sk); iucv_sk 1028 net/iucv/af_iucv.c niucv = iucv_sk(nsk); iucv_sk 1073 net/iucv/af_iucv.c struct iucv_sock *iucv = iucv_sk(sk); iucv_sk 1117 net/iucv/af_iucv.c struct sk_buff_head *list = &iucv_sk(sk)->send_skb_q; iucv_sk 1142 net/iucv/af_iucv.c if (skb_queue_empty(&iucv_sk(sk)->send_skb_q)) { iucv_sk 1154 net/iucv/af_iucv.c if (!list_empty(&iucv_sk(sk)->accept_q))