Home
last modified time | relevance | path

Searched refs:tmpq (Results 1 – 10 of 10) sorted by relevance

/Linux-v5.4/drivers/misc/mic/scif/
Dscif_epd.c73 struct list_head *pos, *tmpq; in scif_find_listen_ep() local
76 list_for_each_safe(pos, tmpq, &scif_info.listen) { in scif_find_listen_ep()
89 struct list_head *pos, *tmpq; in scif_cleanup_zombie_epd() local
93 list_for_each_safe(pos, tmpq, &scif_info.zombie) { in scif_cleanup_zombie_epd()
258 struct list_head *pos, *tmpq; in scif_discnct() local
261 list_for_each_safe(pos, tmpq, &scif_info.connected) { in scif_discnct()
Dscif_nm.c21 struct list_head *pos, *tmpq; in scif_invalidate_ep() local
25 list_for_each_safe(pos, tmpq, &scif_info.disconnected) { in scif_invalidate_ep()
34 list_for_each_safe(pos, tmpq, &scif_info.connected) { in scif_invalidate_ep()
Dscif_api.c88 struct list_head *pos, *tmpq; in scif_disconnect_ep() local
108 list_for_each_safe(pos, tmpq, &scif_info.connected) { in scif_disconnect_ep()
125 list_for_each_safe(pos, tmpq, &scif_info.disconnected) { in scif_disconnect_ep()
158 struct list_head *pos, *tmpq; in scif_close() local
187 list_for_each_safe(pos, tmpq, &scif_info.disconnected) { in scif_close()
221 list_for_each_safe(pos, tmpq, &scif_info.listen) { in scif_close()
232 list_for_each_safe(pos, tmpq, &scif_info.uaccept) { in scif_close()
242 list_for_each_safe(pos, tmpq, &scif_info.connected) { in scif_close()
250 list_for_each_safe(pos, tmpq, &scif_info.disconnected) { in scif_close()
Dscif_fd.c164 struct list_head *pos, *tmpq; in scif_fdioctl() local
172 list_for_each_safe(pos, tmpq, &scif_info.uaccept) { in scif_fdioctl()
188 list_for_each_safe(pos, tmpq, &lisep->li_accept) { in scif_fdioctl()
Dscif_dma.c281 struct list_head *pos, *tmpq; in scif_mmu_notif_handler() local
286 list_for_each_safe(pos, tmpq, &scif_info.mmu_notif_cleanup) { in scif_mmu_notif_handler()
/Linux-v5.4/net/tipc/
Dbcast.c317 struct sk_buff_head tmpq; in tipc_mcast_send_sync() local
347 __skb_queue_head_init(&tmpq); in tipc_mcast_send_sync()
348 __skb_queue_tail(&tmpq, _skb); in tipc_mcast_send_sync()
350 tipc_bcast_xmit(net, &tmpq, cong_link_cnt); in tipc_mcast_send_sync()
352 tipc_rcast_xmit(net, &tmpq, dests, cong_link_cnt); in tipc_mcast_send_sync()
355 __skb_queue_purge(&tmpq); in tipc_mcast_send_sync()
Dlink.c857 struct sk_buff_head tmpq; in link_prepare_wakeup() local
861 __skb_queue_head_init(&tmpq); in link_prepare_wakeup()
872 __skb_queue_tail(&tmpq, skb); in link_prepare_wakeup()
876 skb_queue_splice_tail(&tmpq, inputq); in link_prepare_wakeup()
1224 struct sk_buff_head tmpq; in tipc_link_input() local
1229 skb_queue_head_init(&tmpq); in tipc_link_input()
1233 tipc_data_input(l, iskb, &tmpq); in tipc_link_input()
1234 tipc_skb_queue_splice_tail(&tmpq, inputq); in tipc_link_input()
Dsocket.c1138 struct sk_buff_head tmpq; in tipc_sk_mcast_rcv() local
1144 __skb_queue_head_init(&tmpq); in tipc_sk_mcast_rcv()
1193 __skb_queue_tail(&tmpq, _skb); in tipc_sk_mcast_rcv()
1201 skb_queue_splice_tail_init(&tmpq, inputq); in tipc_sk_mcast_rcv()
1205 __skb_queue_purge(&tmpq); in tipc_sk_mcast_rcv()
/Linux-v5.4/drivers/net/
Dxen-netfront.c999 struct sk_buff_head tmpq; in xennet_poll() local
1006 skb_queue_head_init(&tmpq); in xennet_poll()
1017 err = xennet_get_responses(queue, &rinfo, rp, &tmpq); in xennet_poll()
1021 while ((skb = __skb_dequeue(&tmpq))) in xennet_poll()
1028 skb = __skb_dequeue(&tmpq); in xennet_poll()
1035 __skb_queue_head(&tmpq, skb); in xennet_poll()
1036 queue->rx.rsp_cons += skb_queue_len(&tmpq); in xennet_poll()
1050 if (unlikely(xennet_fill_frags(queue, skb, &tmpq))) in xennet_poll()
/Linux-v5.4/fs/ocfs2/dlm/
Ddlmrecovery.c1802 struct list_head *tmpq = NULL; in dlm_process_recovery_data() local
1831 tmpq = NULL; in dlm_process_recovery_data()
1843 tmpq = dlm_list_idx_to_ptr(res, j); in dlm_process_recovery_data()
1844 list_for_each(iter, tmpq) { in dlm_process_recovery_data()
1891 if (tmpq != queue) { in dlm_process_recovery_data()