Home
last modified time | relevance | path

Searched refs:skbs (Results 1 – 25 of 48) sorted by relevance

12

/Linux-v5.4/drivers/staging/isdn/hysdn/
Dhysdn_net.c45 struct sk_buff *skbs[MAX_SKB_BUFFERS]; /* pointers to tx-skbs */ member
96 dev_kfree_skb(nl->skbs[nl->out_idx++]); /* free skb */ in flush_tx_buffers()
130 lp->skbs[lp->in_idx++] = skb; /* add to buffer list */ in net_send_packet()
174 lp->dev->stats.tx_bytes += lp->skbs[lp->out_idx]->len; in hysdn_tx_netack()
176 dev_kfree_skb(lp->skbs[lp->out_idx++]); /* free skb */ in hysdn_tx_netack()
232 return (lp->skbs[lp->out_idx]); /* next packet to send */ in hysdn_tx_netget()
Dhycapi.c126 cinfo->skbs[cinfo->in_idx++] = skb; /* add to buffer list */ in hycapi_sendmsg_internal()
629 kfree_skb(cinfo->skbs[cinfo->out_idx]); /* free skb */ in hycapi_tx_capiack()
630 cinfo->skbs[cinfo->out_idx++] = NULL; in hycapi_tx_capiack()
656 return (cinfo->skbs[cinfo->out_idx]); /* next packet to send */ in hycapi_tx_capiget()
Dhysdn_defs.h205 struct sk_buff *skbs[HYSDN_MAX_CAPI_SKB]; member
/Linux-v5.4/drivers/net/ethernet/sfc/falcon/
Dselftest.c80 struct sk_buff **skbs; member
426 state->skbs[i] = skb; in ef4_begin_loopback()
480 skb = state->skbs[i]; in ef4_end_loopback()
536 state->skbs = kcalloc(state->packet_count, in ef4_test_loopback()
537 sizeof(state->skbs[0]), GFP_KERNEL); in ef4_test_loopback()
538 if (!state->skbs) in ef4_test_loopback()
559 kfree(state->skbs); in ef4_test_loopback()
/Linux-v5.4/drivers/net/ethernet/sfc/
Dselftest.c80 struct sk_buff **skbs; member
426 state->skbs[i] = skb; in efx_begin_loopback()
480 skb = state->skbs[i]; in efx_end_loopback()
536 state->skbs = kcalloc(state->packet_count, in efx_test_loopback()
537 sizeof(state->skbs[0]), GFP_KERNEL); in efx_test_loopback()
538 if (!state->skbs) in efx_test_loopback()
559 kfree(state->skbs); in efx_test_loopback()
/Linux-v5.4/net/mac80211/
Dtx.c887 __skb_queue_tail(&tx->skbs, tmp); in ieee80211_fragment()
929 __skb_queue_tail(&tx->skbs, skb); in ieee80211_tx_h_fragment()
966 skb_queue_walk(&tx->skbs, skb) { in ieee80211_tx_h_fragment()
972 if (!skb_queue_is_last(&tx->skbs, skb)) { in ieee80211_tx_h_fragment()
1002 skb_queue_walk(&tx->skbs, skb) { in ieee80211_tx_h_stats()
1055 skb_queue_walk(&tx->skbs, skb) { in ieee80211_tx_h_calculate_duration()
1059 if (!skb_queue_is_last(&tx->skbs, skb)) { in ieee80211_tx_h_calculate_duration()
1060 struct sk_buff *next = skb_queue_next(&tx->skbs, skb); in ieee80211_tx_h_calculate_duration()
1167 __skb_queue_head_init(&tx->skbs); in ieee80211_tx_prepare()
1621 struct sk_buff_head *skbs, in ieee80211_tx_frags() argument
[all …]
Dwpa.c254 skb_queue_walk(&tx->skbs, skb) { in ieee80211_crypto_tkip_encrypt()
484 skb_queue_walk(&tx->skbs, skb) { in ieee80211_crypto_ccmp_encrypt()
714 skb_queue_walk(&tx->skbs, skb) { in ieee80211_crypto_gcmp_encrypt()
942 if (WARN_ON(skb_queue_len(&tx->skbs) != 1)) in ieee80211_crypto_aes_cmac_encrypt()
945 skb = skb_peek(&tx->skbs); in ieee80211_crypto_aes_cmac_encrypt()
990 if (WARN_ON(skb_queue_len(&tx->skbs) != 1)) in ieee80211_crypto_aes_cmac_256_encrypt()
993 skb = skb_peek(&tx->skbs); in ieee80211_crypto_aes_cmac_256_encrypt()
1135 if (WARN_ON(skb_queue_len(&tx->skbs) != 1)) in ieee80211_crypto_aes_gmac_encrypt()
1138 skb = skb_peek(&tx->skbs); in ieee80211_crypto_aes_gmac_encrypt()
1239 skb_queue_walk(&tx->skbs, skb) { in ieee80211_crypto_hw_encrypt()
Dwep.c300 skb_queue_walk(&tx->skbs, skb) { in ieee80211_crypto_wep_encrypt()
/Linux-v5.4/drivers/net/ethernet/socionext/
Dsni_ave.c226 struct sk_buff *skbs; member
579 skb = priv->rx.desc[entry].skbs; in ave_rxdesc_prepare()
612 priv->rx.desc[entry].skbs = skb; in ave_rxdesc_prepare()
709 if (priv->tx.desc[done_idx].skbs) { in ave_tx_complete()
712 dev_consume_skb_any(priv->tx.desc[done_idx].skbs); in ave_tx_complete()
713 priv->tx.desc[done_idx].skbs = NULL; in ave_tx_complete()
770 skb = priv->rx.desc[proc_idx].skbs; in ave_rx_receive()
771 priv->rx.desc[proc_idx].skbs = NULL; in ave_rx_receive()
1363 if (!priv->tx.desc[entry].skbs) in ave_stop()
1367 dev_kfree_skb_any(priv->tx.desc[entry].skbs); in ave_stop()
[all …]
/Linux-v5.4/drivers/net/wireless/mediatek/mt7601u/
Ddma.c272 struct sk_buff_head skbs; in mt7601u_tx_tasklet() local
275 __skb_queue_head_init(&skbs); in mt7601u_tx_tasklet()
284 skb_queue_splice_init(&dev->tx_skb_done, &skbs); in mt7601u_tx_tasklet()
288 while (!skb_queue_empty(&skbs)) { in mt7601u_tx_tasklet()
289 struct sk_buff *skb = __skb_dequeue(&skbs); in mt7601u_tx_tasklet()
/Linux-v5.4/include/linux/
Dskb_array.h191 struct sk_buff **skbs, int n) in skb_array_unconsume() argument
193 ptr_ring_unconsume(&a->ring, (void **)skbs, n, __skb_array_destroy_skb); in skb_array_unconsume()
/Linux-v5.4/kernel/bpf/
Dcpumap.c263 void *skbs[CPUMAP_BATCH]; in cpu_map_kthread_run() local
299 m = kmem_cache_alloc_bulk(skbuff_head_cache, gfp, n, skbs); in cpu_map_kthread_run()
302 skbs[i] = NULL; /* effect: xdp_return_frame */ in cpu_map_kthread_run()
309 struct sk_buff *skb = skbs[i]; in cpu_map_kthread_run()
/Linux-v5.4/Documentation/networking/
Dnetdev-features.txt126 ndo_start_xmit can handle skbs with frags in high memory.
130 Those features say that ndo_start_xmit can handle fragmented skbs:
131 NETIF_F_SG --- paged skbs (skb_shinfo()->frags), NETIF_F_FRAGLIST ---
132 chained skbs (skb->next/prev list).
Dsegmentation-offloads.rst60 UFO is deprecated: modern kernels will no longer generate UFO skbs, but can
165 padded and stored as chained skbs, and skb_segment() splits based on those.
181 will check for GSO_BY_FRAGS and WARN if asked to manipulate these skbs.
/Linux-v5.4/drivers/net/ethernet/qlogic/qede/
Dqede_fp.c103 struct sk_buff *skb = txq->sw_tx_ring.skbs[idx].skb; in qede_free_tx_pkt()
108 bool data_split = txq->sw_tx_ring.skbs[idx].flags & QEDE_TSO_SPLIT_BD; in qede_free_tx_pkt()
148 txq->sw_tx_ring.skbs[idx].skb = NULL; in qede_free_tx_pkt()
149 txq->sw_tx_ring.skbs[idx].flags = 0; in qede_free_tx_pkt()
160 struct sk_buff *skb = txq->sw_tx_ring.skbs[idx].skb; in qede_free_failed_tx_pkt()
196 txq->sw_tx_ring.skbs[idx].skb = NULL; in qede_free_failed_tx_pkt()
197 txq->sw_tx_ring.skbs[idx].flags = 0; in qede_free_failed_tx_pkt()
1470 txq->sw_tx_ring.skbs[idx].skb = skb; in qede_start_xmit()
1592 txq->sw_tx_ring.skbs[idx].flags |= QEDE_TSO_SPLIT_BD; in qede_start_xmit()
/Linux-v5.4/drivers/net/wireless/intel/iwlwifi/dvm/
Dtx.c1128 struct sk_buff_head skbs; in iwlagn_rx_reply_tx() local
1149 __skb_queue_head_init(&skbs); in iwlagn_rx_reply_tx()
1178 iwl_trans_reclaim(priv->trans, txq_id, ssn, &skbs); in iwlagn_rx_reply_tx()
1183 skb_queue_walk(&skbs, skb) { in iwlagn_rx_reply_tx()
1254 while (!skb_queue_empty(&skbs)) { in iwlagn_rx_reply_tx()
1255 skb = __skb_dequeue(&skbs); in iwlagn_rx_reply_tx()
/Linux-v5.4/drivers/net/wireless/intel/iwlwifi/
Diwl-trans.h549 struct sk_buff_head *skbs);
953 int ssn, struct sk_buff_head *skbs) in iwl_trans_reclaim() argument
960 trans->ops->reclaim(trans, queue, ssn, skbs); in iwl_trans_reclaim()
/Linux-v5.4/drivers/net/ethernet/natsemi/
Dns83820.c401 struct sk_buff *skbs[NR_RX_DESC]; member
524 BUG_ON(NULL != dev->rx_info.skbs[next_empty]); in ns83820_add_rx_skb()
525 dev->rx_info.skbs[next_empty] = skb; in ns83820_add_rx_skb()
788 struct sk_buff *skb = dev->rx_info.skbs[i]; in ns83820_cleanup_rx()
789 dev->rx_info.skbs[i] = NULL; in ns83820_cleanup_rx()
854 skb = info->skbs[next_rx];
855 info->skbs[next_rx] = NULL;
/Linux-v5.4/drivers/net/wireless/intel/iwlegacy/
D3945.c283 skb = txq->skbs[txq->q.read_ptr]; in il3945_tx_queue_reclaim()
285 txq->skbs[txq->q.read_ptr] = NULL; in il3945_tx_queue_reclaim()
332 info = IEEE80211_SKB_CB(txq->skbs[txq->q.read_ptr]); in il3945_hdl_tx()
667 if (txq->skbs) { in il3945_hw_txq_free_tfd()
668 struct sk_buff *skb = txq->skbs[txq->q.read_ptr]; in il3945_hw_txq_free_tfd()
673 txq->skbs[txq->q.read_ptr] = NULL; in il3945_hw_txq_free_tfd()
/Linux-v5.4/drivers/crypto/chelsio/chtls/
Dchtls_main.c415 struct sk_buff **skbs, const __be64 *rsp) in chtls_recv() argument
417 struct sk_buff *skb = *skbs; in chtls_recv()
/Linux-v5.4/drivers/net/ethernet/chelsio/cxgb3/
Dcxgb3_offload.c480 static int rx_offload_blackhole(struct t3cdev *dev, struct sk_buff **skbs, in rx_offload_blackhole() argument
484 dev_kfree_skb_any(skbs[n]); in rx_offload_blackhole()
1019 static int process_rx(struct t3cdev *dev, struct sk_buff **skbs, int n) in process_rx() argument
1022 struct sk_buff *skb = *skbs++; in process_rx()
Dsge.c1876 struct sk_buff *skbs[], int n) in deliver_partial_bundle() argument
1880 tdev->recv(tdev, skbs, n); in deliver_partial_bundle()
1903 struct sk_buff *skb, *tmp, *skbs[RX_BUNDLE_SIZE]; in ofld_poll() local
1925 skbs[ngathered] = skb; in ofld_poll()
1928 adapter->tdev.recv(&adapter->tdev, skbs, in ofld_poll()
1939 deliver_partial_bundle(&adapter->tdev, q, skbs, ngathered); in ofld_poll()
/Linux-v5.4/drivers/net/wireless/intel/iwlwifi/mvm/
Dtx.c1455 struct sk_buff_head skbs; in iwl_mvm_rx_tx_cmd_single() local
1461 __skb_queue_head_init(&skbs); in iwl_mvm_rx_tx_cmd_single()
1469 iwl_trans_reclaim(mvm->trans, txq_id, ssn, &skbs); in iwl_mvm_rx_tx_cmd_single()
1471 while (!skb_queue_empty(&skbs)) { in iwl_mvm_rx_tx_cmd_single()
1472 struct sk_buff *skb = __skb_dequeue(&skbs); in iwl_mvm_rx_tx_cmd_single()
/Linux-v5.4/net/bluetooth/
Dl2cap_core.c62 struct sk_buff_head *skbs, u8 event);
1872 struct sk_buff_head *skbs) in l2cap_streaming_send() argument
1877 BT_DBG("chan %p, skbs %p", chan, skbs); in l2cap_streaming_send()
1882 skb_queue_splice_tail_init(skbs, &chan->tx_q); in l2cap_streaming_send()
2712 struct sk_buff_head *skbs, u8 event) in l2cap_tx_state_xmit() argument
2714 BT_DBG("chan %p, control %p, skbs %p, event %d", chan, control, skbs, in l2cap_tx_state_xmit()
2720 chan->tx_send_head = skb_peek(skbs); in l2cap_tx_state_xmit()
2722 skb_queue_splice_tail_init(skbs, &chan->tx_q); in l2cap_tx_state_xmit()
2784 struct sk_buff_head *skbs, u8 event) in l2cap_tx_state_wait_f() argument
2786 BT_DBG("chan %p, control %p, skbs %p, event %d", chan, control, skbs, in l2cap_tx_state_wait_f()
[all …]
/Linux-v5.4/drivers/net/ethernet/mscc/
Docelot.h504 struct list_head skbs; member

12