| /Linux-v5.10/drivers/net/ethernet/intel/ice/ |
| D | ice_lib.c | 87 vsi->txq_map = devm_kcalloc(dev, (2 * vsi->alloc_txq), in ice_vsi_alloc_arrays() 88 sizeof(*vsi->txq_map), GFP_KERNEL); in ice_vsi_alloc_arrays() 90 if (!vsi->txq_map) in ice_vsi_alloc_arrays() 113 devm_kfree(dev, vsi->txq_map); in ice_vsi_alloc_arrays() 298 if (vsi->txq_map) { in ice_vsi_free_arrays() 299 devm_kfree(dev, vsi->txq_map); in ice_vsi_free_arrays() 300 vsi->txq_map = NULL; in ice_vsi_free_arrays() 534 .vsi_map = vsi->txq_map, in ice_vsi_get_qs() 575 clear_bit(vsi->txq_map[i], pf->avail_txqs); in ice_vsi_put_qs() 576 vsi->txq_map[i] = ICE_INVAL_Q_INDEX; in ice_vsi_put_qs() [all …]
|
| D | ice_base.c | 743 wr32(hw, QINT_TQCTL(vsi->txq_map[txq]), val); in ice_cfg_txq_interrupt() 747 wr32(hw, QINT_TQCTL(vsi->txq_map[xdp_txq]), in ice_cfg_txq_interrupt()
|
| D | ice.h | 313 u16 *txq_map; /* index in pf->avail_txqs */ member
|
| D | ice_virtchnl_pf.c | 741 reg = (((vsi->txq_map[0] << VPLAN_TX_QBASE_VFFIRSTQ_S) & in ice_ena_vf_q_mappings() 2516 u32 pfq = vsi->txq_map[q_idx]; in ice_vf_ena_txq_interrupt() 2795 qmap = map->txq_map; in ice_cfg_interrupt() 2861 (!vector_id && (map->rxq_map || map->txq_map))) { in ice_vc_cfg_irq_map_msg()
|
| D | ice_main.c | 2274 xdp_ring->reg_idx = vsi->txq_map[xdp_q_idx]; in ice_xdp_alloc_setup_rings() 2332 .vsi_map = vsi->txq_map, in ice_prepare_xdp_rings() 2405 clear_bit(vsi->txq_map[i + vsi->alloc_txq], pf->avail_txqs); in ice_prepare_xdp_rings() 2406 vsi->txq_map[i + vsi->alloc_txq] = ICE_INVAL_Q_INDEX; in ice_prepare_xdp_rings() 2450 clear_bit(vsi->txq_map[i + vsi->alloc_txq], pf->avail_txqs); in ice_destroy_xdp_rings() 2451 vsi->txq_map[i + vsi->alloc_txq] = ICE_INVAL_Q_INDEX; in ice_destroy_xdp_rings() 6568 head = (rd32(hw, QTX_COMM_HEAD(vsi->txq_map[txqueue])) & in ice_tx_timeout()
|
| /Linux-v5.10/drivers/net/wireless/ath/ath9k/ |
| D | link.c | 32 txq = sc->tx.txq_map[i]; in ath_tx_complete_check() 182 txctl.txq = sc->tx.txq_map[IEEE80211_AC_BE]; in ath_paprd_send_frame()
|
| D | debug.h | 195 #define PR_QNUM(_n) sc->tx.txq_map[_n]->axq_qnum
|
| D | tx99.c | 126 txctl.txq = sc->tx.txq_map[IEEE80211_AC_VO]; in ath9k_tx99_init()
|
| D | xmit.c | 173 txq = sc->tx.txq_map[q]; in ath_txq_skb_done() 211 if (tid->txq == sc->tx.txq_map[q]) { in ath_tid_pull() 1960 txq = sc->tx.txq_map[i]; in ath_txq_schedule_all() 2305 if (txq == sc->tx.txq_map[q]) { in ath_tx_start() 2807 tid->txq = sc->tx.txq_map[acno]; in ath_tx_node_init()
|
| D | init.c | 417 sc->tx.txq_map[i] = ath_txq_setup(sc, ATH9K_TX_QUEUE_DATA, i); in ath9k_init_queues() 418 sc->tx.txq_map[i]->mac80211_qnum = i; in ath9k_init_queues()
|
| D | gpio.c | 429 txq = sc->tx.txq_map[IEEE80211_AC_BE]; in ath9k_init_btcoex()
|
| D | beacon.c | 50 txq = sc->tx.txq_map[IEEE80211_AC_BE]; in ath9k_beaconq_config()
|
| D | channel.c | 1016 txctl.txq = sc->tx.txq_map[IEEE80211_AC_VO]; in ath_scan_send_probe() 1137 txctl.txq = sc->tx.txq_map[IEEE80211_AC_VO]; in ath_chanctx_send_vif_ps_frame()
|
| D | ath9k.h | 291 struct ath_txq *txq_map[IEEE80211_NUM_ACS]; member
|
| D | main.c | 808 txctl.txq = sc->tx.txq_map[skb_get_queue_mapping(skb)]; in ath9k_tx() 1650 txq = sc->tx.txq_map[queue]; in ath9k_conf_tx()
|
| D | debug.c | 647 txq = sc->tx.txq_map[i]; in read_file_queues()
|
| /Linux-v5.10/drivers/net/ethernet/cisco/enic/ |
| D | enic_main.c | 825 unsigned int txq_map; in enic_hard_start_xmit() local 833 txq_map = skb_get_queue_mapping(skb) % enic->wq_count; in enic_hard_start_xmit() 834 wq = &enic->wq[txq_map]; in enic_hard_start_xmit() 835 txq = netdev_get_tx_queue(netdev, txq_map); in enic_hard_start_xmit() 849 spin_lock(&enic->wq_lock[txq_map]); in enic_hard_start_xmit() 856 spin_unlock(&enic->wq_lock[txq_map]); in enic_hard_start_xmit() 868 spin_unlock(&enic->wq_lock[txq_map]); in enic_hard_start_xmit()
|
| /Linux-v5.10/include/linux/avf/ |
| D | virtchnl.h | 363 u16 txq_map; member
|
| /Linux-v5.10/drivers/net/ethernet/marvell/ |
| D | mvneta.c | 1457 int rxq_map = 0, txq_map = 0; in mvneta_defaults_set() local 1466 txq_map |= MVNETA_CPU_TXQ_ACCESS(txq); in mvneta_defaults_set() 1473 txq_map = (cpu == pp->rxq_def) ? in mvneta_defaults_set() 1477 txq_map = MVNETA_CPU_TXQ_ACCESS_ALL_MASK; in mvneta_defaults_set() 1481 mvreg_write(pp, MVNETA_CPU_MAP(cpu), rxq_map | txq_map); in mvneta_defaults_set() 4147 int rxq_map = 0, txq_map = 0; in mvneta_percpu_elect() local 4165 txq_map = (cpu == elected_cpu) ? in mvneta_percpu_elect() 4168 txq_map = mvreg_read(pp, MVNETA_CPU_MAP(cpu)) & in mvneta_percpu_elect() 4171 mvreg_write(pp, MVNETA_CPU_MAP(cpu), rxq_map | txq_map); in mvneta_percpu_elect()
|
| /Linux-v5.10/drivers/net/ethernet/intel/iavf/ |
| D | iavf_virtchnl.c | 382 vecmap->txq_map = q_vector->ring_mask; in iavf_map_queues() 391 vecmap->txq_map = 0; in iavf_map_queues()
|
| /Linux-v5.10/drivers/net/ethernet/intel/i40e/ |
| D | i40e_virtchnl_pf.c | 304 if (vecmap->rxq_map == 0 && vecmap->txq_map == 0) { in i40e_config_irq_link_list() 315 tempmap = vecmap->txq_map; in i40e_config_irq_link_list() 2321 if (i40e_validate_queue_map(vf, vsi_id, map->txq_map)) { in i40e_vc_config_irq_map_msg()
|