/Linux-v5.15/block/ |
D | blk-mq-cpumap.c | 18 static int queue_index(struct blk_mq_queue_map *qmap, in queue_index() argument 21 return qmap->queue_offset + (q % nr_queues); in queue_index() 35 int blk_mq_map_queues(struct blk_mq_queue_map *qmap) in blk_mq_map_queues() argument 37 unsigned int *map = qmap->mq_map; in blk_mq_map_queues() 38 unsigned int nr_queues = qmap->nr_queues; in blk_mq_map_queues() 51 map[cpu] = queue_index(qmap, nr_queues, q++); in blk_mq_map_queues() 64 map[cpu] = queue_index(qmap, nr_queues, q++); in blk_mq_map_queues() 68 map[cpu] = queue_index(qmap, nr_queues, q++); in blk_mq_map_queues() 86 int blk_mq_hw_queue_to_node(struct blk_mq_queue_map *qmap, unsigned int index) in blk_mq_hw_queue_to_node() argument 91 if (index == qmap->mq_map[i]) in blk_mq_hw_queue_to_node()
|
D | blk-mq-pci.c | 26 int blk_mq_pci_map_queues(struct blk_mq_queue_map *qmap, struct pci_dev *pdev, in blk_mq_pci_map_queues() argument 32 for (queue = 0; queue < qmap->nr_queues; queue++) { in blk_mq_pci_map_queues() 38 qmap->mq_map[cpu] = qmap->queue_offset + queue; in blk_mq_pci_map_queues() 44 WARN_ON_ONCE(qmap->nr_queues > 1); in blk_mq_pci_map_queues() 45 blk_mq_clear_mq_map(qmap); in blk_mq_pci_map_queues()
|
D | blk-mq-virtio.c | 24 int blk_mq_virtio_map_queues(struct blk_mq_queue_map *qmap, in blk_mq_virtio_map_queues() argument 33 for (queue = 0; queue < qmap->nr_queues; queue++) { in blk_mq_virtio_map_queues() 39 qmap->mq_map[cpu] = qmap->queue_offset + queue; in blk_mq_virtio_map_queues() 44 return blk_mq_map_queues(qmap); in blk_mq_virtio_map_queues()
|
D | blk-mq.h | 84 extern int blk_mq_hw_queue_to_node(struct blk_mq_queue_map *qmap, unsigned int); 265 static inline void blk_mq_clear_mq_map(struct blk_mq_queue_map *qmap) in blk_mq_clear_mq_map() argument 270 qmap->mq_map[cpu] = 0; in blk_mq_clear_mq_map()
|
/Linux-v5.15/Documentation/ABI/testing/ |
D | sysfs-class-net-qmi | 32 Write a number ranging from 1 to 254 to add a qmap mux 50 created qmap mux based network device. 52 What: /sys/class/net/<qmimux iface>/qmap/mux_id
|
/Linux-v5.15/include/linux/ |
D | blk-mq-pci.h | 8 int blk_mq_pci_map_queues(struct blk_mq_queue_map *qmap, struct pci_dev *pdev,
|
D | blk-mq-virtio.h | 8 int blk_mq_virtio_map_queues(struct blk_mq_queue_map *qmap,
|
D | blk-mq.h | 556 int blk_mq_map_queues(struct blk_mq_queue_map *qmap);
|
/Linux-v5.15/drivers/net/ipa/ |
D | ipa_data-v4.2.c | 105 .qmap = true, 130 .qmap = true,
|
D | ipa_data-v4.11.c | 109 .qmap = true, 133 .qmap = true,
|
D | ipa_data-v3.5.1.c | 115 .qmap = true, 140 .qmap = true,
|
D | ipa_data-v4.5.c | 118 .qmap = true, 142 .qmap = true,
|
D | ipa_data-v4.9.c | 110 .qmap = true, 134 .qmap = true,
|
D | ipa_data.h | 147 bool qmap; member
|
D | ipa_data-v3.1.c | 124 .qmap = true, 148 .qmap = true,
|
D | ipa_endpoint.c | 530 if (endpoint->data->qmap) { in ipa_endpoint_init_hdr() 581 if (endpoint->data->qmap && !endpoint->toward_ipa) { in ipa_endpoint_init_hdr_ext() 597 if (endpoint->data->qmap && !endpoint->toward_ipa) { in ipa_endpoint_init_hdr_ext() 622 if (endpoint->data->qmap) in ipa_endpoint_init_hdr_metadata_mask()
|
D | ipa_modem.c | 130 if (endpoint->data->qmap && skb->protocol != htons(ETH_P_MAP)) in ipa_start_xmit()
|
/Linux-v5.15/drivers/net/ethernet/broadcom/bnxt/ |
D | bnxt_dcb.c | 187 unsigned long qmap = 0; in bnxt_queue_remap() local 196 __set_bit(j, &qmap); in bnxt_queue_remap() 208 j = find_next_zero_bit(&qmap, max, j); in bnxt_queue_remap() 210 __set_bit(j, &qmap); in bnxt_queue_remap()
|
/Linux-v5.15/arch/powerpc/kvm/ |
D | book3s_xive.h | 132 u8 qmap; member
|
D | book3s_xive.c | 376 if (xive->qmap & (1 << prio)) in xive_check_provisioning() 395 xive->qmap |= (1 << prio); in xive_check_provisioning() 1390 if (xive->qmap & (1 << i)) { in kvmppc_xive_connect_vcpu()
|
/Linux-v5.15/drivers/scsi/ |
D | virtio_scsi.c | 717 struct blk_mq_queue_map *qmap = &shost->tag_set.map[HCTX_TYPE_DEFAULT]; in virtscsi_map_queues() local 719 return blk_mq_virtio_map_queues(qmap, vscsi->vdev, 2); in virtscsi_map_queues()
|
/Linux-v5.15/drivers/net/ethernet/intel/ice/ |
D | ice_virtchnl_pf.c | 3420 unsigned long qmap; in ice_cfg_interrupt() local 3425 qmap = map->rxq_map; in ice_cfg_interrupt() 3426 for_each_set_bit(vsi_q_id_idx, &qmap, ICE_MAX_RSS_QS_PER_VF) { in ice_cfg_interrupt() 3439 qmap = map->txq_map; in ice_cfg_interrupt() 3440 for_each_set_bit(vsi_q_id_idx, &qmap, ICE_MAX_RSS_QS_PER_VF) { in ice_cfg_interrupt()
|
D | ice_lib.c | 774 u16 offset = 0, qmap = 0, tx_count = 0, pow = 0; in ice_vsi_setup_q_map() local 833 qmap = ((offset << ICE_AQ_VSI_TC_Q_OFFSET_S) & in ice_vsi_setup_q_map() 839 ctxt->info.tc_mapping[i] = cpu_to_le16(qmap); in ice_vsi_setup_q_map()
|
/Linux-v5.15/drivers/scsi/hisi_sas/ |
D | hisi_sas_v2_hw.c | 3553 struct blk_mq_queue_map *qmap = &shost->tag_set.map[HCTX_TYPE_DEFAULT]; in map_queues_v2_hw() local 3557 for (queue = 0; queue < qmap->nr_queues; queue++) { in map_queues_v2_hw() 3563 qmap->mq_map[cpu] = qmap->queue_offset + queue; in map_queues_v2_hw()
|
/Linux-v5.15/drivers/net/ethernet/intel/i40e/ |
D | i40e_main.c | 1703 u16 qcount = 0, max_qcount, qmap, sections = 0; in i40e_vsi_setup_queue_map_mqprio() local 1719 qmap = (offset << I40E_AQ_VSI_TC_QUE_OFFSET_SHIFT) | in i40e_vsi_setup_queue_map_mqprio() 1749 ctxt->info.tc_mapping[0] = cpu_to_le16(qmap); in i40e_vsi_setup_queue_map_mqprio() 1798 u16 qmap; in i40e_vsi_setup_queue_map() local 1866 qmap = in i40e_vsi_setup_queue_map() 1880 qmap = 0; in i40e_vsi_setup_queue_map() 1882 ctxt->info.tc_mapping[i] = cpu_to_le16(qmap); in i40e_vsi_setup_queue_map() 5882 u16 qcount, qmap, sections = 0; in i40e_channel_setup_queue_map() local 5897 qmap = (offset << I40E_AQ_VSI_TC_QUE_OFFSET_SHIFT) | in i40e_channel_setup_queue_map() 5901 ctxt->info.tc_mapping[0] = cpu_to_le16(qmap); in i40e_channel_setup_queue_map() [all …]
|