/Linux-v5.4/net/sched/ |
D | sch_multiq.c | 25 struct Qdisc **queues; member 54 return q->queues[0]; in multiq_classify() 56 return q->queues[band]; in multiq_classify() 105 qdisc = q->queues[q->curband]; in multiq_dequeue() 137 qdisc = q->queues[curband]; in multiq_peek() 154 qdisc_reset(q->queues[band]); in multiq_reset() 167 qdisc_put(q->queues[band]); in multiq_destroy() 169 kfree(q->queues); in multiq_destroy() 197 if (q->queues[i] != &noop_qdisc) { in multiq_tune() 198 struct Qdisc *child = q->queues[i]; in multiq_tune() [all …]
|
D | sch_prio.c | 26 struct Qdisc *queues[TCQ_PRIO_BANDS]; member 57 return q->queues[q->prio2band[band & TC_PRIO_MAX]]; in prio_classify() 63 return q->queues[q->prio2band[0]]; in prio_classify() 65 return q->queues[band]; in prio_classify() 103 struct Qdisc *qdisc = q->queues[prio]; in prio_peek() 117 struct Qdisc *qdisc = q->queues[prio]; in prio_dequeue() 137 qdisc_reset(q->queues[prio]); in prio_reset() 175 qdisc_put(q->queues[prio]); in prio_destroy() 182 struct Qdisc *queues[TCQ_PRIO_BANDS]; in prio_tune() local 200 queues[i] = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, in prio_tune() [all …]
|
/Linux-v5.4/drivers/scsi/aacraid/ |
D | comminit.c | 374 struct aac_entry * queues; in aac_comm_init() local 376 struct aac_queue_block * comm = dev->queues; in aac_comm_init() 395 queues = (struct aac_entry *)(((ulong)headers) + hdrsize); in aac_comm_init() 398 comm->queue[HostNormCmdQueue].base = queues; in aac_comm_init() 400 queues += HOST_NORM_CMD_ENTRIES; in aac_comm_init() 404 comm->queue[HostHighCmdQueue].base = queues; in aac_comm_init() 407 queues += HOST_HIGH_CMD_ENTRIES; in aac_comm_init() 411 comm->queue[AdapNormCmdQueue].base = queues; in aac_comm_init() 414 queues += ADAP_NORM_CMD_ENTRIES; in aac_comm_init() 418 comm->queue[AdapHighCmdQueue].base = queues; in aac_comm_init() [all …]
|
/Linux-v5.4/tools/perf/util/ |
D | arm-spe.c | 30 struct auxtrace_queues queues; member 122 err = auxtrace_queues__add_event(&spe->queues, session, event, in arm_spe_process_auxtrace_event() 158 struct auxtrace_queues *queues = &spe->queues; in arm_spe_free_events() local 161 for (i = 0; i < queues->nr_queues; i++) { in arm_spe_free_events() 162 arm_spe_free_queue(queues->queue_array[i].priv); in arm_spe_free_events() 163 queues->queue_array[i].priv = NULL; in arm_spe_free_events() 165 auxtrace_queues__free(queues); in arm_spe_free_events() 207 err = auxtrace_queues__init(&spe->queues); in arm_spe_process_auxtrace_info()
|
D | intel-bts.c | 46 struct auxtrace_queues queues; member 211 for (i = 0; i < bts->queues.nr_queues; i++) { in intel_bts_setup_queues() 212 ret = intel_bts_setup_queue(bts, &bts->queues.queue_array[i], in intel_bts_setup_queues() 222 if (bts->queues.new_data) { in intel_bts_update_queues() 223 bts->queues.new_data = false; in intel_bts_update_queues() 465 queue = &btsq->bts->queues.queue_array[btsq->queue_nr]; in intel_bts_process_queue() 539 struct auxtrace_queues *queues = &bts->queues; in intel_bts_process_tid_exit() local 542 for (i = 0; i < queues->nr_queues; i++) { in intel_bts_process_tid_exit() 543 struct auxtrace_queue *queue = &bts->queues.queue_array[i]; in intel_bts_process_tid_exit() 568 queue = &bts->queues.queue_array[queue_nr]; in intel_bts_process_queues() [all …]
|
D | s390-cpumsf.c | 169 struct auxtrace_queues queues; member 202 if (!sf->use_logfile || sf->queues.nr_queues <= sample->cpu) in s390_cpumcf_dumpctr() 205 q = &sf->queues.queue_array[sample->cpu]; in s390_cpumcf_dumpctr() 700 queue = &sfq->sf->queues.queue_array[sfq->queue_nr]; in s390_cpumsf_run_decoder() 824 for (i = 0; i < sf->queues.nr_queues; i++) { in s390_cpumsf_setup_queues() 825 ret = s390_cpumsf_setup_queue(sf, &sf->queues.queue_array[i], in s390_cpumsf_setup_queues() 835 if (!sf->queues.new_data) in s390_cpumsf_update_queues() 838 sf->queues.new_data = false; in s390_cpumsf_update_queues() 859 queue = &sf->queues.queue_array[queue_nr]; in s390_cpumsf_process_queues() 984 err = auxtrace_queues__add_event(&sf->queues, session, event, in s390_cpumsf_process_auxtrace_event() [all …]
|
D | auxtrace.c | 170 int auxtrace_queues__init(struct auxtrace_queues *queues) in auxtrace_queues__init() argument 172 queues->nr_queues = AUXTRACE_INIT_NR_QUEUES; in auxtrace_queues__init() 173 queues->queue_array = auxtrace_alloc_queue_array(queues->nr_queues); in auxtrace_queues__init() 174 if (!queues->queue_array) in auxtrace_queues__init() 179 static int auxtrace_queues__grow(struct auxtrace_queues *queues, in auxtrace_queues__grow() argument 182 unsigned int nr_queues = queues->nr_queues; in auxtrace_queues__grow() 192 if (nr_queues < queues->nr_queues || nr_queues < new_nr_queues) in auxtrace_queues__grow() 199 for (i = 0; i < queues->nr_queues; i++) { in auxtrace_queues__grow() 200 list_splice_tail(&queues->queue_array[i].head, in auxtrace_queues__grow() 202 queue_array[i].tid = queues->queue_array[i].tid; in auxtrace_queues__grow() [all …]
|
/Linux-v5.4/Documentation/ABI/testing/ |
D | sysfs-class-net-queues | 1 What: /sys/class/<iface>/queues/rx-<queue>/rps_cpus 11 What: /sys/class/<iface>/queues/rx-<queue>/rps_flow_cnt 19 What: /sys/class/<iface>/queues/tx-<queue>/tx_timeout 27 What: /sys/class/<iface>/queues/tx-<queue>/tx_maxrate 35 What: /sys/class/<iface>/queues/tx-<queue>/xps_cpus 45 What: /sys/class/<iface>/queues/tx-<queue>/xps_rxqs 56 What: /sys/class/<iface>/queues/tx-<queue>/byte_queue_limits/hold_time 65 What: /sys/class/<iface>/queues/tx-<queue>/byte_queue_limits/inflight 73 What: /sys/class/<iface>/queues/tx-<queue>/byte_queue_limits/limit 82 What: /sys/class/<iface>/queues/tx-<queue>/byte_queue_limits/limit_max [all …]
|
/Linux-v5.4/drivers/nvme/target/ |
D | loop.c | 30 struct nvme_loop_queue *queues; member 72 return queue - queue->ctrl->queues; in nvme_loop_queue_idx() 177 struct nvme_loop_queue *queue = &ctrl->queues[0]; in nvme_loop_submit_async_event() 199 iod->queue = &ctrl->queues[queue_idx]; in nvme_loop_init_iod() 219 struct nvme_loop_queue *queue = &ctrl->queues[hctx_idx + 1]; in nvme_loop_init_hctx() 231 struct nvme_loop_queue *queue = &ctrl->queues[0]; in nvme_loop_init_admin_hctx() 255 clear_bit(NVME_LOOP_Q_LIVE, &ctrl->queues[0].flags); in nvme_loop_destroy_admin_queue() 256 nvmet_sq_destroy(&ctrl->queues[0].nvme_sq); in nvme_loop_destroy_admin_queue() 277 kfree(ctrl->queues); in nvme_loop_free_ctrl() 288 clear_bit(NVME_LOOP_Q_LIVE, &ctrl->queues[i].flags); in nvme_loop_destroy_io_queues() [all …]
|
/Linux-v5.4/Documentation/devicetree/bindings/soc/ti/ |
D | keystone-navigator-qmss.txt | 9 management of the packet queues. Packets are queued/de-queued by writing or 32 -- managed-queues : the actual queues managed by each queue manager 33 instance, specified as <"base queue #" "# of queues">. 51 - qpend : pool of qpend(interruptible) queues 52 - general-purpose : pool of general queues, primarily used 53 as free descriptor queues or the 54 transmit DMA queues. 55 - accumulator : pool of queues on PDSP accumulator channel 57 -- qrange : number of queues to use per queue range, specified as 58 <"base queue #" "# of queues">. [all …]
|
/Linux-v5.4/Documentation/networking/device_drivers/ti/ |
D | cpsw.txt | 21 - TX queues must be rated starting from txq0 that has highest priority 23 - CBS shapers should be used with rated queues 25 potential incoming rate, thus, rate of all incoming tx queues has 139 // Add 4 tx queues, for interface Eth0, and 1 tx queue for Eth1 144 // Check if num of queues is set correctly: 159 // TX queues must be rated starting from 0, so set bws for tx0 and tx1 162 // Leave last 2 tx queues not rated. 163 $ echo 40 > /sys/class/net/eth0/queues/tx-0/tx_maxrate 164 $ echo 20 > /sys/class/net/eth0/queues/tx-1/tx_maxrate 167 // Check maximum rate of tx (cpdma) queues: [all …]
|
/Linux-v5.4/Documentation/arm/keystone/ |
D | knav-qmss.rst | 15 management of the packet queues. Packets are queued/de-queued by writing or 24 knav qmss driver provides a set of APIs to drivers to open/close qmss queues, 25 allocate descriptor pools, map the descriptors, push/pop to queues etc. For 31 Accumulator QMSS queues using PDSP firmware 34 queue or multiple contiguous queues. drivers/soc/ti/knav_qmss_acc.c is the 37 1 or 32 queues per channel. More description on the firmware is available in 56 Use of accumulated queues requires the firmware image to be present in the 57 file system. The driver doesn't acc queues to the supported queue range if
|
/Linux-v5.4/include/linux/ |
D | ptr_ring.h | 624 void ***queues; in ptr_ring_resize_multiple() local 627 queues = kmalloc_array(nrings, sizeof(*queues), gfp); in ptr_ring_resize_multiple() 628 if (!queues) in ptr_ring_resize_multiple() 632 queues[i] = __ptr_ring_init_queue_alloc(size, gfp); in ptr_ring_resize_multiple() 633 if (!queues[i]) in ptr_ring_resize_multiple() 640 queues[i] = __ptr_ring_swap_queue(rings[i], queues[i], in ptr_ring_resize_multiple() 647 kvfree(queues[i]); in ptr_ring_resize_multiple() 649 kfree(queues); in ptr_ring_resize_multiple() 655 kvfree(queues[i]); in ptr_ring_resize_multiple() 657 kfree(queues); in ptr_ring_resize_multiple()
|
/Linux-v5.4/drivers/gpu/drm/msm/adreno/ |
D | a6xx_hfi.c | 85 struct a6xx_hfi_queue *queue = &gmu->queues[HFI_RESPONSE_QUEUE]; in a6xx_hfi_wait_for_ack() 154 struct a6xx_hfi_queue *queue = &gmu->queues[HFI_COMMAND_QUEUE]; in a6xx_hfi_send_msg() 313 for (i = 0; i < ARRAY_SIZE(gmu->queues); i++) { in a6xx_hfi_stop() 314 struct a6xx_hfi_queue *queue = &gmu->queues[i]; in a6xx_hfi_stop() 364 table_size += (ARRAY_SIZE(gmu->queues) * in a6xx_hfi_init() 372 table->num_queues = ARRAY_SIZE(gmu->queues); in a6xx_hfi_init() 373 table->active_queues = ARRAY_SIZE(gmu->queues); in a6xx_hfi_init() 377 a6xx_hfi_queue_init(&gmu->queues[0], &headers[0], hfi->virt + offset, in a6xx_hfi_init() 382 a6xx_hfi_queue_init(&gmu->queues[1], &headers[1], hfi->virt + offset, in a6xx_hfi_init()
|
/Linux-v5.4/Documentation/networking/ |
D | multiqueue.txt | 18 the subqueue memory, as well as netdev configuration of where the queues 21 The base driver will also need to manage the queues as it does the global 34 A new round-robin qdisc, sch_multiq also supports multiple hardware queues. The 36 bands and queues based on the value in skb->queue_mapping. Use this field in 43 On qdisc load, the number of bands is based on the number of queues on the 57 The qdisc will allocate the number of bands to equal the number of queues that 59 queues, the band mapping would look like:
|
D | scaling.rst | 27 Contemporary NICs support multiple receive and transmit descriptor queues 29 queues to distribute processing among CPUs. The NIC distributes packets by 47 Some advanced NICs allow steering packets to queues based on 57 module parameter for specifying the number of hardware queues to 60 for each CPU if the device supports enough queues, or otherwise at least 66 default mapping is to distribute the queues evenly in the table, but the 69 indirection table could be done to give different queues different 80 of queues to IRQs can be determined from /proc/interrupts. By default, 95 is to allocate as many queues as there are CPUs in the system (or the 97 is likely the one with the smallest number of receive queues where no [all …]
|
/Linux-v5.4/drivers/staging/qlge/ |
D | TODO | 2 queues become empty because of allocation failures (ex. under memory 16 * rename "rx" queues to "completion" queues. Calling tx completion queues "rx 17 queues" is confusing. 27 frames, resets the link, device and driver buffer queues become
|
/Linux-v5.4/Documentation/devicetree/bindings/dma/ |
D | fsl-qdma.txt | 22 - fsl,dma-queues: Should contain number of queues supported. 28 based on queues 52 fsl,dma-queues = <2>;
|
/Linux-v5.4/Documentation/networking/device_drivers/freescale/dpaa2/ |
D | ethernet-driver.rst | 25 - queues, channels 32 hardware resources, like queues, do not have a corresponding MC object and 98 queues ---------------------- | | Buffer pool | 108 Frames are transmitted and received through hardware frame queues, which can be 110 enqueues TX frames on egress queues and after transmission is complete a TX 113 When frames are available on ingress queues, a data availability notification 115 queues in the same channel have available frames, only one notification is sent. 118 Each network interface can have multiple Rx, Tx and confirmation queues affined
|
/Linux-v5.4/drivers/net/ethernet/netronome/nfp/ |
D | nfp_net_debugfs.c | 150 struct dentry *queues, *tx, *rx, *xdp; in nfp_net_debugfs_vnic_add() local 164 queues = debugfs_create_dir("queue", nn->debugfs_dir); in nfp_net_debugfs_vnic_add() 166 rx = debugfs_create_dir("rx", queues); in nfp_net_debugfs_vnic_add() 167 tx = debugfs_create_dir("tx", queues); in nfp_net_debugfs_vnic_add() 168 xdp = debugfs_create_dir("xdp", queues); in nfp_net_debugfs_vnic_add()
|
/Linux-v5.4/Documentation/devicetree/bindings/mfd/ |
D | fsl-imx25-tsadc.txt | 3 This device combines two general purpose conversion queues one used for general 15 conversion queues. 20 This device includes two conversion queues which can be added as subnodes.
|
/Linux-v5.4/Documentation/devicetree/bindings/net/ |
D | fsl-fec.txt | 14 - fsl,num-tx-queues : The property is valid for enet-avb IP, which supports 15 hw multi queues. Should specify the tx queue number, otherwise set tx queue 17 - fsl,num-rx-queues : The property is valid for enet-avb IP, which supports 18 hw multi queues. Should specify the rx queue number, otherwise set rx queue 34 tx/rx queues 1 and 2. "int0" will be used for queue 0 and ENET_MII interrupts. 35 For imx6sx, "int0" handles all 3 queues and ENET_MII. "pps" is for the pulse
|
D | brcm,systemport.txt | 10 interrupts, and the second cell should be for the transmit queues. An 21 - systemport,num-txq: number of HW transmit queues, an integer 22 - systemport,num-rxq: number of HW receive queues, an integer
|
/Linux-v5.4/drivers/net/xen-netback/ |
D | interface.c | 202 queue = &vif->queues[index]; in xenvif_start_xmit() 254 queue = &vif->queues[index]; in xenvif_get_stats() 278 queue = &vif->queues[queue_index]; in xenvif_up() 294 queue = &vif->queues[queue_index]; in xenvif_down() 407 void *vif_stats = &vif->queues[queue_index].stats; in xenvif_get_ethtool_stats() 483 vif->queues = NULL; in xenvif_alloc() 747 queue = &vif->queues[queue_index]; in xenvif_disconnect_data() 804 struct xenvif_queue *queues = vif->queues; in xenvif_free() local 812 xenvif_deinit_queue(&queues[queue_index]); in xenvif_free() 813 vfree(queues); in xenvif_free()
|
/Linux-v5.4/Documentation/networking/device_drivers/freescale/ |
D | dpaa.txt | 81 Tx FQs = transmission frame queues 137 confirmation frame queues. The driver is then responsible for freeing the 158 strict priority levels. Each traffic class contains NR_CPU TX queues. By 159 default, only one traffic class is enabled and the lowest priority Tx queues 176 Traffic coming on the DPAA Rx queues or on the DPAA Tx confirmation 177 queues is seen by the CPU as ingress traffic on a certain portal. 183 hardware frame queues using a hash on IP v4/v6 source and destination 187 queues are configured to put the received traffic into a pool channel 189 The default frame queues have the HOLDACTIVE option set, ensuring that 196 128 Rx frame queues that are configured to dedicated channels, in a [all …]
|