/Linux-v4.19/tools/bpf/bpftool/ |
D | map_perf_ring.c | 179 struct event_ring_info *rings; in do_event_pipe() local 245 rings = calloc(nfds, sizeof(rings[0])); in do_event_pipe() 246 if (!rings) in do_event_pipe() 254 rings[i].cpu = cpu + i; in do_event_pipe() 255 rings[i].key = index + i; in do_event_pipe() 257 rings[i].fd = bpf_perf_event_open(map_fd, rings[i].key, in do_event_pipe() 258 rings[i].cpu); in do_event_pipe() 259 if (rings[i].fd < 0) in do_event_pipe() 262 rings[i].mem = perf_event_mmap(rings[i].fd); in do_event_pipe() 263 if (!rings[i].mem) in do_event_pipe() [all …]
|
/Linux-v4.19/drivers/net/wireless/broadcom/brcm80211/brcmfmac/ |
D | flowring.c | 152 if (flow->rings[i] == NULL) in brcmf_flowring_create() 170 flow->rings[i] = ring; in brcmf_flowring_create() 182 ring = flow->rings[flowid]; in brcmf_flowring_tid() 202 ring = flow->rings[flowid]; in brcmf_flowring_block() 211 if ((flow->rings[i]) && (i != flowid)) { in brcmf_flowring_block() 212 ring = flow->rings[i]; in brcmf_flowring_block() 222 flow->rings[flowid]->blocked = blocked; in brcmf_flowring_block() 246 ring = flow->rings[flowid]; in brcmf_flowring_delete() 257 flow->rings[flowid] = NULL; in brcmf_flowring_delete() 274 ring = flow->rings[flowid]; in brcmf_flowring_enqueue() [all …]
|
D | flowring.h | 51 struct brcmf_flowring_ring **rings; member
|
/Linux-v4.19/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_ctx.c | 73 ctx->rings[i].sequence = 1; in amdgpu_ctx_init() 74 ctx->rings[i].fences = &ctx->fences[amdgpu_sched_jobs * i]; in amdgpu_ctx_init() 85 struct amdgpu_ring *ring = adev->rings[i]; in amdgpu_ctx_init() 93 r = drm_sched_entity_init(&ctx->rings[i].entity, in amdgpu_ctx_init() 107 drm_sched_entity_destroy(&ctx->rings[j].entity); in amdgpu_ctx_init() 124 dma_fence_put(ctx->rings[i].fences[j]); in amdgpu_ctx_fini() 177 if (ctx->adev->rings[i] == &ctx->adev->gfx.kiq.ring) in amdgpu_ctx_do_release() 180 drm_sched_entity_destroy(&ctx->rings[i].entity); in amdgpu_ctx_do_release() 340 struct amdgpu_ctx_ring *cring = & ctx->rings[ring->idx]; in amdgpu_ctx_add_fence() 367 struct amdgpu_ctx_ring *cring = & ctx->rings[ring->idx]; in amdgpu_ctx_get_fence() [all …]
|
D | amdgpu_fence.c | 492 struct amdgpu_ring *ring = adev->rings[i]; in amdgpu_fence_driver_fini() 526 struct amdgpu_ring *ring = adev->rings[i]; in amdgpu_fence_driver_suspend() 560 struct amdgpu_ring *ring = adev->rings[i]; in amdgpu_fence_driver_resume() 664 struct amdgpu_ring *ring = adev->rings[i]; in amdgpu_debugfs_fence_info()
|
D | amdgpu_ring.c | 261 adev->rings[ring->idx] = ring; in amdgpu_ring_init() 349 if (!(ring->adev) || !(ring->adev->rings[ring->idx])) in amdgpu_ring_fini() 368 ring->adev->rings[ring->idx] = NULL; in amdgpu_ring_fini()
|
D | amdgpu_test.c | 49 if (adev->rings[i]) in amdgpu_do_test_moves() 50 n -= adev->rings[i]->ring_size; in amdgpu_do_test_moves()
|
/Linux-v4.19/net/9p/ |
D | trans_xen.c | 81 struct xen_9pfs_dataring *rings; member 158 ring = &priv->rings[num]; in p9_xen_request() 292 if (!priv->rings[i].intf) in xen_9pfs_front_free() 294 if (priv->rings[i].irq > 0) in xen_9pfs_front_free() 295 unbind_from_irqhandler(priv->rings[i].irq, priv->dev); in xen_9pfs_front_free() 296 if (priv->rings[i].data.in) { in xen_9pfs_front_free() 300 ref = priv->rings[i].intf->ref[j]; in xen_9pfs_front_free() 303 free_pages((unsigned long)priv->rings[i].data.in, in xen_9pfs_front_free() 307 gnttab_end_foreign_access(priv->rings[i].ref, 0, 0); in xen_9pfs_front_free() 308 free_page((unsigned long)priv->rings[i].intf); in xen_9pfs_front_free() [all …]
|
/Linux-v4.19/tools/testing/selftests/net/ |
D | psock_fanout.c | 235 static int sock_fanout_read(int fds[], char *rings[], const int expect[]) in sock_fanout_read() argument 239 ret[0] = sock_fanout_read_ring(fds[0], rings[0]); in sock_fanout_read() 240 ret[1] = sock_fanout_read_ring(fds[1], rings[1]); in sock_fanout_read() 362 char *rings[2]; in test_datapath() local 380 rings[0] = sock_fanout_open_ring(fds[0]); in test_datapath() 381 rings[1] = sock_fanout_open_ring(fds[1]); in test_datapath() 384 sock_fanout_read(fds, rings, expect0); in test_datapath() 389 ret = sock_fanout_read(fds, rings, expect1); in test_datapath() 394 ret |= sock_fanout_read(fds, rings, expect2); in test_datapath() 396 if (munmap(rings[1], RING_NUM_FRAMES * getpagesize()) || in test_datapath() [all …]
|
/Linux-v4.19/Documentation/networking/ |
D | af_xdp.rst | 24 syscall. Associated with each XSK are two rings: the RX ring and the 26 packets on the TX ring. These rings are registered and sized with the 28 to have at least one of these rings for each socket. An RX or TX 37 one of the rings references a frame by referencing its addr. The addr 42 UMEM also has two rings: the FILL ring and the COMPLETION ring. The 50 TX ring. In summary, the RX and FILL rings are used for the RX path 51 and the TX and COMPLETION rings are used for the TX path. 59 corresponding two rings, sets the XDP_SHARED_UMEM flag in the bind 65 process has to create its own socket with associated RX and TX rings, 67 reason that there is only one set of FILL and COMPLETION rings per [all …]
|
/Linux-v4.19/drivers/block/xen-blkback/ |
D | xenbus.c | 92 if (!blkif->rings || !blkif->rings[0].irq || !blkif->vbd.bdev) in xen_update_blkif_status() 118 ring = &blkif->rings[i]; in xen_update_blkif_status() 132 ring = &blkif->rings[i]; in xen_update_blkif_status() 142 blkif->rings = kcalloc(blkif->nr_rings, sizeof(struct xen_blkif_ring), in xen_blkif_alloc_rings() 144 if (!blkif->rings) in xen_blkif_alloc_rings() 148 struct xen_blkif_ring *ring = &blkif->rings[r]; in xen_blkif_alloc_rings() 251 struct xen_blkif_ring *ring = &blkif->rings[r]; in xen_blkif_disconnect() 315 kfree(blkif->rings); in xen_blkif_disconnect() 316 blkif->rings = NULL; in xen_blkif_disconnect() 359 if (!blkif->rings) \ [all …]
|
/Linux-v4.19/include/linux/ |
D | ptr_ring.h | 621 static inline int ptr_ring_resize_multiple(struct ptr_ring **rings, in ptr_ring_resize_multiple() argument 641 spin_lock_irqsave(&(rings[i])->consumer_lock, flags); in ptr_ring_resize_multiple() 642 spin_lock(&(rings[i])->producer_lock); in ptr_ring_resize_multiple() 643 queues[i] = __ptr_ring_swap_queue(rings[i], queues[i], in ptr_ring_resize_multiple() 645 spin_unlock(&(rings[i])->producer_lock); in ptr_ring_resize_multiple() 646 spin_unlock_irqrestore(&(rings[i])->consumer_lock, flags); in ptr_ring_resize_multiple()
|
D | skb_array.h | 205 static inline int skb_array_resize_multiple(struct skb_array **rings, in skb_array_resize_multiple() argument 210 return ptr_ring_resize_multiple((struct ptr_ring **)rings, in skb_array_resize_multiple()
|
/Linux-v4.19/Documentation/devicetree/bindings/soc/qcom/ |
D | qcom,wcnss.txt | 67 Definition: should reference the tx-enable and tx-rings-empty SMEM states 72 Definition: must contain "tx-enable" and "tx-rings-empty" 111 qcom,smem-state-names = "tx-enable", "tx-rings-empty";
|
/Linux-v4.19/drivers/crypto/qat/qat_common/ |
D | adf_transport.c | 277 ring = &bank->rings[ring_num]; in adf_create_ring() 340 adf_handle_response(&bank->rings[i]); in adf_ring_response_handler() 413 ring = &bank->rings[i]; in adf_init_bank() 427 tx_ring = &bank->rings[i - hw_data->tx_rx_gap]; in adf_init_bank() 442 ring = &bank->rings[i]; in adf_init_bank() 523 struct adf_etr_ring_data *ring = &bank->rings[i]; in cleanup_bank()
|
D | adf_transport_internal.h | 76 struct adf_etr_ring_data rings[ADF_ETR_MAX_RINGS_PER_BANK]; member
|
/Linux-v4.19/drivers/crypto/inside-secure/ |
D | safexcel.c | 219 for (i = 0; i < priv->config.rings; i++) { in safexcel_hw_setup_cdesc_rings() 257 for (i = 0; i < priv->config.rings; i++) { in safexcel_hw_setup_rdesc_rings() 363 GENMASK(priv->config.rings - 1, 0), in safexcel_hw_init() 415 for (i = 0; i < priv->config.rings; i++) { in safexcel_hw_init() 441 for (i = 0; i < priv->config.rings; i++) { in safexcel_hw_init() 465 writel(EIP197_DxE_THR_CTRL_EN | GENMASK(priv->config.rings - 1, 0), in safexcel_hw_init() 469 writel(EIP197_DxE_THR_CTRL_EN | GENMASK(priv->config.rings - 1, 0), in safexcel_hw_init() 937 priv->config.rings = min_t(u32, val & GENMASK(3, 0), max_rings); in safexcel_configure() 1047 priv->ring = devm_kcalloc(dev, priv->config.rings, in safexcel_probe() 1055 for (i = 0; i < priv->config.rings; i++) { in safexcel_probe() [all …]
|
/Linux-v4.19/Documentation/devicetree/bindings/net/ |
D | opencores-ethoc.txt | 6 first region is for the device registers and descriptor rings,
|
/Linux-v4.19/Documentation/devicetree/bindings/crypto/ |
D | mediatek-crypto.txt | 7 order. These are global system and four descriptor rings.
|
D | brcm,spu-crypto.txt | 15 Mailbox channels correspond to DMA rings on the device.
|
D | inside-secure-safexcel.txt | 8 - interrupts: Interrupt numbers for the rings and engine.
|
/Linux-v4.19/drivers/net/ |
D | tap.c | 1222 struct ptr_ring **rings; in tap_queue_resize() local 1226 rings = kmalloc_array(n, sizeof(*rings), GFP_KERNEL); in tap_queue_resize() 1227 if (!rings) in tap_queue_resize() 1231 rings[i++] = &q->ring; in tap_queue_resize() 1233 ret = ptr_ring_resize_multiple(rings, n, in tap_queue_resize() 1237 kfree(rings); in tap_queue_resize()
|
/Linux-v4.19/drivers/mailbox/ |
D | bcm-flexrm-mailbox.c | 295 struct flexrm_ring *rings; member 943 ring = &mbox->rings[i]; in flexrm_write_config_in_seqfile() 971 ring = &mbox->rings[i]; in flexrm_write_stats_in_seqfile() 1491 struct flexrm_ring *ring = &mbox->rings[desc->platform.msi_index]; in flexrm_mbox_msi_write() 1552 mbox->rings = ring; in flexrm_mbox_probe() 1557 ring = &mbox->rings[index]; in flexrm_mbox_probe() 1617 ring = &mbox->rings[desc->platform.msi_index]; in flexrm_mbox_probe() 1665 mbox->controller.chans[index].con_priv = &mbox->rings[index]; in flexrm_mbox_probe()
|
/Linux-v4.19/tools/bpf/bpftool/Documentation/ |
D | bpftool-map.rst | 83 Install perf rings into a perf event array map and dump 94 receiving events if it installed its rings earlier.
|
/Linux-v4.19/drivers/crypto/caam/ |
D | Kconfig | 101 number of job rings assigned to the kernel. The number of portals 103 job rings.
|