Home
last modified time | relevance | path

Searched refs:rings (Results 1 – 25 of 66) sorted by relevance

123

/Linux-v4.19/tools/bpf/bpftool/
Dmap_perf_ring.c179 struct event_ring_info *rings; in do_event_pipe() local
245 rings = calloc(nfds, sizeof(rings[0])); in do_event_pipe()
246 if (!rings) in do_event_pipe()
254 rings[i].cpu = cpu + i; in do_event_pipe()
255 rings[i].key = index + i; in do_event_pipe()
257 rings[i].fd = bpf_perf_event_open(map_fd, rings[i].key, in do_event_pipe()
258 rings[i].cpu); in do_event_pipe()
259 if (rings[i].fd < 0) in do_event_pipe()
262 rings[i].mem = perf_event_mmap(rings[i].fd); in do_event_pipe()
263 if (!rings[i].mem) in do_event_pipe()
[all …]
/Linux-v4.19/drivers/net/wireless/broadcom/brcm80211/brcmfmac/
Dflowring.c152 if (flow->rings[i] == NULL) in brcmf_flowring_create()
170 flow->rings[i] = ring; in brcmf_flowring_create()
182 ring = flow->rings[flowid]; in brcmf_flowring_tid()
202 ring = flow->rings[flowid]; in brcmf_flowring_block()
211 if ((flow->rings[i]) && (i != flowid)) { in brcmf_flowring_block()
212 ring = flow->rings[i]; in brcmf_flowring_block()
222 flow->rings[flowid]->blocked = blocked; in brcmf_flowring_block()
246 ring = flow->rings[flowid]; in brcmf_flowring_delete()
257 flow->rings[flowid] = NULL; in brcmf_flowring_delete()
274 ring = flow->rings[flowid]; in brcmf_flowring_enqueue()
[all …]
Dflowring.h51 struct brcmf_flowring_ring **rings; member
/Linux-v4.19/drivers/gpu/drm/amd/amdgpu/
Damdgpu_ctx.c73 ctx->rings[i].sequence = 1; in amdgpu_ctx_init()
74 ctx->rings[i].fences = &ctx->fences[amdgpu_sched_jobs * i]; in amdgpu_ctx_init()
85 struct amdgpu_ring *ring = adev->rings[i]; in amdgpu_ctx_init()
93 r = drm_sched_entity_init(&ctx->rings[i].entity, in amdgpu_ctx_init()
107 drm_sched_entity_destroy(&ctx->rings[j].entity); in amdgpu_ctx_init()
124 dma_fence_put(ctx->rings[i].fences[j]); in amdgpu_ctx_fini()
177 if (ctx->adev->rings[i] == &ctx->adev->gfx.kiq.ring) in amdgpu_ctx_do_release()
180 drm_sched_entity_destroy(&ctx->rings[i].entity); in amdgpu_ctx_do_release()
340 struct amdgpu_ctx_ring *cring = & ctx->rings[ring->idx]; in amdgpu_ctx_add_fence()
367 struct amdgpu_ctx_ring *cring = & ctx->rings[ring->idx]; in amdgpu_ctx_get_fence()
[all …]
Damdgpu_fence.c492 struct amdgpu_ring *ring = adev->rings[i]; in amdgpu_fence_driver_fini()
526 struct amdgpu_ring *ring = adev->rings[i]; in amdgpu_fence_driver_suspend()
560 struct amdgpu_ring *ring = adev->rings[i]; in amdgpu_fence_driver_resume()
664 struct amdgpu_ring *ring = adev->rings[i]; in amdgpu_debugfs_fence_info()
Damdgpu_ring.c261 adev->rings[ring->idx] = ring; in amdgpu_ring_init()
349 if (!(ring->adev) || !(ring->adev->rings[ring->idx])) in amdgpu_ring_fini()
368 ring->adev->rings[ring->idx] = NULL; in amdgpu_ring_fini()
Damdgpu_test.c49 if (adev->rings[i]) in amdgpu_do_test_moves()
50 n -= adev->rings[i]->ring_size; in amdgpu_do_test_moves()
/Linux-v4.19/net/9p/
Dtrans_xen.c81 struct xen_9pfs_dataring *rings; member
158 ring = &priv->rings[num]; in p9_xen_request()
292 if (!priv->rings[i].intf) in xen_9pfs_front_free()
294 if (priv->rings[i].irq > 0) in xen_9pfs_front_free()
295 unbind_from_irqhandler(priv->rings[i].irq, priv->dev); in xen_9pfs_front_free()
296 if (priv->rings[i].data.in) { in xen_9pfs_front_free()
300 ref = priv->rings[i].intf->ref[j]; in xen_9pfs_front_free()
303 free_pages((unsigned long)priv->rings[i].data.in, in xen_9pfs_front_free()
307 gnttab_end_foreign_access(priv->rings[i].ref, 0, 0); in xen_9pfs_front_free()
308 free_page((unsigned long)priv->rings[i].intf); in xen_9pfs_front_free()
[all …]
/Linux-v4.19/tools/testing/selftests/net/
Dpsock_fanout.c235 static int sock_fanout_read(int fds[], char *rings[], const int expect[]) in sock_fanout_read() argument
239 ret[0] = sock_fanout_read_ring(fds[0], rings[0]); in sock_fanout_read()
240 ret[1] = sock_fanout_read_ring(fds[1], rings[1]); in sock_fanout_read()
362 char *rings[2]; in test_datapath() local
380 rings[0] = sock_fanout_open_ring(fds[0]); in test_datapath()
381 rings[1] = sock_fanout_open_ring(fds[1]); in test_datapath()
384 sock_fanout_read(fds, rings, expect0); in test_datapath()
389 ret = sock_fanout_read(fds, rings, expect1); in test_datapath()
394 ret |= sock_fanout_read(fds, rings, expect2); in test_datapath()
396 if (munmap(rings[1], RING_NUM_FRAMES * getpagesize()) || in test_datapath()
[all …]
/Linux-v4.19/Documentation/networking/
Daf_xdp.rst24 syscall. Associated with each XSK are two rings: the RX ring and the
26 packets on the TX ring. These rings are registered and sized with the
28 to have at least one of these rings for each socket. An RX or TX
37 one of the rings references a frame by referencing its addr. The addr
42 UMEM also has two rings: the FILL ring and the COMPLETION ring. The
50 TX ring. In summary, the RX and FILL rings are used for the RX path
51 and the TX and COMPLETION rings are used for the TX path.
59 corresponding two rings, sets the XDP_SHARED_UMEM flag in the bind
65 process has to create its own socket with associated RX and TX rings,
67 reason that there is only one set of FILL and COMPLETION rings per
[all …]
/Linux-v4.19/drivers/block/xen-blkback/
Dxenbus.c92 if (!blkif->rings || !blkif->rings[0].irq || !blkif->vbd.bdev) in xen_update_blkif_status()
118 ring = &blkif->rings[i]; in xen_update_blkif_status()
132 ring = &blkif->rings[i]; in xen_update_blkif_status()
142 blkif->rings = kcalloc(blkif->nr_rings, sizeof(struct xen_blkif_ring), in xen_blkif_alloc_rings()
144 if (!blkif->rings) in xen_blkif_alloc_rings()
148 struct xen_blkif_ring *ring = &blkif->rings[r]; in xen_blkif_alloc_rings()
251 struct xen_blkif_ring *ring = &blkif->rings[r]; in xen_blkif_disconnect()
315 kfree(blkif->rings); in xen_blkif_disconnect()
316 blkif->rings = NULL; in xen_blkif_disconnect()
359 if (!blkif->rings) \
[all …]
/Linux-v4.19/include/linux/
Dptr_ring.h621 static inline int ptr_ring_resize_multiple(struct ptr_ring **rings, in ptr_ring_resize_multiple() argument
641 spin_lock_irqsave(&(rings[i])->consumer_lock, flags); in ptr_ring_resize_multiple()
642 spin_lock(&(rings[i])->producer_lock); in ptr_ring_resize_multiple()
643 queues[i] = __ptr_ring_swap_queue(rings[i], queues[i], in ptr_ring_resize_multiple()
645 spin_unlock(&(rings[i])->producer_lock); in ptr_ring_resize_multiple()
646 spin_unlock_irqrestore(&(rings[i])->consumer_lock, flags); in ptr_ring_resize_multiple()
Dskb_array.h205 static inline int skb_array_resize_multiple(struct skb_array **rings, in skb_array_resize_multiple() argument
210 return ptr_ring_resize_multiple((struct ptr_ring **)rings, in skb_array_resize_multiple()
/Linux-v4.19/Documentation/devicetree/bindings/soc/qcom/
Dqcom,wcnss.txt67 Definition: should reference the tx-enable and tx-rings-empty SMEM states
72 Definition: must contain "tx-enable" and "tx-rings-empty"
111 qcom,smem-state-names = "tx-enable", "tx-rings-empty";
/Linux-v4.19/drivers/crypto/qat/qat_common/
Dadf_transport.c277 ring = &bank->rings[ring_num]; in adf_create_ring()
340 adf_handle_response(&bank->rings[i]); in adf_ring_response_handler()
413 ring = &bank->rings[i]; in adf_init_bank()
427 tx_ring = &bank->rings[i - hw_data->tx_rx_gap]; in adf_init_bank()
442 ring = &bank->rings[i]; in adf_init_bank()
523 struct adf_etr_ring_data *ring = &bank->rings[i]; in cleanup_bank()
Dadf_transport_internal.h76 struct adf_etr_ring_data rings[ADF_ETR_MAX_RINGS_PER_BANK]; member
/Linux-v4.19/drivers/crypto/inside-secure/
Dsafexcel.c219 for (i = 0; i < priv->config.rings; i++) { in safexcel_hw_setup_cdesc_rings()
257 for (i = 0; i < priv->config.rings; i++) { in safexcel_hw_setup_rdesc_rings()
363 GENMASK(priv->config.rings - 1, 0), in safexcel_hw_init()
415 for (i = 0; i < priv->config.rings; i++) { in safexcel_hw_init()
441 for (i = 0; i < priv->config.rings; i++) { in safexcel_hw_init()
465 writel(EIP197_DxE_THR_CTRL_EN | GENMASK(priv->config.rings - 1, 0), in safexcel_hw_init()
469 writel(EIP197_DxE_THR_CTRL_EN | GENMASK(priv->config.rings - 1, 0), in safexcel_hw_init()
937 priv->config.rings = min_t(u32, val & GENMASK(3, 0), max_rings); in safexcel_configure()
1047 priv->ring = devm_kcalloc(dev, priv->config.rings, in safexcel_probe()
1055 for (i = 0; i < priv->config.rings; i++) { in safexcel_probe()
[all …]
/Linux-v4.19/Documentation/devicetree/bindings/net/
Dopencores-ethoc.txt6 first region is for the device registers and descriptor rings,
/Linux-v4.19/Documentation/devicetree/bindings/crypto/
Dmediatek-crypto.txt7 order. These are global system and four descriptor rings.
Dbrcm,spu-crypto.txt15 Mailbox channels correspond to DMA rings on the device.
Dinside-secure-safexcel.txt8 - interrupts: Interrupt numbers for the rings and engine.
/Linux-v4.19/drivers/net/
Dtap.c1222 struct ptr_ring **rings; in tap_queue_resize() local
1226 rings = kmalloc_array(n, sizeof(*rings), GFP_KERNEL); in tap_queue_resize()
1227 if (!rings) in tap_queue_resize()
1231 rings[i++] = &q->ring; in tap_queue_resize()
1233 ret = ptr_ring_resize_multiple(rings, n, in tap_queue_resize()
1237 kfree(rings); in tap_queue_resize()
/Linux-v4.19/drivers/mailbox/
Dbcm-flexrm-mailbox.c295 struct flexrm_ring *rings; member
943 ring = &mbox->rings[i]; in flexrm_write_config_in_seqfile()
971 ring = &mbox->rings[i]; in flexrm_write_stats_in_seqfile()
1491 struct flexrm_ring *ring = &mbox->rings[desc->platform.msi_index]; in flexrm_mbox_msi_write()
1552 mbox->rings = ring; in flexrm_mbox_probe()
1557 ring = &mbox->rings[index]; in flexrm_mbox_probe()
1617 ring = &mbox->rings[desc->platform.msi_index]; in flexrm_mbox_probe()
1665 mbox->controller.chans[index].con_priv = &mbox->rings[index]; in flexrm_mbox_probe()
/Linux-v4.19/tools/bpf/bpftool/Documentation/
Dbpftool-map.rst83 Install perf rings into a perf event array map and dump
94 receiving events if it installed its rings earlier.
/Linux-v4.19/drivers/crypto/caam/
DKconfig101 number of job rings assigned to the kernel. The number of portals
103 job rings.

123