Home
last modified time | relevance | path

Searched refs:ndw (Results 1 – 17 of 17) sorted by relevance

/Linux-v5.4/drivers/gpu/drm/amd/amdgpu/
Damdgpu_vm_sdma.c64 unsigned int ndw = AMDGPU_VM_SDMA_MIN_NUM_DW; in amdgpu_vm_sdma_prepare() local
67 r = amdgpu_job_alloc_with_ib(p->adev, ndw * 4, &p->job); in amdgpu_vm_sdma_prepare()
80 p->num_dw_left = ndw; in amdgpu_vm_sdma_prepare()
200 unsigned int i, ndw, nptes; in amdgpu_vm_sdma_update() local
205 ndw = p->num_dw_left; in amdgpu_vm_sdma_update()
206 ndw -= p->job->ibs->length_dw; in amdgpu_vm_sdma_update()
208 if (ndw < 32) { in amdgpu_vm_sdma_update()
214 ndw = 32; in amdgpu_vm_sdma_update()
216 ndw += count * 2; in amdgpu_vm_sdma_update()
217 ndw = max(ndw, AMDGPU_VM_SDMA_MIN_NUM_DW); in amdgpu_vm_sdma_update()
[all …]
Dsi_dma.c343 unsigned ndw = count * 2; in si_dma_vm_write_pte() local
345 ib->ptr[ib->length_dw++] = DMA_PACKET(DMA_PACKET_WRITE, 0, 0, 0, ndw); in si_dma_vm_write_pte()
348 for (; ndw > 0; ndw -= 2) { in si_dma_vm_write_pte()
373 unsigned ndw; in si_dma_vm_set_pte_pde() local
376 ndw = count * 2; in si_dma_vm_set_pte_pde()
377 if (ndw > 0xFFFFE) in si_dma_vm_set_pte_pde()
378 ndw = 0xFFFFE; in si_dma_vm_set_pte_pde()
386 ib->ptr[ib->length_dw++] = DMA_PTE_PDE_PACKET(ndw); in si_dma_vm_set_pte_pde()
395 pe += ndw * 4; in si_dma_vm_set_pte_pde()
396 addr += (ndw / 2) * incr; in si_dma_vm_set_pte_pde()
[all …]
Damdgpu_ring.c65 int amdgpu_ring_alloc(struct amdgpu_ring *ring, unsigned ndw) in amdgpu_ring_alloc() argument
69 ndw = (ndw + ring->funcs->align_mask) & ~ring->funcs->align_mask; in amdgpu_ring_alloc()
74 if (WARN_ON_ONCE(ndw > ring->max_dw)) in amdgpu_ring_alloc()
77 ring->count_dw = ndw; in amdgpu_ring_alloc()
Dsdma_v2_4.c698 unsigned ndw = count * 2; in sdma_v2_4_vm_write_pte() local
704 ib->ptr[ib->length_dw++] = ndw; in sdma_v2_4_vm_write_pte()
705 for (; ndw > 0; ndw -= 2) { in sdma_v2_4_vm_write_pte()
Dcik_sdma.c759 unsigned ndw = count * 2; in cik_sdma_vm_write_pte() local
765 ib->ptr[ib->length_dw++] = ndw; in cik_sdma_vm_write_pte()
766 for (; ndw > 0; ndw -= 2) { in cik_sdma_vm_write_pte()
Dsdma_v3_0.c969 unsigned ndw = count * 2; in sdma_v3_0_vm_write_pte() local
975 ib->ptr[ib->length_dw++] = ndw; in sdma_v3_0_vm_write_pte()
976 for (; ndw > 0; ndw -= 2) { in sdma_v3_0_vm_write_pte()
Dsdma_v5_0.c1044 unsigned ndw = count * 2; in sdma_v5_0_vm_write_pte() local
1050 ib->ptr[ib->length_dw++] = ndw - 1; in sdma_v5_0_vm_write_pte()
1051 for (; ndw > 0; ndw -= 2) { in sdma_v5_0_vm_write_pte()
Damdgpu_ring.h256 int amdgpu_ring_alloc(struct amdgpu_ring *ring, unsigned ndw);
Dsdma_v4_0.c1526 unsigned ndw = count * 2; in sdma_v4_0_vm_write_pte() local
1532 ib->ptr[ib->length_dw++] = ndw - 1; in sdma_v4_0_vm_write_pte()
1533 for (; ndw > 0; ndw -= 2) { in sdma_v4_0_vm_write_pte()
/Linux-v5.4/drivers/gpu/drm/radeon/
Dni_dma.c320 unsigned ndw; in cayman_dma_vm_copy_pages() local
323 ndw = count * 2; in cayman_dma_vm_copy_pages()
324 if (ndw > 0xFFFFE) in cayman_dma_vm_copy_pages()
325 ndw = 0xFFFFE; in cayman_dma_vm_copy_pages()
328 0, 0, ndw); in cayman_dma_vm_copy_pages()
334 pe += ndw * 4; in cayman_dma_vm_copy_pages()
335 src += ndw * 4; in cayman_dma_vm_copy_pages()
336 count -= ndw / 2; in cayman_dma_vm_copy_pages()
360 unsigned ndw; in cayman_dma_vm_write_pages() local
363 ndw = count * 2; in cayman_dma_vm_write_pages()
[all …]
Dsi_dma.c112 unsigned ndw; in si_dma_vm_write_pages() local
115 ndw = count * 2; in si_dma_vm_write_pages()
116 if (ndw > 0xFFFFE) in si_dma_vm_write_pages()
117 ndw = 0xFFFFE; in si_dma_vm_write_pages()
120 ib->ptr[ib->length_dw++] = DMA_PACKET(DMA_PACKET_WRITE, 0, 0, 0, ndw); in si_dma_vm_write_pages()
123 for (; ndw > 0; ndw -= 2, --count, pe += 8) { in si_dma_vm_write_pages()
159 unsigned ndw; in si_dma_vm_set_pages() local
162 ndw = count * 2; in si_dma_vm_set_pages()
163 if (ndw > 0xFFFFE) in si_dma_vm_set_pages()
164 ndw = 0xFFFFE; in si_dma_vm_set_pages()
[all …]
Dradeon_ring.c108 int radeon_ring_alloc(struct radeon_device *rdev, struct radeon_ring *ring, unsigned ndw) in radeon_ring_alloc() argument
113 if (ndw > (ring->ring_size / 4)) in radeon_ring_alloc()
118 ndw = (ndw + ring->align_mask) & ~ring->align_mask; in radeon_ring_alloc()
119 while (ndw > (ring->ring_free_dw - 1)) { in radeon_ring_alloc()
121 if (ndw < ring->ring_free_dw) { in radeon_ring_alloc()
128 ring->count_dw = ndw; in radeon_ring_alloc()
144 int radeon_ring_lock(struct radeon_device *rdev, struct radeon_ring *ring, unsigned ndw) in radeon_ring_lock() argument
149 r = radeon_ring_alloc(rdev, ring, ndw); in radeon_ring_lock()
Dcik_sdma.c848 unsigned ndw; in cik_sdma_vm_write_pages() local
851 ndw = count * 2; in cik_sdma_vm_write_pages()
852 if (ndw > 0xFFFFE) in cik_sdma_vm_write_pages()
853 ndw = 0xFFFFE; in cik_sdma_vm_write_pages()
860 ib->ptr[ib->length_dw++] = ndw; in cik_sdma_vm_write_pages()
861 for (; ndw > 0; ndw -= 2, --count, pe += 8) { in cik_sdma_vm_write_pages()
897 unsigned ndw; in cik_sdma_vm_set_pages() local
900 ndw = count; in cik_sdma_vm_set_pages()
901 if (ndw > 0x7FFFF) in cik_sdma_vm_set_pages()
902 ndw = 0x7FFFF; in cik_sdma_vm_set_pages()
[all …]
Dradeon_vm.c648 unsigned count = 0, pt_idx, ndw; in radeon_vm_update_page_directory() local
653 ndw = 64; in radeon_vm_update_page_directory()
656 ndw += vm->max_pde_used * 6; in radeon_vm_update_page_directory()
659 if (ndw > 0xfffff) in radeon_vm_update_page_directory()
662 r = radeon_ib_get(rdev, R600_RING_TYPE_DMA_INDEX, &ib, NULL, ndw * 4); in radeon_vm_update_page_directory()
706 WARN_ON(ib.length_dw > ndw); in radeon_vm_update_page_directory()
918 unsigned nptes, ncmds, ndw; in radeon_vm_bo_update() local
974 ndw = 64; in radeon_vm_bo_update()
979 ndw += ncmds * 7; in radeon_vm_bo_update()
983 ndw += ncmds * 4; in radeon_vm_bo_update()
[all …]
Dr100.c902 unsigned ndw; in r100_copy_blit() local
914 ndw = 64 + (10 * num_loops); in r100_copy_blit()
915 r = radeon_ring_lock(rdev, ring, ndw); in r100_copy_blit()
917 DRM_ERROR("radeon: moving bo (%d) asking for %u dw.\n", r, ndw); in r100_copy_blit()
Dradeon.h1018 int radeon_ring_alloc(struct radeon_device *rdev, struct radeon_ring *cp, unsigned ndw);
1019 int radeon_ring_lock(struct radeon_device *rdev, struct radeon_ring *cp, unsigned ndw);
/Linux-v5.4/kernel/rcu/
Dtree_plugin.h2160 int ndw; in do_nocb_deferred_wakeup_common() local
2167 ndw = READ_ONCE(rdp->nocb_defer_wakeup); in do_nocb_deferred_wakeup_common()
2169 wake_nocb_gp(rdp, ndw == RCU_NOCB_WAKE_FORCE, flags); in do_nocb_deferred_wakeup_common()