Lines Matching refs:sdma

49 	u32 me = (ring == &adev->sdma.instance[0].ring) ? 0 : 1;  in si_dma_ring_get_wptr()
57 u32 me = (ring == &adev->sdma.instance[0].ring) ? 0 : 1; in si_dma_ring_set_wptr()
118 for (i = 0; i < adev->sdma.num_instances; i++) { in si_dma_stop()
119 ring = &adev->sdma.instance[i].ring; in si_dma_stop()
138 for (i = 0; i < adev->sdma.num_instances; i++) { in si_dma_start()
139 ring = &adev->sdma.instance[i].ring; in si_dma_start()
470 adev->sdma.num_instances = 2; in si_dma_early_init()
488 &adev->sdma.trap_irq); in si_dma_sw_init()
494 &adev->sdma.trap_irq); in si_dma_sw_init()
498 for (i = 0; i < adev->sdma.num_instances; i++) { in si_dma_sw_init()
499 ring = &adev->sdma.instance[i].ring; in si_dma_sw_init()
504 &adev->sdma.trap_irq, in si_dma_sw_init()
520 for (i = 0; i < adev->sdma.num_instances; i++) in si_dma_sw_fini()
521 amdgpu_ring_fini(&adev->sdma.instance[i].ring); in si_dma_sw_fini()
637 amdgpu_fence_process(&adev->sdma.instance[0].ring); in si_dma_process_trap_irq()
639 amdgpu_fence_process(&adev->sdma.instance[1].ring); in si_dma_process_trap_irq()
654 for (i = 0; i < adev->sdma.num_instances; i++) { in si_dma_set_clockgating_state()
666 for (i = 0; i < adev->sdma.num_instances; i++) { in si_dma_set_clockgating_state()
748 for (i = 0; i < adev->sdma.num_instances; i++) in si_dma_set_ring_funcs()
749 adev->sdma.instance[i].ring.funcs = &si_dma_ring_funcs; in si_dma_set_ring_funcs()
759 adev->sdma.trap_irq.num_types = AMDGPU_SDMA_IRQ_LAST; in si_dma_set_irq_funcs()
760 adev->sdma.trap_irq.funcs = &si_dma_trap_irq_funcs; in si_dma_set_irq_funcs()
824 adev->mman.buffer_funcs_ring = &adev->sdma.instance[0].ring; in si_dma_set_buffer_funcs()
841 for (i = 0; i < adev->sdma.num_instances; i++) { in si_dma_set_vm_pte_funcs()
842 sched = &adev->sdma.instance[i].ring.sched; in si_dma_set_vm_pte_funcs()
846 adev->vm_manager.vm_pte_num_rqs = adev->sdma.num_instances; in si_dma_set_vm_pte_funcs()