Lines Matching refs:sdma
49 u32 me = (ring == &adev->sdma.instance[0].ring) ? 0 : 1; in si_dma_ring_get_wptr()
57 u32 me = (ring == &adev->sdma.instance[0].ring) ? 0 : 1; in si_dma_ring_set_wptr()
120 for (i = 0; i < adev->sdma.num_instances; i++) { in si_dma_stop()
135 for (i = 0; i < adev->sdma.num_instances; i++) { in si_dma_start()
136 ring = &adev->sdma.instance[i].ring; in si_dma_start()
471 adev->sdma.num_instances = 2; in si_dma_early_init()
489 &adev->sdma.trap_irq); in si_dma_sw_init()
495 &adev->sdma.trap_irq); in si_dma_sw_init()
499 for (i = 0; i < adev->sdma.num_instances; i++) { in si_dma_sw_init()
500 ring = &adev->sdma.instance[i].ring; in si_dma_sw_init()
505 &adev->sdma.trap_irq, in si_dma_sw_init()
521 for (i = 0; i < adev->sdma.num_instances; i++) in si_dma_sw_fini()
522 amdgpu_ring_fini(&adev->sdma.instance[i].ring); in si_dma_sw_fini()
638 amdgpu_fence_process(&adev->sdma.instance[0].ring); in si_dma_process_trap_irq()
640 amdgpu_fence_process(&adev->sdma.instance[1].ring); in si_dma_process_trap_irq()
655 for (i = 0; i < adev->sdma.num_instances; i++) { in si_dma_set_clockgating_state()
667 for (i = 0; i < adev->sdma.num_instances; i++) { in si_dma_set_clockgating_state()
749 for (i = 0; i < adev->sdma.num_instances; i++) in si_dma_set_ring_funcs()
750 adev->sdma.instance[i].ring.funcs = &si_dma_ring_funcs; in si_dma_set_ring_funcs()
760 adev->sdma.trap_irq.num_types = AMDGPU_SDMA_IRQ_LAST; in si_dma_set_irq_funcs()
761 adev->sdma.trap_irq.funcs = &si_dma_trap_irq_funcs; in si_dma_set_irq_funcs()
827 adev->mman.buffer_funcs_ring = &adev->sdma.instance[0].ring; in si_dma_set_buffer_funcs()
843 for (i = 0; i < adev->sdma.num_instances; i++) { in si_dma_set_vm_pte_funcs()
845 &adev->sdma.instance[i].ring.sched; in si_dma_set_vm_pte_funcs()
847 adev->vm_manager.vm_pte_num_scheds = adev->sdma.num_instances; in si_dma_set_vm_pte_funcs()