Lines Matching refs:rlc

2486 	for (sect = adev->gfx.rlc.cs_data; sect->section != NULL; ++sect) {  in gfx_v7_0_cp_gfx_start()
3238 adev->gfx.rlc.reg_list = spectre_rlc_save_restore_register_list; in gfx_v7_0_rlc_init()
3239 adev->gfx.rlc.reg_list_size = in gfx_v7_0_rlc_init()
3242 adev->gfx.rlc.reg_list = kalindi_rlc_save_restore_register_list; in gfx_v7_0_rlc_init()
3243 adev->gfx.rlc.reg_list_size = in gfx_v7_0_rlc_init()
3247 adev->gfx.rlc.cs_data = ci_cs_data; in gfx_v7_0_rlc_init()
3248 adev->gfx.rlc.cp_table_size = ALIGN(CP_ME_TABLE_SIZE * 5 * 4, 2048); /* CP JT */ in gfx_v7_0_rlc_init()
3249 adev->gfx.rlc.cp_table_size += 64 * 1024; /* GDS */ in gfx_v7_0_rlc_init()
3251 src_ptr = adev->gfx.rlc.reg_list; in gfx_v7_0_rlc_init()
3252 dws = adev->gfx.rlc.reg_list_size; in gfx_v7_0_rlc_init()
3255 cs_data = adev->gfx.rlc.cs_data; in gfx_v7_0_rlc_init()
3271 if (adev->gfx.rlc.cp_table_size) { in gfx_v7_0_rlc_init()
3278 if (adev->gfx.rlc.funcs->update_spm_vmid) in gfx_v7_0_rlc_init()
3279 adev->gfx.rlc.funcs->update_spm_vmid(adev, 0xf); in gfx_v7_0_rlc_init()
3326 static void gfx_v7_0_update_rlc(struct amdgpu_device *adev, u32 rlc) in gfx_v7_0_update_rlc() argument
3331 if (tmp != rlc) in gfx_v7_0_update_rlc()
3332 WREG32(mmRLC_CNTL, rlc); in gfx_v7_0_update_rlc()
3463 adev->gfx.rlc.funcs->stop(adev); in gfx_v7_0_rlc_resume()
3469 adev->gfx.rlc.funcs->reset(adev); in gfx_v7_0_rlc_resume()
3500 adev->gfx.rlc.funcs->start(adev); in gfx_v7_0_rlc_resume()
3833 if (adev->gfx.rlc.cs_data) { in gfx_v7_0_init_gfx_cgpg()
3835 WREG32(mmRLC_GPM_SCRATCH_DATA, upper_32_bits(adev->gfx.rlc.clear_state_gpu_addr)); in gfx_v7_0_init_gfx_cgpg()
3836 WREG32(mmRLC_GPM_SCRATCH_DATA, lower_32_bits(adev->gfx.rlc.clear_state_gpu_addr)); in gfx_v7_0_init_gfx_cgpg()
3837 WREG32(mmRLC_GPM_SCRATCH_DATA, adev->gfx.rlc.clear_state_size); in gfx_v7_0_init_gfx_cgpg()
3843 if (adev->gfx.rlc.reg_list) { in gfx_v7_0_init_gfx_cgpg()
3845 for (i = 0; i < adev->gfx.rlc.reg_list_size; i++) in gfx_v7_0_init_gfx_cgpg()
3846 WREG32(mmRLC_GPM_SCRATCH_DATA, adev->gfx.rlc.reg_list[i]); in gfx_v7_0_init_gfx_cgpg()
3854 WREG32(mmRLC_SAVE_AND_RESTORE_BASE, adev->gfx.rlc.save_restore_gpu_addr >> 8); in gfx_v7_0_init_gfx_cgpg()
3855 WREG32(mmRLC_JUMP_TABLE_RESTORE, adev->gfx.rlc.cp_table_gpu_addr >> 8); in gfx_v7_0_init_gfx_cgpg()
3890 if (adev->gfx.rlc.cs_data == NULL) in gfx_v7_0_get_csb_size()
3898 for (sect = adev->gfx.rlc.cs_data; sect->section != NULL; ++sect) { in gfx_v7_0_get_csb_size()
3923 if (adev->gfx.rlc.cs_data == NULL) in gfx_v7_0_get_csb_buffer()
3935 for (sect = adev->gfx.rlc.cs_data; sect->section != NULL; ++sect) { in gfx_v7_0_get_csb_buffer()
4185 adev->gfx.rlc.funcs = &gfx_v7_0_rlc_funcs; in gfx_v7_0_early_init()
4429 r = adev->gfx.rlc.funcs->init(adev); in gfx_v7_0_sw_init()
4493 amdgpu_bo_free_kernel(&adev->gfx.rlc.clear_state_obj, in gfx_v7_0_sw_fini()
4494 &adev->gfx.rlc.clear_state_gpu_addr, in gfx_v7_0_sw_fini()
4495 (void **)&adev->gfx.rlc.cs_ptr); in gfx_v7_0_sw_fini()
4496 if (adev->gfx.rlc.cp_table_size) { in gfx_v7_0_sw_fini()
4497 amdgpu_bo_free_kernel(&adev->gfx.rlc.cp_table_obj, in gfx_v7_0_sw_fini()
4498 &adev->gfx.rlc.cp_table_gpu_addr, in gfx_v7_0_sw_fini()
4499 (void **)&adev->gfx.rlc.cp_table_ptr); in gfx_v7_0_sw_fini()
4514 adev->gfx.rlc.funcs->get_csb_buffer(adev, adev->gfx.rlc.cs_ptr); in gfx_v7_0_hw_init()
4516 r = adev->gfx.rlc.funcs->resume(adev); in gfx_v7_0_hw_init()
4534 adev->gfx.rlc.funcs->stop(adev); in gfx_v7_0_hw_fini()
4619 adev->gfx.rlc.funcs->stop(adev); in gfx_v7_0_soft_reset()