| /linux/drivers/gpu/drm/amd/pm/swsmu/smu13/ |
| A D | aldebaran_ppt.c | 292 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in aldebaran_allocate_dpm_context() local 297 if (!smu_dpm->dpm_context) in aldebaran_allocate_dpm_context() 301 smu_dpm->dpm_policies = in aldebaran_allocate_dpm_context() 304 if (!smu_dpm->dpm_policies) in aldebaran_allocate_dpm_context() 540 smu->smu_dpm.dpm_context; in aldebaran_populate_umd_state_clk() 778 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in aldebaran_emit_clk_levels() local 793 dpm_context = smu_dpm->dpm_context; in aldebaran_emit_clk_levels() 935 smu->smu_dpm.dpm_context; in aldebaran_upload_dpm_level() 1268 struct smu_dpm_context *smu_dpm = &(smu->smu_dpm); in aldebaran_set_performance_level() local 1304 struct smu_dpm_context *smu_dpm = &(smu->smu_dpm); in aldebaran_set_soft_freq_limited_range() local [all …]
|
| A D | smu_v13_0_6_ppt.c | 444 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_v13_0_6_allocate_dpm_context() local 447 smu_dpm->dpm_context = in smu_v13_0_6_allocate_dpm_context() 449 if (!smu_dpm->dpm_context) in smu_v13_0_6_allocate_dpm_context() 453 smu_dpm->dpm_policies = in smu_v13_0_6_allocate_dpm_context() 455 if (!smu_dpm->dpm_policies) { in smu_v13_0_6_allocate_dpm_context() 456 kfree(smu_dpm->dpm_context); in smu_v13_0_6_allocate_dpm_context() 784 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_v13_0_6_set_default_dpm_table() local 1121 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_v13_0_6_print_clk_levels() local 1707 struct smu_dpm_context *smu_dpm = &(smu->smu_dpm); in smu_v13_0_6_set_performance_level() local 1763 struct smu_dpm_context *smu_dpm = &(smu->smu_dpm); in smu_v13_0_6_set_soft_freq_limited_range() local [all …]
|
| A D | smu_v13_0.c | 503 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_v13_0_fini_smc_tables() local 530 kfree(smu_dpm->dpm_policies); in smu_v13_0_fini_smc_tables() 531 kfree(smu_dpm->dpm_context); in smu_v13_0_fini_smc_tables() 532 kfree(smu_dpm->golden_dpm_context); in smu_v13_0_fini_smc_tables() 533 kfree(smu_dpm->dpm_current_power_state); in smu_v13_0_fini_smc_tables() 535 smu_dpm->dpm_policies = NULL; in smu_v13_0_fini_smc_tables() 536 smu_dpm->dpm_context = NULL; in smu_v13_0_fini_smc_tables() 537 smu_dpm->golden_dpm_context = NULL; in smu_v13_0_fini_smc_tables() 538 smu_dpm->dpm_context_size = 0; in smu_v13_0_fini_smc_tables() 1688 smu->smu_dpm.dpm_context; in smu_v13_0_set_performance_level() [all …]
|
| A D | smu_v13_0_7_ppt.c | 547 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_v13_0_7_allocate_dpm_context() local 549 smu_dpm->dpm_context = kzalloc(sizeof(struct smu_13_0_dpm_context), in smu_v13_0_7_allocate_dpm_context() 551 if (!smu_dpm->dpm_context) in smu_v13_0_7_allocate_dpm_context() 554 smu_dpm->dpm_context_size = sizeof(struct smu_13_0_dpm_context); in smu_v13_0_7_allocate_dpm_context() 576 struct smu_13_0_dpm_context *dpm_context = smu->smu_dpm.dpm_context; in smu_v13_0_7_set_default_dpm_table() 894 smu->smu_dpm.dpm_context; in smu_v13_0_7_get_dpm_ultimate_freq() 1193 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_v13_0_7_print_clk_levels() local 1194 struct smu_13_0_dpm_context *dpm_context = smu_dpm->dpm_context; in smu_v13_0_7_print_clk_levels() 1903 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_v13_0_7_force_clk_levels() local 1904 struct smu_13_0_dpm_context *dpm_context = smu_dpm->dpm_context; in smu_v13_0_7_force_clk_levels() [all …]
|
| A D | smu_v13_0_0_ppt.c | 539 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_v13_0_0_allocate_dpm_context() local 541 smu_dpm->dpm_context = kzalloc(sizeof(struct smu_13_0_dpm_context), in smu_v13_0_0_allocate_dpm_context() 543 if (!smu_dpm->dpm_context) in smu_v13_0_0_allocate_dpm_context() 546 smu_dpm->dpm_context_size = sizeof(struct smu_13_0_dpm_context); in smu_v13_0_0_allocate_dpm_context() 568 struct smu_13_0_dpm_context *dpm_context = smu->smu_dpm.dpm_context; in smu_v13_0_0_set_default_dpm_table() 905 smu->smu_dpm.dpm_context; in smu_v13_0_0_get_dpm_ultimate_freq() 1204 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_v13_0_0_print_clk_levels() local 1205 struct smu_13_0_dpm_context *dpm_context = smu_dpm->dpm_context; in smu_v13_0_0_print_clk_levels() 1914 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_v13_0_0_force_clk_levels() local 1915 struct smu_13_0_dpm_context *dpm_context = smu_dpm->dpm_context; in smu_v13_0_0_force_clk_levels() [all …]
|
| A D | smu_v13_0_5_ppt.c | 514 struct smu_dpm_context *smu_dpm = &(smu->smu_dpm); in smu_v13_0_5_od_edit_dpm_table() local 518 if (smu_dpm->dpm_level != AMD_DPM_FORCED_LEVEL_MANUAL) in smu_v13_0_5_od_edit_dpm_table()
|
| A D | yellow_carp_ppt.c | 645 struct smu_dpm_context *smu_dpm = &(smu->smu_dpm); in yellow_carp_od_edit_dpm_table() local 649 if (smu_dpm->dpm_level != AMD_DPM_FORCED_LEVEL_MANUAL) in yellow_carp_od_edit_dpm_table()
|
| /linux/drivers/gpu/drm/amd/pm/swsmu/smu14/ |
| A D | smu_v14_0.c | 490 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_v14_0_fini_smc_tables() local 515 kfree(smu_dpm->dpm_context); in smu_v14_0_fini_smc_tables() 516 kfree(smu_dpm->golden_dpm_context); in smu_v14_0_fini_smc_tables() 517 kfree(smu_dpm->dpm_current_power_state); in smu_v14_0_fini_smc_tables() 518 kfree(smu_dpm->dpm_request_power_state); in smu_v14_0_fini_smc_tables() 519 smu_dpm->dpm_context = NULL; in smu_v14_0_fini_smc_tables() 520 smu_dpm->golden_dpm_context = NULL; in smu_v14_0_fini_smc_tables() 521 smu_dpm->dpm_context_size = 0; in smu_v14_0_fini_smc_tables() 522 smu_dpm->dpm_current_power_state = NULL; in smu_v14_0_fini_smc_tables() 1182 smu->smu_dpm.dpm_context; in smu_v14_0_set_performance_level() [all …]
|
| A D | smu_v14_0_2_ppt.c | 470 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_v14_0_2_allocate_dpm_context() local 472 smu_dpm->dpm_context = kzalloc(sizeof(struct smu_14_0_dpm_context), in smu_v14_0_2_allocate_dpm_context() 474 if (!smu_dpm->dpm_context) in smu_v14_0_2_allocate_dpm_context() 477 smu_dpm->dpm_context_size = sizeof(struct smu_14_0_dpm_context); in smu_v14_0_2_allocate_dpm_context() 499 struct smu_14_0_dpm_context *dpm_context = smu->smu_dpm.dpm_context; in smu_v14_0_2_set_default_dpm_table() 827 smu->smu_dpm.dpm_context; in smu_v14_0_2_get_dpm_ultimate_freq() 1073 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_v14_0_2_print_clk_levels() local 1074 struct smu_14_0_dpm_context *dpm_context = smu_dpm->dpm_context; in smu_v14_0_2_print_clk_levels() 1396 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_v14_0_2_force_clk_levels() local 1397 struct smu_14_0_dpm_context *dpm_context = smu_dpm->dpm_context; in smu_v14_0_2_force_clk_levels() [all …]
|
| /linux/drivers/gpu/drm/amd/pm/swsmu/smu11/ |
| A D | arcturus_ppt.c | 307 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in arcturus_allocate_dpm_context() local 312 if (!smu_dpm->dpm_context) in arcturus_allocate_dpm_context() 316 smu_dpm->dpm_policies = in arcturus_allocate_dpm_context() 319 if (!smu_dpm->dpm_policies) in arcturus_allocate_dpm_context() 322 policy = &(smu_dpm->dpm_policies->policies[0]); in arcturus_allocate_dpm_context() 442 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in arcturus_set_default_dpm_table() local 444 smu_dpm->dpm_policies->policy_mask &= in arcturus_set_default_dpm_table() 574 smu->smu_dpm.dpm_context; in arcturus_populate_umd_state_clk() 806 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in arcturus_emit_clk_levels() local 819 dpm_context = smu_dpm->dpm_context; in arcturus_emit_clk_levels() [all …]
|
| A D | smu_v11_0.c | 427 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_v11_0_fini_smc_tables() local 456 kfree(smu_dpm->dpm_context); in smu_v11_0_fini_smc_tables() 457 kfree(smu_dpm->golden_dpm_context); in smu_v11_0_fini_smc_tables() 458 kfree(smu_dpm->dpm_current_power_state); in smu_v11_0_fini_smc_tables() 459 kfree(smu_dpm->dpm_request_power_state); in smu_v11_0_fini_smc_tables() 460 smu_dpm->dpm_context = NULL; in smu_v11_0_fini_smc_tables() 461 smu_dpm->golden_dpm_context = NULL; in smu_v11_0_fini_smc_tables() 462 smu_dpm->dpm_context_size = 0; in smu_v11_0_fini_smc_tables() 463 smu_dpm->dpm_current_power_state = NULL; in smu_v11_0_fini_smc_tables() 464 smu_dpm->dpm_request_power_state = NULL; in smu_v11_0_fini_smc_tables() [all …]
|
| A D | navi10_ppt.c | 939 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in navi10_allocate_dpm_context() local 941 smu_dpm->dpm_context = kzalloc(sizeof(struct smu_11_0_dpm_context), in navi10_allocate_dpm_context() 943 if (!smu_dpm->dpm_context) in navi10_allocate_dpm_context() 946 smu_dpm->dpm_context_size = sizeof(struct smu_11_0_dpm_context); in navi10_allocate_dpm_context() 968 struct smu_11_0_dpm_context *dpm_context = smu->smu_dpm.dpm_context; in navi10_set_default_dpm_table() 1267 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in navi10_emit_clk_levels() local 1268 struct smu_11_0_dpm_context *dpm_context = smu_dpm->dpm_context; in navi10_emit_clk_levels() 1476 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in navi10_print_clk_levels() local 1477 struct smu_11_0_dpm_context *dpm_context = smu_dpm->dpm_context; in navi10_print_clk_levels() 1710 smu->smu_dpm.dpm_context; in navi10_populate_umd_state_clk() [all …]
|
| A D | vangogh_ppt.c | 430 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in vangogh_allocate_dpm_context() local 432 smu_dpm->dpm_context = kzalloc(sizeof(struct smu_11_0_dpm_context), in vangogh_allocate_dpm_context() 434 if (!smu_dpm->dpm_context) in vangogh_allocate_dpm_context() 437 smu_dpm->dpm_context_size = sizeof(struct smu_11_0_dpm_context); in vangogh_allocate_dpm_context() 565 struct smu_dpm_context *smu_dpm_ctx = &(smu->smu_dpm); in vangogh_print_legacy_clk_levels() 667 struct smu_dpm_context *smu_dpm_ctx = &(smu->smu_dpm); in vangogh_print_clk_levels() 2034 struct smu_dpm_context *smu_dpm_ctx = &(smu->smu_dpm); in vangogh_od_edit_dpm_table()
|
| A D | sienna_cichlid_ppt.c | 921 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in sienna_cichlid_allocate_dpm_context() local 923 smu_dpm->dpm_context = kzalloc(sizeof(struct smu_11_0_dpm_context), in sienna_cichlid_allocate_dpm_context() 925 if (!smu_dpm->dpm_context) in sienna_cichlid_allocate_dpm_context() 928 smu_dpm->dpm_context_size = sizeof(struct smu_11_0_dpm_context); in sienna_cichlid_allocate_dpm_context() 956 struct smu_11_0_dpm_context *dpm_context = smu->smu_dpm.dpm_context; in sienna_cichlid_set_default_dpm_table() 1279 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in sienna_cichlid_print_clk_levels() local 1280 struct smu_11_0_dpm_context *dpm_context = smu_dpm->dpm_context; in sienna_cichlid_print_clk_levels() 1490 smu->smu_dpm.dpm_context; in sienna_cichlid_populate_umd_state_clk() 2108 struct smu_11_0_dpm_context *dpm_context = smu->smu_dpm.dpm_context; in sienna_cichlid_update_pcie_parameters()
|
| /linux/drivers/gpu/drm/amd/pm/swsmu/ |
| A D | amdgpu_smu.c | 456 struct smu_dpm_context *smu_dpm_ctx = &(smu->smu_dpm); in smu_restore_dpm_user_profile() 889 smu->smu_dpm.dpm_level, in smu_late_init() 1213 struct smu_dpm_context *dpm_ctxt = &(smu->smu_dpm); in smu_init_xgmi_plpd_mode() 1291 smu->smu_dpm.dpm_level = AMD_DPM_FORCED_LEVEL_AUTO; in smu_sw_init() 2198 struct smu_dpm_context *smu_dpm_ctx = &(smu->smu_dpm); in smu_enable_umd_pstate() 2247 struct smu_dpm_context *smu_dpm_ctx = &(smu->smu_dpm); in smu_adjust_power_state_dynamic() 2329 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in smu_handle_dpm_task() local 2340 struct smu_dpm_context *smu_dpm_ctx = &(smu->smu_dpm); in smu_switch_power_profile() 3559 struct smu_dpm_context *dpm_ctxt = &smu->smu_dpm; in smu_get_pm_policy_info() 3587 struct smu_dpm_context *dpm_ctxt = &smu->smu_dpm; in smu_get_pm_policy() [all …]
|
| /linux/drivers/gpu/drm/amd/pm/swsmu/smu12/ |
| A D | renoir_ppt.c | 388 struct smu_dpm_context *smu_dpm_ctx = &(smu->smu_dpm); in renoir_od_edit_dpm_table() 500 struct smu_dpm_context *smu_dpm_ctx = &(smu->smu_dpm); in renoir_print_clk_levels() 621 struct smu_dpm_context *smu_dpm_ctx = &(smu->smu_dpm); in renoir_get_current_power_state()
|
| /linux/drivers/gpu/drm/amd/pm/swsmu/inc/ |
| A D | amdgpu_smu.h | 527 struct smu_dpm_context smu_dpm; member
|