| /drivers/gpu/drm/nouveau/nvkm/subdev/pmu/ |
| A D | base.c | 32 struct nvkm_pmu *pmu = device->pmu; in nvkm_pmu_fan_controlled() local 37 if (pmu && pmu->func->code.size) in nvkm_pmu_fan_controlled() 50 if (pmu && pmu->func->pgob) in nvkm_pmu_pgob() 51 pmu->func->pgob(pmu, enable); in nvkm_pmu_pgob() 58 return pmu->func->recv(pmu); in nvkm_pmu_recv() 65 if (!pmu || !pmu->func->send) in nvkm_pmu_send() 76 pmu->func->intr(pmu); in nvkm_pmu_intr() 88 pmu->func->fini(pmu); in nvkm_pmu_fini() 101 return pmu->func->init(pmu); in nvkm_pmu_init() 144 ret = nvkm_falcon_ctor(pmu->func->flcn, &pmu->subdev, pmu->subdev.name, in nvkm_pmu_ctor() [all …]
|
| A D | gt215.c | 37 mutex_lock(&pmu->send.mutex); in gt215_pmu_send() 54 pmu->recv.message = message; in gt215_pmu_send() 55 pmu->recv.process = process; in gt215_pmu_send() 65 pmu->send.base)); in gt215_pmu_send() 77 wait_event(pmu->recv.wait, (pmu->recv.process == 0)); in gt215_pmu_send() 105 pmu->recv.base)); in gt215_pmu_recv() 116 if (pmu->recv.process) { in gt215_pmu_recv() 121 pmu->recv.process = 0; in gt215_pmu_recv() 122 wake_up(&pmu->recv.wait); in gt215_pmu_recv() 207 if (pmu->func->enabled(pmu)) { in gt215_pmu_init() [all …]
|
| A D | gk20a.c | 116 struct gk20a_pmu *pmu = in gk20a_pmu_dvfs_work() local 149 gk20a_pmu_dvfs_target(pmu, &state); in gk20a_pmu_dvfs_work() 153 gk20a_pmu_dvfs_reset_dev_status(pmu); in gk20a_pmu_dvfs_work() 158 gk20a_pmu_fini(struct nvkm_pmu *pmu) in gk20a_pmu_fini() argument 160 struct gk20a_pmu *gpmu = gk20a_pmu(pmu); in gk20a_pmu_fini() 163 nvkm_falcon_put(&pmu->falcon, &pmu->subdev); in gk20a_pmu_fini() 167 gk20a_pmu_init(struct nvkm_pmu *pmu) in gk20a_pmu_init() argument 215 struct gk20a_pmu *pmu; in gk20a_pmu_new() local 218 if (!(pmu = kzalloc(sizeof(*pmu), GFP_KERNEL))) in gk20a_pmu_new() 220 *ppmu = &pmu->base; in gk20a_pmu_new() [all …]
|
| A D | gm20b.c | 42 struct nvkm_pmu *pmu = container_of(falcon, typeof(*pmu), falcon); in gm20b_pmu_acr_bootstrap_falcon() local 129 struct nvkm_pmu *pmu = priv; in gm20b_pmu_acr_init_wpr_callback() local 139 complete_all(&pmu->wpr_ready); in gm20b_pmu_acr_init_wpr_callback() 187 if (!pmu->initmsg_received) { in gm20b_pmu_recv() 188 int ret = pmu->func->initmsg(pmu); in gm20b_pmu_recv() 194 pmu->initmsg_received = true; in gm20b_pmu_recv() 205 flush_work(&pmu->recv.work); in gm20b_pmu_fini() 211 nvkm_falcon_put(&pmu->falcon, &pmu->subdev); in gm20b_pmu_fini() 222 ret = nvkm_falcon_get(&pmu->falcon, &pmu->subdev); in gm20b_pmu_init() 226 pmu->initmsg_received = false; in gm20b_pmu_init() [all …]
|
| A D | Kbuild | 2 nvkm-y += nvkm/subdev/pmu/base.o 3 nvkm-y += nvkm/subdev/pmu/memx.o 4 nvkm-y += nvkm/subdev/pmu/gt215.o 5 nvkm-y += nvkm/subdev/pmu/gf100.o 6 nvkm-y += nvkm/subdev/pmu/gf119.o 7 nvkm-y += nvkm/subdev/pmu/gk104.o 8 nvkm-y += nvkm/subdev/pmu/gk110.o 9 nvkm-y += nvkm/subdev/pmu/gk208.o 10 nvkm-y += nvkm/subdev/pmu/gk20a.o 11 nvkm-y += nvkm/subdev/pmu/gm107.o [all …]
|
| /drivers/gpu/drm/i915/ |
| A D | i915_pmu.c | 38 return container_of(pmu, struct drm_i915_private, pmu); in pmu_to_i915() 222 struct i915_pmu *pmu = &i915->pmu; in get_rc6() local 281 struct i915_pmu *pmu = >->i915->pmu; in park_rc6() local 300 struct i915_pmu *pmu = >->i915->pmu; in i915_pmu_gt_parked() local 322 struct i915_pmu *pmu = >->i915->pmu; in i915_pmu_gt_unparked() local 358 struct intel_engine_pmu *pmu = &engine->pmu; in gen3_engine_sample() local 393 struct intel_engine_pmu *pmu = &engine->pmu; in gen2_engine_sample() local 462 struct i915_pmu *pmu = &i915->pmu; in frequency_sample() local 836 pmu->timer_enabled &= pmu_needs_timer(pmu); in i915_pmu_disable() 1149 struct i915_pmu *pmu = &i915->pmu; in i915_pmu_register() local [all …]
|
| /drivers/perf/ |
| A D | fsl_imx8_ddr_perf.c | 113 struct pmu pmu; member 344 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_is_enhanced_filtered() local 397 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_init() local 417 if (event->group_leader->pmu != event->pmu && in ddr_perf_event_init() 431 if (sibling->pmu != event->pmu && in ddr_perf_event_init() 501 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_update() local 535 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_start() local 552 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_add() local 614 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_stop() local 652 .pmu = (struct pmu) { in ddr_perf_init() [all …]
|
| A D | fsl_imx9_ddr_perf.c | 84 struct pmu pmu; member 539 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_update() local 553 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_init() local 573 if (event->group_leader->pmu != event->pmu && in ddr_perf_event_init() 578 if (sibling->pmu != event->pmu && in ddr_perf_event_init() 591 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_start() local 625 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_add() local 662 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_stop() local 674 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_del() local 703 .pmu = (struct pmu) { in ddr_perf_init() [all …]
|
| A D | marvell_cn10k_ddr_pmu.c | 148 struct pmu pmu; member 474 struct cn10k_ddr_pmu *pmu = to_cn10k_ddr_pmu(event->pmu); in cn10k_ddr_perf_event_init() local 491 if (event->group_leader->pmu != event->pmu && in cn10k_ddr_perf_event_init() 592 struct cn10k_ddr_pmu *pmu = to_cn10k_ddr_pmu(event->pmu); in cn10k_ddr_perf_event_update() local 697 static void cn10k_ddr_perf_pmu_enable(struct pmu *pmu) in cn10k_ddr_perf_pmu_enable() argument 706 static void cn10k_ddr_perf_pmu_disable(struct pmu *pmu) in cn10k_ddr_perf_pmu_disable() argument 786 cn10k_ddr_perf_pmu_disable(&pmu->pmu); in ddr_pmu_overflow_hander() 787 cn10k_ddr_perf_pmu_enable(&pmu->pmu); in ddr_pmu_overflow_hander() 927 perf_pmu_migrate_context(&pmu->pmu, cpu, target); in cn10k_ddr_pmu_offline_cpu() 1026 ddr_pmu->pmu = (struct pmu) { in cn10k_ddr_perf_probe() [all …]
|
| A D | arm_pmu_platform.c | 37 ret = info->init(pmu); in probe_current_pmu() 112 pmu->pmu.capabilities |= PERF_PMU_CAP_NO_INTERRUPT; in pmu_parse_irqs() 192 struct arm_pmu *pmu; in arm_pmu_device_probe() local 195 pmu = armpmu_alloc(); in arm_pmu_device_probe() 196 if (!pmu) in arm_pmu_device_probe() 199 pmu->pmu.parent = &pdev->dev; in arm_pmu_device_probe() 200 pmu->plat_device = pdev; in arm_pmu_device_probe() 202 ret = pmu_parse_irqs(pmu); in arm_pmu_device_probe() 217 ret = init_fn(pmu); in arm_pmu_device_probe() 241 armpmu_free_irqs(pmu); in arm_pmu_device_probe() [all …]
|
| A D | arm_pmu.c | 386 if (event->pmu != pmu) in validate_event() 529 static void armpmu_enable(struct pmu *pmu) in armpmu_enable() argument 543 static void armpmu_disable(struct pmu *pmu) in armpmu_disable() argument 559 static bool armpmu_filter(struct pmu *pmu, int cpu) in armpmu_filter() argument 718 pmu->reset(pmu); in arm_perf_starting_cpu() 864 pmu = kzalloc(sizeof(*pmu), GFP_KERNEL); in armpmu_alloc() 865 if (!pmu) in armpmu_alloc() 874 pmu->pmu = (struct pmu) { in armpmu_alloc() 906 return pmu; in armpmu_alloc() 929 pmu->pmu.capabilities |= PERF_PMU_CAP_NO_EXCLUDE; in armpmu_register() [all …]
|
| A D | riscv_pmu_legacy.c | 117 pmu->ctr_stop = NULL; in pmu_legacy_init() 121 pmu->ctr_clear_idx = NULL; in pmu_legacy_init() 122 pmu->ctr_read = pmu_legacy_read_ctr; in pmu_legacy_init() 126 pmu->pmu.capabilities |= PERF_PMU_CAP_NO_INTERRUPT; in pmu_legacy_init() 127 pmu->pmu.capabilities |= PERF_PMU_CAP_NO_EXCLUDE; in pmu_legacy_init() 129 perf_pmu_register(&pmu->pmu, "cpu", PERF_TYPE_RAW); in pmu_legacy_init() 134 struct riscv_pmu *pmu = NULL; in pmu_legacy_device_probe() local 136 pmu = riscv_pmu_alloc(); in pmu_legacy_device_probe() 137 if (!pmu) in pmu_legacy_device_probe() 139 pmu->pmu.parent = &pdev->dev; in pmu_legacy_device_probe() [all …]
|
| A D | arm_pmu_acpi.c | 252 struct arm_pmu *pmu; in arm_pmu_acpi_find_pmu() local 257 if (!pmu || pmu->acpi_cpuid != cpuid) in arm_pmu_acpi_find_pmu() 260 return pmu; in arm_pmu_acpi_find_pmu() 322 struct arm_pmu *pmu; in arm_pmu_acpi_cpu_starting() local 329 if (!pmu) { in arm_pmu_acpi_cpu_starting() 387 if (pmu) in arm_pmu_acpi_probe() 390 pmu = armpmu_alloc(); in arm_pmu_acpi_probe() 391 if (!pmu) { in arm_pmu_acpi_probe() 402 ret = init_fn(pmu); in arm_pmu_acpi_probe() 413 if (!pmu->name) { in arm_pmu_acpi_probe() [all …]
|
| A D | marvell_pem_pmu.c | 81 struct pmu pmu; member 191 struct pem_pmu *pmu = to_pem_pmu(event->pmu); in pem_perf_event_init() local 210 if (event->group_leader->pmu != event->pmu && in pem_perf_event_init() 215 if (sibling->pmu != event->pmu && in pem_perf_event_init() 223 event->cpu = pmu->cpu; in pem_perf_event_init() 236 struct pem_pmu *pmu = to_pem_pmu(event->pmu); in pem_perf_event_update() local 250 struct pem_pmu *pmu = to_pem_pmu(event->pmu); in pem_perf_event_start() local 302 if (cpu != pmu->cpu) in pem_pmu_offline_cpu() 309 perf_pmu_migrate_context(&pmu->pmu, cpu, target); in pem_pmu_offline_cpu() 310 pmu->cpu = target; in pem_pmu_offline_cpu() [all …]
|
| A D | arm_dsu_pmu.c | 114 struct pmu pmu; member 128 static inline struct dsu_pmu *to_dsu_pmu(struct pmu *pmu) in to_dsu_pmu() argument 130 return container_of(pmu, struct dsu_pmu, pmu); in to_dsu_pmu() 146 struct pmu *pmu = dev_get_drvdata(dev); in dsu_pmu_cpumask_show() local 192 struct pmu *pmu = dev_get_drvdata(kobj_to_dev(kobj)); in dsu_pmu_event_attr_is_visible() local 242 dev_err(event->pmu->dev, in dsu_pmu_read_counter() 461 static void dsu_pmu_enable(struct pmu *pmu) in dsu_pmu_enable() argument 478 static void dsu_pmu_disable(struct pmu *pmu) in dsu_pmu_disable() argument 491 static bool dsu_pmu_validate_event(struct pmu *pmu, in dsu_pmu_validate_event() argument 498 if (event->pmu != pmu) in dsu_pmu_validate_event() [all …]
|
| /drivers/soc/dove/ |
| A D | pmu.c | 108 pmu->reset.of_node = pmu->of_node; in pmu_reset_init() 146 struct pmu_data *pmu = pmu_dom->pmu; in pmu_domain_power_off() local 180 struct pmu_data *pmu = pmu_dom->pmu; in pmu_domain_power_on() local 314 pmu = kzalloc(sizeof(*pmu), GFP_KERNEL); in dove_init_pmu_legacy() 315 if (!pmu) in dove_init_pmu_legacy() 329 domain->pmu = pmu; in dove_init_pmu_legacy() 389 pmu = kzalloc(sizeof(*pmu), GFP_KERNEL); in dove_init_pmu() 395 pmu->pmc_base = of_iomap(pmu->of_node, 0); in dove_init_pmu() 396 pmu->pmu_base = of_iomap(pmu->of_node, 1); in dove_init_pmu() 397 if (!pmu->pmc_base || !pmu->pmu_base) { in dove_init_pmu() [all …]
|
| /drivers/pmdomain/starfive/ |
| A D | jh71xx-pmu.c | 87 struct jh71xx_pmu *pmu = pmd->pmu; in jh71xx_pmu_get_state() local 92 *is_on = readl(pmu->base + pmu->match_data->pmu_status) & mask; in jh71xx_pmu_get_state() 99 struct jh71xx_pmu *pmu = pmd->pmu; in jh7110_pmu_set_state() local 167 struct jh71xx_pmu *pmu = pmd->pmu; in jh7110_aon_pmu_set_state() local 187 struct jh71xx_pmu *pmu = pmd->pmu; in jh71xx_pmu_set_state() local 299 pmd->pmu = pmu; in jh71xx_pmu_init_domain() 328 pmu = devm_kzalloc(dev, sizeof(*pmu), GFP_KERNEL); in jh71xx_pmu_probe() 329 if (!pmu) in jh71xx_pmu_probe() 353 if (!pmu->genpd) in jh71xx_pmu_probe() 356 pmu->dev = dev; in jh71xx_pmu_probe() [all …]
|
| /drivers/perf/amlogic/ |
| A D | meson_ddr_pmu_core.c | 21 struct pmu pmu; member 41 pmu->info.hw_info->enable(&pmu->info); in dmc_pmu_enable() 49 pmu->info.hw_info->disable(&pmu->info); in dmc_pmu_disable() 56 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in meson_ddr_set_axi_filter() local 87 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in meson_ddr_perf_event_update() local 94 pmu->info.hw_info->get_counters(&pmu->info, &dc); in meson_ddr_perf_event_update() 120 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in meson_ddr_perf_event_init() local 145 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in meson_ddr_perf_event_start() local 147 memset(&pmu->counters, 0, sizeof(pmu->counters)); in meson_ddr_perf_event_start() 175 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in meson_ddr_perf_event_stop() local [all …]
|
| /drivers/gpu/drm/xe/ |
| A D | xe_pmu.c | 93 struct xe_device *xe = container_of(event->pmu, typeof(*xe), pmu.base); in event_to_gt() 163 struct xe_device *xe = container_of(pmu, typeof(*xe), pmu); in event_supported() 235 struct xe_pmu *pmu = &xe->pmu; in xe_pmu_event_init() local 339 struct xe_pmu *pmu = &xe->pmu; in xe_pmu_event_read() local 362 struct xe_pmu *pmu = &xe->pmu; in xe_pmu_event_start() local 374 struct xe_pmu *pmu = &xe->pmu; in xe_pmu_event_stop() local 386 struct xe_pmu *pmu = &xe->pmu; in xe_pmu_event_add() local 499 struct xe_device *xe = container_of(pmu, typeof(*xe), pmu); in set_supported_events() 521 struct xe_device *xe = container_of(pmu, typeof(*xe), pmu); in xe_pmu_unregister() 540 struct xe_device *xe = container_of(pmu, typeof(*xe), pmu); in xe_pmu_register() [all …]
|
| /drivers/pmdomain/rockchip/ |
| A D | pm-domains.c | 340 struct rockchip_pmu *pmu = pd->pmu; in rockchip_pmu_domain_is_idle() local 359 struct rockchip_pmu *pmu = pd->pmu; in rockchip_pmu_ungate_clk() local 381 struct rockchip_pmu *pmu = pd->pmu; in rockchip_pmu_set_idle_request() local 475 struct rockchip_pmu *pmu = pd->pmu; in rockchip_pmu_domain_is_on() local 496 struct rockchip_pmu *pmu = pd->pmu; in rockchip_pmu_domain_is_mem_on() local 508 struct rockchip_pmu *pmu = pd->pmu; in rockchip_pmu_domain_is_chain_on() local 520 struct rockchip_pmu *pmu = pd->pmu; in rockchip_pmu_domain_mem_reset() local 568 struct rockchip_pmu *pmu = pd->pmu; in rockchip_do_pmu_set_power_domain() local 617 struct rockchip_pmu *pmu = pd->pmu; in rockchip_pd_power() local 669 struct rockchip_pmu *pmu = pd->pmu; in rockchip_pd_regulator_enable() local [all …]
|
| /drivers/gpu/drm/nouveau/nvkm/subdev/devinit/ |
| A D | gm200.c | 88 struct nvbios_pmuR pmu; in pmu_load() local 91 if (!nvbios_pmuRm(bios, type, &pmu)) in pmu_load() 94 if (!post || !subdev->device->pmu) in pmu_load() 101 pmu_code(init, pmu.boot_addr_pmu, pmu.boot_addr, pmu.boot_size, false); in pmu_load() 102 pmu_code(init, pmu.code_addr_pmu, pmu.code_addr, pmu.code_size, true); in pmu_load() 103 pmu_data(init, pmu.data_addr_pmu, pmu.data_addr, pmu.data_size); in pmu_load() 106 *init_addr_pmu = pmu.init_addr_pmu; in pmu_load() 107 *args_addr_pmu = pmu.args_addr_pmu; in pmu_load() 111 return pmu_exec(init, pmu.init_addr_pmu), 0; in pmu_load() 152 pmu_data(init, pmu, img, len); in gm200_devinit_post() [all …]
|
| /drivers/cxl/core/ |
| A D | pmu.c | 16 kfree(pmu); in cxl_pmu_release() 32 struct cxl_pmu *pmu; in devm_cxl_pmu_add() local 36 pmu = kzalloc(sizeof(*pmu), GFP_KERNEL); in devm_cxl_pmu_add() 37 if (!pmu) in devm_cxl_pmu_add() 40 pmu->assoc_id = assoc_id; in devm_cxl_pmu_add() 41 pmu->index = index; in devm_cxl_pmu_add() 42 pmu->type = type; in devm_cxl_pmu_add() 43 pmu->base = regs->pmu; in devm_cxl_pmu_add() 44 dev = &pmu->dev; in devm_cxl_pmu_add() 50 switch (pmu->type) { in devm_cxl_pmu_add() [all …]
|
| /drivers/nvdimm/ |
| A D | nd_perf.c | 121 struct pmu *pmu = dev_get_drvdata(dev); in nvdimm_pmu_cpumask_show() local 124 nd_pmu = container_of(pmu, struct nvdimm_pmu, pmu); in nvdimm_pmu_cpumask_show() 281 if (WARN_ON_ONCE(!(nd_pmu->pmu.event_init && nd_pmu->pmu.add && in register_nvdimm_pmu() 282 nd_pmu->pmu.del && nd_pmu->pmu.read && nd_pmu->pmu.name))) in register_nvdimm_pmu() 287 if (!nd_pmu->pmu.attr_groups) in register_nvdimm_pmu() 305 kfree(nd_pmu->pmu.attr_groups); in register_nvdimm_pmu() 309 rc = perf_pmu_register(&nd_pmu->pmu, nd_pmu->pmu.name, -1); in register_nvdimm_pmu() 312 kfree(nd_pmu->pmu.attr_groups); in register_nvdimm_pmu() 317 nd_pmu->pmu.name); in register_nvdimm_pmu() 325 perf_pmu_unregister(&nd_pmu->pmu); in unregister_nvdimm_pmu() [all …]
|
| /drivers/dma/idxd/ |
| A D | perfmon.h | 22 struct pmu *pmu; in event_to_pmu() local 24 pmu = event->pmu; in event_to_pmu() 25 idxd_pmu = container_of(pmu, struct idxd_pmu, pmu); in event_to_pmu() 33 struct pmu *pmu; in event_to_idxd() local 35 pmu = event->pmu; in event_to_idxd() 36 idxd_pmu = container_of(pmu, struct idxd_pmu, pmu); in event_to_idxd() 41 static inline struct idxd_device *pmu_to_idxd(struct pmu *pmu) in pmu_to_idxd() argument 45 idxd_pmu = container_of(pmu, struct idxd_pmu, pmu); in pmu_to_idxd()
|
| /drivers/perf/hisilicon/ |
| A D | hisi_uncore_pmu.c | 109 if (leader->pmu != event->pmu) in hisi_validate_event_group() 120 if (sibling->pmu != event->pmu) in hisi_validate_event_group() 211 if (event->attr.type != event->pmu->type) in hisi_uncore_pmu_event_init() 236 hisi_pmu = to_hisi_pmu(event->pmu); in hisi_uncore_pmu_event_init() 417 void hisi_uncore_pmu_enable(struct pmu *pmu) in hisi_uncore_pmu_enable() argument 419 struct hisi_pmu *hisi_pmu = to_hisi_pmu(pmu); in hisi_uncore_pmu_enable() 430 void hisi_uncore_pmu_disable(struct pmu *pmu) in hisi_uncore_pmu_disable() argument 432 struct hisi_pmu *hisi_pmu = to_hisi_pmu(pmu); in hisi_uncore_pmu_disable() 605 struct pmu *pmu = &hisi_pmu->pmu; in hisi_pmu_init() local 607 pmu->module = module; in hisi_pmu_init() [all …]
|