Home
last modified time | relevance | path

Searched refs:pmu (Results 1 – 25 of 122) sorted by relevance

12345

/drivers/gpu/drm/nouveau/nvkm/subdev/pmu/
A Dbase.c32 struct nvkm_pmu *pmu = device->pmu; in nvkm_pmu_fan_controlled() local
37 if (pmu && pmu->func->code.size) in nvkm_pmu_fan_controlled()
50 if (pmu && pmu->func->pgob) in nvkm_pmu_pgob()
51 pmu->func->pgob(pmu, enable); in nvkm_pmu_pgob()
58 return pmu->func->recv(pmu); in nvkm_pmu_recv()
65 if (!pmu || !pmu->func->send) in nvkm_pmu_send()
76 pmu->func->intr(pmu); in nvkm_pmu_intr()
88 pmu->func->fini(pmu); in nvkm_pmu_fini()
101 return pmu->func->init(pmu); in nvkm_pmu_init()
144 ret = nvkm_falcon_ctor(pmu->func->flcn, &pmu->subdev, pmu->subdev.name, in nvkm_pmu_ctor()
[all …]
A Dgt215.c37 mutex_lock(&pmu->send.mutex); in gt215_pmu_send()
54 pmu->recv.message = message; in gt215_pmu_send()
55 pmu->recv.process = process; in gt215_pmu_send()
65 pmu->send.base)); in gt215_pmu_send()
77 wait_event(pmu->recv.wait, (pmu->recv.process == 0)); in gt215_pmu_send()
105 pmu->recv.base)); in gt215_pmu_recv()
116 if (pmu->recv.process) { in gt215_pmu_recv()
121 pmu->recv.process = 0; in gt215_pmu_recv()
122 wake_up(&pmu->recv.wait); in gt215_pmu_recv()
207 if (pmu->func->enabled(pmu)) { in gt215_pmu_init()
[all …]
A Dgk20a.c116 struct gk20a_pmu *pmu = in gk20a_pmu_dvfs_work() local
149 gk20a_pmu_dvfs_target(pmu, &state); in gk20a_pmu_dvfs_work()
153 gk20a_pmu_dvfs_reset_dev_status(pmu); in gk20a_pmu_dvfs_work()
158 gk20a_pmu_fini(struct nvkm_pmu *pmu) in gk20a_pmu_fini() argument
160 struct gk20a_pmu *gpmu = gk20a_pmu(pmu); in gk20a_pmu_fini()
163 nvkm_falcon_put(&pmu->falcon, &pmu->subdev); in gk20a_pmu_fini()
167 gk20a_pmu_init(struct nvkm_pmu *pmu) in gk20a_pmu_init() argument
215 struct gk20a_pmu *pmu; in gk20a_pmu_new() local
218 if (!(pmu = kzalloc(sizeof(*pmu), GFP_KERNEL))) in gk20a_pmu_new()
220 *ppmu = &pmu->base; in gk20a_pmu_new()
[all …]
A Dgm20b.c42 struct nvkm_pmu *pmu = container_of(falcon, typeof(*pmu), falcon); in gm20b_pmu_acr_bootstrap_falcon() local
129 struct nvkm_pmu *pmu = priv; in gm20b_pmu_acr_init_wpr_callback() local
139 complete_all(&pmu->wpr_ready); in gm20b_pmu_acr_init_wpr_callback()
187 if (!pmu->initmsg_received) { in gm20b_pmu_recv()
188 int ret = pmu->func->initmsg(pmu); in gm20b_pmu_recv()
194 pmu->initmsg_received = true; in gm20b_pmu_recv()
205 flush_work(&pmu->recv.work); in gm20b_pmu_fini()
211 nvkm_falcon_put(&pmu->falcon, &pmu->subdev); in gm20b_pmu_fini()
222 ret = nvkm_falcon_get(&pmu->falcon, &pmu->subdev); in gm20b_pmu_init()
226 pmu->initmsg_received = false; in gm20b_pmu_init()
[all …]
A DKbuild2 nvkm-y += nvkm/subdev/pmu/base.o
3 nvkm-y += nvkm/subdev/pmu/memx.o
4 nvkm-y += nvkm/subdev/pmu/gt215.o
5 nvkm-y += nvkm/subdev/pmu/gf100.o
6 nvkm-y += nvkm/subdev/pmu/gf119.o
7 nvkm-y += nvkm/subdev/pmu/gk104.o
8 nvkm-y += nvkm/subdev/pmu/gk110.o
9 nvkm-y += nvkm/subdev/pmu/gk208.o
10 nvkm-y += nvkm/subdev/pmu/gk20a.o
11 nvkm-y += nvkm/subdev/pmu/gm107.o
[all …]
/drivers/gpu/drm/i915/
A Di915_pmu.c38 return container_of(pmu, struct drm_i915_private, pmu); in pmu_to_i915()
222 struct i915_pmu *pmu = &i915->pmu; in get_rc6() local
281 struct i915_pmu *pmu = &gt->i915->pmu; in park_rc6() local
300 struct i915_pmu *pmu = &gt->i915->pmu; in i915_pmu_gt_parked() local
322 struct i915_pmu *pmu = &gt->i915->pmu; in i915_pmu_gt_unparked() local
358 struct intel_engine_pmu *pmu = &engine->pmu; in gen3_engine_sample() local
393 struct intel_engine_pmu *pmu = &engine->pmu; in gen2_engine_sample() local
462 struct i915_pmu *pmu = &i915->pmu; in frequency_sample() local
836 pmu->timer_enabled &= pmu_needs_timer(pmu); in i915_pmu_disable()
1149 struct i915_pmu *pmu = &i915->pmu; in i915_pmu_register() local
[all …]
/drivers/perf/
A Dfsl_imx8_ddr_perf.c113 struct pmu pmu; member
344 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_is_enhanced_filtered() local
397 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_init() local
417 if (event->group_leader->pmu != event->pmu && in ddr_perf_event_init()
431 if (sibling->pmu != event->pmu && in ddr_perf_event_init()
501 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_update() local
535 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_start() local
552 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_add() local
614 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_stop() local
652 .pmu = (struct pmu) { in ddr_perf_init()
[all …]
A Dfsl_imx9_ddr_perf.c84 struct pmu pmu; member
539 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_update() local
553 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_init() local
573 if (event->group_leader->pmu != event->pmu && in ddr_perf_event_init()
578 if (sibling->pmu != event->pmu && in ddr_perf_event_init()
591 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_start() local
625 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_add() local
662 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_stop() local
674 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_del() local
703 .pmu = (struct pmu) { in ddr_perf_init()
[all …]
A Dmarvell_cn10k_ddr_pmu.c148 struct pmu pmu; member
474 struct cn10k_ddr_pmu *pmu = to_cn10k_ddr_pmu(event->pmu); in cn10k_ddr_perf_event_init() local
491 if (event->group_leader->pmu != event->pmu && in cn10k_ddr_perf_event_init()
592 struct cn10k_ddr_pmu *pmu = to_cn10k_ddr_pmu(event->pmu); in cn10k_ddr_perf_event_update() local
697 static void cn10k_ddr_perf_pmu_enable(struct pmu *pmu) in cn10k_ddr_perf_pmu_enable() argument
706 static void cn10k_ddr_perf_pmu_disable(struct pmu *pmu) in cn10k_ddr_perf_pmu_disable() argument
786 cn10k_ddr_perf_pmu_disable(&pmu->pmu); in ddr_pmu_overflow_hander()
787 cn10k_ddr_perf_pmu_enable(&pmu->pmu); in ddr_pmu_overflow_hander()
927 perf_pmu_migrate_context(&pmu->pmu, cpu, target); in cn10k_ddr_pmu_offline_cpu()
1026 ddr_pmu->pmu = (struct pmu) { in cn10k_ddr_perf_probe()
[all …]
A Darm_pmu_platform.c37 ret = info->init(pmu); in probe_current_pmu()
112 pmu->pmu.capabilities |= PERF_PMU_CAP_NO_INTERRUPT; in pmu_parse_irqs()
192 struct arm_pmu *pmu; in arm_pmu_device_probe() local
195 pmu = armpmu_alloc(); in arm_pmu_device_probe()
196 if (!pmu) in arm_pmu_device_probe()
199 pmu->pmu.parent = &pdev->dev; in arm_pmu_device_probe()
200 pmu->plat_device = pdev; in arm_pmu_device_probe()
202 ret = pmu_parse_irqs(pmu); in arm_pmu_device_probe()
217 ret = init_fn(pmu); in arm_pmu_device_probe()
241 armpmu_free_irqs(pmu); in arm_pmu_device_probe()
[all …]
A Darm_pmu.c386 if (event->pmu != pmu) in validate_event()
529 static void armpmu_enable(struct pmu *pmu) in armpmu_enable() argument
543 static void armpmu_disable(struct pmu *pmu) in armpmu_disable() argument
559 static bool armpmu_filter(struct pmu *pmu, int cpu) in armpmu_filter() argument
718 pmu->reset(pmu); in arm_perf_starting_cpu()
864 pmu = kzalloc(sizeof(*pmu), GFP_KERNEL); in armpmu_alloc()
865 if (!pmu) in armpmu_alloc()
874 pmu->pmu = (struct pmu) { in armpmu_alloc()
906 return pmu; in armpmu_alloc()
929 pmu->pmu.capabilities |= PERF_PMU_CAP_NO_EXCLUDE; in armpmu_register()
[all …]
A Driscv_pmu_legacy.c117 pmu->ctr_stop = NULL; in pmu_legacy_init()
121 pmu->ctr_clear_idx = NULL; in pmu_legacy_init()
122 pmu->ctr_read = pmu_legacy_read_ctr; in pmu_legacy_init()
126 pmu->pmu.capabilities |= PERF_PMU_CAP_NO_INTERRUPT; in pmu_legacy_init()
127 pmu->pmu.capabilities |= PERF_PMU_CAP_NO_EXCLUDE; in pmu_legacy_init()
129 perf_pmu_register(&pmu->pmu, "cpu", PERF_TYPE_RAW); in pmu_legacy_init()
134 struct riscv_pmu *pmu = NULL; in pmu_legacy_device_probe() local
136 pmu = riscv_pmu_alloc(); in pmu_legacy_device_probe()
137 if (!pmu) in pmu_legacy_device_probe()
139 pmu->pmu.parent = &pdev->dev; in pmu_legacy_device_probe()
[all …]
A Darm_pmu_acpi.c252 struct arm_pmu *pmu; in arm_pmu_acpi_find_pmu() local
257 if (!pmu || pmu->acpi_cpuid != cpuid) in arm_pmu_acpi_find_pmu()
260 return pmu; in arm_pmu_acpi_find_pmu()
322 struct arm_pmu *pmu; in arm_pmu_acpi_cpu_starting() local
329 if (!pmu) { in arm_pmu_acpi_cpu_starting()
387 if (pmu) in arm_pmu_acpi_probe()
390 pmu = armpmu_alloc(); in arm_pmu_acpi_probe()
391 if (!pmu) { in arm_pmu_acpi_probe()
402 ret = init_fn(pmu); in arm_pmu_acpi_probe()
413 if (!pmu->name) { in arm_pmu_acpi_probe()
[all …]
A Dmarvell_pem_pmu.c81 struct pmu pmu; member
191 struct pem_pmu *pmu = to_pem_pmu(event->pmu); in pem_perf_event_init() local
210 if (event->group_leader->pmu != event->pmu && in pem_perf_event_init()
215 if (sibling->pmu != event->pmu && in pem_perf_event_init()
223 event->cpu = pmu->cpu; in pem_perf_event_init()
236 struct pem_pmu *pmu = to_pem_pmu(event->pmu); in pem_perf_event_update() local
250 struct pem_pmu *pmu = to_pem_pmu(event->pmu); in pem_perf_event_start() local
302 if (cpu != pmu->cpu) in pem_pmu_offline_cpu()
309 perf_pmu_migrate_context(&pmu->pmu, cpu, target); in pem_pmu_offline_cpu()
310 pmu->cpu = target; in pem_pmu_offline_cpu()
[all …]
A Darm_dsu_pmu.c114 struct pmu pmu; member
128 static inline struct dsu_pmu *to_dsu_pmu(struct pmu *pmu) in to_dsu_pmu() argument
130 return container_of(pmu, struct dsu_pmu, pmu); in to_dsu_pmu()
146 struct pmu *pmu = dev_get_drvdata(dev); in dsu_pmu_cpumask_show() local
192 struct pmu *pmu = dev_get_drvdata(kobj_to_dev(kobj)); in dsu_pmu_event_attr_is_visible() local
242 dev_err(event->pmu->dev, in dsu_pmu_read_counter()
461 static void dsu_pmu_enable(struct pmu *pmu) in dsu_pmu_enable() argument
478 static void dsu_pmu_disable(struct pmu *pmu) in dsu_pmu_disable() argument
491 static bool dsu_pmu_validate_event(struct pmu *pmu, in dsu_pmu_validate_event() argument
498 if (event->pmu != pmu) in dsu_pmu_validate_event()
[all …]
/drivers/soc/dove/
A Dpmu.c108 pmu->reset.of_node = pmu->of_node; in pmu_reset_init()
146 struct pmu_data *pmu = pmu_dom->pmu; in pmu_domain_power_off() local
180 struct pmu_data *pmu = pmu_dom->pmu; in pmu_domain_power_on() local
314 pmu = kzalloc(sizeof(*pmu), GFP_KERNEL); in dove_init_pmu_legacy()
315 if (!pmu) in dove_init_pmu_legacy()
329 domain->pmu = pmu; in dove_init_pmu_legacy()
389 pmu = kzalloc(sizeof(*pmu), GFP_KERNEL); in dove_init_pmu()
395 pmu->pmc_base = of_iomap(pmu->of_node, 0); in dove_init_pmu()
396 pmu->pmu_base = of_iomap(pmu->of_node, 1); in dove_init_pmu()
397 if (!pmu->pmc_base || !pmu->pmu_base) { in dove_init_pmu()
[all …]
/drivers/pmdomain/starfive/
A Djh71xx-pmu.c87 struct jh71xx_pmu *pmu = pmd->pmu; in jh71xx_pmu_get_state() local
92 *is_on = readl(pmu->base + pmu->match_data->pmu_status) & mask; in jh71xx_pmu_get_state()
99 struct jh71xx_pmu *pmu = pmd->pmu; in jh7110_pmu_set_state() local
167 struct jh71xx_pmu *pmu = pmd->pmu; in jh7110_aon_pmu_set_state() local
187 struct jh71xx_pmu *pmu = pmd->pmu; in jh71xx_pmu_set_state() local
299 pmd->pmu = pmu; in jh71xx_pmu_init_domain()
328 pmu = devm_kzalloc(dev, sizeof(*pmu), GFP_KERNEL); in jh71xx_pmu_probe()
329 if (!pmu) in jh71xx_pmu_probe()
353 if (!pmu->genpd) in jh71xx_pmu_probe()
356 pmu->dev = dev; in jh71xx_pmu_probe()
[all …]
/drivers/perf/amlogic/
A Dmeson_ddr_pmu_core.c21 struct pmu pmu; member
41 pmu->info.hw_info->enable(&pmu->info); in dmc_pmu_enable()
49 pmu->info.hw_info->disable(&pmu->info); in dmc_pmu_disable()
56 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in meson_ddr_set_axi_filter() local
87 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in meson_ddr_perf_event_update() local
94 pmu->info.hw_info->get_counters(&pmu->info, &dc); in meson_ddr_perf_event_update()
120 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in meson_ddr_perf_event_init() local
145 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in meson_ddr_perf_event_start() local
147 memset(&pmu->counters, 0, sizeof(pmu->counters)); in meson_ddr_perf_event_start()
175 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in meson_ddr_perf_event_stop() local
[all …]
/drivers/gpu/drm/xe/
A Dxe_pmu.c93 struct xe_device *xe = container_of(event->pmu, typeof(*xe), pmu.base); in event_to_gt()
163 struct xe_device *xe = container_of(pmu, typeof(*xe), pmu); in event_supported()
235 struct xe_pmu *pmu = &xe->pmu; in xe_pmu_event_init() local
339 struct xe_pmu *pmu = &xe->pmu; in xe_pmu_event_read() local
362 struct xe_pmu *pmu = &xe->pmu; in xe_pmu_event_start() local
374 struct xe_pmu *pmu = &xe->pmu; in xe_pmu_event_stop() local
386 struct xe_pmu *pmu = &xe->pmu; in xe_pmu_event_add() local
499 struct xe_device *xe = container_of(pmu, typeof(*xe), pmu); in set_supported_events()
521 struct xe_device *xe = container_of(pmu, typeof(*xe), pmu); in xe_pmu_unregister()
540 struct xe_device *xe = container_of(pmu, typeof(*xe), pmu); in xe_pmu_register()
[all …]
/drivers/pmdomain/rockchip/
A Dpm-domains.c340 struct rockchip_pmu *pmu = pd->pmu; in rockchip_pmu_domain_is_idle() local
359 struct rockchip_pmu *pmu = pd->pmu; in rockchip_pmu_ungate_clk() local
381 struct rockchip_pmu *pmu = pd->pmu; in rockchip_pmu_set_idle_request() local
475 struct rockchip_pmu *pmu = pd->pmu; in rockchip_pmu_domain_is_on() local
496 struct rockchip_pmu *pmu = pd->pmu; in rockchip_pmu_domain_is_mem_on() local
508 struct rockchip_pmu *pmu = pd->pmu; in rockchip_pmu_domain_is_chain_on() local
520 struct rockchip_pmu *pmu = pd->pmu; in rockchip_pmu_domain_mem_reset() local
568 struct rockchip_pmu *pmu = pd->pmu; in rockchip_do_pmu_set_power_domain() local
617 struct rockchip_pmu *pmu = pd->pmu; in rockchip_pd_power() local
669 struct rockchip_pmu *pmu = pd->pmu; in rockchip_pd_regulator_enable() local
[all …]
/drivers/gpu/drm/nouveau/nvkm/subdev/devinit/
A Dgm200.c88 struct nvbios_pmuR pmu; in pmu_load() local
91 if (!nvbios_pmuRm(bios, type, &pmu)) in pmu_load()
94 if (!post || !subdev->device->pmu) in pmu_load()
101 pmu_code(init, pmu.boot_addr_pmu, pmu.boot_addr, pmu.boot_size, false); in pmu_load()
102 pmu_code(init, pmu.code_addr_pmu, pmu.code_addr, pmu.code_size, true); in pmu_load()
103 pmu_data(init, pmu.data_addr_pmu, pmu.data_addr, pmu.data_size); in pmu_load()
106 *init_addr_pmu = pmu.init_addr_pmu; in pmu_load()
107 *args_addr_pmu = pmu.args_addr_pmu; in pmu_load()
111 return pmu_exec(init, pmu.init_addr_pmu), 0; in pmu_load()
152 pmu_data(init, pmu, img, len); in gm200_devinit_post()
[all …]
/drivers/cxl/core/
A Dpmu.c16 kfree(pmu); in cxl_pmu_release()
32 struct cxl_pmu *pmu; in devm_cxl_pmu_add() local
36 pmu = kzalloc(sizeof(*pmu), GFP_KERNEL); in devm_cxl_pmu_add()
37 if (!pmu) in devm_cxl_pmu_add()
40 pmu->assoc_id = assoc_id; in devm_cxl_pmu_add()
41 pmu->index = index; in devm_cxl_pmu_add()
42 pmu->type = type; in devm_cxl_pmu_add()
43 pmu->base = regs->pmu; in devm_cxl_pmu_add()
44 dev = &pmu->dev; in devm_cxl_pmu_add()
50 switch (pmu->type) { in devm_cxl_pmu_add()
[all …]
/drivers/nvdimm/
A Dnd_perf.c121 struct pmu *pmu = dev_get_drvdata(dev); in nvdimm_pmu_cpumask_show() local
124 nd_pmu = container_of(pmu, struct nvdimm_pmu, pmu); in nvdimm_pmu_cpumask_show()
281 if (WARN_ON_ONCE(!(nd_pmu->pmu.event_init && nd_pmu->pmu.add && in register_nvdimm_pmu()
282 nd_pmu->pmu.del && nd_pmu->pmu.read && nd_pmu->pmu.name))) in register_nvdimm_pmu()
287 if (!nd_pmu->pmu.attr_groups) in register_nvdimm_pmu()
305 kfree(nd_pmu->pmu.attr_groups); in register_nvdimm_pmu()
309 rc = perf_pmu_register(&nd_pmu->pmu, nd_pmu->pmu.name, -1); in register_nvdimm_pmu()
312 kfree(nd_pmu->pmu.attr_groups); in register_nvdimm_pmu()
317 nd_pmu->pmu.name); in register_nvdimm_pmu()
325 perf_pmu_unregister(&nd_pmu->pmu); in unregister_nvdimm_pmu()
[all …]
/drivers/dma/idxd/
A Dperfmon.h22 struct pmu *pmu; in event_to_pmu() local
24 pmu = event->pmu; in event_to_pmu()
25 idxd_pmu = container_of(pmu, struct idxd_pmu, pmu); in event_to_pmu()
33 struct pmu *pmu; in event_to_idxd() local
35 pmu = event->pmu; in event_to_idxd()
36 idxd_pmu = container_of(pmu, struct idxd_pmu, pmu); in event_to_idxd()
41 static inline struct idxd_device *pmu_to_idxd(struct pmu *pmu) in pmu_to_idxd() argument
45 idxd_pmu = container_of(pmu, struct idxd_pmu, pmu); in pmu_to_idxd()
/drivers/perf/hisilicon/
A Dhisi_uncore_pmu.c109 if (leader->pmu != event->pmu) in hisi_validate_event_group()
120 if (sibling->pmu != event->pmu) in hisi_validate_event_group()
211 if (event->attr.type != event->pmu->type) in hisi_uncore_pmu_event_init()
236 hisi_pmu = to_hisi_pmu(event->pmu); in hisi_uncore_pmu_event_init()
417 void hisi_uncore_pmu_enable(struct pmu *pmu) in hisi_uncore_pmu_enable() argument
419 struct hisi_pmu *hisi_pmu = to_hisi_pmu(pmu); in hisi_uncore_pmu_enable()
430 void hisi_uncore_pmu_disable(struct pmu *pmu) in hisi_uncore_pmu_disable() argument
432 struct hisi_pmu *hisi_pmu = to_hisi_pmu(pmu); in hisi_uncore_pmu_disable()
605 struct pmu *pmu = &hisi_pmu->pmu; in hisi_pmu_init() local
607 pmu->module = module; in hisi_pmu_init()
[all …]

Completed in 76 milliseconds

12345