Lines Matching refs:event

108 static inline u64 arm_pmu_event_max_period(struct perf_event *event)  in arm_pmu_event_max_period()  argument
110 if (event->hw.flags & ARMPMU_EVT_64BIT) in arm_pmu_event_max_period()
170 armpmu_map_event(struct perf_event *event, in armpmu_map_event() argument
178 u64 config = event->attr.config; in armpmu_map_event()
179 int type = event->attr.type; in armpmu_map_event()
181 if (type == event->pmu->type) in armpmu_map_event()
196 int armpmu_event_set_period(struct perf_event *event) in armpmu_event_set_period() argument
198 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_event_set_period()
199 struct hw_perf_event *hwc = &event->hw; in armpmu_event_set_period()
205 max_period = arm_pmu_event_max_period(event); in armpmu_event_set_period()
231 armpmu->write_counter(event, (u64)(-left) & max_period); in armpmu_event_set_period()
233 perf_event_update_userpage(event); in armpmu_event_set_period()
238 u64 armpmu_event_update(struct perf_event *event) in armpmu_event_update() argument
240 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_event_update()
241 struct hw_perf_event *hwc = &event->hw; in armpmu_event_update()
243 u64 max_period = arm_pmu_event_max_period(event); in armpmu_event_update()
247 new_raw_count = armpmu->read_counter(event); in armpmu_event_update()
255 local64_add(delta, &event->count); in armpmu_event_update()
262 armpmu_read(struct perf_event *event) in armpmu_read() argument
264 armpmu_event_update(event); in armpmu_read()
268 armpmu_stop(struct perf_event *event, int flags) in armpmu_stop() argument
270 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_stop()
271 struct hw_perf_event *hwc = &event->hw; in armpmu_stop()
278 armpmu->disable(event); in armpmu_stop()
279 armpmu_event_update(event); in armpmu_stop()
284 static void armpmu_start(struct perf_event *event, int flags) in armpmu_start() argument
286 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_start()
287 struct hw_perf_event *hwc = &event->hw; in armpmu_start()
304 armpmu_event_set_period(event); in armpmu_start()
305 armpmu->enable(event); in armpmu_start()
309 armpmu_del(struct perf_event *event, int flags) in armpmu_del() argument
311 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_del()
313 struct hw_perf_event *hwc = &event->hw; in armpmu_del()
316 armpmu_stop(event, PERF_EF_UPDATE); in armpmu_del()
318 armpmu->clear_event_idx(hw_events, event); in armpmu_del()
319 perf_event_update_userpage(event); in armpmu_del()
325 armpmu_add(struct perf_event *event, int flags) in armpmu_add() argument
327 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_add()
329 struct hw_perf_event *hwc = &event->hw; in armpmu_add()
337 idx = armpmu->get_event_idx(hw_events, event); in armpmu_add()
345 event->hw.idx = idx; in armpmu_add()
346 armpmu->disable(event); in armpmu_add()
347 hw_events->events[idx] = event; in armpmu_add()
351 armpmu_start(event, PERF_EF_RELOAD); in armpmu_add()
354 perf_event_update_userpage(event); in armpmu_add()
361 struct perf_event *event) in validate_event() argument
365 if (is_software_event(event)) in validate_event()
373 if (event->pmu != pmu) in validate_event()
376 if (event->state < PERF_EVENT_STATE_OFF) in validate_event()
379 if (event->state == PERF_EVENT_STATE_OFF && !event->attr.enable_on_exec) in validate_event()
382 armpmu = to_arm_pmu(event->pmu); in validate_event()
383 return armpmu->get_event_idx(hw_events, event) >= 0; in validate_event()
387 validate_group(struct perf_event *event) in validate_group() argument
389 struct perf_event *sibling, *leader = event->group_leader; in validate_group()
398 if (!validate_event(event->pmu, &fake_pmu, leader)) in validate_group()
402 if (!validate_event(event->pmu, &fake_pmu, sibling)) in validate_group()
406 if (!validate_event(event->pmu, &fake_pmu, event)) in validate_group()
437 __hw_perf_event_init(struct perf_event *event) in __hw_perf_event_init() argument
439 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in __hw_perf_event_init()
440 struct hw_perf_event *hwc = &event->hw; in __hw_perf_event_init()
444 mapping = armpmu->map_event(event); in __hw_perf_event_init()
447 pr_debug("event %x:%llx not supported\n", event->attr.type, in __hw_perf_event_init()
448 event->attr.config); in __hw_perf_event_init()
467 armpmu->set_event_filter(hwc, &event->attr)) { in __hw_perf_event_init()
478 if (!is_sampling_event(event)) { in __hw_perf_event_init()
485 hwc->sample_period = arm_pmu_event_max_period(event) >> 1; in __hw_perf_event_init()
490 if (event->group_leader != event) { in __hw_perf_event_init()
491 if (validate_group(event) != 0) in __hw_perf_event_init()
498 static int armpmu_event_init(struct perf_event *event) in armpmu_event_init() argument
500 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_event_init()
509 if (event->cpu != -1 && in armpmu_event_init()
510 !cpumask_test_cpu(event->cpu, &armpmu->supported_cpus)) in armpmu_event_init()
514 if (has_branch_stack(event)) in armpmu_event_init()
517 if (armpmu->map_event(event) == -ENOENT) in armpmu_event_init()
520 return __hw_perf_event_init(event); in armpmu_event_init()
553 static int armpmu_filter_match(struct perf_event *event) in armpmu_filter_match() argument
555 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_filter_match()
561 return armpmu->filter_match(event); in armpmu_filter_match()
746 struct perf_event *event; in cpu_pm_pmu_setup() local
750 event = hw_events->events[idx]; in cpu_pm_pmu_setup()
751 if (!event) in cpu_pm_pmu_setup()
759 armpmu_stop(event, PERF_EF_UPDATE); in cpu_pm_pmu_setup()
775 RCU_NONIDLE(armpmu_start(event, PERF_EF_RELOAD)); in cpu_pm_pmu_setup()