| /drivers/gpu/drm/nouveau/nvkm/core/ |
| A D | event.c | 34 if (--event->refs[index * event->types_nr + type] == 0) { in nvkm_event_put() 37 event->func->fini(event, 1 << type, index); in nvkm_event_put() 51 if (++event->refs[index * event->types_nr + type] == 1) { in nvkm_event_get() 54 event->func->init(event, 1 << type, index); in nvkm_event_get() 62 struct nvkm_event *event = ntfy->event; in nvkm_event_ntfy_state() local 114 if (ntfy->event) in nvkm_event_ntfy_block() 133 struct nvkm_event *event = ntfy->event; in nvkm_event_ntfy_del() local 135 if (!event) in nvkm_event_ntfy_del() 151 ntfy->event = event; in nvkm_event_ntfy_add() 175 if (!event->refs || WARN_ON(id >= event->index_nr)) in nvkm_event_ntfy() [all …]
|
| /drivers/perf/ |
| A D | riscv_pmu.c | 27 (event->hw.idx != -1); in riscv_perf_user_access() 48 userpg->pmc_width = to_riscv_pmu(event->pmu)->ctr_get_width(event->hw.idx) + 1; in arch_perf_update_userpage() 194 rvpmu->ctr_stop(event, 0); in riscv_pmu_stop() 197 riscv_pmu_event_update(event); in riscv_pmu_stop() 264 idx = rvpmu->ctr_get_idx(event); in riscv_pmu_add() 269 cpuc->events[idx] = event; in riscv_pmu_add() 294 rvpmu->ctr_clear_idx(event); in riscv_pmu_del() 301 riscv_pmu_event_update(event); in riscv_pmu_read() 313 if (has_branch_stack(event)) in riscv_pmu_event_init() 320 event->attr.config); in riscv_pmu_event_init() [all …]
|
| A D | qcom_l3_pmu.c | 196 int idx = event->hw.idx; in qcom_l3_cache__64bit_counter_start() 228 int idx = event->hw.idx; in qcom_l3_cache__64bit_counter_stop() 242 int idx = event->hw.idx; in qcom_l3_cache__64bit_counter_update() 277 int idx = event->hw.idx; in qcom_l3_cache__32bit_counter_start() 406 if (!event) in qcom_l3_cache__handle_irq() 416 ops->update(event); in qcom_l3_cache__handle_irq() 486 if (event->attr.type != event->pmu->type) in qcom_l3_cache__event_init() 499 if (event->cpu < 0) in qcom_l3_cache__event_init() 530 ops->start(event); in qcom_l3_cache__event_start() 543 ops->update(event); in qcom_l3_cache__event_stop() [all …]
|
| A D | arm_pmu.c | 282 armpmu->disable(event); in armpmu_stop() 309 armpmu->enable(event); in armpmu_start() 359 event->hw.idx = idx; in armpmu_add() 386 if (event->pmu != pmu) in validate_event() 392 if (event->state == PERF_EVENT_STATE_OFF && !event->attr.enable_on_exec) in validate_event() 414 if (event == leader) in validate_group() 422 if (!validate_event(event->pmu, &fake_pmu, event)) in validate_group() 464 event->attr.config); in __hw_perf_event_init() 519 if (event->cpu != -1 && in armpmu_event_init() 750 struct perf_event *event; in cpu_pm_pmu_setup() local [all …]
|
| A D | arm_dsu_pmu.c | 235 int idx = event->hw.idx; in dsu_pmu_read_counter() 261 int idx = event->hw.idx; in dsu_pmu_write_counter() 317 int idx = event->hw.idx; in dsu_pmu_set_event() 387 if (!event) in dsu_pmu_handle_irq() 407 event->hw.state = 0; in dsu_pmu_start() 498 if (event->pmu != pmu) in dsu_pmu_validate_event() 512 if (event->group_leader == event) in dsu_pmu_validate_group() 522 return dsu_pmu_validate_event(event->pmu, &fake_hw, event); in dsu_pmu_validate_group() 529 if (event->attr.type != event->pmu->type) in dsu_pmu_event_init() 539 if (event->cpu < 0 || event->attach_state & PERF_ATTACH_TASK) { in dsu_pmu_event_init() [all …]
|
| A D | alibaba_uncore_drw_pmu.c | 111 #define GET_DRW_EVENTID(event) FIELD_GET(DRW_CONFIG_EVENTID, (event)->attr.config) argument 337 int counter = event->hw.idx; in ali_drw_pmu_enable_counter() 356 int counter = event->hw.idx; in ali_drw_pmu_disable_counter() 379 struct perf_event *event; in ali_drw_pmu_isr() local 384 if (!event) in ali_drw_pmu_isr() 410 if (!event) in ali_drw_pmu_isr() 532 if (event->attr.type != event->pmu->type) in ali_drw_pmu_event_init() 545 event->cpu = drw_pmu->cpu; in ali_drw_pmu_event_init() 546 if (event->cpu < 0) { in ali_drw_pmu_event_init() 551 if (event->group_leader != event && in ali_drw_pmu_event_init() [all …]
|
| A D | thunderx2_pmu.c | 137 TX2_PMU_FORMAT_ATTR(event, event, "config:0-4"); 479 GET_COUNTERID(event, cmask), in tx2_uncore_event_update() 525 if (is_software_event(event)) in tx2_uncore_validate_event() 528 if (event->pmu != pmu) in tx2_uncore_validate_event() 545 if (event->group_leader == event) in tx2_uncore_validate_event_group() 556 if (!tx2_uncore_validate_event(event->pmu, event, &counters)) in tx2_uncore_validate_event_group() 573 if (event->attr.type != event->pmu->type) in tx2_uncore_event_init() 581 if (is_sampling_event(event) || event->attach_state & PERF_ATTACH_TASK) in tx2_uncore_event_init() 584 if (event->cpu < 0) in tx2_uncore_event_init() 590 event->cpu = tx2_pmu->cpu; in tx2_uncore_event_init() [all …]
|
| A D | arm_dmc620_pmu.c | 217 GEN_PMU_FORMAT_ATTR(event); 375 struct perf_event *event; in dmc620_pmu_handle_irq() local 385 if (!event) in dmc620_pmu_handle_irq() 397 if (WARN_ON_ONCE(!event)) in dmc620_pmu_handle_irq() 413 if (!event) in dmc620_pmu_handle_irq() 518 if (event->attr.type != event->pmu->type) in dmc620_pmu_event_init() 525 if (is_sampling_event(event) || in dmc620_pmu_event_init() 542 if (event->cpu < 0) in dmc620_pmu_event_init() 547 if (event->group_leader == event) in dmc620_pmu_event_init() 558 if (sibling != event && in dmc620_pmu_event_init() [all …]
|
| A D | starfive_starlink_pmu.c | 159 int idx = event->hw.idx; in starlink_pmu_set_event_period() 179 int idx = event->hw.idx; in starlink_pmu_counter_start() 216 int idx = event->hw.idx; in starlink_pmu_counter_stop() 286 starlink_pmu_update(event); in starlink_pmu_stop() 319 hw_events->events[idx] = event; in starlink_pmu_add() 354 if (event->group_leader->pmu != event->pmu && in starlink_pmu_validate_event_group() 384 if (event->cpu < 0 || event->attach_state & PERF_ATTACH_TASK) in starlink_pmu_event_init() 409 if (!event) in starlink_pmu_handle_irq() 420 starlink_pmu_update(event); in starlink_pmu_handle_irq() 455 struct perf_event *event; in starlink_pmu_pm_notify() local [all …]
|
| A D | qcom_l2_pmu.c | 75 #define L2_EVT_CODE(event) (((event) & L2_EVT_CODE_MASK) >> L2_EVT_CODE_SHIFT) argument 76 #define L2_EVT_GROUP(event) (((event) & L2_EVT_GRP_MASK) >> L2_EVT_GRP_SHIFT) argument 407 hwc = &event->hw; in l2_cache_handle_irq() 445 if (event->attr.type != event->pmu->type) in l2_cache_event_init() 472 if (event->group_leader->pmu != event->pmu && in l2_cache_event_init() 497 if ((event->group_leader != event) && in l2_cache_event_init() 501 event->cpu, event->group_leader->cpu); in l2_cache_event_init() 505 if ((event != event->group_leader) && in l2_cache_event_init() 551 cluster = get_cluster_pmu(to_l2cache_pmu(event->pmu), event->cpu); in l2_cache_event_start() 595 cluster = get_cluster_pmu(to_l2cache_pmu(event->pmu), event->cpu); in l2_cache_event_add() [all …]
|
| A D | arm_pmuv3.c | 322 GEN_PMU_FORMAT_ATTR(event); 506 int idx = event->hw.idx; in armv8pmu_event_is_chained() 547 int idx = event->hw.idx; in armv8pmu_read_hw_counter() 617 int idx = event->hw.idx; in armv8pmu_write_hw_counter() 684 int counter = event->hw.idx; in armv8pmu_event_cnten_mask() 872 struct perf_event *event, in read_branch_records() argument 915 if (!event) in armv8pmu_handle_irq() 925 hwc = &event->hw; in armv8pmu_handle_irq() 990 !armv8pmu_event_get_threshold(&event->attr) && !has_branch_stack(event)) { in armv8pmu_get_event_idx() 1024 int idx = event->hw.idx; in armv8pmu_clear_event_idx() [all …]
|
| A D | fsl_imx8_ddr_perf.c | 323 return event->attr.config == 0x41 || event->attr.config == 0x42; in ddr_perf_is_filtered() 328 return event->attr.config1; in ddr_perf_filter_val() 348 ddr_perf_is_filtered(event); in ddr_perf_is_enhanced_filtered() 401 if (event->attr.type != event->pmu->type) in ddr_perf_event_init() 404 if (is_sampling_event(event) || event->attach_state & PERF_ATTACH_TASK) in ddr_perf_event_init() 407 if (event->cpu < 0) { in ddr_perf_event_init() 417 if (event->group_leader->pmu != event->pmu && in ddr_perf_event_init() 422 if (!ddr_perf_filters_compatible(event, event->group_leader)) in ddr_perf_event_init() 436 event->cpu = pmu->cpu; in ddr_perf_event_init() 679 struct perf_event *event; in ddr_perf_irq_handler() local [all …]
|
| A D | arm-ccn.c | 259 u32 event; member 323 if (event->event) in arm_ccn_pmu_event_show() 324 res += sysfs_emit_at(buf, res, ",event=0x%x", event->event); in arm_ccn_pmu_event_show() 325 if (event->def) in arm_ccn_pmu_event_show() 327 if (event->mask) in arm_ccn_pmu_event_show() 336 if (event->event == CCN_EVENT_WATCHPOINT) in arm_ccn_pmu_event_show() 641 ccn->dt.pmu_counters[CCN_IDX_PMU_CYCLE_COUNTER].event = event; in arm_ccn_pmu_event_alloc() 675 ccn->dt.pmu_counters[hw->idx].event = event; in arm_ccn_pmu_event_alloc() 713 if (event->attr.type != event->pmu->type) in arm_ccn_pmu_event_init() 822 if (event->group_leader->pmu != event->pmu && in arm_ccn_pmu_event_init() [all …]
|
| A D | dwc_pcie_pmu.c | 66 #define DWC_PCIE_EVENT_ID(event) FIELD_GET(DWC_PCIE_CONFIG_EVENTID, (event)->attr.config) argument 67 #define DWC_PCIE_EVENT_TYPE(event) FIELD_GET(DWC_PCIE_CONFIG_TYPE, (event)->attr.config) argument 68 #define DWC_PCIE_EVENT_LANE(event) FIELD_GET(DWC_PCIE_CONFIG_LANE, (event)->attr.config) argument 359 if (event->attr.type != event->pmu->type) in dwc_pcie_pmu_event_init() 363 if (is_sampling_event(event)) in dwc_pcie_pmu_event_init() 367 if (event->cpu < 0 || event->attach_state & PERF_ATTACH_TASK) in dwc_pcie_pmu_event_init() 370 if (event->group_leader != event && in dwc_pcie_pmu_event_init() 388 event->cpu = pcie_pmu->on_cpu; in dwc_pcie_pmu_event_init() 438 if (pcie_pmu->event[type]) in dwc_pcie_pmu_event_add() 441 pcie_pmu->event[type] = event; in dwc_pcie_pmu_event_add() [all …]
|
| /drivers/firmware/ |
| A D | arm_sdei.c | 84 arg.event = event; \ 208 event = kzalloc(sizeof(*event), GFP_KERNEL); in sdei_event_create() 209 if (!event) { in sdei_event_create() 227 event->type = result; in sdei_event_create() 268 return event; in sdei_event_create() 271 kfree(event); in sdei_event_create() 287 kfree(event); in sdei_event_destroy_llocked() 406 if (!event) { in sdei_event_enable() 452 if (!event) { in sdei_event_disable() 496 if (!event) { in sdei_event_unregister() [all …]
|
| /drivers/md/ |
| A D | dm-uevent.c | 49 if (!event) in dm_uevent_alloc() 53 event->md = md; in dm_uevent_alloc() 55 return event; in dm_uevent_alloc() 68 if (!event) { in dm_build_path_uevent() 106 return event; in dm_build_path_uevent() 133 if (dm_copy_name_and_uuid(event->md, event->name, in dm_send_uevents() 134 event->uuid)) { in dm_send_uevents() 140 if (add_uevent_var(&event->ku_env, "DM_NAME=%s", event->name)) { in dm_send_uevents() 146 if (add_uevent_var(&event->ku_env, "DM_UUID=%s", event->uuid)) { in dm_send_uevents() 152 r = kobject_uevent_env(kobj, event->action, event->ku_env.envp); in dm_send_uevents() [all …]
|
| /drivers/perf/hisilicon/ |
| A D | hisi_uncore_pmu.c | 109 if (leader->pmu != event->pmu) in hisi_validate_event_group() 113 if (leader != event) in hisi_validate_event_group() 120 if (sibling->pmu != event->pmu) in hisi_validate_event_group() 155 struct perf_event *event; in hisi_uncore_pmu_isr() local 172 if (!event) in hisi_uncore_pmu_isr() 211 if (event->attr.type != event->pmu->type) in hisi_uncore_pmu_event_init() 219 if (is_sampling_event(event) || event->attach_state & PERF_ATTACH_TASK) in hisi_uncore_pmu_event_init() 226 if (event->cpu < 0) in hisi_uncore_pmu_event_init() 254 event->cpu = hisi_pmu->on_cpu; in hisi_uncore_pmu_event_init() 270 HISI_GET_EVENTID(event)); in hisi_uncore_pmu_enable_event() [all …]
|
| A D | hisi_pcie_pmu.c | 328 struct perf_event *event) in hisi_pcie_pmu_cmp_event() argument 343 if (leader->pmu != event->pmu) in hisi_pcie_pmu_validate_event_group() 346 if (leader != event && !hisi_pcie_pmu_cmp_event(leader, event)) in hisi_pcie_pmu_validate_event_group() 354 if (sibling->pmu != event->pmu) in hisi_pcie_pmu_validate_event_group() 387 if (event->attr.type != event->pmu->type) in hisi_pcie_pmu_event_init() 396 if (is_sampling_event(event) || event->attach_state & PERF_ATTACH_TASK) in hisi_pcie_pmu_event_init() 405 event->cpu = pcie_pmu->on_cpu; in hisi_pcie_pmu_event_init() 413 u32 idx = event->hw.idx; in hisi_pcie_pmu_read_counter() 423 struct perf_event *event) in hisi_pcie_pmu_get_event_idx() argument 638 struct perf_event *event; in hisi_pcie_pmu_irq() local [all …]
|
| A D | hns3_pmu.c | 297 u32 event; member 373 event = eattr->var; in hns3_pmu_event_show() 375 return sysfs_emit(buf, "config=0x%x\n", event->event); in hns3_pmu_event_show() 387 event = eattr->var; in hns3_pmu_filter_mode_show() 894 if (event == pmu_event->event) in hns3_pmu_get_pmu_event() 1066 if (leader != event && !hns3_pmu_cmp_event(leader, event)) in hns3_pmu_validate_event_group() 1214 return hns3_pmu_readq(hns3_pmu, event->hw.event_base, event->hw.idx); in hns3_pmu_read_counter() 1220 u32 idx = event->hw.idx; in hns3_pmu_write_counter() 1241 if (event->attr.type != event->pmu->type) in hns3_pmu_event_init() 1245 if (is_sampling_event(event) || event->attach_state & PERF_ATTACH_TASK) in hns3_pmu_event_init() [all …]
|
| /drivers/gpu/drm/xe/ |
| A D | xe_pmu.c | 155 event->pmu_private = fw_ref; in event_gt_forcewake() 197 if (!event_to_hwe(event)) in event_param_valid() 224 event->pmu_private = NULL; in xe_pmu_event_destroy() 241 if (event->attr.type != event->pmu->type) in xe_pmu_event_init() 248 if (event->cpu < 0) in xe_pmu_event_init() 262 if (!event->parent) { in xe_pmu_event_init() 285 hwe = event_to_hwe(event); in read_engine_events() 346 xe_pmu_event_update(event); in xe_pmu_event_read() 356 local64_set(&event->hw.prev_count, __xe_pmu_event_read(event)); in xe_pmu_enable() 367 xe_pmu_enable(event); in xe_pmu_event_start() [all …]
|
| /drivers/dma/idxd/ |
| A D | perfmon.c | 59 struct perf_event *event; in perfmon_collect_events() local 105 struct perf_event *event) in perfmon_assign_event() argument 173 idxd = event_to_idxd(event); in perfmon_pmu_event_init() 174 event->hw.idx = -1; in perfmon_pmu_event_init() 176 if (event->attr.type != event->pmu->type) in perfmon_pmu_event_init() 183 if (event->cpu < 0) in perfmon_pmu_event_init() 190 event->hw.config = event->attr.config; in perfmon_pmu_event_init() 192 if (event->group_leader != event) in perfmon_pmu_event_init() 233 struct perf_event *event; in perfmon_counter_overflow() local 301 event->hw.idx = hwc->idx; in perfmon_pmu_event_start() [all …]
|
| /drivers/net/ethernet/huawei/hinic/ |
| A D | hinic_devlink.c | 325 fault_level = (event->event.chip.err_level < FAULT_LEVEL_MAX) ? in chip_fault_show() 326 event->event.chip.err_level : FAULT_LEVEL_MAX; in chip_fault_show() 329 (u32)event->event.chip.func_id); in chip_fault_show() 334 event->event.chip.err_csr_addr); in chip_fault_show() 336 event->event.chip.err_csr_value); in chip_fault_show() 351 sizeof(event->event.val)); in fault_report_show() 361 devlink_fmsg_u8_pair_put(fmsg, "epc", event->event.ucode.epc); in fault_report_show() 366 event->event.mem_timeout.err_csr_ctrl); in fault_report_show() 368 event->event.mem_timeout.err_csr_data); in fault_report_show() 370 event->event.mem_timeout.ctrl_tab); in fault_report_show() [all …]
|
| /drivers/iommu/intel/ |
| A D | perfmon.c | 283 if (event->attr.type != event->pmu->type) in iommu_pmu_event_init() 290 if (event->cpu < 0) in iommu_pmu_event_init() 418 event->attr.config1); in iommu_pmu_assign_event() 421 event->attr.config1); in iommu_pmu_assign_event() 424 event->attr.config1); in iommu_pmu_assign_event() 427 event->attr.config1); in iommu_pmu_assign_event() 448 iommu_pmu_start(event, 0); in iommu_pmu_add() 456 int idx = event->hw.idx; in iommu_pmu_del() 467 event->hw.idx = -1; in iommu_pmu_del() 491 struct perf_event *event; in iommu_pmu_counter_overflow() local [all …]
|
| /drivers/accel/ivpu/ |
| A D | ivpu_trace.h | 21 TP_PROTO(const char *event), 22 TP_ARGS(event), 24 TP_fast_assign(__entry->event = event;), 25 TP_printk("%s", __entry->event) 30 TP_ARGS(event, job), 31 TP_STRUCT__entry(__field(const char *, event) 36 TP_fast_assign(__entry->event = event; 41 __entry->event, 49 TP_ARGS(event, msg), 56 TP_fast_assign(__entry->event = event; [all …]
|
| /drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/ |
| A D | device.c | 35 if (event->func) { in r535_gsp_event_dtor() 36 list_del(&event->head); in r535_gsp_event_dtor() 37 event->func = NULL; in r535_gsp_event_dtor() 41 nvkm_gsp_rm_free(&event->object); in r535_gsp_event_dtor() 42 event->device = NULL; in r535_gsp_event_dtor() 56 ctrl->event = event->id; in r535_gsp_device_event_get() 72 &event->object); in r535_gsp_device_event_ctor() 86 event->device = device; in r535_gsp_device_event_ctor() 87 event->id = id; in r535_gsp_device_event_ctor() 91 nvkm_gsp_event_dtor(event); in r535_gsp_device_event_ctor() [all …]
|