| /drivers/net/ethernet/mellanox/mlx5/core/diag/ |
| A D | fw_tracer_tracepoint.h | 45 bool lost, u8 event_id, const char *msg), 47 TP_ARGS(tracer, trace_timestamp, lost, event_id, msg), 53 __field(u8, event_id) 61 __entry->event_id = event_id; 68 __entry->lost, __entry->event_id,
|
| A D | fw_tracer.h | 62 u8 event_id; member 122 u8 event_id; member 164 u8 event_id; member 175 u8 event_id[0x8]; member 182 u8 event_id[0x8]; member 190 u8 event_id[0x8]; member
|
| A D | fw_tracer.c | 442 u8 event_id, u32 tmsn) in mlx5_tracer_message_find() argument 447 if (message->event_id == event_id && message->tmsn == tmsn) in mlx5_tracer_message_find() 467 tracer_event->event_id = MLX5_GET(tracer_event, trace, event_id); in poll_trace() 471 switch (tracer_event->event_id) { in poll_trace() 515 u8 event_id; in get_block_timestamp() local 517 event_id = MLX5_GET(tracer_event, ts_event, event_id); in get_block_timestamp() 550 u8 event_id, char *msg) in mlx5_fw_tracer_save_trace() argument 558 trace_data->event_id = event_id; in mlx5_fw_tracer_save_trace() 583 str_frmt->event_id, tmp); in mlx5_tracer_print_trace() 601 cur_string->event_id = tracer_event->event_id; in mlx5_tracer_handle_raw_string() [all …]
|
| /drivers/clocksource/ |
| A D | samsung_pwm_timer.c | 70 unsigned int event_id; member 200 samsung_time_setup(pwm.event_id, cycles); in samsung_set_next_event() 201 samsung_time_start(pwm.event_id, false); in samsung_set_next_event() 208 samsung_time_stop(pwm.event_id); in samsung_shutdown() 214 samsung_time_stop(pwm.event_id); in samsung_set_periodic() 216 samsung_time_start(pwm.event_id, true); in samsung_set_periodic() 226 u32 mask = (1 << pwm.event_id); in samsung_clockevent_resume() 250 u32 mask = (1 << pwm.event_id); in samsung_clock_event_isr() 278 irq_number = pwm.irq[pwm.event_id]; in samsung_clockevent_init() 285 u32 mask = (1 << pwm.event_id); in samsung_clockevent_init() [all …]
|
| /drivers/gpu/drm/amd/amdkfd/ |
| A D | kfd_events.c | 125 ev->event_id = id; in allocate_event_notification_slot() 212 p->signal_event_count, ev->event_id, in create_signal_event() 237 ev->event_id = id; in create_other_event() 424 *event_slot_index = ev->event_id; in kfd_event_create() 433 *event_id = ev->event_id; in kfd_event_create() 434 *event_trigger_data = ev->event_id; in kfd_event_create() 557 ev_priv->event_id = ev->event_id; in kfd_criu_checkpoint_events() 573 ev_priv->event_id, in kfd_criu_checkpoint_events() 613 ev = lookup_event_by_id(p, event_id); in kfd_event_destroy() 654 ev = lookup_event_by_id(p, event_id); in kfd_set_event() [all …]
|
| A D | kfd_int_process_v9.c | 151 u64 event_id; in event_interrupt_poison_consumption_v9() local 231 event_id = amdgpu_ras_acquire_event_id(dev->adev, type); in event_interrupt_poison_consumption_v9() 233 RAS_EVENT_LOG(dev->adev, event_id, in event_interrupt_poison_consumption_v9()
|
| A D | kfd_events.h | 55 u32 event_id; member
|
| /drivers/net/wwan/t7xx/ |
| A D | t7xx_state_monitor.c | 147 dev_warn(dev, "Unhandled event %d\n", event->event_id); in fsm_flush_event_cmd_qs() 170 event_received = event->event_id == event_expected; in fsm_wait_for_event() 171 if (event_received || event->event_id == event_ignore) { in fsm_wait_for_event() 523 int t7xx_fsm_append_event(struct t7xx_fsm_ctl *ctl, enum t7xx_fsm_event_state event_id, in t7xx_fsm_append_event() argument 530 if (event_id <= FSM_EVENT_INVALID || event_id >= FSM_EVENT_MAX) { in t7xx_fsm_append_event() 531 dev_err(dev, "Invalid event %d\n", event_id); in t7xx_fsm_append_event() 541 event->event_id = event_id; in t7xx_fsm_append_event() 555 void t7xx_fsm_clr_event(struct t7xx_fsm_ctl *ctl, enum t7xx_fsm_event_state event_id) in t7xx_fsm_clr_event() argument 562 if (event->event_id == event_id) in t7xx_fsm_clr_event()
|
| A D | t7xx_state_monitor.h | 104 enum t7xx_fsm_event_state event_id; member 126 int t7xx_fsm_append_event(struct t7xx_fsm_ctl *ctl, enum t7xx_fsm_event_state event_id, 128 void t7xx_fsm_clr_event(struct t7xx_fsm_ctl *ctl, enum t7xx_fsm_event_state event_id);
|
| A D | t7xx_modem_ops.c | 184 static void t7xx_host_event_notify(struct t7xx_pci_dev *t7xx_dev, unsigned int event_id) in t7xx_host_event_notify() argument 190 value |= FIELD_PREP(HOST_EVENT_MASK, event_id); in t7xx_host_event_notify() 350 static int t7xx_wait_hif_ex_hk_event(struct t7xx_modem *md, int event_id) in t7xx_wait_hif_ex_hk_event() argument 355 if (md->exp_id & event_id) in t7xx_wait_hif_ex_hk_event() 503 enum t7xx_fsm_event_state event_id, in t7xx_core_hk_handler() argument 518 if (event->event_id == err_detect) { in t7xx_core_hk_handler() 523 } else if (event->event_id == event_id) { in t7xx_core_hk_handler()
|
| /drivers/accel/amdxdna/ |
| A D | aie2_error.c | 71 __u8 event_id; member 82 u8 event_id; member 147 aie_get_error_category(u8 row, u8 event_id, enum aie_module_type mod_type) in aie_get_error_category() argument 176 if (event_id != lut[i].event_id) in aie_get_error_category() 196 cat = aie_get_error_category(err->row, err->event_id, err->mod_type); in aie2_error_backtrack() 199 err->event_id, cat); in aie2_error_backtrack()
|
| /drivers/irqchip/ |
| A D | irq-gic-v5-its.c | 121 u16 event_id) in gicv5_its_itt_cache_inv() argument 127 eidr = FIELD_PREP(GICV5_ITS_EIDR_EVENTID, event_id); in gicv5_its_itt_cache_inv() 342 u16 event_id) in gicv5_its_device_get_itte_ref() argument 350 return &itt[event_id]; in gicv5_its_device_get_itte_ref() 354 l1_idx = event_id >> l2_bits; in gicv5_its_device_get_itte_ref() 355 l2_idx = event_id & GENMASK(l2_bits - 1, 0); in gicv5_its_device_get_itte_ref() 852 itte = gicv5_its_device_get_itte_ref(its_dev, event_id); in gicv5_its_map_event() 1021 u16 event_id; in gicv5_its_irq_domain_activate() local 1027 return gicv5_its_map_event(its_dev, event_id, lpi); in gicv5_its_irq_domain_activate() 1034 u16 event_id; in gicv5_its_irq_domain_deactivate() local [all …]
|
| A D | irq-gic-v3-its.c | 438 u32 event_id; member 443 u32 event_id; member 448 u32 event_id; member 464 u32 event_id; member 470 u32 event_id; member 1281 desc.its_int_cmd.event_id = event_id; in BUILD_SINGLE_CMD_FUNC() 1291 desc.its_clear_cmd.event_id = event_id; in its_send_clear() 1301 desc.its_inv_cmd.event_id = event_id; in its_send_inv() 1465 desc.its_inv_cmd.event_id = event_id; in its_send_vinv() 1479 desc.its_int_cmd.event_id = event_id; in its_send_vint() [all …]
|
| /drivers/gpu/drm/amd/amdgpu/ |
| A D | amdgpu_mca.c | 277 u64 event_id = qctx ? qctx->evid.event_id : RAS_EVENT_INVALID_ID; in amdgpu_mca_smu_mca_bank_dump() local 279 RAS_EVENT_LOG(adev, event_id, HW_ERR "Accelerator Check Architecture events logged\n"); in amdgpu_mca_smu_mca_bank_dump() 280 RAS_EVENT_LOG(adev, event_id, HW_ERR "aca entry[%02d].STATUS=0x%016llx\n", in amdgpu_mca_smu_mca_bank_dump() 282 RAS_EVENT_LOG(adev, event_id, HW_ERR "aca entry[%02d].ADDR=0x%016llx\n", in amdgpu_mca_smu_mca_bank_dump() 284 RAS_EVENT_LOG(adev, event_id, HW_ERR "aca entry[%02d].MISC0=0x%016llx\n", in amdgpu_mca_smu_mca_bank_dump() 286 RAS_EVENT_LOG(adev, event_id, HW_ERR "aca entry[%02d].IPID=0x%016llx\n", in amdgpu_mca_smu_mca_bank_dump() 288 RAS_EVENT_LOG(adev, event_id, HW_ERR "aca entry[%02d].SYND=0x%016llx\n", in amdgpu_mca_smu_mca_bank_dump() 557 qctx.evid.event_id = RAS_EVENT_INVALID_ID; in mca_dump_show()
|
| A D | amdgpu_ras.c | 1062 u64 event_id = qctx->evid.event_id; in amdgpu_ras_error_print_error_data() local 1150 u64 event_id = qctx->evid.event_id; in amdgpu_ras_error_generate_report() local 1230 u64 event_id = qctx->evid.event_id; in amdgpu_ras_virt_error_generate_report() local 2199 u64 event_id; in amdgpu_ras_interrupt_poison_consumption_handler() local 2234 event_id = amdgpu_ras_acquire_event_id(adev, type); in amdgpu_ras_interrupt_poison_consumption_handler() 2235 RAS_EVENT_LOG(adev, event_id, in amdgpu_ras_interrupt_poison_consumption_handler() 2250 u64 event_id; in amdgpu_ras_interrupt_poison_creation_handler() local 2257 event_id = amdgpu_ras_acquire_event_id(adev, type); in amdgpu_ras_interrupt_poison_creation_handler() 4542 u64 event_id; in amdgpu_ras_global_ras_isr() local 4550 event_id = amdgpu_ras_acquire_event_id(adev, type); in amdgpu_ras_global_ras_isr() [all …]
|
| /drivers/acpi/acpica/ |
| A D | utdecode.c | 123 const char *acpi_ut_get_event_name(u32 event_id) in acpi_ut_get_event_name() argument 126 if (event_id > ACPI_EVENT_MAX) { in acpi_ut_get_event_name() 130 return (acpi_gbl_event_types[event_id]); in acpi_ut_get_event_name()
|
| /drivers/char/tpm/eventlog/ |
| A D | tpm1.c | 185 switch (do_endian_conversion(pc_event->event_id)) { in get_event_name() 194 (pc_event->event_id)]; in get_event_name() 204 (pc_event->event_id)]; in get_event_name()
|
| /drivers/perf/ |
| A D | dwc_pcie_pmu.c | 291 int event_id = DWC_PCIE_EVENT_ID(event); in dwc_pcie_pmu_read_time_based_counter() local 322 if (event_id >= 0x20 && event_id <= 0x23) in dwc_pcie_pmu_read_time_based_counter() 432 int event_id = DWC_PCIE_EVENT_ID(event); in dwc_pcie_pmu_event_add() local 446 ctrl = FIELD_PREP(DWC_PCIE_CNT_EVENT_SEL, event_id) | in dwc_pcie_pmu_event_add() 458 ctrl = FIELD_PREP(DWC_PCIE_TIME_BASED_REPORT_SEL, event_id) | in dwc_pcie_pmu_event_add()
|
| A D | fsl_imx9_ddr_perf.c | 630 int event_id, counter; in ddr_perf_event_add() local 632 event_id = FIELD_GET(CONFIG_EVENT_MASK, cfg); in ddr_perf_event_add() 635 counter = ddr_perf_alloc_counter(pmu, event_id, counter); in ddr_perf_event_add() 648 imx93_ddr_perf_monitor_config(pmu, event_id, counter, cfg1, cfg2); in ddr_perf_event_add() 652 imx95_ddr_perf_monitor_config(pmu, event_id, counter, cfg1, cfg2); in ddr_perf_event_add()
|
| /drivers/platform/x86/intel/ |
| A D | ishtp_eclite.c | 50 unsigned int event_id; member 135 header.event = opr_dev->opr_context.cmd_area.event_id; in ecl_ish_cl_read() 187 message.header.event = opr_dev->opr_context.cmd_area.event_id; in ecl_ish_cl_write() 240 case cmd_opr_offsetof(event_id): in ecl_opregion_cmd_handler() 241 cmd->event_id = (u32)*value64; in ecl_opregion_cmd_handler()
|
| /drivers/net/wireless/quantenna/qtnfmac/ |
| A D | event.c | 682 u16 event_id; in qtnf_event_parse() local 687 event_id = le16_to_cpu(event->event_id); in qtnf_event_parse() 698 switch (event_id) { in qtnf_event_parse() 749 pr_warn("unknown event type: %x\n", event_id); in qtnf_event_parse() 773 le16_to_cpu(event->event_id), le16_to_cpu(event->mhdr.len), in qtnf_event_process_skb()
|
| /drivers/media/test-drivers/vidtv/ |
| A D | vidtv_channel.c | 167 u16 event_id; in vidtv_channel_eit_event_cat_into_new() local 181 event_id = be16_to_cpu(curr->event_id); in vidtv_channel_eit_event_cat_into_new() 182 tail = vidtv_psi_eit_event_init(tail, event_id); in vidtv_channel_eit_event_cat_into_new()
|
| /drivers/net/ethernet/broadcom/bnxt/ |
| A D | bnxt_ulp.c | 346 u16 event_id = le16_to_cpu(cmpl->event_id); in bnxt_ulp_async_events() local 360 if (!ulp->async_events_bmap || event_id > ulp->max_async_event_id) in bnxt_ulp_async_events() 366 if (test_bit(event_id, ulp->async_events_bmap)) in bnxt_ulp_async_events()
|
| /drivers/pci/switch/ |
| A D | switchtec.c | 994 int event_id, int index) in event_hdr_addr() argument 998 if (event_id < 0 || event_id >= SWITCHTEC_IOCTL_MAX_EVENTS) in event_hdr_addr() 1001 off = event_regs[event_id].offset; in event_hdr_addr() 1003 if (event_regs[event_id].map_reg == part_ev_reg) { in event_hdr_addr() 1008 } else if (event_regs[event_id].map_reg == pff_ev_reg) { in event_hdr_addr() 1013 return event_regs[event_id].map_reg(stdev, off, index); in event_hdr_addr() 1023 reg = event_hdr_addr(stdev, ctl->event_id, ctl->index); in event_ctl() 1083 if (ctl.event_id >= SWITCHTEC_IOCTL_MAX_EVENTS) in ioctl_event_ctl() 1090 if (event_regs[ctl.event_id].map_reg == global_ev_reg) in ioctl_event_ctl() 1092 else if (event_regs[ctl.event_id].map_reg == part_ev_reg) in ioctl_event_ctl() [all …]
|
| /drivers/s390/char/ |
| A D | sclp_sdias.c | 122 sccb->evbuf.event_id = 4712; in sclp_sdias_blk_count() 183 sccb->evbuf.event_id = 4712; in sclp_sdias_copy()
|