| /arch/x86/events/intel/ |
| A D | uncore_nhmex.c | 565 if (!atomic_read(&er->ref) || er->config == config) { in nhmex_mbox_get_shared_reg() 567 er->config = config; in nhmex_mbox_get_shared_reg() 599 if (!atomic_read(&er->ref) || !((er->config ^ config) & mask)) { in nhmex_mbox_get_shared_reg() 607 er->config &= ~mask; in nhmex_mbox_get_shared_reg() 763 return er->idx; in nhmex_mbox_extra_reg_idx() 764 return er->idx + (er->event >> NHMEX_M_PMON_CTL_INC_SEL_SHIFT) - 0xd; in nhmex_mbox_extra_reg_idx() 772 struct extra_reg *er; in nhmex_mbox_hw_config() local 780 for (er = nhmex_uncore_mbox_extra_regs; er->msr; er++) { in nhmex_mbox_hw_config() 781 if (er->event != (event->hw.config & er->config_mask)) in nhmex_mbox_hw_config() 832 config = er->config; in nhmex_mbox_shared_reg_config() [all …]
|
| A D | uncore_snbep.c | 1024 for (er = snbep_uncore_cbox_extra_regs; er->msr; er++) { in snbep_cbox_hw_config() 1025 if (er->event != (event->hw.config & er->config_mask)) in snbep_cbox_hw_config() 1027 idx |= er->idx; in snbep_cbox_hw_config() 1765 for (er = ivbep_uncore_cbox_extra_regs; er->msr; er++) { in ivbep_cbox_hw_config() 1766 if (er->event != (event->hw.config & er->config_mask)) in ivbep_cbox_hw_config() 2193 for (er = knl_uncore_cha_extra_regs; er->msr; er++) { in knl_cha_hw_config() 2194 if (er->event != (event->hw.config & er->config_mask)) in knl_cha_hw_config() 2748 for (er = hswep_uncore_cbox_extra_regs; er->msr; er++) { in hswep_cbox_hw_config() 2749 if (er->event != (event->hw.config & er->config_mask)) in hswep_cbox_hw_config() 3637 for (er = skx_uncore_cha_extra_regs; er->msr; er++) { in skx_cha_hw_config() [all …]
|
| A D | uncore.c | 183 struct intel_uncore_extra_reg *er; in uncore_get_constraint() local 198 er = &box->shared_regs[reg1->idx]; in uncore_get_constraint() 200 if (!atomic_read(&er->ref) || in uncore_get_constraint() 201 (er->config1 == reg1->config && er->config2 == reg2->config)) { in uncore_get_constraint() 202 atomic_inc(&er->ref); in uncore_get_constraint() 203 er->config1 = reg1->config; in uncore_get_constraint() 204 er->config2 = reg2->config; in uncore_get_constraint() 220 struct intel_uncore_extra_reg *er; in uncore_put_constraint() local 235 atomic_dec(&er->ref); in uncore_put_constraint() 245 er = &box->shared_regs[idx]; in uncore_shared_reg_config() [all …]
|
| A D | core.c | 6705 struct extra_reg *er; in intel_pmu_check_extra_regs() local 6715 for (er = extra_regs; er->msr; er++) { in intel_pmu_check_extra_regs() 6716 er->extra_msr_access = check_msr(er->msr, 0x11UL); in intel_pmu_check_extra_regs() 6718 if ((er->idx == EXTRA_REG_LBR) && !er->extra_msr_access) in intel_pmu_check_extra_regs()
|
| /arch/powerpc/platforms/44x/ |
| A D | uic.c | 57 u32 er, sr; in uic_unmask_irq() local 64 er = mfdcr(uic->dcrbase + UIC_ER); in uic_unmask_irq() 65 er |= sr; in uic_unmask_irq() 66 mtdcr(uic->dcrbase + UIC_ER, er); in uic_unmask_irq() 75 u32 er; in uic_mask_irq() local 78 er = mfdcr(uic->dcrbase + UIC_ER); in uic_mask_irq() 79 er &= ~(1 << (31 - src)); in uic_mask_irq() 80 mtdcr(uic->dcrbase + UIC_ER, er); in uic_mask_irq() 100 u32 er, sr; in uic_mask_ack_irq() local 105 er &= ~sr; in uic_mask_ack_irq() [all …]
|
| /arch/arm/mach-ux500/ |
| A D | pm.c | 88 u32 er; /* Enable register */ in prcmu_gic_pending_irq() local 95 er = readl_relaxed(dist_base + GIC_DIST_ENABLE_SET + i * 4); in prcmu_gic_pending_irq() 97 if (pr & er) in prcmu_gic_pending_irq() 142 u32 er; /* Enable register */ in prcmu_copy_gic_settings() local 147 er = readl_relaxed(dist_base + in prcmu_copy_gic_settings() 149 writel(er, PRCM_ARMITMSK31TO0 + i * 4); in prcmu_copy_gic_settings()
|
| /arch/loongarch/kvm/ |
| A D | exit.c | 258 er = EMULATE_FAIL; in kvm_complete_iocsr_read() 262 return er; in kvm_complete_iocsr_read() 289 er = EMULATE_FAIL; in kvm_trap_handle_gspr() 302 er = EMULATE_DONE; in kvm_trap_handle_gspr() 314 er = EMULATE_FAIL; in kvm_trap_handle_gspr() 319 er = EMULATE_FAIL; in kvm_trap_handle_gspr() 324 er = EMULATE_FAIL; in kvm_trap_handle_gspr() 329 if (er == EMULATE_FAIL) { in kvm_trap_handle_gspr() 337 return er; in kvm_trap_handle_gspr() 533 er = EMULATE_FAIL; in kvm_complete_mmio_read() [all …]
|
| /arch/mips/kvm/ |
| A D | vz.c | 892 return er; in kvm_vz_gpsi_cop0() 1071 return er; in kvm_vz_gpsi_cop0() 1091 return er; in kvm_vz_gpsi_cache() 1159 return er; in kvm_vz_gpsi_lwc2() 1196 er = EMULATE_FAIL; in kvm_vz_gpsi_lwc2() 1208 return er; in kvm_vz_gpsi_lwc2() 1291 return er; in kvm_trap_vz_handle_gpsi() 1417 return er; in kvm_trap_vz_handle_gsfc() 1453 return er; in kvm_trap_vz_handle_hc() 1459 return er; in kvm_trap_vz_handle_hc() [all …]
|
| A D | emulate.c | 973 enum emulation_result er; in kvm_mips_emulate_store() local 985 er = update_pc(vcpu, cause); in kvm_mips_emulate_store() 986 if (er == EMULATE_FAIL) in kvm_mips_emulate_store() 987 return er; in kvm_mips_emulate_store() 1273 enum emulation_result er; in kvm_mips_emulate_load() local 1287 er = update_pc(vcpu, cause); in kvm_mips_emulate_load() 1288 if (er == EMULATE_FAIL) in kvm_mips_emulate_load() 1289 return er; in kvm_mips_emulate_load() 1513 enum emulation_result er = EMULATE_DONE; in kvm_mips_complete_mmio_load() local 1517 er = EMULATE_FAIL; in kvm_mips_complete_mmio_load() [all …]
|
| A D | mips.c | 1175 enum emulation_result er = EMULATE_DONE; in __kvm_mips_handle_exit() local 1311 if (er == EMULATE_DONE && !(ret & RESUME_HOST)) in __kvm_mips_handle_exit()
|
| /arch/x86/events/ |
| A D | core.c | 171 struct extra_reg *er; in x86_pmu_extra_regs() local 178 for (er = extra_regs; er->msr; er++) { in x86_pmu_extra_regs() 179 if (er->event != (config & er->config_mask)) in x86_pmu_extra_regs() 181 if (event->attr.config1 & ~er->valid_mask) in x86_pmu_extra_regs() 184 if (!er->extra_msr_access) in x86_pmu_extra_regs() 187 reg->idx = er->idx; in x86_pmu_extra_regs() 189 reg->reg = er->msr; in x86_pmu_extra_regs()
|
| /arch/powerpc/kvm/ |
| A D | book3s_pr.c | 969 enum emulation_result er = EMULATE_FAIL; in kvmppc_emulate_fac() local 972 er = kvmppc_emulate_instruction(vcpu); in kvmppc_emulate_fac() 974 if ((er != EMULATE_DONE) && (er != EMULATE_AGAIN)) { in kvmppc_emulate_fac() 1075 enum emulation_result er; in kvmppc_exit_pr_progint() local 1107 er = kvmppc_emulate_instruction(vcpu); in kvmppc_exit_pr_progint() 1108 switch (er) { in kvmppc_exit_pr_progint()
|
| A D | booke.c | 832 enum emulation_result er; in emulation_exit() local 834 er = kvmppc_emulate_instruction(vcpu); in emulation_exit() 835 switch (er) { in emulation_exit()
|
| A D | powerpc.c | 284 enum emulation_result er; in kvmppc_emulate_mmio() local 287 er = kvmppc_emulate_loadstore(vcpu); in kvmppc_emulate_mmio() 288 switch (er) { in kvmppc_emulate_mmio()
|
| /arch/powerpc/boot/dts/fsl/ |
| A D | e6500_power_isa.dtsi | 61 fsl,eref-er; // Enhanced Reservations (Load and Reserve and Store Cond.)
|
| /arch/arm/boot/dts/ti/omap/ |
| A D | am335x-osd3358-sm-red.dts | 44 * https://www.ti.com/lit/er/sprz360i/sprz360i.pdf
|
| /arch/arm64/boot/dts/rockchip/ |
| A D | rk3568-pinctrl.dtsi | 585 gmac0_rx_er: gmac0-rx-er { 654 gmac1m0_rx_er: gmac1m0-rx-er { 721 gmac1m1_rx_er: gmac1m1-rx-er {
|