| /arch/sparc/kernel/ |
| A D | nmi.c | 107 if (__this_cpu_read(nmi_touch)) { in perfctr_irq() 111 if (!touched && __this_cpu_read(last_irq_sum) == sum) { in perfctr_irq() 113 if (__this_cpu_read(alert_counter) == 30 * nmi_hz) in perfctr_irq() 120 if (__this_cpu_read(wd_enabled)) { in perfctr_irq() 160 if (!__this_cpu_read(wd_enabled)) in stop_nmi_watchdog() 217 if (__this_cpu_read(wd_enabled)) in start_nmi_watchdog() 231 if (!__this_cpu_read(wd_enabled)) in nmi_adjust_hz_one()
|
| /arch/mips/kernel/ |
| A D | mips-r2-to-r6-emul.c | 2243 (unsigned long)__this_cpu_read(mipsr2emustats.movs), in mipsr2_emul_show() 2246 (unsigned long)__this_cpu_read(mipsr2emustats.hilo), in mipsr2_emul_show() 2249 (unsigned long)__this_cpu_read(mipsr2emustats.muls), in mipsr2_emul_show() 2252 (unsigned long)__this_cpu_read(mipsr2emustats.divs), in mipsr2_emul_show() 2255 (unsigned long)__this_cpu_read(mipsr2emustats.dsps), in mipsr2_emul_show() 2258 (unsigned long)__this_cpu_read(mipsr2emustats.bops), in mipsr2_emul_show() 2261 (unsigned long)__this_cpu_read(mipsr2emustats.traps), in mipsr2_emul_show() 2264 (unsigned long)__this_cpu_read(mipsr2emustats.fpus), in mipsr2_emul_show() 2267 (unsigned long)__this_cpu_read(mipsr2emustats.loads), in mipsr2_emul_show() 2270 (unsigned long)__this_cpu_read(mipsr2emustats.stores), in mipsr2_emul_show() [all …]
|
| /arch/x86/coco/sev/ |
| A D | sev-nmi.c | 57 new_ist = old_ist = __this_cpu_read(cpu_tss_rw.x86_tss.ist[IST_INDEX_VC]); in __sev_es_ist_enter() 83 ist = __this_cpu_read(cpu_tss_rw.x86_tss.ist[IST_INDEX_VC]); in __sev_es_ist_exit()
|
| /arch/x86/kernel/ |
| A D | hw_breakpoint.c | 485 set_debugreg(__this_cpu_read(cpu_debugreg[0]), 0); in hw_breakpoint_restore() 486 set_debugreg(__this_cpu_read(cpu_debugreg[1]), 1); in hw_breakpoint_restore() 487 set_debugreg(__this_cpu_read(cpu_debugreg[2]), 2); in hw_breakpoint_restore() 488 set_debugreg(__this_cpu_read(cpu_debugreg[3]), 3); in hw_breakpoint_restore() 490 set_debugreg(__this_cpu_read(cpu_dr7), 7); in hw_breakpoint_restore()
|
| A D | irq_32.c | 75 irqstk = __this_cpu_read(hardirq_stack_ptr); in execute_on_irq_stack() 135 irqstk = __this_cpu_read(softirq_stack_ptr); in do_softirq_own_stack()
|
| A D | irq.c | 261 struct irq_desc *desc = __this_cpu_read(vector_irq[vector]); in reevaluate_vector() 275 struct irq_desc *desc = __this_cpu_read(vector_irq[vector]); in call_irq_handler() 508 if (IS_ERR_OR_NULL(__this_cpu_read(vector_irq[vector]))) in fixup_irqs() 512 desc = __this_cpu_read(vector_irq[vector]); in fixup_irqs() 523 if (__this_cpu_read(vector_irq[vector]) != VECTOR_RETRIGGERED) in fixup_irqs()
|
| A D | fred.c | 59 wrmsrq(MSR_IA32_FRED_RSP0, __this_cpu_read(fred_rsp0)); in cpu_init_fred_exceptions()
|
| A D | kvm.c | 250 if (__this_cpu_read(async_pf_enabled)) { in kvm_read_and_reset_apf_flags() 251 flags = __this_cpu_read(apf_reason.flags); in kvm_read_and_reset_apf_flags() 301 if (__this_cpu_read(async_pf_enabled)) { in DEFINE_IDTENTRY_SYSVEC() 302 token = __this_cpu_read(apf_reason.token); in DEFINE_IDTENTRY_SYSVEC() 389 if (!__this_cpu_read(async_pf_enabled)) in kvm_pv_disable_apf()
|
| A D | dumpstack_64.c | 103 begin = (unsigned long)__this_cpu_read(cea_exception_stacks); in in_exception_stack()
|
| /arch/powerpc/lib/ |
| A D | code-patching.c | 292 patching_mm = __this_cpu_read(cpu_patching_context.mm); in __do_patch_mem_mm() 293 text_poke_addr = __this_cpu_read(cpu_patching_context.addr); in __do_patch_mem_mm() 335 text_poke_addr = (unsigned long)__this_cpu_read(cpu_patching_context.addr) & PAGE_MASK; in __do_patch_mem() 338 pte = __this_cpu_read(cpu_patching_context.pte); in __do_patch_mem() 478 patching_mm = __this_cpu_read(cpu_patching_context.mm); in __do_patch_instructions_mm() 479 text_poke_addr = __this_cpu_read(cpu_patching_context.addr); in __do_patch_instructions_mm() 527 text_poke_addr = (unsigned long)__this_cpu_read(cpu_patching_context.addr) & PAGE_MASK; in __do_patch_instructions() 530 pte = __this_cpu_read(cpu_patching_context.pte); in __do_patch_instructions()
|
| /arch/x86/include/asm/ |
| A D | cpu_entry_area.h | 148 CEA_ESTACK_TOP(__this_cpu_read(cea_exception_stacks), name) 151 CEA_ESTACK_BOT(__this_cpu_read(cea_exception_stacks), name)
|
| A D | irq_stack.h | 119 call_on_stack(__this_cpu_read(hardirq_stack_ptr), \ 138 if (user_mode(regs) || __this_cpu_read(hardirq_stack_inuse)) { \
|
| A D | hardirq.h | 88 return __this_cpu_read(irq_stat.kvm_cpu_l1tf_flush_l1d); in kvm_get_cpu_l1tf_flush_l1d()
|
| A D | debugreg.h | 123 return __this_cpu_read(cpu_dr7) & DR_GLOBAL_ENABLE_MASK; in hw_breakpoint_active()
|
| A D | fred.h | 103 if (cpu_feature_enabled(X86_FEATURE_FRED) && (__this_cpu_read(fred_rsp0) != rsp0)) { in fred_update_rsp0()
|
| /arch/arm64/include/asm/ |
| A D | arch_timer.h | 27 __wa = __this_cpu_read(timer_unstable_counter_workaround); \ 34 __wa = __this_cpu_read(timer_unstable_counter_workaround); \
|
| A D | percpu.h | 265 #undef __this_cpu_read 266 #define __this_cpu_read raw_cpu_read macro
|
| /arch/powerpc/kernel/ |
| A D | hw_breakpoint.c | 236 struct perf_event *bp = __this_cpu_read(bp_per_reg[i]); in thread_change_pc() 246 info = counter_arch_bp(__this_cpu_read(bp_per_reg[i])); in thread_change_pc() 408 bp[i] = __this_cpu_read(bp_per_reg[i]); in hw_breakpoint_handler() 519 bp = __this_cpu_read(bp_per_reg[i]); in single_step_dabr_instruction()
|
| /arch/x86/xen/ |
| A D | time.c | 53 src = &__this_cpu_read(xen_vcpu)->time; in xen_clocksource_read() 69 src = &__this_cpu_read(xen_vcpu)->time; in xen_sched_clock() 547 pvti = &__this_cpu_read(xen_vcpu)->time; in xen_time_init() 628 if (!__this_cpu_read(xen_vcpu)) { in xen_hvm_init_time_ops()
|
| A D | spinlock.c | 38 int irq = __this_cpu_read(lock_kicker_irq); in xen_qlock_wait()
|
| /arch/powerpc/include/asm/ |
| A D | time.h | 99 return __this_cpu_read(decrementers_next_tb); in timer_get_next_tb()
|
| /arch/arm/kernel/ |
| A D | irq.c | 83 __this_cpu_read(irq_stack_ptr)); in do_softirq_own_stack()
|
| /arch/x86/kernel/acpi/ |
| A D | sleep.c | 95 if (__this_cpu_read(cpu_info.cpuid_level) >= 0) { in x86_acpi_suspend_lowlevel()
|
| /arch/powerpc/sysdev/xive/ |
| A D | common.c | 346 struct xive_cpu *xc = __this_cpu_read(xive_cpu); in xive_get_irq() 441 struct xive_cpu *xc = __this_cpu_read(xive_cpu); in xive_irq_eoi() 1079 struct xive_cpu *xc = __this_cpu_read(xive_cpu); in xive_ipi_eoi() 1514 struct xive_cpu *xc = __this_cpu_read(xive_cpu); in xive_setup_cpu() 1606 struct xive_cpu *xc = __this_cpu_read(xive_cpu); in xive_smp_disable_cpu() 1626 struct xive_cpu *xc = __this_cpu_read(xive_cpu); in xive_flush_interrupt() 1639 struct xive_cpu *xc = __this_cpu_read(xive_cpu); in xive_teardown_cpu()
|
| /arch/x86/kernel/cpu/mce/ |
| A D | threshold.c | 52 return __this_cpu_read(storm_desc.poll_mode); in mce_get_storm_mode()
|