/linux/arch/x86/mm/ |
A D | tlb.c | 202 if (asid == this_cpu_read(cpu_tlbstate.loaded_mm_asid)) in clear_asid_other() 227 if (this_cpu_read(cpu_tlbstate.invalidate_other)) in choose_new_asid() 231 if (this_cpu_read(cpu_tlbstate.ctxs[asid].ctx_id) != in choose_new_asid() 315 WARN_ON(!this_cpu_read(cpu_tlbstate_shared.is_lazy)); in leave_mm() 359 if (this_cpu_read(cpu_info.smt_active)) { in l1d_flush_evaluate() 388 prev_mm = this_cpu_read(cpu_tlbstate.last_user_mm_spec); in cond_mitigation() 485 cr4_update_pce_mm(this_cpu_read(cpu_tlbstate.loaded_mm)); in cr4_update_pce() 673 if (this_cpu_read(cpu_tlbstate.loaded_mm) == &init_mm) in enter_lazy_tlb() 774 if (this_cpu_read(cpu_tlbstate_shared.is_lazy)) { in flush_tlb_func() 1090 build_cr3(this_cpu_read(cpu_tlbstate.loaded_mm)->pgd, in __get_current_cr3_fast() [all …]
|
/linux/kernel/trace/ |
A D | trace_preemptirq.c | 44 if (this_cpu_read(tracing_irq_cpu)) { in trace_hardirqs_on_prepare() 55 if (this_cpu_read(tracing_irq_cpu)) { in trace_hardirqs_on() 75 if (!this_cpu_read(tracing_irq_cpu)) { in trace_hardirqs_off_finish() 89 if (!this_cpu_read(tracing_irq_cpu)) { in trace_hardirqs_off()
|
/linux/drivers/irqchip/ |
A D | irq-csky-mpintc.c | 56 (this_cpu_read(intcl_reg) + INTCL_CFGR) : (INTCG_base + INTCG_CICFGR))) 75 void __iomem *reg_base = this_cpu_read(intcl_reg); in csky_mpintc_handler() 83 void __iomem *reg_base = this_cpu_read(intcl_reg); in csky_mpintc_unmask() 92 void __iomem *reg_base = this_cpu_read(intcl_reg); in csky_mpintc_mask() 99 void __iomem *reg_base = this_cpu_read(intcl_reg); in csky_mpintc_eoi() 215 void __iomem *reg_base = this_cpu_read(intcl_reg); in csky_mpintc_send_ipi()
|
/linux/arch/x86/include/asm/xen/ |
A D | hypervisor.h | 79 enum xen_lazy_mode old_mode = this_cpu_read(xen_lazy_mode); in enter_lazy() 93 BUG_ON(this_cpu_read(xen_lazy_mode) != mode); in leave_lazy() 95 if (this_cpu_read(xen_lazy_nesting) == 0) in leave_lazy()
|
/linux/arch/arm64/kernel/ |
A D | topology.c | 158 prev_const_cnt = this_cpu_read(arch_const_cycles_prev); in amu_scale_freq_tick() 159 prev_core_cnt = this_cpu_read(arch_core_cycles_prev); in amu_scale_freq_tick() 163 const_cnt = this_cpu_read(arch_const_cycles_prev); in amu_scale_freq_tick() 164 core_cnt = this_cpu_read(arch_core_cycles_prev); in amu_scale_freq_tick() 179 scale *= this_cpu_read(arch_max_freq_scale); in amu_scale_freq_tick()
|
/linux/arch/x86/include/asm/ |
A D | kvmclock.h | 11 return &this_cpu_read(hv_clock_per_cpu)->pvti; in this_cpu_pvti() 16 return this_cpu_read(hv_clock_per_cpu); in this_cpu_hvclock()
|
A D | switch_to.h | 58 if (unlikely(this_cpu_read(cpu_tss_rw.x86_tss.ss1) == thread->sysenter_cs)) in refresh_sysenter_cs()
|
/linux/arch/x86/kernel/ |
A D | dumpstack_32.c | 40 unsigned long *begin = (unsigned long *)this_cpu_read(pcpu_hot.hardirq_stack_ptr); in in_hardirq_stack() 65 unsigned long *begin = (unsigned long *)this_cpu_read(pcpu_hot.softirq_stack_ptr); in in_softirq_stack() 102 info->next_sp = (unsigned long *)this_cpu_read(cpu_tss_rw.x86_tss.sp); in in_doublefault_stack()
|
A D | nmi.c | 512 if (this_cpu_read(nmi_state) != NMI_NOT_RUNNING) { in DEFINE_IDTENTRY_RAW() 554 local_db_restore(this_cpu_read(nmi_dr7)); in DEFINE_IDTENTRY_RAW() 558 if (unlikely(this_cpu_read(nmi_cr2) != read_cr2())) in DEFINE_IDTENTRY_RAW() 559 write_cr2(this_cpu_read(nmi_cr2)); in DEFINE_IDTENTRY_RAW() 692 if (unlikely(this_cpu_read(nmi_cr2) != read_cr2())) in DEFINE_FREDENTRY_NMI() 693 write_cr2(this_cpu_read(nmi_cr2)); in DEFINE_FREDENTRY_NMI()
|
A D | tsc.c | 78 seq = this_cpu_read(cyc2ns.seq.seqcount.sequence); in __cyc2ns_read() 81 data->cyc2ns_offset = this_cpu_read(cyc2ns.data[idx].cyc2ns_offset); in __cyc2ns_read() 82 data->cyc2ns_mul = this_cpu_read(cyc2ns.data[idx].cyc2ns_mul); in __cyc2ns_read() 83 data->cyc2ns_shift = this_cpu_read(cyc2ns.data[idx].cyc2ns_shift); in __cyc2ns_read() 85 } while (unlikely(seq != this_cpu_read(cyc2ns.seq.seqcount.sequence))); in __cyc2ns_read()
|
/linux/arch/loongarch/kernel/ |
A D | kfpu.c | 27 WARN_ON(this_cpu_read(in_kernel_fpu)); in kernel_fpu_begin() 55 WARN_ON(!this_cpu_read(in_kernel_fpu)); in kernel_fpu_end()
|
/linux/include/linux/ |
A D | lockdep.h | 573 #define __lockdep_enabled (debug_locks && !this_cpu_read(lockdep_recursion)) 577 WARN_ON_ONCE(__lockdep_enabled && !this_cpu_read(hardirqs_enabled)); \ 582 WARN_ON_ONCE(__lockdep_enabled && this_cpu_read(hardirqs_enabled)); \ 587 WARN_ON_ONCE(__lockdep_enabled && !this_cpu_read(hardirq_context)); \ 592 WARN_ON_ONCE(__lockdep_enabled && (this_cpu_read(hardirq_context) || \ 593 !this_cpu_read(hardirqs_enabled))); \ 601 !this_cpu_read(hardirqs_enabled))); \ 609 this_cpu_read(hardirqs_enabled))); \
|
/linux/kernel/ |
A D | watchdog_perf.c | 169 perf_event_enable(this_cpu_read(watchdog_ev)); in watchdog_hardlockup_enable() 178 struct perf_event *event = this_cpu_read(watchdog_ev); in watchdog_hardlockup_disable() 275 perf_event_release_kernel(this_cpu_read(watchdog_ev)); in watchdog_hardlockup_probe()
|
/linux/arch/x86/kernel/cpu/mce/ |
A D | amd.c | 420 if (!this_cpu_read(threshold_banks) && !tr->set_lvt_off) in threshold_restart_bank() 679 for (bank = 0; bank < this_cpu_read(mce_num_banks); ++bank) { in mce_amd_feature_init() 882 for (bank = 0; bank < this_cpu_read(mce_num_banks); ++bank) in amd_deferred_error_interrupt() 921 struct threshold_bank **bp = this_cpu_read(threshold_banks); in amd_threshold_interrupt() 932 for (bank = 0; bank < this_cpu_read(mce_num_banks); ++bank) { in amd_threshold_interrupt() 1224 struct device *dev = this_cpu_read(mce_device); in threshold_create_bank() 1352 unsigned int bank, numbanks = this_cpu_read(mce_num_banks); in __threshold_remove_device() 1366 struct threshold_bank **bp = this_cpu_read(threshold_banks); in mce_threshold_remove_device() 1401 bp = this_cpu_read(threshold_banks); in mce_threshold_create_device() 1405 numbanks = this_cpu_read(mce_num_banks); in mce_threshold_create_device() [all …]
|
A D | core.c | 705 for (i = 0; i < this_cpu_read(mce_num_banks); i++) { in machine_check_poll() 914 for (i = 0; i < this_cpu_read(mce_num_banks); i++) { in mce_no_way_out() 1229 for (i = 0; i < this_cpu_read(mce_num_banks); i++) { in mce_clear_state() 1279 for (i = 0; i < this_cpu_read(mce_num_banks); i++) { in __mc_scan_banks() 1771 u8 n_banks = this_cpu_read(mce_num_banks); in __mcheck_cpu_mce_banks_init() 1846 for (i = 0; i < this_cpu_read(mce_num_banks); i++) { in __mcheck_cpu_init_clear_banks() 1872 for (i = 0; i < this_cpu_read(mce_num_banks); i++) { in __mcheck_cpu_check_banks() 2052 if (this_cpu_read(mce_num_banks) > 8) in mce_zhaoxin_feature_init() 2301 if (bank >= this_cpu_read(mce_num_banks)) { in mce_disable_bank() 2390 for (i = 0; i < this_cpu_read(mce_num_banks); i++) { in mce_disable_error_reporting() [all …]
|
/linux/Documentation/translations/zh_CN/core-api/ |
A D | this_cpu_ops.rst | 46 this_cpu_read(pcp) 75 z = this_cpu_read(x); 232 且更容易与 ``()`` 结合。第二种形式也与 ``this_cpu_read()`` 和大家的使用方式一致。
|
/linux/kernel/printk/ |
A D | printk_safe.c | 48 this_cpu_read(printk_context) || in is_printk_legacy_deferred()
|
/linux/tools/testing/shared/linux/ |
A D | percpu.h | 7 #define this_cpu_read(var) var macro
|
/linux/include/crypto/internal/ |
A D | simd.h | 56 (may_use_simd() && !this_cpu_read(crypto_simd_disabled_for_test))
|
/linux/arch/x86/xen/ |
A D | suspend.c | 42 wrmsrl(MSR_IA32_SPEC_CTRL, this_cpu_read(spec_ctrl)); in xen_vcpu_notify_restore()
|
/linux/arch/x86/kernel/fpu/ |
A D | context.h | 38 return fpu == this_cpu_read(fpu_fpregs_owner_ctx) && cpu == fpu->last_cpu; in fpregs_state_valid()
|
/linux/drivers/accessibility/speakup/ |
A D | fakekey.c | 86 return this_cpu_read(reporting_keystroke); in speakup_fake_key_pressed()
|
/linux/arch/x86/coco/sev/ |
A D | core.c | 255 data = this_cpu_read(runtime_data); in __sev_get_ghcb() 626 return this_cpu_read(svsm_caa); in svsm_get_caa() 639 return this_cpu_read(svsm_caa_pa); in svsm_get_caa_pa() 651 data = this_cpu_read(runtime_data); in __sev_put_ghcb() 1023 call.caa = this_cpu_read(svsm_caa); in snp_set_vmsa() 1348 regs->ax = lower_32_bits(this_cpu_read(svsm_caa_pa)); in vc_handle_msr() 1349 regs->dx = upper_32_bits(this_cpu_read(svsm_caa_pa)); in vc_handle_msr() 1375 data = this_cpu_read(runtime_data); in snp_register_per_cpu_ghcb() 1567 call.rcx = this_cpu_read(svsm_caa_pa); in sev_es_init_vc_handling() 1793 struct sev_es_runtime_data *data = this_cpu_read(runtime_data); in vc_handle_dr7_write() [all …]
|
/linux/net/ipv6/netfilter/ |
A D | nf_dup_ipv6.c | 51 if (this_cpu_read(nf_skb_duplicated)) in nf_dup_ipv6()
|
/linux/net/ipv4/netfilter/ |
A D | nf_dup_ipv4.c | 57 if (this_cpu_read(nf_skb_duplicated)) in nf_dup_ipv4()
|