| /arch/powerpc/include/asm/ |
| A D | idle.h | 14 *this_cpu_ptr(&idle_entry_purr_snap) = mfspr(SPRN_PURR); in snapshot_purr_idle_entry() 19 *this_cpu_ptr(&idle_entry_spurr_snap) = mfspr(SPRN_SPURR); in snapshot_spurr_idle_entry() 25 u64 in_purr = *this_cpu_ptr(&idle_entry_purr_snap); in update_idle_purr_accounting() 34 u64 *idle_spurr_cycles_ptr = this_cpu_ptr(&idle_spurr_cycles); in update_idle_spurr_accounting() 35 u64 in_spurr = *this_cpu_ptr(&idle_entry_spurr_snap); in update_idle_spurr_accounting() 89 return *this_cpu_ptr(&idle_spurr_cycles); in read_this_idle_spurr()
|
| A D | xics.h | 114 struct xics_cppr *os_cppr = this_cpu_ptr(&xics_cppr); in xics_push_cppr() 127 struct xics_cppr *os_cppr = this_cpu_ptr(&xics_cppr); in xics_pop_cppr() 137 struct xics_cppr *os_cppr = this_cpu_ptr(&xics_cppr); in xics_set_base_cppr() 149 struct xics_cppr *os_cppr = this_cpu_ptr(&xics_cppr); in xics_cppr_top()
|
| /arch/arm64/kvm/hyp/nvhe/ |
| A D | stacktrace.c | 28 struct kvm_nvhe_stacktrace_info *stacktrace_info = this_cpu_ptr(&kvm_stacktrace_info); in hyp_prepare_backtrace() 29 struct kvm_nvhe_init_params *params = this_cpu_ptr(&kvm_init_params); in hyp_prepare_backtrace() 32 stacktrace_info->overflow_stack_base = (unsigned long)this_cpu_ptr(overflow_stack); in hyp_prepare_backtrace() 44 unsigned long low = (unsigned long)this_cpu_ptr(overflow_stack); in stackinfo_get_overflow() 55 struct kvm_nvhe_init_params *params = this_cpu_ptr(&kvm_init_params); in stackinfo_get_hyp() 96 unsigned long *stacktrace = this_cpu_ptr(pkvm_stacktrace); in pkvm_save_backtrace_entry()
|
| A D | psci-relay.c | 160 boot_args = this_cpu_ptr(&suspend_args); in psci_cpu_suspend() 161 init_params = this_cpu_ptr(&kvm_init_params); in psci_cpu_suspend() 187 boot_args = this_cpu_ptr(&suspend_args); in psci_system_suspend() 188 init_params = this_cpu_ptr(&kvm_init_params); in psci_system_suspend() 211 boot_args = this_cpu_ptr(&cpu_on_args); in __kvm_host_psci_cpu_entry() 213 boot_args = this_cpu_ptr(&suspend_args); in __kvm_host_psci_cpu_entry()
|
| /arch/x86/kernel/cpu/mce/ |
| A D | intel.c | 166 machine_check_poll(MCP_TIMESTAMP, this_cpu_ptr(&mce_banks_owned)); in intel_threshold_interrupt() 178 unsigned long *owned = (void *)this_cpu_ptr(&mce_banks_owned); in cmci_skip_bank() 192 __clear_bit(bank, this_cpu_ptr(mce_poll_banks)); in cmci_skip_bank() 232 struct mca_storm_desc *storm = this_cpu_ptr(&storm_desc); in cmci_claim_bank() 240 WARN_ON(!test_bit(bank, this_cpu_ptr(mce_poll_banks))); in cmci_claim_bank() 246 set_bit(bank, (void *)this_cpu_ptr(&mce_banks_owned)); in cmci_claim_bank() 253 __clear_bit(bank, this_cpu_ptr(mce_poll_banks)); in cmci_claim_bank() 316 machine_check_poll(0, this_cpu_ptr(&mce_banks_owned)); in cmci_recheck() 325 if (!test_bit(bank, this_cpu_ptr(mce_banks_owned))) in __cmci_disable_bank() 330 __clear_bit(bank, this_cpu_ptr(mce_banks_owned)); in __cmci_disable_bank() [all …]
|
| A D | threshold.c | 37 struct mca_storm_desc *storm = this_cpu_ptr(&storm_desc); in mce_inherit_storm() 71 struct mca_storm_desc *storm = this_cpu_ptr(&storm_desc); in cmci_storm_begin() 73 __set_bit(bank, this_cpu_ptr(mce_poll_banks)); in cmci_storm_begin() 86 struct mca_storm_desc *storm = this_cpu_ptr(&storm_desc); in cmci_storm_end() 88 __clear_bit(bank, this_cpu_ptr(mce_poll_banks)); in cmci_storm_end() 99 struct mca_storm_desc *storm = this_cpu_ptr(&storm_desc); in mce_track_storm()
|
| A D | core.c | 437 *(u64 *)((char *)this_cpu_ptr(&injectm) + offset) = v; in mce_wrmsrq() 1564 final = this_cpu_ptr(&hw_errs_seen); in do_machine_check() 1738 machine_check_poll(0, this_cpu_ptr(&mce_poll_banks)); in mc_poll_banks_default() 1750 struct timer_list *cpu_t = this_cpu_ptr(&mce_timer); in mce_timer_fn() 1757 if (mce_available(this_cpu_ptr(&cpu_info))) in mce_timer_fn() 1784 struct timer_list *t = this_cpu_ptr(&mce_timer); in mce_timer_kick() 2172 struct timer_list *t = this_cpu_ptr(&mce_timer); in __mcheck_cpu_setup_timer() 2179 struct timer_list *t = this_cpu_ptr(&mce_timer); in __mcheck_cpu_init_timer() 2344 __clear_bit(bank, this_cpu_ptr(mce_poll_banks)); in __mce_disable_bank() 2807 struct timer_list *t = this_cpu_ptr(&mce_timer); in mce_cpu_online() [all …]
|
| /arch/x86/events/amd/ |
| A D | brs.c | 205 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in amd_brs_enable() 221 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in amd_brs_enable_all() 228 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in amd_brs_disable() 257 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in amd_brs_disable_all() 283 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in amd_brs_drain() 387 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in amd_pmu_brs_sched_task() 408 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in perf_amd_brs_lopwr_cb()
|
| A D | lbr.c | 100 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in amd_pmu_lbr_filter() 163 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in amd_pmu_lbr_read() 323 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in amd_pmu_lbr_reset() 342 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in amd_pmu_lbr_add() 362 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in amd_pmu_lbr_del() 378 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in amd_pmu_lbr_sched_task() 391 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in amd_pmu_lbr_enable_all() 414 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in amd_pmu_lbr_disable_all()
|
| /arch/xtensa/kernel/ |
| A D | hw_breakpoint.c | 178 i = alloc_slot(this_cpu_ptr(bp_on_reg), XCHAL_NUM_IBREAK, bp); in arch_install_hw_breakpoint() 185 i = alloc_slot(this_cpu_ptr(wp_on_reg), XCHAL_NUM_DBREAK, bp); in arch_install_hw_breakpoint() 216 i = free_slot(this_cpu_ptr(bp_on_reg), XCHAL_NUM_IBREAK, bp); in arch_uninstall_hw_breakpoint() 224 i = free_slot(this_cpu_ptr(wp_on_reg), XCHAL_NUM_DBREAK, bp); in arch_uninstall_hw_breakpoint() 269 struct perf_event *bp = this_cpu_ptr(wp_on_reg)[i]; in restore_dbreak() 281 struct perf_event **bp = this_cpu_ptr(bp_on_reg); in check_hw_breakpoint() 290 struct perf_event **bp = this_cpu_ptr(wp_on_reg); in check_hw_breakpoint()
|
| /arch/x86/events/intel/ |
| A D | bts.c | 235 struct bts_ctx *bts = this_cpu_ptr(bts_ctx); in __bts_event_start() 263 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in bts_event_start() 264 struct bts_ctx *bts = this_cpu_ptr(bts_ctx); in bts_event_start() 294 struct bts_ctx *bts = this_cpu_ptr(bts_ctx); in __bts_event_stop() 308 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in bts_event_stop() 309 struct bts_ctx *bts = this_cpu_ptr(bts_ctx); in bts_event_stop() 348 bts = this_cpu_ptr(bts_ctx); in intel_bts_enable_local() 372 bts = this_cpu_ptr(bts_ctx); in intel_bts_disable_local() 463 struct debug_store *ds = this_cpu_ptr(&cpu_hw_events)->ds; in intel_bts_interrupt() 473 bts = this_cpu_ptr(bts_ctx); in intel_bts_interrupt() [all …]
|
| /arch/arm64/kvm/ |
| A D | vmid.c | 135 atomic64_set(this_cpu_ptr(&active_vmids), VMID_ACTIVE_INVALID); in kvm_arm_vmid_clear_active() 155 old_active_vmid = atomic64_read(this_cpu_ptr(&active_vmids)); in kvm_arm_vmid_update() 157 0 != atomic64_cmpxchg_relaxed(this_cpu_ptr(&active_vmids), in kvm_arm_vmid_update() 168 atomic64_set(this_cpu_ptr(&active_vmids), vmid); in kvm_arm_vmid_update()
|
| /arch/powerpc/platforms/powernv/ |
| A D | opal-tracepoints.c | 54 depth = this_cpu_ptr(&opal_trace_depth); in __trace_opal_entry() 75 depth = this_cpu_ptr(&opal_trace_depth); in __trace_opal_exit()
|
| A D | subcore.c | 300 if (this_cpu_ptr(&split_state)->master) { in cpu_update_split_mode() 314 if (this_cpu_ptr(&split_state)->master) { in cpu_update_split_mode() 363 this_cpu_ptr(&split_state)->master = 1; in set_subcores_per_core()
|
| /arch/arm64/include/asm/ |
| A D | hardirq.h | 36 ___ctx = this_cpu_ptr(&nmi_contexts); \ 69 ___ctx = this_cpu_ptr(&nmi_contexts); \
|
| /arch/powerpc/include/asm/book3s/64/ |
| A D | tlbflush-hash.h | 40 batch = this_cpu_ptr(&ppc64_tlb_batch); in arch_enter_lazy_mmu_mode() 50 batch = this_cpu_ptr(&ppc64_tlb_batch); in arch_leave_lazy_mmu_mode()
|
| /arch/x86/xen/ |
| A D | multicalls.c | 68 return this_cpu_ptr(mc_debug_data_ptr); in get_mc_debug() 144 struct mc_buffer *b = this_cpu_ptr(&mc_buffer); in xen_mc_flush() 221 struct mc_buffer *b = this_cpu_ptr(&mc_buffer); in __xen_mc_entry() 256 struct mc_buffer *b = this_cpu_ptr(&mc_buffer); in xen_mc_extend_args() 286 struct mc_buffer *b = this_cpu_ptr(&mc_buffer); in xen_mc_callback()
|
| /arch/arm64/kernel/ |
| A D | hw_breakpoint.c | 238 slots = this_cpu_ptr(bp_on_reg); in hw_breakpoint_control() 245 slots = this_cpu_ptr(wp_on_reg); in hw_breakpoint_control() 590 slots = this_cpu_ptr(bp_on_reg); in toggle_bp_registers() 594 slots = this_cpu_ptr(wp_on_reg); in toggle_bp_registers() 631 slots = this_cpu_ptr(bp_on_reg); in do_breakpoint() 681 kernel_step = this_cpu_ptr(&stepping_kernel_bp); in do_breakpoint() 763 slots = this_cpu_ptr(wp_on_reg); in do_watchpoint() 829 kernel_step = this_cpu_ptr(&stepping_kernel_bp); in do_watchpoint() 850 int *kernel_step = this_cpu_ptr(&stepping_kernel_bp); in try_step_suspended_breakpoints() 950 for (slots = this_cpu_ptr(bp_on_reg), i = 0; i < core_num_brps; ++i) { in hw_breakpoint_reset() [all …]
|
| A D | paravirt.c | 83 reg = this_cpu_ptr(&stolen_time_region); in stolen_time_cpu_down_prepare() 100 reg = this_cpu_ptr(&stolen_time_region); in stolen_time_cpu_online()
|
| /arch/sh/kernel/ |
| A D | kprobes.c | 97 struct kprobe *saved = this_cpu_ptr(&saved_next_opcode); in arch_remove_kprobe() 106 saved = this_cpu_ptr(&saved_next_opcode2); in arch_remove_kprobe() 148 op1 = this_cpu_ptr(&saved_next_opcode); in prepare_singlestep() 149 op2 = this_cpu_ptr(&saved_next_opcode2); in prepare_singlestep() 321 p = this_cpu_ptr(&saved_next_opcode); in post_kprobe_handler() 333 p = this_cpu_ptr(&saved_next_opcode2); in post_kprobe_handler()
|
| A D | perf_event.c | 201 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in sh_pmu_stop() 219 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in sh_pmu_start() 236 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in sh_pmu_del() 246 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in sh_pmu_add()
|
| /arch/loongarch/kernel/ |
| A D | kfpu.c | 48 euen_curr = this_cpu_ptr(&euen_current); in kernel_fpu_begin() 75 euen_curr = this_cpu_ptr(&euen_current); in kernel_fpu_end()
|
| /arch/x86/kernel/ |
| A D | tsc_sync.c | 53 struct tsc_adjust *adj = this_cpu_ptr(&tsc_adjust); in tsc_verify_tsc_adjust() 159 struct tsc_adjust *cur = this_cpu_ptr(&tsc_adjust); in tsc_store_and_check_tsc_adjust() 183 struct tsc_adjust *ref, *cur = this_cpu_ptr(&tsc_adjust); in tsc_store_and_check_tsc_adjust() 434 struct tsc_adjust *cur = this_cpu_ptr(&tsc_adjust); in check_tsc_sync_target()
|
| /arch/x86/hyperv/ |
| A D | nested.c | 33 flush = *this_cpu_ptr(hyperv_pcpu_input_arg); in hyperv_flush_guest_mapping() 101 flush = *this_cpu_ptr(hyperv_pcpu_input_arg); in hyperv_flush_guest_mapping_range()
|
| /arch/s390/kernel/ |
| A D | idle.c | 26 struct s390_idle_data *idle = this_cpu_ptr(&s390_idle); in account_idle_time_irq() 54 struct s390_idle_data *idle = this_cpu_ptr(&s390_idle); in arch_cpu_idle()
|