| /linux/arch/sparc/kernel/ |
| A D | nmi.c | 107 if (__this_cpu_read(nmi_touch)) { in perfctr_irq() 111 if (!touched && __this_cpu_read(last_irq_sum) == sum) { in perfctr_irq() 113 if (__this_cpu_read(alert_counter) == 30 * nmi_hz) in perfctr_irq() 120 if (__this_cpu_read(wd_enabled)) { in perfctr_irq() 160 if (!__this_cpu_read(wd_enabled)) in stop_nmi_watchdog() 217 if (__this_cpu_read(wd_enabled)) in start_nmi_watchdog() 231 if (!__this_cpu_read(wd_enabled)) in nmi_adjust_hz_one()
|
| /linux/arch/mips/kernel/ |
| A D | mips-r2-to-r6-emul.c | 2243 (unsigned long)__this_cpu_read(mipsr2emustats.movs), in mipsr2_emul_show() 2246 (unsigned long)__this_cpu_read(mipsr2emustats.hilo), in mipsr2_emul_show() 2249 (unsigned long)__this_cpu_read(mipsr2emustats.muls), in mipsr2_emul_show() 2252 (unsigned long)__this_cpu_read(mipsr2emustats.divs), in mipsr2_emul_show() 2255 (unsigned long)__this_cpu_read(mipsr2emustats.dsps), in mipsr2_emul_show() 2258 (unsigned long)__this_cpu_read(mipsr2emustats.bops), in mipsr2_emul_show() 2261 (unsigned long)__this_cpu_read(mipsr2emustats.traps), in mipsr2_emul_show() 2264 (unsigned long)__this_cpu_read(mipsr2emustats.fpus), in mipsr2_emul_show() 2267 (unsigned long)__this_cpu_read(mipsr2emustats.loads), in mipsr2_emul_show() 2270 (unsigned long)__this_cpu_read(mipsr2emustats.stores), in mipsr2_emul_show() [all …]
|
| /linux/kernel/ |
| A D | watchdog.c | 390 u8 tail = __this_cpu_read(cpustat_tail); in update_cpustat() 396 old_stat = __this_cpu_read(cpustat_old[i]); in update_cpustat() 409 u8 tail = __this_cpu_read(cpustat_tail); in print_cpustat() 426 __this_cpu_read(cpustat_util[group][STATS_SYSTEM]), in print_cpustat() 427 __this_cpu_read(cpustat_util[group][STATS_SOFTIRQ]), in print_cpustat() 428 __this_cpu_read(cpustat_util[group][STATS_HARDIRQ]), in print_cpustat() 429 __this_cpu_read(cpustat_util[group][STATS_IDLE])); in print_cpustat() 462 int tail = __this_cpu_read(cpustat_tail); in need_counting_irqs() 471 if (!__this_cpu_read(snapshot_taken)) { in start_counting_irqs() 489 if (__this_cpu_read(snapshot_taken)) { in print_irq_counts() [all …]
|
| A D | softirq.c | 78 struct task_struct *tsk = __this_cpu_read(ksoftirqd); in wakeup_softirqd() 141 return __this_cpu_read(softirq_ctrl.cnt) != 0; in local_bh_blocked() 215 curcnt = __this_cpu_read(softirq_ctrl.cnt); in __local_bh_enable_ip() 421 if (!force_irqthreads() || !__this_cpu_read(ksoftirqd)) { in invoke_softirq() 945 *__this_cpu_read(tasklet_vec.tail) = per_cpu(tasklet_vec, cpu).head; in takeover_tasklets() 953 *__this_cpu_read(tasklet_hi_vec.tail) = per_cpu(tasklet_hi_vec, cpu).head; in takeover_tasklets()
|
| /linux/kernel/time/ |
| A D | tick-oneshot.c | 25 struct clock_event_device *dev = __this_cpu_read(tick_cpu_device.evtdev); in tick_program_event() 52 struct clock_event_device *dev = __this_cpu_read(tick_cpu_device.evtdev); in tick_resume_oneshot() 112 ret = __this_cpu_read(tick_cpu_device.mode) == TICKDEV_MODE_ONESHOT; in tick_oneshot_mode_active()
|
| /linux/include/linux/ |
| A D | context_tracking_state.h | 80 return __this_cpu_read(context_tracking.nesting); in ct_nesting() 92 return __this_cpu_read(context_tracking.nmi_nesting); in ct_nmi_nesting() 118 return context_tracking_enabled() && __this_cpu_read(context_tracking.active); in context_tracking_enabled_this_cpu()
|
| /linux/include/asm-generic/ |
| A D | irq_regs.h | 21 return __this_cpu_read(__irq_regs); in get_irq_regs() 28 old_regs = __this_cpu_read(__irq_regs); in set_irq_regs()
|
| /linux/lib/ |
| A D | percpu_test.c | 11 WARN(__this_cpu_read(pcp) != (expected), \ 13 __this_cpu_read(pcp), __this_cpu_read(pcp), \
|
| A D | percpu_counter.c | 106 count = __this_cpu_read(*fbc->counters); in percpu_counter_add_batch() 126 count = __this_cpu_read(*fbc->counters) + amount; in percpu_counter_add_batch() 152 count = __this_cpu_read(*fbc->counters); in percpu_counter_sync() 340 count = __this_cpu_read(*fbc->counters); in __percpu_counter_limited_add() 385 count = __this_cpu_read(*fbc->counters); in __percpu_counter_limited_add()
|
| A D | cpumask.c | 175 prev = __this_cpu_read(distribute_cpu_mask_prev); in cpumask_any_and_distribute() 197 prev = __this_cpu_read(distribute_cpu_mask_prev); in cpumask_any_distribute()
|
| /linux/arch/powerpc/lib/ |
| A D | code-patching.c | 292 patching_mm = __this_cpu_read(cpu_patching_context.mm); in __do_patch_mem_mm() 293 text_poke_addr = __this_cpu_read(cpu_patching_context.addr); in __do_patch_mem_mm() 335 text_poke_addr = (unsigned long)__this_cpu_read(cpu_patching_context.addr) & PAGE_MASK; in __do_patch_mem() 338 pte = __this_cpu_read(cpu_patching_context.pte); in __do_patch_mem() 478 patching_mm = __this_cpu_read(cpu_patching_context.mm); in __do_patch_instructions_mm() 479 text_poke_addr = __this_cpu_read(cpu_patching_context.addr); in __do_patch_instructions_mm() 525 text_poke_addr = (unsigned long)__this_cpu_read(cpu_patching_context.addr) & PAGE_MASK; in __do_patch_instructions() 528 pte = __this_cpu_read(cpu_patching_context.pte); in __do_patch_instructions()
|
| /linux/arch/x86/kernel/ |
| A D | hw_breakpoint.c | 485 set_debugreg(__this_cpu_read(cpu_debugreg[0]), 0); in hw_breakpoint_restore() 486 set_debugreg(__this_cpu_read(cpu_debugreg[1]), 1); in hw_breakpoint_restore() 487 set_debugreg(__this_cpu_read(cpu_debugreg[2]), 2); in hw_breakpoint_restore() 488 set_debugreg(__this_cpu_read(cpu_debugreg[3]), 3); in hw_breakpoint_restore() 490 set_debugreg(__this_cpu_read(cpu_dr7), 7); in hw_breakpoint_restore()
|
| A D | irq_32.c | 77 irqstk = __this_cpu_read(pcpu_hot.hardirq_stack_ptr); in execute_on_irq_stack() 138 irqstk = __this_cpu_read(pcpu_hot.softirq_stack_ptr); in do_softirq_own_stack()
|
| A D | irq.c | 257 desc = __this_cpu_read(vector_irq[vector]); in call_irq_handler() 511 if (IS_ERR_OR_NULL(__this_cpu_read(vector_irq[vector]))) in fixup_irqs() 515 desc = __this_cpu_read(vector_irq[vector]); in fixup_irqs() 526 if (__this_cpu_read(vector_irq[vector]) != VECTOR_RETRIGGERED) in fixup_irqs()
|
| /linux/drivers/cpuidle/ |
| A D | cpuidle-psci.c | 50 return __this_cpu_read(domain_state); in psci_get_domain_state() 106 struct device *pd_dev = __this_cpu_read(psci_cpuidle_data.dev); in psci_idle_cpuhp_up() 116 struct device *pd_dev = __this_cpu_read(psci_cpuidle_data.dev); in psci_idle_cpuhp_down() 194 u32 *state = __this_cpu_read(psci_cpuidle_data.psci_states); in psci_enter_idle_state()
|
| /linux/drivers/xen/events/ |
| A D | events_2l.c | 123 struct vcpu_info *vcpu_info = __this_cpu_read(xen_vcpu); in evtchn_2l_unmask() 173 struct vcpu_info *vcpu_info = __this_cpu_read(xen_vcpu); in evtchn_2l_handle_events() 192 start_word_idx = __this_cpu_read(current_word_idx); in evtchn_2l_handle_events() 193 start_bit_idx = __this_cpu_read(current_bit_idx); in evtchn_2l_handle_events()
|
| /linux/arch/x86/include/asm/ |
| A D | cpu_entry_area.h | 148 CEA_ESTACK_TOP(__this_cpu_read(cea_exception_stacks), name) 151 CEA_ESTACK_BOT(__this_cpu_read(cea_exception_stacks), name)
|
| A D | irq_stack.h | 119 call_on_stack(__this_cpu_read(pcpu_hot.hardirq_stack_ptr), \ 138 if (user_mode(regs) || __this_cpu_read(pcpu_hot.hardirq_stack_inuse)) { \
|
| /linux/arch/powerpc/kernel/ |
| A D | hw_breakpoint.c | 236 struct perf_event *bp = __this_cpu_read(bp_per_reg[i]); in thread_change_pc() 246 info = counter_arch_bp(__this_cpu_read(bp_per_reg[i])); in thread_change_pc() 408 bp[i] = __this_cpu_read(bp_per_reg[i]); in hw_breakpoint_handler() 519 bp = __this_cpu_read(bp_per_reg[i]); in single_step_dabr_instruction()
|
| /linux/kernel/rcu/ |
| A D | tree_plugin.h | 300 if (__this_cpu_read(rcu_data.cpu_no_qs.b.norm)) { in rcu_qs() 302 __this_cpu_read(rcu_data.gp_seq), in rcu_qs() 599 return (__this_cpu_read(rcu_data.cpu_no_qs.b.exp) || in rcu_preempt_need_deferred_qs() 752 __this_cpu_read(rcu_data.core_needs_qs) && in rcu_flavor_sched_clock_irq() 753 __this_cpu_read(rcu_data.cpu_no_qs.b.norm) && in rcu_flavor_sched_clock_irq() 862 if (!__this_cpu_read(rcu_data.cpu_no_qs.s)) in rcu_qs() 865 __this_cpu_read(rcu_data.gp_seq), TPS("cpuqs")); in rcu_qs() 867 if (__this_cpu_read(rcu_data.cpu_no_qs.b.exp)) in rcu_qs()
|
| /linux/arch/arm64/include/asm/ |
| A D | arch_timer.h | 27 __wa = __this_cpu_read(timer_unstable_counter_workaround); \ 34 __wa = __this_cpu_read(timer_unstable_counter_workaround); \
|
| /linux/arch/x86/xen/ |
| A D | time.c | 53 src = &__this_cpu_read(xen_vcpu)->time; in xen_clocksource_read() 69 src = &__this_cpu_read(xen_vcpu)->time; in xen_sched_clock() 547 pvti = &__this_cpu_read(xen_vcpu)->time; in xen_time_init() 628 if (!__this_cpu_read(xen_vcpu)) { in xen_hvm_init_time_ops()
|
| /linux/drivers/irqchip/ |
| A D | irq-xtensa-mx.c | 83 mask = __this_cpu_read(cached_irq_mask) & ~mask; in xtensa_mx_irq_mask() 101 mask |= __this_cpu_read(cached_irq_mask); in xtensa_mx_irq_unmask()
|
| /linux/mm/ |
| A D | vmstat.c | 359 x = delta + __this_cpu_read(*p); in __mod_zone_page_state() 361 t = __this_cpu_read(pcp->stat_threshold); in __mod_zone_page_state() 395 x = delta + __this_cpu_read(*p); in __mod_node_page_state() 397 t = __this_cpu_read(pcp->stat_threshold); in __mod_node_page_state() 442 t = __this_cpu_read(pcp->stat_threshold); in __inc_zone_state() 465 t = __this_cpu_read(pcp->stat_threshold); in __inc_node_state() 498 t = __this_cpu_read(pcp->stat_threshold); in __dec_zone_state() 521 t = __this_cpu_read(pcp->stat_threshold); in __dec_node_state() 848 if (!__this_cpu_read(pcp->expire) || in refresh_cpu_vm_stats() 849 !__this_cpu_read(pcp->count)) in refresh_cpu_vm_stats() [all …]
|
| /linux/drivers/gpu/drm/amd/display/amdgpu_dm/ |
| A D | dc_fpu.c | 56 depth = __this_cpu_read(fpu_recursion_depth); in dc_assert_fp_enabled()
|