| /arch/powerpc/kernel/ |
| A D | irq.c | 75 EXPORT_PER_CPU_SYMBOL(irq_stat); 167 u64 sum = per_cpu(irq_stat, cpu).timer_irqs_event; in arch_irq_stat_cpu() 169 sum += per_cpu(irq_stat, cpu).broadcast_irqs_event; in arch_irq_stat_cpu() 170 sum += per_cpu(irq_stat, cpu).pmu_irqs; in arch_irq_stat_cpu() 171 sum += per_cpu(irq_stat, cpu).mce_exceptions; in arch_irq_stat_cpu() 172 sum += per_cpu(irq_stat, cpu).spurious_irqs; in arch_irq_stat_cpu() 173 sum += per_cpu(irq_stat, cpu).timer_irqs_others; in arch_irq_stat_cpu() 177 sum += per_cpu(irq_stat, cpu).sreset_irqs; in arch_irq_stat_cpu() 179 sum += per_cpu(irq_stat, cpu).soft_nmi_irqs; in arch_irq_stat_cpu() 182 sum += per_cpu(irq_stat, cpu).doorbell_irqs; in arch_irq_stat_cpu() [all …]
|
| A D | dbell.c | 34 __this_cpu_inc(irq_stat.doorbell_irqs); in DEFINE_INTERRUPT_HANDLER_ASYNC()
|
| A D | time.c | 575 __this_cpu_inc(irq_stat.timer_irqs_event); in DEFINE_INTERRUPT_HANDLER_ASYNC() 581 __this_cpu_inc(irq_stat.timer_irqs_others); in DEFINE_INTERRUPT_HANDLER_ASYNC() 594 __this_cpu_inc(irq_stat.broadcast_irqs_event); in timer_broadcast_interrupt()
|
| A D | watchdog.c | 381 __this_cpu_inc(irq_stat.soft_nmi_irqs); in DEFINE_INTERRUPT_HANDLER_NMI()
|
| A D | traps.c | 463 __this_cpu_inc(irq_stat.sreset_irqs); in DEFINE_INTERRUPT_HANDLER_NMI() 820 __this_cpu_inc(irq_stat.mce_exceptions); in __machine_check_exception() 1935 __this_cpu_inc(irq_stat.pmu_irqs); in DEFINE_INTERRUPT_HANDLER_NMI() 1946 __this_cpu_inc(irq_stat.pmu_irqs); in DEFINE_INTERRUPT_HANDLER_ASYNC()
|
| /arch/x86/include/asm/ |
| A D | hardirq.h | 51 DECLARE_PER_CPU_SHARED_ALIGNED(irq_cpustat_t, irq_stat); 58 #define inc_irq_stat(member) this_cpu_inc(irq_stat.member) 78 __this_cpu_write(irq_stat.kvm_cpu_l1tf_flush_l1d, 1); in kvm_set_cpu_l1tf_flush_l1d() 83 __this_cpu_write(irq_stat.kvm_cpu_l1tf_flush_l1d, 0); in kvm_clear_cpu_l1tf_flush_l1d() 88 return __this_cpu_read(irq_stat.kvm_cpu_l1tf_flush_l1d); in kvm_get_cpu_l1tf_flush_l1d()
|
| /arch/parisc/include/asm/ |
| A D | hardirq.h | 28 DECLARE_PER_CPU_SHARED_ALIGNED(irq_cpustat_t, irq_stat); 31 #define inc_irq_stat(member) this_cpu_inc(irq_stat.member) 32 #define __inc_irq_stat(member) __this_cpu_inc(irq_stat.member)
|
| /arch/s390/include/asm/ |
| A D | irq.h | 73 struct irq_stat { struct 77 DECLARE_PER_CPU_SHARED_ALIGNED(struct irq_stat, irq_stat); argument 81 __this_cpu_inc(irq_stat.irqs[irq]); in inc_irq_stat()
|
| /arch/loongarch/kernel/ |
| A D | paravirt.c | 58 irq_cpustat_t *info = &per_cpu(irq_stat, cpu); in pv_send_ipi_single() 91 info = &per_cpu(irq_stat, i); in pv_send_ipi_mask() 131 info = this_cpu_ptr(&irq_stat); in pv_ipi_interrupt() 167 r = request_percpu_irq(swi, pv_ipi_interrupt, "SWI0-IPI", &irq_stat); in pv_init_ipi()
|
| A D | irq.c | 25 DEFINE_PER_CPU_SHARED_ALIGNED(irq_cpustat_t, irq_stat); 26 EXPORT_PER_CPU_SYMBOL(irq_stat);
|
| A D | smp.c | 93 seq_put_decimal_ull_width(p, " ", per_cpu(irq_stat, cpu).ipi_irqs[i], 10); in show_ipi_list() 275 per_cpu(irq_stat, cpu).ipi_irqs[IPI_RESCHEDULE]++; in loongson_ipi_interrupt() 280 per_cpu(irq_stat, cpu).ipi_irqs[IPI_CALL_FUNCTION]++; in loongson_ipi_interrupt() 285 per_cpu(irq_stat, cpu).ipi_irqs[IPI_IRQ_WORK]++; in loongson_ipi_interrupt() 290 per_cpu(irq_stat, cpu).ipi_irqs[IPI_CLEAR_VECTOR]++; in loongson_ipi_interrupt() 305 r = request_percpu_irq(ipi_irq, loongson_ipi_interrupt, "IPI", &irq_stat); in loongson_init_ipi()
|
| /arch/loongarch/include/asm/ |
| A D | hardirq.h | 30 DECLARE_PER_CPU_SHARED_ALIGNED(irq_cpustat_t, irq_stat);
|
| /arch/powerpc/include/asm/ |
| A D | hardirq.h | 25 DECLARE_PER_CPU_SHARED_ALIGNED(irq_cpustat_t, irq_stat);
|
| /arch/s390/kernel/ |
| A D | irq.c | 38 DEFINE_PER_CPU_SHARED_ALIGNED(struct irq_stat, irq_stat); 39 EXPORT_PER_CPU_SYMBOL_GPL(irq_stat); 271 per_cpu(irq_stat, cpu).irqs[irq]); in show_interrupts()
|
| /arch/parisc/kernel/ |
| A D | irq.c | 125 DEFINE_PER_CPU_SHARED_ALIGNED(irq_cpustat_t, irq_stat); 126 #define irq_stats(x) (&per_cpu(irq_stat, x)) 406 last_usage = &per_cpu(irq_stat.irq_stack_usage, cpu); in stack_overflow_check() 422 last_usage = &per_cpu(irq_stat.kernel_stack_usage, cpu); in stack_overflow_check()
|
| A D | smp.c | 342 memset(&per_cpu(irq_stat, cpuid), 0, sizeof(irq_cpustat_t)); in smp_boot_one_cpu()
|
| /arch/x86/kernel/ |
| A D | irq.c | 33 DEFINE_PER_CPU_SHARED_ALIGNED(irq_cpustat_t, irq_stat); 34 EXPORT_PER_CPU_SYMBOL(irq_stat); 64 #define irq_stats(x) (&per_cpu(irq_stat, x))
|
| /arch/sh/kernel/ |
| A D | traps.c | 182 this_cpu_inc(irq_stat.__nmi_count); in BUILD_TRAP_HANDLER()
|
| A D | irq.c | 48 seq_put_decimal_ull_width(p, " ", per_cpu(irq_stat.__nmi_count, j), 10); in arch_show_interrupts()
|
| /arch/arm64/kernel/ |
| A D | smp.c | 1097 err = request_percpu_nmi(irq, ipi_handler, "IPI", &irq_stat); in ipi_setup_sgi() 1100 err = request_percpu_irq(irq, ipi_handler, "IPI", &irq_stat); in ipi_setup_sgi()
|
| /arch/arm/kernel/ |
| A D | smp.c | 735 "IPI", &irq_stat); in set_smp_ipi_range()
|