1 #ifndef _ASM_X86_HARDIRQ_H
2 #define _ASM_X86_HARDIRQ_H
4 #include <linux/threads.h>
8 unsigned int __softirq_pending;
9 unsigned int __nmi_count; /* arch dependent */
10 #ifdef CONFIG_X86_LOCAL_APIC
11 unsigned int apic_timer_irqs; /* arch dependent */
12 unsigned int irq_spurious_count;
13 unsigned int icr_read_retry_count;
15 unsigned int x86_platform_ipis; /* arch dependent */
16 unsigned int apic_perf_irqs;
17 unsigned int apic_irq_work_irqs;
19 unsigned int irq_resched_count;
20 unsigned int irq_call_count;
22 * irq_tlb_count is double-counted in irq_call_count, so it must be
23 * subtracted from irq_call_count when displaying irq_call_count
25 unsigned int irq_tlb_count;
27 #ifdef CONFIG_X86_THERMAL_VECTOR
28 unsigned int irq_thermal_count;
30 #ifdef CONFIG_X86_MCE_THRESHOLD
31 unsigned int irq_threshold_count;
33 } ____cacheline_aligned irq_cpustat_t;
35 DECLARE_PER_CPU_SHARED_ALIGNED(irq_cpustat_t, irq_stat);
37 /* We can have at most NR_VECTORS irqs routed to a cpu at a time */
38 #define MAX_HARDIRQS_PER_CPU NR_VECTORS
40 #define __ARCH_IRQ_STAT
42 #define inc_irq_stat(member) this_cpu_inc(irq_stat.member)
44 #define local_softirq_pending() this_cpu_read(irq_stat.__softirq_pending)
46 #define __ARCH_SET_SOFTIRQ_PENDING
48 #define set_softirq_pending(x) \
49 this_cpu_write(irq_stat.__softirq_pending, (x))
50 #define or_softirq_pending(x) this_cpu_or(irq_stat.__softirq_pending, (x))
52 extern void ack_bad_irq(unsigned int irq);
54 extern u64 arch_irq_stat_cpu(unsigned int cpu);
55 #define arch_irq_stat_cpu arch_irq_stat_cpu
57 extern u64 arch_irq_stat(void);
58 #define arch_irq_stat arch_irq_stat
60 #endif /* _ASM_X86_HARDIRQ_H */