1 /**************************************************************************** 2 **************************************************************************** 3 *** 4 *** This header was automatically generated from a Linux kernel header 5 *** of the same name, to make information necessary for userspace to 6 *** call into the kernel available to libc. It contains only constants, 7 *** structures, and macros generated from the original header, and thus, 8 *** contains no copyrightable information. 9 *** 10 **************************************************************************** 11 ****************************************************************************/ 12 #ifndef LINUX_HARDIRQ_H 13 #define LINUX_HARDIRQ_H 14 15 #include <linux/preempt.h> 16 #include <linux/smp_lock.h> 17 #include <linux/lockdep.h> 18 #include <asm/hardirq.h> 19 #include <asm/system.h> 20 21 #define PREEMPT_BITS 8 22 #define SOFTIRQ_BITS 8 23 24 #ifndef HARDIRQ_BITS 25 #define HARDIRQ_BITS 12 26 27 #if 1 << HARDIRQ_BITS < NR_IRQS 28 #error HARDIRQ_BITS is too low! 29 #endif 30 #endif 31 32 #define PREEMPT_SHIFT 0 33 #define SOFTIRQ_SHIFT (PREEMPT_SHIFT + PREEMPT_BITS) 34 #define HARDIRQ_SHIFT (SOFTIRQ_SHIFT + SOFTIRQ_BITS) 35 36 #define __IRQ_MASK(x) ((1UL << (x))-1) 37 38 #define PREEMPT_MASK (__IRQ_MASK(PREEMPT_BITS) << PREEMPT_SHIFT) 39 #define SOFTIRQ_MASK (__IRQ_MASK(SOFTIRQ_BITS) << SOFTIRQ_SHIFT) 40 #define HARDIRQ_MASK (__IRQ_MASK(HARDIRQ_BITS) << HARDIRQ_SHIFT) 41 42 #define PREEMPT_OFFSET (1UL << PREEMPT_SHIFT) 43 #define SOFTIRQ_OFFSET (1UL << SOFTIRQ_SHIFT) 44 #define HARDIRQ_OFFSET (1UL << HARDIRQ_SHIFT) 45 46 #if PREEMPT_ACTIVE < 1 << HARDIRQ_SHIFT + HARDIRQ_BITS 47 #error PREEMPT_ACTIVE is too low! 48 #endif 49 50 #define hardirq_count() (preempt_count() & HARDIRQ_MASK) 51 #define softirq_count() (preempt_count() & SOFTIRQ_MASK) 52 #define irq_count() (preempt_count() & (HARDIRQ_MASK | SOFTIRQ_MASK)) 53 54 #define in_irq() (hardirq_count()) 55 #define in_softirq() (softirq_count()) 56 #define in_interrupt() (irq_count()) 57 58 #define in_atomic() ((preempt_count() & ~PREEMPT_ACTIVE) != 0) 59 60 #define preemptible() 0 61 #define IRQ_EXIT_OFFSET HARDIRQ_OFFSET 62 63 #define synchronize_irq(irq) barrier() 64 65 struct task_struct; 66 67 #define irq_enter() do { account_system_vtime(current); add_preempt_count(HARDIRQ_OFFSET); trace_hardirq_enter(); } while (0) 68 #define __irq_exit() do { trace_hardirq_exit(); account_system_vtime(current); sub_preempt_count(HARDIRQ_OFFSET); } while (0) 69 70 #define nmi_enter() do { lockdep_off(); irq_enter(); } while (0) 71 #define nmi_exit() do { __irq_exit(); lockdep_on(); } while (0) 72 73 #endif 74