irq_stack
DECLARE_PER_CPU(unsigned long, irq_stack);
unsigned long low = per_cpu(irq_stack, cpu);
per_cpu(irq_stack, i) = (unsigned long)page_address(page);
per_cpu(irq_stack, i), per_cpu(irq_stack, i) + IRQ_STACK_SIZE);
DEFINE_PER_CPU(unsigned long, irq_stack);
unsigned long begin = (unsigned long)this_cpu_read(irq_stack);
stack = per_cpu(irq_stack, cpu) + IRQ_STACK_START;
extern void *irq_stack[NR_CPUS];
unsigned long low = (unsigned long)irq_stack[cpu];
void *irq_stack[NR_CPUS];
irq_stack[i] = s;
irq_stack[i], irq_stack[i] + IRQ_STACK_SIZE);
stack_page = (unsigned long)irq_stack[cpu];
unsigned long irq_stack;
irq_stack = (unsigned long) &union_ptr->stack;
irq_stack = ALIGN(irq_stack + sizeof(irq_stack_union.slock),
call_on_stack(param1, func, irq_stack);
DEFINE_PER_CPU_ALIGNED(ulong [IRQ_STACK_SIZE/sizeof(ulong)], irq_stack);
per_cpu(irq_stack_ptr, cpu) = per_cpu(irq_stack, cpu);
DECLARE_PER_CPU_CACHE_HOT(struct irq_stack *, hardirq_stack_ptr);
DECLARE_PER_CPU_CACHE_HOT(struct irq_stack *, softirq_stack_ptr);
DEFINE_PER_CPU_CACHE_HOT(struct irq_stack *, hardirq_stack_ptr);
struct irq_stack *irqstk;
DEFINE_PER_CPU_CACHE_HOT(struct irq_stack *, softirq_stack_ptr);
struct irq_stack *curstk, *irqstk;
curstk = (struct irq_stack *) current_stack();
DEFINE_PER_CPU_PAGE_ALIGNED(struct irq_stack, irq_stack_backing_store) __visible;