early_cpu_to_node
scs_alloc(early_cpu_to_node(cpu));
p = arch_alloc_vmap_stack(IRQ_STACK_SIZE, early_cpu_to_node(cpu));
extern int early_cpu_to_node(int cpu);
nid = early_cpu_to_node(cpu);
return early_cpu_to_node(cpu);
if (early_cpu_to_node(from) == early_cpu_to_node(to))
node = early_cpu_to_node(cpu);
limit, early_cpu_to_node(i));
nid = early_cpu_to_node(cpu);
early_cpu_to_node(cpu));
if (early_cpu_to_node(from) == early_cpu_to_node(to))
return early_cpu_to_node(cpu);
struct xive_ipi_desc *xid = &xive_ipis[early_cpu_to_node(cpu)];
return xive_ipis[early_cpu_to_node(cpu)].irq;
extern int early_cpu_to_node(int cpu);
unsigned long cpudata = vdso_encode_cpunode(cpu, early_cpu_to_node(cpu));
early_cpu_to_node(cpu) != NUMA_NO_NODE)
set_numa_node(early_cpu_to_node(cpu));
set_cpu_numa_node(cpu, early_cpu_to_node(cpu));
int node = early_cpu_to_node(cpu);
if (early_cpu_to_node(from) == early_cpu_to_node(to))
return early_cpu_to_node(cpu);
int node = early_cpu_to_node(cpu);
early_cpu_to_node(cpu));
if (early_cpu_to_node(i) != NUMA_NO_NODE)
int nid = early_cpu_to_node(i);
cpumask_set_cpu(cpu, node_to_cpumask_map[early_cpu_to_node(cpu)]);
cpumask_clear_cpu(cpu, node_to_cpumask_map[early_cpu_to_node(cpu)]);
debug_cpumask_set_cpu(cpu, early_cpu_to_node(cpu), enable);
return node_distance(early_cpu_to_node(from), early_cpu_to_node(to));
early_cpu_to_node);
rc = pcpu_page_first_chunk(PERCPU_MODULE_RESERVE, early_cpu_to_node);
node = early_cpu_to_node(eiointc_priv[nr_pics - 1]->node * CORES_PER_EIO_NODE);
node = early_cpu_to_node(acpi_eiointc->node * CORES_PER_EIO_NODE);
int early_cpu_to_node(int cpu);
set_cpu_numa_node(cpu, early_cpu_to_node(cpu));
nid = early_cpu_to_node(cpu);
nid = early_cpu_to_node(cpu);