current_cpu
cd->current_cpu = cpu;
cd->current_cpu = -1;
cpu = cd->current_cpu;
cd->current_cpu = cpu;
int current_cpu; /* Next CPU expected to take this irq */
int i, j, rc, cpu, current_cpu, setup_count = 0;
current_cpu = get_cpu();
if (cpu == current_cpu && idc->hart_index) {
int current_cpu = smp_processor_id();
return cpumask_test_cpu(current_cpu, c->aff_mask);
int current_cpu;
current_cpu = raw_smp_processor_id();
current_infop = lpfc_get_ctx_list(phba, current_cpu, idx);
if (idx != current_cpu)
current_cpu, idx);
mod_delayed_work_on(current_cpu(), mp->m_inodegc_wq, &gc->work,
per_cpu_ptr(xfsstats.xs_stats, current_cpu())->s.v++; \
per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->s.v++; \
per_cpu_ptr(xfsstats.xs_stats, current_cpu())->s.v--; \
per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->s.v--; \
per_cpu_ptr(xfsstats.xs_stats, current_cpu())->s.v += (inc); \
per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->s.v += (inc); \
per_cpu_ptr(xfsstats.xs_stats, current_cpu())->a[off]++; \
per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->a[off]++; \
per_cpu_ptr(xfsstats.xs_stats, current_cpu())->a[off]; \
per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->a[off]; \
per_cpu_ptr(xfsstats.xs_stats, current_cpu())->a[off] += (inc); \
per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->a[off] += (inc); \
int current_cpu = raw_smp_processor_id();
if (cpu == current_cpu)
int current_cpu = raw_smp_processor_id();
if (cpu == current_cpu)
if (cpu == current_cpu)
int current_cpu;
current_cpu = get_cpu();
cpu == current_cpu) {
int current_cpu;
current_cpu = sched_getcpu();
if (current_cpu < 0)
sprintf(pathname, "/sys/devices/system/cpu/cpu%d/topology/thread_siblings", current_cpu);
struct cpumask *tmp, int current_cpu)
if (cpu != -1 && cpu != current_cpu)