for_each_cpu_wrap
for_each_cpu_wrap(ref_cpu, policy->cpus, cpu + 1) {
for_each_cpu_wrap(cpu, &xmon_batch_cpus, xmon_batch_start_cpu) {
for_each_cpu_wrap(cpu, cpu_online_mask, start) {
for_each_cpu_wrap(tgt_cpu, &stor_device->alloced_cpus,
for_each_cpu_wrap(tgt_cpu, &stor_device->alloced_cpus,
for_each_cpu_wrap((cpu), cpu_possible_mask, (start))
for_each_cpu_wrap((cpu), cpu_online_mask, (start))
for_each_cpu_wrap(cpu, cpu_possible_mask, raw_smp_processor_id()) {
for_each_cpu_wrap(cpu, cpu_possible_mask, raw_smp_processor_id()) {
for_each_cpu_wrap(i, smt_mask, cpu) {
for_each_cpu_wrap(i, sched_domain_span(sd), cpu + 1) {
for_each_cpu_wrap(balance_cpu, nohz.idle_cpus_mask, this_cpu+1) {
for_each_cpu_wrap(cpu, cpumask_of_node(env->dst_nid), start + 1) {
for_each_cpu_wrap(cpu, sched_group_span(sg), target + 1) {
for_each_cpu_wrap(cpu, cpus, target + 1) {
for_each_cpu_wrap(cpu, cpus, target) {
for_each_cpu_wrap(i, span, cpu) {
for_each_cpu_wrap(i, span, cpu) {
for_each_cpu_wrap(cpu, m, nr_cpu_ids / 2) \