mm/swap.c
190
local_lock_irqsave(&cpu_fbatches.lock_irq, flags);
mm/swap.c
192
local_lock(&cpu_fbatches.lock);
mm/swap.c
199
local_unlock_irqrestore(&cpu_fbatches.lock_irq, flags);
mm/swap.c
201
local_unlock(&cpu_fbatches.lock);
mm/swap.c
206
&cpu_fbatches.op, \
mm/swap.c
209
offsetof(struct cpu_fbatches, op) >= \
mm/swap.c
210
offsetof(struct cpu_fbatches, lock_irq) \
mm/swap.c
323
struct folio_batch *fbatch = &per_cpu(cpu_fbatches.lru_activate, cpu);
mm/swap.c
362
local_lock(&cpu_fbatches.lock);
mm/swap.c
363
fbatch = this_cpu_ptr(&cpu_fbatches.lru_add);
mm/swap.c
384
local_unlock(&cpu_fbatches.lock);
mm/swap.c
644
struct cpu_fbatches *fbatches = &per_cpu(cpu_fbatches, cpu);
mm/swap.c
656
local_lock_irqsave(&cpu_fbatches.lock_irq, flags);
mm/swap.c
658
local_unlock_irqrestore(&cpu_fbatches.lock_irq, flags);
mm/swap.c
68
static DEFINE_PER_CPU(struct cpu_fbatches, cpu_fbatches) = {
mm/swap.c
736
local_lock(&cpu_fbatches.lock);
mm/swap.c
738
local_unlock(&cpu_fbatches.lock);
mm/swap.c
750
local_lock(&cpu_fbatches.lock);
mm/swap.c
752
local_unlock(&cpu_fbatches.lock);
mm/swap.c
759
local_lock(&cpu_fbatches.lock);
mm/swap.c
762
local_unlock(&cpu_fbatches.lock);
mm/swap.c
777
struct cpu_fbatches *fbatches = &per_cpu(cpu_fbatches, cpu);