Symbol: nodemask_t
arch/loongarch/include/asm/numa.h
23
extern nodemask_t numa_nodes_parsed __initdata;
arch/mips/sgi-ip27/ip27-klnuma.c
23
static nodemask_t ktext_repmask;
arch/x86/include/asm/numa.h
24
extern nodemask_t numa_nodes_parsed __initdata;
arch/x86/include/asm/numa.h
25
extern nodemask_t numa_phys_nodes_parsed __initdata;
arch/x86/kernel/cpu/sgx/main.c
41
static nodemask_t sgx_numa_mask;
arch/x86/mm/numa.c
51
nodemask_t numa_phys_nodes_parsed __initdata;
drivers/acpi/numa/srat.c
24
static nodemask_t nodes_found_map = NODE_MASK_NONE;
drivers/acpi/numa/srat.c
99
nodemask_t nodes_to_enable;
drivers/cxl/core/region.c
39
static nodemask_t nodemask_region_seen = NODE_MASK_NONE;
drivers/irqchip/irq-loongson-eiointc.c
106
static void eiointc_set_irq_route(int pos, unsigned int cpu, unsigned int mnode, nodemask_t *node_map)
drivers/irqchip/irq-loongson-eiointc.c
75
nodemask_t node_map;
drivers/net/ethernet/hisilicon/hns3/hnae3.h
926
nodemask_t numa_node_mask; /* for multi-chip support */
drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.h
899
nodemask_t numa_node_mask;
drivers/net/ethernet/hisilicon/hns3/hns3vf/hclgevf_main.h
239
nodemask_t numa_node_mask;
drivers/nvme/host/multipath.c
1105
nodemask_t numa_nodes;
include/asm-generic/numa.h
12
extern nodemask_t numa_nodes_parsed __initdata;
include/linux/cpuset.h
166
static inline void set_mems_allowed(nodemask_t nodemask)
include/linux/cpuset.h
179
extern void cpuset_nodes_allowed(struct cgroup *cgroup, nodemask_t *mask);
include/linux/cpuset.h
219
static inline nodemask_t cpuset_mems_allowed(struct task_struct *p)
include/linux/cpuset.h
227
static inline int cpuset_nodemask_valid_mems_allowed(nodemask_t *nodemask)
include/linux/cpuset.h
288
static inline void set_mems_allowed(nodemask_t nodemask)
include/linux/cpuset.h
302
static inline void cpuset_nodes_allowed(struct cgroup *cgroup, nodemask_t *mask)
include/linux/cpuset.h
83
extern nodemask_t cpuset_mems_allowed(struct task_struct *p);
include/linux/cpuset.h
86
int cpuset_nodemask_valid_mems_allowed(nodemask_t *nodemask);
include/linux/gfp.h
230
nodemask_t *nodemask);
include/linux/gfp.h
234
nodemask_t *nodemask);
include/linux/gfp.h
238
nodemask_t *nodemask, int nr_pages,
include/linux/gfp.h
457
gfp_t gfp_mask, int nid, nodemask_t *nodemask);
include/linux/gfp.h
462
int nid, nodemask_t *nodemask);
include/linux/hugetlb.h
1127
nodemask_t *nmask, gfp_t gfp_mask)
include/linux/hugetlb.h
1134
nodemask_t *nmask, gfp_t gfp_mask,
include/linux/hugetlb.h
177
extern nodemask_t hugetlb_bootmem_nodes;
include/linux/hugetlb.h
710
nodemask_t *nmask, gfp_t gfp_mask,
include/linux/hugetlb.h
713
nodemask_t *nmask, gfp_t gfp_mask);
include/linux/memcontrol.h
1761
void mem_cgroup_node_filter_allowed(struct mem_cgroup *memcg, nodemask_t *mask);
include/linux/memcontrol.h
1833
nodemask_t *mask)
include/linux/memory-tiers.h
104
static inline int next_demotion_node(int node, const nodemask_t *allowed_mask)
include/linux/memory-tiers.h
109
static inline void node_get_allowed_targets(pg_data_t *pgdat, nodemask_t *targets)
include/linux/memory-tiers.h
32
nodemask_t nodes;
include/linux/memory-tiers.h
41
extern nodemask_t default_dram_nodes;
include/linux/memory-tiers.h
56
int next_demotion_node(int node, const nodemask_t *allowed_mask);
include/linux/memory-tiers.h
57
void node_get_allowed_targets(pg_data_t *pgdat, nodemask_t *targets);
include/linux/memory-tiers.h
60
static inline int next_demotion_node(int node, const nodemask_t *allowed_mask)
include/linux/memory-tiers.h
65
static inline void node_get_allowed_targets(pg_data_t *pgdat, nodemask_t *targets)
include/linux/mempolicy.h
140
extern void mpol_rebind_task(struct task_struct *tsk, const nodemask_t *new);
include/linux/mempolicy.h
141
extern void mpol_rebind_mm(struct mm_struct *mm, nodemask_t *new);
include/linux/mempolicy.h
145
struct mempolicy **mpol, nodemask_t **nodemask);
include/linux/mempolicy.h
146
extern bool init_nodemask_of_mempolicy(nodemask_t *mask);
include/linux/mempolicy.h
148
const nodemask_t *mask);
include/linux/mempolicy.h
159
int do_migrate_pages(struct mm_struct *mm, const nodemask_t *from,
include/linux/mempolicy.h
160
const nodemask_t *to, int flags);
include/linux/mempolicy.h
251
const nodemask_t *new)
include/linux/mempolicy.h
255
static inline void mpol_rebind_mm(struct mm_struct *mm, nodemask_t *new)
include/linux/mempolicy.h
261
struct mempolicy **mpol, nodemask_t **nodemask)
include/linux/mempolicy.h
268
static inline bool init_nodemask_of_mempolicy(nodemask_t *m)
include/linux/mempolicy.h
273
static inline int do_migrate_pages(struct mm_struct *mm, const nodemask_t *from,
include/linux/mempolicy.h
274
const nodemask_t *to, int flags)
include/linux/mempolicy.h
51
nodemask_t nodes; /* interleave/bind/preferred/etc */
include/linux/mempolicy.h
55
nodemask_t cpuset_mems_allowed; /* relative to these nodes */
include/linux/mempolicy.h
56
nodemask_t user_nodemask; /* nodemask passed by user */
include/linux/mm.h
3767
extern void __show_mem(unsigned int flags, nodemask_t *nodemask, int max_zone_idx);
include/linux/mm.h
3777
void warn_alloc(gfp_t gfp_mask, nodemask_t *nodemask, const char *fmt, ...);
include/linux/mmzone.h
1738
nodemask_t *nodes);
include/linux/mmzone.h
1757
nodemask_t *nodes)
include/linux/mmzone.h
1783
nodemask_t *nodes)
include/linux/mmzone.h
1826
static inline bool movable_only_nodes(nodemask_t *nodes)
include/linux/nodemask.h
106
static __always_inline unsigned int __nodemask_pr_numnodes(const nodemask_t *m)
include/linux/nodemask.h
110
static __always_inline const unsigned long *__nodemask_pr_bits(const nodemask_t *m)
include/linux/nodemask.h
125
static __always_inline void __node_set(int node, volatile nodemask_t *dstp)
include/linux/nodemask.h
131
static __always_inline void __node_clear(int node, volatile nodemask_t *dstp)
include/linux/nodemask.h
137
static __always_inline void __nodes_setall(nodemask_t *dstp, unsigned int nbits)
include/linux/nodemask.h
143
static __always_inline void __nodes_clear(nodemask_t *dstp, unsigned int nbits)
include/linux/nodemask.h
153
static __always_inline bool __node_test_and_set(int node, nodemask_t *addr)
include/linux/nodemask.h
160
static __always_inline bool __nodes_and(nodemask_t *dstp, const nodemask_t *src1p,
include/linux/nodemask.h
161
const nodemask_t *src2p, unsigned int nbits)
include/linux/nodemask.h
168
static __always_inline void __nodes_or(nodemask_t *dstp, const nodemask_t *src1p,
include/linux/nodemask.h
169
const nodemask_t *src2p, unsigned int nbits)
include/linux/nodemask.h
176
static __always_inline void __nodes_xor(nodemask_t *dstp, const nodemask_t *src1p,
include/linux/nodemask.h
177
const nodemask_t *src2p, unsigned int nbits)
include/linux/nodemask.h
184
static __always_inline bool __nodes_andnot(nodemask_t *dstp, const nodemask_t *src1p,
include/linux/nodemask.h
185
const nodemask_t *src2p, unsigned int nbits)
include/linux/nodemask.h
191
static __always_inline void __nodes_copy(nodemask_t *dstp,
include/linux/nodemask.h
192
const nodemask_t *srcp, unsigned int nbits)
include/linux/nodemask.h
199
static __always_inline void __nodes_complement(nodemask_t *dstp,
include/linux/nodemask.h
200
const nodemask_t *srcp, unsigned int nbits)
include/linux/nodemask.h
207
static __always_inline bool __nodes_equal(const nodemask_t *src1p,
include/linux/nodemask.h
208
const nodemask_t *src2p, unsigned int nbits)
include/linux/nodemask.h
215
static __always_inline bool __nodes_intersects(const nodemask_t *src1p,
include/linux/nodemask.h
216
const nodemask_t *src2p, unsigned int nbits)
include/linux/nodemask.h
223
static __always_inline bool __nodes_subset(const nodemask_t *src1p,
include/linux/nodemask.h
224
const nodemask_t *src2p, unsigned int nbits)
include/linux/nodemask.h
230
static __always_inline bool __nodes_empty(const nodemask_t *srcp, unsigned int nbits)
include/linux/nodemask.h
236
static __always_inline bool __nodes_full(const nodemask_t *srcp, unsigned int nbits)
include/linux/nodemask.h
242
static __always_inline int __nodes_weight(const nodemask_t *srcp, unsigned int nbits)
include/linux/nodemask.h
251
static __always_inline unsigned int __first_node(const nodemask_t *srcp)
include/linux/nodemask.h
257
static __always_inline unsigned int __next_node(int n, const nodemask_t *srcp)
include/linux/nodemask.h
267
static __always_inline unsigned int __next_node_in(int node, const nodemask_t *srcp)
include/linux/nodemask.h
276
static __always_inline void init_nodemask_of_node(nodemask_t *mask, int node)
include/linux/nodemask.h
294
static __always_inline unsigned int __first_unset_node(const nodemask_t *maskp)
include/linux/nodemask.h
304
((nodemask_t) { { \
include/linux/nodemask.h
311
((nodemask_t) { { \
include/linux/nodemask.h
319
((nodemask_t) { { \
include/linux/nodemask.h
328
nodemask_t *dstp, int nbits)
include/linux/nodemask.h
334
static __always_inline int __nodelist_parse(const char *buf, nodemask_t *dstp, int nbits)
include/linux/nodemask.h
342
const nodemask_t *oldp, const nodemask_t *newp, int nbits)
include/linux/nodemask.h
349
static __always_inline void __nodes_remap(nodemask_t *dstp, const nodemask_t *srcp,
include/linux/nodemask.h
350
const nodemask_t *oldp, const nodemask_t *newp, int nbits)
include/linux/nodemask.h
357
static __always_inline void __nodes_onto(nodemask_t *dstp, const nodemask_t *origp,
include/linux/nodemask.h
358
const nodemask_t *relmapp, int nbits)
include/linux/nodemask.h
365
static __always_inline void __nodes_fold(nodemask_t *dstp, const nodemask_t *origp,
include/linux/nodemask.h
404
extern nodemask_t node_states[NR_NODE_STATES];
include/linux/nodemask.h
491
static __always_inline int node_random(const nodemask_t *maskp)
include/linux/nodemask.h
530
nodemask_t mask1;
include/linux/nodemask.h
531
nodemask_t mask2;
include/linux/nodemask.h
96
extern nodemask_t _unused_nodemask_arg_;
include/linux/numa.h
34
int nearest_node_nodemask(int node, nodemask_t *mask);
include/linux/numa.h
52
static inline int nearest_node_nodemask(int node, nodemask_t *mask)
include/linux/oom.h
33
nodemask_t *nodemask;
include/linux/sched.h
1314
nodemask_t mems_allowed;
include/linux/swap.h
373
gfp_t gfp_mask, nodemask_t *mask);
include/trace/events/sched.h
777
TP_PROTO(struct task_struct *tsk, nodemask_t *mem_allowed_ptr),
include/trace/events/sched.h
794
BUILD_BUG_ON(sizeof(nodemask_t) != \
kernel/cgroup/cpuset-internal.h
102
nodemask_t mems_allowed;
kernel/cgroup/cpuset-internal.h
106
nodemask_t effective_mems;
kernel/cgroup/cpuset-internal.h
146
nodemask_t old_mems_allowed;
kernel/cgroup/cpuset-internal.h
321
struct cpumask *new_cpus, nodemask_t *new_mems,
kernel/cgroup/cpuset-internal.h
335
struct cpumask *new_cpus, nodemask_t *new_mems,
kernel/cgroup/cpuset-v1.c
286
struct cpumask *new_cpus, nodemask_t *new_mems,
kernel/cgroup/cpuset.c
2520
nodemask_t from;
kernel/cgroup/cpuset.c
2521
nodemask_t to;
kernel/cgroup/cpuset.c
2535
static void cpuset_migrate_mm(struct mm_struct *mm, const nodemask_t *from,
kernel/cgroup/cpuset.c
2536
const nodemask_t *to)
kernel/cgroup/cpuset.c
2588
nodemask_t *newmems)
kernel/cgroup/cpuset.c
2617
static nodemask_t newmems; /* protected by cpuset_mutex */
kernel/cgroup/cpuset.c
2678
static void update_nodemasks_hier(struct cpuset *cs, nodemask_t *new_mems)
kernel/cgroup/cpuset.c
3099
static nodemask_t cpuset_attach_nodemask_to;
kernel/cgroup/cpuset.c
358
static inline void check_insane_mems_config(nodemask_t *nodes)
kernel/cgroup/cpuset.c
3704
struct cpumask *new_cpus, nodemask_t *new_mems,
kernel/cgroup/cpuset.c
3741
static nodemask_t new_mems;
kernel/cgroup/cpuset.c
3856
static nodemask_t new_mems;
kernel/cgroup/cpuset.c
4129
nodemask_t cpuset_mems_allowed(struct task_struct *tsk)
kernel/cgroup/cpuset.c
4131
nodemask_t mask;
kernel/cgroup/cpuset.c
4147
int cpuset_nodemask_valid_mems_allowed(nodemask_t *nodemask)
kernel/cgroup/cpuset.c
4255
void cpuset_nodes_allowed(struct cgroup *cgroup, nodemask_t *mask)
kernel/cgroup/cpuset.c
498
static void guarantee_online_mems(struct cpuset *cs, nodemask_t *pmask)
kernel/sched/ext_idle.c
145
static DEFINE_PER_CPU(nodemask_t, per_cpu_unvisited);
kernel/sched/ext_idle.c
152
nodemask_t *unvisited;
kernel/sched/fair.c
2871
nodemask_t nodes;
kernel/sched/fair.c
2911
nodemask_t max_group = NODE_MASK_NONE;
kernel/sched/fair.c
2920
nodemask_t this_group;
kernel/vmcore_info.c
194
VMCOREINFO_SIZE(nodemask_t);
lib/group_cpus.c
231
const nodemask_t nodemsk,
lib/group_cpus.c
412
nodemask_t nodemsk = NODE_MASK_NONE;
lib/group_cpus.c
86
const struct cpumask *mask, nodemask_t *nodemsk)
mm/hugetlb.c
1329
int nid, nodemask_t *nmask)
mm/hugetlb.c
1380
nodemask_t *nodemask;
mm/hugetlb.c
1414
int nid, nodemask_t *nodemask)
mm/hugetlb.c
1431
nodemask_t *nodemask)
mm/hugetlb.c
1845
int nid, nodemask_t *nmask, nodemask_t *node_alloc_noretry)
mm/hugetlb.c
1890
gfp_t gfp_mask, int nid, nodemask_t *nmask,
mm/hugetlb.c
1891
nodemask_t *node_alloc_noretry)
mm/hugetlb.c
1917
gfp_t gfp_mask, int nid, nodemask_t *nmask)
mm/hugetlb.c
1950
nodemask_t *nodes_allowed,
mm/hugetlb.c
1951
nodemask_t *node_alloc_noretry,
mm/hugetlb.c
1977
nodemask_t *nodes_allowed, bool acct_surplus)
mm/hugetlb.c
2130
gfp_t gfp_mask, int nid, nodemask_t *nmask)
mm/hugetlb.c
2178
int nid, nodemask_t *nmask)
mm/hugetlb.c
2215
nodemask_t *nodemask;
mm/hugetlb.c
2234
nodemask_t *nmask, gfp_t gfp_mask)
mm/hugetlb.c
2255
nodemask_t *nmask, gfp_t gfp_mask, bool allow_alloc_fallback)
mm/hugetlb.c
2277
static nodemask_t *policy_mbind_nodemask(gfp_t gfp)
mm/hugetlb.c
2307
nodemask_t *mbind_nodemask, alloc_nodemask;
mm/hugetlb.c
3496
nodemask_t node_alloc_noretry;
mm/hugetlb.c
3709
nodemask_t *nodes_allowed)
mm/hugetlb.c
3741
nodemask_t *nodes_allowed)
mm/hugetlb.c
3751
static int adjust_pool_surplus(struct hstate *h, nodemask_t *nodes_allowed,
mm/hugetlb.c
3781
nodemask_t *nodes_allowed)
mm/hugetlb.c
3788
NODEMASK_ALLOC(nodemask_t, node_alloc_noretry, GFP_KERNEL);
mm/hugetlb.c
4028
long demote_pool_huge_page(struct hstate *src, nodemask_t *nodes_allowed,
mm/hugetlb.c
4102
nodemask_t nodes_allowed, *n_mask;
mm/hugetlb.c
4551
nodemask_t *mbind_nodemask;
mm/hugetlb.c
59
__initdata nodemask_t hugetlb_bootmem_nodes;
mm/hugetlb.c
6171
nodemask_t *nodemask;
mm/hugetlb_cma.c
28
int nid, nodemask_t *nodemask)
mm/hugetlb_cma.h
21
gfp_t gfp_mask, int nid, nodemask_t *nodemask)
mm/hugetlb_cma.h
8
int nid, nodemask_t *nodemask);
mm/hugetlb_internal.h
103
nodemask_t *nodes_allowed,
mm/hugetlb_internal.h
31
static inline int next_node_allowed(int nid, nodemask_t *nodes_allowed)
mm/hugetlb_internal.h
39
static inline int get_valid_node_allowed(int nid, nodemask_t *nodes_allowed)
mm/hugetlb_internal.h
53
nodemask_t *nodes_allowed)
mm/hugetlb_internal.h
71
static inline int hstate_next_node_to_free(struct hstate *h, nodemask_t *nodes_allowed)
mm/hugetlb_sysfs.c
186
nodemask_t nodes_allowed, *n_mask;
mm/internal.h
1288
extern int find_next_best_node(int node, nodemask_t *used_node_mask);
mm/internal.h
1297
static inline int find_next_best_node(int node, nodemask_t *used_node_mask)
mm/internal.h
1421
nodemask_t *nmask;
mm/internal.h
659
nodemask_t *nodemask;
mm/internal.h
896
nodemask_t *);
mm/khugepaged.c
104
nodemask_t alloc_nmask;
mm/memcontrol.c
5651
void mem_cgroup_node_filter_allowed(struct mem_cgroup *memcg, nodemask_t *mask)
mm/memcontrol.c
5653
nodemask_t allowed;
mm/memory-tiers.c
148
static __always_inline nodemask_t get_memtier_nodemask(struct memory_tier *memtier)
mm/memory-tiers.c
150
nodemask_t nodes = NODE_MASK_NONE;
mm/memory-tiers.c
173
nodemask_t nmask;
mm/memory-tiers.c
26
nodemask_t lower_tier_mask;
mm/memory-tiers.c
30
nodemask_t preferred;
mm/memory-tiers.c
302
void node_get_allowed_targets(pg_data_t *pgdat, nodemask_t *targets)
mm/memory-tiers.c
330
int next_demotion_node(int node, const nodemask_t *allowed_mask)
mm/memory-tiers.c
333
nodemask_t mask;
mm/memory-tiers.c
406
nodemask_t preferred = node_demotion[node].preferred;
mm/memory-tiers.c
431
nodemask_t tier_nodes, lower_tier;
mm/memory-tiers.c
47
nodemask_t default_dram_nodes __initdata = NODE_MASK_NONE;
mm/memory_hotplug.c
1844
nodemask_t nmask = node_states[N_MEMORY];
mm/mempolicy.c
1068
nodemask_t *nodes)
mm/mempolicy.c
1111
static void get_policy_nodemask(struct mempolicy *pol, nodemask_t *nodes)
mm/mempolicy.c
1147
static long do_get_mempolicy(int *policy, nodemask_t *nmask,
mm/mempolicy.c
1285
nodemask_t nmask;
mm/mempolicy.c
1336
int do_migrate_pages(struct mm_struct *mm, const nodemask_t *from,
mm/mempolicy.c
1337
const nodemask_t *to, int flags)
mm/mempolicy.c
1341
nodemask_t tmp;
mm/mempolicy.c
1448
nodemask_t *nodemask;
mm/mempolicy.c
1473
int do_migrate_pages(struct mm_struct *mm, const nodemask_t *from,
mm/mempolicy.c
1474
const nodemask_t *to, int flags)
mm/mempolicy.c
1488
nodemask_t *nmask, unsigned long flags)
mm/mempolicy.c
1658
static int get_nodes(nodemask_t *nodes, const unsigned long __user *nmask,
mm/mempolicy.c
1695
nodemask_t *nodes)
mm/mempolicy.c
1744
nodemask_t nodes;
mm/mempolicy.c
1839
nodemask_t nodes;
mm/mempolicy.c
1866
nodemask_t task_nodes;
mm/mempolicy.c
1868
nodemask_t *old;
mm/mempolicy.c
1869
nodemask_t *new;
mm/mempolicy.c
1962
nodemask_t nodes;
mm/mempolicy.c
2190
nodemask_t *mask)
mm/mempolicy.c
2198
memcpy(mask, &pol->nodes, sizeof(nodemask_t));
mm/mempolicy.c
2206
nodemask_t nodemask;
mm/mempolicy.c
2250
nodemask_t nodemask;
mm/mempolicy.c
2269
static nodemask_t *policy_nodemask(gfp_t gfp, struct mempolicy *pol,
mm/mempolicy.c
2272
nodemask_t *nodemask = NULL;
mm/mempolicy.c
2328
struct mempolicy **mpol, nodemask_t **nodemask)
mm/mempolicy.c
2355
bool init_nodemask_of_mempolicy(nodemask_t *mask)
mm/mempolicy.c
2397
const nodemask_t *mask)
mm/mempolicy.c
2415
int nid, nodemask_t *nodemask)
mm/mempolicy.c
2448
nodemask_t *nodemask;
mm/mempolicy.c
2644
nodemask_t nodes;
mm/mempolicy.c
2778
nodemask_t *nodemask;
mm/mempolicy.c
2840
nodemask_t mems = cpuset_mems_allowed(current);
mm/mempolicy.c
324
int nearest_node_nodemask(int node, nodemask_t *mask)
mm/mempolicy.c
3337
nodemask_t interleave_nodes;
mm/mempolicy.c
3423
nodemask_t nodes;
mm/mempolicy.c
3558
nodemask_t nodes = NODE_MASK_NONE;
mm/mempolicy.c
361
int (*create)(struct mempolicy *pol, const nodemask_t *nodes);
mm/mempolicy.c
362
void (*rebind)(struct mempolicy *pol, const nodemask_t *nodes);
mm/mempolicy.c
370
static void mpol_relative_nodemask(nodemask_t *ret, const nodemask_t *orig,
mm/mempolicy.c
371
const nodemask_t *rel)
mm/mempolicy.c
373
nodemask_t tmp;
mm/mempolicy.c
378
static int mpol_new_nodemask(struct mempolicy *pol, const nodemask_t *nodes)
mm/mempolicy.c
386
static int mpol_new_preferred(struct mempolicy *pol, const nodemask_t *nodes)
mm/mempolicy.c
405
const nodemask_t *nodes, struct nodemask_scratch *nsc)
mm/mempolicy.c
442
nodemask_t *nodes)
mm/mempolicy.c
500
static void mpol_rebind_default(struct mempolicy *pol, const nodemask_t *nodes)
mm/mempolicy.c
504
static void mpol_rebind_nodemask(struct mempolicy *pol, const nodemask_t *nodes)
mm/mempolicy.c
506
nodemask_t tmp;
mm/mempolicy.c
525
const nodemask_t *nodes)
mm/mempolicy.c
537
static void mpol_rebind_policy(struct mempolicy *pol, const nodemask_t *newmask)
mm/mempolicy.c
554
void mpol_rebind_task(struct task_struct *tsk, const nodemask_t *new)
mm/mempolicy.c
564
void mpol_rebind_mm(struct mm_struct *mm, nodemask_t *new)
mm/mempolicy.c
608
static nodemask_t *policy_nodemask(gfp_t gfp, struct mempolicy *pol,
mm/mempolicy.c
629
nodemask_t *nmask;
mm/mempolicy.c
981
nodemask_t *nodes, unsigned long flags,
mm/migrate.c
2331
static int do_pages_move(struct mm_struct *mm, nodemask_t task_nodes,
mm/migrate.c
2526
static struct mm_struct *find_mm_struct(pid_t pid, nodemask_t *mem_nodes)
mm/migrate.c
2578
nodemask_t task_nodes;
mm/mm_init.c
363
nodemask_t saved_node_state = node_states[N_MEMORY];
mm/mmzone.c
46
static inline int zref_in_nodemask(struct zoneref *zref, nodemask_t *nodes)
mm/mmzone.c
58
nodemask_t *nodes)
mm/numa_emulation.c
225
nodemask_t physnode_mask = numa_nodes_parsed;
mm/numa_emulation.c
383
nodemask_t physnode_mask = numa_nodes_parsed;
mm/numa_emulation.c
90
nodemask_t physnode_mask = numa_nodes_parsed;
mm/numa_memblks.c
15
nodemask_t numa_nodes_parsed __initdata;
mm/numa_memblks.c
23
static void __init numa_nodemask_from_meminfo(nodemask_t *nodemask,
mm/numa_memblks.c
337
nodemask_t reserved_nodemask = NODE_MASK_NONE;
mm/numa_memblks.c
53
nodemask_t nodes_parsed;
mm/oom_kill.c
95
const nodemask_t *mask = oc->nodemask;
mm/page_alloc.c
224
nodemask_t node_states[NR_NODE_STATES] __read_mostly = {
mm/page_alloc.c
4005
static void warn_alloc_show_mem(gfp_t gfp_mask, nodemask_t *nodemask)
mm/page_alloc.c
4025
void warn_alloc(gfp_t gfp_mask, nodemask_t *nodemask, const char *fmt, ...)
mm/page_alloc.c
4997
int preferred_nid, nodemask_t *nodemask,
mm/page_alloc.c
5066
nodemask_t *nodemask, int nr_pages,
mm/page_alloc.c
5215
int preferred_nid, nodemask_t *nodemask)
mm/page_alloc.c
5280
int preferred_nid, nodemask_t *nodemask)
mm/page_alloc.c
5292
nodemask_t *nodemask)
mm/page_alloc.c
5616
int find_next_best_node(int node, nodemask_t *used_node_mask)
mm/page_alloc.c
5708
nodemask_t used_mask = NODE_MASK_NONE;
mm/page_alloc.c
7248
gfp_t gfp_mask, int nid, nodemask_t *nodemask)
mm/page_alloc.c
7320
int nid, nodemask_t *nodemask)
mm/show_mem.c
119
static bool show_mem_node_skip(unsigned int flags, int nid, nodemask_t *nodemask)
mm/show_mem.c
180
static void show_free_areas(unsigned int filter, nodemask_t *nodemask, int max_zone_idx)
mm/show_mem.c
401
void __show_mem(unsigned int filter, nodemask_t *nodemask, int max_zone_idx)
mm/slub.c
475
static nodemask_t slab_nodes;
mm/slub.c
8641
nodemask_t nodes;
mm/vmscan.c
1027
nodemask_t allowed_mask;
mm/vmscan.c
347
nodemask_t allowed_mask;
mm/vmscan.c
6487
nodemask_t *nodemask)
mm/vmscan.c
6567
gfp_t gfp_mask, nodemask_t *nodemask)
mm/vmscan.c
82
nodemask_t *nodemask;
mm/vmscan.c
990
nodemask_t *allowed_mask;