Symbol: mn
arch/powerpc/kexec/crash.c
428
static void update_crash_elfcorehdr(struct kimage *image, struct memory_notify *mn)
arch/powerpc/kexec/crash.c
451
base_addr = PFN_PHYS(mn->start_pfn);
arch/powerpc/kexec/crash.c
452
size = mn->nr_pages * PAGE_SIZE;
arch/powerpc/kexec/crash.c
573
struct memory_notify *mn;
arch/powerpc/kexec/crash.c
585
mn = (struct memory_notify *)arg;
arch/powerpc/kexec/crash.c
586
update_crash_elfcorehdr(image, mn);
arch/x86/kernel/apic/x2apic_uv_x.c
1168
static void get_mn(struct mn *mnp)
arch/x86/kernel/apic/x2apic_uv_x.c
1190
struct mn mn;
arch/x86/kernel/apic/x2apic_uv_x.c
1192
get_mn(&mn);
arch/x86/kernel/apic/x2apic_uv_x.c
1193
hi->gpa_mask = mn.m_val ?
arch/x86/kernel/apic/x2apic_uv_x.c
1194
(1UL << (mn.m_val + mn.n_val)) - 1 :
arch/x86/kernel/apic/x2apic_uv_x.c
1197
hi->m_val = mn.m_val;
arch/x86/kernel/apic/x2apic_uv_x.c
1198
hi->n_val = mn.n_val;
arch/x86/kernel/apic/x2apic_uv_x.c
1199
hi->m_shift = mn.m_shift;
arch/x86/kernel/apic/x2apic_uv_x.c
1200
hi->n_lshift = mn.n_lshift ? mn.n_lshift : 0;
arch/x86/kernel/apic/x2apic_uv_x.c
1214
uv_cpuid.gnode_shift = max_t(unsigned int, uv_cpuid.gnode_shift, mn.n_val);
arch/x86/kernel/apic/x2apic_uv_x.c
1216
if (mn.m_val)
arch/x86/kernel/apic/x2apic_uv_x.c
1217
hi->gnode_upper = (u64)hi->gnode_extra << mn.m_val;
arch/x86/kernel/cpu/sgx/encl.c
774
static void sgx_mmu_notifier_release(struct mmu_notifier *mn,
arch/x86/kernel/cpu/sgx/encl.c
777
struct sgx_encl_mm *encl_mm = container_of(mn, struct sgx_encl_mm, mmu_notifier);
arch/x86/kernel/cpu/sgx/encl.c
797
mmu_notifier_put(mn);
arch/x86/kernel/cpu/sgx/encl.c
801
static void sgx_mmu_notifier_free(struct mmu_notifier *mn)
arch/x86/kernel/cpu/sgx/encl.c
803
struct sgx_encl_mm *encl_mm = container_of(mn, struct sgx_encl_mm, mmu_notifier);
arch/x86/virt/vmx/tdx/tdx.c
1370
struct memory_notify *mn = v;
arch/x86/virt/vmx/tdx/tdx.c
1387
if (is_tdx_memory(mn->start_pfn, mn->start_pfn + mn->nr_pages))
drivers/clk/qcom/clk-rcg.c
106
static u32 md_to_m(struct mn *mn, u32 md)
drivers/clk/qcom/clk-rcg.c
108
md >>= mn->m_val_shift;
drivers/clk/qcom/clk-rcg.c
109
md &= BIT(mn->width) - 1;
drivers/clk/qcom/clk-rcg.c
132
static u32 mn_to_md(struct mn *mn, u32 m, u32 n, u32 md)
drivers/clk/qcom/clk-rcg.c
136
mask_w = BIT(mn->width) - 1;
drivers/clk/qcom/clk-rcg.c
137
mask = (mask_w << mn->m_val_shift) | mask_w;
drivers/clk/qcom/clk-rcg.c
141
m <<= mn->m_val_shift;
drivers/clk/qcom/clk-rcg.c
149
static u32 ns_m_to_n(struct mn *mn, u32 ns, u32 m)
drivers/clk/qcom/clk-rcg.c
151
ns = ~ns >> mn->n_val_shift;
drivers/clk/qcom/clk-rcg.c
152
ns &= BIT(mn->width) - 1;
drivers/clk/qcom/clk-rcg.c
156
static u32 reg_to_mnctr_mode(struct mn *mn, u32 val)
drivers/clk/qcom/clk-rcg.c
158
val >>= mn->mnctr_mode_shift;
drivers/clk/qcom/clk-rcg.c
163
static u32 mn_to_ns(struct mn *mn, u32 m, u32 n, u32 ns)
drivers/clk/qcom/clk-rcg.c
167
mask = BIT(mn->width) - 1;
drivers/clk/qcom/clk-rcg.c
168
mask <<= mn->n_val_shift;
drivers/clk/qcom/clk-rcg.c
174
n &= BIT(mn->width) - 1;
drivers/clk/qcom/clk-rcg.c
175
n <<= mn->n_val_shift;
drivers/clk/qcom/clk-rcg.c
182
static u32 mn_to_reg(struct mn *mn, u32 m, u32 n, u32 val)
drivers/clk/qcom/clk-rcg.c
186
mask = MNCTR_MODE_MASK << mn->mnctr_mode_shift;
drivers/clk/qcom/clk-rcg.c
187
mask |= BIT(mn->mnctr_en_bit);
drivers/clk/qcom/clk-rcg.c
191
val |= BIT(mn->mnctr_en_bit);
drivers/clk/qcom/clk-rcg.c
192
val |= MNCTR_MODE_DUAL << mn->mnctr_mode_shift;
drivers/clk/qcom/clk-rcg.c
202
struct mn *mn;
drivers/clk/qcom/clk-rcg.c
207
bool banked_mn = !!rcg->mn[1].width;
drivers/clk/qcom/clk-rcg.c
225
mn = &rcg->mn[new_bank];
drivers/clk/qcom/clk-rcg.c
228
ns |= BIT(mn->mnctr_reset_bit);
drivers/clk/qcom/clk-rcg.c
236
md = mn_to_md(mn, f->m, f->n, md);
drivers/clk/qcom/clk-rcg.c
240
ns = mn_to_ns(mn, f->m, f->n, ns);
drivers/clk/qcom/clk-rcg.c
247
ns = mn_to_reg(mn, f->m, f->n, ns);
drivers/clk/qcom/clk-rcg.c
252
reg = mn_to_reg(mn, f->m, f->n, reg);
drivers/clk/qcom/clk-rcg.c
259
ns &= ~BIT(mn->mnctr_reset_bit);
drivers/clk/qcom/clk-rcg.c
297
bool banked_mn = !!rcg->mn[1].width;
drivers/clk/qcom/clk-rcg.c
307
f.m = md_to_m(&rcg->mn[bank], md);
drivers/clk/qcom/clk-rcg.c
308
f.n = ns_m_to_n(&rcg->mn[bank], ns, f.m);
drivers/clk/qcom/clk-rcg.c
346
struct mn *mn = &rcg->mn;
drivers/clk/qcom/clk-rcg.c
351
if (rcg->mn.width) {
drivers/clk/qcom/clk-rcg.c
353
m = md_to_m(mn, md);
drivers/clk/qcom/clk-rcg.c
354
n = ns_m_to_n(mn, ns, m);
drivers/clk/qcom/clk-rcg.c
360
mode = reg_to_mnctr_mode(mn, mode);
drivers/clk/qcom/clk-rcg.c
372
struct mn *mn;
drivers/clk/qcom/clk-rcg.c
374
bool banked_mn = !!rcg->mn[1].width;
drivers/clk/qcom/clk-rcg.c
383
mn = &rcg->mn[bank];
drivers/clk/qcom/clk-rcg.c
385
m = md_to_m(mn, md);
drivers/clk/qcom/clk-rcg.c
386
n = ns_m_to_n(mn, ns, m);
drivers/clk/qcom/clk-rcg.c
390
mode = reg_to_mnctr_mode(mn, reg);
drivers/clk/qcom/clk-rcg.c
477
struct mn *mn = &rcg->mn;
drivers/clk/qcom/clk-rcg.c
481
if (rcg->mn.reset_in_cc)
drivers/clk/qcom/clk-rcg.c
486
if (rcg->mn.width) {
drivers/clk/qcom/clk-rcg.c
487
mask = BIT(mn->mnctr_reset_bit);
drivers/clk/qcom/clk-rcg.c
491
md = mn_to_md(mn, f->m, f->n, md);
drivers/clk/qcom/clk-rcg.c
498
ctl = mn_to_reg(mn, f->m, f->n, ctl);
drivers/clk/qcom/clk-rcg.c
501
ns = mn_to_reg(mn, f->m, f->n, ns);
drivers/clk/qcom/clk-rcg.c
503
ns = mn_to_ns(mn, f->m, f->n, ns);
drivers/clk/qcom/clk-rcg.h
135
struct mn mn[2];
drivers/clk/qcom/clk-rcg.h
96
struct mn mn;
drivers/clk/qcom/gcc-ipq806x.c
1141
.mn = {
drivers/clk/qcom/gcc-ipq806x.c
1192
.mn = {
drivers/clk/qcom/gcc-ipq806x.c
1243
.mn = {
drivers/clk/qcom/gcc-ipq806x.c
1364
.mn = {
drivers/clk/qcom/gcc-ipq806x.c
1414
.mn = {
drivers/clk/qcom/gcc-ipq806x.c
1499
.mn = {
drivers/clk/qcom/gcc-ipq806x.c
2091
.mn = {
drivers/clk/qcom/gcc-ipq806x.c
2165
.mn = {
drivers/clk/qcom/gcc-ipq806x.c
2239
.mn = {
drivers/clk/qcom/gcc-ipq806x.c
2305
.mn = {
drivers/clk/qcom/gcc-ipq806x.c
2424
.mn[0] = {
drivers/clk/qcom/gcc-ipq806x.c
2432
.mn[1] = {
drivers/clk/qcom/gcc-ipq806x.c
2496
.mn[0] = {
drivers/clk/qcom/gcc-ipq806x.c
2504
.mn[1] = {
drivers/clk/qcom/gcc-ipq806x.c
2568
.mn[0] = {
drivers/clk/qcom/gcc-ipq806x.c
2576
.mn[1] = {
drivers/clk/qcom/gcc-ipq806x.c
2640
.mn[0] = {
drivers/clk/qcom/gcc-ipq806x.c
2648
.mn[1] = {
drivers/clk/qcom/gcc-ipq806x.c
2786
.mn[0] = {
drivers/clk/qcom/gcc-ipq806x.c
2794
.mn[1] = {
drivers/clk/qcom/gcc-ipq806x.c
2839
.mn[0] = {
drivers/clk/qcom/gcc-ipq806x.c
2847
.mn[1] = {
drivers/clk/qcom/gcc-ipq806x.c
414
.mn = {
drivers/clk/qcom/gcc-ipq806x.c
465
.mn = {
drivers/clk/qcom/gcc-ipq806x.c
516
.mn = {
drivers/clk/qcom/gcc-ipq806x.c
567
.mn = {
drivers/clk/qcom/gcc-ipq806x.c
618
.mn = {
drivers/clk/qcom/gcc-ipq806x.c
669
.mn = {
drivers/clk/qcom/gcc-ipq806x.c
733
.mn = {
drivers/clk/qcom/gcc-ipq806x.c
784
.mn = {
drivers/clk/qcom/gcc-ipq806x.c
835
.mn = {
drivers/clk/qcom/gcc-ipq806x.c
886
.mn = {
drivers/clk/qcom/gcc-ipq806x.c
937
.mn = {
drivers/clk/qcom/gcc-ipq806x.c
988
.mn = {
drivers/clk/qcom/gcc-mdm9615.c
1041
.mn = {
drivers/clk/qcom/gcc-mdm9615.c
1092
.mn = {
drivers/clk/qcom/gcc-mdm9615.c
1148
.mn = {
drivers/clk/qcom/gcc-mdm9615.c
1204
.mn = {
drivers/clk/qcom/gcc-mdm9615.c
1260
.mn = {
drivers/clk/qcom/gcc-mdm9615.c
188
.mn = {
drivers/clk/qcom/gcc-mdm9615.c
239
.mn = {
drivers/clk/qcom/gcc-mdm9615.c
290
.mn = {
drivers/clk/qcom/gcc-mdm9615.c
341
.mn = {
drivers/clk/qcom/gcc-mdm9615.c
392
.mn = {
drivers/clk/qcom/gcc-mdm9615.c
455
.mn = {
drivers/clk/qcom/gcc-mdm9615.c
506
.mn = {
drivers/clk/qcom/gcc-mdm9615.c
557
.mn = {
drivers/clk/qcom/gcc-mdm9615.c
608
.mn = {
drivers/clk/qcom/gcc-mdm9615.c
659
.mn = {
drivers/clk/qcom/gcc-mdm9615.c
716
.mn = {
drivers/clk/qcom/gcc-mdm9615.c
767
.mn = {
drivers/clk/qcom/gcc-mdm9615.c
818
.mn = {
drivers/clk/qcom/gcc-mdm9615.c
936
.mn = {
drivers/clk/qcom/gcc-mdm9615.c
986
.mn = {
drivers/clk/qcom/gcc-msm8660.c
1037
.mn = {
drivers/clk/qcom/gcc-msm8660.c
106
.mn = {
drivers/clk/qcom/gcc-msm8660.c
1088
.mn = {
drivers/clk/qcom/gcc-msm8660.c
1139
.mn = {
drivers/clk/qcom/gcc-msm8660.c
1190
.mn = {
drivers/clk/qcom/gcc-msm8660.c
1241
.mn = {
drivers/clk/qcom/gcc-msm8660.c
1292
.mn = {
drivers/clk/qcom/gcc-msm8660.c
1356
.mn = {
drivers/clk/qcom/gcc-msm8660.c
1407
.mn = {
drivers/clk/qcom/gcc-msm8660.c
1458
.mn = {
drivers/clk/qcom/gcc-msm8660.c
157
.mn = {
drivers/clk/qcom/gcc-msm8660.c
1573
.mn = {
drivers/clk/qcom/gcc-msm8660.c
1623
.mn = {
drivers/clk/qcom/gcc-msm8660.c
1673
.mn = {
drivers/clk/qcom/gcc-msm8660.c
1723
.mn = {
drivers/clk/qcom/gcc-msm8660.c
1773
.mn = {
drivers/clk/qcom/gcc-msm8660.c
1828
.mn = {
drivers/clk/qcom/gcc-msm8660.c
1884
.mn = {
drivers/clk/qcom/gcc-msm8660.c
1935
.mn = {
drivers/clk/qcom/gcc-msm8660.c
2004
.mn = {
drivers/clk/qcom/gcc-msm8660.c
208
.mn = {
drivers/clk/qcom/gcc-msm8660.c
259
.mn = {
drivers/clk/qcom/gcc-msm8660.c
310
.mn = {
drivers/clk/qcom/gcc-msm8660.c
361
.mn = {
drivers/clk/qcom/gcc-msm8660.c
412
.mn = {
drivers/clk/qcom/gcc-msm8660.c
463
.mn = {
drivers/clk/qcom/gcc-msm8660.c
514
.mn = {
drivers/clk/qcom/gcc-msm8660.c
565
.mn = {
drivers/clk/qcom/gcc-msm8660.c
616
.mn = {
drivers/clk/qcom/gcc-msm8660.c
667
.mn = {
drivers/clk/qcom/gcc-msm8660.c
731
.mn = {
drivers/clk/qcom/gcc-msm8660.c
782
.mn = {
drivers/clk/qcom/gcc-msm8660.c
833
.mn = {
drivers/clk/qcom/gcc-msm8660.c
884
.mn = {
drivers/clk/qcom/gcc-msm8660.c
935
.mn = {
drivers/clk/qcom/gcc-msm8660.c
986
.mn = {
drivers/clk/qcom/gcc-msm8960.c
1027
.mn = {
drivers/clk/qcom/gcc-msm8960.c
1078
.mn = {
drivers/clk/qcom/gcc-msm8960.c
1129
.mn = {
drivers/clk/qcom/gcc-msm8960.c
1180
.mn = {
drivers/clk/qcom/gcc-msm8960.c
1231
.mn = {
drivers/clk/qcom/gcc-msm8960.c
1282
.mn = {
drivers/clk/qcom/gcc-msm8960.c
1333
.mn = {
drivers/clk/qcom/gcc-msm8960.c
1384
.mn = {
drivers/clk/qcom/gcc-msm8960.c
1435
.mn = {
drivers/clk/qcom/gcc-msm8960.c
1486
.mn = {
drivers/clk/qcom/gcc-msm8960.c
1537
.mn = {
drivers/clk/qcom/gcc-msm8960.c
1601
.mn = {
drivers/clk/qcom/gcc-msm8960.c
1652
.mn = {
drivers/clk/qcom/gcc-msm8960.c
1703
.mn = {
drivers/clk/qcom/gcc-msm8960.c
1821
.mn = {
drivers/clk/qcom/gcc-msm8960.c
1871
.mn = {
drivers/clk/qcom/gcc-msm8960.c
1921
.mn = {
drivers/clk/qcom/gcc-msm8960.c
1971
.mn = {
drivers/clk/qcom/gcc-msm8960.c
2021
.mn = {
drivers/clk/qcom/gcc-msm8960.c
2076
.mn = {
drivers/clk/qcom/gcc-msm8960.c
2132
.mn = {
drivers/clk/qcom/gcc-msm8960.c
2183
.mn = {
drivers/clk/qcom/gcc-msm8960.c
2234
.mn = {
drivers/clk/qcom/gcc-msm8960.c
2285
.mn = {
drivers/clk/qcom/gcc-msm8960.c
2384
.mn = {
drivers/clk/qcom/gcc-msm8960.c
2453
.mn = {
drivers/clk/qcom/gcc-msm8960.c
351
.mn = {
drivers/clk/qcom/gcc-msm8960.c
402
.mn = {
drivers/clk/qcom/gcc-msm8960.c
453
.mn = {
drivers/clk/qcom/gcc-msm8960.c
504
.mn = {
drivers/clk/qcom/gcc-msm8960.c
555
.mn = {
drivers/clk/qcom/gcc-msm8960.c
606
.mn = {
drivers/clk/qcom/gcc-msm8960.c
657
.mn = {
drivers/clk/qcom/gcc-msm8960.c
708
.mn = {
drivers/clk/qcom/gcc-msm8960.c
759
.mn = {
drivers/clk/qcom/gcc-msm8960.c
810
.mn = {
drivers/clk/qcom/gcc-msm8960.c
861
.mn = {
drivers/clk/qcom/gcc-msm8960.c
912
.mn = {
drivers/clk/qcom/gcc-msm8960.c
976
.mn = {
drivers/clk/qcom/lcc-ipq806x.c
113
.mn = {
drivers/clk/qcom/lcc-ipq806x.c
230
.mn = {
drivers/clk/qcom/lcc-ipq806x.c
314
.mn = {
drivers/clk/qcom/lcc-ipq806x.c
371
.mn = {
drivers/clk/qcom/lcc-msm8960.c
258
.mn = {
drivers/clk/qcom/lcc-msm8960.c
328
.mn = {
drivers/clk/qcom/lcc-msm8960.c
96
.mn = { \
drivers/clk/qcom/mmcc-msm8960.c
1011
.mn[0] = {
drivers/clk/qcom/mmcc-msm8960.c
1019
.mn[1] = {
drivers/clk/qcom/mmcc-msm8960.c
1091
.mn[0] = {
drivers/clk/qcom/mmcc-msm8960.c
1099
.mn[1] = {
drivers/clk/qcom/mmcc-msm8960.c
1183
.mn = {
drivers/clk/qcom/mmcc-msm8960.c
1306
.mn[0] = {
drivers/clk/qcom/mmcc-msm8960.c
1314
.mn[1] = {
drivers/clk/qcom/mmcc-msm8960.c
1485
.mn = {
drivers/clk/qcom/mmcc-msm8960.c
1659
.mn[0] = {
drivers/clk/qcom/mmcc-msm8960.c
1667
.mn[1] = {
drivers/clk/qcom/mmcc-msm8960.c
1791
.mn = {
drivers/clk/qcom/mmcc-msm8960.c
189
.mn = {
drivers/clk/qcom/mmcc-msm8960.c
2130
.mn = {
drivers/clk/qcom/mmcc-msm8960.c
2180
.mn = {
drivers/clk/qcom/mmcc-msm8960.c
2392
.mn = {
drivers/clk/qcom/mmcc-msm8960.c
240
.mn = {
drivers/clk/qcom/mmcc-msm8960.c
2441
.mn = {
drivers/clk/qcom/mmcc-msm8960.c
291
.mn = {
drivers/clk/qcom/mmcc-msm8960.c
349
.mn = {
drivers/clk/qcom/mmcc-msm8960.c
417
.mn = {
drivers/clk/qcom/mmcc-msm8960.c
485
.mn = {
drivers/clk/qcom/mmcc-msm8960.c
740
.mn = {
drivers/clk/qcom/mmcc-msm8960.c
846
.mn[0] = {
drivers/clk/qcom/mmcc-msm8960.c
854
.mn[1] = {
drivers/clk/qcom/mmcc-msm8960.c
908
.mn[0] = {
drivers/clk/qcom/mmcc-msm8960.c
916
.mn[1] = {
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.h
68
struct amdgpu_mn *mn;
drivers/gpu/drm/amd/amdgpu/amdgpu_ip.h
77
#define IP_VERSION_FULL(mj, mn, rv, var, srev) \
drivers/gpu/drm/amd/amdgpu/amdgpu_ip.h
78
(((mj) << 24) | ((mn) << 16) | ((rv) << 8) | ((var) << 4) | (srev))
drivers/gpu/drm/amd/amdgpu/amdgpu_ip.h
79
#define IP_VERSION(mj, mn, rv) IP_VERSION_FULL(mj, mn, rv, 0, 0)
drivers/gpu/drm/amd/amdkfd/kfd_process.c
1277
static void kfd_process_free_notifier(struct mmu_notifier *mn)
drivers/gpu/drm/amd/amdkfd/kfd_process.c
1279
kfd_unref_process(container_of(mn, struct kfd_process, mmu_notifier));
drivers/gpu/drm/amd/amdkfd/kfd_process.c
1351
static void kfd_process_notifier_release(struct mmu_notifier *mn,
drivers/gpu/drm/amd/amdkfd/kfd_process.c
1360
p = container_of(mn, struct kfd_process, mmu_notifier);
drivers/gpu/drm/amd/amdkfd/kfd_process.c
1592
struct mmu_notifier *mn;
drivers/gpu/drm/amd/amdkfd/kfd_process.c
1651
mn = mmu_notifier_get(&kfd_process_mmu_notifier_ops, process->mm);
drivers/gpu/drm/amd/amdkfd/kfd_process.c
1652
if (IS_ERR(mn)) {
drivers/gpu/drm/amd/amdkfd/kfd_process.c
1653
err = PTR_ERR(mn);
drivers/gpu/drm/amd/amdkfd/kfd_process.c
1656
BUG_ON(mn != &process->mmu_notifier);
drivers/gpu/drm/i915/gem/i915_gem_mman.c
634
struct i915_mmap_offset *mmo, *mn;
drivers/gpu/drm/i915/gem/i915_gem_mman.c
640
rbtree_postorder_for_each_entry_safe(mmo, mn,
drivers/gpu/drm/i915/gem/i915_gem_object.c
244
struct i915_mmap_offset *mmo, *mn;
drivers/gpu/drm/i915/gem/i915_gem_object.c
268
rbtree_postorder_for_each_entry_safe(mmo, mn, &obj->mmo.offsets, offset)
drivers/gpu/drm/i915/gem/i915_gem_object.c
320
struct i915_mmap_offset *mmo, *mn;
drivers/gpu/drm/i915/gem/i915_gem_object.c
324
rbtree_postorder_for_each_entry_safe(mmo, mn,
drivers/gpu/drm/nouveau/include/nvkm/core/memory.h
10
struct nvkm_mm_node *mn;
drivers/gpu/drm/nouveau/nouveau_svm.c
252
nouveau_svmm_invalidate_range_start(struct mmu_notifier *mn,
drivers/gpu/drm/nouveau/nouveau_svm.c
256
container_of(mn, struct nouveau_svmm, notifier);
drivers/gpu/drm/nouveau/nouveau_svm.c
292
static void nouveau_svmm_free_notifier(struct mmu_notifier *mn)
drivers/gpu/drm/nouveau/nouveau_svm.c
294
kfree(container_of(mn, struct nouveau_svmm, notifier));
drivers/gpu/drm/nouveau/nvkm/core/memory.c
38
nvkm_mm_free(&fb->tags.mm, &tags->mn);
drivers/gpu/drm/nouveau/nvkm/core/memory.c
61
if (tags->mn && tags->mn->length != nr) {
drivers/gpu/drm/nouveau/nvkm/core/memory.c
77
if (!nvkm_mm_head(&fb->tags.mm, 0, 1, nr, nr, 1, &tags->mn)) {
drivers/gpu/drm/nouveau/nvkm/core/memory.c
79
clr(device, tags->mn->offset, tags->mn->length);
drivers/gpu/drm/nouveau/nvkm/core/memory.c
90
tags->mn = NULL;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
101
kfree(vram->mn);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
138
vram->mn = kzalloc_obj(*vram->mn);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
139
if (!vram->mn)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
142
vram->mn->offset = addr >> NVKM_RAM_MM_SHIFT;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
143
vram->mn->length = size >> NVKM_RAM_MM_SHIFT;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
174
node = &vram->mn;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
35
struct nvkm_mm_node *mn;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
52
.mem = vram->mn,
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
61
return (u64)nvkm_mm_size(nvkm_vram(memory)->mn) << NVKM_RAM_MM_SHIFT;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
68
if (!nvkm_mm_contiguous(vram->mn))
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
70
return (u64)nvkm_mm_addr(vram->mn) << NVKM_RAM_MM_SHIFT;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
89
struct nvkm_mm_node *next = vram->mn;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
130
return (u64)gk20a_instobj(memory)->mn->offset << 12;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
136
return (u64)gk20a_instobj(memory)->mn->length << 12;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
288
.mem = node->mn,
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
304
dma_free_attrs(dev, (u64)node->base.mn->length << PAGE_SHIFT,
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
317
struct nvkm_mm_node *r = node->base.mn;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
335
for (i = 0; i < node->base.mn->length; i++) {
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
416
node->base.mn = &node->r;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
494
node->base.mn = r;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
53
struct nvkm_mm_node *mn;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
543
size, align, (u64)node->mn->offset << 12);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
300
if (!map->no_comp && map->tags->mn) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
301
u64 tags = map->tags->mn->offset + (map->offset >> 17);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
307
if (map->tags->mn) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
308
u32 tags = map->tags->mn->offset +
drivers/gpu/drm/radeon/radeon_mn.c
49
static bool radeon_mn_invalidate(struct mmu_interval_notifier *mn,
drivers/gpu/drm/radeon/radeon_mn.c
53
struct radeon_bo *bo = container_of(mn, struct radeon_bo, notifier);
drivers/infiniband/hw/hfi1/mmu_rb.c
115
mmdrop(handler->mn.mm);
drivers/infiniband/hw/hfi1/mmu_rb.c
129
if (current->mm != handler->mn.mm)
drivers/infiniband/hw/hfi1/mmu_rb.c
233
if (current->mm != handler->mn.mm)
drivers/infiniband/hw/hfi1/mmu_rb.c
261
static int mmu_notifier_range_start(struct mmu_notifier *mn,
drivers/infiniband/hw/hfi1/mmu_rb.c
265
container_of(mn, struct mmu_rb_handler, mn);
drivers/infiniband/hw/hfi1/mmu_rb.c
59
INIT_HLIST_NODE(&h->mn.hlist);
drivers/infiniband/hw/hfi1/mmu_rb.c
61
h->mn.ops = &mn_opts;
drivers/infiniband/hw/hfi1/mmu_rb.c
68
ret = mmu_notifier_register(&h->mn, current->mm);
drivers/infiniband/hw/hfi1/mmu_rb.c
86
mmgrab(handler->mn.mm);
drivers/infiniband/hw/hfi1/mmu_rb.c
89
mmu_notifier_unregister(&handler->mn, handler->mn.mm);
drivers/infiniband/hw/hfi1/mmu_rb.h
39
struct mmu_notifier mn;
drivers/infiniband/hw/hfi1/pin_system.c
74
return node->rb.handler->mn.mm;
drivers/iommu/amd/amd_iommu_types.h
554
struct mmu_notifier mn; /* mmu notifier for the SVA domain */
drivers/iommu/amd/pasid.c
171
if (sva_pdom->mn.ops)
drivers/iommu/amd/pasid.c
172
mmu_notifier_unregister(&sva_pdom->mn, domain->mm);
drivers/iommu/amd/pasid.c
193
pdom->mn.ops = &sva_mn;
drivers/iommu/amd/pasid.c
196
ret = mmu_notifier_register(&pdom->mn, mm);
drivers/iommu/amd/pasid.c
59
static void sva_arch_invalidate_secondary_tlbs(struct mmu_notifier *mn,
drivers/iommu/amd/pasid.c
67
sva_pdom = container_of(mn, struct protection_domain, mn);
drivers/iommu/amd/pasid.c
80
static void sva_mn_release(struct mmu_notifier *mn, struct mm_struct *mm)
drivers/iommu/amd/pasid.c
86
sva_pdom = container_of(mn, struct protection_domain, mn);
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3-sva.c
134
static void arm_smmu_mm_arch_invalidate_secondary_tlbs(struct mmu_notifier *mn,
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3-sva.c
140
container_of(mn, struct arm_smmu_domain, mmu_notifier);
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3-sva.c
166
static void arm_smmu_mm_release(struct mmu_notifier *mn, struct mm_struct *mm)
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3-sva.c
169
container_of(mn, struct arm_smmu_domain, mmu_notifier);
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3-sva.c
198
static void arm_smmu_mmu_notifier_free(struct mmu_notifier *mn)
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3-sva.c
200
kfree(container_of(mn, struct arm_smmu_domain, mmu_notifier));
drivers/iommu/intel/svm.c
101
struct dmar_domain *domain = container_of(mn, struct dmar_domain, notifier);
drivers/iommu/intel/svm.c
51
static void intel_arch_invalidate_secondary_tlbs(struct mmu_notifier *mn,
drivers/iommu/intel/svm.c
55
struct dmar_domain *domain = container_of(mn, struct dmar_domain, notifier);
drivers/iommu/intel/svm.c
70
static void intel_mm_release(struct mmu_notifier *mn, struct mm_struct *mm)
drivers/iommu/intel/svm.c
72
struct dmar_domain *domain = container_of(mn, struct dmar_domain, notifier);
drivers/iommu/intel/svm.c
99
static void intel_mm_free_notifier(struct mmu_notifier *mn)
drivers/md/persistent-data/dm-btree.c
743
struct btree_node *ln, *rn, *mn, *pn;
drivers/md/persistent-data/dm-btree.c
768
mn = dm_block_data(middle);
drivers/md/persistent-data/dm-btree.c
771
mn->header.nr_entries = cpu_to_le32(0);
drivers/md/persistent-data/dm-btree.c
772
mn->header.flags = ln->header.flags;
drivers/md/persistent-data/dm-btree.c
773
mn->header.max_entries = ln->header.max_entries;
drivers/md/persistent-data/dm-btree.c
774
mn->header.value_size = ln->header.value_size;
drivers/md/persistent-data/dm-btree.c
776
redistribute3(ln, mn, rn);
drivers/md/persistent-data/dm-btree.c
783
le64_to_cpu(mn->keys[0]), &location);
drivers/md/persistent-data/dm-btree.c
798
if (key < le64_to_cpu(mn->keys[0])) {
drivers/misc/ocxl/link.c
494
static void arch_invalidate_secondary_tlbs(struct mmu_notifier *mn,
drivers/misc/ocxl/link.c
498
struct pe_data *pe_data = container_of(mn, struct pe_data, mmu_notifier);
drivers/misc/sgi-gru/grutlbpurge.c
207
static int gru_invalidate_range_start(struct mmu_notifier *mn,
drivers/misc/sgi-gru/grutlbpurge.c
210
struct gru_mm_struct *gms = container_of(mn, struct gru_mm_struct,
drivers/misc/sgi-gru/grutlbpurge.c
222
static void gru_invalidate_range_end(struct mmu_notifier *mn,
drivers/misc/sgi-gru/grutlbpurge.c
225
struct gru_mm_struct *gms = container_of(mn, struct gru_mm_struct,
drivers/misc/sgi-gru/grutlbpurge.c
250
static void gru_free_notifier(struct mmu_notifier *mn)
drivers/misc/sgi-gru/grutlbpurge.c
252
kfree(container_of(mn, struct gru_mm_struct, ms_notifier));
drivers/misc/sgi-gru/grutlbpurge.c
265
struct mmu_notifier *mn;
drivers/misc/sgi-gru/grutlbpurge.c
267
mn = mmu_notifier_get_locked(&gru_mmuops, current->mm);
drivers/misc/sgi-gru/grutlbpurge.c
268
if (IS_ERR(mn))
drivers/misc/sgi-gru/grutlbpurge.c
269
return ERR_CAST(mn);
drivers/misc/sgi-gru/grutlbpurge.c
271
return container_of(mn, struct gru_mm_struct, ms_notifier);
drivers/net/ethernet/chelsio/cxgb4/cudbg_entity.h
230
u8 mn[ID_LEN + 1];
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
2737
memcpy(vpd_data->mn, vpd.id, ID_LEN + 1);
drivers/net/ethernet/mellanox/mlx5/core/diag/en_rep_tracepoint.h
24
TP_fast_assign(const struct mlx5e_neigh *mn = &nhe->m_neigh;
drivers/net/ethernet/mellanox/mlx5/core/diag/en_rep_tracepoint.h
34
if (mn->family == AF_INET) {
drivers/net/ethernet/mellanox/mlx5/core/diag/en_rep_tracepoint.h
35
*p32 = mn->dst_ip.v4;
drivers/net/ethernet/mellanox/mlx5/core/diag/en_rep_tracepoint.h
37
} else if (mn->family == AF_INET6) {
drivers/net/ethernet/mellanox/mlx5/core/diag/en_rep_tracepoint.h
38
*pin6 = mn->dst_ip.v6;
drivers/net/ethernet/mellanox/mlx5/core/diag/en_tc_tracepoint.h
85
TP_fast_assign(const struct mlx5e_neigh *mn = &nhe->m_neigh;
drivers/net/ethernet/mellanox/mlx5/core/diag/en_tc_tracepoint.h
94
if (mn->family == AF_INET) {
drivers/net/ethernet/mellanox/mlx5/core/diag/en_tc_tracepoint.h
95
*p32 = mn->dst_ip.v4;
drivers/net/ethernet/mellanox/mlx5/core/diag/en_tc_tracepoint.h
97
} else if (mn->family == AF_INET6) {
drivers/net/ethernet/mellanox/mlx5/core/diag/en_tc_tracepoint.h
98
*pin6 = mn->dst_ip.v6;
drivers/net/ethernet/ni/nixge.c
1293
struct device_node *mn, *phy_node;
drivers/net/ethernet/ni/nixge.c
1349
mn = of_get_child_by_name(pdev->dev.of_node, "mdio");
drivers/net/ethernet/ni/nixge.c
1350
if (mn) {
drivers/net/ethernet/ni/nixge.c
1351
err = nixge_mdio_setup(priv, mn);
drivers/net/ethernet/ni/nixge.c
1352
of_node_put(mn);
drivers/nvme/host/core.c
2983
const char *mn;
drivers/nvme/host/core.c
2995
.mn = "THNSF5256GPUK TOSHIBA",
drivers/nvme/host/core.c
3019
.mn = "KCD6XVUL6T40",
drivers/nvme/host/core.c
3031
.mn = "Samsung Portable SSD X5",
drivers/nvme/host/core.c
3063
string_matches(id->mn, q->mn, sizeof(id->mn)) &&
drivers/nvme/host/core.c
3094
memcpy(subsys->subnqn + off, id->mn, sizeof(id->mn));
drivers/nvme/host/core.c
3095
off += sizeof(id->mn);
drivers/nvme/host/core.c
3218
memcpy(subsys->model, id->mn, sizeof(subsys->model));
drivers/nvme/target/admin-cmd.c
672
memcpy_and_pad(id->mn, sizeof(id->mn), subsys->model_number,
drivers/nvme/target/discovery.c
274
memcpy_and_pad(id->mn, sizeof(id->mn), ctrl->subsys->model_number,
drivers/scsi/qla2xxx/qla_bsg.c
1021
struct verify_chip_entry_84xx *mn = NULL;
drivers/scsi/qla2xxx/qla_bsg.c
1067
mn = dma_pool_zalloc(ha->s_dma_pool, GFP_KERNEL, &mn_dma);
drivers/scsi/qla2xxx/qla_bsg.c
1068
if (!mn) {
drivers/scsi/qla2xxx/qla_bsg.c
1078
mn->entry_type = VERIFY_CHIP_IOCB_TYPE;
drivers/scsi/qla2xxx/qla_bsg.c
1079
mn->entry_count = 1;
drivers/scsi/qla2xxx/qla_bsg.c
1085
mn->options = cpu_to_le16(options);
drivers/scsi/qla2xxx/qla_bsg.c
1086
mn->fw_ver = cpu_to_le32(fw_ver);
drivers/scsi/qla2xxx/qla_bsg.c
1087
mn->fw_size = cpu_to_le32(data_len);
drivers/scsi/qla2xxx/qla_bsg.c
1088
mn->fw_seq_size = cpu_to_le32(data_len);
drivers/scsi/qla2xxx/qla_bsg.c
1089
put_unaligned_le64(fw_dma, &mn->dsd.address);
drivers/scsi/qla2xxx/qla_bsg.c
1090
mn->dsd.length = cpu_to_le32(data_len);
drivers/scsi/qla2xxx/qla_bsg.c
1091
mn->data_seg_cnt = cpu_to_le16(1);
drivers/scsi/qla2xxx/qla_bsg.c
1093
rval = qla2x00_issue_iocb_timeout(vha, mn, mn_dma, 0, 120);
drivers/scsi/qla2xxx/qla_bsg.c
1108
dma_pool_free(ha->s_dma_pool, mn, mn_dma);
drivers/scsi/qla2xxx/qla_bsg.c
1131
struct access_chip_84xx *mn = NULL;
drivers/scsi/qla2xxx/qla_bsg.c
1146
mn = dma_pool_zalloc(ha->s_dma_pool, GFP_KERNEL, &mn_dma);
drivers/scsi/qla2xxx/qla_bsg.c
1147
if (!mn) {
drivers/scsi/qla2xxx/qla_bsg.c
1153
mn->entry_type = ACCESS_CHIP_IOCB_TYPE;
drivers/scsi/qla2xxx/qla_bsg.c
1154
mn->entry_count = 1;
drivers/scsi/qla2xxx/qla_bsg.c
1192
mn->options = cpu_to_le16(ACO_DUMP_MEMORY);
drivers/scsi/qla2xxx/qla_bsg.c
1193
mn->parameter1 =
drivers/scsi/qla2xxx/qla_bsg.c
1198
mn->options = cpu_to_le16(ACO_REQUEST_INFO);
drivers/scsi/qla2xxx/qla_bsg.c
1199
mn->parameter1 =
drivers/scsi/qla2xxx/qla_bsg.c
1202
mn->parameter2 =
drivers/scsi/qla2xxx/qla_bsg.c
1244
mn->options = cpu_to_le16(ACO_LOAD_MEMORY);
drivers/scsi/qla2xxx/qla_bsg.c
1245
mn->parameter1 =
drivers/scsi/qla2xxx/qla_bsg.c
1250
mn->options = cpu_to_le16(ACO_CHANGE_CONFIG_PARAM);
drivers/scsi/qla2xxx/qla_bsg.c
1251
mn->parameter1 =
drivers/scsi/qla2xxx/qla_bsg.c
1254
mn->parameter2 =
drivers/scsi/qla2xxx/qla_bsg.c
1257
mn->parameter3 =
drivers/scsi/qla2xxx/qla_bsg.c
1267
mn->total_byte_cnt = cpu_to_le32(ql84_mgmt->mgmt.len);
drivers/scsi/qla2xxx/qla_bsg.c
1268
mn->dseg_count = cpu_to_le16(1);
drivers/scsi/qla2xxx/qla_bsg.c
1269
put_unaligned_le64(mgmt_dma, &mn->dsd.address);
drivers/scsi/qla2xxx/qla_bsg.c
1270
mn->dsd.length = cpu_to_le32(ql84_mgmt->mgmt.len);
drivers/scsi/qla2xxx/qla_bsg.c
1273
rval = qla2x00_issue_iocb(vha, mn, mn_dma, 0);
drivers/scsi/qla2xxx/qla_bsg.c
1311
dma_pool_free(ha->s_dma_pool, mn, mn_dma);
drivers/scsi/qla2xxx/qla_mbx.c
4459
struct cs84xx_mgmt_cmd *mn;
drivers/scsi/qla2xxx/qla_mbx.c
4468
mn = dma_pool_alloc(ha->s_dma_pool, GFP_KERNEL, &mn_dma);
drivers/scsi/qla2xxx/qla_mbx.c
4469
if (mn == NULL) {
drivers/scsi/qla2xxx/qla_mbx.c
4482
memset(mn, 0, sizeof(*mn));
drivers/scsi/qla2xxx/qla_mbx.c
4483
mn->p.req.entry_type = VERIFY_CHIP_IOCB_TYPE;
drivers/scsi/qla2xxx/qla_mbx.c
4484
mn->p.req.entry_count = 1;
drivers/scsi/qla2xxx/qla_mbx.c
4485
mn->p.req.options = cpu_to_le16(options);
drivers/scsi/qla2xxx/qla_mbx.c
4490
mn, sizeof(*mn));
drivers/scsi/qla2xxx/qla_mbx.c
4492
rval = qla2x00_issue_iocb_timeout(vha, mn, mn_dma, 0, 120);
drivers/scsi/qla2xxx/qla_mbx.c
4502
mn, sizeof(*mn));
drivers/scsi/qla2xxx/qla_mbx.c
4504
status[0] = le16_to_cpu(mn->p.rsp.comp_status);
drivers/scsi/qla2xxx/qla_mbx.c
4506
le16_to_cpu(mn->p.rsp.failure_code) : 0;
drivers/scsi/qla2xxx/qla_mbx.c
4523
le32_to_cpu(mn->p.rsp.fw_ver));
drivers/scsi/qla2xxx/qla_mbx.c
4528
le32_to_cpu(mn->p.rsp.fw_ver);
drivers/scsi/qla2xxx/qla_mbx.c
4535
dma_pool_free(ha->s_dma_pool, mn, mn_dma);
drivers/video/fbdev/mmp/hw/mmp_ctrl.h
430
#define CFG_HWC_1BITENA(mn) ((mn)<<25)
drivers/video/fbdev/pxa168fb.h
221
#define CFG_HWC_1BITENA(mn) ((mn) << 25)
drivers/xen/gntdev.c
539
static bool gntdev_invalidate(struct mmu_interval_notifier *mn,
drivers/xen/gntdev.c
544
container_of(mn, struct gntdev_grant_map, notifier);
include/linux/nvme.h
313
char mn[40];
lib/maple_tree.c
1311
struct maple_node *mn;
lib/maple_tree.c
1318
mn = mas_mn(mas);
lib/maple_tree.c
1319
slots = ma_slots(mn, mt);
lib/maple_tree.c
1341
pivots = ma_pivots(mn, mt);
lib/maple_tree.c
1350
max_piv = ma_data_end(mn, mt, pivots, mas->max) - 1;
lib/maple_tree.c
332
static inline void mte_set_node_dead(struct maple_enode *mn)
lib/maple_tree.c
334
mte_to_node(mn)->parent = ma_parent_ptr(mte_to_node(mn));
lib/maple_tree.c
6640
struct maple_enode *p, *mn = mas->node;
lib/maple_tree.c
6647
if (mte_is_root(mn))
lib/maple_tree.c
6650
mas->node = mn;
lib/maple_tree.c
678
static inline void mte_set_pivot(struct maple_enode *mn, unsigned char piv,
lib/maple_tree.c
681
struct maple_node *node = mte_to_node(mn);
lib/maple_tree.c
682
enum maple_type type = mte_node_type(mn);
lib/maple_tree.c
706
static inline void __rcu **ma_slots(struct maple_node *mn, enum maple_type mt)
lib/maple_tree.c
710
return mn->ma64.slot;
lib/maple_tree.c
713
return mn->mr64.slot;
lib/maple_tree.c
715
return mn->slot;
lib/maple_tree.c
799
static inline struct maple_metadata *ma_meta(struct maple_node *mn,
lib/maple_tree.c
804
return &mn->ma64.meta;
lib/maple_tree.c
806
return &mn->mr64.meta;
lib/maple_tree.c
817
static inline void ma_set_meta(struct maple_node *mn, enum maple_type mt,
lib/maple_tree.c
820
struct maple_metadata *meta = ma_meta(mn, mt);
lib/maple_tree.c
832
static inline void mt_clear_meta(struct maple_tree *mt, struct maple_node *mn,
lib/maple_tree.c
842
pivots = mn->mr64.pivot;
lib/maple_tree.c
844
slots = mn->mr64.slot;
lib/maple_tree.c
853
meta = ma_meta(mn, type);
lib/maple_tree.c
868
static inline unsigned char ma_meta_end(struct maple_node *mn,
lib/maple_tree.c
871
struct maple_metadata *meta = ma_meta(mn, mt);
lib/maple_tree.c
880
static inline unsigned char ma_meta_gap(struct maple_node *mn)
lib/maple_tree.c
882
return mn->ma64.meta.gap;
lib/maple_tree.c
891
static inline void ma_set_meta_gap(struct maple_node *mn, enum maple_type mt,
lib/maple_tree.c
895
struct maple_metadata *meta = ma_meta(mn, mt);
lib/test_maple_tree.c
2163
struct maple_enode *mn;
lib/test_maple_tree.c
2263
mn = mas.node;
lib/test_maple_tree.c
2269
MT_BUG_ON(mt, mn == mas.node);
mm/ksm.c
3376
struct memory_notify *mn = arg;
mm/ksm.c
3400
ksm_check_stable_tree(mn->start_pfn,
mm/ksm.c
3401
mn->start_pfn + mn->nr_pages);
mm/page_ext.c
427
struct memory_notify *mn = arg;
mm/page_ext.c
432
ret = online_page_ext(mn->start_pfn, mn->nr_pages);
mm/page_ext.c
435
offline_page_ext(mn->start_pfn,
mm/page_ext.c
436
mn->nr_pages);
mm/page_ext.c
439
offline_page_ext(mn->start_pfn,
mm/page_ext.c
440
mn->nr_pages);
net/sched/sch_teql.c
232
struct neighbour *mn;
net/sched/sch_teql.c
234
mn = __neigh_lookup_errno(n->tbl, n->primary_key, dev);
net/sched/sch_teql.c
236
if (IS_ERR(mn))
net/sched/sch_teql.c
237
return PTR_ERR(mn);
net/sched/sch_teql.c
238
n = mn;
tools/testing/radix-tree/maple.c
35105
struct maple_node *mn;
tools/testing/radix-tree/maple.c
35147
mn = mas_pop_node(&mas);
tools/testing/radix-tree/maple.c
35149
mn->parent = ma_parent_ptr(mn);
tools/testing/radix-tree/maple.c
35150
ma_free_rcu(mn);
tools/testing/radix-tree/maple.c
35161
mn = mas_pop_node(&mas);
tools/testing/radix-tree/maple.c
35167
mn->parent = ma_parent_ptr(mn);
tools/testing/radix-tree/maple.c
35168
ma_free_rcu(mn);
virt/kvm/kvm_main.c
505
static inline struct kvm *mmu_notifier_to_kvm(struct mmu_notifier *mn)
virt/kvm/kvm_main.c
507
return container_of(mn, struct kvm, mmu_notifier);
virt/kvm/kvm_main.c
650
static __always_inline int kvm_age_hva_range(struct mmu_notifier *mn,
virt/kvm/kvm_main.c
656
struct kvm *kvm = mmu_notifier_to_kvm(mn);
virt/kvm/kvm_main.c
670
static __always_inline int kvm_age_hva_range_no_flush(struct mmu_notifier *mn,
virt/kvm/kvm_main.c
675
return kvm_age_hva_range(mn, start, end, handler, false);
virt/kvm/kvm_main.c
726
static int kvm_mmu_notifier_invalidate_range_start(struct mmu_notifier *mn,
virt/kvm/kvm_main.c
729
struct kvm *kvm = mmu_notifier_to_kvm(mn);
virt/kvm/kvm_main.c
802
static void kvm_mmu_notifier_invalidate_range_end(struct mmu_notifier *mn,
virt/kvm/kvm_main.c
805
struct kvm *kvm = mmu_notifier_to_kvm(mn);
virt/kvm/kvm_main.c
833
static int kvm_mmu_notifier_clear_flush_young(struct mmu_notifier *mn,
virt/kvm/kvm_main.c
840
return kvm_age_hva_range(mn, start, end, kvm_age_gfn,
virt/kvm/kvm_main.c
844
static int kvm_mmu_notifier_clear_young(struct mmu_notifier *mn,
virt/kvm/kvm_main.c
864
return kvm_age_hva_range_no_flush(mn, start, end, kvm_age_gfn);
virt/kvm/kvm_main.c
867
static int kvm_mmu_notifier_test_young(struct mmu_notifier *mn,
virt/kvm/kvm_main.c
873
return kvm_age_hva_range_no_flush(mn, address, address + 1,
virt/kvm/kvm_main.c
877
static void kvm_mmu_notifier_release(struct mmu_notifier *mn,
virt/kvm/kvm_main.c
880
struct kvm *kvm = mmu_notifier_to_kvm(mn);