Symbol: pt_ops
arch/riscv/include/asm/pgtable.h
165
extern struct pt_alloc_ops pt_ops __meminitdata;
arch/riscv/mm/init.c
1015
pt_ops.alloc_pte = alloc_pte_early;
arch/riscv/mm/init.c
1016
pt_ops.get_pte_virt = get_pte_virt_early;
arch/riscv/mm/init.c
1018
pt_ops.alloc_pmd = alloc_pmd_early;
arch/riscv/mm/init.c
1019
pt_ops.get_pmd_virt = get_pmd_virt_early;
arch/riscv/mm/init.c
1020
pt_ops.alloc_pud = alloc_pud_early;
arch/riscv/mm/init.c
1021
pt_ops.get_pud_virt = get_pud_virt_early;
arch/riscv/mm/init.c
1022
pt_ops.alloc_p4d = alloc_p4d_early;
arch/riscv/mm/init.c
1023
pt_ops.get_p4d_virt = get_p4d_virt_early;
arch/riscv/mm/init.c
1037
pt_ops.alloc_pte = kernel_mapping_pa_to_va(alloc_pte_fixmap);
arch/riscv/mm/init.c
1038
pt_ops.get_pte_virt = kernel_mapping_pa_to_va(get_pte_virt_fixmap);
arch/riscv/mm/init.c
1040
pt_ops.alloc_pmd = kernel_mapping_pa_to_va(alloc_pmd_fixmap);
arch/riscv/mm/init.c
1041
pt_ops.get_pmd_virt = kernel_mapping_pa_to_va(get_pmd_virt_fixmap);
arch/riscv/mm/init.c
1042
pt_ops.alloc_pud = kernel_mapping_pa_to_va(alloc_pud_fixmap);
arch/riscv/mm/init.c
1043
pt_ops.get_pud_virt = kernel_mapping_pa_to_va(get_pud_virt_fixmap);
arch/riscv/mm/init.c
1044
pt_ops.alloc_p4d = kernel_mapping_pa_to_va(alloc_p4d_fixmap);
arch/riscv/mm/init.c
1045
pt_ops.get_p4d_virt = kernel_mapping_pa_to_va(get_p4d_virt_fixmap);
arch/riscv/mm/init.c
1055
pt_ops.alloc_pte = alloc_pte_late;
arch/riscv/mm/init.c
1056
pt_ops.get_pte_virt = get_pte_virt_late;
arch/riscv/mm/init.c
1058
pt_ops.alloc_pmd = alloc_pmd_late;
arch/riscv/mm/init.c
1059
pt_ops.get_pmd_virt = get_pmd_virt_late;
arch/riscv/mm/init.c
1060
pt_ops.alloc_pud = alloc_pud_late;
arch/riscv/mm/init.c
1061
pt_ops.get_pud_virt = get_pud_virt_late;
arch/riscv/mm/init.c
1062
pt_ops.alloc_p4d = alloc_p4d_late;
arch/riscv/mm/init.c
1063
pt_ops.get_p4d_virt = get_p4d_virt_late;
arch/riscv/mm/init.c
352
struct pt_alloc_ops pt_ops __meminitdata;
arch/riscv/mm/init.c
361
#define pt_ops (*(struct pt_alloc_ops *)XIP_FIXUP(&pt_ops))
arch/riscv/mm/init.c
542
pte_phys = pt_ops.alloc_pte(va);
arch/riscv/mm/init.c
544
ptep = pt_ops.get_pte_virt(pte_phys);
arch/riscv/mm/init.c
548
ptep = pt_ops.get_pte_virt(pte_phys);
arch/riscv/mm/init.c
644
next_phys = pt_ops.alloc_pmd(va);
arch/riscv/mm/init.c
646
nextp = pt_ops.get_pmd_virt(next_phys);
arch/riscv/mm/init.c
650
nextp = pt_ops.get_pmd_virt(next_phys);
arch/riscv/mm/init.c
670
next_phys = pt_ops.alloc_pud(va);
arch/riscv/mm/init.c
672
nextp = pt_ops.get_pud_virt(next_phys);
arch/riscv/mm/init.c
676
nextp = pt_ops.get_pud_virt(next_phys);
arch/riscv/mm/init.c
684
pt_ops.alloc_p4d(__va) : (pgtable_l4_enabled ? \
arch/riscv/mm/init.c
685
pt_ops.alloc_pud(__va) : pt_ops.alloc_pmd(__va)))
arch/riscv/mm/init.c
687
pt_ops.get_p4d_virt(__pa) : (pgd_next_t *)(pgtable_l4_enabled ? \
arch/riscv/mm/init.c
688
pt_ops.get_pud_virt(__pa) : (pud_t *)pt_ops.get_pmd_virt(__pa)))
arch/riscv/mm/init.c
703
#define alloc_pgd_next(__va) pt_ops.alloc_pte(__va)
arch/riscv/mm/init.c
704
#define get_pgd_next_virt(__pa) pt_ops.get_pte_virt(__pa)
arch/riscv/mm/kasan_init.c
174
base_pud = pt_ops.get_pud_virt(pfn_to_phys(_p4d_pfn(p4dp_get(p4dp))));
arch/riscv/mm/kasan_init.c
199
base_p4d = pt_ops.get_p4d_virt(pfn_to_phys(_pgd_pfn(pgdp_get(pgdp))));
arch/riscv/mm/kasan_init.c
245
base_pud = pt_ops.get_pud_virt(pfn_to_phys(_p4d_pfn(p4dp_get(p4dp))));
arch/riscv/mm/kasan_init.c
283
base_p4d = pt_ops.get_p4d_virt(pfn_to_phys(_pgd_pfn(pgdp_get(pgdp))));
drivers/gpu/drm/xe/tests/xe_migrate.c
249
expected = m->q->vm->pt_ops->pte_encode_bo(pt, 0, xe->pat.idx[XE_CACHE_WB], 0);
drivers/gpu/drm/xe/xe_ggtt.c
1010
pte = ggtt->pt_ops->ggtt_get_pte(ggtt, start);
drivers/gpu/drm/xe/xe_ggtt.c
1051
ggtt->pt_ops->ggtt_set_pte(ggtt, start, vfid_pte);
drivers/gpu/drm/xe/xe_ggtt.c
1131
return ggtt->pt_ops->pte_encode_flags(bo, pat_index);
drivers/gpu/drm/xe/xe_ggtt.c
137
const struct xe_ggtt_pt_ops *pt_ops;
drivers/gpu/drm/xe/xe_ggtt.c
270
ggtt->pt_ops->pte_encode_flags(ggtt->scratch,
drivers/gpu/drm/xe/xe_ggtt.c
276
ggtt->pt_ops->ggtt_set_pte(ggtt, start, scratch_pte);
drivers/gpu/drm/xe/xe_ggtt.c
430
ggtt->pt_ops =
drivers/gpu/drm/xe/xe_ggtt.c
435
ggtt->pt_ops = &xelp_pt_ops;
drivers/gpu/drm/xe/xe_ggtt.c
704
ggtt->pt_ops->ggtt_set_pte(ggtt, end - cur.remaining,
drivers/gpu/drm/xe/xe_ggtt.c
712
ggtt->pt_ops->ggtt_set_pte(ggtt, end - cur.remaining,
drivers/gpu/drm/xe/xe_ggtt.c
731
pte = ggtt->pt_ops->pte_encode_flags(bo, pat_index);
drivers/gpu/drm/xe/xe_ggtt.c
773
transform(ggtt, node, pte_flags, ggtt->pt_ops->ggtt_set_pte, arg);
drivers/gpu/drm/xe/xe_ggtt.c
845
u64 pte = ggtt->pt_ops->pte_encode_flags(bo, pat_index);
drivers/gpu/drm/xe/xe_ggtt.c
959
ggtt->pt_ops->ggtt_set_pte(ggtt, start, pte);
drivers/gpu/drm/xe/xe_migrate.c
154
flags = vm->pt_ops->pte_encode_addr(xe, 0, pat_index, level,
drivers/gpu/drm/xe/xe_migrate.c
167
entry = vm->pt_ops->pde_encode_bo(bo, pt_2m_ofs);
drivers/gpu/drm/xe/xe_migrate.c
170
flags = vm->pt_ops->pte_encode_addr(xe, 0,
drivers/gpu/drm/xe/xe_migrate.c
1932
addr = vm->pt_ops->pte_encode_bo(pt_bo, 0,
drivers/gpu/drm/xe/xe_migrate.c
2127
pte = m->q->vm->pt_ops->pte_encode_addr(m->tile->xe,
drivers/gpu/drm/xe/xe_migrate.c
231
entry = vm->pt_ops->pde_encode_bo(bo, pt29_ofs);
drivers/gpu/drm/xe/xe_migrate.c
238
entry = vm->pt_ops->pte_encode_bo(bo, i * XE_PAGE_SIZE,
drivers/gpu/drm/xe/xe_migrate.c
255
entry = vm->pt_ops->pte_encode_bo(batch, i,
drivers/gpu/drm/xe/xe_migrate.c
272
entry = vm->pt_ops->pte_encode_bo(batch, i,
drivers/gpu/drm/xe/xe_migrate.c
298
entry = vm->pt_ops->pde_encode_bo(bo, map_ofs + (u64)(level - 1) *
drivers/gpu/drm/xe/xe_migrate.c
306
entry = vm->pt_ops->pde_encode_bo(bo, (u64)i * XE_PAGE_SIZE);
drivers/gpu/drm/xe/xe_migrate.c
321
vm->pt_ops->pte_encode_addr(xe, 0, pat_index, level, IS_DGFX(xe), 0)
drivers/gpu/drm/xe/xe_migrate.c
676
addr = vm->pt_ops->pte_encode_addr(m->tile->xe,
drivers/gpu/drm/xe/xe_pt.c
548
pte = vm->pt_ops->pte_encode_vma(is_null_or_purged ? 0 :
drivers/gpu/drm/xe/xe_pt.c
628
pte = vm->pt_ops->pde_encode_bo(xe_child->bo, 0) | flags;
drivers/gpu/drm/xe/xe_pt.c
72
return vm->pt_ops->pde_encode_bo(vm->scratch_pt[id][level - 1]->bo,
drivers/gpu/drm/xe/xe_pt.c
75
return vm->pt_ops->pte_encode_addr(xe, 0, pat_index, level, IS_DGFX(xe), 0) |
drivers/gpu/drm/xe/xe_vm.c
1651
vm->pt_ops = &xelp_pt_ops;
drivers/gpu/drm/xe/xe_vm.c
2012
return vm->pt_ops->pde_encode_bo(vm->pt_root[tile->id]->bo, 0);
drivers/gpu/drm/xe/xe_vm_types.h
300
const struct xe_pt_ops *pt_ops;