Symbol: pte_index
arch/arm/mm/mmu.c
362
return &bm_pte[pte_index(addr)];
arch/arm64/include/asm/pgtable.h
855
#define pte_offset_phys(dir,addr) (pmd_page_paddr(READ_ONCE(*(dir))) + pte_index(addr) * sizeof(pte_t))
arch/arm64/mm/fixmap.c
40
return &bm_pte[BM_PTE_TABLE_IDX(addr)][pte_index(addr)];
arch/arm64/mm/mmu.c
220
ptep += pte_index(addr);
arch/arm64/mm/mmu.c
239
ptep += pte_index(next) - pte_index(addr);
arch/mips/kvm/mmu.c
157
int i_min = pte_index(start_gpa);
arch/mips/kvm/mmu.c
158
int i_max = pte_index(end_gpa);
arch/mips/kvm/mmu.c
282
int i_min = pte_index(start); \
arch/mips/kvm/mmu.c
283
int i_max = pte_index(end); \
arch/powerpc/include/asm/kvm_book3s.h
240
unsigned long *rmap, long pte_index, int realmode);
arch/powerpc/include/asm/kvm_book3s.h
244
unsigned long pte_index);
arch/powerpc/include/asm/kvm_book3s.h
246
unsigned long pte_index);
arch/powerpc/include/asm/kvm_book3s.h
252
long pte_index, unsigned long pteh, unsigned long ptel,
arch/powerpc/include/asm/kvm_book3s.h
255
unsigned long pte_index, unsigned long avpn,
arch/powerpc/include/asm/kvm_book3s_64.h
299
unsigned long pte_index)
arch/powerpc/include/asm/kvm_book3s_64.h
326
va_low = pte_index >> 3;
arch/powerpc/include/asm/kvm_host.h
471
unsigned long pte_index;
arch/powerpc/include/asm/kvm_ppc.h
830
long pte_index, unsigned long pteh, unsigned long ptel);
arch/powerpc/include/asm/kvm_ppc.h
832
unsigned long pte_index, unsigned long avpn);
arch/powerpc/include/asm/kvm_ppc.h
835
unsigned long pte_index, unsigned long avpn);
arch/powerpc/include/asm/kvm_ppc.h
837
unsigned long pte_index);
arch/powerpc/include/asm/kvm_ppc.h
839
unsigned long pte_index);
arch/powerpc/include/asm/kvm_ppc.h
841
unsigned long pte_index);
arch/powerpc/kvm/book3s_64_mmu_hv.c
293
long pte_index, unsigned long pteh,
arch/powerpc/kvm/book3s_64_mmu_hv.c
299
ret = kvmppc_do_h_enter(kvm, flags, pte_index, pteh, ptel,
arch/powerpc/kvm/book3s_64_mmu_hv.c
48
long pte_index, unsigned long pteh,
arch/powerpc/kvm/book3s_hv_rm_mmu.c
1017
unsigned long pte_index)
arch/powerpc/kvm/book3s_hv_rm_mmu.c
1029
rb = compute_tlbie_rb(hp0, hp1, pte_index);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
1035
unsigned long pte_index)
arch/powerpc/kvm/book3s_hv_rm_mmu.c
1047
rb = compute_tlbie_rb(hp0, hp1, pte_index);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
1211
index = cache_entry->pte_index;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
1287
cache_entry->pte_index = index;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
147
static void remove_revmap_chain(struct kvm *kvm, long pte_index,
arch/powerpc/kvm/book3s_hv_rm_mmu.c
170
if (head == pte_index) {
arch/powerpc/kvm/book3s_hv_rm_mmu.c
172
if (head == pte_index)
arch/powerpc/kvm/book3s_hv_rm_mmu.c
185
long pte_index, unsigned long pteh, unsigned long ptel,
arch/powerpc/kvm/book3s_hv_rm_mmu.c
303
if (pte_index >= kvmppc_hpt_npte(&kvm->arch.hpt))
arch/powerpc/kvm/book3s_hv_rm_mmu.c
306
pte_index &= ~7UL;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
307
hpte = (__be64 *)(kvm->arch.hpt.virt + (pte_index << 4));
arch/powerpc/kvm/book3s_hv_rm_mmu.c
336
pte_index += i;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
338
hpte = (__be64 *)(kvm->arch.hpt.virt + (pte_index << 4));
arch/powerpc/kvm/book3s_hv_rm_mmu.c
355
rev = &kvm->arch.hpt.rev[pte_index];
arch/powerpc/kvm/book3s_hv_rm_mmu.c
376
kvmppc_add_revmap_chain(kvm, rev, rmap, pte_index,
arch/powerpc/kvm/book3s_hv_rm_mmu.c
396
*pte_idx_ret = pte_index;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
402
long pte_index, unsigned long pteh, unsigned long ptel)
arch/powerpc/kvm/book3s_hv_rm_mmu.c
404
return kvmppc_do_h_enter(vcpu->kvm, flags, pte_index, pteh, ptel,
arch/powerpc/kvm/book3s_hv_rm_mmu.c
486
unsigned long pte_index, unsigned long avpn,
arch/powerpc/kvm/book3s_hv_rm_mmu.c
496
if (pte_index >= kvmppc_hpt_npte(&kvm->arch.hpt))
arch/powerpc/kvm/book3s_hv_rm_mmu.c
498
hpte = (__be64 *)(kvm->arch.hpt.virt + (pte_index << 4));
arch/powerpc/kvm/book3s_hv_rm_mmu.c
514
rev = real_vmalloc_addr(&kvm->arch.hpt.rev[pte_index]);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
518
rb = compute_tlbie_rb(v, pte_r, pte_index);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
529
remove_revmap_chain(kvm, pte_index, rev, v,
arch/powerpc/kvm/book3s_hv_rm_mmu.c
548
unsigned long pte_index, unsigned long avpn)
arch/powerpc/kvm/book3s_hv_rm_mmu.c
550
return kvmppc_do_h_remove(vcpu->kvm, flags, pte_index, avpn,
arch/powerpc/kvm/book3s_hv_rm_mmu.c
562
unsigned long flags, req, pte_index, rcbits;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
575
pte_index = args[j];
arch/powerpc/kvm/book3s_hv_rm_mmu.c
576
flags = pte_index >> 56;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
577
pte_index &= ((1ul << 56) - 1);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
585
pte_index >= kvmppc_hpt_npte(&kvm->arch.hpt)) {
arch/powerpc/kvm/book3s_hv_rm_mmu.c
587
args[j] = ((0xa0 | flags) << 56) + pte_index;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
591
hp = (__be64 *) (kvm->arch.hpt.virt + (pte_index << 4));
arch/powerpc/kvm/book3s_hv_rm_mmu.c
623
args[j] = ((0x90 | flags) << 56) + pte_index;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
627
args[j] = ((0x80 | flags) << 56) + pte_index;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
628
rev = real_vmalloc_addr(&kvm->arch.hpt.rev[pte_index]);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
643
tlbrb[n] = compute_tlbie_rb(hp0, hp1, pte_index);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
659
pte_index = args[j] & ((1ul << 56) - 1);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
662
remove_revmap_chain(kvm, pte_index, rev,
arch/powerpc/kvm/book3s_hv_rm_mmu.c
675
unsigned long pte_index, unsigned long avpn)
arch/powerpc/kvm/book3s_hv_rm_mmu.c
685
if (pte_index >= kvmppc_hpt_npte(&kvm->arch.hpt))
arch/powerpc/kvm/book3s_hv_rm_mmu.c
688
hpte = (__be64 *)(kvm->arch.hpt.virt + (pte_index << 4));
arch/powerpc/kvm/book3s_hv_rm_mmu.c
69
unsigned long *rmap, long pte_index, int realmode)
arch/powerpc/kvm/book3s_hv_rm_mmu.c
708
rev = real_vmalloc_addr(&kvm->arch.hpt.rev[pte_index]);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
727
rb = compute_tlbie_rb(v, r, pte_index);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
746
unsigned long pte_index)
arch/powerpc/kvm/book3s_hv_rm_mmu.c
756
if (pte_index >= kvmppc_hpt_npte(&kvm->arch.hpt))
arch/powerpc/kvm/book3s_hv_rm_mmu.c
759
pte_index &= ~3;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
762
rev = real_vmalloc_addr(&kvm->arch.hpt.rev[pte_index]);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
763
for (i = 0; i < n; ++i, ++pte_index) {
arch/powerpc/kvm/book3s_hv_rm_mmu.c
764
hpte = (__be64 *)(kvm->arch.hpt.virt + (pte_index << 4));
arch/powerpc/kvm/book3s_hv_rm_mmu.c
787
unsigned long pte_index)
arch/powerpc/kvm/book3s_hv_rm_mmu.c
798
if (pte_index >= kvmppc_hpt_npte(&kvm->arch.hpt))
arch/powerpc/kvm/book3s_hv_rm_mmu.c
801
rev = real_vmalloc_addr(&kvm->arch.hpt.rev[pte_index]);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
802
hpte = (__be64 *)(kvm->arch.hpt.virt + (pte_index << 4));
arch/powerpc/kvm/book3s_hv_rm_mmu.c
818
kvmppc_clear_ref_hpte(kvm, hpte, pte_index);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
836
unsigned long pte_index)
arch/powerpc/kvm/book3s_hv_rm_mmu.c
84
tail->forw = pte_index;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
846
if (pte_index >= kvmppc_hpt_npte(&kvm->arch.hpt))
arch/powerpc/kvm/book3s_hv_rm_mmu.c
849
rev = real_vmalloc_addr(&kvm->arch.hpt.rev[pte_index]);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
85
head->back = pte_index;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
850
hpte = (__be64 *)(kvm->arch.hpt.virt + (pte_index << 4));
arch/powerpc/kvm/book3s_hv_rm_mmu.c
866
kvmppc_invalidate_hpte(kvm, hpte, pte_index);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
87
rev->forw = rev->back = pte_index;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
89
pte_index | KVMPPC_RMAP_PRESENT | KVMPPC_RMAP_HPT;
arch/powerpc/kvm/book3s_pr_papr.c
113
rb = compute_tlbie_rb(pte[0], pte[1], pte_index);
arch/powerpc/kvm/book3s_pr_papr.c
217
unsigned long pte_index = kvmppc_get_gpr(vcpu, 5);
arch/powerpc/kvm/book3s_pr_papr.c
223
pteg = get_pteg_addr(vcpu, pte_index);
arch/powerpc/kvm/book3s_pr_papr.c
23
static unsigned long get_pteg_addr(struct kvm_vcpu *vcpu, long pte_index)
arch/powerpc/kvm/book3s_pr_papr.c
246
rb = compute_tlbie_rb(v, r, pte_index);
arch/powerpc/kvm/book3s_pr_papr.c
28
pte_index <<= 4;
arch/powerpc/kvm/book3s_pr_papr.c
29
pte_index &= ((1 << ((vcpu_book3s->sdr1 & 0x1f) + 11)) - 1) << 7 | 0x70;
arch/powerpc/kvm/book3s_pr_papr.c
31
pteg_addr |= pte_index;
arch/powerpc/kvm/book3s_pr_papr.c
39
long pte_index = kvmppc_get_gpr(vcpu, 5);
arch/powerpc/kvm/book3s_pr_papr.c
45
i = pte_index & 7;
arch/powerpc/kvm/book3s_pr_papr.c
46
pte_index &= ~7UL;
arch/powerpc/kvm/book3s_pr_papr.c
47
pteg_addr = get_pteg_addr(vcpu, pte_index);
arch/powerpc/kvm/book3s_pr_papr.c
76
kvmppc_set_gpr(vcpu, 4, pte_index | i);
arch/powerpc/kvm/book3s_pr_papr.c
89
unsigned long pte_index = kvmppc_get_gpr(vcpu, 5);
arch/powerpc/kvm/book3s_pr_papr.c
95
pteg = get_pteg_addr(vcpu, pte_index);
arch/powerpc/mm/book3s64/radix_pgtable.c
794
pte = pte_start + pte_index(addr);
arch/riscv/mm/init.c
1661
ptep = pte_base + pte_index(addr);
arch/riscv/mm/init.c
397
ptep = &fixmap_pte[pte_index(addr)];
arch/riscv/mm/init.c
452
uintptr_t pte_idx = pte_index(va);
arch/x86/boot/compressed/ident_map_64.c
229
return pte + pte_index(__address);
arch/x86/hyperv/hv_crash.c
528
pte = hv_crash_ptpgs[3] + pte_index(addr) * sizeof(pte);
arch/x86/mm/init_32.c
117
int pte_idx = pte_index(vaddr);
arch/x86/mm/init_32.c
342
pte_ofs = pte_index((pfn<<PAGE_SHIFT) + PAGE_OFFSET);
arch/x86/mm/init_64.c
1102
pte = pte_start + pte_index(addr);
arch/x86/mm/init_64.c
480
pte = pte_page + pte_index(paddr);
arch/x86/mm/init_64.c
481
i = pte_index(paddr);
arch/x86/mm/ioremap.c
840
return &bm_pte[pte_index(addr)];
arch/x86/power/hibernate_32.c
164
set_pte(pte + pte_index(restore_jump_address),
arch/x86/xen/mmu_pv.c
1940
pte = native_make_pte(xen_read_phys_ulong(pa + pte_index(vaddr) *
drivers/iommu/rockchip-iommu.c
538
u32 dte_index, pte_index, page_offset;
drivers/iommu/rockchip-iommu.c
550
pte_index = rk_iova_pte_index(iova);
drivers/iommu/rockchip-iommu.c
563
pte_addr_phys = rk_ops->pt_address(dte) + (pte_index * 4);
drivers/iommu/rockchip-iommu.c
575
&iova, dte_index, pte_index, page_offset);
drivers/iommu/rockchip-iommu.c
830
u32 dte_index, pte_index;
drivers/iommu/rockchip-iommu.c
849
pte_index = rk_iova_pte_index(iova);
drivers/iommu/rockchip-iommu.c
850
pte_addr = &page_table[pte_index];
drivers/iommu/rockchip-iommu.c
852
pte_dma = rk_ops->pt_address(dte_index) + pte_index * sizeof(u32);
drivers/iommu/sun50i-iommu.c
601
u32 pte_index;
drivers/iommu/sun50i-iommu.c
619
pte_index = sun50i_iova_get_pte_index(iova);
drivers/iommu/sun50i-iommu.c
620
pte_addr = &page_table[pte_index];
include/linux/pgtable.h
105
((pte_t *)kmap_local_page(pmd_page(*(pmd))) + pte_index((address)))
include/linux/pgtable.h
98
return (pte_t *)pmd_page_vaddr(*pmd) + pte_index(address);
mm/hugetlb_vmemmap.c
126
struct page *page = head ? head + pte_index(addr) :
mm/memory.c
2435
remaining_pages_total, PTRS_PER_PTE - pte_index(addr));
mm/memory.c
4654
if (can_swapin_thp(vmf, pte + pte_index(addr), 1 << order))
mm/memory.c
5169
if (pte_range_none(pte + pte_index(addr), 1 << order))
mm/memory.c
5628
pgoff_t pte_off = pte_index(vmf->address);
mm/memory.c
5736
pgoff_t pte_off = pte_index(vmf->address);
mm/vmscan.c
3527
for (i = pte_index(start), addr = start; addr != end; i++, addr += PAGE_SIZE) {
tools/testing/selftests/kvm/lib/arm64/processor.c
164
ptep = addr_gpa2hva(vm, pte_addr(vm, *ptep)) + pte_index(vm, vaddr) * 8;
tools/testing/selftests/kvm/lib/arm64/processor.c
213
ptep = addr_gpa2hva(vm, pte_addr(vm, *ptep)) + pte_index(vm, gva) * 8;
tools/testing/selftests/kvm/lib/riscv/processor.c
107
pte_index(vm, vaddr, level) * 8;
tools/testing/selftests/kvm/lib/riscv/processor.c
130
ptep = addr_gpa2hva(vm, vm->mmu.pgd) + pte_index(vm, gva, level) * 8;
tools/testing/selftests/kvm/lib/riscv/processor.c
137
pte_index(vm, gva, level) * 8;
tools/testing/selftests/kvm/lib/riscv/processor.c
97
ptep = addr_gpa2hva(vm, vm->mmu.pgd) + pte_index(vm, vaddr, level) * 8;