lib/libkvm/kvm_amd64.c
203
amd64_pte_t pte;
lib/libkvm/kvm_amd64.c
304
if (s < sizeof(pte)) {
lib/libkvm/kvm_amd64.c
308
if (pread(kd->pmfd, &pte, sizeof(pte), ofs) != sizeof(pte)) {
lib/libkvm/kvm_amd64.c
312
if ((pte & AMD64_PG_V) == 0) {
lib/libkvm/kvm_amd64.c
317
a = (pte & AMD64_PG_FRAME) + offset;
lib/libkvm/kvm_arm.c
188
#define l2pte_valid(pte) ((pte) != 0)
lib/libkvm/kvm_arm.c
197
arm_pt_entry_t pte;
lib/libkvm/kvm_arm.c
211
pte_pa = (pd & ARM_L1_C_ADDR_MASK) + l2pte_index(va) * sizeof(pte);
lib/libkvm/kvm_arm.c
213
if (pread(kd->pmfd, &pte, sizeof(pte), pte_off) != sizeof(pte)) {
lib/libkvm/kvm_arm.c
217
pte = _kvm32toh(kd, pte);
lib/libkvm/kvm_arm.c
218
if (!l2pte_valid(pte)) {
lib/libkvm/kvm_arm.c
221
if ((pte & ARM_L2_TYPE_MASK) == ARM_L2_TYPE_L) {
lib/libkvm/kvm_arm.c
222
*pa = (pte & ARM_L2_L_FRAME) | (va & ARM_L2_L_OFFSET);
lib/libkvm/kvm_arm.c
225
*pa = (pte & ARM_L2_S_FRAME) | (va & ARM_L2_S_OFFSET);
lib/libkvm/kvm_i386.c
218
i386_pte_t pte;
lib/libkvm/kvm_i386.c
269
pte_pa = (pde & I386_PG_FRAME) + (pteindex * sizeof(pte));
lib/libkvm/kvm_i386.c
272
if (s < sizeof(pte)) {
lib/libkvm/kvm_i386.c
278
if (pread(kd->pmfd, &pte, sizeof(pte), ofs) != sizeof(pte)) {
lib/libkvm/kvm_i386.c
282
pte = le32toh(pte);
lib/libkvm/kvm_i386.c
283
if ((pte & I386_PG_V) == 0) {
lib/libkvm/kvm_i386.c
288
a = (pte & I386_PG_FRAME) + offset;
lib/libkvm/kvm_i386.c
308
i386_pte_pae_t pte;
lib/libkvm/kvm_i386.c
362
if (s < sizeof(pte)) {
lib/libkvm/kvm_i386.c
368
if (pread(kd->pmfd, &pte, sizeof(pte), ofs) != sizeof(pte)) {
lib/libkvm/kvm_i386.c
372
pte = le64toh(pte);
lib/libkvm/kvm_i386.c
373
if ((pte & I386_PG_V) == 0) {
lib/libkvm/kvm_i386.c
378
a = (pte & I386_PG_FRAME_PAE) + offset;
lib/libkvm/kvm_minidump_aarch64.c
253
_aarch64_entry_to_prot(aarch64_pte_t pte)
lib/libkvm/kvm_minidump_aarch64.c
258
if ((pte & AARCH64_ATTR_AP(AARCH64_ATTR_AP_RO)) == 0)
lib/libkvm/kvm_minidump_aarch64.c
260
if ((pte & AARCH64_ATTR_XN) == 0)
lib/libkvm/kvm_minidump_aarch64.c
279
aarch64_pte_t pte = _aarch64_pte_get(kd, pteindex);
lib/libkvm/kvm_minidump_aarch64.c
281
if ((pte & AARCH64_ATTR_DESCR_MASK) != AARCH64_L3_PAGE)
lib/libkvm/kvm_minidump_aarch64.c
285
pa = pte & ~AARCH64_ATTR_MASK;
lib/libkvm/kvm_minidump_aarch64.c
288
_aarch64_entry_to_prot(pte), kd->vmst->page_size, 0)) {
lib/libkvm/kvm_minidump_aarch64.c
60
aarch64_pte_t *pte = _kvm_pmap_get(kd, pteindex, sizeof(*pte));
lib/libkvm/kvm_minidump_aarch64.c
62
return le64toh(*pte);
lib/libkvm/kvm_minidump_amd64.c
188
amd64_pte_t pte;
lib/libkvm/kvm_minidump_amd64.c
198
if (pteindex >= vm->hdr.pmapsize / sizeof(pte))
lib/libkvm/kvm_minidump_amd64.c
200
pte = _amd64_pte_get(kd, pteindex);
lib/libkvm/kvm_minidump_amd64.c
201
if ((pte & AMD64_PG_V) == 0) {
lib/libkvm/kvm_minidump_amd64.c
206
a = pte & AMD64_PG_FRAME;
lib/libkvm/kvm_minidump_amd64.c
246
amd64_pte_t pte;
lib/libkvm/kvm_minidump_amd64.c
284
pte = le64toh(pt[pteindex]);
lib/libkvm/kvm_minidump_amd64.c
285
if ((pte & AMD64_PG_V) == 0) {
lib/libkvm/kvm_minidump_amd64.c
290
a = pte & AMD64_PG_FRAME;
lib/libkvm/kvm_minidump_amd64.c
393
amd64_pte_t pte = (u_long)ptes[i];
lib/libkvm/kvm_minidump_amd64.c
395
pa = pte & AMD64_PG_FRAME;
lib/libkvm/kvm_minidump_amd64.c
397
if ((pte & AMD64_PG_V) != 0) {
lib/libkvm/kvm_minidump_amd64.c
401
_amd64_entry_to_prot(pte), pgsz, 0)) {
lib/libkvm/kvm_minidump_amd64.c
86
amd64_pte_t *pte = _kvm_pmap_get(kd, pteindex, sizeof(*pte));
lib/libkvm/kvm_minidump_amd64.c
88
return le64toh(*pte);
lib/libkvm/kvm_minidump_arm.c
153
arm_pt_entry_t pte;
lib/libkvm/kvm_minidump_arm.c
167
if (pteindex >= vm->hdr.ptesize / sizeof(pte))
lib/libkvm/kvm_minidump_arm.c
169
pte = _arm_pte_get(kd, pteindex);
lib/libkvm/kvm_minidump_arm.c
170
if ((pte & ARM_L2_TYPE_MASK) == ARM_L2_TYPE_INV) {
lib/libkvm/kvm_minidump_arm.c
175
if ((pte & ARM_L2_TYPE_MASK) == ARM_L2_TYPE_L) {
lib/libkvm/kvm_minidump_arm.c
178
a = (pte & ARM_L2_L_FRAME) +
lib/libkvm/kvm_minidump_arm.c
182
(pte & ARM_L2_TYPE_MASK) == ARM_L2_TYPE_T) {
lib/libkvm/kvm_minidump_arm.c
189
a = pte & ARM_L2_S_FRAME;
lib/libkvm/kvm_minidump_arm.c
212
_arm_entry_to_prot(kvm_t *kd, arm_pt_entry_t pte)
lib/libkvm/kvm_minidump_arm.c
219
if (pte & ARM_L2_S_PROT_W)
lib/libkvm/kvm_minidump_arm.c
225
if ((pte & ARM_PTE2_RO) == 0)
lib/libkvm/kvm_minidump_arm.c
227
if ((pte & ARM_PTE2_NX) == 0)
lib/libkvm/kvm_minidump_arm.c
240
arm_pt_entry_t pte = _arm_pte_get(kd, pteindex);
lib/libkvm/kvm_minidump_arm.c
242
if ((pte & ARM_L2_TYPE_MASK) == ARM_L2_TYPE_INV)
lib/libkvm/kvm_minidump_arm.c
246
if ((pte & ARM_L2_TYPE_MASK) == ARM_L2_TYPE_L) {
lib/libkvm/kvm_minidump_arm.c
248
pa = (pte & ARM_L2_L_FRAME) +
lib/libkvm/kvm_minidump_arm.c
252
(pte & ARM_L2_TYPE_MASK) == ARM_L2_TYPE_T) {
lib/libkvm/kvm_minidump_arm.c
256
pa = pte & ARM_L2_S_FRAME;
lib/libkvm/kvm_minidump_arm.c
261
_arm_entry_to_prot(kd, pte), ARM_PAGE_SIZE, 0))
lib/libkvm/kvm_minidump_arm.c
61
arm_pt_entry_t *pte = _kvm_pmap_get(kd, pteindex, sizeof(*pte));
lib/libkvm/kvm_minidump_arm.c
63
return _kvm32toh(kd, *pte);
lib/libkvm/kvm_minidump_i386.c
150
i386_pte_pae_t pte;
lib/libkvm/kvm_minidump_i386.c
160
if (pteindex >= vm->hdr.ptesize / sizeof(pte))
lib/libkvm/kvm_minidump_i386.c
162
pte = _i386_pte_pae_get(kd, pteindex);
lib/libkvm/kvm_minidump_i386.c
163
if ((pte & I386_PG_V) == 0) {
lib/libkvm/kvm_minidump_i386.c
168
a = pte & I386_PG_FRAME_PAE;
lib/libkvm/kvm_minidump_i386.c
195
i386_pte_t pte;
lib/libkvm/kvm_minidump_i386.c
205
if (pteindex >= vm->hdr.ptesize / sizeof(pte))
lib/libkvm/kvm_minidump_i386.c
207
pte = _i386_pte_get(kd, pteindex);
lib/libkvm/kvm_minidump_i386.c
208
if ((pte & I386_PG_V) == 0) {
lib/libkvm/kvm_minidump_i386.c
213
a = pte & I386_PG_FRAME;
lib/libkvm/kvm_minidump_i386.c
250
_i386_entry_to_prot(uint64_t pte)
lib/libkvm/kvm_minidump_i386.c
255
if (pte & I386_PG_RW)
lib/libkvm/kvm_minidump_i386.c
257
if ((pte & I386_PG_NX) == 0)
lib/libkvm/kvm_minidump_i386.c
58
i386_pte_pae_t *pte = _kvm_pmap_get(kd, pteindex, sizeof(*pte));
lib/libkvm/kvm_minidump_i386.c
60
return le64toh(*pte);
lib/libkvm/kvm_minidump_i386.c
66
i386_pte_t *pte = _kvm_pmap_get(kd, pteindex, sizeof(*pte));
lib/libkvm/kvm_minidump_i386.c
68
return le32toh(*pte);
lib/libkvm/kvm_minidump_powerpc64_hpt.c
311
ppc64_pt_entry_t pte, *p;
lib/libkvm/kvm_minidump_powerpc64_hpt.c
313
p = _kvm_pmap_get(kd, ptex, sizeof(pte));
lib/libkvm/kvm_minidump_powerpc64_hpt.c
314
pte.pte_hi = be64toh(p->pte_hi);
lib/libkvm/kvm_minidump_powerpc64_hpt.c
315
pte.pte_lo = be64toh(p->pte_lo);
lib/libkvm/kvm_minidump_powerpc64_hpt.c
316
return (pte);
lib/libkvm/kvm_minidump_powerpc64_hpt.c
328
ppc64_pt_entry_t pte;
lib/libkvm/kvm_minidump_powerpc64_hpt.c
379
pte = pte_get(kd, ptex);
lib/libkvm/kvm_minidump_powerpc64_hpt.c
382
if ((pte.pte_hi & LPTEH_HID) != hid ||
lib/libkvm/kvm_minidump_powerpc64_hpt.c
383
(pte.pte_hi & LPTEH_VALID) == 0 ||
lib/libkvm/kvm_minidump_powerpc64_hpt.c
384
(pte.pte_hi & LPTEH_B_MASK) != LPTEH_B_256MB)
lib/libkvm/kvm_minidump_powerpc64_hpt.c
388
if (PTEH_AVA_VSID(pte.pte_hi) != va_vsid ||
lib/libkvm/kvm_minidump_powerpc64_hpt.c
389
(PTEH_AVA_PAGE(pte.pte_hi) >> ava_pg_shift) !=
lib/libkvm/kvm_minidump_powerpc64_hpt.c
399
if (pte.pte_hi & LPTEH_BIG)
lib/libkvm/kvm_minidump_powerpc64_hpt.c
401
} else if ((pte.pte_hi & LPTEH_BIG) == 0)
lib/libkvm/kvm_minidump_powerpc64_hpt.c
407
__func__, (uintmax_t)ptex, (uintmax_t)pte.pte_hi,
lib/libkvm/kvm_minidump_powerpc64_hpt.c
408
(uintmax_t)pte.pte_lo);
lib/libkvm/kvm_minidump_powerpc64_hpt.c
425
*p = pte;
lib/libkvm/kvm_minidump_powerpc64_hpt.c
430
pte_lookup(kvm_t *kd, kvaddr_t ea, ppc64_pt_entry_t *pte)
lib/libkvm/kvm_minidump_powerpc64_hpt.c
439
return (pte_search(kd, slb, 0, ea, pte));
lib/libkvm/kvm_minidump_powerpc64_hpt.c
447
ppc64_pt_entry_t pte;
lib/libkvm/kvm_minidump_powerpc64_hpt.c
484
if ((err = pte_lookup(kd, va, &pte)) == -1) {
lib/libkvm/kvm_minidump_powerpc64_hpt.c
490
if (pte.pte_hi & LPTEH_BIG)
lib/libkvm/kvm_minidump_powerpc64_hpt.c
491
pgpa = (pte.pte_lo & PTEL_LP_PA_MASK) |
lib/libkvm/kvm_minidump_powerpc64_hpt.c
494
pgpa = pte.pte_lo & PTEL_PA_MASK;
lib/libkvm/kvm_minidump_powerpc64_hpt.c
519
entry_to_prot(ppc64_pt_entry_t *pte)
lib/libkvm/kvm_minidump_powerpc64_hpt.c
523
if (pte->pte_lo & LPTEL_RW)
lib/libkvm/kvm_minidump_powerpc64_hpt.c
525
if ((pte->pte_lo & LPTEL_NOEXEC) != 0)
lib/libkvm/kvm_minidump_powerpc64_hpt.c
561
get_ea(kvm_t *kd, ppc64_pt_entry_t *pte, u_long ptex)
lib/libkvm/kvm_minidump_powerpc64_hpt.c
568
vsid = PTEH_AVA_VSID(pte->pte_hi);
lib/libkvm/kvm_minidump_powerpc64_hpt.c
589
ea |= (PTEH_AVA_PAGE(pte->pte_hi) >> shift) <<
lib/libkvm/kvm_minidump_powerpc64_hpt.c
595
if (pte->pte_hi & LPTEH_HID)
lib/libkvm/kvm_minidump_powerpc64_hpt.c
617
ppc64_pt_entry_t pte = pte_get(kd, ptex);
lib/libkvm/kvm_minidump_powerpc64_hpt.c
618
if ((pte.pte_hi & LPTEH_VALID) == 0)
lib/libkvm/kvm_minidump_powerpc64_hpt.c
622
vsid = PTEH_AVA_VSID(pte.pte_hi);
lib/libkvm/kvm_minidump_powerpc64_hpt.c
628
if ((va = get_ea(kd, &pte, ptex)) == ~0UL)
lib/libkvm/kvm_minidump_powerpc64_hpt.c
632
if (pte.pte_hi & LPTEH_BIG) {
lib/libkvm/kvm_minidump_powerpc64_hpt.c
633
pa = pte.pte_lo & PTEL_LP_PA_MASK;
lib/libkvm/kvm_minidump_powerpc64_hpt.c
636
pa = pte.pte_lo & PTEL_PA_MASK;
lib/libkvm/kvm_minidump_powerpc64_hpt.c
644
entry_to_prot(&pte), pagesz, 0))
lib/libkvm/kvm_minidump_riscv.c
223
_riscv_entry_to_prot(riscv_pt_entry_t pte)
lib/libkvm/kvm_minidump_riscv.c
227
if ((pte & RISCV_PTE_W) != 0)
lib/libkvm/kvm_minidump_riscv.c
229
if ((pte & RISCV_PTE_X) != 0)
lib/libkvm/kvm_minidump_riscv.c
248
riscv_pt_entry_t pte = _riscv_pte_get(kd, pteindex);
lib/libkvm/kvm_minidump_riscv.c
250
if (((pte & RISCV_PTE_V) == 0) ||
lib/libkvm/kvm_minidump_riscv.c
251
((pte & RISCV_PTE_RWX) == 0))
lib/libkvm/kvm_minidump_riscv.c
255
pa = (pte >> RISCV_PTE_PPN0_S) << RISCV_L3_SHIFT;
lib/libkvm/kvm_minidump_riscv.c
258
_riscv_entry_to_prot(pte), RISCV_PAGE_SIZE, 0)) {
lib/libkvm/kvm_minidump_riscv.c
58
riscv_pt_entry_t *pte = _kvm_pmap_get(kd, pteindex, sizeof(*pte));
lib/libkvm/kvm_minidump_riscv.c
60
return le64toh(*pte);
lib/libvmmapi/vmmapi.c
938
vm_get_gpa_pmap(struct vmctx *ctx, uint64_t gpa, uint64_t *pte, int *num)
lib/libvmmapi/vmmapi.c
951
pte[i] = gpapte.pte[i];
lib/libvmmapi/vmmapi.h
134
int vm_get_gpa_pmap(struct vmctx *, uint64_t gpa, uint64_t *pte, int *num);
sys/amd64/amd64/efirt_machdep.c
122
pt_entry_t *pte;
sys/amd64/amd64/efirt_machdep.c
174
pte = (pt_entry_t *)PHYS_TO_DMAP(mphys);
sys/amd64/amd64/efirt_machdep.c
175
pte += pmap_pte_index(va);
sys/amd64/amd64/efirt_machdep.c
176
KASSERT(*pte == 0, ("va %#jx *pt %#jx", va, *pte));
sys/amd64/amd64/efirt_machdep.c
178
return (pte);
sys/amd64/amd64/efirt_machdep.c
185
pt_entry_t *pte;
sys/amd64/amd64/efirt_machdep.c
261
pte = efi_1t1_pte(va);
sys/amd64/amd64/efirt_machdep.c
262
pte_store(pte, va | bits);
sys/amd64/amd64/efirt_machdep.c
284
pte = efi_1t1_pte(0);
sys/amd64/amd64/efirt_machdep.c
288
pte_store(pte, bits);
sys/amd64/amd64/kexec_support.c
120
pt_entry_t *pte =
sys/amd64/amd64/kexec_support.c
122
pte[pmap_pte_index(pg)] =
sys/amd64/amd64/machdep.c
1029
*pte = pa | PG_V | PG_RW | PG_NC_PWT | PG_NC_PCD;
sys/amd64/amd64/machdep.c
1111
*pte = 0;
sys/amd64/amd64/machdep.c
875
pt_entry_t *pte;
sys/amd64/amd64/machdep.c
972
pte = CMAP1;
sys/amd64/amd64/minidump_machdep.c
168
uint64_t pdpe, pde, pte;
sys/amd64/amd64/minidump_machdep.c
245
pte = atomic_load_64(&pt[k]);
sys/amd64/amd64/minidump_machdep.c
246
if ((pte & PG_V) == 0)
sys/amd64/amd64/minidump_machdep.c
248
pa = pte & PG_FRAME;
sys/amd64/amd64/pmap.c
10021
pt_entry_t pte, PG_A, PG_M, PG_RW, PG_V;
sys/amd64/amd64/pmap.c
10031
pte = 0;
sys/amd64/amd64/pmap.c
10039
pte = *pdpe;
sys/amd64/amd64/pmap.c
10040
pa = ((pte & PG_PS_PDP_FRAME) | (addr & PDPMASK)) &
sys/amd64/amd64/pmap.c
10047
pte = *pdep;
sys/amd64/amd64/pmap.c
10049
pa = ((pte & PG_PS_FRAME) | (addr &
sys/amd64/amd64/pmap.c
10053
pte = *pmap_pde_to_pte(pdep, addr);
sys/amd64/amd64/pmap.c
10054
pa = pte & PG_FRAME;
sys/amd64/amd64/pmap.c
10060
if ((pte & PG_V) != 0) {
sys/amd64/amd64/pmap.c
10062
if ((pte & (PG_M | PG_RW)) == (PG_M | PG_RW))
sys/amd64/amd64/pmap.c
10064
if ((pte & PG_A) != 0)
sys/amd64/amd64/pmap.c
10069
(pte & (PG_MANAGED | PG_V)) == (PG_MANAGED | PG_V)) {
sys/amd64/amd64/pmap.c
10355
pt_entry_t *pte, PG_A, PG_M, PG_RW, PG_V;
sys/amd64/amd64/pmap.c
10387
pte = pmap_pde_to_pte(pde, va);
sys/amd64/amd64/pmap.c
10388
if ((*pte & PG_V) == 0)
sys/amd64/amd64/pmap.c
10392
if ((*pte & PG_RW) == 0)
sys/amd64/amd64/pmap.c
10403
*pte |= PG_M | PG_A;
sys/amd64/amd64/pmap.c
10405
*pte |= PG_A;
sys/amd64/amd64/pmap.c
10415
m = PHYS_TO_VM_PAGE(*pte & PG_FRAME);
sys/amd64/amd64/pmap.c
10447
pt_entry_t *pte, PG_V;
sys/amd64/amd64/pmap.c
10471
pte = pmap_pde_to_pte(pde, va);
sys/amd64/amd64/pmap.c
10472
ptr[idx++] = *pte;
sys/amd64/amd64/pmap.c
10719
pt_entry_t *pte, pt;
sys/amd64/amd64/pmap.c
10740
pte = pmap_pde_to_pte(pde, va);
sys/amd64/amd64/pmap.c
10741
pt = *pte;
sys/amd64/amd64/pmap.c
10743
("invalid pte va %#lx pte %#lx pt %#lx", va, (u_long)pte, pt));
sys/amd64/amd64/pmap.c
10767
pt_entry_t *pte;
sys/amd64/amd64/pmap.c
10829
pte = pmap_large_map_pte(va);
sys/amd64/amd64/pmap.c
10830
MPASS(*pte == 0);
sys/amd64/amd64/pmap.c
10831
*pte = pa | pg_g | X86_PG_RW | X86_PG_V |
sys/amd64/amd64/pmap.c
10834
PHYS_TO_VM_PAGE(DMAP_TO_PHYS((uintptr_t)pte))->
sys/amd64/amd64/pmap.c
10853
pt_entry_t *pte;
sys/amd64/amd64/pmap.c
10911
pte = pmap_pde_to_pte(pde, va);
sys/amd64/amd64/pmap.c
10912
KASSERT((*pte & X86_PG_V) != 0,
sys/amd64/amd64/pmap.c
10914
(u_long)pte, *pte));
sys/amd64/amd64/pmap.c
10915
pte_clear(pte);
sys/amd64/amd64/pmap.c
10917
m = PHYS_TO_VM_PAGE(DMAP_TO_PHYS((vm_offset_t)pte));
sys/amd64/amd64/pmap.c
11220
pmap_pti_wire_pte(void *pte)
sys/amd64/amd64/pmap.c
11225
m = PHYS_TO_VM_PAGE(DMAP_TO_PHYS((uintptr_t)pte));
sys/amd64/amd64/pmap.c
11241
pmap_pti_unwire_pte(void *pte, vm_offset_t va)
sys/amd64/amd64/pmap.c
11247
m = PHYS_TO_VM_PAGE(DMAP_TO_PHYS((uintptr_t)pte));
sys/amd64/amd64/pmap.c
11293
pt_entry_t *pte;
sys/amd64/amd64/pmap.c
11321
pte = (pt_entry_t *)PHYS_TO_DMAP(mphys);
sys/amd64/amd64/pmap.c
11322
pte += pmap_pte_index(va);
sys/amd64/amd64/pmap.c
11324
return (pte);
sys/amd64/amd64/pmap.c
11332
pt_entry_t *pte, ptev;
sys/amd64/amd64/pmap.c
11342
pte = pmap_pti_pte(sva, &unwire_pde);
sys/amd64/amd64/pmap.c
11347
if (*pte == 0) {
sys/amd64/amd64/pmap.c
11348
pte_store(pte, ptev);
sys/amd64/amd64/pmap.c
11349
pmap_pti_wire_pte(pte);
sys/amd64/amd64/pmap.c
11353
sva, *pte, ptev));
sys/amd64/amd64/pmap.c
11354
KASSERT(*pte == ptev,
sys/amd64/amd64/pmap.c
11356
sva, *pte, ptev));
sys/amd64/amd64/pmap.c
11379
pt_entry_t *pte;
sys/amd64/amd64/pmap.c
11390
pte = pmap_pti_pte(va, NULL);
sys/amd64/amd64/pmap.c
11391
KASSERT((*pte & X86_PG_V) != 0,
sys/amd64/amd64/pmap.c
11393
(u_long)pte, *pte));
sys/amd64/amd64/pmap.c
11394
pte_clear(pte);
sys/amd64/amd64/pmap.c
11395
pmap_pti_unwire_pte(pte, va);
sys/amd64/amd64/pmap.c
11471
pmap_pkru_same(pmap_t pmap, vm_offset_t sva, vm_offset_t eva, pt_entry_t *pte)
sys/amd64/amd64/pmap.c
11478
KASSERT(pmap->pm_type != PT_X86 || (*pte & X86_PG_PKU_MASK) == 0,
sys/amd64/amd64/pmap.c
11479
("pte %p has unexpected PKU %ld", pte, *pte & X86_PG_PKU_MASK));
sys/amd64/amd64/pmap.c
11494
*pte |= X86_PG_PKU(keyidx);
sys/amd64/amd64/pmap.c
11556
pt_entry_t newpte, *ptep, pte;
sys/amd64/amd64/pmap.c
11610
pte = *ptep;
sys/amd64/amd64/pmap.c
11611
if ((pte & X86_PG_V) == 0)
sys/amd64/amd64/pmap.c
11613
newpte = (pte & ~X86_PG_PKU_MASK) | X86_PG_PKU(keyidx);
sys/amd64/amd64/pmap.c
11614
if (newpte != pte) {
sys/amd64/amd64/pmap.c
11732
pt_entry_t *pte;
sys/amd64/amd64/pmap.c
11785
pte = (pt_entry_t *)(*pde & PG_FRAME) + pmap_pte_index(va);
sys/amd64/amd64/pmap.c
11786
if (*pte != 0)
sys/amd64/amd64/pmap.c
11789
*pte = (pt_entry_t)(pa | X86_PG_A | X86_PG_M | X86_PG_RW | X86_PG_V);
sys/amd64/amd64/pmap.c
11820
pt_entry_t *pte;
sys/amd64/amd64/pmap.c
11853
pte = pmap_pde_to_pte(pde, va);
sys/amd64/amd64/pmap.c
11854
if ((*pte & X86_PG_V) != 0)
sys/amd64/amd64/pmap.c
11857
*pte = (pt_entry_t)(VM_PAGE_TO_PHYS(m) | X86_PG_RW | X86_PG_V |
sys/amd64/amd64/pmap.c
11966
pd_entry_t pde, pt_entry_t pte)
sys/amd64/amd64/pmap.c
11988
} else if (pte != 0) {
sys/amd64/amd64/pmap.c
11989
attrs |= pte & pg_nx;
sys/amd64/amd64/pmap.c
11990
attrs &= pg_nx | (pte & (X86_PG_RW | X86_PG_U));
sys/amd64/amd64/pmap.c
11991
attrs |= pte & (X86_PG_G | X86_PG_PTE_CACHE);
sys/amd64/amd64/pmap.c
12013
pt_entry_t *pt, pte;
sys/amd64/amd64/pmap.c
12149
pte = pt[l];
sys/amd64/amd64/pmap.c
12150
if ((pte & X86_PG_V) == 0) {
sys/amd64/amd64/pmap.c
12156
pml5e, pml4e, pdpe, pde, pte);
sys/amd64/amd64/pmap.c
12173
DB_SHOW_COMMAND(pte, pmap_print_pte)
sys/amd64/amd64/pmap.c
12180
pt_entry_t *pte, PG_V;
sys/amd64/amd64/pmap.c
12225
pte = pmap_pde_to_pte(pde, va);
sys/amd64/amd64/pmap.c
12226
db_printf(" pte@0x%016lx 0x%016lx\n", (uint64_t)pte, *pte);
sys/amd64/amd64/pmap.c
12249
ptpages_show_complain(int level, int idx, uint64_t pte)
sys/amd64/amd64/pmap.c
12251
db_printf("l %d i %d pte %#lx\n", level, idx, pte);
sys/amd64/amd64/pmap.c
1376
static void pmap_pte_props(pt_entry_t *pte, u_long bits, u_long mask);
sys/amd64/amd64/pmap.c
1381
static void pmap_pti_wire_pte(void *pte);
sys/amd64/amd64/pmap.c
1555
pt_entry_t *pte;
sys/amd64/amd64/pmap.c
1559
pte = (pt_entry_t *)PHYS_TO_DMAP(*pde & PG_FRAME);
sys/amd64/amd64/pmap.c
1560
return (&pte[pmap_pte_index(va)]);
sys/amd64/amd64/pmap.c
2069
pt_entry_t *pte, *pcpu_pte;
sys/amd64/amd64/pmap.c
2173
v = (c)va; va += ((n)*PAGE_SIZE); p = pte; pte += (n);
sys/amd64/amd64/pmap.c
2176
pte = vtopte(va);
sys/amd64/amd64/pmap.c
2811
pmap_pat_index(pmap_t pmap, pt_entry_t pte, bool is_pde)
sys/amd64/amd64/pmap.c
2822
if ((pte & pat_flag) != 0)
sys/amd64/amd64/pmap.c
2824
if ((pte & PG_NC_PCD) != 0)
sys/amd64/amd64/pmap.c
2826
if ((pte & PG_NC_PWT) != 0)
sys/amd64/amd64/pmap.c
2830
if ((pte & EPT_PG_IGNORE_PAT) != 0)
sys/amd64/amd64/pmap.c
2831
panic("EPT PTE %#lx has no PAT memory type", pte);
sys/amd64/amd64/pmap.c
2832
pat_idx = (pte & EPT_PG_MEMORY_TYPE(0x7)) >> 3;
sys/amd64/amd64/pmap.c
312
safe_to_clear_referenced(pmap_t pmap, pt_entry_t pte)
sys/amd64/amd64/pmap.c
325
if ((pte & EPT_PG_WRITE) != 0)
sys/amd64/amd64/pmap.c
331
if ((pte & EPT_PG_EXECUTE) == 0 ||
sys/amd64/amd64/pmap.c
3707
pt_entry_t *pte;
sys/amd64/amd64/pmap.c
3730
pte = vtopte(vaddr);
sys/amd64/amd64/pmap.c
3733
pte_store(pte, spa | pte_bits);
sys/amd64/amd64/pmap.c
3753
pt_entry_t *pte, PG_V;
sys/amd64/amd64/pmap.c
3770
pte = pmap_pde_to_pte(pde, va);
sys/amd64/amd64/pmap.c
3771
pa = (*pte & PG_FRAME) |
sys/amd64/amd64/pmap.c
3793
pt_entry_t pte, PG_RW, PG_V;
sys/amd64/amd64/pmap.c
3821
pte = *pmap_pde_to_pte(pdep, va);
sys/amd64/amd64/pmap.c
3822
if ((pte & PG_V) == 0 ||
sys/amd64/amd64/pmap.c
3823
((pte & PG_RW) == 0 && (prot & VM_PROT_WRITE) != 0))
sys/amd64/amd64/pmap.c
3825
m = PHYS_TO_VM_PAGE(pte & PG_FRAME);
sys/amd64/amd64/pmap.c
3882
pt_entry_t *pte;
sys/amd64/amd64/pmap.c
3884
pte = vtopte(va);
sys/amd64/amd64/pmap.c
3885
pte_store(pte, pa | pg_g | pg_nx | X86_PG_A | X86_PG_M |
sys/amd64/amd64/pmap.c
3892
pt_entry_t *pte;
sys/amd64/amd64/pmap.c
3895
pte = vtopte(va);
sys/amd64/amd64/pmap.c
3897
pte_store(pte, pa | pg_g | pg_nx | X86_PG_A | X86_PG_M |
sys/amd64/amd64/pmap.c
3908
pt_entry_t *pte;
sys/amd64/amd64/pmap.c
3910
pte = vtopte(va);
sys/amd64/amd64/pmap.c
3911
pte_clear(pte);
sys/amd64/amd64/pmap.c
3944
pt_entry_t *endpte, oldpte, pa, *pte;
sys/amd64/amd64/pmap.c
3949
pte = vtopte(sva);
sys/amd64/amd64/pmap.c
3950
endpte = pte + count;
sys/amd64/amd64/pmap.c
3951
while (pte < endpte) {
sys/amd64/amd64/pmap.c
3955
if ((*pte & (PG_FRAME | X86_PG_PTE_CACHE)) != pa) {
sys/amd64/amd64/pmap.c
3956
oldpte |= *pte;
sys/amd64/amd64/pmap.c
3957
pte_store(pte, pa | pg_g | pg_nx | X86_PG_A |
sys/amd64/amd64/pmap.c
3960
pte++;
sys/amd64/amd64/pmap.c
4452
pmap_allocpte_free_unref(pmap_t pmap, vm_offset_t va, pt_entry_t *pte)
sys/amd64/amd64/pmap.c
4457
mpg = PHYS_TO_VM_PAGE(*pte & PG_FRAME);
sys/amd64/amd64/pmap.c
4907
pt_entry_t *pte;
sys/amd64/amd64/pmap.c
4923
pte = (pt_entry_t *)PHYS_TO_DMAP(dummypt);
sys/amd64/amd64/pmap.c
4925
pte[i] = (pt_entry_t)(dummypa | X86_PG_V | X86_PG_RW |
sys/amd64/amd64/pmap.c
5205
pt_entry_t *pte, tpte;
sys/amd64/amd64/pmap.c
5308
pte = pmap_pde_to_pte(pde, va);
sys/amd64/amd64/pmap.c
5309
if ((*pte & PG_W) != 0)
sys/amd64/amd64/pmap.c
5311
tpte = pte_load_clear(pte);
sys/amd64/amd64/pmap.c
5880
pt_entry_t *pte;
sys/amd64/amd64/pmap.c
5882
for (pte = firstpte; pte < firstpte + NPTEPG; pte++) {
sys/amd64/amd64/pmap.c
5883
*pte = newpte;
sys/amd64/amd64/pmap.c
6255
pt_entry_t *pte, PG_V;
sys/amd64/amd64/pmap.c
6261
pte = pmap_pde_to_pte(pde, va);
sys/amd64/amd64/pmap.c
6262
if ((*pte & PG_V) == 0)
sys/amd64/amd64/pmap.c
6265
pmap_remove_pte(pmap, pte, va, *pde, free, &lock);
sys/amd64/amd64/pmap.c
6278
pt_entry_t PG_G, *pte;
sys/amd64/amd64/pmap.c
6286
for (pte = pmap_pde_to_pte(pde, sva); sva != eva; pte++,
sys/amd64/amd64/pmap.c
6288
if (*pte == 0) {
sys/amd64/amd64/pmap.c
6295
if ((*pte & PG_G) == 0)
sys/amd64/amd64/pmap.c
6299
if (pmap_remove_pte(pmap, pte, sva, *pde, free, lockp)) {
sys/amd64/amd64/pmap.c
638
pt_entry_t *pte);
sys/amd64/amd64/pmap.c
6510
pt_entry_t *pte, tpte, PG_A, PG_M, PG_RW;
sys/amd64/amd64/pmap.c
6561
pte = pmap_pde_to_pte(pde, pv->pv_va);
sys/amd64/amd64/pmap.c
6562
tpte = pte_load_clear(pte);
sys/amd64/amd64/pmap.c
6646
pt_entry_t *pte, PG_G, PG_M, PG_RW, PG_V;
sys/amd64/amd64/pmap.c
6778
for (pte = pmap_pde_to_pte(pde, sva); sva != va_next; pte++,
sys/amd64/amd64/pmap.c
6781
obits = pbits = *pte;
sys/amd64/amd64/pmap.c
6797
if (!atomic_cmpset_long(pte, obits, pbits))
sys/amd64/amd64/pmap.c
6833
pt_entry_t *firstpte, oldpte, pa, *pte;
sys/amd64/amd64/pmap.c
6901
for (pte = firstpte + NPTEPG - 1; pte > firstpte; pte--) {
sys/amd64/amd64/pmap.c
6902
oldpte = *pte;
sys/amd64/amd64/pmap.c
6915
if (!atomic_fcmpset_long(pte, &oldpte, oldpte & ~PG_RW))
sys/amd64/amd64/pmap.c
7110
pt_entry_t *pte, PG_G, PG_A, PG_M, PG_RW, PG_V;
sys/amd64/amd64/pmap.c
7190
pte = pmap_pde_to_pte(pde, va);
sys/amd64/amd64/pmap.c
7211
origpte = *pte;
sys/amd64/amd64/pmap.c
7267
origpte = pte_load_clear(pte);
sys/amd64/amd64/pmap.c
7332
origpte = pte_load_store(pte, newpte);
sys/amd64/amd64/pmap.c
7354
pte_store(pte, newpte);
sys/amd64/amd64/pmap.c
7411
pt_entry_t *pt_end, *pte;
sys/amd64/amd64/pmap.c
7414
pte = (pt_entry_t *)PHYS_TO_DMAP(pa);
sys/amd64/amd64/pmap.c
7415
for (pt_end = pte + NPTEPG; pte < pt_end; pte++) {
sys/amd64/amd64/pmap.c
7416
if (*pte != 0)
sys/amd64/amd64/pmap.c
7705
pt_entry_t newpte, *pte, PG_V;
sys/amd64/amd64/pmap.c
7760
pte = (pt_entry_t *)PHYS_TO_DMAP(VM_PAGE_TO_PHYS(mpte));
sys/amd64/amd64/pmap.c
7761
pte = &pte[pmap_pte_index(va)];
sys/amd64/amd64/pmap.c
7764
pte = vtopte(va);
sys/amd64/amd64/pmap.c
7766
if (*pte) {
sys/amd64/amd64/pmap.c
7795
pte_store(pte, newpte);
sys/amd64/amd64/pmap.c
7951
pt_entry_t *pte, PG_V, PG_G __diagused;
sys/amd64/amd64/pmap.c
8008
for (pte = pmap_pde_to_pte(pde, sva); sva != va_next; pte++,
sys/amd64/amd64/pmap.c
8010
if ((*pte & PG_V) == 0)
sys/amd64/amd64/pmap.c
8012
if ((*pte & PG_W) == 0)
sys/amd64/amd64/pmap.c
8014
(uintmax_t)*pte);
sys/amd64/amd64/pmap.c
8021
atomic_clear_long(pte, PG_W);
sys/amd64/amd64/pmap.c
8387
pt_entry_t *pte;
sys/amd64/amd64/pmap.c
8409
pte = pmap_pte(pmap, pv->pv_va);
sys/amd64/amd64/pmap.c
8410
if ((*pte & PG_W) != 0)
sys/amd64/amd64/pmap.c
8430
pte = pmap_pde(pmap, pv->pv_va);
sys/amd64/amd64/pmap.c
8431
if ((*pte & PG_W) != 0)
sys/amd64/amd64/pmap.c
8490
pt_entry_t *pte, tpte;
sys/amd64/amd64/pmap.c
8550
pte = pmap_pdpe(pmap, pv->pv_va);
sys/amd64/amd64/pmap.c
8551
ptepde = *pte;
sys/amd64/amd64/pmap.c
8552
pte = pmap_pdpe_to_pde(pte, pv->pv_va);
sys/amd64/amd64/pmap.c
8553
tpte = *pte;
sys/amd64/amd64/pmap.c
8557
pte = (pt_entry_t *)PHYS_TO_DMAP(tpte &
sys/amd64/amd64/pmap.c
8559
pte = &pte[pmap_pte_index(pv->pv_va)];
sys/amd64/amd64/pmap.c
8560
tpte = *pte;
sys/amd64/amd64/pmap.c
8596
pte_clear(pte);
sys/amd64/amd64/pmap.c
8689
pt_entry_t *pte, mask;
sys/amd64/amd64/pmap.c
8711
pte = pmap_pte(pmap, pv->pv_va);
sys/amd64/amd64/pmap.c
8723
rv = (*pte & mask) == mask;
sys/amd64/amd64/pmap.c
8744
pte = pmap_pde(pmap, pv->pv_va);
sys/amd64/amd64/pmap.c
8756
rv = (*pte & mask) == mask;
sys/amd64/amd64/pmap.c
8798
pt_entry_t *pte, PG_V;
sys/amd64/amd64/pmap.c
8811
pte = pmap_pde_to_pte(pde, addr);
sys/amd64/amd64/pmap.c
8812
rv = (*pte & PG_V) == 0;
sys/amd64/amd64/pmap.c
8844
pt_entry_t oldpte, *pte, PG_M, PG_RW;
sys/amd64/amd64/pmap.c
8902
pte = pmap_pde_to_pte(pde, pv->pv_va);
sys/amd64/amd64/pmap.c
8903
oldpte = *pte;
sys/amd64/amd64/pmap.c
8905
while (!atomic_fcmpset_long(pte, &oldpte, oldpte &
sys/amd64/amd64/pmap.c
8947
pt_entry_t *pte, PG_A, PG_M, PG_RW;
sys/amd64/amd64/pmap.c
9034
pte = pmap_pde_to_pte(pde, va);
sys/amd64/amd64/pmap.c
9035
pmap_remove_pte(pmap, pte, va, *pde,
sys/amd64/amd64/pmap.c
9094
pte = pmap_pde_to_pte(pde, pv->pv_va);
sys/amd64/amd64/pmap.c
9095
if ((*pte & (PG_M | PG_RW)) == (PG_M | PG_RW))
sys/amd64/amd64/pmap.c
9097
if ((*pte & PG_A) != 0) {
sys/amd64/amd64/pmap.c
9098
if (safe_to_clear_referenced(pmap, *pte)) {
sys/amd64/amd64/pmap.c
9099
atomic_clear_long(pte, PG_A);
sys/amd64/amd64/pmap.c
9102
} else if ((*pte & PG_W) == 0) {
sys/amd64/amd64/pmap.c
9109
pmap_remove_pte(pmap, pte, pv->pv_va,
sys/amd64/amd64/pmap.c
9149
pt_entry_t *pte, PG_A, PG_G, PG_M, PG_RW, PG_V;
sys/amd64/amd64/pmap.c
9229
pte = pmap_pde_to_pte(pde, va);
sys/amd64/amd64/pmap.c
9230
KASSERT((*pte & PG_V) != 0,
sys/amd64/amd64/pmap.c
9232
pmap_remove_pte(pmap, pte, va, *pde, NULL,
sys/amd64/amd64/pmap.c
9242
for (pte = pmap_pde_to_pte(pde, sva); sva != va_next; pte++,
sys/amd64/amd64/pmap.c
9244
if ((*pte & (PG_MANAGED | PG_V)) != (PG_MANAGED | PG_V))
sys/amd64/amd64/pmap.c
9246
else if ((*pte & (PG_M | PG_RW)) == (PG_M | PG_RW)) {
sys/amd64/amd64/pmap.c
9253
m = PHYS_TO_VM_PAGE(*pte & PG_FRAME);
sys/amd64/amd64/pmap.c
9256
atomic_clear_long(pte, PG_M | PG_A);
sys/amd64/amd64/pmap.c
9257
} else if ((*pte & PG_A) != 0)
sys/amd64/amd64/pmap.c
9258
atomic_clear_long(pte, PG_A);
sys/amd64/amd64/pmap.c
9262
if ((*pte & PG_G) != 0) {
sys/amd64/amd64/pmap.c
9293
pt_entry_t *pte, PG_M, PG_RW;
sys/amd64/amd64/pmap.c
9335
pte = pmap_pde_to_pte(pde, va);
sys/amd64/amd64/pmap.c
9336
atomic_clear_long(pte, PG_M | PG_RW);
sys/amd64/amd64/pmap.c
9360
pte = pmap_pde_to_pte(pde, pv->pv_va);
sys/amd64/amd64/pmap.c
9361
if ((*pte & (PG_M | PG_RW)) == (PG_M | PG_RW)) {
sys/amd64/amd64/pmap.c
9362
atomic_clear_long(pte, PG_M);
sys/amd64/amd64/pmap.c
9376
pmap_pte_props(pt_entry_t *pte, u_long bits, u_long mask)
sys/amd64/amd64/pmap.c
9380
opte = *(u_long *)pte;
sys/amd64/amd64/pmap.c
9384
} while (npte != opte && !atomic_fcmpset_long((u_long *)pte, &opte,
sys/amd64/amd64/pmap.c
9699
pt_entry_t *pte, pte_bits, pte_mask;
sys/amd64/amd64/pmap.c
9808
pte = pmap_pde_to_pte(pde, tmpva);
sys/amd64/amd64/pmap.c
9809
if (*pte == 0) {
sys/amd64/amd64/pmap.c
9884
pte = pmap_pde_to_pte(pde, tmpva);
sys/amd64/amd64/pmap.c
9885
if ((*pte & pte_mask) != pte_bits) {
sys/amd64/amd64/pmap.c
9886
pmap_pte_props(pte, pte_bits, pte_mask);
sys/amd64/amd64/pmap.c
9890
(*pte & PG_FRAME) < dmaplimit) {
sys/amd64/amd64/pmap.c
9893
pa_start = *pte & PG_FRAME;
sys/amd64/amd64/pmap.c
9895
} else if (pa_end == (*pte & PG_FRAME))
sys/amd64/amd64/pmap.c
9906
pa_start = *pte & PG_FRAME;
sys/amd64/include/pmap.h
285
#define pte_load_store(ptep, pte) atomic_swap_long(ptep, pte)
sys/amd64/include/pmap.h
287
#define pte_store(ptep, pte) do { \
sys/amd64/include/pmap.h
288
*(u_long *)(ptep) = (u_long)(pte); \
sys/amd64/include/vmm_dev.h
191
uint64_t pte[4]; /* out */
sys/amd64/vmm/vmm_dev_machdep.c
486
gpapte->pte, &gpapte->ptenum);
sys/amd64/vmm/vmm_instruction_emul.c
2027
pf_error_code(int usermode, int prot, int rsvd, uint64_t pte)
sys/amd64/vmm/vmm_instruction_emul.c
2031
if (pte & PG_V)
sys/amd64/vmm/vmm_instruction_emul.c
2070
uint64_t *ptpbase, ptpphys, pte, pgsize;
sys/amd64/vmm/vmm_instruction_emul.c
2177
pte = ptpbase[ptpindex];
sys/amd64/vmm/vmm_instruction_emul.c
2179
if ((pte & PG_V) == 0) {
sys/amd64/vmm/vmm_instruction_emul.c
2181
pfcode = pf_error_code(usermode, prot, 0, pte);
sys/amd64/vmm/vmm_instruction_emul.c
2187
ptpphys = pte;
sys/amd64/vmm/vmm_instruction_emul.c
2208
pte = ptpbase[ptpindex];
sys/amd64/vmm/vmm_instruction_emul.c
2210
if ((pte & PG_V) == 0 ||
sys/amd64/vmm/vmm_instruction_emul.c
2211
(usermode && (pte & PG_U) == 0) ||
sys/amd64/vmm/vmm_instruction_emul.c
2212
(writable && (pte & PG_RW) == 0)) {
sys/amd64/vmm/vmm_instruction_emul.c
2214
pfcode = pf_error_code(usermode, prot, 0, pte);
sys/amd64/vmm/vmm_instruction_emul.c
2221
if (!check_only && (pte & PG_A) == 0) {
sys/amd64/vmm/vmm_instruction_emul.c
2223
pte, pte | PG_A) == 0) {
sys/amd64/vmm/vmm_instruction_emul.c
2228
if (nlevels > 0 && (pte & PG_PS) != 0) {
sys/amd64/vmm/vmm_instruction_emul.c
2232
pte);
sys/amd64/vmm/vmm_instruction_emul.c
2240
ptpphys = pte;
sys/amd64/vmm/vmm_instruction_emul.c
2244
if (!check_only && writable && (pte & PG_M) == 0) {
sys/amd64/vmm/vmm_instruction_emul.c
2245
if (atomic_cmpset_64(&ptpbase[ptpindex], pte, pte | PG_M) == 0)
sys/amd64/vmm/vmm_instruction_emul.c
2250
pte >>= ptpshift; pte <<= (ptpshift + 12); pte >>= 12;
sys/amd64/vmm/vmm_instruction_emul.c
2251
*gpa = pte | (gla & (pgsize - 1));
sys/arm64/arm64/kexec_support.c
164
pt_entry_t *pte;
sys/arm64/arm64/kexec_support.c
172
pte = pagetable_l0_ttbr0_bootstrap;
sys/arm64/arm64/kexec_support.c
173
pte -= (Ln_ENTRIES * 2); /* move to start of L2 pages */
sys/arm64/arm64/kexec_support.c
180
pte[pmap_l2_index(tmp)] = (tmp | L2_BLOCK | ATTR_AF | ATTR_S1_UXN);
sys/arm64/arm64/kexec_support.c
182
pte[pmap_l2_index(tmp)] = (tmp | L2_BLOCK | ATTR_AF | ATTR_S1_UXN);
sys/arm64/arm64/kexec_support.c
185
pte[pmap_l2_index(tmp)] = (tmp | L2_BLOCK | ATTR_AF | ATTR_S1_UXN);
sys/arm64/arm64/pmap.c
2087
pt_entry_t *pte, tpte;
sys/arm64/arm64/pmap.c
2097
pte = pmap_pte(pmap, va, &lvl);
sys/arm64/arm64/pmap.c
2098
if (pte != NULL) {
sys/arm64/arm64/pmap.c
2099
tpte = pmap_load(pte);
sys/arm64/arm64/pmap.c
2137
pt_entry_t *pte, tpte;
sys/arm64/arm64/pmap.c
2145
pte = pmap_pte(pmap, va, &lvl);
sys/arm64/arm64/pmap.c
2146
if (pte != NULL) {
sys/arm64/arm64/pmap.c
2147
tpte = pmap_load(pte);
sys/arm64/arm64/pmap.c
2198
pt_entry_t *pte, tpte;
sys/arm64/arm64/pmap.c
2209
pte = pmap_l1(kernel_pmap, sva);
sys/arm64/arm64/pmap.c
2210
if (pte == NULL)
sys/arm64/arm64/pmap.c
2212
tpte = pmap_load(pte);
sys/arm64/arm64/pmap.c
2220
pte = pmap_l1_to_l2(&tpte, sva);
sys/arm64/arm64/pmap.c
2221
tpte = pmap_load(pte);
sys/arm64/arm64/pmap.c
2228
pte = pmap_l2_to_l3(&tpte, sva);
sys/arm64/arm64/pmap.c
2229
tpte = pmap_load(pte);
sys/arm64/arm64/pmap.c
2253
pt_entry_t *pte, tpte;
sys/arm64/arm64/pmap.c
2278
pte = pmap_l1(kernel_pmap, va);
sys/arm64/arm64/pmap.c
2279
if (pte == NULL)
sys/arm64/arm64/pmap.c
2288
tpte = pmap_load(pte);
sys/arm64/arm64/pmap.c
2296
pte = pmap_l1_to_l2(&tpte, va);
sys/arm64/arm64/pmap.c
2297
tpte = pmap_load(pte);
sys/arm64/arm64/pmap.c
2305
pte = pmap_l2_to_l3(&tpte, va);
sys/arm64/arm64/pmap.c
2306
tpte = pmap_load(pte);
sys/arm64/arm64/pmap.c
2341
pt_entry_t attr, old_l3e, *pte;
sys/arm64/arm64/pmap.c
2409
pte = pmap_l2_to_l3(pde, va);
sys/arm64/arm64/pmap.c
2410
old_l3e |= pmap_load_store(pte, PHYS_TO_PTE(pa) | attr |
sys/arm64/arm64/pmap.c
2441
pt_entry_t *pte;
sys/arm64/arm64/pmap.c
2443
pte = pmap_pte_exists(kernel_pmap, va, 3, __func__);
sys/arm64/arm64/pmap.c
2444
KASSERT((pmap_load(pte) & ATTR_CONTIGUOUS) == 0,
sys/arm64/arm64/pmap.c
2446
pmap_clear(pte);
sys/arm64/arm64/pmap.c
2553
pt_entry_t attr, old_l3e, *pte;
sys/arm64/arm64/pmap.c
2571
pte = pmap_l2_to_l3(pde, va);
sys/arm64/arm64/pmap.c
2572
old_l3e |= pmap_load_store(pte, VM_PAGE_TO_PTE(m) | attr);
sys/arm64/arm64/pmap.c
2594
pt_entry_t *pte;
sys/arm64/arm64/pmap.c
2603
pte = pmap_pte_exists(kernel_pmap, va, 3, NULL);
sys/arm64/arm64/pmap.c
2604
if (pte != NULL) {
sys/arm64/arm64/pmap.c
2605
pmap_clear(pte);
sys/arm64/arm64/pmap.c
290
#define PTE_TO_VM_PAGE(pte) PHYS_TO_VM_PAGE(PTE_TO_PHYS(pte))
sys/arm64/arm64/pmap.c
3022
pt_entry_t *pte;
sys/arm64/arm64/pmap.c
3047
pte = pmap_l0_to_l1(pde, va);
sys/arm64/arm64/pmap.c
3048
KASSERT(pmap_load(pte) == 0,
sys/arm64/arm64/pmap.c
3054
pte = pmap_l1_to_l2(pde, va);
sys/arm64/arm64/pmap.c
3055
KASSERT(pmap_load(pte) == 0,
sys/arm64/arm64/pmap.c
3304
pt_entry_t *pte, tpte;
sys/arm64/arm64/pmap.c
3386
pte = pmap_l2_to_l3(pde, va);
sys/arm64/arm64/pmap.c
3387
tpte = pmap_load(pte);
sys/arm64/arm64/pmap.c
3391
(void)pmap_demote_l3c(pmap, pte, va);
sys/arm64/arm64/pmap.c
3392
tpte = pmap_load_clear(pte);
sys/arm64/arm64/pmap.c
4390
pt_entry_t *pte, tpte;
sys/arm64/arm64/pmap.c
4415
pte = pmap_pte_exists(pmap, va, 2, __func__);
sys/arm64/arm64/pmap.c
4416
pmap_demote_l2_locked(pmap, pte, va, &lock);
sys/arm64/arm64/pmap.c
4441
pte = pmap_l2_to_l3(pde, pv->pv_va);
sys/arm64/arm64/pmap.c
4442
tpte = pmap_load(pte);
sys/arm64/arm64/pmap.c
4444
(void)pmap_demote_l3c(pmap, pte, pv->pv_va);
sys/arm64/arm64/pmap.c
4445
tpte = pmap_load_clear(pte);
sys/arm64/arm64/pmap.c
521
static void pmap_update_entry(pmap_t pmap, pd_entry_t *pte, pd_entry_t newpte,
sys/arm64/arm64/pmap.c
5216
pmap_enter_largepage(pmap_t pmap, vm_offset_t va, pt_entry_t pte, int flags,
sys/arm64/arm64/pmap.c
5225
KASSERT((PTE_TO_PHYS(pte) & (pagesizes[psind] - 1)) == 0,
sys/arm64/arm64/pmap.c
5227
PTE_TO_PHYS(pte), pte, psind));
sys/arm64/arm64/pmap.c
5230
newpte = pte;
sys/arm64/arm64/pmap.c
527
pt_entry_t *pte);
sys/arm64/arm64/pmap.c
5750
pt_entry_t *pt_end, *pte;
sys/arm64/arm64/pmap.c
5753
pte = (pt_entry_t *)PHYS_TO_DMAP(pa);
sys/arm64/arm64/pmap.c
5754
for (pt_end = pte + Ln_ENTRIES; pte < pt_end; pte++) {
sys/arm64/arm64/pmap.c
5755
if (*pte != 0)
sys/arm64/arm64/pmap.c
6995
pt_entry_t *pte;
sys/arm64/arm64/pmap.c
7017
pte = pmap_pte_exists(pmap, pv->pv_va, 3, __func__);
sys/arm64/arm64/pmap.c
7018
if ((pmap_load(pte) & ATTR_SW_WIRED) != 0)
sys/arm64/arm64/pmap.c
7038
pte = pmap_pte_exists(pmap, pv->pv_va, 2, __func__);
sys/arm64/arm64/pmap.c
7039
if ((pmap_load(pte) & ATTR_SW_WIRED) != 0)
sys/arm64/arm64/pmap.c
7089
pt_entry_t *pte, tpte;
sys/arm64/arm64/pmap.c
7127
pte = pmap_l1_to_l2(pde, pv->pv_va);
sys/arm64/arm64/pmap.c
7128
tpte = pmap_load(pte);
sys/arm64/arm64/pmap.c
7135
pte = pmap_l2_to_l3(pde, pv->pv_va);
sys/arm64/arm64/pmap.c
7136
tpte = pmap_load(pte);
sys/arm64/arm64/pmap.c
7168
pmap_clear(pte);
sys/arm64/arm64/pmap.c
7278
pt_entry_t l3e, mask, *pte, value;
sys/arm64/arm64/pmap.c
7300
pte = pmap_pte_exists(pmap, pv->pv_va, 3, __func__);
sys/arm64/arm64/pmap.c
7311
l3e = pmap_load(pte);
sys/arm64/arm64/pmap.c
7313
l3e = pmap_load_l3c(pte);
sys/arm64/arm64/pmap.c
7336
pte = pmap_pte_exists(pmap, pv->pv_va, 2, __func__);
sys/arm64/arm64/pmap.c
7347
rv = (pmap_load(pte) & mask) == value;
sys/arm64/arm64/pmap.c
7389
pt_entry_t *pte;
sys/arm64/arm64/pmap.c
7401
pte = pmap_l2_to_l3(pde, addr);
sys/arm64/arm64/pmap.c
7402
rv = pmap_load(pte) == 0;
sys/arm64/arm64/pmap.c
7433
pt_entry_t oldpte, *pte, set, clear, mask, val;
sys/arm64/arm64/pmap.c
7461
pte = pmap_pte_exists(pmap, va, 2, __func__);
sys/arm64/arm64/pmap.c
7462
if ((pmap_load(pte) & ATTR_SW_DBM) != 0)
sys/arm64/arm64/pmap.c
7463
(void)pmap_demote_l2_locked(pmap, pte, va, &lock);
sys/arm64/arm64/pmap.c
7483
pte = pmap_pte_exists(pmap, pv->pv_va, 3, __func__);
sys/arm64/arm64/pmap.c
7484
oldpte = pmap_load(pte);
sys/arm64/arm64/pmap.c
7487
(void)pmap_demote_l3c(pmap, pte, pv->pv_va);
sys/arm64/arm64/pmap.c
7493
oldpte = pmap_load(pte);
sys/arm64/arm64/pmap.c
7507
while (!atomic_fcmpset_64(pte, &oldpte,
sys/arm64/arm64/pmap.c
7544
pt_entry_t *pte, tpte;
sys/arm64/arm64/pmap.c
7577
pte = pmap_pte_exists(pmap, va, 2, __func__);
sys/arm64/arm64/pmap.c
7578
tpte = pmap_load(pte);
sys/arm64/arm64/pmap.c
7611
pmap_clear_bits(pte, ATTR_AF);
sys/arm64/arm64/pmap.c
7646
pte = pmap_pte_exists(pmap, pv->pv_va, 3, __func__);
sys/arm64/arm64/pmap.c
7647
tpte = pmap_load(pte);
sys/arm64/arm64/pmap.c
7656
pmap_clear_bits(pte, ATTR_AF);
sys/arm64/arm64/pmap.c
7662
(pmap_load_l3c(pte) & ATTR_AF) != 0) {
sys/arm64/arm64/pmap.c
8232
pt_entry_t pte, *ptep, *newpte;
sys/arm64/arm64/pmap.c
8364
pte = pmap_load(ptep);
sys/arm64/arm64/pmap.c
8365
pte &= ~mask;
sys/arm64/arm64/pmap.c
8366
pte |= bits;
sys/arm64/arm64/pmap.c
8371
L2C_ENTRIES, pte, tmpva, L2_SIZE, L2C_SIZE);
sys/arm64/arm64/pmap.c
8375
L3C_ENTRIES, pte, tmpva, L3_SIZE, L3C_SIZE);
sys/arm64/arm64/pmap.c
8384
pmap_update_entry(kernel_pmap, ptep, pte, tmpva,
sys/arm64/arm64/pmap.c
8389
pa = PTE_TO_PHYS(pte);
sys/arm64/arm64/pmap.c
8939
pt_entry_t *pte, tpte;
sys/arm64/arm64/pmap.c
8946
pte = pmap_pte(pmap, addr, &lvl);
sys/arm64/arm64/pmap.c
8947
if (pte != NULL) {
sys/arm64/arm64/pmap.c
8948
tpte = pmap_load(pte);
sys/arm64/arm64/pmap.c
909
pmap_pte_dirty(pmap_t pmap, pt_entry_t pte)
sys/arm64/arm64/pmap.c
912
KASSERT((pte & ATTR_SW_MANAGED) != 0, ("pte %#lx is unmanaged", pte));
sys/arm64/arm64/pmap.c
915
KASSERT((pte & (ATTR_S1_AP_RW_BIT | ATTR_SW_DBM)) != 0,
sys/arm64/arm64/pmap.c
916
("pte %#lx is writeable and missing ATTR_SW_DBM", pte));
sys/arm64/arm64/pmap.c
918
return ((pte & (ATTR_S1_AP_RW_BIT | ATTR_SW_DBM)) ==
sys/arm64/arm64/pmap.c
922
return ((pte & ATTR_S2_S2AP(ATTR_S2_S2AP_WRITE)) ==
sys/arm64/arm64/pmap.c
9336
pt_entry_t *ptep, pte;
sys/arm64/arm64/pmap.c
9377
if (ptep != NULL && (pte = pmap_load(ptep)) != 0) {
sys/arm64/arm64/pmap.c
9385
if ((pte & ATTR_S2_XN_MASK) !=
sys/arm64/arm64/pmap.c
9402
pt_entry_t pte, *ptep;
sys/arm64/arm64/pmap.c
9449
((pte = pmap_load(ptep)) & ATTR_SW_DBM) != 0) {
sys/arm64/arm64/pmap.c
9450
if ((pte & ATTR_S1_AP_RW_BIT) ==
sys/arm64/arm64/pmap.c
9689
pmap_bti_same(pmap_t pmap, vm_offset_t sva, vm_offset_t eva, pt_entry_t *pte)
sys/arm64/arm64/pmap.c
9699
KASSERT((*pte & ATTR_S1_GP) == 0,
sys/arm64/arm64/pmap.c
9700
("%s: pte %lx has ATTR_S1_GP preset", __func__, *pte));
sys/arm64/arm64/pmap.c
9703
*pte |= ATTR_KERN_GP;
sys/arm64/arm64/pmap.c
9716
*pte |= ATTR_S1_GP;
sys/arm64/iommu/iommu_pmap.c
666
pt_entry_t *pte;
sys/arm64/iommu/iommu_pmap.c
680
pte = smmu_pmap_l2_to_l3(pde, va);
sys/arm64/iommu/iommu_pmap.c
683
smmu_pmap_clear(pte);
sys/arm64/iommu/iommu_pmap.c
684
cpu_dcache_wb_range(pte, sizeof(pt_entry_t));
sys/arm64/iommu/iommu_pmap.c
762
pt_entry_t *pte;
sys/arm64/iommu/iommu_pmap.c
768
pte = smmu_pmap_pte(pmap, va, &lvl);
sys/arm64/iommu/iommu_pmap.c
772
if (pte != NULL) {
sys/arm64/iommu/iommu_pmap.c
774
smmu_pmap_clear(pte);
sys/arm64/vmm/vmm_arm64.c
1003
pte = ptep[idx];
sys/arm64/vmm/vmm_arm64.c
1011
if ((pte & ATTR_DESCR_MASK) != L0_TABLE)
sys/arm64/vmm/vmm_arm64.c
1016
switch (pte & ATTR_DESCR_MASK) {
sys/arm64/vmm/vmm_arm64.c
1021
(pte & TATTR_AP_TABLE_NO_EL0) != 0)
sys/arm64/vmm/vmm_arm64.c
1025
(pte & TATTR_AP_TABLE_RO) != 0)
sys/arm64/vmm/vmm_arm64.c
1030
(pte & TATTR_UXN_TABLE) != 0) ||
sys/arm64/vmm/vmm_arm64.c
1032
(pte & TATTR_PXN_TABLE) != 0))
sys/arm64/vmm/vmm_arm64.c
1035
pte_addr = pte & ~ATTR_MASK;
sys/arm64/vmm/vmm_arm64.c
1044
if ((pte & ATTR_DESCR_MASK) == L3_PAGE)
sys/arm64/vmm/vmm_arm64.c
1052
if (is_el0 && (pte & ATTR_S1_AP(ATTR_S1_AP_USER)) == 0)
sys/arm64/vmm/vmm_arm64.c
1054
if ((prot & PROT_WRITE) != 0 && (pte & ATTR_S1_AP_RW_BIT) != 0)
sys/arm64/vmm/vmm_arm64.c
1057
if ((is_el0 && (pte & ATTR_S1_UXN) != 0) ||
sys/arm64/vmm/vmm_arm64.c
1058
(!is_el0 && (pte & ATTR_S1_PXN) != 0))
sys/arm64/vmm/vmm_arm64.c
1062
*gpa = (pte & ~ATTR_MASK) | (gla & mask);
sys/arm64/vmm/vmm_arm64.c
859
uint64_t mask, *ptep, pte, pte_addr;
sys/arm64/vmm/vmm_arm64.c
995
while (idx > PAGE_SIZE / sizeof(pte)) {
sys/arm64/vmm/vmm_arm64.c
996
idx -= PAGE_SIZE / sizeof(pte);
sys/dev/agp/agp_i810.c
131
static void agp_i810_write_gtt(device_t dev, u_int index, uint32_t pte);
sys/dev/agp/agp_i810.c
132
static void agp_i915_write_gtt(device_t dev, u_int index, uint32_t pte);
sys/dev/agp/agp_i810.c
133
static void agp_i965_write_gtt(device_t dev, u_int index, uint32_t pte);
sys/dev/agp/agp_i810.c
134
static void agp_g4x_write_gtt(device_t dev, u_int index, uint32_t pte);
sys/dev/agp/agp_i810.c
1474
uint32_t pte;
sys/dev/agp/agp_i810.c
1476
pte = (u_int32_t)physical | I810_PTE_VALID;
sys/dev/agp/agp_i810.c
1478
pte |= I810_PTE_LOCAL;
sys/dev/agp/agp_i810.c
1480
pte |= I830_PTE_SYSTEM_CACHED;
sys/dev/agp/agp_i810.c
1481
agp_i810_write_gtt(dev, index, pte);
sys/dev/agp/agp_i810.c
1485
agp_i810_write_gtt(device_t dev, u_int index, uint32_t pte)
sys/dev/agp/agp_i810.c
1490
bus_write_4(sc->sc_res[0], AGP_I810_GTT + index * 4, pte);
sys/dev/agp/agp_i810.c
1491
CTR2(KTR_AGP_I810, "810_pte %x %x", index, pte);
sys/dev/agp/agp_i810.c
1498
uint32_t pte;
sys/dev/agp/agp_i810.c
1500
pte = (u_int32_t)physical | I810_PTE_VALID;
sys/dev/agp/agp_i810.c
1502
pte |= I830_PTE_SYSTEM_CACHED;
sys/dev/agp/agp_i810.c
1503
agp_i810_write_gtt(dev, index, pte);
sys/dev/agp/agp_i810.c
1510
uint32_t pte;
sys/dev/agp/agp_i810.c
1512
pte = (u_int32_t)physical | I810_PTE_VALID;
sys/dev/agp/agp_i810.c
1514
pte |= I830_PTE_SYSTEM_CACHED;
sys/dev/agp/agp_i810.c
1515
pte |= (physical & 0x0000000f00000000ull) >> 28;
sys/dev/agp/agp_i810.c
1516
agp_i915_write_gtt(dev, index, pte);
sys/dev/agp/agp_i810.c
1520
agp_i915_write_gtt(device_t dev, u_int index, uint32_t pte)
sys/dev/agp/agp_i810.c
1525
bus_write_4(sc->sc_res[1], index * 4, pte);
sys/dev/agp/agp_i810.c
1526
CTR2(KTR_AGP_I810, "915_pte %x %x", index, pte);
sys/dev/agp/agp_i810.c
1533
uint32_t pte;
sys/dev/agp/agp_i810.c
1535
pte = (u_int32_t)physical | I810_PTE_VALID;
sys/dev/agp/agp_i810.c
1537
pte |= I830_PTE_SYSTEM_CACHED;
sys/dev/agp/agp_i810.c
1538
pte |= (physical & 0x0000000f00000000ull) >> 28;
sys/dev/agp/agp_i810.c
1539
agp_i965_write_gtt(dev, index, pte);
sys/dev/agp/agp_i810.c
1543
agp_i965_write_gtt(device_t dev, u_int index, uint32_t pte)
sys/dev/agp/agp_i810.c
1548
bus_write_4(sc->sc_res[0], index * 4 + (512 * 1024), pte);
sys/dev/agp/agp_i810.c
1549
CTR2(KTR_AGP_I810, "965_pte %x %x", index, pte);
sys/dev/agp/agp_i810.c
1556
uint32_t pte;
sys/dev/agp/agp_i810.c
1558
pte = (u_int32_t)physical | I810_PTE_VALID;
sys/dev/agp/agp_i810.c
1560
pte |= I830_PTE_SYSTEM_CACHED;
sys/dev/agp/agp_i810.c
1561
pte |= (physical & 0x0000000f00000000ull) >> 28;
sys/dev/agp/agp_i810.c
1562
agp_g4x_write_gtt(dev, index, pte);
sys/dev/agp/agp_i810.c
1566
agp_g4x_write_gtt(device_t dev, u_int index, uint32_t pte)
sys/dev/agp/agp_i810.c
1571
bus_write_4(sc->sc_res[0], index * 4 + (2 * 1024 * 1024), pte);
sys/dev/agp/agp_i810.c
1572
CTR2(KTR_AGP_I810, "g4x_pte %x %x", index, pte);
sys/dev/agp/agp_i810.c
1618
u_int32_t pte;
sys/dev/agp/agp_i810.c
1621
pte = bus_read_4(sc->sc_res[0], AGP_I810_GTT + index * 4);
sys/dev/agp/agp_i810.c
1622
return (pte);
sys/dev/agp/agp_i810.c
1629
u_int32_t pte;
sys/dev/agp/agp_i810.c
1632
pte = bus_read_4(sc->sc_res[1], index * 4);
sys/dev/agp/agp_i810.c
1633
return (pte);
sys/dev/agp/agp_i810.c
1640
u_int32_t pte;
sys/dev/agp/agp_i810.c
1643
pte = bus_read_4(sc->sc_res[0], index * 4 + (512 * 1024));
sys/dev/agp/agp_i810.c
1644
return (pte);
sys/dev/agp/agp_i810.c
1651
u_int32_t pte;
sys/dev/agp/agp_i810.c
1654
pte = bus_read_4(sc->sc_res[0], index * 4 + (2 * 1024 * 1024));
sys/dev/agp/agp_i810.c
1655
return (pte);
sys/dev/agp/agp_i810.c
1662
u_int32_t pte;
sys/dev/agp/agp_i810.c
1666
pte = sc->match->driver->read_gtt_pte(dev, index);
sys/dev/agp/agp_i810.c
1667
res = pte & ~PAGE_MASK;
sys/dev/agp/agp_i810.c
1675
u_int32_t pte;
sys/dev/agp/agp_i810.c
1679
pte = sc->match->driver->read_gtt_pte(dev, index);
sys/dev/agp/agp_i810.c
1680
res = (pte & ~PAGE_MASK) | ((pte & 0xf0) << 28);
sys/dev/xen/grant_table/grant_table.c
479
map_pte_fn(pte_t *pte, struct page *pmd_page,
sys/dev/xen/grant_table/grant_table.c
484
set_pte_at(&init_mm, addr, pte, pfn_pte_ma((*frames)[0], PAGE_KERNEL));
sys/dev/xen/grant_table/grant_table.c
490
unmap_pte_fn(pte_t *pte, struct page *pmd_page,
sys/dev/xen/grant_table/grant_table.c
494
set_pte_at(&init_mm, addr, pte, __pte(0));
sys/i386/i386/minidump_machdep_base.c
159
pt_entry_t *pt, pte;
sys/i386/i386/minidump_machdep_base.c
198
pte = pte_load(&pt[k]);
sys/i386/i386/minidump_machdep_base.c
199
if ((pte & PG_V) == PG_V) {
sys/i386/i386/minidump_machdep_base.c
200
pa = pte & PG_FRAME;
sys/i386/i386/pmap.c
1114
pmap_pat_index(pmap_t pmap, pt_entry_t pte, bool is_pde)
sys/i386/i386/pmap.c
1125
if ((pte & pat_flag) != 0)
sys/i386/i386/pmap.c
1127
if ((pte & PG_NC_PCD) != 0)
sys/i386/i386/pmap.c
1129
if ((pte & PG_NC_PWT) != 0)
sys/i386/i386/pmap.c
1496
__CONCAT(PMTYPE, pte)(pmap_t pmap, vm_offset_t va)
sys/i386/i386/pmap.c
1525
pmap_pte_release(pt_entry_t *pte)
sys/i386/i386/pmap.c
1528
if ((pt_entry_t *)((vm_offset_t)pte & ~PAGE_MASK) == PADDR2)
sys/i386/i386/pmap.c
1630
pt_entry_t *eh_ptep, pte, *ptep;
sys/i386/i386/pmap.c
1642
pte = *ptep;
sys/i386/i386/pmap.c
1644
return (pte);
sys/i386/i386/pmap.c
1684
pt_entry_t pte;
sys/i386/i386/pmap.c
1694
pte = pmap_pte_ufast(pmap, va, pde);
sys/i386/i386/pmap.c
1695
rtval = (pte & PG_FRAME) | (va & PAGE_MASK);
sys/i386/i386/pmap.c
1713
pt_entry_t pte;
sys/i386/i386/pmap.c
1725
pte = pmap_pte_ufast(pmap, va, pde);
sys/i386/i386/pmap.c
1726
if (pte != 0 &&
sys/i386/i386/pmap.c
1727
((pte & PG_RW) || (prot & VM_PROT_WRITE) == 0))
sys/i386/i386/pmap.c
1728
m = PHYS_TO_VM_PAGE(pte & PG_FRAME);
sys/i386/i386/pmap.c
1750
pt_entry_t *pte;
sys/i386/i386/pmap.c
1752
pte = vtopte(va);
sys/i386/i386/pmap.c
1753
pte_store(pte, pa | PG_RW | PG_V);
sys/i386/i386/pmap.c
1759
pt_entry_t *pte;
sys/i386/i386/pmap.c
1761
pte = vtopte(va);
sys/i386/i386/pmap.c
1762
pte_store(pte, pa | PG_RW | PG_V | pmap_cache_bits(kernel_pmap,
sys/i386/i386/pmap.c
1775
pt_entry_t *pte;
sys/i386/i386/pmap.c
1777
pte = vtopte(va);
sys/i386/i386/pmap.c
1778
pte_clear(pte);
sys/i386/i386/pmap.c
1850
pt_entry_t *endpte, oldpte, pa, *pte;
sys/i386/i386/pmap.c
1854
pte = vtopte(sva);
sys/i386/i386/pmap.c
1855
endpte = pte + count;
sys/i386/i386/pmap.c
1856
while (pte < endpte) {
sys/i386/i386/pmap.c
1860
if ((*pte & (PG_FRAME | PG_PTE_CACHE)) != pa) {
sys/i386/i386/pmap.c
1861
oldpte |= *pte;
sys/i386/i386/pmap.c
1862
pte_store(pte, pa | pg_nx | PG_RW | PG_V);
sys/i386/i386/pmap.c
1864
pte++;
sys/i386/i386/pmap.c
195
#define pmap_pde_v(pte) ((*(int *)pte & PG_V) != 0)
sys/i386/i386/pmap.c
196
#define pmap_pte_w(pte) ((*(int *)pte & PG_W) != 0)
sys/i386/i386/pmap.c
197
#define pmap_pte_m(pte) ((*(int *)pte & PG_M) != 0)
sys/i386/i386/pmap.c
198
#define pmap_pte_u(pte) ((*(int *)pte & PG_A) != 0)
sys/i386/i386/pmap.c
199
#define pmap_pte_v(pte) ((*(int *)pte & PG_V) != 0)
sys/i386/i386/pmap.c
201
#define pmap_pte_set_w(pte, v) ((v) ? atomic_set_int((u_int *)(pte), PG_W) : \
sys/i386/i386/pmap.c
202
atomic_clear_int((u_int *)(pte), PG_W))
sys/i386/i386/pmap.c
203
#define pmap_pte_set_prot(pte, v) ((*(int *)pte &= ~PG_PROT), (*(int *)pte |= (v)))
sys/i386/i386/pmap.c
2310
pt_entry_t *pte, tpte;
sys/i386/i386/pmap.c
2356
pte = __CONCAT(PMTYPE, pte)(pmap, va);
sys/i386/i386/pmap.c
2357
tpte = *pte;
sys/i386/i386/pmap.c
2359
tpte = pte_load_clear(pte);
sys/i386/i386/pmap.c
2360
pmap_pte_release(pte);
sys/i386/i386/pmap.c
2741
pt_entry_t *pte;
sys/i386/i386/pmap.c
2743
for (pte = firstpte; pte < firstpte + NPTEPG; pte++) {
sys/i386/i386/pmap.c
2744
*pte = newpte;
sys/i386/i386/pmap.c
3037
pt_entry_t *pte;
sys/i386/i386/pmap.c
3042
if ((pte = pmap_pte_quick(pmap, va)) == NULL || *pte == 0)
sys/i386/i386/pmap.c
3044
pmap_remove_pte(pmap, pte, va, free);
sys/i386/i386/pmap.c
3055
pt_entry_t *pte;
sys/i386/i386/pmap.c
3062
for (pte = pmap_pte_quick(pmap, sva); sva != eva; pte++,
sys/i386/i386/pmap.c
3064
if (*pte == 0)
sys/i386/i386/pmap.c
3071
if ((*pte & PG_G) == 0)
sys/i386/i386/pmap.c
3074
if (pmap_remove_pte(pmap, pte, sva, free))
sys/i386/i386/pmap.c
3204
pt_entry_t *pte, tpte;
sys/i386/i386/pmap.c
3233
pte = pmap_pte_quick(pmap, pv->pv_va);
sys/i386/i386/pmap.c
3234
tpte = pte_load_clear(pte);
sys/i386/i386/pmap.c
324
static void pmap_pte_attr(pt_entry_t *pte, int cache_bits);
sys/i386/i386/pmap.c
3314
pt_entry_t *pte;
sys/i386/i386/pmap.c
3404
for (pte = pmap_pte_quick(pmap, sva); sva != pdnxt; pte++,
sys/i386/i386/pmap.c
3414
obits = pbits = *pte;
sys/i386/i386/pmap.c
3433
if (!atomic_cmpset_64(pte, obits, pbits))
sys/i386/i386/pmap.c
3436
if (!atomic_cmpset_int((u_int *)pte, obits,
sys/i386/i386/pmap.c
346
static void pmap_pte_release(pt_entry_t *pte);
sys/i386/i386/pmap.c
3473
pt_entry_t allpte_PG_A, *firstpte, oldpte, pa, *pte;
sys/i386/i386/pmap.c
3539
for (pte = firstpte + NPTEPG - 1; pte > firstpte; pte--) {
sys/i386/i386/pmap.c
3541
oldpte = *pte;
sys/i386/i386/pmap.c
3553
if (!atomic_cmpset_int((u_int *)pte, oldpte,
sys/i386/i386/pmap.c
3649
pt_entry_t *pte;
sys/i386/i386/pmap.c
3729
pte = pmap_pte_quick(pmap, va);
sys/i386/i386/pmap.c
3736
if (pte == NULL) {
sys/i386/i386/pmap.c
3741
origpte = *pte;
sys/i386/i386/pmap.c
3795
origpte = pte_load_clear(pte);
sys/i386/i386/pmap.c
3857
origpte = pte_load_store(pte, newpte);
sys/i386/i386/pmap.c
3882
pte_store_zero(pte, newpte);
sys/i386/i386/pmap.c
3938
pt_entry_t *pt_end, *pte;
sys/i386/i386/pmap.c
3941
pte = vtopte(va);
sys/i386/i386/pmap.c
3942
for (pt_end = pte + NPTEPG; pte < pt_end; pte++) {
sys/i386/i386/pmap.c
3943
if (*pte != 0)
sys/i386/i386/pmap.c
4156
pt_entry_t newpte, *pte;
sys/i386/i386/pmap.c
4208
pte = pmap_pte_quick(pmap, va);
sys/i386/i386/pmap.c
4209
if (*pte) {
sys/i386/i386/pmap.c
4242
pte_store_zero(pte, newpte);
sys/i386/i386/pmap.c
4372
pt_entry_t *pte;
sys/i386/i386/pmap.c
4426
for (pte = pmap_pte_quick(pmap, sva); sva != pdnxt; pte++,
sys/i386/i386/pmap.c
4428
if ((*pte & PG_V) == 0)
sys/i386/i386/pmap.c
4430
if ((*pte & PG_W) == 0)
sys/i386/i386/pmap.c
4432
(uintmax_t)*pte);
sys/i386/i386/pmap.c
4441
atomic_clear_int((u_int *)pte, PG_W);
sys/i386/i386/pmap.c
4795
pt_entry_t *pte;
sys/i386/i386/pmap.c
4803
pte = pmap_pte_quick(pmap, pv->pv_va);
sys/i386/i386/pmap.c
4804
if ((*pte & PG_W) != 0)
sys/i386/i386/pmap.c
4842
pt_entry_t *pte, tpte;
sys/i386/i386/pmap.c
4874
pte = pmap_pde(pmap, pv->pv_va);
sys/i386/i386/pmap.c
4875
tpte = *pte;
sys/i386/i386/pmap.c
4877
pte = pmap_pte_quick(pmap, pv->pv_va);
sys/i386/i386/pmap.c
4878
tpte = *pte & ~PG_PTE_PAT;
sys/i386/i386/pmap.c
4884
pte, pv->pv_va);
sys/i386/i386/pmap.c
4907
pte_clear(pte);
sys/i386/i386/pmap.c
5005
pt_entry_t *pte;
sys/i386/i386/pmap.c
5015
pte = pmap_pte_quick(pmap, pv->pv_va);
sys/i386/i386/pmap.c
5016
rv = (*pte & (PG_M | PG_RW)) == (PG_M | PG_RW);
sys/i386/i386/pmap.c
5075
pt_entry_t *pte;
sys/i386/i386/pmap.c
5085
pte = pmap_pte_quick(pmap, pv->pv_va);
sys/i386/i386/pmap.c
5086
rv = (*pte & (PG_A | PG_V)) == (PG_A | PG_V);
sys/i386/i386/pmap.c
5105
pt_entry_t oldpte, *pte;
sys/i386/i386/pmap.c
5135
pte = pmap_pte_quick(pmap, pv->pv_va);
sys/i386/i386/pmap.c
5137
oldpte = *pte;
sys/i386/i386/pmap.c
5144
if (!atomic_cmpset_int((u_int *)pte, oldpte,
sys/i386/i386/pmap.c
5181
pt_entry_t *pte;
sys/i386/i386/pmap.c
5254
pte = pmap_pte_quick(pmap, pv->pv_va);
sys/i386/i386/pmap.c
5255
if ((*pte & (PG_M | PG_RW)) == (PG_M | PG_RW))
sys/i386/i386/pmap.c
5257
if ((*pte & PG_A) != 0) {
sys/i386/i386/pmap.c
5258
atomic_clear_int((u_int *)pte, PG_A);
sys/i386/i386/pmap.c
5286
pt_entry_t *pte;
sys/i386/i386/pmap.c
5347
pte = pmap_pte_quick(pmap, va);
sys/i386/i386/pmap.c
5348
KASSERT((*pte & PG_V) != 0,
sys/i386/i386/pmap.c
5350
pmap_remove_pte(pmap, pte, va, NULL);
sys/i386/i386/pmap.c
5357
for (pte = pmap_pte_quick(pmap, sva); sva != pdnxt; pte++,
sys/i386/i386/pmap.c
5359
if ((*pte & (PG_MANAGED | PG_V)) != (PG_MANAGED | PG_V))
sys/i386/i386/pmap.c
5361
else if ((*pte & (PG_M | PG_RW)) == (PG_M | PG_RW)) {
sys/i386/i386/pmap.c
5368
m = PHYS_TO_VM_PAGE(*pte & PG_FRAME);
sys/i386/i386/pmap.c
5371
atomic_clear_int((u_int *)pte, PG_M | PG_A);
sys/i386/i386/pmap.c
5372
} else if ((*pte & PG_A) != 0)
sys/i386/i386/pmap.c
5373
atomic_clear_int((u_int *)pte, PG_A);
sys/i386/i386/pmap.c
5376
if ((*pte & PG_G) != 0) {
sys/i386/i386/pmap.c
5410
pt_entry_t *pte;
sys/i386/i386/pmap.c
5439
pte = pmap_pte_quick(pmap, va);
sys/i386/i386/pmap.c
5445
atomic_clear_int((u_int *)pte, PG_M | PG_RW);
sys/i386/i386/pmap.c
5458
pte = pmap_pte_quick(pmap, pv->pv_va);
sys/i386/i386/pmap.c
5459
if ((*pte & (PG_M | PG_RW)) == (PG_M | PG_RW)) {
sys/i386/i386/pmap.c
5465
atomic_clear_int((u_int *)pte, PG_M);
sys/i386/i386/pmap.c
5480
pmap_pte_attr(pt_entry_t *pte, int cache_bits)
sys/i386/i386/pmap.c
5489
opte = *(u_int *)pte;
sys/i386/i386/pmap.c
5492
} while (npte != opte && !atomic_cmpset_int((u_int *)pte, opte, npte));
sys/i386/i386/pmap.c
5715
pt_entry_t *pte;
sys/i386/i386/pmap.c
5772
pte = vtopte(tmpva);
sys/i386/i386/pmap.c
5773
if (*pte == 0) {
sys/i386/i386/pmap.c
5794
pte = vtopte(tmpva);
sys/i386/i386/pmap.c
5795
if ((*pte & PG_PTE_CACHE) != cache_bits_pte) {
sys/i386/i386/pmap.c
5796
pmap_pte_attr(pte, cache_bits_pte);
sys/i386/i386/pmap.c
5823
pt_entry_t pte;
sys/i386/i386/pmap.c
5831
pte = pde;
sys/i386/i386/pmap.c
5837
pte = pmap_pte_ufast(pmap, addr, pde);
sys/i386/i386/pmap.c
5838
pa = pte & PG_FRAME;
sys/i386/i386/pmap.c
5842
pte = 0;
sys/i386/i386/pmap.c
5846
if ((pte & PG_V) != 0) {
sys/i386/i386/pmap.c
5848
if ((pte & (PG_M | PG_RW)) == (PG_M | PG_RW))
sys/i386/i386/pmap.c
5850
if ((pte & PG_A) != 0)
sys/i386/i386/pmap.c
5855
(pte & (PG_MANAGED | PG_V)) == (PG_MANAGED | PG_V)) {
sys/i386/i386/pmap.c
5935
pt_entry_t *pte;
sys/i386/i386/pmap.c
5939
pte = vtopte(qaddr);
sys/i386/i386/pmap.c
5941
KASSERT(*pte == 0,
sys/i386/i386/pmap.c
5942
("pmap_quick_enter_page: PTE busy %#jx", (uintmax_t)*pte));
sys/i386/i386/pmap.c
5943
*pte = PG_V | PG_RW | VM_PAGE_TO_PHYS(m) | PG_A | PG_M |
sys/i386/i386/pmap.c
5954
pt_entry_t *pte;
sys/i386/i386/pmap.c
5957
pte = vtopte(qaddr);
sys/i386/i386/pmap.c
5959
KASSERT(*pte != 0, ("pmap_quick_remove_page: PTE not in use"));
sys/i386/i386/pmap.c
5962
*pte = 0;
sys/i386/i386/pmap.c
610
pt_entry_t *pte, *unused __unused;
sys/i386/i386/pmap.c
6110
pt_entry_t *pte;
sys/i386/i386/pmap.c
6113
for (i = 0, pte = vtopte(kaddr); i < plen; i++, pte++) {
sys/i386/i386/pmap.c
6114
*pte = PG_V | PG_RW | PG_A | PG_M | VM_PAGE_TO_PHYS(ma[i]) |
sys/i386/i386/pmap.c
6146
pt_entry_t *pte;
sys/i386/i386/pmap.c
6148
pte = CMAP3;
sys/i386/i386/pmap.c
6149
*pte = pa | pte_bits;
sys/i386/i386/pmap.c
6157
pt_entry_t *pte;
sys/i386/i386/pmap.c
6166
pte = (pt_entry_t *)vm86paddr;
sys/i386/i386/pmap.c
6168
pte[i] = (i << PAGE_SHIFT) | PG_V | PG_RW | PG_U;
sys/i386/i386/pmap.c
6172
pt_entry_t *pte;
sys/i386/i386/pmap.c
6186
h->pte = (pt_entry_t *)malloc(PAGE_SIZE, M_TEMP, M_WAITOK);
sys/i386/i386/pmap.c
6188
*h->pte = vm86phystk | PG_RW | PG_V;
sys/i386/i386/pmap.c
6190
*h->ptd = vtophys(h->pte) | PG_RW | PG_V;
sys/i386/i386/pmap.c
6206
free(h->pte, M_TEMP); /* ... and free it */
sys/i386/i386/pmap.c
6308
vm_offset_t va, pd_entry_t pde, pt_entry_t pte)
sys/i386/i386/pmap.c
6316
} else if (pte != 0) {
sys/i386/i386/pmap.c
6317
attrs |= pte & pg_nx;
sys/i386/i386/pmap.c
6318
attrs &= pg_nx | (pte & (PG_RW | PG_U));
sys/i386/i386/pmap.c
6319
attrs |= pte & (PG_G | PG_PTE_CACHE);
sys/i386/i386/pmap.c
6338
pt_entry_t *pt, pte;
sys/i386/i386/pmap.c
6385
pte = *pt;
sys/i386/i386/pmap.c
6386
if ((pte & PG_V) == 0) {
sys/i386/i386/pmap.c
6390
sysctl_kmaps_check(sb, &range, sva, pde, pte);
sys/i386/i386/pmap.c
653
v = (c)va; va += ((n)*PAGE_SIZE); p = pte; pte += (n);
sys/i386/i386/pmap.c
656
pte = vtopte(va);
sys/i386/i386/pmap.c
668
SYSMAP(vm_offset_t, pte, pc->pc_qmap_addr, 1)
sys/i386/i386/pmap.c
907
pt_entry_t *pte;
sys/i386/i386/pmap.c
913
pte = vtopte(va);
sys/i386/i386/pmap.c
914
*head = *pte;
sys/i386/i386/pmap.c
917
*pte = 0;
sys/i386/i386/pmap.c
924
pt_entry_t *pte;
sys/i386/i386/pmap.c
928
pte = vtopte(va);
sys/i386/i386/pmap.c
929
*pte = *head; /* virtual! PG_V is 0 though */
sys/i386/include/pmap_nopae.h
80
#define pte_load_store(ptep, pte) atomic_swap_int(ptep, pte)
sys/i386/include/pmap_nopae.h
82
#define pte_store(ptep, pte) do { \
sys/i386/include/pmap_nopae.h
83
*(u_int *)(ptep) = (u_int)(pte); \
sys/i386/include/pmap_nopae.h
85
#define pte_store_zero(ptep, pte) pte_store(ptep, pte)
sys/i386/include/pmap_pae.h
100
#define pte_store_zero(ptep, pte) \
sys/i386/include/pmap_pae.h
106
*(p + 1) = (uint32_t)(pte >> 32); \
sys/i386/include/pmap_pae.h
108
*p = (uint32_t)pte; \
sys/i386/include/pmap_pae.h
97
#define pte_load_store(ptep, pte) atomic_swap_64_i586(ptep, pte)
sys/i386/include/pmap_pae.h
99
#define pte_store(ptep, pte) atomic_store_rel_64_i586(ptep, pte)
sys/powerpc/aim/mmu_oea.c
1303
if (pvo != NULL && (pvo->pvo_pte.pte.pte_hi & PTE_VALID) &&
sys/powerpc/aim/mmu_oea.c
1304
((pvo->pvo_pte.pte.pte_lo & PTE_PP) == PTE_RW ||
sys/powerpc/aim/mmu_oea.c
1368
rv = pvo == NULL || (pvo->pvo_pte.pte.pte_hi & PTE_VALID) == 0;
sys/powerpc/aim/mmu_oea.c
1395
struct pte *pt;
sys/powerpc/aim/mmu_oea.c
1411
if ((pvo->pvo_pte.pte.pte_lo & PTE_PP) != PTE_BR) {
sys/powerpc/aim/mmu_oea.c
1413
pvo->pvo_pte.pte.pte_lo &= ~PTE_PP;
sys/powerpc/aim/mmu_oea.c
1414
pvo->pvo_pte.pte.pte_lo |= PTE_BR;
sys/powerpc/aim/mmu_oea.c
1416
moea_pte_synch(pt, &pvo->pvo_pte.pte);
sys/powerpc/aim/mmu_oea.c
1417
lo |= pvo->pvo_pte.pte.pte_lo;
sys/powerpc/aim/mmu_oea.c
1418
pvo->pvo_pte.pte.pte_lo &= ~PTE_CHG;
sys/powerpc/aim/mmu_oea.c
1419
moea_pte_change(pt, &pvo->pvo_pte.pte,
sys/powerpc/aim/mmu_oea.c
1467
struct pte *pt;
sys/powerpc/aim/mmu_oea.c
1487
pvo->pvo_pte.pte.pte_lo &= ~PTE_WIMG;
sys/powerpc/aim/mmu_oea.c
1488
pvo->pvo_pte.pte.pte_lo |= lo;
sys/powerpc/aim/mmu_oea.c
1490
moea_pte_change(pt, &pvo->pvo_pte.pte,
sys/powerpc/aim/mmu_oea.c
162
#define PVO_PADDR(pvo) ((pvo)->pvo_pte.pte.pte_lo & PTE_RPGN)
sys/powerpc/aim/mmu_oea.c
1803
struct pte *pt;
sys/powerpc/aim/mmu_oea.c
1828
pvo->pvo_pte.pte.pte_lo &= ~PTE_PP;
sys/powerpc/aim/mmu_oea.c
1829
pvo->pvo_pte.pte.pte_lo |= PTE_BR;
sys/powerpc/aim/mmu_oea.c
1835
moea_pte_change(pt, &pvo->pvo_pte.pte, pvo->pvo_vaddr);
sys/powerpc/aim/mmu_oea.c
2068
(pvo->pvo_pte.pte.pte_lo & PTE_PP) ==
sys/powerpc/aim/mmu_oea.c
2125
moea_pte_create(&pvo->pvo_pte.pte, sr, va, pa | pte_lo);
sys/powerpc/aim/mmu_oea.c
2144
i = moea_pte_insert(ptegidx, &pvo->pvo_pte.pte);
sys/powerpc/aim/mmu_oea.c
2160
struct pte *pt;
sys/powerpc/aim/mmu_oea.c
2168
moea_pte_unset(pt, &pvo->pvo_pte.pte, pvo->pvo_vaddr);
sys/powerpc/aim/mmu_oea.c
2197
moea_attr_save(pg, pvo->pvo_pte.pte.pte_lo &
sys/powerpc/aim/mmu_oea.c
2227
if (pvo->pvo_pte.pte.pte_hi & PTE_HID)
sys/powerpc/aim/mmu_oea.c
2257
static struct pte *
sys/powerpc/aim/mmu_oea.c
2260
struct pte *pt;
sys/powerpc/aim/mmu_oea.c
2277
if ((pvo->pvo_pte.pte.pte_hi & PTE_VALID) && !PVO_PTEGIDX_ISSET(pvo)) {
sys/powerpc/aim/mmu_oea.c
2282
if ((pvo->pvo_pte.pte.pte_hi & PTE_VALID) == 0 && PVO_PTEGIDX_ISSET(pvo)) {
sys/powerpc/aim/mmu_oea.c
2287
if ((pt->pte_hi ^ (pvo->pvo_pte.pte.pte_hi & ~PTE_VALID)) == PTE_VALID) {
sys/powerpc/aim/mmu_oea.c
2288
if ((pvo->pvo_pte.pte.pte_hi & PTE_VALID) == 0) {
sys/powerpc/aim/mmu_oea.c
2293
if (((pt->pte_lo ^ pvo->pvo_pte.pte.pte_lo) & ~(PTE_CHG|PTE_REF))
sys/powerpc/aim/mmu_oea.c
2303
if (pvo->pvo_pte.pte.pte_hi & PTE_VALID) {
sys/powerpc/aim/mmu_oea.c
2305
"moea_pteg_table but valid in pvo: %8x, %8x", pvo, pt, pvo->pvo_pte.pte.pte_hi, pt->pte_hi);
sys/powerpc/aim/mmu_oea.c
2323
struct pte *pt;
sys/powerpc/aim/mmu_oea.c
2347
moea_pte_match(&pvo->pvo_pte.pte, sr, addr,
sys/powerpc/aim/mmu_oea.c
2348
pvo->pvo_pte.pte.pte_hi & PTE_HID)) {
sys/powerpc/aim/mmu_oea.c
2353
j = moea_pte_insert(ptegidx, &pvo->pvo_pte.pte);
sys/powerpc/aim/mmu_oea.c
2373
moea_pte_compare(pt, &pvo->pvo_pte.pte)) {
sys/powerpc/aim/mmu_oea.c
2400
if (moea_pte_compare(pt, &pvo->pvo_pte.pte)) {
sys/powerpc/aim/mmu_oea.c
2416
source_pvo->pvo_pte.pte.pte_hi &= ~PTE_HID;
sys/powerpc/aim/mmu_oea.c
2418
moea_pte_unset(pt, &victim_pvo->pvo_pte.pte, victim_pvo->pvo_vaddr);
sys/powerpc/aim/mmu_oea.c
2419
moea_pte_set(pt, &source_pvo->pvo_pte.pte);
sys/powerpc/aim/mmu_oea.c
2432
struct pte *pt;
sys/powerpc/aim/mmu_oea.c
2439
if (!(pvo_walk->pvo_pte.pte.pte_hi & PTE_VALID))
sys/powerpc/aim/mmu_oea.c
2458
moea_pte_insert(u_int ptegidx, struct pte *pvo_pt)
sys/powerpc/aim/mmu_oea.c
2460
struct pte *pt;
sys/powerpc/aim/mmu_oea.c
2520
if (pt->pte_hi != victim_pvo->pvo_pte.pte.pte_hi)
sys/powerpc/aim/mmu_oea.c
2521
panic("Victim PVO doesn't match PTE! PVO: %8x, PTE: %8x", victim_pvo->pvo_pte.pte.pte_hi, pt->pte_hi);
sys/powerpc/aim/mmu_oea.c
2526
moea_pte_unset(pt, &victim_pvo->pvo_pte.pte, victim_pvo->pvo_vaddr);
sys/powerpc/aim/mmu_oea.c
253
static int moea_pte_insert(u_int, struct pte *);
sys/powerpc/aim/mmu_oea.c
2538
struct pte *pt;
sys/powerpc/aim/mmu_oea.c
2549
if (pvo->pvo_pte.pte.pte_lo & ptebit) {
sys/powerpc/aim/mmu_oea.c
2569
moea_pte_synch(pt, &pvo->pvo_pte.pte);
sys/powerpc/aim/mmu_oea.c
2571
if (pvo->pvo_pte.pte.pte_lo & ptebit) {
sys/powerpc/aim/mmu_oea.c
2586
struct pte *pt;
sys/powerpc/aim/mmu_oea.c
2612
moea_pte_synch(pt, &pvo->pvo_pte.pte);
sys/powerpc/aim/mmu_oea.c
2613
if (pvo->pvo_pte.pte.pte_lo & ptebit) {
sys/powerpc/aim/mmu_oea.c
2619
pvo->pvo_pte.pte.pte_lo &= ~ptebit;
sys/powerpc/aim/mmu_oea.c
262
static struct pte *moea_pvo_to_pte(const struct pvo_entry *, int);
sys/powerpc/aim/mmu_oea.c
2821
if (pvo != NULL && (pvo->pvo_pte.pte.pte_hi & PTE_VALID))
sys/powerpc/aim/mmu_oea.c
2836
!(pvo->pvo_pte.pte.pte_hi & PTE_VALID))
sys/powerpc/aim/mmu_oea.c
527
moea_pte_compare(const struct pte *pt, const struct pte *pvo_pt)
sys/powerpc/aim/mmu_oea.c
536
moea_pte_match(struct pte *pt, u_int sr, vm_offset_t va, int which)
sys/powerpc/aim/mmu_oea.c
544
moea_pte_create(struct pte *pt, u_int sr, vm_offset_t va, u_int pte_lo)
sys/powerpc/aim/mmu_oea.c
561
moea_pte_synch(struct pte *pt, struct pte *pvo_pt)
sys/powerpc/aim/mmu_oea.c
569
moea_pte_clear(struct pte *pt, vm_offset_t va, int ptebit)
sys/powerpc/aim/mmu_oea.c
582
moea_pte_set(struct pte *pt, struct pte *pvo_pt)
sys/powerpc/aim/mmu_oea.c
601
moea_pte_unset(struct pte *pt, struct pte *pvo_pt, vm_offset_t va)
sys/powerpc/aim/mmu_oea.c
627
moea_pte_change(struct pte *pt, struct pte *pvo_pt, vm_offset_t va)
sys/powerpc/aim/mmu_radix.c
1017
if (!atomic_cmpset_long(pte, htobe64(origpte), htobe64(newpte)))
sys/powerpc/aim/mmu_radix.c
1380
pt_entry_t *pte, tpte;
sys/powerpc/aim/mmu_radix.c
1459
pte = pmap_l3e_to_pte(l3e, va);
sys/powerpc/aim/mmu_radix.c
1460
if ((be64toh(*pte) & PG_W) != 0)
sys/powerpc/aim/mmu_radix.c
1462
tpte = be64toh(pte_load_clear(pte));
sys/powerpc/aim/mmu_radix.c
1783
pt_entry_t *pte, pteval;
sys/powerpc/aim/mmu_radix.c
1790
pte = pmap_pml1e(kernel_pmap, PHYS_TO_DMAP(start));
sys/powerpc/aim/mmu_radix.c
1791
if ((be64toh(*pte) & RPTE_VALID) == 0) {
sys/powerpc/aim/mmu_radix.c
1793
pde_store(pte, page);
sys/powerpc/aim/mmu_radix.c
1795
pte = pmap_l1e_to_l2e(pte, PHYS_TO_DMAP(start));
sys/powerpc/aim/mmu_radix.c
1800
} else if ((be64toh(*pte) & RPTE_VALID) == 0) {
sys/powerpc/aim/mmu_radix.c
1802
pde_store(pte, page);
sys/powerpc/aim/mmu_radix.c
1805
pte = pmap_l2e_to_l3e(pte, PHYS_TO_DMAP(start));
sys/powerpc/aim/mmu_radix.c
1810
} else if ((be64toh(*pte) & RPTE_VALID) == 0) {
sys/powerpc/aim/mmu_radix.c
1812
pde_store(pte, page);
sys/powerpc/aim/mmu_radix.c
1814
pte = pmap_l3e_to_pte(pte, PHYS_TO_DMAP(start));
sys/powerpc/aim/mmu_radix.c
1817
pte_store(pte, pteval);
sys/powerpc/aim/mmu_radix.c
1841
pt_entry_t *pte;
sys/powerpc/aim/mmu_radix.c
1864
pte = pmap_pml1e(kernel_pmap, VM_MIN_KERNEL_ADDRESS);
sys/powerpc/aim/mmu_radix.c
1865
*pte = htobe64(pages | RPTE_VALID | RPTE_SHIFT);
sys/powerpc/aim/mmu_radix.c
1867
pte = pmap_l1e_to_l2e(pte, VM_MIN_KERNEL_ADDRESS);
sys/powerpc/aim/mmu_radix.c
1868
*pte = htobe64(pages | RPTE_VALID | RPTE_SHIFT);
sys/powerpc/aim/mmu_radix.c
1870
pte = pmap_l2e_to_l3e(pte, VM_MIN_KERNEL_ADDRESS);
sys/powerpc/aim/mmu_radix.c
1883
for (int i = 0; i < nkpt; i++, pte++, pages += PAGE_SIZE)
sys/powerpc/aim/mmu_radix.c
1884
*pte = htobe64(pages | RPTE_VALID | RPTE_SHIFT);
sys/powerpc/aim/mmu_radix.c
2224
pt_entry_t *pte;
sys/powerpc/aim/mmu_radix.c
2285
pte = pmap_l3e_to_pte(l3e, va);
sys/powerpc/aim/mmu_radix.c
2286
KASSERT((be64toh(*pte) & PG_V) != 0,
sys/powerpc/aim/mmu_radix.c
2288
pmap_remove_pte(pmap, pte, va, be64toh(*l3e), NULL,
sys/powerpc/aim/mmu_radix.c
2298
for (pte = pmap_l3e_to_pte(l3e, sva); sva != va_next;
sys/powerpc/aim/mmu_radix.c
2299
pte++, sva += PAGE_SIZE) {
sys/powerpc/aim/mmu_radix.c
2300
MPASS(pte == pmap_pte(pmap, sva));
sys/powerpc/aim/mmu_radix.c
2302
if ((be64toh(*pte) & (PG_MANAGED | PG_V)) != (PG_MANAGED | PG_V))
sys/powerpc/aim/mmu_radix.c
2304
else if ((be64toh(*pte) & (PG_M | PG_RW)) == (PG_M | PG_RW)) {
sys/powerpc/aim/mmu_radix.c
2311
m = PHYS_TO_VM_PAGE(be64toh(*pte) & PG_FRAME);
sys/powerpc/aim/mmu_radix.c
2314
atomic_clear_long(pte, htobe64(PG_M | PG_A));
sys/powerpc/aim/mmu_radix.c
2315
} else if ((be64toh(*pte) & PG_A) != 0)
sys/powerpc/aim/mmu_radix.c
2316
atomic_clear_long(pte, htobe64(PG_A));
sys/powerpc/aim/mmu_radix.c
2430
pt_entry_t oldpte, *pte;
sys/powerpc/aim/mmu_radix.c
2477
pte = pmap_l3e_to_pte(l3e, va);
sys/powerpc/aim/mmu_radix.c
2478
oldpte = be64toh(*pte);
sys/powerpc/aim/mmu_radix.c
2479
while (!atomic_cmpset_long(pte,
sys/powerpc/aim/mmu_radix.c
2482
oldpte = be64toh(*pte);
sys/powerpc/aim/mmu_radix.c
2504
pte = pmap_l3e_to_pte(l3e, pv->pv_va);
sys/powerpc/aim/mmu_radix.c
2505
if ((be64toh(*pte) & (PG_M | PG_RW)) == (PG_M | PG_RW)) {
sys/powerpc/aim/mmu_radix.c
2506
atomic_clear_long(pte, htobe64(PG_M));
sys/powerpc/aim/mmu_radix.c
2726
pt_entry_t *firstpte, oldpte, pa, *pte;
sys/powerpc/aim/mmu_radix.c
2760
for (pte = firstpte + NPTEPG - 1; pte > firstpte; pte--) {
sys/powerpc/aim/mmu_radix.c
2762
oldpte = be64toh(*pte);
sys/powerpc/aim/mmu_radix.c
2773
if (!atomic_cmpset_long(pte, htobe64(oldpte), htobe64((oldpte | RPTE_EAA_R) & ~RPTE_EAA_W)))
sys/powerpc/aim/mmu_radix.c
2830
pt_entry_t *pte;
sys/powerpc/aim/mmu_radix.c
2901
pte = pmap_l3e_to_pte(l3e, va);
sys/powerpc/aim/mmu_radix.c
2925
origpte = be64toh(*pte);
sys/powerpc/aim/mmu_radix.c
2977
if (!atomic_cmpset_long(pte, htobe64(origpte), htobe64(newpte)))
sys/powerpc/aim/mmu_radix.c
3002
origpte = be64toh(pte_load_clear(pte));
sys/powerpc/aim/mmu_radix.c
3084
origpte = be64toh(pte_load_store(pte, htobe64(newpte)));
sys/powerpc/aim/mmu_radix.c
3108
pte_store(pte, newpte);
sys/powerpc/aim/mmu_radix.c
330
pt_entry_t *pte;
sys/powerpc/aim/mmu_radix.c
334
pte = (pt_entry_t *)PHYS_TO_DMAP(ptepa);
sys/powerpc/aim/mmu_radix.c
335
return (&pte[pmap_pte_index(va)]);
sys/powerpc/aim/mmu_radix.c
3378
pt_entry_t *pte;
sys/powerpc/aim/mmu_radix.c
3427
pte = (pt_entry_t *)PHYS_TO_DMAP(VM_PAGE_TO_PHYS(mpte));
sys/powerpc/aim/mmu_radix.c
3428
pte = &pte[pmap_pte_index(va)];
sys/powerpc/aim/mmu_radix.c
3431
pte = pmap_pte(pmap, va);
sys/powerpc/aim/mmu_radix.c
3433
if (be64toh(*pte)) {
sys/powerpc/aim/mmu_radix.c
3476
pte_store(pte, pa);
sys/powerpc/aim/mmu_radix.c
3504
pt_entry_t *pte;
sys/powerpc/aim/mmu_radix.c
3522
pte = pmap_l3e_to_pte(l3e, va);
sys/powerpc/aim/mmu_radix.c
3523
if (__predict_false(pte == NULL))
sys/powerpc/aim/mmu_radix.c
3525
pa = be64toh(*pte);
sys/powerpc/aim/mmu_radix.c
3536
pt_entry_t pte;
sys/powerpc/aim/mmu_radix.c
3550
pte = be64toh(*pmap_l3e_to_pte(l3ep, va));
sys/powerpc/aim/mmu_radix.c
3551
if ((pte & PG_V) &&
sys/powerpc/aim/mmu_radix.c
3552
((pte & PG_RW) || (prot & VM_PROT_WRITE) == 0))
sys/powerpc/aim/mmu_radix.c
3553
m = PHYS_TO_VM_PAGE(pte & PG_FRAME);
sys/powerpc/aim/mmu_radix.c
3760
pt_entry_t *pte, mask;
sys/powerpc/aim/mmu_radix.c
3781
pte = pmap_pte(pmap, pv->pv_va);
sys/powerpc/aim/mmu_radix.c
3787
rv = (be64toh(*pte) & mask) == mask;
sys/powerpc/aim/mmu_radix.c
3808
pte = pmap_pml3e(pmap, pv->pv_va);
sys/powerpc/aim/mmu_radix.c
3814
rv = (be64toh(*pte) & mask) == mask;
sys/powerpc/aim/mmu_radix.c
3851
pt_entry_t *pte;
sys/powerpc/aim/mmu_radix.c
3859
pte = pmap_l3e_to_pte(l3e, addr);
sys/powerpc/aim/mmu_radix.c
3860
rv = (be64toh(*pte) & PG_V) == 0;
sys/powerpc/aim/mmu_radix.c
3903
pt_entry_t *pte;
sys/powerpc/aim/mmu_radix.c
4010
pte = pmap_l3e_to_pte(l3e, pv->pv_va);
sys/powerpc/aim/mmu_radix.c
4011
if ((be64toh(*pte) & (PG_M | PG_RW)) == (PG_M | PG_RW))
sys/powerpc/aim/mmu_radix.c
4013
if ((be64toh(*pte) & PG_A) != 0) {
sys/powerpc/aim/mmu_radix.c
4014
atomic_clear_long(pte, htobe64(PG_A));
sys/powerpc/aim/mmu_radix.c
4185
pt_entry_t *pte;
sys/powerpc/aim/mmu_radix.c
4208
pte = pmap_pte(pmap, pv->pv_va);
sys/powerpc/aim/mmu_radix.c
4209
if ((be64toh(*pte) & PG_W) != 0)
sys/powerpc/aim/mmu_radix.c
4229
pte = pmap_pml3e(pmap, pv->pv_va);
sys/powerpc/aim/mmu_radix.c
4230
if ((be64toh(*pte) & PG_W) != 0)
sys/powerpc/aim/mmu_radix.c
4548
pt_entry_t *pte;
sys/powerpc/aim/mmu_radix.c
4625
for (pte = pmap_l3e_to_pte(l3e, sva); sva != va_next; pte++,
sys/powerpc/aim/mmu_radix.c
4631
MPASS(pte == pmap_pte(pmap, sva));
sys/powerpc/aim/mmu_radix.c
4632
obits = pbits = be64toh(*pte);
sys/powerpc/aim/mmu_radix.c
4649
if (!atomic_cmpset_long(pte, htobe64(obits), htobe64(pbits)))
sys/powerpc/aim/mmu_radix.c
4672
pt_entry_t oldpte, pa, *pte;
sys/powerpc/aim/mmu_radix.c
4680
pte = kvtopte(va);
sys/powerpc/aim/mmu_radix.c
4683
pte = kvtopte(va);
sys/powerpc/aim/mmu_radix.c
4684
MPASS(pte == pmap_pte(kernel_pmap, va));
sys/powerpc/aim/mmu_radix.c
4695
if (be64toh(*pte) != pa) {
sys/powerpc/aim/mmu_radix.c
4696
oldpte |= be64toh(*pte);
sys/powerpc/aim/mmu_radix.c
4697
pte_store(pte, pa);
sys/powerpc/aim/mmu_radix.c
4700
pte++;
sys/powerpc/aim/mmu_radix.c
4713
pt_entry_t *pte;
sys/powerpc/aim/mmu_radix.c
4719
pte = kvtopte(va);
sys/powerpc/aim/mmu_radix.c
4722
pte = kvtopte(va);
sys/powerpc/aim/mmu_radix.c
4723
pte_clear(pte);
sys/powerpc/aim/mmu_radix.c
4724
pte++;
sys/powerpc/aim/mmu_radix.c
4909
pt_entry_t *pte;
sys/powerpc/aim/mmu_radix.c
4911
for (pte = firstpte; pte < firstpte + NPTEPG; pte++) {
sys/powerpc/aim/mmu_radix.c
4912
*pte = htobe64(newpte);
sys/powerpc/aim/mmu_radix.c
5159
pt_entry_t *pte;
sys/powerpc/aim/mmu_radix.c
5166
pte = pmap_l3e_to_pte(l3e, va);
sys/powerpc/aim/mmu_radix.c
5167
if ((be64toh(*pte) & RPTE_VALID) == 0) {
sys/powerpc/aim/mmu_radix.c
5172
invalidate_all = pmap_remove_pte(pmap, pte, va, be64toh(*l3e), free, &lock);
sys/powerpc/aim/mmu_radix.c
5187
pt_entry_t *pte;
sys/powerpc/aim/mmu_radix.c
5194
for (pte = pmap_l3e_to_pte(l3e, sva); sva != eva; pte++,
sys/powerpc/aim/mmu_radix.c
5196
MPASS(pte == pmap_pte(pmap, sva));
sys/powerpc/aim/mmu_radix.c
5197
if (*pte == 0) {
sys/powerpc/aim/mmu_radix.c
5206
if (pmap_remove_pte(pmap, pte, sva, be64toh(*l3e), free, lockp)) {
sys/powerpc/aim/mmu_radix.c
5343
pt_entry_t *pte, tpte;
sys/powerpc/aim/mmu_radix.c
5394
pte = pmap_l3e_to_pte(l3e, pv->pv_va);
sys/powerpc/aim/mmu_radix.c
5395
tpte = be64toh(pte_load_clear(pte));
sys/powerpc/aim/mmu_radix.c
5450
pt_entry_t *pte, tpte;
sys/powerpc/aim/mmu_radix.c
5493
pte = pmap_pml2e(pmap, pv->pv_va);
sys/powerpc/aim/mmu_radix.c
5494
ptel3e = be64toh(*pte);
sys/powerpc/aim/mmu_radix.c
5495
pte = pmap_l2e_to_l3e(pte, pv->pv_va);
sys/powerpc/aim/mmu_radix.c
5496
tpte = be64toh(*pte);
sys/powerpc/aim/mmu_radix.c
5500
pte = (pt_entry_t *)PHYS_TO_DMAP(tpte &
sys/powerpc/aim/mmu_radix.c
5502
pte = &pte[pmap_pte_index(pv->pv_va)];
sys/powerpc/aim/mmu_radix.c
5503
tpte = be64toh(*pte);
sys/powerpc/aim/mmu_radix.c
5547
pte_clear(pte);
sys/powerpc/aim/mmu_radix.c
5628
pt_entry_t oldpte, *pte;
sys/powerpc/aim/mmu_radix.c
5683
pte = pmap_l3e_to_pte(l3e, pv->pv_va);
sys/powerpc/aim/mmu_radix.c
5685
oldpte = be64toh(*pte);
sys/powerpc/aim/mmu_radix.c
5687
if (!atomic_cmpset_long(pte, htobe64(oldpte),
sys/powerpc/aim/mmu_radix.c
5719
pt_entry_t *pte;
sys/powerpc/aim/mmu_radix.c
5763
for (pte = pmap_l3e_to_pte(l3e, sva); sva != va_next; pte++,
sys/powerpc/aim/mmu_radix.c
5765
MPASS(pte == pmap_pte(pmap, sva));
sys/powerpc/aim/mmu_radix.c
5766
if ((be64toh(*pte) & PG_V) == 0)
sys/powerpc/aim/mmu_radix.c
5768
if ((be64toh(*pte) & PG_W) == 0)
sys/powerpc/aim/mmu_radix.c
5770
(uintmax_t)(be64toh(*pte)));
sys/powerpc/aim/mmu_radix.c
5777
atomic_clear_long(pte, htobe64(PG_W));
sys/powerpc/aim/mmu_radix.c
5809
pt_entry_t pte;
sys/powerpc/aim/mmu_radix.c
5819
pte = be64toh(*l3ep);
sys/powerpc/aim/mmu_radix.c
5826
pte = be64toh(*pmap_l3e_to_pte(l3ep, addr));
sys/powerpc/aim/mmu_radix.c
5827
pa = pte & PG_FRAME;
sys/powerpc/aim/mmu_radix.c
5831
pte = 0;
sys/powerpc/aim/mmu_radix.c
5835
if ((pte & PG_V) != 0) {
sys/powerpc/aim/mmu_radix.c
5837
if ((pte & (PG_M | PG_RW)) == (PG_M | PG_RW))
sys/powerpc/aim/mmu_radix.c
5839
if ((pte & PG_A) != 0)
sys/powerpc/aim/mmu_radix.c
5844
(pte & (PG_MANAGED | PG_V)) == (PG_MANAGED | PG_V)) {
sys/powerpc/aim/mmu_radix.c
6001
pmap_pte_attr(pt_entry_t *pte, uint64_t cache_bits, uint64_t mask)
sys/powerpc/aim/mmu_radix.c
6010
opte = be64toh(*pte);
sys/powerpc/aim/mmu_radix.c
6013
} while (npte != opte && !atomic_cmpset_long(pte, htobe64(opte), htobe64(npte)));
sys/powerpc/aim/mmu_radix.c
6125
pt_entry_t *pte, pteval;
sys/powerpc/aim/mmu_radix.c
6128
pte = kvtopte(va);
sys/powerpc/aim/mmu_radix.c
6129
MPASS(pte != NULL);
sys/powerpc/aim/mmu_radix.c
6132
pte_store(pte, pteval | cache_bits);
sys/powerpc/aim/mmu_radix.c
6138
pt_entry_t *pte;
sys/powerpc/aim/mmu_radix.c
6142
pte = kvtopte(va);
sys/powerpc/aim/mmu_radix.c
6143
pte_clear(pte);
sys/powerpc/aim/mmu_radix.c
6224
pt_entry_t *pte;
sys/powerpc/aim/mmu_radix.c
6304
pte = pmap_l3e_to_pte(l3e, tmpva);
sys/powerpc/aim/mmu_radix.c
6305
if (*pte == 0)
sys/powerpc/aim/mmu_radix.c
6376
pte = pmap_l3e_to_pte(l3e, tmpva);
sys/powerpc/aim/mmu_radix.c
6377
if ((be64toh(*pte) & RPTE_ATTR_MASK) != cache_bits) {
sys/powerpc/aim/mmu_radix.c
6378
pmap_pte_attr(pte, cache_bits,
sys/powerpc/aim/mmu_radix.c
6383
(be64toh(*pte) & PG_FRAME) < dmaplimit) {
sys/powerpc/aim/mmu_radix.c
6386
pa_start = be64toh(*pte) & PG_FRAME;
sys/powerpc/aim/mmu_radix.c
6388
} else if (pa_end == (be64toh(*pte) & PG_FRAME))
sys/powerpc/aim/mmu_radix.c
6398
pa_start = be64toh(*pte) & PG_FRAME;
sys/powerpc/aim/mmu_radix.c
6486
pt_entry_t *pte;
sys/powerpc/aim/mmu_radix.c
6506
pte = pmap_l3e_to_pte(l3e, va);
sys/powerpc/aim/mmu_radix.c
6507
db_printf(" pte %#016lx\n", be64toh(*pte));
sys/powerpc/aim/mmu_radix.c
6530
DB_SHOW_COMMAND(pte, pmap_print_pte)
sys/powerpc/aim/mmu_radix.c
840
#define pte_load_store(ptep, pte) atomic_swap_long(ptep, pte)
sys/powerpc/aim/mmu_radix.c
842
#define pte_store(ptep, pte) do { \
sys/powerpc/aim/mmu_radix.c
843
MPASS((pte) & (RPTE_EAA_R | RPTE_EAA_W | RPTE_EAA_X)); \
sys/powerpc/aim/mmu_radix.c
844
*(u_long *)(ptep) = htobe64((u_long)((pte) | PG_V | RPTE_LEAF)); \
sys/powerpc/aim/mmu_radix.c
918
pt_entry_t *pte;
sys/powerpc/aim/mmu_radix.c
920
pte = kvtopte(va);
sys/powerpc/aim/mmu_radix.c
921
MPASS(pte != NULL);
sys/powerpc/aim/mmu_radix.c
922
*pte = htobe64(pa | RPTE_VALID | RPTE_LEAF | RPTE_EAA_R | \
sys/powerpc/aim/mmu_radix.c
936
pt_entry_t *pte;
sys/powerpc/aim/mmu_radix.c
949
pte = pmap_l3e_to_pte(l3e, va);
sys/powerpc/aim/mmu_radix.c
950
if (pte == NULL || (be64toh(*pte) & PG_V) == 0)
sys/powerpc/aim/mmu_radix.c
952
return (pte);
sys/powerpc/aim/mmu_radix.c
958
pt_entry_t *pte;
sys/powerpc/aim/mmu_radix.c
965
if ((pte = pmap_nofault_pte(pmap, va, &is_l3e)) == NULL)
sys/powerpc/aim/mmu_radix.c
967
origpte = newpte = be64toh(*pte);
sys/powerpc/aim/mmu_radix.c
987
if ((pte = pmap_nofault_pte(pmap, va, &is_l3e)) == NULL ||
sys/powerpc/aim/mmu_radix.c
988
be64toh(*pte) != origpte) {
sys/powerpc/aim/moea64_native.c
702
atomic_pte_lock(volatile struct lpte *pte, uint64_t bitmask, uint64_t *oldhi)
sys/powerpc/aim/moea64_native.c
734
: "=&r" (ret), "=&r"(temp), "=m" (pte->pte_hi)
sys/powerpc/aim/moea64_native.c
735
: "r" ((volatile char *)&pte->pte_hi),
sys/powerpc/aim/moea64_native.c
737
"m" (pte->pte_hi)
sys/powerpc/aim/moea64_native.c
756
: "=&r" (ret), "=&r"(oldhihalf), "=m" (pte->pte_hi)
sys/powerpc/aim/moea64_native.c
757
: "r" ((volatile char *)&pte->pte_hi + 4),
sys/powerpc/aim/moea64_native.c
759
"m" (pte->pte_hi)
sys/powerpc/aim/moea64_native.c
762
*oldhi = (pte->pte_hi & 0xffffffff00000000ULL) | oldhihalf;
sys/powerpc/booke/pmap.c
1142
pte_t *pte;
sys/powerpc/booke/pmap.c
1151
pte = pte_find(kernel_pmap, va);
sys/powerpc/booke/pmap.c
1152
KASSERT((pte != NULL), ("mmu_booke_kenter: invalid va. NULL PTE"));
sys/powerpc/booke/pmap.c
1157
if (PTE_ISVALID(pte)) {
sys/powerpc/booke/pmap.c
1164
*pte = PTE_RPN_FROM_PA(pa) | flags;
sys/powerpc/booke/pmap.c
1184
pte_t *pte;
sys/powerpc/booke/pmap.c
1192
pte = pte_find(kernel_pmap, va);
sys/powerpc/booke/pmap.c
1194
if (!PTE_ISVALID(pte)) {
sys/powerpc/booke/pmap.c
1205
*pte = 0;
sys/powerpc/booke/pmap.c
1278
pte_t *pte;
sys/powerpc/booke/pmap.c
1313
if (((pte = pte_find(pmap, va)) != NULL) &&
sys/powerpc/booke/pmap.c
1314
(PTE_ISVALID(pte)) && (PTE_PA(pte) == pa)) {
sys/powerpc/booke/pmap.c
1320
flags = *pte;
sys/powerpc/booke/pmap.c
1325
if (!PTE_ISWIRED(pte)) {
sys/powerpc/booke/pmap.c
1330
if (PTE_ISWIRED(pte)) {
sys/powerpc/booke/pmap.c
1354
if (PTE_ISMODIFIED(pte))
sys/powerpc/booke/pmap.c
1368
if ((*pte & (PTE_UX | PTE_SX)) == 0)
sys/powerpc/booke/pmap.c
1382
*pte &= ~PTE_FLAGS_MASK;
sys/powerpc/booke/pmap.c
1383
*pte |= flags;
sys/powerpc/booke/pmap.c
1505
pte_t *pte;
sys/powerpc/booke/pmap.c
1533
pte = pte_find_next(pmap, &va);
sys/powerpc/booke/pmap.c
1534
if ((pte == NULL) || !PTE_ISVALID(pte))
sys/powerpc/booke/pmap.c
1672
pte_t *pte;
sys/powerpc/booke/pmap.c
1684
if ((pte = pte_find(pmap, va)) != NULL) {
sys/powerpc/booke/pmap.c
1685
if (PTE_ISVALID(pte)) {
sys/powerpc/booke/pmap.c
1686
m = PHYS_TO_VM_PAGE(PTE_PA(pte));
sys/powerpc/booke/pmap.c
1692
if (PTE_ISMODIFIED(pte) && PTE_ISMANAGED(pte))
sys/powerpc/booke/pmap.c
1696
*pte &= ~(PTE_UW | PTE_SW | PTE_MODIFIED);
sys/powerpc/booke/pmap.c
1713
pte_t *pte;
sys/powerpc/booke/pmap.c
1724
if ((pte = pte_find(pv->pv_pmap, pv->pv_va)) != NULL) {
sys/powerpc/booke/pmap.c
1725
if (PTE_ISVALID(pte)) {
sys/powerpc/booke/pmap.c
1726
m = PHYS_TO_VM_PAGE(PTE_PA(pte));
sys/powerpc/booke/pmap.c
1732
if (PTE_ISMODIFIED(pte))
sys/powerpc/booke/pmap.c
1736
*pte &= ~(PTE_UW | PTE_SW | PTE_MODIFIED);
sys/powerpc/booke/pmap.c
1757
pte_t *pte;
sys/powerpc/booke/pmap.c
1763
pte = pte_find(pmap, va);
sys/powerpc/booke/pmap.c
1764
if ((pte != NULL) && PTE_ISVALID(pte)) {
sys/powerpc/booke/pmap.c
1770
if ((*pte & pte_wbit) != 0 || (prot & VM_PROT_WRITE) == 0) {
sys/powerpc/booke/pmap.c
1771
m = PHYS_TO_VM_PAGE(PTE_PA(pte));
sys/powerpc/booke/pmap.c
1798
pte_t *pte;
sys/powerpc/booke/pmap.c
1815
if ((pte = pte_find(pv->pv_pmap, pv->pv_va)) != NULL &&
sys/powerpc/booke/pmap.c
1816
PTE_ISVALID(pte)) {
sys/powerpc/booke/pmap.c
1817
if (PTE_ISMODIFIED(pte))
sys/powerpc/booke/pmap.c
1846
pte_t *pte;
sys/powerpc/booke/pmap.c
1856
if ((pte = pte_find(pv->pv_pmap, pv->pv_va)) != NULL &&
sys/powerpc/booke/pmap.c
1857
PTE_ISVALID(pte)) {
sys/powerpc/booke/pmap.c
1858
if (PTE_ISREFERENCED(pte))
sys/powerpc/booke/pmap.c
1875
pte_t *pte;
sys/powerpc/booke/pmap.c
1888
if ((pte = pte_find(pv->pv_pmap, pv->pv_va)) != NULL &&
sys/powerpc/booke/pmap.c
1889
PTE_ISVALID(pte)) {
sys/powerpc/booke/pmap.c
1893
if (*pte & (PTE_SW | PTE_UW | PTE_MODIFIED)) {
sys/powerpc/booke/pmap.c
1895
*pte &= ~(PTE_SW | PTE_UW | PTE_MODIFIED |
sys/powerpc/booke/pmap.c
1924
pte_t *pte;
sys/powerpc/booke/pmap.c
1934
if ((pte = pte_find(pv->pv_pmap, pv->pv_va)) != NULL &&
sys/powerpc/booke/pmap.c
1935
PTE_ISVALID(pte)) {
sys/powerpc/booke/pmap.c
1936
if (PTE_ISMODIFIED(pte))
sys/powerpc/booke/pmap.c
1938
if (PTE_ISREFERENCED(pte)) {
sys/powerpc/booke/pmap.c
1943
*pte &= ~PTE_REFERENCED;
sys/powerpc/booke/pmap.c
1973
pte_t *pte;
sys/powerpc/booke/pmap.c
1977
if ((pte = pte_find(pmap, va)) != NULL &&
sys/powerpc/booke/pmap.c
1978
PTE_ISVALID(pte)) {
sys/powerpc/booke/pmap.c
1979
if (!PTE_ISWIRED(pte))
sys/powerpc/booke/pmap.c
1981
pte);
sys/powerpc/booke/pmap.c
1982
*pte &= ~PTE_WIRED;
sys/powerpc/booke/pmap.c
2028
pte_t *pte;
sys/powerpc/booke/pmap.c
2036
if ((pte = pte_find(pv->pv_pmap, pv->pv_va)) != NULL)
sys/powerpc/booke/pmap.c
2037
if (PTE_ISVALID(pte) && PTE_ISWIRED(pte))
sys/powerpc/booke/pmap.c
2134
pte_t *pte;
sys/powerpc/booke/pmap.c
2170
pte = pte_find(kernel_pmap, va);
sys/powerpc/booke/pmap.c
2171
if (pte != NULL && PTE_ISVALID(pte))
sys/powerpc/booke/pmap.c
2183
pte = pte_find(kernel_pmap, va);
sys/powerpc/booke/pmap.c
2184
if (pte == NULL || !PTE_ISVALID(pte))
sys/powerpc/booke/pmap.c
2374
pte_t *pte;
sys/powerpc/booke/pmap.c
2428
pte = pte_find(kernel_pmap, va);
sys/powerpc/booke/pmap.c
2429
if (pte == NULL || !PTE_ISVALID(pte))
sys/powerpc/booke/pmap.c
2436
pte = pte_find(kernel_pmap, va);
sys/powerpc/booke/pmap.c
2437
*pte &= ~(PTE_MAS2_MASK << PTE_MAS2_SHIFT);
sys/powerpc/booke/pmap.c
2438
*pte |= tlb_calc_wimg(PTE_PA(pte), mode) << PTE_MAS2_SHIFT;
sys/powerpc/booke/pmap_32.c
435
pte_t *pte;
sys/powerpc/booke/pmap_32.c
444
pte = &ptbl[ptbl_idx];
sys/powerpc/booke/pmap_32.c
446
if (pte == NULL || !PTE_ISVALID(pte))
sys/powerpc/booke/pmap_32.c
449
if (PTE_ISWIRED(pte))
sys/powerpc/booke/pmap_32.c
453
m = PHYS_TO_VM_PAGE(PTE_PA(pte));
sys/powerpc/booke/pmap_32.c
456
if (PTE_ISMANAGED(pte)) {
sys/powerpc/booke/pmap_32.c
457
if (PTE_ISMODIFIED(pte))
sys/powerpc/booke/pmap_32.c
460
if (PTE_ISREFERENCED(pte))
sys/powerpc/booke/pmap_32.c
479
*pte = 0;
sys/powerpc/booke/pmap_32.c
504
pte_t *ptbl, *pte, pte_tmp;
sys/powerpc/booke/pmap_32.c
520
pte = &ptbl[ptbl_idx];
sys/powerpc/booke/pmap_32.c
526
pte = &pmap->pm_pdir[pdir_idx][ptbl_idx];
sys/powerpc/booke/pmap_32.c
527
if (PTE_ISVALID(pte)) {
sys/powerpc/booke/pmap_32.c
559
*pte = pte_tmp;
sys/powerpc/booke/pmap_32.c
571
pte_t *pte;
sys/powerpc/booke/pmap_32.c
573
pte = pte_find(pmap, va);
sys/powerpc/booke/pmap_32.c
574
if ((pte != NULL) && PTE_ISVALID(pte))
sys/powerpc/booke/pmap_32.c
575
pa = (PTE_PA(pte) | (va & PTE_PA_MASK));
sys/powerpc/booke/pmap_32.c
600
pte_t *pte;
sys/powerpc/booke/pmap_32.c
613
pte = &pdir[i][j];
sys/powerpc/booke/pmap_32.c
614
if (!PTE_ISVALID(pte))
sys/powerpc/booke/pmap_32.c
617
return (pte);
sys/powerpc/booke/pmap_32.c
627
pte_t *pte;
sys/powerpc/booke/pmap_32.c
650
pte = &(kernel_pmap->pm_pdir[PDIR_IDX(va)][PTBL_IDX(va)]);
sys/powerpc/booke/pmap_32.c
652
*pte = PTE_RPN_FROM_PA(kernload + (va - kernstart));
sys/powerpc/booke/pmap_32.c
653
*pte |= PTE_M | PTE_SR | PTE_SW | PTE_SX | PTE_WIRED |
sys/powerpc/booke/pmap_32.c
726
pte_t *pte;
sys/powerpc/booke/pmap_32.c
739
pte = pte_find(pm, va);
sys/powerpc/booke/pmap_32.c
740
valid = (pte != NULL && PTE_ISVALID(pte)) ? 1 : 0;
sys/powerpc/booke/pmap_32.c
742
pa = PTE_PA(pte);
sys/powerpc/booke/pmap_32.c
880
pte_t *pte;
sys/powerpc/booke/pmap_32.c
891
pte = pte_find(kernel_pmap, qaddr);
sys/powerpc/booke/pmap_32.c
893
KASSERT(*pte == 0, ("mmu_booke_quick_enter_page: PTE busy"));
sys/powerpc/booke/pmap_32.c
903
*pte = PTE_RPN_FROM_PA(paddr) | flags;
sys/powerpc/booke/pmap_32.c
915
pte_t *pte;
sys/powerpc/booke/pmap_32.c
917
pte = pte_find(kernel_pmap, addr);
sys/powerpc/booke/pmap_32.c
921
KASSERT(*pte != 0,
sys/powerpc/booke/pmap_32.c
924
*pte = 0;
sys/powerpc/booke/pmap_64.c
206
pte_t *pte;
sys/powerpc/booke/pmap_64.c
230
pte = &pm_root[i][j][k][l];
sys/powerpc/booke/pmap_64.c
231
if (!PTE_ISVALID(pte))
sys/powerpc/booke/pmap_64.c
237
return (pte);
sys/powerpc/booke/pmap_64.c
388
pte_t *pte;
sys/powerpc/booke/pmap_64.c
390
pte = pte_find(pmap, va);
sys/powerpc/booke/pmap_64.c
391
KASSERT(pte != NULL, ("%s: NULL pte for va %#jx, pmap %p",
sys/powerpc/booke/pmap_64.c
394
if (!PTE_ISVALID(pte))
sys/powerpc/booke/pmap_64.c
398
m = PHYS_TO_VM_PAGE(PTE_PA(pte));
sys/powerpc/booke/pmap_64.c
400
if (PTE_ISWIRED(pte))
sys/powerpc/booke/pmap_64.c
404
if (PTE_ISMANAGED(pte)) {
sys/powerpc/booke/pmap_64.c
406
if (PTE_ISMODIFIED(pte))
sys/powerpc/booke/pmap_64.c
410
if (PTE_ISREFERENCED(pte))
sys/powerpc/booke/pmap_64.c
424
*pte = 0;
sys/powerpc/booke/pmap_64.c
445
pte_t *ptbl, *pte, pte_tmp;
sys/powerpc/booke/pmap_64.c
455
pte = &ptbl[ptbl_idx];
sys/powerpc/booke/pmap_64.c
461
pte = &ptbl[ptbl_idx];
sys/powerpc/booke/pmap_64.c
462
if (PTE_ISVALID(pte)) {
sys/powerpc/booke/pmap_64.c
494
*pte = pte_tmp;
sys/powerpc/booke/pmap_64.c
507
pte_t *pte;
sys/powerpc/booke/pmap_64.c
509
pte = pte_find(pmap, va);
sys/powerpc/booke/pmap_64.c
510
if ((pte != NULL) && PTE_ISVALID(pte))
sys/powerpc/booke/pmap_64.c
511
pa = (PTE_PA(pte) | (va & PTE_PA_MASK));
sys/powerpc/booke/pmap_64.c
519
pte_t *pte;
sys/powerpc/booke/pmap_64.c
564
pte = &(kernel_pmap->pm_root[PG_ROOT_IDX(va)][PDIR_L1_IDX(va)][PDIR_IDX(va)][PTBL_IDX(va)]);
sys/powerpc/booke/pmap_64.c
565
*pte = PTE_RPN_FROM_PA(kernload + (va - kernstart));
sys/powerpc/booke/pmap_64.c
566
*pte |= PTE_M | PTE_SR | PTE_SW | PTE_SX | PTE_WIRED |
sys/powerpc/booke/pmap_64.c
636
pte_t *pte;
sys/powerpc/booke/pmap_64.c
642
pte = pte_find(pm, va);
sys/powerpc/booke/pmap_64.c
643
valid = (pte != NULL && PTE_ISVALID(pte)) ? 1 : 0;
sys/powerpc/booke/pmap_64.c
645
pa = PTE_PA(pte);
sys/powerpc/include/pmap.h
122
pte_t pte;
sys/powerpc/include/pte.h
297
#define PTE_TSIZE(pte) (int)((*pte >> PTE_TSIZE_SHIFT) & PTE_TSIZE_MASK)
sys/powerpc/include/pte.h
298
#define PTE_TSIZE_DIRECT(pte) (int)((*pte >> PTE_TSIZE_SHIFT_DIRECT) & PTE_TSIZE_MASK_DIRECT)
sys/powerpc/include/pte.h
304
#define PTE_PA(pte) ((vm_paddr_t)(*pte >> PTE_ARPN_SHIFT) & ~PAGE_MASK)
sys/powerpc/include/pte.h
305
#define PTE_ISVALID(pte) ((*pte) & PTE_VALID)
sys/powerpc/include/pte.h
306
#define PTE_ISWIRED(pte) ((*pte) & PTE_WIRED)
sys/powerpc/include/pte.h
307
#define PTE_ISMANAGED(pte) ((*pte) & PTE_MANAGED)
sys/powerpc/include/pte.h
308
#define PTE_ISMODIFIED(pte) ((*pte) & PTE_MODIFIED)
sys/powerpc/include/pte.h
309
#define PTE_ISREFERENCED(pte) ((*pte) & PTE_REFERENCED)
sys/powerpc/include/pte.h
53
struct pte pt[8];
sys/powerpc/include/pte.h
78
typedef struct pte pte_t;
sys/powerpc/ps3/mmu_ps3.c
200
struct lpte pte;
sys/powerpc/ps3/mmu_ps3.c
210
moea64_pte_from_pvo(pvo, &pte);
sys/powerpc/ps3/mmu_ps3.c
212
pte.pte_lo |= refchg;
sys/powerpc/ps3/mmu_ps3.c
213
pte.pte_lo &= ~ptebit;
sys/powerpc/ps3/mmu_ps3.c
215
lv1_write_htab_entry(mps3_vas_id, pvo->pvo_pte.slot, pte.pte_hi,
sys/powerpc/ps3/mmu_ps3.c
216
pte.pte_lo);
sys/powerpc/ps3/mmu_ps3.c
246
struct lpte pte, evicted;
sys/powerpc/ps3/mmu_ps3.c
256
moea64_pte_from_pvo(pvo, &pte);
sys/powerpc/ps3/mmu_ps3.c
261
pte.pte_hi, pte.pte_lo, LPTE_LOCKED | LPTE_WIRED, 0,
sys/powerpc/pseries/mmu_phyp.c
298
struct lpte pte;
sys/powerpc/pseries/mmu_phyp.c
302
phyp_pft_hcall(H_READ, 0, pvo->pvo_pte.slot, 0, 0, &pte.pte_hi,
sys/powerpc/pseries/mmu_phyp.c
303
&pte.pte_lo, &junk);
sys/powerpc/pseries/mmu_phyp.c
304
if ((pte.pte_hi & LPTE_AVPN_MASK) !=
sys/powerpc/pseries/mmu_phyp.c
308
if (!(pte.pte_hi & LPTE_VALID))
sys/powerpc/pseries/mmu_phyp.c
311
return (pte.pte_lo & (LPTE_CHG | LPTE_REF));
sys/powerpc/pseries/mmu_phyp.c
371
struct lpte pte;
sys/powerpc/pseries/mmu_phyp.c
377
moea64_pte_from_pvo(pvo, &pte);
sys/powerpc/pseries/mmu_phyp.c
380
pte.pte_hi & LPTE_AVPN_MASK, 0, &pte.pte_hi, &pte.pte_lo,
sys/powerpc/pseries/mmu_phyp.c
390
return (pte.pte_lo & (LPTE_REF | LPTE_CHG));
sys/powerpc/pseries/mmu_phyp.c
430
mphyp_pte_insert_locked(struct pvo_entry *pvo, struct lpte *pte)
sys/powerpc/pseries/mmu_phyp.c
440
result = phyp_pft_hcall(H_ENTER, 0, pvo->pvo_pte.slot, pte->pte_hi,
sys/powerpc/pseries/mmu_phyp.c
441
pte->pte_lo, &index, &evicted.pte_lo, &junk);
sys/powerpc/pseries/mmu_phyp.c
448
moea64_pteg_count, pte->pte_hi, pte->pte_lo));
sys/powerpc/pseries/mmu_phyp.c
454
pte->pte_hi ^= LPTE_HID;
sys/powerpc/pseries/mmu_phyp.c
458
pte->pte_hi, pte->pte_lo, &index, &evicted.pte_lo, &junk);
sys/powerpc/pseries/mmu_phyp.c
471
mphyp_pte_evict_and_insert_locked(struct pvo_entry *pvo, struct lpte *pte)
sys/powerpc/pseries/mmu_phyp.c
483
pte->pte_hi ^= LPTE_HID;
sys/powerpc/pseries/mmu_phyp.c
508
result = phyp_pft_hcall(H_ENTER, H_EXACT, index, pte->pte_hi,
sys/powerpc/pseries/mmu_phyp.c
509
pte->pte_lo, &index, &evicted.pte_lo, &junk);
sys/powerpc/pseries/mmu_phyp.c
525
struct lpte pte;
sys/powerpc/pseries/mmu_phyp.c
530
moea64_pte_from_pvo(pvo, &pte);
sys/powerpc/pseries/mmu_phyp.c
535
ret = mphyp_pte_insert_locked(pvo, &pte);
sys/powerpc/pseries/mmu_phyp.c
545
ret = mphyp_pte_evict_and_insert_locked(pvo, &pte);
sys/powerpc/pseries/mmu_phyp.c
585
struct lpte pte;
sys/powerpc/pseries/mmu_phyp.c
601
moea64_pte_from_pvo(pvo, &pte);
sys/powerpc/pseries/mmu_phyp.c
604
pte.pte_hi & LPTE_AVPN_MASK, 0, &pte.pte_hi, &pte.pte_lo,
sys/powerpc/pseries/mmu_phyp.c
611
refchg |= pte.pte_lo & (LPTE_REF | LPTE_CHG);
sys/powerpc/pseries/mmu_phyp.c
622
struct lpte pte;
sys/powerpc/pseries/mmu_phyp.c
639
moea64_pte_from_pvo(pvo, &pte);
sys/powerpc/pseries/mmu_phyp.c
641
ret = mphyp_pte_insert_locked(pvo, &pte);
sys/powerpc/pseries/mmu_phyp.c
650
mphyp_pte_evict_and_insert_locked(pvo, &pte);
sys/riscv/riscv/pmap.c
2912
pt_entry_t bits, *pte, oldpte;
sys/riscv/riscv/pmap.c
2923
pte = pmap_l2_to_l3(l2, va);
sys/riscv/riscv/pmap.c
2924
if (((oldpte = pmap_load(pte)) & PTE_V) == 0)
sys/riscv/riscv/pmap.c
2927
pte = l2;
sys/riscv/riscv/pmap.c
2947
pmap_store_bits(pte, bits);
sys/riscv/riscv/pmap.c
344
#define pmap_clear(pte) pmap_store(pte, 0)
sys/riscv/riscv/pmap.c
345
#define pmap_clear_bits(pte, bits) atomic_clear_64(pte, bits)
sys/riscv/riscv/pmap.c
346
#define pmap_load_store(pte, entry) atomic_swap_64(pte, entry)
sys/riscv/riscv/pmap.c
347
#define pmap_load_clear(pte) pmap_load_store(pte, 0)
sys/riscv/riscv/pmap.c
348
#define pmap_load(pte) atomic_load_64(pte)
sys/riscv/riscv/pmap.c
349
#define pmap_store(pte, entry) atomic_store_64(pte, entry)
sys/riscv/riscv/pmap.c
350
#define pmap_store_bits(pte, bits) atomic_set_64(pte, bits)
sys/riscv/riscv/pmap.c
3543
pt_entry_t *pt_end, *pte;
sys/riscv/riscv/pmap.c
3546
pte = (pt_entry_t *)PHYS_TO_DMAP(pa);
sys/riscv/riscv/pmap.c
3547
for (pt_end = pte + Ln_ENTRIES; pte < pt_end; pte++) {
sys/riscv/riscv/pmap.c
3548
if (*pte != 0)
sys/riscv/riscv/pmap.c
375
#define PTE_TO_PHYS(pte) \
sys/riscv/riscv/pmap.c
376
((((pte) & ~PTE_HI_MASK) >> PTE_PPN0_S) * PAGE_SIZE)
sys/riscv/riscv/pmap.c
381
#define PTE_TO_VM_PAGE(pte) PHYS_TO_VM_PAGE(PTE_TO_PHYS(pte))
sys/riscv/riscv/pmap.c
4323
pt_entry_t *pte, tpte;
sys/riscv/riscv/pmap.c
4350
pte = pmap_l1(pmap, pv->pv_va);
sys/riscv/riscv/pmap.c
4351
ptepde = pmap_load(pte);
sys/riscv/riscv/pmap.c
4352
pte = pmap_l1_to_l2(pte, pv->pv_va);
sys/riscv/riscv/pmap.c
4353
tpte = pmap_load(pte);
sys/riscv/riscv/pmap.c
4362
pte = pmap_l2_to_l3(pte, pv->pv_va);
sys/riscv/riscv/pmap.c
4363
tpte = pmap_load(pte);
sys/riscv/riscv/pmap.c
4382
pmap_clear(pte);
sys/riscv/riscv/pmap.c
919
pt_entry_t *pte;
sys/riscv/riscv/pmap.c
981
pte = pmap_l3(kernel_pmap, dpcpu);
sys/riscv/riscv/pmap.c
982
KASSERT(pte != NULL, ("Bootstrap pages missing"));
sys/riscv/riscv/pmap.c
984
pmap_store(&pte[i], L3_PTE(pa + ptoa(i), PTE_KERN |
sys/riscv/riscv/pmap.c
995
pte = pmap_l3(kernel_pmap, msgbufpv);
sys/riscv/riscv/pmap.c
996
KASSERT(pte != NULL, ("Bootstrap pages missing"));
sys/riscv/riscv/pmap.c
998
pmap_store(&pte[i], L3_PTE(pa + ptoa(i), PTE_KERN |
sys/x86/iommu/amd_idpgtbl.c
134
iommu_pte_t *pte, *ptep;
sys/x86/iommu/amd_idpgtbl.c
141
pte = (iommu_pte_t *)sf_buf_kva(*sf);
sys/x86/iommu/amd_idpgtbl.c
147
pte = iommu_map_pgtbl(domain->pgtbl_obj, idx, flags, sf);
sys/x86/iommu/amd_idpgtbl.c
148
if (pte == NULL) {
sys/x86/iommu/amd_idpgtbl.c
174
ptep->pte = VM_PAGE_TO_PHYS(m) | AMDIOMMU_PTE_IR |
sys/x86/iommu/amd_idpgtbl.c
184
pte += pglvl_pgtbl_pte_off(domain->pglvl, base, lvl);
sys/x86/iommu/amd_idpgtbl.c
185
return (pte);
sys/x86/iommu/amd_idpgtbl.c
193
iommu_pte_t *pte;
sys/x86/iommu/amd_idpgtbl.c
203
pte = NULL;
sys/x86/iommu/amd_idpgtbl.c
211
pte = amdiommu_pgtbl_map_pte(domain, base, domain->pglvl - 1,
sys/x86/iommu/amd_idpgtbl.c
213
if (pte == NULL) {
sys/x86/iommu/amd_idpgtbl.c
223
pte->pte = VM_PAGE_TO_PHYS(ma[pi]) | pflags | AMDIOMMU_PTE_PR;
sys/x86/iommu/amd_idpgtbl.c
305
int lvl, int flags, iommu_pte_t *pte, struct sf_buf **sf,
sys/x86/iommu/amd_idpgtbl.c
310
pte->pte = 0;
sys/x86/iommu/amd_idpgtbl.c
332
iommu_pte_t *pte;
sys/x86/iommu/amd_idpgtbl.c
364
pte = amdiommu_pgtbl_map_pte(domain, base,
sys/x86/iommu/amd_idpgtbl.c
366
KASSERT(pte != NULL,
sys/x86/iommu/amd_idpgtbl.c
370
flags, pte, &sf, entry, false);
sys/x86/iommu/amd_idpgtbl.c
72
iommu_gaddr_t base, int lvl, int flags, iommu_pte_t *pte,
sys/x86/iommu/intel_idpgtbl.c
116
iommu_pte_t *pte;
sys/x86/iommu/intel_idpgtbl.c
134
pte = iommu_map_pgtbl(tbl->pgtbl_obj, idx, IOMMU_PGF_WAITOK, &sf);
sys/x86/iommu/intel_idpgtbl.c
139
pte[i].pte = (DMAR_PTE_ADDR_MASK & f) |
sys/x86/iommu/intel_idpgtbl.c
149
pte[i].pte = (DMAR_PTE_ADDR_MASK &
sys/x86/iommu/intel_idpgtbl.c
333
iommu_pte_t *pte, *ptep;
sys/x86/iommu/intel_idpgtbl.c
341
pte = (iommu_pte_t *)sf_buf_kva(*sf);
sys/x86/iommu/intel_idpgtbl.c
347
pte = iommu_map_pgtbl(domain->pgtbl_obj, idx, flags, sf);
sys/x86/iommu/intel_idpgtbl.c
348
if (pte == NULL) {
sys/x86/iommu/intel_idpgtbl.c
381
dmar_pte_store(&ptep->pte, DMAR_PTE_R | DMAR_PTE_W |
sys/x86/iommu/intel_idpgtbl.c
391
pte += pglvl_pgtbl_pte_off(domain->pglvl, base, lvl);
sys/x86/iommu/intel_idpgtbl.c
392
return (pte);
sys/x86/iommu/intel_idpgtbl.c
400
iommu_pte_t *pte;
sys/x86/iommu/intel_idpgtbl.c
451
pte = dmar_pgtbl_map_pte(domain, base, lvl, flags, &idx, &sf);
sys/x86/iommu/intel_idpgtbl.c
452
if (pte == NULL) {
sys/x86/iommu/intel_idpgtbl.c
462
dmar_pte_store(&pte->pte, VM_PAGE_TO_PHYS(ma[pi]) | pflags |
sys/x86/iommu/intel_idpgtbl.c
464
dmar_flush_pte_to_ram(domain->dmar, pte);
sys/x86/iommu/intel_idpgtbl.c
547
iommu_gaddr_t base, int lvl, int flags, iommu_pte_t *pte,
sys/x86/iommu/intel_idpgtbl.c
566
int flags, iommu_pte_t *pte, struct sf_buf **sf,
sys/x86/iommu/intel_idpgtbl.c
571
dmar_pte_clear(&pte->pte);
sys/x86/iommu/intel_idpgtbl.c
572
dmar_flush_pte_to_ram(domain->dmar, pte);
sys/x86/iommu/intel_idpgtbl.c
597
iommu_pte_t *pte;
sys/x86/iommu/intel_idpgtbl.c
638
pte = dmar_pgtbl_map_pte(domain, base, lvl, flags,
sys/x86/iommu/intel_idpgtbl.c
640
KASSERT(pte != NULL,
sys/x86/iommu/intel_idpgtbl.c
643
if ((pte->pte & DMAR_PTE_SP) != 0 ||
sys/x86/iommu/intel_idpgtbl.c
646
flags, pte, &sf, entry, false);
sys/x86/iommu/x86_iommu.h
36
uint64_t pte;
usr.sbin/bhyvectl/amd64/bhyvectl_machdep.c
1549
uint64_t info[2], pteval[4], *pte;
usr.sbin/bhyvectl/amd64/bhyvectl_machdep.c
1849
pte = &pteval[0];
usr.sbin/bhyvectl/amd64/bhyvectl_machdep.c
1851
printf(" %#lx", *pte++);