lib/libkvm/kvm_alpha.c
133
if (_kvm_pread(kd, kd->pmfd, (char *)&pte, sizeof(pte),
lib/libkvm/kvm_alpha.c
134
(off_t)_kvm_pa2off(kd, pteoff)) != sizeof(pte)) {
lib/libkvm/kvm_alpha.c
140
if ((pte & ALPHA_PTE_VALID) == 0) {
lib/libkvm/kvm_alpha.c
144
pteoff = ALPHA_PTE_TO_PFN(pte) * cpu_kh->page_size +
lib/libkvm/kvm_alpha.c
146
if (_kvm_pread(kd, kd->pmfd, (char *)&pte, sizeof(pte),
lib/libkvm/kvm_alpha.c
147
(off_t)_kvm_pa2off(kd, pteoff)) != sizeof(pte)) {
lib/libkvm/kvm_alpha.c
153
if ((pte & ALPHA_PTE_VALID) == 0) {
lib/libkvm/kvm_alpha.c
157
pteoff = ALPHA_PTE_TO_PFN(pte) * cpu_kh->page_size +
lib/libkvm/kvm_alpha.c
159
if (_kvm_pread(kd, kd->pmfd, (char *)&pte, sizeof(pte),
lib/libkvm/kvm_alpha.c
160
(off_t)_kvm_pa2off(kd, pteoff)) != sizeof(pte)) {
lib/libkvm/kvm_alpha.c
166
if ((pte & ALPHA_PTE_VALID) == 0) {
lib/libkvm/kvm_alpha.c
170
*pa = ALPHA_PTE_TO_PFN(pte) * cpu_kh->page_size + page_off;
lib/libkvm/kvm_alpha.c
97
alpha_pt_entry_t pte;
lib/libkvm/kvm_amd64.c
163
if (pread(kd->pmfd, (void *) &pte, sizeof(pte),
lib/libkvm/kvm_amd64.c
164
_kvm_pa2off(kd, pte_pa)) != sizeof(pte)) {
lib/libkvm/kvm_amd64.c
171
if ((pte & PG_V) == 0) {
lib/libkvm/kvm_amd64.c
175
*pa = (pte & PG_FRAME) + page_off;
lib/libkvm/kvm_amd64.c
88
pt_entry_t pte;
lib/libkvm/kvm_i386.c
184
ptd_entry_t pte;
lib/libkvm/kvm_i386.c
214
if (_kvm_pread(kd, kd->pmfd, &pte, sizeof pte,
lib/libkvm/kvm_i386.c
215
_kvm_pa2off(kd, pte_pa)) != sizeof pte)
lib/libkvm/kvm_i386.c
218
if ((pte & PG_V) == 0)
lib/libkvm/kvm_i386.c
220
*pa = (pte & PG_FRAME) + offset;
lib/libkvm/kvm_mips64.c
143
pt_entry_t pte;
lib/libkvm/kvm_mips64.c
182
if (_kvm_pread(kd, kd->pmfd, (char *)&pte, sizeof(pte),
lib/libkvm/kvm_mips64.c
185
if (!(pte & PG_V))
lib/libkvm/kvm_mips64.c
187
*pa = (pte & PG_FRAME) | (paddr_t)offset;
regress/usr.sbin/bgpd/unittests/rde_decide_test.c
406
pt_getaddr(struct pt_entry *pte, struct bgpd_addr *addr)
sys/arch/alpha/alpha/pmap.c
1468
pt_entry_t *pte, npte, opte;
sys/arch/alpha/alpha/pmap.c
1514
pte = PMAP_KERNEL_PTE(va);
sys/arch/alpha/alpha/pmap.c
1579
pte = pmap_l3pte(pmap, va, l2pte);
sys/arch/alpha/alpha/pmap.c
1583
opte = *pte;
sys/arch/alpha/alpha/pmap.c
1589
if (pmap_pte_v(pte) == 0) {
sys/arch/alpha/alpha/pmap.c
1608
pmap_physpage_addref(pte);
sys/arch/alpha/alpha/pmap.c
1613
opa = pmap_pte_pa(pte);
sys/arch/alpha/alpha/pmap.c
1614
hadasm = (pmap_pte_asm(pte) != 0);
sys/arch/alpha/alpha/pmap.c
1621
if (pmap_pte_w_chg(pte, wired ? PG_WIRED : 0)) {
sys/arch/alpha/alpha/pmap.c
1657
pmap_physpage_addref(pte);
sys/arch/alpha/alpha/pmap.c
1659
needisync |= pmap_remove_mapping(pmap, va, pte, TRUE, cpu_id);
sys/arch/alpha/alpha/pmap.c
1666
error = pmap_pv_enter(pmap, pg, va, pte, TRUE);
sys/arch/alpha/alpha/pmap.c
1668
pmap_l3pt_delref(pmap, va, pte, cpu_id);
sys/arch/alpha/alpha/pmap.c
1732
PMAP_SET_PTE(pte, npte);
sys/arch/alpha/alpha/pmap.c
1765
pt_entry_t *pte, npte;
sys/arch/alpha/alpha/pmap.c
1785
pte = PMAP_KERNEL_PTE(va);
sys/arch/alpha/alpha/pmap.c
1787
if (pmap_pte_v(pte) == 0)
sys/arch/alpha/alpha/pmap.c
1789
if (pmap_pte_w(pte) == 0)
sys/arch/alpha/alpha/pmap.c
1792
if ((prot & PROT_EXEC) != 0 || pmap_pte_exec(pte))
sys/arch/alpha/alpha/pmap.c
1804
PMAP_SET_PTE(pte, npte);
sys/arch/alpha/alpha/pmap.c
1830
pt_entry_t *pte;
sys/arch/alpha/alpha/pmap.c
1848
pte = PMAP_KERNEL_PTE(va);
sys/arch/alpha/alpha/pmap.c
1849
if (pmap_pte_v(pte)) {
sys/arch/alpha/alpha/pmap.c
1851
if (pmap_pte_pv(pte))
sys/arch/alpha/alpha/pmap.c
1855
if (pmap_pte_exec(pte))
sys/arch/alpha/alpha/pmap.c
1859
PMAP_SET_PTE(pte, PG_NV);
sys/arch/alpha/alpha/pmap.c
1888
pt_entry_t *pte;
sys/arch/alpha/alpha/pmap.c
1897
pte = pmap_l3pte(pmap, va, NULL);
sys/arch/alpha/alpha/pmap.c
1899
if (pte == NULL || pmap_pte_v(pte) == 0)
sys/arch/alpha/alpha/pmap.c
1908
if (pmap_pte_w_chg(pte, 0)) {
sys/arch/alpha/alpha/pmap.c
1909
pmap_pte_set_w(pte, FALSE);
sys/arch/alpha/alpha/pmap.c
2280
pmap_remove_mapping(pmap_t pmap, vaddr_t va, pt_entry_t *pte,
sys/arch/alpha/alpha/pmap.c
2294
pmap, va, pte, dolock, cpu_id);
sys/arch/alpha/alpha/pmap.c
2300
if (pte == PT_ENTRY_NULL) {
sys/arch/alpha/alpha/pmap.c
2301
pte = pmap_l3pte(pmap, va, NULL);
sys/arch/alpha/alpha/pmap.c
2302
if (pmap_pte_v(pte) == 0)
sys/arch/alpha/alpha/pmap.c
2306
pa = pmap_pte_pa(pte);
sys/arch/alpha/alpha/pmap.c
2307
onpv = (pmap_pte_pv(pte) != 0);
sys/arch/alpha/alpha/pmap.c
2318
hadasm = (pmap_pte_asm(pte) != 0);
sys/arch/alpha/alpha/pmap.c
2327
if (pmap_pte_exec(pte)) {
sys/arch/alpha/alpha/pmap.c
2339
if (pmap_pte_w(pte))
sys/arch/alpha/alpha/pmap.c
2348
printf("remove: invalidating pte at %p\n", pte);
sys/arch/alpha/alpha/pmap.c
2350
PMAP_SET_PTE(pte, PG_NV);
sys/arch/alpha/alpha/pmap.c
2366
pmap_l3pt_delref(pmap, va, pte, cpu_id);
sys/arch/alpha/alpha/pmap.c
2386
pt_entry_t *pte, npte;
sys/arch/alpha/alpha/pmap.c
2405
pte = pv->pv_pte;
sys/arch/alpha/alpha/pmap.c
2406
npte = (*pte | set) & mask;
sys/arch/alpha/alpha/pmap.c
2407
if (*pte != npte) {
sys/arch/alpha/alpha/pmap.c
2408
hadasm = (pmap_pte_asm(pte) != 0);
sys/arch/alpha/alpha/pmap.c
2410
PMAP_SET_PTE(pte, npte);
sys/arch/alpha/alpha/pmap.c
2432
pt_entry_t faultoff, *pte;
sys/arch/alpha/alpha/pmap.c
2454
pte = PMAP_KERNEL_PTE(v);
sys/arch/alpha/alpha/pmap.c
2465
pte = pmap_l3pte(pmap, v, NULL);
sys/arch/alpha/alpha/pmap.c
2470
if (pte == NULL || !pmap_pte_v(pte)) {
sys/arch/alpha/alpha/pmap.c
2475
exec = pmap_pte_exec(pte);
sys/arch/alpha/alpha/pmap.c
2483
printf("\tpte = %p, ", pte);
sys/arch/alpha/alpha/pmap.c
2484
printf("*pte = 0x%lx\n", *pte);
sys/arch/alpha/alpha/pmap.c
2508
if (!(*pte & (user ? PG_UWE : PG_UWE | PG_KWE))) {
sys/arch/alpha/alpha/pmap.c
2511
user, type, *pte);
sys/arch/alpha/alpha/pmap.c
2513
if (!(*pte & PG_FOW)) {
sys/arch/alpha/alpha/pmap.c
2516
user, type, *pte);
sys/arch/alpha/alpha/pmap.c
2519
if (!(*pte & (user ? PG_URE : PG_URE | PG_KRE))) {
sys/arch/alpha/alpha/pmap.c
2522
user, type, *pte);
sys/arch/alpha/alpha/pmap.c
2524
if (!(*pte & (PG_FOR | PG_FOE))) {
sys/arch/alpha/alpha/pmap.c
2527
user, type, *pte);
sys/arch/alpha/alpha/pmap.c
2533
pa = pmap_pte_pa(pte);
sys/arch/alpha/alpha/pmap.c
2553
p, v, user, type, pa, pte, *pte);
sys/arch/alpha/alpha/pmap.c
2619
pt_entry_t *pte;
sys/arch/alpha/alpha/pmap.c
2627
pte = PMAP_KERNEL_PTE(vaddr);
sys/arch/alpha/alpha/pmap.c
2628
if (pmap_pte_v(pte))
sys/arch/alpha/alpha/pmap.c
2629
paddr = pmap_pte_pa(pte) | (vaddr & PGOFSET);
sys/arch/alpha/alpha/pmap.c
2647
pmap_pv_enter(pmap_t pmap, struct vm_page *pg, vaddr_t va, pt_entry_t *pte,
sys/arch/alpha/alpha/pmap.c
2660
newpv->pv_pte = pte;
sys/arch/alpha/alpha/pmap.c
2873
pt_entry_t *l1pte, *l2pte, pte;
sys/arch/alpha/alpha/pmap.c
2906
pte = (atop(ptaddr) << PG_SHIFT) |
sys/arch/alpha/alpha/pmap.c
2908
*l1pte = pte;
sys/arch/alpha/alpha/pmap.c
2922
pm->pm_lev1map[l1idx] = pte;
sys/arch/alpha/alpha/pmap.c
3024
pt_entry_t pte;
sys/arch/alpha/alpha/pmap.c
3042
pte = ((ALPHA_K0SEG_TO_PHYS((vaddr_t) l1pt) >> PGSHIFT) << PG_SHIFT) |
sys/arch/alpha/alpha/pmap.c
3044
l1pt[l1pte_index(VPTBASE)] = pte;
sys/arch/alpha/alpha/pmap.c
3092
pmap_ptpage_alloc(pmap_t pmap, pt_entry_t *pte, int usage)
sys/arch/alpha/alpha/pmap.c
3105
PMAP_SET_PTE(pte, ((ptpa >> PGSHIFT) << PG_SHIFT) |
sys/arch/alpha/alpha/pmap.c
3121
pmap_ptpage_free(pmap_t pmap, pt_entry_t *pte)
sys/arch/alpha/alpha/pmap.c
3129
ptpa = pmap_pte_pa(pte);
sys/arch/alpha/alpha/pmap.c
3130
PMAP_SET_PTE(pte, PG_NV);
sys/arch/alpha/alpha/pmap.c
3407
pmap_t pmap, vaddr_t va, pt_entry_t pte)
sys/arch/alpha/alpha/pmap.c
3425
pj->pj_pte = pte;
sys/arch/alpha/alpha/pmap.c
3435
atomic_setbits_ulong(&pq->pq_pte, (1UL << 32) | pte);
sys/arch/alpha/alpha/pmap.c
3439
pmap_tlb_shootdown(pmap_t pmap, vaddr_t va, pt_entry_t pte, u_long *cpumaskp)
sys/arch/alpha/alpha/pmap.c
3476
pmap_tlb_shootdown_job(pq, pmap, va, pte);
sys/arch/alpha/alpha/pmap.c
3507
unsigned long pte;
sys/arch/alpha/alpha/pmap.c
3509
pte = atomic_swap_ulong(&pq->pq_pte, 0);
sys/arch/alpha/alpha/pmap.c
3510
if (pte != 0) {
sys/arch/alpha/alpha/pmap.c
3520
if (pte & PG_ASM)
sys/arch/alpha/alpha/pmap.c
709
pt_entry_t pte;
sys/arch/alpha/alpha/pmap.c
763
pte = (ALPHA_K0SEG_TO_PHYS(((vaddr_t)lev2map) +
sys/arch/alpha/alpha/pmap.c
765
pte |= PG_V | PG_ASM | PG_KRE | PG_KWE | PG_WIRED;
sys/arch/alpha/alpha/pmap.c
767
(i*PAGE_SIZE*NPTEPG*NPTEPG))] = pte;
sys/arch/alpha/alpha/pmap.c
771
pte = (ALPHA_K0SEG_TO_PHYS((vaddr_t)kernel_lev1map) >> PGSHIFT)
sys/arch/alpha/alpha/pmap.c
773
pte |= PG_V | PG_KRE | PG_KWE; /* NOTE NO ASM */
sys/arch/alpha/alpha/pmap.c
774
kernel_lev1map[l1pte_index(VPTBASE)] = pte;
sys/arch/alpha/alpha/pmap.c
782
pte = (ALPHA_K0SEG_TO_PHYS(((vaddr_t)lev3map) +
sys/arch/alpha/alpha/pmap.c
784
pte |= PG_V | PG_ASM | PG_KRE | PG_KWE | PG_WIRED;
sys/arch/alpha/alpha/pmap.c
786
(i*PAGE_SIZE*NPTEPG))] = pte;
sys/arch/alpha/dev/sgmap_typedep.c
146
pte = &page_table[pteidx * SGMAP_PTE_SPACING];
sys/arch/alpha/dev/sgmap_typedep.c
151
"pte = %p (pt = %p)\n", sgva, pteidx, pte,
sys/arch/alpha/dev/sgmap_typedep.c
167
pte = &page_table[pteidx * SGMAP_PTE_SPACING]) {
sys/arch/alpha/dev/sgmap_typedep.c
172
*pte = (pa >> SGPTE_PGADDR_SHIFT) | SGPTE_VALID;
sys/arch/alpha/dev/sgmap_typedep.c
176
"*pte = 0x%lx\n", pa, pte, (u_long)(*pte));
sys/arch/alpha/dev/sgmap_typedep.c
182
*pte = __C(SGMAP_TYPE,_prefetch_spill_page_pte);
sys/arch/alpha/dev/sgmap_typedep.c
186
"*pte = 0x%lx\n", pte, *pte);
sys/arch/alpha/dev/sgmap_typedep.c
398
SGMAP_PTE_TYPE *pte, *page_table = sgmap->aps_pt;
sys/arch/alpha/dev/sgmap_typedep.c
424
pte = &page_table[pteidx * SGMAP_PTE_SPACING];
sys/arch/alpha/dev/sgmap_typedep.c
428
"*pte = 0x%lx\n", pte, (u_long)(*pte));
sys/arch/alpha/dev/sgmap_typedep.c
430
*pte = 0;
sys/arch/alpha/dev/sgmap_typedep.c
64
SGMAP_PTE_TYPE *pte, *page_table = sgmap->aps_pt;
sys/arch/alpha/include/alpha_cpu.h
228
#define ALPHA_PTE_TO_PFN(pte) ((pte) >> 32)
sys/arch/alpha/include/pmap.h
145
#define PMAP_TLB_SHOOTDOWN(pm, va, pte) \
sys/arch/alpha/include/pmap.h
146
pmap_tlb_shootdown((pm), (va), (pte), &shootset)
sys/arch/alpha/include/pmap.h
151
#define PMAP_TLB_SHOOTDOWN(pm, va, pte) /* nothing */
sys/arch/alpha/include/pmap.h
185
#define pmap_pte_pa(pte) (PG_PFNUM(*(pte)) << PAGE_SHIFT)
sys/arch/alpha/include/pmap.h
186
#define pmap_pte_prot(pte) (*(pte) & PG_PROT)
sys/arch/alpha/include/pmap.h
187
#define pmap_pte_w(pte) (*(pte) & PG_WIRED)
sys/arch/alpha/include/pmap.h
188
#define pmap_pte_v(pte) (*(pte) & PG_V)
sys/arch/alpha/include/pmap.h
189
#define pmap_pte_pv(pte) (*(pte) & PG_PVLIST)
sys/arch/alpha/include/pmap.h
190
#define pmap_pte_asm(pte) (*(pte) & PG_ASM)
sys/arch/alpha/include/pmap.h
191
#define pmap_pte_exec(pte) (*(pte) & PG_EXEC)
sys/arch/alpha/include/pmap.h
193
#define pmap_pte_set_w(pte, v) \
sys/arch/alpha/include/pmap.h
196
*(pte) |= PG_WIRED; \
sys/arch/alpha/include/pmap.h
198
*(pte) &= ~PG_WIRED; \
sys/arch/alpha/include/pmap.h
201
#define pmap_pte_w_chg(pte, nw) ((nw) ^ pmap_pte_w(pte))
sys/arch/alpha/include/pmap.h
203
#define pmap_pte_set_prot(pte, np) \
sys/arch/alpha/include/pmap.h
205
*(pte) &= ~PG_PROT; \
sys/arch/alpha/include/pmap.h
206
*(pte) |= (np); \
sys/arch/alpha/include/pmap.h
209
#define pmap_pte_prot_chg(pte, np) ((np) ^ pmap_pte_prot(pte))
sys/arch/amd64/amd64/db_memrw.c
109
pte = kvtopte(addr);
sys/arch/amd64/amd64/db_memrw.c
110
oldpte = *pte;
sys/arch/amd64/amd64/db_memrw.c
139
*pte = tmppte;
sys/arch/amd64/amd64/db_memrw.c
152
*pte = oldpte;
sys/arch/amd64/amd64/db_memrw.c
95
pt_entry_t *pte, oldpte, tmppte;
sys/arch/amd64/amd64/lapic.c
178
pt_entry_t *pte;
sys/arch/amd64/amd64/lapic.c
222
pte = kvtopte(va);
sys/arch/amd64/amd64/lapic.c
223
*pte = lapic_base | PG_RW | PG_V | PG_N | PG_G | pg_nx;
sys/arch/amd64/amd64/pmap.c
1565
pt_entry_t *ptes, pte;
sys/arch/amd64/amd64/pmap.c
1578
pte = ptes[offs];
sys/arch/amd64/amd64/pmap.c
1583
if (__predict_true(level == 0 && pmap_valid_entry(pte))) {
sys/arch/amd64/amd64/pmap.c
1585
*pap = (pte & pg_frame) | (va & PAGE_MASK);
sys/arch/amd64/amd64/pmap.c
1588
if (level == 1 && (pte & (PG_PS|PG_V)) == (PG_PS|PG_V)) {
sys/arch/amd64/amd64/pmap.c
1590
*pap = (pte & pg_lgframe) | (va & PAGE_MASK_L2);
sys/arch/amd64/amd64/pmap.c
1658
pt_entry_t *pte = (pt_entry_t *) ptpva;
sys/arch/amd64/amd64/pmap.c
1672
; pte++, startva += PAGE_SIZE) {
sys/arch/amd64/amd64/pmap.c
1673
if (!pmap_valid_entry(*pte))
sys/arch/amd64/amd64/pmap.c
1675
if ((flags & PMAP_REMOVE_SKIPWIRED) && (*pte & PG_W)) {
sys/arch/amd64/amd64/pmap.c
1680
opte = pmap_pte_set(pte, 0);
sys/arch/amd64/amd64/pmap.c
1733
pmap_remove_pte(struct pmap *pmap, struct vm_page *ptp, pt_entry_t *pte,
sys/arch/amd64/amd64/pmap.c
1740
if (!pmap_valid_entry(*pte))
sys/arch/amd64/amd64/pmap.c
1742
if ((flags & PMAP_REMOVE_SKIPWIRED) && (*pte & PG_W)) {
sys/arch/amd64/amd64/pmap.c
1747
opte = pmap_pte_set(pte, 0);
sys/arch/amd64/amd64/pmap.c
2419
pt_entry_t *pte;
sys/arch/amd64/amd64/pmap.c
2426
pte = (pt_entry_t *)pmap->pm_pdir;
sys/arch/amd64/amd64/pmap.c
2427
memset(pte, 0, PAGE_SIZE);
sys/arch/amd64/amd64/pmap.c
400
pmap_pte2flags(u_long pte)
sys/arch/amd64/amd64/pmap.c
402
return (((pte & PG_U) ? PG_PMAP_REF : 0) |
sys/arch/amd64/amd64/pmap.c
403
((pte & PG_M) ? PG_PMAP_MOD : 0));
sys/arch/amd64/amd64/pmap.c
407
pmap_sync_flags_pte(struct vm_page *pg, u_long pte)
sys/arch/amd64/amd64/pmap.c
409
if (pte & (PG_U|PG_M)) {
sys/arch/amd64/amd64/pmap.c
410
atomic_setbits_int(&pg->pg_flags, pmap_pte2flags(pte));
sys/arch/amd64/amd64/pmap.c
507
pt_entry_t *pte, opte, npte;
sys/arch/amd64/amd64/pmap.c
509
pte = kvtopte(va);
sys/arch/amd64/amd64/pmap.c
522
opte = pmap_pte_set(pte, npte);
sys/arch/amd64/amd64/pmap.c
552
pt_entry_t *pte, opte;
sys/arch/amd64/amd64/pmap.c
558
pte = kvtopte(va);
sys/arch/amd64/amd64/pmap.c
560
opte = pmap_pte_set(pte, 0);
sys/arch/amd64/amd64/pmap.c
593
pt_entry_t *pml4e, *pte;
sys/arch/amd64/amd64/pmap.c
605
pte = (pt_entry_t *)PMAP_DIRECT_MAP(early_pte_pages);
sys/arch/amd64/amd64/pmap.c
606
memset(pte, 0, 3 * NBPG);
sys/arch/amd64/amd64/pmap.c
608
pte[0] = (early_pte_pages + NBPG) | PG_V | PG_RW | pg_crypt;
sys/arch/amd64/amd64/pmap.c
609
pte[1] = (early_pte_pages + 2 * NBPG) | PG_V | PG_RW | pg_crypt;
sys/arch/amd64/amd64/pmap.c
611
pte = (pt_entry_t *)PMAP_DIRECT_MAP(early_pte_pages + NBPG);
sys/arch/amd64/amd64/pmap.c
616
pte[(i * 512) + j] = curpa | PG_V | PG_RW | PG_PS |
sys/arch/amd64/amd64/pmap.c
635
pt_entry_t *pml4e, *pte;
sys/arch/amd64/amd64/pmap.c
637
pte = (pt_entry_t *)PMAP_DIRECT_MAP(early_pte_pages);
sys/arch/amd64/amd64/pmap.c
638
memset(pte, 0, 3 * NBPG);
sys/arch/amd64/amd64/vmm_machdep.c
3510
uint64_t pte, pt_paddr, pte_paddr, mask, low_mask, high_mask;
sys/arch/amd64/amd64/vmm_machdep.c
3589
pte = *hva;
sys/arch/amd64/amd64/vmm_machdep.c
3591
pte = *(uint32_t *)hva;
sys/arch/amd64/amd64/vmm_machdep.c
3594
pte);
sys/arch/amd64/amd64/vmm_machdep.c
3597
if (!(pte & PG_V))
sys/arch/amd64/amd64/vmm_machdep.c
3601
if ((mode == PROT_WRITE) && !(pte & PG_RW))
sys/arch/amd64/amd64/vmm_machdep.c
3604
if ((vcpu->vc_exit.cpl > 0) && !(pte & PG_u))
sys/arch/amd64/amd64/vmm_machdep.c
3607
pte = pte | PG_U;
sys/arch/amd64/amd64/vmm_machdep.c
3609
pte = pte | PG_M;
sys/arch/amd64/amd64/vmm_machdep.c
3610
*hva = pte;
sys/arch/amd64/amd64/vmm_machdep.c
3613
if (pte & PG_PS)
sys/arch/amd64/amd64/vmm_machdep.c
3617
pt_paddr = pte & PG_FRAME;
sys/arch/amd64/amd64/vmm_machdep.c
3625
*pa = (pte & high_mask) | (va & low_mask);
sys/arch/arm/arm/bus_dma.c
905
pt_entry_t pte;
sys/arch/arm/arm/bus_dma.c
940
pte = *ptep;
sys/arch/arm/arm/bus_dma.c
941
KDASSERT((pte & L2_TYPE_MASK) != L2_TYPE_INV);
sys/arch/arm/arm/bus_dma.c
942
if (__predict_false((pte & L2_TYPE_MASK)
sys/arch/arm/arm/bus_dma.c
944
curaddr = (pte & L2_L_FRAME) |
sys/arch/arm/arm/bus_dma.c
946
if (pte & L2_L_COHERENT) {
sys/arch/arm/arm/bus_dma.c
951
curaddr = (pte & L2_S_FRAME) |
sys/arch/arm/arm/bus_dma.c
953
if (pte & L2_S_COHERENT) {
sys/arch/arm/arm/db_interface.c
218
pt_entry_t *pte, oldpte, tmppte;
sys/arch/arm/arm/db_interface.c
233
if (pmap_get_pde_pte(pmap, (vaddr_t) dst, &pde, &pte) == FALSE)
sys/arch/arm/arm/db_interface.c
249
if (pte == NULL)
sys/arch/arm/arm/db_interface.c
251
oldpte = *pte;
sys/arch/arm/arm/db_interface.c
253
*pte = tmppte;
sys/arch/arm/arm/db_interface.c
254
PTE_SYNC(pte);
sys/arch/arm/arm/db_interface.c
287
*pte = oldpte;
sys/arch/arm/arm/db_interface.c
288
PTE_SYNC(pte);
sys/arch/arm/arm/pmap7.c
1322
pt_entry_t pte;
sys/arch/arm/arm/pmap7.c
1325
pte = *ptep;
sys/arch/arm/arm/pmap7.c
1327
if (pte == L2_TYPE_INV) {
sys/arch/arm/arm/pmap7.c
1337
pa = l2pte_pa(pte);
sys/arch/arm/arm/pmap7.c
1362
if (pte & L2_V7_AF)
sys/arch/arm/arm/pmap7.c
1479
pt_entry_t *ptep, pte;
sys/arch/arm/arm/pmap7.c
1508
pte = *ptep;
sys/arch/arm/arm/pmap7.c
1510
if (pte == L2_TYPE_INV)
sys/arch/arm/arm/pmap7.c
1513
switch (pte & L2_TYPE_MASK) {
sys/arch/arm/arm/pmap7.c
1515
pa = (pte & L2_L_FRAME) | (va & L2_L_OFFSET);
sys/arch/arm/arm/pmap7.c
1522
pa = (pte & L2_S_FRAME) | (va & L2_S_OFFSET);
sys/arch/arm/arm/pmap7.c
1697
pt_entry_t *ptep, pte;
sys/arch/arm/arm/pmap7.c
1725
pte = *ptep;
sys/arch/arm/arm/pmap7.c
1726
KASSERT(pte != L2_TYPE_INV);
sys/arch/arm/arm/pmap7.c
1728
pa = l2pte_pa(pte);
sys/arch/arm/arm/pmap7.c
1733
KASSERT((pte & L2_V7_AF) == 0);
sys/arch/arm/arm/pmap7.c
1745
pte |= L2_V7_AF;
sys/arch/arm/arm/pmap7.c
1747
*ptep = pte;
sys/arch/arm/arm/pmap7.c
1778
pt_entry_t *ptep, pte;
sys/arch/arm/arm/pmap7.c
1788
pte = *ptep;
sys/arch/arm/arm/pmap7.c
1791
pa = l2pte_pa(pte);
sys/arch/arm/arm/pmap7.c
2530
pt_entry_t *pte;
sys/arch/arm/arm/pmap7.c
2551
pte = (pt_entry_t *)
sys/arch/arm/arm/pmap7.c
2554
pte = (pt_entry_t *) kernel_pt_lookup(pde[L1_IDX(va)] & L1_C_ADDR_MASK);
sys/arch/arm/arm/pmap7.c
2556
if (pte == NULL)
sys/arch/arm/arm/pmap7.c
2560
pte[(va >> PGSHIFT) & 0x3ff] = L2_S_PROTO | pa | L2_V7_AF |
sys/arch/arm/arm/pmap7.c
2562
PTE_SYNC(&pte[(va >> PGSHIFT) & 0x3ff]);
sys/arch/arm/arm/pmap7.c
2564
pte[l2pte_index(va)] = L2_S_PROTO | pa | L2_V7_AF |
sys/arch/arm/arm/pmap7.c
2566
PTE_SYNC(&pte[l2pte_index(va)]);
sys/arch/arm/arm/pmap7.c
2607
pt_entry_t *pte, f1, f2s, f2l;
sys/arch/arm/arm/pmap7.c
2668
pte = (pt_entry_t *)
sys/arch/arm/arm/pmap7.c
2671
pte = (pt_entry_t *) kernel_pt_lookup(
sys/arch/arm/arm/pmap7.c
2674
if (pte == NULL)
sys/arch/arm/arm/pmap7.c
2685
pte[((va >> PGSHIFT) & 0x3f0) + i] =
sys/arch/arm/arm/pmap7.c
2688
PTE_SYNC(&pte[((va >> PGSHIFT) & 0x3f0) + i]);
sys/arch/arm/arm/pmap7.c
2690
pte[l2pte_index(va) + i] =
sys/arch/arm/arm/pmap7.c
2693
PTE_SYNC(&pte[l2pte_index(va) + i]);
sys/arch/arm/arm/pmap7.c
2707
pte[(va >> PGSHIFT) & 0x3ff] = L2_S_PROTO | pa | L2_V7_AF |
sys/arch/arm/arm/pmap7.c
2709
PTE_SYNC(&pte[(va >> PGSHIFT) & 0x3ff]);
sys/arch/arm/arm/pmap7.c
2711
pte[l2pte_index(va)] = L2_S_PROTO | pa | L2_V7_AF |
sys/arch/arm/arm/pmap7.c
2713
PTE_SYNC(&pte[l2pte_index(va)]);
sys/arch/arm/include/pmap.h
293
#define PTE_SYNC(pte) \
sys/arch/arm/include/pmap.h
298
cpu_dcache_wb_range((vaddr_t)(pte), sizeof(pt_entry_t));\
sys/arch/arm/include/pmap.h
300
(void)pmap_extract(pmap_kernel(), (vaddr_t)(pte), &pa); \
sys/arch/arm/include/pmap.h
301
cpu_sdcache_wb_range((vaddr_t)(pte), (paddr_t)(pa), \
sys/arch/arm/include/pmap.h
308
#define PTE_SYNC_RANGE(pte, cnt) \
sys/arch/arm/include/pmap.h
313
cpu_dcache_wb_range((vaddr_t)(pte), \
sys/arch/arm/include/pmap.h
316
(void)pmap_extract(pmap_kernel(), (vaddr_t)(pte), &pa);\
sys/arch/arm/include/pmap.h
317
cpu_sdcache_wb_range((vaddr_t)(pte), (paddr_t)(pa), \
sys/arch/arm/include/pmap.h
330
#define l2pte_valid(pte) (((pte) & L2_TYPE_MASK) != L2_TYPE_INV)
sys/arch/arm/include/pmap.h
331
#define l2pte_pa(pte) ((pte) & L2_S_FRAME)
sys/arch/arm/include/pmap.h
435
pt_entry_t pte;
sys/arch/arm/include/pmap.h
438
pte = (pr & PROT_WRITE) ? L1_S_PROT_UW : L1_S_PROT_UR;
sys/arch/arm/include/pmap.h
440
pte = (pr & PROT_WRITE) ? L1_S_PROT_KW : L1_S_PROT_KR;
sys/arch/arm/include/pmap.h
443
pte |= L1_S_V7_XN;
sys/arch/arm/include/pmap.h
445
return pte;
sys/arch/arm/include/pmap.h
450
pt_entry_t pte;
sys/arch/arm/include/pmap.h
453
pte = (pr & PROT_WRITE) ? L2_L_PROT_UW : L2_L_PROT_UR;
sys/arch/arm/include/pmap.h
455
pte = (pr & PROT_WRITE) ? L2_L_PROT_KW : L2_L_PROT_KR;
sys/arch/arm/include/pmap.h
458
pte |= L2_V7_L_XN;
sys/arch/arm/include/pmap.h
460
return pte;
sys/arch/arm/include/pmap.h
465
pt_entry_t pte;
sys/arch/arm/include/pmap.h
468
pte = (pr & PROT_WRITE) ? L2_S_PROT_UW : L2_S_PROT_UR;
sys/arch/arm/include/pmap.h
470
pte = (pr & PROT_WRITE) ? L2_S_PROT_KW : L2_S_PROT_KR;
sys/arch/arm/include/pmap.h
473
pte |= L2_V7_S_XN;
sys/arch/arm/include/pmap.h
475
return pte;
sys/arch/arm/include/pmap.h
479
l2pte_is_writeable(pt_entry_t pte, struct pmap *pm)
sys/arch/arm/include/pmap.h
481
return (pte & L2_V7_AP(0x4)) == 0;
sys/arch/arm64/arm64/pmap.c
1781
uint64_t pte, access_bits;
sys/arch/arm64/arm64/pmap.c
1822
pte = (pted->pted_pte & PTE_RPGN) | attr | access_bits | L3_P;
sys/arch/arm64/arm64/pmap.c
1823
*pl3 = access_bits ? pte : 0;
sys/arch/arm64/dev/smmu.c
1114
uint64_t pte, access_bits;
sys/arch/arm64/dev/smmu.c
1174
pte = (pted & PTE_RPGN) | attr | access_bits | L3_P;
sys/arch/arm64/dev/smmu.c
1175
*pl3 = pte;
sys/arch/hppa/hppa/pmap.c
1000
pt_entry_t pte;
sys/arch/hppa/hppa/pmap.c
1009
pte = pmap_pte_get(pde, va);
sys/arch/hppa/hppa/pmap.c
1010
attrs = pmap_pvh_attrs(pte);
sys/arch/hppa/hppa/pmap.c
1012
pmap_pte_flush(pmap, va, pte);
sys/arch/hppa/hppa/pmap.c
1013
if (pte & PTE_PROT(TLB_WIRED))
sys/arch/hppa/hppa/pmap.c
1034
pt_entry_t pte = 0;
sys/arch/hppa/hppa/pmap.c
1040
pte = pmap_pte_get(pde, va);
sys/arch/hppa/hppa/pmap.c
1042
if (pte & PTE_PROT(TLB_WIRED)) {
sys/arch/hppa/hppa/pmap.c
1043
pte &= ~PTE_PROT(TLB_WIRED);
sys/arch/hppa/hppa/pmap.c
1045
pmap_pte_set(pde, va, pte);
sys/arch/hppa/hppa/pmap.c
105
pmap_pvh_attrs(pt_entry_t pte)
sys/arch/hppa/hppa/pmap.c
1053
if (!pte)
sys/arch/hppa/hppa/pmap.c
1075
pt_entry_t opte, pte;
sys/arch/hppa/hppa/pmap.c
1078
opte = pte = pmap_pte_get(pde, va);
sys/arch/hppa/hppa/pmap.c
108
if (pte & PTE_PROT(TLB_DIRTY))
sys/arch/hppa/hppa/pmap.c
1080
if (!pte) {
sys/arch/hppa/hppa/pmap.c
1086
pte &= ~clear;
sys/arch/hppa/hppa/pmap.c
1087
pte |= set;
sys/arch/hppa/hppa/pmap.c
1088
attrs |= pmap_pvh_attrs(pte);
sys/arch/hppa/hppa/pmap.c
1091
if (opte != pte) {
sys/arch/hppa/hppa/pmap.c
1093
pmap_pte_set(pde, va, pte);
sys/arch/hppa/hppa/pmap.c
110
if ((pte & PTE_PROT(TLB_REFTRAP)) == 0)
sys/arch/hppa/hppa/pmap.c
1111
pt_entry_t pte;
sys/arch/hppa/hppa/pmap.c
1119
pte = pmap_vp_find(pve->pv_pmap, pve->pv_va);
sys/arch/hppa/hppa/pmap.c
1120
atomic_setbits_int(&pg->pg_flags, pmap_pvh_attrs(pte));
sys/arch/hppa/hppa/pmap.c
1131
pt_entry_t pte;
sys/arch/hppa/hppa/pmap.c
1136
pte = pmap_vp_find(pmap, va);
sys/arch/hppa/hppa/pmap.c
1139
if (pte) {
sys/arch/hppa/hppa/pmap.c
1141
*pap = (pte & ~PGOFSET) | (va & PGOFSET);
sys/arch/hppa/hppa/pmap.c
116
void pmap_pte_flush(struct pmap *pmap, vaddr_t va, pt_entry_t pte);
sys/arch/hppa/hppa/pmap.c
1215
pt_entry_t pte, opte;
sys/arch/hppa/hppa/pmap.c
122
pt_entry_t pte);
sys/arch/hppa/hppa/pmap.c
1224
pte = pa | PTE_PROT(TLB_WIRED | TLB_REFTRAP |
sys/arch/hppa/hppa/pmap.c
1227
pte |= PTE_PROT(TLB_UNCACHABLE);
sys/arch/hppa/hppa/pmap.c
1230
pmap_pte_set(pde, va, pte);
sys/arch/hppa/hppa/pmap.c
1238
if (pmap_initialized && (pg = PHYS_TO_VM_PAGE(PTE_PAGE(pte)))) {
sys/arch/hppa/hppa/pmap.c
1239
if (pmap_check_alias(pg, va, pte))
sys/arch/hppa/hppa/pmap.c
1253
pt_entry_t pte;
sys/arch/hppa/hppa/pmap.c
1273
if (!(pte = pmap_pte_get(pde, va))) {
sys/arch/hppa/hppa/pmap.c
1280
pmap_pte_flush(pmap_kernel(), va, pte);
sys/arch/hppa/hppa/pmap.c
1282
if (pmap_initialized && (pg = PHYS_TO_VM_PAGE(PTE_PAGE(pte)))) {
sys/arch/hppa/hppa/pmap.c
1283
atomic_setbits_int(&pg->pg_flags, pmap_pvh_attrs(pte));
sys/arch/hppa/hppa/pmap.c
301
pmap_pte_set(volatile pt_entry_t *pde, vaddr_t va, pt_entry_t pte)
sys/arch/hppa/hppa/pmap.c
304
pde, va, pte));
sys/arch/hppa/hppa/pmap.c
314
pde[(va >> 12) & 0x3ff] = pte;
sys/arch/hppa/hppa/pmap.c
318
pmap_pte_flush(struct pmap *pmap, vaddr_t va, pt_entry_t pte)
sys/arch/hppa/hppa/pmap.c
321
if (pte & PTE_PROT(TLB_EXECUTE)) {
sys/arch/hppa/hppa/pmap.c
356
pt_entry_t pte;
sys/arch/hppa/hppa/pmap.c
376
if (!(pte = pmap_pte_get(pde, va)))
sys/arch/hppa/hppa/pmap.c
379
printf("0x%08lx-0x%08x:%b\n", va, pte & ~PAGE_MASK,
sys/arch/hppa/hppa/pmap.c
380
TLB_PROT(pte & PAGE_MASK), TLB_BITS);
sys/arch/hppa/hppa/pmap.c
400
pmap_check_alias(struct vm_page *pg, vaddr_t va, pt_entry_t pte)
sys/arch/hppa/hppa/pmap.c
408
pte |= pmap_vp_find(pve->pv_pmap, pve->pv_va);
sys/arch/hppa/hppa/pmap.c
410
(pte & PTE_PROT(TLB_GATEWAY)) == 0 &&
sys/arch/hppa/hppa/pmap.c
411
(pte & PTE_PROT(TLB_WRITE))) {
sys/arch/hppa/hppa/pmap.c
750
pt_entry_t pte;
sys/arch/hppa/hppa/pmap.c
772
if ((pte = pmap_pte_get(pde, va))) {
sys/arch/hppa/hppa/pmap.c
774
("pmap_enter: remapping 0x%x -> 0x%lx\n", pte, pa));
sys/arch/hppa/hppa/pmap.c
776
pmap_pte_flush(pmap, va, pte);
sys/arch/hppa/hppa/pmap.c
777
if (wired && !(pte & PTE_PROT(TLB_WIRED)))
sys/arch/hppa/hppa/pmap.c
779
else if (!wired && (pte & PTE_PROT(TLB_WIRED)))
sys/arch/hppa/hppa/pmap.c
782
if (PTE_PAGE(pte) == pa) {
sys/arch/hppa/hppa/pmap.c
788
pg = PHYS_TO_VM_PAGE(PTE_PAGE(pte));
sys/arch/hppa/hppa/pmap.c
791
atomic_setbits_int(&pg->pg_flags, pmap_pvh_attrs(pte));
sys/arch/hppa/hppa/pmap.c
796
pte = PTE_PROT(TLB_REFTRAP);
sys/arch/hppa/hppa/pmap.c
812
pte |= PTE_PROT(pmap_prot(pmap, prot));
sys/arch/hppa/hppa/pmap.c
813
if (pmap_check_alias(pg, va, pte))
sys/arch/hppa/hppa/pmap.c
821
pte = pa | PTE_PROT(pmap_prot(pmap, prot)) |
sys/arch/hppa/hppa/pmap.c
822
(pte & PTE_PROT(TLB_UNCACHABLE|TLB_DIRTY|TLB_REFTRAP));
sys/arch/hppa/hppa/pmap.c
824
pte |= PTE_PROT(TLB_UNCACHABLE);
sys/arch/hppa/hppa/pmap.c
826
pte |= PTE_PROT(TLB_WIRED);
sys/arch/hppa/hppa/pmap.c
827
pmap_pte_set(pde, va, pte);
sys/arch/hppa/hppa/pmap.c
840
pt_entry_t pte;
sys/arch/hppa/hppa/pmap.c
862
if ((pte = pmap_pte_get(pde, sva))) {
sys/arch/hppa/hppa/pmap.c
868
pmap_pte_flush(pmap, sva, pte);
sys/arch/hppa/hppa/pmap.c
869
if (pte & PTE_PROT(TLB_WIRED))
sys/arch/hppa/hppa/pmap.c
878
(pg = PHYS_TO_VM_PAGE(PTE_PAGE(pte)))) {
sys/arch/hppa/hppa/pmap.c
880
pmap_pvh_attrs(pte));
sys/arch/hppa/hppa/pmap.c
884
if (IS_IOPAGE(PTE_PAGE(pte))) {
sys/arch/hppa/hppa/pmap.c
911
pt_entry_t opte, pte;
sys/arch/hppa/hppa/pmap.c
914
opte = pte = pmap_pte_get(pde, va);
sys/arch/hppa/hppa/pmap.c
915
if (pte & TLB_GATEWAY)
sys/arch/hppa/hppa/pmap.c
917
pte &= ~TLB_WRITE;
sys/arch/hppa/hppa/pmap.c
918
attrs |= pmap_pvh_attrs(pte);
sys/arch/hppa/hppa/pmap.c
920
if (opte != pte) {
sys/arch/hppa/hppa/pmap.c
922
pmap_pte_set(pde, va, pte);
sys/arch/hppa/hppa/pmap.c
939
pt_entry_t pte;
sys/arch/hppa/hppa/pmap.c
957
if ((pte = pmap_pte_get(pde, sva))) {
sys/arch/hppa/hppa/pmap.c
961
sva, pte));
sys/arch/hppa/hppa/pmap.c
966
if ((pte & PTE_PROT(TLB_AR_MASK)) == tlbprot)
sys/arch/hppa/hppa/pmap.c
969
pg = PHYS_TO_VM_PAGE(PTE_PAGE(pte));
sys/arch/hppa/hppa/pmap.c
972
pmap_pvh_attrs(pte));
sys/arch/hppa/hppa/pmap.c
975
pmap_pte_flush(pmap, sva, pte);
sys/arch/hppa/hppa/pmap.c
976
pte &= ~PTE_PROT(TLB_AR_MASK);
sys/arch/hppa/hppa/pmap.c
977
pte |= tlbprot;
sys/arch/hppa/hppa/pmap.c
978
pmap_pte_set(pde, sva, pte);
sys/arch/hppa/include/pte.h
32
#define TLB_PROT(pte) ((pte) << PTE_PROT_SHIFT)
sys/arch/hppa/include/pte.h
36
#define PTE_PAGE(pte) ((pte) & ~PAGE_MASK)
sys/arch/i386/i386/hibernate_machdep.c
250
pt_entry_t *pte, npte;
sys/arch/i386/i386/hibernate_machdep.c
252
pte = s4pte_4k(va);
sys/arch/i386/i386/hibernate_machdep.c
254
*pte = npte;
sys/arch/i386/i386/pmap.c
1019
pte = PTE_BASE + atop(virtual_avail);
sys/arch/i386/i386/pmap.c
1030
pmap_csrcp = (caddr_t) virtual_avail; csrc_pte = pte;
sys/arch/i386/i386/pmap.c
1032
pmap_cdstp = (caddr_t) virtual_avail+PAGE_SIZE; cdst_pte = pte+1;
sys/arch/i386/i386/pmap.c
1034
pmap_zerop = (caddr_t) virtual_avail+PAGE_SIZE*2; zero_pte = pte+2;
sys/arch/i386/i386/pmap.c
1036
pmap_ptpp = (caddr_t) virtual_avail+PAGE_SIZE*3; ptp_pte = pte+3;
sys/arch/i386/i386/pmap.c
1038
pmap_flshp = (caddr_t) virtual_avail+PAGE_SIZE*4; flsh_pte = pte+4;
sys/arch/i386/i386/pmap.c
1041
pte += MAXCPUS * NPTECL;
sys/arch/i386/i386/pmap.c
1043
pmap_csrcp = (caddr_t) virtual_avail; csrc_pte = pte; /* allocate */
sys/arch/i386/i386/pmap.c
1044
virtual_avail += PAGE_SIZE; pte++; /* advance */
sys/arch/i386/i386/pmap.c
1046
pmap_cdstp = (caddr_t) virtual_avail; cdst_pte = pte;
sys/arch/i386/i386/pmap.c
1047
virtual_avail += PAGE_SIZE; pte++;
sys/arch/i386/i386/pmap.c
1049
pmap_zerop = (caddr_t) virtual_avail; zero_pte = pte;
sys/arch/i386/i386/pmap.c
1050
virtual_avail += PAGE_SIZE; pte++;
sys/arch/i386/i386/pmap.c
1052
pmap_ptpp = (caddr_t) virtual_avail; ptp_pte = pte;
sys/arch/i386/i386/pmap.c
1053
virtual_avail += PAGE_SIZE; pte++;
sys/arch/i386/i386/pmap.c
1055
pmap_flshp = (caddr_t) virtual_avail; flsh_pte = pte;
sys/arch/i386/i386/pmap.c
1056
virtual_avail += PAGE_SIZE; pte++;
sys/arch/i386/i386/pmap.c
1064
virtual_avail += round_page(MSGBUFSIZE); pte++;
sys/arch/i386/i386/pmap.c
1067
virtual_avail += round_page(bootargc); pte++;
sys/arch/i386/i386/pmap.c
1104
pt_entry_t *pte, npte;
sys/arch/i386/i386/pmap.c
1114
pte = vtopte(ptpva);
sys/arch/i386/i386/pmap.c
1117
i386_atomic_testset_ul(pte, npte);
sys/arch/i386/i386/pmap.c
1549
pt_entry_t *ptes, pte;
sys/arch/i386/i386/pmap.c
1553
pte = ptes[atop(va)];
sys/arch/i386/i386/pmap.c
1555
if (!pmap_valid_entry(pte))
sys/arch/i386/i386/pmap.c
1558
*pap = (pte & PG_FRAME) | (va & ~PG_FRAME);
sys/arch/i386/i386/pmap.c
1637
pt_entry_t *pte;
sys/arch/i386/i386/pmap.c
1647
pte = PTESLEW(flsh_pte, id);
sys/arch/i386/i386/pmap.c
1651
if (*pte)
sys/arch/i386/i386/pmap.c
1655
*pte = (pa & PG_FRAME) | PG_V | PG_RW;
sys/arch/i386/i386/pmap.c
1658
*pte = 0;
sys/arch/i386/i386/pmap.c
1711
pt_entry_t *pte = (pt_entry_t *) ptpva;
sys/arch/i386/i386/pmap.c
1725
; pte++, startva += NBPG) {
sys/arch/i386/i386/pmap.c
1726
if (!pmap_valid_entry(*pte))
sys/arch/i386/i386/pmap.c
1729
if ((flags & PMAP_REMOVE_SKIPWIRED) && (*pte & PG_W))
sys/arch/i386/i386/pmap.c
1733
opte = i386_atomic_testset_ul(pte, 0);
sys/arch/i386/i386/pmap.c
2008
pt_entry_t *ptes, pte;
sys/arch/i386/i386/pmap.c
2023
pte = ptes[ptei(pve->pv_va)];
sys/arch/i386/i386/pmap.c
2025
mybits |= (pte & testbits);
sys/arch/i386/i386/pmap.c
408
#define PTESLEW(pte, id) ((pte)+(id)*NPTECL)
sys/arch/i386/i386/pmap.c
411
#define PTESLEW(pte, id) (pte)
sys/arch/i386/i386/pmap.c
484
pmap_pte2flags(pt_entry_t pte)
sys/arch/i386/i386/pmap.c
486
return (((pte & PG_U) ? PG_PMAP_REF : 0) |
sys/arch/i386/i386/pmap.c
487
((pte & PG_M) ? PG_PMAP_MOD : 0));
sys/arch/i386/i386/pmap.c
491
pmap_sync_flags_pte_86(struct vm_page *pg, pt_entry_t pte)
sys/arch/i386/i386/pmap.c
493
if (pte & (PG_U|PG_M)) {
sys/arch/i386/i386/pmap.c
494
atomic_setbits_int(&pg->pg_flags, pmap_pte2flags(pte));
sys/arch/i386/i386/pmap.c
666
pt_entry_t pte, *ptep = vtopte(va);
sys/arch/i386/i386/pmap.c
670
pte = i386_atomic_testset_ul(ptep, pa | bits); /* zap! */
sys/arch/i386/i386/pmap.c
671
return (pte & ~PG_FRAME);
sys/arch/i386/i386/pmap.c
678
pt_entry_t pte = *ptep;
sys/arch/i386/i386/pmap.c
680
*ptep = (pte | set) & ~clr;
sys/arch/i386/i386/pmap.c
681
return (pte & ~PG_FRAME);
sys/arch/i386/i386/pmap.c
929
pt_entry_t *pte;
sys/arch/i386/i386/pmapae.c
1054
pt_entry_t *ptes, pte;
sys/arch/i386/i386/pmapae.c
1058
pte = ptes[atop(va)];
sys/arch/i386/i386/pmapae.c
1060
if (!pmap_valid_entry(pte))
sys/arch/i386/i386/pmapae.c
1063
*pap = (pte & PG_FRAME) | (va & ~PG_FRAME);
sys/arch/i386/i386/pmapae.c
1145
pt_entry_t *pte = (pt_entry_t *) ptpva;
sys/arch/i386/i386/pmapae.c
1159
; pte++, startva += NBPG) {
sys/arch/i386/i386/pmapae.c
1160
if (!pmap_valid_entry(*pte))
sys/arch/i386/i386/pmapae.c
1163
if ((flags & PMAP_REMOVE_SKIPWIRED) && (*pte & PG_W))
sys/arch/i386/i386/pmapae.c
1167
opte = i386_atomic_testset_uq(pte, 0);
sys/arch/i386/i386/pmapae.c
1439
pt_entry_t *ptes, pte;
sys/arch/i386/i386/pmapae.c
1454
pte = ptes[ptei(pve->pv_va)];
sys/arch/i386/i386/pmapae.c
1456
mybits |= (pte & testbits);
sys/arch/i386/i386/pmapae.c
2050
pt_entry_t *pte, npte;
sys/arch/i386/i386/pmapae.c
2054
pte = vtopte(ptpva);
sys/arch/i386/i386/pmapae.c
2057
i386_atomic_testset_uq(pte, npte);
sys/arch/i386/i386/pmapae.c
2120
pt_entry_t *pte = PTESLEW(flsh_pte, id);
sys/arch/i386/i386/pmapae.c
2125
if (*pte)
sys/arch/i386/i386/pmapae.c
2129
*pte = (pa & PG_FRAME) | PG_V | PG_RW;
sys/arch/i386/i386/pmapae.c
2132
*pte = 0;
sys/arch/i386/i386/pmapae.c
448
#define PTESLEW(pte, id) ((pte)+(id)*NPTECL)
sys/arch/i386/i386/pmapae.c
451
#define PTESLEW(pte, id) (pte)
sys/arch/i386/i386/pmapae.c
484
pmap_pte2flags(pt_entry_t pte)
sys/arch/i386/i386/pmapae.c
486
return (((pte & PG_U) ? PG_PMAP_REF : 0) |
sys/arch/i386/i386/pmapae.c
487
((pte & PG_M) ? PG_PMAP_MOD : 0));
sys/arch/i386/i386/pmapae.c
491
pmap_sync_flags_pte_pae(struct vm_page *pg, pt_entry_t pte)
sys/arch/i386/i386/pmapae.c
493
if (pte & (PG_U|PG_M)) {
sys/arch/i386/i386/pmapae.c
494
atomic_setbits_int(&pg->pg_flags, pmap_pte2flags(pte));
sys/arch/i386/i386/pmapae.c
568
pt_entry_t pte, *ptep = vtopte(va);
sys/arch/i386/i386/pmapae.c
578
pte = i386_atomic_testset_uq(ptep, pa | bits | nx); /* zap! */
sys/arch/i386/i386/pmapae.c
579
return (pte & ~PG_FRAME);
sys/arch/i386/i386/pmapae.c
586
pt_entry_t pte = *ptep;
sys/arch/i386/i386/pmapae.c
588
i386_atomic_testset_uq(ptep, (pte | set) & ~(pt_entry_t)clr);
sys/arch/i386/i386/pmapae.c
589
return (pte & ~PG_FRAME);
sys/arch/i386/i386/pmapae.c
652
pt_entry_t pte;
sys/arch/i386/i386/pmapae.c
822
pte = PDE(kpm, pdei(va));
sys/arch/i386/i386/pmapae.c
823
PDE(kpm, pdei(va)) = pte | PG_NX;
sys/arch/i386/i386/pmapae.c
827
pte = PDE(kpm, pdei(va));
sys/arch/i386/i386/pmapae.c
828
PDE(kpm, pdei(va)) = pte | PG_NX;
sys/arch/landisk/dev/obio.c
233
pt_entry_t *pte;
sys/arch/landisk/dev/obio.c
273
pte = __pmap_kpte_lookup(va);
sys/arch/landisk/dev/obio.c
274
KDASSERT(pte);
sys/arch/landisk/dev/obio.c
275
*pte |= m; /* PTEA PCMCIA assistant bit */
sys/arch/landisk/dev/obio.c
276
sh_tlb_update(0, va, *pte);
sys/arch/m88k/include/cmmu.h
95
#define cmmu_tlbis(cpu, va, pte) (cmmu->tlb_inv_s)(cpu, va, pte)
sys/arch/m88k/include/cmmu.h
96
#define cmmu_tlbiu(cpu, va, pte) (cmmu->tlb_inv_u)(cpu, va, pte)
sys/arch/m88k/include/mmu.h
123
#define PDT_VALID(pte) (*(pte) & PG_V)
sys/arch/m88k/m88k/db_interface.c
495
pt_entry_t *pte, opte, npte;
sys/arch/m88k/m88k/db_interface.c
503
pte = NULL;
sys/arch/m88k/m88k/db_interface.c
506
pte = pmap_pte(pmap_kernel(), va);
sys/arch/m88k/m88k/db_interface.c
507
if (pte != NULL) {
sys/arch/m88k/m88k/db_interface.c
508
opte = *pte;
sys/arch/m88k/m88k/db_interface.c
516
if (pte != NULL && (opte & PG_RO)) {
sys/arch/m88k/m88k/db_interface.c
518
*pte = npte;
sys/arch/m88k/m88k/db_interface.c
523
if (pte != NULL && (opte & PG_RO)) {
sys/arch/m88k/m88k/db_interface.c
524
*pte = opte;
sys/arch/m88k/m88k/db_interface.c
527
if (pte != NULL && (opte & (CACHE_INH | CACHE_WT)) == 0) {
sys/arch/m88k/m88k/m8820x_machdep.c
773
m8820x_tlbis(cpuid_t cpu, vaddr_t va, pt_entry_t pte)
sys/arch/m88k/m88k/m8820x_machdep.c
791
m8820x_tlbiu(cpuid_t cpu, vaddr_t va, pt_entry_t pte)
sys/arch/m88k/m88k/pmap.c
1001
if (PDT_VALID(pte))
sys/arch/m88k/m88k/pmap.c
1002
pmap_remove_pte(pmap, va, pte, NULL, FALSE);
sys/arch/m88k/m88k/pmap.c
1077
npte |= invalidate_pte(pte) & PG_M_U;
sys/arch/m88k/m88k/pmap.c
1079
*pte = npte;
sys/arch/m88k/m88k/pmap.c
1108
pt_entry_t *pte, npte;
sys/arch/m88k/m88k/pmap.c
1121
if ((pte = pmap_pte(pmap_kernel(), va)) == NULL)
sys/arch/m88k/m88k/pmap.c
1122
pte = pmap_expand_kmap(va, 0);
sys/arch/m88k/m88k/pmap.c
1130
invalidate_pte(pte);
sys/arch/m88k/m88k/pmap.c
1132
*pte = npte;
sys/arch/m88k/m88k/pmap.c
1143
pmap_remove_pte(pmap_t pmap, vaddr_t va, pt_entry_t *pte, struct vm_page *pg,
sys/arch/m88k/m88k/pmap.c
1157
if (pmap_pte_w(pte))
sys/arch/m88k/m88k/pmap.c
1160
pa = ptoa(PG_PFNUM(*pte));
sys/arch/m88k/m88k/pmap.c
1166
opte = invalidate_pte(pte) & PG_M_U;
sys/arch/m88k/m88k/pmap.c
1186
pmap, (void *)va, pte, (void *)pa, pg, flush);
sys/arch/m88k/m88k/pmap.c
1198
pmap, (void *)va, pte, pg, flush, (void *)pa, head);
sys/arch/m88k/m88k/pmap.c
1240
pt_entry_t *pte;
sys/arch/m88k/m88k/pmap.c
1260
pte = sdt_pte(sdt, va);
sys/arch/m88k/m88k/pmap.c
1262
if (PDT_VALID(pte))
sys/arch/m88k/m88k/pmap.c
1263
pmap_remove_pte(pmap, va, pte, NULL,
sys/arch/m88k/m88k/pmap.c
1266
pte++;
sys/arch/m88k/m88k/pmap.c
1302
pt_entry_t *pte, opte;
sys/arch/m88k/m88k/pmap.c
1314
pte = sdt_pte(sdt, va);
sys/arch/m88k/m88k/pmap.c
1316
if (PDT_VALID(pte)) {
sys/arch/m88k/m88k/pmap.c
1321
opte = invalidate_pte(pte);
sys/arch/m88k/m88k/pmap.c
1334
pte++;
sys/arch/m88k/m88k/pmap.c
1347
pt_entry_t *pte;
sys/arch/m88k/m88k/pmap.c
1362
pte = pmap_pte(pmap, va);
sys/arch/m88k/m88k/pmap.c
1364
if (pte == NULL || !PDT_VALID(pte)) {
sys/arch/m88k/m88k/pmap.c
1369
pmap_remove_pte(pmap, va, pte, pg, TRUE);
sys/arch/m88k/m88k/pmap.c
1386
pt_entry_t *pte, ap, opte, npte;
sys/arch/m88k/m88k/pmap.c
1413
pte = sdt_pte(sdt, va);
sys/arch/m88k/m88k/pmap.c
1415
if (PDT_VALID(pte)) {
sys/arch/m88k/m88k/pmap.c
1422
opte = invalidate_pte(pte);
sys/arch/m88k/m88k/pmap.c
1424
*pte = npte;
sys/arch/m88k/m88k/pmap.c
1428
pte++;
sys/arch/m88k/m88k/pmap.c
1442
pt_entry_t *pte;
sys/arch/m88k/m88k/pmap.c
1444
pte = pmap_pte(pmap, va);
sys/arch/m88k/m88k/pmap.c
1445
if (pmap_pte_w(pte)) {
sys/arch/m88k/m88k/pmap.c
1447
*pte &= ~PG_W;
sys/arch/m88k/m88k/pmap.c
1497
pt_entry_t *pte, npte, opte;
sys/arch/m88k/m88k/pmap.c
1512
pte = pmap_pte(pmap, va);
sys/arch/m88k/m88k/pmap.c
1517
if (pte == NULL || !PDT_VALID(pte))
sys/arch/m88k/m88k/pmap.c
1520
if (ptoa(PG_PFNUM(*pte)) != VM_PAGE_TO_PHYS(pg))
sys/arch/m88k/m88k/pmap.c
1522
*pte, pmap, pg, VM_PAGE_TO_PHYS(pg));
sys/arch/m88k/m88k/pmap.c
1528
opte = *pte;
sys/arch/m88k/m88k/pmap.c
1537
invalidate_pte(pte);
sys/arch/m88k/m88k/pmap.c
1538
*pte = npte;
sys/arch/m88k/m88k/pmap.c
1558
pt_entry_t *pte;
sys/arch/m88k/m88k/pmap.c
1580
pte = pmap_pte(pmap, pvep->pv_va);
sys/arch/m88k/m88k/pmap.c
1581
if (pte == NULL || !PDT_VALID(pte))
sys/arch/m88k/m88k/pmap.c
1585
if (ptoa(PG_PFNUM(*pte)) != VM_PAGE_TO_PHYS(pg))
sys/arch/m88k/m88k/pmap.c
1587
*pte, pmap, pg, VM_PAGE_TO_PHYS(pg));
sys/arch/m88k/m88k/pmap.c
1590
if ((*pte & bit) != 0) {
sys/arch/m88k/m88k/pmap.c
1616
pt_entry_t *pte, opte, npte;
sys/arch/m88k/m88k/pmap.c
1639
pte = pmap_pte(pmap, va);
sys/arch/m88k/m88k/pmap.c
1644
if (pte == NULL || !PDT_VALID(pte))
sys/arch/m88k/m88k/pmap.c
1647
if (ptoa(PG_PFNUM(*pte)) != VM_PAGE_TO_PHYS(pg))
sys/arch/m88k/m88k/pmap.c
1649
*pte, pmap, pg, VM_PAGE_TO_PHYS(pg));
sys/arch/m88k/m88k/pmap.c
1655
opte = *pte;
sys/arch/m88k/m88k/pmap.c
1662
invalidate_pte(pte);
sys/arch/m88k/m88k/pmap.c
1664
*pte = npte;
sys/arch/m88k/m88k/pmap.c
176
invalidate_pte(pt_entry_t *pte)
sys/arch/m88k/m88k/pmap.c
1806
pt_entry_t *pte;
sys/arch/m88k/m88k/pmap.c
1810
pte = pmap_pte(pmap, va);
sys/arch/m88k/m88k/pmap.c
1812
if (pte == NULL)
sys/arch/m88k/m88k/pmap.c
1817
if ((*pte & (PG_M | PG_RO)) != 0)
sys/arch/m88k/m88k/pmap.c
182
("xmem %0, %2, %%r0" : "+r"(oldpte), "+m"(*pte) : "r"(pte));
sys/arch/m88k/m88k/pmap.c
1821
*pte |= PG_M;
sys/arch/m88k/m88k/pmap.c
1822
pa = *pte & PG_FRAME;
sys/arch/m88k/m88k/pmap.c
1831
tlb_flush(pmap, va, *pte);
sys/arch/m88k/m88k/pmap.c
1846
pt_entry_t *pte, opte, npte;
sys/arch/m88k/m88k/pmap.c
1856
if ((pte = pmap_pte(pmap_kernel(), va)) == NULL)
sys/arch/m88k/m88k/pmap.c
1858
DPRINTF(CD_CACHE, ("cache_ctrl: pte@%p\n", pte));
sys/arch/m88k/m88k/pmap.c
1865
if (((*pte & CACHE_INH) == 0 && (mode & CACHE_INH) != 0) ||
sys/arch/m88k/m88k/pmap.c
1866
((*pte & CACHE_WT) == 0 && (mode & CACHE_WT) != 0)) {
sys/arch/m88k/m88k/pmap.c
1867
pa = ptoa(PG_PFNUM(*pte));
sys/arch/m88k/m88k/pmap.c
1891
opte = invalidate_pte(pte);
sys/arch/m88k/m88k/pmap.c
1893
*pte = npte;
sys/arch/m88k/m88k/pmap.c
191
#define pmap_pte_w(pte) (*(pte) & PG_W)
sys/arch/m88k/m88k/pmap.c
1910
pt_entry_t *pte, opte, npte;
sys/arch/m88k/m88k/pmap.c
1920
pte = pmap_pte(pmap, va);
sys/arch/m88k/m88k/pmap.c
1922
if (pte == NULL || !PDT_VALID(pte))
sys/arch/m88k/m88k/pmap.c
1924
opte = *pte;
sys/arch/m88k/m88k/pmap.c
1937
invalidate_pte(pte);
sys/arch/m88k/m88k/pmap.c
1939
*pte = npte;
sys/arch/m88k/m88k/pmap.c
259
pt_entry_t *pte;
sys/arch/m88k/m88k/pmap.c
351
pte = pmap_pte(pmap, va);
sys/arch/m88k/m88k/pmap.c
352
if (pte != NULL && PDT_VALID(pte)) {
sys/arch/m88k/m88k/pmap.c
353
*pap = ptoa(PG_PFNUM(*pte)) | (va & PAGE_MASK);
sys/arch/m88k/m88k/pmap.c
354
*ti = (*pte | pmap->pm_apr) & CACHE_MASK;
sys/arch/m88k/m88k/pmap.c
377
tlb_flush(pmap_t pmap, vaddr_t va, pt_entry_t pte)
sys/arch/m88k/m88k/pmap.c
392
cmmu_tlbis(ci->ci_cpuid, va, pte);
sys/arch/m88k/m88k/pmap.c
394
cmmu_tlbiu(ci->ci_cpuid, va, pte);
sys/arch/m88k/m88k/pmap.c
404
tlb_kflush(vaddr_t va, pt_entry_t pte)
sys/arch/m88k/m88k/pmap.c
413
cmmu_tlbis(ci->ci_cpuid, va, pte);
sys/arch/m88k/m88k/pmap.c
417
cmmu_tlbis(ci->ci_cpuid, va, pte);
sys/arch/m88k/m88k/pmap.c
588
pt_entry_t *pte, npte;
sys/arch/m88k/m88k/pmap.c
612
if ((pte = pmap_pte(pmap_kernel(), curpa)) == NULL)
sys/arch/m88k/m88k/pmap.c
613
pte = pmap_expand_kmap(curpa, 0);
sys/arch/m88k/m88k/pmap.c
615
*pte = npte | curpa;
sys/arch/m88k/m88k/pmap.c
674
pt_entry_t *pte, template;
sys/arch/m88k/m88k/pmap.c
734
pte = (pt_entry_t *)ptepa;
sys/arch/m88k/m88k/pmap.c
748
DPRINTF(CD_BOOT, ("kernel pte %p", pte));
sys/arch/m88k/m88k/pmap.c
751
*pte++ = PG_NV;
sys/arch/m88k/m88k/pmap.c
755
*pte++ = pa | PG_SO | PG_RO | PG_W | PG_V;
sys/arch/m88k/m88k/pmap.c
761
*pte++ = pa | PG_SO | PG_RW | PG_M_U | PG_W | PG_V | CACHE_WT;
sys/arch/m88k/m88k/pmap.c
763
*pte++ = pa | PG_SO | PG_RW | PG_M_U | PG_W | PG_V;
sys/arch/m88k/m88k/pmap.c
769
*pte++ = pa | PG_SO | PG_RW | PG_M_U | PG_V;
sys/arch/m88k/m88k/pmap.c
772
DPRINTF(CD_BOOT, ("-%p, pa %08lx\n", pte, pa));
sys/arch/m88k/m88k/pmap.c
773
for (i = (pt_entry_t *)round_page((vaddr_t)pte) - pte; i != 0; i--)
sys/arch/m88k/m88k/pmap.c
774
*pte++ = PG_NV;
sys/arch/m88k/m88k/pmap.c
780
pte = (pt_entry_t *)ptepa + atop(pa);
sys/arch/m88k/m88k/pmap.c
782
*pte++ = pa | template;
sys/arch/m88k/m88k/pmap.c
957
pt_entry_t *pte, npte;
sys/arch/m88k/m88k/pmap.c
971
if ((pte = pmap_pte(pmap, va)) == NULL) {
sys/arch/m88k/m88k/pmap.c
973
pte = pmap_expand_kmap(va, flags & PMAP_CANFAIL);
sys/arch/m88k/m88k/pmap.c
975
pte = pmap_expand(pmap, va, flags & PMAP_CANFAIL);
sys/arch/m88k/m88k/pmap.c
978
if (pte == NULL) {
sys/arch/m88k/m88k/pmap.c
987
old_pa = ptoa(PG_PFNUM(*pte));
sys/arch/m88k/m88k/pmap.c
988
DPRINTF(CD_ENT, ("pmap_enter: old_pa %lx pte %x\n", old_pa, *pte));
sys/arch/m88k/m88k/pmap.c
995
if (wired && !(pmap_pte_w(pte)))
sys/arch/m88k/m88k/pmap.c
997
else if (!wired && pmap_pte_w(pte))
sys/arch/mips64/include/pmap.h
192
pt_entry_t **pde, *pte;
sys/arch/mips64/include/pmap.h
196
if ((pte = pde[uvtopde(va)]) == NULL)
sys/arch/mips64/include/pmap.h
198
return pte + uvtopte(va);
sys/arch/mips64/mips64/pmap.c
1025
pte = kvtopte(va);
sys/arch/mips64/mips64/pmap.c
1043
if ((pte = pde[uvtopde(va)]) == NULL) {
sys/arch/mips64/mips64/pmap.c
1044
pte = pool_get(&pmap_pg_pool, PR_NOWAIT | PR_ZERO);
sys/arch/mips64/mips64/pmap.c
1045
if (pte == NULL) {
sys/arch/mips64/mips64/pmap.c
1052
pde[uvtopde(va)] = pte;
sys/arch/mips64/mips64/pmap.c
1054
pte += uvtopte(va);
sys/arch/mips64/mips64/pmap.c
1057
if ((*pte & PG_V) && pa != pfn_to_pad(*pte)) {
sys/arch/mips64/mips64/pmap.c
1062
if ((*pte & PG_V) == 0) {
sys/arch/mips64/mips64/pmap.c
1067
if ((*pte & PG_WIRED) != 0 && wired == 0)
sys/arch/mips64/mips64/pmap.c
1069
else if ((*pte & PG_WIRED) == 0 && wired != 0)
sys/arch/mips64/mips64/pmap.c
1137
opte = *pte;
sys/arch/mips64/mips64/pmap.c
1138
*pte = npte;
sys/arch/mips64/mips64/pmap.c
1176
opte = *pte;
sys/arch/mips64/mips64/pmap.c
1177
*pte = npte;
sys/arch/mips64/mips64/pmap.c
1209
pt_entry_t *pte, npte, opte;
sys/arch/mips64/mips64/pmap.c
1224
pte = kvtopte(va);
sys/arch/mips64/mips64/pmap.c
1225
if ((*pte & PG_V) == 0) {
sys/arch/mips64/mips64/pmap.c
1229
if ((*pte & PG_WIRED) == 0)
sys/arch/mips64/mips64/pmap.c
1232
opte = *pte;
sys/arch/mips64/mips64/pmap.c
1233
*pte = npte;
sys/arch/mips64/mips64/pmap.c
1247
pt_entry_t *pte, entry;
sys/arch/mips64/mips64/pmap.c
1260
pte = kvtopte(va);
sys/arch/mips64/mips64/pmap.c
1261
for (; va < eva; va += PAGE_SIZE, pte++) {
sys/arch/mips64/mips64/pmap.c
1262
entry = *pte;
sys/arch/mips64/mips64/pmap.c
1267
*pte = PG_NV | PG_G;
sys/arch/mips64/mips64/pmap.c
1278
pt_entry_t *pte;
sys/arch/mips64/mips64/pmap.c
1283
pte = kvtopte(va);
sys/arch/mips64/mips64/pmap.c
1285
pte = pmap_pte_lookup(pmap, va);
sys/arch/mips64/mips64/pmap.c
1286
if (pte == NULL)
sys/arch/mips64/mips64/pmap.c
1290
if (*pte & PG_V) {
sys/arch/mips64/mips64/pmap.c
1291
if (*pte & PG_WIRED) {
sys/arch/mips64/mips64/pmap.c
1292
*pte &= ~PG_WIRED;
sys/arch/mips64/mips64/pmap.c
1311
pt_entry_t *pte;
sys/arch/mips64/mips64/pmap.c
1329
pte = kvtopte(va);
sys/arch/mips64/mips64/pmap.c
1330
if (*pte & PG_V)
sys/arch/mips64/mips64/pmap.c
1331
pa = pfn_to_pad(*pte) | (va & PAGE_MASK);
sys/arch/mips64/mips64/pmap.c
1337
pte = pmap_pte_lookup(pmap, va);
sys/arch/mips64/mips64/pmap.c
1338
if (pte != NULL && (*pte & PG_V) != 0)
sys/arch/mips64/mips64/pmap.c
1339
pa = pfn_to_pad(*pte) | (va & PAGE_MASK);
sys/arch/mips64/mips64/pmap.c
1447
pt_entry_t *pte, entry;
sys/arch/mips64/mips64/pmap.c
1471
pte = kvtopte(pv->pv_va);
sys/arch/mips64/mips64/pmap.c
1472
entry = *pte;
sys/arch/mips64/mips64/pmap.c
1479
*pte = entry;
sys/arch/mips64/mips64/pmap.c
1485
pte = pmap_pte_lookup(pv->pv_pmap, pv->pv_va);
sys/arch/mips64/mips64/pmap.c
1486
if (pte == NULL)
sys/arch/mips64/mips64/pmap.c
1488
entry = *pte;
sys/arch/mips64/mips64/pmap.c
1494
*pte = entry;
sys/arch/mips64/mips64/pmap.c
1562
pt_entry_t *pte, entry;
sys/arch/mips64/mips64/pmap.c
1573
pte = kvtopte(va);
sys/arch/mips64/mips64/pmap.c
1575
pte = pmap_pte_lookup(pmap, va);
sys/arch/mips64/mips64/pmap.c
1576
if (pte == NULL)
sys/arch/mips64/mips64/pmap.c
1580
entry = *pte;
sys/arch/mips64/mips64/pmap.c
1601
if (pmap_pte_cas(pte, old_entry, entry) != old_entry) {
sys/arch/mips64/mips64/pmap.c
1607
*pte = entry;
sys/arch/mips64/mips64/pmap.c
1634
pt_entry_t *pte, entry;
sys/arch/mips64/mips64/pmap.c
1650
pte = kvtopte(pv->pv_va);
sys/arch/mips64/mips64/pmap.c
1651
entry = *pte;
sys/arch/mips64/mips64/pmap.c
1654
*pte = entry;
sys/arch/mips64/mips64/pmap.c
1660
pte = pmap_pte_lookup(pv->pv_pmap, pv->pv_va);
sys/arch/mips64/mips64/pmap.c
1661
if (pte == NULL)
sys/arch/mips64/mips64/pmap.c
1663
entry = *pte;
sys/arch/mips64/mips64/pmap.c
1666
*pte = entry;
sys/arch/mips64/mips64/pmap.c
2017
pt_entry_t *pte;
sys/arch/mips64/mips64/pmap.c
2034
pte = pmap_pte_lookup(pmap, uva);
sys/arch/mips64/mips64/pmap.c
2035
if (pte != NULL && (*pte & PG_V) != 0 && (*pte & pg_xi) == 0) {
sys/arch/mips64/mips64/pmap.c
2036
pa = pfn_to_pad(*pte) | (uva & PAGE_MASK);
sys/arch/mips64/mips64/pmap.c
231
pmap_pte_cas(pt_entry_t *pte, pt_entry_t o, pt_entry_t n)
sys/arch/mips64/mips64/pmap.c
234
return atomic_cas_ulong((unsigned long *)pte, o, n);
sys/arch/mips64/mips64/pmap.c
236
return atomic_cas_uint((unsigned int *)pte, o, n);
sys/arch/mips64/mips64/pmap.c
537
pt_entry_t **pde, *pte;
sys/arch/mips64/mips64/pmap.c
556
if ((pte = pde[j]) == NULL)
sys/arch/mips64/mips64/pmap.c
560
if (pte[k] != PG_NV)
sys/arch/mips64/mips64/pmap.c
563
"index %u", pmap, pte, k);
sys/arch/mips64/mips64/pmap.c
566
pool_put(&pmap_pg_pool, pte);
sys/arch/mips64/mips64/pmap.c
637
pt_entry_t ***seg, **pde, *pte, entry;
sys/arch/mips64/mips64/pmap.c
657
pte = kvtopte(sva);
sys/arch/mips64/mips64/pmap.c
658
for (va = sva; va < eva; va += PAGE_SIZE, pte++) {
sys/arch/mips64/mips64/pmap.c
659
entry = *pte;
sys/arch/mips64/mips64/pmap.c
669
*pte = PG_NV | PG_G;
sys/arch/mips64/mips64/pmap.c
697
pte = *pde + uvtopte(va);
sys/arch/mips64/mips64/pmap.c
699
va += PAGE_SIZE, pte++) {
sys/arch/mips64/mips64/pmap.c
700
entry = *pte;
sys/arch/mips64/mips64/pmap.c
713
*pte = PG_NV;
sys/arch/mips64/mips64/pmap.c
743
pt_entry_t *pte, entry, p;
sys/arch/mips64/mips64/pmap.c
757
pte = kvtopte(pv->pv_va);
sys/arch/mips64/mips64/pmap.c
758
entry = *pte;
sys/arch/mips64/mips64/pmap.c
766
*pte = entry;
sys/arch/mips64/mips64/pmap.c
771
pte = pmap_pte_lookup(pv->pv_pmap, pv->pv_va);
sys/arch/mips64/mips64/pmap.c
772
if (pte == NULL)
sys/arch/mips64/mips64/pmap.c
774
entry = *pte;
sys/arch/mips64/mips64/pmap.c
784
*pte = entry;
sys/arch/mips64/mips64/pmap.c
883
pt_entry_t ***seg, **pde, *pte, entry, p;
sys/arch/mips64/mips64/pmap.c
917
pte = kvtopte(sva);
sys/arch/mips64/mips64/pmap.c
918
for (va = sva; va < eva; va += PAGE_SIZE, pte++) {
sys/arch/mips64/mips64/pmap.c
919
entry = *pte;
sys/arch/mips64/mips64/pmap.c
927
*pte = entry;
sys/arch/mips64/mips64/pmap.c
955
pte = *pde + uvtopte(va);
sys/arch/mips64/mips64/pmap.c
957
va += PAGE_SIZE, pte++) {
sys/arch/mips64/mips64/pmap.c
958
entry = *pte;
sys/arch/mips64/mips64/pmap.c
974
*pte = entry;
sys/arch/mips64/mips64/pmap.c
999
pt_entry_t **pde, *pte, npte, opte;
sys/arch/powerpc/powerpc/pmap.c
1011
void *pte;
sys/arch/powerpc/powerpc/pmap.c
1014
if ((pte = pmap_ptedinhash(pted)) != NULL) {
sys/arch/powerpc/powerpc/pmap.c
1016
struct pte_64 *ptp64 = pte;
sys/arch/powerpc/powerpc/pmap.c
1027
struct pte_32 *ptp32 = pte;
sys/arch/powerpc/powerpc/pmap.c
1996
void *pte;
sys/arch/powerpc/powerpc/pmap.c
2018
if ((pte = pmap_ptedinhash(pted)) != NULL) {
sys/arch/powerpc/powerpc/pmap.c
2019
struct pte_64 *ptp64 = pte;
sys/arch/powerpc/powerpc/pmap.c
2043
void *pte;
sys/arch/powerpc/powerpc/pmap.c
2059
if ((pte = pmap_ptedinhash(pted)) != NULL) {
sys/arch/powerpc/powerpc/pmap.c
2060
struct pte_32 *ptp32 = pte;
sys/arch/powerpc/powerpc/pmap.c
2089
void *pte;
sys/arch/powerpc/powerpc/pmap.c
2123
if ((pte = pmap_ptedinhash(pted)) != NULL)
sys/arch/powerpc/powerpc/pmap.c
2124
pte_zap(pte, pted);
sys/arch/powerpc/powerpc/pmap.c
2349
void *pte;
sys/arch/powerpc/powerpc/pmap.c
2353
if ((pte = pmap_ptedinhash(pted)) != NULL)
sys/arch/powerpc/powerpc/pmap.c
2354
pte_zap(pte, pted);
sys/arch/powerpc/powerpc/pmap.c
2456
void *pte;
sys/arch/powerpc/powerpc/pmap.c
2460
if ((pte = pmap_ptedinhash(pted)) != NULL)
sys/arch/powerpc/powerpc/pmap.c
2461
pte_zap(pte, pted);
sys/arch/powerpc/powerpc/pmap.c
508
pmap_pte2flags(u_int32_t pte)
sys/arch/powerpc/powerpc/pmap.c
510
return (((pte & PTE_REF_32) ? PG_PMAP_REF : 0) |
sys/arch/powerpc/powerpc/pmap.c
511
((pte & PTE_CHG_32) ? PG_PMAP_MOD : 0));
sys/arch/powerpc/powerpc/pmap.c
663
void *pte;
sys/arch/powerpc/powerpc/pmap.c
672
if ((pte = pmap_ptedinhash(pted)) != NULL)
sys/arch/powerpc/powerpc/pmap.c
673
pte_zap(pte, pted);
sys/arch/powerpc/powerpc/pmap.c
802
struct pte_64 *pte = pmap_ptable64;
sys/arch/powerpc/powerpc/pmap.c
804
pte += (idx ^ (PTED_HID(pted) ? pmap_ptab_mask : 0)) * 8;
sys/arch/powerpc/powerpc/pmap.c
805
pte += PTED_PTEGIDX(pted);
sys/arch/powerpc/powerpc/pmap.c
814
(PTED_HID(pted) ? PTE_HID_64 : 0)) == pte->pte_hi)
sys/arch/powerpc/powerpc/pmap.c
815
return (pte);
sys/arch/powerpc/powerpc/pmap.c
817
struct pte_32 *pte = pmap_ptable32;
sys/arch/powerpc/powerpc/pmap.c
819
pte += (idx ^ (PTED_HID(pted) ? pmap_ptab_mask : 0)) * 8;
sys/arch/powerpc/powerpc/pmap.c
820
pte += PTED_PTEGIDX(pted);
sys/arch/powerpc/powerpc/pmap.c
829
(PTED_HID(pted) ? PTE_HID_32 : 0)) == pte->pte_hi)
sys/arch/powerpc/powerpc/pmap.c
830
return (pte);
sys/arch/powerpc/powerpc/pmap.c
842
pte_del(void *pte, vaddr_t va)
sys/arch/powerpc/powerpc/pmap.c
845
((struct pte_64 *)pte)->pte_hi &= ~PTE_VALID_64;
sys/arch/powerpc/powerpc/pmap.c
847
((struct pte_32 *)pte)->pte_hi &= ~PTE_VALID_32;
sys/arch/powerpc/powerpc/pmap.c
857
pte_zap(void *pte, struct pte_desc *pted)
sys/arch/powerpc/powerpc/pmap.c
859
pte_del(pte, pted->pted_va);
sys/arch/powerpc/powerpc/pmap.c
866
((struct pte_64 *)pte)->pte_lo & (PTE_REF_64|PTE_CHG_64));
sys/arch/powerpc/powerpc/pmap.c
869
((struct pte_32 *)pte)->pte_lo & (PTE_REF_32|PTE_CHG_32));
sys/arch/powerpc/powerpc/pmap.c
972
void *pte;
sys/arch/powerpc/powerpc/pmap.c
975
if ((pte = pmap_ptedinhash(pted)) != NULL) {
sys/arch/powerpc/powerpc/pmap.c
977
struct pte_64 *ptp64 = pte;
sys/arch/powerpc/powerpc/pmap.c
980
struct pte_32 *ptp32 = pte;
sys/arch/powerpc64/include/pmap.h
82
struct pte;
sys/arch/powerpc64/include/pmap.h
83
struct pte *pmap_get_kernel_pte(vaddr_t);
sys/arch/powerpc64/powerpc64/db_memrw.c
115
old_pte_lo = pte->pte_lo;
sys/arch/powerpc64/powerpc64/db_memrw.c
116
pte->pte_lo = (old_pte_lo & ~PTE_PP) | PTE_RW;
sys/arch/powerpc64/powerpc64/db_memrw.c
123
pte->pte_lo = old_pte_lo;
sys/arch/powerpc64/powerpc64/db_memrw.c
81
struct pte *pte;
sys/arch/powerpc64/powerpc64/db_memrw.c
93
pte = pmap_get_kernel_pte((vaddr_t)dst);
sys/arch/powerpc64/powerpc64/db_memrw.c
94
if (pte == NULL) {
sys/arch/powerpc64/powerpc64/pmap.c
117
struct pte *pmap_ptable;
sys/arch/powerpc64/powerpc64/pmap.c
1201
struct pte *pte;
sys/arch/powerpc64/powerpc64/pmap.c
121
#define HTABMEMSZ (pmap_ptab_cnt * 8 * sizeof(struct pte))
sys/arch/powerpc64/powerpc64/pmap.c
1222
if ((pte = pmap_ptedinhash(pted)) != NULL) {
sys/arch/powerpc64/powerpc64/pmap.c
1223
pte_del(pte, pmap_pted2ava(pted));
sys/arch/powerpc64/powerpc64/pmap.c
1226
pte->pte_lo = pted->pted_pte.pte_lo;
sys/arch/powerpc64/powerpc64/pmap.c
1228
pte->pte_hi |= PTE_VALID;
sys/arch/powerpc64/powerpc64/pmap.c
1244
void *pte;
sys/arch/powerpc64/powerpc64/pmap.c
1278
if ((pte = pmap_ptedinhash(pted)) != NULL)
sys/arch/powerpc64/powerpc64/pmap.c
1279
pte_zap(pte, pted);
sys/arch/powerpc64/powerpc64/pmap.c
132
struct pte pted_pte;
sys/arch/powerpc64/powerpc64/pmap.c
1352
struct pte *pte;
sys/arch/powerpc64/powerpc64/pmap.c
1360
pte = pte_lookup(vsid, va);
sys/arch/powerpc64/powerpc64/pmap.c
1361
if (pte)
sys/arch/powerpc64/powerpc64/pmap.c
1362
pte_del(pte, pmap_ava(vsid, va));
sys/arch/powerpc64/powerpc64/pmap.c
1365
if (pte)
sys/arch/powerpc64/powerpc64/pmap.c
1394
struct pte *pte;
sys/arch/powerpc64/powerpc64/pmap.c
1397
if ((pte = pmap_ptedinhash(pted)) != NULL)
sys/arch/powerpc64/powerpc64/pmap.c
1398
pte_zap(pte, pted);
sys/arch/powerpc64/powerpc64/pmap.c
1416
struct pte *pte;
sys/arch/powerpc64/powerpc64/pmap.c
1422
if ((pte = pmap_ptedinhash(pted)) != NULL) {
sys/arch/powerpc64/powerpc64/pmap.c
1423
pte_zap(pte, pted);
sys/arch/powerpc64/powerpc64/pmap.c
1426
pte->pte_lo = pted->pted_pte.pte_lo;
sys/arch/powerpc64/powerpc64/pmap.c
1428
pte->pte_hi |= PTE_VALID;
sys/arch/powerpc64/powerpc64/pmap.c
1441
struct pte *pte;
sys/arch/powerpc64/powerpc64/pmap.c
1458
pte = pte_lookup(vsid, va);
sys/arch/powerpc64/powerpc64/pmap.c
1459
if (pte)
sys/arch/powerpc64/powerpc64/pmap.c
1460
*pa = (pte->pte_lo & PTE_RPGN) | (va & PAGE_MASK);
sys/arch/powerpc64/powerpc64/pmap.c
1463
return (pte != NULL);
sys/arch/powerpc64/powerpc64/pmap.c
1743
struct pte *
sys/arch/powerpc64/powerpc64/pmap.c
295
struct pte *
sys/arch/powerpc64/powerpc64/pmap.c
298
struct pte *pte;
sys/arch/powerpc64/powerpc64/pmap.c
309
pte = pmap_ptable + (idx * 8);
sys/arch/powerpc64/powerpc64/pmap.c
310
pte += PTED_PTEGIDX(pted); /* increment by index into pteg */
sys/arch/powerpc64/powerpc64/pmap.c
319
(PTED_HID(pted) ? PTE_HID : 0)) == pte->pte_hi)
sys/arch/powerpc64/powerpc64/pmap.c
320
return pte;
sys/arch/powerpc64/powerpc64/pmap.c
657
struct pte *
sys/arch/powerpc64/powerpc64/pmap.c
661
struct pte *pte;
sys/arch/powerpc64/powerpc64/pmap.c
667
pte = pmap_ptable + (idx * 8);
sys/arch/powerpc64/powerpc64/pmap.c
673
if ((pte[i].pte_hi & ~PTE_WIRED) == pte_hi)
sys/arch/powerpc64/powerpc64/pmap.c
674
return &pte[i];
sys/arch/powerpc64/powerpc64/pmap.c
679
pte = pmap_ptable + (idx * 8);
sys/arch/powerpc64/powerpc64/pmap.c
683
if ((pte[i].pte_hi & ~PTE_WIRED) == pte_hi)
sys/arch/powerpc64/powerpc64/pmap.c
684
return &pte[i];
sys/arch/powerpc64/powerpc64/pmap.c
696
pte_del(struct pte *pte, uint64_t ava)
sys/arch/powerpc64/powerpc64/pmap.c
698
pte->pte_hi &= ~PTE_VALID;
sys/arch/powerpc64/powerpc64/pmap.c
707
pte_zap(struct pte *pte, struct pte_desc *pted)
sys/arch/powerpc64/powerpc64/pmap.c
709
pte_del(pte, pmap_pted2ava(pted));
sys/arch/powerpc64/powerpc64/pmap.c
735
struct pte *pte = &pted->pted_pte;
sys/arch/powerpc64/powerpc64/pmap.c
742
pte->pte_hi = (pmap_pted2avpn(pted) & PTE_AVPN) | PTE_VALID;
sys/arch/powerpc64/powerpc64/pmap.c
743
pte->pte_lo = (pa & PTE_RPGN);
sys/arch/powerpc64/powerpc64/pmap.c
746
pte->pte_hi |= PTE_WIRED;
sys/arch/powerpc64/powerpc64/pmap.c
749
pte->pte_lo |= PTE_RW;
sys/arch/powerpc64/powerpc64/pmap.c
751
pte->pte_lo |= PTE_RO;
sys/arch/powerpc64/powerpc64/pmap.c
754
pte->pte_lo |= pmap_acwimgn(prot, cache);
sys/arch/powerpc64/powerpc64/pmap.c
760
struct pte *pte;
sys/arch/powerpc64/powerpc64/pmap.c
768
if ((pte = pmap_ptedinhash(pted)) != NULL)
sys/arch/powerpc64/powerpc64/pmap.c
769
pte_zap(pte, pted);
sys/arch/powerpc64/powerpc64/pmap.c
787
pte = pmap_ptable + (idx * 8);
sys/arch/powerpc64/powerpc64/pmap.c
789
if (pte[i].pte_hi & PTE_VALID)
sys/arch/powerpc64/powerpc64/pmap.c
795
pte[i].pte_hi = pted->pted_pte.pte_hi & ~PTE_VALID;
sys/arch/powerpc64/powerpc64/pmap.c
796
pte[i].pte_lo = pted->pted_pte.pte_lo;
sys/arch/powerpc64/powerpc64/pmap.c
798
pte[i].pte_hi |= PTE_VALID;
sys/arch/powerpc64/powerpc64/pmap.c
805
pte = pmap_ptable + (idx ^ pmap_ptab_mask) * 8;
sys/arch/powerpc64/powerpc64/pmap.c
807
if (pte[i].pte_hi & PTE_VALID)
sys/arch/powerpc64/powerpc64/pmap.c
813
pte[i].pte_hi = pted->pted_pte.pte_hi & ~PTE_VALID;
sys/arch/powerpc64/powerpc64/pmap.c
814
pte[i].pte_lo = pted->pted_pte.pte_lo;
sys/arch/powerpc64/powerpc64/pmap.c
816
pte[i].pte_hi |= (PTE_HID|PTE_VALID);
sys/arch/powerpc64/powerpc64/pmap.c
829
pte = pmap_ptable;
sys/arch/powerpc64/powerpc64/pmap.c
830
pte += (idx ^ (PTED_HID(pted) ? pmap_ptab_mask : 0)) * 8;
sys/arch/powerpc64/powerpc64/pmap.c
831
pte += PTED_PTEGIDX(pted); /* increment by index into pteg */
sys/arch/powerpc64/powerpc64/pmap.c
833
if ((pte->pte_hi & PTE_WIRED) == 0)
sys/arch/powerpc64/powerpc64/pmap.c
844
if (pte->pte_hi & PTE_VALID) {
sys/arch/powerpc64/powerpc64/pmap.c
847
avpn = pte->pte_hi & PTE_AVPN;
sys/arch/powerpc64/powerpc64/pmap.c
852
idx ^= ((pte->pte_hi & PTE_HID) ? pmap_ptab_mask : 0);
sys/arch/powerpc64/powerpc64/pmap.c
855
pte_del(pte, vpn << PAGE_SHIFT);
sys/arch/powerpc64/powerpc64/pmap.c
859
pte->pte_hi = pted->pted_pte.pte_hi & ~PTE_VALID;
sys/arch/powerpc64/powerpc64/pmap.c
861
pte->pte_hi |= PTE_HID;
sys/arch/powerpc64/powerpc64/pmap.c
862
pte->pte_lo = pted->pted_pte.pte_lo;
sys/arch/powerpc64/powerpc64/pmap.c
864
pte->pte_hi |= PTE_VALID;
sys/arch/powerpc64/powerpc64/pmap.c
874
struct pte *pte;
sys/arch/powerpc64/powerpc64/pmap.c
888
if ((pte = pmap_ptedinhash(pted)) != NULL)
sys/arch/powerpc64/powerpc64/pmap.c
889
pte_zap(pte, pted);
sys/arch/riscv64/riscv64/pmap.c
1633
uint64_t pte, access_bits;
sys/arch/riscv64/riscv64/pmap.c
1656
pte = VP_Lx(pted->pted_pte) | attr | access_bits | PTE_V;
sys/arch/riscv64/riscv64/pmap.c
1657
*pl3 = access_bits ? pte : 0;
sys/arch/sh/include/mmu.h
104
#define sh_tlb_update(a, va, pte) sh4_tlb_update(a, va, pte)
sys/arch/sh/include/mmu.h
88
#define sh_tlb_update(a, va, pte) (*__sh_tlb_update)(a, va, pte)
sys/arch/sh/include/mmu.h
96
#define sh_tlb_update(a, va, pte) sh3_tlb_update(a, va, pte)
sys/arch/sh/sh/locore_c.c
170
pt_entry_t *pte;
sys/arch/sh/sh/locore_c.c
178
pte = __pmap_kpte_lookup(vpn);
sys/arch/sh/sh/locore_c.c
179
KDASSERT(pte && *pte != 0);
sys/arch/sh/sh/locore_c.c
182
md_upte->data = (*pte & PG_HW_BITS) | PG_D | PG_V;
sys/arch/sh/sh/locore_c.c
195
pt_entry_t *pte;
sys/arch/sh/sh/locore_c.c
204
pte = __pmap_kpte_lookup(vpn);
sys/arch/sh/sh/locore_c.c
205
KDASSERT(pte && *pte != 0);
sys/arch/sh/sh/locore_c.c
212
md_upte->data = (*pte & PG_HW_BITS) |
sys/arch/sh/sh/mmu_sh3.c
109
sh3_tlb_update(int asid, vaddr_t va, uint32_t pte)
sys/arch/sh/sh/mmu_sh3.c
113
KDASSERT(asid < 0x100 && (pte & ~PGOFSET) != 0 && va != 0);
sys/arch/sh/sh/mmu_sh3.c
123
_reg_write_4(SH3_PTEL, pte & PG_HW_BITS);
sys/arch/sh/sh/mmu_sh4.c
145
sh4_tlb_update(int asid, vaddr_t va, uint32_t pte)
sys/arch/sh/sh/mmu_sh4.c
151
KDASSERT(asid < 0x100 && (pte & ~PGOFSET) != 0 && va != 0);
sys/arch/sh/sh/mmu_sh4.c
163
ptel = pte & PG_HW_BITS;
sys/arch/sh/sh/mmu_sh4.c
164
if (pte & _PG_PCMCIA) {
sys/arch/sh/sh/mmu_sh4.c
166
(pte >> _PG_PCMCIA_SHIFT) & SH4_PTEA_SA_MASK);
sys/arch/sh/sh/pmap.c
1016
pt_entry_t *pte;
sys/arch/sh/sh/pmap.c
1023
if (((pte = __pmap_pte_lookup(pmap, va)) == NULL) ||
sys/arch/sh/sh/pmap.c
1024
((entry = *pte) == 0))
sys/arch/sh/sh/pmap.c
1036
*pte = entry;
sys/arch/sh/sh/pmap.c
248
pt_entry_t *pte = (pt_entry_t *)va;
sys/arch/sh/sh/pmap.c
249
for (j = 0; j < __PMAP_PTP_PG_N; j++, pte++)
sys/arch/sh/sh/pmap.c
250
KDASSERT(*pte == 0);
sys/arch/sh/sh/pmap.c
294
pt_entry_t entry, *pte;
sys/arch/sh/sh/pmap.c
356
pte = __pmap_kpte_lookup(va);
sys/arch/sh/sh/pmap.c
358
pte = __pmap_pte_alloc(pmap, va);
sys/arch/sh/sh/pmap.c
359
if (pte == NULL) {
sys/arch/sh/sh/pmap.c
369
*pte = entry;
sys/arch/sh/sh/pmap.c
396
pt_entry_t *pte, oentry;
sys/arch/sh/sh/pmap.c
399
if ((pte = __pmap_pte_lookup(pmap, va)) == NULL ||
sys/arch/sh/sh/pmap.c
400
((oentry = *pte) == 0))
sys/arch/sh/sh/pmap.c
419
*pte = entry;
sys/arch/sh/sh/pmap.c
487
pt_entry_t *pte, entry;
sys/arch/sh/sh/pmap.c
493
if ((pte = __pmap_pte_lookup(pmap, va)) == NULL ||
sys/arch/sh/sh/pmap.c
494
(entry = *pte) == 0)
sys/arch/sh/sh/pmap.c
503
*pte = 0;
sys/arch/sh/sh/pmap.c
556
pt_entry_t *pte, entry;
sys/arch/sh/sh/pmap.c
570
pte = __pmap_kpte_lookup(va);
sys/arch/sh/sh/pmap.c
572
KDASSERT(*pte == 0);
sys/arch/sh/sh/pmap.c
573
*pte = entry;
sys/arch/sh/sh/pmap.c
581
pt_entry_t *pte;
sys/arch/sh/sh/pmap.c
589
pte = __pmap_kpte_lookup(va);
sys/arch/sh/sh/pmap.c
590
KDASSERT(pte != NULL);
sys/arch/sh/sh/pmap.c
591
if (*pte == 0)
sys/arch/sh/sh/pmap.c
594
if (SH_HAS_VIRTUAL_ALIAS && PHYS_TO_VM_PAGE(*pte & PG_PPN))
sys/arch/sh/sh/pmap.c
596
*pte = 0;
sys/arch/sh/sh/pmap.c
605
pt_entry_t *pte;
sys/arch/sh/sh/pmap.c
614
pte = __pmap_pte_lookup(pmap, va);
sys/arch/sh/sh/pmap.c
615
if (pte == NULL || *pte == 0)
sys/arch/sh/sh/pmap.c
619
*pap = (*pte & PG_PPN) | (va & PGOFSET);
sys/arch/sh/sh/pmap.c
628
pt_entry_t *pte, entry, protbits;
sys/arch/sh/sh/pmap.c
660
if (((pte = __pmap_pte_lookup(pmap, va)) == NULL) ||
sys/arch/sh/sh/pmap.c
661
(entry = *pte) == 0)
sys/arch/sh/sh/pmap.c
672
*pte = entry;
sys/arch/sh/sh/pmap.c
749
pt_entry_t *pte, entry;
sys/arch/sh/sh/pmap.c
751
if ((pte = __pmap_pte_lookup(pmap, va)) == NULL ||
sys/arch/sh/sh/pmap.c
752
(entry = *pte) == 0 ||
sys/arch/sh/sh/pmap.c
756
*pte = entry & ~_PG_WIRED;
sys/arch/sh/sh/pmap.c
811
pt_entry_t *pte;
sys/arch/sh/sh/pmap.c
827
if ((pte = __pmap_pte_lookup(pmap, va)) == NULL)
sys/arch/sh/sh/pmap.c
829
if ((*pte & PG_V) == 0)
sys/arch/sh/sh/pmap.c
831
*pte &= ~PG_V;
sys/arch/sh/sh/pmap.c
853
pt_entry_t *pte, entry;
sys/arch/sh/sh/pmap.c
877
if ((pte = __pmap_pte_lookup(pmap, va)) == NULL)
sys/arch/sh/sh/pmap.c
879
entry = *pte;
sys/arch/sh/sh/pmap.c
886
*pte = entry & ~PG_D;
sys/arch/sh/sh/pmap.c
957
pt_entry_t *ptp, *pte;
sys/arch/sh/sh/pmap.c
959
if ((pte = __pmap_pte_lookup(pmap, va)) != NULL)
sys/arch/sh/sh/pmap.c
960
return (pte);
sys/dev/acpi/acpidmar.c
136
struct pte_entry *pte;
sys/dev/acpi/acpidmar.c
2089
dom->pte = iommu_alloc_page(iommu, &dom->ptep, 0);
sys/dev/acpi/acpidmar.c
2090
if (dom->pte == NULL) {
sys/dev/acpi/acpidmar.c
2144
iommu_free_page(dom->pte);
sys/dev/acpi/acpidmar.c
2716
_dumppte(struct pte_entry *pte, int lvl, vaddr_t va)
sys/dev/acpi/acpidmar.c
2724
if (pte[i].val & PTE_P) {
sys/dev/acpi/acpidmar.c
2726
npte = (void *)PMAP_DIRECT_MAP((pte[i].val & PTE_PADDR_MASK));
sys/dev/acpi/acpidmar.c
2728
pte[i].val, (pte[i].val >> 9) & 7);
sys/dev/acpi/acpidmar.c
2732
pte[i].val, va | sh);
sys/dev/acpi/acpidmar.c
2759
_dumppte(dom->pte, 3, 0);
sys/dev/acpi/acpidmar.c
3747
struct pte_entry *pte;
sys/dev/acpi/acpidmar.c
3749
pte = (void *)PMAP_DIRECT_MAP(ptep);
sys/dev/acpi/acpidmar.c
3751
if (!(pte[i].val & PTE_P))
sys/dev/acpi/acpidmar.c
3754
pb = pte[i].val & ~VTD_PAGE_MASK;
sys/dev/acpi/acpidmar.c
3758
pte[i].val == PTE_R ? 'r' : ' ',
sys/dev/acpi/acpidmar.c
3759
pte[i].val & PTE_W ? 'w' : ' ',
sys/dev/acpi/acpidmar.c
530
struct pte_entry *pte, *npte;
sys/dev/acpi/acpidmar.c
541
npte = dom->pte;
sys/dev/acpi/acpidmar.c
545
pte = &npte[idx];
sys/dev/acpi/acpidmar.c
549
pte->val = create ? (pa | flags) : 0;
sys/dev/acpi/acpidmar.c
550
iommu_flush_cache(iommu, pte, sizeof(*pte));
sys/dev/acpi/acpidmar.c
554
val = pte->val;
sys/dev/acpi/acpidmar.c
569
val = pte->val;
sys/dev/acpi/acpidmar.c
571
pte->val = paddr | PTE_P | PTE_R | PTE_W;
sys/dev/acpi/acpidmar.c
572
iommu_flush_cache(iommu, pte, sizeof(*pte));
sys/dev/acpi/acpidmar.c
573
val = pte->val;
sys/dev/acpi/acpidmar.c
602
pte_lvl(struct domain *dom, struct pte_entry *pte, vaddr_t va,
sys/dev/acpi/acpidmar.c
612
val = pte[idx].val;
sys/dev/acpi/acpidmar.c
627
val = pte[idx].val;
sys/dev/acpi/acpidmar.c
629
pte[idx].val = paddr | flags;
sys/dev/acpi/acpidmar.c
630
iommu_flush_cache(iommu, &pte[idx], sizeof(pte[idx]));
sys/dev/acpi/acpidmar.c
631
val = pte[idx].val;
sys/dev/acpi/acpidmar.c
647
struct pte_entry *pte;
sys/dev/acpi/acpidmar.c
658
pte = dom->pte;
sys/dev/acpi/acpidmar.c
659
pte = pte_lvl(dom, pte, va, 30,
sys/dev/acpi/acpidmar.c
662
if (pte == NULL)
sys/dev/acpi/acpidmar.c
664
pte = pte_lvl(dom, pte, va, 21,
sys/dev/acpi/acpidmar.c
667
if (pte == NULL)
sys/dev/acpi/acpidmar.c
677
pte[idx].val = create ? (pa | flags) : 0;
sys/dev/acpi/acpidmar.c
678
iommu_flush_cache(iommu, &pte[idx], sizeof(pte[idx]));
sys/dev/fdt/qcsmem.c
160
struct qcsmem_ptable_entry *pte;
sys/dev/fdt/qcsmem.c
218
pte = &ptable->entry[i];
sys/dev/fdt/qcsmem.c
219
if (!pte->offset || !pte->size)
sys/dev/fdt/qcsmem.c
221
if (pte->host[0] == QCSMEM_GLOBAL_HOST &&
sys/dev/fdt/qcsmem.c
222
pte->host[1] == QCSMEM_GLOBAL_HOST)
sys/dev/fdt/qcsmem.c
224
else if (pte->host[0] == QCSMEM_LOCAL_HOST &&
sys/dev/fdt/qcsmem.c
225
pte->host[1] < QCSMEM_HOST_COUNT)
sys/dev/fdt/qcsmem.c
226
part = &sc->sc_partitions[pte->host[1]];
sys/dev/fdt/qcsmem.c
227
else if (pte->host[1] == QCSMEM_LOCAL_HOST &&
sys/dev/fdt/qcsmem.c
228
pte->host[0] < QCSMEM_HOST_COUNT)
sys/dev/fdt/qcsmem.c
229
part = &sc->sc_partitions[pte->host[0]];
sys/dev/fdt/qcsmem.c
235
pte->offset;
sys/dev/fdt/qcsmem.c
241
if (pte->host[0] != phdr->host[0] ||
sys/dev/fdt/qcsmem.c
242
pte->host[1] != phdr->host[1]) {
sys/dev/fdt/qcsmem.c
244
pte->host[0], phdr->host[0],
sys/dev/fdt/qcsmem.c
245
pte->host[1], phdr->host[1]);
sys/dev/fdt/qcsmem.c
248
if (pte->size != phdr->size) {
sys/dev/fdt/qcsmem.c
250
pte->size, phdr->size);
sys/dev/fdt/qcsmem.c
259
part->size = pte->size;
sys/dev/fdt/qcsmem.c
260
part->cacheline = pte->cacheline;
sys/dev/pci/agp_i810.c
618
u_int32_t pte = 0;
sys/dev/pci/agp_i810.c
628
pte = v | INTEL_ENABLED;
sys/dev/pci/agp_i810.c
636
pte |= (v & 0x0000000f00000000ULL) >> 28;
sys/dev/pci/agp_i810.c
642
bus_space_write_4(isc->gtt_bst, isc->gtt_bsh, wroff, pte);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm_sdma.c
228
uint64_t *pte;
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm_sdma.c
276
pte = (uint64_t *)&(p->job->ibs->ptr[p->num_dw_left]);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm_sdma.c
278
pte[i] = amdgpu_vm_map_gart(p->pages_addr, addr);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm_sdma.c
279
pte[i] |= flags;
sys/dev/pci/drm/amd/display/dc/dce/dce_mem_input.c
142
const struct pte_setting *pte = &pte_settings[mi_tiling][mi_bpp];
sys/dev/pci/drm/amd/display/dc/dce/dce_mem_input.c
144
unsigned int page_width = log_2(pte->page_width);
sys/dev/pci/drm/amd/display/dc/dce/dce_mem_input.c
145
unsigned int page_height = log_2(pte->page_height);
sys/dev/pci/drm/amd/display/dc/dce/dce_mem_input.c
147
pte->min_pte_before_flip_vert_scan :
sys/dev/pci/drm/amd/display/dc/dce/dce_mem_input.c
148
pte->min_pte_before_flip_horiz_scan;
sys/dev/pci/drm/amd/display/dc/dce/dce_mem_input.c
159
DVMM_PTE_REQ_PER_CHUNK, pte->pte_req_per_chunk,
sys/dev/pci/drm/amd/display/dc/dce110/dce110_mem_input_v.c
570
const unsigned int *pte = get_dvmm_hw_setting(tiling_info, format, false);
sys/dev/pci/drm/amd/display/dc/dce110/dce110_mem_input_v.c
577
unsigned int temp_page_width = pte[1];
sys/dev/pci/drm/amd/display/dc/dce110/dce110_mem_input_v.c
578
unsigned int temp_page_height = pte[2];
sys/dev/pci/drm/amd/display/dc/dce110/dce110_mem_input_v.c
598
min_pte_before_flip = pte[4];
sys/dev/pci/drm/amd/display/dc/dce110/dce110_mem_input_v.c
602
min_pte_before_flip = pte[3];
sys/dev/pci/drm/amd/display/dc/dce110/dce110_mem_input_v.c
620
set_reg_field_value(value, pte[5], UNP_DVMM_PTE_ARB_CONTROL, DVMM_PTE_REQ_PER_CHUNK);
sys/dev/pci/drm/i915/display/intel_dpt.c
36
static void gen8_set_pte(void __iomem *addr, gen8_pte_t pte)
sys/dev/pci/drm/i915/display/intel_dpt.c
38
writeq(pte, addr);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1678
static int check_present_pte(pte_t *pte, unsigned long addr, void *data)
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1680
pte_t ptent = ptep_get(pte);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1691
static int check_absent_pte(pte_t *pte, unsigned long addr, void *data)
sys/dev/pci/drm/i915/gem/selftests/i915_gem_mman.c
1693
pte_t ptent = ptep_get(pte);
sys/dev/pci/drm/i915/gt/gen6_ppgtt.c
104
memset32(vaddr + pte, scratch_pte, count);
sys/dev/pci/drm/i915/gt/gen6_ppgtt.c
106
pte = 0;
sys/dev/pci/drm/i915/gt/gen6_ppgtt.c
81
unsigned int pte = first_entry % GEN6_PTES;
sys/dev/pci/drm/i915/gt/gen6_ppgtt.c
87
const unsigned int count = min(num_entries, GEN6_PTES - pte);
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
272
unsigned int pte = gen8_pd_index(start, 0);
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
286
GEM_BUG_ON(pte % 16);
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
288
pte /= 16;
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
292
memset64(vaddr + pte,
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
36
gen8_pte_t pte = addr | GEN8_PAGE_PRESENT | GEN8_PAGE_RW;
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
39
pte &= ~GEN8_PAGE_RW;
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
48
pte |= PPAT_UNCACHED;
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
51
pte |= PPAT_DISPLAY_ELLC;
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
54
pte |= PPAT_CACHED;
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
58
return pte;
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
65
gen8_pte_t pte = addr | GEN8_PAGE_PRESENT | GEN8_PAGE_RW;
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
68
pte &= ~GEN8_PAGE_RW;
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
71
pte |= GEN12_PPGTT_PTE_LM;
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
74
pte |= GEN12_PPGTT_PTE_PAT0;
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
77
pte |= GEN12_PPGTT_PTE_PAT1;
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
80
pte |= GEN12_PPGTT_PTE_PAT2;
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
83
pte |= MTL_PPGTT_PTE_PAT3;
sys/dev/pci/drm/i915/gt/gen8_ppgtt.c
85
return pte;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
1478
gen6_pte_t pte = GEN6_PTE_ADDR_ENCODE(addr) | GEN6_PTE_VALID;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
1483
pte |= GEN6_PTE_CACHE_LLC;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
1486
pte |= GEN6_PTE_UNCACHED;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
1492
return pte;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
1499
gen6_pte_t pte = GEN6_PTE_ADDR_ENCODE(addr) | GEN6_PTE_VALID;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
1503
pte |= GEN7_PTE_CACHE_L3_LLC;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
1506
pte |= GEN6_PTE_CACHE_LLC;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
1509
pte |= GEN6_PTE_UNCACHED;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
1515
return pte;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
1522
gen6_pte_t pte = GEN6_PTE_ADDR_ENCODE(addr) | GEN6_PTE_VALID;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
1525
pte |= BYT_PTE_WRITEABLE;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
1528
pte |= BYT_PTE_SNOOPED_BY_CPU_CACHES;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
1530
return pte;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
1537
gen6_pte_t pte = HSW_PTE_ADDR_ENCODE(addr) | GEN6_PTE_VALID;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
1540
pte |= HSW_WB_LLC_AGE3;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
1542
return pte;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
1549
gen6_pte_t pte = HSW_PTE_ADDR_ENCODE(addr) | GEN6_PTE_VALID;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
1555
pte |= HSW_WT_ELLC_LLC_AGE3;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
1558
pte |= HSW_WB_ELLC_LLC_AGE3;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
1562
return pte;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
1565
static dma_addr_t gen6_pte_decode(u64 pte, bool *is_present, bool *is_local)
sys/dev/pci/drm/i915/gt/intel_ggtt.c
1567
*is_present = pte & GEN6_PTE_VALID;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
1570
return ((pte & 0xff0) << 28) | (pte & ~0xfff);
sys/dev/pci/drm/i915/gt/intel_ggtt.c
294
gen8_pte_t pte = addr | GEN8_PAGE_PRESENT;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
299
pte |= GEN12_GGTT_PTE_LM;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
302
pte |= MTL_GGTT_PTE_PAT0;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
305
pte |= MTL_GGTT_PTE_PAT1;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
307
return pte;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
314
gen8_pte_t pte = addr | GEN8_PAGE_PRESENT;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
317
pte |= GEN12_GGTT_PTE_LM;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
319
return pte;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
322
static dma_addr_t gen8_ggtt_pte_decode(u64 pte, bool *is_present, bool *is_local)
sys/dev/pci/drm/i915/gt/intel_ggtt.c
324
*is_present = pte & GEN8_PAGE_PRESENT;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
325
*is_local = pte & GEN12_GGTT_PTE_LM;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
327
return pte & GEN12_GGTT_PTE_ADDR_MASK;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
371
const gen8_pte_t pte)
sys/dev/pci/drm/i915/gt/intel_ggtt.c
428
*cs++ = lower_32_bits(pte | addr);
sys/dev/pci/drm/i915/gt/intel_ggtt.c
429
*cs++ = upper_32_bits(pte | addr);
sys/dev/pci/drm/i915/gt/intel_ggtt.c
439
memset64((u64 *)cs, pte, n_ptes);
sys/dev/pci/drm/i915/gt/intel_ggtt.c
471
static void gen8_set_pte(void __iomem *addr, gen8_pte_t pte)
sys/dev/pci/drm/i915/gt/intel_ggtt.c
473
writeq(pte, addr);
sys/dev/pci/drm/i915/gt/intel_ggtt.c
488
gen8_pte_t __iomem *pte =
sys/dev/pci/drm/i915/gt/intel_ggtt.c
491
gen8_set_pte(pte, ggtt->vm.pte_encode(addr, pat_index, flags));
sys/dev/pci/drm/i915/gt/intel_ggtt.c
500
gen8_pte_t __iomem *pte =
sys/dev/pci/drm/i915/gt/intel_ggtt.c
503
return ggtt->vm.pte_decode(gen8_get_pte(pte), is_present, is_local);
sys/dev/pci/drm/i915/gt/intel_ggtt.c
511
gen8_pte_t pte;
sys/dev/pci/drm/i915/gt/intel_ggtt.c
513
pte = ggtt->vm.pte_encode(addr, pat_index, flags);
sys/dev/pci/drm/i915/gt/intel_ggtt.c
515
gen8_ggtt_bind_ptes(ggtt, offset, NULL, 1, pte))
sys/dev/pci/drm/i915/gt/intel_ggtt.c
653
gen6_pte_t __iomem *pte =
sys/dev/pci/drm/i915/gt/intel_ggtt.c
656
iowrite32(vm->pte_encode(addr, pat_index, flags), pte);
sys/dev/pci/drm/i915/gt/intel_ggtt.c
666
gen6_pte_t __iomem *pte =
sys/dev/pci/drm/i915/gt/intel_ggtt.c
669
return vm->pte_decode(ioread32(pte), is_present, is_local);
sys/dev/pci/drm/i915/gt/intel_gtt.h
315
dma_addr_t (*pte_decode)(u64 pte, bool *is_present, bool *is_local);
sys/dev/pci/drm/i915/gvt/dmabuf.c
44
#define GEN8_DECODE_PTE(pte) (pte & GENMASK_ULL(63, 12))
sys/dev/pci/drm/i915/gvt/gtt.c
230
static void write_pte64(struct i915_ggtt *ggtt, unsigned long index, u64 pte)
sys/dev/pci/drm/i915/gvt/gtt.c
234
writeq(pte, addr);
sys/dev/pci/drm/i915/gvt/gtt.c
2770
gen8_pte_t pte;
sys/dev/pci/drm/i915/gvt/gtt.c
2780
pte = mm->ggtt_mm.host_ggtt_aperture[idx];
sys/dev/pci/drm/i915/gvt/gtt.c
2781
if (pte & GEN8_PAGE_PRESENT)
sys/dev/pci/drm/i915/gvt/gtt.c
2782
write_pte64(vgpu->gvt->gt->ggtt, offset + idx, pte);
sys/dev/pci/drm/i915/gvt/gtt.c
2788
pte = mm->ggtt_mm.host_ggtt_hidden[idx];
sys/dev/pci/drm/i915/gvt/gtt.c
2789
if (pte & GEN8_PAGE_PRESENT)
sys/dev/pci/drm/i915/gvt/gtt.c
2790
write_pte64(vgpu->gvt->gt->ggtt, offset + idx, pte);
sys/dev/pci/drm/i915/gvt/gtt.c
430
DEFINE_PPGTT_GMA_TO_INDEX(gen8, pte, (gma >> 12 & 0x1ff));
sys/dev/pci/drm/i915/i915_mm.c
53
static int remap_sg(pte_t *pte, unsigned long addr, void *data)
sys/dev/pci/drm/i915/i915_mm.c
61
set_pte_at(r->mm, addr, pte,
sys/dev/pci/drm/i915/i915_mm.c
76
static int remap_pfn(pte_t *pte, unsigned long addr, void *data)
sys/dev/pci/drm/i915/i915_mm.c
81
set_pte_at(r->mm, addr, pte, pte_mkspecial(pfn_pte(r->pfn, r->prot)));
sys/dev/pci/drm/radeon/radeon_vm.c
828
uint64_t pte;
sys/dev/pci/drm/radeon/radeon_vm.c
841
pte = radeon_bo_gpu_offset(pt);
sys/dev/pci/drm/radeon/radeon_vm.c
842
pte += (addr & mask) * 8;
sys/dev/pci/drm/radeon/radeon_vm.c
844
if ((last_pte + 8 * count) != pte) {
sys/dev/pci/drm/radeon/radeon_vm.c
853
last_pte = pte;
sys/dev/usb/xhci.c
1760
uint64_t *pte;
sys/dev/usb/xhci.c
1765
(void **)&pte, npage * sizeof(uint64_t), XHCI_SPAD_TABLE_ALIGN,
sys/dev/usb/xhci.c
1779
pte[i] = htole64(
usr.sbin/bgpd/rde.c
3036
rde_dump_adjout_as(struct rde_peer *peer, struct pt_entry *pte,
usr.sbin/bgpd/rde.c
3062
rib.exit_nexthop.aid = pte->aid;
usr.sbin/bgpd/rde.c
3063
rib.true_nexthop.aid = pte->aid;
usr.sbin/bgpd/rde.c
3065
pt_getaddr(pte, &rib.prefix);
usr.sbin/bgpd/rde.c
3066
rib.prefixlen = pte->prefixlen;
usr.sbin/bgpd/rde.c
3075
if (peer_has_add_path(peer, pte->aid, CAPA_AP_SEND)) {
usr.sbin/bgpd/rde.c
3173
rde_dump_adjout_filter(struct rde_peer *peer, struct pt_entry *pte,
usr.sbin/bgpd/rde.c
3195
rde_dump_adjout_as(peer, pte, p, attrs, req->pid, req->flags);
usr.sbin/bgpd/rde.c
3211
rde_dump_adjout_upcall(struct rde_peer *peer, struct pt_entry *pte,
usr.sbin/bgpd/rde.c
3216
rde_dump_adjout_filter(peer, pte, p, &ctx->req);
usr.sbin/bgpd/rde.c
3345
struct pt_entry *pte;
usr.sbin/bgpd/rde.c
3350
pte = pt_get(&req->prefix,
usr.sbin/bgpd/rde.c
3352
if (pte == NULL)
usr.sbin/bgpd/rde.c
3356
peer, pte);
usr.sbin/bgpd/rde.c
3359
peer, pte, p)) {
usr.sbin/bgpd/rde.c
3361
peer, pte, p, ctx);
usr.sbin/bgpd/rde.c
3366
pte = pt_lookup(&req->prefix);
usr.sbin/bgpd/rde.c
3368
pte = pt_get(&req->prefix,
usr.sbin/bgpd/rde.c
3371
if (pte == NULL)
usr.sbin/bgpd/rde.c
3375
for (p = adjout_prefix_first(peer, pte);
usr.sbin/bgpd/rde.c
3377
p = adjout_prefix_next(peer, pte, p)) {
usr.sbin/bgpd/rde.c
3378
rde_dump_adjout_upcall(peer, pte, p,
usr.sbin/bgpd/rde.c
3630
rde_up_flush_upcall(struct rde_peer *peer, struct pt_entry *pte,
usr.sbin/bgpd/rde.c
3633
adjout_prefix_withdraw(peer, pte, p);
usr.sbin/bgpd/rde.c
4859
struct pt_entry *pte;
usr.sbin/bgpd/rde.c
4866
pte = pt_get_flow(f);
usr.sbin/bgpd/rde.c
4867
if (pte == NULL)
usr.sbin/bgpd/rde.c
4868
pte = pt_add_flow(f);
usr.sbin/bgpd/rde.c
4870
if (prefix_flowspec_update(peerself, state, pte, path_id_tx) == 1)
usr.sbin/bgpd/rde.c
4877
struct pt_entry *pte;
usr.sbin/bgpd/rde.c
4879
pte = pt_get_flow(f);
usr.sbin/bgpd/rde.c
4880
if (pte == NULL)
usr.sbin/bgpd/rde.c
4883
if (prefix_flowspec_withdraw(peerself, pte) == 1)
usr.sbin/bgpd/rde_adjout.c
419
adjout_prefix_index(struct pt_entry *pte, struct adjout_prefix *p)
usr.sbin/bgpd/rde_adjout.c
421
ptrdiff_t idx = p - pte->adjout;
usr.sbin/bgpd/rde_adjout.c
423
if (idx < 0 || idx > pte->adjoutlen)
usr.sbin/bgpd/rde_adjout.c
435
struct pt_entry *pte)
usr.sbin/bgpd/rde_adjout.c
440
for (i = 0; i < pte->adjoutlen; i++) {
usr.sbin/bgpd/rde_adjout.c
441
p = &pte->adjout[i];
usr.sbin/bgpd/rde_adjout.c
458
adjout_prefix_with_attrs(struct pt_entry *pte, uint32_t path_id_tx,
usr.sbin/bgpd/rde_adjout.c
464
for (i = 0; i < pte->adjoutlen; i++) {
usr.sbin/bgpd/rde_adjout.c
465
p = &pte->adjout[i];
usr.sbin/bgpd/rde_adjout.c
482
adjout_prefix_first(struct rde_peer *peer, struct pt_entry *pte)
usr.sbin/bgpd/rde_adjout.c
488
if (peer_has_add_path(peer, pte->aid, CAPA_AP_SEND))
usr.sbin/bgpd/rde_adjout.c
491
for (i = 0; i < pte->adjoutlen; i++) {
usr.sbin/bgpd/rde_adjout.c
492
p = &pte->adjout[i];
usr.sbin/bgpd/rde_adjout.c
506
adjout_prefix_next(struct rde_peer *peer, struct pt_entry *pte,
usr.sbin/bgpd/rde_adjout.c
512
if (!peer_has_add_path(peer, pte->aid, CAPA_AP_SEND))
usr.sbin/bgpd/rde_adjout.c
515
i = adjout_prefix_index(pte, last);
usr.sbin/bgpd/rde_adjout.c
516
for (; i < pte->adjoutlen; i++)
usr.sbin/bgpd/rde_adjout.c
517
if (pte->adjout[i].path_id_tx != last->path_id_tx)
usr.sbin/bgpd/rde_adjout.c
519
for (; i < pte->adjoutlen; i++) {
usr.sbin/bgpd/rde_adjout.c
520
p = &pte->adjout[i];
usr.sbin/bgpd/rde_adjout.c
533
struct filterstate *state, struct pt_entry *pte, uint32_t path_id_tx)
usr.sbin/bgpd/rde_adjout.c
559
adjout_prefix_unlink(p, pte, peer);
usr.sbin/bgpd/rde_adjout.c
564
adjout_prefix_link(pte, peer, attrs, path_id_tx);
usr.sbin/bgpd/rde_adjout.c
568
pend_prefix_add(peer, attrs, pte, path_id_tx);
usr.sbin/bgpd/rde_adjout.c
576
adjout_prefix_withdraw(struct rde_peer *peer, struct pt_entry *pte,
usr.sbin/bgpd/rde_adjout.c
583
pend_prefix_add(peer, NULL, pte, p->path_id_tx);
usr.sbin/bgpd/rde_adjout.c
585
adjout_prefix_unlink(p, pte, peer);
usr.sbin/bgpd/rde_adjout.c
598
struct pt_entry *pte = NULL;
usr.sbin/bgpd/rde_adjout.c
606
pte = ctx->ctx_pt;
usr.sbin/bgpd/rde_adjout.c
607
if (pte->refcnt == 1)
usr.sbin/bgpd/rde_adjout.c
608
pte = pt_next(pte);
usr.sbin/bgpd/rde_adjout.c
612
return pte;
usr.sbin/bgpd/rde_adjout.c
625
struct pt_entry *pte, *next;
usr.sbin/bgpd/rde_adjout.c
634
pte = pt_first(ctx->ctx_aid);
usr.sbin/bgpd/rde_adjout.c
636
pte = prefix_restart(ctx);
usr.sbin/bgpd/rde_adjout.c
638
for (i = 0; pte != NULL; pte = next) {
usr.sbin/bgpd/rde_adjout.c
639
next = pt_next(pte);
usr.sbin/bgpd/rde_adjout.c
641
ctx->ctx_aid != pte->aid)
usr.sbin/bgpd/rde_adjout.c
645
pt_getaddr(pte, &addr);
usr.sbin/bgpd/rde_adjout.c
653
ctx->ctx_pt = pt_ref(pte);
usr.sbin/bgpd/rde_adjout.c
656
p = adjout_prefix_first(peer, pte);
usr.sbin/bgpd/rde_adjout.c
659
ctx->ctx_prefix_call(peer, pte, p, ctx->ctx_arg);
usr.sbin/bgpd/rde_adjout.c
738
adjout_prefix_link(struct pt_entry *pte, struct rde_peer *peer,
usr.sbin/bgpd/rde_adjout.c
748
if ((p = adjout_prefix_with_attrs(pte, path_id_tx, attrs)) == NULL) {
usr.sbin/bgpd/rde_adjout.c
749
p = adjout_prefix_alloc(pte, path_id_tx);
usr.sbin/bgpd/rde_adjout.c
761
adjout_prefix_unlink(struct adjout_prefix *p, struct pt_entry *pte,
usr.sbin/bgpd/rde_adjout.c
770
adjout_prefix_free(pte, p);
usr.sbin/bgpd/rde_adjout.c
775
adjout_prefix_resize(struct pt_entry *pte)
usr.sbin/bgpd/rde_adjout.c
780
avail = pte->adjoutavail;
usr.sbin/bgpd/rde_adjout.c
782
if ((new = reallocarray(pte->adjout, newlen, sizeof(*new))) == NULL)
usr.sbin/bgpd/rde_adjout.c
787
pte->adjout = new;
usr.sbin/bgpd/rde_adjout.c
788
pte->adjoutavail = newlen;
usr.sbin/bgpd/rde_adjout.c
796
adjout_prefix_alloc(struct pt_entry *pte, uint32_t path_id_tx)
usr.sbin/bgpd/rde_adjout.c
801
if (pte->adjoutlen + 1 > pte->adjoutavail)
usr.sbin/bgpd/rde_adjout.c
802
adjout_prefix_resize(pte);
usr.sbin/bgpd/rde_adjout.c
805
for (i = 0; i < pte->adjoutlen; i++) {
usr.sbin/bgpd/rde_adjout.c
806
if (pte->adjout[i].path_id_tx > path_id_tx)
usr.sbin/bgpd/rde_adjout.c
810
p = &pte->adjout[i];
usr.sbin/bgpd/rde_adjout.c
812
for (i = pte->adjoutlen; &pte->adjout[i] > p; i--)
usr.sbin/bgpd/rde_adjout.c
813
pte->adjout[i] = pte->adjout[i - 1];
usr.sbin/bgpd/rde_adjout.c
820
pte->adjoutlen++;
usr.sbin/bgpd/rde_adjout.c
827
adjout_prefix_free(struct pt_entry *pte, struct adjout_prefix *p)
usr.sbin/bgpd/rde_adjout.c
833
idx = adjout_prefix_index(pte, p);
usr.sbin/bgpd/rde_adjout.c
834
for (i = idx + 1; i < pte->adjoutlen; i++)
usr.sbin/bgpd/rde_adjout.c
835
pte->adjout[i - 1] = pte->adjout[i];
usr.sbin/bgpd/rde_adjout.c
837
p = &pte->adjout[pte->adjoutlen - 1];
usr.sbin/bgpd/rde_adjout.c
839
pte->adjoutlen--;
usr.sbin/bgpd/rde_peer.c
515
peer_reaper_upcall(struct rde_peer *peer, struct pt_entry *pte,
usr.sbin/bgpd/rde_peer.c
518
adjout_prefix_withdraw(peer, pte, p);
usr.sbin/bgpd/rde_peer.c
619
peer_blast_upcall(struct rde_peer *peer, struct pt_entry *pte,
usr.sbin/bgpd/rde_peer.c
622
pend_prefix_add(peer, p->attrs, pte, p->path_id_tx);
usr.sbin/bgpd/rde_prefix.c
168
pt_getaddr(struct pt_entry *pte, struct bgpd_addr *addr)
usr.sbin/bgpd/rde_prefix.c
174
addr->aid = pte->aid;
usr.sbin/bgpd/rde_prefix.c
177
addr->v4 = ((struct pt_entry4 *)pte)->prefix4;
usr.sbin/bgpd/rde_prefix.c
180
addr->v6 = ((struct pt_entry6 *)pte)->prefix6;
usr.sbin/bgpd/rde_prefix.c
184
addr->v4 = ((struct pt_entry_vpn4 *)pte)->prefix4;
usr.sbin/bgpd/rde_prefix.c
185
addr->rd = ((struct pt_entry_vpn4 *)pte)->rd;
usr.sbin/bgpd/rde_prefix.c
186
addr->labellen = ((struct pt_entry_vpn4 *)pte)->labellen;
usr.sbin/bgpd/rde_prefix.c
188
((struct pt_entry_vpn4 *)pte)->labelstack,
usr.sbin/bgpd/rde_prefix.c
192
addr->v6 = ((struct pt_entry_vpn6 *)pte)->prefix6;
usr.sbin/bgpd/rde_prefix.c
193
addr->rd = ((struct pt_entry_vpn6 *)pte)->rd;
usr.sbin/bgpd/rde_prefix.c
194
addr->labellen = ((struct pt_entry_vpn6 *)pte)->labellen;
usr.sbin/bgpd/rde_prefix.c
196
((struct pt_entry_vpn6 *)pte)->labelstack,
usr.sbin/bgpd/rde_prefix.c
200
evpn = (struct pt_entry_evpn *)pte;
usr.sbin/bgpd/rde_prefix.c
220
pflow = (struct pt_entry_flow *)pte;
usr.sbin/bgpd/rde_prefix.c
231
pt_getflowspec(struct pt_entry *pte, uint8_t **flow)
usr.sbin/bgpd/rde_prefix.c
235
switch (pte->aid) {
usr.sbin/bgpd/rde_prefix.c
238
pflow = (struct pt_entry_flow *)pte;
usr.sbin/bgpd/rde_prefix.c
343
struct pt_entry *pte;
usr.sbin/bgpd/rde_prefix.c
345
pte = pt_fill(prefix, prefixlen);
usr.sbin/bgpd/rde_prefix.c
346
return RB_FIND(pt_tree, &pttable, pte);
usr.sbin/bgpd/rde_prefix.c
352
struct pt_entry *pte;
usr.sbin/bgpd/rde_prefix.c
354
pte = pt_fill(prefix, prefixlen);
usr.sbin/bgpd/rde_prefix.c
355
return RB_NFIND(pt_tree, &pttable, pte);
usr.sbin/bgpd/rde_prefix.c
417
struct pt_entry *pte;
usr.sbin/bgpd/rde_prefix.c
423
pte = pt_fill(&addr, 0);
usr.sbin/bgpd/rde_prefix.c
424
return RB_NFIND(pt_tree, &pttable, pte);
usr.sbin/bgpd/rde_prefix.c
428
pt_next(struct pt_entry *pte)
usr.sbin/bgpd/rde_prefix.c
430
return RB_NEXT(pt_tree, &pttable, pte);
usr.sbin/bgpd/rde_prefix.c
434
pt_remove(struct pt_entry *pte)
usr.sbin/bgpd/rde_prefix.c
436
if (pte->refcnt != 0)
usr.sbin/bgpd/rde_prefix.c
439
if (RB_REMOVE(pt_tree, &pttable, pte) == NULL)
usr.sbin/bgpd/rde_prefix.c
441
pt_free(pte);
usr.sbin/bgpd/rde_prefix.c
622
pt_free(struct pt_entry *pte)
usr.sbin/bgpd/rde_prefix.c
624
rdemem.pt_cnt[pte->aid]--;
usr.sbin/bgpd/rde_prefix.c
625
rdemem.pt_size[pte->aid] -= pte->len;
usr.sbin/bgpd/rde_prefix.c
626
free(pte);
usr.sbin/bgpd/rde_prefix.c
631
pt_writebuf(struct ibuf *buf, struct pt_entry *pte, int withdraw,
usr.sbin/bgpd/rde_prefix.c
634
struct pt_entry_vpn4 *pvpn4 = (struct pt_entry_vpn4 *)pte;
usr.sbin/bgpd/rde_prefix.c
635
struct pt_entry_vpn6 *pvpn6 = (struct pt_entry_vpn6 *)pte;
usr.sbin/bgpd/rde_prefix.c
636
struct pt_entry_flow *pflow = (struct pt_entry_flow *)pte;
usr.sbin/bgpd/rde_prefix.c
637
struct pt_entry_evpn *pevpn = (struct pt_entry_evpn *)pte;
usr.sbin/bgpd/rde_prefix.c
650
switch (pte->aid) {
usr.sbin/bgpd/rde_prefix.c
653
plen = pte->prefixlen;
usr.sbin/bgpd/rde_prefix.c
656
if (ibuf_add(tmp, pte->data, PREFIX_SIZE(plen) - 1) == -1)
usr.sbin/bgpd/rde_prefix.c
802
fatalx("%s: unknown aid %d", __func__, pte->aid);
usr.sbin/bgpd/rde_rib.c
1017
struct pt_entry *pte, uint32_t path_id_tx)
usr.sbin/bgpd/rde_rib.c
1025
re = rib_get(&flowrib, pte);
usr.sbin/bgpd/rde_rib.c
1027
re = rib_add(&flowrib, pte);
usr.sbin/bgpd/rde_rib.c
1062
prefix_flowspec_withdraw(struct rde_peer *peer, struct pt_entry *pte)
usr.sbin/bgpd/rde_rib.c
1067
re = rib_get(&flowrib, pte);
usr.sbin/bgpd/rde_rib.c
265
rib_get(struct rib *rib, struct pt_entry *pte)
usr.sbin/bgpd/rde_rib.c
270
xre.prefix = pte;
usr.sbin/bgpd/rde_rib.c
316
rib_add(struct rib *rib, struct pt_entry *pte)
usr.sbin/bgpd/rde_rib.c
324
re->prefix = pt_ref(pte);
usr.sbin/bgpd/rde_rib.c
917
struct pt_entry *pte;
usr.sbin/bgpd/rde_rib.c
921
pte = pt_get(prefix, prefixlen);
usr.sbin/bgpd/rde_rib.c
922
if (pte == NULL)
usr.sbin/bgpd/rde_rib.c
923
pte = pt_add(prefix, prefixlen);
usr.sbin/bgpd/rde_rib.c
924
re = rib_get(rib, pte);
usr.sbin/bgpd/rde_rib.c
926
re = rib_add(rib, pte);
usr.sbin/bgpd/rde_update.c
415
struct pt_entry *pte;
usr.sbin/bgpd/rde_update.c
446
pte = pt_get(&addr, 0);
usr.sbin/bgpd/rde_update.c
447
if (pte == NULL)
usr.sbin/bgpd/rde_update.c
448
pte = pt_add(&addr, 0);
usr.sbin/bgpd/rde_update.c
449
p = adjout_prefix_first(peer, pte);
usr.sbin/bgpd/rde_update.c
450
adjout_prefix_update(p, peer, &state, pte, 0);
usr.sbin/vmd/x86_vm.c
1020
uint64_t pte, pt_paddr, pte_paddr, mask, low_mask, high_mask;
usr.sbin/vmd/x86_vm.c
1075
if (read_mem(pte_paddr, &pte, pte_size)) {
usr.sbin/vmd/x86_vm.c
1081
pte);
usr.sbin/vmd/x86_vm.c
1084
if (!(pte & PG_V))
usr.sbin/vmd/x86_vm.c
1088
if ((mode == PROT_WRITE) && !(pte & PG_RW))
usr.sbin/vmd/x86_vm.c
1091
if ((exit->cpl > 0) && !(pte & PG_u))
usr.sbin/vmd/x86_vm.c
1094
pte = pte | PG_U;
usr.sbin/vmd/x86_vm.c
1096
pte = pte | PG_M;
usr.sbin/vmd/x86_vm.c
1097
if (write_mem(pte_paddr, &pte, pte_size)) {
usr.sbin/vmd/x86_vm.c
1104
if (pte & PG_PS)
usr.sbin/vmd/x86_vm.c
1108
pt_paddr = pte & PG_FRAME;
usr.sbin/vmd/x86_vm.c
1116
*pa = (pte & high_mask) | (va & low_mask);