Symbol: PG_A
sys/amd64/amd64/pmap.c
10021
pt_entry_t pte, PG_A, PG_M, PG_RW, PG_V;
sys/amd64/amd64/pmap.c
10025
PG_A = pmap_accessed_bit(pmap);
sys/amd64/amd64/pmap.c
10064
if ((pte & PG_A) != 0)
sys/amd64/amd64/pmap.c
10355
pt_entry_t *pte, PG_A, PG_M, PG_RW, PG_V;
sys/amd64/amd64/pmap.c
10363
PG_A = pmap_accessed_bit(pmap);
sys/amd64/amd64/pmap.c
10381
*pde |= PG_A;
sys/amd64/amd64/pmap.c
10403
*pte |= PG_M | PG_A;
sys/amd64/amd64/pmap.c
10405
*pte |= PG_A;
sys/amd64/amd64/pmap.c
4587
pt_entry_t PG_A, PG_M, PG_RW, PG_V;
sys/amd64/amd64/pmap.c
4591
PG_A = pmap_accessed_bit(pmap);
sys/amd64/amd64/pmap.c
4615
*pml5 = VM_PAGE_TO_PHYS(m) | PG_U | PG_RW | PG_V | PG_A | PG_M;
sys/amd64/amd64/pmap.c
4623
PG_A | PG_M;
sys/amd64/amd64/pmap.c
4635
*pml4 = VM_PAGE_TO_PHYS(m) | PG_U | PG_RW | PG_V | PG_A | PG_M;
sys/amd64/amd64/pmap.c
4652
PG_A | PG_M;
sys/amd64/amd64/pmap.c
4663
*pdp = VM_PAGE_TO_PHYS(m) | PG_U | PG_RW | PG_V | PG_A | PG_M;
sys/amd64/amd64/pmap.c
4691
*pd = VM_PAGE_TO_PHYS(m) | PG_U | PG_RW | PG_V | PG_A | PG_M;
sys/amd64/amd64/pmap.c
5206
pt_entry_t PG_G, PG_A, PG_M, PG_RW;
sys/amd64/amd64/pmap.c
5219
PG_G = PG_A = PG_M = PG_RW = 0;
sys/amd64/amd64/pmap.c
5288
PG_A = pmap_accessed_bit(pmap);
sys/amd64/amd64/pmap.c
5317
if ((tpte & PG_A) != 0)
sys/amd64/amd64/pmap.c
5962
pt_entry_t PG_A, PG_G, PG_M, PG_PKU_MASK, PG_RW, PG_V;
sys/amd64/amd64/pmap.c
5967
PG_A = pmap_accessed_bit(pmap);
sys/amd64/amd64/pmap.c
5987
if ((oldpde & PG_A) == 0) {
sys/amd64/amd64/pmap.c
6039
newpde = mptepa | PG_M | PG_A | (oldpde & PG_U) | PG_RW | PG_V;
sys/amd64/amd64/pmap.c
6152
pt_entry_t PG_G, PG_A, PG_M, PG_RW;
sys/amd64/amd64/pmap.c
6155
PG_A = pmap_accessed_bit(pmap);
sys/amd64/amd64/pmap.c
6177
if (oldpde & PG_A)
sys/amd64/amd64/pmap.c
6216
pt_entry_t oldpte, PG_A, PG_M, PG_RW;
sys/amd64/amd64/pmap.c
6219
PG_A = pmap_accessed_bit(pmap);
sys/amd64/amd64/pmap.c
6232
if (oldpte & PG_A)
sys/amd64/amd64/pmap.c
6510
pt_entry_t *pte, tpte, PG_A, PG_M, PG_RW;
sys/amd64/amd64/pmap.c
6554
PG_A = pmap_accessed_bit(pmap);
sys/amd64/amd64/pmap.c
6565
if (tpte & PG_A)
sys/amd64/amd64/pmap.c
6834
pt_entry_t allpte_PG_A, PG_A, PG_G, PG_M, PG_PKU_MASK, PG_RW, PG_V;
sys/amd64/amd64/pmap.c
6841
PG_A = pmap_accessed_bit(pmap);
sys/amd64/amd64/pmap.c
6899
allpte_PG_A = newpde & PG_A;
sys/amd64/amd64/pmap.c
6938
newpde &= ~PG_A | allpte_PG_A;
sys/amd64/amd64/pmap.c
6944
KASSERT((newpde & PG_A) != 0 || safe_to_clear_referenced(pmap, newpde),
sys/amd64/amd64/pmap.c
7110
pt_entry_t *pte, PG_G, PG_A, PG_M, PG_RW, PG_V;
sys/amd64/amd64/pmap.c
7118
PG_A = pmap_accessed_bit(pmap);
sys/amd64/amd64/pmap.c
7136
newpte = (pt_entry_t)(pa | PG_A | PG_V);
sys/amd64/amd64/pmap.c
7252
if (((origpte ^ newpte) & ~(PG_M | PG_A)) == 0)
sys/amd64/amd64/pmap.c
7280
if ((origpte & PG_A) != 0) {
sys/amd64/amd64/pmap.c
7351
if ((origpte & PG_A) != 0)
sys/amd64/amd64/pmap.c
7847
pt_entry_t PG_A, PG_M, PG_RW, PG_V;
sys/amd64/amd64/pmap.c
7852
PG_A = pmap_accessed_bit(pmap);
sys/amd64/amd64/pmap.c
7915
pde_store(pde, pa | PG_PS | PG_M | PG_A |
sys/amd64/amd64/pmap.c
8043
pt_entry_t *dst_pte, PG_A, PG_M, PG_V, ptetemp, *src_pte;
sys/amd64/amd64/pmap.c
8074
PG_A = pmap_accessed_bit(dst_pmap);
sys/amd64/amd64/pmap.c
8157
srcptepaddr &= ~PG_A;
sys/amd64/amd64/pmap.c
8205
*dst_pte = ptetemp & ~(PG_W | PG_M | PG_A);
sys/amd64/amd64/pmap.c
8690
pt_entry_t PG_A, PG_M, PG_RW, PG_V;
sys/amd64/amd64/pmap.c
8719
PG_A = pmap_accessed_bit(pmap);
sys/amd64/amd64/pmap.c
8721
mask |= PG_V | PG_A;
sys/amd64/amd64/pmap.c
8752
PG_A = pmap_accessed_bit(pmap);
sys/amd64/amd64/pmap.c
8754
mask |= PG_V | PG_A;
sys/amd64/amd64/pmap.c
8947
pt_entry_t *pte, PG_A, PG_M, PG_RW;
sys/amd64/amd64/pmap.c
8981
PG_A = pmap_accessed_bit(pmap);
sys/amd64/amd64/pmap.c
8995
if ((oldpde & PG_A) != 0) {
sys/amd64/amd64/pmap.c
9018
atomic_clear_long(pde, PG_A);
sys/amd64/amd64/pmap.c
9087
PG_A = pmap_accessed_bit(pmap);
sys/amd64/amd64/pmap.c
9097
if ((*pte & PG_A) != 0) {
sys/amd64/amd64/pmap.c
9099
atomic_clear_long(pte, PG_A);
sys/amd64/amd64/pmap.c
9149
pt_entry_t *pte, PG_A, PG_G, PG_M, PG_RW, PG_V;
sys/amd64/amd64/pmap.c
9166
PG_A = pmap_accessed_bit(pmap);
sys/amd64/amd64/pmap.c
9256
atomic_clear_long(pte, PG_M | PG_A);
sys/amd64/amd64/pmap.c
9257
} else if ((*pte & PG_A) != 0)
sys/amd64/amd64/pmap.c
9258
atomic_clear_long(pte, PG_A);
sys/amd64/amd64/pmap.c
9535
pt_entry_t PG_A, PG_M, PG_RW, PG_V;
sys/amd64/amd64/pmap.c
9539
PG_A = pmap_accessed_bit(pmap);
sys/amd64/amd64/pmap.c
9564
newpdpe = pdpgpa | PG_M | PG_A | (oldpdpe & PG_U) | PG_RW | PG_V;
sys/amd64/amd64/pmap.c
9565
KASSERT((oldpdpe & PG_A) != 0,
sys/amd64/vmm/vmm_instruction_emul.c
2138
if (!check_only && (pte32 & PG_A) == 0) {
sys/amd64/vmm/vmm_instruction_emul.c
2140
pte32, pte32 | PG_A) == 0) {
sys/amd64/vmm/vmm_instruction_emul.c
2221
if (!check_only && (pte & PG_A) == 0) {
sys/amd64/vmm/vmm_instruction_emul.c
2223
pte, pte | PG_A) == 0) {
sys/i386/i386/minidump_machdep_base.c
287
fakept[k] = (pa + (k * PAGE_SIZE)) | PG_V | PG_RW | PG_A | PG_M;
sys/i386/i386/pmap.c
1511
*PMAP2 = newpf | PG_RW | PG_V | PG_A | PG_M;
sys/i386/i386/pmap.c
1573
*PMAP1 = newpf | PG_RW | PG_V | PG_A | PG_M;
sys/i386/i386/pmap.c
1607
*PMAP3 = newpf | PG_RW | PG_V | PG_A | PG_M;
sys/i386/i386/pmap.c
1637
*eh_ptep = pde | PG_RW | PG_V | PG_A | PG_M;
sys/i386/i386/pmap.c
198
#define pmap_pte_u(pte) ((*(int *)pte & PG_A) != 0)
sys/i386/i386/pmap.c
2150
(pd_entry_t)(ptepa | PG_U | PG_RW | PG_V | PG_A | PG_M);
sys/i386/i386/pmap.c
2268
newpdir = (pd_entry_t) (ptppaddr | PG_V | PG_RW | PG_A | PG_M);
sys/i386/i386/pmap.c
2371
if ((tpte & PG_A) != 0)
sys/i386/i386/pmap.c
2767
if ((oldpde & PG_A) == 0 || (mpte = pmap_remove_pt_page(pmap, va)) ==
sys/i386/i386/pmap.c
2778
if ((oldpde & PG_A) == 0 ||
sys/i386/i386/pmap.c
2808
*PMAP1 = mptepa | PG_RW | PG_V | PG_A | PG_M;
sys/i386/i386/pmap.c
2827
*PMAP2 = mptepa | PG_RW | PG_V | PG_A | PG_M;
sys/i386/i386/pmap.c
2833
newpde = mptepa | PG_M | PG_A | (oldpde & PG_U) | PG_RW | PG_V;
sys/i386/i386/pmap.c
2834
KASSERT((oldpde & PG_A) != 0,
sys/i386/i386/pmap.c
2915
newpde = mptepa | PG_M | PG_A | PG_RW | PG_V;
sys/i386/i386/pmap.c
2973
if (oldpde & PG_A)
sys/i386/i386/pmap.c
3024
if (oldpte & PG_A)
sys/i386/i386/pmap.c
3239
if (tpte & PG_A)
sys/i386/i386/pmap.c
3537
allpte_PG_A = newpde & PG_A;
sys/i386/i386/pmap.c
3580
newpde &= ~PG_A | allpte_PG_A;
sys/i386/i386/pmap.c
3671
newpte = (pt_entry_t)(pa | PG_A | PG_V);
sys/i386/i386/pmap.c
3780
if (((origpte ^ newpte) & ~(PG_M | PG_A)) == 0)
sys/i386/i386/pmap.c
3808
if ((origpte & PG_A) != 0) {
sys/i386/i386/pmap.c
3879
if ((origpte & PG_A) != 0)
sys/i386/i386/pmap.c
399
*pt = pa | PG_V | PG_RW | PG_A | PG_M;
sys/i386/i386/pmap.c
4345
pde_store(pde, pa | PG_PS | PG_M | PG_A |
sys/i386/i386/pmap.c
4545
PG_A);
sys/i386/i386/pmap.c
4597
*cmap_pte2 = PG_V | PG_RW | VM_PAGE_TO_PHYS(m) | PG_A | PG_M |
sys/i386/i386/pmap.c
4628
*cmap_pte2 = PG_V | PG_RW | VM_PAGE_TO_PHYS(m) | PG_A | PG_M |
sys/i386/i386/pmap.c
4658
*cmap_pte1 = PG_V | VM_PAGE_TO_PHYS(src) | PG_A |
sys/i386/i386/pmap.c
4661
*cmap_pte2 = PG_V | PG_RW | VM_PAGE_TO_PHYS(dst) | PG_A | PG_M |
sys/i386/i386/pmap.c
4698
*cmap_pte1 = PG_V | VM_PAGE_TO_PHYS(a_pg) | PG_A |
sys/i386/i386/pmap.c
470
IdlePTD[a] = (KPTphys + ptoa(a)) | PG_V | PG_RW | PG_A | PG_M;
sys/i386/i386/pmap.c
4701
*cmap_pte2 = PG_V | PG_RW | VM_PAGE_TO_PHYS(b_pg) | PG_A |
sys/i386/i386/pmap.c
5086
rv = (*pte & (PG_A | PG_V)) == (PG_A | PG_V);
sys/i386/i386/pmap.c
519
*(pt_entry_t *)vm86pa = 0 | PG_RW | PG_U | PG_A | PG_M | PG_V;
sys/i386/i386/pmap.c
5207
if ((*pde & PG_A) != 0) {
sys/i386/i386/pmap.c
5229
atomic_clear_int((u_int *)pde, PG_A);
sys/i386/i386/pmap.c
524
*pt = (ISA_HOLE_START + ptoa(a)) | PG_RW | PG_U | PG_A |
sys/i386/i386/pmap.c
5257
if ((*pte & PG_A) != 0) {
sys/i386/i386/pmap.c
5258
atomic_clear_int((u_int *)pte, PG_A);
sys/i386/i386/pmap.c
537
IdlePTD[a >> PDRSHIFT] = a | PG_PS | PG_A | PG_M |
sys/i386/i386/pmap.c
5371
atomic_clear_int((u_int *)pte, PG_M | PG_A);
sys/i386/i386/pmap.c
5372
} else if ((*pte & PG_A) != 0)
sys/i386/i386/pmap.c
5373
atomic_clear_int((u_int *)pte, PG_A);
sys/i386/i386/pmap.c
5667
PG_A | PG_M | pmap_cache_bits(kernel_pmap, m->md.pat_mode,
sys/i386/i386/pmap.c
5850
if ((pte & PG_A) != 0)
sys/i386/i386/pmap.c
5943
*pte = PG_V | PG_RW | VM_PAGE_TO_PHYS(m) | PG_A | PG_M |
sys/i386/i386/pmap.c
5993
PG_M | PG_A | PG_RW | PG_V | pgeflag |
sys/i386/i386/pmap.c
6012
PTD[TRPTDI] = VM_PAGE_TO_PHYS(pd_m) | PG_M | PG_A | PG_RW | PG_V |
sys/i386/i386/pmap.c
6099
if ((opte & (PG_V | PG_A)) == (PG_V | PG_A))
sys/i386/i386/pmap.c
6102
if ((opte & (PG_V | PG_A)) == (PG_V | PG_A))
sys/i386/i386/pmap.c
6114
*pte = PG_V | PG_RW | PG_A | PG_M | VM_PAGE_TO_PHYS(ma[i]) |
sys/powerpc/aim/mmu_radix.c
1012
newpte |= PG_A;
sys/powerpc/aim/mmu_radix.c
1466
if ((tpte & PG_A) != 0)
sys/powerpc/aim/mmu_radix.c
1768
#define DMAP_PAGE_BITS (RPTE_VALID | RPTE_LEAF | RPTE_EAA_MASK | PG_M | PG_A)
sys/powerpc/aim/mmu_radix.c
2314
atomic_clear_long(pte, htobe64(PG_M | PG_A));
sys/powerpc/aim/mmu_radix.c
2315
} else if ((be64toh(*pte) & PG_A) != 0)
sys/powerpc/aim/mmu_radix.c
2316
atomic_clear_long(pte, htobe64(PG_A));
sys/powerpc/aim/mmu_radix.c
2632
PG_A));
sys/powerpc/aim/mmu_radix.c
2739
if ((newpde & ((PG_FRAME & L3_PAGE_MASK) | PG_A | PG_V)) != (PG_A | PG_V)) {
sys/powerpc/aim/mmu_radix.c
2759
pa = (newpde & (PG_PS_FRAME | PG_A | PG_V)) + L3_PAGE_SIZE - PAGE_SIZE;
sys/powerpc/aim/mmu_radix.c
2763
if ((oldpte & (PG_FRAME | PG_A | PG_V)) != pa) {
sys/powerpc/aim/mmu_radix.c
2852
newpte = (pt_entry_t)(pa | PG_A | PG_V | RPTE_LEAF);
sys/powerpc/aim/mmu_radix.c
2856
newpte |= PG_A;
sys/powerpc/aim/mmu_radix.c
2975
if (((origpte ^ newpte) & ~(PG_M | PG_A)) == 0) {
sys/powerpc/aim/mmu_radix.c
2976
if ((newpte & (PG_A|PG_M)) != (origpte & (PG_A|PG_M))) {
sys/powerpc/aim/mmu_radix.c
2981
if ((newpte & PG_A) != (origpte & PG_A))
sys/powerpc/aim/mmu_radix.c
2986
if (((origpte ^ newpte) & ~(PG_M | PG_A)) == 0)
sys/powerpc/aim/mmu_radix.c
3015
if ((origpte & PG_A) != 0)
sys/powerpc/aim/mmu_radix.c
3037
if ((origpte & PG_A) != 0)
sys/powerpc/aim/mmu_radix.c
3105
if ((origpte & PG_A) != 0)
sys/powerpc/aim/mmu_radix.c
3786
mask |= PG_V | PG_A;
sys/powerpc/aim/mmu_radix.c
3813
mask |= PG_V | PG_A;
sys/powerpc/aim/mmu_radix.c
3946
if ((oldl3e & PG_A) != 0) {
sys/powerpc/aim/mmu_radix.c
3968
atomic_clear_long(l3e, htobe64(PG_A));
sys/powerpc/aim/mmu_radix.c
4013
if ((be64toh(*pte) & PG_A) != 0) {
sys/powerpc/aim/mmu_radix.c
4014
atomic_clear_long(pte, htobe64(PG_A));
sys/powerpc/aim/mmu_radix.c
4112
pa |= PG_M | PG_A | PG_RW;
sys/powerpc/aim/mmu_radix.c
4651
if (obits & (PG_A|PG_M)) {
sys/powerpc/aim/mmu_radix.c
4678
attr_bits = RPTE_EAA_R | RPTE_EAA_W | RPTE_EAA_P | PG_M | PG_A;
sys/powerpc/aim/mmu_radix.c
4946
if ((oldpde & PG_A) == 0 || (mpte = pmap_remove_pt_page(pmap, va)) ==
sys/powerpc/aim/mmu_radix.c
4964
if ((oldpde & PG_A) == 0 || (mpte = vm_page_alloc_noobj(
sys/powerpc/aim/mmu_radix.c
4982
KASSERT((oldpde & PG_A) != 0,
sys/powerpc/aim/mmu_radix.c
5095
if (oldpde & PG_A)
sys/powerpc/aim/mmu_radix.c
5137
if (oldpte & PG_A)
sys/powerpc/aim/mmu_radix.c
5398
if (tpte & PG_A)
sys/powerpc/aim/mmu_radix.c
5839
if ((pte & PG_A) != 0)
sys/powerpc/aim/mmu_radix.c
6040
KASSERT((oldpdpe & PG_A) != 0,
sys/powerpc/aim/mmu_radix.c
6130
pteval = pa | RPTE_EAA_R | RPTE_EAA_W | RPTE_EAA_P | PG_M | PG_A;
sys/powerpc/aim/mmu_radix.c
864
PG_M | PG_A | RPTE_EAA_MASK | PG_V)
sys/powerpc/aim/mmu_radix.c
923
RPTE_EAA_W | RPTE_EAA_P | PG_M | PG_A);
sys/powerpc/aim/mmu_radix.c
971
((flags & VM_PROT_READ) && (startpte & PG_A))) {
sys/powerpc/aim/mmu_radix.c
998
newpte |= PG_A;