Symbol: PG_M
sys/amd64/amd64/pmap.c
10021
pt_entry_t pte, PG_A, PG_M, PG_RW, PG_V;
sys/amd64/amd64/pmap.c
10026
PG_M = pmap_modified_bit(pmap);
sys/amd64/amd64/pmap.c
10062
if ((pte & (PG_M | PG_RW)) == (PG_M | PG_RW))
sys/amd64/amd64/pmap.c
10355
pt_entry_t *pte, PG_A, PG_M, PG_RW, PG_V;
sys/amd64/amd64/pmap.c
10364
PG_M = pmap_modified_bit(pmap);
sys/amd64/amd64/pmap.c
10403
*pte |= PG_M | PG_A;
sys/amd64/amd64/pmap.c
4587
pt_entry_t PG_A, PG_M, PG_RW, PG_V;
sys/amd64/amd64/pmap.c
4592
PG_M = pmap_modified_bit(pmap);
sys/amd64/amd64/pmap.c
4615
*pml5 = VM_PAGE_TO_PHYS(m) | PG_U | PG_RW | PG_V | PG_A | PG_M;
sys/amd64/amd64/pmap.c
4623
PG_A | PG_M;
sys/amd64/amd64/pmap.c
4635
*pml4 = VM_PAGE_TO_PHYS(m) | PG_U | PG_RW | PG_V | PG_A | PG_M;
sys/amd64/amd64/pmap.c
4652
PG_A | PG_M;
sys/amd64/amd64/pmap.c
4663
*pdp = VM_PAGE_TO_PHYS(m) | PG_U | PG_RW | PG_V | PG_A | PG_M;
sys/amd64/amd64/pmap.c
4691
*pd = VM_PAGE_TO_PHYS(m) | PG_U | PG_RW | PG_V | PG_A | PG_M;
sys/amd64/amd64/pmap.c
5206
pt_entry_t PG_G, PG_A, PG_M, PG_RW;
sys/amd64/amd64/pmap.c
5219
PG_G = PG_A = PG_M = PG_RW = 0;
sys/amd64/amd64/pmap.c
5289
PG_M = pmap_modified_bit(pmap);
sys/amd64/amd64/pmap.c
5315
if ((tpte & (PG_M | PG_RW)) == (PG_M | PG_RW))
sys/amd64/amd64/pmap.c
5962
pt_entry_t PG_A, PG_G, PG_M, PG_PKU_MASK, PG_RW, PG_V;
sys/amd64/amd64/pmap.c
5969
PG_M = pmap_modified_bit(pmap);
sys/amd64/amd64/pmap.c
6039
newpde = mptepa | PG_M | PG_A | (oldpde & PG_U) | PG_RW | PG_V;
sys/amd64/amd64/pmap.c
6040
KASSERT((oldpde & (PG_M | PG_RW)) != PG_RW,
sys/amd64/amd64/pmap.c
6152
pt_entry_t PG_G, PG_A, PG_M, PG_RW;
sys/amd64/amd64/pmap.c
6156
PG_M = pmap_modified_bit(pmap);
sys/amd64/amd64/pmap.c
6175
if ((oldpde & (PG_M | PG_RW)) == (PG_M | PG_RW))
sys/amd64/amd64/pmap.c
6216
pt_entry_t oldpte, PG_A, PG_M, PG_RW;
sys/amd64/amd64/pmap.c
6220
PG_M = pmap_modified_bit(pmap);
sys/amd64/amd64/pmap.c
6230
if ((oldpte & (PG_M | PG_RW)) == (PG_M | PG_RW))
sys/amd64/amd64/pmap.c
6510
pt_entry_t *pte, tpte, PG_A, PG_M, PG_RW;
sys/amd64/amd64/pmap.c
6555
PG_M = pmap_modified_bit(pmap);
sys/amd64/amd64/pmap.c
6571
if ((tpte & (PG_M | PG_RW)) == (PG_M | PG_RW))
sys/amd64/amd64/pmap.c
6595
pt_entry_t PG_G, PG_M, PG_RW;
sys/amd64/amd64/pmap.c
6598
PG_M = pmap_modified_bit(pmap);
sys/amd64/amd64/pmap.c
6608
if ((oldpde & (PG_MANAGED | PG_M | PG_RW)) ==
sys/amd64/amd64/pmap.c
6609
(PG_MANAGED | PG_M | PG_RW)) {
sys/amd64/amd64/pmap.c
6614
newpde &= ~(PG_RW | PG_M);
sys/amd64/amd64/pmap.c
6646
pt_entry_t *pte, PG_G, PG_M, PG_RW, PG_V;
sys/amd64/amd64/pmap.c
6661
PG_M = pmap_modified_bit(pmap);
sys/amd64/amd64/pmap.c
6712
pbits &= ~(PG_RW | PG_M);
sys/amd64/amd64/pmap.c
6761
pbits &= ~(PG_RW | PG_M);
sys/amd64/amd64/pmap.c
6786
if ((pbits & (PG_MANAGED | PG_M | PG_RW)) ==
sys/amd64/amd64/pmap.c
6787
(PG_MANAGED | PG_M | PG_RW)) {
sys/amd64/amd64/pmap.c
6791
pbits &= ~(PG_RW | PG_M);
sys/amd64/amd64/pmap.c
6834
pt_entry_t allpte_PG_A, PG_A, PG_G, PG_M, PG_PKU_MASK, PG_RW, PG_V;
sys/amd64/amd64/pmap.c
6843
PG_M = pmap_modified_bit(pmap);
sys/amd64/amd64/pmap.c
6882
if ((newpde & (PG_M | PG_RW)) == PG_RW) {
sys/amd64/amd64/pmap.c
6910
if ((oldpte & (PG_M | PG_RW)) == PG_RW) {
sys/amd64/amd64/pmap.c
7110
pt_entry_t *pte, PG_G, PG_A, PG_M, PG_RW, PG_V;
sys/amd64/amd64/pmap.c
7120
PG_M = pmap_modified_bit(pmap);
sys/amd64/amd64/pmap.c
7138
newpte |= PG_M;
sys/amd64/amd64/pmap.c
7141
KASSERT((newpte & (PG_M | PG_RW)) != PG_M,
sys/amd64/amd64/pmap.c
7160
newpte |= PG_M;
sys/amd64/amd64/pmap.c
7252
if (((origpte ^ newpte) & ~(PG_M | PG_A)) == 0)
sys/amd64/amd64/pmap.c
7278
if ((origpte & (PG_M | PG_RW)) == (PG_M | PG_RW))
sys/amd64/amd64/pmap.c
7335
if ((newpte & PG_M) == 0 && (origpte & (PG_M | PG_RW)) ==
sys/amd64/amd64/pmap.c
7336
(PG_M | PG_RW)) {
sys/amd64/amd64/pmap.c
7847
pt_entry_t PG_A, PG_M, PG_RW, PG_V;
sys/amd64/amd64/pmap.c
7853
PG_M = pmap_modified_bit(pmap);
sys/amd64/amd64/pmap.c
7915
pde_store(pde, pa | PG_PS | PG_M | PG_A |
sys/amd64/amd64/pmap.c
8043
pt_entry_t *dst_pte, PG_A, PG_M, PG_V, ptetemp, *src_pte;
sys/amd64/amd64/pmap.c
8075
PG_M = pmap_modified_bit(dst_pmap);
sys/amd64/amd64/pmap.c
8205
*dst_pte = ptetemp & ~(PG_W | PG_M | PG_A);
sys/amd64/amd64/pmap.c
8491
pt_entry_t PG_M, PG_RW, PG_V;
sys/amd64/amd64/pmap.c
8528
PG_M = pmap_modified_bit(pmap);
sys/amd64/amd64/pmap.c
8617
if ((tpte & (PG_M | PG_RW)) == (PG_M | PG_RW)) {
sys/amd64/amd64/pmap.c
8690
pt_entry_t PG_A, PG_M, PG_RW, PG_V;
sys/amd64/amd64/pmap.c
8714
PG_M = pmap_modified_bit(pmap);
sys/amd64/amd64/pmap.c
8716
mask |= PG_RW | PG_M;
sys/amd64/amd64/pmap.c
8747
PG_M = pmap_modified_bit(pmap);
sys/amd64/amd64/pmap.c
8749
mask |= PG_RW | PG_M;
sys/amd64/amd64/pmap.c
8844
pt_entry_t oldpte, *pte, PG_M, PG_RW;
sys/amd64/amd64/pmap.c
8896
PG_M = pmap_modified_bit(pmap);
sys/amd64/amd64/pmap.c
8906
~(PG_RW | PG_M)))
sys/amd64/amd64/pmap.c
8908
if ((oldpte & PG_M) != 0)
sys/amd64/amd64/pmap.c
8947
pt_entry_t *pte, PG_A, PG_M, PG_RW;
sys/amd64/amd64/pmap.c
8982
PG_M = pmap_modified_bit(pmap);
sys/amd64/amd64/pmap.c
8987
if ((oldpde & (PG_M | PG_RW)) == (PG_M | PG_RW)) {
sys/amd64/amd64/pmap.c
9088
PG_M = pmap_modified_bit(pmap);
sys/amd64/amd64/pmap.c
9095
if ((*pte & (PG_M | PG_RW)) == (PG_M | PG_RW))
sys/amd64/amd64/pmap.c
9149
pt_entry_t *pte, PG_A, PG_G, PG_M, PG_RW, PG_V;
sys/amd64/amd64/pmap.c
9168
PG_M = pmap_modified_bit(pmap);
sys/amd64/amd64/pmap.c
9246
else if ((*pte & (PG_M | PG_RW)) == (PG_M | PG_RW)) {
sys/amd64/amd64/pmap.c
9256
atomic_clear_long(pte, PG_M | PG_A);
sys/amd64/amd64/pmap.c
9293
pt_entry_t *pte, PG_M, PG_RW;
sys/amd64/amd64/pmap.c
9321
PG_M = pmap_modified_bit(pmap);
sys/amd64/amd64/pmap.c
9336
atomic_clear_long(pte, PG_M | PG_RW);
sys/amd64/amd64/pmap.c
9355
PG_M = pmap_modified_bit(pmap);
sys/amd64/amd64/pmap.c
9361
if ((*pte & (PG_M | PG_RW)) == (PG_M | PG_RW)) {
sys/amd64/amd64/pmap.c
9362
atomic_clear_long(pte, PG_M);
sys/amd64/amd64/pmap.c
9535
pt_entry_t PG_A, PG_M, PG_RW, PG_V;
sys/amd64/amd64/pmap.c
9540
PG_M = pmap_modified_bit(pmap);
sys/amd64/amd64/pmap.c
9564
newpdpe = pdpgpa | PG_M | PG_A | (oldpdpe & PG_U) | PG_RW | PG_V;
sys/amd64/amd64/pmap.c
9567
KASSERT((oldpdpe & (PG_M | PG_RW)) != PG_RW,
sys/amd64/include/pmap.h
85
PG_M | PG_U | PG_RW | PG_V | PG_PKU_MASK)
sys/amd64/vmm/vmm_instruction_emul.c
2153
if (!check_only && writable && (pte32 & PG_M) == 0) {
sys/amd64/vmm/vmm_instruction_emul.c
2155
pte32, pte32 | PG_M) == 0) {
sys/amd64/vmm/vmm_instruction_emul.c
2244
if (!check_only && writable && (pte & PG_M) == 0) {
sys/amd64/vmm/vmm_instruction_emul.c
2245
if (atomic_cmpset_64(&ptpbase[ptpindex], pte, pte | PG_M) == 0)
sys/i386/i386/minidump_machdep_base.c
287
fakept[k] = (pa + (k * PAGE_SIZE)) | PG_V | PG_RW | PG_A | PG_M;
sys/i386/i386/pmap.c
1511
*PMAP2 = newpf | PG_RW | PG_V | PG_A | PG_M;
sys/i386/i386/pmap.c
1573
*PMAP1 = newpf | PG_RW | PG_V | PG_A | PG_M;
sys/i386/i386/pmap.c
1607
*PMAP3 = newpf | PG_RW | PG_V | PG_A | PG_M;
sys/i386/i386/pmap.c
1637
*eh_ptep = pde | PG_RW | PG_V | PG_A | PG_M;
sys/i386/i386/pmap.c
197
#define pmap_pte_m(pte) ((*(int *)pte & PG_M) != 0)
sys/i386/i386/pmap.c
2150
(pd_entry_t)(ptepa | PG_U | PG_RW | PG_V | PG_A | PG_M);
sys/i386/i386/pmap.c
2268
newpdir = (pd_entry_t) (ptppaddr | PG_V | PG_RW | PG_A | PG_M);
sys/i386/i386/pmap.c
2369
if ((tpte & (PG_M | PG_RW)) == (PG_M | PG_RW))
sys/i386/i386/pmap.c
2808
*PMAP1 = mptepa | PG_RW | PG_V | PG_A | PG_M;
sys/i386/i386/pmap.c
2827
*PMAP2 = mptepa | PG_RW | PG_V | PG_A | PG_M;
sys/i386/i386/pmap.c
2833
newpde = mptepa | PG_M | PG_A | (oldpde & PG_U) | PG_RW | PG_V;
sys/i386/i386/pmap.c
2836
KASSERT((oldpde & (PG_M | PG_RW)) != PG_RW,
sys/i386/i386/pmap.c
2915
newpde = mptepa | PG_M | PG_A | PG_RW | PG_V;
sys/i386/i386/pmap.c
2971
if ((oldpde & (PG_M | PG_RW)) == (PG_M | PG_RW))
sys/i386/i386/pmap.c
3022
if ((oldpte & (PG_M | PG_RW)) == (PG_M | PG_RW))
sys/i386/i386/pmap.c
3245
if ((tpte & (PG_M | PG_RW)) == (PG_M | PG_RW))
sys/i386/i386/pmap.c
3276
if ((oldpde & (PG_MANAGED | PG_M | PG_RW)) ==
sys/i386/i386/pmap.c
3277
(PG_MANAGED | PG_M | PG_RW)) {
sys/i386/i386/pmap.c
3282
newpde &= ~(PG_RW | PG_M);
sys/i386/i386/pmap.c
3419
if ((pbits & (PG_MANAGED | PG_M | PG_RW)) ==
sys/i386/i386/pmap.c
3420
(PG_MANAGED | PG_M | PG_RW)) {
sys/i386/i386/pmap.c
3424
pbits &= ~(PG_RW | PG_M);
sys/i386/i386/pmap.c
3519
if ((newpde & (PG_M | PG_RW)) == PG_RW) {
sys/i386/i386/pmap.c
3548
if ((oldpte & (PG_M | PG_RW)) == PG_RW) {
sys/i386/i386/pmap.c
3673
newpte |= PG_M;
sys/i386/i386/pmap.c
3676
KASSERT((newpte & (PG_M | PG_RW)) != PG_M,
sys/i386/i386/pmap.c
3780
if (((origpte ^ newpte) & ~(PG_M | PG_A)) == 0)
sys/i386/i386/pmap.c
3806
if ((origpte & (PG_M | PG_RW)) == (PG_M | PG_RW))
sys/i386/i386/pmap.c
3860
if ((newpte & PG_M) == 0 && (origpte & (PG_M | PG_RW)) ==
sys/i386/i386/pmap.c
3861
(PG_M | PG_RW)) {
sys/i386/i386/pmap.c
3972
KASSERT((newpde & (PG_M | PG_RW)) != PG_RW,
sys/i386/i386/pmap.c
399
*pt = pa | PG_V | PG_RW | PG_A | PG_M;
sys/i386/i386/pmap.c
4345
pde_store(pde, pa | PG_PS | PG_M | PG_A |
sys/i386/i386/pmap.c
4544
*dst_pte = ptetemp & ~(PG_W | PG_M |
sys/i386/i386/pmap.c
4597
*cmap_pte2 = PG_V | PG_RW | VM_PAGE_TO_PHYS(m) | PG_A | PG_M |
sys/i386/i386/pmap.c
4628
*cmap_pte2 = PG_V | PG_RW | VM_PAGE_TO_PHYS(m) | PG_A | PG_M |
sys/i386/i386/pmap.c
4661
*cmap_pte2 = PG_V | PG_RW | VM_PAGE_TO_PHYS(dst) | PG_A | PG_M |
sys/i386/i386/pmap.c
470
IdlePTD[a] = (KPTphys + ptoa(a)) | PG_V | PG_RW | PG_A | PG_M;
sys/i386/i386/pmap.c
4702
PG_M | pmap_cache_bits(kernel_pmap, b_pg->md.pat_mode,
sys/i386/i386/pmap.c
4912
if ((tpte & (PG_M | PG_RW)) == (PG_M | PG_RW)) {
sys/i386/i386/pmap.c
5016
rv = (*pte & (PG_M | PG_RW)) == (PG_M | PG_RW);
sys/i386/i386/pmap.c
5145
oldpte & ~(PG_RW | PG_M)))
sys/i386/i386/pmap.c
5147
if ((oldpte & PG_M) != 0)
sys/i386/i386/pmap.c
519
*(pt_entry_t *)vm86pa = 0 | PG_RW | PG_U | PG_A | PG_M | PG_V;
sys/i386/i386/pmap.c
5199
if ((*pde & (PG_M | PG_RW)) == (PG_M | PG_RW)) {
sys/i386/i386/pmap.c
525
PG_M | PG_V;
sys/i386/i386/pmap.c
5255
if ((*pte & (PG_M | PG_RW)) == (PG_M | PG_RW))
sys/i386/i386/pmap.c
5361
else if ((*pte & (PG_M | PG_RW)) == (PG_M | PG_RW)) {
sys/i386/i386/pmap.c
537
IdlePTD[a >> PDRSHIFT] = a | PG_PS | PG_A | PG_M |
sys/i386/i386/pmap.c
5371
atomic_clear_int((u_int *)pte, PG_M | PG_A);
sys/i386/i386/pmap.c
5445
atomic_clear_int((u_int *)pte, PG_M | PG_RW);
sys/i386/i386/pmap.c
5459
if ((*pte & (PG_M | PG_RW)) == (PG_M | PG_RW)) {
sys/i386/i386/pmap.c
5465
atomic_clear_int((u_int *)pte, PG_M);
sys/i386/i386/pmap.c
5667
PG_A | PG_M | pmap_cache_bits(kernel_pmap, m->md.pat_mode,
sys/i386/i386/pmap.c
5848
if ((pte & (PG_M | PG_RW)) == (PG_M | PG_RW))
sys/i386/i386/pmap.c
5943
*pte = PG_V | PG_RW | VM_PAGE_TO_PHYS(m) | PG_A | PG_M |
sys/i386/i386/pmap.c
5993
PG_M | PG_A | PG_RW | PG_V | pgeflag |
sys/i386/i386/pmap.c
6012
PTD[TRPTDI] = VM_PAGE_TO_PHYS(pd_m) | PG_M | PG_A | PG_RW | PG_V |
sys/i386/i386/pmap.c
6114
*pte = PG_V | PG_RW | PG_A | PG_M | VM_PAGE_TO_PHYS(ma[i]) |
sys/i386/include/pmap.h
83
PG_M | PG_NC_PCD | PG_NC_PWT | PG_U | PG_RW | PG_V)
sys/powerpc/aim/mmu_radix.c
1006
newpte |= PG_M;
sys/powerpc/aim/mmu_radix.c
1464
if ((tpte & (PG_M | PG_RW)) == (PG_M | PG_RW))
sys/powerpc/aim/mmu_radix.c
1768
#define DMAP_PAGE_BITS (RPTE_VALID | RPTE_LEAF | RPTE_EAA_MASK | PG_M | PG_A)
sys/powerpc/aim/mmu_radix.c
2304
else if ((be64toh(*pte) & (PG_M | PG_RW)) == (PG_M | PG_RW)) {
sys/powerpc/aim/mmu_radix.c
2314
atomic_clear_long(pte, htobe64(PG_M | PG_A));
sys/powerpc/aim/mmu_radix.c
2481
htobe64((oldpte | RPTE_EAA_R) & ~(PG_M | PG_RW))))
sys/powerpc/aim/mmu_radix.c
2505
if ((be64toh(*pte) & (PG_M | PG_RW)) == (PG_M | PG_RW)) {
sys/powerpc/aim/mmu_radix.c
2506
atomic_clear_long(pte, htobe64(PG_M));
sys/powerpc/aim/mmu_radix.c
2631
*dst_pte = htobe64(ptetemp & ~(PG_W | PG_M |
sys/powerpc/aim/mmu_radix.c
2744
if ((newpde & (PG_M | PG_RW)) == PG_RW) {
sys/powerpc/aim/mmu_radix.c
2768
if ((oldpte & (PG_M | PG_RW)) == PG_RW) {
sys/powerpc/aim/mmu_radix.c
2854
newpte |= PG_M;
sys/powerpc/aim/mmu_radix.c
2861
KASSERT((newpte & (PG_M | PG_RW)) != PG_M,
sys/powerpc/aim/mmu_radix.c
2878
newpte |= PG_M;
sys/powerpc/aim/mmu_radix.c
2975
if (((origpte ^ newpte) & ~(PG_M | PG_A)) == 0) {
sys/powerpc/aim/mmu_radix.c
2976
if ((newpte & (PG_A|PG_M)) != (origpte & (PG_A|PG_M))) {
sys/powerpc/aim/mmu_radix.c
2979
if ((newpte & PG_M) != (origpte & PG_M))
sys/powerpc/aim/mmu_radix.c
2986
if (((origpte ^ newpte) & ~(PG_M | PG_A)) == 0)
sys/powerpc/aim/mmu_radix.c
3013
if ((origpte & (PG_M | PG_RW)) == (PG_M | PG_RW))
sys/powerpc/aim/mmu_radix.c
3087
if ((newpte & PG_M) == 0 && (origpte & (PG_M | PG_RW)) ==
sys/powerpc/aim/mmu_radix.c
3088
(PG_M | PG_RW)) {
sys/powerpc/aim/mmu_radix.c
3210
KASSERT((newpde & (PG_M | PG_RW)) != PG_RW,
sys/powerpc/aim/mmu_radix.c
3784
mask |= PG_RW | PG_M;
sys/powerpc/aim/mmu_radix.c
3811
mask |= PG_RW | PG_M;
sys/powerpc/aim/mmu_radix.c
3938
if ((oldl3e & (PG_M | PG_RW)) == (PG_M | PG_RW)) {
sys/powerpc/aim/mmu_radix.c
4011
if ((be64toh(*pte) & (PG_M | PG_RW)) == (PG_M | PG_RW))
sys/powerpc/aim/mmu_radix.c
4112
pa |= PG_M | PG_A | PG_RW;
sys/powerpc/aim/mmu_radix.c
4514
if ((oldpde & (PG_MANAGED | PG_M | PG_RW)) ==
sys/powerpc/aim/mmu_radix.c
4515
(PG_MANAGED | PG_M | PG_RW)) {
sys/powerpc/aim/mmu_radix.c
4522
newpde &= ~(PG_RW | PG_M);
sys/powerpc/aim/mmu_radix.c
4637
if ((pbits & (PG_MANAGED | PG_M | PG_RW)) ==
sys/powerpc/aim/mmu_radix.c
4638
(PG_MANAGED | PG_M | PG_RW)) {
sys/powerpc/aim/mmu_radix.c
4642
pbits &= ~(PG_RW | PG_M);
sys/powerpc/aim/mmu_radix.c
4651
if (obits & (PG_A|PG_M)) {
sys/powerpc/aim/mmu_radix.c
4678
attr_bits = RPTE_EAA_R | RPTE_EAA_W | RPTE_EAA_P | PG_M | PG_A;
sys/powerpc/aim/mmu_radix.c
4984
KASSERT((oldpde & (PG_M | PG_RW)) != PG_RW,
sys/powerpc/aim/mmu_radix.c
5093
if ((oldpde & (PG_M | PG_RW)) == (PG_M | PG_RW))
sys/powerpc/aim/mmu_radix.c
5135
if ((oldpte & (PG_M | PG_RW)) == (PG_M | PG_RW))
sys/powerpc/aim/mmu_radix.c
5404
if ((tpte & (PG_M | PG_RW)) == (PG_M | PG_RW))
sys/powerpc/aim/mmu_radix.c
5552
if ((tpte & (PG_M | PG_RW)) == (PG_M | PG_RW)) {
sys/powerpc/aim/mmu_radix.c
5688
htobe64((oldpte | RPTE_EAA_R) & ~(PG_RW | PG_M))))
sys/powerpc/aim/mmu_radix.c
5690
if ((oldpte & PG_M) != 0)
sys/powerpc/aim/mmu_radix.c
5837
if ((pte & (PG_M | PG_RW)) == (PG_M | PG_RW))
sys/powerpc/aim/mmu_radix.c
6042
KASSERT((oldpdpe & (PG_M | PG_RW)) != PG_RW,
sys/powerpc/aim/mmu_radix.c
6130
pteval = pa | RPTE_EAA_R | RPTE_EAA_W | RPTE_EAA_P | PG_M | PG_A;
sys/powerpc/aim/mmu_radix.c
864
PG_M | PG_A | RPTE_EAA_MASK | PG_V)
sys/powerpc/aim/mmu_radix.c
923
RPTE_EAA_W | RPTE_EAA_P | PG_M | PG_A);
sys/powerpc/aim/mmu_radix.c
970
if (((flags & VM_PROT_WRITE) && (startpte & PG_M)) ||