arch/alpha/include/asm/page.h
25
typedef struct { unsigned long pte; } pte_t;
arch/alpha/include/asm/page.h
30
#define pte_val(x) ((x).pte)
arch/alpha/include/asm/pgalloc.h
17
pmd_populate(struct mm_struct *mm, pmd_t *pmd, pgtable_t pte)
arch/alpha/include/asm/pgalloc.h
19
pmd_set(pmd, (pte_t *)(page_to_pa(pte) + PAGE_OFFSET));
arch/alpha/include/asm/pgalloc.h
23
pmd_populate_kernel(struct mm_struct *mm, pmd_t *pmd, pte_t *pte)
arch/alpha/include/asm/pgalloc.h
25
pmd_set(pmd, pte);
arch/alpha/include/asm/pgtable.h
170
#define pte_pfn(pte) (pte_val(pte) >> PFN_PTE_SHIFT)
arch/alpha/include/asm/pgtable.h
172
#define pte_page(pte) pfn_to_page(pte_pfn(pte))
arch/alpha/include/asm/pgtable.h
175
{ pte_t pte; pte_val(pte) = (PHYS_TWIDDLE(physpfn) << 32) | pgprot_val(pgprot); return pte; }
arch/alpha/include/asm/pgtable.h
177
extern inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
arch/alpha/include/asm/pgtable.h
178
{ pte_val(pte) = (pte_val(pte) & _PAGE_CHG_MASK) | pgprot_val(newprot); return pte; }
arch/alpha/include/asm/pgtable.h
205
extern inline int pte_none(pte_t pte) { return !pte_val(pte); }
arch/alpha/include/asm/pgtable.h
206
extern inline int pte_present(pte_t pte) { return pte_val(pte) & _PAGE_VALID; }
arch/alpha/include/asm/pgtable.h
226
extern inline int pte_write(pte_t pte) { return !(pte_val(pte) & _PAGE_FOW); }
arch/alpha/include/asm/pgtable.h
227
extern inline int pte_dirty(pte_t pte) { return pte_val(pte) & _PAGE_DIRTY; }
arch/alpha/include/asm/pgtable.h
228
extern inline int pte_young(pte_t pte) { return pte_val(pte) & _PAGE_ACCESSED; }
arch/alpha/include/asm/pgtable.h
230
extern inline pte_t pte_wrprotect(pte_t pte) { pte_val(pte) |= _PAGE_FOW; return pte; }
arch/alpha/include/asm/pgtable.h
231
extern inline pte_t pte_mkclean(pte_t pte) { pte_val(pte) &= ~(__DIRTY_BITS); return pte; }
arch/alpha/include/asm/pgtable.h
232
extern inline pte_t pte_mkold(pte_t pte) { pte_val(pte) &= ~(__ACCESS_BITS); return pte; }
arch/alpha/include/asm/pgtable.h
233
extern inline pte_t pte_mkwrite_novma(pte_t pte){ pte_val(pte) &= ~_PAGE_FOW; return pte; }
arch/alpha/include/asm/pgtable.h
234
extern inline pte_t pte_mkdirty(pte_t pte) { pte_val(pte) |= __DIRTY_BITS; return pte; }
arch/alpha/include/asm/pgtable.h
235
extern inline pte_t pte_mkyoung(pte_t pte) { pte_val(pte) |= __ACCESS_BITS; return pte; }
arch/alpha/include/asm/pgtable.h
278
pte_t pte = READ_ONCE(*ptep);
arch/alpha/include/asm/pgtable.h
281
return pte;
arch/alpha/include/asm/pgtable.h
290
pte_t pte = ptep_get_and_clear(mm, addr, ptep);
arch/alpha/include/asm/pgtable.h
292
page_table_check_pte_clear(mm, addr, pte);
arch/alpha/include/asm/pgtable.h
294
return pte;
arch/alpha/include/asm/pgtable.h
330
{ pte_t pte; pte_val(pte) = ((type & 0x7f) << 32) | (offset << 40); return pte; }
arch/alpha/include/asm/pgtable.h
335
#define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) })
arch/alpha/include/asm/pgtable.h
338
static inline bool pte_swp_exclusive(pte_t pte)
arch/alpha/include/asm/pgtable.h
340
return pte_val(pte) & _PAGE_SWP_EXCLUSIVE;
arch/alpha/include/asm/pgtable.h
343
static inline pte_t pte_swp_mkexclusive(pte_t pte)
arch/alpha/include/asm/pgtable.h
345
pte_val(pte) |= _PAGE_SWP_EXCLUSIVE;
arch/alpha/include/asm/pgtable.h
346
return pte;
arch/alpha/include/asm/pgtable.h
349
static inline pte_t pte_swp_clear_exclusive(pte_t pte)
arch/alpha/include/asm/pgtable.h
351
pte_val(pte) &= ~_PAGE_SWP_EXCLUSIVE;
arch/alpha/include/asm/pgtable.h
352
return pte;
arch/alpha/include/asm/tlb.h
7
#define __pte_free_tlb(tlb, pte, address) tlb_remove_ptdesc((tlb), page_ptdesc(pte))
arch/alpha/kernel/core_cia.c
330
unsigned long *ppte, pte;
arch/alpha/kernel/core_cia.c
335
pte = (virt_to_phys(ppte) >> (PAGE_SHIFT - 1)) | 1;
arch/alpha/kernel/core_cia.c
338
ppte[i] = pte;
arch/alpha/kernel/core_irongate.c
317
u32 *mmio_regs, *gatt_pages, *cur_gatt, pte;
arch/alpha/kernel/core_irongate.c
369
pte = cur_gatt[GET_GATT_OFF(baddr)] & ~1;
arch/alpha/kernel/core_irongate.c
371
cur_gatt, pte);
arch/alpha/kernel/core_irongate.c
386
pte = cur_gatt[GET_GATT_OFF(baddr)] & ~1;
arch/alpha/kernel/core_irongate.c
389
pte, PAGE_SIZE, 0)) {
arch/alpha/kernel/core_marvel.c
1000
pte = aper->arena->ptes[baddr >> PAGE_SHIFT];
arch/alpha/kernel/core_marvel.c
1001
if (!(pte & 1)) {
arch/alpha/kernel/core_marvel.c
1005
return (pte >> 1) << PAGE_SHIFT;
arch/alpha/kernel/core_marvel.c
992
unsigned long pte;
arch/alpha/kernel/core_titan.c
703
unsigned long pte;
arch/alpha/kernel/core_titan.c
711
pte = aper->arena->ptes[baddr >> PAGE_SHIFT];
arch/alpha/kernel/core_titan.c
712
if (!(pte & 1)) {
arch/alpha/kernel/core_titan.c
717
return (pte >> 1) << PAGE_SHIFT;
arch/arc/include/asm/arcregs.h
315
unsigned int pad:6, fbe:2, tqe:2, ts:4, ft:1, rse:2, pte:3, bce:3, ver:8;
arch/arc/include/asm/arcregs.h
317
unsigned int ver:8, bce:3, pte:3, rse:2, ft:1, ts:4, tqe:2, fbe:2, pad:6;
arch/arc/include/asm/hugepage.h
25
static inline pmd_t pte_pmd(pte_t pte)
arch/arc/include/asm/hugepage.h
27
return __pmd(pte_val(pte));
arch/arc/include/asm/page.h
70
unsigned long long pte;
arch/arc/include/asm/page.h
72
unsigned long pte;
arch/arc/include/asm/page.h
76
#define pte_val(x) ((x).pte)
arch/arc/include/asm/pgalloc.h
37
pmd_populate_kernel(struct mm_struct *mm, pmd_t *pmd, pte_t *pte)
arch/arc/include/asm/pgalloc.h
46
set_pmd(pmd, __pmd((unsigned long)pte));
arch/arc/include/asm/pgalloc.h
90
#define __pte_free_tlb(tlb, pte, addr) pte_free((tlb)->mm, pte)
arch/arc/include/asm/pgtable-bits-arcv2.h
100
return __pte((pte_val(pte) & _PAGE_CHG_MASK) | pgprot_val(newprot));
arch/arc/include/asm/pgtable-bits-arcv2.h
130
#define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) })
arch/arc/include/asm/pgtable-bits-arcv2.h
133
static inline bool pte_swp_exclusive(pte_t pte)
arch/arc/include/asm/pgtable-bits-arcv2.h
135
return pte_val(pte) & _PAGE_SWP_EXCLUSIVE;
arch/arc/include/asm/pgtable-bits-arcv2.h
80
#define pte_write(pte) (pte_val(pte) & _PAGE_WRITE)
arch/arc/include/asm/pgtable-bits-arcv2.h
81
#define pte_dirty(pte) (pte_val(pte) & _PAGE_DIRTY)
arch/arc/include/asm/pgtable-bits-arcv2.h
82
#define pte_young(pte) (pte_val(pte) & _PAGE_ACCESSED)
arch/arc/include/asm/pgtable-bits-arcv2.h
83
#define pte_special(pte) (pte_val(pte) & _PAGE_SPECIAL)
arch/arc/include/asm/pgtable-bits-arcv2.h
86
static inline pte_t pte_##fn(pte_t pte) { pte_val(pte) op; return pte; }
arch/arc/include/asm/pgtable-bits-arcv2.h
98
static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
arch/arc/include/asm/pgtable-levels.h
175
#define pte_page(pte) pfn_to_page(pte_pfn(pte))
arch/arc/include/asm/pgtable-levels.h
176
#define set_pte(ptep, pte) ((*(ptep)) = (pte))
arch/arc/include/asm/pgtable-levels.h
177
#define pte_pfn(pte) (pte_val(pte) >> PAGE_SHIFT)
arch/arc/kernel/setup.c
229
bpu_pred = 2048 << bpu.pte;
arch/arc/mm/tlb.c
533
pte_t pte = __pte(pmd_val(*pmd));
arch/arc/mm/tlb.c
534
update_mmu_cache_range(NULL, vma, addr, &pte, HPAGE_PMD_NR);
arch/arm/include/asm/kfence.h
15
pte_t *pte = pte_alloc_one_kernel(&init_mm);
arch/arm/include/asm/kfence.h
17
if (!pte)
arch/arm/include/asm/kfence.h
21
set_pte_ext(pte + i, pfn_pte(pfn + i, PAGE_KERNEL), 0);
arch/arm/include/asm/kfence.h
22
pmd_populate_kernel(&init_mm, pmd, pte);
arch/arm/include/asm/pgalloc.h
100
if (!pte)
arch/arm/include/asm/pgalloc.h
102
if (!PageHighMem(pte))
arch/arm/include/asm/pgalloc.h
103
clean_pte_table(page_address(pte));
arch/arm/include/asm/pgalloc.h
104
return pte;
arch/arm/include/asm/pgalloc.h
107
static inline void __pmd_populate(pmd_t *pmdp, phys_addr_t pte,
arch/arm/include/asm/pgalloc.h
110
pmdval_t pmdval = (pte + PTE_HWTABLE_OFF) | prot;
arch/arm/include/asm/pgalloc.h
41
#define pud_populate(mm,pmd,pte) do { } while (0)
arch/arm/include/asm/pgalloc.h
43
#define pud_populate(mm,pmd,pte) BUG()
arch/arm/include/asm/pgalloc.h
50
static inline void clean_pte_table(pte_t *pte)
arch/arm/include/asm/pgalloc.h
52
clean_dcache_area(pte + PTE_HWTABLE_PTRS, PTE_HWTABLE_SIZE);
arch/arm/include/asm/pgalloc.h
80
pte_t *pte = __pte_alloc_one_kernel(mm);
arch/arm/include/asm/pgalloc.h
82
if (pte)
arch/arm/include/asm/pgalloc.h
83
clean_pte_table(pte);
arch/arm/include/asm/pgalloc.h
85
return pte;
arch/arm/include/asm/pgalloc.h
97
struct page *pte;
arch/arm/include/asm/pgalloc.h
99
pte = __pte_alloc_one(mm, GFP_PGTABLE_USER | PGTABLE_HIGHMEM);
arch/arm/include/asm/pgtable-2level-types.h
21
typedef struct { pteval_t pte; } pte_t;
arch/arm/include/asm/pgtable-2level-types.h
26
#define pte_val(x) ((x).pte)
arch/arm/include/asm/pgtable-2level.h
237
#define set_pte_ext(ptep,pte,ext) cpu_set_pte_ext(ptep,pte,ext)
arch/arm/include/asm/pgtable-3level-types.h
24
typedef struct { pteval_t pte; } pte_t;
arch/arm/include/asm/pgtable-3level-types.h
29
#define pte_val(x) ((x).pte)
arch/arm/include/asm/pgtable-3level.h
171
#define set_pte_ext(ptep,pte,ext) cpu_set_pte_ext(ptep,__pte(pte_val(pte)|(ext)))
arch/arm/include/asm/pgtable-3level.h
173
#define pte_huge(pte) (pte_val(pte) && !(pte_val(pte) & PTE_TABLE_BIT))
arch/arm/include/asm/pgtable-3level.h
174
#define pte_mkhuge(pte) (__pte(pte_val(pte) & ~PTE_TABLE_BIT))
arch/arm/include/asm/pgtable-3level.h
182
#define pte_special(pte) (pte_isset((pte), L_PTE_SPECIAL))
arch/arm/include/asm/pgtable-3level.h
183
static inline pte_t pte_mkspecial(pte_t pte)
arch/arm/include/asm/pgtable-3level.h
185
pte_val(pte) |= L_PTE_SPECIAL;
arch/arm/include/asm/pgtable-3level.h
186
return pte;
arch/arm/include/asm/pgtable.h
167
#define pte_pfn(pte) ((pte_val(pte) & PHYS_MASK) >> PAGE_SHIFT)
arch/arm/include/asm/pgtable.h
170
#define pte_page(pte) pfn_to_page(pte_pfn(pte))
arch/arm/include/asm/pgtable.h
174
#define pte_isset(pte, val) ((u32)(val) == (val) ? pte_val(pte) & (val) \
arch/arm/include/asm/pgtable.h
175
: !!(pte_val(pte) & (val)))
arch/arm/include/asm/pgtable.h
176
#define pte_isclear(pte, val) (!(pte_val(pte) & (val)))
arch/arm/include/asm/pgtable.h
178
#define pte_none(pte) (!pte_val(pte))
arch/arm/include/asm/pgtable.h
179
#define pte_present(pte) (pte_isset((pte), L_PTE_PRESENT))
arch/arm/include/asm/pgtable.h
180
#define pte_valid(pte) (pte_isset((pte), L_PTE_VALID))
arch/arm/include/asm/pgtable.h
181
#define pte_accessible(mm, pte) (mm_tlb_flush_pending(mm) ? pte_present(pte) : pte_valid(pte))
arch/arm/include/asm/pgtable.h
182
#define pte_write(pte) (pte_isclear((pte), L_PTE_RDONLY))
arch/arm/include/asm/pgtable.h
183
#define pte_dirty(pte) (pte_isset((pte), L_PTE_DIRTY))
arch/arm/include/asm/pgtable.h
184
#define pte_young(pte) (pte_isset((pte), L_PTE_YOUNG))
arch/arm/include/asm/pgtable.h
185
#define pte_exec(pte) (pte_isclear((pte), L_PTE_XN))
arch/arm/include/asm/pgtable.h
187
#define pte_valid_user(pte) \
arch/arm/include/asm/pgtable.h
188
(pte_valid(pte) && pte_isset((pte), L_PTE_USER) && pte_young(pte))
arch/arm/include/asm/pgtable.h
190
static inline bool pte_access_permitted(pte_t pte, bool write)
arch/arm/include/asm/pgtable.h
198
return (pte_val(pte) & mask) == needed;
arch/arm/include/asm/pgtable.h
216
static inline pte_t clear_pte_bit(pte_t pte, pgprot_t prot)
arch/arm/include/asm/pgtable.h
218
pte_val(pte) &= ~pgprot_val(prot);
arch/arm/include/asm/pgtable.h
219
return pte;
arch/arm/include/asm/pgtable.h
222
static inline pte_t set_pte_bit(pte_t pte, pgprot_t prot)
arch/arm/include/asm/pgtable.h
224
pte_val(pte) |= pgprot_val(prot);
arch/arm/include/asm/pgtable.h
225
return pte;
arch/arm/include/asm/pgtable.h
228
static inline pte_t pte_wrprotect(pte_t pte)
arch/arm/include/asm/pgtable.h
230
return set_pte_bit(pte, __pgprot(L_PTE_RDONLY));
arch/arm/include/asm/pgtable.h
233
static inline pte_t pte_mkwrite_novma(pte_t pte)
arch/arm/include/asm/pgtable.h
235
return clear_pte_bit(pte, __pgprot(L_PTE_RDONLY));
arch/arm/include/asm/pgtable.h
238
static inline pte_t pte_mkclean(pte_t pte)
arch/arm/include/asm/pgtable.h
240
return clear_pte_bit(pte, __pgprot(L_PTE_DIRTY));
arch/arm/include/asm/pgtable.h
243
static inline pte_t pte_mkdirty(pte_t pte)
arch/arm/include/asm/pgtable.h
245
return set_pte_bit(pte, __pgprot(L_PTE_DIRTY));
arch/arm/include/asm/pgtable.h
248
static inline pte_t pte_mkold(pte_t pte)
arch/arm/include/asm/pgtable.h
250
return clear_pte_bit(pte, __pgprot(L_PTE_YOUNG));
arch/arm/include/asm/pgtable.h
253
static inline pte_t pte_mkyoung(pte_t pte)
arch/arm/include/asm/pgtable.h
255
return set_pte_bit(pte, __pgprot(L_PTE_YOUNG));
arch/arm/include/asm/pgtable.h
258
static inline pte_t pte_mkexec(pte_t pte)
arch/arm/include/asm/pgtable.h
260
return clear_pte_bit(pte, __pgprot(L_PTE_XN));
arch/arm/include/asm/pgtable.h
263
static inline pte_t pte_mknexec(pte_t pte)
arch/arm/include/asm/pgtable.h
265
return set_pte_bit(pte, __pgprot(L_PTE_XN));
arch/arm/include/asm/pgtable.h
268
static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
arch/arm/include/asm/pgtable.h
272
pte_val(pte) = (pte_val(pte) & ~mask) | (pgprot_val(newprot) & mask);
arch/arm/include/asm/pgtable.h
273
return pte;
arch/arm/include/asm/pgtable.h
301
#define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) })
arch/arm/include/asm/pgtable.h
304
static inline bool pte_swp_exclusive(pte_t pte)
arch/arm/include/asm/pgtable.h
306
return pte_isset(pte, L_PTE_SWP_EXCLUSIVE);
arch/arm/include/asm/pgtable.h
309
static inline pte_t pte_swp_mkexclusive(pte_t pte)
arch/arm/include/asm/pgtable.h
311
return set_pte_bit(pte, __pgprot(L_PTE_SWP_EXCLUSIVE));
arch/arm/include/asm/pgtable.h
314
static inline pte_t pte_swp_clear_exclusive(pte_t pte)
arch/arm/include/asm/pgtable.h
316
return clear_pte_bit(pte, __pgprot(L_PTE_SWP_EXCLUSIVE));
arch/arm/include/asm/pgtable.h
60
#define pte_ERROR(pte) __pte_error(__FILE__, __LINE__, pte)
arch/arm/include/asm/proc-fns.h
70
void (*set_pte_ext)(pte_t *ptep, pte_t pte);
arch/arm/include/asm/proc-fns.h
72
void (*set_pte_ext)(pte_t *ptep, pte_t pte, unsigned int ext);
arch/arm/include/asm/proc-fns.h
92
extern void cpu_set_pte_ext(pte_t *ptep, pte_t pte);
arch/arm/include/asm/proc-fns.h
94
extern void cpu_set_pte_ext(pte_t *ptep, pte_t pte, unsigned int ext);
arch/arm/include/asm/tlb.h
33
__pte_free_tlb(struct mmu_gather *tlb, pgtable_t pte, unsigned long addr)
arch/arm/include/asm/tlb.h
35
struct ptdesc *ptdesc = page_ptdesc(pte);
arch/arm/kernel/efi.c
17
pte_t pte = *ptep;
arch/arm/kernel/efi.c
20
pte = set_pte_bit(pte, __pgprot(L_PTE_RDONLY));
arch/arm/kernel/efi.c
22
pte = set_pte_bit(pte, __pgprot(L_PTE_XN));
arch/arm/kernel/efi.c
23
set_pte_ext(ptep, pte, PTE_EXT_NG);
arch/arm/kernel/traps.c
757
void __pte_error(const char *file, int line, pte_t pte)
arch/arm/kernel/traps.c
759
pr_err("%s:%d: bad pte %08llx.\n", file, line, (long long)pte_val(pte));
arch/arm/lib/uaccess_with_memcpy.c
104
pte_t *pte;
arch/arm/lib/uaccess_with_memcpy.c
108
while (!pin_page_for_write(to, &pte, &ptl)) {
arch/arm/lib/uaccess_with_memcpy.c
128
if (pte)
arch/arm/lib/uaccess_with_memcpy.c
129
pte_unmap_unlock(pte, ptl);
arch/arm/lib/uaccess_with_memcpy.c
168
pte_t *pte;
arch/arm/lib/uaccess_with_memcpy.c
172
while (!pin_page_for_write(addr, &pte, &ptl)) {
arch/arm/lib/uaccess_with_memcpy.c
189
if (pte)
arch/arm/lib/uaccess_with_memcpy.c
190
pte_unmap_unlock(pte, ptl);
arch/arm/lib/uaccess_with_memcpy.c
29
pte_t *pte;
arch/arm/lib/uaccess_with_memcpy.c
76
pte = pte_offset_map_lock(current->mm, pmd, addr, &ptl);
arch/arm/lib/uaccess_with_memcpy.c
77
if (unlikely(!pte))
arch/arm/lib/uaccess_with_memcpy.c
80
if (unlikely(!pte_present(*pte) || !pte_young(*pte) ||
arch/arm/lib/uaccess_with_memcpy.c
81
!pte_write(*pte) || !pte_dirty(*pte))) {
arch/arm/lib/uaccess_with_memcpy.c
82
pte_unmap_unlock(pte, ptl);
arch/arm/lib/uaccess_with_memcpy.c
86
*ptep = pte;
arch/arm/mm/dma-mapping.c
310
static int __dma_update_pte(pte_t *pte, unsigned long addr, void *data)
arch/arm/mm/dma-mapping.c
315
set_pte_ext(pte, mk_pte(page, prot), 0);
arch/arm/mm/dump.c
313
pte_t *pte = pte_offset_kernel(pmd, 0);
arch/arm/mm/dump.c
317
for (i = 0; i < PTRS_PER_PTE; i++, pte++) {
arch/arm/mm/dump.c
319
note_page(st, addr, 5, pte_val(*pte), domain);
arch/arm/mm/fault-armv.c
109
pte_unmap_unlock(pte, ptl);
arch/arm/mm/fault-armv.c
114
ret = do_adjust_pte(vma, address, pfn, pte);
arch/arm/mm/fault-armv.c
118
pte_unmap(pte);
arch/arm/mm/fault-armv.c
72
pte_t *pte;
arch/arm/mm/fault-armv.c
98
pte = pte_offset_map_rw_nolock(vma->vm_mm, pmd, address, &pmdval, &ptl);
arch/arm/mm/fault-armv.c
99
if (!pte)
arch/arm/mm/fault.c
101
(long long)pte_val(pte[PTE_HWTABLE_PTRS]));
arch/arm/mm/fault.c
103
pte_unmap(pte);
arch/arm/mm/fault.c
55
pte_t *pte;
arch/arm/mm/fault.c
94
pte = pte_offset_map(pmd, addr);
arch/arm/mm/fault.c
95
if (!pte)
arch/arm/mm/fault.c
98
pr_cont(", *pte=%08llx", (long long)pte_val(*pte));
arch/arm/mm/mm.h
25
static inline void set_top_pte(unsigned long va, pte_t pte)
arch/arm/mm/mm.h
28
set_pte_ext(ptep, pte, 0);
arch/arm/mm/mmu.c
104
.pte = L_PTE_MT_WRITEALLOC,
arch/arm/mm/mmu.c
1732
pte_t *pte;
arch/arm/mm/mmu.c
1736
pte = pte_offset_early_fixmap(pmd_off_k(map.virtual), map.virtual);
arch/arm/mm/mmu.c
1739
if (pte_none(*pte) ||
arch/arm/mm/mmu.c
1740
(pte_val(*pte) & L_PTE_MT_MASK) != L_PTE_MT_DEV_SHARED)
arch/arm/mm/mmu.c
1743
map.pfn = pte_pfn(*pte);
arch/arm/mm/mmu.c
400
pte_t *pte = pte_offset_fixmap(pmd_off_k(vaddr), vaddr);
arch/arm/mm/mmu.c
412
set_pte_at(NULL, vaddr, pte,
arch/arm/mm/mmu.c
415
pte_clear(NULL, vaddr, pte);
arch/arm/mm/mmu.c
570
vecs_pgprot = kern_pgprot = user_pgprot = cp->pte;
arch/arm/mm/mmu.c
748
pte_t *pte = alloc(PTE_HWTABLE_OFF + PTE_HWTABLE_SIZE);
arch/arm/mm/mmu.c
749
__pmd_populate(pmd, __pa(pte), prot);
arch/arm/mm/mmu.c
76
pteval_t pte;
arch/arm/mm/mmu.c
767
pte_t *pte = arm_pte_alloc(pmd, addr, type->prot_l1, alloc);
arch/arm/mm/mmu.c
769
set_pte_ext(pte, pfn_pte(pfn, __pgprot(type->prot_pte)),
arch/arm/mm/mmu.c
772
} while (pte++, addr += PAGE_SIZE, addr != end);
arch/arm/mm/mmu.c
84
.pte = L_PTE_MT_UNCACHED,
arch/arm/mm/mmu.c
89
.pte = L_PTE_MT_BUFFERABLE,
arch/arm/mm/mmu.c
94
.pte = L_PTE_MT_WRITETHROUGH,
arch/arm/mm/mmu.c
99
.pte = L_PTE_MT_WRITEBACK,
arch/arm/mm/pageattr.c
19
pte_t pte = *ptep;
arch/arm/mm/pageattr.c
21
pte = clear_pte_bit(pte, cdata->clear_mask);
arch/arm/mm/pageattr.c
22
pte = set_pte_bit(pte, cdata->set_mask);
arch/arm/mm/pageattr.c
24
set_pte_ext(ptep, pte, 0);
arch/arm/mm/pgd.c
146
pgtable_t pte;
arch/arm/mm/pgd.c
167
pte = pmd_pgtable(*pmd);
arch/arm/mm/pgd.c
169
pte_free(mm, pte);
arch/arm/mm/proc.c
113
void cpu_arm922_set_pte_ext(pte_t *ptep, pte_t pte, unsigned int ext);
arch/arm/mm/proc.c
130
void cpu_arm925_set_pte_ext(pte_t *ptep, pte_t pte, unsigned int ext);
arch/arm/mm/proc.c
147
void cpu_arm926_set_pte_ext(pte_t *ptep, pte_t pte, unsigned int ext);
arch/arm/mm/proc.c
200
void cpu_fa526_set_pte_ext(pte_t *ptep, pte_t pte, unsigned int ext);
arch/arm/mm/proc.c
217
void cpu_arm1020_set_pte_ext(pte_t *ptep, pte_t pte, unsigned int ext);
arch/arm/mm/proc.c
234
void cpu_arm1020e_set_pte_ext(pte_t *ptep, pte_t pte, unsigned int ext);
arch/arm/mm/proc.c
251
void cpu_arm1022_set_pte_ext(pte_t *ptep, pte_t pte, unsigned int ext);
arch/arm/mm/proc.c
268
void cpu_arm1026_set_pte_ext(pte_t *ptep, pte_t pte, unsigned int ext);
arch/arm/mm/proc.c
285
void cpu_sa110_set_pte_ext(pte_t *ptep, pte_t pte, unsigned int ext);
arch/arm/mm/proc.c
302
void cpu_sa1100_set_pte_ext(pte_t *ptep, pte_t pte, unsigned int ext);
arch/arm/mm/proc.c
325
void cpu_xscale_set_pte_ext(pte_t *ptep, pte_t pte, unsigned int ext);
arch/arm/mm/proc.c
348
void cpu_xsc3_set_pte_ext(pte_t *ptep, pte_t pte, unsigned int ext);
arch/arm/mm/proc.c
371
void cpu_mohawk_set_pte_ext(pte_t *ptep, pte_t pte, unsigned int ext);
arch/arm/mm/proc.c
394
void cpu_feroceon_set_pte_ext(pte_t *ptep, pte_t pte, unsigned int ext);
arch/arm/mm/proc.c
417
void cpu_v6_set_pte_ext(pte_t *ptep, pte_t pte, unsigned int ext);
arch/arm/mm/proc.c
43
void cpu_arm720_set_pte_ext(pte_t *ptep, pte_t pte, unsigned int ext);
arch/arm/mm/proc.c
450
void cpu_v7_set_pte_ext(pte_t *ptep, pte_t pte);
arch/arm/mm/proc.c
452
void cpu_v7_set_pte_ext(pte_t *ptep, pte_t pte, unsigned int ext);
arch/arm/mm/proc.c
488
void cpu_v7m_set_pte_ext(pte_t *ptep, pte_t pte, unsigned int ext);
arch/arm/mm/proc.c
90
void cpu_arm920_set_pte_ext(pte_t *ptep, pte_t pte, unsigned int ext);
arch/arm64/include/asm/assembler.h
625
.macro phys_to_pte, pte, phys
arch/arm64/include/asm/hugetlb.h
39
pte_t *ptep, pte_t pte, unsigned long sz);
arch/arm64/include/asm/hugetlb.h
43
pte_t pte, int dirty);
arch/arm64/include/asm/kvm_pgtable.h
111
static inline bool kvm_pte_valid(kvm_pte_t pte)
arch/arm64/include/asm/kvm_pgtable.h
113
return pte & KVM_PTE_VALID;
arch/arm64/include/asm/kvm_pgtable.h
116
static inline u64 kvm_pte_to_phys(kvm_pte_t pte)
arch/arm64/include/asm/kvm_pgtable.h
121
pa = pte & KVM_PTE_ADDR_MASK_LPA2;
arch/arm64/include/asm/kvm_pgtable.h
122
pa |= FIELD_GET(KVM_PTE_ADDR_51_50_LPA2, pte) << 50;
arch/arm64/include/asm/kvm_pgtable.h
124
pa = pte & KVM_PTE_ADDR_MASK;
arch/arm64/include/asm/kvm_pgtable.h
126
pa |= FIELD_GET(KVM_PTE_ADDR_51_48, pte) << 48;
arch/arm64/include/asm/kvm_pgtable.h
134
kvm_pte_t pte;
arch/arm64/include/asm/kvm_pgtable.h
137
pte = pa & KVM_PTE_ADDR_MASK_LPA2;
arch/arm64/include/asm/kvm_pgtable.h
139
pte |= FIELD_PREP(KVM_PTE_ADDR_51_50_LPA2, pa >> 50);
arch/arm64/include/asm/kvm_pgtable.h
141
pte = pa & KVM_PTE_ADDR_MASK;
arch/arm64/include/asm/kvm_pgtable.h
144
pte |= FIELD_PREP(KVM_PTE_ADDR_51_48, pa >> 48);
arch/arm64/include/asm/kvm_pgtable.h
148
return pte;
arch/arm64/include/asm/kvm_pgtable.h
151
static inline kvm_pfn_t kvm_pte_to_pfn(kvm_pte_t pte)
arch/arm64/include/asm/kvm_pgtable.h
153
return __phys_to_pfn(kvm_pte_to_phys(pte));
arch/arm64/include/asm/kvm_pgtable.h
861
enum kvm_pgtable_prot kvm_pgtable_stage2_pte_prot(kvm_pte_t pte);
arch/arm64/include/asm/kvm_pgtable.h
871
enum kvm_pgtable_prot kvm_pgtable_hyp_pte_prot(kvm_pte_t pte);
arch/arm64/include/asm/mte.h
131
static inline void mte_sync_tags(pte_t pte, unsigned int nr_pages)
arch/arm64/include/asm/mte.h
99
void mte_sync_tags(pte_t pte, unsigned int nr_pages);
arch/arm64/include/asm/pgtable-prot.h
132
#define pte_pi_index(pte) ( \
arch/arm64/include/asm/pgtable-prot.h
133
((pte & BIT(PTE_PI_IDX_3)) >> (PTE_PI_IDX_3 - 3)) | \
arch/arm64/include/asm/pgtable-prot.h
134
((pte & BIT(PTE_PI_IDX_2)) >> (PTE_PI_IDX_2 - 2)) | \
arch/arm64/include/asm/pgtable-prot.h
135
((pte & BIT(PTE_PI_IDX_1)) >> (PTE_PI_IDX_1 - 1)) | \
arch/arm64/include/asm/pgtable-prot.h
136
((pte & BIT(PTE_PI_IDX_0)) >> (PTE_PI_IDX_0 - 0)))
arch/arm64/include/asm/pgtable-types.h
31
typedef struct { pteval_t pte; } pte_t;
arch/arm64/include/asm/pgtable-types.h
32
#define pte_val(x) ((x).pte)
arch/arm64/include/asm/pgtable.h
120
static inline phys_addr_t __pte_to_phys(pte_t pte)
arch/arm64/include/asm/pgtable.h
122
pte_val(pte) &= ~PTE_MAYBE_SHARED;
arch/arm64/include/asm/pgtable.h
1221
static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
arch/arm64/include/asm/pgtable.h
123
return (pte_val(pte) & PTE_ADDR_LOW) |
arch/arm64/include/asm/pgtable.h
1232
if (pte_hw_dirty(pte))
arch/arm64/include/asm/pgtable.h
1233
pte = set_pte_bit(pte, __pgprot(PTE_DIRTY));
arch/arm64/include/asm/pgtable.h
1235
pte_val(pte) = (pte_val(pte) & ~mask) | (pgprot_val(newprot) & mask);
arch/arm64/include/asm/pgtable.h
124
((pte_val(pte) & PTE_ADDR_HIGH) << PTE_ADDR_HIGH_SHIFT);
arch/arm64/include/asm/pgtable.h
1240
if (pte_sw_dirty(pte))
arch/arm64/include/asm/pgtable.h
1241
pte = pte_mkdirty(pte);
arch/arm64/include/asm/pgtable.h
1242
return pte;
arch/arm64/include/asm/pgtable.h
1266
static inline bool pte_user_accessible_page(pte_t pte, unsigned long addr)
arch/arm64/include/asm/pgtable.h
1268
return pte_valid(pte) && (pte_user(pte) || pte_user_exec(pte));
arch/arm64/include/asm/pgtable.h
1296
pte_t old_pte, pte;
arch/arm64/include/asm/pgtable.h
1298
pte = __ptep_get(ptep);
arch/arm64/include/asm/pgtable.h
1300
old_pte = pte;
arch/arm64/include/asm/pgtable.h
1301
pte = pte_mkold(pte);
arch/arm64/include/asm/pgtable.h
1302
pte_val(pte) = cmpxchg_relaxed(&pte_val(*ptep),
arch/arm64/include/asm/pgtable.h
1303
pte_val(old_pte), pte_val(pte));
arch/arm64/include/asm/pgtable.h
1304
} while (pte_val(pte) != pte_val(old_pte));
arch/arm64/include/asm/pgtable.h
1306
return pte_young(pte);
arch/arm64/include/asm/pgtable.h
131
static inline phys_addr_t __pte_to_phys(pte_t pte)
arch/arm64/include/asm/pgtable.h
133
return pte_val(pte) & PTE_ADDR_LOW;
arch/arm64/include/asm/pgtable.h
1346
pte_t pte = __pte(xchg_relaxed(&pte_val(*ptep), 0));
arch/arm64/include/asm/pgtable.h
1350
page_table_check_pte_clear(mm, address, pte);
arch/arm64/include/asm/pgtable.h
1353
page_table_check_pmd_clear(mm, address, pte_pmd(pte));
arch/arm64/include/asm/pgtable.h
1357
page_table_check_pud_clear(mm, address, pte_pud(pte));
arch/arm64/include/asm/pgtable.h
1364
return pte;
arch/arm64/include/asm/pgtable.h
1389
pte_t pte, tmp_pte;
arch/arm64/include/asm/pgtable.h
1391
pte = __ptep_get_and_clear(mm, addr, ptep);
arch/arm64/include/asm/pgtable.h
1397
pte = pte_mkdirty(pte);
arch/arm64/include/asm/pgtable.h
1399
pte = pte_mkyoung(pte);
arch/arm64/include/asm/pgtable.h
1401
return pte;
arch/arm64/include/asm/pgtable.h
1415
pte_t pte)
arch/arm64/include/asm/pgtable.h
142
#define pte_pfn(pte) (__pte_to_phys(pte) >> PAGE_SHIFT)
arch/arm64/include/asm/pgtable.h
1420
old_pte = pte;
arch/arm64/include/asm/pgtable.h
1421
pte = pte_wrprotect(pte);
arch/arm64/include/asm/pgtable.h
1422
pte_val(pte) = cmpxchg_relaxed(&pte_val(*ptep),
arch/arm64/include/asm/pgtable.h
1423
pte_val(old_pte), pte_val(pte));
arch/arm64/include/asm/pgtable.h
1424
} while (pte_val(pte) != pte_val(old_pte));
arch/arm64/include/asm/pgtable.h
1448
pte_t pte, cydp_t flags)
arch/arm64/include/asm/pgtable.h
1453
old_pte = pte;
arch/arm64/include/asm/pgtable.h
1456
pte = pte_mkold(pte);
arch/arm64/include/asm/pgtable.h
1458
pte = pte_mkclean(pte);
arch/arm64/include/asm/pgtable.h
146
#define pte_none(pte) (!pte_val(pte))
arch/arm64/include/asm/pgtable.h
1460
pte_val(pte) = cmpxchg_relaxed(&pte_val(*ptep),
arch/arm64/include/asm/pgtable.h
1461
pte_val(old_pte), pte_val(pte));
arch/arm64/include/asm/pgtable.h
1462
} while (pte_val(pte) != pte_val(old_pte));
arch/arm64/include/asm/pgtable.h
1469
pte_t pte;
arch/arm64/include/asm/pgtable.h
147
#define pte_page(pte) (pfn_to_page(pte_pfn(pte)))
arch/arm64/include/asm/pgtable.h
1472
pte = __ptep_get(ptep);
arch/arm64/include/asm/pgtable.h
1475
__set_pte(ptep, pte_mkclean(pte_mkold(pte)));
arch/arm64/include/asm/pgtable.h
1477
__clear_young_dirty_pte(vma, addr, ptep, pte, flags);
arch/arm64/include/asm/pgtable.h
152
#define pte_present(pte) (pte_valid(pte) || pte_present_invalid(pte))
arch/arm64/include/asm/pgtable.h
1523
#define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) })
arch/arm64/include/asm/pgtable.h
153
#define pte_young(pte) (!!(pte_val(pte) & PTE_AF))
arch/arm64/include/asm/pgtable.h
154
#define pte_special(pte) (!!(pte_val(pte) & PTE_SPECIAL))
arch/arm64/include/asm/pgtable.h
155
#define pte_write(pte) (!!(pte_val(pte) & PTE_WRITE))
arch/arm64/include/asm/pgtable.h
156
#define pte_rdonly(pte) (!!(pte_val(pte) & PTE_RDONLY))
arch/arm64/include/asm/pgtable.h
157
#define pte_user(pte) (!!(pte_val(pte) & PTE_USER))
arch/arm64/include/asm/pgtable.h
158
#define pte_user_exec(pte) (!(pte_val(pte) & PTE_UXN))
arch/arm64/include/asm/pgtable.h
159
#define pte_cont(pte) (!!(pte_val(pte) & PTE_CONT))
arch/arm64/include/asm/pgtable.h
160
#define pte_tagged(pte) ((pte_val(pte) & PTE_ATTRINDX_MASK) == \
arch/arm64/include/asm/pgtable.h
1633
pte_t *ptep, pte_t old_pte, pte_t pte,
arch/arm64/include/asm/pgtable.h
1644
pte_t *ptep, pte_t pte);
arch/arm64/include/asm/pgtable.h
1646
pte_t *ptep, pte_t pte);
arch/arm64/include/asm/pgtable.h
1650
pte_t *ptep, pte_t pte, unsigned int nr);
arch/arm64/include/asm/pgtable.h
1670
unsigned long addr, pte_t *ptep, pte_t pte)
arch/arm64/include/asm/pgtable.h
1684
bool palign = (pte_pfn(pte) & contmask) == contmask;
arch/arm64/include/asm/pgtable.h
1687
pte_valid(pte) && !pte_cont(pte) && !pte_special(pte)))
arch/arm64/include/asm/pgtable.h
1688
__contpte_try_fold(mm, addr, ptep, pte);
arch/arm64/include/asm/pgtable.h
1693
unsigned long addr, pte_t *ptep, pte_t pte)
arch/arm64/include/asm/pgtable.h
1695
if (unlikely(pte_valid_cont(pte)))
arch/arm64/include/asm/pgtable.h
1696
__contpte_try_unfold(mm, addr, ptep, pte);
arch/arm64/include/asm/pgtable.h
1700
static inline unsigned int pte_batch_hint(pte_t *ptep, pte_t pte)
arch/arm64/include/asm/pgtable.h
1702
if (!pte_valid_cont(pte))
arch/arm64/include/asm/pgtable.h
1729
pte_t pte = __ptep_get(ptep);
arch/arm64/include/asm/pgtable.h
173
#define pte_hw_dirty(pte) (pte_write(pte) && !pte_rdonly(pte))
arch/arm64/include/asm/pgtable.h
1731
if (likely(!pte_valid_cont(pte)))
arch/arm64/include/asm/pgtable.h
1732
return pte;
arch/arm64/include/asm/pgtable.h
1734
return contpte_ptep_get(ptep, pte);
arch/arm64/include/asm/pgtable.h
174
#define pte_sw_dirty(pte) (!!(pte_val(pte) & PTE_DIRTY))
arch/arm64/include/asm/pgtable.h
1740
pte_t pte = __ptep_get(ptep);
arch/arm64/include/asm/pgtable.h
1742
if (likely(!pte_valid_cont(pte)))
arch/arm64/include/asm/pgtable.h
1743
return pte;
arch/arm64/include/asm/pgtable.h
1748
static inline void set_pte(pte_t *ptep, pte_t pte)
arch/arm64/include/asm/pgtable.h
175
#define pte_dirty(pte) (pte_sw_dirty(pte) || pte_hw_dirty(pte))
arch/arm64/include/asm/pgtable.h
1760
__set_pte(ptep, pte_mknoncont(pte));
arch/arm64/include/asm/pgtable.h
1765
pte_t *ptep, pte_t pte, unsigned int nr)
arch/arm64/include/asm/pgtable.h
1767
pte = pte_mknoncont(pte);
arch/arm64/include/asm/pgtable.h
177
#define pte_valid(pte) (!!(pte_val(pte) & PTE_VALID))
arch/arm64/include/asm/pgtable.h
1771
__set_ptes(mm, addr, ptep, pte, 1);
arch/arm64/include/asm/pgtable.h
1772
contpte_try_fold(mm, addr, ptep, pte);
arch/arm64/include/asm/pgtable.h
1774
contpte_set_ptes(mm, addr, ptep, pte, nr);
arch/arm64/include/asm/pgtable.h
178
#define pte_present_invalid(pte) \
arch/arm64/include/asm/pgtable.h
179
((pte_val(pte) & (PTE_VALID | PTE_PRESENT_INVALID)) == PTE_PRESENT_INVALID)
arch/arm64/include/asm/pgtable.h
1802
pte_t pte;
arch/arm64/include/asm/pgtable.h
1806
pte = __get_and_clear_full_ptes(mm, addr, ptep, nr, full);
arch/arm64/include/asm/pgtable.h
1808
pte = contpte_get_and_clear_full_ptes(mm, addr, ptep, nr, full);
arch/arm64/include/asm/pgtable.h
1811
return pte;
arch/arm64/include/asm/pgtable.h
184
#define pte_valid_not_user(pte) \
arch/arm64/include/asm/pgtable.h
185
((pte_val(pte) & (PTE_VALID | PTE_USER | PTE_UXN)) == (PTE_VALID | PTE_UXN))
arch/arm64/include/asm/pgtable.h
189
#define pte_valid_cont(pte) (pte_valid(pte) && pte_cont(pte))
arch/arm64/include/asm/pgtable.h
198
#define pte_accessible(mm, pte) \
arch/arm64/include/asm/pgtable.h
199
(mm_tlb_flush_pending(mm) ? pte_present(pte) : pte_valid(pte))
arch/arm64/include/asm/pgtable.h
226
#define pte_access_permitted_no_overlay(pte, write) \
arch/arm64/include/asm/pgtable.h
227
(((pte_val(pte) & (PTE_VALID | PTE_USER)) == (PTE_VALID | PTE_USER)) && (!(write) || pte_write(pte)))
arch/arm64/include/asm/pgtable.h
228
#define pte_access_permitted(pte, write) \
arch/arm64/include/asm/pgtable.h
229
(pte_access_permitted_no_overlay(pte, write) && \
arch/arm64/include/asm/pgtable.h
230
por_el0_allows_pkey(FIELD_GET(PTE_PO_IDX_MASK, pte_val(pte)), write, false))
arch/arm64/include/asm/pgtable.h
236
static inline pte_t clear_pte_bit(pte_t pte, pgprot_t prot)
arch/arm64/include/asm/pgtable.h
238
pte_val(pte) &= ~pgprot_val(prot);
arch/arm64/include/asm/pgtable.h
239
return pte;
arch/arm64/include/asm/pgtable.h
242
static inline pte_t set_pte_bit(pte_t pte, pgprot_t prot)
arch/arm64/include/asm/pgtable.h
244
pte_val(pte) |= pgprot_val(prot);
arch/arm64/include/asm/pgtable.h
245
return pte;
arch/arm64/include/asm/pgtable.h
260
static inline pte_t pte_mkwrite_novma(pte_t pte)
arch/arm64/include/asm/pgtable.h
262
pte = set_pte_bit(pte, __pgprot(PTE_WRITE));
arch/arm64/include/asm/pgtable.h
263
if (pte_sw_dirty(pte))
arch/arm64/include/asm/pgtable.h
264
pte = clear_pte_bit(pte, __pgprot(PTE_RDONLY));
arch/arm64/include/asm/pgtable.h
265
return pte;
arch/arm64/include/asm/pgtable.h
268
static inline pte_t pte_mkclean(pte_t pte)
arch/arm64/include/asm/pgtable.h
270
pte = clear_pte_bit(pte, __pgprot(PTE_DIRTY));
arch/arm64/include/asm/pgtable.h
271
pte = set_pte_bit(pte, __pgprot(PTE_RDONLY));
arch/arm64/include/asm/pgtable.h
273
return pte;
arch/arm64/include/asm/pgtable.h
276
static inline pte_t pte_mkdirty(pte_t pte)
arch/arm64/include/asm/pgtable.h
278
pte = set_pte_bit(pte, __pgprot(PTE_DIRTY));
arch/arm64/include/asm/pgtable.h
280
if (pte_write(pte))
arch/arm64/include/asm/pgtable.h
281
pte = clear_pte_bit(pte, __pgprot(PTE_RDONLY));
arch/arm64/include/asm/pgtable.h
283
return pte;
arch/arm64/include/asm/pgtable.h
286
static inline pte_t pte_wrprotect(pte_t pte)
arch/arm64/include/asm/pgtable.h
292
if (pte_hw_dirty(pte))
arch/arm64/include/asm/pgtable.h
293
pte = set_pte_bit(pte, __pgprot(PTE_DIRTY));
arch/arm64/include/asm/pgtable.h
295
pte = clear_pte_bit(pte, __pgprot(PTE_WRITE));
arch/arm64/include/asm/pgtable.h
296
pte = set_pte_bit(pte, __pgprot(PTE_RDONLY));
arch/arm64/include/asm/pgtable.h
297
return pte;
arch/arm64/include/asm/pgtable.h
300
static inline pte_t pte_mkold(pte_t pte)
arch/arm64/include/asm/pgtable.h
302
return clear_pte_bit(pte, __pgprot(PTE_AF));
arch/arm64/include/asm/pgtable.h
305
static inline pte_t pte_mkyoung(pte_t pte)
arch/arm64/include/asm/pgtable.h
307
return set_pte_bit(pte, __pgprot(PTE_AF));
arch/arm64/include/asm/pgtable.h
310
static inline pte_t pte_mkspecial(pte_t pte)
arch/arm64/include/asm/pgtable.h
312
return set_pte_bit(pte, __pgprot(PTE_SPECIAL));
arch/arm64/include/asm/pgtable.h
315
static inline pte_t pte_mkcont(pte_t pte)
arch/arm64/include/asm/pgtable.h
317
return set_pte_bit(pte, __pgprot(PTE_CONT));
arch/arm64/include/asm/pgtable.h
320
static inline pte_t pte_mknoncont(pte_t pte)
arch/arm64/include/asm/pgtable.h
322
return clear_pte_bit(pte, __pgprot(PTE_CONT));
arch/arm64/include/asm/pgtable.h
325
static inline pte_t pte_mkvalid(pte_t pte)
arch/arm64/include/asm/pgtable.h
327
return set_pte_bit(pte, __pgprot(PTE_VALID));
arch/arm64/include/asm/pgtable.h
330
static inline pte_t pte_mkinvalid(pte_t pte)
arch/arm64/include/asm/pgtable.h
332
pte = set_pte_bit(pte, __pgprot(PTE_PRESENT_INVALID));
arch/arm64/include/asm/pgtable.h
333
pte = clear_pte_bit(pte, __pgprot(PTE_VALID));
arch/arm64/include/asm/pgtable.h
334
return pte;
arch/arm64/include/asm/pgtable.h
348
static inline int pte_uffd_wp(pte_t pte)
arch/arm64/include/asm/pgtable.h
350
return !!(pte_val(pte) & PTE_UFFD_WP);
arch/arm64/include/asm/pgtable.h
353
static inline pte_t pte_mkuffd_wp(pte_t pte)
arch/arm64/include/asm/pgtable.h
355
return pte_wrprotect(set_pte_bit(pte, __pgprot(PTE_UFFD_WP)));
arch/arm64/include/asm/pgtable.h
358
static inline pte_t pte_clear_uffd_wp(pte_t pte)
arch/arm64/include/asm/pgtable.h
360
return clear_pte_bit(pte, __pgprot(PTE_UFFD_WP));
arch/arm64/include/asm/pgtable.h
364
static inline void __set_pte_nosync(pte_t *ptep, pte_t pte)
arch/arm64/include/asm/pgtable.h
366
WRITE_ONCE(*ptep, pte);
arch/arm64/include/asm/pgtable.h
369
static inline void __set_pte_complete(pte_t pte)
arch/arm64/include/asm/pgtable.h
375
if (pte_valid_not_user(pte))
arch/arm64/include/asm/pgtable.h
379
static inline void __set_pte(pte_t *ptep, pte_t pte)
arch/arm64/include/asm/pgtable.h
381
__set_pte_nosync(ptep, pte);
arch/arm64/include/asm/pgtable.h
382
__set_pte_complete(pte);
arch/arm64/include/asm/pgtable.h
410
pte_t pte)
arch/arm64/include/asm/pgtable.h
419
if (!pte_valid(old_pte) || !pte_valid(pte))
arch/arm64/include/asm/pgtable.h
429
VM_WARN_ONCE(!pte_young(pte),
arch/arm64/include/asm/pgtable.h
431
__func__, pte_val(old_pte), pte_val(pte));
arch/arm64/include/asm/pgtable.h
432
VM_WARN_ONCE(pte_write(old_pte) && !pte_dirty(pte),
arch/arm64/include/asm/pgtable.h
434
__func__, pte_val(old_pte), pte_val(pte));
arch/arm64/include/asm/pgtable.h
435
VM_WARN_ONCE(!pgattr_change_is_safe(pte_val(old_pte), pte_val(pte)),
arch/arm64/include/asm/pgtable.h
437
__func__, pte_val(old_pte), pte_val(pte));
arch/arm64/include/asm/pgtable.h
440
static inline void __sync_cache_and_tags(pte_t pte, unsigned int nr_pages)
arch/arm64/include/asm/pgtable.h
442
if (pte_present(pte) && pte_user_exec(pte) && !pte_special(pte))
arch/arm64/include/asm/pgtable.h
443
__sync_icache_dcache(pte);
arch/arm64/include/asm/pgtable.h
452
if (system_supports_mte() && pte_access_permitted_no_overlay(pte, false) &&
arch/arm64/include/asm/pgtable.h
453
!pte_special(pte) && pte_tagged(pte))
arch/arm64/include/asm/pgtable.h
454
mte_sync_tags(pte, nr_pages);
arch/arm64/include/asm/pgtable.h
461
static inline pgprot_t pte_pgprot(pte_t pte)
arch/arm64/include/asm/pgtable.h
463
unsigned long pfn = pte_pfn(pte);
arch/arm64/include/asm/pgtable.h
465
return __pgprot(pte_val(pfn_pte(pfn, __pgprot(0))) ^ pte_val(pte));
arch/arm64/include/asm/pgtable.h
469
static inline pte_t pte_advance_pfn(pte_t pte, unsigned long nr)
arch/arm64/include/asm/pgtable.h
471
return pfn_pte(pte_pfn(pte) + nr, pte_pgprot(pte));
arch/arm64/include/asm/pgtable.h
498
static inline pud_t pte_pud(pte_t pte)
arch/arm64/include/asm/pgtable.h
500
return __pud(pte_val(pte));
arch/arm64/include/asm/pgtable.h
513
static inline pmd_t pte_pmd(pte_t pte)
arch/arm64/include/asm/pgtable.h
515
return __pmd(pte_val(pte));
arch/arm64/include/asm/pgtable.h
528
static inline pte_t pte_swp_mkexclusive(pte_t pte)
arch/arm64/include/asm/pgtable.h
530
return set_pte_bit(pte, __pgprot(PTE_SWP_EXCLUSIVE));
arch/arm64/include/asm/pgtable.h
533
static inline bool pte_swp_exclusive(pte_t pte)
arch/arm64/include/asm/pgtable.h
535
return pte_val(pte) & PTE_SWP_EXCLUSIVE;
arch/arm64/include/asm/pgtable.h
538
static inline pte_t pte_swp_clear_exclusive(pte_t pte)
arch/arm64/include/asm/pgtable.h
540
return clear_pte_bit(pte, __pgprot(PTE_SWP_EXCLUSIVE));
arch/arm64/include/asm/pgtable.h
544
static inline pte_t pte_swp_mkuffd_wp(pte_t pte)
arch/arm64/include/asm/pgtable.h
546
return set_pte_bit(pte, __pgprot(PTE_SWP_UFFD_WP));
arch/arm64/include/asm/pgtable.h
549
static inline int pte_swp_uffd_wp(pte_t pte)
arch/arm64/include/asm/pgtable.h
551
return !!(pte_val(pte) & PTE_SWP_UFFD_WP);
arch/arm64/include/asm/pgtable.h
554
static inline pte_t pte_swp_clear_uffd_wp(pte_t pte)
arch/arm64/include/asm/pgtable.h
556
return clear_pte_bit(pte, __pgprot(PTE_SWP_UFFD_WP));
arch/arm64/include/asm/pgtable.h
564
static inline int pte_protnone(pte_t pte)
arch/arm64/include/asm/pgtable.h
575
return pte_present_invalid(pte) && !pte_user(pte) && !pte_user_exec(pte);
arch/arm64/include/asm/pgtable.h
624
#define pmd_special(pte) (!!((pmd_val(pte) & PTE_SPECIAL)))
arch/arm64/include/asm/pgtable.h
675
pte_t *ptep, pte_t pte, unsigned int nr,
arch/arm64/include/asm/pgtable.h
682
page_table_check_ptes_set(mm, addr, ptep, pte, nr);
arch/arm64/include/asm/pgtable.h
686
pte_pmd(pte), nr);
arch/arm64/include/asm/pgtable.h
691
pte_pud(pte), nr);
arch/arm64/include/asm/pgtable.h
698
__sync_cache_and_tags(pte, nr * stride);
arch/arm64/include/asm/pgtable.h
701
__check_safe_pte_update(mm, ptep, pte);
arch/arm64/include/asm/pgtable.h
702
__set_pte_nosync(ptep, pte);
arch/arm64/include/asm/pgtable.h
706
pte = pte_advance_pfn(pte, stride);
arch/arm64/include/asm/pgtable.h
709
__set_pte_complete(pte);
arch/arm64/include/asm/pgtable.h
713
pte_t *ptep, pte_t pte, unsigned int nr)
arch/arm64/include/asm/pgtable.h
715
__set_ptes_anysz(mm, addr, ptep, pte, nr, PAGE_SIZE);
arch/arm64/include/asm/pgtable.h
788
#define pte_leaf_size(pte) (pte_cont(pte) ? CONT_PTE_SIZE : PAGE_SIZE)
arch/arm64/include/asm/ptdump.h
65
void note_page_pte(struct ptdump_state *st, unsigned long addr, pte_t pte);
arch/arm64/include/asm/ptdump.h
81
static inline void note_page_pte(struct ptdump_state *st, unsigned long addr, pte_t pte) { }
arch/arm64/include/asm/tlb.h
75
static inline void __pte_free_tlb(struct mmu_gather *tlb, pgtable_t pte,
arch/arm64/include/asm/tlb.h
78
struct ptdesc *ptdesc = page_ptdesc(pte);
arch/arm64/kernel/efi.c
118
pte_t pte = __ptep_get(ptep);
arch/arm64/kernel/efi.c
121
pte = set_pte_bit(pte, __pgprot(PTE_RDONLY));
arch/arm64/kernel/efi.c
123
pte = set_pte_bit(pte, __pgprot(PTE_PXN));
arch/arm64/kernel/efi.c
125
pte = set_pte_bit(pte, __pgprot(PTE_GP));
arch/arm64/kernel/efi.c
126
__set_pte(ptep, pte);
arch/arm64/kernel/mte.c
38
void mte_sync_tags(pte_t pte, unsigned int nr_pages)
arch/arm64/kernel/mte.c
40
struct page *page = pte_page(pte);
arch/arm64/kernel/pi/map_range.c
29
void __init map_range(phys_addr_t *pte, u64 start, u64 end, phys_addr_t pa,
arch/arm64/kernel/pi/map_range.c
60
*tbl = __pte(__phys_to_pte_val(*pte) |
arch/arm64/kernel/pi/map_range.c
62
*pte += PTRS_PER_PTE * sizeof(pte_t);
arch/arm64/kernel/pi/map_range.c
64
map_range(pte, start, next, pa, prot, level + 1,
arch/arm64/kernel/pi/pi.h
32
void map_range(phys_addr_t *pte, u64 start, u64 end, phys_addr_t pa,
arch/arm64/kvm/hyp/nvhe/mem_protect.c
1033
kvm_pte_t pte;
arch/arm64/kvm/hyp/nvhe/mem_protect.c
1038
ret = kvm_pgtable_get_leaf(&vm->pgt, ipa, &pte, &level);
arch/arm64/kvm/hyp/nvhe/mem_protect.c
1041
if (!kvm_pte_valid(pte))
arch/arm64/kvm/hyp/nvhe/mem_protect.c
1049
state = guest_get_page_state(pte, ipa);
arch/arm64/kvm/hyp/nvhe/mem_protect.c
1053
phys = kvm_pte_to_phys(pte);
arch/arm64/kvm/hyp/nvhe/mem_protect.c
496
kvm_pte_t pte;
arch/arm64/kvm/hyp/nvhe/mem_protect.c
502
ret = kvm_pgtable_get_leaf(&host_mmu.pgt, addr, &pte, &level);
arch/arm64/kvm/hyp/nvhe/mem_protect.c
506
if (kvm_pte_valid(pte))
arch/arm64/kvm/hyp/nvhe/mem_protect.c
509
if (pte) {
arch/arm64/kvm/hyp/nvhe/mem_protect.c
637
enum pkvm_page_state (*get_page_state)(kvm_pte_t pte, u64 addr);
arch/arm64/kvm/hyp/nvhe/mem_protect.c
710
static enum pkvm_page_state guest_get_page_state(kvm_pte_t pte, u64 addr)
arch/arm64/kvm/hyp/nvhe/mem_protect.c
712
if (!kvm_pte_valid(pte))
arch/arm64/kvm/hyp/nvhe/mem_protect.c
715
return pkvm_getstate(kvm_pgtable_stage2_pte_prot(pte));
arch/arm64/kvm/hyp/nvhe/mm.c
234
kvm_pte_t pte, *ptep = slot->ptep;
arch/arm64/kvm/hyp/nvhe/mm.c
236
pte = *ptep;
arch/arm64/kvm/hyp/nvhe/mm.c
237
pte &= ~kvm_phys_to_pte(KVM_PHYS_INVALID);
arch/arm64/kvm/hyp/nvhe/mm.c
238
pte |= kvm_phys_to_pte(phys) | KVM_PTE_VALID;
arch/arm64/kvm/hyp/nvhe/mm.c
239
WRITE_ONCE(*ptep, pte);
arch/arm64/kvm/hyp/pgtable.c
100
pte |= KVM_PTE_VALID;
arch/arm64/kvm/hyp/pgtable.c
101
return pte;
arch/arm64/kvm/hyp/pgtable.c
106
kvm_pte_t pte = kvm_phys_to_pte(pa);
arch/arm64/kvm/hyp/pgtable.c
110
pte |= attr & (KVM_PTE_LEAF_ATTR_LO | KVM_PTE_LEAF_ATTR_HI);
arch/arm64/kvm/hyp/pgtable.c
111
pte |= FIELD_PREP(KVM_PTE_TYPE, type);
arch/arm64/kvm/hyp/pgtable.c
112
pte |= KVM_PTE_VALID;
arch/arm64/kvm/hyp/pgtable.c
114
return pte;
arch/arm64/kvm/hyp/pgtable.c
1209
kvm_pte_t pte;
arch/arm64/kvm/hyp/pgtable.c
1216
kvm_pte_t pte = ctx->old;
arch/arm64/kvm/hyp/pgtable.c
1224
data->pte = pte;
arch/arm64/kvm/hyp/pgtable.c
1225
pte &= ~data->attr_clr;
arch/arm64/kvm/hyp/pgtable.c
1226
pte |= data->attr_set;
arch/arm64/kvm/hyp/pgtable.c
1233
if (data->pte != pte) {
arch/arm64/kvm/hyp/pgtable.c
1239
stage2_pte_executable(pte) && !stage2_pte_executable(ctx->old))
arch/arm64/kvm/hyp/pgtable.c
1240
mm_ops->icache_inval_pou(kvm_pte_follow(pte, mm_ops),
arch/arm64/kvm/hyp/pgtable.c
1243
if (!stage2_try_set_pte(ctx, pte))
arch/arm64/kvm/hyp/pgtable.c
1272
*orig_pte = data.pte;
arch/arm64/kvm/hyp/pgtable.c
1486
kvm_pte_t pte = ctx->old, new, *childp;
arch/arm64/kvm/hyp/pgtable.c
1498
if (!kvm_pte_valid(pte))
arch/arm64/kvm/hyp/pgtable.c
1525
phys = kvm_pte_to_phys(pte);
arch/arm64/kvm/hyp/pgtable.c
1526
prot = kvm_pgtable_stage2_pte_prot(pte);
arch/arm64/kvm/hyp/pgtable.c
290
kvm_pte_t pte;
arch/arm64/kvm/hyp/pgtable.c
299
data->pte = ctx->old;
arch/arm64/kvm/hyp/pgtable.c
320
*ptep = data.pte;
arch/arm64/kvm/hyp/pgtable.c
379
enum kvm_pgtable_prot kvm_pgtable_hyp_pte_prot(kvm_pte_t pte)
arch/arm64/kvm/hyp/pgtable.c
381
enum kvm_pgtable_prot prot = pte & KVM_PTE_LEAF_ATTR_HI_SW;
arch/arm64/kvm/hyp/pgtable.c
384
if (!kvm_pte_valid(pte))
arch/arm64/kvm/hyp/pgtable.c
388
if (!(pte & KVM_PTE_LEAF_ATTR_HI_S1_PXN))
arch/arm64/kvm/hyp/pgtable.c
390
if (!(pte & KVM_PTE_LEAF_ATTR_HI_S1_UXN))
arch/arm64/kvm/hyp/pgtable.c
393
if (!(pte & KVM_PTE_LEAF_ATTR_HI_S1_XN))
arch/arm64/kvm/hyp/pgtable.c
397
ap = FIELD_GET(KVM_PTE_LEAF_ATTR_LO_S1_AP, pte);
arch/arm64/kvm/hyp/pgtable.c
74
static bool kvm_pte_table(kvm_pte_t pte, s8 level)
arch/arm64/kvm/hyp/pgtable.c
752
enum kvm_pgtable_prot kvm_pgtable_stage2_pte_prot(kvm_pte_t pte)
arch/arm64/kvm/hyp/pgtable.c
754
enum kvm_pgtable_prot prot = pte & KVM_PTE_LEAF_ATTR_HI_SW;
arch/arm64/kvm/hyp/pgtable.c
756
if (!kvm_pte_valid(pte))
arch/arm64/kvm/hyp/pgtable.c
759
if (pte & KVM_PTE_LEAF_ATTR_LO_S2_S2AP_R)
arch/arm64/kvm/hyp/pgtable.c
761
if (pte & KVM_PTE_LEAF_ATTR_LO_S2_S2AP_W)
arch/arm64/kvm/hyp/pgtable.c
764
switch (FIELD_GET(KVM_PTE_LEAF_ATTR_HI_S2_XN, pte)) {
arch/arm64/kvm/hyp/pgtable.c
789
static bool stage2_pte_is_counted(kvm_pte_t pte)
arch/arm64/kvm/hyp/pgtable.c
79
if (!kvm_pte_valid(pte))
arch/arm64/kvm/hyp/pgtable.c
796
return !!pte;
arch/arm64/kvm/hyp/pgtable.c
799
static bool stage2_pte_is_locked(kvm_pte_t pte)
arch/arm64/kvm/hyp/pgtable.c
801
return !kvm_pte_valid(pte) && (pte & KVM_INVALID_PTE_LOCKED);
arch/arm64/kvm/hyp/pgtable.c
82
return FIELD_GET(KVM_PTE_TYPE, pte) == KVM_PTE_TYPE_TABLE;
arch/arm64/kvm/hyp/pgtable.c
85
static kvm_pte_t *kvm_pte_follow(kvm_pte_t pte, struct kvm_pgtable_mm_ops *mm_ops)
arch/arm64/kvm/hyp/pgtable.c
87
return mm_ops->phys_to_virt(kvm_pte_to_phys(pte));
arch/arm64/kvm/hyp/pgtable.c
918
static bool stage2_pte_cacheable(struct kvm_pgtable *pgt, kvm_pte_t pte)
arch/arm64/kvm/hyp/pgtable.c
920
u64 memattr = pte & KVM_PTE_LEAF_ATTR_LO_S2_MEMATTR;
arch/arm64/kvm/hyp/pgtable.c
921
return kvm_pte_valid(pte) && memattr == KVM_S2_MEMATTR(pgt, NORMAL);
arch/arm64/kvm/hyp/pgtable.c
924
static bool stage2_pte_executable(kvm_pte_t pte)
arch/arm64/kvm/hyp/pgtable.c
926
return kvm_pte_valid(pte) && !(pte & KVM_PTE_LEAF_ATTR_HI_S2_XN);
arch/arm64/kvm/hyp/pgtable.c
97
kvm_pte_t pte = kvm_phys_to_pte(mm_ops->virt_to_phys(childp));
arch/arm64/kvm/hyp/pgtable.c
99
pte |= FIELD_PREP(KVM_PTE_TYPE, KVM_PTE_TYPE_TABLE);
arch/arm64/kvm/mmu.c
829
kvm_pte_t pte = 0; /* Keep GCC quiet... */
arch/arm64/kvm/mmu.c
839
ret = kvm_pgtable_get_leaf(&pgt, addr, &pte, &level);
arch/arm64/kvm/mmu.c
855
if (!kvm_pte_valid(pte))
arch/arm64/kvm/nested.c
516
kvm_pte_t pte;
arch/arm64/kvm/nested.c
562
if (kvm_pgtable_get_leaf(mmu->pgt, tmp, &pte, NULL))
arch/arm64/kvm/nested.c
564
if (!(pte & PTE_VALID))
arch/arm64/kvm/nested.c
566
level = FIELD_GET(KVM_NV_GUEST_MAP_SZ, pte);
arch/arm64/mm/contpte.c
230
__set_ptes(mm, start_addr, start_ptep, pte, CONT_PTES);
arch/arm64/mm/contpte.c
234
pte_t *ptep, pte_t pte)
arch/arm64/mm/contpte.c
268
page = pte_page(pte);
arch/arm64/mm/contpte.c
278
pfn = ALIGN_DOWN(pte_pfn(pte), CONT_PTES);
arch/arm64/mm/contpte.c
279
prot = pte_pgprot(pte_mkold(pte_mkclean(pte)));
arch/arm64/mm/contpte.c
292
pte = pte_mkcont(pte);
arch/arm64/mm/contpte.c
293
contpte_convert(mm, addr, orig_ptep, pte);
arch/arm64/mm/contpte.c
298
pte_t *ptep, pte_t pte)
arch/arm64/mm/contpte.c
307
pte = pte_mknoncont(pte);
arch/arm64/mm/contpte.c
308
contpte_convert(mm, addr, ptep, pte);
arch/arm64/mm/contpte.c
321
pte_t pte;
arch/arm64/mm/contpte.c
327
pte = __ptep_get(ptep);
arch/arm64/mm/contpte.c
329
if (pte_dirty(pte)) {
arch/arm64/mm/contpte.c
332
pte = __ptep_get(ptep);
arch/arm64/mm/contpte.c
333
if (pte_young(pte)) {
arch/arm64/mm/contpte.c
341
if (pte_young(pte)) {
arch/arm64/mm/contpte.c
346
pte = __ptep_get(ptep);
arch/arm64/mm/contpte.c
347
if (pte_dirty(pte)) {
arch/arm64/mm/contpte.c
360
static inline bool contpte_is_consistent(pte_t pte, unsigned long pfn,
arch/arm64/mm/contpte.c
363
pgprot_t prot = pte_pgprot(pte_mkold(pte_mkclean(pte)));
arch/arm64/mm/contpte.c
365
return pte_valid_cont(pte) && pte_pfn(pte) == pfn &&
arch/arm64/mm/contpte.c
392
pte_t pte;
arch/arm64/mm/contpte.c
406
pte = __ptep_get(ptep);
arch/arm64/mm/contpte.c
408
if (!contpte_is_consistent(pte, pfn, orig_prot))
arch/arm64/mm/contpte.c
411
if (pte_dirty(pte)) {
arch/arm64/mm/contpte.c
414
pte = __ptep_get(ptep);
arch/arm64/mm/contpte.c
416
if (!contpte_is_consistent(pte, pfn, orig_prot))
arch/arm64/mm/contpte.c
419
if (pte_young(pte)) {
arch/arm64/mm/contpte.c
427
if (pte_young(pte)) {
arch/arm64/mm/contpte.c
433
pte = __ptep_get(ptep);
arch/arm64/mm/contpte.c
435
if (!contpte_is_consistent(pte, pfn, orig_prot))
arch/arm64/mm/contpte.c
438
if (pte_dirty(pte)) {
arch/arm64/mm/contpte.c
452
pte_t *ptep, pte_t pte, unsigned int nr)
arch/arm64/mm/contpte.c
468
return __set_ptes(mm, addr, ptep, pte, nr);
arch/arm64/mm/contpte.c
471
pfn = pte_pfn(pte);
arch/arm64/mm/contpte.c
472
prot = pte_pgprot(pte);
arch/arm64/mm/contpte.c
477
pte = pfn_pte(pfn, prot);
arch/arm64/mm/contpte.c
480
pte = pte_mkcont(pte);
arch/arm64/mm/contpte.c
482
pte = pte_mknoncont(pte);
arch/arm64/mm/contpte.c
484
__set_ptes(mm, addr, ptep, pte, nr);
arch/arm64/mm/contpte.c
70
pte_t *ptep, pte_t pte)
arch/arm64/mm/contpte.c
79
pte = pfn_pte(ALIGN_DOWN(pte_pfn(pte), CONT_PTES), pte_pgprot(pte));
arch/arm64/mm/contpte.c
85
pte = pte_mkdirty(pte);
arch/arm64/mm/contpte.c
88
pte = pte_mkyoung(pte);
arch/arm64/mm/fault.c
162
pte_t *ptep, pte;
arch/arm64/mm/fault.c
189
pte = __ptep_get(ptep);
arch/arm64/mm/fault.c
190
pr_cont(", pte=%016llx", pte_val(pte));
arch/arm64/mm/fault.c
212
pte_t pte = __ptep_get(ptep);
arch/arm64/mm/fault.c
214
if (pte_same(pte, entry))
arch/arm64/mm/fault.c
227
pteval = pte_val(pte);
arch/arm64/mm/flush.c
52
void __sync_icache_dcache(pte_t pte)
arch/arm64/mm/flush.c
54
struct folio *folio = page_folio(pte_page(pte));
arch/arm64/mm/hugetlbpage.c
130
pte_t pte = __ptep_get(ptep);
arch/arm64/mm/hugetlbpage.c
132
if (pte_dirty(pte))
arch/arm64/mm/hugetlbpage.c
135
if (pte_young(pte))
arch/arm64/mm/hugetlbpage.c
155
pte_t pte, tmp_pte;
arch/arm64/mm/hugetlbpage.c
158
pte = __ptep_get_and_clear_anysz(mm, addr, ptep, pgsize);
arch/arm64/mm/hugetlbpage.c
159
present = pte_present(pte);
arch/arm64/mm/hugetlbpage.c
166
pte = pte_mkdirty(pte);
arch/arm64/mm/hugetlbpage.c
168
pte = pte_mkyoung(pte);
arch/arm64/mm/hugetlbpage.c
171
return pte;
arch/arm64/mm/hugetlbpage.c
216
pte_t *ptep, pte_t pte, unsigned long sz)
arch/arm64/mm/hugetlbpage.c
224
if (!pte_present(pte)) {
arch/arm64/mm/hugetlbpage.c
226
__set_ptes_anysz(mm, addr, ptep, pte, 1, pgsize);
arch/arm64/mm/hugetlbpage.c
231
if (pte_cont(pte) && pte_valid(__ptep_get(ptep)))
arch/arm64/mm/hugetlbpage.c
234
__set_ptes_anysz(mm, addr, ptep, pte, ncontig, pgsize);
arch/arm64/mm/hugetlbpage.c
400
static int __cont_access_flags_changed(pte_t *ptep, pte_t pte, int ncontig)
arch/arm64/mm/hugetlbpage.c
404
if (pte_write(pte) != pte_write(__ptep_get(ptep)))
arch/arm64/mm/hugetlbpage.c
410
if (pte_dirty(pte) != pte_dirty(orig_pte))
arch/arm64/mm/hugetlbpage.c
413
if (pte_young(pte) != pte_young(orig_pte))
arch/arm64/mm/hugetlbpage.c
422
pte_t pte, int dirty)
arch/arm64/mm/hugetlbpage.c
429
VM_WARN_ON(!pte_present(pte));
arch/arm64/mm/hugetlbpage.c
431
if (!pte_cont(pte))
arch/arm64/mm/hugetlbpage.c
432
return __ptep_set_access_flags(vma, addr, ptep, pte, dirty);
arch/arm64/mm/hugetlbpage.c
436
if (!__cont_access_flags_changed(ptep, pte, ncontig))
arch/arm64/mm/hugetlbpage.c
444
pte = pte_mkdirty(pte);
arch/arm64/mm/hugetlbpage.c
447
pte = pte_mkyoung(pte);
arch/arm64/mm/hugetlbpage.c
449
__set_ptes_anysz(mm, addr, ptep, pte, ncontig, pgsize);
arch/arm64/mm/hugetlbpage.c
458
pte_t pte;
arch/arm64/mm/hugetlbpage.c
460
pte = __ptep_get(ptep);
arch/arm64/mm/hugetlbpage.c
461
VM_WARN_ON(!pte_present(pte));
arch/arm64/mm/hugetlbpage.c
463
if (!pte_cont(pte)) {
arch/arm64/mm/hugetlbpage.c
470
pte = get_clear_contig_flush(mm, addr, ptep, pgsize, ncontig);
arch/arm64/mm/hugetlbpage.c
471
pte = pte_wrprotect(pte);
arch/arm64/mm/hugetlbpage.c
473
__set_ptes_anysz(mm, addr, ptep, pte, ncontig, pgsize);
arch/arm64/mm/hugetlbpage.c
534
pte_t old_pte, pte_t pte)
arch/arm64/mm/hugetlbpage.c
538
set_huge_pte_at(vma->vm_mm, addr, ptep, pte, psize);
arch/arm64/mm/mmu.c
1367
void __pi_map_range(phys_addr_t *pte, u64 start, u64 end, phys_addr_t pa,
arch/arm64/mm/mmu.c
1451
pte_t *ptep, pte;
arch/arm64/mm/mmu.c
1455
pte = __ptep_get(ptep);
arch/arm64/mm/mmu.c
1456
if (pte_none(pte))
arch/arm64/mm/mmu.c
1459
WARN_ON(!pte_present(pte));
arch/arm64/mm/mmu.c
1463
free_hotplug_page_range(pte_page(pte),
arch/arm64/mm/mmu.c
1583
pte_t *ptep, pte;
arch/arm64/mm/mmu.c
1588
pte = __ptep_get(ptep);
arch/arm64/mm/mmu.c
1594
WARN_ON(!pte_none(pte));
arch/arm64/mm/mmu.c
2142
pte_t pte = get_and_clear_ptes(vma->vm_mm, addr, ptep, nr);
arch/arm64/mm/mmu.c
2150
if (pte_accessible(vma->vm_mm, pte) && pte_user_exec(pte))
arch/arm64/mm/mmu.c
2155
return pte;
arch/arm64/mm/mmu.c
2164
pte_t *ptep, pte_t old_pte, pte_t pte,
arch/arm64/mm/mmu.c
2167
set_ptes(vma->vm_mm, addr, ptep, pte, nr);
arch/arm64/mm/mmu.c
2171
pte_t old_pte, pte_t pte)
arch/arm64/mm/mmu.c
2173
modify_prot_commit_ptes(vma, addr, ptep, old_pte, pte, 1);
arch/arm64/mm/mmu.c
673
pte_t *ptep, pte;
arch/arm64/mm/mmu.c
748
pte = __ptep_get(ptep);
arch/arm64/mm/mmu.c
749
if (!pte_present(pte))
arch/arm64/mm/mmu.c
751
if (pte_cont(pte))
arch/arm64/mm/mmu.c
868
pte_t pte = __ptep_get(ptep);
arch/arm64/mm/mmu.c
870
if (pte_cont(pte))
arch/arm64/mm/pageattr.c
66
static int pageattr_pte_entry(pte_t *pte, unsigned long addr,
arch/arm64/mm/pageattr.c
69
pte_t val = __ptep_get(pte);
arch/arm64/mm/pageattr.c
72
__set_pte(pte, val);
arch/arm64/mm/ptdump.c
254
void note_page_pte(struct ptdump_state *pt_st, unsigned long addr, pte_t pte)
arch/arm64/mm/ptdump.c
256
note_page(pt_st, addr, 4, pte_val(pte));
arch/arm64/mm/trans_pgd.c
36
pte_t pte = __ptep_get(src_ptep);
arch/arm64/mm/trans_pgd.c
38
if (pte_valid(pte)) {
arch/arm64/mm/trans_pgd.c
44
__set_pte(dst_ptep, pte_mkwrite_novma(pte));
arch/arm64/mm/trans_pgd.c
45
} else if (!pte_none(pte)) {
arch/arm64/mm/trans_pgd.c
58
BUG_ON(!pfn_valid(pte_pfn(pte)));
arch/arm64/mm/trans_pgd.c
60
__set_pte(dst_ptep, pte_mkvalid(pte_mkwrite_novma(pte)));
arch/csky/abiv2/cacheflush.c
11
unsigned long address, pte_t *pte, unsigned int nr)
arch/csky/abiv2/cacheflush.c
13
unsigned long pfn = pte_pfn(*pte);
arch/csky/include/asm/pgalloc.h
14
pte_t *pte)
arch/csky/include/asm/pgalloc.h
16
set_pmd(pmd, __pmd(__pa(pte)));
arch/csky/include/asm/pgalloc.h
20
pgtable_t pte)
arch/csky/include/asm/pgalloc.h
22
set_pmd(pmd, __pmd(__pa(page_address(pte))));
arch/csky/include/asm/pgalloc.h
29
pte_t *pte;
arch/csky/include/asm/pgalloc.h
32
pte = __pte_alloc_one_kernel(mm);
arch/csky/include/asm/pgalloc.h
33
if (!pte)
arch/csky/include/asm/pgalloc.h
37
(pte + i)->pte_low = _PAGE_GLOBAL;
arch/csky/include/asm/pgalloc.h
39
return pte;
arch/csky/include/asm/pgalloc.h
64
#define __pte_free_tlb(tlb, pte, address) \
arch/csky/include/asm/pgalloc.h
65
tlb_remove_ptdesc((tlb), page_ptdesc(pte))
arch/csky/include/asm/pgtable.h
141
static inline int pte_read(pte_t pte)
arch/csky/include/asm/pgtable.h
143
return pte.pte_low & _PAGE_READ;
arch/csky/include/asm/pgtable.h
146
static inline int pte_write(pte_t pte)
arch/csky/include/asm/pgtable.h
148
return (pte).pte_low & _PAGE_WRITE;
arch/csky/include/asm/pgtable.h
151
static inline int pte_dirty(pte_t pte)
arch/csky/include/asm/pgtable.h
153
return (pte).pte_low & _PAGE_MODIFIED;
arch/csky/include/asm/pgtable.h
156
static inline int pte_young(pte_t pte)
arch/csky/include/asm/pgtable.h
158
return (pte).pte_low & _PAGE_ACCESSED;
arch/csky/include/asm/pgtable.h
161
static inline pte_t pte_wrprotect(pte_t pte)
arch/csky/include/asm/pgtable.h
163
pte_val(pte) &= ~(_PAGE_WRITE | _PAGE_DIRTY);
arch/csky/include/asm/pgtable.h
164
return pte;
arch/csky/include/asm/pgtable.h
167
static inline pte_t pte_mkclean(pte_t pte)
arch/csky/include/asm/pgtable.h
169
pte_val(pte) &= ~(_PAGE_MODIFIED|_PAGE_DIRTY);
arch/csky/include/asm/pgtable.h
170
return pte;
arch/csky/include/asm/pgtable.h
173
static inline pte_t pte_mkold(pte_t pte)
arch/csky/include/asm/pgtable.h
175
pte_val(pte) &= ~(_PAGE_ACCESSED|_PAGE_VALID);
arch/csky/include/asm/pgtable.h
176
return pte;
arch/csky/include/asm/pgtable.h
179
static inline pte_t pte_mkwrite_novma(pte_t pte)
arch/csky/include/asm/pgtable.h
181
pte_val(pte) |= _PAGE_WRITE;
arch/csky/include/asm/pgtable.h
182
if (pte_val(pte) & _PAGE_MODIFIED)
arch/csky/include/asm/pgtable.h
183
pte_val(pte) |= _PAGE_DIRTY;
arch/csky/include/asm/pgtable.h
184
return pte;
arch/csky/include/asm/pgtable.h
187
static inline pte_t pte_mkdirty(pte_t pte)
arch/csky/include/asm/pgtable.h
189
pte_val(pte) |= _PAGE_MODIFIED;
arch/csky/include/asm/pgtable.h
190
if (pte_val(pte) & _PAGE_WRITE)
arch/csky/include/asm/pgtable.h
191
pte_val(pte) |= _PAGE_DIRTY;
arch/csky/include/asm/pgtable.h
192
return pte;
arch/csky/include/asm/pgtable.h
195
static inline pte_t pte_mkyoung(pte_t pte)
arch/csky/include/asm/pgtable.h
197
pte_val(pte) |= _PAGE_ACCESSED;
arch/csky/include/asm/pgtable.h
198
if (pte_val(pte) & _PAGE_READ)
arch/csky/include/asm/pgtable.h
199
pte_val(pte) |= _PAGE_VALID;
arch/csky/include/asm/pgtable.h
200
return pte;
arch/csky/include/asm/pgtable.h
203
static inline bool pte_swp_exclusive(pte_t pte)
arch/csky/include/asm/pgtable.h
205
return pte_val(pte) & _PAGE_SWP_EXCLUSIVE;
arch/csky/include/asm/pgtable.h
208
static inline pte_t pte_swp_mkexclusive(pte_t pte)
arch/csky/include/asm/pgtable.h
210
pte_val(pte) |= _PAGE_SWP_EXCLUSIVE;
arch/csky/include/asm/pgtable.h
211
return pte;
arch/csky/include/asm/pgtable.h
214
static inline pte_t pte_swp_clear_exclusive(pte_t pte)
arch/csky/include/asm/pgtable.h
216
pte_val(pte) &= ~_PAGE_SWP_EXCLUSIVE;
arch/csky/include/asm/pgtable.h
217
return pte;
arch/csky/include/asm/pgtable.h
252
static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
arch/csky/include/asm/pgtable.h
254
return __pte((pte_val(pte) & _PAGE_CHG_MASK) |
arch/csky/include/asm/pgtable.h
262
unsigned long address, pte_t *pte, unsigned int nr);
arch/csky/include/asm/pgtable.h
36
#define pte_none(pte) (!(pte_val(pte) & ~_PAGE_GLOBAL))
arch/csky/include/asm/pgtable.h
37
#define pte_present(pte) (pte_val(pte) & _PAGE_PRESENT)
arch/csky/include/asm/pgtable.h
42
#define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) })
arch/csky/include/asm/pgtable.h
85
static inline void set_pte(pte_t *p, pte_t pte)
arch/csky/include/asm/pgtable.h
87
*p = pte;
arch/csky/mm/highmem.c
26
pte_t *pte;
arch/csky/mm/highmem.c
34
pte = pte_offset_kernel(pmd, vaddr);
arch/csky/mm/highmem.c
35
pkmap_page_table = pte;
arch/csky/mm/init.c
109
pte = (pte_t *) memblock_alloc_low(PAGE_SIZE, PAGE_SIZE);
arch/csky/mm/init.c
110
if (!pte)
arch/csky/mm/init.c
115
set_pmd(pmd, __pmd(__pa(pte)));
arch/csky/mm/init.c
116
BUG_ON(pte != pte_offset_kernel(pmd, 0));
arch/csky/mm/init.c
93
pte_t *pte;
arch/hexagon/include/asm/page.h
63
typedef struct { unsigned long pte; } pte_t;
arch/hexagon/include/asm/page.h
68
#define pte_val(x) ((x).pte)
arch/hexagon/include/asm/pgalloc.h
45
pgtable_t pte)
arch/hexagon/include/asm/pgalloc.h
51
set_pmd(pmd, __pmd(((unsigned long)page_to_pfn(pte) << PAGE_SHIFT) |
arch/hexagon/include/asm/pgalloc.h
65
pte_t *pte)
arch/hexagon/include/asm/pgalloc.h
77
set_pmd(pmd, __pmd(((unsigned long)__pa(pte)) | HEXAGON_L1_PTE_SIZE));
arch/hexagon/include/asm/pgalloc.h
85
set_pmd(ppmd, __pmd(((unsigned long)__pa(pte)) | HEXAGON_L1_PTE_SIZE));
arch/hexagon/include/asm/pgalloc.h
90
#define __pte_free_tlb(tlb, pte, addr) \
arch/hexagon/include/asm/pgalloc.h
91
tlb_remove_ptdesc((tlb), page_ptdesc(pte))
arch/hexagon/include/asm/pgtable.h
136
#define pte_mkhuge(pte) __pte((pte_val(pte) & ~0x3) | HVM_HUGEPAGE_SIZE)
arch/hexagon/include/asm/pgtable.h
143
extern void sync_icache_dcache(pte_t pte);
arch/hexagon/include/asm/pgtable.h
145
#define pte_present_exec_user(pte) \
arch/hexagon/include/asm/pgtable.h
146
((pte_val(pte) & (_PAGE_EXECUTE | _PAGE_USER)) == \
arch/hexagon/include/asm/pgtable.h
228
static inline int pte_none(pte_t pte)
arch/hexagon/include/asm/pgtable.h
230
return pte_val(pte) == _NULL_PTE;
arch/hexagon/include/asm/pgtable.h
236
static inline int pte_present(pte_t pte)
arch/hexagon/include/asm/pgtable.h
238
return pte_val(pte) & _PAGE_PRESENT;
arch/hexagon/include/asm/pgtable.h
245
static inline pte_t pte_mkold(pte_t pte)
arch/hexagon/include/asm/pgtable.h
247
pte_val(pte) &= ~_PAGE_ACCESSED;
arch/hexagon/include/asm/pgtable.h
248
return pte;
arch/hexagon/include/asm/pgtable.h
252
static inline pte_t pte_mkyoung(pte_t pte)
arch/hexagon/include/asm/pgtable.h
254
pte_val(pte) |= _PAGE_ACCESSED;
arch/hexagon/include/asm/pgtable.h
255
return pte;
arch/hexagon/include/asm/pgtable.h
259
static inline pte_t pte_mkclean(pte_t pte)
arch/hexagon/include/asm/pgtable.h
261
pte_val(pte) &= ~_PAGE_DIRTY;
arch/hexagon/include/asm/pgtable.h
262
return pte;
arch/hexagon/include/asm/pgtable.h
266
static inline pte_t pte_mkdirty(pte_t pte)
arch/hexagon/include/asm/pgtable.h
268
pte_val(pte) |= _PAGE_DIRTY;
arch/hexagon/include/asm/pgtable.h
269
return pte;
arch/hexagon/include/asm/pgtable.h
273
static inline int pte_young(pte_t pte)
arch/hexagon/include/asm/pgtable.h
275
return pte_val(pte) & _PAGE_ACCESSED;
arch/hexagon/include/asm/pgtable.h
279
static inline int pte_dirty(pte_t pte)
arch/hexagon/include/asm/pgtable.h
281
return pte_val(pte) & _PAGE_DIRTY;
arch/hexagon/include/asm/pgtable.h
285
static inline pte_t pte_modify(pte_t pte, pgprot_t prot)
arch/hexagon/include/asm/pgtable.h
287
pte_val(pte) &= PAGE_MASK;
arch/hexagon/include/asm/pgtable.h
288
pte_val(pte) |= pgprot_val(prot);
arch/hexagon/include/asm/pgtable.h
289
return pte;
arch/hexagon/include/asm/pgtable.h
293
static inline pte_t pte_wrprotect(pte_t pte)
arch/hexagon/include/asm/pgtable.h
295
pte_val(pte) &= ~_PAGE_WRITE;
arch/hexagon/include/asm/pgtable.h
296
return pte;
arch/hexagon/include/asm/pgtable.h
300
static inline pte_t pte_mkwrite_novma(pte_t pte)
arch/hexagon/include/asm/pgtable.h
302
pte_val(pte) |= _PAGE_WRITE;
arch/hexagon/include/asm/pgtable.h
303
return pte;
arch/hexagon/include/asm/pgtable.h
307
static inline pte_t pte_mkexec(pte_t pte)
arch/hexagon/include/asm/pgtable.h
309
pte_val(pte) |= _PAGE_EXECUTE;
arch/hexagon/include/asm/pgtable.h
310
return pte;
arch/hexagon/include/asm/pgtable.h
314
static inline int pte_read(pte_t pte)
arch/hexagon/include/asm/pgtable.h
316
return pte_val(pte) & _PAGE_READ;
arch/hexagon/include/asm/pgtable.h
320
static inline int pte_write(pte_t pte)
arch/hexagon/include/asm/pgtable.h
322
return pte_val(pte) & _PAGE_WRITE;
arch/hexagon/include/asm/pgtable.h
327
static inline int pte_exec(pte_t pte)
arch/hexagon/include/asm/pgtable.h
329
return pte_val(pte) & _PAGE_EXECUTE;
arch/hexagon/include/asm/pgtable.h
333
#define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) })
arch/hexagon/include/asm/pgtable.h
343
#define pte_pfn(pte) (pte_val(pte) >> PAGE_SHIFT)
arch/hexagon/include/asm/pgtable.h
390
static inline bool pte_swp_exclusive(pte_t pte)
arch/hexagon/include/asm/pgtable.h
392
return pte_val(pte) & _PAGE_SWP_EXCLUSIVE;
arch/hexagon/include/asm/pgtable.h
395
static inline pte_t pte_swp_mkexclusive(pte_t pte)
arch/hexagon/include/asm/pgtable.h
397
pte_val(pte) |= _PAGE_SWP_EXCLUSIVE;
arch/hexagon/include/asm/pgtable.h
398
return pte;
arch/hexagon/include/asm/pgtable.h
401
static inline pte_t pte_swp_clear_exclusive(pte_t pte)
arch/hexagon/include/asm/pgtable.h
403
pte_val(pte) &= ~_PAGE_SWP_EXCLUSIVE;
arch/hexagon/include/asm/pgtable.h
404
return pte;
arch/hexagon/include/asm/vm_mmu.h
47
#define __HVM_PTE_CVAL(pte) (((pte) & __HVM_PTE_C) >> 6)
arch/hexagon/mm/init.c
46
void sync_icache_dcache(pte_t pte)
arch/hexagon/mm/init.c
51
page = pte_page(pte);
arch/loongarch/include/asm/hugetlb.h
29
pte_t pte = ptep_get(ptep);
arch/loongarch/include/asm/hugetlb.h
33
return pte;
arch/loongarch/include/asm/hugetlb.h
40
pte_t pte;
arch/loongarch/include/asm/hugetlb.h
43
pte = huge_ptep_get_and_clear(vma->vm_mm, addr, ptep, sz);
arch/loongarch/include/asm/hugetlb.h
45
return pte;
arch/loongarch/include/asm/hugetlb.h
49
static inline int huge_pte_none(pte_t pte)
arch/loongarch/include/asm/hugetlb.h
51
unsigned long val = pte_val(pte) & ~_PAGE_GLOBAL;
arch/loongarch/include/asm/hugetlb.h
58
pte_t *ptep, pte_t pte,
arch/loongarch/include/asm/hugetlb.h
61
int changed = !pte_same(ptep_get(ptep), pte);
arch/loongarch/include/asm/hugetlb.h
64
set_pte_at(vma->vm_mm, addr, ptep, pte);
arch/loongarch/include/asm/kfence.h
54
pte_t *pte = virt_to_kpte(addr);
arch/loongarch/include/asm/kfence.h
56
if (WARN_ON(!pte) || pte_none(ptep_get(pte)))
arch/loongarch/include/asm/kfence.h
60
set_pte(pte, __pte(pte_val(ptep_get(pte)) & ~(_PAGE_VALID | _PAGE_PRESENT)));
arch/loongarch/include/asm/kfence.h
62
set_pte(pte, __pte(pte_val(ptep_get(pte)) | (_PAGE_VALID | _PAGE_PRESENT)));
arch/loongarch/include/asm/kvm_mmu.h
33
typedef int (*kvm_pte_ops)(kvm_pte_t *pte, phys_addr_t addr, kvm_ptw_ctx *ctx);
arch/loongarch/include/asm/kvm_mmu.h
62
static inline int kvm_pte_young(kvm_pte_t pte) { return pte & _PAGE_ACCESSED; }
arch/loongarch/include/asm/kvm_mmu.h
63
static inline int kvm_pte_huge(kvm_pte_t pte) { return pte & _PAGE_HUGE; }
arch/loongarch/include/asm/kvm_mmu.h
64
static inline int kvm_pte_dirty(kvm_pte_t pte) { return pte & __WRITEABLE; }
arch/loongarch/include/asm/kvm_mmu.h
65
static inline int kvm_pte_writeable(kvm_pte_t pte) { return pte & KVM_PAGE_WRITEABLE; }
arch/loongarch/include/asm/kvm_mmu.h
67
static inline kvm_pte_t kvm_pte_mkyoung(kvm_pte_t pte)
arch/loongarch/include/asm/kvm_mmu.h
69
return pte | _PAGE_ACCESSED;
arch/loongarch/include/asm/kvm_mmu.h
72
static inline kvm_pte_t kvm_pte_mkold(kvm_pte_t pte)
arch/loongarch/include/asm/kvm_mmu.h
74
return pte & ~_PAGE_ACCESSED;
arch/loongarch/include/asm/kvm_mmu.h
77
static inline kvm_pte_t kvm_pte_mkdirty(kvm_pte_t pte)
arch/loongarch/include/asm/kvm_mmu.h
79
return pte | __WRITEABLE;
arch/loongarch/include/asm/kvm_mmu.h
82
static inline kvm_pte_t kvm_pte_mkclean(kvm_pte_t pte)
arch/loongarch/include/asm/kvm_mmu.h
84
return pte & ~__WRITEABLE;
arch/loongarch/include/asm/kvm_mmu.h
87
static inline kvm_pte_t kvm_pte_mkhuge(kvm_pte_t pte)
arch/loongarch/include/asm/kvm_mmu.h
89
return pte | _PAGE_HUGE;
arch/loongarch/include/asm/kvm_mmu.h
92
static inline kvm_pte_t kvm_pte_mksmall(kvm_pte_t pte)
arch/loongarch/include/asm/kvm_mmu.h
94
return pte & ~_PAGE_HUGE;
arch/loongarch/include/asm/kvm_mmu.h
97
static inline kvm_pte_t kvm_pte_mkwriteable(kvm_pte_t pte)
arch/loongarch/include/asm/kvm_mmu.h
99
return pte | KVM_PAGE_WRITEABLE;
arch/loongarch/include/asm/page.h
44
typedef struct { unsigned long pte; } pte_t;
arch/loongarch/include/asm/page.h
45
#define pte_val(x) ((x).pte)
arch/loongarch/include/asm/pgalloc.h
17
pmd_t *pmd, pte_t *pte)
arch/loongarch/include/asm/pgalloc.h
19
set_pmd(pmd, __pmd((unsigned long)pte));
arch/loongarch/include/asm/pgalloc.h
22
static inline void pmd_populate(struct mm_struct *mm, pmd_t *pmd, pgtable_t pte)
arch/loongarch/include/asm/pgalloc.h
24
set_pmd(pmd, __pmd((unsigned long)page_address(pte)));
arch/loongarch/include/asm/pgalloc.h
50
pte_t *pte = __pte_alloc_one_kernel(mm);
arch/loongarch/include/asm/pgalloc.h
52
if (pte)
arch/loongarch/include/asm/pgalloc.h
53
kernel_pte_init(pte);
arch/loongarch/include/asm/pgalloc.h
55
return pte;
arch/loongarch/include/asm/pgalloc.h
58
#define __pte_free_tlb(tlb, pte, address) tlb_remove_ptdesc((tlb), page_ptdesc(pte))
arch/loongarch/include/asm/pgtable.h
276
#define pte_pfn(x) ((unsigned long)(((x).pte & _PFN_MASK) >> PFN_PTE_SHIFT))
arch/loongarch/include/asm/pgtable.h
325
pte_t pte;
arch/loongarch/include/asm/pgtable.h
326
pte_val(pte) = ((type & __SWP_TYPE_MASK) << __SWP_TYPE_SHIFT) | (offset << __SWP_OFFSET_SHIFT);
arch/loongarch/include/asm/pgtable.h
327
return pte;
arch/loongarch/include/asm/pgtable.h
336
#define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) })
arch/loongarch/include/asm/pgtable.h
339
static inline bool pte_swp_exclusive(pte_t pte)
arch/loongarch/include/asm/pgtable.h
341
return pte_val(pte) & _PAGE_SWP_EXCLUSIVE;
arch/loongarch/include/asm/pgtable.h
344
static inline pte_t pte_swp_mkexclusive(pte_t pte)
arch/loongarch/include/asm/pgtable.h
346
pte_val(pte) |= _PAGE_SWP_EXCLUSIVE;
arch/loongarch/include/asm/pgtable.h
347
return pte;
arch/loongarch/include/asm/pgtable.h
350
static inline pte_t pte_swp_clear_exclusive(pte_t pte)
arch/loongarch/include/asm/pgtable.h
352
pte_val(pte) &= ~_PAGE_SWP_EXCLUSIVE;
arch/loongarch/include/asm/pgtable.h
353
return pte;
arch/loongarch/include/asm/pgtable.h
356
#define pte_none(pte) (!(pte_val(pte) & ~_PAGE_GLOBAL))
arch/loongarch/include/asm/pgtable.h
357
#define pte_present(pte) (pte_val(pte) & (_PAGE_PRESENT | _PAGE_PROTNONE))
arch/loongarch/include/asm/pgtable.h
358
#define pte_no_exec(pte) (pte_val(pte) & _PAGE_NO_EXEC)
arch/loongarch/include/asm/pgtable.h
372
pte_t pte = ptep_get(ptep);
arch/loongarch/include/asm/pgtable.h
373
pte_val(pte) &= _PAGE_GLOBAL;
arch/loongarch/include/asm/pgtable.h
374
set_pte(ptep, pte);
arch/loongarch/include/asm/pgtable.h
388
static inline int pte_write(pte_t pte) { return pte_val(pte) & _PAGE_WRITE; }
arch/loongarch/include/asm/pgtable.h
389
static inline int pte_young(pte_t pte) { return pte_val(pte) & _PAGE_ACCESSED; }
arch/loongarch/include/asm/pgtable.h
390
static inline int pte_dirty(pte_t pte) { return pte_val(pte) & (_PAGE_DIRTY | _PAGE_MODIFIED); }
arch/loongarch/include/asm/pgtable.h
392
static inline pte_t pte_mkold(pte_t pte)
arch/loongarch/include/asm/pgtable.h
394
pte_val(pte) &= ~_PAGE_ACCESSED;
arch/loongarch/include/asm/pgtable.h
395
return pte;
arch/loongarch/include/asm/pgtable.h
398
static inline pte_t pte_mkyoung(pte_t pte)
arch/loongarch/include/asm/pgtable.h
400
pte_val(pte) |= _PAGE_ACCESSED;
arch/loongarch/include/asm/pgtable.h
401
return pte;
arch/loongarch/include/asm/pgtable.h
404
static inline pte_t pte_mkclean(pte_t pte)
arch/loongarch/include/asm/pgtable.h
406
pte_val(pte) &= ~(_PAGE_DIRTY | _PAGE_MODIFIED);
arch/loongarch/include/asm/pgtable.h
407
return pte;
arch/loongarch/include/asm/pgtable.h
410
static inline pte_t pte_mkdirty(pte_t pte)
arch/loongarch/include/asm/pgtable.h
412
pte_val(pte) |= _PAGE_MODIFIED;
arch/loongarch/include/asm/pgtable.h
413
if (pte_val(pte) & _PAGE_WRITE)
arch/loongarch/include/asm/pgtable.h
414
pte_val(pte) |= _PAGE_DIRTY;
arch/loongarch/include/asm/pgtable.h
415
return pte;
arch/loongarch/include/asm/pgtable.h
418
static inline pte_t pte_mkwrite_novma(pte_t pte)
arch/loongarch/include/asm/pgtable.h
420
pte_val(pte) |= _PAGE_WRITE;
arch/loongarch/include/asm/pgtable.h
421
if (pte_val(pte) & _PAGE_MODIFIED)
arch/loongarch/include/asm/pgtable.h
422
pte_val(pte) |= _PAGE_DIRTY;
arch/loongarch/include/asm/pgtable.h
423
return pte;
arch/loongarch/include/asm/pgtable.h
426
static inline pte_t pte_wrprotect(pte_t pte)
arch/loongarch/include/asm/pgtable.h
428
pte_val(pte) &= ~(_PAGE_WRITE | _PAGE_DIRTY);
arch/loongarch/include/asm/pgtable.h
429
return pte;
arch/loongarch/include/asm/pgtable.h
432
static inline int pte_huge(pte_t pte) { return pte_val(pte) & _PAGE_HUGE; }
arch/loongarch/include/asm/pgtable.h
434
static inline pte_t pte_mkhuge(pte_t pte)
arch/loongarch/include/asm/pgtable.h
436
pte_val(pte) |= _PAGE_HUGE;
arch/loongarch/include/asm/pgtable.h
437
return pte;
arch/loongarch/include/asm/pgtable.h
441
static inline int pte_special(pte_t pte) { return pte_val(pte) & _PAGE_SPECIAL; }
arch/loongarch/include/asm/pgtable.h
442
static inline pte_t pte_mkspecial(pte_t pte) { pte_val(pte) |= _PAGE_SPECIAL; return pte; }
arch/loongarch/include/asm/pgtable.h
458
static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
arch/loongarch/include/asm/pgtable.h
460
if (pte_val(pte) & _PAGE_DIRTY)
arch/loongarch/include/asm/pgtable.h
461
pte_val(pte) |= _PAGE_MODIFIED;
arch/loongarch/include/asm/pgtable.h
463
return __pte((pte_val(pte) & _PAGE_CHG_MASK) |
arch/loongarch/include/asm/pgtable.h
619
static inline long pte_protnone(pte_t pte)
arch/loongarch/include/asm/pgtable.h
621
return (pte_val(pte) & _PAGE_PROTNONE);
arch/loongarch/kvm/mmu.c
41
static int kvm_mkold_pte(kvm_pte_t *pte, phys_addr_t addr, kvm_ptw_ctx *ctx)
arch/loongarch/kvm/mmu.c
43
if (kvm_pte_young(*pte)) {
arch/loongarch/kvm/mmu.c
44
*pte = kvm_pte_mkold(*pte);
arch/loongarch/kvm/mmu.c
55
static int kvm_mkclean_pte(kvm_pte_t *pte, phys_addr_t addr, kvm_ptw_ctx *ctx)
arch/loongarch/kvm/mmu.c
60
val = *pte;
arch/loongarch/kvm/mmu.c
78
*pte = kvm_pte_mkclean(val);
arch/loongarch/kvm/mmu.c
88
static int kvm_flush_pte(kvm_pte_t *pte, phys_addr_t addr, kvm_ptw_ctx *ctx)
arch/loongarch/kvm/mmu.c
98
*pte = ctx->invalid_entry;
arch/loongarch/mm/fault.c
40
pte_t *pte;
arch/loongarch/mm/fault.c
64
pte = pte_offset_kernel(pmd, address);
arch/loongarch/mm/fault.c
65
if (!pte_present(ptep_get(pte)))
arch/loongarch/mm/fault.c
68
return write ? pte_write(ptep_get(pte)) : 1;
arch/loongarch/mm/hugetlbpage.c
22
pte_t *pte = NULL;
arch/loongarch/mm/hugetlbpage.c
28
pte = (pte_t *)pmd_alloc(mm, pud, addr);
arch/loongarch/mm/hugetlbpage.c
30
return pte;
arch/loongarch/mm/init.c
170
pte_t *pte;
arch/loongarch/mm/init.c
172
pte = memblock_alloc_or_panic(PAGE_SIZE, PAGE_SIZE);
arch/loongarch/mm/init.c
173
pmd_populate_kernel(&init_mm, pmd, pte);
arch/loongarch/mm/init.c
174
kernel_pte_init(pte);
arch/loongarch/mm/kasan_init.c
40
#define __pte_none(early, pte) (early ? pte_none(pte) : \
arch/loongarch/mm/kasan_init.c
41
((pte_val(pte) & _PFN_MASK) == (unsigned long)__pa(kasan_early_shadow_page)))
arch/loongarch/mm/pageattr.c
167
pte_t *pte;
arch/loongarch/mm/pageattr.c
197
pte = pte_offset_kernel(pmd, addr);
arch/loongarch/mm/pageattr.c
198
return pte_present(ptep_get(pte));
arch/loongarch/mm/pageattr.c
80
static int pageattr_pte_entry(pte_t *pte, unsigned long addr,
arch/loongarch/mm/pageattr.c
83
pte_t val = ptep_get(pte);
arch/loongarch/mm/pageattr.c
86
set_pte(pte, val);
arch/m68k/include/asm/mcf_pgalloc.h
10
pagetable_dtor_free(virt_to_ptdesc(pte));
arch/m68k/include/asm/mcf_pgalloc.h
35
#define pmd_populate(mm, pmd, pte) (pmd_val(*pmd) = (unsigned long)(pte))
arch/m68k/include/asm/mcf_pgalloc.h
51
pte_t *pte;
arch/m68k/include/asm/mcf_pgalloc.h
60
pte = ptdesc_address(ptdesc);
arch/m68k/include/asm/mcf_pgalloc.h
61
return pte;
arch/m68k/include/asm/mcf_pgalloc.h
8
static inline void pte_free_kernel(struct mm_struct *mm, pte_t *pte)
arch/m68k/include/asm/mcf_pgtable.h
101
pte_val(pte) = (pte_val(pte) & CF_PAGE_CHG_MASK) | pgprot_val(newprot);
arch/m68k/include/asm/mcf_pgtable.h
102
return pte;
arch/m68k/include/asm/mcf_pgtable.h
112
#define __pte_page(pte) ((void *) (pte_val(pte) & PAGE_MASK))
arch/m68k/include/asm/mcf_pgtable.h
115
static inline int pte_none(pte_t pte)
arch/m68k/include/asm/mcf_pgtable.h
117
return !pte_val(pte);
arch/m68k/include/asm/mcf_pgtable.h
120
static inline int pte_present(pte_t pte)
arch/m68k/include/asm/mcf_pgtable.h
122
return pte_val(pte) & CF_PAGE_VALID;
arch/m68k/include/asm/mcf_pgtable.h
131
#define pte_page(pte) virt_to_page(__pte_page(pte))
arch/m68k/include/asm/mcf_pgtable.h
152
static inline int pte_read(pte_t pte)
arch/m68k/include/asm/mcf_pgtable.h
154
return pte_val(pte) & CF_PAGE_READABLE;
arch/m68k/include/asm/mcf_pgtable.h
157
static inline int pte_write(pte_t pte)
arch/m68k/include/asm/mcf_pgtable.h
159
return pte_val(pte) & CF_PAGE_WRITABLE;
arch/m68k/include/asm/mcf_pgtable.h
162
static inline int pte_exec(pte_t pte)
arch/m68k/include/asm/mcf_pgtable.h
164
return pte_val(pte) & CF_PAGE_EXEC;
arch/m68k/include/asm/mcf_pgtable.h
167
static inline int pte_dirty(pte_t pte)
arch/m68k/include/asm/mcf_pgtable.h
169
return pte_val(pte) & CF_PAGE_DIRTY;
arch/m68k/include/asm/mcf_pgtable.h
172
static inline int pte_young(pte_t pte)
arch/m68k/include/asm/mcf_pgtable.h
174
return pte_val(pte) & CF_PAGE_ACCESSED;
arch/m68k/include/asm/mcf_pgtable.h
177
static inline pte_t pte_wrprotect(pte_t pte)
arch/m68k/include/asm/mcf_pgtable.h
179
pte_val(pte) &= ~CF_PAGE_WRITABLE;
arch/m68k/include/asm/mcf_pgtable.h
180
return pte;
arch/m68k/include/asm/mcf_pgtable.h
183
static inline pte_t pte_rdprotect(pte_t pte)
arch/m68k/include/asm/mcf_pgtable.h
185
pte_val(pte) &= ~CF_PAGE_READABLE;
arch/m68k/include/asm/mcf_pgtable.h
186
return pte;
arch/m68k/include/asm/mcf_pgtable.h
189
static inline pte_t pte_exprotect(pte_t pte)
arch/m68k/include/asm/mcf_pgtable.h
191
pte_val(pte) &= ~CF_PAGE_EXEC;
arch/m68k/include/asm/mcf_pgtable.h
192
return pte;
arch/m68k/include/asm/mcf_pgtable.h
195
static inline pte_t pte_mkclean(pte_t pte)
arch/m68k/include/asm/mcf_pgtable.h
197
pte_val(pte) &= ~CF_PAGE_DIRTY;
arch/m68k/include/asm/mcf_pgtable.h
198
return pte;
arch/m68k/include/asm/mcf_pgtable.h
201
static inline pte_t pte_mkold(pte_t pte)
arch/m68k/include/asm/mcf_pgtable.h
203
pte_val(pte) &= ~CF_PAGE_ACCESSED;
arch/m68k/include/asm/mcf_pgtable.h
204
return pte;
arch/m68k/include/asm/mcf_pgtable.h
207
static inline pte_t pte_mkwrite_novma(pte_t pte)
arch/m68k/include/asm/mcf_pgtable.h
209
pte_val(pte) |= CF_PAGE_WRITABLE;
arch/m68k/include/asm/mcf_pgtable.h
210
return pte;
arch/m68k/include/asm/mcf_pgtable.h
213
static inline pte_t pte_mkread(pte_t pte)
arch/m68k/include/asm/mcf_pgtable.h
215
pte_val(pte) |= CF_PAGE_READABLE;
arch/m68k/include/asm/mcf_pgtable.h
216
return pte;
arch/m68k/include/asm/mcf_pgtable.h
219
static inline pte_t pte_mkexec(pte_t pte)
arch/m68k/include/asm/mcf_pgtable.h
221
pte_val(pte) |= CF_PAGE_EXEC;
arch/m68k/include/asm/mcf_pgtable.h
222
return pte;
arch/m68k/include/asm/mcf_pgtable.h
225
static inline pte_t pte_mkdirty(pte_t pte)
arch/m68k/include/asm/mcf_pgtable.h
227
pte_val(pte) |= CF_PAGE_DIRTY;
arch/m68k/include/asm/mcf_pgtable.h
228
return pte;
arch/m68k/include/asm/mcf_pgtable.h
231
static inline pte_t pte_mkyoung(pte_t pte)
arch/m68k/include/asm/mcf_pgtable.h
233
pte_val(pte) |= CF_PAGE_ACCESSED;
arch/m68k/include/asm/mcf_pgtable.h
234
return pte;
arch/m68k/include/asm/mcf_pgtable.h
237
static inline pte_t pte_mknocache(pte_t pte)
arch/m68k/include/asm/mcf_pgtable.h
239
pte_val(pte) |= 0x80 | (pte_val(pte) & ~0x40);
arch/m68k/include/asm/mcf_pgtable.h
240
return pte;
arch/m68k/include/asm/mcf_pgtable.h
243
static inline pte_t pte_mkcache(pte_t pte)
arch/m68k/include/asm/mcf_pgtable.h
245
pte_val(pte) &= ~CF_PAGE_NOCACHE;
arch/m68k/include/asm/mcf_pgtable.h
246
return pte;
arch/m68k/include/asm/mcf_pgtable.h
268
#define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) })
arch/m68k/include/asm/mcf_pgtable.h
271
static inline bool pte_swp_exclusive(pte_t pte)
arch/m68k/include/asm/mcf_pgtable.h
273
return pte_val(pte) & _PAGE_SWP_EXCLUSIVE;
arch/m68k/include/asm/mcf_pgtable.h
276
static inline pte_t pte_swp_mkexclusive(pte_t pte)
arch/m68k/include/asm/mcf_pgtable.h
278
pte_val(pte) |= _PAGE_SWP_EXCLUSIVE;
arch/m68k/include/asm/mcf_pgtable.h
279
return pte;
arch/m68k/include/asm/mcf_pgtable.h
282
static inline pte_t pte_swp_clear_exclusive(pte_t pte)
arch/m68k/include/asm/mcf_pgtable.h
284
pte_val(pte) &= ~_PAGE_SWP_EXCLUSIVE;
arch/m68k/include/asm/mcf_pgtable.h
285
return pte;
arch/m68k/include/asm/mcf_pgtable.h
293
#define pte_pfn(pte) (pte_val(pte) >> PAGE_SHIFT)
arch/m68k/include/asm/mcf_pgtable.h
99
static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
arch/m68k/include/asm/mmu_context.h
102
pte_t *pte = NULL;
arch/m68k/include/asm/mmu_context.h
140
pte = (mmuar >= PAGE_OFFSET) ? pte_offset_kernel(pmd, mmuar)
arch/m68k/include/asm/mmu_context.h
142
if (!pte || pte_none(*pte) || !pte_present(*pte))
arch/m68k/include/asm/mmu_context.h
145
set_pte(pte, pte_mkyoung(*pte));
arch/m68k/include/asm/mmu_context.h
147
if (!pte_dirty(*pte) && mmuar <= PAGE_OFFSET)
arch/m68k/include/asm/mmu_context.h
148
set_pte(pte, pte_wrprotect(*pte));
arch/m68k/include/asm/mmu_context.h
151
(((int)(pte->pte) & (int)CF_PAGE_MMUTR_MASK)
arch/m68k/include/asm/mmu_context.h
154
mmu_write(MMUDR, (pte_val(*pte) & PAGE_MASK) |
arch/m68k/include/asm/mmu_context.h
155
((pte->pte) & CF_PAGE_MMUDR_MASK) | MMUDR_SZ_8KB | MMUDR_X);
arch/m68k/include/asm/mmu_context.h
164
if (pte && mmuar < PAGE_OFFSET)
arch/m68k/include/asm/mmu_context.h
165
pte_unmap(pte);
arch/m68k/include/asm/motorola_pgalloc.h
32
static inline void pte_free_kernel(struct mm_struct *mm, pte_t *pte)
arch/m68k/include/asm/motorola_pgalloc.h
34
free_pointer_table(pte, TABLE_PTE);
arch/m68k/include/asm/motorola_pgalloc.h
82
static inline void pmd_populate_kernel(struct mm_struct *mm, pmd_t *pmd, pte_t *pte)
arch/m68k/include/asm/motorola_pgalloc.h
84
pmd_set(pmd, pte);
arch/m68k/include/asm/motorola_pgtable.h
100
#define __pte_page(pte) ((unsigned long)__va(pte_val(pte) & PAGE_MASK))
arch/m68k/include/asm/motorola_pgtable.h
105
#define pte_none(pte) (!pte_val(pte))
arch/m68k/include/asm/motorola_pgtable.h
106
#define pte_present(pte) (pte_val(pte) & (_PAGE_PRESENT | _PAGE_PROTNONE))
arch/m68k/include/asm/motorola_pgtable.h
110
#define pte_page(pte) virt_to_page(__va(pte_val(pte)))
arch/m68k/include/asm/motorola_pgtable.h
111
#define pte_pfn(pte) (pte_val(pte) >> PAGE_SHIFT)
arch/m68k/include/asm/motorola_pgtable.h
146
static inline int pte_write(pte_t pte) { return !(pte_val(pte) & _PAGE_RONLY); }
arch/m68k/include/asm/motorola_pgtable.h
147
static inline int pte_dirty(pte_t pte) { return pte_val(pte) & _PAGE_DIRTY; }
arch/m68k/include/asm/motorola_pgtable.h
148
static inline int pte_young(pte_t pte) { return pte_val(pte) & _PAGE_ACCESSED; }
arch/m68k/include/asm/motorola_pgtable.h
150
static inline pte_t pte_wrprotect(pte_t pte) { pte_val(pte) |= _PAGE_RONLY; return pte; }
arch/m68k/include/asm/motorola_pgtable.h
151
static inline pte_t pte_mkclean(pte_t pte) { pte_val(pte) &= ~_PAGE_DIRTY; return pte; }
arch/m68k/include/asm/motorola_pgtable.h
152
static inline pte_t pte_mkold(pte_t pte) { pte_val(pte) &= ~_PAGE_ACCESSED; return pte; }
arch/m68k/include/asm/motorola_pgtable.h
153
static inline pte_t pte_mkwrite_novma(pte_t pte){ pte_val(pte) &= ~_PAGE_RONLY; return pte; }
arch/m68k/include/asm/motorola_pgtable.h
154
static inline pte_t pte_mkdirty(pte_t pte) { pte_val(pte) |= _PAGE_DIRTY; return pte; }
arch/m68k/include/asm/motorola_pgtable.h
155
static inline pte_t pte_mkyoung(pte_t pte) { pte_val(pte) |= _PAGE_ACCESSED; return pte; }
arch/m68k/include/asm/motorola_pgtable.h
156
static inline pte_t pte_mknocache(pte_t pte)
arch/m68k/include/asm/motorola_pgtable.h
158
pte_val(pte) = (pte_val(pte) & _CACHEMASK040) | m68k_pgtable_cachemode;
arch/m68k/include/asm/motorola_pgtable.h
159
return pte;
arch/m68k/include/asm/motorola_pgtable.h
161
static inline pte_t pte_mkcache(pte_t pte)
arch/m68k/include/asm/motorola_pgtable.h
163
pte_val(pte) = (pte_val(pte) & _CACHEMASK040) | m68k_supervisor_cachemode;
arch/m68k/include/asm/motorola_pgtable.h
164
return pte;
arch/m68k/include/asm/motorola_pgtable.h
185
#define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) })
arch/m68k/include/asm/motorola_pgtable.h
188
static inline bool pte_swp_exclusive(pte_t pte)
arch/m68k/include/asm/motorola_pgtable.h
190
return pte_val(pte) & _PAGE_SWP_EXCLUSIVE;
arch/m68k/include/asm/motorola_pgtable.h
193
static inline pte_t pte_swp_mkexclusive(pte_t pte)
arch/m68k/include/asm/motorola_pgtable.h
195
pte_val(pte) |= _PAGE_SWP_EXCLUSIVE;
arch/m68k/include/asm/motorola_pgtable.h
196
return pte;
arch/m68k/include/asm/motorola_pgtable.h
199
static inline pte_t pte_swp_clear_exclusive(pte_t pte)
arch/m68k/include/asm/motorola_pgtable.h
201
pte_val(pte) &= ~_PAGE_SWP_EXCLUSIVE;
arch/m68k/include/asm/motorola_pgtable.h
202
return pte;
arch/m68k/include/asm/motorola_pgtable.h
84
static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
arch/m68k/include/asm/motorola_pgtable.h
86
pte_val(pte) = (pte_val(pte) & _PAGE_CHG_MASK) | pgprot_val(newprot);
arch/m68k/include/asm/motorola_pgtable.h
87
return pte;
arch/m68k/include/asm/page.h
24
typedef struct { unsigned long pte; } pte_t;
arch/m68k/include/asm/page.h
39
#define pte_val(x) ((x).pte)
arch/m68k/include/asm/sun3_pgalloc.h
20
#define __pte_free_tlb(tlb, pte, addr) \
arch/m68k/include/asm/sun3_pgalloc.h
21
tlb_remove_ptdesc((tlb), page_ptdesc(pte))
arch/m68k/include/asm/sun3_pgalloc.h
23
static inline void pmd_populate_kernel(struct mm_struct *mm, pmd_t *pmd, pte_t *pte)
arch/m68k/include/asm/sun3_pgalloc.h
25
pmd_val(*pmd) = __pa((unsigned long)pte);
arch/m68k/include/asm/sun3_pgtable.h
103
#define pte_pfn(pte) (pte_val(pte) & SUN3_PAGE_PGNUM_MASK)
arch/m68k/include/asm/sun3_pgtable.h
107
#define pte_page(pte) virt_to_page(__pte_page(pte))
arch/m68k/include/asm/sun3_pgtable.h
134
static inline int pte_write(pte_t pte) { return pte_val(pte) & SUN3_PAGE_WRITEABLE; }
arch/m68k/include/asm/sun3_pgtable.h
135
static inline int pte_dirty(pte_t pte) { return pte_val(pte) & SUN3_PAGE_MODIFIED; }
arch/m68k/include/asm/sun3_pgtable.h
136
static inline int pte_young(pte_t pte) { return pte_val(pte) & SUN3_PAGE_ACCESSED; }
arch/m68k/include/asm/sun3_pgtable.h
138
static inline pte_t pte_wrprotect(pte_t pte) { pte_val(pte) &= ~SUN3_PAGE_WRITEABLE; return pte; }
arch/m68k/include/asm/sun3_pgtable.h
139
static inline pte_t pte_mkclean(pte_t pte) { pte_val(pte) &= ~SUN3_PAGE_MODIFIED; return pte; }
arch/m68k/include/asm/sun3_pgtable.h
140
static inline pte_t pte_mkold(pte_t pte) { pte_val(pte) &= ~SUN3_PAGE_ACCESSED; return pte; }
arch/m68k/include/asm/sun3_pgtable.h
141
static inline pte_t pte_mkwrite_novma(pte_t pte){ pte_val(pte) |= SUN3_PAGE_WRITEABLE; return pte; }
arch/m68k/include/asm/sun3_pgtable.h
142
static inline pte_t pte_mkdirty(pte_t pte) { pte_val(pte) |= SUN3_PAGE_MODIFIED; return pte; }
arch/m68k/include/asm/sun3_pgtable.h
143
static inline pte_t pte_mkyoung(pte_t pte) { pte_val(pte) |= SUN3_PAGE_ACCESSED; return pte; }
arch/m68k/include/asm/sun3_pgtable.h
144
static inline pte_t pte_mknocache(pte_t pte) { pte_val(pte) |= SUN3_PAGE_NOCACHE; return pte; }
arch/m68k/include/asm/sun3_pgtable.h
148
static inline pte_t pte_mkcache(pte_t pte) { return pte; }
arch/m68k/include/asm/sun3_pgtable.h
169
#define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) })
arch/m68k/include/asm/sun3_pgtable.h
172
static inline bool pte_swp_exclusive(pte_t pte)
arch/m68k/include/asm/sun3_pgtable.h
174
return pte_val(pte) & _PAGE_SWP_EXCLUSIVE;
arch/m68k/include/asm/sun3_pgtable.h
177
static inline pte_t pte_swp_mkexclusive(pte_t pte)
arch/m68k/include/asm/sun3_pgtable.h
179
pte_val(pte) |= _PAGE_SWP_EXCLUSIVE;
arch/m68k/include/asm/sun3_pgtable.h
180
return pte;
arch/m68k/include/asm/sun3_pgtable.h
183
static inline pte_t pte_swp_clear_exclusive(pte_t pte)
arch/m68k/include/asm/sun3_pgtable.h
185
pte_val(pte) &= ~_PAGE_SWP_EXCLUSIVE;
arch/m68k/include/asm/sun3_pgtable.h
186
return pte;
arch/m68k/include/asm/sun3_pgtable.h
79
static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
arch/m68k/include/asm/sun3_pgtable.h
81
pte_val(pte) = (pte_val(pte) & SUN3_PAGE_CHG_MASK) | pgprot_val(newprot);
arch/m68k/include/asm/sun3_pgtable.h
82
return pte;
arch/m68k/include/asm/sun3_pgtable.h
87
#define __pte_page(pte) \
arch/m68k/include/asm/sun3_pgtable.h
88
(__va ((pte_val (pte) & SUN3_PAGE_PGNUM_MASK) << PAGE_SHIFT))
arch/m68k/include/asm/sun3_pgtable.h
95
static inline int pte_none (pte_t pte) { return !pte_val (pte); }
arch/m68k/include/asm/sun3_pgtable.h
96
static inline int pte_present (pte_t pte) { return pte_val (pte) & SUN3_PAGE_VALID; }
arch/m68k/kernel/sys_m68k.c
473
pte_t *pte;
arch/m68k/kernel/sys_m68k.c
490
pte = pte_offset_map_lock(mm, pmd, (unsigned long)mem, &ptl);
arch/m68k/kernel/sys_m68k.c
491
if (!pte)
arch/m68k/kernel/sys_m68k.c
493
if (!pte_present(*pte) || !pte_dirty(*pte)
arch/m68k/kernel/sys_m68k.c
494
|| !pte_write(*pte)) {
arch/m68k/kernel/sys_m68k.c
495
pte_unmap_unlock(pte, ptl);
arch/m68k/kernel/sys_m68k.c
507
pte_unmap_unlock(pte, ptl);
arch/m68k/mm/mcfmmu.c
114
pte = (KMAPAREA(mmuar)) ? pte_offset_kernel(pmd, mmuar)
arch/m68k/mm/mcfmmu.c
116
if (!pte || pte_none(*pte) || !pte_present(*pte))
arch/m68k/mm/mcfmmu.c
120
if (!pte_write(*pte))
arch/m68k/mm/mcfmmu.c
122
set_pte(pte, pte_mkdirty(*pte));
arch/m68k/mm/mcfmmu.c
125
set_pte(pte, pte_mkyoung(*pte));
arch/m68k/mm/mcfmmu.c
127
if (!pte_dirty(*pte) && !KMAPAREA(mmuar))
arch/m68k/mm/mcfmmu.c
128
set_pte(pte, pte_wrprotect(*pte));
arch/m68k/mm/mcfmmu.c
132
mmutr |= (pte->pte & CF_PAGE_MMUTR_MASK) >> CF_PAGE_MMUTR_SHIFT;
arch/m68k/mm/mcfmmu.c
135
mmu_write(MMUDR, (pte_val(*pte) & PAGE_MASK) |
arch/m68k/mm/mcfmmu.c
136
((pte->pte) & CF_PAGE_MMUDR_MASK) | MMUDR_SZ_8KB | MMUDR_X);
arch/m68k/mm/mcfmmu.c
144
if (pte && !KMAPAREA(mmuar))
arch/m68k/mm/mcfmmu.c
145
pte_unmap(pte);
arch/m68k/mm/mcfmmu.c
64
pte_t pte = pfn_pte(virt_to_pfn((void *)address),
arch/m68k/mm/mcfmmu.c
67
pte_val(pte) = 0;
arch/m68k/mm/mcfmmu.c
69
set_pte(pg_table, pte);
arch/m68k/mm/mcfmmu.c
85
pte_t *pte = NULL;
arch/m68k/mm/sun3kmap.c
30
unsigned long pte;
arch/m68k/mm/sun3kmap.c
34
pte = pte_val(ptep);
arch/m68k/mm/sun3kmap.c
35
pte |= type;
arch/m68k/mm/sun3kmap.c
37
sun3_put_pte(virt, pte);
arch/m68k/mm/sun3mmu.c
71
pte_t pte = pfn_pte(virt_to_pfn((void *)address), PAGE_INIT);
arch/m68k/mm/sun3mmu.c
73
pte_val (pte) = 0;
arch/m68k/mm/sun3mmu.c
74
set_pte (pg_table, pte);
arch/m68k/sun3/dvma.c
25
unsigned long pte;
arch/m68k/sun3/dvma.c
33
pte = pte_val(ptep);
arch/m68k/sun3/dvma.c
35
if(ptelist[(vaddr & 0xff000) >> PAGE_SHIFT] != pte) {
arch/m68k/sun3/dvma.c
36
sun3_put_pte(vaddr, pte);
arch/m68k/sun3/dvma.c
37
ptelist[(vaddr & 0xff000) >> PAGE_SHIFT] = pte;
arch/m68k/sun3/mmu_emu.c
359
pte_t *pte;
arch/m68k/sun3/mmu_emu.c
385
pte = (pte_t *) pgd_val (*(crp + segment));
arch/m68k/sun3/mmu_emu.c
388
if (!pte) {
arch/m68k/sun3/mmu_emu.c
393
pte = (pte_t *) __va ((unsigned long)(pte + offset));
arch/m68k/sun3/mmu_emu.c
396
if (!(pte_val (*pte) & SUN3_PAGE_VALID))
arch/m68k/sun3/mmu_emu.c
404
sun3_put_pte (vaddr&PAGE_MASK, pte_val (*pte));
arch/m68k/sun3/mmu_emu.c
412
if (pte_val (*pte) & SUN3_PAGE_WRITEABLE)
arch/m68k/sun3/mmu_emu.c
413
pte_val (*pte) |= (SUN3_PAGE_ACCESSED
arch/m68k/sun3/mmu_emu.c
418
pte_val (*pte) |= SUN3_PAGE_ACCESSED;
arch/m68k/sun3/mmu_emu.c
72
static void print_pte(pte_t pte)
arch/m68k/sun3/mmu_emu.c
76
unsigned long val = pte_val (pte);
arch/m68k/sun3/mmu_emu.c
94
unsigned long val = pte_val (pte);
arch/m68k/sun3x/dvma.c
112
pte_t *pte;
arch/m68k/sun3x/dvma.c
115
if((pte = pte_alloc_kernel(pmd, vaddr)) == NULL) {
arch/m68k/sun3x/dvma.c
128
set_pte(pte, pfn_pte(virt_to_pfn((void *)kaddr),
arch/m68k/sun3x/dvma.c
130
pte++;
arch/microblaze/include/asm/page.h
55
typedef struct { unsigned long pte; } pte_t;
arch/microblaze/include/asm/page.h
60
# define pte_val(x) ((x).pte)
arch/microblaze/include/asm/pgalloc.h
28
#define __pte_free_tlb(tlb, pte, addr) pte_free((tlb)->mm, (pte))
arch/microblaze/include/asm/pgalloc.h
30
#define pmd_populate(mm, pmd, pte) \
arch/microblaze/include/asm/pgalloc.h
31
(pmd_val(*(pmd)) = (unsigned long)page_address(pte))
arch/microblaze/include/asm/pgalloc.h
33
#define pmd_populate_kernel(mm, pmd, pte) \
arch/microblaze/include/asm/pgalloc.h
34
(pmd_val(*(pmd)) = (unsigned long) (pte))
arch/microblaze/include/asm/pgtable.h
220
#define pte_none(pte) ((pte_val(pte) & ~_PTE_NONE_MASK) == 0)
arch/microblaze/include/asm/pgtable.h
221
#define pte_present(pte) (pte_val(pte) & _PAGE_PRESENT)
arch/microblaze/include/asm/pgtable.h
244
static inline int pte_read(pte_t pte) { return pte_val(pte) & _PAGE_USER; }
arch/microblaze/include/asm/pgtable.h
245
static inline int pte_write(pte_t pte) { return pte_val(pte) & _PAGE_RW; }
arch/microblaze/include/asm/pgtable.h
246
static inline int pte_exec(pte_t pte) { return pte_val(pte) & _PAGE_EXEC; }
arch/microblaze/include/asm/pgtable.h
247
static inline int pte_dirty(pte_t pte) { return pte_val(pte) & _PAGE_DIRTY; }
arch/microblaze/include/asm/pgtable.h
248
static inline int pte_young(pte_t pte) { return pte_val(pte) & _PAGE_ACCESSED; }
arch/microblaze/include/asm/pgtable.h
250
static inline void pte_uncache(pte_t pte) { pte_val(pte) |= _PAGE_NO_CACHE; }
arch/microblaze/include/asm/pgtable.h
251
static inline void pte_cache(pte_t pte) { pte_val(pte) &= ~_PAGE_NO_CACHE; }
arch/microblaze/include/asm/pgtable.h
253
static inline pte_t pte_rdprotect(pte_t pte) \
arch/microblaze/include/asm/pgtable.h
254
{ pte_val(pte) &= ~_PAGE_USER; return pte; }
arch/microblaze/include/asm/pgtable.h
255
static inline pte_t pte_wrprotect(pte_t pte) \
arch/microblaze/include/asm/pgtable.h
256
{ pte_val(pte) &= ~(_PAGE_RW | _PAGE_HWWRITE); return pte; }
arch/microblaze/include/asm/pgtable.h
257
static inline pte_t pte_exprotect(pte_t pte) \
arch/microblaze/include/asm/pgtable.h
258
{ pte_val(pte) &= ~_PAGE_EXEC; return pte; }
arch/microblaze/include/asm/pgtable.h
259
static inline pte_t pte_mkclean(pte_t pte) \
arch/microblaze/include/asm/pgtable.h
260
{ pte_val(pte) &= ~(_PAGE_DIRTY | _PAGE_HWWRITE); return pte; }
arch/microblaze/include/asm/pgtable.h
261
static inline pte_t pte_mkold(pte_t pte) \
arch/microblaze/include/asm/pgtable.h
262
{ pte_val(pte) &= ~_PAGE_ACCESSED; return pte; }
arch/microblaze/include/asm/pgtable.h
264
static inline pte_t pte_mkread(pte_t pte) \
arch/microblaze/include/asm/pgtable.h
265
{ pte_val(pte) |= _PAGE_USER; return pte; }
arch/microblaze/include/asm/pgtable.h
266
static inline pte_t pte_mkexec(pte_t pte) \
arch/microblaze/include/asm/pgtable.h
267
{ pte_val(pte) |= _PAGE_USER | _PAGE_EXEC; return pte; }
arch/microblaze/include/asm/pgtable.h
268
static inline pte_t pte_mkwrite_novma(pte_t pte) \
arch/microblaze/include/asm/pgtable.h
269
{ pte_val(pte) |= _PAGE_RW; return pte; }
arch/microblaze/include/asm/pgtable.h
270
static inline pte_t pte_mkdirty(pte_t pte) \
arch/microblaze/include/asm/pgtable.h
271
{ pte_val(pte) |= _PAGE_DIRTY; return pte; }
arch/microblaze/include/asm/pgtable.h
272
static inline pte_t pte_mkyoung(pte_t pte) \
arch/microblaze/include/asm/pgtable.h
273
{ pte_val(pte) |= _PAGE_ACCESSED; return pte; }
arch/microblaze/include/asm/pgtable.h
282
pte_t pte;
arch/microblaze/include/asm/pgtable.h
283
pte_val(pte) = physpage | pgprot_val(pgprot);
arch/microblaze/include/asm/pgtable.h
284
return pte;
arch/microblaze/include/asm/pgtable.h
287
static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
arch/microblaze/include/asm/pgtable.h
289
pte_val(pte) = (pte_val(pte) & _PAGE_CHG_MASK) | pgprot_val(newprot);
arch/microblaze/include/asm/pgtable.h
290
return pte;
arch/microblaze/include/asm/pgtable.h
324
static inline void set_pte(pte_t *ptep, pte_t pte)
arch/microblaze/include/asm/pgtable.h
326
*ptep = pte;
arch/microblaze/include/asm/pgtable.h
397
#define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) >> 2 })
arch/microblaze/include/asm/pgtable.h
400
static inline bool pte_swp_exclusive(pte_t pte)
arch/microblaze/include/asm/pgtable.h
402
return pte_val(pte) & _PAGE_SWP_EXCLUSIVE;
arch/microblaze/include/asm/pgtable.h
405
static inline pte_t pte_swp_mkexclusive(pte_t pte)
arch/microblaze/include/asm/pgtable.h
407
pte_val(pte) |= _PAGE_SWP_EXCLUSIVE;
arch/microblaze/include/asm/pgtable.h
408
return pte;
arch/microblaze/include/asm/pgtable.h
411
static inline pte_t pte_swp_clear_exclusive(pte_t pte)
arch/microblaze/include/asm/pgtable.h
413
pte_val(pte) &= ~_PAGE_SWP_EXCLUSIVE;
arch/microblaze/include/asm/pgtable.h
414
return pte;
arch/microblaze/include/asm/tlbflush.h
37
#define update_mmu_cache(vma, addr, pte) \
arch/microblaze/mm/pgtable.c
201
pte_t *pte;
arch/microblaze/mm/pgtable.c
210
pte = pte_offset_kernel(pmd, addr & PAGE_MASK);
arch/microblaze/mm/pgtable.c
211
if (pte) {
arch/microblaze/mm/pgtable.c
213
*ptep = pte;
arch/microblaze/mm/pgtable.c
227
pte_t *pte;
arch/microblaze/mm/pgtable.c
239
if (get_pteptr(mm, addr, &pte))
arch/microblaze/mm/pgtable.c
240
pa = (pte_val(*pte) & PAGE_MASK) | (addr & ~PAGE_MASK);
arch/mips/include/asm/hugetlb.h
20
pte_t pte = *ptep;
arch/mips/include/asm/hugetlb.h
24
return pte;
arch/mips/include/asm/hugetlb.h
31
pte_t pte;
arch/mips/include/asm/hugetlb.h
39
pte = huge_ptep_get_and_clear(vma->vm_mm, addr, ptep, sz);
arch/mips/include/asm/hugetlb.h
41
return pte;
arch/mips/include/asm/hugetlb.h
45
static inline int huge_pte_none(pte_t pte)
arch/mips/include/asm/hugetlb.h
47
unsigned long val = pte_val(pte) & ~_PAGE_GLOBAL;
arch/mips/include/asm/hugetlb.h
54
pte_t *ptep, pte_t pte,
arch/mips/include/asm/hugetlb.h
57
int changed = !pte_same(*ptep, pte);
arch/mips/include/asm/hugetlb.h
60
set_pte_at(vma->vm_mm, addr, ptep, pte);
arch/mips/include/asm/page.h
110
typedef struct { unsigned long long pte; } pte_t;
arch/mips/include/asm/page.h
111
#define pte_val(x) ((x).pte)
arch/mips/include/asm/page.h
115
typedef struct { unsigned long pte; } pte_t;
arch/mips/include/asm/page.h
116
#define pte_val(x) ((x).pte)
arch/mips/include/asm/pgalloc.h
21
pte_t *pte)
arch/mips/include/asm/pgalloc.h
23
set_pmd(pmd, __pmd((unsigned long)pte));
arch/mips/include/asm/pgalloc.h
27
pgtable_t pte)
arch/mips/include/asm/pgalloc.h
29
set_pmd(pmd, __pmd((unsigned long)page_address(pte)));
arch/mips/include/asm/pgalloc.h
51
#define __pte_free_tlb(tlb, pte, address) tlb_remove_ptdesc((tlb), page_ptdesc(pte))
arch/mips/include/asm/pgtable-32.h
160
pte_t pte;
arch/mips/include/asm/pgtable-32.h
162
pte.pte_low = (pfn >> _PAGE_PRESENT_SHIFT) |
arch/mips/include/asm/pgtable-32.h
164
pte.pte_high = (pfn << PFN_PTE_SHIFT) |
arch/mips/include/asm/pgtable-32.h
166
return pte;
arch/mips/include/asm/pgtable-32.h
176
pte_t pte;
arch/mips/include/asm/pgtable-32.h
178
pte.pte_high = (pfn << 6) | (pgprot_val(prot) & 0x3f);
arch/mips/include/asm/pgtable-32.h
179
pte.pte_low = pgprot_val(prot);
arch/mips/include/asm/pgtable-32.h
181
return pte;
arch/mips/include/asm/pgtable-32.h
187
#define pte_pfn(x) ((unsigned long)((x).pte >> PFN_PTE_SHIFT))
arch/mips/include/asm/pgtable-32.h
214
#define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) })
arch/mips/include/asm/pgtable-32.h
242
#define __pte_to_swp_entry(pte) ((swp_entry_t) { (pte).pte_high })
arch/mips/include/asm/pgtable-32.h
271
#define __pte_to_swp_entry(pte) ((swp_entry_t) { (pte).pte_high })
arch/mips/include/asm/pgtable-32.h
295
#define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) })
arch/mips/include/asm/pgtable-64.h
301
#define pte_pfn(x) ((unsigned long)((x).pte >> PFN_PTE_SHIFT))
arch/mips/include/asm/pgtable-64.h
341
{ pte_t pte; pte_val(pte) = ((type & 0x7f) << 16) | (offset << 24); return pte; }
arch/mips/include/asm/pgtable-64.h
346
#define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) })
arch/mips/include/asm/pgtable.h
111
# define pte_none(pte) (!(((pte).pte_high) & ~_PAGE_GLOBAL))
arch/mips/include/asm/pgtable.h
113
# define pte_none(pte) (!(((pte).pte_low | (pte).pte_high) & ~_PAGE_GLOBAL))
arch/mips/include/asm/pgtable.h
116
#define pte_present(pte) ((pte).pte_low & _PAGE_PRESENT)
arch/mips/include/asm/pgtable.h
117
#define pte_no_exec(pte) ((pte).pte_low & _PAGE_NO_EXEC)
arch/mips/include/asm/pgtable.h
119
static inline void set_pte(pte_t *ptep, pte_t pte)
arch/mips/include/asm/pgtable.h
121
ptep->pte_high = pte.pte_high;
arch/mips/include/asm/pgtable.h
123
ptep->pte_low = pte.pte_low;
arch/mips/include/asm/pgtable.h
126
if (pte.pte_high & _PAGE_GLOBAL) {
arch/mips/include/asm/pgtable.h
128
if (pte.pte_low & _PAGE_GLOBAL) {
arch/mips/include/asm/pgtable.h
162
#define pte_none(pte) (!(pte_val(pte) & ~_PAGE_GLOBAL))
arch/mips/include/asm/pgtable.h
163
#define pte_present(pte) (pte_val(pte) & _PAGE_PRESENT)
arch/mips/include/asm/pgtable.h
164
#define pte_no_exec(pte) (pte_val(pte) & _PAGE_NO_EXEC)
arch/mips/include/asm/pgtable.h
182
cmpxchg64(&buddy->pte, 0, _PAGE_GLOBAL);
arch/mips/include/asm/pgtable.h
184
cmpxchg(&buddy->pte, 0, _PAGE_GLOBAL);
arch/mips/include/asm/pgtable.h
205
pte_t *ptep, pte_t pte, unsigned int nr)
arch/mips/include/asm/pgtable.h
211
if (!pte_present(pte))
arch/mips/include/asm/pgtable.h
214
(pte_pfn(ptep[i]) == pte_pfn(pte)))
arch/mips/include/asm/pgtable.h
220
__update_cache(addr, pte);
arch/mips/include/asm/pgtable.h
223
set_pte(ptep, pte);
arch/mips/include/asm/pgtable.h
227
pte = __pte(pte_val(pte) + (1UL << PFN_PTE_SHIFT));
arch/mips/include/asm/pgtable.h
262
static inline int pte_special(pte_t pte)
arch/mips/include/asm/pgtable.h
264
return pte.pte_low & _PAGE_SPECIAL;
arch/mips/include/asm/pgtable.h
267
static inline pte_t pte_mkspecial(pte_t pte)
arch/mips/include/asm/pgtable.h
269
pte.pte_low |= _PAGE_SPECIAL;
arch/mips/include/asm/pgtable.h
270
return pte;
arch/mips/include/asm/pgtable.h
273
static inline int pte_special(pte_t pte)
arch/mips/include/asm/pgtable.h
275
return pte_val(pte) & _PAGE_SPECIAL;
arch/mips/include/asm/pgtable.h
278
static inline pte_t pte_mkspecial(pte_t pte)
arch/mips/include/asm/pgtable.h
280
pte_val(pte) |= _PAGE_SPECIAL;
arch/mips/include/asm/pgtable.h
281
return pte;
arch/mips/include/asm/pgtable.h
291
static inline int pte_write(pte_t pte) { return pte.pte_low & _PAGE_WRITE; }
arch/mips/include/asm/pgtable.h
292
static inline int pte_dirty(pte_t pte) { return pte.pte_low & _PAGE_MODIFIED; }
arch/mips/include/asm/pgtable.h
293
static inline int pte_young(pte_t pte) { return pte.pte_low & _PAGE_ACCESSED; }
arch/mips/include/asm/pgtable.h
295
static inline pte_t pte_wrprotect(pte_t pte)
arch/mips/include/asm/pgtable.h
297
pte.pte_low &= ~_PAGE_WRITE;
arch/mips/include/asm/pgtable.h
299
pte.pte_low &= ~_PAGE_SILENT_WRITE;
arch/mips/include/asm/pgtable.h
300
pte.pte_high &= ~_PAGE_SILENT_WRITE;
arch/mips/include/asm/pgtable.h
301
return pte;
arch/mips/include/asm/pgtable.h
304
static inline pte_t pte_mkclean(pte_t pte)
arch/mips/include/asm/pgtable.h
306
pte.pte_low &= ~_PAGE_MODIFIED;
arch/mips/include/asm/pgtable.h
308
pte.pte_low &= ~_PAGE_SILENT_WRITE;
arch/mips/include/asm/pgtable.h
309
pte.pte_high &= ~_PAGE_SILENT_WRITE;
arch/mips/include/asm/pgtable.h
310
return pte;
arch/mips/include/asm/pgtable.h
313
static inline pte_t pte_mkold(pte_t pte)
arch/mips/include/asm/pgtable.h
315
pte.pte_low &= ~_PAGE_ACCESSED;
arch/mips/include/asm/pgtable.h
317
pte.pte_low &= ~_PAGE_SILENT_READ;
arch/mips/include/asm/pgtable.h
318
pte.pte_high &= ~_PAGE_SILENT_READ;
arch/mips/include/asm/pgtable.h
319
return pte;
arch/mips/include/asm/pgtable.h
322
static inline pte_t pte_mkwrite_novma(pte_t pte)
arch/mips/include/asm/pgtable.h
324
pte.pte_low |= _PAGE_WRITE;
arch/mips/include/asm/pgtable.h
325
if (pte.pte_low & _PAGE_MODIFIED) {
arch/mips/include/asm/pgtable.h
327
pte.pte_low |= _PAGE_SILENT_WRITE;
arch/mips/include/asm/pgtable.h
328
pte.pte_high |= _PAGE_SILENT_WRITE;
arch/mips/include/asm/pgtable.h
330
return pte;
arch/mips/include/asm/pgtable.h
333
static inline pte_t pte_mkdirty(pte_t pte)
arch/mips/include/asm/pgtable.h
335
pte.pte_low |= _PAGE_MODIFIED;
arch/mips/include/asm/pgtable.h
336
if (pte.pte_low & _PAGE_WRITE) {
arch/mips/include/asm/pgtable.h
338
pte.pte_low |= _PAGE_SILENT_WRITE;
arch/mips/include/asm/pgtable.h
339
pte.pte_high |= _PAGE_SILENT_WRITE;
arch/mips/include/asm/pgtable.h
341
return pte;
arch/mips/include/asm/pgtable.h
344
static inline pte_t pte_mkyoung(pte_t pte)
arch/mips/include/asm/pgtable.h
346
pte.pte_low |= _PAGE_ACCESSED;
arch/mips/include/asm/pgtable.h
347
if (!(pte.pte_low & _PAGE_NO_READ)) {
arch/mips/include/asm/pgtable.h
349
pte.pte_low |= _PAGE_SILENT_READ;
arch/mips/include/asm/pgtable.h
350
pte.pte_high |= _PAGE_SILENT_READ;
arch/mips/include/asm/pgtable.h
352
return pte;
arch/mips/include/asm/pgtable.h
355
static inline int pte_write(pte_t pte) { return pte_val(pte) & _PAGE_WRITE; }
arch/mips/include/asm/pgtable.h
356
static inline int pte_dirty(pte_t pte) { return pte_val(pte) & _PAGE_MODIFIED; }
arch/mips/include/asm/pgtable.h
357
static inline int pte_young(pte_t pte) { return pte_val(pte) & _PAGE_ACCESSED; }
arch/mips/include/asm/pgtable.h
359
static inline pte_t pte_wrprotect(pte_t pte)
arch/mips/include/asm/pgtable.h
361
pte_val(pte) &= ~(_PAGE_WRITE | _PAGE_SILENT_WRITE);
arch/mips/include/asm/pgtable.h
362
return pte;
arch/mips/include/asm/pgtable.h
365
static inline pte_t pte_mkclean(pte_t pte)
arch/mips/include/asm/pgtable.h
367
pte_val(pte) &= ~(_PAGE_MODIFIED | _PAGE_SILENT_WRITE);
arch/mips/include/asm/pgtable.h
368
return pte;
arch/mips/include/asm/pgtable.h
371
static inline pte_t pte_mkold(pte_t pte)
arch/mips/include/asm/pgtable.h
373
pte_val(pte) &= ~(_PAGE_ACCESSED | _PAGE_SILENT_READ);
arch/mips/include/asm/pgtable.h
374
return pte;
arch/mips/include/asm/pgtable.h
377
static inline pte_t pte_mkwrite_novma(pte_t pte)
arch/mips/include/asm/pgtable.h
379
pte_val(pte) |= _PAGE_WRITE;
arch/mips/include/asm/pgtable.h
380
if (pte_val(pte) & _PAGE_MODIFIED)
arch/mips/include/asm/pgtable.h
381
pte_val(pte) |= _PAGE_SILENT_WRITE;
arch/mips/include/asm/pgtable.h
382
return pte;
arch/mips/include/asm/pgtable.h
385
static inline pte_t pte_mkdirty(pte_t pte)
arch/mips/include/asm/pgtable.h
387
pte_val(pte) |= _PAGE_MODIFIED | _PAGE_SOFT_DIRTY;
arch/mips/include/asm/pgtable.h
388
if (pte_val(pte) & _PAGE_WRITE)
arch/mips/include/asm/pgtable.h
389
pte_val(pte) |= _PAGE_SILENT_WRITE;
arch/mips/include/asm/pgtable.h
390
return pte;
arch/mips/include/asm/pgtable.h
393
static inline pte_t pte_mkyoung(pte_t pte)
arch/mips/include/asm/pgtable.h
395
pte_val(pte) |= _PAGE_ACCESSED;
arch/mips/include/asm/pgtable.h
396
if (!(pte_val(pte) & _PAGE_NO_READ))
arch/mips/include/asm/pgtable.h
397
pte_val(pte) |= _PAGE_SILENT_READ;
arch/mips/include/asm/pgtable.h
398
return pte;
arch/mips/include/asm/pgtable.h
404
static inline int pte_huge(pte_t pte) { return pte_val(pte) & _PAGE_HUGE; }
arch/mips/include/asm/pgtable.h
406
static inline pte_t pte_mkhuge(pte_t pte)
arch/mips/include/asm/pgtable.h
408
pte_val(pte) |= _PAGE_HUGE;
arch/mips/include/asm/pgtable.h
409
return pte;
arch/mips/include/asm/pgtable.h
428
static inline bool pte_soft_dirty(pte_t pte)
arch/mips/include/asm/pgtable.h
430
return pte_val(pte) & _PAGE_SOFT_DIRTY;
arch/mips/include/asm/pgtable.h
434
static inline pte_t pte_mksoft_dirty(pte_t pte)
arch/mips/include/asm/pgtable.h
436
pte_val(pte) |= _PAGE_SOFT_DIRTY;
arch/mips/include/asm/pgtable.h
437
return pte;
arch/mips/include/asm/pgtable.h
441
static inline pte_t pte_clear_soft_dirty(pte_t pte)
arch/mips/include/asm/pgtable.h
443
pte_val(pte) &= ~(_PAGE_SOFT_DIRTY);
arch/mips/include/asm/pgtable.h
444
return pte;
arch/mips/include/asm/pgtable.h
45
extern void __update_cache(unsigned long address, pte_t pte);
arch/mips/include/asm/pgtable.h
508
static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
arch/mips/include/asm/pgtable.h
510
pte.pte_low &= (_PAGE_MODIFIED | _PAGE_ACCESSED | _PFNX_MASK);
arch/mips/include/asm/pgtable.h
511
pte.pte_high &= (_PFN_MASK | _CACHE_MASK);
arch/mips/include/asm/pgtable.h
512
pte.pte_low |= pgprot_val(newprot) & ~_PFNX_MASK;
arch/mips/include/asm/pgtable.h
513
pte.pte_high |= pgprot_val(newprot) & ~(_PFN_MASK | _CACHE_MASK);
arch/mips/include/asm/pgtable.h
514
return pte;
arch/mips/include/asm/pgtable.h
517
static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
arch/mips/include/asm/pgtable.h
519
pte.pte_low &= _PAGE_CHG_MASK;
arch/mips/include/asm/pgtable.h
520
pte.pte_high &= (_PFN_MASK | _CACHE_MASK);
arch/mips/include/asm/pgtable.h
521
pte.pte_low |= pgprot_val(newprot);
arch/mips/include/asm/pgtable.h
522
pte.pte_high |= pgprot_val(newprot) & ~(_PFN_MASK | _CACHE_MASK);
arch/mips/include/asm/pgtable.h
523
return pte;
arch/mips/include/asm/pgtable.h
526
static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
arch/mips/include/asm/pgtable.h
528
pte_val(pte) &= _PAGE_CHG_MASK;
arch/mips/include/asm/pgtable.h
529
pte_val(pte) |= pgprot_val(newprot) & ~_PAGE_CHG_MASK;
arch/mips/include/asm/pgtable.h
530
if ((pte_val(pte) & _PAGE_ACCESSED) && !(pte_val(pte) & _PAGE_NO_READ))
arch/mips/include/asm/pgtable.h
531
pte_val(pte) |= _PAGE_SILENT_READ;
arch/mips/include/asm/pgtable.h
532
return pte;
arch/mips/include/asm/pgtable.h
537
static inline bool pte_swp_exclusive(pte_t pte)
arch/mips/include/asm/pgtable.h
539
return pte.pte_low & _PAGE_SWP_EXCLUSIVE;
arch/mips/include/asm/pgtable.h
542
static inline pte_t pte_swp_mkexclusive(pte_t pte)
arch/mips/include/asm/pgtable.h
544
pte.pte_low |= _PAGE_SWP_EXCLUSIVE;
arch/mips/include/asm/pgtable.h
545
return pte;
arch/mips/include/asm/pgtable.h
548
static inline pte_t pte_swp_clear_exclusive(pte_t pte)
arch/mips/include/asm/pgtable.h
550
pte.pte_low &= ~_PAGE_SWP_EXCLUSIVE;
arch/mips/include/asm/pgtable.h
551
return pte;
arch/mips/include/asm/pgtable.h
554
static inline bool pte_swp_exclusive(pte_t pte)
arch/mips/include/asm/pgtable.h
556
return pte_val(pte) & _PAGE_SWP_EXCLUSIVE;
arch/mips/include/asm/pgtable.h
559
static inline pte_t pte_swp_mkexclusive(pte_t pte)
arch/mips/include/asm/pgtable.h
561
pte_val(pte) |= _PAGE_SWP_EXCLUSIVE;
arch/mips/include/asm/pgtable.h
562
return pte;
arch/mips/include/asm/pgtable.h
565
static inline pte_t pte_swp_clear_exclusive(pte_t pte)
arch/mips/include/asm/pgtable.h
567
pte_val(pte) &= ~_PAGE_SWP_EXCLUSIVE;
arch/mips/include/asm/pgtable.h
568
return pte;
arch/mips/include/asm/pgtable.h
573
pte_t pte);
arch/mips/include/asm/pgtable.h
580
pte_t pte = *ptep;
arch/mips/include/asm/pgtable.h
581
__update_tlb(vma, address, pte);
arch/mips/include/asm/pgtable.h
597
pte_t pte = *(pte_t *)pmdp;
arch/mips/include/asm/pgtable.h
599
__update_tlb(vma, address, pte);
arch/mips/kvm/mmu.c
154
static bool kvm_mips_flush_gpa_pte(pte_t *pte, unsigned long start_gpa,
arch/mips/kvm/mmu.c
163
if (!pte_present(pte[i]))
arch/mips/kvm/mmu.c
166
set_pte(pte + i, __pte(0));
arch/mips/kvm/mmu.c
174
pte_t *pte;
arch/mips/kvm/mmu.c
185
pte = pte_offset_kernel(pmd + i, 0);
arch/mips/kvm/mmu.c
189
if (kvm_mips_flush_gpa_pte(pte, start_gpa, end)) {
arch/mips/kvm/mmu.c
191
pte_free_kernel(NULL, pte);
arch/mips/kvm/mmu.c
278
static int kvm_mips_##name##_pte(pte_t *pte, unsigned long start, \
arch/mips/kvm/mmu.c
288
if (!pte_present(pte[i])) \
arch/mips/kvm/mmu.c
291
old = pte[i]; \
arch/mips/kvm/mmu.c
295
set_pte(pte + i, new); \
arch/mips/kvm/mmu.c
306
pte_t *pte; \
arch/mips/kvm/mmu.c
316
pte = pte_offset_kernel(pmd + i, 0); \
arch/mips/kvm/mmu.c
320
ret |= kvm_mips_##name##_pte(pte, start, cur_end); \
arch/mips/mm/cache.c
146
void __update_cache(unsigned long address, pte_t pte)
arch/mips/mm/cache.c
150
int exec = !pte_no_exec(pte) && !cpu_has_ic_fills_f_dc;
arch/mips/mm/cache.c
153
pfn = pte_pfn(pte);
arch/mips/mm/hugetlbpage.c
30
pte_t *pte = NULL;
arch/mips/mm/hugetlbpage.c
36
pte = (pte_t *)pmd_alloc(mm, pud, addr);
arch/mips/mm/hugetlbpage.c
38
return pte;
arch/mips/mm/init.c
112
entrylo = (pte.pte_low & _PFNX_MASK);
arch/mips/mm/init.c
234
pte_t *pte;
arch/mips/mm/init.c
250
pte = (pte_t *) memblock_alloc_low(PAGE_SIZE,
arch/mips/mm/init.c
252
if (!pte)
arch/mips/mm/init.c
257
set_pmd(pmd, __pmd((unsigned long)pte));
arch/mips/mm/init.c
258
BUG_ON(pte != pte_offset_kernel(pmd, 0));
arch/mips/mm/init.c
82
pte_t pte;
arch/mips/mm/init.c
92
pte = mk_pte(page, prot);
arch/mips/mm/init.c
94
entrylo = pte_to_entrylo(pte.pte_high);
arch/mips/mm/init.c
96
entrylo = pte.pte_high;
arch/mips/mm/init.c
98
entrylo = pte_to_entrylo(pte_val(pte));
arch/mips/mm/pgtable-32.c
50
pte_t *pte;
arch/mips/mm/pgtable-32.c
76
pte = pte_offset_kernel(pmd, vaddr);
arch/mips/mm/pgtable-32.c
77
pkmap_page_table = pte;
arch/mips/mm/tlb-r3k.c
179
void __update_tlb(struct vm_area_struct *vma, unsigned long address, pte_t pte)
arch/mips/mm/tlb-r3k.c
206
write_c0_entrylo0(pte_val(pte));
arch/mips/mm/tlb-r4k.c
297
void __update_tlb(struct vm_area_struct * vma, unsigned long address, pte_t pte)
arch/mips/mm/tlbex.c
1613
iPTE_LW(u32 **p, unsigned int pte, unsigned int ptr)
arch/mips/mm/tlbex.c
1620
uasm_i_lld(p, pte, 0, ptr);
arch/mips/mm/tlbex.c
1623
UASM_i_LL(p, pte, 0, ptr);
arch/mips/mm/tlbex.c
1627
uasm_i_ld(p, pte, 0, ptr);
arch/mips/mm/tlbex.c
1630
UASM_i_LW(p, pte, 0, ptr);
arch/mips/mm/tlbex.c
1635
iPTE_SW(u32 **p, struct uasm_reloc **r, unsigned int pte, unsigned int ptr,
arch/mips/mm/tlbex.c
1643
uasm_i_or(p, pte, pte, scratch);
arch/mips/mm/tlbex.c
1646
uasm_i_ori(p, pte, pte, mode);
arch/mips/mm/tlbex.c
1652
uasm_i_scd(p, pte, 0, ptr);
arch/mips/mm/tlbex.c
1655
UASM_i_SC(p, pte, 0, ptr);
arch/mips/mm/tlbex.c
1658
uasm_il_beqzl(p, r, pte, label_smp_pgtable_change);
arch/mips/mm/tlbex.c
1660
uasm_il_beqz(p, r, pte, label_smp_pgtable_change);
arch/mips/mm/tlbex.c
1665
uasm_i_ll(p, pte, sizeof(pte_t) / 2, ptr);
arch/mips/mm/tlbex.c
1666
uasm_i_ori(p, pte, pte, hwmode);
arch/mips/mm/tlbex.c
1668
uasm_i_sc(p, pte, sizeof(pte_t) / 2, ptr);
arch/mips/mm/tlbex.c
1669
uasm_il_beqz(p, r, pte, label_smp_pgtable_change);
arch/mips/mm/tlbex.c
1671
uasm_i_lw(p, pte, 0, ptr);
arch/mips/mm/tlbex.c
1680
uasm_i_sd(p, pte, 0, ptr);
arch/mips/mm/tlbex.c
1683
UASM_i_SW(p, pte, 0, ptr);
arch/mips/mm/tlbex.c
1687
uasm_i_lw(p, pte, sizeof(pte_t) / 2, ptr);
arch/mips/mm/tlbex.c
1688
uasm_i_ori(p, pte, pte, hwmode);
arch/mips/mm/tlbex.c
1690
uasm_i_sw(p, pte, sizeof(pte_t) / 2, ptr);
arch/mips/mm/tlbex.c
1691
uasm_i_lw(p, pte, 0, ptr);
arch/mips/mm/tlbex.c
1704
int pte, int ptr, int scratch, enum label_id lid)
arch/mips/mm/tlbex.c
1706
int t = scratch >= 0 ? scratch : pte;
arch/mips/mm/tlbex.c
1707
int cur = pte;
arch/mips/mm/tlbex.c
1711
uasm_il_bbit0(p, r, pte, ilog2(_PAGE_PRESENT), lid);
arch/mips/mm/tlbex.c
1720
if (pte == t)
arch/mips/mm/tlbex.c
1722
iPTE_LW(p, pte, ptr);
arch/mips/mm/tlbex.c
1733
if (pte == t)
arch/mips/mm/tlbex.c
1735
iPTE_LW(p, pte, ptr);
arch/mips/mm/tlbex.c
1741
build_make_valid(u32 **p, struct uasm_reloc **r, unsigned int pte,
arch/mips/mm/tlbex.c
1746
iPTE_SW(p, r, pte, ptr, mode, scratch);
arch/mips/mm/tlbex.c
1755
unsigned int pte, unsigned int ptr, int scratch,
arch/mips/mm/tlbex.c
1758
int t = scratch >= 0 ? scratch : pte;
arch/mips/mm/tlbex.c
1759
int cur = pte;
arch/mips/mm/tlbex.c
1770
if (pte == t)
arch/mips/mm/tlbex.c
1772
iPTE_LW(p, pte, ptr);
arch/mips/mm/tlbex.c
1781
build_make_write(u32 **p, struct uasm_reloc **r, unsigned int pte,
arch/mips/mm/tlbex.c
1787
iPTE_SW(p, r, pte, ptr, mode, scratch);
arch/mips/mm/tlbex.c
1796
unsigned int pte, unsigned int ptr, int scratch,
arch/mips/mm/tlbex.c
1800
uasm_il_bbit0(p, r, pte, ilog2(_PAGE_WRITE), lid);
arch/mips/mm/tlbex.c
1803
int t = scratch >= 0 ? scratch : pte;
arch/mips/mm/tlbex.c
1804
uasm_i_srl(p, t, pte, _PAGE_WRITE_SHIFT);
arch/mips/mm/tlbex.c
1807
if (pte == t)
arch/mips/mm/tlbex.c
1809
iPTE_LW(p, pte, ptr);
arch/mips/mm/tlbex.c
1825
build_r3000_pte_reload_tlbwi(u32 **p, unsigned int pte, unsigned int tmp)
arch/mips/mm/tlbex.c
1827
uasm_i_mtc0(p, pte, C0_ENTRYLO0); /* cp0 delay */
arch/mips/mm/tlbex.c
1842
struct uasm_reloc **r, unsigned int pte,
arch/mips/mm/tlbex.c
1846
uasm_i_mtc0(p, pte, C0_ENTRYLO0); /* cp0 delay */
arch/mips/mm/tlbex.c
1859
build_r3000_tlbchange_handler_head(u32 **p, unsigned int pte,
arch/mips/mm/tlbex.c
1864
uasm_i_mfc0(p, pte, C0_BADVADDR);
arch/mips/mm/tlbex.c
1867
uasm_i_srl(p, pte, pte, 22); /* load delay */
arch/mips/mm/tlbex.c
1868
uasm_i_sll(p, pte, pte, 2);
arch/mips/mm/tlbex.c
1869
uasm_i_addu(p, ptr, ptr, pte);
arch/mips/mm/tlbex.c
1870
uasm_i_mfc0(p, pte, C0_CONTEXT);
arch/mips/mm/tlbex.c
1872
uasm_i_andi(p, pte, pte, 0xffc); /* load delay */
arch/mips/mm/tlbex.c
1873
uasm_i_addu(p, ptr, ptr, pte);
arch/mips/mm/tlbex.c
1874
uasm_i_lw(p, pte, 0, ptr);
arch/mips/mm/tlbex.c
690
static void build_huge_update_entries(u32 **p, unsigned int pte,
arch/mips/mm/tlbex.c
710
build_convert_pte_to_entrylo(p, pte);
arch/mips/mm/tlbex.c
711
UASM_i_MTC0(p, pte, C0_ENTRYLO0); /* load it */
arch/mips/mm/tlbex.c
714
UASM_i_ADDIU(p, pte, pte, HPAGE_SIZE >> 7);
arch/mips/mm/tlbex.c
716
UASM_i_ADDU(p, pte, pte, tmp);
arch/mips/mm/tlbex.c
718
UASM_i_MTC0(p, pte, C0_ENTRYLO1); /* load it */
arch/mips/mm/tlbex.c
723
unsigned int pte,
arch/mips/mm/tlbex.c
728
UASM_i_SC(p, pte, 0, ptr);
arch/mips/mm/tlbex.c
729
uasm_il_beqz(p, r, pte, label_tlb_huge_update);
arch/mips/mm/tlbex.c
730
UASM_i_LW(p, pte, 0, ptr); /* Needed because SC killed our PTE */
arch/mips/mm/tlbex.c
732
UASM_i_SW(p, pte, 0, ptr);
arch/mips/mm/tlbex.c
744
build_huge_update_entries(p, pte, ptr);
arch/mips/mm/tlbex.c
745
build_huge_tlb_write_entry(p, l, r, pte, tlb_random, 0);
arch/mips/mm/tlbex.c
750
build_huge_update_entries(p, pte, ptr);
arch/mips/mm/tlbex.c
751
build_huge_tlb_write_entry(p, l, r, pte, tlb_indexed, 0);
arch/mips/sgi-ip22/ip28-berr.c
316
unsigned long pte = (lo >> 6) << 12; /* PTEBase */
arch/mips/sgi-ip22/ip28-berr.c
317
pte += 8*((vaddr >> pgsz) & 0x1ff);
arch/mips/sgi-ip22/ip28-berr.c
318
if (page_is_ram(PFN_DOWN(pte))) {
arch/mips/sgi-ip22/ip28-berr.c
325
PHYS_TO_XKSEG_UNCACHED(pte);
arch/nios2/include/asm/page.h
57
typedef struct { unsigned long pte; } pte_t;
arch/nios2/include/asm/page.h
61
#define pte_val(x) ((x).pte)
arch/nios2/include/asm/pgalloc.h
18
pte_t *pte)
arch/nios2/include/asm/pgalloc.h
20
set_pmd(pmd, __pmd((unsigned long)pte));
arch/nios2/include/asm/pgalloc.h
24
pgtable_t pte)
arch/nios2/include/asm/pgalloc.h
26
set_pmd(pmd, __pmd((unsigned long)page_address(pte)));
arch/nios2/include/asm/pgalloc.h
31
#define __pte_free_tlb(tlb, pte, addr) \
arch/nios2/include/asm/pgalloc.h
32
tlb_remove_ptdesc((tlb), page_ptdesc(pte))
arch/nios2/include/asm/pgtable.h
105
static inline int pte_none(pte_t pte)
arch/nios2/include/asm/pgtable.h
107
return !(pte_val(pte) & ~(_PAGE_GLOBAL|0xf));
arch/nios2/include/asm/pgtable.h
110
static inline int pte_present(pte_t pte) \
arch/nios2/include/asm/pgtable.h
111
{ return pte_val(pte) & _PAGE_PRESENT; }
arch/nios2/include/asm/pgtable.h
117
static inline pte_t pte_wrprotect(pte_t pte)
arch/nios2/include/asm/pgtable.h
119
pte_val(pte) &= ~_PAGE_WRITE;
arch/nios2/include/asm/pgtable.h
120
return pte;
arch/nios2/include/asm/pgtable.h
123
static inline pte_t pte_mkclean(pte_t pte)
arch/nios2/include/asm/pgtable.h
125
pte_val(pte) &= ~_PAGE_DIRTY;
arch/nios2/include/asm/pgtable.h
126
return pte;
arch/nios2/include/asm/pgtable.h
129
static inline pte_t pte_mkold(pte_t pte)
arch/nios2/include/asm/pgtable.h
131
pte_val(pte) &= ~_PAGE_ACCESSED;
arch/nios2/include/asm/pgtable.h
132
return pte;
arch/nios2/include/asm/pgtable.h
135
static inline pte_t pte_mkwrite_novma(pte_t pte)
arch/nios2/include/asm/pgtable.h
137
pte_val(pte) |= _PAGE_WRITE;
arch/nios2/include/asm/pgtable.h
138
return pte;
arch/nios2/include/asm/pgtable.h
141
static inline pte_t pte_mkdirty(pte_t pte)
arch/nios2/include/asm/pgtable.h
143
pte_val(pte) |= _PAGE_DIRTY;
arch/nios2/include/asm/pgtable.h
144
return pte;
arch/nios2/include/asm/pgtable.h
147
static inline pte_t pte_mkyoung(pte_t pte)
arch/nios2/include/asm/pgtable.h
149
pte_val(pte) |= _PAGE_ACCESSED;
arch/nios2/include/asm/pgtable.h
150
return pte;
arch/nios2/include/asm/pgtable.h
153
static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
arch/nios2/include/asm/pgtable.h
157
pte_val(pte) = (pte_val(pte) & ~mask) | (pgprot_val(newprot) & mask);
arch/nios2/include/asm/pgtable.h
158
return pte;
arch/nios2/include/asm/pgtable.h
172
#define pte_pfn(pte) (pte_val(pte) & 0xfffff)
arch/nios2/include/asm/pgtable.h
174
#define pte_page(pte) (pfn_to_page(pte_pfn(pte)))
arch/nios2/include/asm/pgtable.h
187
pte_t *ptep, pte_t pte, unsigned int nr)
arch/nios2/include/asm/pgtable.h
189
unsigned long paddr = (unsigned long)page_to_virt(pte_page(pte));
arch/nios2/include/asm/pgtable.h
193
set_pte(ptep, pte);
arch/nios2/include/asm/pgtable.h
197
pte_val(pte) += 1;
arch/nios2/include/asm/pgtable.h
260
#define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) })
arch/nios2/include/asm/pgtable.h
262
static inline bool pte_swp_exclusive(pte_t pte)
arch/nios2/include/asm/pgtable.h
264
return pte_val(pte) & _PAGE_SWP_EXCLUSIVE;
arch/nios2/include/asm/pgtable.h
267
static inline pte_t pte_swp_mkexclusive(pte_t pte)
arch/nios2/include/asm/pgtable.h
269
pte_val(pte) |= _PAGE_SWP_EXCLUSIVE;
arch/nios2/include/asm/pgtable.h
270
return pte;
arch/nios2/include/asm/pgtable.h
273
static inline pte_t pte_swp_clear_exclusive(pte_t pte)
arch/nios2/include/asm/pgtable.h
275
pte_val(pte) &= ~_PAGE_SWP_EXCLUSIVE;
arch/nios2/include/asm/pgtable.h
276
return pte;
arch/nios2/include/asm/pgtable.h
87
static inline int pte_write(pte_t pte) \
arch/nios2/include/asm/pgtable.h
88
{ return pte_val(pte) & _PAGE_WRITE; }
arch/nios2/include/asm/pgtable.h
89
static inline int pte_dirty(pte_t pte) \
arch/nios2/include/asm/pgtable.h
90
{ return pte_val(pte) & _PAGE_DIRTY; }
arch/nios2/include/asm/pgtable.h
91
static inline int pte_young(pte_t pte) \
arch/nios2/include/asm/pgtable.h
92
{ return pte_val(pte) & _PAGE_ACCESSED; }
arch/nios2/include/asm/tlbflush.h
42
pte_t pte);
arch/nios2/mm/cacheflush.c
212
pte_t pte = *ptep;
arch/nios2/mm/cacheflush.c
213
unsigned long pfn = pte_pfn(pte);
arch/nios2/mm/cacheflush.c
217
reload_tlb_page(vma, address, pte);
arch/nios2/mm/ioremap.c
22
static inline void remap_area_pte(pte_t *pte, unsigned long address,
arch/nios2/mm/ioremap.c
39
if (!pte_none(*pte)) {
arch/nios2/mm/ioremap.c
43
set_pte(pte, pfn_pte(pfn, pgprot));
arch/nios2/mm/ioremap.c
46
pte++;
arch/nios2/mm/ioremap.c
64
pte_t *pte = pte_alloc_kernel(pmd, address);
arch/nios2/mm/ioremap.c
66
if (!pte)
arch/nios2/mm/ioremap.c
68
remap_area_pte(pte, address, end - address, address + phys_addr,
arch/nios2/mm/tlb.c
113
void reload_tlb_page(struct vm_area_struct *vma, unsigned long addr, pte_t pte)
arch/nios2/mm/tlb.c
117
reload_tlb_one_pid(addr, mmu_pid, pte);
arch/nios2/mm/tlb.c
95
static void reload_tlb_one_pid(unsigned long addr, unsigned long mmu_pid, pte_t pte)
arch/nios2/mm/tlb.c
99
replace_tlb_one_pid(addr, mmu_pid, pte_val(pte));
arch/openrisc/include/asm/page.h
39
unsigned long pte;
arch/openrisc/include/asm/page.h
49
#define pte_val(x) ((x).pte)
arch/openrisc/include/asm/pgalloc.h
28
#define pmd_populate_kernel(mm, pmd, pte) \
arch/openrisc/include/asm/pgalloc.h
29
set_pmd(pmd, __pmd(_KERNPG_TABLE + __pa(pte)))
arch/openrisc/include/asm/pgalloc.h
32
struct page *pte)
arch/openrisc/include/asm/pgalloc.h
35
((unsigned long)page_to_pfn(pte) <<
arch/openrisc/include/asm/pgalloc.h
67
#define __pte_free_tlb(tlb, pte, addr) \
arch/openrisc/include/asm/pgalloc.h
68
tlb_remove_ptdesc((tlb), page_ptdesc(pte))
arch/openrisc/include/asm/pgtable.h
200
static inline int pte_read(pte_t pte) { return pte_val(pte) & _PAGE_READ; }
arch/openrisc/include/asm/pgtable.h
201
static inline int pte_write(pte_t pte) { return pte_val(pte) & _PAGE_WRITE; }
arch/openrisc/include/asm/pgtable.h
202
static inline int pte_exec(pte_t pte) { return pte_val(pte) & _PAGE_EXEC; }
arch/openrisc/include/asm/pgtable.h
203
static inline int pte_dirty(pte_t pte) { return pte_val(pte) & _PAGE_DIRTY; }
arch/openrisc/include/asm/pgtable.h
204
static inline int pte_young(pte_t pte) { return pte_val(pte) & _PAGE_ACCESSED; }
arch/openrisc/include/asm/pgtable.h
206
static inline pte_t pte_wrprotect(pte_t pte)
arch/openrisc/include/asm/pgtable.h
208
pte_val(pte) &= ~(_PAGE_WRITE);
arch/openrisc/include/asm/pgtable.h
209
return pte;
arch/openrisc/include/asm/pgtable.h
212
static inline pte_t pte_rdprotect(pte_t pte)
arch/openrisc/include/asm/pgtable.h
214
pte_val(pte) &= ~(_PAGE_READ);
arch/openrisc/include/asm/pgtable.h
215
return pte;
arch/openrisc/include/asm/pgtable.h
218
static inline pte_t pte_exprotect(pte_t pte)
arch/openrisc/include/asm/pgtable.h
220
pte_val(pte) &= ~(_PAGE_EXEC);
arch/openrisc/include/asm/pgtable.h
221
return pte;
arch/openrisc/include/asm/pgtable.h
224
static inline pte_t pte_mkclean(pte_t pte)
arch/openrisc/include/asm/pgtable.h
226
pte_val(pte) &= ~(_PAGE_DIRTY);
arch/openrisc/include/asm/pgtable.h
227
return pte;
arch/openrisc/include/asm/pgtable.h
230
static inline pte_t pte_mkold(pte_t pte)
arch/openrisc/include/asm/pgtable.h
232
pte_val(pte) &= ~(_PAGE_ACCESSED);
arch/openrisc/include/asm/pgtable.h
233
return pte;
arch/openrisc/include/asm/pgtable.h
236
static inline pte_t pte_mkwrite_novma(pte_t pte)
arch/openrisc/include/asm/pgtable.h
238
pte_val(pte) |= _PAGE_WRITE;
arch/openrisc/include/asm/pgtable.h
239
return pte;
arch/openrisc/include/asm/pgtable.h
242
static inline pte_t pte_mkread(pte_t pte)
arch/openrisc/include/asm/pgtable.h
244
pte_val(pte) |= _PAGE_READ;
arch/openrisc/include/asm/pgtable.h
245
return pte;
arch/openrisc/include/asm/pgtable.h
248
static inline pte_t pte_mkexec(pte_t pte)
arch/openrisc/include/asm/pgtable.h
250
pte_val(pte) |= _PAGE_EXEC;
arch/openrisc/include/asm/pgtable.h
251
return pte;
arch/openrisc/include/asm/pgtable.h
254
static inline pte_t pte_mkdirty(pte_t pte)
arch/openrisc/include/asm/pgtable.h
256
pte_val(pte) |= _PAGE_DIRTY;
arch/openrisc/include/asm/pgtable.h
257
return pte;
arch/openrisc/include/asm/pgtable.h
260
static inline pte_t pte_mkyoung(pte_t pte)
arch/openrisc/include/asm/pgtable.h
262
pte_val(pte) |= _PAGE_ACCESSED;
arch/openrisc/include/asm/pgtable.h
263
return pte;
arch/openrisc/include/asm/pgtable.h
279
pte_t pte;
arch/openrisc/include/asm/pgtable.h
281
pte_val(pte) = __pa(page) | pgprot_val(pgprot);
arch/openrisc/include/asm/pgtable.h
282
return pte;
arch/openrisc/include/asm/pgtable.h
293
static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
arch/openrisc/include/asm/pgtable.h
295
pte_val(pte) = (pte_val(pte) & _PAGE_CHG_MASK) | pgprot_val(newprot);
arch/openrisc/include/asm/pgtable.h
296
return pte;
arch/openrisc/include/asm/pgtable.h
307
static inline unsigned long __pte_page(pte_t pte)
arch/openrisc/include/asm/pgtable.h
310
return (unsigned long)__va(pte_val(pte) & PAGE_MASK);
arch/openrisc/include/asm/pgtable.h
313
#define pte_pagenr(pte) ((__pte_page(pte) - PAGE_OFFSET) >> PAGE_SHIFT)
arch/openrisc/include/asm/pgtable.h
318
#define pte_page(pte) (mem_map+pte_pagenr(pte))
arch/openrisc/include/asm/pgtable.h
342
#define pte_pfn(x) ((unsigned long)(((x).pte)) >> PAGE_SHIFT)
arch/openrisc/include/asm/pgtable.h
357
unsigned long address, pte_t *pte)
arch/openrisc/include/asm/pgtable.h
362
unsigned long address, pte_t *pte);
arch/openrisc/include/asm/pgtable.h
394
#define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) })
arch/openrisc/include/asm/pgtable.h
397
static inline bool pte_swp_exclusive(pte_t pte)
arch/openrisc/include/asm/pgtable.h
399
return pte_val(pte) & _PAGE_SWP_EXCLUSIVE;
arch/openrisc/include/asm/pgtable.h
402
static inline pte_t pte_swp_mkexclusive(pte_t pte)
arch/openrisc/include/asm/pgtable.h
404
pte_val(pte) |= _PAGE_SWP_EXCLUSIVE;
arch/openrisc/include/asm/pgtable.h
405
return pte;
arch/openrisc/include/asm/pgtable.h
408
static inline pte_t pte_swp_clear_exclusive(pte_t pte)
arch/openrisc/include/asm/pgtable.h
410
pte_val(pte) &= ~_PAGE_SWP_EXCLUSIVE;
arch/openrisc/include/asm/pgtable.h
411
return pte;
arch/openrisc/kernel/dma.c
25
page_set_nocache(pte_t *pte, unsigned long addr,
arch/openrisc/kernel/dma.c
28
pte_val(*pte) |= _PAGE_CI;
arch/openrisc/kernel/dma.c
47
page_clear_nocache(pte_t *pte, unsigned long addr,
arch/openrisc/kernel/dma.c
50
pte_val(*pte) &= ~_PAGE_CI;
arch/openrisc/mm/cache.c
82
pte_t *pte)
arch/openrisc/mm/cache.c
84
unsigned long pfn = pte_val(*pte) >> PAGE_SHIFT;
arch/openrisc/mm/init.c
103
v += PAGE_SIZE, p += PAGE_SIZE, j++, pte++) {
arch/openrisc/mm/init.c
110
set_pte(pte, mk_pte_phys(p, prot));
arch/openrisc/mm/init.c
204
pte_t *pte;
arch/openrisc/mm/init.c
209
pte = pte_alloc_kernel(pmd, va);
arch/openrisc/mm/init.c
211
if (pte == NULL)
arch/openrisc/mm/init.c
215
set_pte_at(&init_mm, va, pte, pfn_pte(pa >> PAGE_SHIFT, prot));
arch/openrisc/mm/init.c
217
pte_clear(&init_mm, va, pte);
arch/openrisc/mm/init.c
67
pte_t *pte;
arch/openrisc/mm/init.c
95
pte = memblock_alloc_raw(PAGE_SIZE, PAGE_SIZE);
arch/openrisc/mm/init.c
96
if (!pte)
arch/openrisc/mm/init.c
99
set_pmd(pme, __pmd(_KERNPG_TABLE + __pa(pte)));
arch/openrisc/mm/ioremap.c
36
pte_t *pte;
arch/openrisc/mm/ioremap.c
39
pte = __pte_alloc_one_kernel(mm);
arch/openrisc/mm/ioremap.c
41
pte = memblock_alloc_or_panic(PAGE_SIZE, PAGE_SIZE);
arch/openrisc/mm/ioremap.c
44
return pte;
arch/parisc/include/asm/hugetlb.h
29
pte_t pte, int dirty);
arch/parisc/include/asm/hugetlb.h
9
pte_t *ptep, pte_t pte, unsigned long sz);
arch/parisc/include/asm/kfence.h
24
pte_t *pte = virt_to_kpte(addr);
arch/parisc/include/asm/kfence.h
26
if (WARN_ON(!pte))
arch/parisc/include/asm/kfence.h
35
set_pte(pte, __pte(pte_val(*pte) & ~_PAGE_PRESENT));
arch/parisc/include/asm/kfence.h
37
set_pte(pte, __pte(pte_val(*pte) | _PAGE_PRESENT));
arch/parisc/include/asm/page.h
33
typedef struct { unsigned long pte; } pte_t; /* either 32 or 64bit */
arch/parisc/include/asm/page.h
47
#define pte_val(x) ((x).pte)
arch/parisc/include/asm/pgalloc.h
51
pmd_populate_kernel(struct mm_struct *mm, pmd_t *pmd, pte_t *pte)
arch/parisc/include/asm/pgalloc.h
54
+ (__u32)(__pa((unsigned long)pte) >> PxD_VALUE_SHIFT)));
arch/parisc/include/asm/pgtable.h
315
static inline int pte_dirty(pte_t pte) { return pte_val(pte) & _PAGE_DIRTY; }
arch/parisc/include/asm/pgtable.h
316
static inline int pte_young(pte_t pte) { return pte_val(pte) & _PAGE_ACCESSED; }
arch/parisc/include/asm/pgtable.h
317
static inline int pte_write(pte_t pte) { return pte_val(pte) & _PAGE_WRITE; }
arch/parisc/include/asm/pgtable.h
318
static inline int pte_special(pte_t pte) { return pte_val(pte) & _PAGE_SPECIAL; }
arch/parisc/include/asm/pgtable.h
320
static inline pte_t pte_mkclean(pte_t pte) { pte_val(pte) &= ~_PAGE_DIRTY; return pte; }
arch/parisc/include/asm/pgtable.h
321
static inline pte_t pte_mkold(pte_t pte) { pte_val(pte) &= ~_PAGE_ACCESSED; return pte; }
arch/parisc/include/asm/pgtable.h
322
static inline pte_t pte_wrprotect(pte_t pte) { pte_val(pte) &= ~_PAGE_WRITE; return pte; }
arch/parisc/include/asm/pgtable.h
323
static inline pte_t pte_mkdirty(pte_t pte) { pte_val(pte) |= _PAGE_DIRTY; return pte; }
arch/parisc/include/asm/pgtable.h
324
static inline pte_t pte_mkyoung(pte_t pte) { pte_val(pte) |= _PAGE_ACCESSED; return pte; }
arch/parisc/include/asm/pgtable.h
325
static inline pte_t pte_mkwrite_novma(pte_t pte) { pte_val(pte) |= _PAGE_WRITE; return pte; }
arch/parisc/include/asm/pgtable.h
326
static inline pte_t pte_mkspecial(pte_t pte) { pte_val(pte) |= _PAGE_SPECIAL; return pte; }
arch/parisc/include/asm/pgtable.h
332
#define pte_huge(pte) (pte_val(pte) & _PAGE_HUGE)
arch/parisc/include/asm/pgtable.h
333
#define pte_mkhuge(pte) (__pte(pte_val(pte) | \
arch/parisc/include/asm/pgtable.h
336
#define pte_huge(pte) (0)
arch/parisc/include/asm/pgtable.h
337
#define pte_mkhuge(pte) (pte)
arch/parisc/include/asm/pgtable.h
352
pte_t pte;
arch/parisc/include/asm/pgtable.h
353
pte_val(pte) = (pfn << PFN_PTE_SHIFT) | pgprot_val(pgprot);
arch/parisc/include/asm/pgtable.h
354
return pte;
arch/parisc/include/asm/pgtable.h
357
static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
arch/parisc/include/asm/pgtable.h
358
{ pte_val(pte) = (pte_val(pte) & _PAGE_CHG_MASK) | pgprot_val(newprot); return pte; }
arch/parisc/include/asm/pgtable.h
364
#define pte_page(pte) (pfn_to_page(pte_pfn(pte)))
arch/parisc/include/asm/pgtable.h
380
pte_t *ptep, pte_t pte, unsigned int nr)
arch/parisc/include/asm/pgtable.h
382
if (pte_present(pte) && pte_user(pte))
arch/parisc/include/asm/pgtable.h
383
__update_cache(pte);
arch/parisc/include/asm/pgtable.h
385
*ptep = pte;
arch/parisc/include/asm/pgtable.h
390
pte_val(pte) += 1 << PFN_PTE_SHIFT;
arch/parisc/include/asm/pgtable.h
395
#define set_pte_at(mm, addr, ptep, pte) set_ptes(mm, addr, ptep, pte, 1)
arch/parisc/include/asm/pgtable.h
426
#define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) })
arch/parisc/include/asm/pgtable.h
429
static inline bool pte_swp_exclusive(pte_t pte)
arch/parisc/include/asm/pgtable.h
431
return pte_val(pte) & _PAGE_SWP_EXCLUSIVE;
arch/parisc/include/asm/pgtable.h
434
static inline pte_t pte_swp_mkexclusive(pte_t pte)
arch/parisc/include/asm/pgtable.h
436
pte_val(pte) |= _PAGE_SWP_EXCLUSIVE;
arch/parisc/include/asm/pgtable.h
437
return pte;
arch/parisc/include/asm/pgtable.h
440
static inline pte_t pte_swp_clear_exclusive(pte_t pte)
arch/parisc/include/asm/pgtable.h
442
pte_val(pte) &= ~_PAGE_SWP_EXCLUSIVE;
arch/parisc/include/asm/pgtable.h
443
return pte;
arch/parisc/include/asm/pgtable.h
454
pte_t pte;
arch/parisc/include/asm/pgtable.h
456
pte = ptep_get(ptep);
arch/parisc/include/asm/pgtable.h
457
if (!pte_young(pte)) {
arch/parisc/include/asm/pgtable.h
460
set_pte_at(vma->vm_mm, addr, ptep, pte_mkold(pte));
arch/parisc/include/asm/pgtable.h
64
extern void __update_cache(pte_t pte);
arch/parisc/include/asm/tlb.h
10
#define __pte_free_tlb(tlb, pte, addr) tlb_remove_ptdesc((tlb), page_ptdesc(pte))
arch/parisc/kernel/cache.c
109
void __update_cache(pte_t pte)
arch/parisc/kernel/cache.c
111
unsigned long pfn = pte_pfn(pte);
arch/parisc/kernel/cache.c
432
static inline bool pte_needs_cache_flush(pte_t pte)
arch/parisc/kernel/cache.c
434
return (pte_val(pte) & (_PAGE_PRESENT | _PAGE_ACCESSED | _PAGE_NO_CACHE))
arch/parisc/kernel/cache.c
628
pte_t *ptep, pte;
arch/parisc/kernel/cache.c
632
pte = ptep_get(ptep);
arch/parisc/kernel/cache.c
633
needs_flush = pte_needs_cache_flush(pte);
arch/parisc/kernel/cache.c
637
__flush_cache_page(vma, vmaddr, PFN_PHYS(pte_pfn(pte)));
arch/parisc/kernel/cache.c
787
pte_t pte = ptep_get(ptep);
arch/parisc/kernel/cache.c
789
if (!pte_young(pte))
arch/parisc/kernel/cache.c
791
set_pte(ptep, pte_mkold(pte));
arch/parisc/kernel/cache.c
793
__flush_cache_page(vma, addr, PFN_PHYS(pte_pfn(pte)));
arch/parisc/kernel/cache.c
808
pte_t pte = ptep_get_and_clear(mm, addr, ptep);
arch/parisc/kernel/cache.c
809
unsigned long pfn = pte_pfn(pte);
arch/parisc/kernel/cache.c
813
else if (pte_accessible(mm, pte))
arch/parisc/kernel/cache.c
816
return pte;
arch/parisc/kernel/pci-dma.c
115
pte_t * pte = pte_alloc_kernel(pmd, vaddr);
arch/parisc/kernel/pci-dma.c
116
if (!pte)
arch/parisc/kernel/pci-dma.c
118
if (map_pte_uncached(pte, orig_vaddr, end - vaddr, paddr_ptr))
arch/parisc/kernel/pci-dma.c
156
pte_t * pte;
arch/parisc/kernel/pci-dma.c
167
pte = pte_offset_kernel(pmd, vaddr);
arch/parisc/kernel/pci-dma.c
174
pte_t page = *pte;
arch/parisc/kernel/pci-dma.c
176
pte_clear(&init_mm, vaddr, pte);
arch/parisc/kernel/pci-dma.c
182
pte++;
arch/parisc/kernel/pci-dma.c
76
static inline int map_pte_uncached(pte_t * pte,
arch/parisc/kernel/pci-dma.c
90
if (!pte_none(*pte))
arch/parisc/kernel/pci-dma.c
93
set_pte(pte, __mk_pte(*paddr_ptr, PAGE_KERNEL_UNC));
arch/parisc/kernel/pci-dma.c
99
pte++;
arch/parisc/mm/fixmap.c
20
pte_t *pte;
arch/parisc/mm/fixmap.c
22
pte = pte_offset_kernel(pmd, vaddr);
arch/parisc/mm/fixmap.c
23
set_pte_at(&init_mm, vaddr, pte, __mk_pte(phys, PAGE_KERNEL_RWX));
arch/parisc/mm/fixmap.c
30
pte_t *pte = virt_to_kpte(vaddr);
arch/parisc/mm/fixmap.c
32
if (WARN_ON(pte_none(*pte)))
arch/parisc/mm/fixmap.c
35
pte_clear(&init_mm, vaddr, pte);
arch/parisc/mm/hugetlbpage.c
151
pte_t pte, int dirty)
arch/parisc/mm/hugetlbpage.c
156
changed = !pte_same(*ptep, pte);
arch/parisc/mm/hugetlbpage.c
158
__set_huge_pte_at(mm, addr, ptep, pte);
arch/parisc/mm/hugetlbpage.c
33
pte_t *pte = NULL;
arch/parisc/mm/hugetlbpage.c
48
pte = pte_alloc_huge(mm, pmd, addr);
arch/parisc/mm/hugetlbpage.c
50
return pte;
arch/parisc/mm/hugetlbpage.c
60
pte_t *pte = NULL;
arch/parisc/mm/hugetlbpage.c
72
pte = pte_offset_huge(pmd, addr);
arch/parisc/mm/hugetlbpage.c
76
return pte;
arch/parisc/mm/init.c
395
pte_t pte;
arch/parisc/mm/init.c
417
pte = __mk_pte(address, prot);
arch/parisc/mm/init.c
419
pte = pte_mkhuge(pte);
arch/parisc/mm/init.c
424
set_pte(pg_table, pte);
arch/parisc/mm/init.c
688
pte_t *pte = memblock_alloc_or_panic(PAGE_SIZE, PAGE_SIZE);
arch/parisc/mm/init.c
690
pmd_populate_kernel(&init_mm, pmd, pte);
arch/powerpc/include/asm/book3s/32/pgalloc.h
35
pte_t *pte)
arch/powerpc/include/asm/book3s/32/pgalloc.h
37
*pmdp = __pmd(__pa(pte) | _PMD_PRESENT);
arch/powerpc/include/asm/book3s/32/pgtable.h
366
#define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) >> 3 })
arch/powerpc/include/asm/book3s/32/pgtable.h
369
static inline bool pte_swp_exclusive(pte_t pte)
arch/powerpc/include/asm/book3s/32/pgtable.h
371
return pte_val(pte) & _PAGE_SWP_EXCLUSIVE;
arch/powerpc/include/asm/book3s/32/pgtable.h
374
static inline pte_t pte_swp_mkexclusive(pte_t pte)
arch/powerpc/include/asm/book3s/32/pgtable.h
376
return __pte(pte_val(pte) | _PAGE_SWP_EXCLUSIVE);
arch/powerpc/include/asm/book3s/32/pgtable.h
379
static inline pte_t pte_swp_clear_exclusive(pte_t pte)
arch/powerpc/include/asm/book3s/32/pgtable.h
381
return __pte(pte_val(pte) & ~_PAGE_SWP_EXCLUSIVE);
arch/powerpc/include/asm/book3s/32/pgtable.h
385
static inline bool pte_read(pte_t pte)
arch/powerpc/include/asm/book3s/32/pgtable.h
387
return !!(pte_val(pte) & _PAGE_READ);
arch/powerpc/include/asm/book3s/32/pgtable.h
390
static inline bool pte_write(pte_t pte)
arch/powerpc/include/asm/book3s/32/pgtable.h
392
return !!(pte_val(pte) & _PAGE_WRITE);
arch/powerpc/include/asm/book3s/32/pgtable.h
395
static inline int pte_dirty(pte_t pte) { return !!(pte_val(pte) & _PAGE_DIRTY); }
arch/powerpc/include/asm/book3s/32/pgtable.h
396
static inline int pte_young(pte_t pte) { return !!(pte_val(pte) & _PAGE_ACCESSED); }
arch/powerpc/include/asm/book3s/32/pgtable.h
397
static inline int pte_special(pte_t pte) { return !!(pte_val(pte) & _PAGE_SPECIAL); }
arch/powerpc/include/asm/book3s/32/pgtable.h
398
static inline int pte_none(pte_t pte) { return (pte_val(pte) & ~_PTE_NONE_MASK) == 0; }
arch/powerpc/include/asm/book3s/32/pgtable.h
399
static inline bool pte_exec(pte_t pte) { return pte_val(pte) & _PAGE_EXEC; }
arch/powerpc/include/asm/book3s/32/pgtable.h
401
static inline int pte_present(pte_t pte)
arch/powerpc/include/asm/book3s/32/pgtable.h
403
return pte_val(pte) & _PAGE_PRESENT;
arch/powerpc/include/asm/book3s/32/pgtable.h
406
static inline bool pte_hw_valid(pte_t pte)
arch/powerpc/include/asm/book3s/32/pgtable.h
408
return pte_val(pte) & _PAGE_PRESENT;
arch/powerpc/include/asm/book3s/32/pgtable.h
411
static inline bool pte_hashpte(pte_t pte)
arch/powerpc/include/asm/book3s/32/pgtable.h
413
return !!(pte_val(pte) & _PAGE_HASHPTE);
arch/powerpc/include/asm/book3s/32/pgtable.h
416
static inline bool pte_ci(pte_t pte)
arch/powerpc/include/asm/book3s/32/pgtable.h
418
return !!(pte_val(pte) & _PAGE_NO_CACHE);
arch/powerpc/include/asm/book3s/32/pgtable.h
426
static inline bool pte_access_permitted(pte_t pte, bool write)
arch/powerpc/include/asm/book3s/32/pgtable.h
432
if (!pte_present(pte) || !pte_read(pte))
arch/powerpc/include/asm/book3s/32/pgtable.h
435
if (write && !pte_write(pte))
arch/powerpc/include/asm/book3s/32/pgtable.h
441
static inline bool pte_user_accessible_page(pte_t pte, unsigned long addr)
arch/powerpc/include/asm/book3s/32/pgtable.h
443
return pte_present(pte) && !is_kernel_addr(addr);
arch/powerpc/include/asm/book3s/32/pgtable.h
459
static inline pte_t pte_wrprotect(pte_t pte)
arch/powerpc/include/asm/book3s/32/pgtable.h
461
return __pte(pte_val(pte) & ~_PAGE_WRITE);
arch/powerpc/include/asm/book3s/32/pgtable.h
464
static inline pte_t pte_exprotect(pte_t pte)
arch/powerpc/include/asm/book3s/32/pgtable.h
466
return __pte(pte_val(pte) & ~_PAGE_EXEC);
arch/powerpc/include/asm/book3s/32/pgtable.h
469
static inline pte_t pte_mkclean(pte_t pte)
arch/powerpc/include/asm/book3s/32/pgtable.h
471
return __pte(pte_val(pte) & ~_PAGE_DIRTY);
arch/powerpc/include/asm/book3s/32/pgtable.h
474
static inline pte_t pte_mkold(pte_t pte)
arch/powerpc/include/asm/book3s/32/pgtable.h
476
return __pte(pte_val(pte) & ~_PAGE_ACCESSED);
arch/powerpc/include/asm/book3s/32/pgtable.h
479
static inline pte_t pte_mkexec(pte_t pte)
arch/powerpc/include/asm/book3s/32/pgtable.h
481
return __pte(pte_val(pte) | _PAGE_EXEC);
arch/powerpc/include/asm/book3s/32/pgtable.h
484
static inline pte_t pte_mkpte(pte_t pte)
arch/powerpc/include/asm/book3s/32/pgtable.h
486
return pte;
arch/powerpc/include/asm/book3s/32/pgtable.h
489
static inline pte_t pte_mkwrite_novma(pte_t pte)
arch/powerpc/include/asm/book3s/32/pgtable.h
494
return __pte(pte_val(pte) | _PAGE_RW);
arch/powerpc/include/asm/book3s/32/pgtable.h
497
static inline pte_t pte_mkdirty(pte_t pte)
arch/powerpc/include/asm/book3s/32/pgtable.h
499
return __pte(pte_val(pte) | _PAGE_DIRTY);
arch/powerpc/include/asm/book3s/32/pgtable.h
502
static inline pte_t pte_mkyoung(pte_t pte)
arch/powerpc/include/asm/book3s/32/pgtable.h
504
return __pte(pte_val(pte) | _PAGE_ACCESSED);
arch/powerpc/include/asm/book3s/32/pgtable.h
507
static inline pte_t pte_mkspecial(pte_t pte)
arch/powerpc/include/asm/book3s/32/pgtable.h
509
return __pte(pte_val(pte) | _PAGE_SPECIAL);
arch/powerpc/include/asm/book3s/32/pgtable.h
512
static inline pte_t pte_mkhuge(pte_t pte)
arch/powerpc/include/asm/book3s/32/pgtable.h
514
return pte;
arch/powerpc/include/asm/book3s/32/pgtable.h
517
static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
arch/powerpc/include/asm/book3s/32/pgtable.h
519
return __pte((pte_val(pte) & _PAGE_CHG_MASK) | pgprot_val(newprot));
arch/powerpc/include/asm/book3s/32/pgtable.h
547
pte_t *ptep, pte_t pte, int percpu)
arch/powerpc/include/asm/book3s/32/pgtable.h
551
(pte_val(pte) & ~_PAGE_HASHPTE));
arch/powerpc/include/asm/book3s/32/pgtable.h
558
"r" (pte) : "memory");
arch/powerpc/include/asm/book3s/32/pgtable.h
560
pte_update(mm, addr, ptep, ~_PAGE_HASHPTE, pte_val(pte), 0);
arch/powerpc/include/asm/book3s/64/hash-4k.h
103
#define pte_pagesize_index(mm, addr, pte) MMU_PAGE_4K
arch/powerpc/include/asm/book3s/64/hash-4k.h
80
static inline real_pte_t __real_pte(pte_t pte, pte_t *ptep, int offset)
arch/powerpc/include/asm/book3s/64/hash-4k.h
82
return (real_pte_t){pte};
arch/powerpc/include/asm/book3s/64/hash-4k.h
85
#define __rpte_to_pte(r) ((r).pte)
arch/powerpc/include/asm/book3s/64/hash-64k.h
152
#define __rpte_to_pte(r) ((r).pte)
arch/powerpc/include/asm/book3s/64/hash-64k.h
170
#define pte_pagesize_index(mm, addr, pte) \
arch/powerpc/include/asm/book3s/64/hash-64k.h
171
(((pte) & H_PAGE_COMBO)? MMU_PAGE_4K: MMU_PAGE_64K)
arch/powerpc/include/asm/book3s/64/hash-64k.h
92
static inline real_pte_t __real_pte(pte_t pte, pte_t *ptep, int offset)
arch/powerpc/include/asm/book3s/64/hash-64k.h
97
rpte.pte = pte;
arch/powerpc/include/asm/book3s/64/hash.h
162
pte_t *ptep, unsigned long pte, int huge);
arch/powerpc/include/asm/book3s/64/hash.h
251
static inline int hash__pte_none(pte_t pte)
arch/powerpc/include/asm/book3s/64/hash.h
253
return (pte_val(pte) & ~H_PTE_NONE_MASK) == 0;
arch/powerpc/include/asm/book3s/64/hash.h
265
pte_t *ptep, pte_t pte, int percpu)
arch/powerpc/include/asm/book3s/64/hash.h
271
*ptep = pte;
arch/powerpc/include/asm/book3s/64/hugetlb.h
17
pte_t old_pte, pte_t pte);
arch/powerpc/include/asm/book3s/64/mmu-hash.h
466
unsigned int hash_page_do_lazy_icache(unsigned int pp, pte_t pte, int trap);
arch/powerpc/include/asm/book3s/64/mmu-hash.h
494
unsigned long pte);
arch/powerpc/include/asm/book3s/64/pgalloc.h
159
pte_t *pte)
arch/powerpc/include/asm/book3s/64/pgalloc.h
161
*pmd = __pmd(__pgtable_ptr_val(pte) | PMD_VAL_BITS);
arch/powerpc/include/asm/book3s/64/pgtable.h
1039
static inline pmd_t pte_pmd(pte_t pte)
arch/powerpc/include/asm/book3s/64/pgtable.h
1041
return __pmd_raw(pte_raw(pte));
arch/powerpc/include/asm/book3s/64/pgtable.h
391
static inline int pte_write(pte_t pte)
arch/powerpc/include/asm/book3s/64/pgtable.h
393
return !!(pte_raw(pte) & cpu_to_be64(_PAGE_WRITE));
arch/powerpc/include/asm/book3s/64/pgtable.h
396
static inline int pte_read(pte_t pte)
arch/powerpc/include/asm/book3s/64/pgtable.h
398
return !!(pte_raw(pte) & cpu_to_be64(_PAGE_READ));
arch/powerpc/include/asm/book3s/64/pgtable.h
455
static inline int pte_dirty(pte_t pte)
arch/powerpc/include/asm/book3s/64/pgtable.h
457
return !!(pte_raw(pte) & cpu_to_be64(_PAGE_DIRTY));
arch/powerpc/include/asm/book3s/64/pgtable.h
460
static inline int pte_young(pte_t pte)
arch/powerpc/include/asm/book3s/64/pgtable.h
462
return !!(pte_raw(pte) & cpu_to_be64(_PAGE_ACCESSED));
arch/powerpc/include/asm/book3s/64/pgtable.h
465
static inline int pte_special(pte_t pte)
arch/powerpc/include/asm/book3s/64/pgtable.h
467
return !!(pte_raw(pte) & cpu_to_be64(_PAGE_SPECIAL));
arch/powerpc/include/asm/book3s/64/pgtable.h
470
static inline bool pte_exec(pte_t pte)
arch/powerpc/include/asm/book3s/64/pgtable.h
472
return !!(pte_raw(pte) & cpu_to_be64(_PAGE_EXEC));
arch/powerpc/include/asm/book3s/64/pgtable.h
477
static inline bool pte_soft_dirty(pte_t pte)
arch/powerpc/include/asm/book3s/64/pgtable.h
479
return !!(pte_raw(pte) & cpu_to_be64(_PAGE_SOFT_DIRTY));
arch/powerpc/include/asm/book3s/64/pgtable.h
482
static inline pte_t pte_mksoft_dirty(pte_t pte)
arch/powerpc/include/asm/book3s/64/pgtable.h
484
return __pte_raw(pte_raw(pte) | cpu_to_be64(_PAGE_SOFT_DIRTY));
arch/powerpc/include/asm/book3s/64/pgtable.h
487
static inline pte_t pte_clear_soft_dirty(pte_t pte)
arch/powerpc/include/asm/book3s/64/pgtable.h
489
return __pte_raw(pte_raw(pte) & cpu_to_be64(~_PAGE_SOFT_DIRTY));
arch/powerpc/include/asm/book3s/64/pgtable.h
494
static inline int pte_protnone(pte_t pte)
arch/powerpc/include/asm/book3s/64/pgtable.h
496
return (pte_raw(pte) & cpu_to_be64(_PAGE_PRESENT | _PAGE_PTE | _PAGE_RWX)) ==
arch/powerpc/include/asm/book3s/64/pgtable.h
501
static inline bool pte_hw_valid(pte_t pte)
arch/powerpc/include/asm/book3s/64/pgtable.h
503
return (pte_raw(pte) & cpu_to_be64(_PAGE_PRESENT | _PAGE_PTE)) ==
arch/powerpc/include/asm/book3s/64/pgtable.h
507
static inline int pte_present(pte_t pte)
arch/powerpc/include/asm/book3s/64/pgtable.h
516
if (pte_hw_valid(pte))
arch/powerpc/include/asm/book3s/64/pgtable.h
518
return (pte_raw(pte) & cpu_to_be64(_PAGE_INVALID | _PAGE_PTE)) ==
arch/powerpc/include/asm/book3s/64/pgtable.h
523
extern bool arch_pte_access_permitted(u64 pte, bool write, bool execute);
arch/powerpc/include/asm/book3s/64/pgtable.h
525
static inline bool arch_pte_access_permitted(u64 pte, bool write, bool execute)
arch/powerpc/include/asm/book3s/64/pgtable.h
531
static inline bool pte_user(pte_t pte)
arch/powerpc/include/asm/book3s/64/pgtable.h
533
return !(pte_raw(pte) & cpu_to_be64(_PAGE_PRIVILEGED));
arch/powerpc/include/asm/book3s/64/pgtable.h
537
static inline bool pte_access_permitted(pte_t pte, bool write)
arch/powerpc/include/asm/book3s/64/pgtable.h
543
if (!pte_present(pte) || !pte_user(pte) || !pte_read(pte))
arch/powerpc/include/asm/book3s/64/pgtable.h
546
if (write && !pte_write(pte))
arch/powerpc/include/asm/book3s/64/pgtable.h
549
return arch_pte_access_permitted(pte_val(pte), write, 0);
arch/powerpc/include/asm/book3s/64/pgtable.h
552
static inline bool pte_user_accessible_page(pte_t pte, unsigned long addr)
arch/powerpc/include/asm/book3s/64/pgtable.h
554
return pte_present(pte) && pte_user(pte);
arch/powerpc/include/asm/book3s/64/pgtable.h
573
static inline pte_t pte_wrprotect(pte_t pte)
arch/powerpc/include/asm/book3s/64/pgtable.h
575
return __pte_raw(pte_raw(pte) & cpu_to_be64(~_PAGE_WRITE));
arch/powerpc/include/asm/book3s/64/pgtable.h
578
static inline pte_t pte_exprotect(pte_t pte)
arch/powerpc/include/asm/book3s/64/pgtable.h
580
return __pte_raw(pte_raw(pte) & cpu_to_be64(~_PAGE_EXEC));
arch/powerpc/include/asm/book3s/64/pgtable.h
583
static inline pte_t pte_mkclean(pte_t pte)
arch/powerpc/include/asm/book3s/64/pgtable.h
585
return __pte_raw(pte_raw(pte) & cpu_to_be64(~_PAGE_DIRTY));
arch/powerpc/include/asm/book3s/64/pgtable.h
588
static inline pte_t pte_mkold(pte_t pte)
arch/powerpc/include/asm/book3s/64/pgtable.h
590
return __pte_raw(pte_raw(pte) & cpu_to_be64(~_PAGE_ACCESSED));
arch/powerpc/include/asm/book3s/64/pgtable.h
593
static inline pte_t pte_mkexec(pte_t pte)
arch/powerpc/include/asm/book3s/64/pgtable.h
595
return __pte_raw(pte_raw(pte) | cpu_to_be64(_PAGE_EXEC));
arch/powerpc/include/asm/book3s/64/pgtable.h
598
static inline pte_t pte_mkwrite_novma(pte_t pte)
arch/powerpc/include/asm/book3s/64/pgtable.h
603
return __pte_raw(pte_raw(pte) | cpu_to_be64(_PAGE_RW));
arch/powerpc/include/asm/book3s/64/pgtable.h
606
static inline pte_t pte_mkdirty(pte_t pte)
arch/powerpc/include/asm/book3s/64/pgtable.h
608
return __pte_raw(pte_raw(pte) | cpu_to_be64(_PAGE_DIRTY | _PAGE_SOFT_DIRTY));
arch/powerpc/include/asm/book3s/64/pgtable.h
611
static inline pte_t pte_mkyoung(pte_t pte)
arch/powerpc/include/asm/book3s/64/pgtable.h
613
return __pte_raw(pte_raw(pte) | cpu_to_be64(_PAGE_ACCESSED));
arch/powerpc/include/asm/book3s/64/pgtable.h
616
static inline pte_t pte_mkspecial(pte_t pte)
arch/powerpc/include/asm/book3s/64/pgtable.h
618
return __pte_raw(pte_raw(pte) | cpu_to_be64(_PAGE_SPECIAL));
arch/powerpc/include/asm/book3s/64/pgtable.h
621
static inline pte_t pte_mkhuge(pte_t pte)
arch/powerpc/include/asm/book3s/64/pgtable.h
623
return pte;
arch/powerpc/include/asm/book3s/64/pgtable.h
626
static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
arch/powerpc/include/asm/book3s/64/pgtable.h
629
return __pte_raw((pte_raw(pte) & cpu_to_be64(_PAGE_CHG_MASK)) |
arch/powerpc/include/asm/book3s/64/pgtable.h
657
#define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val((pte)) & ~_PAGE_PTE })
arch/powerpc/include/asm/book3s/64/pgtable.h
671
static inline pte_t pte_swp_mksoft_dirty(pte_t pte)
arch/powerpc/include/asm/book3s/64/pgtable.h
673
return __pte_raw(pte_raw(pte) | cpu_to_be64(_PAGE_SWP_SOFT_DIRTY));
arch/powerpc/include/asm/book3s/64/pgtable.h
676
static inline bool pte_swp_soft_dirty(pte_t pte)
arch/powerpc/include/asm/book3s/64/pgtable.h
678
return !!(pte_raw(pte) & cpu_to_be64(_PAGE_SWP_SOFT_DIRTY));
arch/powerpc/include/asm/book3s/64/pgtable.h
681
static inline pte_t pte_swp_clear_soft_dirty(pte_t pte)
arch/powerpc/include/asm/book3s/64/pgtable.h
683
return __pte_raw(pte_raw(pte) & cpu_to_be64(~_PAGE_SWP_SOFT_DIRTY));
arch/powerpc/include/asm/book3s/64/pgtable.h
687
static inline pte_t pte_swp_mkexclusive(pte_t pte)
arch/powerpc/include/asm/book3s/64/pgtable.h
689
return __pte_raw(pte_raw(pte) | cpu_to_be64(_PAGE_SWP_EXCLUSIVE));
arch/powerpc/include/asm/book3s/64/pgtable.h
692
static inline bool pte_swp_exclusive(pte_t pte)
arch/powerpc/include/asm/book3s/64/pgtable.h
694
return !!(pte_raw(pte) & cpu_to_be64(_PAGE_SWP_EXCLUSIVE));
arch/powerpc/include/asm/book3s/64/pgtable.h
697
static inline pte_t pte_swp_clear_exclusive(pte_t pte)
arch/powerpc/include/asm/book3s/64/pgtable.h
699
return __pte_raw(pte_raw(pte) & cpu_to_be64(~_PAGE_SWP_EXCLUSIVE));
arch/powerpc/include/asm/book3s/64/pgtable.h
740
static inline int pte_none(pte_t pte)
arch/powerpc/include/asm/book3s/64/pgtable.h
743
return radix__pte_none(pte);
arch/powerpc/include/asm/book3s/64/pgtable.h
744
return hash__pte_none(pte);
arch/powerpc/include/asm/book3s/64/pgtable.h
748
pte_t *ptep, pte_t pte, int percpu)
arch/powerpc/include/asm/book3s/64/pgtable.h
751
VM_WARN_ON(!(pte_raw(pte) & cpu_to_be64(_PAGE_PTE)));
arch/powerpc/include/asm/book3s/64/pgtable.h
756
pte = __pte_raw(pte_raw(pte) | cpu_to_be64(_PAGE_PTE));
arch/powerpc/include/asm/book3s/64/pgtable.h
759
return radix__set_pte_at(mm, addr, ptep, pte, percpu);
arch/powerpc/include/asm/book3s/64/pgtable.h
760
return hash__set_pte_at(mm, addr, ptep, pte, percpu);
arch/powerpc/include/asm/book3s/64/pgtable.h
793
static inline bool pte_ci(pte_t pte)
arch/powerpc/include/asm/book3s/64/pgtable.h
795
__be64 pte_v = pte_raw(pte);
arch/powerpc/include/asm/book3s/64/pgtable.h
887
static inline pud_t pte_pud(pte_t pte)
arch/powerpc/include/asm/book3s/64/pgtable.h
889
return __pud_raw(pte_raw(pte));
arch/powerpc/include/asm/book3s/64/pgtable.h
961
static inline p4d_t pte_p4d(pte_t pte)
arch/powerpc/include/asm/book3s/64/pgtable.h
963
return __p4d_raw(pte_raw(pte));
arch/powerpc/include/asm/book3s/64/radix.h
152
pte_t old_pte, pte_t pte);
arch/powerpc/include/asm/book3s/64/radix.h
207
static inline int radix__pte_none(pte_t pte)
arch/powerpc/include/asm/book3s/64/radix.h
209
return (pte_val(pte) & ~RADIX_PTE_NONE_MASK) == 0;
arch/powerpc/include/asm/book3s/64/radix.h
213
pte_t *ptep, pte_t pte, int percpu)
arch/powerpc/include/asm/book3s/64/radix.h
215
*ptep = pte;
arch/powerpc/include/asm/book3s/64/tlbflush-hash.h
17
real_pte_t pte[PPC64_TLB_BATCH_NR];
arch/powerpc/include/asm/book3s/64/tlbflush-hash.h
60
extern void flush_hash_page(unsigned long vpn, real_pte_t pte, int psize,
arch/powerpc/include/asm/hugetlb.h
44
pte_t pte, unsigned long sz);
arch/powerpc/include/asm/hugetlb.h
58
pte_t pte;
arch/powerpc/include/asm/hugetlb.h
61
pte = huge_ptep_get_and_clear(vma->vm_mm, addr, ptep, sz);
arch/powerpc/include/asm/hugetlb.h
63
return pte;
arch/powerpc/include/asm/hugetlb.h
69
pte_t pte, int dirty);
arch/powerpc/include/asm/kasan.h
67
void kasan_update_early_region(unsigned long k_start, unsigned long k_end, pte_t pte);
arch/powerpc/include/asm/kvm_book3s.h
153
extern int kvmppc_mmu_map_page(struct kvm_vcpu *vcpu, struct kvmppc_pte *pte,
arch/powerpc/include/asm/kvm_book3s.h
155
extern void kvmppc_mmu_unmap_page(struct kvm_vcpu *vcpu, struct kvmppc_pte *pte);
arch/powerpc/include/asm/kvm_book3s.h
166
extern void kvmppc_mmu_hpte_cache_map(struct kvm_vcpu *vcpu, struct hpte_cache *pte);
arch/powerpc/include/asm/kvm_book3s.h
168
extern void kvmppc_mmu_hpte_cache_free(struct hpte_cache *pte);
arch/powerpc/include/asm/kvm_book3s.h
171
extern void kvmppc_mmu_invalidate_pte(struct kvm_vcpu *vcpu, struct hpte_cache *pte);
arch/powerpc/include/asm/kvm_book3s.h
196
extern void kvmppc_unmap_pte(struct kvm *kvm, pte_t *pte, unsigned long gpa,
arch/powerpc/include/asm/kvm_book3s.h
58
struct kvmppc_pte pte;
arch/powerpc/include/asm/kvm_book3s_64.h
625
extern int kvmppc_create_pte(struct kvm *kvm, pgd_t *pgtable, pte_t pte,
arch/powerpc/include/asm/kvm_book3s_64.h
643
pte_t *pte;
arch/powerpc/include/asm/kvm_book3s_64.h
645
pte = __find_linux_pte(kvm->arch.pgtable, ea, NULL, hshift);
arch/powerpc/include/asm/kvm_book3s_64.h
646
return pte;
arch/powerpc/include/asm/kvm_book3s_64.h
652
pte_t *pte;
arch/powerpc/include/asm/kvm_book3s_64.h
656
pte = __find_linux_pte(kvm->arch.pgtable, ea, NULL, hshift);
arch/powerpc/include/asm/kvm_book3s_64.h
658
return pte;
arch/powerpc/include/asm/kvm_book3s_64.h
664
pte_t *pte;
arch/powerpc/include/asm/kvm_book3s_64.h
672
pte = __find_linux_pte(kvm->mm->pgd, ea, NULL, hshift);
arch/powerpc/include/asm/kvm_book3s_64.h
674
return pte;
arch/powerpc/include/asm/kvm_host.h
401
struct kvmppc_pte *pte, bool data, bool iswrite);
arch/powerpc/include/asm/kvm_ppc.h
118
struct kvmppc_pte *pte);
arch/powerpc/include/asm/nohash/32/hugetlb-8xx.h
20
pte_t pte, unsigned long sz);
arch/powerpc/include/asm/nohash/32/pgalloc.h
18
pte_t *pte)
arch/powerpc/include/asm/nohash/32/pgalloc.h
21
*pmdp = __pmd((unsigned long)pte | _PMD_PRESENT);
arch/powerpc/include/asm/nohash/32/pgalloc.h
23
*pmdp = __pmd(__pa(pte) | _PMD_PRESENT);
arch/powerpc/include/asm/nohash/32/pgtable.h
194
#define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) >> 3 })
arch/powerpc/include/asm/nohash/32/pte-8xx.h
101
static inline int pte_write(pte_t pte)
arch/powerpc/include/asm/nohash/32/pte-8xx.h
103
return !(pte_val(pte) & _PAGE_RO);
arch/powerpc/include/asm/nohash/32/pte-8xx.h
108
static inline pte_t pte_mkwrite_novma(pte_t pte)
arch/powerpc/include/asm/nohash/32/pte-8xx.h
110
return __pte(pte_val(pte) & ~_PAGE_RO);
arch/powerpc/include/asm/nohash/32/pte-8xx.h
115
static inline pte_t pte_mkhuge(pte_t pte)
arch/powerpc/include/asm/nohash/32/pte-8xx.h
117
return __pte(pte_val(pte) | _PAGE_SPS | _PAGE_HUGE);
arch/powerpc/include/asm/nohash/32/pte-8xx.h
144
static inline unsigned long __pte_leaf_size(pmd_t pmd, pte_t pte)
arch/powerpc/include/asm/nohash/32/pte-8xx.h
146
pte_basic_t val = pte_val(pte);
arch/powerpc/include/asm/nohash/32/pte-8xx.h
231
pte_basic_t val = READ_ONCE(ptep->pte);
arch/powerpc/include/asm/nohash/32/pte-8xx.h
232
pte_t pte = {val, val, val, val};
arch/powerpc/include/asm/nohash/32/pte-8xx.h
234
return pte;
arch/powerpc/include/asm/nohash/32/pte-8xx.h
87
static inline pte_t pte_wrprotect(pte_t pte)
arch/powerpc/include/asm/nohash/32/pte-8xx.h
89
return __pte(pte_val(pte) | _PAGE_RO);
arch/powerpc/include/asm/nohash/32/pte-8xx.h
94
static inline int pte_read(pte_t pte)
arch/powerpc/include/asm/nohash/32/pte-8xx.h
96
return (pte_val(pte) & _PAGE_RO) != _PAGE_NA;
arch/powerpc/include/asm/nohash/64/pgalloc.h
40
pte_t *pte)
arch/powerpc/include/asm/nohash/64/pgalloc.h
42
pmd_set(pmd, (unsigned long)pte);
arch/powerpc/include/asm/nohash/64/pgtable-4k.h
77
static inline p4d_t pte_p4d(pte_t pte)
arch/powerpc/include/asm/nohash/64/pgtable-4k.h
79
return __p4d(pte_val(pte));
arch/powerpc/include/asm/nohash/64/pgtable.h
136
static inline pud_t pte_pud(pte_t pte)
arch/powerpc/include/asm/nohash/64/pgtable.h
138
return __pud(pte_val(pte));
arch/powerpc/include/asm/nohash/64/pgtable.h
195
#define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val((pte)) })
arch/powerpc/include/asm/nohash/pgtable.h
159
static inline pte_t pte_mkwrite_novma(pte_t pte)
arch/powerpc/include/asm/nohash/pgtable.h
164
return __pte(pte_val(pte) | _PAGE_RW);
arch/powerpc/include/asm/nohash/pgtable.h
168
static inline pte_t pte_mkdirty(pte_t pte)
arch/powerpc/include/asm/nohash/pgtable.h
170
return __pte(pte_val(pte) | _PAGE_DIRTY);
arch/powerpc/include/asm/nohash/pgtable.h
173
static inline pte_t pte_mkyoung(pte_t pte)
arch/powerpc/include/asm/nohash/pgtable.h
175
return __pte(pte_val(pte) | _PAGE_ACCESSED);
arch/powerpc/include/asm/nohash/pgtable.h
179
static inline pte_t pte_wrprotect(pte_t pte)
arch/powerpc/include/asm/nohash/pgtable.h
181
return __pte(pte_val(pte) & ~_PAGE_WRITE);
arch/powerpc/include/asm/nohash/pgtable.h
186
static inline pte_t pte_mkexec(pte_t pte)
arch/powerpc/include/asm/nohash/pgtable.h
188
return __pte(pte_val(pte) | _PAGE_EXEC);
arch/powerpc/include/asm/nohash/pgtable.h
193
static inline int pte_write(pte_t pte)
arch/powerpc/include/asm/nohash/pgtable.h
195
return pte_val(pte) & _PAGE_WRITE;
arch/powerpc/include/asm/nohash/pgtable.h
198
static inline int pte_dirty(pte_t pte) { return pte_val(pte) & _PAGE_DIRTY; }
arch/powerpc/include/asm/nohash/pgtable.h
199
static inline int pte_special(pte_t pte) { return pte_val(pte) & _PAGE_SPECIAL; }
arch/powerpc/include/asm/nohash/pgtable.h
200
static inline int pte_none(pte_t pte) { return (pte_val(pte) & ~_PTE_NONE_MASK) == 0; }
arch/powerpc/include/asm/nohash/pgtable.h
201
static inline bool pte_hashpte(pte_t pte) { return false; }
arch/powerpc/include/asm/nohash/pgtable.h
202
static inline bool pte_ci(pte_t pte) { return pte_val(pte) & _PAGE_NO_CACHE; }
arch/powerpc/include/asm/nohash/pgtable.h
203
static inline bool pte_exec(pte_t pte) { return pte_val(pte) & _PAGE_EXEC; }
arch/powerpc/include/asm/nohash/pgtable.h
205
static inline int pte_present(pte_t pte)
arch/powerpc/include/asm/nohash/pgtable.h
207
return pte_val(pte) & _PAGE_PRESENT;
arch/powerpc/include/asm/nohash/pgtable.h
210
static inline bool pte_hw_valid(pte_t pte)
arch/powerpc/include/asm/nohash/pgtable.h
212
return pte_val(pte) & _PAGE_PRESENT;
arch/powerpc/include/asm/nohash/pgtable.h
215
static inline int pte_young(pte_t pte)
arch/powerpc/include/asm/nohash/pgtable.h
217
return pte_val(pte) & _PAGE_ACCESSED;
arch/powerpc/include/asm/nohash/pgtable.h
226
static inline bool pte_read(pte_t pte)
arch/powerpc/include/asm/nohash/pgtable.h
228
return (pte_val(pte) & _PAGE_READ) == _PAGE_READ;
arch/powerpc/include/asm/nohash/pgtable.h
237
static inline bool pte_access_permitted(pte_t pte, bool write)
arch/powerpc/include/asm/nohash/pgtable.h
243
if (!pte_present(pte) || !pte_read(pte))
arch/powerpc/include/asm/nohash/pgtable.h
246
if (write && !pte_write(pte))
arch/powerpc/include/asm/nohash/pgtable.h
252
static inline bool pte_user_accessible_page(pte_t pte, unsigned long addr)
arch/powerpc/include/asm/nohash/pgtable.h
254
return pte_present(pte) && !is_kernel_addr(addr);
arch/powerpc/include/asm/nohash/pgtable.h
268
static inline pte_t pte_exprotect(pte_t pte)
arch/powerpc/include/asm/nohash/pgtable.h
270
return __pte(pte_val(pte) & ~_PAGE_EXEC);
arch/powerpc/include/asm/nohash/pgtable.h
273
static inline pte_t pte_mkclean(pte_t pte)
arch/powerpc/include/asm/nohash/pgtable.h
275
return __pte(pte_val(pte) & ~_PAGE_DIRTY);
arch/powerpc/include/asm/nohash/pgtable.h
278
static inline pte_t pte_mkold(pte_t pte)
arch/powerpc/include/asm/nohash/pgtable.h
280
return __pte(pte_val(pte) & ~_PAGE_ACCESSED);
arch/powerpc/include/asm/nohash/pgtable.h
283
static inline pte_t pte_mkspecial(pte_t pte)
arch/powerpc/include/asm/nohash/pgtable.h
285
return __pte(pte_val(pte) | _PAGE_SPECIAL);
arch/powerpc/include/asm/nohash/pgtable.h
289
static inline pte_t pte_mkhuge(pte_t pte)
arch/powerpc/include/asm/nohash/pgtable.h
291
return __pte(pte_val(pte));
arch/powerpc/include/asm/nohash/pgtable.h
295
static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
arch/powerpc/include/asm/nohash/pgtable.h
297
return __pte((pte_val(pte) & _PAGE_CHG_MASK) | pgprot_val(newprot));
arch/powerpc/include/asm/nohash/pgtable.h
300
static inline bool pte_swp_exclusive(pte_t pte)
arch/powerpc/include/asm/nohash/pgtable.h
302
return pte_val(pte) & _PAGE_SWP_EXCLUSIVE;
arch/powerpc/include/asm/nohash/pgtable.h
305
static inline pte_t pte_swp_mkexclusive(pte_t pte)
arch/powerpc/include/asm/nohash/pgtable.h
307
return __pte(pte_val(pte) | _PAGE_SWP_EXCLUSIVE);
arch/powerpc/include/asm/nohash/pgtable.h
310
static inline pte_t pte_swp_clear_exclusive(pte_t pte)
arch/powerpc/include/asm/nohash/pgtable.h
312
return __pte(pte_val(pte) & ~_PAGE_SWP_EXCLUSIVE);
arch/powerpc/include/asm/nohash/pgtable.h
321
pte_t *ptep, pte_t pte, int percpu)
arch/powerpc/include/asm/nohash/pgtable.h
334
: "r" (pte) : "memory");
arch/powerpc/include/asm/nohash/pgtable.h
341
ptep->pte3 = ptep->pte2 = ptep->pte1 = ptep->pte = pte_val(pte);
arch/powerpc/include/asm/nohash/pgtable.h
343
*ptep = pte;
arch/powerpc/include/asm/nohash/pgtable.h
37
static inline unsigned long pte_huge_size(pte_t pte)
arch/powerpc/include/asm/nohash/pte-e500.h
90
static inline pte_t pte_mkexec(pte_t pte)
arch/powerpc/include/asm/nohash/pte-e500.h
92
return __pte((pte_val(pte) & ~_PAGE_BAP_SX) | _PAGE_BAP_UX);
arch/powerpc/include/asm/nohash/pte-e500.h
96
static inline unsigned long pte_huge_size(pte_t pte)
arch/powerpc/include/asm/nohash/pte-e500.h
98
pte_basic_t val = pte_val(pte);
arch/powerpc/include/asm/pgalloc.h
38
static inline void pte_free_kernel(struct mm_struct *mm, pte_t *pte)
arch/powerpc/include/asm/pgalloc.h
40
pte_fragment_free((unsigned long *)pte, 1);
arch/powerpc/include/asm/pgtable-be-types.h
13
return be64_to_cpu(x.pte);
arch/powerpc/include/asm/pgtable-be-types.h
18
return x.pte;
arch/powerpc/include/asm/pgtable-be-types.h
76
typedef struct { pte_t pte; unsigned long hidx; } real_pte_t;
arch/powerpc/include/asm/pgtable-be-types.h
78
typedef struct { pte_t pte; } real_pte_t;
arch/powerpc/include/asm/pgtable-be-types.h
8
typedef struct { __be64 pte; } pte_t;
arch/powerpc/include/asm/pgtable-types.h
11
typedef struct { pte_basic_t pte, pte1, pte2, pte3; } pte_t;
arch/powerpc/include/asm/pgtable-types.h
13
typedef struct { pte_basic_t pte; } pte_t;
arch/powerpc/include/asm/pgtable-types.h
23
return x.pte;
arch/powerpc/include/asm/pgtable-types.h
79
typedef struct { pte_t pte; unsigned long hidx; } real_pte_t;
arch/powerpc/include/asm/pgtable-types.h
81
typedef struct { pte_t pte; } real_pte_t;
arch/powerpc/include/asm/pgtable.h
35
pte_t pte, unsigned int nr);
arch/powerpc/include/asm/pgtable.h
38
pte_t *ptep, pte_t pte);
arch/powerpc/include/asm/pgtable.h
49
static inline unsigned long pte_pfn(pte_t pte)
arch/powerpc/include/asm/pgtable.h
51
return (pte_val(pte) & PTE_RPN_MASK) >> PTE_RPN_SHIFT;
arch/powerpc/include/asm/pgtable.h
58
static inline pgprot_t pte_pgprot(pte_t pte)
arch/powerpc/include/asm/pgtable.h
62
pte_flags = pte_val(pte) & ~PTE_RPN_MASK;
arch/powerpc/include/asm/pte-walk.h
13
pte_t *pte;
arch/powerpc/include/asm/pte-walk.h
16
pte = __find_linux_pte(pgdir, ea, is_thp, hshift);
arch/powerpc/include/asm/pte-walk.h
26
return pte;
arch/powerpc/kernel/mce_power.c
31
pte_t *ptep, pte;
arch/powerpc/kernel/mce_power.c
47
pte = READ_ONCE(*ptep);
arch/powerpc/kernel/mce_power.c
49
if (!pte_present(pte) || pte_special(pte)) {
arch/powerpc/kernel/mce_power.c
55
pfn = pte_pfn(pte);
arch/powerpc/kernel/mce_power.c
58
pfn = pte_pfn(__pte(pte_val(pte) | (addr & rpnmask)));
arch/powerpc/kvm/book3s.c
452
enum xlate_readwrite xlrw, struct kvmppc_pte *pte)
arch/powerpc/kvm/book3s.c
460
r = vcpu->arch.mmu.xlate(vcpu, eaddr, pte, data, iswrite);
arch/powerpc/kvm/book3s.c
462
pte->eaddr = eaddr;
arch/powerpc/kvm/book3s.c
463
pte->raddr = eaddr & KVM_PAM;
arch/powerpc/kvm/book3s.c
464
pte->vpage = VSID_REAL | eaddr >> 12;
arch/powerpc/kvm/book3s.c
465
pte->may_read = true;
arch/powerpc/kvm/book3s.c
466
pte->may_write = true;
arch/powerpc/kvm/book3s.c
467
pte->may_execute = true;
arch/powerpc/kvm/book3s.c
474
pte->raddr &= ~SPLIT_HACK_MASK;
arch/powerpc/kvm/book3s_32_mmu.c
128
struct kvmppc_pte *pte, bool data,
arch/powerpc/kvm/book3s_32_mmu.c
160
pte->vpage = (((u64)eaddr >> 12) & 0xffff) | vsid;
arch/powerpc/kvm/book3s_32_mmu.c
162
pte->raddr = bat->brpn | (eaddr & ~bat->bepi_mask);
arch/powerpc/kvm/book3s_32_mmu.c
163
pte->may_read = bat->pp;
arch/powerpc/kvm/book3s_32_mmu.c
164
pte->may_write = bat->pp > 1;
arch/powerpc/kvm/book3s_32_mmu.c
165
pte->may_execute = true;
arch/powerpc/kvm/book3s_32_mmu.c
166
if (!pte->may_read) {
arch/powerpc/kvm/book3s_32_mmu.c
170
if (iswrite && !pte->may_write) {
arch/powerpc/kvm/book3s_32_mmu.c
183
struct kvmppc_pte *pte, bool data,
arch/powerpc/kvm/book3s_32_mmu.c
199
pte->vpage = kvmppc_mmu_book3s_32_ea_to_vp(vcpu, eaddr, data);
arch/powerpc/kvm/book3s_32_mmu.c
221
pte->raddr = (pte1 & ~(0xFFFULL)) | (eaddr & 0xFFF);
arch/powerpc/kvm/book3s_32_mmu.c
228
pte->may_write = false;
arch/powerpc/kvm/book3s_32_mmu.c
229
pte->may_read = false;
arch/powerpc/kvm/book3s_32_mmu.c
230
pte->may_execute = true;
arch/powerpc/kvm/book3s_32_mmu.c
236
pte->may_write = true;
arch/powerpc/kvm/book3s_32_mmu.c
241
pte->may_read = true;
arch/powerpc/kvm/book3s_32_mmu.c
262
if (pte->may_read && !(pte_r & PTEG_FLAG_ACCESSED)) {
arch/powerpc/kvm/book3s_32_mmu.c
266
if (iswrite && pte->may_write && !(pte_r & PTEG_FLAG_DIRTY)) {
arch/powerpc/kvm/book3s_32_mmu.c
270
if (!pte->may_read || (iswrite && !pte->may_write))
arch/powerpc/kvm/book3s_32_mmu.c
291
struct kvmppc_pte *pte, bool data,
arch/powerpc/kvm/book3s_32_mmu.c
297
pte->eaddr = eaddr;
arch/powerpc/kvm/book3s_32_mmu.c
298
pte->page_size = MMU_PAGE_4K;
arch/powerpc/kvm/book3s_32_mmu.c
304
pte->vpage = kvmppc_mmu_book3s_32_ea_to_vp(vcpu, eaddr, data);
arch/powerpc/kvm/book3s_32_mmu.c
305
pte->raddr = vcpu->arch.magic_page_pa | (pte->raddr & 0xfff);
arch/powerpc/kvm/book3s_32_mmu.c
306
pte->raddr &= KVM_PAM;
arch/powerpc/kvm/book3s_32_mmu.c
307
pte->may_execute = true;
arch/powerpc/kvm/book3s_32_mmu.c
308
pte->may_read = true;
arch/powerpc/kvm/book3s_32_mmu.c
309
pte->may_write = true;
arch/powerpc/kvm/book3s_32_mmu.c
314
r = kvmppc_mmu_book3s_32_xlate_bat(vcpu, eaddr, pte, data, iswrite);
arch/powerpc/kvm/book3s_32_mmu.c
316
r = kvmppc_mmu_book3s_32_xlate_pte(vcpu, eaddr, pte,
arch/powerpc/kvm/book3s_32_mmu.c
319
r = kvmppc_mmu_book3s_32_xlate_pte(vcpu, eaddr, pte,
arch/powerpc/kvm/book3s_32_mmu.c
70
struct kvmppc_pte *pte, bool data,
arch/powerpc/kvm/book3s_32_mmu.c
84
struct kvmppc_pte pte;
arch/powerpc/kvm/book3s_32_mmu.c
86
if (!kvmppc_mmu_book3s_32_xlate_bat(vcpu, eaddr, &pte, data, false))
arch/powerpc/kvm/book3s_32_mmu.c
87
return pte.vpage;
arch/powerpc/kvm/book3s_32_mmu_host.c
142
struct hpte_cache *pte;
arch/powerpc/kvm/book3s_32_mmu_host.c
232
pte = kvmppc_mmu_hpte_cache_next(vcpu);
arch/powerpc/kvm/book3s_32_mmu_host.c
233
if (!pte) {
arch/powerpc/kvm/book3s_32_mmu_host.c
245
pte->slot = (ulong)&pteg[rr];
arch/powerpc/kvm/book3s_32_mmu_host.c
246
pte->host_vpn = vpn;
arch/powerpc/kvm/book3s_32_mmu_host.c
247
pte->pte = *orig_pte;
arch/powerpc/kvm/book3s_32_mmu_host.c
248
pte->pfn = hpaddr >> PAGE_SHIFT;
arch/powerpc/kvm/book3s_32_mmu_host.c
250
kvmppc_mmu_hpte_cache_map(vcpu, pte);
arch/powerpc/kvm/book3s_32_mmu_host.c
257
void kvmppc_mmu_unmap_page(struct kvm_vcpu *vcpu, struct kvmppc_pte *pte)
arch/powerpc/kvm/book3s_32_mmu_host.c
259
kvmppc_mmu_pte_vflush(vcpu, pte->vpage, 0xfffffffffULL);
arch/powerpc/kvm/book3s_32_mmu_host.c
49
void kvmppc_mmu_invalidate_pte(struct kvm_vcpu *vcpu, struct hpte_cache *pte)
arch/powerpc/kvm/book3s_32_mmu_host.c
54
pteg = (u32*)pte->slot;
arch/powerpc/kvm/book3s_32_mmu_host.c
59
asm volatile ("tlbie %0" : : "r" (pte->pte.eaddr) : "memory");
arch/powerpc/kvm/book3s_64_mmu_host.c
194
cpte->pte = *orig_pte;
arch/powerpc/kvm/book3s_64_mmu_host.c
214
void kvmppc_mmu_unmap_page(struct kvm_vcpu *vcpu, struct kvmppc_pte *pte)
arch/powerpc/kvm/book3s_64_mmu_host.c
219
vcpu->arch.mmu.esid_to_vsid(vcpu, pte->eaddr >> SID_SHIFT, &vsid);
arch/powerpc/kvm/book3s_64_mmu_host.c
222
kvmppc_mmu_pte_vflush(vcpu, pte->vpage, mask);
arch/powerpc/kvm/book3s_64_mmu_host.c
24
void kvmppc_mmu_invalidate_pte(struct kvm_vcpu *vcpu, struct hpte_cache *pte)
arch/powerpc/kvm/book3s_64_mmu_host.c
26
mmu_hash_ops.hpte_invalidate(pte->slot, pte->host_vpn,
arch/powerpc/kvm/book3s_64_mmu_host.c
27
pte->pagesize, pte->pagesize,
arch/powerpc/kvm/book3s_64_mmu_hv.c
526
pte_t pte, *ptep;
arch/powerpc/kvm/book3s_64_mmu_hv.c
617
pte = __pte(0);
arch/powerpc/kvm/book3s_64_mmu_hv.c
619
pte = READ_ONCE(*ptep);
arch/powerpc/kvm/book3s_64_mmu_hv.c
625
if (!pte_present(pte)) {
arch/powerpc/kvm/book3s_64_mmu_hv.c
630
hpa = pte_pfn(pte) << PAGE_SHIFT;
arch/powerpc/kvm/book3s_64_mmu_hv.c
634
is_ci = pte_ci(pte);
arch/powerpc/kvm/book3s_64_mmu_radix.c
1082
pte_t *ptep, pte;
arch/powerpc/kvm/book3s_64_mmu_radix.c
1098
pte = READ_ONCE(*ptep);
arch/powerpc/kvm/book3s_64_mmu_radix.c
1099
if (pte_present(pte) && pte_dirty(pte)) {
arch/powerpc/kvm/book3s_64_mmu_radix.c
1104
if (pte_val(pte) != pte_val(*ptep)) {
arch/powerpc/kvm/book3s_64_mmu_radix.c
1296
unsigned long pte;
arch/powerpc/kvm/book3s_64_mmu_radix.c
1378
pte = pud_val(pud);
arch/powerpc/kvm/book3s_64_mmu_radix.c
1390
pte = pmd_val(pmd);
arch/powerpc/kvm/book3s_64_mmu_radix.c
1396
pte = pte_val(READ_ONCE(*ptep));
arch/powerpc/kvm/book3s_64_mmu_radix.c
1397
if (!(pte & _PAGE_PRESENT)) {
arch/powerpc/kvm/book3s_64_mmu_radix.c
1404
" %lx: %lx %d\n", gpa, pte, shift);
arch/powerpc/kvm/book3s_64_mmu_radix.c
149
u64 pte, base, gpa;
arch/powerpc/kvm/book3s_64_mmu_radix.c
187
pte = __be64_to_cpu(rpte);
arch/powerpc/kvm/book3s_64_mmu_radix.c
188
if (!(pte & _PAGE_PRESENT))
arch/powerpc/kvm/book3s_64_mmu_radix.c
191
if (pte & _PAGE_PTE)
arch/powerpc/kvm/book3s_64_mmu_radix.c
194
base = pte & RPDB_MASK;
arch/powerpc/kvm/book3s_64_mmu_radix.c
195
bits = pte & RPDS_MASK;
arch/powerpc/kvm/book3s_64_mmu_radix.c
204
gpa = pte & 0x01fffffffffff000ul;
arch/powerpc/kvm/book3s_64_mmu_radix.c
218
gpte->may_read = !!(pte & _PAGE_READ);
arch/powerpc/kvm/book3s_64_mmu_radix.c
219
gpte->may_write = !!(pte & _PAGE_WRITE);
arch/powerpc/kvm/book3s_64_mmu_radix.c
220
gpte->may_execute = !!(pte & _PAGE_EXEC);
arch/powerpc/kvm/book3s_64_mmu_radix.c
222
gpte->rc = pte & (_PAGE_ACCESSED | _PAGE_DIRTY);
arch/powerpc/kvm/book3s_64_mmu_radix.c
225
*pte_ret_p = pte;
arch/powerpc/kvm/book3s_64_mmu_radix.c
273
u64 pte;
arch/powerpc/kvm/book3s_64_mmu_radix.c
289
vcpu->kvm->arch.process_table, pid, &pte);
arch/powerpc/kvm/book3s_64_mmu_radix.c
295
if (pte & _PAGE_PRIVILEGED) {
arch/powerpc/kvm/book3s_64_mmu_radix.c
301
if (!(pte & _PAGE_PRIVILEGED)) {
arch/powerpc/kvm/book3s_64_mmu_radix.c
382
pte_t *ptep, pte_t pte)
arch/powerpc/kvm/book3s_64_mmu_radix.c
384
radix__set_pte_at(kvm->mm, addr, ptep, pte, 0);
arch/powerpc/kvm/book3s_64_mmu_radix.c
392
pte_t *pte;
arch/powerpc/kvm/book3s_64_mmu_radix.c
394
pte = kmem_cache_alloc(kvm_pte_cache, GFP_KERNEL);
arch/powerpc/kvm/book3s_64_mmu_radix.c
396
kmemleak_ignore(pte);
arch/powerpc/kvm/book3s_64_mmu_radix.c
398
return pte;
arch/powerpc/kvm/book3s_64_mmu_radix.c
423
void kvmppc_unmap_pte(struct kvm *kvm, pte_t *pte, unsigned long gpa,
arch/powerpc/kvm/book3s_64_mmu_radix.c
434
old = kvmppc_radix_update_pte(kvm, pte, ~0UL, 0, gpa, shift);
arch/powerpc/kvm/book3s_64_mmu_radix.c
476
static void kvmppc_unmap_free_pte(struct kvm *kvm, pte_t *pte, bool full,
arch/powerpc/kvm/book3s_64_mmu_radix.c
480
memset(pte, 0, sizeof(long) << RADIX_PTE_INDEX_SIZE);
arch/powerpc/kvm/book3s_64_mmu_radix.c
482
pte_t *p = pte;
arch/powerpc/kvm/book3s_64_mmu_radix.c
494
kvmppc_pte_free(pte);
arch/powerpc/kvm/book3s_64_mmu_radix.c
516
pte_t *pte;
arch/powerpc/kvm/book3s_64_mmu_radix.c
518
pte = pte_offset_kernel(p, 0);
arch/powerpc/kvm/book3s_64_mmu_radix.c
519
kvmppc_unmap_free_pte(kvm, pte, full, lpid);
arch/powerpc/kvm/book3s_64_mmu_radix.c
577
pte_t *pte = pte_offset_kernel(pmd, 0);
arch/powerpc/kvm/book3s_64_mmu_radix.c
587
kvmppc_unmap_free_pte(kvm, pte, false, lpid);
arch/powerpc/kvm/book3s_64_mmu_radix.c
615
int kvmppc_create_pte(struct kvm *kvm, pgd_t *pgtable, pte_t pte,
arch/powerpc/kvm/book3s_64_mmu_radix.c
666
if (pud_raw(*pud) == pte_raw(pte)) {
arch/powerpc/kvm/book3s_64_mmu_radix.c
671
WARN_ON_ONCE((pud_val(*pud) ^ pte_val(pte)) &
arch/powerpc/kvm/book3s_64_mmu_radix.c
674
0, pte_val(pte), hgpa, PUD_SHIFT);
arch/powerpc/kvm/book3s_64_mmu_radix.c
699
kvmppc_radix_set_pte_at(kvm, gpa, (pte_t *)pud, pte);
arch/powerpc/kvm/book3s_64_mmu_radix.c
717
if (pmd_raw(*pmd) == pte_raw(pte)) {
arch/powerpc/kvm/book3s_64_mmu_radix.c
722
WARN_ON_ONCE((pmd_val(*pmd) ^ pte_val(pte)) &
arch/powerpc/kvm/book3s_64_mmu_radix.c
725
0, pte_val(pte), lgpa, PMD_SHIFT);
arch/powerpc/kvm/book3s_64_mmu_radix.c
751
kvmppc_radix_set_pte_at(kvm, gpa, pmdp_ptep(pmd), pte);
arch/powerpc/kvm/book3s_64_mmu_radix.c
766
if (pte_raw(*ptep) == pte_raw(pte)) {
arch/powerpc/kvm/book3s_64_mmu_radix.c
771
WARN_ON_ONCE((pte_val(*ptep) ^ pte_val(pte)) &
arch/powerpc/kvm/book3s_64_mmu_radix.c
773
kvmppc_radix_update_pte(kvm, ptep, 0, pte_val(pte), gpa, 0);
arch/powerpc/kvm/book3s_64_mmu_radix.c
777
kvmppc_radix_set_pte_at(kvm, gpa, ptep, pte);
arch/powerpc/kvm/book3s_64_mmu_radix.c
832
pte_t pte, *ptep;
arch/powerpc/kvm/book3s_64_mmu_radix.c
854
pte = __pte(0);
arch/powerpc/kvm/book3s_64_mmu_radix.c
856
pte = READ_ONCE(*ptep);
arch/powerpc/kvm/book3s_64_mmu_radix.c
862
if (!pte_present(pte)) {
arch/powerpc/kvm/book3s_64_mmu_radix.c
889
pte = __pte(pte_val(pte) | (hva & rpnmask));
arch/powerpc/kvm/book3s_64_mmu_radix.c
893
pte = __pte(pte_val(pte) | _PAGE_EXEC | _PAGE_ACCESSED);
arch/powerpc/kvm/book3s_64_mmu_radix.c
895
if (pte_val(pte) & _PAGE_WRITE)
arch/powerpc/kvm/book3s_64_mmu_radix.c
896
pte = __pte(pte_val(pte) | _PAGE_DIRTY);
arch/powerpc/kvm/book3s_64_mmu_radix.c
898
pte = __pte(pte_val(pte) & ~(_PAGE_WRITE | _PAGE_DIRTY));
arch/powerpc/kvm/book3s_64_mmu_radix.c
902
ret = kvmppc_create_pte(kvm, kvm->arch.pgtable, pte, gpa, level,
arch/powerpc/kvm/book3s_64_mmu_radix.c
905
*inserted_pte = pte;
arch/powerpc/kvm/book3s_64_mmu_radix.c
910
if (!ret && (pte_val(pte) & _PAGE_WRITE))
arch/powerpc/kvm/book3s_hv_nested.c
1530
pte_t pte, *pte_p;
arch/powerpc/kvm/book3s_hv_nested.c
1626
pte = __pte(0);
arch/powerpc/kvm/book3s_hv_nested.c
1632
pte = *pte_p;
arch/powerpc/kvm/book3s_hv_nested.c
1635
if (!pte_present(pte) || (writing && !(pte_val(pte) & _PAGE_WRITE))) {
arch/powerpc/kvm/book3s_hv_nested.c
1638
writing, &pte, &level);
arch/powerpc/kvm/book3s_hv_nested.c
1657
pte = __pte(pte_val(pte) & ~perm);
arch/powerpc/kvm/book3s_hv_nested.c
1666
pte = __pte(pte_val(pte) | (gpa & mask));
arch/powerpc/kvm/book3s_hv_nested.c
1680
ret = kvmppc_create_pte(kvm, gp->shadow_pgtable, pte, n_gpa, level,
arch/powerpc/kvm/book3s_hv_nested.c
903
pte_t *pte;
arch/powerpc/kvm/book3s_hv_nested.c
911
pte = __find_linux_pte(gp->shadow_pgtable, ea, NULL, hshift);
arch/powerpc/kvm/book3s_hv_nested.c
913
return pte;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
251
pte_t pte;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
266
pte = kvmppc_read_update_linux_pte(ptep, writing);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
267
if (pte_present(pte) && !pte_protnone(pte)) {
arch/powerpc/kvm/book3s_hv_rm_mmu.c
268
if (writing && !pte_write(pte))
arch/powerpc/kvm/book3s_hv_rm_mmu.c
271
is_ci = pte_ci(pte);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
272
pa = pte_pfn(pte) << PAGE_SHIFT;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
324
u64 pte;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
327
pte = be64_to_cpu(hpte[0]);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
328
if (!(pte & (HPTE_V_VALID | HPTE_V_ABSENT)))
arch/powerpc/kvm/book3s_hv_rm_mmu.c
330
__unlock_hpte(hpte, pte);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
342
u64 pte;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
346
pte = be64_to_cpu(hpte[0]);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
347
if (pte & (HPTE_V_VALID | HPTE_V_ABSENT)) {
arch/powerpc/kvm/book3s_hv_rm_mmu.c
348
__unlock_hpte(hpte, pte);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
492
u64 pte, orig_pte, pte_r;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
501
pte = orig_pte = be64_to_cpu(hpte[0]);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
504
pte = hpte_new_to_old_v(pte, pte_r);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
507
if ((pte & (HPTE_V_ABSENT | HPTE_V_VALID)) == 0 ||
arch/powerpc/kvm/book3s_hv_rm_mmu.c
508
((flags & H_AVPN) && (pte & ~0x7fUL) != avpn) ||
arch/powerpc/kvm/book3s_hv_rm_mmu.c
509
((flags & H_ANDCOND) && (pte & avpn) != 0)) {
arch/powerpc/kvm/book3s_hv_rm_mmu.c
515
v = pte & ~HPTE_V_HVLOCK;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
891
pte_t *ptep, pte;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
906
pte = kvmppc_read_update_linux_pte(ptep, writing);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
907
if (!pte_present(pte))
arch/powerpc/kvm/book3s_hv_rm_mmu.c
913
pa = pte_pfn(pte) << PAGE_SHIFT;
arch/powerpc/kvm/book3s_mmu_hpte.c
102
kvmppc_mmu_invalidate_pte(vcpu, pte);
arch/powerpc/kvm/book3s_mmu_hpte.c
107
if (hlist_unhashed(&pte->list_pte)) {
arch/powerpc/kvm/book3s_mmu_hpte.c
112
hlist_del_init_rcu(&pte->list_pte);
arch/powerpc/kvm/book3s_mmu_hpte.c
113
hlist_del_init_rcu(&pte->list_pte_long);
arch/powerpc/kvm/book3s_mmu_hpte.c
114
hlist_del_init_rcu(&pte->list_vpte);
arch/powerpc/kvm/book3s_mmu_hpte.c
115
hlist_del_init_rcu(&pte->list_vpte_long);
arch/powerpc/kvm/book3s_mmu_hpte.c
117
hlist_del_init_rcu(&pte->list_vpte_64k);
arch/powerpc/kvm/book3s_mmu_hpte.c
123
kfree_rcu(pte, rcu_head);
arch/powerpc/kvm/book3s_mmu_hpte.c
129
struct hpte_cache *pte;
arch/powerpc/kvm/book3s_mmu_hpte.c
137
hlist_for_each_entry_rcu(pte, list, list_vpte_long)
arch/powerpc/kvm/book3s_mmu_hpte.c
138
invalidate_pte(vcpu, pte);
arch/powerpc/kvm/book3s_mmu_hpte.c
148
struct hpte_cache *pte;
arch/powerpc/kvm/book3s_mmu_hpte.c
156
hlist_for_each_entry_rcu(pte, list, list_pte)
arch/powerpc/kvm/book3s_mmu_hpte.c
157
if ((pte->pte.eaddr & ~0xfffUL) == guest_ea)
arch/powerpc/kvm/book3s_mmu_hpte.c
158
invalidate_pte(vcpu, pte);
arch/powerpc/kvm/book3s_mmu_hpte.c
167
struct hpte_cache *pte;
arch/powerpc/kvm/book3s_mmu_hpte.c
176
hlist_for_each_entry_rcu(pte, list, list_pte_long)
arch/powerpc/kvm/book3s_mmu_hpte.c
177
if ((pte->pte.eaddr & 0x0ffff000UL) == guest_ea)
arch/powerpc/kvm/book3s_mmu_hpte.c
178
invalidate_pte(vcpu, pte);
arch/powerpc/kvm/book3s_mmu_hpte.c
210
struct hpte_cache *pte;
arch/powerpc/kvm/book3s_mmu_hpte.c
218
hlist_for_each_entry_rcu(pte, list, list_vpte)
arch/powerpc/kvm/book3s_mmu_hpte.c
219
if ((pte->pte.vpage & vp_mask) == guest_vp)
arch/powerpc/kvm/book3s_mmu_hpte.c
220
invalidate_pte(vcpu, pte);
arch/powerpc/kvm/book3s_mmu_hpte.c
231
struct hpte_cache *pte;
arch/powerpc/kvm/book3s_mmu_hpte.c
240
hlist_for_each_entry_rcu(pte, list, list_vpte_64k)
arch/powerpc/kvm/book3s_mmu_hpte.c
241
if ((pte->pte.vpage & vp_mask) == guest_vp)
arch/powerpc/kvm/book3s_mmu_hpte.c
242
invalidate_pte(vcpu, pte);
arch/powerpc/kvm/book3s_mmu_hpte.c
253
struct hpte_cache *pte;
arch/powerpc/kvm/book3s_mmu_hpte.c
262
hlist_for_each_entry_rcu(pte, list, list_vpte_long)
arch/powerpc/kvm/book3s_mmu_hpte.c
263
if ((pte->pte.vpage & vp_mask) == guest_vp)
arch/powerpc/kvm/book3s_mmu_hpte.c
264
invalidate_pte(vcpu, pte);
arch/powerpc/kvm/book3s_mmu_hpte.c
295
struct hpte_cache *pte;
arch/powerpc/kvm/book3s_mmu_hpte.c
305
hlist_for_each_entry_rcu(pte, list, list_vpte_long)
arch/powerpc/kvm/book3s_mmu_hpte.c
306
if ((pte->pte.raddr >= pa_start) &&
arch/powerpc/kvm/book3s_mmu_hpte.c
307
(pte->pte.raddr < pa_end))
arch/powerpc/kvm/book3s_mmu_hpte.c
308
invalidate_pte(vcpu, pte);
arch/powerpc/kvm/book3s_mmu_hpte.c
317
struct hpte_cache *pte;
arch/powerpc/kvm/book3s_mmu_hpte.c
322
pte = kmem_cache_zalloc(hpte_cache, GFP_KERNEL);
arch/powerpc/kvm/book3s_mmu_hpte.c
324
return pte;
arch/powerpc/kvm/book3s_mmu_hpte.c
327
void kvmppc_mmu_hpte_cache_free(struct hpte_cache *pte)
arch/powerpc/kvm/book3s_mmu_hpte.c
329
kmem_cache_free(hpte_cache, pte);
arch/powerpc/kvm/book3s_mmu_hpte.c
56
void kvmppc_mmu_hpte_cache_map(struct kvm_vcpu *vcpu, struct hpte_cache *pte)
arch/powerpc/kvm/book3s_mmu_hpte.c
61
trace_kvm_book3s_mmu_map(pte);
arch/powerpc/kvm/book3s_mmu_hpte.c
66
index = kvmppc_mmu_hash_pte(pte->pte.eaddr);
arch/powerpc/kvm/book3s_mmu_hpte.c
67
hlist_add_head_rcu(&pte->list_pte, &vcpu3s->hpte_hash_pte[index]);
arch/powerpc/kvm/book3s_mmu_hpte.c
70
index = kvmppc_mmu_hash_pte_long(pte->pte.eaddr);
arch/powerpc/kvm/book3s_mmu_hpte.c
71
hlist_add_head_rcu(&pte->list_pte_long,
arch/powerpc/kvm/book3s_mmu_hpte.c
75
index = kvmppc_mmu_hash_vpte(pte->pte.vpage);
arch/powerpc/kvm/book3s_mmu_hpte.c
76
hlist_add_head_rcu(&pte->list_vpte, &vcpu3s->hpte_hash_vpte[index]);
arch/powerpc/kvm/book3s_mmu_hpte.c
79
index = kvmppc_mmu_hash_vpte_long(pte->pte.vpage);
arch/powerpc/kvm/book3s_mmu_hpte.c
80
hlist_add_head_rcu(&pte->list_vpte_long,
arch/powerpc/kvm/book3s_mmu_hpte.c
85
index = kvmppc_mmu_hash_vpte_64k(pte->pte.vpage);
arch/powerpc/kvm/book3s_mmu_hpte.c
86
hlist_add_head_rcu(&pte->list_vpte_64k,
arch/powerpc/kvm/book3s_mmu_hpte.c
95
static void invalidate_pte(struct kvm_vcpu *vcpu, struct hpte_cache *pte)
arch/powerpc/kvm/book3s_mmu_hpte.c
99
trace_kvm_book3s_mmu_invalidate(pte);
arch/powerpc/kvm/book3s_pr.c
640
static void kvmppc_patch_dcbz(struct kvm_vcpu *vcpu, struct kvmppc_pte *pte)
arch/powerpc/kvm/book3s_pr.c
647
r = kvm_vcpu_map(vcpu, pte->raddr >> PAGE_SHIFT, &map);
arch/powerpc/kvm/book3s_pr.c
651
hpage_offset = pte->raddr & ~PAGE_MASK;
arch/powerpc/kvm/book3s_pr.c
688
struct kvmppc_pte pte = { 0 };
arch/powerpc/kvm/book3s_pr.c
699
page_found = vcpu->arch.mmu.xlate(vcpu, eaddr, &pte, data, iswrite);
arch/powerpc/kvm/book3s_pr.c
701
pte.may_execute = true;
arch/powerpc/kvm/book3s_pr.c
702
pte.may_read = true;
arch/powerpc/kvm/book3s_pr.c
703
pte.may_write = true;
arch/powerpc/kvm/book3s_pr.c
704
pte.raddr = eaddr & KVM_PAM;
arch/powerpc/kvm/book3s_pr.c
705
pte.eaddr = eaddr;
arch/powerpc/kvm/book3s_pr.c
706
pte.vpage = eaddr >> 12;
arch/powerpc/kvm/book3s_pr.c
707
pte.page_size = MMU_PAGE_64K;
arch/powerpc/kvm/book3s_pr.c
708
pte.wimg = HPTE_R_M;
arch/powerpc/kvm/book3s_pr.c
713
pte.vpage |= ((u64)VSID_REAL << (SID_SHIFT - 12));
arch/powerpc/kvm/book3s_pr.c
718
((pte.raddr & SPLIT_HACK_MASK) == SPLIT_HACK_OFFS))
arch/powerpc/kvm/book3s_pr.c
719
pte.raddr &= ~SPLIT_HACK_MASK;
arch/powerpc/kvm/book3s_pr.c
725
pte.vpage |= ((u64)VSID_REAL_DR << (SID_SHIFT - 12));
arch/powerpc/kvm/book3s_pr.c
727
pte.vpage |= ((u64)VSID_REAL_IR << (SID_SHIFT - 12));
arch/powerpc/kvm/book3s_pr.c
728
pte.vpage |= vsid;
arch/powerpc/kvm/book3s_pr.c
742
pte.may_execute = !data;
arch/powerpc/kvm/book3s_pr.c
763
} else if (kvmppc_visible_gpa(vcpu, pte.raddr)) {
arch/powerpc/kvm/book3s_pr.c
770
kvmppc_mmu_unmap_page(vcpu, &pte);
arch/powerpc/kvm/book3s_pr.c
773
if (kvmppc_mmu_map_page(vcpu, &pte, iswrite) == -EIO) {
arch/powerpc/kvm/book3s_pr.c
782
kvmppc_patch_dcbz(vcpu, &pte);
arch/powerpc/kvm/book3s_pr.c
786
vcpu->arch.paddr_accessed = pte.raddr;
arch/powerpc/kvm/book3s_pr.c
787
vcpu->arch.vaddr_accessed = pte.eaddr;
arch/powerpc/kvm/book3s_pr_papr.c
100
pte[0] = be64_to_cpu((__force __be64)pte[0]);
arch/powerpc/kvm/book3s_pr_papr.c
101
pte[1] = be64_to_cpu((__force __be64)pte[1]);
arch/powerpc/kvm/book3s_pr_papr.c
104
if ((pte[0] & HPTE_V_VALID) == 0 ||
arch/powerpc/kvm/book3s_pr_papr.c
105
((flags & H_AVPN) && (pte[0] & ~0x7fUL) != avpn) ||
arch/powerpc/kvm/book3s_pr_papr.c
106
((flags & H_ANDCOND) && (pte[0] & avpn) != 0))
arch/powerpc/kvm/book3s_pr_papr.c
113
rb = compute_tlbie_rb(pte[0], pte[1], pte_index);
arch/powerpc/kvm/book3s_pr_papr.c
117
kvmppc_set_gpr(vcpu, 4, pte[0]);
arch/powerpc/kvm/book3s_pr_papr.c
118
kvmppc_set_gpr(vcpu, 5, pte[1]);
arch/powerpc/kvm/book3s_pr_papr.c
156
unsigned long pte[2];
arch/powerpc/kvm/book3s_pr_papr.c
179
if (copy_from_user(pte, (void __user *)pteg, sizeof(pte))) {
arch/powerpc/kvm/book3s_pr_papr.c
183
pte[0] = be64_to_cpu((__force __be64)pte[0]);
arch/powerpc/kvm/book3s_pr_papr.c
184
pte[1] = be64_to_cpu((__force __be64)pte[1]);
arch/powerpc/kvm/book3s_pr_papr.c
189
if ((pte[0] & HPTE_V_VALID) == 0 ||
arch/powerpc/kvm/book3s_pr_papr.c
190
((flags & H_AVPN) && (pte[0] & ~0x7fUL) != tsl) ||
arch/powerpc/kvm/book3s_pr_papr.c
191
((flags & H_ANDCOND) && (pte[0] & tsl) != 0)) {
arch/powerpc/kvm/book3s_pr_papr.c
200
rb = compute_tlbie_rb(pte[0], pte[1],
arch/powerpc/kvm/book3s_pr_papr.c
204
tsh |= (pte[1] & (HPTE_R_C | HPTE_R_R)) << 43;
arch/powerpc/kvm/book3s_pr_papr.c
220
unsigned long pte[2];
arch/powerpc/kvm/book3s_pr_papr.c
226
if (copy_from_user(pte, (void __user *)pteg, sizeof(pte)))
arch/powerpc/kvm/book3s_pr_papr.c
228
pte[0] = be64_to_cpu((__force __be64)pte[0]);
arch/powerpc/kvm/book3s_pr_papr.c
229
pte[1] = be64_to_cpu((__force __be64)pte[1]);
arch/powerpc/kvm/book3s_pr_papr.c
232
if ((pte[0] & HPTE_V_VALID) == 0 ||
arch/powerpc/kvm/book3s_pr_papr.c
233
((flags & H_AVPN) && (pte[0] & ~0x7fUL) != avpn))
arch/powerpc/kvm/book3s_pr_papr.c
236
v = pte[0];
arch/powerpc/kvm/book3s_pr_papr.c
237
r = pte[1];
arch/powerpc/kvm/book3s_pr_papr.c
244
pte[1] = r;
arch/powerpc/kvm/book3s_pr_papr.c
248
pte[0] = (__force u64)cpu_to_be64(pte[0]);
arch/powerpc/kvm/book3s_pr_papr.c
249
pte[1] = (__force u64)cpu_to_be64(pte[1]);
arch/powerpc/kvm/book3s_pr_papr.c
251
if (copy_to_user((void __user *)pteg, pte, sizeof(pte)))
arch/powerpc/kvm/book3s_pr_papr.c
92
unsigned long pte[2];
arch/powerpc/kvm/book3s_pr_papr.c
98
if (copy_from_user(pte, (void __user *)pteg, sizeof(pte)))
arch/powerpc/kvm/booke.c
1981
enum xlate_readwrite xlrw, struct kvmppc_pte *pte)
arch/powerpc/kvm/booke.c
1989
pte->eaddr = eaddr;
arch/powerpc/kvm/booke.c
1990
pte->raddr = (vcpu->arch.magic_page_pa & PAGE_MASK) |
arch/powerpc/kvm/booke.c
1992
pte->vpage = eaddr >> PAGE_SHIFT;
arch/powerpc/kvm/booke.c
1993
pte->may_read = true;
arch/powerpc/kvm/booke.c
1994
pte->may_write = true;
arch/powerpc/kvm/booke.c
1995
pte->may_execute = true;
arch/powerpc/kvm/booke.c
2019
pte->eaddr = eaddr;
arch/powerpc/kvm/booke.c
2020
pte->raddr = (gpaddr & PAGE_MASK) | (eaddr & ~PAGE_MASK);
arch/powerpc/kvm/booke.c
2021
pte->vpage = eaddr >> PAGE_SHIFT;
arch/powerpc/kvm/booke.c
2024
pte->may_read = true;
arch/powerpc/kvm/booke.c
2025
pte->may_write = true;
arch/powerpc/kvm/booke.c
2026
pte->may_execute = true;
arch/powerpc/kvm/e500_mmu_host.c
380
pte_t pte = READ_ONCE(*ptep);
arch/powerpc/kvm/e500_mmu_host.c
382
if (pte_present(pte)) {
arch/powerpc/kvm/e500_mmu_host.c
383
wimg = (pte_val(pte) >> PTE_WIMGE_SHIFT) &
arch/powerpc/kvm/powerpc.c
352
struct kvmppc_pte pte;
arch/powerpc/kvm/powerpc.c
365
XLATE_WRITE, &pte);
arch/powerpc/kvm/powerpc.c
369
*eaddr = pte.raddr;
arch/powerpc/kvm/powerpc.c
371
if (!pte.may_write)
arch/powerpc/kvm/powerpc.c
376
((pte.raddr & KVM_PAM & PAGE_MASK) == mp_pa) &&
arch/powerpc/kvm/powerpc.c
379
magic += pte.eaddr & 0xfff;
arch/powerpc/kvm/powerpc.c
384
if (kvm_write_guest(vcpu->kvm, pte.raddr, ptr, size))
arch/powerpc/kvm/powerpc.c
395
struct kvmppc_pte pte;
arch/powerpc/kvm/powerpc.c
408
XLATE_READ, &pte);
arch/powerpc/kvm/powerpc.c
412
*eaddr = pte.raddr;
arch/powerpc/kvm/powerpc.c
414
if (!pte.may_read)
arch/powerpc/kvm/powerpc.c
417
if (!data && !pte.may_execute)
arch/powerpc/kvm/powerpc.c
422
((pte.raddr & KVM_PAM & PAGE_MASK) == mp_pa) &&
arch/powerpc/kvm/powerpc.c
425
magic += pte.eaddr & 0xfff;
arch/powerpc/kvm/powerpc.c
431
rc = kvm_read_guest(vcpu->kvm, pte.raddr, ptr, size);
arch/powerpc/kvm/trace_pr.h
106
__entry->host_vpn = pte->host_vpn;
arch/powerpc/kvm/trace_pr.h
107
__entry->pfn = pte->pfn;
arch/powerpc/kvm/trace_pr.h
108
__entry->eaddr = pte->pte.eaddr;
arch/powerpc/kvm/trace_pr.h
109
__entry->vpage = pte->pte.vpage;
arch/powerpc/kvm/trace_pr.h
110
__entry->raddr = pte->pte.raddr;
arch/powerpc/kvm/trace_pr.h
111
__entry->flags = (pte->pte.may_read ? 0x4 : 0) |
arch/powerpc/kvm/trace_pr.h
112
(pte->pte.may_write ? 0x2 : 0) |
arch/powerpc/kvm/trace_pr.h
113
(pte->pte.may_execute ? 0x1 : 0);
arch/powerpc/kvm/trace_pr.h
64
TP_PROTO(struct hpte_cache *pte),
arch/powerpc/kvm/trace_pr.h
65
TP_ARGS(pte),
arch/powerpc/kvm/trace_pr.h
77
__entry->host_vpn = pte->host_vpn;
arch/powerpc/kvm/trace_pr.h
78
__entry->pfn = pte->pfn;
arch/powerpc/kvm/trace_pr.h
79
__entry->eaddr = pte->pte.eaddr;
arch/powerpc/kvm/trace_pr.h
80
__entry->vpage = pte->pte.vpage;
arch/powerpc/kvm/trace_pr.h
81
__entry->raddr = pte->pte.raddr;
arch/powerpc/kvm/trace_pr.h
82
__entry->flags = (pte->pte.may_read ? 0x4 : 0) |
arch/powerpc/kvm/trace_pr.h
83
(pte->pte.may_write ? 0x2 : 0) |
arch/powerpc/kvm/trace_pr.h
84
(pte->pte.may_execute ? 0x1 : 0);
arch/powerpc/kvm/trace_pr.h
93
TP_PROTO(struct hpte_cache *pte),
arch/powerpc/kvm/trace_pr.h
94
TP_ARGS(pte),
arch/powerpc/lib/code-patching.c
128
this_cpu_write(cpu_patching_context.pte, virt_to_kpte(addr));
arch/powerpc/lib/code-patching.c
138
this_cpu_write(cpu_patching_context.pte, NULL);
arch/powerpc/lib/code-patching.c
155
pte_t *pte;
arch/powerpc/lib/code-patching.c
177
pte = get_locked_pte(mm, addr, &ptl);
arch/powerpc/lib/code-patching.c
178
if (!pte)
arch/powerpc/lib/code-patching.c
180
pte_unmap_unlock(pte, ptl);
arch/powerpc/lib/code-patching.c
286
pte_t *pte;
arch/powerpc/lib/code-patching.c
296
pte = get_locked_pte(patching_mm, text_poke_addr, &ptl);
arch/powerpc/lib/code-patching.c
297
if (!pte)
arch/powerpc/lib/code-patching.c
300
__set_pte_at(patching_mm, text_poke_addr, pte, pfn_pte(pfn, PAGE_KERNEL), 0);
arch/powerpc/lib/code-patching.c
315
pte_clear(patching_mm, text_poke_addr, pte);
arch/powerpc/lib/code-patching.c
322
pte_unmap_unlock(pte, ptl);
arch/powerpc/lib/code-patching.c
332
pte_t *pte;
arch/powerpc/lib/code-patching.c
338
pte = __this_cpu_read(cpu_patching_context.pte);
arch/powerpc/lib/code-patching.c
339
__set_pte_at(&init_mm, text_poke_addr, pte, pfn_pte(pfn, PAGE_KERNEL), 0);
arch/powerpc/lib/code-patching.c
346
pte_clear(&init_mm, text_poke_addr, pte);
arch/powerpc/lib/code-patching.c
475
pte_t *pte;
arch/powerpc/lib/code-patching.c
482
pte = get_locked_pte(patching_mm, text_poke_addr, &ptl);
arch/powerpc/lib/code-patching.c
483
if (!pte)
arch/powerpc/lib/code-patching.c
486
__set_pte_at(patching_mm, text_poke_addr, pte, pfn_pte(pfn, PAGE_KERNEL), 0);
arch/powerpc/lib/code-patching.c
503
pte_clear(patching_mm, text_poke_addr, pte);
arch/powerpc/lib/code-patching.c
510
pte_unmap_unlock(pte, ptl);
arch/powerpc/lib/code-patching.c
524
pte_t *pte;
arch/powerpc/lib/code-patching.c
530
pte = __this_cpu_read(cpu_patching_context.pte);
arch/powerpc/lib/code-patching.c
531
__set_pte_at(&init_mm, text_poke_addr, pte, pfn_pte(pfn, PAGE_KERNEL), 0);
arch/powerpc/lib/code-patching.c
538
pte_clear(&init_mm, text_poke_addr, pte);
arch/powerpc/lib/code-patching.c
58
pte_t *pte;
arch/powerpc/mm/book3s64/hash_4k.c
36
pte_t pte = READ_ONCE(*ptep);
arch/powerpc/mm/book3s64/hash_4k.c
38
old_pte = pte_val(pte);
arch/powerpc/mm/book3s64/hash_64k.c
243
pte_t pte = READ_ONCE(*ptep);
arch/powerpc/mm/book3s64/hash_64k.c
245
old_pte = pte_val(pte);
arch/powerpc/mm/book3s64/hash_64k.c
257
unlikely(pte_ci(pte)))
arch/powerpc/mm/book3s64/hash_64k.c
54
pte_t pte = READ_ONCE(*ptep);
arch/powerpc/mm/book3s64/hash_64k.c
56
old_pte = pte_val(pte);
arch/powerpc/mm/book3s64/hash_native.c
783
real_pte_t pte;
arch/powerpc/mm/book3s64/hash_native.c
793
pte = batch->pte[i];
arch/powerpc/mm/book3s64/hash_native.c
795
pte_iterate_hashed_subpages(pte, psize, vpn, index, shift) {
arch/powerpc/mm/book3s64/hash_native.c
797
hidx = __rpte_to_hidx(pte, index);
arch/powerpc/mm/book3s64/hash_native.c
827
pte = batch->pte[i];
arch/powerpc/mm/book3s64/hash_native.c
829
pte_iterate_hashed_subpages(pte, psize,
arch/powerpc/mm/book3s64/hash_native.c
844
pte = batch->pte[i];
arch/powerpc/mm/book3s64/hash_native.c
846
pte_iterate_hashed_subpages(pte, psize,
arch/powerpc/mm/book3s64/hash_tlb.c
130
batch->pte[i] = rpte;
arch/powerpc/mm/book3s64/hash_tlb.c
152
flush_hash_page(batch->vpn[0], batch->pte[0],
arch/powerpc/mm/book3s64/hash_tlb.c
213
unsigned long pte;
arch/powerpc/mm/book3s64/hash_tlb.c
217
pte = pte_val(*ptep);
arch/powerpc/mm/book3s64/hash_tlb.c
218
if (!(pte & H_PAGE_HASHPTE))
arch/powerpc/mm/book3s64/hash_tlb.c
220
hpte_need_flush(&init_mm, start, ptep, pte, hugepage_shift);
arch/powerpc/mm/book3s64/hash_tlb.c
228
pte_t *pte;
arch/powerpc/mm/book3s64/hash_tlb.c
246
for (pte = start_pte; pte < start_pte + PTRS_PER_PTE; pte++) {
arch/powerpc/mm/book3s64/hash_tlb.c
247
unsigned long pteval = pte_val(*pte);
arch/powerpc/mm/book3s64/hash_tlb.c
249
hpte_need_flush(mm, addr, pte, pteval, 0);
arch/powerpc/mm/book3s64/hash_tlb.c
42
pte_t *ptep, unsigned long pte, int huge)
arch/powerpc/mm/book3s64/hash_tlb.c
73
psize = pte_pagesize_index(mm, addr, pte); /* shutup gcc */
arch/powerpc/mm/book3s64/hash_tlb.c
76
psize = pte_pagesize_index(mm, addr, pte);
arch/powerpc/mm/book3s64/hash_tlb.c
98
rpte = __real_pte(__pte(pte), ptep, offset);
arch/powerpc/mm/book3s64/hash_utils.c
1574
unsigned int hash_page_do_lazy_icache(unsigned int pp, pte_t pte, int trap)
arch/powerpc/mm/book3s64/hash_utils.c
1578
if (!pfn_valid(pte_pfn(pte)))
arch/powerpc/mm/book3s64/hash_utils.c
1581
folio = page_folio(pte_page(pte));
arch/powerpc/mm/book3s64/hash_utils.c
1687
int ssize, int psize, int lpsize, unsigned long pte)
arch/powerpc/mm/book3s64/hash_utils.c
1694
trap, vsid, ssize, psize, lpsize, pte);
arch/powerpc/mm/book3s64/hash_utils.c
2200
void flush_hash_page(unsigned long vpn, real_pte_t pte, int psize, int ssize,
arch/powerpc/mm/book3s64/hash_utils.c
2207
pte_iterate_hashed_subpages(pte, psize, vpn, index, shift) {
arch/powerpc/mm/book3s64/hash_utils.c
2208
gslot = pte_get_hash_gslot(vpn, shift, ssize, pte, index);
arch/powerpc/mm/book3s64/hash_utils.c
2289
flush_hash_page(batch->vpn[i], batch->pte[i],
arch/powerpc/mm/book3s64/hugetlbpage.c
154
pte_t *ptep, pte_t old_pte, pte_t pte)
arch/powerpc/mm/book3s64/hugetlbpage.c
160
old_pte, pte);
arch/powerpc/mm/book3s64/hugetlbpage.c
163
set_huge_pte_at(vma->vm_mm, addr, ptep, pte, psize);
arch/powerpc/mm/book3s64/pgtable.c
555
pte_t *ptep, pte_t old_pte, pte_t pte)
arch/powerpc/mm/book3s64/pgtable.c
559
ptep, old_pte, pte);
arch/powerpc/mm/book3s64/pgtable.c
560
set_pte_at_unchecked(vma->vm_mm, addr, ptep, pte);
arch/powerpc/mm/book3s64/pkeys.c
431
bool arch_pte_access_permitted(u64 pte, bool write, bool execute)
arch/powerpc/mm/book3s64/pkeys.c
436
return pkey_access_permitted(pte_to_pkey_bits(pte), write, execute);
arch/powerpc/mm/book3s64/radix_hugetlbpage.c
47
pte_t old_pte, pte_t pte)
arch/powerpc/mm/book3s64/radix_hugetlbpage.c
58
is_pte_rw_upgrade(pte_val(old_pte), pte_val(pte)) &&
arch/powerpc/mm/book3s64/radix_hugetlbpage.c
62
set_huge_pte_at(vma->vm_mm, addr, ptep, pte, psize);
arch/powerpc/mm/book3s64/radix_pgtable.c
1020
pte_t *pte = pte_offset_kernel(pmdp, addr);
arch/powerpc/mm/book3s64/radix_pgtable.c
1022
if (pte_none(*pte)) {
arch/powerpc/mm/book3s64/radix_pgtable.c
1057
set_pte_at(&init_mm, addr, pte, entry);
arch/powerpc/mm/book3s64/radix_pgtable.c
1060
return pte;
arch/powerpc/mm/book3s64/radix_pgtable.c
1099
pte_t *pte;
arch/powerpc/mm/book3s64/radix_pgtable.c
1104
pte = early_alloc_pgtable(PAGE_SIZE, node, 0, 0);
arch/powerpc/mm/book3s64/radix_pgtable.c
1105
pmd_populate(&init_mm, pmdp, pte);
arch/powerpc/mm/book3s64/radix_pgtable.c
1123
pte_t *pte;
arch/powerpc/mm/book3s64/radix_pgtable.c
1206
pte = vmemmap_pte_alloc(pmd, node, addr);
arch/powerpc/mm/book3s64/radix_pgtable.c
1207
if (!pte)
arch/powerpc/mm/book3s64/radix_pgtable.c
1210
pte = radix__vmemmap_pte_populate(pmd, addr, node, altmap, NULL);
arch/powerpc/mm/book3s64/radix_pgtable.c
1211
if (!pte)
arch/powerpc/mm/book3s64/radix_pgtable.c
1214
vmemmap_verify(pte, node, addr, addr + PAGE_SIZE);
arch/powerpc/mm/book3s64/radix_pgtable.c
1228
pte_t *pte;
arch/powerpc/mm/book3s64/radix_pgtable.c
1244
pte = vmemmap_pte_alloc(pmd, node, addr);
arch/powerpc/mm/book3s64/radix_pgtable.c
1245
if (!pte)
arch/powerpc/mm/book3s64/radix_pgtable.c
1248
vmemmap_verify(pte, node, addr, addr + PAGE_SIZE);
arch/powerpc/mm/book3s64/radix_pgtable.c
1250
return pte;
arch/powerpc/mm/book3s64/radix_pgtable.c
1260
pte_t *pte;
arch/powerpc/mm/book3s64/radix_pgtable.c
1279
pte = vmemmap_pte_alloc(pmd, node, map_addr);
arch/powerpc/mm/book3s64/radix_pgtable.c
1280
if (!pte)
arch/powerpc/mm/book3s64/radix_pgtable.c
1285
if (pte_none(*pte)) {
arch/powerpc/mm/book3s64/radix_pgtable.c
1291
pte = radix__vmemmap_populate_address(map_addr - PAGE_SIZE, node, NULL, NULL);
arch/powerpc/mm/book3s64/radix_pgtable.c
1292
if (!pte)
arch/powerpc/mm/book3s64/radix_pgtable.c
1297
pte = radix__vmemmap_pte_populate(pmd, map_addr, node, NULL, NULL);
arch/powerpc/mm/book3s64/radix_pgtable.c
1298
if (!pte)
arch/powerpc/mm/book3s64/radix_pgtable.c
1300
vmemmap_verify(pte, node, map_addr, map_addr + PAGE_SIZE);
arch/powerpc/mm/book3s64/radix_pgtable.c
1301
return pte;
arch/powerpc/mm/book3s64/radix_pgtable.c
1303
return pte;
arch/powerpc/mm/book3s64/radix_pgtable.c
1323
pte_t *pte;
arch/powerpc/mm/book3s64/radix_pgtable.c
1342
pte = vmemmap_pte_alloc(pmd, node, addr);
arch/powerpc/mm/book3s64/radix_pgtable.c
1343
if (!pte)
arch/powerpc/mm/book3s64/radix_pgtable.c
1345
if (!pte_none(*pte)) {
arch/powerpc/mm/book3s64/radix_pgtable.c
1365
pte = radix__vmemmap_pte_populate(pmd, addr, node, NULL, NULL);
arch/powerpc/mm/book3s64/radix_pgtable.c
1366
if (!pte)
arch/powerpc/mm/book3s64/radix_pgtable.c
1368
vmemmap_verify(pte, node, addr, addr + PAGE_SIZE);
arch/powerpc/mm/book3s64/radix_pgtable.c
1375
pte = radix__vmemmap_populate_address(addr + PAGE_SIZE, node, NULL, NULL);
arch/powerpc/mm/book3s64/radix_pgtable.c
1376
if (!pte)
arch/powerpc/mm/book3s64/radix_pgtable.c
1390
pte = radix__vmemmap_pte_populate(pmd, addr, node, NULL, NULL);
arch/powerpc/mm/book3s64/radix_pgtable.c
1391
if (!pte)
arch/powerpc/mm/book3s64/radix_pgtable.c
1393
vmemmap_verify(pte, node, addr, addr + PAGE_SIZE);
arch/powerpc/mm/book3s64/radix_pgtable.c
1400
pte = radix__vmemmap_pte_populate(pmd, addr, node, NULL, pte_page(*tail_page_pte));
arch/powerpc/mm/book3s64/radix_pgtable.c
1401
if (!pte)
arch/powerpc/mm/book3s64/radix_pgtable.c
1403
vmemmap_verify(pte, node, addr, addr + PAGE_SIZE);
arch/powerpc/mm/book3s64/radix_pgtable.c
1598
pte_t old_pte, pte_t pte)
arch/powerpc/mm/book3s64/radix_pgtable.c
1608
is_pte_rw_upgrade(pte_val(old_pte), pte_val(pte)) &&
arch/powerpc/mm/book3s64/radix_pgtable.c
1612
set_pte_at_unchecked(mm, addr, ptep, pte);
arch/powerpc/mm/book3s64/radix_pgtable.c
1650
pte_t *pte;
arch/powerpc/mm/book3s64/radix_pgtable.c
1651
pte = (pte_t *)pmd_page_vaddr(pmd[i]);
arch/powerpc/mm/book3s64/radix_pgtable.c
1653
pte_free_kernel(&init_mm, pte);
arch/powerpc/mm/book3s64/radix_pgtable.c
1687
pte_t *pte;
arch/powerpc/mm/book3s64/radix_pgtable.c
1689
pte = (pte_t *)pmd_page_vaddr(*pmd);
arch/powerpc/mm/book3s64/radix_pgtable.c
1694
pte_free_kernel(&init_mm, pte);
arch/powerpc/mm/book3s64/radix_pgtable.c
695
pte_t *pte;
arch/powerpc/mm/book3s64/radix_pgtable.c
699
pte = pte_start + i;
arch/powerpc/mm/book3s64/radix_pgtable.c
700
if (!pte_none(*pte))
arch/powerpc/mm/book3s64/radix_pgtable.c
792
pte_t *pte;
arch/powerpc/mm/book3s64/radix_pgtable.c
794
pte = pte_start + pte_index(addr);
arch/powerpc/mm/book3s64/radix_pgtable.c
795
for (; addr < end; addr = next, pte++) {
arch/powerpc/mm/book3s64/radix_pgtable.c
800
if (!pte_present(*pte))
arch/powerpc/mm/book3s64/radix_pgtable.c
805
free_vmemmap_pages(pte_page(*pte), altmap, 0);
arch/powerpc/mm/book3s64/radix_pgtable.c
806
pte_clear(&init_mm, addr, pte);
arch/powerpc/mm/book3s64/radix_pgtable.c
811
free_vmemmap_pages(pte_page(*pte), altmap, 0);
arch/powerpc/mm/book3s64/radix_pgtable.c
812
pte_clear(&init_mm, addr, pte);
arch/powerpc/mm/book3s64/subpage_prot.c
60
pte_t *pte;
arch/powerpc/mm/book3s64/subpage_prot.c
73
pte = pte_offset_map_lock(mm, pmd, addr, &ptl);
arch/powerpc/mm/book3s64/subpage_prot.c
74
if (!pte)
arch/powerpc/mm/book3s64/subpage_prot.c
78
pte_update(mm, addr, pte, 0, 0, 0);
arch/powerpc/mm/book3s64/subpage_prot.c
80
++pte;
arch/powerpc/mm/book3s64/subpage_prot.c
83
pte_unmap_unlock(pte - 1, ptl);
arch/powerpc/mm/ioremap.c
39
pte_t pte = __pte(pgprot_val(prot));
arch/powerpc/mm/ioremap.c
43
if (pte_write(pte))
arch/powerpc/mm/ioremap.c
44
pte = pte_mkdirty(pte);
arch/powerpc/mm/ioremap.c
46
return __ioremap_caller(addr, size, pte_pgprot(pte), caller);
arch/powerpc/mm/kasan/8xx.c
30
pte_t pte = pte_mkhuge(pfn_pte(PHYS_PFN(__pa(block + i * PAGE_SIZE)), PAGE_KERNEL));
arch/powerpc/mm/kasan/8xx.c
32
__set_pte_at(&init_mm, k_cur, ptep + i, pte, 1);
arch/powerpc/mm/kasan/8xx.c
68
pte_t pte = pfn_pte(PHYS_PFN(__pa(va)), PAGE_KERNEL);
arch/powerpc/mm/kasan/8xx.c
71
pte = pte_mkhuge(pte);
arch/powerpc/mm/kasan/8xx.c
73
__set_pte_at(&init_mm, k_cur, pte_offset_kernel(pmd, k_cur), pte, 0);
arch/powerpc/mm/kasan/book3s_32.c
52
pte_t pte = pfn_pte(PHYS_PFN(phys + k_cur - k_nobat), PAGE_KERNEL);
arch/powerpc/mm/kasan/book3s_32.c
54
__set_pte_at(&init_mm, k_cur, pte_offset_kernel(pmd, k_cur), pte, 0);
arch/powerpc/mm/kasan/init_32.c
75
pte_t pte = pfn_pte(PHYS_PFN(__pa(va)), PAGE_KERNEL);
arch/powerpc/mm/kasan/init_32.c
77
__set_pte_at(&init_mm, k_cur, pte_offset_kernel(pmd, k_cur), pte, 0);
arch/powerpc/mm/kasan/init_32.c
84
kasan_update_early_region(unsigned long k_start, unsigned long k_end, pte_t pte)
arch/powerpc/mm/kasan/init_32.c
95
__set_pte_at(&init_mm, k_cur, ptep, pte, 0);
arch/powerpc/mm/nohash/e500_hugetlbpage.c
119
book3e_hugetlb_preload(struct vm_area_struct *vma, unsigned long ea, pte_t pte)
arch/powerpc/mm/nohash/e500_hugetlbpage.c
156
mas2 |= (pte_val(pte) >> PTE_WIMGE_SHIFT) & MAS2_WIMGE_MASK;
arch/powerpc/mm/nohash/e500_hugetlbpage.c
157
mas7_3 = (u64)pte_pfn(pte) << PAGE_SHIFT;
arch/powerpc/mm/nohash/e500_hugetlbpage.c
158
mas7_3 |= (pte_val(pte) >> PTE_BAP_SHIFT) & MAS3_BAP_MASK;
arch/powerpc/mm/nohash/e500_hugetlbpage.c
159
if (!pte_dirty(pte))
arch/powerpc/mm/pgtable-frag.c
100
pte = get_pte_from_cache(mm);
arch/powerpc/mm/pgtable-frag.c
101
if (pte)
arch/powerpc/mm/pgtable-frag.c
102
return pte;
arch/powerpc/mm/pgtable-frag.c
98
pte_t *pte;
arch/powerpc/mm/pgtable.c
101
static pte_t set_pte_filter_hash(pte_t pte, unsigned long addr) { return pte; }
arch/powerpc/mm/pgtable.c
111
static inline pte_t set_pte_filter(pte_t pte, unsigned long addr)
arch/powerpc/mm/pgtable.c
116
return pte;
arch/powerpc/mm/pgtable.c
119
return set_pte_filter_hash(pte, addr);
arch/powerpc/mm/pgtable.c
122
if (!pte_exec(pte) || !pte_looks_normal(pte, addr))
arch/powerpc/mm/pgtable.c
123
return pte;
arch/powerpc/mm/pgtable.c
126
folio = maybe_pte_to_folio(pte);
arch/powerpc/mm/pgtable.c
128
return pte;
arch/powerpc/mm/pgtable.c
132
return pte;
arch/powerpc/mm/pgtable.c
138
return pte;
arch/powerpc/mm/pgtable.c
142
return pte_exprotect(pte);
arch/powerpc/mm/pgtable.c
145
static pte_t set_access_flags_filter(pte_t pte, struct vm_area_struct *vma,
arch/powerpc/mm/pgtable.c
151
return pte;
arch/powerpc/mm/pgtable.c
154
return pte;
arch/powerpc/mm/pgtable.c
161
if (dirty || pte_exec(pte) || !is_exec_fault())
arch/powerpc/mm/pgtable.c
162
return pte;
arch/powerpc/mm/pgtable.c
170
return pte;
arch/powerpc/mm/pgtable.c
174
folio = maybe_pte_to_folio(pte);
arch/powerpc/mm/pgtable.c
187
return pte_mkexec(pte);
arch/powerpc/mm/pgtable.c
194
pte_t pte, unsigned int nr)
arch/powerpc/mm/pgtable.c
202
pte = set_pte_filter(pte, addr);
arch/powerpc/mm/pgtable.c
211
page_table_check_ptes_set(mm, addr, ptep, pte, nr);
arch/powerpc/mm/pgtable.c
222
__set_pte_at(mm, addr, ptep, pte, 0);
arch/powerpc/mm/pgtable.c
227
pte = pte_next_pfn(pte);
arch/powerpc/mm/pgtable.c
232
pte_t *ptep, pte_t pte)
arch/powerpc/mm/pgtable.c
235
pte = set_pte_filter(pte, addr);
arch/powerpc/mm/pgtable.c
236
__set_pte_at(mm, addr, ptep, pte, 0);
arch/powerpc/mm/pgtable.c
272
pte_t pte, int dirty)
arch/powerpc/mm/pgtable.c
280
ptep_set_access_flags(vma, addr, ptep, pte, dirty);
arch/powerpc/mm/pgtable.c
285
pte = set_access_flags_filter(pte, vma, dirty);
arch/powerpc/mm/pgtable.c
286
changed = !pte_same(*(ptep), pte);
arch/powerpc/mm/pgtable.c
305
__ptep_set_access_flags(vma, ptep, pte, addr, psize);
arch/powerpc/mm/pgtable.c
336
pte_t pte, unsigned long sz)
arch/powerpc/mm/pgtable.c
340
pte = set_pte_filter(pte, addr);
arch/powerpc/mm/pgtable.c
346
__set_huge_pte_at(pmdp, pte_offset_kernel(pmdp, 0), pte_val(pte));
arch/powerpc/mm/pgtable.c
347
__set_huge_pte_at(pmdp, pte_offset_kernel(pmdp + 1, 0), pte_val(pte) + SZ_4M);
arch/powerpc/mm/pgtable.c
349
__set_huge_pte_at(pmdp, ptep, pte_val(pte));
arch/powerpc/mm/pgtable.c
354
pte_t pte, unsigned long sz)
arch/powerpc/mm/pgtable.c
359
pte = set_pte_filter(pte, addr);
arch/powerpc/mm/pgtable.c
379
__set_pte_at(mm, addr, ptep, pte, 0);
arch/powerpc/mm/pgtable.c
380
pte = __pte(pte_val(pte) + ((unsigned long long)pdsize / PAGE_SIZE << PFN_PTE_SHIFT));
arch/powerpc/mm/pgtable.c
393
pte_t *pte;
arch/powerpc/mm/pgtable.c
413
pte = pte_offset_map_ro_nolock(mm, pmd, addr, &ptl);
arch/powerpc/mm/pgtable.c
414
BUG_ON(!pte);
arch/powerpc/mm/pgtable.c
416
pte_unmap(pte);
arch/powerpc/mm/pgtable.c
50
static inline int pte_looks_normal(pte_t pte, unsigned long addr)
arch/powerpc/mm/pgtable.c
53
if (pte_present(pte) && !pte_special(pte)) {
arch/powerpc/mm/pgtable.c
54
if (pte_ci(pte))
arch/powerpc/mm/pgtable.c
62
static struct folio *maybe_pte_to_folio(pte_t pte)
arch/powerpc/mm/pgtable.c
64
unsigned long pfn = pte_pfn(pte);
arch/powerpc/mm/pgtable.c
83
static pte_t set_pte_filter_hash(pte_t pte, unsigned long addr)
arch/powerpc/mm/pgtable.c
85
pte = __pte(pte_val(pte) & ~_PAGE_HPTEFLAGS);
arch/powerpc/mm/pgtable.c
86
if (pte_looks_normal(pte, addr) && !(cpu_has_feature(CPU_FTR_COHERENT_ICACHE) ||
arch/powerpc/mm/pgtable.c
88
struct folio *folio = maybe_pte_to_folio(pte);
arch/powerpc/mm/pgtable.c
90
return pte;
arch/powerpc/mm/pgtable.c
96
return pte;
arch/powerpc/mm/ptdump/hashpagetable.c
155
*flag, u64 pte, int num)
arch/powerpc/mm/ptdump/hashpagetable.c
168
val = pte & flag->val;
arch/powerpc/mm/ptdump/hashpagetable.c
173
if ((pte & flag->mask) == flag->val)
arch/powerpc/mm/ptdump/hashpagetable.c
382
pte_t *pte = pte_offset_kernel(pmd, 0);
arch/powerpc/mm/ptdump/hashpagetable.c
386
for (i = 0; i < PTRS_PER_PTE; i++, pte++) {
arch/powerpc/mm/ptdump/hashpagetable.c
388
pteval = pte_val(*pte);
arch/powerpc/mm/ptdump/ptdump.c
139
*flag, u64 pte, int num)
arch/powerpc/mm/ptdump/ptdump.c
152
val = pte & flag->val;
arch/powerpc/mm/ptdump/ptdump.c
157
if ((pte & flag->mask) == flag->val)
arch/powerpc/mm/ptdump/ptdump.c
186
pte_t pte = __pte(st->current_flags);
arch/powerpc/mm/ptdump/ptdump.c
191
if (!pte_write(pte) || !pte_exec(pte))
arch/powerpc/mm/ptdump/ptdump.c
302
static void note_page_pte(struct ptdump_state *pt_st, unsigned long addr, pte_t pte)
arch/powerpc/mm/ptdump/ptdump.c
304
note_page(pt_st, addr, 4, pte_val(pte));
arch/powerpc/platforms/pseries/lpar.c
1329
static inline unsigned long compute_slot(real_pte_t pte,
arch/powerpc/platforms/pseries/lpar.c
1338
hidx = __rpte_to_hidx(pte, index);
arch/powerpc/platforms/pseries/lpar.c
1356
real_pte_t pte;
arch/powerpc/platforms/pseries/lpar.c
1364
pte = batch->pte[i];
arch/powerpc/platforms/pseries/lpar.c
1365
pte_iterate_hashed_subpages(pte, psize, vpn, index, shift) {
arch/powerpc/platforms/pseries/lpar.c
1385
slot = compute_slot(pte, vpn, index, shift, ssize);
arch/powerpc/platforms/pseries/lpar.c
1535
real_pte_t pte;
arch/powerpc/platforms/pseries/lpar.c
1551
pte = batch->pte[i];
arch/powerpc/platforms/pseries/lpar.c
1552
pte_iterate_hashed_subpages(pte, psize, vpn, index, shift) {
arch/powerpc/platforms/pseries/lpar.c
1553
slot = compute_slot(pte, vpn, index, shift, ssize);
arch/powerpc/xmon/xmon.c
3269
static void format_pte(void *ptep, unsigned long pte)
arch/powerpc/xmon/xmon.c
3271
pte_t entry = __pte(pte);
arch/powerpc/xmon/xmon.c
3273
printf("ptep @ 0x%016lx = 0x%016lx\n", (unsigned long)ptep, pte);
arch/powerpc/xmon/xmon.c
3274
printf("Maps physical address = 0x%016lx\n", pte & PTE_RPN_MASK);
arch/riscv/include/asm/hugetlb.h
26
unsigned long addr, pte_t *ptep, pte_t pte,
arch/riscv/include/asm/hugetlb.h
45
pte_t pte, int dirty);
arch/riscv/include/asm/kfence.h
18
pte_t *pte = virt_to_kpte(addr);
arch/riscv/include/asm/kfence.h
21
set_pte(pte, __pte(pte_val(ptep_get(pte)) & ~_PAGE_PRESENT));
arch/riscv/include/asm/kfence.h
23
set_pte(pte, __pte(pte_val(ptep_get(pte)) | _PAGE_PRESENT));
arch/riscv/include/asm/kvm_gstage.h
22
pte_t pte;
arch/riscv/include/asm/page.h
66
unsigned long pte;
arch/riscv/include/asm/page.h
75
#define pte_val(x) ((x).pte)
arch/riscv/include/asm/pgalloc.h
133
static inline void __pte_free_tlb(struct mmu_gather *tlb, pgtable_t pte,
arch/riscv/include/asm/pgalloc.h
136
tlb_remove_ptdesc(tlb, page_ptdesc(pte));
arch/riscv/include/asm/pgalloc.h
19
pmd_t *pmd, pte_t *pte)
arch/riscv/include/asm/pgalloc.h
21
unsigned long pfn = virt_to_pfn(pte);
arch/riscv/include/asm/pgalloc.h
27
pmd_t *pmd, pgtable_t pte)
arch/riscv/include/asm/pgalloc.h
29
unsigned long pfn = virt_to_pfn(page_address(pte));
arch/riscv/include/asm/pgtable-64.h
102
#define napot_cont_order(val) (__builtin_ctzl((val.pte >> _PAGE_PFN_SHIFT) << 1))
arch/riscv/include/asm/pgtable.h
1198
#define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) })
arch/riscv/include/asm/pgtable.h
1201
static inline bool pte_swp_exclusive(pte_t pte)
arch/riscv/include/asm/pgtable.h
1203
return pte_val(pte) & _PAGE_SWP_EXCLUSIVE;
arch/riscv/include/asm/pgtable.h
1206
static inline pte_t pte_swp_mkexclusive(pte_t pte)
arch/riscv/include/asm/pgtable.h
1208
return __pte(pte_val(pte) | _PAGE_SWP_EXCLUSIVE);
arch/riscv/include/asm/pgtable.h
1211
static inline pte_t pte_swp_clear_exclusive(pte_t pte)
arch/riscv/include/asm/pgtable.h
1213
return __pte(pte_val(pte) & ~_PAGE_SWP_EXCLUSIVE);
arch/riscv/include/asm/pgtable.h
296
static inline unsigned long pte_napot(pte_t pte)
arch/riscv/include/asm/pgtable.h
298
return pte_val(pte) & _PAGE_NAPOT;
arch/riscv/include/asm/pgtable.h
301
static inline pte_t pte_mknapot(pte_t pte, unsigned int order)
arch/riscv/include/asm/pgtable.h
307
return __pte((pte_val(pte) & napot_mask) | napot_bit | _PAGE_NAPOT);
arch/riscv/include/asm/pgtable.h
314
static inline unsigned long pte_napot(pte_t pte)
arch/riscv/include/asm/pgtable.h
322
static inline unsigned long pte_pfn(pte_t pte)
arch/riscv/include/asm/pgtable.h
324
unsigned long res = __page_val_to_pfn(pte_val(pte));
arch/riscv/include/asm/pgtable.h
326
if (has_svnapot() && pte_napot(pte))
arch/riscv/include/asm/pgtable.h
345
static inline pgprot_t pte_pgprot(pte_t pte)
arch/riscv/include/asm/pgtable.h
347
unsigned long pfn = pte_pfn(pte);
arch/riscv/include/asm/pgtable.h
349
return __pgprot(pte_val(pfn_pte(pfn, __pgprot(0))) ^ pte_val(pte));
arch/riscv/include/asm/pgtable.h
352
static inline int pte_present(pte_t pte)
arch/riscv/include/asm/pgtable.h
354
return (pte_val(pte) & (_PAGE_PRESENT | _PAGE_PROT_NONE));
arch/riscv/include/asm/pgtable.h
370
static inline int pte_none(pte_t pte)
arch/riscv/include/asm/pgtable.h
372
return (pte_val(pte) == 0);
arch/riscv/include/asm/pgtable.h
375
static inline int pte_write(pte_t pte)
arch/riscv/include/asm/pgtable.h
377
return pte_val(pte) & _PAGE_WRITE;
arch/riscv/include/asm/pgtable.h
380
static inline int pte_exec(pte_t pte)
arch/riscv/include/asm/pgtable.h
382
return pte_val(pte) & _PAGE_EXEC;
arch/riscv/include/asm/pgtable.h
385
static inline int pte_user(pte_t pte)
arch/riscv/include/asm/pgtable.h
387
return pte_val(pte) & _PAGE_USER;
arch/riscv/include/asm/pgtable.h
390
static inline int pte_huge(pte_t pte)
arch/riscv/include/asm/pgtable.h
392
return pte_present(pte) && (pte_val(pte) & _PAGE_LEAF);
arch/riscv/include/asm/pgtable.h
395
static inline int pte_dirty(pte_t pte)
arch/riscv/include/asm/pgtable.h
397
return pte_val(pte) & _PAGE_DIRTY;
arch/riscv/include/asm/pgtable.h
400
static inline int pte_young(pte_t pte)
arch/riscv/include/asm/pgtable.h
402
return pte_val(pte) & _PAGE_ACCESSED;
arch/riscv/include/asm/pgtable.h
405
static inline int pte_special(pte_t pte)
arch/riscv/include/asm/pgtable.h
407
return pte_val(pte) & _PAGE_SPECIAL;
arch/riscv/include/asm/pgtable.h
412
static inline pte_t pte_wrprotect(pte_t pte)
arch/riscv/include/asm/pgtable.h
414
return __pte((pte_val(pte) & ~(_PAGE_WRITE)) | (_PAGE_READ));
arch/riscv/include/asm/pgtable.h
421
static inline bool pte_uffd_wp(pte_t pte)
arch/riscv/include/asm/pgtable.h
423
return !!(pte_val(pte) & _PAGE_UFFD_WP);
arch/riscv/include/asm/pgtable.h
426
static inline pte_t pte_mkuffd_wp(pte_t pte)
arch/riscv/include/asm/pgtable.h
428
return pte_wrprotect(__pte(pte_val(pte) | _PAGE_UFFD_WP));
arch/riscv/include/asm/pgtable.h
431
static inline pte_t pte_clear_uffd_wp(pte_t pte)
arch/riscv/include/asm/pgtable.h
433
return __pte(pte_val(pte) & ~(_PAGE_UFFD_WP));
arch/riscv/include/asm/pgtable.h
436
static inline bool pte_swp_uffd_wp(pte_t pte)
arch/riscv/include/asm/pgtable.h
438
return !!(pte_val(pte) & _PAGE_SWP_UFFD_WP);
arch/riscv/include/asm/pgtable.h
441
static inline pte_t pte_swp_mkuffd_wp(pte_t pte)
arch/riscv/include/asm/pgtable.h
443
return __pte(pte_val(pte) | _PAGE_SWP_UFFD_WP);
arch/riscv/include/asm/pgtable.h
446
static inline pte_t pte_swp_clear_uffd_wp(pte_t pte)
arch/riscv/include/asm/pgtable.h
448
return __pte(pte_val(pte) & ~(_PAGE_SWP_UFFD_WP));
arch/riscv/include/asm/pgtable.h
455
pte_t pte_mkwrite(pte_t pte, struct vm_area_struct *vma);
arch/riscv/include/asm/pgtable.h
458
static inline pte_t pte_mkwrite_novma(pte_t pte)
arch/riscv/include/asm/pgtable.h
460
return __pte(pte_val(pte) | _PAGE_WRITE);
arch/riscv/include/asm/pgtable.h
463
static inline pte_t pte_mkwrite_shstk(pte_t pte)
arch/riscv/include/asm/pgtable.h
465
return __pte((pte_val(pte) & ~(_PAGE_LEAF)) | _PAGE_WRITE);
arch/riscv/include/asm/pgtable.h
470
static inline pte_t pte_mkdirty(pte_t pte)
arch/riscv/include/asm/pgtable.h
472
return __pte(pte_val(pte) | _PAGE_DIRTY | _PAGE_SOFT_DIRTY);
arch/riscv/include/asm/pgtable.h
475
static inline pte_t pte_mkclean(pte_t pte)
arch/riscv/include/asm/pgtable.h
477
return __pte(pte_val(pte) & ~(_PAGE_DIRTY));
arch/riscv/include/asm/pgtable.h
480
static inline pte_t pte_mkyoung(pte_t pte)
arch/riscv/include/asm/pgtable.h
482
return __pte(pte_val(pte) | _PAGE_ACCESSED);
arch/riscv/include/asm/pgtable.h
485
static inline pte_t pte_mkold(pte_t pte)
arch/riscv/include/asm/pgtable.h
487
return __pte(pte_val(pte) & ~(_PAGE_ACCESSED));
arch/riscv/include/asm/pgtable.h
490
static inline pte_t pte_mkspecial(pte_t pte)
arch/riscv/include/asm/pgtable.h
492
return __pte(pte_val(pte) | _PAGE_SPECIAL);
arch/riscv/include/asm/pgtable.h
495
static inline pte_t pte_mkhuge(pte_t pte)
arch/riscv/include/asm/pgtable.h
497
return pte;
arch/riscv/include/asm/pgtable.h
505
static inline bool pte_soft_dirty(pte_t pte)
arch/riscv/include/asm/pgtable.h
507
return !!(pte_val(pte) & _PAGE_SOFT_DIRTY);
arch/riscv/include/asm/pgtable.h
510
static inline pte_t pte_mksoft_dirty(pte_t pte)
arch/riscv/include/asm/pgtable.h
512
return __pte(pte_val(pte) | _PAGE_SOFT_DIRTY);
arch/riscv/include/asm/pgtable.h
515
static inline pte_t pte_clear_soft_dirty(pte_t pte)
arch/riscv/include/asm/pgtable.h
517
return __pte(pte_val(pte) & ~(_PAGE_SOFT_DIRTY));
arch/riscv/include/asm/pgtable.h
520
static inline bool pte_swp_soft_dirty(pte_t pte)
arch/riscv/include/asm/pgtable.h
522
return !!(pte_val(pte) & _PAGE_SWP_SOFT_DIRTY);
arch/riscv/include/asm/pgtable.h
525
static inline pte_t pte_swp_mksoft_dirty(pte_t pte)
arch/riscv/include/asm/pgtable.h
527
return __pte(pte_val(pte) | _PAGE_SWP_SOFT_DIRTY);
arch/riscv/include/asm/pgtable.h
530
static inline pte_t pte_swp_clear_soft_dirty(pte_t pte)
arch/riscv/include/asm/pgtable.h
532
return __pte(pte_val(pte) & ~(_PAGE_SWP_SOFT_DIRTY));
arch/riscv/include/asm/pgtable.h
537
#define pte_leaf_size(pte) (pte_napot(pte) ? \
arch/riscv/include/asm/pgtable.h
538
napot_cont_size(napot_cont_order(pte)) :\
arch/riscv/include/asm/pgtable.h
546
static inline int pte_protnone(pte_t pte)
arch/riscv/include/asm/pgtable.h
548
return (pte_val(pte) & (_PAGE_PRESENT | _PAGE_PROT_NONE)) == _PAGE_PROT_NONE;
arch/riscv/include/asm/pgtable.h
558
static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
arch/riscv/include/asm/pgtable.h
564
return __pte((pte_val(pte) & _PAGE_CHG_MASK) | newprot_val);
arch/riscv/include/asm/pgtable.h
625
void flush_icache_pte(struct mm_struct *mm, pte_t pte);
arch/riscv/include/asm/pgtable.h
670
pte_t pte = __pte(xchg(&ptep->pte, 0));
arch/riscv/include/asm/pgtable.h
672
pte_t pte = *ptep;
arch/riscv/include/asm/pgtable.h
677
page_table_check_pte_clear(mm, address, pte);
arch/riscv/include/asm/pgtable.h
679
return pte;
arch/riscv/include/asm/pgtable.h
763
static inline pmd_t pte_pmd(pte_t pte)
arch/riscv/include/asm/pgtable.h
765
return __pmd(pte_val(pte));
arch/riscv/include/asm/pgtable.h
768
static inline pud_t pte_pud(pte_t pte)
arch/riscv/include/asm/pgtable.h
770
return __pud(pte_val(pte));
arch/riscv/include/asm/pgtable.h
862
static inline pmd_t pmd_mkwrite_shstk(pmd_t pte)
arch/riscv/include/asm/pgtable.h
864
return __pmd((pmd_val(pte) & ~(_PAGE_LEAF)) | _PAGE_WRITE);
arch/riscv/include/asm/pgtable.h
987
static inline bool pte_user_accessible_page(pte_t pte, unsigned long addr)
arch/riscv/include/asm/pgtable.h
989
return pte_present(pte) && pte_user(pte);
arch/riscv/kernel/efi.c
63
pte_t pte = ptep_get(ptep);
arch/riscv/kernel/efi.c
67
val = pte_val(pte) & ~_PAGE_WRITE;
arch/riscv/kernel/efi.c
69
pte = __pte(val);
arch/riscv/kernel/efi.c
72
val = pte_val(pte) & ~_PAGE_EXEC;
arch/riscv/kernel/efi.c
73
pte = __pte(val);
arch/riscv/kernel/efi.c
75
set_pte(ptep, pte);
arch/riscv/kernel/hibernate.c
186
pte_t pte = READ_ONCE(*src_ptep);
arch/riscv/kernel/hibernate.c
188
if (pte_present(pte))
arch/riscv/kernel/hibernate.c
189
set_pte(dst_ptep, __pte(pte_val(pte) | pgprot_val(prot)));
arch/riscv/kvm/gstage.c
157
if (pte_val(*ptep) != pte_val(map->pte)) {
arch/riscv/kvm/gstage.c
158
set_pte(ptep, map->pte);
arch/riscv/kvm/gstage.c
206
out_map->pte = pfn_pte(PFN_DOWN(hpa), prot);
arch/riscv/kvm/gstage.c
207
out_map->pte = pte_mkdirty(out_map->pte);
arch/riscv/kvm/gstage.c
38
static inline unsigned long gstage_pte_page_vaddr(pte_t pte)
arch/riscv/kvm/gstage.c
40
return (unsigned long)pfn_to_virt(__page_val_to_pfn(pte_val(pte)));
arch/riscv/kvm/mmu.c
63
map.pte = pfn_pte(pfn, prot);
arch/riscv/kvm/mmu.c
64
map.pte = pte_mkdirty(map.pte);
arch/riscv/kvm/mmu.c
68
map.pte = pte_wrprotect(map.pte);
arch/riscv/mm/cacheflush.c
100
void flush_icache_pte(struct mm_struct *mm, pte_t pte)
arch/riscv/mm/cacheflush.c
102
struct folio *folio = page_folio(pte_page(pte));
arch/riscv/mm/fault.c
34
pte_t *ptep, pte;
arch/riscv/mm/fault.c
72
pte = ptep_get(ptep);
arch/riscv/mm/fault.c
73
pr_cont(", pte=%016lx", pte_val(pte));
arch/riscv/mm/hugetlbpage.c
121
pte = pte_offset_huge(pmd, addr & napot_cont_mask(order));
arch/riscv/mm/hugetlbpage.c
125
return pte;
arch/riscv/mm/hugetlbpage.c
153
pte_t pte, tmp_pte;
arch/riscv/mm/hugetlbpage.c
156
pte = ptep_get_and_clear(mm, addr, ptep);
arch/riscv/mm/hugetlbpage.c
157
present = pte_present(pte);
arch/riscv/mm/hugetlbpage.c
164
pte = pte_mkdirty(pte);
arch/riscv/mm/hugetlbpage.c
166
pte = pte_mkyoung(pte);
arch/riscv/mm/hugetlbpage.c
169
return pte;
arch/riscv/mm/hugetlbpage.c
18
pte_t pte = ptep_get(ptep);
arch/riscv/mm/hugetlbpage.c
20
if (pte_dirty(pte))
arch/riscv/mm/hugetlbpage.c
23
if (pte_young(pte))
arch/riscv/mm/hugetlbpage.c
249
pte_t pte,
arch/riscv/mm/hugetlbpage.c
257
if (!pte_present(pte)) {
arch/riscv/mm/hugetlbpage.c
259
set_ptes(mm, addr, ptep, pte, 1);
arch/riscv/mm/hugetlbpage.c
263
if (!pte_napot(pte)) {
arch/riscv/mm/hugetlbpage.c
264
set_ptes(mm, addr, ptep, pte, 1);
arch/riscv/mm/hugetlbpage.c
271
set_pte_at(mm, addr, ptep, pte);
arch/riscv/mm/hugetlbpage.c
277
pte_t pte,
arch/riscv/mm/hugetlbpage.c
285
if (!pte_napot(pte))
arch/riscv/mm/hugetlbpage.c
286
return ptep_set_access_flags(vma, addr, ptep, pte, dirty);
arch/riscv/mm/hugetlbpage.c
288
order = napot_cont_order(pte);
arch/riscv/mm/hugetlbpage.c
294
pte = pte_mkdirty(pte);
arch/riscv/mm/hugetlbpage.c
297
pte = pte_mkyoung(pte);
arch/riscv/mm/hugetlbpage.c
300
set_pte_at(mm, addr, ptep, pte);
arch/riscv/mm/hugetlbpage.c
325
pte_t pte = ptep_get(ptep);
arch/riscv/mm/hugetlbpage.c
330
if (!pte_napot(pte)) {
arch/riscv/mm/hugetlbpage.c
335
order = napot_cont_order(pte);
arch/riscv/mm/hugetlbpage.c
350
pte_t pte = ptep_get(ptep);
arch/riscv/mm/hugetlbpage.c
353
if (!pte_napot(pte))
arch/riscv/mm/hugetlbpage.c
356
pte_num = napot_pte_num(napot_cont_order(pte));
arch/riscv/mm/hugetlbpage.c
36
pte_t *pte = NULL;
arch/riscv/mm/hugetlbpage.c
367
pte_t pte = ptep_get(ptep);
arch/riscv/mm/hugetlbpage.c
370
if (!pte_napot(pte)) {
arch/riscv/mm/hugetlbpage.c
52
pte = (pte_t *)pud;
arch/riscv/mm/hugetlbpage.c
58
pte = huge_pmd_share(mm, vma, addr, pud);
arch/riscv/mm/hugetlbpage.c
60
pte = (pte_t *)pmd_alloc(mm, pud, addr);
arch/riscv/mm/hugetlbpage.c
70
pte = pte_alloc_huge(mm, pmd, addr & napot_cont_mask(order));
arch/riscv/mm/hugetlbpage.c
76
if (pte) {
arch/riscv/mm/hugetlbpage.c
77
pte_t pteval = ptep_get_lockless(pte);
arch/riscv/mm/hugetlbpage.c
81
return pte;
arch/riscv/mm/hugetlbpage.c
89
pte_t *pte = NULL;
arch/riscv/mm/init.c
1571
pte_t *pte;
arch/riscv/mm/init.c
1575
pte = pte_start + i;
arch/riscv/mm/init.c
1576
if (!pte_none(*pte))
arch/riscv/mm/init.c
1654
pte_t *ptep, pte;
arch/riscv/mm/init.c
1662
pte = ptep_get(ptep);
arch/riscv/mm/init.c
1668
free_vmemmap_storage(pte_page(pte), PAGE_SIZE, altmap);
arch/riscv/mm/pageattr.c
405
static int debug_pagealloc_set_page(pte_t *pte, unsigned long addr, void *data)
arch/riscv/mm/pageattr.c
409
unsigned long val = pte_val(ptep_get(pte));
arch/riscv/mm/pageattr.c
416
set_pte(pte, __pte(val));
arch/riscv/mm/pageattr.c
442
pte_t *pte;
arch/riscv/mm/pageattr.c
468
pte = pte_offset_kernel(pmd, addr);
arch/riscv/mm/pageattr.c
469
return pte_present(ptep_get(pte));
arch/riscv/mm/pageattr.c
68
static int pageattr_pte_entry(pte_t *pte, unsigned long addr,
arch/riscv/mm/pageattr.c
71
pte_t val = ptep_get(pte);
arch/riscv/mm/pageattr.c
74
set_pte(pte, val);
arch/riscv/mm/pgtable.c
101
pte_free_kernel(NULL, pte);
arch/riscv/mm/pgtable.c
128
pte_t *pte = (pte_t *)pmd_page_vaddr(pmdp_get(pmd));
arch/riscv/mm/pgtable.c
133
pte_free_kernel(NULL, pte);
arch/riscv/mm/pgtable.c
169
pte_t pte_mkwrite(pte_t pte, struct vm_area_struct *vma)
arch/riscv/mm/pgtable.c
172
return pte_mkwrite_shstk(pte);
arch/riscv/mm/pgtable.c
174
return pte_mkwrite_novma(pte);
arch/riscv/mm/pgtable.c
99
pte_t *pte = (pte_t *)pmd_page_vaddr(pmd[i]);
arch/riscv/mm/ptdump.c
320
static void note_page_pte(struct ptdump_state *pt_st, unsigned long addr, pte_t pte)
arch/riscv/mm/ptdump.c
322
note_page(pt_st, addr, 4, pte_val(pte));
arch/s390/boot/vmem.c
181
static bool kasan_pte_populate_zero_shadow(pte_t *pte, enum populate_mode mode)
arch/s390/boot/vmem.c
184
set_pte(pte, pte_z);
arch/s390/boot/vmem.c
219
static bool kasan_pte_populate_zero_shadow(pte_t *pte, enum populate_mode mode)
arch/s390/boot/vmem.c
247
pte_t *pte;
arch/s390/boot/vmem.c
249
pte = (void *)physmem_alloc_or_die(RR_VMEM, PAGE_SIZE, PAGE_SIZE);
arch/s390/boot/vmem.c
250
__arch_set_page_dat(pte, 1);
arch/s390/boot/vmem.c
251
memset64((u64 *)pte, _PAGE_INVALID, PTRS_PER_PTE);
arch/s390/boot/vmem.c
252
return pte;
arch/s390/boot/vmem.c
337
pte_t *pte, entry;
arch/s390/boot/vmem.c
339
pte = pte_offset_kernel(pmd, addr);
arch/s390/boot/vmem.c
340
for (; addr < end; addr += PAGE_SIZE, pte++) {
arch/s390/boot/vmem.c
341
if (pte_none(*pte)) {
arch/s390/boot/vmem.c
342
if (kasan_pte_populate_zero_shadow(pte, mode))
arch/s390/boot/vmem.c
346
set_pte(pte, entry);
arch/s390/boot/vmem.c
359
pte_t *pte;
arch/s390/boot/vmem.c
375
pte = boot_pte_alloc();
arch/s390/boot/vmem.c
376
pmd_populate(&init_mm, pmd, pte);
arch/s390/include/asm/hugetlb.h
22
pte_t *ptep, pte_t pte, unsigned long sz);
arch/s390/include/asm/hugetlb.h
24
pte_t *ptep, pte_t pte);
arch/s390/include/asm/hugetlb.h
60
pte_t pte, int dirty)
arch/s390/include/asm/hugetlb.h
62
int changed = !pte_same(huge_ptep_get(vma->vm_mm, addr, ptep), pte);
arch/s390/include/asm/hugetlb.h
66
__set_huge_pte_at(vma->vm_mm, addr, ptep, pte);
arch/s390/include/asm/hugetlb.h
75
pte_t pte = __huge_ptep_get_and_clear(mm, addr, ptep);
arch/s390/include/asm/hugetlb.h
77
__set_huge_pte_at(mm, addr, ptep, pte_wrprotect(pte));
arch/s390/include/asm/hugetlb.h
81
static inline pte_t huge_pte_mkuffd_wp(pte_t pte)
arch/s390/include/asm/hugetlb.h
83
return pte;
arch/s390/include/asm/hugetlb.h
87
static inline pte_t huge_pte_clear_uffd_wp(pte_t pte)
arch/s390/include/asm/hugetlb.h
89
return pte;
arch/s390/include/asm/hugetlb.h
93
static inline int huge_pte_uffd_wp(pte_t pte)
arch/s390/include/asm/page.h
110
DEFINE_PGVAL_FUNC(pte)
arch/s390/include/asm/page.h
80
typedef struct { unsigned long pte; } pte_t;
arch/s390/include/asm/pgalloc.h
152
pmd_t *pmd, pgtable_t pte)
arch/s390/include/asm/pgalloc.h
154
set_pmd(pmd, __pmd(_SEGMENT_ENTRY | __pa(pte)));
arch/s390/include/asm/pgalloc.h
157
#define pmd_populate_kernel(mm, pmd, pte) pmd_populate(mm, pmd, pte)
arch/s390/include/asm/pgalloc.h
165
#define pte_free_kernel(mm, pte) page_table_free(mm, (unsigned long *) pte)
arch/s390/include/asm/pgalloc.h
166
#define pte_free(mm, pte) page_table_free(mm, (unsigned long *) pte)
arch/s390/include/asm/pgtable.h
1016
static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
arch/s390/include/asm/pgtable.h
1018
pte = clear_pte_bit(pte, __pgprot(~_PAGE_CHG_MASK));
arch/s390/include/asm/pgtable.h
1019
pte = set_pte_bit(pte, newprot);
arch/s390/include/asm/pgtable.h
1024
if ((pte_val(pte) & _PAGE_YOUNG) && (pte_val(pte) & _PAGE_READ))
arch/s390/include/asm/pgtable.h
1025
pte = clear_pte_bit(pte, __pgprot(_PAGE_INVALID));
arch/s390/include/asm/pgtable.h
1030
if ((pte_val(pte) & _PAGE_DIRTY) && (pte_val(pte) & _PAGE_WRITE))
arch/s390/include/asm/pgtable.h
1031
pte = clear_pte_bit(pte, __pgprot(_PAGE_PROTECT));
arch/s390/include/asm/pgtable.h
1032
return pte;
arch/s390/include/asm/pgtable.h
1035
static inline pte_t pte_wrprotect(pte_t pte)
arch/s390/include/asm/pgtable.h
1037
pte = clear_pte_bit(pte, __pgprot(_PAGE_WRITE));
arch/s390/include/asm/pgtable.h
1038
return set_pte_bit(pte, __pgprot(_PAGE_PROTECT));
arch/s390/include/asm/pgtable.h
1041
static inline pte_t pte_mkwrite_novma(pte_t pte)
arch/s390/include/asm/pgtable.h
1043
pte = set_pte_bit(pte, __pgprot(_PAGE_WRITE));
arch/s390/include/asm/pgtable.h
1044
if (pte_val(pte) & _PAGE_DIRTY)
arch/s390/include/asm/pgtable.h
1045
pte = clear_pte_bit(pte, __pgprot(_PAGE_PROTECT));
arch/s390/include/asm/pgtable.h
1046
return pte;
arch/s390/include/asm/pgtable.h
1049
static inline pte_t pte_mkclean(pte_t pte)
arch/s390/include/asm/pgtable.h
1051
pte = clear_pte_bit(pte, __pgprot(_PAGE_DIRTY));
arch/s390/include/asm/pgtable.h
1052
return set_pte_bit(pte, __pgprot(_PAGE_PROTECT));
arch/s390/include/asm/pgtable.h
1055
static inline pte_t pte_mkdirty(pte_t pte)
arch/s390/include/asm/pgtable.h
1057
pte = set_pte_bit(pte, __pgprot(_PAGE_DIRTY | _PAGE_SOFT_DIRTY));
arch/s390/include/asm/pgtable.h
1058
if (pte_val(pte) & _PAGE_WRITE)
arch/s390/include/asm/pgtable.h
1059
pte = clear_pte_bit(pte, __pgprot(_PAGE_PROTECT));
arch/s390/include/asm/pgtable.h
1060
return pte;
arch/s390/include/asm/pgtable.h
1063
static inline pte_t pte_mkold(pte_t pte)
arch/s390/include/asm/pgtable.h
1065
pte = clear_pte_bit(pte, __pgprot(_PAGE_YOUNG));
arch/s390/include/asm/pgtable.h
1066
return set_pte_bit(pte, __pgprot(_PAGE_INVALID));
arch/s390/include/asm/pgtable.h
1069
static inline pte_t pte_mkyoung(pte_t pte)
arch/s390/include/asm/pgtable.h
1071
pte = set_pte_bit(pte, __pgprot(_PAGE_YOUNG));
arch/s390/include/asm/pgtable.h
1072
if (pte_val(pte) & _PAGE_READ)
arch/s390/include/asm/pgtable.h
1073
pte = clear_pte_bit(pte, __pgprot(_PAGE_INVALID));
arch/s390/include/asm/pgtable.h
1074
return pte;
arch/s390/include/asm/pgtable.h
1077
static inline pte_t pte_mkspecial(pte_t pte)
arch/s390/include/asm/pgtable.h
1079
return set_pte_bit(pte, __pgprot(_PAGE_SPECIAL));
arch/s390/include/asm/pgtable.h
1083
static inline pte_t pte_mkhuge(pte_t pte)
arch/s390/include/asm/pgtable.h
1085
return set_pte_bit(pte, __pgprot(_PAGE_LARGE));
arch/s390/include/asm/pgtable.h
1170
pte_t pte = *ptep;
arch/s390/include/asm/pgtable.h
1172
pte = ptep_xchg_direct(vma->vm_mm, addr, ptep, pte_mkold(pte));
arch/s390/include/asm/pgtable.h
1173
return pte_young(pte);
arch/s390/include/asm/pgtable.h
1258
pte_t pte = *ptep;
arch/s390/include/asm/pgtable.h
1260
if (pte_write(pte))
arch/s390/include/asm/pgtable.h
1261
ptep_xchg_lazy(mm, addr, ptep, pte_wrprotect(pte));
arch/s390/include/asm/pgtable.h
1833
#define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) })
arch/s390/include/asm/pgtable.h
1907
pte_t pte;
arch/s390/include/asm/pgtable.h
1910
pte = mk_swap_pte(__swp_type_rste(arch_entry), __swp_offset_rste(arch_entry));
arch/s390/include/asm/pgtable.h
1911
return __pte_to_swp_entry(pte);
arch/s390/include/asm/pgtable.h
556
static inline pte_t clear_pte_bit(pte_t pte, pgprot_t prot)
arch/s390/include/asm/pgtable.h
558
return __pte(pte_val(pte) & ~pgprot_val(prot));
arch/s390/include/asm/pgtable.h
561
static inline pte_t set_pte_bit(pte_t pte, pgprot_t prot)
arch/s390/include/asm/pgtable.h
563
return __pte(pte_val(pte) | pgprot_val(prot));
arch/s390/include/asm/pgtable.h
813
static inline int pte_present(pte_t pte)
arch/s390/include/asm/pgtable.h
816
return (pte_val(pte) & _PAGE_PRESENT) != 0;
arch/s390/include/asm/pgtable.h
819
static inline int pte_none(pte_t pte)
arch/s390/include/asm/pgtable.h
822
return pte_val(pte) == _PAGE_INVALID;
arch/s390/include/asm/pgtable.h
825
static inline int pte_swap(pte_t pte)
arch/s390/include/asm/pgtable.h
828
return (pte_val(pte) & (_PAGE_PROTECT | _PAGE_PRESENT))
arch/s390/include/asm/pgtable.h
832
static inline int pte_special(pte_t pte)
arch/s390/include/asm/pgtable.h
834
return (pte_val(pte) & _PAGE_SPECIAL);
arch/s390/include/asm/pgtable.h
844
static inline int pte_protnone(pte_t pte)
arch/s390/include/asm/pgtable.h
846
return pte_present(pte) && !(pte_val(pte) & _PAGE_READ);
arch/s390/include/asm/pgtable.h
856
static inline bool pte_swp_exclusive(pte_t pte)
arch/s390/include/asm/pgtable.h
858
return pte_val(pte) & _PAGE_SWP_EXCLUSIVE;
arch/s390/include/asm/pgtable.h
861
static inline pte_t pte_swp_mkexclusive(pte_t pte)
arch/s390/include/asm/pgtable.h
863
return set_pte_bit(pte, __pgprot(_PAGE_SWP_EXCLUSIVE));
arch/s390/include/asm/pgtable.h
866
static inline pte_t pte_swp_clear_exclusive(pte_t pte)
arch/s390/include/asm/pgtable.h
868
return clear_pte_bit(pte, __pgprot(_PAGE_SWP_EXCLUSIVE));
arch/s390/include/asm/pgtable.h
871
static inline int pte_soft_dirty(pte_t pte)
arch/s390/include/asm/pgtable.h
873
return pte_val(pte) & _PAGE_SOFT_DIRTY;
arch/s390/include/asm/pgtable.h
877
static inline pte_t pte_mksoft_dirty(pte_t pte)
arch/s390/include/asm/pgtable.h
879
return set_pte_bit(pte, __pgprot(_PAGE_SOFT_DIRTY));
arch/s390/include/asm/pgtable.h
883
static inline pte_t pte_clear_soft_dirty(pte_t pte)
arch/s390/include/asm/pgtable.h
885
return clear_pte_bit(pte, __pgprot(_PAGE_SOFT_DIRTY));
arch/s390/include/asm/pgtable.h
914
static inline int pte_write(pte_t pte)
arch/s390/include/asm/pgtable.h
916
return (pte_val(pte) & _PAGE_WRITE) != 0;
arch/s390/include/asm/pgtable.h
919
static inline int pte_dirty(pte_t pte)
arch/s390/include/asm/pgtable.h
921
return (pte_val(pte) & _PAGE_DIRTY) != 0;
arch/s390/include/asm/pgtable.h
924
static inline int pte_young(pte_t pte)
arch/s390/include/asm/pgtable.h
926
return (pte_val(pte) & _PAGE_YOUNG) != 0;
arch/s390/include/asm/pgtable.h
930
static inline int pte_unused(pte_t pte)
arch/s390/include/asm/pgtable.h
932
return pte_val(pte) & _PAGE_UNUSED;
arch/s390/include/asm/pgtable.h
942
static inline pgprot_t pte_pgprot(pte_t pte)
arch/s390/include/asm/pgtable.h
944
unsigned long pte_flags = pte_val(pte) & _PAGE_CHG_MASK;
arch/s390/include/asm/pgtable.h
946
if (pte_write(pte))
arch/s390/include/asm/pgtable.h
950
pte_flags |= pte_val(pte) & mio_wb_bit_mask;
arch/s390/include/asm/pgtable.h
979
static inline void set_pte(pte_t *ptep, pte_t pte)
arch/s390/include/asm/pgtable.h
981
WRITE_ONCE(*ptep, pte);
arch/s390/include/asm/tlb.h
78
static inline void pte_free_tlb(struct mmu_gather *tlb, pgtable_t pte,
arch/s390/include/asm/tlb.h
85
tlb_remove_ptdesc(tlb, virt_to_ptdesc(pte));
arch/s390/include/asm/uv.h
632
int uv_destroy_pte(pte_t pte);
arch/s390/include/asm/uv.h
633
int uv_convert_from_secure_pte(pte_t pte);
arch/s390/kernel/uv.c
156
int uv_destroy_pte(pte_t pte)
arch/s390/kernel/uv.c
158
VM_WARN_ON(!pte_present(pte));
arch/s390/kernel/uv.c
159
return uv_destroy_folio(pfn_folio(pte_pfn(pte)));
arch/s390/kernel/uv.c
206
int uv_convert_from_secure_pte(pte_t pte)
arch/s390/kernel/uv.c
208
VM_WARN_ON(!pte_present(pte));
arch/s390/kernel/uv.c
209
return uv_convert_from_secure_folio(pfn_folio(pte_pfn(pte)));
arch/s390/kvm/dat.c
1018
static long dat_set_pn_pte(union pte *ptep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/dat.c
1067
union pte *ptep;
arch/s390/kvm/dat.c
1144
static long dat_reset_cmma_pte(union pte *ptep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/dat.c
1176
static long __dat_peek_cmma_pte(union pte *ptep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/dat.c
1216
static long __dat_get_cmma_pte(union pte *ptep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/dat.c
1271
static long __dat_set_cmma_pte(union pte *ptep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/dat.c
1310
union pte *ptep;
arch/s390/kvm/dat.c
162
static void dat_set_storage_key_from_pgste(union pte pte, union pgste pgste)
arch/s390/kvm/dat.c
166
page_set_storage_key(pte_origin(pte), nkey.skey, 0);
arch/s390/kvm/dat.c
169
static void dat_move_storage_key(union pte old, union pte new)
arch/s390/kvm/dat.c
174
static union pgste dat_save_storage_key_into_pgste(union pte pte, union pgste pgste)
arch/s390/kvm/dat.c
178
skey.skey = page_get_storage_key(pte_origin(pte));
arch/s390/kvm/dat.c
188
union pgste __dat_ptep_xchg(union pte *ptep, union pgste pgste, union pte new, gfn_t gfn,
arch/s390/kvm/dat.c
191
union pte old = READ_ONCE(*ptep);
arch/s390/kvm/dat.c
239
union pte init;
arch/s390/kvm/dat.c
379
int walk_level, union crste **last, union pte **ptepp)
arch/s390/kvm/dat.c
617
union pte *ptep;
arch/s390/kvm/dat.c
647
static void dat_update_ptep_sd(union pgste old, union pgste pgste, union pte *ptep)
arch/s390/kvm/dat.c
658
union pte *ptep;
arch/s390/kvm/dat.c
714
union pte *ptep;
arch/s390/kvm/dat.c
759
union pte *ptep;
arch/s390/kvm/dat.c
788
static long dat_reset_skeys_pte(union pte *ptep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/dat.c
841
static long _dat_slot_pte(union pte *ptep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/dat.c
845
union pte new_pte, pte = READ_ONCE(*ptep);
arch/s390/kvm/dat.c
850
if (pte.val == new_pte.val)
arch/s390/kvm/dat.c
907
static void pgste_set_unlock_multiple(union pte *first, int n, union pgste *pgstes)
arch/s390/kvm/dat.c
918
static bool pgste_get_trylock_multiple(union pte *first, int n, union pgste *pgstes)
arch/s390/kvm/dat.c
968
static long _dat_test_young_pte(union pte *ptep, gfn_t start, gfn_t end, struct dat_walk *walk)
arch/s390/kvm/dat.h
304
static_assert(sizeof(union pte) == sizeof(unsigned long));
arch/s390/kvm/dat.h
339
union pte ptes[_PAGE_ENTRIES];
arch/s390/kvm/dat.h
360
long (*pte_entry)(union pte *pte, gfn_t gfn, gfn_t next, struct dat_walk *w);
arch/s390/kvm/dat.h
366
union pte *last_pte;
arch/s390/kvm/dat.h
393
static inline union pte _pte(kvm_pfn_t pfn, bool writable, bool dirty, bool special)
arch/s390/kvm/dat.h
395
union pte res = { .val = PFN_PHYS(pfn) };
arch/s390/kvm/dat.h
492
union pte *ptep; /* Used to resolve the fault, or NULL */
arch/s390/kvm/dat.h
50
#define _PTE_TOK(t, p) ((union pte) { .tok.i = 1, .tok.type = (t), .tok.par = (p) })
arch/s390/kvm/dat.h
514
union pgste __must_check __dat_ptep_xchg(union pte *ptep, union pgste pgste, union pte new,
arch/s390/kvm/dat.h
524
int walk_level, union crste **last, union pte **ptepp);
arch/s390/kvm/dat.h
581
static inline struct page_table *pte_table_start(union pte *ptep)
arch/s390/kvm/dat.h
723
static inline unsigned long pte_origin(union pte pte)
arch/s390/kvm/dat.h
725
return pte.val & PAGE_MASK;
arch/s390/kvm/dat.h
753
static inline union pgste *pgste_of(union pte *pte)
arch/s390/kvm/dat.h
755
return (union pgste *)(pte + _PAGE_ENTRIES);
arch/s390/kvm/dat.h
758
static inline bool pte_hole(union pte pte)
arch/s390/kvm/dat.h
760
return pte.h.i && !pte.tok.pr && pte.tok.type != _DAT_TOKEN_NONE;
arch/s390/kvm/dat.h
844
static inline bool pgste_get_trylock(union pte *ptep, union pgste *res)
arch/s390/kvm/dat.h
859
static inline union pgste pgste_get_lock(union pte *ptep)
arch/s390/kvm/dat.h
868
static inline void pgste_set_unlock(union pte *ptep, union pgste pgste)
arch/s390/kvm/dat.h
875
static inline void dat_ptep_xchg(union pte *ptep, union pte new, gfn_t gfn, union asce asce,
arch/s390/kvm/dat.h
885
static inline void dat_ptep_clear(union pte *ptep, gfn_t gfn, union asce asce, bool has_skeys)
arch/s390/kvm/dat.h
949
static inline int get_level(union crste *crstep, union pte *ptep)
arch/s390/kvm/gaccess.c
1410
if (table.pte.i)
arch/s390/kvm/gaccess.c
1412
if (table.pte.z)
arch/s390/kvm/gaccess.c
1414
w->p |= table.pte.p;
arch/s390/kvm/gaccess.c
1419
return kvm_s390_get_guest_page(kvm, entries + LEVEL_MEM, table.pte.pfra, wr);
arch/s390/kvm/gaccess.c
1422
static int _do_shadow_pte(struct gmap *sg, gpa_t raddr, union pte *ptep_h, union pte *ptep,
arch/s390/kvm/gaccess.c
1426
union pte newpte;
arch/s390/kvm/gaccess.c
1511
union pte *ptep, *ptep_h;
arch/s390/kvm/gaccess.c
34
union page_table_entry pte;
arch/s390/kvm/gaccess.c
476
union page_table_entry pte;
arch/s390/kvm/gaccess.c
608
if (deref_table(vcpu->kvm, ptr, &pte.val))
arch/s390/kvm/gaccess.c
610
if (pte.i)
arch/s390/kvm/gaccess.c
612
if (pte.z)
arch/s390/kvm/gaccess.c
614
dat_protection |= pte.p;
arch/s390/kvm/gaccess.c
615
iep_protection = pte.iep;
arch/s390/kvm/gaccess.c
616
raddr.pfra = pte.pfra;
arch/s390/kvm/gmap.c
1011
union pte *ptep;
arch/s390/kvm/gmap.c
1012
union pte pte;
arch/s390/kvm/gmap.c
1032
pte = ptep->s.pr ? *ptep : _pte(pfn, wr, false, false);
arch/s390/kvm/gmap.c
1033
pte.h.p = 1;
arch/s390/kvm/gmap.c
1034
pgste = _gmap_ptep_xchg(sg->parent, ptep, pte, pgste, p_gfn, false);
arch/s390/kvm/gmap.c
1041
static long __set_cmma_dirty_pte(union pte *ptep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/gmap.c
1068
union pte *ptep;
arch/s390/kvm/gmap.c
283
static long gmap_clear_young_pte(union pte *ptep, gfn_t gfn, gfn_t end, struct dat_walk *walk)
arch/s390/kvm/gmap.c
287
union pte pte, new;
arch/s390/kvm/gmap.c
289
pte = READ_ONCE(*ptep);
arch/s390/kvm/gmap.c
291
if (!pte.s.pr || (!pte.s.y && pte.h.i))
arch/s390/kvm/gmap.c
296
new = pte;
arch/s390/kvm/gmap.c
300
folio_set_dirty(pfn_folio(pte.h.pfra));
arch/s390/kvm/gmap.c
372
static long _gmap_unmap_pte(union pte *ptep, gfn_t gfn, gfn_t next, struct dat_walk *w)
arch/s390/kvm/gmap.c
442
static union pgste __pte_test_and_clear_softdirty(union pte *ptep, union pgste pgste, gfn_t gfn,
arch/s390/kvm/gmap.c
445
union pte pte = READ_ONCE(*ptep);
arch/s390/kvm/gmap.c
447
if (!pte.s.pr || (pte.h.p && !pte.s.sd))
arch/s390/kvm/gmap.c
455
pte.h.p = 1;
arch/s390/kvm/gmap.c
456
pte.s.sd = 0;
arch/s390/kvm/gmap.c
457
pgste = gmap_ptep_xchg(gmap, ptep, pte, pgste, gfn);
arch/s390/kvm/gmap.c
465
static long _pte_test_and_clear_softdirty(union pte *ptep, gfn_t gfn, gfn_t end,
arch/s390/kvm/gmap.c
558
union pte newpte, oldpte = READ_ONCE(*f->ptep);
arch/s390/kvm/gmap.c
635
union pte newpte, oldpte;
arch/s390/kvm/gmap.c
706
union pte *ptep;
arch/s390/kvm/gmap.c
740
union pte *ptep;
arch/s390/kvm/gmap.c
834
union pte *ptep;
arch/s390/kvm/gmap.c
925
static long _destroy_pages_pte(union pte *ptep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/gmap.h
169
static inline union pgste _gmap_ptep_xchg(struct gmap *gmap, union pte *ptep, union pte newpte,
arch/s390/kvm/gmap.h
194
static inline union pgste gmap_ptep_xchg(struct gmap *gmap, union pte *ptep, union pte newpte,
arch/s390/kvm/priv.c
1186
union pte *ptep;
arch/s390/mm/dump_pagetables.c
150
static void note_page_pte(struct ptdump_state *pt_st, unsigned long addr, pte_t pte)
arch/s390/mm/dump_pagetables.c
152
note_page(pt_st, addr, 4, pte_val(pte));
arch/s390/mm/gmap_helpers.c
177
static int find_zeropage_pte_entry(pte_t *pte, unsigned long addr,
arch/s390/mm/gmap_helpers.c
183
if (is_zero_pfn(pte_pfn(*pte))) {
arch/s390/mm/hugetlbpage.c
131
pte = mk_swap_pte(__swp_type_rste(arch_entry), __swp_offset_rste(arch_entry));
arch/s390/mm/hugetlbpage.c
132
pteval = pte_val(pte);
arch/s390/mm/hugetlbpage.c
139
pte_t *ptep, pte_t pte)
arch/s390/mm/hugetlbpage.c
143
rste = __pte_to_rste(pte);
arch/s390/mm/hugetlbpage.c
147
if (likely(pte_present(pte)))
arch/s390/mm/hugetlbpage.c
150
} else if (likely(pte_present(pte)))
arch/s390/mm/hugetlbpage.c
157
pte_t *ptep, pte_t pte, unsigned long sz)
arch/s390/mm/hugetlbpage.c
159
__set_huge_pte_at(mm, addr, ptep, pte);
arch/s390/mm/hugetlbpage.c
170
pte_t pte = huge_ptep_get(mm, addr, ptep);
arch/s390/mm/hugetlbpage.c
178
return pte;
arch/s390/mm/hugetlbpage.c
25
static inline unsigned long __pte_to_rste(pte_t pte)
arch/s390/mm/hugetlbpage.c
50
if (pte_present(pte)) {
arch/s390/mm/hugetlbpage.c
51
rste = pte_val(pte) & PAGE_MASK;
arch/s390/mm/hugetlbpage.c
53
rste |= move_set_bit(pte_val(pte), _PAGE_READ,
arch/s390/mm/hugetlbpage.c
55
rste |= move_set_bit(pte_val(pte), _PAGE_WRITE,
arch/s390/mm/hugetlbpage.c
57
rste |= move_set_bit(pte_val(pte), _PAGE_INVALID,
arch/s390/mm/hugetlbpage.c
59
rste |= move_set_bit(pte_val(pte), _PAGE_PROTECT,
arch/s390/mm/hugetlbpage.c
61
rste |= move_set_bit(pte_val(pte), _PAGE_DIRTY,
arch/s390/mm/hugetlbpage.c
63
rste |= move_set_bit(pte_val(pte), _PAGE_YOUNG,
arch/s390/mm/hugetlbpage.c
66
rste |= move_set_bit(pte_val(pte), _PAGE_SOFT_DIRTY,
arch/s390/mm/hugetlbpage.c
69
rste |= move_set_bit(pte_val(pte), _PAGE_NOEXEC,
arch/s390/mm/hugetlbpage.c
71
} else if (!pte_none(pte)) {
arch/s390/mm/hugetlbpage.c
73
arch_entry = __pte_to_swp_entry(pte);
arch/s390/mm/hugetlbpage.c
85
pte_t pte;
arch/s390/mm/maccess.c
101
set_pte(memcpy_real_ptep, pte);
arch/s390/mm/maccess.c
88
pte_t pte;
arch/s390/mm/maccess.c
96
pte = mk_pte_phys(phys, PAGE_KERNEL_RO);
arch/s390/mm/maccess.c
99
if (pte_val(pte) != pte_val(*memcpy_real_ptep)) {
arch/s390/mm/pageattr.c
425
static void ipte_range(pte_t *pte, unsigned long address, int nr)
arch/s390/mm/pageattr.c
430
__ptep_ipte_range(address, nr - 1, pte, IPTE_GLOBAL);
arch/s390/mm/pageattr.c
434
__ptep_ipte(address, pte, 0, 0, IPTE_GLOBAL);
arch/s390/mm/pageattr.c
436
pte++;
arch/s390/mm/pageattr.c
443
pte_t *ptep, pte;
arch/s390/mm/pageattr.c
454
pte = clear_pte_bit(*ptep, __pgprot(_PAGE_INVALID));
arch/s390/mm/pageattr.c
455
set_pte(ptep, pte);
arch/s390/mm/pgalloc.c
234
unsigned long *pte, next;
arch/s390/mm/pgalloc.c
238
pte = origin;
arch/s390/mm/pgalloc.c
239
pte += (addr & _PAGE_INDEX) >> PAGE_SHIFT;
arch/s390/mm/pgalloc.c
242
*pte = base_lra(addr);
arch/s390/mm/pgalloc.c
243
} while (pte++, addr = next, addr < end);
arch/s390/mm/pgtable.c
176
pte_t *ptep, pte_t old_pte, pte_t pte)
arch/s390/mm/pgtable.c
178
set_pte(ptep, pte);
arch/s390/mm/vmem.c
174
pte_t *pte;
arch/s390/mm/vmem.c
177
pte = pte_offset_kernel(pmd, addr);
arch/s390/mm/vmem.c
178
for (; addr < end; addr += PAGE_SIZE, pte++) {
arch/s390/mm/vmem.c
180
if (pte_none(*pte))
arch/s390/mm/vmem.c
183
vmem_free_pages((unsigned long)pfn_to_virt(pte_pfn(*pte)), get_order(PAGE_SIZE), altmap);
arch/s390/mm/vmem.c
184
pte_clear(&init_mm, addr, pte);
arch/s390/mm/vmem.c
185
} else if (pte_none(*pte)) {
arch/s390/mm/vmem.c
191
set_pte(pte, __pte(__pa(new_page) | prot));
arch/s390/mm/vmem.c
193
set_pte(pte, __pte(__pa(addr) | prot));
arch/s390/mm/vmem.c
209
pte_t *pte;
arch/s390/mm/vmem.c
213
pte = pte_offset_kernel(pmd, start);
arch/s390/mm/vmem.c
214
for (i = 0; i < PTRS_PER_PTE; i++, pte++) {
arch/s390/mm/vmem.c
215
if (!pte_none(*pte))
arch/s390/mm/vmem.c
230
pte_t *pte;
arch/s390/mm/vmem.c
280
pte = vmem_pte_alloc();
arch/s390/mm/vmem.c
281
if (!pte)
arch/s390/mm/vmem.c
283
pmd_populate(&init_mm, pmd, pte);
arch/s390/mm/vmem.c
582
pte_t *pte;
arch/s390/mm/vmem.c
617
pte = vmem_pte_alloc();
arch/s390/mm/vmem.c
618
if (!pte)
arch/s390/mm/vmem.c
620
pmd_populate(&init_mm, pmd, pte);
arch/s390/mm/vmem.c
631
pte_t *ptep, pte;
arch/s390/mm/vmem.c
639
pte = mk_pte_phys(phys, prot);
arch/s390/mm/vmem.c
640
set_pte(ptep, pte);
arch/s390/mm/vmem.c
74
pte_t *pte;
arch/s390/mm/vmem.c
77
pte = (pte_t *)page_table_alloc(&init_mm);
arch/s390/mm/vmem.c
79
pte = (pte_t *)memblock_alloc(PAGE_SIZE, PAGE_SIZE);
arch/s390/mm/vmem.c
80
if (!pte)
arch/s390/mm/vmem.c
82
memset64((u64 *)pte, _PAGE_INVALID, PTRS_PER_PTE);
arch/s390/mm/vmem.c
83
__arch_set_page_dat(pte, 1);
arch/s390/mm/vmem.c
84
return pte;
arch/sh/include/asm/pgalloc.h
24
pte_t *pte)
arch/sh/include/asm/pgalloc.h
26
set_pmd(pmd, __pmd((unsigned long)pte));
arch/sh/include/asm/pgalloc.h
30
pgtable_t pte)
arch/sh/include/asm/pgalloc.h
32
set_pmd(pmd, __pmd((unsigned long)page_address(pte)));
arch/sh/include/asm/pgalloc.h
35
#define __pte_free_tlb(tlb, pte, addr) \
arch/sh/include/asm/pgalloc.h
36
tlb_remove_ptdesc((tlb), page_ptdesc(pte))
arch/sh/include/asm/pgtable.h
101
unsigned long address, pte_t pte);
arch/sh/include/asm/pgtable.h
103
unsigned long address, pte_t pte);
arch/sh/include/asm/pgtable.h
109
pte_t pte = *ptep;
arch/sh/include/asm/pgtable.h
110
__update_cache(vma, address, pte);
arch/sh/include/asm/pgtable.h
111
__update_tlb(vma, address, pte);
arch/sh/include/asm/pgtable.h
121
static inline bool __pte_access_permitted(pte_t pte, u64 prot)
arch/sh/include/asm/pgtable.h
123
return (pte_val(pte) & (prot | _PAGE_SPECIAL)) == prot;
arch/sh/include/asm/pgtable.h
127
static inline bool pte_access_permitted(pte_t pte, bool write)
arch/sh/include/asm/pgtable.h
134
return __pte_access_permitted(pte, prot);
arch/sh/include/asm/pgtable.h
137
static inline bool pte_access_permitted(pte_t pte, bool write)
arch/sh/include/asm/pgtable.h
143
return __pte_access_permitted(pte, prot);
arch/sh/include/asm/pgtable_32.h
300
static inline void set_pte(pte_t *ptep, pte_t pte)
arch/sh/include/asm/pgtable_32.h
302
ptep->pte_high = pte.pte_high;
arch/sh/include/asm/pgtable_32.h
304
ptep->pte_low = pte.pte_low;
arch/sh/include/asm/pgtable_32.h
339
#define pte_not_present(pte) (!((pte).pte_low & _PAGE_PRESENT))
arch/sh/include/asm/pgtable_32.h
340
#define pte_dirty(pte) ((pte).pte_low & _PAGE_DIRTY)
arch/sh/include/asm/pgtable_32.h
341
#define pte_young(pte) ((pte).pte_low & _PAGE_ACCESSED)
arch/sh/include/asm/pgtable_32.h
342
#define pte_special(pte) ((pte).pte_low & _PAGE_SPECIAL)
arch/sh/include/asm/pgtable_32.h
345
#define pte_write(pte) \
arch/sh/include/asm/pgtable_32.h
346
((pte).pte_high & (_PAGE_EXT_USER_WRITE | _PAGE_EXT_KERN_WRITE))
arch/sh/include/asm/pgtable_32.h
348
#define pte_write(pte) ((pte).pte_low & _PAGE_RW)
arch/sh/include/asm/pgtable_32.h
352
static inline pte_t pte_##fn(pte_t pte) { pte.pte_##h op; return pte; }
arch/sh/include/asm/pgtable_32.h
383
static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
arch/sh/include/asm/pgtable_32.h
385
pte.pte_low &= _PAGE_CHG_MASK;
arch/sh/include/asm/pgtable_32.h
386
pte.pte_low |= pgprot_val(newprot);
arch/sh/include/asm/pgtable_32.h
389
pte.pte_high |= pgprot_val(newprot) >> 32;
arch/sh/include/asm/pgtable_32.h
392
return pte;
arch/sh/include/asm/pgtable_32.h
449
#define __pte_to_swp_entry(pte) ((swp_entry_t){ (pte).pte_high })
arch/sh/include/asm/pgtable_32.h
466
#define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) >> 1 })
arch/sh/include/asm/pgtable_32.h
473
static inline bool pte_swp_exclusive(pte_t pte)
arch/sh/include/asm/pgtable_32.h
475
return pte.pte_low & _PAGE_SWP_EXCLUSIVE;
arch/sh/include/asm/tlb.h
17
unsigned long addr, pte_t pte)
arch/sh/mm/cache-sh4.c
222
pte_t *pte;
arch/sh/mm/cache-sh4.c
235
pte = pte_offset_kernel(pmd, address);
arch/sh/mm/cache-sh4.c
238
if (!(pte_val(*pte) & _PAGE_PRESENT))
arch/sh/mm/cache.c
142
unsigned long address, pte_t pte)
arch/sh/mm/cache.c
144
unsigned long pfn = pte_pfn(pte);
arch/sh/mm/fault.c
112
pte = pte_offset_kernel(pmd, addr);
arch/sh/mm/fault.c
113
pr_cont(", *pte=%0*llx", (u32)(sizeof(*pte) * 2),
arch/sh/mm/fault.c
114
(u64)pte_val(*pte));
arch/sh/mm/fault.c
59
pte_t *pte;
arch/sh/mm/hugetlbpage.c
31
pte_t *pte = NULL;
arch/sh/mm/hugetlbpage.c
41
pte = pte_alloc_huge(mm, pmd, addr);
arch/sh/mm/hugetlbpage.c
46
return pte;
arch/sh/mm/hugetlbpage.c
56
pte_t *pte = NULL;
arch/sh/mm/hugetlbpage.c
66
pte = pte_offset_huge(pmd, addr);
arch/sh/mm/hugetlbpage.c
71
return pte;
arch/sh/mm/init.c
100
pte_t *pte;
arch/sh/mm/init.c
102
pte = __get_pte_phys(addr);
arch/sh/mm/init.c
107
set_pte(pte, pfn_pte(0, __pgprot(0)));
arch/sh/mm/init.c
151
pte_t *pte;
arch/sh/mm/init.c
153
pte = memblock_alloc_or_panic(PAGE_SIZE, PAGE_SIZE);
arch/sh/mm/init.c
154
pmd_populate_kernel(&init_mm, pmd, pte);
arch/sh/mm/init.c
155
BUG_ON(pte != pte_offset_kernel(pmd, 0));
arch/sh/mm/init.c
161
static pte_t * __init page_table_kmap_check(pte_t *pte, pmd_t *pmd,
arch/sh/mm/init.c
164
return pte;
arch/sh/mm/init.c
173
pte_t *pte = NULL;
arch/sh/mm/init.c
191
pte = page_table_kmap_check(one_page_table_init(pmd),
arch/sh/mm/init.c
192
pmd, vaddr, pte);
arch/sh/mm/init.c
83
pte_t *pte;
arch/sh/mm/init.c
85
pte = __get_pte_phys(addr);
arch/sh/mm/init.c
86
if (!pte_none(*pte)) {
arch/sh/mm/init.c
87
pte_ERROR(*pte);
arch/sh/mm/init.c
91
set_pte(pte, pfn_pte(phys >> PAGE_SHIFT, prot));
arch/sh/mm/init.c
95
tlb_wire_entry(NULL, addr, *pte);
arch/sh/mm/nommu.c
74
void __update_tlb(struct vm_area_struct *vma, unsigned long address, pte_t pte)
arch/sh/mm/tlb-pteaex.c
18
void __update_tlb(struct vm_area_struct *vma, unsigned long address, pte_t pte)
arch/sh/mm/tlb-pteaex.c
37
pteval = pte.pte_low;
arch/sh/mm/tlb-pteaex.c
47
__raw_writel(pte.pte_high, MMU_PTEA);
arch/sh/mm/tlb-sh3.c
27
void __update_tlb(struct vm_area_struct *vma, unsigned long address, pte_t pte)
arch/sh/mm/tlb-sh3.c
43
pteval = pte_val(pte);
arch/sh/mm/tlb-sh4.c
16
void __update_tlb(struct vm_area_struct *vma, unsigned long address, pte_t pte)
arch/sh/mm/tlb-sh4.c
32
pteval = pte.pte_low;
arch/sh/mm/tlb-sh4.c
42
__raw_writel(pte.pte_high, MMU_PTEA);
arch/sh/mm/tlb-urb.c
20
void tlb_wire_entry(struct vm_area_struct *vma, unsigned long addr, pte_t pte)
arch/sh/mm/tlb-urb.c
47
__update_tlb(vma, addr, pte);
arch/sh/mm/tlbex_32.c
30
pte_t *pte;
arch/sh/mm/tlbex_32.c
56
pte = pte_offset_kernel(pmd, address);
arch/sh/mm/tlbex_32.c
57
entry = *pte;
arch/sh/mm/tlbex_32.c
67
set_pte(pte, entry);
arch/sh/mm/tlbex_32.c
80
update_mmu_cache(NULL, address, pte);
arch/sparc/include/asm/hugetlb.h
17
pte_t *ptep, pte_t pte, unsigned long sz);
arch/sparc/include/asm/hugetlb.h
19
pte_t *ptep, pte_t pte);
arch/sparc/include/asm/hugetlb.h
43
pte_t pte, int dirty)
arch/sparc/include/asm/hugetlb.h
45
int changed = !pte_same(*ptep, pte);
arch/sparc/include/asm/hugetlb.h
47
__set_huge_pte_at(vma->vm_mm, addr, ptep, pte);
arch/sparc/include/asm/mmu_64.h
68
unsigned long pte;
arch/sparc/include/asm/mmu_64.h
71
void __tsb_insert(unsigned long ent, unsigned long tag, unsigned long pte);
arch/sparc/include/asm/page_32.h
53
typedef struct { unsigned long pte; } pte_t;
arch/sparc/include/asm/page_32.h
61
#define pte_val(x) ((x).pte)
arch/sparc/include/asm/page_64.h
67
typedef struct { unsigned long pte; } pte_t;
arch/sparc/include/asm/page_64.h
74
#define pte_val(x) ((x).pte)
arch/sparc/include/asm/pgalloc_32.h
53
#define pmd_populate(mm, pmd, pte) pmd_set(pmd, pte)
arch/sparc/include/asm/pgalloc_32.h
67
static inline void free_pte_fast(pte_t *pte)
arch/sparc/include/asm/pgalloc_32.h
69
srmmu_free_nocache(pte, SRMMU_PTE_TABLE_SIZE);
arch/sparc/include/asm/pgalloc_32.h
72
#define pte_free_kernel(mm, pte) free_pte_fast(pte)
arch/sparc/include/asm/pgalloc_32.h
74
void pte_free(struct mm_struct * mm, pgtable_t pte);
arch/sparc/include/asm/pgalloc_32.h
75
#define __pte_free_tlb(tlb, pte, addr) pte_free((tlb)->mm, pte)
arch/sparc/include/asm/pgalloc_64.h
106
static inline void __pte_free_tlb(struct mmu_gather *tlb, pte_t *pte,
arch/sparc/include/asm/pgalloc_64.h
109
pgtable_free_tlb(tlb, pte, true);
arch/sparc/include/asm/pgalloc_64.h
65
void pte_free_kernel(struct mm_struct *mm, pte_t *pte);
arch/sparc/include/asm/pgtable_32.h
148
static inline int pte_present(pte_t pte)
arch/sparc/include/asm/pgtable_32.h
150
return ((pte_val(pte) & SRMMU_ET_MASK) == SRMMU_ET_PTE);
arch/sparc/include/asm/pgtable_32.h
153
static inline int pte_none(pte_t pte)
arch/sparc/include/asm/pgtable_32.h
155
return !pte_val(pte);
arch/sparc/include/asm/pgtable_32.h
212
static inline int pte_write(pte_t pte)
arch/sparc/include/asm/pgtable_32.h
214
return pte_val(pte) & SRMMU_WRITE;
arch/sparc/include/asm/pgtable_32.h
217
static inline int pte_dirty(pte_t pte)
arch/sparc/include/asm/pgtable_32.h
219
return pte_val(pte) & SRMMU_DIRTY;
arch/sparc/include/asm/pgtable_32.h
222
static inline int pte_young(pte_t pte)
arch/sparc/include/asm/pgtable_32.h
224
return pte_val(pte) & SRMMU_REF;
arch/sparc/include/asm/pgtable_32.h
227
static inline pte_t pte_wrprotect(pte_t pte)
arch/sparc/include/asm/pgtable_32.h
229
return __pte(pte_val(pte) & ~SRMMU_WRITE);
arch/sparc/include/asm/pgtable_32.h
232
static inline pte_t pte_mkclean(pte_t pte)
arch/sparc/include/asm/pgtable_32.h
234
return __pte(pte_val(pte) & ~SRMMU_DIRTY);
arch/sparc/include/asm/pgtable_32.h
237
static inline pte_t pte_mkold(pte_t pte)
arch/sparc/include/asm/pgtable_32.h
239
return __pte(pte_val(pte) & ~SRMMU_REF);
arch/sparc/include/asm/pgtable_32.h
242
static inline pte_t pte_mkwrite_novma(pte_t pte)
arch/sparc/include/asm/pgtable_32.h
244
return __pte(pte_val(pte) | SRMMU_WRITE);
arch/sparc/include/asm/pgtable_32.h
247
static inline pte_t pte_mkdirty(pte_t pte)
arch/sparc/include/asm/pgtable_32.h
249
return __pte(pte_val(pte) | SRMMU_DIRTY);
arch/sparc/include/asm/pgtable_32.h
252
static inline pte_t pte_mkyoung(pte_t pte)
arch/sparc/include/asm/pgtable_32.h
254
return __pte(pte_val(pte) | SRMMU_REF);
arch/sparc/include/asm/pgtable_32.h
264
static inline unsigned long pte_pfn(pte_t pte)
arch/sparc/include/asm/pgtable_32.h
266
if (srmmu_device_memory(pte_val(pte))) {
arch/sparc/include/asm/pgtable_32.h
274
return (pte_val(pte) & SRMMU_PTE_PMASK) >> PFN_PTE_SHIFT;
arch/sparc/include/asm/pgtable_32.h
277
#define pte_page(pte) pfn_to_page(pte_pfn(pte))
arch/sparc/include/asm/pgtable_32.h
296
static pte_t pte_modify(pte_t pte, pgprot_t newprot) __attribute_const__;
arch/sparc/include/asm/pgtable_32.h
297
static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
arch/sparc/include/asm/pgtable_32.h
299
return __pte((pte_val(pte) & SRMMU_CHG_MASK) |
arch/sparc/include/asm/pgtable_32.h
348
#define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) })
arch/sparc/include/asm/pgtable_32.h
351
static inline bool pte_swp_exclusive(pte_t pte)
arch/sparc/include/asm/pgtable_32.h
353
return pte_val(pte) & SRMMU_SWP_EXCLUSIVE;
arch/sparc/include/asm/pgtable_32.h
356
static inline pte_t pte_swp_mkexclusive(pte_t pte)
arch/sparc/include/asm/pgtable_32.h
358
return __pte(pte_val(pte) | SRMMU_SWP_EXCLUSIVE);
arch/sparc/include/asm/pgtable_32.h
361
static inline pte_t pte_swp_clear_exclusive(pte_t pte)
arch/sparc/include/asm/pgtable_32.h
363
return __pte(pte_val(pte) & ~SRMMU_SWP_EXCLUSIVE);
arch/sparc/include/asm/pgtable_64.h
1023
#define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) })
arch/sparc/include/asm/pgtable_64.h
1026
static inline bool pte_swp_exclusive(pte_t pte)
arch/sparc/include/asm/pgtable_64.h
1028
return pte_val(pte) & _PAGE_SWP_EXCLUSIVE;
arch/sparc/include/asm/pgtable_64.h
1031
static inline pte_t pte_swp_mkexclusive(pte_t pte)
arch/sparc/include/asm/pgtable_64.h
1033
return __pte(pte_val(pte) | _PAGE_SWP_EXCLUSIVE);
arch/sparc/include/asm/pgtable_64.h
1036
static inline pte_t pte_swp_clear_exclusive(pte_t pte)
arch/sparc/include/asm/pgtable_64.h
1038
return __pte(pte_val(pte) & ~_PAGE_SWP_EXCLUSIVE);
arch/sparc/include/asm/pgtable_64.h
1052
unsigned long addr, pte_t pte);
arch/sparc/include/asm/pgtable_64.h
1061
pte_t pte, pte_t oldpte)
arch/sparc/include/asm/pgtable_64.h
1070
if (adi_state.enabled && (pte_val(pte) & _PAGE_MCD_4V))
arch/sparc/include/asm/pgtable_64.h
1071
adi_restore_tags(mm, vma, addr, pte);
arch/sparc/include/asm/pgtable_64.h
1120
static inline bool pte_access_permitted(pte_t pte, bool write)
arch/sparc/include/asm/pgtable_64.h
1134
return (pte_val(pte) & (prot | _PAGE_SPECIAL)) == prot;
arch/sparc/include/asm/pgtable_64.h
1172
extern unsigned long pte_leaf_size(pte_t pte);
arch/sparc/include/asm/pgtable_64.h
232
pte_t pte = pfn_pte(page_nr, pgprot);
arch/sparc/include/asm/pgtable_64.h
234
return __pmd(pte_val(pte));
arch/sparc/include/asm/pgtable_64.h
239
static inline unsigned long pte_pfn(pte_t pte)
arch/sparc/include/asm/pgtable_64.h
252
: "r" (pte_val(pte)),
arch/sparc/include/asm/pgtable_64.h
260
static inline pte_t pte_modify(pte_t pte, pgprot_t prot)
arch/sparc/include/asm/pgtable_64.h
315
return __pte((pte_val(pte) & mask) | (pgprot_val(prot) & ~mask));
arch/sparc/include/asm/pgtable_64.h
321
pte_t pte = __pte(pmd_val(pmd));
arch/sparc/include/asm/pgtable_64.h
323
pte = pte_modify(pte, newprot);
arch/sparc/include/asm/pgtable_64.h
325
return __pmd(pte_val(pte));
arch/sparc/include/asm/pgtable_64.h
359
static inline unsigned long pte_dirty(pte_t pte)
arch/sparc/include/asm/pgtable_64.h
374
return (pte_val(pte) & mask);
arch/sparc/include/asm/pgtable_64.h
377
static inline unsigned long pte_write(pte_t pte)
arch/sparc/include/asm/pgtable_64.h
392
return (pte_val(pte) & mask);
arch/sparc/include/asm/pgtable_64.h
416
static inline pte_t pte_mkhuge(pte_t pte)
arch/sparc/include/asm/pgtable_64.h
418
return __pte(pte_val(pte) | __pte_default_huge_mask());
arch/sparc/include/asm/pgtable_64.h
421
static inline bool is_default_hugetlb_pte(pte_t pte)
arch/sparc/include/asm/pgtable_64.h
425
return (pte_val(pte) & mask) == mask;
arch/sparc/include/asm/pgtable_64.h
441
pte_t pte = __pte(pmd_val(pmd));
arch/sparc/include/asm/pgtable_64.h
443
pte = pte_mkhuge(pte);
arch/sparc/include/asm/pgtable_64.h
444
pte_val(pte) |= _PAGE_PMD_HUGE;
arch/sparc/include/asm/pgtable_64.h
446
return __pmd(pte_val(pte));
arch/sparc/include/asm/pgtable_64.h
450
static inline bool is_hugetlb_pte(pte_t pte)
arch/sparc/include/asm/pgtable_64.h
456
static inline pte_t __pte_mkhwwrite(pte_t pte)
arch/sparc/include/asm/pgtable_64.h
458
unsigned long val = pte_val(pte);
arch/sparc/include/asm/pgtable_64.h
476
static inline pte_t pte_mkdirty(pte_t pte)
arch/sparc/include/asm/pgtable_64.h
478
unsigned long val = pte_val(pte), mask;
arch/sparc/include/asm/pgtable_64.h
491
pte = __pte(val | mask);
arch/sparc/include/asm/pgtable_64.h
492
return pte_write(pte) ? __pte_mkhwwrite(pte) : pte;
arch/sparc/include/asm/pgtable_64.h
495
static inline pte_t pte_mkclean(pte_t pte)
arch/sparc/include/asm/pgtable_64.h
497
unsigned long val = pte_val(pte), tmp;
arch/sparc/include/asm/pgtable_64.h
519
static inline pte_t pte_mkwrite_novma(pte_t pte)
arch/sparc/include/asm/pgtable_64.h
521
unsigned long val = pte_val(pte), mask;
arch/sparc/include/asm/pgtable_64.h
534
pte = __pte(val | mask);
arch/sparc/include/asm/pgtable_64.h
535
return pte_dirty(pte) ? __pte_mkhwwrite(pte) : pte;
arch/sparc/include/asm/pgtable_64.h
538
static inline pte_t pte_wrprotect(pte_t pte)
arch/sparc/include/asm/pgtable_64.h
540
unsigned long val = pte_val(pte), tmp;
arch/sparc/include/asm/pgtable_64.h
562
static inline pte_t pte_mkold(pte_t pte)
arch/sparc/include/asm/pgtable_64.h
579
return __pte(pte_val(pte) & ~mask);
arch/sparc/include/asm/pgtable_64.h
582
static inline pte_t pte_mkyoung(pte_t pte)
arch/sparc/include/asm/pgtable_64.h
599
return __pte(pte_val(pte) | mask);
arch/sparc/include/asm/pgtable_64.h
602
static inline pte_t pte_mkspecial(pte_t pte)
arch/sparc/include/asm/pgtable_64.h
604
pte_val(pte) |= _PAGE_SPECIAL;
arch/sparc/include/asm/pgtable_64.h
605
return pte;
arch/sparc/include/asm/pgtable_64.h
608
static inline pte_t pte_mkmcd(pte_t pte)
arch/sparc/include/asm/pgtable_64.h
610
pte_val(pte) |= _PAGE_MCD_4V;
arch/sparc/include/asm/pgtable_64.h
611
return pte;
arch/sparc/include/asm/pgtable_64.h
614
static inline pte_t pte_mknotmcd(pte_t pte)
arch/sparc/include/asm/pgtable_64.h
616
pte_val(pte) &= ~_PAGE_MCD_4V;
arch/sparc/include/asm/pgtable_64.h
617
return pte;
arch/sparc/include/asm/pgtable_64.h
620
static inline unsigned long pte_young(pte_t pte)
arch/sparc/include/asm/pgtable_64.h
635
return (pte_val(pte) & mask);
arch/sparc/include/asm/pgtable_64.h
638
static inline unsigned long pte_exec(pte_t pte)
arch/sparc/include/asm/pgtable_64.h
651
return (pte_val(pte) & mask);
arch/sparc/include/asm/pgtable_64.h
654
static inline unsigned long pte_present(pte_t pte)
arch/sparc/include/asm/pgtable_64.h
656
unsigned long val = pte_val(pte);
arch/sparc/include/asm/pgtable_64.h
676
static inline unsigned long pte_special(pte_t pte)
arch/sparc/include/asm/pgtable_64.h
678
return pte_val(pte) & _PAGE_SPECIAL;
arch/sparc/include/asm/pgtable_64.h
684
pte_t pte = __pte(pmd_val(pmd));
arch/sparc/include/asm/pgtable_64.h
686
return pte_val(pte) & _PAGE_PMD_HUGE;
arch/sparc/include/asm/pgtable_64.h
691
pte_t pte = __pte(pmd_val(pmd));
arch/sparc/include/asm/pgtable_64.h
693
return pte_pfn(pte);
arch/sparc/include/asm/pgtable_64.h
699
pte_t pte = __pte(pmd_val(pmd));
arch/sparc/include/asm/pgtable_64.h
701
return pte_write(pte);
arch/sparc/include/asm/pgtable_64.h
710
pte_t pte = __pte(pmd_val(pmd));
arch/sparc/include/asm/pgtable_64.h
712
return pte_dirty(pte);
arch/sparc/include/asm/pgtable_64.h
718
pte_t pte = __pte(pmd_val(pmd));
arch/sparc/include/asm/pgtable_64.h
720
return pte_young(pte);
arch/sparc/include/asm/pgtable_64.h
725
pte_t pte = __pte(pmd_val(pmd));
arch/sparc/include/asm/pgtable_64.h
727
return pte_val(pte) & _PAGE_PMD_HUGE;
arch/sparc/include/asm/pgtable_64.h
732
pte_t pte = __pte(pmd_val(pmd));
arch/sparc/include/asm/pgtable_64.h
734
pte = pte_mkold(pte);
arch/sparc/include/asm/pgtable_64.h
736
return __pmd(pte_val(pte));
arch/sparc/include/asm/pgtable_64.h
741
pte_t pte = __pte(pmd_val(pmd));
arch/sparc/include/asm/pgtable_64.h
743
pte = pte_wrprotect(pte);
arch/sparc/include/asm/pgtable_64.h
745
return __pmd(pte_val(pte));
arch/sparc/include/asm/pgtable_64.h
750
pte_t pte = __pte(pmd_val(pmd));
arch/sparc/include/asm/pgtable_64.h
752
pte = pte_mkdirty(pte);
arch/sparc/include/asm/pgtable_64.h
754
return __pmd(pte_val(pte));
arch/sparc/include/asm/pgtable_64.h
759
pte_t pte = __pte(pmd_val(pmd));
arch/sparc/include/asm/pgtable_64.h
761
pte = pte_mkclean(pte);
arch/sparc/include/asm/pgtable_64.h
763
return __pmd(pte_val(pte));
arch/sparc/include/asm/pgtable_64.h
768
pte_t pte = __pte(pmd_val(pmd));
arch/sparc/include/asm/pgtable_64.h
770
pte = pte_mkyoung(pte);
arch/sparc/include/asm/pgtable_64.h
772
return __pmd(pte_val(pte));
arch/sparc/include/asm/pgtable_64.h
777
pte_t pte = __pte(pmd_val(pmd));
arch/sparc/include/asm/pgtable_64.h
779
pte = pte_mkwrite_novma(pte);
arch/sparc/include/asm/pgtable_64.h
781
return __pmd(pte_val(pte));
arch/sparc/include/asm/pgtable_64.h
838
pte_t pte = __pte(pmd_val(pmd));
arch/sparc/include/asm/pgtable_64.h
841
pfn = pte_pfn(pte);
arch/sparc/include/asm/pgtable_64.h
848
pte_t pte = __pte(pud_val(pud));
arch/sparc/include/asm/pgtable_64.h
851
pfn = pte_pfn(pte);
arch/sparc/include/asm/pgtable_64.h
872
pte_t pte = __pte(pud_val(pud));
arch/sparc/include/asm/pgtable_64.h
874
return pte_val(pte) & _PAGE_PMD_HUGE;
arch/sparc/include/asm/pgtable_64.h
880
pte_t pte = __pte(pud_val(pud));
arch/sparc/include/asm/pgtable_64.h
882
return pte_pfn(pte);
arch/sparc/include/asm/pgtable_64.h
886
#define pte_none(pte) (!pte_val(pte))
arch/sparc/include/asm/pgtable_64.h
924
pte_t *ptep, pte_t pte, int fullmm)
arch/sparc/include/asm/pgtable_64.h
928
*ptep = pte;
arch/sparc/include/asm/pgtable_64.h
935
pte_t *ptep, pte_t pte, unsigned int nr)
arch/sparc/include/asm/pgtable_64.h
938
__set_pte_at(mm, addr, ptep, pte, 0);
arch/sparc/include/asm/pgtable_64.h
942
pte_val(pte) += PAGE_SIZE;
arch/sparc/include/asm/pgtable_64.h
957
#define move_pte(pte, old_addr, new_addr) \
arch/sparc/include/asm/pgtable_64.h
959
pte_t newpte = (pte); \
arch/sparc/include/asm/pgtable_64.h
960
if (tlb_type != hypervisor && pte_present(pte)) { \
arch/sparc/include/asm/pgtable_64.h
961
unsigned long this_pfn = pte_pfn(pte); \
arch/sparc/kernel/adi_64.c
328
unsigned long addr, pte_t pte)
arch/sparc/kernel/adi_64.c
343
paddr = pte_val(pte) & _PAGE_PADDR_4V;
arch/sparc/kernel/signal32.c
298
pte_t *ptep, pte;
arch/sparc/kernel/signal32.c
333
pte = *ptep;
arch/sparc/kernel/signal32.c
334
if (!pte_present(pte))
arch/sparc/kernel/signal32.c
337
paddr = (unsigned long) page_address(pte_page(pte));
arch/sparc/mm/fault_64.c
121
pte = *ptep;
arch/sparc/mm/fault_64.c
122
if (pte_present(pte)) {
arch/sparc/mm/fault_64.c
123
pa = (pte_pfn(pte) << PAGE_SHIFT);
arch/sparc/mm/fault_64.c
85
pte_t *ptep, pte;
arch/sparc/mm/hugetlbpage.c
100
return pte_mkmcd(pte);
arch/sparc/mm/hugetlbpage.c
102
return pte_mknotmcd(pte);
arch/sparc/mm/hugetlbpage.c
104
return pte;
arch/sparc/mm/hugetlbpage.c
177
static unsigned long huge_tte_to_size(pte_t pte)
arch/sparc/mm/hugetlbpage.c
179
unsigned long size = 1UL << huge_tte_to_shift(pte);
arch/sparc/mm/hugetlbpage.c
188
unsigned long pte_leaf_size(pte_t pte) { return 1UL << tte_to_shift(pte); }
arch/sparc/mm/hugetlbpage.c
91
pte_t pte;
arch/sparc/mm/hugetlbpage.c
94
pte = hugepage_shift_to_tte(entry, shift);
arch/sparc/mm/init_64.c
1626
pte_t *pte;
arch/sparc/mm/init_64.c
1663
pte = pte_offset_kernel(pmd, addr);
arch/sparc/mm/init_64.c
1664
if (pte_none(*pte))
arch/sparc/mm/init_64.c
1667
return pfn_valid(pte_pfn(*pte));
arch/sparc/mm/init_64.c
1777
pte_t *pte;
arch/sparc/mm/init_64.c
1834
pte = pte_offset_kernel(pmd, vstart);
arch/sparc/mm/init_64.c
1840
pte_val(*pte) = (paddr | pgprot_val(prot));
arch/sparc/mm/init_64.c
1844
pte++;
arch/sparc/mm/init_64.c
2589
unsigned long pte;
arch/sparc/mm/init_64.c
2606
pte = pmd_val(*pmd);
arch/sparc/mm/init_64.c
2607
if (!(pte & _PAGE_VALID)) {
arch/sparc/mm/init_64.c
274
static inline void tsb_insert(struct tsb *ent, unsigned long tag, unsigned long pte)
arch/sparc/mm/init_64.c
2785
pte_t pte;
arch/sparc/mm/init_64.c
2787
pte_val(pte) = page | pgprot_val(pgprot_noncached(prot));
arch/sparc/mm/init_64.c
2788
pte_val(pte) |= (((unsigned long)space) << 32);
arch/sparc/mm/init_64.c
2789
pte_val(pte) |= pte_sz_bits(page_size);
arch/sparc/mm/init_64.c
2791
return pte;
arch/sparc/mm/init_64.c
281
__tsb_insert(tsb_addr, tag, pte);
arch/sparc/mm/init_64.c
2891
static void __pte_free(pgtable_t pte)
arch/sparc/mm/init_64.c
2893
struct ptdesc *ptdesc = virt_to_ptdesc(pte);
arch/sparc/mm/init_64.c
2899
void pte_free_kernel(struct mm_struct *mm, pte_t *pte)
arch/sparc/mm/init_64.c
2901
__pte_free(pte);
arch/sparc/mm/init_64.c
2904
void pte_free(struct mm_struct *mm, pgtable_t pte)
arch/sparc/mm/init_64.c
2906
__pte_free(pte);
arch/sparc/mm/init_64.c
2937
unsigned long pte, flags;
arch/sparc/mm/init_64.c
2944
pte = pmd_val(entry);
arch/sparc/mm/init_64.c
2947
if (!(pte & _PAGE_VALID))
arch/sparc/mm/init_64.c
2951
pte |= (addr & (1UL << REAL_HPAGE_SHIFT));
arch/sparc/mm/init_64.c
2959
addr, pte);
arch/sparc/mm/init_64.c
397
pte_t pte = *ptep;
arch/sparc/mm/init_64.c
401
unsigned long pfn = pte_pfn(pte);
arch/sparc/mm/init_64.c
410
if (!pte_accessible(mm, pte))
arch/sparc/mm/init_64.c
429
pte_val(pte) &= ~mask;
arch/sparc/mm/init_64.c
430
pte_val(pte) |= (address & mask);
arch/sparc/mm/init_64.c
435
pte_val(pte) |= (address & (1UL << REAL_HPAGE_SHIFT));
arch/sparc/mm/init_64.c
440
REAL_HPAGE_SHIFT, address, pte_val(pte));
arch/sparc/mm/init_64.c
448
address, pte_val(pte));
arch/sparc/mm/init_64.c
450
pte_val(pte) += PAGE_SIZE;
arch/sparc/mm/init_64.c
659
unsigned long pte,
arch/sparc/mm/init_64.c
662
unsigned long ret = sun4v_mmu_map_perm_addr(vaddr, 0, pte, mmu);
arch/sparc/mm/init_64.c
666
"errors with %lx\n", vaddr, 0, pte, mmu, ret);
arch/sparc/mm/leon_mm.c
122
pte = ped;
arch/sparc/mm/leon_mm.c
144
pte = ptr;
arch/sparc/mm/leon_mm.c
155
(vaddr & ~(-1 << LEON_PTE_SH)) | ((pte & ~0xff) << 4);
arch/sparc/mm/leon_mm.c
159
(vaddr & ~(-1 << LEON_PMD_SH)) | ((pte & ~0xff) << 4);
arch/sparc/mm/leon_mm.c
163
(vaddr & ~(-1 << LEON_PGD_SH)) | ((pte & ~0xff) << 4);
arch/sparc/mm/leon_mm.c
174
return pte;
arch/sparc/mm/leon_mm.c
42
unsigned int lvl, pte;
arch/sparc/mm/leon_mm.c
73
pte = pgd;
arch/sparc/mm/leon_mm.c
95
pte = pmd;
arch/sparc/mm/srmmu.c
112
pte_t pte;
arch/sparc/mm/srmmu.c
114
pte = __pte((SRMMU_ET_PTD | (__nocache_pa(pgdp) >> 4)));
arch/sparc/mm/srmmu.c
115
set_pte((pte_t *)ctxp, pte);
arch/sparc/mm/srmmu.c
274
pte_t *pte;
arch/sparc/mm/srmmu.c
303
pte = pte_offset_kernel(__nocache_fix(pmd), vaddr);
arch/sparc/mm/srmmu.c
310
set_pte(__nocache_fix(pte), __pte(pteval));
arch/sparc/mm/srmmu.c
903
pte_t *pte;
arch/sparc/mm/srmmu.c
966
pte = pte_offset_kernel(pmd, PKMAP_BASE);
arch/sparc/mm/srmmu.c
967
pkmap_page_table = pte;
arch/sparc/mm/tlb.c
160
pte_t *pte;
arch/sparc/mm/tlb.c
162
pte = pte_offset_map(&pmd, vaddr);
arch/sparc/mm/tlb.c
163
if (!pte)
arch/sparc/mm/tlb.c
167
if (pte_val(*pte) & _PAGE_VALID) {
arch/sparc/mm/tlb.c
168
bool exec = pte_exec(*pte);
arch/sparc/mm/tlb.c
172
pte++;
arch/sparc/mm/tlb.c
175
pte_unmap(pte);
arch/um/include/asm/page.h
31
typedef struct { unsigned long pte; } pte_t;
arch/um/include/asm/page.h
49
#define pte_val(x) ((x).pte)
arch/um/include/asm/page.h
51
#define pte_get_bits(p, bits) ((p).pte & (bits))
arch/um/include/asm/page.h
52
#define pte_set_bits(p, bits) ((p).pte |= (bits))
arch/um/include/asm/page.h
53
#define pte_clear_bits(p, bits) ((p).pte &= ~(bits))
arch/um/include/asm/page.h
54
#define pte_copy(to, from) ((to).pte = (from).pte)
arch/um/include/asm/page.h
55
#define pte_is_zero(p) (!((p).pte & ~_PAGE_NEEDSYNC))
arch/um/include/asm/page.h
56
#define pte_set_val(p, phys, prot) (p).pte = (phys | pgprot_val(prot))
arch/um/include/asm/pgalloc.h
15
#define pmd_populate_kernel(mm, pmd, pte) \
arch/um/include/asm/pgalloc.h
16
set_pmd(pmd, __pmd(_PAGE_TABLE + (unsigned long) __pa(pte)))
arch/um/include/asm/pgalloc.h
18
#define pmd_populate(mm, pmd, pte) \
arch/um/include/asm/pgalloc.h
20
((unsigned long long)page_to_pfn(pte) << \
arch/um/include/asm/pgalloc.h
28
#define __pte_free_tlb(tlb, pte, address) \
arch/um/include/asm/pgalloc.h
29
tlb_remove_ptdesc((tlb), page_ptdesc(pte))
arch/um/include/asm/pgtable-4level.h
100
static inline unsigned long pte_pfn(pte_t pte)
arch/um/include/asm/pgtable-4level.h
102
return phys_to_pfn(pte_val(pte));
arch/um/include/asm/pgtable.h
113
static inline int pte_none(pte_t pte)
arch/um/include/asm/pgtable.h
115
return pte_is_zero(pte);
arch/um/include/asm/pgtable.h
122
static inline int pte_read(pte_t pte)
arch/um/include/asm/pgtable.h
124
return((pte_get_bits(pte, _PAGE_USER)) &&
arch/um/include/asm/pgtable.h
125
!(pte_get_bits(pte, _PAGE_PROTNONE)));
arch/um/include/asm/pgtable.h
128
static inline int pte_exec(pte_t pte){
arch/um/include/asm/pgtable.h
129
return((pte_get_bits(pte, _PAGE_USER)) &&
arch/um/include/asm/pgtable.h
130
!(pte_get_bits(pte, _PAGE_PROTNONE)));
arch/um/include/asm/pgtable.h
133
static inline int pte_write(pte_t pte)
arch/um/include/asm/pgtable.h
135
return((pte_get_bits(pte, _PAGE_RW)) &&
arch/um/include/asm/pgtable.h
136
!(pte_get_bits(pte, _PAGE_PROTNONE)));
arch/um/include/asm/pgtable.h
139
static inline int pte_dirty(pte_t pte)
arch/um/include/asm/pgtable.h
141
return pte_get_bits(pte, _PAGE_DIRTY);
arch/um/include/asm/pgtable.h
144
static inline int pte_young(pte_t pte)
arch/um/include/asm/pgtable.h
146
return pte_get_bits(pte, _PAGE_ACCESSED);
arch/um/include/asm/pgtable.h
149
static inline int pte_needsync(pte_t pte)
arch/um/include/asm/pgtable.h
151
return pte_get_bits(pte, _PAGE_NEEDSYNC);
arch/um/include/asm/pgtable.h
160
static inline pte_t pte_mkclean(pte_t pte)
arch/um/include/asm/pgtable.h
162
pte_clear_bits(pte, _PAGE_DIRTY);
arch/um/include/asm/pgtable.h
163
return(pte);
arch/um/include/asm/pgtable.h
166
static inline pte_t pte_mkold(pte_t pte)
arch/um/include/asm/pgtable.h
168
pte_clear_bits(pte, _PAGE_ACCESSED);
arch/um/include/asm/pgtable.h
169
return(pte);
arch/um/include/asm/pgtable.h
172
static inline pte_t pte_wrprotect(pte_t pte)
arch/um/include/asm/pgtable.h
174
pte_clear_bits(pte, _PAGE_RW);
arch/um/include/asm/pgtable.h
175
return pte;
arch/um/include/asm/pgtable.h
178
static inline pte_t pte_mkread(pte_t pte)
arch/um/include/asm/pgtable.h
180
pte_set_bits(pte, _PAGE_USER);
arch/um/include/asm/pgtable.h
181
return pte;
arch/um/include/asm/pgtable.h
184
static inline pte_t pte_mkdirty(pte_t pte)
arch/um/include/asm/pgtable.h
186
pte_set_bits(pte, _PAGE_DIRTY);
arch/um/include/asm/pgtable.h
187
return(pte);
arch/um/include/asm/pgtable.h
190
static inline pte_t pte_mkyoung(pte_t pte)
arch/um/include/asm/pgtable.h
192
pte_set_bits(pte, _PAGE_ACCESSED);
arch/um/include/asm/pgtable.h
193
return(pte);
arch/um/include/asm/pgtable.h
196
static inline pte_t pte_mkwrite_novma(pte_t pte)
arch/um/include/asm/pgtable.h
198
pte_set_bits(pte, _PAGE_RW);
arch/um/include/asm/pgtable.h
199
return pte;
arch/um/include/asm/pgtable.h
202
static inline pte_t pte_mkuptodate(pte_t pte)
arch/um/include/asm/pgtable.h
204
pte_clear_bits(pte, _PAGE_NEEDSYNC);
arch/um/include/asm/pgtable.h
205
return pte;
arch/um/include/asm/pgtable.h
208
static inline pte_t pte_mkneedsync(pte_t pte)
arch/um/include/asm/pgtable.h
210
pte_set_bits(pte, _PAGE_NEEDSYNC);
arch/um/include/asm/pgtable.h
211
return(pte);
arch/um/include/asm/pgtable.h
245
pte_t *ptep, pte_t pte, int nr)
arch/um/include/asm/pgtable.h
251
set_pte(ptep, pte);
arch/um/include/asm/pgtable.h
255
pte = __pte(pte_val(pte) + (nr << PFN_PTE_SHIFT));
arch/um/include/asm/pgtable.h
272
pte_t pte;
arch/um/include/asm/pgtable.h
274
pte_set_val(pte, pfn_to_phys(pfn), pgprot);
arch/um/include/asm/pgtable.h
276
return pte;
arch/um/include/asm/pgtable.h
279
static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
arch/um/include/asm/pgtable.h
281
pte_set_val(pte, (pte_val(pte) & _PAGE_CHG_MASK), newprot);
arch/um/include/asm/pgtable.h
282
return pte;
arch/um/include/asm/pgtable.h
317
#define __pte_to_swp_entry(pte) \
arch/um/include/asm/pgtable.h
318
((swp_entry_t) { pte_val(pte_mkuptodate(pte)) })
arch/um/include/asm/pgtable.h
321
static inline bool pte_swp_exclusive(pte_t pte)
arch/um/include/asm/pgtable.h
323
return pte_get_bits(pte, _PAGE_SWP_EXCLUSIVE);
arch/um/include/asm/pgtable.h
326
static inline pte_t pte_swp_mkexclusive(pte_t pte)
arch/um/include/asm/pgtable.h
328
pte_set_bits(pte, _PAGE_SWP_EXCLUSIVE);
arch/um/include/asm/pgtable.h
329
return pte;
arch/um/include/asm/pgtable.h
332
static inline pte_t pte_swp_clear_exclusive(pte_t pte)
arch/um/include/asm/pgtable.h
334
pte_clear_bits(pte, _PAGE_SWP_EXCLUSIVE);
arch/um/include/asm/pgtable.h
335
return pte;
arch/um/kernel/skas/uaccess.c
256
pte_t *pte;
arch/um/kernel/skas/uaccess.c
262
pte = maybe_map(addr, 1);
arch/um/kernel/skas/uaccess.c
263
if (pte == NULL)
arch/um/kernel/skas/uaccess.c
266
page = pte_page(*pte);
arch/um/kernel/skas/uaccess.c
334
pte_t *pte;
arch/um/kernel/skas/uaccess.c
341
pte = maybe_map((unsigned long) uaddr, 1);
arch/um/kernel/skas/uaccess.c
342
if (pte == NULL)
arch/um/kernel/skas/uaccess.c
345
page = pte_page(*pte);
arch/um/kernel/skas/uaccess.c
48
pte_t *pte = virt_to_pte(current->mm, virt);
arch/um/kernel/skas/uaccess.c
51
if ((pte == NULL) || !pte_present(*pte) ||
arch/um/kernel/skas/uaccess.c
52
(is_write && !pte_write(*pte))) {
arch/um/kernel/skas/uaccess.c
56
pte = virt_to_pte(current->mm, virt);
arch/um/kernel/skas/uaccess.c
58
if (!pte_present(*pte))
arch/um/kernel/skas/uaccess.c
59
pte = NULL;
arch/um/kernel/skas/uaccess.c
61
return pte;
arch/um/kernel/skas/uaccess.c
68
pte_t *pte;
arch/um/kernel/skas/uaccess.c
71
pte = maybe_map(addr, is_write);
arch/um/kernel/skas/uaccess.c
72
if (pte == NULL)
arch/um/kernel/skas/uaccess.c
75
page = pte_page(*pte);
arch/um/kernel/tlb.c
55
pte_t *pte;
arch/um/kernel/tlb.c
58
pte = pte_offset_kernel(pmd, addr);
arch/um/kernel/tlb.c
60
if (!pte_needsync(*pte))
arch/um/kernel/tlb.c
63
if (pte_present(*pte)) {
arch/um/kernel/tlb.c
65
unsigned long phys = pte_val(*pte) & PAGE_MASK;
arch/um/kernel/tlb.c
69
r = pte_read(*pte);
arch/um/kernel/tlb.c
70
w = pte_write(*pte);
arch/um/kernel/tlb.c
71
x = pte_exec(*pte);
arch/um/kernel/tlb.c
72
if (!pte_young(*pte)) {
arch/um/kernel/tlb.c
75
} else if (!pte_dirty(*pte))
arch/um/kernel/tlb.c
87
*pte = pte_mkuptodate(*pte);
arch/um/kernel/tlb.c
88
} while (pte++, addr += PAGE_SIZE, ((addr < end) && !ret));
arch/um/kernel/trap.c
146
pte_t *pte;
arch/um/kernel/trap.c
207
pte = pte_offset_kernel(pmd, address);
arch/um/kernel/trap.c
208
} while (!pte_present(*pte));
arch/um/kernel/trap.c
219
WARN_ON(!pte_young(*pte) || (is_write && !pte_dirty(*pte)));
arch/x86/boot/compressed/ident_map_64.c
197
pte_t *pte;
arch/x86/boot/compressed/ident_map_64.c
201
pte = (pte_t *)info->alloc_pgt_page(info->context);
arch/x86/boot/compressed/ident_map_64.c
202
if (!pte)
arch/x86/boot/compressed/ident_map_64.c
211
set_pte(&pte[i], __pte(address | page_flags));
arch/x86/boot/compressed/ident_map_64.c
224
pmd = __pmd((unsigned long)pte | info->kernpg_flag);
arch/x86/boot/compressed/ident_map_64.c
229
return pte + pte_index(__address);
arch/x86/boot/compressed/ident_map_64.c
263
pte_t *ptep, pte;
arch/x86/boot/compressed/ident_map_64.c
311
pte = *ptep;
arch/x86/boot/compressed/ident_map_64.c
312
pte = pte_set_flags(pte, set);
arch/x86/boot/compressed/ident_map_64.c
313
pte = pte_clear_flags(pte, clr);
arch/x86/boot/compressed/ident_map_64.c
314
set_pte(ptep, pte);
arch/x86/boot/startup/sme.c
167
pte_t *pte;
arch/x86/boot/startup/sme.c
175
pte = ppd->pgtable_area;
arch/x86/boot/startup/sme.c
176
memset(pte, 0, sizeof(*pte) * PTRS_PER_PTE);
arch/x86/boot/startup/sme.c
177
ppd->pgtable_area += sizeof(*pte) * PTRS_PER_PTE;
arch/x86/boot/startup/sme.c
178
set_pmd(pmd, __pmd(PMD_FLAGS | __pa(pte)));
arch/x86/boot/startup/sme.c
184
pte = pte_offset_kernel(pmd, ppd->vaddr);
arch/x86/boot/startup/sme.c
185
if (pte_none(*pte))
arch/x86/boot/startup/sme.c
186
set_pte(pte, __pte(ppd->paddr | ppd->pte_flags));
arch/x86/coco/sev/core.c
554
pte_t *pte;
arch/x86/coco/sev/core.c
562
pte = lookup_address(addr, &level);
arch/x86/coco/sev/core.c
567
if (!pte || !pte_decrypted(*pte) || pte_none(*pte)) {
arch/x86/coco/sev/core.c
589
set_pte_enc(pte, level, (void *)addr);
arch/x86/coco/sev/core.c
601
pte = lookup_address(addr, &level);
arch/x86/coco/sev/core.c
602
if (!pte || !pte_decrypted(*pte) || pte_none(*pte))
arch/x86/coco/sev/core.c
605
set_pte_enc(pte, level, (void *)addr);
arch/x86/coco/sev/core.c
697
pte_t *pte;
arch/x86/coco/sev/core.c
721
pte = lookup_address((unsigned long)ghcb, &level);
arch/x86/coco/sev/core.c
725
set_pte_enc(pte, level, (void *)addr);
arch/x86/coco/sev/vc-handle.c
46
pte_t *pte;
arch/x86/coco/sev/vc-handle.c
50
pte = lookup_address_in_pgd(pgd, va, &level);
arch/x86/coco/sev/vc-handle.c
51
if (!pte) {
arch/x86/coco/sev/vc-handle.c
62
if (WARN_ON_ONCE(pte_val(*pte) & _PAGE_ENC))
arch/x86/coco/sev/vc-handle.c
66
pa = (phys_addr_t)pte_pfn(*pte) << PAGE_SHIFT;
arch/x86/coco/tdx/tdx.c
1043
pte_t *pte;
arch/x86/coco/tdx/tdx.c
1045
pte = lookup_address(addr, &level);
arch/x86/coco/tdx/tdx.c
1048
if (pte && pte_decrypted(*pte)) {
arch/x86/coco/tdx/tdx.c
1058
set_pte(pte, __pte(0));
arch/x86/hyperv/hv_crash.c
512
pte_t *pte;
arch/x86/hyperv/hv_crash.c
528
pte = hv_crash_ptpgs[3] + pte_index(addr) * sizeof(pte);
arch/x86/hyperv/hv_crash.c
529
set_pte(pte, pfn_pte(addr >> PAGE_SHIFT, PAGE_KERNEL_EXEC));
arch/x86/include/asm/fixmap.h
159
void __native_set_fixmap(enum fixed_addresses idx, pte_t pte);
arch/x86/include/asm/kexec.h
166
pte_t *pte;
arch/x86/include/asm/kfence.h
44
pte_t *pte = lookup_address(addr, &level);
arch/x86/include/asm/kfence.h
47
if (WARN_ON(!pte || level != PG_LEVEL_4K))
arch/x86/include/asm/kfence.h
50
val = pte_val(*pte);
arch/x86/include/asm/kfence.h
65
set_pte(pte, __pte(flip_protnone_guard(val, new, PTE_PFN_MASK)));
arch/x86/include/asm/paravirt.h
341
static inline pteval_t pte_val(pte_t pte)
arch/x86/include/asm/paravirt.h
343
return PVOP_ALT_CALLEE1(pteval_t, pv_ops, mmu.pte_val, pte.pte,
arch/x86/include/asm/paravirt.h
367
return (pte_t) { .pte = ret };
arch/x86/include/asm/paravirt.h
371
pte_t *ptep, pte_t old_pte, pte_t pte)
arch/x86/include/asm/paravirt.h
374
PVOP_VCALL4(pv_ops, mmu.ptep_modify_prot_commit, vma, addr, ptep, pte.pte);
arch/x86/include/asm/paravirt.h
377
static inline void set_pte(pte_t *ptep, pte_t pte)
arch/x86/include/asm/paravirt.h
379
PVOP_VCALL2(pv_ops, mmu.set_pte, ptep, pte.pte);
arch/x86/include/asm/paravirt.h
468
static inline void set_pte_atomic(pte_t *ptep, pte_t pte)
arch/x86/include/asm/paravirt.h
470
set_pte(ptep, pte);
arch/x86/include/asm/paravirt_types.h
153
pte_t *ptep, pte_t pte);
arch/x86/include/asm/pgalloc.h
54
extern void ___pte_free_tlb(struct mmu_gather *tlb, struct page *pte);
arch/x86/include/asm/pgalloc.h
56
static inline void __pte_free_tlb(struct mmu_gather *tlb, struct page *pte,
arch/x86/include/asm/pgalloc.h
59
___pte_free_tlb(tlb, pte);
arch/x86/include/asm/pgalloc.h
63
pmd_t *pmd, pte_t *pte)
arch/x86/include/asm/pgalloc.h
65
paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT);
arch/x86/include/asm/pgalloc.h
66
set_pmd(pmd, __pmd(__pa(pte) | _PAGE_TABLE));
arch/x86/include/asm/pgalloc.h
70
pmd_t *pmd, pte_t *pte)
arch/x86/include/asm/pgalloc.h
72
paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT);
arch/x86/include/asm/pgalloc.h
73
set_pmd_safe(pmd, __pmd(__pa(pte) | _PAGE_TABLE));
arch/x86/include/asm/pgalloc.h
77
struct page *pte)
arch/x86/include/asm/pgalloc.h
79
unsigned long pfn = page_to_pfn(pte);
arch/x86/include/asm/pgtable-2level.h
108
#define __pte_to_swp_entry(pte) ((swp_entry_t) { (pte).pte_low })
arch/x86/include/asm/pgtable-2level.h
109
#define __swp_entry_to_pte(x) ((pte_t) { .pte = (x).val })
arch/x86/include/asm/pgtable-2level.h
15
static inline void native_set_pte(pte_t *ptep , pte_t pte)
arch/x86/include/asm/pgtable-2level.h
17
*ptep = pte;
arch/x86/include/asm/pgtable-2level.h
29
static inline void native_set_pte_atomic(pte_t *ptep, pte_t pte)
arch/x86/include/asm/pgtable-2level.h
31
native_set_pte(ptep, pte);
arch/x86/include/asm/pgtable-2level_types.h
16
pteval_t pte;
arch/x86/include/asm/pgtable-3level.h
105
return pxx_xchg64(pte, ptep, 0ULL);
arch/x86/include/asm/pgtable-3level.h
190
#define __swp_entry_to_pte(x) ((pte_t){ .pte = \
arch/x86/include/asm/pgtable-3level.h
198
#define __pteval_swp_type(x) ((unsigned long)((x).pte >> (64 - SWP_TYPE_BITS)))
arch/x86/include/asm/pgtable-3level.h
199
#define __pteval_swp_offset(x) ((unsigned long)(~((x).pte) << SWP_TYPE_BITS >> SWP_OFFSET_SHIFT))
arch/x86/include/asm/pgtable-3level.h
201
#define __pte_to_swp_entry(pte) (__swp_entry(__pteval_swp_type(pte), \
arch/x86/include/asm/pgtable-3level.h
202
__pteval_swp_offset(pte)))
arch/x86/include/asm/pgtable-3level.h
36
static inline void native_set_pte(pte_t *ptep, pte_t pte)
arch/x86/include/asm/pgtable-3level.h
38
WRITE_ONCE(ptep->pte_high, pte.pte_high);
arch/x86/include/asm/pgtable-3level.h
40
WRITE_ONCE(ptep->pte_low, pte.pte_low);
arch/x86/include/asm/pgtable-3level.h
43
static inline void native_set_pte_atomic(pte_t *ptep, pte_t pte)
arch/x86/include/asm/pgtable-3level.h
45
pxx_xchg64(pte, ptep, native_pte_val(pte));
arch/x86/include/asm/pgtable-3level_types.h
19
pteval_t pte;
arch/x86/include/asm/pgtable.h
1001
static inline int pte_protnone(pte_t pte)
arch/x86/include/asm/pgtable.h
1003
return (pte_flags(pte) & (_PAGE_PROTNONE | _PAGE_PRESENT))
arch/x86/include/asm/pgtable.h
1254
pte_t pte = native_ptep_get_and_clear(ptep);
arch/x86/include/asm/pgtable.h
1255
page_table_check_pte_clear(mm, addr, pte);
arch/x86/include/asm/pgtable.h
1256
return pte;
arch/x86/include/asm/pgtable.h
1264
pte_t pte;
arch/x86/include/asm/pgtable.h
1270
pte = native_local_ptep_get_and_clear(ptep);
arch/x86/include/asm/pgtable.h
1271
page_table_check_pte_clear(mm, addr, pte);
arch/x86/include/asm/pgtable.h
1273
pte = ptep_get_and_clear(mm, addr, ptep);
arch/x86/include/asm/pgtable.h
1275
return pte;
arch/x86/include/asm/pgtable.h
1292
} while (!try_cmpxchg((long *)&ptep->pte, (long *)&old_pte, *(long *)&new_pte));
arch/x86/include/asm/pgtable.h
1513
static inline pte_t pte_swp_mkexclusive(pte_t pte)
arch/x86/include/asm/pgtable.h
1515
return pte_set_flags(pte, _PAGE_SWP_EXCLUSIVE);
arch/x86/include/asm/pgtable.h
1518
static inline bool pte_swp_exclusive(pte_t pte)
arch/x86/include/asm/pgtable.h
1520
return pte_flags(pte) & _PAGE_SWP_EXCLUSIVE;
arch/x86/include/asm/pgtable.h
1523
static inline pte_t pte_swp_clear_exclusive(pte_t pte)
arch/x86/include/asm/pgtable.h
1525
return pte_clear_flags(pte, _PAGE_SWP_EXCLUSIVE);
arch/x86/include/asm/pgtable.h
1529
static inline pte_t pte_swp_mksoft_dirty(pte_t pte)
arch/x86/include/asm/pgtable.h
1531
return pte_set_flags(pte, _PAGE_SWP_SOFT_DIRTY);
arch/x86/include/asm/pgtable.h
1534
static inline int pte_swp_soft_dirty(pte_t pte)
arch/x86/include/asm/pgtable.h
1536
return pte_flags(pte) & _PAGE_SWP_SOFT_DIRTY;
arch/x86/include/asm/pgtable.h
1539
static inline pte_t pte_swp_clear_soft_dirty(pte_t pte)
arch/x86/include/asm/pgtable.h
1541
return pte_clear_flags(pte, _PAGE_SWP_SOFT_DIRTY);
arch/x86/include/asm/pgtable.h
156
static inline bool pte_dirty(pte_t pte)
arch/x86/include/asm/pgtable.h
1563
static inline pte_t pte_swp_mkuffd_wp(pte_t pte)
arch/x86/include/asm/pgtable.h
1565
return pte_set_flags(pte, _PAGE_SWP_UFFD_WP);
arch/x86/include/asm/pgtable.h
1568
static inline int pte_swp_uffd_wp(pte_t pte)
arch/x86/include/asm/pgtable.h
1570
return pte_flags(pte) & _PAGE_SWP_UFFD_WP;
arch/x86/include/asm/pgtable.h
1573
static inline pte_t pte_swp_clear_uffd_wp(pte_t pte)
arch/x86/include/asm/pgtable.h
1575
return pte_clear_flags(pte, _PAGE_SWP_UFFD_WP);
arch/x86/include/asm/pgtable.h
158
return pte_flags(pte) & _PAGE_DIRTY_BITS;
arch/x86/include/asm/pgtable.h
161
static inline bool pte_shstk(pte_t pte)
arch/x86/include/asm/pgtable.h
164
(pte_flags(pte) & (_PAGE_RW | _PAGE_DIRTY)) == _PAGE_DIRTY;
arch/x86/include/asm/pgtable.h
1640
static inline bool pte_access_permitted(pte_t pte, bool write)
arch/x86/include/asm/pgtable.h
1642
return __pte_access_permitted(pte_val(pte), write);
arch/x86/include/asm/pgtable.h
1666
void arch_check_zapped_pte(struct vm_area_struct *vma, pte_t pte);
arch/x86/include/asm/pgtable.h
167
static inline int pte_young(pte_t pte)
arch/x86/include/asm/pgtable.h
1683
static inline bool pte_user_accessible_page(pte_t pte, unsigned long addr)
arch/x86/include/asm/pgtable.h
1685
return (pte_val(pte) & _PAGE_PRESENT) && (pte_val(pte) & _PAGE_USER);
arch/x86/include/asm/pgtable.h
169
return pte_flags(pte) & _PAGE_ACCESSED;
arch/x86/include/asm/pgtable.h
1715
#define set_pte_safe(ptep, pte) \
arch/x86/include/asm/pgtable.h
1717
WARN_ON_ONCE(pte_present(*ptep) && !pte_same(*ptep, pte)); \
arch/x86/include/asm/pgtable.h
1718
set_pte(ptep, pte); \
arch/x86/include/asm/pgtable.h
172
static inline bool pte_decrypted(pte_t pte)
arch/x86/include/asm/pgtable.h
174
return cc_mkdec(pte_val(pte)) == pte_val(pte);
arch/x86/include/asm/pgtable.h
213
static inline int pte_write(pte_t pte)
arch/x86/include/asm/pgtable.h
219
return (pte_flags(pte) & _PAGE_RW) || pte_shstk(pte);
arch/x86/include/asm/pgtable.h
238
static inline int pte_huge(pte_t pte)
arch/x86/include/asm/pgtable.h
240
return pte_flags(pte) & _PAGE_PSE;
arch/x86/include/asm/pgtable.h
243
static inline int pte_global(pte_t pte)
arch/x86/include/asm/pgtable.h
245
return pte_flags(pte) & _PAGE_GLOBAL;
arch/x86/include/asm/pgtable.h
248
static inline int pte_exec(pte_t pte)
arch/x86/include/asm/pgtable.h
250
return !(pte_flags(pte) & _PAGE_NX);
arch/x86/include/asm/pgtable.h
253
static inline int pte_special(pte_t pte)
arch/x86/include/asm/pgtable.h
255
return pte_flags(pte) & _PAGE_SPECIAL;
arch/x86/include/asm/pgtable.h
264
static inline unsigned long pte_pfn(pte_t pte)
arch/x86/include/asm/pgtable.h
266
phys_addr_t pfn = pte_val(pte);
arch/x86/include/asm/pgtable.h
296
#define pte_page(pte) pfn_to_page(pte_pfn(pte))
arch/x86/include/asm/pgtable.h
299
static inline bool pmd_leaf(pmd_t pte)
arch/x86/include/asm/pgtable.h
301
return pmd_flags(pte) & _PAGE_PSE;
arch/x86/include/asm/pgtable.h
348
static inline pte_t pte_set_flags(pte_t pte, pteval_t set)
arch/x86/include/asm/pgtable.h
350
pteval_t v = native_pte_val(pte);
arch/x86/include/asm/pgtable.h
355
static inline pte_t pte_clear_flags(pte_t pte, pteval_t clear)
arch/x86/include/asm/pgtable.h
357
pteval_t v = native_pte_val(pte);
arch/x86/include/asm/pgtable.h
393
static inline pte_t pte_mksaveddirty(pte_t pte)
arch/x86/include/asm/pgtable.h
395
pteval_t v = native_pte_val(pte);
arch/x86/include/asm/pgtable.h
401
static inline pte_t pte_clear_saveddirty(pte_t pte)
arch/x86/include/asm/pgtable.h
403
pteval_t v = native_pte_val(pte);
arch/x86/include/asm/pgtable.h
409
static inline pte_t pte_wrprotect(pte_t pte)
arch/x86/include/asm/pgtable.h
411
pte = pte_clear_flags(pte, _PAGE_RW);
arch/x86/include/asm/pgtable.h
418
return pte_mksaveddirty(pte);
arch/x86/include/asm/pgtable.h
422
static inline int pte_uffd_wp(pte_t pte)
arch/x86/include/asm/pgtable.h
424
return pte_flags(pte) & _PAGE_UFFD_WP;
arch/x86/include/asm/pgtable.h
427
static inline pte_t pte_mkuffd_wp(pte_t pte)
arch/x86/include/asm/pgtable.h
429
return pte_wrprotect(pte_set_flags(pte, _PAGE_UFFD_WP));
arch/x86/include/asm/pgtable.h
432
static inline pte_t pte_clear_uffd_wp(pte_t pte)
arch/x86/include/asm/pgtable.h
434
return pte_clear_flags(pte, _PAGE_UFFD_WP);
arch/x86/include/asm/pgtable.h
438
static inline pte_t pte_mkclean(pte_t pte)
arch/x86/include/asm/pgtable.h
440
return pte_clear_flags(pte, _PAGE_DIRTY_BITS);
arch/x86/include/asm/pgtable.h
443
static inline pte_t pte_mkold(pte_t pte)
arch/x86/include/asm/pgtable.h
445
return pte_clear_flags(pte, _PAGE_ACCESSED);
arch/x86/include/asm/pgtable.h
448
static inline pte_t pte_mkexec(pte_t pte)
arch/x86/include/asm/pgtable.h
450
return pte_clear_flags(pte, _PAGE_NX);
arch/x86/include/asm/pgtable.h
453
static inline pte_t pte_mkdirty(pte_t pte)
arch/x86/include/asm/pgtable.h
455
pte = pte_set_flags(pte, _PAGE_DIRTY | _PAGE_SOFT_DIRTY);
arch/x86/include/asm/pgtable.h
457
return pte_mksaveddirty(pte);
arch/x86/include/asm/pgtable.h
460
static inline pte_t pte_mkwrite_shstk(pte_t pte)
arch/x86/include/asm/pgtable.h
462
pte = pte_clear_flags(pte, _PAGE_RW);
arch/x86/include/asm/pgtable.h
464
return pte_set_flags(pte, _PAGE_DIRTY);
arch/x86/include/asm/pgtable.h
467
static inline pte_t pte_mkyoung(pte_t pte)
arch/x86/include/asm/pgtable.h
469
return pte_set_flags(pte, _PAGE_ACCESSED);
arch/x86/include/asm/pgtable.h
472
static inline pte_t pte_mkwrite_novma(pte_t pte)
arch/x86/include/asm/pgtable.h
474
return pte_set_flags(pte, _PAGE_RW);
arch/x86/include/asm/pgtable.h
478
pte_t pte_mkwrite(pte_t pte, struct vm_area_struct *vma);
arch/x86/include/asm/pgtable.h
481
static inline pte_t pte_mkhuge(pte_t pte)
arch/x86/include/asm/pgtable.h
483
return pte_set_flags(pte, _PAGE_PSE);
arch/x86/include/asm/pgtable.h
486
static inline pte_t pte_clrhuge(pte_t pte)
arch/x86/include/asm/pgtable.h
488
return pte_clear_flags(pte, _PAGE_PSE);
arch/x86/include/asm/pgtable.h
491
static inline pte_t pte_mkglobal(pte_t pte)
arch/x86/include/asm/pgtable.h
493
return pte_set_flags(pte, _PAGE_GLOBAL);
arch/x86/include/asm/pgtable.h
496
static inline pte_t pte_clrglobal(pte_t pte)
arch/x86/include/asm/pgtable.h
498
return pte_clear_flags(pte, _PAGE_GLOBAL);
arch/x86/include/asm/pgtable.h
501
static inline pte_t pte_mkspecial(pte_t pte)
arch/x86/include/asm/pgtable.h
503
return pte_set_flags(pte, _PAGE_SPECIAL);
arch/x86/include/asm/pgtable.h
660
static inline int pte_soft_dirty(pte_t pte)
arch/x86/include/asm/pgtable.h
662
return pte_flags(pte) & _PAGE_SOFT_DIRTY;
arch/x86/include/asm/pgtable.h
675
static inline pte_t pte_mksoft_dirty(pte_t pte)
arch/x86/include/asm/pgtable.h
677
return pte_set_flags(pte, _PAGE_SOFT_DIRTY);
arch/x86/include/asm/pgtable.h
68
#define set_pte(ptep, pte) native_set_pte(ptep, pte)
arch/x86/include/asm/pgtable.h
690
static inline pte_t pte_clear_soft_dirty(pte_t pte)
arch/x86/include/asm/pgtable.h
692
return pte_clear_flags(pte, _PAGE_SOFT_DIRTY);
arch/x86/include/asm/pgtable.h
70
#define set_pte_atomic(ptep, pte) \
arch/x86/include/asm/pgtable.h
71
native_set_pte_atomic(ptep, pte)
arch/x86/include/asm/pgtable.h
779
static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
arch/x86/include/asm/pgtable.h
781
pteval_t val = pte_val(pte), oldval = val;
arch/x86/include/asm/pgtable.h
948
static inline int pte_none(pte_t pte)
arch/x86/include/asm/pgtable.h
950
return !(pte.pte & ~(_PAGE_KNL_ERRATUM_MASK));
arch/x86/include/asm/pgtable.h
956
return a.pte == b.pte;
arch/x86/include/asm/pgtable.h
959
static inline pte_t pte_advance_pfn(pte_t pte, unsigned long nr)
arch/x86/include/asm/pgtable.h
961
if (__pte_needs_invert(pte_val(pte)))
arch/x86/include/asm/pgtable.h
962
return __pte(pte_val(pte) - (nr << PFN_PTE_SHIFT));
arch/x86/include/asm/pgtable.h
963
return __pte(pte_val(pte) + (nr << PFN_PTE_SHIFT));
arch/x86/include/asm/pgtable_64.h
235
#define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val((pte)) })
arch/x86/include/asm/pgtable_64.h
61
static inline void native_set_pte(pte_t *ptep, pte_t pte)
arch/x86/include/asm/pgtable_64.h
63
WRITE_ONCE(*ptep, pte);
arch/x86/include/asm/pgtable_64.h
72
static inline void native_set_pte_atomic(pte_t *ptep, pte_t pte)
arch/x86/include/asm/pgtable_64.h
74
native_set_pte(ptep, pte);
arch/x86/include/asm/pgtable_64.h
90
return native_make_pte(xchg(&xp->pte, 0));
arch/x86/include/asm/pgtable_64_types.h
21
typedef struct { pteval_t pte; } pte_t;
arch/x86/include/asm/pgtable_types.h
471
return (pte_t) { .pte = val };
arch/x86/include/asm/pgtable_types.h
474
static inline pteval_t native_pte_val(pte_t pte)
arch/x86/include/asm/pgtable_types.h
476
return pte.pte;
arch/x86/include/asm/pgtable_types.h
479
static inline pteval_t pte_flags(pte_t pte)
arch/x86/include/asm/pgtable_types.h
481
return native_pte_val(pte) & PTE_FLAGS_MASK;
arch/x86/include/asm/pgtable_types.h
536
void set_pte_vaddr(unsigned long vaddr, pte_t pte);
arch/x86/include/asm/xen/hypercall.h
313
return _hypercall3(int, update_va_mapping, va, new_val.pte, flags);
arch/x86/include/asm/xen/hypercall.h
337
mcl->args[1] = new_val.pte;
arch/x86/include/asm/xen/page.h
309
static inline unsigned long pte_mfn(pte_t pte)
arch/x86/include/asm/xen/page.h
311
return (pte.pte & XEN_PTE_MFN_MASK) >> PAGE_SHIFT;
arch/x86/include/asm/xen/page.h
316
pte_t pte;
arch/x86/include/asm/xen/page.h
318
pte.pte = ((phys_addr_t)page_nr << PAGE_SHIFT) |
arch/x86/include/asm/xen/page.h
321
return pte;
arch/x86/include/asm/xen/page.h
324
static inline pteval_t pte_val_ma(pte_t pte)
arch/x86/include/asm/xen/page.h
326
return pte.pte;
arch/x86/include/asm/xen/page.h
331
return (pte_t) { .pte = x };
arch/x86/kernel/alternative.c
2540
pte_t pte, *ptep;
arch/x86/kernel/alternative.c
2584
pte = mk_pte(pages[0], pgprot);
arch/x86/kernel/alternative.c
2585
set_pte_at(text_poke_mm, text_poke_mm_addr, ptep, pte);
arch/x86/kernel/alternative.c
2588
pte = mk_pte(pages[1], pgprot);
arch/x86/kernel/alternative.c
2589
set_pte_at(text_poke_mm, text_poke_mm_addr + PAGE_SIZE, ptep + 1, pte);
arch/x86/kernel/cpu/sgx/encl.c
1114
pte_t pte;
arch/x86/kernel/cpu/sgx/encl.c
1119
pte = pte_mkold(*ptep);
arch/x86/kernel/cpu/sgx/encl.c
1120
set_pte_at((struct mm_struct *)data, addr, ptep, pte);
arch/x86/kernel/espfix_64.c
131
pte_t pte, *pte_p;
arch/x86/kernel/espfix_64.c
192
pte = __pte(__pa(stack_page) | ((__PAGE_KERNEL_RO | _PAGE_ENC) & ptemask));
arch/x86/kernel/espfix_64.c
194
set_pte(&pte_p[n*PTE_STRIDE], pte);
arch/x86/kernel/head32.c
105
static __init __no_stack_protector pte_t init_map(pte_t pte, pte_t **ptep, pl2_t **pl2p,
arch/x86/kernel/head32.c
108
while ((pte.pte & PTE_PFN_MASK) < limit) {
arch/x86/kernel/head32.c
119
**ptep = pte;
arch/x86/kernel/head32.c
120
pte.pte += PAGE_SIZE;
arch/x86/kernel/head32.c
125
return pte;
arch/x86/kernel/head32.c
132
pte_t pte, *ptep = (pte_t *)__pa_nodebug(__brk_base);
arch/x86/kernel/head32.c
137
pte.pte = PTE_IDENT_ATTR;
arch/x86/kernel/head32.c
138
pte = init_map(pte, &ptep, &pl2p, limit);
arch/x86/kernel/head32.c
142
*ptr = (pte.pte & PTE_PFN_MASK) >> PAGE_SHIFT;
arch/x86/kernel/head32.c
154
*ptr = (pte.pte & PTE_PFN_MASK) + PAGE_OFFSET;
arch/x86/kernel/head32.c
162
pte.pte = PTE_IDENT_ATTR | PFN_ALIGN(limit);
arch/x86/kernel/head32.c
165
init_map(pte, &ptep, &pl2p, limit);
arch/x86/kernel/head32.c
42
*pl2p = (pte_t){ .pte = 0 };
arch/x86/kernel/head32.c
45
*(pl2p + ((PAGE_OFFSET >> PGDIR_SHIFT))) = (pte_t) {.pte = 0};
arch/x86/kernel/ldt.c
316
pte_t pte, *ptep;
arch/x86/kernel/ldt.c
337
pte = pfn_pte(pfn, pte_prot);
arch/x86/kernel/ldt.c
338
set_pte_at(mm, va, ptep, pte);
arch/x86/kernel/machine_kexec_32.c
80
pgd_t *pgd, pmd_t *pmd, pte_t *pte,
arch/x86/kernel/machine_kexec_32.c
95
set_pmd(pmd, __pmd(__pa(pte) | _PAGE_TABLE));
arch/x86/kernel/machine_kexec_32.c
96
pte = pte_offset_kernel(pmd, vaddr);
arch/x86/kernel/machine_kexec_32.c
97
set_pte(pte, pfn_pte(paddr >> PAGE_SHIFT, PAGE_KERNEL_EXEC));
arch/x86/kernel/machine_kexec_64.c
159
free_page((unsigned long)image->arch.pte);
arch/x86/kernel/machine_kexec_64.c
160
image->arch.pte = NULL;
arch/x86/kernel/machine_kexec_64.c
172
pte_t *pte;
arch/x86/kernel/machine_kexec_64.c
207
pte = (pte_t *)get_zeroed_page(GFP_KERNEL);
arch/x86/kernel/machine_kexec_64.c
208
if (!pte)
arch/x86/kernel/machine_kexec_64.c
210
image->arch.pte = pte;
arch/x86/kernel/machine_kexec_64.c
211
set_pmd(pmd, __pmd(__pa(pte) | _KERNPG_TABLE));
arch/x86/kernel/machine_kexec_64.c
213
pte = pte_offset_kernel(pmd, vaddr);
arch/x86/kernel/machine_kexec_64.c
218
set_pte(pte, pfn_pte(paddr >> PAGE_SHIFT, prot));
arch/x86/kernel/tboot.c
120
pte_t *pte;
arch/x86/kernel/tboot.c
132
pte = pte_alloc_map(&tboot_mm, pmd, vaddr);
arch/x86/kernel/tboot.c
133
if (!pte)
arch/x86/kernel/tboot.c
135
set_pte_at(&tboot_mm, vaddr, pte, pfn_pte(pfn, prot));
arch/x86/kernel/tboot.c
136
pte_unmap(pte);
arch/x86/kvm/mmu/mmu.c
2606
u64 pte;
arch/x86/kvm/mmu/mmu.c
2609
pte = *spte;
arch/x86/kvm/mmu/mmu.c
2610
if (is_shadow_present_pte(pte)) {
arch/x86/kvm/mmu/mmu.c
2611
if (is_last_spte(pte, sp->role.level)) {
arch/x86/kvm/mmu/mmu.c
2614
child = spte_to_child_sp(pte);
arch/x86/kvm/mmu/mmu.c
2628
} else if (is_mmio_spte(kvm, pte)) {
arch/x86/kvm/mmu/mmu.c
3058
u64 pte = *sptep;
arch/x86/kvm/mmu/mmu.c
3060
child = spte_to_child_sp(pte);
arch/x86/kvm/mmu/mmutrace.h
103
__entry->pte = pte;
arch/x86/kvm/mmu/mmutrace.h
107
TP_printk("pte %llx level %u", __entry->pte, __entry->level)
arch/x86/kvm/mmu/mmutrace.h
94
TP_PROTO(u64 pte, int level),
arch/x86/kvm/mmu/mmutrace.h
95
TP_ARGS(pte, level),
arch/x86/kvm/mmu/mmutrace.h
98
__field(__u64, pte)
arch/x86/kvm/mmu/paging_tmpl.h
127
static inline int FNAME(is_present_gpte)(unsigned long pte)
arch/x86/kvm/mmu/paging_tmpl.h
130
return pte & PT_PRESENT_MASK;
arch/x86/kvm/mmu/paging_tmpl.h
132
return pte & 7;
arch/x86/kvm/mmu/paging_tmpl.h
203
pt_element_t pte, orig_pte;
arch/x86/kvm/mmu/paging_tmpl.h
213
pte = orig_pte = walker->ptes[level - 1];
arch/x86/kvm/mmu/paging_tmpl.h
217
if (!(pte & PT_GUEST_ACCESSED_MASK)) {
arch/x86/kvm/mmu/paging_tmpl.h
218
trace_kvm_mmu_set_accessed_bit(table_gfn, index, sizeof(pte));
arch/x86/kvm/mmu/paging_tmpl.h
219
pte |= PT_GUEST_ACCESSED_MASK;
arch/x86/kvm/mmu/paging_tmpl.h
222
!(pte & PT_GUEST_DIRTY_MASK)) {
arch/x86/kvm/mmu/paging_tmpl.h
223
trace_kvm_mmu_set_dirty_bit(table_gfn, index, sizeof(pte));
arch/x86/kvm/mmu/paging_tmpl.h
228
pte |= PT_GUEST_DIRTY_MASK;
arch/x86/kvm/mmu/paging_tmpl.h
230
if (pte == orig_pte)
arch/x86/kvm/mmu/paging_tmpl.h
249
ret = __try_cmpxchg_user(ptep_user, &orig_pte, pte, fault);
arch/x86/kvm/mmu/paging_tmpl.h
254
walker->ptes[level - 1] = pte;
arch/x86/kvm/mmu/paging_tmpl.h
263
pte_t pte = {.pte = gpte};
arch/x86/kvm/mmu/paging_tmpl.h
265
pkeys = pte_flags_pkey(pte_flags(pte));
arch/x86/kvm/mmu/paging_tmpl.h
307
pt_element_t pte;
arch/x86/kvm/mmu/paging_tmpl.h
327
pte = kvm_mmu_get_guest_pgd(vcpu, mmu);
arch/x86/kvm/mmu/paging_tmpl.h
333
pte = mmu->get_pdptr(vcpu, (addr >> 30) & 3);
arch/x86/kvm/mmu/paging_tmpl.h
334
trace_kvm_mmu_paging_element(pte, walker->level);
arch/x86/kvm/mmu/paging_tmpl.h
335
if (!FNAME(is_present_gpte)(pte))
arch/x86/kvm/mmu/paging_tmpl.h
371
table_gfn = gpte_to_gfn(pte);
arch/x86/kvm/mmu/paging_tmpl.h
405
if (unlikely(get_user(pte, ptep_user)))
arch/x86/kvm/mmu/paging_tmpl.h
409
trace_kvm_mmu_paging_element(pte, walker->level);
arch/x86/kvm/mmu/paging_tmpl.h
415
pte_access = pt_access & (pte ^ walk_nx_mask);
arch/x86/kvm/mmu/paging_tmpl.h
417
if (unlikely(!FNAME(is_present_gpte)(pte)))
arch/x86/kvm/mmu/paging_tmpl.h
420
if (unlikely(FNAME(is_rsvd_bits_set)(mmu, pte, walker->level))) {
arch/x86/kvm/mmu/paging_tmpl.h
425
walker->ptes[walker->level - 1] = pte;
arch/x86/kvm/mmu/paging_tmpl.h
429
} while (!FNAME(is_last_gpte)(mmu, walker->level, pte));
arch/x86/kvm/mmu/paging_tmpl.h
431
pte_pkey = FNAME(gpte_pkeys)(vcpu, pte);
arch/x86/kvm/mmu/paging_tmpl.h
440
gfn = gpte_to_gfn_lvl(pte, walker->level);
arch/x86/kvm/mmu/paging_tmpl.h
445
gfn += pse36_gfn_delta(pte);
arch/x86/kvm/mmu/paging_tmpl.h
455
FNAME(protect_clean_gpte)(mmu, &walker->pte_access, pte);
arch/x86/kvm/mmu/paging_tmpl.h
462
accessed_dirty &= pte >>
arch/x86/kvm/mmu/paging_tmpl.h
74
#define gpte_to_gfn(pte) gpte_to_gfn_lvl((pte), PG_LEVEL_4K)
arch/x86/kvm/mmu/spte.h
309
static inline bool is_shadow_present_pte(u64 pte)
arch/x86/kvm/mmu/spte.h
311
return !!(pte & SPTE_MMU_PRESENT_MASK);
arch/x86/kvm/mmu/spte.h
348
static inline bool is_large_pte(u64 pte)
arch/x86/kvm/mmu/spte.h
350
return pte & PT_PAGE_SIZE_MASK;
arch/x86/kvm/mmu/spte.h
353
static inline bool is_last_spte(u64 pte, int level)
arch/x86/kvm/mmu/spte.h
355
return (level == PG_LEVEL_4K) || is_large_pte(pte);
arch/x86/kvm/mmu/spte.h
363
static inline kvm_pfn_t spte_to_pfn(u64 pte)
arch/x86/kvm/mmu/spte.h
365
return (pte & SPTE_BASE_ADDR_MASK) >> PAGE_SHIFT;
arch/x86/kvm/mmu/spte.h
373
static inline u64 get_rsvd_bits(struct rsvd_bits_validate *rsvd_check, u64 pte,
arch/x86/kvm/mmu/spte.h
376
int bit7 = (pte >> 7) & 1;
arch/x86/kvm/mmu/spte.h
382
u64 pte, int level)
arch/x86/kvm/mmu/spte.h
384
return pte & get_rsvd_bits(rsvd_check, pte, level);
arch/x86/kvm/mmu/spte.h
388
u64 pte)
arch/x86/kvm/mmu/spte.h
390
return rsvd_check->bad_mt_xwr & BIT_ULL(pte & 0x3f);
arch/x86/kvm/mmu/spte.h
466
static inline bool is_writable_pte(unsigned long pte)
arch/x86/kvm/mmu/spte.h
468
return pte & PT_WRITABLE_MASK;
arch/x86/kvm/mmu/tdp_iter.h
136
tdp_ptep_t spte_to_child_pt(u64 pte, int level);
arch/x86/mm/cpu_entry_area.c
82
pte_t pte = pfn_pte(pa >> PAGE_SHIFT, flags);
arch/x86/mm/cpu_entry_area.c
93
pte = pte_set_flags(pte, _PAGE_GLOBAL);
arch/x86/mm/cpu_entry_area.c
95
set_pte_vaddr(va, pte);
arch/x86/mm/dump_pagetables.c
269
static void effective_prot_pte(struct ptdump_state *st, pte_t pte)
arch/x86/mm/dump_pagetables.c
271
effective_prot(st, 4, pte_val(pte));
arch/x86/mm/dump_pagetables.c
391
static void note_page_pte(struct ptdump_state *pt_st, unsigned long addr, pte_t pte)
arch/x86/mm/dump_pagetables.c
393
note_page(pt_st, addr, 4, pte_val(pte));
arch/x86/mm/fault.c
1027
pte = pte_offset_kernel(pmd, address);
arch/x86/mm/fault.c
1028
if (!pte_present(*pte))
arch/x86/mm/fault.c
1031
ret = spurious_kernel_fault_check(error_code, pte);
arch/x86/mm/fault.c
301
pte_t *pte;
arch/x86/mm/fault.c
326
pte = pte_offset_kernel(pmd, address);
arch/x86/mm/fault.c
327
pr_cont("*pte = %0*Lx ", sizeof(*pte) * 2, (u64)pte_val(*pte));
arch/x86/mm/fault.c
357
pte_t *pte;
arch/x86/mm/fault.c
391
pte = pte_offset_kernel(pmd, address);
arch/x86/mm/fault.c
392
if (bad_address(pte))
arch/x86/mm/fault.c
395
pr_cont("PTE %lx", pte_val(*pte));
arch/x86/mm/fault.c
518
pte_t *pte;
arch/x86/mm/fault.c
523
pte = lookup_address_in_pgd_attr(pgd, address, &level, &nx, &rw);
arch/x86/mm/fault.c
525
if (pte && pte_present(*pte) && (!pte_exec(*pte) || nx))
arch/x86/mm/fault.c
528
if (pte && pte_present(*pte) && pte_exec(*pte) && !nx &&
arch/x86/mm/fault.c
947
static int spurious_kernel_fault_check(unsigned long error_code, pte_t *pte)
arch/x86/mm/fault.c
949
if ((error_code & X86_PF_WRITE) && !pte_write(*pte))
arch/x86/mm/fault.c
952
if ((error_code & X86_PF_INSTR) && !pte_exec(*pte))
arch/x86/mm/fault.c
986
pte_t *pte;
arch/x86/mm/ident_map.c
11
info->free_pgt_page(pte, info->context);
arch/x86/mm/ident_map.c
9
pte_t *pte = pte_offset_kernel(pmd, 0);
arch/x86/mm/init_32.c
155
static pte_t *__init page_table_kmap_check(pte_t *pte, pmd_t *pmd,
arch/x86/mm/init_32.c
178
set_pte(newpte + i, pte[i]);
arch/x86/mm/init_32.c
185
pte = newpte;
arch/x86/mm/init_32.c
189
&& lastpte && lastpte + PTRS_PER_PTE != pte);
arch/x86/mm/init_32.c
191
return pte;
arch/x86/mm/init_32.c
210
pte_t *pte = NULL;
arch/x86/mm/init_32.c
227
pte = page_table_kmap_check(one_page_table_init(pmd),
arch/x86/mm/init_32.c
228
pmd, vaddr, pte, &adr);
arch/x86/mm/init_32.c
262
pte_t *pte;
arch/x86/mm/init_32.c
340
pte = one_page_table_init(pmd);
arch/x86/mm/init_32.c
343
pte += pte_ofs;
arch/x86/mm/init_32.c
345
pte++, pfn++, pte_ofs++, addr += PAGE_SIZE) {
arch/x86/mm/init_32.c
358
set_pte(pte, pfn_pte(pfn, init_prot));
arch/x86/mm/init_32.c
361
set_pte(pte, pfn_pte(pfn, prot));
arch/x86/mm/init_32.c
425
pte_t *pte;
arch/x86/mm/init_32.c
455
pte = pte_offset_kernel(pmd, va);
arch/x86/mm/init_32.c
456
if (!pte_present(*pte))
arch/x86/mm/init_32.c
460
pfn, pmd, __pa(pmd), pte, __pa(pte));
arch/x86/mm/init_32.c
461
pte_clear(NULL, va, pte);
arch/x86/mm/init_64.c
1042
pte_t *pte;
arch/x86/mm/init_64.c
1046
pte = pte_start + i;
arch/x86/mm/init_64.c
1047
if (!pte_none(*pte))
arch/x86/mm/init_64.c
1099
pte_t *pte;
arch/x86/mm/init_64.c
1102
pte = pte_start + pte_index(addr);
arch/x86/mm/init_64.c
1103
for (; addr < end; addr = next, pte++) {
arch/x86/mm/init_64.c
1108
if (!pte_present(*pte))
arch/x86/mm/init_64.c
1116
phys_addr = pte_val(*pte) + (addr & PAGE_MASK);
arch/x86/mm/init_64.c
1121
free_pagetable(pte_page(*pte), 0);
arch/x86/mm/init_64.c
1124
pte_clear(&init_mm, addr, pte);
arch/x86/mm/init_64.c
1596
pte_t *pte = NULL;
arch/x86/mm/init_64.c
1630
pte = pte_offset_kernel(pmd, addr);
arch/x86/mm/init_64.c
1631
if (pte_none(*pte))
arch/x86/mm/init_64.c
1633
get_page_bootmem(section_nr, pte_page(*pte),
arch/x86/mm/init_64.c
306
pte_t *pte = (pte_t *) spp_getpage();
arch/x86/mm/init_64.c
307
pmd_populate_kernel(&init_mm, pmd, pte);
arch/x86/mm/init_64.c
308
if (pte != pte_offset_kernel(pmd, 0))
arch/x86/mm/init_64.c
317
pte_t *pte = fill_pte(pmd, vaddr);
arch/x86/mm/init_64.c
319
set_pte(pte, new_pte);
arch/x86/mm/init_64.c
477
pte_t *pte;
arch/x86/mm/init_64.c
480
pte = pte_page + pte_index(paddr);
arch/x86/mm/init_64.c
483
for (; i < PTRS_PER_PTE; i++, paddr = paddr_next, pte++) {
arch/x86/mm/init_64.c
491
set_pte_init(pte, __pte(0), init);
arch/x86/mm/init_64.c
501
if (!pte_none(*pte)) {
arch/x86/mm/init_64.c
508
set_pte_init(pte, pfn_pte(paddr >> PAGE_SHIFT, prot), init);
arch/x86/mm/init_64.c
533
pte_t *pte;
arch/x86/mm/init_64.c
550
pte = (pte_t *)pmd_page_vaddr(*pmd);
arch/x86/mm/init_64.c
551
paddr_last = phys_pte_init(pte, paddr,
arch/x86/mm/init_64.c
589
pte = alloc_low_page();
arch/x86/mm/init_64.c
590
paddr_last = phys_pte_init(pte, paddr, paddr_end, new_prot, init);
arch/x86/mm/init_64.c
593
pmd_populate_kernel_init(&init_mm, pmd, pte, init);
arch/x86/mm/init_64.c
76
DEFINE_POPULATE(pmd_populate_kernel, pmd, pte, init)
arch/x86/mm/init_64.c
91
DEFINE_ENTRY(pte, pte, init)
arch/x86/mm/ioremap.c
891
pte_t *pte;
arch/x86/mm/ioremap.c
897
pte = early_ioremap_pte(addr);
arch/x86/mm/ioremap.c
903
set_pte(pte, pfn_pte(phys >> PAGE_SHIFT, flags));
arch/x86/mm/ioremap.c
905
pte_clear(&init_mm, addr, pte);
arch/x86/mm/kasan_init_64.c
40
pte_t *pte;
arch/x86/mm/kasan_init_64.c
441
pte_t pte;
arch/x86/mm/kasan_init_64.c
447
pte = __pte(__pa(kasan_early_shadow_page) | pgprot_val(prot));
arch/x86/mm/kasan_init_64.c
448
set_pte(&kasan_early_shadow_pte[i], pte);
arch/x86/mm/kasan_init_64.c
58
pte = pte_offset_kernel(pmd, addr);
arch/x86/mm/kasan_init_64.c
63
if (!pte_none(*pte))
arch/x86/mm/kasan_init_64.c
68
set_pte_at(&init_mm, addr, pte, entry);
arch/x86/mm/kasan_init_64.c
69
} while (pte++, addr += PAGE_SIZE, addr != end);
arch/x86/mm/kmmio.c
120
pte_t *pte = lookup_address(addr, &l);
arch/x86/mm/kmmio.c
122
if (!pte)
arch/x86/mm/kmmio.c
147
static void clear_pte_presence(pte_t *pte, bool clear, pteval_t *old)
arch/x86/mm/kmmio.c
149
pteval_t v = pte_val(*pte);
arch/x86/mm/kmmio.c
153
pte_clear(&init_mm, 0, pte);
arch/x86/mm/kmmio.c
156
set_pte_atomic(pte, __pte(*old));
arch/x86/mm/kmmio.c
163
pte_t *pte = lookup_address(f->addr, &level);
arch/x86/mm/kmmio.c
165
if (!pte) {
arch/x86/mm/kmmio.c
172
clear_pmd_presence((pmd_t *)pte, clear, &f->old_presence);
arch/x86/mm/kmmio.c
175
clear_pte_presence(pte, clear, &f->old_presence);
arch/x86/mm/kmmio.c
243
pte_t *pte = lookup_address(addr, &l);
arch/x86/mm/kmmio.c
244
if (!pte)
arch/x86/mm/kmmio.c
444
pte_t *pte;
arch/x86/mm/kmmio.c
453
pte = lookup_address(addr, &l);
arch/x86/mm/kmmio.c
454
if (!pte) {
arch/x86/mm/kmmio.c
545
pte_t *pte;
arch/x86/mm/kmmio.c
547
pte = lookup_address(addr, &l);
arch/x86/mm/kmmio.c
548
if (!pte)
arch/x86/mm/kmmio.c
83
pte_t *pte = lookup_address(addr, &l);
arch/x86/mm/kmmio.c
85
if (!pte)
arch/x86/mm/mmio-mod.c
82
pte_t *pte = lookup_address(address, &level);
arch/x86/mm/mmio-mod.c
84
if (!pte) {
arch/x86/mm/mmio-mod.c
97
(unsigned long long)pte_val(*pte),
arch/x86/mm/mmio-mod.c
98
(unsigned long long)pte_val(*pte) & _PAGE_PRESENT);
arch/x86/mm/pat/cpa-test.c
121
pte_t *pte, pte0;
arch/x86/mm/pat/cpa-test.c
148
pte = NULL;
arch/x86/mm/pat/cpa-test.c
152
pte = lookup_address(addr[i] + k*PAGE_SIZE, &level);
arch/x86/mm/pat/cpa-test.c
153
if (!pte || pgprot_val(pte_pgprot(*pte)) == 0 ||
arch/x86/mm/pat/cpa-test.c
154
!(pte_val(*pte) & _PAGE_PRESENT)) {
arch/x86/mm/pat/cpa-test.c
159
pte0 = *pte;
arch/x86/mm/pat/cpa-test.c
161
if (pgprot_val(pte_pgprot(*pte)) !=
arch/x86/mm/pat/cpa-test.c
175
if (!addr[i] || !pte || !k) {
arch/x86/mm/pat/cpa-test.c
200
pte = lookup_address(addr[i], &level);
arch/x86/mm/pat/cpa-test.c
201
if (!pte || !pte_testbit(*pte) || pte_huge(*pte)) {
arch/x86/mm/pat/cpa-test.c
203
pte ? (u64)pte_val(*pte) : 0ULL);
arch/x86/mm/pat/cpa-test.c
220
pte = lookup_address(addr[i], &level);
arch/x86/mm/pat/cpa-test.c
221
if (!pte) {
arch/x86/mm/pat/cpa-test.c
231
pte = lookup_address(addr[i], &level);
arch/x86/mm/pat/cpa-test.c
232
if (!pte || pte_testbit(*pte)) {
arch/x86/mm/pat/cpa-test.c
234
addr[i], pte ? (u64)pte_val(*pte) : 0ULL);
arch/x86/mm/pat/cpa-test.c
39
static int pte_testbit(pte_t pte)
arch/x86/mm/pat/cpa-test.c
41
return pte_flags(pte) & _PAGE_SOFTW1;
arch/x86/mm/pat/cpa-test.c
60
pte_t *pte;
arch/x86/mm/pat/cpa-test.c
62
pte = lookup_address(addr, &level);
arch/x86/mm/pat/cpa-test.c
63
if (!pte) {
arch/x86/mm/pat/cpa-test.c
73
if ((pte_val(*pte) & _PAGE_PRESENT) && !(pte_val(*pte) & _PAGE_PSE)) {
arch/x86/mm/pat/cpa-test.c
76
addr, level, (u64)pte_val(*pte));
arch/x86/mm/pat/cpa-test.c
85
if (!(pte_val(*pte) & _PAGE_NX)) {
arch/x86/mm/pat/set_memory.c
1084
static void split_set_pte(struct cpa_data *cpa, pte_t *pte, unsigned long pfn,
arch/x86/mm/pat/set_memory.c
1117
set_pte(pte, pfn_pte(pfn, ref_prot));
arch/x86/mm/pat/set_memory.c
1249
pte_t *pte, first;
arch/x86/mm/pat/set_memory.c
1258
pte = pte_offset_kernel(pmd, addr);
arch/x86/mm/pat/set_memory.c
1259
first = *pte;
arch/x86/mm/pat/set_memory.c
1271
for (i = 1, pte++; i < PTRS_PER_PTE; i++, pte++) {
arch/x86/mm/pat/set_memory.c
1272
pte_t entry = *pte;
arch/x86/mm/pat/set_memory.c
1403
static bool try_to_free_pte_page(pte_t *pte)
arch/x86/mm/pat/set_memory.c
1408
if (!pte_none(pte[i]))
arch/x86/mm/pat/set_memory.c
1411
free_page((unsigned long)pte);
arch/x86/mm/pat/set_memory.c
1429
pte_t *pte = pte_offset_kernel(pmd, start);
arch/x86/mm/pat/set_memory.c
1432
set_pte(pte, __pte(0));
arch/x86/mm/pat/set_memory.c
1435
pte++;
arch/x86/mm/pat/set_memory.c
1542
pte_t *pte = (pte_t *)get_zeroed_page(GFP_KERNEL);
arch/x86/mm/pat/set_memory.c
1543
if (!pte)
arch/x86/mm/pat/set_memory.c
1546
set_pmd(pmd, __pmd(__pa(pte) | _KERNPG_TABLE));
arch/x86/mm/pat/set_memory.c
1564
pte_t *pte;
arch/x86/mm/pat/set_memory.c
1566
pte = pte_offset_kernel(pmd, start);
arch/x86/mm/pat/set_memory.c
1571
set_pte(pte, pfn_pte(cpa->pfn, pgprot));
arch/x86/mm/pat/set_memory.c
1575
pte++;
arch/x86/mm/pat/set_memory.c
2683
pte_t *pte;
arch/x86/mm/pat/set_memory.c
2688
pte = lookup_address((unsigned long)page_address(page), &level);
arch/x86/mm/pat/set_memory.c
2689
return (pte_val(*pte) & _PAGE_PRESENT);
arch/x86/mm/pat/set_memory.c
463
pte_t *pte = lookup_address(addr, &level);
arch/x86/mm/pat/set_memory.c
468
if (pte && (pte_val(*pte) & _PAGE_PRESENT))
arch/x86/mm/pat/set_memory.c
850
pte_t *pte;
arch/x86/mm/pat/set_memory.c
852
pte = lookup_address(virt_addr, &level);
arch/x86/mm/pat/set_memory.c
853
BUG_ON(!pte);
arch/x86/mm/pat/set_memory.c
862
phys_addr = (phys_addr_t)pud_pfn(*(pud_t *)pte) << PAGE_SHIFT;
arch/x86/mm/pat/set_memory.c
866
phys_addr = (phys_addr_t)pmd_pfn(*(pmd_t *)pte) << PAGE_SHIFT;
arch/x86/mm/pat/set_memory.c
870
phys_addr = (phys_addr_t)pte_pfn(*pte) << PAGE_SHIFT;
arch/x86/mm/pat/set_memory.c
881
static void __set_pmd_pte(pte_t *kpte, unsigned long address, pte_t pte)
arch/x86/mm/pat/set_memory.c
884
set_pte_atomic(kpte, pte);
arch/x86/mm/pat/set_memory.c
899
set_pte_atomic((pte_t *)pmd, pte);
arch/x86/mm/pgtable.c
21
void ___pte_free_tlb(struct mmu_gather *tlb, struct page *pte)
arch/x86/mm/pgtable.c
23
paravirt_release_pte(page_to_pfn(pte));
arch/x86/mm/pgtable.c
24
tlb_remove_ptdesc(tlb, page_ptdesc(pte));
arch/x86/mm/pgtable.c
453
(unsigned long *) &ptep->pte);
arch/x86/mm/pgtable.c
564
void __native_set_fixmap(enum fixed_addresses idx, pte_t pte)
arch/x86/mm/pgtable.c
581
set_pte_vaddr(address, pte);
arch/x86/mm/pgtable.c
802
pte_t pte_mkwrite(pte_t pte, struct vm_area_struct *vma)
arch/x86/mm/pgtable.c
805
return pte_mkwrite_shstk(pte);
arch/x86/mm/pgtable.c
807
pte = pte_mkwrite_novma(pte);
arch/x86/mm/pgtable.c
809
return pte_clear_saveddirty(pte);
arch/x86/mm/pgtable.c
822
void arch_check_zapped_pte(struct vm_area_struct *vma, pte_t pte)
arch/x86/mm/pgtable.c
832
pte_shstk(pte));
arch/x86/mm/pgtable_32.c
33
pte_t *pte;
arch/x86/mm/pgtable_32.c
55
pte = pte_offset_kernel(pmd, vaddr);
arch/x86/mm/pgtable_32.c
57
set_pte_at(&init_mm, vaddr, pte, pteval);
arch/x86/mm/pgtable_32.c
59
pte_clear(&init_mm, vaddr, pte);
arch/x86/mm/pti.c
255
pte_t *pte;
arch/x86/mm/pti.c
280
pte = pte_offset_kernel(pmd, address);
arch/x86/mm/pti.c
281
if (pte_flags(*pte) & _PAGE_USER) {
arch/x86/mm/pti.c
285
return pte;
arch/x86/mm/pti.c
291
pte_t *pte, *target_pte;
arch/x86/mm/pti.c
294
pte = lookup_address(VSYSCALL_ADDR, &level);
arch/x86/mm/pti.c
295
if (!pte || WARN_ON(level != PG_LEVEL_4K) || pte_none(*pte))
arch/x86/mm/pti.c
302
*target_pte = *pte;
arch/x86/mm/pti.c
325
pte_t *pte, *target_pte;
arch/x86/mm/pti.c
394
pte = pte_offset_kernel(pmd, addr);
arch/x86/mm/pti.c
395
if (pte_none(*pte)) {
arch/x86/mm/pti.c
401
if (WARN_ON(!(pte_flags(*pte) & _PAGE_PRESENT)))
arch/x86/mm/pti.c
411
*pte = pte_set_flags(*pte, _PAGE_GLOBAL);
arch/x86/mm/pti.c
414
*target_pte = *pte;
arch/x86/power/hibernate.c
160
pte_t *pte;
arch/x86/power/hibernate.c
186
pte = pte_offset_kernel(pmd, relocated_restore_code);
arch/x86/power/hibernate.c
187
set_pte(pte, __pte(pte_val(*pte) & ~_PAGE_NX));
arch/x86/power/hibernate_32.c
115
pte = resume_one_page_table_init(pmd);
arch/x86/power/hibernate_32.c
116
if (!pte)
arch/x86/power/hibernate_32.c
119
max_pte = pte + PTRS_PER_PTE;
arch/x86/power/hibernate_32.c
120
for (; pte < max_pte; pte++, pfn++) {
arch/x86/power/hibernate_32.c
124
set_pte(pte, pfn_pte(pfn, PAGE_KERNEL_EXEC));
arch/x86/power/hibernate_32.c
149
pte_t *pte;
arch/x86/power/hibernate_32.c
161
pte = resume_one_page_table_init(pmd);
arch/x86/power/hibernate_32.c
162
if (!pte)
arch/x86/power/hibernate_32.c
164
set_pte(pte + pte_index(restore_jump_address),
arch/x86/power/hibernate_32.c
86
pte_t *pte;
arch/x86/virt/svm/sev.c
844
pte_t *pte;
arch/x86/virt/svm/sev.c
848
pte = lookup_address_in_pgd(pgd, hva, &level);
arch/x86/virt/svm/sev.c
850
if (!pte) {
arch/x86/virt/svm/sev.c
855
paddr = PFN_PHYS(pte_pfn(*pte)) | (hva & ~page_level_mask(level));
arch/x86/virt/svm/sev.c
915
pte_t *pte;
arch/x86/virt/svm/sev.c
943
pte = lookup_address(vaddr, &level);
arch/x86/virt/svm/sev.c
944
if (!pte || pte_none(*pte))
arch/x86/xen/enlighten_pv.c
458
pte_t pte;
arch/x86/xen/enlighten_pv.c
467
pte = pfn_pte(pfn, prot);
arch/x86/xen/enlighten_pv.c
493
if (HYPERVISOR_update_va_mapping((unsigned long)v, pte, 0))
arch/x86/xen/enlighten_pv.c
498
if (va != v && HYPERVISOR_update_va_mapping((unsigned long)va, pte, 0))
arch/x86/xen/enlighten_pv.c
592
pte_t pte;
arch/x86/xen/enlighten_pv.c
601
pte = pfn_pte(pfn, PAGE_KERNEL_RO);
arch/x86/xen/enlighten_pv.c
603
if (HYPERVISOR_update_va_mapping((unsigned long)va, pte, 0))
arch/x86/xen/grant-table.c
94
static int gnttab_apply(pte_t *pte, unsigned long addr, void *data)
arch/x86/xen/grant-table.c
98
area->ptes[area->idx++] = pte;
arch/x86/xen/mmu.c
21
pte_t *pte;
arch/x86/xen/mmu.c
33
pte = lookup_address(address, &level);
arch/x86/xen/mmu.c
34
BUG_ON(pte == NULL);
arch/x86/xen/mmu.c
36
return XMADDR(((phys_addr_t)pte_mfn(*pte) << PAGE_SHIFT) + offset);
arch/x86/xen/mmu_pv.c
101
pte_t xen_make_pte_init(pteval_t pte);
arch/x86/xen/mmu_pv.c
1506
static void __init xen_set_pte_init(pte_t *ptep, pte_t pte)
arch/x86/xen/mmu_pv.c
1509
__xen_set_pte(ptep, pte);
arch/x86/xen/mmu_pv.c
1511
native_set_pte(ptep, pte);
arch/x86/xen/mmu_pv.c
1514
__visible pte_t xen_make_pte_init(pteval_t pte)
arch/x86/xen/mmu_pv.c
1524
pfn = (pte & PTE_PFN_MASK) >> PAGE_SHIFT;
arch/x86/xen/mmu_pv.c
1528
pte &= ~_PAGE_RW;
arch/x86/xen/mmu_pv.c
1530
pte = pte_pfn_to_mfn(pte);
arch/x86/xen/mmu_pv.c
1531
return native_make_pte(pte);
arch/x86/xen/mmu_pv.c
1704
pte_t pte = pfn_pte(pfn, prot);
arch/x86/xen/mmu_pv.c
1706
if (HYPERVISOR_update_va_mapping((unsigned long)addr, pte, flags))
arch/x86/xen/mmu_pv.c
1728
pte_t *pte = v;
arch/x86/xen/mmu_pv.c
1734
pte[i] = xen_make_pte(pte[i].pte);
arch/x86/xen/mmu_pv.c
191
pte_t *pte, ptev;
arch/x86/xen/mmu_pv.c
1915
pte_t pte;
arch/x86/xen/mmu_pv.c
1940
pte = native_make_pte(xen_read_phys_ulong(pa + pte_index(vaddr) *
arch/x86/xen/mmu_pv.c
1941
sizeof(pte)));
arch/x86/xen/mmu_pv.c
1942
if (!pte_present(pte))
arch/x86/xen/mmu_pv.c
1944
pa = pte_pfn(pte) << PAGE_SHIFT;
arch/x86/xen/mmu_pv.c
195
pte = lookup_address(address, &level);
arch/x86/xen/mmu_pv.c
196
if (pte == NULL)
arch/x86/xen/mmu_pv.c
199
ptev = pte_wrprotect(*pte);
arch/x86/xen/mmu_pv.c
207
pte_t *pte, ptev;
arch/x86/xen/mmu_pv.c
2092
pte_t pte;
arch/x86/xen/mmu_pv.c
2103
pte = pfn_pte(phys, prot);
arch/x86/xen/mmu_pv.c
2108
pte = pfn_pte(PFN_DOWN(__pa(dummy_mapping)), PAGE_KERNEL);
arch/x86/xen/mmu_pv.c
211
pte = lookup_address(address, &level);
arch/x86/xen/mmu_pv.c
2118
pte = pfn_pte(PFN_DOWN(__pa(dummy_mapping)), PAGE_KERNEL);
arch/x86/xen/mmu_pv.c
212
if (pte == NULL)
arch/x86/xen/mmu_pv.c
2125
pte = mfn_pte(phys, prot);
arch/x86/xen/mmu_pv.c
2130
pte = mfn_pte(phys, prot);
arch/x86/xen/mmu_pv.c
2135
if (HYPERVISOR_update_va_mapping(vaddr, pte, UVMF_INVLPG))
arch/x86/xen/mmu_pv.c
2142
set_pte_vaddr_pud(level3_user_vsyscall, vaddr, pte);
arch/x86/xen/mmu_pv.c
215
ptev = pte_mkwrite_novma(*pte);
arch/x86/xen/mmu_pv.c
2459
pte_t pte = pte_mkspecial(mfn_pte(*rmd->pfn, rmd->prot));
arch/x86/xen/mmu_pv.c
2474
rmd->mmu_update->val = pte_val_ma(pte);
arch/x86/xen/mmu_pv.c
364
pte_t *ptep, pte_t pte)
arch/x86/xen/mmu_pv.c
368
trace_xen_mmu_ptep_modify_prot_commit(vma->vm_mm, addr, ptep, pte);
arch/x86/xen/mmu_pv.c
372
u.val = pte_val_ma(pte);
arch/x86/xen/mmu_pv.c
421
__visible pteval_t xen_pte_val(pte_t pte)
arch/x86/xen/mmu_pv.c
423
pteval_t pteval = pte.pte;
arch/x86/xen/mmu_pv.c
435
__visible pte_t xen_make_pte(pteval_t pte)
arch/x86/xen/mmu_pv.c
437
pte = pte_pfn_to_mfn(pte);
arch/x86/xen/mmu_pv.c
439
return native_make_pte(pte);
arch/x86/xen/mmu_pv.c
91
pteval_t xen_pte_val(pte_t pte);
arch/x86/xen/mmu_pv.c
96
pte_t xen_make_pte(pteval_t pte);
arch/x86/xen/p2m.c
687
pte_t *pte;
arch/x86/xen/p2m.c
710
pte = (pte_t *)(mfn_to_virt(PFN_DOWN(map_ops[i].host_addr)) +
arch/x86/xen/p2m.c
712
mfn = pte_mfn(*pte);
arch/xtensa/include/asm/page.h
93
typedef struct { unsigned long pte; } pte_t; /* page table entry */
arch/xtensa/include/asm/page.h
98
#define pte_val(x) ((x).pte)
arch/xtensa/include/asm/pgtable.h
234
# define pte_none(pte) (pte_val(pte) == (_PAGE_CA_INVALID | _PAGE_USER))
arch/xtensa/include/asm/pgtable.h
236
# define pte_present(pte) ((pte_val(pte) & _PAGE_CA_MASK) != _PAGE_CA_INVALID)
arch/xtensa/include/asm/pgtable.h
238
# define pte_present(pte) \
arch/xtensa/include/asm/pgtable.h
239
(((pte_val(pte) & _PAGE_CA_MASK) != _PAGE_CA_INVALID) \
arch/xtensa/include/asm/pgtable.h
240
|| ((pte_val(pte) & _PAGE_ATTRIB_MASK) == _PAGE_NONE))
arch/xtensa/include/asm/pgtable.h
250
static inline int pte_write(pte_t pte) { return pte_val(pte) & _PAGE_WRITABLE; }
arch/xtensa/include/asm/pgtable.h
251
static inline int pte_dirty(pte_t pte) { return pte_val(pte) & _PAGE_DIRTY; }
arch/xtensa/include/asm/pgtable.h
252
static inline int pte_young(pte_t pte) { return pte_val(pte) & _PAGE_ACCESSED; }
arch/xtensa/include/asm/pgtable.h
254
static inline pte_t pte_wrprotect(pte_t pte)
arch/xtensa/include/asm/pgtable.h
255
{ pte_val(pte) &= ~(_PAGE_WRITABLE | _PAGE_HW_WRITE); return pte; }
arch/xtensa/include/asm/pgtable.h
256
static inline pte_t pte_mkclean(pte_t pte)
arch/xtensa/include/asm/pgtable.h
257
{ pte_val(pte) &= ~(_PAGE_DIRTY | _PAGE_HW_WRITE); return pte; }
arch/xtensa/include/asm/pgtable.h
258
static inline pte_t pte_mkold(pte_t pte)
arch/xtensa/include/asm/pgtable.h
259
{ pte_val(pte) &= ~_PAGE_ACCESSED; return pte; }
arch/xtensa/include/asm/pgtable.h
260
static inline pte_t pte_mkdirty(pte_t pte)
arch/xtensa/include/asm/pgtable.h
261
{ pte_val(pte) |= _PAGE_DIRTY; return pte; }
arch/xtensa/include/asm/pgtable.h
262
static inline pte_t pte_mkyoung(pte_t pte)
arch/xtensa/include/asm/pgtable.h
263
{ pte_val(pte) |= _PAGE_ACCESSED; return pte; }
arch/xtensa/include/asm/pgtable.h
264
static inline pte_t pte_mkwrite_novma(pte_t pte)
arch/xtensa/include/asm/pgtable.h
265
{ pte_val(pte) |= _PAGE_WRITABLE; return pte; }
arch/xtensa/include/asm/pgtable.h
272
#define pte_pfn(pte) (pte_val(pte) >> PAGE_SHIFT)
arch/xtensa/include/asm/pgtable.h
277
static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
arch/xtensa/include/asm/pgtable.h
279
return __pte((pte_val(pte) & _PAGE_CHG_MASK) | pgprot_val(newprot));
arch/xtensa/include/asm/pgtable.h
298
static inline void set_pte(pte_t *ptep, pte_t pte)
arch/xtensa/include/asm/pgtable.h
300
update_pte(ptep, pte);
arch/xtensa/include/asm/pgtable.h
315
pte_t pte = *ptep;
arch/xtensa/include/asm/pgtable.h
316
if (!pte_young(pte))
arch/xtensa/include/asm/pgtable.h
318
update_pte(ptep, pte_mkold(pte));
arch/xtensa/include/asm/pgtable.h
325
pte_t pte = *ptep;
arch/xtensa/include/asm/pgtable.h
327
return pte;
arch/xtensa/include/asm/pgtable.h
333
pte_t pte = *ptep;
arch/xtensa/include/asm/pgtable.h
334
update_pte(ptep, pte_wrprotect(pte));
arch/xtensa/include/asm/pgtable.h
348
#define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) })
arch/xtensa/include/asm/pgtable.h
351
static inline bool pte_swp_exclusive(pte_t pte)
arch/xtensa/include/asm/pgtable.h
353
return pte_val(pte) & _PAGE_SWP_EXCLUSIVE;
arch/xtensa/include/asm/pgtable.h
356
static inline pte_t pte_swp_mkexclusive(pte_t pte)
arch/xtensa/include/asm/pgtable.h
358
pte_val(pte) |= _PAGE_SWP_EXCLUSIVE;
arch/xtensa/include/asm/pgtable.h
359
return pte;
arch/xtensa/include/asm/pgtable.h
362
static inline pte_t pte_swp_clear_exclusive(pte_t pte)
arch/xtensa/include/asm/pgtable.h
364
pte_val(pte) &= ~_PAGE_SWP_EXCLUSIVE;
arch/xtensa/include/asm/pgtable.h
365
return pte;
arch/xtensa/include/asm/tlb.h
19
#define __pte_free_tlb(tlb, pte, address) pte_free((tlb)->mm, pte)
arch/xtensa/mm/kasan_init.c
42
pte_t *pte = memblock_alloc_or_panic(n_pages * sizeof(pte_t), PAGE_SIZE);
arch/xtensa/mm/kasan_init.c
58
set_pte(pte + j, pfn_pte(PHYS_PFN(phys), PAGE_KERNEL));
arch/xtensa/mm/kasan_init.c
62
for (i = 0; i < n_pmds ; ++i, pte += PTRS_PER_PTE)
arch/xtensa/mm/kasan_init.c
63
set_pmd(pmd + i, __pmd((unsigned long)pte));
arch/xtensa/mm/mmu.c
27
pte_t *pte;
arch/xtensa/mm/mmu.c
35
pte = memblock_alloc_low(n_pages * sizeof(pte_t), PAGE_SIZE);
arch/xtensa/mm/mmu.c
36
if (!pte)
arch/xtensa/mm/mmu.c
41
pte_clear(NULL, 0, pte + i);
arch/xtensa/mm/mmu.c
44
pte_t *cur_pte = pte + i;
arch/xtensa/mm/mmu.c
52
return pte;
arch/xtensa/mm/tlb.c
182
pte_t *pte;
arch/xtensa/mm/tlb.c
199
pte = pte_offset_map(pmd, vaddr);
arch/xtensa/mm/tlb.c
200
if (!pte)
arch/xtensa/mm/tlb.c
202
pteval = pte_val(*pte);
arch/xtensa/mm/tlb.c
203
pte_unmap(pte);
arch/xtensa/mm/tlb.c
240
unsigned pte = get_pte_for_vaddr(vpn);
arch/xtensa/mm/tlb.c
254
if ((pte ^ r1) & PAGE_MASK) {
arch/xtensa/mm/tlb.c
256
dtlb ? 'D' : 'I', w, e, r0, r1, pte);
arch/xtensa/mm/tlb.c
257
if (pte == 0 || !pte_present(__pte(pte))) {
block/partitions/efi.c
274
gpt_entry *pte;
block/partitions/efi.c
283
pte = kmalloc(count, GFP_KERNEL);
block/partitions/efi.c
284
if (!pte)
block/partitions/efi.c
288
(u8 *) pte, count) < count) {
block/partitions/efi.c
289
kfree(pte);
block/partitions/efi.c
290
pte=NULL;
block/partitions/efi.c
293
return pte;
block/partitions/efi.c
460
is_pte_valid(const gpt_entry *pte, const u64 lastlba)
block/partitions/efi.c
462
if ((!efi_guidcmp(pte->partition_type_guid, NULL_GUID)) ||
block/partitions/efi.c
463
le64_to_cpu(pte->starting_lba) > lastlba ||
block/partitions/efi.c
464
le64_to_cpu(pte->ending_lba) > lastlba)
drivers/accel/ivpu/ivpu_mmu_context.c
242
u64 *pte;
drivers/accel/ivpu/ivpu_mmu_context.c
263
pte = ivpu_mmu_ensure_pte(vdev, &ctx->pgtable, pgd_idx, pud_idx, pmd_idx);
drivers/accel/ivpu/ivpu_mmu_context.c
264
if (!pte)
drivers/accel/ivpu/ivpu_mmu_context.c
268
pte[pte_idx] = dma_addr | prot;
drivers/char/agp/amd64-agp.c
51
u32 pte;
drivers/char/agp/amd64-agp.c
87
pte = (tmp & 0x000000ff00000000ULL) >> 28;
drivers/char/agp/amd64-agp.c
88
pte |=(tmp & 0x00000000fffff000ULL);
drivers/char/agp/amd64-agp.c
89
pte |= GPTE_VALID | GPTE_COHERENT;
drivers/char/agp/amd64-agp.c
91
writel(pte, agp_bridge->gatt_table+j);
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_sdma.c
228
uint64_t *pte;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_sdma.c
276
pte = (uint64_t *)&(p->job->ibs->ptr[p->num_dw_left]);
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_sdma.c
278
pte[i] = amdgpu_vm_map_gart(p->pages_addr, addr);
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_sdma.c
279
pte[i] |= flags;
drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.c
142
const struct pte_setting *pte = &pte_settings[mi_tiling][mi_bpp];
drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.c
144
unsigned int page_width = log_2(pte->page_width);
drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.c
145
unsigned int page_height = log_2(pte->page_height);
drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.c
147
pte->min_pte_before_flip_vert_scan :
drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.c
148
pte->min_pte_before_flip_horiz_scan;
drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.c
159
DVMM_PTE_REQ_PER_CHUNK, pte->pte_req_per_chunk,
drivers/gpu/drm/amd/display/dc/dce110/dce110_mem_input_v.c
570
const unsigned int *pte = get_dvmm_hw_setting(tiling_info, format, false);
drivers/gpu/drm/amd/display/dc/dce110/dce110_mem_input_v.c
577
unsigned int temp_page_width = pte[1];
drivers/gpu/drm/amd/display/dc/dce110/dce110_mem_input_v.c
578
unsigned int temp_page_height = pte[2];
drivers/gpu/drm/amd/display/dc/dce110/dce110_mem_input_v.c
598
min_pte_before_flip = pte[4];
drivers/gpu/drm/amd/display/dc/dce110/dce110_mem_input_v.c
602
min_pte_before_flip = pte[3];
drivers/gpu/drm/amd/display/dc/dce110/dce110_mem_input_v.c
620
set_reg_field_value(value, pte[5], UNP_DVMM_PTE_ARB_CONTROL, DVMM_PTE_REQ_PER_CHUNK);
drivers/gpu/drm/gma500/gem.c
315
uint32_t pte;
drivers/gpu/drm/gma500/gem.c
324
pte = psb_gtt_mask_pte(pfn_base + i, PSB_MMU_CACHED_MEMORY);
drivers/gpu/drm/gma500/gem.c
325
iowrite32(pte, pdev->gtt_map + i);
drivers/gpu/drm/gma500/gtt.c
108
u32 pte;
drivers/gpu/drm/gma500/gtt.c
114
pte = psb_gtt_mask_pte(page_to_pfn(pdev->scratch_page), PSB_MMU_CACHED_MEMORY);
drivers/gpu/drm/gma500/gtt.c
120
iowrite32(pte, gtt_slot);
drivers/gpu/drm/gma500/gtt.c
174
uint32_t pte;
drivers/gpu/drm/gma500/gtt.c
177
pte = psb_gtt_mask_pte(pfn_base, PSB_MMU_CACHED_MEMORY);
drivers/gpu/drm/gma500/gtt.c
180
iowrite32(pte, pdev->gtt_map + i);
drivers/gpu/drm/gma500/gtt.c
83
u32 pte;
drivers/gpu/drm/gma500/gtt.c
93
pte = psb_gtt_mask_pte(page_to_pfn(pages[i]), PSB_MMU_CACHED_MEMORY);
drivers/gpu/drm/gma500/gtt.c
94
iowrite32(pte, gtt_slot);
drivers/gpu/drm/gma500/mmu.c
388
uint32_t pte)
drivers/gpu/drm/gma500/mmu.c
390
pt->v[psb_mmu_pt_index(addr)] = pte;
drivers/gpu/drm/gma500/mmu.c
626
uint32_t pte;
drivers/gpu/drm/gma500/mmu.c
646
pte = psb_mmu_mask_pte(start_pfn++, type);
drivers/gpu/drm/gma500/mmu.c
647
psb_mmu_set_pte(pt, addr, pte);
drivers/gpu/drm/gma500/mmu.c
675
uint32_t pte;
drivers/gpu/drm/gma500/mmu.c
708
pte = psb_mmu_mask_pte(page_to_pfn(*pages++),
drivers/gpu/drm/gma500/mmu.c
710
psb_mmu_set_pte(pt, addr, pte);
drivers/gpu/drm/i915/display/intel_dpt.c
38
static void gen8_set_pte(void __iomem *addr, gen8_pte_t pte)
drivers/gpu/drm/i915/display/intel_dpt.c
40
writeq(pte, addr);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
1680
static int check_present_pte(pte_t *pte, unsigned long addr, void *data)
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
1682
pte_t ptent = ptep_get(pte);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
1693
static int check_absent_pte(pte_t *pte, unsigned long addr, void *data)
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
1695
pte_t ptent = ptep_get(pte);
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
104
memset32(vaddr + pte, scratch_pte, count);
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
106
pte = 0;
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
81
unsigned int pte = first_entry % GEN6_PTES;
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
87
const unsigned int count = min(num_entries, GEN6_PTES - pte);
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
272
unsigned int pte = gen8_pd_index(start, 0);
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
286
GEM_BUG_ON(pte % 16);
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
288
pte /= 16;
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
292
memset64(vaddr + pte,
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
36
gen8_pte_t pte = addr | GEN8_PAGE_PRESENT | GEN8_PAGE_RW;
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
39
pte &= ~GEN8_PAGE_RW;
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
48
pte |= PPAT_UNCACHED;
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
51
pte |= PPAT_DISPLAY_ELLC;
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
54
pte |= PPAT_CACHED;
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
58
return pte;
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
65
gen8_pte_t pte = addr | GEN8_PAGE_PRESENT | GEN8_PAGE_RW;
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
68
pte &= ~GEN8_PAGE_RW;
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
71
pte |= GEN12_PPGTT_PTE_LM;
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
74
pte |= GEN12_PPGTT_PTE_PAT0;
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
77
pte |= GEN12_PPGTT_PTE_PAT1;
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
80
pte |= GEN12_PPGTT_PTE_PAT2;
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
83
pte |= MTL_PPGTT_PTE_PAT3;
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
85
return pte;
drivers/gpu/drm/i915/gt/intel_ggtt.c
1352
gen6_pte_t pte = GEN6_PTE_ADDR_ENCODE(addr) | GEN6_PTE_VALID;
drivers/gpu/drm/i915/gt/intel_ggtt.c
1357
pte |= GEN6_PTE_CACHE_LLC;
drivers/gpu/drm/i915/gt/intel_ggtt.c
1360
pte |= GEN6_PTE_UNCACHED;
drivers/gpu/drm/i915/gt/intel_ggtt.c
1366
return pte;
drivers/gpu/drm/i915/gt/intel_ggtt.c
1373
gen6_pte_t pte = GEN6_PTE_ADDR_ENCODE(addr) | GEN6_PTE_VALID;
drivers/gpu/drm/i915/gt/intel_ggtt.c
1377
pte |= GEN7_PTE_CACHE_L3_LLC;
drivers/gpu/drm/i915/gt/intel_ggtt.c
1380
pte |= GEN6_PTE_CACHE_LLC;
drivers/gpu/drm/i915/gt/intel_ggtt.c
1383
pte |= GEN6_PTE_UNCACHED;
drivers/gpu/drm/i915/gt/intel_ggtt.c
1389
return pte;
drivers/gpu/drm/i915/gt/intel_ggtt.c
1396
gen6_pte_t pte = GEN6_PTE_ADDR_ENCODE(addr) | GEN6_PTE_VALID;
drivers/gpu/drm/i915/gt/intel_ggtt.c
1399
pte |= BYT_PTE_WRITEABLE;
drivers/gpu/drm/i915/gt/intel_ggtt.c
1402
pte |= BYT_PTE_SNOOPED_BY_CPU_CACHES;
drivers/gpu/drm/i915/gt/intel_ggtt.c
1404
return pte;
drivers/gpu/drm/i915/gt/intel_ggtt.c
1411
gen6_pte_t pte = HSW_PTE_ADDR_ENCODE(addr) | GEN6_PTE_VALID;
drivers/gpu/drm/i915/gt/intel_ggtt.c
1414
pte |= HSW_WB_LLC_AGE3;
drivers/gpu/drm/i915/gt/intel_ggtt.c
1416
return pte;
drivers/gpu/drm/i915/gt/intel_ggtt.c
1423
gen6_pte_t pte = HSW_PTE_ADDR_ENCODE(addr) | GEN6_PTE_VALID;
drivers/gpu/drm/i915/gt/intel_ggtt.c
1429
pte |= HSW_WT_ELLC_LLC_AGE3;
drivers/gpu/drm/i915/gt/intel_ggtt.c
1432
pte |= HSW_WB_ELLC_LLC_AGE3;
drivers/gpu/drm/i915/gt/intel_ggtt.c
1436
return pte;
drivers/gpu/drm/i915/gt/intel_ggtt.c
1439
static dma_addr_t gen6_pte_decode(u64 pte, bool *is_present, bool *is_local)
drivers/gpu/drm/i915/gt/intel_ggtt.c
1441
*is_present = pte & GEN6_PTE_VALID;
drivers/gpu/drm/i915/gt/intel_ggtt.c
1444
return ((pte & 0xff0) << 28) | (pte & ~0xfff);
drivers/gpu/drm/i915/gt/intel_ggtt.c
265
gen8_pte_t pte = addr | GEN8_PAGE_PRESENT;
drivers/gpu/drm/i915/gt/intel_ggtt.c
270
pte |= GEN12_GGTT_PTE_LM;
drivers/gpu/drm/i915/gt/intel_ggtt.c
273
pte |= MTL_GGTT_PTE_PAT0;
drivers/gpu/drm/i915/gt/intel_ggtt.c
276
pte |= MTL_GGTT_PTE_PAT1;
drivers/gpu/drm/i915/gt/intel_ggtt.c
278
return pte;
drivers/gpu/drm/i915/gt/intel_ggtt.c
285
gen8_pte_t pte = addr | GEN8_PAGE_PRESENT;
drivers/gpu/drm/i915/gt/intel_ggtt.c
288
pte |= GEN12_GGTT_PTE_LM;
drivers/gpu/drm/i915/gt/intel_ggtt.c
290
return pte;
drivers/gpu/drm/i915/gt/intel_ggtt.c
293
static dma_addr_t gen8_ggtt_pte_decode(u64 pte, bool *is_present, bool *is_local)
drivers/gpu/drm/i915/gt/intel_ggtt.c
295
*is_present = pte & GEN8_PAGE_PRESENT;
drivers/gpu/drm/i915/gt/intel_ggtt.c
296
*is_local = pte & GEN12_GGTT_PTE_LM;
drivers/gpu/drm/i915/gt/intel_ggtt.c
298
return pte & GEN12_GGTT_PTE_ADDR_MASK;
drivers/gpu/drm/i915/gt/intel_ggtt.c
342
const gen8_pte_t pte)
drivers/gpu/drm/i915/gt/intel_ggtt.c
399
*cs++ = lower_32_bits(pte | addr);
drivers/gpu/drm/i915/gt/intel_ggtt.c
400
*cs++ = upper_32_bits(pte | addr);
drivers/gpu/drm/i915/gt/intel_ggtt.c
410
memset64((u64 *)cs, pte, n_ptes);
drivers/gpu/drm/i915/gt/intel_ggtt.c
442
static void gen8_set_pte(void __iomem *addr, gen8_pte_t pte)
drivers/gpu/drm/i915/gt/intel_ggtt.c
444
writeq(pte, addr);
drivers/gpu/drm/i915/gt/intel_ggtt.c
459
gen8_pte_t __iomem *pte =
drivers/gpu/drm/i915/gt/intel_ggtt.c
462
gen8_set_pte(pte, ggtt->vm.pte_encode(addr, pat_index, flags));
drivers/gpu/drm/i915/gt/intel_ggtt.c
471
gen8_pte_t __iomem *pte =
drivers/gpu/drm/i915/gt/intel_ggtt.c
474
return ggtt->vm.pte_decode(gen8_get_pte(pte), is_present, is_local);
drivers/gpu/drm/i915/gt/intel_ggtt.c
482
gen8_pte_t pte;
drivers/gpu/drm/i915/gt/intel_ggtt.c
484
pte = ggtt->vm.pte_encode(addr, pat_index, flags);
drivers/gpu/drm/i915/gt/intel_ggtt.c
486
gen8_ggtt_bind_ptes(ggtt, offset, NULL, 1, pte))
drivers/gpu/drm/i915/gt/intel_ggtt.c
624
gen6_pte_t __iomem *pte =
drivers/gpu/drm/i915/gt/intel_ggtt.c
627
iowrite32(vm->pte_encode(addr, pat_index, flags), pte);
drivers/gpu/drm/i915/gt/intel_ggtt.c
637
gen6_pte_t __iomem *pte =
drivers/gpu/drm/i915/gt/intel_ggtt.c
640
return vm->pte_decode(ioread32(pte), is_present, is_local);
drivers/gpu/drm/i915/gt/intel_gtt.h
315
dma_addr_t (*pte_decode)(u64 pte, bool *is_present, bool *is_local);
drivers/gpu/drm/i915/gvt/dmabuf.c
45
#define GEN8_DECODE_PTE(pte) (pte & GENMASK_ULL(63, 12))
drivers/gpu/drm/i915/gvt/gtt.c
233
static void write_pte64(struct i915_ggtt *ggtt, unsigned long index, u64 pte)
drivers/gpu/drm/i915/gvt/gtt.c
237
writeq(pte, addr);
drivers/gpu/drm/i915/gvt/gtt.c
2773
gen8_pte_t pte;
drivers/gpu/drm/i915/gvt/gtt.c
2783
pte = mm->ggtt_mm.host_ggtt_aperture[idx];
drivers/gpu/drm/i915/gvt/gtt.c
2784
if (pte & GEN8_PAGE_PRESENT)
drivers/gpu/drm/i915/gvt/gtt.c
2785
write_pte64(vgpu->gvt->gt->ggtt, offset + idx, pte);
drivers/gpu/drm/i915/gvt/gtt.c
2791
pte = mm->ggtt_mm.host_ggtt_hidden[idx];
drivers/gpu/drm/i915/gvt/gtt.c
2792
if (pte & GEN8_PAGE_PRESENT)
drivers/gpu/drm/i915/gvt/gtt.c
2793
write_pte64(vgpu->gvt->gt->ggtt, offset + idx, pte);
drivers/gpu/drm/i915/gvt/gtt.c
433
DEFINE_PPGTT_GMA_TO_INDEX(gen8, pte, (gma >> 12 & 0x1ff));
drivers/gpu/drm/i915/i915_mm.c
51
static int remap_sg(pte_t *pte, unsigned long addr, void *data)
drivers/gpu/drm/i915/i915_mm.c
59
set_pte_at(r->mm, addr, pte,
drivers/gpu/drm/i915/i915_mm.c
73
static int remap_pfn(pte_t *pte, unsigned long addr, void *data)
drivers/gpu/drm/i915/i915_mm.c
78
set_pte_at(r->mm, addr, pte, pte_mkspecial(pfn_pte(r->pfn, r->prot)));
drivers/gpu/drm/imagination/pvr_vm_mips.c
183
u32 pte;
drivers/gpu/drm/imagination/pvr_vm_mips.c
192
pte = ((dma_addr >> ROGUE_MIPSFW_LOG2_PAGE_SIZE_4K)
drivers/gpu/drm/imagination/pvr_vm_mips.c
194
pte |= pte_flags;
drivers/gpu/drm/imagination/pvr_vm_mips.c
196
WRITE_ONCE(mips_data->pt[pfn], pte);
drivers/gpu/drm/lima/lima_vm.c
274
u32 pte = *pt++;
drivers/gpu/drm/lima/lima_vm.c
276
if (pte)
drivers/gpu/drm/lima/lima_vm.c
277
printk(KERN_INFO " pt %03x:%08x\n", k, pte);
drivers/gpu/drm/nouveau/nouveau_dmem.c
213
if (vmf->pte) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fbsr.c
90
int pte = 0, idx;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fbsr.c
94
rpc->pteDesc.pte_pde[pte++].pte = (sg_dma_address(sgl) >> 12) + i;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fbsr.c
99
rpc->pteDesc.pte_pde[i].pte = (phys >> 12) + i;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1662
u64 *pte;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1680
pte = rx3->lvl0.data;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1681
*pte = rx3->lvl1.addr;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1684
pte = rx3->lvl1.data;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1686
*pte++ = sg_page_iter_dma_address(&sg_dma_iter);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1692
pte = sg_virt(sg);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1693
sgl_end = (void *)pte + sg->length;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1696
*pte++ = sg_page_iter_dma_address(&sg_dma_iter);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1700
if ((void *)pte >= sgl_end)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/nvrm/fbsr.h
66
NvU64 pte; // PTE when IDR==0; PDE when IDR > 0
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
103
buf += sprintf(buf, "%05x:", it->pte[lvl]);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
146
u32 pdei = it->pte[it->lvl + 1];
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
211
pgt->pte[lpti].s.sptes -= pten;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
221
if (pgt->pte[pteb].s.sptes) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
223
if (!(pgt->pte[ptei].s.sptes))
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
235
pgt->pte[ptei].s.spte_valid = false;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
237
if (pgt->pte[ptei].s.sptes)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
239
pgt->pte[ptei].s.spte_valid = false;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
242
if (pgt->pte[pteb].s.sparse) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
245
} else if (!pgt->pte[pteb].s.lpte_valid) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
288
pgt->pte[lpti].s.lptes--;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
289
if (pgt->pte[lpti].s.lptes == 0)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
290
pgt->pte[lpti].s.lpte_valid = false;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
322
pgt->pte[lpti].s.sptes += pten;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
332
if (pgt->pte[pteb].s.spte_valid) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
334
if (!pgt->pte[ptei].s.spte_valid)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
346
pgt->pte[ptei].s.spte_valid = true;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
347
pgt->pte[ptei].s.lpte_valid = false;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
349
if (pgt->pte[ptei].s.spte_valid)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
351
pgt->pte[ptei].s.spte_valid = true;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
352
pgt->pte[ptei].s.lpte_valid = false;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
355
if (pgt->pte[pteb].s.sparse) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
393
pgt->pte[lpti].s.spte_valid = false;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
394
pgt->pte[lpti].s.lpte_valid = true;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
395
pgt->pte[lpti].s.lptes++;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
413
memset32(&pgt->pte[ptei].u, sparse.u, ptes);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
425
memset32(&pt->pte[ptei].u, 0x00, ptes);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
472
bool spte = !!pgt->pte[ptei].s.sptes;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
474
bool next = !!pgt->pte[ptei].s.sptes;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
484
memset32(&pgt->pte[pteb].u, 0x00, ptes);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
488
pgt->pte[pteb++].s.spte_valid = true;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
543
it.pte[it.lvl] = bits & ((1 << desc[it.lvl].bits) - 1);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
559
const u32 ptei = it.pte[0];
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
56
if (!(pgt = kzalloc(sizeof(*pgt) + (sizeof(pgt->pte[0]) * lpte), GFP_KERNEL)))
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
564
const u32 pdei = it.pte[it.lvl];
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
599
it.pte[it.lvl] += ptes;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
602
while (it.pte[it.lvl] == (1 << desc[it.lvl].bits)) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
603
it.pte[it.lvl++] = 0;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
604
it.pte[it.lvl]++;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
616
addr = it.pte[it.max--];
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
619
addr |= it.pte[it.max];
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
78
u32 pte[NVKM_VMM_LEVELS_MAX];
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h
58
union nvkm_pte_tracker pte[];
drivers/gpu/drm/radeon/radeon_vm.c
827
uint64_t pte;
drivers/gpu/drm/radeon/radeon_vm.c
840
pte = radeon_bo_gpu_offset(pt);
drivers/gpu/drm/radeon/radeon_vm.c
841
pte += (addr & mask) * 8;
drivers/gpu/drm/radeon/radeon_vm.c
843
if ((last_pte + 8 * count) != pte) {
drivers/gpu/drm/radeon/radeon_vm.c
852
last_pte = pte;
drivers/gpu/drm/xe/display/xe_fb_pin.c
132
u64 pte = xe_ggtt_encode_pte_flags(ggtt, bo, xe->pat.idx[XE_CACHE_NONE]);
drivers/gpu/drm/xe/display/xe_fb_pin.c
138
iosys_map_wr(&dpt->vmap, x * 8, u64, pte | addr);
drivers/gpu/drm/xe/display/xe_fb_pin.c
230
u64 pte, size;
drivers/gpu/drm/xe/display/xe_fb_pin.c
258
pte = xe_ggtt_encode_pte_flags(ggtt, bo, xe->pat.idx[XE_CACHE_NONE]);
drivers/gpu/drm/xe/display/xe_fb_pin.c
259
vma->node = xe_ggtt_node_insert_transform(ggtt, bo, pte,
drivers/gpu/drm/xe/display/xe_fb_pin.c
28
u64 pte = xe_ggtt_encode_pte_flags(ggtt, bo, xe->pat.idx[XE_CACHE_NONE]);
drivers/gpu/drm/xe/display/xe_fb_pin.c
40
iosys_map_wr(map, *dpt_ofs, u64, pte | addr);
drivers/gpu/drm/xe/display/xe_fb_pin.c
61
u64 pte = xe_ggtt_encode_pte_flags(ggtt, bo, xe->pat.idx[XE_CACHE_NONE]);
drivers/gpu/drm/xe/display/xe_fb_pin.c
68
iosys_map_wr(map, *dpt_ofs, u64, pte | addr);
drivers/gpu/drm/xe/display/xe_initial_plane.c
60
u64 pte = xe_ggtt_read_pte(tile0->mem.ggtt, base);
drivers/gpu/drm/xe/display/xe_initial_plane.c
62
if (!(pte & XE_GGTT_PTE_DM)) {
drivers/gpu/drm/xe/display/xe_initial_plane.c
68
phys_base = pte & ~(page_size - 1);
drivers/gpu/drm/xe/xe_ggtt.c
100
u64 pte;
drivers/gpu/drm/xe/xe_ggtt.c
1008
u64 pte = xe_encode_vfid_pte(vfid);
drivers/gpu/drm/xe/xe_ggtt.c
1016
ggtt->pt_ops->ggtt_set_pte(ggtt, start, pte);
drivers/gpu/drm/xe/xe_ggtt.c
102
pte = xelp_ggtt_pte_flags(bo, pat_index);
drivers/gpu/drm/xe/xe_ggtt.c
1053
u64 pte;
drivers/gpu/drm/xe/xe_ggtt.c
1068
pte = ggtt->pt_ops->ggtt_get_pte(ggtt, start);
drivers/gpu/drm/xe/xe_ggtt.c
1069
if (vfid != u64_get_bits(pte, GGTT_PTE_VFID))
drivers/gpu/drm/xe/xe_ggtt.c
107
pte |= XELPG_GGTT_PTE_PAT0;
drivers/gpu/drm/xe/xe_ggtt.c
1072
*buf++ = u64_replace_bits(pte, 0, GGTT_PTE_VFID);
drivers/gpu/drm/xe/xe_ggtt.c
110
pte |= XELPG_GGTT_PTE_PAT1;
drivers/gpu/drm/xe/xe_ggtt.c
112
return pte;
drivers/gpu/drm/xe/xe_ggtt.c
179
static void xe_ggtt_set_pte(struct xe_ggtt *ggtt, u64 addr, u64 pte)
drivers/gpu/drm/xe/xe_ggtt.c
184
writeq(pte, &ggtt->gsm[addr >> XE_PTE_SHIFT]);
drivers/gpu/drm/xe/xe_ggtt.c
187
static void xe_ggtt_set_pte_and_flush(struct xe_ggtt *ggtt, u64 addr, u64 pte)
drivers/gpu/drm/xe/xe_ggtt.c
189
xe_ggtt_set_pte(ggtt, addr, pte);
drivers/gpu/drm/xe/xe_ggtt.c
763
struct xe_bo *bo, u64 pte)
drivers/gpu/drm/xe/xe_ggtt.c
780
pte | xe_res_dma(&cur));
drivers/gpu/drm/xe/xe_ggtt.c
783
pte |= vram_region_gpu_offset(bo->ttm.resource);
drivers/gpu/drm/xe/xe_ggtt.c
788
pte + cur.start);
drivers/gpu/drm/xe/xe_ggtt.c
803
u64 pte;
drivers/gpu/drm/xe/xe_ggtt.c
806
pte = ggtt->pt_ops->pte_encode_flags(bo, pat_index);
drivers/gpu/drm/xe/xe_ggtt.c
807
xe_ggtt_map_bo(ggtt, bo->ggtt_node[ggtt->tile->id], bo, pte);
drivers/gpu/drm/xe/xe_ggtt.c
89
u64 pte = XE_PAGE_PRESENT;
drivers/gpu/drm/xe/xe_ggtt.c
900
u64 pte = ggtt->pt_ops->pte_encode_flags(bo, pat_index);
drivers/gpu/drm/xe/xe_ggtt.c
902
xe_ggtt_map_bo(ggtt, bo->ggtt_node[tile_id], bo, pte);
drivers/gpu/drm/xe/xe_ggtt.c
92
pte |= XE_GGTT_PTE_DM;
drivers/gpu/drm/xe/xe_ggtt.c
94
return pte;
drivers/gpu/drm/xe/xe_ggtt.h
32
struct xe_bo *bo, u64 pte,
drivers/gpu/drm/xe/xe_ggtt_types.h
60
typedef void (*xe_ggtt_set_pte_fn)(struct xe_ggtt *ggtt, u64 addr, u64 pte);
drivers/gpu/drm/xe/xe_lmtt.c
300
u64 pte, unsigned int idx)
drivers/gpu/drm/xe/xe_lmtt.c
305
lmtt_debug(lmtt, "WRITE level=%u index=%u pte=%#llx\n", level, idx, pte);
drivers/gpu/drm/xe/xe_lmtt.c
309
lmtt_assert(lmtt, !overflows_type(pte, u32));
drivers/gpu/drm/xe/xe_lmtt.c
310
lmtt_assert(lmtt, !pte || !iosys_map_rd(&pt->bo->vmap, idx * sizeof(u32), u32));
drivers/gpu/drm/xe/xe_lmtt.c
312
xe_map_wr(lmtt_to_xe(lmtt), &pt->bo->vmap, idx * sizeof(u32), u32, pte);
drivers/gpu/drm/xe/xe_lmtt.c
315
lmtt_assert(lmtt, !pte || !iosys_map_rd(&pt->bo->vmap, idx * sizeof(u64), u64));
drivers/gpu/drm/xe/xe_lmtt.c
317
xe_map_wr(lmtt_to_xe(lmtt), &pt->bo->vmap, idx * sizeof(u64), u64, pte);
drivers/gpu/drm/xe/xe_migrate.c
2050
u64 pte;
drivers/gpu/drm/xe/xe_migrate.c
2059
pte = m->q->vm->pt_ops->pte_encode_addr(m->tile->xe,
drivers/gpu/drm/xe/xe_migrate.c
2062
bb->cs[bb->len++] = lower_32_bits(pte);
drivers/gpu/drm/xe/xe_migrate.c
2063
bb->cs[bb->len++] = upper_32_bits(pte);
drivers/gpu/drm/xe/xe_pt.c
1587
u64 pte, struct xe_pt *xe_child)
drivers/gpu/drm/xe/xe_pt.c
1591
u64 phys_addr = pte & XE_PTE_ADDR_MASK;
drivers/gpu/drm/xe/xe_pt.c
1609
if (xe_child->level == 0 && !(pte & XE_PTE_PS64)) {
drivers/gpu/drm/xe/xe_pt.c
1617
} else if (xe_child->level == 1 && pte & XE_PDE_PS_2M) {
drivers/gpu/drm/xe/xe_pt.c
1624
xe_child->level, pte);
drivers/gpu/drm/xe/xe_pt.c
1636
xe_child->level, pte, reclamation_size, num_entries);
drivers/gpu/drm/xe/xe_pt.c
1663
u64 pte;
drivers/gpu/drm/xe/xe_pt.c
1685
pte = xe_map_rd(xe, leaf_map, (first + i) * sizeof(u64), u64);
drivers/gpu/drm/xe/xe_pt.c
1691
if (!pte) {
drivers/gpu/drm/xe/xe_pt.c
1699
(pte & (XE_PDE_PS_2M | XE_PDPE_PS_1G)));
drivers/gpu/drm/xe/xe_pt.c
1702
if (pte & XE_PTE_PS64)
drivers/gpu/drm/xe/xe_pt.c
1708
pte, xe_child);
drivers/gpu/drm/xe/xe_pt.c
1715
pte);
drivers/gpu/drm/xe/xe_pt.c
386
pgoff_t offset, struct xe_pt *xe_child, u64 pte)
drivers/gpu/drm/xe/xe_pt.c
416
xe_pt_write(xe_walk->vm->xe, map, offset, pte);
drivers/gpu/drm/xe/xe_pt.c
425
entry->pt_entries[idx].pte = pte;
drivers/gpu/drm/xe/xe_pt.c
529
u64 pte;
drivers/gpu/drm/xe/xe_pt.c
540
pte = 0;
drivers/gpu/drm/xe/xe_pt.c
542
pte = vm->pt_ops->pte_encode_vma(is_null ? 0 :
drivers/gpu/drm/xe/xe_pt.c
548
pte |= is_vram ? xe_walk->default_vram_pte :
drivers/gpu/drm/xe/xe_pt.c
559
pte |= XE_PTE_PS64;
drivers/gpu/drm/xe/xe_pt.c
567
ret = xe_pt_insert_entry(xe_walk, xe_parent, offset, NULL, pte);
drivers/gpu/drm/xe/xe_pt.c
622
pte = vm->pt_ops->pde_encode_bo(xe_child->bo, 0) | flags;
drivers/gpu/drm/xe/xe_pt.c
624
pte);
drivers/gpu/drm/xe/xe_pt.c
995
sizeof(u64), u64, ptes[i].pte);
drivers/gpu/drm/xe/xe_pt.c
997
ptr[i] = ptes[i].pte;
drivers/gpu/drm/xe/xe_pt_types.h
45
u64 (*pte_encode_vma)(u64 pte, struct xe_vma *vma,
drivers/gpu/drm/xe/xe_pt_types.h
55
u64 pte;
drivers/gpu/drm/xe/xe_vm.c
1256
u64 pte = 0;
drivers/gpu/drm/xe/xe_vm.c
1259
pte |= XE_PPGTT_PTE_PAT0;
drivers/gpu/drm/xe/xe_vm.c
1262
pte |= XE_PPGTT_PTE_PAT1;
drivers/gpu/drm/xe/xe_vm.c
1264
return pte;
drivers/gpu/drm/xe/xe_vm.c
1269
u64 pte = 0;
drivers/gpu/drm/xe/xe_vm.c
1272
pte |= XE_PPGTT_PTE_PAT0;
drivers/gpu/drm/xe/xe_vm.c
1275
pte |= XE_PPGTT_PTE_PAT1;
drivers/gpu/drm/xe/xe_vm.c
1279
pte |= XE_PPGTT_PDE_PDPE_PAT2;
drivers/gpu/drm/xe/xe_vm.c
1281
pte |= XE_PPGTT_PTE_PAT2;
drivers/gpu/drm/xe/xe_vm.c
1285
pte |= XELPG_PPGTT_PTE_PAT3;
drivers/gpu/drm/xe/xe_vm.c
1288
pte |= XE2_PPGTT_PTE_PAT4;
drivers/gpu/drm/xe/xe_vm.c
1290
return pte;
drivers/gpu/drm/xe/xe_vm.c
1345
u64 pte;
drivers/gpu/drm/xe/xe_vm.c
1347
pte = xe_bo_addr(bo, bo_offset, XE_PAGE_SIZE);
drivers/gpu/drm/xe/xe_vm.c
1348
pte |= XE_PAGE_PRESENT | XE_PAGE_RW;
drivers/gpu/drm/xe/xe_vm.c
1349
pte |= pte_encode_pat_index(pat_index, pt_level);
drivers/gpu/drm/xe/xe_vm.c
1350
pte |= pte_encode_ps(pt_level);
drivers/gpu/drm/xe/xe_vm.c
1353
pte |= XE_PPGTT_PTE_DM;
drivers/gpu/drm/xe/xe_vm.c
1355
return pte;
drivers/gpu/drm/xe/xe_vm.c
1358
static u64 xelp_pte_encode_vma(u64 pte, struct xe_vma *vma,
drivers/gpu/drm/xe/xe_vm.c
1361
pte |= XE_PAGE_PRESENT;
drivers/gpu/drm/xe/xe_vm.c
1364
pte |= XE_PAGE_RW;
drivers/gpu/drm/xe/xe_vm.c
1366
pte |= pte_encode_pat_index(pat_index, pt_level);
drivers/gpu/drm/xe/xe_vm.c
1367
pte |= pte_encode_ps(pt_level);
drivers/gpu/drm/xe/xe_vm.c
1370
pte |= XE_PTE_NULL;
drivers/gpu/drm/xe/xe_vm.c
1372
return pte;
drivers/gpu/drm/xe/xe_vm.c
1379
u64 pte;
drivers/gpu/drm/xe/xe_vm.c
1384
pte = addr;
drivers/gpu/drm/xe/xe_vm.c
1385
pte |= XE_PAGE_PRESENT | XE_PAGE_RW;
drivers/gpu/drm/xe/xe_vm.c
1386
pte |= pte_encode_pat_index(pat_index, pt_level);
drivers/gpu/drm/xe/xe_vm.c
1387
pte |= pte_encode_ps(pt_level);
drivers/gpu/drm/xe/xe_vm.c
1390
pte |= XE_PPGTT_PTE_DM;
drivers/gpu/drm/xe/xe_vm.c
1392
pte |= flags;
drivers/gpu/drm/xe/xe_vm.c
1394
return pte;
drivers/iommu/amd/iommu.c
1991
u64 *pte;
drivers/iommu/amd/iommu.c
1998
pte = &root[index];
drivers/iommu/amd/iommu.c
2003
if (!(*pte & GCR3_VALID)) {
drivers/iommu/amd/iommu.c
2011
*pte = iommu_virt_to_phys(root) | GCR3_VALID;
drivers/iommu/amd/iommu.c
2014
root = iommu_phys_to_virt(*pte & PAGE_MASK);
drivers/iommu/amd/iommu.c
2019
return pte;
drivers/iommu/amd/iommu.c
2026
u64 *pte;
drivers/iommu/amd/iommu.c
2028
pte = __get_gcr3_pte(gcr3_info, pasid, true);
drivers/iommu/amd/iommu.c
2029
if (pte == NULL)
drivers/iommu/amd/iommu.c
2033
*pte = (gcr3 & PAGE_MASK) | GCR3_VALID;
drivers/iommu/amd/iommu.c
2035
*pte = 0;
drivers/iommu/generic_pt/fmt/amdv1.h
344
u64 pte = 0;
drivers/iommu/generic_pt/fmt/amdv1.h
347
pte |= AMDV1PT_FMT_FC;
drivers/iommu/generic_pt/fmt/amdv1.h
349
pte |= AMDV1PT_FMT_IR;
drivers/iommu/generic_pt/fmt/amdv1.h
351
pte |= AMDV1PT_FMT_IW;
drivers/iommu/generic_pt/fmt/amdv1.h
359
pte = __sme_set(pte);
drivers/iommu/generic_pt/fmt/amdv1.h
361
attrs->descriptor_bits = pte;
drivers/iommu/generic_pt/fmt/vtdss.h
219
u64 pte = 0;
drivers/iommu/generic_pt/fmt/vtdss.h
229
pte |= VTDSS_FMT_R;
drivers/iommu/generic_pt/fmt/vtdss.h
231
pte |= VTDSS_FMT_W;
drivers/iommu/generic_pt/fmt/vtdss.h
233
pte |= VTDSS_FMT_SNP;
drivers/iommu/generic_pt/fmt/vtdss.h
242
attrs->descriptor_bits = pte;
drivers/iommu/generic_pt/fmt/x86_64.h
220
u64 pte;
drivers/iommu/generic_pt/fmt/x86_64.h
222
pte = X86_64_FMT_U | X86_64_FMT_A;
drivers/iommu/generic_pt/fmt/x86_64.h
224
pte |= X86_64_FMT_RW | X86_64_FMT_D;
drivers/iommu/generic_pt/fmt/x86_64.h
232
pte = __sme_set(pte);
drivers/iommu/generic_pt/fmt/x86_64.h
234
attrs->descriptor_bits = pte;
drivers/iommu/intel/iommu.c
556
struct dma_pte *pte;
drivers/iommu/intel/iommu.c
561
pte = &parent[offset];
drivers/iommu/intel/iommu.c
563
pr_info("pte level: %d, pte value: 0x%016llx\n", level, pte->val);
drivers/iommu/intel/iommu.c
565
if (!dma_pte_present(pte)) {
drivers/iommu/intel/iommu.c
570
if (level == 1 || dma_pte_superpage(pte))
drivers/iommu/intel/iommu.c
573
parent = phys_to_virt(dma_pte_addr(pte));
drivers/iommu/intel/iommu.c
582
struct pasid_entry *entries, *pte;
drivers/iommu/intel/iommu.c
649
pte = &entries[index];
drivers/iommu/intel/iommu.c
650
for (i = 0; i < ARRAY_SIZE(pte->val); i++)
drivers/iommu/intel/iommu.c
651
pr_info("pasid table entry[%d]: 0x%016llx\n", i, pte->val[i]);
drivers/iommu/intel/iommu.c
653
if (!pasid_pte_is_present(pte)) {
drivers/iommu/intel/iommu.c
658
if (pasid_pte_get_pgtt(pte) == PASID_ENTRY_PGTT_FL_ONLY) {
drivers/iommu/intel/iommu.c
659
level = pte->val[2] & BIT_ULL(2) ? 5 : 4;
drivers/iommu/intel/iommu.c
660
pgtable = phys_to_virt(pte->val[2] & VTD_PAGE_MASK);
drivers/iommu/intel/iommu.c
662
level = agaw_to_level((pte->val[0] >> 2) & 0x7);
drivers/iommu/intel/iommu.c
663
pgtable = phys_to_virt(pte->val[0] & VTD_PAGE_MASK);
drivers/iommu/intel/iommu.h
850
static inline u64 dma_pte_addr(struct dma_pte *pte)
drivers/iommu/intel/iommu.h
853
return pte->val & VTD_PAGE_MASK;
drivers/iommu/intel/iommu.h
856
return __cmpxchg64(&pte->val, 0ULL, 0ULL) & VTD_PAGE_MASK;
drivers/iommu/intel/iommu.h
860
static inline bool dma_pte_present(struct dma_pte *pte)
drivers/iommu/intel/iommu.h
862
return (pte->val & 3) != 0;
drivers/iommu/intel/iommu.h
865
static inline bool dma_pte_superpage(struct dma_pte *pte)
drivers/iommu/intel/iommu.h
867
return (pte->val & DMA_PTE_LARGE_PAGE);
drivers/iommu/intel/pasid.c
244
struct pasid_entry *pte;
drivers/iommu/intel/pasid.c
248
pte = intel_pasid_get_entry(dev, pasid);
drivers/iommu/intel/pasid.c
249
if (WARN_ON(!pte)) {
drivers/iommu/intel/pasid.c
254
if (!pasid_pte_is_present(pte)) {
drivers/iommu/intel/pasid.c
255
if (!pasid_pte_is_fault_disabled(pte)) {
drivers/iommu/intel/pasid.c
256
WARN_ON(READ_ONCE(pte->val[0]) != 0);
drivers/iommu/intel/pasid.c
267
pasid_clear_entry(pte);
drivers/iommu/intel/pasid.c
274
did = pasid_get_domain_id(pte);
drivers/iommu/intel/pasid.c
275
pgtt = pasid_pte_get_pgtt(pte);
drivers/iommu/intel/pasid.c
276
pasid_clear_present(pte);
drivers/iommu/intel/pasid.c
280
clflush_cache_range(pte, sizeof(*pte));
drivers/iommu/intel/pasid.c
292
clflush_cache_range(pte, sizeof(*pte));
drivers/iommu/intel/pasid.c
303
struct pasid_entry *pte,
drivers/iommu/intel/pasid.c
307
clflush_cache_range(pte, sizeof(*pte));
drivers/iommu/intel/pasid.c
328
struct pasid_entry *pte)
drivers/iommu/intel/pasid.c
331
clflush_cache_range(pte, sizeof(*pte));
drivers/iommu/intel/pasid.c
355
struct pasid_entry *pte,
drivers/iommu/intel/pasid.c
361
pasid_clear_entry(pte);
drivers/iommu/intel/pasid.c
364
pasid_set_flptr(pte, fsptptr);
drivers/iommu/intel/pasid.c
367
pasid_set_flpm(pte, 1);
drivers/iommu/intel/pasid.c
370
pasid_set_pgsnp(pte);
drivers/iommu/intel/pasid.c
372
pasid_set_domain_id(pte, did);
drivers/iommu/intel/pasid.c
373
pasid_set_address_width(pte, iommu->agaw);
drivers/iommu/intel/pasid.c
374
pasid_set_page_snoop(pte, flags & PASID_FLAG_PWSNP);
drivers/iommu/intel/pasid.c
377
pasid_set_translation_type(pte, PASID_ENTRY_PGTT_FL_ONLY);
drivers/iommu/intel/pasid.c
378
pasid_set_present(pte);
drivers/iommu/intel/pasid.c
385
struct pasid_entry *pte;
drivers/iommu/intel/pasid.c
400
pte = intel_pasid_get_entry(dev, pasid);
drivers/iommu/intel/pasid.c
401
if (!pte) {
drivers/iommu/intel/pasid.c
406
if (pasid_pte_is_present(pte)) {
drivers/iommu/intel/pasid.c
411
pasid_pte_config_first_level(iommu, pte, fsptptr, did, flags);
drivers/iommu/intel/pasid.c
415
pasid_flush_caches(iommu, pte, pasid, did);
drivers/iommu/intel/pasid.c
424
struct pasid_entry *pte,
drivers/iommu/intel/pasid.c
432
pasid_clear_entry(pte);
drivers/iommu/intel/pasid.c
433
pasid_set_domain_id(pte, did);
drivers/iommu/intel/pasid.c
434
pasid_set_slptr(pte, pt_info.ssptptr);
drivers/iommu/intel/pasid.c
435
pasid_set_address_width(pte, pt_info.aw);
drivers/iommu/intel/pasid.c
436
pasid_set_translation_type(pte, PASID_ENTRY_PGTT_SL_ONLY);
drivers/iommu/intel/pasid.c
437
pasid_set_fault_enable(pte);
drivers/iommu/intel/pasid.c
438
pasid_set_page_snoop(pte, !(domain->sspt.vtdss_pt.common.features &
drivers/iommu/intel/pasid.c
441
pasid_set_ssade(pte);
drivers/iommu/intel/pasid.c
443
pasid_set_present(pte);
drivers/iommu/intel/pasid.c
450
struct pasid_entry *pte;
drivers/iommu/intel/pasid.c
467
pte = intel_pasid_get_entry(dev, pasid);
drivers/iommu/intel/pasid.c
468
if (!pte) {
drivers/iommu/intel/pasid.c
473
if (pasid_pte_is_present(pte)) {
drivers/iommu/intel/pasid.c
478
pasid_pte_config_second_level(iommu, pte, domain, did);
drivers/iommu/intel/pasid.c
481
pasid_flush_caches(iommu, pte, pasid, did);
drivers/iommu/intel/pasid.c
493
struct pasid_entry *pte;
drivers/iommu/intel/pasid.c
498
pte = intel_pasid_get_entry(dev, pasid);
drivers/iommu/intel/pasid.c
499
if (!pte) {
drivers/iommu/intel/pasid.c
506
did = pasid_get_domain_id(pte);
drivers/iommu/intel/pasid.c
507
pgtt = pasid_pte_get_pgtt(pte);
drivers/iommu/intel/pasid.c
518
if (pasid_get_ssade(pte) == enabled) {
drivers/iommu/intel/pasid.c
524
pasid_set_ssade(pte);
drivers/iommu/intel/pasid.c
526
pasid_clear_ssade(pte);
drivers/iommu/intel/pasid.c
530
clflush_cache_range(pte, sizeof(*pte));
drivers/iommu/intel/pasid.c
559
struct pasid_entry *pte, u16 did)
drivers/iommu/intel/pasid.c
563
pasid_clear_entry(pte);
drivers/iommu/intel/pasid.c
564
pasid_set_domain_id(pte, did);
drivers/iommu/intel/pasid.c
565
pasid_set_address_width(pte, iommu->agaw);
drivers/iommu/intel/pasid.c
566
pasid_set_translation_type(pte, PASID_ENTRY_PGTT_PT);
drivers/iommu/intel/pasid.c
567
pasid_set_fault_enable(pte);
drivers/iommu/intel/pasid.c
568
pasid_set_page_snoop(pte, !!ecap_smpwc(iommu->ecap));
drivers/iommu/intel/pasid.c
569
pasid_set_present(pte);
drivers/iommu/intel/pasid.c
576
struct pasid_entry *pte;
drivers/iommu/intel/pasid.c
579
pte = intel_pasid_get_entry(dev, pasid);
drivers/iommu/intel/pasid.c
580
if (!pte) {
drivers/iommu/intel/pasid.c
585
if (pasid_pte_is_present(pte)) {
drivers/iommu/intel/pasid.c
590
pasid_pte_config_pass_through(iommu, pte, did);
drivers/iommu/intel/pasid.c
593
pasid_flush_caches(iommu, pte, pasid, did);
drivers/iommu/intel/pasid.c
604
struct pasid_entry *pte;
drivers/iommu/intel/pasid.c
608
pte = intel_pasid_get_entry(dev, pasid);
drivers/iommu/intel/pasid.c
609
if (WARN_ON(!pte || !pasid_pte_is_present(pte))) {
drivers/iommu/intel/pasid.c
614
pasid_set_pgsnp(pte);
drivers/iommu/intel/pasid.c
615
did = pasid_get_domain_id(pte);
drivers/iommu/intel/pasid.c
618
intel_pasid_flush_present(iommu, dev, pasid, did, pte);
drivers/iommu/intel/pasid.c
622
struct pasid_entry *pte,
drivers/iommu/intel/pasid.c
633
pasid_clear_entry(pte);
drivers/iommu/intel/pasid.c
636
pasid_set_flpm(pte, 1);
drivers/iommu/intel/pasid.c
638
pasid_set_flptr(pte, s1_cfg->pgtbl_addr);
drivers/iommu/intel/pasid.c
641
pasid_set_sre(pte);
drivers/iommu/intel/pasid.c
643
pasid_set_wpe(pte);
drivers/iommu/intel/pasid.c
647
pasid_set_eafe(pte);
drivers/iommu/intel/pasid.c
650
pasid_set_pgsnp(pte);
drivers/iommu/intel/pasid.c
652
pasid_set_slptr(pte, pt_info.ssptptr);
drivers/iommu/intel/pasid.c
653
pasid_set_fault_enable(pte);
drivers/iommu/intel/pasid.c
654
pasid_set_domain_id(pte, did);
drivers/iommu/intel/pasid.c
655
pasid_set_address_width(pte, pt_info.aw);
drivers/iommu/intel/pasid.c
656
pasid_set_page_snoop(pte, !(s2_domain->sspt.vtdss_pt.common.features &
drivers/iommu/intel/pasid.c
659
pasid_set_ssade(pte);
drivers/iommu/intel/pasid.c
660
pasid_set_translation_type(pte, PASID_ENTRY_PGTT_NESTED);
drivers/iommu/intel/pasid.c
661
pasid_set_present(pte);
drivers/iommu/intel/pasid.c
681
struct pasid_entry *pte;
drivers/iommu/intel/pasid.c
713
pte = intel_pasid_get_entry(dev, pasid);
drivers/iommu/intel/pasid.c
714
if (!pte) {
drivers/iommu/intel/pasid.c
718
if (pasid_pte_is_present(pte)) {
drivers/iommu/intel/pasid.c
723
pasid_pte_config_nestd(iommu, pte, s1_cfg, s2_domain, did);
drivers/iommu/intel/pasid.c
726
pasid_flush_caches(iommu, pte, pasid, did);
drivers/iommu/intel/pasid.h
71
static inline bool pasid_pte_is_present(struct pasid_entry *pte)
drivers/iommu/intel/pasid.h
73
return READ_ONCE(pte->val[0]) & PASID_PTE_PRESENT;
drivers/iommu/intel/pasid.h
77
static inline bool pasid_pte_is_fault_disabled(struct pasid_entry *pte)
drivers/iommu/intel/pasid.h
79
return READ_ONCE(pte->val[0]) & PASID_PTE_FPD;
drivers/iommu/intel/pasid.h
83
static inline u16 pasid_pte_get_pgtt(struct pasid_entry *pte)
drivers/iommu/intel/pasid.h
85
return (u16)((READ_ONCE(pte->val[0]) >> 6) & 0x7);
drivers/iommu/io-pgtable-arm-v7s.c
171
static bool arm_v7s_pte_is_cont(arm_v7s_iopte pte, int lvl);
drivers/iommu/io-pgtable-arm-v7s.c
184
static arm_v7s_iopte to_mtk_iopte(phys_addr_t paddr, arm_v7s_iopte pte)
drivers/iommu/io-pgtable-arm-v7s.c
187
pte |= ARM_V7S_ATTR_MTK_PA_BIT32;
drivers/iommu/io-pgtable-arm-v7s.c
189
pte |= ARM_V7S_ATTR_MTK_PA_BIT33;
drivers/iommu/io-pgtable-arm-v7s.c
191
pte |= ARM_V7S_ATTR_MTK_PA_BIT34;
drivers/iommu/io-pgtable-arm-v7s.c
192
return pte;
drivers/iommu/io-pgtable-arm-v7s.c
198
arm_v7s_iopte pte = paddr & ARM_V7S_LVL_MASK(lvl);
drivers/iommu/io-pgtable-arm-v7s.c
201
return to_mtk_iopte(paddr, pte);
drivers/iommu/io-pgtable-arm-v7s.c
203
return pte;
drivers/iommu/io-pgtable-arm-v7s.c
206
static phys_addr_t iopte_to_paddr(arm_v7s_iopte pte, int lvl,
drivers/iommu/io-pgtable-arm-v7s.c
212
if (ARM_V7S_PTE_IS_TABLE(pte, lvl))
drivers/iommu/io-pgtable-arm-v7s.c
214
else if (arm_v7s_pte_is_cont(pte, lvl))
drivers/iommu/io-pgtable-arm-v7s.c
219
paddr = pte & mask;
drivers/iommu/io-pgtable-arm-v7s.c
223
if (pte & ARM_V7S_ATTR_MTK_PA_BIT32)
drivers/iommu/io-pgtable-arm-v7s.c
225
if (pte & ARM_V7S_ATTR_MTK_PA_BIT33)
drivers/iommu/io-pgtable-arm-v7s.c
227
if (pte & ARM_V7S_ATTR_MTK_PA_BIT34)
drivers/iommu/io-pgtable-arm-v7s.c
232
static arm_v7s_iopte *iopte_deref(arm_v7s_iopte pte, int lvl,
drivers/iommu/io-pgtable-arm-v7s.c
235
return phys_to_virt(iopte_to_paddr(pte, lvl, &data->iop.cfg));
drivers/iommu/io-pgtable-arm-v7s.c
323
static void __arm_v7s_set_pte(arm_v7s_iopte *ptep, arm_v7s_iopte pte,
drivers/iommu/io-pgtable-arm-v7s.c
329
ptep[i] = pte;
drivers/iommu/io-pgtable-arm-v7s.c
338
arm_v7s_iopte pte = ARM_V7S_ATTR_NG | ARM_V7S_ATTR_S;
drivers/iommu/io-pgtable-arm-v7s.c
341
pte |= ARM_V7S_ATTR_TEX(1);
drivers/iommu/io-pgtable-arm-v7s.c
343
pte |= ARM_V7S_PTE_AF;
drivers/iommu/io-pgtable-arm-v7s.c
345
pte |= ARM_V7S_PTE_AP_UNPRIV;
drivers/iommu/io-pgtable-arm-v7s.c
347
pte |= ARM_V7S_PTE_AP_RDONLY;
drivers/iommu/io-pgtable-arm-v7s.c
349
pte <<= ARM_V7S_ATTR_SHIFT(lvl);
drivers/iommu/io-pgtable-arm-v7s.c
352
pte |= ARM_V7S_ATTR_XN(lvl);
drivers/iommu/io-pgtable-arm-v7s.c
354
pte |= ARM_V7S_ATTR_B;
drivers/iommu/io-pgtable-arm-v7s.c
356
pte |= ARM_V7S_ATTR_B | ARM_V7S_ATTR_C;
drivers/iommu/io-pgtable-arm-v7s.c
358
pte |= ARM_V7S_PTE_TYPE_PAGE;
drivers/iommu/io-pgtable-arm-v7s.c
360
pte |= ARM_V7S_ATTR_NS_SECTION;
drivers/iommu/io-pgtable-arm-v7s.c
362
return pte;
drivers/iommu/io-pgtable-arm-v7s.c
365
static arm_v7s_iopte arm_v7s_pte_to_cont(arm_v7s_iopte pte, int lvl)
drivers/iommu/io-pgtable-arm-v7s.c
368
pte |= ARM_V7S_CONT_SECTION;
drivers/iommu/io-pgtable-arm-v7s.c
370
arm_v7s_iopte xn = pte & ARM_V7S_ATTR_XN(lvl);
drivers/iommu/io-pgtable-arm-v7s.c
371
arm_v7s_iopte tex = pte & ARM_V7S_CONT_PAGE_TEX_MASK;
drivers/iommu/io-pgtable-arm-v7s.c
373
pte ^= xn | tex | ARM_V7S_PTE_TYPE_PAGE;
drivers/iommu/io-pgtable-arm-v7s.c
374
pte |= (xn << ARM_V7S_CONT_PAGE_XN_SHIFT) |
drivers/iommu/io-pgtable-arm-v7s.c
378
return pte;
drivers/iommu/io-pgtable-arm-v7s.c
381
static bool arm_v7s_pte_is_cont(arm_v7s_iopte pte, int lvl)
drivers/iommu/io-pgtable-arm-v7s.c
383
if (lvl == 1 && !ARM_V7S_PTE_IS_TABLE(pte, lvl))
drivers/iommu/io-pgtable-arm-v7s.c
384
return pte & ARM_V7S_CONT_SECTION;
drivers/iommu/io-pgtable-arm-v7s.c
386
return !(pte & ARM_V7S_PTE_TYPE_PAGE);
drivers/iommu/io-pgtable-arm-v7s.c
399
arm_v7s_iopte pte;
drivers/iommu/io-pgtable-arm-v7s.c
421
pte = arm_v7s_prot_to_pte(prot, lvl, cfg);
drivers/iommu/io-pgtable-arm-v7s.c
423
pte = arm_v7s_pte_to_cont(pte, lvl);
drivers/iommu/io-pgtable-arm-v7s.c
425
pte |= paddr_to_iopte(paddr, lvl, cfg);
drivers/iommu/io-pgtable-arm-v7s.c
427
__arm_v7s_set_pte(ptep, pte, num_entries, cfg);
drivers/iommu/io-pgtable-arm-v7s.c
465
arm_v7s_iopte pte, *cptep;
drivers/iommu/io-pgtable-arm-v7s.c
481
pte = READ_ONCE(*ptep);
drivers/iommu/io-pgtable-arm-v7s.c
482
if (!pte) {
drivers/iommu/io-pgtable-arm-v7s.c
487
pte = arm_v7s_install_table(cptep, ptep, 0, cfg);
drivers/iommu/io-pgtable-arm-v7s.c
488
if (pte)
drivers/iommu/io-pgtable-arm-v7s.c
495
if (ARM_V7S_PTE_IS_TABLE(pte, lvl)) {
drivers/iommu/io-pgtable-arm-v7s.c
496
cptep = iopte_deref(pte, lvl, data);
drivers/iommu/io-pgtable-arm-v7s.c
497
} else if (pte) {
drivers/iommu/io-pgtable-arm-v7s.c
546
arm_v7s_iopte pte = data->pgd[i];
drivers/iommu/io-pgtable-arm-v7s.c
548
if (ARM_V7S_PTE_IS_TABLE(pte, 1))
drivers/iommu/io-pgtable-arm-v7s.c
549
__arm_v7s_free_table(iopte_deref(pte, 1, data),
drivers/iommu/io-pgtable-arm-v7s.c
562
arm_v7s_iopte pte[ARM_V7S_CONT_PAGES];
drivers/iommu/io-pgtable-arm-v7s.c
573
pte[i] = READ_ONCE(ptep[i]);
drivers/iommu/io-pgtable-arm-v7s.c
574
if (WARN_ON(!ARM_V7S_PTE_IS_VALID(pte[i])))
drivers/iommu/io-pgtable-arm-v7s.c
588
if (num_entries <= 1 && arm_v7s_pte_is_cont(pte[0], lvl)) {
drivers/iommu/io-pgtable-arm-v7s.c
600
if (ARM_V7S_PTE_IS_TABLE(pte[i], lvl)) {
drivers/iommu/io-pgtable-arm-v7s.c
604
ptep = iopte_deref(pte[i], lvl, data);
drivers/iommu/io-pgtable-arm-v7s.c
612
} else if (lvl == 1 && !ARM_V7S_PTE_IS_TABLE(pte[0], lvl)) {
drivers/iommu/io-pgtable-arm-v7s.c
618
ptep = iopte_deref(pte[0], lvl, data);
drivers/iommu/io-pgtable-arm-v7s.c
648
arm_v7s_iopte *ptep = data->pgd, pte;
drivers/iommu/io-pgtable-arm-v7s.c
654
pte = READ_ONCE(*ptep);
drivers/iommu/io-pgtable-arm-v7s.c
655
ptep = iopte_deref(pte, lvl, data);
drivers/iommu/io-pgtable-arm-v7s.c
656
} while (ARM_V7S_PTE_IS_TABLE(pte, lvl));
drivers/iommu/io-pgtable-arm-v7s.c
658
if (!ARM_V7S_PTE_IS_VALID(pte))
drivers/iommu/io-pgtable-arm-v7s.c
662
if (arm_v7s_pte_is_cont(pte, lvl))
drivers/iommu/io-pgtable-arm-v7s.c
664
return iopte_to_paddr(pte, lvl, &data->iop.cfg) | (iova & ~mask);
drivers/iommu/io-pgtable-arm-v7s.c
83
#define ARM_V7S_PTE_IS_VALID(pte) (((pte) & 0x3) != 0)
drivers/iommu/io-pgtable-arm-v7s.c
84
#define ARM_V7S_PTE_IS_TABLE(pte, lvl) \
drivers/iommu/io-pgtable-arm-v7s.c
85
((lvl) == 1 && (((pte) & 0x3) == ARM_V7S_PTE_TYPE_TABLE))
drivers/iommu/io-pgtable-arm.c
146
#define iopte_deref(pte,d) __va(iopte_to_paddr(pte, d))
drivers/iommu/io-pgtable-arm.c
148
#define iopte_type(pte) \
drivers/iommu/io-pgtable-arm.c
149
(((pte) >> ARM_LPAE_PTE_TYPE_SHIFT) & ARM_LPAE_PTE_TYPE_MASK)
drivers/iommu/io-pgtable-arm.c
151
#define iopte_writeable_dirty(pte) \
drivers/iommu/io-pgtable-arm.c
152
(((pte) & ARM_LPAE_PTE_AP_WR_CLEAN_MASK) == ARM_LPAE_PTE_DBM)
drivers/iommu/io-pgtable-arm.c
169
static inline bool iopte_leaf(arm_lpae_iopte pte, int lvl,
drivers/iommu/io-pgtable-arm.c
173
return iopte_type(pte) == ARM_LPAE_PTE_TYPE_PAGE;
drivers/iommu/io-pgtable-arm.c
175
return iopte_type(pte) == ARM_LPAE_PTE_TYPE_BLOCK;
drivers/iommu/io-pgtable-arm.c
178
static inline bool iopte_table(arm_lpae_iopte pte, int lvl)
drivers/iommu/io-pgtable-arm.c
182
return iopte_type(pte) == ARM_LPAE_PTE_TYPE_TABLE;
drivers/iommu/io-pgtable-arm.c
188
arm_lpae_iopte pte = paddr;
drivers/iommu/io-pgtable-arm.c
191
return (pte | (pte >> (48 - 12))) & ARM_LPAE_PTE_ADDR_MASK;
drivers/iommu/io-pgtable-arm.c
194
static phys_addr_t iopte_to_paddr(arm_lpae_iopte pte,
drivers/iommu/io-pgtable-arm.c
197
u64 paddr = pte & ARM_LPAE_PTE_ADDR_MASK;
drivers/iommu/io-pgtable-arm.c
341
arm_lpae_iopte pte = prot;
drivers/iommu/io-pgtable-arm.c
347
pte |= ARM_LPAE_PTE_TYPE_PAGE;
drivers/iommu/io-pgtable-arm.c
349
pte |= ARM_LPAE_PTE_TYPE_BLOCK;
drivers/iommu/io-pgtable-arm.c
352
ptep[i] = pte | paddr_to_iopte(paddr + i * sz, data);
drivers/iommu/io-pgtable-arm.c
427
arm_lpae_iopte *cptep, pte;
drivers/iommu/io-pgtable-arm.c
453
pte = READ_ONCE(*ptep);
drivers/iommu/io-pgtable-arm.c
454
if (!pte) {
drivers/iommu/io-pgtable-arm.c
459
pte = arm_lpae_install_table(cptep, ptep, 0, data);
drivers/iommu/io-pgtable-arm.c
460
if (pte)
drivers/iommu/io-pgtable-arm.c
462
} else if (!cfg->coherent_walk && !(pte & ARM_LPAE_PTE_SW_SYNC)) {
drivers/iommu/io-pgtable-arm.c
466
if (pte && !iopte_leaf(pte, lvl, data->iop.fmt)) {
drivers/iommu/io-pgtable-arm.c
467
cptep = iopte_deref(pte, data);
drivers/iommu/io-pgtable-arm.c
468
} else if (pte) {
drivers/iommu/io-pgtable-arm.c
482
arm_lpae_iopte pte;
drivers/iommu/io-pgtable-arm.c
486
pte = ARM_LPAE_PTE_nG;
drivers/iommu/io-pgtable-arm.c
488
pte |= ARM_LPAE_PTE_AP_RDONLY;
drivers/iommu/io-pgtable-arm.c
490
pte |= ARM_LPAE_PTE_DBM;
drivers/iommu/io-pgtable-arm.c
492
pte |= ARM_LPAE_PTE_AP_UNPRIV;
drivers/iommu/io-pgtable-arm.c
494
pte = ARM_LPAE_PTE_HAP_FAULT;
drivers/iommu/io-pgtable-arm.c
496
pte |= ARM_LPAE_PTE_HAP_READ;
drivers/iommu/io-pgtable-arm.c
498
pte |= ARM_LPAE_PTE_HAP_WRITE;
drivers/iommu/io-pgtable-arm.c
508
pte |= ARM_LPAE_PTE_MEMATTR_DEV;
drivers/iommu/io-pgtable-arm.c
511
pte |= ARM_LPAE_PTE_MEMATTR_FWB_WB;
drivers/iommu/io-pgtable-arm.c
513
pte |= ARM_LPAE_PTE_MEMATTR_OIWB;
drivers/iommu/io-pgtable-arm.c
515
pte |= ARM_LPAE_PTE_MEMATTR_NC;
drivers/iommu/io-pgtable-arm.c
519
pte |= (ARM_LPAE_MAIR_ATTR_IDX_DEV
drivers/iommu/io-pgtable-arm.c
522
pte |= (ARM_LPAE_MAIR_ATTR_IDX_CACHE
drivers/iommu/io-pgtable-arm.c
533
pte |= ARM_LPAE_PTE_SH_IS;
drivers/iommu/io-pgtable-arm.c
535
pte |= ARM_LPAE_PTE_SH_OS;
drivers/iommu/io-pgtable-arm.c
538
pte |= ARM_LPAE_PTE_XN;
drivers/iommu/io-pgtable-arm.c
541
pte |= ARM_LPAE_PTE_NS;
drivers/iommu/io-pgtable-arm.c
544
pte |= ARM_LPAE_PTE_AF;
drivers/iommu/io-pgtable-arm.c
546
return pte;
drivers/iommu/io-pgtable-arm.c
603
arm_lpae_iopte pte = *ptep++;
drivers/iommu/io-pgtable-arm.c
605
if (!pte || iopte_leaf(pte, lvl, data->iop.fmt))
drivers/iommu/io-pgtable-arm.c
608
__arm_lpae_free_pgtable(data, lvl + 1, iopte_deref(pte, data));
drivers/iommu/io-pgtable-arm.c
627
arm_lpae_iopte pte;
drivers/iommu/io-pgtable-arm.c
637
pte = READ_ONCE(*ptep);
drivers/iommu/io-pgtable-arm.c
638
if (!pte) {
drivers/iommu/io-pgtable-arm.c
650
pte = READ_ONCE(ptep[i]);
drivers/iommu/io-pgtable-arm.c
651
if (!pte) {
drivers/iommu/io-pgtable-arm.c
656
if (!iopte_leaf(pte, lvl, iop->fmt)) {
drivers/iommu/io-pgtable-arm.c
662
__arm_lpae_free_pgtable(data, lvl + 1, iopte_deref(pte, data));
drivers/iommu/io-pgtable-arm.c
674
} else if (iopte_leaf(pte, lvl, iop->fmt)) {
drivers/iommu/io-pgtable-arm.c
680
ptep = iopte_deref(pte, data);
drivers/iommu/io-pgtable-arm.c
721
arm_lpae_iopte pte;
drivers/iommu/io-pgtable-arm.c
729
data->pte = *ptep;
drivers/iommu/io-pgtable-arm.c
752
return iopte_to_paddr(d.pte, data) | iova;
drivers/iommu/io-pgtable-arm.c
782
arm_lpae_iopte pte = READ_ONCE(*ptep);
drivers/iommu/io-pgtable-arm.c
789
if (iopte_leaf(pte, lvl, iop->fmt)) {
drivers/iommu/io-pgtable-arm.c
794
if (!iopte_table(pte, lvl)) {
drivers/iommu/io-pgtable-arm.c
798
ptep = iopte_deref(pte, data);
drivers/iommu/io-pgtable-dart.c
103
return pte & APPLE_DART1_PADDR_MASK;
drivers/iommu/io-pgtable-dart.c
106
paddr = pte & APPLE_DART2_PADDR_MASK;
drivers/iommu/io-pgtable-dart.c
118
dart_iopte pte = prot;
drivers/iommu/io-pgtable-dart.c
129
pte |= FIELD_PREP(APPLE_DART_PTE_SUBPAGE_START, 0);
drivers/iommu/io-pgtable-dart.c
130
pte |= FIELD_PREP(APPLE_DART_PTE_SUBPAGE_END, 0xfff);
drivers/iommu/io-pgtable-dart.c
132
pte |= APPLE_DART_PTE_VALID;
drivers/iommu/io-pgtable-dart.c
135
ptep[i] = pte | paddr_to_iopte(paddr + i * sz, data);
drivers/iommu/io-pgtable-dart.c
176
dart_iopte pte, *ptep;
drivers/iommu/io-pgtable-dart.c
189
pte = READ_ONCE(*ptep);
drivers/iommu/io-pgtable-dart.c
192
if (!pte)
drivers/iommu/io-pgtable-dart.c
196
ptep = iopte_deref(pte, data);
drivers/iommu/io-pgtable-dart.c
205
dart_iopte pte = 0;
drivers/iommu/io-pgtable-dart.c
208
pte |= APPLE_DART1_PTE_PROT_SP_DIS;
drivers/iommu/io-pgtable-dart.c
210
pte |= APPLE_DART1_PTE_PROT_NO_WRITE;
drivers/iommu/io-pgtable-dart.c
212
pte |= APPLE_DART1_PTE_PROT_NO_READ;
drivers/iommu/io-pgtable-dart.c
216
pte |= APPLE_DART2_PTE_PROT_NO_WRITE;
drivers/iommu/io-pgtable-dart.c
218
pte |= APPLE_DART2_PTE_PROT_NO_READ;
drivers/iommu/io-pgtable-dart.c
220
pte |= APPLE_DART2_PTE_PROT_NO_CACHE;
drivers/iommu/io-pgtable-dart.c
223
return pte;
drivers/iommu/io-pgtable-dart.c
234
dart_iopte pte, *cptep, *ptep;
drivers/iommu/io-pgtable-dart.c
255
pte = READ_ONCE(*ptep);
drivers/iommu/io-pgtable-dart.c
258
if (!pte) {
drivers/iommu/io-pgtable-dart.c
263
pte = dart_install_table(cptep, ptep, 0, data);
drivers/iommu/io-pgtable-dart.c
264
if (pte)
drivers/iommu/io-pgtable-dart.c
268
pte = READ_ONCE(*ptep);
drivers/iommu/io-pgtable-dart.c
271
ptep = iopte_deref(pte, data);
drivers/iommu/io-pgtable-dart.c
300
dart_iopte pte, *ptep;
drivers/iommu/io-pgtable-dart.c
318
pte = READ_ONCE(*ptep);
drivers/iommu/io-pgtable-dart.c
319
if (WARN_ON(!pte))
drivers/iommu/io-pgtable-dart.c
340
dart_iopte pte, *ptep;
drivers/iommu/io-pgtable-dart.c
350
pte = READ_ONCE(*ptep);
drivers/iommu/io-pgtable-dart.c
352
if (pte) {
drivers/iommu/io-pgtable-dart.c
354
return iopte_to_paddr(pte, data) | iova;
drivers/iommu/io-pgtable-dart.c
460
dart_iopte pte = *ptep++;
drivers/iommu/io-pgtable-dart.c
462
if (pte)
drivers/iommu/io-pgtable-dart.c
463
apple_dart_free_pgtables(data, iopte_deref(pte, data), level - 1);
drivers/iommu/io-pgtable-dart.c
67
#define iopte_deref(pte, d) __va(iopte_to_paddr(pte, d))
drivers/iommu/io-pgtable-dart.c
85
dart_iopte pte;
drivers/iommu/io-pgtable-dart.c
91
pte = paddr >> APPLE_DART2_PADDR_SHIFT;
drivers/iommu/io-pgtable-dart.c
92
pte &= APPLE_DART2_PADDR_MASK;
drivers/iommu/io-pgtable-dart.c
94
return pte;
drivers/iommu/io-pgtable-dart.c
97
static phys_addr_t iopte_to_paddr(dart_iopte pte,
drivers/iommu/omap-iommu.c
1602
u32 *pgd, *pte;
drivers/iommu/omap-iommu.c
1609
iopgtable_lookup_entry(oiommu, da, &pgd, &pte);
drivers/iommu/omap-iommu.c
1611
if (pte) {
drivers/iommu/omap-iommu.c
1612
if (iopte_is_small(*pte))
drivers/iommu/omap-iommu.c
1613
ret = omap_iommu_translate(*pte, da, IOPTE_MASK);
drivers/iommu/omap-iommu.c
1614
else if (iopte_is_large(*pte))
drivers/iommu/omap-iommu.c
1615
ret = omap_iommu_translate(*pte, da, IOLARGE_MASK);
drivers/iommu/omap-iommu.c
1617
dev_err(dev, "bogus pte 0x%x, da 0x%llx", *pte,
drivers/iommu/riscv/iommu-bits.h
693
u64 pte;
drivers/iommu/riscv/iommu.c
1085
#define _io_pte_present(pte) ((pte) & (_PAGE_PRESENT | _PAGE_PROT_NONE))
drivers/iommu/riscv/iommu.c
1086
#define _io_pte_leaf(pte) ((pte) & _PAGE_LEAF)
drivers/iommu/riscv/iommu.c
1087
#define _io_pte_none(pte) ((pte) == 0)
drivers/iommu/riscv/iommu.c
1091
unsigned long pte,
drivers/iommu/riscv/iommu.c
1097
if (!_io_pte_present(pte) || _io_pte_leaf(pte))
drivers/iommu/riscv/iommu.c
1100
ptr = (unsigned long *)pfn_to_virt(__page_val_to_pfn(pte));
drivers/iommu/riscv/iommu.c
1104
pte = READ_ONCE(ptr[i]);
drivers/iommu/riscv/iommu.c
1105
if (!_io_pte_none(pte) && cmpxchg_relaxed(ptr + i, pte, 0) == pte)
drivers/iommu/riscv/iommu.c
1106
riscv_iommu_pte_free(domain, pte, freelist);
drivers/iommu/riscv/iommu.c
1120
unsigned long pte, old;
drivers/iommu/riscv/iommu.c
1136
pte = READ_ONCE(*ptr);
drivers/iommu/riscv/iommu.c
1142
if (_io_pte_present(pte) && _io_pte_leaf(pte))
drivers/iommu/riscv/iommu.c
1148
if (_io_pte_none(pte)) {
drivers/iommu/riscv/iommu.c
1153
old = pte;
drivers/iommu/riscv/iommu.c
1154
pte = _io_pte_entry(virt_to_pfn(addr), _PAGE_TABLE);
drivers/iommu/riscv/iommu.c
1155
if (cmpxchg_relaxed(ptr, old, pte) != old) {
drivers/iommu/riscv/iommu.c
1160
ptr = (unsigned long *)pfn_to_virt(__page_val_to_pfn(pte));
drivers/iommu/riscv/iommu.c
1170
unsigned long pte;
drivers/iommu/riscv/iommu.c
1177
pte = READ_ONCE(*ptr);
drivers/iommu/riscv/iommu.c
1178
if (_io_pte_present(pte) && _io_pte_leaf(pte)) {
drivers/iommu/riscv/iommu.c
1182
if (_io_pte_none(pte))
drivers/iommu/riscv/iommu.c
1184
ptr = (unsigned long *)pfn_to_virt(__page_val_to_pfn(pte));
drivers/iommu/riscv/iommu.c
1198
unsigned long pte, old, pte_prot;
drivers/iommu/riscv/iommu.c
1217
pte = _io_pte_entry(phys_to_pfn(phys), pte_prot);
drivers/iommu/riscv/iommu.c
1218
if (cmpxchg_relaxed(ptr, old, pte) != old)
drivers/iommu/rockchip-iommu.c
264
static inline bool rk_pte_is_page_valid(u32 pte)
drivers/iommu/rockchip-iommu.c
266
return pte & RK_PTE_PAGE_VALID;
drivers/iommu/rockchip-iommu.c
300
static u32 rk_mk_pte_invalid(u32 pte)
drivers/iommu/rockchip-iommu.c
302
return pte & ~RK_PTE_PAGE_VALID;
drivers/iommu/rockchip-iommu.c
545
u32 pte = 0;
drivers/iommu/rockchip-iommu.c
565
pte = *pte_addr;
drivers/iommu/rockchip-iommu.c
567
if (!rk_pte_is_page_valid(pte))
drivers/iommu/rockchip-iommu.c
570
page_addr_phys = rk_ops->pt_address(pte) + page_offset;
drivers/iommu/rockchip-iommu.c
571
page_flags = pte & RK_PTE_PAGE_FLAGS_MASK;
drivers/iommu/rockchip-iommu.c
578
rk_dte_is_pt_valid(dte), &pte_addr_phys, pte,
drivers/iommu/rockchip-iommu.c
579
rk_pte_is_page_valid(pte), &page_addr_phys, page_flags);
drivers/iommu/rockchip-iommu.c
657
u32 dte, pte;
drivers/iommu/rockchip-iommu.c
668
pte = page_table[rk_iova_pte_index(iova)];
drivers/iommu/rockchip-iommu.c
669
if (!rk_pte_is_page_valid(pte))
drivers/iommu/rockchip-iommu.c
672
phys = rk_ops->pt_address(pte) + rk_iova_page_offset(iova);
drivers/iommu/rockchip-iommu.c
765
u32 pte = pte_addr[pte_count];
drivers/iommu/rockchip-iommu.c
766
if (!rk_pte_is_page_valid(pte))
drivers/iommu/rockchip-iommu.c
769
pte_addr[pte_count] = rk_mk_pte_invalid(pte);
drivers/iommu/rockchip-iommu.c
788
u32 pte = pte_addr[pte_count];
drivers/iommu/rockchip-iommu.c
790
if (rk_pte_is_page_valid(pte))
drivers/iommu/s390-iommu.c
1016
pte = READ_ONCE(pto[px]);
drivers/iommu/s390-iommu.c
1017
if (pt_entry_isvalid(pte))
drivers/iommu/s390-iommu.c
1018
phys = pte & ZPCI_PTE_ADDR_MASK;
drivers/iommu/s390-iommu.c
477
unsigned long pte;
drivers/iommu/s390-iommu.c
479
pte = READ_ONCE(*ptep);
drivers/iommu/s390-iommu.c
481
invalidate_pt_entry(&pte);
drivers/iommu/s390-iommu.c
483
set_pt_pfaa(&pte, page_addr);
drivers/iommu/s390-iommu.c
484
validate_pt_entry(&pte);
drivers/iommu/s390-iommu.c
488
entry_set_protected(&pte);
drivers/iommu/s390-iommu.c
490
entry_clr_protected(&pte);
drivers/iommu/s390-iommu.c
492
xchg(ptep, pte);
drivers/iommu/s390-iommu.c
994
unsigned long ste, pte, rte;
drivers/iommu/sun50i-iommu.c
257
static phys_addr_t sun50i_pte_get_page_address(u32 pte)
drivers/iommu/sun50i-iommu.c
259
return (phys_addr_t)pte & SUN50I_PTE_PAGE_ADDRESS_MASK;
drivers/iommu/sun50i-iommu.c
262
static enum sun50i_iommu_aci sun50i_get_pte_aci(u32 pte)
drivers/iommu/sun50i-iommu.c
264
return FIELD_GET(SUN50I_PTE_ACI_MASK, pte);
drivers/iommu/sun50i-iommu.c
267
static bool sun50i_pte_is_page_valid(u32 pte)
drivers/iommu/sun50i-iommu.c
269
return pte & SUN50I_PTE_PAGE_VALID;
drivers/iommu/sun50i-iommu.c
668
u32 dte, pte;
drivers/iommu/sun50i-iommu.c
676
pte = page_table[sun50i_iova_get_pte_index(iova)];
drivers/iommu/sun50i-iommu.c
677
if (!sun50i_pte_is_page_valid(pte))
drivers/iommu/sun50i-iommu.c
680
return sun50i_pte_get_page_address(pte) +
drivers/iommu/tegra-smmu.c
667
u32 *pte, dma_addr_t pte_dma, u32 val)
drivers/iommu/tegra-smmu.c
670
unsigned long offset = SMMU_OFFSET_IN_PAGE(pte);
drivers/iommu/tegra-smmu.c
672
*pte = val;
drivers/iommu/tegra-smmu.c
729
u32 *pte;
drivers/iommu/tegra-smmu.c
735
pte = as_get_pte(as, iova, &pte_dma, pt);
drivers/iommu/tegra-smmu.c
736
if (!pte)
drivers/iommu/tegra-smmu.c
740
if (*pte == 0)
drivers/iommu/tegra-smmu.c
751
tegra_smmu_set_pte(as, iova, pte, pte_dma,
drivers/iommu/tegra-smmu.c
763
u32 *pte;
drivers/iommu/tegra-smmu.c
765
pte = tegra_smmu_pte_lookup(as, iova, &pte_dma);
drivers/iommu/tegra-smmu.c
766
if (!pte || !*pte)
drivers/iommu/tegra-smmu.c
769
tegra_smmu_set_pte(as, iova, pte, pte_dma, 0);
drivers/iommu/tegra-smmu.c
812
u32 *pte;
drivers/iommu/tegra-smmu.c
814
pte = tegra_smmu_pte_lookup(as, iova, &pte_dma);
drivers/iommu/tegra-smmu.c
815
if (!pte || !*pte)
drivers/iommu/tegra-smmu.c
818
pfn = *pte & as->smmu->pfn_mask;
drivers/md/dm-switch.c
180
region_table_slot_t pte;
drivers/md/dm-switch.c
184
pte = sctx->region_table[region_index];
drivers/md/dm-switch.c
185
pte &= ~((((region_table_slot_t)1 << sctx->region_table_entry_bits) - 1) << bit);
drivers/md/dm-switch.c
186
pte |= (region_table_slot_t)value << bit;
drivers/md/dm-switch.c
187
sctx->region_table[region_index] = pte;
drivers/misc/sgi-gru/grufault.c
212
pte_t pte;
drivers/misc/sgi-gru/grufault.c
231
pte = ptep_get((pte_t *)pmdp);
drivers/misc/sgi-gru/grufault.c
234
pte = *pte_offset_kernel(pmdp, vaddr);
drivers/misc/sgi-gru/grufault.c
236
if (unlikely(!pte_present(pte) ||
drivers/misc/sgi-gru/grufault.c
237
(write && (!pte_write(pte) || !pte_dirty(pte)))))
drivers/misc/sgi-gru/grufault.c
240
*paddr = pte_pfn(pte) << PAGE_SHIFT;
drivers/net/ethernet/mellanox/mlxsw/reg.h
897
MLXSW_ITEM32(reg, spvm, pte, 0x00, 30, 1);
drivers/staging/media/atomisp/include/mmu/isp_mmu.h
112
#define ISP_PTE_VALID(mmu, pte) \
drivers/staging/media/atomisp/include/mmu/isp_mmu.h
113
((pte) & ISP_PTE_VALID_MASK(mmu))
drivers/staging/media/atomisp/include/mmu/isp_mmu.h
95
unsigned int pte);
drivers/staging/media/atomisp/pci/mmu/isp_mmu.c
153
unsigned int isp_virt, unsigned int pte)
drivers/staging/media/atomisp/pci/mmu/isp_mmu.c
163
pte);
drivers/staging/media/atomisp/pci/mmu/isp_mmu.c
168
unsigned int isp_virt, unsigned int pte)
drivers/staging/media/atomisp/pci/mmu/isp_mmu.c
175
pte);
drivers/staging/media/atomisp/pci/mmu/isp_mmu.c
178
static void mmu_unmap_l1_pt_error(struct isp_mmu *mmu, unsigned int pte)
drivers/staging/media/atomisp/pci/mmu/isp_mmu.c
181
"L1PT = 0x%x\n", (unsigned int)pte);
drivers/staging/media/atomisp/pci/mmu/isp_mmu.c
194
unsigned int pte;
drivers/staging/media/atomisp/pci/mmu/isp_mmu.c
206
pte = atomisp_get_pte(l2_pt, idx);
drivers/staging/media/atomisp/pci/mmu/isp_mmu.c
208
if (ISP_PTE_VALID(mmu, pte)) {
drivers/staging/media/atomisp/pci/mmu/isp_mmu.c
210
l2_pt, idx, ptr, pte, phys);
drivers/staging/media/atomisp/pci/mmu/isp_mmu.c
218
pte = isp_pgaddr_to_pte_valid(mmu, phys);
drivers/staging/media/atomisp/pci/mmu/isp_mmu.c
220
atomisp_set_pte(l2_pt, idx, pte);
drivers/staging/media/atomisp/pci/mmu/isp_mmu.c
358
unsigned int pte;
drivers/staging/media/atomisp/pci/mmu/isp_mmu.c
369
pte = atomisp_get_pte(l2_pt, idx);
drivers/staging/media/atomisp/pci/mmu/isp_mmu.c
371
if (!ISP_PTE_VALID(mmu, pte))
drivers/staging/media/atomisp/pci/mmu/isp_mmu.c
373
l2_pt, idx, ptr, pte);
drivers/staging/media/atomisp/pci/mmu/isp_mmu.c
531
unsigned int pte;
drivers/staging/media/atomisp/pci/mmu/isp_mmu.c
546
pte = atomisp_get_pte(l1_pt, idx);
drivers/staging/media/atomisp/pci/mmu/isp_mmu.c
548
if (ISP_PTE_VALID(mmu, pte)) {
drivers/staging/media/atomisp/pci/mmu/isp_mmu.c
549
l2_pt = isp_pte_to_pgaddr(mmu, pte);
drivers/staging/media/atomisp/pci/mmu/isp_mmu.c
57
unsigned int idx, unsigned int pte)
drivers/staging/media/atomisp/pci/mmu/isp_mmu.c
60
*(pt_virt + idx) = pte;
drivers/staging/media/atomisp/pci/mmu/isp_mmu.c
69
unsigned int pte)
drivers/staging/media/atomisp/pci/mmu/isp_mmu.c
71
return mmu->driver->pte_to_phys(mmu, pte);
drivers/staging/media/atomisp/pci/mmu/isp_mmu.c
77
unsigned int pte = mmu->driver->phys_to_pte(mmu, phys);
drivers/staging/media/atomisp/pci/mmu/isp_mmu.c
79
return (unsigned int)(pte | ISP_PTE_VALID_MASK(mmu));
drivers/staging/media/atomisp/pci/mmu/sh_mmu_mrfld.c
27
unsigned int pte)
drivers/staging/media/atomisp/pci/mmu/sh_mmu_mrfld.c
31
return (phys_addr_t)((pte & ~mask) << ISP_PAGE_OFFSET);
drivers/staging/media/atomisp/pci/mmu/sh_mmu_mrfld.c
37
unsigned int pte = sh_phys_to_pte(mmu, phys);
drivers/staging/media/atomisp/pci/mmu/sh_mmu_mrfld.c
39
return HOST_ADDRESS(pte);
drivers/staging/media/ipu3/ipu3-mmu.c
124
int pte;
drivers/staging/media/ipu3/ipu3-mmu.c
130
for (pte = 0; pte < IPU3_PT_PTES; pte++)
drivers/staging/media/ipu3/ipu3-mmu.c
131
pt[pte] = pteval;
drivers/staging/media/ipu3/ipu3-mmu.c
29
#define IPU3_PTE2ADDR(pte) ((phys_addr_t)(pte) << IPU3_PAGE_SHIFT)
drivers/target/sbp/sbp_target.c
1244
struct sbp_page_table_entry *pte;
drivers/target/sbp/sbp_target.c
1272
pte = req->pg_tbl;
drivers/target/sbp/sbp_target.c
1278
pte = NULL;
drivers/target/sbp/sbp_target.c
1290
offset = (u64)be16_to_cpu(pte->segment_base_hi) << 32 |
drivers/target/sbp/sbp_target.c
1291
be32_to_cpu(pte->segment_base_lo);
drivers/target/sbp/sbp_target.c
1292
length = be16_to_cpu(pte->segment_length);
drivers/target/sbp/sbp_target.c
1294
pte++;
drivers/xen/gntdev.c
304
static int find_grant_ptes(pte_t *pte, unsigned long addr, void *data)
drivers/xen/gntdev.c
313
pte_maddr = arbitrary_virt_to_machine(pte).maddr;
drivers/xen/privcmd.c
1654
static int is_mapped_fn(pte_t *pte, unsigned long addr, void *data)
drivers/xen/privcmd.c
1656
return pte_none(ptep_get(pte)) ? 0 : -EBUSY;
drivers/xen/xenbus/xenbus_client.c
752
static int map_ring_apply(pte_t *pte, unsigned long addr, void *data)
drivers/xen/xenbus/xenbus_client.c
756
info->phys_addrs[info->idx++] = arbitrary_virt_to_machine(pte).maddr;
drivers/xen/xlate_mmu.c
100
pte_t pte = pte_mkspecial(pfn_pte(page_to_pfn(page), info->prot));
drivers/xen/xlate_mmu.c
138
set_pte_at(info->vma->vm_mm, addr, ptep, pte);
drivers/xen/xlate_mmu.c
276
pte_t pte = pte_mkspecial(pfn_pte(page_to_pfn(page), r->prot));
drivers/xen/xlate_mmu.c
278
set_pte_at(r->mm, addr, ptep, pte);
fs/hugetlbfs/inode.c
350
pte_t *ptep, pte;
fs/hugetlbfs/inode.c
356
pte = huge_ptep_get(vma->vm_mm, addr, ptep);
fs/hugetlbfs/inode.c
357
if (huge_pte_none(pte) || !pte_present(pte))
fs/hugetlbfs/inode.c
360
if (pte_pfn(pte) == pfn)
fs/proc/task_mmu.c
1018
static void smaps_pte_entry(pte_t *pte, unsigned long addr,
fs/proc/task_mmu.c
1026
pte_t ptent = ptep_get(pte);
fs/proc/task_mmu.c
1112
pte_t *pte;
fs/proc/task_mmu.c
1122
pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl);
fs/proc/task_mmu.c
1123
if (!pte) {
fs/proc/task_mmu.c
1127
for (; addr != end; pte++, addr += PAGE_SIZE)
fs/proc/task_mmu.c
1128
smaps_pte_entry(pte, addr, walk);
fs/proc/task_mmu.c
1129
pte_unmap_unlock(pte - 1, ptl);
fs/proc/task_mmu.c
1232
static int smaps_hugetlb_range(pte_t *pte, unsigned long hmask,
fs/proc/task_mmu.c
1243
ptl = huge_pte_lock(hstate_vma(vma), walk->mm, pte);
fs/proc/task_mmu.c
1244
ptent = huge_ptep_get(walk->mm, addr, pte);
fs/proc/task_mmu.c
1258
hugetlb_pmd_shared(pte))
fs/proc/task_mmu.c
1600
static inline bool pte_is_pinned(struct vm_area_struct *vma, unsigned long addr, pte_t pte)
fs/proc/task_mmu.c
1604
if (!pte_write(pte))
fs/proc/task_mmu.c
1610
folio = vm_normal_folio(vma, addr, pte);
fs/proc/task_mmu.c
1617
unsigned long addr, pte_t *pte)
fs/proc/task_mmu.c
1627
pte_t ptent = ptep_get(pte);
fs/proc/task_mmu.c
1637
old_pte = ptep_modify_prot_start(vma, addr, pte);
fs/proc/task_mmu.c
1640
ptep_modify_prot_commit(vma, addr, pte, old_pte, ptent);
fs/proc/task_mmu.c
1643
set_pte_at(vma->vm_mm, addr, pte, ptent);
fs/proc/task_mmu.c
1685
pte_t *pte, ptent;
fs/proc/task_mmu.c
1710
pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl);
fs/proc/task_mmu.c
1711
if (!pte) {
fs/proc/task_mmu.c
1715
for (; addr != end; pte++, addr += PAGE_SIZE) {
fs/proc/task_mmu.c
1716
ptent = ptep_get(pte);
fs/proc/task_mmu.c
1719
clear_soft_dirty(vma, addr, pte);
fs/proc/task_mmu.c
1731
ptep_test_and_clear_young(vma, addr, pte);
fs/proc/task_mmu.c
1735
pte_unmap_unlock(pte - 1, ptl);
fs/proc/task_mmu.c
1935
struct vm_area_struct *vma, unsigned long addr, pte_t pte)
fs/proc/task_mmu.c
1941
if (pte_none(pte))
fs/proc/task_mmu.c
1944
if (pte_present(pte)) {
fs/proc/task_mmu.c
1946
frame = pte_pfn(pte);
fs/proc/task_mmu.c
1948
page = vm_normal_page(vma, addr, pte);
fs/proc/task_mmu.c
1949
if (pte_soft_dirty(pte))
fs/proc/task_mmu.c
1951
if (pte_uffd_wp(pte))
fs/proc/task_mmu.c
1956
if (pte_swp_soft_dirty(pte))
fs/proc/task_mmu.c
1958
if (pte_swp_uffd_wp(pte))
fs/proc/task_mmu.c
1960
entry = softleaf_from_pte(pte);
fs/proc/task_mmu.c
2085
pte_t *pte, *orig_pte;
fs/proc/task_mmu.c
2101
orig_pte = pte = pte_offset_map_lock(walk->mm, pmdp, addr, &ptl);
fs/proc/task_mmu.c
2102
if (!pte) {
fs/proc/task_mmu.c
2106
for (; addr < end; pte++, addr += PAGE_SIZE) {
fs/proc/task_mmu.c
2109
pme = pte_to_pagemap_entry(pm, vma, addr, ptep_get(pte));
fs/proc/task_mmu.c
2132
pte_t pte;
fs/proc/task_mmu.c
2138
pte = huge_ptep_get(walk->mm, addr, ptep);
fs/proc/task_mmu.c
2139
if (pte_present(pte)) {
fs/proc/task_mmu.c
2140
struct folio *folio = page_folio(pte_page(pte));
fs/proc/task_mmu.c
2149
if (huge_pte_uffd_wp(pte))
fs/proc/task_mmu.c
2154
frame = pte_pfn(pte) +
fs/proc/task_mmu.c
2156
} else if (pte_swp_uffd_wp_any(pte)) {
fs/proc/task_mmu.c
2345
unsigned long addr, pte_t pte)
fs/proc/task_mmu.c
2349
if (pte_none(pte))
fs/proc/task_mmu.c
2352
if (pte_present(pte)) {
fs/proc/task_mmu.c
2357
if (!pte_uffd_wp(pte))
fs/proc/task_mmu.c
2361
page = vm_normal_page(vma, addr, pte);
fs/proc/task_mmu.c
2366
if (is_zero_pfn(pte_pfn(pte)))
fs/proc/task_mmu.c
2368
if (pte_soft_dirty(pte))
fs/proc/task_mmu.c
2375
if (!pte_swp_uffd_wp_any(pte))
fs/proc/task_mmu.c
2378
entry = softleaf_from_pte(pte);
fs/proc/task_mmu.c
2386
if (pte_swp_soft_dirty(pte))
fs/proc/task_mmu.c
2394
unsigned long addr, pte_t *pte, pte_t ptent)
fs/proc/task_mmu.c
2399
old_pte = ptep_modify_prot_start(vma, addr, pte);
fs/proc/task_mmu.c
2401
ptep_modify_prot_commit(vma, addr, pte, old_pte, ptent);
fs/proc/task_mmu.c
2403
set_pte_at(vma->vm_mm, addr, pte,
fs/proc/task_mmu.c
2407
set_pte_at(vma->vm_mm, addr, pte, ptent);
fs/proc/task_mmu.c
2474
static unsigned long pagemap_hugetlb_category(pte_t pte)
fs/proc/task_mmu.c
2478
if (pte_none(pte))
fs/proc/task_mmu.c
2486
if (pte_present(pte)) {
fs/proc/task_mmu.c
2489
if (!huge_pte_uffd_wp(pte))
fs/proc/task_mmu.c
2491
if (!PageAnon(pte_page(pte)))
fs/proc/task_mmu.c
2493
if (is_zero_pfn(pte_pfn(pte)))
fs/proc/task_mmu.c
2495
if (pte_soft_dirty(pte))
fs/proc/task_mmu.c
2500
if (!pte_swp_uffd_wp_any(pte))
fs/proc/task_mmu.c
2502
if (pte_swp_soft_dirty(pte))
fs/proc/task_mmu.c
2740
pte_t *pte, *start_pte;
fs/proc/task_mmu.c
2749
start_pte = pte = pte_offset_map_lock(vma->vm_mm, pmd, start, &ptl);
fs/proc/task_mmu.c
2750
if (!pte) {
fs/proc/task_mmu.c
2759
for (addr = start; addr != end; pte++, addr += PAGE_SIZE) {
fs/proc/task_mmu.c
2760
pte_t ptent = ptep_get(pte);
fs/proc/task_mmu.c
2765
make_uffd_wp_pte(vma, addr, pte, ptent);
fs/proc/task_mmu.c
2776
for (addr = start; addr < end; pte++, addr += PAGE_SIZE) {
fs/proc/task_mmu.c
2778
pte_t ptent = ptep_get(pte);
fs/proc/task_mmu.c
2789
make_uffd_wp_pte(vma, addr, pte, ptent);
fs/proc/task_mmu.c
2797
for (addr = start; addr != end; pte++, addr += PAGE_SIZE) {
fs/proc/task_mmu.c
2798
pte_t ptent = ptep_get(pte);
fs/proc/task_mmu.c
2815
make_uffd_wp_pte(vma, addr, pte, ptent);
fs/proc/task_mmu.c
2842
pte_t pte;
fs/proc/task_mmu.c
2847
pte = huge_ptep_get(walk->mm, start, ptep);
fs/proc/task_mmu.c
2848
categories = p->cur_vma_category | pagemap_hugetlb_category(pte);
fs/proc/task_mmu.c
2859
pte = huge_ptep_get(walk->mm, start, ptep);
fs/proc/task_mmu.c
2860
categories = p->cur_vma_category | pagemap_hugetlb_category(pte);
fs/proc/task_mmu.c
2880
make_uffd_wp_huge_pte(vma, start, ptep, pte);
fs/proc/task_mmu.c
3169
static struct page *can_gather_numa_stats(pte_t pte, struct vm_area_struct *vma,
fs/proc/task_mmu.c
3175
if (!pte_present(pte))
fs/proc/task_mmu.c
3178
page = vm_normal_page(vma, addr, pte);
fs/proc/task_mmu.c
3225
pte_t *pte;
fs/proc/task_mmu.c
3240
orig_pte = pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl);
fs/proc/task_mmu.c
3241
if (!pte) {
fs/proc/task_mmu.c
3246
pte_t ptent = ptep_get(pte);
fs/proc/task_mmu.c
3252
} while (pte++, addr += PAGE_SIZE, addr != end);
fs/proc/task_mmu.c
3258
static int gather_hugetlb_stats(pte_t *pte, unsigned long hmask,
fs/proc/task_mmu.c
3266
ptl = huge_pte_lock(hstate_vma(walk->vma), walk->mm, pte);
fs/proc/task_mmu.c
3267
huge_pte = huge_ptep_get(walk->mm, addr, pte);
fs/proc/task_mmu.c
3281
static int gather_hugetlb_stats(pte_t *pte, unsigned long hmask,
fs/userfaultfd.c
235
pte_t *ptep, pte;
fs/userfaultfd.c
243
pte = huge_ptep_get(vma->vm_mm, vmf->address, ptep);
fs/userfaultfd.c
251
if (huge_pte_none(pte))
fs/userfaultfd.c
254
if (pte_is_uffd_marker(pte))
fs/userfaultfd.c
260
if (!huge_pte_write(pte) && (reason & VM_UFFD_WP))
fs/userfaultfd.c
293
pte_t *pte;
fs/userfaultfd.c
325
pte = pte_offset_map(pmd, address);
fs/userfaultfd.c
326
if (!pte)
fs/userfaultfd.c
333
ptent = ptep_get(pte);
fs/userfaultfd.c
351
pte_unmap(pte);
include/asm-generic/hugetlb.h
10
return pte_write(pte);
include/asm-generic/hugetlb.h
111
pte_t pte, int dirty)
include/asm-generic/hugetlb.h
113
return ptep_set_access_flags(vma, addr, ptep, pte, dirty);
include/asm-generic/hugetlb.h
13
static inline unsigned long huge_pte_dirty(pte_t pte)
include/asm-generic/hugetlb.h
15
return pte_dirty(pte);
include/asm-generic/hugetlb.h
18
static inline pte_t huge_pte_mkwrite(pte_t pte)
include/asm-generic/hugetlb.h
20
return pte_mkwrite_novma(pte);
include/asm-generic/hugetlb.h
24
static inline pte_t huge_pte_wrprotect(pte_t pte)
include/asm-generic/hugetlb.h
26
return pte_wrprotect(pte);
include/asm-generic/hugetlb.h
30
static inline pte_t huge_pte_mkdirty(pte_t pte)
include/asm-generic/hugetlb.h
32
return pte_mkdirty(pte);
include/asm-generic/hugetlb.h
35
static inline pte_t huge_pte_modify(pte_t pte, pgprot_t newprot)
include/asm-generic/hugetlb.h
37
return pte_modify(pte, newprot);
include/asm-generic/hugetlb.h
41
static inline pte_t huge_pte_mkuffd_wp(pte_t pte)
include/asm-generic/hugetlb.h
43
return huge_pte_wrprotect(pte_mkuffd_wp(pte));
include/asm-generic/hugetlb.h
48
static inline pte_t huge_pte_clear_uffd_wp(pte_t pte)
include/asm-generic/hugetlb.h
50
return pte_clear_uffd_wp(pte);
include/asm-generic/hugetlb.h
55
static inline int huge_pte_uffd_wp(pte_t pte)
include/asm-generic/hugetlb.h
57
return pte_uffd_wp(pte);
include/asm-generic/hugetlb.h
71
pte_t *ptep, pte_t pte, unsigned long sz)
include/asm-generic/hugetlb.h
73
set_pte_at(mm, addr, ptep, pte);
include/asm-generic/hugetlb.h
8
static inline unsigned long huge_pte_write(pte_t pte)
include/asm-generic/hugetlb.h
94
static inline int huge_pte_none(pte_t pte)
include/asm-generic/hugetlb.h
96
return pte_none(pte);
include/asm-generic/pgalloc.h
55
static inline void pte_free_kernel(struct mm_struct *mm, pte_t *pte)
include/asm-generic/pgalloc.h
57
pagetable_dtor_free(virt_to_ptdesc(pte));
include/asm-generic/pgtable-nopmd.h
38
#define pud_populate(mm, pmd, pte) do { } while (0)
include/asm-generic/pgtable_uffd.h
22
static __always_inline int pte_uffd_wp(pte_t pte)
include/asm-generic/pgtable_uffd.h
32
static __always_inline pte_t pte_mkuffd_wp(pte_t pte)
include/asm-generic/pgtable_uffd.h
34
return pte;
include/asm-generic/pgtable_uffd.h
42
static __always_inline pte_t pte_clear_uffd_wp(pte_t pte)
include/asm-generic/pgtable_uffd.h
44
return pte;
include/asm-generic/pgtable_uffd.h
52
static __always_inline pte_t pte_swp_mkuffd_wp(pte_t pte)
include/asm-generic/pgtable_uffd.h
54
return pte;
include/asm-generic/pgtable_uffd.h
57
static __always_inline int pte_swp_uffd_wp(pte_t pte)
include/asm-generic/pgtable_uffd.h
62
static __always_inline pte_t pte_swp_clear_uffd_wp(pte_t pte)
include/asm-generic/pgtable_uffd.h
64
return pte;
include/linux/hugetlb.h
1013
return pud_lockptr(mm, (pud_t *) pte);
include/linux/hugetlb.h
1015
return pmd_lockptr(mm, (pmd_t *) pte);
include/linux/hugetlb.h
1017
return ptep_lockptr(mm, pte);
include/linux/hugetlb.h
1061
pte_t old_pte, pte_t pte)
include/linux/hugetlb.h
1065
set_huge_pte_at(vma->vm_mm, addr, ptep, pte, psize);
include/linux/hugetlb.h
1258
struct mm_struct *mm, pte_t *pte)
include/linux/hugetlb.h
1286
pte_t *ptep, pte_t pte, unsigned long sz)
include/linux/hugetlb.h
1310
struct mm_struct *mm, pte_t *pte)
include/linux/hugetlb.h
1314
ptl = huge_pte_lockptr(h, mm, pte);
include/linux/hugetlb.h
1328
static inline bool hugetlb_pmd_shared(pte_t *pte)
include/linux/hugetlb.h
1330
return ptdesc_pmd_is_shared(virt_to_ptdesc(pte));
include/linux/hugetlb.h
1333
static inline bool hugetlb_pmd_shared(pte_t *pte)
include/linux/hugetlb.h
985
struct mm_struct *mm, pte_t *pte)
include/linux/ksm.h
130
static inline void ksm_might_unmap_zero_page(struct mm_struct *mm, pte_t pte)
include/linux/ksm.h
33
#define is_ksm_zero_pte(pte) (is_zero_pfn(pte_pfn(pte)) && pte_dirty(pte))
include/linux/ksm.h
43
static inline void ksm_might_unmap_zero_page(struct mm_struct *mm, pte_t pte)
include/linux/ksm.h
45
if (is_ksm_zero_pte(pte)) {
include/linux/leafops.h
525
static inline bool pte_is_marker(pte_t pte)
include/linux/leafops.h
527
return softleaf_is_marker(softleaf_from_pte(pte));
include/linux/leafops.h
537
static inline bool pte_is_uffd_wp_marker(pte_t pte)
include/linux/leafops.h
539
const softleaf_t entry = softleaf_from_pte(pte);
include/linux/leafops.h
55
static inline softleaf_t softleaf_from_pte(pte_t pte)
include/linux/leafops.h
554
static inline bool pte_is_uffd_marker(pte_t pte)
include/linux/leafops.h
556
const softleaf_t entry = softleaf_from_pte(pte);
include/linux/leafops.h
59
if (pte_present(pte) || pte_none(pte))
include/linux/leafops.h
62
pte = pte_swp_clear_flags(pte);
include/linux/leafops.h
63
arch_entry = __pte_to_swp_entry(pte);
include/linux/mm.h
1690
static inline pte_t maybe_mkwrite(pte_t pte, struct vm_area_struct *vma)
include/linux/mm.h
1693
pte = pte_mkwrite(pte, vma);
include/linux/mm.h
1694
return pte;
include/linux/mm.h
2825
pte_t pte);
include/linux/mm.h
2827
pte_t pte);
include/linux/mm.h
3041
pte_t pte);
include/linux/mm.h
3164
static inline int pte_special(pte_t pte)
include/linux/mm.h
3169
static inline pte_t pte_mkspecial(pte_t pte)
include/linux/mm.h
3171
return pte;
include/linux/mm.h
3477
static inline spinlock_t *ptep_lockptr(struct mm_struct *mm, pte_t *pte)
include/linux/mm.h
3481
return ptlock_ptr(virt_to_ptdesc(pte));
include/linux/mm.h
3508
static inline spinlock_t *ptep_lockptr(struct mm_struct *mm, pte_t *pte)
include/linux/mm.h
3565
#define pte_unmap_unlock(pte, ptl) do { \
include/linux/mm.h
3567
pte_unmap(pte); \
include/linux/mm.h
4242
typedef int (*pte_fn_t)(pte_t *pte, unsigned long addr, void *data);
include/linux/mm.h
727
pte_t *pte; /* Pointer to pte entry matching
include/linux/mm_inline.h
571
pte_t *pte, pte_t pteval)
include/linux/mm_inline.h
579
WARN_ON_ONCE(!pte_none(ptep_get(pte)));
include/linux/mm_inline.h
602
set_pte_at(vma->vm_mm, addr, pte,
include/linux/page_table_check.h
125
unsigned long addr, pte_t pte)
include/linux/page_table_check.h
141
pte_t pte, unsigned int nr)
include/linux/page_table_check.h
18
pte_t pte);
include/linux/page_table_check.h
24
pte_t *ptep, pte_t pte, unsigned int nr);
include/linux/page_table_check.h
50
unsigned long addr, pte_t pte)
include/linux/page_table_check.h
55
__page_table_check_pte_clear(mm, addr, pte);
include/linux/page_table_check.h
78
pte_t pte, unsigned int nr)
include/linux/page_table_check.h
83
__page_table_check_ptes_set(mm, addr, ptep, pte, nr);
include/linux/pagewalk.h
187
pte_t pte;
include/linux/pagewalk.h
79
int (*pte_entry)(pte_t *pte, unsigned long addr,
include/linux/pagewalk.h
83
int (*hugetlb_entry)(pte_t *pte, unsigned long hmask,
include/linux/pgtable.h
1018
static inline pte_t pte_mkwrite(pte_t pte, struct vm_area_struct *vma)
include/linux/pgtable.h
1020
return pte_mkwrite_novma(pte);
include/linux/pgtable.h
106
#define pte_unmap(pte) do { \
include/linux/pgtable.h
107
kunmap_local((pte)); \
include/linux/pgtable.h
1115
static inline pte_t pte_sw_mkyoung(pte_t pte)
include/linux/pgtable.h
1117
return pte;
include/linux/pgtable.h
115
static inline void pte_unmap(pte_t *pte)
include/linux/pgtable.h
1241
static inline int pte_unused(pte_t pte)
include/linux/pgtable.h
1248
#define pte_access_permitted(pte, write) \
include/linux/pgtable.h
1249
(pte_present(pte) && (!(write) || pte_write(pte)))
include/linux/pgtable.h
1305
pte_t pte, pte_t oldpte,
include/linux/pgtable.h
1322
pte_t pte, pte_t oldpte,
include/linux/pgtable.h
1327
pte_advance_pfn(pte, i),
include/linux/pgtable.h
1380
#define move_pte(pte, old_addr, new_addr) (pte)
include/linux/pgtable.h
1384
# define pte_accessible(mm, pte) ((void)(pte), 1)
include/linux/pgtable.h
1506
pte_t *ptep, pte_t pte)
include/linux/pgtable.h
1512
set_pte_at(vma->vm_mm, addr, ptep, pte);
include/linux/pgtable.h
1545
pte_t *ptep, pte_t old_pte, pte_t pte)
include/linux/pgtable.h
1547
__ptep_modify_prot_commit(vma, addr, ptep, pte);
include/linux/pgtable.h
1578
pte_t pte, tmp_pte;
include/linux/pgtable.h
1580
pte = ptep_modify_prot_start(vma, addr, ptep);
include/linux/pgtable.h
1586
pte = pte_mkdirty(pte);
include/linux/pgtable.h
1588
pte = pte_mkyoung(pte);
include/linux/pgtable.h
1590
return pte;
include/linux/pgtable.h
1616
pte_t *ptep, pte_t old_pte, pte_t pte, unsigned int nr)
include/linux/pgtable.h
1621
ptep_modify_prot_commit(vma, addr, ptep, old_pte, pte);
include/linux/pgtable.h
1625
pte = pte_next_pfn(pte);
include/linux/pgtable.h
1765
static inline int pte_soft_dirty(pte_t pte)
include/linux/pgtable.h
1775
static inline pte_t pte_mksoft_dirty(pte_t pte)
include/linux/pgtable.h
1777
return pte;
include/linux/pgtable.h
1785
static inline pte_t pte_clear_soft_dirty(pte_t pte)
include/linux/pgtable.h
1787
return pte;
include/linux/pgtable.h
1795
static inline pte_t pte_swp_mksoft_dirty(pte_t pte)
include/linux/pgtable.h
1797
return pte;
include/linux/pgtable.h
1800
static inline int pte_swp_soft_dirty(pte_t pte)
include/linux/pgtable.h
1805
static inline pte_t pte_swp_clear_soft_dirty(pte_t pte)
include/linux/pgtable.h
1807
return pte;
include/linux/pgtable.h
2017
static inline int pte_protnone(pte_t pte)
include/linux/pgtable.h
376
static inline unsigned int pte_batch_hint(pte_t *ptep, pte_t pte)
include/linux/pgtable.h
383
static inline pte_t pte_advance_pfn(pte_t pte, unsigned long nr)
include/linux/pgtable.h
385
return __pte(pte_val(pte) + (nr << PFN_PTE_SHIFT));
include/linux/pgtable.h
389
#define pte_next_pfn(pte) pte_advance_pfn(pte, 1)
include/linux/pgtable.h
411
pte_t *ptep, pte_t pte, unsigned int nr)
include/linux/pgtable.h
413
page_table_check_ptes_set(mm, addr, ptep, pte, nr);
include/linux/pgtable.h
416
set_pte(ptep, pte);
include/linux/pgtable.h
420
pte = pte_next_pfn(pte);
include/linux/pgtable.h
424
#define set_pte_at(mm, addr, ptep, pte) set_ptes(mm, addr, ptep, pte, 1)
include/linux/pgtable.h
498
pte_t pte = ptep_get(ptep);
include/linux/pgtable.h
500
if (!pte_young(pte))
include/linux/pgtable.h
503
set_pte_at(vma->vm_mm, address, ptep, pte_mkold(pte));
include/linux/pgtable.h
593
pte_t pte)
include/linux/pgtable.h
616
pte_t pte = ptep_get(ptep);
include/linux/pgtable.h
618
page_table_check_pte_clear(mm, address, pte);
include/linux/pgtable.h
619
return pte;
include/linux/pgtable.h
646
pte_t pte;
include/linux/pgtable.h
652
pte = ptep_get_and_clear(vma->vm_mm, addr, ptep);
include/linux/pgtable.h
654
pte = pte_mkold(pte);
include/linux/pgtable.h
656
pte = pte_mkclean(pte);
include/linux/pgtable.h
657
set_pte_at(vma->vm_mm, addr, ptep, pte);
include/linux/pgtable.h
670
pte_t pte = ptep_get(ptep);
include/linux/pgtable.h
677
page_table_check_pte_clear(mm, addr, pte);
include/linux/pgtable.h
711
pte_t pte;
include/linux/pgtable.h
714
pte.pte_low = ptep->pte_low;
include/linux/pgtable.h
716
pte.pte_high = ptep->pte_high;
include/linux/pgtable.h
718
} while (unlikely(pte.pte_low != ptep->pte_low));
include/linux/pgtable.h
720
return pte;
include/linux/pgtable.h
844
pte_t pte, tmp_pte;
include/linux/pgtable.h
846
pte = ptep_get_and_clear_full(mm, addr, ptep, full);
include/linux/pgtable.h
852
pte = pte_mkdirty(pte);
include/linux/pgtable.h
854
pte = pte_mkyoung(pte);
include/linux/pgtable.h
856
return pte;
include/linux/ptdump.h
14
void (*note_page_pte)(struct ptdump_state *st, unsigned long addr, pte_t pte);
include/linux/ptdump.h
20
void (*effective_prot_pte)(struct ptdump_state *st, pte_t pte);
include/linux/rmap.h
871
pte_t *pte;
include/linux/rmap.h
889
if (pvmw->pte && !is_vm_hugetlb_page(pvmw->vma))
include/linux/rmap.h
890
pte_unmap(pvmw->pte);
include/linux/rmap.h
907
WARN_ON_ONCE(!pvmw->pmd && !pvmw->pte);
include/linux/rmap.h
916
pvmw->pte = NULL;
include/linux/swapops.h
218
extern void migration_entry_wait_huge(struct vm_area_struct *vma, unsigned long addr, pte_t *pte);
include/linux/swapops.h
238
unsigned long addr, pte_t *pte) { }
include/linux/swapops.h
70
static inline pte_t pte_swp_clear_flags(pte_t pte)
include/linux/swapops.h
72
if (pte_swp_exclusive(pte))
include/linux/swapops.h
73
pte = pte_swp_clear_exclusive(pte);
include/linux/swapops.h
74
if (pte_swp_soft_dirty(pte))
include/linux/swapops.h
75
pte = pte_swp_clear_soft_dirty(pte);
include/linux/swapops.h
76
if (pte_swp_uffd_wp(pte))
include/linux/swapops.h
77
pte = pte_swp_clear_uffd_wp(pte);
include/linux/swapops.h
78
return pte;
include/linux/userfaultfd_k.h
195
pte_t pte)
include/linux/userfaultfd_k.h
197
return userfaultfd_wp(vma) && pte_uffd_wp(pte);
include/linux/userfaultfd_k.h
314
static inline bool pte_swp_uffd_wp_any(pte_t pte)
include/linux/userfaultfd_k.h
319
if (pte_present(pte))
include/linux/userfaultfd_k.h
322
if (pte_swp_uffd_wp(pte))
include/linux/userfaultfd_k.h
325
if (pte_is_uffd_wp_marker(pte))
include/linux/userfaultfd_k.h
368
pte_t pte)
include/linux/userfaultfd_k.h
463
static inline bool pte_swp_uffd_wp_any(pte_t pte)
include/trace/events/migrate.h
115
TP_PROTO(unsigned long addr, unsigned long pte, int order),
include/trace/events/migrate.h
117
TP_ARGS(addr, pte, order),
include/trace/events/migrate.h
121
__field(unsigned long, pte)
include/trace/events/migrate.h
127
__entry->pte = pte;
include/trace/events/migrate.h
131
TP_printk("addr=%lx, pte=%lx order=%d", __entry->addr, __entry->pte, __entry->order)
include/trace/events/migrate.h
135
TP_PROTO(unsigned long addr, unsigned long pte, int order),
include/trace/events/migrate.h
136
TP_ARGS(addr, pte, order)
include/trace/events/migrate.h
140
TP_PROTO(unsigned long addr, unsigned long pte, int order),
include/trace/events/migrate.h
141
TP_ARGS(addr, pte, order)
include/trace/events/thp.h
14
TP_PROTO(unsigned long addr, unsigned long pte),
include/trace/events/thp.h
15
TP_ARGS(addr, pte),
include/trace/events/thp.h
18
__field(unsigned long, pte)
include/trace/events/thp.h
23
__entry->pte = pte;
include/trace/events/thp.h
26
TP_printk("Set page table entry with 0x%lx with 0x%lx", __entry->addr, __entry->pte)
include/trace/events/thp.h
41
TP_PROTO(unsigned long addr, unsigned long pte, unsigned long clr, unsigned long set),
include/trace/events/thp.h
42
TP_ARGS(addr, pte, clr, set),
include/trace/events/thp.h
45
__field(unsigned long, pte)
include/trace/events/thp.h
52
__entry->pte = pte;
include/trace/events/thp.h
58
TP_printk("hugepage update at addr 0x%lx and pte = 0x%lx clr = 0x%lx, set = 0x%lx", __entry->addr, __entry->pte, __entry->clr, __entry->set)
include/trace/events/xen.h
142
__entry->pteval = pteval.pte),
include/trace/events/xen.h
271
__entry->pteval = pteval.pte),
kernel/bpf/arena.c
121
static int apply_range_set_cb(pte_t *pte, unsigned long addr, void *data)
kernel/bpf/arena.c
129
if (unlikely(!pte_none(ptep_get(pte))))
kernel/bpf/arena.c
137
set_pte_at(&init_mm, addr, pte, mk_pte(page, PAGE_KERNEL));
kernel/bpf/arena.c
147
static int apply_range_clear_cb(pte_t *pte, unsigned long addr, void *free_pages)
kernel/bpf/arena.c
153
old_pte = ptep_get(pte);
kernel/bpf/arena.c
161
pte_clear(&init_mm, addr, pte);
kernel/bpf/arena.c
248
pte_t pte;
kernel/bpf/arena.c
250
pte = ptep_get(ptep);
kernel/bpf/arena.c
251
if (!pte_present(pte)) /* sanity check */
kernel/bpf/arena.c
253
page = pte_page(pte);
kernel/events/core.c
8420
pte_t *ptep, pte;
kernel/events/core.c
8459
pte = ptep_get_lockless(ptep);
kernel/events/core.c
8460
if (pte_present(pte))
kernel/events/core.c
8461
size = __pte_leaf_size(pmd, pte);
kernel/events/uprobes.c
419
if (!pte_write(fw->pte)) {
kernel/events/uprobes.c
422
if (unlikely(userfaultfd_pte_wp(vma, fw->pte)))
kernel/events/uprobes.c
431
flush_cache_page(vma, vaddr, pte_pfn(fw->pte));
kernel/events/uprobes.c
432
fw->pte = ptep_clear_flush(vma, vaddr, fw->ptep);
kernel/events/uprobes.c
467
set_pte_at(vma->vm_mm, vaddr, fw->ptep, pte_mkdirty(fw->pte));
lib/test_hmm.c
1642
if (vmf->pte) {
mm/damon/ops-common.c
163
if (pvmw.pte)
mm/damon/ops-common.c
164
damon_ptep_mkold(pvmw.pte, vma, addr);
mm/damon/ops-common.c
196
pte_t pte;
mm/damon/ops-common.c
201
if (pvmw.pte) {
mm/damon/ops-common.c
202
pte = ptep_get(pvmw.pte);
mm/damon/ops-common.c
209
*accessed = (pte_present(pte) && pte_young(pte)) ||
mm/damon/ops-common.c
44
void damon_ptep_mkold(pte_t *pte, struct vm_area_struct *vma, unsigned long addr)
mm/damon/ops-common.c
46
pte_t pteval = ptep_get(pte);
mm/damon/ops-common.c
66
young |= ptep_test_and_clear_young(vma, addr, pte);
mm/damon/ops-common.h
12
void damon_ptep_mkold(pte_t *pte, struct vm_area_struct *vma, unsigned long addr);
mm/damon/vaddr.c
309
pte_t *pte;
mm/damon/vaddr.c
322
pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl);
mm/damon/vaddr.c
323
if (!pte)
mm/damon/vaddr.c
325
if (!pte_present(ptep_get(pte)))
mm/damon/vaddr.c
327
damon_ptep_mkold(pte, walk->vma, addr);
mm/damon/vaddr.c
329
pte_unmap_unlock(pte, ptl);
mm/damon/vaddr.c
334
static void damon_hugetlb_mkold(pte_t *pte, struct mm_struct *mm,
mm/damon/vaddr.c
338
pte_t entry = huge_ptep_get(mm, addr, pte);
mm/damon/vaddr.c
347
set_huge_pte_at(mm, addr, pte, entry, psize);
mm/damon/vaddr.c
361
static int damon_mkold_hugetlb_entry(pte_t *pte, unsigned long hmask,
mm/damon/vaddr.c
369
ptl = huge_pte_lock(h, walk->mm, pte);
mm/damon/vaddr.c
370
entry = huge_ptep_get(walk->mm, addr, pte);
mm/damon/vaddr.c
374
damon_hugetlb_mkold(pte, walk->mm, walk->vma, addr);
mm/damon/vaddr.c
434
pte_t *pte;
mm/damon/vaddr.c
461
pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl);
mm/damon/vaddr.c
462
if (!pte)
mm/damon/vaddr.c
464
ptent = ptep_get(pte);
mm/damon/vaddr.c
475
pte_unmap_unlock(pte, ptl);
mm/damon/vaddr.c
480
static int damon_young_hugetlb_entry(pte_t *pte, unsigned long hmask,
mm/damon/vaddr.c
490
ptl = huge_pte_lock(h, walk->mm, pte);
mm/damon/vaddr.c
491
entry = huge_ptep_get(walk->mm, addr, pte);
mm/damon/vaddr.c
709
pte_t *start_pte, *pte, ptent;
mm/damon/vaddr.c
732
start_pte = pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl);
mm/damon/vaddr.c
733
if (!pte)
mm/damon/vaddr.c
736
for (; addr < next; pte += nr, addr += nr * PAGE_SIZE) {
mm/damon/vaddr.c
738
ptent = ptep_get(pte);
mm/damon/vaddr.c
745
if (damos_va_filter_out(s, folio, walk->vma, addr, pte, NULL))
mm/damon/vaddr.c
872
pte_t *start_pte, *pte, ptent;
mm/damon/vaddr.c
897
start_pte = pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl);
mm/damon/vaddr.c
901
for (; addr < next; pte += nr, addr += nr * PAGE_SIZE) {
mm/damon/vaddr.c
903
ptent = ptep_get(pte);
mm/damon/vaddr.c
913
if (!damos_va_filter_out(s, folio, vma, addr, pte, NULL))
mm/debug_vm_pgtable.c
100
WARN_ON(!pte_write(pte_mkwrite(pte_wrprotect(pte), args->vma)));
mm/debug_vm_pgtable.c
101
WARN_ON(pte_young(pte_mkold(pte_mkyoung(pte))));
mm/debug_vm_pgtable.c
102
WARN_ON(pte_dirty(pte_mkclean(pte_mkdirty(pte))));
mm/debug_vm_pgtable.c
103
WARN_ON(pte_write(pte_wrprotect(pte_mkwrite(pte, args->vma))));
mm/debug_vm_pgtable.c
104
WARN_ON(pte_dirty(pte_wrprotect(pte_mkclean(pte))));
mm/debug_vm_pgtable.c
105
WARN_ON(!pte_dirty(pte_wrprotect(pte_mkdirty(pte))));
mm/debug_vm_pgtable.c
107
WARN_ON(!pte_dirty(pte_mkwrite_novma(pte_mkdirty(pte))));
mm/debug_vm_pgtable.c
108
WARN_ON(pte_dirty(pte_mkwrite_novma(pte_mkclean(pte))));
mm/debug_vm_pgtable.c
109
WARN_ON(!pte_write(pte_mkdirty(pte_mkwrite_novma(pte))));
mm/debug_vm_pgtable.c
110
WARN_ON(!pte_write(pte_mkwrite_novma(pte_wrprotect(pte))));
mm/debug_vm_pgtable.c
111
WARN_ON(pte_write(pte_wrprotect(pte_mkwrite_novma(pte))));
mm/debug_vm_pgtable.c
117
pte_t pte;
mm/debug_vm_pgtable.c
138
pte = pfn_pte(args->pte_pfn, args->page_prot);
mm/debug_vm_pgtable.c
139
set_pte_at(args->mm, args->vaddr, args->ptep, pte);
mm/debug_vm_pgtable.c
142
pte = ptep_get(args->ptep);
mm/debug_vm_pgtable.c
143
WARN_ON(pte_write(pte));
mm/debug_vm_pgtable.c
145
pte = ptep_get(args->ptep);
mm/debug_vm_pgtable.c
146
WARN_ON(!pte_none(pte));
mm/debug_vm_pgtable.c
148
pte = pfn_pte(args->pte_pfn, args->page_prot);
mm/debug_vm_pgtable.c
149
pte = pte_wrprotect(pte);
mm/debug_vm_pgtable.c
150
pte = pte_mkclean(pte);
mm/debug_vm_pgtable.c
151
set_pte_at(args->mm, args->vaddr, args->ptep, pte);
mm/debug_vm_pgtable.c
153
pte = pte_mkwrite(pte, args->vma);
mm/debug_vm_pgtable.c
154
pte = pte_mkdirty(pte);
mm/debug_vm_pgtable.c
155
ptep_set_access_flags(args->vma, args->vaddr, args->ptep, pte, 1);
mm/debug_vm_pgtable.c
156
pte = ptep_get(args->ptep);
mm/debug_vm_pgtable.c
157
WARN_ON(!(pte_write(pte) && pte_dirty(pte)));
mm/debug_vm_pgtable.c
159
pte = ptep_get(args->ptep);
mm/debug_vm_pgtable.c
160
WARN_ON(!pte_none(pte));
mm/debug_vm_pgtable.c
162
pte = pfn_pte(args->pte_pfn, args->page_prot);
mm/debug_vm_pgtable.c
163
pte = pte_mkyoung(pte);
mm/debug_vm_pgtable.c
164
set_pte_at(args->mm, args->vaddr, args->ptep, pte);
mm/debug_vm_pgtable.c
167
pte = ptep_get(args->ptep);
mm/debug_vm_pgtable.c
168
WARN_ON(pte_young(pte));
mm/debug_vm_pgtable.c
609
pte_t pte = pfn_pte(args->pte_pfn, args->page_prot);
mm/debug_vm_pgtable.c
626
set_pte_at(args->mm, args->vaddr, args->ptep, pte);
mm/debug_vm_pgtable.c
627
WARN_ON(pte_none(pte));
mm/debug_vm_pgtable.c
631
pte = ptep_get(args->ptep);
mm/debug_vm_pgtable.c
632
WARN_ON(!pte_none(pte));
mm/debug_vm_pgtable.c
662
pte_t pte = pfn_pte(args->fixed_pte_pfn, args->page_prot);
mm/debug_vm_pgtable.c
668
WARN_ON(!pte_special(pte_mkspecial(pte)));
mm/debug_vm_pgtable.c
673
pte_t pte = pfn_pte(args->fixed_pte_pfn, args->page_prot_none);
mm/debug_vm_pgtable.c
679
WARN_ON(!pte_protnone(pte));
mm/debug_vm_pgtable.c
680
WARN_ON(!pte_present(pte));
mm/debug_vm_pgtable.c
705
pte_t pte = pfn_pte(args->fixed_pte_pfn, args->page_prot);
mm/debug_vm_pgtable.c
711
WARN_ON(!pte_soft_dirty(pte_mksoft_dirty(pte)));
mm/debug_vm_pgtable.c
712
WARN_ON(pte_soft_dirty(pte_clear_soft_dirty(pte)));
mm/debug_vm_pgtable.c
717
pte_t pte;
mm/debug_vm_pgtable.c
724
pte = swp_entry_to_pte(args->swp_entry);
mm/debug_vm_pgtable.c
725
entry = softleaf_from_pte(pte);
mm/debug_vm_pgtable.c
728
WARN_ON(!pte_swp_soft_dirty(pte_swp_mksoft_dirty(pte)));
mm/debug_vm_pgtable.c
729
WARN_ON(pte_swp_soft_dirty(pte_swp_clear_soft_dirty(pte)));
mm/debug_vm_pgtable.c
777
pte_t pte;
mm/debug_vm_pgtable.c
782
pte = swp_entry_to_pte(entry);
mm/debug_vm_pgtable.c
783
softleaf = softleaf_from_pte(pte);
mm/debug_vm_pgtable.c
785
WARN_ON(pte_swp_exclusive(pte));
mm/debug_vm_pgtable.c
789
pte = pte_swp_mkexclusive(pte);
mm/debug_vm_pgtable.c
790
softleaf = softleaf_from_pte(pte);
mm/debug_vm_pgtable.c
792
WARN_ON(!pte_swp_exclusive(pte));
mm/debug_vm_pgtable.c
794
WARN_ON(pte_swp_soft_dirty(pte));
mm/debug_vm_pgtable.c
797
pte = pte_swp_clear_exclusive(pte);
mm/debug_vm_pgtable.c
798
softleaf = softleaf_from_pte(pte);
mm/debug_vm_pgtable.c
800
WARN_ON(pte_swp_exclusive(pte));
mm/debug_vm_pgtable.c
83
pte_t pte = pfn_pte(args->fixed_pte_pfn, prot);
mm/debug_vm_pgtable.c
887
pte_t pte;
mm/debug_vm_pgtable.c
890
pte = pfn_pte(args->fixed_pmd_pfn, args->page_prot);
mm/debug_vm_pgtable.c
891
pte = arch_make_huge_pte(pte, PMD_SHIFT, VM_ACCESS_FLAGS);
mm/debug_vm_pgtable.c
894
WARN_ON(!pte_huge(pte));
mm/debug_vm_pgtable.c
896
WARN_ON(!huge_pte_dirty(huge_pte_mkdirty(pte)));
mm/debug_vm_pgtable.c
897
WARN_ON(!huge_pte_write(huge_pte_mkwrite(huge_pte_wrprotect(pte))));
mm/debug_vm_pgtable.c
898
WARN_ON(huge_pte_write(huge_pte_wrprotect(huge_pte_mkwrite(pte))));
mm/debug_vm_pgtable.c
95
WARN_ON(pte_dirty(pte_wrprotect(pte)));
mm/debug_vm_pgtable.c
97
WARN_ON(!pte_same(pte, pte));
mm/debug_vm_pgtable.c
98
WARN_ON(!pte_young(pte_mkyoung(pte_mkold(pte))));
mm/debug_vm_pgtable.c
99
WARN_ON(!pte_dirty(pte_mkdirty(pte_mkclean(pte))));
mm/filemap.c
3757
pte_t *old_ptep = vmf->pte;
mm/filemap.c
3772
vmf->pte -= start;
mm/filemap.c
3797
if (!pte_none(ptep_get(&vmf->pte[count])))
mm/filemap.c
3814
vmf->pte += count;
mm/filemap.c
3828
vmf->pte = old_ptep;
mm/filemap.c
3855
if (!pte_none(ptep_get(vmf->pte)))
mm/filemap.c
3913
vmf->pte = pte_offset_map_lock(vma->vm_mm, vmf->pmd, addr, &vmf->ptl);
mm/filemap.c
3914
if (!vmf->pte) {
mm/filemap.c
3925
vmf->pte += xas.xa_index - last_pgoff;
mm/filemap.c
3941
pte_unmap_unlock(vmf->pte, vmf->ptl);
mm/gup.c
1038
pte_t *pte;
mm/gup.c
1057
pte = pte_offset_map(pmd, address);
mm/gup.c
1058
if (!pte)
mm/gup.c
1060
entry = ptep_get(pte);
mm/gup.c
1078
pte_unmap(pte);
mm/gup.c
2840
pte_t pte = ptep_get_lockless(ptep);
mm/gup.c
2851
if (pte_protnone(pte))
mm/gup.c
2854
if (!pte_access_permitted(pte, flags & FOLL_WRITE))
mm/gup.c
2857
if (pte_special(pte))
mm/gup.c
2861
VM_WARN_ON_ONCE(!pfn_valid(pte_pfn(pte)));
mm/gup.c
2862
page = pte_page(pte);
mm/gup.c
2869
unlikely(pte_val(pte) != pte_val(ptep_get(ptep)))) {
mm/gup.c
2879
if (!pte_write(pte) && gup_must_unshare(NULL, flags, page)) {
mm/gup.c
764
pte_t *pte, unsigned int flags)
mm/gup.c
767
pte_t orig_entry = ptep_get(pte);
mm/gup.c
775
set_pte_at(vma->vm_mm, address, pte, entry);
mm/gup.c
776
update_mmu_cache(vma, address, pte);
mm/gup.c
785
static inline bool can_follow_write_pte(pte_t pte, struct page *page,
mm/gup.c
790
if (pte_write(pte))
mm/gup.c
797
if (pte_needs_soft_dirty_wp(vma, pte))
mm/gup.c
799
return !userfaultfd_pte_wp(vma, pte);
mm/gup.c
809
pte_t *ptep, pte;
mm/gup.c
815
pte = ptep_get(ptep);
mm/gup.c
816
if (!pte_present(pte))
mm/gup.c
818
if (pte_protnone(pte) && !gup_can_follow_protnone(vma, flags))
mm/gup.c
821
page = vm_normal_page(vma, address, pte);
mm/gup.c
827
!can_follow_write_pte(pte, page, vma, flags)) {
mm/gup.c
839
if (is_zero_pfn(pte_pfn(pte))) {
mm/gup.c
840
page = pte_page(pte);
mm/gup.c
849
if (!pte_write(pte) && gup_must_unshare(vma, flags, page)) {
mm/gup.c
879
!pte_dirty(pte) && !folio_test_dirty(folio))
mm/gup.c
893
if (!pte_none(pte))
mm/hmm.c
228
pte_t pte)
mm/hmm.c
230
if (pte_none(pte) || !pte_present(pte) || pte_protnone(pte))
mm/hmm.c
232
return pte_write(pte) ? (HMM_PFN_VALID | HMM_PFN_WRITE) : HMM_PFN_VALID;
mm/hmm.c
243
pte_t pte = ptep_get(ptep);
mm/hmm.c
252
if (pte_none(pte) || pte_is_uffd_wp_marker(pte)) {
mm/hmm.c
260
if (!pte_present(pte)) {
mm/hmm.c
261
const softleaf_t entry = softleaf_from_pte(pte);
mm/hmm.c
303
cpu_flags = pte_to_hmm_pfn_flags(range, pte);
mm/hmm.c
313
if (!vm_normal_page(walk->vma, addr, pte) &&
mm/hmm.c
314
!is_zero_pfn(pte_pfn(pte))) {
mm/hmm.c
323
new_pfn_flags = pte_pfn(pte) | cpu_flags;
mm/hmm.c
543
static int hmm_vma_walk_hugetlb_entry(pte_t *pte, unsigned long hmask,
mm/hmm.c
557
ptl = huge_pte_lock(hstate_vma(vma), walk->mm, pte);
mm/hmm.c
558
entry = huge_ptep_get(walk->mm, addr, pte);
mm/huge_memory.c
1399
vmf->pte = NULL;
mm/huge_memory.c
2954
pte_t *pte;
mm/huge_memory.c
2970
pte = pte_offset_map(&_pmd, haddr);
mm/huge_memory.c
2971
VM_BUG_ON(!pte);
mm/huge_memory.c
2979
VM_BUG_ON(!pte_none(ptep_get(pte)));
mm/huge_memory.c
2980
set_pte_at(mm, addr, pte, entry);
mm/huge_memory.c
2981
pte++;
mm/huge_memory.c
2983
pte_unmap(pte - 1);
mm/huge_memory.c
2999
pte_t *pte;
mm/huge_memory.c
3177
pte = pte_offset_map(&_pmd, haddr);
mm/huge_memory.c
3178
VM_BUG_ON(!pte);
mm/huge_memory.c
3207
VM_WARN_ON(!pte_none(ptep_get(pte + i)));
mm/huge_memory.c
3208
set_pte_at(mm, addr, pte + i, entry);
mm/huge_memory.c
3234
VM_WARN_ON(!pte_none(ptep_get(pte + i)));
mm/huge_memory.c
3235
set_pte_at(mm, addr, pte + i, entry);
mm/huge_memory.c
3254
VM_WARN_ON(!pte_none(ptep_get(pte + i)));
mm/huge_memory.c
3256
set_ptes(mm, haddr, pte, entry, HPAGE_PMD_NR);
mm/huge_memory.c
3258
pte_unmap(pte);
mm/huge_memory.c
4874
if (!(pvmw->pmd && !pvmw->pte))
mm/huge_memory.c
4925
if (!(pvmw->pmd && !pvmw->pte))
mm/hugetlb.c
5068
pte_t pte;
mm/hugetlb.c
5080
pte = huge_ptep_get_and_clear(mm, old_addr, src_pte, sz);
mm/hugetlb.c
5082
if (need_clear_uffd_wp && pte_is_uffd_wp_marker(pte)) {
mm/hugetlb.c
5086
if (pte_present(pte))
mm/hugetlb.c
5087
pte = huge_pte_clear_uffd_wp(pte);
mm/hugetlb.c
5089
pte = pte_swp_clear_uffd_wp(pte);
mm/hugetlb.c
5091
set_huge_pte_at(mm, new_addr, dst_pte, pte, sz);
mm/hugetlb.c
5172
pte_t pte;
mm/hugetlb.c
5206
pte = huge_ptep_get(mm, address, ptep);
mm/hugetlb.c
5207
if (huge_pte_none(pte)) {
mm/hugetlb.c
5216
if (unlikely(!pte_present(pte))) {
mm/hugetlb.c
5223
if (pte_swp_uffd_wp_any(pte) &&
mm/hugetlb.c
5240
if (folio != page_folio(pte_page(pte))) {
mm/hugetlb.c
5251
folio = page_folio(pte_page(pte));
mm/hugetlb.c
5254
pte = huge_ptep_get_and_clear(mm, address, ptep, sz);
mm/hugetlb.c
5256
if (huge_pte_dirty(pte))
mm/hugetlb.c
5259
if (huge_pte_uffd_wp(pte) &&
mm/hugetlb.c
5447
pte_t pte = huge_ptep_get(mm, vmf->address, vmf->pte);
mm/hugetlb.c
5463
if (!unshare && huge_pte_uffd_wp(pte))
mm/hugetlb.c
5468
set_huge_ptep_writable(vma, vmf->address, vmf->pte);
mm/hugetlb.c
5472
old_folio = page_folio(pte_page(pte));
mm/hugetlb.c
5495
vmf->pte);
mm/hugetlb.c
5559
vmf->pte = hugetlb_walk(vma, vmf->address,
mm/hugetlb.c
5561
if (likely(vmf->pte &&
mm/hugetlb.c
5562
pte_same(huge_ptep_get(mm, vmf->address, vmf->pte), pte)))
mm/hugetlb.c
5599
vmf->pte = hugetlb_walk(vma, vmf->address, huge_page_size(h));
mm/hugetlb.c
5600
if (likely(vmf->pte && pte_same(huge_ptep_get(mm, vmf->address, vmf->pte), pte))) {
mm/hugetlb.c
5604
huge_ptep_clear_flush(vma, vmf->address, vmf->pte);
mm/hugetlb.c
5607
if (huge_pte_uffd_wp(pte))
mm/hugetlb.c
5609
set_huge_pte_at(mm, vmf->address, vmf->pte, newpte,
mm/hugetlb.c
5769
if (!hugetlb_pte_stable(h, mm, vmf->address, vmf->pte, vmf->orig_pte)) {
mm/hugetlb.c
5798
if (hugetlb_pte_stable(h, mm, vmf->address, vmf->pte, vmf->orig_pte))
mm/hugetlb.c
5846
if (!hugetlb_pte_stable(h, mm, vmf->address, vmf->pte, vmf->orig_pte)) {
mm/hugetlb.c
5870
vmf->ptl = huge_pte_lock(h, mm, vmf->pte);
mm/hugetlb.c
5873
if (!pte_same(huge_ptep_get(mm, vmf->address, vmf->pte), vmf->orig_pte))
mm/hugetlb.c
5887
set_huge_pte_at(mm, vmf->address, vmf->pte, new_pte, huge_page_size(h));
mm/hugetlb.c
6003
vmf.pte = huge_pte_alloc(mm, vma, vmf.address, huge_page_size(h));
mm/hugetlb.c
6004
if (!vmf.pte) {
mm/hugetlb.c
6010
vmf.orig_pte = huge_ptep_get(mm, vmf.address, vmf.pte);
mm/hugetlb.c
6050
migration_entry_wait_huge(vma, vmf.address, vmf.pte);
mm/hugetlb.c
6077
vmf.ptl = huge_pte_lock(h, mm, vmf.pte);
mm/hugetlb.c
6080
if (unlikely(!pte_same(vmf.orig_pte, huge_ptep_get(mm, vmf.address, vmf.pte))))
mm/hugetlb.c
6084
if (userfaultfd_wp(vma) && huge_pte_uffd_wp(huge_ptep_get(mm, vmf.address, vmf.pte)) &&
mm/hugetlb.c
6094
set_huge_pte_at(mm, vmf.address, vmf.pte, vmf.orig_pte,
mm/hugetlb.c
6122
if (huge_ptep_set_access_flags(vma, vmf.address, vmf.pte, vmf.orig_pte,
mm/hugetlb.c
6124
update_mmu_cache(vma, vmf.address, vmf.pte);
mm/hugetlb.c
6413
pte_t pte;
mm/hugetlb.c
6472
pte = huge_ptep_get(mm, address, ptep);
mm/hugetlb.c
6473
if (huge_pte_none(pte)) {
mm/hugetlb.c
6482
entry = softleaf_from_pte(pte);
mm/hugetlb.c
6487
pte_t newpte = pte;
mm/hugetlb.c
6504
if (!pte_same(pte, newpte))
mm/hugetlb.c
6506
} else if (unlikely(pte_is_marker(pte))) {
mm/hugetlb.c
6513
if (pte_is_uffd_wp_marker(pte) && uffd_wp_resolve)
mm/hugetlb.c
6521
pte = huge_pte_modify(old_pte, newprot);
mm/hugetlb.c
6522
pte = arch_make_huge_pte(pte, shift, vma->vm_flags);
mm/hugetlb.c
6524
pte = huge_pte_mkuffd_wp(pte);
mm/hugetlb.c
6526
pte = huge_pte_clear_uffd_wp(pte);
mm/hugetlb.c
6527
huge_ptep_modify_prot_commit(vma, address, ptep, old_pte, pte);
mm/hugetlb.c
6879
pte_t *pte;
mm/hugetlb.c
6910
pte = (pte_t *)pmd_alloc(mm, pud, addr);
mm/hugetlb.c
6912
return pte;
mm/hugetlb.c
7013
pte_t *pte = NULL;
mm/hugetlb.c
7022
pte = (pte_t *)pud;
mm/hugetlb.c
7026
pte = huge_pmd_share(mm, vma, addr, pud);
mm/hugetlb.c
7028
pte = (pte_t *)pmd_alloc(mm, pud, addr);
mm/hugetlb.c
7032
if (pte) {
mm/hugetlb.c
7033
pte_t pteval = ptep_get_lockless(pte);
mm/hugetlb.c
7038
return pte;
mm/hugetlb_vmemmap.c
139
static int vmemmap_pte_entry(pte_t *pte, unsigned long addr,
mm/hugetlb_vmemmap.c
149
vmemmap_walk->reuse_page = pte_page(ptep_get(pte));
mm/hugetlb_vmemmap.c
151
vmemmap_walk->remap_pte(pte, addr, vmemmap_walk);
mm/hugetlb_vmemmap.c
208
static void vmemmap_remap_pte(pte_t *pte, unsigned long addr,
mm/hugetlb_vmemmap.c
216
struct page *page = pte_page(ptep_get(pte));
mm/hugetlb_vmemmap.c
234
set_pte_at(&init_mm, addr, pte, entry);
mm/hugetlb_vmemmap.c
256
static void vmemmap_restore_pte(pte_t *pte, unsigned long addr,
mm/hugetlb_vmemmap.c
263
BUG_ON(pte_page(ptep_get(pte)) != walk->reuse_page);
mm/hugetlb_vmemmap.c
276
set_pte_at(&init_mm, addr, pte, mk_pte(page, pgprot));
mm/hugetlb_vmemmap.c
36
void (*remap_pte)(pte_t *pte, unsigned long addr,
mm/hugetlb_vmemmap.c
67
pte_t entry, *pte;
mm/hugetlb_vmemmap.c
71
pte = pte_offset_kernel(&__pmd, addr);
mm/hugetlb_vmemmap.c
72
set_pte_at(&init_mm, addr, pte, entry);
mm/internal.h
1652
static inline bool pte_needs_soft_dirty_wp(struct vm_area_struct *vma, pte_t pte)
mm/internal.h
1654
return vma_soft_dirty_enabled(vma) && !pte_soft_dirty(pte);
mm/internal.h
296
static inline pte_t __pte_batch_clear_ignored(pte_t pte, fpb_t flags)
mm/internal.h
299
pte = pte_mkclean(pte);
mm/internal.h
301
pte = pte_clear_soft_dirty(pte);
mm/internal.h
303
pte = pte_wrprotect(pte);
mm/internal.h
304
return pte_mkold(pte);
mm/internal.h
342
pte_t expected_pte, pte = *ptentp;
mm/internal.h
345
VM_WARN_ON_FOLIO(!pte_present(pte), folio);
mm/internal.h
347
VM_WARN_ON_FOLIO(page_folio(pfn_to_page(pte_pfn(pte))) != folio, folio);
mm/internal.h
357
folio_pfn(folio) + folio_nr_pages(folio) - pte_pfn(pte));
mm/internal.h
359
nr = pte_batch_hint(ptep, pte);
mm/internal.h
360
expected_pte = __pte_batch_clear_ignored(pte_advance_pfn(pte, nr), flags);
mm/internal.h
364
pte = ptep_get(ptep);
mm/internal.h
366
if (!pte_same(__pte_batch_clear_ignored(pte, flags), expected_pte))
mm/internal.h
370
any_writable |= pte_write(pte);
mm/internal.h
372
any_young |= pte_young(pte);
mm/internal.h
373
any_dirty |= pte_dirty(pte);
mm/internal.h
376
cur_nr = pte_batch_hint(ptep, pte);
mm/internal.h
392
unsigned int folio_pte_batch(struct folio *folio, pte_t *ptep, pte_t pte,
mm/internal.h
405
static inline pte_t pte_move_swp_offset(pte_t pte, long delta)
mm/internal.h
407
const softleaf_t entry = softleaf_from_pte(pte);
mm/internal.h
411
if (pte_swp_soft_dirty(pte))
mm/internal.h
413
if (pte_swp_exclusive(pte))
mm/internal.h
415
if (pte_swp_uffd_wp(pte))
mm/internal.h
429
static inline pte_t pte_next_swp_offset(pte_t pte)
mm/internal.h
431
return pte_move_swp_offset(pte, 1);
mm/internal.h
449
static inline int swap_pte_batch(pte_t *start_ptep, int max_nr, pte_t pte)
mm/internal.h
451
pte_t expected_pte = pte_next_swp_offset(pte);
mm/internal.h
453
const softleaf_t entry = softleaf_from_pte(pte);
mm/internal.h
464
pte = ptep_get(ptep);
mm/internal.h
466
if (!pte_same(pte, expected_pte))
mm/internal.h
468
entry = softleaf_from_pte(pte);
mm/internal.h
517
void pmd_install(struct mm_struct *mm, pmd_t *pmd, pgtable_t *pte);
mm/kasan/init.c
103
set_pte_at(&init_mm, addr, pte, zero_pte);
mm/kasan/init.c
105
pte = pte_offset_kernel(pmd, addr);
mm/kasan/init.c
286
pte_t *pte;
mm/kasan/init.c
290
pte = pte_start + i;
mm/kasan/init.c
291
if (!pte_none(ptep_get(pte)))
mm/kasan/init.c
344
static void kasan_remove_pte_table(pte_t *pte, unsigned long addr,
mm/kasan/init.c
350
for (; addr < end; addr = next, pte++) {
mm/kasan/init.c
355
ptent = ptep_get(pte);
mm/kasan/init.c
362
pte_clear(&init_mm, addr, pte);
mm/kasan/init.c
372
pte_t *pte;
mm/kasan/init.c
386
pte = pte_offset_kernel(pmd, addr);
mm/kasan/init.c
387
kasan_remove_pte_table(pte, addr, next);
mm/kasan/init.c
75
static inline bool kasan_early_shadow_page_entry(pte_t pte)
mm/kasan/init.c
77
return pte_page(pte) == virt_to_page(lm_alias(kasan_early_shadow_page));
mm/kasan/init.c
95
pte_t *pte = pte_offset_kernel(pmd, addr);
mm/kasan/shadow.c
192
pte_t *pte;
mm/kasan/shadow.c
209
pte = pte_offset_kernel(pmd, addr);
mm/kasan/shadow.c
210
return !pte_none(ptep_get(pte));
mm/kasan/shadow.c
305
pte_t pte;
mm/kasan/shadow.c
313
pte = pfn_pte(page_to_pfn(page), PAGE_KERNEL);
mm/kasan/shadow.c
317
set_pte_at(&init_mm, addr, ptep, pte);
mm/kasan/shadow.c
471
pte_t pte;
mm/kasan/shadow.c
477
pte = ptep_get(ptep);
mm/kasan/shadow.c
478
none = pte_none(pte);
mm/kasan/shadow.c
484
__free_page(pfn_to_page(pte_pfn(pte)));
mm/khugepaged.c
1000
if (!pte) {
mm/khugepaged.c
1007
vmf.orig_pte = ptep_get_lockless(pte);
mm/khugepaged.c
1012
vmf.pte = pte;
mm/khugepaged.c
1016
pte = NULL;
mm/khugepaged.c
1037
if (pte)
mm/khugepaged.c
1038
pte_unmap(pte);
mm/khugepaged.c
1083
pte_t *pte;
mm/khugepaged.c
1169
pte = pte_offset_map_lock(mm, &_pmd, address, &pte_ptl);
mm/khugepaged.c
1170
if (pte) {
mm/khugepaged.c
1171
result = __collapse_huge_page_isolate(vma, address, pte, cc,
mm/khugepaged.c
1179
if (pte)
mm/khugepaged.c
1180
pte_unmap(pte);
mm/khugepaged.c
1200
result = __collapse_huge_page_copy(pte, folio, pmd, _pmd,
mm/khugepaged.c
1203
pte_unmap(pte);
mm/khugepaged.c
1238
pte_t *pte, *_pte;
mm/khugepaged.c
1255
pte = pte_offset_map_lock(mm, pmd, start_addr, &ptl);
mm/khugepaged.c
1256
if (!pte) {
mm/khugepaged.c
1261
for (addr = start_addr, _pte = pte; _pte < pte + HPAGE_PMD_NR;
mm/khugepaged.c
1390
pte_unmap_unlock(pte, ptl);
mm/khugepaged.c
1475
pte_t *start_pte, *pte;
mm/khugepaged.c
1536
for (i = 0, addr = haddr, pte = start_pte;
mm/khugepaged.c
1537
i < HPAGE_PMD_NR; i++, addr += PAGE_SIZE, pte++) {
mm/khugepaged.c
1539
pte_t ptent = ptep_get(pte);
mm/khugepaged.c
1591
for (i = 0, addr = haddr, pte = start_pte; i < HPAGE_PMD_NR;
mm/khugepaged.c
1593
pte += nr_batch_ptes) {
mm/khugepaged.c
1596
pte_t ptent = ptep_get(pte);
mm/khugepaged.c
1617
nr_batch_ptes = folio_pte_batch(folio, pte, ptent, max_nr_batch_ptes);
mm/khugepaged.c
1624
clear_ptes(mm, addr, pte, nr_batch_ptes);
mm/khugepaged.c
336
static bool pte_none_or_zero(pte_t pte)
mm/khugepaged.c
338
if (pte_none(pte))
mm/khugepaged.c
340
return pte_present(pte) && is_zero_pfn(pte_pfn(pte));
mm/khugepaged.c
504
static void release_pte_pages(pte_t *pte, pte_t *_pte,
mm/khugepaged.c
509
while (--_pte >= pte) {
mm/khugepaged.c
532
unsigned long start_addr, pte_t *pte, struct collapse_control *cc,
mm/khugepaged.c
542
for (_pte = pte; _pte < pte + HPAGE_PMD_NR;
mm/khugepaged.c
664
release_pte_pages(pte, _pte, compound_pagelist);
mm/khugepaged.c
670
static void __collapse_huge_page_copy_succeeded(pte_t *pte,
mm/khugepaged.c
682
for (_pte = pte; _pte < pte + HPAGE_PMD_NR; _pte += nr_ptes,
mm/khugepaged.c
734
static void __collapse_huge_page_copy_failed(pte_t *pte,
mm/khugepaged.c
755
release_pte_pages(pte, pte + HPAGE_PMD_NR, compound_pagelist);
mm/khugepaged.c
773
static enum scan_result __collapse_huge_page_copy(pte_t *pte, struct folio *folio,
mm/khugepaged.c
785
pte_t pteval = ptep_get(pte + i);
mm/khugepaged.c
802
__collapse_huge_page_copy_succeeded(pte, vma, address, ptl,
mm/khugepaged.c
805
__collapse_huge_page_copy_failed(pte, pmd, orig_pmd, vma,
mm/khugepaged.c
982
pte_t *pte = NULL;
mm/khugepaged.c
994
if (!pte++) {
mm/khugepaged.c
999
pte = pte_offset_map_ro_nolock(mm, pmd, addr, &ptl);
mm/ksm.c
1296
if (WARN_ONCE(!pvmw.pte, "Unexpected PMD mapping?"))
mm/ksm.c
1299
entry = ptep_get(pvmw.pte);
mm/ksm.c
1326
entry = ptep_clear_flush(vma, pvmw.address, pvmw.pte);
mm/ksm.c
1332
set_pte_at(mm, pvmw.address, pvmw.pte, entry);
mm/ksm.c
1339
set_pte_at(mm, pvmw.address, pvmw.pte, entry);
mm/ksm.c
1350
set_pte_at(mm, pvmw.address, pvmw.pte, entry);
mm/ksm.c
2496
pte_t *start_ptep = NULL, *ptep, pte;
mm/ksm.c
2538
pte = ptep_get(ptep);
mm/ksm.c
2540
if (!pte_present(pte))
mm/ksm.c
2543
page = vm_normal_page(vma, addr, pte);
mm/ksm.c
629
pte_t pte = ptep_get(ptep);
mm/ksm.c
632
if (pte_present(pte)) {
mm/ksm.c
633
folio = vm_normal_folio(walk->vma, addr, pte);
mm/ksm.c
634
} else if (!pte_none(pte)) {
mm/ksm.c
635
const softleaf_t entry = softleaf_from_pte(pte);
mm/ksm.c
646
(pte_present(pte) && is_ksm_zero_pte(pte));
mm/madvise.c
1092
static int guard_install_pte_entry(pte_t *pte, unsigned long addr,
mm/madvise.c
1095
pte_t pteval = ptep_get(pte);
mm/madvise.c
1236
static int guard_remove_pte_entry(pte_t *pte, unsigned long addr,
mm/madvise.c
1239
pte_t ptent = ptep_get(pte);
mm/madvise.c
1243
pte_clear_not_present_full(walk->mm, addr, pte, false);
mm/madvise.c
1244
update_mmu_cache(walk->vma, addr, pte);
mm/madvise.c
195
pte_t pte;
mm/madvise.c
205
pte = ptep_get(ptep);
mm/madvise.c
206
entry = softleaf_from_pte(pte);
mm/madvise.c
361
pte_t *start_pte, *pte, ptent;
mm/madvise.c
450
start_pte = pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl);
mm/madvise.c
455
for (; addr < end; pte += nr, addr += nr * PAGE_SIZE) {
mm/madvise.c
457
ptent = ptep_get(pte);
mm/madvise.c
487
nr = madvise_folio_pte_batch(addr, end, folio, pte, &ptent);
mm/madvise.c
504
start_pte = pte =
mm/madvise.c
530
clear_young_dirty_ptes(vma, addr, pte, nr,
mm/madvise.c
532
tlb_remove_tlb_entries(tlb, pte, nr, addr);
mm/madvise.c
660
pte_t *start_pte, *pte, ptent;
mm/madvise.c
672
start_pte = pte = pte_offset_map_lock(mm, pmd, addr, &ptl);
mm/madvise.c
677
for (; addr != end; pte += nr, addr += PAGE_SIZE * nr) {
mm/madvise.c
679
ptent = ptep_get(pte);
mm/madvise.c
693
nr = swap_pte_batch(pte, max_nr, ptent);
mm/madvise.c
696
clear_not_present_full_ptes(mm, addr, pte, nr, tlb->fullmm);
mm/madvise.c
699
pte_clear_not_present_full(mm, addr, pte, tlb->fullmm);
mm/madvise.c
716
nr = madvise_folio_pte_batch(addr, end, folio, pte, &ptent);
mm/madvise.c
731
pte = pte_offset_map_lock(mm, pmd, addr, &ptl);
mm/madvise.c
732
start_pte = pte;
mm/madvise.c
767
clear_young_dirty_ptes(vma, addr, pte, nr, cydp_flags);
mm/madvise.c
768
tlb_remove_tlb_entries(tlb, pte, nr, addr);
mm/mapping_dirty_helpers.c
102
ptep_modify_prot_commit(walk->vma, addr, pte, old_pte, ptent);
mm/mapping_dirty_helpers.c
34
static int wp_pte(pte_t *pte, unsigned long addr, unsigned long end,
mm/mapping_dirty_helpers.c
38
pte_t ptent = ptep_get(pte);
mm/mapping_dirty_helpers.c
41
pte_t old_pte = ptep_modify_prot_start(walk->vma, addr, pte);
mm/mapping_dirty_helpers.c
44
ptep_modify_prot_commit(walk->vma, addr, pte, old_pte, ptent);
mm/mapping_dirty_helpers.c
89
static int clean_record_pte(pte_t *pte, unsigned long addr,
mm/mapping_dirty_helpers.c
94
pte_t ptent = ptep_get(pte);
mm/mapping_dirty_helpers.c
99
pte_t old_pte = ptep_modify_prot_start(walk->vma, addr, pte);
mm/memory-failure.c
345
pte_t *pte;
mm/memory-failure.c
365
pte = pte_offset_map(pmd, address);
mm/memory-failure.c
366
if (!pte)
mm/memory-failure.c
368
ptent = ptep_get(pte);
mm/memory-failure.c
371
pte_unmap(pte);
mm/memory-failure.c
691
static int check_hwpoisoned_entry(pte_t pte, unsigned long addr, short shift,
mm/memory-failure.c
698
if (pte_present(pte)) {
mm/memory-failure.c
699
pfn = pte_pfn(pte);
mm/memory-failure.c
701
const softleaf_t entry = softleaf_from_pte(pte);
mm/memory-failure.c
782
pte_t pte;
mm/memory-failure.c
786
pte = huge_ptep_get(walk->mm, addr, ptep);
mm/memory-failure.c
787
ret = check_hwpoisoned_entry(pte, addr, huge_page_shift(h),
mm/memory.c
1017
pte = swp_entry_to_pte(entry);
mm/memory.c
1019
pte = pte_swp_mkuffd_wp(pte);
mm/memory.c
1020
set_pte_at(src_mm, addr, src_pte, pte);
mm/memory.c
1042
pte = pte_swp_clear_uffd_wp(pte);
mm/memory.c
1043
set_pte_at(dst_mm, addr, dst_pte, pte);
mm/memory.c
1065
pte_t pte;
mm/memory.c
1086
pte = folio_mk_pte(new_folio, dst_vma->vm_page_prot);
mm/memory.c
1087
pte = maybe_mkwrite(pte_mkdirty(pte), dst_vma);
mm/memory.c
1090
pte = pte_mkuffd_wp(pte);
mm/memory.c
1091
set_pte_at(dst_vma->vm_mm, addr, dst_pte, pte);
mm/memory.c
1097
pte_t pte, unsigned long addr, int nr)
mm/memory.c
1102
if (is_cow_mapping(src_vma->vm_flags) && pte_write(pte)) {
mm/memory.c
1104
pte = pte_wrprotect(pte);
mm/memory.c
1109
pte = pte_mkclean(pte);
mm/memory.c
1110
pte = pte_mkold(pte);
mm/memory.c
1113
pte = pte_clear_uffd_wp(pte);
mm/memory.c
1115
set_ptes(dst_vma->vm_mm, addr, dst_pte, pte, nr);
mm/memory.c
1127
pte_t *dst_pte, pte_t *src_pte, pte_t pte, unsigned long addr,
mm/memory.c
1135
page = vm_normal_page(src_vma, addr, pte);
mm/memory.c
1152
nr = folio_pte_batch_flags(folio, src_vma, src_pte, &pte, max_nr, flags);
mm/memory.c
1166
__copy_present_ptes(dst_vma, src_vma, dst_pte, src_pte, pte,
mm/memory.c
1194
__copy_present_ptes(dst_vma, src_vma, dst_pte, src_pte, pte, addr, 1);
mm/memory.c
1605
unsigned long addr, pte_t *pte, int nr,
mm/memory.c
1622
if (pte_install_uffd_wp_if_needed(vma, addr, pte, pteval))
mm/memory.c
1626
pte++;
mm/memory.c
1635
struct page *page, pte_t *pte, pte_t ptent, unsigned int nr,
mm/memory.c
1643
ptent = get_and_clear_full_ptes(mm, addr, pte, nr, tlb->fullmm);
mm/memory.c
1656
clear_full_ptes(mm, addr, pte, nr, tlb->fullmm);
mm/memory.c
1661
tlb_remove_tlb_entries(tlb, pte, nr, addr);
mm/memory.c
1663
*any_skipped = zap_install_uffd_wp_if_needed(vma, addr, pte,
mm/memory.c
1685
struct vm_area_struct *vma, pte_t *pte, pte_t ptent,
mm/memory.c
1698
ptep_get_and_clear_full(mm, addr, pte, tlb->fullmm);
mm/memory.c
1700
tlb_remove_tlb_entry(tlb, pte, addr);
mm/memory.c
1703
pte, 1, details, ptent);
mm/memory.c
1719
nr = folio_pte_batch(folio, pte, ptent, max_nr);
mm/memory.c
1720
zap_present_folio_ptes(tlb, vma, folio, page, pte, ptent, nr,
mm/memory.c
1725
zap_present_folio_ptes(tlb, vma, folio, page, pte, ptent, 1, addr,
mm/memory.c
1731
struct vm_area_struct *vma, pte_t *pte, pte_t ptent,
mm/memory.c
1762
nr = swap_pte_batch(pte, max_nr, ptent);
mm/memory.c
1795
clear_not_present_full_ptes(vma->vm_mm, addr, pte, nr, tlb->fullmm);
mm/memory.c
1796
*any_skipped = zap_install_uffd_wp_if_needed(vma, addr, pte, nr, details, ptent);
mm/memory.c
1802
struct vm_area_struct *vma, pte_t *pte,
mm/memory.c
1808
pte_t ptent = ptep_get(pte);
mm/memory.c
1815
ptent = ptep_get(pte + nr);
mm/memory.c
1822
pte += nr;
mm/memory.c
1827
nr += zap_present_ptes(tlb, vma, pte, ptent, max_nr, addr,
mm/memory.c
1831
nr += zap_nonpresent_ptes(tlb, vma, pte, ptent, max_nr, addr,
mm/memory.c
1865
pte_t *start_pte, *pte;
mm/memory.c
1875
for (i = 0, pte = start_pte; i < PTRS_PER_PTE; i++, pte++) {
mm/memory.c
1876
if (!pte_none(ptep_get(pte)))
mm/memory.c
1906
pte_t *pte;
mm/memory.c
1915
start_pte = pte = pte_offset_map_lock(mm, pmd, addr, &ptl);
mm/memory.c
1916
if (!pte)
mm/memory.c
1929
nr = do_zap_pte_range(tlb, vma, pte, addr, end, details, rss,
mm/memory.c
1938
} while (pte += nr, addr += PAGE_SIZE * nr, addr != end);
mm/memory.c
2341
static int insert_page_into_pte_locked(struct vm_area_struct *vma, pte_t *pte,
mm/memory.c
2346
pte_t pteval = ptep_get(pte);
mm/memory.c
2359
if (ptep_set_access_flags(vma, addr, pte, pteval, 1))
mm/memory.c
2360
update_mmu_cache(vma, addr, pte);
mm/memory.c
2378
set_pte_at(vma->vm_mm, addr, pte, pteval);
mm/memory.c
2386
pte_t *pte;
mm/memory.c
2393
pte = get_locked_pte(vma->vm_mm, addr, &ptl);
mm/memory.c
2394
if (!pte)
mm/memory.c
2396
retval = insert_page_into_pte_locked(vma, pte, addr, page, prot,
mm/memory.c
2398
pte_unmap_unlock(pte, ptl);
mm/memory.c
2403
static int insert_page_in_batch_locked(struct vm_area_struct *vma, pte_t *pte,
mm/memory.c
2411
return insert_page_into_pte_locked(vma, pte, addr, page, prot, false);
mm/memory.c
2421
pte_t *start_pte, *pte;
mm/memory.c
2451
for (pte = start_pte; pte_idx < batch_size; ++pte, ++pte_idx) {
mm/memory.c
2452
int err = insert_page_in_batch_locked(vma, pte,
mm/memory.c
2630
pte_t *pte, entry;
mm/memory.c
2633
pte = get_locked_pte(mm, addr, &ptl);
mm/memory.c
2634
if (!pte)
mm/memory.c
2636
entry = ptep_get(pte);
mm/memory.c
2655
if (ptep_set_access_flags(vma, addr, pte, entry, 1))
mm/memory.c
2656
update_mmu_cache(vma, addr, pte);
mm/memory.c
2669
set_pte_at(mm, addr, pte, entry);
mm/memory.c
2670
update_mmu_cache(vma, addr, pte); /* XXX: why not for insert_page? */
mm/memory.c
2673
pte_unmap_unlock(pte, ptl);
mm/memory.c
2871
pte_t *pte, *mapped_pte;
mm/memory.c
2875
mapped_pte = pte = pte_alloc_map_lock(mm, pmd, addr, &ptl);
mm/memory.c
2876
if (!pte)
mm/memory.c
2880
BUG_ON(!pte_none(ptep_get(pte)));
mm/memory.c
2885
set_pte_at(mm, addr, pte, pte_mkspecial(pfn_pte(pfn, prot)));
mm/memory.c
2887
} while (pte++, addr += PAGE_SIZE, addr != end);
mm/memory.c
3220
pte_t *pte, *mapped_pte;
mm/memory.c
3225
mapped_pte = pte = (mm == &init_mm) ?
mm/memory.c
3228
if (!pte)
mm/memory.c
3231
mapped_pte = pte = (mm == &init_mm) ?
mm/memory.c
3234
if (!pte)
mm/memory.c
3242
if (create || !pte_none(ptep_get(pte))) {
mm/memory.c
3243
err = fn(pte, addr, data);
mm/memory.c
3247
} while (pte++, addr += PAGE_SIZE, addr != end);
mm/memory.c
3445
same = pte_same(ptep_get(vmf->pte), vmf->orig_pte);
mm/memory.c
3449
pte_unmap(vmf->pte);
mm/memory.c
3450
vmf->pte = NULL;
mm/memory.c
3490
vmf->pte = NULL;
mm/memory.c
3494
vmf->pte = pte_offset_map_lock(mm, vmf->pmd, addr, &vmf->ptl);
mm/memory.c
3495
if (unlikely(!vmf->pte || !pte_same(ptep_get(vmf->pte), vmf->orig_pte))) {
mm/memory.c
3500
if (vmf->pte)
mm/memory.c
3501
update_mmu_tlb(vma, addr, vmf->pte);
mm/memory.c
3507
if (ptep_set_access_flags(vma, addr, vmf->pte, entry, 0))
mm/memory.c
3508
update_mmu_cache_range(vmf, vma, addr, vmf->pte, 1);
mm/memory.c
3518
if (vmf->pte)
mm/memory.c
3522
vmf->pte = pte_offset_map_lock(mm, vmf->pmd, addr, &vmf->ptl);
mm/memory.c
3523
if (unlikely(!vmf->pte || !pte_same(ptep_get(vmf->pte), vmf->orig_pte))) {
mm/memory.c
3525
if (vmf->pte)
mm/memory.c
3526
update_mmu_tlb(vma, addr, vmf->pte);
mm/memory.c
3549
if (vmf->pte)
mm/memory.c
3550
pte_unmap_unlock(vmf->pte, vmf->ptl);
mm/memory.c
3687
if (ptep_set_access_flags(vma, vmf->address, vmf->pte, entry, 1))
mm/memory.c
3688
update_mmu_cache_range(vmf, vma, vmf->address, vmf->pte, 1);
mm/memory.c
3689
pte_unmap_unlock(vmf->pte, vmf->ptl);
mm/memory.c
3816
vmf->pte = pte_offset_map_lock(mm, vmf->pmd, vmf->address, &vmf->ptl);
mm/memory.c
3817
if (likely(vmf->pte && pte_same(ptep_get(vmf->pte), vmf->orig_pte))) {
mm/memory.c
3846
ptep_clear_flush(vma, vmf->address, vmf->pte);
mm/memory.c
3850
set_pte_at(mm, vmf->address, vmf->pte, entry);
mm/memory.c
3851
update_mmu_cache_range(vmf, vma, vmf->address, vmf->pte, 1);
mm/memory.c
3881
pte_unmap_unlock(vmf->pte, vmf->ptl);
mm/memory.c
3882
} else if (vmf->pte) {
mm/memory.c
3883
update_mmu_tlb(vma, vmf->address, vmf->pte);
mm/memory.c
3884
pte_unmap_unlock(vmf->pte, vmf->ptl);
mm/memory.c
3929
vmf->pte = pte_offset_map_lock(vmf->vma->vm_mm, vmf->pmd, vmf->address,
mm/memory.c
3931
if (!vmf->pte)
mm/memory.c
3937
if (!pte_same(ptep_get(vmf->pte), vmf->orig_pte)) {
mm/memory.c
3938
update_mmu_tlb(vmf->vma, vmf->address, vmf->pte);
mm/memory.c
3939
pte_unmap_unlock(vmf->pte, vmf->ptl);
mm/memory.c
3957
pte_unmap_unlock(vmf->pte, vmf->ptl);
mm/memory.c
3983
pte_unmap_unlock(vmf->pte, vmf->ptl);
mm/memory.c
4155
pte_t pte;
mm/memory.c
4158
if (userfaultfd_pte_wp(vma, ptep_get(vmf->pte))) {
mm/memory.c
4160
pte_unmap_unlock(vmf->pte, vmf->ptl);
mm/memory.c
4169
pte = pte_clear_uffd_wp(ptep_get(vmf->pte));
mm/memory.c
4171
set_pte_at(vma->vm_mm, vmf->address, vmf->pte, pte);
mm/memory.c
4176
vmf->orig_pte = pte;
mm/memory.c
4224
pte_unmap_unlock(vmf->pte, vmf->ptl);
mm/memory.c
4236
pte_unmap_unlock(vmf->pte, vmf->ptl);
mm/memory.c
438
void pmd_install(struct mm_struct *mm, pmd_t *pmd, pgtable_t *pte)
mm/memory.c
4403
vmf->pte = pte_offset_map_lock(vma->vm_mm, vmf->pmd, vmf->address,
mm/memory.c
4405
if (likely(vmf->pte && pte_same(ptep_get(vmf->pte), vmf->orig_pte)))
mm/memory.c
4407
vmf->pte, vmf->orig_pte);
mm/memory.c
4409
if (vmf->pte)
mm/memory.c
4410
pte_unmap_unlock(vmf->pte, vmf->ptl);
mm/memory.c
4454
vmf->pte = pte_offset_map_lock(vmf->vma->vm_mm, vmf->pmd,
mm/memory.c
4456
if (!vmf->pte)
mm/memory.c
4466
if (pte_same(vmf->orig_pte, ptep_get(vmf->pte)))
mm/memory.c
4467
pte_clear(vmf->vma->vm_mm, vmf->address, vmf->pte);
mm/memory.c
4468
pte_unmap_unlock(vmf->pte, vmf->ptl);
mm/memory.c
4553
pte_t pte;
mm/memory.c
4557
pte = ptep_get(ptep);
mm/memory.c
4559
if (!pte_same(pte, pte_move_swp_offset(vmf->orig_pte, -idx)))
mm/memory.c
4561
entry = softleaf_from_pte(pte);
mm/memory.c
4562
if (swap_pte_batch(ptep, nr_pages, pte) != nr_pages)
mm/memory.c
458
pmd_populate(mm, pmd, *pte);
mm/memory.c
459
*pte = NULL;
mm/memory.c
4609
pte_t *pte;
mm/memory.c
4642
pte = pte_offset_map_lock(vmf->vma->vm_mm, vmf->pmd,
mm/memory.c
4644
if (unlikely(!pte))
mm/memory.c
4654
if (can_swapin_thp(vmf, pte + pte_index(addr), 1 << order))
mm/memory.c
4659
pte_unmap_unlock(pte, ptl);
mm/memory.c
4715
pte_t pte;
mm/memory.c
4745
vmf->pte = pte_offset_map_lock(vma->vm_mm, vmf->pmd,
mm/memory.c
4747
if (unlikely(!vmf->pte ||
mm/memory.c
4748
!pte_same(ptep_get(vmf->pte),
mm/memory.c
4760
pte_unmap_unlock(vmf->pte, vmf->ptl);
mm/memory.c
4766
pte_unmap(vmf->pte);
mm/memory.c
4810
vmf->pte = pte_offset_map_lock(vma->vm_mm, vmf->pmd,
mm/memory.c
4812
if (likely(vmf->pte &&
mm/memory.c
4813
pte_same(ptep_get(vmf->pte), vmf->orig_pte)))
mm/memory.c
4881
vmf->pte = pte_offset_map_lock(vma->vm_mm, vmf->pmd, vmf->address,
mm/memory.c
4883
if (unlikely(!vmf->pte || !pte_same(ptep_get(vmf->pte), vmf->orig_pte)))
mm/memory.c
4894
ptep = vmf->pte;
mm/memory.c
4908
folio_ptep = vmf->pte - idx;
mm/memory.c
5004
pte = mk_pte(page, vma->vm_page_prot);
mm/memory.c
5006
pte = pte_mksoft_dirty(pte);
mm/memory.c
5008
pte = pte_mkuffd_wp(pte);
mm/memory.c
5018
if ((vma->vm_flags & VM_WRITE) && !userfaultfd_pte_wp(vma, pte) &&
mm/memory.c
5019
!pte_needs_soft_dirty_wp(vma, pte)) {
mm/memory.c
5020
pte = pte_mkwrite(pte, vma);
mm/memory.c
5022
pte = pte_mkdirty(pte);
mm/memory.c
5030
vmf->orig_pte = pte_advance_pfn(pte, page_idx);
mm/memory.c
5054
(pte_write(pte) && !PageAnonExclusive(page)));
mm/memory.c
5055
set_ptes(vma->vm_mm, address, ptep, pte, nr_pages);
mm/memory.c
5057
pte, pte, nr_pages);
mm/memory.c
5091
if (vmf->pte)
mm/memory.c
5092
pte_unmap_unlock(vmf->pte, vmf->ptl);
mm/memory.c
5098
if (vmf->pte)
mm/memory.c
5099
pte_unmap_unlock(vmf->pte, vmf->ptl);
mm/memory.c
5115
static bool pte_range_none(pte_t *pte, int nr_pages)
mm/memory.c
5120
if (!pte_none(ptep_get_lockless(pte + i)))
mm/memory.c
5134
pte_t *pte;
mm/memory.c
5157
pte = pte_offset_map(vmf->pmd, vmf->address & PMD_MASK);
mm/memory.c
5158
if (!pte)
mm/memory.c
5169
if (pte_range_none(pte + pte_index(addr), 1 << order))
mm/memory.c
5174
pte_unmap(pte);
mm/memory.c
5242
vmf->pte = pte_offset_map_lock(vma->vm_mm, vmf->pmd,
mm/memory.c
5244
if (!vmf->pte)
mm/memory.c
5247
update_mmu_tlb(vma, vmf->address, vmf->pte);
mm/memory.c
5255
pte_unmap_unlock(vmf->pte, vmf->ptl);
mm/memory.c
5287
vmf->pte = pte_offset_map_lock(vma->vm_mm, vmf->pmd, addr, &vmf->ptl);
mm/memory.c
5288
if (!vmf->pte)
mm/memory.c
5291
update_mmu_tlb(vma, addr, vmf->pte);
mm/memory.c
5293
} else if (nr_pages > 1 && !pte_range_none(vmf->pte, nr_pages)) {
mm/memory.c
5294
update_mmu_tlb_range(vma, addr, vmf->pte, nr_pages);
mm/memory.c
5304
pte_unmap_unlock(vmf->pte, vmf->ptl);
mm/memory.c
5317
set_ptes(vma->vm_mm, addr, vmf->pte, entry, nr_pages);
mm/memory.c
5320
update_mmu_cache_range(vmf, vma, addr, vmf->pte, nr_pages);
mm/memory.c
5322
if (vmf->pte)
mm/memory.c
5323
pte_unmap_unlock(vmf->pte, vmf->ptl);
mm/memory.c
5527
set_ptes(vma->vm_mm, addr, vmf->pte, entry, nr);
mm/memory.c
5530
update_mmu_cache_range(vmf, vma, addr, vmf->pte, nr);
mm/memory.c
5536
return !pte_same(ptep_get(vmf->pte), vmf->orig_pte);
mm/memory.c
5538
return !pte_none(ptep_get(vmf->pte));
mm/memory.c
5646
vmf->pte = pte_offset_map_lock(vma->vm_mm, vmf->pmd,
mm/memory.c
5648
if (!vmf->pte)
mm/memory.c
5653
update_mmu_tlb(vma, addr, vmf->pte);
mm/memory.c
5656
} else if (nr_pages > 1 && !pte_range_none(vmf->pte, nr_pages)) {
mm/memory.c
5658
pte_unmap_unlock(vmf->pte, vmf->ptl);
mm/memory.c
5669
pte_unmap_unlock(vmf->pte, vmf->ptl);
mm/memory.c
5913
vmf->pte = pte_offset_map_lock(vmf->vma->vm_mm, vmf->pmd,
mm/memory.c
5915
if (unlikely(!vmf->pte))
mm/memory.c
5925
if (unlikely(pte_none(ptep_get(vmf->pte))))
mm/memory.c
5930
pte_unmap_unlock(vmf->pte, vmf->ptl);
mm/memory.c
5998
pte_t pte, old_pte;
mm/memory.c
6001
pte = pte_modify(old_pte, vma->vm_page_prot);
mm/memory.c
6002
pte = pte_mkyoung(pte);
mm/memory.c
6004
pte = pte_mkwrite(pte, vma);
mm/memory.c
6005
ptep_modify_prot_commit(vma, fault_addr, fault_pte, old_pte, pte);
mm/memory.c
6023
start_ptep = vmf->pte - ((addr - start) >> PAGE_SHIFT);
mm/memory.c
6057
pte_t pte, old_pte;
mm/memory.c
6066
old_pte = ptep_get(vmf->pte);
mm/memory.c
6069
pte_unmap_unlock(vmf->pte, vmf->ptl);
mm/memory.c
6073
pte = pte_modify(old_pte, vma->vm_page_prot);
mm/memory.c
6079
writable = pte_write(pte);
mm/memory.c
6081
can_change_pte_writable(vma, vmf->address, pte))
mm/memory.c
6084
folio = vm_normal_folio(vma, vmf->address, pte);
mm/memory.c
6100
pte_unmap_unlock(vmf->pte, vmf->ptl);
mm/memory.c
6113
vmf->pte = pte_offset_map_lock(vma->vm_mm, vmf->pmd,
mm/memory.c
6115
if (unlikely(!vmf->pte))
mm/memory.c
6117
if (unlikely(!pte_same(ptep_get(vmf->pte), vmf->orig_pte))) {
mm/memory.c
6118
pte_unmap_unlock(vmf->pte, vmf->ptl);
mm/memory.c
6127
numa_rebuild_large_mapping(vmf, vma, folio, pte, ignore_writable,
mm/memory.c
6130
numa_rebuild_single_mapping(vmf, vma, vmf->address, vmf->pte,
mm/memory.c
6132
pte_unmap_unlock(vmf->pte, vmf->ptl);
mm/memory.c
6252
vmf->pte);
mm/memory.c
628
#define print_bad_pte(vma, addr, pte, page) \
mm/memory.c
6284
vmf->pte = NULL;
mm/memory.c
629
print_bad_page_map(vma, addr, pte_val(pte), page, PGTABLE_LEVEL_PTE)
mm/memory.c
6302
vmf->pte = pte_offset_map_rw_nolock(vmf->vma->vm_mm, vmf->pmd,
mm/memory.c
6305
if (unlikely(!vmf->pte))
mm/memory.c
6307
vmf->orig_pte = ptep_get_lockless(vmf->pte);
mm/memory.c
6311
pte_unmap(vmf->pte);
mm/memory.c
6312
vmf->pte = NULL;
mm/memory.c
6316
if (!vmf->pte)
mm/memory.c
6327
if (unlikely(!pte_same(ptep_get(vmf->pte), entry))) {
mm/memory.c
6328
update_mmu_tlb(vmf->vma, vmf->address, vmf->pte);
mm/memory.c
6338
if (ptep_set_access_flags(vmf->vma, vmf->address, vmf->pte, entry,
mm/memory.c
6341
vmf->pte, 1);
mm/memory.c
6345
pte_unmap_unlock(vmf->pte, vmf->ptl);
mm/memory.c
6797
pte_t *ptep, pte;
mm/memory.c
6861
pte = ptep_get(ptep);
mm/memory.c
6862
if (!pte_present(pte))
mm/memory.c
6864
pfnmap_args_setup(args, lock, ptep, pte_pgprot(pte),
mm/memory.c
6865
pte_pfn(pte), PAGE_MASK, pte_write(pte),
mm/memory.c
6866
pte_special(pte));
mm/memory.c
765
pte_t pte)
mm/memory.c
767
return __vm_normal_page(vma, addr, pte_pfn(pte), pte_special(pte),
mm/memory.c
768
pte_val(pte), PGTABLE_LEVEL_PTE);
mm/memory.c
784
pte_t pte)
mm/memory.c
786
struct page *page = vm_normal_page(vma, addr, pte);
mm/memory.c
885
pte_t pte;
mm/memory.c
889
pte = pte_mkold(mk_pte(page, READ_ONCE(vma->vm_page_prot)));
mm/memory.c
891
pte = pte_mksoft_dirty(pte);
mm/memory.c
894
pte = pte_mkuffd_wp(pte);
mm/memory.c
897
can_change_pte_writable(vma, address, pte)) {
mm/memory.c
899
pte = pte_mkdirty(pte);
mm/memory.c
900
pte = pte_mkwrite(pte, vma);
mm/memory.c
902
set_pte_at(vma->vm_mm, address, ptep, pte);
mm/memory.c
945
pte_t pte = orig_pte;
mm/memory.c
963
pte = pte_swp_clear_exclusive(orig_pte);
mm/memory.c
964
set_pte_at(src_mm, addr, src_pte, pte);
mm/memory.c
981
pte = softleaf_to_pte(entry);
mm/memory.c
983
pte = pte_swp_mksoft_dirty(pte);
mm/memory.c
985
pte = pte_swp_mkuffd_wp(pte);
mm/memory.c
986
set_pte_at(src_mm, addr, src_pte, pte);
mm/mempolicy.c
691
pte_t *pte, *mapped_pte;
mm/mempolicy.c
703
mapped_pte = pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl);
mm/mempolicy.c
704
if (!pte) {
mm/mempolicy.c
708
for (; addr != end; pte += nr, addr += nr * PAGE_SIZE) {
mm/mempolicy.c
711
ptent = ptep_get(pte);
mm/mempolicy.c
725
nr = folio_pte_batch(folio, pte, ptent, max_nr);
mm/mempolicy.c
771
static int queue_folios_hugetlb(pte_t *pte, unsigned long hmask,
mm/mempolicy.c
782
ptl = huge_pte_lock(hstate_vma(walk->vma), walk->mm, pte);
mm/mempolicy.c
783
ptep = huge_ptep_get(walk->mm, addr, pte);
mm/mempolicy.c
810
(!folio_maybe_mapped_shared(folio) && !hugetlb_pmd_shared(pte)))
mm/migrate.c
332
set_pte_at(pvmw->vma->vm_mm, pvmw->address, pvmw->pte, newpte);
mm/migrate.c
355
pte_t pte;
mm/migrate.c
367
if (!pvmw.pte) {
mm/migrate.c
374
old_pte = ptep_get(pvmw.pte);
mm/migrate.c
380
pte = mk_pte(new, READ_ONCE(vma->vm_page_prot));
mm/migrate.c
384
pte = pte_mkold(pte);
mm/migrate.c
386
pte = pte_mkdirty(pte);
mm/migrate.c
388
pte = pte_mksoft_dirty(pte);
mm/migrate.c
390
pte = pte_clear_soft_dirty(pte);
mm/migrate.c
393
pte = pte_mkwrite(pte, vma);
mm/migrate.c
395
pte = pte_mkuffd_wp(pte);
mm/migrate.c
401
if (pte_write(pte))
mm/migrate.c
407
pte = softleaf_to_pte(entry);
mm/migrate.c
409
pte = pte_swp_mksoft_dirty(pte);
mm/migrate.c
411
pte = pte_swp_mkuffd_wp(pte);
mm/migrate.c
420
pte = arch_make_huge_pte(pte, shift, vma->vm_flags);
mm/migrate.c
426
set_huge_pte_at(vma->vm_mm, pvmw.address, pvmw.pte, pte,
mm/migrate.c
436
set_pte_at(vma->vm_mm, pvmw.address, pvmw.pte, pte);
mm/migrate.c
441
trace_remove_migration_pte(pvmw.address, pte_val(pte),
mm/migrate.c
445
update_mmu_cache(vma, pvmw.address, pvmw.pte);
mm/migrate.c
486
pte_t pte;
mm/migrate.c
493
pte = ptep_get(ptep);
mm/migrate.c
496
if (pte_none(pte) || pte_present(pte))
mm/migrate.c
499
entry = softleaf_from_pte(pte);
mm/migrate.c
520
pte_t pte;
mm/migrate.c
524
pte = huge_ptep_get(vma->vm_mm, addr, ptep);
mm/migrate.c
526
if (huge_pte_none(pte))
mm/migrate.c
529
entry = softleaf_from_pte(pte);
mm/migrate_device.c
283
pte_t pte;
mm/migrate_device.c
285
pte = ptep_get(ptep);
mm/migrate_device.c
287
if (pte_none(pte)) {
mm/migrate_device.c
295
if (!pte_present(pte)) {
mm/migrate_device.c
301
entry = softleaf_from_pte(pte);
mm/migrate_device.c
336
pfn = pte_pfn(pte);
mm/migrate_device.c
343
page = vm_normal_page(migrate->vma, addr, pte);
mm/migrate_device.c
374
mpfn |= pte_write(pte) ? MIGRATE_PFN_WRITE : 0;
mm/migrate_device.c
410
flush_cache_page(vma, addr, pte_pfn(pte));
mm/migrate_device.c
414
pte = ptep_clear_flush(vma, addr, ptep);
mm/migrate_device.c
417
set_pte_at(mm, addr, ptep, pte);
mm/migrate_device.c
425
pte = ptep_get_and_clear(mm, addr, ptep);
mm/migrate_device.c
431
if (pte_dirty(pte))
mm/migrate_device.c
444
if (pte_present(pte)) {
mm/migrate_device.c
445
if (pte_young(pte))
mm/migrate_device.c
447
if (pte_dirty(pte))
mm/migrate_device.c
451
if (pte_present(pte)) {
mm/migrate_device.c
452
if (pte_soft_dirty(pte))
mm/migrate_device.c
454
if (pte_uffd_wp(pte))
mm/migrate_device.c
457
if (pte_swp_soft_dirty(pte))
mm/migrate_device.c
459
if (pte_swp_uffd_wp(pte))
mm/migrate_device.c
472
if (pte_present(pte))
mm/mincore.c
186
pte_t pte = ptep_get(ptep);
mm/mincore.c
190
if (pte_none(pte) || pte_is_marker(pte))
mm/mincore.c
193
else if (pte_present(pte)) {
mm/mincore.c
194
unsigned int batch = pte_batch_hint(ptep, pte);
mm/mincore.c
205
const softleaf_t entry = softleaf_from_pte(pte);
mm/mincore.c
26
static int mincore_hugetlb(pte_t *pte, unsigned long hmask, unsigned long addr,
mm/mincore.c
34
ptl = huge_pte_lock(hstate_vma(walk->vma), walk->mm, pte);
mm/mincore.c
40
if (!pte) {
mm/mincore.c
43
const pte_t ptep = huge_ptep_get(walk->mm, addr, pte);
mm/mlock.c
308
pte_t *pte, unsigned long addr, unsigned long end)
mm/mlock.c
311
pte_t ptent = ptep_get(pte);
mm/mlock.c
316
return folio_pte_batch(folio, pte, ptent, count);
mm/mlock.c
356
pte_t *start_pte, *pte;
mm/mlock.c
384
for (pte = start_pte; addr != end; pte++, addr += PAGE_SIZE) {
mm/mlock.c
385
ptent = ptep_get(pte);
mm/mlock.c
392
step = folio_mlock_step(folio, pte, addr, end);
mm/mlock.c
402
pte += step - 1;
mm/mprotect.c
101
return can_change_private_pte_writable(vma, addr, pte);
mm/mprotect.c
103
return can_change_shared_pte_writable(vma, pte);
mm/mprotect.c
107
pte_t pte, int max_nr_ptes, fpb_t flags)
mm/mprotect.c
116
return folio_pte_batch_flags(folio, NULL, ptep, &pte, max_nr_ptes, flags);
mm/mprotect.c
218
pte_t *pte, oldpte;
mm/mprotect.c
228
pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl);
mm/mprotect.c
229
if (!pte)
mm/mprotect.c
239
oldpte = ptep_get(pte);
mm/mprotect.c
265
pte, oldpte, max_nr_ptes, /* flags = */ 0);
mm/mprotect.c
269
nr_ptes = mprotect_folio_pte_batch(folio, pte, oldpte, max_nr_ptes, flags);
mm/mprotect.c
271
oldpte = modify_prot_start_ptes(vma, addr, pte, nr_ptes);
mm/mprotect.c
295
addr, pte, oldpte, ptent, nr_ptes, tlb);
mm/mprotect.c
297
prot_commit_flush_ptes(vma, addr, pte, oldpte, ptent,
mm/mprotect.c
315
set_pte_at(vma->vm_mm, addr, pte,
mm/mprotect.c
363
pte_clear(vma->vm_mm, addr, pte);
mm/mprotect.c
377
set_pte_at(vma->vm_mm, addr, pte, newpte);
mm/mprotect.c
381
} while (pte += nr_ptes, addr += nr_ptes * PAGE_SIZE, addr != end);
mm/mprotect.c
383
pte_unmap_unlock(pte - 1, ptl);
mm/mprotect.c
41
static bool maybe_change_pte_writable(struct vm_area_struct *vma, pte_t pte)
mm/mprotect.c
47
if (pte_protnone(pte))
mm/mprotect.c
51
if (pte_needs_soft_dirty_wp(vma, pte))
mm/mprotect.c
55
if (userfaultfd_pte_wp(vma, pte))
mm/mprotect.c
62
unsigned long addr, pte_t pte)
mm/mprotect.c
66
if (!maybe_change_pte_writable(vma, pte))
mm/mprotect.c
664
static int prot_none_pte_entry(pte_t *pte, unsigned long addr,
mm/mprotect.c
667
return pfn_modify_allowed(pte_pfn(ptep_get(pte)),
mm/mprotect.c
672
static int prot_none_hugetlb_entry(pte_t *pte, unsigned long hmask,
mm/mprotect.c
676
return pfn_modify_allowed(pte_pfn(ptep_get(pte)),
mm/mprotect.c
75
page = vm_normal_page(vma, addr, pte);
mm/mprotect.c
80
pte_t pte)
mm/mprotect.c
82
if (!maybe_change_pte_writable(vma, pte))
mm/mprotect.c
85
VM_WARN_ON_ONCE(is_zero_pfn(pte_pfn(pte)) && pte_dirty(pte));
mm/mprotect.c
94
return pte_dirty(pte);
mm/mprotect.c
98
pte_t pte)
mm/mremap.c
159
static pte_t move_soft_dirty_pte(pte_t pte)
mm/mremap.c
161
if (pte_none(pte))
mm/mremap.c
162
return pte;
mm/mremap.c
169
if (pte_present(pte))
mm/mremap.c
170
pte = pte_mksoft_dirty(pte);
mm/mremap.c
172
pte = pte_swp_mksoft_dirty(pte);
mm/mremap.c
175
return pte;
mm/mremap.c
179
pte_t *ptep, pte_t pte, int max_nr)
mm/mremap.c
187
if (pte_batch_hint(ptep, pte) == 1)
mm/mremap.c
190
folio = vm_normal_folio(vma, addr, pte);
mm/mremap.c
194
return folio_pte_batch_flags(folio, NULL, ptep, &pte, max_nr, FPB_RESPECT_WRITE);
mm/mremap.c
204
pte_t old_pte, pte;
mm/mremap.c
291
pte = get_and_clear_ptes(mm, old_addr, old_ptep, nr_ptes);
mm/mremap.c
292
pte = move_pte(pte, old_addr, new_addr);
mm/mremap.c
293
pte = move_soft_dirty_pte(pte);
mm/mremap.c
295
if (need_clear_uffd_wp && pte_is_uffd_wp_marker(pte))
mm/mremap.c
299
if (pte_present(pte))
mm/mremap.c
300
pte = pte_clear_uffd_wp(pte);
mm/mremap.c
302
pte = pte_swp_clear_uffd_wp(pte);
mm/mremap.c
304
set_ptes(mm, new_addr, new_ptep, pte, nr_ptes);
mm/page_idle.c
61
if (pvmw.pte) {
mm/page_idle.c
70
if (likely(pte_present(ptep_get(pvmw.pte))))
mm/page_idle.c
71
referenced |= ptep_test_and_clear_young(vma, addr, pvmw.pte);
mm/page_table_check.c
149
pte_t pte)
mm/page_table_check.c
154
if (pte_user_accessible_page(pte, addr)) {
mm/page_table_check.c
155
page_table_check_clear(pte_pfn(pte), PAGE_SIZE >> PAGE_SHIFT);
mm/page_table_check.c
191
static void page_table_check_pte_flags(pte_t pte)
mm/page_table_check.c
193
if (pte_present(pte)) {
mm/page_table_check.c
194
WARN_ON_ONCE(pte_uffd_wp(pte) && pte_write(pte));
mm/page_table_check.c
195
} else if (pte_swp_uffd_wp(pte)) {
mm/page_table_check.c
196
const softleaf_t entry = softleaf_from_pte(pte);
mm/page_table_check.c
203
pte_t *ptep, pte_t pte, unsigned int nr)
mm/page_table_check.c
210
page_table_check_pte_flags(pte);
mm/page_table_check.c
214
if (pte_user_accessible_page(pte, addr))
mm/page_table_check.c
215
page_table_check_set(pte_pfn(pte), nr, pte_write(pte));
mm/page_vma_mapped.c
110
pte_t ptent = ptep_get(pvmw->pte);
mm/page_vma_mapped.c
192
if (pvmw->pmd && !pvmw->pte)
mm/page_vma_mapped.c
199
if (pvmw->pte)
mm/page_vma_mapped.c
206
pvmw->pte = hugetlb_walk(vma, pvmw->address, size);
mm/page_vma_mapped.c
207
if (!pvmw->pte)
mm/page_vma_mapped.c
210
pvmw->ptl = huge_pte_lock(hstate, mm, pvmw->pte);
mm/page_vma_mapped.c
217
if (pvmw->pte)
mm/page_vma_mapped.c
24
pvmw->pte = pte_offset_map_lock(pvmw->vma->vm_mm, pvmw->pmd,
mm/page_vma_mapped.c
27
return !!pvmw->pte;
mm/page_vma_mapped.c
296
if (!pvmw->pte)
mm/page_vma_mapped.c
314
pte_unmap(pvmw->pte);
mm/page_vma_mapped.c
315
pvmw->pte = NULL;
mm/page_vma_mapped.c
319
pvmw->pte++;
mm/page_vma_mapped.c
320
} while (pte_none(ptep_get(pvmw->pte)));
mm/page_vma_mapped.c
325
pte_unmap_unlock(pvmw->pte, ptl);
mm/page_vma_mapped.c
326
pvmw->pte = NULL;
mm/page_vma_mapped.c
39
pvmw->pte = pte_offset_map_rw_nolock(pvmw->vma->vm_mm, pvmw->pmd,
mm/page_vma_mapped.c
41
if (!pvmw->pte)
mm/page_vma_mapped.c
44
ptent = ptep_get(pvmw->pte);
mm/page_vma_mapped.c
77
pte_unmap_unlock(pvmw->pte, *ptlp);
mm/pagewalk.c
1012
pte = ptep_get(ptep);
mm/pagewalk.c
1017
fw->pte = pte;
mm/pagewalk.c
1019
if (pte_present(pte)) {
mm/pagewalk.c
1020
page = vm_normal_page(vma, addr, pte);
mm/pagewalk.c
1024
is_zero_pfn(pte_pfn(pte))) {
mm/pagewalk.c
1025
page = pfn_to_page(pte_pfn(pte));
mm/pagewalk.c
1029
} else if (!pte_none(pte)) {
mm/pagewalk.c
1030
const softleaf_t entry = softleaf_from_pte(pte);
mm/pagewalk.c
30
static int walk_pte_range_inner(pte_t *pte, unsigned long addr,
mm/pagewalk.c
347
pte_t *pte;
mm/pagewalk.c
354
pte = hugetlb_walk(vma, addr & hmask, sz);
mm/pagewalk.c
355
if (pte)
mm/pagewalk.c
356
err = ops->hugetlb_entry(pte, hmask, addr, next, walk);
mm/pagewalk.c
37
if (ops->install_pte && pte_none(ptep_get(pte))) {
mm/pagewalk.c
45
set_pte_at(walk->mm, addr, pte, new_pte);
mm/pagewalk.c
48
update_mmu_cache(walk->vma, addr, pte);
mm/pagewalk.c
50
err = ops->pte_entry(pte, addr, addr + PAGE_SIZE, walk);
mm/pagewalk.c
57
pte++;
mm/pagewalk.c
65
pte_t *pte;
mm/pagewalk.c
77
pte = pte_offset_kernel(pmd, addr);
mm/pagewalk.c
79
pte = pte_offset_map(pmd, addr);
mm/pagewalk.c
80
if (pte) {
mm/pagewalk.c
81
err = walk_pte_range_inner(pte, addr, end, walk);
mm/pagewalk.c
83
pte_unmap(pte);
mm/pagewalk.c
86
pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl);
mm/pagewalk.c
87
if (pte) {
mm/pagewalk.c
88
err = walk_pte_range_inner(pte, addr, end, walk);
mm/pagewalk.c
89
pte_unmap_unlock(pte, ptl);
mm/pagewalk.c
911
pte_t *ptep, pte;
mm/pagewalk.c
92
if (!pte)
mm/pgtable-generic.c
100
pte_t pte;
mm/pgtable-generic.c
101
pte = ptep_get_and_clear(mm, address, ptep);
mm/pgtable-generic.c
102
if (pte_accessible(mm, pte))
mm/pgtable-generic.c
104
return pte;
mm/pgtable-generic.c
313
pte_t *pte;
mm/pgtable-generic.c
315
pte = __pte_offset_map(pmd, addr, &pmdval);
mm/pgtable-generic.c
316
if (likely(pte))
mm/pgtable-generic.c
318
return pte;
mm/pgtable-generic.c
325
pte_t *pte;
mm/pgtable-generic.c
328
pte = __pte_offset_map(pmd, addr, pmdvalp);
mm/pgtable-generic.c
329
if (likely(pte))
mm/pgtable-generic.c
331
return pte;
mm/pgtable-generic.c
396
pte_t *pte;
mm/pgtable-generic.c
398
pte = __pte_offset_map(pmd, addr, &pmdval);
mm/pgtable-generic.c
399
if (unlikely(!pte))
mm/pgtable-generic.c
400
return pte;
mm/pgtable-generic.c
405
return pte;
mm/pgtable-generic.c
407
pte_unmap_unlock(pte, ptl);
mm/ptdump.c
120
static int ptdump_pte_entry(pte_t *pte, unsigned long addr,
mm/ptdump.c
124
pte_t val = ptep_get(pte);
mm/rmap.c
1116
if (pvmw->pte) {
mm/rmap.c
1117
pte_t *pte = pvmw->pte;
mm/rmap.c
1118
pte_t entry = ptep_get(pte);
mm/rmap.c
1132
entry = ptep_clear_flush(vma, address, pte);
mm/rmap.c
1135
set_pte_at(vma->vm_mm, address, pte, entry);
mm/rmap.c
1940
enum ttu_flags flags, pte_t pte)
mm/rmap.c
1959
if (pte_unused(pte))
mm/rmap.c
1971
return folio_pte_batch_flags(folio, vma, pvmw->pte, &pte, max_nr,
mm/rmap.c
2044
if (pvmw.pte && ptes != pvmw.nr_pages)
mm/rmap.c
2062
if (!pvmw.pte) {
mm/rmap.c
2088
VM_BUG_ON_FOLIO(!pvmw.pte, folio);
mm/rmap.c
2094
pteval = ptep_get(pvmw.pte);
mm/rmap.c
2144
if (huge_pmd_unshare(&tlb, vma, address, pvmw.pte)) {
mm/rmap.c
2157
pteval = huge_ptep_clear_flush(vma, address, pvmw.pte);
mm/rmap.c
2166
pteval = get_and_clear_ptes(mm, address, pvmw.pte, nr_pages);
mm/rmap.c
2182
pte_clear(mm, address, pvmw.pte);
mm/rmap.c
2190
pte_install_uffd_wp_if_needed(vma, address, pvmw.pte, pteval);
mm/rmap.c
2199
set_huge_pte_at(mm, address, pvmw.pte, pteval,
mm/rmap.c
2203
set_pte_at(mm, address, pvmw.pte, pteval);
mm/rmap.c
2256
set_ptes(mm, address, pvmw.pte, pteval, nr_pages);
mm/rmap.c
2269
set_ptes(mm, address, pvmw.pte, pteval, nr_pages);
mm/rmap.c
2277
set_pte_at(mm, address, pvmw.pte, pteval);
mm/rmap.c
2288
set_pte_at(mm, address, pvmw.pte, pteval);
mm/rmap.c
2296
set_pte_at(mm, address, pvmw.pte, pteval);
mm/rmap.c
2321
set_pte_at(mm, address, pvmw.pte, swp_pte);
mm/rmap.c
2455
if (!pvmw.pte) {
mm/rmap.c
2494
VM_BUG_ON_FOLIO(!pvmw.pte, folio);
mm/rmap.c
2500
pteval = ptep_get(pvmw.pte);
mm/rmap.c
2548
if (huge_pmd_unshare(&tlb, vma, address, pvmw.pte)) {
mm/rmap.c
2563
pteval = huge_ptep_clear_flush(vma, address, pvmw.pte);
mm/rmap.c
2579
pteval = ptep_get_and_clear(mm, address, pvmw.pte);
mm/rmap.c
2583
pteval = ptep_clear_flush(vma, address, pvmw.pte);
mm/rmap.c
2591
pte_clear(mm, address, pvmw.pte);
mm/rmap.c
2608
set_huge_pte_at(mm, address, pvmw.pte, pteval,
mm/rmap.c
2612
set_pte_at(mm, address, pvmw.pte, pteval);
mm/rmap.c
2638
set_huge_pte_at(mm, address, pvmw.pte,
mm/rmap.c
2641
set_pte_at(mm, address, pvmw.pte, pteval);
mm/rmap.c
2651
set_huge_pte_at(mm, address, pvmw.pte,
mm/rmap.c
2659
set_pte_at(mm, address, pvmw.pte, pteval);
mm/rmap.c
2697
set_huge_pte_at(mm, address, pvmw.pte, swp_pte,
mm/rmap.c
2700
set_pte_at(mm, address, pvmw.pte, swp_pte);
mm/rmap.c
2868
fw.level != FW_LEVEL_PTE || !pte_write(fw.pte)) {
mm/rmap.c
2879
fw.pte = ptep_clear_flush(vma, addr, fw.ptep);
mm/rmap.c
2882
if (pte_dirty(fw.pte))
mm/rmap.c
2892
if (pte_soft_dirty(fw.pte))
mm/rmap.c
934
if (pvmw.pte && ptes != pvmw.nr_pages)
mm/rmap.c
968
if (lru_gen_enabled() && pvmw.pte) {
mm/rmap.c
971
} else if (pvmw.pte) {
mm/rmap.c
975
pte_t pteval = ptep_get(pvmw.pte);
mm/rmap.c
977
nr = folio_pte_batch(folio, pvmw.pte,
mm/rmap.c
982
if (clear_flush_young_ptes_notify(vma, address, pvmw.pte, nr))
mm/rmap.c
985
pvmw.pte += nr - 1;
mm/sparse-vmemmap.c
143
void __meminit vmemmap_verify(pte_t *pte, int node,
mm/sparse-vmemmap.c
146
unsigned long pfn = pte_pfn(ptep_get(pte));
mm/sparse-vmemmap.c
158
pte_t *pte = pte_offset_kernel(pmd, addr);
mm/sparse-vmemmap.c
159
if (pte_none(ptep_get(pte))) {
mm/sparse-vmemmap.c
182
set_pte_at(&init_mm, addr, pte, entry);
mm/sparse-vmemmap.c
184
return pte;
mm/sparse-vmemmap.c
258
pte_t *pte;
mm/sparse-vmemmap.c
272
pte = vmemmap_pte_populate(pmd, addr, node, altmap, ptpfn, flags);
mm/sparse-vmemmap.c
273
if (!pte)
mm/sparse-vmemmap.c
275
vmemmap_verify(pte, node, addr, addr + PAGE_SIZE);
mm/sparse-vmemmap.c
277
return pte;
mm/sparse-vmemmap.c
287
pte_t *pte;
mm/sparse-vmemmap.c
290
pte = vmemmap_populate_address(addr, node, altmap,
mm/sparse-vmemmap.c
292
if (!pte)
mm/sparse-vmemmap.c
323
pte_t *pte;
mm/sparse-vmemmap.c
338
pte = virt_to_kpte(maddr);
mm/sparse-vmemmap.c
339
pte_clear(&init_mm, maddr, pte);
mm/sparse-vmemmap.c
347
pte = virt_to_kpte(maddr);
mm/sparse-vmemmap.c
348
pfn = pte_pfn(ptep_get(pte));
mm/sparse-vmemmap.c
349
pte_clear(&init_mm, maddr, pte);
mm/sparse-vmemmap.c
373
pte_t *pte;
mm/sparse-vmemmap.c
376
pte = virt_to_kpte(maddr);
mm/sparse-vmemmap.c
377
ptep_set_wrprotect(&init_mm, maddr, pte);
mm/sparse-vmemmap.c
389
pte_t *pte;
mm/sparse-vmemmap.c
393
pte = vmemmap_populate_address(maddr, node, NULL, -1, 0);
mm/sparse-vmemmap.c
394
if (!pte)
mm/sparse-vmemmap.c
402
pte_pfn(ptep_get(pte)), 0);
mm/sparse-vmemmap.c
491
pte_t *pte;
mm/sparse-vmemmap.c
499
pte = pte_offset_kernel(pmd_off_k(addr), addr);
mm/sparse-vmemmap.c
500
if (!pte)
mm/sparse-vmemmap.c
503
return pte;
mm/sparse-vmemmap.c
512
pte_t *pte;
mm/sparse-vmemmap.c
516
pte = compound_section_tail_page(start);
mm/sparse-vmemmap.c
517
if (!pte)
mm/sparse-vmemmap.c
525
pte_pfn(ptep_get(pte)),
mm/sparse-vmemmap.c
534
pte = vmemmap_populate_address(addr, node, NULL, -1, 0);
mm/sparse-vmemmap.c
535
if (!pte)
mm/sparse-vmemmap.c
540
pte = vmemmap_populate_address(next, node, NULL, -1, 0);
mm/sparse-vmemmap.c
541
if (!pte)
mm/sparse-vmemmap.c
550
pte_pfn(ptep_get(pte)),
mm/swap_state.c
834
pte_t *pte = NULL, pentry;
mm/swap_state.c
851
if (!pte++) {
mm/swap_state.c
852
pte = pte_offset_map(vmf->pmd, addr);
mm/swap_state.c
853
if (!pte)
mm/swap_state.c
856
pentry = ptep_get_lockless(pte);
mm/swap_state.c
861
pte_unmap(pte);
mm/swap_state.c
862
pte = NULL;
mm/swap_state.c
887
if (pte)
mm/swap_state.c
888
pte_unmap(pte);
mm/swapfile.c
2070
static inline int pte_same_as_swp(pte_t pte, pte_t swp_pte)
mm/swapfile.c
2072
return pte_same(pte_swp_clear_flags(pte), swp_pte);
mm/swapfile.c
2086
pte_t *pte, new_pte, old_pte;
mm/swapfile.c
2110
pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl);
mm/swapfile.c
2111
if (unlikely(!pte || !pte_same_as_swp(ptep_get(pte),
mm/swapfile.c
2117
old_pte = ptep_get(pte);
mm/swapfile.c
2176
set_pte_at(vma->vm_mm, addr, pte, new_pte);
mm/swapfile.c
2179
if (pte)
mm/swapfile.c
2180
pte_unmap_unlock(pte, ptl);
mm/swapfile.c
2192
pte_t *pte = NULL;
mm/swapfile.c
2204
if (!pte++) {
mm/swapfile.c
2205
pte = pte_offset_map(pmd, addr);
mm/swapfile.c
2206
if (!pte)
mm/swapfile.c
2210
ptent = ptep_get_lockless(pte);
mm/swapfile.c
2219
pte_unmap(pte);
mm/swapfile.c
2220
pte = NULL;
mm/swapfile.c
2255
if (pte)
mm/swapfile.c
2256
pte_unmap(pte);
mm/swapfile.c
3144
const pte_t pte = softleaf_to_pte(entry);
mm/swapfile.c
3150
entry = __pte_to_swp_entry(pte);
mm/util.c
1429
unsigned int folio_pte_batch(struct folio *folio, pte_t *ptep, pte_t pte,
mm/util.c
1432
return folio_pte_batch_flags(folio, NULL, ptep, &pte, max_nr, 0);
mm/vmalloc.c
107
pte = pte_alloc_kernel_track(pmd, addr, mask);
mm/vmalloc.c
108
if (!pte)
mm/vmalloc.c
114
if (unlikely(!pte_none(ptep_get(pte)))) {
mm/vmalloc.c
128
set_huge_pte_at(&init_mm, addr, pte, entry, size);
mm/vmalloc.c
133
set_pte_at(&init_mm, addr, pte, pfn_pte(pfn, prot));
mm/vmalloc.c
135
} while (pte += PFN_DOWN(size), addr += size, addr != end);
mm/vmalloc.c
3581
static int vmap_pfn_apply(pte_t *pte, unsigned long addr, void *private)
mm/vmalloc.c
3591
set_pte_at(&init_mm, addr, pte, ptent);
mm/vmalloc.c
369
pte_t *pte;
mm/vmalloc.c
373
pte = pte_offset_kernel(pmd, addr);
mm/vmalloc.c
378
size = arch_vmap_pte_range_unmap_size(addr, pte);
mm/vmalloc.c
382
pte = PTR_ALIGN_DOWN(pte, sizeof(*pte) * (size >> PAGE_SHIFT));
mm/vmalloc.c
384
ptent = huge_ptep_get_and_clear(&init_mm, addr, pte, size);
mm/vmalloc.c
389
ptent = ptep_get_and_clear(&init_mm, addr, pte);
mm/vmalloc.c
391
} while (pte += (size >> PAGE_SHIFT), addr += size, addr != end);
mm/vmalloc.c
530
pte_t *pte;
mm/vmalloc.c
537
pte = pte_alloc_kernel_track(pmd, addr, mask);
mm/vmalloc.c
538
if (!pte)
mm/vmalloc.c
546
if (WARN_ON(!pte_none(ptep_get(pte)))) {
mm/vmalloc.c
559
set_pte_at(&init_mm, addr, pte, mk_pte(page, prot));
mm/vmalloc.c
561
} while (pte++, addr += PAGE_SIZE, addr != end);
mm/vmalloc.c
807
pte_t *ptep, pte;
mm/vmalloc.c
847
pte = ptep_get(ptep);
mm/vmalloc.c
848
if (pte_present(pte))
mm/vmalloc.c
849
page = pte_page(pte);
mm/vmalloc.c
98
pte_t *pte;
mm/vmscan.c
3397
static unsigned long get_pte_pfn(pte_t pte, struct vm_area_struct *vma, unsigned long addr,
mm/vmscan.c
3400
unsigned long pfn = pte_pfn(pte);
mm/vmscan.c
3404
if (!pte_present(pte) || is_zero_pfn(pfn))
mm/vmscan.c
3407
if (WARN_ON_ONCE(pte_special(pte)))
mm/vmscan.c
3410
if (!pte_young(pte) && !mm_has_notifiers(vma->vm_mm))
mm/vmscan.c
3498
pte_t *pte;
mm/vmscan.c
3511
pte = pte_offset_map_rw_nolock(args->mm, pmd, start & PMD_MASK, &pmdval, &ptl);
mm/vmscan.c
3512
if (!pte)
mm/vmscan.c
3516
pte_unmap(pte);
mm/vmscan.c
3521
pte_unmap_unlock(pte, ptl);
mm/vmscan.c
3530
pte_t ptent = ptep_get(pte + i);
mm/vmscan.c
3543
if (!ptep_clear_young_notify(args->vma, addr, pte + i))
mm/vmscan.c
3567
pte_unmap_unlock(pte, ptl);
mm/vmscan.c
4210
pte_t *pte = pvmw->pte;
mm/vmscan.c
4224
if (!ptep_clear_young_notify(vma, addr, pte))
mm/vmscan.c
4256
pte -= (addr - start) / PAGE_SIZE;
mm/vmscan.c
4260
pte_t ptent = ptep_get(pte + i);
mm/vmscan.c
4270
if (!ptep_clear_young_notify(vma, addr, pte + i))
sound/soc/codecs/cs35l56-shared.c
891
u8 pte[2];
sound/soc/codecs/cs35l56-shared.c
901
struct cs35l56_pte pte;
sound/soc/codecs/cs35l56-shared.c
905
ret = regmap_raw_read(cs35l56_base->regmap, CS35L56_OTP_MEM_53, &pte, sizeof(pte));
sound/soc/codecs/cs35l56-shared.c
911
unique_id = (u32)pte.lot[2] | ((u32)pte.lot[1] << 8) | ((u32)pte.lot[0] << 16);
sound/soc/codecs/cs35l56-shared.c
913
unique_id |= (u32)pte.x | ((u32)pte.y << 8) | ((u32)pte.wafer_id << 16) |
sound/soc/codecs/cs35l56-shared.c
914
((u32)pte.dvs << 24);
tools/testing/selftests/kvm/include/x86/processor.h
1486
#define is_present_pte(mmu, pte) \
tools/testing/selftests/kvm/include/x86/processor.h
1488
!!(*(pte) & PTE_PRESENT_MASK(mmu)) : \
tools/testing/selftests/kvm/include/x86/processor.h
1489
!!(*(pte) & (PTE_READABLE_MASK(mmu) | PTE_EXECUTABLE_MASK(mmu))))
tools/testing/selftests/kvm/include/x86/processor.h
1490
#define is_executable_pte(mmu, pte) \
tools/testing/selftests/kvm/include/x86/processor.h
1491
((*(pte) & (PTE_EXECUTABLE_MASK(mmu) | PTE_NX_MASK(mmu))) == PTE_EXECUTABLE_MASK(mmu))
tools/testing/selftests/kvm/include/x86/processor.h
1492
#define is_writable_pte(mmu, pte) (!!(*(pte) & PTE_WRITABLE_MASK(mmu)))
tools/testing/selftests/kvm/include/x86/processor.h
1493
#define is_user_pte(mmu, pte) (!!(*(pte) & PTE_USER_MASK(mmu)))
tools/testing/selftests/kvm/include/x86/processor.h
1494
#define is_accessed_pte(mmu, pte) (!!(*(pte) & PTE_ACCESSED_MASK(mmu)))
tools/testing/selftests/kvm/include/x86/processor.h
1495
#define is_dirty_pte(mmu, pte) (!!(*(pte) & PTE_DIRTY_MASK(mmu)))
tools/testing/selftests/kvm/include/x86/processor.h
1496
#define is_huge_pte(mmu, pte) (!!(*(pte) & PTE_HUGE_MASK(mmu)))
tools/testing/selftests/kvm/include/x86/processor.h
1497
#define is_nx_pte(mmu, pte) (!is_executable_pte(mmu, pte))
tools/testing/selftests/kvm/include/x86/processor.h
376
#define PTE_GET_PA(pte) ((pte) & PHYSICAL_PAGE_MASK)
tools/testing/selftests/kvm/include/x86/processor.h
377
#define PTE_GET_PFN(pte) (PTE_GET_PA(pte) >> PAGE_SHIFT)
tools/testing/selftests/kvm/lib/arm64/processor.c
244
uint64_t pte, *ptep;
tools/testing/selftests/kvm/lib/arm64/processor.c
249
for (pte = page; pte < page + ptrs_per_pte(vm) * 8; pte += 8) {
tools/testing/selftests/kvm/lib/arm64/processor.c
250
ptep = addr_gpa2hva(vm, pte);
tools/testing/selftests/kvm/lib/arm64/processor.c
253
fprintf(stream, "%*s%s: %lx: %lx at %p\n", indent, "", type[level], pte, *ptep, ptep);
tools/testing/selftests/kvm/lib/arm64/processor.c
68
uint64_t pte;
tools/testing/selftests/kvm/lib/arm64/processor.c
71
pte = pa & PTE_ADDR_MASK_LPA2(vm->page_shift);
tools/testing/selftests/kvm/lib/arm64/processor.c
72
pte |= FIELD_GET(GENMASK(51, 50), pa) << PTE_ADDR_51_50_LPA2_SHIFT;
tools/testing/selftests/kvm/lib/arm64/processor.c
75
pte = pa & PTE_ADDR_MASK(vm->page_shift);
tools/testing/selftests/kvm/lib/arm64/processor.c
77
pte |= FIELD_GET(GENMASK(51, 48), pa) << PTE_ADDR_51_48_SHIFT;
tools/testing/selftests/kvm/lib/arm64/processor.c
79
pte |= attrs;
tools/testing/selftests/kvm/lib/arm64/processor.c
81
return pte;
tools/testing/selftests/kvm/lib/arm64/processor.c
84
static uint64_t pte_addr(struct kvm_vm *vm, uint64_t pte)
tools/testing/selftests/kvm/lib/arm64/processor.c
89
pa = pte & PTE_ADDR_MASK_LPA2(vm->page_shift);
tools/testing/selftests/kvm/lib/arm64/processor.c
90
pa |= FIELD_GET(PTE_ADDR_51_50_LPA2, pte) << 50;
tools/testing/selftests/kvm/lib/arm64/processor.c
92
pa = pte & PTE_ADDR_MASK(vm->page_shift);
tools/testing/selftests/kvm/lib/arm64/processor.c
94
pa |= FIELD_GET(PTE_ADDR_51_48, pte) << 48;
tools/testing/selftests/kvm/lib/loongarch/processor.c
144
uint64_t pte, *ptep;
tools/testing/selftests/kvm/lib/loongarch/processor.c
150
for (pte = page; pte < page + ptrs_per_pte(vm) * 8; pte += 8) {
tools/testing/selftests/kvm/lib/loongarch/processor.c
151
ptep = addr_gpa2hva(vm, pte);
tools/testing/selftests/kvm/lib/loongarch/processor.c
155
indent, "", type[level], pte, *ptep, ptep);
tools/testing/selftests/kvm/lib/riscv/processor.c
156
uint64_t pte, *ptep;
tools/testing/selftests/kvm/lib/riscv/processor.c
161
for (pte = page; pte < page + ptrs_per_pte(vm) * 8; pte += 8) {
tools/testing/selftests/kvm/lib/riscv/processor.c
162
ptep = addr_gpa2hva(vm, pte);
tools/testing/selftests/kvm/lib/riscv/processor.c
166
type[level], pte, *ptep, ptep);
tools/testing/selftests/kvm/lib/s390/processor.c
117
uint64_t *pte, ptea;
tools/testing/selftests/kvm/lib/s390/processor.c
120
pte = addr_gpa2hva(vm, ptea);
tools/testing/selftests/kvm/lib/s390/processor.c
121
if (*pte & PAGE_INVALID)
tools/testing/selftests/kvm/lib/s390/processor.c
124
indent, "", ptea, *pte);
tools/testing/selftests/kvm/lib/x86/processor.c
229
uint64_t *pte = virt_get_pte(vm, mmu, parent_pte, vaddr, current_level);
tools/testing/selftests/kvm/lib/x86/processor.c
233
if (!is_present_pte(mmu, pte)) {
tools/testing/selftests/kvm/lib/x86/processor.c
234
*pte = PTE_PRESENT_MASK(mmu) | PTE_READABLE_MASK(mmu) |
tools/testing/selftests/kvm/lib/x86/processor.c
238
*pte |= PTE_HUGE_MASK(mmu) | (paddr & PHYSICAL_PAGE_MASK);
tools/testing/selftests/kvm/lib/x86/processor.c
240
*pte |= vm_alloc_page_table(vm) & PHYSICAL_PAGE_MASK;
tools/testing/selftests/kvm/lib/x86/processor.c
250
TEST_ASSERT(!is_huge_pte(mmu, pte),
tools/testing/selftests/kvm/lib/x86/processor.c
254
return pte;
tools/testing/selftests/kvm/lib/x86/processor.c
261
uint64_t *pte = &mmu->pgd;
tools/testing/selftests/kvm/lib/x86/processor.c
292
pte = virt_create_upper_pte(vm, mmu, pte, vaddr, paddr,
tools/testing/selftests/kvm/lib/x86/processor.c
294
if (is_huge_pte(mmu, pte))
tools/testing/selftests/kvm/lib/x86/processor.c
299
pte = virt_get_pte(vm, mmu, pte, vaddr, PG_LEVEL_4K);
tools/testing/selftests/kvm/lib/x86/processor.c
300
TEST_ASSERT(!is_present_pte(mmu, pte),
tools/testing/selftests/kvm/lib/x86/processor.c
302
*pte = PTE_PRESENT_MASK(mmu) | PTE_READABLE_MASK(mmu) |
tools/testing/selftests/kvm/lib/x86/processor.c
311
*pte |= PTE_C_BIT_MASK(mmu);
tools/testing/selftests/kvm/lib/x86/processor.c
313
*pte |= PTE_S_BIT_MASK(mmu);
tools/testing/selftests/kvm/lib/x86/processor.c
342
static bool vm_is_target_pte(struct kvm_mmu *mmu, uint64_t *pte,
tools/testing/selftests/kvm/lib/x86/processor.c
345
if (is_huge_pte(mmu, pte)) {
tools/testing/selftests/kvm/lib/x86/processor.c
361
uint64_t *pte = &mmu->pgd;
tools/testing/selftests/kvm/lib/x86/processor.c
386
pte = virt_get_pte(vm, mmu, pte, vaddr, current_level);
tools/testing/selftests/kvm/lib/x86/processor.c
387
if (vm_is_target_pte(mmu, pte, level, current_level))
tools/testing/selftests/kvm/lib/x86/processor.c
388
return pte;
tools/testing/selftests/kvm/lib/x86/processor.c
391
return virt_get_pte(vm, mmu, pte, vaddr, PG_LEVEL_4K);
tools/testing/selftests/kvm/lib/x86/processor.c
414
uint64_t *pte, *pte_start;
tools/testing/selftests/kvm/lib/x86/processor.c
463
pte = &pte_start[n4];
tools/testing/selftests/kvm/lib/x86/processor.c
464
if (!is_present_pte(mmu, pte))
tools/testing/selftests/kvm/lib/x86/processor.c
470
pte - pte_start, pte,
tools/testing/selftests/kvm/lib/x86/processor.c
471
addr_hva2gpa(vm, pte),
tools/testing/selftests/kvm/lib/x86/processor.c
472
PTE_GET_PFN(*pte),
tools/testing/selftests/kvm/lib/x86/processor.c
473
is_writable_pte(mmu, pte),
tools/testing/selftests/kvm/lib/x86/processor.c
474
is_nx_pte(mmu, pte),
tools/testing/selftests/kvm/lib/x86/processor.c
475
is_dirty_pte(mmu, pte),
tools/testing/selftests/kvm/lib/x86/processor.c
622
uint64_t *pte = __vm_get_page_table_entry(vm, &vm->mmu, gva, &level);
tools/testing/selftests/kvm/lib/x86/processor.c
624
TEST_ASSERT(is_present_pte(&vm->mmu, pte),
tools/testing/selftests/kvm/lib/x86/processor.c
631
return vm_untag_gpa(vm, PTE_GET_PA(*pte)) | (gva & ~HUGEPAGE_MASK(level));
tools/testing/selftests/kvm/x86/hyperv_tlb_flush.c
586
uint64_t *pte;
tools/testing/selftests/kvm/x86/hyperv_tlb_flush.c
622
pte = vm_get_pte(vm, data->test_pages + i * PAGE_SIZE);
tools/testing/selftests/kvm/x86/hyperv_tlb_flush.c
623
gpa = addr_hva2gpa(vm, pte);
tools/testing/selftests/mm/uffd-wp-mremap.c
131
uint64_t pte;
tools/testing/selftests/mm/uffd-wp-mremap.c
135
pte = pagemap_get_entry(pagemap_fd, addr);
tools/testing/selftests/mm/uffd-wp-mremap.c
136
if (!!(pte & PM_UFFD_WP) != expect) {
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
173
printf("PTE: 0x%016lx\n", mapping.pte);
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
177
ASSERT_NE(0, mapping.pte);
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
180
ASSERT_EQ(0, mapping.pte);
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
184
ASSERT_EQ(0, mapping.pte);
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
23
u64 pte;
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
72
parse_next_value(&rest, &mapping->pte);