Symbol: PAGE_SHIFT
arch/alpha/include/asm/mmu_context.h
203
= ((unsigned long)mm->pgd - IDENT_ADDR) >> PAGE_SHIFT;
arch/alpha/include/asm/mmu_context.h
212
= ((unsigned long)mm->pgd - IDENT_ADDR) >> PAGE_SHIFT;
arch/alpha/include/asm/page.h
81
#define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
arch/alpha/include/asm/page.h
82
#define virt_addr_valid(kaddr) pfn_valid((__pa(kaddr) >> PAGE_SHIFT))
arch/alpha/include/asm/pgtable.h
150
#define KSEG_PFN (0xc0000000000UL >> PAGE_SHIFT)
arch/alpha/include/asm/pgtable.h
152
((((pfn) & KSEG_PFN) == (0x40000000000UL >> PAGE_SHIFT)) \
arch/alpha/include/asm/pgtable.h
162
#define page_to_pa(page) (page_to_pfn(page) << PAGE_SHIFT)
arch/alpha/include/asm/pgtable.h
175
{ pmd_val(*pmdp) = _PAGE_TABLE | ((((unsigned long) ptep) - PAGE_OFFSET) << (32-PAGE_SHIFT)); }
arch/alpha/include/asm/pgtable.h
178
{ pud_val(*pudp) = _PAGE_TABLE | ((((unsigned long) pmdp) - PAGE_OFFSET) << (32-PAGE_SHIFT)); }
arch/alpha/include/asm/pgtable.h
187
return ((pmd_val(pmd) & _PFN_MASK) >> (32-PAGE_SHIFT)) + PAGE_OFFSET;
arch/alpha/include/asm/pgtable.h
196
return (pmd_t *)(PAGE_OFFSET + ((pud_val(pgd) & _PFN_MASK) >> (32-PAGE_SHIFT)));
arch/alpha/include/asm/pgtable.h
257
+ ((address >> PAGE_SHIFT) & (PTRS_PER_PAGE - 1));
arch/alpha/include/asm/pgtable.h
32
#define PMD_SHIFT (PAGE_SHIFT + (PAGE_SHIFT-3))
arch/alpha/include/asm/pgtable.h
37
#define PGDIR_SHIFT (PAGE_SHIFT + 2*(PAGE_SHIFT-3))
arch/alpha/include/asm/pgtable.h
45
#define PTRS_PER_PTE (1UL << (PAGE_SHIFT-3))
arch/alpha/include/asm/pgtable.h
46
#define PTRS_PER_PMD (1UL << (PAGE_SHIFT-3))
arch/alpha/include/asm/pgtable.h
47
#define PTRS_PER_PGD (1UL << (PAGE_SHIFT-3))
arch/alpha/include/asm/pgtable.h
51
#define PTRS_PER_PAGE (1UL << (PAGE_SHIFT-3))
arch/alpha/kernel/core_cia.c
335
pte = (virt_to_phys(ppte) >> (PAGE_SHIFT - 1)) | 1;
arch/alpha/kernel/core_cia.c
374
pte0 = (virt_to_phys(page) >> (PAGE_SHIFT - 1)) | 1;
arch/alpha/kernel/core_cia.c
760
} else if (max_low_pfn > (0x100000000UL >> PAGE_SHIFT)) {
arch/alpha/kernel/core_irongate.c
204
unsigned long memtop = max_low_pfn << PAGE_SHIFT;
arch/alpha/kernel/core_marvel.c
1000
pte = aper->arena->ptes[baddr >> PAGE_SHIFT];
arch/alpha/kernel/core_marvel.c
1005
return (pte >> 1) << PAGE_SHIFT;
arch/alpha/kernel/core_marvel.c
746
pfn = ptes[baddr >> PAGE_SHIFT];
arch/alpha/kernel/core_marvel.c
755
pfn << PAGE_SHIFT,
arch/alpha/kernel/core_titan.c
522
pfn = ptes[baddr >> PAGE_SHIFT];
arch/alpha/kernel/core_titan.c
531
pfn << PAGE_SHIFT,
arch/alpha/kernel/core_titan.c
711
pte = aper->arena->ptes[baddr >> PAGE_SHIFT];
arch/alpha/kernel/core_titan.c
717
return (pte >> 1) << PAGE_SHIFT;
arch/alpha/kernel/osf_sys.c
192
ret = ksys_mmap_pgoff(addr, len, prot, flags, fd, off >> PAGE_SHIFT);
arch/alpha/kernel/pci-sysfs.c
264
size = ((res_size - 1) >> PAGE_SHIFT) + 1;
arch/alpha/kernel/pci-sysfs.c
29
vma->vm_pgoff += base >> PAGE_SHIFT;
arch/alpha/kernel/pci-sysfs.c
44
size = ((pci_resource_len(pdev, num) - 1) >> (PAGE_SHIFT - shift)) + 1;
arch/alpha/kernel/pci-sysfs.c
89
vma->vm_pgoff += bar.start >> (PAGE_SHIFT - (sparse ? 5 : 0));
arch/alpha/kernel/pci_iommu.c
109
base = arena->dma_base >> PAGE_SHIFT;
arch/alpha/kernel/pci_iommu.c
110
boundary_size = dma_get_seg_boundary_nr_pages(dev, PAGE_SHIFT);
arch/alpha/kernel/pci_iommu.c
114
nent = arena->size >> PAGE_SHIFT;
arch/alpha/kernel/pci_iommu.c
373
dma_ofs = (dma_addr - arena->dma_base) >> PAGE_SHIFT;
arch/alpha/kernel/pci_iommu.c
44
return (paddr >> (PAGE_SHIFT-1)) | 1;
arch/alpha/kernel/pci_iommu.c
53
unsigned long mem = max_low_pfn << PAGE_SHIFT;
arch/alpha/kernel/pci_iommu.c
761
ofs = (addr - arena->dma_base) >> PAGE_SHIFT;
arch/alpha/kernel/pci_iommu.c
772
if ((fend - arena->dma_base) >> PAGE_SHIFT >= arena->next_entry)
arch/alpha/kernel/pci_iommu.c
794
__direct_map_base + (max_low_pfn << PAGE_SHIFT) - 1 <= mask))
arch/alpha/kernel/setup.c
1168
if (maxsize > (max_low_pfn + 1) << PAGE_SHIFT)
arch/alpha/kernel/setup.c
1169
maxsize = 1 << (ilog2(max_low_pfn + 1) + PAGE_SHIFT);
arch/alpha/kernel/setup.c
261
return end >> PAGE_SHIFT; /* Return the PFN of the limit. */
arch/alpha/kernel/setup.c
309
cluster->numpages << PAGE_SHIFT);
arch/alpha/kernel/setup.c
316
cluster->numpages << PAGE_SHIFT);
arch/alpha/kernel/setup.c
335
mem_size_limit = (32ul * 1024 * 1024 * 1024) >> PAGE_SHIFT;
arch/alpha/kernel/setup.c
340
mem_size_limit << (PAGE_SHIFT - 10),
arch/alpha/kernel/setup.c
341
max_low_pfn << (PAGE_SHIFT - 10));
arch/alpha/kernel/setup.c
502
get_mem_size_limit(p+9) << PAGE_SHIFT;
arch/alpha/kernel/sys_nautilus.c
205
unsigned long memtop = max_low_pfn << PAGE_SHIFT;
arch/alpha/mm/fault.c
52
pcb->ptbr = ((unsigned long) next_mm->pgd - IDENT_ADDR) >> PAGE_SHIFT;
arch/alpha/mm/init.c
179
console_remap_vm.size = nr_pages << PAGE_SHIFT;
arch/alpha/mm/init.c
187
unsigned long pfn = crb->map[i].pa >> PAGE_SHIFT;
arch/alpha/mm/init.c
215
dma_pfn = virt_to_phys((char *)MAX_DMA_ADDRESS) >> PAGE_SHIFT;
arch/alpha/mm/init.c
82
newptbr = ((unsigned long) swapper_pg_dir - PAGE_OFFSET) >> PAGE_SHIFT;
arch/arc/include/asm/highmem.h
24
#define FIXADDR_TOP (FIXMAP_BASE + (FIX_KMAP_END << PAGE_SHIFT))
arch/arc/include/asm/highmem.h
30
#define __fix_to_virt(x) (FIXADDR_TOP - ((x) << PAGE_SHIFT))
arch/arc/include/asm/highmem.h
31
#define __virt_to_fix(x) (((FIXADDR_TOP - ((x) & PAGE_MASK))) >> PAGE_SHIFT)
arch/arc/include/asm/highmem.h
35
#define LAST_PKMAP (PKMAP_SIZE >> PAGE_SHIFT)
arch/arc/include/asm/highmem.h
37
#define PKMAP_ADDR(nr) (PKMAP_BASE + ((nr) << PAGE_SHIFT))
arch/arc/include/asm/highmem.h
38
#define PKMAP_NR(virt) (((virt) - PKMAP_BASE) >> PAGE_SHIFT)
arch/arc/include/asm/hugepage.h
45
#define pfn_pmd(pfn, prot) (__pmd(((pfn) << PAGE_SHIFT) | pgprot_val(prot)))
arch/arc/include/asm/page.h
127
return __pa(kaddr) >> PAGE_SHIFT;
arch/arc/include/asm/pgtable-levels.h
143
#define pmd_pfn(pmd) ((pmd_val(pmd) & PMD_MASK) >> PAGE_SHIFT)
arch/arc/include/asm/pgtable-levels.h
144
#define pfn_pmd(pfn,prot) __pmd(((pfn) << PAGE_SHIFT) | pgprot_val(prot))
arch/arc/include/asm/pgtable-levels.h
160
#define pmd_pfn(pmd) ((pmd_val(pmd) & PAGE_MASK) >> PAGE_SHIFT)
arch/arc/include/asm/pgtable-levels.h
171
#define PFN_PTE_SHIFT PAGE_SHIFT
arch/arc/include/asm/pgtable-levels.h
177
#define pte_pfn(pte) (pte_val(pte) >> PAGE_SHIFT)
arch/arc/include/asm/pgtable-levels.h
86
#define PTRS_PER_PTE BIT(PMD_SHIFT - PAGE_SHIFT)
arch/arc/include/asm/thread_info.h
25
#define THREAD_SHIFT (PAGE_SHIFT << THREAD_SIZE_ORDER)
arch/arc/include/uapi/asm/page.h
27
#define PAGE_SIZE _BITUL(PAGE_SHIFT) /* Default 8K */
arch/arc/mm/cache.c
829
phy = (pfn << PAGE_SHIFT) + off;
arch/arc/mm/mmap.c
38
(addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1))
arch/arc/mm/mmap.c
58
info.align_offset = pgoff << PAGE_SHIFT;
arch/arm/include/asm/cacheflush.h
17
#define CACHE_COLOUR(vaddr) ((vaddr & (SHMLBA - 1)) >> PAGE_SHIFT)
arch/arm/include/asm/highmem.h
11
#define PKMAP_NR(virt) (((virt) - PKMAP_BASE) >> PAGE_SHIFT)
arch/arm/include/asm/highmem.h
12
#define PKMAP_ADDR(nr) (PKMAP_BASE + ((nr) << PAGE_SHIFT))
arch/arm/include/asm/kexec.h
71
return page_to_pfn(page) + (arch_phys_to_idmap_offset >> PAGE_SHIFT);
arch/arm/include/asm/kexec.h
77
return pfn_to_page(boot_pfn - (arch_phys_to_idmap_offset >> PAGE_SHIFT));
arch/arm/include/asm/memory.h
192
#define PHYS_OFFSET ((phys_addr_t)__pv_phys_pfn_offset << PAGE_SHIFT)
arch/arm/include/asm/memory.h
276
#define PHYS_PFN_OFFSET ((unsigned long)(PHYS_OFFSET >> PAGE_SHIFT))
arch/arm/include/asm/memory.h
293
return (((kaddr - PAGE_OFFSET) >> PAGE_SHIFT) +
arch/arm/include/asm/memory.h
330
#define pfn_to_kaddr(pfn) __va((phys_addr_t)(pfn) << PAGE_SHIFT)
arch/arm/include/asm/pgtable-3level.h
210
#define pmd_pfn(pmd) (((pmd_val(pmd) & PMD_MASK) & PHYS_MASK) >> PAGE_SHIFT)
arch/arm/include/asm/pgtable-3level.h
211
#define pfn_pmd(pfn,prot) (__pmd(((phys_addr_t)(pfn) << PAGE_SHIFT) | pgprot_val(prot)))
arch/arm/include/asm/pgtable-3level.h
60
#define HUGETLB_PAGE_ORDER (HPAGE_SHIFT - PAGE_SHIFT)
arch/arm/include/asm/pgtable.h
158
#define pte_pfn(pte) ((pte_val(pte) & PHYS_MASK) >> PAGE_SHIFT)
arch/arm/include/asm/pgtable.h
201
#define PFN_PTE_SHIFT PAGE_SHIFT
arch/arm/kernel/process.c
360
last = TASK_SIZE - (npages << PAGE_SHIFT);
arch/arm/kernel/process.c
370
slots = ((last - first) >> PAGE_SHIFT) + 1;
arch/arm/kernel/process.c
374
addr = first + (offset << PAGE_SHIFT);
arch/arm/kernel/process.c
415
addr = get_unmapped_area(NULL, hint, npages << PAGE_SHIFT, 0, 0);
arch/arm/kernel/setup.c
988
return total << PAGE_SHIFT;
arch/arm/kernel/vdso.c
179
text_pages = (vdso_end - vdso_start) >> PAGE_SHIFT;
arch/arm/kernel/vdso.c
223
len = (vdso_total_pages - VDSO_NR_PAGES) << PAGE_SHIFT;
arch/arm/mm/cache-feroceon-l2.c
110
BUG_ON((start ^ end) >> PAGE_SHIFT);
arch/arm/mm/cache-feroceon-l2.c
51
void *vaddr = kmap_atomic_pfn(paddr >> PAGE_SHIFT);
arch/arm/mm/cache-feroceon-l2.c
79
BUG_ON((start ^ end) >> PAGE_SHIFT);
arch/arm/mm/cache-xsc3l2.c
70
unsigned long pa_offset = pa << (32 - PAGE_SHIFT);
arch/arm/mm/cache-xsc3l2.c
71
if (unlikely(pa_offset < (prev_va << (32 - PAGE_SHIFT)))) {
arch/arm/mm/cache-xsc3l2.c
78
va = (unsigned long)kmap_atomic_pfn(pa >> PAGE_SHIFT);
arch/arm/mm/cache-xsc3l2.c
80
return va + (pa_offset >> (32 - PAGE_SHIFT));
arch/arm/mm/copypage-v6.c
106
unsigned long to = COPYPAGE_V6_TO + (CACHE_COLOUR(vaddr) << PAGE_SHIFT);
arch/arm/mm/copypage-v6.c
88
kfrom = COPYPAGE_V6_FROM + (offset << PAGE_SHIFT);
arch/arm/mm/copypage-v6.c
89
kto = COPYPAGE_V6_TO + (offset << PAGE_SHIFT);
arch/arm/mm/dma-mapping.c
1118
unsigned long nr_pages = PAGE_ALIGN(size) >> PAGE_SHIFT;
arch/arm/mm/dma-mapping.c
1170
unsigned int count = PAGE_ALIGN(size) >> PAGE_SHIFT;
arch/arm/mm/dma-mapping.c
1201
for (count = 0, s = sg; count < (size >> PAGE_SHIFT); s = sg_next(s)) {
arch/arm/mm/dma-mapping.c
1214
count += len >> PAGE_SHIFT;
arch/arm/mm/dma-mapping.c
1489
unsigned int bits = size >> PAGE_SHIFT;
arch/arm/mm/dma-mapping.c
155
for (p = page + (size >> PAGE_SHIFT), e = page + (1 << order); p < e; p++)
arch/arm/mm/dma-mapping.c
168
struct page *e = page + (size >> PAGE_SHIFT);
arch/arm/mm/dma-mapping.c
207
atomic_pool = gen_pool_create(PAGE_SHIFT, -1);
arch/arm/mm/dma-mapping.c
397
size_t count = size >> PAGE_SHIFT;
arch/arm/mm/dma-mapping.c
435
dma_release_from_contiguous(dev, page, size >> PAGE_SHIFT);
arch/arm/mm/dma-mapping.c
760
size_t mapping_size = mapping->bits << PAGE_SHIFT;
arch/arm/mm/dma-mapping.c
768
count = PAGE_ALIGN(size) >> PAGE_SHIFT;
arch/arm/mm/dma-mapping.c
807
iova += start << PAGE_SHIFT;
arch/arm/mm/dma-mapping.c
816
size_t mapping_size = mapping->bits << PAGE_SHIFT;
arch/arm/mm/dma-mapping.c
829
start = (addr - bitmap_base) >> PAGE_SHIFT;
arch/arm/mm/dma-mapping.c
840
count = size >> PAGE_SHIFT;
arch/arm/mm/dma-mapping.c
855
int count = size >> PAGE_SHIFT;
arch/arm/mm/dma-mapping.c
941
int count = size >> PAGE_SHIFT;
arch/arm/mm/dma-mapping.c
964
unsigned int count = PAGE_ALIGN(size) >> PAGE_SHIFT;
arch/arm/mm/dma-mapping.c
984
len = (j - i) << PAGE_SHIFT;
arch/arm/mm/fault-armv.c
135
pgoff = vma->vm_pgoff + ((addr - vma->vm_start) >> PAGE_SHIFT);
arch/arm/mm/fault-armv.c
161
offset = (pgoff - mpnt->vm_pgoff) << PAGE_SHIFT;
arch/arm/mm/fault-armv.c
53
outer_flush_range((pfn << PAGE_SHIFT),
arch/arm/mm/fault-armv.c
54
(pfn << PAGE_SHIFT) + PAGE_SIZE);
arch/arm/mm/fault.c
91
if (PageHighMem(pfn_to_page(pmd_val(*pmd) >> PAGE_SHIFT)))
arch/arm/mm/flush.c
40
unsigned long to = FLUSH_ALIAS_START + (CACHE_COLOUR(vaddr) << PAGE_SHIFT);
arch/arm/mm/flush.c
54
unsigned long va = FLUSH_ALIAS_START + (CACHE_COLOUR(vaddr) << PAGE_SHIFT);
arch/arm/mm/init.c
106
arm_dma_pfn_limit = arm_dma_limit >> PAGE_SHIFT;
arch/arm/mm/init.c
208
early_memtest((phys_addr_t)min_low_pfn << PAGE_SHIFT,
arch/arm/mm/init.c
209
(phys_addr_t)max_low_pfn << PAGE_SHIFT);
arch/arm/mm/ioremap.c
227
pfn += SZ_1M >> PAGE_SHIFT;
arch/arm/mm/ioremap.c
229
pfn += SZ_1M >> PAGE_SHIFT;
arch/arm/mm/ioremap.c
256
super_pmd_val |= ((pfn >> (32 - PAGE_SHIFT)) & 0xf) << 20;
arch/arm/mm/ioremap.c
267
pfn += SUPERSECTION_SIZE >> PAGE_SHIFT;
arch/arm/mm/mm.h
84
#define arm_dma_pfn_limit (~0ul >> PAGE_SHIFT)
arch/arm/mm/mmap.c
105
(addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1))
arch/arm/mm/mmap.c
127
info.align_offset = pgoff << PAGE_SHIFT;
arch/arm/mm/mmap.c
166
return (pfn + (size >> PAGE_SHIFT)) <= (1 + (PHYS_MASK >> PAGE_SHIFT));
arch/arm/mm/mmap.c
18
(((pgoff)<<PAGE_SHIFT) & (SHMLBA-1)))
arch/arm/mm/mmap.c
52
(addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1))
arch/arm/mm/mmap.c
76
info.align_offset = pgoff << PAGE_SHIFT;
arch/arm/mm/mmu.c
406
pfn_pte(phys >> PAGE_SHIFT, prot));
arch/arm/mm/mmu.c
899
phys |= (((md->pfn >> (32 - PAGE_SHIFT)) & 0xF) << 20);
arch/arm/mm/nommu.c
189
if (pfn >= (0x100000000ULL >> PAGE_SHIFT))
arch/arm/mm/nommu.c
191
return (void __iomem *) (offset + (pfn << PAGE_SHIFT));
arch/arm64/include/asm/elf.h
192
0x7ff >> (PAGE_SHIFT - 12) : \
arch/arm64/include/asm/elf.h
193
0x3ffff >> (PAGE_SHIFT - 12))
arch/arm64/include/asm/elf.h
195
#define STACK_RND_MASK (0x3ffff >> (PAGE_SHIFT - 12))
arch/arm64/include/asm/fixmap.h
102
#define FIXADDR_SIZE (__end_of_permanent_fixed_addresses << PAGE_SHIFT)
arch/arm64/include/asm/fixmap.h
104
#define FIXADDR_TOT_SIZE (__end_of_fixed_addresses << PAGE_SHIFT)
arch/arm64/include/asm/kernel-pgtable.h
25
#define SWAPPER_BLOCK_SHIFT PAGE_SHIFT
arch/arm64/include/asm/kvm_arm.h
259
#define ARM64_VTTBR_X(ipa, levels) ((ipa) - ((levels) * (PAGE_SHIFT - 3)))
arch/arm64/include/asm/kvm_pgtable.h
125
if (PAGE_SHIFT == 16)
arch/arm64/include/asm/kvm_pgtable.h
142
if (PAGE_SHIFT == 16) {
arch/arm64/include/asm/kvm_pgtable.h
34
(IS_ENABLED(CONFIG_ARM64_PA_BITS_52) && PAGE_SHIFT == 16))
arch/arm64/include/asm/kvm_pgtable.h
55
#define KVM_PTE_ADDR_MASK GENMASK(47, PAGE_SHIFT)
arch/arm64/include/asm/kvm_pgtable.h
57
#define KVM_PTE_ADDR_MASK_LPA2 GENMASK(49, PAGE_SHIFT)
arch/arm64/include/asm/kvm_pkvm.h
100
return res >> PAGE_SHIFT;
arch/arm64/include/asm/kvm_pkvm.h
105
return PAGE_ALIGN(KVM_MAX_PVMS * sizeof(void *)) >> PAGE_SHIFT;
arch/arm64/include/asm/kvm_pkvm.h
129
res += __hyp_pgtable_max_pages(reg->size >> PAGE_SHIFT);
arch/arm64/include/asm/kvm_pkvm.h
142
res += __hyp_pgtable_max_pages(SZ_1G >> PAGE_SHIFT);
arch/arm64/include/asm/kvm_pkvm.h
158
res += __hyp_pgtable_max_pages(SZ_1G >> PAGE_SHIFT);
arch/arm64/include/asm/kvm_pkvm.h
80
unsigned long nr_pages = reg->size >> PAGE_SHIFT;
arch/arm64/include/asm/kvm_pkvm.h
83
start = (reg->base >> PAGE_SHIFT) * vmemmap_entry_size;
arch/arm64/include/asm/memory.h
121
#if (MIN_THREAD_SHIFT < PAGE_SHIFT)
arch/arm64/include/asm/memory.h
122
#define THREAD_SHIFT PAGE_SHIFT
arch/arm64/include/asm/memory.h
127
#if THREAD_SHIFT >= PAGE_SHIFT
arch/arm64/include/asm/memory.h
128
#define THREAD_SIZE_ORDER (THREAD_SHIFT - PAGE_SHIFT)
arch/arm64/include/asm/memory.h
144
#define NVHE_STACK_SHIFT PAGE_SHIFT
arch/arm64/include/asm/memory.h
280
#define PHYS_PFN_OFFSET (PHYS_OFFSET >> PAGE_SHIFT)
arch/arm64/include/asm/memory.h
34
#define VMEMMAP_SIZE ((VMEMMAP_RANGE >> PAGE_SHIFT) * sizeof(struct page))
arch/arm64/include/asm/memory.h
393
#define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT)
arch/arm64/include/asm/pgtable-hwdef.h
13
#define PTDESC_TABLE_SHIFT (PAGE_SHIFT - PTDESC_ORDER)
arch/arm64/include/asm/pgtable-hwdef.h
179
#define PTE_ADDR_LOW (((_AT(pteval_t, 1) << (50 - PAGE_SHIFT)) - 1) << PAGE_SHIFT)
arch/arm64/include/asm/pgtable-hwdef.h
90
#define CONT_PTE_SHIFT (CONFIG_ARM64_CONT_PTE_SHIFT + PAGE_SHIFT)
arch/arm64/include/asm/pgtable-hwdef.h
91
#define CONT_PTES (1 << (CONT_PTE_SHIFT - PAGE_SHIFT))
arch/arm64/include/asm/pgtable.h
136
#define pte_pfn(pte) (__pte_to_phys(pte) >> PAGE_SHIFT)
arch/arm64/include/asm/pgtable.h
138
__pte(__phys_to_pte_val((phys_addr_t)(pfn) << PAGE_SHIFT) | pgprot_val(prot))
arch/arm64/include/asm/pgtable.h
1619
#define exec_folio_order() ilog2(SZ_64K >> PAGE_SHIFT)
arch/arm64/include/asm/pgtable.h
1692
bool valign = ((addr >> PAGE_SHIFT) & contmask) == contmask;
arch/arm64/include/asm/pgtable.h
27
#define VMEMMAP_UNUSED_NPAGES ((_PAGE_OFFSET(vabits_actual) - PAGE_OFFSET) >> PAGE_SHIFT)
arch/arm64/include/asm/pgtable.h
31
#define vmemmap ((struct page *)VMEMMAP_START - (memstart_addr >> PAGE_SHIFT))
arch/arm64/include/asm/pgtable.h
477
#define HUGETLB_PAGE_ORDER (HPAGE_SHIFT - PAGE_SHIFT)
arch/arm64/include/asm/pgtable.h
630
#define pmd_pfn(pmd) ((__pmd_to_phys(pmd) & PMD_MASK) >> PAGE_SHIFT)
arch/arm64/include/asm/pgtable.h
631
#define pfn_pmd(pfn,prot) __pmd(__phys_to_pmd_val((phys_addr_t)(pfn) << PAGE_SHIFT) | pgprot_val(prot))
arch/arm64/include/asm/pgtable.h
654
#define pud_pfn(pud) ((__pud_to_phys(pud) & PUD_MASK) >> PAGE_SHIFT)
arch/arm64/include/asm/pgtable.h
655
#define pfn_pud(pfn,prot) __pud(__phys_to_pud_val((phys_addr_t)(pfn) << PAGE_SHIFT) | pgprot_val(prot))
arch/arm64/include/asm/pgtable.h
677
unsigned long stride = pgsize >> PAGE_SHIFT;
arch/arm64/include/asm/tlbflush.h
433
arg |= FIELD_PREP(TLBIR_BADDR_MASK, addr >> (lpa2 ? 16 : PAGE_SHIFT));
arch/arm64/include/asm/tlbflush.h
454
pages = (end - addr) >> PAGE_SHIFT;
arch/arm64/include/asm/tlbflush.h
465
addr += __TLBI_RANGE_PAGES(num, scale) << PAGE_SHIFT;
arch/arm64/include/asm/tlbflush.h
494
return pages >= (MAX_DVM_OPS * stride) >> PAGE_SHIFT;
arch/arm64/include/asm/tlbflush.h
522
pages = (end - start) >> PAGE_SHIFT;
arch/arm64/include/asm/tlbflush.h
611
pages = (end - start) >> PAGE_SHIFT;
arch/arm64/include/asm/vmalloc.h
63
return PAGE_SHIFT;
arch/arm64/kernel/elfcore.c
20
return (m->dump_size >> PAGE_SHIFT) * MTE_PAGE_TAG_STORAGE;
arch/arm64/kernel/image.h
50
#define __HEAD_FLAG_PAGE_SIZE ((PAGE_SHIFT - 10) / 2)
arch/arm64/kernel/pi/map_range.c
42
tbl += (start >> (lshift + PAGE_SHIFT)) % PTRS_PER_PTE;
arch/arm64/kernel/sys.c
28
return ksys_mmap_pgoff(addr, len, prot, flags, fd, off >> PAGE_SHIFT);
arch/arm64/kernel/sys32.c
57
off_4k >>= (PAGE_SHIFT - 12);
arch/arm64/kernel/traps.c
612
val |= (PAGE_SHIFT - 2) & CTR_EL0_IminLine_MASK;
arch/arm64/kernel/vdso.c
110
vdso_text_len = vdso_info[abi].vdso_pages << PAGE_SHIFT;
arch/arm64/kernel/vdso.c
82
PAGE_SHIFT;
arch/arm64/kvm/arm.c
2443
free_pages(per_cpu(kvm_arm_hyp_stack_base, cpu), NVHE_STACK_SHIFT - PAGE_SHIFT);
arch/arm64/kvm/arm.c
2650
stack_base = __get_free_pages(GFP_KERNEL, NVHE_STACK_SHIFT - PAGE_SHIFT);
arch/arm64/kvm/at.c
1777
gfn = ipa >> PAGE_SHIFT;
arch/arm64/kvm/hyp/include/nvhe/memory.h
82
#define hyp_phys_to_pfn(phys) ((phys) >> PAGE_SHIFT)
arch/arm64/kvm/hyp/include/nvhe/memory.h
83
#define hyp_pfn_to_phys(pfn) ((phys_addr_t)((pfn) << PAGE_SHIFT))
arch/arm64/kvm/hyp/nvhe/early_alloc.c
21
return (cur - base) >> PAGE_SHIFT;
arch/arm64/kvm/hyp/nvhe/early_alloc.c
26
unsigned long size = (nr_pages << PAGE_SHIFT);
arch/arm64/kvm/hyp/nvhe/mem_protect.c
263
nr_pages = kvm_pgtable_stage2_pgd_size(mmu->vtcr) >> PAGE_SHIFT;
arch/arm64/kvm/hyp/nvhe/mem_protect.c
380
u64 limit = BIT(kvm_phys_shift(&host_mmu.arch.mmu) - PAGE_SHIFT);
arch/arm64/kvm/hyp/nvhe/mem_protect.c
65
*__e = __p + ((__sz) >> PAGE_SHIFT); \
arch/arm64/kvm/hyp/nvhe/mem_protect.c
953
if (nr_pages != block_size >> PAGE_SHIFT)
arch/arm64/kvm/hyp/nvhe/mm.c
315
#if PAGE_SHIFT < 16
arch/arm64/kvm/hyp/nvhe/page_alloc.c
232
get_order(nr_pages << PAGE_SHIFT));
arch/arm64/kvm/hyp/nvhe/page_alloc.c
236
pool->range_end = phys + (nr_pages << PAGE_SHIFT);
arch/arm64/kvm/hyp/nvhe/pkvm.c
646
PAGE_ALIGN(size) >> PAGE_SHIFT))
arch/arm64/kvm/hyp/nvhe/pkvm.c
666
PAGE_ALIGN(size) >> PAGE_SHIFT));
arch/arm64/kvm/hyp/nvhe/setup.c
102
unsigned long pgt_size = hyp_s1_pgtable_pages() << PAGE_SHIFT;
arch/arm64/kvm/hyp/pgtable.c
51
u64 mask = BIT(PAGE_SHIFT - 3) - 1;
arch/arm64/kvm/hyp/pgtable.c
660
pages = size >> PAGE_SHIFT;
arch/arm64/kvm/hyp/pgtable.c
665
addr += inval_pages << PAGE_SHIFT;
arch/arm64/kvm/hyp/vhe/switch.c
92
va |= __vcpu_sys_reg(vcpu, VNCR_EL2) & GENMASK(PAGE_SHIFT - 1, 0);
arch/arm64/kvm/mmu.c
1031
phys_addr_t addr = memslot->base_gfn << PAGE_SHIFT;
arch/arm64/kvm/mmu.c
1248
start = memslot->base_gfn << PAGE_SHIFT;
arch/arm64/kvm/mmu.c
1249
end = (memslot->base_gfn + memslot->npages) << PAGE_SHIFT;
arch/arm64/kvm/mmu.c
1278
start = memslot->base_gfn << PAGE_SHIFT;
arch/arm64/kvm/mmu.c
1279
end = (memslot->base_gfn + memslot->npages) << PAGE_SHIFT;
arch/arm64/kvm/mmu.c
1302
phys_addr_t start = (base_gfn + __ffs(mask)) << PAGE_SHIFT;
arch/arm64/kvm/mmu.c
1303
phys_addr_t end = (base_gfn + __fls(mask) + 1) << PAGE_SHIFT;
arch/arm64/kvm/mmu.c
1346
gpa_start = memslot->base_gfn << PAGE_SHIFT;
arch/arm64/kvm/mmu.c
1441
return PAGE_SHIFT;
arch/arm64/kvm/mmu.c
1445
pa = (vma->vm_pgoff << PAGE_SHIFT) + (hva - vma->vm_start);
arch/arm64/kvm/mmu.c
1459
return PAGE_SHIFT;
arch/arm64/kvm/mmu.c
1473
unsigned long i, nr_pages = size >> PAGE_SHIFT;
arch/arm64/kvm/mmu.c
1588
gfn = kvm_s2_trans_output(nested) >> PAGE_SHIFT;
arch/arm64/kvm/mmu.c
1590
gfn = fault_ipa >> PAGE_SHIFT;
arch/arm64/kvm/mmu.c
1699
vma_shift = PAGE_SHIFT;
arch/arm64/kvm/mmu.c
1718
vma_shift = PAGE_SHIFT;
arch/arm64/kvm/mmu.c
1721
case PAGE_SHIFT:
arch/arm64/kvm/mmu.c
1765
gfn = ipa >> PAGE_SHIFT;
arch/arm64/kvm/mmu.c
187
u64 size = nr_pages << PAGE_SHIFT;
arch/arm64/kvm/mmu.c
188
u64 addr = gfn << PAGE_SHIFT;
arch/arm64/kvm/mmu.c
2148
gfn = ipa >> PAGE_SHIFT;
arch/arm64/kvm/mmu.c
219
kvm_account_pgtable_pages(virt, (size >> PAGE_SHIFT));
arch/arm64/kvm/mmu.c
2229
__unmap_stage2_range(&kvm->arch.mmu, range->start << PAGE_SHIFT,
arch/arm64/kvm/mmu.c
2230
(range->end - range->start) << PAGE_SHIFT,
arch/arm64/kvm/mmu.c
2239
u64 size = (range->end - range->start) << PAGE_SHIFT;
arch/arm64/kvm/mmu.c
2245
range->start << PAGE_SHIFT,
arch/arm64/kvm/mmu.c
225
kvm_account_pgtable_pages(virt, -(size >> PAGE_SHIFT));
arch/arm64/kvm/mmu.c
2255
u64 size = (range->end - range->start) << PAGE_SHIFT;
arch/arm64/kvm/mmu.c
2261
range->start << PAGE_SHIFT,
arch/arm64/kvm/mmu.c
2422
if ((new->base_gfn + new->npages) > (kvm_phys_size(&kvm->arch.mmu) >> PAGE_SHIFT))
arch/arm64/kvm/mmu.c
2433
reg_end = hva + (new->npages << PAGE_SHIFT);
arch/arm64/kvm/mmu.c
2493
gpa_t gpa = slot->base_gfn << PAGE_SHIFT;
arch/arm64/kvm/mmu.c
2494
phys_addr_t size = slot->npages << PAGE_SHIFT;
arch/arm64/kvm/mmu.c
354
phys_addr_t addr = memslot->base_gfn << PAGE_SHIFT;
arch/arm64/kvm/nested.c
1304
gfn = vt->wr.pa >> PAGE_SHIFT;
arch/arm64/kvm/nested.c
1329
vt->hpa = pfn << PAGE_SHIFT;
arch/arm64/kvm/pkvm.c
366
u64 gfn = addr >> PAGE_SHIFT;
arch/arm64/kvm/pkvm.c
367
u64 pfn = phys >> PAGE_SHIFT;
arch/arm64/kvm/pkvm.c
461
return kvm_call_hyp_nvhe(__pkvm_host_relax_perms_guest, addr >> PAGE_SHIFT, prot);
arch/arm64/kvm/pkvm.c
467
WARN_ON(kvm_call_hyp_nvhe(__pkvm_host_mkyoung_guest, addr >> PAGE_SHIFT));
arch/arm64/kvm/pkvm.c
71
hyp_mem_size = hyp_mem_pages << PAGE_SHIFT;
arch/arm64/kvm/pvtime.c
97
if (kvm_is_error_hva(gfn_to_hva(kvm, ipa >> PAGE_SHIFT)))
arch/arm64/kvm/vgic/vgic-its.c
850
gfn_t gfn = gpa >> PAGE_SHIFT;
arch/arm64/mm/contpte.c
471
end = addr + (nr << PAGE_SHIFT);
arch/arm64/mm/contpte.c
477
nr = (next - addr) >> PAGE_SHIFT;
arch/arm64/mm/contpte.c
480
if (((addr | next | (pfn << PAGE_SHIFT)) & ~CONT_PTE_MASK) == 0)
arch/arm64/mm/fault.c
774
lsb = PAGE_SHIFT;
arch/arm64/mm/fixmap.c
130
__set_pte(ptep, pfn_pte(phys >> PAGE_SHIFT, flags));
arch/arm64/mm/hugetlbpage.c
42
return PUD_SHIFT - PAGE_SHIFT;
arch/arm64/mm/hugetlbpage.c
44
return CONT_PMD_SHIFT - PAGE_SHIFT;
arch/arm64/mm/hugetlbpage.c
502
hugetlb_add_hstate(PUD_SHIFT - PAGE_SHIFT);
arch/arm64/mm/hugetlbpage.c
504
hugetlb_add_hstate(CONT_PMD_SHIFT - PAGE_SHIFT);
arch/arm64/mm/hugetlbpage.c
505
hugetlb_add_hstate(PMD_SHIFT - PAGE_SHIFT);
arch/arm64/mm/hugetlbpage.c
506
hugetlb_add_hstate(CONT_PTE_SHIFT - PAGE_SHIFT);
arch/arm64/mm/init.c
307
early_memtest(min << PAGE_SHIFT, max << PAGE_SHIFT);
arch/arm64/mm/kasan_init.c
250
return (addr & ~_PAGE_OFFSET(vabits)) >> (shift + PAGE_SHIFT);
arch/arm64/mm/kasan_init.c
274
return (addr >> (shift + PAGE_SHIFT)) % PTRS_PER_PTE;
arch/arm64/mm/mmap.c
64
return !(((pfn << PAGE_SHIFT) + size) & ~PHYS_MASK);
arch/arm64/mm/mmu.c
1435
vmem_altmap_free(altmap, size >> PAGE_SHIFT);
arch/arm64/mm/mmu.c
2010
ret = __add_pages(nid, start >> PAGE_SHIFT, size >> PAGE_SHIFT,
arch/arm64/mm/mmu.c
2029
unsigned long start_pfn = start >> PAGE_SHIFT;
arch/arm64/mm/mmu.c
2030
unsigned long nr_pages = size >> PAGE_SHIFT;
arch/arm64/mm/mmu.c
632
unsigned int step = PMD_SIZE >> PAGE_SHIFT;
arch/arm64/mm/pageattr.c
196
>> PAGE_SHIFT;
arch/arm64/mm/trans_pgd.c
219
int bits_mapped = PAGE_SHIFT - 4;
arch/csky/abiv1/mmap.c
14
(((pgoff)<<PAGE_SHIFT) & (SHMLBA-1)))
arch/csky/abiv1/mmap.c
36
.align_offset = pgoff << PAGE_SHIFT
arch/csky/abiv1/mmap.c
50
(addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1))
arch/csky/abiv2/inc/abi/entry.h
272
bmaski r7, (PAGE_SHIFT + 1) /* r7 = 0x1fff */
arch/csky/include/asm/fixmap.h
24
#define FIXADDR_SIZE (__end_of_fixed_addresses << PAGE_SHIFT)
arch/csky/include/asm/highmem.h
29
#define PKMAP_NR(virt) ((virt-PKMAP_BASE) >> PAGE_SHIFT)
arch/csky/include/asm/highmem.h
30
#define PKMAP_ADDR(nr) (PKMAP_BASE + ((nr) << PAGE_SHIFT))
arch/csky/include/asm/page.h
14
#define THREAD_SHIFT (PAGE_SHIFT + 1)
arch/csky/include/asm/page.h
75
return __pa(kaddr) >> PAGE_SHIFT;
arch/csky/include/asm/pgtable.h
31
#define PFN_PTE_SHIFT PAGE_SHIFT
arch/csky/include/asm/pgtable.h
32
#define pmd_pfn(pmd) (pmd_phys(pmd) >> PAGE_SHIFT)
arch/csky/include/asm/pgtable.h
33
#define pmd_page(pmd) (pfn_to_page(pmd_phys(pmd) >> PAGE_SHIFT))
arch/csky/include/asm/pgtable.h
38
#define pte_pfn(x) ((unsigned long)((x).pte_low >> PAGE_SHIFT))
arch/csky/include/asm/pgtable.h
39
#define pfn_pte(pfn, prot) __pte(((unsigned long long)(pfn) << PAGE_SHIFT) \
arch/csky/include/asm/pgtable.h
46
#define __mk_pte(page_nr, pgprot) __pte(((page_nr) << PAGE_SHIFT) | \
arch/csky/include/asm/thread_info.h
34
#define THREAD_SIZE_ORDER (THREAD_SHIFT - PAGE_SHIFT)
arch/csky/kernel/syscall.c
29
offset >> (PAGE_SHIFT - 12));
arch/csky/kernel/vdso.c
21
vdso_pages = (vdso_end - vdso_start) >> PAGE_SHIFT;
arch/csky/kernel/vdso.c
32
pg = virt_to_page(vdso_start + (i << PAGE_SHIFT));
arch/csky/kernel/vdso.c
51
vdso_len = vdso_pages << PAGE_SHIFT;
arch/csky/kernel/vdso.c
69
_install_special_mapping(mm, vdso_base, vdso_pages << PAGE_SHIFT,
arch/csky/kernel/vdso.c
79
vdso_base += (vdso_pages << PAGE_SHIFT);
arch/hexagon/include/asm/mem-layout.h
38
#define PHYS_PFN_OFFSET (PHYS_OFFSET >> PAGE_SHIFT)
arch/hexagon/include/asm/mem-layout.h
76
#define FIXADDR_SIZE (__end_of_fixed_addresses << PAGE_SHIFT)
arch/hexagon/include/asm/mem-layout.h
86
#define PKMAP_NR(virt) ((virt - PKMAP_BASE) >> PAGE_SHIFT)
arch/hexagon/include/asm/mem-layout.h
87
#define PKMAP_ADDR(nr) (PKMAP_BASE + ((nr) << PAGE_SHIFT))
arch/hexagon/include/asm/page.h
120
return __pa(kaddr) >> PAGE_SHIFT;
arch/hexagon/include/asm/page.h
44
#define HUGETLB_PAGE_ORDER (HPAGE_SHIFT-PAGE_SHIFT)
arch/hexagon/include/asm/page.h
95
#define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
arch/hexagon/include/asm/pgalloc.h
51
set_pmd(pmd, __pmd(((unsigned long)page_to_pfn(pte) << PAGE_SHIFT) |
arch/hexagon/include/asm/pgtable.h
214
#define pmd_pfn(pmd) (pmd_val(pmd) >> PAGE_SHIFT)
arch/hexagon/include/asm/pgtable.h
219
#define pmd_page(pmd) (pfn_to_page(pmd_val(pmd) >> PAGE_SHIFT))
arch/hexagon/include/asm/pgtable.h
335
#define PFN_PTE_SHIFT PAGE_SHIFT
arch/hexagon/include/asm/pgtable.h
337
#define pfn_pte(pfn, pgprot) __pte((pfn << PAGE_SHIFT) | pgprot_val(pgprot))
arch/hexagon/include/asm/pgtable.h
340
#define pte_pfn(pte) (pte_val(pte) >> PAGE_SHIFT)
arch/hexagon/include/asm/thread_info.h
21
#define THREAD_SIZE_ORDER (THREAD_SHIFT - PAGE_SHIFT)
arch/hexagon/kernel/asm-offsets.c
29
DEFINE(_PAGE_SHIFT, PAGE_SHIFT);
arch/hexagon/mm/init.c
121
bootmem_lastpg = PFN_DOWN((bootmem_lastpg << PAGE_SHIFT) &
arch/hexagon/mm/init.c
125
(bootmem_lastpg - ARCH_PFN_OFFSET) << PAGE_SHIFT);
arch/hexagon/mm/init.c
129
(bootmem_startpg - ARCH_PFN_OFFSET) << PAGE_SHIFT);
arch/hexagon/mm/init.c
160
segtable += (bootmem_lastpg-ARCH_PFN_OFFSET)>>(22-PAGE_SHIFT);
arch/loongarch/include/asm/dma.h
9
#define MAX_DMA32_PFN (1UL << (32 - PAGE_SHIFT))
arch/loongarch/include/asm/fixmap.h
19
#define FIXADDR_SIZE (__end_of_fixed_addresses << PAGE_SHIFT)
arch/loongarch/include/asm/page.h
13
#define HPAGE_SHIFT (PAGE_SHIFT + PAGE_SHIFT - PTRLOG)
arch/loongarch/include/asm/page.h
16
#define HUGETLB_PAGE_ORDER (HPAGE_SHIFT - PAGE_SHIFT)
arch/loongarch/include/asm/page.h
69
#define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT)
arch/loongarch/include/asm/pgtable-bits.h
85
#define PFN_PTE_SHIFT (PAGE_SHIFT - 12 + _PAGE_PFN_SHIFT)
arch/loongarch/include/asm/pgtable.h
176
#define p4d_page(p4d) (pfn_to_page(p4d_phys(p4d) >> PAGE_SHIFT))
arch/loongarch/include/asm/pgtable.h
222
#define pud_page(pud) (pfn_to_page(pud_phys(pud) >> PAGE_SHIFT))
arch/loongarch/include/asm/pgtable.h
260
#define pmd_page(pmd) (pfn_to_page(pmd_phys(pmd) >> PAGE_SHIFT))
arch/loongarch/include/asm/pgtable.h
27
#define PGDIR_SHIFT (PAGE_SHIFT + (PAGE_SHIFT - PTRLOG))
arch/loongarch/include/asm/pgtable.h
29
#define PMD_SHIFT (PAGE_SHIFT + (PAGE_SHIFT - PTRLOG))
arch/loongarch/include/asm/pgtable.h
32
#define PGDIR_SHIFT (PMD_SHIFT + (PAGE_SHIFT - PTRLOG))
arch/loongarch/include/asm/pgtable.h
34
#define PMD_SHIFT (PAGE_SHIFT + (PAGE_SHIFT - PTRLOG))
arch/loongarch/include/asm/pgtable.h
37
#define PUD_SHIFT (PMD_SHIFT + (PAGE_SHIFT - PTRLOG))
arch/loongarch/include/asm/pgtable.h
40
#define PGDIR_SHIFT (PUD_SHIFT + (PAGE_SHIFT - PTRLOG))
arch/loongarch/include/asm/pgtable.h
49
#define VA_BITS (PGDIR_SHIFT + (PAGE_SHIFT - PTRLOG))
arch/loongarch/include/asm/pgtable.h
573
return pfn_to_page(pmd_phys(pmd) >> PAGE_SHIFT);
arch/loongarch/include/asm/sparsemem.h
15
#define VMEMMAP_SIZE (sizeof(struct page) * (1UL << (cpu_pabits + 1 - PAGE_SHIFT)))
arch/loongarch/include/asm/vdso/vdso.h
17
#define VVAR_SIZE (VDSO_NR_PAGES << PAGE_SHIFT)
arch/loongarch/kernel/asm-offsets.c
214
DEFINE(_PAGE_SHIFT, PAGE_SHIFT);
arch/loongarch/kernel/numa.c
183
num_physpages += (mem_size >> PAGE_SHIFT);
arch/loongarch/kernel/numa.c
187
mem_start >> PAGE_SHIFT, mem_end >> PAGE_SHIFT, num_physpages);
arch/loongarch/kernel/numa.c
192
num_physpages += (mem_size >> PAGE_SHIFT);
arch/loongarch/kernel/numa.c
196
mem_start >> PAGE_SHIFT, mem_end >> PAGE_SHIFT, num_physpages);
arch/loongarch/kernel/syscall.c
34
return ksys_mmap_pgoff(addr, len, prot, flags, fd, offset >> PAGE_SHIFT);
arch/loongarch/kernel/syscall.c
43
return ksys_mmap_pgoff(addr, len, prot, flags, fd, offset >> (PAGE_SHIFT - 12));
arch/loongarch/kvm/mmu.c
204
size = 0x1UL << (ctx->pgtable_shift + PAGE_SHIFT - 3);
arch/loongarch/kvm/mmu.c
248
size = 0x1UL << (ctx->pgtable_shift + PAGE_SHIFT - 3);
arch/loongarch/kvm/mmu.c
305
ret = kvm_ptw_top(kvm->arch.pgd, start_gfn << PAGE_SHIFT,
arch/loongarch/kvm/mmu.c
306
end_gfn << PAGE_SHIFT, &ctx);
arch/loongarch/kvm/mmu.c
309
ret = kvm_ptw_top(kvm->arch.pgd, start_gfn << PAGE_SHIFT,
arch/loongarch/kvm/mmu.c
310
end_gfn << PAGE_SHIFT, &ctx);
arch/loongarch/kvm/mmu.c
348
return kvm_ptw_top(kvm->arch.pgd, start_gfn << PAGE_SHIFT, end_gfn << PAGE_SHIFT, &ctx);
arch/loongarch/kvm/mmu.c
376
kvm_ptw_top(kvm->arch.pgd, start << PAGE_SHIFT, end << PAGE_SHIFT, &ctx);
arch/loongarch/kvm/mmu.c
392
if ((new->base_gfn + new->npages) > (kvm->arch.gpa_size >> PAGE_SHIFT))
arch/loongarch/kvm/mmu.c
397
gpa_start = new->base_gfn << PAGE_SHIFT;
arch/loongarch/kvm/mmu.c
489
kvm_flush_range(kvm, 0, kvm->arch.gpa_size >> PAGE_SHIFT, 0);
arch/loongarch/kvm/mmu.c
510
return kvm_ptw_top(kvm->arch.pgd, range->start << PAGE_SHIFT,
arch/loongarch/kvm/mmu.c
511
range->end << PAGE_SHIFT, &ctx);
arch/loongarch/kvm/mmu.c
522
return kvm_ptw_top(kvm->arch.pgd, range->start << PAGE_SHIFT,
arch/loongarch/kvm/mmu.c
523
range->end << PAGE_SHIFT, &ctx);
arch/loongarch/kvm/mmu.c
528
gpa_t gpa = range->start << PAGE_SHIFT;
arch/loongarch/kvm/mmu.c
556
gfn_t gfn = gpa >> PAGE_SHIFT;
arch/loongarch/kvm/mmu.c
68
offset = (addr >> PAGE_SHIFT) - ctx->gfn;
arch/loongarch/kvm/mmu.c
779
gfn_t gfn = gpa >> PAGE_SHIFT;
arch/loongarch/kvm/vcpu.c
1204
if (kvm_is_error_hva(gfn_to_hva(kvm, gpa >> PAGE_SHIFT)))
arch/loongarch/kvm/vm.c
98
kvm->arch.pte_shifts[i] = PAGE_SHIFT + i * (PAGE_SHIFT - 3);
arch/loongarch/mm/init.c
192
set_pte(ptep, pfn_pte(phys >> PAGE_SHIFT, flags));
arch/loongarch/mm/init.c
76
unsigned long start_pfn = start >> PAGE_SHIFT;
arch/loongarch/mm/init.c
77
unsigned long nr_pages = size >> PAGE_SHIFT;
arch/loongarch/mm/init.c
91
unsigned long start_pfn = start >> PAGE_SHIFT;
arch/loongarch/mm/init.c
92
unsigned long nr_pages = size >> PAGE_SHIFT;
arch/loongarch/mm/mmap.c
153
return !(((pfn << PAGE_SHIFT) + size) & ~(GENMASK_ULL(cpu_pabits, 0)));
arch/loongarch/mm/mmap.c
17
+ (((pgoff) << PAGE_SHIFT) & SHM_ALIGN_MASK))
arch/loongarch/mm/mmap.c
44
((addr - (pgoff << PAGE_SHIFT)) & SHM_ALIGN_MASK))
arch/loongarch/mm/mmap.c
67
info.align_offset = pgoff << PAGE_SHIFT;
arch/loongarch/mm/tlb.c
214
pgd_w = PAGE_SHIFT - 3;
arch/loongarch/mm/tlb.c
217
pud_w = PAGE_SHIFT - 3;
arch/loongarch/mm/tlb.c
221
pmd_w = PAGE_SHIFT - 3;
arch/loongarch/mm/tlb.c
223
pte_i = PAGE_SHIFT;
arch/loongarch/mm/tlb.c
224
pte_w = PAGE_SHIFT - 3;
arch/loongarch/mm/tlb.c
69
size = (end - start) >> (PAGE_SHIFT + 1);
arch/loongarch/mm/tlb.c
93
size = (end - start + (PAGE_SIZE - 1)) >> PAGE_SHIFT;
arch/m68k/include/asm/mcf_pgtable.h
288
#define PFN_PTE_SHIFT PAGE_SHIFT
arch/m68k/include/asm/mcf_pgtable.h
289
#define pmd_pfn(pmd) (pmd_val(pmd) >> PAGE_SHIFT)
arch/m68k/include/asm/mcf_pgtable.h
290
#define pmd_page(pmd) (pfn_to_page(pmd_val(pmd) >> PAGE_SHIFT))
arch/m68k/include/asm/mcf_pgtable.h
292
#define pfn_pte(pfn, prot) __pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot))
arch/m68k/include/asm/mcf_pgtable.h
293
#define pte_pfn(pte) (pte_val(pte) >> PAGE_SHIFT)
arch/m68k/include/asm/mcf_pgtable.h
97
#define pmd_pgtable(pmd) pfn_to_virt(pmd_val(pmd) >> PAGE_SHIFT)
arch/m68k/include/asm/motorola_pgtable.h
109
#define PFN_PTE_SHIFT PAGE_SHIFT
arch/m68k/include/asm/motorola_pgtable.h
111
#define pte_pfn(pte) (pte_val(pte) >> PAGE_SHIFT)
arch/m68k/include/asm/motorola_pgtable.h
112
#define pfn_pte(pfn, prot) __pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot))
arch/m68k/include/asm/motorola_pgtable.h
119
#define pmd_pfn(pmd) ((pmd_val(pmd) & _TABLE_MASK) >> PAGE_SHIFT)
arch/m68k/include/asm/motorola_pgtable.h
132
#define pud_page(pud) (mem_map + ((unsigned long)(__va(pud_val(pud)) - PAGE_OFFSET) >> PAGE_SHIFT))
arch/m68k/include/asm/page_mm.h
125
return __pa(kaddr) >> PAGE_SHIFT;
arch/m68k/include/asm/page_mm.h
130
return __va(pfn << PAGE_SHIFT);
arch/m68k/include/asm/page_mm.h
142
#define ARCH_PFN_OFFSET (m68k_memory[0].addr >> PAGE_SHIFT)
arch/m68k/include/asm/page_no.h
23
return __pa(kaddr) >> PAGE_SHIFT;
arch/m68k/include/asm/page_no.h
28
return __va(pfn << PAGE_SHIFT);
arch/m68k/include/asm/page_no.h
31
#define virt_to_page(addr) (mem_map + (((unsigned long)(addr)-PAGE_OFFSET) >> PAGE_SHIFT))
arch/m68k/include/asm/page_no.h
32
#define page_to_virt(page) __va(((((page) - mem_map) << PAGE_SHIFT) + PAGE_OFFSET))
arch/m68k/include/asm/sun3_pgtable.h
108
#define pmd_pfn(pmd) (pmd_val(pmd) >> PAGE_SHIFT)
arch/m68k/include/asm/sun3_pgtable.h
88
(__va ((pte_val (pte) & SUN3_PAGE_PGNUM_MASK) << PAGE_SHIFT))
arch/m68k/include/asm/thread_info.h
13
#if PAGE_SHIFT < 13
arch/m68k/kernel/sys_m68k.c
181
for (len >>= PAGE_SHIFT; len--; addr += PAGE_SIZE)
arch/m68k/kernel/sys_m68k.c
344
for (len >>= PAGE_SHIFT; len--; addr += PAGE_SIZE)
arch/m68k/mm/motorola.c
479
min_low_pfn = availmem >> PAGE_SHIFT;
arch/m68k/mm/motorola.c
480
max_pfn = max_low_pfn = (max_addr >> PAGE_SHIFT) + 1;
arch/m68k/mm/sun3kmap.c
33
ptep = pfn_pte(phys >> PAGE_SHIFT, PAGE_KERNEL);
arch/m68k/sun3/config.c
116
max_pfn = num_pages = __pa(memory_end) >> PAGE_SHIFT;
arch/m68k/sun3/dvma.c
35
if(ptelist[(vaddr & 0xff000) >> PAGE_SHIFT] != pte) {
arch/m68k/sun3/dvma.c
37
ptelist[(vaddr & 0xff000) >> PAGE_SHIFT] = pte;
arch/m68k/sun3/mmu_emu.c
114
val, (val & SUN3_PAGE_PGNUM_MASK) << PAGE_SHIFT, flags, type);
arch/m68k/sun3/mmu_emu.c
78
val, (val & SUN3_PAGE_PGNUM_MASK) << PAGE_SHIFT);
arch/microblaze/include/asm/cacheflush.h
79
unsigned long addr = folio_pfn(folio) << PAGE_SHIFT;
arch/microblaze/include/asm/cacheflush.h
86
flush_dcache_range(pfn << PAGE_SHIFT, (pfn << PAGE_SHIFT) + PAGE_SIZE);
arch/microblaze/include/asm/fixmap.h
58
#define __FIXADDR_SIZE (__end_of_fixed_addresses << PAGE_SHIFT)
arch/microblaze/include/asm/highmem.h
49
#define PKMAP_NR(virt) ((virt - PKMAP_BASE) >> PAGE_SHIFT)
arch/microblaze/include/asm/highmem.h
50
#define PKMAP_ADDR(nr) (PKMAP_BASE + ((nr) << PAGE_SHIFT))
arch/microblaze/include/asm/page.h
101
# define ARCH_PFN_OFFSET (memory_start >> PAGE_SHIFT)
arch/microblaze/include/asm/page.h
26
#define PTE_SHIFT (PAGE_SHIFT - 2) /* 1024 ptes per page */
arch/microblaze/include/asm/page.h
98
# define virt_to_page(kaddr) (pfn_to_page(__pa(kaddr) >> PAGE_SHIFT))
arch/microblaze/include/asm/page.h
99
# define page_to_virt(page) __va(page_to_pfn(page) << PAGE_SHIFT)
arch/microblaze/include/asm/pgtable.h
221
((pte_val(x) - memory_start) >> PAGE_SHIFT))
arch/microblaze/include/asm/pgtable.h
222
#define PFN_PTE_SHIFT PAGE_SHIFT
arch/microblaze/include/asm/pgtable.h
364
#define pmd_pfn(pmd) (__pa(pmd_val(pmd)) >> PAGE_SHIFT)
arch/microblaze/include/asm/pgtable.h
367
#define pmd_page(pmd) (pfn_to_page(__pa(pmd_val(pmd)) >> PAGE_SHIFT))
arch/microblaze/include/asm/pgtable.h
89
#define PGDIR_SHIFT (PAGE_SHIFT + PTE_SHIFT)
arch/microblaze/kernel/sys_microblaze.c
43
return ksys_mmap_pgoff(addr, len, prot, flags, fd, pgoff >> PAGE_SHIFT);
arch/microblaze/kernel/sys_microblaze.c
54
pgoff >> (PAGE_SHIFT - 12));
arch/microblaze/mm/init.c
95
min_low_pfn = memory_start >> PAGE_SHIFT; /* minimum for allocation */
arch/microblaze/mm/init.c
96
max_low_pfn = ((u64)memory_start + (u64)lowmem_size) >> PAGE_SHIFT;
arch/microblaze/mm/init.c
97
max_pfn = ((u64)memory_start + (u64)memory_size) >> PAGE_SHIFT;
arch/microblaze/mm/pgtable.c
155
set_pte_at(&init_mm, va, pg, pfn_pte(pa >> PAGE_SHIFT,
arch/mips/alchemy/common/setup.c
101
return phys_addr >> PAGE_SHIFT;
arch/mips/alchemy/common/setup.c
99
phys_addr_t phys_addr = fixup_bigphys_addr(pfn << PAGE_SHIFT, size);
arch/mips/include/asm/dma.h
96
#define MAX_DMA32_PFN (1UL << (32 - PAGE_SHIFT))
arch/mips/include/asm/fixmap.h
67
#define FIXADDR_SIZE (__end_of_fixed_addresses << PAGE_SHIFT)
arch/mips/include/asm/highmem.h
46
#define PKMAP_NR(virt) ((virt-PKMAP_BASE) >> PAGE_SHIFT)
arch/mips/include/asm/highmem.h
47
#define PKMAP_ADDR(nr) (PKMAP_BASE + ((nr) << PAGE_SHIFT))
arch/mips/include/asm/kvm_host.h
273
#define TLB_LO_IDX(x, va) (((va) >> PAGE_SHIFT) & 1)
arch/mips/include/asm/mach-cavium-octeon/cpu-feature-overrides.h
75
#define MAX_DMA32_PFN (((1ULL << 32) - (1ULL << 28)) >> PAGE_SHIFT)
arch/mips/include/asm/page.h
207
#define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT)
arch/mips/include/asm/page.h
35
return (PAGE_SHIFT - 10) / 2;
arch/mips/include/asm/page.h
43
#define HPAGE_SHIFT (PAGE_SHIFT + PAGE_SHIFT - 3)
arch/mips/include/asm/page.h
46
#define HUGETLB_PAGE_ORDER (HPAGE_SHIFT - PAGE_SHIFT)
arch/mips/include/asm/pgtable-32.h
65
# define PGDIR_SHIFT (2 * PAGE_SHIFT - PTE_T_LOG2 - 1)
arch/mips/include/asm/pgtable-32.h
67
# define PGDIR_SHIFT (2 * PAGE_SHIFT - PTE_T_LOG2)
arch/mips/include/asm/pgtable-32.h
78
# define __PGD_TABLE_ORDER (32 - 3 * PAGE_SHIFT + PGD_T_LOG2 + PTE_T_LOG2 + 1)
arch/mips/include/asm/pgtable-32.h
80
# define __PGD_TABLE_ORDER (32 - 3 * PAGE_SHIFT + PGD_T_LOG2 + PTE_T_LOG2)
arch/mips/include/asm/pgtable-32.h
98
#define PKMAP_END ((FIXADDR_START) & ~((LAST_PKMAP << PAGE_SHIFT)-1))
arch/mips/include/asm/pgtable-64.h
213
#define p4d_page(p4d) (pfn_to_page(p4d_phys(p4d) >> PAGE_SHIFT))
arch/mips/include/asm/pgtable-64.h
311
#define pud_page(pud) (pfn_to_page(pud_phys(pud) >> PAGE_SHIFT))
arch/mips/include/asm/pgtable-64.h
45
#define PGDIR_SHIFT (PAGE_SHIFT + PAGE_SHIFT - 3)
arch/mips/include/asm/pgtable-64.h
49
#define PMD_SHIFT (PAGE_SHIFT + (PAGE_SHIFT - 3))
arch/mips/include/asm/pgtable-64.h
54
# define PGDIR_SHIFT (PMD_SHIFT + (PAGE_SHIFT + PMD_TABLE_ORDER - 3))
arch/mips/include/asm/pgtable-64.h
59
#define PUD_SHIFT (PMD_SHIFT + (PAGE_SHIFT + PMD_TABLE_ORDER - 3))
arch/mips/include/asm/pgtable-64.h
62
#define PGDIR_SHIFT (PUD_SHIFT + (PAGE_SHIFT + PUD_TABLE_ORDER - 3))
arch/mips/include/asm/pgtable-bits.h
185
# define PFN_PTE_SHIFT PAGE_SHIFT
arch/mips/include/asm/pgtable-bits.h
188
# define PFN_PTE_SHIFT (PAGE_SHIFT - 12 + _CACHE_SHIFT + 3)
arch/mips/include/asm/pgtable.h
423
return pfn_to_page(pmd_phys(pmd) >> PAGE_SHIFT);
arch/mips/include/asm/pgtable.h
73
#define pmd_page(pmd) (pfn_to_page(pmd_phys(pmd) >> PAGE_SHIFT))
arch/mips/include/asm/tlb.h
9
(((base) + ((idx) << (PAGE_SHIFT + 1))) | \
arch/mips/kernel/asm-offsets.c
222
DEFINE(_PAGE_SHIFT, PAGE_SHIFT);
arch/mips/kernel/syscall.c
69
offset >> PAGE_SHIFT);
arch/mips/kernel/syscall.c
80
pgoff >> (PAGE_SHIFT - 12));
arch/mips/kvm/mmu.c
273
start_gfn << PAGE_SHIFT,
arch/mips/kvm/mmu.c
274
end_gfn << PAGE_SHIFT);
arch/mips/kvm/mmu.c
399
start_gfn << PAGE_SHIFT,
arch/mips/kvm/mmu.c
400
end_gfn << PAGE_SHIFT);
arch/mips/kvm/mmu.c
437
start_gfn << PAGE_SHIFT,
arch/mips/kvm/mmu.c
438
end_gfn << PAGE_SHIFT);
arch/mips/kvm/mmu.c
454
gpa_t gpa = range->start << PAGE_SHIFT;
arch/mips/kvm/mmu.c
485
gfn_t gfn = gpa >> PAGE_SHIFT;
arch/mips/kvm/mmu.c
553
gfn_t gfn = gpa >> PAGE_SHIFT;
arch/mips/loongson64/numa.c
111
memblock_reserve(kernel_start_pfn << PAGE_SHIFT,
arch/mips/loongson64/numa.c
112
((kernel_end_pfn - kernel_start_pfn) << PAGE_SHIFT));
arch/mips/loongson64/numa.c
115
if (node_end_pfn(0) >= (0xffffffff >> PAGE_SHIFT))
arch/mips/mm/c-r3k.c
239
unsigned long kaddr = KSEG0ADDR(pfn << PAGE_SHIFT);
arch/mips/mm/cache.c
159
address -= offset_in_folio(folio, pfn << PAGE_SHIFT);
arch/mips/mm/dma-noncoherent.c
98
struct page *page = pfn_to_page(paddr >> PAGE_SHIFT);
arch/mips/mm/init.c
409
(highend_pfn - max_low_pfn) << (PAGE_SHIFT - 10));
arch/mips/mm/init.c
447
BUILD_BUG_ON(IS_ENABLED(CONFIG_32BIT) && (PFN_PTE_SHIFT > PAGE_SHIFT));
arch/mips/mm/init.c
86
idx = (addr >> PAGE_SHIFT) & (FIX_N_COLOURS - 1);
arch/mips/mm/mmap.c
25
(((pgoff) << PAGE_SHIFT) & shm_align_mask))
arch/mips/mm/mmap.c
52
((addr - (pgoff << PAGE_SHIFT)) & shm_align_mask))
arch/mips/mm/mmap.c
76
info.align_offset = pgoff << PAGE_SHIFT;
arch/mips/mm/sc-ip22.c
167
size <<= PAGE_SHIFT;
arch/mips/mm/tlb-r3k.c
119
size = (end - start + (PAGE_SIZE - 1)) >> PAGE_SHIFT;
arch/mips/mm/tlb-r3k.c
82
size = (end - start + (PAGE_SIZE - 1)) >> PAGE_SHIFT;
arch/mips/mm/tlb-r4k.c
122
size = (end - start) >> (PAGE_SHIFT + 1);
arch/mips/mm/tlb-r4k.c
175
size = (end - start + (PAGE_SIZE - 1)) >> PAGE_SHIFT;
arch/mips/mm/tlbex.c
1076
PGDIR_SHIFT + PGD_TABLE_ORDER + PAGE_SHIFT - 3);
arch/mips/mm/tlbex.c
1220
} else if (PAGE_SHIFT == 14 || PAGE_SHIFT == 13) {
arch/mips/mm/tlbex.c
1445
pmd_w = PMD_SHIFT - PAGE_SHIFT;
arch/mips/mm/tlbex.c
1447
pgd_w = PGDIR_SHIFT - PAGE_SHIFT + PGD_TABLE_ORDER;
arch/mips/mm/tlbex.c
1450
pt_i = PAGE_SHIFT; /* 3rd level PTE */
arch/mips/mm/tlbex.c
1451
pt_w = PAGE_SHIFT - 3;
arch/mips/mm/tlbex.c
1486
PGDIR_SHIFT + PGD_TABLE_ORDER + PAGE_SHIFT - 3);
arch/mips/mm/tlbex.c
2015
UASM_i_SRL(p, wr.r1, wr.r1, PAGE_SHIFT - PTE_T_LOG2);
arch/mips/mm/tlbex.c
2418
pwfield |= PAGE_SHIFT << MIPS_PWFIELD_PTI_SHIFT;
arch/mips/mm/tlbex.c
2545
check_for_high_segbits = current_cpu_data.vmbits > (PGDIR_SHIFT + PGD_TABLE_ORDER + PAGE_SHIFT - 3);
arch/mips/mm/tlbex.c
783
uasm_i_dsrl_safe(p, ptr, tmp, PGDIR_SHIFT + PGD_TABLE_ORDER + PAGE_SHIFT - 3);
arch/mips/mm/tlbex.c
957
unsigned int shift = 4 - (PTE_T_LOG2 + 1) + PAGE_SHIFT - 12;
arch/mips/sgi-ip27/ip27-klnuma.c
126
return TO_NODE(nasid, offset) >> PAGE_SHIFT;
arch/mips/sgi-ip27/ip27-klnuma.c
128
return KDM_TO_PHYS(PAGE_ALIGN(SYMMON_STK_ADDR(nasid, 0))) >> PAGE_SHIFT;
arch/mips/sgi-ip27/ip27-memory.c
259
return size >> PAGE_SHIFT;
arch/mips/sgi-ip27/ip27-memory.c
265
return size >> PAGE_SHIFT;
arch/mips/sgi-ip27/ip27-memory.c
336
if ((nodebytes >> PAGE_SHIFT) * (sizeof(struct page)) >
arch/mips/sgi-ip27/ip27-memory.c
337
(slot0sz << PAGE_SHIFT)) {
arch/mips/sgi-ip27/ip27-memory.c
34
#define SLOT_PFNSHIFT (SLOT_SHIFT - PAGE_SHIFT)
arch/mips/sgi-ip27/ip27-memory.c
35
#define PFN_NASIDSHFT (NASID_SHFT - PAGE_SHIFT)
arch/mips/sgi-ip27/ip27-memory.c
361
__node_data[node] = __va(slot_freepfn << PAGE_SHIFT);
arch/mips/sgi-ip27/ip27-memory.c
373
memblock_reserve(slot_firstpfn << PAGE_SHIFT,
arch/mips/sgi-ip27/ip27-memory.c
374
((slot_freepfn - slot_firstpfn) << PAGE_SHIFT));
arch/nios2/include/asm/page.h
81
((void *)(((page) - mem_map) << PAGE_SHIFT) + PAGE_OFFSET)
arch/nios2/include/asm/page.h
83
# define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT)
arch/nios2/include/asm/pgtable.h
208
pte_val(null) = (addr >> PAGE_SHIFT) & 0xf;
arch/nios2/include/asm/pgtable.h
218
#define pmd_pfn(pmd) (pmd_phys(pmd) >> PAGE_SHIFT)
arch/nios2/include/asm/pgtable.h
219
#define pmd_page(pmd) (pfn_to_page(pmd_phys(pmd) >> PAGE_SHIFT))
arch/nios2/mm/cacheflush.c
93
start = vma->vm_start + ((pgoff - vma->vm_pgoff) << PAGE_SHIFT);
arch/nios2/mm/tlb.c
134
WRCTL(CTL_PTEADDR, (addr >> PAGE_SHIFT) << 2);
arch/nios2/mm/tlb.c
144
if (((pteaddr >> 2) & 0xfffff) != (addr >> PAGE_SHIFT))
arch/nios2/mm/tlb.c
174
line << (PAGE_SHIFT + cpuinfo.tlb_num_ways_log2));
arch/nios2/mm/tlb.c
191
if ((tlbacc << PAGE_SHIFT) != 0) {
arch/nios2/mm/tlb.c
194
(pteaddr << (PAGE_SHIFT-2)),
arch/nios2/mm/tlb.c
195
(tlbacc << PAGE_SHIFT),
arch/nios2/mm/tlb.c
23
<< PAGE_SHIFT)
arch/nios2/mm/tlb.c
38
return ((addr | 0xC0000000UL) >> PAGE_SHIFT) << 2;
arch/nios2/mm/tlb.c
53
WRCTL(CTL_PTEADDR, (addr >> PAGE_SHIFT) << 2);
arch/nios2/mm/tlb.c
64
if (((pteaddr >> 2) & 0xfffff) != (addr >> PAGE_SHIFT))
arch/openrisc/include/asm/fixmap.h
35
#define FIXADDR_SIZE (__end_of_fixed_addresses << PAGE_SHIFT)
arch/openrisc/include/asm/page.h
67
return __pa(kaddr) >> PAGE_SHIFT;
arch/openrisc/include/asm/page.h
71
(mem_map + (((unsigned long)(addr)-PAGE_OFFSET) >> PAGE_SHIFT))
arch/openrisc/include/asm/pgalloc.h
36
(unsigned long) PAGE_SHIFT)));
arch/openrisc/include/asm/pgtable.h
309
#define pte_pagenr(pte) ((__pte_page(pte) - PAGE_OFFSET) >> PAGE_SHIFT)
arch/openrisc/include/asm/pgtable.h
313
#define __page_address(page) (PAGE_OFFSET + (((page) - mem_map) << PAGE_SHIFT))
arch/openrisc/include/asm/pgtable.h
326
#define pmd_pfn(pmd) (pmd_val(pmd) >> PAGE_SHIFT)
arch/openrisc/include/asm/pgtable.h
327
#define pmd_page(pmd) (pfn_to_page(pmd_val(pmd) >> PAGE_SHIFT))
arch/openrisc/include/asm/pgtable.h
337
#define PFN_PTE_SHIFT PAGE_SHIFT
arch/openrisc/include/asm/pgtable.h
338
#define pte_pfn(x) ((unsigned long)(((x).pte)) >> PAGE_SHIFT)
arch/openrisc/include/asm/pgtable.h
339
#define pfn_pte(pfn, prot) __pte((((pfn) << PAGE_SHIFT)) | pgprot_val(prot))
arch/openrisc/include/asm/pgtable.h
56
#define PGDIR_SHIFT (PAGE_SHIFT + (PAGE_SHIFT-2))
arch/openrisc/include/asm/pgtable.h
66
#define PTRS_PER_PTE (1UL << (PAGE_SHIFT-2))
arch/openrisc/mm/cache.c
46
unsigned long paddr = page_to_pfn(page) << PAGE_SHIFT;
arch/openrisc/mm/cache.c
84
unsigned long pfn = pte_val(*pte) >> PAGE_SHIFT;
arch/openrisc/mm/fault.c
27
#define TLB_OFFSET(add) (((add) >> PAGE_SHIFT) & (NUM_TLB_ENTRIES-1))
arch/openrisc/mm/init.c
212
set_pte_at(&init_mm, va, pte, pfn_pte(pa >> PAGE_SHIFT, prot));
arch/openrisc/mm/tlb.c
35
#define DTLB_OFFSET(addr) (((addr) >> PAGE_SHIFT) & (NUM_DTLB_SETS-1))
arch/openrisc/mm/tlb.c
36
#define ITLB_OFFSET(addr) (((addr) >> PAGE_SHIFT) & (NUM_ITLB_SETS-1))
arch/parisc/include/asm/fixmap.h
34
#define FIXMAP_SIZE (FIX_BITMAP_COUNT << PAGE_SHIFT)
arch/parisc/include/asm/fixmap.h
57
#define __fix_to_virt(_x) (FIXMAP_START + ((_x) << PAGE_SHIFT))
arch/parisc/include/asm/page.h
152
#define HUGETLB_PAGE_ORDER (HPAGE_SHIFT - PAGE_SHIFT)
arch/parisc/include/asm/page.h
166
#define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
arch/parisc/include/asm/page.h
168
#define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
arch/parisc/include/asm/pgtable.h
105
#define PLD_SHIFT PAGE_SHIFT
arch/parisc/include/asm/pgtable.h
107
#define BITS_PER_PTE (PAGE_SHIFT - BITS_PER_PTE_ENTRY)
arch/parisc/include/asm/pgtable.h
115
#define BITS_PER_PMD (PAGE_SHIFT + PMD_TABLE_ORDER - BITS_PER_PMD_ENTRY)
arch/parisc/include/asm/pgtable.h
123
#if (PGDIR_SHIFT + PAGE_SHIFT + PGD_TABLE_ORDER - BITS_PER_PGD_ENTRY) > BITS_PER_LONG
arch/parisc/include/asm/pgtable.h
126
#define BITS_PER_PGD (PAGE_SHIFT + PGD_TABLE_ORDER - BITS_PER_PGD_ENTRY)
arch/parisc/include/asm/pgtable.h
334
pte_val(__pte) = ((((addr)>>PAGE_SHIFT)<<PFN_PTE_SHIFT) + pgprot_val(pgprot)); \
arch/parisc/include/asm/pgtable.h
360
#define pmd_pfn(pmd) (pmd_address(pmd) >> PAGE_SHIFT)
arch/parisc/include/asm/ropes.h
192
#define IOVP_SHIFT PAGE_SHIFT
arch/parisc/include/asm/thread_info.h
36
#define THREAD_SHIFT (PAGE_SHIFT + THREAD_SIZE_ORDER)
arch/parisc/kernel/inventory.c
127
#define PDC_PAGE_ADJ_SHIFT (PAGE_SHIFT - 12) /* pdc pages are always 4k */
arch/parisc/kernel/inventory.c
146
pmem_ptr->start_pfn = (start >> PAGE_SHIFT);
arch/parisc/kernel/inventory.c
167
npages = (PAGE_ALIGN(PAGE0->imm_max_mem) >> PAGE_SHIFT);
arch/parisc/kernel/pci-dma.c
253
unsigned int pages_needed = size >> PAGE_SHIFT;
arch/parisc/kernel/pci-dma.c
293
return (pcxl_dma_start + (res_idx << (PAGE_SHIFT + 3)));
arch/parisc/kernel/pci-dma.c
308
unsigned int res_idx = (vaddr - pcxl_dma_start) >> (PAGE_SHIFT + 3);
arch/parisc/kernel/pci-dma.c
309
unsigned int pages_mapped = size >> PAGE_SHIFT;
arch/parisc/kernel/pci-dma.c
379
pcxl_res_size = PCXL_DMA_MAP_SIZE >> (PAGE_SHIFT + 3);
arch/parisc/kernel/pci-dma.c
412
size = 1 << (order + PAGE_SHIFT);
arch/parisc/kernel/pci-dma.c
431
size = 1 << (order + PAGE_SHIFT);
arch/parisc/kernel/pdt.c
238
num_poisoned_pages_inc(addr >> PAGE_SHIFT);
arch/parisc/kernel/pdt.c
334
memory_failure(pde >> PAGE_SHIFT, 0);
arch/parisc/kernel/pdt.c
336
soft_offline_page(pde >> PAGE_SHIFT, 0);
arch/parisc/kernel/sys_parisc.c
192
pgoff >> (PAGE_SHIFT - 12));
arch/parisc/kernel/sys_parisc.c
201
offset >> PAGE_SHIFT);
arch/parisc/kernel/sys_parisc.c
35
& ((SHM_COLOUR-1) >> PAGE_SHIFT) : 0UL)
arch/parisc/kernel/sys_parisc.c
40
return (filp_pgoff + pgoff) << PAGE_SHIFT;
arch/parisc/kernel/sys_parisc.c
92
stack_base += (STACK_RND_MASK << PAGE_SHIFT);
arch/parisc/kernel/vdso.c
101
int pages = (end - start) >> PAGE_SHIFT;
arch/parisc/mm/fault.c
418
lsb = PAGE_SHIFT;
arch/parisc/mm/init.c
110
#define MAX_GAP (0x40000000UL >> PAGE_SHIFT)
arch/parisc/mm/init.c
171
size = (pmem_ranges[i].pages << PAGE_SHIFT);
arch/parisc/mm/init.c
172
start = (pmem_ranges[i].start_pfn << PAGE_SHIFT);
arch/parisc/mm/init.c
201
rsize = pmem_ranges[i].pages << PAGE_SHIFT;
arch/parisc/mm/init.c
207
pmem_ranges[i].pages = (mem_limit >> PAGE_SHIFT)
arch/parisc/mm/init.c
208
- (mem_max >> PAGE_SHIFT);
arch/parisc/mm/init.c
257
start = start_pfn << PAGE_SHIFT;
arch/parisc/mm/init.c
258
size = npages << PAGE_SHIFT;
arch/parisc/mm/init.c
295
memblock_reserve((pmem_holes[i].start_pfn << PAGE_SHIFT),
arch/parisc/mm/init.c
296
(pmem_holes[i].pages << PAGE_SHIFT));
arch/parisc/mm/init.c
369
start_pte = ((start_vaddr >> PAGE_SHIFT) & (PTRS_PER_PTE - 1));
arch/parisc/mm/init.c
551
BUILD_BUG_ON(PAGE_SHIFT + BITS_PER_PTE + BITS_PER_PMD + BITS_PER_PGD
arch/parisc/mm/init.c
625
start_paddr = pmem_ranges[range].start_pfn << PAGE_SHIFT;
arch/parisc/mm/init.c
626
size = pmem_ranges[range].pages << PAGE_SHIFT;
arch/parisc/mm/init.c
731
if ((size >> PAGE_SHIFT) >= min_num_pages)
arch/parisc/mm/init.c
732
pdc_btlb_insert(start >> PAGE_SHIFT, __pa(start) >> PAGE_SHIFT,
arch/parisc/mm/init.c
733
size >> PAGE_SHIFT, entry_info, *slot);
arch/powerpc/boot/page.h
17
#define PAGE_SIZE (ASM_CONST(1) << PAGE_SHIFT)
arch/powerpc/include/asm/book3s/32/pgtable.h
129
#define PGDIR_SHIFT (PAGE_SHIFT + PTE_INDEX_SIZE)
arch/powerpc/include/asm/book3s/32/pgtable.h
345
#define pmd_pfn(pmd) (pmd_val(pmd) >> PAGE_SHIFT)
arch/powerpc/include/asm/book3s/32/pgtable.h
57
#define PTE_RPN_SHIFT (PAGE_SHIFT)
arch/powerpc/include/asm/book3s/64/hash-64k.h
160
unsigned long __end = vpn + (1UL << (PAGE_SHIFT - VPN_SHIFT)); \
arch/powerpc/include/asm/book3s/64/hash-64k.h
178
if (pfn > (PTE_RPN_MASK >> PAGE_SHIFT)) {
arch/powerpc/include/asm/book3s/64/hash.h
35
H_PUD_INDEX_SIZE + H_PGD_INDEX_SIZE + PAGE_SHIFT)
arch/powerpc/include/asm/book3s/64/mmu-hash.h
690
#define SBP_L1_BITS (PAGE_SHIFT - 2)
arch/powerpc/include/asm/book3s/64/mmu-hash.h
691
#define SBP_L2_BITS (PAGE_SHIFT - 3)
arch/powerpc/include/asm/book3s/64/mmu-hash.h
694
#define SBP_L2_SHIFT (PAGE_SHIFT + SBP_L1_BITS)
arch/powerpc/include/asm/book3s/64/pgtable.h
104
#define PTE_RPN_SHIFT PAGE_SHIFT
arch/powerpc/include/asm/book3s/64/pgtable.h
208
#define PMD_SHIFT (PAGE_SHIFT + PTE_INDEX_SIZE)
arch/powerpc/include/asm/book3s/64/pgtable.h
253
return 7 + PAGE_SHIFT; /* default from linux/vmalloc.h */
arch/powerpc/include/asm/book3s/64/pgtable.h
566
VM_BUG_ON(pfn >> (64 - PAGE_SHIFT));
arch/powerpc/include/asm/book3s/64/pgtable.h
567
VM_BUG_ON((pfn << PAGE_SHIFT) & ~PTE_RPN_MASK);
arch/powerpc/include/asm/book3s/64/pgtable.h
569
return __pte(((pte_basic_t)pfn << PAGE_SHIFT) | pgprot_val(pgprot) | _PAGE_PTE);
arch/powerpc/include/asm/book3s/64/pgtable.h
648
#define __swp_offset(x) (((x).val & PTE_RPN_MASK) >> PAGE_SHIFT)
arch/powerpc/include/asm/book3s/64/pgtable.h
650
(type) | (((offset) << PAGE_SHIFT) & PTE_RPN_MASK)})
arch/powerpc/include/asm/book3s/64/radix.h
35
#define RADIX_PMD_SHIFT (PAGE_SHIFT + RADIX_PTE_INDEX_SIZE)
arch/powerpc/include/asm/book3s/64/radix.h
47
RADIX_PUD_INDEX_SIZE + RADIX_PGD_INDEX_SIZE + PAGE_SHIFT)
arch/powerpc/include/asm/elf.h
123
(0x7ff >> (PAGE_SHIFT - 12)) : \
arch/powerpc/include/asm/elf.h
124
(0x3ffff >> (PAGE_SHIFT - 12)))
arch/powerpc/include/asm/fixmap.h
81
#define __FIXADDR_SIZE (__end_of_fixed_addresses << PAGE_SHIFT)
arch/powerpc/include/asm/highmem.h
56
#define PKMAP_NR(virt) ((virt-PKMAP_BASE) >> PAGE_SHIFT)
arch/powerpc/include/asm/highmem.h
57
#define PKMAP_ADDR(nr) (PKMAP_BASE + ((nr) << PAGE_SHIFT))
arch/powerpc/include/asm/kvm_book3s_64.h
379
return ((ptel & HPTE_R_RPN) & ~(psize - 1)) >> PAGE_SHIFT;
arch/powerpc/include/asm/kvm_book3s_64.h
491
unsigned long mask = (pagesize >> PAGE_SHIFT) - 1;
arch/powerpc/include/asm/nohash/32/mmu-8xx.h
256
return PAGE_SHIFT;
arch/powerpc/include/asm/nohash/32/pgtable.h
135
#define PTE_RPN_SHIFT (PAGE_SHIFT)
arch/powerpc/include/asm/nohash/32/pgtable.h
168
#define pmd_pfn(pmd) (pmd_val(pmd) >> PAGE_SHIFT)
arch/powerpc/include/asm/nohash/32/pgtable.h
172
#define pmd_pfn(pmd) (__pa(pmd_val(pmd)) >> PAGE_SHIFT)
arch/powerpc/include/asm/nohash/32/pgtable.h
45
#define PGDIR_SHIFT (PAGE_SHIFT + PTE_INDEX_SIZE)
arch/powerpc/include/asm/nohash/64/pgtable-4k.h
30
#define PMD_SHIFT (PAGE_SHIFT + PTE_INDEX_SIZE)
arch/powerpc/include/asm/nohash/64/pgtable.h
19
PUD_INDEX_SIZE + PGD_INDEX_SIZE + PAGE_SHIFT)
arch/powerpc/include/asm/page.h
112
#define ARCH_PFN_OFFSET ((unsigned long)(MEMORY_START >> PAGE_SHIFT))
arch/powerpc/include/asm/page.h
222
return __pa(kaddr) >> PAGE_SHIFT;
arch/powerpc/include/asm/page.h
227
return __va(pfn << PAGE_SHIFT);
arch/powerpc/include/asm/page.h
28
#define HPAGE_SHIFT PAGE_SHIFT
arch/powerpc/include/asm/page.h
39
#define HUGETLB_PAGE_ORDER (HPAGE_SHIFT - PAGE_SHIFT)
arch/powerpc/include/asm/page_32.h
17
#define PTE_SHIFT (PAGE_SHIFT - PTE_T_LOG2 - 2) /* 1/4 of a page */
arch/powerpc/include/asm/page_32.h
19
#define PTE_SHIFT (PAGE_SHIFT - PTE_T_LOG2) /* full page */
arch/powerpc/include/asm/page_64.h
24
#define PAGE_FACTOR (PAGE_SHIFT - HW_PAGE_SHIFT)
arch/powerpc/include/asm/pgtable.h
169
#define PTE_FRAG_SIZE_SHIFT PAGE_SHIFT
arch/powerpc/include/asm/pte-walk.h
56
hugepage_shift = PAGE_SHIFT;
arch/powerpc/include/asm/rtas.h
556
unsigned long paddr = (pfn << PAGE_SHIFT);
arch/powerpc/include/asm/thread_info.h
23
#if defined(CONFIG_VMAP_STACK) && MIN_THREAD_SHIFT < PAGE_SHIFT
arch/powerpc/include/asm/thread_info.h
24
#define THREAD_SHIFT PAGE_SHIFT
arch/powerpc/include/asm/thread_info.h
84
#define THREAD_SIZE_ORDER (THREAD_SHIFT - PAGE_SHIFT)
arch/powerpc/kernel/crash_dump.c
131
free_reserved_page(pfn_to_page(addr >> PAGE_SHIFT));
arch/powerpc/kernel/crash_dump.c
82
paddr = pfn << PAGE_SHIFT;
arch/powerpc/kernel/iommu.c
1109
SetPageDirty(pfn_to_page(*hpa >> PAGE_SHIFT));
arch/powerpc/kernel/iommu.c
513
if (tbl->it_page_shift < PAGE_SHIFT && slen >= PAGE_SIZE &&
arch/powerpc/kernel/iommu.c
515
align = PAGE_SHIFT - tbl->it_page_shift;
arch/powerpc/kernel/iommu.c
872
if (tbl->it_page_shift < PAGE_SHIFT && size >= PAGE_SIZE &&
arch/powerpc/kernel/iommu.c
874
align = PAGE_SHIFT - tbl->it_page_shift;
arch/powerpc/kernel/mce.c
312
PAGE_SHIFT;
arch/powerpc/kernel/mce_power.c
465
instr_addr = (pfn << PAGE_SHIFT) + (regs->nip & ~PAGE_MASK);
arch/powerpc/kernel/mce_power.c
470
*phys_addr = (pfn << PAGE_SHIFT);
arch/powerpc/kernel/mce_power.c
54
if (shift <= PAGE_SHIFT)
arch/powerpc/kernel/mce_power.c
556
(pfn << PAGE_SHIFT);
arch/powerpc/kernel/paca.c
86
shared_lppaca_total_size >> PAGE_SHIFT);
arch/powerpc/kernel/pci-common.c
515
vma->vm_pgoff += (ioaddr + hose->io_base_phys) >> PAGE_SHIFT;
arch/powerpc/kernel/pci-common.c
530
resource_size_t offset = ((resource_size_t)pfn) << PAGE_SHIFT;
arch/powerpc/kernel/pci-common.c
659
((resource_size_t)vma->vm_pgoff) << PAGE_SHIFT;
arch/powerpc/kernel/pci-common.c
698
vma->vm_pgoff = offset >> PAGE_SHIFT;
arch/powerpc/kernel/proc_powerpc.c
40
__pa(pde_data(file_inode(file))) >> PAGE_SHIFT,
arch/powerpc/kernel/prom_init.c
2074
tce_entry = (i << PAGE_SHIFT);
arch/powerpc/kernel/setup-common.c
1001
early_memtest(min_low_pfn << PAGE_SHIFT, max_low_pfn << PAGE_SHIFT);
arch/powerpc/kernel/suspend.c
20
unsigned long nosave_begin_pfn = __pa(&__nosave_begin) >> PAGE_SHIFT;
arch/powerpc/kernel/suspend.c
21
unsigned long nosave_end_pfn = PAGE_ALIGN(__pa(&__nosave_end)) >> PAGE_SHIFT;
arch/powerpc/kernel/syscalls.c
56
return do_mmap2(addr, len, prot, flags, fd, pgoff, PAGE_SHIFT-12);
arch/powerpc/kernel/syscalls.c
65
return do_mmap2(addr, len, prot, flags, fd, off_4k, PAGE_SHIFT-12);
arch/powerpc/kernel/syscalls.c
73
return do_mmap2(addr, len, prot, flags, fd, offset, PAGE_SHIFT);
arch/powerpc/kernel/vdso.c
246
int pages = (end - start) >> PAGE_SHIFT;
arch/powerpc/kvm/book3s.c
428
gfn_t gfn = gpa >> PAGE_SHIFT;
arch/powerpc/kvm/book3s.c
439
pfn = (kvm_pfn_t)virt_to_phys((void*)shared_page) >> PAGE_SHIFT;
arch/powerpc/kvm/book3s_32_mmu.c
115
r = gfn_to_hva(vcpu->kvm, pteg >> PAGE_SHIFT);
arch/powerpc/kvm/book3s_32_mmu_host.c
154
hpaddr <<= PAGE_SHIFT;
arch/powerpc/kvm/book3s_32_mmu_host.c
199
mark_page_dirty(vcpu->kvm, orig_pte->raddr >> PAGE_SHIFT);
arch/powerpc/kvm/book3s_32_mmu_host.c
205
kvmppc_mmu_flush_icache(hpaddr >> PAGE_SHIFT);
arch/powerpc/kvm/book3s_32_mmu_host.c
248
pte->pfn = hpaddr >> PAGE_SHIFT;
arch/powerpc/kvm/book3s_32_mmu_host.c
34
#if PAGE_SHIFT != 12
arch/powerpc/kvm/book3s_64_mmu.c
148
r = gfn_to_hva(vcpu->kvm, pteg >> PAGE_SHIFT);
arch/powerpc/kvm/book3s_64_mmu_host.c
105
hpaddr = pfn << PAGE_SHIFT;
arch/powerpc/kvm/book3s_64_mmu_host.c
89
unsigned long gfn = orig_pte->raddr >> PAGE_SHIFT;
arch/powerpc/kvm/book3s_64_mmu_hv.c
104
kvm_free_hpt_cma(page, 1 << (order - PAGE_SHIFT));
arch/powerpc/kvm/book3s_64_mmu_hv.c
106
free_pages(hpt, order - PAGE_SHIFT);
arch/powerpc/kvm/book3s_64_mmu_hv.c
1072
n = (n + PAGE_SIZE - 1) >> PAGE_SHIFT;
arch/powerpc/kvm/book3s_64_mmu_hv.c
1094
gfn = vpa->gpa >> PAGE_SHIFT;
arch/powerpc/kvm/book3s_64_mmu_hv.c
1131
unsigned long gfn = gpa >> PAGE_SHIFT;
arch/powerpc/kvm/book3s_64_mmu_hv.c
1172
gfn = gpa >> PAGE_SHIFT;
arch/powerpc/kvm/book3s_64_mmu_hv.c
187
1 << (info->order - PAGE_SHIFT));
arch/powerpc/kvm/book3s_64_mmu_hv.c
189
free_pages(info->virt, info->order - PAGE_SHIFT);
arch/powerpc/kvm/book3s_64_mmu_hv.c
220
npages = memslot->npages >> (porder - PAGE_SHIFT);
arch/powerpc/kvm/book3s_64_mmu_hv.c
547
gfn_base = gpa_base >> PAGE_SHIFT;
arch/powerpc/kvm/book3s_64_mmu_hv.c
576
gfn_base = gpa_base >> PAGE_SHIFT;
arch/powerpc/kvm/book3s_64_mmu_hv.c
578
gfn = gpa >> PAGE_SHIFT;
arch/powerpc/kvm/book3s_64_mmu_hv.c
630
hpa = pte_pfn(pte) << PAGE_SHIFT;
arch/powerpc/kvm/book3s_64_mmu_hv.c
83
page = kvm_alloc_hpt_cma(1ul << (order - PAGE_SHIFT));
arch/powerpc/kvm/book3s_64_mmu_hv.c
92
|__GFP_NOWARN, order - PAGE_SHIFT);
arch/powerpc/kvm/book3s_64_mmu_radix.c
1015
unsigned long gpa = gfn << PAGE_SHIFT;
arch/powerpc/kvm/book3s_64_mmu_radix.c
1019
uv_page_inval(kvm->arch.lpid, gpa, PAGE_SHIFT);
arch/powerpc/kvm/book3s_64_mmu_radix.c
1034
unsigned long gpa = gfn << PAGE_SHIFT;
arch/powerpc/kvm/book3s_64_mmu_radix.c
1063
unsigned long gpa = gfn << PAGE_SHIFT;
arch/powerpc/kvm/book3s_64_mmu_radix.c
1081
unsigned long gpa = gfn << PAGE_SHIFT;
arch/powerpc/kvm/book3s_64_mmu_radix.c
1171
gpa = memslot->base_gfn << PAGE_SHIFT;
arch/powerpc/kvm/book3s_64_mmu_radix.c
1401
shift = PAGE_SHIFT;
arch/powerpc/kvm/book3s_64_mmu_radix.c
326
pshift = PAGE_SHIFT;
arch/powerpc/kvm/book3s_64_mmu_radix.c
430
unsigned long gfn = gpa >> PAGE_SHIFT;
arch/powerpc/kvm/book3s_64_mmu_radix.c
489
pte_pfn(*p) << PAGE_SHIFT,
arch/powerpc/kvm/book3s_64_mmu_radix.c
490
PAGE_SHIFT, NULL, lpid);
arch/powerpc/kvm/book3s_64_mmu_radix.c
512
pte_pfn(*(pte_t *)p) << PAGE_SHIFT,
arch/powerpc/kvm/book3s_64_mmu_radix.c
830
unsigned long hva, gfn = gpa >> PAGE_SHIFT;
arch/powerpc/kvm/book3s_64_mmu_radix.c
882
if (shift > PAGE_SHIFT) {
arch/powerpc/kvm/book3s_64_mmu_radix.c
952
gfn = gpa >> PAGE_SHIFT;
arch/powerpc/kvm/book3s_64_vio.c
355
unsigned long gfn = tce >> PAGE_SHIFT;
arch/powerpc/kvm/book3s_hv.c
1006
mark_page_dirty(kvm, to >> PAGE_SHIFT);
arch/powerpc/kvm/book3s_hv.c
5341
if ((size >> PAGE_SHIFT) > totalram_pages())
arch/powerpc/kvm/book3s_hv.c
978
from_memslot = gfn_to_memslot(kvm, from >> PAGE_SHIFT);
arch/powerpc/kvm/book3s_hv.c
982
<< PAGE_SHIFT))
arch/powerpc/kvm/book3s_hv.c
984
from_addr = gfn_to_hva_memslot(from_memslot, from >> PAGE_SHIFT);
arch/powerpc/kvm/book3s_hv.c
990
to_memslot = gfn_to_memslot(kvm, to >> PAGE_SHIFT);
arch/powerpc/kvm/book3s_hv.c
994
<< PAGE_SHIFT))
arch/powerpc/kvm/book3s_hv.c
996
to_addr = gfn_to_hva_memslot(to_memslot, to >> PAGE_SHIFT);
arch/powerpc/kvm/book3s_hv_builtin.c
41
#define HPT_ALIGN_PAGES ((1 << 18) >> PAGE_SHIFT) /* 256k */
arch/powerpc/kvm/book3s_hv_builtin.c
60
VM_BUG_ON(order_base_2(nr_pages) < KVM_CMA_CHUNK_ORDER - PAGE_SHIFT);
arch/powerpc/kvm/book3s_hv_builtin.c
96
align_size = HPT_ALIGN_PAGES << PAGE_SHIFT;
arch/powerpc/kvm/book3s_hv_builtin.c
98
KVM_CMA_CHUNK_ORDER - PAGE_SHIFT, false, "kvm_cma",
arch/powerpc/kvm/book3s_hv_nested.c
1046
gfn = (gpa >> PAGE_SHIFT) - memslot->base_gfn;
arch/powerpc/kvm/book3s_hv_nested.c
1047
end_gfn = gfn + (nbytes >> PAGE_SHIFT);
arch/powerpc/kvm/book3s_hv_nested.c
1085
shift = PAGE_SHIFT;
arch/powerpc/kvm/book3s_hv_nested.c
1148
npages = 1UL << (shift - PAGE_SHIFT);
arch/powerpc/kvm/book3s_hv_nested.c
1159
npages -= 1UL << (shadow_shift - PAGE_SHIFT);
arch/powerpc/kvm/book3s_hv_nested.c
1506
return PAGE_SHIFT;
arch/powerpc/kvm/book3s_hv_nested.c
1516
if (shift == PAGE_SHIFT)
arch/powerpc/kvm/book3s_hv_nested.c
1585
if (l1_shift < PAGE_SHIFT) {
arch/powerpc/kvm/book3s_hv_nested.c
1588
l1_shift, PAGE_SHIFT);
arch/powerpc/kvm/book3s_hv_nested.c
1592
gfn = gpa >> PAGE_SHIFT;
arch/powerpc/kvm/book3s_hv_nested.c
1630
shift = PAGE_SHIFT;
arch/powerpc/kvm/book3s_hv_nested.c
1646
gfn = (gpa & ~((1UL << shift) - 1)) >> PAGE_SHIFT;
arch/powerpc/kvm/book3s_hv_nested.c
1662
unsigned int actual_shift = PAGE_SHIFT;
arch/powerpc/kvm/book3s_hv_nested.c
578
!kvm_is_visible_gfn(vcpu->kvm, (ptcr & PRTB_MASK) >> PAGE_SHIFT))
arch/powerpc/kvm/book3s_hv_rm_mmu.c
227
gfn = gpa >> PAGE_SHIFT;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
272
pa = pte_pfn(pte) << PAGE_SHIFT;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
889
unsigned long gfn, hva, pa, psize = PAGE_SHIFT;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
894
gfn = gpa >> PAGE_SHIFT;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
913
pa = pte_pfn(pte) << PAGE_SHIFT;
arch/powerpc/kvm/book3s_hv_rm_mmu.c
947
kvmppc_update_dirty_map(memslot, dest >> PAGE_SHIFT, PAGE_SIZE);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
978
kvmppc_update_dirty_map(dest_memslot, dest >> PAGE_SHIFT, PAGE_SIZE);
arch/powerpc/kvm/book3s_hv_uvmem.c
1003
vmf->address + PAGE_SIZE, PAGE_SHIFT,
arch/powerpc/kvm/book3s_hv_uvmem.c
1021
(kvmppc_uvmem_pgmap.range.start >> PAGE_SHIFT);
arch/powerpc/kvm/book3s_hv_uvmem.c
1031
kvmppc_gfn_remove(pvt->gpa >> PAGE_SHIFT, pvt->kvm);
arch/powerpc/kvm/book3s_hv_uvmem.c
1033
kvmppc_gfn_secure_mem_pfn(pvt->gpa >> PAGE_SHIFT, pvt->kvm);
arch/powerpc/kvm/book3s_hv_uvmem.c
1058
if (page_shift != PAGE_SHIFT)
arch/powerpc/kvm/book3s_hv_uvmem.c
1097
ret = uv_page_in(kvm->arch.lpid, page_to_pfn(page) << PAGE_SHIFT,
arch/powerpc/kvm/book3s_hv_uvmem.c
1098
gfn << PAGE_SHIFT, 0, PAGE_SHIFT);
arch/powerpc/kvm/book3s_hv_uvmem.c
1196
pfn_first = res->start >> PAGE_SHIFT;
arch/powerpc/kvm/book3s_hv_uvmem.c
1197
pfn_last = pfn_first + (resource_size(res) >> PAGE_SHIFT);
arch/powerpc/kvm/book3s_hv_uvmem.c
404
end = start + (memslot->npages << PAGE_SHIFT);
arch/powerpc/kvm/book3s_hv_uvmem.c
450
memslot->base_gfn << PAGE_SHIFT,
arch/powerpc/kvm/book3s_hv_uvmem.c
645
PAGE_SHIFT, kvm, pvt->gpa, NULL))
arch/powerpc/kvm/book3s_hv_uvmem.c
702
pfn_first = kvmppc_uvmem_pgmap.range.start >> PAGE_SHIFT;
arch/powerpc/kvm/book3s_hv_uvmem.c
704
(range_len(&kvmppc_uvmem_pgmap.range) >> PAGE_SHIFT);
arch/powerpc/kvm/book3s_hv_uvmem.c
719
kvmppc_gfn_secure_uvmem_pfn(gpa >> PAGE_SHIFT, uvmem_pfn, kvm);
arch/powerpc/kvm/book3s_hv_uvmem.c
810
end = start + (1UL << PAGE_SHIFT);
arch/powerpc/kvm/book3s_hv_uvmem.c
816
(gfn << PAGE_SHIFT), kvm, PAGE_SHIFT, false);
arch/powerpc/kvm/book3s_hv_uvmem.c
948
if (page_shift != PAGE_SHIFT)
arch/powerpc/kvm/book3s_pr.c
1880
ga = memslot->base_gfn << PAGE_SHIFT;
arch/powerpc/kvm/book3s_pr.c
1881
ga_end = ga + (memslot->npages << PAGE_SHIFT);
arch/powerpc/kvm/book3s_pr.c
441
kvmppc_mmu_pte_pflush(vcpu, range->start << PAGE_SHIFT,
arch/powerpc/kvm/book3s_pr.c
442
range->end << PAGE_SHIFT);
arch/powerpc/kvm/book3s_pr.c
647
r = kvm_vcpu_map(vcpu, pte->raddr >> PAGE_SHIFT, &map);
arch/powerpc/kvm/book3s_pr.c
677
return kvm_is_visible_gfn(vcpu->kvm, gpa >> PAGE_SHIFT);
arch/powerpc/kvm/book3s_xive.c
2756
if (xive->q_order < PAGE_SHIFT)
arch/powerpc/kvm/book3s_xive.c
2759
xive->q_page_order = xive->q_order - PAGE_SHIFT;
arch/powerpc/kvm/book3s_xive_native.c
218
esb_pgoff << PAGE_SHIFT,
arch/powerpc/kvm/book3s_xive_native.c
219
2ull << PAGE_SHIFT, 1);
arch/powerpc/kvm/book3s_xive_native.c
280
vmf_insert_pfn(vma, vmf->address, page >> PAGE_SHIFT);
arch/powerpc/kvm/book3s_xive_native.c
297
vmf_insert_pfn(vma, vmf->address, xive_tima_os >> PAGE_SHIFT);
arch/powerpc/kvm/booke.c
1316
gfn = gpaddr >> PAGE_SHIFT;
arch/powerpc/kvm/booke.c
1364
gfn = gpaddr >> PAGE_SHIFT;
arch/powerpc/kvm/booke.c
1992
pte->vpage = eaddr >> PAGE_SHIFT;
arch/powerpc/kvm/booke.c
2021
pte->vpage = eaddr >> PAGE_SHIFT;
arch/powerpc/kvm/e500.h
266
if (!gfn_to_memslot(vcpu->kvm, gpa >> PAGE_SHIFT))
arch/powerpc/kvm/e500_mmu.c
52
set_base = (addr >> PAGE_SHIFT) & (sets - 1);
arch/powerpc/kvm/e500_mmu_host.c
169
pfn = (kvm_pfn_t)virt_to_phys((void *)shared_page) >> PAGE_SHIFT;
arch/powerpc/kvm/e500_mmu_host.c
178
magic.mas7_3 = ((u64)pfn << PAGE_SHIFT) |
arch/powerpc/kvm/e500_mmu_host.c
316
stlbe->mas7_3 = ((u64)pfn << PAGE_SHIFT) |
arch/powerpc/kvm/e500_mmu_host.c
400
psize_pages = 1UL << (psize - PAGE_SHIFT);
arch/powerpc/kvm/e500_mmu_host.c
419
tsize = min(psize - PAGE_SHIFT + BOOK3E_PAGESZ_4K, tsize);
arch/powerpc/kvm/e500_mmu_host.c
449
gvaddr &= ~((tsize_pages << PAGE_SHIFT) - 1);
arch/powerpc/kvm/e500_mmu_host.c
483
get_tlb_raddr(gtlbe) >> PAGE_SHIFT,
arch/powerpc/kvm/e500_mmu_host.c
571
gfn_t gfn = gpaddr >> PAGE_SHIFT;
arch/powerpc/kvm/e500_mmu_host.c
656
pfn = addr >> PAGE_SHIFT;
arch/powerpc/lib/code-patching.c
167
addr = (1 + (get_random_long() % (DEFAULT_MAP_WINDOW / PAGE_SIZE - 2))) << PAGE_SHIFT;
arch/powerpc/lib/code-patching.c
233
return __pa_symbol(addr) >> PAGE_SHIFT;
arch/powerpc/lib/code-patching.c
243
return map_kernel_page(text_poke_addr, (pfn << PAGE_SHIFT), PAGE_KERNEL);
arch/powerpc/mm/book3s32/tlb.c
67
count = ((pmd_end - start) >> PAGE_SHIFT) + 1;
arch/powerpc/mm/book3s64/hash_4k.c
81
pa = pte_pfn(__pte(old_pte)) << PAGE_SHIFT;
arch/powerpc/mm/book3s64/hash_64k.c
154
pa = pte_pfn(__pte(old_pte)) << PAGE_SHIFT;
arch/powerpc/mm/book3s64/hash_64k.c
291
pa = pte_pfn(__pte(old_pte)) << PAGE_SHIFT;
arch/powerpc/mm/book3s64/hash_hugepage.c
130
pa = pmd_pfn(__pmd(old_pmd)) << PAGE_SHIFT;
arch/powerpc/mm/book3s64/hash_pgtable.c
169
set_pte_at(&init_mm, ea, ptep, pfn_pte(pa >> PAGE_SHIFT, prot));
arch/powerpc/mm/book3s64/hash_utils.c
1663
spp = sbpp[(ea >> PAGE_SHIFT) & (SBP_L1_COUNT - 1)];
arch/powerpc/mm/book3s64/hash_utils.c
287
hash = hpt_hash(vpn, PAGE_SHIFT, mmu_kernel_ssize);
arch/powerpc/mm/book3s64/hash_utils.c
314
hash = hpt_hash(vpn, PAGE_SHIFT, mmu_kernel_ssize);
arch/powerpc/mm/book3s64/hash_utils.c
336
unsigned long linear_map_count = memblock_end_of_DRAM() >> PAGE_SHIFT;
arch/powerpc/mm/book3s64/hash_utils.c
352
linear_map_hash_count = memblock_end_of_DRAM() >> PAGE_SHIFT;
arch/powerpc/mm/book3s64/hash_utils.c
366
if ((paddr >> PAGE_SHIFT) < linear_map_hash_count)
arch/powerpc/mm/book3s64/hash_utils.c
367
linear_map_hash_slots[paddr >> PAGE_SHIFT] = slot | 0x80;
arch/powerpc/mm/book3s64/hash_utils.c
382
lmi = __pa(vaddr) >> PAGE_SHIFT;
arch/powerpc/mm/book3s64/hash_utils.c
419
linear_map_kf_hash_count = KFENCE_POOL_SIZE >> PAGE_SHIFT;
arch/powerpc/mm/book3s64/hash_utils.c
472
>> PAGE_SHIFT;
arch/powerpc/mm/book3s64/hash_utils.c
490
lmi = (vaddr - (unsigned long)__kfence_pool) >> PAGE_SHIFT;
arch/powerpc/mm/book3s64/hugetlbpage.c
108
pa = pte_pfn(__pte(old_pte)) << PAGE_SHIFT;
arch/powerpc/mm/book3s64/iommu_api.c
100
chunk = (1UL << (PAGE_SHIFT + MAX_PAGE_ORDER)) /
arch/powerpc/mm/book3s64/iommu_api.c
106
ret = pin_user_pages(ua + (entry << PAGE_SHIFT), n,
arch/powerpc/mm/book3s64/iommu_api.c
135
if ((mem2->ua < (ua + (entries << PAGE_SHIFT))) &&
arch/powerpc/mm/book3s64/iommu_api.c
137
(mem2->entries << PAGE_SHIFT)))) {
arch/powerpc/mm/book3s64/iommu_api.c
150
pageshift = PAGE_SHIFT;
arch/powerpc/mm/book3s64/iommu_api.c
154
if ((mem->pageshift > PAGE_SHIFT) && PageHuge(page))
arch/powerpc/mm/book3s64/iommu_api.c
161
mem->hpas[i] = page_to_pfn(page) << PAGE_SHIFT;
arch/powerpc/mm/book3s64/iommu_api.c
214
page = pfn_to_page(mem->hpas[i] >> PAGE_SHIFT);
arch/powerpc/mm/book3s64/iommu_api.c
297
(mem->entries << PAGE_SHIFT))) {
arch/powerpc/mm/book3s64/iommu_api.c
333
const long entry = (ua - mem->ua) >> PAGE_SHIFT;
arch/powerpc/mm/book3s64/iommu_api.c
365
end = mem->dev_hpa + (mem->entries << PAGE_SHIFT);
arch/powerpc/mm/book3s64/iommu_api.c
80
mem->pageshift = __ffs(dev_hpa | (entries << PAGE_SHIFT));
arch/powerpc/mm/book3s64/iommu_api.c
91
mem->pageshift = __ffs(ua | (entries << PAGE_SHIFT));
arch/powerpc/mm/book3s64/pgtable.c
241
pmdv = (pfn << PAGE_SHIFT) & PTE_RPN_MASK;
arch/powerpc/mm/book3s64/pgtable.c
250
pudv = (pfn << PAGE_SHIFT) & PTE_RPN_MASK;
arch/powerpc/mm/book3s64/radix_pgtable.c
1008
entry = pfn_pte(__pa(p) >> PAGE_SHIFT, PAGE_KERNEL);
arch/powerpc/mm/book3s64/radix_pgtable.c
1056
entry = pfn_pte(__pa(p) >> PAGE_SHIFT, PAGE_KERNEL);
arch/powerpc/mm/book3s64/radix_pgtable.c
130
unsigned long pfn = pa >> PAGE_SHIFT;
arch/powerpc/mm/book3s64/radix_pgtable.c
1338
addr_pfn += (PMD_SIZE >> PAGE_SHIFT);
arch/powerpc/mm/book3s64/radix_pgtable.c
78
unsigned long pfn = pa >> PAGE_SHIFT;
arch/powerpc/mm/book3s64/radix_pgtable.c
966
int nid = early_pfn_to_nid(phys >> PAGE_SHIFT);
arch/powerpc/mm/book3s64/slice.c
285
int pshift = max_t(int, mmu_psize_defs[psize].shift, PAGE_SHIFT);
arch/powerpc/mm/book3s64/slice.c
327
int pshift = max_t(int, mmu_psize_defs[psize].shift, PAGE_SHIFT);
arch/powerpc/mm/book3s64/slice.c
434
int pshift = max_t(int, mmu_psize_defs[psize].shift, PAGE_SHIFT);
arch/powerpc/mm/book3s64/subpage_prot.c
120
spp += (addr >> PAGE_SHIFT) & (SBP_L1_COUNT - 1);
arch/powerpc/mm/book3s64/subpage_prot.c
122
i = (addr >> PAGE_SHIFT) & (PTRS_PER_PTE - 1);
arch/powerpc/mm/book3s64/subpage_prot.c
124
if (addr + (nw << PAGE_SHIFT) > next)
arch/powerpc/mm/book3s64/subpage_prot.c
125
nw = (next - addr) >> PAGE_SHIFT;
arch/powerpc/mm/book3s64/subpage_prot.c
213
if (!access_ok(map, (len >> PAGE_SHIFT) * sizeof(u32)))
arch/powerpc/mm/book3s64/subpage_prot.c
255
spp += (addr >> PAGE_SHIFT) & (SBP_L1_COUNT - 1);
arch/powerpc/mm/book3s64/subpage_prot.c
261
i = (addr >> PAGE_SHIFT) & (PTRS_PER_PTE - 1);
arch/powerpc/mm/book3s64/subpage_prot.c
263
if (addr + (nw << PAGE_SHIFT) > next)
arch/powerpc/mm/book3s64/subpage_prot.c
264
nw = (next - addr) >> PAGE_SHIFT;
arch/powerpc/mm/dma-noncoherent.c
96
struct page *page = pfn_to_page(paddr >> PAGE_SHIFT);
arch/powerpc/mm/fault.c
157
lsb = PAGE_SHIFT;
arch/powerpc/mm/hugetlbpage.c
163
hugetlb_add_hstate(shift - PAGE_SHIFT);
arch/powerpc/mm/hugetlbpage.c
206
return PUD_SHIFT - PAGE_SHIFT;
arch/powerpc/mm/hugetlbpage.c
211
return mmu_psize_to_shift(MMU_PAGE_16G) - PAGE_SHIFT;
arch/powerpc/mm/init_64.c
253
int nr_pfns = page_size >> PAGE_SHIFT;
arch/powerpc/mm/init_64.c
353
page = pfn_to_page(addr >> PAGE_SHIFT);
arch/powerpc/mm/kasan/init_book3e_64.c
61
__set_pte_at(&init_mm, ea, ptep, pfn_pte(pa >> PAGE_SHIFT, prot), 0);
arch/powerpc/mm/mem.c
139
update_end_of_memory_vars(start_pfn << PAGE_SHIFT,
arch/powerpc/mm/mem.c
140
nr_pages << PAGE_SHIFT);
arch/powerpc/mm/mem.c
148
unsigned long start_pfn = start >> PAGE_SHIFT;
arch/powerpc/mm/mem.c
149
unsigned long nr_pages = size >> PAGE_SHIFT;
arch/powerpc/mm/mem.c
163
unsigned long start_pfn = start >> PAGE_SHIFT;
arch/powerpc/mm/mem.c
164
unsigned long nr_pages = size >> PAGE_SHIFT;
arch/powerpc/mm/mem.c
174
max_low_pfn = max_pfn = memblock_end_of_DRAM() >> PAGE_SHIFT;
arch/powerpc/mm/mem.c
175
min_low_pfn = MEMORY_START >> PAGE_SHIFT;
arch/powerpc/mm/mem.c
177
max_low_pfn = lowmem_end_addr >> PAGE_SHIFT;
arch/powerpc/mm/mem.c
223
max_zone_pfns[ZONE_DMA] = min((zone_dma_limit >> PAGE_SHIFT) + 1, max_low_pfn);
arch/powerpc/mm/nohash/8xx.c
18
#define IMMR_SIZE (FIX_IMMR_SIZE << PAGE_SHIFT)
arch/powerpc/mm/nohash/8xx.c
91
arch_make_huge_pte(pfn_pte(pa >> PAGE_SHIFT, prot), shift, 0),
arch/powerpc/mm/nohash/book3e_pgtable.c
114
set_pte_at(&init_mm, ea, ptep, pfn_pte(pa >> PAGE_SHIFT, prot));
arch/powerpc/mm/nohash/e500_hugetlbpage.c
157
mas7_3 = (u64)pte_pfn(pte) << PAGE_SHIFT;
arch/powerpc/mm/nohash/tlb_64e.c
76
vpte = (vpte >> (PAGE_SHIFT - 3)) & ~0xffful;
arch/powerpc/mm/numa.c
1000
fake_numa_create_new_node(((start + size) >> PAGE_SHIFT), &nid);
arch/powerpc/mm/numa.c
1160
max_low_pfn = max_pfn = memblock_end_of_DRAM() >> PAGE_SHIFT;
arch/powerpc/mm/numa.c
1161
min_low_pfn = MEMORY_START >> PAGE_SHIFT;
arch/powerpc/mm/numa.c
120
if ((end_pfn << PAGE_SHIFT) > mem) {
arch/powerpc/mm/numa.c
884
fake_numa_create_new_node(((base + size) >> PAGE_SHIFT),
arch/powerpc/mm/pgtable_32.c
87
set_pte_at(&init_mm, va, pg, pfn_pte(pa >> PAGE_SHIFT, prot));
arch/powerpc/perf/vpa-dtl.c
540
buf->size = nr_pages << PAGE_SHIFT;
arch/powerpc/platforms/book3s/vas-api.c
439
(paste_addr >> PAGE_SHIFT));
arch/powerpc/platforms/book3s/vas-api.c
565
pfn = paste_addr >> PAGE_SHIFT;
arch/powerpc/platforms/cell/spufs/file.c
237
offset = vmf->pgoff << PAGE_SHIFT;
arch/powerpc/platforms/cell/spufs/file.c
252
pfn = (ctx->spu->local_store_phys + offset) >> PAGE_SHIFT;
arch/powerpc/platforms/cell/spufs/file.c
315
unsigned long area, offset = vmf->pgoff << PAGE_SHIFT;
arch/powerpc/platforms/cell/spufs/file.c
354
(area + offset) >> PAGE_SHIFT);
arch/powerpc/platforms/cell/spufs/inode.c
702
sb->s_blocksize_bits = PAGE_SHIFT;
arch/powerpc/platforms/powernv/memtrace.c
51
unsigned long ent_nrpages = ent->size >> PAGE_SHIFT;
arch/powerpc/platforms/powernv/opal-memory-errors.c
50
memory_failure(paddr_start >> PAGE_SHIFT, 0);
arch/powerpc/platforms/powernv/opal-prd.c
122
addr = vma->vm_pgoff << PAGE_SHIFT;
arch/powerpc/platforms/powernv/opal.c
1144
uint64_t data = vmalloc_to_pfn(vmalloc_addr) << PAGE_SHIFT;
arch/powerpc/platforms/powernv/pci-ioda-tce.c
300
PAGE_SHIFT);
arch/powerpc/platforms/powernv/pci-ioda-tce.c
312
level_shift = max_t(unsigned int, level_shift, PAGE_SHIFT);
arch/powerpc/platforms/powernv/pci-ioda-tce.c
68
shift - PAGE_SHIFT);
arch/powerpc/platforms/powernv/pci-ioda.c
1379
const u64 maxblock = 1UL << (PAGE_SHIFT + MAX_PAGE_ORDER);
arch/powerpc/platforms/powernv/pci-ioda.c
1389
const u64 window_size = min((maxblock * 8) << PAGE_SHIFT, max_memory);
arch/powerpc/platforms/powernv/pci-ioda.c
1391
unsigned long tces_order = ilog2(window_size >> PAGE_SHIFT);
arch/powerpc/platforms/powernv/pci-ioda.c
1403
rc = pnv_pci_ioda2_create_table(&pe->table_group, 0, PAGE_SHIFT,
arch/powerpc/platforms/powernv/pci-ioda.c
1485
table_shift = max_t(unsigned, table_shift, PAGE_SHIFT);
arch/powerpc/platforms/ps3/spu.c
157
result = lv1_construct_logical_spe(PAGE_SHIFT, PAGE_SHIFT, PAGE_SHIFT,
arch/powerpc/platforms/ps3/spu.c
158
PAGE_SHIFT, PAGE_SHIFT, get_vas_id(), SPE_TYPE_LOGICAL,
arch/powerpc/platforms/pseries/cmm.c
39
#define KB2PAGES(_p) ((_p)>>(PAGE_SHIFT-10))
arch/powerpc/platforms/pseries/cmm.c
40
#define PAGES2KB(_p) ((_p)<<(PAGE_SHIFT-10))
arch/powerpc/platforms/pseries/hotplug-memory.c
245
start_pfn = base >> PAGE_SHIFT;
arch/powerpc/platforms/pseries/iommu.c
1671
ret = walk_system_ram_range(0, ddw_memory_hotplug_max() >> PAGE_SHIFT,
arch/powerpc/platforms/pseries/iommu.c
2433
if (window->direct && (arg->start_pfn << PAGE_SHIFT) <
arch/powerpc/platforms/pseries/iommu.c
2446
if (window->direct && (arg->start_pfn << PAGE_SHIFT) <
arch/powerpc/platforms/pseries/iommu.c
449
next = start_pfn << PAGE_SHIFT;
arch/powerpc/platforms/pseries/iommu.c
450
num_tce = num_pfn << PAGE_SHIFT;
arch/powerpc/platforms/pseries/iommu.c
490
unsigned long dmastart = (start_pfn << PAGE_SHIFT) +
arch/powerpc/platforms/pseries/iommu.c
493
unsigned long npages = num_pfn << PAGE_SHIFT >> tceshift;
arch/powerpc/platforms/pseries/iommu.c
494
void *uaddr = __va(start_pfn << PAGE_SHIFT);
arch/powerpc/platforms/pseries/iommu.c
518
next = start_pfn << PAGE_SHIFT;
arch/powerpc/platforms/pseries/iommu.c
519
num_tce = num_pfn << PAGE_SHIFT;
arch/powerpc/platforms/pseries/iommu.c
954
1ULL << (be32_to_cpu(dwp->window_shift) - PAGE_SHIFT), dwp);
arch/powerpc/platforms/pseries/papr_scm.c
1336
!is_zone_device_page(pfn_to_page(phys_addr >> PAGE_SHIFT)))
arch/powerpc/platforms/pseries/ras.c
651
paddr = pfn << PAGE_SHIFT;
arch/powerpc/sysdev/xive/native.c
595
if (val == PAGE_SHIFT)
arch/powerpc/sysdev/xive/spapr.c
863
if (val == PAGE_SHIFT)
arch/powerpc/sysdev/xive/xive-internal.h
72
return (queue_shift > PAGE_SHIFT) ? (queue_shift - PAGE_SHIFT) : 0;
arch/riscv/include/asm/elf.h
57
0x7ff >> (PAGE_SHIFT - 12) : \
arch/riscv/include/asm/elf.h
58
0x3ffff >> (PAGE_SHIFT - 12))
arch/riscv/include/asm/page.h
197
return __va(pfn << PAGE_SHIFT);
arch/riscv/include/asm/page.h
20
#define HUGETLB_PAGE_ORDER (HPAGE_SHIFT - PAGE_SHIFT)
arch/riscv/include/asm/pgtable-64.h
104
#define napot_cont_shift(order) ((order) + PAGE_SHIFT)
arch/riscv/include/asm/pgtable.h
783
#define __pmd_to_phys(pmd) (__page_val_to_pfn(pmd_val(pmd)) << PAGE_SHIFT)
arch/riscv/include/asm/pgtable.h
787
return ((__pmd_to_phys(pmd) & PMD_MASK) >> PAGE_SHIFT);
arch/riscv/include/asm/pgtable.h
790
#define __pud_to_phys(pud) (__page_val_to_pfn(pud_val(pud)) << PAGE_SHIFT)
arch/riscv/include/asm/pgtable.h
795
return ((__pud_to_phys(pud) & PUD_MASK) >> PAGE_SHIFT);
arch/riscv/include/asm/pgtable.h
85
(VA_BITS - PAGE_SHIFT - 1 + STRUCT_PAGE_MAX_SHIFT)
arch/riscv/include/asm/set_memory.h
25
int num_pages = PAGE_ALIGN(end - start) >> PAGE_SHIFT;
arch/riscv/include/asm/thread_info.h
35
#define THREAD_SHIFT (PAGE_SHIFT + THREAD_SIZE_ORDER)
arch/riscv/kernel/sys_riscv.c
30
offset >> (PAGE_SHIFT - page_shift_offset));
arch/riscv/kernel/vdso.c
125
vdso_text_len = vdso_info->vdso_pages << PAGE_SHIFT;
arch/riscv/kernel/vdso.c
20
#define VVAR_SIZE (VDSO_NR_PAGES << PAGE_SHIFT)
arch/riscv/kernel/vdso.c
56
PAGE_SHIFT;
arch/riscv/kvm/gstage.c
111
unsigned long order = PAGE_SHIFT;
arch/riscv/kvm/gstage.c
70
unsigned long page_order = PAGE_SHIFT;
arch/riscv/kvm/mmu.c
108
phys_addr_t start = (base_gfn + __ffs(mask)) << PAGE_SHIFT;
arch/riscv/kvm/mmu.c
109
phys_addr_t end = (base_gfn + __fls(mask) + 1) << PAGE_SHIFT;
arch/riscv/kvm/mmu.c
140
gpa_t gpa = slot->base_gfn << PAGE_SHIFT;
arch/riscv/kvm/mmu.c
141
phys_addr_t size = slot->npages << PAGE_SHIFT;
arch/riscv/kvm/mmu.c
189
(kvm_riscv_gstage_gpa_size >> PAGE_SHIFT))
arch/riscv/kvm/mmu.c
193
size = new->npages << PAGE_SHIFT;
arch/riscv/kvm/mmu.c
23
phys_addr_t start = memslot->base_gfn << PAGE_SHIFT;
arch/riscv/kvm/mmu.c
24
phys_addr_t end = (memslot->base_gfn + memslot->npages) << PAGE_SHIFT;
arch/riscv/kvm/mmu.c
258
kvm_riscv_gstage_unmap_range(&gstage, range->start << PAGE_SHIFT,
arch/riscv/kvm/mmu.c
259
(range->end - range->start) << PAGE_SHIFT,
arch/riscv/kvm/mmu.c
270
u64 size = (range->end - range->start) << PAGE_SHIFT;
arch/riscv/kvm/mmu.c
282
if (!kvm_riscv_gstage_get_leaf(&gstage, range->start << PAGE_SHIFT,
arch/riscv/kvm/mmu.c
293
u64 size = (range->end - range->start) << PAGE_SHIFT;
arch/riscv/kvm/mmu.c
305
if (!kvm_riscv_gstage_get_leaf(&gstage, range->start << PAGE_SHIFT,
arch/riscv/kvm/mmu.c
323
gpa_start = memslot->base_gfn << PAGE_SHIFT;
arch/riscv/kvm/mmu.c
456
gfn_t gfn = gpa >> PAGE_SHIFT;
arch/riscv/kvm/mmu.c
493
vma_pageshift = PAGE_SHIFT;
arch/riscv/kvm/mmu.c
499
gfn = (gpa & huge_page_mask(hstate_vma(vma))) >> PAGE_SHIFT;
arch/riscv/kvm/mmu.c
547
ret = kvm_riscv_gstage_map_page(&gstage, pcache, gpa, hfn << PAGE_SHIFT,
arch/riscv/kvm/mmu.c
550
ret = kvm_riscv_gstage_map_page(&gstage, pcache, gpa, hfn << PAGE_SHIFT,
arch/riscv/kvm/mmu.c
610
hgatp |= (k->pgd_phys >> PAGE_SHIFT) & HGATP_PPN;
arch/riscv/kvm/tlb.c
463
gfn << PAGE_SHIFT, nr_pages << PAGE_SHIFT,
arch/riscv/kvm/tlb.c
464
PAGE_SHIFT, READ_ONCE(kvm->arch.vmid.vmid));
arch/riscv/kvm/vcpu_exit.c
26
gfn = fault_addr >> PAGE_SHIFT;
arch/riscv/kvm/vcpu_sbi_replace.c
112
cp->a2, cp->a3, PAGE_SHIFT, vmid);
arch/riscv/kvm/vcpu_sbi_replace.c
122
cp->a3, PAGE_SHIFT, cp->a4, vmid);
arch/riscv/kvm/vcpu_sbi_sta.c
46
gfn = shmem >> PAGE_SHIFT;
arch/riscv/kvm/vcpu_sbi_system.c
32
hva = kvm_vcpu_gfn_to_hva_prot(vcpu, cp->a1 >> PAGE_SHIFT, NULL);
arch/riscv/kvm/vcpu_sbi_v01.c
86
cp->a2, PAGE_SHIFT, vmid);
arch/riscv/kvm/vcpu_sbi_v01.c
94
cp->a1, cp->a2, PAGE_SHIFT,
arch/riscv/mm/hugetlbpage.c
231
hugepage_shift = PAGE_SHIFT;
arch/riscv/mm/hugetlbpage.c
445
hugetlb_add_hstate(PUD_SHIFT - PAGE_SHIFT);
arch/riscv/mm/hugetlbpage.c
454
return PUD_SHIFT - PAGE_SHIFT;
arch/riscv/mm/init.c
1116
vmemmap_start_pfn = round_down(phys_ram_base, VMEMMAP_ADDR_ALIGN) >> PAGE_SHIFT;
arch/riscv/mm/init.c
1426
early_memtest(min_low_pfn << PAGE_SHIFT, max_low_pfn << PAGE_SHIFT);
arch/riscv/mm/init.c
1632
vmem_altmap_free(altmap, size >> PAGE_SHIFT);
arch/riscv/mm/init.c
1804
ret = __add_pages(nid, start >> PAGE_SHIFT, size >> PAGE_SHIFT, params);
arch/riscv/mm/init.c
1820
__remove_pages(start >> PAGE_SHIFT, size >> PAGE_SHIFT, altmap);
arch/riscv/mm/init.c
245
vmemmap_start_pfn = round_down(phys_ram_base, VMEMMAP_ADDR_ALIGN) >> PAGE_SHIFT;
arch/riscv/mm/init.c
397
set_pte(ptep, pfn_pte(phys >> PAGE_SHIFT, prot));
arch/riscv/mm/init.c
843
mmap_rnd_bits_max = MMAP_VA_BITS - PAGE_SHIFT - 3;
arch/riscv/mm/pageattr.c
165
pfn_pmd(pfn + ((i * PMD_SIZE) >> PAGE_SHIFT), prot));
arch/riscv/mm/pageattr.c
218
pfn_pud(pfn + ((i * PUD_SIZE) >> PAGE_SHIFT), prot));
arch/riscv/mm/pageattr.c
292
page_start = (start - (unsigned long)area->addr) >> PAGE_SHIFT;
arch/s390/appldata/appldata_mem.c
23
#define P2K(x) ((x) << (PAGE_SHIFT - 10)) /* Converts #Pages to KB */
arch/s390/boot/vmem.c
483
__arch_set_page_nodat((void *)start, (end - start) >> PAGE_SHIFT);
arch/s390/hypfs/inode.c
248
sb->s_blocksize_bits = PAGE_SHIFT;
arch/s390/include/asm/kfence.h
16
unsigned long pool_pages = KFENCE_POOL_SIZE >> PAGE_SHIFT;
arch/s390/include/asm/page.h
25
#define HUGETLB_PAGE_ORDER (HPAGE_SHIFT - PAGE_SHIFT)
arch/s390/include/asm/page.h
257
#define phys_to_pfn(phys) ((phys) >> PAGE_SHIFT)
arch/s390/include/asm/page.h
258
#define pfn_to_phys(pfn) ((pfn) << PAGE_SHIFT)
arch/s390/include/asm/pci_dma.h
57
#define ZPCI_ST_SHIFT (ZPCI_PT_BITS + PAGE_SHIFT)
arch/s390/include/asm/pgtable.h
1319
#define PFN_PTE_SHIFT PAGE_SHIFT
arch/s390/include/asm/pgtable.h
1372
return __pa(pmd_deref(pmd)) >> PAGE_SHIFT;
arch/s390/include/asm/pgtable.h
1388
return __pa(pud_deref(pud)) >> PAGE_SHIFT;
arch/s390/include/asm/pgtable.h
1468
#define pfn_pte(pfn, pgprot) mk_pte_phys(((pfn) << PAGE_SHIFT), (pgprot))
arch/s390/include/asm/pgtable.h
1469
#define pte_pfn(x) (pte_val(x) >> PAGE_SHIFT)
arch/s390/include/asm/pgtable.h
1772
#define pfn_pmd(pfn, pgprot) mk_pmd_phys(((pfn) << PAGE_SHIFT), (pgprot))
arch/s390/include/asm/pgtable.h
361
#define _PAGE_INDEX (0xffUL << PAGE_SHIFT)
arch/s390/include/asm/pgtable.h
687
return (pgd_val(pgd) & origin_mask) >> PAGE_SHIFT;
arch/s390/include/asm/pgtable.h
714
return (p4d_val(p4d) & origin_mask) >> PAGE_SHIFT;
arch/s390/include/asm/set_memory.h
51
numpages = (end - start) >> PAGE_SHIFT; \
arch/s390/kernel/crash_dump.c
187
if (pfn < oldmem_data.size >> PAGE_SHIFT) {
arch/s390/kernel/crash_dump.c
188
size_old = min(size, oldmem_data.size - (pfn << PAGE_SHIFT));
arch/s390/kernel/crash_dump.c
190
pfn + (oldmem_data.start >> PAGE_SHIFT),
arch/s390/kernel/crash_dump.c
196
pfn += size_old >> PAGE_SHIFT;
arch/s390/kernel/crash_dump.c
215
if (pfn < hsa_end >> PAGE_SHIFT) {
arch/s390/kernel/crash_dump.c
216
size_hsa = min(size, hsa_end - (pfn << PAGE_SHIFT));
arch/s390/kernel/crash_dump.c
221
pfn += size_hsa >> PAGE_SHIFT;
arch/s390/kernel/early.c
83
page_set_storage_key(init_pfn << PAGE_SHIFT,
arch/s390/kernel/machine_kexec.c
144
free_reserved_page(pfn_to_page(addr >> PAGE_SHIFT));
arch/s390/kernel/machine_kexec.c
160
set_memory_ro(crashk_res.start, size >> PAGE_SHIFT);
arch/s390/kernel/machine_kexec.c
162
set_memory_rw(crashk_res.start, size >> PAGE_SHIFT);
arch/s390/kernel/numa.c
26
NODE_DATA(0)->node_spanned_pages = memblock_end_of_DRAM() >> PAGE_SHIFT;
arch/s390/kernel/perf_cpum_sf.c
1318
perf_aux_output_end(handle, i << PAGE_SHIFT);
arch/s390/kernel/perf_cpum_sf.c
1344
aux->head = handle->head >> PAGE_SHIFT;
arch/s390/kernel/perf_cpum_sf.c
1345
range = (handle->size + 1) >> PAGE_SHIFT;
arch/s390/kernel/perf_cpum_sf.c
1504
size = aux_sdb_num_alert(aux) << PAGE_SHIFT;
arch/s390/kernel/perf_cpum_sf.c
1506
size >> PAGE_SHIFT);
arch/s390/kernel/perf_cpum_sf.c
1523
aux->head = handle->head >> PAGE_SHIFT;
arch/s390/kernel/perf_cpum_sf.c
1524
range = (handle->size + 1) >> PAGE_SHIFT;
arch/s390/kernel/perf_cpum_sf.c
1535
size = range << PAGE_SHIFT;
arch/s390/kernel/perf_cpum_sf.c
1540
size = aux_sdb_num_alert(aux) << PAGE_SHIFT;
arch/s390/kernel/process.c
239
return (get_random_u16() & BRK_RND_MASK) << PAGE_SHIFT;
arch/s390/kernel/vdso.c
101
offset = get_random_u32_below(((end - start) >> PAGE_SHIFT) + 1);
arch/s390/kernel/vdso.c
102
addr = start + (offset << PAGE_SHIFT);
arch/s390/kernel/vdso.c
131
int pages = (end - start) >> PAGE_SHIFT;
arch/s390/kvm/dat.c
131
asce->rsto = __pa(table) >> PAGE_SHIFT;
arch/s390/kvm/dat.c
813
addr = ((max(gfn, walk->start) - gfn) << PAGE_SHIFT) + origin;
arch/s390/kvm/dat.c
814
end = ((min(next, walk->end) - gfn) << PAGE_SHIFT) + origin;
arch/s390/kvm/dat.h
640
return 1ULL << ((asce.dt + 1) * 11 + _SEGMENT_SHIFT - PAGE_SHIFT);
arch/s390/kvm/gaccess.c
384
teid->addr = gva >> PAGE_SHIFT;
arch/s390/kvm/gaccess.c
951
nr_pages = (((ga & ~PAGE_MASK) + len - 1) >> PAGE_SHIFT) + 1;
arch/s390/kvm/gmap.c
1273
new = gmap_new(parent->kvm, asce.r ? 1UL << (64 - PAGE_SHIFT) : asce_end(asce));
arch/s390/kvm/gmap.c
36
if (limit <= _REGION3_SIZE >> PAGE_SHIFT)
arch/s390/kvm/gmap.c
38
if (limit <= _REGION2_SIZE >> PAGE_SHIFT)
arch/s390/kvm/gmap.c
40
if (limit <= _REGION1_SIZE >> PAGE_SHIFT)
arch/s390/kvm/gmap.c
723
dat_set_ptval(pt, PTVAL_VMADDR, p_gfn >> (_SEGMENT_SHIFT - PAGE_SHIFT));
arch/s390/kvm/gmap.c
943
cur = ((max(gfn, walk->start) - gfn) << PAGE_SHIFT) + origin;
arch/s390/kvm/gmap.c
944
end = ((min(next, walk->end) - gfn) << PAGE_SHIFT) + origin;
arch/s390/kvm/interrupt.c
2774
mark_page_dirty(kvm, adapter_int->ind_gaddr >> PAGE_SHIFT);
arch/s390/kvm/interrupt.c
2780
mark_page_dirty(kvm, adapter_int->summary_gaddr >> PAGE_SHIFT);
arch/s390/kvm/pci.c
208
page_limit = rlimit(RLIMIT_MEMLOCK) >> PAGE_SHIFT;
arch/s390/kvm/priv.c
1158
gfn = vcpu->run->s.regs.gprs[r2] >> PAGE_SHIFT;
arch/s390/kvm/priv.c
1173
cbrlo[entries] = gfn << PAGE_SHIFT;
arch/s390/kvm/vsie.c
680
kvm_release_page_dirty(pfn_to_page(hpa >> PAGE_SHIFT));
arch/s390/mm/extmem.c
211
unsigned long start = qout->segstart >> PAGE_SHIFT;
arch/s390/mm/extmem.c
218
if (start != qout->range[i].start >> PAGE_SHIFT) {
arch/s390/mm/extmem.c
222
start = (qout->range[i].end >> PAGE_SHIFT) + 1;
arch/s390/mm/extmem.c
605
seg->range[i].start >> PAGE_SHIFT,
arch/s390/mm/extmem.c
606
seg->range[i].end >> PAGE_SHIFT,
arch/s390/mm/fault.c
119
table += (address & _PAGE_INDEX) >> PAGE_SHIFT;
arch/s390/mm/hugetlbpage.c
238
return PUD_SHIFT - PAGE_SHIFT;
arch/s390/mm/init.c
239
mem_data.start = arg->start_pfn << PAGE_SHIFT;
arch/s390/mm/init.c
240
mem_data.end = mem_data.start + (arg->nr_pages << PAGE_SHIFT);
arch/s390/mm/init.c
281
unsigned long start_pfn = start >> PAGE_SHIFT;
arch/s390/mm/init.c
282
unsigned long nr_pages = size >> PAGE_SHIFT;
arch/s390/mm/mmap.c
105
info.align_offset = pgoff << PAGE_SHIFT;
arch/s390/mm/mmap.c
144
info.align_offset = pgoff << PAGE_SHIFT;
arch/s390/mm/mmap.c
26
return STACK_RND_MASK << PAGE_SHIFT;
arch/s390/mm/mmap.c
40
return (get_random_u32() & MMAP_RND_MASK) << PAGE_SHIFT;
arch/s390/mm/mmap.c
74
return MMAP_ALIGN_MASK << PAGE_SHIFT;
arch/s390/mm/pageattr.c
123
pte_addr = pmd_pfn(*pmdp) << PAGE_SHIFT;
arch/s390/mm/pageattr.c
210
pmd_addr = pud_pfn(*pudp) << PAGE_SHIFT;
arch/s390/mm/pageattr.c
351
offset = (addr - va_start) >> PAGE_SHIFT;
arch/s390/mm/pgalloc.c
239
pte += (addr & _PAGE_INDEX) >> PAGE_SHIFT;
arch/s390/pci/pci_mmio.c
191
io_addr = (void __iomem *)((args.pfn << PAGE_SHIFT) |
arch/s390/pci/pci_mmio.c
343
io_addr = (void __iomem *)((args.pfn << PAGE_SHIFT) |
arch/sh/include/asm/fixmap.h
79
#define FIXADDR_SIZE (__end_of_fixed_addresses << PAGE_SHIFT)
arch/sh/include/asm/page.h
147
#define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT)
arch/sh/include/asm/page.h
155
#define PFN_START (__MEMORY_START >> PAGE_SHIFT)
arch/sh/include/asm/page.h
157
#define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
arch/sh/include/asm/page.h
158
#define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
arch/sh/include/asm/page.h
30
#define HUGETLB_PAGE_ORDER (HPAGE_SHIFT-PAGE_SHIFT)
arch/sh/include/asm/pgtable-2level.h
15
#define PTE_SHIFT PAGE_SHIFT
arch/sh/include/asm/pgtable-3level.h
22
#define PMD_SHIFT (PAGE_SHIFT + (PAGE_SHIFT - PTE_MAGNITUDE))
arch/sh/include/asm/pgtable.h
67
#define PTE_FLAGS_MASK (~(PTE_PHYS_MASK) << PAGE_SHIFT)
arch/sh/include/asm/pgtable.h
87
#define pte_pfn(x) ((unsigned long)(((x).pte_low >> PAGE_SHIFT)))
arch/sh/include/asm/pgtable_32.h
316
#define PFN_PTE_SHIFT PAGE_SHIFT
arch/sh/include/asm/pgtable_32.h
318
__pte(((unsigned long long)(pfn) << PAGE_SHIFT) | pgprot_val(prot))
arch/sh/include/asm/pgtable_32.h
320
__pmd(((unsigned long long)(pfn) << PAGE_SHIFT) | pgprot_val(prot))
arch/sh/include/asm/pgtable_32.h
332
#define pages_to_mb(x) ((x) >> (20-PAGE_SHIFT))
arch/sh/include/asm/pgtable_32.h
400
#define pmd_pfn(pmd) (__pa(pmd_val(pmd)) >> PAGE_SHIFT)
arch/sh/include/asm/thread_info.h
85
#define THREAD_SIZE_ORDER (THREAD_SHIFT - PAGE_SHIFT)
arch/sh/include/mach-dreamcast/mach/maple.h
10
#define MAPLE_DMA_PAGES ((MAPLE_DMA_ORDER > PAGE_SHIFT) ? \
arch/sh/include/mach-dreamcast/mach/maple.h
11
MAPLE_DMA_ORDER - PAGE_SHIFT : 0)
arch/sh/kernel/cpu/sh4/sq.c
170
page = bitmap_find_free_region(sq_bitmap, 0x04000000 >> PAGE_SHIFT,
arch/sh/kernel/cpu/sh4/sq.c
177
map->sq_addr = P4SEG_STORE_QUE + (page << PAGE_SHIFT);
arch/sh/kernel/cpu/sh4/sq.c
183
psz = (size + (PAGE_SIZE - 1)) >> PAGE_SHIFT;
arch/sh/kernel/cpu/sh4/sq.c
222
page = (map->sq_addr - P4SEG_STORE_QUE) >> PAGE_SHIFT;
arch/sh/kernel/cpu/sh4/sq.c
374
unsigned int nr_pages = 0x04000000 >> PAGE_SHIFT;
arch/sh/kernel/crash_dump.c
22
vaddr = ioremap(pfn << PAGE_SHIFT, PAGE_SIZE);
arch/sh/kernel/io_trapped.c
65
n = len >> PAGE_SHIFT;
arch/sh/kernel/setup.c
201
start = start_pfn << PAGE_SHIFT;
arch/sh/kernel/setup.c
202
end = end_pfn << PAGE_SHIFT;
arch/sh/kernel/swsusp.c
21
unsigned long begin_pfn = __pa(&__nosave_begin) >> PAGE_SHIFT;
arch/sh/kernel/swsusp.c
22
unsigned long end_pfn = PAGE_ALIGN(__pa(&__nosave_end)) >> PAGE_SHIFT;
arch/sh/kernel/sys_sh.c
38
return ksys_mmap_pgoff(addr, len, prot, flags, fd, off>>PAGE_SHIFT);
arch/sh/kernel/sys_sh.c
49
if (pgoff & ((1 << (PAGE_SHIFT - 12)) - 1))
arch/sh/kernel/sys_sh.c
52
pgoff >>= PAGE_SHIFT - 12;
arch/sh/mm/cache-sh2a.c
135
if (((end - begin) >> PAGE_SHIFT) >= MAX_OCACHE_PAGES) {
arch/sh/mm/cache-sh2a.c
169
if (((end - start) >> PAGE_SHIFT) >= MAX_ICACHE_PAGES) {
arch/sh/mm/cache-sh2a.c
66
if (((end - begin) >> PAGE_SHIFT) >= MAX_OCACHE_PAGES) {
arch/sh/mm/cache-sh4.c
228
phys = pfn << PAGE_SHIFT;
arch/sh/mm/cache-sh4.c
51
if (((end - start) >> PAGE_SHIFT) >= MAX_ICACHE_PAGES) {
arch/sh/mm/cache-sh7705.c
173
__flush_dcache_page(pfn << PAGE_SHIFT);
arch/sh/mm/cache.c
262
c->n_aliases = c->alias_mask ? (c->alias_mask >> PAGE_SHIFT) + 1 : 0;
arch/sh/mm/fault.c
109
if (PageHighMem(pfn_to_page(pmd_val(*pmd) >> PAGE_SHIFT)))
arch/sh/mm/init.c
293
max_low_pfn = max_pfn = memblock_end_of_DRAM() >> PAGE_SHIFT;
arch/sh/mm/init.c
294
min_low_pfn = __MEMORY_START >> PAGE_SHIFT;
arch/sh/mm/init.c
91
set_pte(pte, pfn_pte(phys >> PAGE_SHIFT, prot));
arch/sh/mm/ioremap_fixed.c
122
nrpages = map->size >> PAGE_SHIFT;
arch/sh/mm/ioremap_fixed.c
79
nrpages = size >> PAGE_SHIFT;
arch/sh/mm/kmap.c
40
(((addr >> PAGE_SHIFT) & (FIX_N_COLOURS - 1)) +
arch/sh/mm/mmap.c
116
((addr - (pgoff << PAGE_SHIFT)) & shm_align_mask))
arch/sh/mm/mmap.c
146
info.align_offset = pgoff << PAGE_SHIFT;
arch/sh/mm/mmap.c
49
unsigned long off = (pgoff << PAGE_SHIFT) & shm_align_mask;
arch/sh/mm/mmap.c
68
((addr - (pgoff << PAGE_SHIFT)) & shm_align_mask))
arch/sh/mm/mmap.c
96
info.align_offset = pgoff << PAGE_SHIFT;
arch/sh/mm/pmb.c
543
pmbe->vpn >> PAGE_SHIFT, pmbe->ppn >> PAGE_SHIFT,
arch/sh/mm/tlbflush_32.c
50
size = (end - start + (PAGE_SIZE - 1)) >> PAGE_SHIFT;
arch/sh/mm/tlbflush_32.c
85
size = (end - start + (PAGE_SIZE - 1)) >> PAGE_SHIFT;
arch/sparc/include/asm/highmem.h
42
#define PKMAP_SIZE (LAST_PKMAP << PAGE_SHIFT)
arch/sparc/include/asm/highmem.h
43
#define PKMAP_BASE PMD_ALIGN(SRMMU_NOCACHE_VADDR + (SRMMU_MAX_NOCACHE_PAGES << PAGE_SHIFT))
arch/sparc/include/asm/highmem.h
46
#define PKMAP_NR(virt) ((virt - PKMAP_BASE) >> PAGE_SHIFT)
arch/sparc/include/asm/highmem.h
47
#define PKMAP_ADDR(nr) (PKMAP_BASE + ((nr) << PAGE_SHIFT))
arch/sparc/include/asm/io-unit.h
45
unsigned long bmap[(IOUNIT_DMA_SIZE >> (PAGE_SHIFT + 3)) / sizeof(unsigned long)];
arch/sparc/include/asm/io-unit.h
53
#define IOUNIT_BMAP1_END (IOUNIT_DMA_SIZE >> (PAGE_SHIFT + 1))
arch/sparc/include/asm/io-unit.h
55
#define IOUNIT_BMAP2_END IOUNIT_BMAP2_START + (IOUNIT_DMA_SIZE >> (PAGE_SHIFT + 2))
arch/sparc/include/asm/io-unit.h
57
#define IOUNIT_BMAPM_END ((IOUNIT_DMA_SIZE - IOUNIT_DVMA_SIZE) >> PAGE_SHIFT)
arch/sparc/include/asm/leon.h
253
#define PFN(x) ((x) >> PAGE_SHIFT)
arch/sparc/include/asm/page_32.h
129
#define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
arch/sparc/include/asm/page_32.h
131
#define virt_addr_valid(kaddr) ((((unsigned long)(kaddr)-PAGE_OFFSET)>>PAGE_SHIFT) < max_mapnr)
arch/sparc/include/asm/page_64.h
12
#if PAGE_SHIFT < 14
arch/sparc/include/asm/page_64.h
148
#define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT)
arch/sparc/include/asm/page_64.h
150
#define virt_to_page(kaddr) pfn_to_page(__pa(kaddr)>>PAGE_SHIFT)
arch/sparc/include/asm/page_64.h
152
#define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
arch/sparc/include/asm/page_64.h
27
#define HUGETLB_PAGE_ORDER (HPAGE_SHIFT - PAGE_SHIFT)
arch/sparc/include/asm/pgtable_32.h
103
return (pmd_val(pmd) & SRMMU_PTD_PMASK) >> (PAGE_SHIFT-4);
arch/sparc/include/asm/pgtable_32.h
249
#define PFN_PTE_SHIFT (PAGE_SHIFT - 4)
arch/sparc/include/asm/pgtable_32.h
395
offset = ((unsigned long long) GET_PFN(pfn)) << PAGE_SHIFT;
arch/sparc/include/asm/pgtable_32.h
399
return phys_base >> PAGE_SHIFT;
arch/sparc/include/asm/pgtable_64.h
1012
#define __swp_type(entry) (((entry).val >> PAGE_SHIFT) & 0x7fUL)
arch/sparc/include/asm/pgtable_64.h
1013
#define __swp_offset(entry) ((entry).val >> (PAGE_SHIFT + 8UL))
arch/sparc/include/asm/pgtable_64.h
1017
((((long)(type) & 0x7fUL) << PAGE_SHIFT) | \
arch/sparc/include/asm/pgtable_64.h
1018
((long)(offset) << (PAGE_SHIFT + 8UL))) \
arch/sparc/include/asm/pgtable_64.h
1084
unsigned long offset = GET_PFN(pfn) << PAGE_SHIFT;
arch/sparc/include/asm/pgtable_64.h
1090
return phys_base >> PAGE_SHIFT;
arch/sparc/include/asm/pgtable_64.h
220
unsigned long paddr = pfn << PAGE_SHIFT;
arch/sparc/include/asm/pgtable_64.h
250
"i" (21), "i" (21 + PAGE_SHIFT),
arch/sparc/include/asm/pgtable_64.h
251
"i" (8), "i" (8 + PAGE_SHIFT));
arch/sparc/include/asm/pgtable_64.h
51
#define PMD_SHIFT (PAGE_SHIFT + (PAGE_SHIFT-3))
arch/sparc/include/asm/pgtable_64.h
54
#define PMD_BITS (PAGE_SHIFT - 3)
arch/sparc/include/asm/pgtable_64.h
62
#define PUD_BITS (PAGE_SHIFT - 3)
arch/sparc/include/asm/pgtable_64.h
68
#define PGDIR_BITS (PAGE_SHIFT - 3)
arch/sparc/include/asm/pgtable_64.h
840
return ((unsigned long) __va(pfn << PAGE_SHIFT));
arch/sparc/include/asm/pgtable_64.h
850
return ((pmd_t *) __va(pfn << PAGE_SHIFT));
arch/sparc/include/asm/pgtable_64.h
926
maybe_tlb_batch_add(mm, addr, ptep, orig, fullmm, PAGE_SHIFT);
arch/sparc/include/asm/pgtable_64.h
929
#define PFN_PTE_SHIFT PAGE_SHIFT
arch/sparc/include/asm/pgtable_64.h
94
#define PTRS_PER_PTE (1UL << (PAGE_SHIFT-3))
arch/sparc/include/asm/processor_64.h
25
#define VPTE_SIZE (1UL << (VA_BITS - PAGE_SHIFT + 3))
arch/sparc/include/asm/processor_64.h
27
#define VPTE_SIZE (1 << (VA_BITS - PAGE_SHIFT + 3))
arch/sparc/include/asm/thread_info_64.h
102
#if PAGE_SHIFT == 13
arch/sparc/include/asm/thread_info_64.h
104
#define THREAD_SHIFT (PAGE_SHIFT + 1)
arch/sparc/include/asm/thread_info_64.h
107
#define THREAD_SHIFT PAGE_SHIFT
arch/sparc/include/asm/thread_info_64.h
131
#if PAGE_SHIFT == 13
arch/sparc/include/asm/tsb.h
160
srlx REG2, 64 - PAGE_SHIFT, REG2; \
arch/sparc/include/asm/tsb.h
165
srlx REG2, 64 - PAGE_SHIFT, REG2; \
arch/sparc/include/asm/tsb.h
177
srlx REG2, 64 - PAGE_SHIFT, REG2; \
arch/sparc/include/asm/tsb.h
192
srlx REG2, 64 - PAGE_SHIFT, REG2; \
arch/sparc/include/asm/tsb.h
273
srlx REG2, 64 - PAGE_SHIFT, REG2; \
arch/sparc/include/asm/tsb.h
278
srlx REG2, 64 - PAGE_SHIFT, REG2; \
arch/sparc/include/asm/tsb.h
284
srlx REG2, 64 - PAGE_SHIFT, REG2; \
arch/sparc/include/asm/tsb.h
289
srlx REG2, 64 - PAGE_SHIFT, REG2; \
arch/sparc/include/asm/tsb.h
348
srlx VADDR, PAGE_SHIFT, REG2; \
arch/sparc/include/asm/vaddrs.h
51
#define FIXADDR_SIZE ((FIX_KMAP_END + 1) << PAGE_SHIFT)
arch/sparc/include/asm/vaddrs.h
54
#define __fix_to_virt(x) (FIXADDR_TOP - ((x) << PAGE_SHIFT))
arch/sparc/include/asm/viking.h
232
vaddr >>= PAGE_SHIFT;
arch/sparc/include/asm/viking.h
242
vaddr >>= PAGE_SHIFT;
arch/sparc/kernel/ldc.c
1061
iommu_tbl_pool_init(iommu, num_tsb_entries, PAGE_SHIFT,
arch/sparc/kernel/ldc.c
2012
(index << PAGE_SHIFT) |
arch/sparc/kernel/ldc.c
2122
unsigned long base = page_to_pfn(sg_page(sg)) << PAGE_SHIFT;
arch/sparc/kernel/ldc.c
2185
fill_cookies(&state, page_to_pfn(sg_page(s)) << PAGE_SHIFT,
arch/sparc/kernel/ldc.c
2238
npages = PAGE_ALIGN(((cookie & ~PAGE_MASK) + size)) >> PAGE_SHIFT;
arch/sparc/kernel/pci.c
764
vma->vm_pgoff += (ioaddr + pbm->io_space.start) >> PAGE_SHIFT;
arch/sparc/kernel/setup_32.c
335
pfn_base = phys_base >> PAGE_SHIFT;
arch/sparc/kernel/sys_sparc_32.c
116
pgoff >> (PAGE_SHIFT - 12));
arch/sparc/kernel/sys_sparc_32.c
124
return ksys_mmap_pgoff(addr, len, prot, flags, fd, off >> PAGE_SHIFT);
arch/sparc/kernel/sys_sparc_32.c
135
(pgoff >> (PAGE_SHIFT - 12)), flags);
arch/sparc/kernel/sys_sparc_32.c
56
((addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1)))
arch/sparc/kernel/sys_sparc_32.c
73
info.align_offset = pgoff << PAGE_SHIFT;
arch/sparc/kernel/sys_sparc_64.c
118
((addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1)))
arch/sparc/kernel/sys_sparc_64.c
149
info.align_offset = pgoff << PAGE_SHIFT;
arch/sparc/kernel/sys_sparc_64.c
186
((addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1)))
arch/sparc/kernel/sys_sparc_64.c
217
info.align_offset = pgoff << PAGE_SHIFT;
arch/sparc/kernel/sys_sparc_64.c
290
rnd = (val % (1UL << (23UL-PAGE_SHIFT)));
arch/sparc/kernel/sys_sparc_64.c
292
rnd = (val % (1UL << (30UL-PAGE_SHIFT)));
arch/sparc/kernel/sys_sparc_64.c
294
return rnd << PAGE_SHIFT;
arch/sparc/kernel/sys_sparc_64.c
482
retval = ksys_mmap_pgoff(addr, len, prot, flags, fd, off >> PAGE_SHIFT);
arch/sparc/kernel/sys_sparc_64.c
86
unsigned long off = (pgoff<<PAGE_SHIFT) & (SHMLBA-1);
arch/sparc/kernel/traps_64.c
1738
if (pfn_valid(afar >> PAGE_SHIFT))
arch/sparc/kernel/traps_64.c
1739
get_page(pfn_to_page(afar >> PAGE_SHIFT));
arch/sparc/kernel/traps_64.c
2175
if (pfn_valid(addr >> PAGE_SHIFT))
arch/sparc/kernel/traps_64.c
2176
get_page(pfn_to_page(addr >> PAGE_SHIFT));
arch/sparc/lib/copy_page.S
204
#if (PAGE_SHIFT == 16)
arch/sparc/lib/copy_page.S
30
#if (PAGE_SHIFT == 13)
arch/sparc/lib/copy_page.S
32
#elif (PAGE_SHIFT == 16)
arch/sparc/mm/fault_64.c
108
pa = pmd_pfn(*pmdp) << PAGE_SHIFT;
arch/sparc/mm/fault_64.c
123
pa = (pte_pfn(pte) << PAGE_SHIFT);
arch/sparc/mm/hugetlbpage.c
130
shift = PAGE_SHIFT;
arch/sparc/mm/hugetlbpage.c
152
shift = PAGE_SHIFT;
arch/sparc/mm/hugetlbpage.c
170
if (shift == PAGE_SHIFT)
arch/sparc/mm/hugetlbpage.c
249
shift = PAGE_SHIFT;
arch/sparc/mm/hugetlbpage.c
255
shift = PAGE_SHIFT;
arch/sparc/mm/hugetlbpage.c
264
orig_shift = pte_none(orig) ? PAGE_SHIFT : huge_tte_to_shift(orig);
arch/sparc/mm/hugetlbpage.c
292
shift = PAGE_SHIFT;
arch/sparc/mm/hugetlbpage.c
298
shift = PAGE_SHIFT;
arch/sparc/mm/hugetlbpage.c
301
orig_shift = pte_none(entry) ? PAGE_SHIFT : huge_tte_to_shift(entry);
arch/sparc/mm/init_32.c
174
start_pfn >>= PAGE_SHIFT;
arch/sparc/mm/init_32.c
176
max_pfn = end_of_phys_memory >> PAGE_SHIFT;
arch/sparc/mm/init_32.c
181
if (max_low_pfn > pfn_base + (SRMMU_MAXMEM >> PAGE_SHIFT)) {
arch/sparc/mm/init_32.c
182
highstart_pfn = pfn_base + (SRMMU_MAXMEM >> PAGE_SHIFT);
arch/sparc/mm/init_32.c
186
high_pages >> (20 - PAGE_SHIFT));
arch/sparc/mm/init_32.c
192
size = (start_pfn << PAGE_SHIFT) - phys_base;
arch/sparc/mm/init_32.c
197
*pages_avail = (size >> PAGE_SHIFT) - high_pages;
arch/sparc/mm/init_32.c
200
memblock_set_current_limit(max_low_pfn << PAGE_SHIFT);
arch/sparc/mm/init_32.c
249
i = last_valid_pfn >> ((20 - PAGE_SHIFT) + 5);
arch/sparc/mm/init_32.c
64
unsigned long start_pfn = sp_banks[i].base_addr >> PAGE_SHIFT;
arch/sparc/mm/init_32.c
65
unsigned long end_pfn = (sp_banks[i].base_addr + sp_banks[i].num_bytes) >> PAGE_SHIFT;
arch/sparc/mm/init_32.c
82
unsigned long tmp = pfn_base + (SRMMU_MAXMEM >> PAGE_SHIFT);
arch/sparc/mm/init_32.c
85
last_pfn = (sp_banks[0].base_addr + sp_banks[0].num_bytes) >> PAGE_SHIFT;
arch/sparc/mm/init_32.c
87
curr_pfn = sp_banks[i].base_addr >> PAGE_SHIFT;
arch/sparc/mm/init_32.c
95
last_pfn = (sp_banks[i].base_addr + sp_banks[i].num_bytes) >> PAGE_SHIFT;
arch/sparc/mm/init_64.c
1597
end_pfn = memblock_end_of_DRAM() >> PAGE_SHIFT;
arch/sparc/mm/init_64.c
1599
min_low_pfn = (phys_base >> PAGE_SHIFT);
arch/sparc/mm/init_64.c
1631
return pfn_valid(pa >> PAGE_SHIFT);
arch/sparc/mm/init_64.c
1905
unsigned long phys_start = page_to_pfn(page) << PAGE_SHIFT;
arch/sparc/mm/init_64.c
333
hugetlb_add_hstate(HPAGE_64K_SHIFT - PAGE_SHIFT);
arch/sparc/mm/init_64.c
334
hugetlb_add_hstate(HPAGE_SHIFT - PAGE_SHIFT);
arch/sparc/mm/init_64.c
335
hugetlb_add_hstate(HPAGE_256MB_SHIFT - PAGE_SHIFT);
arch/sparc/mm/init_64.c
336
hugetlb_add_hstate(HPAGE_2GB_SHIFT - PAGE_SHIFT);
arch/sparc/mm/init_64.c
444
__update_mmu_tsb_insert(mm, MM_TSB_BASE, PAGE_SHIFT,
arch/sparc/mm/io-unit.c
104
npages = (offset_in_page(phys) + size + (PAGE_SIZE - 1)) >> PAGE_SHIFT;
arch/sparc/mm/io-unit.c
138
phys = IOUNIT_DMA_BASE + (scan << PAGE_SHIFT) + offset_in_page(phys);
arch/sparc/mm/io-unit.c
190
len = ((vaddr & ~PAGE_MASK) + len + (PAGE_SIZE-1)) >> PAGE_SHIFT;
arch/sparc/mm/io-unit.c
191
vaddr = (vaddr - IOUNIT_DMA_BASE) >> PAGE_SHIFT;
arch/sparc/mm/io-unit.c
208
len = ((sg->dma_address & ~PAGE_MASK) + sg->length + (PAGE_SIZE-1)) >> PAGE_SHIFT;
arch/sparc/mm/io-unit.c
209
vaddr = (sg->dma_address - IOUNIT_DMA_BASE) >> PAGE_SHIFT;
arch/sparc/mm/io-unit.c
254
i = ((addr - IOUNIT_DMA_BASE) >> PAGE_SHIFT);
arch/sparc/mm/iommu.c
124
iommu->usemap.num_colors = vac_cache_size >> PAGE_SHIFT;
arch/sparc/mm/iommu.c
189
unsigned long npages = (off + len + PAGE_SIZE - 1) >> PAGE_SHIFT;
arch/sparc/mm/iommu.c
219
busa0 = iommu->start + (ioptex << PAGE_SHIFT);
arch/sparc/mm/iommu.c
288
unsigned int npages = (off + len + PAGE_SIZE-1) >> PAGE_SHIFT;
arch/sparc/mm/iommu.c
289
unsigned int ioptex = (busa - iommu->start) >> PAGE_SHIFT;
arch/sparc/mm/iommu.c
342
ioptex = bit_map_string_get(&iommu->usemap, len >> PAGE_SHIFT,
arch/sparc/mm/iommu.c
343
addr >> PAGE_SHIFT);
arch/sparc/mm/iommu.c
385
iommu_flush_iotlb(first, len >> PAGE_SHIFT);
arch/sparc/mm/iommu.c
389
*dma_handle = iommu->start + (ioptex << PAGE_SHIFT);
arch/sparc/mm/iommu.c
403
int ioptex = (busa - iommu->start) >> PAGE_SHIFT;
arch/sparc/mm/iommu.c
420
bit_map_clear(&iommu->usemap, ioptex, len >> PAGE_SHIFT);
arch/sparc/mm/srmmu.c
350
page = pfn_to_page(__nocache_pa((unsigned long)ptep) >> PAGE_SHIFT);
arch/sparc/mm/srmmu.c
366
page = pfn_to_page(__nocache_pa((unsigned long)ptep) >> PAGE_SHIFT);
arch/sparc/mm/srmmu.c
98
#define SRMMU_NOCACHE_BITMAP_SHIFT (PAGE_SHIFT - 4)
arch/sparc/mm/tlb.c
170
tlb_batch_add_one(mm, vaddr, exec, PAGE_SHIFT);
arch/sparc/mm/tsb.c
131
if (tb->hugepage_shift == PAGE_SHIFT)
arch/sparc/mm/tsb.c
132
__flush_tsb_one(tb, PAGE_SHIFT, base, nentries);
arch/sparc/mm/tsb.c
135
__flush_huge_tsb_one(tb, PAGE_SHIFT, base, nentries,
arch/sparc/mm/tsb.c
164
if (hugepage_shift == PAGE_SHIFT)
arch/sparc/mm/tsb.c
165
__flush_tsb_one_entry(base, vaddr, PAGE_SHIFT,
arch/sparc/mm/tsb.c
169
__flush_huge_tsb_one_entry(base, vaddr, PAGE_SHIFT,
arch/sparc/mm/tsb.c
511
PAGE_SHIFT : REAL_HPAGE_SHIFT);
arch/sparc/mm/tsb.c
56
if ((end - start) >> PAGE_SHIFT >= 2 * KERNEL_TSB_NENTRIES)
arch/sparc/mm/tsb.c
60
unsigned long hash = tsb_hash(v, PAGE_SHIFT,
arch/sparc/mm/ultra.S
207
#if (PAGE_SHIFT != 13)
arch/sparc/vdso/vma.c
117
return start + (offset << PAGE_SHIFT);
arch/um/include/asm/page.h
88
#define phys_to_pfn(p) ((p) >> PAGE_SHIFT)
arch/um/include/asm/pgalloc.h
21
(unsigned long long) PAGE_SHIFT)))
arch/um/include/asm/pgtable-4level.h
107
return __pmd((page_nr << PAGE_SHIFT) | pgprot_val(pgprot));
arch/um/include/asm/pgtable.h
216
#define PFN_PTE_SHIFT PAGE_SHIFT
arch/um/include/asm/pgtable.h
91
#define pmd_pfn(pmd) (pmd_val(pmd) >> PAGE_SHIFT)
arch/um/kernel/asm-offsets.c
23
DEFINE(UM_KERN_PAGE_SHIFT, PAGE_SHIFT);
arch/um/kernel/mem.c
82
max_zone_pfns[ZONE_NORMAL] = high_physmem >> PAGE_SHIFT;
arch/um/kernel/physmem.c
97
max_low_pfn = min_low_pfn + (map_size >> PAGE_SHIFT);
arch/x86/boot/startup/sev-shared.c
621
pc->entry[0].pfn = paddr >> PAGE_SHIFT;
arch/x86/boot/startup/sev-shared.c
675
sev_es_wr_ghcb_msr(GHCB_MSR_PSC_REQ_GFN(paddr >> PAGE_SHIFT, desc->op));
arch/x86/boot/startup/sev-startup.c
57
paddr_end = paddr + (npages << PAGE_SHIFT);
arch/x86/coco/sev/core.c
1479
unsigned int npages = PAGE_ALIGN(sz) >> PAGE_SHIFT;
arch/x86/coco/sev/core.c
1496
unsigned int npages = PAGE_ALIGN(sz) >> PAGE_SHIFT;
arch/x86/coco/sev/core.c
344
pfn = __pa(vaddr) >> PAGE_SHIFT;
arch/x86/coco/sev/core.c
404
vaddr_end = vaddr + (npages << PAGE_SHIFT);
arch/x86/coco/sev/core.c
434
npages = (end - start) >> PAGE_SHIFT;
arch/x86/coco/sev/core.c
598
npages = (end - addr) >> PAGE_SHIFT;
arch/x86/coco/sev/core.c
671
p = pfn_to_online_page(pa >> PAGE_SHIFT);
arch/x86/coco/sev/core.c
960
pfn = address >> PAGE_SHIFT;
arch/x86/coco/sev/core.c
970
pfn = address >> PAGE_SHIFT;
arch/x86/coco/sev/vc-handle.c
66
pa = (phys_addr_t)pte_pfn(*pte) << PAGE_SHIFT;
arch/x86/coco/sev/vc-shared.c
613
unsigned long pfn = paddr >> PAGE_SHIFT;
arch/x86/entry/calling.h
157
#define PTI_USER_PGTABLE_BIT PAGE_SHIFT
arch/x86/entry/vdso/vma.c
57
if (!image || (vmf->pgoff << PAGE_SHIFT) >= image->size)
arch/x86/entry/vdso/vma.c
60
vmf->page = virt_to_page(image->data + (vmf->pgoff << PAGE_SHIFT));
arch/x86/entry/vdso/vma.c
98
__pa(pvti) >> PAGE_SHIFT,
arch/x86/events/intel/bts.c
128
offset += __nr_pages << PAGE_SHIFT;
arch/x86/events/intel/bts.c
328
bb->nr_pages << PAGE_SHIFT);
arch/x86/events/intel/bts.c
396
head = handle->head & ((bb->nr_pages << PAGE_SHIFT) - 1);
arch/x86/events/intel/bts.c
89
size_t size = nr_pages << PAGE_SHIFT;
arch/x86/events/intel/pt.c
1021
if (topa->offset + topa->size > (unsigned long)pg << PAGE_SHIFT)
arch/x86/events/intel/pt.c
1140
npages = handle->size >> PAGE_SHIFT;
arch/x86/events/intel/pt.c
1146
idx = (head >> PAGE_SHIFT) + npages;
arch/x86/events/intel/pt.c
1155
wakeup = handle->wakeup >> PAGE_SHIFT;
arch/x86/events/intel/pt.c
1158
idx = (head >> PAGE_SHIFT) + npages - 1;
arch/x86/events/intel/pt.c
1198
head &= (buf->nr_pages << PAGE_SHIFT) - 1;
arch/x86/events/intel/pt.c
1201
pg = (head >> PAGE_SHIFT) & (buf->nr_pages - 1);
arch/x86/events/intel/pt.c
1700
buf->nr_pages << PAGE_SHIFT);
arch/x86/events/intel/pt.c
1738
from = buf->nr_pages << PAGE_SHIFT;
arch/x86/events/intel/pt.c
895
((buf->nr_pages << PAGE_SHIFT) - 1));
arch/x86/events/intel/pt.c
897
base += buf->nr_pages << PAGE_SHIFT;
arch/x86/hyperv/hv_crash.c
529
set_pte(pte, pfn_pte(addr >> PAGE_SHIFT, PAGE_KERNEL_EXEC));
arch/x86/hyperv/hv_init.c
552
src = memremap(hypercall_msr.guest_physical_address << PAGE_SHIFT, PAGE_SIZE,
arch/x86/hyperv/mmu.c
43
gva_list[gva_n] |= (diff - 1) >> PAGE_SHIFT;
arch/x86/include/asm/acpi.h
242
#define ACPI_TABLE_UPGRADE_MAX_PHYS (max_low_pfn_mapped << PAGE_SHIFT)
arch/x86/include/asm/boot.h
13
# define MIN_KERNEL_ALIGN_LG2 (PAGE_SHIFT + THREAD_SIZE_ORDER)
arch/x86/include/asm/dma.h
74
#define MAX_DMA_PFN ((16UL * 1024 * 1024) >> PAGE_SHIFT)
arch/x86/include/asm/dma.h
77
#define MAX_DMA32_PFN (1UL << (32 - PAGE_SHIFT))
arch/x86/include/asm/dma.h
84
#define MAX_DMA_ADDRESS ((unsigned long)__va(MAX_DMA_PFN << PAGE_SHIFT))
arch/x86/include/asm/fixmap.h
150
#define FIXADDR_SIZE (__end_of_permanent_fixed_addresses << PAGE_SHIFT)
arch/x86/include/asm/fixmap.h
152
#define FIXADDR_TOT_SIZE (__end_of_fixed_addresses << PAGE_SHIFT)
arch/x86/include/asm/fixmap.h
86
VSYSCALL_PAGE = (FIXADDR_TOP - VSYSCALL_ADDR) >> PAGE_SHIFT,
arch/x86/include/asm/highmem.h
57
#define PKMAP_NR(virt) ((virt-PKMAP_BASE) >> PAGE_SHIFT)
arch/x86/include/asm/highmem.h
58
#define PKMAP_ADDR(nr) (PKMAP_BASE + ((nr) << PAGE_SHIFT))
arch/x86/include/asm/kmsan.h
94
ret = pfn_valid(x >> PAGE_SHIFT);
arch/x86/include/asm/kvm_host.h
158
#define KVM_HPAGE_SHIFT(x) (PAGE_SHIFT + KVM_HPAGE_GFN_SHIFT(x))
arch/x86/include/asm/page.h
62
#define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
arch/x86/include/asm/page.h
68
return __va(pfn << PAGE_SHIFT);
arch/x86/include/asm/page_types.h
23
#define HUGETLB_PAGE_ORDER (HPAGE_SHIFT - PAGE_SHIFT)
arch/x86/include/asm/page_types.h
60
return (phys_addr_t)max_pfn_mapped << PAGE_SHIFT;
arch/x86/include/asm/pgalloc.h
105
paravirt_alloc_pmd(mm, __pa(pmd) >> PAGE_SHIFT);
arch/x86/include/asm/pgalloc.h
113
paravirt_alloc_pud(mm, __pa(pud) >> PAGE_SHIFT);
arch/x86/include/asm/pgalloc.h
119
paravirt_alloc_pud(mm, __pa(pud) >> PAGE_SHIFT);
arch/x86/include/asm/pgalloc.h
136
paravirt_alloc_p4d(mm, __pa(p4d) >> PAGE_SHIFT);
arch/x86/include/asm/pgalloc.h
144
paravirt_alloc_p4d(mm, __pa(p4d) >> PAGE_SHIFT);
arch/x86/include/asm/pgalloc.h
65
paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT);
arch/x86/include/asm/pgalloc.h
72
paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT);
arch/x86/include/asm/pgalloc.h
82
set_pmd(pmd, __pmd(((pteval_t)pfn << PAGE_SHIFT) | _PAGE_TABLE));
arch/x86/include/asm/pgalloc.h
99
paravirt_alloc_pmd(mm, __pa(pmd) >> PAGE_SHIFT);
arch/x86/include/asm/pgtable.h
1033
return npg >> (20 - PAGE_SHIFT);
arch/x86/include/asm/pgtable.h
1406
#define PTI_PGTABLE_SWITCH_BIT PAGE_SHIFT
arch/x86/include/asm/pgtable.h
1473
return (PAGE_SHIFT - PTE_SHIFT) + level * PTE_SHIFT;
arch/x86/include/asm/pgtable.h
254
#define PFN_PTE_SHIFT PAGE_SHIFT
arch/x86/include/asm/pgtable.h
260
return (pfn & PTE_PFN_MASK) >> PAGE_SHIFT;
arch/x86/include/asm/pgtable.h
267
return (pfn & pmd_pfn_mask(pmd)) >> PAGE_SHIFT;
arch/x86/include/asm/pgtable.h
275
return (pfn & pud_pfn_mask(pud)) >> PAGE_SHIFT;
arch/x86/include/asm/pgtable.h
280
return (p4d_val(p4d) & p4d_pfn_mask(p4d)) >> PAGE_SHIFT;
arch/x86/include/asm/pgtable.h
285
return (pgd_val(pgd) & PTE_PFN_MASK) >> PAGE_SHIFT;
arch/x86/include/asm/pgtable.h
732
phys_addr_t pfn = (phys_addr_t)page_nr << PAGE_SHIFT;
arch/x86/include/asm/pgtable.h
743
phys_addr_t pfn = (phys_addr_t)page_nr << PAGE_SHIFT;
arch/x86/include/asm/pgtable.h
751
phys_addr_t pfn = (phys_addr_t)page_nr << PAGE_SHIFT;
arch/x86/include/asm/pgtable_32.h
72
#define LOWMEM_PAGES ((((_ULL(2)<<31) - __PAGE_OFFSET) >> PAGE_SHIFT))
arch/x86/include/asm/processor.h
231
return BIT_ULL(boot_cpu_data.x86_cache_bits - 1 - PAGE_SHIFT);
arch/x86/include/asm/tlbflush.h
335
: PAGE_SHIFT, true)
arch/x86/include/asm/tlbflush.h
345
flush_tlb_mm_range(vma->vm_mm, a, a + PAGE_SIZE, PAGE_SHIFT, false);
arch/x86/include/asm/vga.h
25
set_memory_decrypted(start, (s) >> PAGE_SHIFT); \
arch/x86/include/asm/xen/page.h
303
#define mfn_to_virt(m) (__va(mfn_to_pfn(m) << PAGE_SHIFT))
arch/x86/include/asm/xen/page.h
307
#define gfn_to_virt(g) (__va(gfn_to_pfn(g) << PAGE_SHIFT))
arch/x86/include/asm/xen/page.h
311
return (pte.pte & XEN_PTE_MFN_MASK) >> PAGE_SHIFT;
arch/x86/include/asm/xen/page.h
318
pte.pte = ((phys_addr_t)page_nr << PAGE_SHIFT) |
arch/x86/kernel/acpi/madt_wakeup.c
89
mstart = pfn_mapped[i].start << PAGE_SHIFT;
arch/x86/kernel/acpi/madt_wakeup.c
90
mend = pfn_mapped[i].end << PAGE_SHIFT;
arch/x86/kernel/alternative.c
2625
PAGE_SHIFT, false);
arch/x86/kernel/amd_gart_64.c
266
iommu_page = (dma_addr - iommu_bus_base)>>PAGE_SHIFT;
arch/x86/kernel/amd_gart_64.c
459
iommu_full(dev, pages << PAGE_SHIFT, dir);
arch/x86/kernel/amd_gart_64.c
648
gatt_size = (aper_size >> PAGE_SHIFT) * sizeof(u32);
arch/x86/kernel/amd_gart_64.c
653
if (set_memory_uc((unsigned long)gatt, gatt_size >> PAGE_SHIFT))
arch/x86/kernel/amd_gart_64.c
747
end_pfn = (aper_base>>PAGE_SHIFT) + (aper_size>>PAGE_SHIFT);
arch/x86/kernel/amd_gart_64.c
751
init_memory_mapping(start_pfn<<PAGE_SHIFT, end_pfn<<PAGE_SHIFT,
arch/x86/kernel/amd_gart_64.c
756
iommu_pages = iommu_size >> PAGE_SHIFT;
arch/x86/kernel/amd_gart_64.c
769
iommu_gatt_base = agp_gatt_table + (iommu_start>>PAGE_SHIFT);
arch/x86/kernel/amd_gart_64.c
781
iommu_size >> PAGE_SHIFT);
arch/x86/kernel/amd_gart_64.c
97
PAGE_SIZE) >> PAGE_SHIFT;
arch/x86/kernel/amd_gart_64.c
98
boundary_size = dma_get_seg_boundary_nr_pages(dev, PAGE_SHIFT);
arch/x86/kernel/aperture_64.c
132
register_nosave_region(addr >> PAGE_SHIFT,
arch/x86/kernel/aperture_64.c
133
(addr+aper_size) >> PAGE_SHIFT);
arch/x86/kernel/aperture_64.c
89
aperture_pfn_start = aper_base >> PAGE_SHIFT;
arch/x86/kernel/aperture_64.c
90
aperture_page_count = (32 * 1024 * 1024) << aper_order >> PAGE_SHIFT;
arch/x86/kernel/cpu/amd.c
116
int mbytes = get_num_physpages() >> (20-PAGE_SHIFT);
arch/x86/kernel/cpu/bugs.c
2818
half_pa = (u64)l1tf_pfn_limit() << PAGE_SHIFT;
arch/x86/kernel/cpu/mce/apei.c
44
lsb = find_first_bit((void *)&mem_err->physical_addr_mask, PAGE_SHIFT);
arch/x86/kernel/cpu/mce/apei.c
46
lsb = PAGE_SHIFT;
arch/x86/kernel/cpu/mce/core.c
1455
pfn = (p->mce_addr & MCI_ADDR_PHYSADDR) >> PAGE_SHIFT;
arch/x86/kernel/cpu/mce/core.c
1484
pfn = (p->mce_addr & MCI_ADDR_PHYSADDR) >> PAGE_SHIFT;
arch/x86/kernel/cpu/mce/core.c
1509
if (count > 1 && (current->mce_addr >> PAGE_SHIFT) != (m->addr >> PAGE_SHIFT))
arch/x86/kernel/cpu/mce/core.c
1708
struct page *p = pfn_to_online_page(m->addr >> PAGE_SHIFT);
arch/x86/kernel/cpu/mce/core.c
340
p = pfn_to_online_page(final->m.addr >> PAGE_SHIFT);
arch/x86/kernel/cpu/mce/core.c
570
return MCI_MISC_ADDR_LSB(m->misc) >= PAGE_SHIFT;
arch/x86/kernel/cpu/mce/core.c
646
pfn = (mce->addr & MCI_ADDR_PHYSADDR) >> PAGE_SHIFT;
arch/x86/kernel/cpu/mce/intel.c
530
if (MCI_MISC_ADDR_LSB(m->misc) > PAGE_SHIFT)
arch/x86/kernel/cpu/microcode/amd.c
713
if (p_addr >> PAGE_SHIFT != p_addr_end >> PAGE_SHIFT)
arch/x86/kernel/cpu/mtrr/amd.c
106
if (type > MTRR_TYPE_WRCOMB || size < (1 << (17 - PAGE_SHIFT))
arch/x86/kernel/cpu/mtrr/amd.c
20
*base = (low & 0xFFFE0000) >> PAGE_SHIFT;
arch/x86/kernel/cpu/mtrr/amd.c
46
*size = (low + 4) << (15 - PAGE_SHIFT);
arch/x86/kernel/cpu/mtrr/amd.c
83
regs[reg] = (-size >> (15 - PAGE_SHIFT) & 0x0001FFFC)
arch/x86/kernel/cpu/mtrr/amd.c
84
| (base << PAGE_SHIFT) | (type + 1);
arch/x86/kernel/cpu/mtrr/centaur.c
52
*base = centaur_mcr[reg].high >> PAGE_SHIFT;
arch/x86/kernel/cpu/mtrr/centaur.c
53
*size = -(centaur_mcr[reg].low & 0xfffff000) >> PAGE_SHIFT;
arch/x86/kernel/cpu/mtrr/centaur.c
74
high = base << PAGE_SHIFT;
arch/x86/kernel/cpu/mtrr/centaur.c
77
low = -size << PAGE_SHIFT | 0x1f;
arch/x86/kernel/cpu/mtrr/centaur.c
80
low = -size << PAGE_SHIFT | 0x02; /* NC */
arch/x86/kernel/cpu/mtrr/centaur.c
82
low = -size << PAGE_SHIFT | 0x09; /* WWO, WC */
arch/x86/kernel/cpu/mtrr/cleanup.c
181
range_state[reg].base_pfn = basek >> (PAGE_SHIFT - 10);
arch/x86/kernel/cpu/mtrr/cleanup.c
182
range_state[reg].size_pfn = sizek >> (PAGE_SHIFT - 10);
arch/x86/kernel/cpu/mtrr/cleanup.c
193
basek = range_state[reg].base_pfn << (PAGE_SHIFT - 10);
arch/x86/kernel/cpu/mtrr/cleanup.c
194
sizek = range_state[reg].size_pfn << (PAGE_SHIFT - 10);
arch/x86/kernel/cpu/mtrr/cleanup.c
392
basek = base_pfn << (PAGE_SHIFT - 10);
arch/x86/kernel/cpu/mtrr/cleanup.c
393
sizek = size_pfn << (PAGE_SHIFT - 10);
arch/x86/kernel/cpu/mtrr/cleanup.c
497
#define PSHIFT (PAGE_SHIFT - 10)
arch/x86/kernel/cpu/mtrr/cleanup.c
511
size_base = range_state[i].size_pfn << (PAGE_SHIFT - 10);
arch/x86/kernel/cpu/mtrr/cleanup.c
516
start_base = range_state[i].base_pfn << (PAGE_SHIFT - 10);
arch/x86/kernel/cpu/mtrr/cleanup.c
697
x_remove_base = 1 << (32 - PAGE_SHIFT);
arch/x86/kernel/cpu/mtrr/cleanup.c
699
x_remove_size = (mtrr_tom2 >> PAGE_SHIFT) - x_remove_base;
arch/x86/kernel/cpu/mtrr/cleanup.c
706
1ULL<<(20 - PAGE_SHIFT));
arch/x86/kernel/cpu/mtrr/cleanup.c
713
range_sums >> (20 - PAGE_SHIFT));
arch/x86/kernel/cpu/mtrr/cleanup.c
82
if (base < (1<<(20-PAGE_SHIFT)) && mtrr_state.have_fixed &&
arch/x86/kernel/cpu/mtrr/cleanup.c
835
trim_start <<= PAGE_SHIFT;
arch/x86/kernel/cpu/mtrr/cleanup.c
838
trim_size <<= PAGE_SHIFT;
arch/x86/kernel/cpu/mtrr/cleanup.c
87
if (base + size <= (1<<(20-PAGE_SHIFT)))
arch/x86/kernel/cpu/mtrr/cleanup.c
89
size -= (1<<(20-PAGE_SHIFT)) - base;
arch/x86/kernel/cpu/mtrr/cleanup.c
90
base = 1<<(20-PAGE_SHIFT);
arch/x86/kernel/cpu/mtrr/cleanup.c
928
range[nr_range].start = (1ULL<<(32 - PAGE_SHIFT));
arch/x86/kernel/cpu/mtrr/cleanup.c
929
range[nr_range].end = mtrr_tom2 >> PAGE_SHIFT;
arch/x86/kernel/cpu/mtrr/cyrix.c
228
base <<= PAGE_SHIFT;
arch/x86/kernel/cpu/mtrr/cyrix.c
35
*base >>= PAGE_SHIFT;
arch/x86/kernel/cpu/mtrr/generic.c
1007
vr->base_lo = base << PAGE_SHIFT | type;
arch/x86/kernel/cpu/mtrr/generic.c
1008
vr->base_hi = (base >> (32 - PAGE_SHIFT)) & ~phys_hi_rsvd;
arch/x86/kernel/cpu/mtrr/generic.c
1009
vr->mask_lo = -size << PAGE_SHIFT | MTRR_PHYSMASK_V;
arch/x86/kernel/cpu/mtrr/generic.c
1010
vr->mask_hi = (-size >> (32 - PAGE_SHIFT)) & ~phys_hi_rsvd;
arch/x86/kernel/cpu/mtrr/generic.c
1031
if (base & ((1 << (22 - PAGE_SHIFT)) - 1)) {
arch/x86/kernel/cpu/mtrr/generic.c
667
high_width = (boot_cpu_data.x86_phys_bits - (32 - PAGE_SHIFT) + 3) / 4;
arch/x86/kernel/cpu/mtrr/generic.c
851
*size = -mask >> PAGE_SHIFT;
arch/x86/kernel/cpu/mtrr/generic.c
852
*base = (u64)base_hi << (32 - PAGE_SHIFT) | base_lo >> PAGE_SHIFT;
arch/x86/kernel/cpu/mtrr/if.c
144
base >>= PAGE_SHIFT;
arch/x86/kernel/cpu/mtrr/if.c
145
size >>= PAGE_SHIFT;
arch/x86/kernel/cpu/mtrr/if.c
255
if (base + size - 1 >= (1UL << (8 * sizeof(gentry.size) - PAGE_SHIFT))
arch/x86/kernel/cpu/mtrr/if.c
256
|| size >= (1UL << (8 * sizeof(gentry.size) - PAGE_SHIFT)))
arch/x86/kernel/cpu/mtrr/if.c
259
gentry.base = base << PAGE_SHIFT;
arch/x86/kernel/cpu/mtrr/if.c
260
gentry.size = size << PAGE_SHIFT;
arch/x86/kernel/cpu/mtrr/if.c
369
if (size < (0x100000 >> PAGE_SHIFT)) {
arch/x86/kernel/cpu/mtrr/if.c
372
size <<= PAGE_SHIFT - 10;
arch/x86/kernel/cpu/mtrr/if.c
375
size >>= 20 - PAGE_SHIFT;
arch/x86/kernel/cpu/mtrr/if.c
379
i, base, base >> (20 - PAGE_SHIFT),
arch/x86/kernel/cpu/mtrr/if.c
54
base >>= PAGE_SHIFT;
arch/x86/kernel/cpu/mtrr/if.c
55
size >>= PAGE_SHIFT;
arch/x86/kernel/cpu/mtrr/if.c
73
base >>= PAGE_SHIFT;
arch/x86/kernel/cpu/mtrr/if.c
74
size >>= PAGE_SHIFT;
arch/x86/kernel/cpu/mtrr/mtrr.c
249
(boot_cpu_data.x86_phys_bits - PAGE_SHIFT)) {
arch/x86/kernel/cpu/mtrr/mtrr.c
377
return mtrr_add_page(base >> PAGE_SHIFT, size >> PAGE_SHIFT, type,
arch/x86/kernel/cpu/mtrr/mtrr.c
465
return mtrr_del_page(reg, base >> PAGE_SHIFT, size >> PAGE_SHIFT);
arch/x86/kernel/cpu/sgx/main.c
629
unsigned long nr_pages = size >> PAGE_SHIFT;
arch/x86/kernel/cpu/sgx/main.c
683
struct sgx_epc_page *page = sgx_paddr_to_page(pfn << PAGE_SHIFT);
arch/x86/kernel/crash.c
552
old_elfcorehdr = kmap_local_page(pfn_to_page(mem >> PAGE_SHIFT));
arch/x86/kernel/crash_dump_64.c
25
vaddr = (__force void *)ioremap_encrypted(pfn << PAGE_SHIFT, PAGE_SIZE);
arch/x86/kernel/crash_dump_64.c
27
vaddr = (__force void *)ioremap_cache(pfn << PAGE_SHIFT, PAGE_SIZE);
arch/x86/kernel/dumpstack_64.c
117
k = (stk - begin) >> PAGE_SHIFT;
arch/x86/kernel/e820.c
695
max_gap_start = (max_pfn << PAGE_SHIFT) + SZ_1M;
arch/x86/kernel/e820.c
845
# define MAX_ARCH_PFN (1ULL<<(36-PAGE_SHIFT))
arch/x86/kernel/e820.c
847
# define MAX_ARCH_PFN (1ULL<<(32-PAGE_SHIFT))
arch/x86/kernel/e820.c
850
# define MAX_ARCH_PFN MAXMEM>>PAGE_SHIFT
arch/x86/kernel/e820.c
871
start_pfn = entry->addr >> PAGE_SHIFT;
arch/x86/kernel/e820.c
872
end_pfn = (entry->addr + entry->size) >> PAGE_SHIFT;
arch/x86/kernel/e820.c
899
return e820__end_ram_pfn(1UL << (32 - PAGE_SHIFT));
arch/x86/kernel/espfix_64.c
169
paravirt_alloc_pmd(&init_mm, __pa(pmd_p) >> PAGE_SHIFT);
arch/x86/kernel/espfix_64.c
181
paravirt_alloc_pte(&init_mm, __pa(pte_p) >> PAGE_SHIFT);
arch/x86/kernel/espfix_64.c
45
#define ESPFIX_PAGE_SPACE (1UL << (P4D_SHIFT-PAGE_SHIFT-16))
arch/x86/kernel/espfix_64.c
82
addr = (page << PAGE_SHIFT) + (slot * ESPFIX_STACK_SIZE);
arch/x86/kernel/head32.c
131
unsigned long limit = __pa_nodebug(_end) + (PAGE_TABLE_SIZE(LOWMEM_PAGES) << PAGE_SHIFT);
arch/x86/kernel/head32.c
142
*ptr = (pte.pte & PTE_PFN_MASK) >> PAGE_SHIFT;
arch/x86/kernel/irq_64.c
44
phys_addr_t pa = per_cpu_ptr_to_phys(stack + (i << PAGE_SHIFT));
arch/x86/kernel/irq_64.c
46
pages[i] = pfn_to_page(pa >> PAGE_SHIFT);
arch/x86/kernel/kvm.c
944
kvm_sev_hypercall3(KVM_HC_MAP_GPA_RANGE, pfn << PAGE_SHIFT, npages,
arch/x86/kernel/ldt.c
312
unsigned long offset = i << PAGE_SHIFT;
arch/x86/kernel/ldt.c
364
unsigned long offset = i << PAGE_SHIFT;
arch/x86/kernel/ldt.c
377
flush_tlb_mm_range(mm, va, va + nr_pages * PAGE_SIZE, PAGE_SHIFT, false);
arch/x86/kernel/machine_kexec_32.c
202
<< PAGE_SHIFT);
arch/x86/kernel/machine_kexec_32.c
97
set_pte(pte, pfn_pte(paddr >> PAGE_SHIFT, PAGE_KERNEL_EXEC));
arch/x86/kernel/machine_kexec_64.c
218
set_pte(pte, pfn_pte(paddr >> PAGE_SHIFT, prot));
arch/x86/kernel/machine_kexec_64.c
264
mstart = pfn_mapped[i].start << PAGE_SHIFT;
arch/x86/kernel/machine_kexec_64.c
265
mend = pfn_mapped[i].end << PAGE_SHIFT;
arch/x86/kernel/machine_kexec_64.c
374
kexec_pa_swap_page = page_to_pfn(image->swap_page) << PAGE_SHIFT;
arch/x86/kernel/machine_kexec_64.c
642
page = pfn_to_page(start >> PAGE_SHIFT);
arch/x86/kernel/machine_kexec_64.c
643
nr_pages = (end >> PAGE_SHIFT) - (start >> PAGE_SHIFT) + 1;
arch/x86/kernel/setup.c
1057
if (max_pfn > (1UL<<(32 - PAGE_SHIFT)))
arch/x86/kernel/setup.c
1105
(max_pfn_mapped<<PAGE_SHIFT) - 1);
arch/x86/kernel/setup.c
1192
dma_contiguous_reserve(max_pfn_mapped << PAGE_SHIFT);
arch/x86/kernel/static_call.c
215
if (((addr >> PAGE_SHIFT) != ((addr + 7) >> PAGE_SHIFT)) &&
arch/x86/kernel/sys_ia32.c
235
a.offset>>PAGE_SHIFT);
arch/x86/kernel/sys_x86_64.c
155
info.align_offset = pgoff << PAGE_SHIFT;
arch/x86/kernel/sys_x86_64.c
210
info.align_offset = pgoff << PAGE_SHIFT;
arch/x86/kernel/sys_x86_64.c
89
return ksys_mmap_pgoff(addr, len, prot, flags, fd, off >> PAGE_SHIFT);
arch/x86/kernel/tboot.c
174
if (map_tboot_pages(map_base << PAGE_SHIFT, map_base, map_size))
arch/x86/kernel/tboot.c
444
TXT_PRIV_CONFIG_REGS_BASE) >> PAGE_SHIFT)
arch/x86/kvm/hyperv.c
771
msg_page_gfn = synic->msg_page >> PAGE_SHIFT;
arch/x86/kvm/mmu.h
76
return (1ULL << (max_gpa_bits - PAGE_SHIFT)) - 1;
arch/x86/kvm/mmu/mmu.c
1448
if (ALIGN(start << PAGE_SHIFT, PMD_SIZE) !=
arch/x86/kvm/mmu/mmu.c
1449
ALIGN(end << PAGE_SHIFT, PMD_SIZE))
arch/x86/kvm/mmu/mmu.c
308
return gpa >> PAGE_SHIFT;
arch/x86/kvm/mmu/mmu.c
3484
send_sig_mceerr(BUS_MCEERR_AR, (void __user *)hva, PAGE_SHIFT, current);
arch/x86/kvm/mmu/mmu.c
3951
root = mmu_alloc_root(vcpu, i << (30 - PAGE_SHIFT), 0,
arch/x86/kvm/mmu/mmu.c
4066
root_gfn = (root_pgd & __PT_BASE_ADDR_MASK) >> PAGE_SHIFT;
arch/x86/kvm/mmu/mmu.c
4083
if (!kvm_vcpu_is_visible_gfn(vcpu, pdptrs[i] >> PAGE_SHIFT))
arch/x86/kvm/mmu/mmu.c
4145
root_gfn = pdptrs[i] >> PAGE_SHIFT;
arch/x86/kvm/mmu/mmu.c
6234
quadrant = page_offset >> PAGE_SHIFT;
arch/x86/kvm/mmu/mmu.c
6247
gfn_t gfn = gpa >> PAGE_SHIFT;
arch/x86/kvm/mmu/mmu.c
645
return sp->shadowed_translation[index] >> PAGE_SHIFT;
arch/x86/kvm/mmu/mmu.c
680
sp->shadowed_translation[index] = (gfn << PAGE_SHIFT) | access;
arch/x86/kvm/mmu/mmu_internal.h
20
(PAGE_SHIFT + ((level) - 1) * (bits_per_level))
arch/x86/kvm/mmu/mmu_internal.h
25
((base_addr_mask) & ~((1ULL << (PAGE_SHIFT + (((level) - 1) * (bits_per_level)))) - 1))
arch/x86/kvm/mmu/mmu_internal.h
28
((base_addr_mask) & ((1ULL << (PAGE_SHIFT + (((level) - 1) * (bits_per_level)))) - 1))
arch/x86/kvm/mmu/mmu_internal.h
335
kvm_prepare_memory_fault_exit(vcpu, fault->gfn << PAGE_SHIFT,
arch/x86/kvm/mmu/mmutrace.h
121
__entry->gpa = ((u64)table_gfn << PAGE_SHIFT)
arch/x86/kvm/mmu/paging_tmpl.h
106
return (gpte & PT_LVL_ADDR_MASK(lvl)) >> PAGE_SHIFT;
arch/x86/kvm/mmu/paging_tmpl.h
441
gfn += (addr & PT_LVL_OFFSET_MASK(walker->level)) >> PAGE_SHIFT;
arch/x86/kvm/mmu/paging_tmpl.h
452
walker->gfn = real_gpa >> PAGE_SHIFT;
arch/x86/kvm/mmu/paging_tmpl.h
98
int shift = 32 - PT32_DIR_PSE36_SHIFT - PAGE_SHIFT;
arch/x86/kvm/mmu/spte.c
253
spte |= (u64)pfn << PAGE_SHIFT;
arch/x86/kvm/mmu/spte.c
349
child_spte |= (index * KVM_PAGES_PER_HPAGE(role.level)) << PAGE_SHIFT;
arch/x86/kvm/mmu/spte.c
547
GENMASK_ULL(low_phys_bits - 1, PAGE_SHIFT);
arch/x86/kvm/mmu/spte.c
96
u64 gpa = gfn << PAGE_SHIFT;
arch/x86/kvm/mmu/spte.h
251
return zero_pfn(0) << PAGE_SHIFT;
arch/x86/kvm/mmu/spte.h
256
return is_zero_pfn(shadow_page >> PAGE_SHIFT);
arch/x86/kvm/mmu/spte.h
261
struct page *page = pfn_to_page((shadow_page) >> PAGE_SHIFT);
arch/x86/kvm/mmu/spte.h
365
return (pte & SPTE_BASE_ADDR_MASK) >> PAGE_SHIFT;
arch/x86/kvm/mmu/tdp_iter.c
118
if (SPTE_INDEX((iter->gfn | iter->gfn_bits) << PAGE_SHIFT, iter->level) ==
arch/x86/kvm/mmu/tdp_iter.c
15
SPTE_INDEX((iter->gfn | iter->gfn_bits) << PAGE_SHIFT, iter->level);
arch/x86/kvm/mmu/tdp_iter.c
73
return (tdp_ptep_t)__va(spte_to_pfn(spte) << PAGE_SHIFT);
arch/x86/kvm/mmu/tdp_mmu.c
1944
gfn_t gfn = addr >> PAGE_SHIFT;
arch/x86/kvm/svm/nested.c
1365
svm->nested.vmcb02.pa = __sme_set(page_to_pfn(vmcb02_page) << PAGE_SHIFT);
arch/x86/kvm/svm/sev.c
1689
data.guest_address = (page_to_pfn(guest_page[0]) << PAGE_SHIFT) + offset;
arch/x86/kvm/svm/sev.c
1875
data.guest_address = (page_to_pfn(guest_page[0]) << PAGE_SHIFT) + offset;
arch/x86/kvm/svm/sev.c
2311
ret = rmp_make_private(pfn, gfn << PAGE_SHIFT, PG_LEVEL_4K,
arch/x86/kvm/svm/sev.c
2455
u64 pfn = __pa(svm->sev_es.vmsa) >> PAGE_SHIFT;
arch/x86/kvm/svm/sev.c
313
snp_leak_pages(pfn, page_level_size(level) >> PAGE_SHIFT);
arch/x86/kvm/svm/sev.c
3269
u64 pfn = __pa(svm->sev_es.vmsa) >> PAGE_SHIFT;
arch/x86/kvm/svm/sev.c
339
data.paddr = __sme_set(pfn << PAGE_SHIFT);
arch/x86/kvm/svm/sev.c
4408
if (kvm_vcpu_map(vcpu, ghcb_gpa >> PAGE_SHIFT, &svm->sev_es.ghcb_map)) {
arch/x86/kvm/svm/sev.c
4853
gfn = gpa >> PAGE_SHIFT;
arch/x86/kvm/svm/sev.c
694
first = (uaddr & PAGE_MASK) >> PAGE_SHIFT;
arch/x86/kvm/svm/sev.c
695
last = ((uaddr + ulen - 1) & PAGE_MASK) >> PAGE_SHIFT;
arch/x86/kvm/svm/sev.c
699
lock_limit = rlimit(RLIMIT_MEMLOCK) >> PAGE_SHIFT;
arch/x86/kvm/svm/svm.c
1291
svm->vmcb01.pa = __sme_set(page_to_pfn(vmcb01_page) << PAGE_SHIFT);
arch/x86/kvm/svm/svm.h
36
return __sme_set(page_to_pfn(page) << PAGE_SHIFT);
arch/x86/kvm/svm/svm.h
41
return pfn_to_page(__sme_clr(pa) >> PAGE_SHIFT);
arch/x86/kvm/vmx/nested.c
5554
vmx->nested.current_vmptr >> PAGE_SHIFT,
arch/x86/kvm/vmx/nested.c
6146
if (kvm_vcpu_read_guest_page(vcpu, vmcs12->eptp_list_address >> PAGE_SHIFT,
arch/x86/kvm/vmx/tdx.c
3182
region.gpa + (region.nr_pages << PAGE_SHIFT) <= region.gpa ||
arch/x86/kvm/vmx/tdx.c
3184
!vt_is_tdx_private_gpa(kvm, region.gpa + (region.nr_pages << PAGE_SHIFT) - 1))
arch/x86/kvm/vmx/vmx.c
6534
kvm_vcpu_mark_page_dirty(vcpu, gpa >> PAGE_SHIFT);
arch/x86/kvm/vmx/vmx.c
7031
const gfn_t gfn = APIC_DEFAULT_PHYS_BASE >> PAGE_SHIFT;
arch/x86/kvm/x86.c
1062
gfn_t pdpt_gfn = cr3 >> PAGE_SHIFT;
arch/x86/kvm/x86.c
13499
ugfn = slot->userspace_addr >> PAGE_SHIFT;
arch/x86/kvm/x86.c
7886
ret = kvm_vcpu_read_guest_page(vcpu, gpa >> PAGE_SHIFT, data,
arch/x86/kvm/x86.c
7921
ret = kvm_vcpu_read_guest_page(vcpu, gpa >> PAGE_SHIFT, val,
arch/x86/kvm/x86.c
8084
*gpa = vcpu->arch.mmio_gfn << PAGE_SHIFT |
arch/x86/kvm/x86.h
408
vcpu->arch.mmio_gfn == gpa >> PAGE_SHIFT)
arch/x86/mm/cpu_entry_area.c
82
pte_t pte = pfn_pte(pa >> PAGE_SHIFT, flags);
arch/x86/mm/fault.c
305
if (!low_pfn(pgd_val(*pgd) >> PAGE_SHIFT) || !pgd_present(*pgd))
arch/x86/mm/fault.c
939
lsb = PAGE_SHIFT;
arch/x86/mm/hugetlbpage.c
39
hugetlb_add_hstate(PUD_SHIFT - PAGE_SHIFT);
arch/x86/mm/hugetlbpage.c
49
return PUD_SHIFT - PAGE_SHIFT;
arch/x86/mm/init.c
1048
l1tf_limit <<= PAGE_SHIFT - SWP_OFFSET_FIRST_BIT;
arch/x86/mm/init.c
134
order = get_order((unsigned long)num << PAGE_SHIFT);
arch/x86/mm/init.c
144
min_pfn_mapped << PAGE_SHIFT,
arch/x86/mm/init.c
145
max_pfn_mapped << PAGE_SHIFT);
arch/x86/mm/init.c
153
pfn = ret >> PAGE_SHIFT;
arch/x86/mm/init.c
162
adr = __va((pfn + i) << PAGE_SHIFT);
arch/x86/mm/init.c
166
return __va(pfn << PAGE_SHIFT);
arch/x86/mm/init.c
194
pgt_buf_start = base >> PAGE_SHIFT;
arch/x86/mm/init.c
196
pgt_buf_top = pgt_buf_start + (tables >> PAGE_SHIFT);
arch/x86/mm/init.c
333
mr[nr_range].start = start_pfn<<PAGE_SHIFT;
arch/x86/mm/init.c
334
mr[nr_range].end = end_pfn<<PAGE_SHIFT;
arch/x86/mm/init.c
358
if ((end >> PAGE_SHIFT) > max_low_pfn)
arch/x86/mm/init.c
512
if (start_pfn < (1UL<<(32-PAGE_SHIFT)))
arch/x86/mm/init.c
514
min(end_pfn, 1UL<<(32-PAGE_SHIFT)));
arch/x86/mm/init.c
552
add_pfn_range_mapped(start >> PAGE_SHIFT, ret >> PAGE_SHIFT);
arch/x86/mm/init.c
554
return ret >> PAGE_SHIFT;
arch/x86/mm/init.c
588
can_use_brk_pgt = max(start, (u64)pgt_buf_end<<PAGE_SHIFT) >=
arch/x86/mm/init.c
589
min(end, (u64)pgt_buf_top<<PAGE_SHIFT);
arch/x86/mm/init.c
614
return step_size << (PMD_SHIFT - PAGE_SHIFT - 1);
arch/x86/mm/init.c
656
min_pfn_mapped = real_end >> PAGE_SHIFT;
arch/x86/mm/init.c
677
min_pfn_mapped = last_start >> PAGE_SHIFT;
arch/x86/mm/init.c
706
min_pfn_mapped = start >> PAGE_SHIFT;
arch/x86/mm/init.c
767
end = max_pfn << PAGE_SHIFT;
arch/x86/mm/init.c
769
end = max_low_pfn << PAGE_SHIFT;
arch/x86/mm/init.c
812
early_memtest(0, max_pfn_mapped << PAGE_SHIFT);
arch/x86/mm/init.c
886
if (iomem_is_exclusive(pagenr << PAGE_SHIFT)) {
arch/x86/mm/init.c
926
set_memory_np(begin, (end - begin) >> PAGE_SHIFT);
arch/x86/mm/init.c
933
set_memory_nx(begin, (end - begin) >> PAGE_SHIFT);
arch/x86/mm/init.c
934
set_memory_rw(begin, (end - begin) >> PAGE_SHIFT);
arch/x86/mm/init.c
950
unsigned long len_pages = (end_ul - begin_ul) >> PAGE_SHIFT;
arch/x86/mm/init_32.c
266
start_pfn = start >> PAGE_SHIFT;
arch/x86/mm/init_32.c
267
end_pfn = end >> PAGE_SHIFT;
arch/x86/mm/init_32.c
291
pgd_idx = pgd_index((pfn<<PAGE_SHIFT) + PAGE_OFFSET);
arch/x86/mm/init_32.c
299
pmd_idx = pmd_index((pfn<<PAGE_SHIFT) + PAGE_OFFSET);
arch/x86/mm/init_32.c
323
pfn &= PMD_MASK >> PAGE_SHIFT;
arch/x86/mm/init_32.c
342
pte_ofs = pte_index((pfn<<PAGE_SHIFT) + PAGE_OFFSET);
arch/x86/mm/init_32.c
359
last_map_addr = (pfn << PAGE_SHIFT) + PAGE_SIZE;
arch/x86/mm/init_32.c
436
for (pfn = max_low_pfn; pfn < 1<<(32-PAGE_SHIFT); pfn++) {
arch/x86/mm/init_32.c
437
va = PAGE_OFFSET + (pfn<<PAGE_SHIFT);
arch/x86/mm/init_32.c
522
highmem_pages = memparse(arg, &arg) >> PAGE_SHIFT;
arch/x86/mm/init_32.c
636
max_pfn_mapped<<PAGE_SHIFT);
arch/x86/mm/init_32.c
637
printk(KERN_INFO " low ram: 0 - %08lx\n", max_low_pfn<<PAGE_SHIFT);
arch/x86/mm/init_32.c
737
set_memory_nx(start, size >> PAGE_SHIFT);
arch/x86/mm/init_32.c
745
set_pages_ro(virt_to_page(start), size >> PAGE_SHIFT);
arch/x86/mm/init_32.c
753
set_pages_rw(virt_to_page(start), size >> PAGE_SHIFT);
arch/x86/mm/init_32.c
756
set_pages_ro(virt_to_page(start), size >> PAGE_SHIFT);
arch/x86/mm/init_64.c
1293
unsigned long start_pfn = start >> PAGE_SHIFT;
arch/x86/mm/init_64.c
1294
unsigned long nr_pages = size >> PAGE_SHIFT;
arch/x86/mm/init_64.c
1403
set_memory_ro(start, (end - start) >> PAGE_SHIFT);
arch/x86/mm/init_64.c
1420
set_memory_nx(text_end, (all_end - text_end) >> PAGE_SHIFT);
arch/x86/mm/init_64.c
1426
set_memory_rw(start, (end-start) >> PAGE_SHIFT);
arch/x86/mm/init_64.c
1429
set_memory_ro(start, (end-start) >> PAGE_SHIFT);
arch/x86/mm/init_64.c
1463
unsigned long boot_mem_end = max_pfn << PAGE_SHIFT;
arch/x86/mm/init_64.c
1523
entry = pfn_pte(__pa(p) >> PAGE_SHIFT,
arch/x86/mm/init_64.c
1637
nr_pmd_pages = (next - addr) >> PAGE_SHIFT;
arch/x86/mm/init_64.c
457
vaddr_end = __START_KERNEL_map + (max_pfn_mapped << PAGE_SHIFT);
arch/x86/mm/init_64.c
508
set_pte_init(pte, pfn_pte(paddr >> PAGE_SHIFT, prot), init);
arch/x86/mm/init_64.c
582
pfn_pmd(paddr >> PAGE_SHIFT, prot_sethuge(prot)),
arch/x86/mm/init_64.c
668
pfn_pud(paddr >> PAGE_SHIFT, prot_sethuge(prot)),
arch/x86/mm/init_64.c
967
unsigned long end = ((start_pfn + nr_pages) << PAGE_SHIFT) - 1;
arch/x86/mm/init_64.c
987
update_end_of_memory_vars(start_pfn << PAGE_SHIFT, nr_pages << PAGE_SHIFT);
arch/x86/mm/init_64.c
995
unsigned long start_pfn = start >> PAGE_SHIFT;
arch/x86/mm/init_64.c
996
unsigned long nr_pages = size >> PAGE_SHIFT;
arch/x86/mm/ioremap.c
48
unsigned long nrpages = size >> PAGE_SHIFT;
arch/x86/mm/ioremap.c
79
start_pfn = (res->start + PAGE_SIZE - 1) >> PAGE_SHIFT;
arch/x86/mm/ioremap.c
80
stop_pfn = (res->end + 1) >> PAGE_SHIFT;
arch/x86/mm/ioremap.c
903
set_pte(pte, pfn_pte(phys >> PAGE_SHIFT, flags));
arch/x86/mm/kaslr.c
113
memory_tb = DIV_ROUND_UP(max_pfn << PAGE_SHIFT, 1UL << TB_SHIFT) +
arch/x86/mm/kaslr.c
130
vmemmap_size = (kaslr_regions[0].size_tb << (TB_SHIFT - PAGE_SHIFT)) *
arch/x86/mm/mem_encrypt_amd.c
279
notify_page_enc_status_changed(pfn, psize >> PAGE_SHIFT, enc);
arch/x86/mm/mem_encrypt_amd.c
309
enc_dec_hypercall(vaddr, npages << PAGE_SHIFT, enc);
arch/x86/mm/mem_encrypt_amd.c
332
d->pa = d->pfn << PAGE_SHIFT;
arch/x86/mm/mem_encrypt_amd.c
550
npages = (vaddr_end - vaddr) >> PAGE_SHIFT;
arch/x86/mm/mem_encrypt_amd.c
60
unsigned long npages = PAGE_ALIGN(sz) >> PAGE_SHIFT;
arch/x86/mm/mmap.c
214
phys_addr_t addr = (phys_addr_t)pfn << PAGE_SHIFT;
arch/x86/mm/mmap.c
46
max <<= PAGE_SHIFT;
arch/x86/mm/mmap.c
74
return (get_random_long() & ((1UL << rndbits) - 1)) << PAGE_SHIFT;
arch/x86/mm/pat/cpa-test.c
141
addr[i] = (unsigned long)__va(pfn << PAGE_SHIFT);
arch/x86/mm/pat/cpa-test.c
58
unsigned long addr = (unsigned long)__va(i << PAGE_SHIFT);
arch/x86/mm/pat/memtype.c
403
unsigned long start_pfn = start >> PAGE_SHIFT;
arch/x86/mm/pat/memtype.c
404
unsigned long end_pfn = (end + PAGE_SIZE - 1) >> PAGE_SHIFT;
arch/x86/mm/pat/memtype.c
414
if (start_pfn < ISA_END_ADDRESS >> PAGE_SHIFT)
arch/x86/mm/pat/memtype.c
415
start_pfn = ISA_END_ADDRESS >> PAGE_SHIFT;
arch/x86/mm/pat/memtype.c
455
for (pfn = (start >> PAGE_SHIFT); pfn < (end >> PAGE_SHIFT); ++pfn) {
arch/x86/mm/pat/memtype.c
473
for (pfn = (start >> PAGE_SHIFT); pfn < (end >> PAGE_SHIFT); ++pfn) {
arch/x86/mm/pat/memtype.c
485
for (pfn = (start >> PAGE_SHIFT); pfn < (end >> PAGE_SHIFT); ++pfn) {
arch/x86/mm/pat/memtype.c
666
page = pfn_to_page(paddr >> PAGE_SHIFT);
arch/x86/mm/pat/memtype.c
772
if (!phys_mem_access_encrypted(pfn << PAGE_SHIFT, size))
arch/x86/mm/pat/memtype.c
818
if (!page_is_ram(base >> PAGE_SHIFT))
arch/x86/mm/pat/memtype.c
910
resource_size_t paddr = (resource_size_t)pfn << PAGE_SHIFT;
arch/x86/mm/pat/memtype.c
932
const resource_size_t paddr = (resource_size_t)pfn << PAGE_SHIFT;
arch/x86/mm/pat/memtype.c
939
const resource_size_t paddr = (resource_size_t)pfn << PAGE_SHIFT;
arch/x86/mm/pat/set_memory.c
1000
numpages = psize >> PAGE_SHIFT;
arch/x86/mm/pat/set_memory.c
1161
pfninc = PMD_SIZE >> PAGE_SHIFT;
arch/x86/mm/pat/set_memory.c
1596
unsigned long pre_end = start + (num_pages << PAGE_SHIFT);
arch/x86/mm/pat/set_memory.c
1600
cur_pages = (pre_end - start) >> PAGE_SHIFT;
arch/x86/mm/pat/set_memory.c
1639
cpa->pfn += PMD_SIZE >> PAGE_SHIFT;
arch/x86/mm/pat/set_memory.c
1640
cur_pages += PMD_SIZE >> PAGE_SHIFT;
arch/x86/mm/pat/set_memory.c
1666
end = start + (cpa->numpages << PAGE_SHIFT);
arch/x86/mm/pat/set_memory.c
1677
cur_pages = (pre_end - start) >> PAGE_SHIFT;
arch/x86/mm/pat/set_memory.c
1712
cpa->pfn += PUD_SIZE >> PAGE_SHIFT;
arch/x86/mm/pat/set_memory.c
1713
cur_pages += PUD_SIZE >> PAGE_SHIFT;
arch/x86/mm/pat/set_memory.c
1789
addr + (cpa->numpages << PAGE_SHIFT));
arch/x86/mm/pat/set_memory.c
1825
PAGE_OFFSET + (max_pfn_mapped << PAGE_SHIFT))) {
arch/x86/mm/pat/set_memory.c
1827
cpa->pfn = __pa(vaddr) >> PAGE_SHIFT;
arch/x86/mm/pat/set_memory.c
1928
unsigned long laddr = (unsigned long)__va(cpa->pfn << PAGE_SHIFT);
arch/x86/mm/pat/set_memory.c
1941
PAGE_OFFSET + (max_pfn_mapped << PAGE_SHIFT)))) {
arch/x86/mm/pat/set_memory.c
1969
unsigned long temp_cpa_vaddr = (cpa->pfn << PAGE_SHIFT) +
arch/x86/mm/pat/set_memory.c
2266
if (arch_is_platform_page(pfn << PAGE_SHIFT))
arch/x86/mm/pat/set_memory.c
2280
decoy_addr = (pfn << PAGE_SHIFT) + (PAGE_OFFSET ^ BIT(63));
arch/x86/mm/pat/set_memory.c
2511
start = page_to_pfn(pages[i]) << PAGE_SHIFT;
arch/x86/mm/pat/set_memory.c
2537
start = page_to_pfn(pages[i]) << PAGE_SHIFT;
arch/x86/mm/pat/set_memory.c
255
return __pa_symbol(_text) >> PAGE_SHIFT;
arch/x86/mm/pat/set_memory.c
2580
start = page_to_pfn(pages[i]) << PAGE_SHIFT;
arch/x86/mm/pat/set_memory.c
261
return __pa_symbol(roundup(_brk_end, PMD_SIZE) - 1) >> PAGE_SHIFT;
arch/x86/mm/pat/set_memory.c
862
phys_addr = (phys_addr_t)pud_pfn(*(pud_t *)pte) << PAGE_SHIFT;
arch/x86/mm/pat/set_memory.c
866
phys_addr = (phys_addr_t)pmd_pfn(*(pmd_t *)pte) << PAGE_SHIFT;
arch/x86/mm/pat/set_memory.c
870
phys_addr = (phys_addr_t)pte_pfn(*pte) << PAGE_SHIFT;
arch/x86/mm/pat/set_memory.c
962
numpages = (lpaddr - address) >> PAGE_SHIFT;
arch/x86/mm/pat/set_memory.c
992
pfn = old_pfn + ((address & (psize - 1)) >> PAGE_SHIFT);
arch/x86/mm/pgtable.c
139
paravirt_alloc_pmd(mm, __pa(pmd) >> PAGE_SHIFT);
arch/x86/mm/pgtable.c
228
paravirt_release_pmd(pgd_val(pgd) >> PAGE_SHIFT);
arch/x86/mm/pgtable.c
30
paravirt_release_pmd(__pa(pmd) >> PAGE_SHIFT);
arch/x86/mm/pgtable.c
44
paravirt_release_pud(__pa(pud) >> PAGE_SHIFT);
arch/x86/mm/pgtable.c
51
paravirt_release_p4d(__pa(p4d) >> PAGE_SHIFT);
arch/x86/mm/pgtable.c
591
__native_set_fixmap(idx, pfn_pte(phys >> PAGE_SHIFT, flags));
arch/x86/mm/pgtable.c
648
(u64)addr >> PAGE_SHIFT,
arch/x86/mm/pgtable.c
680
(u64)addr >> PAGE_SHIFT,
arch/x86/mm/physaddr.c
54
return pfn_valid(x >> PAGE_SHIFT);
arch/x86/mm/physaddr.c
69
VIRTUAL_BUG_ON((phys_addr >> PAGE_SHIFT) > max_low_pfn);
arch/x86/mm/physaddr.c
85
return pfn_valid((x - PAGE_OFFSET) >> PAGE_SHIFT);
arch/x86/mm/pti.c
471
*target_pte = pfn_pte(pa >> PAGE_SHIFT, PAGE_KERNEL);
arch/x86/mm/pti.c
595
set_memory_global(start, (end_global - start) >> PAGE_SHIFT);
arch/x86/mm/pti.c
614
set_memory_nonglobal(start, (end - start) >> PAGE_SHIFT);
arch/x86/mm/tlb.c
1339
(info->end - info->start) >> PAGE_SHIFT);
arch/x86/mm/tlb.c
1488
for (addr = info->start; addr < info->end; addr += nr << PAGE_SHIFT) {
arch/x86/mm/tlb.c
1489
nr = (info->end - addr) >> PAGE_SHIFT;
arch/x86/mm/tlb.c
1534
info = get_flush_tlb_info(NULL, start, end, PAGE_SHIFT, false,
arch/x86/pci/pcbios.c
61
set_memory_x(PAGE_OFFSET + BIOS_BEGIN, (BIOS_END - BIOS_BEGIN) >> PAGE_SHIFT);
arch/x86/platform/efi/efi.c
755
num_pages >>= PAGE_SHIFT;
arch/x86/platform/efi/efi_32.c
43
size = md->num_pages << PAGE_SHIFT;
arch/x86/platform/efi/efi_64.c
193
pfn = pa_memmap >> PAGE_SHIFT;
arch/x86/platform/efi/efi_64.c
242
npages = (_etext - _text) >> PAGE_SHIFT;
arch/x86/platform/efi/efi_64.c
250
npages = (__end_rodata - __start_rodata) >> PAGE_SHIFT;
arch/x86/platform/efi/efi_64.c
252
pfn = rodata >> PAGE_SHIFT;
arch/x86/platform/efi/efi_64.c
261
pfn = tramp >> PAGE_SHIFT;
arch/x86/platform/efi/efi_64.c
301
pfn = md->phys_addr >> PAGE_SHIFT;
arch/x86/platform/efi/efi_64.c
309
unsigned long size = md->num_pages << PAGE_SHIFT;
arch/x86/platform/efi/efi_64.c
378
pfn = md->phys_addr >> PAGE_SHIFT;
arch/x86/power/hibernate.c
52
nosave_begin_pfn = __pa_symbol(&__nosave_begin) >> PAGE_SHIFT;
arch/x86/power/hibernate.c
53
nosave_end_pfn = PAGE_ALIGN(__pa_symbol(&__nosave_end)) >> PAGE_SHIFT;
arch/x86/power/hibernate_64.c
113
mstart = pfn_mapped[i].start << PAGE_SHIFT;
arch/x86/power/hibernate_64.c
114
mend = pfn_mapped[i].end << PAGE_SHIFT;
arch/x86/realmode/init.c
115
set_memory_decrypted((unsigned long)base, size >> PAGE_SHIFT);
arch/x86/realmode/init.c
204
set_memory_nx((unsigned long) base, size >> PAGE_SHIFT);
arch/x86/realmode/init.c
205
set_memory_ro((unsigned long) base, ro_size >> PAGE_SHIFT);
arch/x86/realmode/init.c
206
set_memory_x((unsigned long) text_start, text_size >> PAGE_SHIFT);
arch/x86/um/syscalls_64.c
63
return ksys_mmap_pgoff(addr, len, prot, flags, fd, off >> PAGE_SHIFT);
arch/x86/virt/svm/sev.c
123
#define PFN_PMD_MASK GENMASK_ULL(63, PMD_SHIFT - PAGE_SHIFT)
arch/x86/virt/svm/sev.c
420
ram_pa_max = max_pfn << PAGE_SHIFT;
arch/x86/virt/svm/sev.c
695
paddr = pfn << PAGE_SHIFT;
arch/x86/virt/svm/sev.c
726
: "a" (pfn << PAGE_SHIFT), "c" (e)
arch/x86/virt/svm/sev.c
743
e->gpa = e_raw->gpa << PAGE_SHIFT;
arch/x86/virt/svm/sev.c
880
unsigned long paddr = pfn << PAGE_SHIFT;
arch/x86/virt/svm/sev.c
989
unsigned long paddr = pfn << PAGE_SHIFT;
arch/x86/virt/vmx/tdx/tdx.c
455
nr_pamt_entries = tdmr->size >> PAGE_SHIFT;
arch/x86/virt/vmx/tdx/tdx.c
538
pamt = alloc_contig_pages(tdmr_pamt_size >> PAGE_SHIFT, GFP_KERNEL,
arch/x86/virt/vmx/tdx/tdx.c
547
tdmr_pamt_base = page_to_pfn(pamt) << PAGE_SHIFT;
arch/x86/virt/vmx/tdx/tdx.c
600
free_contig_range(pamt_base >> PAGE_SHIFT, pamt_size >> PAGE_SHIFT);
arch/x86/xen/enlighten_pv.c
202
var[reg].base_lo |= op.u.read_memtype.mfn << PAGE_SHIFT;
arch/x86/xen/enlighten_pv.c
203
var[reg].base_hi = op.u.read_memtype.mfn >> (32 - PAGE_SHIFT);
arch/x86/xen/enlighten_pv.c
204
mask = ~((op.u.read_memtype.nr_mfns << PAGE_SHIFT) - 1);
arch/x86/xen/mmu.c
36
return XMADDR(((phys_addr_t)pte_mfn(*pte) << PAGE_SHIFT) + offset);
arch/x86/xen/mmu_pv.c
1128
pa = pte_pfn(pte_tbl[i]) << PAGE_SHIFT;
arch/x86/xen/mmu_pv.c
1524
pfn = (pte & PTE_PFN_MASK) >> PAGE_SHIFT;
arch/x86/xen/mmu_pv.c
1584
unsigned long addr = (unsigned long)__va(pfn << PAGE_SHIFT);
arch/x86/xen/mmu_pv.c
1688
paddr = mfn_to_pfn(maddr >> PAGE_SHIFT) << PAGE_SHIFT;
arch/x86/xen/mmu_pv.c
1703
unsigned long pfn = __pa(addr) >> PAGE_SHIFT;
arch/x86/xen/mmu_pv.c
1944
pa = pte_pfn(pte) << PAGE_SHIFT;
arch/x86/xen/mmu_pv.c
1965
n_pte = roundup(size, PAGE_SIZE) >> PAGE_SHIFT;
arch/x86/xen/mmu_pv.c
2095
phys >>= PAGE_SHIFT;
arch/x86/xen/mmu_pv.c
2507
range = (unsigned long)batch << PAGE_SHIFT;
arch/x86/xen/mmu_pv.c
382
unsigned long mfn = (val & XEN_PTE_MFN_MASK) >> PAGE_SHIFT;
arch/x86/xen/mmu_pv.c
389
val = ((pteval_t)pfn << PAGE_SHIFT) | flags;
arch/x86/xen/mmu_pv.c
398
unsigned long pfn = (val & PTE_PFN_MASK) >> PAGE_SHIFT;
arch/x86/xen/mmu_pv.c
415
val = ((pteval_t)mfn << PAGE_SHIFT) | flags;
arch/x86/xen/p2m.c
336
paravirt_alloc_pte(&init_mm, __pa(p2m_missing_pte) >> PAGE_SHIFT);
arch/x86/xen/p2m.c
338
paravirt_alloc_pte(&init_mm, __pa(p2m_identity_pte) >> PAGE_SHIFT);
arch/x86/xen/p2m.c
483
paravirt_alloc_pte(&init_mm, __pa(pte_newpg[i]) >> PAGE_SHIFT);
arch/x86/xen/p2m.c
504
paravirt_release_pte(__pa(pte_newpg[i]) >> PAGE_SHIFT);
arch/x86/xen/setup.c
251
.ptr = ((uint64_t)mfn << PAGE_SHIFT) | MMU_MACHPHYS_UPDATE,
arch/x86/xen/setup.c
269
if (HYPERVISOR_update_va_mapping((unsigned long)__va(pfn << PAGE_SHIFT),
arch/x86/xen/setup.c
392
(unsigned long)__va(pfn << PAGE_SHIFT),
arch/x86/xen/setup.c
745
if (dest_len > (NR_FIX_BTMAPS << PAGE_SHIFT) - dest_off)
arch/x86/xen/setup.c
746
dest_len = (NR_FIX_BTMAPS << PAGE_SHIFT) - dest_off;
arch/x86/xen/setup.c
748
if (src_len > (NR_FIX_BTMAPS << PAGE_SHIFT) - src_off)
arch/x86/xen/setup.c
749
src_len = (NR_FIX_BTMAPS << PAGE_SHIFT) - src_off;
arch/xtensa/include/asm/fixmap.h
30
#define FIXADDR_SIZE (__end_of_fixed_addresses << PAGE_SHIFT)
arch/xtensa/include/asm/highmem.h
25
#define PKMAP_NR(virt) (((virt) - PKMAP_BASE) >> PAGE_SHIFT)
arch/xtensa/include/asm/highmem.h
26
#define PKMAP_ADDR(nr) (PKMAP_BASE + ((nr) << PAGE_SHIFT))
arch/xtensa/include/asm/page.h
139
#define ARCH_PFN_OFFSET (PHYS_OFFSET >> PAGE_SHIFT)
arch/xtensa/include/asm/page.h
170
#define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
arch/xtensa/include/asm/page.h
171
#define page_to_virt(page) __va(page_to_pfn(page) << PAGE_SHIFT)
arch/xtensa/include/asm/page.h
172
#define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
arch/xtensa/include/asm/page.h
63
# define DCACHE_ALIAS_ORDER (DCACHE_WAY_SHIFT - PAGE_SHIFT)
arch/xtensa/include/asm/page.h
65
# define DCACHE_ALIAS(a) (((a) & DCACHE_ALIAS_MASK) >> PAGE_SHIFT)
arch/xtensa/include/asm/page.h
74
# define ICACHE_ALIAS_ORDER (ICACHE_WAY_SHIFT - PAGE_SHIFT)
arch/xtensa/include/asm/page.h
76
# define ICACHE_ALIAS(a) (((a) & ICACHE_ALIAS_MASK) >> PAGE_SHIFT)
arch/xtensa/include/asm/pgtable.h
224
#define pmd_pfn(pmd) (__pa(pmd_val(pmd)) >> PAGE_SHIFT)
arch/xtensa/include/asm/pgtable.h
267
#define PFN_PTE_SHIFT PAGE_SHIFT
arch/xtensa/include/asm/pgtable.h
268
#define pte_pfn(pte) (pte_val(pte) >> PAGE_SHIFT)
arch/xtensa/include/asm/pgtable.h
271
#define pfn_pte(pfn, prot) __pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot))
arch/xtensa/include/asm/pgtable.h
380
#define _PTE_INDEX(rt,rs) extui rt, rs, PAGE_SHIFT, PTRS_PER_PTE_SHIFT
arch/xtensa/include/asm/pgtable.h
387
srli pmd, pmd, PAGE_SHIFT; \
arch/xtensa/include/asm/pgtable.h
388
slli pmd, pmd, PAGE_SHIFT; \
arch/xtensa/include/asm/thread_info.h
143
#define THREAD_SIZE_ORDER (KERNEL_STACK_SHIFT - PAGE_SHIFT)
arch/xtensa/kernel/pci.c
88
vma->vm_pgoff += (ioaddr + pci_ctrl->io_space.start) >> PAGE_SHIFT;
arch/xtensa/kernel/syscall.c
37
(((pgoff) << PAGE_SHIFT) & (SHMLBA - 1)))
arch/xtensa/kernel/syscall.c
69
((addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1)))
arch/xtensa/mm/highmem.c
40
return kmap_idx(type, DCACHE_ALIAS(pfn << PAGE_SHIFT));
arch/xtensa/mm/init.c
60
early_memtest((phys_addr_t)min_low_pfn << PAGE_SHIFT,
arch/xtensa/mm/init.c
61
(phys_addr_t)max_low_pfn << PAGE_SHIFT);
arch/xtensa/mm/tlb.c
103
if (end-start + (PAGE_SIZE-1) <= _TLB_ENTRIES << PAGE_SHIFT) {
arch/xtensa/mm/tlb.c
154
end - start < _TLB_ENTRIES << PAGE_SHIFT) {
arch/xtensa/mm/tlb.c
234
unsigned tlbidx = w | (e << PAGE_SHIFT);
arch/xtensa/mm/tlb.c
239
unsigned vpn = (r0 & PAGE_MASK) | (e << PAGE_SHIFT);
arch/xtensa/mm/tlb.c
258
struct page *p = pfn_to_page(r1 >> PAGE_SHIFT);
arch/xtensa/mm/tlb.c
31
int e = w + (i << PAGE_SHIFT);
arch/xtensa/mm/tlb.c
44
int e = w + (i << PAGE_SHIFT);
block/bdev.c
136
lstart >> PAGE_SHIFT,
block/bdev.c
137
lend >> PAGE_SHIFT);
block/blk-map.c
370
unsigned long end = (kaddr + len + PAGE_SIZE - 1) >> PAGE_SHIFT;
block/blk-map.c
371
unsigned long start = kaddr >> PAGE_SHIFT;
block/blk-settings.c
83
io_opt * 2 >> PAGE_SHIFT,
block/blk-sysfs.c
181
ret = queue_var_show(disk->bdi->ra_pages << (PAGE_SHIFT - 10), page);
block/blk-sysfs.c
206
WRITE_ONCE(disk->bdi->ra_pages, ra_kb >> (PAGE_SHIFT - 10));
crypto/ahash.c
229
page += offset >> PAGE_SHIFT;
crypto/ahash.c
91
walk->pg = sg_page(walk->sg) + (walk->offset >> PAGE_SHIFT);
crypto/skcipher.c
196
diff |= (u8 *)(sg_page(walk->in.sg) + (walk->in.offset >> PAGE_SHIFT)) -
crypto/skcipher.c
197
(u8 *)(sg_page(walk->out.sg) + (walk->out.offset >> PAGE_SHIFT));
drivers/accel/amdxdna/amdxdna_gem.c
1047
drm_clflush_pages(abo->base.pages, gobj->size >> PAGE_SHIFT);
drivers/accel/amdxdna/amdxdna_gem.c
304
nr_pages = (PAGE_ALIGN(addr + len) - (addr & PAGE_MASK)) >> PAGE_SHIFT;
drivers/accel/amdxdna/amdxdna_gem.c
57
align = 1 << max(PAGE_SHIFT, xdna->dev_info->dev_mem_buf_shift);
drivers/accel/amdxdna/amdxdna_ubuf.c
177
ubuf->nr_pages = exp_info.size >> PAGE_SHIFT;
drivers/accel/amdxdna/amdxdna_ubuf.c
178
lock_limit = rlimit(RLIMIT_MEMLOCK) >> PAGE_SHIFT;
drivers/accel/amdxdna/amdxdna_ubuf.c
194
npages = va_ent[i].len >> PAGE_SHIFT;
drivers/accel/amdxdna/amdxdna_ubuf.c
35
ubuf->nr_pages << PAGE_SHIFT, GFP_KERNEL);
drivers/accel/amdxdna/amdxdna_ubuf.c
80
pgoff = (vmf->address - vma->vm_start) >> PAGE_SHIFT;
drivers/accel/habanalabs/common/habanalabs.h
3706
(PAGE_SIZE - 1)) >> PAGE_SHIFT;
drivers/accel/habanalabs/common/habanalabs.h
51
#define HL_MMAP_TYPE_SHIFT (59 - PAGE_SHIFT)
drivers/accel/habanalabs/common/habanalabs.h
57
#define HL_MMAP_OFFSET_VALUE_MASK (0x1FFFFFFFFFFFull >> PAGE_SHIFT)
drivers/accel/habanalabs/common/hldio.c
220
PAGE_SHIFT];
drivers/accel/habanalabs/common/hldio.c
243
npages = (io->len_bytes >> PAGE_SHIFT);
drivers/accel/habanalabs/common/hldio.c
389
p2pr->p2ppages = vmalloc((p2pr->size >> PAGE_SHIFT) * sizeof(struct page *));
drivers/accel/habanalabs/common/hldio.c
395
for (i = 0, addr = p2pr->p2pmem ; i < (p2pr->size >> PAGE_SHIFT) ; ++i, addr += PAGE_SIZE) {
drivers/accel/habanalabs/common/memory.c
1422
*handle <<= PAGE_SHIFT;
drivers/accel/habanalabs/common/memory.c
2405
npages = (end - start) >> PAGE_SHIFT;
drivers/accel/habanalabs/common/memory_mgr.c
118
buf = idr_find(&mmg->handles, lower_32_bits(handle >> PAGE_SHIFT));
drivers/accel/habanalabs/common/memory_mgr.c
171
buf->handle = (((u64)rc | buf->behavior->mem_id) << PAGE_SHIFT);
drivers/accel/habanalabs/common/memory_mgr.c
185
idr_remove(&mmg->handles, lower_32_bits(buf->handle >> PAGE_SHIFT));
drivers/accel/habanalabs/common/memory_mgr.c
241
handle = vma->vm_pgoff << PAGE_SHIFT;
drivers/accel/habanalabs/common/memory_mgr.c
25
buf = idr_find(&mmg->handles, lower_32_bits(handle >> PAGE_SHIFT));
drivers/accel/habanalabs/common/memory_mgr.c
292
vma->vm_pgoff = handle >> PAGE_SHIFT;
drivers/accel/habanalabs/common/memory_mgr.c
67
idr_remove(&buf->mmg->handles, lower_32_bits(buf->handle >> PAGE_SHIFT));
drivers/accel/habanalabs/common/memory_mgr.c
86
idr_remove(&buf->mmg->handles, lower_32_bits(buf->handle >> PAGE_SHIFT));
drivers/accel/habanalabs/common/mmu/mmu.c
840
hr_priv->mmu_pgt_pool = gen_pool_create(PAGE_SHIFT, -1);
drivers/accel/habanalabs/gaudi/gaudi.c
4186
virt_to_phys(cpu_addr) >> PAGE_SHIFT,
drivers/accel/habanalabs/gaudi2/gaudi2.c
11631
rc = remap_pfn_range(vma, vma->vm_start, address >> PAGE_SHIFT,
drivers/accel/habanalabs/gaudi2/gaudi2.c
6856
virt_to_phys(cpu_addr) >> PAGE_SHIFT,
drivers/accel/ivpu/ivpu_gem_userptr.c
72
unsigned long nr_pages = size >> PAGE_SHIFT;
drivers/acpi/apei/einj-core.c
766
if (is_zero_pfn(base_addr >> PAGE_SHIFT))
drivers/acpi/osl.c
296
pfn = pg_off >> PAGE_SHIFT;
drivers/acpi/osl.c
309
pfn = pg_off >> PAGE_SHIFT;
drivers/acpi/tables.c
419
#define MAP_CHUNK_SIZE (NR_FIX_BTMAPS << PAGE_SHIFT)
drivers/android/binder/rust_binderfs.c
626
sb->s_blocksize_bits = PAGE_SHIFT;
drivers/android/binder_alloc.c
819
size_t index = buffer_space_offset >> PAGE_SHIFT;
drivers/android/binderfs.c
622
sb->s_blocksize_bits = PAGE_SHIFT;
drivers/ata/libata-sff.c
617
page += offset >> PAGE_SHIFT;
drivers/ata/libata-sff.c
754
page += offset >> PAGE_SHIFT;
drivers/atm/lanai.c
381
int order = get_order(lanai_buf_size(buf)) + (PAGE_SHIFT - 10);
drivers/base/firmware_loader/sysfs.c
250
int page_nr = offset >> PAGE_SHIFT;
drivers/base/firmware_loader/sysfs.c
306
PAGE_ALIGN(min_size) >> PAGE_SHIFT);
drivers/base/memory.c
574
if (phys_addr & ((pages_per_block << PAGE_SHIFT) - 1))
drivers/base/memory.c
614
pfn >>= PAGE_SHIFT;
drivers/base/memory.c
630
pfn >>= PAGE_SHIFT;
drivers/base/node.c
440
#define K(x) ((x) << (PAGE_SHIFT - 10))
drivers/block/drbd/drbd_actlog.c
280
#if (PAGE_SHIFT + 3) < (AL_EXTENT_SHIFT - BM_BLOCK_SHIFT)
drivers/block/drbd/drbd_actlog.c
293
((PAGE_SHIFT + 3) -
drivers/block/drbd/drbd_bitmap.c
1003
on_disk_sector = first_bm_sect + (((sector_t)page_nr) << (PAGE_SHIFT-SECTOR_SHIFT));
drivers/block/drbd/drbd_bitmap.c
1536
first_page = sl >> (3 + PAGE_SHIFT);
drivers/block/drbd/drbd_bitmap.c
1537
last_page = el >> (3 + PAGE_SHIFT);
drivers/block/drbd/drbd_bitmap.c
1642
c += (0 != test_bit_le(bitnr - (page_nr << (PAGE_SHIFT+3)), p_addr));
drivers/block/drbd/drbd_bitmap.c
299
unsigned int page_nr = long_nr >> (PAGE_SHIFT - LN2_BPL + 3);
drivers/block/drbd/drbd_bitmap.c
307
unsigned int page_nr = bitnr >> (PAGE_SHIFT + 3);
drivers/block/drbd/drbd_bitmap.c
474
#define BITS_PER_PAGE (1UL << (PAGE_SHIFT + 3))
drivers/block/drbd/drbd_bitmap.c
477
# if BITS_PER_PAGE != (1UL << (PAGE_SHIFT + 3))
drivers/block/drbd/drbd_int.h
1262
#if DRBD_MAX_BIO_SIZE > (BIO_MAX_VECS << PAGE_SHIFT)
drivers/block/floppy.c
3080
#define MAX_LEN (1UL << MAX_PAGE_ORDER << PAGE_SHIFT)
drivers/block/null_blk/main.c
1127
flushed += one_round << PAGE_SHIFT;
drivers/block/ublk_drv.c
2616
unsigned long pfn, end, phys_off = vma->vm_pgoff << PAGE_SHIFT;
drivers/block/ublk_drv.c
2644
pfn = virt_to_phys(ublk_queue_cmd_buf(ub, q_id)) >> PAGE_SHIFT;
drivers/block/ublk_drv.c
5315
range->base_offset = start << PAGE_SHIFT;
drivers/block/ublk_drv.c
5368
nr_pages = buf_reg.len >> PAGE_SHIFT;
drivers/block/ublk_drv.c
5516
((bv.bv_offset + bv.bv_len - 1) >> PAGE_SHIFT);
drivers/block/ublk_drv.c
900
if (p->logical_bs_shift > PAGE_SHIFT || p->logical_bs_shift < 9)
drivers/block/virtio_blk.c
542
bufsize = min_t(size_t, bufsize, queue_max_segments(q) << PAGE_SHIFT);
drivers/block/zram/zram_drv.c
1240
u64 nr_pages = zram->disksize >> PAGE_SHIFT;
drivers/block/zram/zram_drv.c
1545
unsigned long nr_pages = zram->disksize >> PAGE_SHIFT;
drivers/block/zram/zram_drv.c
1932
orig_size << PAGE_SHIFT,
drivers/block/zram/zram_drv.c
1934
mem_used << PAGE_SHIFT,
drivers/block/zram/zram_drv.c
1935
zram->limit_pages << PAGE_SHIFT,
drivers/block/zram/zram_drv.c
1936
max_used << PAGE_SHIFT,
drivers/block/zram/zram_drv.c
1963
size_t num_pages = disksize >> PAGE_SHIFT;
drivers/block/zram/zram_drv.c
1982
num_pages = disksize >> PAGE_SHIFT;
drivers/block/zram/zram_drv.c
2372
unsigned long nr_pages = zram->disksize >> PAGE_SHIFT;
drivers/block/zram/zram_drv.c
395
zram->limit_pages = PAGE_ALIGN(limit) >> PAGE_SHIFT;
drivers/block/zram/zram_drv.c
428
unsigned long nr_pages = zram->disksize >> PAGE_SHIFT;
drivers/block/zram/zram_drv.c
535
#define FOUR_K(x) ((x) * (1 << (PAGE_SHIFT - 12)))
drivers/block/zram/zram_drv.c
768
nr_pages = i_size_read(inode) >> PAGE_SHIFT;
drivers/block/zram/zram_drv.c
907
zram->bd_wb_limit += 1UL << (PAGE_SHIFT - 12);
drivers/block/zram/zram_drv.c
915
zram->bd_wb_limit -= 1UL << (PAGE_SHIFT - 12);
drivers/block/zram/zram_drv.h
23
#define SECTORS_PER_PAGE_SHIFT (PAGE_SHIFT - SECTOR_SHIFT)
drivers/block/zram/zram_drv.h
39
#define ZRAM_FLAG_SHIFT (PAGE_SHIFT + 1)
drivers/cdx/cdx.c
724
size = ((cdx_resource_len(cdx_dev, num) - 1) >> PAGE_SHIFT) + 1;
drivers/cdx/cdx.c
733
vma->vm_pgoff += (cdx_resource_start(cdx_dev, num) >> PAGE_SHIFT);
drivers/char/agp/backend.c
119
#if PAGE_SHIFT < 20
drivers/char/agp/backend.c
120
memory = totalram_pages() >> (20 - PAGE_SHIFT);
drivers/char/agp/backend.c
122
memory = totalram_pages() << (PAGE_SHIFT - 20);
drivers/char/agp/backend.c
134
result = result << (20 - PAGE_SHIFT);
drivers/char/agp/generic.c
347
num_entries -= agp_memory_reserved>>PAGE_SHIFT;
drivers/char/agp/intel-gtt.c
1490
*gtt_total = intel_private.gtt_total_entries << PAGE_SHIFT;
drivers/char/agp/intel-gtt.c
1492
*mappable_end = intel_private.gtt_mappable_entries << PAGE_SHIFT;
drivers/char/agp/intel-gtt.c
227
dma_addr_t addr = i << PAGE_SHIFT;
drivers/char/agp/intel-gtt.c
562
return aperture_size >> PAGE_SHIFT;
drivers/char/agp/intel-gtt.c
694
aper_size = (intel_private.gtt_mappable_entries << PAGE_SHIFT) / MB(1);
drivers/char/agp/intel-gtt.c
895
len = sg_dma_len(sg) >> PAGE_SHIFT;
drivers/char/agp/intel-gtt.c
897
dma_addr_t addr = sg_dma_address(sg) + (m << PAGE_SHIFT);
drivers/char/agp/parisc-agp.c
204
pa |= (ci >> PAGE_SHIFT) & 0xff;/* move CI (8 bits) into lowest byte */
drivers/char/agp/parisc-agp.c
286
info->gatt = &io_pdir[(PLUTO_IOVA_SIZE/2) >> PAGE_SHIFT];
drivers/char/agp/uninorth-agp.c
187
gp[i] = (page_to_phys(mem->pages[i]) >> PAGE_SHIFT) | 0x80000000UL;
drivers/char/agp/uninorth-agp.c
431
scratch_value = (page_to_phys(agp_bridge->scratch_page_page) >> PAGE_SHIFT) | 0x80000000UL;
drivers/char/bsr.c
128
dev->bsr_addr >> PAGE_SHIFT,
drivers/char/mem.c
123
allowed = page_is_allowed(p >> PAGE_SHIFT);
drivers/char/mem.c
202
allowed = page_is_allowed(p >> PAGE_SHIFT);
drivers/char/mem.c
273
phys_addr_t offset = pfn << PAGE_SHIFT;
drivers/char/mem.c
291
return pgoff << PAGE_SHIFT;
drivers/char/mem.c
334
const phys_addr_t offset = (phys_addr_t)desc->pgoff << PAGE_SHIFT;
drivers/char/mem.c
337
if (offset >> PAGE_SHIFT != desc->pgoff)
drivers/char/uv_mmtimer.c
166
if (remap_pfn_range(vma, vma->vm_start, uv_mmtimer_addr >> PAGE_SHIFT,
drivers/char/xillybus/xillybus_core.c
1850
int idtbuffersize = (1 << PAGE_SHIFT);
drivers/char/xillybus/xillybus_core.c
1860
unsigned char bogus_idt[8] = { 1, 224, (PAGE_SHIFT)-2, 0,
drivers/char/xillybus/xillybus_core.c
1861
3, 192, PAGE_SHIFT, 0 };
drivers/char/xillybus/xillyusb.c
2282
if (LOG2_INITIAL_FIFO_BUF_SIZE > PAGE_SHIFT)
drivers/char/xillybus/xillyusb.c
2283
fifo_buf_order = LOG2_INITIAL_FIFO_BUF_SIZE - PAGE_SHIFT;
drivers/char/xillybus/xillyusb.c
379
log2_fifo_buf_size = fifo_buf_order + PAGE_SHIFT;
drivers/char/xillybus/xillyusb.c
387
buf_order = (log2_size > PAGE_SHIFT) ?
drivers/char/xillybus/xillyusb.c
388
log2_size - PAGE_SHIFT : 0;
drivers/char/xillybus/xillyusb.c
513
ep->buffer_size = 1 << (PAGE_SHIFT + order);
drivers/comedi/comedi_buf.c
139
async->prealloc_bufsz = n_pages << PAGE_SHIFT;
drivers/comedi/comedi_buf.c
160
unsigned long pg = offset >> PAGE_SHIFT;
drivers/comedi/comedi_buf.c
228
unsigned int n_pages = new_size >> PAGE_SHIFT;
drivers/comedi/comedi_buf.c
344
unsigned int page = async->munge_ptr >> PAGE_SHIFT;
drivers/comedi/comedi_buf.c
581
unsigned int page = write_ptr >> PAGE_SHIFT;
drivers/comedi/comedi_buf.c
608
unsigned int page = read_ptr >> PAGE_SHIFT;
drivers/comedi/comedi_fops.c
2465
addr - vma->vm_start + (vma->vm_pgoff << PAGE_SHIFT);
drivers/comedi/comedi_fops.c
2655
unsigned int page = src_offset >> PAGE_SHIFT;
drivers/comedi/comedi_fops.c
2683
unsigned int page = dst_offset >> PAGE_SHIFT;
drivers/comedi/drivers/mite.c
655
unsigned int n_full_links = nbytes >> PAGE_SHIFT;
drivers/comedi/drivers/mite.c
727
n_links = async->prealloc_bufsz >> PAGE_SHIFT;
drivers/comedi/drivers/mite.c
740
return mite_init_ring_descriptors(ring, s, n_links << PAGE_SHIFT);
drivers/crypto/ccp/sev-dev-tio.c
34
u64 pa = pfn << PAGE_SHIFT;
drivers/crypto/ccp/sev-dev-tio.c
389
unsigned int npages = PAGE_ALIGN(len) >> PAGE_SHIFT;
drivers/crypto/ccp/sev-dev-tio.c
429
unsigned long i, npages = PAGE_ALIGN(len) >> PAGE_SHIFT;
drivers/crypto/ccp/sev-dev-tio.c
46
#define sla_to_pfn(sla) (__pa(sla_to_va(sla)) >> PAGE_SHIFT)
drivers/crypto/ccp/sev-dev-tio.c
470
buf->capacity_sz = (npages << PAGE_SHIFT);
drivers/crypto/ccp/sev-dev-tio.c
53
.pfn = pa >> PAGE_SHIFT,
drivers/crypto/ccp/sev-dev.c
1213
range->base = page_to_pfn(entry->page) << PAGE_SHIFT;
drivers/crypto/ccp/sev-dev.c
1344
range->page_count = size >> PAGE_SHIFT;
drivers/crypto/ccp/sev-dev.c
422
unsigned long pfn = __sme_clr(paddr) >> PAGE_SHIFT;
drivers/crypto/ccp/sev-dev.c
701
npages = PAGE_ALIGN(desc->len) >> PAGE_SHIFT;
drivers/crypto/ccp/sev-dev.c
719
npages = PAGE_ALIGN(desc->len) >> PAGE_SHIFT;
drivers/crypto/hisilicon/qm.c
2579
phys_base >> PAGE_SHIFT,
drivers/crypto/hisilicon/qm.c
2908
PAGE_SHIFT;
drivers/crypto/hisilicon/sgl.c
76
block_size = 1 << (PAGE_SHIFT + MAX_PAGE_ORDER < 32 ?
drivers/crypto/hisilicon/sgl.c
77
PAGE_SHIFT + MAX_PAGE_ORDER : 31);
drivers/cxl/core/mce.c
29
pfn = spa >> PAGE_SHIFT;
drivers/cxl/core/mce.c
37
pfn = spa_alias >> PAGE_SHIFT;
drivers/dax/device.c
249
vmf->address & ~((1UL << (order + PAGE_SHIFT)) - 1), order);
drivers/dax/device.c
332
off = pgoff << PAGE_SHIFT;
drivers/dax/device.c
457
order_base_2(dev_dax->align >> PAGE_SHIFT);
drivers/dma-buf/dma-buf.c
1595
dmabuf->size >> PAGE_SHIFT)
drivers/dma-buf/dma-buf.c
243
dmabuf->size >> PAGE_SHIFT)
drivers/dma-buf/heaps/cma_heap.c
304
pgoff_t pagecount = size >> PAGE_SHIFT;
drivers/dma-buf/heaps/cma_heap.c
83
buffer->pagecount << PAGE_SHIFT,
drivers/dma-buf/heaps/system_heap.c
194
unsigned long n = sg->length >> PAGE_SHIFT;
drivers/dma-buf/heaps/system_heap.c
202
unsigned long n = (sg->length >> PAGE_SHIFT) - pgoff;
drivers/dma-buf/heaps/system_heap.c
204
unsigned long size = n << PAGE_SHIFT;
drivers/dma-buf/udmabuf.c
124
ubuf->offsets[pg] >> PAGE_SHIFT);
drivers/dma-buf/udmabuf.c
316
exp_info.size = ubuf->pagecount << PAGE_SHIFT;
drivers/dma-buf/udmabuf.c
333
pgcnt = size >> PAGE_SHIFT;
drivers/dma-buf/udmabuf.c
334
end = start + (pgcnt << PAGE_SHIFT) - 1;
drivers/dma-buf/udmabuf.c
385
pglimit = ((u64)size_limit_mb * 1024 * 1024) >> PAGE_SHIFT;
drivers/dma-buf/udmabuf.c
394
subpgcnt = list[i].size >> PAGE_SHIFT;
drivers/dma-buf/udmabuf.c
59
pfn += ubuf->offsets[pgoff] >> PAGE_SHIFT;
drivers/dma-buf/udmabuf.c
77
pfn += ubuf->offsets[pgoff] >> PAGE_SHIFT;
drivers/dma/idxd/cdev.c
417
IDXD_PORTAL_LIMITED)) >> PAGE_SHIFT;
drivers/dma/idxd/idxd.h
628
return ((wq_id * 4) << PAGE_SHIFT) + idxd_get_wq_portal_offset(prot);
drivers/edac/altera_edac.c
106
err_addr >> PAGE_SHIFT,
drivers/edac/altera_edac.c
386
dimm->nr_pages = ((mem_size - 1) >> PAGE_SHIFT) + 1;
drivers/edac/amd64_edac.c
3057
nr_pages <<= 20 - PAGE_SHIFT;
drivers/edac/amd64_edac.c
3074
nr_pages <<= 20 - PAGE_SHIFT;
drivers/edac/amd64_edac.c
3572
nr_pages <<= 20 - PAGE_SHIFT;
drivers/edac/amd64_edac.c
956
err->page = (u32) (error_address >> PAGE_SHIFT);
drivers/edac/amd76x_edac.c
208
csrow->first_page = mba_base >> PAGE_SHIFT;
drivers/edac/amd76x_edac.c
209
dimm->nr_pages = (mba_mask + 1) >> PAGE_SHIFT;
drivers/edac/amd76x_edac.c
211
csrow->page_mask = mba_mask >> PAGE_SHIFT;
drivers/edac/amd76x_edac.c
212
dimm->grain = dimm->nr_pages << PAGE_SHIFT;
drivers/edac/armada_xp_edac.c
205
addr_val >> PAGE_SHIFT,
drivers/edac/armada_xp_edac.c
213
addr_val >> PAGE_SHIFT,
drivers/edac/aspeed_edac.c
114
page = rec_addr >> PAGE_SHIFT;
drivers/edac/aspeed_edac.c
135
page = un_rec_addr >> PAGE_SHIFT;
drivers/edac/aspeed_edac.c
258
&r, PAGE_SHIFT);
drivers/edac/aspeed_edac.c
260
csrow->first_page = r.start >> PAGE_SHIFT;
drivers/edac/aspeed_edac.c
261
nr_pages = resource_size(&r) >> PAGE_SHIFT;
drivers/edac/cpc925_edac.c
341
nr_pages = row_size >> PAGE_SHIFT;
drivers/edac/cpc925_edac.c
460
pa = mci->csrows[rank]->first_page << PAGE_SHIFT;
drivers/edac/cpc925_edac.c
499
*pfn = pa >> PAGE_SHIFT;
drivers/edac/dmc520_edac.c
457
pages_per_rank = rs >> PAGE_SHIFT;
drivers/edac/e752x_edac.c
1101
cumul_size = value << (25 + drc_drbg - PAGE_SHIFT);
drivers/edac/e752x_edac.c
339
page = sec1_add >> (PAGE_SHIFT - 4);
drivers/edac/e752x_edac.c
401
block_page = error_2b >> (PAGE_SHIFT - 4);
drivers/edac/e752x_edac.c
420
block_page = error_2b >> (PAGE_SHIFT - 4);
drivers/edac/e752x_edac.c
468
page = error_1b >> (PAGE_SHIFT - 4); /* convert the addr to 4k page */
drivers/edac/e7xxx_edac.c
383
cumul_size = value << (25 + drc_drbg - PAGE_SHIFT);
drivers/edac/edac_mc.c
929
(e->page_frame_number << PAGE_SHIFT) | e->offset_in_page,
drivers/edac/edac_mc.h
40
#if PAGE_SHIFT < 20
drivers/edac/edac_mc.h
41
#define PAGES_TO_MiB(pages) ((pages) >> (20 - PAGE_SHIFT))
drivers/edac/edac_mc.h
42
#define MiB_TO_PAGES(mb) ((mb) << (20 - PAGE_SHIFT))
drivers/edac/edac_mc.h
44
#define PAGES_TO_MiB(pages) ((pages) << (PAGE_SHIFT - 20))
drivers/edac/edac_mc.h
45
#define MiB_TO_PAGES(mb) ((mb) >> (PAGE_SHIFT - 20))
drivers/edac/fsl_ddr_edac.c
327
pfn = err_addr >> PAGE_SHIFT;
drivers/edac/fsl_ddr_edac.c
475
start <<= (24 - PAGE_SHIFT);
drivers/edac/fsl_ddr_edac.c
476
end <<= (24 - PAGE_SHIFT);
drivers/edac/fsl_ddr_edac.c
477
end |= (1 << (24 - PAGE_SHIFT)) - 1;
drivers/edac/highbank_mc_edac.c
225
dimm->nr_pages = (~0UL >> PAGE_SHIFT) + 1;
drivers/edac/highbank_mc_edac.c
66
err_addr >> PAGE_SHIFT,
drivers/edac/highbank_mc_edac.c
76
err_addr >> PAGE_SHIFT,
drivers/edac/i3000_edac.c
395
cumul_size = value << (I3000_DRB_SHIFT - PAGE_SHIFT);
drivers/edac/i3000_edac.c
53
deap >>= PAGE_SHIFT;
drivers/edac/i3000_edac.c
54
deap |= (edeap & 1) << (32 - PAGE_SHIFT);
drivers/edac/i3200_edac.c
331
n <<= (I3200_DRB_SHIFT - PAGE_SHIFT);
drivers/edac/i3200_edac.c
405
dimm->grain = nr_pages << PAGE_SHIFT;
drivers/edac/i7core_edac.c
1771
m->addr >> PAGE_SHIFT,
drivers/edac/i82860_edac.c
116
info->eap >>= PAGE_SHIFT;
drivers/edac/i82860_edac.c
167
(I82860_GBA_SHIFT - PAGE_SHIFT);
drivers/edac/i82875p_edac.c
244
info->eap >>= PAGE_SHIFT;
drivers/edac/i82875p_edac.c
365
cumul_size = value << (I82875P_DRB_SHIFT - PAGE_SHIFT);
drivers/edac/i82975x_edac.c
299
page >>= (PAGE_SHIFT - 1);
drivers/edac/i82975x_edac.c
312
& ((1 << PAGE_SHIFT) -
drivers/edac/i82975x_edac.c
387
cumul_size <<= (I82975X_DRB_SHIFT - PAGE_SHIFT);
drivers/edac/ie31200_edac.c
355
info->erraddr >> PAGE_SHIFT, 0, 0,
drivers/edac/ie31200_edac.c
361
info->erraddr >> PAGE_SHIFT, 0,
drivers/edac/igen6_edac.c
829
res->sys_addr >> PAGE_SHIFT,
drivers/edac/npcm_edac.c
133
edac_mc_handle_error(HW_EVENT_ERR_CORRECTED, mci, 1, addr >> PAGE_SHIFT,
drivers/edac/npcm_edac.c
166
edac_mc_handle_error(HW_EVENT_ERR_UNCORRECTED, mci, 1, addr >> PAGE_SHIFT,
drivers/edac/pasemi_edac.c
145
dimm->nr_pages = 128 << (20 - PAGE_SHIFT);
drivers/edac/pasemi_edac.c
148
dimm->nr_pages = 256 << (20 - PAGE_SHIFT);
drivers/edac/pasemi_edac.c
152
dimm->nr_pages = 512 << (20 - PAGE_SHIFT);
drivers/edac/pasemi_edac.c
155
dimm->nr_pages = 1024 << (20 - PAGE_SHIFT);
drivers/edac/pasemi_edac.c
158
dimm->nr_pages = 2048 << (20 - PAGE_SHIFT);
drivers/edac/pnd2_edac.c
1192
edac_mc_handle_error(tp_event, mci, core_err_cnt, m->addr >> PAGE_SHIFT,
drivers/edac/sb_edac.c
3170
m->addr >> PAGE_SHIFT, m->addr & ~PAGE_MASK, 0,
drivers/edac/sb_edac.c
3232
m->addr >> PAGE_SHIFT, m->addr & ~PAGE_MASK, 0,
drivers/edac/skx_common.c
546
dimm->nr_pages = size >> PAGE_SHIFT;
drivers/edac/skx_common.c
724
m->addr >> PAGE_SHIFT, m->addr & ~PAGE_MASK, 0,
drivers/edac/skx_common.c
766
if (!pfn_to_online_page(res.addr >> PAGE_SHIFT) && !arch_is_platform_page(res.addr)) {
drivers/edac/synopsys_edac.c
857
dimm->nr_pages = (size >> PAGE_SHIFT) / csi->nr_channels;
drivers/edac/ti_edac.c
107
err_addr >> PAGE_SHIFT,
drivers/edac/ti_edac.c
115
err_addr >> PAGE_SHIFT,
drivers/edac/ti_edac.c
181
dimm->nr_pages = memsize >> PAGE_SHIFT;
drivers/edac/versal_edac.c
606
dimm->nr_pages = (size >> PAGE_SHIFT) / csi->nr_channels;
drivers/edac/x38_edac.c
313
n <<= (X38_DRB_SHIFT - PAGE_SHIFT);
drivers/edac/x38_edac.c
389
dimm->grain = nr_pages << PAGE_SHIFT;
drivers/firewire/core-cdev.c
1138
buffer_end = client->buffer.page_count << PAGE_SHIFT;
drivers/firewire/core-cdev.c
1800
page_count = size >> PAGE_SHIFT;
drivers/firewire/core-iso.c
135
return (i << PAGE_SHIFT) + offset;
drivers/firewire/ohci.c
3233
end_page = PAGE_ALIGN(payload_index + p->payload_length) >> PAGE_SHIFT;
drivers/firewire/ohci.c
3235
payload_z = end_page - (payload_index >> PAGE_SHIFT);
drivers/firewire/ohci.c
3280
page = payload_index >> PAGE_SHIFT;
drivers/firewire/ohci.c
3282
next_page_index = (page + 1) << PAGE_SHIFT;
drivers/firewire/ohci.c
3334
page = payload >> PAGE_SHIFT;
drivers/firewire/ohci.c
3403
page = payload >> PAGE_SHIFT;
drivers/firmware/efi/efi-init.c
208
size = npages << PAGE_SHIFT;
drivers/firmware/efi/libstub/x86-stub.c
662
entry->size = d->num_pages << PAGE_SHIFT;
drivers/firmware/qcom/qcom_tzmem.c
269
pool->genpool = gen_pool_create(PAGE_SHIFT, -1);
drivers/fpga/dfl-afu-dma-region.c
104
int npages = region->length >> PAGE_SHIFT;
drivers/fpga/dfl-afu-dma-region.c
37
int npages = region->length >> PAGE_SHIFT;
drivers/fpga/dfl-afu-dma-region.c
85
long npages = region->length >> PAGE_SHIFT;
drivers/fpga/dfl-afu-main.c
822
offset = vma->vm_pgoff << PAGE_SHIFT;
drivers/fpga/dfl-afu-main.c
843
(region.phys + (offset - region.offset)) >> PAGE_SHIFT,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.c
495
mem_info->local_mem_size_public = (ttm_tt_pages_limit() << PAGE_SHIFT);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.c
804
tmp = (ttm_tt_pages_limit() << PAGE_SHIFT) / num_online_nodes();
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.c
811
return (ttm_tt_pages_limit() << PAGE_SHIFT);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
131
kfd_mem_limit.max_ttm_mem_limit = ttm_tt_pages_limit() << PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
583
(u64)ttm->num_pages << PAGE_SHIFT,
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
1043
info.alignment = robj->tbo.page_alignment << PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c
207
adev->gmc.gart_size >> PAGE_SHIFT,
drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c
333
drm_mm_init(&mgr->mm, 0, adev->gmc.gart_size >> PAGE_SHIFT);
drivers/gpu/drm/amd/amdgpu/amdgpu_hmm.c
213
hmm_range->hmm_pfns += MAX_WALK_BYTE >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_job.c
271
job->gds_base = amdgpu_bo_gpu_offset(gds) >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_job.c
272
job->gds_size = amdgpu_bo_size(gds) >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_job.c
275
job->gws_base = amdgpu_bo_gpu_offset(gws) >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_job.c
276
job->gws_size = amdgpu_bo_size(gws) >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_job.c
279
job->oa_base = amdgpu_bo_gpu_offset(oa) >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_job.c
280
job->oa_size = amdgpu_bo_size(oa) >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_object.c
119
unsigned int visible_pfn = adev->gmc.visible_vram_size >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_object.c
1502
offset = (bo->tbo.resource->start << PAGE_SHIFT) + fb_base;
drivers/gpu/drm/amd/amdgpu/amdgpu_object.c
1522
offset = (bo->tbo.resource->start << PAGE_SHIFT) +
drivers/gpu/drm/amd/amdgpu/amdgpu_object.c
468
(*bo_ptr)->placements[i].fpfn = offset >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_object.c
469
(*bo_ptr)->placements[i].lpfn = (offset + size) >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_object.c
645
size <<= PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_object.c
650
size = ALIGN(size, 4) << PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_object.c
653
page_align = ALIGN(bp->byte_align, PAGE_SIZE) >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_object.h
217
return (bo->tbo.page_alignment << PAGE_SHIFT) / AMDGPU_GPU_PAGE_SIZE;
drivers/gpu/drm/amd/amdgpu/amdgpu_res_cursor.h
100
cur->size = min((node->size << PAGE_SHIFT) - start, size);
drivers/gpu/drm/amd/amdgpu/amdgpu_res_cursor.h
161
cur->start = node->start << PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_res_cursor.h
162
cur->size = min(node->size << PAGE_SHIFT, cur->remaining);
drivers/gpu/drm/amd/amdgpu/amdgpu_res_cursor.h
96
while (start >= node->size << PAGE_SHIFT)
drivers/gpu/drm/amd/amdgpu/amdgpu_res_cursor.h
97
start -= node++->size << PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_res_cursor.h
99
cur->start = (node->start << PAGE_SHIFT) + start;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
1001
gtt->offset = (u64)tmp->start << PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
147
abo->placements[0].fpfn = adev->gmc.visible_vram_size >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
2183
gtt_size = ttm_tt_pages_limit() << PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
265
dma_addr = &bo->ttm->dma_address[mm_cur->start >> PAGE_SHIFT];
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
2858
pfn = addr >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
2909
pfn = addr >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
631
mem->bus.offset = mem->start << PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
642
mem->bus.offset = mem->start << PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
648
mem->bus.offset = mem->start << PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
665
amdgpu_res_first(bo->resource, (u64)page_offset << PAGE_SHIFT, 0,
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
669
return ((uint64_t)(adev->doorbell.base + cursor.start)) >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
671
return ((uint64_t)(adev->rmmio_remap.bus_addr + cursor.start)) >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
673
return (adev->gmc.aper_base + cursor.start) >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
801
(u64)ttm->num_pages << PAGE_SHIFT,
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
867
gtt->offset + (page_idx << PAGE_SHIFT),
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
952
gtt->offset = (u64)bo_mem->start << PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
989
placements.lpfn = adev->gmc.gart_size >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
384
mapping = amdgpu_vm_bo_lookup_mapping(queue->vm, addr >> PAGE_SHIFT);
drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c
553
abo->placements[i].fpfn = 0 >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c
554
abo->placements[i].lpfn = (256 * 1024 * 1024) >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c
640
fpfn = PAGE_ALIGN(offset) >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c
641
lpfn = 0x100000000ULL >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c
644
lpfn = (0x100000000ULL - PAGE_ALIGN(offset)) >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
1198
uint64_t pfn = cursor.start >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
1223
addr = pages_addr[cursor.start >> PAGE_SHIFT];
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
2087
after->offset += (after->start - tmp->start) << PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
2996
node_id, addr >> PAGE_SHIFT, ts, write_fault)) {
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
962
result = pages_addr[addr >> PAGE_SHIFT];
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
289
if (res->start >= adev->gmc.visible_vram_size >> PAGE_SHIFT)
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
459
lpfn = (u64)place->lpfn << PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
463
fpfn = (u64)place->fpfn << PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
476
pages_per_block = 2UL << (20UL - PAGE_SHIFT);
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
526
min_block_size = (u64)tbo->page_alignment << PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
534
else if ((size >= (u64)pages_per_block << PAGE_SHIFT) &&
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
535
!(size & (((u64)pages_per_block << PAGE_SHIFT) - 1)))
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
536
min_block_size = (u64)pages_per_block << PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
550
pages_per_block = max_t(u32, 2UL << (20UL - PAGE_SHIFT),
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
594
start >>= PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
824
amdgpu_vram_mgr_block_start(block) >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
826
(amdgpu_vram_mgr_block_size(block) >> PAGE_SHIFT);
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
857
amdgpu_vram_mgr_block_start(block) >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
859
(amdgpu_vram_mgr_block_size(block) >> PAGE_SHIFT);
drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c
1038
*addr |= 1 << PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c
1248
if (!page_is_ram(addr >> PAGE_SHIFT)) {
drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c
1252
nid = pfn_to_nid(addr >> PAGE_SHIFT);
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
78
wptr_mapping = amdgpu_vm_bo_lookup_mapping(wptr_vm, wptr >> PAGE_SHIFT);
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1092
args->va_addr >> PAGE_SHIFT,
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1093
(args->va_addr + args->size - 1) >> PAGE_SHIFT)) {
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1105
args->mmap_offset >> PAGE_SHIFT,
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1106
(args->mmap_offset + args->size - 1) >> PAGE_SHIFT)) {
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
3480
address >> PAGE_SHIFT,
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
3500
mmap_offset = vma->vm_pgoff << PAGE_SHIFT;
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1815
mem_in_bytes <<= PAGE_SHIFT;
drivers/gpu/drm/amd/amdkfd/kfd_doorbell.c
143
address >> PAGE_SHIFT,
drivers/gpu/drm/amd/amdkfd/kfd_events.c
1085
pfn >>= PAGE_SHIFT;
drivers/gpu/drm/amd/amdkfd/kfd_events.c
1223
PAGE_SHIFT;
drivers/gpu/drm/amd/amdkfd/kfd_int_process_v10.c
366
exception_data.va = (info.page_addr) << PAGE_SHIFT;
drivers/gpu/drm/amd/amdkfd/kfd_int_process_v11.c
339
exception_data.va = (info.page_addr) << PAGE_SHIFT;
drivers/gpu/drm/amd/amdkfd/kfd_int_process_v12_1.c
326
exception_data.va = (info.page_addr) << PAGE_SHIFT;
drivers/gpu/drm/amd/amdkfd/kfd_int_process_v9.c
560
exception_data.va = (info.page_addr) << PAGE_SHIFT;
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
215
return (addr + adev->kfd.pgmap.range.start) >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
244
addr = page_to_pfn(page) << PAGE_SHIFT;
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
264
page = pfn_to_page(addr >> PAGE_SHIFT);
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
304
npages << PAGE_SHIFT, &cursor);
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
309
dst[i] = cursor.start + (j << PAGE_SHIFT);
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
334
amdgpu_res_next(&cursor, (j + 1) << PAGE_SHIFT);
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
343
src[i] >> PAGE_SHIFT, page_to_pfn(spage));
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
345
if (j >= (cursor.size >> PAGE_SHIFT) - 1 && i < npages - 1) {
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
398
u64 npages = (end - start) >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
427
start >> PAGE_SHIFT, end >> PAGE_SHIFT,
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
465
start >> PAGE_SHIFT, end >> PAGE_SHIFT,
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
520
start = start_mgr << PAGE_SHIFT;
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
521
end = (last_mgr + 1) << PAGE_SHIFT;
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
537
ttm_res_offset = (start_mgr - prange->start + prange->offset) << PAGE_SHIFT;
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
653
dst[i] >> PAGE_SHIFT, page_to_pfn(dpage));
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
697
u64 npages = (end - start) >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
730
start >> PAGE_SHIFT, end >> PAGE_SHIFT,
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
769
start >> PAGE_SHIFT, end >> PAGE_SHIFT,
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
829
start = start_mgr << PAGE_SHIFT;
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
830
end = (last_mgr + 1) << PAGE_SHIFT;
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
978
addr >>= PAGE_SHIFT;
drivers/gpu/drm/amd/amdkfd/kfd_queue.c
158
addr >>= PAGE_SHIFT;
drivers/gpu/drm/amd/amdkfd/kfd_queue.c
159
last = addr + (size >> PAGE_SHIFT) - 1;
drivers/gpu/drm/amd/amdkfd/kfd_queue.c
98
addr >>= PAGE_SHIFT;
drivers/gpu/drm/amd/amdkfd/kfd_queue.c
99
size >>= PAGE_SHIFT;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
116
prange->start << PAGE_SHIFT,
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
117
prange->npages << PAGE_SHIFT,
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
144
prange->notifier.interval_tree.start >> PAGE_SHIFT,
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
145
prange->notifier.interval_tree.last >> PAGE_SHIFT);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1484
(last_start - prange->start) << PAGE_SHIFT,
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1752
start = map_start << PAGE_SHIFT;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1753
end = (map_last + 1) << PAGE_SHIFT;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1769
npages = (next - addr) >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1808
offset = (addr >> PAGE_SHIFT) - prange->start;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
186
addr[i] = (hmm_pfns[i] << PAGE_SHIFT) +
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
200
addr[i] >> PAGE_SHIFT, page_to_pfn(page));
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2306
start = prange->notifier.interval_tree.start >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2307
last = prange->notifier.interval_tree.last >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
248
pr_debug_ratelimited("unmap 0x%llx\n", dma_addr[i] >> PAGE_SHIFT);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2554
prange->start << PAGE_SHIFT);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2662
start = max(start, range->start) >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2663
last = min(last, range->end - 1) >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2665
start, last, range->start >> PAGE_SHIFT,
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2666
(range->end - 1) >> PAGE_SHIFT,
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2667
mni->interval_tree.start >> PAGE_SHIFT,
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2668
mni->interval_tree.last >> PAGE_SHIFT, range->event);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2811
vma = vma_lookup(p->mm, addr << PAGE_SHIFT);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2819
start_limit = max(vma->vm_start >> PAGE_SHIFT,
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
282
uint64_t size = (prange->last - prange->start + 1) << PAGE_SHIFT;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2821
end_limit = min(vma->vm_end >> PAGE_SHIFT,
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2849
vma->vm_start >> PAGE_SHIFT, vma->vm_end >> PAGE_SHIFT,
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2885
start << PAGE_SHIFT,
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2886
last << PAGE_SHIFT,
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2895
*bo_s = userptr >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
3171
vma = vma_lookup(mm, addr << PAGE_SHIFT);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
3279
size = (pchild->last - pchild->start + 1) << PAGE_SHIFT;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
3292
size = (prange->last - prange->start + 1) << PAGE_SHIFT;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
338
amdgpu_amdkfd_reserve_mem_limit(NULL, size << PAGE_SHIFT,
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
3464
start <<= PAGE_SHIFT;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
3465
end = start + (size << PAGE_SHIFT);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
3473
return svm_range_check_vm(p, start_unchg, (end - 1) >> PAGE_SHIFT, NULL,
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
4321
start >>= PAGE_SHIFT;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
4322
size >>= PAGE_SHIFT;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
613
bo->tbo.resource->start << PAGE_SHIFT, bp.size,
drivers/gpu/drm/armada/armada_gem.c
25
unsigned long pfn = obj->phys_addr >> PAGE_SHIFT;
drivers/gpu/drm/armada/armada_gem.c
27
pfn += (vmf->address - vmf->vma->vm_start) >> PAGE_SHIFT;
drivers/gpu/drm/drm_fbdev_shmem.c
88
unsigned int i = offset >> PAGE_SHIFT;
drivers/gpu/drm/drm_gem.c
1329
obj = drm_gem_object_lookup_at_offset(filp, pgoff, len >> PAGE_SHIFT);
drivers/gpu/drm/drm_gem.c
1371
drm_vma_node_size(&obj->vma_node) << PAGE_SHIFT,
drivers/gpu/drm/drm_gem.c
1558
obj->lru->count -= obj->size >> PAGE_SHIFT;
drivers/gpu/drm/drm_gem.c
1601
lru->count += obj->size >> PAGE_SHIFT;
drivers/gpu/drm/drm_gem.c
1689
*remaining += obj->size >> PAGE_SHIFT;
drivers/gpu/drm/drm_gem.c
1694
freed += obj->size >> PAGE_SHIFT;
drivers/gpu/drm/drm_gem.c
685
npages = obj->size >> PAGE_SHIFT;
drivers/gpu/drm/drm_gem.c
756
npages = obj->size >> PAGE_SHIFT;
drivers/gpu/drm/drm_gem_dma_helper.c
365
len >> PAGE_SHIFT);
drivers/gpu/drm/drm_gem_shmem_helper.c
236
set_pages_array_wc(pages, obj->size >> PAGE_SHIFT);
drivers/gpu/drm/drm_gem_shmem_helper.c
261
set_pages_array_wb(shmem->pages, obj->size >> PAGE_SHIFT);
drivers/gpu/drm/drm_gem_shmem_helper.c
395
shmem->vaddr = vmap(shmem->pages, obj->size >> PAGE_SHIFT,
drivers/gpu/drm/drm_gem_shmem_helper.c
562
loff_t num_pages = obj->size >> PAGE_SHIFT;
drivers/gpu/drm/drm_gem_shmem_helper.c
579
unsigned long paddr = pfn << PAGE_SHIFT;
drivers/gpu/drm/drm_gem_shmem_helper.c
586
pfn &= PMD_MASK >> PAGE_SHIFT;
drivers/gpu/drm/drm_gem_shmem_helper.c
613
loff_t num_pages = obj->size >> PAGE_SHIFT;
drivers/gpu/drm/drm_gem_shmem_helper.c
794
return drm_prime_pages_to_sg(obj->dev, shmem->pages, obj->size >> PAGE_SHIFT);
drivers/gpu/drm/drm_gem_vram_helper.c
268
return drm_gem_vram_pg_offset(gbo) << PAGE_SHIFT;
drivers/gpu/drm/drm_gem_vram_helper.c
796
mem->bus.offset = (mem->start << PAGE_SHIFT) + vmm->vram_base;
drivers/gpu/drm/drm_gem_vram_helper.c
865
false, vram_size >> PAGE_SHIFT);
drivers/gpu/drm/drm_gem_vram_helper.c
966
max_fbpages = (vmm->vram_size / 2) >> PAGE_SHIFT;
drivers/gpu/drm/drm_gpusvm.c
270
return (end - start) >> PAGE_SHIFT;
drivers/gpu/drm/drm_pagemap.c
1024
npages = devmem_allocation->size >> PAGE_SHIFT;
drivers/gpu/drm/drm_pagemap.c
346
return (end - start) >> PAGE_SHIFT;
drivers/gpu/drm/drm_panic.c
245
new_page = offset >> PAGE_SHIFT;
drivers/gpu/drm/drm_panic.c
344
new_page = offset >> PAGE_SHIFT;
drivers/gpu/drm/drm_prime.c
866
(unsigned long)nr_pages << PAGE_SHIFT,
drivers/gpu/drm/etnaviv/etnaviv_dump.c
148
n_bomap_pages += obj->base.size >> PAGE_SHIFT;
drivers/gpu/drm/etnaviv/etnaviv_dump.c
216
for (j = 0; j < obj->base.size >> PAGE_SHIFT; j++)
drivers/gpu/drm/etnaviv/etnaviv_gem.c
104
unsigned int npages = etnaviv_obj->base.size >> PAGE_SHIFT;
drivers/gpu/drm/etnaviv/etnaviv_gem.c
191
pgoff = (vmf->address - vma->vm_start) >> PAGE_SHIFT;
drivers/gpu/drm/etnaviv/etnaviv_gem.c
196
pfn, pfn << PAGE_SHIFT);
drivers/gpu/drm/etnaviv/etnaviv_gem.c
366
return vmap(pages, obj->base.size >> PAGE_SHIFT, VM_MAP, prot);
drivers/gpu/drm/etnaviv/etnaviv_gem.c
670
int ret, pinned = 0, npages = etnaviv_obj->base.size >> PAGE_SHIFT;
drivers/gpu/drm/etnaviv/etnaviv_gem.c
714
unsigned int npages = etnaviv_obj->base.size >> PAGE_SHIFT;
drivers/gpu/drm/etnaviv/etnaviv_gem_prime.c
20
unsigned int npages = obj->size >> PAGE_SHIFT;
drivers/gpu/drm/exynos/exynos_drm_g2d.c
472
npages = (end - start) >> PAGE_SHIFT;
drivers/gpu/drm/exynos/exynos_drm_g2d.c
521
if (g2d->current_pool + (npages << PAGE_SHIFT) < g2d->max_pool) {
drivers/gpu/drm/exynos/exynos_drm_g2d.c
522
g2d->current_pool += npages << PAGE_SHIFT;
drivers/gpu/drm/gma500/fbdev.c
26
unsigned long address = vmf->address - (vmf->pgoff << PAGE_SHIFT);
drivers/gpu/drm/gma500/fbdev.c
27
unsigned long pfn = info->fix.smem_start >> PAGE_SHIFT;
drivers/gpu/drm/gma500/fbdev.c
57
if (vma->vm_pgoff > (~0UL >> PAGE_SHIFT))
drivers/gpu/drm/gma500/gem.c
291
page_offset = (vmf->address - vma->vm_start) >> PAGE_SHIFT;
drivers/gpu/drm/gma500/gem.c
295
pfn = (dev_priv->stolen_base + pobj->offset) >> PAGE_SHIFT;
drivers/gpu/drm/gma500/gem.c
317
pfn_base = pdev->stolen_base >> PAGE_SHIFT;
drivers/gpu/drm/gma500/gem.c
318
num_pages = pdev->vram_stolen_size >> PAGE_SHIFT;
drivers/gpu/drm/gma500/gem.c
321
num_pages, pfn_base << PAGE_SHIFT, 0);
drivers/gpu/drm/gma500/gtt.c
116
npages = resource_size(res) >> PAGE_SHIFT;
drivers/gpu/drm/gma500/gtt.c
206
gtt_pages = pci_resource_len(pdev, PSB_GTT_RESOURCE) >> PAGE_SHIFT;
drivers/gpu/drm/gma500/gtt.c
216
gatt_pages = pci_resource_len(pdev, PSB_GATT_RESOURCE) >> PAGE_SHIFT;
drivers/gpu/drm/gma500/gtt.c
228
gatt_pages = (128 * 1024 * 1024) >> PAGE_SHIFT;
drivers/gpu/drm/gma500/gtt.c
269
dev_priv->gtt_map = ioremap(pg->gtt_phys_start, pg->gtt_pages << PAGE_SHIFT);
drivers/gpu/drm/gma500/gtt.c
58
BUG_ON(pfn & ~(0xFFFFFFFF >> PAGE_SHIFT));
drivers/gpu/drm/gma500/gtt.c
67
return (pfn << PAGE_SHIFT) | mask;
drivers/gpu/drm/gma500/gtt.c
74
return pdev->gtt_map + (offset >> PAGE_SHIFT);
drivers/gpu/drm/gma500/gtt.c
89
npages = resource_size(res) >> PAGE_SHIFT;
drivers/gpu/drm/gma500/mmu.c
129
PSB_WSGX32(page_to_pfn(pd->p) << PAGE_SHIFT, offset);
drivers/gpu/drm/gma500/mmu.c
155
return (pfn << PAGE_SHIFT) | mask;
drivers/gpu/drm/gma500/mmu.c
265
uint32_t clflush_add = pd->driver->clflush_add >> PAGE_SHIFT;
drivers/gpu/drm/gma500/mmu.c
501
add = desired_tile_stride << PAGE_SHIFT;
drivers/gpu/drm/gma500/mmu.c
502
row_add = hw_tile_stride << PAGE_SHIFT;
drivers/gpu/drm/gma500/mmu.c
538
end = addr + (num_pages << PAGE_SHIFT);
drivers/gpu/drm/gma500/mmu.c
584
add = desired_tile_stride << PAGE_SHIFT;
drivers/gpu/drm/gma500/mmu.c
585
row_add = hw_tile_stride << PAGE_SHIFT;
drivers/gpu/drm/gma500/mmu.c
636
end = addr + (num_pages << PAGE_SHIFT);
drivers/gpu/drm/gma500/mmu.c
692
add = desired_tile_stride << PAGE_SHIFT;
drivers/gpu/drm/gma500/mmu.c
693
row_add = hw_tile_stride << PAGE_SHIFT;
drivers/gpu/drm/gma500/psb_drv.c
140
stolen_gtt = (pg->stolen_size >> PAGE_SHIFT) * 4;
drivers/gpu/drm/gma500/psb_drv.c
141
stolen_gtt = (stolen_gtt + PAGE_SIZE - 1) >> PAGE_SHIFT;
drivers/gpu/drm/gma500/psb_drv.c
145
(stolen_gtt << PAGE_SHIFT) * 1024;
drivers/gpu/drm/gma500/psb_drv.c
194
dev_priv->vram_stolen_size >> PAGE_SHIFT);
drivers/gpu/drm/gma500/psb_drv.c
355
dev_priv->stolen_base >> PAGE_SHIFT,
drivers/gpu/drm/gma500/psb_drv.c
357
pg->stolen_size >> PAGE_SHIFT, 0);
drivers/gpu/drm/gma500/psb_drv.h
66
#define PSB_TT_PRIV0_PLIMIT (PSB_TT_PRIV0_LIMIT >> PAGE_SHIFT)
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
1172
offset += cache->page << PAGE_SHIFT;
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
1325
offset += page << PAGE_SHIFT;
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
1391
offset >> PAGE_SHIFT);
drivers/gpu/drm/i915/gem/i915_gem_internal.c
42
if (overflows_type(obj->base.size >> PAGE_SHIFT, npages))
drivers/gpu/drm/i915/gem/i915_gem_internal.c
45
npages = obj->base.size >> PAGE_SHIFT;
drivers/gpu/drm/i915/gem/i915_gem_internal.c
46
max_segment = i915_sg_segment_size(i915->drm.dev) >> PAGE_SHIFT;
drivers/gpu/drm/i915/gem/i915_gem_mman.c
134
return i915_gem_object_get_tile_row_size(obj) >> PAGE_SHIFT;
drivers/gpu/drm/i915/gem/i915_gem_mman.c
213
(obj->base.size >> PAGE_SHIFT) - view.partial.offset);
drivers/gpu/drm/i915/gem/i915_gem_mman.c
216
if (chunk >= obj->base.size >> PAGE_SHIFT)
drivers/gpu/drm/i915/gem/i915_gem_mman.c
312
vm_start = area->vm_start >> PAGE_SHIFT;
drivers/gpu/drm/i915/gem/i915_gem_mman.c
313
vm_end = area->vm_end >> PAGE_SHIFT;
drivers/gpu/drm/i915/gem/i915_gem_mman.c
314
vma_size = vma->size >> PAGE_SHIFT;
drivers/gpu/drm/i915/gem/i915_gem_mman.c
330
*start_vaddr = (unsigned long)start << PAGE_SHIFT;
drivers/gpu/drm/i915/gem/i915_gem_mman.c
331
*end_vaddr = (unsigned long)end << PAGE_SHIFT;
drivers/gpu/drm/i915/gem/i915_gem_mman.c
333
*pfn = (gmadr_start + i915_ggtt_offset(vma)) >> PAGE_SHIFT;
drivers/gpu/drm/i915/gem/i915_gem_mman.c
334
*pfn += (*start_vaddr - area->vm_start) >> PAGE_SHIFT;
drivers/gpu/drm/i915/gem/i915_gem_mman.c
340
#define MIN_CHUNK_PAGES (SZ_1M >> PAGE_SHIFT)
drivers/gpu/drm/i915/gem/i915_gem_mman.c
360
page_offset = (vmf->address - area->vm_start) >> PAGE_SHIFT;
drivers/gpu/drm/i915/gem/i915_gem_object.c
480
pgoff_t idx = offset >> PAGE_SHIFT;
drivers/gpu/drm/i915/gem/i915_gem_object.c
495
pgoff_t idx = offset >> PAGE_SHIFT;
drivers/gpu/drm/i915/gem/i915_gem_object.c
537
GEM_BUG_ON(overflows_type(offset >> PAGE_SHIFT, pgoff_t));
drivers/gpu/drm/i915/gem/i915_gem_pages.c
278
unsigned long n_pages = obj->base.size >> PAGE_SHIFT, i;
drivers/gpu/drm/i915/gem/i915_gem_pages.c
337
unsigned long n_pfn = obj->base.size >> PAGE_SHIFT;
drivers/gpu/drm/i915/gem/i915_gem_pages.c
354
pfns[i++] = (iomap + addr) >> PAGE_SHIFT;
drivers/gpu/drm/i915/gem/i915_gem_pages.c
379
unsigned long n_pages = obj->base.size >> PAGE_SHIFT, i;
drivers/gpu/drm/i915/gem/i915_gem_pages.c
422
new_page = offset >> PAGE_SHIFT;
drivers/gpu/drm/i915/gem/i915_gem_pages.c
657
GEM_BUG_ON(n >= obj->base.size >> PAGE_SHIFT);
drivers/gpu/drm/i915/gem/i915_gem_pages.c
796
*len = sg_dma_len(sg) - (offset << PAGE_SHIFT);
drivers/gpu/drm/i915/gem/i915_gem_pages.c
798
return sg_dma_address(sg) + (offset << PAGE_SHIFT);
drivers/gpu/drm/i915/gem/i915_gem_shmem.c
256
obj->base.size >> PAGE_SHIFT);
drivers/gpu/drm/i915/gem/i915_gem_shmem.c
699
totalram_pages() << PAGE_SHIFT,
drivers/gpu/drm/i915/gem/i915_gem_shrinker.c
231
count += obj->base.size >> PAGE_SHIFT;
drivers/gpu/drm/i915/gem/i915_gem_shrinker.c
236
scanned += obj->base.size >> PAGE_SHIFT;
drivers/gpu/drm/i915/gem/i915_gem_shrinker.c
296
count = READ_ONCE(i915->mm.shrink_memory) >> PAGE_SHIFT;
drivers/gpu/drm/i915/gem/i915_gem_shrinker.c
372
unevictable += obj->base.size >> PAGE_SHIFT;
drivers/gpu/drm/i915/gem/i915_gem_shrinker.c
374
available += obj->base.size >> PAGE_SHIFT;
drivers/gpu/drm/i915/gem/i915_gem_shrinker.c
409
unsigned long count = i915_vma_size(vma) >> PAGE_SHIFT;
drivers/gpu/drm/i915/gem/i915_gem_tiling.c
312
obj->bit_17 = bitmap_zalloc(obj->base.size >> PAGE_SHIFT,
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
1335
&i915_sys_placement, page_size >> PAGE_SHIFT,
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
1345
if (size >> PAGE_SHIFT > INT_MAX && ret == -ENOSPC)
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
1373
totalram_pages() << PAGE_SHIFT,
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
143
WARN_ON(overflows_type(offset >> PAGE_SHIFT, place->fpfn));
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
144
place->fpfn = offset >> PAGE_SHIFT;
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
145
WARN_ON(overflows_type(place->fpfn + (size >> PAGE_SHIFT), place->lpfn));
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
146
place->lpfn = place->fpfn + (size >> PAGE_SHIFT);
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
152
WARN_ON(overflows_type(resource_size(&mr->io) >> PAGE_SHIFT, place->lpfn));
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
153
place->lpfn = resource_size(&mr->io) >> PAGE_SHIFT;
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
191
const size_t size = (size_t)ttm->num_pages << PAGE_SHIFT;
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
534
0, (unsigned long)ttm->num_pages << PAGE_SHIFT,
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
574
page_alignment = bo->page_alignment << PAGE_SHIFT;
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
700
return ((base + sg_dma_address(sg)) >> PAGE_SHIFT) + ofs;
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
710
unsigned long page = offset >> PAGE_SHIFT;
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
721
offset -= page << PAGE_SHIFT;
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
834
GEM_BUG_ON(bo->ttm && ((obj->base.size >> PAGE_SHIFT) < bo->ttm->num_pages));
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
845
if (overflows_type(obj->base.size >> PAGE_SHIFT, unsigned int))
drivers/gpu/drm/i915/gem/i915_gem_ttm_move.c
335
arg->num_pages = bo->base.size >> PAGE_SHIFT;
drivers/gpu/drm/i915/gem/i915_gem_userptr.c
108
if (overflows_type(obj->base.size >> PAGE_SHIFT, num_pages))
drivers/gpu/drm/i915/gem/i915_gem_userptr.c
111
num_pages = obj->base.size >> PAGE_SHIFT;
drivers/gpu/drm/i915/gem/i915_gem_userptr.c
126
num_pages << PAGE_SHIFT,
drivers/gpu/drm/i915/gem/i915_gem_userptr.c
236
const unsigned long num_pages = obj->base.size >> PAGE_SHIFT;
drivers/gpu/drm/i915/gem/i915_gem_userptr.c
93
const unsigned long num_pages = obj->base.size >> PAGE_SHIFT;
drivers/gpu/drm/i915/gem/selftests/huge_pages.c
1069
vma->size >> PAGE_SHIFT, val);
drivers/gpu/drm/i915/gem/selftests/huge_pages.c
1084
for (n = 0; n < obj->base.size >> PAGE_SHIFT; ++n) {
drivers/gpu/drm/i915/gem/selftests/huge_pages.c
1110
unsigned long n = obj->base.size >> PAGE_SHIFT;
drivers/gpu/drm/i915/gem/selftests/huge_pages.c
185
if (size >> PAGE_SHIFT > INT_MAX)
drivers/gpu/drm/i915/gem/selftests/huge_pages.c
220
if (overflows_type(obj->base.size >> PAGE_SHIFT, unsigned int))
drivers/gpu/drm/i915/gem/selftests/huge_pages.c
227
if (sg_alloc_table(st, obj->base.size >> PAGE_SHIFT, GFP)) {
drivers/gpu/drm/i915/gem/selftests/huge_pages.c
336
if (size >> PAGE_SHIFT > UINT_MAX)
drivers/gpu/drm/i915/gem/selftests/huge_pages.c
740
max_pages = vm->total >> PAGE_SHIFT;
drivers/gpu/drm/i915/gem/selftests/huge_pages.c
744
u64 size = page_num << PAGE_SHIFT;
drivers/gpu/drm/i915/gem/selftests/huge_pages.c
88
if (overflows_type(obj->base.size >> PAGE_SHIFT, unsigned int))
drivers/gpu/drm/i915/gem/selftests/huge_pages.c
95
if (sg_alloc_table(st, obj->base.size >> PAGE_SHIFT, GFP)) {
drivers/gpu/drm/i915/gem/selftests/i915_gem_coherency.c
35
page = i915_gem_object_get_page(ctx->obj, offset >> PAGE_SHIFT);
drivers/gpu/drm/i915/gem/selftests/i915_gem_coherency.c
66
page = i915_gem_object_get_page(ctx->obj, offset >> PAGE_SHIFT);
drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c
433
return huge_gem_object_phys_size(obj) >> PAGE_SHIFT;
drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c
438
return huge_gem_object_dma_size(obj) >> PAGE_SHIFT;
drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c
468
(dw * real_page_count(obj)) << PAGE_SHIFT |
drivers/gpu/drm/i915/gem/selftests/i915_gem_dmabuf.c
190
vma->size >> PAGE_SHIFT, 0xdeadbeaf);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
1350
io_size >> PAGE_SHIFT);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
1410
resource_size(&mr->io) >> PAGE_SHIFT);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
153
offset = tiled_offset(tile, page << PAGE_SHIFT);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
159
p = i915_gem_object_get_page(obj, offset >> PAGE_SHIFT);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
167
vma->size >> PAGE_SHIFT,
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
1699
(addr - (unsigned long)data) >> PAGE_SHIFT);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
170
offset >> PAGE_SHIFT,
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
1712
(addr - (unsigned long)data) >> PAGE_SHIFT);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
249
offset = tiled_offset(tile, page << PAGE_SHIFT);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
255
p = i915_gem_object_get_page(obj, offset >> PAGE_SHIFT);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
263
vma->size >> PAGE_SHIFT,
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
266
offset >> PAGE_SHIFT,
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
338
nreal << PAGE_SHIFT,
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
339
(1 + next_prime_number(to_gt(i915)->ggtt->vm.total >> PAGE_SHIFT)) << PAGE_SHIFT);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
475
nreal << PAGE_SHIFT,
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
476
(1 + next_prime_number(to_gt(i915)->ggtt->vm.total >> PAGE_SHIFT)) << PAGE_SHIFT);
drivers/gpu/drm/i915/gt/intel_ggtt_fencing.c
817
const unsigned int page_count = obj->base.size >> PAGE_SHIFT;
drivers/gpu/drm/i915/gt/intel_ggtt_gmch.c
28
intel_gmch_gtt_insert_page(addr, offset >> PAGE_SHIFT, flags);
drivers/gpu/drm/i915/gt/intel_ggtt_gmch.c
34
return intel_gmch_gtt_read_entry(offset >> PAGE_SHIFT,
drivers/gpu/drm/i915/gt/intel_ggtt_gmch.c
46
intel_gmch_gtt_insert_sg_entries(vma_res->bi.pages, vma_res->start >> PAGE_SHIFT,
drivers/gpu/drm/i915/gt/intel_ggtt_gmch.c
58
intel_gmch_gtt_clear_range(start >> PAGE_SHIFT, length >> PAGE_SHIFT);
drivers/gpu/drm/i915/gt/intel_gt_buffer_pool.c
24
n = fls(sz >> PAGE_SHIFT) - 1;
drivers/gpu/drm/i915/gt/intel_gtt.h
535
return (address >> PAGE_SHIFT) & mask;
drivers/gpu/drm/i915/gt/intel_gtt.h
64
#define ggtt_total_entries(ggtt) ((ggtt)->vm.total >> PAGE_SHIFT)
drivers/gpu/drm/i915/gt/intel_gtt.h
84
#define NUM_PTE(pde_shift) (1 << (pde_shift - PAGE_SHIFT))
drivers/gpu/drm/i915/gt/intel_migrate.c
480
u32 height = size >> PAGE_SHIFT;
drivers/gpu/drm/i915/gt/intel_migrate.c
596
*cs++ = size >> PAGE_SHIFT << 16 | PAGE_SIZE / 4;
drivers/gpu/drm/i915/gt/intel_migrate.c
607
*cs++ = size >> PAGE_SHIFT << 16 | PAGE_SIZE / 4;
drivers/gpu/drm/i915/gt/intel_migrate.c
618
*cs++ = size >> PAGE_SHIFT << 16 | PAGE_SIZE;
drivers/gpu/drm/i915/gt/intel_migrate.c
926
GEM_BUG_ON(size >> PAGE_SHIFT > S16_MAX);
drivers/gpu/drm/i915/gt/intel_migrate.c
945
*cs++ = size >> PAGE_SHIFT << 16 | PAGE_SIZE / 4;
drivers/gpu/drm/i915/gt/intel_migrate.c
965
*cs++ = size >> PAGE_SHIFT << 16 | PAGE_SIZE / 4;
drivers/gpu/drm/i915/gt/intel_migrate.c
974
*cs++ = size >> PAGE_SHIFT << 16 | PAGE_SIZE / 4;
drivers/gpu/drm/i915/gt/intel_reset.c
864
vma_offset = vma->gtt_view.partial.offset << PAGE_SHIFT;
drivers/gpu/drm/i915/gt/selftest_reset.c
100
((page + 1) << PAGE_SHIFT) - 1))
drivers/gpu/drm/i915/gt/selftest_reset.c
125
dma_addr_t dma = (dma_addr_t)dsm->start + (page << PAGE_SHIFT);
drivers/gpu/drm/i915/gt/selftest_reset.c
148
page << PAGE_SHIFT,
drivers/gpu/drm/i915/gt/selftest_reset.c
149
((page + 1) << PAGE_SHIFT) - 1)) {
drivers/gpu/drm/i915/gt/selftest_reset.c
166
if (max >= I915_GEM_STOLEN_BIAS >> PAGE_SHIFT) {
drivers/gpu/drm/i915/gt/selftest_reset.c
37
num_pages = resource_size(dsm) >> PAGE_SHIFT;
drivers/gpu/drm/i915/gt/selftest_reset.c
83
dma_addr_t dma = (dma_addr_t)dsm->start + (page << PAGE_SHIFT);
drivers/gpu/drm/i915/gt/selftest_reset.c
99
page << PAGE_SHIFT,
drivers/gpu/drm/i915/gt/shmem_utils.c
101
for (pfn = off >> PAGE_SHIFT; len; pfn++) {
drivers/gpu/drm/i915/gt/shmem_utils.c
136
for (pfn = off >> PAGE_SHIFT; len; pfn++) {
drivers/gpu/drm/i915/gt/shmem_utils.c
65
n_pages = file->f_mapping->host->i_size >> PAGE_SHIFT;
drivers/gpu/drm/i915/gt/uc/intel_guc.c
270
offset = intel_guc_ggtt_offset(guc, log->vma) >> PAGE_SHIFT;
drivers/gpu/drm/i915/gt/uc/intel_guc.c
286
u32 ads = intel_guc_ggtt_offset(guc, guc->ads_vma) >> PAGE_SHIFT;
drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c
1289
idx = offset >> PAGE_SHIFT;
drivers/gpu/drm/i915/gvt/dmabuf.c
58
if (overflows_type(obj->base.size >> PAGE_SHIFT, page_num))
drivers/gpu/drm/i915/gvt/dmabuf.c
61
page_num = obj->base.size >> PAGE_SHIFT;
drivers/gpu/drm/i915/gvt/dmabuf.c
80
(fb_info->start >> PAGE_SHIFT);
drivers/gpu/drm/i915/gvt/gtt.c
1119
ops->set_pfn(&sub_se, dma_addr >> PAGE_SHIFT);
drivers/gpu/drm/i915/gvt/gtt.c
1169
ops->set_pfn(&entry, dma_addr >> PAGE_SHIFT);
drivers/gpu/drm/i915/gvt/gtt.c
1221
pte_ops->set_pfn(&se, dma_addr >> PAGE_SHIFT);
drivers/gpu/drm/i915/gvt/gtt.c
1361
spt->guest_page.gfn << PAGE_SHIFT, vgpu);
drivers/gpu/drm/i915/gvt/gtt.c
1863
mm->ggtt_mm.host_ggtt_aperture = vzalloc((vgpu_aperture_sz(vgpu) >> PAGE_SHIFT) * sizeof(u64));
drivers/gpu/drm/i915/gvt/gtt.c
1870
mm->ggtt_mm.host_ggtt_hidden = vzalloc((vgpu_hidden_sz(vgpu) >> PAGE_SHIFT) * sizeof(u64));
drivers/gpu/drm/i915/gvt/gtt.c
2142
intel_gvt_dma_unmap_guest_page(vgpu, pfn << PAGE_SHIFT);
drivers/gpu/drm/i915/gvt/gtt.c
2235
ops->set_pfn(&m, dma_addr >> PAGE_SHIFT);
drivers/gpu/drm/i915/gvt/gtt.c
2737
index = vgpu_aperture_gmadr_base(vgpu) >> PAGE_SHIFT;
drivers/gpu/drm/i915/gvt/gtt.c
2738
num_entries = vgpu_aperture_sz(vgpu) >> PAGE_SHIFT;
drivers/gpu/drm/i915/gvt/gtt.c
2747
index = vgpu_hidden_gmadr_base(vgpu) >> PAGE_SHIFT;
drivers/gpu/drm/i915/gvt/gtt.c
2748
num_entries = vgpu_hidden_sz(vgpu) >> PAGE_SHIFT;
drivers/gpu/drm/i915/gvt/gtt.c
2780
num_low = vgpu_aperture_sz(vgpu) >> PAGE_SHIFT;
drivers/gpu/drm/i915/gvt/gtt.c
2781
offset = vgpu_aperture_gmadr_base(vgpu) >> PAGE_SHIFT;
drivers/gpu/drm/i915/gvt/gtt.c
2788
num_hi = vgpu_hidden_sz(vgpu) >> PAGE_SHIFT;
drivers/gpu/drm/i915/gvt/gtt.c
2789
offset = vgpu_hidden_gmadr_base(vgpu) >> PAGE_SHIFT;
drivers/gpu/drm/i915/gvt/gtt.c
307
pfn = (e->val64 & ADDR_1G_MASK) >> PAGE_SHIFT;
drivers/gpu/drm/i915/gvt/gtt.c
309
pfn = (e->val64 & ADDR_2M_MASK) >> PAGE_SHIFT;
drivers/gpu/drm/i915/gvt/gtt.c
311
pfn = (e->val64 & ADDR_64K_MASK) >> PAGE_SHIFT;
drivers/gpu/drm/i915/gvt/gtt.c
313
pfn = (e->val64 & ADDR_4K_MASK) >> PAGE_SHIFT;
drivers/gpu/drm/i915/gvt/gtt.c
321
pfn &= (ADDR_1G_MASK >> PAGE_SHIFT);
drivers/gpu/drm/i915/gvt/gtt.c
324
pfn &= (ADDR_2M_MASK >> PAGE_SHIFT);
drivers/gpu/drm/i915/gvt/gtt.c
327
pfn &= (ADDR_64K_MASK >> PAGE_SHIFT);
drivers/gpu/drm/i915/gvt/gtt.c
330
pfn &= (ADDR_4K_MASK >> PAGE_SHIFT);
drivers/gpu/drm/i915/gvt/gtt.c
333
e->val64 |= (pfn << PAGE_SHIFT);
drivers/gpu/drm/i915/gvt/gtt.c
575
offset -= (vgpu_aperture_gmadr_base(mm->vgpu) >> PAGE_SHIFT);
drivers/gpu/drm/i915/gvt/gtt.c
578
offset -= (vgpu_hidden_gmadr_base(mm->vgpu) >> PAGE_SHIFT);
drivers/gpu/drm/i915/gvt/gtt.c
935
intel_gvt_dma_unmap_guest_page(vgpu, pfn << PAGE_SHIFT);
drivers/gpu/drm/i915/gvt/gvt.h
407
#define gvt_ggtt_sz(gvt) (gvt_to_ggtt(gvt)->vm.total >> PAGE_SHIFT << 3)
drivers/gpu/drm/i915/gvt/kvmgt.c
1023
index = vma->vm_pgoff >> (VFIO_PCI_OFFSET_SHIFT - PAGE_SHIFT);
drivers/gpu/drm/i915/gvt/kvmgt.c
1038
((1U << (VFIO_PCI_OFFSET_SHIFT - PAGE_SHIFT)) - 1);
drivers/gpu/drm/i915/gvt/kvmgt.c
1039
req_start = pgoff << PAGE_SHIFT;
drivers/gpu/drm/i915/gvt/kvmgt.c
1047
pgoff = (gvt_aperture_pa_base(vgpu->gvt) >> PAGE_SHIFT) + pgoff;
drivers/gpu/drm/i915/gvt/kvmgt.c
132
vfio_unpin_pages(&vgpu->vfio_device, gfn << PAGE_SHIFT,
drivers/gpu/drm/i915/gvt/kvmgt.c
150
dma_addr_t cur_iova = (gfn + npage) << PAGE_SHIFT;
drivers/gpu/drm/i915/gvt/kvmgt.c
1569
if (kvmgt_gfn_is_write_protected(info, gpa >> PAGE_SHIFT))
drivers/gpu/drm/i915/gvt/kvmgt.c
626
u64 iov_pfn = iova >> PAGE_SHIFT;
drivers/gpu/drm/i915/gvt/opregion.c
277
vgpu_opregion(vgpu)->gfn[i] = (gpa >> PAGE_SHIFT) + i;
drivers/gpu/drm/i915/gvt/opregion.c
422
scic_pa = (vgpu_opregion(vgpu)->gfn[0] << PAGE_SHIFT) +
drivers/gpu/drm/i915/gvt/opregion.c
424
parm_pa = (vgpu_opregion(vgpu)->gfn[0] << PAGE_SHIFT) +
drivers/gpu/drm/i915/gvt/page_track.c
166
page_track = intel_vgpu_find_page_track(vgpu, gpa >> PAGE_SHIFT);
drivers/gpu/drm/i915/gvt/page_track.c
172
intel_gvt_page_track_remove(vgpu, gpa >> PAGE_SHIFT);
drivers/gpu/drm/i915/gvt/scheduler.c
224
context_page_num = context_page_num >> PAGE_SHIFT;
drivers/gpu/drm/i915/gvt/scheduler.c
984
context_page_num = context_page_num >> PAGE_SHIFT;
drivers/gpu/drm/i915/i915_cmd_parser.c
1211
for (n = offset >> PAGE_SHIFT; remain; n++) {
drivers/gpu/drm/i915/i915_debugfs.c
219
vma->gtt_view.partial.offset << PAGE_SHIFT,
drivers/gpu/drm/i915/i915_debugfs.c
220
vma->gtt_view.partial.size << PAGE_SHIFT);
drivers/gpu/drm/i915/i915_gem.c
254
for (idx = args->offset >> PAGE_SHIFT; remain; idx++) {
drivers/gpu/drm/i915/i915_gem.c
422
offset >> PAGE_SHIFT),
drivers/gpu/drm/i915/i915_gem.c
603
offset >> PAGE_SHIFT),
drivers/gpu/drm/i915/i915_gem.c
702
for (idx = args->offset >> PAGE_SHIFT; remain; idx++) {
drivers/gpu/drm/i915/i915_gem_gtt.c
48
obj->base.size >> PAGE_SHIFT, NULL,
drivers/gpu/drm/i915/i915_mm.c
111
zap_special_vma_range(vma, addr, (r.pfn - pfn) << PAGE_SHIFT);
drivers/gpu/drm/i915/i915_mm.c
146
while (offset >= r.sgt.max >> PAGE_SHIFT) {
drivers/gpu/drm/i915/i915_mm.c
147
offset -= r.sgt.max >> PAGE_SHIFT;
drivers/gpu/drm/i915/i915_mm.c
152
r.sgt.curr = offset << PAGE_SHIFT;
drivers/gpu/drm/i915/i915_mm.c
159
zap_special_vma_range(vma, addr, r.pfn << PAGE_SHIFT);
drivers/gpu/drm/i915/i915_mm.c
46
return (r->sgt.dma + r->sgt.curr + r->iobase) >> PAGE_SHIFT;
drivers/gpu/drm/i915/i915_mm.c
48
return r->sgt.pfn + (r->sgt.curr >> PAGE_SHIFT);
drivers/gpu/drm/i915/i915_ptr_util.h
38
#define page_mask_bits(ptr) ptr_mask_bits(ptr, PAGE_SHIFT)
drivers/gpu/drm/i915/i915_ptr_util.h
39
#define page_unmask_bits(ptr) ptr_unmask_bits(ptr, PAGE_SHIFT)
drivers/gpu/drm/i915/i915_ptr_util.h
40
#define page_pack_bits(ptr, bits) ptr_pack_bits(ptr, bits, PAGE_SHIFT)
drivers/gpu/drm/i915/i915_ptr_util.h
41
#define page_unpack_bits(ptr, bits) ptr_unpack_bits(ptr, bits, PAGE_SHIFT)
drivers/gpu/drm/i915/i915_scatterlist.c
115
block_size = node->size << PAGE_SHIFT;
drivers/gpu/drm/i915/i915_scatterlist.c
116
offset = node->start << PAGE_SHIFT;
drivers/gpu/drm/i915/i915_scatterlist.c
85
const u32 segment_pages = max_segment >> PAGE_SHIFT;
drivers/gpu/drm/i915/i915_scatterlist.c
97
i915_refct_sgt_init(rsgt, node->size << PAGE_SHIFT);
drivers/gpu/drm/i915/i915_scatterlist.h
114
pfn_to_page((__iter).pfn + ((__iter).curr >> PAGE_SHIFT))); \
drivers/gpu/drm/i915/i915_scatterlist.h
52
return sg->length >> PAGE_SHIFT;
drivers/gpu/drm/i915/i915_scatterlist.h
57
return sg_dma_len(sg) >> PAGE_SHIFT;
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
109
gpu_buddy_block_offset(block) >> PAGE_SHIFT;
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
113
(gpu_buddy_block_size(mm, block) >> PAGE_SHIFT);
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
180
gpu_buddy_block_offset(block) >> PAGE_SHIFT;
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
182
(gpu_buddy_block_size(mm, block) >> PAGE_SHIFT);
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
213
gpu_buddy_block_offset(block) >> PAGE_SHIFT;
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
215
(gpu_buddy_block_size(mm, block) >> PAGE_SHIFT);
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
234
(u64)bman->visible_avail << PAGE_SHIFT >> 20);
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
236
(u64)bman->visible_size << PAGE_SHIFT >> 20);
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
238
(u64)bman->visible_reserved << PAGE_SHIFT >> 20);
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
305
bman->visible_size = visible_size >> PAGE_SHIFT;
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
311
ttm_resource_manager_init(man, bdev, bman->mm.size >> PAGE_SHIFT);
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
376
unsigned long fpfn = start >> PAGE_SHIFT;
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
390
unsigned long lpfn = fpfn + (size >> PAGE_SHIFT);
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
428
*avail = bman->mm.avail >> PAGE_SHIFT;
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
74
min_page_size = bo->page_alignment << PAGE_SHIFT;
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
79
if (size > lpfn << PAGE_SHIFT) {
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
93
err = gpu_buddy_alloc_blocks(mm, (u64)place->fpfn << PAGE_SHIFT,
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
94
(u64)lpfn << PAGE_SHIFT,
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
95
(u64)n_pages << PAGE_SHIFT,
drivers/gpu/drm/i915/i915_vma.c
1132
sg = add_padding_pages(left >> PAGE_SHIFT, st, sg);
drivers/gpu/drm/i915/i915_vma.c
1155
len = min(sg_dma_len(iter) - (offset << PAGE_SHIFT),
drivers/gpu/drm/i915/i915_vma.c
1156
count << PAGE_SHIFT);
drivers/gpu/drm/i915/i915_vma.c
1159
sg_dma_address(iter) + (offset << PAGE_SHIFT);
drivers/gpu/drm/i915/i915_vma.c
1163
count -= len >> PAGE_SHIFT;
drivers/gpu/drm/i915/i915_vma.c
189
obj->base.size >> PAGE_SHIFT));
drivers/gpu/drm/i915/i915_vma.c
191
vma->size <<= PAGE_SHIFT;
drivers/gpu/drm/i915/i915_vma.c
1942
vma_offset = vma->gtt_view.partial.offset << PAGE_SHIFT;
drivers/gpu/drm/i915/i915_vma.c
195
vma->size <<= PAGE_SHIFT;
drivers/gpu/drm/i915/i915_vma.c
198
vma->size <<= PAGE_SHIFT;
drivers/gpu/drm/i915/intel_memory_region.c
316
*avail <<= PAGE_SHIFT;
drivers/gpu/drm/i915/intel_memory_region.c
317
*visible_avail <<= PAGE_SHIFT;
drivers/gpu/drm/i915/intel_memory_region.c
96
return get_random_u32_below(last >> PAGE_SHIFT) << PAGE_SHIFT;
drivers/gpu/drm/i915/intel_region_ttm.c
212
if (WARN_ON(overflows_type(offset >> PAGE_SHIFT, place.fpfn))) {
drivers/gpu/drm/i915/intel_region_ttm.c
216
place.fpfn = offset >> PAGE_SHIFT;
drivers/gpu/drm/i915/intel_region_ttm.c
217
if (WARN_ON(overflows_type(place.fpfn + (size >> PAGE_SHIFT), place.lpfn))) {
drivers/gpu/drm/i915/intel_region_ttm.c
221
place.lpfn = place.fpfn + (size >> PAGE_SHIFT);
drivers/gpu/drm/i915/intel_region_ttm.c
227
if (WARN_ON(overflows_type(resource_size(&mem->io) >> PAGE_SHIFT, place.lpfn))) {
drivers/gpu/drm/i915/intel_region_ttm.c
231
place.lpfn = resource_size(&mem->io) >> PAGE_SHIFT;
drivers/gpu/drm/i915/selftests/i915_gem_gtt.c
1465
const u64 limit = totalram_pages() << PAGE_SHIFT;
drivers/gpu/drm/i915/selftests/i915_gem_gtt.c
180
limit = totalram_pages() << PAGE_SHIFT;
drivers/gpu/drm/i915/selftests/i915_gem_gtt.c
433
const u64 full_size = npages << PAGE_SHIFT;
drivers/gpu/drm/i915/selftests/i915_gem_gtt.c
641
min_t(u64, ULONG_MAX - 1, hole_size >> PAGE_SHIFT);
drivers/gpu/drm/i915/selftests/i915_gem_gtt.c
660
obj = fake_dma_object(vm->i915, size << PAGE_SHIFT);
drivers/gpu/drm/i915/selftests/igt_mmap.c
39
addr = vm_mmap(file, 0, drm_vma_node_size(node) << PAGE_SHIFT,
drivers/gpu/drm/i915/selftests/intel_memory_region.c
672
vma->size >> PAGE_SHIFT, value);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
677
unsigned long n = obj->base.size >> PAGE_SHIFT;
drivers/gpu/drm/i915/selftests/scatterlist.c
199
BIT(12) >> PAGE_SHIFT,
drivers/gpu/drm/i915/selftests/scatterlist.c
200
BIT(16) >> PAGE_SHIFT,
drivers/gpu/drm/i915/selftests/scatterlist.c
201
BIT(21) >> PAGE_SHIFT,
drivers/gpu/drm/imagination/pvr_fw_mips.c
81
u32 src_page_nr = (page_nr * ROGUE_MIPSFW_PAGE_SIZE_4K) >> PAGE_SHIFT;
drivers/gpu/drm/imagination/pvr_mmu.h
52
#define PVR_DEVICE_PAGE_SHIFT (PAGE_SHIFT)
drivers/gpu/drm/imagination/pvr_vm_mips.c
66
mips_data->pt = vmap(mips_data->pt_pages, pt_size >> PAGE_SHIFT, VM_MAP,
drivers/gpu/drm/lima/lima_gem.c
28
(lima_heap_init_nr_pages << PAGE_SHIFT);
drivers/gpu/drm/lima/lima_gem.c
42
pages = kvmalloc_objs(*pages, bo->base.base.size >> PAGE_SHIFT,
drivers/gpu/drm/lima/lima_gem.c
55
for (i = old_size >> PAGE_SHIFT; i < new_size >> PAGE_SHIFT; i++) {
drivers/gpu/drm/lima/lima_gem.c
90
ret = lima_vm_map_bo(vm, bo, old_size >> PAGE_SHIFT);
drivers/gpu/drm/lima/lima_sched.c
365
data = vmap(bo->base.pages, bo->heap_size >> PAGE_SHIFT,
drivers/gpu/drm/lima/lima_vm.c
300
base = bo_va->node.start + (pageoff << PAGE_SHIFT);
drivers/gpu/drm/loongson/lsdc_gem.c
217
drm_err(ddev, "Requesting(%zuMiB) failed\n", (size_t)(args->size >> PAGE_SHIFT));
drivers/gpu/drm/loongson/lsdc_ttm.c
231
mem->bus.offset = (mem->start << PAGE_SHIFT) + ldev->vram_base;
drivers/gpu/drm/loongson/lsdc_ttm.c
267
return resource->start << PAGE_SHIFT;
drivers/gpu/drm/loongson/lsdc_ttm.c
553
num_vram_pages = ldev->vram_size >> PAGE_SHIFT;
drivers/gpu/drm/loongson/lsdc_ttm.c
564
num_gtt_pages = ldev->gtt_size >> PAGE_SHIFT;
drivers/gpu/drm/mgag200/mgag200_mode.c
793
max_fbpages = mdev->vram_available >> PAGE_SHIFT;
drivers/gpu/drm/msm/msm_gem.c
196
size_t npages = obj->size >> PAGE_SHIFT;
drivers/gpu/drm/msm/msm_gem.c
363
pgoff = (vmf->address - vma->vm_start) >> PAGE_SHIFT;
drivers/gpu/drm/msm/msm_gem.c
368
pfn, pfn << PAGE_SHIFT);
drivers/gpu/drm/msm/msm_gem.c
734
msm_obj->vaddr = vmap(pages, obj->size >> PAGE_SHIFT,
drivers/gpu/drm/msm/msm_gem_prime.c
18
size_t npages = obj->size >> PAGE_SHIFT;
drivers/gpu/drm/nouveau/nouveau_bo.c
1098
nvbo->offset = (new_reg->start << PAGE_SHIFT);
drivers/gpu/drm/nouveau/nouveau_bo.c
1109
u64 offset = new_reg->start << PAGE_SHIFT;
drivers/gpu/drm/nouveau/nouveau_bo.c
1272
reg->bus.offset = (reg->start << PAGE_SHIFT) +
drivers/gpu/drm/nouveau/nouveau_bo.c
1286
reg->bus.offset = (reg->start << PAGE_SHIFT) +
drivers/gpu/drm/nouveau/nouveau_bo.c
1378
u32 mappable = device->func->resource_size(device, NVKM_BAR1_FB) >> PAGE_SHIFT;
drivers/gpu/drm/nouveau/nouveau_bo.c
282
(!vmm->page[i].host || vmm->page[i].shift > PAGE_SHIFT))
drivers/gpu/drm/nouveau/nouveau_bo.c
322
(!vmm->page[i].host || vmm->page[i].shift > PAGE_SHIFT))
drivers/gpu/drm/nouveau/nouveau_bo.c
360
&nvbo->placement, align >> PAGE_SHIFT, &ctx,
drivers/gpu/drm/nouveau/nouveau_bo.c
498
fpfn = (vram_size / 2) >> PAGE_SHIFT;
drivers/gpu/drm/nouveau/nouveau_bo.c
502
lpfn = (vram_size / 2) >> PAGE_SHIFT;
drivers/gpu/drm/nouveau/nouveau_bo0039.c
52
u32 src_offset = old_reg->start << PAGE_SHIFT;
drivers/gpu/drm/nouveau/nouveau_bo0039.c
54
u32 dst_offset = new_reg->start << PAGE_SHIFT;
drivers/gpu/drm/nouveau/nouveau_dmem.c
112
unsigned long off = (page_to_pfn(page) << PAGE_SHIFT) -
drivers/gpu/drm/nouveau/nouveau_dmem.c
343
pfn_first = chunk->pagemap.range.start >> PAGE_SHIFT;
drivers/gpu/drm/nouveau/nouveau_dmem.c
477
unsigned long i, npages = range_len(&chunk->pagemap.range) >> PAGE_SHIFT;
drivers/gpu/drm/nouveau/nouveau_dmem.c
486
migrate_device_range(src_pfns, chunk->pagemap.range.start >> PAGE_SHIFT,
drivers/gpu/drm/nouveau/nouveau_dmem.c
52
#define DMEM_CHUNK_NPAGES (DMEM_CHUNK_SIZE >> PAGE_SHIFT)
drivers/gpu/drm/nouveau/nouveau_dmem.c
766
((paddr >> PAGE_SHIFT) << NVIF_VMM_PFNMAP_V0_ADDR_SHIFT);
drivers/gpu/drm/nouveau/nouveau_dmem.c
828
unsigned long npages = (end - start) >> PAGE_SHIFT;
drivers/gpu/drm/nouveau/nouveau_dmem.c
867
if (args.start + (max << PAGE_SHIFT) > end)
drivers/gpu/drm/nouveau/nouveau_dmem.c
870
args.end = args.start + (max << PAGE_SHIFT);
drivers/gpu/drm/nouveau/nouveau_mem.c
117
ret = nvif_mem_ctor_type(mmu, "ttmHostMem", mmu->mem, type, PAGE_SHIFT,
drivers/gpu/drm/nouveau/nouveau_mem.c
159
reg->start = mem->mem.addr >> PAGE_SHIFT;
drivers/gpu/drm/nouveau/nouveau_svm.c
568
PAGE_SHIFT;
drivers/gpu/drm/nouveau/nouveau_svm.c
571
page -= (addr - args->p.addr) >> PAGE_SHIFT;
drivers/gpu/drm/nouveau/nouveau_svm.c
803
args->p.page = PAGE_SHIFT;
drivers/gpu/drm/nouveau/nouveau_svm.c
909
args->p.page = PAGE_SHIFT;
drivers/gpu/drm/nouveau/nouveau_ttm.c
148
(*res)->start = mem->vma[0].addr >> PAGE_SHIFT;
drivers/gpu/drm/nouveau/nouveau_ttm.c
192
drm->gem.vram_available >> PAGE_SHIFT);
drivers/gpu/drm/nouveau/nouveau_ttm.c
198
drm->gem.vram_available >> PAGE_SHIFT);
drivers/gpu/drm/nouveau/nouveau_ttm.c
221
unsigned long size_pages = drm->gem.gart_available >> PAGE_SHIFT;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
497
(!vmm->page[i].host || vmm->page[i].shift > PAGE_SHIFT))
drivers/gpu/drm/nouveau/nouveau_uvmm.c
510
return PAGE_SHIFT;
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
166
return PAGE_SHIFT;
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
261
int pages = len >> PAGE_SHIFT;
drivers/gpu/drm/nouveau/nvkm/engine/device/tegra.c
134
tdev->iommu.pgshift = PAGE_SHIFT;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
209
node->base.vaddr = vmap(node->pages, size >> PAGE_SHIFT, VM_MAP,
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
304
dma_free_attrs(dev, (u64)node->base.mn->length << PAGE_SHIFT,
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
397
node->base.vaddr = dma_alloc_attrs(dev, npages << PAGE_SHIFT,
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
414
node->r.length = (npages << PAGE_SHIFT) >> 12;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
531
ret = gk20a_instobj_ctor_iommu(imem, size >> PAGE_SHIFT,
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
534
ret = gk20a_instobj_ctor_dma(imem, size >> PAGE_SHIFT,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
163
if (page != PAGE_SHIFT)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
183
mem->pages = size >> PAGE_SHIFT;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
192
size = ALIGN(size, PAGE_SIZE) >> PAGE_SHIFT;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
51
return PAGE_SHIFT;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
66
return nvkm_mem(memory)->pages << PAGE_SHIFT;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c
173
page = max_t(u8, page, PAGE_SHIFT);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1570
map->dma += map->offset >> PAGE_SHIFT;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
68
if (map->page->shift == PAGE_SHIFT) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
35
if (map->page->shift == PAGE_SHIFT) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
157
if (map->page->shift == PAGE_SHIFT) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c
49
#if PAGE_SHIFT == 12
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv41.c
48
#if PAGE_SHIFT == 12
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
115
#if PAGE_SHIFT == 12
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
68
if (map->page->shift == PAGE_SHIFT) {
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
575
int num_pages = (size + PAGE_SIZE - 1) >> PAGE_SHIFT;
drivers/gpu/drm/omapdrm/omap_fbdev.c
51
npages = fbi->fix.line_length >> PAGE_SHIFT;
drivers/gpu/drm/omapdrm/omap_gem.c
1011
count = obj->size >> PAGE_SHIFT;
drivers/gpu/drm/omapdrm/omap_gem.c
1081
omap_obj->vaddr = vmap(omap_obj->pages, obj->size >> PAGE_SHIFT,
drivers/gpu/drm/omapdrm/omap_gem.c
1109
u32 npages = obj->size >> PAGE_SHIFT;
drivers/gpu/drm/omapdrm/omap_gem.c
1488
usergart[i].stride_pfn = tiler_stride(fmts[i], 0) >> PAGE_SHIFT;
drivers/gpu/drm/omapdrm/omap_gem.c
1507
usergart[i].stride_pfn << PAGE_SHIFT);
drivers/gpu/drm/omapdrm/omap_gem.c
182
(entry->obj_pgoff << PAGE_SHIFT);
drivers/gpu/drm/omapdrm/omap_gem.c
234
int npages = obj->size >> PAGE_SHIFT;
drivers/gpu/drm/omapdrm/omap_gem.c
305
unsigned int npages = obj->size >> PAGE_SHIFT;
drivers/gpu/drm/omapdrm/omap_gem.c
362
pgoff = (vmf->address - vma->vm_start) >> PAGE_SHIFT;
drivers/gpu/drm/omapdrm/omap_gem.c
369
pfn = (omap_obj->dma_addr >> PAGE_SHIFT) + pgoff;
drivers/gpu/drm/omapdrm/omap_gem.c
373
pfn, pfn << PAGE_SHIFT);
drivers/gpu/drm/omapdrm/omap_gem.c
410
pgoff = (vmf->address - vma->vm_start) >> PAGE_SHIFT;
drivers/gpu/drm/omapdrm/omap_gem.c
421
vaddr = vmf->address - ((pgoff - base_pgoff) << PAGE_SHIFT);
drivers/gpu/drm/omapdrm/omap_gem.c
442
vaddr += off << PAGE_SHIFT;
drivers/gpu/drm/omapdrm/omap_gem.c
464
pfn = entry->dma_addr >> PAGE_SHIFT;
drivers/gpu/drm/omapdrm/omap_gem.c
467
pfn, pfn << PAGE_SHIFT);
drivers/gpu/drm/omapdrm/omap_gem.c
637
u32 npages = obj->size >> PAGE_SHIFT;
drivers/gpu/drm/omapdrm/omap_gem.c
713
int i, npages = obj->size >> PAGE_SHIFT;
drivers/gpu/drm/omapdrm/omap_gem.c
746
u32 npages = obj->size >> PAGE_SHIFT;
drivers/gpu/drm/panfrost/panfrost_drv.c
160
args->offset = mapping->mmnode.start << PAGE_SHIFT;
drivers/gpu/drm/panfrost/panfrost_drv.c
454
args->offset = mapping->mmnode.start << PAGE_SHIFT;
drivers/gpu/drm/panfrost/panfrost_dump.c
142
n_bomap_pages += dbo->size >> PAGE_SHIFT;
drivers/gpu/drm/panfrost/panfrost_dump.c
226
iter.hdr->bomap.iova = mapping->mmnode.start << PAGE_SHIFT;
drivers/gpu/drm/panfrost/panfrost_gem.c
194
align = size >> PAGE_SHIFT;
drivers/gpu/drm/panfrost/panfrost_gem.c
196
align = size >= SZ_2M ? SZ_2M >> PAGE_SHIFT : 0;
drivers/gpu/drm/panfrost/panfrost_gem.c
201
size >> PAGE_SHIFT, align, color, 0);
drivers/gpu/drm/panfrost/panfrost_gem_shrinker.c
30
count += shmem->base.size >> PAGE_SHIFT;
drivers/gpu/drm/panfrost/panfrost_gem_shrinker.c
79
freed += shmem->base.size >> PAGE_SHIFT;
drivers/gpu/drm/panfrost/panfrost_gem_shrinker.c
87
pr_info_ratelimited("Purging %lu bytes\n", freed << PAGE_SHIFT);
drivers/gpu/drm/panfrost/panfrost_mmu.c
481
ret = mmu_map_sg(pfdev, mapping->mmu, mapping->mmnode.start << PAGE_SHIFT,
drivers/gpu/drm/panfrost/panfrost_mmu.c
501
u64 iova = mapping->mmnode.start << PAGE_SHIFT;
drivers/gpu/drm/panfrost/panfrost_mmu.c
502
size_t len = mapping->mmnode.size << PAGE_SHIFT;
drivers/gpu/drm/panfrost/panfrost_mmu.c
526
mapping->mmnode.start << PAGE_SHIFT, len);
drivers/gpu/drm/panfrost/panfrost_mmu.c
555
u64 offset = addr >> PAGE_SHIFT;
drivers/gpu/drm/panfrost/panfrost_mmu.c
606
bomapping->mmnode.start << PAGE_SHIFT);
drivers/gpu/drm/panfrost/panfrost_mmu.c
614
page_offset = addr >> PAGE_SHIFT;
drivers/gpu/drm/panfrost/panfrost_mmu.c
616
nr_pages = bo->base.base.size >> PAGE_SHIFT;
drivers/gpu/drm/panfrost/panfrost_mmu.c
752
#define PFN_4G (SZ_4G >> PAGE_SHIFT)
drivers/gpu/drm/panfrost/panfrost_mmu.c
754
#define PFN_16M (SZ_16M >> PAGE_SHIFT)
drivers/gpu/drm/panfrost/panfrost_mmu.c
805
drm_mm_init(&mmu->mm, SZ_32M >> PAGE_SHIFT, (SZ_4G - SZ_32M) >> PAGE_SHIFT);
drivers/gpu/drm/panfrost/panfrost_perfcnt.c
54
gpuva = pfdev->perfcnt->mapping->mmnode.start << PAGE_SHIFT;
drivers/gpu/drm/panthor/panthor_device.c
413
u64 offset = (u64)vma->vm_pgoff << PAGE_SHIFT;
drivers/gpu/drm/panthor/panthor_device.c
457
u64 offset = (u64)vma->vm_pgoff << PAGE_SHIFT;
drivers/gpu/drm/panthor/panthor_drv.c
1671
u64 offset = (u64)vma->vm_pgoff << PAGE_SHIFT;
drivers/gpu/drm/panthor/panthor_drv.c
1687
vma->vm_pgoff = offset >> PAGE_SHIFT;
drivers/gpu/drm/panthor/panthor_mmu.c
2113
pg = to_panthor_bo(op->gem.obj)->base.pages[bo_offset >> PAGE_SHIFT];
drivers/gpu/drm/qxl/qxl_drv.h
289
return slot->high_bits | ((bo->tbo.resource->start << PAGE_SHIFT) + offset);
drivers/gpu/drm/qxl/qxl_image.c
148
ptr = qxl_bo_kmap_atomic_page(qdev, chunk_bo, page << PAGE_SHIFT);
drivers/gpu/drm/qxl/qxl_object.c
222
offset = bo->tbo.resource->start << PAGE_SHIFT;
drivers/gpu/drm/qxl/qxl_ttm.c
83
mem->bus.offset = (mem->start << PAGE_SHIFT) + qdev->vram_base;
drivers/gpu/drm/qxl/qxl_ttm.c
88
mem->bus.offset = (mem->start << PAGE_SHIFT) +
drivers/gpu/drm/radeon/radeon_agp.c
211
info->memory_allowed = kern->max_memory << PAGE_SHIFT;
drivers/gpu/drm/radeon/radeon_agp.c
212
info->memory_used = kern->current_memory << PAGE_SHIFT;
drivers/gpu/drm/radeon/radeon_gart.c
109
rdev->gart.table_size >> PAGE_SHIFT);
drivers/gpu/drm/radeon/radeon_gart.c
84
rdev->gart.table_size >> PAGE_SHIFT);
drivers/gpu/drm/radeon/radeon_gem.c
303
args->vram_size = (u64)man->size << PAGE_SHIFT;
drivers/gpu/drm/radeon/radeon_object.c
122
rbo->rdev->mc.visible_vram_size >> PAGE_SHIFT;
drivers/gpu/drm/radeon/radeon_object.c
136
unsigned long page_align = roundup(byte_align, PAGE_SIZE) >> PAGE_SHIFT;
drivers/gpu/drm/radeon/radeon_object.c
311
bo->rdev->mc.visible_vram_size >> PAGE_SHIFT;
drivers/gpu/drm/radeon/radeon_object.c
313
bo->placements[i].lpfn = max_offset >> PAGE_SHIFT;
drivers/gpu/drm/radeon/radeon_object.c
584
bo->tbo.resource->start << PAGE_SHIFT,
drivers/gpu/drm/radeon/radeon_object.c
734
offset = bo->resource->start << PAGE_SHIFT;
drivers/gpu/drm/radeon/radeon_object.c
744
lpfn = rdev->mc.visible_vram_size >> PAGE_SHIFT;
drivers/gpu/drm/radeon/radeon_object.c
756
offset = bo->resource->start << PAGE_SHIFT;
drivers/gpu/drm/radeon/radeon_object.c
88
rbo->rdev->mc.visible_vram_size >> PAGE_SHIFT;
drivers/gpu/drm/radeon/radeon_object.h
107
return (bo->tbo.resource->start << PAGE_SHIFT) + start;
drivers/gpu/drm/radeon/radeon_object.h
122
return (bo->tbo.page_alignment << PAGE_SHIFT) / RADEON_GPU_PAGE_SIZE;
drivers/gpu/drm/radeon/radeon_ttm.c
105
bo->resource->start < (rbo->rdev->mc.visible_vram_size >> PAGE_SHIFT)) {
drivers/gpu/drm/radeon/radeon_ttm.c
106
unsigned fpfn = rbo->rdev->mc.visible_vram_size >> PAGE_SHIFT;
drivers/gpu/drm/radeon/radeon_ttm.c
146
old_start = (u64)old_mem->start << PAGE_SHIFT;
drivers/gpu/drm/radeon/radeon_ttm.c
147
new_start = (u64)new_mem->start << PAGE_SHIFT;
drivers/gpu/drm/radeon/radeon_ttm.c
269
mem->bus.offset = (mem->start << PAGE_SHIFT) +
drivers/gpu/drm/radeon/radeon_ttm.c
277
mem->bus.offset = mem->start << PAGE_SHIFT;
drivers/gpu/drm/radeon/radeon_ttm.c
362
(u64)ttm->num_pages << PAGE_SHIFT,
drivers/gpu/drm/radeon/radeon_ttm.c
438
gtt->offset = (unsigned long)(bo_mem->start << PAGE_SHIFT);
drivers/gpu/drm/radeon/radeon_ttm.c
73
false, rdev->mc.real_vram_size >> PAGE_SHIFT);
drivers/gpu/drm/radeon/radeon_ttm.c
769
man->size = size >> PAGE_SHIFT;
drivers/gpu/drm/radeon/radeon_ttm.c
79
true, rdev->mc.gtt_size >> PAGE_SHIFT);
drivers/gpu/drm/radeon/radeon_uvd.c
310
rbo->placements[i].fpfn = 0 >> PAGE_SHIFT;
drivers/gpu/drm/radeon/radeon_uvd.c
311
rbo->placements[i].lpfn = (256 * 1024 * 1024) >> PAGE_SHIFT;
drivers/gpu/drm/radeon/radeon_uvd.c
324
rbo->placements[1].fpfn += (256 * 1024 * 1024) >> PAGE_SHIFT;
drivers/gpu/drm/radeon/radeon_uvd.c
325
rbo->placements[1].lpfn += (256 * 1024 * 1024) >> PAGE_SHIFT;
drivers/gpu/drm/radeon/radeon_vm.c
945
addr = (u64)mem->start << PAGE_SHIFT;
drivers/gpu/drm/rockchip/rockchip_drm_gem.c
221
unsigned int count = obj->size >> PAGE_SHIFT;
drivers/gpu/drm/rockchip/rockchip_drm_gem.c
90
rk_obj->num_pages = rk_obj->base.size >> PAGE_SHIFT;
drivers/gpu/drm/tegra/gem.c
351
bo->num_pages = bo->gem.size >> PAGE_SHIFT;
drivers/gpu/drm/tegra/gem.c
573
offset = (vmf->address - vma->vm_start) >> PAGE_SHIFT;
drivers/gpu/drm/tests/drm_gem_shmem_test.c
152
for (i = 0; i < (shmem->base.size >> PAGE_SHIFT); i++)
drivers/gpu/drm/ttm/tests/ttm_tt_test.c
340
int expected_num_pages = BO_SIZE >> PAGE_SHIFT;
drivers/gpu/drm/ttm/tests/ttm_tt_test.c
47
int num_pages = params->size >> PAGE_SHIFT;
drivers/gpu/drm/ttm/tests/ttm_tt_test.c
73
int num_pages = (size + SZ_4K) >> PAGE_SHIFT;
drivers/gpu/drm/ttm/ttm_backup.c
168
return (u64)get_nr_swap_pages() << PAGE_SHIFT;
drivers/gpu/drm/ttm/ttm_backup.c
35
start <<= PAGE_SHIFT;
drivers/gpu/drm/ttm/ttm_bo_util.c
445
offset = start_page << PAGE_SHIFT;
drivers/gpu/drm/ttm/ttm_bo_util.c
446
size = num_pages << PAGE_SHIFT;
drivers/gpu/drm/ttm/ttm_bo_vm.c
211
page_offset = ((address - vma->vm_start) >> PAGE_SHIFT) +
drivers/gpu/drm/ttm/ttm_bo_vm.c
372
unsigned long page = offset >> PAGE_SHIFT;
drivers/gpu/drm/ttm/ttm_bo_vm.c
379
offset -= page << PAGE_SHIFT;
drivers/gpu/drm/ttm/ttm_bo_vm.c
467
<< PAGE_SHIFT);
drivers/gpu/drm/ttm/ttm_bo_vm.c
94
return (bo->resource->bus.offset >> PAGE_SHIFT) + page_offset;
drivers/gpu/drm/ttm/ttm_device.c
91
num_pages = ((u64)si.totalram * si.mem_unit) >> PAGE_SHIFT;
drivers/gpu/drm/ttm/ttm_device.c
96
>> PAGE_SHIFT;
drivers/gpu/drm/ttm/ttm_device.c
97
num_dma32 = min(num_dma32, 2UL << (30 - PAGE_SHIFT));
drivers/gpu/drm/ttm/ttm_pool.c
1388
pool_node_limit[nid] = (node_size >> PAGE_SHIFT) / 2;
drivers/gpu/drm/ttm/ttm_pool.c
295
dma_unmap_page(pool->dev, dma_addr, (long)num_pages << PAGE_SHIFT,
drivers/gpu/drm/ttm/ttm_resource.c
734
PAGE_SHIFT;
drivers/gpu/drm/ttm/ttm_resource.c
747
<< PAGE_SHIFT));
drivers/gpu/drm/ttm/ttm_tt.c
159
ttm->num_pages = (PAGE_ALIGN(bo->base.size) >> PAGE_SHIFT) + extra_pages;
drivers/gpu/drm/ttm/ttm_tt.c
325
loff_t size = (loff_t)ttm->num_pages << PAGE_SHIFT;
drivers/gpu/drm/ttm/ttm_tt.c
551
ttm_backup_shmem_create(((loff_t)tt->num_pages) << PAGE_SHIFT);
drivers/gpu/drm/v3d/v3d_bo.c
198
bo->vaddr = vmap(obj->pages, obj->base.size >> PAGE_SHIFT, VM_MAP,
drivers/gpu/drm/virtio/virtgpu_vram.c
60
if (check_add_overflow(vma->vm_pgoff << PAGE_SHIFT, vm_size, &vm_end))
drivers/gpu/drm/virtio/virtgpu_vram.c
67
(vram->vram_node.start >> PAGE_SHIFT) + vma->vm_pgoff,
drivers/gpu/drm/vmwgfx/vmwgfx_blit.c
368
u32 dst_page = dst_offset >> PAGE_SHIFT;
drivers/gpu/drm/vmwgfx/vmwgfx_blit.c
369
u32 src_page = src_offset >> PAGE_SHIFT;
drivers/gpu/drm/vmwgfx/vmwgfx_bo.c
282
ptr->offset = bo->resource->start << PAGE_SHIFT;
drivers/gpu/drm/vmwgfx/vmwgfx_cmd.c
588
cmd->body.guestResult.offset = bo->resource->start << PAGE_SHIFT;
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c
1265
drm_mm_init(&man->mm, 0, size >> PAGE_SHIFT);
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c
888
header->size = header->node.size << PAGE_SHIFT;
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c
890
offset = header->node.start << PAGE_SHIFT;
drivers/gpu/drm/vmwgfx/vmwgfx_cursor_plane.c
375
kmap_offset = cmd->dma.guest.ptr.offset >> PAGE_SHIFT;
drivers/gpu/drm/vmwgfx/vmwgfx_cursor_plane.c
376
kmap_num = (VMW_CURSOR_SNOOP_HEIGHT * image_pitch) >> PAGE_SHIFT;
drivers/gpu/drm/vmwgfx/vmwgfx_drv.c
696
dev_priv->vram_size >> PAGE_SHIFT);
drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c
3765
reloc->location->offset += bo->resource->start << PAGE_SHIFT;
drivers/gpu/drm/vmwgfx/vmwgfx_gmr.c
85
*cmd = vmw_piter_dma_addr(iter) >> PAGE_SHIFT;
drivers/gpu/drm/vmwgfx/vmwgfx_gmr.c
88
PAGE_SHIFT;
drivers/gpu/drm/vmwgfx/vmwgfx_gmrid_manager.c
105
((gman->max_gmr_pages) << (PAGE_SHIFT - 10)));
drivers/gpu/drm/vmwgfx/vmwgfx_gmrid_manager.c
98
((new_max_pages) << (PAGE_SHIFT - 10)));
drivers/gpu/drm/vmwgfx/vmwgfx_mob.c
132
vmw_piter_start(&iter, vsgt, offset >> PAGE_SHIFT);
drivers/gpu/drm/vmwgfx/vmwgfx_mob.c
135
mob = vmw_mob_create(otable->size >> PAGE_SHIFT);
drivers/gpu/drm/vmwgfx/vmwgfx_mob.c
149
vmw_mob_pt_setup(mob, iter, otable->size >> PAGE_SHIFT);
drivers/gpu/drm/vmwgfx/vmwgfx_mob.c
163
cmd->body.baseAddress = mob->pt_root_page >> PAGE_SHIFT;
drivers/gpu/drm/vmwgfx/vmwgfx_mob.c
383
return tot_size >> PAGE_SHIFT;
drivers/gpu/drm/vmwgfx/vmwgfx_mob.c
435
*((u64 *) *addr) = val >> PAGE_SHIFT;
drivers/gpu/drm/vmwgfx/vmwgfx_mob.c
441
*(*addr)++ = val >> PAGE_SHIFT;
drivers/gpu/drm/vmwgfx/vmwgfx_mob.c
640
cmd->body.base = mob->pt_root_page >> PAGE_SHIFT;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
202
unmap_shared_mapping_range(mapping, (offset + start) << PAGE_SHIFT,
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
203
(loff_t) (end - start) << PAGE_SHIFT);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
307
res_start >>= PAGE_SHIFT;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
343
res_start >>= PAGE_SHIFT;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
385
res_start >>= PAGE_SHIFT;
drivers/gpu/drm/vmwgfx/vmwgfx_resource.c
1072
res->func->dirty_range_add(res, start << PAGE_SHIFT,
drivers/gpu/drm/vmwgfx/vmwgfx_resource.c
1073
end << PAGE_SHIFT);
drivers/gpu/drm/vmwgfx/vmwgfx_resource.c
1105
unsigned long res_start = start << PAGE_SHIFT;
drivers/gpu/drm/vmwgfx/vmwgfx_resource.c
1106
unsigned long res_end = end << PAGE_SHIFT;
drivers/gpu/drm/vmwgfx/vmwgfx_resource.c
417
pgoff_t start = res->guest_memory_offset >> PAGE_SHIFT;
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
1113
pgoff_t start = res->guest_memory_offset >> PAGE_SHIFT;
drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c
197
(unsigned long)vsgt->num_pages << PAGE_SHIFT,
drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c
449
mem->bus.offset = (mem->start << PAGE_SHIFT) +
drivers/gpu/drm/xe/display/xe_panic.c
50
new_page = offset >> PAGE_SHIFT;
drivers/gpu/drm/xe/tests/xe_bo.c
91
ccs_page = xe_bo_ccs_pages_start(bo) >> PAGE_SHIFT;
drivers/gpu/drm/xe/xe_bo.c
1643
return xe_ttm_stolen_io_offset(bo, page_offset << PAGE_SHIFT) >> PAGE_SHIFT;
drivers/gpu/drm/xe/xe_bo.c
1646
xe_res_first(ttm_bo->resource, (u64)page_offset << PAGE_SHIFT, 0, &cursor);
drivers/gpu/drm/xe/xe_bo.c
1647
return (vram->io_start + cursor.start) >> PAGE_SHIFT;
drivers/gpu/drm/xe/xe_bo.c
2338
alignment = align >> PAGE_SHIFT;
drivers/gpu/drm/xe/xe_bo.c
2342
alignment = SZ_4K >> PAGE_SHIFT;
drivers/gpu/drm/xe/xe_bo.c
2460
place->fpfn = start >> PAGE_SHIFT;
drivers/gpu/drm/xe/xe_bo.c
2461
place->lpfn = end >> PAGE_SHIFT;
drivers/gpu/drm/xe/xe_bo.c
265
place.lpfn = io_size >> PAGE_SHIFT;
drivers/gpu/drm/xe/xe_bo.c
3162
page = offset >> PAGE_SHIFT;
drivers/gpu/drm/xe/xe_bo.c
3168
xe_res_first_sg(xe_bo_sg(bo), page << PAGE_SHIFT,
drivers/gpu/drm/xe/xe_bo.c
3174
xe_res_first(bo->ttm.resource, page << PAGE_SHIFT,
drivers/gpu/drm/xe/xe_bo.c
3210
ret = ttm_bo_kmap(&bo->ttm, 0, xe_bo_size(bo) >> PAGE_SHIFT, &bo->kmap);
drivers/gpu/drm/xe/xe_bo.c
397
(u64)num_pages << PAGE_SHIFT,
drivers/gpu/drm/xe/xe_bo.c
468
global_total_pages << PAGE_SHIFT);
drivers/gpu/drm/xe/xe_bo.c
643
mem->bus.offset = mem->start << PAGE_SHIFT;
drivers/gpu/drm/xe/xe_dma_buf.c
123
obj->size >> PAGE_SHIFT);
drivers/gpu/drm/xe/xe_guc.c
109
u32 offset = guc_bo_ggtt_addr(guc, guc->log.bo) >> PAGE_SHIFT;
drivers/gpu/drm/xe/xe_guc.c
155
u32 ads = guc_bo_ggtt_addr(guc, guc->ads.bo) >> PAGE_SHIFT;
drivers/gpu/drm/xe/xe_migrate.c
2212
batch_size += pte_update_cmd_size(npages << PAGE_SHIFT);
drivers/gpu/drm/xe/xe_migrate.c
2241
sram_addr, npages << PAGE_SHIFT, 1);
drivers/gpu/drm/xe/xe_migrate.c
2244
sram_addr, npages << PAGE_SHIFT, 0);
drivers/gpu/drm/xe/xe_query.c
271
mem_regions->mem_regions[0].total_size = man->size << PAGE_SHIFT;
drivers/gpu/drm/xe/xe_shrinker.c
145
num_pages = ttm_backup_bytes_avail() >> PAGE_SHIFT;
drivers/gpu/drm/xe/xe_svm.c
464
xe_assert(vr->xe, (pfn << PAGE_SHIFT) >= hpa_base);
drivers/gpu/drm/xe/xe_svm.c
466
offset = (pfn << PAGE_SHIFT) - hpa_base;
drivers/gpu/drm/xe/xe_svm.c
799
for (i = 0; i < gpu_buddy_block_size(buddy, block) >> PAGE_SHIFT; ++i)
drivers/gpu/drm/xe/xe_ttm_sys_mgr.c
115
ttm_resource_manager_init(man, &xe->ttm, gtt_size >> PAGE_SHIFT);
drivers/gpu/drm/xe/xe_ttm_sys_mgr.c
44
ttm_resource_manager_usage(man) > (man->size << PAGE_SHIFT)) {
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
112
if (lpfn <= mgr->visible_size >> PAGE_SHIFT && size > mgr->visible_avail) {
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
117
err = gpu_buddy_alloc_blocks(mm, (u64)place->fpfn << PAGE_SHIFT,
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
118
(u64)lpfn << PAGE_SHIFT, size,
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
123
if (lpfn <= mgr->visible_size >> PAGE_SHIFT) {
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
156
vres->base.start = gpu_buddy_block_offset(block) >> PAGE_SHIFT;
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
223
if (!place->fpfn && place->lpfn == mgr->visible_size >> PAGE_SHIFT)
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
228
gpu_buddy_block_offset(block) >> PAGE_SHIFT;
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
230
(gpu_buddy_block_size(mm, block) >> PAGE_SHIFT);
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
253
if (!place->fpfn && place->lpfn == mgr->visible_size >> PAGE_SHIFT)
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
258
gpu_buddy_block_offset(block) >> PAGE_SHIFT;
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
260
(gpu_buddy_block_size(mm, block) >> PAGE_SHIFT);
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
62
if (!lpfn || lpfn > man->size >> PAGE_SHIFT)
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
63
lpfn = man->size >> PAGE_SHIFT;
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
65
if (tbo->base.size >> PAGE_SHIFT > (lpfn - place->fpfn))
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
88
if (place->fpfn || lpfn != man->size >> PAGE_SHIFT)
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
99
min_page_size = (u64)tbo->page_alignment << PAGE_SHIFT;
drivers/gpu/drm/xe/xe_userptr.c
126
xe_vma_size(vma) >> PAGE_SHIFT, &ctx);
drivers/gpu/drm/xe/xe_userptr.c
148
xe_vma_size(vma) >> PAGE_SHIFT, &ctx);
drivers/gpu/drm/xe/xe_userptr.c
415
xe_vma_size(&uvma->vma) >> PAGE_SHIFT);
drivers/hv/channel.c
161
<< PAGE_SHIFT));
drivers/hv/channel.c
189
newchannel->ringbuffer_pagecount = (send_size + recv_size) >> PAGE_SHIFT;
drivers/hv/channel.c
190
newchannel->ringbuffer_send_offset = send_size >> PAGE_SHIFT;
drivers/hv/channel.c
690
(send_pages + recv_pages) << PAGE_SHIFT,
drivers/hv/channel.c
691
newchannel->ringbuffer_send_offset << PAGE_SHIFT,
drivers/hv/channel.c
731
hv_ring_gpadl_send_hvpgoffset(send_pages << PAGE_SHIFT);
drivers/hv/hv_balloon.c
1070
#define MB2PAGES(mb) ((mb) << (20 - PAGE_SHIFT))
drivers/hv/hv_balloon.c
1219
get_order(alloc_unit << PAGE_SHIFT));
drivers/hv/hv_balloon.c
1233
split_page(pg, get_order(alloc_unit << PAGE_SHIFT));
drivers/hv/hv_balloon.c
497
#define HA_BYTES_IN_CHUNK (ha_pages_in_chunk << PAGE_SHIFT)
drivers/hv/mshv_synic.c
482
*event_flags_page = memremap(siefp.base_siefp_gpa << PAGE_SHIFT,
drivers/hv/mshv_synic.c
493
*event_ring_page = memremap(sirbp.base_sirbp_gpa << PAGE_SHIFT,
drivers/hv/mshv_vtl_main.c
1222
is_valid = (vmf->address & mask) == ((vmf->pgoff << PAGE_SHIFT) & mask) &&
drivers/hv/mshv_vtl_main.c
1227
*pfn = vmf->pgoff & ~(mask >> PAGE_SHIFT);
drivers/hv/ring_buffer.c
228
ring_info->ring_size = page_cnt << PAGE_SHIFT;
drivers/hv/vmbus_drv.c
2036
return channel->ringbuffer_pagecount << PAGE_SHIFT;
drivers/hwtracing/coresight/coresight-catu.c
273
size >> PAGE_SHIFT, pages);
drivers/hwtracing/coresight/coresight-etb10.c
414
head = handle->head & ((buf->nr_pages << PAGE_SHIFT) - 1);
drivers/hwtracing/coresight/coresight-tmc-etf.c
461
head = handle->head & (((unsigned long)buf->nr_pages << PAGE_SHIFT) - 1);
drivers/hwtracing/coresight/coresight-tmc-etr.c
1382
size = nr_pages << PAGE_SHIFT;
drivers/hwtracing/coresight/coresight-tmc-etr.c
1590
pg_idx = head >> PAGE_SHIFT;
drivers/hwtracing/coresight/coresight-tmc-etr.c
365
start = offset >> PAGE_SHIFT;
drivers/hwtracing/coresight/coresight-tmc-etr.c
399
int pg_idx = offset >> PAGE_SHIFT;
drivers/hwtracing/coresight/coresight-tmc-etr.c
57
((idx) % ((unsigned long)(buf)->nr_pages << PAGE_SHIFT))
drivers/hwtracing/coresight/coresight-tmc-etr.c
572
int nr_dpages = size >> PAGE_SHIFT;
drivers/hwtracing/coresight/coresight-tmc.h
394
return (unsigned long)sg_table->data_pages.nr_pages << PAGE_SHIFT;
drivers/hwtracing/coresight/coresight-trbe.c
27
((idx) % ((unsigned long)(buf)->nr_pages << PAGE_SHIFT))
drivers/hwtracing/coresight/ultrasoc-smb.c
342
pg_idx = head >> PAGE_SHIFT;
drivers/hwtracing/intel_th/msu.c
1127
nr_blocks << PAGE_SHIFT);
drivers/hwtracing/intel_th/msu.c
1328
ret = msc_buffer_contig_alloc(msc, nr_pages[0] << PAGE_SHIFT);
drivers/hwtracing/intel_th/msu.c
1502
unsigned long size = msc->nr_pages << PAGE_SHIFT, rem = len;
drivers/hwtracing/intel_th/msu.c
1550
size = msc->nr_pages << PAGE_SHIFT;
drivers/hwtracing/intel_th/msu.c
1650
if (size >> PAGE_SHIFT != msc->nr_pages)
drivers/hwtracing/intel_th/msu.c
343
return (size_t)win->nr_blocks << PAGE_SHIFT;
drivers/hwtracing/intel_th/msu.c
789
reg = msc->base_addr >> PAGE_SHIFT;
drivers/hwtracing/intel_th/msu.c
848
msc->single_sz = reg & ((msc->nr_pages << PAGE_SHIFT) - 1);
drivers/hwtracing/intel_th/msu.c
923
unsigned long nr_pages = size >> PAGE_SHIFT;
drivers/hwtracing/intel_th/msu.c
976
for (off = 0; off < msc->nr_pages << PAGE_SHIFT; off += PAGE_SIZE) {
drivers/hwtracing/intel_th/msu.c
998
return virt_to_page(msc->base + (pgoff << PAGE_SHIFT));
drivers/infiniband/core/umem.c
220
lock_limit = rlimit(RLIMIT_MEMLOCK) >> PAGE_SHIFT;
drivers/infiniband/core/umem.c
250
pinned << PAGE_SHIFT, ib_dma_max_seg_size(device),
drivers/infiniband/core/umem_odp.c
103
(end - start) >> PAGE_SHIFT,
drivers/infiniband/core/umem_odp.c
150
umem_odp->page_shift = PAGE_SHIFT;
drivers/infiniband/core/umem_odp.c
193
odp_data->page_shift = PAGE_SHIFT;
drivers/infiniband/core/umem_odp.c
255
umem_odp->page_shift = PAGE_SHIFT;
drivers/infiniband/core/umem_odp.c
356
pfn_start_idx = (range.start - ib_umem_start(umem_odp)) >> PAGE_SHIFT;
drivers/infiniband/core/umem_odp.c
357
num_pfns = (range.end - range.start) >> PAGE_SHIFT;
drivers/infiniband/core/umem_odp.c
391
pfn_index += 1 << (page_shift - PAGE_SHIFT), dma_index++) {
drivers/infiniband/core/umem_odp.c
409
if (hmm_order + PAGE_SHIFT < page_shift) {
drivers/infiniband/core/umem_odp.c
91
nr_entries = (end - start) >> PAGE_SHIFT;
drivers/infiniband/hw/bng_re/bng_fw.c
75
sginfo.pgshft = PAGE_SHIFT;
drivers/infiniband/hw/bng_re/bng_fw.c
752
req.log2_dbr_pg_size = cpu_to_le16(PAGE_SHIFT -
drivers/infiniband/hw/bnxt_re/ib_verbs.c
1157
qplib_qp->sq.sg_info.pgshft = PAGE_SHIFT;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
1170
qplib_qp->rq.sg_info.pgshft = PAGE_SHIFT;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
1264
qp->qplib_qp.sq.sg_info.pgshft = PAGE_SHIFT;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
1276
qp->qplib_qp.rq.sg_info.pgshft = PAGE_SHIFT;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
1336
rq->sg_info.pgshft = PAGE_SHIFT;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
1413
qplqp->sq.sg_info.pgshft = PAGE_SHIFT;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
1880
qplib_srq->sg_info.pgshft = PAGE_SHIFT;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
1936
srq->qplib_srq.sg_info.pgshft = PAGE_SHIFT;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3177
cq->qplib_cq.sg_info.pgshft = PAGE_SHIFT;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3333
cq->qplib_cq.sg_info.pgshft = PAGE_SHIFT;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
4603
pfn = bnxt_entry->mem_offset >> PAGE_SHIFT;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
4609
pfn = bnxt_entry->mem_offset >> PAGE_SHIFT;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
4618
pfn = bnxt_entry->mem_offset >> PAGE_SHIFT;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1101
sginfo.pgshft = PAGE_SHIFT;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
603
sginfo.pgshft = PAGE_SHIFT;
drivers/infiniband/hw/bnxt_re/qplib_rcfw.c
847
req.log2_dbr_pg_size = cpu_to_le16(PAGE_SHIFT -
drivers/infiniband/hw/bnxt_re/qplib_rcfw.c
945
sginfo.pgshft = PAGE_SHIFT;
drivers/infiniband/hw/bnxt_re/qplib_res.c
376
sginfo.pgshft = PAGE_SHIFT;
drivers/infiniband/hw/bnxt_re/qplib_res.c
489
sginfo.pgshft = PAGE_SHIFT;
drivers/infiniband/hw/cxgb4/mem.c
537
shift = PAGE_SHIFT;
drivers/infiniband/hw/cxgb4/provider.c
131
u32 key = vma->vm_pgoff << PAGE_SHIFT;
drivers/infiniband/hw/cxgb4/provider.c
164
ret = io_remap_pfn_range(vma, vma->vm_start, addr >> PAGE_SHIFT,
drivers/infiniband/hw/cxgb4/provider.c
170
addr >> PAGE_SHIFT,
drivers/infiniband/hw/cxgb4/provider.c
175
addr >> PAGE_SHIFT,
drivers/infiniband/hw/efa/efa_verbs.c
2048
pfn = entry->address >> PAGE_SHIFT;
drivers/infiniband/hw/erdma/erdma_main.c
438
req.cfg = FIELD_PREP(ERDMA_CMD_CONFIG_DEVICE_PGSHIFT_MASK, PAGE_SHIFT) |
drivers/infiniband/hw/erdma/erdma_verbs.c
1486
ctx->sdb = dev->func_bar_addr + (ctx->ext_db.sdb_off << PAGE_SHIFT);
drivers/infiniband/hw/erdma/erdma_verbs.c
1487
ctx->cdb = dev->func_bar_addr + (ctx->ext_db.rdb_off << PAGE_SHIFT);
drivers/infiniband/hw/erdma/erdma_verbs.c
1488
ctx->rdb = dev->func_bar_addr + (ctx->ext_db.cdb_off << PAGE_SHIFT);
drivers/infiniband/hw/erdma/erdma_verbs.c
182
PAGE_SHIFT - ERDMA_HW_PAGE_SHIFT);
drivers/infiniband/hw/erdma/erdma_verbs.c
547
dev->func_bar + (ERDMA_SDB_SHARED_PAGE_INDEX << PAGE_SHIFT);
drivers/infiniband/hw/erdma/erdma_verbs.c
672
PAGE_SHIFT;
drivers/infiniband/hw/hfi1/file_ops.c
324
u64 token = vma->vm_pgoff << PAGE_SHIFT,
drivers/infiniband/hw/hfi1/file_ops.c
599
page = vmalloc_to_page((void *)(vmf->pgoff << PAGE_SHIFT));
drivers/infiniband/hw/hfi1/file_ops.c
725
paddr = page_to_pfn(page) << PAGE_SHIFT;
drivers/infiniband/hw/hfi1/tid_rdma.c
1211
(TID_RDMA_MAX_SEGMENT_SIZE >> PAGE_SHIFT));
drivers/infiniband/hw/hfi1/tid_rdma.c
1308
u32 pmtu_pg = flow->req->qp->pmtu >> PAGE_SHIFT;
drivers/infiniband/hw/hfi1/tid_rdma.c
164
(((u64)((p->max_len >> PAGE_SHIFT) - 1) &
drivers/infiniband/hw/hfi1/tid_rdma.c
179
TID_OPFN_MAX_LEN_MASK) + 1) << PAGE_SHIFT;
drivers/infiniband/hw/hfi1/tid_rdma.c
2353
u32 tidlen = EXP_TID_GET(tidentry, LEN) << PAGE_SHIFT;
drivers/infiniband/hw/hfi1/tid_rdma.c
4215
u32 tidlen = EXP_TID_GET(tidentry, LEN) << PAGE_SHIFT;
drivers/infiniband/hw/hfi1/tid_rdma.c
927
PAGE_SHIFT;
drivers/infiniband/hw/hfi1/tid_rdma.c
931
PAGE_SHIFT;
drivers/infiniband/hw/hfi1/tid_rdma.h
19
#define TID_RDMA_MAX_PAGES (BIT(18) >> PAGE_SHIFT)
drivers/infiniband/hw/hfi1/user_exp_rcv.c
605
PAGE_SHIFT;
drivers/infiniband/hw/hfi1/user_exp_rcv.c
609
PAGE_SHIFT;
drivers/infiniband/hw/hfi1/user_exp_rcv.h
48
return 1 + ((epage - spage) >> PAGE_SHIFT);
drivers/infiniband/hw/hfi1/user_sdma.h
27
#define num_pages(x) (1 + ((((x) - 1) & PAGE_MASK) >> PAGE_SHIFT))
drivers/infiniband/hw/hns/hns_roce_cq.c
260
buf_attr.page_shift = hr_dev->caps.cqe_buf_pg_sz + PAGE_SHIFT;
drivers/infiniband/hw/hns/hns_roce_cq.c
266
hr_dev->caps.cqe_ba_pg_sz + PAGE_SHIFT,
drivers/infiniband/hw/hns/hns_roce_device.h
94
#define PG_SHIFT_OFFSET (PAGE_SHIFT - 12)
drivers/infiniband/hw/hns/hns_roce_hem.c
131
+ PAGE_SHIFT);
drivers/infiniband/hw/hns/hns_roce_hem.c
133
+ PAGE_SHIFT);
drivers/infiniband/hw/hns/hns_roce_hem.c
139
+ PAGE_SHIFT);
drivers/infiniband/hw/hns/hns_roce_hem.c
141
+ PAGE_SHIFT);
drivers/infiniband/hw/hns/hns_roce_hem.c
147
+ PAGE_SHIFT);
drivers/infiniband/hw/hns/hns_roce_hem.c
149
+ PAGE_SHIFT);
drivers/infiniband/hw/hns/hns_roce_hem.c
155
+ PAGE_SHIFT);
drivers/infiniband/hw/hns/hns_roce_hem.c
157
+ PAGE_SHIFT);
drivers/infiniband/hw/hns/hns_roce_hem.c
163
+ PAGE_SHIFT);
drivers/infiniband/hw/hns/hns_roce_hem.c
165
+ PAGE_SHIFT);
drivers/infiniband/hw/hns/hns_roce_hem.c
171
+ PAGE_SHIFT);
drivers/infiniband/hw/hns/hns_roce_hem.c
173
+ PAGE_SHIFT);
drivers/infiniband/hw/hns/hns_roce_hem.c
179
+ PAGE_SHIFT);
drivers/infiniband/hw/hns/hns_roce_hem.c
181
+ PAGE_SHIFT);
drivers/infiniband/hw/hns/hns_roce_hem.c
187
PAGE_SHIFT);
drivers/infiniband/hw/hns/hns_roce_hem.c
189
PAGE_SHIFT);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
2791
pg_shift = hr_dev->caps.llm_buf_pg_sz + PAGE_SHIFT;
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
6908
buf_attr.page_shift = hr_dev->caps.eqe_buf_pg_sz + PAGE_SHIFT;
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
6914
hr_dev->caps.eqe_ba_pg_sz + PAGE_SHIFT, NULL,
drivers/infiniband/hw/hns/hns_roce_main.c
412
address = context->uar.pfn << PAGE_SHIFT;
drivers/infiniband/hw/hns/hns_roce_main.c
541
pfn = entry->address >> PAGE_SHIFT;
drivers/infiniband/hw/hns/hns_roce_mr.c
100
buf_attr.page_shift = is_fast ? PAGE_SHIFT :
drivers/infiniband/hw/hns/hns_roce_mr.c
101
hr_dev->caps.pbl_buf_pg_sz + PAGE_SHIFT;
drivers/infiniband/hw/hns/hns_roce_mr.c
114
hr_dev->caps.pbl_ba_pg_sz + PAGE_SHIFT,
drivers/infiniband/hw/hns/hns_roce_mr.c
404
mr->size = max_num_sg * (1 << PAGE_SHIFT);
drivers/infiniband/hw/hns/hns_roce_pd.c
104
uar->pfn = ((pci_resource_start(hr_dev->pci_dev, 2)) >> PAGE_SHIFT);
drivers/infiniband/hw/hns/hns_roce_qp.c
810
PAGE_SHIFT + hr_dev->caps.mtt_ba_pg_sz,
drivers/infiniband/hw/hns/hns_roce_srq.c
175
buf_attr.page_shift = hr_dev->caps.idx_buf_pg_sz + PAGE_SHIFT;
drivers/infiniband/hw/hns/hns_roce_srq.c
182
hr_dev->caps.idx_ba_pg_sz + PAGE_SHIFT,
drivers/infiniband/hw/hns/hns_roce_srq.c
230
buf_attr.page_shift = hr_dev->caps.srqwqe_buf_pg_sz + PAGE_SHIFT;
drivers/infiniband/hw/hns/hns_roce_srq.c
237
hr_dev->caps.srqwqe_ba_pg_sz + PAGE_SHIFT,
drivers/infiniband/hw/ionic/ionic_controlpath.c
1892
(u64)qp->sq_cmb_pgid << PAGE_SHIFT,
drivers/infiniband/hw/ionic/ionic_controlpath.c
2102
(u64)qp->rq_cmb_pgid << PAGE_SHIFT,
drivers/infiniband/hw/ionic/ionic_controlpath.c
303
*addr = dev->lif_cfg.db_phys + ((phys_addr_t)dbpage_num << PAGE_SHIFT);
drivers/infiniband/hw/ionic/ionic_controlpath.c
394
resp.page_shift = PAGE_SHIFT;
drivers/infiniband/hw/ionic/ionic_controlpath.c
451
vma->vm_pgoff << PAGE_SHIFT);
drivers/infiniband/hw/ionic/ionic_queue.c
23
if (q->depth_log2 + q->stride_log2 < PAGE_SHIFT)
drivers/infiniband/hw/ionic/ionic_queue.c
24
q->depth_log2 = PAGE_SHIFT - q->stride_log2;
drivers/infiniband/hw/irdma/verbs.c
143
pci_resource_start(ucontext->iwdev->rf->pcidev, 0)) >> PAGE_SHIFT;
drivers/infiniband/hw/irdma/verbs.c
213
pci_resource_start(ucontext->iwdev->rf->pcidev, 0)) >> PAGE_SHIFT;
drivers/infiniband/hw/mana/main.c
536
PAGE_SHIFT;
drivers/infiniband/hw/mlx4/main.c
1187
PAGE_SHIFT,
drivers/infiniband/hw/mlx4/main.c
2654
ibdev->uar_map = ioremap((phys_addr_t) ibdev->priv_uar.pfn << PAGE_SHIFT,
drivers/infiniband/hw/mlx4/mr.c
256
shift = PAGE_SHIFT;
drivers/infiniband/hw/mlx4/srq.c
124
PAGE_SHIFT, &srq->mtt);
drivers/infiniband/hw/mlx5/devx.c
2290
PAGE_SHIFT),
drivers/infiniband/hw/mlx5/devx.c
2340
min(PAGE_SHIFT, MLX5_ADAPTER_PAGE_SHIFT)));
drivers/infiniband/hw/mlx5/dm.c
17
>> PAGE_SHIFT;
drivers/infiniband/hw/mlx5/dm.c
98
start_page_idx = (addr - hw_start_addr) >> PAGE_SHIFT;
drivers/infiniband/hw/mlx5/main.c
2119
MLX5_ADAPTER_PAGE_SHIFT : PAGE_SHIFT;
drivers/infiniband/hw/mlx5/main.c
2380
return (dev->mdev->bar_addr >> PAGE_SHIFT) + uar_idx / fw_uars_per_page;
drivers/infiniband/hw/mlx5/main.c
2513
pg_off = pg_off >> PAGE_SHIFT;
drivers/infiniband/hw/mlx5/main.c
2687
pfn = (mentry->address >> PAGE_SHIFT);
drivers/infiniband/hw/mlx5/main.c
2707
(index & 0xFF)) << PAGE_SHIFT;
drivers/infiniband/hw/mlx5/main.c
2745
PAGE_SHIFT;
drivers/infiniband/hw/mlx5/mr.c
2358
page_shift = PAGE_SHIFT;
drivers/infiniband/hw/mlx5/mr.c
2381
PAGE_SHIFT, MLX5_MKC_ACCESS_MODE_MTT, in,
drivers/infiniband/hw/mlx5/mr.c
258
MLX5_SET(mkc, mkc, log_page_size, PAGE_SHIFT);
drivers/infiniband/hw/mlx5/odp.c
1043
PAGE_SHIFT;
drivers/infiniband/hw/mlx5/odp.c
1924
u32 log_va_pages = ilog2(TASK_SIZE) - PAGE_SHIFT;
drivers/infiniband/hw/mlx5/odp.c
1928
if (log_va_pages <= 48 - PAGE_SHIFT)
drivers/infiniband/hw/mlx5/odp.c
1931
else if (log_va_pages <= 56 - PAGE_SHIFT)
drivers/infiniband/hw/mlx5/odp.c
1937
mlx5_imr_mtt_bits = mlx5_imr_mtt_shift - PAGE_SHIFT;
drivers/infiniband/hw/mlx5/odp.c
375
dev->odp_max_size = BIT_ULL(MLX5_MAX_UMR_SHIFT + PAGE_SHIFT);
drivers/infiniband/hw/mlx5/odp.c
507
PAGE_SHIFT,
drivers/infiniband/hw/mlx5/odp.c
740
return np << (page_shift - PAGE_SHIFT);
drivers/infiniband/hw/mthca/mthca_allocator.c
115
int p = (index * sizeof (void *)) >> PAGE_SHIFT;
drivers/infiniband/hw/mthca/mthca_allocator.c
125
int p = (index * sizeof (void *)) >> PAGE_SHIFT;
drivers/infiniband/hw/mthca/mthca_allocator.c
142
int p = (index * sizeof (void *)) >> PAGE_SHIFT;
drivers/infiniband/hw/mthca/mthca_allocator.c
202
shift = get_order(size) + PAGE_SHIFT;
drivers/infiniband/hw/mthca/mthca_allocator.c
226
shift = PAGE_SHIFT;
drivers/infiniband/hw/mthca/mthca_cmd.c
1416
u8 uar_page_sz = PAGE_SHIFT - 12;
drivers/infiniband/hw/mthca/mthca_cmd.c
1596
(PAGE_SHIFT - MTHCA_ICM_PAGE_SHIFT);
drivers/infiniband/hw/mthca/mthca_cmd.c
861
(PAGE_SHIFT - MTHCA_ICM_PAGE_SHIFT);
drivers/infiniband/hw/mthca/mthca_eq.c
518
dma_list, PAGE_SHIFT, npages,
drivers/infiniband/hw/mthca/mthca_main.c
714
dev->kar = ioremap((phys_addr_t) dev->driver_uar.pfn << PAGE_SHIFT, PAGE_SIZE);
drivers/infiniband/hw/mthca/mthca_memfree.c
234
table->icm[i] = mthca_alloc_icm(dev, MTHCA_TABLE_CHUNK_SIZE >> PAGE_SHIFT,
drivers/infiniband/hw/mthca/mthca_memfree.c
390
table->icm[i] = mthca_alloc_icm(dev, chunk_size >> PAGE_SHIFT,
drivers/infiniband/hw/mthca/mthca_provider.c
908
err = mthca_mr_alloc(dev, to_mpd(pd)->pd_num, PAGE_SHIFT, virt, length,
drivers/infiniband/hw/mthca/mthca_qp.c
213
return qp->queue.page_list[(n << qp->rq.wqe_shift) >> PAGE_SHIFT].buf +
drivers/infiniband/hw/mthca/mthca_qp.c
225
PAGE_SHIFT].buf +
drivers/infiniband/hw/mthca/mthca_srq.c
79
return srq->queue.page_list[(n << srq->wqe_shift) >> PAGE_SHIFT].buf +
drivers/infiniband/hw/mthca/mthca_uar.c
44
uar->pfn = (pci_resource_start(dev->pdev, 2) >> PAGE_SHIFT) + uar->index;
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
543
unsigned long vm_page = vma->vm_pgoff << PAGE_SHIFT;
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
551
found = ocrdma_search_mmap(ucontext, vma->vm_pgoff << PAGE_SHIFT, len);
drivers/infiniband/hw/qedr/verbs.c
2991
&mr->info.pbl_info, PAGE_SHIFT);
drivers/infiniband/hw/qedr/verbs.c
3016
mr->hw_mr.page_size_log = PAGE_SHIFT;
drivers/infiniband/hw/qedr/verbs.c
414
pfn = entry->io_address >> PAGE_SHIFT;
drivers/infiniband/hw/usnic/usnic_ib_verbs.c
683
vma->vm_pgoff, PAGE_SHIFT, vfid);
drivers/infiniband/hw/usnic/usnic_ib_verbs.c
705
bus_addr >> PAGE_SHIFT,
drivers/infiniband/hw/usnic/usnic_uiom.c
121
npages = PAGE_ALIGN(size + (addr & ~PAGE_MASK)) >> PAGE_SHIFT;
drivers/infiniband/hw/usnic/usnic_uiom.c
127
lock_limit = rlimit(RLIMIT_MEMLOCK) >> PAGE_SHIFT;
drivers/infiniband/hw/usnic/usnic_uiom.c
196
va = interval->start << PAGE_SHIFT;
drivers/infiniband/hw/usnic/usnic_uiom.c
197
size = ((interval->last - interval->start) + 1) << PAGE_SHIFT;
drivers/infiniband/hw/usnic/usnic_uiom.c
218
npages = PAGE_ALIGN(uiomr->length + uiomr->offset) >> PAGE_SHIFT;
drivers/infiniband/hw/usnic/usnic_uiom.c
219
vpn_start = (uiomr->va & PAGE_MASK) >> PAGE_SHIFT;
drivers/infiniband/hw/usnic/usnic_uiom.c
260
if ((va >> PAGE_SHIFT) < interval_node->start)
drivers/infiniband/hw/usnic/usnic_uiom.c
263
if ((va >> PAGE_SHIFT) == interval_node->start) {
drivers/infiniband/hw/usnic/usnic_uiom.c
290
if ((va >> PAGE_SHIFT) == interval_node->last) {
drivers/infiniband/hw/usnic/usnic_uiom.c
349
npages = PAGE_ALIGN(size + offset) >> PAGE_SHIFT;
drivers/infiniband/hw/usnic/usnic_uiom.c
350
vpn_start = (addr & PAGE_MASK) >> PAGE_SHIFT;
drivers/infiniband/hw/usnic/usnic_uiom.c
425
return PAGE_ALIGN(uiomr->length + uiomr->offset) >> PAGE_SHIFT;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_doorbell.c
110
PAGE_SHIFT) + uar->index;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c
849
PAGE_SHIFT;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c
851
ioremap(dev->driver_uar.pfn << PAGE_SHIFT, PAGE_SIZE);
drivers/infiniband/hw/vmw_pvrdma/pvrdma_mr.c
261
mr->page_shift = PAGE_SHIFT;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c
372
unsigned long offset = vma->vm_pgoff << PAGE_SHIFT;
drivers/infiniband/sw/rdmavt/mmap.c
71
unsigned long offset = vma->vm_pgoff << PAGE_SHIFT;
drivers/infiniband/sw/rdmavt/mr.c
373
mr->mr.page_shift = PAGE_SHIFT;
drivers/infiniband/sw/rdmavt/mr.c
600
mr->mr.page_shift = PAGE_SHIFT;
drivers/infiniband/sw/rdmavt/qp.c
245
u32 page = ((unsigned long)address >> PAGE_SHIFT) & wss->pages_mask;
drivers/infiniband/sw/rxe/rxe_mmap.c
65
unsigned long offset = vma->vm_pgoff << PAGE_SHIFT;
drivers/infiniband/sw/rxe/rxe_mr.c
263
u32 i, pages_per_mr = mr_page_size(mr) >> PAGE_SHIFT;
drivers/infiniband/sw/rxe/rxe_mr.c
307
u32 map_mr_pages = (page_size >> PAGE_SHIFT) * mr->num_buf;
drivers/infiniband/sw/rxe/rxe_mr.c
63
mr->page_shift = PAGE_SHIFT;
drivers/infiniband/sw/rxe/rxe_mr.c
94
idx = (iova - (mr->ibmr.iova & mr->page_mask)) >> PAGE_SHIFT;
drivers/infiniband/sw/siw/siw_mem.c
347
num_pages = PAGE_ALIGN(start + len - first_page_va) >> PAGE_SHIFT;
drivers/infiniband/sw/siw/siw_mem.h
60
unsigned int page_idx = (addr - umem->fp_addr) >> PAGE_SHIFT,
drivers/infiniband/sw/siw/siw_qp_tx.c
391
i += PAGE_ALIGN(sge_bytes + offset) >> PAGE_SHIFT;
drivers/iommu/amd/init.c
3982
pfn = __sme_clr(paddr) >> PAGE_SHIFT;
drivers/iommu/amd/iommu.c
2501
pci_prepare_ats(to_pci_dev(dev), PAGE_SHIFT);
drivers/iommu/amd/iommu.c
568
ret = pci_enable_ats(pdev, PAGE_SHIFT);
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.h
200
#define Q_MAX_SZ_SHIFT (PAGE_SHIFT + CONFIG_CMA_ALIGNMENT)
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.h
202
#define Q_MAX_SZ_SHIFT (PAGE_SHIFT + MAX_PAGE_ORDER)
drivers/iommu/arm/arm-smmu/arm-smmu-nvidia.c
280
smmu->pgsize_bitmap &= GENMASK(PAGE_SHIFT, 0);
drivers/iommu/dma-iommu.c
1035
__iommu_dma_free_pages(pages, PAGE_ALIGN(size) >> PAGE_SHIFT);
drivers/iommu/dma-iommu.c
1077
__iommu_dma_free_pages(sh->pages, PAGE_ALIGN(size) >> PAGE_SHIFT);
drivers/iommu/dma-iommu.c
1085
unsigned long count = PAGE_ALIGN(size) >> PAGE_SHIFT;
drivers/iommu/dma-iommu.c
1093
unsigned long count = PAGE_ALIGN(size) >> PAGE_SHIFT;
drivers/iommu/dma-iommu.c
1574
int count = alloc_size >> PAGE_SHIFT;
drivers/iommu/dma-iommu.c
1684
unsigned long nr_pages = PAGE_ALIGN(size) >> PAGE_SHIFT;
drivers/iommu/dma-iommu.c
1723
PAGE_ALIGN(size) >> PAGE_SHIFT,
drivers/iommu/dma-iommu.c
967
count = PAGE_ALIGN(size) >> PAGE_SHIFT;
drivers/iommu/dma-iommu.c
968
pages = __iommu_dma_alloc_pages(dev, count, alloc_sizes >> PAGE_SHIFT,
drivers/iommu/generic_pt/iommu_pt.h
713
pt_install_leaf_entry(&pts, map->oa, PAGE_SHIFT,
drivers/iommu/generic_pt/iommu_pt.h
925
if (log2_mod(iova | paddr, PAGE_SHIFT))
drivers/iommu/generic_pt/iommu_pt.h
927
map.leaf_pgsize_lg2 = PAGE_SHIFT;
drivers/iommu/generic_pt/pt_fmt_defaults.h
139
return PT_GRANULE_LG2SZ == PAGE_SHIFT;
drivers/iommu/intel/dmar.c
668
if (dmar->width < PAGE_SHIFT - 1) {
drivers/iommu/intel/iommu.h
347
#define dma_frcd_page_addr(d) (d & (((u64)-1) << PAGE_SHIFT))
drivers/iommu/intel/iommu.h
39
#define IOVA_PFN(addr) ((addr) >> PAGE_SHIFT)
drivers/iommu/intel/pasid.c
64
1 << (order + PAGE_SHIFT));
drivers/iommu/intel/pasid.c
71
pasid_table->max_pasid = 1 << (order + PAGE_SHIFT + 3);
drivers/iommu/iommu-pages.c
28
return 1UL << (folio_order(ioptdesc_folio(desc)) + PAGE_SHIFT);
drivers/iommu/iommufd/driver.c
296
msi_desc_set_iommu_msi_iova(desc, iova, PAGE_SHIFT);
drivers/iommu/iommufd/driver.c
68
immap->vm_pgoff = startp >> PAGE_SHIFT;
drivers/iommu/iommufd/iova_bitmap.c
306
bytes = (bitmap->mapped.npages << PAGE_SHIFT) - bitmap->mapped.pgoff;
drivers/iommu/iommufd/main.c
580
immap = mtree_load(&ictx->mt_mmap, vma->vm_pgoff << PAGE_SHIFT);
drivers/iommu/iommufd/main.c
602
immap->mmio_addr >> PAGE_SHIFT, length,
drivers/iommu/iommufd/pages.c
831
unsigned long end = start + (npages << PAGE_SHIFT) - 1;
drivers/iommu/iommufd/pages.c
844
offset >>= PAGE_SHIFT;
drivers/iommu/iommufd/pages.c
963
PAGE_SHIFT;
drivers/iommu/iommufd/viommu.c
341
*base_pa = (page_to_pfn(pages[0]) << PAGE_SHIFT) + offset;
drivers/iommu/s390-iommu.c
61
return ((unsigned long)ptr >> PAGE_SHIFT) & ZPCI_PT_MASK;
drivers/md/bcache/bcache.h
795
return meta_bucket_pages(sb) << PAGE_SHIFT;
drivers/md/bcache/super.c
175
SB_OFFSET >> PAGE_SHIFT, GFP_KERNEL);
drivers/md/bcache/sysfs.c
688
ret += 1 << (b->keys.page_order + PAGE_SHIFT);
drivers/md/dm-bufio.c
1190
c->sectors_per_block_bits - (PAGE_SHIFT - SECTOR_SHIFT));
drivers/md/dm-bufio.c
1215
c->sectors_per_block_bits - (PAGE_SHIFT - SECTOR_SHIFT));
drivers/md/dm-bufio.c
2844
DM_BUFIO_MEMORY_PERCENT, 100) << PAGE_SHIFT;
drivers/md/dm-crypt.c
1634
unsigned int nr_iovecs = (size + PAGE_SIZE - 1) >> PAGE_SHIFT;
drivers/md/dm-crypt.c
1655
unsigned remaining_order = __fls((remaining_size + PAGE_SIZE - 1) >> PAGE_SHIFT);
drivers/md/dm-crypt.c
276
val = min(val, BIO_MAX_VECS << PAGE_SHIFT);
drivers/md/dm-crypt.c
3697
BIO_MAX_VECS << PAGE_SHIFT);
drivers/md/dm-crypt.c
3699
BIO_MAX_VECS << PAGE_SHIFT);
drivers/md/dm-flakey.c
434
if (unlikely(bio->bi_iter.bi_size > UIO_MAXIOV << PAGE_SHIFT))
drivers/md/dm-flakey.c
435
dm_accept_partial_bio(bio, UIO_MAXIOV << PAGE_SHIFT >> SECTOR_SHIFT);
drivers/md/dm-flakey.c
438
nr_iovecs = (size + PAGE_SIZE - 1) >> PAGE_SHIFT;
drivers/md/dm-flakey.c
457
unsigned remaining_order = __fls((remaining_size + PAGE_SIZE - 1) >> PAGE_SHIFT);
drivers/md/dm-integrity.c
1062
pl_index = sector >> (PAGE_SHIFT - SECTOR_SHIFT);
drivers/md/dm-integrity.c
1186
pl_index = sector >> (PAGE_SHIFT - SECTOR_SHIFT);
drivers/md/dm-integrity.c
4426
PAGE_SIZE >> SECTOR_SHIFT) >> (PAGE_SHIFT - SECTOR_SHIFT);
drivers/md/dm-integrity.c
5288
pl_index = sector >> (PAGE_SHIFT - SECTOR_SHIFT);
drivers/md/dm-integrity.c
737
*pl_index = sector >> (PAGE_SHIFT - SECTOR_SHIFT);
drivers/md/dm-integrity.c
908
atomic_add(roundup(pl_offset + n_bytes, PAGE_SIZE) >> PAGE_SHIFT, &comp->in_flight);
drivers/md/dm-log-writes.c
909
return dax_zero_page_range(dax_dev, pgoff, nr_pages << PAGE_SHIFT);
drivers/md/dm-pcache/backing_dev.c
346
u32 max_bvecs = (PCACHE_CACHE_SUBTREE_SIZE >> PAGE_SHIFT) + 1;
drivers/md/dm-pcache/cache_dev.c
85
total_pages = bdev_size >> PAGE_SHIFT;
drivers/md/dm-stats.c
91
if (a >> PAGE_SHIFT > totalram_pages() / DM_STATS_MEMORY_FACTOR)
drivers/md/dm-vdo/vio.h
21
MAX_BLOCKS_PER_VIO = (BIO_MAX_VECS << PAGE_SHIFT) / VDO_BLOCK_SIZE,
drivers/md/dm-writecache.c
266
p = s >> PAGE_SHIFT;
drivers/md/dm-writecache.c
271
if (p != s >> PAGE_SHIFT) {
drivers/md/dm-writecache.c
281
offset >>= PAGE_SHIFT - 9;
drivers/md/md-bitmap.c
102
#define PAGE_BIT_SHIFT (PAGE_SHIFT + 3)
drivers/md/md-bitmap.c
434
unsigned int bitmap_limit = (num_pages - pg_index % num_pages) << PAGE_SHIFT;
drivers/md/md-bitmap.c
565
(unsigned long long)index << PAGE_SHIFT);
drivers/md/md-bitmap.c
573
blk_cur = index << (PAGE_SHIFT - inode->i_blkbits);
drivers/md/md-bitmap.c
613
(unsigned long long)index << PAGE_SHIFT,
drivers/md/md-llbitmap.c
1197
int page_start = (start + BITMAP_DATA_OFFSET) >> PAGE_SHIFT;
drivers/md/md-llbitmap.c
1198
int page_end = (end + BITMAP_DATA_OFFSET) >> PAGE_SHIFT;
drivers/md/md-llbitmap.c
1214
int page_start = (start + BITMAP_DATA_OFFSET) >> PAGE_SHIFT;
drivers/md/md-llbitmap.c
1215
int page_end = (end + BITMAP_DATA_OFFSET) >> PAGE_SHIFT;
drivers/md/md-llbitmap.c
1229
int page_start = (start + BITMAP_DATA_OFFSET) >> PAGE_SHIFT;
drivers/md/md-llbitmap.c
1230
int page_end = (end + BITMAP_DATA_OFFSET) >> PAGE_SHIFT;
drivers/md/md-llbitmap.c
1246
int page_start = (start + BITMAP_DATA_OFFSET) >> PAGE_SHIFT;
drivers/md/md-llbitmap.c
1247
int page_end = (end + BITMAP_DATA_OFFSET) >> PAGE_SHIFT;
drivers/md/md-llbitmap.c
424
idx = pos >> PAGE_SHIFT;
drivers/md/md-llbitmap.c
503
idx = pos >> PAGE_SHIFT;
drivers/md/md.c
174
int i, serial_nums = 1 << ((PAGE_SHIFT - ilog2(sizeof(atomic_t))));
drivers/md/md.c
8946
used_pages, stats.pages, used_pages << (PAGE_SHIFT - 10),
drivers/md/raid1.c
1246
unsigned vcnt = (size + PAGE_SIZE - 1) >> PAGE_SHIFT;
drivers/md/raid1.c
2281
vcnt = (r1_bio->sectors + PAGE_SIZE / 512 - 1) >> (PAGE_SHIFT - 9);
drivers/md/raid1.h
26
#define BARRIER_BUCKETS_NR_BITS (PAGE_SHIFT - ilog2(sizeof(atomic_t)))
drivers/md/raid10.c
2365
vcnt = (r10_bio->sectors + (PAGE_SIZE >> 9) - 1) >> (PAGE_SHIFT - 9);
drivers/md/raid10.c
3269
max_sync = RESYNC_PAGES << (PAGE_SHIFT-9);
drivers/md/raid5-cache.c
1040
reserve = (1 + write_disks) << (PAGE_SHIFT - 9);
drivers/md/raid5-cache.c
2013
(le32_to_cpu(payload->size) >> (PAGE_SHIFT - 9));
drivers/md/raid5-cache.c
2023
(le32_to_cpu(payload->size) >> (PAGE_SHIFT - 9));
drivers/md/raid5-cache.c
2135
(le32_to_cpu(payload->size) >> (PAGE_SHIFT - 9));
drivers/md/raid5-cache.c
2921
reserve = (1 + pages) << (PAGE_SHIFT - 9);
drivers/md/raid5-cache.c
831
(PAGE_SHIFT - 9));
drivers/md/raid5-ppl.c
343
le32_add_cpu(&e->data_size, data_disks << PAGE_SHIFT);
drivers/md/raid5.h
485
#define STRIPE_SHIFT (PAGE_SHIFT - 9)
drivers/media/common/videobuf2/videobuf2-core.c
2411
*plane = (offset >> PAGE_SHIFT) & PLANE_INDEX_MASK;
drivers/media/common/videobuf2/videobuf2-core.c
2489
unsigned long offset = vma->vm_pgoff << PAGE_SHIFT;
drivers/media/common/videobuf2/videobuf2-core.c
2563
unsigned long offset = pgoff << PAGE_SHIFT;
drivers/media/common/videobuf2/videobuf2-core.c
35
#define PLANE_INDEX_SHIFT (PAGE_SHIFT + PLANE_INDEX_BITS)
drivers/media/common/videobuf2/videobuf2-core.c
401
vb->planes[plane].m.offset = offset + (plane << PAGE_SHIFT);
drivers/media/common/videobuf2/videobuf2-dma-sg.c
121
buf->num_pages = size >> PAGE_SHIFT;
drivers/media/common/videobuf2/videobuf2-memops.c
46
first = start >> PAGE_SHIFT;
drivers/media/common/videobuf2/videobuf2-memops.c
47
last = (start + length - 1) >> PAGE_SHIFT;
drivers/media/pci/bt8xx/bttv-driver.c
3602
gbuffers, gbufsize >> 10, gbufsize >> PAGE_SHIFT);
drivers/media/pci/cx23885/cx23885-alsa.c
385
(PAGE_ALIGN(chip->dma_size) >> PAGE_SHIFT));
drivers/media/pci/cx23885/cx23885-alsa.c
78
buf->vaddr = vmalloc_32(nr_pages << PAGE_SHIFT);
drivers/media/pci/cx23885/cx23885-alsa.c
85
buf->vaddr, nr_pages << PAGE_SHIFT);
drivers/media/pci/cx23885/cx23885-alsa.c
87
memset(buf->vaddr, 0, nr_pages << PAGE_SHIFT);
drivers/media/pci/cx25821/cx25821-alsa.c
140
buf->vaddr = vmalloc_32(nr_pages << PAGE_SHIFT);
drivers/media/pci/cx25821/cx25821-alsa.c
148
nr_pages << PAGE_SHIFT);
drivers/media/pci/cx25821/cx25821-alsa.c
150
memset(buf->vaddr, 0, nr_pages << PAGE_SHIFT);
drivers/media/pci/cx25821/cx25821-alsa.c
526
(PAGE_ALIGN(chip->dma_size) >> PAGE_SHIFT));
drivers/media/pci/cx88/cx88-alsa.c
280
buf->vaddr = vmalloc_32(nr_pages << PAGE_SHIFT);
drivers/media/pci/cx88/cx88-alsa.c
287
buf->vaddr, nr_pages << PAGE_SHIFT);
drivers/media/pci/cx88/cx88-alsa.c
289
memset(buf->vaddr, 0, nr_pages << PAGE_SHIFT);
drivers/media/pci/cx88/cx88-alsa.c
476
(PAGE_ALIGN(chip->dma_size) >> PAGE_SHIFT));
drivers/media/pci/ivtv/ivtv-udma.c
19
dma_page->first = (first & PAGE_MASK) >> PAGE_SHIFT;
drivers/media/pci/ivtv/ivtv-udma.c
20
dma_page->last = ((first+size-1) & PAGE_MASK) >> PAGE_SHIFT;
drivers/media/pci/saa7134/saa7134-alsa.c
262
dma->vaddr = vmalloc_32(nr_pages << PAGE_SHIFT);
drivers/media/pci/saa7134/saa7134-alsa.c
269
dma->vaddr, nr_pages << PAGE_SHIFT);
drivers/media/pci/saa7134/saa7134-alsa.c
271
memset(dma->vaddr, 0, nr_pages << PAGE_SHIFT);
drivers/media/pci/saa7134/saa7134-alsa.c
349
(dev->dmasound.bufsize + PAGE_SIZE) >> PAGE_SHIFT);
drivers/media/platform/mediatek/vpu/mtk_vpu.c
887
vpu->enable_4GB = !!(totalram_pages() > (SZ_2G >> PAGE_SHIFT));
drivers/media/platform/rockchip/rga/rga-buf.c
144
offsets[i] = curr_desc << PAGE_SHIFT;
drivers/media/platform/samsung/s5p-mfc/s5p_mfc.c
1081
unsigned long offset = vma->vm_pgoff << PAGE_SHIFT;
drivers/media/platform/samsung/s5p-mfc/s5p_mfc.c
1089
vma->vm_pgoff -= (DST_QUEUE_OFF_BASE >> PAGE_SHIFT);
drivers/media/platform/samsung/s5p-mfc/s5p_mfc.c
1229
mfc_dev->mem_bitmap = bitmap_zalloc(mem_size >> PAGE_SHIFT, GFP_KERNEL);
drivers/media/platform/samsung/s5p-mfc/s5p_mfc.c
1252
bitmap_set(mfc_dev->mem_bitmap, 0, offset >> PAGE_SHIFT);
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_opr.c
100
unsigned int count = b->size >> PAGE_SHIFT;
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_opr.c
37
unsigned int bits = dev->mem_size >> PAGE_SHIFT;
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_opr.c
38
unsigned int count = b->size >> PAGE_SHIFT;
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_opr.c
39
unsigned int align = (SZ_64K >> PAGE_SHIFT) - 1;
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_opr.c
50
offset = start << PAGE_SHIFT;
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_opr.c
99
unsigned int start = (b->dma - dev->mem_base) >> PAGE_SHIFT;
drivers/media/usb/go7007/go7007-priv.h
128
#define GO7007_BUF_SIZE (GO7007_BUF_PAGES << PAGE_SHIFT)
drivers/media/v4l2-core/v4l2-mem2mem.c
1001
unsigned long offset = pgoff << PAGE_SHIFT;
drivers/media/v4l2-core/v4l2-mem2mem.c
1008
pgoff -= (DST_QUEUE_OFF_BASE >> PAGE_SHIFT);
drivers/media/v4l2-core/v4l2-mem2mem.c
981
unsigned long offset = vma->vm_pgoff << PAGE_SHIFT;
drivers/media/v4l2-core/v4l2-mem2mem.c
988
vma->vm_pgoff -= (DST_QUEUE_OFF_BASE >> PAGE_SHIFT);
drivers/memstick/core/mspro_block.c
563
sg_page(&(msb->req_sg[msb->current_seg])) + (t_offset >> PAGE_SHIFT),
drivers/memstick/host/jmb38x_ms.c
320
pg = sg_page(&host->req->sg) + (off >> PAGE_SHIFT);
drivers/memstick/host/tifm_ms.c
204
pg = sg_page(&host->req->sg) + (off >> PAGE_SHIFT);
drivers/misc/bcm-vk/bcm_vk_dev.c
1214
>> PAGE_SHIFT) + 1;
drivers/misc/bcm-vk/bcm_vk_dev.c
1219
>> PAGE_SHIFT);
drivers/misc/bcm-vk/bcm_vk_sg.c
58
first = (data & PAGE_MASK) >> PAGE_SHIFT;
drivers/misc/bcm-vk/bcm_vk_sg.c
59
last = ((data + vkdata->size - 1) & PAGE_MASK) >> PAGE_SHIFT;
drivers/misc/fastrpc.c
1050
pg_start = (ctx->args[i].ptr & PAGE_MASK) >> PAGE_SHIFT;
drivers/misc/fastrpc.c
1052
PAGE_SHIFT;
drivers/misc/fastrpc.c
1073
pg_start = (rpra[i].buf.pv & PAGE_MASK) >> PAGE_SHIFT;
drivers/misc/fastrpc.c
1074
pg_end = ((rpra[i].buf.pv + len - 1) & PAGE_MASK) >> PAGE_SHIFT;
drivers/misc/genwqe/card_dev.c
406
__func__, vma->vm_start, vma->vm_pgoff << PAGE_SHIFT,
drivers/misc/genwqe/card_dev.c
467
pfn = virt_to_phys(dma_map->k_vaddr) >> PAGE_SHIFT;
drivers/misc/ibmasm/ibmasmfs.c
113
sb->s_blocksize_bits = PAGE_SHIFT;
drivers/misc/ocxl/afu_irq.c
160
1 << PAGE_SHIFT, 1);
drivers/misc/ocxl/afu_irq.c
22
return (offset - ctx->afu->irq_base_offset) >> PAGE_SHIFT;
drivers/misc/ocxl/afu_irq.c
27
return ctx->afu->irq_base_offset + (irq_id << PAGE_SHIFT);
drivers/misc/ocxl/context.c
108
return vmf_insert_pfn(vma, address, trigger_addr >> PAGE_SHIFT);
drivers/misc/ocxl/context.c
134
ret = vmf_insert_pfn(vma, address, pp_mmio_addr >> PAGE_SHIFT);
drivers/misc/ocxl/context.c
146
offset = vmf->pgoff << PAGE_SHIFT;
drivers/misc/ocxl/context.c
164
int irq_id = ocxl_irq_offset_to_id(ctx, vma->vm_pgoff << PAGE_SHIFT);
drivers/misc/ocxl/context.c
191
(ctx->afu->config.pp_mmio_stride >> PAGE_SHIFT))
drivers/misc/ocxl/context.c
200
if ((vma->vm_pgoff << PAGE_SHIFT) < ctx->afu->irq_base_offset)
drivers/misc/ocxl/link.c
356
spa->spa_order = SPA_SPA_SIZE_LOG - PAGE_SHIFT;
drivers/misc/ocxl/sysfs.c
115
if (vmf->pgoff >= (afu->config.global_mmio_size >> PAGE_SHIFT))
drivers/misc/ocxl/sysfs.c
119
offset += (afu->global_mmio_start >> PAGE_SHIFT);
drivers/misc/ocxl/sysfs.c
134
(afu->config.global_mmio_size >> PAGE_SHIFT))
drivers/misc/sgi-gru/grufault.c
184
*pageshift = is_vm_hugetlb_page(vma) ? HPAGE_SHIFT : PAGE_SHIFT;
drivers/misc/sgi-gru/grufault.c
186
*pageshift = PAGE_SHIFT;
drivers/misc/sgi-gru/grufault.c
240
*paddr = pte_pfn(pte) << PAGE_SHIFT;
drivers/misc/sgi-gru/grufault.c
242
*pageshift = is_vm_hugetlb_page(vma) ? HPAGE_SHIFT : PAGE_SHIFT;
drivers/misc/sgi-gru/grufault.c
244
*pageshift = PAGE_SHIFT;
drivers/misc/sgi-gru/grufault.c
427
if (unlikely(cbe) && pageshift == PAGE_SHIFT) {
drivers/misc/sgi-gru/grumain.c
334
gts->ts_sizeavail = GRU_SIZEAVAIL(PAGE_SHIFT);
drivers/misc/sgi-gru/grumain.c
961
paddr >> PAGE_SHIFT, GRU_GSEG_PAGESIZE,
drivers/misc/sgi-gru/grutlbpurge.c
153
pageshift = PAGE_SHIFT;
drivers/misc/sgi-xp/xpc_uv.c
167
mq->order = pg_order + PAGE_SHIFT;
drivers/misc/sgi-xp/xpc_uv.c
257
pg_order = mq->order - PAGE_SHIFT;
drivers/misc/uacce/uacce.c
380
uacce->qf_pg_num[UACCE_QFRT_MMIO] << PAGE_SHIFT);
drivers/misc/uacce/uacce.c
389
uacce->qf_pg_num[UACCE_QFRT_DUS] << PAGE_SHIFT);
drivers/misc/vmw_balloon.c
258
u64 reserved : PAGE_SHIFT - 5;
drivers/misc/vmw_balloon.c
80
#define VMW_BALLOON_2M_ORDER (PMD_SHIFT - PAGE_SHIFT)
drivers/misc/vmw_vmci/vmci_guest.c
731
vmci_write_reg(vmci_dev, PAGE_SHIFT, VMCI_GUEST_PAGE_SHIFT);
drivers/misc/vmw_vmci/vmci_guest.c
752
vmci_dev->notification_base >> PAGE_SHIFT;
drivers/misc/vmw_vmci/vmci_queue_pair.c
461
produce_q->kernel_if->u.g.pas[i] >> PAGE_SHIFT;
drivers/misc/vmw_vmci/vmci_queue_pair.c
465
consume_q->kernel_if->u.g.pas[i] >> PAGE_SHIFT;
drivers/mmc/core/mmc_test.c
366
order = get_order(max_seg_page_cnt << PAGE_SHIFT);
drivers/mmc/host/bcm2835.c
1072
host->drain_page += host->drain_offset >> PAGE_SHIFT;
drivers/mmc/host/tifm_sd.c
194
pg = sg_page(&sg[host->sg_pos]) + (off >> PAGE_SHIFT);
drivers/mmc/host/tifm_sd.c
241
pg = sg_page(&sg[host->sg_pos]) + (off >> PAGE_SHIFT);
drivers/mmc/host/usdhi6rol0.c
419
(host->page_idx << PAGE_SHIFT) + data->blksz - blk_head)
drivers/mmc/host/usdhi6rol0.c
463
done = (host->page_idx << PAGE_SHIFT) + host->offset;
drivers/mtd/devices/block2mtd.c
110
pgoff_t index = from >> PAGE_SHIFT;
drivers/mtd/devices/block2mtd.c
144
pgoff_t index = to >> PAGE_SHIFT; // page index
drivers/mtd/devices/block2mtd.c
60
pgoff_t index = to >> PAGE_SHIFT; // page index
drivers/mtd/devices/block2mtd.c
61
int pages = len >> PAGE_SHIFT;
drivers/mtd/mtdchar.c
1357
if (len > mtd->size || pgoff >= (mtd->size >> PAGE_SHIFT))
drivers/mtd/mtdchar.c
1360
offset = pgoff << PAGE_SHIFT;
drivers/mtd/mtdswap.c
1108
readpos = (loff_t)realblock << PAGE_SHIFT;
drivers/mtd/mtdswap.c
1280
d->pages_per_eblk = mtd->erasesize >> PAGE_SHIFT;
drivers/mtd/mtdswap.c
1431
mbd_dev->size = swap_size >> PAGE_SHIFT;
drivers/mtd/mtdswap.c
642
writepos = (loff_t)*bp << PAGE_SHIFT;
drivers/mtd/mtdswap.c
686
readpos = (loff_t) oldblock << PAGE_SHIFT;
drivers/mtd/nand/raw/nandsim.c
1316
start_index = pos >> PAGE_SHIFT;
drivers/mtd/nand/raw/nandsim.c
1317
end_index = (pos + count - 1) >> PAGE_SHIFT;
drivers/net/ethernet/amd/pds_core/main.c
139
(u64)page_num << PAGE_SHIFT, PAGE_SIZE);
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
3253
jumbo_factor = PAGE_ALIGN(bn->netdev->mtu - 40) >> PAGE_SHIFT;
drivers/net/ethernet/broadcom/bnge/bnge_rmem.h
41
#if (PAGE_SHIFT < 12)
drivers/net/ethernet/broadcom/bnge/bnge_rmem.h
43
#elif (PAGE_SHIFT <= 13)
drivers/net/ethernet/broadcom/bnge/bnge_rmem.h
44
#define BNGE_PAGE_SHIFT PAGE_SHIFT
drivers/net/ethernet/broadcom/bnge/bnge_rmem.h
45
#elif (PAGE_SHIFT < 16)
drivers/net/ethernet/broadcom/bnge/bnge_rmem.h
52
#if (PAGE_SHIFT > 15)
drivers/net/ethernet/broadcom/bnge/bnge_rmem.h
55
#define BNGE_RX_PAGE_SHIFT PAGE_SHIFT
drivers/net/ethernet/broadcom/bnx2.c
3038
int pages = PAGE_ALIGN(raw_len - hdr_len) >> PAGE_SHIFT;
drivers/net/ethernet/broadcom/bnx2.c
3063
pages = PAGE_ALIGN(frag_size) >> PAGE_SHIFT;
drivers/net/ethernet/broadcom/bnx2.c
3211
pages = PAGE_ALIGN(len - hdr_len) >> PAGE_SHIFT;
drivers/net/ethernet/broadcom/bnx2.c
5385
int pages = PAGE_ALIGN(bp->dev->mtu - 40) >> PAGE_SHIFT;
drivers/net/ethernet/broadcom/bnx2.h
6550
#if (PAGE_SHIFT > 14)
drivers/net/ethernet/broadcom/bnx2.h
6553
#define BNX2_PAGE_BITS PAGE_SHIFT
drivers/net/ethernet/broadcom/bnxt/bnxt.c
4879
jumbo_factor = PAGE_ALIGN(bp->dev->mtu - 40) >> PAGE_SHIFT;
drivers/net/ethernet/broadcom/bnxt/bnxt.h
747
#if (PAGE_SHIFT < 12)
drivers/net/ethernet/broadcom/bnxt/bnxt.h
749
#elif (PAGE_SHIFT <= 13)
drivers/net/ethernet/broadcom/bnxt/bnxt.h
750
#define BNXT_PAGE_SHIFT PAGE_SHIFT
drivers/net/ethernet/broadcom/bnxt/bnxt.h
751
#elif (PAGE_SHIFT < 16)
drivers/net/ethernet/broadcom/bnxt/bnxt.h
760
#if (PAGE_SHIFT > 15)
drivers/net/ethernet/broadcom/bnxt/bnxt.h
763
#define BNXT_RX_PAGE_SHIFT PAGE_SHIFT
drivers/net/ethernet/broadcom/cnic_if.h
29
#if (PAGE_SHIFT > 14)
drivers/net/ethernet/broadcom/cnic_if.h
32
#define CNIC_PAGE_BITS PAGE_SHIFT
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1267
#define SIZE_TO_PAGES(size) (((size) >> PAGE_SHIFT) + ((((size) &\
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1268
(PAGE_SIZE - 1)) + (PAGE_SIZE - 1)) >> PAGE_SHIFT))
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3330
page_count = q_size >> PAGE_SHIFT;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
1242
t3_write_reg(adap, A_ULPRX_TDDP_PSZ, V_HPZ0(PAGE_SHIFT - 12));
drivers/net/ethernet/chelsio/cxgb3/sge.c
3315
V_HOSTPAGESIZE(PAGE_SHIFT - 11) | F_BIGENDIANINGRESS |
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
4217
t4_write_reg(adap, ULP_RX_TDDP_PSZ_A, HPZ0_V(PAGE_SHIFT - 12));
drivers/net/ethernet/chelsio/cxgb4/sge.c
5100
s->fl_pg_order = ilog2(fl_large_pg) - PAGE_SHIFT;
drivers/net/ethernet/chelsio/cxgb4/sge.c
65
#if PAGE_SHIFT >= 16
drivers/net/ethernet/chelsio/cxgb4/sge.c
68
# define FL_PG_ORDER (16 - PAGE_SHIFT)
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
2664
s->fl_pg_order = ilog2(fl_large_pg) - PAGE_SHIFT;
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_main.c
37
static uint send_page_order = (14 - PAGE_SHIFT < 0) ? 0 : 14 - PAGE_SHIFT;
drivers/net/ethernet/cortina/gemini.c
829
fpp_order = PAGE_SHIFT - geth->freeq_frag_order;
drivers/net/ethernet/cortina/gemini.c
868
unsigned int fpp_order = PAGE_SHIFT - geth->freeq_frag_order;
drivers/net/ethernet/cortina/gemini.c
923
unsigned int fpp_order = PAGE_SHIFT - geth->freeq_frag_order;
drivers/net/ethernet/cortina/gemini.c
995
unsigned int fpp_order = PAGE_SHIFT - geth->freeq_frag_order;
drivers/net/ethernet/fungible/funcore/fun_dev.c
117
((PAGE_SHIFT - 12) << NVME_CC_MPS_SHIFT) |
drivers/net/ethernet/fungible/funcore/fun_queue.c
493
funq->cq_vector, 0, 0, PAGE_SHIFT, &funq->rqid,
drivers/net/ethernet/fungible/funeth/funeth_rx.c
730
0, 0, fp->fdev->kern_end_qid, PAGE_SHIFT,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
771
int page_id = index >> (PAGE_SHIFT - GVE_TX_BUF_SHIFT_DQO);
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
32
#define EHEA_HUGEPAGE_PFN_MASK ((EHEA_HUGEPAGE_SIZE - 1) >> PAGE_SHIFT)
drivers/net/ethernet/intel/e1000e/netdev.c
3016
#define PAGE_USE_COUNT(S) (((S) >> PAGE_SHIFT) + \
drivers/net/ethernet/marvell/sky2.c
1161
sky2->rx_nfrags = size >> PAGE_SHIFT;
drivers/net/ethernet/marvell/sky2.c
1165
size -= sky2->rx_nfrags << PAGE_SHIFT;
drivers/net/ethernet/marvell/sky2.h
2198
dma_addr_t frag_addr[ETH_JUMBO_MTU >> PAGE_SHIFT ?: 1];
drivers/net/ethernet/mellanox/mlx4/alloc.c
562
buf->page_shift = get_order(size) + PAGE_SHIFT;
drivers/net/ethernet/mellanox/mlx4/alloc.c
596
buf->page_shift = PAGE_SHIFT;
drivers/net/ethernet/mellanox/mlx4/en_main.c
288
mdev->uar_map = ioremap((phys_addr_t) mdev->priv_uar.pfn << PAGE_SHIFT,
drivers/net/ethernet/mellanox/mlx4/eq.c
1028
err = mlx4_mtt_init(dev, npages, PAGE_SHIFT, &eq->mtt);
drivers/net/ethernet/mellanox/mlx4/eq.c
1040
eq_context->log_page_size = PAGE_SHIFT - MLX4_ICM_PAGE_SHIFT;
drivers/net/ethernet/mellanox/mlx4/fw.c
1733
(PAGE_SHIFT - MLX4_ICM_PAGE_SHIFT);
drivers/net/ethernet/mellanox/mlx4/fw.c
2592
(PAGE_SHIFT - MLX4_ICM_PAGE_SHIFT);
drivers/net/ethernet/mellanox/mlx4/icm.c
270
table->icm[i] = mlx4_alloc_icm(dev, MLX4_TABLE_CHUNK_SIZE >> PAGE_SHIFT,
drivers/net/ethernet/mellanox/mlx4/icm.c
447
table->icm[i] = mlx4_alloc_icm(dev, chunk_size >> PAGE_SHIFT,
drivers/net/ethernet/mellanox/mlx4/main.c
1908
(dev->caps.num_uars << PAGE_SHIFT);
drivers/net/ethernet/mellanox/mlx4/main.c
1910
(dev->caps.num_uars << PAGE_SHIFT);
drivers/net/ethernet/mellanox/mlx4/main.c
2373
PAGE_SHIFT - DEFAULT_UAR_PAGE_SHIFT;
drivers/net/ethernet/mellanox/mlx4/main.c
2377
init_hca->uar_page_sz = PAGE_SHIFT - 12;
drivers/net/ethernet/mellanox/mlx4/main.c
2739
priv->kar = ioremap((phys_addr_t) priv->driver_uar.pfn << PAGE_SHIFT, PAGE_SIZE);
drivers/net/ethernet/mellanox/mlx4/main.c
303
(1 << (PAGE_SHIFT - dev->uar_page_shift)));
drivers/net/ethernet/mellanox/mlx4/main.c
553
dev->uar_page_shift = PAGE_SHIFT;
drivers/net/ethernet/mellanox/mlx4/main.c
964
if (dev->uar_page_shift > PAGE_SHIFT) {
drivers/net/ethernet/mellanox/mlx4/pd.c
159
uar->pfn = (pci_resource_start(dev->persist->pdev, 2) >> PAGE_SHIFT)
drivers/net/ethernet/mellanox/mlx4/pd.c
202
uar->map = ioremap(uar->pfn << PAGE_SHIFT, PAGE_SIZE);
drivers/net/ethernet/mellanox/mlx4/pd.c
209
uar->index << PAGE_SHIFT,
drivers/net/ethernet/mellanox/mlx5/core/alloc.c
81
buf->page_shift = PAGE_SHIFT;
drivers/net/ethernet/mellanox/mlx5/core/diag/fw_tracer.c
206
MLX5_SET(mkc, mkc, log_page_size, PAGE_SHIFT);
drivers/net/ethernet/mellanox/mlx5/core/en.h
92
(PAGE_SHIFT - MLX5E_SHAMPO_LOG_HEADER_ENTRY_SIZE)
drivers/net/ethernet/mellanox/mlx5/core/en/params.c
1182
frame_shift <= PAGE_SHIFT; frame_shift++) {
drivers/net/ethernet/mellanox/mlx5/core/en/params.c
38
req_page_shift = PAGE_SHIFT;
drivers/net/ethernet/mellanox/mlx5/core/en_ethtool.c
366
PAGE_SHIFT,
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
1007
if (WARN_ON(BIT(PAGE_SHIFT + pool_order) / 64 >
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
1021
pp_params.max_len = BIT(PAGE_SHIFT + pool_order);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
5831
mlx5e_check_fragmented_striding_rq_cap(mdev, PAGE_SHIFT,
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
5836
mlx5e_check_fragmented_striding_rq_cap(mdev, PAGE_SHIFT,
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
675
u16 page_order = max_t(s16, rq->mpwqe.page_shift - PAGE_SHIFT, 0);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
676
u32 page_size = BIT(PAGE_SHIFT + page_order);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
693
u16 page_order = max_t(s16, rq->mpwqe.page_shift - PAGE_SHIFT, 0);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
694
u32 page_size = BIT(PAGE_SHIFT + page_order);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
808
MLX5_SET(mkc, mkc, log_page_size, PAGE_SHIFT);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
951
pool_order = rq->mpwqe.page_shift - PAGE_SHIFT;
drivers/net/ethernet/mellanox/mlx5/core/main.c
636
MLX5_SET(cmd_hca_cap, set_hca_cap, log_uar_page_sz, PAGE_SHIFT - 12);
drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c
253
#define MLX5_U64_4K_PAGE_MASK ((~(u64)0U) << PAGE_SHIFT)
drivers/net/ethernet/mellanox/mlx5/core/uar.c
137
up->map = ioremap_wc(pfn << PAGE_SHIFT, PAGE_SIZE);
drivers/net/ethernet/mellanox/mlx5/core/uar.c
143
up->map = ioremap(pfn << PAGE_SHIFT, PAGE_SIZE);
drivers/net/ethernet/mellanox/mlx5/core/uar.c
74
system_page_index = index >> (PAGE_SHIFT - MLX5_ADAPTER_PAGE_SHIFT);
drivers/net/ethernet/mellanox/mlx5/core/uar.c
78
return (mdev->bar_addr >> PAGE_SHIFT) + system_page_index;
drivers/net/ethernet/mellanox/mlx5/core/wq.c
140
u16 rq_npages = rq_byte_size >> PAGE_SHIFT;
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp.c
650
chunk_order = min_t(unsigned int, dma_order, PAGE_SHIFT);
drivers/net/ethernet/pensando/ionic/ionic_bus_pci.c
104
(u64)page_num << PAGE_SHIFT, PAGE_SIZE);
drivers/net/ethernet/sfc/falcon/tx.c
30
&tx_queue->cb_page[index >> (PAGE_SHIFT - EF4_TX_CB_ORDER)];
drivers/net/ethernet/sfc/siena/tx.c
31
&tx_queue->cb_page[index >> (PAGE_SHIFT - EFX_TX_CB_ORDER)];
drivers/net/ethernet/sfc/tx.c
39
&tx_queue->cb_page[index >> (PAGE_SHIFT - EFX_TX_CB_ORDER)];
drivers/net/ethernet/sun/cassini.c
3356
if (PAGE_SHIFT < CAS_JUMBO_PAGE_SHIFT) {
drivers/net/ethernet/sun/cassini.c
3360
PAGE_SHIFT);
drivers/net/ethernet/sun/cassini.c
3362
__free_pages(page, CAS_JUMBO_PAGE_SHIFT - PAGE_SHIFT);
drivers/net/ethernet/sun/cassini.c
3363
cp->page_order = CAS_JUMBO_PAGE_SHIFT - PAGE_SHIFT;
drivers/net/ethernet/sun/niu.c
3291
a >>= PAGE_SHIFT;
drivers/net/ethernet/sun/niu.c
4445
bss = min(PAGE_SHIFT, 15);
drivers/net/ethernet/sun/niu.c
4448
rp->rbr_blocks_per_page = 1 << (PAGE_SHIFT-bss);
drivers/net/wireguard/ratelimiter.c
175
(totalram_pages() << PAGE_SHIFT) /
drivers/net/wireless/intel/iwlwifi/iwl-trans.h
699
u32 sg_offset: PAGE_SHIFT;
drivers/net/xen-netfront.c
543
page += offset >> PAGE_SHIFT;
drivers/nvdimm/pfn_devs.c
665
unsigned long reserve = nd_info_block_reserve() >> PAGE_SHIFT;
drivers/nvdimm/pmem.c
439
u64 len = nr_pages << PAGE_SHIFT;
drivers/nvme/host/zns.c
136
bufsize = min_t(size_t, bufsize, queue_max_segments(q) << PAGE_SHIFT);
drivers/nvme/target/pci-epf.c
36
(NVMET_PCI_EPF_MAX_SEGS << (PAGE_SHIFT - 10))
drivers/nvme/target/rdma.c
194
return 1 + (((len - 1) & PAGE_MASK) >> PAGE_SHIFT);
drivers/parisc/ccio-dma.c
1007
ioc->usg_pages += sg_dma_len(sglist) >> PAGE_SHIFT;
drivers/parisc/ccio-dma.c
1249
if (iova_space_size < (1 << (20 - PAGE_SHIFT))) {
drivers/parisc/ccio-dma.c
1250
iova_space_size = 1 << (20 - PAGE_SHIFT);
drivers/parisc/ccio-dma.c
1252
} else if (iova_space_size > (1 << (30 - PAGE_SHIFT))) {
drivers/parisc/ccio-dma.c
1253
iova_space_size = 1 << (30 - PAGE_SHIFT);
drivers/parisc/ccio-dma.c
1270
iov_order = get_order(iova_space_size << PAGE_SHIFT);
drivers/parisc/ccio-dma.c
1273
iova_space_size = 1 << (iov_order + PAGE_SHIFT);
drivers/parisc/ccio-dma.c
1280
BUG_ON((1 << get_order(ioc->pdir_size)) != (ioc->pdir_size >> PAGE_SHIFT));
drivers/parisc/ccio-dma.c
1284
(unsigned long) totalram_pages() >> (20 - PAGE_SHIFT),
drivers/parisc/ccio-dma.c
1286
iov_order + PAGE_SHIFT);
drivers/parisc/ccio-dma.c
1319
ioc->chainid_shift = get_order(iova_space_size) + PAGE_SHIFT - CCIO_CHAINID_SHIFT;
drivers/parisc/ccio-dma.c
268
#define IOVP_SHIFT PAGE_SHIFT
drivers/parisc/sba_iommu.c
1065
ioc->usg_pages += ((sg_dma_address(sglist) & ~IOVP_MASK) + sg_dma_len(sglist) + IOVP_SIZE - 1) >> PAGE_SHIFT;
drivers/parisc/sba_iommu.c
1296
iov_order = get_order(iova_space_size >> (IOVP_SHIFT - PAGE_SHIFT));
drivers/parisc/sba_iommu.c
1301
iov_order + PAGE_SHIFT);
drivers/parisc/sba_iommu.c
1314
ioc->hint_shift_pdir = iov_order + PAGE_SHIFT;
drivers/parisc/sba_iommu.c
1315
ioc->hint_mask_pdir = ~(0x3 << (iov_order + PAGE_SHIFT));
drivers/parisc/sba_iommu.c
1326
iova_space_mask <<= (iov_order + PAGE_SHIFT);
drivers/parisc/sba_iommu.c
1346
switch (PAGE_SHIFT) {
drivers/parisc/sba_iommu.c
1353
1 << PAGE_SHIFT);
drivers/parisc/sba_iommu.c
1414
if (iova_space_size < (1 << (20 - PAGE_SHIFT))) {
drivers/parisc/sba_iommu.c
1415
iova_space_size = 1 << (20 - PAGE_SHIFT);
drivers/parisc/sba_iommu.c
1417
else if (iova_space_size > (1 << (30 - PAGE_SHIFT))) {
drivers/parisc/sba_iommu.c
1418
iova_space_size = 1 << (30 - PAGE_SHIFT);
drivers/parisc/sba_iommu.c
1426
iov_order = get_order(iova_space_size << PAGE_SHIFT);
drivers/parisc/sba_iommu.c
1429
iova_space_size = 1 << (iov_order + PAGE_SHIFT);
drivers/parisc/sba_iommu.c
1436
(unsigned long) totalram_pages() >> (20 - PAGE_SHIFT),
drivers/parisc/sba_iommu.c
1438
iov_order + PAGE_SHIFT);
drivers/parisc/sba_iommu.c
1447
ioc->hint_shift_pdir = iov_order + PAGE_SHIFT;
drivers/parisc/sba_iommu.c
1448
ioc->hint_mask_pdir = ~(0x3 << (iov_order + PAGE_SHIFT));
drivers/parisc/sba_iommu.c
1458
iova_space_mask <<= (iov_order + PAGE_SHIFT);
drivers/parisc/sba_iommu.c
1488
switch (PAGE_SHIFT) {
drivers/parisc/sba_iommu.c
1495
1 << PAGE_SHIFT);
drivers/parisc/sba_iommu.c
383
unsigned long o = 1 << get_order(bits_wanted << PAGE_SHIFT);
drivers/parisc/sba_iommu.c
579
pba |= (ci >> PAGE_SHIFT) & 0xff; /* move CI (8 bits) into lowest byte */
drivers/parisc/sba_iommu.c
637
iovp |= get_order(byte_cnt) + PAGE_SHIFT;
drivers/pci/iov.c
832
i = PAGE_SHIFT > 12 ? PAGE_SHIFT - 12 : 0;
drivers/pci/mmap.c
31
size = ((pci_resource_len(pdev, bar) - 1) >> PAGE_SHIFT) + 1;
drivers/pci/mmap.c
45
vma->vm_pgoff += (pci_resource_start(pdev, bar) >> PAGE_SHIFT);
drivers/pci/mmap.c
69
size = ((pci_resource_len(pdev, resno) - 1) >> PAGE_SHIFT) + 1;
drivers/pci/mmap.c
73
pci_start >>= PAGE_SHIFT;
drivers/pci/p2pdma.c
343
p2pdma->pool = gen_pool_create(PAGE_SHIFT, dev_to_node(&pdev->dev));
drivers/pci/pci.c
6389
align_order = PAGE_SHIFT;
drivers/pci/pci.c
6392
align_order = PAGE_SHIFT;
drivers/pci/proc.c
289
vma->vm_pgoff -= start >> PAGE_SHIFT;
drivers/perf/arm_spe_pmu.c
104
((idx) % ((unsigned long)(buf)->nr_pages << PAGE_SHIFT))
drivers/platform/goldfish/goldfish_pipe.c
264
int requested_pages = ((last_page - first_page) >> PAGE_SHIFT) + 1;
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1482
(((addrs[k - 1] & ~PAGE_MASK) + 1) << PAGE_SHIFT);
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1653
addrs[k - 1] += ((len + PAGE_SIZE - 1) >> PAGE_SHIFT);
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1656
(((len + PAGE_SIZE - 1) >> PAGE_SHIFT) - 1);
drivers/ptp/ptp_vmclock.c
384
st->res.start >> PAGE_SHIFT, PAGE_SIZE,
drivers/rapidio/devices/rio_mport_cdev.c
2192
baddr = ((dma_addr_t)vma->vm_pgoff << PAGE_SHIFT);
drivers/rapidio/devices/rio_mport_cdev.c
2212
vma->vm_pgoff = offset >> PAGE_SHIFT;
drivers/rapidio/devices/rio_mport_cdev.c
836
nr_pages = PAGE_ALIGN(xfer->length + offset) >> PAGE_SHIFT;
drivers/ras/cec.c
351
ca->array[to] = pfn << PAGE_SHIFT;
drivers/ras/cec.c
362
u64 pfn = ca->array[to] >> PAGE_SHIFT;
drivers/ras/cec.c
546
if (!cec_add_elem(m->addr >> PAGE_SHIFT)) {
drivers/ras/cec.c
81
#define COUNT_BITS (PAGE_SHIFT - DECAY_BITS)
drivers/ras/cec.c
89
#define PFN(e) ((e) >> PAGE_SHIFT)
drivers/s390/block/dcssblk.c
57
memset(kaddr, 0, nr_pages << PAGE_SHIFT);
drivers/s390/block/dcssblk.c
58
dax_flush(dax_dev, kaddr, nr_pages << PAGE_SHIFT);
drivers/s390/char/sclp_mem.c
226
__arch_set_page_nodat((void *)__va(addr), block_size >> PAGE_SHIFT);
drivers/s390/char/vmcp.c
65
nr_pages = ALIGN(session->bufsize, PAGE_SIZE) >> PAGE_SHIFT;
drivers/s390/char/vmcp.c
89
nr_pages = ALIGN(session->bufsize, PAGE_SIZE) >> PAGE_SHIFT;
drivers/s390/cio/itcw.c
125
>> PAGE_SHIFT);
drivers/s390/cio/itcw.c
130
>> PAGE_SHIFT);
drivers/s390/cio/itcw.c
204
>> PAGE_SHIFT);
drivers/s390/cio/itcw.c
209
>> PAGE_SHIFT);
drivers/s390/cio/vfio_ccw_cp.c
178
u64 iova_pfn_start = iova >> PAGE_SHIFT;
drivers/s390/cio/vfio_ccw_cp.c
179
u64 iova_pfn_end = (iova + length - 1) >> PAGE_SHIFT;
drivers/s390/cio/vfio_ccw_cp.c
184
pfn = pa->pa_iova[i] >> PAGE_SHIFT;
drivers/sbus/char/flash.c
62
if ((vma->vm_pgoff << PAGE_SHIFT) > size)
drivers/sbus/char/flash.c
64
addr = vma->vm_pgoff + (addr >> PAGE_SHIFT);
drivers/sbus/char/flash.c
66
if (vma->vm_end - (vma->vm_start + (vma->vm_pgoff << PAGE_SHIFT)) > size)
drivers/sbus/char/flash.c
67
size = vma->vm_end - (vma->vm_start + (vma->vm_pgoff << PAGE_SHIFT));
drivers/sbus/char/oradax.c
393
if (remap_pfn_range(vma, vma->vm_start, ctx->ca_buf_ra >> PAGE_SHIFT,
drivers/scsi/bnx2fc/bnx2fc_hwi.c
94
fcoe_init1.flags = (PAGE_SHIFT <<
drivers/scsi/csiostor/csio_scsi.c
1532
sg_addr = kmap_atomic(sg_page(sg) + (sg_off >> PAGE_SHIFT));
drivers/scsi/csiostor/csio_wr.c
1313
uint32_t s_hps = PAGE_SHIFT - 10;
drivers/scsi/csiostor/csio_wr.c
1419
csio_wr_reg32(hw, HPZ0_V(PAGE_SHIFT - 12), ULP_RX_TDDP_PSZ_A);
drivers/scsi/cxgbi/libcxgbi.c
1292
ttinfo->nr_pages = (xferlen + sgl->offset + (1 << PAGE_SHIFT) - 1) >>
drivers/scsi/cxgbi/libcxgbi.c
1293
PAGE_SHIFT;
drivers/scsi/cxgbi/libcxgbi.h
81
min_t(u32, MAX_SKB_FRAGS << PAGE_SHIFT, CXGBI_ULP2_MAX_ISO_PAYLOAD)
drivers/scsi/fcoe/fcoe_transport.c
327
skb_frag_page(frag) + (off >> PAGE_SHIFT));
drivers/scsi/ipr.c
4251
src = (u8 *)dump->ioa_dump.ioa_data[(off & PAGE_MASK) >> PAGE_SHIFT];
drivers/scsi/libfc/fc_fcp.c
677
page = sg_page(sg) + (off >> PAGE_SHIFT);
drivers/scsi/libfc/fc_libfc.c
132
page_addr = kmap_atomic(sg_page(sg) + (off >> PAGE_SHIFT));
drivers/scsi/mpt3sas/mpt3sas_base.c
3023
#define convert_to_kb(x) ((x) << (PAGE_SHIFT - 10))
drivers/scsi/ncr53c8xx.c
195
#define MEMO_CLUSTER_SHIFT (PAGE_SHIFT+MEMO_PAGE_ORDER)
drivers/scsi/ncr53c8xx.c
224
struct m_link h[PAGE_SHIFT-MEMO_SHIFT+MEMO_PAGE_ORDER+1];
drivers/scsi/scsi_lib.c
3226
page = sg_page(sg) + (*offset >> PAGE_SHIFT);
drivers/scsi/sd_zbc.c
186
bufsize = min_t(size_t, bufsize, max_segments << PAGE_SHIFT);
drivers/scsi/sg.c
1228
offset = vmf->pgoff << PAGE_SHIFT;
drivers/scsi/sg.c
1235
length = 1 << (PAGE_SHIFT + rsv_schp->page_order);
drivers/scsi/sg.c
1240
struct page *page = rsv_schp->pages[k] + (offset >> PAGE_SHIFT);
drivers/scsi/sg.c
1281
length = 1 << (PAGE_SHIFT + rsv_schp->page_order);
drivers/scsi/sg.c
1898
ret_sz = 1 << (PAGE_SHIFT + order);
drivers/scsi/sg.c
1977
num = 1 << (PAGE_SHIFT + schp->page_order);
drivers/scsi/sg.c
2028
num = 1 << (PAGE_SHIFT + rsv_schp->page_order);
drivers/scsi/st.c
4990
unsigned long end = (uaddr + count + PAGE_SIZE - 1) >> PAGE_SHIFT;
drivers/scsi/st.c
4991
unsigned long start = uaddr >> PAGE_SHIFT;
drivers/scsi/sym53c8xx_2/sym_hipd.h
144
#define SYM_MEM_CLUSTER_SHIFT (PAGE_SHIFT+SYM_MEM_PAGE_ORDER)
drivers/scsi/vmw_pvscsi.c
492
cmd.ringsStatePPN = adapter->ringStatePA >> PAGE_SHIFT;
drivers/scsi/vmw_pvscsi.c
498
cmd.reqRingPPNs[i] = base >> PAGE_SHIFT;
drivers/scsi/vmw_pvscsi.c
504
cmd.cmpRingPPNs[i] = base >> PAGE_SHIFT;
drivers/scsi/vmw_pvscsi.c
522
cmd_msg.ringPPNs[i] = base >> PAGE_SHIFT;
drivers/soc/aspeed/aspeed-lpc-ctrl.c
56
if (vma->vm_pgoff + vma_pages(vma) > lpc_ctrl->mem_size >> PAGE_SHIFT)
drivers/soc/aspeed/aspeed-lpc-ctrl.c
63
(lpc_ctrl->mem_base >> PAGE_SHIFT) + vma->vm_pgoff,
drivers/soc/aspeed/aspeed-p2a-ctrl.c
113
if (vma->vm_pgoff + vma_pages(vma) > ctrl->mem_size >> PAGE_SHIFT)
drivers/soc/aspeed/aspeed-p2a-ctrl.c
120
(ctrl->mem_base >> PAGE_SHIFT) + vma->vm_pgoff,
drivers/soc/qcom/rmtfs_mem.c
147
rmtfs_mem->addr >> PAGE_SHIFT,
drivers/staging/media/atomisp/pci/hmm/hmm.c
186
idx = (virt - bo->start) >> PAGE_SHIFT;
drivers/staging/media/atomisp/pci/hmm/hmm.c
187
offset = (virt - bo->start) - (idx << PAGE_SHIFT);
drivers/staging/media/atomisp/pci/hmm/hmm.c
318
idx = (virt - bo->start) >> PAGE_SHIFT;
drivers/staging/media/atomisp/pci/hmm/hmm.c
319
offset = (virt - bo->start) - (idx << PAGE_SHIFT);
drivers/staging/media/atomisp/pci/hmm/hmm.c
389
idx = (virt - bo->start) >> PAGE_SHIFT;
drivers/staging/media/atomisp/pci/hmm/hmm.c
390
offset = (virt - bo->start) - (idx << PAGE_SHIFT);
drivers/staging/media/atomisp/pci/hmm/hmm.c
428
idx = (virt - bo->start) >> PAGE_SHIFT;
drivers/staging/media/atomisp/pci/hmm/hmm.c
429
offset = (virt - bo->start) - (idx << PAGE_SHIFT);
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
785
virt += (1 << PAGE_SHIFT);
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
801
(bo->pgnr << PAGE_SHIFT));
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
865
(bo->pgnr << PAGE_SHIFT));
drivers/staging/media/ipu3/ipu3-dmamap.c
23
int count = size >> PAGE_SHIFT;
drivers/staging/media/ipu3/ipu3-dmamap.c
37
unsigned int i = 0, count = size >> PAGE_SHIFT;
drivers/staging/media/ipu3/ipu3-dmamap.c
72
imgu_dmamap_free_buffer(pages, i << PAGE_SHIFT);
drivers/staging/media/ipu3/ipu3-dmamap.c
99
int count = size >> PAGE_SHIFT;
drivers/staging/media/ipu3/ipu3-mmu.c
26
#define IPU3_PT_ORDER (IPU3_PT_SIZE >> PAGE_SHIFT)
drivers/staging/media/meson/vdec/codec_hevc_common.c
290
first_page = comm->fbc_buffer_paddr[vb->index] >> PAGE_SHIFT;
drivers/staging/media/meson/vdec/codec_hevc_common.c
292
first_page = vb2_dma_contig_plane_dma_addr(vb, 0) >> PAGE_SHIFT;
drivers/staging/vme_user/vme.c
762
phys_addr = image->bus_resource.start + (desc->pgoff << PAGE_SHIFT);
drivers/target/iscsi/cxgbit/cxgbit_ddp.c
192
(1 << PAGE_SHIFT) - 1) >> PAGE_SHIFT;
drivers/target/target_core_pscsi.c
832
PAGE_SIZE - 1) >> PAGE_SHIFT;
drivers/target/target_core_user.c
1910
offset = (vmf->pgoff - mi) << PAGE_SHIFT;
drivers/target/target_core_user.c
2239
data_size = TCMU_MBS_TO_PAGES(udev->data_area_mb) << PAGE_SHIFT;
drivers/target/target_core_user.c
2240
udev->mmap_pages = (data_size + udev->cmdr_size + CMDR_OFF) >> PAGE_SHIFT;
drivers/target/target_core_user.c
77
#define TCMU_MBS_TO_PAGES(_mbs) ((size_t)_mbs << (20 - PAGE_SHIFT))
drivers/target/target_core_user.c
78
#define TCMU_PAGES_TO_MBS(_pages) (_pages >> (20 - PAGE_SHIFT))
drivers/target/tcm_fc/tfc_io.c
141
from = kmap_atomic(page + (mem_off >> PAGE_SHIFT));
drivers/target/tcm_fc/tfc_io.c
300
to = kmap_atomic(page + (mem_off >> PAGE_SHIFT));
drivers/tee/optee/protmem.c
60
rp->gen_pool = gen_pool_create(PAGE_SHIFT, -1);
drivers/tee/tee_heap.c
479
stp->gen_pool = gen_pool_create(PAGE_SHIFT, -1);
drivers/tee/tee_shm.c
579
if (vma->vm_pgoff + vma_pages(vma) > shm->size >> PAGE_SHIFT)
drivers/tee/tee_shm.c
582
return remap_pfn_range(vma, vma->vm_start, shm->paddr >> PAGE_SHIFT,
drivers/thermal/intel/intel_hfi.c
313
hfi_features.nr_table_pages << PAGE_SHIFT);
drivers/thermal/intel/intel_hfi.c
467
hfi_instance->local_table = kzalloc(hfi_features.nr_table_pages << PAGE_SHIFT,
drivers/uio/uio.c
709
offset = (vmf->pgoff - mi) << PAGE_SHIFT;
drivers/uio/uio.c
772
mem->addr >> PAGE_SHIFT,
drivers/uio/uio.c
847
+ idev->info->mem[mi].size + PAGE_SIZE -1) >> PAGE_SHIFT;
drivers/uio/uio_hv_generic.c
165
channel->ringbuffer_pagecount << PAGE_SHIFT);
drivers/uio/uio_hv_generic.c
290
= channel->ringbuffer_pagecount << PAGE_SHIFT;
drivers/usb/core/devio.c
280
virt_to_phys(usbm->mem) >> PAGE_SHIFT,
drivers/usb/gadget/function/f_fs.c
1929
sb->s_blocksize_bits = PAGE_SHIFT;
drivers/usb/gadget/function/f_fs.c
816
n_pages = PAGE_ALIGN(sz) >> PAGE_SHIFT;
drivers/usb/gadget/legacy/inode.c
2036
sb->s_blocksize_bits = PAGE_SHIFT;
drivers/usb/host/xhci-sideband.c
30
n_pages = PAGE_ALIGN(sz) >> PAGE_SHIFT;
drivers/usb/mon/mon_bin.c
1257
offset = vmf->pgoff << PAGE_SHIFT;
drivers/vdpa/mlx5/net/mlx5_vnet.c
3478
if (MLX5_CAP_GEN(mvdev->mdev, log_min_sf_size) + 12 < PAGE_SHIFT)
drivers/vdpa/vdpa_user/iova_domain.c
231
page = pfn_to_page((map->addr + iova - map->start) >> PAGE_SHIFT);
drivers/vdpa/vdpa_user/iova_domain.c
291
if (count != (domain->bounce_size >> PAGE_SHIFT))
drivers/vdpa/vdpa_user/iova_domain.c
339
count = domain->bounce_size >> PAGE_SHIFT;
drivers/vdpa/vdpa_user/iova_domain.c
549
unsigned long iova = vmf->pgoff << PAGE_SHIFT;
drivers/vdpa/vdpa_user/vduse_dev.c
1232
npages = size >> PAGE_SHIFT;
drivers/vfio/cdx/main.c
234
((1U << (VFIO_CDX_OFFSET_SHIFT - PAGE_SHIFT)) - 1);
drivers/vfio/cdx/main.c
235
base = pgoff << PAGE_SHIFT;
drivers/vfio/cdx/main.c
240
vma->vm_pgoff = (region.addr >> PAGE_SHIFT) + pgoff;
drivers/vfio/cdx/main.c
255
index = vma->vm_pgoff >> (VFIO_CDX_OFFSET_SHIFT - PAGE_SHIFT);
drivers/vfio/fsl-mc/vfio_fsl_mc.c
361
((1U << (VFIO_FSL_MC_OFFSET_SHIFT - PAGE_SHIFT)) - 1);
drivers/vfio/fsl-mc/vfio_fsl_mc.c
362
base = pgoff << PAGE_SHIFT;
drivers/vfio/fsl-mc/vfio_fsl_mc.c
372
vma->vm_pgoff = (region.addr >> PAGE_SHIFT) + pgoff;
drivers/vfio/fsl-mc/vfio_fsl_mc.c
386
index = vma->vm_pgoff >> (VFIO_FSL_MC_OFFSET_SHIFT - PAGE_SHIFT);
drivers/vfio/pci/hisilicon/hisi_acc_vfio_pci.c
1364
index = vma->vm_pgoff >> (VFIO_PCI_OFFSET_SHIFT - PAGE_SHIFT);
drivers/vfio/pci/hisilicon/hisi_acc_vfio_pci.c
1372
((1U << (VFIO_PCI_OFFSET_SHIFT - PAGE_SHIFT)) - 1);
drivers/vfio/pci/hisilicon/hisi_acc_vfio_pci.c
1373
req_start = pgoff << PAGE_SHIFT;
drivers/vfio/pci/mlx5/cmd.c
341
MLX5_SET(mkc, mkc, log_page_size, PAGE_SHIFT);
drivers/vfio/pci/nvgrace-gpu/main.c
106
num_pages = region->memlength >> PAGE_SHIFT;
drivers/vfio/pci/nvgrace-gpu/main.c
125
vma->vm_pgoff >> (VFIO_PCI_OFFSET_SHIFT - PAGE_SHIFT);
drivers/vfio/pci/nvgrace-gpu/main.c
127
((1U << (VFIO_PCI_OFFSET_SHIFT - PAGE_SHIFT)) - 1);
drivers/vfio/pci/nvgrace-gpu/main.c
157
nr_pages = region->memlength >> PAGE_SHIFT;
drivers/vfio/pci/nvgrace-gpu/main.c
291
((1U << (VFIO_PCI_OFFSET_SHIFT - PAGE_SHIFT)) - 1);
drivers/vfio/pci/nvgrace-gpu/main.c
293
return ((addr - vma->vm_start) >> PAGE_SHIFT) + pgoff;
drivers/vfio/pci/nvgrace-gpu/main.c
303
vma->vm_pgoff >> (VFIO_PCI_OFFSET_SHIFT - PAGE_SHIFT);
drivers/vfio/pci/nvgrace-gpu/main.c
365
index = vma->vm_pgoff >> (VFIO_PCI_OFFSET_SHIFT - PAGE_SHIFT);
drivers/vfio/pci/nvgrace-gpu/main.c
377
((1U << (VFIO_PCI_OFFSET_SHIFT - PAGE_SHIFT)) - 1);
drivers/vfio/pci/vfio_pci_core.c
1654
int index = vma->vm_pgoff >> (VFIO_PCI_OFFSET_SHIFT - PAGE_SHIFT);
drivers/vfio/pci/vfio_pci_core.c
1658
((1U << (VFIO_PCI_OFFSET_SHIFT - PAGE_SHIFT)) - 1);
drivers/vfio/pci/vfio_pci_core.c
1660
return (pci_resource_start(vdev->pdev, index) >> PAGE_SHIFT) + pgoff;
drivers/vfio/pci/vfio_pci_core.c
1692
unsigned long pgoff = (addr - vma->vm_start) >> PAGE_SHIFT;
drivers/vfio/pci/vfio_pci_core.c
1705
(VFIO_PCI_OFFSET_SHIFT - PAGE_SHIFT),
drivers/vfio/pci/vfio_pci_core.c
1732
index = vma->vm_pgoff >> (VFIO_PCI_OFFSET_SHIFT - PAGE_SHIFT);
drivers/vfio/pci/vfio_pci_core.c
1757
((1U << (VFIO_PCI_OFFSET_SHIFT - PAGE_SHIFT)) - 1);
drivers/vfio/pci/vfio_pci_core.c
1758
req_start = pgoff << PAGE_SHIFT;
drivers/vfio/pci/virtio/migrate.c
87
filled, 0, filled << PAGE_SHIFT, UINT_MAX,
drivers/vfio/platform/vfio_platform_common.c
555
((1U << (VFIO_PLATFORM_OFFSET_SHIFT - PAGE_SHIFT)) - 1);
drivers/vfio/platform/vfio_platform_common.c
556
req_start = pgoff << PAGE_SHIFT;
drivers/vfio/platform/vfio_platform_common.c
562
vma->vm_pgoff = (region.addr >> PAGE_SHIFT) + pgoff;
drivers/vfio/platform/vfio_platform_common.c
574
index = vma->vm_pgoff >> (VFIO_PLATFORM_OFFSET_SHIFT - PAGE_SHIFT);
drivers/vfio/vfio_iommu_spapr_tce.c
115
mem = mm_iommu_get(container->mm, vaddr, size >> PAGE_SHIFT);
drivers/vfio/vfio_iommu_spapr_tce.c
142
unsigned long entries = size >> PAGE_SHIFT;
drivers/vfio/vfio_iommu_spapr_tce.c
189
page = pfn_to_page(hpa >> PAGE_SHIFT);
drivers/vfio/vfio_iommu_spapr_tce.c
293
locked = table_group->tce32_size >> PAGE_SHIFT;
drivers/vfio/vfio_iommu_spapr_tce.c
385
page = pfn_to_page(hpa >> PAGE_SHIFT);
drivers/vfio/vfio_iommu_spapr_tce.c
621
ret = account_locked_vm(container->mm, table_size >> PAGE_SHIFT, true);
drivers/vfio/vfio_iommu_spapr_tce.c
637
unsigned long pages = tbl->it_allocated_size >> PAGE_SHIFT;
drivers/vfio/vfio_iommu_type1.c
1071
entry->phys >> PAGE_SHIFT,
drivers/vfio/vfio_iommu_type1.c
1072
entry->len >> PAGE_SHIFT,
drivers/vfio/vfio_iommu_type1.c
1140
phys >> PAGE_SHIFT,
drivers/vfio/vfio_iommu_type1.c
1141
unmapped >> PAGE_SHIFT,
drivers/vfio/vfio_iommu_type1.c
1561
ret = iommu_map(d->domain, iova, (phys_addr_t)pfn << PAGE_SHIFT,
drivers/vfio/vfio_iommu_type1.c
1562
npage << PAGE_SHIFT, prot | IOMMU_CACHE,
drivers/vfio/vfio_iommu_type1.c
1574
iommu_unmap(d->domain, iova, npage << PAGE_SHIFT);
drivers/vfio/vfio_iommu_type1.c
1589
unsigned long pfn, limit = rlimit(RLIMIT_MEMLOCK) >> PAGE_SHIFT;
drivers/vfio/vfio_iommu_type1.c
1597
size >> PAGE_SHIFT, &pfn, limit,
drivers/vfio/vfio_iommu_type1.c
1615
size -= npage << PAGE_SHIFT;
drivers/vfio/vfio_iommu_type1.c
1616
dma->size += npage << PAGE_SHIFT;
drivers/vfio/vfio_iommu_type1.c
1810
unsigned long limit = rlimit(RLIMIT_MEMLOCK) >> PAGE_SHIFT;
drivers/vfio/vfio_iommu_type1.c
1865
n >> PAGE_SHIFT,
drivers/vfio/vfio_iommu_type1.c
1874
phys = pfn << PAGE_SHIFT;
drivers/vfio/vfio_iommu_type1.c
1875
size = npage << PAGE_SHIFT;
drivers/vfio/vfio_iommu_type1.c
1884
phys >> PAGE_SHIFT,
drivers/vfio/vfio_iommu_type1.c
1885
size >> PAGE_SHIFT,
drivers/vfio/vfio_iommu_type1.c
1939
vfio_unpin_pages_remote(dma, iova, phys >> PAGE_SHIFT,
drivers/vfio/vfio_iommu_type1.c
1940
size >> PAGE_SHIFT, true);
drivers/vfio/vfio_iommu_type1.c
626
epfn = (*pfn | (~addr_mask >> PAGE_SHIFT)) + 1;
drivers/vfio/vfio_iommu_type1.c
642
dma_addr_t iova_end = iova_start + (nr_pages << PAGE_SHIFT);
drivers/vfio/vfio_iommu_type1.c
772
__func__, limit << PAGE_SHIFT);
drivers/vhost/vdpa.c
1072
offset = (vma->vm_pgoff << PAGE_SHIFT) + uaddr - vma->vm_start;
drivers/vhost/vringh.c
1079
u64 pfn = io_addr >> PAGE_SHIFT;
drivers/video/fbdev/aty/atyfb_base.c
1946
if (vma->vm_pgoff > (~0UL >> PAGE_SHIFT))
drivers/video/fbdev/aty/atyfb_base.c
1949
off = vma->vm_pgoff << PAGE_SHIFT;
drivers/video/fbdev/aty/atyfb_base.c
1958
vma->vm_pgoff = off >> PAGE_SHIFT; /* propagate off changes */
drivers/video/fbdev/aty/atyfb_base.c
1988
map_offset >> PAGE_SHIFT, map_size, vma->vm_page_prot))
drivers/video/fbdev/controlfb.c
740
mmio_pgoff = PAGE_ALIGN((start & ~PAGE_MASK) + len) >> PAGE_SHIFT;
drivers/video/fbdev/core/fb_defio.c
138
page = pfn_to_page((info->fix.smem_start + offs) >> PAGE_SHIFT);
drivers/video/fbdev/core/fb_defio.c
151
unsigned long pgoff = offset >> PAGE_SHIFT;
drivers/video/fbdev/core/fb_defio.c
164
pageref->offset = pgoff << PAGE_SHIFT;
drivers/video/fbdev/core/fb_defio.c
239
offset = vmf->pgoff << PAGE_SHIFT;
drivers/video/fbdev/core/fb_defio.c
340
unsigned long offset = vmf->pgoff << PAGE_SHIFT;
drivers/video/fbdev/core/fb_defio.c
398
pgoff_t pgoff = pageref->offset >> PAGE_SHIFT;
drivers/video/fbdev/core/fb_io_fops.c
146
unsigned long mmio_pgoff = PAGE_ALIGN((start & ~PAGE_MASK) + len) >> PAGE_SHIFT;
drivers/video/fbdev/ep93xx-fb.c
312
unsigned int offset = vma->vm_pgoff << PAGE_SHIFT;
drivers/video/fbdev/gbefb.c
1007
if (vma->vm_pgoff > (~0UL >> PAGE_SHIFT))
drivers/video/fbdev/gbefb.c
1035
if (remap_pfn_range(vma, addr, phys_addr >> PAGE_SHIFT,
drivers/video/fbdev/gbefb.c
999
unsigned long offset = vma->vm_pgoff << PAGE_SHIFT;
drivers/video/fbdev/metronomefb.c
476
unsigned long pgoffset = pageref->offset >> PAGE_SHIFT;
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
1106
vma->vm_pgoff << PAGE_SHIFT);
drivers/video/fbdev/ps3fb.c
716
info->fix.smem_start + (vma->vm_pgoff << PAGE_SHIFT),
drivers/video/fbdev/pvr2fb.c
653
nr_pages = (count + PAGE_SIZE - 1) >> PAGE_SHIFT;
drivers/video/fbdev/pvr2fb.c
685
len = nr_pages << PAGE_SHIFT;
drivers/video/fbdev/pvr2fb.c
704
if ((*ppos + (i << PAGE_SHIFT)) > fb_info->fix.smem_len) {
drivers/video/fbdev/pxa3xx-gcu.c
486
case SHARED_SIZE >> PAGE_SHIFT:
drivers/video/fbdev/pxa3xx-gcu.c
495
priv->resource_mem->start >> PAGE_SHIFT,
drivers/video/fbdev/sa1100fb.c
563
unsigned long off = vma->vm_pgoff << PAGE_SHIFT;
drivers/video/fbdev/sbuslib.c
57
if (vma->vm_pgoff > (~0UL >> PAGE_SHIFT))
drivers/video/fbdev/sbuslib.c
60
off = vma->vm_pgoff << PAGE_SHIFT;
drivers/video/fbdev/sbuslib.c
90
map_offset >> PAGE_SHIFT),
drivers/video/fbdev/sh_mobile_lcdcfb.c
1998
ch->fb_size >> PAGE_SHIFT);
drivers/video/fbdev/sh_mobile_lcdcfb.c
441
unsigned int nr_pages_max = ch->fb_size >> PAGE_SHIFT;
drivers/video/fbdev/smscufx.c
780
unsigned long offset = vma->vm_pgoff << PAGE_SHIFT;
drivers/video/fbdev/smscufx.c
788
if (vma->vm_pgoff > (~0UL >> PAGE_SHIFT))
drivers/video/fbdev/udlfb.c
328
unsigned long offset = vma->vm_pgoff << PAGE_SHIFT;
drivers/video/fbdev/udlfb.c
336
if (vma->vm_pgoff > (~0UL >> PAGE_SHIFT))
drivers/video/fbdev/xen-fbfront.c
391
info->nr_pages = (fb_size + PAGE_SIZE - 1) >> PAGE_SHIFT;
drivers/virt/acrn/mm.c
172
nr_pages = memmap->len >> PAGE_SHIFT;
drivers/virt/coco/sev-guest/sev-guest.c
213
npages = report_req->certs_len >> PAGE_SHIFT;
drivers/virt/coco/sev-guest/sev-guest.c
256
report_req->certs_len = req.input.data_npages << PAGE_SHIFT;
drivers/virt/coco/tdx-guest/tdx-guest.c
216
unsigned int count = len >> PAGE_SHIFT;
drivers/virt/coco/tdx-guest/tdx-guest.c
229
unsigned int count = len >> PAGE_SHIFT;
drivers/virt/fsl_hypervisor.c
221
num_pages = (param.count + lb_offset + PAGE_SIZE - 1) >> PAGE_SHIFT;
drivers/virt/vboxguest/vboxguest_core.c
77
pages = kmalloc_objs(*pages, size >> PAGE_SHIFT);
drivers/virt/vboxguest/vboxguest_core.c
85
for (i = 0; i < (size >> PAGE_SHIFT); i++)
drivers/virt/vboxguest/vboxguest_core.c
93
guest_mappings[i] = vmap(pages, (size >> PAGE_SHIFT),
drivers/virt/vboxguest/vboxguest_utils.c
194
return size >> PAGE_SHIFT;
drivers/virtio/virtio_balloon.c
141
BUILD_BUG_ON(PAGE_SHIFT < VIRTIO_BALLOON_PFN_SHIFT);
drivers/virtio/virtio_balloon.c
345
#define pages_to_bytes(x) ((u64)(x) << PAGE_SHIFT)
drivers/virtio/virtio_balloon.c
39
(1 << (VIRTIO_BALLOON_HINT_BLOCK_ORDER + PAGE_SHIFT))
drivers/virtio/virtio_mem.c
1335
order = ilog2(vm->sbm.sb_size) - PAGE_SHIFT;
drivers/virtio/virtio_mmio.c
393
u64 q_pfn = virtqueue_get_desc_addr(vq) >> PAGE_SHIFT;
drivers/virtio/virtio_mmio.c
403
0x1ULL << (32 + PAGE_SHIFT - 30));
drivers/virtio/virtio_mmio.c
631
DMA_BIT_MASK(32 + PAGE_SHIFT));
drivers/virtio/virtio_pci_legacy.c
144
0x1ULL << (32 + PAGE_SHIFT - 30));
drivers/xen/gntalloc.c
513
gref = find_grefs(priv, vma->vm_pgoff << PAGE_SHIFT, count);
drivers/xen/gntdev-dmabuf.c
208
nr_pages << PAGE_SHIFT,
drivers/xen/gntdev-dmabuf.c
383
exp_info.size = args->count << PAGE_SHIFT;
drivers/xen/gntdev-dmabuf.c
603
if (attach->dmabuf->size != gntdev_dmabuf->nr_pages << PAGE_SHIFT) {
drivers/xen/gntdev.c
307
unsigned int pgnr = (addr - map->pages_vm_start) >> PAGE_SHIFT;
drivers/xen/gntdev.c
447
int pgno = (map->notify.addr >> PAGE_SHIFT);
drivers/xen/gntdev.c
528
return map->pages[(addr - map->pages_vm_start) >> PAGE_SHIFT];
drivers/xen/gntdev.c
552
map_end = map->pages_vm_start + (map->count << PAGE_SHIFT);
drivers/xen/gntdev.c
568
unmap_grant_pages(map, (mstart - map_start) >> PAGE_SHIFT,
drivers/xen/gntdev.c
569
(mend - mstart) >> PAGE_SHIFT);
drivers/xen/gntdev.c
672
op.index = map->index << PAGE_SHIFT;
drivers/xen/gntdev.c
693
map = gntdev_find_map_index(priv, op.index >> PAGE_SHIFT, op.count);
drivers/xen/gntdev.c
725
op.offset = map->index << PAGE_SHIFT;
drivers/xen/gntdev.c
769
uint64_t begin = map->index << PAGE_SHIFT;
drivers/xen/gntdev.c
770
uint64_t end = (map->index + map->count) << PAGE_SHIFT;
drivers/xen/gntdev.c
788
map->notify.addr = op.index - (map->index << PAGE_SHIFT);
drivers/xen/grant-table.c
1066
if (args->nr_pages < 0 || args->nr_pages > (INT_MAX >> PAGE_SHIFT))
drivers/xen/grant-table.c
1069
size = args->nr_pages << PAGE_SHIFT;
drivers/xen/grant-table.c
1139
size = args->nr_pages << PAGE_SHIFT;
drivers/xen/grant-table.c
1536
return width > 32 + PAGE_SHIFT;
drivers/xen/mem-reservation.c
51
(unsigned long)__va(pfn << PAGE_SHIFT),
drivers/xen/mem-reservation.c
75
(unsigned long)__va(pfn << PAGE_SHIFT),
drivers/xen/privcmd.c
1664
return apply_to_page_range(vma->vm_mm, addr, nr_pages << PAGE_SHIFT,
drivers/xen/privcmd.c
256
if ((msg->npages > (LONG_MAX >> PAGE_SHIFT)) ||
drivers/xen/privcmd.c
257
((unsigned long)(msg->npages << PAGE_SHIFT) >= -st->va))
drivers/xen/privcmd.c
262
((msg->va+(msg->npages<<PAGE_SHIFT)) > vma->vm_end))
drivers/xen/privcmd.c
273
st->va += msg->npages << PAGE_SHIFT;
drivers/xen/privcmd.c
507
if ((m.num <= 0) || (nr_pages > (LONG_MAX >> PAGE_SHIFT)))
drivers/xen/privcmd.c
549
m.addr + (nr_pages << PAGE_SHIFT) != vma->vm_end) {
drivers/xen/privcmd.c
561
m.addr + (nr_pages << PAGE_SHIFT) > vma->vm_end) {
drivers/xen/privcmd.c
829
rc = xen_remap_vma_range(vma, kdata.addr, kdata.num << PAGE_SHIFT);
drivers/xen/swiotlb-xen.c
120
unsigned int i, dma_bits = order + PAGE_SHIFT;
drivers/xen/swiotlb-xen.c
79
phys_addr_t algn = 1ULL << (get_order(size) + PAGE_SHIFT);
drivers/xen/xen-balloon.c
173
balloon_set_new_target(target_bytes >> PAGE_SHIFT);
drivers/xen/xen-balloon.c
185
<< PAGE_SHIFT);
drivers/xen/xen-balloon.c
200
balloon_set_new_target(target_bytes >> PAGE_SHIFT);
drivers/xen/xen-balloon.c
50
#define PAGES2KB(_p) ((_p)<<(PAGE_SHIFT-10))
drivers/xen/xen-balloon.c
86
new_target >>= PAGE_SHIFT - 10;
drivers/xen/xen-balloon.c
95
static_max >>= PAGE_SHIFT - 10;
drivers/xen/xlate_mmu.c
152
unsigned long range = DIV_ROUND_UP(nr, XEN_PFN_PER_PAGE) << PAGE_SHIFT;
fs/9p/vfs_super.c
62
sb->s_bdi->ra_pages = v9ses->maxdata >> PAGE_SHIFT;
fs/9p/vfs_super.c
63
sb->s_bdi->io_pages = v9ses->maxdata >> PAGE_SHIFT;
fs/affs/file.c
673
index = pos >> PAGE_SHIFT;
fs/affs/file.c
721
tmp = (folio->index << PAGE_SHIFT) + from;
fs/affs/file.c
820
tmp = (folio->index << PAGE_SHIFT) + from;
fs/afs/super.c
458
sb->s_blocksize_bits = PAGE_SHIFT;
fs/btrfs/compression.c
1002
in_folio = filemap_get_folio(mapping, start >> PAGE_SHIFT);
fs/btrfs/compression.c
1205
return (page_pgoff(folio, page) << PAGE_SHIFT) + bvec->bv_offset;
fs/btrfs/compression.c
1553
index = start >> PAGE_SHIFT;
fs/btrfs/compression.c
1554
index_end = end >> PAGE_SHIFT;
fs/btrfs/compression.c
254
pgoff_t index = cb->start >> PAGE_SHIFT;
fs/btrfs/compression.c
255
const pgoff_t end_index = (cb->start + cb->len - 1) >> PAGE_SHIFT;
fs/btrfs/compression.c
408
end_index = (i_size_read(inode) - 1) >> PAGE_SHIFT;
fs/btrfs/compression.c
425
pgoff_t pg_index = cur >> PAGE_SHIFT;
fs/btrfs/compression.c
476
page_end = (pg_index << PAGE_SHIFT) + folio_size(folio) - 1;
fs/btrfs/defrag.c
1202
const unsigned int nr_pages = ((start + len - 1) >> PAGE_SHIFT) -
fs/btrfs/defrag.c
1203
(start >> PAGE_SHIFT) + 1;
fs/btrfs/defrag.c
1215
folios[i] = defrag_prepare_one_folio(inode, cur >> PAGE_SHIFT);
fs/btrfs/defrag.c
1313
ra, NULL, entry->start >> PAGE_SHIFT,
fs/btrfs/defrag.c
1314
((entry->start + range_len - 1) >> PAGE_SHIFT) -
fs/btrfs/defrag.c
1315
(entry->start >> PAGE_SHIFT) + 1);
fs/btrfs/defrag.c
1421
start_index = cur >> PAGE_SHIFT;
fs/btrfs/defrag.c
1436
cluster_end = (((cur >> PAGE_SHIFT) +
fs/btrfs/defrag.c
1437
(SZ_256K >> PAGE_SHIFT)) << PAGE_SHIFT) - 1;
fs/btrfs/direct-io.c
994
invalidate_mapping_pages(file->f_mapping, pos >> PAGE_SHIFT,
fs/btrfs/direct-io.c
995
endbyte >> PAGE_SHIFT);
fs/btrfs/disk-io.c
3400
fs_info->block_min_order = ilog2(round_up(sectorsize, PAGE_SIZE) >> PAGE_SHIFT);
fs/btrfs/disk-io.c
3401
fs_info->block_max_order = ilog2((BITS_PER_LONG << fs_info->sectorsize_bits) >> PAGE_SHIFT);
fs/btrfs/disk-io.c
3797
folio = __filemap_get_folio(mapping, bytenr >> PAGE_SHIFT,
fs/btrfs/disk-io.c
3872
bytenr >> PAGE_SHIFT);
fs/btrfs/extent_io.c
1017
if (folio_contains(folio, last_byte >> PAGE_SHIFT)) {
fs/btrfs/extent_io.c
1176
folio = filemap_get_folio(inode->vfs_inode.i_mapping, cur >> PAGE_SHIFT);
fs/btrfs/extent_io.c
1838
const pgoff_t end_index = i_size >> PAGE_SHIFT;
fs/btrfs/extent_io.c
2332
index = ((mapping->writeback_index << PAGE_SHIFT) >> fs_info->nodesize_bits);
fs/btrfs/extent_io.c
2466
index = wbc->range_start >> PAGE_SHIFT;
fs/btrfs/extent_io.c
2467
end = wbc->range_end >> PAGE_SHIFT;
fs/btrfs/extent_io.c
2625
folio = filemap_get_folio(mapping, cur >> PAGE_SHIFT);
fs/btrfs/extent_io.c
271
pgoff_t index = start >> PAGE_SHIFT;
fs/btrfs/extent_io.c
272
pgoff_t end_index = end >> PAGE_SHIFT;
fs/btrfs/extent_io.c
309
pgoff_t index = start >> PAGE_SHIFT;
fs/btrfs/extent_io.c
310
pgoff_t end_index = end >> PAGE_SHIFT;
fs/btrfs/extent_io.c
3364
const pgoff_t index = eb->start >> PAGE_SHIFT;
fs/btrfs/extent_io.c
605
if (folio_contains(folio, i_size >> PAGE_SHIFT) &&
fs/btrfs/extent_io.c
706
eb->folio_shift = PAGE_SHIFT;
fs/btrfs/extent_io.h
303
return (eb->len >> PAGE_SHIFT) ?: 1;
fs/btrfs/file.c
1950
if (folio_contains(folio, (size - 1) >> PAGE_SHIFT)) {
fs/btrfs/file.c
2244
const pgoff_t start_index = page_lockstart >> PAGE_SHIFT;
fs/btrfs/file.c
2245
const pgoff_t end_index = (page_lockend - 1) >> PAGE_SHIFT;
fs/btrfs/file.c
858
const pgoff_t index = pos >> PAGE_SHIFT;
fs/btrfs/free-space-cache.c
373
last_index = (i_size_read(inode) - 1) >> PAGE_SHIFT;
fs/btrfs/fs.h
978
return 1U << (PAGE_SHIFT + fs_info->block_min_order);
fs/btrfs/inode.c
10016
start >> PAGE_SHIFT,
fs/btrfs/inode.c
10017
end >> PAGE_SHIFT);
fs/btrfs/inode.c
10228
first_ppage = PAGE_ALIGN(bsi->block_start) >> PAGE_SHIFT;
fs/btrfs/inode.c
10229
next_ppage = PAGE_ALIGN_DOWN(bsi->block_start + bsi->block_len) >> PAGE_SHIFT;
fs/btrfs/inode.c
1570
PAGE_SHIFT;
fs/btrfs/inode.c
3466
u32 pindex = offset_in_folio(folio, paddr + i * step) >> PAGE_SHIFT;
fs/btrfs/inode.c
404
pgoff_t index = offset >> PAGE_SHIFT;
fs/btrfs/inode.c
405
const pgoff_t end_index = (offset + bytes - 1) >> PAGE_SHIFT;
fs/btrfs/inode.c
4987
const pgoff_t index = (start >> PAGE_SHIFT);
fs/btrfs/inode.c
5058
pgoff_t index = (offset >> PAGE_SHIFT);
fs/btrfs/inode.c
802
const pgoff_t end_index = end >> PAGE_SHIFT;
fs/btrfs/inode.c
806
for (pgoff_t index = start >> PAGE_SHIFT; index <= end_index; index++) {
fs/btrfs/inode.c
9679
i = (iocb->ki_pos - start) >> PAGE_SHIFT;
fs/btrfs/ioctl.c
4560
index = (priv->iocb.ki_pos - priv->start) >> PAGE_SHIFT;
fs/btrfs/lzo.c
371
const u32 min_folio_shift = PAGE_SHIFT + fs_info->block_min_order;
fs/btrfs/messages.h
195
#define BTRFS_32BIT_MAX_FILE_SIZE (((u64)ULONG_MAX + 1) << PAGE_SHIFT)
fs/btrfs/raid56.c
1065
const unsigned int stripe_npages = BTRFS_STRIPE_LEN >> PAGE_SHIFT;
fs/btrfs/raid56.c
1297
max(BTRFS_STRIPE_LEN >> PAGE_SHIFT, 1),
fs/btrfs/raid56.c
1312
const u32 step_bits = min(fs_info->sectorsize_bits, PAGE_SHIFT);
fs/btrfs/raid56.c
2608
const unsigned int page_index = (i * step) >> PAGE_SHIFT;
fs/btrfs/raid56.c
3046
const unsigned int pindex = cur_off >> PAGE_SHIFT;
fs/btrfs/raid56.c
311
sector_nr = (page_nr << PAGE_SHIFT) >> rbio->bioc->fs_info->sectorsize_bits;
fs/btrfs/raid56.c
364
int page_index = offset >> PAGE_SHIFT;
fs/btrfs/reflink.c
85
folio = __filemap_get_folio(mapping, file_offset >> PAGE_SHIFT,
fs/btrfs/relocation.c
2784
const pgoff_t last_index = (cluster->end - offset) >> PAGE_SHIFT;
fs/btrfs/relocation.c
2785
const pgoff_t index = orig_file_offset >> PAGE_SHIFT;
fs/btrfs/relocation.c
4052
to_copy = min_t(u64, to_copy, BIO_MAX_VECS << PAGE_SHIFT);
fs/btrfs/scrub.c
1871
const u32 min_folio_shift = PAGE_SHIFT + fs_info->block_min_order;
fs/btrfs/scrub.c
357
const u32 min_folio_shift = PAGE_SHIFT + fs_info->block_min_order;
fs/btrfs/scrub.c
687
const u32 min_folio_shift = PAGE_SHIFT + fs_info->block_min_order;
fs/btrfs/scrub.c
700
const u32 min_folio_shift = PAGE_SHIFT + fs_info->block_min_order;
fs/btrfs/send.c
5204
const pgoff_t last_index = ((end - 1) >> PAGE_SHIFT);
fs/btrfs/send.c
5213
pgoff_t index = (cur >> PAGE_SHIFT);
fs/btrfs/send.c
5601
(data_offset >> PAGE_SHIFT),
fs/btrfs/send.c
8099
send_buf_num_pages = sctx->send_max_size >> PAGE_SHIFT;
fs/btrfs/send.c
8108
vmalloc_to_page(sctx->send_buf + (i << PAGE_SHIFT));
fs/btrfs/space-info.c
769
u64 temp = min(delalloc_bytes, to_reclaim) >> PAGE_SHIFT;
fs/btrfs/tests/extent-io-tests.c
159
for (pgoff_t index = 0; index < (total_dirty >> PAGE_SHIFT); index++) {
fs/btrfs/tests/extent-io-tests.c
205
test_start >> PAGE_SHIFT);
fs/btrfs/tests/extent-io-tests.c
240
PAGE_SHIFT);
fs/btrfs/tests/extent-io-tests.c
26
pgoff_t index = start >> PAGE_SHIFT;
fs/btrfs/tests/extent-io-tests.c
27
pgoff_t end_index = end >> PAGE_SHIFT;
fs/btrfs/tests/extent-io-tests.c
291
(max_bytes + SZ_1M) >> PAGE_SHIFT);
fs/btrfs/tests/extent-io-tests.c
673
struct page *page = folio_page(eb->folios[i >> PAGE_SHIFT], 0);
fs/btrfs/tests/extent-io-tests.c
688
for (int i = 0; i < (eb->len >> PAGE_SHIFT); i++) {
fs/btrfs/tests/extent-io-tests.c
691
if (memcmp(memory + (i << PAGE_SHIFT), eb_addr, PAGE_SIZE) != 0) {
fs/btrfs/verity.c
707
u64 off = (u64)index << PAGE_SHIFT;
fs/btrfs/verity.c
715
index += merkle_pos >> PAGE_SHIFT;
fs/btrfs/volumes.c
1368
invalidate_inode_pages2_range(mapping, bytenr >> PAGE_SHIFT,
fs/btrfs/volumes.c
1369
(bytenr + BTRFS_SUPER_INFO_SIZE) >> PAGE_SHIFT);
fs/btrfs/volumes.c
1373
page = read_cache_page_gfp(mapping, bytenr >> PAGE_SHIFT, GFP_NOFS);
fs/btrfs/zoned.c
135
bytenr >> PAGE_SHIFT, GFP_NOFS);
fs/btrfs/zoned.c
766
(u64)lim->max_segments << PAGE_SHIFT),
fs/buffer.c
2137
pgoff_t index = pos >> PAGE_SHIFT;
fs/buffer.c
2401
index = pos >> PAGE_SHIFT;
fs/buffer.c
2404
while (index > (curidx = (curpos = *bytes)>>PAGE_SHIFT)) {
fs/buffer.c
2547
pgoff_t index = from >> PAGE_SHIFT;
fs/ceph/addr.c
1096
ceph_wbc->index = wbc->range_start >> PAGE_SHIFT;
fs/ceph/addr.c
1097
ceph_wbc->end = wbc->range_end >> PAGE_SHIFT;
fs/ceph/addr.c
1215
ceph_wbc->strip_unit_end = folio->index + ((xlen - 1) >> PAGE_SHIFT);
fs/ceph/addr.c
1228
return folio->index == (ceph_wbc->offset + ceph_wbc->len) >> PAGE_SHIFT;
fs/ceph/addr.c
1666
ceph_wbc.max_pages = ceph_wbc.wsize >> PAGE_SHIFT;
fs/ceph/addr.c
192
max_len = max_pages << PAGE_SHIFT;
fs/ceph/addr.c
1954
loff_t off = (loff_t)vmf->pgoff << PAGE_SHIFT;
fs/ceph/addr.c
62
#define CONGESTION_ON_THRESH(congestion_kb) (congestion_kb >> (PAGE_SHIFT-10))
fs/ceph/crypto.c
482
int pgidx = blkoff >> PAGE_SHIFT;
fs/ceph/crypto.c
534
int pgidx = pgsoff >> PAGE_SHIFT;
fs/ceph/crypto.c
590
int pgidx = blkoff >> PAGE_SHIFT;
fs/ceph/crypto.h
125
BUILD_BUG_ON(CEPH_FSCRYPT_BLOCK_SHIFT > PAGE_SHIFT);
fs/ceph/dir.c
139
pgoff_t ptr_pgoff = ptr_pos >> PAGE_SHIFT;
fs/ceph/file.c
1511
pos >> PAGE_SHIFT,
fs/ceph/file.c
1512
(pos + count - 1) >> PAGE_SHIFT);
fs/ceph/file.c
2068
pos >> PAGE_SHIFT,
fs/ceph/file.c
2069
(pos + len - 1) >> PAGE_SHIFT);
fs/ceph/file.c
2533
folio = filemap_lock_folio(inode->i_mapping, offset >> PAGE_SHIFT);
fs/ceph/file.c
3062
dst_off >> PAGE_SHIFT,
fs/ceph/file.c
3063
(dst_off + len) >> PAGE_SHIFT);
fs/ceph/inode.c
1216
inode->i_blkbits = PAGE_SHIFT;
fs/ceph/super.c
1283
sb->s_bdi->ra_pages = fsc->mount_options->rasize >> PAGE_SHIFT;
fs/ceph/super.c
1286
sb->s_bdi->io_pages = fsc->mount_options->rsize >> PAGE_SHIFT;
fs/ceph/super.c
965
(CEPH_MAX_WRITE_SIZE >> PAGE_SHIFT) * sizeof(struct page *));
fs/ceph/super.h
1002
congestion_kb = (16*int_sqrt(totalram_pages())) << (PAGE_SHIFT-10);
fs/configfs/mount.c
69
sb->s_blocksize_bits = PAGE_SHIFT;
fs/cramfs/inode.c
204
blocknr = offset >> PAGE_SHIFT;
fs/cramfs/inode.c
215
blk_offset = (blocknr - buffer_blocknr[i]) << PAGE_SHIFT;
fs/cramfs/inode.c
223
devsize = bdev_nr_bytes(sb->s_bdev) >> PAGE_SHIFT;
fs/cramfs/inode.c
347
last_page = inode->i_size >> PAGE_SHIFT;
fs/cramfs/inode.c
378
max_pages = (inode->i_size + PAGE_SIZE - 1) >> PAGE_SHIFT;
fs/cramfs/inode.c
411
ret = remap_pfn_range(vma, vma->vm_start, address >> PAGE_SHIFT,
fs/cramfs/inode.c
461
pages = (len + PAGE_SIZE - 1) >> PAGE_SHIFT;
fs/cramfs/inode.c
462
max_pages = (inode->i_size + PAGE_SIZE - 1) >> PAGE_SHIFT;
fs/cramfs/inode.c
829
maxblock = (inode->i_size + PAGE_SIZE - 1) >> PAGE_SHIFT;
fs/crypto/bio.c
137
const unsigned int du_per_page_bits = PAGE_SHIFT - du_bits;
fs/crypto/crypto.c
180
u64 index = ((u64)folio->index << (PAGE_SHIFT - du_bits)) +
fs/crypto/crypto.c
262
u64 index = ((u64)folio->index << (PAGE_SHIFT - du_bits)) +
fs/crypto/crypto.c
274
struct page *page = folio_page(folio, i >> PAGE_SHIFT);
fs/dax.c
1201
XA_STATE(xas, &mapping->i_pages, wbc->range_start >> PAGE_SHIFT);
fs/dax.c
1203
pgoff_t end_index = wbc->range_end >> PAGE_SHIFT;
fs/dax.c
1208
if (WARN_ON_ONCE(inode->i_blkbits != PAGE_SHIFT))
fs/dax.c
1434
copy_pos >> PAGE_SHIFT,
fs/dax.c
1435
(copy_pos + copy_len - 1) >> PAGE_SHIFT);
fs/dax.c
1517
iter->pos >> PAGE_SHIFT,
fs/dax.c
1518
(iter->pos + length - 1) >> PAGE_SHIFT);
fs/dax.c
1625
pos >> PAGE_SHIFT,
fs/dax.c
1626
(end - 1) >> PAGE_SHIFT);
fs/dax.c
1628
pos >> PAGE_SHIFT,
fs/dax.c
1629
(end - 1) >> PAGE_SHIFT);
fs/dax.c
1811
loff_t pos = (loff_t)xas->xa_index << PAGE_SHIFT;
fs/dax.c
1868
.pos = (loff_t)vmf->pgoff << PAGE_SHIFT,
fs/dax.c
1951
((vmf->address >> PAGE_SHIFT) & PG_PMD_COLOUR))
fs/dax.c
2028
iter.pos = (loff_t)xas.xa_index << PAGE_SHIFT;
fs/dax.c
2145
loff_t start = ((loff_t)vmf->pgoff) << PAGE_SHIFT;
fs/dax.c
36
#define PG_PMD_COLOUR ((PMD_SIZE >> PAGE_SHIFT) - 1)
fs/dax.c
37
#define PG_PMD_NR (PMD_SIZE >> PAGE_SHIFT)
fs/dax.c
759
pgoff_t start_idx = start >> PAGE_SHIFT;
fs/dax.c
770
end_idx = end >> PAGE_SHIFT;
fs/dax.c
891
pgoff_t start_idx = start >> PAGE_SHIFT;
fs/dax.c
899
end_idx = end >> PAGE_SHIFT;
fs/ecryptfs/mmap.c
137
folio, (lower_offset >> PAGE_SHIFT),
fs/ecryptfs/mmap.c
247
pgoff_t index = pos >> PAGE_SHIFT;
fs/ecryptfs/mmap.c
258
prev_page_end_size = ((loff_t)index << PAGE_SHIFT);
fs/ecryptfs/mmap.c
445
pgoff_t index = pos >> PAGE_SHIFT;
fs/ecryptfs/read_write.c
113
pgoff_t ecryptfs_page_idx = (pos >> PAGE_SHIFT);
fs/efivarfs/super.c
354
sb->s_blocksize_bits = PAGE_SHIFT;
fs/erofs/data.c
31
pgoff_t index = (buf->off + offset) >> PAGE_SHIFT;
fs/erofs/data.c
42
fpos = index << PAGE_SHIFT;
fs/erofs/decompressor.c
156
return out + ((rq->outpages - rq->inpages) << PAGE_SHIFT);
fs/erofs/decompressor.c
324
no = (rq->pageofs_out + cur + pi) >> PAGE_SHIFT;
fs/erofs/internal.h
479
DO_ONCE_LITE_IF(realinode->i_blkbits != PAGE_SHIFT,
fs/erofs/super.c
278
if (sbi->blkszbits < 9 || sbi->blkszbits > PAGE_SHIFT) {
fs/erofs/super.c
655
sbi->blkszbits = PAGE_SHIFT;
fs/erofs/super.c
681
sb->s_blocksize_bits = PAGE_SHIFT;
fs/erofs/super.c
728
if (test_opt(&sbi->opt, DAX_ALWAYS) && sbi->blkszbits != PAGE_SHIFT) {
fs/erofs/zdata.c
1072
.page = folio_page(folio, pgs >> PAGE_SHIFT),
fs/erofs/zdata.c
108
return PAGE_ALIGN(pcl->pageofs_in + pcl->pclustersize) >> PAGE_SHIFT;
fs/erofs/zdata.c
1134
DBG_BUGON((poff >> PAGE_SHIFT) >= be->nr_pages);
fs/erofs/zdata.c
1135
page = be->decompressed_pages + (poff >> PAGE_SHIFT);
fs/erofs/zdata.c
1168
pgnr = (bvi->bvec.offset + cur + off0) >> PAGE_SHIFT;
fs/erofs/zdata.c
1172
((pgnr << PAGE_SHIFT) - off0);
fs/erofs/zdata.c
1265
be->nr_pages = PAGE_ALIGN(pcl->length + pcl->pageofs_out) >> PAGE_SHIFT;
fs/erofs/zdata.c
1578
filemap_add_folio(mc, folio, (pcl->pos >> PAGE_SHIFT) + nr, gfp)) {
fs/erofs/zdata.c
1861
pgoff_t index = cur >> PAGE_SHIFT;
fs/erofs/zdata.c
1875
cur = (index << PAGE_SHIFT) - 1;
fs/erofs/zdata.c
1937
(void)z_erofs_runqueue(&f, nrpages << PAGE_SHIFT);
fs/erofs/zdata.c
254
unsigned int nrpages = PAGE_ALIGN(size) >> PAGE_SHIFT;
fs/erofs/zdata.c
539
pgoff_t poff = pcl->pos >> PAGE_SHIFT;
fs/eventpoll.c
2599
max_user_watches = (((si.totalram - si.totalhigh) / 25) << PAGE_SHIFT) /
fs/exec.c
623
stack_base += (STACK_RND_MASK << PAGE_SHIFT);
fs/exec.c
736
stop = bprm->p >> PAGE_SHIFT;
fs/exfat/exfat_fs.h
85
<< (PAGE_SHIFT - (sb)->s_blocksize_bits))
fs/exfat/file.c
760
new_valid_size = ((loff_t)vmf->pgoff + 1) << PAGE_SHIFT;
fs/ext2/dir.c
264
unsigned long n = pos >> PAGE_SHIFT;
fs/ext2/dir.c
292
ctx->pos = (n<<PAGE_SHIFT) + offset;
fs/ext2/dir.c
384
if (unlikely(n > (dir->i_blocks >> (PAGE_SHIFT - 9)))) {
fs/ext2/dir.c
79
last_byte -= page_nr << PAGE_SHIFT;
fs/ext2/file.c
272
pos >> PAGE_SHIFT,
fs/ext2/file.c
273
endbyte >> PAGE_SHIFT);
fs/ext4/dir.c
197
PAGE_SHIFT;
fs/ext4/dir.c
203
file->f_ra.prev_pos = (loff_t)index << PAGE_SHIFT;
fs/ext4/ext4.h
375
PAGE_SHIFT)
fs/ext4/ext4.h
377
#define EXT4_PG_TO_LBLK(inode, pnum) (((loff_t)(pnum) << PAGE_SHIFT) >> \
fs/ext4/file.c
625
offset >> PAGE_SHIFT,
fs/ext4/file.c
626
endbyte >> PAGE_SHIFT);
fs/ext4/inode.c
1312
index = pos >> PAGE_SHIFT;
fs/ext4/inode.c
1746
index = mpd->start_pos >> PAGE_SHIFT;
fs/ext4/inode.c
1747
end = mpd->next_pos >> PAGE_SHIFT;
fs/ext4/inode.c
2413
mpd->start_pos >> PAGE_SHIFT);
fs/ext4/inode.c
2422
!folio_contains(folio, pos >> PAGE_SHIFT)))
fs/ext4/inode.c
2623
pgoff_t index = mpd->start_pos >> PAGE_SHIFT;
fs/ext4/inode.c
2624
pgoff_t end = mpd->end_pos >> PAGE_SHIFT;
fs/ext4/inode.c
2856
mpd->start_pos = writeback_index << PAGE_SHIFT;
fs/ext4/inode.c
2866
tag_pages_for_writeback(mapping, mpd->start_pos >> PAGE_SHIFT,
fs/ext4/inode.c
2867
mpd->end_pos >> PAGE_SHIFT);
fs/ext4/inode.c
2995
mpd->end_pos = (writeback_index << PAGE_SHIFT) - 1;
fs/ext4/inode.c
3006
mapping->writeback_index = mpd->start_pos >> PAGE_SHIFT;
fs/ext4/inode.c
3131
index = pos >> PAGE_SHIFT;
fs/ext4/inode.c
4027
folio = __filemap_get_folio(mapping, from >> PAGE_SHIFT,
fs/ext4/inode.c
4260
folio = filemap_lock_folio(inode->i_mapping, start >> PAGE_SHIFT);
fs/ext4/inode.c
5751
inode->i_size >> PAGE_SHIFT);
fs/ext4/move_extent.c
211
orig_pos >> PAGE_SHIFT, donor_pos >> PAGE_SHIFT,
fs/ext4/super.c
5087
umin(MAX_PAGECACHE_ORDER, (11 + (sb)->s_blocksize_bits - PAGE_SHIFT))
fs/ext4/verity.c
365
index += ext4_verity_metadata_pos(inode) >> PAGE_SHIFT;
fs/ext4/verity.c
372
index += ext4_verity_metadata_pos(inode) >> PAGE_SHIFT;
fs/ext4/verity.c
48
folio = read_mapping_folio(inode->i_mapping, pos >> PAGE_SHIFT,
fs/f2fs/compress.c
1218
pgoff_t start_idx = from >> (PAGE_SHIFT + log_cluster_size) <<
fs/f2fs/compress.c
1245
loff_t start = (loff_t)folio->index << PAGE_SHIFT;
fs/f2fs/compress.c
1257
round_down(from, 1 << log_cluster_size << PAGE_SHIFT),
fs/f2fs/data.c
1460
((loff_t)(index + 1) << PAGE_SHIFT))
fs/f2fs/data.c
1461
f2fs_i_size_write(inode, ((loff_t)(index + 1) << PAGE_SHIFT));
fs/f2fs/data.c
152
unsigned nr_pages = fi.length >> PAGE_SHIFT;
fs/f2fs/data.c
2535
size_t page_offset = offset << PAGE_SHIFT;
fs/f2fs/data.c
2580
offset << PAGE_SHIFT))
fs/f2fs/data.c
3068
>> PAGE_SHIFT;
fs/f2fs/data.c
3069
loff_t psize = (loff_t)(folio->index + 1) << PAGE_SHIFT;
fs/f2fs/data.c
3285
index = wbc->range_start >> PAGE_SHIFT;
fs/f2fs/data.c
3286
end = wbc->range_end >> PAGE_SHIFT;
fs/f2fs/data.c
3859
pgoff_t index = pos >> PAGE_SHIFT;
fs/f2fs/debug.c
393
si->page_mem += (unsigned long long)npages << PAGE_SHIFT;
fs/f2fs/debug.c
398
si->page_mem += (unsigned long long)npages << PAGE_SHIFT;
fs/f2fs/debug.c
404
si->page_mem += (unsigned long long)npages << PAGE_SHIFT;
fs/f2fs/dir.c
42
>> PAGE_SHIFT;
fs/f2fs/extent_cache.c
904
if ((f_size >> PAGE_SHIFT) == ei->fofs && f_size & (PAGE_SIZE - 1) &&
fs/f2fs/file.c
126
f2fs_zero_post_eof_page(inode, (folio->index + 1) << PAGE_SHIFT, true);
fs/f2fs/file.c
1284
pg_start = ((unsigned long long) offset) >> PAGE_SHIFT;
fs/f2fs/file.c
1285
pg_end = ((unsigned long long) offset + len) >> PAGE_SHIFT;
fs/f2fs/file.c
1315
blk_start = (loff_t)pg_start << PAGE_SHIFT;
fs/f2fs/file.c
1316
blk_end = (loff_t)pg_end << PAGE_SHIFT;
fs/f2fs/file.c
1464
new_size = (loff_t)(dst + i) << PAGE_SHIFT;
fs/f2fs/file.c
1559
pgoff_t start = offset >> PAGE_SHIFT;
fs/f2fs/file.c
1560
pgoff_t end = (offset + len) >> PAGE_SHIFT;
fs/f2fs/file.c
169
if (((loff_t)(folio->index + 1) << PAGE_SHIFT) >
fs/f2fs/file.c
1697
pg_start = ((unsigned long long) offset) >> PAGE_SHIFT;
fs/f2fs/file.c
1698
pg_end = ((unsigned long long) offset + len) >> PAGE_SHIFT;
fs/f2fs/file.c
1718
(loff_t)pg_start << PAGE_SHIFT);
fs/f2fs/file.c
1731
(loff_t)index << PAGE_SHIFT,
fs/f2fs/file.c
1732
((loff_t)pg_end << PAGE_SHIFT) - 1);
fs/f2fs/file.c
1762
(loff_t)index << PAGE_SHIFT);
fs/f2fs/file.c
1821
pg_start = offset >> PAGE_SHIFT;
fs/f2fs/file.c
1822
pg_end = (offset + len) >> PAGE_SHIFT;
fs/f2fs/file.c
1894
pg_start = ((unsigned long long)offset) >> PAGE_SHIFT;
fs/f2fs/file.c
1895
pg_end = ((unsigned long long)offset + len) >> PAGE_SHIFT;
fs/f2fs/file.c
1971
(loff_t)(last_off + 1) << PAGE_SHIFT;
fs/f2fs/file.c
1973
new_size = ((loff_t)pg_end << PAGE_SHIFT) + off_end;
fs/f2fs/file.c
2558
start = offset >> PAGE_SHIFT;
fs/f2fs/file.c
2904
pg_start = range->start >> PAGE_SHIFT;
fs/f2fs/file.c
2906
(range->start + range->len) >> PAGE_SHIFT,
fs/f2fs/file.c
2924
pg_start << PAGE_SHIFT,
fs/f2fs/file.c
2925
(pg_end << PAGE_SHIFT) - 1);
fs/f2fs/file.c
3046
range->len = (u64)total << PAGE_SHIFT;
fs/f2fs/file.c
3074
if (unlikely((range.start + range.len) >> PAGE_SHIFT >
fs/f2fs/file.c
482
pgofs = (pgoff_t)(offset >> PAGE_SHIFT);
fs/f2fs/file.c
484
for (; data_ofs < isize; data_ofs = (loff_t)pgofs << PAGE_SHIFT) {
fs/f2fs/file.c
504
data_ofs = (loff_t)pgofs << PAGE_SHIFT) {
fs/f2fs/file.c
5085
start_pos >> PAGE_SHIFT,
fs/f2fs/file.c
5086
end_pos >> PAGE_SHIFT);
fs/f2fs/file.c
742
pgoff_t index = from >> PAGE_SHIFT;
fs/f2fs/file.c
870
F2FS_I(inode)->i_cluster_size << PAGE_SHIFT);
fs/f2fs/node.c
101
sizeof(struct extent_node)) >> PAGE_SHIFT;
fs/f2fs/node.c
105
sizeof(struct discard_cmd)) >> PAGE_SHIFT;
fs/f2fs/node.c
72
sizeof(struct free_nid)) >> PAGE_SHIFT;
fs/f2fs/node.c
76
sizeof(struct nat_entry)) >> PAGE_SHIFT;
fs/f2fs/node.c
91
mem_size >>= PAGE_SHIFT;
fs/f2fs/recovery.c
714
(i_size_read(inode) <= ((loff_t)index << PAGE_SHIFT)))
fs/f2fs/recovery.c
716
(loff_t)(index + 1) << PAGE_SHIFT);
fs/f2fs/recovery.c
918
(loff_t)MAIN_BLKADDR(sbi) << PAGE_SHIFT, -1);
fs/f2fs/shrinker.c
168
pgoff_t npages = reclaim_caches_kb >> (PAGE_SHIFT - 10);
fs/f2fs/shrinker.c
201
return npages << (PAGE_SHIFT - 10);
fs/f2fs/super.c
3136
folio = mapping_read_folio_gfp(mapping, off >> PAGE_SHIFT,
fs/f2fs/sysfs.c
1944
(loff_t)fi->donate_start << (PAGE_SHIFT - 10),
fs/f2fs/sysfs.c
1945
(loff_t)(fi->donate_end + 1) << (PAGE_SHIFT - 10),
fs/f2fs/sysfs.c
1946
(loff_t)inode->i_mapping->nrpages << (PAGE_SHIFT - 10));
fs/f2fs/verity.c
261
index += f2fs_verity_metadata_pos(inode) >> PAGE_SHIFT;
fs/f2fs/verity.c
268
index += f2fs_verity_metadata_pos(inode) >> PAGE_SHIFT;
fs/f2fs/verity.c
51
page = read_mapping_page(inode->i_mapping, pos >> PAGE_SHIFT,
fs/fat/fatent.c
672
reada_blocks = ra_pages << (PAGE_SHIFT - sb->s_blocksize_bits + 1);
fs/freevxfs/vxfs_lookup.c
223
pp = vxfs_get_page(ip->i_mapping, pos >> PAGE_SHIFT);
fs/freevxfs/vxfs_lookup.c
74
pp = vxfs_get_page(ip->i_mapping, pos >> PAGE_SHIFT);
fs/fs-writeback.c
905
max_time = DIV_ROUND_UP((max_bytes >> PAGE_SHIFT) << WB_FRN_TIME_SHIFT,
fs/fuse/dax.c
846
start_pos >> PAGE_SHIFT,
fs/fuse/dax.c
847
end_pos >> PAGE_SHIFT);
fs/fuse/dev.c
1793
index = outarg.offset >> PAGE_SHIFT;
fs/fuse/dev.c
1814
folio_offset = ((index - folio->index) << PAGE_SHIFT) + offset;
fs/fuse/dev.c
1816
nr_pages = (offset + nr_bytes + PAGE_SIZE - 1) >> PAGE_SHIFT;
fs/fuse/dev.c
1885
num_pages = (num + offset + PAGE_SIZE - 1) >> PAGE_SHIFT;
fs/fuse/dev.c
1887
num = min(num, num_pages << PAGE_SHIFT);
fs/fuse/dev.c
1906
index = outarg->offset >> PAGE_SHIFT;
fs/fuse/dev.c
1918
folio_offset = ((index - folio->index) << PAGE_SHIFT) + offset;
fs/fuse/dev.c
1920
nr_pages = (offset + nr_bytes + PAGE_SIZE - 1) >> PAGE_SHIFT;
fs/fuse/file.c
1258
pgoff_t index = pos >> PAGE_SHIFT;
fs/fuse/file.c
1273
folio_offset = ((index - folio->index) << PAGE_SHIFT) + offset;
fs/fuse/file.c
1327
unsigned int pages = ((pos + len - 1) >> PAGE_SHIFT) -
fs/fuse/file.c
1328
(pos >> PAGE_SHIFT) + 1;
fs/fuse/file.c
1610
(folio_page_idx(folio, pages[i]) << PAGE_SHIFT);
fs/fuse/file.c
1655
pgoff_t idx_from = pos >> PAGE_SHIFT;
fs/fuse/file.c
1656
pgoff_t idx_to = (pos + count - 1) >> PAGE_SHIFT;
fs/fuse/file.c
2176
if ((bytes + PAGE_SIZE - 1) >> PAGE_SHIFT > fc->max_pages)
fs/fuse/file.c
2814
return round_up(off, fc->max_pages << PAGE_SHIFT);
fs/fuse/inode.c
1239
*limit = ((totalram_pages() << PAGE_SHIFT) >> 13) / 392;
fs/fuse/inode.c
1849
sb->s_blocksize_bits = PAGE_SHIFT;
fs/fuse/inode.c
576
pg_start = offset >> PAGE_SHIFT;
fs/fuse/inode.c
580
pg_end = (offset + len - 1) >> PAGE_SHIFT;
fs/fuse/ioctl.c
78
u32 max = fc->max_pages << PAGE_SHIFT;
fs/fuse/readdir.c
344
size_t bufsize = clamp((unsigned int) ctx->count, PAGE_SIZE, fc->max_pages << PAGE_SHIFT);
fs/fuse/readdir.c
518
index = ff->readdir.cache_off >> PAGE_SHIFT;
fs/fuse/readdir.c
520
if (index == (fi->rdc.size >> PAGE_SHIFT))
fs/fuse/readdir.c
56
index = size >> PAGE_SHIFT;
fs/fuse/readdir.c
86
fi->rdc.size = (index << PAGE_SHIFT) + offset + reclen;
fs/fuse/virtio_fs.c
1032
memset(kaddr, 0, nr_pages << PAGE_SHIFT);
fs/fuse/virtio_fs.c
1033
dax_flush(dax_dev, kaddr, nr_pages << PAGE_SHIFT);
fs/gfs2/aops.c
307
index = wbc->range_start >> PAGE_SHIFT;
fs/gfs2/aops.c
308
end = wbc->range_end >> PAGE_SHIFT;
fs/gfs2/aops.c
453
unsigned long index = *pos >> PAGE_SHIFT;
fs/gfs2/file.c
1157
(iocb->ki_pos - buffered) >> PAGE_SHIFT,
fs/gfs2/file.c
1158
(iocb->ki_pos - 1) >> PAGE_SHIFT);
fs/gfs2/file.c
803
size = min_t(size_t, SZ_1M, nr_dirtied << PAGE_SHIFT);
fs/gfs2/lops.c
511
unsigned int shift = PAGE_SHIFT - bsize_shift;
fs/gfs2/meta_io.c
139
shift = PAGE_SHIFT - sdp->sd_sb.sb_bsize_shift;
fs/gfs2/meta_io.c
416
unsigned int shift = PAGE_SHIFT - sdp->sd_sb.sb_bsize_shift;
fs/gfs2/quota.c
738
blk = index << (PAGE_SHIFT - sdp->sd_sb.sb_bsize_shift);
fs/gfs2/quota.c
805
pg_beg = loc >> PAGE_SHIFT;
fs/hfs/bnode.c
385
block = off >> PAGE_SHIFT;
fs/hfs/bnode.c
82
pagenum = off >> PAGE_SHIFT;
fs/hfs/btree.c
153
tree->pages_per_bnode = (tree->node_size + PAGE_SIZE - 1) >> PAGE_SHIFT;
fs/hfs/btree.c
308
pagep = node->page + (off >> PAGE_SHIFT);
fs/hfs/btree.c
353
pagep = node->page + (off >> PAGE_SHIFT);
fs/hfs/btree.c
400
page = node->page[off >> PAGE_SHIFT];
fs/hfs/inode.c
106
nidx = folio->index << (PAGE_SHIFT - tree->node_size_shift);
fs/hfs/inode.c
107
i = 1 << (PAGE_SHIFT - tree->node_size_shift);
fs/hfs/inode.c
93
nidx = folio->index >> (tree->node_size_shift - PAGE_SHIFT);
fs/hfsplus/bnode.c
114
pagep = node->page + (off >> PAGE_SHIFT);
fs/hfsplus/bnode.c
156
pagep = node->page + (off >> PAGE_SHIFT);
fs/hfsplus/bnode.c
185
src_page = src_node->page + (src >> PAGE_SHIFT);
fs/hfsplus/bnode.c
187
dst_page = dst_node->page + (dst >> PAGE_SHIFT);
fs/hfsplus/bnode.c
245
src_page = node->page + (src >> PAGE_SHIFT);
fs/hfsplus/bnode.c
248
dst_page = node->page + (dst >> PAGE_SHIFT);
fs/hfsplus/bnode.c
296
src_page = node->page + (src >> PAGE_SHIFT);
fs/hfsplus/bnode.c
298
dst_page = node->page + (dst >> PAGE_SHIFT);
fs/hfsplus/bnode.c
43
pagep = node->page + (off >> PAGE_SHIFT);
fs/hfsplus/bnode.c
498
block = off >> PAGE_SHIFT;
fs/hfsplus/btree.c
183
ctx->page_idx = page_off >> PAGE_SHIFT;
fs/hfsplus/btree.c
377
PAGE_SHIFT;
fs/hfsplus/inode.c
104
(PAGE_SHIFT - tree->node_size_shift);
fs/hfsplus/inode.c
105
i = 1 << (PAGE_SHIFT - tree->node_size_shift);
fs/hfsplus/inode.c
90
(tree->node_size_shift - PAGE_SHIFT);
fs/hfsplus/super.c
519
PAGE_SHIFT;
fs/hostfs/hostfs_kern.c
453
pgoff_t index = pos >> PAGE_SHIFT;
fs/hugetlbfs/inode.c
137
if (desc->pgoff & (~huge_page_mask(h) >> PAGE_SHIFT))
fs/hugetlbfs/inode.c
141
len = vma_len + ((loff_t)desc->pgoff << PAGE_SHIFT);
fs/hugetlbfs/inode.c
377
offset = (start - vma->vm_pgoff) << PAGE_SHIFT;
fs/hugetlbfs/inode.c
389
t_end = ((end - vma->vm_pgoff) << PAGE_SHIFT) + vma->vm_start;
fs/hugetlbfs/inode.c
576
const pgoff_t end = lend >> PAGE_SHIFT;
fs/hugetlbfs/inode.c
583
next = lstart >> PAGE_SHIFT;
fs/hugetlbfs/inode.c
633
pgoff = offset >> PAGE_SHIFT;
fs/hugetlbfs/inode.c
700
hole_start >> PAGE_SHIFT,
fs/hugetlbfs/inode.c
701
hole_end >> PAGE_SHIFT, 0);
fs/hugetlbfs/inode.c
97
(((1UL << (PAGE_SHIFT + 1)) - 1) << (BITS_PER_LONG - (PAGE_SHIFT + 1)))
fs/iomap/buffered-io.c
1315
start_byte >> PAGE_SHIFT);
fs/iomap/buffered-io.c
1604
pgoff_t pstart = *start >> PAGE_SHIFT;
fs/iomap/buffered-io.c
1605
pgoff_t pend = (end - 1) >> PAGE_SHIFT;
fs/iomap/buffered-io.c
1614
*start = (pstart << PAGE_SHIFT);
fs/iomap/buffered-io.c
1809
pgoff_t end_index = isize >> PAGE_SHIFT;
fs/iomap/buffered-io.c
725
return __filemap_get_folio(iter->inode->i_mapping, pos >> PAGE_SHIFT,
fs/iomap/swapfile.c
45
first_ppage = ALIGN(iomap->addr, PAGE_SIZE) >> PAGE_SHIFT;
fs/iomap/swapfile.c
47
PAGE_SHIFT;
fs/isofs/compress.c
236
WARN_ON(start_off - (full_page << PAGE_SHIFT) !=
fs/isofs/compress.c
270
pages += poffset >> PAGE_SHIFT;
fs/isofs/compress.c
271
pcount -= poffset >> PAGE_SHIFT;
fs/isofs/compress.c
272
full_page -= poffset >> PAGE_SHIFT;
fs/isofs/compress.c
310
PAGE_SHIFT <= zisofs_block_shift ?
fs/isofs/compress.c
311
(1 << (zisofs_block_shift - PAGE_SHIFT)) : 0;
fs/isofs/compress.c
315
end_index = (inode->i_size + PAGE_SIZE - 1) >> PAGE_SHIFT;
fs/isofs/compress.c
325
if (PAGE_SHIFT <= zisofs_block_shift) {
fs/isofs/compress.c
73
return ((loff_t)pcount) << PAGE_SHIFT;
fs/jbd2/journal.c
2677
return 1 << (PAGE_SHIFT + mapping_max_folio_order(inode->i_mapping) -
fs/jffs2/file.c
137
pgoff_t index = pos >> PAGE_SHIFT;
fs/jffs2/file.c
92
__func__, inode->i_ino, folio->index << PAGE_SHIFT);
fs/jffs2/file.c
97
ret = jffs2_read_inode_range(c, f, kaddr, folio->index << PAGE_SHIFT,
fs/jffs2/fs.c
216
buf->f_bsize = 1 << PAGE_SHIFT;
fs/jffs2/fs.c
217
buf->f_blocks = c->flash_size >> PAGE_SHIFT;
fs/jffs2/fs.c
232
buf->f_bavail = buf->f_bfree = avail >> PAGE_SHIFT;
fs/jffs2/fs.c
594
sb->s_blocksize_bits = PAGE_SHIFT;
fs/jffs2/gc.c
1329
folio = read_cache_folio(inode->i_mapping, start >> PAGE_SHIFT,
fs/jffs2/gc.c
555
if((start >> PAGE_SHIFT) < ((end-1) >> PAGE_SHIFT)) {
fs/jffs2/nodelist.c
240
if (lastend && (lastend-1) >> PAGE_SHIFT == newfrag->ofs >> PAGE_SHIFT) {
fs/jfs/jfs_metapage.c
699
l2BlocksPerPage = PAGE_SHIFT - l2bsize;
fs/jfs/jfs_metapage.c
901
int l2BlocksPerPage = PAGE_SHIFT - ip->i_blkbits;
fs/kernfs/mount.c
294
sb->s_blocksize_bits = PAGE_SHIFT;
fs/libfs.c
1044
s->s_blocksize_bits = PAGE_SHIFT;
fs/libfs.c
1616
last_fs_page = (max_bytes >> PAGE_SHIFT) - 1;
fs/libfs.c
1772
inode->i_blkbits = PAGE_SHIFT;
fs/libfs.c
2110
invalidate_mapping_pages(mapping, pos >> PAGE_SHIFT, end >> PAGE_SHIFT);
fs/libfs.c
38
stat->blocks = inode->i_mapping->nrpages << (PAGE_SHIFT - 9);
fs/libfs.c
683
s->s_blocksize_bits = PAGE_SHIFT;
fs/minix/dir.c
38
if (page_nr == (inode->i_size >> PAGE_SHIFT))
fs/minix/dir.c
99
n = pos >> PAGE_SHIFT;
fs/netfs/buffered_read.c
630
pgoff_t index = pos >> PAGE_SHIFT;
fs/netfs/direct_write.c
35
pgoff_t first = wreq->start >> PAGE_SHIFT;
fs/netfs/direct_write.c
36
pgoff_t last = (wreq->start + wreq->transferred - 1) >> PAGE_SHIFT;
fs/netfs/misc.c
51
order = umin(ilog2(size - *_cur_size) - PAGE_SHIFT,
fs/nfs/blocklayout/blocklayout.c
241
int pg_index = header->args.pgbase >> PAGE_SHIFT;
fs/nfs/blocklayout/blocklayout.c
385
int pg_index = header->args.pgbase >> PAGE_SHIFT;
fs/nfs/blocklayout/blocklayout.c
901
return i_size_read(inode) - (idx << PAGE_SHIFT);
fs/nfs/blocklayout/blocklayout.c
903
return (end - idx) << PAGE_SHIFT;
fs/nfs/blocklayout/blocklayout.h
44
#define PAGE_CACHE_SECTOR_SHIFT (PAGE_SHIFT - SECTOR_SHIFT)
fs/nfs/client.c
866
server->rpages = (server->rsize + PAGE_SIZE - 1) >> PAGE_SHIFT;
fs/nfs/dir.c
954
array_size = (dtsize + PAGE_SIZE - 1) >> PAGE_SHIFT;
fs/nfs/direct.c
1014
end = (pos + iov_iter_count(iter) - 1) >> PAGE_SHIFT;
fs/nfs/direct.c
1056
pos >> PAGE_SHIFT, end);
fs/nfs/file.c
294
folio = filemap_lock_folio(mapping, from >> PAGE_SHIFT);
fs/nfs/file.c
399
folio = write_begin_get_folio(iocb, mapping, pos >> PAGE_SHIFT, len);
fs/nfs/fscache.c
310
start = (sreq->start + sreq->transferred) >> PAGE_SHIFT;
fs/nfs/fscache.c
311
last = ((sreq->start + sreq->len - sreq->transferred - 1) >> PAGE_SHIFT);
fs/nfs/internal.h
901
PAGE_SHIFT;
fs/nfs/localio.c
481
pagevec += base >> PAGE_SHIFT;
fs/nfs/nfs3acl.c
201
unsigned int npages = 1 + ((args.len - 1) >> PAGE_SHIFT);
fs/nfs/nfs3xdr.c
1332
NFSACL_MAXPAGES << PAGE_SHIFT,
fs/nfs/nfs42proc.c
393
WARN_ON_ONCE(invalidate_inode_pages2_range(mapping, pos >> PAGE_SHIFT,
fs/nfs/nfs42proc.c
394
end >> PAGE_SHIFT));
fs/nfs/nfs4xdr.c
5223
maxsize >>= PAGE_SHIFT;
fs/nfs/pagelist.c
446
ret = nfs_page_create(l_ctx, pgbase, offset >> PAGE_SHIFT,
fs/nfs/pagelist.c
627
if (((mirror->pg_count + req->wb_bytes) >> PAGE_SHIFT) *
fs/nfs/pnfs.c
1183
size_t npages = (max_reply_sz + PAGE_SIZE - 1) >> PAGE_SHIFT;
fs/nfs/read.c
92
npages = (pgm->pg_bytes_written + PAGE_SIZE - 1) >> PAGE_SHIFT;
fs/nfs/write.c
2172
nfs_congestion_kb = (16*int_sqrt(totalram_pages())) << (PAGE_SHIFT-10);
fs/nfs/write.c
282
#define NFS_CONGESTION_ON_THRESH (nfs_congestion_kb >> (PAGE_SHIFT-10))
fs/nfsd/nfs3proc.c
558
rqstp->rq_next_page += (buf->buflen + PAGE_SIZE - 1) >> PAGE_SHIFT;
fs/nfsd/nfs4state.c
8954
max_delegations = nr_free_buffer_pages() >> (20 - 2 - PAGE_SHIFT);
fs/nfsd/nfscache.c
73
limit = (16 * int_sqrt(low_pages)) << (PAGE_SHIFT-10);
fs/nfsd/nfssvc.c
557
target = (i.totalram - i.totalhigh) << PAGE_SHIFT;
fs/nilfs2/btnode.c
223
if (inode->i_blkbits == PAGE_SHIFT) {
fs/nilfs2/dir.c
173
dir->i_ino, error, (folio->index << PAGE_SHIFT) + offs,
fs/nilfs2/dir.c
181
dir->i_ino, (folio->index << PAGE_SHIFT) + offs,
fs/nilfs2/dir.c
241
unsigned long n = pos >> PAGE_SHIFT;
fs/nilfs2/dir.c
337
if (unlikely(n > (dir->i_blocks >> (PAGE_SHIFT - 9)))) {
fs/nilfs2/dir.c
76
last_byte -= page_nr << PAGE_SHIFT;
fs/nilfs2/file.c
107
nilfs_set_file_dirty(inode, 1 << (PAGE_SHIFT - inode->i_blkbits));
fs/nilfs2/mdt.c
356
pgoff_t index = block >> (PAGE_SHIFT - inode->i_blkbits);
fs/nilfs2/mdt.c
371
(PAGE_SHIFT - inode->i_blkbits);
fs/nilfs2/page.c
39
first_block = (unsigned long)index << (PAGE_SHIFT - blkbits);
fs/nilfs2/page.c
519
index = start_blk >> (PAGE_SHIFT - inode->i_blkbits);
fs/nilfs2/page.c
52
pgoff_t index = blkoff >> (PAGE_SHIFT - blkbits);
fs/nilfs2/page.c
538
b = folio->index << (PAGE_SHIFT - inode->i_blkbits);
fs/nilfs2/segment.c
2162
nilfs->ns_blocksize_bits != PAGE_SHIFT) {
fs/nilfs2/segment.c
715
index = start >> PAGE_SHIFT;
fs/nilfs2/segment.c
716
last = end >> PAGE_SHIFT;
fs/notify/fanotify/fanotify_user.c
2191
max_marks = (((si.totalram - si.totalhigh) / 100) << PAGE_SHIFT) /
fs/notify/inotify/inotify_user.c
827
watches_max = (((si.totalram - si.totalhigh) / 100) << PAGE_SHIFT) /
fs/ntfs3/attrib.c
1573
pgoff_t index = vbo[i] >> PAGE_SHIFT;
fs/ntfs3/bitmap.c
525
ra->ra_pages = (wnd->nbits / 8 + PAGE_SIZE - 1) >> PAGE_SHIFT;
fs/ntfs3/bitmap.c
565
pgoff_t idx = lbo >> PAGE_SHIFT;
fs/ntfs3/file.c
1038
err = ntfs_get_frame_pages(mapping, frame_vbo >> PAGE_SHIFT,
fs/ntfs3/file.c
1057
ip = off >> PAGE_SHIFT;
fs/ntfs3/file.c
1093
index = frame_vbo >> PAGE_SHIFT;
fs/ntfs3/file.c
1126
ip = off >> PAGE_SHIFT;
fs/ntfs3/file.c
1333
invalidate_mapping_pages(inode->i_mapping, offset >> PAGE_SHIFT,
fs/ntfs3/file.c
1334
endbyte >> PAGE_SHIFT);
fs/ntfs3/file.c
253
u64 from = (u64)vma->vm_pgoff << PAGE_SHIFT;
fs/ntfs3/file.c
310
u64 from = (u64)desc->pgoff << PAGE_SHIFT;
fs/ntfs3/file.c
990
u32 pages_per_frame = frame_size >> PAGE_SHIFT;
fs/ntfs3/frecord.c
1915
idx = (vbo - frame_vbo) >> PAGE_SHIFT;
fs/ntfs3/frecord.c
1917
pages_per_frame = frame_size >> PAGE_SHIFT;
fs/ntfs3/frecord.c
1925
index = frame_vbo >> PAGE_SHIFT;
fs/ntfs3/frecord.c
2000
pages_per_frame = frame_size >> PAGE_SHIFT;
fs/ntfs3/frecord.c
2251
frame_size = pages_per_frame << PAGE_SHIFT;
fs/ntfs3/fsntfs.c
1238
pgoff_t index = lbo >> PAGE_SHIFT;
fs/ntfs3/fsntfs.c
1242
ra->prev_pos = (loff_t)index << PAGE_SHIFT;
fs/ntfs3/fsntfs.c
1541
folio = read_mapping_folio(mapping, lbo >> PAGE_SHIFT, NULL);
fs/ntfs3/super.c
1225
if (err == -EINVAL && !boot_block && dev_size0 > PAGE_SHIFT) {
fs/ntfs3/super.c
1728
<< (PAGE_SHIFT - sb->s_blocksize_bits);
fs/ntfs3/super.c
980
sbi->volume.blocks = dev_size >> PAGE_SHIFT;
fs/nullfs.c
17
s->s_blocksize_bits = PAGE_SHIFT;
fs/ocfs2/alloc.c
6916
index = start >> PAGE_SHIFT;
fs/ocfs2/alloc.c
6929
} while (index < (last_page_bytes >> PAGE_SHIFT));
fs/ocfs2/aops.c
1011
target_index = user_pos >> PAGE_SHIFT;
fs/ocfs2/aops.c
1030
end_index = ((last_byte - 1) >> PAGE_SHIFT) + 1;
fs/ocfs2/aops.c
1037
end_index = (user_pos + user_len - 1) >> PAGE_SHIFT;
fs/ocfs2/aops.c
1170
p_blkno += (1 << (PAGE_SHIFT - inode->i_sb->s_blocksize_bits));
fs/ocfs2/aops.c
494
if (unlikely(PAGE_SHIFT > osb->s_clustersize_bits)) {
fs/ocfs2/aops.c
497
cpp = 1 << (PAGE_SHIFT - osb->s_clustersize_bits);
fs/ocfs2/aops.c
838
if (unlikely(PAGE_SHIFT > osb->s_clustersize_bits))
fs/ocfs2/dlmfs/dlmfs.c
512
sb->s_blocksize_bits = PAGE_SHIFT;
fs/ocfs2/file.c
760
unsigned long index = abs_from >> PAGE_SHIFT;
fs/ocfs2/file.c
767
BUG_ON(abs_to > (((u64)index + 1) << PAGE_SHIFT));
fs/ocfs2/mmap.c
60
last_index = (size - 1) >> PAGE_SHIFT;
fs/ocfs2/ocfs2.h
879
if (unlikely(PAGE_SHIFT > cbits))
fs/ocfs2/ocfs2.h
880
clusters = pg_index << (PAGE_SHIFT - cbits);
fs/ocfs2/ocfs2.h
881
else if (PAGE_SHIFT < cbits)
fs/ocfs2/ocfs2.h
882
clusters = pg_index >> (cbits - PAGE_SHIFT);
fs/ocfs2/ocfs2.h
896
if (PAGE_SHIFT > cbits) {
fs/ocfs2/ocfs2.h
897
index = (pgoff_t)clusters >> (PAGE_SHIFT - cbits);
fs/ocfs2/ocfs2.h
898
} else if (PAGE_SHIFT < cbits) {
fs/ocfs2/ocfs2.h
899
index = (pgoff_t)clusters << (cbits - PAGE_SHIFT);
fs/ocfs2/ocfs2.h
910
if (PAGE_SHIFT < cbits)
fs/ocfs2/ocfs2.h
911
pages_per_cluster = 1 << (cbits - PAGE_SHIFT);
fs/ocfs2/refcounttree.c
2924
page_index = offset >> PAGE_SHIFT;
fs/ocfs2/refcounttree.c
2925
map_end = ((loff_t)page_index + 1) << PAGE_SHIFT;
fs/orangefs/super.c
422
sb->s_blocksize_bits = PAGE_SHIFT;
fs/pipe.c
1373
nr_slots = size >> PAGE_SHIFT;
fs/pipe.c
805
pipe_bufs = max_size >> PAGE_SHIFT;
fs/proc/kcore.c
150
ent->size = max_low_pfn << PAGE_SHIFT;
fs/proc/kcore.c
163
unsigned long pfn = __pa(ent->addr) >> PAGE_SHIFT;
fs/proc/kcore.c
164
unsigned long nr_pages = ent->size >> PAGE_SHIFT;
fs/proc/kcore.c
217
ent->size = nr_pages << PAGE_SHIFT;
fs/proc/kcore.c
557
pfn = phys >> PAGE_SHIFT;
fs/proc/meminfo.c
138
atomic_long_read(&num_poisoned_pages) << (PAGE_SHIFT - 10));
fs/proc/meminfo.c
30
seq_put_decimal_ull_width(m, s, num << (PAGE_SHIFT - 10), 8);
fs/proc/meminfo.c
95
(PAGE_SHIFT - 10));
fs/proc/nommu.c
56
((loff_t)region->vm_pgoff) << PAGE_SHIFT,
fs/proc/task_mmu.c
101
>> PAGE_SHIFT;
fs/proc/task_mmu.c
2005
unsigned int idx = (addr & ~PMD_MASK) >> PAGE_SHIFT;
fs/proc/task_mmu.c
2155
((addr & ~hmask) >> PAGE_SHIFT);
fs/proc/task_mmu.c
2240
pm.len = (PAGEMAP_WALK_SIZE >> PAGE_SHIFT);
fs/proc/task_mmu.c
2252
if (svpfn <= (ULONG_MAX >> PAGE_SHIFT)) {
fs/proc/task_mmu.c
2258
start_vaddr = untagged_addr_remote(mm, svpfn << PAGE_SHIFT);
fs/proc/task_mmu.c
2261
end = start_vaddr + ((count / PM_ENTRY_BYTES) << PAGE_SHIFT);
fs/proc/task_mmu.c
2982
p->vec_buf_len = min_t(size_t, PAGEMAP_WALK_SIZE >> PAGE_SHIFT,
fs/proc/task_mmu.c
36
seq_put_decimal_ull_width(m, str, (val) << (PAGE_SHIFT-10), 8)
fs/proc/task_mmu.c
478
pgoff = ((loff_t)vma->vm_pgoff) << PAGE_SHIFT;
fs/proc/task_mmu.c
62
text = min(text, mm->exec_vm << PAGE_SHIFT);
fs/proc/task_mmu.c
63
lib = (mm->exec_vm << PAGE_SHIFT) - text;
fs/proc/task_mmu.c
720
karg.vma_offset = ((__u64)vma->vm_pgoff) << PAGE_SHIFT;
fs/proc/task_nommu.c
114
>> PAGE_SHIFT;
fs/proc/task_nommu.c
116
>> PAGE_SHIFT;
fs/proc/task_nommu.c
118
size >>= PAGE_SHIFT;
fs/proc/task_nommu.c
143
pgoff = (loff_t)vma->vm_pgoff << PAGE_SHIFT;
fs/proc/vmcore.c
485
offset = (loff_t) index << PAGE_SHIFT;
fs/proc/vmcore.c
532
pos_end = pfn + (size >> PAGE_SHIFT);
fs/proc/vmcore.c
543
map_size = (pos - pos_start) << PAGE_SHIFT;
fs/proc/vmcore.c
561
map_size = (pos - pos_start) << PAGE_SHIFT;
fs/proc/vmcore.c
597
start = (u64)vma->vm_pgoff << PAGE_SHIFT;
fs/proc/vmcore.c
615
pfn = __pa(elfcorebuf + start) >> PAGE_SHIFT;
fs/proc/vmcore.c
686
paddr >> PAGE_SHIFT, tsz,
fs/pstore/inode.c
420
sb->s_blocksize_bits = PAGE_SHIFT;
fs/pstore/ram_core.c
449
pages[i] = pfn_to_page(addr >> PAGE_SHIFT);
fs/pstore/ram_core.c
508
if (pfn_valid(start >> PAGE_SHIFT))
fs/pstore/ram_core.c
582
if (pfn_valid(prz->paddr >> PAGE_SHIFT)) {
fs/qnx6/dir.c
123
unsigned long n = pos >> PAGE_SHIFT;
fs/qnx6/dir.c
139
ctx->pos = (n + 1) << PAGE_SHIFT;
fs/qnx6/dir.c
42
last_byte -= page_nr << PAGE_SHIFT;
fs/qnx6/dir.c
54
u32 n = s >> (PAGE_SHIFT - sb->s_blocksize_bits); /* in pages */
fs/qnx6/inode.c
537
n = (ino - 1) >> (PAGE_SHIFT - QNX6_INODE_SIZE_BITS);
fs/ramfs/file-nommu.c
212
lpages = (len + PAGE_SIZE - 1) >> PAGE_SHIFT;
fs/ramfs/file-nommu.c
216
maxpages = (isize + PAGE_SIZE - 1) >> PAGE_SHIFT;
fs/ramfs/file-nommu.c
90
npages = (newsize + PAGE_SIZE - 1) >> PAGE_SHIFT;
fs/ramfs/inode.c
267
sb->s_blocksize_bits = PAGE_SHIFT;
fs/remap_range.c
162
return read_mapping_folio(file->f_mapping, pos >> PAGE_SHIFT, file);
fs/resctrl/pseudo_lock.c
1000
unsigned long off = desc->pgoff << PAGE_SHIFT;
fs/resctrl/pseudo_lock.c
1035
physical = __pa(plr->kmem) >> PAGE_SHIFT;
fs/romfs/mmap-nommu.c
33
lpages = (len + PAGE_SIZE - 1) >> PAGE_SHIFT;
fs/romfs/mmap-nommu.c
35
offset = pgoff << PAGE_SHIFT;
fs/romfs/mmap-nommu.c
37
maxpages = (isize + PAGE_SIZE - 1) >> PAGE_SHIFT;
fs/romfs/mmap-nommu.c
44
if (len > mtd->size || pgoff >= (mtd->size >> PAGE_SHIFT))
fs/smb/server/auth.c
770
PAGE_SIZE - 1) >> PAGE_SHIFT) -
fs/smb/server/auth.c
771
(kaddr >> PAGE_SHIFT);
fs/splice.c
967
len = min_t(size_t, len, p_space << PAGE_SHIFT);
fs/squashfs/block.c
153
read_start >> PAGE_SHIFT,
fs/squashfs/block.c
165
(read_end >> PAGE_SHIFT) - 1,
fs/squashfs/block.c
182
(read_start >> PAGE_SHIFT) + idx,
fs/squashfs/block.c
240
pgoff_t index = (read_start >> PAGE_SHIFT) + i;
fs/squashfs/cache.c
249
cache->pages = block_size >> PAGE_SHIFT;
fs/squashfs/cache.c
415
int pages = (length + PAGE_SIZE - 1) >> PAGE_SHIFT;
fs/squashfs/file.c
399
int i, mask = (1 << (msblk->block_log - PAGE_SHIFT)) - 1;
fs/squashfs/file.c
466
int index = folio->index >> (msblk->block_log - PAGE_SHIFT);
fs/squashfs/file.c
477
PAGE_SHIFT))
fs/squashfs/file.c
579
unsigned short shift = msblk->block_log - PAGE_SHIFT;
fs/squashfs/file.c
605
max_pages = (expected + PAGE_SIZE - 1) >> PAGE_SHIFT;
fs/squashfs/file_direct.c
28
loff_t file_end = (i_size_read(inode) - 1) >> PAGE_SHIFT;
fs/squashfs/file_direct.c
29
int mask = (1 << (msblk->block_log - PAGE_SHIFT)) - 1;
fs/squashfs/file_direct.c
71
start_index << PAGE_SHIFT);
fs/squashfs/page_actor.c
133
actor->next_index = start_index >> PAGE_SHIFT;
fs/squashfs/page_actor.c
70
int max_pages = (actor->length + PAGE_SIZE - 1) >> PAGE_SHIFT;
fs/sync.c
246
if (offset >= (0x100000000ULL << PAGE_SHIFT)) {
fs/sync.c
254
if (endbyte >= (0x100000000ULL << PAGE_SHIFT)) {
fs/ubifs/file.c
1137
pgoff_t index = new_size >> PAGE_SHIFT;
fs/ubifs/file.c
210
pgoff_t index = pos >> PAGE_SHIFT;
fs/ubifs/file.c
415
pgoff_t index = pos >> PAGE_SHIFT;
fs/ubifs/file.c
605
end_index = (i_size - 1) >> PAGE_SHIFT;
fs/ubifs/file.c
760
end_index = ((isize - 1) >> PAGE_SHIFT);
fs/ubifs/file.c
895
ubifs_assert(c, folio->index <= ui->synced_i_size >> PAGE_SHIFT);
fs/ubifs/ubifs.h
48
#define UBIFS_BLOCKS_PER_PAGE_SHIFT (PAGE_SHIFT - UBIFS_BLOCK_SHIFT)
fs/udf/file.c
62
if (folio->index == size >> PAGE_SHIFT)
fs/ufs/balloc.c
240
1 << (PAGE_SHIFT - inode->i_blkbits);
fs/ufs/balloc.c
256
last_index = end >> (PAGE_SHIFT - inode->i_blkbits);
fs/ufs/balloc.c
258
index = i >> (PAGE_SHIFT - inode->i_blkbits);
fs/ufs/dir.c
218
last_byte -= page_nr << PAGE_SHIFT;
fs/ufs/dir.c
418
unsigned long n = pos >> PAGE_SHIFT;
fs/ufs/dir.c
445
ctx->pos = (n<<PAGE_SHIFT) + offset;
fs/ufs/inode.c
1053
(PAGE_SHIFT - inode->i_blkbits));
fs/ufs/inode.c
1059
end = lastfrag & ((1 << (PAGE_SHIFT - inode->i_blkbits)) - 1);
fs/verity/open.c
157
params->tree_pages = PAGE_ALIGN(params->tree_size) >> PAGE_SHIFT;
fs/verity/open.c
83
if (log_blocksize < 10 || log_blocksize > PAGE_SHIFT ||
fs/verity/open.c
92
params->log_blocks_per_page = PAGE_SHIFT - log_blocksize;
fs/verity/read_metadata.c
31
index = offset >> PAGE_SHIFT;
fs/verity/read_metadata.c
32
last_index = (end_offset - 1) >> PAGE_SHIFT;
fs/verity/verify.c
384
u64 pos = (u64)data_folio->index << PAGE_SHIFT;
fs/xfs/libxfs/xfs_ialloc.c
3082
igeo->min_folio_order = mp->m_sb.sb_blocklog - PAGE_SHIFT;
fs/xfs/scrub/xfile.c
130
if (shmem_get_folio(inode, pos >> PAGE_SHIFT, 0, &folio,
fs/xfs/scrub/xfile.c
200
if (shmem_get_folio(inode, pos >> PAGE_SHIFT, 0, &folio,
fs/xfs/scrub/xfile.c
271
error = shmem_get_folio(inode, pos >> PAGE_SHIFT, 0, &folio,
fs/xfs/xfs_buf.c
111
mm_account_reclaimed_pages(howmany(size, PAGE_SHIFT));
fs/xfs/xfs_buf_mem.c
144
error = shmem_get_folio(inode, pos >> PAGE_SHIFT, 0, &folio, SGP_CACHE);
fs/xfs/xfs_buf_mem.h
10
#define XMBUF_BLOCKSHIFT (PAGE_SHIFT)
fs/xfs/xfs_mount.c
146
if (max_bytes >> PAGE_SHIFT > ULONG_MAX)
fs/xfs/xfs_mount.h
658
#define XFS_MIN_IO_LOG PAGE_SHIFT
fs/xfs/xfs_notify_failure.c
121
fserror_report_data_lost(VFS_I(ip), (u64)pgoff << PAGE_SHIFT,
fs/xfs/xfs_notify_failure.c
122
(u64)pgcnt << PAGE_SHIFT, GFP_NOFS);
fs/xfs/xfs_notify_failure.c
50
return pos >> PAGE_SHIFT;
fs/xfs/xfs_notify_failure.c
70
return XFS_FSB_TO_B(mp, end_cross - start_cross) >> PAGE_SHIFT;
fs/xfs/xfs_platform.h
129
#define BLKDEV_IOSHIFT PAGE_SHIFT
fs/xfs/xfs_pnfs.c
298
start >> PAGE_SHIFT,
fs/xfs/xfs_pnfs.c
299
(end - 1) >> PAGE_SHIFT);
fs/xfs/xfs_zone_alloc.c
1403
PAGE_SHIFT;
include/asm-generic/fixmap.h
21
#define __fix_to_virt(x) (FIXADDR_TOP - ((x) << PAGE_SHIFT))
include/asm-generic/fixmap.h
22
#define __virt_to_fix(x) ((FIXADDR_TOP - ((x)&PAGE_MASK)) >> PAGE_SHIFT)
include/asm-generic/getorder.h
33
return BITS_PER_LONG - PAGE_SHIFT;
include/asm-generic/getorder.h
35
if (size < (1UL << PAGE_SHIFT))
include/asm-generic/getorder.h
38
return ilog2((size) - 1) - PAGE_SHIFT + 1;
include/asm-generic/getorder.h
42
size >>= PAGE_SHIFT;
include/asm-generic/tlb.h
546
return PAGE_SHIFT;
include/asm-generic/tlb.h
554
return PAGE_SHIFT;
include/crypto/scatterwalk.h
109
page = base_page + (offset >> PAGE_SHIFT);
include/drm/drm_vma_manager.h
205
return ((__u64)node->vm_node.start) << PAGE_SHIFT;
include/drm/drm_vma_manager.h
226
drm_vma_node_size(node) << PAGE_SHIFT, 1);
include/drm/drm_vma_manager.h
37
#define DRM_FILE_PAGE_OFFSET_START ((0xFFFFFFFFUL >> PAGE_SHIFT) + 1)
include/drm/drm_vma_manager.h
38
#define DRM_FILE_PAGE_OFFSET_SIZE ((0xFFFFFFFFUL >> PAGE_SHIFT) * 256)
include/drm/drm_vma_manager.h
40
#define DRM_FILE_PAGE_OFFSET_START ((0xFFFFFFFUL >> PAGE_SHIFT) + 1)
include/drm/drm_vma_manager.h
41
#define DRM_FILE_PAGE_OFFSET_SIZE ((0xFFFFFFFUL >> PAGE_SHIFT) * 16)
include/linux/blk_types.h
37
#define PAGE_SECTORS_SHIFT (PAGE_SHIFT - SECTOR_SHIFT)
include/linux/bvec.h
223
bv->bv_page = bvec->bv_page + (bvec->bv_offset >> PAGE_SHIFT);
include/linux/ceph/libceph.h
182
return ((off+len+PAGE_SIZE-1) >> PAGE_SHIFT) -
include/linux/ceph/libceph.h
183
(off >> PAGE_SHIFT);
include/linux/f2fs_fs.h
16
#define F2FS_MAX_LOG_SECTOR_SIZE PAGE_SHIFT /* Max is Block Size */
include/linux/f2fs_fs.h
17
#define F2FS_LOG_SECTORS_PER_BLOCK (PAGE_SHIFT - 9) /* log number for sector/blk */
include/linux/f2fs_fs.h
19
#define F2FS_BLKSIZE_BITS PAGE_SHIFT /* bits for F2FS_BLKSIZE */
include/linux/filter.h
1091
return set_memory_rox((unsigned long)hdr, hdr->size >> PAGE_SHIFT);
include/linux/fs.h
1326
#define MAX_LFS_FILESIZE ((loff_t)ULONG_MAX << PAGE_SHIFT)
include/linux/huge_mm.h
117
#define HPAGE_PMD_ORDER (HPAGE_PMD_SHIFT-PAGE_SHIFT)
include/linux/huge_mm.h
122
#define HPAGE_PUD_ORDER (HPAGE_PUD_SHIFT-PAGE_SHIFT)
include/linux/huge_mm.h
228
if (!IS_ALIGNED((vma->vm_start >> PAGE_SHIFT) - vma->vm_pgoff,
include/linux/huge_mm.h
229
hpage_size >> PAGE_SHIFT))
include/linux/hugetlb.h
1201
return PAGE_SHIFT;
include/linux/hugetlb.h
793
return h->order + PAGE_SHIFT;
include/linux/hugetlb.h
882
return hstates[index].order + PAGE_SHIFT;
include/linux/io.h
189
u64 from = ((u64)pfn) << PAGE_SHIFT;
include/linux/kexec.h
506
free_reserved_page(boot_pfn_to_page(addr >> PAGE_SHIFT));
include/linux/kho/abi/kexec_handover.h
252
KHO_ORDER_0_LOG2 = 64 - PAGE_SHIFT,
include/linux/kho/abi/kexec_handover.h
258
KHO_BITMAP_SIZE_LOG2 = PAGE_SHIFT + const_ilog2(BITS_PER_BYTE),
include/linux/kvm_host.h
1351
__kvm_get_guest(__kvm, __gpa >> PAGE_SHIFT, \
include/linux/kvm_host.h
1373
__kvm_put_guest(__kvm, __gpa >> PAGE_SHIFT, \
include/linux/kvm_host.h
1899
gfn_t gfn_offset = (hva - slot->userspace_addr) >> PAGE_SHIFT;
include/linux/kvm_host.h
1906
return (gpa_t)gfn << PAGE_SHIFT;
include/linux/kvm_host.h
1911
return (gfn_t)(gpa >> PAGE_SHIFT);
include/linux/kvm_host.h
1916
return (hpa_t)pfn << PAGE_SHIFT;
include/linux/mlx4/device.h
1093
return buf->page_list[offset >> PAGE_SHIFT].buf +
include/linux/mlx4/device.h
1599
return (index << (PAGE_SHIFT - dev->uar_page_shift));
include/linux/mlx4/device.h
1605
return (128 >> (PAGE_SHIFT - dev->uar_page_shift));
include/linux/mlx5/driver.h
931
fbc->log_frag_strides = PAGE_SHIFT - fbc->log_stride;
include/linux/mm.h
1927
return PAGE_SHIFT + compound_order(page);
include/linux/mm.h
2797
return PAGE_SHIFT + folio_order(folio);
include/linux/mm.h
4242
return (vma->vm_end - vma->vm_start) >> PAGE_SHIFT;
include/linux/mm.h
4257
return vma_desc_size(desc) >> PAGE_SHIFT;
include/linux/mm.h
4810
#define PROCESS_PAGES_NON_PREEMPT_BATCH (SZ_32M >> PAGE_SHIFT)
include/linux/mm.h
4932
nr_vmemmap_pages = ((nr_pages * sizeof(struct page)) >> PAGE_SHIFT);
include/linux/mm.h
5131
return range_contains_unaccepted_memory(pfn << PAGE_SHIFT, PAGE_SIZE);
include/linux/mm.h
82
#define PAGES_TO_MB(pages) ((pages) >> (20 - PAGE_SHIFT))
include/linux/mm.h
83
#define MB_TO_PAGES(mb) ((mb) << (20 - PAGE_SHIFT))
include/linux/mmzone.h
1936
#define PFN_SECTION_SHIFT (SECTION_SIZE_BITS - PAGE_SHIFT)
include/linux/mmzone.h
1946
#if (MAX_PAGE_ORDER + PAGE_SHIFT) > SECTION_SIZE_BITS
include/linux/mmzone.h
1965
#define PFN_SUBSECTION_SHIFT (SUBSECTION_SHIFT - PAGE_SHIFT)
include/linux/mmzone.h
85
#define MAX_FOLIO_ORDER (ilog2(SZ_16G) - PAGE_SHIFT)
include/linux/mmzone.h
87
#define MAX_FOLIO_ORDER (ilog2(SZ_1G) - PAGE_SHIFT)
include/linux/mmzone.h
94
#define MAX_FOLIO_ORDER (PUD_SHIFT - PAGE_SHIFT)
include/linux/net.h
354
struct page *p = page + (offset >> PAGE_SHIFT);
include/linux/nfs_page.h
195
return folio_page(folio, pgbase >> PAGE_SHIFT);
include/linux/nfs_page.h
277
return (((loff_t)req->wb_index) << PAGE_SHIFT) + req->wb_offset;
include/linux/nfsacl.h
20
>> PAGE_SHIFT)
include/linux/pagemap.h
1072
pgoff = (address - vma->vm_start) >> PAGE_SHIFT;
include/linux/pagemap.h
1535
PAGE_SHIFT;
include/linux/pagemap.h
1550
pgoff_t index = size >> PAGE_SHIFT;
include/linux/pagemap.h
398
return 1U << (PAGE_SHIFT + MAX_PAGECACHE_ORDER);
include/linux/pagemap.h
485
return mapping_min_folio_nrpages(mapping) << PAGE_SHIFT;
include/linux/pagemap.h
728
if (shift <= PAGE_SHIFT)
include/linux/pagemap.h
731
return shift - PAGE_SHIFT;
include/linux/pagemap.h
957
return (loff_t)folio_next_index(folio) << PAGE_SHIFT;
include/linux/pfn.h
10
#define PFN_UP(x) (((x) + PAGE_SIZE-1) >> PAGE_SHIFT)
include/linux/pfn.h
11
#define PFN_DOWN(x) ((x) >> PAGE_SHIFT)
include/linux/pfn.h
12
#define PFN_PHYS(x) ((phys_addr_t)(x) << PAGE_SHIFT)
include/linux/pfn.h
13
#define PHYS_PFN(x) ((unsigned long)((x) >> PAGE_SHIFT))
include/linux/pgtable.h
1976
return offset_from_zero_pfn <= (zero_page_mask >> PAGE_SHIFT);
include/linux/pgtable.h
50
return (address >> PAGE_SHIFT) & (PTRS_PER_PTE - 1);
include/linux/pgtable.h
8
#define PMD_ORDER (PMD_SHIFT - PAGE_SHIFT)
include/linux/pgtable.h
9
#define PUD_ORDER (PUD_SHIFT - PAGE_SHIFT)
include/linux/raid/pq.h
44
#ifndef PAGE_SHIFT
include/linux/scatterlist.h
618
(dma_iter->base.sg_pgoffset << PAGE_SHIFT);
include/linux/skbuff.h
437
for (p = skb_frag_page(f) + ((f_off) >> PAGE_SHIFT), \
include/linux/slab.h
584
#define KMALLOC_SHIFT_HIGH (PAGE_SHIFT + 1)
include/linux/slab.h
585
#define KMALLOC_SHIFT_MAX (MAX_PAGE_ORDER + PAGE_SHIFT)
include/linux/slab.h
595
#define KMALLOC_MAX_ORDER (KMALLOC_SHIFT_MAX - PAGE_SHIFT)
include/linux/slab.h
751
static_assert(PAGE_SHIFT <= 20);
include/linux/stackdepot.h
36
#define DEPOT_POOL_SIZE (1LL << (PAGE_SHIFT + DEPOT_POOL_ORDER))
include/linux/stackdepot.h
38
#define DEPOT_OFFSET_BITS (DEPOT_POOL_ORDER + PAGE_SHIFT - DEPOT_STACK_ALIGN)
include/linux/swapops.h
36
#define SWP_PFN_BITS (MAX_PHYSMEM_BITS - PAGE_SHIFT)
include/linux/swapops.h
39
sizeof(phys_addr_t) * 8 - PAGE_SHIFT, \
include/linux/vmalloc.h
132
return PAGE_SHIFT;
include/linux/vmalloc.h
49
#define IOREMAP_MAX_ORDER (7 + PAGE_SHIFT) /* 128 pages */
include/linux/vmstat.h
328
delta >>= PAGE_SHIFT;
include/linux/writeback.h
376
#define MIN_WRITEBACK_PAGES (4096UL >> (PAGE_SHIFT - 10))
include/net/page_pool/helpers.h
433
ret <<= PAGE_SHIFT;
include/net/proto_memory.h
9
#define SK_MEMORY_PCPU_RESERVE (1 << (20 - PAGE_SHIFT))
include/net/sock.h
1568
return (amt + PAGE_SIZE - 1) >> PAGE_SHIFT;
include/net/xsk_buff_pool.h
130
xskb->frame_dma = (dma_pages[addr >> PAGE_SHIFT] & ~XSK_NEXT_PG_CONTIG_MASK) +
include/net/xsk_buff_pool.h
186
!(pool->dma_pages[addr >> PAGE_SHIFT] & XSK_NEXT_PG_CONTIG_MASK);
include/rdma/ib_verbs.h
2377
return (u64)entry->start_pgoff << PAGE_SHIFT;
include/sound/emu10k1.h
1557
#define snd_emu10k1_memblk_offset(blk) (((blk)->mapped_page << PAGE_SHIFT) | ((blk)->mem.offset & (PAGE_SIZE - 1)))
include/sound/memalloc.h
68
return (size + PAGE_SIZE - 1) >> PAGE_SHIFT;
include/trace/events/filemap.h
139
((loff_t)__entry->index) << PAGE_SHIFT
include/trace/events/filemap.h
45
__entry->index << PAGE_SHIFT,
include/trace/events/filemap.h
91
((loff_t)__entry->index) << PAGE_SHIFT,
include/trace/events/filemap.h
92
((((loff_t)__entry->last_index + 1) << PAGE_SHIFT) - 1)
include/trace/events/kmem.h
452
<< PAGE_SHIFT);
include/trace/events/oom.h
10
#define PG_COUNT_TO_KB(x) ((x) << (PAGE_SHIFT - 10))
include/trace/events/writeback.h
605
#define KBps(x) ((x) << (PAGE_SHIFT - 10))
include/vdso/datapage.h
168
#define VDSO_ARCH_DATA_PAGES (VDSO_ARCH_DATA_SIZE >> PAGE_SHIFT)
include/xen/page.h
19
(pfn_to_page((unsigned long)(xen_pfn) >> (PAGE_SHIFT - XEN_PAGE_SHIFT)))
include/xen/page.h
21
((page_to_pfn(page)) << (PAGE_SHIFT - XEN_PAGE_SHIFT))
io_uring/memmap.c
202
if ((reg->size >> PAGE_SHIFT) > INT_MAX)
io_uring/memmap.c
209
nr_pages = reg->size >> PAGE_SHIFT;
io_uring/memmap.c
236
loff_t offset = pgoff << PAGE_SHIFT;
io_uring/memmap.c
299
long offset = vma->vm_pgoff << PAGE_SHIFT;
io_uring/memmap.c
313
page_limit = (sz + PAGE_SIZE - 1) >> PAGE_SHIFT;
io_uring/memmap.c
360
pgoff = addr >> PAGE_SHIFT;
io_uring/memmap.c
51
end = end >> PAGE_SHIFT;
io_uring/memmap.c
52
start = uaddr >> PAGE_SHIFT;
io_uring/memmap.h
49
return (size_t) mr->nr_pages << PAGE_SHIFT;
io_uring/notif.h
47
unsigned nr_pages = (len >> PAGE_SHIFT) + 2;
io_uring/rsrc.c
47
page_limit = rlimit(RLIMIT_MEMLOCK) >> PAGE_SHIFT;
io_uring/rsrc.c
664
imu->acct_pages += page_size(hpage) >> PAGE_SHIFT;
io_uring/rsrc.c
819
imu->folio_shift = PAGE_SHIFT;
io_uring/rsrc.c
830
off += data.first_folio_page_idx << PAGE_SHIFT;
io_uring/rsrc.c
972
imu->folio_shift = PAGE_SHIFT;
io_uring/zcrx.c
1140
if (pp->p.order + PAGE_SHIFT != ifq->niov_shift)
io_uring/zcrx.c
209
0, (unsigned long)nr_pages << PAGE_SHIFT,
io_uring/zcrx.c
443
int buf_size_shift = PAGE_SHIFT;
io_uring/zcrx.c
454
if (!ifq->dev && buf_size_shift != PAGE_SHIFT)
io_uring/zcrx.c
54
niov_pages_shift = area->ifq->niov_shift - PAGE_SHIFT;
ipc/mqueue.c
410
sb->s_blocksize_bits = PAGE_SHIFT;
ipc/shm.c
337
ns->shm_tot -= (shp->shm_segsz + PAGE_SIZE - 1) >> PAGE_SHIFT;
ipc/shm.c
709
size_t numpages = (size + PAGE_SIZE - 1) >> PAGE_SHIFT;
ipc/shm.c
718
if (numpages << PAGE_SHIFT < size)
kernel/bpf/arena.c
113
return (u32)(uaddr - (u32)arena->user_vm_start) >> PAGE_SHIFT;
kernel/bpf/arena.c
540
long page_cnt_max = (arena->user_vm_end - arena->user_vm_start) >> PAGE_SHIFT;
kernel/bpf/arena.c
613
kern_vm_start + uaddr32 + (mapped << PAGE_SHIFT),
kernel/bpf/arena.c
614
this_batch << PAGE_SHIFT, apply_range_set_cb, &data);
kernel/bpf/arena.c
626
flush_vmap_cache(kern_vm_start + uaddr32, mapped << PAGE_SHIFT);
kernel/bpf/arena.c
635
flush_vmap_cache(kern_vm_start + uaddr32, mapped << PAGE_SHIFT);
kernel/bpf/arena.c
679
uaddr_end = min(arena->user_vm_end, full_uaddr + (page_cnt << PAGE_SHIFT));
kernel/bpf/arena.c
683
page_cnt = (uaddr_end - full_uaddr) >> PAGE_SHIFT;
kernel/bpf/arena.c
700
apply_to_existing_page_range(&init_mm, kaddr, page_cnt << PAGE_SHIFT,
kernel/bpf/arena.c
751
long page_cnt_max = (arena->user_vm_end - arena->user_vm_start) >> PAGE_SHIFT;
kernel/bpf/arena.c
815
apply_to_existing_page_range(&init_mm, kaddr, page_cnt << PAGE_SHIFT,
kernel/bpf/arraymap.c
581
pgoff_t pgoff = PAGE_ALIGN(sizeof(*array)) >> PAGE_SHIFT;
kernel/bpf/local_storage.c
493
PAGE_SIZE) >> PAGE_SHIFT;
kernel/bpf/local_storage.c
497
PAGE_SIZE) >> PAGE_SHIFT;
kernel/bpf/ringbuf.c
21
(offsetof(struct bpf_ringbuf, consumer_pos) >> PAGE_SHIFT)
kernel/bpf/ringbuf.c
370
usage += (u64)rb->nr_pages << PAGE_SHIFT;
kernel/bpf/ringbuf.c
372
nr_data_pages = map->max_entries >> PAGE_SHIFT;
kernel/bpf/ringbuf.c
416
return ((void *)hdr - (void *)rb) >> PAGE_SHIFT;
kernel/bpf/ringbuf.c
426
unsigned long off = (unsigned long)hdr->pg_off << PAGE_SHIFT;
kernel/bpf/ringbuf.c
98
int nr_data_pages = data_sz >> PAGE_SHIFT;
kernel/bpf/stackmap.c
204
id_offs[i].offset = (vma->vm_pgoff << PAGE_SHIFT) + ip - vma->vm_start;
kernel/bpf/syscall.c
2501
prog->pages * 1ULL << PAGE_SHIFT,
kernel/bpf/sysfs_btf.c
22
unsigned long pages = PAGE_ALIGN(attr->size) >> PAGE_SHIFT;
kernel/bpf/sysfs_btf.c
25
unsigned long pfn = addr >> PAGE_SHIFT;
kernel/bpf/sysfs_btf.c
39
if ((vm_size >> PAGE_SHIFT) > pages)
kernel/crash_core.c
615
ptr = kmap_local_page(pfn_to_page(mem >> PAGE_SHIFT));
kernel/crash_dump_dm_crypt.c
94
kexec_crash_image->dm_crypt_keys_addr >> PAGE_SHIFT));
kernel/debug/kdb/kdb_main.c
2415
#define K(x) ((x) << (PAGE_SHIFT - 10))
kernel/debug/kdb/kdb_support.c
351
pfn = (addr >> PAGE_SHIFT);
kernel/dma/coherent.c
152
if (unlikely(size > ((dma_addr_t)mem->size << PAGE_SHIFT)))
kernel/dma/coherent.c
163
((dma_addr_t)pageno << PAGE_SHIFT);
kernel/dma/coherent.c
164
ret = mem->virt_base + ((dma_addr_t)pageno << PAGE_SHIFT);
kernel/dma/coherent.c
203
(mem->virt_base + ((dma_addr_t)mem->size << PAGE_SHIFT))) {
kernel/dma/coherent.c
204
int page = (vaddr - mem->virt_base) >> PAGE_SHIFT;
kernel/dma/coherent.c
238
(mem->virt_base + ((dma_addr_t)mem->size << PAGE_SHIFT))) {
kernel/dma/coherent.c
240
int start = (vaddr - mem->virt_base) >> PAGE_SHIFT;
kernel/dma/coherent.c
242
int count = PAGE_ALIGN(size) >> PAGE_SHIFT;
kernel/dma/coherent.c
248
user_count << PAGE_SHIFT,
kernel/dma/coherent.c
42
int pages = size >> PAGE_SHIFT;
kernel/dma/contiguous.c
156
return (total_pages * CONFIG_CMA_SIZE_PERCENTAGE / 100) << PAGE_SHIFT;
kernel/dma/contiguous.c
354
return cma_alloc(cma, size >> PAGE_SHIFT, align, gfp & __GFP_NOWARN);
kernel/dma/contiguous.c
424
unsigned int count = PAGE_ALIGN(size) >> PAGE_SHIFT;
kernel/dma/debug.c
421
#define CACHELINE_PER_PAGE_SHIFT (PAGE_SHIFT - L1_CACHE_SHIFT)
kernel/dma/debug.c
426
return ((entry->paddr >> PAGE_SHIFT) << CACHELINE_PER_PAGE_SHIFT) +
kernel/dma/direct.c
42
phys_addr_t phys = (phys_addr_t)(max_pfn - 1) << PAGE_SHIFT;
kernel/dma/direct.c
522
unsigned long count = PAGE_ALIGN(size) >> PAGE_SHIFT;
kernel/dma/direct.c
538
user_count << PAGE_SHIFT, vma->vm_page_prot);
kernel/dma/direct.c
543
u64 min_mask = (max_pfn - 1) << PAGE_SHIFT;
kernel/dma/mapping.c
761
unsigned long count = PAGE_ALIGN(size) >> PAGE_SHIFT;
kernel/dma/mapping.c
767
vma_pages(vma) << PAGE_SHIFT, vma->vm_page_prot);
kernel/dma/ops_helpers.c
41
unsigned long count = PAGE_ALIGN(size) >> PAGE_SHIFT;
kernel/dma/ops_helpers.c
56
user_count << PAGE_SHIFT, vma->vm_page_prot);
kernel/dma/pool.c
168
pool = gen_pool_create(PAGE_SHIFT, NUMA_NO_NODE);
kernel/dma/pool.c
204
atomic_pool_size = max_t(size_t, pages << PAGE_SHIFT, SZ_128K);
kernel/dma/pool.c
91
pool_size = 1 << (PAGE_SHIFT + order);
kernel/dma/remap.c
29
vaddr = vmap(pages, PAGE_ALIGN(size) >> PAGE_SHIFT,
kernel/dma/remap.c
43
int count = PAGE_ALIGN(size) >> PAGE_SHIFT;
kernel/dma/swiotlb.c
1777
align = (1 << (get_order(size) + PAGE_SHIFT)) - 1;
kernel/dma/swiotlb.c
1848
rmem->size >> PAGE_SHIFT);
kernel/dma/swiotlb.c
265
set_memory_decrypted((unsigned long)mem->vaddr, bytes >> PAGE_SHIFT);
kernel/dma/swiotlb.c
509
(nslabs << IO_TLB_SHIFT) >> PAGE_SHIFT);
kernel/dma/swiotlb.c
542
set_memory_encrypted(tbl_vaddr, tbl_size >> PAGE_SHIFT);
kernel/dma/swiotlb.c
56
#define SLABS_PER_PAGE (1 << (PAGE_SHIFT - IO_TLB_SHIFT))
kernel/events/core.c
7121
atomic_long_sub((size >> PAGE_SHIFT) + 1 - mmap_locked,
kernel/events/core.c
7226
user_lock_limit = sysctl_perf_event_mlock >> (PAGE_SHIFT - 10);
kernel/events/core.c
7250
lock_limit >>= PAGE_SHIFT;
kernel/events/core.c
7367
if (aux_offset != vma->vm_pgoff << PAGE_SHIFT)
kernel/events/core.c
9819
unsigned long off = vma->vm_pgoff << PAGE_SHIFT;
kernel/events/core.c
9909
.pgoff = (u64)vma->vm_pgoff << PAGE_SHIFT,
kernel/events/internal.h
128
return rb->nr_pages << (PAGE_SHIFT + page_order(rb));
kernel/events/internal.h
133
return (unsigned long)rb->aux_nr_pages << PAGE_SHIFT;
kernel/events/ring_buffer.c
241
page_shift = PAGE_SHIFT + page_order(rb);
kernel/events/ring_buffer.c
589
from &= (rb->aux_nr_pages << PAGE_SHIFT) - 1;
kernel/events/ring_buffer.c
590
to &= (rb->aux_nr_pages << PAGE_SHIFT) - 1;
kernel/events/ring_buffer.c
599
addr = rb->aux_pages[from >> PAGE_SHIFT];
kernel/events/ring_buffer.c
608
from &= (rb->aux_nr_pages << PAGE_SHIFT) - 1;
kernel/events/ring_buffer.c
706
(unsigned long)nr_pages << (PAGE_SHIFT - 1));
kernel/events/ring_buffer.c
845
if (order_base_2(size) > PAGE_SHIFT+MAX_PAGE_ORDER)
kernel/events/uprobes.c
1058
page = read_mapping_page(mapping, offset >> PAGE_SHIFT, filp);
kernel/events/uprobes.c
1060
page = shmem_read_mapping_page(mapping, offset >> PAGE_SHIFT);
kernel/events/uprobes.c
1204
unsigned long pgoff = offset >> PAGE_SHIFT;
kernel/events/uprobes.c
147
return vma->vm_start + offset - ((loff_t)vma->vm_pgoff << PAGE_SHIFT);
kernel/events/uprobes.c
1485
offset = (loff_t)vma->vm_pgoff << PAGE_SHIFT;
kernel/events/uprobes.c
152
return ((loff_t)vma->vm_pgoff << PAGE_SHIFT) + (vaddr - vma->vm_start);
kernel/events/uprobes.c
2456
offset = (loff_t)(vma->vm_pgoff << PAGE_SHIFT) + (bp_vaddr - vma->vm_start);
kernel/events/uprobes.c
386
const pgoff_t index = vaddr_to_offset(vma, vaddr) >> PAGE_SHIFT;
kernel/fork.c
2345
p->nr_dirtied_pause = 128 >> (PAGE_SHIFT - 10);
kernel/kexec_core.c
105
#define PAGE_COUNT(x) (((x) + PAGE_SIZE - 1) >> PAGE_SHIFT)
kernel/kexec_core.c
364
addr = pfn << PAGE_SHIFT;
kernel/kexec_core.c
365
eaddr = (epfn << PAGE_SHIFT) - 1;
kernel/kexec_core.c
366
if ((epfn >= (KEXEC_CONTROL_MEMORY_LIMIT >> PAGE_SHIFT)) ||
kernel/kexec_core.c
425
size = (1 << order) << PAGE_SHIFT;
kernel/kexec_core.c
450
pages = pfn_to_page(hole_start >> PAGE_SHIFT);
kernel/kexec_core.c
554
page = boot_pfn_to_page(entry >> PAGE_SHIFT);
kernel/kexec_core.c
564
u32 nr_pages = image->segment[i].memsz >> PAGE_SHIFT;
kernel/kexec_core.c
676
addr = page_to_boot_pfn(page) << PAGE_SHIFT;
kernel/kexec_core.c
692
(KEXEC_SOURCE_MEMORY_LIMIT >> PAGE_SHIFT)) {
kernel/kexec_core.c
696
addr = page_to_boot_pfn(page) << PAGE_SHIFT;
kernel/kexec_core.c
719
old_page = boot_pfn_to_page(old_addr >> PAGE_SHIFT);
kernel/kexec_core.c
832
<< PAGE_SHIFT);
kernel/kexec_core.c
895
page = boot_pfn_to_page(maddr >> PAGE_SHIFT);
kernel/kexec_file.c
672
size_t nr_pages = kbuf->memsz >> PAGE_SHIFT;
kernel/kexec_file.c
690
mem = page_to_boot_pfn(p) << PAGE_SHIFT;
kernel/kexec_file.c
699
kbuf->mem = page_to_boot_pfn(p) << PAGE_SHIFT;
kernel/kexec_file.c
815
zero_buf = __va(page_to_pfn(ZERO_PAGE(0)) << PAGE_SHIFT);
kernel/liveupdate/kexec_handover.c
101
unsigned long l = phys >> (PAGE_SHIFT + order);
kernel/liveupdate/kexec_handover.c
1108
shift = PAGE_SHIFT + order;
kernel/liveupdate/kexec_handover.c
125
phys = key << (PAGE_SHIFT + *order);
kernel/liveupdate/kexec_handover.c
1416
unsigned long count = kho_scratch[i].size >> PAGE_SHIFT;
kernel/liveupdate/kexec_handover.c
481
sz = 1 << (order + PAGE_SHIFT);
kernel/liveupdate/kexec_handover.c
823
if (WARN_ON(kho_scratch_overlap(pfn << PAGE_SHIFT, PAGE_SIZE << order)))
kernel/liveupdate/kexec_handover.c
867
if (WARN_ON(kho_scratch_overlap(start_pfn << PAGE_SHIFT,
kernel/liveupdate/kexec_handover.c
868
nr_pages << PAGE_SHIFT))) {
kernel/liveupdate/luo_file.c
120
((LUO_FILE_PGCNT << PAGE_SHIFT) / sizeof(struct luo_file_ser))
kernel/liveupdate/luo_file.c
183
size = LUO_FILE_PGCNT << PAGE_SHIFT;
kernel/liveupdate/luo_file.c
430
memset(file_set->files, 0, LUO_FILE_PGCNT << PAGE_SHIFT);
kernel/liveupdate/luo_flb.c
550
header_ser = kho_alloc_preserve(LUO_FLB_PGCNT << PAGE_SHIFT);
kernel/liveupdate/luo_flb.c
57
#define LUO_FLB_MAX (((LUO_FLB_PGCNT << PAGE_SHIFT) - \
kernel/liveupdate/luo_session.c
447
header_ser = kho_alloc_preserve(LUO_SESSION_PGCNT << PAGE_SHIFT);
kernel/liveupdate/luo_session.c
74
#define LUO_SESSION_MAX (((LUO_SESSION_PGCNT << PAGE_SHIFT) - \
kernel/module/strict_rwx.c
24
return set_memory((unsigned long)mod_mem->base, mod_mem->size >> PAGE_SHIFT);
kernel/power/power.h
51
#define PAGES_FOR_IO ((4096 * 1024) >> PAGE_SHIFT)
kernel/power/power.h
57
#define SPARE_PAGES ((1024 * 1024) >> PAGE_SHIFT)
kernel/power/snapshot.c
1021
(unsigned long long) start_pfn << PAGE_SHIFT,
kernel/power/snapshot.c
1022
((unsigned long long) end_pfn << PAGE_SHIFT) - 1);
kernel/power/snapshot.c
1093
(unsigned long long) region->start_pfn << PAGE_SHIFT,
kernel/power/snapshot.c
1094
((unsigned long long) region->end_pfn << PAGE_SHIFT)
kernel/power/snapshot.c
2187
info->size <<= PAGE_SHIFT;
kernel/power/snapshot.c
375
#define BM_BLOCK_SHIFT (PAGE_SHIFT + 3)
kernel/power/snapshot.c
425
#define BM_RTREE_LEVEL_SHIFT (PAGE_SHIFT - 2)
kernel/power/snapshot.c
427
#define BM_RTREE_LEVEL_SHIFT (PAGE_SHIFT - 3)
kernel/power/user.c
361
size <<= PAGE_SHIFT;
kernel/power/user.c
367
size <<= PAGE_SHIFT;
kernel/power/user.c
378
offset <<= PAGE_SHIFT;
kernel/relay.c
114
n_pages = *size >> PAGE_SHIFT;
kernel/relay.c
41
page = vmalloc_to_page(buf->start + (pgoff << PAGE_SHIFT));
kernel/resource.c
569
start = (u64) start_pfn << PAGE_SHIFT;
kernel/resource.c
570
end = ((u64)(start_pfn + nr_pages) << PAGE_SHIFT) - 1;
kernel/sched/fair.c
1984
enough_wmark = max(1UL * 1024 * 1024 * 1024 >> PAGE_SHIFT,
kernel/sched/fair.c
3520
pages <<= 20 - PAGE_SHIFT; /* MB in pages */
kernel/sched/fair.c
3643
end = ALIGN(start + (pages << PAGE_SHIFT), HPAGE_SIZE);
kernel/sched/fair.c
3656
pages -= (end - start) >> PAGE_SHIFT;
kernel/sched/fair.c
3657
virtpages -= (end - start) >> PAGE_SHIFT;
kernel/trace/trace_sched_switch.c
211
size = 1 << (order + PAGE_SHIFT);
lib/alloc_tag.c
419
(vm_module_tags->nr_pages << PAGE_SHIFT);
lib/alloc_tag.c
429
more_pages = ALIGN(new_end - phys_end, PAGE_SIZE) >> PAGE_SHIFT;
lib/alloc_tag.c
442
vmap_pages_range(phys_end, phys_end + (nr << PAGE_SHIFT), PAGE_KERNEL,
lib/alloc_tag.c
443
next_page, PAGE_SHIFT) < 0) {
lib/alloc_tag.c
673
get_vm_area_size(vm_module_tags) >> PAGE_SHIFT,
lib/buildid.c
51
r->folio = filemap_get_folio(r->file->f_mapping, file_off >> PAGE_SHIFT);
lib/debugobjects.c
34
#define ODEBUG_CHUNK_SHIFT PAGE_SHIFT
lib/iov_iter.c
1008
index = pos >> PAGE_SHIFT;
lib/iov_iter.c
1571
XA_STATE(xas, i->xarray, pos >> PAGE_SHIFT);
lib/iov_iter.c
354
v += (page - head) << PAGE_SHIFT;
lib/raid6/algos.c
192
(20 - PAGE_SHIFT + RAID6_TIME_JIFFIES_LG2));
lib/raid6/algos.c
212
(20 - PAGE_SHIFT + RAID6_TIME_JIFFIES_LG2));
lib/raid6/algos.c
231
(20 - PAGE_SHIFT + RAID6_TIME_JIFFIES_LG2 + 1));
lib/rhashtable.c
1192
const unsigned int shift = PAGE_SHIFT - ilog2(sizeof(void *));
lib/rhashtable.c
1232
const unsigned int shift = PAGE_SHIFT - ilog2(sizeof(void *));
lib/rhashtable.c
152
const unsigned int shift = PAGE_SHIFT - ilog2(sizeof(void *));
lib/rhashtable.c
77
const unsigned int shift = PAGE_SHIFT - ilog2(sizeof(void *));
lib/scatterlist.c
540
chunk_size = ((j - cur_page) << PAGE_SHIFT) - offset;
lib/scatterlist.c
623
nent = round_up(length, PAGE_SIZE << order) >> (PAGE_SHIFT + order);
lib/scatterlist.c
625
if (length > (nent << (PAGE_SHIFT + order)))
lib/scatterlist.c
741
return PAGE_ALIGN(sg->offset + sg->length) >> PAGE_SHIFT;
lib/scatterlist.c
765
return PAGE_ALIGN(sg->offset + sg_dma_len(sg)) >> PAGE_SHIFT;
lib/scatterlist.c
824
miter->piter.sg_pgoffset += miter->__offset >> PAGE_SHIFT;
lib/scatterlist.c
827
(miter->piter.sg_pgoffset << PAGE_SHIFT) -
lib/stackdepot.c
251
if (scale > PAGE_SHIFT)
lib/stackdepot.c
252
entries >>= (scale - PAGE_SHIFT);
lib/stackdepot.c
254
entries <<= (PAGE_SHIFT - scale);
lib/test_hmm.c
1014
xa_erase(&dmirror->pt, addr >> PAGE_SHIFT);
lib/test_hmm.c
1043
unsigned long size = cmd->npages << PAGE_SHIFT;
lib/test_hmm.c
1072
next = min(end, addr + (PTRS_PER_PTE << PAGE_SHIFT));
lib/test_hmm.c
1108
unsigned long size = cmd->npages << PAGE_SHIFT;
lib/test_hmm.c
1146
next = min(end, addr + (PTRS_PER_PTE << PAGE_SHIFT));
lib/test_hmm.c
1234
if (hmm_pfn_to_map_order(entry) + PAGE_SHIFT == PMD_SHIFT)
lib/test_hmm.c
1236
else if (hmm_pfn_to_map_order(entry) + PAGE_SHIFT == PUD_SHIFT)
lib/test_hmm.c
1314
n = (range->end - range->start) >> PAGE_SHIFT;
lib/test_hmm.c
1329
unsigned long size = cmd->npages << PAGE_SHIFT;
lib/test_hmm.c
1358
next = min(addr + (ARRAY_SIZE(pfns) << PAGE_SHIFT), end);
lib/test_hmm.c
1366
n = (range.end - range.start) >> PAGE_SHIFT;
lib/test_hmm.c
1382
unsigned long start_pfn = chunk->pagemap.range.start >> PAGE_SHIFT;
lib/test_hmm.c
1383
unsigned long end_pfn = chunk->pagemap.range.end >> PAGE_SHIFT;
lib/test_hmm.c
1496
if (cmd.addr >= (cmd.addr + (cmd.npages << PAGE_SHIFT)))
lib/test_hmm.c
1525
cmd.addr + (cmd.npages << PAGE_SHIFT));
lib/test_hmm.c
1655
nr = (args.end - args.start) >> PAGE_SHIFT;
lib/test_hmm.c
215
for (pfn = (range->start >> PAGE_SHIFT);
lib/test_hmm.c
216
pfn < (range->end >> PAGE_SHIFT);
lib/test_hmm.c
255
xa_for_each_range(&dmirror->pt, pfn, entry, start >> PAGE_SHIFT,
lib/test_hmm.c
256
end >> PAGE_SHIFT)
lib/test_hmm.c
352
range.end = min(addr + (ARRAY_SIZE(pfns) << PAGE_SHIFT), end);
lib/test_hmm.c
371
for (pfn = start >> PAGE_SHIFT; pfn < (end >> PAGE_SHIFT); pfn++) {
lib/test_hmm.c
393
unsigned long size = cmd->npages << PAGE_SHIFT;
lib/test_hmm.c
412
start = cmd->addr + (bounce.cpages << PAGE_SHIFT);
lib/test_hmm.c
437
for (pfn = start >> PAGE_SHIFT; pfn < (end >> PAGE_SHIFT); pfn++) {
lib/test_hmm.c
459
unsigned long size = cmd->npages << PAGE_SHIFT;
lib/test_hmm.c
483
start = cmd->addr + (bounce.cpages << PAGE_SHIFT);
lib/test_hmm.c
564
pfn_first = devmem->pagemap.range.start >> PAGE_SHIFT;
lib/test_hmm.c
565
pfn_last = pfn_first + (range_len(&devmem->pagemap.range) >> PAGE_SHIFT);
lib/test_hmm.c
802
for (pfn = start >> PAGE_SHIFT; pfn < (end >> PAGE_SHIFT); pfn++) {
lib/test_hmm.c
822
entry = xa_store(&dmirror->pt, addr >> PAGE_SHIFT, entry, GFP_ATOMIC);
lib/test_hmm.c
840
const unsigned long start_pfn = start >> PAGE_SHIFT;
lib/test_hmm.c
841
const unsigned long end_pfn = end >> PAGE_SHIFT;
lib/test_hmm.c
889
unsigned long size = cmd->npages << PAGE_SHIFT;
lib/test_kho.c
212
max_nr = max_mem >> PAGE_SHIFT;
lib/test_vmalloc.c
333
align = 1 << get_random_u32_inclusive(1, PAGE_SHIFT - 1);
lib/test_xarray.c
825
index = pos >> PAGE_SHIFT;
mm/backing-dev.c
272
bdi->ra_pages = read_ahead_kb >> (PAGE_SHIFT - 10);
mm/bootmem_info.c
61
mapsize = PAGE_ALIGN(mem_section_usage_size()) >> PAGE_SHIFT;
mm/bootmem_info.c
73
nr_pages = PAGE_ALIGN(sizeof(struct pglist_data)) >> PAGE_SHIFT;
mm/cma.c
1126
size >>= PAGE_SHIFT;
mm/cma.c
241
cma->available_count = cma->count = size >> PAGE_SHIFT;
mm/cma.c
472
if (!IS_ALIGNED(size >> PAGE_SHIFT, 1 << order_per_bit))
mm/cma.c
48
return cma->count << PAGE_SHIFT;
mm/cma.c
680
cmrp->count = size >> PAGE_SHIFT;
mm/compaction.c
80
#define COMPACTION_HPAGE_ORDER (PMD_SHIFT - PAGE_SHIFT)
mm/damon/vaddr.c
607
ilx += (addr - vma->vm_start) >> (PAGE_SHIFT + order);
mm/early_ioremap.c
146
nrpages = size >> PAGE_SHIFT;
mm/early_ioremap.c
203
nrpages = PAGE_ALIGN(offset + size) >> PAGE_SHIFT;
mm/early_ioremap.c
254
#define MAX_MAP_CHUNK (NR_FIX_BTMAPS << PAGE_SHIFT)
mm/execmem.c
100
#define PENDING_FREE_MASK (1 << (PAGE_SHIFT - 1))
mm/execmem.c
139
unsigned int nr = PAGE_ALIGN(size) >> PAGE_SHIFT;
mm/execmem.c
152
unsigned int nr = PAGE_ALIGN(size) >> PAGE_SHIFT;
mm/execmem.c
70
area = __get_vm_area_node(size, range->alignment, PAGE_SHIFT, VM_ALLOC,
mm/execmem.c
74
area = __get_vm_area_node(size, range->alignment, PAGE_SHIFT, VM_ALLOC,
mm/fadvise.c
121
start_index = (offset+(PAGE_SIZE-1)) >> PAGE_SHIFT;
mm/fadvise.c
122
end_index = (endbyte >> PAGE_SHIFT);
mm/fadvise.c
98
start_index = offset >> PAGE_SHIFT;
mm/fadvise.c
99
end_index = endbyte >> PAGE_SHIFT;
mm/filemap.c
2631
index = (iocb->ki_pos >> (PAGE_SHIFT + min_order)) << min_order;
mm/filemap.c
2672
pgoff_t index = iocb->ki_pos >> PAGE_SHIFT;
mm/filemap.c
2680
mapping_min_folio_nrbytes(mapping)) >> PAGE_SHIFT;
mm/filemap.c
2920
return invalidate_inode_pages2_range(mapping, pos >> PAGE_SHIFT,
mm/filemap.c
2921
end >> PAGE_SHIFT);
mm/filemap.c
3207
XA_STATE(xas, &mapping->i_pages, start >> PAGE_SHIFT);
mm/filemap.c
3208
pgoff_t max = (end - 1) >> PAGE_SHIFT;
mm/filemap.c
3217
loff_t pos = (u64)xas.xa_index << PAGE_SHIFT;
mm/filemap.c
3235
xas_set(&xas, pos >> PAGE_SHIFT);
mm/filemap.c
3912
addr = vma->vm_start + ((start_pgoff - vma->vm_pgoff) << PAGE_SHIFT);
mm/filemap.c
3924
addr += (xas.xa_index - last_pgoff) << PAGE_SHIFT;
mm/filemap.c
4235
iocb->ki_pos >> PAGE_SHIFT,
mm/filemap.c
4236
(iocb->ki_pos + count - 1) >> PAGE_SHIFT))
mm/filemap.c
4526
pgoff_t first = start >> PAGE_SHIFT;
mm/filemap.c
4527
pgoff_t last = end >> PAGE_SHIFT;
mm/filemap.c
4739
first_index = csr.off >> PAGE_SHIFT;
mm/filemap.c
4741
csr.len == 0 ? ULONG_MAX : (csr.off + csr.len - 1) >> PAGE_SHIFT;
mm/filemap.c
483
XA_STATE(xas, &mapping->i_pages, start_byte >> PAGE_SHIFT);
mm/filemap.c
484
pgoff_t max = end_byte >> PAGE_SHIFT;
mm/filemap.c
513
pgoff_t index = start_byte >> PAGE_SHIFT;
mm/filemap.c
514
pgoff_t end = end_byte >> PAGE_SHIFT;
mm/filemap.c
640
XA_STATE(xas, &mapping->i_pages, start_byte >> PAGE_SHIFT);
mm/filemap.c
641
pgoff_t max = end_byte >> PAGE_SHIFT;
mm/gup.c
1459
page_increm = 1 + (~(start >> PAGE_SHIFT) & page_mask);
mm/gup.c
1724
start += ret << PAGE_SHIFT;
mm/gup.c
2938
refs = (end - addr) >> PAGE_SHIFT;
mm/gup.c
2939
page = pmd_page(orig) + ((addr & ~PMD_MASK) >> PAGE_SHIFT);
mm/gup.c
2981
refs = (end - addr) >> PAGE_SHIFT;
mm/gup.c
2982
page = pud_page(orig) + ((addr & ~PUD_MASK) >> PAGE_SHIFT);
mm/gup.c
3196
len = nr_pages << PAGE_SHIFT;
mm/gup.c
3207
start += nr_pinned << PAGE_SHIFT;
mm/gup.c
3441
unsigned int i, pgshift = PAGE_SHIFT;
mm/gup.c
668
pfn += (addr & ~PUD_MASK) >> PAGE_SHIFT;
mm/gup.c
740
page += (addr & ~HPAGE_PMD_MASK) >> PAGE_SHIFT;
mm/hmm.c
166
i = (addr - range->start) >> PAGE_SHIFT;
mm/hmm.c
167
npages = (end - addr) >> PAGE_SHIFT;
mm/hmm.c
194
hmm_pfn_flags_order(PMD_SHIFT - PAGE_SHIFT);
mm/hmm.c
207
npages = (end - addr) >> PAGE_SHIFT;
mm/hmm.c
214
pfn = pmd_pfn(pmd) + ((addr & ~PMD_MASK) >> PAGE_SHIFT);
mm/hmm.c
341
unsigned long npages = (end - start) >> PAGE_SHIFT;
mm/hmm.c
350
hmm_pfn_flags_order(PMD_SHIFT - PAGE_SHIFT);
mm/hmm.c
388
unsigned long npages = (end - start) >> PAGE_SHIFT;
mm/hmm.c
404
&range->hmm_pfns[(start - range->start) >> PAGE_SHIFT];
mm/hmm.c
405
unsigned long npages = (end - start) >> PAGE_SHIFT;
mm/hmm.c
481
hmm_pfn_flags_order(PUD_SHIFT - PAGE_SHIFT);
mm/hmm.c
511
i = (addr - range->start) >> PAGE_SHIFT;
mm/hmm.c
512
npages = (end - addr) >> PAGE_SHIFT;
mm/hmm.c
523
pfn = pud_pfn(pud) + ((addr & ~PUD_MASK) >> PAGE_SHIFT);
mm/hmm.c
53
unsigned long i = (addr - range->start) >> PAGE_SHIFT;
mm/hmm.c
560
i = (start - range->start) >> PAGE_SHIFT;
mm/hmm.c
584
pfn = pte_pfn(entry) + ((start & ~hmask) >> PAGE_SHIFT);
mm/hmm.c
621
((start - range->start) >> PAGE_SHIFT),
mm/hmm.c
622
(end - start) >> PAGE_SHIFT, 0))
mm/huge_memory.c
1270
off >> PAGE_SHIFT, flags, vm_flags);
mm/huge_memory.c
1300
loff_t off = (loff_t)pgoff << PAGE_SHIFT;
mm/hugetlb.c
3260
start_pfn = virt_to_phys(m) >> PAGE_SHIFT;
mm/hugetlb.c
4393
hugetlb_add_hstate(ilog2(size) - PAGE_SHIFT);
mm/hugetlb.c
4427
hugetlb_add_hstate(ilog2(size) - PAGE_SHIFT);
mm/hugetlb.c
5372
pgoff = ((address - vma->vm_start) >> PAGE_SHIFT) +
mm/hugetlb.c
6763
unsigned long saddr = ((idx - svma->vm_pgoff) << PAGE_SHIFT) +
mm/hugetlb.c
6848
pgoff_t idx = ((addr - vma->vm_start) >> PAGE_SHIFT) +
mm/hugetlb_vmemmap.c
340
unsigned long nr_pages = (end - start) >> PAGE_SHIFT;
mm/hugetlb_vmemmap.c
740
pmd_vmemmap_size = (PMD_SIZE / (sizeof(struct page))) << PAGE_SHIFT;
mm/hugetlb_vmemmap.c
768
psize = nr_pages << PAGE_SHIFT;
mm/internal.h
1167
addr = vma->vm_start + ((pgoff - vma->vm_pgoff) << PAGE_SHIFT);
mm/internal.h
1243
((pgoff - vma->vm_pgoff) << PAGE_SHIFT);
mm/internal.h
1271
address = vma->vm_start + ((pgoff - vma->vm_pgoff) << PAGE_SHIFT);
mm/internal.h
651
#define K(x) ((x) << (PAGE_SHIFT-10))
mm/kasan/kasan.h
140
#define KASAN_MEMORY_PER_SHADOW_PAGE (KASAN_GRANULE_SIZE << PAGE_SHIFT)
mm/kasan/kasan_test_c.c
348
size_t size = (1UL << (PAGE_SHIFT + order));
mm/kasan/quarantine.c
266
total_size = (totalram_pages() << PAGE_SHIFT) /
mm/kasan/shadow.c
223
shadow_size = nr_shadow_pages << PAGE_SHIFT;
mm/khugepaged.c
1627
unsigned int max_nr_batch_ptes = (end - addr) >> PAGE_SHIFT;
mm/khugepaged.c
1788
addr = vma->vm_start + ((pgoff - vma->vm_pgoff) << PAGE_SHIFT);
mm/khugepaged.c
2455
const loff_t lstart = (loff_t)pgoff << PAGE_SHIFT;
mm/khugepaged.c
2737
recommended_min <<= (PAGE_SHIFT-10);
mm/khugepaged.c
721
unsigned int max_nr_ptes = (end - address) >> PAGE_SHIFT;
mm/kmsan/hooks.c
173
PAGE_SHIFT);
mm/kmsan/hooks.c
182
PAGE_SHIFT);
mm/ksm.c
2530
page += ((addr & (PMD_SIZE - 1)) >> PAGE_SHIFT);
mm/madvise.c
1026
+ ((loff_t)vma->vm_pgoff << PAGE_SHIFT);
mm/madvise.c
256
((xas.xa_index - vma->vm_pgoff) << PAGE_SHIFT);
mm/madvise.c
321
+ ((loff_t)vma->vm_pgoff << PAGE_SHIFT);
mm/mapping_dirty_helpers.c
97
pgoff_t pgoff = ((addr - walk->vma->vm_start) >> PAGE_SHIFT) +
mm/memblock.c
2455
unsigned int nr_pages = map->size >> PAGE_SHIFT;
mm/memblock.c
2468
unsigned int nr_pages = map->size >> PAGE_SHIFT;
mm/memblock.c
783
if ((nr_pages << PAGE_SHIFT) > threshold_bytes) {
mm/memblock.c
786
(nr_pages << PAGE_SHIFT) / SZ_1M, mem_size_mb);
mm/memcontrol.c
169
nr_pages = nr_bytes >> PAGE_SHIFT;
mm/memcontrol.c
3058
unsigned int nr_pages = stock->nr_bytes >> PAGE_SHIFT;
mm/memcontrol.c
3137
nr_pages = nr_bytes >> PAGE_SHIFT;
mm/memcontrol.c
3155
nr_pages = stock->nr_bytes >> PAGE_SHIFT;
mm/memcontrol.c
3180
ret = obj_cgroup_charge_pages(objcg, gfp, charge_size >> PAGE_SHIFT);
mm/memory-failure.c
1701
loff_t start = ((loff_t)index << PAGE_SHIFT) & ~(size - 1);
mm/memory-failure.c
2225
tk->size_shift = PAGE_SHIFT;
mm/memory-failure.c
370
ret = PAGE_SHIFT;
mm/memory-failure.c
707
mask = ~((1UL << (shift - PAGE_SHIFT)) - 1);
mm/memory-failure.c
711
hwpoison_vaddr = addr + ((poisoned_pfn - pfn) << PAGE_SHIFT);
mm/memory-failure.c
728
hwpoison_vaddr = addr + ((hwp->pfn - pfn) << PAGE_SHIFT);
mm/memory-failure.c
729
set_to_kill(&hwp->tk, hwpoison_vaddr, PAGE_SHIFT);
mm/memory-failure.c
763
ret = check_hwpoisoned_entry(ptep_get(ptep), addr, PAGE_SHIFT,
mm/memory.c
2929
pfn -= addr >> PAGE_SHIFT;
mm/memory.c
2937
pfn + (addr >> PAGE_SHIFT), prot);
mm/memory.c
2952
pfn -= addr >> PAGE_SHIFT;
mm/memory.c
2959
pfn + (addr >> PAGE_SHIFT), prot);
mm/memory.c
2974
pfn -= addr >> PAGE_SHIFT;
mm/memory.c
2981
pfn + (addr >> PAGE_SHIFT), prot);
mm/memory.c
3022
pfn -= addr >> PAGE_SHIFT;
mm/memory.c
3028
pfn + (addr >> PAGE_SHIFT), prot);
mm/memory.c
3225
pfn = start_phys >> PAGE_SHIFT;
mm/memory.c
3226
pages = (size + ~PAGE_MASK) >> PAGE_SHIFT;
mm/memory.c
3237
if ((vm_len >> PAGE_SHIFT) > pages)
mm/memory.c
4336
start = vma->vm_start + ((start_idx - vma->vm_pgoff) << PAGE_SHIFT);
mm/memory.c
4337
size = (end_idx - start_idx) << PAGE_SHIFT;
mm/memory.c
4430
pgoff_t hba = (pgoff_t)(holebegin) >> PAGE_SHIFT;
mm/memory.c
4431
pgoff_t hlen = ((pgoff_t)(holelen) + PAGE_SIZE - 1) >> PAGE_SHIFT;
mm/memory.c
4436
(holebegin + holelen + PAGE_SIZE - 1) >> PAGE_SHIFT;
mm/memory.c
4666
if ((addr >> PAGE_SHIFT) % nr == swp_offset % nr)
mm/memory.c
5768
65536 >> PAGE_SHIFT;
mm/memory.c
5773
*val = fault_around_pages << PAGE_SHIFT;
mm/memory.c
5791
fault_around_pages = rounddown_pow_of_two(val) >> PAGE_SHIFT;
mm/memory.c
6109
unsigned long addr_start = addr - (nr << PAGE_SHIFT);
mm/memory.c
6117
start_ptep = vmf->pte - ((addr - start) >> PAGE_SHIFT);
mm/memory.c
6829
args->pfn = pfn_base + ((args->address & ~addr_mask) >> PAGE_SHIFT);
mm/memory.c
7013
phys_addr = (resource_size_t)args.pfn << PAGE_SHIFT;
mm/memory.c
7028
(phys_addr != (args.pfn << PAGE_SHIFT)) ||
mm/memory.c
712
unsigned long off = (addr - vma->vm_start) >> PAGE_SHIFT;
mm/memory.c
7309
ip += vma->vm_pgoff << PAGE_SHIFT;
mm/memory.c
7342
~(((unsigned long)nr_pages << PAGE_SHIFT) - 1);
mm/memory_hotplug.c
1242
(unsigned long long) pfn << PAGE_SHIFT,
mm/memory_hotplug.c
1243
(((unsigned long long) pfn + nr_pages) << PAGE_SHIFT) - 1);
mm/memory_hotplug.c
2110
(unsigned long long) start_pfn << PAGE_SHIFT,
mm/memory_hotplug.c
2111
((unsigned long long) end_pfn << PAGE_SHIFT) - 1,
mm/mempolicy.c
2052
*ilx += (addr - vma->vm_start) >> (PAGE_SHIFT + order);
mm/mempolicy.c
3237
sn = sp_alloc(0, MAX_LFS_FILESIZE >> PAGE_SHIFT, npol);
mm/mempolicy.c
3374
if ((total_pages << PAGE_SHIFT) >= (16 << 20))
mm/mempolicy.c
709
max_nr = (end - addr) >> PAGE_SHIFT;
mm/memremap.c
77
return (range->start + range_len(range)) >> PAGE_SHIFT;
mm/migrate_device.c
526
migrate->end = migrate->start + (migrate->npages << PAGE_SHIFT);
mm/migrate_device.c
737
long nr_pages = (args->end - args->start) >> PAGE_SHIFT;
mm/mincore.c
138
unsigned long nr = (end - addr) >> PAGE_SHIFT;
mm/mincore.c
170
int nr = (end - addr) >> PAGE_SHIFT;
mm/mincore.c
197
unsigned int max_nr = (end - addr) >> PAGE_SHIFT;
mm/mincore.c
256
end = min(vma->vm_end, addr + (pages << PAGE_SHIFT));
mm/mincore.c
265
return (end - addr) >> PAGE_SHIFT;
mm/mincore.c
310
pages = len >> PAGE_SHIFT;
mm/mincore.c
338
start += retval << PAGE_SHIFT;
mm/mlock.c
310
unsigned int count = (end - addr) >> PAGE_SHIFT;
mm/mlock.c
403
addr += (step - 1) << PAGE_SHIFT;
mm/mlock.c
495
nr_pages = (end - start) >> PAGE_SHIFT;
mm/mlock.c
603
return count >> PAGE_SHIFT;
mm/mlock.c
633
lock_limit >>= PAGE_SHIFT;
mm/mlock.c
634
locked = len >> PAGE_SHIFT;
mm/mlock.c
764
lock_limit >>= PAGE_SHIFT;
mm/mlock.c
803
locked = (size + PAGE_SIZE - 1) >> PAGE_SHIFT;
mm/mlock.c
806
lock_limit >>= PAGE_SHIFT;
mm/mlock.c
828
dec_rlimit_ucounts(ucounts, UCOUNT_RLIMIT_MEMLOCK, (size + PAGE_SIZE - 1) >> PAGE_SHIFT);
mm/mm_init.c
1530
if (HPAGE_SHIFT > PAGE_SHIFT && HUGETLB_PAGE_ORDER < order)
mm/mm_init.c
1742
(u64)start_pfn << PAGE_SHIFT,
mm/mm_init.c
1743
end_pfn ? ((u64)end_pfn << PAGE_SHIFT) - 1 : 0);
mm/mm_init.c
1876
<< PAGE_SHIFT,
mm/mm_init.c
1878
<< PAGE_SHIFT) - 1);
mm/mm_init.c
1886
(u64)zone_movable_pfn[i] << PAGE_SHIFT);
mm/mm_init.c
1897
(u64)start_pfn << PAGE_SHIFT,
mm/mm_init.c
1898
((u64)end_pfn << PAGE_SHIFT) - 1);
mm/mm_init.c
2381
#define ADAPT_SCALE_NPAGES (ADAPT_SCALE_BASE >> PAGE_SHIFT)
mm/mm_init.c
2427
if (scale > PAGE_SHIFT)
mm/mm_init.c
2428
numentries >>= (scale - PAGE_SHIFT);
mm/mm_init.c
2430
numentries <<= (PAGE_SHIFT - scale);
mm/mm_init.c
2439
max = ((unsigned long long)nr_all_pages << PAGE_SHIFT) >> 4;
mm/mm_init.c
282
WARN_ON((coremem >> PAGE_SHIFT) > ULONG_MAX);
mm/mm_init.c
284
*core = coremem >> PAGE_SHIFT;
mm/mm_init.c
606
set_page_address(page, __va(pfn << PAGE_SHIFT));
mm/mmap.c
1109
if (pgoff + (size >> PAGE_SHIFT) < pgoff)
mm/mmap.c
1336
if (mm->total_vm + npages > rlimit(RLIMIT_AS) >> PAGE_SHIFT)
mm/mmap.c
1340
mm->data_vm + npages > rlimit(RLIMIT_DATA) >> PAGE_SHIFT) {
mm/mmap.c
1343
mm->data_vm + npages <= rlimit_max(RLIMIT_DATA) >> PAGE_SHIFT)
mm/mmap.c
1348
(mm->data_vm + npages) << PAGE_SHIFT,
mm/mmap.c
1477
vm_stat_account(mm, vma->vm_flags, len >> PAGE_SHIFT);
mm/mmap.c
238
locked_pages = bytes >> PAGE_SHIFT;
mm/mmap.c
242
limit_pages >>= PAGE_SHIFT;
mm/mmap.c
274
if (pgoff > maxsize >> PAGE_SHIFT)
mm/mmap.c
375
if ((pgoff + (len >> PAGE_SHIFT)) < pgoff)
mm/mmap.c
539
pgoff = addr >> PAGE_SHIFT;
mm/mmap.c
640
a.offset >> PAGE_SHIFT);
mm/mmap.c
940
unsigned long stack_guard_gap = 256UL<<PAGE_SHIFT;
mm/mmap.c
949
stack_guard_gap = val << PAGE_SHIFT;
mm/mprotect.c
242
int max_nr_ptes = (end - addr) >> PAGE_SHIFT;
mm/mprotect.c
702
long nrpages = (end - start) >> PAGE_SHIFT;
mm/mremap.c
1022
unsigned long pages = bytes >> PAGE_SHIFT;
mm/mremap.c
1201
vm_acct_memory(len >> PAGE_SHIFT);
mm/mremap.c
1259
unsigned long internal_pgoff = internal_offset >> PAGE_SHIFT;
mm/mremap.c
1476
unsigned long pages = vrm->old_len >> PAGE_SHIFT;
mm/mremap.c
1805
pgoff = (addr - vma->vm_start) >> PAGE_SHIFT;
mm/mremap.c
1807
if (pgoff + (new_len >> PAGE_SHIFT) < pgoff)
mm/mremap.c
1816
if (!may_expand_vm(mm, &vma->flags, vrm->delta >> PAGE_SHIFT))
mm/mremap.c
270
max_nr_ptes = (old_end - old_addr) >> PAGE_SHIFT;
mm/mremap.c
951
pgoff_t internal_pgoff = (vrm->addr - vma->vm_start) >> PAGE_SHIFT;
mm/mremap.c
988
charged = vrm->new_len >> PAGE_SHIFT;
mm/mremap.c
990
charged = vrm->delta >> PAGE_SHIFT;
mm/msync.c
89
((loff_t)vma->vm_pgoff << PAGE_SHIFT);
mm/nommu.c
1082
pglen = (len + PAGE_SIZE - 1) >> PAGE_SHIFT;
mm/nommu.c
1100
rpglen = (rpglen + PAGE_SIZE - 1) >> PAGE_SHIFT;
mm/nommu.c
1119
start += (pgoff - pregion->vm_pgoff) << PAGE_SHIFT;
mm/nommu.c
1193
current->mm->total_vm += len >> PAGE_SHIFT;
mm/nommu.c
1298
a.offset >> PAGE_SHIFT);
mm/nommu.c
1335
npages = (addr - vma->vm_start) >> PAGE_SHIFT;
mm/nommu.c
1593
if (addr != (pfn << PAGE_SHIFT))
mm/nommu.c
1603
unsigned long pfn = start >> PAGE_SHIFT;
mm/nommu.c
1619
vma->vm_start = (unsigned long)(addr + (pgoff << PAGE_SHIFT));
mm/nommu.c
1812
low = newsize >> PAGE_SHIFT;
mm/nommu.c
1813
high = (size + PAGE_SIZE - 1) >> PAGE_SHIFT;
mm/nommu.c
1841
r_top = (region->vm_pgoff << PAGE_SHIFT) + r_size;
mm/nommu.c
718
if ((pgoff + (rlen >> PAGE_SHIFT)) < pgoff)
mm/nommu.c
953
point = len >> PAGE_SHIFT;
mm/nommu.c
959
base = alloc_pages_exact(total << PAGE_SHIFT, GFP_KERNEL);
mm/nommu.c
969
region->vm_top = region->vm_start + (total << PAGE_SHIFT);
mm/nommu.c
979
fpos <<= PAGE_SHIFT;
mm/numa.c
27
tnid = early_pfn_to_nid(nd_pa >> PAGE_SHIFT);
mm/numa_emulation.c
109
size = PFN_PHYS((unsigned long)(size >> PAGE_SHIFT) / nr_nodes);
mm/page-writeback.c
2065
ratelimit = min(ratelimit, 32 >> (PAGE_SHIFT - 10));
mm/page-writeback.c
2419
return wbc->range_end >> PAGE_SHIFT;
mm/page-writeback.c
2492
wbc->index = wbc->range_start >> PAGE_SHIFT;
mm/page-writeback.c
55
#define DIRTY_POLL_THRESH (128 >> (PAGE_SHIFT - 10))
mm/page-writeback.c
764
unsigned long pages = min_bytes >> PAGE_SHIFT;
mm/page-writeback.c
785
unsigned long pages = max_bytes >> PAGE_SHIFT;
mm/page_alloc.c
6139
if ((cci->per_cpu_data_slice_size >> PAGE_SHIFT) > 3 * pcp->batch)
mm/page_alloc.c
6412
unsigned long pages_min = min_free_kbytes >> (PAGE_SHIFT - 10);
mm/page_io.c
140
first_block >>= (PAGE_SHIFT - blkbits);
mm/page_table_check.c
155
page_table_check_clear(pte_pfn(pte), PAGE_SIZE >> PAGE_SHIFT);
mm/page_table_check.c
167
page_table_check_clear(pmd_pfn(pmd), PMD_SIZE >> PAGE_SHIFT);
mm/page_table_check.c
179
page_table_check_clear(pud_pfn(pud), PUD_SIZE >> PAGE_SHIFT);
mm/page_table_check.c
234
unsigned long stride = PMD_SIZE >> PAGE_SHIFT;
mm/page_table_check.c
252
unsigned long stride = PUD_SIZE >> PAGE_SHIFT;
mm/pagewalk.c
1021
fw->page = page + ((addr & (entry_size - 1)) >> PAGE_SHIFT);
mm/pagewalk.c
823
start_addr = ((cba - vba) << PAGE_SHIFT) + vma->vm_start;
mm/pagewalk.c
824
end_addr = ((cea - vba) << PAGE_SHIFT) + vma->vm_start;
mm/percpu-km.c
117
nr_pages = (ai->groups[0].nr_units * ai->unit_size) >> PAGE_SHIFT;
mm/percpu-km.c
55
const int nr_pages = pcpu_group_sizes[0] >> PAGE_SHIFT;
mm/percpu-km.c
89
const int nr_pages = pcpu_group_sizes[0] >> PAGE_SHIFT;
mm/percpu-vm.c
137
vunmap_range_noflush(addr, addr + (nr_pages << PAGE_SHIFT));
mm/percpu-vm.c
196
return vmap_pages_range_noflush(addr, addr + (nr_pages << PAGE_SHIFT),
mm/percpu-vm.c
197
PAGE_KERNEL, pages, PAGE_SHIFT, GFP_KERNEL);
mm/percpu.c
1367
BITS_TO_LONGS(region_size >> PAGE_SHIFT));
mm/percpu.c
1376
chunk->nr_pages = region_size >> PAGE_SHIFT;
mm/percpu.c
2668
pcpu_unit_pages = ai->unit_size >> PAGE_SHIFT;
mm/percpu.c
2669
pcpu_unit_size = pcpu_unit_pages << PAGE_SHIFT;
mm/percpu.c
272
return pcpu_unit_offsets[cpu] + (page_idx << PAGE_SHIFT);
mm/percpu.c
3219
unit_pages = ai->unit_size >> PAGE_SHIFT;
mm/percpu.c
3255
pcpu_populate_pte(unit_addr + (i << PAGE_SHIFT));
mm/readahead.c
326
end_index = (isize - 1) >> PAGE_SHIFT;
mm/readahead.c
474
pgoff_t limit = (i_size_read(mapping->host) - 1) >> PAGE_SHIFT;
mm/readahead.c
587
prev_index = (unsigned long long)ra->prev_pos >> PAGE_SHIFT;
mm/rmap.c
1694
address + (nr << PAGE_SHIFT) > vma->vm_end);
mm/rmap.c
1959
max_nr = (end_addr - addr) >> PAGE_SHIFT;
mm/rmap.c
970
const unsigned int max_nr = (end_addr - address) >> PAGE_SHIFT;
mm/secretmem.c
61
if (((loff_t)vmf->pgoff << PAGE_SHIFT) >= i_size_read(inode))
mm/shmem.c
1020
return swapped << PAGE_SHIFT;
mm/shmem.c
1049
return swapped << PAGE_SHIFT;
mm/shmem.c
1113
pgoff_t start = (lstart + PAGE_SIZE - 1) >> PAGE_SHIFT;
mm/shmem.c
1114
pgoff_t end = (lend + 1) >> PAGE_SHIFT;
mm/shmem.c
1162
same_folio = (lstart >> PAGE_SHIFT) == (lend >> PAGE_SHIFT);
mm/shmem.c
1163
folio = shmem_get_partial_folio(inode, lstart >> PAGE_SHIFT);
mm/shmem.c
1178
folio = shmem_get_partial_folio(inode, lend >> PAGE_SHIFT);
mm/shmem.c
2478
if (index > (MAX_LFS_FILESIZE >> PAGE_SHIFT))
mm/shmem.c
2482
((loff_t)index << PAGE_SHIFT) >= i_size_read(inode))
mm/shmem.c
2621
((loff_t)index << PAGE_SHIFT) >= i_size_read(inode)) {
mm/shmem.c
2856
offset = (pgoff << PAGE_SHIFT) & (hpage_size - 1);
mm/shmem.c
2904
index = ((addr - vma->vm_start) >> PAGE_SHIFT) + vma->vm_pgoff;
mm/shmem.c
3306
pgoff_t index = pos >> PAGE_SHIFT;
mm/shmem.c
3383
index = iocb->ki_pos >> PAGE_SHIFT;
mm/shmem.c
3574
index = *ppos >> PAGE_SHIFT;
mm/shmem.c
3714
shmem_falloc.start = (u64)unmap_start >> PAGE_SHIFT;
mm/shmem.c
3715
shmem_falloc.next = (unmap_end + 1) >> PAGE_SHIFT;
mm/shmem.c
3745
start = offset >> PAGE_SHIFT;
mm/shmem.c
3746
end = (offset + len + PAGE_SIZE - 1) >> PAGE_SHIFT;
mm/shmem.c
3793
(loff_t)start << PAGE_SHIFT,
mm/shmem.c
3794
((loff_t)index << PAGE_SHIFT) - 1, true);
mm/shmem.c
4627
size <<= PAGE_SHIFT;
mm/shmem.c
5101
sb->s_blocksize_bits = PAGE_SHIFT;
mm/shmem.c
617
if (i_size >> PAGE_SHIFT >= aligned_index)
mm/shmem.c
89
#define VM_ACCT(size) (PAGE_ALIGN(size) >> PAGE_SHIFT)
mm/sparse-vmemmap.c
128
nr_pfns = size >> PAGE_SHIFT;
mm/sparse.c
132
unsigned long max_sparsemem_pfn = (DIRECT_MAP_PHYSMEM_END + 1) >> PAGE_SHIFT;
mm/swap.h
243
return ((loff_t)swp_offset(entry)) << PAGE_SHIFT;
mm/swap_state.c
48
#define SWAP_RA_WIN_SHIFT (PAGE_SHIFT / 2)
mm/swap_state.c
778
left = faddr - (win << PAGE_SHIFT) + PAGE_SIZE;
mm/swap_state.c
780
left = faddr - (((win - 1) / 2) << PAGE_SHIFT);
mm/swap_state.c
781
right = left + (win << PAGE_SHIFT);
mm/swap_table.h
53
#define SWAP_CACHE_PFN_BITS (MAX_POSSIBLE_PHYSMEM_BITS - PAGE_SHIFT)
mm/swap_table.h
55
#define SWAP_CACHE_PFN_BITS (MAX_PHYSMEM_BITS - PAGE_SHIFT)
mm/swap_table.h
57
#define SWAP_CACHE_PFN_BITS (BITS_PER_LONG - PAGE_SHIFT)
mm/swapfile.c
288
start_block = (se->start_block + 1) << (PAGE_SHIFT - 9);
mm/swapfile.c
289
nr_blocks = ((sector_t)se->nr_pages - 1) << (PAGE_SHIFT - 9);
mm/swapfile.c
299
start_block = se->start_block << (PAGE_SHIFT - 9);
mm/swapfile.c
300
nr_blocks = (sector_t)se->nr_pages << (PAGE_SHIFT - 9);
mm/swapfile.c
3339
swapfilepages = i_size_read(inode) >> PAGE_SHIFT;
mm/swapfile.c
342
return sector << (PAGE_SHIFT - 9);
mm/swapfile.c
364
start_block <<= PAGE_SHIFT - 9;
mm/swapfile.c
365
nr_blocks <<= PAGE_SHIFT - 9;
mm/truncate.c
387
start = (lstart + PAGE_SIZE - 1) >> PAGE_SHIFT;
mm/truncate.c
396
end = (lend + 1) >> PAGE_SHIFT;
mm/truncate.c
412
same_folio = (lstart >> PAGE_SHIFT) == (lend >> PAGE_SHIFT);
mm/truncate.c
413
folio = __filemap_get_folio(mapping, lstart >> PAGE_SHIFT, FGP_LOCK, 0);
mm/truncate.c
427
folio = __filemap_get_folio(mapping, lend >> PAGE_SHIFT,
mm/util.c
1421
len = vma_pages(vma) << PAGE_SHIFT;
mm/util.c
338
#define STACK_RND_MASK (0x7ff >> (PAGE_SHIFT - 12)) /* 8MB of VA */
mm/util.c
348
random_variable <<= PAGE_SHIFT;
mm/util.c
381
range >>= PAGE_SHIFT;
mm/util.c
386
return start + (get_random_long() % range << PAGE_SHIFT);
mm/util.c
410
return rnd << PAGE_SHIFT;
mm/util.c
450
pad += (STACK_RND_MASK << PAGE_SHIFT);
mm/util.c
517
limit = task_rlimit(task, RLIMIT_MEMLOCK) >> PAGE_SHIFT;
mm/util.c
529
(void *)_RET_IP_, (inc) ? '+' : '-', pages << PAGE_SHIFT,
mm/util.c
530
locked_vm << PAGE_SHIFT, task_rlimit(task, RLIMIT_MEMLOCK),
mm/util.c
569
loff_t off = (loff_t)pgoff << PAGE_SHIFT;
mm/util.c
617
return vm_mmap_pgoff(file, addr, len, prot, flag, offset >> PAGE_SHIFT);
mm/util.c
909
allowed = sysctl_overcommit_kbytes >> (PAGE_SHIFT - 10);
mm/util.c
983
allowed -= sysctl_admin_reserve_kbytes >> (PAGE_SHIFT - 10);
mm/util.c
989
long reserve = sysctl_user_reserve_kbytes >> (PAGE_SHIFT - 10);
mm/util.c
997
bytes_failed = pages << PAGE_SHIFT;
mm/vma.c
1881
pgoff = addr >> PAGE_SHIFT;
mm/vma.c
1982
b->vm_pgoff == a->vm_pgoff + ((b->vm_start - a->vm_start) >> PAGE_SHIFT);
mm/vma.c
2894
if (!may_expand_vm(mm, &vma_flags, len >> PAGE_SHIFT))
mm/vma.c
2900
if (security_vm_enough_memory_mm(mm, len >> PAGE_SHIFT))
mm/vma.c
2928
vma_set_range(vma, addr, addr + len, addr >> PAGE_SHIFT);
mm/vma.c
2939
mm->total_vm += len >> PAGE_SHIFT;
mm/vma.c
2940
mm->data_vm += len >> PAGE_SHIFT;
mm/vma.c
2942
mm->locked_vm += (len >> PAGE_SHIFT);
mm/vma.c
2950
vm_unacct_memory(len >> PAGE_SHIFT);
mm/vma.c
3086
grow << PAGE_SHIFT))
mm/vma.c
3169
grow = (address - vma->vm_end) >> PAGE_SHIFT;
mm/vma.c
3172
if (vma->vm_pgoff + (size >> PAGE_SHIFT) >= vma->vm_pgoff) {
mm/vma.c
3248
grow = (vma->vm_start - address) >> PAGE_SHIFT;
mm/vma.c
3321
vma->vm_pgoff = vma->vm_start >> PAGE_SHIFT;
mm/vma.c
524
new->vm_pgoff += ((addr - vma->vm_start) >> PAGE_SHIFT);
mm/vma.c
563
vma->vm_pgoff += (addr - new->vm_start) >> PAGE_SHIFT;
mm/vmalloc.c
106
pfn = phys_addr >> PAGE_SHIFT;
mm/vmalloc.c
2311
unsigned long nr = va_size(va) >> PAGE_SHIFT;
mm/vmalloc.c
2450
nr_lazy = atomic_long_add_return_relaxed(va_size(va) >> PAGE_SHIFT,
mm/vmalloc.c
2684
addr = va_start + (pages_off << PAGE_SHIFT);
mm/vmalloc.c
2914
offset = (addr & (VMAP_BLOCK_SIZE - 1)) >> PAGE_SHIFT;
mm/vmalloc.c
2972
s = va_start + (vb->dirty_min << PAGE_SHIFT);
mm/vmalloc.c
2973
e = va_start + (vb->dirty_max << PAGE_SHIFT);
mm/vmalloc.c
3021
unsigned long size = (unsigned long)count << PAGE_SHIFT;
mm/vmalloc.c
3064
unsigned long size = (unsigned long)count << PAGE_SHIFT;
mm/vmalloc.c
3087
pages, PAGE_SHIFT) < 0) {
mm/vmalloc.c
3213
PAGE_SHIFT, IOREMAP_MAX_ORDER);
mm/vmalloc.c
3251
return __get_vm_area_node(size, 1, PAGE_SHIFT, flags, start, end,
mm/vmalloc.c
3268
return __get_vm_area_node(size, 1, PAGE_SHIFT, flags,
mm/vmalloc.c
3277
return __get_vm_area_node(size, 1, PAGE_SHIFT, flags,
mm/vmalloc.c
3545
size = (unsigned long)count << PAGE_SHIFT;
mm/vmalloc.c
3552
pages, PAGE_SHIFT) < 0) {
mm/vmalloc.c
382
pte = PTR_ALIGN_DOWN(pte, sizeof(*pte) * (size >> PAGE_SHIFT));
mm/vmalloc.c
3838
unsigned int nr_small_pages = size >> PAGE_SHIFT;
mm/vmalloc.c
3867
set_vm_area_page_order(area, page_shift - PAGE_SHIFT);
mm/vmalloc.c
391
} while (pte += (size >> PAGE_SHIFT), addr += size, addr != end);
mm/vmalloc.c
3995
unsigned int shift = PAGE_SHIFT;
mm/vmalloc.c
4000
if ((size >> PAGE_SHIFT) > totalram_pages()) {
mm/vmalloc.c
4096
if (shift > PAGE_SHIFT) {
mm/vmalloc.c
4097
shift = PAGE_SHIFT;
mm/vmalloc.c
4547
n = ((re - rs + 1) << PAGE_SHIFT) - offset;
mm/vmalloc.c
4721
if (check_shl_overflow(pgoff, PAGE_SHIFT, &off))
mm/vmalloc.c
58
ioremap_max_page_shift = PAGE_SHIFT;
mm/vmalloc.c
63
static const unsigned int ioremap_max_page_shift = PAGE_SHIFT;
mm/vmalloc.c
662
unsigned int i, nr = (end - addr) >> PAGE_SHIFT;
mm/vmalloc.c
664
WARN_ON(page_shift < PAGE_SHIFT);
mm/vmalloc.c
667
page_shift == PAGE_SHIFT)
mm/vmalloc.c
670
for (i = 0; i < nr; i += 1U << (page_shift - PAGE_SHIFT)) {
mm/vmalloc.c
736
if ((end - start) >> PAGE_SHIFT > totalram_pages())
mm/vmalloc.c
760
return vmap_pages_range(start, end, PAGE_KERNEL, pages, PAGE_SHIFT);
mm/vmalloc.c
826
return p4d_page(*p4d) + ((addr & ~P4D_MASK) >> PAGE_SHIFT);
mm/vmalloc.c
834
return pud_page(*pud) + ((addr & ~PUD_MASK) >> PAGE_SHIFT);
mm/vmalloc.c
842
return pmd_page(*pmd) + ((addr & ~PMD_MASK) >> PAGE_SHIFT);
mm/vmscan.c
2290
gb = (inactive + active) >> (30 - PAGE_SHIFT);
mm/vmscan.c
3537
const unsigned int max_nr = (end - addr) >> PAGE_SHIFT;
mm/vmscan.c
4275
const unsigned int max_nr = (end - addr) >> PAGE_SHIFT;
mm/vmstat.c
260
mem = zone_managed_pages(zone) >> (27 - PAGE_SHIFT);
mm/vmstat.c
392
delta >>= PAGE_SHIFT;
mm/vmstat.c
634
delta >>= PAGE_SHIFT;
mm/workingset.c
691
lruvec, NR_SLAB_RECLAIMABLE_B) >> PAGE_SHIFT;
mm/workingset.c
693
lruvec, NR_SLAB_UNRECLAIMABLE_B) >> PAGE_SHIFT;
mm/zsmalloc.c
103
MAX(32, (ZS_MAX_PAGES_PER_ZSPAGE << PAGE_SHIFT >> OBJ_INDEX_BITS))
mm/zsmalloc.c
1260
nr_zpdesc = offset >> PAGE_SHIFT;
mm/zsmalloc.c
77
#define _PFN_BITS (MAX_POSSIBLE_PHYSMEM_BITS - PAGE_SHIFT)
mm/zswap.c
1233
nr_backing = memcg_page_state(memcg, MEMCG_ZSWAP_B) >> PAGE_SHIFT;
net/ceph/crypto.c
223
chunk_cnt = chunk_len >> PAGE_SHIFT;
net/ceph/messenger_v2.c
1079
int idx = spos >> PAGE_SHIFT;
net/ceph/messenger_v2.c
859
sg_cnt += PAGE_ALIGN(buf_len) >> PAGE_SHIFT;
net/ceph/messenger_v2.c
955
int idx = dpos >> PAGE_SHIFT;
net/ceph/pagelist.c
107
space = (space + PAGE_SIZE - 1) >> PAGE_SHIFT; /* conv to num pages */
net/ceph/pagevec.c
87
int i = off >> PAGE_SHIFT;
net/core/datagram.c
675
start += (pages[n] - head) << PAGE_SHIFT;
net/core/devmem.c
256
binding->chunk_pool = gen_pool_create(PAGE_SHIFT,
net/core/devmem.c
49
((dma_addr_t)net_iov_idx(niov) << PAGE_SHIFT);
net/core/devmem.h
122
((unsigned long)net_iov_idx(niov) << PAGE_SHIFT);
net/core/page_pool_priv.h
22
netmem_set_dma_addr(netmem, addr >> PAGE_SHIFT);
net/core/page_pool_priv.h
28
<< PAGE_SHIFT;
net/core/skbuff.c
1685
num_pg = (size >> PAGE_SHIFT) + 2; /* worst case */
net/core/skbuff.c
1686
max_pg = rlim >> PAGE_SHIFT;
net/core/skbuff.c
2012
new_frags = (__skb_pagelen(skb) + psize - 1) >> (PAGE_SHIFT + order);
net/core/skbuff.c
4443
pg_idx = (pg_off + st->frag_off) >> PAGE_SHIFT;
net/core/sock.c
1068
sk_forward_alloc_add(sk, pages << PAGE_SHIFT);
net/core/sock.c
1071
sk->sk_reserved_mem + (pages << PAGE_SHIFT));
net/core/sock.c
3411
sk_forward_alloc_add(sk, amt << PAGE_SHIFT);
net/core/sock.c
3414
sk_forward_alloc_add(sk, -(amt << PAGE_SHIFT));
net/core/sock.c
3448
amount >>= PAGE_SHIFT;
net/core/sock.c
3449
sk_forward_alloc_add(sk, -(amount << PAGE_SHIFT));
net/ipv4/inetpeer.c
77
nr_entries = div64_ul((u64)totalram_pages() << PAGE_SHIFT,
net/ipv4/tcp.c
5357
limit = nr_free_buffer_pages() << (PAGE_SHIFT - 7);
net/ipv4/tcp.c
919
sk_forward_alloc_add(sk, amt << PAGE_SHIFT);
net/ipv4/tcp_input.c
5521
int npages = min_t(size_t, size >> PAGE_SHIFT, MAX_SKB_FRAGS);
net/ipv4/tcp_input.c
5523
data_len = npages << PAGE_SHIFT;
net/ipv4/tcp_sigpool.c
350
page = skb_frag_page(f) + (offset >> PAGE_SHIFT);
net/ipv4/udp.c
1615
__sk_mem_reduce_allocated(sk, amt >> PAGE_SHIFT);
net/netfilter/ipvs/ip_vs_conn.c
2006
max_avail = order_base_2(totalram_pages()) + PAGE_SHIFT;
net/netfilter/nf_conntrack_core.c
2638
= (((nr_pages << PAGE_SHIFT) / 16384)
net/netfilter/xt_hashlimit.c
289
size = (nr_pages << PAGE_SHIFT) / 16384 /
net/packet/diag.c
80
pdr.pdr_block_size = ring->pg_vec_pages << PAGE_SHIFT;
net/rds/ib_frmr.c
245
frmr->dma_npages += len >> PAGE_SHIFT;
net/rds/info.c
188
>> PAGE_SHIFT;
net/rds/rdma.c
222
if ((nr_pages - 1) > (RDS_MAX_MSG_SIZE >> PAGE_SHIFT)) {
net/rds/rdma.c
61
return ((vec->addr + vec->bytes + PAGE_SIZE - 1) >> PAGE_SHIFT) -
net/rds/rdma.c
62
(vec->addr >> PAGE_SHIFT);
net/sctp/protocol.c
1508
limit = (sysctl_sctp_mem[1]) << (PAGE_SHIFT - 7);
net/sctp/protocol.c
1524
goal = nr_pages >> (22 - PAGE_SHIFT);
net/sctp/protocol.c
1526
goal = nr_pages >> (24 - PAGE_SHIFT);
net/sunrpc/auth_gss/auth_gss.c
1816
first = snd_buf->page_base >> PAGE_SHIFT;
net/sunrpc/auth_gss/auth_gss.c
1817
last = (snd_buf->page_base + snd_buf->page_len - 1) >> PAGE_SHIFT;
net/sunrpc/auth_gss/auth_gss.c
1863
first = snd_buf->page_base >> PAGE_SHIFT;
net/sunrpc/auth_gss/auth_gss.c
1866
snd_buf->page_base -= first << PAGE_SHIFT;
net/sunrpc/auth_gss/gss_krb5_crypto.c
255
int i = (page_pos + outbuf->page_base) >> PAGE_SHIFT;
net/sunrpc/auth_gss/svcauth_gss.c
1093
pgto = to_offs >> PAGE_SHIFT;
net/sunrpc/auth_gss/svcauth_gss.c
1094
pgfrom = from_offs >> PAGE_SHIFT;
net/sunrpc/rpc_pipe.c
1126
sb->s_blocksize_bits = PAGE_SHIFT;
net/sunrpc/socklib.c
67
struct page **ppage = xdr->pages + (xdr->page_base >> PAGE_SHIFT);
net/sunrpc/svcsock.c
1013
npages = (len + PAGE_SIZE - 1) >> PAGE_SHIFT;
net/sunrpc/svcsock.c
1027
npages = (len + PAGE_SIZE - 1) >> PAGE_SHIFT;
net/sunrpc/svcsock.c
994
npages = (len + PAGE_SIZE - 1) >> PAGE_SHIFT;
net/sunrpc/xdr.c
112
return (buf->page_base + buf->page_len + PAGE_SIZE - 1) >> PAGE_SHIFT;
net/sunrpc/xdr.c
1191
xdr->page_ptr = buf->pages + (new >> PAGE_SHIFT);
net/sunrpc/xdr.c
1346
pgnr = base >> PAGE_SHIFT;
net/sunrpc/xdr.c
1379
newbase = (1 + xdr->page_ptr - xdr->buf->pages) << PAGE_SHIFT;
net/sunrpc/xdr.c
1682
subbuf->pages = &buf->pages[base >> PAGE_SHIFT];
net/sunrpc/xdr.c
2005
ppages = buf->pages + (base >> PAGE_SHIFT);
net/sunrpc/xdr.c
2186
i = (offset + buf->page_base) >> PAGE_SHIFT;
net/sunrpc/xdr.c
251
pgto = pages + (pgto_base >> PAGE_SHIFT);
net/sunrpc/xdr.c
252
pgfrom = pages + (pgfrom_base >> PAGE_SHIFT);
net/sunrpc/xdr.c
319
pgto = pages + (pgto_base >> PAGE_SHIFT);
net/sunrpc/xdr.c
320
pgfrom = pages + (pgfrom_base >> PAGE_SHIFT);
net/sunrpc/xdr.c
377
pgto = pages + (pgbase >> PAGE_SHIFT);
net/sunrpc/xdr.c
424
pgfrom = pages + (pgbase >> PAGE_SHIFT);
net/sunrpc/xdr.c
484
page = pages + (pgbase >> PAGE_SHIFT);
net/sunrpc/xdr.c
515
npages = (pagelen + buf->page_base + PAGE_SIZE - 1) >> PAGE_SHIFT;
net/sunrpc/xdr.c
523
pagelen = i << PAGE_SHIFT;
net/sunrpc/xprtrdma/rpc_rdma.c
1064
(rqst->rq_rcv_buf.page_base >> PAGE_SHIFT);
net/sunrpc/xprtrdma/rpc_rdma.c
1073
npages = PAGE_ALIGN(page_base + pagelist_len) >> PAGE_SHIFT;
net/sunrpc/xprtrdma/rpc_rdma.c
190
ppages = buf->pages + (buf->page_base >> PAGE_SHIFT);
net/sunrpc/xprtrdma/rpc_rdma.c
241
ppages = xdrbuf->pages + (xdrbuf->page_base >> PAGE_SHIFT);
net/sunrpc/xprtrdma/rpc_rdma.c
605
ppages = xdr->pages + (xdr->page_base >> PAGE_SHIFT);
net/sunrpc/xprtrdma/rpc_rdma.c
686
ppages = xdr->pages + (xdr->page_base >> PAGE_SHIFT);
net/sunrpc/xprtrdma/svc_rdma_rw.c
461
page_no = page_off >> PAGE_SHIFT;
net/sunrpc/xprtrdma/svc_rdma_rw.c
508
(write_len >> PAGE_SHIFT) + 2);
net/sunrpc/xprtrdma/svc_rdma_rw.c
760
nr_bvec = PAGE_ALIGN(total) >> PAGE_SHIFT;
net/sunrpc/xprtrdma/svc_rdma_rw.c
857
numpages = PAGE_ALIGN(head->rc_pageoff + remaining) >> PAGE_SHIFT;
net/sunrpc/xprtrdma/svc_rdma_sendto.c
642
ppages = xdr->pages + (xdr->page_base >> PAGE_SHIFT);
net/sunrpc/xprtrdma/svc_rdma_sendto.c
756
ppages = xdr->pages + (xdr->page_base >> PAGE_SHIFT);
net/sunrpc/xprtrdma/svc_rdma_transport.c
474
maxpayload >> PAGE_SHIFT);
net/sunrpc/xprtrdma/transport.c
360
xprt->max_payload = RPCRDMA_MAX_DATA_SEGS << PAGE_SHIFT;
net/sunrpc/xprtrdma/xprt_rdma.h
311
RPCRDMA_MAX_PAGE_SGES = RPCRDMA_MAX_INLINE >> PAGE_SHIFT,
net/sunrpc/xprtsock.c
346
n = (buf->page_base + want + PAGE_SIZE - 1) >> PAGE_SHIFT;
net/tls/tls.h
45
TLS_MAX_PAYLOAD_SIZE >> PAGE_SHIFT))
net/xdp/xdp_umem.c
136
lock_limit = rlimit(RLIMIT_MEMLOCK) >> PAGE_SHIFT;
net/xdp/xsk.c
1832
loff_t offset = (loff_t)vma->vm_pgoff << PAGE_SHIFT;
net/xdp/xsk.c
862
page = pool->umem->pgs[addr >> PAGE_SHIFT];
net/xdp/xsk_buff_pool.c
743
return (pool->dma_pages[addr >> PAGE_SHIFT] &
samples/vfio-mdev/mbochs.c
477
pg = __mbochs_get_page(mdev_state, pos >> PAGE_SHIFT);
samples/vfio-mdev/mbochs.c
535
mdev_state->pagecount = mdev_state->memsize >> PAGE_SHIFT;
samples/vfio-mdev/mbochs.c
777
pgoff_t page_offset = (vmf->address - vma->vm_start) >> PAGE_SHIFT;
samples/vfio-mdev/mbochs.c
798
if (vma->vm_pgoff != MBOCHS_MEMORY_BAR_OFFSET >> PAGE_SHIFT)
samples/vfio-mdev/mbochs.c
948
page_offset = dmabuf->mode.offset >> PAGE_SHIFT;
samples/vfio-mdev/mdpy.c
426
if (vma->vm_pgoff != MDPY_MEMORY_BAR_OFFSET >> PAGE_SHIFT)
security/integrity/ima/ima_kexec.c
170
((kexec_segment_size >> PAGE_SHIFT) > totalram_pages() / 2)) {
security/selinux/selinuxfs.c
451
offset = vmf->pgoff << PAGE_SHIFT;
security/selinux/ss/sidtab.h
37
#define SIDTAB_NODE_ALLOC_SHIFT PAGE_SHIFT
sound/core/memalloc.c
311
set_memory_wc((unsigned long)(p), size >> PAGE_SHIFT);
sound/core/memalloc.c
320
set_memory_wb((unsigned long)(p), size >> PAGE_SHIFT);
sound/core/memalloc.c
340
dmab->addr >> PAGE_SHIFT,
sound/core/memalloc.c
458
dmab->addr >> PAGE_SHIFT,
sound/core/memalloc.c
619
offset >> PAGE_SHIFT);
sound/core/memalloc.c
702
size << PAGE_SHIFT, wc);
sound/core/memalloc.c
726
sgbuf->count = size >> PAGE_SHIFT;
sound/core/memalloc.c
748
npages = chunk >> PAGE_SHIFT;
sound/core/memalloc.c
757
0, sgbuf->count << PAGE_SHIFT, GFP_KERNEL))
sound/core/pcm_native.c
3906
offset = vmf->pgoff << PAGE_SHIFT;
sound/core/pcm_native.c
4017
offset = area->vm_pgoff << PAGE_SHIFT;
sound/core/pcm_native.c
4049
offset = area->vm_pgoff << PAGE_SHIFT;
sound/core/pcm_native.c
4210
unsigned long offset = pgoff << PAGE_SHIFT;
sound/pci/emu10k1/emu10k1_main.c
1566
emu->max_cache_pages = max_cache_bytes >> PAGE_SHIFT;
sound/pci/emu10k1/emupcm.c
390
epcm->start_addr = mapped << PAGE_SHIFT;
sound/pci/emu10k1/memory.c
225
blk = (struct snd_emu10k1_memblk *)__snd_util_memblk_new(emu->memhdr, psize << PAGE_SHIFT, p->prev);
sound/pci/emu10k1/memory.c
31
#define get_aligned_page(offset) ((offset) >> PAGE_SHIFT)
sound/pci/emu10k1/memory.c
324
unsigned long ofs = idx << PAGE_SHIFT;
sound/pci/emu10k1/memory.c
33
#define aligned_page_offset(page) ((page) << PAGE_SHIFT)
sound/pci/riptide/riptide.c
1502
ofs = j << PAGE_SHIFT;
sound/pci/trident/trident_memory.c
183
unsigned long ofs = idx << PAGE_SHIFT;
sound/ppc/snd_ps3.c
902
PAGE_SHIFT);
sound/ppc/snd_ps3.c
910
PAGE_SHIFT, /* use system page size */
sound/soc/amd/acp-pcm-dma.c
1012
rtd->num_of_pages = PAGE_ALIGN(size) >> PAGE_SHIFT;
sound/soc/amd/acp/acp-platform.c
136
int num_pages = (PAGE_ALIGN(size) >> PAGE_SHIFT);
sound/soc/amd/ps/ps-pdm-dma.c
230
rtd->num_pages = (PAGE_ALIGN(size) >> PAGE_SHIFT);
sound/soc/amd/ps/ps-sdw-dma.c
417
stream->num_pages = (PAGE_ALIGN(size) >> PAGE_SHIFT);
sound/soc/amd/raven/acp3x-pcm-dma.c
290
rtd->num_pages = (PAGE_ALIGN(size) >> PAGE_SHIFT);
sound/soc/amd/renoir/acp3x-pdm-dma.c
252
rtd->num_pages = (PAGE_ALIGN(size) >> PAGE_SHIFT);
sound/soc/amd/vangogh/acp5x-pcm-dma.c
285
rtd->num_pages = (PAGE_ALIGN(size) >> PAGE_SHIFT);
sound/soc/amd/yc/acp6x-pdm-dma.c
228
rtd->num_pages = (PAGE_ALIGN(size) >> PAGE_SHIFT);
sound/soc/loongson/loongson_dma.c
301
substream->dma_buffer.addr >> PAGE_SHIFT,
sound/soc/pxa/mmp-sspa.c
410
substream->dma_buffer.addr >> PAGE_SHIFT,
sound/soc/qcom/lpass-platform.c
903
offset = vma->vm_pgoff << PAGE_SHIFT;
sound/soc/qcom/lpass-platform.c
905
(runtime->dma_addr + offset) >> PAGE_SHIFT,
sound/soc/samsung/idma.c
251
offset = vma->vm_pgoff << PAGE_SHIFT;
sound/soc/samsung/idma.c
253
(runtime->dma_addr + offset) >> PAGE_SHIFT,
sound/soc/sof/amd/acp-loader.c
181
page_count = PAGE_ALIGN(size_fw) >> PAGE_SHIFT;
sound/soc/sof/amd/acp-loader.c
232
dma_size = (PAGE_ALIGN(sdev->basefw.fw->size) >> PAGE_SHIFT) * ACP_PAGE_SIZE;
sound/soc/sof/amd/acp-loader.c
64
page_count = PAGE_ALIGN(size_fw) >> PAGE_SHIFT;
sound/soc/sof/sof-utils.c
45
u32 pfn = snd_sgbuf_get_addr(dmab, i * PAGE_SIZE) >> PAGE_SHIFT;
sound/soc/uniphier/aio-dma.c
201
substream->runtime->dma_addr >> PAGE_SHIFT,
sound/usb/usx2y/us122l.c
105
offset = vmf->pgoff << PAGE_SHIFT;
sound/usb/usx2y/us122l.c
179
offset = area->vm_pgoff << PAGE_SHIFT;
sound/usb/usx2y/usX2Yhwdep.c
27
offset = vmf->pgoff << PAGE_SHIFT;
sound/usb/usx2y/usx2yhwdeppcm.c
674
offset = vmf->pgoff << PAGE_SHIFT;
tools/include/linux/mm.h
10
#define PAGE_SIZE (_AC(1, UL) << PAGE_SHIFT)
tools/include/linux/pfn.h
10
#define PHYS_PFN(x) ((unsigned long)((x) >> PAGE_SHIFT))
tools/include/linux/pfn.h
7
#define PFN_UP(x) (((x) + PAGE_SIZE - 1) >> PAGE_SHIFT)
tools/include/linux/pfn.h
8
#define PFN_DOWN(x) ((x) >> PAGE_SHIFT)
tools/include/linux/pfn.h
9
#define PFN_PHYS(x) ((phys_addr_t)(x) << PAGE_SHIFT)
tools/testing/selftests/drivers/net/hw/ncdevmem.c
993
dmabuf_cmsg->frag_offset >> PAGE_SHIFT,
tools/testing/selftests/kvm/include/s390/processor.h
26
#define PAGE_SIZE BIT_ULL(PAGE_SHIFT)
tools/testing/selftests/kvm/include/x86/processor.h
1396
return kvm_hypercall(KVM_HC_MAP_GPA_RANGE, gpa, size >> PAGE_SHIFT, flags, 0);
tools/testing/selftests/kvm/include/x86/processor.h
369
#define PAGE_SIZE (1ULL << PAGE_SHIFT)
tools/testing/selftests/kvm/include/x86/processor.h
372
#define HUGEPAGE_SHIFT(x) (PAGE_SHIFT + (((x) - 1) * 9))
tools/testing/selftests/kvm/include/x86/processor.h
377
#define PTE_GET_PFN(pte) (PTE_GET_PA(pte) >> PAGE_SHIFT)
tools/testing/selftests/kvm/include/x86/sev.h
139
.gfn_start = gpa >> PAGE_SHIFT,
tools/testing/selftests/kvm/lib/s390/processor.c
106
idx = (gva >> PAGE_SHIFT) & 0x0ffu; /* page index */
tools/testing/selftests/kvm/lib/s390/processor.c
82
idx = (gva >> PAGE_SHIFT) & 0x0ffu; /* page index */
tools/testing/selftests/kvm/s390/keyop.c
26
#define BUF_START_ADDR (BUF_START_GFN << PAGE_SHIFT)
tools/testing/selftests/kvm/s390/ucontrol_test.c
155
KVM_S390_SIE_PAGE_OFFSET << PAGE_SHIFT);
tools/testing/selftests/kvm/x86/nested_dirty_log_test.c
162
page_nr = (gva - TEST_MEM_ALIAS_BASE) >> PAGE_SHIFT;
tools/testing/selftests/kvm/x86/nested_dirty_log_test.c
164
page_nr = (gva - TEST_MEM_BASE) >> PAGE_SHIFT;
tools/testing/selftests/kvm/x86/private_mem_conversions_test.c
184
if ((j >> PAGE_SHIFT) & 1) {
tools/testing/vma/include/dup.h
1313
return (vma->vm_end - vma->vm_start) >> PAGE_SHIFT;
tools/testing/vma/include/dup.h
1491
locked_pages = bytes >> PAGE_SHIFT;
tools/testing/vma/include/dup.h
1495
limit_pages >>= PAGE_SHIFT;
tools/testing/vma/include/dup.h
400
#define PHYS_PFN(x) ((unsigned long)((x) >> PAGE_SHIFT))
tools/testing/vma/shared.c
9
unsigned long stack_guard_gap = 256UL<<PAGE_SHIFT;
virt/kvm/guest_memfd.c
233
pgoff_t start = offset >> PAGE_SHIFT;
virt/kvm/guest_memfd.c
234
pgoff_t end = (offset + len) >> PAGE_SHIFT;
virt/kvm/guest_memfd.c
265
start = offset >> PAGE_SHIFT;
virt/kvm/guest_memfd.c
266
end = (offset + len) >> PAGE_SHIFT;
virt/kvm/guest_memfd.c
398
if (((loff_t)vmf->pgoff << PAGE_SHIFT) >= i_size_read(inode))
virt/kvm/guest_memfd.c
446
*pgoff = vma->vm_pgoff + ((addr - vma->vm_start) >> PAGE_SHIFT);
virt/kvm/guest_memfd.c
646
loff_t size = slot->npages << PAGE_SHIFT;
virt/kvm/guest_memfd.c
674
start = offset >> PAGE_SHIFT;
virt/kvm/kvm_main.c
1536
(new->npages << PAGE_SHIFT) - 1;
virt/kvm/kvm_main.c
2039
(mem->memory_size >> PAGE_SHIFT) > KVM_MEM_MAX_NR_PAGES)
virt/kvm/kvm_main.c
2060
base_gfn = (mem->guest_phys_addr >> PAGE_SHIFT);
virt/kvm/kvm_main.c
2061
npages = (mem->memory_size >> PAGE_SHIFT);
virt/kvm/kvm_main.c
2611
start = attrs->address >> PAGE_SHIFT;
virt/kvm/kvm_main.c
2612
end = (attrs->address + attrs->size) >> PAGE_SHIFT;
virt/kvm/kvm_main.c
3216
gfn_t gfn = gpa >> PAGE_SHIFT;
virt/kvm/kvm_main.c
3236
gfn_t gfn = gpa >> PAGE_SHIFT;
virt/kvm/kvm_main.c
3277
gfn_t gfn = gpa >> PAGE_SHIFT;
virt/kvm/kvm_main.c
3327
gfn_t gfn = gpa >> PAGE_SHIFT;
virt/kvm/kvm_main.c
3348
gfn_t gfn = gpa >> PAGE_SHIFT;
virt/kvm/kvm_main.c
3371
gfn_t start_gfn = gpa >> PAGE_SHIFT;
virt/kvm/kvm_main.c
3372
gfn_t end_gfn = (gpa + len - 1) >> PAGE_SHIFT;
virt/kvm/kvm_main.c
3440
mark_page_dirty_in_slot(kvm, ghc->memslot, gpa >> PAGE_SHIFT);
virt/kvm/kvm_main.c
3493
gfn_t gfn = gpa >> PAGE_SHIFT;
virt/kvm/kvm_main.c
597
slot->userspace_addr + (slot->npages << PAGE_SHIFT));