Symbol: pgoff_t
arch/arm/mm/fault-armv.c
132
pgoff_t pgoff;
arch/arm/mm/flush.c
242
pgoff_t pgoff, pgoff_end;
arch/arm64/include/asm/mte.h
30
void mte_invalidate_tags(int type, pgoff_t offset);
arch/arm64/include/asm/pgtable.h
1554
static inline void arch_swap_invalidate_page(int type, pgoff_t offset)
arch/arm64/mm/mteswap.c
63
void mte_invalidate_tags(int type, pgoff_t offset)
arch/nios2/mm/cacheflush.c
79
pgoff_t pgoff;
arch/parisc/kernel/cache.c
481
pgoff_t pgoff;
arch/powerpc/kvm/book3s_xive_native.c
204
pgoff_t esb_pgoff = KVM_XIVE_ESB_PAGE_OFFSET + irq * 2;
arch/x86/kernel/cpu/sgx/encl.c
110
static inline pgoff_t sgx_encl_get_backing_page_pcmd_offset(struct sgx_encl *encl,
arch/x86/kernel/cpu/sgx/encl.c
113
pgoff_t epc_end_off = encl->size + sizeof(struct sgx_secs);
arch/x86/kernel/cpu/sgx/encl.c
138
pgoff_t page_index, page_pcmd_off;
arch/x86/kernel/cpu/sgx/encl.c
951
pgoff_t index)
arch/x86/kernel/cpu/sgx/encl.c
975
pgoff_t page_pcmd_off = sgx_encl_get_backing_page_pcmd_offset(encl, page_index);
arch/x86/kernel/cpu/sgx/main.c
306
pgoff_t page_index;
drivers/accel/amdxdna/amdxdna_ubuf.c
77
pgoff_t pgoff;
drivers/android/binder_alloc.c
1342
pgoff_t pgoff;
drivers/android/binder_alloc.c
1374
pgoff_t pgoff;
drivers/android/binder_alloc.c
814
pgoff_t *pgoffp)
drivers/android/binder_alloc.c
818
pgoff_t pgoff = buffer_space_offset & ~PAGE_MASK;
drivers/android/binder_alloc.c
842
pgoff_t pgoff;
drivers/block/brd.c
124
pgoff_t idx;
drivers/dax/dax-private.h
115
phys_addr_t dax_pgoff_to_phys(struct dev_dax *dev_dax, pgoff_t pgoff, unsigned long size);
drivers/dax/device.c
149
pgoff_t pgoff;
drivers/dax/device.c
194
pgoff_t pgoff;
drivers/dax/device.c
61
__weak phys_addr_t dax_pgoff_to_phys(struct dev_dax *dev_dax, pgoff_t pgoff,
drivers/dax/device.c
89
pgoff_t pgoff;
drivers/dax/super.c
149
long dax_direct_access(struct dax_device *dax_dev, pgoff_t pgoff, long nr_pages,
drivers/dax/super.c
171
size_t dax_copy_from_iter(struct dax_device *dax_dev, pgoff_t pgoff, void *addr,
drivers/dax/super.c
187
size_t dax_copy_to_iter(struct dax_device *dax_dev, pgoff_t pgoff, void *addr,
drivers/dax/super.c
203
int dax_zero_page_range(struct dax_device *dax_dev, pgoff_t pgoff,
drivers/dax/super.c
223
size_t dax_recovery_write(struct dax_device *dax_dev, pgoff_t pgoff,
drivers/dma-buf/heaps/cma_heap.c
304
pgoff_t pagecount = size >> PAGE_SHIFT;
drivers/dma-buf/heaps/cma_heap.c
309
pgoff_t pg;
drivers/dma-buf/heaps/cma_heap.c
58
pgoff_t pagecount;
drivers/dma-buf/udmabuf.c
114
pgoff_t pg;
drivers/dma-buf/udmabuf.c
200
pgoff_t i;
drivers/dma-buf/udmabuf.c
208
static __always_inline int init_udmabuf(struct udmabuf *ubuf, pgoff_t pgcnt)
drivers/dma-buf/udmabuf.c
28
pgoff_t pagecount;
drivers/dma-buf/udmabuf.c
326
pgoff_t nr_pinned = ubuf->nr_pinned;
drivers/dma-buf/udmabuf.c
327
pgoff_t upgcnt = ubuf->pagecount;
drivers/dma-buf/udmabuf.c
329
pgoff_t pgoff, pgcnt;
drivers/dma-buf/udmabuf.c
341
pgoff_t subpgoff = pgoff;
drivers/dma-buf/udmabuf.c
375
pgoff_t pgcnt = 0, pglimit;
drivers/dma-buf/udmabuf.c
387
pgoff_t subpgcnt;
drivers/dma-buf/udmabuf.c
39
pgoff_t nr_pinned;
drivers/dma-buf/udmabuf.c
44
pgoff_t *offsets;
drivers/dma-buf/udmabuf.c
51
pgoff_t pgoff = vmf->pgoff;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
1206
pgoff_t i;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
1246
pgoff_t i;
drivers/gpu/drm/drm_gem_shmem_helper.c
563
pgoff_t page_offset = vmf->pgoff - vma->vm_pgoff; /* page offset within VMA */
drivers/gpu/drm/drm_gem_shmem_helper.c
616
pgoff_t page_offset = vmf->pgoff - vma->vm_pgoff; /* page offset within VMA */
drivers/gpu/drm/etnaviv/etnaviv_gem.c
170
pgoff_t pgoff;
drivers/gpu/drm/gma500/gem.c
263
pgoff_t page_offset;
drivers/gpu/drm/i915/gem/i915_gem_mman.c
201
pgoff_t page_offset,
drivers/gpu/drm/i915/gem/i915_gem_mman.c
354
pgoff_t page_offset;
drivers/gpu/drm/i915/gem/i915_gem_object.c
480
pgoff_t idx = offset >> PAGE_SHIFT;
drivers/gpu/drm/i915/gem/i915_gem_object.c
495
pgoff_t idx = offset >> PAGE_SHIFT;
drivers/gpu/drm/i915/gem/i915_gem_object.c
537
GEM_BUG_ON(overflows_type(offset >> PAGE_SHIFT, pgoff_t));
drivers/gpu/drm/i915/gem/i915_gem_object.h
377
pgoff_t n,
drivers/gpu/drm/i915/gem/i915_gem_object.h
401
static_assert(castable_to_type(n, pgoff_t)); \
drivers/gpu/drm/i915/gem/i915_gem_object.h
424
__i915_gem_object_get_sg(struct drm_i915_gem_object *obj, pgoff_t n,
drivers/gpu/drm/i915/gem/i915_gem_object.h
446
static_assert(castable_to_type(n, pgoff_t)); \
drivers/gpu/drm/i915/gem/i915_gem_object.h
469
__i915_gem_object_get_sg_dma(struct drm_i915_gem_object *obj, pgoff_t n,
drivers/gpu/drm/i915/gem/i915_gem_object.h
491
static_assert(castable_to_type(n, pgoff_t)); \
drivers/gpu/drm/i915/gem/i915_gem_object.h
511
__i915_gem_object_get_page(struct drm_i915_gem_object *obj, pgoff_t n);
drivers/gpu/drm/i915/gem/i915_gem_object.h
527
static_assert(castable_to_type(n, pgoff_t)); \
drivers/gpu/drm/i915/gem/i915_gem_object.h
546
__i915_gem_object_get_dirty_page(struct drm_i915_gem_object *obj, pgoff_t n);
drivers/gpu/drm/i915/gem/i915_gem_object.h
562
static_assert(castable_to_type(n, pgoff_t)); \
drivers/gpu/drm/i915/gem/i915_gem_object.h
580
__i915_gem_object_get_dma_address_len(struct drm_i915_gem_object *obj, pgoff_t n,
drivers/gpu/drm/i915/gem/i915_gem_object.h
600
static_assert(castable_to_type(n, pgoff_t)); \
drivers/gpu/drm/i915/gem/i915_gem_object.h
617
__i915_gem_object_get_dma_address(struct drm_i915_gem_object *obj, pgoff_t n);
drivers/gpu/drm/i915/gem/i915_gem_object.h
635
static_assert(castable_to_type(n, pgoff_t)); \
drivers/gpu/drm/i915/gem/i915_gem_pages.c
647
pgoff_t n,
drivers/gpu/drm/i915/gem/i915_gem_pages.c
762
__i915_gem_object_get_page(struct drm_i915_gem_object *obj, pgoff_t n)
drivers/gpu/drm/i915/gem/i915_gem_pages.c
775
__i915_gem_object_get_dirty_page(struct drm_i915_gem_object *obj, pgoff_t n)
drivers/gpu/drm/i915/gem/i915_gem_pages.c
788
pgoff_t n, unsigned int *len)
drivers/gpu/drm/i915/gem/i915_gem_pages.c
802
__i915_gem_object_get_dma_address(struct drm_i915_gem_object *obj, pgoff_t n)
drivers/gpu/drm/i915/i915_gem.c
232
pgoff_t idx;
drivers/gpu/drm/i915/i915_gem.c
672
pgoff_t idx;
drivers/gpu/drm/i915/i915_vma.c
1142
pgoff_t obj_offset,
drivers/gpu/drm/i915/i915_vma.c
1175
pgoff_t obj_offset, unsigned int alignment_pad,
drivers/gpu/drm/i915/i915_vma.c
977
pgoff_t src_idx;
drivers/gpu/drm/msm/msm_gem.c
336
pgoff_t pgoff;
drivers/gpu/drm/omapdrm/omap_gem.c
118
pgoff_t obj_pgoff; /* page offset of obj currently
drivers/gpu/drm/omapdrm/omap_gem.c
359
pgoff_t pgoff;
drivers/gpu/drm/omapdrm/omap_gem.c
388
pgoff_t pgoff, base_pgoff;
drivers/gpu/drm/panfrost/panfrost_mmu.c
595
pgoff_t page_offset, nr_pages;
drivers/gpu/drm/panfrost/panfrost_mmu.c
654
for (pgoff_t pg = page_offset; pg < page_offset + NUM_FAULT_PAGES;) {
drivers/gpu/drm/panthor/panthor_mmu.c
2110
pgoff_t bo_offset;
drivers/gpu/drm/tegra/gem.c
568
pgoff_t offset;
drivers/gpu/drm/ttm/ttm_backup.c
16
static unsigned long ttm_backup_shmem_idx_to_handle(pgoff_t idx)
drivers/gpu/drm/ttm/ttm_backup.c
21
static pgoff_t ttm_backup_handle_to_shmem_idx(pgoff_t handle)
drivers/gpu/drm/ttm/ttm_backup.c
31
void ttm_backup_drop(struct file *backup, pgoff_t handle)
drivers/gpu/drm/ttm/ttm_backup.c
53
pgoff_t handle, bool intr, gfp_t additional_gfp)
drivers/gpu/drm/ttm/ttm_backup.c
57
pgoff_t idx = ttm_backup_handle_to_shmem_idx(handle);
drivers/gpu/drm/ttm/ttm_backup.c
99
bool writeback, pgoff_t idx, gfp_t page_gfp,
drivers/gpu/drm/ttm/ttm_bo.c
1135
pgoff_t num_pages = tt->num_pages;
drivers/gpu/drm/ttm/ttm_bo_util.c
96
pgoff_t i;
drivers/gpu/drm/ttm/ttm_bo_vm.c
185
pgoff_t num_prefault)
drivers/gpu/drm/ttm/ttm_bo_vm.c
195
pgoff_t i;
drivers/gpu/drm/ttm/ttm_pool.c
1037
pgoff_t shrunken = 0;
drivers/gpu/drm/ttm/ttm_pool.c
1038
pgoff_t i, num_pages;
drivers/gpu/drm/ttm/ttm_pool.c
111
pgoff_t alloced_pages;
drivers/gpu/drm/ttm/ttm_pool.c
112
pgoff_t restored_pages;
drivers/gpu/drm/ttm/ttm_pool.c
470
pgoff_t nr;
drivers/gpu/drm/ttm/ttm_pool.c
510
static pgoff_t ttm_pool_unmap_and_free(struct ttm_pool *pool, struct page *page,
drivers/gpu/drm/ttm/ttm_pool.c
515
pgoff_t nr;
drivers/gpu/drm/ttm/ttm_pool.c
541
pgoff_t nr)
drivers/gpu/drm/ttm/ttm_pool.c
543
pgoff_t i;
drivers/gpu/drm/ttm/ttm_pool.c
569
pgoff_t i, nr = 1UL << restore->order;
drivers/gpu/drm/ttm/ttm_pool.c
713
pgoff_t start_page, pgoff_t end_page)
drivers/gpu/drm/ttm/ttm_pool.c
717
pgoff_t i, nr;
drivers/gpu/drm/ttm/ttm_pool.c
764
pgoff_t caching_divide;
drivers/gpu/drm/ttm/ttm_pool.c
84
pgoff_t remaining_pages;
drivers/gpu/drm/ttm/ttm_pool.c
966
pgoff_t start_page = 0;
drivers/gpu/drm/ttm/ttm_pool.c
990
pgoff_t mid = restore->snapshot_alloc.caching_divide - tt->pages;
drivers/gpu/drm/ttm/ttm_resource.c
722
pgoff_t i)
drivers/gpu/drm/ttm/ttm_resource.c
804
pgoff_t i)
drivers/gpu/drm/ttm/ttm_tt.c
489
pgoff_t i)
drivers/gpu/drm/vmwgfx/vmwgfx_drv.h
1393
pgoff_t start, pgoff_t end);
drivers/gpu/drm/vmwgfx/vmwgfx_drv.h
818
void vmw_resource_dirty_update(struct vmw_resource *res, pgoff_t start,
drivers/gpu/drm/vmwgfx/vmwgfx_drv.h
819
pgoff_t end);
drivers/gpu/drm/vmwgfx/vmwgfx_drv.h
821
int vmw_resources_clean(struct vmw_bo *vbo, pgoff_t start,
drivers/gpu/drm/vmwgfx/vmwgfx_drv.h
822
pgoff_t end, pgoff_t *num_prefault);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
111
pgoff_t num_marked;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
127
pgoff_t start = 0;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
128
pgoff_t end = dirty->bitmap_size;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
171
pgoff_t start, pgoff_t end)
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
196
pgoff_t start, pgoff_t end)
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
219
pgoff_t num_pages = PFN_UP(vbo->tbo.resource->size);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
243
pgoff_t offset = drm_vma_node_start(&vbo->tbo.base.vma_node);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
302
pgoff_t start, cur, end;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
338
pgoff_t start, cur, end;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
448
pgoff_t num_prefault;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
460
pgoff_t allowed_prefault;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
71
pgoff_t offset = drm_vma_node_start(&vbo->tbo.base.vma_node);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
73
pgoff_t num_marked;
drivers/gpu/drm/vmwgfx/vmwgfx_resource.c
1068
void vmw_resource_dirty_update(struct vmw_resource *res, pgoff_t start,
drivers/gpu/drm/vmwgfx/vmwgfx_resource.c
1069
pgoff_t end)
drivers/gpu/drm/vmwgfx/vmwgfx_resource.c
1100
int vmw_resources_clean(struct vmw_bo *vbo, pgoff_t start,
drivers/gpu/drm/vmwgfx/vmwgfx_resource.c
1101
pgoff_t end, pgoff_t *num_prefault)
drivers/gpu/drm/vmwgfx/vmwgfx_resource.c
417
pgoff_t start = res->guest_memory_offset >> PAGE_SHIFT;
drivers/gpu/drm/vmwgfx/vmwgfx_resource.c
418
pgoff_t end = __KERNEL_DIV_ROUND_UP
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
1113
pgoff_t start = res->guest_memory_offset >> PAGE_SHIFT;
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
1114
pgoff_t end = __KERNEL_DIV_ROUND_UP(res->guest_memory_offset +
drivers/gpu/drm/xe/tests/xe_bo.c
32
pgoff_t ccs_page;
drivers/gpu/drm/xe/xe_pt.c
1667
static int xe_pt_stage_unbind_entry(struct xe_ptw *parent, pgoff_t offset,
drivers/gpu/drm/xe/xe_pt.c
1677
pgoff_t first = xe_pt_offset(addr, xe_child->level, walk);
drivers/gpu/drm/xe/xe_pt.c
1686
pgoff_t count = xe_pt_num_entries(addr, next, xe_child->level, walk);
drivers/gpu/drm/xe/xe_pt.c
1688
for (pgoff_t i = 0; i < count; i++) {
drivers/gpu/drm/xe/xe_pt.c
1771
xe_pt_stage_unbind_post_descend(struct xe_ptw *parent, pgoff_t offset,
drivers/gpu/drm/xe/xe_pt.c
1780
pgoff_t end_offset;
drivers/gpu/drm/xe/xe_pt.c
340
pgoff_t offset, bool alloc_entries)
drivers/gpu/drm/xe/xe_pt.c
386
pgoff_t offset, struct xe_pt *xe_child, u64 pte)
drivers/gpu/drm/xe/xe_pt.c
515
xe_pt_stage_bind_entry(struct xe_ptw *parent, pgoff_t offset,
drivers/gpu/drm/xe/xe_pt.c
838
pgoff_t *offset, pgoff_t *end_offset)
drivers/gpu/drm/xe/xe_pt.c
879
static int xe_pt_zap_ptes_entry(struct xe_ptw *parent, pgoff_t offset,
drivers/gpu/drm/xe/xe_pt.c
888
pgoff_t end_offset;
drivers/gpu/drm/xe/xe_pt_walk.c
36
static bool xe_pt_next(pgoff_t *offset, u64 *addr, u64 next, u64 end,
drivers/gpu/drm/xe/xe_pt_walk.c
39
pgoff_t step = 1;
drivers/gpu/drm/xe/xe_pt_walk.c
76
pgoff_t offset = xe_pt_offset(addr, level, walk);
drivers/gpu/drm/xe/xe_pt_walk.h
124
static inline pgoff_t
drivers/gpu/drm/xe/xe_pt_walk.h
143
static inline pgoff_t
drivers/gpu/drm/xe/xe_pt_walk.h
63
typedef int (*xe_pt_entry_fn)(struct xe_ptw *parent, pgoff_t offset,
drivers/hwtracing/stm/core.c
669
static int stm_mmap_mapped(unsigned long start, unsigned long end, pgoff_t pgoff,
drivers/iommu/iommu-pages.h
26
pgoff_t __index;
drivers/md/dm-linear.c
162
static struct dax_device *linear_dax_pgoff(struct dm_target *ti, pgoff_t *pgoff)
drivers/md/dm-linear.c
171
static long linear_dax_direct_access(struct dm_target *ti, pgoff_t pgoff,
drivers/md/dm-linear.c
180
static int linear_dax_zero_page_range(struct dm_target *ti, pgoff_t pgoff,
drivers/md/dm-linear.c
188
static size_t linear_dax_recovery_write(struct dm_target *ti, pgoff_t pgoff,
drivers/md/dm-log-writes.c
887
pgoff_t *pgoff)
drivers/md/dm-log-writes.c
895
static long log_writes_dax_direct_access(struct dm_target *ti, pgoff_t pgoff,
drivers/md/dm-log-writes.c
904
static int log_writes_dax_zero_page_range(struct dm_target *ti, pgoff_t pgoff,
drivers/md/dm-log-writes.c
913
pgoff_t pgoff, void *addr, size_t bytes, struct iov_iter *i)
drivers/md/dm-stripe.c
302
static struct dax_device *stripe_dax_pgoff(struct dm_target *ti, pgoff_t *pgoff)
drivers/md/dm-stripe.c
317
static long stripe_dax_direct_access(struct dm_target *ti, pgoff_t pgoff,
drivers/md/dm-stripe.c
326
static int stripe_dax_zero_page_range(struct dm_target *ti, pgoff_t pgoff,
drivers/md/dm-stripe.c
334
static size_t stripe_dax_recovery_write(struct dm_target *ti, pgoff_t pgoff,
drivers/md/dm-target.c
256
static long io_err_dax_direct_access(struct dm_target *ti, pgoff_t pgoff,
drivers/md/dm.c
1210
static long dm_dax_direct_access(struct dax_device *dax_dev, pgoff_t pgoff,
drivers/md/dm.c
1238
static int dm_dax_zero_page_range(struct dax_device *dax_dev, pgoff_t pgoff,
drivers/md/dm.c
1265
static size_t dm_dax_recovery_write(struct dax_device *dax_dev, pgoff_t pgoff,
drivers/mtd/devices/block2mtd.c
110
pgoff_t index = from >> PAGE_SHIFT;
drivers/mtd/devices/block2mtd.c
144
pgoff_t index = to >> PAGE_SHIFT; // page index
drivers/mtd/devices/block2mtd.c
50
static struct page *page_read(struct address_space *mapping, pgoff_t index)
drivers/mtd/devices/block2mtd.c
60
pgoff_t index = to >> PAGE_SHIFT; // page index
drivers/mtd/nand/raw/nandsim.c
1312
pgoff_t index, start_index, end_index;
drivers/nvdimm/pmem.c
242
__weak long __pmem_direct_access(struct pmem_device *pmem, pgoff_t pgoff,
drivers/nvdimm/pmem.c
293
static int pmem_dax_zero_page_range(struct dax_device *dax_dev, pgoff_t pgoff,
drivers/nvdimm/pmem.c
304
pgoff_t pgoff, long nr_pages, enum dax_access_mode mode,
drivers/nvdimm/pmem.c
325
static size_t pmem_recovery_write(struct dax_device *dax_dev, pgoff_t pgoff,
drivers/nvdimm/pmem.h
30
long __pmem_direct_access(struct pmem_device *pmem, pgoff_t pgoff,
drivers/s390/block/dcssblk.c
32
static long dcssblk_dax_direct_access(struct dax_device *dax_dev, pgoff_t pgoff,
drivers/s390/block/dcssblk.c
47
pgoff_t pgoff, size_t nr_pages)
drivers/s390/block/dcssblk.c
934
__dcssblk_direct_access(struct dcssblk_dev_info *dev_info, pgoff_t pgoff,
drivers/s390/block/dcssblk.c
950
dcssblk_dax_direct_access(struct dax_device *dax_dev, pgoff_t pgoff,
drivers/staging/vme_user/vme_user.c
449
static int vme_user_vm_mapped(unsigned long start, unsigned long end, pgoff_t pgoff,
drivers/target/target_core_user.c
1863
static int tcmu_vma_mapped(unsigned long start, unsigned long end, pgoff_t pgoff,
drivers/vfio/pci/nvgrace-gpu/main.c
121
pgoff_t *pgoff)
drivers/vfio/pci/nvgrace-gpu/main.c
126
pgoff_t vma_offset_in_region = vma->vm_pgoff &
drivers/vfio/pci/nvgrace-gpu/main.c
128
pgoff_t pfn_offset_in_region;
drivers/vfio/pci/nvgrace-gpu/main.c
96
pgoff_t *pfn_offset_in_region)
drivers/video/fbdev/core/fb_defio.c
398
pgoff_t pgoff = pageref->offset >> PAGE_SHIFT;
fs/affs/file.c
659
pgoff_t index;
fs/afs/file.c
30
static vm_fault_t afs_vm_map_pages(struct vm_fault *vmf, pgoff_t start_pgoff, pgoff_t end_pgoff);
fs/afs/file.c
31
static int afs_mapped(unsigned long start, unsigned long end, pgoff_t pgoff,
fs/afs/file.c
510
static int afs_mapped(unsigned long start, unsigned long end, pgoff_t pgoff,
fs/afs/file.c
535
static vm_fault_t afs_vm_map_pages(struct vm_fault *vmf, pgoff_t start_pgoff, pgoff_t end_pgoff)
fs/aio.c
466
pgoff_t idx;
fs/aio.c
487
if (idx < (pgoff_t)ctx->nr_pages) {
fs/btrfs/compression.c
1537
pgoff_t index, index_end;
fs/btrfs/compression.c
254
pgoff_t index = cb->start >> PAGE_SHIFT;
fs/btrfs/compression.c
255
const pgoff_t end_index = (cb->start + cb->len - 1) >> PAGE_SHIFT;
fs/btrfs/compression.c
375
pgoff_t end_index;
fs/btrfs/compression.c
424
pgoff_t page_end;
fs/btrfs/compression.c
425
pgoff_t pg_index = cur >> PAGE_SHIFT;
fs/btrfs/defrag.c
1371
pgoff_t start_index;
fs/btrfs/defrag.c
845
static struct folio *defrag_prepare_one_folio(struct btrfs_inode *inode, pgoff_t index)
fs/btrfs/extent_io.c
1838
const pgoff_t end_index = i_size >> PAGE_SHIFT;
fs/btrfs/extent_io.c
2437
pgoff_t index;
fs/btrfs/extent_io.c
2438
pgoff_t end; /* Inclusive */
fs/btrfs/extent_io.c
2439
pgoff_t done_index;
fs/btrfs/extent_io.c
271
pgoff_t index = start >> PAGE_SHIFT;
fs/btrfs/extent_io.c
272
pgoff_t end_index = end >> PAGE_SHIFT;
fs/btrfs/extent_io.c
309
pgoff_t index = start >> PAGE_SHIFT;
fs/btrfs/extent_io.c
310
pgoff_t end_index = end >> PAGE_SHIFT;
fs/btrfs/extent_io.c
3364
const pgoff_t index = eb->start >> PAGE_SHIFT;
fs/btrfs/file.c
2244
const pgoff_t start_index = page_lockstart >> PAGE_SHIFT;
fs/btrfs/file.c
2245
const pgoff_t end_index = (page_lockend - 1) >> PAGE_SHIFT;
fs/btrfs/file.c
2246
pgoff_t tmp = start_index;
fs/btrfs/file.c
858
const pgoff_t index = pos >> PAGE_SHIFT;
fs/btrfs/free-space-cache.c
370
pgoff_t last_index;
fs/btrfs/inode.c
404
pgoff_t index = offset >> PAGE_SHIFT;
fs/btrfs/inode.c
405
const pgoff_t end_index = (offset + bytes - 1) >> PAGE_SHIFT;
fs/btrfs/inode.c
4987
const pgoff_t index = (start >> PAGE_SHIFT);
fs/btrfs/inode.c
5058
pgoff_t index = (offset >> PAGE_SHIFT);
fs/btrfs/inode.c
802
const pgoff_t end_index = end >> PAGE_SHIFT;
fs/btrfs/inode.c
806
for (pgoff_t index = start >> PAGE_SHIFT; index <= end_index; index++) {
fs/btrfs/ioctl.c
4543
pgoff_t index;
fs/btrfs/relocation.c
2784
const pgoff_t last_index = (cluster->end - offset) >> PAGE_SHIFT;
fs/btrfs/relocation.c
2785
const pgoff_t index = orig_file_offset >> PAGE_SHIFT;
fs/btrfs/send.c
5204
const pgoff_t last_index = ((end - 1) >> PAGE_SHIFT);
fs/btrfs/send.c
5213
pgoff_t index = (cur >> PAGE_SHIFT);
fs/btrfs/tests/extent-io-tests.c
159
for (pgoff_t index = 0; index < (total_dirty >> PAGE_SHIFT); index++) {
fs/btrfs/tests/extent-io-tests.c
26
pgoff_t index = start >> PAGE_SHIFT;
fs/btrfs/tests/extent-io-tests.c
27
pgoff_t end_index = end >> PAGE_SHIFT;
fs/btrfs/verity.c
704
pgoff_t index)
fs/buffer.c
1633
pgoff_t index = ((loff_t)block << blkbits) / PAGE_SIZE;
fs/buffer.c
1634
pgoff_t end;
fs/buffer.c
185
pgoff_t index;
fs/buffer.c
2137
pgoff_t index = pos >> PAGE_SHIFT;
fs/buffer.c
2396
pgoff_t index, curidx;
fs/buffer.c
2547
pgoff_t index = from >> PAGE_SHIFT;
fs/buffer.c
940
pgoff_t index, unsigned size, gfp_t gfp)
fs/ceph/addr.c
1603
(pgoff_t)-1,
fs/ceph/addr.c
593
pgoff_t start_index;
fs/ceph/addr.c
594
pgoff_t index;
fs/ceph/addr.c
595
pgoff_t end;
fs/ceph/addr.c
598
pgoff_t strip_unit_end;
fs/ceph/dir.c
139
pgoff_t ptr_pgoff = ptr_pos >> PAGE_SHIFT;
fs/ceph/inode.c
1920
pgoff_t pgoff = ctl->index / nsize;
fs/dax.c
1004
pgoff_t pgoff = dax_iomap_pgoff(&iter->iomap, iter->pos);
fs/dax.c
1203
pgoff_t end_index = wbc->range_end >> PAGE_SHIFT;
fs/dax.c
1242
pgoff_t pgoff = dax_iomap_pgoff(iomap, pos);
fs/dax.c
128
pgoff_t entry_start;
fs/dax.c
1482
pgoff_t pgoff = dax_iomap_pgoff(iomap, pos);
fs/dax.c
1523
pgoff_t pgoff = dax_iomap_pgoff(iomap, pos);
fs/dax.c
1636
pgoff_t pgoff = dax_iomap_pgoff(iomap, pos);
fs/dax.c
1939
pgoff_t max_pgoff)
fs/dax.c
1982
pgoff_t max_pgoff;
fs/dax.c
566
dax_entry_t dax_lock_mapping_entry(struct address_space *mapping, pgoff_t index,
fs/dax.c
609
void dax_unlock_mapping_entry(struct address_space *mapping, pgoff_t index,
fs/dax.c
759
pgoff_t start_idx = start >> PAGE_SHIFT;
fs/dax.c
760
pgoff_t end_idx;
fs/dax.c
815
pgoff_t index, bool trunc)
fs/dax.c
840
pgoff_t start, pgoff_t end)
fs/dax.c
872
int dax_delete_mapping_entry(struct address_space *mapping, pgoff_t index)
fs/dax.c
891
pgoff_t start_idx = start >> PAGE_SHIFT;
fs/dax.c
892
pgoff_t end_idx;
fs/dax.c
992
pgoff_t index)
fs/dax.c
997
static pgoff_t dax_iomap_pgoff(const struct iomap *iomap, loff_t pos)
fs/ecryptfs/crypto.c
301
pgoff_t page_index,
fs/ecryptfs/ecryptfs_kernel.h
652
pgoff_t page_index,
fs/ecryptfs/mmap.c
247
pgoff_t index = pos >> PAGE_SHIFT;
fs/ecryptfs/mmap.c
445
pgoff_t index = pos >> PAGE_SHIFT;
fs/ecryptfs/read_write.c
113
pgoff_t ecryptfs_page_idx = (pos >> PAGE_SHIFT);
fs/ecryptfs/read_write.c
247
pgoff_t page_index,
fs/erofs/data.c
31
pgoff_t index = (buf->off + offset) >> PAGE_SHIFT;
fs/erofs/dir.c
55
pgoff_t ra_pages = DIV_ROUND_UP_POW2(
fs/erofs/dir.c
57
pgoff_t nr_pages = DIV_ROUND_UP_POW2(dir->i_size, PAGE_SIZE);
fs/erofs/dir.c
74
pgoff_t idx = DIV_ROUND_UP_POW2(ctx->pos, PAGE_SIZE);
fs/erofs/dir.c
75
pgoff_t pages = min(nr_pages - idx, ra_pages);
fs/erofs/fscache.c
46
pgoff_t start_page = req->start / PAGE_SIZE;
fs/erofs/fscache.c
47
pgoff_t last_page = ((req->start + req->len) / PAGE_SIZE) - 1;
fs/erofs/internal.h
356
pgoff_t index)
fs/erofs/zdata.c
1861
pgoff_t index = cur >> PAGE_SHIFT;
fs/erofs/zdata.c
539
pgoff_t poff = pcl->pos >> PAGE_SHIFT;
fs/ext4/dir.c
196
pgoff_t index = map.m_pblk << inode->i_blkbits >>
fs/ext4/inode.c
1298
pgoff_t index;
fs/ext4/inode.c
1721
pgoff_t index, end;
fs/ext4/inode.c
2306
pgoff_t start, end;
fs/ext4/inode.c
2623
pgoff_t index = mpd->start_pos >> PAGE_SHIFT;
fs/ext4/inode.c
2624
pgoff_t end = mpd->end_pos >> PAGE_SHIFT;
fs/ext4/inode.c
2766
pgoff_t writeback_index = 0;
fs/ext4/inode.c
3124
pgoff_t index;
fs/ext4/move_extent.c
61
pgoff_t index1, pgoff_t index2, size_t len,
fs/ext4/verity.c
363
pgoff_t index)
fs/ext4/verity.c
369
static void ext4_readahead_merkle_tree(struct inode *inode, pgoff_t index,
fs/f2fs/checkpoint.c
247
struct folio *f2fs_grab_meta_folio(struct f2fs_sb_info *sbi, pgoff_t index)
fs/f2fs/checkpoint.c
263
static struct folio *__get_meta_folio(struct f2fs_sb_info *sbi, pgoff_t index,
fs/f2fs/checkpoint.c
316
struct folio *f2fs_get_meta_folio(struct f2fs_sb_info *sbi, pgoff_t index)
fs/f2fs/checkpoint.c
321
struct folio *f2fs_get_meta_folio_retry(struct f2fs_sb_info *sbi, pgoff_t index)
fs/f2fs/checkpoint.c
338
struct folio *f2fs_get_tmp_folio(struct f2fs_sb_info *sbi, pgoff_t index)
fs/f2fs/checkpoint.c
531
void f2fs_ra_meta_pages_cond(struct f2fs_sb_info *sbi, pgoff_t index,
fs/f2fs/checkpoint.c
621
pgoff_t index = 0, prev = ULONG_MAX;
fs/f2fs/checkpoint.c
633
(pgoff_t)-1,
fs/f2fs/compress.c
1011
int f2fs_is_compressed_cluster(struct inode *inode, pgoff_t index)
fs/f2fs/compress.c
1019
bool f2fs_is_sparse_cluster(struct inode *inode, pgoff_t index)
fs/f2fs/compress.c
1087
struct page **pagep, pgoff_t index, void **fsdata)
fs/f2fs/compress.c
1094
pgoff_t start_idx = start_idx_of_cluster(cc);
fs/f2fs/compress.c
1175
struct page **pagep, pgoff_t index, void **fsdata)
fs/f2fs/compress.c
1191
pgoff_t index, unsigned copied)
fs/f2fs/compress.c
1218
pgoff_t start_idx = from >> (PAGE_SHIFT + log_cluster_size) <<
fs/f2fs/compress.c
1295
pgoff_t start_idx = start_idx_of_cluster(cc);
fs/f2fs/compress.c
1697
pgoff_t start_idx = start_idx_of_cluster(cc);
fs/f2fs/compress.c
1995
pgoff_t index = 0;
fs/f2fs/compress.c
1996
pgoff_t end = MAX_BLKADDR(sbi);
fs/f2fs/compress.c
59
static unsigned int offset_in_cluster(struct compress_ctx *cc, pgoff_t index)
fs/f2fs/compress.c
64
static pgoff_t cluster_idx(struct compress_ctx *cc, pgoff_t index)
fs/f2fs/compress.c
69
static pgoff_t start_idx_of_cluster(struct compress_ctx *cc)
fs/f2fs/compress.c
820
static bool is_page_in_cluster(struct compress_ctx *cc, pgoff_t index)
fs/f2fs/compress.c
837
bool f2fs_cluster_can_merge_page(struct compress_ctx *cc, pgoff_t index)
fs/f2fs/compress.c
87
struct inode *inode, pgoff_t index, void *data)
fs/f2fs/data.c
1105
pgoff_t first_idx, bool for_write)
fs/f2fs/data.c
1245
int f2fs_reserve_block(struct dnode_of_data *dn, pgoff_t index)
fs/f2fs/data.c
1262
pgoff_t idx)
fs/f2fs/data.c
1269
struct folio *f2fs_get_read_data_folio(struct inode *inode, pgoff_t index,
fs/f2fs/data.c
1270
blk_opf_t op_flags, bool for_write, pgoff_t *next_pgofs)
fs/f2fs/data.c
1282
pgoff_t folio_index = mapping_align_index(mapping, index);
fs/f2fs/data.c
1353
struct folio *f2fs_find_data_folio(struct inode *inode, pgoff_t index,
fs/f2fs/data.c
1354
pgoff_t *next_pgofs)
fs/f2fs/data.c
1387
struct folio *f2fs_get_lock_data_folio(struct inode *inode, pgoff_t index,
fs/f2fs/data.c
1416
struct folio *ifolio, pgoff_t index, bool new_i_size)
fs/f2fs/data.c
1522
int f2fs_get_block_locked(struct dnode_of_data *dn, pgoff_t index)
fs/f2fs/data.c
1539
pgoff_t pgoff)
fs/f2fs/data.c
1564
pgoff_t pgoff = (pgoff_t)map->m_lblk;
fs/f2fs/data.c
1571
map->m_len = min((pgoff_t)maxblocks, ei.fofs + ei.len - pgoff);
fs/f2fs/data.c
1625
pgoff_t pgofs, end_offset, end;
fs/f2fs/data.c
1652
pgofs = (pgoff_t)map->m_lblk;
fs/f2fs/data.c
1996
pgoff_t next_pgofs;
fs/f2fs/data.c
2174
pgoff_t index = folio->index;
fs/f2fs/data.c
2471
pgoff_t index, offset, next_pgofs = 0;
fs/f2fs/data.c
2632
pgoff_t nc_cluster_idx = NULL_CLUSTER;
fs/f2fs/data.c
2633
pgoff_t index;
fs/f2fs/data.c
2635
pgoff_t next_pgofs = 0;
fs/f2fs/data.c
3067
const pgoff_t end_index = ((unsigned long long)i_size)
fs/f2fs/data.c
3255
pgoff_t index;
fs/f2fs/data.c
3256
pgoff_t end; /* Inclusive */
fs/f2fs/data.c
3257
pgoff_t done_index;
fs/f2fs/data.c
3666
pgoff_t index = folio->index;
fs/f2fs/data.c
3750
static int __find_data_block(struct inode *inode, pgoff_t index,
fs/f2fs/data.c
3777
static int __reserve_data_block(struct inode *inode, pgoff_t index,
fs/f2fs/data.c
3814
pgoff_t index = folio->index;
fs/f2fs/data.c
3859
pgoff_t index = pos >> PAGE_SHIFT;
fs/f2fs/data.c
4509
pgoff_t next_pgofs = 0;
fs/f2fs/data.c
521
pgoff_t first_idx,
fs/f2fs/data.c
536
pgoff_t next_idx,
fs/f2fs/dir.c
1086
pgoff_t next_pgofs;
fs/f2fs/dir.c
1098
min(npages - n, (pgoff_t)MAX_DIR_RA_PAGES));
fs/f2fs/dir.c
286
pgoff_t next_pgofs;
fs/f2fs/dir.c
886
pgoff_t index = folio->index;
fs/f2fs/dir.c
941
pgoff_t next_pgofs;
fs/f2fs/extent_cache.c
1048
bool f2fs_lookup_read_extent_cache(struct inode *inode, pgoff_t pgofs,
fs/f2fs/extent_cache.c
1057
bool f2fs_lookup_read_extent_cache_block(struct inode *inode, pgoff_t index,
fs/f2fs/extent_cache.c
1074
pgoff_t fofs, block_t blkaddr, unsigned int len)
fs/f2fs/extent_cache.c
1097
bool f2fs_lookup_age_extent_cache(struct inode *inode, pgoff_t pgofs,
fs/f2fs/extent_cache.c
1112
pgoff_t fofs, unsigned int len)
fs/f2fs/extent_cache.c
401
pgoff_t fofs, unsigned int len)
fs/f2fs/extent_cache.c
403
if (fofs < (pgoff_t)et->largest.fofs + et->largest.len &&
fs/f2fs/extent_cache.c
480
static bool __lookup_extent_tree(struct inode *inode, pgoff_t pgofs,
fs/f2fs/extent_cache.c
498
(pgoff_t)et->largest.fofs + et->largest.len > pgofs) {
fs/f2fs/extent_cache.c
831
pgoff_t fofs, block_t blkaddr, unsigned int llen,
fs/f2fs/f2fs.h
1657
pgoff_t cluster_idx; /* cluster index number */
fs/f2fs/f2fs.h
1688
pgoff_t cluster_idx; /* cluster index number */
fs/f2fs/f2fs.h
1768
pgoff_t page_eio_ofs[NR_PAGE_TYPE]; /* EIO page offset */
fs/f2fs/f2fs.h
2202
pgoff_t index)
fs/f2fs/f2fs.h
2204
pgoff_t idx_in_folio = index % folio_nr_pages(folio);
fs/f2fs/f2fs.h
3025
pgoff_t index, bool for_write)
fs/f2fs/f2fs.h
3057
struct address_space *mapping, pgoff_t index,
fs/f2fs/f2fs.h
3783
int f2fs_truncate_hole(struct inode *inode, pgoff_t pg_start, pgoff_t pg_end);
fs/f2fs/f2fs.h
3931
pgoff_t f2fs_get_next_page_offset(struct dnode_of_data *dn, pgoff_t pgofs);
fs/f2fs/f2fs.h
3932
int f2fs_get_dnode_of_data(struct dnode_of_data *dn, pgoff_t index, int mode);
fs/f2fs/f2fs.h
3933
int f2fs_truncate_inode_blocks(struct inode *inode, pgoff_t from);
fs/f2fs/f2fs.h
3941
struct folio *f2fs_get_node_folio(struct f2fs_sb_info *sbi, pgoff_t nid,
fs/f2fs/f2fs.h
3944
struct folio *folio, pgoff_t nid,
fs/f2fs/f2fs.h
3946
struct folio *f2fs_get_inode_folio(struct f2fs_sb_info *sbi, pgoff_t ino);
fs/f2fs/f2fs.h
3947
struct folio *f2fs_get_xnode_folio(struct f2fs_sb_info *sbi, pgoff_t xnid);
fs/f2fs/f2fs.h
4087
struct folio *f2fs_grab_meta_folio(struct f2fs_sb_info *sbi, pgoff_t index);
fs/f2fs/f2fs.h
4088
struct folio *f2fs_get_meta_folio(struct f2fs_sb_info *sbi, pgoff_t index);
fs/f2fs/f2fs.h
4089
struct folio *f2fs_get_meta_folio_retry(struct f2fs_sb_info *sbi, pgoff_t index);
fs/f2fs/f2fs.h
4090
struct folio *f2fs_get_tmp_folio(struct f2fs_sb_info *sbi, pgoff_t index);
fs/f2fs/f2fs.h
4097
void f2fs_ra_meta_pages_cond(struct f2fs_sb_info *sbi, pgoff_t index,
fs/f2fs/f2fs.h
4160
int f2fs_get_block_locked(struct dnode_of_data *dn, pgoff_t index);
fs/f2fs/f2fs.h
4161
int f2fs_reserve_block(struct dnode_of_data *dn, pgoff_t index);
fs/f2fs/f2fs.h
4162
struct folio *f2fs_get_read_data_folio(struct inode *inode, pgoff_t index,
fs/f2fs/f2fs.h
4163
blk_opf_t op_flags, bool for_write, pgoff_t *next_pgofs);
fs/f2fs/f2fs.h
4164
struct folio *f2fs_find_data_folio(struct inode *inode, pgoff_t index,
fs/f2fs/f2fs.h
4165
pgoff_t *next_pgofs);
fs/f2fs/f2fs.h
4166
struct folio *f2fs_get_lock_data_folio(struct inode *inode, pgoff_t index,
fs/f2fs/f2fs.h
4169
struct folio *ifolio, pgoff_t index, bool new_i_size);
fs/f2fs/f2fs.h
4541
bool f2fs_lookup_read_extent_cache(struct inode *inode, pgoff_t pgofs,
fs/f2fs/f2fs.h
4543
bool f2fs_lookup_read_extent_cache_block(struct inode *inode, pgoff_t index,
fs/f2fs/f2fs.h
4547
pgoff_t fofs, block_t blkaddr, unsigned int len);
fs/f2fs/f2fs.h
4553
bool f2fs_lookup_age_extent_cache(struct inode *inode, pgoff_t pgofs,
fs/f2fs/f2fs.h
4557
pgoff_t fofs, unsigned int len);
fs/f2fs/f2fs.h
4623
struct page **pagep, pgoff_t index, void **fsdata);
fs/f2fs/f2fs.h
4625
pgoff_t index, unsigned copied);
fs/f2fs/f2fs.h
4636
bool f2fs_cluster_can_merge_page(struct compress_ctx *cc, pgoff_t index);
fs/f2fs/f2fs.h
4645
int f2fs_is_compressed_cluster(struct inode *inode, pgoff_t index);
fs/f2fs/f2fs.h
4646
bool f2fs_is_sparse_cluster(struct inode *inode, pgoff_t index);
fs/f2fs/f2fs.h
4648
pgoff_t fofs, block_t blkaddr,
fs/f2fs/f2fs.h
4732
pgoff_t index) { return 0; }
fs/f2fs/f2fs.h
4735
pgoff_t index) { return true; }
fs/f2fs/f2fs.h
4738
pgoff_t fofs, block_t blkaddr,
fs/f2fs/f2fs.h
5060
pgoff_t ofs = folio->index;
fs/f2fs/f2fs.h
830
pgoff_t *m_next_pgofs; /* point next possible non-hole pgofs */
fs/f2fs/f2fs.h
831
pgoff_t *m_next_extent; /* point to next possible extent */
fs/f2fs/f2fs.h
972
pgoff_t donate_start, donate_end; /* inclusive */
fs/f2fs/file.c
1214
static int fill_zero(struct inode *inode, pgoff_t index,
fs/f2fs/file.c
1240
int f2fs_truncate_hole(struct inode *inode, pgoff_t pg_start, pgoff_t pg_end)
fs/f2fs/file.c
1246
pgoff_t end_offset, count;
fs/f2fs/file.c
1274
pgoff_t pg_start, pg_end;
fs/f2fs/file.c
1336
int *do_replace, pgoff_t off, pgoff_t len)
fs/f2fs/file.c
1350
done = min((pgoff_t)ADDRS_PER_BLOCK(inode) -
fs/f2fs/file.c
1357
done = min((pgoff_t)ADDRS_PER_PAGE(dn.node_folio, inode) -
fs/f2fs/file.c
1391
int *do_replace, pgoff_t off, int len)
fs/f2fs/file.c
1416
pgoff_t src, pgoff_t dst, pgoff_t len, bool full)
fs/f2fs/file.c
1419
pgoff_t i = 0;
fs/f2fs/file.c
1432
pgoff_t ilen;
fs/f2fs/file.c
1445
ilen = min((pgoff_t)
fs/f2fs/file.c
1503
struct inode *dst_inode, pgoff_t src, pgoff_t dst,
fs/f2fs/file.c
1504
pgoff_t len, bool full)
fs/f2fs/file.c
1508
pgoff_t olen;
fs/f2fs/file.c
1512
olen = min((pgoff_t)4 * ADDRS_PER_BLOCK(src_inode), len);
fs/f2fs/file.c
1558
pgoff_t nrpages = DIV_ROUND_UP(i_size_read(inode), PAGE_SIZE);
fs/f2fs/file.c
1559
pgoff_t start = offset >> PAGE_SHIFT;
fs/f2fs/file.c
1560
pgoff_t end = (offset + len) >> PAGE_SHIFT;
fs/f2fs/file.c
1620
static int f2fs_do_zero_range(struct dnode_of_data *dn, pgoff_t start,
fs/f2fs/file.c
1621
pgoff_t end)
fs/f2fs/file.c
1624
pgoff_t index = start;
fs/f2fs/file.c
1678
pgoff_t index, pg_start, pg_end;
fs/f2fs/file.c
1725
pgoff_t end;
fs/f2fs/file.c
1788
pgoff_t nr, pg_start, pg_end, delta, idx;
fs/f2fs/file.c
1876
pgoff_t pg_start, pg_end;
fs/f2fs/file.c
1962
pgoff_t last_off;
fs/f2fs/file.c
2895
pgoff_t pg_start, pg_end, next_pgofs;
fs/f2fs/file.c
2905
pg_end = min_t(pgoff_t,
fs/f2fs/file.c
2934
if ((pgoff_t)ei.fofs + ei.len >= pg_end)
fs/f2fs/file.c
2990
pgoff_t idx;
fs/f2fs/file.c
3620
pgoff_t m_next_extent;
fs/f2fs/file.c
3793
static int release_compress_blocks(struct dnode_of_data *dn, pgoff_t count)
fs/f2fs/file.c
3852
pgoff_t page_idx = 0, last_idx;
fs/f2fs/file.c
3904
pgoff_t end_offset, count;
fs/f2fs/file.c
3963
static int reserve_compress_blocks(struct dnode_of_data *dn, pgoff_t count,
fs/f2fs/file.c
4048
pgoff_t page_idx = 0, last_idx;
fs/f2fs/file.c
4083
pgoff_t end_offset, count;
fs/f2fs/file.c
4147
pgoff_t off, block_t block, block_t len, u32 flags)
fs/f2fs/file.c
4182
pgoff_t index, pg_end, prev_index = 0;
fs/f2fs/file.c
422
struct dnode_of_data *dn, pgoff_t index, int whence)
fs/f2fs/file.c
4253
pgoff_t end_offset, count;
fs/f2fs/file.c
4434
static int redirty_blocks(struct inode *inode, pgoff_t page_idx, int len)
fs/f2fs/file.c
4439
pgoff_t redirty_idx = page_idx;
fs/f2fs/file.c
4479
pgoff_t page_idx = 0, last_idx, cluster_idx;
fs/f2fs/file.c
4560
pgoff_t page_idx = 0, last_idx, cluster_idx;
fs/f2fs/file.c
460
pgoff_t pgofs, end_offset;
fs/f2fs/file.c
482
pgofs = (pgoff_t)(offset >> PAGE_SHIFT);
fs/f2fs/file.c
720
pgoff_t fofs;
fs/f2fs/file.c
742
pgoff_t index = from >> PAGE_SHIFT;
fs/f2fs/file.c
779
pgoff_t free_from;
fs/f2fs/file.c
791
free_from = (pgoff_t)F2FS_BLK_ALIGN(from);
fs/f2fs/gc.c
1216
static int ra_data_block(struct inode *inode, pgoff_t index)
fs/f2fs/node.c
1181
int f2fs_truncate_inode_blocks(struct inode *inode, pgoff_t from)
fs/f2fs/node.c
147
pgoff_t dst_off;
fs/f2fs/node.c
1515
struct folio *folio, pgoff_t nid,
fs/f2fs/node.c
1561
static struct folio *__get_node_folio(struct f2fs_sb_info *sbi, pgoff_t nid,
fs/f2fs/node.c
1615
struct folio *f2fs_get_node_folio(struct f2fs_sb_info *sbi, pgoff_t nid,
fs/f2fs/node.c
1621
struct folio *f2fs_get_inode_folio(struct f2fs_sb_info *sbi, pgoff_t ino)
fs/f2fs/node.c
1626
struct folio *f2fs_get_xnode_folio(struct f2fs_sb_info *sbi, pgoff_t xnid)
fs/f2fs/node.c
1677
pgoff_t index;
fs/f2fs/node.c
1686
(pgoff_t)-1, PAGECACHE_TAG_DIRTY,
fs/f2fs/node.c
1885
pgoff_t index;
fs/f2fs/node.c
1904
(pgoff_t)-1, PAGECACHE_TAG_DIRTY,
fs/f2fs/node.c
2043
pgoff_t index = 0;
fs/f2fs/node.c
2050
(pgoff_t)-1, PAGECACHE_TAG_DIRTY,
fs/f2fs/node.c
2086
pgoff_t index;
fs/f2fs/node.c
2099
&index, (pgoff_t)-1, PAGECACHE_TAG_DIRTY,
fs/f2fs/node.c
573
pgoff_t index;
fs/f2fs/node.c
686
pgoff_t f2fs_get_next_page_offset(struct dnode_of_data *dn, pgoff_t pgofs)
fs/f2fs/node.c
694
pgoff_t base = 0;
fs/f2fs/node.c
804
int f2fs_get_dnode_of_data(struct dnode_of_data *dn, pgoff_t index, int mode)
fs/f2fs/node.c
912
pgoff_t fofs = index;
fs/f2fs/node.c
957
pgoff_t index;
fs/f2fs/node.h
196
static inline pgoff_t current_nat_addr(struct f2fs_sb_info *sbi, nid_t start)
fs/f2fs/node.h
199
pgoff_t block_off;
fs/f2fs/node.h
200
pgoff_t block_addr;
fs/f2fs/node.h
209
block_addr = (pgoff_t)(nm_i->nat_blkaddr +
fs/f2fs/node.h
219
static inline pgoff_t next_nat_addr(struct f2fs_sb_info *sbi,
fs/f2fs/node.h
220
pgoff_t block_addr)
fs/f2fs/segment.c
224
static int __replace_atomic_write_block(struct inode *inode, pgoff_t index,
fs/f2fs/segment.c
288
pgoff_t start_index = 0;
fs/f2fs/segment.c
317
pgoff_t len = DIV_ROUND_UP(i_size_read(inode), PAGE_SIZE);
fs/f2fs/segment.c
318
pgoff_t off = 0, blen, index;
fs/f2fs/segment.c
324
blen = min_t(pgoff_t, ADDRS_PER_BLOCK(cow_inode), len);
fs/f2fs/segment.c
337
blen = min((pgoff_t)ADDRS_PER_PAGE(dn.node_folio, cow_inode),
fs/f2fs/segment.c
3654
static int __get_age_segment_type(struct inode *inode, pgoff_t pgofs)
fs/f2fs/segment.c
4576
pgoff_t src_off, dst_off;
fs/f2fs/segment.h
202
pgoff_t index;
fs/f2fs/segment.h
894
static inline pgoff_t current_sit_addr(struct f2fs_sb_info *sbi,
fs/f2fs/segment.h
916
static inline pgoff_t next_sit_addr(struct f2fs_sb_info *sbi,
fs/f2fs/segment.h
917
pgoff_t block_addr)
fs/f2fs/shrinker.c
168
pgoff_t npages = reclaim_caches_kb >> (PAGE_SHIFT - 10);
fs/f2fs/shrinker.c
171
pgoff_t len;
fs/f2fs/super.c
3840
pgoff_t index, bool update)
fs/f2fs/super.c
3873
struct folio *folio, pgoff_t index)
fs/f2fs/super.c
3963
struct folio *folio, pgoff_t index)
fs/f2fs/super.c
4549
pgoff_t index;
fs/f2fs/verity.c
259
pgoff_t index)
fs/f2fs/verity.c
265
static void f2fs_readahead_merkle_tree(struct inode *inode, pgoff_t index,
fs/fuse/dev.c
1767
pgoff_t index;
fs/fuse/dev.c
1864
pgoff_t index;
fs/fuse/file.c
1258
pgoff_t index = pos >> PAGE_SHIFT;
fs/fuse/file.c
1655
pgoff_t idx_from = pos >> PAGE_SHIFT;
fs/fuse/file.c
1656
pgoff_t idx_to = (pos + count - 1) >> PAGE_SHIFT;
fs/fuse/inode.c
561
pgoff_t pg_start;
fs/fuse/inode.c
562
pgoff_t pg_end;
fs/fuse/readdir.c
37
pgoff_t index;
fs/fuse/readdir.c
452
pgoff_t index;
fs/fuse/virtio_fs.c
1006
static long virtio_fs_direct_access(struct dax_device *dax_dev, pgoff_t pgoff,
fs/fuse/virtio_fs.c
1022
pgoff_t pgoff, size_t nr_pages)
fs/gfs2/aops.c
193
pgoff_t *done_index)
fs/gfs2/aops.c
289
pgoff_t writeback_index;
fs/gfs2/aops.c
290
pgoff_t index;
fs/gfs2/aops.c
291
pgoff_t end;
fs/gfs2/aops.c
292
pgoff_t done_index;
fs/gfs2/dir.c
1493
f_ra->start = max((pgoff_t)index, f_ra->start);
fs/hfsplus/super.c
522
(last_fs_page > (pgoff_t)(~0ULL))) {
fs/hostfs/hostfs_kern.c
453
pgoff_t index = pos >> PAGE_SHIFT;
fs/hugetlbfs/inode.c
372
static unsigned long vma_offset_start(struct vm_area_struct *vma, pgoff_t start)
fs/hugetlbfs/inode.c
382
static unsigned long vma_offset_end(struct vm_area_struct *vma, pgoff_t end)
fs/hugetlbfs/inode.c
401
struct folio *folio, pgoff_t index)
fs/hugetlbfs/inode.c
409
pgoff_t start, end;
fs/hugetlbfs/inode.c
483
hugetlb_vmdelete_list(struct rb_root_cached *root, pgoff_t start, pgoff_t end,
fs/hugetlbfs/inode.c
519
pgoff_t index, bool truncate_op)
fs/hugetlbfs/inode.c
576
const pgoff_t end = lend >> PAGE_SHIFT;
fs/hugetlbfs/inode.c
578
pgoff_t next, index;
fs/hugetlbfs/inode.c
628
pgoff_t pgoff;
fs/hugetlbfs/inode.c
649
pgoff_t idx = start >> huge_page_shift(h);
fs/hugetlbfs/inode.c
731
pgoff_t start, index, end;
fs/iomap/buffered-io.c
1604
pgoff_t pstart = *start >> PAGE_SHIFT;
fs/iomap/buffered-io.c
1605
pgoff_t pend = (end - 1) >> PAGE_SHIFT;
fs/iomap/buffered-io.c
1809
pgoff_t end_index = isize >> PAGE_SHIFT;
fs/isofs/compress.c
313
pgoff_t index = folio->index, end_index;
fs/jffs2/file.c
137
pgoff_t index = pos >> PAGE_SHIFT;
fs/libfs.c
1625
(last_fs_page > (pgoff_t)(~0ULL))) {
fs/netfs/buffered_read.c
630
pgoff_t index = pos >> PAGE_SHIFT;
fs/netfs/buffered_write.c
40
pgoff_t index = pos / PAGE_SIZE;
fs/netfs/direct_write.c
35
pgoff_t first = wreq->start >> PAGE_SHIFT;
fs/netfs/direct_write.c
36
pgoff_t last = (wreq->start + wreq->transferred - 1) >> PAGE_SHIFT;
fs/netfs/fscache_io.c
176
pgoff_t first = start / PAGE_SIZE;
fs/netfs/fscache_io.c
177
pgoff_t last = (start + len - 1) / PAGE_SIZE;
fs/netfs/iterator.c
198
pgoff_t index = pos / PAGE_SIZE;
fs/nfs/blocklayout/blocklayout.c
887
static u64 pnfs_num_cont_bytes(struct inode *inode, pgoff_t idx)
fs/nfs/blocklayout/blocklayout.c
890
pgoff_t end;
fs/nfs/dir.c
160
pgoff_t folio_index;
fs/nfs/dir.c
161
pgoff_t folio_index_max;
fs/nfs/dir.c
360
static pgoff_t nfs_readdir_folio_cookie_hash(u64 cookie)
fs/nfs/dir.c
402
pgoff_t index = nfs_readdir_folio_cookie_hash(cookie);
fs/nfs/dir.c
446
pgoff_t index = nfs_readdir_folio_cookie_hash(cookie);
fs/nfs/fscache.c
307
pgoff_t start, last;
fs/nfs/internal.h
875
pgoff_t index = folio->index >> folio_order(folio);
fs/nfs/internal.h
876
pgoff_t end_index = (i_size - 1) >> folio_shift(folio);
fs/nfs/nfstrace.h
352
pgoff_t page_index,
fs/nfs/nfstrace.h
365
__field(pgoff_t, index)
fs/nfs/nfstrace.h
405
pgoff_t page_index, \
fs/nfs/pagelist.c
1431
void nfs_pageio_cond_complete(struct nfs_pageio_descriptor *desc, pgoff_t index)
fs/nfs/pagelist.c
378
unsigned int pgbase, pgoff_t index,
fs/nfs/write.c
193
pgoff_t end_index;
fs/nilfs2/btnode.c
163
pgoff_t index = folio->index;
fs/nilfs2/btree.c
2174
pgoff_t index = 0;
fs/nilfs2/btree.c
2184
while (filemap_get_folios_tag(btcache, &index, (pgoff_t)-1,
fs/nilfs2/mdt.c
356
pgoff_t index = block >> (PAGE_SHIFT - inode->i_blkbits);
fs/nilfs2/page.c
248
pgoff_t index = 0;
fs/nilfs2/page.c
253
if (!filemap_get_folios_tag(smap, &index, (pgoff_t)-1,
fs/nilfs2/page.c
29
unsigned long block, pgoff_t index, int blkbits,
fs/nilfs2/page.c
303
pgoff_t start = 0;
fs/nilfs2/page.c
313
pgoff_t index = folio->index;
fs/nilfs2/page.c
365
pgoff_t index = 0;
fs/nilfs2/page.c
369
while (filemap_get_folios_tag(mapping, &index, (pgoff_t)-1,
fs/nilfs2/page.c
511
pgoff_t index;
fs/nilfs2/page.c
52
pgoff_t index = blkoff >> (PAGE_SHIFT - blkbits);
fs/nilfs2/segment.c
705
pgoff_t index = 0, last = ULONG_MAX;
fs/nilfs2/segment.c
770
pgoff_t index = 0;
fs/nilfs2/segment.c
777
(pgoff_t)-1, PAGECACHE_TAG_DIRTY, &fbatch)) {
fs/ntfs3/attrib.c
1573
pgoff_t index = vbo[i] >> PAGE_SHIFT;
fs/ntfs3/bitmap.c
565
pgoff_t idx = lbo >> PAGE_SHIFT;
fs/ntfs3/file.c
936
static int ntfs_get_frame_pages(struct address_space *mapping, pgoff_t index,
fs/ntfs3/file.c
994
pgoff_t index;
fs/ntfs3/frecord.c
1856
pgoff_t index, gfp_t gfp)
fs/ntfs3/frecord.c
1889
pgoff_t index;
fs/ntfs3/frecord.c
1979
pgoff_t index;
fs/ntfs3/fsntfs.c
1238
pgoff_t index = lbo >> PAGE_SHIFT;
fs/ntfs3/inode.c
1066
pgoff_t idx;
fs/ocfs2/mmap.c
55
pgoff_t last_index;
fs/ocfs2/ocfs2.h
890
static inline pgoff_t ocfs2_align_clusters_to_page_index(struct super_block *sb,
fs/ocfs2/ocfs2.h
894
pgoff_t index = clusters;
fs/ocfs2/ocfs2.h
897
index = (pgoff_t)clusters >> (PAGE_SHIFT - cbits);
fs/ocfs2/ocfs2.h
899
index = (pgoff_t)clusters << (cbits - PAGE_SHIFT);
fs/ocfs2/refcounttree.c
2905
pgoff_t page_index;
fs/proc/task_mmu.c
1962
pgoff_t offset;
fs/proc/task_mmu.c
3307
pgoff_t ilx;
fs/proc/vmcore.c
474
pgoff_t index = vmf->pgoff;
fs/smb/client/cifsfs.c
1310
pgoff_t index = pos / PAGE_SIZE;
fs/squashfs/block.c
197
pgoff_t index)
fs/squashfs/block.c
240
pgoff_t index = (read_start >> PAGE_SHIFT) + i;
fs/squashfs/page_actor.h
25
pgoff_t next_index;
fs/sync.c
245
if (sizeof(pgoff_t) == 4) {
fs/ubifs/file.c
1137
pgoff_t index = new_size >> PAGE_SHIFT;
fs/ubifs/file.c
210
pgoff_t index = pos >> PAGE_SHIFT;
fs/ubifs/file.c
415
pgoff_t index = pos >> PAGE_SHIFT;
fs/ubifs/file.c
600
pgoff_t end_index;
fs/ubifs/file.c
699
pgoff_t offset = folio1->index, end_index;
fs/ubifs/file.c
763
pgoff_t page_offset = offset + page_idx;
fs/ubifs/file.c
813
pgoff_t index = folio->index, last_page_read = ui->last_page_read;
fs/ubifs/ubifs.h
416
pgoff_t last_page_read;
fs/ubifs/ubifs.h
417
pgoff_t read_in_a_row;
fs/ufs/balloc.c
243
pgoff_t index, cur_index, last_index;
fs/ufs/util.c
198
pgoff_t index)
fs/ufs/util.h
271
struct folio *ufs_get_locked_folio(struct address_space *mapping, pgoff_t index);
fs/verity/pagecache.c
19
struct page *generic_read_merkle_tree_page(struct inode *inode, pgoff_t index)
fs/verity/pagecache.c
41
void generic_readahead_merkle_tree(struct inode *inode, pgoff_t index,
fs/verity/read_metadata.c
23
pgoff_t index, last_index;
fs/verity/verify.c
221
pgoff_t hpage_idx;
fs/verity/verify.c
51
void fsverity_readahead(struct fsverity_info *vi, pgoff_t index,
fs/verity/verify.c
68
pgoff_t start_idx = (level_start + next_start_hidx) >>
fs/verity/verify.c
70
pgoff_t end_idx = (level_start + next_end_hidx) >>
fs/xfs/xfs_notify_failure.c
39
static pgoff_t
fs/xfs/xfs_notify_failure.c
83
pgoff_t pgoff;
include/drm/ttm/ttm_backup.h
56
void ttm_backup_drop(struct file *backup, pgoff_t handle);
include/drm/ttm/ttm_backup.h
59
pgoff_t handle, bool intr, gfp_t additional_gfp);
include/drm/ttm/ttm_backup.h
63
bool writeback, pgoff_t idx, gfp_t page_gfp,
include/drm/ttm/ttm_bo.h
440
pgoff_t num_prefault);
include/drm/ttm/ttm_kmap_iter.h
32
struct iosys_map *dmap, pgoff_t i);
include/drm/ttm/ttm_resource.h
373
pgoff_t i;
include/drm/ttm/ttm_resource.h
374
pgoff_t end;
include/drm/ttm/ttm_resource.h
375
pgoff_t offs;
include/linux/dax.h
121
pgoff_t pgoff, void *addr, size_t bytes, struct iov_iter *i)
include/linux/dax.h
171
static inline struct page *dax_layout_busy_page_range(struct address_space *mapping, pgoff_t start, pgoff_t nr_pages)
include/linux/dax.h
245
long dax_direct_access(struct dax_device *dax_dev, pgoff_t pgoff, long nr_pages,
include/linux/dax.h
247
size_t dax_copy_from_iter(struct dax_device *dax_dev, pgoff_t pgoff, void *addr,
include/linux/dax.h
249
size_t dax_copy_to_iter(struct dax_device *dax_dev, pgoff_t pgoff, void *addr,
include/linux/dax.h
251
int dax_zero_page_range(struct dax_device *dax_dev, pgoff_t pgoff,
include/linux/dax.h
264
int dax_delete_mapping_entry(struct address_space *mapping, pgoff_t index);
include/linux/dax.h
268
pgoff_t index);
include/linux/dax.h
28
long (*direct_access)(struct dax_device *, pgoff_t, long,
include/linux/dax.h
31
int (*zero_page_range)(struct dax_device *, pgoff_t, size_t);
include/linux/dax.h
36
size_t (*recovery_write)(struct dax_device *dax_dev, pgoff_t pgoff,
include/linux/dax.h
63
size_t dax_recovery_write(struct dax_device *dax_dev, pgoff_t pgoff,
include/linux/device-mapper.h
157
typedef long (*dm_dax_direct_access_fn) (struct dm_target *ti, pgoff_t pgoff,
include/linux/device-mapper.h
160
typedef int (*dm_dax_zero_page_range_fn)(struct dm_target *ti, pgoff_t pgoff,
include/linux/device-mapper.h
168
typedef size_t (*dm_dax_recovery_write_fn)(struct dm_target *ti, pgoff_t pgoff,
include/linux/fs.h
1214
pgoff_t start;
include/linux/fs.h
1226
static inline int ra_has_index(struct file_ra_state *ra, pgoff_t index)
include/linux/fs.h
485
pgoff_t writeback_index;
include/linux/fsverity.h
102
pgoff_t index);
include/linux/fsverity.h
117
void (*readahead_merkle_tree)(struct inode *inode, pgoff_t index,
include/linux/fsverity.h
198
void fsverity_readahead(struct fsverity_info *vi, pgoff_t index,
include/linux/fsverity.h
260
static inline void fsverity_readahead(struct fsverity_info *vi, pgoff_t index,
include/linux/fsverity.h
314
struct page *generic_read_merkle_tree_page(struct inode *inode, pgoff_t index);
include/linux/fsverity.h
315
void generic_readahead_merkle_tree(struct inode *inode, pgoff_t index,
include/linux/gfp.h
325
struct mempolicy *mpol, pgoff_t ilx, int nid);
include/linux/gfp.h
338
struct mempolicy *mpol, pgoff_t ilx, int nid)
include/linux/hugetlb.h
1111
struct address_space *mapping, pgoff_t idx)
include/linux/hugetlb.h
162
u32 hugetlb_fault_mutex_hash(struct address_space *mapping, pgoff_t idx);
include/linux/hugetlb.h
717
pgoff_t idx);
include/linux/hugetlb.h
804
static inline pgoff_t hugetlb_linear_page_index(struct vm_area_struct *vma,
include/linux/hugetlb.h
834
struct address_space *mapping, pgoff_t idx)
include/linux/iov_iter.h
210
pgoff_t index = start / PAGE_SIZE;
include/linux/kvm_host.h
613
pgoff_t pgoff;
include/linux/memfd.h
11
struct folio *memfd_alloc_folio(struct file *memfd, pgoff_t idx);
include/linux/memfd.h
28
static inline struct folio *memfd_alloc_folio(struct file *memfd, pgoff_t idx)
include/linux/memory-failure.h
11
unsigned long pfn, pgoff_t *pgoff);
include/linux/mempolicy.h
119
pgoff_t start, end;
include/linux/mempolicy.h
129
pgoff_t idx);
include/linux/mempolicy.h
133
unsigned long addr, pgoff_t *ilx);
include/linux/mempolicy.h
135
unsigned long addr, int order, pgoff_t *ilx);
include/linux/mempolicy.h
224
mpol_shared_policy_lookup(struct shared_policy *sp, pgoff_t idx)
include/linux/mempolicy.h
230
unsigned long addr, int order, pgoff_t *ilx)
include/linux/mm.h
3146
pgoff_t start, pgoff_t nr, bool even_cows);
include/linux/mm.h
3166
pgoff_t start, pgoff_t nr, bool even_cows) { }
include/linux/mm.h
3239
pgoff_t *offset);
include/linux/mm.h
4172
pgoff_t start_pgoff, pgoff_t end_pgoff);
include/linux/mm.h
4964
int mf_dax_kill_procs(struct address_space *mapping, pgoff_t index,
include/linux/mm.h
5088
pgoff_t first_index, pgoff_t nr,
include/linux/mm.h
5089
pgoff_t bitmap_pgoff,
include/linux/mm.h
5091
pgoff_t *start,
include/linux/mm.h
5092
pgoff_t *end);
include/linux/mm.h
5095
pgoff_t first_index, pgoff_t nr);
include/linux/mm.h
719
pgoff_t pgoff; /* Logical page offset based on vma */
include/linux/mm.h
793
int (*mapped)(unsigned long start, unsigned long end, pgoff_t pgoff,
include/linux/mm.h
808
pgoff_t start_pgoff, pgoff_t end_pgoff);
include/linux/mm.h
851
unsigned long addr, pgoff_t *ilx);
include/linux/mm_types.h
105
pgoff_t __folio_index; /* Our offset within mapping. */
include/linux/mm_types.h
423
pgoff_t index;
include/linux/mm_types.h
586
pgoff_t pt_index;
include/linux/mm_types.h
841
pgoff_t pgoff;
include/linux/mm_types.h
907
pgoff_t pgoff;
include/linux/netfs.h
255
pgoff_t no_unlock_folio; /* Don't unlock this folio after read */
include/linux/nfs_fs.h
127
pgoff_t page_index;
include/linux/nfs_page.h
149
extern void nfs_pageio_cond_complete(struct nfs_pageio_descriptor *, pgoff_t);
include/linux/nfs_page.h
50
pgoff_t wb_index; /* Offset >> PAGE_SHIFT */
include/linux/pagemap.h
1001
struct folio *mapping_read_folio_gfp(struct address_space *, pgoff_t index,
include/linux/pagemap.h
1003
struct page *read_cache_page(struct address_space *, pgoff_t index,
include/linux/pagemap.h
1006
pgoff_t index, gfp_t gfp_mask);
include/linux/pagemap.h
1009
pgoff_t index, struct file *file)
include/linux/pagemap.h
1015
pgoff_t index, struct file *file)
include/linux/pagemap.h
1035
static inline pgoff_t page_pgoff(const struct folio *folio,
include/linux/pagemap.h
1063
static inline pgoff_t folio_pgoff(const struct folio *folio)
include/linux/pagemap.h
1068
static inline pgoff_t linear_page_index(const struct vm_area_struct *vma,
include/linux/pagemap.h
1071
pgoff_t pgoff;
include/linux/pagemap.h
1286
pgoff_t index, gfp_t gfp);
include/linux/pagemap.h
1288
pgoff_t index, gfp_t gfp);
include/linux/pagemap.h
1300
pgoff_t index, gfp_t gfp, void **shadowp);
include/linux/pagemap.h
1352
pgoff_t _index;
include/linux/pagemap.h
1393
struct file_ra_state *ra, struct file *file, pgoff_t index,
include/linux/pagemap.h
1509
static inline pgoff_t readahead_index(const struct readahead_control *rac)
include/linux/pagemap.h
1550
pgoff_t index = size >> PAGE_SHIFT;
include/linux/pagemap.h
22
pgoff_t start, pgoff_t end);
include/linux/pagemap.h
32
pgoff_t start, pgoff_t end);
include/linux/pagemap.h
497
static inline pgoff_t mapping_align_index(const struct address_space *mapping,
include/linux/pagemap.h
498
pgoff_t index)
include/linux/pagemap.h
679
pgoff_t page_cache_next_miss(struct address_space *mapping,
include/linux/pagemap.h
680
pgoff_t index, unsigned long max_scan);
include/linux/pagemap.h
681
pgoff_t page_cache_prev_miss(struct address_space *mapping,
include/linux/pagemap.h
682
pgoff_t index, unsigned long max_scan);
include/linux/pagemap.h
754
void *filemap_get_entry(struct address_space *mapping, pgoff_t index);
include/linux/pagemap.h
756
pgoff_t index, fgf_t fgf_flags, gfp_t gfp, struct mempolicy *policy);
include/linux/pagemap.h
757
struct page *pagecache_get_page(struct address_space *mapping, pgoff_t index,
include/linux/pagemap.h
761
pgoff_t index, fgf_t fgf_flags, gfp_t gfp)
include/linux/pagemap.h
780
struct address_space *mapping, pgoff_t index, size_t len)
include/linux/pagemap.h
805
pgoff_t index)
include/linux/pagemap.h
823
pgoff_t index)
include/linux/pagemap.h
841
pgoff_t index)
include/linux/pagemap.h
859
pgoff_t offset)
include/linux/pagemap.h
865
pgoff_t offset, fgf_t fgp_flags)
include/linux/pagemap.h
884
pgoff_t index)
include/linux/pagemap.h
909
pgoff_t index, gfp_t gfp_mask)
include/linux/pagemap.h
931
pgoff_t index)
include/linux/pagemap.h
944
static inline pgoff_t folio_next_index(const struct folio *folio)
include/linux/pagemap.h
970
static inline struct page *folio_file_page(struct folio *folio, pgoff_t index)
include/linux/pagemap.h
984
static inline bool folio_contains(const struct folio *folio, pgoff_t index)
include/linux/pagemap.h
990
unsigned filemap_get_folios(struct address_space *mapping, pgoff_t *start,
include/linux/pagemap.h
991
pgoff_t end, struct folio_batch *fbatch);
include/linux/pagemap.h
993
pgoff_t *start, pgoff_t end, struct folio_batch *fbatch);
include/linux/pagemap.h
994
unsigned filemap_get_folios_tag(struct address_space *mapping, pgoff_t *start,
include/linux/pagemap.h
995
pgoff_t end, xa_mark_t tag, struct folio_batch *fbatch);
include/linux/pagemap.h
997
pgoff_t *start, pgoff_t end, struct folio_batch *fbatch);
include/linux/pagemap.h
999
struct folio *read_cache_folio(struct address_space *, pgoff_t index,
include/linux/pagewalk.h
145
int walk_page_mapping(struct address_space *mapping, pgoff_t first_index,
include/linux/pagewalk.h
146
pgoff_t nr, const struct mm_walk_ops *ops,
include/linux/pgtable.h
1400
static inline void arch_swap_invalidate_page(int type, pgoff_t offset)
include/linux/rmap.h
867
pgoff_t pgoff;
include/linux/rmap.h
931
int mapping_wrprotect_range(struct address_space *mapping, pgoff_t pgoff,
include/linux/rmap.h
934
int pfn_mkclean_range(unsigned long pfn, unsigned long nr_pages, pgoff_t pgoff,
include/linux/shmem_fs.h
124
pgoff_t index, gfp_t gfp_mask);
include/linux/shmem_fs.h
132
struct vm_area_struct *vma, pgoff_t index,
include/linux/shmem_fs.h
137
struct vm_area_struct *vma, pgoff_t index,
include/linux/shmem_fs.h
163
pgoff_t start, pgoff_t end);
include/linux/shmem_fs.h
174
int shmem_get_folio(struct inode *inode, pgoff_t index, loff_t write_end,
include/linux/shmem_fs.h
177
pgoff_t index, gfp_t gfp);
include/linux/shmem_fs.h
180
pgoff_t index)
include/linux/shmem_fs.h
186
struct address_space *mapping, pgoff_t index)
include/linux/shmem_fs.h
217
static inline pgoff_t shmem_fallocend(struct inode *inode, pgoff_t eof)
include/linux/shmem_fs.h
52
pgoff_t fallocend; /* highest fallocate endindex */
include/linux/swap.h
191
pgoff_t start_page;
include/linux/swap.h
192
pgoff_t nr_pages;
include/linux/swap.h
439
extern sector_t swapdev_block(int, pgoff_t);
include/linux/swapops.h
105
static inline pgoff_t swp_offset(swp_entry_t entry)
include/linux/swapops.h
136
static inline swp_entry_t make_readable_device_private_entry(pgoff_t offset)
include/linux/swapops.h
141
static inline swp_entry_t make_writable_device_private_entry(pgoff_t offset)
include/linux/swapops.h
146
static inline swp_entry_t make_device_exclusive_entry(pgoff_t offset)
include/linux/swapops.h
152
static inline swp_entry_t make_readable_device_private_entry(pgoff_t offset)
include/linux/swapops.h
157
static inline swp_entry_t make_writable_device_private_entry(pgoff_t offset)
include/linux/swapops.h
162
static inline swp_entry_t make_device_exclusive_entry(pgoff_t offset)
include/linux/swapops.h
171
static inline swp_entry_t make_readable_migration_entry(pgoff_t offset)
include/linux/swapops.h
176
static inline swp_entry_t make_readable_exclusive_migration_entry(pgoff_t offset)
include/linux/swapops.h
181
static inline swp_entry_t make_writable_migration_entry(pgoff_t offset)
include/linux/swapops.h
220
static inline swp_entry_t make_readable_migration_entry(pgoff_t offset)
include/linux/swapops.h
225
static inline swp_entry_t make_readable_exclusive_migration_entry(pgoff_t offset)
include/linux/swapops.h
230
static inline swp_entry_t make_writable_migration_entry(pgoff_t offset)
include/linux/swapops.h
84
static inline swp_entry_t swp_entry(unsigned long type, pgoff_t offset)
include/linux/writeback.h
364
pgoff_t start, pgoff_t end);
include/linux/writeback.h
76
pgoff_t index;
include/trace/events/btrfs.h
685
__field( pgoff_t, index )
include/trace/events/erofs.h
117
TP_PROTO(struct inode *inode, pgoff_t start, unsigned int nrpage,
include/trace/events/erofs.h
125
__field(pgoff_t, start )
include/trace/events/erofs.h
92
__field(pgoff_t, index )
include/trace/events/ext4.h
458
__field( pgoff_t, writeback_index )
include/trace/events/ext4.h
585
__field( pgoff_t, writeback_index )
include/trace/events/ext4.h
618
__field( pgoff_t, index )
include/trace/events/ext4.h
656
__field( pgoff_t, index )
include/trace/events/f2fs.h
1180
__field(pgoff_t, index)
include/trace/events/f2fs.h
1366
__field(pgoff_t, index)
include/trace/events/f2fs.h
1367
__field(pgoff_t, nrpages)
include/trace/events/f2fs.h
1434
TP_PROTO(struct inode *inode, struct inode *cow_inode, pgoff_t index,
include/trace/events/f2fs.h
1442
__field(pgoff_t, index)
include/trace/events/f2fs.h
1471
TP_PROTO(struct inode *inode, pgoff_t index,
include/trace/events/f2fs.h
1478
__field(pgoff_t, index)
include/trace/events/f2fs.h
1501
TP_PROTO(struct inode *inode, pgoff_t index,
include/trace/events/f2fs.h
1509
TP_PROTO(struct inode *inode, pgoff_t index,
include/trace/events/f2fs.h
1527
__field(pgoff_t, writeback_index)
include/trace/events/f2fs.h
1578
TP_PROTO(struct inode *inode, pgoff_t start, unsigned int nrpage),
include/trace/events/f2fs.h
1584
__field(pgoff_t, start)
include/trace/events/f2fs.h
2024
TP_PROTO(struct inode *inode, pgoff_t cluster_idx,
include/trace/events/f2fs.h
2031
__field(pgoff_t, idx)
include/trace/events/f2fs.h
2055
TP_PROTO(struct inode *inode, pgoff_t cluster_idx,
include/trace/events/f2fs.h
2062
__field(pgoff_t, idx)
include/trace/events/f2fs.h
2086
TP_PROTO(struct inode *inode, pgoff_t cluster_idx,
include/trace/events/f2fs.h
2094
TP_PROTO(struct inode *inode, pgoff_t cluster_idx,
include/trace/events/f2fs.h
2102
TP_PROTO(struct inode *inode, pgoff_t cluster_idx,
include/trace/events/f2fs.h
2110
TP_PROTO(struct inode *inode, pgoff_t cluster_idx,
include/trace/events/filemap.h
100
pgoff_t last_index
include/trace/events/filemap.h
108
pgoff_t index,
include/trace/events/filemap.h
109
pgoff_t last_index
include/trace/events/filemap.h
115
TP_PROTO(struct address_space *mapping, pgoff_t index),
include/trace/events/filemap.h
63
pgoff_t index,
include/trace/events/filemap.h
64
pgoff_t last_index
include/trace/events/filemap.h
99
pgoff_t index,
include/trace/events/fs_dax.h
112
__field(pgoff_t, pgoff)
include/trace/events/fs_dax.h
12
pgoff_t max_pgoff, int result),
include/trace/events/fs_dax.h
150
TP_PROTO(struct inode *inode, pgoff_t start_index, pgoff_t end_index),
include/trace/events/fs_dax.h
154
__field(pgoff_t, start_index)
include/trace/events/fs_dax.h
155
__field(pgoff_t, end_index)
include/trace/events/fs_dax.h
175
TP_PROTO(struct inode *inode, pgoff_t start_index, pgoff_t end_index),\
include/trace/events/fs_dax.h
182
TP_PROTO(struct inode *inode, pgoff_t pgoff, pgoff_t pglen),
include/trace/events/fs_dax.h
186
__field(pgoff_t, pgoff)
include/trace/events/fs_dax.h
187
__field(pgoff_t, pglen)
include/trace/events/fs_dax.h
20
__field(pgoff_t, pgoff)
include/trace/events/fs_dax.h
21
__field(pgoff_t, max_pgoff)
include/trace/events/fs_dax.h
57
pgoff_t max_pgoff, int result), \
include/trace/events/huge_memory.h
204
TP_PROTO(struct mm_struct *mm, struct folio *new_folio, pgoff_t index,
include/trace/events/huge_memory.h
211
__field(pgoff_t, index)
include/trace/events/netfs.h
490
__field(pgoff_t, index)
include/trace/events/readahead.h
118
TP_PROTO(struct inode *inode, pgoff_t index, struct file_ra_state *ra,
include/trace/events/readahead.h
124
TP_PROTO(struct inode *inode, pgoff_t index, struct file_ra_state *ra,
include/trace/events/readahead.h
15
TP_PROTO(struct inode *inode, pgoff_t index, unsigned long nr_to_read,
include/trace/events/readahead.h
23
__field(pgoff_t, index)
include/trace/events/readahead.h
44
TP_PROTO(struct inode *inode, pgoff_t index, struct file_ra_state *ra),
include/trace/events/readahead.h
51
__field(pgoff_t, index)
include/trace/events/readahead.h
77
TP_PROTO(struct inode *inode, pgoff_t index, struct file_ra_state *ra,
include/trace/events/readahead.h
85
__field(pgoff_t, index)
include/trace/events/writeback.h
71
__field(pgoff_t, index)
ipc/shm.c
582
unsigned long addr, pgoff_t *ilx)
kernel/bpf/arraymap.c
581
pgoff_t pgoff = PAGE_ALIGN(sizeof(*array)) >> PAGE_SHIFT;
kernel/events/internal.h
85
pgoff_t pgoff, int nr_pages, long watermark, int flags);
kernel/events/ring_buffer.c
678
pgoff_t pgoff, int nr_pages, long watermark, int flags)
kernel/events/uprobes.c
386
const pgoff_t index = vaddr_to_offset(vma, vaddr) >> PAGE_SHIFT;
kernel/power/swap.c
266
static int hib_submit_io_sync(blk_opf_t opf, pgoff_t page_off, void *addr)
kernel/power/swap.c
272
static int hib_submit_io_async(blk_opf_t opf, pgoff_t page_off, void *addr,
kernel/relay.c
36
pgoff_t pgoff = vmf->pgoff;
lib/iov_iter.c
1004
pgoff_t index;
lib/iov_iter.c
973
pgoff_t index, unsigned int nr_pages)
lib/scatterlist.c
1345
pgoff_t index = start / PAGE_SIZE;
mm/damon/vaddr.c
590
pgoff_t ilx;
mm/fadvise.c
37
pgoff_t start_index;
mm/fadvise.c
38
pgoff_t end_index;
mm/filemap.c
1806
pgoff_t page_cache_next_miss(struct address_space *mapping,
mm/filemap.c
1807
pgoff_t index, unsigned long max_scan)
mm/filemap.c
1843
pgoff_t page_cache_prev_miss(struct address_space *mapping,
mm/filemap.c
1844
pgoff_t index, unsigned long max_scan)
mm/filemap.c
1892
void *filemap_get_entry(struct address_space *mapping, pgoff_t index)
mm/filemap.c
1941
pgoff_t index, fgf_t fgp_flags, gfp_t gfp, struct mempolicy *policy)
mm/filemap.c
2060
static inline struct folio *find_get_entry(struct xa_state *xas, pgoff_t max,
mm/filemap.c
2115
unsigned find_get_entries(struct address_space *mapping, pgoff_t *start,
mm/filemap.c
2116
pgoff_t end, struct folio_batch *fbatch, pgoff_t *indices)
mm/filemap.c
2164
unsigned find_lock_entries(struct address_space *mapping, pgoff_t *start,
mm/filemap.c
2165
pgoff_t end, struct folio_batch *fbatch, pgoff_t *indices)
mm/filemap.c
2232
unsigned filemap_get_folios(struct address_space *mapping, pgoff_t *start,
mm/filemap.c
2233
pgoff_t end, struct folio_batch *fbatch)
mm/filemap.c
2255
pgoff_t *start, pgoff_t end, struct folio_batch *fbatch)
mm/filemap.c
2330
unsigned filemap_get_folios_tag(struct address_space *mapping, pgoff_t *start,
mm/filemap.c
2331
pgoff_t end, xa_mark_t tag, struct folio_batch *fbatch)
mm/filemap.c
2357
if (end == (pgoff_t)-1)
mm/filemap.c
2358
*start = (pgoff_t)-1;
mm/filemap.c
2385
unsigned filemap_get_folios_dirty(struct address_space *mapping, pgoff_t *start,
mm/filemap.c
2386
pgoff_t end, struct folio_batch *fbatch)
mm/filemap.c
2416
if (end == (pgoff_t)-1)
mm/filemap.c
2417
*start = (pgoff_t)-1;
mm/filemap.c
2456
pgoff_t index, pgoff_t max, struct folio_batch *fbatch)
mm/filemap.c
2606
pgoff_t index;
mm/filemap.c
2655
pgoff_t last_index)
mm/filemap.c
2672
pgoff_t index = iocb->ki_pos >> PAGE_SHIFT;
mm/filemap.c
2673
pgoff_t last_index;
mm/filemap.c
3208
pgoff_t max = (end - 1) >> PAGE_SHIFT;
mm/filemap.c
3519
pgoff_t max_idx, index = vmf->pgoff;
mm/filemap.c
3674
pgoff_t start)
mm/filemap.c
3702
struct address_space *mapping, pgoff_t end_pgoff)
mm/filemap.c
3750
pgoff_t file_end)
mm/filemap.c
3872
pgoff_t start_pgoff, pgoff_t end_pgoff)
mm/filemap.c
3877
pgoff_t file_end, last_pgoff = start_pgoff;
mm/filemap.c
4059
pgoff_t index, filler_t filler, struct file *file, gfp_t gfp)
mm/filemap.c
4135
struct folio *read_cache_folio(struct address_space *mapping, pgoff_t index,
mm/filemap.c
4161
pgoff_t index, gfp_t gfp)
mm/filemap.c
4168
pgoff_t index, filler_t *filler, struct file *file, gfp_t gfp)
mm/filemap.c
4179
pgoff_t index, filler_t *filler, struct file *file)
mm/filemap.c
4202
pgoff_t index,
mm/filemap.c
4526
pgoff_t first = start >> PAGE_SHIFT;
mm/filemap.c
4527
pgoff_t last = end >> PAGE_SHIFT;
mm/filemap.c
4528
pgoff_t nr = end == LLONG_MAX ? ULONG_MAX : last - first + 1;
mm/filemap.c
4569
pgoff_t first_index, pgoff_t last_index, struct cachestat *cs)
mm/filemap.c
4581
pgoff_t folio_first_index, folio_last_index;
mm/filemap.c
4720
pgoff_t first_index, last_index;
mm/filemap.c
484
pgoff_t max = end_byte >> PAGE_SHIFT;
mm/filemap.c
513
pgoff_t index = start_byte >> PAGE_SHIFT;
mm/filemap.c
514
pgoff_t end = end_byte >> PAGE_SHIFT;
mm/filemap.c
641
pgoff_t max = end_byte >> PAGE_SHIFT;
mm/filemap.c
815
pgoff_t offset = old->index;
mm/filemap.c
849
struct folio *folio, pgoff_t index, gfp_t gfp, void **shadowp)
mm/filemap.c
950
pgoff_t index, gfp_t gfp)
mm/folio-compat.c
70
pgoff_t index, gfp_t gfp)
mm/folio-compat.c
77
struct page *pagecache_get_page(struct address_space *mapping, pgoff_t index,
mm/gup.c
3438
pgoff_t *offset)
mm/gup.c
3442
pgoff_t start_idx, end_idx;
mm/huge_memory.c
3869
pgoff_t end, int *nr_shmem_dropped)
mm/huge_memory.c
4044
pgoff_t end = 0;
mm/huge_memory.c
4799
static int split_huge_pages_in_file(const char *file_path, pgoff_t off_start,
mm/huge_memory.c
4800
pgoff_t off_end, unsigned int new_order,
mm/huge_memory.c
4805
pgoff_t index;
mm/huge_memory.c
4900
pgoff_t off_start = 0, off_end = 0;
mm/hugetlb.c
1013
static pgoff_t vma_hugecache_offset(struct hstate *h,
mm/hugetlb.c
2452
pgoff_t idx;
mm/hugetlb.c
5365
pgoff_t pgoff;
mm/hugetlb.c
5510
pgoff_t idx;
mm/hugetlb.c
5615
pgoff_t idx = linear_page_index(vma, address);
mm/hugetlb.c
5626
pgoff_t idx)
mm/hugetlb.c
5915
u32 hugetlb_fault_mutex_hash(struct address_space *mapping, pgoff_t idx)
mm/hugetlb.c
5932
u32 hugetlb_fault_mutex_hash(struct address_space *mapping, pgoff_t idx)
mm/hugetlb.c
6172
pgoff_t idx = vma_hugecache_offset(h, dst_vma, dst_addr);
mm/hugetlb.c
6761
unsigned long addr, pgoff_t idx)
mm/hugetlb.c
6848
pgoff_t idx = ((addr - vma->vm_start) >> PAGE_SHIFT) +
mm/internal.h
1148
pgoff_t pgoff, addr;
mm/internal.h
1237
pgoff_t pgoff, unsigned long nr_pages)
mm/internal.h
1263
pgoff_t pgoff;
mm/internal.h
1713
pgoff_t pgoff)
mm/internal.h
1758
pgoff_t index, void *expected, gfp_t gfp);
mm/internal.h
550
struct file *file, pgoff_t index, unsigned long nr_to_read)
mm/internal.h
556
unsigned find_lock_entries(struct address_space *mapping, pgoff_t *start,
mm/internal.h
557
pgoff_t end, struct folio_batch *fbatch, pgoff_t *indices);
mm/internal.h
558
unsigned find_get_entries(struct address_space *mapping, pgoff_t *start,
mm/internal.h
559
pgoff_t end, struct folio_batch *fbatch, pgoff_t *indices);
mm/internal.h
566
pgoff_t start, pgoff_t end, unsigned long *nr_failed);
mm/khugepaged.c
1774
static void retract_page_tables(struct address_space *mapping, pgoff_t pgoff)
mm/khugepaged.c
1883
struct file *file, pgoff_t start, struct collapse_control *cc)
mm/khugepaged.c
1888
pgoff_t index = 0, end = start + HPAGE_PMD_NR;
mm/khugepaged.c
2314
unsigned long addr, struct file *file, pgoff_t start,
mm/khugepaged.c
2432
pgoff_t pgoff;
mm/madvise.c
239
pgoff_t end_index = linear_page_index(vma, end) - 1;
mm/mapping_dirty_helpers.c
264
pgoff_t first_index, pgoff_t nr)
mm/mapping_dirty_helpers.c
314
pgoff_t first_index, pgoff_t nr,
mm/mapping_dirty_helpers.c
315
pgoff_t bitmap_pgoff,
mm/mapping_dirty_helpers.c
317
pgoff_t *start,
mm/mapping_dirty_helpers.c
318
pgoff_t *end)
mm/mapping_dirty_helpers.c
67
pgoff_t bitmap_pgoff;
mm/mapping_dirty_helpers.c
69
pgoff_t start;
mm/mapping_dirty_helpers.c
70
pgoff_t end;
mm/mapping_dirty_helpers.c
97
pgoff_t pgoff = ((addr - walk->vma->vm_start) >> PAGE_SHIFT) +
mm/memfd.c
67
struct folio *memfd_alloc_folio(struct file *memfd, pgoff_t idx)
mm/memfd_luo.c
97
pgoff_t offset;
mm/memory-failure.c
1685
struct address_space *mapping, pgoff_t index, int flags)
mm/memory-failure.c
1784
int mf_dax_kill_procs(struct address_space *mapping, pgoff_t index,
mm/memory-failure.c
2213
pgoff_t pgoff)
mm/memory-failure.c
2255
pgoff_t pgoff;
mm/memory-failure.c
551
pgoff_t pgoff;
mm/memory-failure.c
590
pgoff_t pgoff;
mm/memory-failure.c
623
struct list_head *to_kill, pgoff_t pgoff)
mm/memory-failure.c
633
struct address_space *mapping, pgoff_t pgoff,
mm/memory.c
2990
unsigned long pfn, pgoff_t *vm_pgoff_p)
mm/memory.c
3210
static int __simple_ioremap_prep(unsigned long vm_len, pgoff_t vm_pgoff,
mm/memory.c
4324
pgoff_t first_index,
mm/memory.c
4325
pgoff_t last_index,
mm/memory.c
4333
const pgoff_t start_idx = max(first_index, vma->vm_pgoff);
mm/memory.c
4334
const pgoff_t end_idx = min(last_index, vma_last_pgoff(vma)) + 1;
mm/memory.c
4360
pgoff_t first_index;
mm/memory.c
4361
pgoff_t last_index;
mm/memory.c
4391
void unmap_mapping_pages(struct address_space *mapping, pgoff_t start,
mm/memory.c
4392
pgoff_t nr, bool even_cows)
mm/memory.c
4395
pgoff_t first_index = start;
mm/memory.c
4396
pgoff_t last_index = start + nr - 1;
mm/memory.c
4430
pgoff_t hba = (pgoff_t)(holebegin) >> PAGE_SHIFT;
mm/memory.c
4431
pgoff_t hlen = ((pgoff_t)(holelen) + PAGE_SIZE - 1) >> PAGE_SHIFT;
mm/memory.c
4651
static inline unsigned long thp_swap_suitable_orders(pgoff_t swp_offset,
mm/memory.c
5684
pgoff_t file_end;
mm/memory.c
5718
pgoff_t idx = folio_page_idx(folio, page);
mm/memory.c
5720
pgoff_t vma_off = vmf->pgoff - vmf->vma->vm_pgoff;
mm/memory.c
5722
pgoff_t pte_off = pte_index(vmf->address);
mm/memory.c
5829
pgoff_t nr_pages = READ_ONCE(fault_around_pages);
mm/memory.c
5830
pgoff_t pte_off = pte_index(vmf->address);
mm/memory.c
5832
pgoff_t vma_off = vmf->pgoff - vmf->vma->vm_pgoff;
mm/memory.c
5833
pgoff_t from_pte, to_pte;
mm/memory.c
5841
to_pte = min3(from_pte + nr_pages, (pgoff_t)PTRS_PER_PTE,
mm/memory.c
591
pgoff_t index;
mm/mempolicy.c
1170
pgoff_t ilx; /* ignored here */
mm/mempolicy.c
1439
pgoff_t ilx = mmpol->ilx;
mm/mempolicy.c
2019
unsigned long addr, pgoff_t *ilx)
mm/mempolicy.c
2042
unsigned long addr, int order, pgoff_t *ilx)
mm/mempolicy.c
2063
pgoff_t ilx; /* ignored here */
mm/mempolicy.c
2203
static unsigned int weighted_interleave_nid(struct mempolicy *pol, pgoff_t ilx)
mm/mempolicy.c
2248
static unsigned int interleave_nid(struct mempolicy *pol, pgoff_t ilx)
mm/mempolicy.c
2270
pgoff_t ilx, int *nid)
mm/mempolicy.c
2330
pgoff_t ilx;
mm/mempolicy.c
2446
struct mempolicy *pol, pgoff_t ilx, int nid)
mm/mempolicy.c
2507
struct mempolicy *pol, pgoff_t ilx, int nid)
mm/mempolicy.c
2537
pgoff_t ilx;
mm/mempolicy.c
2891
pgoff_t start, pgoff_t end)
mm/mempolicy.c
2946
pgoff_t idx)
mm/mempolicy.c
2988
pgoff_t ilx;
mm/mempolicy.c
3134
static int shared_policy_replace(struct shared_policy *sp, pgoff_t start,
mm/mempolicy.c
3135
pgoff_t end, struct sp_node *new)
mm/mempolicy.c
609
pgoff_t ilx, int *nid);
mm/mempolicy.c
623
pgoff_t ilx;
mm/mincore.c
108
static unsigned char mincore_page(struct address_space *mapping, pgoff_t index)
mm/mincore.c
142
pgoff_t pgoff;
mm/mmap.c
1428
pgoff_t pgoff;
mm/mremap.c
951
pgoff_t internal_pgoff = (vrm->addr - vma->vm_start) >> PAGE_SHIFT;
mm/mremap.c
952
pgoff_t pgoff = vma->vm_pgoff + internal_pgoff;
mm/nommu.c
1634
pgoff_t start_pgoff, pgoff_t end_pgoff)
mm/nommu.c
1809
pgoff_t low, high;
mm/page-writeback.c
2361
pgoff_t start, pgoff_t end)
mm/page-writeback.c
2415
static pgoff_t wbc_end(struct writeback_control *wbc)
mm/pagewalk.c
796
int walk_page_mapping(struct address_space *mapping, pgoff_t first_index,
mm/pagewalk.c
797
pgoff_t nr, const struct mm_walk_ops *ops,
mm/pagewalk.c
805
pgoff_t vba, vea, cba, cea;
mm/readahead.c
321
pgoff_t end_index; /* The last page we want to read */
mm/readahead.c
445
static inline int ra_alloc_folio(struct readahead_control *ractl, pgoff_t index,
mm/readahead.c
446
pgoff_t mark, unsigned int order, gfp_t gfp)
mm/readahead.c
471
pgoff_t start = readahead_index(ractl);
mm/readahead.c
472
pgoff_t index = start;
mm/readahead.c
474
pgoff_t limit = (i_size_read(mapping->host) - 1) >> PAGE_SHIFT;
mm/readahead.c
475
pgoff_t mark = index + ra->size - ra->async_size;
mm/readahead.c
560
pgoff_t index = readahead_index(ractl);
mm/readahead.c
564
pgoff_t prev_index, miss;
mm/readahead.c
638
pgoff_t index = readahead_index(ractl);
mm/readahead.c
639
pgoff_t expected, start, end, aligned_end, align;
mm/readahead.c
771
pgoff_t new_index, new_nr_pages;
mm/rmap.c
1226
pgoff_t pgoff;
mm/rmap.c
1250
pgoff_t pgoff_start, unsigned long nr_pages,
mm/rmap.c
1273
int mapping_wrprotect_range(struct address_space *mapping, pgoff_t pgoff,
mm/rmap.c
1310
int pfn_mkclean_range(unsigned long pfn, unsigned long nr_pages, pgoff_t pgoff,
mm/rmap.c
2966
pgoff_t pgoff_start, pgoff_end;
mm/rmap.c
3030
pgoff_t pgoff_start, unsigned long nr_pages,
mm/rmap.c
3033
pgoff_t pgoff_end = pgoff_start + nr_pages - 1;
mm/secretmem.c
54
pgoff_t offset = vmf->pgoff;
mm/shmem.c
1062
pgoff_t index = 0;
mm/shmem.c
107
pgoff_t start; /* start of range currently being fallocated */
mm/shmem.c
1076
static struct folio *shmem_get_partial_folio(struct inode *inode, pgoff_t index)
mm/shmem.c
108
pgoff_t next; /* the next page offset to be fallocated */
mm/shmem.c
109
pgoff_t nr_falloced; /* how many new pages have been fallocated */
mm/shmem.c
110
pgoff_t nr_unswapped; /* how often writeout refused to swap out */
mm/shmem.c
1113
pgoff_t start = (lstart + PAGE_SIZE - 1) >> PAGE_SHIFT;
mm/shmem.c
1114
pgoff_t end = (lend + 1) >> PAGE_SHIFT;
mm/shmem.c
1116
pgoff_t indices[FOLIO_BATCH_SIZE];
mm/shmem.c
1120
pgoff_t index;
mm/shmem.c
1214
pgoff_t base = indices[i];
mm/shmem.c
1442
pgoff_t start, struct folio_batch *fbatch,
mm/shmem.c
1443
pgoff_t *indices, unsigned int type)
mm/shmem.c
1484
struct folio_batch *fbatch, pgoff_t *indices)
mm/shmem.c
1514
pgoff_t start = 0;
mm/shmem.c
1516
pgoff_t indices[FOLIO_BATCH_SIZE];
mm/shmem.c
160
static int shmem_swapin_folio(struct inode *inode, pgoff_t index,
mm/shmem.c
1600
pgoff_t index;
mm/shmem.c
1778
pgoff_t index, unsigned int order, pgoff_t *ilx);
mm/shmem.c
1781
struct shmem_inode_info *info, pgoff_t index)
mm/shmem.c
1784
pgoff_t ilx;
mm/shmem.c
1837
struct vm_area_struct *vma, pgoff_t index,
mm/shmem.c
1881
struct address_space *mapping, pgoff_t index,
mm/shmem.c
1885
pgoff_t aligned_index;
mm/shmem.c
1918
struct address_space *mapping, pgoff_t index,
mm/shmem.c
1926
struct shmem_inode_info *info, pgoff_t index)
mm/shmem.c
1929
pgoff_t ilx;
mm/shmem.c
1940
gfp_t gfp, struct inode *inode, pgoff_t index,
mm/shmem.c
1947
pgoff_t aligned_index;
mm/shmem.c
2044
struct vm_area_struct *vma, pgoff_t index,
mm/shmem.c
2126
struct shmem_inode_info *info, pgoff_t index,
mm/shmem.c
2185
static void shmem_set_folio_swapin_error(struct inode *inode, pgoff_t index,
mm/shmem.c
2212
static int shmem_split_large_entry(struct inode *inode, pgoff_t index,
mm/shmem.c
2226
pgoff_t swap_index;
mm/shmem.c
2244
pgoff_t aligned_index =
mm/shmem.c
2246
pgoff_t swap_offset = aligned_index - swap_index;
mm/shmem.c
2290
static int shmem_swapin_folio(struct inode *inode, pgoff_t index,
mm/shmem.c
2303
pgoff_t offset;
mm/shmem.c
2464
static int shmem_get_folio_gfp(struct inode *inode, pgoff_t index,
mm/shmem.c
2667
int shmem_get_folio(struct inode *inode, pgoff_t index, loff_t write_end,
mm/shmem.c
2892
unsigned long addr, pgoff_t *ilx)
mm/shmem.c
2895
pgoff_t index;
mm/shmem.c
2909
pgoff_t index, unsigned int order, pgoff_t *ilx)
mm/shmem.c
2921
pgoff_t index, unsigned int order, pgoff_t *ilx)
mm/shmem.c
3191
pgoff_t pgoff = linear_page_index(dst_vma, dst_addr);
mm/shmem.c
3195
pgoff_t max_off;
mm/shmem.c
3306
pgoff_t index = pos >> PAGE_SHIFT;
mm/shmem.c
3367
pgoff_t index;
mm/shmem.c
3568
pgoff_t index;
mm/shmem.c
3688
pgoff_t start, index, end, undo_fallocend;
mm/shmem.c
498
pgoff_t index, void *expected, void *replacement)
mm/shmem.c
520
static int shmem_confirm_swap(struct address_space *mapping, pgoff_t index,
mm/shmem.c
5993
pgoff_t index, gfp_t gfp)
mm/shmem.c
6017
pgoff_t index, gfp_t gfp)
mm/shmem.c
605
unsigned long within_size_orders, pgoff_t index,
mm/shmem.c
608
pgoff_t aligned_index;
mm/shmem.c
626
static unsigned int shmem_huge_global_enabled(struct inode *inode, pgoff_t index,
mm/shmem.c
762
pgoff_t next, end;
mm/shmem.c
861
static unsigned int shmem_huge_global_enabled(struct inode *inode, pgoff_t index,
mm/shmem.c
883
pgoff_t index, void *expected, gfp_t gfp)
mm/shmem.c
965
pgoff_t index, pgoff_t end, void *radswap)
mm/shmem.c
969
pgoff_t base;
mm/shmem.c
998
pgoff_t start, pgoff_t end)
mm/swap.h
284
struct mempolicy *mpol, pgoff_t ilx,
mm/swap.h
300
struct mempolicy *mpol, pgoff_t ilx);
mm/swap.h
358
struct swap_info_struct *si, pgoff_t offset, bool irq)
mm/swap.h
425
gfp_t gfp_mask, struct mempolicy *mpol, pgoff_t ilx)
mm/swap.h
85
struct swap_info_struct *si, pgoff_t offset)
mm/swap_cgroup.c
25
pgoff_t offset)
mm/swap_cgroup.c
37
pgoff_t offset,
mm/swap_cgroup.c
69
pgoff_t offset, end;
mm/swap_cgroup.c
95
pgoff_t offset, end;
mm/swap_state.c
530
struct mempolicy *mpol, pgoff_t ilx,
mm/swap_state.c
577
pgoff_t offset = swp_offset(entry);
mm/swap_state.c
600
pgoff_t ilx;
mm/swap_state.c
699
struct mempolicy *mpol, pgoff_t ilx)
mm/swap_state.c
807
struct mempolicy *mpol, pgoff_t targ_ilx, struct vm_fault *vmf)
mm/swap_state.c
815
pgoff_t ilx;
mm/swap_state.c
895
pgoff_t ilx;
mm/swapfile.c
1544
pgoff_t offset, int nr,
mm/swapfile.c
1549
pgoff_t end = offset + nr;
mm/swapfile.c
1651
pgoff_t offset, int nr)
mm/swapfile.c
1914
pgoff_t offset = swp_offset(entry);
mm/swapfile.c
2125
pgoff_t offset = swp_offset(entry);
mm/swapfile.c
2198
sector_t swapdev_block(int type, pgoff_t offset)
mm/swapfile.c
337
pgoff_t offset;
mm/swapfile.c
350
pgoff_t start_page, pgoff_t nr_pages)
mm/swapfile.c
355
pgoff_t offset = start_page - se->start_page;
mm/truncate.c
369
pgoff_t start; /* inclusive */
mm/truncate.c
370
pgoff_t end; /* exclusive */
mm/truncate.c
372
pgoff_t indices[FOLIO_BATCH_SIZE];
mm/truncate.c
373
pgoff_t index;
mm/truncate.c
535
pgoff_t start, pgoff_t end, unsigned long *nr_failed)
mm/truncate.c
537
pgoff_t indices[FOLIO_BATCH_SIZE];
mm/truncate.c
539
pgoff_t index = start;
mm/truncate.c
600
pgoff_t start, pgoff_t end)
mm/truncate.c
61
struct folio_batch *fbatch, pgoff_t *indices)
mm/truncate.c
673
pgoff_t start, pgoff_t end)
mm/truncate.c
675
pgoff_t indices[FOLIO_BATCH_SIZE];
mm/truncate.c
677
pgoff_t index;
mm/userfaultfd.c
151
pgoff_t offset, max_off;
mm/userfaultfd.c
390
pgoff_t pgoff = linear_page_index(dst_vma, dst_addr);
mm/userfaultfd.c
507
pgoff_t idx;
mm/vma.c
1248
unsigned long start, unsigned long end, pgoff_t pgoff)
mm/vma.c
16
pgoff_t pgoff;
mm/vma.c
1865
unsigned long addr, unsigned long len, pgoff_t pgoff,
mm/vma.c
200
pgoff_t pglen = PHYS_PFN(vmg->end - vmg->start);
mm/vma.c
708
pgoff_t pgoff;
mm/vma.c
968
pgoff_t pglen = PHYS_PFN(vmg->end - vmg->start);
mm/vma.h
234
static inline pgoff_t vma_pgoff_offset(struct vm_area_struct *vma,
mm/vma.h
279
unsigned long start, unsigned long end, pgoff_t pgoff);
mm/vma.h
451
unsigned long addr, unsigned long len, pgoff_t pgoff,
mm/vma.h
99
pgoff_t pgoff;
mm/zswap.c
1559
pgoff_t offset = swp_offset(swp);
mm/zswap.c
1597
pgoff_t offset = swp_offset(swp);
mm/zswap.c
1649
pgoff_t offset = swp_offset(swp);
mm/zswap.c
998
pgoff_t offset = swp_offset(swpentry);
samples/vfio-mdev/mbochs.c
161
pgoff_t pagecount;
samples/vfio-mdev/mbochs.c
181
pgoff_t pagecount;
samples/vfio-mdev/mbochs.c
212
pgoff_t pgoff);
samples/vfio-mdev/mbochs.c
214
pgoff_t pgoff);
samples/vfio-mdev/mbochs.c
726
pgoff_t pgoff)
samples/vfio-mdev/mbochs.c
742
pgoff_t pgoff)
samples/vfio-mdev/mbochs.c
777
pgoff_t page_offset = (vmf->address - vma->vm_start) >> PAGE_SHIFT;
samples/vfio-mdev/mbochs.c
907
pgoff_t pg;
samples/vfio-mdev/mbochs.c
932
pgoff_t page_offset, pg;
tools/testing/nvdimm/dax-dev.c
9
phys_addr_t dax_pgoff_to_phys(struct dev_dax *dev_dax, pgoff_t pgoff,
tools/testing/nvdimm/pmem-dax.c
11
long __pmem_direct_access(struct pmem_device *pmem, pgoff_t pgoff,
tools/testing/vma/include/dup.h
480
pgoff_t pgoff;
tools/testing/vma/include/dup.h
535
pgoff_t pgoff;
tools/testing/vma/include/dup.h
672
int (*mapped)(unsigned long start, unsigned long end, pgoff_t pgoff,
tools/testing/vma/include/dup.h
687
pgoff_t start_pgoff, pgoff_t end_pgoff);
tools/testing/vma/include/dup.h
730
unsigned long addr, pgoff_t *ilx);
tools/testing/vma/shared.c
126
pgoff_t pgoff)
tools/testing/vma/shared.c
17
pgoff_t pgoff, vma_flags_t vma_flags)
tools/testing/vma/shared.c
41
pgoff_t pgoff, vma_flags_t vma_flags)
tools/testing/vma/shared.h
105
pgoff_t pgoff, vma_flags_t vma_flags);
tools/testing/vma/shared.h
132
pgoff_t pgoff);
tools/testing/vma/shared.h
97
pgoff_t pgoff, vma_flags_t vma_flags);
tools/testing/vma/tests/merge.c
36
unsigned long end, pgoff_t pgoff, vma_flags_t vma_flags)
tools/testing/vma/tests/merge.c
59
unsigned long end, pgoff_t pgoff, vma_flags_t vma_flags,
tools/testing/vma/tests/merge.c
74
unsigned long end, pgoff_t pgoff, vma_flags_t vma_flags,
virt/kvm/guest_memfd.c
120
static struct folio *kvm_gmem_get_folio(struct inode *inode, pgoff_t index)
virt/kvm/guest_memfd.c
161
static void __kvm_gmem_invalidate_begin(struct gmem_file *f, pgoff_t start,
virt/kvm/guest_memfd.c
162
pgoff_t end,
virt/kvm/guest_memfd.c
171
pgoff_t pgoff = slot->gmem.pgoff;
virt/kvm/guest_memfd.c
198
static void kvm_gmem_invalidate_begin(struct inode *inode, pgoff_t start,
virt/kvm/guest_memfd.c
199
pgoff_t end)
virt/kvm/guest_memfd.c
210
static void __kvm_gmem_invalidate_end(struct gmem_file *f, pgoff_t start,
virt/kvm/guest_memfd.c
211
pgoff_t end)
virt/kvm/guest_memfd.c
222
static void kvm_gmem_invalidate_end(struct inode *inode, pgoff_t start,
virt/kvm/guest_memfd.c
223
pgoff_t end)
virt/kvm/guest_memfd.c
233
pgoff_t start = offset >> PAGE_SHIFT;
virt/kvm/guest_memfd.c
234
pgoff_t end = (offset + len) >> PAGE_SHIFT;
virt/kvm/guest_memfd.c
256
pgoff_t start, index, end;
virt/kvm/guest_memfd.c
442
unsigned long addr, pgoff_t *pgoff)
virt/kvm/guest_memfd.c
501
pgoff_t start, end;
virt/kvm/guest_memfd.c
53
static inline kvm_pfn_t folio_file_pfn(struct folio *folio, pgoff_t index)
virt/kvm/guest_memfd.c
58
static pgoff_t kvm_gmem_get_index(struct kvm_memory_slot *slot, gfn_t gfn)
virt/kvm/guest_memfd.c
64
pgoff_t index, struct folio *folio)
virt/kvm/guest_memfd.c
755
pgoff_t index, kvm_pfn_t *pfn,
virt/kvm/guest_memfd.c
793
pgoff_t index = kvm_gmem_get_index(slot, gfn);
virt/kvm/guest_memfd.c
829
pgoff_t index = kvm_gmem_get_index(slot, gfn);
virt/kvm/guest_memfd.c
89
pgoff_t index;