get_pages
.get_pages = etnaviv_gem_shmem_get_pages,
.get_pages = etnaviv_gem_userptr_get_pages,
ret = etnaviv_obj->ops->get_pages(etnaviv_obj);
int (*get_pages)(struct etnaviv_gem_object *);
.get_pages = i915_gem_object_get_pages_dmabuf,
.get_pages = i915_gem_object_get_pages_internal,
int (*get_pages)(struct drm_i915_gem_object *obj);
err = obj->ops->get_pages(obj);
.get_pages = shmem_get_pages,
.get_pages = i915_gem_object_get_pages_stolen,
.get_pages = i915_ttm_get_pages,
.get_pages = i915_gem_userptr_get_pages,
.get_pages = huge_get_pages,
.get_pages = get_huge_pages,
.get_pages = fake_get_huge_pages,
.get_pages = fake_get_huge_pages_single,
.get_pages = pd_dummy_obj_get_pages,
.get_pages = vgpu_gem_get_pages,
.get_pages = fake_get_pages,
.get_pages = mock_region_get_pages,
return get_pages(obj);
pages = get_pages(obj);
DECL_SVM_RANGE_US_STATS(get_pages, GET_PAGES)
goto get_pages;
get_pages:
get_pages(sbi, is_dir ?
get_pages(sbi, is_dir ?
get_pages(sbi, is_dir ?
s64 total = get_pages(sbi, F2FS_DIRTY_IMETA);
} else if (get_pages(sbi, F2FS_DIRTY_QDATA)) {
if (get_pages(sbi, F2FS_DIRTY_DENTS)) {
if (get_pages(sbi, F2FS_DIRTY_IMETA)) {
if (get_pages(sbi, F2FS_DIRTY_NODES)) {
if (!get_pages(sbi, type))
if (get_pages(sbi, F2FS_DIRTY_NODES) ||
get_pages(sbi, F2FS_DIRTY_IMETA))
f2fs_bug_on(sbi, get_pages(sbi, F2FS_DIRTY_DENTS));
get_pages(sbi, F2FS_DIRTY_META) <
wbc->pages_skipped += get_pages(sbi, F2FS_DIRTY_META);
if (type == F2FS_WB_CP_DATA && !get_pages(sbi, type) &&
si->ndirty_node = get_pages(sbi, F2FS_DIRTY_NODES);
si->ndirty_dent = get_pages(sbi, F2FS_DIRTY_DENTS);
si->ndirty_meta = get_pages(sbi, F2FS_DIRTY_META);
si->ndirty_data = get_pages(sbi, F2FS_DIRTY_DATA);
si->ndirty_qdata = get_pages(sbi, F2FS_DIRTY_QDATA);
si->ndirty_imeta = get_pages(sbi, F2FS_DIRTY_IMETA);
si->nr_dio_read = get_pages(sbi, F2FS_DIO_READ);
si->nr_dio_write = get_pages(sbi, F2FS_DIO_WRITE);
si->nr_wb_cp_data = get_pages(sbi, F2FS_WB_CP_DATA);
si->nr_wb_data = get_pages(sbi, F2FS_WB_DATA);
si->nr_rd_data = get_pages(sbi, F2FS_RD_DATA);
si->nr_rd_node = get_pages(sbi, F2FS_RD_NODE);
si->nr_rd_meta = get_pages(sbi, F2FS_RD_META);
return div_u64(get_pages(sbi, block_type) + BLKS_PER_SEC(sbi) - 1,
if (get_pages(sbi, F2FS_RD_DATA) || get_pages(sbi, F2FS_RD_NODE) ||
get_pages(sbi, F2FS_RD_META) || get_pages(sbi, F2FS_WB_DATA) ||
get_pages(sbi, F2FS_WB_CP_DATA) ||
get_pages(sbi, F2FS_DIO_READ) ||
get_pages(sbi, F2FS_DIO_WRITE))
return get_pages(sbi, F2FS_RD_DATA) || get_pages(sbi, F2FS_DIO_READ);
get_pages(F2FS_I_SB(inode), F2FS_DIO_WRITE) &&
get_pages(sbi, F2FS_DIRTY_NODES),
get_pages(sbi, F2FS_DIRTY_DENTS),
get_pages(sbi, F2FS_DIRTY_IMETA),
get_pages(sbi, F2FS_DIRTY_NODES),
get_pages(sbi, F2FS_DIRTY_DENTS),
get_pages(sbi, F2FS_DIRTY_IMETA),
get_pages(sbi, F2FS_DIRTY_NODES) <
wbc->pages_skipped += get_pages(sbi, F2FS_DIRTY_NODES);
mem_size = get_pages(sbi, F2FS_DIRTY_DENTS);
unsigned int dents = get_pages(sbi, F2FS_DIRTY_DENTS);
unsigned int qdata = get_pages(sbi, F2FS_DIRTY_QDATA);
unsigned int nodes = get_pages(sbi, F2FS_DIRTY_NODES);
unsigned int meta = get_pages(sbi, F2FS_DIRTY_META);
unsigned int imeta = get_pages(sbi, F2FS_DIRTY_IMETA);
unsigned int total_node_blocks = get_pages(sbi, F2FS_DIRTY_NODES) +
get_pages(sbi, F2FS_DIRTY_DENTS) +
get_pages(sbi, F2FS_DIRTY_IMETA);
unsigned int total_dent_blocks = get_pages(sbi, F2FS_DIRTY_DENTS);
total_data_blocks = get_pages(sbi, F2FS_DIRTY_DATA);
if (!get_pages(sbi, i))
"umount, type: %d, count: %lld", i, get_pages(sbi, i));
get_pages(sbi, F2FS_DIRTY_META),
get_pages(sbi, F2FS_DIRTY_NODES),
get_pages(sbi, F2FS_DIRTY_DATA));
skipped_write = get_pages(sbi, F2FS_SKIPPED_WRITE);
dirty_data = get_pages(sbi, F2FS_DIRTY_DATA);
get_pages(sbi, F2FS_DIRTY_DATA),
get_pages(sbi, F2FS_SKIPPED_WRITE),
if (!get_pages(sbi, F2FS_DIRTY_DATA))
if (get_pages(sbi, F2FS_SKIPPED_WRITE) == skipped_write)
if (unlikely(get_pages(sbi, F2FS_DIRTY_DATA) ||
get_pages(sbi, F2FS_SKIPPED_WRITE)))
get_pages(sbi, F2FS_DIRTY_DATA),
get_pages(sbi, F2FS_SKIPPED_WRITE), retry);
if (get_pages(sbi, F2FS_SKIPPED_WRITE))