Symbol: page_count
arch/alpha/kernel/core_marvel.c
976
mem->page_count, mem->pages);
arch/alpha/kernel/core_marvel.c
984
mem->page_count);
arch/alpha/kernel/core_titan.c
687
mem->page_count, mem->pages);
arch/alpha/kernel/core_titan.c
695
mem->page_count);
arch/arm64/include/asm/kvm_pgtable.h
225
int (*page_count)(void *addr);
arch/arm64/kvm/hyp/nvhe/mem_protect.c
121
.page_count = hyp_page_count,
arch/arm64/kvm/hyp/nvhe/mem_protect.c
275
.page_count = hyp_page_count,
arch/arm64/kvm/hyp/nvhe/setup.c
311
.page_count = hyp_page_count,
arch/arm64/kvm/hyp/pgtable.c
1166
if (mm_ops->page_count(childp) != 1)
arch/arm64/kvm/hyp/pgtable.c
1613
if (mm_ops->page_count(childp) != 1)
arch/arm64/kvm/hyp/pgtable.c
1696
WARN_ON(mm_ops->page_count(pgtable) != 1);
arch/arm64/kvm/hyp/pgtable.c
488
if (mm_ops->page_count(childp) != 1)
arch/arm64/kvm/hyp/pgtable.c
523
if (!pgt->mm_ops->page_count)
arch/arm64/kvm/mmu.c
262
if (page_count(p) == 1)
arch/arm64/kvm/mmu.c
269
return page_count(virt_to_page(addr));
arch/arm64/kvm/mmu.c
868
.page_count = kvm_host_page_count,
arch/riscv/include/asm/kvm_nacl.h
46
unsigned long page_count);
arch/riscv/kvm/nacl.c
20
unsigned long page_count)
arch/riscv/kvm/nacl.c
50
*entp = cpu_to_lelong(page_count);
arch/s390/include/asm/kvm_host.h
607
int page_count;
arch/s390/kernel/diag/diag310.c
132
union diag310_req_size req_size = { .page_count = pages };
arch/s390/kernel/diag/diag310.c
59
u64 page_count : 32;
arch/s390/kvm/vsie.c
1442
if (kvm->arch.vsie.page_count < nr_vcpus) {
arch/s390/kvm/vsie.c
1449
kvm->arch.vsie.pages[kvm->arch.vsie.page_count] = vsie_page;
arch/s390/kvm/vsie.c
1450
kvm->arch.vsie.page_count++;
arch/s390/kvm/vsie.c
1558
for (i = 0; i < kvm->arch.vsie.page_count; i++) {
arch/s390/kvm/vsie.c
1570
kvm->arch.vsie.page_count = 0;
drivers/android/binder_alloc.c
1018
page_count++;
drivers/android/binder_alloc.c
1028
__func__, alloc->pid, buffers, page_count);
drivers/android/binder_alloc.c
966
int buffers, page_count;
drivers/android/binder_alloc.c
997
page_count = 0;
drivers/block/drbd/drbd_receiver.c
148
if (page_count(page) == 1)
drivers/block/ps3vram.c
357
for (i = 0; i < cache->page_count; i++) {
drivers/block/ps3vram.c
377
for (i = 0; i < cache->page_count; i++) {
drivers/block/ps3vram.c
388
i = (jiffies + (counter++)) % cache->page_count;
drivers/block/ps3vram.c
402
priv->cache.page_count = CACHE_PAGE_COUNT;
drivers/block/ps3vram.c
62
unsigned int page_count;
drivers/char/agp/agp.h
194
struct agp_memory *agp_generic_alloc_by_type(size_t page_count, int type);
drivers/char/agp/agp.h
198
struct agp_memory *memory, size_t page_count);
drivers/char/agp/agp.h
215
struct agp_memory *agp_generic_alloc_user(size_t page_count, int type);
drivers/char/agp/ali-agp.c
128
int i, page_count;
drivers/char/agp/ali-agp.c
133
page_count = 1 << A_SIZE_32(agp_bridge->current_size)->page_order;
drivers/char/agp/ali-agp.c
134
for (i = 0; i < PAGE_SIZE * page_count; i += PAGE_SIZE) {
drivers/char/agp/alpha-agp.c
96
if ((pg_start + mem->page_count) > num_entries)
drivers/char/agp/amd-k7-agp.c
296
if ((pg_start + mem->page_count) > num_entries)
drivers/char/agp/amd-k7-agp.c
300
while (j < (pg_start + mem->page_count)) {
drivers/char/agp/amd-k7-agp.c
313
for (i = 0, j = pg_start; i < mem->page_count; i++, j++) {
drivers/char/agp/amd-k7-agp.c
336
for (i = pg_start; i < (mem->page_count + pg_start); i++) {
drivers/char/agp/amd64-agp.c
64
if (((unsigned long)pg_start + mem->page_count) > num_entries)
drivers/char/agp/amd64-agp.c
70
while (j < (pg_start + mem->page_count)) {
drivers/char/agp/amd64-agp.c
81
for (i = 0, j = pg_start; i < mem->page_count; i++, j++) {
drivers/char/agp/ati-agp.c
272
if (mem->page_count == 0)
drivers/char/agp/ati-agp.c
275
if ((pg_start + mem->page_count) > num_entries)
drivers/char/agp/ati-agp.c
279
while (j < (pg_start + mem->page_count)) {
drivers/char/agp/ati-agp.c
293
for (i = 0, j = pg_start; i < mem->page_count; i++, j++) {
drivers/char/agp/ati-agp.c
318
if (mem->page_count == 0)
drivers/char/agp/ati-agp.c
321
for (i = pg_start; i < (mem->page_count + pg_start); i++) {
drivers/char/agp/efficeon-agp.c
238
int i, count = mem->page_count, num_entries;
drivers/char/agp/efficeon-agp.c
246
if ((pg_start + mem->page_count) > num_entries)
drivers/char/agp/efficeon-agp.c
287
int i, count = mem->page_count, num_entries;
drivers/char/agp/efficeon-agp.c
293
if ((pg_start + mem->page_count) > num_entries)
drivers/char/agp/generic.c
1038
if (mem->page_count == 0)
drivers/char/agp/generic.c
1076
if (((pg_start + mem->page_count) > num_entries) ||
drivers/char/agp/generic.c
1077
((pg_start + mem->page_count) < pg_start))
drivers/char/agp/generic.c
1082
while (j < (pg_start + mem->page_count)) {
drivers/char/agp/generic.c
1093
for (i = 0, j = pg_start; i < mem->page_count; i++, j++) {
drivers/char/agp/generic.c
1117
if (mem->page_count == 0)
drivers/char/agp/generic.c
1124
if (((pg_start + mem->page_count) > num_entries) ||
drivers/char/agp/generic.c
1125
((pg_start + mem->page_count) < pg_start))
drivers/char/agp/generic.c
1135
for (i = pg_start; i < (mem->page_count + pg_start); i++) {
drivers/char/agp/generic.c
1145
struct agp_memory *agp_generic_alloc_by_type(size_t page_count, int type)
drivers/char/agp/generic.c
1159
struct agp_memory *agp_generic_alloc_user(size_t page_count, int type)
drivers/char/agp/generic.c
1165
pages = (page_count + ENTRIES_PER_PAGE - 1) / ENTRIES_PER_PAGE;
drivers/char/agp/generic.c
1166
new = agp_create_user_memory(page_count);
drivers/char/agp/generic.c
1170
for (i = 0; i < page_count; i++)
drivers/char/agp/generic.c
1172
new->page_count = 0;
drivers/char/agp/generic.c
1205
mem->page_count++;
drivers/char/agp/generic.c
1242
set_pages_array_wb(mem->pages, mem->page_count);
drivers/char/agp/generic.c
1245
for (i = 0; i < mem->page_count; i++) {
drivers/char/agp/generic.c
181
if (curr->page_count != 0) {
drivers/char/agp/generic.c
186
for (i = 0; i < curr->page_count; i++) {
drivers/char/agp/generic.c
191
for (i = 0; i < curr->page_count; i++) {
drivers/char/agp/generic.c
219
size_t page_count, u32 type)
drivers/char/agp/generic.c
230
if ((cur_memory + page_count > bridge->max_memory_agp) ||
drivers/char/agp/generic.c
231
(cur_memory + page_count < page_count))
drivers/char/agp/generic.c
235
new = agp_generic_alloc_user(page_count, type);
drivers/char/agp/generic.c
242
new = bridge->driver->alloc_by_type(page_count, type);
drivers/char/agp/generic.c
248
scratch_pages = (page_count + ENTRIES_PER_PAGE - 1) / ENTRIES_PER_PAGE;
drivers/char/agp/generic.c
256
if (bridge->driver->agp_alloc_pages(bridge, new, page_count)) {
drivers/char/agp/generic.c
264
for (i = 0; i < page_count; i++) {
drivers/char/agp/generic.c
272
new->page_count++;
drivers/char/agp/intel-gtt.c
1002
intel_gmch_gtt_clear_range(pg_start, mem->page_count);
drivers/char/agp/intel-gtt.c
1027
new->page_count = pg_count;
drivers/char/agp/intel-gtt.c
129
DBG("try unmapping %lu pages\n", (unsigned long)mem->page_count);
drivers/char/agp/intel-gtt.c
219
if ((pg_start + mem->page_count)
drivers/char/agp/intel-gtt.c
226
for (i = pg_start; i < (pg_start + mem->page_count); i++) {
drivers/char/agp/intel-gtt.c
271
new->page_count = pg_count;
drivers/char/agp/intel-gtt.c
282
if (curr->page_count == 4)
drivers/char/agp/intel-gtt.c
946
if (mem->page_count == 0)
drivers/char/agp/intel-gtt.c
949
if (pg_start + mem->page_count > intel_private.gtt_total_entries)
drivers/char/agp/intel-gtt.c
964
ret = intel_gtt_map_memory(mem->pages, mem->page_count, &st);
drivers/char/agp/intel-gtt.c
972
intel_gmch_gtt_insert_pages(pg_start, mem->page_count, mem->pages,
drivers/char/agp/intel-gtt.c
999
if (mem->page_count == 0)
drivers/char/agp/nvidia-agp.c
211
if (mem->page_count == 0)
drivers/char/agp/nvidia-agp.c
214
if ((pg_start + mem->page_count) >
drivers/char/agp/nvidia-agp.c
218
for (j = pg_start; j < (pg_start + mem->page_count); j++) {
drivers/char/agp/nvidia-agp.c
227
for (i = 0, j = pg_start; i < mem->page_count; i++, j++) {
drivers/char/agp/nvidia-agp.c
251
if (mem->page_count == 0)
drivers/char/agp/nvidia-agp.c
254
for (i = pg_start; i < (mem->page_count + pg_start); i++)
drivers/char/agp/parisc-agp.c
137
io_pg_count = info->io_pages_per_kpage * mem->page_count;
drivers/char/agp/parisc-agp.c
154
for (i = 0, j = io_pg_start; i < mem->page_count; i++) {
drivers/char/agp/parisc-agp.c
185
io_pg_count = info->io_pages_per_kpage * mem->page_count;
drivers/char/agp/sworks-agp.c
328
if ((pg_start + mem->page_count) > num_entries) {
drivers/char/agp/sworks-agp.c
333
while (j < (pg_start + mem->page_count)) {
drivers/char/agp/sworks-agp.c
346
for (i = 0, j = pg_start; i < mem->page_count; i++, j++) {
drivers/char/agp/sworks-agp.c
371
for (i = pg_start; i < (mem->page_count + pg_start); i++) {
drivers/char/agp/uninorth-agp.c
166
if (mem->page_count == 0)
drivers/char/agp/uninorth-agp.c
172
if ((pg_start + mem->page_count) > num_entries)
drivers/char/agp/uninorth-agp.c
176
for (i = 0; i < mem->page_count; ++i) {
drivers/char/agp/uninorth-agp.c
185
for (i = 0; i < mem->page_count; i++) {
drivers/char/agp/uninorth-agp.c
215
if (mem->page_count == 0)
drivers/char/agp/uninorth-agp.c
219
for (i = 0; i < mem->page_count; ++i) {
drivers/crypto/ccp/sev-dev.c
1219
range->page_count = 1 << entry->order;
drivers/crypto/ccp/sev-dev.c
1343
range->page_count = size >> PAGE_SHIFT;
drivers/firewire/core-cdev.c
1138
buffer_end = client->buffer.page_count << PAGE_SHIFT;
drivers/firewire/core-cdev.c
1783
int page_count, ret;
drivers/firewire/core-cdev.c
1800
page_count = size >> PAGE_SHIFT;
drivers/firewire/core-cdev.c
1804
ret = fw_iso_buffer_alloc(&client->buffer, page_count);
drivers/firewire/core-cdev.c
1820
client->buffer.page_count);
drivers/firewire/core-iso.c
110
for (int i = 0; i < buffer->page_count; ++i) {
drivers/firewire/core-iso.c
119
release_pages(buffer->pages, buffer->page_count);
drivers/firewire/core-iso.c
124
buffer->page_count = 0;
drivers/firewire/core-iso.c
131
for (int i = 0; i < buffer->page_count; i++) {
drivers/firewire/core-iso.c
31
int fw_iso_buffer_alloc(struct fw_iso_buffer *buffer, int page_count)
drivers/firewire/core-iso.c
34
page_count);
drivers/firewire/core-iso.c
43
page_count, page_array);
drivers/firewire/core-iso.c
44
if (nr_populated != page_count) {
drivers/firewire/core-iso.c
50
buffer->page_count = page_count;
drivers/firewire/core-iso.c
60
buffer->page_count);
drivers/firewire/core-iso.c
68
for (i = 0; i < buffer->page_count; i++) {
drivers/firewire/core-iso.c
77
if (i < buffer->page_count) {
drivers/firewire/core-iso.c
90
int page_count, enum dma_data_direction direction)
drivers/firewire/core-iso.c
94
ret = fw_iso_buffer_alloc(buffer, page_count);
drivers/firewire/core.h
166
int fw_iso_buffer_alloc(struct fw_iso_buffer *buffer, int page_count);
drivers/firewire/ohci.c
3410
if (WARN_ON(offset & 3 || rest & 3 || page + z > buffer->page_count))
drivers/gpu/drm/i915/gem/i915_gem_shmem.c
103
for (i = 0; i < page_count; i++) {
drivers/gpu/drm/i915/gem/i915_gem_shmem.c
123
i915_gem_shrink(NULL, i915, 2 * page_count, NULL, *s++);
drivers/gpu/drm/i915/gem/i915_gem_shmem.c
158
page_count - i,
drivers/gpu/drm/i915/gem/i915_gem_shmem.c
70
unsigned int page_count; /* restricted by sg_alloc_table */
drivers/gpu/drm/i915/gem/i915_gem_shmem.c
77
if (overflows_type(size / PAGE_SIZE, page_count))
drivers/gpu/drm/i915/gem/i915_gem_shmem.c
80
page_count = size / PAGE_SIZE;
drivers/gpu/drm/i915/gem/i915_gem_shmem.c
88
if (sg_alloc_table(st, page_count, GFP_KERNEL | __GFP_NOWARN))
drivers/gpu/drm/i915/gt/intel_ggtt_fencing.c
816
const unsigned int page_count = obj->base.size >> PAGE_SHIFT;
drivers/gpu/drm/i915/gt/intel_ggtt_fencing.c
822
obj->bit_17 = bitmap_zalloc(page_count, GFP_KERNEL);
drivers/gpu/drm/i915/selftests/scatterlist.c
198
static unsigned int page_count[] = {
drivers/gpu/drm/i915/selftests/scatterlist.c
204
return page_count[(prandom_u32_state(rnd) % 3)];
drivers/gpu/drm/nouveau/nouveau_bo0039.c
55
u32 page_count = PFN_UP(new_reg->size);
drivers/gpu/drm/nouveau/nouveau_bo0039.c
65
page_count = PFN_UP(new_reg->size);
drivers/gpu/drm/nouveau/nouveau_bo0039.c
66
while (page_count) {
drivers/gpu/drm/nouveau/nouveau_bo0039.c
67
int line_count = (page_count > 2047) ? 2047 : page_count;
drivers/gpu/drm/nouveau/nouveau_bo0039.c
88
page_count -= line_count;
drivers/gpu/drm/nouveau/nouveau_bo85b5.c
47
u32 page_count = PFN_UP(new_reg->size);
drivers/gpu/drm/nouveau/nouveau_bo85b5.c
50
page_count = PFN_UP(new_reg->size);
drivers/gpu/drm/nouveau/nouveau_bo85b5.c
51
while (page_count) {
drivers/gpu/drm/nouveau/nouveau_bo85b5.c
52
int line_count = (page_count > 8191) ? 8191 : page_count;
drivers/gpu/drm/nouveau/nouveau_bo85b5.c
68
page_count -= line_count;
drivers/gpu/drm/nouveau/nouveau_bo9039.c
45
u32 page_count = PFN_UP(new_reg->size);
drivers/gpu/drm/nouveau/nouveau_bo9039.c
48
page_count = PFN_UP(new_reg->size);
drivers/gpu/drm/nouveau/nouveau_bo9039.c
49
while (page_count) {
drivers/gpu/drm/nouveau/nouveau_bo9039.c
50
int line_count = (page_count > 2047) ? 2047 : page_count;
drivers/gpu/drm/nouveau/nouveau_bo9039.c
78
page_count -= line_count;
drivers/gpu/drm/nouveau/nouveau_bo90b5.c
40
u32 page_count = PFN_UP(new_reg->size);
drivers/gpu/drm/nouveau/nouveau_bo90b5.c
43
page_count = PFN_UP(new_reg->size);
drivers/gpu/drm/nouveau/nouveau_bo90b5.c
44
while (page_count) {
drivers/gpu/drm/nouveau/nouveau_bo90b5.c
45
int line_count = (page_count > 8191) ? 8191 : page_count;
drivers/gpu/drm/nouveau/nouveau_bo90b5.c
61
page_count -= line_count;
drivers/gpu/drm/ttm/ttm_agp_backend.c
66
mem->page_count = 0;
drivers/gpu/drm/ttm/ttm_agp_backend.c
73
mem->pages[mem->page_count++] = page;
drivers/hv/hv_balloon.c
1001
resp.page_count = process_hot_add(pg_start, pfn_cnt,
drivers/hv/hv_balloon.c
1004
dm->num_pages_added += resp.page_count;
drivers/hv/hv_balloon.c
1022
if (resp.page_count > 0)
drivers/hv/hv_balloon.c
1029
if (!do_hot_add || resp.page_count == 0) {
drivers/hv/hv_balloon.c
369
__u32 page_count;
drivers/hv/hv_proc.c
24
int i, j, page_count;
drivers/hv/hv_proc.c
82
for (i = 0, page_count = 0; i < num_allocations; ++i) {
drivers/hv/hv_proc.c
84
for (j = 0; j < counts[i]; ++j, ++page_count)
drivers/hv/hv_proc.c
85
input_page->gpa_page_list[page_count] = base_pfn + j;
drivers/hv/hv_proc.c
88
page_count, 0, input_page, NULL);
drivers/hv/mshv_regions.c
100
for (count = stride; count < page_count; count += stride) {
drivers/hv/mshv_regions.c
109
page_count - count))
drivers/hv/mshv_regions.c
141
u64 page_offset, u64 page_count,
drivers/hv/mshv_regions.c
145
u64 page_count,
drivers/hv/mshv_regions.c
150
if (page_offset + page_count > region->nr_pages)
drivers/hv/mshv_regions.c
153
while (page_count) {
drivers/hv/mshv_regions.c
157
page_count--;
drivers/hv/mshv_regions.c
163
page_count,
drivers/hv/mshv_regions.c
169
page_count -= ret;
drivers/hv/mshv_regions.c
200
u64 page_offset, u64 page_count,
drivers/hv/mshv_regions.c
208
page_count,
drivers/hv/mshv_regions.c
225
u64 page_offset, u64 page_count,
drivers/hv/mshv_regions.c
233
page_count, 0,
drivers/hv/mshv_regions.c
248
u64 page_offset, u64 page_count,
drivers/hv/mshv_regions.c
256
page_count, flags,
drivers/hv/mshv_regions.c
262
u64 page_offset, u64 page_count)
drivers/hv/mshv_regions.c
265
page_offset, page_count,
drivers/hv/mshv_regions.c
278
u64 page_offset, u64 page_count)
drivers/hv/mshv_regions.c
281
unpin_user_pages(region->mreg_pages + page_offset, page_count);
drivers/hv/mshv_regions.c
284
page_count * sizeof(struct page *));
drivers/hv/mshv_regions.c
332
u64 page_offset, u64 page_count,
drivers/hv/mshv_regions.c
340
page_count, flags);
drivers/hv/mshv_regions.c
35
u64 gfn, u64 page_count)
drivers/hv/mshv_regions.c
440
u64 page_offset, u64 page_count)
drivers/hv/mshv_regions.c
450
pfns = kmalloc_array(page_count, sizeof(*pfns), GFP_KERNEL);
drivers/hv/mshv_regions.c
456
range.end = range.start + page_count * HV_HYP_PAGE_SIZE;
drivers/hv/mshv_regions.c
46
!IS_ALIGNED(page_count, PTRS_PER_PMD))
drivers/hv/mshv_regions.c
465
for (i = 0; i < page_count; i++)
drivers/hv/mshv_regions.c
469
page_offset, page_count);
drivers/hv/mshv_regions.c
479
u64 page_offset, page_count;
drivers/hv/mshv_regions.c
487
page_count = min(region->nr_pages - page_offset,
drivers/hv/mshv_regions.c
490
ret = mshv_region_range_fault(region, page_offset, page_count);
drivers/hv/mshv_regions.c
496
gfn, page_offset, page_count);
drivers/hv/mshv_regions.c
526
u64 page_offset, page_count;
drivers/hv/mshv_regions.c
535
page_count = HVPFN_DOWN(mend - mstart);
drivers/hv/mshv_regions.c
545
page_offset, page_count);
drivers/hv/mshv_regions.c
549
mshv_region_invalidate_pages(region, page_offset, page_count);
drivers/hv/mshv_regions.c
563
page_offset, page_offset + page_count - 1, (u64)range->mm, ret);
drivers/hv/mshv_regions.c
79
u64 page_offset, u64 page_count,
drivers/hv/mshv_regions.c
83
u64 page_count,
drivers/hv/mshv_regions.c
95
stride = mshv_chunk_stride(page, gfn, page_count);
drivers/hv/mshv_root.h
284
int hv_call_map_gpa_pages(u64 partition_id, u64 gpa_target, u64 page_count,
drivers/hv/mshv_root.h
286
int hv_call_unmap_gpa_pages(u64 partition_id, u64 gpa_target, u64 page_count,
drivers/hv/mshv_root.h
300
u64 page_count, struct page **pages,
drivers/hv/mshv_root.h
304
struct hv_vp_state_data state_data, u64 page_count,
drivers/hv/mshv_root_hv_call.c
1008
u64 page_count = page_struct_count;
drivers/hv/mshv_root_hv_call.c
1012
if (page_count == 0)
drivers/hv/mshv_root_hv_call.c
1016
if (!HV_PAGE_COUNT_2M_ALIGNED(page_count))
drivers/hv/mshv_root_hv_call.c
1019
page_count >>= large_shift;
drivers/hv/mshv_root_hv_call.c
1022
while (done < page_count) {
drivers/hv/mshv_root_hv_call.c
1023
ulong i, completed, remain = page_count - done;
drivers/hv/mshv_root_hv_call.c
190
u64 page_count = page_struct_count;
drivers/hv/mshv_root_hv_call.c
192
if (page_count == 0 || (pages && mmio_spa))
drivers/hv/mshv_root_hv_call.c
199
if (!HV_PAGE_COUNT_2M_ALIGNED(page_count))
drivers/hv/mshv_root_hv_call.c
203
page_count >>= large_shift;
drivers/hv/mshv_root_hv_call.c
206
while (done < page_count) {
drivers/hv/mshv_root_hv_call.c
207
ulong i, completed, remain = page_count - done;
drivers/hv/mshv_root_hv_call.c
267
int hv_call_map_gpa_pages(u64 partition_id, u64 gpa_target, u64 page_count,
drivers/hv/mshv_root_hv_call.c
270
return hv_do_map_gpa_hcall(partition_id, gpa_target, page_count,
drivers/hv/mshv_root_hv_call.c
293
u64 status, page_count = page_count_4k;
drivers/hv/mshv_root_hv_call.c
297
if (page_count == 0)
drivers/hv/mshv_root_hv_call.c
301
if (!HV_PAGE_COUNT_2M_ALIGNED(page_count))
drivers/hv/mshv_root_hv_call.c
305
page_count >>= large_shift;
drivers/hv/mshv_root_hv_call.c
308
while (done < page_count) {
drivers/hv/mshv_root_hv_call.c
309
ulong completed, remain = page_count - done;
drivers/hv/mshv_root_hv_call.c
422
u64 page_count, struct page **pages,
drivers/hv/mshv_root_hv_call.c
433
if (page_count > HV_GET_VP_STATE_BATCH_SIZE)
drivers/hv/mshv_root_hv_call.c
436
if (!page_count && !ret_output)
drivers/hv/mshv_root_hv_call.c
449
for (i = 0; i < page_count; i++)
drivers/hv/mshv_root_hv_call.c
453
(page_count << HV_HYPERCALL_VARHEAD_OFFSET);
drivers/hv/mshv_root_hv_call.c
475
struct hv_vp_state_data state_data, u64 page_count,
drivers/hv/mshv_root_hv_call.c
486
if (page_count > HV_SET_VP_STATE_BATCH_SIZE)
drivers/hv/mshv_root_hv_call.c
494
else if (page_count)
drivers/hv/mshv_root_hv_call.c
495
varhead_sz = page_count;
drivers/hv/mshv_root_hv_call.c
510
for (i = 0; i < page_count; i++)
drivers/hv/mshv_root_main.c
1461
mshv_field_nonzero(args, rsvd) || !args.page_count ||
drivers/hv/mshv_root_main.c
1465
if (check_mul_overflow(args.page_count, sizeof(*states), &states_buf_sz))
drivers/hv/mshv_root_main.c
1469
bitmap_buf_sz = DIV_ROUND_UP(args.page_count, 8);
drivers/hv/mshv_root_main.c
1502
ret = hv_call_get_gpa_access_states(partition->pt_id, args.page_count,
drivers/hv/mshv_root_main.c
704
unsigned long user_pfn, size_t page_count,
drivers/hv/mshv_root_main.c
711
if (page_count > INT_MAX)
drivers/hv/mshv_root_main.c
718
if (check_add_overflow(user_pfn, (page_count - 1), &check))
drivers/hv/mshv_root_main.c
724
pages = kzalloc_objs(struct page *, page_count);
drivers/hv/mshv_root_main.c
728
for (completed = 0; completed < page_count; completed += ret) {
drivers/hv/mshv_root_main.c
730
int remaining = page_count - completed;
drivers/hv/mshv_root_main.c
744
state_data, page_count, pages,
drivers/hv/mshv_root_main.c
749
state_data, page_count, pages,
drivers/hv/mshv_root_main.c
834
size_t page_count = PFN_DOWN(args.buf_sz);
drivers/hv/mshv_root_main.c
837
page_count, is_set);
drivers/infiniband/hw/hns/hns_roce_mr.c
555
static inline int mtr_check_direct_pages(dma_addr_t *pages, int page_count,
drivers/infiniband/hw/hns/hns_roce_mr.c
561
for (i = 1; i < page_count; i++)
drivers/infiniband/hw/hns/hns_roce_mr.c
644
int page_count = cal_mtr_pg_cnt(mtr);
drivers/infiniband/hw/hns/hns_roce_mr.c
653
pages = kvzalloc_objs(dma_addr_t, page_count);
drivers/infiniband/hw/hns/hns_roce_mr.c
658
npage = hns_roce_get_umem_bufs(pages, page_count,
drivers/infiniband/hw/hns/hns_roce_mr.c
661
npage = hns_roce_get_kmem_bufs(hr_dev, pages, page_count,
drivers/infiniband/hw/hns/hns_roce_mr.c
664
if (npage != page_count) {
drivers/infiniband/hw/hns/hns_roce_mr.c
666
page_count);
drivers/infiniband/hw/hns/hns_roce_mr.c
681
ret = hns_roce_mtr_map(hr_dev, mtr, pages, page_count);
drivers/infiniband/hw/mana/main.c
412
create_req->page_count = num_pages_total;
drivers/infiniband/hw/mthca/mthca_cmd.c
1564
int mthca_UNMAP_ICM(struct mthca_dev *dev, u64 virt, u32 page_count)
drivers/infiniband/hw/mthca/mthca_cmd.c
1567
page_count, (unsigned long long) virt);
drivers/infiniband/hw/mthca/mthca_cmd.c
1569
return mthca_cmd(dev, virt, page_count, 0,
drivers/infiniband/hw/mthca/mthca_cmd.h
279
int mthca_UNMAP_ICM(struct mthca_dev *dev, u64 virt, u32 page_count);
drivers/md/dm-vdo/block-map.c
1350
for (info = cache->infos; info < cache->infos + cache->page_count; info++) {
drivers/md/dm-vdo/block-map.c
1359
return vdo_int_map_create(cache->page_count, &cache->page_map);
drivers/md/dm-vdo/block-map.c
186
for (info = cache->infos; info < cache->infos + cache->page_count; info++) {
drivers/md/dm-vdo/block-map.c
221
u64 size = cache->page_count * (u64) VDO_BLOCK_SIZE;
drivers/md/dm-vdo/block-map.c
224
result = vdo_allocate(cache->page_count, struct page_info, "page infos",
drivers/md/dm-vdo/block-map.c
233
result = vdo_int_map_create(cache->page_count, &cache->page_map);
drivers/md/dm-vdo/block-map.c
267
if (cache->waiter_count > cache->page_count) {
drivers/md/dm-vdo/block-map.c
2790
zone->page_cache.page_count = cache_size / map->zone_count;
drivers/md/dm-vdo/block-map.c
2791
zone->page_cache.stats.free_pages = zone->page_cache.page_count;
drivers/md/dm-vdo/block-map.c
2860
for (info = cache->infos; info < cache->infos + cache->page_count; info++)
drivers/md/dm-vdo/block-map.c
589
for (info = cache->infos; info < cache->infos + cache->page_count; info++) {
drivers/md/dm-vdo/block-map.h
53
page_count_t page_count;
drivers/md/dm-vdo/repair.c
1050
if (completion == (&repair->page_completions[repair->page_count]))
drivers/md/dm-vdo/repair.c
1139
for (i = 0; i < repair->page_count; i++) {
drivers/md/dm-vdo/repair.c
1703
page_count_t page_count = min_t(page_count_t,
drivers/md/dm-vdo/repair.c
1718
result = vdo_allocate_extended(struct repair_completion, page_count,
drivers/md/dm-vdo/repair.c
1730
repair->page_count = page_count;
drivers/md/dm-vdo/repair.c
593
for (i = 0; i < repair->page_count; i++) {
drivers/md/dm-vdo/repair.c
926
for (i = 0; i < repair->page_count; i++) {
drivers/md/dm-vdo/repair.c
95
page_count_t page_count;
drivers/media/pci/ivtv/ivtv-driver.h
271
int page_count;
drivers/media/pci/ivtv/ivtv-driver.h
291
int page_count;
drivers/media/pci/ivtv/ivtv-udma.c
102
if (dma->SG_length || dma->page_count) {
drivers/media/pci/ivtv/ivtv-udma.c
104
dma->SG_length, dma->page_count);
drivers/media/pci/ivtv/ivtv-udma.c
110
if (user_dma.page_count <= 0) {
drivers/media/pci/ivtv/ivtv-udma.c
112
user_dma.page_count, size_in_bytes, user_dma.offset);
drivers/media/pci/ivtv/ivtv-udma.c
117
err = pin_user_pages_unlocked(user_dma.uaddr, user_dma.page_count,
drivers/media/pci/ivtv/ivtv-udma.c
120
if (user_dma.page_count != err) {
drivers/media/pci/ivtv/ivtv-udma.c
122
err, user_dma.page_count);
drivers/media/pci/ivtv/ivtv-udma.c
130
dma->page_count = user_dma.page_count;
drivers/media/pci/ivtv/ivtv-udma.c
136
unpin_user_pages(dma->map, dma->page_count);
drivers/media/pci/ivtv/ivtv-udma.c
137
dma->page_count = 0;
drivers/media/pci/ivtv/ivtv-udma.c
143
dma->page_count, DMA_TO_DEVICE);
drivers/media/pci/ivtv/ivtv-udma.c
146
unpin_user_pages(dma->map, dma->page_count);
drivers/media/pci/ivtv/ivtv-udma.c
147
dma->page_count = 0;
drivers/media/pci/ivtv/ivtv-udma.c
158
return dma->page_count;
drivers/media/pci/ivtv/ivtv-udma.c
168
if (dma->page_count == 0)
drivers/media/pci/ivtv/ivtv-udma.c
173
dma_unmap_sg(&itv->pdev->dev, dma->SGlist, dma->page_count,
drivers/media/pci/ivtv/ivtv-udma.c
180
unpin_user_pages(dma->map, dma->page_count);
drivers/media/pci/ivtv/ivtv-udma.c
181
dma->page_count = 0;
drivers/media/pci/ivtv/ivtv-udma.c
197
itv->udma.page_count, DMA_TO_DEVICE);
drivers/media/pci/ivtv/ivtv-udma.c
21
dma_page->page_count = dma_page->last - dma_page->first + 1;
drivers/media/pci/ivtv/ivtv-udma.c
22
if (dma_page->page_count == 1) dma_page->tail -= dma_page->offset;
drivers/media/pci/ivtv/ivtv-udma.c
36
for (i = 0; i < dma_page->page_count; i++) {
drivers/media/pci/ivtv/ivtv-udma.c
37
unsigned int len = (i == dma_page->page_count - 1) ?
drivers/media/pci/ivtv/ivtv-yuv.c
106
dma->page_count = y_pages + uv_pages;
drivers/media/pci/ivtv/ivtv-yuv.c
111
unpin_user_pages(dma->map, dma->page_count);
drivers/media/pci/ivtv/ivtv-yuv.c
112
dma->page_count = 0;
drivers/media/pci/ivtv/ivtv-yuv.c
116
dma->page_count, DMA_TO_DEVICE);
drivers/media/pci/ivtv/ivtv-yuv.c
119
unpin_user_pages(dma->map, dma->page_count);
drivers/media/pci/ivtv/ivtv-yuv.c
120
dma->page_count = 0;
drivers/media/pci/ivtv/ivtv-yuv.c
54
if (dma->SG_length || dma->page_count) {
drivers/media/pci/ivtv/ivtv-yuv.c
57
dma->SG_length, dma->page_count);
drivers/media/pci/ivtv/ivtv-yuv.c
66
y_dma.page_count, &dma->map[0], 0);
drivers/media/pci/ivtv/ivtv-yuv.c
68
if (y_pages == y_dma.page_count) {
drivers/media/pci/ivtv/ivtv-yuv.c
70
uv_dma.page_count, &dma->map[y_pages], 0);
drivers/media/pci/ivtv/ivtv-yuv.c
73
if (y_pages != y_dma.page_count || uv_pages != uv_dma.page_count) {
drivers/media/pci/ivtv/ivtv-yuv.c
76
if (y_pages == y_dma.page_count) {
drivers/media/pci/ivtv/ivtv-yuv.c
79
uv_pages, uv_dma.page_count);
drivers/media/pci/ivtv/ivtv-yuv.c
90
y_pages, y_dma.page_count);
drivers/media/pci/ivtv/ivtvfb.c
286
__func__, size_in_bytes, itv->udma.page_count);
drivers/media/pci/ivtv/ivtvfb.c
293
size_in_bytes, itv->udma.page_count);
drivers/mtd/nand/raw/denali.c
551
const int page_count = 1;
drivers/mtd/nand/raw/denali.c
563
(write ? BIT(8) : 0) | page_count);
drivers/mtd/nand/raw/denali.c
576
const int page_count = 1;
drivers/mtd/nand/raw/denali.c
584
0x2000 | (write ? BIT(8) : 0) | page_count);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1297
(bfi_q)->pages = htons((u16)(bna_qpt)->page_count); \
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1855
u32 page_count,
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1869
rxq->qpt.page_count = page_count;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1878
for (i = 0; i < rxq->qpt.page_count; i++) {
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1893
u32 page_count,
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1907
rxp->cq.qpt.page_count = page_count;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1916
for (i = 0; i < rxp->cq.qpt.page_count; i++) {
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2278
u32 page_count;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2298
page_count = res_info[BNA_RX_RES_MEM_T_CQPT_PAGE].res_u.mem_info.len /
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2464
bna_rxp_cqpt_setup(rxp, page_count, PAGE_SIZE,
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3161
bna_txq_qpt_setup(struct bna_txq *txq, int page_count, int page_size,
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3174
txq->qpt.page_count = page_count;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3183
for (i = 0; i < page_count; i++) {
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3319
u32 page_count;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3330
page_count = q_size >> PAGE_SHIFT;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3335
mem_info->len = page_count * sizeof(struct bna_dma_addr);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3341
mem_info->len = page_count * sizeof(void *);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3347
mem_info->len = PAGE_SIZE * page_count;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3372
int page_count;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3376
page_count = (res_info[BNA_TX_RES_MEM_T_PAGE].res_u.mem_info.len) /
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3479
bna_txq_qpt_setup(txq, page_count, PAGE_SIZE,
drivers/net/ethernet/brocade/bna/bna_types.h
300
u32 page_count;
drivers/net/ethernet/cavium/liquidio/octeon_network.h
341
if (unlikely(page_count(pg_info->page) != 1) ||
drivers/net/ethernet/freescale/gianfar.c
2328
if (unlikely(page_count(page) != 1 || page_is_pfmemalloc(page)))
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
13
return page_count(bs->page_info.page) - bs->page_info.pagecnt_bias;
drivers/net/ethernet/google/gve/gve_rx.c
503
int pagecount = page_count(page_info->page);
drivers/net/ethernet/google/gve/gve_utils.c
96
int pagecount = page_count(page_info->page);
drivers/net/ethernet/hisilicon/hns/hns_enet.c
464
if (likely(page_count(desc_cb->priv) == 1)) {
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
3768
return page_count(cb->priv) == cb->pagecnt_bias;
drivers/net/ethernet/intel/fm10k/fm10k_main.c
214
if (unlikely(page_count(page) != 1))
drivers/net/ethernet/intel/i40e/i40e_txrx.c
1934
if (unlikely((rx_buffer->page_count - pagecnt_bias) > 1)) {
drivers/net/ethernet/intel/i40e/i40e_txrx.c
1988
rx_buffer->page_count =
drivers/net/ethernet/intel/i40e/i40e_txrx.c
1990
page_count(rx_buffer->page);
drivers/net/ethernet/intel/i40e/i40e_txrx.h
283
__u32 page_count;
drivers/net/ethernet/intel/igb/igb_main.c
8946
page_count(rx_buffer->page);
drivers/net/ethernet/intel/igbvf/netdev.c
317
(page_count(buffer_info->page) != 1))
drivers/net/ethernet/intel/igc/igc_main.c
1895
page_count(rx_buffer->page);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
2235
page_count(rx_buffer->page);
drivers/net/ethernet/mellanox/mlx4/en_rx.c
468
release = page_count(page) != 1 ||
drivers/net/ethernet/mellanox/mlx4/icm.c
240
static int mlx4_UNMAP_ICM(struct mlx4_dev *dev, u64 virt, u32 page_count)
drivers/net/ethernet/mellanox/mlx4/icm.c
242
return mlx4_cmd(dev, virt, page_count, 0, MLX4_CMD_UNMAP_ICM,
drivers/net/ethernet/microsoft/mana/gdma_main.c
1023
req->page_count = num_page;
drivers/net/ethernet/sfc/falcon/rx.c
126
if (page_count(page) == 1) {
drivers/net/ethernet/sfc/rx_common.c
55
if (page_count(page) == 1) {
drivers/net/ethernet/sfc/siena/rx_common.c
58
if (page_count(page) == 1) {
drivers/net/ethernet/sfc/siena/siena_sriov.c
810
u64 page_count = req->u.set_status_page.peer_page_count;
drivers/net/ethernet/sfc/siena/siena_sriov.c
816
if (!req->u.set_status_page.dma_addr || page_count > max_page_count) {
drivers/net/ethernet/sfc/siena/siena_sriov.c
832
if (page_count) {
drivers/net/ethernet/sfc/siena/siena_sriov.c
833
vf->peer_page_addrs = kcalloc(page_count, sizeof(u64),
drivers/net/ethernet/sfc/siena/siena_sriov.c
838
page_count * sizeof(u64));
drivers/net/ethernet/sfc/siena/siena_sriov.c
839
vf->peer_page_count = page_count;
drivers/net/ethernet/sun/cassini.c
1339
if (page_count(page->buffer) == 1)
drivers/net/ethernet/sun/cassini.c
1359
if (page_count(page0[index]->buffer) > 1) {
drivers/net/ethernet/sun/cassini.c
2182
if (page_count(page[entry]->buffer) > 1) {
drivers/net/ethernet/sun/cassini.c
563
if (page_count(page->buffer) > 1)
drivers/net/fjes/fjes_hw.c
1213
int page_count;
drivers/net/fjes/fjes_hw.c
1229
page_count = hw->hw_info.trace_size / FJES_DEBUG_PAGE_SIZE;
drivers/net/fjes/fjes_hw.c
1230
for (i = 0; i < page_count; i++) {
drivers/net/fjes/fjes_hw.c
509
int page_count;
drivers/net/fjes/fjes_hw.c
528
page_count = buf_pair->tx.size / EP_BUFFER_INFO_SIZE;
drivers/net/fjes/fjes_hw.c
529
for (i = 0; i < page_count; i++) {
drivers/net/fjes/fjes_hw.c
538
page_count = buf_pair->rx.size / EP_BUFFER_INFO_SIZE;
drivers/net/fjes/fjes_hw.c
539
for (i = 0; i < page_count; i++) {
drivers/net/hyperv/netvsc.c
1021
u32 page_count = packet->page_buf_cnt;
drivers/net/hyperv/netvsc.c
1028
packet->dma_range = kzalloc_objs(*packet->dma_range, page_count,
drivers/net/hyperv/netvsc.c
1033
for (i = 0; i < page_count; i++) {
drivers/net/hyperv/netvsc.c
956
u32 page_count = packet->cp_partial ? 1 : packet->page_buf_cnt;
drivers/net/hyperv/netvsc.c
967
for (i = 0; i < page_count; i++) {
drivers/net/usb/r8152.c
2500
if (page_count(agg->page) == 1) {
drivers/net/usb/r8152.c
2679
WARN_ON(!agg_free && page_count(agg->page) > 1);
drivers/net/usb/r8152.c
2683
if (page_count(agg->page) == 1) {
drivers/net/usb/r8152.c
3121
if (page_count(agg->page) > 1)
drivers/net/wireless/mediatek/mt76/mt7603/init.c
102
mt76_wr(dev, MT_GROUP_THRESH(0), page_count);
drivers/net/wireless/mediatek/mt76/mt7603/init.c
44
int page_count;
drivers/net/wireless/mediatek/mt76/mt7603/init.c
55
page_count = mt76_get_field(dev, MT_PSE_FC_P0,
drivers/net/wireless/mediatek/mt76/mt7603/init.c
73
mt76_wr(dev, MT_SCH_1, page_count | (2 << 28));
drivers/net/wireless/mediatek/mt76/mt7603/init.c
92
mt76_wr(dev, MT_RSV_MAX_THRESH, page_count - reserved_count);
drivers/net/wireless/mediatek/mt76/mt7603/init.c
96
page_count - beacon_pages - mcu_pages);
drivers/s390/net/qeth_core_main.c
2901
if (page_count(entry->elements[i]) > 1) {
drivers/s390/net/qeth_core_main.c
2916
if (page_count(entry->elements[i]) > 1) {
drivers/scsi/cxgbi/libcxgbi.c
1831
if (page_count(sg_page(sg)) < 1) {
drivers/scsi/elx/libefc_sli/sli4.c
434
u32 payload_size, page_count;
drivers/scsi/elx/libefc_sli/sli4.c
439
page_count = sli_page_count(qs[0]->dma.size, SLI_PAGE_SIZE) * num_rqs;
drivers/scsi/elx/libefc_sli/sli4.c
443
SZ_DMAADDR * page_count,
drivers/scsi/elx/libefc_sli/sli4.c
458
len = SLI4_RQST_PYLD_LEN_VAR(rq_create_v2, SZ_DMAADDR * page_count);
drivers/scsi/elx/libefc_sli/sli4.c
5054
u32 page_count, payload_size;
drivers/scsi/elx/libefc_sli/sli4.c
5056
page_count = sli_page_count(dma->size, SLI_PAGE_SIZE);
drivers/scsi/elx/libefc_sli/sli4.c
5059
(page_count * SZ_DMAADDR)) - sizeof(struct sli4_rqst_hdr));
drivers/scsi/elx/libefc_sli/sli4.c
5061
if (page_count > 16) {
drivers/scsi/elx/libefc_sli/sli4.c
5091
req->page_count = cpu_to_le16(page_count);
drivers/scsi/elx/libefc_sli/sli4.c
5093
for (i = 0; i < page_count; i++) {
drivers/scsi/elx/libefc_sli/sli4.h
3476
__le16 page_count;
drivers/scsi/lpfc/lpfc_init.c
8933
rpi_hdr->page_count = 1;
drivers/scsi/lpfc/lpfc_sli.c
15914
queue->page_count = pgcnt;
drivers/scsi/lpfc/lpfc_sli.c
15922
for (x = 0; x < queue->page_count; x++) {
drivers/scsi/lpfc/lpfc_sli.c
16170
eq->page_count);
drivers/scsi/lpfc/lpfc_sli.c
16366
cq->page_count);
drivers/scsi/lpfc/lpfc_sli.c
16400
cq->page_count);
drivers/scsi/lpfc/lpfc_sli.c
16513
length += ((numcq * cqp[0]->page_count) *
drivers/scsi/lpfc/lpfc_sli.c
16546
&cq_set->u.request, cq->page_count);
drivers/scsi/lpfc/lpfc_sli.c
16745
mq->page_count);
drivers/scsi/lpfc/lpfc_sli.c
16827
&mq_create_ext->u.request, mq->page_count);
drivers/scsi/lpfc/lpfc_sli.c
16987
wq->page_count);
drivers/scsi/lpfc/lpfc_sli.c
17297
hrq->page_count);
drivers/scsi/lpfc/lpfc_sli.c
17440
drq->page_count);
drivers/scsi/lpfc/lpfc_sli.c
17530
length += ((2 * numrq * hrqp[0]->page_count) *
drivers/scsi/lpfc/lpfc_sli.c
17572
hrq->page_count);
drivers/scsi/lpfc/lpfc_sli.c
19777
hdr_tmpl, rpi_page->page_count);
drivers/scsi/lpfc/lpfc_sli4.h
1029
uint32_t page_count;
drivers/scsi/lpfc/lpfc_sli4.h
223
uint16_t page_count; /* Number of pages allocated for this queue */
drivers/target/target_core_rd.c
100
rd_dev->rd_host->rd_host_id, rd_dev->rd_dev_id, page_count,
drivers/target/target_core_rd.c
101
rd_dev->sg_table_count, (unsigned long)page_count * PAGE_SIZE);
drivers/target/target_core_rd.c
215
u32 page_count;
drivers/target/target_core_rd.c
220
page_count = rd_release_sgl_table(rd_dev, rd_dev->sg_prot_array,
drivers/target/target_core_rd.c
225
rd_dev->rd_host->rd_host_id, rd_dev->rd_dev_id, page_count,
drivers/target/target_core_rd.c
226
rd_dev->sg_table_count, (unsigned long)page_count * PAGE_SIZE);
drivers/target/target_core_rd.c
68
u32 i, j, page_count = 0, sg_per_table;
drivers/target/target_core_rd.c
78
page_count++;
drivers/target/target_core_rd.c
85
return page_count;
drivers/target/target_core_rd.c
90
u32 page_count;
drivers/target/target_core_rd.c
95
page_count = rd_release_sgl_table(rd_dev, rd_dev->sg_table_array,
drivers/tee/optee/protmem.c
19
size_t page_count;
drivers/tee/optee/protmem.c
325
rp->page_count = min_size / PAGE_SIZE;
drivers/tee/optee/protmem.c
39
rp->protmem = tee_shm_alloc_dma_mem(rp->optee->ctx, rp->page_count);
drivers/tee/optee/smc_abi.c
685
size_t page_count;
drivers/tee/optee/smc_abi.c
725
pages = tee_shm_get_pages(shm, &page_count);
drivers/tee/optee/smc_abi.c
729
pages_list = optee_allocate_pages_list(page_count);
drivers/tee/optee/smc_abi.c
736
call_ctx->num_entries = page_count;
drivers/tee/optee/smc_abi.c
748
optee_fill_pages_list(pages_list, pages, page_count,
drivers/tee/tee_shm.c
285
size_t page_count)
drivers/tee/tee_shm.c
295
page = dma_alloc_pages(&teedev->dev, page_count * PAGE_SIZE,
drivers/tee/tee_shm.c
309
dma_mem->shm.size = page_count * PAGE_SIZE;
drivers/tee/tee_shm.c
317
dma_free_pages(&teedev->dev, page_count * PAGE_SIZE, page, dma_addr,
drivers/tee/tee_shm.c
327
size_t page_count)
drivers/vfio/pci/pds/dirty.c
143
u32 page_count)
drivers/vfio/pci/pds/dirty.c
152
max_sge = DIV_ROUND_UP(page_count, PAGE_SIZE * 8);
drivers/vfio/pci/pds/dirty.c
197
u32 page_count;
drivers/vfio/pci/pds/dirty.c
200
page_count = le32_to_cpu(ri->page_count);
drivers/vfio/pci/pds/dirty.c
202
region_size = page_count * region_page_size;
drivers/vfio/pci/pds/dirty.c
205
page_count / BITS_PER_BYTE);
drivers/vfio/pci/pds/dirty.c
212
err = pds_vfio_dirty_alloc_sgl(pds_vfio, region, page_count);
drivers/vfio/pci/pds/dirty.c
224
dev_bmp_offset_byte += page_count / BITS_PER_BYTE;
drivers/vfio/pci/pds/dirty.c
302
u32 page_count;
drivers/vfio/pci/pds/dirty.c
304
page_count = DIV_ROUND_UP(region_size, region_page_size);
drivers/vfio/pci/pds/dirty.c
307
ri->page_count = cpu_to_le32(page_count);
drivers/vfio/pci/pds/dirty.c
312
i, region_start, node->last, region_size, page_count,
drivers/vfio/pci/pds/dirty.c
65
le32_to_cpu(region_info[i].page_count),
drivers/virt/vboxguest/vboxguest_utils.c
199
u32 page_count;
drivers/virt/vboxguest/vboxguest_utils.c
201
page_count = hgcm_call_buf_size_in_pages(buf, len);
drivers/virt/vboxguest/vboxguest_utils.c
202
*extra += offsetof(struct vmmdev_hgcm_pagelist, pages[page_count]);
drivers/virt/vboxguest/vboxguest_utils.c
340
u32 i, page_count;
drivers/virt/vboxguest/vboxguest_utils.c
351
page_count = hgcm_call_buf_size_in_pages(buf, len);
drivers/virt/vboxguest/vboxguest_utils.c
359
dst_pg_lst->page_count = page_count;
drivers/virt/vboxguest/vboxguest_utils.c
361
for (i = 0; i < page_count; i++) {
drivers/virt/vboxguest/vboxguest_utils.c
371
*off_extra += offsetof(struct vmmdev_hgcm_pagelist, pages[page_count]);
drivers/w1/slaves/w1_ds2433.c
337
if (data->cfg->page_count > W1_VALIDCRC_MAX) {
drivers/w1/slaves/w1_ds2433.c
347
bitmap_zero(data->validcrc, data->cfg->page_count);
drivers/w1/slaves/w1_ds2433.c
44
unsigned int page_count; /* number of 256 bits pages */
drivers/w1/slaves/w1_ds2433.c
50
.page_count = 16,
drivers/w1/slaves/w1_ds2433.c
56
.page_count = 80,
drivers/w1/slaves/w1_ds2433.c
94
bitmap_zero(data->validcrc, data->cfg->page_count);
drivers/xen/grant-table.c
1329
if (page_count(item->pages[pc]) > 1) {
drivers/xen/privcmd.c
616
int page_count;
drivers/xen/privcmd.c
624
page_count = pin_user_pages_fast(
drivers/xen/privcmd.c
627
if (page_count <= 0)
drivers/xen/privcmd.c
628
return page_count ? : -EFAULT;
drivers/xen/privcmd.c
630
*pinned += page_count;
drivers/xen/privcmd.c
631
nr_pages -= page_count;
drivers/xen/privcmd.c
632
pages += page_count;
drivers/xen/privcmd.c
634
off = (requested == page_count) ? 0 : off + page_count;
fs/ntfs3/fslog.c
1729
page_cnt = le16_to_cpu(best_page->page_count);
fs/ntfs3/fslog.c
1741
u16 bppc = le16_to_cpu(best_page->page_count);
fs/ntfs3/fslog.c
1834
le16_to_cpu(page->page_count) == page_cnt))) {
fs/ntfs3/fslog.c
1957
(!reuse_page || page->page_pos != page->page_count)) {
fs/ntfs3/fslog.c
1966
} else if (le16_to_cpu(page->page_count) != page_cnt ||
fs/ntfs3/fslog.c
2023
page_cnt = le16_to_cpu(page->page_count);
fs/ntfs3/fslog.c
2119
page_cnt = le16_to_cpu(tst_page->page_count) + 1;
fs/ntfs3/fslog.c
2210
page->page_pos = page->page_count = cpu_to_le16(1);
fs/ntfs3/fslog.c
270
__le16 page_count; // 0x14:
fs/orangefs/orangefs-bufmap.c
152
int page_count;
fs/orangefs/orangefs-bufmap.c
171
unpin_user_pages(bufmap->page_array, bufmap->page_count);
fs/orangefs/orangefs-bufmap.c
226
bufmap->page_count = bufmap->total_size / PAGE_SIZE;
fs/orangefs/orangefs-bufmap.c
230
kzalloc_objs(struct page *, bufmap->page_count);
fs/orangefs/orangefs-bufmap.c
255
bufmap->page_count, FOLL_WRITE, bufmap->page_array);
fs/orangefs/orangefs-bufmap.c
260
if (ret != bufmap->page_count) {
fs/orangefs/orangefs-bufmap.c
262
bufmap->page_count, ret);
fs/orangefs/orangefs-bufmap.c
275
for (i = 0; i < bufmap->page_count; i++)
fs/pipe.c
130
if (page_count(page) == 1) {
fs/pipe.c
155
if (page_count(page) != 1)
fs/pipe.c
184
if (page_count(page) == 1) {
fs/pstore/ram_core.c
418
unsigned int page_count;
fs/pstore/ram_core.c
424
page_count = DIV_ROUND_UP(size + offset_in_page(start), PAGE_SIZE);
fs/pstore/ram_core.c
441
pages = kmalloc_objs(struct page *, page_count);
fs/pstore/ram_core.c
444
__func__, page_count);
fs/pstore/ram_core.c
448
for (i = 0; i < page_count; i++) {
fs/pstore/ram_core.c
456
vaddr = vmap(pages, page_count, VM_MAP | VM_IOREMAP, prot);
fs/smb/server/transport_rdma.c
1670
int page_count;
fs/smb/server/transport_rdma.c
1676
page_count = get_buf_page_count(v->iov_base, v->iov_len);
fs/smb/server/transport_rdma.c
1677
if (page_count > possible_vecs) {
fs/smb/server/transport_rdma.c
1705
page_count = get_buf_page_count(v->iov_base, v->iov_len);
fs/smb/server/transport_rdma.c
1706
if (WARN_ON_ONCE(page_count > possible_vecs)) {
fs/smb/server/transport_rdma.c
1715
possible_vecs -= page_count;
fs/squashfs/block.c
114
else if (idx == page_count - 1 && index + length != read_end)
fs/squashfs/block.c
178
for (idx = 0; idx < page_count; idx++) {
fs/squashfs/block.c
227
const int page_count = DIV_ROUND_UP(total_len + offset, PAGE_SIZE);
fs/squashfs/block.c
231
bio = bio_kmalloc(page_count, GFP_NOIO);
fs/squashfs/block.c
234
bio_init_inline(bio, sb->s_bdev, page_count, REQ_OP_READ);
fs/squashfs/block.c
237
for (i = 0; i < page_count; ++i) {
fs/squashfs/block.c
264
page_count);
fs/squashfs/block.c
81
u64 read_start, u64 read_end, int page_count)
fs/squashfs/block.c
91
struct folio **cache_folios = kmalloc_objs(*cache_folios, page_count,
include/linux/agp_backend.h
74
size_t page_count;
include/linux/ceph/messenger.h
248
unsigned short page_count; /* pages in array */
include/linux/firewire.h
530
int page_count;
include/linux/firewire.h
534
int page_count, enum dma_data_direction direction);
include/linux/net.h
320
return !PageSlab(page) && page_count(page) >= 1;
include/linux/page_ref.h
288
VM_BUG_ON_PAGE(page_count(page) != 0, page);
include/linux/pds/pds_adminq.h
1010
__le32 page_count;
include/linux/psp-sev.h
810
u32 page_count;
include/linux/relay.h
64
unsigned int page_count; /* number of current buffer pages */
include/linux/tee_core.h
342
size_t page_count);
include/net/mana/gdma.h
797
u32 page_count;
include/uapi/linux/mshv.h
217
__u64 page_count;
include/uapi/linux/vbox_vmmdev_types.h
284
__u16 page_count; /** Number of pages. */
kernel/power/snapshot.c
1246
unsigned long pfn, max_zone_pfn, page_count = WD_PAGE_COUNT;
kernel/power/snapshot.c
1260
if (!--page_count) {
kernel/power/snapshot.c
1262
page_count = WD_PAGE_COUNT;
kernel/power/snapshot.c
1279
if (!--page_count) {
kernel/power/snapshot.c
1281
page_count = WD_PAGE_COUNT;
kernel/relay.c
130
buf->page_count = n_pages;
kernel/relay.c
198
for (i = 0; i < buf->page_count; i++)
mm/internal.h
733
VM_BUG_ON_PAGE(page_count(buddy) != 0, buddy);
mm/kmemleak.c
1770
if (page_count(page) == 0)
mm/memory-failure.c
1404
if (page_count(p)) {
mm/memory-failure.c
2486
VM_BUG_ON_PAGE(!page_count(p), p);
mm/memory-failure.c
2877
pfn, msg_page[huge], page_count(page), &page->flags.f);
mm/memory-failure.c
981
int count = page_count(p) - 1;
mm/memory_hotplug.c
1769
if (PageOffline(page) && page_count(page))
mm/page_alloc.c
3141
VM_WARN_ON_PAGE(!page_count(page), page);
mm/page_alloc.c
7456
BUG_ON(page_count(page));
mm/page_alloc.c
7463
BUG_ON(page_count(page));
mm/page_alloc.c
7564
if (page_count(page_head) > 0)
mm/page_isolation.c
579
PageOffline(page) && !page_count(page))
mm/util.c
1216
else if (page_count(page) == 0 && is_free_buddy_page(page))
net/ceph/messenger.c
1006
VM_BUG_ON_PAGE(!PageWriteback(page) && page_count(page) < 2, page);
net/ceph/messenger.c
836
int page_count;
net/ceph/messenger.c
844
page_count = calc_pages_for(data->alignment, (u64)data->length);
net/ceph/messenger.c
847
BUG_ON(page_count > (int)USHRT_MAX);
net/ceph/messenger.c
848
cursor->page_count = (unsigned short)page_count;
net/ceph/messenger.c
860
BUG_ON(cursor->page_index >= cursor->page_count);
net/ceph/messenger.c
887
BUG_ON(cursor->page_index >= cursor->page_count);
sound/soc/sof/amd/acp-loader.c
171
u32 page_count, dma_size;
sound/soc/sof/amd/acp-loader.c
181
page_count = PAGE_ALIGN(size_fw) >> PAGE_SHIFT;
sound/soc/sof/amd/acp-loader.c
182
adata->fw_bin_page_count = page_count;
sound/soc/sof/amd/acp-loader.c
184
configure_pte_for_fw_loading(FW_BIN, page_count, adata);
sound/soc/sof/amd/acp-loader.c
193
src_addr = ACP_SYSTEM_MEMORY_WINDOW + (page_count * ACP_PAGE_SIZE);
sound/soc/sof/amd/acp-loader.c
208
(page_count * ACP_PAGE_SIZE);
sound/soc/sof/amd/acp-loader.c
55
u32 dma_size, page_count;
sound/soc/sof/amd/acp-loader.c
64
page_count = PAGE_ALIGN(size_fw) >> PAGE_SHIFT;
sound/soc/sof/amd/acp-loader.c
65
dma_size = page_count * ACP_PAGE_SIZE;
tools/testing/radix-tree/regression2.c
61
unsigned long page_count = 0;
tools/testing/radix-tree/regression2.c
71
p->index = page_count++;
tools/testing/selftests/kvm/s390/cmma_test.c
83
[page_count] "L"(TEST_DATA_PAGE_COUNT)
tools/testing/selftests/mm/ksm_tests.c
321
long page_count, int timeout, size_t page_size)
tools/testing/selftests/mm/ksm_tests.c
332
map_ptr = allocate_memory(NULL, prot, mapping, '*', page_size * page_count);
tools/testing/selftests/mm/ksm_tests.c
336
if (ksm_merge_pages(merge_type, map_ptr, page_size * page_count, start_time, timeout))
tools/testing/selftests/mm/ksm_tests.c
340
if (assert_ksm_pages_count(page_count)) {
tools/testing/selftests/mm/ksm_tests.c
342
munmap(map_ptr, page_size * page_count);
tools/testing/selftests/mm/ksm_tests.c
350
munmap(map_ptr, page_size * page_count);
tools/testing/selftests/mm/ksm_tests.c
358
int page_count = 2;
tools/testing/selftests/mm/ksm_tests.c
366
map_ptr = allocate_memory(NULL, prot, mapping, '*', page_size * page_count);
tools/testing/selftests/mm/ksm_tests.c
370
if (ksm_merge_pages(merge_type, map_ptr, page_size * page_count, start_time, timeout))
tools/testing/selftests/mm/ksm_tests.c
384
munmap(map_ptr, page_size * page_count);
tools/testing/selftests/mm/ksm_tests.c
390
munmap(map_ptr, page_size * page_count);
tools/testing/selftests/mm/ksm_tests.c
394
static int check_ksm_zero_page_merge(int merge_type, int mapping, int prot, long page_count,
tools/testing/selftests/mm/ksm_tests.c
409
map_ptr = allocate_memory(NULL, prot, mapping, 0, page_size * page_count);
tools/testing/selftests/mm/ksm_tests.c
413
if (ksm_merge_pages(merge_type, map_ptr, page_size * page_count, start_time, timeout))
tools/testing/selftests/mm/ksm_tests.c
425
else if (!use_zero_pages && !assert_ksm_pages_count(page_count))
tools/testing/selftests/mm/ksm_tests.c
429
munmap(map_ptr, page_size * page_count);
tools/testing/selftests/mm/ksm_tests.c
434
munmap(map_ptr, page_size * page_count);
tools/testing/selftests/mm/ksm_tests.c
464
int page_count = 2;
tools/testing/selftests/mm/ksm_tests.c
506
if (merge_across_nodes && !assert_ksm_pages_count(page_count))
tools/testing/selftests/mm/ksm_tests.c
691
size_t page_count = 4000;
tools/testing/selftests/mm/ksm_tests.c
693
map_ptr = allocate_memory(NULL, prot, mapping, '*', page_size * page_count);
tools/testing/selftests/mm/ksm_tests.c
701
for (size_t i = 0; i < page_count - 1; i = i + 2)
tools/testing/selftests/mm/ksm_tests.c
711
printf("Total size: %lu MiB\n\n", (page_size * page_count) / MB);
tools/testing/selftests/mm/ksm_tests.c
715
printf("Average speed: %.3f MiB/s\n\n", ((page_size * (page_count / 2)) / MB) /
tools/testing/selftests/mm/ksm_tests.c
719
for (size_t i = 0; i < page_count - 1; i = i + 2) {
tools/testing/selftests/mm/ksm_tests.c
723
if (ksm_merge_pages(merge_type, map_ptr, page_size * page_count, start_time, timeout))
tools/testing/selftests/mm/ksm_tests.c
730
for (size_t i = 0; i < page_count - 1; i = i + 2)
tools/testing/selftests/mm/ksm_tests.c
743
printf("Average speed: %.3f MiB/s\n", ((page_size * (page_count / 2)) / MB) /
tools/testing/selftests/mm/ksm_tests.c
746
munmap(map_ptr, page_size * page_count);
tools/testing/selftests/mm/ksm_tests.c
751
munmap(map_ptr, page_size * page_count);
tools/testing/selftests/mm/ksm_tests.c
761
long page_count = KSM_PAGE_COUNT_DEFAULT;
tools/testing/selftests/mm/ksm_tests.c
775
page_count = atol(optarg);
tools/testing/selftests/mm/ksm_tests.c
776
if (page_count <= 0) {
tools/testing/selftests/mm/ksm_tests.c
869
ksm_write_sysfs(KSM_FP("pages_to_scan"), page_count))
tools/testing/selftests/mm/ksm_tests.c
874
ret = check_ksm_merge(merge_type, MAP_PRIVATE | MAP_ANONYMOUS, prot, page_count,
tools/testing/selftests/mm/ksm_tests.c
883
page_count, ksm_scan_limit_sec, use_zero_pages,