Symbol: page_array
arch/x86/kernel/cpu/sgx/driver.c
27
xa_init(&encl->page_array);
arch/x86/kernel/cpu/sgx/encl.c
287
entry = xa_load(&encl->page_array, PFN_DOWN(addr));
arch/x86/kernel/cpu/sgx/encl.c
307
entry = xa_load(&encl->page_array, PFN_DOWN(addr));
arch/x86/kernel/cpu/sgx/encl.c
379
ret = xa_insert(&encl->page_array, PFN_DOWN(encl_page->desc),
arch/x86/kernel/cpu/sgx/encl.c
417
xa_erase(&encl->page_array, PFN_DOWN(encl_page->desc));
arch/x86/kernel/cpu/sgx/encl.c
457
(!xa_load(&encl->page_array, PFN_DOWN(addr))))
arch/x86/kernel/cpu/sgx/encl.c
530
XA_STATE(xas, &encl->page_array, PFN_DOWN(start));
arch/x86/kernel/cpu/sgx/encl.c
708
XA_STATE(xas, &encl->page_array, PFN_DOWN(encl->base));
arch/x86/kernel/cpu/sgx/encl.c
741
xa_destroy(&encl->page_array);
arch/x86/kernel/cpu/sgx/encl.c
86
entry = xa_load(&encl->page_array, PFN_DOWN(addr));
arch/x86/kernel/cpu/sgx/encl.h
58
struct xarray page_array;
arch/x86/kernel/cpu/sgx/ioctl.c
1135
xa_erase(&encl->page_array, PFN_DOWN(entry->desc));
arch/x86/kernel/cpu/sgx/ioctl.c
307
ret = xa_insert(&encl->page_array, PFN_DOWN(encl_page->desc),
arch/x86/kernel/cpu/sgx/ioctl.c
339
xa_erase(&encl->page_array, PFN_DOWN(encl_page->desc));
arch/x86/kernel/cpu/sgx/virt.c
160
xa_for_each(&vepc->page_array, index, entry) {
arch/x86/kernel/cpu/sgx/virt.c
197
xa_for_each(&vepc->page_array, index, entry) {
arch/x86/kernel/cpu/sgx/virt.c
207
xa_erase(&vepc->page_array, index);
arch/x86/kernel/cpu/sgx/virt.c
215
xa_for_each(&vepc->page_array, index, entry) {
arch/x86/kernel/cpu/sgx/virt.c
226
xa_erase(&vepc->page_array, index);
arch/x86/kernel/cpu/sgx/virt.c
23
struct xarray page_array;
arch/x86/kernel/cpu/sgx/virt.c
256
xa_destroy(&vepc->page_array);
arch/x86/kernel/cpu/sgx/virt.c
271
xa_init(&vepc->page_array);
arch/x86/kernel/cpu/sgx/virt.c
46
epc_page = xa_load(&vepc->page_array, index);
arch/x86/kernel/cpu/sgx/virt.c
54
ret = xa_err(xa_store(&vepc->page_array, index, epc_page, GFP_KERNEL));
arch/x86/kernel/cpu/sgx/virt.c
69
xa_erase(&vepc->page_array, index);
drivers/firewire/core-iso.c
33
struct page **page_array __free(kfree) = kzalloc_objs(page_array[0],
drivers/firewire/core-iso.c
36
if (!page_array)
drivers/firewire/core-iso.c
43
page_count, page_array);
drivers/firewire/core-iso.c
46
release_pages(page_array, nr_populated);
drivers/firewire/core-iso.c
51
buffer->pages = no_free_ptr(page_array);
drivers/infiniband/sw/siw/siw_qp_tx.c
443
struct page *page_array[MAX_ARRAY];
drivers/infiniband/sw/siw/siw_qp_tx.c
512
page_array[seg] = p;
drivers/infiniband/sw/siw/siw_qp_tx.c
540
page_array[seg] = ib_virt_dma_to_page(va);
drivers/infiniband/sw/siw/siw_qp_tx.c
593
rv = siw_0copy_tx(s, page_array, &wqe->sqe.sge[c_tx->sge_idx],
drivers/s390/cio/vfio_ccw_cp.c
131
static int page_array_pin(struct page_array *pa, struct vfio_device *vdev, bool unaligned)
drivers/s390/cio/vfio_ccw_cp.c
169
static void page_array_unpin_free(struct page_array *pa, struct vfio_device *vdev, bool unaligned)
drivers/s390/cio/vfio_ccw_cp.c
176
static bool page_array_iova_pinned(struct page_array *pa, u64 iova, u64 length)
drivers/s390/cio/vfio_ccw_cp.c
192
static inline void page_array_idal_create_words(struct page_array *pa,
drivers/s390/cio/vfio_ccw_cp.c
39
struct page_array *ch_pa;
drivers/s390/cio/vfio_ccw_cp.c
58
static int page_array_alloc(struct page_array *pa, unsigned int len)
drivers/s390/cio/vfio_ccw_cp.c
608
struct page_array *pa,
drivers/s390/cio/vfio_ccw_cp.c
685
struct page_array *pa,
drivers/s390/cio/vfio_ccw_cp.c
816
struct page_array *pa;
drivers/s390/cio/vfio_ccw_cp.c
91
static void page_array_unpin(struct page_array *pa,
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
606
static void free_pages_bulk_array(unsigned long nr_pages, struct page **page_array)
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
611
__free_pages(page_array[i], 0);
fs/btrfs/extent_io.c
672
int btrfs_alloc_page_array(unsigned int nr_pages, struct page **page_array,
fs/btrfs/extent_io.c
681
allocated = alloc_pages_bulk(gfp, nr_pages, page_array);
fs/btrfs/extent_io.c
685
__free_page(page_array[i]);
fs/btrfs/extent_io.c
686
page_array[i] = NULL;
fs/btrfs/extent_io.c
701
struct page *page_array[INLINE_EXTENT_BUFFER_PAGES] = { 0 };
fs/btrfs/extent_io.c
705
ret = btrfs_alloc_page_array(num_pages, page_array, nofail);
fs/btrfs/extent_io.c
710
eb->folios[i] = page_folio(page_array[i]);
fs/btrfs/extent_io.h
365
int btrfs_alloc_page_array(unsigned int nr_pages, struct page **page_array,
fs/nfs/blocklayout/blocklayout.c
248
header->page_array.npages, f_offset,
fs/nfs/blocklayout/blocklayout.c
260
for (i = pg_index; i < header->page_array.npages; i++) {
fs/nfs/blocklayout/blocklayout.c
293
header->page_array.npages - i,
fs/nfs/blocklayout/blocklayout.c
408
for (i = pg_index; i < header->page_array.npages; i++) {
fs/nfs/blocklayout/blocklayout.c
422
bio = do_add_page_to_bio(bio, header->page_array.npages - i,
fs/nfs/localio.c
323
iocb->bvec = kmalloc_objs(struct bio_vec, hdr->page_array.npages, flags);
fs/nfs/localio.c
473
struct page **pagevec = hdr->page_array.pagevec;
fs/nfs/localio.c
483
while (total && v < hdr->page_array.npages) {
fs/nfs/pagelist.c
659
if (hdr->page_array.pagevec != hdr->page_array.page_array)
fs/nfs/pagelist.c
660
kfree(hdr->page_array.pagevec);
fs/nfs/pagelist.c
696
hdr->args.pages = hdr->page_array.pagevec;
fs/nfs/pagelist.c
885
struct nfs_page_array *pg_array = &hdr->page_array;
fs/nfs/pagelist.c
893
if (pagecount <= ARRAY_SIZE(pg_array->page_array))
fs/nfs/pagelist.c
894
pg_array->pagevec = pg_array->page_array;
fs/nfs/pagelist.c
907
pages = hdr->page_array.pagevec;
fs/orangefs/orangefs-bufmap.c
142
struct page **page_array; /* array of mapped pages */
fs/orangefs/orangefs-bufmap.c
154
struct page **page_array;
fs/orangefs/orangefs-bufmap.c
171
unpin_user_pages(bufmap->page_array, bufmap->page_count);
fs/orangefs/orangefs-bufmap.c
177
kfree(bufmap->page_array);
fs/orangefs/orangefs-bufmap.c
229
bufmap->page_array =
fs/orangefs/orangefs-bufmap.c
231
if (!bufmap->page_array)
fs/orangefs/orangefs-bufmap.c
255
bufmap->page_count, FOLL_WRITE, bufmap->page_array);
fs/orangefs/orangefs-bufmap.c
265
unpin_user_page(bufmap->page_array[i]);
fs/orangefs/orangefs-bufmap.c
276
flush_dcache_page(bufmap->page_array[i]);
fs/orangefs/orangefs-bufmap.c
280
bufmap->desc_array[i].page_array = &bufmap->page_array[offset];
fs/orangefs/orangefs-bufmap.c
482
struct page *page = to->page_array[i];
fs/orangefs/orangefs-bufmap.c
511
struct page *page = from->page_array[i];
include/linux/bpf.h
2658
unsigned long nr_pages, struct page **page_array);
include/linux/gfp.h
239
struct page **page_array);
include/linux/gfp.h
244
struct page **page_array);
include/linux/gfp.h
254
struct page **page_array)
include/linux/gfp.h
259
return alloc_pages_bulk_noprof(gfp, nid, NULL, nr_pages, page_array);
include/linux/nfs_xdr.h
1631
struct page *page_array[NFS_PAGEVEC_SIZE];
include/linux/nfs_xdr.h
1682
struct nfs_page_array page_array;
include/linux/relay.h
63
struct page **page_array; /* array of current buffer pages */
io_uring/rsrc.c
680
struct page **page_array = *pages, **new_array = NULL;
io_uring/rsrc.c
691
struct page *p = compound_head(page_array[j]);
io_uring/rsrc.c
695
WARN_ON_ONCE(i > 0 && p != page_array[j]);
io_uring/rsrc.c
709
kvfree(page_array);
io_uring/rsrc.c
715
bool io_check_coalesce_buffer(struct page **page_array, int nr_pages,
io_uring/rsrc.c
718
struct folio *folio = page_folio(page_array[0]);
io_uring/rsrc.c
724
data->first_folio_page_idx = folio_page_idx(folio, page_array[0]);
io_uring/rsrc.c
731
if (page_folio(page_array[i]) == folio &&
io_uring/rsrc.c
732
page_array[i] == page_array[i-1] + 1) {
io_uring/rsrc.c
738
if (folio_page_idx(folio, page_array[i-1]) !=
io_uring/rsrc.c
747
folio = page_folio(page_array[i]);
io_uring/rsrc.c
749
folio_page_idx(folio, page_array[i]) != 0)
io_uring/rsrc.h
93
bool io_check_coalesce_buffer(struct page **page_array, int nr_pages,
kernel/relay.c
116
buf->page_array = relay_alloc_page_array(n_pages);
kernel/relay.c
117
if (!buf->page_array)
kernel/relay.c
121
buf->page_array[i] = alloc_page(GFP_KERNEL | __GFP_ZERO);
kernel/relay.c
122
if (unlikely(!buf->page_array[i]))
kernel/relay.c
124
set_page_private(buf->page_array[i], (unsigned long)buf);
kernel/relay.c
126
mem = vmap(buf->page_array, n_pages, VM_MAP, PAGE_KERNEL);
kernel/relay.c
135
__free_page(buf->page_array[j]);
kernel/relay.c
136
relay_free_page_array(buf->page_array);
kernel/relay.c
199
__free_page(buf->page_array[i]);
kernel/relay.c
200
relay_free_page_array(buf->page_array);
mm/kasan/shadow.c
342
struct page **page_array = pages;
mm/kasan/shadow.c
347
___free_pages_bulk(page_array, nr_total - nr_pages);
mm/mempolicy.c
2597
struct page **page_array)
mm/mempolicy.c
2615
page_array);
mm/mempolicy.c
2620
nr_pages_per_node, page_array);
mm/mempolicy.c
2623
page_array += nr_allocated;
mm/mempolicy.c
2632
struct page **page_array)
mm/mempolicy.c
2670
page_array);
mm/mempolicy.c
2671
page_array += nr_allocated;
mm/mempolicy.c
2735
page_array);
mm/mempolicy.c
2736
page_array += nr_allocated;
mm/mempolicy.c
2750
struct page **page_array)
mm/mempolicy.c
2759
nr_pages, page_array);
mm/mempolicy.c
2764
page_array + nr_allocated);
mm/mempolicy.c
2775
unsigned long nr_pages, struct page **page_array)
mm/mempolicy.c
2786
nr_pages, page_array);
mm/mempolicy.c
2790
gfp, pol, nr_pages, page_array);
mm/mempolicy.c
2794
numa_node_id(), pol, nr_pages, page_array);
mm/mempolicy.c
2799
nr_pages, page_array);
mm/page_alloc.c
5067
struct page **page_array)
mm/page_alloc.c
5084
while (nr_populated < nr_pages && page_array[nr_populated])
mm/page_alloc.c
5173
if (page_array[nr_populated]) {
mm/page_alloc.c
5192
page_array[nr_populated++] = page;
mm/page_alloc.c
5206
page_array[nr_populated++] = page;