drivers/md/dm-crypt.c
1708
pages = mempool_alloc(&cc->page_pool, gfp_mask);
drivers/md/dm-crypt.c
1747
mempool_free(&fi.folio->page, &cc->page_pool);
drivers/md/dm-crypt.c
230
mempool_t page_pool;
drivers/md/dm-crypt.c
2756
mempool_exit(&cc->page_pool);
drivers/md/dm-crypt.c
3312
ret = mempool_init(&cc->page_pool, BIO_MAX_VECS, crypt_page_alloc, crypt_page_free, cc);
drivers/net/ethernet/airoha/airoha_eth.c
1436
if (qdma->q_rx[i].page_pool)
drivers/net/ethernet/airoha/airoha_eth.c
1437
page_pool_destroy(qdma->q_rx[i].page_pool);
drivers/net/ethernet/airoha/airoha_eth.c
553
page = page_pool_dev_alloc_frag(q->page_pool, &offset,
drivers/net/ethernet/airoha/airoha_eth.c
606
enum dma_data_direction dir = page_pool_get_dma_dir(q->page_pool);
drivers/net/ethernet/airoha/airoha_eth.c
701
page_pool_put_full_page(q->page_pool, page, true);
drivers/net/ethernet/airoha/airoha_eth.c
763
q->page_pool = page_pool_create(&pp_params);
drivers/net/ethernet/airoha/airoha_eth.c
764
if (IS_ERR(q->page_pool)) {
drivers/net/ethernet/airoha/airoha_eth.c
765
int err = PTR_ERR(q->page_pool);
drivers/net/ethernet/airoha/airoha_eth.c
767
q->page_pool = NULL;
drivers/net/ethernet/airoha/airoha_eth.c
807
page_pool_get_dma_dir(q->page_pool));
drivers/net/ethernet/airoha/airoha_eth.c
808
page_pool_put_full_page(q->page_pool, page, false);
drivers/net/ethernet/airoha/airoha_eth.h
198
struct page_pool *page_pool;
drivers/net/ethernet/apm/xgene/xgene_enet_cle.c
709
if (pdata->rx_ring[idx]->page_pool) {
drivers/net/ethernet/apm/xgene/xgene_enet_cle.c
710
pool_id = pdata->rx_ring[idx]->page_pool->id;
drivers/net/ethernet/apm/xgene/xgene_enet_cle.c
786
if (pdata->rx_ring[0]->page_pool) {
drivers/net/ethernet/apm/xgene/xgene_enet_cle.c
787
pool_id = pdata->rx_ring[0]->page_pool->id;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
1062
struct xgene_enet_desc_ring *buf_pool, *page_pool;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
1081
page_pool = ring->page_pool;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
1082
if (page_pool) {
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
1083
xgene_enet_delete_pagepool(page_pool);
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
1084
xgene_enet_delete_ring(page_pool);
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
1085
pdata->port_ops->clear(pdata, page_pool);
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
1149
struct xgene_enet_desc_ring *page_pool;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
1179
page_pool = ring->page_pool;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
1180
if (page_pool) {
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
1181
p = page_pool->frag_page;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
1185
p = page_pool->frag_dma_addr;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
1309
struct xgene_enet_desc_ring *page_pool = NULL;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
1370
page_pool = xgene_enet_create_desc_ring(ndev, ring_num++,
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
1373
if (!page_pool) {
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
1378
slots = page_pool->slots;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
1379
page_pool->frag_page = devm_kcalloc(dev, slots,
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
1382
if (!page_pool->frag_page) {
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
1387
page_pool->frag_dma_addr = devm_kcalloc(dev, slots,
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
1390
if (!page_pool->frag_dma_addr) {
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
1395
page_pool->dst_ring_num = xgene_enet_dst_ring_num(page_pool);
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
1396
rx_ring->page_pool = page_pool;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
1814
struct xgene_enet_desc_ring *page_pool;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
1835
page_pool = pdata->rx_ring[i]->page_pool;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
1836
xgene_enet_init_bufpool(page_pool);
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
1843
ret = xgene_enet_refill_pagepool(page_pool, count);
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
1869
page_pool = pdata->rx_ring[0]->page_pool;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
1870
ring_id = (page_pool) ? page_pool->id : 0;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
672
struct xgene_enet_desc_ring *buf_pool, *page_pool;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
690
page_pool = rx_ring->page_pool;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
719
xgene_enet_free_pagepool(page_pool, raw_desc, exp_desc);
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
729
slots = page_pool->slots - 1;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
730
head = page_pool->head;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
741
page = page_pool->frag_page[head];
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
747
page_pool->frag_page[head] = NULL;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
751
page_pool->head = head;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
764
ret = xgene_enet_refill_pagepool(page_pool, NUM_NXTBUFPOOL);
drivers/net/ethernet/apm/xgene/xgene_enet_main.h
120
struct xgene_enet_desc_ring *page_pool;
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
1052
netmem = page_pool_alloc_frag_netmem(rxr->page_pool, offset,
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
1055
netmem = page_pool_alloc_netmems(rxr->page_pool, gfp);
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
374
page_pool_recycle_direct_netmem(rxr->page_pool, netmem);
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
545
page_pool_destroy(rxr->page_pool);
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
547
rxr->page_pool = rxr->head_pool = NULL;
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
568
struct page_pool *pool;
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
583
rxr->page_pool = pool;
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
599
page_pool_destroy(rxr->page_pool);
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
600
rxr->page_pool = NULL;
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
607
page_pool_enable_direct_recycling(rxr->page_pool, &rxr->bnapi->napi);
drivers/net/ethernet/broadcom/bnge/bnge_netdev.h
441
struct page_pool *page_pool;
drivers/net/ethernet/broadcom/bnge/bnge_netdev.h
442
struct page_pool *head_pool;
drivers/net/ethernet/broadcom/bnge/bnge_txrx.c
223
page_pool_dma_sync_netmem_for_cpu(rxr->page_pool, netmem, 0,
drivers/net/ethernet/broadcom/bnx2x/bnx2x.h
618
struct bnx2x_alloc_pool page_pool;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
552
struct bnx2x_alloc_pool *pool = &fp->page_pool;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h
1015
if (!fp->page_pool.page)
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h
1024
bnx2x_free_rx_mem_pool(bp, &fp->page_pool);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
1160
page_pool_recycle_direct(rxr->page_pool, page);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
1198
page_pool_recycle_direct(rxr->page_pool, page);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
1328
page_pool_dma_sync_netmem_for_cpu(rxr->page_pool, netmem, 0,
drivers/net/ethernet/broadcom/bnxt/bnxt.c
16049
rc = bnxt_alloc_rx_page_pool(bp, clone, rxr->page_pool->p.nid);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
16059
clone->page_pool);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
16105
page_pool_destroy(clone->page_pool);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
16107
clone->page_pool = NULL;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
16123
page_pool_destroy(rxr->page_pool);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
16125
rxr->page_pool = NULL;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
16216
rxr->page_pool = clone->page_pool;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
16299
page_pool_disable_direct_recycling(rxr->page_pool);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
3503
page_pool_recycle_direct(rxr->page_pool, data);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
3525
page_pool_recycle_direct_netmem(rxr->page_pool, netmem);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
3815
page_pool_destroy(rxr->page_pool);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
3817
rxr->page_pool = rxr->head_pool = NULL;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
3852
struct page_pool *pool;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
3871
rxr->page_pool = pool;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
3891
page_pool_destroy(rxr->page_pool);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
3892
rxr->page_pool = NULL;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
3899
page_pool_enable_direct_recycling(rxr->page_pool, &rxr->bnapi->napi);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
3948
rxr->page_pool);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
4296
rxr->page_pool->p.napi = NULL;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
4297
rxr->page_pool = NULL;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
916
page = page_pool_dev_alloc_frag(rxr->page_pool, offset,
drivers/net/ethernet/broadcom/bnxt/bnxt.c
919
page = page_pool_dev_alloc_pages(rxr->page_pool);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
937
netmem = page_pool_alloc_frag_netmem(rxr->page_pool, offset,
drivers/net/ethernet/broadcom/bnxt/bnxt.c
940
netmem = page_pool_alloc_netmems(rxr->page_pool, gfp);
drivers/net/ethernet/broadcom/bnxt/bnxt.h
1121
struct page_pool *page_pool;
drivers/net/ethernet/broadcom/bnxt/bnxt.h
1122
struct page_pool *head_pool;
drivers/net/ethernet/broadcom/bnxt/bnxt.h
42
struct page_pool;
drivers/net/ethernet/broadcom/bnxt/bnxt_xdp.c
157
page_pool_recycle_direct(rxr->page_pool, tx_buf->page);
drivers/net/ethernet/broadcom/bnxt/bnxt_xdp.c
215
page_pool_recycle_direct(rxr->page_pool, page);
drivers/net/ethernet/broadcom/bnxt/bnxt_xdp.c
310
page_pool_recycle_direct(rxr->page_pool, page);
drivers/net/ethernet/cisco/enic/enic.h
186
struct page_pool *pool;
drivers/net/ethernet/engleder/tsnep.h
129
struct page_pool *page_pool;
drivers/net/ethernet/engleder/tsnep_main.c
1018
rx->page_pool = page_pool_create(&pp_params);
drivers/net/ethernet/engleder/tsnep_main.c
1019
if (IS_ERR(rx->page_pool)) {
drivers/net/ethernet/engleder/tsnep_main.c
1020
retval = PTR_ERR(rx->page_pool);
drivers/net/ethernet/engleder/tsnep_main.c
1021
rx->page_pool = NULL;
drivers/net/ethernet/engleder/tsnep_main.c
1086
page_pool_put_full_page(rx->page_pool, *page, false);
drivers/net/ethernet/engleder/tsnep_main.c
1100
rx->page_buffer[i] = page_pool_dev_alloc_pages(rx->page_pool);
drivers/net/ethernet/engleder/tsnep_main.c
1125
page = page_pool_dev_alloc_pages(rx->page_pool);
drivers/net/ethernet/engleder/tsnep_main.c
1324
page_pool_put_page(rx->page_pool, virt_to_head_page(xdp->data),
drivers/net/ethernet/engleder/tsnep_main.c
1428
page_pool_recycle_direct(rx->page_pool, page);
drivers/net/ethernet/engleder/tsnep_main.c
1450
dma_dir = page_pool_get_dma_dir(rx->page_pool);
drivers/net/ethernet/engleder/tsnep_main.c
1627
page = page_pool_dev_alloc_pages(rx->page_pool);
drivers/net/ethernet/engleder/tsnep_main.c
1943
rx->page_pool);
drivers/net/ethernet/engleder/tsnep_main.c
960
page_pool_put_full_page(rx->page_pool, entry->page,
drivers/net/ethernet/engleder/tsnep_main.c
968
if (rx->page_pool)
drivers/net/ethernet/engleder/tsnep_main.c
969
page_pool_destroy(rx->page_pool);
drivers/net/ethernet/freescale/fec.h
567
struct page_pool *page_pool;
drivers/net/ethernet/freescale/fec_main.c
1757
new_page = page_pool_dev_alloc_pages(rxq->page_pool);
drivers/net/ethernet/freescale/fec_main.c
1848
page_pool_recycle_direct(rxq->page_pool, page);
drivers/net/ethernet/freescale/fec_main.c
2006
page_pool_put_page(rxq->page_pool, page, sync, true);
drivers/net/ethernet/freescale/fec_main.c
3519
if (!rxq->page_pool)
drivers/net/ethernet/freescale/fec_main.c
3522
page_pool_get_stats(rxq->page_pool, &stats);
drivers/net/ethernet/freescale/fec_main.c
3827
allocator = rxq->xsk_pool ? NULL : rxq->page_pool;
drivers/net/ethernet/freescale/fec_main.c
3865
page_pool_put_full_page(rxq->page_pool,
drivers/net/ethernet/freescale/fec_main.c
3872
page_pool_destroy(rxq->page_pool);
drivers/net/ethernet/freescale/fec_main.c
3873
rxq->page_pool = NULL;
drivers/net/ethernet/freescale/fec_main.c
4023
page = page_pool_dev_alloc_pages(rxq->page_pool);
drivers/net/ethernet/freescale/fec_main.c
4481
buf[i].page = page_pool_dev_alloc_pages(rxq->page_pool);
drivers/net/ethernet/freescale/fec_main.c
4493
page_pool_put_full_page(rxq->page_pool,
drivers/net/ethernet/freescale/fec_main.c
4496
page_pool_destroy(rxq->page_pool);
drivers/net/ethernet/freescale/fec_main.c
486
rxq->page_pool = page_pool_create(&pp_params);
drivers/net/ethernet/freescale/fec_main.c
487
if (IS_ERR(rxq->page_pool)) {
drivers/net/ethernet/freescale/fec_main.c
488
err = PTR_ERR(rxq->page_pool);
drivers/net/ethernet/freescale/fec_main.c
489
rxq->page_pool = NULL;
drivers/net/ethernet/google/gve/gve.h
1301
struct page_pool *gve_rx_create_page_pool(struct gve_priv *priv,
drivers/net/ethernet/google/gve/gve.h
308
struct page_pool *page_pool;
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
229
netmem = page_pool_alloc_netmem(rx->dqo.page_pool,
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
240
buf_state->page_info.pad = rx->dqo.page_pool->p.offset;
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
245
struct page_pool *gve_rx_create_page_pool(struct gve_priv *priv,
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
274
if (rx->dqo.page_pool) {
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
286
if (rx->dqo.page_pool) {
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
316
} else if (rx->dqo.page_pool) {
drivers/net/ethernet/google/gve/gve_main.c
1268
rx->dqo.page_pool);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
100
if (rx->dqo.page_pool)
drivers/net/ethernet/google/gve/gve_rx_dqo.c
119
if (rx->dqo.page_pool)
drivers/net/ethernet/google/gve/gve_rx_dqo.c
120
page_pool_disable_direct_recycling(rx->dqo.page_pool);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
149
if (rx->dqo.page_pool)
drivers/net/ethernet/google/gve/gve_rx_dqo.c
183
if (rx->dqo.page_pool) {
drivers/net/ethernet/google/gve/gve_rx_dqo.c
184
page_pool_destroy(rx->dqo.page_pool);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
185
rx->dqo.page_pool = NULL;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
220
struct page_pool *pool;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
281
rx->dqo.page_pool = pool;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
550
if (rx->dqo.page_pool) {
drivers/net/ethernet/google/gve/gve_rx_dqo.c
582
if (rx->dqo.page_pool)
drivers/net/ethernet/google/gve/gve_rx_dqo.c
761
if (rx->dqo.page_pool) {
drivers/net/ethernet/google/gve/gve_rx_dqo.c
762
page_pool_dma_sync_netmem_for_cpu(rx->dqo.page_pool,
drivers/net/ethernet/google/gve/gve_rx_dqo.c
819
if (rx->dqo.page_pool) {
drivers/net/ethernet/google/gve/gve_rx_dqo.c
839
if (rx->dqo.page_pool)
drivers/net/ethernet/google/gve/gve_rx_dqo.c
849
} else if (!rx->ctx.skb_head && rx->dqo.page_pool &&
drivers/net/ethernet/google/gve/gve_rx_dqo.c
902
!(rx->dqo.page_pool &&
drivers/net/ethernet/google/gve/gve_rx_dqo.c
930
if (rx->dqo.page_pool)
drivers/net/ethernet/hisilicon/hibmcge/hbg_common.h
88
struct page_pool *page_pool; /* only for rx */
drivers/net/ethernet/hisilicon/hibmcge/hbg_txrx.c
539
if (!ring->page_pool)
drivers/net/ethernet/hisilicon/hibmcge/hbg_txrx.c
542
page_pool_destroy(ring->page_pool);
drivers/net/ethernet/hisilicon/hibmcge/hbg_txrx.c
543
ring->page_pool = NULL;
drivers/net/ethernet/hisilicon/hibmcge/hbg_txrx.c
562
ring->page_pool = page_pool_create(&pp_params);
drivers/net/ethernet/hisilicon/hibmcge/hbg_txrx.c
563
if (IS_ERR(ring->page_pool)) {
drivers/net/ethernet/hisilicon/hibmcge/hbg_txrx.c
564
ret = PTR_ERR(ring->page_pool);
drivers/net/ethernet/hisilicon/hibmcge/hbg_txrx.c
567
ring->page_pool = NULL;
drivers/net/ethernet/hisilicon/hibmcge/hbg_txrx.c
80
page_pool_put_full_page(ring->page_pool, buffer->page, false);
drivers/net/ethernet/hisilicon/hibmcge/hbg_txrx.c
95
if (unlikely(!ring->page_pool))
drivers/net/ethernet/hisilicon/hibmcge/hbg_txrx.c
98
buffer->page = page_pool_dev_alloc_frag(ring->page_pool, &offset, len);
drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c
778
READ_ONCE(ring->page_pool->pages_state_hold_cnt),
drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c
779
atomic_read(&ring->page_pool->pages_state_release_cnt),
drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c
780
ring->page_pool->p.pool_size,
drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c
781
ring->page_pool->p.order,
drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c
782
ring->page_pool->p.nid,
drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c
783
ring->page_pool->p.max_len / 1024);
drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c
798
if (!priv->ring[h->kinfo.num_tqps].page_pool) {
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
3418
if (ring->page_pool) {
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
3419
p = page_pool_dev_alloc_frag(ring->page_pool,
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
3459
page_pool_put_full_page(ring->page_pool, cb->priv,
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
3549
if (ret || ring->page_pool)
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
3811
if (ring->page_pool) {
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4109
page_pool_put_full_page(ring->page_pool, desc_cb->priv,
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4119
if (ring->page_pool)
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4159
if (ring->page_pool)
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
5026
ring->page_pool = page_pool_create(&pp_params);
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
5027
if (IS_ERR(ring->page_pool)) {
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
5029
PTR_ERR(ring->page_pool));
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
5030
ring->page_pool = NULL;
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
5097
if (!HNAE3_IS_TX_RING(ring) && ring->page_pool) {
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
5098
page_pool_destroy(ring->page_pool);
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
5099
ring->page_pool = NULL;
drivers/net/ethernet/hisilicon/hns3/hns3_enet.h
474
struct page_pool *page_pool;
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
105
page = page_pool_dev_alloc_frag(page_pool, &page_offset, buf_len);
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
156
err = rx_alloc_mapped_page(rxq->page_pool, rx_info,
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
186
err = rx_alloc_mapped_page(rqres->page_pool,
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
206
page_pool_put_full_page(rqres->page_pool,
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
224
page_pool_dma_sync_for_cpu(rxq->page_pool, page, rx_info->page_offset,
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
230
page_pool_put_full_page(rxq->page_pool, page, false);
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
442
rqres->page_pool = page_pool_create(&pp_params);
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
443
if (IS_ERR(rqres->page_pool)) {
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
462
page_pool_destroy(rqres->page_pool);
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
487
page_pool_destroy(rqres->page_pool);
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
522
rxq->page_pool = rqres->page_pool;
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
96
static int rx_alloc_mapped_page(struct page_pool *page_pool,
drivers/net/ethernet/huawei/hinic3/hinic3_rx.h
111
struct page_pool *page_pool;
drivers/net/ethernet/huawei/hinic3/hinic3_rx.h
89
struct page_pool *page_pool;
drivers/net/ethernet/intel/iavf/iavf_txrx.h
222
struct page_pool *pp; /* Used on Rx for buffer management */
drivers/net/ethernet/intel/ice/ice_txrx.h
272
struct page_pool *pp;
drivers/net/ethernet/intel/ice/ice_txrx.h
292
struct page_pool *hdr_pp;
drivers/net/ethernet/intel/idpf/idpf_txrx.h
533
struct page_pool *pp;
drivers/net/ethernet/intel/idpf/idpf_txrx.h
761
struct page_pool *pp;
drivers/net/ethernet/intel/idpf/idpf_txrx.h
769
struct page_pool *hdr_pp;
drivers/net/ethernet/intel/idpf/xdp.c
73
const struct page_pool *pp;
drivers/net/ethernet/intel/libeth/rx.c
164
struct page_pool *pool;
drivers/net/ethernet/marvell/mvneta.c
1943
page = page_pool_alloc_pages(rxq->page_pool,
drivers/net/ethernet/marvell/mvneta.c
2016
page_pool_put_full_page(rxq->page_pool, data, false);
drivers/net/ethernet/marvell/mvneta.c
2020
page_pool_destroy(rxq->page_pool);
drivers/net/ethernet/marvell/mvneta.c
2021
rxq->page_pool = NULL;
drivers/net/ethernet/marvell/mvneta.c
2082
page_pool_put_full_page(rxq->page_pool,
drivers/net/ethernet/marvell/mvneta.c
2086
page_pool_put_page(rxq->page_pool, virt_to_head_page(xdp->data),
drivers/net/ethernet/marvell/mvneta.c
2334
dma_dir = page_pool_get_dma_dir(rxq->page_pool);
drivers/net/ethernet/marvell/mvneta.c
2367
dma_dir = page_pool_get_dma_dir(rxq->page_pool);
drivers/net/ethernet/marvell/mvneta.c
2389
page_pool_put_full_page(rxq->page_pool, page, true);
drivers/net/ethernet/marvell/mvneta.c
2395
mvneta_swbm_build_skb(struct mvneta_port *pp, struct page_pool *pool,
drivers/net/ethernet/marvell/mvneta.c
2476
page_pool_put_full_page(rxq->page_pool, page,
drivers/net/ethernet/marvell/mvneta.c
2498
skb = mvneta_swbm_build_skb(pp, rxq->page_pool, &xdp_buf, desc_status);
drivers/net/ethernet/marvell/mvneta.c
3372
rxq->page_pool = page_pool_create(&pp_params);
drivers/net/ethernet/marvell/mvneta.c
3373
if (IS_ERR(rxq->page_pool)) {
drivers/net/ethernet/marvell/mvneta.c
3374
err = PTR_ERR(rxq->page_pool);
drivers/net/ethernet/marvell/mvneta.c
3375
rxq->page_pool = NULL;
drivers/net/ethernet/marvell/mvneta.c
3385
rxq->page_pool);
drivers/net/ethernet/marvell/mvneta.c
3394
page_pool_destroy(rxq->page_pool);
drivers/net/ethernet/marvell/mvneta.c
3395
rxq->page_pool = NULL;
drivers/net/ethernet/marvell/mvneta.c
4973
if (pp->rxqs[i].page_pool)
drivers/net/ethernet/marvell/mvneta.c
4974
page_pool_get_stats(pp->rxqs[i].page_pool, &stats);
drivers/net/ethernet/marvell/mvneta.c
719
struct page_pool *page_pool;
drivers/net/ethernet/marvell/mvpp2/mvpp2.h
1109
struct page_pool *page_pool[MVPP2_PORT_MAX_RXQ];
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
104
static struct page_pool *
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
1049
struct page_pool *pp = NULL;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
1071
pp = port->priv->page_pool[bm_pool->id];
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3001
priv->page_pool[rxq->logic_rxq]);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3007
priv->page_pool[rxq->logic_rxq +
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3589
struct page_pool *page_pool, int pool)
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3595
buf = mvpp2_buf_alloc(port, bm_pool, page_pool,
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
361
struct page_pool *page_pool)
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
363
if (page_pool)
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
364
return page_pool_dev_alloc_pages(page_pool);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
373
struct page_pool *page_pool, void *data)
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
375
if (page_pool)
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
376
page_pool_put_full_page(page_pool, virt_to_head_page(data), false);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3811
struct xdp_buff *xdp, struct page_pool *pp,
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3918
struct page_pool *pp = NULL;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3943
pp = port->priv->page_pool[pool];
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
495
struct page_pool *pp = NULL;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
505
pp = priv->page_pool[bm_pool->id];
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
5124
if (!priv->page_pool[0])
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
5136
if (priv->page_pool[0]->p.dma_dir != dma_dir)
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
570
page_pool_destroy(priv->page_pool[bm_pool->id]);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
571
priv->page_pool[bm_pool->id] = NULL;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
662
priv->page_pool[i] =
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
667
if (IS_ERR(priv->page_pool[i])) {
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
671
page_pool_destroy(priv->page_pool[j]);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
672
priv->page_pool[j] = NULL;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
674
return PTR_ERR(priv->page_pool[i]);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
763
struct page_pool *page_pool,
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
772
data = mvpp2_frag_alloc(bm_pool, page_pool);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
776
if (page_pool) {
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
380
pool->page_pool = NULL;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
391
pool->page_pool = page_pool_create(&pp_params);
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
392
if (IS_ERR(pool->page_pool)) {
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
394
return PTR_ERR(pool->page_pool);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1074
} else if (pool->page_pool) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1077
pool->page_pool);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1298
if (pool->page_pool) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1299
page_pool_put_full_page(pool->page_pool, page, true);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1361
page_pool_destroy(pool->page_pool);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1518
pool->page_pool = page_pool_create(&pp_params);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1519
if (IS_ERR(pool->page_pool)) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1521
return PTR_ERR(pool->page_pool);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
541
page = page_pool_alloc_frag(pool->page_pool, &offset, sz, GFP_ATOMIC);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
557
if (pool->page_pool)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
1288
if (pool->page_pool)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
1575
page_pool_recycle_direct(pool->page_pool, page);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.h
133
struct page_pool *page_pool;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1870
static struct page_pool *mtk_create_page_pool(struct mtk_eth *eth,
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1883
struct page_pool *pp;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1911
static void *mtk_page_pool_get_buff(struct page_pool *pp, dma_addr_t *dma_addr,
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1926
if (ring->page_pool)
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1927
page_pool_put_full_page(ring->page_pool,
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2165
page_pool_put_full_page(ring->page_pool,
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2243
if (ring->page_pool) {
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2248
new_data = mtk_page_pool_get_buff(ring->page_pool,
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2258
pktlen, page_pool_get_dma_dir(ring->page_pool));
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2274
page_pool_put_full_page(ring->page_pool,
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2800
struct page_pool *pp;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2807
ring->page_pool = pp;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2823
if (ring->page_pool) {
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2824
data = mtk_page_pool_get_buff(ring->page_pool,
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2934
if (ring->page_pool) {
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2937
page_pool_destroy(ring->page_pool);
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2938
ring->page_pool = NULL;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
4596
if (!ring->page_pool)
drivers/net/ethernet/mediatek/mtk_eth_soc.c
4599
page_pool_get_stats(ring->page_pool, &stats);
drivers/net/ethernet/mediatek/mtk_eth_soc.h
973
struct page_pool *page_pool;
drivers/net/ethernet/mellanox/mlx4/en_tx.c
354
struct page_pool *pool = ring->recycle_ring->pp;
drivers/net/ethernet/mellanox/mlx4/mlx4_en.h
333
struct page_pool *pp;
drivers/net/ethernet/mellanox/mlx5/core/en.h
67
struct page_pool;
drivers/net/ethernet/mellanox/mlx5/core/en.h
728
struct page_pool *page_pool;
drivers/net/ethernet/mellanox/mlx5/core/en.h
729
struct page_pool *hd_page_pool;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
1035
page_pool_destroy(rq->page_pool);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
1062
page_pool_destroy(rq->page_pool);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
981
rq->page_pool = page_pool_create(&pp_params);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
982
if (IS_ERR(rq->page_pool)) {
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
983
err = PTR_ERR(rq->page_pool);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
984
rq->page_pool = NULL;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
988
rq->hd_page_pool = rq->page_pool;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
991
MEM_TYPE_PAGE_POOL, rq->page_pool);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1898
if (unlikely(mlx5e_page_alloc_fragmented(rq->page_pool,
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1969
mlx5e_page_release_fragmented(rq->page_pool,
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1988
mlx5e_page_release_fragmented(rq->page_pool,
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1995
mlx5e_page_release_fragmented(rq->page_pool, &wi->linear_page);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
277
static int mlx5e_page_alloc_fragmented(struct page_pool *pp,
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
295
static void mlx5e_page_release_fragmented(struct page_pool *pp,
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
316
err = mlx5e_page_alloc_fragmented(rq->page_pool,
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
336
mlx5e_page_release_fragmented(rq->page_pool, frag->frag_page);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
591
mlx5e_page_release_fragmented(rq->page_pool,
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
637
err = mlx5e_page_alloc_fragmented(rq->page_pool, frag_page);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
685
mlx5e_page_release_fragmented(rq->page_pool, frag_page);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
720
if (rq->page_pool)
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
721
page_pool_nid_changed(rq->page_pool, numa_mem_id());
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
911
if (rq->page_pool)
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
912
page_pool_nid_changed(rq->page_pool, numa_mem_id());
drivers/net/ethernet/mellanox/mlx5/core/en_stats.c
496
struct page_pool *pool = c->rq.page_pool;
drivers/net/ethernet/mellanox/mlxsw/pci.c
1010
page_pool = page_pool_create(&pp_params);
drivers/net/ethernet/mellanox/mlxsw/pci.c
1011
if (IS_ERR(page_pool))
drivers/net/ethernet/mellanox/mlxsw/pci.c
1012
return PTR_ERR(page_pool);
drivers/net/ethernet/mellanox/mlxsw/pci.c
1014
q->u.cq.page_pool = page_pool;
drivers/net/ethernet/mellanox/mlxsw/pci.c
1024
page_pool_destroy(q->u.cq.page_pool);
drivers/net/ethernet/mellanox/mlxsw/pci.c
399
struct page_pool *page_pool;
drivers/net/ethernet/mellanox/mlxsw/pci.c
410
page_pool = cq->u.cq.page_pool;
drivers/net/ethernet/mellanox/mlxsw/pci.c
411
page_pool_dma_sync_for_cpu(page_pool, pages[page_index],
drivers/net/ethernet/mellanox/mlxsw/pci.c
436
page_pool_dma_sync_for_cpu(page_pool, page, 0, frag_size);
drivers/net/ethernet/mellanox/mlxsw/pci.c
454
page = page_pool_dev_alloc_pages(cq->u.cq.page_pool);
drivers/net/ethernet/mellanox/mlxsw/pci.c
469
page_pool_put_page(cq->u.cq.page_pool, elem_info->pages[index], -1,
drivers/net/ethernet/mellanox/mlxsw/pci.c
521
page_pool_recycle_direct(cq->u.cq.page_pool, old_pages[i]);
drivers/net/ethernet/mellanox/mlxsw/pci.c
534
page_pool_recycle_direct(cq->u.cq.page_pool, pages[i]);
drivers/net/ethernet/mellanox/mlxsw/pci.c
89
struct page_pool *page_pool;
drivers/net/ethernet/mellanox/mlxsw/pci.c
997
struct page_pool *page_pool;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1015
page_pool_dma_sync_netmem_for_cpu(qt->sub1.page_pool, netmem,
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1044
page_pool_put_full_netmem(qt->sub1.page_pool, netmem,
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1050
page_pool_put_full_page(qt->sub0.page_pool, page, !!budget);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1513
page_pool_destroy(qt->sub0.page_pool);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1514
page_pool_destroy(qt->sub1.page_pool);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1567
struct page_pool *pp;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1585
qt->sub0.page_pool = pp;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1596
qt->sub1.page_pool = pp;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2020
qt->sub0.page_pool);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2736
page_pool_enable_direct_recycling(qt->sub0.page_pool,
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2738
page_pool_enable_direct_recycling(qt->sub1.page_pool,
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2937
page_pool_disable_direct_recycling(real->sub0.page_pool);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2938
page_pool_disable_direct_recycling(real->sub1.page_pool);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
769
page_pool_put_unrefed_netmem(ring->page_pool, netmem, -1,
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
906
netmem = page_pool_dev_alloc_netmems(bdq->page_pool);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.h
129
struct page_pool *page_pool;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
104
rx->page_pool);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
107
return PTR_ERR_OR_ZERO(rx->page_pool);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
117
return PTR_ERR(rx->page_pool);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
16
page = page_pool_dev_alloc_pages(rx->page_pool);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
375
page_pool_recycle_direct(rx->page_pool,
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
468
page_pool_recycle_direct(rx->page_pool, page);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
58
page_pool_put_full_page(rx->page_pool,
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
72
page_pool_recycle_direct(rx->page_pool, page);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
811
struct page_pool *page_pool;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
817
page_pool = lan966x->rx.page_pool;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
834
page_pool_destroy(page_pool);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
841
lan966x->rx.page_pool = page_pool;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
93
rx->page_pool = page_pool_create(&pp_params);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
979
page_pool_destroy(lan966x->rx.page_pool);
drivers/net/ethernet/microchip/lan966x/lan966x_main.h
216
struct page_pool *page_pool;
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
145
page_pool_recycle_direct(rx->page_pool, page);
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
168
rx->page_pool = page_pool_create(&pp_params);
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
169
if (IS_ERR(rx->page_pool))
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
170
return PTR_ERR(rx->page_pool);
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
31
page = page_pool_dev_alloc_pages(rx->page_pool);
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
401
page_pool_destroy(rx->page_pool);
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
95
page_pool_put_full_page(rx->page_pool,
drivers/net/ethernet/microchip/sparx5/sparx5_main.h
162
struct page_pool *page_pool;
drivers/net/ethernet/microsoft/mana/mana_en.c
2013
page_pool_recycle_direct(rxq->page_pool,
drivers/net/ethernet/microsoft/mana/mana_en.c
2016
page_pool_free_va(rxq->page_pool, buf_va, true);
drivers/net/ethernet/microsoft/mana/mana_en.c
2046
page = page_pool_dev_alloc_pages(rxq->page_pool);
drivers/net/ethernet/microsoft/mana/mana_en.c
2064
page = page_pool_dev_alloc_frag(rxq->page_pool, &offset,
drivers/net/ethernet/microsoft/mana/mana_en.c
2508
page_pool_free_va(rxq->page_pool, rx_oob->buf_va, true);
drivers/net/ethernet/microsoft/mana/mana_en.c
2514
page_pool_destroy(rxq->page_pool);
drivers/net/ethernet/microsoft/mana/mana_en.c
2632
rxq->page_pool = page_pool_create(&pprm);
drivers/net/ethernet/microsoft/mana/mana_en.c
2634
if (IS_ERR(rxq->page_pool)) {
drivers/net/ethernet/microsoft/mana/mana_en.c
2635
ret = PTR_ERR(rxq->page_pool);
drivers/net/ethernet/microsoft/mana/mana_en.c
2636
rxq->page_pool = NULL;
drivers/net/ethernet/microsoft/mana/mana_en.c
2749
rxq->page_pool));
drivers/net/ethernet/microsoft/mana/mana_en.c
66
page_pool_put_full_page(rxq->page_pool, page, false);
drivers/net/ethernet/pensando/ionic/ionic_dev.h
267
struct page_pool *page_pool;
drivers/net/ethernet/pensando/ionic/ionic_lif.c
2713
err = xdp_rxq_info_reg_mem_model(rxq_info, MEM_TYPE_PAGE_POOL, q->page_pool);
drivers/net/ethernet/pensando/ionic/ionic_lif.c
2890
swap(a->q.page_pool, b->q.page_pool);
drivers/net/ethernet/pensando/ionic/ionic_lif.c
2891
a->q.page_pool->p.napi = &a->napi;
drivers/net/ethernet/pensando/ionic/ionic_lif.c
2892
if (b->q.page_pool) /* is NULL when increasing queue count */
drivers/net/ethernet/pensando/ionic/ionic_lif.c
2893
b->q.page_pool->p.napi = &b->napi;
drivers/net/ethernet/pensando/ionic/ionic_lif.c
451
page_pool_destroy(qcq->q.page_pool);
drivers/net/ethernet/pensando/ionic/ionic_lif.c
452
qcq->q.page_pool = NULL;
drivers/net/ethernet/pensando/ionic/ionic_lif.c
604
new->q.page_pool = page_pool_create(&pp_params);
drivers/net/ethernet/pensando/ionic/ionic_lif.c
605
if (IS_ERR(new->q.page_pool)) {
drivers/net/ethernet/pensando/ionic/ionic_lif.c
607
err = PTR_ERR(new->q.page_pool);
drivers/net/ethernet/pensando/ionic/ionic_lif.c
608
new->q.page_pool = NULL;
drivers/net/ethernet/pensando/ionic/ionic_lif.c
747
page_pool_destroy(new->q.page_pool);
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
136
page_pool_put_full_page(q->page_pool, buf_info->page, recycle_direct);
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
162
page_pool_dma_sync_for_cpu(q->page_pool,
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
255
page_pool_dma_sync_for_cpu(q->page_pool,
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
491
page_pool_dma_sync_for_cpu(rxq->page_pool, buf_info->page,
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
522
page_pool_dma_sync_for_cpu(rxq->page_pool, bi->page,
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
824
buf_info->page = page_pool_alloc(q->page_pool,
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
851
buf_info->page = page_pool_alloc(q->page_pool,
drivers/net/ethernet/realtek/rtase/rtase.h
329
struct page_pool *page_pool;
drivers/net/ethernet/realtek/rtase/rtase_main.c
378
page = page_pool_dev_alloc_pages(tp->page_pool);
drivers/net/ethernet/realtek/rtase/rtase_main.c
425
static void rtase_rx_ring_clear(struct page_pool *page_pool,
drivers/net/ethernet/realtek/rtase/rtase_main.c
437
page_pool_put_full_page(page_pool, page, true);
drivers/net/ethernet/realtek/rtase/rtase_main.c
615
rtase_rx_ring_clear(tp->page_pool, &tp->rx_ring[i]);
drivers/net/ethernet/realtek/rtase/rtase_main.c
617
page_pool_destroy(tp->page_pool);
drivers/net/ethernet/realtek/rtase/rtase_main.c
618
tp->page_pool = NULL;
drivers/net/ethernet/realtek/rtase/rtase_main.c
625
struct page_pool *page_pool;
drivers/net/ethernet/realtek/rtase/rtase_main.c
638
page_pool = page_pool_create(&pp_params);
drivers/net/ethernet/realtek/rtase/rtase_main.c
639
if (IS_ERR(page_pool)) {
drivers/net/ethernet/realtek/rtase/rtase_main.c
644
tp->page_pool = page_pool;
drivers/net/ethernet/renesas/ravb.h
1101
struct page_pool *rx_pool[NUM_RX_QUEUE];
drivers/net/ethernet/socionext/netsec.c
1041
page_pool_put_page(dring->page_pool, page, pkt_len,
drivers/net/ethernet/socionext/netsec.c
1218
page_pool_put_full_page(dring->page_pool, page, false);
drivers/net/ethernet/socionext/netsec.c
1230
page_pool_destroy(dring->page_pool);
drivers/net/ethernet/socionext/netsec.c
1313
dring->page_pool = page_pool_create(&pp_params);
drivers/net/ethernet/socionext/netsec.c
1314
if (IS_ERR(dring->page_pool)) {
drivers/net/ethernet/socionext/netsec.c
1315
err = PTR_ERR(dring->page_pool);
drivers/net/ethernet/socionext/netsec.c
1316
dring->page_pool = NULL;
drivers/net/ethernet/socionext/netsec.c
1325
dring->page_pool);
drivers/net/ethernet/socionext/netsec.c
286
struct page_pool *page_pool;
drivers/net/ethernet/socionext/netsec.c
742
page = page_pool_dev_alloc_pages(dring->page_pool);
drivers/net/ethernet/socionext/netsec.c
863
page_pool_get_dma_dir(rx_ring->page_pool);
drivers/net/ethernet/socionext/netsec.c
922
page_pool_put_page(dring->page_pool, page, sync, true);
drivers/net/ethernet/socionext/netsec.c
932
page_pool_put_page(dring->page_pool, page, sync, true);
drivers/net/ethernet/socionext/netsec.c
944
page_pool_put_page(dring->page_pool, page, sync, true);
drivers/net/ethernet/socionext/netsec.c
966
dma_dir = page_pool_get_dma_dir(dring->page_pool);
drivers/net/ethernet/stmicro/stmmac/stmmac.h
124
struct page_pool *page_pool;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1613
buf->page = page_pool_alloc_pages(rx_q->page_pool, gfp);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1620
buf->sec_page = page_pool_alloc_pages(rx_q->page_pool, gfp);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1653
page_pool_put_full_page(rx_q->page_pool, buf->page, false);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1657
page_pool_put_full_page(rx_q->page_pool, buf->sec_page, false);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1859
rx_q->page_pool));
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2104
if (rx_q->page_pool)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2105
page_pool_destroy(rx_q->page_pool);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2210
rx_q->page_pool = page_pool_create(&pp_params);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2211
if (IS_ERR(rx_q->page_pool)) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2212
ret = PTR_ERR(rx_q->page_pool);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2213
rx_q->page_pool = NULL;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4958
buf->page = page_pool_alloc_pages(rx_q->page_pool, gfp);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4964
buf->sec_page = page_pool_alloc_pages(rx_q->page_pool, gfp);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5610
dma_dir = page_pool_get_dma_dir(rx_q->page_pool);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5682
page_pool_put_page(rx_q->page_pool, buf->page, 0, true);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5748
page_pool_put_page(rx_q->page_pool,
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5784
page_pool_recycle_direct(rx_q->page_pool,
drivers/net/ethernet/tehuti/tn40.c
142
priv->page_pool = page_pool_create(&pp);
drivers/net/ethernet/tehuti/tn40.c
143
if (IS_ERR(priv->page_pool))
drivers/net/ethernet/tehuti/tn40.c
144
return PTR_ERR(priv->page_pool);
drivers/net/ethernet/tehuti/tn40.c
172
page_pool_destroy(priv->page_pool);
drivers/net/ethernet/tehuti/tn40.c
189
page_pool_put_full_page(priv->page_pool, dm->page,
drivers/net/ethernet/tehuti/tn40.c
203
page_pool_destroy(priv->page_pool);
drivers/net/ethernet/tehuti/tn40.c
250
page = page_pool_dev_alloc_pages(priv->page_pool);
drivers/net/ethernet/tehuti/tn40.h
146
struct page_pool *page_pool;
drivers/net/ethernet/ti/am65-cpsw-nuss.c
1364
new_page = page_pool_dev_alloc_pages(flow->page_pool);
drivers/net/ethernet/ti/am65-cpsw-nuss.c
2399
flow->page_pool = NULL;
drivers/net/ethernet/ti/am65-cpsw-nuss.c
531
if (flow->page_pool) {
drivers/net/ethernet/ti/am65-cpsw-nuss.c
532
page_pool_destroy(flow->page_pool);
drivers/net/ethernet/ti/am65-cpsw-nuss.c
533
flow->page_pool = NULL;
drivers/net/ethernet/ti/am65-cpsw-nuss.c
571
struct page_pool *pool;
drivers/net/ethernet/ti/am65-cpsw-nuss.c
583
flow->page_pool = pool;
drivers/net/ethernet/ti/am65-cpsw-nuss.c
607
page = page_pool_dev_alloc_pages(flow->page_pool);
drivers/net/ethernet/ti/am65-cpsw-nuss.c
770
page_pool_put_full_page(flow->page_pool, page, allow_direct);
drivers/net/ethernet/ti/am65-cpsw-nuss.h
103
struct page_pool *page_pool;
drivers/net/ethernet/ti/cpsw.c
375
struct page_pool *pool;
drivers/net/ethernet/ti/cpsw.c
387
pool = cpsw->page_pool[ch];
drivers/net/ethernet/ti/cpsw_new.c
311
struct page_pool *pool;
drivers/net/ethernet/ti/cpsw_new.c
331
pool = cpsw->page_pool[ch];
drivers/net/ethernet/ti/cpsw_priv.c
1086
struct page_pool *pool;
drivers/net/ethernet/ti/cpsw_priv.c
1093
pool = cpsw->page_pool[ch];
drivers/net/ethernet/ti/cpsw_priv.c
1127
static struct page_pool *cpsw_create_page_pool(struct cpsw_common *cpsw,
drivers/net/ethernet/ti/cpsw_priv.c
1131
struct page_pool *pool;
drivers/net/ethernet/ti/cpsw_priv.c
1149
struct page_pool *pool;
drivers/net/ethernet/ti/cpsw_priv.c
1157
cpsw->page_pool[ch] = pool;
drivers/net/ethernet/ti/cpsw_priv.c
1166
struct page_pool *pool;
drivers/net/ethernet/ti/cpsw_priv.c
1169
pool = cpsw->page_pool[ch];
drivers/net/ethernet/ti/cpsw_priv.c
1207
page_pool_destroy(cpsw->page_pool[ch]);
drivers/net/ethernet/ti/cpsw_priv.c
1208
cpsw->page_pool[ch] = NULL;
drivers/net/ethernet/ti/cpsw_priv.c
1362
page_pool_recycle_direct(cpsw->page_pool[ch], page);
drivers/net/ethernet/ti/cpsw_priv.h
360
struct page_pool *page_pool[CPSW_MAX_QUEUES];
drivers/net/ethernet/ti/icssg/icssg_common.c
1023
struct page_pool *pool;
drivers/net/ethernet/ti/icssg/icssg_common.c
1135
struct page_pool *pool;
drivers/net/ethernet/ti/icssg/icssg_common.c
468
static struct page_pool *prueth_create_page_pool(struct prueth_emac *emac,
drivers/net/ethernet/ti/icssg/icssg_common.c
473
struct page_pool *pool;
drivers/net/ethernet/ti/icssg/icssg_common.c
500
struct page_pool *pool;
drivers/net/ethernet/ti/icssg/icssg_prueth.c
595
struct page_pool *pool = emac->rx_chns.pg_pool;
drivers/net/ethernet/ti/icssg/icssg_prueth.h
143
struct page_pool *pg_pool;
drivers/net/ethernet/wangxun/libwx/wx_lib.c
257
page_pool_put_full_page(rx_ring->page_pool, rx_buffer->page, true);
drivers/net/ethernet/wangxun/libwx/wx_lib.c
2593
page_pool_put_full_page(rx_ring->page_pool, rx_buffer->page, false);
drivers/net/ethernet/wangxun/libwx/wx_lib.c
2645
if (rx_ring->page_pool) {
drivers/net/ethernet/wangxun/libwx/wx_lib.c
2646
page_pool_destroy(rx_ring->page_pool);
drivers/net/ethernet/wangxun/libwx/wx_lib.c
2647
rx_ring->page_pool = NULL;
drivers/net/ethernet/wangxun/libwx/wx_lib.c
2810
rx_ring->page_pool = page_pool_create(&pp_params);
drivers/net/ethernet/wangxun/libwx/wx_lib.c
2811
if (IS_ERR(rx_ring->page_pool)) {
drivers/net/ethernet/wangxun/libwx/wx_lib.c
2812
ret = PTR_ERR(rx_ring->page_pool);
drivers/net/ethernet/wangxun/libwx/wx_lib.c
2813
rx_ring->page_pool = NULL;
drivers/net/ethernet/wangxun/libwx/wx_lib.c
298
page = page_pool_dev_alloc_pages(rx_ring->page_pool);
drivers/net/ethernet/wangxun/libwx/wx_type.h
1067
struct page_pool *page_pool;
drivers/net/netdevsim/netdev.c
465
static int nsim_create_page_pool(struct page_pool **p, struct napi_struct *napi)
drivers/net/netdevsim/netdev.c
476
struct page_pool *pool;
drivers/net/netdevsim/netdev.c
501
err = nsim_create_page_pool(&rq->page_pool, &rq->napi);
drivers/net/netdevsim/netdev.c
510
page_pool_destroy(ns->rq[i]->page_pool);
drivers/net/netdevsim/netdev.c
511
ns->rq[i]->page_pool = NULL;
drivers/net/netdevsim/netdev.c
586
page_pool_destroy(ns->rq[i]->page_pool);
drivers/net/netdevsim/netdev.c
587
ns->rq[i]->page_pool = NULL;
drivers/net/netdevsim/netdev.c
760
struct page_pool *pp;
drivers/net/netdevsim/netdev.c
785
err = nsim_create_page_pool(&qmem->rq->page_pool, &qmem->rq->napi);
drivers/net/netdevsim/netdev.c
809
page_pool_destroy(qmem->rq->page_pool);
drivers/net/netdevsim/netdev.c
824
ns->rq[idx]->page_pool = qmem->pp;
drivers/net/netdevsim/netdev.c
858
qmem->pp = ns->rq[idx]->page_pool;
drivers/net/netdevsim/netdev.c
951
ns->page = page_pool_dev_alloc_pages(ns->rq[0]->page_pool);
drivers/net/netdevsim/netdevsim.h
98
struct page_pool *page_pool;
drivers/net/veth.c
1016
rq->page_pool = page_pool_create(&pp_params);
drivers/net/veth.c
1017
if (IS_ERR(rq->page_pool)) {
drivers/net/veth.c
1018
int err = PTR_ERR(rq->page_pool);
drivers/net/veth.c
1020
rq->page_pool = NULL;
drivers/net/veth.c
1061
page_pool_destroy(priv->rq[i].page_pool);
drivers/net/veth.c
1062
priv->rq[i].page_pool = NULL;
drivers/net/veth.c
1095
page_pool_destroy(priv->rq[i].page_pool);
drivers/net/veth.c
1096
priv->rq[i].page_pool = NULL;
drivers/net/veth.c
189
if (!priv->rq[i].page_pool)
drivers/net/veth.c
191
page_pool_get_stats(priv->rq[i].page_pool, &pp_stats);
drivers/net/veth.c
71
struct page_pool *page_pool;
drivers/net/veth.c
761
if (skb_pp_cow_data(rq->page_pool, pskb, XDP_PACKET_HEADROOM))
drivers/net/vmxnet3/vmxnet3_drv.c
1391
struct page_pool *pp;
drivers/net/vmxnet3/vmxnet3_drv.c
1407
rq->page_pool = pp;
drivers/net/vmxnet3/vmxnet3_drv.c
1420
vmxnet3_pp_get_buff(struct page_pool *pp, dma_addr_t *dma_addr,
drivers/net/vmxnet3/vmxnet3_drv.c
2041
page_pool_recycle_direct(rq->page_pool,
drivers/net/vmxnet3/vmxnet3_drv.c
2069
page_pool_destroy(rq->page_pool);
drivers/net/vmxnet3/vmxnet3_drv.c
2070
rq->page_pool = NULL;
drivers/net/vmxnet3/vmxnet3_drv.c
2200
page_pool_destroy(rq->page_pool);
drivers/net/vmxnet3/vmxnet3_drv.c
2201
rq->page_pool = NULL;
drivers/net/vmxnet3/vmxnet3_drv.c
690
void *data = vmxnet3_pp_get_buff(rq->page_pool,
drivers/net/vmxnet3/vmxnet3_int.h
341
struct page_pool *page_pool;
drivers/net/vmxnet3/vmxnet3_xdp.c
282
page_pool_recycle_direct(rq->page_pool, page);
drivers/net/vmxnet3/vmxnet3_xdp.c
290
page_pool_recycle_direct(rq->page_pool, page);
drivers/net/vmxnet3/vmxnet3_xdp.c
307
page_pool_recycle_direct(rq->page_pool, page);
drivers/net/vmxnet3/vmxnet3_xdp.c
320
page_pool_recycle_direct(rq->page_pool, page);
drivers/net/vmxnet3/vmxnet3_xdp.c
345
page = page_pool_alloc_pages(rq->page_pool, GFP_ATOMIC);
drivers/net/vmxnet3/vmxnet3_xdp.c
352
xdp_prepare_buff(&xdp, page_address(page), rq->page_pool->p.offset,
drivers/net/vmxnet3/vmxnet3_xdp.c
395
rq->page_pool->p.offset, rbi->len,
drivers/net/vmxnet3/vmxnet3_xdp.c
396
page_pool_get_dma_dir(rq->page_pool));
drivers/net/vmxnet3/vmxnet3_xdp.c
399
xdp_prepare_buff(&xdp, page_address(page), rq->page_pool->p.offset,
drivers/net/vmxnet3/vmxnet3_xdp.c
417
new_data = vmxnet3_pp_get_buff(rq->page_pool, &new_dma_addr,
drivers/net/vmxnet3/vmxnet3_xdp.h
39
void *vmxnet3_pp_get_buff(struct page_pool *pp, dma_addr_t *dma_addr,
drivers/net/wireless/mediatek/mt76/dma.c
1178
page_pool_destroy(q->page_pool);
drivers/net/wireless/mediatek/mt76/dma.c
468
page_pool_get_dma_dir(q->page_pool));
drivers/net/wireless/mediatek/mt76/dma.c
527
page_pool_get_dma_dir(q->page_pool));
drivers/net/wireless/mediatek/mt76/dma.c
539
page_pool_get_dma_dir(q->page_pool));
drivers/net/wireless/mediatek/mt76/mac80211.c
2029
page_pool_get_stats(dev->q_rx[i].page_pool, &stats);
drivers/net/wireless/mediatek/mt76/mac80211.c
653
q->page_pool = page_pool_create(&pp_params);
drivers/net/wireless/mediatek/mt76/mac80211.c
654
if (IS_ERR(q->page_pool)) {
drivers/net/wireless/mediatek/mt76/mac80211.c
655
int err = PTR_ERR(q->page_pool);
drivers/net/wireless/mediatek/mt76/mac80211.c
657
q->page_pool = NULL;
drivers/net/wireless/mediatek/mt76/mt76.h
2017
page = page_pool_alloc_frag(q->page_pool, offset, size,
drivers/net/wireless/mediatek/mt76/mt76.h
265
struct page_pool *page_pool;
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1950
page_pool_get_dma_dir(p->q->page_pool));
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1969
page_pool_get_dma_dir(q->page_pool));
drivers/net/wireless/mediatek/mt76/npu.c
54
page_pool_get_dma_dir(q->page_pool));
drivers/net/wireless/mediatek/mt76/npu.c
85
page_pool_get_dma_dir(q->page_pool));
drivers/net/wireless/mediatek/mt76/usb.c
707
page_pool_destroy(q->page_pool);
drivers/net/wireless/mediatek/mt76/usb.c
708
q->page_pool = NULL;
drivers/net/xen-netfront.c
156
struct page_pool *page_pool;
drivers/net/xen-netfront.c
1853
page_pool_destroy(queue->page_pool);
drivers/net/xen-netfront.c
2179
queue->page_pool = page_pool_create(&pp_params);
drivers/net/xen-netfront.c
2180
if (IS_ERR(queue->page_pool)) {
drivers/net/xen-netfront.c
2181
err = PTR_ERR(queue->page_pool);
drivers/net/xen-netfront.c
2182
queue->page_pool = NULL;
drivers/net/xen-netfront.c
2194
MEM_TYPE_PAGE_POOL, queue->page_pool);
drivers/net/xen-netfront.c
2204
page_pool_destroy(queue->page_pool);
drivers/net/xen-netfront.c
2205
queue->page_pool = NULL;
drivers/net/xen-netfront.c
282
page = page_pool_alloc_pages(queue->page_pool,
include/linux/mm_types.h
123
struct page_pool *pp;
include/linux/netdevice.h
3575
struct page_pool *pool;
include/linux/skbuff.h
3710
int skb_pp_cow_data(struct page_pool *pool, struct sk_buff **pskb,
include/linux/skbuff.h
3712
int skb_cow_data_for_xdp(struct page_pool *pool, struct sk_buff **pskb,
include/net/libeth/rx.h
77
struct page_pool *pp;
include/net/mana/mana.h
356
struct page_pool *page_pool;
include/net/netmem.h
107
struct page_pool *pp;
include/net/netmem.h
312
static inline struct page_pool *__netmem_get_pp(netmem_ref netmem)
include/net/netmem.h
317
static inline struct page_pool *netmem_get_pp(netmem_ref netmem)
include/net/netmem.h
35
struct page_pool *pp;
include/net/page_pool/helpers.h
109
static inline struct page *page_pool_dev_alloc_frag(struct page_pool *pool,
include/net/page_pool/helpers.h
118
static inline netmem_ref page_pool_alloc_netmem(struct page_pool *pool,
include/net/page_pool/helpers.h
147
static inline netmem_ref page_pool_dev_alloc_netmem(struct page_pool *pool,
include/net/page_pool/helpers.h
156
static inline netmem_ref page_pool_dev_alloc_netmems(struct page_pool *pool)
include/net/page_pool/helpers.h
163
static inline struct page *page_pool_alloc(struct page_pool *pool,
include/net/page_pool/helpers.h
182
static inline struct page *page_pool_dev_alloc(struct page_pool *pool,
include/net/page_pool/helpers.h
191
static inline void *page_pool_alloc_va(struct page_pool *pool,
include/net/page_pool/helpers.h
216
static inline void *page_pool_dev_alloc_va(struct page_pool *pool,
include/net/page_pool/helpers.h
232
page_pool_get_dma_dir(const struct page_pool *pool)
include/net/page_pool/helpers.h
331
static inline void page_pool_put_netmem(struct page_pool *pool,
include/net/page_pool/helpers.h
360
static inline void page_pool_put_page(struct page_pool *pool,
include/net/page_pool/helpers.h
369
static inline void page_pool_put_full_netmem(struct page_pool *pool,
include/net/page_pool/helpers.h
385
static inline void page_pool_put_full_page(struct page_pool *pool,
include/net/page_pool/helpers.h
399
static inline void page_pool_recycle_direct(struct page_pool *pool,
include/net/page_pool/helpers.h
405
static inline void page_pool_recycle_direct_netmem(struct page_pool *pool,
include/net/page_pool/helpers.h
422
static inline void page_pool_free_va(struct page_pool *pool, void *va,
include/net/page_pool/helpers.h
450
static inline void __page_pool_dma_sync_for_cpu(const struct page_pool *pool,
include/net/page_pool/helpers.h
471
static inline void page_pool_dma_sync_for_cpu(const struct page_pool *pool,
include/net/page_pool/helpers.h
480
page_pool_dma_sync_netmem_for_cpu(const struct page_pool *pool,
include/net/page_pool/helpers.h
492
static inline void page_pool_get(struct page_pool *pool)
include/net/page_pool/helpers.h
497
static inline bool page_pool_put(struct page_pool *pool)
include/net/page_pool/helpers.h
502
static inline void page_pool_nid_changed(struct page_pool *pool, int new_nid)
include/net/page_pool/helpers.h
520
static inline bool page_pool_is_unreadable(struct page_pool *pool)
include/net/page_pool/helpers.h
67
bool page_pool_get_stats(const struct page_pool *pool,
include/net/page_pool/helpers.h
92
static inline struct page *page_pool_dev_alloc_pages(struct page_pool *pool)
include/net/page_pool/memory_provider.h
13
netmem_ref (*alloc_netmems)(struct page_pool *pool, gfp_t gfp);
include/net/page_pool/memory_provider.h
14
bool (*release_netmem)(struct page_pool *pool, netmem_ref netmem);
include/net/page_pool/memory_provider.h
15
int (*init)(struct page_pool *pool);
include/net/page_pool/memory_provider.h
16
void (*destroy)(struct page_pool *pool);
include/net/page_pool/memory_provider.h
23
void net_mp_niov_set_page_pool(struct page_pool *pool, struct net_iov *niov);
include/net/page_pool/memory_provider.h
45
static inline void net_mp_netmem_place_in_cache(struct page_pool *pool,
include/net/page_pool/types.h
255
struct page *page_pool_alloc_pages(struct page_pool *pool, gfp_t gfp);
include/net/page_pool/types.h
256
netmem_ref page_pool_alloc_netmems(struct page_pool *pool, gfp_t gfp);
include/net/page_pool/types.h
257
struct page *page_pool_alloc_frag(struct page_pool *pool, unsigned int *offset,
include/net/page_pool/types.h
259
netmem_ref page_pool_alloc_frag_netmem(struct page_pool *pool,
include/net/page_pool/types.h
262
struct page_pool *page_pool_create(const struct page_pool_params *params);
include/net/page_pool/types.h
263
struct page_pool *page_pool_create_percpu(const struct page_pool_params *params,
include/net/page_pool/types.h
269
void page_pool_enable_direct_recycling(struct page_pool *pool,
include/net/page_pool/types.h
271
void page_pool_disable_direct_recycling(struct page_pool *pool);
include/net/page_pool/types.h
272
void page_pool_destroy(struct page_pool *pool);
include/net/page_pool/types.h
273
void page_pool_use_xdp_mem(struct page_pool *pool, void (*disconnect)(void *),
include/net/page_pool/types.h
277
static inline void page_pool_destroy(struct page_pool *pool)
include/net/page_pool/types.h
281
static inline void page_pool_use_xdp_mem(struct page_pool *pool,
include/net/page_pool/types.h
292
void page_pool_put_unrefed_netmem(struct page_pool *pool, netmem_ref netmem,
include/net/page_pool/types.h
295
void page_pool_put_unrefed_page(struct page_pool *pool, struct page *page,
include/net/page_pool/types.h
309
void page_pool_update_nid(struct page_pool *pool, int new_nid);
include/net/xdp.h
500
int xdp_reg_page_pool(struct page_pool *pool);
include/net/xdp.h
501
void xdp_unreg_page_pool(const struct page_pool *pool);
include/net/xdp.h
503
const struct page_pool *pool);
include/net/xdp.h
60
struct page_pool;
include/net/xdp_priv.h
13
struct page_pool *page_pool;
include/trace/events/page_pool.h
101
__field(const struct page_pool *, pool)
include/trace/events/page_pool.h
16
TP_PROTO(const struct page_pool *pool,
include/trace/events/page_pool.h
22
__field(const struct page_pool *, pool)
include/trace/events/page_pool.h
3
#define TRACE_SYSTEM page_pool
include/trace/events/page_pool.h
44
TP_PROTO(const struct page_pool *pool,
include/trace/events/page_pool.h
50
__field(const struct page_pool *, pool)
include/trace/events/page_pool.h
70
TP_PROTO(const struct page_pool *pool,
include/trace/events/page_pool.h
76
__field(const struct page_pool *, pool)
include/trace/events/page_pool.h
96
TP_PROTO(const struct page_pool *pool, int new_nid),
io_uring/zcrx.c
1004
static netmem_ref io_pp_zc_alloc_netmems(struct page_pool *pp, gfp_t gfp)
io_uring/zcrx.c
1023
static bool io_pp_zc_release_netmem(struct page_pool *pp, netmem_ref netmem)
io_uring/zcrx.c
1036
static int io_pp_zc_init(struct page_pool *pp)
io_uring/zcrx.c
1060
static void io_pp_zc_destroy(struct page_pool *pp)
io_uring/zcrx.c
1334
struct page_pool *pp;
io_uring/zcrx.c
311
static void io_zcrx_sync_for_device(struct page_pool *pool,
io_uring/zcrx.c
35
static inline struct io_zcrx_ifq *io_pp_to_ifq(struct page_pool *pp)
io_uring/zcrx.c
949
static void io_zcrx_ring_refill(struct page_pool *pp,
io_uring/zcrx.c
988
static void io_zcrx_refill_slow(struct page_pool *pp, struct io_zcrx_ifq *ifq)
net/bpf/test_run.c
104
struct page_pool *pp;
net/bpf/test_run.c
152
struct page_pool *pp;
net/core/dev.c
13167
struct page_pool *pp_ptr;
net/core/dev.c
13312
struct page_pool *pp_ptr;
net/core/devmem.c
432
int mp_dmabuf_devmem_init(struct page_pool *pool)
net/core/devmem.c
452
netmem_ref mp_dmabuf_devmem_alloc_netmems(struct page_pool *pool, gfp_t gfp)
net/core/devmem.c
471
void mp_dmabuf_devmem_destroy(struct page_pool *pool)
net/core/devmem.c
478
bool mp_dmabuf_devmem_release_page(struct page_pool *pool, netmem_ref netmem)
net/core/mp_dmabuf_devmem.h
14
int mp_dmabuf_devmem_init(struct page_pool *pool);
net/core/mp_dmabuf_devmem.h
16
netmem_ref mp_dmabuf_devmem_alloc_netmems(struct page_pool *pool, gfp_t gfp);
net/core/mp_dmabuf_devmem.h
18
void mp_dmabuf_devmem_destroy(struct page_pool *pool);
net/core/mp_dmabuf_devmem.h
20
bool mp_dmabuf_devmem_release_page(struct page_pool *pool, netmem_ref netmem);
net/core/mp_dmabuf_devmem.h
22
static inline int mp_dmabuf_devmem_init(struct page_pool *pool)
net/core/mp_dmabuf_devmem.h
28
mp_dmabuf_devmem_alloc_netmems(struct page_pool *pool, gfp_t gfp)
net/core/mp_dmabuf_devmem.h
33
static inline void mp_dmabuf_devmem_destroy(struct page_pool *pool)
net/core/mp_dmabuf_devmem.h
38
mp_dmabuf_devmem_release_page(struct page_pool *pool, netmem_ref netmem)
net/core/netmem_priv.h
28
static inline void netmem_set_pp(netmem_ref netmem, struct page_pool *pool)
net/core/page_pool.c
1022
static netmem_ref page_pool_drain_frag(struct page_pool *pool,
net/core/page_pool.c
1040
static void page_pool_free_frag(struct page_pool *pool)
net/core/page_pool.c
1053
netmem_ref page_pool_alloc_frag_netmem(struct page_pool *pool,
net/core/page_pool.c
1098
struct page *page_pool_alloc_frag(struct page_pool *pool, unsigned int *offset,
net/core/page_pool.c
1106
static void page_pool_empty_ring(struct page_pool *pool)
net/core/page_pool.c
1121
static void __page_pool_destroy(struct page_pool *pool)
net/core/page_pool.c
1137
static void page_pool_empty_alloc_cache_once(struct page_pool *pool)
net/core/page_pool.c
1154
static void page_pool_scrub(struct page_pool *pool)
net/core/page_pool.c
1186
static int page_pool_release(struct page_pool *pool)
net/core/page_pool.c
1205
struct page_pool *pool = container_of(dwq, typeof(*pool), release_dw);
net/core/page_pool.c
1234
void page_pool_use_xdp_mem(struct page_pool *pool, void (*disconnect)(void *),
net/core/page_pool.c
1258
void page_pool_enable_direct_recycling(struct page_pool *pool,
net/core/page_pool.c
1271
void page_pool_disable_direct_recycling(struct page_pool *pool)
net/core/page_pool.c
1289
void page_pool_destroy(struct page_pool *pool)
net/core/page_pool.c
1313
void page_pool_update_nid(struct page_pool *pool, int new_nid)
net/core/page_pool.c
1336
void net_mp_niov_set_page_pool(struct page_pool *pool, struct net_iov *niov)
net/core/page_pool.c
161
static bool page_pool_producer_lock(struct page_pool *pool)
net/core/page_pool.c
174
static void page_pool_producer_unlock(struct page_pool *pool,
net/core/page_pool.c
186
CACHELINE_ASSERT_GROUP_MEMBER(struct page_pool, frag, frag_users);
net/core/page_pool.c
187
CACHELINE_ASSERT_GROUP_MEMBER(struct page_pool, frag, frag_page);
net/core/page_pool.c
188
CACHELINE_ASSERT_GROUP_MEMBER(struct page_pool, frag, frag_offset);
net/core/page_pool.c
189
CACHELINE_ASSERT_GROUP_SIZE(struct page_pool, frag,
net/core/page_pool.c
193
static int page_pool_init(struct page_pool *pool,
net/core/page_pool.c
321
static void page_pool_uninit(struct page_pool *pool)
net/core/page_pool.c
337
struct page_pool *
net/core/page_pool.c
340
struct page_pool *pool;
net/core/page_pool.c
370
struct page_pool *page_pool_create(const struct page_pool_params *params)
net/core/page_pool.c
376
static void page_pool_return_netmem(struct page_pool *pool, netmem_ref netmem);
net/core/page_pool.c
378
static noinline netmem_ref page_pool_refill_alloc_cache(struct page_pool *pool)
net/core/page_pool.c
431
static netmem_ref __page_pool_get_cached(struct page_pool *pool)
net/core/page_pool.c
447
static void __page_pool_dma_sync_for_device(const struct page_pool *pool,
net/core/page_pool.c
461
page_pool_dma_sync_for_device(const struct page_pool *pool,
net/core/page_pool.c
475
static int page_pool_register_dma_index(struct page_pool *pool,
net/core/page_pool.c
500
static int page_pool_release_dma_index(struct page_pool *pool,
net/core/page_pool.c
525
static bool page_pool_dma_map(struct page_pool *pool, netmem_ref netmem, gfp_t gfp)
net/core/page_pool.c
564
static struct page *__page_pool_alloc_page_order(struct page_pool *pool,
net/core/page_pool.c
590
static noinline netmem_ref __page_pool_alloc_netmems_slow(struct page_pool *pool,
net/core/page_pool.c
654
netmem_ref page_pool_alloc_netmems(struct page_pool *pool, gfp_t gfp)
net/core/page_pool.c
673
struct page *page_pool_alloc_pages(struct page_pool *pool, gfp_t gfp)
net/core/page_pool.c
684
s32 page_pool_inflight(const struct page_pool *pool, bool strict)
net/core/page_pool.c
703
void page_pool_set_pp_info(struct page_pool *pool, netmem_ref netmem)
net/core/page_pool.c
725
static __always_inline void __page_pool_release_netmem_dma(struct page_pool *pool,
net/core/page_pool.c
753
static void page_pool_return_netmem(struct page_pool *pool, netmem_ref netmem)
net/core/page_pool.c
780
static bool page_pool_recycle_in_ring(struct page_pool *pool, netmem_ref netmem)
net/core/page_pool.c
800
struct page_pool *pool)
net/core/page_pool.c
827
__page_pool_put_page(struct page_pool *pool, netmem_ref netmem,
net/core/page_pool.c
85
bool page_pool_get_stats(const struct page_pool *pool,
net/core/page_pool.c
872
static bool page_pool_napi_local(const struct page_pool *pool)
net/core/page_pool.c
899
void page_pool_put_unrefed_netmem(struct page_pool *pool, netmem_ref netmem,
net/core/page_pool.c
915
void page_pool_put_unrefed_page(struct page_pool *pool, struct page *page,
net/core/page_pool.c
923
static void page_pool_recycle_ring_bulk(struct page_pool *pool,
net/core/page_pool.c
984
struct page_pool *pool = NULL;
net/core/page_pool.c
991
struct page_pool *netmem_pp;
net/core/page_pool_priv.h
12
s32 page_pool_inflight(const struct page_pool *pool, bool strict);
net/core/page_pool_priv.h
14
int page_pool_list(struct page_pool *pool);
net/core/page_pool_priv.h
15
void page_pool_detached(struct page_pool *pool);
net/core/page_pool_priv.h
16
void page_pool_unlist(struct page_pool *pool);
net/core/page_pool_priv.h
41
void page_pool_set_pp_info(struct page_pool *pool, netmem_ref netmem);
net/core/page_pool_priv.h
46
static inline void page_pool_set_pp_info(struct page_pool *pool,
net/core/page_pool_user.c
114
page_pool_nl_stats_fill(struct sk_buff *rsp, const struct page_pool *pool,
net/core/page_pool_user.c
216
page_pool_nl_fill(struct sk_buff *rsp, const struct page_pool *pool,
net/core/page_pool_user.c
262
static void netdev_nl_page_pool_event(const struct page_pool *pool, u32 cmd)
net/core/page_pool_user.c
311
int page_pool_list(struct page_pool *pool)
net/core/page_pool_user.c
337
void page_pool_detached(struct page_pool *pool)
net/core/page_pool_user.c
345
void page_pool_unlist(struct page_pool *pool)
net/core/page_pool_user.c
359
struct page_pool *pool;
net/core/page_pool_user.c
36
typedef int (*pp_nl_fill_cb)(struct sk_buff *rsp, const struct page_pool *pool,
net/core/page_pool_user.c
381
struct page_pool *pool;
net/core/page_pool_user.c
394
struct page_pool *pool, *last;
net/core/page_pool_user.c
42
struct page_pool *pool;
net/core/page_pool_user.c
88
struct page_pool *pool;
net/core/skbuff.c
1021
int skb_cow_data_for_xdp(struct page_pool *pool, struct sk_buff **pskb,
net/core/skbuff.c
939
int skb_pp_cow_data(struct page_pool *pool, struct sk_buff **pskb,
net/core/xdp.c
132
page_pool_destroy(xa->page_pool);
net/core/xdp.c
381
int xdp_reg_page_pool(struct page_pool *pool)
net/core/xdp.c
396
void xdp_unreg_page_pool(const struct page_pool *pool)
net/core/xdp.c
416
const struct page_pool *pool)
net/core/xdp.c
689
struct page_pool *pp)
net/core/xdp.c
745
struct page_pool *pp;
tools/testing/selftests/net/bench/page_pool/bench_page_pool_simple.c
106
static void pp_fill_ptr_ring(struct page_pool *pp, int elems)
tools/testing/selftests/net/bench/page_pool/bench_page_pool_simple.c
133
struct page_pool *pp;