Symbol: page_idx
arch/x86/include/asm/sev.h
586
int page_idx;
arch/x86/include/asm/sev.h
594
for (page_idx = 0; page_idx < npages; page_idx++) {
arch/x86/include/asm/sev.h
595
val = bytes[page_idx * PAGE_SIZE];
arch/x86/include/asm/sev.h
596
val = bytes[page_idx * PAGE_SIZE + PAGE_SIZE - 1];
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
854
uint64_t page_idx, pages_per_xcc;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
860
for (i = 0, page_idx = 0; i < num_xcc; i++, page_idx += pages_per_xcc) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
862
gtt->offset + (page_idx << PAGE_SHIFT),
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
863
pages_per_xcc, &gtt->ttm.dma_address[page_idx],
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1663
int ret, i, page_idx = 0;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1695
for_each_sgtable_dma_page(sgt, &sg_dma_iter, page_idx) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1697
page_idx++;
drivers/infiniband/hw/mlx5/dm.c
177
u16 page_idx;
drivers/infiniband/hw/mlx5/dm.c
180
page_idx = op_entry->mentry.rdma_entry.start_pgoff & 0xFFFF;
drivers/infiniband/hw/mlx5/dm.c
183
&page_idx, sizeof(page_idx));
drivers/infiniband/hw/mlx5/dm.c
24
u64 page_idx = 0;
drivers/infiniband/hw/mlx5/dm.c
281
u16 page_idx;
drivers/infiniband/hw/mlx5/dm.c
317
page_idx = dm->mentry.rdma_entry.start_pgoff & 0xFFFF;
drivers/infiniband/hw/mlx5/dm.c
319
&page_idx, sizeof(page_idx));
drivers/infiniband/hw/mlx5/dm.c
44
while (page_idx < num_memic_hw_pages) {
drivers/infiniband/hw/mlx5/dm.c
46
page_idx = bitmap_find_next_zero_area(dm->memic_alloc_pages,
drivers/infiniband/hw/mlx5/dm.c
48
page_idx,
drivers/infiniband/hw/mlx5/dm.c
51
if (page_idx < num_memic_hw_pages)
drivers/infiniband/hw/mlx5/dm.c
513
u16 page_idx;
drivers/infiniband/hw/mlx5/dm.c
520
page_idx = memic->mentry.rdma_entry.start_pgoff & 0xFFFF;
drivers/infiniband/hw/mlx5/dm.c
522
&page_idx, sizeof(page_idx));
drivers/infiniband/hw/mlx5/dm.c
53
page_idx, num_pages);
drivers/infiniband/hw/mlx5/dm.c
57
if (page_idx >= num_memic_hw_pages)
drivers/infiniband/hw/mlx5/dm.c
61
hw_start_addr + (page_idx * PAGE_SIZE));
drivers/infiniband/hw/mlx5/dm.c
67
page_idx, num_pages);
drivers/infiniband/hw/mlx5/dm.c
71
page_idx++;
drivers/infiniband/hw/mlx5/main.c
2535
clear_bit(mentry->page_idx, var_table->bitmap);
drivers/infiniband/hw/mlx5/main.c
2541
mlx5_cmd_uar_dealloc(dev->mdev, mentry->page_idx,
drivers/infiniband/hw/mlx5/main.c
4147
u32 page_idx;
drivers/infiniband/hw/mlx5/main.c
4156
page_idx = find_first_zero_bit(var_table->bitmap,
drivers/infiniband/hw/mlx5/main.c
4158
if (page_idx >= var_table->num_var_hw_entries) {
drivers/infiniband/hw/mlx5/main.c
4164
set_bit(page_idx, var_table->bitmap);
drivers/infiniband/hw/mlx5/main.c
4168
(page_idx * var_table->stride_size);
drivers/infiniband/hw/mlx5/main.c
4169
entry->page_idx = page_idx;
drivers/infiniband/hw/mlx5/main.c
4181
clear_bit(page_idx, var_table->bitmap);
drivers/infiniband/hw/mlx5/main.c
4218
&entry->page_idx, sizeof(entry->page_idx));
drivers/infiniband/hw/mlx5/main.c
4281
entry->page_idx = uar_index;
drivers/infiniband/hw/mlx5/main.c
4345
&entry->page_idx, sizeof(entry->page_idx));
drivers/infiniband/hw/mlx5/mlx5_ib.h
630
u32 page_idx;
drivers/infiniband/hw/mthca/mthca_memfree.h
106
iter->page_idx = 0;
drivers/infiniband/hw/mthca/mthca_memfree.h
116
if (++iter->page_idx >= iter->chunk->nsg) {
drivers/infiniband/hw/mthca/mthca_memfree.h
124
iter->page_idx = 0;
drivers/infiniband/hw/mthca/mthca_memfree.h
130
return sg_dma_address(&iter->chunk->mem[iter->page_idx]);
drivers/infiniband/hw/mthca/mthca_memfree.h
135
return sg_dma_len(&iter->chunk->mem[iter->page_idx]);
drivers/infiniband/hw/mthca/mthca_memfree.h
77
int page_idx;
drivers/infiniband/sw/siw/siw_mem.h
60
unsigned int page_idx = (addr - umem->fp_addr) >> PAGE_SHIFT,
drivers/infiniband/sw/siw/siw_mem.h
61
chunk_idx = page_idx >> CHUNK_SHIFT,
drivers/infiniband/sw/siw/siw_mem.h
62
page_in_chunk = page_idx & ~CHUNK_MASK;
drivers/infiniband/sw/siw/siw_mem.h
64
if (likely(page_idx < umem->num_pages))
drivers/input/touchscreen/raydium_i2c_ts.c
647
u16 page_idx, const void *data, size_t len)
drivers/input/touchscreen/raydium_i2c_ts.c
658
buf[BL_PAGE_STR] = page_idx ? 0xff : 0;
drivers/input/touchscreen/raydium_i2c_ts.c
672
page_idx, i, error);
drivers/iommu/iommufd/iova_bitmap.c
425
unsigned int page_idx = cur_bit / BITS_PER_PAGE;
drivers/iommu/iommufd/iova_bitmap.c
431
if (unlikely(page_idx > last_page_idx)) {
drivers/iommu/iommufd/iova_bitmap.c
440
kaddr = kmap_local_page(mapped->pages[page_idx]);
drivers/md/md-llbitmap.c
687
static void llbitmap_raise_barrier(struct llbitmap *llbitmap, int page_idx)
drivers/md/md-llbitmap.c
689
struct llbitmap_page_ctl *pctl = llbitmap->pctl[page_idx];
drivers/md/md-llbitmap.c
701
static void llbitmap_release_barrier(struct llbitmap *llbitmap, int page_idx)
drivers/md/md-llbitmap.c
703
struct llbitmap_page_ctl *pctl = llbitmap->pctl[page_idx];
drivers/md/md-llbitmap.c
708
static int llbitmap_suspend_timeout(struct llbitmap *llbitmap, int page_idx)
drivers/md/md-llbitmap.c
710
struct llbitmap_page_ctl *pctl = llbitmap->pctl[page_idx];
drivers/md/md-llbitmap.c
723
static void llbitmap_resume(struct llbitmap *llbitmap, int page_idx)
drivers/md/md-llbitmap.c
725
struct llbitmap_page_ctl *pctl = llbitmap->pctl[page_idx];
drivers/md/raid1.c
2779
int page_idx = 0;
drivers/md/raid1.c
3005
page = resync_fetch_page(rp, page_idx);
drivers/md/raid1.c
3017
} while (++page_idx < RESYNC_PAGES);
drivers/md/raid10.c
3159
int page_idx = 0;
drivers/md/raid10.c
3636
page = resync_fetch_page(rp, page_idx);
drivers/md/raid10.c
3646
} while (++page_idx < RESYNC_PAGES);
drivers/mmc/host/usdhi6rol0.c
1011
host->page_idx = 0;
drivers/mmc/host/usdhi6rol0.c
1722
data->flags & MMC_DATA_READ ? 'R' : 'W', host->page_idx,
drivers/mmc/host/usdhi6rol0.c
181
int page_idx; /* page index within an SG segment */
drivers/mmc/host/usdhi6rol0.c
419
(host->page_idx << PAGE_SHIFT) + data->blksz - blk_head)
drivers/mmc/host/usdhi6rol0.c
443
host->page_idx++;
drivers/mmc/host/usdhi6rol0.c
453
host->page_idx++;
drivers/mmc/host/usdhi6rol0.c
463
done = (host->page_idx << PAGE_SHIFT) + host->offset;
drivers/mmc/host/usdhi6rol0.c
489
host->page_idx = 0;
drivers/mmc/host/usdhi6rol0.c
506
host->pg.page = sg_page(host->sg) + host->page_idx;
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
137
static int wqs_allocate_page(struct hinic_wqs *wqs, int page_idx)
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
139
return queue_alloc_page(wqs->hwif, &wqs->page_vaddr[page_idx],
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
140
&wqs->page_paddr[page_idx],
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
141
&wqs->shadow_page_vaddr[page_idx],
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
150
static void wqs_free_page(struct hinic_wqs *wqs, int page_idx)
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
156
wqs->page_vaddr[page_idx],
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
157
(dma_addr_t)wqs->page_paddr[page_idx]);
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
158
vfree(wqs->shadow_page_vaddr[page_idx]);
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
231
static int wqs_next_block(struct hinic_wqs *wqs, int *page_idx,
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
249
*page_idx = wqs->free_blocks[pos].page_idx;
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
252
wqs->free_blocks[pos].page_idx = -1;
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
259
static void wqs_return_block(struct hinic_wqs *wqs, int page_idx,
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
269
wqs->free_blocks[pos].page_idx = page_idx;
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
279
int page_idx, blk_idx, pos = 0;
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
281
for (page_idx = 0; page_idx < wqs->num_pages; page_idx++) {
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
283
wqs->free_blocks[pos].page_idx = page_idx;
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
308
int err, i, page_idx;
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
325
for (page_idx = 0; page_idx < wqs->num_pages; page_idx++) {
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
326
err = wqs_allocate_page(wqs, page_idx);
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
345
for (i = 0; i < page_idx; i++)
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
360
int page_idx;
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
364
for (page_idx = 0; page_idx < wqs->num_pages; page_idx++)
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
365
wqs_free_page(wqs, page_idx);
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
45
((void *)((wqs)->page_vaddr[(wq)->page_idx]) \
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
49
((wqs)->page_paddr[(wq)->page_idx] \
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
53
((void *)((wqs)->shadow_page_vaddr[(wq)->page_idx]) \
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
539
err = wqs_next_block(wqs, &wq->page_idx, &wq->block_idx);
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
570
wqs_return_block(wqs, wq->page_idx, wq->block_idx);
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
583
wqs_return_block(wqs, wq->page_idx, wq->block_idx);
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
645
wq[i].page_idx = 0;
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.h
18
int page_idx;
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.h
25
int page_idx;
drivers/net/ethernet/huawei/hinic3/hinic3_queue_common.h
40
u32 page_idx, elem_idx, elem_per_pg, ofs;
drivers/net/ethernet/huawei/hinic3/hinic3_queue_common.h
44
page_idx = (idx >> shift) & (qpages->num_pages - 1);
drivers/net/ethernet/huawei/hinic3/hinic3_queue_common.h
50
page = qpages->pages + page_idx;
drivers/net/ethernet/mediatek/mtk_wed.c
644
int i, page_idx = 0, n_pages, ring_size;
drivers/net/ethernet/mediatek/mtk_wed.c
691
page_list[page_idx].p = page;
drivers/net/ethernet/mediatek/mtk_wed.c
692
page_list[page_idx++].phy_addr = page_phys;
drivers/net/ethernet/mediatek/mtk_wed.c
742
int i, page_idx = 0;
drivers/net/ethernet/mediatek/mtk_wed.c
751
dma_addr_t page_phy = page_list[page_idx].phy_addr;
drivers/net/ethernet/mediatek/mtk_wed.c
752
void *page = page_list[page_idx++].p;
drivers/net/ethernet/mediatek/mtk_wed.c
778
int i, page_idx = 0;
drivers/net/ethernet/mediatek/mtk_wed.c
814
page_list[page_idx].p = page;
drivers/net/ethernet/mediatek/mtk_wed.c
815
page_list[page_idx++].phy_addr = page_phys;
drivers/net/ethernet/mediatek/mtk_wed.c
859
int i, page_idx = 0;
drivers/net/ethernet/mediatek/mtk_wed.c
871
dma_addr_t buf_addr = page_list[page_idx].phy_addr;
drivers/net/ethernet/mediatek/mtk_wed.c
872
void *page = page_list[page_idx++].p;
drivers/net/ethernet/mellanox/mlx4/icm.h
103
iter->page_idx = 0;
drivers/net/ethernet/mellanox/mlx4/icm.h
113
if (++iter->page_idx >= iter->chunk->nsg) {
drivers/net/ethernet/mellanox/mlx4/icm.h
121
iter->page_idx = 0;
drivers/net/ethernet/mellanox/mlx4/icm.h
128
return iter->chunk->buf[iter->page_idx].dma_addr;
drivers/net/ethernet/mellanox/mlx4/icm.h
130
return sg_dma_address(&iter->chunk->sg[iter->page_idx]);
drivers/net/ethernet/mellanox/mlx4/icm.h
136
return iter->chunk->buf[iter->page_idx].size;
drivers/net/ethernet/mellanox/mlx4/icm.h
138
return sg_dma_len(&iter->chunk->sg[iter->page_idx]);
drivers/net/ethernet/mellanox/mlx4/icm.h
75
int page_idx;
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/rx.c
255
u32 page_idx)
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/rx.c
257
struct mlx5e_xdp_buff *mxbuf = xsk_buff_to_mxbuf(wi->alloc_units.xsk_buffs[page_idx]);
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/rx.c
297
__set_bit(page_idx, wi->skip_release_bitmap); /* non-atomic */
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/rx.h
19
u32 page_idx);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1792
u32 page_idx = wqe_offset >> rq->mpwqe.page_shift;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1818
rq, wi, cqe, cqe_bcnt, head_offset, page_idx);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1866
u32 page_idx)
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1868
struct mlx5e_frag_page *frag_page = &wi->alloc_units.frag_pages[page_idx];
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2046
u32 page_idx)
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2048
struct mlx5e_frag_page *frag_page = &wi->alloc_units.frag_pages[page_idx];
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2184
u32 page_idx = wqe_offset >> PAGE_SHIFT;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2219
frag_page = &wi->alloc_units.frag_pages[page_idx];
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2228
page_idx);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2256
frag_page = &wi->alloc_units.frag_pages[page_idx];
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2293
u32 page_idx = wqe_offset >> rq->mpwqe.page_shift;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2321
page_idx);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
67
u32 page_idx);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
71
u32 page_idx);
drivers/scsi/lpfc/lpfc_sli.c
16498
int cnt, idx, numcq, page_idx = 0;
drivers/scsi/lpfc/lpfc_sli.c
16675
cnt = page_idx + dmabuf->buffer_tag;
drivers/scsi/lpfc/lpfc_sli.c
16682
page_idx += rc;
drivers/scsi/lpfc/lpfc_sli.c
17513
int cnt, idx, numrq, page_idx = 0;
drivers/scsi/lpfc/lpfc_sli.c
17599
cnt = page_idx + dmabuf->buffer_tag;
drivers/scsi/lpfc/lpfc_sli.c
17606
page_idx += rc;
drivers/scsi/lpfc/lpfc_sli.c
17611
cnt = page_idx + dmabuf->buffer_tag;
drivers/scsi/lpfc/lpfc_sli.c
17618
page_idx += rc;
drivers/tty/vt/ucs.c
222
u8 page_idx = cp >> 8, offset = cp;
drivers/tty/vt/ucs.c
235
page = __inline_bsearch(&page_idx, ucs_fallback_pages,
fs/f2fs/file.c
3853
pgoff_t page_idx = 0, last_idx;
fs/f2fs/file.c
3903
while (page_idx < last_idx) {
fs/f2fs/file.c
3910
ret = f2fs_get_dnode_of_data(&dn, page_idx, LOOKUP_NODE);
fs/f2fs/file.c
3914
page_idx = f2fs_get_next_page_offset(&dn,
fs/f2fs/file.c
3915
page_idx);
fs/f2fs/file.c
3923
count = min(end_offset - dn.ofs_in_node, last_idx - page_idx);
fs/f2fs/file.c
3935
page_idx += count;
fs/f2fs/file.c
4049
pgoff_t page_idx = 0, last_idx;
fs/f2fs/file.c
4081
while (page_idx < last_idx) {
fs/f2fs/file.c
4089
ret = f2fs_get_dnode_of_data(&dn, page_idx, LOOKUP_NODE);
fs/f2fs/file.c
4093
page_idx = f2fs_get_next_page_offset(&dn,
fs/f2fs/file.c
4094
page_idx);
fs/f2fs/file.c
4102
count = min(end_offset - dn.ofs_in_node, last_idx - page_idx);
fs/f2fs/file.c
4114
page_idx += count;
fs/f2fs/file.c
4433
static int redirty_blocks(struct inode *inode, pgoff_t page_idx, int len)
fs/f2fs/file.c
4435
DEFINE_READAHEAD(ractl, NULL, NULL, inode->i_mapping, page_idx);
fs/f2fs/file.c
4438
pgoff_t redirty_idx = page_idx;
fs/f2fs/file.c
4446
folio = read_cache_folio(mapping, page_idx, NULL, NULL);
fs/f2fs/file.c
4451
page_len += folio_nr_pages(folio) - (page_idx - folio->index);
fs/f2fs/file.c
4452
page_idx = folio_next_index(folio);
fs/f2fs/file.c
4468
} while (redirty_idx < page_idx);
fs/f2fs/file.c
4478
pgoff_t page_idx = 0, last_idx, cluster_idx;
fs/f2fs/file.c
4517
page_idx = cluster_idx << fi->i_log_cluster_size;
fs/f2fs/file.c
4519
if (!f2fs_is_compressed_cluster(inode, page_idx))
fs/f2fs/file.c
4522
ret = redirty_blocks(inode, page_idx, fi->i_cluster_size);
fs/f2fs/file.c
4559
pgoff_t page_idx = 0, last_idx, cluster_idx;
fs/f2fs/file.c
4597
page_idx = cluster_idx << fi->i_log_cluster_size;
fs/f2fs/file.c
4599
if (f2fs_is_sparse_cluster(inode, page_idx))
fs/f2fs/file.c
4602
ret = redirty_blocks(inode, page_idx, fi->i_cluster_size);
fs/ntfs3/fslog.c
1148
u32 page_idx = vbo >> log->page_bits;
fs/ntfs3/fslog.c
1152
u32 page_vbo = page_idx << log->page_bits;
fs/ubifs/file.c
703
int err, page_idx, page_cnt, ret = 0, n = 0;
fs/ubifs/file.c
762
for (page_idx = 1; page_idx < page_cnt; page_idx++) {
fs/ubifs/file.c
763
pgoff_t page_offset = offset + page_idx;
fs/ubifs/file.c
781
ui->last_page_read = offset + page_idx - 1;
mm/memory.c
4718
unsigned long page_idx;
mm/memory.c
4892
page_idx = 0;
mm/memory.c
4914
page_idx = idx;
mm/memory.c
5030
vmf->orig_pte = pte_advance_pfn(pte, page_idx);
mm/percpu-vm.c
14
unsigned int cpu, int page_idx)
mm/percpu-vm.c
19
return vmalloc_to_page((void *)pcpu_chunk_addr(chunk, cpu, page_idx));
mm/percpu.c
265
static int __maybe_unused pcpu_page_idx(unsigned int cpu, int page_idx)
mm/percpu.c
267
return pcpu_unit_map[cpu] * pcpu_unit_pages + page_idx;
mm/percpu.c
270
static unsigned long pcpu_unit_page_offset(unsigned int cpu, int page_idx)
mm/percpu.c
272
return pcpu_unit_offsets[cpu] + (page_idx << PAGE_SHIFT);
mm/percpu.c
276
unsigned int cpu, int page_idx)
mm/percpu.c
279
pcpu_unit_page_offset(cpu, page_idx);
sound/soc/amd/acp-pcm-dma.c
301
u16 page_idx;
sound/soc/amd/acp-pcm-dma.c
307
for (page_idx = 0; page_idx < (num_of_pages); page_idx++) {
sound/soc/amd/acp-pcm-dma.c
309
acp_reg_write((offset + (page_idx * 8)),
sound/soc/amd/acp-pcm-dma.c
318
acp_reg_write((offset + (page_idx * 8) + 4),
sound/soc/amd/acp/acp-platform.c
138
u16 page_idx;
sound/soc/amd/acp/acp-platform.c
176
for (page_idx = 0; page_idx < num_pages; page_idx++) {
sound/soc/amd/ps/ps-pdm-dma.c
159
u16 page_idx;
sound/soc/amd/ps/ps-pdm-dma.c
169
for (page_idx = 0; page_idx < rtd->num_pages; page_idx++) {
sound/soc/amd/ps/ps-sdw-dma.c
209
u16 page_idx;
sound/soc/amd/ps/ps-sdw-dma.c
221
for (page_idx = 0; page_idx < stream->num_pages; page_idx++) {
sound/soc/amd/raven/acp3x-pcm-dma.c
109
u16 page_idx;
sound/soc/amd/raven/acp3x-pcm-dma.c
141
for (page_idx = 0; page_idx < rtd->num_pages; page_idx++) {
sound/soc/amd/renoir/acp3x-pdm-dma.c
178
u16 page_idx;
sound/soc/amd/renoir/acp3x-pdm-dma.c
191
for (page_idx = 0; page_idx < rtd->num_pages; page_idx++) {
sound/soc/amd/vangogh/acp5x-pcm-dma.c
105
u16 page_idx;
sound/soc/amd/vangogh/acp5x-pcm-dma.c
136
for (page_idx = 0; page_idx < rtd->num_pages; page_idx++) {
sound/soc/amd/yc/acp6x-pdm-dma.c
155
u16 page_idx;
sound/soc/amd/yc/acp6x-pdm-dma.c
167
for (page_idx = 0; page_idx < rtd->num_pages; page_idx++) {
sound/soc/sof/amd/acp-loader.c
123
u16 page_idx;
sound/soc/sof/amd/acp-loader.c
150
for (page_idx = 0; page_idx < num_pages; page_idx++) {
sound/soc/sof/amd/acp-stream.c
106
for (page_idx = 0; page_idx < stream->num_pages; page_idx++) {
sound/soc/sof/amd/acp-stream.c
107
addr = snd_sgbuf_get_addr(stream->dmab, page_idx * PAGE_SIZE);
sound/soc/sof/amd/acp-stream.c
34
int page_idx;
tools/testing/selftests/kvm/s390/keyop.c
71
static int do_keyop(struct kvm_vcpu *vcpu, int op, unsigned long page_idx, unsigned char skey)
tools/testing/selftests/kvm/s390/keyop.c
74
.guest_addr = BUF_START_ADDR + page_idx * PAGE_SIZE,
tools/testing/selftests/kvm/s390/keyop.c
84
page_idx, skey, keyop.key);