Symbol: vmap
arch/arm/lib/uaccess_with_memcpy.c
243
user_ptr = vmap(&dst_page, 1, VM_IOREMAP, __pgprot(__PAGE_COPY));
arch/arm/mm/fault-armv.c
253
p1 = vmap(&page, 1, VM_IOREMAP, prot);
arch/arm/mm/fault-armv.c
254
p2 = vmap(&page, 1, VM_IOREMAP, prot);
arch/hexagon/kernel/vdso.c
29
vdso = vmap(&vdso_page, 1, 0, PAGE_KERNEL);
arch/powerpc/kvm/e500_mmu.c
789
virt = vmap(pages, num_pages, VM_MAP, PAGE_KERNEL);
arch/powerpc/perf/vpa-dtl.c
533
buf->base = vmap(pglist, nr_pages, VM_MAP, PAGE_KERNEL);
arch/sh/kernel/io_trapped.c
73
tiop->virt_base = vmap(pages, n, VM_MAP, PAGE_NONE);
arch/x86/kernel/irq_64.c
49
va = vmap(pages, IRQ_STACK_SIZE / PAGE_SIZE, VM_MAP, PAGE_KERNEL);
drivers/accel/amdxdna/amdxdna_gem.c
389
.vmap = drm_gem_dmabuf_vmap,
drivers/accel/amdxdna/amdxdna_gem.c
487
.vmap = drm_gem_shmem_object_vmap,
drivers/accel/amdxdna/amdxdna_ubuf.c
113
kva = vmap(ubuf->pages, ubuf->nr_pages, VM_MAP, PAGE_KERNEL);
drivers/accel/amdxdna/amdxdna_ubuf.c
131
.vmap = amdxdna_ubuf_vmap,
drivers/accel/ethosu/ethosu_gem.c
36
.vmap = drm_gem_dma_object_vmap,
drivers/accel/ivpu/ivpu_gem.c
354
.vmap = drm_gem_shmem_object_vmap,
drivers/accel/ivpu/ivpu_mmu_context.c
63
cpu = vmap(&page, 1, VM_MAP, pgprot_writecombine(PAGE_KERNEL));
drivers/accel/rocket/rocket_gem.c
41
.vmap = drm_gem_shmem_object_vmap,
drivers/base/firmware_loader/main.c
298
fw_priv->data = vmap(fw_priv->pages, fw_priv->nr_pages, 0,
drivers/char/agp/uninorth-agp.c
423
bridge->gatt_table = vmap(uninorth_priv.pages_arr, (1 << page_order), 0, PAGE_KERNEL_NCG);
drivers/dma-buf/dma-buf.c
1593
if (!dmabuf->ops->vmap)
drivers/dma-buf/dma-buf.c
1605
ret = dmabuf->ops->vmap(dmabuf, &ptr);
drivers/dma-buf/heaps/cma_heap.c
217
vaddr = vmap(buffer->pages, buffer->pagecount, VM_MAP, PAGE_KERNEL);
drivers/dma-buf/heaps/cma_heap.c
290
.vmap = cma_heap_vmap,
drivers/dma-buf/heaps/system_heap.c
238
vaddr = vmap(pages, npages, VM_MAP, PAGE_KERNEL);
drivers/dma-buf/heaps/system_heap.c
313
.vmap = system_heap_vmap,
drivers/dma-buf/udmabuf.c
282
.vmap = vmap_udmabuf,
drivers/firewire/ohci.c
875
vaddr = vmap(pages, ARRAY_SIZE(pages), VM_MAP, PAGE_KERNEL);
drivers/firmware/efi/capsule-loader.c
119
cap_info->capsule = vmap(cap_info->pages, cap_info->index,
drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c
364
.vmap = amdgpu_dma_buf_vmap,
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
393
.vmap = drm_gem_ttm_vmap,
drivers/gpu/drm/drm_gem.c
1384
if (!obj->funcs->vmap)
drivers/gpu/drm/drm_gem.c
1387
ret = obj->funcs->vmap(obj, map);
drivers/gpu/drm/drm_gem_dma_helper.c
54
.vmap = drm_gem_dma_object_vmap,
drivers/gpu/drm/drm_gem_shmem_helper.c
393
shmem->vaddr = vmap(shmem->pages, obj->size >> PAGE_SHIFT,
drivers/gpu/drm/drm_gem_shmem_helper.c
48
.vmap = drm_gem_shmem_object_vmap,
drivers/gpu/drm/drm_gem_vram_helper.c
694
.vmap = drm_gem_vram_object_vmap,
drivers/gpu/drm/drm_prime.c
834
.vmap = drm_gem_dmabuf_vmap,
drivers/gpu/drm/etnaviv/etnaviv_gem.c
337
etnaviv_obj->vaddr = etnaviv_obj->ops->vmap(etnaviv_obj);
drivers/gpu/drm/etnaviv/etnaviv_gem.c
366
return vmap(pages, obj->base.size >> PAGE_SHIFT, VM_MAP, prot);
drivers/gpu/drm/etnaviv/etnaviv_gem.c
497
.vmap = etnaviv_gem_vmap_impl,
drivers/gpu/drm/etnaviv/etnaviv_gem.c
566
.vmap = etnaviv_gem_prime_vmap,
drivers/gpu/drm/etnaviv/etnaviv_gem.c
730
.vmap = etnaviv_gem_vmap_impl,
drivers/gpu/drm/etnaviv/etnaviv_gem.h
70
void *(*vmap)(struct etnaviv_gem_object *);
drivers/gpu/drm/etnaviv/etnaviv_gem_prime.c
108
.vmap = etnaviv_gem_prime_vmap_impl,
drivers/gpu/drm/gud/gud_pipe.c
60
struct iosys_map dst_map, vmap;
drivers/gpu/drm/gud/gud_pipe.c
77
iosys_map_set_vaddr(&vmap, src);
drivers/gpu/drm/gud/gud_pipe.c
78
drm_fb_xrgb8888_to_gray8(&dst_map, NULL, &vmap, fb, rect, fmtcnv_state);
drivers/gpu/drm/hyperv/hyperv_drm_modeset.c
29
const struct iosys_map *vmap,
drivers/gpu/drm/hyperv/hyperv_drm_modeset.c
40
drm_fb_memcpy(&dst, fb->pitches, vmap, fb, rect);
drivers/gpu/drm/i915/gem/i915_gem_dmabuf.c
210
.vmap = i915_gem_dmabuf_vmap,
drivers/gpu/drm/i915/gem/i915_gem_object.c
925
.vmap = i915_gem_vmap_object,
drivers/gpu/drm/i915/gem/i915_gem_pages.c
325
vaddr = vmap(pages, n_pages, 0, pgprot);
drivers/gpu/drm/i915/gem/selftests/mock_dmabuf.c
95
.vmap = mock_dmabuf_vmap,
drivers/gpu/drm/i915/gt/shmem_utils.c
77
vaddr = vmap(pages, n_pages, VM_MAP_PUT_PAGES, PAGE_KERNEL);
drivers/gpu/drm/imagination/pvr_gem.c
59
.vmap = drm_gem_shmem_object_vmap,
drivers/gpu/drm/imagination/pvr_mmu.c
277
host_ptr = vmap(&raw_page, 1, VM_MAP, prot);
drivers/gpu/drm/imagination/pvr_vm_mips.c
66
mips_data->pt = vmap(mips_data->pt_pages, pt_size >> PAGE_SHIFT, VM_MAP,
drivers/gpu/drm/lima/lima_gem.c
219
.vmap = lima_gem_vmap,
drivers/gpu/drm/lima/lima_sched.c
365
data = vmap(bo->base.pages, bo->heap_size >> PAGE_SHIFT,
drivers/gpu/drm/loongson/lsdc_gem.c
137
.vmap = lsdc_gem_object_vmap,
drivers/gpu/drm/mediatek/mtk_gem.c
72
.vmap = drm_gem_dma_object_vmap,
drivers/gpu/drm/mgag200/mgag200_mode.c
450
static void mgag200_handle_damage(struct mga_device *mdev, const struct iosys_map *vmap,
drivers/gpu/drm/mgag200/mgag200_mode.c
456
drm_fb_memcpy(&dst, fb->pitches, vmap, fb, clip);
drivers/gpu/drm/msm/msm_gem.c
1191
.vmap = msm_gem_prime_vmap,
drivers/gpu/drm/msm/msm_gem.c
731
msm_obj->vaddr = vmap(pages, obj->size >> PAGE_SHIFT,
drivers/gpu/drm/msm/msm_gem_prime.c
61
.vmap = drm_gem_dmabuf_vmap,
drivers/gpu/drm/nouveau/nouveau_gem.c
225
.vmap = drm_gem_ttm_vmap,
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c
101
info->min = nvbios_rd32(bios, vmap + 0x00);
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c
102
info->max = nvbios_rd32(bios, vmap + 0x04);
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c
103
info->arg[0] = nvbios_rd32(bios, vmap + 0x08);
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c
104
info->arg[1] = nvbios_rd32(bios, vmap + 0x0c);
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c
105
info->arg[2] = nvbios_rd32(bios, vmap + 0x10);
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c
108
info->mode = nvbios_rd08(bios, vmap + 0x00);
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c
109
info->link = nvbios_rd08(bios, vmap + 0x01);
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c
110
info->min = nvbios_rd32(bios, vmap + 0x02);
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c
111
info->max = nvbios_rd32(bios, vmap + 0x06);
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c
112
info->arg[0] = nvbios_rd32(bios, vmap + 0x0a);
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c
113
info->arg[1] = nvbios_rd32(bios, vmap + 0x0e);
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c
114
info->arg[2] = nvbios_rd32(bios, vmap + 0x12);
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c
115
info->arg[3] = nvbios_rd32(bios, vmap + 0x16);
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c
116
info->arg[4] = nvbios_rd32(bios, vmap + 0x1a);
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c
117
info->arg[5] = nvbios_rd32(bios, vmap + 0x1e);
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c
120
return vmap;
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c
32
u32 vmap = 0;
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c
36
vmap = nvbios_rd32(bios, bit_P.offset + 0x20);
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c
37
if (vmap) {
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c
38
*ver = nvbios_rd08(bios, vmap + 0);
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c
42
*hdr = nvbios_rd08(bios, vmap + 1);
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c
43
*cnt = nvbios_rd08(bios, vmap + 3);
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c
44
*len = nvbios_rd08(bios, vmap + 2);
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c
45
return vmap;
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c
60
u32 vmap = nvbios_vmap_table(bios, ver, hdr, cnt, len);
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c
62
switch (!!vmap * *ver) {
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c
69
info->max0 = nvbios_rd08(bios, vmap + 0x7);
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c
70
info->max1 = nvbios_rd08(bios, vmap + 0x8);
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c
72
info->max2 = nvbios_rd08(bios, vmap + 0xc);
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c
77
return vmap;
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c
84
u32 vmap = nvbios_vmap_table(bios, ver, &hdr, &cnt, len);
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c
85
if (vmap && idx < cnt) {
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c
86
vmap = vmap + hdr + (idx * *len);
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c
87
return vmap;
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c
96
u32 vmap = nvbios_vmap_entry(bios, idx, ver, len);
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c
98
switch (!!vmap * *ver) {
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
209
node->base.vaddr = vmap(node->pages, size >> PAGE_SHIFT, VM_MAP,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
137
*pmap = vmap(mem->mem, mem->pages, VM_MAP, PAGE_KERNEL);
drivers/gpu/drm/nouveau/nvkm/subdev/volt/base.c
109
u32 vmap;
drivers/gpu/drm/nouveau/nvkm/subdev/volt/base.c
111
vmap = nvbios_vmap_entry_parse(bios, id, &ver, &len, &info);
drivers/gpu/drm/nouveau/nvkm/subdev/volt/base.c
112
if (vmap) {
drivers/gpu/drm/nouveau/nvkm/subdev/volt/base.c
295
struct nvbios_vmap vmap;
drivers/gpu/drm/nouveau/nvkm/subdev/volt/base.c
301
if (nvbios_vmap_parse(bios, &ver, &hdr, &cnt, &len, &vmap)) {
drivers/gpu/drm/nouveau/nvkm/subdev/volt/base.c
302
volt->max0_id = vmap.max0;
drivers/gpu/drm/nouveau/nvkm/subdev/volt/base.c
303
volt->max1_id = vmap.max1;
drivers/gpu/drm/nouveau/nvkm/subdev/volt/base.c
304
volt->max2_id = vmap.max2;
drivers/gpu/drm/nouveau/nvkm/subdev/volt/base.c
87
u32 vmap;
drivers/gpu/drm/nouveau/nvkm/subdev/volt/base.c
89
vmap = nvbios_vmap_entry_parse(bios, id, &ver, &len, &info);
drivers/gpu/drm/nouveau/nvkm/subdev/volt/base.c
90
if (vmap) {
drivers/gpu/drm/omapdrm/omap_gem.c
1081
omap_obj->vaddr = vmap(omap_obj->pages, obj->size >> PAGE_SHIFT,
drivers/gpu/drm/panfrost/panfrost_gem.c
362
.vmap = drm_gem_dmabuf_vmap,
drivers/gpu/drm/panfrost/panfrost_gem.c
411
.vmap = drm_gem_shmem_object_vmap,
drivers/gpu/drm/panthor/panthor_gem.c
321
.vmap = drm_gem_dmabuf_vmap,
drivers/gpu/drm/panthor/panthor_gem.c
382
.vmap = drm_gem_shmem_object_vmap,
drivers/gpu/drm/qxl/qxl_object.c
98
.vmap = qxl_gem_prime_vmap,
drivers/gpu/drm/radeon/radeon_gem.c
285
.vmap = drm_gem_ttm_vmap,
drivers/gpu/drm/renesas/rcar-du/rcar_du_kms.c
366
.vmap = drm_gem_dma_object_vmap,
drivers/gpu/drm/rockchip/rockchip_drm_gem.c
139
rk_obj->kvaddr = vmap(rk_obj->pages, rk_obj->num_pages, VM_MAP,
drivers/gpu/drm/rockchip/rockchip_drm_gem.c
278
.vmap = rockchip_gem_prime_vmap,
drivers/gpu/drm/rockchip/rockchip_drm_gem.c
524
vaddr = vmap(rk_obj->pages, rk_obj->num_pages, VM_MAP,
drivers/gpu/drm/sitronix/st7571.c
154
const struct iosys_map *vmap,
drivers/gpu/drm/sitronix/st7571.c
168
drm_fb_xrgb8888_to_mono(&dst, &dst_pitch, vmap, fb, rect, fmtcnv_state);
drivers/gpu/drm/sitronix/st7571.c
173
memcpy(st7571->hwbuf, vmap->vaddr, size);
drivers/gpu/drm/sitronix/st7571.c
179
const struct iosys_map *vmap,
drivers/gpu/drm/sitronix/st7571.c
193
drm_fb_xrgb8888_to_gray2(&dst, &dst_pitch, vmap, fb, rect, fmtcnv_state);
drivers/gpu/drm/sitronix/st7571.c
198
memcpy(st7571->hwbuf, vmap->vaddr, size);
drivers/gpu/drm/sitronix/st7571.c
203
memcpy(st7571->hwbuf, vmap->vaddr, size);
drivers/gpu/drm/sitronix/st7571.h
43
const struct iosys_map *vmap,
drivers/gpu/drm/sitronix/st7586.c
75
struct iosys_map dst_map, vmap;
drivers/gpu/drm/sitronix/st7586.c
82
iosys_map_set_vaddr(&vmap, vaddr);
drivers/gpu/drm/sitronix/st7586.c
83
drm_fb_xrgb8888_to_gray8(&dst_map, NULL, &vmap, fb, clip, fmtcnv_state);
drivers/gpu/drm/sitronix/st7920.c
316
const struct iosys_map *vmap,
drivers/gpu/drm/sitronix/st7920.c
333
drm_fb_xrgb8888_to_mono(&dst, &dst_pitch, vmap, fb, rect, fmtcnv_state);
drivers/gpu/drm/solomon/ssd130x.c
1002
const struct iosys_map *vmap,
drivers/gpu/drm/solomon/ssd130x.c
1019
drm_fb_xrgb8888_to_mono(&dst, &dst_pitch, vmap, fb, rect, fmtcnv_state);
drivers/gpu/drm/solomon/ssd130x.c
1027
const struct iosys_map *vmap,
drivers/gpu/drm/solomon/ssd130x.c
1045
drm_fb_xrgb8888_to_gray8(&dst, &dst_pitch, vmap, fb, rect, fmtcnv_state);
drivers/gpu/drm/solomon/ssd130x.c
1053
const struct iosys_map *vmap,
drivers/gpu/drm/solomon/ssd130x.c
1069
drm_fb_xrgb8888_to_rgb332(&dst, &dst_pitch, vmap, fb, rect, fmtcnv_state);
drivers/gpu/drm/tegra/fbdev.c
123
bo->vaddr = vmap(bo->pages, bo->num_pages, VM_MAP,
drivers/gpu/drm/tegra/gem.c
195
vaddr = vmap(obj->pages, obj->num_pages, VM_MAP,
drivers/gpu/drm/tegra/gem.c
754
.vmap = tegra_gem_prime_vmap,
drivers/gpu/drm/tiny/repaper.c
514
static int repaper_fb_dirty(struct drm_framebuffer *fb, const struct iosys_map *vmap,
drivers/gpu/drm/tiny/repaper.c
549
drm_fb_xrgb8888_to_mono(&dst, &dst_pitch, vmap, fb, &clip, fmtcnv_state);
drivers/gpu/drm/tiny/sharp-memory.c
129
const struct iosys_map *vmap,
drivers/gpu/drm/tiny/sharp-memory.c
143
drm_fb_xrgb8888_to_mono(&dst, &pitch, vmap, fb, &clip, fmtcnv_state);
drivers/gpu/drm/tiny/sharp-memory.c
150
const struct iosys_map *vmap,
drivers/gpu/drm/tiny/sharp-memory.c
166
sharp_memory_set_tx_buffer_data(&tx_buffer[2], fb, vmap, clip, pitch, fmtcnv_state);
drivers/gpu/drm/tiny/sharp-memory.c
209
static void sharp_memory_fb_dirty(struct drm_framebuffer *fb, const struct iosys_map *vmap,
drivers/gpu/drm/tiny/sharp-memory.c
222
sharp_memory_update_display(smd, fb, vmap, clip, fmtconv_state);
drivers/gpu/drm/ttm/ttm_bo_util.c
376
map->virtual = vmap(ttm->pages + start_page, num_pages,
drivers/gpu/drm/ttm/ttm_bo_util.c
545
vaddr = vmap(ttm->pages, ttm->num_pages, 0, prot);
drivers/gpu/drm/v3d/v3d_bo.c
198
bo->vaddr = vmap(obj->pages, obj->base.size >> PAGE_SHIFT, VM_MAP,
drivers/gpu/drm/v3d/v3d_bo.c
71
.vmap = drm_gem_shmem_object_vmap,
drivers/gpu/drm/vc4/vc4_bo.c
761
.vmap = drm_gem_dma_object_vmap,
drivers/gpu/drm/virtio/virtgpu_object.c
135
.vmap = drm_gem_shmem_object_vmap,
drivers/gpu/drm/virtio/virtgpu_prime.c
84
.vmap = drm_gem_dmabuf_vmap,
drivers/gpu/drm/vmwgfx/vmwgfx_gem.c
151
.vmap = vmw_gem_vmap,
drivers/gpu/drm/xe/display/intel_fbdev_fb.c
85
XE_WARN_ON(iosys_map_is_null(&obj->vmap));
drivers/gpu/drm/xe/display/intel_fbdev_fb.c
87
info->screen_base = obj->vmap.vaddr_iomem;
drivers/gpu/drm/xe/display/xe_dsb_buffer.c
24
iosys_map_wr(&dsb_buf->bo->vmap, idx * 4, u32, val);
drivers/gpu/drm/xe/display/xe_dsb_buffer.c
29
return iosys_map_rd(&dsb_buf->bo->vmap, idx * 4, u32);
drivers/gpu/drm/xe/display/xe_dsb_buffer.c
36
iosys_map_memset(&dsb_buf->bo->vmap, idx * 4, val, size);
drivers/gpu/drm/xe/display/xe_fb_pin.c
138
iosys_map_wr(&dpt->vmap, x * 8, u64, pte | addr);
drivers/gpu/drm/xe/display/xe_fb_pin.c
145
write_dpt_remapped(bo, &dpt->vmap, &dpt_ofs,
drivers/gpu/drm/xe/display/xe_fb_pin.c
157
write_dpt_rotated(bo, &dpt->vmap, &dpt_ofs,
drivers/gpu/drm/xe/display/xe_fb_pin.c
476
*map = vma->bo->vmap;
drivers/gpu/drm/xe/display/xe_hdcp_gsc.c
131
struct iosys_map *map = &gsc_context->hdcp_bo->vmap;
drivers/gpu/drm/xe/display/xe_hdcp_gsc.c
171
addr_in_wr_off = xe_gsc_emit_header(xe, &gsc_context->hdcp_bo->vmap,
drivers/gpu/drm/xe/display/xe_hdcp_gsc.c
174
xe_map_memcpy_to(xe, &gsc_context->hdcp_bo->vmap, addr_in_wr_off,
drivers/gpu/drm/xe/display/xe_hdcp_gsc.c
197
xe_map_memcpy_from(xe, msg_out, &gsc_context->hdcp_bo->vmap,
drivers/gpu/drm/xe/display/xe_hdcp_gsc.c
81
xe_map_memset(xe, &bo->vmap, 0, 0, xe_bo_size(bo));
drivers/gpu/drm/xe/display/xe_panic.c
16
struct iosys_map vmap;
drivers/gpu/drm/xe/display/xe_panic.c
23
if (!panic->vmap.is_iomem && iosys_map_is_set(&panic->vmap)) {
drivers/gpu/drm/xe/display/xe_panic.c
24
drm_clflush_virt_range(panic->vmap.vaddr, PAGE_SIZE);
drivers/gpu/drm/xe/display/xe_panic.c
25
kunmap_local(panic->vmap.vaddr);
drivers/gpu/drm/xe/display/xe_panic.c
27
iosys_map_clear(&panic->vmap);
drivers/gpu/drm/xe/display/xe_panic.c
63
iosys_map_set_vaddr_iomem(&panic->vmap,
drivers/gpu/drm/xe/display/xe_panic.c
67
iosys_map_set_vaddr(&panic->vmap,
drivers/gpu/drm/xe/display/xe_panic.c
74
if (iosys_map_is_set(&panic->vmap))
drivers/gpu/drm/xe/display/xe_panic.c
75
iosys_map_wr(&panic->vmap, offset, u32, color);
drivers/gpu/drm/xe/tests/xe_guc_buf_kunit.c
242
iosys_map_memcpy_from(d, &cache->sam->bo->vmap, 0, bytes);
drivers/gpu/drm/xe/tests/xe_guc_buf_kunit.c
36
iosys_map_set_vaddr(&bo->vmap, buf);
drivers/gpu/drm/xe/tests/xe_guc_g2g_test.c
394
xe_map_memset(xe, &bo->vmap, 0, 0, g2g_size);
drivers/gpu/drm/xe/tests/xe_guc_g2g_test.c
422
xe_map_memset(xe, &bo->vmap, 0, 0, g2g_size);
drivers/gpu/drm/xe/tests/xe_migrate.c
109
xe_map_memset(xe, &remote->vmap, 0, 0xd0, xe_bo_size(remote));
drivers/gpu/drm/xe/tests/xe_migrate.c
114
retval = xe_map_rd(xe, &remote->vmap, 0, u64);
drivers/gpu/drm/xe/tests/xe_migrate.c
117
retval = xe_map_rd(xe, &remote->vmap, xe_bo_size(remote) - 8, u64);
drivers/gpu/drm/xe/tests/xe_migrate.c
124
xe_map_memset(xe, &remote->vmap, 0, 0xc0, xe_bo_size(remote));
drivers/gpu/drm/xe/tests/xe_migrate.c
125
xe_map_memset(xe, &bo->vmap, 0, 0xd0, xe_bo_size(bo));
drivers/gpu/drm/xe/tests/xe_migrate.c
132
retval = xe_map_rd(xe, &bo->vmap, 0, u64);
drivers/gpu/drm/xe/tests/xe_migrate.c
135
retval = xe_map_rd(xe, &bo->vmap, xe_bo_size(bo) - 8, u64);
drivers/gpu/drm/xe/tests/xe_migrate.c
142
xe_map_memset(xe, &remote->vmap, 0, 0xd0, xe_bo_size(remote));
drivers/gpu/drm/xe/tests/xe_migrate.c
143
xe_map_memset(xe, &bo->vmap, 0, 0xc0, xe_bo_size(bo));
drivers/gpu/drm/xe/tests/xe_migrate.c
149
retval = xe_map_rd(xe, &remote->vmap, 0, u64);
drivers/gpu/drm/xe/tests/xe_migrate.c
152
retval = xe_map_rd(xe, &remote->vmap, xe_bo_size(bo) - 8, u64);
drivers/gpu/drm/xe/tests/xe_migrate.c
247
xe_map_wr(xe, &bo->vmap, XE_PAGE_SIZE * (NUM_KERNEL_PDE - 1), u64,
drivers/gpu/drm/xe/tests/xe_migrate.c
262
retval = xe_map_rd(xe, &bo->vmap, XE_PAGE_SIZE * (NUM_KERNEL_PDE - 1),
drivers/gpu/drm/xe/tests/xe_migrate.c
269
xe_map_wr(xe, &pt->vmap, 0, u32, 0xdeaddead);
drivers/gpu/drm/xe/tests/xe_migrate.c
277
retval = xe_map_rd(xe, &pt->vmap, 0, u32);
drivers/gpu/drm/xe/tests/xe_migrate.c
284
xe_map_memset(xe, &tiny->vmap, 0, 0x22, xe_bo_size(tiny));
drivers/gpu/drm/xe/tests/xe_migrate.c
292
retval = xe_map_rd(xe, &tiny->vmap, 0, u32);
drivers/gpu/drm/xe/tests/xe_migrate.c
294
retval = xe_map_rd(xe, &tiny->vmap, xe_bo_size(tiny) - 4, u32);
drivers/gpu/drm/xe/tests/xe_migrate.c
306
xe_map_memset(xe, &big->vmap, 0, 0x11, xe_bo_size(big));
drivers/gpu/drm/xe/tests/xe_migrate.c
314
retval = xe_map_rd(xe, &big->vmap, 0, u32);
drivers/gpu/drm/xe/tests/xe_migrate.c
316
retval = xe_map_rd(xe, &big->vmap, xe_bo_size(big) - 4, u32);
drivers/gpu/drm/xe/tests/xe_migrate.c
504
xe_map_memset(xe, &sys_bo->vmap, 0, 0xd0, xe_bo_size(sys_bo));
drivers/gpu/drm/xe/tests/xe_migrate.c
508
retval = xe_map_rd(xe, &vram_bo->vmap, 0, u64);
drivers/gpu/drm/xe/tests/xe_migrate.c
527
retval = xe_map_rd(xe, &vram_bo->vmap, 0, u64);
drivers/gpu/drm/xe/tests/xe_migrate.c
529
retval = xe_map_rd(xe, &vram_bo->vmap, xe_bo_size(vram_bo) - 8, u64);
drivers/gpu/drm/xe/tests/xe_migrate.c
535
retval = xe_map_rd(xe, &ccs_bo->vmap, 0, u64);
drivers/gpu/drm/xe/tests/xe_migrate.c
538
retval = xe_map_rd(xe, &ccs_bo->vmap, xe_bo_size(ccs_bo) - 8, u64);
drivers/gpu/drm/xe/tests/xe_migrate.c
566
retval = xe_map_rd(xe, &vram_bo->vmap, 0, u64);
drivers/gpu/drm/xe/tests/xe_migrate.c
568
retval = xe_map_rd(xe, &vram_bo->vmap, xe_bo_size(vram_bo) - 8, u64);
drivers/gpu/drm/xe/tests/xe_migrate.c
574
retval = xe_map_rd(xe, &ccs_bo->vmap, 0, u64);
drivers/gpu/drm/xe/tests/xe_migrate.c
576
retval = xe_map_rd(xe, &ccs_bo->vmap, xe_bo_size(ccs_bo) - 8, u64);
drivers/gpu/drm/xe/tests/xe_migrate.c
589
xe_map_memset(xe, &sys_bo->vmap, 0, 0xd0, xe_bo_size(sys_bo));
drivers/gpu/drm/xe/tests/xe_migrate.c
593
retval = xe_map_rd(xe, &vram_bo->vmap, 0, u64);
drivers/gpu/drm/xe/tests/xe_migrate.c
601
retval = xe_map_rd(xe, &sys_bo->vmap, 0, u64);
drivers/gpu/drm/xe/tests/xe_migrate.c
603
retval = xe_map_rd(xe, &sys_bo->vmap, xe_bo_size(sys_bo) - 8, u64);
drivers/gpu/drm/xe/tests/xe_migrate.c
619
retval = xe_map_rd(xe, &sys_bo->vmap, 0, u64);
drivers/gpu/drm/xe/tests/xe_migrate.c
621
retval = xe_map_rd(xe, &sys_bo->vmap, xe_bo_size(sys_bo) - 8, u64);
drivers/gpu/drm/xe/tests/xe_migrate.c
629
retval = xe_map_rd(xe, &sys_bo->vmap, 0, u64);
drivers/gpu/drm/xe/tests/xe_migrate.c
631
retval = xe_map_rd(xe, &sys_bo->vmap, xe_bo_size(sys_bo) - 8, u64);
drivers/gpu/drm/xe/xe_bo.c
1296
if (iosys_map_is_null(&bo->vmap)) {
drivers/gpu/drm/xe/xe_bo.c
1303
xe_map_memcpy_from(xe, backup->vmap.vaddr, &bo->vmap, 0,
drivers/gpu/drm/xe/xe_bo.c
1446
if (iosys_map_is_null(&bo->vmap)) {
drivers/gpu/drm/xe/xe_bo.c
1453
xe_map_memcpy_to(xe, &bo->vmap, 0, backup->vmap.vaddr,
drivers/gpu/drm/xe/xe_bo.c
1640
struct iosys_map vmap;
drivers/gpu/drm/xe/xe_bo.c
1669
iosys_map_set_vaddr_iomem(&vmap, (u8 __iomem *)vram->mapping +
drivers/gpu/drm/xe/xe_bo.c
1672
xe_map_memcpy_to(xe, &vmap, page_offset, buf, byte_count);
drivers/gpu/drm/xe/xe_bo.c
1674
xe_map_memcpy_from(xe, buf, &vmap, page_offset, byte_count);
drivers/gpu/drm/xe/xe_bo.c
2700
xe_map_memcpy_to(xe, &bo->vmap, 0, data, size);
drivers/gpu/drm/xe/xe_bo.c
2727
xe_assert(xe, !(*src)->vmap.is_iomem);
drivers/gpu/drm/xe/xe_bo.c
2729
bo = xe_managed_bo_create_from_data(xe, tile, (*src)->vmap.vaddr,
drivers/gpu/drm/xe/xe_bo.c
3029
if (!iosys_map_is_null(&bo->vmap))
drivers/gpu/drm/xe/xe_bo.c
3045
iosys_map_set_vaddr_iomem(&bo->vmap, (void __iomem *)virtual);
drivers/gpu/drm/xe/xe_bo.c
3047
iosys_map_set_vaddr(&bo->vmap, virtual);
drivers/gpu/drm/xe/xe_bo.c
3054
if (!iosys_map_is_null(&bo->vmap)) {
drivers/gpu/drm/xe/xe_bo.c
3055
iosys_map_clear(&bo->vmap);
drivers/gpu/drm/xe/xe_bo_types.h
51
struct iosys_map vmap;
drivers/gpu/drm/xe/xe_dma_buf.c
210
.vmap = drm_gem_dmabuf_vmap,
drivers/gpu/drm/xe/xe_eu_stall.c
769
xecore_buf->vaddr = stream->bo->vmap.vaddr + vaddr_offset;
drivers/gpu/drm/xe/xe_exec_queue.c
700
xe_map_memset(xe, &bo->vmap, 0, 0, SZ_4K);
drivers/gpu/drm/xe/xe_ggtt.c
501
xe_map_memset(xe, &ggtt->scratch->vmap, 0, 0, xe_bo_size(ggtt->scratch));
drivers/gpu/drm/xe/xe_gsc.c
150
wr_offset = xe_gsc_emit_header(xe, &bo->vmap, 0, HECI_MEADDRESS_MKHI, 0,
drivers/gpu/drm/xe/xe_gsc.c
152
wr_offset = emit_version_query_msg(xe, &bo->vmap, wr_offset);
drivers/gpu/drm/xe/xe_gsc.c
164
err = xe_gsc_read_out_header(xe, &bo->vmap, GSC_VER_PKT_SZ,
drivers/gpu/drm/xe/xe_gsc.c
172
compat->major = version_query_rd(xe, &bo->vmap, rd_offset, proj_major);
drivers/gpu/drm/xe/xe_gsc.c
173
compat->minor = version_query_rd(xe, &bo->vmap, rd_offset, compat_major);
drivers/gpu/drm/xe/xe_gsc.c
174
compat->patch = version_query_rd(xe, &bo->vmap, rd_offset, compat_minor);
drivers/gpu/drm/xe/xe_gsc.c
60
xe_map_memcpy_from(xe, storage, &gsc->fw.bo->vmap, 0, fw_size);
drivers/gpu/drm/xe/xe_gsc.c
61
xe_map_memcpy_to(xe, &gsc->private->vmap, 0, storage, fw_size);
drivers/gpu/drm/xe/xe_gsc.c
62
xe_map_memset(xe, &gsc->private->vmap, fw_size, 0,
drivers/gpu/drm/xe/xe_gsc_proxy.c
430
gsc->proxy.to_gsc = IOSYS_MAP_INIT_OFFSET(&bo->vmap, 0);
drivers/gpu/drm/xe/xe_gsc_proxy.c
431
gsc->proxy.from_gsc = IOSYS_MAP_INIT_OFFSET(&bo->vmap, GSC_PROXY_BUFFER_SIZE);
drivers/gpu/drm/xe/xe_gt.c
439
&q->lrc[0]->bo->vmap,
drivers/gpu/drm/xe/xe_guc.c
498
xe_map_memset(xe, &bo->vmap, 0, 0, g2g_size);
drivers/gpu/drm/xe/xe_guc_ads.c
57
return &ads->bo->vmap;
drivers/gpu/drm/xe/xe_guc_capture.c
1048
xe_map_memcpy_from(guc_to_xe(guc), out + fill_size, &guc->log.bo->vmap,
drivers/gpu/drm/xe/xe_guc_capture.c
1392
xe_map_memcpy_from(guc_to_xe(guc), &log_buf_state_local, &guc->log.bo->vmap,
drivers/gpu/drm/xe/xe_guc_capture.c
1435
xe_map_wr(guc_to_xe(guc), &guc->log.bo->vmap,
drivers/gpu/drm/xe/xe_guc_capture.c
1445
xe_map_wr(guc_to_xe(guc), &guc->log.bo->vmap,
drivers/gpu/drm/xe/xe_guc_ct.c
2015
xe_map_memcpy_from(xe, snapshot->ctb, &ct->bo->vmap, 0, snapshot->ctb_size);
drivers/gpu/drm/xe/xe_guc_ct.c
608
xe_map_memset(xe, &ct->bo->vmap, 0, 0, xe_bo_size(ct->bo));
drivers/gpu/drm/xe/xe_guc_ct.c
609
guc_ct_ctb_h2g_init(xe, &ct->ctbs.h2g, &ct->bo->vmap);
drivers/gpu/drm/xe/xe_guc_ct.c
610
guc_ct_ctb_g2h_init(xe, &ct->ctbs.g2h, &ct->bo->vmap);
drivers/gpu/drm/xe/xe_guc_ct.c
627
xe_map_memset(xe, &ct->bo->vmap, CTB_H2G_BUFFER_OFFSET, 0,
drivers/gpu/drm/xe/xe_guc_engine_activity.c
44
return IOSYS_MAP_INIT_OFFSET(&buffer->activity_bo->vmap, offset);
drivers/gpu/drm/xe/xe_guc_engine_activity.c
62
return IOSYS_MAP_INIT_OFFSET(&buffer->metadata_bo->vmap, offset);
drivers/gpu/drm/xe/xe_guc_hwconfig.c
104
xe_map_memcpy_from(xe, dst, &guc->hwconfig.bo->vmap, 0,
drivers/gpu/drm/xe/xe_guc_log.c
208
xe_map_memcpy_from(xe, snapshot->copy[i], &log->bo->vmap,
drivers/gpu/drm/xe/xe_guc_log.c
638
xe_map_memset(xe, &bo->vmap, 0, 0, xe_bo_size(bo));
drivers/gpu/drm/xe/xe_guc_pc.c
119
return &pc->bo->vmap;
drivers/gpu/drm/xe/xe_guc_pc.c
1258
xe_map_memset(xe, &pc->bo->vmap, 0, 0, size);
drivers/gpu/drm/xe/xe_guc_submit.c
732
xe_map_wr(xe, &group->cgp_bo->vmap, 0, u32,
drivers/gpu/drm/xe/xe_guc_submit.c
735
xe_map_wr(xe, &group->cgp_bo->vmap,
drivers/gpu/drm/xe/xe_guc_submit.c
739
xe_map_wr(xe, &group->cgp_bo->vmap,
drivers/gpu/drm/xe/xe_guc_submit.c
744
xe_map_wr(xe, &group->cgp_bo->vmap, 17 * sizeof(u32),
drivers/gpu/drm/xe/xe_guc_submit.c
746
xe_map_wr(xe, &group->cgp_bo->vmap, 16 * sizeof(u32), u32, 0);
drivers/gpu/drm/xe/xe_guc_submit.c
748
xe_map_wr(xe, &group->cgp_bo->vmap, 16 * sizeof(u32),
drivers/gpu/drm/xe/xe_guc_submit.c
750
xe_map_wr(xe, &group->cgp_bo->vmap, 17 * sizeof(u32), u32, 0);
drivers/gpu/drm/xe/xe_huc.c
173
wr_offset = xe_gsc_emit_header(xe, &pkt->vmap, 0, HECI_MEADDRESS_PXP, 0,
drivers/gpu/drm/xe/xe_huc.c
175
wr_offset = huc_emit_pxp_auth_msg(xe, &pkt->vmap, wr_offset,
drivers/gpu/drm/xe/xe_huc.c
185
if (xe_gsc_check_and_update_pending(xe, &pkt->vmap, 0, &pkt->vmap,
drivers/gpu/drm/xe/xe_huc.c
197
err = xe_gsc_read_out_header(xe, &pkt->vmap, PXP43_HUC_AUTH_INOUT_SIZE,
drivers/gpu/drm/xe/xe_huc.c
209
out_status = huc_auth_msg_rd(xe, &pkt->vmap, rd_offset, header.status);
drivers/gpu/drm/xe/xe_lmtt.c
310
lmtt_assert(lmtt, !pte || !iosys_map_rd(&pt->bo->vmap, idx * sizeof(u32), u32));
drivers/gpu/drm/xe/xe_lmtt.c
312
xe_map_wr(lmtt_to_xe(lmtt), &pt->bo->vmap, idx * sizeof(u32), u32, pte);
drivers/gpu/drm/xe/xe_lmtt.c
315
lmtt_assert(lmtt, !pte || !iosys_map_rd(&pt->bo->vmap, idx * sizeof(u64), u64));
drivers/gpu/drm/xe/xe_lmtt.c
317
xe_map_wr(lmtt_to_xe(lmtt), &pt->bo->vmap, idx * sizeof(u64), u64, pte);
drivers/gpu/drm/xe/xe_lmtt.c
87
xe_map_memset(lmtt_to_xe(lmtt), &bo->vmap, 0, 0, xe_bo_size(bo));
drivers/gpu/drm/xe/xe_lrc.c
1237
if (state->lrc->bo->vmap.is_iomem) {
drivers/gpu/drm/xe/xe_lrc.c
1241
state->ptr = state->lrc->bo->vmap.vaddr + state->offset;
drivers/gpu/drm/xe/xe_lrc.c
1271
if (!state->lrc->bo->vmap.is_iomem)
drivers/gpu/drm/xe/xe_lrc.c
1274
xe_map_memcpy_to(gt_to_xe(state->lrc->gt), &state->lrc->bo->vmap,
drivers/gpu/drm/xe/xe_lrc.c
1326
if (lrc->bo->vmap.is_iomem) {
drivers/gpu/drm/xe/xe_lrc.c
1373
if (lrc->bo->vmap.is_iomem) {
drivers/gpu/drm/xe/xe_lrc.c
807
struct iosys_map map = lrc->bo->vmap; \
drivers/gpu/drm/xe/xe_memirq.c
191
iosys_map_memset(&bo->vmap, 0, 0, bo_size);
drivers/gpu/drm/xe/xe_memirq.c
194
memirq->source = IOSYS_MAP_INIT_OFFSET(&bo->vmap, XE_MEMIRQ_SOURCE_OFFSET(0));
drivers/gpu/drm/xe/xe_memirq.c
195
memirq->status = IOSYS_MAP_INIT_OFFSET(&bo->vmap, XE_MEMIRQ_STATUS_OFFSET(0));
drivers/gpu/drm/xe/xe_memirq.c
196
memirq->mask = IOSYS_MAP_INIT_OFFSET(&bo->vmap, XE_MEMIRQ_ENABLE_OFFSET);
drivers/gpu/drm/xe/xe_memirq.c
472
struct iosys_map src_offset = IOSYS_MAP_INIT_OFFSET(&memirq->bo->vmap,
drivers/gpu/drm/xe/xe_memirq.c
477
IOSYS_MAP_INIT_OFFSET(&memirq->bo->vmap,
drivers/gpu/drm/xe/xe_migrate.c
167
xe_map_wr(xe, &bo->vmap, ofs, u64, entry);
drivers/gpu/drm/xe/xe_migrate.c
1742
&update->pt_bo->vmap, NULL,
drivers/gpu/drm/xe/xe_migrate.c
1747
&update->pt_bo->vmap, NULL,
drivers/gpu/drm/xe/xe_migrate.c
176
xe_map_wr(xe, &bo->vmap, ofs, u64, pos | flags);
drivers/gpu/drm/xe/xe_migrate.c
180
xe_map_wr(xe, &bo->vmap, ofs, u64, pos | flags);
drivers/gpu/drm/xe/xe_migrate.c
221
xe_pt_write(xe, &vm->pt_root[id]->bo->vmap, 0, entry);
drivers/gpu/drm/xe/xe_migrate.c
230
xe_map_wr(xe, &bo->vmap, map_ofs + level * 8, u64, entry);
drivers/gpu/drm/xe/xe_migrate.c
247
xe_map_wr(xe, &bo->vmap, map_ofs + level * 8, u64,
drivers/gpu/drm/xe/xe_migrate.c
264
xe_map_wr(xe, &bo->vmap, map_ofs + level * 8, u64,
drivers/gpu/drm/xe/xe_migrate.c
289
xe_map_wr(xe, &bo->vmap, map_ofs + XE_PAGE_SIZE * level, u64,
drivers/gpu/drm/xe/xe_migrate.c
297
xe_map_wr(xe, &bo->vmap, map_ofs + XE_PAGE_SIZE +
drivers/gpu/drm/xe/xe_migrate.c
309
xe_map_wr(xe, &bo->vmap, map_ofs + XE_PAGE_SIZE * level + 255 * 8, u64,
drivers/gpu/drm/xe/xe_oa.c
895
xe_assert(stream->oa->xe, bo->vmap.is_iomem == 0);
drivers/gpu/drm/xe/xe_oa.c
896
stream->oa_buffer.vaddr = bo->vmap.vaddr;
drivers/gpu/drm/xe/xe_pt.c
1659
struct iosys_map *leaf_map = &xe_child->bo->vmap;
drivers/gpu/drm/xe/xe_pt.c
166
struct iosys_map *map = &pt->bo->vmap;
drivers/gpu/drm/xe/xe_pt.c
238
struct iosys_map *map = &pt->bo->vmap;
drivers/gpu/drm/xe/xe_pt.c
409
struct iosys_map *map = &parent->bo->vmap;
drivers/gpu/drm/xe/xe_pt.c
874
xe_map_memset(tile_to_xe(xe_walk->tile), &xe_child->bo->vmap,
drivers/gpu/drm/xe/xe_pxp_submit.c
157
gsc_res->batch = IOSYS_MAP_INIT_OFFSET(&bo->vmap, 0);
drivers/gpu/drm/xe/xe_pxp_submit.c
158
gsc_res->msg_in = IOSYS_MAP_INIT_OFFSET(&bo->vmap, PXP_BB_SIZE);
drivers/gpu/drm/xe/xe_pxp_submit.c
159
gsc_res->msg_out = IOSYS_MAP_INIT_OFFSET(&bo->vmap, PXP_BB_SIZE + inout_size);
drivers/gpu/drm/xe/xe_pxp_submit.c
310
offset = pxp_emit_session_termination(pxp->xe, &pxp->vcs_exec.bo->vmap, offset, id);
drivers/gpu/drm/xe/xe_pxp_submit.c
311
offset = pxp_emit_wait(pxp->xe, &pxp->vcs_exec.bo->vmap, offset);
drivers/gpu/drm/xe/xe_pxp_submit.c
312
emit_cmd(pxp->xe, &pxp->vcs_exec.bo->vmap, offset, MI_BATCH_BUFFER_END);
drivers/gpu/drm/xe/xe_sa.c
130
if (!sa_manager->bo->vmap.is_iomem)
drivers/gpu/drm/xe/xe_sa.c
131
sa_manager->cpu_ptr = sa_manager->bo->vmap.vaddr;
drivers/gpu/drm/xe/xe_sa.c
150
xe_map_memcpy_to(xe, &sa_manager->shadow->vmap,
drivers/gpu/drm/xe/xe_sa.c
187
if (!sa_manager->bo->vmap.is_iomem)
drivers/gpu/drm/xe/xe_sa.c
190
xe_map_memcpy_to(xe, &sa_manager->bo->vmap, drm_suballoc_soffset(sa_bo),
drivers/gpu/drm/xe/xe_sa.c
204
if (!sa_manager->bo->vmap.is_iomem)
drivers/gpu/drm/xe/xe_sa.c
207
xe_map_memcpy_from(xe, xe_sa_bo_cpu_addr(sa_bo), &sa_manager->bo->vmap,
drivers/gpu/drm/xe/xe_sa.c
74
sa_manager->is_iomem = bo->vmap.is_iomem;
drivers/gpu/drm/xe/xe_sa.c
76
if (bo->vmap.is_iomem) {
drivers/gpu/drm/xe/xe_sa.c
81
sa_manager->cpu_ptr = bo->vmap.vaddr;
drivers/gpu/drm/xe/xe_sriov_packet.c
139
data->vaddr = bo->vmap.vaddr;
drivers/gpu/drm/xe/xe_sriov_vf_ccs.c
164
xe_map_memset(xe, &sa_manager->bo->vmap, offset, MI_NOOP,
drivers/gpu/drm/xe/xe_sriov_vf_ccs.c
166
xe_map_memset(xe, &sa_manager->shadow->vmap, offset, MI_NOOP,
drivers/gpu/drm/xe/xe_sriov_vf_ccs.c
170
xe_map_wr(xe, &sa_manager->bo->vmap, offset, u32, MI_BATCH_BUFFER_END);
drivers/gpu/drm/xe/xe_sriov_vf_ccs.c
171
xe_map_wr(xe, &sa_manager->shadow->vmap, offset, u32, MI_BATCH_BUFFER_END);
drivers/gpu/drm/xe/xe_sriov_vf_ccs.c
396
xe_map_wr(xe, &lrc->bo->vmap, XE_SRIOV_VF_CCS_RW_BB_ADDR_OFFSET, u32, addr);
drivers/gpu/drm/xe/xe_uc_fw.c
317
xe_map_memcpy_from(xe, dst, &uc_fw->bo->vmap,
drivers/gpu/drm/xen/xen_drm_front_gem.c
108
.vmap = xen_drm_front_gem_prime_vmap,
drivers/gpu/drm/xen/xen_drm_front_gem.c
292
vaddr = vmap(xen_obj->pages, xen_obj->num_pages,
drivers/hv/ring_buffer.c
209
vmap(pages_wraparound, page_cnt * 2 - 1, VM_MAP,
drivers/hwtracing/coresight/coresight-tmc-etr.c
283
sg_table->table_vaddr = vmap(table_pages->pages,
drivers/hwtracing/coresight/coresight-tmc-etr.c
303
sg_table->data_vaddr = vmap(sg_table->data_pages.pages,
drivers/hwtracing/coresight/coresight-trbe.c
763
buf->trbe_base = (unsigned long)vmap(pglist, nr_pages, VM_MAP, PAGE_KERNEL);
drivers/hwtracing/ptt/hisi_ptt.c
1057
buf->base = vmap(pagelist, nr_pages, VM_MAP, PAGE_KERNEL);
drivers/iommu/dma-iommu.c
1076
return vmap(sgt_handle(sgt)->pages, count, VM_MAP, PAGE_KERNEL);
drivers/md/dm-pcache/cache_dev.c
51
*vaddr = vmap(pages, total_pages, VM_MAP, PAGE_KERNEL);
drivers/md/dm-writecache.c
327
wc->memory_map = vmap(pages, p, VM_MAP, PAGE_KERNEL);
drivers/media/common/videobuf2/videobuf2-dma-contig.c
469
.vmap = vb2_dc_dmabuf_ops_vmap,
drivers/media/common/videobuf2/videobuf2-dma-sg.c
516
.vmap = vb2_dma_sg_dmabuf_ops_vmap,
drivers/media/common/videobuf2/videobuf2-vmalloc.c
331
.vmap = vb2_vmalloc_dmabuf_ops_vmap,
drivers/media/pci/intel/ipu6/ipu6-dma.c
209
info->vaddr = vmap(pages, count, VM_USERMAP, PAGE_KERNEL);
drivers/misc/fastrpc.c
770
.vmap = fastrpc_vmap,
drivers/misc/vmw_vmci/vmci_queue_pair.c
761
produce_q->q_header = vmap(headers, 2, VM_MAP, PAGE_KERNEL);
drivers/net/ethernet/google/gve/gve_tx.c
41
fifo->base = vmap(fifo->qpl->pages, fifo->qpl->num_entries, VM_MAP,
drivers/net/vrf.c
110
struct vrf_map vmap;
drivers/net/vrf.c
136
return &nn_vrf->vmap;
drivers/net/vrf.c
1814
static int vrf_map_init(struct vrf_map *vmap)
drivers/net/vrf.c
1816
spin_lock_init(&vmap->vmap_lock);
drivers/net/vrf.c
1817
hash_init(vmap->ht);
drivers/net/vrf.c
1819
vmap->strict_mode = false;
drivers/net/vrf.c
182
static struct vrf_map_elem *vrf_map_lookup_elem(struct vrf_map *vmap,
drivers/net/vrf.c
1825
static bool vrf_strict_mode(struct vrf_map *vmap)
drivers/net/vrf.c
1829
vrf_map_lock(vmap);
drivers/net/vrf.c
1830
strict_mode = vmap->strict_mode;
drivers/net/vrf.c
1831
vrf_map_unlock(vmap);
drivers/net/vrf.c
1836
static int vrf_strict_mode_change(struct vrf_map *vmap, bool new_mode)
drivers/net/vrf.c
1841
vrf_map_lock(vmap);
drivers/net/vrf.c
1843
cur_mode = &vmap->strict_mode;
drivers/net/vrf.c
1851
if (vmap->shared_tables) {
drivers/net/vrf.c
1866
vrf_map_unlock(vmap);
drivers/net/vrf.c
1875
struct vrf_map *vmap = netns_vrf_map(net);
drivers/net/vrf.c
1888
proc_strict_mode = vrf_strict_mode(vmap);
drivers/net/vrf.c
189
hash_for_each_possible(vmap->ht, me, hnode, key) {
drivers/net/vrf.c
1893
ret = vrf_strict_mode_change(vmap, (bool)proc_strict_mode);
drivers/net/vrf.c
1957
vrf_map_init(&nn_vrf->vmap);
drivers/net/vrf.c
197
static void vrf_map_add_elem(struct vrf_map *vmap, struct vrf_map_elem *me)
drivers/net/vrf.c
203
hash_add(vmap->ht, &me->hnode, key);
drivers/net/vrf.c
211
static void vrf_map_lock(struct vrf_map *vmap) __acquires(&vmap->vmap_lock)
drivers/net/vrf.c
213
spin_lock(&vmap->vmap_lock);
drivers/net/vrf.c
216
static void vrf_map_unlock(struct vrf_map *vmap) __releases(&vmap->vmap_lock)
drivers/net/vrf.c
218
spin_unlock(&vmap->vmap_lock);
drivers/net/vrf.c
225
struct vrf_map *vmap = netns_vrf_map_by_dev(dev);
drivers/net/vrf.c
242
vrf_map_lock(vmap);
drivers/net/vrf.c
244
me = vrf_map_lookup_elem(vmap, table_id);
drivers/net/vrf.c
247
vrf_map_add_elem(vmap, me);
drivers/net/vrf.c
255
if (vmap->strict_mode) {
drivers/net/vrf.c
265
++vmap->shared_tables;
drivers/net/vrf.c
272
vrf_map_unlock(vmap);
drivers/net/vrf.c
284
struct vrf_map *vmap = netns_vrf_map_by_dev(dev);
drivers/net/vrf.c
290
vrf_map_lock(vmap);
drivers/net/vrf.c
292
me = vrf_map_lookup_elem(vmap, table_id);
drivers/net/vrf.c
300
--vmap->shared_tables;
drivers/net/vrf.c
309
vrf_map_unlock(vmap);
drivers/net/vrf.c
315
struct vrf_map *vmap = netns_vrf_map(net);
drivers/net/vrf.c
319
vrf_map_lock(vmap);
drivers/net/vrf.c
321
if (!vmap->strict_mode) {
drivers/net/vrf.c
326
me = vrf_map_lookup_elem(vmap, table_id);
drivers/net/vrf.c
335
vrf_map_unlock(vmap);
drivers/perf/arm_spe_pmu.c
1030
buf->base = vmap(pglist, nr_pages, VM_MAP, PAGE_KERNEL);
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
912
bo->vmap_addr = vmap(bo->pages, bo->pgnr, VM_MAP,
drivers/staging/media/ipu3/ipu3-dmamap.c
127
map->vaddr = vmap(pages, count, VM_USERMAP, PAGE_KERNEL);
drivers/staging/media/ipu7/ipu7-dma.c
209
info->vaddr = vmap(pages, count, VM_USERMAP, PAGE_KERNEL);
drivers/target/target_core_transport.c
2752
cmd->t_data_vmap = vmap(pages, cmd->t_data_nents, VM_MAP, PAGE_KERNEL);
drivers/vdpa/alibaba/eni_vdpa.c
500
eni_vdpa->vdpa.vmap.dma_dev = &pdev->dev;
drivers/vdpa/ifcvf/ifcvf_main.c
717
adapter->vdpa.vmap.dma_dev = &pdev->dev;
drivers/vdpa/mlx5/core/mr.c
381
struct device *dma = mvdev->vdev.vmap.dma_dev;
drivers/vdpa/mlx5/core/mr.c
435
struct device *dma = mvdev->vdev.vmap.dma_dev;
drivers/vdpa/mlx5/net/mlx5_vnet.c
3415
map.dma_dev = mvdev->vdev.vmap.dma_dev;
drivers/vdpa/mlx5/net/mlx5_vnet.c
3978
mvdev->vdev.vmap.dma_dev = &mdev->pdev->dev;
drivers/vdpa/octeon_ep/octep_vdpa_main.c
519
oct_vdpa->vdpa.vmap.dma_dev = &pdev->dev;
drivers/vdpa/pds/vdpa_dev.c
647
pdsv->vdpa_dev.vmap.dma_dev = dma_dev;
drivers/vdpa/solidrun/snet_main.c
1056
snet->vdpa.vmap.dma_dev = &pdev->dev;
drivers/vdpa/vdpa_sim/vdpa_sim.c
273
vdpasim->vdpa.vmap.dma_dev = dev;
drivers/vdpa/virtio_pci/vp_vdpa.c
524
vp_vdpa->vdpa.vmap.dma_dev = &pdev->dev;
drivers/virt/acrn/mm.c
254
remap_vaddr = vmap(pages, nr_pages, VM_MAP, PAGE_KERNEL);
drivers/virt/vboxguest/vboxguest_core.c
93
guest_mappings[i] = vmap(pages, (size >> PAGE_SHIFT),
drivers/virtio/virtio_ring.c
366
vdev->map->max_mapping_size(vdev->vmap);
drivers/virtio/virtio_vdpa.c
198
vdev->vmap = map;
drivers/xen/xenbus/xenbus_client.c
701
addr = vmap(node->hvm.pages, nr_pages, VM_MAP | VM_IOREMAP,
drivers/xen/xlate_mmu.c
248
vaddr = vmap(pages, nr_pages, 0, PAGE_KERNEL);
fs/erofs/zutil.c
96
ptr = vmap(tmp_pages, nrpages, VM_MAP, PAGE_KERNEL);
fs/ntfs3/frecord.c
2252
frame_mem = vmap(pages, pages_per_frame, VM_MAP, PAGE_KERNEL);
fs/ntfs3/frecord.c
2535
frame_mem = vmap(pages, pages_per_frame, VM_MAP, PAGE_KERNEL_RO);
fs/pstore/ram_core.c
456
vaddr = vmap(pages, page_count, VM_MAP | VM_IOREMAP, prot);
include/drm/drm_gem.h
171
int (*vmap)(struct drm_gem_object *obj, struct iosys_map *map);
include/linux/dma-buf.h
278
int (*vmap)(struct dma_buf *dmabuf, struct iosys_map *map);
include/linux/mm.h
2963
struct vm_area_struct **vmap)
include/linux/mm.h
2983
*vmap = vma;
include/linux/vdpa.h
532
return vdev->vmap;
include/linux/vdpa.h
92
union virtio_map vmap;
include/linux/virtio.h
184
union virtio_map vmap;
include/linux/vmalloc.h
217
extern void *vmap(struct page **pages, unsigned int count,
io_uring/memmap.c
125
ptr = vmap(mr->pages, mr->nr_pages, VM_MAP, PAGE_KERNEL);
kernel/bpf/ringbuf.c
138
rb = vmap(pages, nr_meta_pages + 2 * nr_data_pages,
kernel/crash_core.c
78
safecopy = vmap(vmcoreinfo_pages, nr_pages, VM_MAP, PAGE_KERNEL);
kernel/dma/remap.c
29
vaddr = vmap(pages, PAGE_ALIGN(size) >> PAGE_SHIFT,
kernel/dma/remap.c
53
vaddr = vmap(pages, count, VM_DMA_COHERENT, prot);
kernel/kexec_core.c
1003
vaddr = vmap(src_pages, npages, VM_MAP, PAGE_KERNEL);
kernel/module/decompress.c
317
info->hdr = vmap(info->pages, info->used_pages, VM_MAP, PAGE_KERNEL);
kernel/relay.c
126
mem = vmap(buf->page_array, n_pages, VM_MAP, PAGE_KERNEL);
lib/tests/kunit_iov_iter.c
66
buffer = vmap(pages, npages, VM_MAP | VM_MAP_PUT_PAGES, PAGE_KERNEL);
mm/kasan/kasan_test_c.c
1935
v_ptr = vmap(&p_page, 1, VM_MAP, PAGE_KERNEL);
mm/khugepaged.c
892
bool expect_anon, struct vm_area_struct **vmap, struct collapse_control *cc)
mm/khugepaged.c
901
*vmap = vma = find_vma(mm, address);
mm/khugepaged.c
916
if (expect_anon && (!(*vmap)->anon_vma || !vma_is_anonymous(*vmap)))
mm/kmsan/kmsan_test.c
328
vbuf = vmap(pages, npages, VM_MAP, PAGE_KERNEL);
mm/nommu.c
313
EXPORT_SYMBOL(vmap);
mm/vma.c
1844
struct vm_area_struct *copy_vma(struct vm_area_struct **vmap,
mm/vma.c
1848
struct vm_area_struct *vma = *vmap;
mm/vma.c
1900
*vmap = vma = new_vma;
mm/vma.c
2506
static int __mmap_new_vma(struct mmap_state *map, struct vm_area_struct **vmap)
mm/vma.c
2563
*vmap = vma;
mm/vma.h
447
struct vm_area_struct *copy_vma(struct vm_area_struct **vmap,
mm/vma.h
691
int create_init_stack_vma(struct mm_struct *mm, struct vm_area_struct **vmap,
mm/vma_exec.c
107
int create_init_stack_vma(struct mm_struct *mm, struct vm_area_struct **vmap,
mm/vma_exec.c
152
*vmap = vma;
mm/vma_exec.c
161
*vmap = NULL;
mm/vmalloc.c
3572
EXPORT_SYMBOL(vmap);
net/xdp/xdp_umem.c
49
umem->addrs = vmap(pages, nr_pages, VM_MAP, PAGE_KERNEL);
sound/core/memalloc.c
763
p = vmap(sgbuf->pages, sgbuf->count, VM_MAP, PAGE_KERNEL);