vmap
if (alter && vmap[v].is_const) {
s->k = vmap[v].const_val;
free((void *)vmap);
vmap = reallocarray(NULL, maxval, sizeof(*vmap));
if (vmap == NULL || vnode_base == NULL)
struct vmapinfo *vmap;
memset((char *)vmap, 0, maxval * sizeof(*vmap));
vmap[val].const_val = v0;
vmap[val].is_const = 1;
a = vmap[v0].const_val;
b = vmap[v1].const_val;
if (vmap[val].is_const) {
b->s.k += vmap[val].const_val;
if (vmap[val].is_const && BPF_SRC(b->s.code) == BPF_K) {
bpf_int32 v = vmap[val].const_val;
if (alter && vmap[v].is_const) {
s->k += vmap[v].const_val;
if (alter && vmap[val[A_ATOM]].is_const) {
s->k = -vmap[val[A_ATOM]].const_val;
if (vmap[val[A_ATOM]].is_const) {
if (alter && vmap[val[X_ATOM]].is_const) {
if (vmap[val[A_ATOM]].is_const) {
s->k = vmap[val[X_ATOM]].const_val;
if (alter && vmap[val[A_ATOM]].is_const
&& vmap[val[A_ATOM]].const_val == 0) {
if (alter && vmap[v].is_const) {
s->k = vmap[v].const_val;
.vmap = amdgpu_dma_buf_vmap,
.vmap = drm_gem_ttm_vmap,
if (!obj->funcs->vmap)
ret = obj->funcs->vmap(obj, map);
.vmap = drm_gem_dma_vmap,
.vmap = drm_gem_dmabuf_vmap,
.vmap = i915_gem_dmabuf_vmap,
.vmap = i915_gem_vmap_object,
vaddr = vmap(pages, n_pages, 0, pgprot);
.vmap = mock_dmabuf_vmap,
vaddr = vmap(pages, n_pages, VM_MAP_PUT_PAGES, PAGE_KERNEL);
int (*vmap)(struct drm_gem_object *obj, struct iosys_map *map);
void *vmap(struct vm_page **, unsigned int, unsigned long, pgprot_t);
.vmap = drm_gem_ttm_vmap,
map->virtual = vmap(ttm->pages + start_page, num_pages,
vaddr = vmap(ttm->pages, ttm->num_pages, 0, prot);