Symbol: evict
arch/powerpc/kvm/book3s_32_mmu_host.c
141
bool evict = false;
arch/powerpc/kvm/book3s_32_mmu_host.c
171
evict = true;
arch/powerpc/kvm/book3s_32_mmu_host.c
178
if (!evict && (pteg[rr] & PTE_V)) {
drivers/gpu/drm/amd/amdgpu/amdgpu_object.c
1261
bool evict,
drivers/gpu/drm/amd/amdgpu/amdgpu_object.c
1271
amdgpu_vm_bo_move(abo, new_mem, evict);
drivers/gpu/drm/amd/amdgpu/amdgpu_object.h
297
bool evict,
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
383
bool evict,
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
430
r = ttm_bo_move_accel_cleanup(bo, fence, evict, true, new_mem);
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
498
static int amdgpu_bo_move(struct ttm_buffer_object *bo, bool evict,
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
520
amdgpu_bo_move_notify(bo, evict, new_mem);
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
527
amdgpu_bo_move_notify(bo, evict, new_mem);
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
539
amdgpu_bo_move_notify(bo, evict, new_mem);
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
556
amdgpu_bo_move_notify(bo, evict, new_mem);
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
582
amdgpu_bo_move_notify(bo, evict, new_mem);
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
584
r = amdgpu_move_blit(bo, evict, new_mem, old_mem);
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
602
if (evict)
drivers/gpu/drm/drm_gem.c
1723
if (obj->funcs->evict)
drivers/gpu/drm/drm_gem.c
1724
return obj->funcs->evict(obj);
drivers/gpu/drm/drm_gem_vram_helper.c
504
bool evict,
drivers/gpu/drm/drm_gem_vram_helper.c
767
bool evict,
drivers/gpu/drm/drm_gem_vram_helper.c
787
return drm_gem_vram_bo_driver_move(gbo, evict, ctx, new_mem);
drivers/gpu/drm/drm_gpuvm.c
1107
INIT_LIST_HEAD(&gpuvm->evict.list);
drivers/gpu/drm/drm_gpuvm.c
1108
spin_lock_init(&gpuvm->evict.lock);
drivers/gpu/drm/drm_gpuvm.c
1151
drm_WARN(gpuvm->drm, !list_empty(&gpuvm->evict.list),
drivers/gpu/drm/drm_gpuvm.c
1258
drm_gpuvm_bo_list_add(vm_bo, evict, false);
drivers/gpu/drm/drm_gpuvm.c
1470
LIST_HEAD(evict);
drivers/gpu/drm/drm_gpuvm.c
1473
for_each_vm_bo_in_list(gpuvm, evict, &evict, vm_bo) {
drivers/gpu/drm/drm_gpuvm.c
1480
restore_vm_bo_list(gpuvm, evict);
drivers/gpu/drm/drm_gpuvm.c
1494
list_for_each_entry_safe(vm_bo, next, &gpuvm->evict.list,
drivers/gpu/drm/drm_gpuvm.c
1495
list.entry.evict) {
drivers/gpu/drm/drm_gpuvm.c
1505
drm_gpuvm_bo_list_del_init(vm_bo, evict, false);
drivers/gpu/drm/drm_gpuvm.c
1598
INIT_LIST_HEAD(&vm_bo->list.entry.evict);
drivers/gpu/drm/drm_gpuvm.c
1653
drm_gpuvm_bo_list_del(vm_bo, evict, lock);
drivers/gpu/drm/drm_gpuvm.c
1705
drm_gpuvm_bo_list_del(vm_bo, evict, true);
drivers/gpu/drm/drm_gpuvm.c
1787
drm_gpuvm_bo_list_del(vm_bo, evict, false);
drivers/gpu/drm/drm_gpuvm.c
1958
drm_gpuvm_bo_evict(struct drm_gpuvm_bo *vm_bo, bool evict)
drivers/gpu/drm/drm_gpuvm.c
1965
vm_bo->evicted = evict;
drivers/gpu/drm/drm_gpuvm.c
1974
if (evict)
drivers/gpu/drm/drm_gpuvm.c
1975
drm_gpuvm_bo_list_add(vm_bo, evict, lock);
drivers/gpu/drm/drm_gpuvm.c
1977
drm_gpuvm_bo_list_del_init(vm_bo, evict, lock);
drivers/gpu/drm/i915/gem/i915_gem_ttm_move.c
570
int i915_ttm_move(struct ttm_buffer_object *bo, bool evict,
drivers/gpu/drm/i915/gem/i915_gem_ttm_move.c
660
if (I915_SELFTEST_ONLY(evict && fail_gpu_migration))
drivers/gpu/drm/i915/gem/i915_gem_ttm_move.c
663
ret = ttm_bo_move_accel_cleanup(bo, migration_fence, evict,
drivers/gpu/drm/i915/gem/i915_gem_ttm_move.h
33
int i915_ttm_move(struct ttm_buffer_object *bo, bool evict,
drivers/gpu/drm/i915/gt/selftest_hangcheck.c
1382
struct drm_mm_node evict = arg->vma->node;
drivers/gpu/drm/i915/gt/selftest_hangcheck.c
1388
err = i915_gem_evict_for_node(vm, NULL, &evict, 0);
drivers/gpu/drm/i915/selftests/i915_live_selftests.h
39
selftest(evict, i915_gem_evict_live_selftests)
drivers/gpu/drm/i915/selftests/i915_mock_selftests.h
32
selftest(evict, i915_gem_evict_mock_selftests)
drivers/gpu/drm/loongson/lsdc_ttm.c
160
bool evict,
drivers/gpu/drm/msm/msm_gem_shrinker.c
161
return evict(obj, ticket);
drivers/gpu/drm/msm/msm_gem_shrinker.c
178
{ &priv->lru.willneed, evict, can_swap() },
drivers/gpu/drm/nouveau/nouveau_bo.c
1144
nouveau_bo_move(struct ttm_buffer_object *bo, bool evict,
drivers/gpu/drm/nouveau/nouveau_bo.c
1162
drm_gpuvm_bo_gem_evict(obj, evict);
drivers/gpu/drm/nouveau/nouveau_bo.c
1207
ret = nouveau_bo_move_m2mf(bo, evict, ctx,
drivers/gpu/drm/nouveau/nouveau_bo.c
1227
drm_gpuvm_bo_gem_evict(obj, !evict);
drivers/gpu/drm/nouveau/nouveau_bo.c
548
bool force = false, evict = false;
drivers/gpu/drm/nouveau/nouveau_bo.c
558
evict = true;
drivers/gpu/drm/nouveau/nouveau_bo.c
563
bool error = evict;
drivers/gpu/drm/nouveau/nouveau_bo.c
586
if (evict) {
drivers/gpu/drm/nouveau/nouveau_bo.c
936
nouveau_bo_move_m2mf(struct ttm_buffer_object *bo, int evict,
drivers/gpu/drm/nouveau/nouveau_bo.c
982
ret = ttm_bo_move_accel_cleanup(bo, &fence->base, evict, false,
drivers/gpu/drm/qxl/qxl_ttm.c
137
static int qxl_bo_move(struct ttm_buffer_object *bo, bool evict,
drivers/gpu/drm/radeon/radeon_ttm.c
134
bool evict,
drivers/gpu/drm/radeon/radeon_ttm.c
183
r = ttm_bo_move_accel_cleanup(bo, &fence->base, evict, false, new_mem);
drivers/gpu/drm/radeon/radeon_ttm.c
188
static int radeon_bo_move(struct ttm_buffer_object *bo, bool evict,
drivers/gpu/drm/radeon/radeon_ttm.c
238
r = radeon_move_blit(bo, evict, new_mem, old_mem);
drivers/gpu/drm/ttm/tests/ttm_kunit_helpers.c
63
static int mock_move(struct ttm_buffer_object *bo, bool evict,
drivers/gpu/drm/ttm/ttm_bo.c
121
struct ttm_resource *mem, bool evict,
drivers/gpu/drm/ttm/ttm_bo.c
157
ret = bdev->funcs->move(bo, evict, ctx, mem, hop);
drivers/gpu/drm/ttm/ttm_bo_util.c
702
bool evict,
drivers/gpu/drm/ttm/ttm_bo_util.c
712
if (!evict)
drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c
497
bool evict,
drivers/gpu/drm/xe/xe_bo.c
838
static int xe_bo_move(struct ttm_buffer_object *ttm_bo, bool evict,
drivers/gpu/drm/xe/xe_bo.c
998
ret = ttm_bo_move_accel_cleanup(ttm_bo, fence, evict, true,
drivers/gpu/drm/xe/xe_vm.c
378
} while (!list_empty(&vm->gpuvm.evict.list));
drivers/infiniband/hw/hfi1/mmu_rb.c
244
if (handler->ops->evict(handler->ops_arg, rbnode, evict_arg,
drivers/infiniband/hw/hfi1/mmu_rb.h
27
int (*evict)(void *ops_arg, struct mmu_rb_node *mnode,
drivers/infiniband/hw/hfi1/pin_system.c
31
.evict = sdma_rb_evict,
fs/inode.c
1966
evict(inode);
fs/inode.c
888
evict(inode);
include/drm/drm_gem.h
206
int (*evict)(struct drm_gem_object *obj);
include/drm/drm_gpuvm.h
335
} evict;
include/drm/drm_gpuvm.h
723
struct list_head evict;
include/drm/drm_gpuvm.h
769
void drm_gpuvm_bo_evict(struct drm_gpuvm_bo *vm_bo, bool evict);
include/drm/drm_gpuvm.h
780
drm_gpuvm_bo_gem_evict(struct drm_gem_object *obj, bool evict)
include/drm/drm_gpuvm.h
786
drm_gpuvm_bo_evict(vm_bo, evict);
include/drm/ttm/ttm_bo.h
458
struct dma_fence *fence, bool evict,
include/drm/ttm/ttm_device.h
150
int (*move)(struct ttm_buffer_object *bo, bool evict,