arch/alpha/boot/tools/objstrip.c
198
aout = (struct exec *) buf;
arch/alpha/boot/tools/objstrip.c
58
struct exec * aout; /* includes file & aout header */
arch/alpha/include/uapi/asm/a.out.h
90
(sizeof(struct exec) + (x).fh.f_nscns*SCNHSZ + SCNROUND - 1) & ~(SCNROUND - 1))
arch/arc/mm/fault.c
107
exec = 1;
arch/arc/mm/fault.c
127
if (exec)
arch/arc/mm/fault.c
79
unsigned int write = 0, exec = 0, mask;
arch/mips/include/asm/sgiarcs.h
192
LONG exec; /* Load and begin execution of a
arch/mips/kernel/smp.c
632
int exec = vma->vm_flags & VM_EXEC;
arch/mips/kernel/smp.c
642
set_cpu_context(cpu, mm, !exec);
arch/mips/mm/c-r3k.c
240
int exec = vma->vm_flags & VM_EXEC;
arch/mips/mm/c-r3k.c
260
if (exec)
arch/mips/mm/c-r4k.c
473
int exec = vma->vm_flags & VM_EXEC;
arch/mips/mm/c-r4k.c
483
if (cpu_has_dc_aliases || (exec && !cpu_has_ic_fills_f_dc))
arch/mips/mm/c-r4k.c
486
if (exec)
arch/mips/mm/c-r4k.c
493
int exec = vma->vm_flags & VM_EXEC;
arch/mips/mm/c-r4k.c
495
if (cpu_has_dc_aliases || exec)
arch/mips/mm/c-r4k.c
543
int exec = vma->vm_flags & VM_EXEC;
arch/mips/mm/c-r4k.c
586
if (cpu_has_dc_aliases || (exec && !cpu_has_ic_fills_f_dc)) {
arch/mips/mm/c-r4k.c
589
if (exec && !cpu_icache_snoops_remote_store)
arch/mips/mm/c-r4k.c
592
if (exec) {
arch/mips/mm/cache.c
150
int exec = !pte_no_exec(pte) && !cpu_has_ic_fills_f_dc;
arch/mips/mm/cache.c
165
if (exec || pages_do_alias(addr, address))
arch/parisc/include/asm/processor.h
245
elf_caddr_t *argv = (elf_caddr_t *)bprm->exec + 1; \
arch/powerpc/mm/book3s64/radix_pgtable.c
259
print_mapping(unsigned long start, unsigned long end, unsigned long size, bool exec)
arch/powerpc/mm/book3s64/radix_pgtable.c
269
exec ? " (exec)" : "");
arch/powerpc/mm/book3s64/radix_pgtable.c
303
bool prev_exec, exec = false;
arch/powerpc/mm/book3s64/radix_pgtable.c
324
prev_exec = exec;
arch/powerpc/mm/book3s64/radix_pgtable.c
344
exec = true;
arch/powerpc/mm/book3s64/radix_pgtable.c
347
exec = false;
arch/powerpc/mm/book3s64/radix_pgtable.c
350
if (mapping_size != previous_size || exec != prev_exec) {
arch/powerpc/mm/book3s64/radix_pgtable.c
362
print_mapping(start, addr, mapping_size, exec);
arch/sparc/include/uapi/asm/oradax.h
41
struct ccb_exec_result exec;
arch/sparc/mm/tlb.c
168
bool exec = pte_exec(*pte);
arch/sparc/mm/tlb.c
170
tlb_batch_add_one(mm, vaddr, exec, PAGE_SHIFT);
arch/sparc/mm/tlb.c
220
bool exec = pte_exec(orig_pte);
arch/sparc/mm/tlb.c
222
tlb_batch_add_one(mm, addr, exec, REAL_HPAGE_SHIFT);
arch/sparc/mm/tlb.c
223
tlb_batch_add_one(mm, addr + REAL_HPAGE_SIZE, exec,
arch/sparc/mm/tlb.c
79
bool exec, unsigned int hugepage_shift)
arch/sparc/mm/tlb.c
85
if (exec)
arch/x86/include/asm/elf.h
351
if (exec->e_machine == EM_X86_64) \
arch/x86/kvm/mmu/mmu.c
3385
fault->huge_page_disallowed = fault->exec && fault->nx_huge_page_workaround_enabled;
arch/x86/kvm/mmu/mmu_internal.h
233
const bool exec;
arch/x86/kvm/mmu/mmu_internal.h
336
PAGE_SIZE, fault->write, fault->exec,
arch/x86/kvm/mmu/mmu_internal.h
347
.exec = err & PFERR_FETCH_MASK,
arch/x86/kvm/mmu/spte.h
495
if (fault->exec)
arch/x86/kvm/vmx/vmcs.h
60
u32 exec;
arch/x86/kvm/vmx/vmx.h
617
BUILD_CONTROLS_SHADOW(exec, CPU_BASED_VM_EXEC_CONTROL, 32)
arch/x86/mm/pat/cpa-test.c
45
long lpg, gpg, spg, exec;
arch/x86/mm/pat/cpa-test.c
54
s->lpg = s->gpg = s->spg = s->exec = 0;
arch/x86/mm/pat/cpa-test.c
86
s->exec++;
arch/x86/mm/pat/cpa-test.c
96
s->spg, s->lpg, s->gpg, s->exec,
drivers/accel/qaic/qaic_data.c
1200
struct qaic_execute_entry *exec, unsigned int count,
drivers/accel/qaic/qaic_data.c
1204
struct qaic_partial_execute_entry *pexec = (struct qaic_partial_execute_entry *)exec;
drivers/accel/qaic/qaic_data.c
1218
is_partial ? pexec[i].handle : exec[i].handle);
drivers/accel/qaic/qaic_data.c
1294
struct qaic_execute_entry *exec, unsigned int count,
drivers/accel/qaic/qaic_data.c
1297
struct qaic_partial_execute_entry *pexec = (struct qaic_partial_execute_entry *)exec;
drivers/accel/qaic/qaic_data.c
1311
is_partial ? pexec[i].handle : exec[i].handle);
drivers/accel/qaic/qaic_data.c
1327
struct qaic_execute_entry *exec;
drivers/accel/qaic/qaic_data.c
1343
size = is_partial ? sizeof(struct qaic_partial_execute_entry) : sizeof(*exec);
drivers/accel/qaic/qaic_data.c
1347
exec = memdup_array_user(u64_to_user_ptr(args->data), args->hdr.count, size);
drivers/accel/qaic/qaic_data.c
1348
if (IS_ERR(exec))
drivers/accel/qaic/qaic_data.c
1349
return PTR_ERR(exec);
drivers/accel/qaic/qaic_data.c
1398
ret = send_bo_list_to_device(qdev, file_priv, exec, args->hdr.count, is_partial, dbc,
drivers/accel/qaic/qaic_data.c
1408
update_profiling_data(file_priv, exec, args->hdr.count, is_partial, received_ts,
drivers/accel/qaic/qaic_data.c
1423
kfree(exec);
drivers/bus/mhi/host/main.c
155
u32 exec;
drivers/bus/mhi/host/main.c
156
int ret = mhi_read_reg(mhi_cntrl, mhi_cntrl->bhi, BHI_EXECENV, &exec);
drivers/bus/mhi/host/main.c
158
return (ret) ? MHI_EE_MAX : exec;
drivers/connector/cn_proc.c
161
ev->event_data.exec.process_pid = task->pid;
drivers/connector/cn_proc.c
162
ev->event_data.exec.process_tgid = task->tgid;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1146
struct drm_exec exec;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1176
drm_exec_init(&ctx->exec, DRM_EXEC_INTERRUPTIBLE_WAIT, 0);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1177
drm_exec_until_all_locked(&ctx->exec) {
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1178
ret = amdgpu_vm_lock_pd(vm, &ctx->exec, 2);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1179
drm_exec_retry_on_contention(&ctx->exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1183
ret = drm_exec_prepare_obj(&ctx->exec, &bo->tbo.base, 1);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1184
drm_exec_retry_on_contention(&ctx->exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1192
drm_exec_fini(&ctx->exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1215
drm_exec_init(&ctx->exec, DRM_EXEC_INTERRUPTIBLE_WAIT |
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1217
drm_exec_until_all_locked(&ctx->exec) {
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1226
&ctx->exec, 2);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1227
drm_exec_retry_on_contention(&ctx->exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1233
ret = drm_exec_prepare_obj(&ctx->exec, &bo->tbo.base, 1);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1234
drm_exec_retry_on_contention(&ctx->exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1242
drm_exec_fini(&ctx->exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1264
drm_exec_fini(&ctx->exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2665
struct drm_exec exec;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2674
drm_exec_init(&exec, 0, 0);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2676
drm_exec_until_all_locked(&exec) {
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2680
ret = amdgpu_vm_lock_pd(peer_vm, &exec, 2);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2681
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2692
ret = drm_exec_prepare_obj(&exec, gobj, 1);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2693
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2748
drm_exec_fini(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2925
struct drm_exec exec;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2932
drm_exec_init(&exec, DRM_EXEC_IGNORE_DUPLICATES, 0);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2933
drm_exec_until_all_locked(&exec) {
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2936
ret = amdgpu_vm_lock_pd(peer_vm, &exec, 2);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2937
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2952
ret = drm_exec_prepare_obj(&exec, gobj, 1);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2953
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
3001
ret = process_validate_vms(process_info, &exec.ticket);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
3042
ret = amdgpu_vm_handle_moved(adev, peer_vm, &exec.ticket);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
3122
drm_exec_fini(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
881
struct drm_exec exec;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
962
drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT, 0);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
963
drm_exec_until_all_locked(&exec) {
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
964
ret = amdgpu_vm_lock_pd(vm, &exec, 0);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
965
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
968
ret = drm_exec_lock_obj(&exec, &bo[i]->tbo.base);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
969
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
980
drm_exec_fini(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
1168
r = amdgpu_vm_handle_moved(adev, vm, &p->exec.ticket);
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
1222
drm_exec_for_each_locked_object(&p->exec, index, obj) {
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
1340
drm_exec_for_each_locked_object(&p->exec, index, gobj) {
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
1371
amdgpu_vm_bo_trace_cs(&fpriv->vm, &p->exec.ticket);
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
1392
drm_exec_fini(&parser->exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
1799
if (dma_resv_locking_ctx((*bo)->tbo.base.resv) != &parser->exec.ticket)
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
69
drm_exec_init(&p->exec, DRM_EXEC_INTERRUPTIBLE_WAIT |
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
911
drm_exec_until_all_locked(&p->exec) {
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
912
r = amdgpu_vm_lock_pd(&fpriv->vm, &p->exec, 1 + p->gang_size);
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
913
drm_exec_retry_on_contention(&p->exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
919
r = drm_exec_prepare_obj(&p->exec, &e->bo->tbo.base,
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
921
drm_exec_retry_on_contention(&p->exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
929
r = drm_exec_prepare_obj(&p->exec, &p->uf_bo->tbo.base,
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
931
drm_exec_retry_on_contention(&p->exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
972
drm_exec_for_each_locked_object(&p->exec, index, obj) {
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.h
66
struct drm_exec exec;
drivers/gpu/drm/amd/amdgpu/amdgpu_csa.c
103
drm_exec_fini(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_csa.c
111
struct drm_exec exec;
drivers/gpu/drm/amd/amdgpu/amdgpu_csa.c
114
drm_exec_init(&exec, 0, 0);
drivers/gpu/drm/amd/amdgpu/amdgpu_csa.c
115
drm_exec_until_all_locked(&exec) {
drivers/gpu/drm/amd/amdgpu/amdgpu_csa.c
116
r = amdgpu_vm_lock_pd(vm, &exec, 0);
drivers/gpu/drm/amd/amdgpu/amdgpu_csa.c
118
r = drm_exec_lock_obj(&exec, &bo->tbo.base);
drivers/gpu/drm/amd/amdgpu/amdgpu_csa.c
119
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_csa.c
137
drm_exec_fini(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_csa.c
69
struct drm_exec exec;
drivers/gpu/drm/amd/amdgpu/amdgpu_csa.c
72
drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT, 0);
drivers/gpu/drm/amd/amdgpu/amdgpu_csa.c
73
drm_exec_until_all_locked(&exec) {
drivers/gpu/drm/amd/amdgpu/amdgpu_csa.c
74
r = amdgpu_vm_lock_pd(vm, &exec, 0);
drivers/gpu/drm/amd/amdgpu/amdgpu_csa.c
76
r = drm_exec_lock_obj(&exec, &bo->tbo.base);
drivers/gpu/drm/amd/amdgpu/amdgpu_csa.c
77
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_eviction_fence.c
48
struct drm_exec *exec)
drivers/gpu/drm/amd/amdgpu/amdgpu_eviction_fence.c
80
drm_exec_for_each_locked_object(exec, index, obj) {
drivers/gpu/drm/amd/amdgpu/amdgpu_eviction_fence.h
68
struct drm_exec *exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
1001
drm_exec_fini(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
1014
struct drm_exec exec;
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
1027
drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT |
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
1029
drm_exec_until_all_locked(&exec) {
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
1030
r = drm_exec_lock_obj(&exec, gobj);
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
1031
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
1036
r = amdgpu_vm_lock_pd(&fpriv->vm, &exec, 0);
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
1037
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
1052
drm_exec_fini(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
1084
drm_exec_fini(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
1122
drm_exec_fini(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
1134
drm_exec_fini(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
1141
drm_exec_fini(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
235
struct drm_exec exec;
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
246
drm_exec_init(&exec, DRM_EXEC_IGNORE_DUPLICATES, 0);
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
247
drm_exec_until_all_locked(&exec) {
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
248
r = drm_exec_prepare_obj(&exec, &abo->tbo.base, 1);
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
249
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
253
r = amdgpu_vm_lock_pd(vm, &exec, 0);
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
254
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
273
drm_exec_fini(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
309
drm_exec_fini(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
323
struct drm_exec exec;
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
326
drm_exec_init(&exec, DRM_EXEC_IGNORE_DUPLICATES, 0);
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
327
drm_exec_until_all_locked(&exec) {
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
328
r = drm_exec_prepare_obj(&exec, &bo->tbo.base, 1);
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
329
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
333
r = amdgpu_vm_lock_pd(vm, &exec, 0);
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
334
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
364
drm_exec_fini(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
839
struct drm_exec exec;
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
908
drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT |
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
910
drm_exec_until_all_locked(&exec) {
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
912
r = drm_exec_lock_obj(&exec, gobj);
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
913
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
918
r = amdgpu_vm_lock_pd(&fpriv->vm, &exec, 2);
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
919
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
1448
struct drm_exec exec;
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
1488
drm_exec_init(&exec, DRM_EXEC_IGNORE_DUPLICATES, 0);
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
1489
drm_exec_until_all_locked(&exec) {
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
1490
r = amdgpu_vm_lock_pd(&fpriv->vm, &exec, 0);
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
1491
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
1497
drm_exec_fini(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_seq64.c
114
drm_exec_fini(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_seq64.c
130
struct drm_exec exec;
drivers/gpu/drm/amd/amdgpu/amdgpu_seq64.c
142
drm_exec_init(&exec, 0, 0);
drivers/gpu/drm/amd/amdgpu/amdgpu_seq64.c
143
drm_exec_until_all_locked(&exec) {
drivers/gpu/drm/amd/amdgpu/amdgpu_seq64.c
144
r = amdgpu_vm_lock_pd(vm, &exec, 0);
drivers/gpu/drm/amd/amdgpu/amdgpu_seq64.c
146
r = drm_exec_lock_obj(&exec, &bo->tbo.base);
drivers/gpu/drm/amd/amdgpu/amdgpu_seq64.c
147
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_seq64.c
157
drm_exec_fini(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_seq64.c
71
struct drm_exec exec;
drivers/gpu/drm/amd/amdgpu/amdgpu_seq64.c
79
drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT, 0);
drivers/gpu/drm/amd/amdgpu/amdgpu_seq64.c
80
drm_exec_until_all_locked(&exec) {
drivers/gpu/drm/amd/amdgpu/amdgpu_seq64.c
81
r = amdgpu_vm_lock_pd(vm, &exec, 0);
drivers/gpu/drm/amd/amdgpu/amdgpu_seq64.c
83
r = drm_exec_lock_obj(&exec, &bo->tbo.base);
drivers/gpu/drm/amd/amdgpu/amdgpu_seq64.c
84
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1091
amdgpu_userq_bo_validate(struct amdgpu_device *adev, struct drm_exec *exec,
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1107
ret = drm_exec_prepare_obj(exec, &bo->tbo.base, 2);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1141
struct drm_exec exec;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1148
drm_exec_init(&exec, DRM_EXEC_IGNORE_DUPLICATES, 0);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1149
drm_exec_until_all_locked(&exec) {
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1150
ret = amdgpu_vm_lock_pd(vm, &exec, 1);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1151
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1155
ret = amdgpu_vm_lock_done_list(vm, &exec, 1);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1156
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1168
ret = amdgpu_userq_bo_validate(adev, &exec, vm);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1169
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1227
drm_exec_fini(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1255
ret = amdgpu_eviction_fence_replace_fence(&fpriv->evf_mgr, &exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1260
drm_exec_fini(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
1000
drm_exec_fini(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
477
struct drm_exec exec;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
586
drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT,
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
590
drm_exec_until_all_locked(&exec) {
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
591
r = drm_exec_prepare_array(&exec, gobj_read, num_read_bo_handles, 1);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
592
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
598
r = drm_exec_prepare_array(&exec, gobj_write, num_write_bo_handles, 1);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
599
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
630
drm_exec_fini(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
672
struct drm_exec exec;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
747
drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT,
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
751
drm_exec_until_all_locked(&exec) {
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
752
r = drm_exec_prepare_array(&exec, gobj_read, num_read_bo_handles, 1);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
753
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
755
drm_exec_fini(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
759
r = drm_exec_prepare_array(&exec, gobj_write, num_write_bo_handles, 1);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
760
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
762
drm_exec_fini(&exec);
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
459
int amdgpu_vm_lock_pd(struct amdgpu_vm *vm, struct drm_exec *exec,
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
463
return drm_exec_prepare_obj(exec, &vm->root.bo->tbo.base,
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
475
int amdgpu_vm_lock_done_list(struct amdgpu_vm *vm, struct drm_exec *exec,
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
493
ret = drm_exec_prepare_obj(exec, &bo->tbo.base, 1);
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
508
int amdgpu_vm_lock_pd(struct amdgpu_vm *vm, struct drm_exec *exec,
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
510
int amdgpu_vm_lock_done_list(struct amdgpu_vm *vm, struct drm_exec *exec,
drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c
2118
bool exec = flags & AMDGPU_FENCE_FLAG_EXEC;
drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c
2140
(exec ? EOP_EXEC : 0)));
drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c
6140
bool exec = flags & AMDGPU_FENCE_FLAG_EXEC;
drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c
6165
(exec ? EOP_EXEC : 0)));
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
5542
bool exec = flags & AMDGPU_FENCE_FLAG_EXEC;
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
5556
if (exec)
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1571
struct drm_exec exec;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1581
drm_exec_init(&ctx->exec, intr ? DRM_EXEC_INTERRUPTIBLE_WAIT: 0, 0);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1582
drm_exec_until_all_locked(&ctx->exec) {
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1592
r = amdgpu_vm_lock_pd(vm, &ctx->exec, 2);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1593
drm_exec_retry_on_contention(&ctx->exec);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1621
drm_exec_fini(&ctx->exec);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1627
drm_exec_fini(&ctx->exec);
drivers/gpu/drm/drm_exec.c
104
void drm_exec_fini(struct drm_exec *exec)
drivers/gpu/drm/drm_exec.c
106
drm_exec_unlock_all(exec);
drivers/gpu/drm/drm_exec.c
107
kvfree(exec->objects);
drivers/gpu/drm/drm_exec.c
108
if (exec->contended != DRM_EXEC_DUMMY) {
drivers/gpu/drm/drm_exec.c
109
drm_gem_object_put(exec->contended);
drivers/gpu/drm/drm_exec.c
110
ww_acquire_fini(&exec->ticket);
drivers/gpu/drm/drm_exec.c
123
bool drm_exec_cleanup(struct drm_exec *exec)
drivers/gpu/drm/drm_exec.c
125
if (likely(!exec->contended)) {
drivers/gpu/drm/drm_exec.c
126
ww_acquire_done(&exec->ticket);
drivers/gpu/drm/drm_exec.c
130
if (likely(exec->contended == DRM_EXEC_DUMMY)) {
drivers/gpu/drm/drm_exec.c
131
exec->contended = NULL;
drivers/gpu/drm/drm_exec.c
132
ww_acquire_init(&exec->ticket, &reservation_ww_class);
drivers/gpu/drm/drm_exec.c
136
drm_exec_unlock_all(exec);
drivers/gpu/drm/drm_exec.c
137
exec->num_objects = 0;
drivers/gpu/drm/drm_exec.c
143
static int drm_exec_obj_locked(struct drm_exec *exec,
drivers/gpu/drm/drm_exec.c
146
if (unlikely(exec->num_objects == exec->max_objects)) {
drivers/gpu/drm/drm_exec.c
147
size_t size = exec->max_objects * sizeof(void *);
drivers/gpu/drm/drm_exec.c
150
tmp = kvrealloc(exec->objects, size + PAGE_SIZE, GFP_KERNEL);
drivers/gpu/drm/drm_exec.c
154
exec->objects = tmp;
drivers/gpu/drm/drm_exec.c
155
exec->max_objects += PAGE_SIZE / sizeof(void *);
drivers/gpu/drm/drm_exec.c
158
exec->objects[exec->num_objects++] = obj;
drivers/gpu/drm/drm_exec.c
164
static int drm_exec_lock_contended(struct drm_exec *exec)
drivers/gpu/drm/drm_exec.c
166
struct drm_gem_object *obj = exec->contended;
drivers/gpu/drm/drm_exec.c
173
exec->contended = NULL;
drivers/gpu/drm/drm_exec.c
174
if (exec->flags & DRM_EXEC_INTERRUPTIBLE_WAIT) {
drivers/gpu/drm/drm_exec.c
176
&exec->ticket);
drivers/gpu/drm/drm_exec.c
180
dma_resv_lock_slow(obj->resv, &exec->ticket);
drivers/gpu/drm/drm_exec.c
183
ret = drm_exec_obj_locked(exec, obj);
drivers/gpu/drm/drm_exec.c
187
exec->prelocked = obj;
drivers/gpu/drm/drm_exec.c
209
int drm_exec_lock_obj(struct drm_exec *exec, struct drm_gem_object *obj)
drivers/gpu/drm/drm_exec.c
213
ret = drm_exec_lock_contended(exec);
drivers/gpu/drm/drm_exec.c
217
if (exec->prelocked == obj) {
drivers/gpu/drm/drm_exec.c
218
drm_gem_object_put(exec->prelocked);
drivers/gpu/drm/drm_exec.c
219
exec->prelocked = NULL;
drivers/gpu/drm/drm_exec.c
223
if (exec->flags & DRM_EXEC_INTERRUPTIBLE_WAIT)
drivers/gpu/drm/drm_exec.c
224
ret = dma_resv_lock_interruptible(obj->resv, &exec->ticket);
drivers/gpu/drm/drm_exec.c
226
ret = dma_resv_lock(obj->resv, &exec->ticket);
drivers/gpu/drm/drm_exec.c
230
exec->contended = obj;
drivers/gpu/drm/drm_exec.c
235
exec->flags & DRM_EXEC_IGNORE_DUPLICATES)
drivers/gpu/drm/drm_exec.c
241
ret = drm_exec_obj_locked(exec, obj);
drivers/gpu/drm/drm_exec.c
262
void drm_exec_unlock_obj(struct drm_exec *exec, struct drm_gem_object *obj)
drivers/gpu/drm/drm_exec.c
266
for (i = exec->num_objects; i--;) {
drivers/gpu/drm/drm_exec.c
267
if (exec->objects[i] == obj) {
drivers/gpu/drm/drm_exec.c
269
for (++i; i < exec->num_objects; ++i)
drivers/gpu/drm/drm_exec.c
270
exec->objects[i - 1] = exec->objects[i];
drivers/gpu/drm/drm_exec.c
271
--exec->num_objects;
drivers/gpu/drm/drm_exec.c
291
int drm_exec_prepare_obj(struct drm_exec *exec, struct drm_gem_object *obj,
drivers/gpu/drm/drm_exec.c
296
ret = drm_exec_lock_obj(exec, obj);
drivers/gpu/drm/drm_exec.c
302
drm_exec_unlock_obj(exec, obj);
drivers/gpu/drm/drm_exec.c
323
int drm_exec_prepare_array(struct drm_exec *exec,
drivers/gpu/drm/drm_exec.c
331
ret = drm_exec_prepare_obj(exec, objects[i], num_fences);
drivers/gpu/drm/drm_exec.c
56
static void drm_exec_unlock_all(struct drm_exec *exec)
drivers/gpu/drm/drm_exec.c
61
drm_exec_for_each_locked_object_reverse(exec, index, obj) {
drivers/gpu/drm/drm_exec.c
66
drm_gem_object_put(exec->prelocked);
drivers/gpu/drm/drm_exec.c
67
exec->prelocked = NULL;
drivers/gpu/drm/drm_exec.c
81
void drm_exec_init(struct drm_exec *exec, u32 flags, unsigned nr)
drivers/gpu/drm/drm_exec.c
86
exec->flags = flags;
drivers/gpu/drm/drm_exec.c
87
exec->objects = kvmalloc_array(nr, sizeof(void *), GFP_KERNEL);
drivers/gpu/drm/drm_exec.c
90
exec->max_objects = exec->objects ? nr : 0;
drivers/gpu/drm/drm_exec.c
91
exec->num_objects = 0;
drivers/gpu/drm/drm_exec.c
92
exec->contended = DRM_EXEC_DUMMY;
drivers/gpu/drm/drm_exec.c
93
exec->prelocked = NULL;
drivers/gpu/drm/drm_gpuvm.c
1189
exec_prepare_obj(struct drm_exec *exec, struct drm_gem_object *obj,
drivers/gpu/drm/drm_gpuvm.c
1192
return num_fences ? drm_exec_prepare_obj(exec, obj, num_fences) :
drivers/gpu/drm/drm_gpuvm.c
1193
drm_exec_lock_obj(exec, obj);
drivers/gpu/drm/drm_gpuvm.c
1212
struct drm_exec *exec,
drivers/gpu/drm/drm_gpuvm.c
1215
return exec_prepare_obj(exec, gpuvm->r_obj, num_fences);
drivers/gpu/drm/drm_gpuvm.c
1221
struct drm_exec *exec,
drivers/gpu/drm/drm_gpuvm.c
1229
ret = exec_prepare_obj(exec, vm_bo->obj, num_fences);
drivers/gpu/drm/drm_gpuvm.c
1242
struct drm_exec *exec,
drivers/gpu/drm/drm_gpuvm.c
1253
ret = exec_prepare_obj(exec, vm_bo->obj, num_fences);
drivers/gpu/drm/drm_gpuvm.c
1289
struct drm_exec *exec,
drivers/gpu/drm/drm_gpuvm.c
1293
return drm_gpuvm_prepare_objects_locked(gpuvm, exec,
drivers/gpu/drm/drm_gpuvm.c
1296
return __drm_gpuvm_prepare_objects(gpuvm, exec, num_fences);
drivers/gpu/drm/drm_gpuvm.c
1315
drm_gpuvm_prepare_range(struct drm_gpuvm *gpuvm, struct drm_exec *exec,
drivers/gpu/drm/drm_gpuvm.c
1325
ret = exec_prepare_obj(exec, obj, num_fences);
drivers/gpu/drm/drm_gpuvm.c
1352
struct drm_exec *exec = &vm_exec->exec;
drivers/gpu/drm/drm_gpuvm.c
1356
drm_exec_init(exec, vm_exec->flags, 0);
drivers/gpu/drm/drm_gpuvm.c
1358
drm_exec_until_all_locked(exec) {
drivers/gpu/drm/drm_gpuvm.c
1359
ret = drm_gpuvm_prepare_vm(gpuvm, exec, num_fences);
drivers/gpu/drm/drm_gpuvm.c
1360
drm_exec_retry_on_contention(exec);
drivers/gpu/drm/drm_gpuvm.c
1364
ret = drm_gpuvm_prepare_objects(gpuvm, exec, num_fences);
drivers/gpu/drm/drm_gpuvm.c
1365
drm_exec_retry_on_contention(exec);
drivers/gpu/drm/drm_gpuvm.c
1371
drm_exec_retry_on_contention(exec);
drivers/gpu/drm/drm_gpuvm.c
1380
drm_exec_fini(exec);
drivers/gpu/drm/drm_gpuvm.c
1393
return drm_exec_prepare_array(&vm_exec->exec, args->objs,
drivers/gpu/drm/drm_gpuvm.c
1444
struct drm_exec *exec = &vm_exec->exec;
drivers/gpu/drm/drm_gpuvm.c
1447
drm_exec_init(exec, vm_exec->flags, 0);
drivers/gpu/drm/drm_gpuvm.c
1449
drm_exec_until_all_locked(exec) {
drivers/gpu/drm/drm_gpuvm.c
1450
ret = drm_gpuvm_prepare_range(gpuvm, exec, addr, range,
drivers/gpu/drm/drm_gpuvm.c
1452
drm_exec_retry_on_contention(exec);
drivers/gpu/drm/drm_gpuvm.c
1460
drm_exec_fini(exec);
drivers/gpu/drm/drm_gpuvm.c
1466
__drm_gpuvm_validate(struct drm_gpuvm *gpuvm, struct drm_exec *exec)
drivers/gpu/drm/drm_gpuvm.c
1474
ret = ops->vm_bo_validate(vm_bo, exec);
drivers/gpu/drm/drm_gpuvm.c
1486
drm_gpuvm_validate_locked(struct drm_gpuvm *gpuvm, struct drm_exec *exec)
drivers/gpu/drm/drm_gpuvm.c
1499
ret = ops->vm_bo_validate(vm_bo, exec);
drivers/gpu/drm/drm_gpuvm.c
1522
drm_gpuvm_validate(struct drm_gpuvm *gpuvm, struct drm_exec *exec)
drivers/gpu/drm/drm_gpuvm.c
1530
return drm_gpuvm_validate_locked(gpuvm, exec);
drivers/gpu/drm/drm_gpuvm.c
1532
return __drm_gpuvm_validate(gpuvm, exec);
drivers/gpu/drm/drm_gpuvm.c
1547
struct drm_exec *exec,
drivers/gpu/drm/drm_gpuvm.c
1555
drm_exec_for_each_locked_object(exec, index, obj) {
drivers/gpu/drm/drm_gpuvm.c
2729
struct drm_exec *exec = priv;
drivers/gpu/drm/drm_gpuvm.c
2734
return drm_exec_lock_obj(exec, op->remap.unmap->va->gem.obj);
drivers/gpu/drm/drm_gpuvm.c
2738
return drm_exec_lock_obj(exec, op->unmap.va->gem.obj);
drivers/gpu/drm/drm_gpuvm.c
2807
struct drm_exec *exec, unsigned int num_fences,
drivers/gpu/drm/drm_gpuvm.c
2813
int ret = drm_exec_prepare_obj(exec, req_obj, num_fences);
drivers/gpu/drm/drm_gpuvm.c
2818
return __drm_gpuvm_sm_map(gpuvm, &lock_ops, exec, req, false);
drivers/gpu/drm/drm_gpuvm.c
2838
drm_gpuvm_sm_unmap_exec_lock(struct drm_gpuvm *gpuvm, struct drm_exec *exec,
drivers/gpu/drm/drm_gpuvm.c
2841
return __drm_gpuvm_sm_unmap(gpuvm, &lock_ops, exec,
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
1011
struct drm_i915_gem_exec_object2 *entry = &eb->exec[i];
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
1041
eb_vma_misplaced(&eb->exec[i], vma, ev->flags));
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
1517
const struct drm_i915_gem_exec_object2 *entry = ev->exec;
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
1600
const struct drm_i915_gem_exec_object2 *entry = ev->exec;
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
1655
const unsigned int nreloc = eb->exec[i].relocation_count;
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
1663
err = check_relocations(&eb->exec[i]);
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
1667
urelocs = u64_to_user_ptr(eb->exec[i].relocs_ptr);
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
1709
eb->exec[i].relocs_ptr = (uintptr_t)relocs;
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
1721
relocs = u64_to_ptr(typeof(*relocs), eb->exec[i].relocs_ptr);
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
1722
if (eb->exec[i].relocation_count)
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
1736
err = check_relocations(&eb->exec[i]);
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
1878
&eb->exec[i];
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
2183
struct drm_i915_gem_execbuffer2 *exec)
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
2185
if (exec->flags & __I915_EXEC_ILLEGAL_FLAGS)
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
2189
if (!(exec->flags & (I915_EXEC_FENCE_ARRAY |
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
2191
if (exec->num_cliprects || exec->cliprects_ptr)
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
2195
if (exec->DR4 == 0xffffffff) {
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
2197
exec->DR4 = 0;
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
2199
if (exec->DR1 || exec->DR4)
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
2202
if ((exec->batch_start_offset | exec->batch_len) & 0x7)
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
249
struct drm_i915_gem_exec_object2 *exec; /** ioctl execobj[] */
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
3344
struct drm_i915_gem_exec_object2 *exec)
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
3363
eb.exec = exec;
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
3364
eb.vma = (struct eb_vma *)(exec + args->buffer_count + 1);
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
41
struct drm_i915_gem_exec_object2 *exec;
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
567
struct drm_i915_gem_exec_object2 *entry = &eb->exec[i];
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
571
ev->exec = entry;
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
659
struct drm_i915_gem_exec_object2 *entry = ev->exec;
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
953
vma = eb_lookup_vma(eb, eb->exec[i].handle);
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
959
err = eb_validate_vma(eb, &eb->exec[i], vma);
drivers/gpu/drm/imagination/pvr_job.c
531
prepare_fw_obj_resv(struct drm_exec *exec, struct pvr_fw_object *fw_obj)
drivers/gpu/drm/imagination/pvr_job.c
533
return drm_exec_prepare_obj(exec, gem_from_pvr_gem(fw_obj->gem), 1);
drivers/gpu/drm/imagination/pvr_job.c
537
jobs_lock_all_objs(struct drm_exec *exec, struct pvr_job_data *job_data,
drivers/gpu/drm/imagination/pvr_job.c
546
int err = drm_exec_lock_obj(exec,
drivers/gpu/drm/imagination/pvr_job.c
553
err = prepare_fw_obj_resv(exec,
drivers/gpu/drm/imagination/pvr_job.c
564
prepare_job_resvs_for_each(struct drm_exec *exec, struct pvr_job_data *job_data,
drivers/gpu/drm/imagination/pvr_job.c
567
drm_exec_until_all_locked(exec) {
drivers/gpu/drm/imagination/pvr_job.c
568
int err = jobs_lock_all_objs(exec, job_data, job_count);
drivers/gpu/drm/imagination/pvr_job.c
570
drm_exec_retry_on_contention(exec);
drivers/gpu/drm/imagination/pvr_job.c
711
struct drm_exec exec;
drivers/gpu/drm/imagination/pvr_job.c
743
drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT | DRM_EXEC_IGNORE_DUPLICATES, 0);
drivers/gpu/drm/imagination/pvr_job.c
752
err = prepare_job_resvs_for_each(&exec, job_data, args->jobs.count);
drivers/gpu/drm/imagination/pvr_job.c
769
drm_exec_fini(&exec);
drivers/gpu/drm/imagination/pvr_vm.c
701
return drm_exec_lock_obj(&vm_exec->exec, gem_from_pvr_gem(pvr_obj));
drivers/gpu/drm/msm/msm_gem.c
1045
struct drm_exec exec;
drivers/gpu/drm/msm/msm_gem.c
1073
drm_exec_init(&exec, 0, 0);
drivers/gpu/drm/msm/msm_gem.c
1074
drm_exec_until_all_locked (&exec) {
drivers/gpu/drm/msm/msm_gem.c
1077
drm_exec_lock_obj(&exec,
drivers/gpu/drm/msm/msm_gem.c
1079
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/msm/msm_gem.c
108
struct drm_exec exec;
drivers/gpu/drm/msm/msm_gem.c
1083
drm_exec_fini(&exec); /* drop locks */
drivers/gpu/drm/msm/msm_gem.c
110
msm_gem_lock_vm_and_obj(&exec, obj, priv->kms->vm);
drivers/gpu/drm/msm/msm_gem.c
112
drm_exec_fini(&exec); /* drop locks */
drivers/gpu/drm/msm/msm_gem.c
55
struct drm_exec exec;
drivers/gpu/drm/msm/msm_gem.c
556
struct drm_exec exec;
drivers/gpu/drm/msm/msm_gem.c
559
msm_gem_lock_vm_and_obj(&exec, obj, vm);
drivers/gpu/drm/msm/msm_gem.c
561
drm_exec_fini(&exec); /* drop locks */
drivers/gpu/drm/msm/msm_gem.c
581
struct drm_exec exec;
drivers/gpu/drm/msm/msm_gem.c
584
msm_gem_lock_vm_and_obj(&exec, obj, vm);
drivers/gpu/drm/msm/msm_gem.c
591
drm_exec_fini(&exec); /* drop locks */
drivers/gpu/drm/msm/msm_gem.c
620
struct drm_exec exec;
drivers/gpu/drm/msm/msm_gem.c
623
msm_gem_lock_vm_and_obj(&exec, obj, vm);
drivers/gpu/drm/msm/msm_gem.c
636
drm_exec_fini(&exec); /* drop locks */
drivers/gpu/drm/msm/msm_gem.c
660
struct drm_exec exec;
drivers/gpu/drm/msm/msm_gem.c
662
msm_gem_lock_vm_and_obj(&exec, obj, vm);
drivers/gpu/drm/msm/msm_gem.c
669
drm_exec_fini(&exec); /* drop locks */
drivers/gpu/drm/msm/msm_gem.c
83
msm_gem_lock_vm_and_obj(&exec, obj, ctx->vm);
drivers/gpu/drm/msm/msm_gem.c
85
drm_exec_fini(&exec); /* drop locks */
drivers/gpu/drm/msm/msm_gem.h
358
msm_gem_lock_vm_and_obj(struct drm_exec *exec,
drivers/gpu/drm/msm/msm_gem.h
364
drm_exec_init(exec, 0, 2);
drivers/gpu/drm/msm/msm_gem.h
365
drm_exec_until_all_locked (exec) {
drivers/gpu/drm/msm/msm_gem.h
366
ret = drm_exec_lock_obj(exec, drm_gpuvm_resv_obj(vm));
drivers/gpu/drm/msm/msm_gem.h
368
ret = drm_exec_lock_obj(exec, obj);
drivers/gpu/drm/msm/msm_gem.h
369
drm_exec_retry_on_contention(exec);
drivers/gpu/drm/msm/msm_gem.h
436
struct drm_exec exec;
drivers/gpu/drm/msm/msm_gem_submit.c
277
struct drm_exec *exec = &submit->exec;
drivers/gpu/drm/msm/msm_gem_submit.c
280
drm_exec_init(&submit->exec, flags, submit->nr_bos);
drivers/gpu/drm/msm/msm_gem_submit.c
282
drm_exec_until_all_locked (&submit->exec) {
drivers/gpu/drm/msm/msm_gem_submit.c
283
ret = drm_gpuvm_prepare_vm(submit->vm, exec, 1);
drivers/gpu/drm/msm/msm_gem_submit.c
284
drm_exec_retry_on_contention(exec);
drivers/gpu/drm/msm/msm_gem_submit.c
288
ret = drm_gpuvm_prepare_objects(submit->vm, exec, 1);
drivers/gpu/drm/msm/msm_gem_submit.c
289
drm_exec_retry_on_contention(exec);
drivers/gpu/drm/msm/msm_gem_submit.c
306
drm_exec_init(&submit->exec, flags, submit->nr_bos);
drivers/gpu/drm/msm/msm_gem_submit.c
308
drm_exec_until_all_locked (&submit->exec) {
drivers/gpu/drm/msm/msm_gem_submit.c
309
ret = drm_exec_lock_obj(&submit->exec,
drivers/gpu/drm/msm/msm_gem_submit.c
311
drm_exec_retry_on_contention(&submit->exec);
drivers/gpu/drm/msm/msm_gem_submit.c
316
ret = drm_exec_prepare_obj(&submit->exec, obj, 1);
drivers/gpu/drm/msm/msm_gem_submit.c
317
drm_exec_retry_on_contention(&submit->exec);
drivers/gpu/drm/msm/msm_gem_submit.c
413
drm_gpuvm_resv_add_fence(submit->vm, &submit->exec,
drivers/gpu/drm/msm/msm_gem_submit.c
526
if (submit->exec.objects)
drivers/gpu/drm/msm/msm_gem_submit.c
527
drm_exec_fini(&submit->exec);
drivers/gpu/drm/msm/msm_gem_submit.c
795
ret = drm_gpuvm_validate(submit->vm, &submit->exec);
drivers/gpu/drm/msm/msm_gem_vma.c
1227
vm_bind_job_lock_objects(struct msm_vm_bind_job *job, struct drm_exec *exec)
drivers/gpu/drm/msm/msm_gem_vma.c
1232
drm_exec_until_all_locked (exec) {
drivers/gpu/drm/msm/msm_gem_vma.c
1233
ret = drm_exec_lock_obj(exec, drm_gpuvm_resv_obj(job->vm));
drivers/gpu/drm/msm/msm_gem_vma.c
1234
drm_exec_retry_on_contention(exec);
drivers/gpu/drm/msm/msm_gem_vma.c
1243
ret = drm_gpuvm_sm_unmap_exec_lock(job->vm, exec,
drivers/gpu/drm/msm/msm_gem_vma.c
1256
ret = drm_gpuvm_sm_map_exec_lock(job->vm, exec, 1, &map_req);
drivers/gpu/drm/msm/msm_gem_vma.c
1267
drm_exec_retry_on_contention(exec);
drivers/gpu/drm/msm/msm_gem_vma.c
1529
struct drm_exec exec;
drivers/gpu/drm/msm/msm_gem_vma.c
1531
drm_exec_init(&exec, flags, nr_bos + 1);
drivers/gpu/drm/msm/msm_gem_vma.c
1533
ret = vm_bind_job_lock_objects(job, &exec);
drivers/gpu/drm/msm/msm_gem_vma.c
1577
drm_exec_fini(&exec);
drivers/gpu/drm/msm/msm_gem_vma.c
439
msm_gem_vm_bo_validate(struct drm_gpuvm_bo *vm_bo, struct drm_exec *exec)
drivers/gpu/drm/msm/msm_gem_vma.c
899
struct drm_exec exec;
drivers/gpu/drm/msm/msm_gem_vma.c
916
drm_exec_init(&exec, 0, 2);
drivers/gpu/drm/msm/msm_gem_vma.c
917
drm_exec_until_all_locked (&exec) {
drivers/gpu/drm/msm/msm_gem_vma.c
918
drm_exec_lock_obj(&exec, drm_gpuvm_resv_obj(gpuvm));
drivers/gpu/drm/msm/msm_gem_vma.c
919
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/msm/msm_gem_vma.c
932
drm_exec_lock_obj(&exec, obj);
drivers/gpu/drm/msm/msm_gem_vma.c
933
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/msm/msm_gem_vma.c
940
drm_exec_unlock_obj(&exec, obj);
drivers/gpu/drm/msm/msm_gem_vma.c
944
drm_exec_fini(&exec);
drivers/gpu/drm/msm/msm_gpu.c
262
struct drm_exec exec;
drivers/gpu/drm/msm/msm_gpu.c
266
drm_exec_init(&exec, DRM_EXEC_IGNORE_DUPLICATES, 0);
drivers/gpu/drm/msm/msm_gpu.c
267
drm_exec_until_all_locked(&exec) {
drivers/gpu/drm/msm/msm_gpu.c
270
drm_exec_lock_obj(&exec, drm_gpuvm_resv_obj(submit->vm));
drivers/gpu/drm/msm/msm_gpu.c
271
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/msm/msm_gpu.c
278
drm_exec_lock_obj(&exec, vma->gem.obj);
drivers/gpu/drm/msm/msm_gpu.c
279
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/msm/msm_gpu.c
301
drm_exec_fini(&exec);
drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/dcb.h
66
int dcb_outp_foreach(struct nvkm_bios *, void *data, int (*exec)
drivers/gpu/drm/nouveau/include/nvkm/subdev/bus.h
14
int nvkm_hwsq_fini(struct nvkm_hwsq **, bool exec);
drivers/gpu/drm/nouveau/include/nvkm/subdev/pmu.h
59
int nvkm_memx_fini(struct nvkm_memx **, bool exec);
drivers/gpu/drm/nouveau/include/nvkm/subdev/timer.h
8
struct list_head exec;
drivers/gpu/drm/nouveau/nouveau_bios.c
1328
int (*exec)(struct drm_device *, void *, int idx, u8 *outp))
drivers/gpu/drm/nouveau/nouveau_bios.c
1343
ret = exec(dev, data, idx, outp);
drivers/gpu/drm/nouveau/nouveau_bo.c
1056
drm->ttm.move = mthd->exec;
drivers/gpu/drm/nouveau/nouveau_bo.c
1061
} while ((++mthd)->exec);
drivers/gpu/drm/nouveau/nouveau_bo.c
998
int (*exec)(struct nouveau_channel *,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1220
bind_lock_validate(struct nouveau_job *job, struct drm_exec *exec,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1239
ret = drm_exec_prepare_obj(exec, obj, num_fences);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1265
struct drm_exec *exec = &vme->exec;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1421
drm_exec_init(exec, vme->flags, 0);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1422
drm_exec_until_all_locked(exec) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1423
ret = bind_lock_validate(job, exec, vme->num_fences);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1424
drm_exec_retry_on_contention(exec);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1873
nouveau_uvmm_bo_validate(struct drm_gpuvm_bo *vm_bo, struct drm_exec *exec)
drivers/gpu/drm/nouveau/nvkm/subdev/bios/dcb.c
212
int (*exec)(struct nvkm_bios *, void *, int, u16))
drivers/gpu/drm/nouveau/nvkm/subdev/bios/dcb.c
229
ret = exec(bios, data, idx, outp);
drivers/gpu/drm/nouveau/nvkm/subdev/bios/init.c
2228
void (*exec)(struct nvbios_init *);
drivers/gpu/drm/nouveau/nvkm/subdev/bios/init.c
2310
!init_opcode[opcode].exec) {
drivers/gpu/drm/nouveau/nvkm/subdev/bios/init.c
2315
init_opcode[opcode].exec(init);
drivers/gpu/drm/nouveau/nvkm/subdev/bios/init.c
65
init_exec_set(struct nvbios_init *init, bool exec)
drivers/gpu/drm/nouveau/nvkm/subdev/bios/init.c
67
if (exec) init->execute &= 0xfd;
drivers/gpu/drm/nouveau/nvkm/subdev/bios/init.c
78
init_exec_force(struct nvbios_init *init, bool exec)
drivers/gpu/drm/nouveau/nvkm/subdev/bios/init.c
80
if (exec) init->execute |= 0x04;
drivers/gpu/drm/nouveau/nvkm/subdev/bus/hwsq.c
61
nvkm_hwsq_fini(struct nvkm_hwsq **phwsq, bool exec)
drivers/gpu/drm/nouveau/nvkm/subdev/bus/hwsq.c
70
if (exec)
drivers/gpu/drm/nouveau/nvkm/subdev/bus/hwsq.h
75
hwsq_exec(struct hwsq *ram, bool exec)
drivers/gpu/drm/nouveau/nvkm/subdev/bus/hwsq.h
79
ret = nvkm_hwsq_fini(&ram->hwsq, exec);
drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c
420
void (*exec)(struct gf100_clk *, int);
drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c
434
stage[i].exec(clk, j);
drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c
453
void (*exec)(struct gk104_clk *, int);
drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c
471
stage[i].exec(clk, j);
drivers/gpu/drm/nouveau/nvkm/subdev/devinit/gm200.c
131
u32 exec, args;
drivers/gpu/drm/nouveau/nvkm/subdev/devinit/gm200.c
141
ret = pmu_load(init, 0x04, post, &exec, &args);
drivers/gpu/drm/nouveau/nvkm/subdev/devinit/gm200.c
166
pmu_exec(init, exec);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramfuc.h
71
ramfuc_exec(struct ramfuc *ram, bool exec)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramfuc.h
75
ret = nvkm_memx_fini(&ram->memx, exec);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
888
bool exec = nvkm_boolopt(device->cfgopt, "NvMemExec", true);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
890
if (exec) {
drivers/gpu/drm/nouveau/nvkm/subdev/mxm/base.c
200
bool (*exec)(struct nvkm_mxm *, u8 version);
drivers/gpu/drm/nouveau/nvkm/subdev/mxm/base.c
218
if (shadow->exec(mxm, version)) {
drivers/gpu/drm/nouveau/nvkm/subdev/mxm/mxms.c
148
if (mxm->subdev.debug >= NV_DBG_DEBUG && (exec == NULL)) {
drivers/gpu/drm/nouveau/nvkm/subdev/mxm/mxms.c
170
if (!exec(mxm, desc, info))
drivers/gpu/drm/nouveau/nvkm/subdev/mxm/mxms.c
97
bool (*exec)(struct nvkm_mxm *, u8 *, void *), void *info)
drivers/gpu/drm/nouveau/nvkm/subdev/pmu/memx.c
72
nvkm_memx_fini(struct nvkm_memx **pmemx, bool exec)
drivers/gpu/drm/nouveau/nvkm/subdev/pmu/memx.c
88
if (exec) {
drivers/gpu/drm/nouveau/nvkm/subdev/timer/base.c
100
list_del(&alarm->exec);
drivers/gpu/drm/nouveau/nvkm/subdev/timer/base.c
73
LIST_HEAD(exec);
drivers/gpu/drm/nouveau/nvkm/subdev/timer/base.c
90
list_add(&alarm->exec, &exec);
drivers/gpu/drm/nouveau/nvkm/subdev/timer/base.c
99
list_for_each_entry_safe(alarm, atemp, &exec, exec) {
drivers/gpu/drm/panthor/panthor_drv.c
1051
drm_exec_until_all_locked(&ctx.exec) {
drivers/gpu/drm/panthor/panthor_drv.c
1052
ret = panthor_vm_prepare_mapped_bos_resvs(&ctx.exec, vm,
drivers/gpu/drm/panthor/panthor_drv.c
1253
drm_exec_until_all_locked(&ctx.exec) {
drivers/gpu/drm/panthor/panthor_drv.c
1255
ret = panthor_vm_bind_job_prepare_resvs(&ctx.exec, ctx.jobs[i].job);
drivers/gpu/drm/panthor/panthor_drv.c
1256
drm_exec_retry_on_contention(&ctx.exec);
drivers/gpu/drm/panthor/panthor_drv.c
311
struct drm_exec exec;
drivers/gpu/drm/panthor/panthor_drv.c
705
upd_resvs(&ctx->exec, ctx->jobs[i].job);
drivers/gpu/drm/panthor/panthor_drv.c
734
drm_exec_init(&ctx->exec,
drivers/gpu/drm/panthor/panthor_drv.c
751
drm_exec_fini(&ctx->exec);
drivers/gpu/drm/panthor/panthor_mmu.c
2646
int panthor_vm_bind_job_prepare_resvs(struct drm_exec *exec,
drivers/gpu/drm/panthor/panthor_mmu.c
2653
ret = drm_gpuvm_prepare_vm(&job->vm->base, exec, 1);
drivers/gpu/drm/panthor/panthor_mmu.c
2659
ret = drm_exec_prepare_obj(exec, job->ctx.map.vm_bo->obj, 1);
drivers/gpu/drm/panthor/panthor_mmu.c
2672
void panthor_vm_bind_job_update_resvs(struct drm_exec *exec,
drivers/gpu/drm/panthor/panthor_mmu.c
2678
drm_gpuvm_resv_add_fence(&job->vm->base, exec,
drivers/gpu/drm/panthor/panthor_mmu.c
2684
void panthor_vm_update_resvs(struct panthor_vm *vm, struct drm_exec *exec,
drivers/gpu/drm/panthor/panthor_mmu.c
2689
drm_gpuvm_resv_add_fence(&vm->base, exec, fence, private_usage, extobj_usage);
drivers/gpu/drm/panthor/panthor_mmu.c
2794
int panthor_vm_prepare_mapped_bos_resvs(struct drm_exec *exec, struct panthor_vm *vm,
drivers/gpu/drm/panthor/panthor_mmu.c
2800
ret = drm_gpuvm_prepare_vm(&vm->base, exec, slot_count);
drivers/gpu/drm/panthor/panthor_mmu.c
2804
return drm_gpuvm_prepare_objects(&vm->base, exec, slot_count);
drivers/gpu/drm/panthor/panthor_mmu.h
49
int panthor_vm_prepare_mapped_bos_resvs(struct drm_exec *exec,
drivers/gpu/drm/panthor/panthor_mmu.h
90
int panthor_vm_bind_job_prepare_resvs(struct drm_exec *exec,
drivers/gpu/drm/panthor/panthor_mmu.h
92
void panthor_vm_bind_job_update_resvs(struct drm_exec *exec, struct drm_sched_job *job);
drivers/gpu/drm/panthor/panthor_mmu.h
94
void panthor_vm_update_resvs(struct panthor_vm *vm, struct drm_exec *exec,
drivers/gpu/drm/panthor/panthor_sched.c
4042
void panthor_job_update_resvs(struct drm_exec *exec, struct drm_sched_job *sched_job)
drivers/gpu/drm/panthor/panthor_sched.c
4046
panthor_vm_update_resvs(job->group->vm, exec, &sched_job->s_fence->finished,
drivers/gpu/drm/panthor/panthor_sched.h
38
void panthor_job_update_resvs(struct drm_exec *exec, struct drm_sched_job *job);
drivers/gpu/drm/qxl/qxl_drv.h
153
struct drm_exec exec;
drivers/gpu/drm/qxl/qxl_release.c
223
drm_exec_init(&release->exec, no_intr ? 0 :
drivers/gpu/drm/qxl/qxl_release.c
225
drm_exec_until_all_locked(&release->exec) {
drivers/gpu/drm/qxl/qxl_release.c
227
ret = drm_exec_prepare_obj(&release->exec,
drivers/gpu/drm/qxl/qxl_release.c
230
drm_exec_retry_on_contention(&release->exec);
drivers/gpu/drm/qxl/qxl_release.c
243
drm_exec_fini(&release->exec);
drivers/gpu/drm/qxl/qxl_release.c
254
drm_exec_fini(&release->exec);
drivers/gpu/drm/qxl/qxl_release.c
442
drm_exec_fini(&release->exec);
drivers/gpu/drm/radeon/radeon.h
1032
struct drm_exec exec;
drivers/gpu/drm/radeon/radeon_cs.c
196
r = radeon_bo_list_validate(p->rdev, &p->exec, &p->validated, p->ring);
drivers/gpu/drm/radeon/radeon_cs.c
274
drm_exec_init(&p->exec, DRM_EXEC_INTERRUPTIBLE_WAIT, 0);
drivers/gpu/drm/radeon/radeon_cs.c
442
drm_exec_fini(&parser->exec);
drivers/gpu/drm/radeon/radeon_gem.c
610
struct drm_exec exec;
drivers/gpu/drm/radeon/radeon_gem.c
620
drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT, 0);
drivers/gpu/drm/radeon/radeon_gem.c
621
drm_exec_until_all_locked(&exec) {
drivers/gpu/drm/radeon/radeon_gem.c
623
r = drm_exec_prepare_obj(&exec, &entry->robj->tbo.base,
drivers/gpu/drm/radeon/radeon_gem.c
625
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/radeon/radeon_gem.c
630
r = drm_exec_prepare_obj(&exec, &bo_va->bo->tbo.base, 1);
drivers/gpu/drm/radeon/radeon_gem.c
631
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/radeon/radeon_gem.c
656
drm_exec_fini(&exec);
drivers/gpu/drm/radeon/radeon_object.c
467
struct drm_exec *exec,
drivers/gpu/drm/radeon/radeon_object.c
476
drm_exec_until_all_locked(exec) {
drivers/gpu/drm/radeon/radeon_object.c
478
r = drm_exec_prepare_obj(exec, &lobj->robj->tbo.base,
drivers/gpu/drm/radeon/radeon_object.c
480
drm_exec_retry_on_contention(exec);
drivers/gpu/drm/radeon/radeon_object.h
155
struct drm_exec *exec,
drivers/gpu/drm/tests/drm_exec_test.c
105
struct drm_exec exec;
drivers/gpu/drm/tests/drm_exec_test.c
110
drm_exec_init(&exec, DRM_EXEC_IGNORE_DUPLICATES, 0);
drivers/gpu/drm/tests/drm_exec_test.c
111
drm_exec_until_all_locked(&exec) {
drivers/gpu/drm/tests/drm_exec_test.c
112
ret = drm_exec_lock_obj(&exec, &gobj);
drivers/gpu/drm/tests/drm_exec_test.c
113
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/tests/drm_exec_test.c
118
ret = drm_exec_lock_obj(&exec, &gobj);
drivers/gpu/drm/tests/drm_exec_test.c
119
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/tests/drm_exec_test.c
124
drm_exec_unlock_obj(&exec, &gobj);
drivers/gpu/drm/tests/drm_exec_test.c
125
drm_exec_fini(&exec);
drivers/gpu/drm/tests/drm_exec_test.c
132
struct drm_exec exec;
drivers/gpu/drm/tests/drm_exec_test.c
137
drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT, 0);
drivers/gpu/drm/tests/drm_exec_test.c
138
drm_exec_until_all_locked(&exec) {
drivers/gpu/drm/tests/drm_exec_test.c
139
ret = drm_exec_prepare_obj(&exec, &gobj, 1);
drivers/gpu/drm/tests/drm_exec_test.c
140
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/tests/drm_exec_test.c
145
drm_exec_fini(&exec);
drivers/gpu/drm/tests/drm_exec_test.c
159
struct drm_exec exec;
drivers/gpu/drm/tests/drm_exec_test.c
170
drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT, 0);
drivers/gpu/drm/tests/drm_exec_test.c
171
drm_exec_until_all_locked(&exec)
drivers/gpu/drm/tests/drm_exec_test.c
172
ret = drm_exec_prepare_array(&exec, array, ARRAY_SIZE(array),
drivers/gpu/drm/tests/drm_exec_test.c
175
drm_exec_fini(&exec);
drivers/gpu/drm/tests/drm_exec_test.c
183
struct drm_exec exec;
drivers/gpu/drm/tests/drm_exec_test.c
185
drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT, 0);
drivers/gpu/drm/tests/drm_exec_test.c
186
drm_exec_until_all_locked(&exec)
drivers/gpu/drm/tests/drm_exec_test.c
190
drm_exec_fini(&exec);
drivers/gpu/drm/tests/drm_exec_test.c
192
drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT, 0);
drivers/gpu/drm/tests/drm_exec_test.c
193
drm_exec_until_all_locked(&exec)
drivers/gpu/drm/tests/drm_exec_test.c
197
drm_exec_fini(&exec);
drivers/gpu/drm/tests/drm_exec_test.c
47
struct drm_exec exec;
drivers/gpu/drm/tests/drm_exec_test.c
49
drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT, 0);
drivers/gpu/drm/tests/drm_exec_test.c
50
drm_exec_fini(&exec);
drivers/gpu/drm/tests/drm_exec_test.c
58
struct drm_exec exec;
drivers/gpu/drm/tests/drm_exec_test.c
63
drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT, 0);
drivers/gpu/drm/tests/drm_exec_test.c
64
drm_exec_until_all_locked(&exec) {
drivers/gpu/drm/tests/drm_exec_test.c
65
ret = drm_exec_lock_obj(&exec, &gobj);
drivers/gpu/drm/tests/drm_exec_test.c
66
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/tests/drm_exec_test.c
71
drm_exec_fini(&exec);
drivers/gpu/drm/tests/drm_exec_test.c
78
struct drm_exec exec;
drivers/gpu/drm/tests/drm_exec_test.c
83
drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT, 0);
drivers/gpu/drm/tests/drm_exec_test.c
84
drm_exec_until_all_locked(&exec) {
drivers/gpu/drm/tests/drm_exec_test.c
85
ret = drm_exec_lock_obj(&exec, &gobj);
drivers/gpu/drm/tests/drm_exec_test.c
86
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/tests/drm_exec_test.c
91
drm_exec_unlock_obj(&exec, &gobj);
drivers/gpu/drm/tests/drm_exec_test.c
92
ret = drm_exec_lock_obj(&exec, &gobj);
drivers/gpu/drm/tests/drm_exec_test.c
93
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/tests/drm_exec_test.c
98
drm_exec_fini(&exec);
drivers/gpu/drm/vc4/vc4_drv.h
1000
void vc4_move_job_to_render(struct drm_device *dev, struct vc4_exec_info *exec);
drivers/gpu/drm/vc4/vc4_drv.h
1069
struct vc4_exec_info *exec);
drivers/gpu/drm/vc4/vc4_drv.h
1072
vc4_validate_shader_recs(struct drm_device *dev, struct vc4_exec_info *exec);
drivers/gpu/drm/vc4/vc4_drv.h
1074
struct drm_gem_dma_object *vc4_use_bo(struct vc4_exec_info *exec,
drivers/gpu/drm/vc4/vc4_drv.h
1077
int vc4_get_rcl(struct drm_device *dev, struct vc4_exec_info *exec);
drivers/gpu/drm/vc4/vc4_drv.h
1079
bool vc4_check_tex_size(struct vc4_exec_info *exec,
drivers/gpu/drm/vc4/vc4_gem.c
1016
struct vc4_exec_info *exec;
drivers/gpu/drm/vc4/vc4_gem.c
1046
exec = kzalloc_objs(*exec, 1);
drivers/gpu/drm/vc4/vc4_gem.c
1047
if (!exec)
drivers/gpu/drm/vc4/vc4_gem.c
1050
exec->dev = vc4;
drivers/gpu/drm/vc4/vc4_gem.c
1054
kfree(exec);
drivers/gpu/drm/vc4/vc4_gem.c
1058
exec->args = args;
drivers/gpu/drm/vc4/vc4_gem.c
1059
INIT_LIST_HEAD(&exec->unref_list);
drivers/gpu/drm/vc4/vc4_gem.c
1061
ret = vc4_cl_lookup_bos(dev, file_priv, exec);
drivers/gpu/drm/vc4/vc4_gem.c
1066
exec->perfmon = vc4_perfmon_find(vc4file,
drivers/gpu/drm/vc4/vc4_gem.c
1068
if (!exec->perfmon) {
drivers/gpu/drm/vc4/vc4_gem.c
1097
if (exec->args->bin_cl_size != 0) {
drivers/gpu/drm/vc4/vc4_gem.c
1098
ret = vc4_get_bcl(dev, exec);
drivers/gpu/drm/vc4/vc4_gem.c
1102
exec->ct0ca = 0;
drivers/gpu/drm/vc4/vc4_gem.c
1103
exec->ct0ea = 0;
drivers/gpu/drm/vc4/vc4_gem.c
1106
ret = vc4_get_rcl(dev, exec);
drivers/gpu/drm/vc4/vc4_gem.c
1110
ret = vc4_lock_bo_reservations(exec, &exec_ctx);
drivers/gpu/drm/vc4/vc4_gem.c
1131
exec->args = NULL;
drivers/gpu/drm/vc4/vc4_gem.c
1133
ret = vc4_queue_submit(dev, exec, &exec_ctx, out_sync);
drivers/gpu/drm/vc4/vc4_gem.c
1152
vc4_complete_exec(&vc4->base, exec);
drivers/gpu/drm/vc4/vc4_gem.c
159
struct vc4_exec_info *exec[2];
drivers/gpu/drm/vc4/vc4_gem.c
171
exec[0] = vc4_first_bin_job(vc4);
drivers/gpu/drm/vc4/vc4_gem.c
172
exec[1] = vc4_first_render_job(vc4);
drivers/gpu/drm/vc4/vc4_gem.c
173
if (!exec[0] && !exec[1]) {
drivers/gpu/drm/vc4/vc4_gem.c
181
if (!exec[i])
drivers/gpu/drm/vc4/vc4_gem.c
185
list_for_each_entry(bo, &exec[i]->unref_list, unref_head)
drivers/gpu/drm/vc4/vc4_gem.c
187
state->bo_count += exec[i]->bo_count + unref_list_count;
drivers/gpu/drm/vc4/vc4_gem.c
200
if (!exec[i])
drivers/gpu/drm/vc4/vc4_gem.c
203
for (j = 0; j < exec[i]->bo_count; j++) {
drivers/gpu/drm/vc4/vc4_gem.c
204
bo = to_vc4_bo(exec[i]->bo[j]);
drivers/gpu/drm/vc4/vc4_gem.c
212
drm_gem_object_get(exec[i]->bo[j]);
drivers/gpu/drm/vc4/vc4_gem.c
213
kernel_state->bo[k++] = exec[i]->bo[j];
drivers/gpu/drm/vc4/vc4_gem.c
216
list_for_each_entry(bo, &exec[i]->unref_list, unref_head) {
drivers/gpu/drm/vc4/vc4_gem.c
227
if (exec[0])
drivers/gpu/drm/vc4/vc4_gem.c
228
state->start_bin = exec[0]->ct0ca;
drivers/gpu/drm/vc4/vc4_gem.c
229
if (exec[1])
drivers/gpu/drm/vc4/vc4_gem.c
230
state->start_render = exec[1]->ct1ca;
drivers/gpu/drm/vc4/vc4_gem.c
477
struct vc4_exec_info *exec;
drivers/gpu/drm/vc4/vc4_gem.c
483
exec = vc4_first_bin_job(vc4);
drivers/gpu/drm/vc4/vc4_gem.c
484
if (!exec)
drivers/gpu/drm/vc4/vc4_gem.c
492
if (exec->perfmon && vc4->active_perfmon != exec->perfmon)
drivers/gpu/drm/vc4/vc4_gem.c
493
vc4_perfmon_start(vc4, exec->perfmon);
drivers/gpu/drm/vc4/vc4_gem.c
498
if (exec->ct0ca != exec->ct0ea) {
drivers/gpu/drm/vc4/vc4_gem.c
499
trace_vc4_submit_cl(dev, false, exec->seqno, exec->ct0ca,
drivers/gpu/drm/vc4/vc4_gem.c
500
exec->ct0ea);
drivers/gpu/drm/vc4/vc4_gem.c
501
submit_cl(dev, 0, exec->ct0ca, exec->ct0ea);
drivers/gpu/drm/vc4/vc4_gem.c
505
vc4_move_job_to_render(dev, exec);
drivers/gpu/drm/vc4/vc4_gem.c
513
if (next && next->perfmon == exec->perfmon)
drivers/gpu/drm/vc4/vc4_gem.c
522
struct vc4_exec_info *exec = vc4_first_render_job(vc4);
drivers/gpu/drm/vc4/vc4_gem.c
524
if (!exec)
drivers/gpu/drm/vc4/vc4_gem.c
538
trace_vc4_submit_cl(dev, true, exec->seqno, exec->ct1ca, exec->ct1ea);
drivers/gpu/drm/vc4/vc4_gem.c
539
submit_cl(dev, 1, exec->ct1ca, exec->ct1ea);
drivers/gpu/drm/vc4/vc4_gem.c
543
vc4_move_job_to_render(struct drm_device *dev, struct vc4_exec_info *exec)
drivers/gpu/drm/vc4/vc4_gem.c
551
list_move_tail(&exec->head, &vc4->render_job_list);
drivers/gpu/drm/vc4/vc4_gem.c
557
vc4_attach_fences(struct vc4_exec_info *exec)
drivers/gpu/drm/vc4/vc4_gem.c
562
for (i = 0; i < exec->bo_count; i++) {
drivers/gpu/drm/vc4/vc4_gem.c
563
bo = to_vc4_bo(exec->bo[i]);
drivers/gpu/drm/vc4/vc4_gem.c
564
dma_resv_add_fence(bo->base.base.resv, exec->fence,
drivers/gpu/drm/vc4/vc4_gem.c
568
for (i = 0; i < exec->rcl_write_bo_count; i++) {
drivers/gpu/drm/vc4/vc4_gem.c
569
bo = to_vc4_bo(&exec->rcl_write_bo[i]->base);
drivers/gpu/drm/vc4/vc4_gem.c
570
dma_resv_add_fence(bo->base.base.resv, exec->fence,
drivers/gpu/drm/vc4/vc4_gem.c
583
vc4_lock_bo_reservations(struct vc4_exec_info *exec,
drivers/gpu/drm/vc4/vc4_gem.c
591
drm_exec_init(exec_ctx, DRM_EXEC_INTERRUPTIBLE_WAIT, exec->bo_count);
drivers/gpu/drm/vc4/vc4_gem.c
593
ret = drm_exec_prepare_array(exec_ctx, exec->bo,
drivers/gpu/drm/vc4/vc4_gem.c
594
exec->bo_count, 1);
drivers/gpu/drm/vc4/vc4_gem.c
615
vc4_queue_submit(struct drm_device *dev, struct vc4_exec_info *exec,
drivers/gpu/drm/vc4/vc4_gem.c
633
exec->seqno = seqno;
drivers/gpu/drm/vc4/vc4_gem.c
636
vc4->dma_fence_context, exec->seqno);
drivers/gpu/drm/vc4/vc4_gem.c
637
fence->seqno = exec->seqno;
drivers/gpu/drm/vc4/vc4_gem.c
638
exec->fence = &fence->base;
drivers/gpu/drm/vc4/vc4_gem.c
641
drm_syncobj_replace_fence(out_sync, exec->fence);
drivers/gpu/drm/vc4/vc4_gem.c
643
vc4_attach_fences(exec);
drivers/gpu/drm/vc4/vc4_gem.c
647
list_add_tail(&exec->head, &vc4->bin_job_list);
drivers/gpu/drm/vc4/vc4_gem.c
655
if (vc4_first_bin_job(vc4) == exec &&
drivers/gpu/drm/vc4/vc4_gem.c
656
(!renderjob || renderjob->perfmon == exec->perfmon)) {
drivers/gpu/drm/vc4/vc4_gem.c
680
struct vc4_exec_info *exec)
drivers/gpu/drm/vc4/vc4_gem.c
682
struct drm_vc4_submit_cl *args = exec->args;
drivers/gpu/drm/vc4/vc4_gem.c
686
exec->bo_count = args->bo_handle_count;
drivers/gpu/drm/vc4/vc4_gem.c
688
if (!exec->bo_count) {
drivers/gpu/drm/vc4/vc4_gem.c
697
exec->bo_count, &exec->bo);
drivers/gpu/drm/vc4/vc4_gem.c
702
for (i = 0; i < exec->bo_count; i++) {
drivers/gpu/drm/vc4/vc4_gem.c
703
ret = vc4_bo_inc_usecnt(to_vc4_bo(exec->bo[i]));
drivers/gpu/drm/vc4/vc4_gem.c
720
vc4_bo_dec_usecnt(to_vc4_bo(exec->bo[i]));
drivers/gpu/drm/vc4/vc4_gem.c
724
for (i = 0; i < exec->bo_count && exec->bo[i]; i++)
drivers/gpu/drm/vc4/vc4_gem.c
725
drm_gem_object_put(exec->bo[i]);
drivers/gpu/drm/vc4/vc4_gem.c
727
kvfree(exec->bo);
drivers/gpu/drm/vc4/vc4_gem.c
728
exec->bo = NULL;
drivers/gpu/drm/vc4/vc4_gem.c
733
vc4_get_bcl(struct drm_device *dev, struct vc4_exec_info *exec)
drivers/gpu/drm/vc4/vc4_gem.c
735
struct drm_vc4_submit_cl *args = exec->args;
drivers/gpu/drm/vc4/vc4_gem.c
775
exec->shader_rec_u = temp + shader_rec_offset;
drivers/gpu/drm/vc4/vc4_gem.c
776
exec->uniforms_u = temp + uniforms_offset;
drivers/gpu/drm/vc4/vc4_gem.c
777
exec->shader_state = temp + exec_size;
drivers/gpu/drm/vc4/vc4_gem.c
778
exec->shader_state_size = args->shader_rec_count;
drivers/gpu/drm/vc4/vc4_gem.c
787
if (copy_from_user(exec->shader_rec_u,
drivers/gpu/drm/vc4/vc4_gem.c
794
if (copy_from_user(exec->uniforms_u,
drivers/gpu/drm/vc4/vc4_gem.c
807
exec->exec_bo = &bo->base;
drivers/gpu/drm/vc4/vc4_gem.c
809
list_add_tail(&to_vc4_bo(&exec->exec_bo->base)->unref_head,
drivers/gpu/drm/vc4/vc4_gem.c
810
&exec->unref_list);
drivers/gpu/drm/vc4/vc4_gem.c
812
exec->ct0ca = exec->exec_bo->dma_addr + bin_offset;
drivers/gpu/drm/vc4/vc4_gem.c
814
exec->bin_u = bin;
drivers/gpu/drm/vc4/vc4_gem.c
816
exec->shader_rec_v = exec->exec_bo->vaddr + shader_rec_offset;
drivers/gpu/drm/vc4/vc4_gem.c
817
exec->shader_rec_p = exec->exec_bo->dma_addr + shader_rec_offset;
drivers/gpu/drm/vc4/vc4_gem.c
818
exec->shader_rec_size = args->shader_rec_size;
drivers/gpu/drm/vc4/vc4_gem.c
820
exec->uniforms_v = exec->exec_bo->vaddr + uniforms_offset;
drivers/gpu/drm/vc4/vc4_gem.c
821
exec->uniforms_p = exec->exec_bo->dma_addr + uniforms_offset;
drivers/gpu/drm/vc4/vc4_gem.c
822
exec->uniforms_size = args->uniforms_size;
drivers/gpu/drm/vc4/vc4_gem.c
825
exec->exec_bo->vaddr + bin_offset,
drivers/gpu/drm/vc4/vc4_gem.c
827
exec);
drivers/gpu/drm/vc4/vc4_gem.c
831
ret = vc4_validate_shader_recs(dev, exec);
drivers/gpu/drm/vc4/vc4_gem.c
835
if (exec->found_tile_binning_mode_config_packet) {
drivers/gpu/drm/vc4/vc4_gem.c
836
ret = vc4_v3d_bin_bo_get(vc4, &exec->bin_bo_used);
drivers/gpu/drm/vc4/vc4_gem.c
847
vc4_complete_exec(struct drm_device *dev, struct vc4_exec_info *exec)
drivers/gpu/drm/vc4/vc4_gem.c
856
if (exec->fence) {
drivers/gpu/drm/vc4/vc4_gem.c
857
dma_fence_signal(exec->fence);
drivers/gpu/drm/vc4/vc4_gem.c
858
dma_fence_put(exec->fence);
drivers/gpu/drm/vc4/vc4_gem.c
861
if (exec->bo) {
drivers/gpu/drm/vc4/vc4_gem.c
862
for (i = 0; i < exec->bo_count; i++) {
drivers/gpu/drm/vc4/vc4_gem.c
863
struct vc4_bo *bo = to_vc4_bo(exec->bo[i]);
drivers/gpu/drm/vc4/vc4_gem.c
866
drm_gem_object_put(exec->bo[i]);
drivers/gpu/drm/vc4/vc4_gem.c
868
kvfree(exec->bo);
drivers/gpu/drm/vc4/vc4_gem.c
871
while (!list_empty(&exec->unref_list)) {
drivers/gpu/drm/vc4/vc4_gem.c
872
struct vc4_bo *bo = list_first_entry(&exec->unref_list,
drivers/gpu/drm/vc4/vc4_gem.c
880
vc4->bin_alloc_used &= ~exec->bin_slots;
drivers/gpu/drm/vc4/vc4_gem.c
884
if (exec->bin_bo_used)
drivers/gpu/drm/vc4/vc4_gem.c
888
vc4_perfmon_put(exec->perfmon);
drivers/gpu/drm/vc4/vc4_gem.c
892
kfree(exec);
drivers/gpu/drm/vc4/vc4_gem.c
905
struct vc4_exec_info *exec =
drivers/gpu/drm/vc4/vc4_gem.c
908
list_del(&exec->head);
drivers/gpu/drm/vc4/vc4_gem.c
911
vc4_complete_exec(&vc4->base, exec);
drivers/gpu/drm/vc4/vc4_irq.c
121
struct vc4_exec_info *next, *exec = vc4_first_bin_job(vc4);
drivers/gpu/drm/vc4/vc4_irq.c
123
if (!exec)
drivers/gpu/drm/vc4/vc4_irq.c
126
trace_vc4_bcl_end_irq(dev, exec->seqno);
drivers/gpu/drm/vc4/vc4_irq.c
128
vc4_move_job_to_render(dev, exec);
drivers/gpu/drm/vc4/vc4_irq.c
135
if (next && next->perfmon == exec->perfmon)
drivers/gpu/drm/vc4/vc4_irq.c
143
struct vc4_exec_info *exec = vc4_first_bin_job(vc4);
drivers/gpu/drm/vc4/vc4_irq.c
145
if (!exec)
drivers/gpu/drm/vc4/vc4_irq.c
149
if (exec->perfmon)
drivers/gpu/drm/vc4/vc4_irq.c
150
vc4_perfmon_stop(vc4, exec->perfmon, false);
drivers/gpu/drm/vc4/vc4_irq.c
152
list_move_tail(&exec->head, &vc4->bin_job_list);
drivers/gpu/drm/vc4/vc4_irq.c
160
struct vc4_exec_info *exec = vc4_first_render_job(vc4);
drivers/gpu/drm/vc4/vc4_irq.c
163
if (!exec)
drivers/gpu/drm/vc4/vc4_irq.c
166
trace_vc4_rcl_end_irq(dev, exec->seqno);
drivers/gpu/drm/vc4/vc4_irq.c
169
list_move_tail(&exec->head, &vc4->job_done_list);
drivers/gpu/drm/vc4/vc4_irq.c
177
if (exec->perfmon && !nextrender &&
drivers/gpu/drm/vc4/vc4_irq.c
178
(!nextbin || nextbin->perfmon != exec->perfmon))
drivers/gpu/drm/vc4/vc4_irq.c
179
vc4_perfmon_stop(vc4, exec->perfmon, true);
drivers/gpu/drm/vc4/vc4_irq.c
190
else if (nextbin && nextbin->perfmon != exec->perfmon)
drivers/gpu/drm/vc4/vc4_irq.c
193
if (exec->fence) {
drivers/gpu/drm/vc4/vc4_irq.c
194
dma_fence_signal_locked(exec->fence);
drivers/gpu/drm/vc4/vc4_irq.c
195
dma_fence_put(exec->fence);
drivers/gpu/drm/vc4/vc4_irq.c
196
exec->fence = NULL;
drivers/gpu/drm/vc4/vc4_irq.c
68
struct vc4_exec_info *exec;
drivers/gpu/drm/vc4/vc4_irq.c
93
exec = vc4_first_bin_job(vc4);
drivers/gpu/drm/vc4/vc4_irq.c
94
if (!exec)
drivers/gpu/drm/vc4/vc4_irq.c
95
exec = vc4_last_render_job(vc4);
drivers/gpu/drm/vc4/vc4_irq.c
96
if (exec) {
drivers/gpu/drm/vc4/vc4_irq.c
97
exec->bin_slots |= vc4->bin_alloc_overflow;
drivers/gpu/drm/vc4/vc4_render_cl.c
101
static uint32_t vc4_full_res_offset(struct vc4_exec_info *exec,
drivers/gpu/drm/vc4/vc4_render_cl.c
107
(DIV_ROUND_UP(exec->args->width, 32) * y + x);
drivers/gpu/drm/vc4/vc4_render_cl.c
125
static void emit_tile(struct vc4_exec_info *exec,
drivers/gpu/drm/vc4/vc4_render_cl.c
129
struct drm_vc4_submit_cl *args = exec->args;
drivers/gpu/drm/vc4/vc4_render_cl.c
141
vc4_full_res_offset(exec, setup->color_read,
drivers/gpu/drm/vc4/vc4_render_cl.c
163
vc4_full_res_offset(exec, setup->zs_read,
drivers/gpu/drm/vc4/vc4_render_cl.c
187
rcl_u32(setup, (exec->tile_alloc_offset +
drivers/gpu/drm/vc4/vc4_render_cl.c
188
(y * exec->bin_tiles_x + x) * 32));
drivers/gpu/drm/vc4/vc4_render_cl.c
203
vc4_full_res_offset(exec, setup->msaa_color_write,
drivers/gpu/drm/vc4/vc4_render_cl.c
221
vc4_full_res_offset(exec, setup->msaa_zs_write,
drivers/gpu/drm/vc4/vc4_render_cl.c
255
static int vc4_create_rcl_bo(struct drm_device *dev, struct vc4_exec_info *exec,
drivers/gpu/drm/vc4/vc4_render_cl.c
258
struct drm_vc4_submit_cl *args = exec->args;
drivers/gpu/drm/vc4/vc4_render_cl.c
337
&exec->unref_list);
drivers/gpu/drm/vc4/vc4_render_cl.c
374
emit_tile(exec, setup, x, y, first, last);
drivers/gpu/drm/vc4/vc4_render_cl.c
379
exec->ct1ca = setup->rcl->dma_addr;
drivers/gpu/drm/vc4/vc4_render_cl.c
380
exec->ct1ea = setup->rcl->dma_addr + setup->next_offset;
drivers/gpu/drm/vc4/vc4_render_cl.c
385
static int vc4_full_res_bounds_check(struct vc4_exec_info *exec,
drivers/gpu/drm/vc4/vc4_render_cl.c
389
struct drm_vc4_submit_cl *args = exec->args;
drivers/gpu/drm/vc4/vc4_render_cl.c
390
u32 render_tiles_stride = DIV_ROUND_UP(exec->args->width, 32);
drivers/gpu/drm/vc4/vc4_render_cl.c
411
static int vc4_rcl_msaa_surface_setup(struct vc4_exec_info *exec,
drivers/gpu/drm/vc4/vc4_render_cl.c
423
*obj = vc4_use_bo(exec, surf->hindex);
drivers/gpu/drm/vc4/vc4_render_cl.c
427
exec->rcl_write_bo[exec->rcl_write_bo_count++] = *obj;
drivers/gpu/drm/vc4/vc4_render_cl.c
434
return vc4_full_res_bounds_check(exec, *obj, surf);
drivers/gpu/drm/vc4/vc4_render_cl.c
437
static int vc4_rcl_surface_setup(struct vc4_exec_info *exec,
drivers/gpu/drm/vc4/vc4_render_cl.c
459
*obj = vc4_use_bo(exec, surf->hindex);
drivers/gpu/drm/vc4/vc4_render_cl.c
464
exec->rcl_write_bo[exec->rcl_write_bo_count++] = *obj;
drivers/gpu/drm/vc4/vc4_render_cl.c
467
if (surf == &exec->args->zs_write) {
drivers/gpu/drm/vc4/vc4_render_cl.c
478
ret = vc4_full_res_bounds_check(exec, *obj, surf);
drivers/gpu/drm/vc4/vc4_render_cl.c
527
if (!vc4_check_tex_size(exec, *obj, surf->offset, tiling,
drivers/gpu/drm/vc4/vc4_render_cl.c
528
exec->args->width, exec->args->height, cpp)) {
drivers/gpu/drm/vc4/vc4_render_cl.c
536
vc4_rcl_render_config_surface_setup(struct vc4_exec_info *exec,
drivers/gpu/drm/vc4/vc4_render_cl.c
564
*obj = vc4_use_bo(exec, surf->hindex);
drivers/gpu/drm/vc4/vc4_render_cl.c
568
exec->rcl_write_bo[exec->rcl_write_bo_count++] = *obj;
drivers/gpu/drm/vc4/vc4_render_cl.c
588
if (!vc4_check_tex_size(exec, *obj, surf->offset, tiling,
drivers/gpu/drm/vc4/vc4_render_cl.c
589
exec->args->width, exec->args->height, cpp)) {
drivers/gpu/drm/vc4/vc4_render_cl.c
596
int vc4_get_rcl(struct drm_device *dev, struct vc4_exec_info *exec)
drivers/gpu/drm/vc4/vc4_render_cl.c
600
struct drm_vc4_submit_cl *args = exec->args;
drivers/gpu/drm/vc4/vc4_render_cl.c
616
(args->max_x_tile > exec->bin_tiles_x ||
drivers/gpu/drm/vc4/vc4_render_cl.c
617
args->max_y_tile > exec->bin_tiles_y)) {
drivers/gpu/drm/vc4/vc4_render_cl.c
621
exec->bin_tiles_x, exec->bin_tiles_y);
drivers/gpu/drm/vc4/vc4_render_cl.c
625
ret = vc4_rcl_render_config_surface_setup(exec, &setup,
drivers/gpu/drm/vc4/vc4_render_cl.c
631
ret = vc4_rcl_surface_setup(exec, &setup.color_read, &args->color_read,
drivers/gpu/drm/vc4/vc4_render_cl.c
636
ret = vc4_rcl_surface_setup(exec, &setup.zs_read, &args->zs_read,
drivers/gpu/drm/vc4/vc4_render_cl.c
641
ret = vc4_rcl_surface_setup(exec, &setup.zs_write, &args->zs_write,
drivers/gpu/drm/vc4/vc4_render_cl.c
646
ret = vc4_rcl_msaa_surface_setup(exec, &setup.msaa_color_write,
drivers/gpu/drm/vc4/vc4_render_cl.c
651
ret = vc4_rcl_msaa_surface_setup(exec, &setup.msaa_zs_write,
drivers/gpu/drm/vc4/vc4_render_cl.c
665
return vc4_create_rcl_bo(dev, exec, &setup);
drivers/gpu/drm/vc4/vc4_v3d.c
181
struct vc4_exec_info *exec;
drivers/gpu/drm/vc4/vc4_v3d.c
200
exec = vc4_last_render_job(vc4);
drivers/gpu/drm/vc4/vc4_v3d.c
201
if (exec)
drivers/gpu/drm/vc4/vc4_v3d.c
202
seqno = exec->seqno;
drivers/gpu/drm/vc4/vc4_validate.c
108
vc4_use_bo(struct vc4_exec_info *exec, uint32_t hindex)
drivers/gpu/drm/vc4/vc4_validate.c
110
struct vc4_dev *vc4 = exec->dev;
drivers/gpu/drm/vc4/vc4_validate.c
117
if (hindex >= exec->bo_count) {
drivers/gpu/drm/vc4/vc4_validate.c
119
hindex, exec->bo_count);
drivers/gpu/drm/vc4/vc4_validate.c
122
obj = to_drm_gem_dma_obj(exec->bo[hindex]);
drivers/gpu/drm/vc4/vc4_validate.c
135
vc4_use_handle(struct vc4_exec_info *exec, uint32_t gem_handles_packet_index)
drivers/gpu/drm/vc4/vc4_validate.c
137
return vc4_use_bo(exec, exec->bo_index[gem_handles_packet_index]);
drivers/gpu/drm/vc4/vc4_validate.c
141
validate_bin_pos(struct vc4_exec_info *exec, void *untrusted, uint32_t pos)
drivers/gpu/drm/vc4/vc4_validate.c
146
return (untrusted - 1 == exec->bin_u + pos);
drivers/gpu/drm/vc4/vc4_validate.c
165
vc4_check_tex_size(struct vc4_exec_info *exec, struct drm_gem_dma_object *fbo,
drivers/gpu/drm/vc4/vc4_validate.c
169
struct vc4_dev *vc4 = exec->dev;
drivers/gpu/drm/vc4/vc4_validate.c
226
if (!validate_bin_pos(exec, untrusted, exec->args->bin_cl_size - 1)) {
drivers/gpu/drm/vc4/vc4_validate.c
230
exec->found_flush = true;
drivers/gpu/drm/vc4/vc4_validate.c
238
if (exec->found_start_tile_binning_packet) {
drivers/gpu/drm/vc4/vc4_validate.c
242
exec->found_start_tile_binning_packet = true;
drivers/gpu/drm/vc4/vc4_validate.c
244
if (!exec->found_tile_binning_mode_config_packet) {
drivers/gpu/drm/vc4/vc4_validate.c
255
if (!validate_bin_pos(exec, untrusted, exec->args->bin_cl_size - 2)) {
drivers/gpu/drm/vc4/vc4_validate.c
260
exec->found_increment_semaphore_packet = true;
drivers/gpu/drm/vc4/vc4_validate.c
276
if (exec->shader_state_count == 0) {
drivers/gpu/drm/vc4/vc4_validate.c
280
shader_state = &exec->shader_state[exec->shader_state_count - 1];
drivers/gpu/drm/vc4/vc4_validate.c
285
ib = vc4_use_handle(exec, 0);
drivers/gpu/drm/vc4/vc4_validate.c
310
if (exec->shader_state_count == 0) {
drivers/gpu/drm/vc4/vc4_validate.c
314
shader_state = &exec->shader_state[exec->shader_state_count - 1];
drivers/gpu/drm/vc4/vc4_validate.c
331
uint32_t i = exec->shader_state_count++;
drivers/gpu/drm/vc4/vc4_validate.c
333
if (i >= exec->shader_state_size) {
drivers/gpu/drm/vc4/vc4_validate.c
338
exec->shader_state[i].addr = *(uint32_t *)untrusted;
drivers/gpu/drm/vc4/vc4_validate.c
339
exec->shader_state[i].max_index = 0;
drivers/gpu/drm/vc4/vc4_validate.c
341
if (exec->shader_state[i].addr & ~0xf) {
drivers/gpu/drm/vc4/vc4_validate.c
346
*(uint32_t *)validated = (exec->shader_rec_p +
drivers/gpu/drm/vc4/vc4_validate.c
347
exec->shader_state[i].addr);
drivers/gpu/drm/vc4/vc4_validate.c
349
exec->shader_rec_p +=
drivers/gpu/drm/vc4/vc4_validate.c
350
roundup(gl_shader_rec_size(exec->shader_state[i].addr), 16);
drivers/gpu/drm/vc4/vc4_validate.c
358
struct drm_device *dev = exec->exec_bo->base.dev;
drivers/gpu/drm/vc4/vc4_validate.c
365
if (exec->found_tile_binning_mode_config_packet) {
drivers/gpu/drm/vc4/vc4_validate.c
369
exec->found_tile_binning_mode_config_packet = true;
drivers/gpu/drm/vc4/vc4_validate.c
371
exec->bin_tiles_x = *(uint8_t *)(untrusted + 12);
drivers/gpu/drm/vc4/vc4_validate.c
372
exec->bin_tiles_y = *(uint8_t *)(untrusted + 13);
drivers/gpu/drm/vc4/vc4_validate.c
373
tile_count = exec->bin_tiles_x * exec->bin_tiles_y;
drivers/gpu/drm/vc4/vc4_validate.c
376
if (exec->bin_tiles_x == 0 ||
drivers/gpu/drm/vc4/vc4_validate.c
377
exec->bin_tiles_y == 0) {
drivers/gpu/drm/vc4/vc4_validate.c
379
exec->bin_tiles_x, exec->bin_tiles_y);
drivers/gpu/drm/vc4/vc4_validate.c
401
exec->bin_slots |= BIT(bin_slot);
drivers/gpu/drm/vc4/vc4_validate.c
410
exec->tile_alloc_offset = bin_addr + roundup(tile_state_size, 4096);
drivers/gpu/drm/vc4/vc4_validate.c
422
*(uint32_t *)(validated + 0) = exec->tile_alloc_offset;
drivers/gpu/drm/vc4/vc4_validate.c
425
exec->tile_alloc_offset);
drivers/gpu/drm/vc4/vc4_validate.c
435
memcpy(exec->bo_index, untrusted, sizeof(exec->bo_index));
drivers/gpu/drm/vc4/vc4_validate.c
445
int (*func)(struct vc4_exec_info *exec, void *validated,
drivers/gpu/drm/vc4/vc4_validate.c
490
struct vc4_exec_info *exec)
drivers/gpu/drm/vc4/vc4_validate.c
493
uint32_t len = exec->args->bin_cl_size;
drivers/gpu/drm/vc4/vc4_validate.c
53
struct vc4_exec_info *exec, \
drivers/gpu/drm/vc4/vc4_validate.c
530
if (info->func && info->func(exec,
drivers/gpu/drm/vc4/vc4_validate.c
548
exec->ct0ea = exec->ct0ca + dst_offset;
drivers/gpu/drm/vc4/vc4_validate.c
550
if (!exec->found_start_tile_binning_packet) {
drivers/gpu/drm/vc4/vc4_validate.c
562
if (!exec->found_increment_semaphore_packet || !exec->found_flush) {
drivers/gpu/drm/vc4/vc4_validate.c
572
reloc_tex(struct vc4_exec_info *exec,
drivers/gpu/drm/vc4/vc4_validate.c
584
uint32_t *validated_p0 = exec->uniforms_v + sample->p_offset[0];
drivers/gpu/drm/vc4/vc4_validate.c
594
tex = vc4_use_bo(exec, texture_handle_index);
drivers/gpu/drm/vc4/vc4_validate.c
689
if (!vc4_check_tex_size(exec, tex, offset + cube_map_stride * 5,
drivers/gpu/drm/vc4/vc4_validate.c
751
struct vc4_exec_info *exec,
drivers/gpu/drm/vc4/vc4_validate.c
772
if (nr_relocs * 4 > exec->shader_rec_size) {
drivers/gpu/drm/vc4/vc4_validate.c
775
nr_relocs, exec->shader_rec_size);
drivers/gpu/drm/vc4/vc4_validate.c
778
src_handles = exec->shader_rec_u;
drivers/gpu/drm/vc4/vc4_validate.c
779
exec->shader_rec_u += nr_relocs * 4;
drivers/gpu/drm/vc4/vc4_validate.c
780
exec->shader_rec_size -= nr_relocs * 4;
drivers/gpu/drm/vc4/vc4_validate.c
782
if (packet_size > exec->shader_rec_size) {
drivers/gpu/drm/vc4/vc4_validate.c
785
packet_size, exec->shader_rec_size);
drivers/gpu/drm/vc4/vc4_validate.c
788
pkt_u = exec->shader_rec_u;
drivers/gpu/drm/vc4/vc4_validate.c
789
pkt_v = exec->shader_rec_v;
drivers/gpu/drm/vc4/vc4_validate.c
791
exec->shader_rec_u += packet_size;
drivers/gpu/drm/vc4/vc4_validate.c
798
exec->shader_rec_v += roundup(packet_size, 16);
drivers/gpu/drm/vc4/vc4_validate.c
799
exec->shader_rec_size -= packet_size;
drivers/gpu/drm/vc4/vc4_validate.c
802
if (src_handles[i] > exec->bo_count) {
drivers/gpu/drm/vc4/vc4_validate.c
807
bo[i] = to_drm_gem_dma_obj(exec->bo[src_handles[i]]);
drivers/gpu/drm/vc4/vc4_validate.c
812
bo[i] = vc4_use_bo(exec, src_handles[i]);
drivers/gpu/drm/vc4/vc4_validate.c
850
exec->uniforms_size) {
drivers/gpu/drm/vc4/vc4_validate.c
855
texture_handles_u = exec->uniforms_u;
drivers/gpu/drm/vc4/vc4_validate.c
859
memcpy(exec->uniforms_v, uniform_data_u,
drivers/gpu/drm/vc4/vc4_validate.c
865
if (!reloc_tex(exec,
drivers/gpu/drm/vc4/vc4_validate.c
882
((uint32_t *)exec->uniforms_v)[o] = exec->uniforms_p;
drivers/gpu/drm/vc4/vc4_validate.c
885
*(uint32_t *)(pkt_v + o + 4) = exec->uniforms_p;
drivers/gpu/drm/vc4/vc4_validate.c
887
exec->uniforms_u += validated_shader->uniforms_src_size;
drivers/gpu/drm/vc4/vc4_validate.c
888
exec->uniforms_v += validated_shader->uniforms_size;
drivers/gpu/drm/vc4/vc4_validate.c
889
exec->uniforms_p += validated_shader->uniforms_size;
drivers/gpu/drm/vc4/vc4_validate.c
930
struct vc4_exec_info *exec)
drivers/gpu/drm/vc4/vc4_validate.c
939
for (i = 0; i < exec->shader_state_count; i++) {
drivers/gpu/drm/vc4/vc4_validate.c
940
ret = validate_gl_shader_rec(dev, exec, &exec->shader_state[i]);
drivers/gpu/drm/xe/display/xe_fb_pin.c
280
struct drm_exec exec;
drivers/gpu/drm/xe/display/xe_fb_pin.c
308
xe_validation_guard(&ctx, &xe->val, &exec, (struct xe_val_flags) {.interruptible = true},
drivers/gpu/drm/xe/display/xe_fb_pin.c
310
ret = drm_exec_lock_obj(&exec, &bo->ttm.base);
drivers/gpu/drm/xe/display/xe_fb_pin.c
311
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/xe/display/xe_fb_pin.c
316
ret = xe_bo_migrate(bo, XE_PL_VRAM0, NULL, &exec);
drivers/gpu/drm/xe/display/xe_fb_pin.c
318
ret = xe_bo_validate(bo, NULL, true, &exec);
drivers/gpu/drm/xe/display/xe_fb_pin.c
319
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/xe/tests/xe_bo.c
136
struct drm_exec *exec = XE_VALIDATION_OPT_OUT;
drivers/gpu/drm/xe/tests/xe_bo.c
144
bo_flags, exec);
drivers/gpu/drm/xe/tests/xe_bo.c
154
test, exec);
drivers/gpu/drm/xe/tests/xe_bo.c
160
0xdeadbeefdeadbeefULL, test, exec);
drivers/gpu/drm/xe/tests/xe_bo.c
165
ret = ccs_test_migrate(tile, bo, true, 0ULL, 0ULL, test, exec);
drivers/gpu/drm/xe/tests/xe_bo.c
212
struct drm_exec *exec = XE_VALIDATION_OPT_OUT;
drivers/gpu/drm/xe/tests/xe_bo.c
223
bo_flags, exec);
drivers/gpu/drm/xe/tests/xe_bo.c
239
err = xe_bo_pin_external(external, false, exec);
drivers/gpu/drm/xe/tests/xe_bo.c
27
struct kunit *test, struct drm_exec *exec)
drivers/gpu/drm/xe/tests/xe_bo.c
297
err = xe_bo_validate(bo, bo->vm, false, exec);
drivers/gpu/drm/xe/tests/xe_bo.c
306
err = xe_bo_validate(external, NULL, false, exec);
drivers/gpu/drm/xe/tests/xe_bo.c
39
ret = xe_bo_validate(bo, NULL, false, exec);
drivers/gpu/drm/xe/tests/xe_bo.c
64
ret = xe_bo_evict(bo, exec);
drivers/gpu/drm/xe/tests/xe_dma_buf.c
159
struct drm_exec *exec = XE_VALIDATION_OPT_OUT;
drivers/gpu/drm/xe/tests/xe_dma_buf.c
164
err = xe_bo_validate(import_bo, NULL, false, exec);
drivers/gpu/drm/xe/tests/xe_dma_buf.c
177
check_residency(test, bo, import_bo, dmabuf, exec);
drivers/gpu/drm/xe/tests/xe_dma_buf.c
31
struct drm_exec *exec)
drivers/gpu/drm/xe/tests/xe_dma_buf.c
67
ret = xe_bo_evict(exported, exec);
drivers/gpu/drm/xe/tests/xe_dma_buf.c
82
ret = xe_bo_validate(imported, NULL, false, exec);
drivers/gpu/drm/xe/tests/xe_migrate.c
165
struct drm_exec *exec, struct kunit *test)
drivers/gpu/drm/xe/tests/xe_migrate.c
167
test_copy(m, bo, test, XE_BO_FLAG_SYSTEM, exec);
drivers/gpu/drm/xe/tests/xe_migrate.c
171
struct drm_exec *exec, struct kunit *test)
drivers/gpu/drm/xe/tests/xe_migrate.c
182
test_copy(m, bo, test, region, exec);
drivers/gpu/drm/xe/tests/xe_migrate.c
186
struct drm_exec *exec)
drivers/gpu/drm/xe/tests/xe_migrate.c
208
exec);
drivers/gpu/drm/xe/tests/xe_migrate.c
217
exec);
drivers/gpu/drm/xe/tests/xe_migrate.c
228
exec);
drivers/gpu/drm/xe/tests/xe_migrate.c
298
test_copy_sysmem(m, tiny, exec, test);
drivers/gpu/drm/xe/tests/xe_migrate.c
301
test_copy_vram(m, tiny, exec, test);
drivers/gpu/drm/xe/tests/xe_migrate.c
320
test_copy_sysmem(m, big, exec, test);
drivers/gpu/drm/xe/tests/xe_migrate.c
323
test_copy_vram(m, big, exec, test);
drivers/gpu/drm/xe/tests/xe_migrate.c
350
struct drm_exec *exec = XE_VALIDATION_OPT_OUT;
drivers/gpu/drm/xe/tests/xe_migrate.c
354
xe_migrate_sanity_test(m, test, exec);
drivers/gpu/drm/xe/tests/xe_migrate.c
496
struct drm_exec *exec, struct kunit *test)
drivers/gpu/drm/xe/tests/xe_migrate.c
515
ret = xe_bo_evict(vram_bo, exec);
drivers/gpu/drm/xe/tests/xe_migrate.c
544
ret = xe_bo_validate(vram_bo, NULL, false, exec);
drivers/gpu/drm/xe/tests/xe_migrate.c
642
struct drm_exec *exec;
drivers/gpu/drm/xe/tests/xe_migrate.c
657
exec = XE_VALIDATION_OPT_OUT;
drivers/gpu/drm/xe/tests/xe_migrate.c
659
ret = xe_bo_validate(sys_bo, NULL, false, exec);
drivers/gpu/drm/xe/tests/xe_migrate.c
684
ret = xe_bo_validate(ccs_bo, NULL, false, exec);
drivers/gpu/drm/xe/tests/xe_migrate.c
708
ret = xe_bo_validate(vram_bo, NULL, false, exec);
drivers/gpu/drm/xe/tests/xe_migrate.c
721
test_migrate(xe, tile, sys_bo, vram_bo, ccs_bo, exec, test);
drivers/gpu/drm/xe/tests/xe_migrate.c
73
struct kunit *test, u32 region, struct drm_exec *exec)
drivers/gpu/drm/xe/tests/xe_migrate.c
88
exec);
drivers/gpu/drm/xe/tests/xe_migrate.c
95
err = xe_bo_validate(remote, NULL, false, exec);
drivers/gpu/drm/xe/xe_bo.c
1196
struct drm_exec exec;
drivers/gpu/drm/xe/xe_bo.c
1200
xe_validation_guard(&ctx, &xe->val, &exec, (struct xe_val_flags) {.exclusive = true}, ret) {
drivers/gpu/drm/xe/xe_bo.c
1201
ret = drm_exec_lock_obj(&exec, &bo->ttm.base);
drivers/gpu/drm/xe/xe_bo.c
1202
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/xe/xe_bo.c
1223
XE_BO_FLAG_PINNED, &exec);
drivers/gpu/drm/xe/xe_bo.c
1225
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/xe/xe_bo.c
1333
struct drm_exec exec;
drivers/gpu/drm/xe/xe_bo.c
1338
xe_validation_guard(&ctx, &xe->val, &exec, (struct xe_val_flags) {.exclusive = true}, ret) {
drivers/gpu/drm/xe/xe_bo.c
1339
ret = drm_exec_lock_obj(&exec, &bo->ttm.base);
drivers/gpu/drm/xe/xe_bo.c
1340
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/xe/xe_bo.c
1364
XE_BO_FLAG_PINNED, &exec);
drivers/gpu/drm/xe/xe_bo.c
1366
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/xe/xe_bo.c
1804
struct drm_exec *exec)
drivers/gpu/drm/xe/xe_bo.c
1815
err = xe_bo_migrate(bo, XE_PL_TT, ctx, exec);
drivers/gpu/drm/xe/xe_bo.c
1939
struct drm_exec exec;
drivers/gpu/drm/xe/xe_bo.c
1979
xe_validation_guard(&ctx, &xe->val, &exec, (struct xe_val_flags) {.interruptible = true},
drivers/gpu/drm/xe/xe_bo.c
1987
err = drm_exec_lock_obj(&exec, &tbo->base);
drivers/gpu/drm/xe/xe_bo.c
1988
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/xe/xe_bo.c
1998
err = xe_bo_fault_migrate(bo, &tctx, &exec);
drivers/gpu/drm/xe/xe_bo.c
2000
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/xe/xe_bo.c
2136
u32 flags, struct drm_exec *exec)
drivers/gpu/drm/xe/xe_bo.c
2205
xe_validation_assert_exec(xe, exec, &bo->ttm.base);
drivers/gpu/drm/xe/xe_bo.c
2310
u64 alignment, struct drm_exec *exec)
drivers/gpu/drm/xe/xe_bo.c
2335
cpu_caching, type, flags, exec);
drivers/gpu/drm/xe/xe_bo.c
2370
exec);
drivers/gpu/drm/xe/xe_bo.c
2372
err = xe_ggtt_insert_bo(t->mem.ggtt, bo, exec);
drivers/gpu/drm/xe/xe_bo.c
2407
struct drm_exec *exec)
drivers/gpu/drm/xe/xe_bo.c
2410
flags, 0, exec);
drivers/gpu/drm/xe/xe_bo.c
2419
struct drm_exec exec;
drivers/gpu/drm/xe/xe_bo.c
2423
xe_validation_guard(&ctx, &xe->val, &exec, (struct xe_val_flags) {.interruptible = intr},
drivers/gpu/drm/xe/xe_bo.c
2426
cpu_caching, type, flags, alignment, &exec);
drivers/gpu/drm/xe/xe_bo.c
2427
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/xe/xe_bo.c
2456
u32 flags, struct drm_exec *exec)
drivers/gpu/drm/xe/xe_bo.c
2462
if (vm || exec) {
drivers/gpu/drm/xe/xe_bo.c
2463
xe_assert(xe, exec);
drivers/gpu/drm/xe/xe_bo.c
2466
flags, 0, exec);
drivers/gpu/drm/xe/xe_bo.c
2499
struct drm_exec exec;
drivers/gpu/drm/xe/xe_bo.c
2503
xe_validation_guard(&ctx, &xe->val, &exec, (struct xe_val_flags) {}, err) {
drivers/gpu/drm/xe/xe_bo.c
2505
0, type, flags, 0, &exec);
drivers/gpu/drm/xe/xe_bo.c
2507
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/xe/xe_bo.c
2513
err = xe_bo_pin(bo, &exec);
drivers/gpu/drm/xe/xe_bo.c
2517
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/xe/xe_bo.c
2531
u64 alignment, struct drm_exec *exec)
drivers/gpu/drm/xe/xe_bo.c
2544
alignment, exec);
drivers/gpu/drm/xe/xe_bo.c
2548
err = xe_bo_pin(bo, exec);
drivers/gpu/drm/xe/xe_bo.c
2593
struct drm_exec exec;
drivers/gpu/drm/xe/xe_bo.c
2597
xe_validation_guard(&ctx, &xe->val, &exec, (struct xe_val_flags) {.interruptible = intr},
drivers/gpu/drm/xe/xe_bo.c
2600
type, flags, alignment, &exec);
drivers/gpu/drm/xe/xe_bo.c
2602
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/xe/xe_bo.c
2634
struct drm_exec *exec)
drivers/gpu/drm/xe/xe_bo.c
2637
0, exec);
drivers/gpu/drm/xe/xe_bo.c
2772
int xe_bo_pin_external(struct xe_bo *bo, bool in_place, struct drm_exec *exec)
drivers/gpu/drm/xe/xe_bo.c
2782
err = xe_bo_validate(bo, NULL, false, exec);
drivers/gpu/drm/xe/xe_bo.c
2815
int xe_bo_pin(struct xe_bo *bo, struct drm_exec *exec)
drivers/gpu/drm/xe/xe_bo.c
2837
err = xe_bo_validate(bo, NULL, false, exec);
drivers/gpu/drm/xe/xe_bo.c
2941
struct drm_exec *exec)
drivers/gpu/drm/xe/xe_bo.c
2963
xe_validation_assert_exec(xe_bo_device(bo), exec, &bo->ttm.base);
drivers/gpu/drm/xe/xe_bo.c
3160
struct drm_exec exec;
drivers/gpu/drm/xe/xe_bo.c
3243
xe_validation_guard(&ctx, &xe->val, &exec, (struct xe_val_flags) {.interruptible = true},
drivers/gpu/drm/xe/xe_bo.c
3246
err = xe_vm_drm_exec_lock(vm, &exec);
drivers/gpu/drm/xe/xe_bo.c
3247
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/xe/xe_bo.c
3252
bo_flags, &exec);
drivers/gpu/drm/xe/xe_bo.c
3253
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/xe/xe_bo.c
3425
struct drm_exec *exec)
drivers/gpu/drm/xe/xe_bo.c
3465
xe_validation_assert_exec(xe_bo_device(bo), exec, &bo->ttm.base);
drivers/gpu/drm/xe/xe_bo.c
3479
int xe_bo_evict(struct xe_bo *bo, struct drm_exec *exec)
drivers/gpu/drm/xe/xe_bo.h
103
struct drm_exec *exec);
drivers/gpu/drm/xe/xe_bo.h
105
u16 cpu_caching, u32 flags, struct drm_exec *exec);
drivers/gpu/drm/xe/xe_bo.h
109
struct drm_exec *exec);
drivers/gpu/drm/xe/xe_bo.h
201
int xe_bo_pin_external(struct xe_bo *bo, bool in_place, struct drm_exec *exec);
drivers/gpu/drm/xe/xe_bo.h
202
int xe_bo_pin(struct xe_bo *bo, struct drm_exec *exec);
drivers/gpu/drm/xe/xe_bo.h
206
struct drm_exec *exec);
drivers/gpu/drm/xe/xe_bo.h
290
struct drm_exec *exec);
drivers/gpu/drm/xe/xe_bo.h
291
int xe_bo_evict(struct xe_bo *bo, struct drm_exec *exec);
drivers/gpu/drm/xe/xe_bo.h
99
u32 flags, struct drm_exec *exec);
drivers/gpu/drm/xe/xe_dma_buf.c
107
struct drm_exec *exec = XE_VALIDATION_UNSUPPORTED;
drivers/gpu/drm/xe/xe_dma_buf.c
116
r = xe_bo_migrate(bo, XE_PL_TT, NULL, exec);
drivers/gpu/drm/xe/xe_dma_buf.c
118
r = xe_bo_validate(bo, NULL, false, exec);
drivers/gpu/drm/xe/xe_dma_buf.c
178
struct drm_exec exec;
drivers/gpu/drm/xe/xe_dma_buf.c
185
xe_validation_guard(&ctx, &xe_bo_device(bo)->val, &exec, (struct xe_val_flags) {}, ret) {
drivers/gpu/drm/xe/xe_dma_buf.c
186
ret = drm_exec_lock_obj(&exec, &bo->ttm.base);
drivers/gpu/drm/xe/xe_dma_buf.c
187
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/xe/xe_dma_buf.c
191
ret = xe_bo_migrate(bo, XE_PL_TT, NULL, &exec);
drivers/gpu/drm/xe/xe_dma_buf.c
192
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/xe/xe_dma_buf.c
249
struct drm_exec exec;
drivers/gpu/drm/xe/xe_dma_buf.c
258
xe_validation_guard(&ctx, &xe->val, &exec, (struct xe_val_flags) {}, ret) {
drivers/gpu/drm/xe/xe_dma_buf.c
259
ret = drm_exec_lock_obj(&exec, dummy_obj);
drivers/gpu/drm/xe/xe_dma_buf.c
260
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/xe/xe_dma_buf.c
266
ttm_bo_type_sg, XE_BO_FLAG_SYSTEM, &exec);
drivers/gpu/drm/xe/xe_dma_buf.c
267
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/xe/xe_dma_buf.c
283
struct drm_exec *exec = XE_VALIDATION_UNSUPPORTED;
drivers/gpu/drm/xe/xe_dma_buf.c
285
XE_WARN_ON(xe_bo_evict(bo, exec));
drivers/gpu/drm/xe/xe_dma_buf.c
55
struct drm_exec *exec = XE_VALIDATION_UNSUPPORTED;
drivers/gpu/drm/xe/xe_dma_buf.c
77
ret = xe_bo_migrate(bo, XE_PL_TT, NULL, exec);
drivers/gpu/drm/xe/xe_dma_buf.c
87
ret = xe_bo_pin_external(bo, !allow_vram, exec);
drivers/gpu/drm/xe/xe_exec.c
105
xe_vm_set_validation_exec(vm, &vm_exec->exec);
drivers/gpu/drm/xe/xe_exec.c
106
ret = xe_vm_validate_rebind(vm, &vm_exec->exec, 1);
drivers/gpu/drm/xe/xe_exec.c
122
struct drm_exec *exec = &vm_exec.exec;
drivers/gpu/drm/xe/xe_exec.c
325
drm_gpuvm_resv_add_fence(&vm->gpuvm, exec, &job->drm.s_fence->finished,
drivers/gpu/drm/xe/xe_ggtt.c
863
u64 start, u64 end, struct drm_exec *exec)
drivers/gpu/drm/xe/xe_ggtt.c
878
err = xe_bo_validate(bo, NULL, false, exec);
drivers/gpu/drm/xe/xe_ggtt.c
926
u64 start, u64 end, struct drm_exec *exec)
drivers/gpu/drm/xe/xe_ggtt.c
928
return __xe_ggtt_insert_bo_at(ggtt, bo, start, end, exec);
drivers/gpu/drm/xe/xe_ggtt.c
940
struct drm_exec *exec)
drivers/gpu/drm/xe/xe_ggtt.c
942
return __xe_ggtt_insert_bo_at(ggtt, bo, 0, U64_MAX, exec);
drivers/gpu/drm/xe/xe_ggtt.h
39
int xe_ggtt_insert_bo(struct xe_ggtt *ggtt, struct xe_bo *bo, struct drm_exec *exec);
drivers/gpu/drm/xe/xe_ggtt.h
41
u64 start, u64 end, struct drm_exec *exec);
drivers/gpu/drm/xe/xe_gt_sriov_pf_migration.c
524
struct drm_exec exec;
drivers/gpu/drm/xe/xe_gt_sriov_pf_migration.c
527
drm_exec_init(&exec, 0, 0);
drivers/gpu/drm/xe/xe_gt_sriov_pf_migration.c
528
drm_exec_until_all_locked(&exec) {
drivers/gpu/drm/xe/xe_gt_sriov_pf_migration.c
529
err = drm_exec_lock_obj(&exec, &vram->ttm.base);
drivers/gpu/drm/xe/xe_gt_sriov_pf_migration.c
530
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/xe/xe_gt_sriov_pf_migration.c
536
err = drm_exec_lock_obj(&exec, &sysmem->ttm.base);
drivers/gpu/drm/xe/xe_gt_sriov_pf_migration.c
537
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/xe/xe_gt_sriov_pf_migration.c
548
drm_exec_fini(&exec);
drivers/gpu/drm/xe/xe_migrate.c
187
struct xe_vm *vm, struct drm_exec *exec)
drivers/gpu/drm/xe/xe_migrate.c
214
XE_BO_FLAG_PAGETABLE, exec);
drivers/gpu/drm/xe/xe_migrate.c
417
struct drm_exec exec;
drivers/gpu/drm/xe/xe_migrate.c
420
xe_validation_guard(&ctx, &xe->val, &exec, (struct xe_val_flags) {}, err) {
drivers/gpu/drm/xe/xe_migrate.c
421
err = xe_vm_drm_exec_lock(vm, &exec);
drivers/gpu/drm/xe/xe_migrate.c
422
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/xe/xe_migrate.c
423
err = xe_migrate_prepare_vm(tile, m, vm, &exec);
drivers/gpu/drm/xe/xe_migrate.c
424
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/xe/xe_pagefault.c
104
xe_validation_ctx_init(&ctx, &vm->xe->val, &exec, (struct xe_val_flags) {});
drivers/gpu/drm/xe/xe_pagefault.c
105
drm_exec_until_all_locked(&exec) {
drivers/gpu/drm/xe/xe_pagefault.c
106
err = xe_pagefault_begin(&exec, vma, tile->mem.vram,
drivers/gpu/drm/xe/xe_pagefault.c
108
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/xe/xe_pagefault.c
115
xe_vm_set_validation_exec(vm, &exec);
drivers/gpu/drm/xe/xe_pagefault.c
48
static int xe_pagefault_begin(struct drm_exec *exec, struct xe_vma *vma,
drivers/gpu/drm/xe/xe_pagefault.c
55
err = xe_vm_lock_vma(exec, vma);
drivers/gpu/drm/xe/xe_pagefault.c
62
return need_vram_move ? xe_bo_migrate(bo, vram->placement, NULL, exec) :
drivers/gpu/drm/xe/xe_pagefault.c
63
xe_bo_validate(bo, vm, true, exec);
drivers/gpu/drm/xe/xe_pagefault.c
72
struct drm_exec exec;
drivers/gpu/drm/xe/xe_pt.c
104
unsigned int level, struct drm_exec *exec)
drivers/gpu/drm/xe/xe_pt.c
129
drm_WARN_ON(&vm->xe->drm, IS_ERR_OR_NULL(exec));
drivers/gpu/drm/xe/xe_pt.c
132
bo_flags, exec);
drivers/gpu/drm/xe/xe_pt.h
33
unsigned int level, struct drm_exec *exec);
drivers/gpu/drm/xe/xe_pxp_submit.c
113
xe_validation_guard(&ctx, &xe->val, &exec, (struct xe_val_flags){}, err) {
drivers/gpu/drm/xe/xe_pxp_submit.c
114
err = xe_vm_drm_exec_lock(vm, &exec);
drivers/gpu/drm/xe/xe_pxp_submit.c
115
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/xe/xe_pxp_submit.c
122
XE_BO_FLAG_NEEDS_UC, &exec);
drivers/gpu/drm/xe/xe_pxp_submit.c
123
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/xe/xe_pxp_submit.c
93
struct drm_exec exec;
drivers/gpu/drm/xe/xe_svm.c
1039
struct drm_exec exec;
drivers/gpu/drm/xe/xe_svm.c
1048
xe_validation_guard(&vctx, &xe->val, &exec, (struct xe_val_flags) {}, err) {
drivers/gpu/drm/xe/xe_svm.c
1052
XE_BO_FLAG_CPU_ADDR_MIRROR, &exec);
drivers/gpu/drm/xe/xe_svm.c
1053
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/xe/xe_svm.c
1200
struct drm_exec exec;
drivers/gpu/drm/xe/xe_svm.c
1309
xe_validation_guard(&vctx, &vm->xe->val, &exec, (struct xe_val_flags) {}, err) {
drivers/gpu/drm/xe/xe_svm.c
1310
err = xe_vm_drm_exec_lock(vm, &exec);
drivers/gpu/drm/xe/xe_svm.c
1311
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/xe/xe_svm.c
1313
xe_vm_set_validation_exec(vm, &exec);
drivers/gpu/drm/xe/xe_svm.c
1317
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/xe/xe_sync.c
122
bool exec = flags & SYNC_PARSE_FLAG_EXEC;
drivers/gpu/drm/xe/xe_sync.c
200
if (exec) {
drivers/gpu/drm/xe/xe_validation.c
126
struct drm_exec *exec, const struct xe_val_flags flags)
drivers/gpu/drm/xe/xe_validation.c
130
ctx->exec = exec;
drivers/gpu/drm/xe/xe_validation.c
146
if (exec) {
drivers/gpu/drm/xe/xe_validation.c
151
drm_exec_init(exec, ctx->exec_flags, ctx->nr);
drivers/gpu/drm/xe/xe_validation.c
165
static bool xe_validation_contention_injected(struct drm_exec *exec)
drivers/gpu/drm/xe/xe_validation.c
167
return !!exec->ticket.contending_lock;
drivers/gpu/drm/xe/xe_validation.c
172
static bool xe_validation_contention_injected(struct drm_exec *exec)
drivers/gpu/drm/xe/xe_validation.c
183
xe_validation_contention_injected(ctx->exec)) ||
drivers/gpu/drm/xe/xe_validation.c
213
ctx->exec = &vm_exec->exec;
drivers/gpu/drm/xe/xe_validation.c
243
if (ctx->exec)
drivers/gpu/drm/xe/xe_validation.c
244
drm_exec_fini(ctx->exec);
drivers/gpu/drm/xe/xe_validation.c
266
drm_exec_fini(ctx->exec);
drivers/gpu/drm/xe/xe_validation.c
272
drm_exec_init(ctx->exec, ctx->exec_flags, ctx->nr);
drivers/gpu/drm/xe/xe_validation.c
28
const struct drm_exec *exec,
drivers/gpu/drm/xe/xe_validation.c
31
xe_assert(xe, exec);
drivers/gpu/drm/xe/xe_validation.c
32
if (IS_ERR(exec)) {
drivers/gpu/drm/xe/xe_validation.c
33
switch (PTR_ERR(exec)) {
drivers/gpu/drm/xe/xe_validation.h
114
struct drm_exec *exec;
drivers/gpu/drm/xe/xe_validation.h
125
struct drm_exec *exec, const struct xe_val_flags flags);
drivers/gpu/drm/xe/xe_validation.h
62
void xe_validation_assert_exec(const struct xe_device *xe, const struct drm_exec *exec,
drivers/gpu/drm/xe/xe_vm.c
1158
int xe_vm_lock_vma(struct drm_exec *exec, struct xe_vma *vma)
drivers/gpu/drm/xe/xe_vm.c
1166
err = drm_exec_lock_obj(exec, xe_vm_obj(vm));
drivers/gpu/drm/xe/xe_vm.c
1168
err = drm_exec_lock_obj(exec, &bo->ttm.base);
drivers/gpu/drm/xe/xe_vm.c
1177
struct drm_exec exec;
drivers/gpu/drm/xe/xe_vm.c
1180
xe_validation_guard(&ctx, &xe->val, &exec, (struct xe_val_flags) {}, err) {
drivers/gpu/drm/xe/xe_vm.c
1181
err = xe_vm_lock_vma(&exec, vma);
drivers/gpu/drm/xe/xe_vm.c
1182
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/xe/xe_vm.c
1422
struct xe_vm *vm, struct drm_exec *exec)
drivers/gpu/drm/xe/xe_vm.c
1428
vm->scratch_pt[id][i] = xe_pt_create(vm, tile, i, exec);
drivers/gpu/drm/xe/xe_vm.c
1481
struct drm_exec exec;
drivers/gpu/drm/xe/xe_vm.c
1570
xe_validation_guard(&ctx, &xe->val, &exec, (struct xe_val_flags) {.interruptible = true},
drivers/gpu/drm/xe/xe_vm.c
1572
err = xe_vm_drm_exec_lock(vm, &exec);
drivers/gpu/drm/xe/xe_vm.c
1573
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/xe/xe_vm.c
1584
&exec);
drivers/gpu/drm/xe/xe_vm.c
1589
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/xe/xe_vm.c
1602
err = xe_vm_create_scratch(xe, tile, vm, &exec);
drivers/gpu/drm/xe/xe_vm.c
1606
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/xe/xe_vm.c
199
struct drm_exec *exec)
drivers/gpu/drm/xe/xe_vm.c
209
drm_gpuvm_resv_add_fence(&vm->gpuvm, exec, q->lr.pfence,
drivers/gpu/drm/xe/xe_vm.c
221
struct drm_exec *exec = &vm_exec.exec;
drivers/gpu/drm/xe/xe_vm.c
2432
struct drm_exec exec;
drivers/gpu/drm/xe/xe_vm.c
2440
xe_validation_guard(&ctx, &vm->xe->val, &exec,
drivers/gpu/drm/xe/xe_vm.c
2443
err = drm_exec_lock_obj(&exec, xe_vm_obj(vm));
drivers/gpu/drm/xe/xe_vm.c
2444
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/xe/xe_vm.c
2447
err = drm_exec_lock_obj(&exec, &bo->ttm.base);
drivers/gpu/drm/xe/xe_vm.c
2448
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/xe/xe_vm.c
247
drm_gpuvm_resv_add_fence(&vm->gpuvm, exec, pfence,
drivers/gpu/drm/xe/xe_vm.c
2889
static int vma_lock_and_validate(struct drm_exec *exec, struct xe_vma *vma,
drivers/gpu/drm/xe/xe_vm.c
2898
err = drm_exec_lock_obj(exec, &bo->ttm.base);
drivers/gpu/drm/xe/xe_vm.c
2902
res_evict, exec);
drivers/gpu/drm/xe/xe_vm.c
2978
static int op_lock_and_prep(struct drm_exec *exec, struct xe_vm *vm,
drivers/gpu/drm/xe/xe_vm.c
2994
err = vma_lock_and_validate(exec, op->map.vma,
drivers/gpu/drm/xe/xe_vm.c
3004
err = vma_lock_and_validate(exec,
drivers/gpu/drm/xe/xe_vm.c
3008
err = vma_lock_and_validate(exec, op->remap.prev,
drivers/gpu/drm/xe/xe_vm.c
3011
err = vma_lock_and_validate(exec, op->remap.next,
drivers/gpu/drm/xe/xe_vm.c
3019
err = vma_lock_and_validate(exec,
drivers/gpu/drm/xe/xe_vm.c
3034
err = vma_lock_and_validate(exec,
drivers/gpu/drm/xe/xe_vm.c
3041
exec);
drivers/gpu/drm/xe/xe_vm.c
3070
static int vm_bind_ioctl_ops_lock_and_prep(struct drm_exec *exec,
drivers/gpu/drm/xe/xe_vm.c
3077
err = drm_exec_lock_obj(exec, xe_vm_obj(vm));
drivers/gpu/drm/xe/xe_vm.c
3082
err = op_lock_and_prep(exec, vm, vops, op);
drivers/gpu/drm/xe/xe_vm.c
326
static int xe_gpuvm_validate(struct drm_gpuvm_bo *vm_bo, struct drm_exec *exec)
drivers/gpu/drm/xe/xe_vm.c
3318
struct drm_exec exec;
drivers/gpu/drm/xe/xe_vm.c
3324
xe_validation_guard(&ctx, &vm->xe->val, &exec,
drivers/gpu/drm/xe/xe_vm.c
3329
err = vm_bind_ioctl_ops_lock_and_prep(&exec, vm, vops);
drivers/gpu/drm/xe/xe_vm.c
3330
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/xe/xe_vm.c
3335
xe_vm_set_validation_exec(vm, &exec);
drivers/gpu/drm/xe/xe_vm.c
340
ret = xe_bo_validate(gem_to_xe_bo(vm_bo->obj), vm, false, exec);
drivers/gpu/drm/xe/xe_vm.c
363
int xe_vm_validate_rebind(struct xe_vm *vm, struct drm_exec *exec,
drivers/gpu/drm/xe/xe_vm.c
371
ret = drm_gpuvm_validate(&vm->gpuvm, exec);
drivers/gpu/drm/xe/xe_vm.c
380
drm_exec_for_each_locked_object(exec, index, obj) {
drivers/gpu/drm/xe/xe_vm.c
389
static int xe_preempt_work_begin(struct drm_exec *exec, struct xe_vm *vm,
drivers/gpu/drm/xe/xe_vm.c
394
err = drm_gpuvm_prepare_vm(&vm->gpuvm, exec, 0);
drivers/gpu/drm/xe/xe_vm.c
409
err = drm_gpuvm_prepare_objects(&vm->gpuvm, exec, 0);
drivers/gpu/drm/xe/xe_vm.c
423
return xe_vm_validate_rebind(vm, exec, vm->preempt.num_exec_queues);
drivers/gpu/drm/xe/xe_vm.c
457
struct drm_exec exec;
drivers/gpu/drm/xe/xe_vm.c
489
err = xe_validation_ctx_init(&ctx, &vm->xe->val, &exec,
drivers/gpu/drm/xe/xe_vm.c
494
drm_exec_until_all_locked(&exec) {
drivers/gpu/drm/xe/xe_vm.c
497
err = xe_preempt_work_begin(&exec, vm, &done);
drivers/gpu/drm/xe/xe_vm.c
498
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/xe/xe_vm.c
510
xe_vm_set_validation_exec(vm, &exec);
drivers/gpu/drm/xe/xe_vm.c
545
resume_and_reinstall_preempt_fences(vm, &exec);
drivers/gpu/drm/xe/xe_vm.c
58
int xe_vm_drm_exec_lock(struct xe_vm *vm, struct drm_exec *exec)
drivers/gpu/drm/xe/xe_vm.c
60
return drm_exec_lock_obj(exec, xe_vm_obj(vm));
drivers/gpu/drm/xe/xe_vm.h
266
int xe_vm_lock_vma(struct drm_exec *exec, struct xe_vma *vma);
drivers/gpu/drm/xe/xe_vm.h
268
int xe_vm_validate_rebind(struct xe_vm *vm, struct drm_exec *exec,
drivers/gpu/drm/xe/xe_vm.h
296
int xe_vm_drm_exec_lock(struct xe_vm *vm, struct drm_exec *exec);
drivers/gpu/drm/xe/xe_vm.h
377
static inline void xe_vm_set_validation_exec(struct xe_vm *vm, struct drm_exec *exec)
drivers/gpu/drm/xe/xe_vm.h
380
xe_assert(vm->xe, !!exec ^ !!vm->validation._exec);
drivers/gpu/drm/xe/xe_vm.h
381
vm->validation._exec = exec;
drivers/gpu/drm/xe/xe_vm_madvise.c
422
struct drm_exec exec;
drivers/gpu/drm/xe/xe_vm_madvise.c
467
drm_exec_init(&exec, DRM_EXEC_IGNORE_DUPLICATES | DRM_EXEC_INTERRUPTIBLE_WAIT, 0);
drivers/gpu/drm/xe/xe_vm_madvise.c
468
drm_exec_until_all_locked(&exec) {
drivers/gpu/drm/xe/xe_vm_madvise.c
474
err = drm_exec_lock_obj(&exec, &bo->ttm.base);
drivers/gpu/drm/xe/xe_vm_madvise.c
475
drm_exec_retry_on_contention(&exec);
drivers/gpu/drm/xe/xe_vm_madvise.c
500
drm_exec_fini(&exec);
drivers/leds/leds-lp55xx-common.c
141
u8 mode, exec;
drivers/leds/leds-lp55xx-common.c
149
ret = lp55xx_read(chip, cfg->reg_exec.addr, &exec);
drivers/leds/leds-lp55xx-common.c
160
exec &= ~LP55xx_EXEC_ENGn_MASK(i, cfg->reg_exec.shift);
drivers/leds/leds-lp55xx-common.c
161
exec |= LP55xx_EXEC_RUN_ENG << LP55xx_EXEC_ENGn_SHIFT(i, cfg->reg_exec.shift);
drivers/leds/leds-lp55xx-common.c
166
lp55xx_write(chip, cfg->reg_exec.addr, exec);
drivers/mtd/nand/raw/nand_base.c
2557
ret = pattern->exec(chip, &ctx.subop);
drivers/mtd/nand/raw/qcom_nandc.c
1718
nandc->regs->exec = cpu_to_le32(1);
drivers/mtd/nand/raw/qcom_nandc.c
1721
qcom_write_reg_dma(nandc, &nandc->regs->exec, NAND_EXEC_CMD, 1, NAND_BAM_NEXT_SGL);
drivers/mtd/nand/raw/qcom_nandc.c
1778
nandc->regs->exec = cpu_to_le32(1);
drivers/mtd/nand/raw/qcom_nandc.c
1781
qcom_write_reg_dma(nandc, &nandc->regs->exec, NAND_EXEC_CMD, 1, NAND_BAM_NEXT_SGL);
drivers/mtd/nand/raw/qcom_nandc.c
1835
nandc->regs->exec = cpu_to_le32(1);
drivers/mtd/nand/raw/qcom_nandc.c
1841
qcom_write_reg_dma(nandc, &nandc->regs->exec, NAND_EXEC_CMD, 1, NAND_BAM_NEXT_SGL);
drivers/mtd/nand/raw/qcom_nandc.c
1912
nandc->regs->exec = cpu_to_le32(1);
drivers/mtd/nand/raw/qcom_nandc.c
297
nandc->regs->exec = cpu_to_le32(1);
drivers/mtd/nand/raw/qcom_nandc.c
341
qcom_write_reg_dma(nandc, &nandc->regs->exec, NAND_EXEC_CMD, 1, NAND_BAM_NEXT_SGL);
drivers/mtd/nand/raw/qcom_nandc.c
388
qcom_write_reg_dma(nandc, &nandc->regs->exec, NAND_EXEC_CMD, 1, NAND_BAM_NEXT_SGL);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
57
exe_q_execute exec,
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
74
o->execute = exec;
drivers/net/ethernet/brocade/bna/bfa_ioc.c
1162
if (swab32(fwhdr.exec) == BFI_FWBOOT_TYPE_NORMAL) {
drivers/net/ethernet/brocade/bna/bfi.h
257
u32 exec; /*!< exec vector */
drivers/net/wireless/realtek/rtlwifi/base.c
2325
void rtl_c2hcmd_launcher(struct ieee80211_hw *hw, int exec)
drivers/net/wireless/realtek/rtlwifi/base.c
2344
if (exec)
drivers/net/wireless/realtek/rtlwifi/base.h
111
void rtl_c2hcmd_launcher(struct ieee80211_hw *hw, int exec);
drivers/sbus/char/oradax.c
860
ctx->result.exec.status = DAX_SUBMIT_ERR_THR_INIT;
drivers/sbus/char/oradax.c
867
ctx->result.exec.status = DAX_SUBMIT_ERR_NO_CA_AVAIL;
drivers/sbus/char/oradax.c
877
ctx->result.exec.status = DAX_SUBMIT_ERR_CCB_ARR_MMU_MISS;
drivers/sbus/char/oradax.c
885
ctx->result.exec.status = DAX_SUBMIT_ERR_NO_CA_AVAIL;
drivers/sbus/char/oradax.c
891
ctx->result.exec.status = dax_preprocess_usr_ccbs(ctx, idx, nccbs);
drivers/sbus/char/oradax.c
892
if (ctx->result.exec.status != DAX_SUBMIT_OK)
drivers/sbus/char/oradax.c
895
ctx->result.exec.status = dax_lock_pages(ctx, idx, nccbs,
drivers/sbus/char/oradax.c
896
&ctx->result.exec.status_data);
drivers/sbus/char/oradax.c
897
if (ctx->result.exec.status != DAX_SUBMIT_OK)
drivers/sbus/char/oradax.c
905
&accepted_len, &ctx->result.exec.status_data);
drivers/sbus/char/oradax.c
918
ctx->result.exec.status = DAX_SUBMIT_OK;
drivers/sbus/char/oradax.c
926
ctx->result.exec.status = DAX_SUBMIT_ERR_WOULDBLOCK;
drivers/sbus/char/oradax.c
934
ctx->result.exec.status = DAX_SUBMIT_ERR_NOMAP;
drivers/sbus/char/oradax.c
944
ctx->result.exec.status = DAX_SUBMIT_ERR_CCB_INVAL;
drivers/sbus/char/oradax.c
953
ctx->result.exec.status = DAX_SUBMIT_ERR_NOACCESS;
drivers/sbus/char/oradax.c
962
ctx->result.exec.status = DAX_SUBMIT_ERR_UNAVAIL;
drivers/sbus/char/oradax.c
965
ctx->result.exec.status = DAX_SUBMIT_ERR_INTERNAL;
drivers/sbus/char/oradax.c
982
hv_rv, accepted_len, ctx->result.exec.status_data,
drivers/sbus/char/oradax.c
983
ctx->result.exec.status);
drivers/scsi/bfa/bfa_ioc.c
724
if (swab32(fwhdr.exec) == BFI_FWBOOT_TYPE_NORMAL) {
drivers/scsi/bfa/bfa_ioc.c
741
bfa_trc(iocpf->ioc, swab32(fwhdr.exec));
drivers/scsi/bfa/bfi.h
350
u32 exec; /* exec vector */
drivers/spi/spi-qpic-snand.c
1018
qcom_write_reg_dma(snandc, &snandc->regs->exec, NAND_EXEC_CMD, 1, NAND_BAM_NEXT_SGL);
drivers/spi/spi-qpic-snand.c
1057
snandc->regs->exec = cpu_to_le32(1);
drivers/spi/spi-qpic-snand.c
1140
snandc->regs->exec = cpu_to_le32(1);
drivers/spi/spi-qpic-snand.c
1215
snandc->regs->exec = cpu_to_le32(1);
drivers/spi/spi-qpic-snand.c
1357
snandc->regs->exec = cpu_to_le32(1);
drivers/spi/spi-qpic-snand.c
1362
qcom_write_reg_dma(snandc, &snandc->regs->exec, NAND_EXEC_CMD, 1, NAND_BAM_NEXT_SGL);
drivers/spi/spi-qpic-snand.c
492
qcom_write_reg_dma(snandc, &snandc->regs->exec, NAND_EXEC_CMD, 1, NAND_BAM_NEXT_SGL);
drivers/spi/spi-qpic-snand.c
520
snandc->regs->exec = cpu_to_le32(1);
drivers/spi/spi-qpic-snand.c
524
qcom_write_reg_dma(snandc, &snandc->regs->exec, NAND_EXEC_CMD, 1, NAND_BAM_NEXT_SGL);
drivers/spi/spi-qpic-snand.c
554
qcom_write_reg_dma(snandc, &snandc->regs->exec, NAND_EXEC_CMD, 1, NAND_BAM_NEXT_SGL);
drivers/spi/spi-qpic-snand.c
604
snandc->regs->exec = cpu_to_le32(1);
drivers/spi/spi-qpic-snand.c
737
snandc->regs->exec = cpu_to_le32(1);
drivers/spi/spi-qpic-snand.c
851
snandc->regs->exec = cpu_to_le32(1);
drivers/spi/spi-qpic-snand.c
940
snandc->regs->exec = cpu_to_le32(1);
drivers/spi/spi-wpcm-fiu.c
137
int (*exec)(struct spi_mem *mem, const struct spi_mem_op *op);
drivers/spi/spi-wpcm-fiu.c
301
{ .match = wpcm_fiu_normal_match, .exec = wpcm_fiu_normal_exec },
drivers/spi/spi-wpcm-fiu.c
302
{ .match = wpcm_fiu_fast_read_match, .exec = wpcm_fiu_fast_read_exec },
drivers/spi/spi-wpcm-fiu.c
303
{ .match = wpcm_fiu_4ba_match, .exec = wpcm_fiu_4ba_exec },
drivers/spi/spi-wpcm-fiu.c
304
{ .match = wpcm_fiu_rdid_match, .exec = wpcm_fiu_rdid_exec },
drivers/spi/spi-wpcm-fiu.c
305
{ .match = wpcm_fiu_dummy_match, .exec = wpcm_fiu_dummy_exec },
drivers/spi/spi-wpcm-fiu.c
360
return shape->exec(mem, op);
fs/binfmt_elf.c
165
create_elf_tables(struct linux_binprm *bprm, const struct elfhdr *exec,
fs/binfmt_elf.c
255
NEW_AUX_ENT(AT_PHNUM, exec->e_phnum);
fs/binfmt_elf.c
276
NEW_AUX_ENT(AT_EXECFN, bprm->exec);
fs/binfmt_elf.c
309
bprm->exec = (unsigned long)sp; /* XXX: PARISC HACK */
fs/binfmt_elf_fdpic.c
657
NEW_AUX_ENT(AT_EXECFN, bprm->exec);
fs/exec.c
1075
void __set_task_comm(struct task_struct *tsk, const char *buf, bool exec)
fs/exec.c
1082
perf_event_comm(tsk, exec);
fs/exec.c
1692
struct file *exec;
fs/exec.c
1702
exec = bprm->file;
fs/exec.c
1706
exe_file_allow_write_access(exec);
fs/exec.c
1709
fput(exec);
fs/exec.c
1712
bprm->executable = exec;
fs/exec.c
1714
fput(exec);
fs/exec.c
1820
bprm->exec = bprm->p;
fs/exec.c
1882
bprm->exec = bprm->p;
fs/exec.c
648
bprm->exec -= stack_shift;
fs/exec.c
750
bprm->exec += *sp_location - MAX_ARG_PAGES * PAGE_SIZE;
include/drm/drm_exec.h
105
#define drm_exec_until_all_locked(exec) \
include/drm/drm_exec.h
110
drm_exec_cleanup(exec); \
include/drm/drm_exec.h
120
#define drm_exec_retry_on_contention(exec) \
include/drm/drm_exec.h
122
if (unlikely(drm_exec_is_contended(exec))) \
include/drm/drm_exec.h
133
static inline bool drm_exec_is_contended(struct drm_exec *exec)
include/drm/drm_exec.h
135
return !!exec->contended;
include/drm/drm_exec.h
138
void drm_exec_init(struct drm_exec *exec, u32 flags, unsigned nr);
include/drm/drm_exec.h
139
void drm_exec_fini(struct drm_exec *exec);
include/drm/drm_exec.h
140
bool drm_exec_cleanup(struct drm_exec *exec);
include/drm/drm_exec.h
141
int drm_exec_lock_obj(struct drm_exec *exec, struct drm_gem_object *obj);
include/drm/drm_exec.h
142
void drm_exec_unlock_obj(struct drm_exec *exec, struct drm_gem_object *obj);
include/drm/drm_exec.h
143
int drm_exec_prepare_obj(struct drm_exec *exec, struct drm_gem_object *obj,
include/drm/drm_exec.h
145
int drm_exec_prepare_array(struct drm_exec *exec,
include/drm/drm_exec.h
63
drm_exec_obj(struct drm_exec *exec, unsigned long index)
include/drm/drm_exec.h
65
return index < exec->num_objects ? exec->objects[index] : NULL;
include/drm/drm_exec.h
76
#define drm_exec_for_each_locked_object(exec, index, obj) \
include/drm/drm_exec.h
77
for ((index) = 0; ((obj) = drm_exec_obj(exec, index)); ++(index))
include/drm/drm_exec.h
90
#define drm_exec_for_each_locked_object_reverse(exec, index, obj) \
include/drm/drm_exec.h
91
for ((index) = (exec)->num_objects - 1; \
include/drm/drm_exec.h
92
((obj) = drm_exec_obj(exec, index)); --(index))
include/drm/drm_gpuvm.h
1206
struct drm_exec *exec);
include/drm/drm_gpuvm.h
1260
struct drm_exec *exec, unsigned int num_fences,
include/drm/drm_gpuvm.h
1263
int drm_gpuvm_sm_unmap_exec_lock(struct drm_gpuvm *gpuvm, struct drm_exec *exec,
include/drm/drm_gpuvm.h
537
struct drm_exec exec;
include/drm/drm_gpuvm.h
574
struct drm_exec *exec,
include/drm/drm_gpuvm.h
578
struct drm_exec *exec,
include/drm/drm_gpuvm.h
582
struct drm_exec *exec,
include/drm/drm_gpuvm.h
607
drm_exec_fini(&vm_exec->exec);
include/drm/drm_gpuvm.h
610
int drm_gpuvm_validate(struct drm_gpuvm *gpuvm, struct drm_exec *exec);
include/drm/drm_gpuvm.h
612
struct drm_exec *exec,
include/drm/drm_gpuvm.h
632
drm_gpuvm_resv_add_fence(vm_exec->vm, &vm_exec->exec, fence,
include/drm/drm_gpuvm.h
647
return drm_gpuvm_validate(vm_exec->vm, &vm_exec->exec);
include/linux/binfmts.h
67
unsigned long exec;
include/linux/mtd/nand-qpic-common.h
301
__le32 exec;
include/linux/mtd/rawnand.h
966
int (*exec)(struct nand_chip *chip, const struct nand_subop *subop);
include/linux/mtd/rawnand.h
971
.exec = _exec, \
include/linux/perf_event.h
1725
extern void perf_event_comm(struct task_struct *tsk, bool exec);
include/linux/perf_event.h
2002
static inline void perf_event_comm(struct task_struct *tsk, bool exec) { }
include/linux/sched.h
1995
extern void __set_task_comm(struct task_struct *tsk, const char *from, bool exec);
include/uapi/linux/a.out.h
44
#define N_MAGIC(exec) ((exec).a_info & 0xffff)
include/uapi/linux/a.out.h
46
#define N_MACHTYPE(exec) ((enum machine_type)(((exec).a_info >> 16) & 0xff))
include/uapi/linux/a.out.h
47
#define N_FLAGS(exec) (((exec).a_info >> 24) & 0xff)
include/uapi/linux/a.out.h
48
#define N_SET_INFO(exec, magic, type, flags) \
include/uapi/linux/a.out.h
49
((exec).a_info = ((magic) & 0xffff) \
include/uapi/linux/a.out.h
52
#define N_SET_MAGIC(exec, magic) \
include/uapi/linux/a.out.h
53
((exec).a_info = (((exec).a_info & 0xffff0000) | ((magic) & 0xffff)))
include/uapi/linux/a.out.h
55
#define N_SET_MACHTYPE(exec, machtype) \
include/uapi/linux/a.out.h
56
((exec).a_info = \
include/uapi/linux/a.out.h
57
((exec).a_info&0xff00ffff) | ((((int)(machtype))&0xff) << 16))
include/uapi/linux/a.out.h
59
#define N_SET_FLAGS(exec, flags) \
include/uapi/linux/a.out.h
60
((exec).a_info = \
include/uapi/linux/a.out.h
61
((exec).a_info&0x00ffffff) | (((flags) & 0xff) << 24))
include/uapi/linux/a.out.h
83
#define _N_HDROFF(x) (1024 - sizeof (struct exec))
include/uapi/linux/a.out.h
87
(N_MAGIC(x) == ZMAGIC ? _N_HDROFF((x)) + sizeof (struct exec) : \
include/uapi/linux/a.out.h
88
(N_MAGIC(x) == QMAGIC ? 0 : sizeof (struct exec)))
include/uapi/linux/cn_proc.h
108
} exec;
include/uapi/linux/openvswitch.h
676
bool exec; /* When true, actions in sample will not
kernel/events/core.c
9325
void perf_event_comm(struct task_struct *task, bool exec)
kernel/events/core.c
9339
.misc = exec ? PERF_RECORD_MISC_COMM_EXEC : 0,
net/openvswitch/actions.c
967
clone_flow_key = !arg->exec;
net/openvswitch/flow_netlink.c
2562
arg.exec = last || !actions_may_change_flow(actions);
net/openvswitch/flow_netlink.c
2650
u32 exec;
net/openvswitch/flow_netlink.c
2659
exec = last || !actions_may_change_flow(attr);
net/openvswitch/flow_netlink.c
2661
err = ovs_nla_add_action(sfa, OVS_CLONE_ATTR_EXEC, &exec,
net/openvswitch/flow_netlink.c
2662
sizeof(exec), log);
scripts/kconfig/qconf.cc
1708
switch (mb.exec()) {
scripts/kconfig/qconf.cc
1855
configApp->exec();
scripts/kconfig/qconf.cc
895
headerPopup->exec(e->globalPos());
tools/perf/util/comm.c
192
struct comm *comm__new(const char *str, u64 timestamp, bool exec)
tools/perf/util/comm.c
200
comm->exec = exec;
tools/perf/util/comm.c
211
int comm__override(struct comm *comm, const char *str, u64 timestamp, bool exec)
tools/perf/util/comm.c
222
if (exec)
tools/perf/util/comm.c
223
comm->exec = true;
tools/perf/util/comm.h
15
bool exec;
tools/perf/util/comm.h
23
struct comm *comm__new(const char *str, u64 timestamp, bool exec);
tools/perf/util/comm.h
26
bool exec);
tools/perf/util/machine.c
607
bool exec = event->header.misc & PERF_RECORD_MISC_COMM_EXEC;
tools/perf/util/machine.c
610
if (exec)
tools/perf/util/machine.c
617
__thread__set_comm(thread, event->comm.comm, sample->time, exec)) {
tools/perf/util/probe-event.c
222
static int convert_exec_to_group(const char *exec, char **result)
tools/perf/util/probe-event.c
228
exec_copy = strdup(exec);
tools/perf/util/probe-event.c
623
static int get_text_start_address(const char *exec, u64 *address,
tools/perf/util/probe-event.c
633
fd = open(exec, O_RDONLY);
tools/perf/util/probe-event.c
776
int ntevs, const char *exec,
tools/perf/util/probe-event.c
782
if (!exec)
tools/perf/util/probe-event.c
785
ret = get_text_start_address(exec, &stext, nsi);
tools/perf/util/probe-event.c
792
tevs[i].point.module = strdup(exec);
tools/perf/util/scripting-engines/trace-event-python.c
1213
tuple_set_s32(t, 4, comm->exec);
tools/perf/util/thread.c
229
if (comm->exec) {
tools/perf/util/thread.c
251
u64 timestamp, bool exec)
tools/perf/util/thread.c
258
int err = comm__override(curr, str, timestamp, exec);
tools/perf/util/thread.c
262
new = comm__new(str, timestamp, exec);
tools/perf/util/thread.c
267
if (exec)
tools/perf/util/thread.c
277
bool exec)
tools/perf/util/thread.c
282
ret = ____thread__set_comm(thread, str, timestamp, exec);
tools/perf/util/thread.h
103
bool exec);
tools/testing/selftests/arm64/gcs/libc-gcs.c
673
FIXTURE_VARIANT_ADD(invalid_mprotect, exec)
tools/testing/selftests/bpf/progs/local_storage.c
193
void BPF_PROG(exec, struct linux_binprm *bprm)
tools/testing/selftests/bpf/progs/task_kfunc_failure.c
305
int BPF_PROG(task_access_comm4, struct task_struct *task, const char *buf, bool exec)
tools/testing/selftests/bpf/progs/test_overhead.c
10
int BPF_KPROBE(prog1, struct task_struct *tsk, const char *buf, bool exec)
tools/testing/selftests/bpf/progs/test_overhead.c
28
int BPF_PROG(prog4, struct task_struct *tsk, const char *buf, bool exec)
tools/testing/selftests/bpf/progs/test_overhead.c
34
int BPF_PROG(prog5, struct task_struct *tsk, const char *buf, bool exec)
tools/testing/selftests/connector/proc_filter.c
164
event->event_data.exec.process_pid,
tools/testing/selftests/connector/proc_filter.c
165
event->event_data.exec.process_tgid);
tools/testing/selftests/exec/check-exec.c
269
const int exec = variant->mount_exec && variant->file_exec;
tools/testing/selftests/exec/check-exec.c
271
test_exec_path(_metadata, reg_file_path, exec ? 0 : EACCES);
tools/testing/selftests/exec/check-exec.c
274
test_exec_path(_metadata, reg_file_path, exec ? 0 : EACCES);
tools/testing/selftests/exec/check-exec.c
279
const int exec = variant->mount_exec && variant->file_exec;
tools/testing/selftests/exec/check-exec.c
283
test_exec_path(_metadata, reg_file_path, exec ? 0 : EACCES);
tools/testing/selftests/exec/check-exec.c
286
test_exec_path(_metadata, reg_file_path, exec ? 0 : EACCES);
tools/testing/selftests/exec/check-exec.c
292
const int exec = variant->file_exec;
tools/testing/selftests/exec/check-exec.c
294
test_exec_fd(_metadata, self->memfd, exec ? 0 : EACCES);
tools/testing/selftests/exec/check-exec.c
297
test_exec_fd(_metadata, self->memfd, exec ? 0 : EACCES);
tools/testing/selftests/exec/check-exec.c
302
const int exec = variant->file_exec;
tools/testing/selftests/exec/check-exec.c
306
test_exec_fd(_metadata, self->memfd, exec ? 0 : EACCES);
tools/testing/selftests/exec/check-exec.c
309
test_exec_fd(_metadata, self->memfd, exec ? 0 : EACCES);
tools/testing/selftests/exec/check-exec.c
375
TEST_F(secbits, exec)
tools/testing/selftests/exec/null-argv.c
10
#define FORK(exec) \
tools/testing/selftests/exec/null-argv.c
15
exec; /* Some kind of exec */ \
tools/testing/selftests/exec/null-argv.c
16
perror("# " #exec); \
tools/testing/selftests/exec/null-argv.c
19
check_result(pid, #exec); \