arch/alpha/include/asm/core_titan.h
358
u64 fault; /* 0x28 */
arch/alpha/kernel/err_titan.c
620
emchk->fault,
arch/alpha/kernel/osf_sys.c
1128
goto fault;
arch/alpha/kernel/osf_sys.c
1137
goto fault;
arch/alpha/kernel/osf_sys.c
1141
fault:
arch/alpha/mm/fault.c
145
fault = handle_mm_fault(vma, address, flags, regs);
arch/alpha/mm/fault.c
147
if (fault_signal_pending(fault, regs)) {
arch/alpha/mm/fault.c
154
if (fault & VM_FAULT_COMPLETED)
arch/alpha/mm/fault.c
157
if (unlikely(fault & VM_FAULT_ERROR)) {
arch/alpha/mm/fault.c
158
if (fault & VM_FAULT_OOM)
arch/alpha/mm/fault.c
160
else if (fault & VM_FAULT_SIGSEGV)
arch/alpha/mm/fault.c
162
else if (fault & VM_FAULT_SIGBUS)
arch/alpha/mm/fault.c
167
if (fault & VM_FAULT_RETRY) {
arch/alpha/mm/fault.c
92
vm_fault_t fault;
arch/arc/include/asm/disasm.h
87
int fault, write;
arch/arc/kernel/disasm.c
40
goto fault;
arch/arc/kernel/disasm.c
429
fault: state->fault = 1;
arch/arc/kernel/disasm.c
53
goto fault;
arch/arc/kernel/process.c
72
goto fault;
arch/arc/kernel/process.c
79
goto fault;
arch/arc/kernel/process.c
87
fault:
arch/arc/kernel/unaligned.c
126
goto fault; \
arch/arc/kernel/unaligned.c
160
fault: state->fault = 1;
arch/arc/kernel/unaligned.c
180
goto fault;
arch/arc/kernel/unaligned.c
192
fault: state->fault = 1;
arch/arc/kernel/unaligned.c
225
if (state.fault)
arch/arc/kernel/unaligned.c
226
goto fault;
arch/arc/kernel/unaligned.c
230
goto fault;
arch/arc/kernel/unaligned.c
237
if (state.fault)
arch/arc/kernel/unaligned.c
238
goto fault;
arch/arc/kernel/unaligned.c
257
fault:
arch/arc/kernel/unaligned.c
51
goto fault; \
arch/arc/kernel/unaligned.c
66
goto fault; \
arch/arc/kernel/unaligned.c
93
goto fault; \
arch/arc/mm/fault.c
135
fault = handle_mm_fault(vma, address, flags, regs);
arch/arc/mm/fault.c
138
if (fault_signal_pending(fault, regs)) {
arch/arc/mm/fault.c
145
if (fault & VM_FAULT_COMPLETED)
arch/arc/mm/fault.c
151
if (unlikely(fault & VM_FAULT_RETRY)) {
arch/arc/mm/fault.c
164
if (likely(!(fault & VM_FAULT_ERROR)))
arch/arc/mm/fault.c
171
if (fault & VM_FAULT_OOM) {
arch/arc/mm/fault.c
176
if (fault & VM_FAULT_SIGBUS) {
arch/arc/mm/fault.c
80
vm_fault_t fault = VM_FAULT_SIGSEGV; /* handle_mm_fault() output */
arch/arm/mm/alignment.c
223
goto fault; \
arch/arm/mm/alignment.c
244
goto fault; \
arch/arm/mm/alignment.c
276
goto fault; \
arch/arm/mm/alignment.c
318
goto fault; \
arch/arm/mm/alignment.c
385
fault:
arch/arm/mm/alignment.c
447
fault:
arch/arm/mm/alignment.c
483
fault:
arch/arm/mm/alignment.c
583
fault:
arch/arm/mm/alignment.c
773
int fault;
arch/arm/mm/alignment.c
776
fault = get_user(instr, ip);
arch/arm/mm/alignment.c
778
fault = get_kernel_nofault(instr, ip);
arch/arm/mm/alignment.c
782
return fault;
arch/arm/mm/alignment.c
788
int fault;
arch/arm/mm/alignment.c
791
fault = get_user(instr, ip);
arch/arm/mm/alignment.c
793
fault = get_kernel_nofault(instr, ip);
arch/arm/mm/alignment.c
797
return fault;
arch/arm/mm/alignment.c
811
int fault;
arch/arm/mm/alignment.c
824
fault = alignment_get_thumb(regs, ptr, &tinstr);
arch/arm/mm/alignment.c
825
if (!fault) {
arch/arm/mm/alignment.c
830
fault = alignment_get_thumb(regs, ptr + 1, &tinst2);
arch/arm/mm/alignment.c
839
fault = alignment_get_arm(regs, (void *)instrptr, &instr);
arch/arm/mm/alignment.c
842
if (fault) {
arch/arm/mm/fault.c
298
vm_fault_t fault;
arch/arm/mm/fault.c
359
fault = 0;
arch/arm/mm/fault.c
363
fault = handle_mm_fault(vma, addr, flags | FAULT_FLAG_VMA_LOCK, regs);
arch/arm/mm/fault.c
364
if (!(fault & (VM_FAULT_RETRY | VM_FAULT_COMPLETED)))
arch/arm/mm/fault.c
367
if (!(fault & VM_FAULT_RETRY)) {
arch/arm/mm/fault.c
372
if (fault & VM_FAULT_MAJOR)
arch/arm/mm/fault.c
376
if (fault_signal_pending(fault, regs)) {
arch/arm/mm/fault.c
386
fault = 0;
arch/arm/mm/fault.c
397
fault = 0;
arch/arm/mm/fault.c
402
fault = handle_mm_fault(vma, addr & PAGE_MASK, flags, regs);
arch/arm/mm/fault.c
408
if (fault_signal_pending(fault, regs)) {
arch/arm/mm/fault.c
415
if (fault & VM_FAULT_COMPLETED)
arch/arm/mm/fault.c
418
if (!(fault & VM_FAULT_ERROR)) {
arch/arm/mm/fault.c
419
if (fault & VM_FAULT_RETRY) {
arch/arm/mm/fault.c
429
if (likely(!(fault & VM_FAULT_ERROR)))
arch/arm/mm/fault.c
441
if (fault & VM_FAULT_OOM) {
arch/arm/mm/fault.c
451
if (fault & VM_FAULT_SIGBUS) {
arch/arm64/include/asm/kvm_emulate.h
317
return vcpu->arch.fault.esr_el2;
arch/arm64/include/asm/kvm_emulate.h
345
return vcpu->arch.fault.far_el2;
arch/arm64/include/asm/kvm_emulate.h
350
u64 hpfar = vcpu->arch.fault.hpfar_el2;
arch/arm64/include/asm/kvm_emulate.h
360
return vcpu->arch.fault.disr_el1;
arch/arm64/include/asm/kvm_host.h
855
struct kvm_vcpu_fault_info fault;
arch/arm64/kernel/armv8_deprecated.c
203
goto fault;
arch/arm64/kernel/armv8_deprecated.c
208
goto fault;
arch/arm64/kernel/armv8_deprecated.c
224
fault:
arch/arm64/kernel/asm-offsets.c
106
DEFINE(VCPU_FAULT_DISR, offsetof(struct kvm_vcpu, arch.fault.disr_el1));
arch/arm64/kernel/compat_alignment.c
287
int fault;
arch/arm64/kernel/compat_alignment.c
289
fault = get_user(instr, ip);
arch/arm64/kernel/compat_alignment.c
290
if (fault)
arch/arm64/kernel/compat_alignment.c
291
return fault;
arch/arm64/kernel/compat_alignment.c
300
int fault;
arch/arm64/kernel/compat_alignment.c
302
fault = get_user(instr, ip);
arch/arm64/kernel/compat_alignment.c
303
if (fault)
arch/arm64/kernel/compat_alignment.c
304
return fault;
arch/arm64/kvm/handle_exit.c
206
run->debug.arch.far = vcpu->arch.fault.far_el2;
arch/arm64/kvm/hyp/include/hyp/fault.h
100
fault->hpfar_el2 = hpfar | HPFAR_EL2_NS;
arch/arm64/kvm/hyp/include/hyp/fault.h
82
static inline bool __get_fault_info(u64 esr, struct kvm_vcpu_fault_info *fault)
arch/arm64/kvm/hyp/include/hyp/fault.h
86
fault->far_el2 = read_sysreg_el2(SYS_FAR);
arch/arm64/kvm/hyp/include/hyp/fault.h
87
fault->hpfar_el2 = 0;
arch/arm64/kvm/hyp/include/hyp/fault.h
93
else if (!__translate_far_to_hpfar(fault->far_el2, &hpfar))
arch/arm64/kvm/hyp/include/hyp/switch.h
413
return __get_fault_info(vcpu->arch.fault.esr_el2, &vcpu->arch.fault);
arch/arm64/kvm/hyp/include/hyp/switch.h
419
arm64_mops_reset_regs(vcpu_gp_regs(vcpu), vcpu->arch.fault.esr_el2);
arch/arm64/kvm/hyp/include/hyp/switch.h
882
vcpu->arch.fault.esr_el2 = read_sysreg_el2(SYS_ESR);
arch/arm64/kvm/hyp/nvhe/hyp-main.c
155
host_vcpu->arch.fault = hyp_vcpu->vcpu.arch.fault;
arch/arm64/kvm/hyp/nvhe/mem_protect.c
610
struct kvm_vcpu_fault_info fault;
arch/arm64/kvm/hyp/nvhe/mem_protect.c
615
if (!__get_fault_info(esr, &fault)) {
arch/arm64/kvm/hyp/nvhe/mem_protect.c
628
BUG_ON(!(fault.hpfar_el2 & HPFAR_EL2_NS));
arch/arm64/kvm/hyp/nvhe/mem_protect.c
629
addr = FIELD_GET(HPFAR_EL2_FIPA, fault.hpfar_el2) << 12;
arch/arm64/kvm/hyp/vhe/switch.c
515
vcpu->arch.fault.esr_el2 = FIELD_PREP(ESR_ELx_EC_MASK, ESR_ELx_EC_SYS64) |
arch/arm64/kvm/nested.c
856
vcpu_write_sys_reg(vcpu, vcpu->arch.fault.far_el2, FAR_EL2);
arch/arm64/kvm/nested.c
857
vcpu_write_sys_reg(vcpu, vcpu->arch.fault.hpfar_el2, HPFAR_EL2);
arch/arm64/mm/fault.c
561
vm_fault_t fault;
arch/arm64/mm/fault.c
635
fault = 0;
arch/arm64/mm/fault.c
642
fault = 0;
arch/arm64/mm/fault.c
651
fault = 0;
arch/arm64/mm/fault.c
657
fault = handle_mm_fault(vma, addr, mm_flags | FAULT_FLAG_VMA_LOCK, regs);
arch/arm64/mm/fault.c
658
if (!(fault & (VM_FAULT_RETRY | VM_FAULT_COMPLETED)))
arch/arm64/mm/fault.c
661
if (!(fault & VM_FAULT_RETRY)) {
arch/arm64/mm/fault.c
666
if (fault & VM_FAULT_MAJOR)
arch/arm64/mm/fault.c
670
if (fault_signal_pending(fault, regs)) {
arch/arm64/mm/fault.c
680
fault = 0;
arch/arm64/mm/fault.c
687
fault = 0;
arch/arm64/mm/fault.c
695
fault = 0;
arch/arm64/mm/fault.c
700
fault = handle_mm_fault(vma, addr, mm_flags, regs);
arch/arm64/mm/fault.c
703
if (fault_signal_pending(fault, regs)) {
arch/arm64/mm/fault.c
710
if (fault & VM_FAULT_COMPLETED)
arch/arm64/mm/fault.c
713
if (fault & VM_FAULT_RETRY) {
arch/arm64/mm/fault.c
721
if (likely(!(fault & VM_FAULT_ERROR)))
arch/arm64/mm/fault.c
733
if (fault & VM_FAULT_OOM) {
arch/arm64/mm/fault.c
745
if (fault & VM_FAULT_SIGBUS) {
arch/arm64/mm/fault.c
751
} else if (fault & (VM_FAULT_HWPOISON_LARGE | VM_FAULT_HWPOISON)) {
arch/arm64/mm/fault.c
755
if (fault & VM_FAULT_HWPOISON_LARGE)
arch/arm64/mm/fault.c
756
lsb = hstate_index_to_shift(VM_FAULT_GET_HINDEX(fault));
arch/csky/abiv2/fpu.c
22
int fault;
arch/csky/abiv2/fpu.c
32
fault = __get_user(instr_low, (u16 *)instrptr);
arch/csky/abiv2/fpu.c
33
if (fault)
arch/csky/abiv2/fpu.c
36
fault = __get_user(instr_hi, (u16 *)(instrptr + 2));
arch/csky/abiv2/fpu.c
37
if (fault)
arch/csky/mm/fault.c
195
vm_fault_t fault;
arch/csky/mm/fault.c
263
fault = handle_mm_fault(vma, addr, flags, regs);
arch/csky/mm/fault.c
270
if (fault_signal_pending(fault, regs)) {
arch/csky/mm/fault.c
277
if (fault & VM_FAULT_COMPLETED)
arch/csky/mm/fault.c
280
if (unlikely(fault & VM_FAULT_RETRY)) {
arch/csky/mm/fault.c
293
if (unlikely(fault & VM_FAULT_ERROR)) {
arch/csky/mm/fault.c
294
mm_fault_error(regs, addr, fault);
arch/csky/mm/fault.c
73
static inline void mm_fault_error(struct pt_regs *regs, unsigned long addr, vm_fault_t fault)
arch/csky/mm/fault.c
77
if (fault & VM_FAULT_OOM) {
arch/csky/mm/fault.c
88
} else if (fault & VM_FAULT_SIGBUS) {
arch/hexagon/mm/vm_fault.c
113
if (fault & VM_FAULT_OOM) {
arch/hexagon/mm/vm_fault.c
121
if (fault & VM_FAULT_SIGBUS) {
arch/hexagon/mm/vm_fault.c
43
vm_fault_t fault;
arch/hexagon/mm/vm_fault.c
84
fault = handle_mm_fault(vma, address, flags, regs);
arch/hexagon/mm/vm_fault.c
86
if (fault_signal_pending(fault, regs)) {
arch/hexagon/mm/vm_fault.c
93
if (fault & VM_FAULT_COMPLETED)
arch/hexagon/mm/vm_fault.c
97
if (likely(!(fault & VM_FAULT_ERROR))) {
arch/hexagon/mm/vm_fault.c
98
if (fault & VM_FAULT_RETRY) {
arch/loongarch/include/asm/kvm_vcpu.h
50
int kvm_handle_fault(struct kvm_vcpu *vcpu, int fault);
arch/loongarch/kernel/unaligned.c
445
goto fault;
arch/loongarch/kernel/unaligned.c
469
goto fault;
arch/loongarch/kernel/unaligned.c
483
fault:
arch/loongarch/kvm/exit.c
975
int kvm_handle_fault(struct kvm_vcpu *vcpu, int fault)
arch/loongarch/kvm/exit.c
977
return kvm_fault_tables[fault](vcpu, fault);
arch/loongarch/mm/fault.c
183
vm_fault_t fault;
arch/loongarch/mm/fault.c
249
fault = handle_mm_fault(vma, address, flags | FAULT_FLAG_VMA_LOCK, regs);
arch/loongarch/mm/fault.c
250
if (!(fault & (VM_FAULT_RETRY | VM_FAULT_COMPLETED)))
arch/loongarch/mm/fault.c
253
if (!(fault & VM_FAULT_RETRY)) {
arch/loongarch/mm/fault.c
259
if (fault & VM_FAULT_MAJOR)
arch/loongarch/mm/fault.c
263
if (fault_signal_pending(fault, regs)) {
arch/loongarch/mm/fault.c
309
fault = handle_mm_fault(vma, address, flags, regs);
arch/loongarch/mm/fault.c
311
if (fault_signal_pending(fault, regs)) {
arch/loongarch/mm/fault.c
318
if (fault & VM_FAULT_COMPLETED)
arch/loongarch/mm/fault.c
321
if (unlikely(fault & VM_FAULT_RETRY)) {
arch/loongarch/mm/fault.c
334
if (unlikely(fault & VM_FAULT_ERROR)) {
arch/loongarch/mm/fault.c
335
if (fault & VM_FAULT_OOM) {
arch/loongarch/mm/fault.c
338
} else if (fault & VM_FAULT_SIGSEGV) {
arch/loongarch/mm/fault.c
341
} else if (fault & (VM_FAULT_SIGBUS|VM_FAULT_HWPOISON|VM_FAULT_HWPOISON_LARGE)) {
arch/m68k/ifpsp060/src/fpsp.S
19600
# if an immediate data access fails, the resulting fault status #
arch/m68k/mm/fault.c
141
fault = handle_mm_fault(vma, address, flags, regs);
arch/m68k/mm/fault.c
142
pr_debug("handle_mm_fault returns %x\n", fault);
arch/m68k/mm/fault.c
144
if (fault_signal_pending(fault, regs)) {
arch/m68k/mm/fault.c
151
if (fault & VM_FAULT_COMPLETED)
arch/m68k/mm/fault.c
154
if (unlikely(fault & VM_FAULT_ERROR)) {
arch/m68k/mm/fault.c
155
if (fault & VM_FAULT_OOM)
arch/m68k/mm/fault.c
157
else if (fault & VM_FAULT_SIGSEGV)
arch/m68k/mm/fault.c
159
else if (fault & VM_FAULT_SIGBUS)
arch/m68k/mm/fault.c
164
if (fault & VM_FAULT_RETRY) {
arch/m68k/mm/fault.c
75
vm_fault_t fault;
arch/microblaze/mm/fault.c
221
fault = handle_mm_fault(vma, address, flags, regs);
arch/microblaze/mm/fault.c
223
if (fault_signal_pending(fault, regs)) {
arch/microblaze/mm/fault.c
230
if (fault & VM_FAULT_COMPLETED)
arch/microblaze/mm/fault.c
233
if (unlikely(fault & VM_FAULT_ERROR)) {
arch/microblaze/mm/fault.c
234
if (fault & VM_FAULT_OOM)
arch/microblaze/mm/fault.c
236
else if (fault & VM_FAULT_SIGSEGV)
arch/microblaze/mm/fault.c
238
else if (fault & VM_FAULT_SIGBUS)
arch/microblaze/mm/fault.c
243
if (fault & VM_FAULT_RETRY) {
arch/microblaze/mm/fault.c
93
vm_fault_t fault;
arch/mips/kernel/unaligned.c
1020
goto fault;
arch/mips/kernel/unaligned.c
1026
goto fault;
arch/mips/kernel/unaligned.c
1041
goto fault;
arch/mips/kernel/unaligned.c
1047
goto fault;
arch/mips/kernel/unaligned.c
1093
goto fault;
arch/mips/kernel/unaligned.c
1103
goto fault;
arch/mips/kernel/unaligned.c
1113
goto fault;
arch/mips/kernel/unaligned.c
1131
goto fault;
arch/mips/kernel/unaligned.c
1153
goto fault;
arch/mips/kernel/unaligned.c
1168
goto fault;
arch/mips/kernel/unaligned.c
1178
goto fault;
arch/mips/kernel/unaligned.c
1196
goto fault;
arch/mips/kernel/unaligned.c
1211
fault:
arch/mips/kernel/unaligned.c
1354
goto fault;
arch/mips/kernel/unaligned.c
1365
goto fault;
arch/mips/kernel/unaligned.c
1378
goto fault;
arch/mips/kernel/unaligned.c
1397
goto fault;
arch/mips/kernel/unaligned.c
1421
goto fault;
arch/mips/kernel/unaligned.c
1438
goto fault;
arch/mips/kernel/unaligned.c
1451
goto fault;
arch/mips/kernel/unaligned.c
1471
goto fault;
arch/mips/kernel/unaligned.c
1492
fault:
arch/mips/kernel/unaligned.c
175
goto fault;
arch/mips/kernel/unaligned.c
184
goto fault;
arch/mips/kernel/unaligned.c
193
goto fault;
arch/mips/kernel/unaligned.c
213
goto fault;
arch/mips/kernel/unaligned.c
222
goto fault;
arch/mips/kernel/unaligned.c
243
goto fault;
arch/mips/kernel/unaligned.c
252
goto fault;
arch/mips/kernel/unaligned.c
261
goto fault;
arch/mips/kernel/unaligned.c
272
goto fault;
arch/mips/kernel/unaligned.c
281
goto fault;
arch/mips/kernel/unaligned.c
299
goto fault;
arch/mips/kernel/unaligned.c
314
goto fault;
arch/mips/kernel/unaligned.c
329
goto fault;
arch/mips/kernel/unaligned.c
348
goto fault;
arch/mips/kernel/unaligned.c
371
goto fault;
arch/mips/kernel/unaligned.c
393
goto fault;
arch/mips/kernel/unaligned.c
409
goto fault;
arch/mips/kernel/unaligned.c
428
goto fault;
arch/mips/kernel/unaligned.c
501
goto fault;
arch/mips/kernel/unaligned.c
534
goto fault;
arch/mips/kernel/unaligned.c
584
fault:
arch/mips/kernel/unaligned.c
656
goto fault;
arch/mips/kernel/unaligned.c
663
goto fault;
arch/mips/kernel/unaligned.c
698
goto fault;
arch/mips/kernel/unaligned.c
703
goto fault;
arch/mips/kernel/unaligned.c
718
goto fault;
arch/mips/kernel/unaligned.c
723
goto fault;
arch/mips/kernel/unaligned.c
737
goto fault;
arch/mips/kernel/unaligned.c
742
goto fault;
arch/mips/kernel/unaligned.c
761
goto fault;
arch/mips/kernel/unaligned.c
766
goto fault;
arch/mips/kernel/unaligned.c
789
goto fault;
arch/mips/kernel/unaligned.c
796
goto fault;
arch/mips/kernel/unaligned.c
803
goto fault;
arch/mips/kernel/unaligned.c
826
goto fault;
arch/mips/kernel/unaligned.c
833
goto fault;
arch/mips/kernel/unaligned.c
840
goto fault;
arch/mips/kernel/unaligned.c
863
goto fault;
arch/mips/kernel/unaligned.c
870
goto fault;
arch/mips/kernel/unaligned.c
877
goto fault;
arch/mips/kernel/unaligned.c
905
goto fault;
arch/mips/kernel/unaligned.c
912
goto fault;
arch/mips/kernel/unaligned.c
919
goto fault;
arch/mips/loongson64/cop2-ex.c
118
goto fault;
arch/mips/loongson64/cop2-ex.c
123
goto fault;
arch/mips/loongson64/cop2-ex.c
135
goto fault;
arch/mips/loongson64/cop2-ex.c
141
goto fault;
arch/mips/loongson64/cop2-ex.c
165
goto fault;
arch/mips/loongson64/cop2-ex.c
176
goto fault;
arch/mips/loongson64/cop2-ex.c
187
goto fault;
arch/mips/loongson64/cop2-ex.c
201
goto fault;
arch/mips/loongson64/cop2-ex.c
217
goto fault;
arch/mips/loongson64/cop2-ex.c
247
goto fault;
arch/mips/loongson64/cop2-ex.c
259
goto fault;
arch/mips/loongson64/cop2-ex.c
271
goto fault;
arch/mips/loongson64/cop2-ex.c
287
goto fault;
arch/mips/loongson64/cop2-ex.c
305
goto fault;
arch/mips/loongson64/cop2-ex.c
317
fault:
arch/mips/loongson64/cop2-ex.c
76
goto fault;
arch/mips/loongson64/cop2-ex.c
80
goto fault;
arch/mips/loongson64/cop2-ex.c
92
goto fault;
arch/mips/loongson64/cop2-ex.c
96
goto fault;
arch/mips/mm/fault.c
150
fault = handle_mm_fault(vma, address, flags, regs);
arch/mips/mm/fault.c
152
if (fault_signal_pending(fault, regs)) {
arch/mips/mm/fault.c
159
if (fault & VM_FAULT_COMPLETED)
arch/mips/mm/fault.c
162
if (unlikely(fault & VM_FAULT_ERROR)) {
arch/mips/mm/fault.c
163
if (fault & VM_FAULT_OOM)
arch/mips/mm/fault.c
165
else if (fault & VM_FAULT_SIGSEGV)
arch/mips/mm/fault.c
167
else if (fault & VM_FAULT_SIGBUS)
arch/mips/mm/fault.c
172
if (fault & VM_FAULT_RETRY) {
arch/mips/mm/fault.c
47
vm_fault_t fault;
arch/nios2/kernel/misaligned.c
111
fault |= __put_user(d0, (u8 *)(addr+0));
arch/nios2/kernel/misaligned.c
112
fault |= __put_user(d1, (u8 *)(addr+1));
arch/nios2/kernel/misaligned.c
116
fault |= __get_user(d0, (u8 *)(addr+0));
arch/nios2/kernel/misaligned.c
117
fault |= __get_user(d1, (u8 *)(addr+1));
arch/nios2/kernel/misaligned.c
133
fault |= __put_user(d0, (u8 *)(addr+0));
arch/nios2/kernel/misaligned.c
134
fault |= __put_user(d1, (u8 *)(addr+1));
arch/nios2/kernel/misaligned.c
135
fault |= __put_user(d2, (u8 *)(addr+2));
arch/nios2/kernel/misaligned.c
136
fault |= __put_user(d3, (u8 *)(addr+3));
arch/nios2/kernel/misaligned.c
140
fault |= __get_user(d0, (u8 *)(addr+0));
arch/nios2/kernel/misaligned.c
141
fault |= __get_user(d1, (u8 *)(addr+1));
arch/nios2/kernel/misaligned.c
142
fault |= __get_user(d2, (u8 *)(addr+2));
arch/nios2/kernel/misaligned.c
143
fault |= __get_user(d3, (u8 *)(addr+3));
arch/nios2/kernel/misaligned.c
153
if (fault) {
arch/nios2/kernel/misaligned.c
72
unsigned int fault;
arch/nios2/kernel/misaligned.c
85
fault = 0;
arch/nios2/kernel/misaligned.c
98
fault |= __get_user(d0, (u8 *)(addr+0));
arch/nios2/kernel/misaligned.c
99
fault |= __get_user(d1, (u8 *)(addr+1));
arch/nios2/mm/fault.c
124
fault = handle_mm_fault(vma, address, flags, regs);
arch/nios2/mm/fault.c
126
if (fault_signal_pending(fault, regs)) {
arch/nios2/mm/fault.c
133
if (fault & VM_FAULT_COMPLETED)
arch/nios2/mm/fault.c
136
if (unlikely(fault & VM_FAULT_ERROR)) {
arch/nios2/mm/fault.c
137
if (fault & VM_FAULT_OOM)
arch/nios2/mm/fault.c
139
else if (fault & VM_FAULT_SIGSEGV)
arch/nios2/mm/fault.c
141
else if (fault & VM_FAULT_SIGBUS)
arch/nios2/mm/fault.c
146
if (fault & VM_FAULT_RETRY) {
arch/nios2/mm/fault.c
50
vm_fault_t fault;
arch/openrisc/mm/fault.c
166
fault = handle_mm_fault(vma, address, flags, regs);
arch/openrisc/mm/fault.c
168
if (fault_signal_pending(fault, regs)) {
arch/openrisc/mm/fault.c
175
if (fault & VM_FAULT_COMPLETED)
arch/openrisc/mm/fault.c
178
if (unlikely(fault & VM_FAULT_ERROR)) {
arch/openrisc/mm/fault.c
179
if (fault & VM_FAULT_OOM)
arch/openrisc/mm/fault.c
181
else if (fault & VM_FAULT_SIGSEGV)
arch/openrisc/mm/fault.c
183
else if (fault & VM_FAULT_SIGBUS)
arch/openrisc/mm/fault.c
189
if (fault & VM_FAULT_RETRY) {
arch/openrisc/mm/fault.c
53
vm_fault_t fault;
arch/parisc/mm/fault.c
274
vm_fault_t fault = 0;
arch/parisc/mm/fault.c
318
fault = handle_mm_fault(vma, address, flags, regs);
arch/parisc/mm/fault.c
320
if (fault_signal_pending(fault, regs)) {
arch/parisc/mm/fault.c
329
if (fault & VM_FAULT_COMPLETED)
arch/parisc/mm/fault.c
332
if (unlikely(fault & VM_FAULT_ERROR)) {
arch/parisc/mm/fault.c
338
if (fault & VM_FAULT_OOM)
arch/parisc/mm/fault.c
340
else if (fault & VM_FAULT_SIGSEGV)
arch/parisc/mm/fault.c
342
else if (fault & (VM_FAULT_SIGBUS|VM_FAULT_HWPOISON|
arch/parisc/mm/fault.c
347
if (fault & VM_FAULT_RETRY) {
arch/parisc/mm/fault.c
405
if (fault & (VM_FAULT_HWPOISON|VM_FAULT_HWPOISON_LARGE)) {
arch/parisc/mm/fault.c
415
if (fault & VM_FAULT_HWPOISON_LARGE)
arch/parisc/mm/fault.c
416
lsb = hstate_index_to_shift(VM_FAULT_GET_HINDEX(fault));
arch/parisc/mm/fault.c
417
else if (fault & VM_FAULT_HWPOISON)
arch/powerpc/kvm/book3s_64_vio.c
247
.fault = kvm_spapr_tce_fault,
arch/powerpc/kvm/book3s_xive_native.c
285
.fault = xive_native_esb_fault,
arch/powerpc/kvm/book3s_xive_native.c
306
.fault = xive_native_tima_fault,
arch/powerpc/mm/fault.c
141
vm_fault_t fault)
arch/powerpc/mm/fault.c
148
if (fault & (VM_FAULT_HWPOISON|VM_FAULT_HWPOISON_LARGE)) {
arch/powerpc/mm/fault.c
154
if (fault & VM_FAULT_HWPOISON_LARGE)
arch/powerpc/mm/fault.c
155
lsb = hstate_index_to_shift(VM_FAULT_GET_HINDEX(fault));
arch/powerpc/mm/fault.c
156
if (fault & VM_FAULT_HWPOISON)
arch/powerpc/mm/fault.c
169
vm_fault_t fault)
arch/powerpc/mm/fault.c
179
if (fault & VM_FAULT_OOM) {
arch/powerpc/mm/fault.c
188
if (fault & (VM_FAULT_SIGBUS|VM_FAULT_HWPOISON|
arch/powerpc/mm/fault.c
190
return do_sigbus(regs, addr, fault);
arch/powerpc/mm/fault.c
191
else if (fault & VM_FAULT_SIGSEGV)
arch/powerpc/mm/fault.c
423
vm_fault_t fault, major = 0;
arch/powerpc/mm/fault.c
505
fault = handle_mm_fault(vma, address, flags | FAULT_FLAG_VMA_LOCK, regs);
arch/powerpc/mm/fault.c
506
if (!(fault & (VM_FAULT_RETRY | VM_FAULT_COMPLETED)))
arch/powerpc/mm/fault.c
509
if (!(fault & VM_FAULT_RETRY)) {
arch/powerpc/mm/fault.c
514
if (fault & VM_FAULT_MAJOR)
arch/powerpc/mm/fault.c
517
if (fault_signal_pending(fault, regs))
arch/powerpc/mm/fault.c
548
fault = handle_mm_fault(vma, address, flags, regs);
arch/powerpc/mm/fault.c
550
major |= fault & VM_FAULT_MAJOR;
arch/powerpc/mm/fault.c
552
if (fault_signal_pending(fault, regs))
arch/powerpc/mm/fault.c
556
if (fault & VM_FAULT_COMPLETED)
arch/powerpc/mm/fault.c
563
if (unlikely(fault & VM_FAULT_RETRY)) {
arch/powerpc/mm/fault.c
571
if (unlikely(fault & VM_FAULT_ERROR))
arch/powerpc/mm/fault.c
572
return mm_fault_error(regs, address, fault);
arch/powerpc/platforms/book3s/vas-api.c
401
vm_fault_t fault;
arch/powerpc/platforms/book3s/vas-api.c
438
fault = vmf_insert_pfn(vma, vma->vm_start,
arch/powerpc/platforms/book3s/vas-api.c
440
return fault;
arch/powerpc/platforms/book3s/vas-api.c
502
.fault = vas_mmap_fault,
arch/powerpc/platforms/cell/spufs/file.c
1031
.fault = spufs_signal1_mmap_fault,
arch/powerpc/platforms/cell/spufs/file.c
1165
.fault = spufs_signal2_mmap_fault,
arch/powerpc/platforms/cell/spufs/file.c
1283
.fault = spufs_mss_mmap_fault,
arch/powerpc/platforms/cell/spufs/file.c
1344
.fault = spufs_psmap_mmap_fault,
arch/powerpc/platforms/cell/spufs/file.c
1403
.fault = spufs_mfc_mmap_fault,
arch/powerpc/platforms/cell/spufs/file.c
285
.fault = spufs_mem_mmap_fault,
arch/powerpc/platforms/cell/spufs/file.c
373
.fault = spufs_cntl_mmap_fault,
arch/powerpc/platforms/pseries/vas.h
114
__be32 fault; /* Interrupt when NX sees fault */
arch/riscv/kernel/usercfi.c
140
_ASM_EXTABLE(1b, %l[fault])
arch/riscv/kernel/usercfi.c
145
: fault
arch/riscv/kernel/usercfi.c
149
fault:
arch/riscv/mm/fault.c
117
static inline void mm_fault_error(struct pt_regs *regs, unsigned long addr, vm_fault_t fault)
arch/riscv/mm/fault.c
124
if (fault & VM_FAULT_OOM) {
arch/riscv/mm/fault.c
131
} else if (fault & (VM_FAULT_SIGBUS | VM_FAULT_HWPOISON | VM_FAULT_HWPOISON_LARGE)) {
arch/riscv/mm/fault.c
135
} else if (fault & VM_FAULT_SIGSEGV) {
arch/riscv/mm/fault.c
286
vm_fault_t fault;
arch/riscv/mm/fault.c
362
fault = handle_mm_fault(vma, addr, flags | FAULT_FLAG_VMA_LOCK, regs);
arch/riscv/mm/fault.c
363
if (!(fault & (VM_FAULT_RETRY | VM_FAULT_COMPLETED)))
arch/riscv/mm/fault.c
366
if (!(fault & VM_FAULT_RETRY)) {
arch/riscv/mm/fault.c
371
if (fault & VM_FAULT_MAJOR)
arch/riscv/mm/fault.c
374
if (fault_signal_pending(fault, regs)) {
arch/riscv/mm/fault.c
406
fault = handle_mm_fault(vma, addr, flags, regs);
arch/riscv/mm/fault.c
413
if (fault_signal_pending(fault, regs)) {
arch/riscv/mm/fault.c
420
if (fault & VM_FAULT_COMPLETED)
arch/riscv/mm/fault.c
423
if (unlikely(fault & VM_FAULT_RETRY)) {
arch/riscv/mm/fault.c
437
if (unlikely(fault & VM_FAULT_ERROR)) {
arch/riscv/mm/fault.c
439
mm_fault_error(regs, addr, fault);
arch/s390/kvm/gaccess.c
1147
struct guest_fault fault = {
arch/s390/kvm/gaccess.c
1160
rc = kvm_s390_faultin_gfn(NULL, kvm, &fault);
arch/s390/kvm/gaccess.c
892
struct guest_fault fault = {
arch/s390/kvm/gaccess.c
903
rc = kvm_s390_faultin_gfn(NULL, kvm, &fault);
arch/s390/kvm/gmap.c
593
int gmap_try_fixup_minor(struct gmap *gmap, struct guest_fault *fault)
arch/s390/kvm/gmap.c
600
rc = dat_entry_walk(NULL, fault->gfn, gmap->asce, DAT_WALK_LEAF, TABLE_TYPE_PAGE_TABLE,
arch/s390/kvm/gmap.c
601
&fault->crstep, &fault->ptep);
arch/s390/kvm/gmap.c
606
if (fault->ptep) {
arch/s390/kvm/gmap.c
607
pgste = pgste_get_lock(fault->ptep);
arch/s390/kvm/gmap.c
608
rc = _gmap_handle_minor_pte_fault(gmap, &pgste, fault);
arch/s390/kvm/gmap.c
609
if (!rc && fault->callback)
arch/s390/kvm/gmap.c
610
fault->callback(fault);
arch/s390/kvm/gmap.c
611
pgste_set_unlock(fault->ptep, pgste);
arch/s390/kvm/gmap.c
613
rc = gmap_handle_minor_crste_fault(gmap, fault);
arch/s390/kvm/gmap.c
614
if (!rc && fault->callback)
arch/s390/kvm/gmap.c
615
fault->callback(fault);
arch/s390/kvm/gmap.h
88
int gmap_try_fixup_minor(struct gmap *gmap, struct guest_fault *fault);
arch/s390/kvm/gmap.h
93
int gmap_link(struct kvm_s390_mmu_cache *mc, struct gmap *gmap, struct guest_fault *fault);
arch/s390/mm/fault.c
272
vm_fault_t fault;
arch/s390/mm/fault.c
305
fault = handle_mm_fault(vma, address, flags | FAULT_FLAG_VMA_LOCK, regs);
arch/s390/mm/fault.c
306
if (!(fault & (VM_FAULT_RETRY | VM_FAULT_COMPLETED)))
arch/s390/mm/fault.c
308
if (!(fault & VM_FAULT_RETRY)) {
arch/s390/mm/fault.c
313
if (fault & VM_FAULT_MAJOR)
arch/s390/mm/fault.c
316
if (fault_signal_pending(fault, regs)) {
arch/s390/mm/fault.c
328
fault = handle_mm_fault(vma, address, flags, regs);
arch/s390/mm/fault.c
329
if (fault_signal_pending(fault, regs)) {
arch/s390/mm/fault.c
335
if (fault & VM_FAULT_COMPLETED)
arch/s390/mm/fault.c
337
if (fault & VM_FAULT_RETRY) {
arch/s390/mm/fault.c
343
if (!(fault & VM_FAULT_ERROR))
arch/s390/mm/fault.c
345
if (fault & VM_FAULT_OOM) {
arch/s390/mm/fault.c
350
} else if (fault & VM_FAULT_SIGSEGV) {
arch/s390/mm/fault.c
355
} else if (fault & (VM_FAULT_SIGBUS | VM_FAULT_HWPOISON |
arch/s390/mm/fault.c
362
pr_emerg("Unexpected fault flags: %08x\n", fault);
arch/s390/mm/gmap_helpers.c
225
vm_fault_t fault;
arch/s390/mm/gmap_helpers.c
247
fault = handle_mm_fault(vma, addr,
arch/s390/mm/gmap_helpers.c
250
if (fault & VM_FAULT_OOM)
arch/sh/mm/fault.c
314
unsigned long address, vm_fault_t fault)
arch/sh/mm/fault.c
320
if (fault_signal_pending(fault, regs)) {
arch/sh/mm/fault.c
327
if (!(fault & VM_FAULT_RETRY))
arch/sh/mm/fault.c
330
if (!(fault & VM_FAULT_ERROR))
arch/sh/mm/fault.c
333
if (fault & VM_FAULT_OOM) {
arch/sh/mm/fault.c
347
if (fault & VM_FAULT_SIGBUS)
arch/sh/mm/fault.c
349
else if (fault & VM_FAULT_SIGSEGV)
arch/sh/mm/fault.c
397
vm_fault_t fault;
arch/sh/mm/fault.c
469
fault = handle_mm_fault(vma, address, flags, regs);
arch/sh/mm/fault.c
471
if (unlikely(fault & (VM_FAULT_RETRY | VM_FAULT_ERROR)))
arch/sh/mm/fault.c
472
if (mm_fault_error(regs, error_code, address, fault))
arch/sh/mm/fault.c
476
if (fault & VM_FAULT_COMPLETED)
arch/sh/mm/fault.c
479
if (fault & VM_FAULT_RETRY) {
arch/sparc/mm/fault_32.c
120
vm_fault_t fault;
arch/sparc/mm/fault_32.c
179
fault = handle_mm_fault(vma, address, flags, regs);
arch/sparc/mm/fault_32.c
181
if (fault_signal_pending(fault, regs)) {
arch/sparc/mm/fault_32.c
188
if (fault & VM_FAULT_COMPLETED)
arch/sparc/mm/fault_32.c
191
if (unlikely(fault & VM_FAULT_ERROR)) {
arch/sparc/mm/fault_32.c
192
if (fault & VM_FAULT_OOM)
arch/sparc/mm/fault_32.c
194
else if (fault & VM_FAULT_SIGSEGV)
arch/sparc/mm/fault_32.c
196
else if (fault & VM_FAULT_SIGBUS)
arch/sparc/mm/fault_32.c
201
if (fault & VM_FAULT_RETRY) {
arch/sparc/mm/fault_64.c
274
vm_fault_t fault;
arch/sparc/mm/fault_64.c
429
fault = handle_mm_fault(vma, address, flags, regs);
arch/sparc/mm/fault_64.c
431
if (fault_signal_pending(fault, regs)) {
arch/sparc/mm/fault_64.c
440
if (fault & VM_FAULT_COMPLETED)
arch/sparc/mm/fault_64.c
443
if (unlikely(fault & VM_FAULT_ERROR)) {
arch/sparc/mm/fault_64.c
444
if (fault & VM_FAULT_OOM)
arch/sparc/mm/fault_64.c
446
else if (fault & VM_FAULT_SIGSEGV)
arch/sparc/mm/fault_64.c
448
else if (fault & VM_FAULT_SIGBUS)
arch/sparc/mm/fault_64.c
453
if (fault & VM_FAULT_RETRY) {
arch/um/kernel/trap.c
178
vm_fault_t fault;
arch/um/kernel/trap.c
180
fault = handle_mm_fault(vma, address, flags, NULL);
arch/um/kernel/trap.c
182
if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current))
arch/um/kernel/trap.c
186
if (fault & VM_FAULT_COMPLETED)
arch/um/kernel/trap.c
189
if (unlikely(fault & VM_FAULT_ERROR)) {
arch/um/kernel/trap.c
190
if (fault & VM_FAULT_OOM) {
arch/um/kernel/trap.c
192
} else if (fault & VM_FAULT_SIGSEGV) {
arch/um/kernel/trap.c
194
} else if (fault & VM_FAULT_SIGBUS) {
arch/um/kernel/trap.c
200
if (fault & VM_FAULT_RETRY) {
arch/x86/coco/sev/vc-handle.c
228
goto fault;
arch/x86/coco/sev/vc-handle.c
237
goto fault;
arch/x86/coco/sev/vc-handle.c
246
goto fault;
arch/x86/coco/sev/vc-handle.c
255
goto fault;
arch/x86/coco/sev/vc-handle.c
265
fault:
arch/x86/coco/sev/vc-handle.c
306
goto fault;
arch/x86/coco/sev/vc-handle.c
315
goto fault;
arch/x86/coco/sev/vc-handle.c
324
goto fault;
arch/x86/coco/sev/vc-handle.c
332
goto fault;
arch/x86/coco/sev/vc-handle.c
343
fault:
arch/x86/coco/sev/vc-handle.c
84
goto fault;
arch/x86/coco/sev/vc-handle.c
88
goto fault;
arch/x86/coco/sev/vc-handle.c
94
fault:
arch/x86/entry/vdso/vma.c
120
.fault = vdso_fault,
arch/x86/entry/vdso/vma.c
125
.fault = vvar_vclock_fault,
arch/x86/hyperv/nested.c
105
goto fault;
arch/x86/hyperv/nested.c
114
goto fault;
arch/x86/hyperv/nested.c
126
fault:
arch/x86/hyperv/nested.c
29
goto fault;
arch/x86/hyperv/nested.c
37
goto fault;
arch/x86/hyperv/nested.c
50
fault:
arch/x86/hyperv/nested.c
97
goto fault;
arch/x86/include/asm/kvm_host.h
2257
void kvm_inject_page_fault(struct kvm_vcpu *vcpu, struct x86_exception *fault);
arch/x86/include/asm/kvm_host.h
2259
struct x86_exception *fault);
arch/x86/include/asm/kvm_host.h
466
int (*page_fault)(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault);
arch/x86/include/asm/kvm_host.h
468
struct x86_exception *fault);
arch/x86/kernel/cpu/sgx/encl.c
687
.fault = sgx_vma_fault,
arch/x86/kernel/cpu/sgx/virt.c
97
.fault = sgx_vepc_fault,
arch/x86/kvm/kvm_emulate.h
131
struct x86_exception *fault, bool system);
arch/x86/kvm/kvm_emulate.h
143
struct x86_exception *fault, bool system);
arch/x86/kvm/kvm_emulate.h
153
struct x86_exception *fault);
arch/x86/kvm/kvm_emulate.h
163
struct x86_exception *fault);
arch/x86/kvm/kvm_emulate.h
175
struct x86_exception *fault);
arch/x86/kvm/kvm_emulate.h
190
struct x86_exception *fault);
arch/x86/kvm/mmu.h
209
bool fault;
arch/x86/kvm/mmu.h
213
fault = (mmu->permissions[index] >> pte_access) & 1;
arch/x86/kvm/mmu.h
232
fault |= (pkru_bits != 0);
arch/x86/kvm/mmu.h
235
return -(u32)fault & errcode;
arch/x86/kvm/mmu/mmu.c
3032
kvm_pfn_t pfn, struct kvm_page_fault *fault)
arch/x86/kvm/mmu/mmu.c
3043
bool host_writable = !fault || fault->map_writable;
arch/x86/kvm/mmu/mmu.c
3044
bool prefetch = !fault || fault->prefetch;
arch/x86/kvm/mmu/mmu.c
3045
bool write_fault = fault && fault->write;
arch/x86/kvm/mmu/mmu.c
3312
static u8 kvm_gmem_max_mapping_level(struct kvm *kvm, struct kvm_page_fault *fault,
arch/x86/kvm/mmu/mmu.c
3320
if (fault) {
arch/x86/kvm/mmu/mmu.c
3321
pfn = fault->pfn;
arch/x86/kvm/mmu/mmu.c
3322
max_level = fault->max_level;
arch/x86/kvm/mmu/mmu.c
3345
int kvm_mmu_max_mapping_level(struct kvm *kvm, struct kvm_page_fault *fault,
arch/x86/kvm/mmu/mmu.c
3354
if (fault) {
arch/x86/kvm/mmu/mmu.c
3355
max_level = fault->max_level;
arch/x86/kvm/mmu/mmu.c
3356
is_private = fault->is_private;
arch/x86/kvm/mmu/mmu.c
3373
host_level = kvm_gmem_max_mapping_level(kvm, fault, slot, gfn,
arch/x86/kvm/mmu/mmu.c
3380
void kvm_mmu_hugepage_adjust(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault)
arch/x86/kvm/mmu/mmu.c
3382
struct kvm_memory_slot *slot = fault->slot;
arch/x86/kvm/mmu/mmu.c
3385
fault->huge_page_disallowed = fault->exec && fault->nx_huge_page_workaround_enabled;
arch/x86/kvm/mmu/mmu.c
3387
if (unlikely(fault->max_level == PG_LEVEL_4K))
arch/x86/kvm/mmu/mmu.c
3390
if (is_error_noslot_pfn(fault->pfn))
arch/x86/kvm/mmu/mmu.c
3400
fault->req_level = kvm_mmu_max_mapping_level(vcpu->kvm, fault,
arch/x86/kvm/mmu/mmu.c
3401
fault->slot, fault->gfn);
arch/x86/kvm/mmu/mmu.c
3402
if (fault->req_level == PG_LEVEL_4K || fault->huge_page_disallowed)
arch/x86/kvm/mmu/mmu.c
3409
fault->goal_level = fault->req_level;
arch/x86/kvm/mmu/mmu.c
3410
mask = KVM_PAGES_PER_HPAGE(fault->goal_level) - 1;
arch/x86/kvm/mmu/mmu.c
3411
VM_BUG_ON((fault->gfn & mask) != (fault->pfn & mask));
arch/x86/kvm/mmu/mmu.c
3412
fault->pfn &= ~mask;
arch/x86/kvm/mmu/mmu.c
3415
void disallowed_hugepage_adjust(struct kvm_page_fault *fault, u64 spte, int cur_level)
arch/x86/kvm/mmu/mmu.c
3418
cur_level == fault->goal_level &&
arch/x86/kvm/mmu/mmu.c
3431
fault->pfn |= fault->gfn & page_mask;
arch/x86/kvm/mmu/mmu.c
3432
fault->goal_level--;
arch/x86/kvm/mmu/mmu.c
3436
static int direct_map(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault)
arch/x86/kvm/mmu/mmu.c
3441
gfn_t base_gfn = fault->gfn;
arch/x86/kvm/mmu/mmu.c
3443
kvm_mmu_hugepage_adjust(vcpu, fault);
arch/x86/kvm/mmu/mmu.c
3445
trace_kvm_mmu_spte_requested(fault);
arch/x86/kvm/mmu/mmu.c
3446
for_each_shadow_entry(vcpu, fault->addr, it) {
arch/x86/kvm/mmu/mmu.c
3451
if (fault->nx_huge_page_workaround_enabled)
arch/x86/kvm/mmu/mmu.c
3452
disallowed_hugepage_adjust(fault, *it.sptep, it.level);
arch/x86/kvm/mmu/mmu.c
3454
base_gfn = gfn_round_for_level(fault->gfn, it.level);
arch/x86/kvm/mmu/mmu.c
3455
if (it.level == fault->goal_level)
arch/x86/kvm/mmu/mmu.c
3463
if (fault->huge_page_disallowed)
arch/x86/kvm/mmu/mmu.c
3465
fault->req_level >= it.level);
arch/x86/kvm/mmu/mmu.c
3468
if (WARN_ON_ONCE(it.level != fault->goal_level))
arch/x86/kvm/mmu/mmu.c
3471
ret = mmu_set_spte(vcpu, fault->slot, it.sptep, ACC_ALL,
arch/x86/kvm/mmu/mmu.c
3472
base_gfn, fault->pfn, fault);
arch/x86/kvm/mmu/mmu.c
3487
static int kvm_handle_error_pfn(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault)
arch/x86/kvm/mmu/mmu.c
3489
if (is_sigpending_pfn(fault->pfn)) {
arch/x86/kvm/mmu/mmu.c
3499
if (fault->pfn == KVM_PFN_ERR_RO_FAULT)
arch/x86/kvm/mmu/mmu.c
3502
if (fault->pfn == KVM_PFN_ERR_HWPOISON) {
arch/x86/kvm/mmu/mmu.c
3503
kvm_send_hwpoison_signal(fault->slot, fault->gfn);
arch/x86/kvm/mmu/mmu.c
3511
struct kvm_page_fault *fault,
arch/x86/kvm/mmu/mmu.c
3514
gva_t gva = fault->is_tdp ? 0 : fault->addr;
arch/x86/kvm/mmu/mmu.c
3516
if (fault->is_private) {
arch/x86/kvm/mmu/mmu.c
3517
kvm_mmu_prepare_memory_fault_exit(vcpu, fault);
arch/x86/kvm/mmu/mmu.c
3521
vcpu_cache_mmio_info(vcpu, gva, fault->gfn,
arch/x86/kvm/mmu/mmu.c
3524
fault->slot = NULL;
arch/x86/kvm/mmu/mmu.c
3525
fault->pfn = KVM_PFN_NOSLOT;
arch/x86/kvm/mmu/mmu.c
3526
fault->map_writable = false;
arch/x86/kvm/mmu/mmu.c
3543
if (unlikely(fault->gfn > kvm_mmu_max_gfn()))
arch/x86/kvm/mmu/mmu.c
3549
static bool page_fault_can_be_fast(struct kvm *kvm, struct kvm_page_fault *fault)
arch/x86/kvm/mmu/mmu.c
3557
if (fault->rsvd)
arch/x86/kvm/mmu/mmu.c
3577
fault->is_private != kvm_mem_is_private(kvm, fault->gfn))
arch/x86/kvm/mmu/mmu.c
3595
if (!fault->present)
arch/x86/kvm/mmu/mmu.c
3602
return fault->write;
arch/x86/kvm/mmu/mmu.c
3610
struct kvm_page_fault *fault,
arch/x86/kvm/mmu/mmu.c
3629
mark_page_dirty_in_slot(vcpu->kvm, fault->slot, fault->gfn);
arch/x86/kvm/mmu/mmu.c
3660
static int fast_page_fault(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault)
arch/x86/kvm/mmu/mmu.c
3668
if (!page_fault_can_be_fast(vcpu->kvm, fault))
arch/x86/kvm/mmu/mmu.c
3677
sptep = kvm_tdp_mmu_fast_pf_get_last_sptep(vcpu, fault->gfn, &spte);
arch/x86/kvm/mmu/mmu.c
3679
sptep = fast_pf_get_last_sptep(vcpu, fault->addr, &spte);
arch/x86/kvm/mmu/mmu.c
3706
if (is_access_allowed(fault, spte)) {
arch/x86/kvm/mmu/mmu.c
3734
if (fault->write && is_mmu_writable_spte(spte)) {
arch/x86/kvm/mmu/mmu.c
3748
kvm_slot_dirty_track_enabled(fault->slot))
arch/x86/kvm/mmu/mmu.c
3754
!is_access_allowed(fault, new_spte))
arch/x86/kvm/mmu/mmu.c
3762
if (fast_pf_fix_direct_spte(vcpu, fault, sptep, spte, new_spte)) {
arch/x86/kvm/mmu/mmu.c
3774
trace_fast_page_fault(vcpu, fault, sptep, spte, ret);
arch/x86/kvm/mmu/mmu.c
4478
struct kvm_page_fault *fault)
arch/x86/kvm/mmu/mmu.c
4480
if (unlikely(fault->rsvd))
arch/x86/kvm/mmu/mmu.c
4483
if (!fault->present || !fault->write)
arch/x86/kvm/mmu/mmu.c
4490
if (kvm_gfn_is_write_tracked(vcpu->kvm, fault->slot, fault->gfn))
arch/x86/kvm/mmu/mmu.c
4519
struct kvm_page_fault *fault)
arch/x86/kvm/mmu/mmu.c
4524
arch.gfn = fault->gfn;
arch/x86/kvm/mmu/mmu.c
4525
arch.error_code = fault->error_code;
arch/x86/kvm/mmu/mmu.c
4532
return kvm_setup_async_pf(vcpu, fault->addr,
arch/x86/kvm/mmu/mmu.c
4533
kvm_vcpu_gfn_to_hva(vcpu, fault->gfn), &arch);
arch/x86/kvm/mmu/mmu.c
4568
struct kvm_page_fault *fault, int r)
arch/x86/kvm/mmu/mmu.c
4570
kvm_release_faultin_page(vcpu->kvm, fault->refcounted_page,
arch/x86/kvm/mmu/mmu.c
4571
r == RET_PF_RETRY, fault->map_writable);
arch/x86/kvm/mmu/mmu.c
4575
struct kvm_page_fault *fault)
arch/x86/kvm/mmu/mmu.c
4579
if (!kvm_slot_has_gmem(fault->slot)) {
arch/x86/kvm/mmu/mmu.c
4580
kvm_mmu_prepare_memory_fault_exit(vcpu, fault);
arch/x86/kvm/mmu/mmu.c
4584
r = kvm_gmem_get_pfn(vcpu->kvm, fault->slot, fault->gfn, &fault->pfn,
arch/x86/kvm/mmu/mmu.c
4585
&fault->refcounted_page, &max_order);
arch/x86/kvm/mmu/mmu.c
4587
kvm_mmu_prepare_memory_fault_exit(vcpu, fault);
arch/x86/kvm/mmu/mmu.c
4591
fault->map_writable = !(fault->slot->flags & KVM_MEM_READONLY);
arch/x86/kvm/mmu/mmu.c
4592
fault->max_level = kvm_max_level_for_order(max_order);
arch/x86/kvm/mmu/mmu.c
4598
struct kvm_page_fault *fault)
arch/x86/kvm/mmu/mmu.c
4600
unsigned int foll = fault->write ? FOLL_WRITE : 0;
arch/x86/kvm/mmu/mmu.c
4602
if (fault->is_private || kvm_memslot_is_gmem_only(fault->slot))
arch/x86/kvm/mmu/mmu.c
4603
return kvm_mmu_faultin_pfn_gmem(vcpu, fault);
arch/x86/kvm/mmu/mmu.c
4606
fault->pfn = __kvm_faultin_pfn(fault->slot, fault->gfn, foll,
arch/x86/kvm/mmu/mmu.c
4607
&fault->map_writable, &fault->refcounted_page);
arch/x86/kvm/mmu/mmu.c
4615
if (fault->pfn != KVM_PFN_ERR_NEEDS_IO)
arch/x86/kvm/mmu/mmu.c
4618
if (!fault->prefetch && kvm_can_do_async_pf(vcpu)) {
arch/x86/kvm/mmu/mmu.c
4619
trace_kvm_try_async_get_page(fault->addr, fault->gfn);
arch/x86/kvm/mmu/mmu.c
4620
if (kvm_find_async_pf_gfn(vcpu, fault->gfn)) {
arch/x86/kvm/mmu/mmu.c
4621
trace_kvm_async_pf_repeated_fault(fault->addr, fault->gfn);
arch/x86/kvm/mmu/mmu.c
4624
} else if (kvm_arch_setup_async_pf(vcpu, fault)) {
arch/x86/kvm/mmu/mmu.c
4636
fault->pfn = __kvm_faultin_pfn(fault->slot, fault->gfn, foll,
arch/x86/kvm/mmu/mmu.c
4637
&fault->map_writable, &fault->refcounted_page);
arch/x86/kvm/mmu/mmu.c
4643
struct kvm_page_fault *fault, unsigned int access)
arch/x86/kvm/mmu/mmu.c
4645
struct kvm_memory_slot *slot = fault->slot;
arch/x86/kvm/mmu/mmu.c
4649
if (KVM_BUG_ON(kvm_is_gfn_alias(kvm, fault->gfn), kvm))
arch/x86/kvm/mmu/mmu.c
4658
fault->mmu_seq = vcpu->kvm->mmu_invalidate_seq;
arch/x86/kvm/mmu/mmu.c
4665
if (fault->is_private != kvm_mem_is_private(kvm, fault->gfn)) {
arch/x86/kvm/mmu/mmu.c
4666
kvm_mmu_prepare_memory_fault_exit(vcpu, fault);
arch/x86/kvm/mmu/mmu.c
4671
return kvm_handle_noslot_fault(vcpu, fault, access);
arch/x86/kvm/mmu/mmu.c
4681
if (fault->prefetch)
arch/x86/kvm/mmu/mmu.c
4700
return kvm_handle_noslot_fault(vcpu, fault, access);
arch/x86/kvm/mmu/mmu.c
4733
if (mmu_invalidate_retry_gfn_unsafe(kvm, fault->mmu_seq, fault->gfn))
arch/x86/kvm/mmu/mmu.c
4736
ret = __kvm_mmu_faultin_pfn(vcpu, fault);
arch/x86/kvm/mmu/mmu.c
4740
if (unlikely(is_error_pfn(fault->pfn)))
arch/x86/kvm/mmu/mmu.c
4741
return kvm_handle_error_pfn(vcpu, fault);
arch/x86/kvm/mmu/mmu.c
4743
if (WARN_ON_ONCE(!fault->slot || is_noslot_pfn(fault->pfn)))
arch/x86/kvm/mmu/mmu.c
4744
return kvm_handle_noslot_fault(vcpu, fault, access);
arch/x86/kvm/mmu/mmu.c
4753
if (mmu_invalidate_retry_gfn_unsafe(kvm, fault->mmu_seq, fault->gfn)) {
arch/x86/kvm/mmu/mmu.c
4754
kvm_mmu_finish_page_fault(vcpu, fault, RET_PF_RETRY);
arch/x86/kvm/mmu/mmu.c
4766
struct kvm_page_fault *fault)
arch/x86/kvm/mmu/mmu.c
4790
return fault->slot &&
arch/x86/kvm/mmu/mmu.c
4791
mmu_invalidate_retry_gfn(vcpu->kvm, fault->mmu_seq, fault->gfn);
arch/x86/kvm/mmu/mmu.c
4794
static int direct_page_fault(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault)
arch/x86/kvm/mmu/mmu.c
4802
if (page_fault_handle_page_track(vcpu, fault))
arch/x86/kvm/mmu/mmu.c
4805
r = fast_page_fault(vcpu, fault);
arch/x86/kvm/mmu/mmu.c
4813
r = kvm_mmu_faultin_pfn(vcpu, fault, ACC_ALL);
arch/x86/kvm/mmu/mmu.c
4820
if (is_page_fault_stale(vcpu, fault))
arch/x86/kvm/mmu/mmu.c
4827
r = direct_map(vcpu, fault);
arch/x86/kvm/mmu/mmu.c
4830
kvm_mmu_finish_page_fault(vcpu, fault, r);
arch/x86/kvm/mmu/mmu.c
4836
struct kvm_page_fault *fault)
arch/x86/kvm/mmu/mmu.c
4839
fault->max_level = PG_LEVEL_2M;
arch/x86/kvm/mmu/mmu.c
4840
return direct_page_fault(vcpu, fault);
arch/x86/kvm/mmu/mmu.c
4889
struct kvm_page_fault *fault)
arch/x86/kvm/mmu/mmu.c
4893
if (page_fault_handle_page_track(vcpu, fault))
arch/x86/kvm/mmu/mmu.c
4896
r = fast_page_fault(vcpu, fault);
arch/x86/kvm/mmu/mmu.c
4904
r = kvm_mmu_faultin_pfn(vcpu, fault, ACC_ALL);
arch/x86/kvm/mmu/mmu.c
4911
if (is_page_fault_stale(vcpu, fault))
arch/x86/kvm/mmu/mmu.c
4914
r = kvm_tdp_mmu_map(vcpu, fault);
arch/x86/kvm/mmu/mmu.c
4917
kvm_mmu_finish_page_fault(vcpu, fault, r);
arch/x86/kvm/mmu/mmu.c
4923
int kvm_tdp_page_fault(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault)
arch/x86/kvm/mmu/mmu.c
4927
return kvm_tdp_mmu_page_fault(vcpu, fault);
arch/x86/kvm/mmu/mmu.c
4930
return direct_page_fault(vcpu, fault);
arch/x86/kvm/mmu/mmu.c
5038
struct kvm_page_fault fault = {
arch/x86/kvm/mmu/mmu.c
5065
kvm_assert_gmem_invalidate_lock_held(fault.slot);
arch/x86/kvm/mmu/mmu.c
5070
if (kvm_gfn_is_write_tracked(kvm, fault.slot, fault.gfn))
arch/x86/kvm/mmu/mmu.c
5092
r = kvm_tdp_mmu_map(vcpu, &fault);
arch/x86/kvm/mmu/mmu_internal.h
293
int kvm_tdp_page_fault(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault);
arch/x86/kvm/mmu/mmu_internal.h
333
struct kvm_page_fault *fault)
arch/x86/kvm/mmu/mmu_internal.h
335
kvm_prepare_memory_fault_exit(vcpu, fault->gfn << PAGE_SHIFT,
arch/x86/kvm/mmu/mmu_internal.h
336
PAGE_SIZE, fault->write, fault->exec,
arch/x86/kvm/mmu/mmu_internal.h
337
fault->is_private);
arch/x86/kvm/mmu/mmu_internal.h
344
struct kvm_page_fault fault = {
arch/x86/kvm/mmu/mmu_internal.h
372
fault.gfn = gpa_to_gfn(fault.addr) & ~kvm_gfn_direct_bits(vcpu->kvm);
arch/x86/kvm/mmu/mmu_internal.h
373
fault.slot = kvm_vcpu_gfn_to_memslot(vcpu, fault.gfn);
arch/x86/kvm/mmu/mmu_internal.h
380
if (IS_ENABLED(CONFIG_MITIGATION_RETPOLINE) && fault.is_tdp)
arch/x86/kvm/mmu/mmu_internal.h
381
r = kvm_tdp_page_fault(vcpu, &fault);
arch/x86/kvm/mmu/mmu_internal.h
383
r = vcpu->arch.mmu->page_fault(vcpu, &fault);
arch/x86/kvm/mmu/mmu_internal.h
390
if (r == RET_PF_EMULATE && fault.is_private) {
arch/x86/kvm/mmu/mmu_internal.h
392
kvm_mmu_prepare_memory_fault_exit(vcpu, &fault);
arch/x86/kvm/mmu/mmu_internal.h
396
if (fault.write_fault_to_shadow_pgtable && emulation_type)
arch/x86/kvm/mmu/mmu_internal.h
399
*level = fault.goal_level;
arch/x86/kvm/mmu/mmu_internal.h
404
int kvm_mmu_max_mapping_level(struct kvm *kvm, struct kvm_page_fault *fault,
arch/x86/kvm/mmu/mmu_internal.h
406
void kvm_mmu_hugepage_adjust(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault);
arch/x86/kvm/mmu/mmu_internal.h
407
void disallowed_hugepage_adjust(struct kvm_page_fault *fault, u64 spte, int cur_level);
arch/x86/kvm/mmu/mmutrace.h
260
TP_PROTO(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault,
arch/x86/kvm/mmu/mmutrace.h
262
TP_ARGS(vcpu, fault, sptep, old_spte, ret),
arch/x86/kvm/mmu/mmutrace.h
276
__entry->cr2_or_gpa = fault->addr;
arch/x86/kvm/mmu/mmutrace.h
277
__entry->error_code = fault->error_code;
arch/x86/kvm/mmu/mmutrace.h
375
TP_PROTO(struct kvm_page_fault *fault),
arch/x86/kvm/mmu/mmutrace.h
376
TP_ARGS(fault),
arch/x86/kvm/mmu/mmutrace.h
385
__entry->gfn = fault->gfn;
arch/x86/kvm/mmu/mmutrace.h
386
__entry->pfn = fault->pfn | (fault->gfn & (KVM_PAGES_PER_HPAGE(fault->goal_level) - 1));
arch/x86/kvm/mmu/mmutrace.h
387
__entry->level = fault->goal_level;
arch/x86/kvm/mmu/paging_tmpl.h
249
ret = __try_cmpxchg_user(ptep_user, &orig_pte, pte, fault);
arch/x86/kvm/mmu/paging_tmpl.h
380
nested_access, &walker->fault);
arch/x86/kvm/mmu/paging_tmpl.h
448
real_gpa = kvm_translate_gpa(vcpu, mmu, gfn_to_gpa(gfn), access, &walker->fault);
arch/x86/kvm/mmu/paging_tmpl.h
481
walker->fault.vector = PF_VECTOR;
arch/x86/kvm/mmu/paging_tmpl.h
482
walker->fault.error_code_valid = true;
arch/x86/kvm/mmu/paging_tmpl.h
483
walker->fault.error_code = errcode;
arch/x86/kvm/mmu/paging_tmpl.h
500
walker->fault.exit_qualification = 0;
arch/x86/kvm/mmu/paging_tmpl.h
503
walker->fault.exit_qualification |= EPT_VIOLATION_ACC_WRITE;
arch/x86/kvm/mmu/paging_tmpl.h
505
walker->fault.exit_qualification |= EPT_VIOLATION_ACC_READ;
arch/x86/kvm/mmu/paging_tmpl.h
507
walker->fault.exit_qualification |= EPT_VIOLATION_ACC_INSTR;
arch/x86/kvm/mmu/paging_tmpl.h
513
walker->fault.exit_qualification |= EPT_VIOLATION_RWX_TO_PROT(pte_access);
arch/x86/kvm/mmu/paging_tmpl.h
516
walker->fault.address = addr;
arch/x86/kvm/mmu/paging_tmpl.h
517
walker->fault.nested_page_fault = mmu != vcpu->arch.walk_mmu;
arch/x86/kvm/mmu/paging_tmpl.h
518
walker->fault.async_page_fault = false;
arch/x86/kvm/mmu/paging_tmpl.h
520
trace_kvm_mmu_walker_error(walker->fault.error_code);
arch/x86/kvm/mmu/paging_tmpl.h
614
static int FNAME(fetch)(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault,
arch/x86/kvm/mmu/paging_tmpl.h
621
gfn_t base_gfn = fault->gfn;
arch/x86/kvm/mmu/paging_tmpl.h
652
for_each_shadow_entry(vcpu, fault->addr, it) {
arch/x86/kvm/mmu/paging_tmpl.h
700
if (fault->write && table_gfn == fault->gfn)
arch/x86/kvm/mmu/paging_tmpl.h
701
fault->write_fault_to_shadow_pgtable = true;
arch/x86/kvm/mmu/paging_tmpl.h
710
kvm_mmu_hugepage_adjust(vcpu, fault);
arch/x86/kvm/mmu/paging_tmpl.h
712
trace_kvm_mmu_spte_requested(fault);
arch/x86/kvm/mmu/paging_tmpl.h
719
if (fault->nx_huge_page_workaround_enabled)
arch/x86/kvm/mmu/paging_tmpl.h
720
disallowed_hugepage_adjust(fault, *it.sptep, it.level);
arch/x86/kvm/mmu/paging_tmpl.h
722
base_gfn = gfn_round_for_level(fault->gfn, it.level);
arch/x86/kvm/mmu/paging_tmpl.h
723
if (it.level == fault->goal_level)
arch/x86/kvm/mmu/paging_tmpl.h
734
if (fault->huge_page_disallowed)
arch/x86/kvm/mmu/paging_tmpl.h
736
fault->req_level >= it.level);
arch/x86/kvm/mmu/paging_tmpl.h
739
if (WARN_ON_ONCE(it.level != fault->goal_level))
arch/x86/kvm/mmu/paging_tmpl.h
742
ret = mmu_set_spte(vcpu, fault->slot, it.sptep, gw->pte_access,
arch/x86/kvm/mmu/paging_tmpl.h
743
base_gfn, fault->pfn, fault);
arch/x86/kvm/mmu/paging_tmpl.h
765
static int FNAME(page_fault)(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault)
arch/x86/kvm/mmu/paging_tmpl.h
770
WARN_ON_ONCE(fault->is_tdp);
arch/x86/kvm/mmu/paging_tmpl.h
777
r = FNAME(walk_addr)(&walker, vcpu, fault->addr,
arch/x86/kvm/mmu/paging_tmpl.h
778
fault->error_code & ~PFERR_RSVD_MASK);
arch/x86/kvm/mmu/paging_tmpl.h
784
if (!fault->prefetch)
arch/x86/kvm/mmu/paging_tmpl.h
785
kvm_inject_emulated_page_fault(vcpu, &walker.fault);
arch/x86/kvm/mmu/paging_tmpl.h
790
fault->gfn = walker.gfn;
arch/x86/kvm/mmu/paging_tmpl.h
791
fault->max_level = walker.level;
arch/x86/kvm/mmu/paging_tmpl.h
792
fault->slot = kvm_vcpu_gfn_to_memslot(vcpu, fault->gfn);
arch/x86/kvm/mmu/paging_tmpl.h
794
if (page_fault_handle_page_track(vcpu, fault)) {
arch/x86/kvm/mmu/paging_tmpl.h
795
shadow_page_table_clear_flood(vcpu, fault->addr);
arch/x86/kvm/mmu/paging_tmpl.h
803
r = kvm_mmu_faultin_pfn(vcpu, fault, walker.pte_access);
arch/x86/kvm/mmu/paging_tmpl.h
814
if (fault->write && !(walker.pte_access & ACC_WRITE_MASK) &&
arch/x86/kvm/mmu/paging_tmpl.h
815
!is_cr0_wp(vcpu->arch.mmu) && !fault->user && fault->slot) {
arch/x86/kvm/mmu/paging_tmpl.h
833
if (is_page_fault_stale(vcpu, fault))
arch/x86/kvm/mmu/paging_tmpl.h
839
r = FNAME(fetch)(vcpu, fault, &walker);
arch/x86/kvm/mmu/paging_tmpl.h
842
kvm_mmu_finish_page_fault(vcpu, fault, r);
arch/x86/kvm/mmu/paging_tmpl.h
879
*exception = walker.fault;
arch/x86/kvm/mmu/paging_tmpl.h
92
struct x86_exception fault;
arch/x86/kvm/mmu/spte.h
493
static inline bool is_access_allowed(struct kvm_page_fault *fault, u64 spte)
arch/x86/kvm/mmu/spte.h
495
if (fault->exec)
arch/x86/kvm/mmu/spte.h
498
if (fault->write)
arch/x86/kvm/mmu/tdp_mmu.c
1169
struct kvm_page_fault *fault,
arch/x86/kvm/mmu/tdp_mmu.c
1177
if (WARN_ON_ONCE(sp->role.level != fault->goal_level))
arch/x86/kvm/mmu/tdp_mmu.c
1181
(fault->prefetch || is_access_allowed(fault, iter->old_spte)) &&
arch/x86/kvm/mmu/tdp_mmu.c
1183
WARN_ON_ONCE(fault->pfn != spte_to_pfn(iter->old_spte));
arch/x86/kvm/mmu/tdp_mmu.c
1187
if (unlikely(!fault->slot))
arch/x86/kvm/mmu/tdp_mmu.c
1190
wrprot = make_spte(vcpu, sp, fault->slot, ACC_ALL, iter->gfn,
arch/x86/kvm/mmu/tdp_mmu.c
1191
fault->pfn, iter->old_spte, fault->prefetch,
arch/x86/kvm/mmu/tdp_mmu.c
1192
false, fault->map_writable, &new_spte);
arch/x86/kvm/mmu/tdp_mmu.c
1208
if (wrprot && fault->write)
arch/x86/kvm/mmu/tdp_mmu.c
1263
int kvm_tdp_mmu_map(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault)
arch/x86/kvm/mmu/tdp_mmu.c
1265
struct kvm_mmu_page *root = tdp_mmu_get_root_for_fault(vcpu, fault);
arch/x86/kvm/mmu/tdp_mmu.c
1273
kvm_mmu_hugepage_adjust(vcpu, fault);
arch/x86/kvm/mmu/tdp_mmu.c
1275
trace_kvm_mmu_spte_requested(fault);
arch/x86/kvm/mmu/tdp_mmu.c
1279
for_each_tdp_pte(iter, kvm, root, fault->gfn, fault->gfn + 1) {
arch/x86/kvm/mmu/tdp_mmu.c
1282
if (fault->nx_huge_page_workaround_enabled)
arch/x86/kvm/mmu/tdp_mmu.c
1283
disallowed_hugepage_adjust(fault, iter.old_spte, iter.level);
arch/x86/kvm/mmu/tdp_mmu.c
1292
if (iter.level == fault->goal_level)
arch/x86/kvm/mmu/tdp_mmu.c
1309
sp->nx_huge_page_disallowed = fault->huge_page_disallowed;
arch/x86/kvm/mmu/tdp_mmu.c
1328
if (fault->huge_page_disallowed &&
arch/x86/kvm/mmu/tdp_mmu.c
1329
fault->req_level >= iter.level) {
arch/x86/kvm/mmu/tdp_mmu.c
1341
WARN_ON_ONCE(iter.level == fault->goal_level);
arch/x86/kvm/mmu/tdp_mmu.c
1345
ret = tdp_mmu_map_handle_target_level(vcpu, fault, &iter);
arch/x86/kvm/mmu/tdp_mmu.h
49
struct kvm_page_fault *fault)
arch/x86/kvm/mmu/tdp_mmu.h
51
if (unlikely(!kvm_is_addr_direct(vcpu->kvm, fault->addr)))
arch/x86/kvm/mmu/tdp_mmu.h
74
int kvm_tdp_mmu_map(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault);
arch/x86/kvm/svm/nested.c
37
struct x86_exception *fault)
arch/x86/kvm/svm/nested.c
49
vmcb->control.exit_info_2 = fault->address;
arch/x86/kvm/svm/nested.c
53
vmcb->control.exit_info_1 |= fault->error_code;
arch/x86/kvm/svm/svm_ops.h
12
_ASM_EXTABLE(1b, %l[fault]) \
arch/x86/kvm/svm/svm_ops.h
13
::: clobber : fault); \
arch/x86/kvm/svm/svm_ops.h
15
fault: \
arch/x86/kvm/svm/svm_ops.h
22
_ASM_EXTABLE(1b, %l[fault]) \
arch/x86/kvm/svm/svm_ops.h
23
:: op1 : clobber : fault); \
arch/x86/kvm/svm/svm_ops.h
25
fault: \
arch/x86/kvm/svm/svm_ops.h
32
_ASM_EXTABLE(1b, %l[fault]) \
arch/x86/kvm/svm/svm_ops.h
33
:: op1, op2 : clobber : fault); \
arch/x86/kvm/svm/svm_ops.h
35
fault: \
arch/x86/kvm/vmx/nested.c
414
struct x86_exception *fault)
arch/x86/kvm/vmx/nested.c
442
if (fault->error_code & PFERR_RSVD_MASK) {
arch/x86/kvm/vmx/nested.c
446
exit_qualification = fault->exit_qualification;
arch/x86/kvm/vmx/nested.c
461
fault->address);
arch/x86/kvm/vmx/nested.c
465
vmcs12->guest_physical_address = fault->address;
arch/x86/kvm/vmx/sgx.c
28
bool fault;
arch/x86/kvm/vmx/sgx.c
38
fault = true;
arch/x86/kvm/vmx/sgx.c
41
fault = is_noncanonical_address(*gva, vcpu, 0);
arch/x86/kvm/vmx/sgx.c
44
fault = (s.unusable) ||
arch/x86/kvm/vmx/sgx.c
50
if (fault)
arch/x86/kvm/vmx/sgx.c
52
return fault ? -EINVAL : 0;
arch/x86/kvm/vmx/vmx.c
2994
_ASM_EXTABLE(1b, %l[fault])
arch/x86/kvm/vmx/vmx.c
2996
: : fault);
arch/x86/kvm/vmx/vmx.c
2999
fault:
arch/x86/kvm/vmx/vmx.c
541
noinstr void vmread_error_trampoline2(unsigned long field, bool fault)
arch/x86/kvm/vmx/vmx.c
543
if (fault) {
arch/x86/kvm/vmx/vmx.c
801
_ASM_EXTABLE(1b, %l[fault])
arch/x86/kvm/vmx/vmx.c
802
::: "cc", "memory" : fault);
arch/x86/kvm/vmx/vmx.c
807
fault:
arch/x86/kvm/vmx/vmx_ops.h
196
_ASM_EXTABLE(1b, %l[fault]) \
arch/x86/kvm/vmx/vmx_ops.h
197
: : op1 : "cc" : error, fault); \
arch/x86/kvm/vmx/vmx_ops.h
204
fault: \
arch/x86/kvm/vmx/vmx_ops.h
213
_ASM_EXTABLE(1b, %l[fault]) \
arch/x86/kvm/vmx/vmx_ops.h
214
: : op1, op2 : "cc" : error, fault); \
arch/x86/kvm/vmx/vmx_ops.h
221
fault: \
arch/x86/kvm/vmx/vmx_ops.h
39
void vmread_error_trampoline2(unsigned long field, bool fault);
arch/x86/kvm/x86.c
1001
if ((fault->error_code & PFERR_PRESENT_MASK) &&
arch/x86/kvm/x86.c
1002
!(fault->error_code & PFERR_RSVD_MASK))
arch/x86/kvm/x86.c
1003
kvm_mmu_invalidate_addr(vcpu, fault_mmu, fault->address,
arch/x86/kvm/x86.c
1006
fault_mmu->inject_page_fault(vcpu, fault);
arch/x86/kvm/x86.c
13947
struct x86_exception fault;
arch/x86/kvm/x86.c
13954
fault.vector = PF_VECTOR;
arch/x86/kvm/x86.c
13955
fault.error_code_valid = true;
arch/x86/kvm/x86.c
13956
fault.error_code = 0;
arch/x86/kvm/x86.c
13957
fault.nested_page_fault = false;
arch/x86/kvm/x86.c
13958
fault.address = work->arch.token;
arch/x86/kvm/x86.c
13959
fault.async_page_fault = true;
arch/x86/kvm/x86.c
13960
kvm_inject_page_fault(vcpu, &fault);
arch/x86/kvm/x86.c
14113
struct x86_exception fault;
arch/x86/kvm/x86.c
14118
mmu->gva_to_gpa(vcpu, mmu, gva, access, &fault) != INVALID_GPA) {
arch/x86/kvm/x86.c
14124
fault.vector = PF_VECTOR;
arch/x86/kvm/x86.c
14125
fault.error_code_valid = true;
arch/x86/kvm/x86.c
14126
fault.error_code = error_code;
arch/x86/kvm/x86.c
14127
fault.nested_page_fault = false;
arch/x86/kvm/x86.c
14128
fault.address = gva;
arch/x86/kvm/x86.c
14129
fault.async_page_fault = false;
arch/x86/kvm/x86.c
14131
vcpu->arch.walk_mmu->inject_page_fault(vcpu, &fault);
arch/x86/kvm/x86.c
971
void kvm_inject_page_fault(struct kvm_vcpu *vcpu, struct x86_exception *fault)
arch/x86/kvm/x86.c
979
if (is_guest_mode(vcpu) && fault->async_page_fault)
arch/x86/kvm/x86.c
981
true, fault->error_code,
arch/x86/kvm/x86.c
982
true, fault->address);
arch/x86/kvm/x86.c
984
kvm_queue_exception_e_p(vcpu, PF_VECTOR, fault->error_code,
arch/x86/kvm/x86.c
985
fault->address);
arch/x86/kvm/x86.c
989
struct x86_exception *fault)
arch/x86/kvm/x86.c
992
WARN_ON_ONCE(fault->vector != PF_VECTOR);
arch/x86/kvm/x86.c
994
fault_mmu = fault->nested_page_fault ? vcpu->arch.mmu :
arch/x86/mm/fault.c
1214
vm_fault_t fault;
arch/x86/mm/fault.c
1334
fault = handle_mm_fault(vma, address, flags | FAULT_FLAG_VMA_LOCK, regs);
arch/x86/mm/fault.c
1335
if (!(fault & (VM_FAULT_RETRY | VM_FAULT_COMPLETED)))
arch/x86/mm/fault.c
1338
if (!(fault & VM_FAULT_RETRY)) {
arch/x86/mm/fault.c
1343
if (fault & VM_FAULT_MAJOR)
arch/x86/mm/fault.c
1347
if (fault_signal_pending(fault, regs)) {
arch/x86/mm/fault.c
1385
fault = handle_mm_fault(vma, address, flags, regs);
arch/x86/mm/fault.c
1387
if (fault_signal_pending(fault, regs)) {
arch/x86/mm/fault.c
1400
if (fault & VM_FAULT_COMPLETED)
arch/x86/mm/fault.c
1408
if (unlikely(fault & VM_FAULT_RETRY)) {
arch/x86/mm/fault.c
1415
if (likely(!(fault & VM_FAULT_ERROR)))
arch/x86/mm/fault.c
1424
if (fault & VM_FAULT_OOM) {
arch/x86/mm/fault.c
1440
if (fault & (VM_FAULT_SIGBUS|VM_FAULT_HWPOISON|
arch/x86/mm/fault.c
1442
do_sigbus(regs, error_code, address, fault);
arch/x86/mm/fault.c
1443
else if (fault & VM_FAULT_SIGSEGV)
arch/x86/mm/fault.c
908
vm_fault_t fault)
arch/x86/mm/fault.c
929
if (fault & (VM_FAULT_HWPOISON|VM_FAULT_HWPOISON_LARGE)) {
arch/x86/mm/fault.c
936
if (fault & VM_FAULT_HWPOISON_LARGE)
arch/x86/mm/fault.c
937
lsb = hstate_index_to_shift(VM_FAULT_GET_HINDEX(fault));
arch/x86/mm/fault.c
938
if (fault & VM_FAULT_HWPOISON)
arch/xtensa/include/asm/futex.h
158
[fault] "I" (-EFAULT)
arch/xtensa/include/asm/futex.h
45
[fault] "I" (-EFAULT) \
arch/xtensa/include/asm/futex.h
69
: [oparg] "r" (arg), [fault] "I" (-EFAULT) \
arch/xtensa/mm/fault.c
159
fault = handle_mm_fault(vma, address, flags, regs);
arch/xtensa/mm/fault.c
161
if (fault_signal_pending(fault, regs)) {
arch/xtensa/mm/fault.c
168
if (fault & VM_FAULT_COMPLETED)
arch/xtensa/mm/fault.c
171
if (unlikely(fault & VM_FAULT_ERROR)) {
arch/xtensa/mm/fault.c
172
if (fault & VM_FAULT_OOM)
arch/xtensa/mm/fault.c
174
else if (fault & VM_FAULT_SIGSEGV)
arch/xtensa/mm/fault.c
176
else if (fault & VM_FAULT_SIGBUS)
arch/xtensa/mm/fault.c
181
if (fault & VM_FAULT_RETRY) {
arch/xtensa/mm/fault.c
97
vm_fault_t fault;
drivers/accel/amdxdna/amdxdna_ubuf.c
94
.fault = amdxdna_ubuf_vm_fault,
drivers/android/binder.c
6023
.fault = binder_vm_fault,
drivers/android/tests/binder_alloc_kunit.c
475
.fault = binder_vm_fault,
drivers/char/agp/alpha-agp.c
42
.fault = alpha_core_agp_vm_fault,
drivers/crypto/marvell/octeontx2/otx2_cpt_hw_types.h
327
u64 fault:1;
drivers/crypto/marvell/octeontx2/otx2_cpt_hw_types.h
347
u64 fault:1;
drivers/crypto/marvell/octeontx2/otx2_cpt_hw_types.h
435
u64 fault:1;
drivers/crypto/marvell/octeontx2/otx2_cptlf.c
169
irq_misc.s.fault = 0x1;
drivers/crypto/marvell/octeontx2/otx2_cptlf.c
210
if (irq_misc.s.fault) {
drivers/crypto/marvell/octeontx2/otx2_cptlf.c
213
irq_misc_ack.s.fault = 0x1;
drivers/dax/device.c
290
.fault = dev_dax_fault,
drivers/dma-buf/heaps/cma_heap.c
195
.fault = cma_heap_vm_fault,
drivers/dma-buf/udmabuf.c
93
.fault = udmabuf_vm_fault,
drivers/dma/idxd/irq.c
224
struct idxd_evl_fault *fault = container_of(work, struct idxd_evl_fault, work);
drivers/dma/idxd/irq.c
225
struct idxd_wq *wq = fault->wq;
drivers/dma/idxd/irq.c
229
struct __evl_entry *entry_head = fault->entry;
drivers/dma/idxd/irq.c
237
switch (fault->status) {
drivers/dma/idxd/irq.c
262
dev_dbg_ratelimited(dev, "Unrecognized error code: %#x\n", fault->status);
drivers/dma/idxd/irq.c
284
switch (fault->status) {
drivers/dma/idxd/irq.c
308
kmem_cache_free(idxd->evl_cache, fault);
drivers/dma/idxd/irq.c
325
struct idxd_evl_fault *fault;
drivers/dma/idxd/irq.c
334
fault = kmem_cache_alloc(idxd->evl_cache, GFP_ATOMIC);
drivers/dma/idxd/irq.c
335
if (fault) {
drivers/dma/idxd/irq.c
338
fault->wq = wq;
drivers/dma/idxd/irq.c
339
fault->status = status;
drivers/dma/idxd/irq.c
340
memcpy(&fault->entry, entry_head, ent_size);
drivers/dma/idxd/irq.c
341
INIT_WORK(&fault->work, idxd_evl_fault_work);
drivers/dma/idxd/irq.c
342
queue_work(wq->wq, &fault->work);
drivers/firmware/cirrus/cs_dsp.c
3256
unsigned int fault[6];
drivers/firmware/cirrus/cs_dsp.c
3267
fault);
drivers/firmware/cirrus/cs_dsp.c
3274
*fault & HALO_AHBM_FLAGS_ERR_MASK,
drivers/firmware/cirrus/cs_dsp.c
3275
(*fault & HALO_AHBM_CORE_ERR_ADDR_MASK) >>
drivers/firmware/cirrus/cs_dsp.c
3279
fault);
drivers/firmware/cirrus/cs_dsp.c
3285
cs_dsp_warn(dsp, "AHB: SYS_ADDR: 0x%x\n", *fault);
drivers/firmware/cirrus/cs_dsp.c
3288
fault, ARRAY_SIZE(fault));
drivers/firmware/cirrus/cs_dsp.c
3294
cs_dsp_warn(dsp, "XM: STATUS:0x%x ADDR:0x%x\n", fault[1], fault[0]);
drivers/firmware/cirrus/cs_dsp.c
3295
cs_dsp_warn(dsp, "YM: STATUS:0x%x ADDR:0x%x\n", fault[3], fault[2]);
drivers/firmware/cirrus/cs_dsp.c
3296
cs_dsp_warn(dsp, "PM: STATUS:0x%x ADDR:0x%x\n", fault[5], fault[4]);
drivers/gpib/common/gpib_os.c
1008
fault = copy_from_user(board->buffer, userbuf, (board->buffer_length < remain) ?
drivers/gpib/common/gpib_os.c
1010
if (fault) {
drivers/gpib/common/gpib_os.c
1030
if (fault == 0)
drivers/gpib/common/gpib_os.c
1031
fault = copy_to_user((void __user *)arg, &cmd, sizeof(cmd));
drivers/gpib/common/gpib_os.c
1039
if (!no_clear_io_in_prog || fault)
drivers/gpib/common/gpib_os.c
1044
if (fault)
drivers/gpib/common/gpib_os.c
1057
int fault;
drivers/gpib/common/gpib_os.c
1060
fault = copy_from_user(&write_cmd, (void __user *)arg, sizeof(write_cmd));
drivers/gpib/common/gpib_os.c
1061
if (fault)
drivers/gpib/common/gpib_os.c
1095
fault = copy_from_user(board->buffer, userbuf, (board->buffer_length < remain) ?
drivers/gpib/common/gpib_os.c
1097
if (fault) {
drivers/gpib/common/gpib_os.c
1117
if (fault == 0)
drivers/gpib/common/gpib_os.c
1118
fault = copy_to_user((void __user *)arg, &write_cmd, sizeof(write_cmd));
drivers/gpib/common/gpib_os.c
1124
if (fault)
drivers/gpib/common/gpib_os.c
963
int fault = 0;
drivers/gpio/gpio-max3191x.c
168
__assign_bit(i, max3191x->fault, val);
drivers/gpio/gpio-max3191x.c
182
if (!max3191x->ignore_uv && test_bit(chipnum, max3191x->fault))
drivers/gpio/gpio-max3191x.c
364
max3191x->fault = devm_kcalloc(dev, n, sizeof(long), GFP_KERNEL);
drivers/gpio/gpio-max3191x.c
369
!max3191x->fault || !max3191x->xfer.rx_buf)
drivers/gpio/gpio-max3191x.c
87
unsigned long *fault;
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
149
.fault = amdgpu_gem_fault,
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
423
struct amdgpu_gmc_fault *fault;
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
438
fault = &gmc->fault_ring[gmc->fault_hash[hash].idx];
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
439
while (fault->timestamp >= stamp) {
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
442
if (atomic64_read(&fault->key) == key) {
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
450
if (fault->timestamp_expiry != 0 &&
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
451
amdgpu_ih_ts_after(fault->timestamp_expiry,
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
458
tmp = fault->timestamp;
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
459
fault = &gmc->fault_ring[fault->next];
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
462
if (fault->timestamp >= tmp)
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
467
fault = &gmc->fault_ring[gmc->last_fault];
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
468
atomic64_set(&fault->key, key);
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
469
fault->timestamp = timestamp;
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
472
fault->next = gmc->fault_hash[hash].idx;
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
493
struct amdgpu_gmc_fault *fault;
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
516
fault = &gmc->fault_ring[gmc->fault_hash[hash].idx];
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
518
if (atomic64_read(&fault->key) == key) {
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
523
fault->timestamp_expiry = last_ts;
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
527
tmp = fault->timestamp;
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
528
fault = &gmc->fault_ring[fault->next];
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
529
} while (fault->timestamp < tmp);
drivers/gpu/drm/armada/armada_gem.c
32
.fault = armada_gem_vm_fault,
drivers/gpu/drm/drm_gem_shmem_helper.c
648
.fault = drm_gem_shmem_fault,
drivers/gpu/drm/etnaviv/etnaviv_gem.c
556
.fault = etnaviv_gem_fault,
drivers/gpu/drm/gma500/fbdev.c
46
.fault = psb_fbdev_vm_fault,
drivers/gpu/drm/gma500/gem.c
129
.fault = psb_gem_fault,
drivers/gpu/drm/i915/display/intel_display_irq.c
101
u32 fault;
drivers/gpu/drm/i915/display/intel_display_irq.c
1178
{ .fault = MTL_PLANE_ATS_FAULT, .handle = handle_plane_ats_fault, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1179
{ .fault = MTL_PIPEDMC_ATS_FAULT, .handle = handle_pipedmc_ats_fault, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1180
{ .fault = GEN12_PIPEDMC_FAULT, .handle = handle_pipedmc_fault, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1181
{ .fault = GEN11_PIPE_PLANE5_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_5, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1182
{ .fault = GEN9_PIPE_PLANE4_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_4, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1183
{ .fault = GEN9_PIPE_PLANE3_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_3, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1184
{ .fault = GEN9_PIPE_PLANE2_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_2, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1185
{ .fault = GEN9_PIPE_PLANE1_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_1, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1186
{ .fault = GEN9_PIPE_CURSOR_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_CURSOR, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1191
{ .fault = GEN12_PIPEDMC_FAULT, .handle = handle_pipedmc_fault, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1192
{ .fault = GEN11_PIPE_PLANE7_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_7, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1193
{ .fault = GEN11_PIPE_PLANE6_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_6, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1194
{ .fault = GEN11_PIPE_PLANE5_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_5, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1195
{ .fault = GEN9_PIPE_PLANE4_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_4, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1196
{ .fault = GEN9_PIPE_PLANE3_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_3, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1197
{ .fault = GEN9_PIPE_PLANE2_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_2, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1198
{ .fault = GEN9_PIPE_PLANE1_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_1, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1199
{ .fault = GEN9_PIPE_CURSOR_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_CURSOR, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1204
{ .fault = GEN11_PIPE_PLANE7_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_7, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1205
{ .fault = GEN11_PIPE_PLANE6_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_6, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1206
{ .fault = GEN11_PIPE_PLANE5_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_5, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1207
{ .fault = GEN9_PIPE_PLANE4_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_4, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1208
{ .fault = GEN9_PIPE_PLANE3_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_3, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1209
{ .fault = GEN9_PIPE_PLANE2_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_2, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1210
{ .fault = GEN9_PIPE_PLANE1_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_1, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1211
{ .fault = GEN9_PIPE_CURSOR_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_CURSOR, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1216
{ .fault = GEN9_PIPE_PLANE4_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_4, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1217
{ .fault = GEN9_PIPE_PLANE3_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_3, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1218
{ .fault = GEN9_PIPE_PLANE2_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_2, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1219
{ .fault = GEN9_PIPE_PLANE1_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_1, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1220
{ .fault = GEN9_PIPE_CURSOR_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_CURSOR, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1225
{ .fault = GEN8_PIPE_SPRITE_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_SPRITE0, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1226
{ .fault = GEN8_PIPE_PRIMARY_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_PRIMARY, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1227
{ .fault = GEN8_PIPE_CURSOR_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_CURSOR, },
drivers/gpu/drm/i915/display/intel_display_irq.c
133
for (handler = handlers; handler && handler->fault; handler++) {
drivers/gpu/drm/i915/display/intel_display_irq.c
134
if ((fault_errors & handler->fault) == 0)
drivers/gpu/drm/i915/display/intel_display_irq.c
138
fault_errors &= ~handler->fault;
drivers/gpu/drm/i915/display/intel_display_irq.c
1848
{ .fault = SPRITEB_INVALID_GTT_STATUS, .handle = handle_plane_fault, .plane_id = PLANE_SPRITE1, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1849
{ .fault = SPRITEA_INVALID_GTT_STATUS, .handle = handle_plane_fault, .plane_id = PLANE_SPRITE0, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1850
{ .fault = PLANEA_INVALID_GTT_STATUS, .handle = handle_plane_fault, .plane_id = PLANE_PRIMARY, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1851
{ .fault = CURSORA_INVALID_GTT_STATUS, .handle = handle_plane_fault, .plane_id = PLANE_CURSOR, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1852
{ .fault = SPRITED_INVALID_GTT_STATUS, .handle = handle_plane_fault, .plane_id = PLANE_SPRITE1, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1853
{ .fault = SPRITEC_INVALID_GTT_STATUS, .handle = handle_plane_fault, .plane_id = PLANE_SPRITE0, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1854
{ .fault = PLANEB_INVALID_GTT_STATUS, .handle = handle_plane_fault, .plane_id = PLANE_PRIMARY, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1855
{ .fault = CURSORB_INVALID_GTT_STATUS, .handle = handle_plane_fault, .plane_id = PLANE_CURSOR, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1856
{ .fault = SPRITEF_INVALID_GTT_STATUS, .handle = handle_plane_fault, .plane_id = PLANE_SPRITE1, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1857
{ .fault = SPRITEE_INVALID_GTT_STATUS, .handle = handle_plane_fault, .plane_id = PLANE_SPRITE0, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1858
{ .fault = PLANEC_INVALID_GTT_STATUS, .handle = handle_plane_fault, .plane_id = PLANE_PRIMARY, },
drivers/gpu/drm/i915/display/intel_display_irq.c
1859
{ .fault = CURSORC_INVALID_GTT_STATUS, .handle = handle_plane_fault, .plane_id = PLANE_CURSOR, },
drivers/gpu/drm/i915/display/intel_display_irq.c
744
{ .fault = ERR_INT_SPRITE_A_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_SPRITE0, },
drivers/gpu/drm/i915/display/intel_display_irq.c
745
{ .fault = ERR_INT_PRIMARY_A_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_PRIMARY, },
drivers/gpu/drm/i915/display/intel_display_irq.c
746
{ .fault = ERR_INT_CURSOR_A_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_CURSOR, },
drivers/gpu/drm/i915/display/intel_display_irq.c
747
{ .fault = ERR_INT_SPRITE_B_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_SPRITE0, },
drivers/gpu/drm/i915/display/intel_display_irq.c
748
{ .fault = ERR_INT_PRIMARY_B_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_PRIMARY, },
drivers/gpu/drm/i915/display/intel_display_irq.c
749
{ .fault = ERR_INT_CURSOR_B_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_CURSOR, },
drivers/gpu/drm/i915/display/intel_display_irq.c
750
{ .fault = ERR_INT_SPRITE_C_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_SPRITE0, },
drivers/gpu/drm/i915/display/intel_display_irq.c
751
{ .fault = ERR_INT_PRIMARY_C_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_PRIMARY, },
drivers/gpu/drm/i915/display/intel_display_irq.c
752
{ .fault = ERR_INT_CURSOR_C_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_CURSOR, },
drivers/gpu/drm/i915/display/intel_display_irq.c
861
{ .fault = GTT_FAULT_SPRITE_A_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_SPRITE0, },
drivers/gpu/drm/i915/display/intel_display_irq.c
862
{ .fault = GTT_FAULT_SPRITE_B_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_SPRITE0, },
drivers/gpu/drm/i915/display/intel_display_irq.c
863
{ .fault = GTT_FAULT_PRIMARY_A_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_PRIMARY, },
drivers/gpu/drm/i915/display/intel_display_irq.c
864
{ .fault = GTT_FAULT_PRIMARY_B_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_PRIMARY, },
drivers/gpu/drm/i915/display/intel_display_irq.c
865
{ .fault = GTT_FAULT_CURSOR_A_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_CURSOR, },
drivers/gpu/drm/i915/display/intel_display_irq.c
866
{ .fault = GTT_FAULT_CURSOR_B_FAULT, .handle = handle_plane_fault, .plane_id = PLANE_CURSOR, },
drivers/gpu/drm/i915/gem/i915_gem_mman.c
940
.fault = vm_fault_gtt,
drivers/gpu/drm/i915/gem/i915_gem_mman.c
947
.fault = vm_fault_cpu,
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
1184
.fault = vm_fault_ttm,
drivers/gpu/drm/i915/gt/intel_gt.c
307
u32 fault;
drivers/gpu/drm/i915/gt/intel_gt.c
309
fault = GEN6_RING_FAULT_REG_READ(engine);
drivers/gpu/drm/i915/gt/intel_gt.c
311
if (fault & RING_FAULT_VALID) {
drivers/gpu/drm/i915/gt/intel_gt.c
317
fault & RING_FAULT_VADDR_MASK,
drivers/gpu/drm/i915/gt/intel_gt.c
318
fault & RING_FAULT_GTTSEL_MASK ?
drivers/gpu/drm/i915/gt/intel_gt.c
320
REG_FIELD_GET(RING_FAULT_SRCID_MASK, fault),
drivers/gpu/drm/i915/gt/intel_gt.c
321
REG_FIELD_GET(RING_FAULT_FAULT_TYPE_MASK, fault));
drivers/gpu/drm/i915/gt/intel_gt.c
326
static void gen8_report_fault(struct intel_gt *gt, u32 fault,
drivers/gpu/drm/i915/gt/intel_gt.c
342
REG_FIELD_GET(RING_FAULT_ENGINE_ID_MASK, fault),
drivers/gpu/drm/i915/gt/intel_gt.c
343
REG_FIELD_GET(RING_FAULT_SRCID_MASK, fault),
drivers/gpu/drm/i915/gt/intel_gt.c
344
REG_FIELD_GET(RING_FAULT_FAULT_TYPE_MASK, fault));
drivers/gpu/drm/i915/gt/intel_gt.c
349
u32 fault;
drivers/gpu/drm/i915/gt/intel_gt.c
358
fault = intel_gt_mcr_read_any(gt, XEHP_RING_FAULT_REG);
drivers/gpu/drm/i915/gt/intel_gt.c
359
if (fault & RING_FAULT_VALID)
drivers/gpu/drm/i915/gt/intel_gt.c
360
gen8_report_fault(gt, fault,
drivers/gpu/drm/i915/gt/intel_gt.c
369
u32 fault;
drivers/gpu/drm/i915/gt/intel_gt.c
381
fault = intel_uncore_read(uncore, fault_reg);
drivers/gpu/drm/i915/gt/intel_gt.c
382
if (fault & RING_FAULT_VALID)
drivers/gpu/drm/i915/gt/intel_gt.c
383
gen8_report_fault(gt, fault,
drivers/gpu/drm/lima/lima_mmu.c
42
u32 fault = mmu_read(LIMA_MMU_PAGE_FAULT_ADDR);
drivers/gpu/drm/lima/lima_mmu.c
45
lima_ip_name(ip), fault, LIMA_MMU_STATUS_BUS_ID(status),
drivers/gpu/drm/msm/msm_gem.c
1178
.fault = msm_gem_fault,
drivers/gpu/drm/nouveau/include/nvkm/core/layout.h
22
NVKM_LAYOUT_ONCE(NVKM_SUBDEV_FAULT , struct nvkm_fault , fault)
drivers/gpu/drm/nouveau/include/nvkm/subdev/top.h
16
int fault;
drivers/gpu/drm/nouveau/include/nvkm/subdev/top.h
29
struct nvkm_subdev *nvkm_top_fault(struct nvkm_device *, int fault);
drivers/gpu/drm/nouveau/nouveau_gem.c
70
.fault = nouveau_ttm_fault,
drivers/gpu/drm/nouveau/nouveau_svm.c
1013
buffer->fault = kvzalloc_objs(*buffer->fault, buffer->entries);
drivers/gpu/drm/nouveau/nouveau_svm.c
1014
if (!buffer->fault)
drivers/gpu/drm/nouveau/nouveau_svm.c
412
struct nouveau_svm_fault *fault)
drivers/gpu/drm/nouveau/nouveau_svm.c
414
nouveau_svm_fault_cancel(svm, fault->inst,
drivers/gpu/drm/nouveau/nouveau_svm.c
415
fault->hub,
drivers/gpu/drm/nouveau/nouveau_svm.c
416
fault->gpc,
drivers/gpu/drm/nouveau/nouveau_svm.c
417
fault->client);
drivers/gpu/drm/nouveau/nouveau_svm.c
421
nouveau_svm_fault_priority(u8 fault)
drivers/gpu/drm/nouveau/nouveau_svm.c
423
switch (fault) {
drivers/gpu/drm/nouveau/nouveau_svm.c
469
struct nouveau_svm_fault *fault;
drivers/gpu/drm/nouveau/nouveau_svm.c
477
if (!buffer->fault[buffer->fault_nr]) {
drivers/gpu/drm/nouveau/nouveau_svm.c
478
fault = kmalloc_obj(*fault);
drivers/gpu/drm/nouveau/nouveau_svm.c
479
if (WARN_ON(!fault)) {
drivers/gpu/drm/nouveau/nouveau_svm.c
483
buffer->fault[buffer->fault_nr] = fault;
drivers/gpu/drm/nouveau/nouveau_svm.c
486
fault = buffer->fault[buffer->fault_nr++];
drivers/gpu/drm/nouveau/nouveau_svm.c
487
fault->inst = inst;
drivers/gpu/drm/nouveau/nouveau_svm.c
488
fault->addr = (u64)addrhi << 32 | addrlo;
drivers/gpu/drm/nouveau/nouveau_svm.c
489
fault->time = (u64)timehi << 32 | timelo;
drivers/gpu/drm/nouveau/nouveau_svm.c
490
fault->engine = engine;
drivers/gpu/drm/nouveau/nouveau_svm.c
491
fault->gpc = gpc;
drivers/gpu/drm/nouveau/nouveau_svm.c
492
fault->hub = hub;
drivers/gpu/drm/nouveau/nouveau_svm.c
493
fault->access = (info & 0x000f0000) >> 16;
drivers/gpu/drm/nouveau/nouveau_svm.c
494
fault->client = client;
drivers/gpu/drm/nouveau/nouveau_svm.c
495
fault->fault = (info & 0x0000001f);
drivers/gpu/drm/nouveau/nouveau_svm.c
498
fault->inst, fault->addr, fault->access);
drivers/gpu/drm/nouveau/nouveau_svm.c
66
u8 fault;
drivers/gpu/drm/nouveau/nouveau_svm.c
68
} **fault;
drivers/gpu/drm/nouveau/nouveau_svm.c
754
sort(buffer->fault, buffer->fault_nr, sizeof(*buffer->fault),
drivers/gpu/drm/nouveau/nouveau_svm.c
760
if (!svmm || buffer->fault[fi]->inst != inst) {
drivers/gpu/drm/nouveau/nouveau_svm.c
762
nouveau_ivmm_find(svm, buffer->fault[fi]->inst);
drivers/gpu/drm/nouveau/nouveau_svm.c
764
inst = buffer->fault[fi]->inst;
drivers/gpu/drm/nouveau/nouveau_svm.c
767
buffer->fault[fi]->svmm = svmm;
drivers/gpu/drm/nouveau/nouveau_svm.c
783
if (!(svmm = buffer->fault[fi]->svmm)) {
drivers/gpu/drm/nouveau/nouveau_svm.c
784
nouveau_svm_fault_cancel_fault(svm, buffer->fault[fi]);
drivers/gpu/drm/nouveau/nouveau_svm.c
787
SVMM_DBG(svmm, "addr %016llx", buffer->fault[fi]->addr);
drivers/gpu/drm/nouveau/nouveau_svm.c
792
start = buffer->fault[fi]->addr;
drivers/gpu/drm/nouveau/nouveau_svm.c
809
switch (buffer->fault[fi]->access) {
drivers/gpu/drm/nouveau/nouveau_svm.c
826
nouveau_svm_fault_cancel_fault(svm, buffer->fault[fi]);
drivers/gpu/drm/nouveau/nouveau_svm.c
852
if (buffer->fault[fn]->svmm != svmm ||
drivers/gpu/drm/nouveau/nouveau_svm.c
853
buffer->fault[fn]->addr >= limit ||
drivers/gpu/drm/nouveau/nouveau_svm.c
854
(buffer->fault[fi]->access == FAULT_ACCESS_READ &&
drivers/gpu/drm/nouveau/nouveau_svm.c
856
(buffer->fault[fi]->access != FAULT_ACCESS_READ &&
drivers/gpu/drm/nouveau/nouveau_svm.c
857
buffer->fault[fi]->access != FAULT_ACCESS_PREFETCH &&
drivers/gpu/drm/nouveau/nouveau_svm.c
859
(buffer->fault[fi]->access != FAULT_ACCESS_READ &&
drivers/gpu/drm/nouveau/nouveau_svm.c
860
buffer->fault[fi]->access != FAULT_ACCESS_WRITE &&
drivers/gpu/drm/nouveau/nouveau_svm.c
861
buffer->fault[fi]->access != FAULT_ACCESS_PREFETCH &&
drivers/gpu/drm/nouveau/nouveau_svm.c
869
struct nouveau_svm_fault *fault =
drivers/gpu/drm/nouveau/nouveau_svm.c
870
buffer->fault[fi++];
drivers/gpu/drm/nouveau/nouveau_svm.c
872
nouveau_svm_fault_cancel_fault(svm, fault);
drivers/gpu/drm/nouveau/nouveau_svm.c
974
if (buffer->fault) {
drivers/gpu/drm/nouveau/nouveau_svm.c
975
for (i = 0; buffer->fault[i] && i < buffer->entries; i++)
drivers/gpu/drm/nouveau/nouveau_svm.c
976
kfree(buffer->fault[i]);
drivers/gpu/drm/nouveau/nouveau_svm.c
977
kvfree(buffer->fault);
drivers/gpu/drm/nouveau/nvkm/engine/device/base.c
2083
.fault = { 0x00000001, gp100_fault_new },
drivers/gpu/drm/nouveau/nvkm/engine/device/base.c
2116
.fault = { 0x00000001, gp100_fault_new },
drivers/gpu/drm/nouveau/nvkm/engine/device/base.c
2150
.fault = { 0x00000001, gp100_fault_new },
drivers/gpu/drm/nouveau/nvkm/engine/device/base.c
2184
.fault = { 0x00000001, gp100_fault_new },
drivers/gpu/drm/nouveau/nvkm/engine/device/base.c
2218
.fault = { 0x00000001, gp100_fault_new },
drivers/gpu/drm/nouveau/nvkm/engine/device/base.c
2252
.fault = { 0x00000001, gp100_fault_new },
drivers/gpu/drm/nouveau/nvkm/engine/device/base.c
2284
.fault = { 0x00000001, gp10b_fault_new },
drivers/gpu/drm/nouveau/nvkm/engine/device/base.c
2310
.fault = { 0x00000001, gv100_fault_new },
drivers/gpu/drm/nouveau/nvkm/engine/device/base.c
2345
.fault = { 0x00000001, tu102_fault_new },
drivers/gpu/drm/nouveau/nvkm/engine/device/base.c
2380
.fault = { 0x00000001, tu102_fault_new },
drivers/gpu/drm/nouveau/nvkm/engine/device/base.c
2415
.fault = { 0x00000001, tu102_fault_new },
drivers/gpu/drm/nouveau/nvkm/engine/device/base.c
2450
.fault = { 0x00000001, tu102_fault_new },
drivers/gpu/drm/nouveau/nvkm/engine/device/base.c
2485
.fault = { 0x00000001, tu102_fault_new },
drivers/gpu/drm/nouveau/nvkm/engine/device/base.c
2518
.fault = { 0x00000001, tu102_fault_new },
drivers/gpu/drm/nouveau/nvkm/engine/device/base.c
2542
.fault = { 0x00000001, tu102_fault_new },
drivers/gpu/drm/nouveau/nvkm/engine/device/base.c
2572
.fault = { 0x00000001, tu102_fault_new },
drivers/gpu/drm/nouveau/nvkm/engine/device/base.c
2602
.fault = { 0x00000001, tu102_fault_new },
drivers/gpu/drm/nouveau/nvkm/engine/device/base.c
2632
.fault = { 0x00000001, tu102_fault_new },
drivers/gpu/drm/nouveau/nvkm/engine/device/base.c
2662
.fault = { 0x00000001, tu102_fault_new },
drivers/gpu/drm/nouveau/nvkm/engine/device/base.c
2689
.fault = { 0x00000001, tu102_fault_new },
drivers/gpu/drm/nouveau/nvkm/engine/device/base.c
2707
.fault = { 0x00000001, tu102_fault_new },
drivers/gpu/drm/nouveau/nvkm/engine/device/base.c
2728
.fault = { 0x00000001, tu102_fault_new },
drivers/gpu/drm/nouveau/nvkm/engine/device/base.c
2749
.fault = { 0x00000001, tu102_fault_new },
drivers/gpu/drm/nouveau/nvkm/engine/device/base.c
2770
.fault = { 0x00000001, tu102_fault_new },
drivers/gpu/drm/nouveau/nvkm/engine/device/base.c
2791
.fault = { 0x00000001, tu102_fault_new },
drivers/gpu/drm/nouveau/nvkm/engine/device/user.c
298
else if (device->fault && index-- == 0)
drivers/gpu/drm/nouveau/nvkm/engine/device/user.c
299
sclass = &device->fault->user;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c
228
ENGN_DEBUG(engn, "triggering mmu fault on 0x%02x", engn->fault);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c
233
nvkm_wr32(device, 0x002a30 + (engn->id * 4), 0x00000100 | engn->fault);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c
548
engn = nvkm_runl_find_engn(engn, runl, engn->fault == info->engine);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c
654
if (WARN_ON(engn->fault < 0) || !engn->func->chsw(engn))
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runl.c
359
engn->fault = -1;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runl.c
364
engn->fault = nvkm_top_fault_id(device, engine->subdev.type, engine->subdev.inst);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runl.c
366
if (engn->fault < 0 && fifo->func->mmu_fault) {
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runl.c
371
engn->fault = map->value;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runl.h
30
int fault;
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
105
struct nvkm_fault *fault = nvkm_fault(subdev);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
108
for (i = 0; i < ARRAY_SIZE(fault->buffer); i++) {
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
109
if (i < fault->func->buffer.nr) {
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
110
ret = nvkm_fault_oneinit_buffer(fault, i);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
113
fault->buffer_nr = i + 1;
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
117
ret = nvkm_event_init(&nvkm_fault_ntfy, subdev, 1, fault->buffer_nr, &fault->event);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
121
if (fault->func->oneinit)
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
122
ret = fault->func->oneinit(fault);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
129
struct nvkm_fault *fault = nvkm_fault(subdev);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
132
nvkm_event_ntfy_del(&fault->nrpfb);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
133
nvkm_event_fini(&fault->event);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
135
for (i = 0; i < fault->buffer_nr; i++) {
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
136
if (fault->buffer[i]) {
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
137
nvkm_memory_unref(&fault->buffer[i]->mem);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
138
kfree(fault->buffer[i]);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
142
return fault;
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
158
struct nvkm_fault *fault;
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
159
if (!(fault = *pfault = kzalloc_obj(*fault)))
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
161
nvkm_subdev_ctor(&nvkm_fault, device, type, inst, &fault->subdev);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
162
fault->func = func;
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
163
fault->user.ctor = nvkm_ufault_new;
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
164
fault->user.base = func->user.base;
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
29
struct nvkm_fault *fault = container_of(event, typeof(*fault), event);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
30
fault->func->buffer.intr(fault->buffer[index], false);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
36
struct nvkm_fault *fault = container_of(event, typeof(*fault), event);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
37
fault->func->buffer.intr(fault->buffer[index], true);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
49
struct nvkm_fault *fault = nvkm_fault(subdev);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
50
return fault->func->intr(fault);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
56
struct nvkm_fault *fault = nvkm_fault(subdev);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
57
if (fault->func->fini)
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
58
fault->func->fini(fault);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
65
struct nvkm_fault *fault = nvkm_fault(subdev);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
66
if (fault->func->init)
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
67
fault->func->init(fault);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
72
nvkm_fault_oneinit_buffer(struct nvkm_fault *fault, int id)
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
74
struct nvkm_subdev *subdev = &fault->subdev;
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
81
buffer->fault = fault;
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
83
fault->func->buffer.info(buffer);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
84
fault->buffer[id] = buffer;
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
89
fault->func->buffer.entry_size, 0x1000, true,
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
95
buffer->addr = fault->func->buffer.pin(buffer);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gp100.c
32
struct nvkm_device *device = buffer->fault->subdev.device;
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gp100.c
39
struct nvkm_device *device = buffer->fault->subdev.device;
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gp100.c
46
struct nvkm_device *device = buffer->fault->subdev.device;
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gp100.c
60
buffer->entries = nvkm_rd32(buffer->fault->subdev.device, 0x002a78);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gp100.c
66
gp100_fault_intr(struct nvkm_fault *fault)
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gp100.c
68
nvkm_event_ntfy(&fault->event, 0, NVKM_FAULT_BUFFER_EVENT_PENDING);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
109
struct nvkm_device *device = buffer->fault->subdev.device;
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
122
struct nvkm_fault *fault = container_of(ntfy, typeof(*fault), nrpfb);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
124
schedule_work(&fault->nrpfb_work);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
129
gv100_fault_intr_fault(struct nvkm_fault *fault)
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
131
struct nvkm_subdev *subdev = &fault->subdev;
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
155
gv100_fault_intr(struct nvkm_fault *fault)
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
157
struct nvkm_subdev *subdev = &fault->subdev;
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
162
gv100_fault_intr_fault(fault);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
168
if (fault->buffer[0]) {
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
169
nvkm_event_ntfy(&fault->event, 0, NVKM_FAULT_BUFFER_EVENT_PENDING);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
175
if (fault->buffer[1]) {
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
176
nvkm_event_ntfy(&fault->event, 1, NVKM_FAULT_BUFFER_EVENT_PENDING);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
187
gv100_fault_fini(struct nvkm_fault *fault)
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
189
nvkm_event_ntfy_block(&fault->nrpfb);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
190
flush_work(&fault->nrpfb_work);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
192
if (fault->buffer[0])
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
193
fault->func->buffer.fini(fault->buffer[0]);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
195
nvkm_mask(fault->subdev.device, 0x100a34, 0x80000000, 0x80000000);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
199
gv100_fault_init(struct nvkm_fault *fault)
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
201
nvkm_mask(fault->subdev.device, 0x100a2c, 0x80000000, 0x80000000);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
202
fault->func->buffer.init(fault->buffer[0]);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
203
nvkm_event_ntfy_allow(&fault->nrpfb);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
207
gv100_fault_oneinit(struct nvkm_fault *fault)
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
209
nvkm_event_ntfy_add(&fault->event, 0, NVKM_FAULT_BUFFER_EVENT_PENDING, true,
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
210
gv100_fault_ntfy_nrpfb, &fault->nrpfb);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
33
struct nvkm_fault *fault = container_of(work, typeof(*fault), nrpfb_work);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
34
struct nvkm_fault_buffer *buffer = fault->buffer[0];
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
35
struct nvkm_device *device = fault->subdev.device;
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
44
const u32 base = get * buffer->fault->func->buffer.entry_size;
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
78
struct nvkm_device *device = buffer->fault->subdev.device;
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
89
struct nvkm_device *device = buffer->fault->subdev.device;
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
97
struct nvkm_device *device = buffer->fault->subdev.device;
drivers/gpu/drm/nouveau/nvkm/subdev/fault/priv.h
12
struct nvkm_fault *fault;
drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c
116
tu102_fault_fini(struct nvkm_fault *fault)
drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c
118
nvkm_event_ntfy_block(&fault->nrpfb);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c
119
flush_work(&fault->nrpfb_work);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c
121
if (fault->buffer[0])
drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c
122
fault->func->buffer.fini(fault->buffer[0]);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c
124
nvkm_inth_block(&fault->info_fault);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c
128
tu102_fault_init(struct nvkm_fault *fault)
drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c
130
nvkm_inth_allow(&fault->info_fault);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c
132
fault->func->buffer.init(fault->buffer[0]);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c
133
nvkm_event_ntfy_allow(&fault->nrpfb);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c
137
tu102_fault_oneinit(struct nvkm_fault *fault)
drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c
139
struct nvkm_device *device = fault->subdev.device;
drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c
144
NVKM_INTR_PRIO_NORMAL, &fault->subdev, tu102_fault_info_fault,
drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c
145
&fault->info_fault);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c
149
for (i = 0; i < fault->buffer_nr; i++) {
drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c
151
NVKM_INTR_PRIO_NORMAL, &fault->subdev,
drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c
152
tu102_fault_buffer_notify, &fault->buffer[i]->inth);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c
157
return gv100_fault_oneinit(fault);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c
38
nvkm_event_ntfy(&buffer->fault->event, buffer->id, NVKM_FAULT_BUFFER_EVENT_PENDING);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c
54
struct nvkm_device *device = buffer->fault->subdev.device;
drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c
63
struct nvkm_device *device = buffer->fault->subdev.device;
drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c
75
struct nvkm_device *device = buffer->fault->subdev.device;
drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c
88
struct nvkm_fault *fault = container_of(inth, typeof(*fault), info_fault);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/tu102.c
89
struct nvkm_subdev *subdev = &fault->subdev;
drivers/gpu/drm/nouveau/nvkm/subdev/fault/user.c
42
return nvkm_uevent_add(uevent, &buffer->fault->event, buffer->id,
drivers/gpu/drm/nouveau/nvkm/subdev/fault/user.c
51
struct nvkm_device *device = buffer->fault->subdev.device;
drivers/gpu/drm/nouveau/nvkm/subdev/fault/user.c
62
buffer->fault->func->buffer.fini(buffer);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/user.c
70
buffer->fault->func->buffer.init(buffer);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/user.c
96
struct nvkm_fault *fault = device->fault;
drivers/gpu/drm/nouveau/nvkm/subdev/fault/user.c
97
struct nvkm_fault_buffer *buffer = fault->buffer[fault->func->user.rp];
drivers/gpu/drm/nouveau/nvkm/subdev/top/base.c
100
return info->fault;
drivers/gpu/drm/nouveau/nvkm/subdev/top/base.c
107
nvkm_top_fault(struct nvkm_device *device, int fault)
drivers/gpu/drm/nouveau/nvkm/subdev/top/base.c
113
if (info->fault == fault)
drivers/gpu/drm/nouveau/nvkm/subdev/top/base.c
34
info->fault = -1;
drivers/gpu/drm/nouveau/nvkm/subdev/top/base.c
99
if (info->type == type && info->inst == inst && info->fault >= 0)
drivers/gpu/drm/nouveau/nvkm/subdev/top/ga100.c
52
info->fault = (data & 0x0000007f);
drivers/gpu/drm/nouveau/nvkm/subdev/top/ga100.c
92
info->addr, info->fault, info->runlist < 0 ? 0 : info->runlist,
drivers/gpu/drm/nouveau/nvkm/subdev/top/gk104.c
103
info->addr, info->fault, info->engine, info->runlist,
drivers/gpu/drm/nouveau/nvkm/subdev/top/gk104.c
54
info->fault = (data & 0x000003f8) >> 3;
drivers/gpu/drm/omapdrm/omap_gem.c
1268
.fault = omap_gem_fault,
drivers/gpu/drm/panthor/panthor_device.c
452
.fault = panthor_mmio_vm_fault,
drivers/gpu/drm/panthor/panthor_fw.h
115
u32 fault;
drivers/gpu/drm/panthor/panthor_sched.c
1532
u32 fault;
drivers/gpu/drm/panthor/panthor_sched.c
1538
fault = cs_iface->output->fault;
drivers/gpu/drm/panthor/panthor_sched.c
1569
(unsigned int)CS_EXCEPTION_TYPE(fault),
drivers/gpu/drm/panthor/panthor_sched.c
1570
panthor_exception_name(ptdev, CS_EXCEPTION_TYPE(fault)),
drivers/gpu/drm/panthor/panthor_sched.c
1571
(unsigned int)CS_EXCEPTION_DATA(fault),
drivers/gpu/drm/radeon/radeon_gem.c
77
.fault = radeon_gem_fault,
drivers/gpu/drm/tegra/gem.c
580
.fault = tegra_bo_fault,
drivers/gpu/drm/ttm/ttm_bo_vm.c
474
.fault = ttm_bo_vm_fault,
drivers/gpu/drm/vc4/vc4_bo.c
752
.fault = vc4_fault,
drivers/gpu/drm/vmwgfx/vmwgfx_gem.c
138
.fault = vmw_bo_vm_fault,
drivers/gpu/drm/xe/xe_bo.c
2063
.fault = xe_bo_cpu_fault,
drivers/gpu/drm/xe/xe_device.c
312
.fault = barrier_fault,
drivers/gpu/drm/xe/xe_mmio_gem.c
45
.fault = xe_mmio_gem_vm_fault,
drivers/gpu/drm/xe/xe_svm.c
1153
DECL_SVM_RANGE_COUNT_STATS(fault, PAGEFAULT)
drivers/gpu/drm/xe/xe_svm.c
1183
DECL_SVM_RANGE_US_STATS(fault, PAGEFAULT)
drivers/hid/usbhid/hiddev.c
444
goto fault;
drivers/hid/usbhid/hiddev.c
447
goto fault;
drivers/hid/usbhid/hiddev.c
471
goto fault;
drivers/hid/usbhid/hiddev.c
526
goto fault;
drivers/hid/usbhid/hiddev.c
545
goto fault;
drivers/hid/usbhid/hiddev.c
557
fault:
drivers/hsi/clients/cmt_speech.c
1100
.fault = cs_char_vma_fault,
drivers/hv/mshv_root_main.c
76
.fault = mshv_vp_fault,
drivers/hv/mshv_vtl_main.c
1262
.fault = mshv_vtl_low_fault,
drivers/hv/mshv_vtl_main.c
822
.fault = mshv_vtl_fault,
drivers/hwmon/emc2103.c
207
bool fault = (data->temp[nr].degrees == -128);
drivers/hwmon/emc2103.c
208
return sprintf(buf, "%d\n", fault ? 1 : 0);
drivers/hwmon/emc2103.c
412
bool fault = ((data->fan_tach & 0x1fe0) == 0x1fe0);
drivers/hwmon/emc2103.c
413
return sprintf(buf, "%d\n", fault ? 1 : 0);
drivers/hwmon/ltc2945.c
352
unsigned int fault;
drivers/hwmon/ltc2945.c
355
ret = regmap_read(regmap, LTC2945_FAULT, &fault);
drivers/hwmon/ltc2945.c
359
fault &= attr->index;
drivers/hwmon/ltc2945.c
360
if (fault) /* Clear reported faults in chip register */
drivers/hwmon/ltc2945.c
363
return sysfs_emit(buf, "%d\n", !!fault);
drivers/hwmon/ltc4222.c
105
unsigned int fault;
drivers/hwmon/ltc4222.c
108
ret = regmap_read(regmap, attr->nr, &fault);
drivers/hwmon/ltc4222.c
111
fault &= attr->index;
drivers/hwmon/ltc4222.c
112
if (fault) /* Clear reported faults in chip register */
drivers/hwmon/ltc4222.c
115
return sysfs_emit(buf, "%d\n", !!fault);
drivers/hwmon/ltc4260.c
101
return sysfs_emit(buf, "%d\n", !!fault);
drivers/hwmon/ltc4260.c
90
unsigned int fault;
drivers/hwmon/ltc4260.c
93
ret = regmap_read(regmap, LTC4260_FAULT, &fault);
drivers/hwmon/ltc4260.c
97
fault &= attr->index;
drivers/hwmon/ltc4260.c
98
if (fault) /* Clear reported faults in chip register */
drivers/hwmon/ltc4261.c
141
u8 fault;
drivers/hwmon/ltc4261.c
146
fault = data->regs[LTC4261_FAULT] & attr->index;
drivers/hwmon/ltc4261.c
147
if (fault) /* Clear reported faults in chip register */
drivers/hwmon/ltc4261.c
148
i2c_smbus_write_byte_data(data->client, LTC4261_FAULT, ~fault);
drivers/hwmon/ltc4261.c
150
return sysfs_emit(buf, "%d\n", fault ? 1 : 0);
drivers/hwmon/max16065.c
165
WRITE_ONCE(data->fault[i],
drivers/hwmon/max16065.c
173
WRITE_ONCE(data->fault[0],
drivers/hwmon/max16065.c
174
data->fault[0] | data->fault[1]);
drivers/hwmon/max16065.c
188
int val = READ_ONCE(data->fault[attr2->nr]);
drivers/hwmon/max16065.c
96
int fault[2];
drivers/hwmon/max6620.c
139
u8 fault;
drivers/hwmon/max6620.c
204
data->fault |= (ret >> 4) & (ret & 0x0F);
drivers/hwmon/max6620.c
257
*val = !!(data->fault & BIT(channel));
drivers/hwmon/max6620.c
272
data->fault &= ~BIT(channel);
drivers/hwmon/max6620.c
369
data->fault &= ~BIT(channel);
drivers/hwmon/max6639.c
102
static int max6639_temp_read_fault(struct device *dev, int channel, long *fault)
drivers/hwmon/max6639.c
112
*fault = val & 1;
drivers/hwtracing/intel_th/msu.c
1626
.fault = msc_mmap_fault,
drivers/iio/dac/ad5421.c
207
unsigned int fault;
drivers/iio/dac/ad5421.c
211
fault = ad5421_read(indio_dev, AD5421_REG_FAULT);
drivers/iio/dac/ad5421.c
212
if (!fault)
drivers/iio/dac/ad5421.c
234
if (fault == 0xffff)
drivers/iio/dac/ad5421.c
235
fault = 0;
drivers/iio/dac/ad5421.c
238
events = (old_fault ^ fault) & fault;
drivers/iio/dac/ad5421.c
268
old_fault = fault;
drivers/iio/dac/ad5421.c
269
fault = ad5421_read(indio_dev, AD5421_REG_FAULT);
drivers/iio/dac/ad5421.c
272
if (fault & AD5421_FAULT_TRIGGER_IRQ)
drivers/iio/dac/ad5421.c
275
} while (fault & AD5421_FAULT_TRIGGER_IRQ);
drivers/iio/dac/ad8460.c
375
enum ad8460_fault_type fault,
drivers/iio/dac/ad8460.c
378
return regmap_update_bits(state->regmap, AD8460_CTRL_REG(0x08 + fault),
drivers/iio/dac/ad8460.c
384
enum ad8460_fault_type fault,
drivers/iio/dac/ad8460.c
390
ret = regmap_read(state->regmap, AD8460_CTRL_REG(0x08 + fault), &val);
drivers/iio/dac/ad8460.c
400
enum ad8460_fault_type fault, bool en)
drivers/iio/dac/ad8460.c
402
return regmap_update_bits(state->regmap, AD8460_CTRL_REG(0x08 + fault),
drivers/iio/dac/ad8460.c
408
enum ad8460_fault_type fault, bool *en)
drivers/iio/dac/ad8460.c
413
ret = regmap_read(state->regmap, AD8460_CTRL_REG(0x08 + fault), &val);
drivers/iio/dac/ad8460.c
542
int fault;
drivers/iio/dac/ad8460.c
550
fault = ad8460_select_fault_type(chan->type, dir);
drivers/iio/dac/ad8460.c
551
if (fault < 0)
drivers/iio/dac/ad8460.c
552
return fault;
drivers/iio/dac/ad8460.c
554
return ad8460_set_fault_threshold(state, fault, val);
drivers/iio/dac/ad8460.c
564
int fault;
drivers/iio/dac/ad8460.c
572
fault = ad8460_select_fault_type(chan->type, dir);
drivers/iio/dac/ad8460.c
573
if (fault < 0)
drivers/iio/dac/ad8460.c
574
return fault;
drivers/iio/dac/ad8460.c
576
return ad8460_get_fault_threshold(state, fault, val);
drivers/iio/dac/ad8460.c
585
int fault;
drivers/iio/dac/ad8460.c
590
fault = ad8460_select_fault_type(chan->type, dir);
drivers/iio/dac/ad8460.c
591
if (fault < 0)
drivers/iio/dac/ad8460.c
592
return fault;
drivers/iio/dac/ad8460.c
594
return ad8460_set_fault_threshold_en(state, fault, val);
drivers/iio/dac/ad8460.c
603
int fault, ret;
drivers/iio/dac/ad8460.c
609
fault = ad8460_select_fault_type(chan->type, dir);
drivers/iio/dac/ad8460.c
610
if (fault < 0)
drivers/iio/dac/ad8460.c
611
return fault;
drivers/iio/dac/ad8460.c
613
ret = ad8460_get_fault_threshold_en(state, fault, &en);
drivers/iio/resolver/ad2s1210.c
1339
st->sample.fault = reg_val;
drivers/iio/resolver/ad2s1210.c
1342
ad2s1210_push_events(indio_dev, st->sample.fault, pf->timestamp);
drivers/iio/resolver/ad2s1210.c
162
u8 fault;
drivers/iio/resolver/ad2s1210.c
487
st->sample.fault = reg_val;
drivers/iio/resolver/ad2s1210.c
520
ad2s1210_push_events(indio_dev, st->sample.fault, timestamp);
drivers/iio/temperature/max31856.c
317
bool fault;
drivers/iio/temperature/max31856.c
323
fault = reg_val & faultbit;
drivers/iio/temperature/max31856.c
325
return sysfs_emit(buf, "%d\n", fault);
drivers/iio/temperature/max31865.c
203
bool fault;
drivers/iio/temperature/max31865.c
211
fault = data->buf[0] & faultbit;
drivers/iio/temperature/max31865.c
213
return sysfs_emit(buf, "%d\n", fault);
drivers/infiniband/core/umem_odp.c
325
u64 bcnt, u64 access_mask, bool fault)
drivers/infiniband/core/umem_odp.c
358
if (fault) {
drivers/infiniband/core/umem_odp.c
397
WARN_ON(fault && range.hmm_pfns[pfn_index] & HMM_PFN_ERROR);
drivers/infiniband/core/umem_odp.c
398
WARN_ON(fault && !(range.hmm_pfns[pfn_index] & HMM_PFN_VALID));
drivers/infiniband/core/uverbs_main.c
827
.fault = rdma_umap_fault,
drivers/infiniband/hw/hfi1/fault.c
135
bitmap_zero(fault->opcodes, sizeof(fault->opcodes) *
drivers/infiniband/hw/hfi1/fault.c
145
clear_bit(i, fault->opcodes);
drivers/infiniband/hw/hfi1/fault.c
147
set_bit(i, fault->opcodes);
drivers/infiniband/hw/hfi1/fault.c
166
struct fault *fault = file->private_data;
drivers/infiniband/hw/hfi1/fault.c
167
size_t bitsize = sizeof(fault->opcodes) * BITS_PER_BYTE;
drivers/infiniband/hw/hfi1/fault.c
172
bit = find_first_bit(fault->opcodes, bitsize);
drivers/infiniband/hw/hfi1/fault.c
174
zero = find_next_zero_bit(fault->opcodes, bitsize, bit);
drivers/infiniband/hw/hfi1/fault.c
183
bit = find_next_bit(fault->opcodes, bitsize, zero);
drivers/infiniband/hw/hfi1/fault.c
201
if (ibd->fault)
drivers/infiniband/hw/hfi1/fault.c
202
debugfs_remove_recursive(ibd->fault->dir);
drivers/infiniband/hw/hfi1/fault.c
203
kfree(ibd->fault);
drivers/infiniband/hw/hfi1/fault.c
204
ibd->fault = NULL;
drivers/infiniband/hw/hfi1/fault.c
212
ibd->fault = kzalloc_obj(*ibd->fault);
drivers/infiniband/hw/hfi1/fault.c
213
if (!ibd->fault)
drivers/infiniband/hw/hfi1/fault.c
216
ibd->fault->attr.interval = 1;
drivers/infiniband/hw/hfi1/fault.c
217
ibd->fault->attr.require_end = ULONG_MAX;
drivers/infiniband/hw/hfi1/fault.c
218
ibd->fault->attr.stacktrace_depth = 32;
drivers/infiniband/hw/hfi1/fault.c
219
ibd->fault->attr.dname = NULL;
drivers/infiniband/hw/hfi1/fault.c
220
ibd->fault->attr.verbose = 0;
drivers/infiniband/hw/hfi1/fault.c
221
ibd->fault->enable = false;
drivers/infiniband/hw/hfi1/fault.c
222
ibd->fault->opcode = false;
drivers/infiniband/hw/hfi1/fault.c
223
ibd->fault->fault_skip = 0;
drivers/infiniband/hw/hfi1/fault.c
224
ibd->fault->skip = 0;
drivers/infiniband/hw/hfi1/fault.c
225
ibd->fault->direction = HFI1_FAULT_DIR_TXRX;
drivers/infiniband/hw/hfi1/fault.c
226
ibd->fault->suppress_err = false;
drivers/infiniband/hw/hfi1/fault.c
227
bitmap_zero(ibd->fault->opcodes,
drivers/infiniband/hw/hfi1/fault.c
228
sizeof(ibd->fault->opcodes) * BITS_PER_BYTE);
drivers/infiniband/hw/hfi1/fault.c
231
fault_create_debugfs_attr("fault", parent, &ibd->fault->attr);
drivers/infiniband/hw/hfi1/fault.c
233
kfree(ibd->fault);
drivers/infiniband/hw/hfi1/fault.c
234
ibd->fault = NULL;
drivers/infiniband/hw/hfi1/fault.c
237
ibd->fault->dir = fault_dir;
drivers/infiniband/hw/hfi1/fault.c
241
debugfs_create_bool("enable", 0600, fault_dir, &ibd->fault->enable);
drivers/infiniband/hw/hfi1/fault.c
243
&ibd->fault->suppress_err);
drivers/infiniband/hw/hfi1/fault.c
245
&ibd->fault->opcode);
drivers/infiniband/hw/hfi1/fault.c
246
debugfs_create_file("opcodes", 0600, fault_dir, ibd->fault,
drivers/infiniband/hw/hfi1/fault.c
249
&ibd->fault->fault_skip);
drivers/infiniband/hw/hfi1/fault.c
251
&ibd->fault->fault_skip_usec);
drivers/infiniband/hw/hfi1/fault.c
252
debugfs_create_u8("direction", 0600, fault_dir, &ibd->fault->direction);
drivers/infiniband/hw/hfi1/fault.c
259
if (ibd->fault)
drivers/infiniband/hw/hfi1/fault.c
260
return ibd->fault->suppress_err;
drivers/infiniband/hw/hfi1/fault.c
269
if (!ibd->fault || !ibd->fault->enable)
drivers/infiniband/hw/hfi1/fault.c
271
if (!(ibd->fault->direction & direction))
drivers/infiniband/hw/hfi1/fault.c
273
if (ibd->fault->opcode) {
drivers/infiniband/hw/hfi1/fault.c
274
if (bitmap_empty(ibd->fault->opcodes,
drivers/infiniband/hw/hfi1/fault.c
275
(sizeof(ibd->fault->opcodes) *
drivers/infiniband/hw/hfi1/fault.c
278
if (!(test_bit(opcode, ibd->fault->opcodes)))
drivers/infiniband/hw/hfi1/fault.c
281
if (ibd->fault->fault_skip_usec &&
drivers/infiniband/hw/hfi1/fault.c
282
time_before(jiffies, ibd->fault->skip_usec))
drivers/infiniband/hw/hfi1/fault.c
284
if (ibd->fault->fault_skip && ibd->fault->skip) {
drivers/infiniband/hw/hfi1/fault.c
285
ibd->fault->skip--;
drivers/infiniband/hw/hfi1/fault.c
288
ret = should_fail(&ibd->fault->attr, 1);
drivers/infiniband/hw/hfi1/fault.c
290
ibd->fault->skip = ibd->fault->fault_skip;
drivers/infiniband/hw/hfi1/fault.c
291
ibd->fault->skip_usec = jiffies +
drivers/infiniband/hw/hfi1/fault.c
292
usecs_to_jiffies(ibd->fault->fault_skip_usec);
drivers/infiniband/hw/hfi1/fault.c
303
ibd->fault->n_txfaults[opcode]++;
drivers/infiniband/hw/hfi1/fault.c
315
ibd->fault->n_rxfaults[packet->opcode]++;
drivers/infiniband/hw/hfi1/fault.c
69
if (!ibd->fault->n_rxfaults[i] && !ibd->fault->n_txfaults[i])
drivers/infiniband/hw/hfi1/fault.c
74
(unsigned long long)ibd->fault->n_rxfaults[i],
drivers/infiniband/hw/hfi1/fault.c
75
(unsigned long long)ibd->fault->n_txfaults[i]);
drivers/infiniband/hw/hfi1/fault.c
96
struct fault *fault = file->private_data;
drivers/infiniband/hw/hfi1/file_ops.c
89
.fault = vma_fault,
drivers/infiniband/hw/hfi1/verbs.h
260
struct fault *fault;
drivers/infiniband/hw/mlx5/odp.c
700
bool fault = !(flags & MLX5_PF_FLAGS_SNAPSHOT);
drivers/infiniband/hw/mlx5/odp.c
715
np = ib_umem_odp_map_dma_and_lock(odp, user_va, bcnt, access_mask, fault);
drivers/infiniband/sw/rxe/rxe_odp.c
47
bool fault = !(flags & RXE_PAGEFAULT_SNAPSHOT);
drivers/infiniband/sw/rxe/rxe_odp.c
60
access_mask, fault);
drivers/input/touchscreen/atmel_mxt_ts.c
2511
goto fault;
drivers/input/touchscreen/atmel_mxt_ts.c
2527
goto fault;
drivers/input/touchscreen/atmel_mxt_ts.c
2533
fault:
drivers/iommu/amd/ppr.c
127
event.fault.type = IOMMU_FAULT_PAGE_REQ;
drivers/iommu/amd/ppr.c
128
event.fault.prm.perm = ppr_flag_to_fault_perm(PPR_FLAGS(raw[0]));
drivers/iommu/amd/ppr.c
129
event.fault.prm.addr = (u64)(raw[1] & PAGE_MASK);
drivers/iommu/amd/ppr.c
130
event.fault.prm.pasid = PPR_PASID(raw[0]);
drivers/iommu/amd/ppr.c
131
event.fault.prm.grpid = PPR_TAG(raw[0]) & 0x1FF;
drivers/iommu/amd/ppr.c
138
if (event.fault.prm.pasid == 0 ||
drivers/iommu/amd/ppr.c
139
event.fault.prm.pasid >= dev_data->max_pasids) {
drivers/iommu/amd/ppr.c
141
event.fault.prm.pasid, pdev->dev.id);
drivers/iommu/amd/ppr.c
145
event.fault.prm.flags |= IOMMU_FAULT_PAGE_RESPONSE_NEEDS_PASID;
drivers/iommu/amd/ppr.c
146
event.fault.prm.flags |= IOMMU_FAULT_PAGE_REQUEST_PASID_VALID;
drivers/iommu/amd/ppr.c
148
event.fault.prm.flags |= IOMMU_FAULT_PAGE_REQUEST_LAST_PAGE;
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
1883
struct iommu_fault *flt = &fault_evt.fault;
drivers/iommu/exynos-iommu.c
278
struct sysmmu_fault *fault);
drivers/iommu/exynos-iommu.c
315
struct sysmmu_fault *fault)
drivers/iommu/exynos-iommu.c
323
fault->addr = readl(data->sfrbase + finfo->addr_reg);
drivers/iommu/exynos-iommu.c
324
fault->name = finfo->name;
drivers/iommu/exynos-iommu.c
325
fault->type = finfo->type;
drivers/iommu/exynos-iommu.c
332
struct sysmmu_fault *fault)
drivers/iommu/exynos-iommu.c
337
fault->type = IOMMU_FAULT_READ;
drivers/iommu/exynos-iommu.c
340
fault->type = IOMMU_FAULT_WRITE;
drivers/iommu/exynos-iommu.c
347
fault->name = sysmmu_v5_fault_names[itype];
drivers/iommu/exynos-iommu.c
348
fault->addr = readl(data->sfrbase + addr_reg);
drivers/iommu/exynos-iommu.c
355
struct sysmmu_fault *fault)
drivers/iommu/exynos-iommu.c
359
fault->addr = readl(SYSMMU_REG(data, fault_va));
drivers/iommu/exynos-iommu.c
360
fault->name = sysmmu_v7_fault_names[itype % 4];
drivers/iommu/exynos-iommu.c
361
fault->type = (info & BIT(20)) ? IOMMU_FAULT_WRITE : IOMMU_FAULT_READ;
drivers/iommu/exynos-iommu.c
550
const struct sysmmu_fault *fault)
drivers/iommu/exynos-iommu.c
556
fault->type == IOMMU_FAULT_READ ? "READ" : "WRITE",
drivers/iommu/exynos-iommu.c
557
fault->name, fault->addr);
drivers/iommu/exynos-iommu.c
559
ent = section_entry(phys_to_virt(data->pgtable), fault->addr);
drivers/iommu/exynos-iommu.c
562
ent = page_entry(ent, fault->addr);
drivers/iommu/exynos-iommu.c
571
struct sysmmu_fault fault;
drivers/iommu/exynos-iommu.c
580
ret = data->variant->get_fault_info(data, itype, &fault);
drivers/iommu/exynos-iommu.c
585
show_fault_information(data, &fault);
drivers/iommu/exynos-iommu.c
589
fault.addr, fault.type);
drivers/iommu/intel/dmar.c
1240
static void qi_dump_fault(struct intel_iommu *iommu, u32 fault)
drivers/iommu/intel/dmar.c
1246
if (fault & DMA_FSTS_IQE)
drivers/iommu/intel/dmar.c
1249
if (fault & DMA_FSTS_ITE)
drivers/iommu/intel/dmar.c
1252
if (fault & DMA_FSTS_ICE)
drivers/iommu/intel/dmar.c
1273
u32 fault;
drivers/iommu/intel/dmar.c
1283
fault = readl(iommu->reg + DMAR_FSTS_REG);
drivers/iommu/intel/dmar.c
1284
if (fault & (DMA_FSTS_IQE | DMA_FSTS_ITE | DMA_FSTS_ICE))
drivers/iommu/intel/dmar.c
1285
qi_dump_fault(iommu, fault);
drivers/iommu/intel/dmar.c
1292
if (fault & DMA_FSTS_IQE) {
drivers/iommu/intel/dmar.c
1314
if (fault & DMA_FSTS_ITE) {
drivers/iommu/intel/dmar.c
1353
if (fault & DMA_FSTS_ICE) {
drivers/iommu/intel/prq.c
181
event.fault.type = IOMMU_FAULT_PAGE_REQ;
drivers/iommu/intel/prq.c
182
event.fault.prm.addr = (u64)desc->addr << VTD_PAGE_SHIFT;
drivers/iommu/intel/prq.c
183
event.fault.prm.pasid = desc->pasid;
drivers/iommu/intel/prq.c
184
event.fault.prm.grpid = desc->prg_index;
drivers/iommu/intel/prq.c
185
event.fault.prm.perm = prq_to_iommu_prot(desc);
drivers/iommu/intel/prq.c
188
event.fault.prm.flags |= IOMMU_FAULT_PAGE_REQUEST_LAST_PAGE;
drivers/iommu/intel/prq.c
190
event.fault.prm.flags |= IOMMU_FAULT_PAGE_REQUEST_PASID_VALID;
drivers/iommu/intel/prq.c
191
event.fault.prm.flags |= IOMMU_FAULT_PAGE_RESPONSE_NEEDS_PASID;
drivers/iommu/intel/prq.c
383
prm = &evt->fault.prm;
drivers/iommu/io-pgfault.c
106
if (iopf->fault.prm.grpid == evt->fault.prm.grpid)
drivers/iommu/io-pgfault.c
121
struct iommu_fault *fault = &evt->fault;
drivers/iommu/io-pgfault.c
124
if (fault->prm.flags & IOMMU_FAULT_PAGE_REQUEST_PASID_VALID) {
drivers/iommu/io-pgfault.c
126
fault->prm.pasid, 0);
drivers/iommu/io-pgfault.c
161
struct iommu_fault *fault = &evt->fault;
drivers/iommu/io-pgfault.c
163
.pasid = fault->prm.pasid,
drivers/iommu/io-pgfault.c
164
.grpid = fault->prm.grpid,
drivers/iommu/io-pgfault.c
217
struct iommu_fault *fault = &evt->fault;
drivers/iommu/io-pgfault.c
234
if (!(fault->prm.flags & IOMMU_FAULT_PAGE_REQUEST_LAST_PAGE)) {
drivers/iommu/io-pgfault.c
237
ret = report_partial_fault(iopf_param, fault);
drivers/iommu/io-pgfault.c
269
fault->prm.pasid);
drivers/iommu/io-pgfault.c
279
if (fault->type == IOMMU_FAULT_PAGE_REQ)
drivers/iommu/io-pgfault.c
330
.pasid = iopf->fault.prm.pasid,
drivers/iommu/io-pgfault.c
331
.grpid = iopf->fault.prm.grpid,
drivers/iommu/io-pgfault.c
47
if (!(iopf->fault.prm.flags & IOMMU_FAULT_PAGE_REQUEST_LAST_PAGE))
drivers/iommu/io-pgfault.c
474
.pasid = iopf->fault.prm.pasid,
drivers/iommu/io-pgfault.c
475
.grpid = iopf->fault.prm.grpid,
drivers/iommu/io-pgfault.c
64
struct iommu_fault *fault)
drivers/iommu/io-pgfault.c
72
iopf->fault = *fault;
drivers/iommu/io-pgfault.c
98
group->last_fault.fault = evt->fault;
drivers/iommu/iommu-sva.c
222
iommu_sva_handle_mm(struct iommu_fault *fault, struct mm_struct *mm)
drivers/iommu/iommu-sva.c
228
struct iommu_fault_page_request *prm = &fault->prm;
drivers/iommu/iommu-sva.c
290
status = iommu_sva_handle_mm(&iopf->fault,
drivers/iommu/iommufd/device.c
483
if (!hwpt->fault || !dev_is_pci(idev->dev))
drivers/iommu/iommufd/eventq.c
101
iommufd_fault_deliver_fetch(struct iommufd_fault *fault)
drivers/iommu/iommufd/eventq.c
103
struct list_head *list = &fault->common.deliver;
drivers/iommu/iommufd/eventq.c
106
spin_lock(&fault->common.lock);
drivers/iommu/iommufd/eventq.c
111
spin_unlock(&fault->common.lock);
drivers/iommu/iommufd/eventq.c
116
static void iommufd_fault_deliver_restore(struct iommufd_fault *fault,
drivers/iommu/iommufd/eventq.c
119
spin_lock(&fault->common.lock);
drivers/iommu/iommufd/eventq.c
120
list_add(&group->node, &fault->common.deliver);
drivers/iommu/iommufd/eventq.c
121
spin_unlock(&fault->common.lock);
drivers/iommu/iommufd/eventq.c
129
struct iommufd_fault *fault = eventq_to_fault(eventq);
drivers/iommu/iommufd/eventq.c
140
mutex_lock(&fault->mutex);
drivers/iommu/iommufd/eventq.c
141
while ((group = iommufd_fault_deliver_fetch(fault))) {
drivers/iommu/iommufd/eventq.c
144
iommufd_fault_deliver_restore(fault, group);
drivers/iommu/iommufd/eventq.c
148
rc = xa_alloc(&fault->response, &group->cookie, group,
drivers/iommu/iommufd/eventq.c
151
iommufd_fault_deliver_restore(fault, group);
drivers/iommu/iommufd/eventq.c
157
iommufd_compose_fault_message(&iopf->fault,
drivers/iommu/iommufd/eventq.c
161
xa_erase(&fault->response, group->cookie);
drivers/iommu/iommufd/eventq.c
162
iommufd_fault_deliver_restore(fault, group);
drivers/iommu/iommufd/eventq.c
169
mutex_unlock(&fault->mutex);
drivers/iommu/iommufd/eventq.c
179
struct iommufd_fault *fault = eventq_to_fault(eventq);
drivers/iommu/iommufd/eventq.c
188
mutex_lock(&fault->mutex);
drivers/iommu/iommufd/eventq.c
204
group = xa_erase(&fault->response, response.cookie);
drivers/iommu/iommufd/eventq.c
214
mutex_unlock(&fault->mutex);
drivers/iommu/iommufd/eventq.c
22
struct iommufd_fault *fault = hwpt->fault;
drivers/iommu/iommufd/eventq.c
27
if (!fault || !handle)
drivers/iommu/iommufd/eventq.c
31
mutex_lock(&fault->mutex);
drivers/iommu/iommufd/eventq.c
32
spin_lock(&fault->common.lock);
drivers/iommu/iommufd/eventq.c
33
list_for_each_entry_safe(group, next, &fault->common.deliver, node) {
drivers/iommu/iommufd/eventq.c
38
spin_unlock(&fault->common.lock);
drivers/iommu/iommufd/eventq.c
420
struct iommufd_fault *fault;
drivers/iommu/iommufd/eventq.c
427
fault = __iommufd_object_alloc_ucmd(ucmd, fault, IOMMUFD_OBJ_FAULT,
drivers/iommu/iommufd/eventq.c
429
if (IS_ERR(fault))
drivers/iommu/iommufd/eventq.c
430
return PTR_ERR(fault);
drivers/iommu/iommufd/eventq.c
432
xa_init_flags(&fault->response, XA_FLAGS_ALLOC1);
drivers/iommu/iommufd/eventq.c
433
mutex_init(&fault->mutex);
drivers/iommu/iommufd/eventq.c
435
fdno = iommufd_eventq_init(&fault->common, "[iommufd-pgfault]",
drivers/iommu/iommufd/eventq.c
440
cmd->out_fault_id = fault->common.obj.id;
drivers/iommu/iommufd/eventq.c
447
fd_install(fdno, fault->common.filep);
drivers/iommu/iommufd/eventq.c
458
struct iommufd_fault *fault;
drivers/iommu/iommufd/eventq.c
46
xa_for_each(&fault->response, index, group) {
drivers/iommu/iommufd/eventq.c
461
fault = hwpt->fault;
drivers/iommu/iommufd/eventq.c
463
spin_lock(&fault->common.lock);
drivers/iommu/iommufd/eventq.c
464
list_add_tail(&group->node, &fault->common.deliver);
drivers/iommu/iommufd/eventq.c
465
spin_unlock(&fault->common.lock);
drivers/iommu/iommufd/eventq.c
467
wake_up_interruptible(&fault->common.wait_queue);
drivers/iommu/iommufd/eventq.c
49
xa_erase(&fault->response, index);
drivers/iommu/iommufd/eventq.c
53
mutex_unlock(&fault->mutex);
drivers/iommu/iommufd/eventq.c
60
struct iommufd_fault *fault = eventq_to_fault(eventq);
drivers/iommu/iommufd/eventq.c
70
list_for_each_entry_safe(group, next, &fault->common.deliver, node) {
drivers/iommu/iommufd/eventq.c
75
xa_for_each(&fault->response, index, group) {
drivers/iommu/iommufd/eventq.c
76
xa_erase(&fault->response, index);
drivers/iommu/iommufd/eventq.c
80
xa_destroy(&fault->response);
drivers/iommu/iommufd/eventq.c
81
mutex_destroy(&fault->mutex);
drivers/iommu/iommufd/eventq.c
84
static void iommufd_compose_fault_message(struct iommu_fault *fault,
drivers/iommu/iommufd/eventq.c
89
hwpt_fault->flags = fault->prm.flags;
drivers/iommu/iommufd/eventq.c
91
hwpt_fault->pasid = fault->prm.pasid;
drivers/iommu/iommufd/eventq.c
92
hwpt_fault->grpid = fault->prm.grpid;
drivers/iommu/iommufd/eventq.c
93
hwpt_fault->perm = fault->prm.perm;
drivers/iommu/iommufd/eventq.c
94
hwpt_fault->addr = fault->prm.addr;
drivers/iommu/iommufd/hw_pagetable.c
16
if (hwpt->fault)
drivers/iommu/iommufd/hw_pagetable.c
17
refcount_dec(&hwpt->fault->common.obj.users);
drivers/iommu/iommufd/hw_pagetable.c
412
struct iommufd_fault *fault;
drivers/iommu/iommufd/hw_pagetable.c
414
fault = iommufd_get_fault(ucmd, cmd->fault_id);
drivers/iommu/iommufd/hw_pagetable.c
415
if (IS_ERR(fault)) {
drivers/iommu/iommufd/hw_pagetable.c
416
rc = PTR_ERR(fault);
drivers/iommu/iommufd/hw_pagetable.c
419
hwpt->fault = fault;
drivers/iommu/iommufd/hw_pagetable.c
421
refcount_inc(&fault->common.obj.users);
drivers/iommu/iommufd/hw_pagetable.c
422
iommufd_put_object(ucmd->ictx, &fault->common.obj);
drivers/iommu/iommufd/iommufd_private.h
366
struct iommufd_fault *fault;
drivers/iommu/iommufd/main.c
417
struct iommu_fault_alloc fault;
drivers/iommu/iommufd/selftest.c
1778
event.fault.prm.flags = IOMMU_FAULT_PAGE_REQUEST_LAST_PAGE;
drivers/iommu/iommufd/selftest.c
1780
event.fault.prm.flags |= IOMMU_FAULT_PAGE_REQUEST_PASID_VALID;
drivers/iommu/iommufd/selftest.c
1781
event.fault.type = IOMMU_FAULT_PAGE_REQ;
drivers/iommu/iommufd/selftest.c
1782
event.fault.prm.addr = cmd->trigger_iopf.addr;
drivers/iommu/iommufd/selftest.c
1783
event.fault.prm.pasid = cmd->trigger_iopf.pasid;
drivers/iommu/iommufd/selftest.c
1784
event.fault.prm.grpid = cmd->trigger_iopf.grpid;
drivers/iommu/iommufd/selftest.c
1785
event.fault.prm.perm = cmd->trigger_iopf.perm;
drivers/iommu/virtio-iommu.c
594
struct virtio_iommu_fault *fault)
drivers/iommu/virtio-iommu.c
598
u8 reason = fault->reason;
drivers/iommu/virtio-iommu.c
599
u32 flags = le32_to_cpu(fault->flags);
drivers/iommu/virtio-iommu.c
600
u32 endpoint = le32_to_cpu(fault->endpoint);
drivers/iommu/virtio-iommu.c
601
u64 address = le64_to_cpu(fault->address);
drivers/iommu/virtio-iommu.c
643
viommu_fault_handler(viommu, &evt->fault);
drivers/iommu/virtio-iommu.c
95
struct virtio_iommu_fault fault;
drivers/leds/flash/leds-as3645a.c
251
static int as3645a_get_fault(struct led_classdev_flash *fled, u32 *fault)
drivers/leds/flash/leds-as3645a.c
262
*fault |= LED_FAULT_OVER_CURRENT;
drivers/leds/flash/leds-as3645a.c
265
*fault |= LED_FAULT_INDICATOR;
drivers/leds/flash/leds-as3645a.c
271
*fault |= LED_FAULT_TIMEOUT;
drivers/leds/flash/leds-as3645a.c
274
*fault |= LED_FAULT_OVER_TEMPERATURE;
drivers/leds/flash/leds-as3645a.c
277
*fault |= LED_FAULT_OVER_CURRENT;
drivers/leds/flash/leds-as3645a.c
280
*fault |= LED_FAULT_INPUT_VOLTAGE;
drivers/leds/flash/leds-as3645a.c
383
u32 fault = 0;
drivers/leds/flash/leds-as3645a.c
406
rval = as3645a_get_fault(&flash->fled, &fault);
drivers/leds/flash/leds-lm3601x.c
302
u32 *fault)
drivers/leds/flash/leds-lm3601x.c
308
*fault = led->last_flag;
drivers/leds/flash/leds-max77693.c
551
u32 *fault)
drivers/leds/flash/leds-max77693.c
555
*fault = sub_led->flash_faults;
drivers/leds/flash/leds-mt6360.c
340
static int mt6360_fault_get(struct led_classdev_flash *fl_cdev, u32 *fault)
drivers/leds/flash/leds-mt6360.c
380
*fault = rfault;
drivers/leds/flash/leds-mt6370-flash.c
282
static int mt6370_fault_get(struct led_classdev_flash *fl_cdev, u32 *fault)
drivers/leds/flash/leds-mt6370-flash.c
330
*fault = rfault;
drivers/leds/flash/leds-qcom-flash.c
508
static int qcom_flash_fault_get(struct led_classdev_flash *fled_cdev, u32 *fault)
drivers/leds/flash/leds-qcom-flash.c
580
*fault = fault_sts;
drivers/leds/flash/leds-rt4505.c
187
static int rt4505_fault_get(struct led_classdev_flash *fled_cdev, u32 *fault)
drivers/leds/flash/leds-rt4505.c
210
*fault = led_faults;
drivers/leds/flash/leds-sy7802.c
260
static int sy7802_fault_get(struct led_classdev_flash *fl_cdev, u32 *fault)
drivers/leds/flash/leds-sy7802.c
281
*fault = led_faults;
drivers/leds/flash/leds-tps6131x.c
479
static int tps6131x_flash_fault_get(struct led_classdev_flash *fled_cdev, u32 *fault)
drivers/leds/flash/leds-tps6131x.c
485
*fault = 0;
drivers/leds/flash/leds-tps6131x.c
500
*fault |= LED_FAULT_SHORT_CIRCUIT;
drivers/leds/flash/leds-tps6131x.c
503
*fault |= LED_FAULT_TIMEOUT;
drivers/leds/flash/leds-tps6131x.c
506
*fault |= LED_FAULT_OVER_TEMPERATURE;
drivers/leds/flash/leds-tps6131x.c
509
*fault |= LED_FAULT_LED_OVER_TEMPERATURE;
drivers/leds/flash/leds-tps6131x.c
512
*fault |= LED_FAULT_UNDER_VOLTAGE;
drivers/leds/led-class-flash.c
190
u32 fault, mask = 0x1;
drivers/leds/led-class-flash.c
194
ret = led_get_flash_fault(fled_cdev, &fault);
drivers/leds/led-class-flash.c
201
if (fault & mask) {
drivers/leds/led-class-flash.c
403
int led_get_flash_fault(struct led_classdev_flash *fled_cdev, u32 *fault)
drivers/leds/led-class-flash.c
405
return call_flash_op(fled_cdev, fault_get, fault);
drivers/leds/leds-lp8860.c
158
int ret, fault;
drivers/leds/leds-lp8860.c
165
fault = read_buf;
drivers/leds/leds-lp8860.c
171
fault |= read_buf;
drivers/leds/leds-lp8860.c
174
if (fault)
drivers/md/md-autodetect.c
62
int minor, level, factor, fault, partitioned = 0;
drivers/md/md-autodetect.c
93
get_option(&str, &fault) != 2) {
drivers/media/i2c/adp1653.c
103
return flash->fault;
drivers/media/i2c/adp1653.c
152
if (flash->fault & ADP1653_REG_FAULT_FLT_SCP)
drivers/media/i2c/adp1653.c
154
if (flash->fault & ADP1653_REG_FAULT_FLT_OT)
drivers/media/i2c/adp1653.c
156
if (flash->fault & ADP1653_REG_FAULT_FLT_TMR)
drivers/media/i2c/adp1653.c
158
if (flash->fault & ADP1653_REG_FAULT_FLT_OV)
drivers/media/i2c/adp1653.c
161
flash->fault = 0;
drivers/media/i2c/adp1653.c
200
struct v4l2_ctrl *fault;
drivers/media/i2c/adp1653.c
241
fault = v4l2_ctrl_new_std(&flash->ctrls, &adp1653_ctrl_ops,
drivers/media/i2c/adp1653.c
250
fault->flags |= V4L2_CTRL_FLAG_VOLATILE;
drivers/media/i2c/adp1653.c
275
flash->fault = 0;
drivers/media/i2c/adp1653.c
80
int fault;
drivers/media/i2c/adp1653.c
83
fault = i2c_smbus_read_byte_data(client, ADP1653_REG_FAULT);
drivers/media/i2c/adp1653.c
84
if (fault < 0)
drivers/media/i2c/adp1653.c
85
return fault;
drivers/media/i2c/adp1653.c
87
flash->fault |= fault;
drivers/media/i2c/adp1653.c
89
if (!flash->fault)
drivers/media/i2c/lm3560.c
168
s32 fault = 0;
drivers/media/i2c/lm3560.c
174
fault |= V4L2_FLASH_FAULT_SHORT_CIRCUIT;
drivers/media/i2c/lm3560.c
176
fault |= V4L2_FLASH_FAULT_OVER_TEMPERATURE;
drivers/media/i2c/lm3560.c
178
fault |= V4L2_FLASH_FAULT_TIMEOUT;
drivers/media/i2c/lm3560.c
179
ctrl->cur.val = fault;
drivers/media/i2c/lm3560.c
281
struct v4l2_ctrl *fault;
drivers/media/i2c/lm3560.c
323
fault = v4l2_ctrl_new_std(hdl, ops, V4L2_CID_FLASH_FAULT, 0,
drivers/media/i2c/lm3560.c
328
if (fault != NULL)
drivers/media/i2c/lm3560.c
329
fault->flags |= V4L2_CTRL_FLAG_VOLATILE;
drivers/media/i2c/lm3646.c
205
struct v4l2_ctrl *fault;
drivers/media/i2c/lm3646.c
245
fault = v4l2_ctrl_new_std(hdl, ops, V4L2_CID_FLASH_FAULT, 0,
drivers/media/i2c/lm3646.c
250
if (fault != NULL)
drivers/media/i2c/lm3646.c
251
fault->flags |= V4L2_CTRL_FLAG_VOLATILE;
drivers/misc/enclosure.c
484
return sysfs_emit(buf, "%d\n", ecomp->fault);
drivers/misc/enclosure.c
644
static DEVICE_ATTR(fault, S_IRUGO | S_IWUSR, get_component_fault,
drivers/misc/ocxl/context.c
158
.fault = ocxl_mmap_fault,
drivers/misc/ocxl/link.c
142
struct xsl_fault *fault = container_of(fault_work, struct xsl_fault,
drivers/misc/ocxl/link.c
144
struct spa *spa = container_of(fault, struct spa, xsl_fault);
drivers/misc/ocxl/link.c
152
rc = copro_handle_mm_fault(fault->pe_data.mm, fault->dar, fault->dsisr,
drivers/misc/ocxl/link.c
156
if (fault->pe_data.xsl_err_cb) {
drivers/misc/ocxl/link.c
157
fault->pe_data.xsl_err_cb(
drivers/misc/ocxl/link.c
158
fault->pe_data.xsl_err_data,
drivers/misc/ocxl/link.c
159
fault->dar, fault->dsisr);
drivers/misc/ocxl/link.c
172
if (fault->dsisr & SPA_XSL_S)
drivers/misc/ocxl/link.c
175
if (get_region_id(fault->dar) != USER_REGION_ID)
drivers/misc/ocxl/link.c
179
hash_page_mm(fault->pe_data.mm, fault->dar, access, 0x300,
drivers/misc/ocxl/link.c
185
mmput(fault->pe_data.mm);
drivers/misc/ocxl/sysfs.c
124
.fault = global_mmio_fault,
drivers/misc/sgi-gru/grufile.c
523
.fault = gru_fault,
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_utils.h
131
u8 fault;
drivers/net/ethernet/intel/fm10k/fm10k_pci.c
1218
struct fm10k_fault *fault)
drivers/net/ethernet/intel/fm10k/fm10k_pci.c
1227
switch (fault->type) {
drivers/net/ethernet/intel/fm10k/fm10k_pci.c
1241
switch (fault->type) {
drivers/net/ethernet/intel/fm10k/fm10k_pci.c
1250
switch (fault->type) {
drivers/net/ethernet/intel/fm10k/fm10k_pci.c
1274
error, fault->address, fault->specinfo,
drivers/net/ethernet/intel/fm10k/fm10k_pci.c
1275
PCI_SLOT(fault->func), PCI_FUNC(fault->func));
drivers/net/ethernet/intel/fm10k/fm10k_pci.c
1286
if (fault->func && iov_data) {
drivers/net/ethernet/intel/fm10k/fm10k_pci.c
1287
int vf = fault->func - 1;
drivers/net/ethernet/intel/fm10k/fm10k_pci.c
1305
struct fm10k_fault fault = { 0 };
drivers/net/ethernet/intel/fm10k/fm10k_pci.c
1316
err = hw->mac.ops.get_fault(hw, type, &fault);
drivers/net/ethernet/intel/fm10k/fm10k_pci.c
1323
fm10k_handle_fault(interface, type, &fault);
drivers/net/ethernet/intel/fm10k/fm10k_pf.c
1426
struct fm10k_fault *fault)
drivers/net/ethernet/intel/fm10k/fm10k_pf.c
1446
fault->address = fm10k_read_reg(hw, type + FM10K_FAULT_ADDR_HI);
drivers/net/ethernet/intel/fm10k/fm10k_pf.c
1447
fault->address <<= 32;
drivers/net/ethernet/intel/fm10k/fm10k_pf.c
1448
fault->address |= fm10k_read_reg(hw, type + FM10K_FAULT_ADDR_LO);
drivers/net/ethernet/intel/fm10k/fm10k_pf.c
1449
fault->specinfo = fm10k_read_reg(hw, type + FM10K_FAULT_SPECINFO);
drivers/net/ethernet/intel/fm10k/fm10k_pf.c
1456
fault->func = 0;
drivers/net/ethernet/intel/fm10k/fm10k_pf.c
1458
fault->func = 1 + FIELD_GET(FM10K_FAULT_FUNC_VF_MASK, func);
drivers/net/ethernet/intel/fm10k/fm10k_pf.c
1461
fault->type = func & FM10K_FAULT_FUNC_TYPE_MASK;
drivers/net/ethernet/mellanox/mlxsw/core_hwmon.c
181
bool fault;
drivers/net/ethernet/mellanox/mlxsw/core_hwmon.c
189
mlxsw_reg_fore_unpack(fore_pl, mlxsw_hwmon_attr->type_index, &fault);
drivers/net/ethernet/mellanox/mlxsw/core_hwmon.c
191
return sprintf(buf, "%u\n", fault);
drivers/net/ethernet/mellanox/mlxsw/core_hwmon.c
290
u8 module, fault;
drivers/net/ethernet/mellanox/mlxsw/core_hwmon.c
311
fault = 1;
drivers/net/ethernet/mellanox/mlxsw/core_hwmon.c
317
fault = 0;
drivers/net/ethernet/mellanox/mlxsw/core_hwmon.c
321
return sprintf(buf, "%u\n", fault);
drivers/net/ethernet/mellanox/mlxsw/reg.h
9432
bool *fault)
drivers/net/ethernet/mellanox/mlxsw/reg.h
9436
if (fault) {
drivers/net/ethernet/mellanox/mlxsw/reg.h
9438
*fault = limit & BIT(tacho);
drivers/net/phy/microchip_t1.c
922
bool fault = false, check_a = false, check_b = false;
drivers/net/phy/microchip_t1.c
968
fault = true;
drivers/net/phy/microchip_t1.c
971
if (!fault)
drivers/platform/chrome/cros_usbpd_logger.c
134
fault = r->data < ARRAY_SIZE(fault_names) ? fault_names[r->data]
drivers/platform/chrome/cros_usbpd_logger.c
136
len += append_str(buf, len, "Power supply fault: %s", fault);
drivers/platform/chrome/cros_usbpd_logger.c
85
const char *fault, *role, *chg_type;
drivers/power/supply/bq24257_charger.c
335
switch (state.fault) {
drivers/power/supply/bq24257_charger.c
431
state->fault = ret;
drivers/power/supply/bq24257_charger.c
444
switch (state->fault) {
drivers/power/supply/bq24257_charger.c
464
bq->state.fault != new_state->fault ||
drivers/power/supply/bq24257_charger.c
632
} else if (new_state->fault == FAULT_NO_BAT) {
drivers/power/supply/bq24257_charger.c
634
} else if (new_state->fault == FAULT_TIMER) {
drivers/power/supply/bq24257_charger.c
658
state.status, state.fault, state.power_good);
drivers/power/supply/bq24257_charger.c
724
else if (state.fault != FAULT_NO_BAT)
drivers/power/supply/bq24257_charger.c
73
u8 fault;
drivers/power/supply/da9030_battery.c
108
uint8_t fault;
drivers/power/supply/da9030_battery.c
216
da903x_read(charger->master, DA9030_FAULT_LOG, &charger->fault);
drivers/power/supply/da9030_battery.c
321
if (charger->fault & DA9030_FAULT_LOG_OVER_TEMP)
drivers/power/supply/da9030_battery.c
323
else if (charger->fault & DA9030_FAULT_LOG_VBAT_OVER)
drivers/power/supply/max8903_charger.c
200
bool fault;
drivers/power/supply/max8903_charger.c
209
fault = gpiod_get_value(data->flt);
drivers/power/supply/max8903_charger.c
211
if (fault == data->fault)
drivers/power/supply/max8903_charger.c
214
data->fault = fault;
drivers/power/supply/max8903_charger.c
216
if (fault)
drivers/power/supply/max8903_charger.c
319
data->fault = false;
drivers/power/supply/max8903_charger.c
34
bool fault;
drivers/power/supply/max8903_charger.c
71
if (data->fault)
drivers/power/supply/mp2629_charger.c
258
if (!charger->fault)
drivers/power/supply/mp2629_charger.c
260
if (MP2629_FAULT_BATTERY & charger->fault)
drivers/power/supply/mp2629_charger.c
262
else if (MP2629_FAULT_THERMAL & charger->fault)
drivers/power/supply/mp2629_charger.c
264
else if (MP2629_FAULT_INPUT & charger->fault)
drivers/power/supply/mp2629_charger.c
451
charger->fault = rval;
drivers/power/supply/mp2629_charger.c
78
int fault;
drivers/regulator/wm8350-regulator.c
572
u16 stop, u16 fault)
drivers/regulator/wm8350-regulator.c
613
(fault << WM8350_DC1_ERRACT_SHIFT));
drivers/scsi/libiscsi.c
1772
goto fault;
drivers/scsi/libiscsi.c
1814
goto fault;
drivers/scsi/libiscsi.c
1821
goto fault;
drivers/scsi/libiscsi.c
1827
goto fault;
drivers/scsi/libiscsi.c
1880
fault:
drivers/scsi/lpfc/lpfc.h
435
uint8_t fault;
drivers/scsi/lpfc/lpfc_attr.c
1264
trunk_errmsg[link.link0.fault]);
drivers/scsi/lpfc/lpfc_attr.c
1271
trunk_errmsg[link.link1.fault]);
drivers/scsi/lpfc/lpfc_attr.c
1278
trunk_errmsg[link.link2.fault]);
drivers/scsi/lpfc/lpfc_attr.c
1285
trunk_errmsg[link.link3.fault]);
drivers/scsi/lpfc/lpfc_init.c
5469
phba->sli4_hba.link_state.fault =
drivers/scsi/lpfc/lpfc_init.c
5484
phba->sli4_hba.link_state.fault);
drivers/scsi/lpfc/lpfc_init.c
6154
phba->trunk_link.link0.fault = port_fault & 0x1 ? err : 0;
drivers/scsi/lpfc/lpfc_init.c
6161
phba->trunk_link.link1.fault = port_fault & 0x2 ? err : 0;
drivers/scsi/lpfc/lpfc_init.c
6168
phba->trunk_link.link2.fault = port_fault & 0x4 ? err : 0;
drivers/scsi/lpfc/lpfc_init.c
6175
phba->trunk_link.link3.fault = port_fault & 0x8 ? err : 0;
drivers/scsi/lpfc/lpfc_init.c
6257
phba->sli4_hba.link_state.fault =
drivers/scsi/lpfc/lpfc_init.c
6286
phba->sli4_hba.link_state.fault,
drivers/scsi/lpfc/lpfc_sli4.h
291
uint8_t fault;
drivers/scsi/mpi3mr/mpi/mpi30_transport.h
54
__le32 fault;
drivers/scsi/mpi3mr/mpi3mr.h
908
u16 fault;
drivers/scsi/mpi3mr/mpi3mr_fw.c
1062
u32 fault;
drivers/scsi/mpi3mr/mpi3mr_fw.c
1064
fault = (readl(&mrioc->sysif_regs->fault) &
drivers/scsi/mpi3mr/mpi3mr_fw.c
1067
switch (fault) {
drivers/scsi/mpi3mr/mpi3mr_fw.c
1100
code = readl(&mrioc->sysif_regs->fault);
drivers/scsi/mpi3mr/mpi3mr_fw.c
1127
mrioc->saved_fault_code = readl(&mrioc->sysif_regs->fault) &
drivers/scsi/mpi3mr/mpi3mr_fw.c
1699
u32 fault;
drivers/scsi/mpi3mr/mpi3mr_fw.c
1703
fault = readl(&mrioc->sysif_regs->fault) & MPI3_SYSIF_FAULT_CODE_MASK;
drivers/scsi/mpi3mr/mpi3mr_fw.c
1704
if (fault == MPI3_SYSIF_FAULT_CODE_DIAG_FAULT_RESET) {
drivers/scsi/mpi3mr/mpi3mr_fw.c
2666
trigger_data.fault = (readl(&mrioc->sysif_regs->fault) &
drivers/scsi/mpi3mr/mpi3mr_fw.c
2681
trigger_data.fault = (readl(&mrioc->sysif_regs->fault) &
drivers/scsi/mpi3mr/mpi3mr_fw.c
2928
trigger_data.fault = readl(&mrioc->sysif_regs->fault) & MPI3_SYSIF_FAULT_CODE_MASK;
drivers/scsi/mpi3mr/mpi3mr_fw.c
2953
switch (trigger_data.fault) {
drivers/scsi/mpi3mr/mpi3mr_fw.c
5582
trigger_data.fault = (readl(&mrioc->sysif_regs->fault) &
drivers/scsi/ses.c
244
ecomp->fault = 0;
drivers/scsi/ses.c
249
ecomp->fault = (desc[3] & 0x60) >> 4;
drivers/scsi/sg.c
1253
.fault = sg_vma_fault,
drivers/spi/spi-tle62x0.c
102
unsigned long fault = 0;
drivers/spi/spi-tle62x0.c
115
fault <<= 8;
drivers/spi/spi-tle62x0.c
116
fault |= ((unsigned long)buff[ptr]);
drivers/spi/spi-tle62x0.c
122
bp += sprintf(bp, "%s ", decode_fault(fault >> (ptr * 2)));
drivers/staging/greybus/light.c
686
u32 *fault)
drivers/staging/greybus/light.c
709
*fault = le32_to_cpu(resp.fault);
drivers/target/target_core_user.c
1924
.fault = tcmu_vma_fault,
drivers/uio/uio.c
726
.fault = uio_vma_fault,
drivers/usb/mon/mon_bin.c
1273
.fault = mon_bin_vma_fault,
drivers/vdpa/vdpa_user/iova_domain.c
569
.fault = vduse_domain_mmap_fault,
drivers/vfio/pci/nvgrace-gpu/main.c
339
.fault = nvgrace_gpu_vfio_pci_fault,
drivers/vfio/pci/vfio_pci_core.c
1722
.fault = vfio_pci_mmap_page_fault,
drivers/vhost/vdpa.c
1500
.fault = vhost_vdpa_fault,
drivers/video/backlight/sky81452-backlight.c
167
static DEVICE_ATTR(fault, S_IRUGO, sky81452_bl_show_fault, NULL);
drivers/video/fbdev/core/fb_defio.c
239
.fault = fb_deferred_io_fault,
drivers/xen/privcmd-buf.c
130
.fault = privcmd_buf_vma_fault,
drivers/xen/privcmd.c
1634
.fault = privcmd_fault
fs/9p/vfs_file.c
504
.fault = filemap_fault,
fs/afs/file.c
66
.fault = filemap_fault,
fs/aio.c
389
.fault = filemap_fault,
fs/btrfs/file.c
2036
.fault = filemap_fault,
fs/ceph/addr.c
2344
.fault = ceph_filemap_fault,
fs/erofs/data.c
467
.fault = erofs_dax_fault,
fs/exfat/file.c
736
.fault = filemap_fault,
fs/ext2/file.c
116
.fault = ext2_dax_fault,
fs/ext4/file.c
788
.fault = ext4_dax_fault,
fs/ext4/file.c
798
.fault = filemap_fault,
fs/f2fs/file.c
194
.fault = f2fs_filemap_fault,
fs/fuse/dax.c
817
.fault = fuse_dax_fault,
fs/fuse/file.c
2363
.fault = filemap_fault,
fs/gfs2/file.c
574
.fault = gfs2_fault,
fs/kernfs/file.c
400
if (of->vm_ops->fault)
fs/kernfs/file.c
401
ret = of->vm_ops->fault(vmf);
fs/kernfs/file.c
452
.fault = kernfs_vma_fault,
fs/nfs/file.c
692
.fault = filemap_fault,
fs/nilfs2/file.c
124
.fault = filemap_fault,
fs/ntfs3/file.c
266
.fault = filemap_fault,
fs/ocfs2/mmap.c
158
.fault = ocfs2_fault,
fs/orangefs/file.c
393
.fault = orangefs_fault,
fs/proc/vmcore.c
507
.fault = mmap_vmcore_fault,
fs/smb/client/file.c
3064
.fault = filemap_fault,
fs/ubifs/file.c
1579
.fault = filemap_fault,
fs/udf/file.c
84
.fault = filemap_fault,
fs/vboxsf/file.c
164
.fault = filemap_fault,
fs/xfs/xfs_file.c
1994
.fault = xfs_filemap_fault,
fs/zonefs/file.c
320
.fault = filemap_fault,
include/linux/cdrom.h
139
__u8 fault : 1;
include/linux/cdrom.h
148
__u8 fault : 1;
include/linux/ceph/messenger.h
44
void (*fault) (struct ceph_connection *con);
include/linux/enclosure.h
90
int fault;
include/linux/greybus/greybus_protocols.h
1777
__le32 fault;
include/linux/iommu.h
125
struct iommu_fault fault;
include/linux/led-class-flash.h
215
int led_get_flash_fault(struct led_classdev_flash *fled_cdev, u32 *fault);
include/linux/led-class-flash.h
47
int (*fault_get)(struct led_classdev_flash *fled_cdev, u32 *fault);
include/linux/mfd/wm8350/pmic.h
759
u16 stop, u16 fault);
include/linux/mm.h
766
vm_fault_t (*fault)(struct vm_fault *vmf);
include/linux/mm_types.h
1673
vm_fault_t (*fault)(const struct vm_special_mapping *sm,
include/media/i2c/adp1653.h
111
int fault;
include/rdma/ib_umem_odp.h
78
u64 bcnt, u64 access_mask, bool fault);
ipc/shm.c
545
return sfd->vm_ops->fault(vmf);
ipc/shm.c
614
WARN_ON(!sfd->vm_ops->fault);
ipc/shm.c
686
.fault = shm_fault,
kernel/bpf/arena.c
421
.fault = arena_vm_fault,
kernel/bpf/syscall.c
5090
bool fault;
kernel/bpf/syscall.c
5098
fault = copy_to_user(uinsns, insns_sanitized, ulen);
kernel/bpf/syscall.c
5100
if (fault)
kernel/events/uprobes.c
1693
.fault = xol_fault,
kernel/relay.c
54
.fault = relay_buf_fault,
kernel/trace/trace_events_user.c
445
struct user_event_enabler_fault *fault = container_of(
kernel/trace/trace_events_user.c
447
struct user_event_enabler *enabler = fault->enabler;
kernel/trace/trace_events_user.c
448
struct user_event_mm *mm = fault->mm;
kernel/trace/trace_events_user.c
450
int attempt = fault->attempt;
kernel/trace/trace_events_user.c
490
kmem_cache_free(fault_cache, fault);
kernel/trace/trace_events_user.c
497
struct user_event_enabler_fault *fault;
kernel/trace/trace_events_user.c
499
fault = kmem_cache_zalloc(fault_cache, GFP_NOWAIT);
kernel/trace/trace_events_user.c
501
if (!fault)
kernel/trace/trace_events_user.c
504
INIT_WORK(&fault->work, user_event_enabler_fault_fixup);
kernel/trace/trace_events_user.c
505
fault->mm = user_event_mm_get(mm);
kernel/trace/trace_events_user.c
506
fault->enabler = enabler;
kernel/trace/trace_events_user.c
507
fault->attempt = attempt;
kernel/trace/trace_events_user.c
512
if (!schedule_work(&fault->work)) {
kernel/trace/trace_events_user.c
517
kmem_cache_free(fault_cache, fault);
lib/vdso/datastore.c
96
.fault = vvar_fault,
mm/filemap.c
3983
.fault = filemap_fault,
mm/gup.c
629
} else if ((vma_is_anonymous(vma) || !vma->vm_ops->fault)) {
mm/hmm.c
256
goto fault;
mm/hmm.c
283
goto fault;
mm/hmm.c
286
goto fault;
mm/hmm.c
289
goto fault;
mm/hmm.c
307
goto fault;
mm/hmm.c
328
fault:
mm/hugetlb.c
4829
.fault = hugetlb_vm_op_fault,
mm/hugetlb_cgroup.c
107
fault = hugetlb_cgroup_counter_from_cgroup(h_cgroup, idx);
mm/hugetlb_cgroup.c
110
page_counter_init(fault, fault_parent, false);
mm/hugetlb_cgroup.c
114
fault->track_failcnt = true;
mm/hugetlb_cgroup.c
121
VM_BUG_ON(page_counter_set_max(fault, limit));
mm/hugetlb_cgroup.c
97
struct page_counter *fault, *fault_parent = NULL;
mm/memory.c
5364
ret = vma->vm_ops->fault(vmf);
mm/memory.c
5912
if (!vma->vm_ops->fault) {
mm/memory.c
621
vma->vm_ops ? vma->vm_ops->fault : NULL,
mm/mmap.c
1418
.fault = special_mapping_fault,
mm/mmap.c
1433
if (sm->fault)
mm/mmap.c
1434
return sm->fault(sm, vmf->vma, vmf);
mm/secretmem.c
112
.fault = secretmem_fault,
mm/shmem.c
5310
.fault = shmem_fault,
mm/shmem.c
5319
.fault = shmem_fault,
mm/slub.c
1297
u8 *fault;
mm/slub.c
1302
fault = memchr_inv(kasan_reset_tag(start), value, bytes);
mm/slub.c
1304
if (!fault)
mm/slub.c
1308
while (end > fault && end[-1] == value)
mm/slub.c
1315
what, fault, end - 1, fault - addr, fault[0], value);
mm/slub.c
1321
restore_bytes(s, what, value, fault, end);
mm/slub.c
1412
u8 *fault;
mm/slub.c
1438
fault = memchr_inv(kasan_reset_tag(pad), POISON_INUSE, remainder);
mm/slub.c
1440
if (!fault)
mm/slub.c
1442
while (end > fault && end[-1] == POISON_INUSE)
mm/slub.c
1446
fault, end - 1, fault - start);
mm/slub.c
1450
restore_bytes(s, "slab padding", POISON_INUSE, fault, end);
net/bluetooth/hidp/core.c
769
goto fault;
net/bluetooth/hidp/core.c
805
fault:
net/bridge/br_ioctl.c
121
goto fault;
net/bridge/br_ioctl.c
130
goto fault;
net/bridge/br_ioctl.c
135
fault:
net/ceph/messenger.c
1535
if (con->ops->fault)
net/ceph/messenger.c
1536
con->ops->fault(con);
net/ceph/messenger.c
1546
bool fault;
net/ceph/messenger.c
1552
if ((fault = con_sock_closed(con))) {
net/ceph/messenger.c
1583
fault = true;
net/ceph/messenger.c
1596
fault = true;
net/ceph/messenger.c
1601
if (fault)
net/ceph/messenger.c
1605
if (fault)
net/ceph/mon_client.c
1590
.fault = mon_fault,
net/ceph/osd_client.c
5911
.fault = osd_fault,
net/core/datagram.c
460
goto fault;
net/core/datagram.c
475
fault:
net/core/datagram.c
481
goto fault;
net/core/datagram.c
562
goto fault;
net/core/datagram.c
585
goto fault;
net/core/datagram.c
606
goto fault;
net/core/datagram.c
616
fault:
net/core/datagram.c
900
goto fault;
net/core/datagram.c
905
goto fault;
net/core/datagram.c
917
fault:
net/core/skbuff.c
3017
goto fault;
net/core/skbuff.c
3031
goto fault;
net/core/skbuff.c
3074
goto fault;
net/core/skbuff.c
3086
fault:
net/core/skbuff.c
3446
goto fault;
net/core/skbuff.c
3459
goto fault;
net/core/skbuff.c
3503
goto fault;
net/core/skbuff.c
3514
fault:
samples/vfio-mdev/mbochs.c
790
.fault = mbochs_region_vm_fault,
samples/vfio-mdev/mbochs.c
826
.fault = mbochs_dmabuf_vm_fault,
security/selinux/selinuxfs.c
464
.fault = sel_mmap_policy_fault,
sound/core/pcm_native.c
3740
.fault = snd_pcm_mmap_status_fault,
sound/core/pcm_native.c
3778
.fault = snd_pcm_mmap_control_fault,
sound/core/pcm_native.c
3910
.fault = snd_pcm_mmap_data_fault,
sound/usb/usx2y/us122l.c
126
.fault = usb_stream_hwdep_vm_fault,
sound/usb/usx2y/usX2Yhwdep.c
37
.fault = snd_us428ctls_vm_fault,
sound/usb/usx2y/usx2yhwdeppcm.c
684
.fault = snd_usx2y_hwdep_pcm_vm_fault,
tools/objtool/signal.c
21
unsigned long fault = (unsigned long)fault_addr;
tools/objtool/signal.c
24
return fault < stack_limit && fault >= stack_limit - 4096;
tools/testing/selftests/arm64/mte/check_child_memory.c
31
int i, parent_tag, child_tag, fault, child_status;
tools/testing/selftests/arm64/mte/check_child_memory.c
35
fault = 0;
tools/testing/selftests/arm64/mte/check_child_memory.c
47
fault = 1;
tools/testing/selftests/arm64/mte/check_child_memory.c
54
fault = 1;
tools/testing/selftests/arm64/mte/check_child_memory.c
62
fault = 1;
tools/testing/selftests/arm64/mte/check_child_memory.c
69
fault = 1;
tools/testing/selftests/arm64/mte/check_child_memory.c
73
_exit(fault);
tools/testing/selftests/arm64/mte/check_child_memory.c
78
fault = WEXITSTATUS(child_status);
tools/testing/selftests/arm64/mte/check_child_memory.c
80
fault = 1;
tools/testing/selftests/arm64/mte/check_child_memory.c
81
return (fault) ? KSFT_FAIL : KSFT_PASS;
tools/testing/selftests/arm64/mte/check_hugetlb_options.c
100
fault = 0;
tools/testing/selftests/arm64/mte/check_hugetlb_options.c
112
fault = 1;
tools/testing/selftests/arm64/mte/check_hugetlb_options.c
119
fault = 1;
tools/testing/selftests/arm64/mte/check_hugetlb_options.c
124
_exit(fault);
tools/testing/selftests/arm64/mte/check_hugetlb_options.c
129
fault = WEXITSTATUS(child_status);
tools/testing/selftests/arm64/mte/check_hugetlb_options.c
131
fault = 1;
tools/testing/selftests/arm64/mte/check_hugetlb_options.c
132
return (fault) ? KSFT_FAIL : KSFT_PASS;
tools/testing/selftests/arm64/mte/check_hugetlb_options.c
96
int i, parent_tag, child_tag, fault, child_status;
tools/testing/selftests/iommu/iommufd_utils.h
944
struct iommu_hwpt_pgfault fault = {};
tools/testing/selftests/iommu/iommufd_utils.h
952
bytes = read(fault_fd, &fault, sizeof(fault));
tools/testing/selftests/iommu/iommufd_utils.h
956
response.cookie = fault.cookie;
tools/testing/selftests/kvm/arm64/at.c
30
bool fault; \
tools/testing/selftests/kvm/arm64/at.c
38
fault = par & SYS_PAR_EL1_F; \
tools/testing/selftests/kvm/arm64/at.c
41
__GUEST_ASSERT((expect_fault) == fault, \
tools/testing/selftests/kvm/include/x86/apic.h
109
uint8_t fault = x2apic_write_reg_safe(reg, value);
tools/testing/selftests/kvm/include/x86/apic.h
111
__GUEST_ASSERT(!fault, "Unexpected fault 0x%x on WRMSR(%x) = %lx\n",
tools/testing/selftests/kvm/include/x86/apic.h
112
fault, APIC_BASE_MSR + (reg >> 4), value);
tools/testing/selftests/kvm/include/x86/apic.h
117
uint8_t fault = x2apic_write_reg_safe(reg, value);
tools/testing/selftests/kvm/include/x86/apic.h
119
__GUEST_ASSERT(fault == GP_VECTOR,
tools/testing/selftests/kvm/include/x86/apic.h
121
APIC_BASE_MSR + (reg >> 4), value, fault);
tools/testing/selftests/mm/khugepaged.c
1011
ops->fault(p, 0, (hpage_pmd_nr - max_ptes_shared) *
tools/testing/selftests/mm/khugepaged.c
1045
ops->fault(p, 0, hpage_pmd_size);
tools/testing/selftests/mm/khugepaged.c
1067
ops->fault(p, 0, size);
tools/testing/selftests/mm/khugepaged.c
308
ops->fault(p, 0, hpage_pmd_size);
tools/testing/selftests/mm/khugepaged.c
47
void (*fault)(void *p, unsigned long start, unsigned long end);
tools/testing/selftests/mm/khugepaged.c
475
.fault = &anon_fault,
tools/testing/selftests/mm/khugepaged.c
483
.fault = &file_fault,
tools/testing/selftests/mm/khugepaged.c
491
.fault = &anon_fault,
tools/testing/selftests/mm/khugepaged.c
584
ops->fault(p, 0, nr_hpages * hpage_pmd_size);
tools/testing/selftests/mm/khugepaged.c
648
ops->fault(p, 0, size);
tools/testing/selftests/mm/khugepaged.c
669
ops->fault(p, 0, page_size);
tools/testing/selftests/mm/khugepaged.c
694
ops->fault(p, 0, (hpage_pmd_nr - max_ptes_none - fault_nr_pages) * page_size);
tools/testing/selftests/mm/khugepaged.c
700
ops->fault(p, 0, (hpage_pmd_nr - max_ptes_none) * page_size);
tools/testing/selftests/mm/khugepaged.c
716
ops->fault(p, 0, hpage_pmd_size);
tools/testing/selftests/mm/khugepaged.c
743
ops->fault(p, 0, hpage_pmd_size);
tools/testing/selftests/mm/khugepaged.c
762
ops->fault(p, 0, hpage_pmd_size);
tools/testing/selftests/mm/khugepaged.c
842
ops->fault(BASE_ADDR, 0, hpage_pmd_size);
tools/testing/selftests/mm/khugepaged.c
871
ops->fault(p, 0, hpage_pmd_size);
tools/testing/selftests/mm/khugepaged.c
892
ops->fault(p, 0, page_size);
tools/testing/selftests/mm/khugepaged.c
909
ops->fault(p, page_size, 2 * page_size);
tools/testing/selftests/mm/khugepaged.c
954
ops->fault(p, 0, page_size);
tools/testing/selftests/mm/khugepaged.c
999
ops->fault(p, 0, (hpage_pmd_nr - max_ptes_shared - 1) * page_size);
tools/testing/selftests/net/tcp_ao/connect-deny.c
190
if (fault(KEYREJECT) && ret == -EKEYREJECTED) {
tools/testing/selftests/net/tcp_ao/connect-deny.c
192
} else if (ret == -ETIMEDOUT && fault(TIMEOUT)) {
tools/testing/selftests/net/tcp_ao/connect-deny.c
195
(fault(TIMEOUT) || fault(KEYREJECT))) {
tools/testing/selftests/net/tcp_ao/connect-deny.c
203
if (fault(TIMEOUT) || fault(KEYREJECT))
tools/testing/selftests/net/tcp_ao/connect-deny.c
58
if (!fault(TIMEOUT))
tools/testing/selftests/net/tcp_ao/connect-deny.c
61
if (!fault(KEYREJECT))
tools/testing/selftests/net/tcp_ao/connect-deny.c
66
if (fault(TIMEOUT))
tools/testing/selftests/net/tcp_ao/connect-deny.c
73
if (fault(TIMEOUT))
tools/testing/selftests/net/tcp_ao/key-management.c
121
if ((err == -EBUSY && fault(BUSY)) || (err == -EINVAL && fault(CURRNEXT))) {
tools/testing/selftests/net/tcp_ao/key-management.c
125
if (err && fault(FIXME)) {
tools/testing/selftests/net/tcp_ao/key-management.c
131
if (fault(BUSY) || fault(CURRNEXT)) {
tools/testing/selftests/net/tcp_ao/key-management.c
223
if (!err && !fault(CURRNEXT)) {
tools/testing/selftests/net/tcp_ao/key-management.c
227
if (err == -EINVAL && fault(CURRNEXT)) {
tools/testing/selftests/net/tcp_ao/restore.c
147
if (fault(TIMEOUT))
tools/testing/selftests/net/tcp_ao/restore.c
168
if (fault(TIMEOUT))
tools/testing/selftests/net/tcp_ao/restore.c
173
if (fault(TIMEOUT))
tools/testing/selftests/net/tcp_ao/restore.c
26
if (fault(TIMEOUT))
tools/testing/selftests/net/tcp_ao/restore.c
55
if (fault(TIMEOUT)) {
tools/testing/selftests/net/tcp_ao/unsigned-md5.c
295
if (fault(KEYREJECT) && ret == -EKEYREJECTED)
tools/testing/selftests/net/tcp_ao/unsigned-md5.c
297
else if (ret == -ETIMEDOUT && fault(TIMEOUT))
tools/testing/selftests/net/tcp_ao/unsigned-md5.c
300
(fault(TIMEOUT) || fault(KEYREJECT)))
tools/testing/selftests/net/tcp_ao/unsigned-md5.c
307
if (fault(TIMEOUT) || fault(KEYREJECT))
tools/testing/selftests/net/tcp_ao/unsigned-md5.c
349
expected_to_fail = fault(PREINSTALL_AO) && added_ao;
tools/testing/selftests/net/tcp_ao/unsigned-md5.c
350
expected_to_fail |= fault(PREINSTALL_MD5) && !added_ao;
tools/testing/selftests/net/tcp_ao/unsigned-md5.c
467
if (fault(POSTINSTALL)) {
tools/testing/selftests/net/tcp_ao/unsigned-md5.c
473
} else if (fault(POSTINSTALL)) {
tools/testing/selftests/net/tcp_ao/unsigned-md5.c
482
if (fault(POSTINSTALL)) {
tools/testing/selftests/net/tcp_ao/unsigned-md5.c
488
} else if (fault(POSTINSTALL)) {
tools/testing/selftests/net/tcp_ao/unsigned-md5.c
76
if (!fault(TIMEOUT))
tools/testing/selftests/net/tcp_ao/unsigned-md5.c
79
if (!fault(KEYREJECT))
tools/testing/selftests/net/tcp_ao/unsigned-md5.c
84
if (fault(TIMEOUT))
tools/testing/selftests/net/tcp_ao/unsigned-md5.c
91
if (fault(TIMEOUT))
tools/testing/vma/include/dup.h
625
vm_fault_t (*fault)(struct vm_fault *vmf);
virt/kvm/guest_memfd.c
461
.fault = kvm_gmem_fault_user_mapping,
virt/kvm/kvm_main.c
4080
.fault = kvm_vcpu_fault,