arch/arm64/include/asm/runtime-const.h
16
asm_inline("1:\t" \
arch/arm64/include/asm/runtime-const.h
29
asm_inline("1:\t" \
arch/loongarch/include/asm/bug.h
42
asm_inline volatile (__stringify(ASM_BUG_FLAGS(cond_str, flags)) extra);
arch/riscv/include/asm/runtime-const.h
107
asm_inline(RISCV_RUNTIME_CONST_64_PREAMBLE \
arch/riscv/include/asm/runtime-const.h
121
asm_inline(RISCV_RUNTIME_CONST_64_PREAMBLE \
arch/riscv/include/asm/runtime-const.h
135
asm_inline(RISCV_RUNTIME_CONST_64_PREAMBLE \
arch/riscv/include/asm/runtime-const.h
147
asm_inline(".option push\n\t" \
arch/riscv/include/asm/runtime-const.h
22
asm_inline(".option push\n\t" \
arch/riscv/include/asm/runtime-const.h
91
asm_inline(RISCV_RUNTIME_CONST_64_PREAMBLE \
arch/s390/boot/ipl_parm.c
41
asm_inline volatile(
arch/s390/boot/physmem_info.c
108
asm_inline volatile(
arch/s390/boot/physmem_info.c
129
asm_inline volatile(
arch/s390/boot/physmem_info.c
70
asm_inline volatile(
arch/s390/boot/startup.c
152
asm_inline volatile(
arch/s390/boot/startup.c
91
asm_inline volatile(
arch/s390/include/asm/alternative.h
164
asm_inline volatile(ALTERNATIVE(oldinstr, altinstr, feature) : : : "memory")
arch/s390/include/asm/alternative.h
167
asm_inline volatile(ALTERNATIVE_2(oldinstr, altinstr1, feature1, \
arch/s390/include/asm/alternative.h
172
asm_inline volatile (ALTERNATIVE(oldinstr, newinstr, feature) \
arch/s390/include/asm/alternative.h
177
asm_inline volatile(ALTERNATIVE(oldinstr, altinstr, feature) \
arch/s390/include/asm/bug.h
39
asm_inline volatile("\n" \
arch/s390/include/asm/bug.h
67
asm_inline volatile("\n" \
arch/s390/include/asm/cpu_mf.h
174
asm_inline volatile (
arch/s390/include/asm/cpu_mf.h
188
asm_inline volatile (
arch/s390/include/asm/cpu_mf.h
203
asm_inline volatile (
arch/s390/include/asm/current.h
23
asm_inline(
arch/s390/include/asm/diag.h
69
asm_inline volatile(
arch/s390/include/asm/fpu-insn.h
105
asm_inline volatile(
arch/s390/include/asm/futex.h
21
asm_inline volatile( \
arch/s390/include/asm/futex.h
86
asm_inline volatile(
arch/s390/include/asm/lowcore.h
225
asm_inline(
arch/s390/include/asm/preempt.h
121
asm_inline(
arch/s390/include/asm/preempt.h
34
asm_inline(
arch/s390/include/asm/preempt.h
90
asm_inline(
arch/s390/include/asm/processor.h
413
asm_inline volatile(
arch/s390/include/asm/runtime-const.h
11
asm_inline( \
arch/s390/include/asm/runtime-const.h
27
asm_inline( \
arch/s390/include/asm/smp.h
22
asm_inline(
arch/s390/include/asm/spinlock.h
26
asm_inline(
arch/s390/include/asm/spinlock.h
99
asm_inline volatile(
arch/s390/include/asm/uaccess.h
151
asm_inline volatile( \
arch/s390/include/asm/uaccess.h
267
asm_inline volatile( \
arch/s390/include/asm/uaccess.h
375
asm_inline volatile(
arch/s390/include/asm/uaccess.h
44
asm_inline volatile(
arch/s390/include/asm/uaccess.h
449
asm_inline volatile( \
arch/s390/include/asm/uaccess.h
72
asm_inline volatile(
arch/s390/include/asm/word-at-a-time.h
55
asm_inline volatile(
arch/s390/kernel/cert_store.c
238
asm_inline volatile(
arch/s390/kernel/diag/diag.c
198
asm_inline volatile(
arch/s390/kernel/diag/diag.c
289
asm_inline volatile("\n"
arch/s390/kernel/ipl.c
189
asm_inline volatile(
arch/s390/kernel/traps.c
298
asm_inline volatile(
arch/s390/kvm/gaccess.c
844
asm_inline volatile(
arch/s390/lib/spinlock.c
110
asm_inline volatile(
arch/s390/lib/spinlock.c
125
asm_inline volatile(
arch/s390/lib/spinlock.c
97
asm_inline volatile(
arch/s390/lib/uaccess.c
131
asm_inline volatile(
arch/s390/lib/uaccess.c
158
asm_inline volatile(
arch/s390/lib/uaccess.c
185
asm_inline volatile(
arch/s390/lib/uaccess.c
48
asm_inline volatile(
arch/s390/mm/pfault.c
60
asm_inline volatile(
arch/s390/mm/pfault.c
82
asm_inline volatile(
arch/s390/pci/pci_clp.c
57
asm_inline volatile (
arch/s390/pci/pci_clp.c
80
asm_inline volatile (
arch/s390/pci/pci_insn.c
163
asm_inline volatile (
arch/s390/pci/pci_insn.c
232
asm_inline volatile (
arch/s390/pci/pci_insn.c
270
asm_inline volatile (
arch/s390/pci/pci_insn.c
324
asm_inline volatile (
arch/s390/pci/pci_insn.c
359
asm_inline volatile (
arch/s390/pci/pci_insn.c
413
asm_inline volatile (
arch/s390/pci/pci_mmio.c
226
asm_inline volatile (
arch/s390/pci/pci_mmio.c
39
asm_inline volatile (
arch/s390/pci/pci_mmio.c
72
asm_inline volatile (
arch/x86/include/asm/alternative.h
247
asm_inline volatile(ALTERNATIVE(oldinstr, newinstr, ft_flags) : : : "memory")
arch/x86/include/asm/alternative.h
250
asm_inline volatile(ALTERNATIVE_2(oldinstr, newinstr1, ft_flags1, newinstr2, ft_flags2) ::: "memory")
arch/x86/include/asm/alternative.h
261
asm_inline volatile(ALTERNATIVE(oldinstr, newinstr, ft_flags) \
arch/x86/include/asm/alternative.h
266
asm_inline volatile(ALTERNATIVE(oldinstr, newinstr, ft_flags) \
arch/x86/include/asm/alternative.h
280
asm_inline volatile(ALTERNATIVE("call %c[old]", "call %c[new]", ft_flags) \
arch/x86/include/asm/alternative.h
294
asm_inline volatile(ALTERNATIVE_2("call %c[old]", "call %c[new1]", ft_flags1, \
arch/x86/include/asm/arch_hweight.h
19
asm_inline (ALTERNATIVE("call __sw_hweight32",
arch/x86/include/asm/arch_hweight.h
48
asm_inline (ALTERNATIVE("call __sw_hweight64",
arch/x86/include/asm/atomic.h
119
asm_inline volatile(LOCK_PREFIX "andl %1, %0"
arch/x86/include/asm/atomic.h
137
asm_inline volatile(LOCK_PREFIX "orl %1, %0"
arch/x86/include/asm/atomic.h
155
asm_inline volatile(LOCK_PREFIX "xorl %1, %0"
arch/x86/include/asm/atomic.h
33
asm_inline volatile(LOCK_PREFIX "addl %1, %0"
arch/x86/include/asm/atomic.h
40
asm_inline volatile(LOCK_PREFIX "subl %1, %0"
arch/x86/include/asm/atomic.h
53
asm_inline volatile(LOCK_PREFIX "incl %0"
arch/x86/include/asm/atomic.h
60
asm_inline volatile(LOCK_PREFIX "decl %0"
arch/x86/include/asm/atomic64_64.h
113
asm_inline volatile(LOCK_PREFIX "andq %1, %0"
arch/x86/include/asm/atomic64_64.h
131
asm_inline volatile(LOCK_PREFIX "orq %1, %0"
arch/x86/include/asm/atomic64_64.h
149
asm_inline volatile(LOCK_PREFIX "xorq %1, %0"
arch/x86/include/asm/atomic64_64.h
25
asm_inline volatile(LOCK_PREFIX "addq %1, %0"
arch/x86/include/asm/atomic64_64.h
32
asm_inline volatile(LOCK_PREFIX "subq %1, %0"
arch/x86/include/asm/atomic64_64.h
45
asm_inline volatile(LOCK_PREFIX "incq %0"
arch/x86/include/asm/atomic64_64.h
53
asm_inline volatile(LOCK_PREFIX "decq %0"
arch/x86/include/asm/bitops.h
101
asm_inline volatile(LOCK_PREFIX "xorb %2,%1"
arch/x86/include/asm/bitops.h
124
asm_inline volatile(LOCK_PREFIX "xorb %b1,%0"
arch/x86/include/asm/bitops.h
128
asm_inline volatile(LOCK_PREFIX __ASM_SIZE(btc) " %1,%0"
arch/x86/include/asm/bitops.h
55
asm_inline volatile(LOCK_PREFIX "orb %b1,%0"
arch/x86/include/asm/bitops.h
60
asm_inline volatile(LOCK_PREFIX __ASM_SIZE(bts) " %1,%0"
arch/x86/include/asm/bitops.h
75
asm_inline volatile(LOCK_PREFIX "andb %b1,%0"
arch/x86/include/asm/bitops.h
79
asm_inline volatile(LOCK_PREFIX __ASM_SIZE(btr) " %1,%0"
arch/x86/include/asm/bug.h
160
asm_inline volatile("lea (2f)(%%rip), %[addr]\n1:\n" \
arch/x86/include/asm/bug.h
88
asm_inline volatile("1:\t" ins "\n" \
arch/x86/include/asm/cmpxchg.h
103
asm_inline volatile(lock "cmpxchgw %2, %1" \
arch/x86/include/asm/cmpxchg.h
112
asm_inline volatile(lock "cmpxchgl %2, %1" \
arch/x86/include/asm/cmpxchg.h
121
asm_inline volatile(lock "cmpxchgq %2, %1" \
arch/x86/include/asm/cmpxchg.h
168
asm_inline volatile(lock "cmpxchgb %[new], %[ptr]" \
arch/x86/include/asm/cmpxchg.h
179
asm_inline volatile(lock "cmpxchgw %[new], %[ptr]" \
arch/x86/include/asm/cmpxchg.h
190
asm_inline volatile(lock "cmpxchgl %[new], %[ptr]" \
arch/x86/include/asm/cmpxchg.h
201
asm_inline volatile(lock "cmpxchgq %[new], %[ptr]" \
arch/x86/include/asm/cmpxchg.h
47
asm_inline volatile (lock #op "b %b0, %1" \
arch/x86/include/asm/cmpxchg.h
52
asm_inline volatile (lock #op "w %w0, %1" \
arch/x86/include/asm/cmpxchg.h
57
asm_inline volatile (lock #op "l %0, %1" \
arch/x86/include/asm/cmpxchg.h
62
asm_inline volatile (lock #op "q %q0, %1" \
arch/x86/include/asm/cmpxchg.h
94
asm_inline volatile(lock "cmpxchgb %2, %1" \
arch/x86/include/asm/cmpxchg_32.h
123
asm_inline volatile( \
arch/x86/include/asm/cmpxchg_32.h
22
asm_inline volatile(_lock "cmpxchg8b %[ptr]" \
arch/x86/include/asm/cmpxchg_32.h
48
asm_inline volatile(_lock "cmpxchg8b %[ptr]" \
arch/x86/include/asm/cmpxchg_32.h
93
asm_inline volatile( \
arch/x86/include/asm/cmpxchg_64.h
41
asm_inline volatile(_lock "cmpxchg16b %[ptr]" \
arch/x86/include/asm/cmpxchg_64.h
68
asm_inline volatile(_lock "cmpxchg16b %[ptr]" \
arch/x86/include/asm/futex.h
83
asm_inline volatile("\n"
arch/x86/include/asm/gsseg.h
23
asm_inline volatile("1: " LKGS_DI
arch/x86/include/asm/irq_stack.h
87
asm_inline volatile( \
arch/x86/include/asm/nospec-branch.h
540
asm_inline volatile(ALTERNATIVE("", "call write_ibpb", X86_FEATURE_IBPB)
arch/x86/include/asm/percpu.h
338
asm_inline qual ( \
arch/x86/include/asm/percpu.h
367
asm_inline qual ( \
arch/x86/include/asm/percpu.h
406
asm_inline qual ( \
arch/x86/include/asm/percpu.h
435
asm_inline qual ( \
arch/x86/include/asm/rmwcc.h
12
asm_inline volatile (fullop \
arch/x86/include/asm/runtime-const.h
23
asm_inline("mov %1,%0\n1:\n" \
arch/x86/include/asm/runtime-const.h
37
asm_inline("shrl $12,%k0\n1:\n" \
arch/x86/include/asm/special_insns.h
201
asm_inline volatile(ALTERNATIVE_2(
arch/x86/include/asm/uaccess_64.h
29
asm_inline (ALTERNATIVE("", "and " __percpu_arg([mask]) ", %[addr]",
arch/x86/include/asm/vdso/sys_call.h
50
asm_inline volatile( \
arch/x86/include/asm/vmware.h
111
asm_inline volatile (VMWARE_HYPERCALL
arch/x86/include/asm/vmware.h
136
asm_inline volatile (VMWARE_HYPERCALL
arch/x86/include/asm/vmware.h
161
asm_inline volatile (VMWARE_HYPERCALL
arch/x86/include/asm/vmware.h
187
asm_inline volatile (VMWARE_HYPERCALL
arch/x86/include/asm/vmware.h
215
asm_inline volatile (VMWARE_HYPERCALL
arch/x86/include/asm/vmware.h
243
asm_inline volatile (VMWARE_HYPERCALL
arch/x86/include/asm/vmware.h
277
asm_inline volatile (
arch/x86/include/asm/vmware.h
305
asm_inline volatile (
arch/x86/kernel/alternative.c
2374
asm_inline volatile (
arch/x86/kernel/cpu/vmware.c
103
asm_inline volatile ("movw %[port], %%dx; inl (%%dx), %%eax"
arch/x86/kernel/cpu/vmware.c
79
asm_inline volatile ("vmcall"
arch/x86/kernel/cpu/vmware.c
91
asm_inline volatile ("vmmcall"
drivers/gpu/drm/vmwgfx/vmwgfx_msg_arm64.h
113
asm_inline volatile (
drivers/gpu/drm/vmwgfx/vmwgfx_msg_arm64.h
143
asm_inline volatile (
drivers/gpu/drm/vmwgfx/vmwgfx_msg_arm64.h
173
asm_inline volatile (
drivers/gpu/drm/vmwgfx/vmwgfx_msg_arm64.h
62
asm_inline volatile (
drivers/gpu/drm/vmwgfx/vmwgfx_msg_arm64.h
87
asm_inline volatile (
drivers/s390/block/dasd_diag.c
79
asm_inline volatile(
drivers/s390/char/diag_ftp.c
108
asm_inline volatile(
drivers/s390/char/sclp.h
321
asm_inline volatile(
drivers/s390/cio/ioasm.c
109
asm_inline volatile(
drivers/s390/cio/ioasm.c
181
asm_inline volatile(
drivers/s390/cio/ioasm.c
25
asm_inline volatile(
drivers/s390/cio/ioasm.c
55
asm_inline volatile(
samples/bpf/asm_goto_workaround.h
22
#ifdef asm_inline
tools/testing/selftests/rseq/slice_test.c
40
#ifndef asm_inline