Symbol: LOCK_PREFIX
arch/x86/include/asm/alternative.h
323
.macro LOCK_PREFIX
arch/x86/include/asm/alternative.h
331
.macro LOCK_PREFIX
arch/x86/include/asm/atomic.h
119
asm_inline volatile(LOCK_PREFIX "andl %1, %0"
arch/x86/include/asm/atomic.h
137
asm_inline volatile(LOCK_PREFIX "orl %1, %0"
arch/x86/include/asm/atomic.h
155
asm_inline volatile(LOCK_PREFIX "xorl %1, %0"
arch/x86/include/asm/atomic.h
33
asm_inline volatile(LOCK_PREFIX "addl %1, %0"
arch/x86/include/asm/atomic.h
40
asm_inline volatile(LOCK_PREFIX "subl %1, %0"
arch/x86/include/asm/atomic.h
47
return GEN_BINARY_RMWcc(LOCK_PREFIX "subl", v->counter, e, "er", i);
arch/x86/include/asm/atomic.h
53
asm_inline volatile(LOCK_PREFIX "incl %0"
arch/x86/include/asm/atomic.h
60
asm_inline volatile(LOCK_PREFIX "decl %0"
arch/x86/include/asm/atomic.h
67
return GEN_UNARY_RMWcc(LOCK_PREFIX "decl", v->counter, e);
arch/x86/include/asm/atomic.h
73
return GEN_UNARY_RMWcc(LOCK_PREFIX "incl", v->counter, e);
arch/x86/include/asm/atomic.h
79
return GEN_BINARY_RMWcc(LOCK_PREFIX "addl", v->counter, s, "er", i);
arch/x86/include/asm/atomic64_64.h
113
asm_inline volatile(LOCK_PREFIX "andq %1, %0"
arch/x86/include/asm/atomic64_64.h
131
asm_inline volatile(LOCK_PREFIX "orq %1, %0"
arch/x86/include/asm/atomic64_64.h
149
asm_inline volatile(LOCK_PREFIX "xorq %1, %0"
arch/x86/include/asm/atomic64_64.h
25
asm_inline volatile(LOCK_PREFIX "addq %1, %0"
arch/x86/include/asm/atomic64_64.h
32
asm_inline volatile(LOCK_PREFIX "subq %1, %0"
arch/x86/include/asm/atomic64_64.h
39
return GEN_BINARY_RMWcc(LOCK_PREFIX "subq", v->counter, e, "er", i);
arch/x86/include/asm/atomic64_64.h
45
asm_inline volatile(LOCK_PREFIX "incq %0"
arch/x86/include/asm/atomic64_64.h
53
asm_inline volatile(LOCK_PREFIX "decq %0"
arch/x86/include/asm/atomic64_64.h
61
return GEN_UNARY_RMWcc(LOCK_PREFIX "decq", v->counter, e);
arch/x86/include/asm/atomic64_64.h
67
return GEN_UNARY_RMWcc(LOCK_PREFIX "incq", v->counter, e);
arch/x86/include/asm/atomic64_64.h
73
return GEN_BINARY_RMWcc(LOCK_PREFIX "addq", v->counter, s, "er", i);
arch/x86/include/asm/bitops.h
101
asm_inline volatile(LOCK_PREFIX "xorb %2,%1"
arch/x86/include/asm/bitops.h
124
asm_inline volatile(LOCK_PREFIX "xorb %b1,%0"
arch/x86/include/asm/bitops.h
128
asm_inline volatile(LOCK_PREFIX __ASM_SIZE(btc) " %1,%0"
arch/x86/include/asm/bitops.h
136
return GEN_BINARY_RMWcc(LOCK_PREFIX __ASM_SIZE(bts), *addr, c, "Ir", nr);
arch/x86/include/asm/bitops.h
159
return GEN_BINARY_RMWcc(LOCK_PREFIX __ASM_SIZE(btr), *addr, c, "Ir", nr);
arch/x86/include/asm/bitops.h
196
return GEN_BINARY_RMWcc(LOCK_PREFIX __ASM_SIZE(btc), *addr, c, "Ir", nr);
arch/x86/include/asm/bitops.h
55
asm_inline volatile(LOCK_PREFIX "orb %b1,%0"
arch/x86/include/asm/bitops.h
60
asm_inline volatile(LOCK_PREFIX __ASM_SIZE(bts) " %1,%0"
arch/x86/include/asm/bitops.h
75
asm_inline volatile(LOCK_PREFIX "andb %b1,%0"
arch/x86/include/asm/bitops.h
79
asm_inline volatile(LOCK_PREFIX __ASM_SIZE(btr) " %1,%0"
arch/x86/include/asm/cmpxchg.h
134
__raw_cmpxchg((ptr), (old), (new), (size), LOCK_PREFIX)
arch/x86/include/asm/cmpxchg.h
218
__raw_try_cmpxchg((ptr), (pold), (new), (size), LOCK_PREFIX)
arch/x86/include/asm/cmpxchg.h
242
#define xadd(ptr, inc) __xadd((ptr), (inc), LOCK_PREFIX)
arch/x86/include/asm/cmpxchg_32.h
34
return __arch_cmpxchg64(ptr, old, new, LOCK_PREFIX);
arch/x86/include/asm/cmpxchg_32.h
63
return __arch_try_cmpxchg64(ptr, oldp, new, LOCK_PREFIX);
arch/x86/include/asm/cmpxchg_64.h
52
return __arch_cmpxchg128(ptr, old, new, LOCK_PREFIX);
arch/x86/include/asm/cmpxchg_64.h
83
return __arch_try_cmpxchg128(ptr, oldp, new, LOCK_PREFIX);
arch/x86/include/asm/futex.h
35
"3:\t" LOCK_PREFIX "cmpxchgl %3, %2\n" \
arch/x86/include/asm/futex.h
57
unsafe_atomic_op1(LOCK_PREFIX "xaddl %0, %2", oval, uaddr, oparg, Efault);
arch/x86/include/asm/futex.h
84
"1:\t" LOCK_PREFIX "cmpxchgl %3, %2\n"
arch/x86/include/asm/qspinlock.h
26
val = GEN_BINARY_RMWcc(LOCK_PREFIX "btsl", lock->val.counter, c,
arch/x86/include/asm/qspinlock_paravirt.h
45
LOCK_PREFIX "cmpxchg %dl,(%rdi)\n\t" \
arch/x86/include/asm/uaccess.h
379
"1: " LOCK_PREFIX "cmpxchg"itype" %[new], %[ptr]\n"\
arch/x86/include/asm/uaccess.h
398
"1: " LOCK_PREFIX "cmpxchg8b %[ptr]\n" \
arch/x86/include/asm/uaccess.h
419
"1: " LOCK_PREFIX "cmpxchg"itype" %[new], %[ptr]\n"\
arch/x86/include/asm/uaccess.h
449
"1: " LOCK_PREFIX "cmpxchg8b %[ptr]\n" \
arch/x86/kvm/xen.c
667
asm volatile(LOCK_PREFIX "orq %0, %1\n"
arch/x86/kvm/xen.c
669
LOCK_PREFIX "andq %0, %2\n"
arch/x86/kvm/xen.c
679
asm volatile(LOCK_PREFIX "orl %0, %1\n"
arch/x86/kvm/xen.c
681
LOCK_PREFIX "andl %0, %2\n"
tools/arch/x86/include/asm/atomic.h
52
asm volatile(LOCK_PREFIX "incl %0"
tools/arch/x86/include/asm/atomic.h
66
GEN_UNARY_RMWcc(LOCK_PREFIX "decl", v->counter, "%0", "e");
tools/arch/x86/include/asm/atomic.h
76
GEN_BINARY_RMWcc(LOCK_PREFIX __ASM_SIZE(bts), *addr, "Ir", nr, "%0", "c");
tools/arch/x86/include/asm/atomic.h
81
GEN_BINARY_RMWcc(LOCK_PREFIX __ASM_SIZE(btc), *addr, "Ir", nr, "%0", "c");
tools/arch/x86/include/asm/cmpxchg.h
84
__raw_cmpxchg((ptr), (old), (new), (size), LOCK_PREFIX)
tools/testing/selftests/kvm/include/x86/kvm_util_arch.h
67
__asm__ __volatile__(KVM_FEP LOCK_PREFIX "cmpxchg %[new], %[ptr]" \