Searched refs:arch_try_cmpxchg (Results 1 – 11 of 11) sorted by relevance
| /linux/arch/s390/include/asm/ |
| H A D | cmpxchg.h | 110 #define arch_try_cmpxchg(ptr, oldp, new) \ macro 155 #define arch_try_cmpxchg(ptr, oldp, new) \ macro 170 #define arch_try_cmpxchg64 arch_try_cmpxchg 171 #define arch_try_cmpxchg_local arch_try_cmpxchg 172 #define arch_try_cmpxchg64_local arch_try_cmpxchg 187 } while (!arch_try_cmpxchg((u32 *)ptr, &old, new)); in __arch_xchg1() 202 } while (!arch_try_cmpxchg((u32 *)ptr, &old, new)); in __arch_xchg2() 217 } while (!arch_try_cmpxchg((u32 *)ptr, &old, x & 0xffffffff)); in __arch_xchg() 224 } while (!arch_try_cmpxchg((u64 *)ptr, &old, x)); in __arch_xchg()
|
| H A D | spinlock.h | 79 return likely(arch_try_cmpxchg(&lp->lock, &old, spinlock_lockval())); in arch_spin_trylock_once() 142 if (!arch_try_cmpxchg(&rw->cnts, &old, 0x30000)) in arch_write_lock() 157 return (!(old & 0xffff0000) && arch_try_cmpxchg(&rw->cnts, &old, old + 1)); in arch_read_trylock() 165 return !old && arch_try_cmpxchg(&rw->cnts, &old, 0x30000); in arch_write_trylock()
|
| H A D | atomic.h | 119 return arch_try_cmpxchg(&v->counter, old, new); in arch_atomic_try_cmpxchg() 199 return arch_try_cmpxchg(&v->counter, old, new); in arch_atomic64_try_cmpxchg()
|
| H A D | preempt.h | 52 } while (!arch_try_cmpxchg(&get_lowcore()->preempt_count, &old, new)); in preempt_count_set()
|
| /linux/arch/s390/lib/ |
| H A D | spinlock.c | 181 if (arch_try_cmpxchg(&lp->lock, &old, new)) in arch_spin_lock_queued() 189 if (arch_try_cmpxchg(&lp->lock, &old, new)) in arch_spin_lock_queued() 226 if (arch_try_cmpxchg(&lp->lock, &old, new)) in arch_spin_lock_queued() 300 if (arch_try_cmpxchg(&lp->lock, &owner, cpu)) in arch_spin_trylock_retry() 342 arch_try_cmpxchg(&rw->cnts, &old, old | 0x10000)) in arch_write_lock_wait()
|
| /linux/include/linux/atomic/ |
| H A D | atomic-arch-fallback.h | 183 #if defined(arch_try_cmpxchg) 184 #define raw_try_cmpxchg arch_try_cmpxchg 187 __atomic_op_fence(arch_try_cmpxchg, __VA_ARGS__) 203 __atomic_op_acquire(arch_try_cmpxchg, __VA_ARGS__) 204 #elif defined(arch_try_cmpxchg) 205 #define raw_try_cmpxchg_acquire arch_try_cmpxchg 221 __atomic_op_release(arch_try_cmpxchg, __VA_ARGS__) 222 #elif defined(arch_try_cmpxchg) 223 #define raw_try_cmpxchg_release arch_try_cmpxchg 237 #elif defined(arch_try_cmpxchg) [all...] |
| /linux/arch/x86/include/asm/ |
| H A D | cmpxchg_64.h | 20 arch_try_cmpxchg((ptr), (po), (n)); \
|
| H A D | atomic.h | 107 return arch_try_cmpxchg(&v->counter, old, new); in arch_atomic_try_cmpxchg()
|
| H A D | atomic64_64.h | 101 return arch_try_cmpxchg(&v->counter, old, new); in arch_atomic64_try_cmpxchg()
|
| H A D | cmpxchg.h | 226 #define arch_try_cmpxchg(ptr, pold, new) \ macro
|
| /linux/arch/s390/kvm/ |
| H A D | dat.c | 156 return arch_try_cmpxchg((long *)crstep, &old.val, new.val); in dat_crstep_xchg_atomic()
|