| /src/contrib/llvm-project/libcxx/include/__atomic/ |
| H A D | aliases.h | 28 using atomic_bool = atomic<bool>; 29 using atomic_char = atomic<char>; 30 using atomic_schar = atomic<signed char>; 31 using atomic_uchar = atomic<unsigned char>; 32 using atomic_short = atomic<short>; 33 using atomic_ushort = atomic<unsigned short>; 34 using atomic_int = atomic<int>; 35 using atomic_uint = atomic<unsigned int>; 36 using atomic_long = atomic<long>; 37 using atomic_ulong = atomic<unsigned long>; [all …]
|
| H A D | atomic.h | 36 struct atomic : public __atomic_base<_Tp> { struct 42 _LIBCPP_HIDE_FROM_ABI atomic() = default; argument 44 _LIBCPP_HIDE_FROM_ABI atomic() _NOEXCEPT = default; 47 _LIBCPP_HIDE_FROM_ABI _LIBCPP_CONSTEXPR atomic(_Tp __d) _NOEXCEPT : __base(__d) {} in atomic() argument 58 atomic& operator=(const atomic&) = delete; argument 59 atomic& operator=(const atomic&) volatile = delete; argument 65 struct atomic<_Tp*> : public __atomic_base<_Tp*> { 70 _LIBCPP_HIDE_FROM_ABI atomic() _NOEXCEPT = default; 72 _LIBCPP_HIDE_FROM_ABI _LIBCPP_CONSTEXPR atomic(_Tp* __d) _NOEXCEPT : __base(__d) {} 120 atomic& operator=(const atomic&) = delete; [all …]
|
| /src/contrib/llvm-project/libcxx/include/ |
| H A D | atomic | 14 atomic synopsis 62 struct atomic 70 atomic() noexcept = default; // until C++20 71 constexpr atomic() noexcept(is_nothrow_default_constructible_v<T>); // since C++20 72 constexpr atomic(T desr) noexcept; 73 atomic(const atomic&) = delete; 74 atomic& operator=(const atomic&) = delete; 75 atomic& operator=(const atomic&) volatile = delete; 113 struct atomic<integral> 122 atomic() noexcept = default; [all …]
|
| /src/contrib/llvm-project/llvm/lib/Target/WebAssembly/ |
| H A D | WebAssemblyInstrAtomics.td | 40 "memory.atomic.notify \t$dst, ${off}(${addr})${p2align}, $count", 41 "memory.atomic.notify \t${off}${p2align}", 0x00, false>; 46 "memory.atomic.notify \t$dst, ${off}(${addr})${p2align}, $count", 47 "memory.atomic.notify \t${off}${p2align}", 0x00, true>; 54 "memory.atomic.wait32 \t$dst, ${off}(${addr})${p2align}, $exp, $timeout", 55 "memory.atomic.wait32 \t${off}${p2align}", 0x01, false>; 61 "memory.atomic.wait32 \t$dst, ${off}(${addr})${p2align}, $exp, $timeout", 62 "memory.atomic.wait32 \t${off}${p2align}", 0x01, true>; 68 "memory.atomic.wait64 \t$dst, ${off}(${addr})${p2align}, $exp, $timeout", 69 "memory.atomic.wait64 \t${off}${p2align}", 0x02, false>; [all …]
|
| /src/sys/dev/qat/qat_api/common/utils/ |
| H A D | lac_lock_free_stack.h | 12 uint64_t atomic; member 41 old_top.atomic = stack->top.atomic; in pop() 48 } while (!__sync_bool_compare_and_swap(&stack->top.atomic, in pop() 49 old_top.atomic, in pop() 50 new_top.atomic)); in pop() 62 old_top.atomic = stack->top.atomic; in push() 66 } while (!__sync_bool_compare_and_swap(&stack->top.atomic, in push() 67 old_top.atomic, in push() 68 new_top.atomic)); in push()
|
| /src/contrib/llvm-project/lldb/include/lldb/Utility/ |
| H A D | Timer.h | 33 std::atomic<uint64_t> m_nanos; 34 std::atomic<uint64_t> m_nanos_total; 35 std::atomic<uint64_t> m_count; 36 std::atomic<Category *> m_next; 71 static std::atomic<bool> g_quiet; 72 static std::atomic<unsigned> g_display_depth;
|
| /src/contrib/llvm-project/compiler-rt/lib/xray/ |
| H A D | xray_x86_64.cpp | 161 reinterpret_cast<std::atomic<uint16_t> *>(Address), MovR10Seq, in patchFunctionEntry() 165 reinterpret_cast<std::atomic<uint16_t> *>(Address), Jmp9Seq, in patchFunctionEntry() 208 reinterpret_cast<std::atomic<uint16_t> *>(Address), MovR10Seq, in patchFunctionExit() 212 reinterpret_cast<std::atomic<uint8_t> *>(Address), RetOpCode, in patchFunctionExit() 238 reinterpret_cast<std::atomic<uint16_t> *>(Address), MovR10Seq, in patchFunctionTailExit() 242 reinterpret_cast<std::atomic<uint16_t> *>(Address), Jmp9Seq, in patchFunctionTailExit() 267 reinterpret_cast<std::atomic<uint16_t> *>(Address), NopwSeq, in patchCustomEvent() 271 reinterpret_cast<std::atomic<uint16_t> *>(Address), Jmp15Seq, in patchCustomEvent() 299 reinterpret_cast<std::atomic<uint16_t> *>(Address), NopwSeq, in patchTypedEvent() 303 reinterpret_cast<std::atomic<uint16_t> *>(Address), Jmp20Seq, in patchTypedEvent()
|
| H A D | xray_AArch64.cpp | 78 reinterpret_cast<std::atomic<uint32_t> *>(FirstAddress), in patchSled() 82 reinterpret_cast<std::atomic<uint32_t> *>(FirstAddress), in patchSled() 124 reinterpret_cast<std::atomic<uint32_t> *>(Sled.address()), Inst, in patchCustomEvent() 135 reinterpret_cast<std::atomic<uint32_t> *>(Sled.address()), Inst, in patchTypedEvent()
|
| /src/contrib/llvm-project/clang/lib/Tooling/Inclusions/Stdlib/ |
| H A D | StdSpecialSymbolMap.inc | 245 // Only add headers for the generic atomic template. 247 SYMBOL(atomic, std::, <atomic>) 249 SYMBOL(atomic_bool, std::, <atomic>) 251 SYMBOL(atomic_char, std::, <atomic>) 253 SYMBOL(atomic_char16_t, std::, <atomic>) 255 SYMBOL(atomic_char32_t, std::, <atomic>) 257 SYMBOL(atomic_char8_t, std::, <atomic>) 259 SYMBOL(atomic_int, std::, <atomic>) 261 SYMBOL(atomic_int16_t, std::, <atomic>) 263 SYMBOL(atomic_int32_t, std::, <atomic>) [all …]
|
| /src/contrib/llvm-project/llvm/lib/DWARFLinker/Parallel/ |
| H A D | ArrayList.h | 111 std::atomic<ItemsGroup *> Next = nullptr; 117 std::atomic<size_t> ItemsCount = 0; 131 bool allocateNewGroup(std::atomic<ItemsGroup *> &AtomicGroup) { in allocateNewGroup() 158 std::atomic<ItemsGroup *> GroupsHead = nullptr; 159 std::atomic<ItemsGroup *> LastGroup = nullptr;
|
| H A D | DWARFLinkerCompileUnit.h | 142 std::atomic<bool> &HasNewInterconnectedCUs); 178 std::atomic<uint16_t> Flags = {0}; 316 return reinterpret_cast<std::atomic<uint64_t> *>(&OutDieOffsetArray[Idx]) in getDieOutOffset() 323 return reinterpret_cast<std::atomic<TypeEntry *> *>(&TypeEntries[Idx]) in getDieTypeEntry() 330 return reinterpret_cast<std::atomic<uint64_t> *>( in getDieOutOffset() 338 return reinterpret_cast<std::atomic<TypeEntry *> *>( in getDieTypeEntry() 346 reinterpret_cast<std::atomic<uint64_t> *>(&OutDieOffsetArray[Idx]) in rememberDieOutOffset() 353 reinterpret_cast<std::atomic<TypeEntry *> *>(&TypeEntries[Idx]) in setDieTypeEntry() 361 reinterpret_cast<std::atomic<TypeEntry *> *>( in setDieTypeEntry() 717 std::atomic<Stage> Stage;
|
| H A D | TypePool.h | 27 using TypeEntry = StringMapEntry<std::atomic<TypeEntryBody *>>; 59 std::atomic<DIE *> Die = {nullptr}; 62 std::atomic<DIE *> DeclarationDie = {nullptr}; 65 std::atomic<bool> ParentIsDeclaration = {true};
|
| H A D | DependencyTracker.h | 43 std::atomic<bool> &HasNewInterconnectedCUs); 226 std::atomic<bool> &HasNewInterconnectedCUs); 233 std::atomic<bool> &HasNewInterconnectedCUs); 246 std::atomic<bool> &HasNewInterconnectedCUs);
|
| H A D | DWARFLinkerImpl.h | 197 std::atomic<bool> HasNewInterconnectedCUs = {false}; 199 std::atomic<bool> HasNewGlobalDependency = {false}; 202 std::atomic<size_t> &UniqueUnitID; 206 std::atomic<size_t> &UniqueUnitID); 340 std::atomic<size_t> UniqueUnitID;
|
| /src/contrib/llvm-project/openmp/runtime/src/ |
| H A D | kmp_wait_release.h | 109 std::atomic<bool> *sleepLoc; 116 kmp_flag(std::atomic<bool> *sloc) 155 kmp_flag_native(volatile PtrType *p, PtrType c, std::atomic<bool> *sloc) 236 std::atomic<PtrType> *loc; /**< Pointer to flag location to wait on */ 241 kmp_flag_atomic(std::atomic<PtrType> *p) : kmp_flag<FlagType>(), loc(p) {} 242 kmp_flag_atomic(std::atomic<PtrType> *p, kmp_info_t *thr) 246 kmp_flag_atomic(std::atomic<PtrType> *p, PtrType c) 248 kmp_flag_atomic(std::atomic<PtrType> *p, PtrType c, std::atomic<bool> *sloc) 251 std::atomic<PtrType> *get() { return loc; } 255 void set(std::atomic<PtrType> *new_loc) { loc = new_loc; } [all …]
|
| H A D | kmp_global.cpp | 66 std::atomic<kmp_int32> __kmp_team_counter = 0; 67 std::atomic<kmp_int32> __kmp_task_counter = 0; 393 std::atomic<int> __kmp_debug_count = 462 std::atomic<int> __kmp_thread_pool_active_nth = 0; 575 std::atomic<kmp_int32> __kmp_tdg_task_id = 0;
|
| H A D | kmp_lock.h | 128 std::atomic<kmp_int32> poll; 130 std::atomic<kmp_int32> poll; 251 std::atomic<bool> initialized; 254 std::atomic<unsigned> 256 std::atomic<unsigned> 258 std::atomic<int> owner_id; // (gtid+1) of owning thread, 0 if unlocked 259 std::atomic<int> depth_locked; // depth locked, for nested locks only 452 std::atomic<std::atomic<kmp_uint64> *> polls; 453 std::atomic<kmp_uint64> mask; // is 2**num_polls-1 for mod op 455 std::atomic<kmp_uint64> *old_polls; // will deallocate old_polls [all …]
|
| /src/contrib/llvm-project/libcxx/include/__stop_token/ |
| H A D | atomic_unique_lock.h | 34 std::atomic<_State>& __state_; 38 _LIBCPP_HIDE_FROM_ABI explicit __atomic_unique_lock(std::atomic<_State>& __state) noexcept in __atomic_unique_lock() 44 …_LIBCPP_HIDE_FROM_ABI __atomic_unique_lock(std::atomic<_State>& __state, _Pred&& __give_up_locking… in __atomic_unique_lock() 51 std::atomic<_State>& __state, in __atomic_unique_lock()
|
| H A D | stop_state.h | 36 atomic<bool> __completed_ = false; 54 atomic<uint32_t> __state_ = 0; 59 atomic<uint32_t> __ref_count_ = 0; 227 _LIBCPP_HIDE_FROM_ABI static atomic<uint32_t>& __get_atomic_ref_count(__stop_state& __state) {
|
| /src/crypto/openssl/doc/man3/ |
| H A D | CRYPTO_THREAD_run_once.pod | 94 result of the operation in I<*ret>. I<lock> will be locked, unless atomic 97 be the only way that the variable is modified. If atomic operations are not 103 result of the operation in I<*ret>. I<lock> will be locked, unless atomic 106 be the only way that the variable is modified. If atomic operations are not 111 CRYPTO_atomic_and() performs an atomic bitwise and of I<op> and I<*val> and stores 113 I<*ret>. I<lock> will be locked, unless atomic operations are supported on the 116 be the only way that the variable is modified. If atomic operations are not 121 CRYPTO_atomic_or() performs an atomic bitwise or of I<op> and I<*val> and stores 123 I<*ret>. I<lock> will be locked, unless atomic operations are supported on the 126 be the only way that the variable is modified. If atomic operations are not [all …]
|
| /src/contrib/llvm-project/llvm/include/llvm/IR/ |
| H A D | IntrinsicsBPF.td | 44 llvm_i8_ty, // atomic order 54 ImmArg <ArgIndex<2>>, // atomic order 64 llvm_i8_ty, // atomic order 74 ImmArg <ArgIndex<3>>, // atomic order
|
| /src/sys/dev/cxgbe/ |
| H A D | offload.h | 38 #define INIT_ULPTX_WRH(w, wrlen, atomic, tid) do { \ argument 39 (w)->wr_hi = htonl(V_FW_WR_OP(FW_ULPTX_WR) | V_FW_WR_ATOMIC(atomic)); \ 45 #define INIT_ULPTX_WR(w, wrlen, atomic, tid) \ argument 46 INIT_ULPTX_WRH(&((w)->wr), wrlen, atomic, tid)
|
| /src/cddl/lib/libspl/ |
| H A D | Makefile | 37 atomic.c \ 61 CFLAGS.atomic.c+= -Wno-error=atomic-alignment
|
| /src/contrib/libcxxrt/ |
| H A D | guard.cc | 128 atomic<GuardWord> val; 205 atomic<uint32_t> init_word; 210 atomic<uint32_t> lock_word;
|
| /src/contrib/llvm-project/llvm/include/llvm/Target/GlobalISel/ |
| H A D | SelectionDAGCompat.td | 26 // SelectionDAG has separate nodes for atomic and non-atomic memory operations 211 // G_LOAD handles both atomic and non-atomic loads where as SelectionDAG had 212 // separate nodes for them. This GINodeEquiv maps the non-atomic loads to 213 // G_LOAD with a non-atomic MachineMemOperand. 228 // G_STORE handles both atomic and non-atomic stores where as SelectionDAG had 229 // separate nodes for them. This GINodeEquiv maps the non-atomic stores to 230 // G_STORE with a non-atomic MachineMemOperand.
|