/linux/kernel/futex/ |
H A D | requeue.c | 109 old = atomic_read_acquire(&q->requeue_state); in futex_requeue_pi_prepare() 136 old = atomic_read_acquire(&q->requeue_state); in futex_requeue_pi_complete() 167 old = atomic_read_acquire(&q->requeue_state); in futex_requeue_pi_wakeup_sync()
|
/linux/include/linux/ |
H A D | call_once.h | 49 if (atomic_read_acquire(&once->state) == ONCE_COMPLETED) in call_once()
|
H A D | context_tracking_state.h | 75 return atomic_read_acquire(&ct->state) & CT_RCU_WATCHING_MASK; in ct_rcu_watching_cpu_acquire()
|
H A D | page-flags.h | 260 if (atomic_read_acquire(&page->_refcount) == u) in page_count_writable()
|
/linux/lib/ |
H A D | test_objpool.c | 161 if (atomic_read_acquire(&test->data.stop)) in ot_hrtimer_handler() 228 if (atomic_read_acquire(&test->data.stop)) in ot_thread_worker() 425 WARN_ON(!atomic_read_acquire(&test->data.stop)); in ot_fini_async_rcu() 515 stop = atomic_read_acquire(&test->data.stop); in ot_bulk_async()
|
/linux/kernel/power/ |
H A D | swap.c | 601 wait_event(d->go, atomic_read_acquire(&d->ready) || in crc32_threadfn() 648 wait_event(d->go, atomic_read_acquire(&d->ready) || in compress_threadfn() 832 atomic_read_acquire(&data[thr].stop)); in save_compressed_image() 871 wait_event(crc->done, atomic_read_acquire(&crc->stop)); in save_compressed_image() 1164 wait_event(d->go, atomic_read_acquire(&d->ready) || in decompress_threadfn() 1386 wait_event(crc->done, atomic_read_acquire(&crc->stop)); in load_compressed_image() 1441 atomic_read_acquire(&data[thr].stop)); in load_compressed_image() 1486 wait_event(crc->done, atomic_read_acquire(&crc->stop)); in load_compressed_image()
|
/linux/drivers/net/ethernet/google/gve/ |
H A D | gve_tx_dqo.c | 35 atomic_read_acquire(&tx->dqo_compl.free_tx_qpl_buf_cnt); in gve_has_free_tx_qpl_bufs() 87 s16 old_head = atomic_read_acquire(&tx->dqo_compl.free_tx_qpl_buf_head); in gve_free_tx_qpl_bufs() 109 if (atomic_read_acquire(&tx->dqo_compl.free_pending_packets) != -1) in gve_has_pending_packet() 160 s16 old_head = atomic_read_acquire(&tx->dqo_compl.free_pending_packets); in gve_free_pending_packet() 480 tx->dqo_tx.head = atomic_read_acquire(&tx->dqo_compl.hw_tx_head); in gve_has_tx_slots_available() 1027 atomic_read_acquire(&tx->dqo_tx.xsk_reorder_queue_tail); in gve_xsk_reorder_queue_head()
|
/linux/kernel/locking/ |
H A D | percpu-rwsem.c | 73 if (likely(!atomic_read_acquire(&sem->block))) in __percpu_down_read_trylock()
|
H A D | rwbase_rt.c | 229 if (!atomic_read_acquire(&rwb->readers)) { in __rwbase_write_trylock()
|
/linux/kernel/rcu/ |
H A D | refscale.c | 915 while (atomic_read_acquire(&n_started)) in ref_scale_reader() 925 while (atomic_read_acquire(&n_warmedup)) in ref_scale_reader() 941 while (atomic_read_acquire(&n_cooleddown)) in ref_scale_reader()
|
/linux/tools/memory-model/Documentation/ |
H A D | herd-representation.txt | 40 | atomic_read_acquire | |
|
H A D | glossary.txt | 32 but atomic_read_acquire() and atomic_xchg_acquire() also include
|
H A D | ordering.txt | 320 Acquire operations include smp_load_acquire(), atomic_read_acquire(), 339 smp_load_acquire(), but also atomic_read_acquire(), and
|
/linux/drivers/gpu/drm/vmwgfx/ |
H A D | vmwgfx_irq.c | 130 u32 last_read_seqno = atomic_read_acquire(&dev_priv->last_read_seqno); in vmw_seqno_passed()
|
H A D | vmwgfx_fence.c | 497 arg->passed_seqno = atomic_read_acquire(&dev_priv->last_read_seqno); in vmw_fence_obj_signaled_ioctl()
|
/linux/kernel/kcsan/ |
H A D | kcsan_test.c | 592 KCSAN_EXPECT_READ_BARRIER(atomic_read_acquire(&dummy), false); in test_barrier_nothreads() 637 KCSAN_EXPECT_WRITE_BARRIER(atomic_read_acquire(&dummy), false); in test_barrier_nothreads() 682 KCSAN_EXPECT_RW_BARRIER(atomic_read_acquire(&dummy), false); in test_barrier_nothreads()
|
/linux/drivers/net/wireguard/ |
H A D | send.c | 270 (state = atomic_read_acquire(&PACKET_CB(first)->state)) != in wg_packet_tx_worker()
|
/linux/kernel/bpf/ |
H A D | rqspinlock.c | 94 if (!(atomic_read_acquire(&lock->val) & (mask))) in is_lock_released()
|
/linux/tools/memory-model/ |
H A D | linux-kernel.def | 63 atomic_read_acquire(X) smp_load_acquire(X)
|
/linux/kernel/ |
H A D | scftorture.c | 500 while (atomic_read_acquire(&n_started)) { in scftorture_invoker()
|
/linux/Documentation/ |
H A D | atomic_t.txt | 17 atomic_read_acquire(), atomic_set_release()
|
/linux/kernel/time/ |
H A D | timer_migration.c | 1261 curstate.state = atomic_read_acquire(&group->migr_state); in tmigr_inactive_up()
|
/linux/fs/ |
H A D | file.c | 1164 * atomic_read_acquire() pairs with atomic_dec_and_test() in in __fget_light() 1167 if (likely(atomic_read_acquire(&files->count) == 1)) { in __fget_light()
|
/linux/include/linux/atomic/ |
H A D | atomic-instrumented.h | 37 * atomic_read_acquire() - atomic load with acquire ordering 47 atomic_read_acquire(const atomic_t *v) in atomic_read_acquire() function
|
/linux/kernel/sched/ |
H A D | rt.c | 2137 next = atomic_read_acquire(&rd->rto_loop_next); in rto_next_cpu()
|