| /linux/drivers/scsi/elx/efct/ |
| H A D | efct_scsi.c | 436 atomic_sub_return(1, &xport->io_pending_count); in efct_scsi_dispatch_pending() 469 atomic_sub_return(1, &xport->io_pending_recursing); in efct_scsi_check_pending() 477 atomic_sub_return(1, &xport->io_pending_recursing); in efct_scsi_check_pending() 497 atomic_sub_return(1, &xport->io_pending_count); in efct_scsi_check_pending() 513 atomic_sub_return(1, &xport->io_pending_recursing); in efct_scsi_check_pending()
|
| H A D | efct_io.c | 166 atomic_sub_return(1, &efct->xport->io_active_count); in efct_io_pool_io_free()
|
| H A D | efct_lio.c | 336 atomic_sub_return(1, &efct->tgt_efct.ios_in_use); in efct_lio_release_cmd() 1310 ini_count = atomic_sub_return(1, &efct->tgt_efct.initiator_count); in efct_scsi_del_initiator()
|
| /linux/drivers/crypto/virtio/ |
| H A D | virtio_crypto_mgr.c | 116 if (atomic_sub_return(1, &vcrypto_dev->ref_count) == 0) in virtcrypto_dev_put()
|
| /linux/arch/arm64/include/asm/ |
| H A D | atomic.h | 53 ATOMIC_FETCH_OPS(atomic_sub_return)
|
| /linux/include/linux/ |
| H A D | page_ref.h | 144 int ret = atomic_sub_return(nr, &folio->_refcount); in folio_ref_sub_return()
|
| /linux/io_uring/ |
| H A D | waitid.c | 169 if (!atomic_sub_return(1, &iw->refs)) in io_waitid_drop_issue_ref()
|
| H A D | poll.c | 307 } while (atomic_sub_return(v, &req->poll_refs) & IO_POLL_REF_MASK); in io_poll_check_events()
|
| /linux/drivers/crypto/intel/qat/qat_common/ |
| H A D | adf_dev_mgr.c | 411 if (atomic_sub_return(1, &accel_dev->ref_count) == 0) in adf_dev_put()
|
| /linux/tools/memory-model/ |
| H A D | linux-kernel.def | 108 atomic_sub_return(V,X) __atomic_op_return{MB}(X,-,V)
|
| /linux/drivers/tty/ |
| H A D | tty_buffer.c | 203 WARN_ON(atomic_sub_return(b->size, &buf->mem_used) < 0); in tty_buffer_free()
|
| /linux/drivers/gpu/drm/i915/gt/ |
| H A D | gen6_ppgtt.c | 93 if (!atomic_sub_return(count, &pt->used)) in gen6_ppgtt_clear_range()
|
| /linux/drivers/s390/cio/ |
| H A D | qdio_main.c | 467 if (atomic_sub_return(count, &q->nr_buf_used) == 0) in get_inbound_buffer_frontier() 479 if (atomic_sub_return(count, &q->nr_buf_used) == 0) in get_inbound_buffer_frontier()
|
| /linux/net/9p/ |
| H A D | trans_rdma.c | 437 if ((atomic_sub_return(1, &rdma->excess_rc) >= 0)) { in rdma_request()
|
| /linux/drivers/net/wwan/t7xx/ |
| H A D | t7xx_hif_dpmaif_tx.c | 311 if (atomic_sub_return(send_cnt, &txq->tx_budget) <= (MAX_SKB_FRAGS + 2)) in t7xx_dpmaif_add_skb_to_ring()
|
| /linux/drivers/w1/ |
| H A D | w1_netlink.c | 64 if (atomic_sub_return(1, &block->refcnt) == 0) { in w1_unref_block()
|
| H A D | w1.c | 782 refcnt = atomic_sub_return(1, &sl->refcnt); in w1_unref_slave()
|
| /linux/net/vmw_vsock/ |
| H A D | virtio_transport.c | 295 new_cnt = atomic_sub_return(cnt, &vsock->queued_replies); in virtio_transport_cancel_pkt()
|
| /linux/rust/helpers/ |
| H A D | atomic.c | 106 return atomic_sub_return(i, v); in rust_helper_atomic_sub_return()
|
| /linux/drivers/vhost/ |
| H A D | vsock.c | 318 new_cnt = atomic_sub_return(cnt, &vsock->queued_replies); in vhost_transport_cancel_pkt()
|
| /linux/net/sunrpc/xprtrdma/ |
| H A D | svc_rdma_sendto.c | 367 if (atomic_sub_return(sqecount, &rdma->sc_sq_avail) < 0) { in svc_rdma_post_send()
|
| H A D | svc_rdma_rw.c | 388 if (atomic_sub_return(cc->cc_sqecount, in svc_rdma_post_chunk_ctxt()
|
| /linux/drivers/net/ethernet/huawei/hinic/ |
| H A D | hinic_hw_wq.c | 750 if (atomic_sub_return(num_wqebbs, &wq->delta) <= 0) { in hinic_get_wqe()
|
| /linux/net/rds/ |
| H A D | ib_send.c | 231 if ((atomic_sub_return(nr, &ic->i_signaled_sends) == 0) && in rds_ib_sub_signaled()
|
| /linux/drivers/xen/ |
| H A D | gntdev.c | 440 live_grants = atomic_sub_return(successful_unmaps, &map->live_grants); in __unmap_grant_pages_done()
|