| /linux/include/trace/events/ |
| H A D | scmi.h | 39 bool poll, int inflight), 40 TP_ARGS(transfer_id, msg_id, protocol_id, seq, poll, inflight), 48 __field(int, inflight) 57 __entry->inflight = inflight; 62 __entry->transfer_id, __entry->poll, __entry->inflight) 95 int status, int inflight), 96 TP_ARGS(transfer_id, msg_id, protocol_id, seq, status, inflight), 104 __field(int, inflight) 113 __entry->inflight = inflight; 118 __entry->transfer_id, __entry->status, __entry->inflight)
|
| H A D | page_pool.h | 17 s32 inflight, u32 hold, u32 release), 19 TP_ARGS(pool, inflight, hold, release), 23 __field(s32, inflight) 31 __entry->inflight = inflight; 38 __entry->pool, __entry->inflight, __entry->hold,
|
| H A D | wbt.h | 132 int step, unsigned int inflight), 134 TP_ARGS(bdi, status, step, inflight), 140 __field(unsigned int, inflight) 148 __entry->inflight = inflight; 152 __entry->status, __entry->step, __entry->inflight)
|
| /linux/drivers/firmware/arm_scmi/transports/ |
| H A D | smc.c | 68 atomic_t inflight; member 100 atomic_set(&scmi_info->inflight, INFLIGHT_NONE); in smc_channel_lock_init() 105 static bool smc_xfer_inflight(struct scmi_xfer *xfer, atomic_t *inflight) in smc_xfer_inflight() argument 109 ret = atomic_cmpxchg(inflight, INFLIGHT_NONE, xfer->hdr.seq); in smc_xfer_inflight() 119 spin_until_cond(smc_xfer_inflight(xfer, &scmi_info->inflight)); in smc_channel_lock_acquire() 127 atomic_set(&scmi_info->inflight, INFLIGHT_NONE); in smc_channel_lock_release()
|
| /linux/block/ |
| H A D | genhd.c | 129 unsigned int inflight[2], bool mq_driver) in bdev_count_inflight_rw() 136 blk_mq_in_driver_rw(part, inflight); in bdev_count_inflight_rw() 150 inflight[READ] = read > 0 ? read : 0; in bdev_count_inflight_rw() 151 inflight[WRITE] = write > 0 ? write : 0; in bdev_count_inflight_rw() 165 unsigned int inflight[2] = {0}; in bdev_count_inflight() local 167 bdev_count_inflight_rw(part, inflight, false); in bdev_count_inflight() 169 return inflight[READ] + inflight[WRITE]; in bdev_count_inflight() 1067 unsigned int inflight; in part_stat_show() local 1069 inflight = bdev_count_inflight(bdev); in part_stat_show() 1070 if (inflight) { in part_stat_show() [all …]
|
| H A D | blk-wbt.c | 201 int inflight, limit; in wbt_rqw_done() local 203 inflight = atomic_dec_return(&rqw->inflight); in wbt_rqw_done() 221 if (inflight && inflight >= limit) in wbt_rqw_done() 225 int diff = limit - inflight; in wbt_rqw_done() 227 if (!inflight || diff >= rwb->wb_background / 2) in wbt_rqw_done() 295 ret += atomic_read(&rwb->rq_wait[i].inflight); in wbt_inflight() 429 unsigned int inflight = wbt_inflight(rwb); in wb_timer_fn() local 437 trace_wbt_timer(rwb->rqos.disk->bdi, status, rqd->scale_step, inflight); in wb_timer_fn() 479 if (rqd->scale_step || inflight) in wb_timer_fn() 877 atomic_read(&rwb->rq_wait[i].inflight)); in wbt_inflight_show()
|
| H A D | blk-iolatency.c | 276 atomic_dec(&rqw->inflight); in iolat_cleanup_cb() 305 atomic_inc(&rqw->inflight); in __blkcg_iolatency_throttle() 591 int inflight = 0; in blkcg_iolatency_done_bio() local 613 inflight = atomic_dec_return(&rqw->inflight); in blkcg_iolatency_done_bio() 614 WARN_ON_ONCE(inflight < 0); in blkcg_iolatency_done_bio()
|
| H A D | blk-rq-qos.h | 24 atomic_t inflight; member 84 atomic_set(&rq_wait->inflight, 0); in rq_wait_init()
|
| /linux/drivers/crypto/chelsio/ |
| H A D | chcr_core.c | 56 if (atomic_read(&dev->inflight)) { in detach_work_fn() 60 atomic_read(&dev->inflight)); in detach_work_fn() 65 atomic_read(&dev->inflight)); in detach_work_fn() 103 atomic_set(&dev->inflight, 0); in chcr_dev_add() 122 atomic_set(&dev->inflight, 0); in chcr_dev_init() 232 if (atomic_read(&dev->inflight) != 0) { in chcr_detach_device()
|
| /linux/net/core/ |
| H A D | page_pool.c | 688 s32 inflight; in page_pool_inflight() local 690 inflight = _distance(hold_cnt, release_cnt); in page_pool_inflight() 693 trace_page_pool_release(pool, inflight, hold_cnt, release_cnt); in page_pool_inflight() 694 WARN(inflight < 0, "Negative(%d) inflight packet-pages", in page_pool_inflight() 695 inflight); in page_pool_inflight() 697 inflight = max(0, inflight); in page_pool_inflight() 700 return inflight; in page_pool_inflight() 1189 int inflight; in page_pool_release() local 1192 inflight = page_pool_inflight(pool, true); in page_pool_release() 1196 if (!inflight) in page_pool_release() [all …]
|
| H A D | page_pool_user.c | 219 size_t inflight, refsz; in page_pool_nl_fill() local 240 inflight = page_pool_inflight(pool, false); in page_pool_nl_fill() 242 if (nla_put_uint(rsp, NETDEV_A_PAGE_POOL_INFLIGHT, inflight) || in page_pool_nl_fill() 244 inflight * refsz)) in page_pool_nl_fill()
|
| /linux/tools/net/ynl/ynltool/ |
| H A D | page-pool.c | 52 if (pp->_present.inflight) in count_pool() 53 s->live[l].refs += pp->inflight; in count_pool() 227 if (pp->_present.inflight) in print_json_pool_list() 228 jsonw_uint_field(json_wtr, "refs", pp->inflight); in print_json_pool_list() 276 if (pp->_present.inflight || pp->_present.inflight_mem) { in print_plain_pool_list() 278 if (pp->_present.inflight) in print_plain_pool_list() 279 printf(" %llu pages", pp->inflight); in print_plain_pool_list()
|
| /linux/tools/testing/selftests/net/af_unix/ |
| H A D | scm_rights.c | 270 int inflight, int receiver) in __send_fd() argument 275 self->fd[inflight * 2], in __send_fd() 276 self->fd[inflight * 2], in __send_fd() 314 #define send_fd(inflight, receiver) \ argument 315 __send_fd(_metadata, self, variant, inflight, receiver)
|
| /linux/net/atm/ |
| H A D | pppoatm.c | 64 atomic_t inflight; member 139 atomic_dec(&pvcc->inflight); in pppoatm_pop() 244 atomic_inc_not_zero(&pvcc->inflight)) in pppoatm_may_send() 274 atomic_inc_not_zero(&pvcc->inflight)) in pppoatm_may_send() 406 atomic_set(&pvcc->inflight, NONE_INFLIGHT); in pppoatm_assign_vcc()
|
| /linux/net/unix/ |
| H A D | garbage.c | 214 struct unix_sock *inflight = unix_get_socket(fpl->fp[j++]); in unix_add_edges() local 217 if (!inflight) in unix_add_edges() 221 edge->predecessor = inflight; in unix_add_edges() 233 fpl->inflight = true; in unix_add_edges() 263 fpl->inflight = false; in unix_del_edges() 314 if (fpl->inflight) in unix_destroy_fpl()
|
| /linux/drivers/gpu/drm/i915/gt/ |
| H A D | intel_context_types.h | 90 struct intel_engine_cs *inflight; member 94 __intel_context_inflight(READ_ONCE((ce)->inflight)) 96 __intel_context_inflight_count(READ_ONCE((ce)->inflight))
|
| H A D | intel_execlists_submission.c | 535 old = ce->inflight; in execlists_schedule_in() 538 WRITE_ONCE(ce->inflight, ptr_inc(old)); in execlists_schedule_in() 599 GEM_BUG_ON(ce->inflight != engine); in __execlists_schedule_out() 644 WRITE_ONCE(ce->inflight, NULL); in __execlists_schedule_out() 654 GEM_BUG_ON(!ce->inflight); in execlists_schedule_out() 655 ce->inflight = ptr_dec(ce->inflight); in execlists_schedule_out() 656 if (!__intel_context_inflight_count(ce->inflight)) in execlists_schedule_out() 997 const struct intel_engine_cs *inflight; in virtual_matches() local 1014 inflight = intel_context_inflight(&ve->context); in virtual_matches() 1015 if (inflight && inflight != engine) in virtual_matches() [all …]
|
| H A D | intel_timeline.c | 424 unsigned long count, ready, inflight; in intel_gt_show_timelines() local 441 inflight = 0; in intel_gt_show_timelines() 450 inflight++; in intel_gt_show_timelines() 455 count, ready, inflight); in intel_gt_show_timelines()
|
| /linux/drivers/infiniband/ulp/rtrs/ |
| H A D | rtrs-clt-stats.c | 102 atomic_read(&stats->inflight), sum.failover_cnt); in rtrs_clt_stats_rdma_to_str() 158 atomic_set(&s->inflight, 0); in rtrs_clt_reset_all_stats() 182 atomic_inc(&stats->inflight); in rtrs_clt_update_all_stats()
|
| H A D | README | 100 corresponding path is disconnected, all the inflight IO are failed over to a 131 inflight IO and for the error code. 149 inflight IO and for the error code. The new rkey is sent back using 171 outstanding inflight IO and the error code. 192 outstanding inflight IO and the error code. The new rkey is sent back using
|
| /linux/net/ipv4/ |
| H A D | tcp_bbr.c | 415 u32 inflight; in bbr_inflight() local 417 inflight = bbr_bdp(sk, bw, gain); in bbr_inflight() 418 inflight = bbr_quantization_budget(sk, inflight); in bbr_inflight() 420 return inflight; in bbr_inflight() 562 u32 inflight, bw; in bbr_is_next_cycle_phase() local 570 inflight = bbr_packets_in_net_at_edt(sk, rs->prior_in_flight); in bbr_is_next_cycle_phase() 581 inflight >= bbr_inflight(sk, bw, bbr->pacing_gain)); in bbr_is_next_cycle_phase() 588 inflight <= bbr_inflight(sk, bw, BBR_UNIT); in bbr_is_next_cycle_phase()
|
| /linux/drivers/vhost/ |
| H A D | scsi.c | 144 struct vhost_scsi_inflight *inflight; member 265 struct vhost_scsi_inflight *inflight; member 299 struct vhost_scsi_inflight *inflight; in vhost_scsi_done_inflight() local 301 inflight = container_of(kref, struct vhost_scsi_inflight, kref); in vhost_scsi_done_inflight() 302 complete(&inflight->comp); in vhost_scsi_done_inflight() 335 struct vhost_scsi_inflight *inflight; in vhost_scsi_get_inflight() local 339 inflight = &svq->inflights[svq->inflight_idx]; in vhost_scsi_get_inflight() 340 kref_get(&inflight->kref); in vhost_scsi_get_inflight() 342 return inflight; in vhost_scsi_get_inflight() 345 static void vhost_scsi_put_inflight(struct vhost_scsi_inflight *inflight) in vhost_scsi_put_inflight() argument [all …]
|
| /linux/io_uring/ |
| H A D | cancel.c | 570 return percpu_counter_sum(&tctx->inflight); in tctx_inflight() 583 s64 inflight; in io_uring_cancel_generic() local 602 inflight = tctx_inflight(tctx, false); in io_uring_cancel_generic() 603 if (!inflight) in io_uring_cancel_generic() 644 if (inflight == tctx_inflight(tctx, !cancel_all)) in io_uring_cancel_generic()
|
| H A D | tctx.c | 66 percpu_counter_destroy(&tctx->inflight); in __io_uring_free() 87 ret = percpu_counter_init(&tctx->inflight, 0, GFP_KERNEL); in io_uring_alloc_task_context() 96 percpu_counter_destroy(&tctx->inflight); in io_uring_alloc_task_context()
|
| /linux/drivers/crypto/cavium/cpt/ |
| H A D | cpt_hw_types.h | 443 u64 inflight:8; member 445 u64 inflight:8;
|