Home
last modified time | relevance | path

Searched refs:inflight (Results 1 – 25 of 60) sorted by relevance

123

/linux/include/trace/events/
H A Dscmi.h39 bool poll, int inflight),
40 TP_ARGS(transfer_id, msg_id, protocol_id, seq, poll, inflight),
48 __field(int, inflight)
57 __entry->inflight = inflight;
62 __entry->transfer_id, __entry->poll, __entry->inflight)
95 int status, int inflight),
96 TP_ARGS(transfer_id, msg_id, protocol_id, seq, status, inflight),
104 __field(int, inflight)
113 __entry->inflight = inflight;
118 __entry->transfer_id, __entry->status, __entry->inflight)
H A Dpage_pool.h17 s32 inflight, u32 hold, u32 release),
19 TP_ARGS(pool, inflight, hold, release),
23 __field(s32, inflight)
31 __entry->inflight = inflight;
38 __entry->pool, __entry->inflight, __entry->hold,
H A Dwbt.h132 int step, unsigned int inflight),
134 TP_ARGS(bdi, status, step, inflight),
140 __field(unsigned int, inflight)
148 __entry->inflight = inflight;
152 __entry->status, __entry->step, __entry->inflight)
/linux/drivers/firmware/arm_scmi/transports/
H A Dsmc.c68 atomic_t inflight; member
100 atomic_set(&scmi_info->inflight, INFLIGHT_NONE); in smc_channel_lock_init()
105 static bool smc_xfer_inflight(struct scmi_xfer *xfer, atomic_t *inflight) in smc_xfer_inflight() argument
109 ret = atomic_cmpxchg(inflight, INFLIGHT_NONE, xfer->hdr.seq); in smc_xfer_inflight()
119 spin_until_cond(smc_xfer_inflight(xfer, &scmi_info->inflight)); in smc_channel_lock_acquire()
127 atomic_set(&scmi_info->inflight, INFLIGHT_NONE); in smc_channel_lock_release()
/linux/block/
H A Dgenhd.c129 unsigned int inflight[2], bool mq_driver) in bdev_count_inflight_rw()
136 blk_mq_in_driver_rw(part, inflight); in bdev_count_inflight_rw()
150 inflight[READ] = read > 0 ? read : 0; in bdev_count_inflight_rw()
151 inflight[WRITE] = write > 0 ? write : 0; in bdev_count_inflight_rw()
165 unsigned int inflight[2] = {0}; in bdev_count_inflight() local
167 bdev_count_inflight_rw(part, inflight, false); in bdev_count_inflight()
169 return inflight[READ] + inflight[WRITE]; in bdev_count_inflight()
1067 unsigned int inflight; in part_stat_show() local
1069 inflight = bdev_count_inflight(bdev); in part_stat_show()
1070 if (inflight) { in part_stat_show()
[all …]
H A Dblk-wbt.c201 int inflight, limit; in wbt_rqw_done() local
203 inflight = atomic_dec_return(&rqw->inflight); in wbt_rqw_done()
221 if (inflight && inflight >= limit) in wbt_rqw_done()
225 int diff = limit - inflight; in wbt_rqw_done()
227 if (!inflight || diff >= rwb->wb_background / 2) in wbt_rqw_done()
295 ret += atomic_read(&rwb->rq_wait[i].inflight); in wbt_inflight()
429 unsigned int inflight = wbt_inflight(rwb); in wb_timer_fn() local
437 trace_wbt_timer(rwb->rqos.disk->bdi, status, rqd->scale_step, inflight); in wb_timer_fn()
479 if (rqd->scale_step || inflight) in wb_timer_fn()
877 atomic_read(&rwb->rq_wait[i].inflight)); in wbt_inflight_show()
H A Dblk-iolatency.c276 atomic_dec(&rqw->inflight); in iolat_cleanup_cb()
305 atomic_inc(&rqw->inflight); in __blkcg_iolatency_throttle()
591 int inflight = 0; in blkcg_iolatency_done_bio() local
613 inflight = atomic_dec_return(&rqw->inflight); in blkcg_iolatency_done_bio()
614 WARN_ON_ONCE(inflight < 0); in blkcg_iolatency_done_bio()
H A Dblk-rq-qos.h24 atomic_t inflight; member
84 atomic_set(&rq_wait->inflight, 0); in rq_wait_init()
/linux/drivers/crypto/chelsio/
H A Dchcr_core.c56 if (atomic_read(&dev->inflight)) { in detach_work_fn()
60 atomic_read(&dev->inflight)); in detach_work_fn()
65 atomic_read(&dev->inflight)); in detach_work_fn()
103 atomic_set(&dev->inflight, 0); in chcr_dev_add()
122 atomic_set(&dev->inflight, 0); in chcr_dev_init()
232 if (atomic_read(&dev->inflight) != 0) { in chcr_detach_device()
/linux/net/core/
H A Dpage_pool.c688 s32 inflight; in page_pool_inflight() local
690 inflight = _distance(hold_cnt, release_cnt); in page_pool_inflight()
693 trace_page_pool_release(pool, inflight, hold_cnt, release_cnt); in page_pool_inflight()
694 WARN(inflight < 0, "Negative(%d) inflight packet-pages", in page_pool_inflight()
695 inflight); in page_pool_inflight()
697 inflight = max(0, inflight); in page_pool_inflight()
700 return inflight; in page_pool_inflight()
1189 int inflight; in page_pool_release() local
1192 inflight = page_pool_inflight(pool, true); in page_pool_release()
1196 if (!inflight) in page_pool_release()
[all …]
H A Dpage_pool_user.c219 size_t inflight, refsz; in page_pool_nl_fill() local
240 inflight = page_pool_inflight(pool, false); in page_pool_nl_fill()
242 if (nla_put_uint(rsp, NETDEV_A_PAGE_POOL_INFLIGHT, inflight) || in page_pool_nl_fill()
244 inflight * refsz)) in page_pool_nl_fill()
/linux/tools/net/ynl/ynltool/
H A Dpage-pool.c52 if (pp->_present.inflight) in count_pool()
53 s->live[l].refs += pp->inflight; in count_pool()
227 if (pp->_present.inflight) in print_json_pool_list()
228 jsonw_uint_field(json_wtr, "refs", pp->inflight); in print_json_pool_list()
276 if (pp->_present.inflight || pp->_present.inflight_mem) { in print_plain_pool_list()
278 if (pp->_present.inflight) in print_plain_pool_list()
279 printf(" %llu pages", pp->inflight); in print_plain_pool_list()
/linux/tools/testing/selftests/net/af_unix/
H A Dscm_rights.c270 int inflight, int receiver) in __send_fd() argument
275 self->fd[inflight * 2], in __send_fd()
276 self->fd[inflight * 2], in __send_fd()
314 #define send_fd(inflight, receiver) \ argument
315 __send_fd(_metadata, self, variant, inflight, receiver)
/linux/net/atm/
H A Dpppoatm.c64 atomic_t inflight; member
139 atomic_dec(&pvcc->inflight); in pppoatm_pop()
244 atomic_inc_not_zero(&pvcc->inflight)) in pppoatm_may_send()
274 atomic_inc_not_zero(&pvcc->inflight)) in pppoatm_may_send()
406 atomic_set(&pvcc->inflight, NONE_INFLIGHT); in pppoatm_assign_vcc()
/linux/net/unix/
H A Dgarbage.c214 struct unix_sock *inflight = unix_get_socket(fpl->fp[j++]); in unix_add_edges() local
217 if (!inflight) in unix_add_edges()
221 edge->predecessor = inflight; in unix_add_edges()
233 fpl->inflight = true; in unix_add_edges()
263 fpl->inflight = false; in unix_del_edges()
314 if (fpl->inflight) in unix_destroy_fpl()
/linux/drivers/gpu/drm/i915/gt/
H A Dintel_context_types.h90 struct intel_engine_cs *inflight; member
94 __intel_context_inflight(READ_ONCE((ce)->inflight))
96 __intel_context_inflight_count(READ_ONCE((ce)->inflight))
H A Dintel_execlists_submission.c535 old = ce->inflight; in execlists_schedule_in()
538 WRITE_ONCE(ce->inflight, ptr_inc(old)); in execlists_schedule_in()
599 GEM_BUG_ON(ce->inflight != engine); in __execlists_schedule_out()
644 WRITE_ONCE(ce->inflight, NULL); in __execlists_schedule_out()
654 GEM_BUG_ON(!ce->inflight); in execlists_schedule_out()
655 ce->inflight = ptr_dec(ce->inflight); in execlists_schedule_out()
656 if (!__intel_context_inflight_count(ce->inflight)) in execlists_schedule_out()
997 const struct intel_engine_cs *inflight; in virtual_matches() local
1014 inflight = intel_context_inflight(&ve->context); in virtual_matches()
1015 if (inflight && inflight != engine) in virtual_matches()
[all …]
H A Dintel_timeline.c424 unsigned long count, ready, inflight; in intel_gt_show_timelines() local
441 inflight = 0; in intel_gt_show_timelines()
450 inflight++; in intel_gt_show_timelines()
455 count, ready, inflight); in intel_gt_show_timelines()
/linux/drivers/infiniband/ulp/rtrs/
H A Drtrs-clt-stats.c102 atomic_read(&stats->inflight), sum.failover_cnt); in rtrs_clt_stats_rdma_to_str()
158 atomic_set(&s->inflight, 0); in rtrs_clt_reset_all_stats()
182 atomic_inc(&stats->inflight); in rtrs_clt_update_all_stats()
H A DREADME100 corresponding path is disconnected, all the inflight IO are failed over to a
131 inflight IO and for the error code.
149 inflight IO and for the error code. The new rkey is sent back using
171 outstanding inflight IO and the error code.
192 outstanding inflight IO and the error code. The new rkey is sent back using
/linux/net/ipv4/
H A Dtcp_bbr.c415 u32 inflight; in bbr_inflight() local
417 inflight = bbr_bdp(sk, bw, gain); in bbr_inflight()
418 inflight = bbr_quantization_budget(sk, inflight); in bbr_inflight()
420 return inflight; in bbr_inflight()
562 u32 inflight, bw; in bbr_is_next_cycle_phase() local
570 inflight = bbr_packets_in_net_at_edt(sk, rs->prior_in_flight); in bbr_is_next_cycle_phase()
581 inflight >= bbr_inflight(sk, bw, bbr->pacing_gain)); in bbr_is_next_cycle_phase()
588 inflight <= bbr_inflight(sk, bw, BBR_UNIT); in bbr_is_next_cycle_phase()
/linux/drivers/vhost/
H A Dscsi.c144 struct vhost_scsi_inflight *inflight; member
265 struct vhost_scsi_inflight *inflight; member
299 struct vhost_scsi_inflight *inflight; in vhost_scsi_done_inflight() local
301 inflight = container_of(kref, struct vhost_scsi_inflight, kref); in vhost_scsi_done_inflight()
302 complete(&inflight->comp); in vhost_scsi_done_inflight()
335 struct vhost_scsi_inflight *inflight; in vhost_scsi_get_inflight() local
339 inflight = &svq->inflights[svq->inflight_idx]; in vhost_scsi_get_inflight()
340 kref_get(&inflight->kref); in vhost_scsi_get_inflight()
342 return inflight; in vhost_scsi_get_inflight()
345 static void vhost_scsi_put_inflight(struct vhost_scsi_inflight *inflight) in vhost_scsi_put_inflight() argument
[all …]
/linux/io_uring/
H A Dcancel.c570 return percpu_counter_sum(&tctx->inflight); in tctx_inflight()
583 s64 inflight; in io_uring_cancel_generic() local
602 inflight = tctx_inflight(tctx, false); in io_uring_cancel_generic()
603 if (!inflight) in io_uring_cancel_generic()
644 if (inflight == tctx_inflight(tctx, !cancel_all)) in io_uring_cancel_generic()
H A Dtctx.c66 percpu_counter_destroy(&tctx->inflight); in __io_uring_free()
87 ret = percpu_counter_init(&tctx->inflight, 0, GFP_KERNEL); in io_uring_alloc_task_context()
96 percpu_counter_destroy(&tctx->inflight); in io_uring_alloc_task_context()
/linux/drivers/crypto/cavium/cpt/
H A Dcpt_hw_types.h443 u64 inflight:8; member
445 u64 inflight:8;

123