| /linux/sound/core/seq/oss/ |
| H A D | seq_oss_readq.c | 48 q->qlen = 0; in snd_seq_oss_readq_new() 76 if (q->qlen) { in snd_seq_oss_readq_clear() 77 q->qlen = 0; in snd_seq_oss_readq_clear() 144 if (q->qlen >= q->maxlen - 1) in snd_seq_oss_readq_put_event() 149 q->qlen++; in snd_seq_oss_readq_put_event() 165 if (q->qlen == 0) in snd_seq_oss_readq_pick() 178 (q->qlen > 0 || q->head == q->tail), in snd_seq_oss_readq_wait() 189 if (q->qlen > 0) { in snd_seq_oss_readq_free() 191 q->qlen--; in snd_seq_oss_readq_free() 203 return q->qlen; in snd_seq_oss_readq_poll() [all …]
|
| /linux/net/sched/ |
| H A D | sch_sfq.c | 97 sfq_index qlen; /* number of skbs in skblist */ member 200 int qlen = slot->qlen; in sfq_link() local 202 p = qlen + SFQ_MAX_FLOWS; in sfq_link() 203 n = q->dep[qlen].next; in sfq_link() 208 q->dep[qlen].next = x; /* sfq_dep_head(q, p)->next = x */ in sfq_link() 228 d = q->slots[x].qlen--; in sfq_dec() 241 d = ++q->slots[x].qlen; in sfq_inc() 303 sch->q.qlen--; in sfq_drop() 346 sfq_index x, qlen; in sfq_enqueue() local 420 if (slot->qlen >= q->maxdepth) { in sfq_enqueue() [all …]
|
| H A D | sch_sfb.c | 40 u16 qlen; /* length of virtual queue */ member 132 if (b[hash].qlen < 0xFFFF) in increment_one_qlen() 133 b[hash].qlen++; in increment_one_qlen() 161 if (b[hash].qlen > 0) in decrement_one_qlen() 162 b[hash].qlen--; in decrement_one_qlen() 201 u32 qlen = 0, prob = 0, totalpm = 0; in sfb_compute_qlen() local 205 if (qlen < b->qlen) in sfb_compute_qlen() 206 qlen = b->qlen; in sfb_compute_qlen() 214 return qlen; in sfb_compute_qlen() 296 if (unlikely(sch->q.qlen >= q->limit)) { in sfb_enqueue() [all …]
|
| H A D | sch_choke.c | 126 --sch->q.qlen; in choke_drop_by_idx() 220 q->vars.qavg = red_calc_qavg(p, &q->vars, sch->q.qlen); in choke_enqueue() 267 if (sch->q.qlen < q->limit) { in choke_enqueue() 270 ++sch->q.qlen; in choke_enqueue() 297 --sch->q.qlen; in choke_dequeue() 380 unsigned int oqlen = sch->q.qlen, tail = 0; in choke_change() 395 --sch->q.qlen; in choke_change() 398 qdisc_tree_reduce_backlog(sch, oqlen - sch->q.qlen, dropped); in choke_change()
|
| H A D | sch_drr.c | 267 __u32 qlen = qdisc_qlen_sum(cl->qdisc); in drr_dump_class_stats() local 272 if (qlen) in drr_dump_class_stats() 277 gnet_stats_copy_queue(d, cl_q->cpu_qstats, &cl_q->qstats, qlen) < 0) in drr_dump_class_stats() 369 sch->q.qlen++; in drr_enqueue() 396 if (cl->qdisc->q.qlen == 0) in drr_dequeue() 402 sch->q.qlen--; in drr_dequeue() 437 if (cl->qdisc->q.qlen) in drr_reset_qdisc()
|
| /linux/sound/core/seq/ |
| H A D | seq_midi_event.c | 50 int qlen; member 141 dev->qlen = 0; in reset_encode() 193 dev->qlen = status_event[dev->type].qlen; in snd_midi_event_encode_byte() 195 if (dev->qlen > 0) { in snd_midi_event_encode_byte() 199 dev->qlen--; in snd_midi_event_encode_byte() 203 dev->qlen = status_event[dev->type].qlen - 1; in snd_midi_event_encode_byte() 207 if (dev->qlen == 0) { in snd_midi_event_encode_byte() 312 int qlen; in snd_midi_event_decode() local 323 qlen = status_event[type].qlen + 1; in snd_midi_event_decode() 328 qlen = status_event[type].qlen; in snd_midi_event_decode() [all …]
|
| /linux/include/net/ |
| H A D | sch_generic.h | 62 __u32 qlen; member 193 return !READ_ONCE(qdisc->q.qlen); in qdisc_is_empty() 539 return q->q.qlen; in qdisc_qlen() 544 __u32 qlen = q->qstats.qlen; in qdisc_qlen_sum() local 549 qlen += per_cpu_ptr(q->cpu_qstats, i)->qlen; in qdisc_qlen_sum() 551 qlen += q->q.qlen; in qdisc_qlen_sum() 554 return qlen; in qdisc_qlen_sum() 981 this_cpu_inc(sch->cpu_qstats->qlen); in qdisc_qstats_cpu_qlen_inc() 986 this_cpu_dec(sch->cpu_qstats->qlen); in qdisc_qstats_cpu_qlen_dec() 1026 __u32 qlen = qdisc_qlen_sum(sch); in qdisc_qstats_copy() local [all …]
|
| H A D | request_sock.h | 164 int qlen; /* # of pending (TCP_SYN_RECV) reqs */ member 182 atomic_t qlen; member 222 atomic_dec(&queue->qlen); in reqsk_queue_removed() 228 atomic_inc(&queue->qlen); in reqsk_queue_added() 233 return atomic_read(&queue->qlen); in reqsk_queue_len()
|
| /linux/drivers/md/ |
| H A D | dm-ps-queue-length.c | 40 atomic_t qlen; /* the number of in-flight I/Os */ member 101 DMEMIT("%d ", atomic_read(&pi->qlen)); in ql_status() 153 atomic_set(&pi->qlen, 0); in ql_add_path() 204 (atomic_read(&pi->qlen) < atomic_read(&best->qlen))) in ql_select_path() 207 if (!atomic_read(&best->qlen)) in ql_select_path() 228 atomic_inc(&pi->qlen); in ql_start_io() 238 atomic_dec(&pi->qlen); in ql_end_io()
|
| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | bpf_qdisc_fifo.c | 26 if (sch->q.qlen == sch->limit) in BPF_PROG() 35 sch->q.qlen++; in BPF_PROG() 72 sch->q.qlen--; in BPF_PROG() 93 bpf_for(i, 0, sch->q.qlen) { in BPF_PROG() 109 sch->q.qlen = 0; in BPF_PROG()
|
| H A D | bpf_qdisc_fq.c | 78 u32 qlen; member 248 flow->qlen = 0, in fq_new_flow() 317 if (sch->q.qlen >= sch->limit) in BPF_PROG() 344 if (flow->qlen >= q.flow_plimit) { in BPF_PROG() 381 flow->qlen++; in BPF_PROG() 384 sch->q.qlen++; in BPF_PROG() 537 flow->qlen--; in fq_dequeue_nonprio_flows() 595 if (!sch->q.qlen) in BPF_PROG() 612 sch->q.qlen--; in BPF_PROG() 698 sch->q.qlen = 0; in BPF_PROG()
|
| /linux/drivers/usb/gadget/legacy/ |
| H A D | gmidi.c | 53 static unsigned int qlen = 32; variable 54 module_param(qlen, uint, S_IRUGO); 55 MODULE_PARM_DESC(qlen, "USB read and write request queue length"); 157 midi_opts->qlen = qlen; in midi_bind()
|
| H A D | printer.c | 51 static unsigned qlen = 10; variable 52 module_param(qlen, uint, S_IRUGO|S_IWUSR); 53 MODULE_PARM_DESC(qlen, "The number of 8k buffers to use per endpoint"); 55 #define QLEN qlen
|
| H A D | zero.c | 66 .qlen = GZERO_QLEN, 263 module_param_named(qlen, gzero_options.qlen, uint, S_IRUGO|S_IWUSR); 264 MODULE_PARM_DESC(qlen, "depth of loopback queue"); 322 lb_opts->qlen = gzero_options.qlen; in zero_bind()
|
| /linux/net/core/ |
| H A D | gen_stats.c | 348 qstats->qlen += qcpu->qlen; in gnet_stats_add_queue_cpu() 363 qstats->qlen += q->qlen; in gnet_stats_add_queue() 389 struct gnet_stats_queue *q, __u32 qlen) in gnet_stats_copy_queue() argument 394 qstats.qlen = qlen; in gnet_stats_copy_queue() 398 d->tc_stats.qlen = qstats.qlen; in gnet_stats_copy_queue()
|
| /linux/include/trace/events/ |
| H A D | rcu.h | 514 TP_PROTO(const char *rcuname, struct rcu_head *rhp, long qlen), 516 TP_ARGS(rcuname, rhp, qlen), 522 __field(long, qlen) 529 __entry->qlen = qlen; 534 __entry->qlen) 572 TP_PROTO(const char *rcuname, long qlen, long blimit), 574 TP_ARGS(rcuname, qlen, blimit), 578 __field(long, qlen) 584 __entry->qlen = qlen; 589 __entry->rcuname, __entry->qlen, __entry->blimit)
|
| /linux/drivers/crypto/cavium/cpt/ |
| H A D | cptvf_main.c | 99 pqinfo->qlen = 0; in free_pending_queues() 103 static int alloc_pending_queues(struct pending_qinfo *pqinfo, u32 qlen, in alloc_pending_queues() argument 111 pqinfo->qlen = qlen; in alloc_pending_queues() 114 queue->head = kzalloc_objs(*queue->head, qlen); in alloc_pending_queues() 136 static int init_pending_queues(struct cpt_vf *cptvf, u32 qlen, u32 nr_queues) in init_pending_queues() argument 144 ret = alloc_pending_queues(&cptvf->pqinfo, qlen, nr_queues); in init_pending_queues() 203 u32 qlen) in alloc_command_queues() argument 213 cptvf->qsize = min(qlen, cqinfo->qchunksize) * in alloc_command_queues() 216 q_size = qlen * cqinfo->cmd_size; in alloc_command_queues() 277 static int init_command_queues(struct cpt_vf *cptvf, u32 qlen) in init_command_queues() argument [all …]
|
| /linux/drivers/usb/gadget/function/ |
| H A D | f_loopback.c | 34 unsigned qlen; member 321 for (i = 0; i < loop->qlen && result == 0; i++) { in alloc_requests() 439 loop->qlen = lb_opts->qlen; in loopback_alloc() 440 if (!loop->qlen) in loopback_alloc() 441 loop->qlen = 32; in loopback_alloc() 477 result = sprintf(page, "%d\n", opts->qlen); in f_lb_opts_qlen_show() 500 opts->qlen = num; in f_lb_opts_qlen_store() 507 CONFIGFS_ATTR(f_lb_opts_, qlen); 577 lb_opts->qlen = GZERO_QLEN; in loopback_alloc_instance()
|
| H A D | g_zero.h | 24 unsigned qlen; member 54 unsigned qlen; member
|
| /linux/drivers/nfc/st-nci/ |
| H A D | ndlc.c | 94 if (ndlc->send_q.qlen) in llt_ndlc_send_queue() 96 ndlc->send_q.qlen, ndlc->ack_pending_q.qlen); in llt_ndlc_send_queue() 98 while (ndlc->send_q.qlen) { in llt_ndlc_send_queue() 153 if (ndlc->rcv_q.qlen) in llt_ndlc_rcv_queue() 154 pr_debug("rcvQlen=%d\n", ndlc->rcv_q.qlen); in llt_ndlc_rcv_queue()
|
| /linux/drivers/crypto/marvell/octeontx/ |
| H A D | otx_cptvf_main.c | 88 queue->qlen = 0; in free_pending_queues() 93 static int alloc_pending_queues(struct otx_cpt_pending_qinfo *pqinfo, u32 qlen, in alloc_pending_queues() argument 103 queue->head = kzalloc_objs(*queue->head, qlen); in alloc_pending_queues() 112 queue->qlen = qlen; in alloc_pending_queues() 125 static int init_pending_queues(struct otx_cptvf *cptvf, u32 qlen, in init_pending_queues() argument 134 ret = alloc_pending_queues(&cptvf->pqinfo, qlen, num_queues); in init_pending_queues() 188 u32 qlen) in alloc_command_queues() argument 199 cptvf->qsize = min(qlen, cqinfo->qchunksize) * in alloc_command_queues() 202 q_size = qlen * OTX_CPT_INST_SIZE; in alloc_command_queues() 263 static int init_command_queues(struct otx_cptvf *cptvf, u32 qlen) in init_command_queues() argument [all …]
|
| /linux/tools/testing/selftests/net/tcp_ao/lib/ |
| H A D | repair.c | 42 static void test_sock_checkpoint_queue(int sk, int queue, int qlen, in test_sock_checkpoint_queue() argument 56 if (!qlen) { in test_sock_checkpoint_queue() 61 q->buf = malloc(qlen); in test_sock_checkpoint_queue() 64 ret = recv(sk, q->buf, qlen, MSG_PEEK | MSG_DONTWAIT); in test_sock_checkpoint_queue() 65 if (ret != qlen) in test_sock_checkpoint_queue() 66 test_error("recv(%d): %d", qlen, ret); in test_sock_checkpoint_queue()
|
| /linux/drivers/crypto/marvell/octeontx2/ |
| H A D | otx2_cptvf_reqmgr.c | 47 int qlen) in get_free_pending_entry() argument 56 if (unlikely(q->rear == qlen)) in get_free_pending_entry() 115 pentry = get_free_pending_entry(pqueue, pqueue->qlen); in process_request() 121 pentry = get_free_pending_entry(pqueue, pqueue->qlen); in process_request() 135 pqueue->pending_count > (pqueue->qlen - CPT_IQ_STOP_MARGIN)) { in process_request() 349 resume_index = modulo_inc(pqueue->front, pqueue->qlen, in process_pending_queue() 375 pqueue->front = modulo_inc(pqueue->front, pqueue->qlen, 1); in process_pending_queue()
|
| /linux/net/kcm/ |
| H A D | kcmproc.c | 119 kcm->sk.sk_receive_queue.qlen, in kcm_format_sock() 121 kcm->sk.sk_write_queue.qlen, in kcm_format_sock() 149 psock->sk->sk_receive_queue.qlen, in kcm_format_psock() 151 psock->sk->sk_write_queue.qlen, in kcm_format_psock() 167 if (psock->sk->sk_receive_queue.qlen) { in kcm_format_psock()
|
| /linux/net/nfc/hci/ |
| H A D | llc_shdlc.c | 322 if (shdlc->send_q.qlen == 0) { in llc_shdlc_rcv_s_frame() 462 if (shdlc->rcv_q.qlen) in llc_shdlc_handle_rcv_queue() 463 pr_debug("rcvQlen=%d\n", shdlc->rcv_q.qlen); in llc_shdlc_handle_rcv_queue() 518 if (shdlc->send_q.qlen) in llc_shdlc_handle_send_queue() 520 shdlc->send_q.qlen, shdlc->ns, shdlc->dnr, in llc_shdlc_handle_send_queue() 523 shdlc->ack_pending_q.qlen); in llc_shdlc_handle_send_queue() 525 while (shdlc->send_q.qlen && shdlc->ack_pending_q.qlen < shdlc->w && in llc_shdlc_handle_send_queue()
|