Home
last modified time | relevance | path

Searched refs:qs (Results 1 – 25 of 39) sorted by relevance

12

/linux/fs/ocfs2/cluster/ !
H A Dquorum.c94 struct o2quo_state *qs = &o2quo_state; in o2quo_make_decision() local
96 spin_lock_bh(&qs->qs_lock); in o2quo_make_decision()
98 lowest_hb = find_first_bit(qs->qs_hb_bm, O2NM_MAX_NODES); in o2quo_make_decision()
100 lowest_reachable = test_bit(lowest_hb, qs->qs_conn_bm); in o2quo_make_decision()
103 "lowest: %d (%sreachable)\n", qs->qs_heartbeating, in o2quo_make_decision()
104 qs->qs_connected, lowest_hb, lowest_reachable ? "" : "un"); in o2quo_make_decision()
106 if (!test_bit(o2nm_this_node(), qs->qs_hb_bm) || in o2quo_make_decision()
107 qs->qs_heartbeating == 1) in o2quo_make_decision()
110 if (qs->qs_heartbeating & 1) { in o2quo_make_decision()
113 quorum = (qs in o2quo_make_decision()
162 o2quo_set_hold(struct o2quo_state * qs,u8 node) o2quo_set_hold() argument
174 o2quo_clear_hold(struct o2quo_state * qs,u8 node) o2quo_clear_hold() argument
197 struct o2quo_state *qs = &o2quo_state; o2quo_hb_up() local
221 struct o2quo_state *qs = &o2quo_state; o2quo_hb_down() local
246 struct o2quo_state *qs = &o2quo_state; o2quo_hb_still_up() local
265 struct o2quo_state *qs = &o2quo_state; o2quo_conn_up() local
291 struct o2quo_state *qs = &o2quo_state; o2quo_conn_err() local
315 struct o2quo_state *qs = &o2quo_state; o2quo_init() local
323 struct o2quo_state *qs = &o2quo_state; o2quo_exit() local
[all...]
/linux/kernel/bpf/ !
H A Dqueue_stack_maps.c31 static bool queue_stack_map_is_empty(struct bpf_queue_stack *qs) in queue_stack_map_is_empty() argument
33 return qs->head == qs->tail; in queue_stack_map_is_empty()
36 static bool queue_stack_map_is_full(struct bpf_queue_stack *qs) in queue_stack_map_is_full() argument
38 u32 head = qs->head + 1; in queue_stack_map_is_full()
40 if (unlikely(head >= qs->size)) in queue_stack_map_is_full()
43 return head == qs->tail; in queue_stack_map_is_full()
68 struct bpf_queue_stack *qs; in queue_stack_map_alloc() local
72 queue_size = sizeof(*qs) + size * attr->value_size; in queue_stack_map_alloc()
74 qs in queue_stack_map_alloc()
90 struct bpf_queue_stack *qs = bpf_queue_stack(map); queue_stack_map_free() local
97 struct bpf_queue_stack *qs = bpf_queue_stack(map); __queue_map_get() local
127 struct bpf_queue_stack *qs = bpf_queue_stack(map); __stack_map_get() local
185 struct bpf_queue_stack *qs = bpf_queue_stack(map); queue_stack_map_push_elem() local
[all...]
/linux/drivers/net/ethernet/chelsio/cxgb3/ !
H A Dsge.c721 * @qs: the queue set
726 static void init_qset_cntxt(struct sge_qset *qs, unsigned int id) in init_qset_cntxt() argument
728 qs->rspq.cntxt_id = id; in init_qset_cntxt()
729 qs->fl[0].cntxt_id = 2 * id; in init_qset_cntxt()
730 qs->fl[1].cntxt_id = 2 * id + 1; in init_qset_cntxt()
731 qs->txq[TXQ_ETH].cntxt_id = FW_TUNNEL_SGEEC_START + id; in init_qset_cntxt()
732 qs->txq[TXQ_ETH].token = FW_TUNNEL_TID_START + id; in init_qset_cntxt()
733 qs->txq[TXQ_OFLD].cntxt_id = FW_OFLD_SGEEC_START + id; in init_qset_cntxt()
734 qs->txq[TXQ_CTRL].cntxt_id = FW_CTRL_SGEEC_START + id; in init_qset_cntxt()
735 qs in init_qset_cntxt()
1246 t3_stop_tx_queue(struct netdev_queue * txq,struct sge_qset * qs,struct sge_txq * q) t3_stop_tx_queue() argument
1267 struct sge_qset *qs; t3_eth_xmit() local
1427 struct sge_qset *qs = txq_to_qset(q, qid); check_desc_avail() local
1523 struct sge_qset *qs = container_of(w, struct sge_qset, restart_ctrlq() local
1743 struct sge_qset *qs = container_of(w, struct sge_qset, restart_offloadq() local
1836 struct sge_qset *qs = &adap->sge.qs[queue_set(skb)]; t3_offload_tx() local
1860 struct sge_qset *qs = rspq_to_qset(q); offload_enqueue() local
1898 struct sge_qset *qs = container_of(napi, struct sge_qset, napi); ofld_poll() local
1985 restart_tx(struct sge_qset * qs) restart_tx() argument
2088 struct sge_qset *qs = rspq_to_qset(rq); rx_eth() local
2134 lro_add_page(struct adapter * adap,struct sge_qset * qs,struct sge_fl * fl,int len,int complete) lro_add_page() argument
2216 handle_rsp_cntrl_info(struct sge_qset * qs,u32 flags) handle_rsp_cntrl_info() argument
2252 check_ring_db(struct adapter * adap,struct sge_qset * qs,unsigned int sleeping) check_ring_db() argument
2322 process_responses(struct adapter * adap,struct sge_qset * qs,int budget) process_responses() argument
2476 struct sge_qset *qs = container_of(napi, struct sge_qset, napi); napi_rx_handler() local
2518 process_pure_responses(struct adapter * adap,struct sge_qset * qs,struct rsp_desc * r) process_pure_responses() argument
2577 struct sge_qset *qs = rspq_to_qset(q); handle_responses() local
2598 struct sge_qset *qs = cookie; t3_sge_intr_msix() local
2617 struct sge_qset *qs = cookie; t3_sge_intr_msix_napi() local
2665 rspq_check_napi(struct sge_qset * qs) rspq_check_napi() argument
2909 struct sge_qset *qs = timer_container_of(qs, t, tx_reclaim_timer); sge_timer_tx() local
2950 struct sge_qset *qs = timer_container_of(qs, t, rx_reclaim_timer); sge_timer_rx() local
2998 t3_update_qset_coalesce(struct sge_qset * qs,const struct qset_params * p) t3_update_qset_coalesce() argument
3292 struct sge_qset *qs = &adap->sge.qs[i]; t3_sge_stop() local
[all...]
H A Dcxgb3_main.c409 adap->sge.qs[qidx]. in request_msix_data_irqs()
412 &adap->sge.qs[qidx]); in request_msix_data_irqs()
416 &adap->sge.qs[qidx]); in request_msix_data_irqs()
436 &adapter->sge.qs[i]); in free_irq_resources()
446 while (adap->sge.qs[0].rspq.offload_pkts < init_cnt + n) { in await_mgmt_replies()
459 unsigned long cnt = adap->sge.qs[0].rspq.offload_pkts; in init_tp_parity()
595 struct sge_qset *qs = &adap->sge.qs[i]; in ring_dbs() local
597 if (qs->adap) in ring_dbs()
599 t3_write_reg(adap, A_SG_KDOORBELL, F_SELEGRCNTX | V_EGRCNTX(qs in ring_dbs()
608 struct sge_qset *qs = &adap->sge.qs[i]; init_napi() local
2009 struct sge_qset *qs; set_coalesce() local
2218 struct sge_qset *qs = cxgb_siocdevprivate() local
2620 struct sge_qset *qs = &adapter->sge.qs[qidx]; cxgb_netpoll() local
2768 struct sge_qset *qs = &adapter->sge.qs[0]; t3_adap_check_task() local
[all...]
H A Dadapter.h68 struct sge_qset *qs; member
216 struct sge_qset qs[SGE_QSETS]; member
325 void t3_update_qset_coalesce(struct sge_qset *qs, const struct qset_params *p);
/linux/drivers/net/ethernet/cavium/thunder/ !
H A Dnicvf_queues.c382 struct queue_set *qs = nic->qs; in nicvf_refill_rbdr() local
383 int rbdr_idx = qs->rbdr_cnt; in nicvf_refill_rbdr()
395 rbdr = &qs->rbdr[rbdr_idx]; in nicvf_refill_rbdr()
404 if (qcount >= (qs->rbdr_len - 1)) in nicvf_refill_rbdr()
407 refill_rb_cnt = qs->rbdr_len - qcount - 1; in nicvf_refill_rbdr()
628 struct queue_set *qs, int qidx) in nicvf_reclaim_snd_queue() argument
640 struct queue_set *qs, int qidx) in nicvf_reclaim_rcv_queue() argument
650 struct queue_set *qs, int qidx) in nicvf_reclaim_cmp_queue() argument
745 static void nicvf_rcv_queue_config(struct nicvf *nic, struct queue_set *qs, in nicvf_rcv_queue_config() argument
818 nicvf_cmp_queue_config(struct nicvf * nic,struct queue_set * qs,int qidx,bool enable) nicvf_cmp_queue_config() argument
859 nicvf_snd_queue_config(struct nicvf * nic,struct queue_set * qs,int qidx,bool enable) nicvf_snd_queue_config() argument
917 nicvf_rbdr_config(struct nicvf * nic,struct queue_set * qs,int qidx,bool enable) nicvf_rbdr_config() argument
957 struct queue_set *qs = nic->qs; nicvf_qset_config() local
992 struct queue_set *qs = nic->qs; nicvf_free_resources() local
1010 struct queue_set *qs = nic->qs; nicvf_alloc_resources() local
1039 struct queue_set *qs; nicvf_set_qset_resources() local
1067 struct queue_set *qs = nic->qs; nicvf_config_data_transfer() local
[all...]
H A Dnicvf_main.c319 mbx.cpi_cfg.rq_cnt = nic->qs->rq_cnt; in nicvf_config_cpi()
441 nic->snicvf[sqs]->qs->rq_cnt = MAX_RCV_QUEUES_PER_QS; in nicvf_request_sqs()
444 nic->snicvf[sqs]->qs->rq_cnt = rx_queues; in nicvf_request_sqs()
449 nic->snicvf[sqs]->qs->sq_cnt = MAX_SND_QUEUES_PER_QS; in nicvf_request_sqs()
452 nic->snicvf[sqs]->qs->sq_cnt = tx_queues; in nicvf_request_sqs()
456 nic->snicvf[sqs]->qs->cq_cnt = in nicvf_request_sqs()
457 max(nic->snicvf[sqs]->qs->rq_cnt, nic->snicvf[sqs]->qs->sq_cnt); in nicvf_request_sqs()
667 sq = &nic->qs->sq[cqe_tx->sq_idx]; in nicvf_snd_pkt_handler()
852 struct queue_set *qs in nicvf_cq_intr_handler() local
988 struct queue_set *qs = nic->qs; nicvf_handle_qs_err() local
1328 struct queue_set *qs = nic->qs; nicvf_stop() local
1453 struct queue_set *qs = nic->qs; nicvf_open() local
1656 struct queue_set *qs = nic->qs; nicvf_update_stats() local
[all...]
H A Dnic.h283 struct queue_set *qs; member
587 struct qs_cfg_msg qs; member
/linux/drivers/soc/qcom/ !
H A Dsocinfo.c788 struct qcom_socinfo *qs; in qcom_socinfo_probe() local
799 qs = devm_kzalloc(&pdev->dev, sizeof(*qs), GFP_KERNEL); in qcom_socinfo_probe()
800 if (!qs) in qcom_socinfo_probe()
803 qs->attr.family = "Snapdragon"; in qcom_socinfo_probe()
804 qs->attr.machine = socinfo_machine(&pdev->dev, in qcom_socinfo_probe()
806 qs->attr.soc_id = devm_kasprintf(&pdev->dev, GFP_KERNEL, "%u", in qcom_socinfo_probe()
808 qs->attr.revision = devm_kasprintf(&pdev->dev, GFP_KERNEL, "%u.%u", in qcom_socinfo_probe()
811 if (!qs->attr.soc_id || !qs in qcom_socinfo_probe()
838 struct qcom_socinfo *qs = platform_get_drvdata(pdev); qcom_socinfo_remove() local
[all...]
/linux/fs/qnx4/ !
H A Dinode.c47 struct qnx4_sb_info *qs; in qnx4_reconfigure() local
50 qs = qnx4_sb(sb); in qnx4_reconfigure()
51 qs->Version = QNX4_VERSION; in qnx4_reconfigure()
197 struct qnx4_sb_info *qs; in qnx4_fill_super() local
200 qs = kzalloc(sizeof(struct qnx4_sb_info), GFP_KERNEL); in qnx4_fill_super()
201 if (!qs) in qnx4_fill_super()
203 s->s_fs_info = qs; in qnx4_fill_super()
259 struct qnx4_sb_info *qs = qnx4_sb(sb); in qnx4_kill_sb() local
261 if (qs) { in qnx4_kill_sb()
262 kfree(qs in qnx4_kill_sb()
[all...]
/linux/drivers/net/ethernet/hisilicon/hns/ !
H A Dhns_ae_adapt.c84 vf_cb = kzalloc(struct_size(vf_cb, ae_handle.qs, qnum_per_vf), in hns_ae_get_handle()
112 ae_handle->qs[i] = &ring_pair_cb->q; in hns_ae_get_handle()
113 ae_handle->qs[i]->rx_ring.q = ae_handle->qs[i]; in hns_ae_get_handle()
114 ae_handle->qs[i]->tx_ring.q = ae_handle->qs[i]; in hns_ae_get_handle()
144 hns_ae_get_ring_pair(handle->qs[i])->used_by_vf = 0; in hns_ae_put_handle()
158 ret = hns_rcb_wait_tx_ring_clean(handle->qs[i]); in hns_ae_wait_flow_down()
190 hns_rcb_ring_enable_hw(handle->qs[i], val); in hns_ae_ring_enable_all()
316 q = handle->qs[ in hns_ae_set_mtu()
[all...]
H A Dhnae.c299 hnae_fini_queue(handle->qs[i]); in hnae_reinit_handle()
305 ret = hnae_init_queue(handle, handle->qs[i], handle->dev); in hnae_reinit_handle()
312 hnae_fini_queue(handle->qs[j]); in hnae_reinit_handle()
351 ret = hnae_init_queue(handle, handle->qs[i], dev); in hnae_get_handle()
364 hnae_fini_queue(handle->qs[j]); in hnae_get_handle()
378 hnae_fini_queue(h->qs[i]); in hnae_put_handle()
H A Dhns_dsaf_rcb.h133 void hns_rcb_wait_fbd_clean(struct hnae_queue **qs, int q_num, u32 flag);
134 int hns_rcb_wait_tx_ring_clean(struct hnae_queue *qs);
H A Dhns_dsaf_rcb.c32 *@qs: ring struct pointer array
36 void hns_rcb_wait_fbd_clean(struct hnae_queue **qs, int q_num, u32 flag) in hns_rcb_wait_fbd_clean() argument
45 fbd_num += dsaf_read_dev(qs[i], in hns_rcb_wait_fbd_clean()
48 fbd_num += dsaf_read_dev(qs[i], in hns_rcb_wait_fbd_clean()
57 dev_err(qs[i]->handle->owner_dev, in hns_rcb_wait_fbd_clean()
61 int hns_rcb_wait_tx_ring_clean(struct hnae_queue *qs) in hns_rcb_wait_tx_ring_clean() argument
66 tail = dsaf_read_dev(&qs->tx_ring, RCB_REG_TAIL); in hns_rcb_wait_tx_ring_clean()
69 head = dsaf_read_dev(&qs->tx_ring, RCB_REG_HEAD); in hns_rcb_wait_tx_ring_clean()
77 dev_err(qs->dev->dev, "rcb wait timeout, head not equal to tail.\n"); in hns_rcb_wait_tx_ring_clean()
H A Dhnae.h561 struct hnae_queue *qs[]; /* flexible array of all queues */ member
665 ring = &h->qs[i]->rx_ring; in hnae_reinit_all_ring_desc()
680 ring = &h->qs[i]->rx_ring; in hnae_reinit_all_ring_page_off()
/linux/fs/qnx6/ !
H A Dinode.c299 struct qnx6_sb_info *qs; in qnx6_fill_super() local
305 qs = kzalloc(sizeof(struct qnx6_sb_info), GFP_KERNEL); in qnx6_fill_super()
306 if (!qs) in qnx6_fill_super()
308 s->s_fs_info = qs; in qnx6_fill_super()
309 qs->s_mount_opt = ctx->s_mount_opts; in qnx6_fill_super()
317 if (qs->s_mount_opt == QNX6_MOUNT_MMI_FS) { in qnx6_fill_super()
469 kfree(qs); in qnx6_fill_super()
476 struct qnx6_sb_info *qs = QNX6_SB(sb); in qnx6_put_super() local
477 brelse(qs->sb_buf); in qnx6_put_super()
478 iput(qs in qnx6_put_super()
[all...]
/linux/drivers/scsi/elx/efct/ !
H A Defct_hw_queues.c199 struct sli4_queue *qs[SLI4_MAX_CQ_SET_COUNT]; in efct_hw_new_cq_set() local
217 qs[i] = cq->queue; in efct_hw_new_cq_set()
222 if (sli_cq_alloc_set(sli4, qs, num_cqs, entry_count, assefct)) { in efct_hw_new_cq_set()
315 struct sli4_queue *qs[SLI4_MAX_RQ_SET_COUNT * 2] = { NULL }; in efct_hw_new_rq_set() local
342 qs[q_count] = rq->hdr; in efct_hw_new_rq_set()
349 qs[q_count + 1] = rq->data; in efct_hw_new_rq_set()
354 if (sli_fc_rq_set_alloc(&hw->sli, num_rq_pairs, qs, in efct_hw_new_rq_set()
/linux/tools/testing/selftests/drivers/net/ !
H A Dstats.py68 for qs in stats:
69 if qs["ifindex"]== test.ifindex:
70 return qs
/linux/drivers/md/ !
H A Ddm-cache-policy-smq.c258 struct ilist qs[MAX_LEVELS]; member
279 l_init(q->qs + i); in q_init()
303 l_add_tail(q->es, q->qs + e->level, e); in q_push()
313 l_add_head(q->es, q->qs + e->level, e); in q_push_front()
323 l_add_before(q->es, q->qs + e->level, old, e); in q_push_before()
328 l_del(q->es, q->qs + e->level, e); in q_del()
344 for (e = l_head(q->es, q->qs + level); e; e = l_next(q->es, e)) { in q_peek()
378 for (e = l_head(q->es, q->qs + level); e; e = l_next(q->es, e)) in __redist_pop_from()
380 l_del(q->es, q->qs + e->level, e); in __redist_pop_from()
438 l = q->qs in q_redistribute()
[all...]
/linux/arch/s390/include/uapi/asm/ !
H A Druntime_instr.h22 __u32 qs : 1; member
/linux/drivers/s390/cio/ !
H A Dqdio.h96 " .insn rsy,0xeb000000008a,%[qs],%[ccq],0(%[state])" in do_sqbs()
97 : [ccq] "+&d" (_ccq), [qs] "+&d" (_queuestart) in do_sqbs()
115 " .insn rrf,0xb99c0000,%[qs],%[state],%[ccq],0" in do_eqbs()
116 : [ccq] "+&d" (_ccq), [qs] "+&d" (_queuestart), in do_eqbs()
/linux/include/uapi/sound/ !
H A Dhdspm.h27 qs enumerator
41 __u8 speed; /* enum {ss, ds, qs} */
/linux/net/sched/ !
H A Dsch_sfq.c909 struct gnet_stats_queue qs = { 0 }; in sfq_dump_class_stats() local
916 qs.qlen = slot->qlen; in sfq_dump_class_stats()
917 qs.backlog = slot->backlog; in sfq_dump_class_stats()
919 if (gnet_stats_copy_queue(d, NULL, &qs, qs.qlen) < 0) in sfq_dump_class_stats()
/linux/drivers/scsi/elx/libefc_sli/ !
H A Dsli4.c428 struct sli4_queue *qs[], u32 base_cq_id, in sli_cmd_rq_create_v2() argument
439 page_count = sli_page_count(qs[0]->dma.size, SLI_PAGE_SIZE) * num_rqs; in sli_cmd_rq_create_v2()
463 num_pages = sli_page_count(qs[0]->dma.size, SLI_PAGE_SIZE); in sli_cmd_rq_create_v2()
465 req->rqe_count = cpu_to_le16(qs[0]->dma.size / SLI4_RQE_SIZE); in sli_cmd_rq_create_v2()
474 for (p = 0, addr = qs[i]->dma.phys; p < num_pages; in sli_cmd_rq_create_v2()
582 struct sli4_queue *qs[], u32 base_cq_id, in sli_fc_rq_set_alloc() argument
593 if (__sli_queue_init(sli4, qs[i], SLI4_QTYPE_RQ, in sli_fc_rq_set_alloc()
600 if (sli_cmd_rq_create_v2(sli4, num_rqs, qs, base_cq_id, in sli_fc_rq_set_alloc()
624 qs[i]->id = i + le16_to_cpu(rsp->q_id); in sli_fc_rq_set_alloc()
625 if ((qs[ in sli_fc_rq_set_alloc()
817 sli_cmd_cq_set_create(struct sli4 * sli4,struct sli4_queue * qs[],u32 num_cqs,struct sli4_queue * eqs[],struct efc_dma * dma) sli_cmd_cq_set_create() argument
917 sli_cq_alloc_set(struct sli4 * sli4,struct sli4_queue * qs[],u32 num_cqs,u32 n_entries,struct sli4_queue * eqs[]) sli_cq_alloc_set() argument
[all...]
/linux/arch/mips/boot/dts/mscc/ !
H A Docelot.dtsi141 reg-names = "sys", "rew", "qs", "ptp", "port0", "port1",

12