/linux-6.8/fs/ocfs2/cluster/ |
D | quorum.c | 94 struct o2quo_state *qs = &o2quo_state; in o2quo_make_decision() local 96 spin_lock_bh(&qs->qs_lock); in o2quo_make_decision() 98 lowest_hb = find_first_bit(qs->qs_hb_bm, O2NM_MAX_NODES); in o2quo_make_decision() 100 lowest_reachable = test_bit(lowest_hb, qs->qs_conn_bm); in o2quo_make_decision() 103 "lowest: %d (%sreachable)\n", qs->qs_heartbeating, in o2quo_make_decision() 104 qs->qs_connected, lowest_hb, lowest_reachable ? "" : "un"); in o2quo_make_decision() 106 if (!test_bit(o2nm_this_node(), qs->qs_hb_bm) || in o2quo_make_decision() 107 qs->qs_heartbeating == 1) in o2quo_make_decision() 110 if (qs->qs_heartbeating & 1) { in o2quo_make_decision() 113 quorum = (qs->qs_heartbeating + 1)/2; in o2quo_make_decision() [all …]
|
/linux-6.8/kernel/bpf/ |
D | queue_stack_maps.c | 30 static bool queue_stack_map_is_empty(struct bpf_queue_stack *qs) in queue_stack_map_is_empty() argument 32 return qs->head == qs->tail; in queue_stack_map_is_empty() 35 static bool queue_stack_map_is_full(struct bpf_queue_stack *qs) in queue_stack_map_is_full() argument 37 u32 head = qs->head + 1; in queue_stack_map_is_full() 39 if (unlikely(head >= qs->size)) in queue_stack_map_is_full() 42 return head == qs->tail; in queue_stack_map_is_full() 67 struct bpf_queue_stack *qs; in queue_stack_map_alloc() local 71 queue_size = sizeof(*qs) + size * attr->value_size; in queue_stack_map_alloc() 73 qs = bpf_map_area_alloc(queue_size, numa_node); in queue_stack_map_alloc() 74 if (!qs) in queue_stack_map_alloc() [all …]
|
/linux-6.8/drivers/net/ethernet/chelsio/cxgb3/ |
D | sge.c | 721 * @qs: the queue set 726 static void init_qset_cntxt(struct sge_qset *qs, unsigned int id) in init_qset_cntxt() argument 728 qs->rspq.cntxt_id = id; in init_qset_cntxt() 729 qs->fl[0].cntxt_id = 2 * id; in init_qset_cntxt() 730 qs->fl[1].cntxt_id = 2 * id + 1; in init_qset_cntxt() 731 qs->txq[TXQ_ETH].cntxt_id = FW_TUNNEL_SGEEC_START + id; in init_qset_cntxt() 732 qs->txq[TXQ_ETH].token = FW_TUNNEL_TID_START + id; in init_qset_cntxt() 733 qs->txq[TXQ_OFLD].cntxt_id = FW_OFLD_SGEEC_START + id; in init_qset_cntxt() 734 qs->txq[TXQ_CTRL].cntxt_id = FW_CTRL_SGEEC_START + id; in init_qset_cntxt() 735 qs->txq[TXQ_CTRL].token = FW_CTRL_TID_START + id; in init_qset_cntxt() [all …]
|
D | cxgb3_main.c | 409 adap->sge.qs[qidx]. in request_msix_data_irqs() 412 &adap->sge.qs[qidx]); in request_msix_data_irqs() 416 &adap->sge.qs[qidx]); in request_msix_data_irqs() 436 &adapter->sge.qs[i]); in free_irq_resources() 446 while (adap->sge.qs[0].rspq.offload_pkts < init_cnt + n) { in await_mgmt_replies() 459 unsigned long cnt = adap->sge.qs[0].rspq.offload_pkts; in init_tp_parity() 595 struct sge_qset *qs = &adap->sge.qs[i]; in ring_dbs() local 597 if (qs->adap) in ring_dbs() 599 t3_write_reg(adap, A_SG_KDOORBELL, F_SELEGRCNTX | V_EGRCNTX(qs->txq[j].cntxt_id)); in ring_dbs() 608 struct sge_qset *qs = &adap->sge.qs[i]; in init_napi() local [all …]
|
/linux-6.8/drivers/net/ethernet/cavium/thunder/ |
D | nicvf_queues.c | 382 struct queue_set *qs = nic->qs; in nicvf_refill_rbdr() local 383 int rbdr_idx = qs->rbdr_cnt; in nicvf_refill_rbdr() 395 rbdr = &qs->rbdr[rbdr_idx]; in nicvf_refill_rbdr() 404 if (qcount >= (qs->rbdr_len - 1)) in nicvf_refill_rbdr() 407 refill_rb_cnt = qs->rbdr_len - qcount - 1; in nicvf_refill_rbdr() 628 struct queue_set *qs, int qidx) in nicvf_reclaim_snd_queue() argument 640 struct queue_set *qs, int qidx) in nicvf_reclaim_rcv_queue() argument 650 struct queue_set *qs, int qidx) in nicvf_reclaim_cmp_queue() argument 745 static void nicvf_rcv_queue_config(struct nicvf *nic, struct queue_set *qs, in nicvf_rcv_queue_config() argument 752 rq = &qs->rq[qidx]; in nicvf_rcv_queue_config() [all …]
|
D | nicvf_ethtool.c | 217 for (qidx = 0; qidx < nic->qs->rq_cnt; qidx++) { in nicvf_get_qset_strings() 225 for (qidx = 0; qidx < nic->qs->sq_cnt; qidx++) { in nicvf_get_qset_strings() 282 (nic->qs->rq_cnt + nic->qs->sq_cnt); in nicvf_get_sset_count() 290 (snic->qs->rq_cnt + snic->qs->sq_cnt); in nicvf_get_sset_count() 306 for (qidx = 0; qidx < nic->qs->rq_cnt; qidx++) { in nicvf_get_qset_stats() 309 *((*data)++) = ((u64 *)&nic->qs->rq[qidx].stats) in nicvf_get_qset_stats() 313 for (qidx = 0; qidx < nic->qs->sq_cnt; qidx++) { in nicvf_get_qset_stats() 316 *((*data)++) = ((u64 *)&nic->qs->sq[qidx].stats) in nicvf_get_qset_stats() 475 struct queue_set *qs = nic->qs; in nicvf_get_ringparam() local 478 ring->rx_pending = qs->cq_len; in nicvf_get_ringparam() [all …]
|
D | nicvf_main.c | 319 mbx.cpi_cfg.rq_cnt = nic->qs->rq_cnt; in nicvf_config_cpi() 441 nic->snicvf[sqs]->qs->rq_cnt = MAX_RCV_QUEUES_PER_QS; in nicvf_request_sqs() 444 nic->snicvf[sqs]->qs->rq_cnt = rx_queues; in nicvf_request_sqs() 449 nic->snicvf[sqs]->qs->sq_cnt = MAX_SND_QUEUES_PER_QS; in nicvf_request_sqs() 452 nic->snicvf[sqs]->qs->sq_cnt = tx_queues; in nicvf_request_sqs() 456 nic->snicvf[sqs]->qs->cq_cnt = in nicvf_request_sqs() 457 max(nic->snicvf[sqs]->qs->rq_cnt, nic->snicvf[sqs]->qs->sq_cnt); in nicvf_request_sqs() 667 sq = &nic->qs->sq[cqe_tx->sq_idx]; in nicvf_snd_pkt_handler() 852 struct queue_set *qs = nic->qs; in nicvf_cq_intr_handler() local 853 struct cmp_queue *cq = &qs->cq[cq_idx]; in nicvf_cq_intr_handler() [all …]
|
D | nicvf_queues.h | 58 /* Default queue count per QS, its lengths and threshold values */ 244 u8 cq_qs; /* CQ's QS to which this RQ is assigned */ 245 u8 cq_idx; /* CQ index (0 to 7) in the QS */ 246 u8 cont_rbdr_qs; /* Continue buffer ptrs - QS num */ 247 u8 cont_qs_rbdr_idx; /* RBDR idx in the cont QS */ 248 u8 start_rbdr_qs; /* First buffer ptrs - QS num */ 249 u8 start_qs_rbdr_idx; /* RBDR idx in the above QS */ 266 u8 cq_qs; /* CQ's QS to which this SQ is pointing */ 267 u8 cq_idx; /* CQ index (0 to 7) in the above QS */ 335 void nicvf_cmp_queue_config(struct nicvf *nic, struct queue_set *qs,
|
/linux-6.8/drivers/net/ethernet/hisilicon/hns/ |
D | hns_ae_adapt.c | 84 vf_cb = kzalloc(struct_size(vf_cb, ae_handle.qs, qnum_per_vf), in hns_ae_get_handle() 112 ae_handle->qs[i] = &ring_pair_cb->q; in hns_ae_get_handle() 113 ae_handle->qs[i]->rx_ring.q = ae_handle->qs[i]; in hns_ae_get_handle() 114 ae_handle->qs[i]->tx_ring.q = ae_handle->qs[i]; in hns_ae_get_handle() 144 hns_ae_get_ring_pair(handle->qs[i])->used_by_vf = 0; in hns_ae_put_handle() 158 ret = hns_rcb_wait_tx_ring_clean(handle->qs[i]); in hns_ae_wait_flow_down() 190 hns_rcb_ring_enable_hw(handle->qs[i], val); in hns_ae_ring_enable_all() 316 q = handle->qs[i]; in hns_ae_set_mtu() 344 hns_rcb_int_clr_hw(handle->qs[k], in hns_ae_start() 347 hns_rcbv2_int_clr_hw(handle->qs[k], in hns_ae_start() [all …]
|
D | hnae.c | 297 hnae_fini_queue(handle->qs[i]); in hnae_reinit_handle() 303 ret = hnae_init_queue(handle, handle->qs[i], handle->dev); in hnae_reinit_handle() 310 hnae_fini_queue(handle->qs[j]); in hnae_reinit_handle() 349 ret = hnae_init_queue(handle, handle->qs[i], dev); in hnae_get_handle() 362 hnae_fini_queue(handle->qs[j]); in hnae_get_handle() 376 hnae_fini_queue(h->qs[i]); in hnae_put_handle()
|
D | hns_dsaf_rcb.c | 32 *@qs: ring struct pointer array 36 void hns_rcb_wait_fbd_clean(struct hnae_queue **qs, int q_num, u32 flag) in hns_rcb_wait_fbd_clean() argument 45 fbd_num += dsaf_read_dev(qs[i], in hns_rcb_wait_fbd_clean() 48 fbd_num += dsaf_read_dev(qs[i], in hns_rcb_wait_fbd_clean() 57 dev_err(qs[i]->handle->owner_dev, in hns_rcb_wait_fbd_clean() 61 int hns_rcb_wait_tx_ring_clean(struct hnae_queue *qs) in hns_rcb_wait_tx_ring_clean() argument 66 tail = dsaf_read_dev(&qs->tx_ring, RCB_REG_TAIL); in hns_rcb_wait_tx_ring_clean() 69 head = dsaf_read_dev(&qs->tx_ring, RCB_REG_HEAD); in hns_rcb_wait_tx_ring_clean() 77 dev_err(qs->dev->dev, "rcb wait timeout, head not equal to tail.\n"); in hns_rcb_wait_tx_ring_clean()
|
D | hns_dsaf_rcb.h | 134 void hns_rcb_wait_fbd_clean(struct hnae_queue **qs, int q_num, u32 flag); 135 int hns_rcb_wait_tx_ring_clean(struct hnae_queue *qs);
|
/linux-6.8/fs/qnx4/ |
D | inode.c | 46 struct qnx4_sb_info *qs; in qnx4_remount() local 49 qs = qnx4_sb(sb); in qnx4_remount() 50 qs->Version = QNX4_VERSION; in qnx4_remount() 191 struct qnx4_sb_info *qs; in qnx4_fill_super() local 193 qs = kzalloc(sizeof(struct qnx4_sb_info), GFP_KERNEL); in qnx4_fill_super() 194 if (!qs) in qnx4_fill_super() 196 s->s_fs_info = qs; in qnx4_fill_super() 240 struct qnx4_sb_info *qs = qnx4_sb(sb); in qnx4_kill_sb() local 242 if (qs) { in qnx4_kill_sb() 243 kfree(qs->BitMap); in qnx4_kill_sb() [all …]
|
/linux-6.8/drivers/soc/qcom/ |
D | socinfo.c | 759 struct qcom_socinfo *qs; in qcom_socinfo_probe() local 770 qs = devm_kzalloc(&pdev->dev, sizeof(*qs), GFP_KERNEL); in qcom_socinfo_probe() 771 if (!qs) in qcom_socinfo_probe() 774 qs->attr.family = "Snapdragon"; in qcom_socinfo_probe() 775 qs->attr.machine = socinfo_machine(&pdev->dev, in qcom_socinfo_probe() 777 qs->attr.soc_id = devm_kasprintf(&pdev->dev, GFP_KERNEL, "%u", in qcom_socinfo_probe() 779 qs->attr.revision = devm_kasprintf(&pdev->dev, GFP_KERNEL, "%u.%u", in qcom_socinfo_probe() 783 qs->attr.serial_number = devm_kasprintf(&pdev->dev, GFP_KERNEL, in qcom_socinfo_probe() 787 qs->soc_dev = soc_device_register(&qs->attr); in qcom_socinfo_probe() 788 if (IS_ERR(qs->soc_dev)) in qcom_socinfo_probe() [all …]
|
/linux-6.8/drivers/net/ethernet/intel/ice/ |
D | ice_eswitch.c | 459 pf->eswitch.qs.value = 1; in ice_eswitch_enable_switchdev() 492 pf->eswitch.qs.is_reaching = false; in ice_eswitch_disable_switchdev() 623 if (eswitch->qs.is_reaching) { in ice_eswitch_cp_change_queues() 624 if (eswitch->qs.to_reach >= eswitch->qs.value + change) { in ice_eswitch_cp_change_queues() 625 queues = eswitch->qs.to_reach; in ice_eswitch_cp_change_queues() 626 eswitch->qs.is_reaching = false; in ice_eswitch_cp_change_queues() 630 } else if ((change > 0 && cp->alloc_txq <= eswitch->qs.value) || in ice_eswitch_cp_change_queues() 646 eswitch->qs.value += change; in ice_eswitch_cp_change_queues() 665 pf->eswitch.qs.to_reach -= 1; in ice_eswitch_attach() 769 if (pf->eswitch.qs.value + change < 0) in ice_eswitch_reserve_cp_queues() [all …]
|
/linux-6.8/drivers/net/ethernet/chelsio/cxgb4vf/ |
D | cxgb4vf_main.c | 367 int qs, msi; in name_msix_vecs() local 369 for (qs = 0, msi = MSIX_IQFLINT; qs < pi->nqsets; qs++, msi++) { in name_msix_vecs() 371 "%s-%d", dev->name, qs); in name_msix_vecs() 634 int qs; in setup_sge_queues() local 636 for (qs = 0; qs < pi->nqsets; qs++, rxq++, txq++) { in setup_sge_queues() 644 netdev_get_tx_queue(dev, qs), in setup_sge_queues() 649 rxq->rspq.idx = qs; in setup_sge_queues() 665 int qs; in setup_sge_queues() local 667 for (qs = 0; qs < pi->nqsets; qs++, rxq++, txq++) { in setup_sge_queues() 707 int qs, err; in setup_rss() local [all …]
|
/linux-6.8/arch/s390/include/asm/ |
D | atomic_ops.h | 55 : [old] "=d" (old), [ptr] "+QS" (*ptr) \ 83 : [ptr] "+QS" (*ptr) : [val] "i" (val) : "cc", "memory");\ 134 : [old] "=d" (old), [new] "=&d" (new), [ptr] "+QS" (*ptr)\ 183 : [old] "+d" (old), [ptr] "+QS" (*ptr) in __atomic64_cmpxchg() 195 : [old] "+d" (old), [ptr] "+QS" (*ptr) in __atomic64_cmpxchg_bool()
|
D | cmpxchg.h | 66 : "=&d" (old), "+QS" (*(long *) address) in __arch_xchg() 169 [address] "+QS" (*(long *)address) in __cmpxchg() 199 : [old] "+d" (old), [ptr] "+QS" (*ptr) in arch_cmpxchg128()
|
/linux-6.8/Documentation/devicetree/bindings/net/ |
D | mscc,vsc7514-switch.yaml | 60 - description: qs target 85 - const: qs 153 reg-names = "sys", "rew", "qs", "ptp", "port0", "port1", 200 reg-names = "sys", "rew", "qs", "ptp", "port0", "port1",
|
/linux-6.8/fs/qnx6/ |
D | inode.c | 303 struct qnx6_sb_info *qs; in qnx6_fill_super() local 308 qs = kzalloc(sizeof(struct qnx6_sb_info), GFP_KERNEL); in qnx6_fill_super() 309 if (!qs) in qnx6_fill_super() 311 s->s_fs_info = qs; in qnx6_fill_super() 476 kfree(qs); in qnx6_fill_super() 483 struct qnx6_sb_info *qs = QNX6_SB(sb); in qnx6_put_super() local 484 brelse(qs->sb_buf); in qnx6_put_super() 485 iput(qs->longfile); in qnx6_put_super() 486 iput(qs->inodes); in qnx6_put_super() 487 kfree(qs); in qnx6_put_super()
|
/linux-6.8/drivers/staging/fieldbus/anybuss/ |
D | host.c | 382 struct kfifo qs[3]; member 875 static bool qs_have_work(struct kfifo *qs, size_t num) in qs_have_work() argument 881 for (i = 0; i < num; i++, qs++) { in qs_have_work() 882 ret = kfifo_out_peek(qs, &t, sizeof(t)); in qs_have_work() 892 struct kfifo *qs = cd->qs; in process_qs() local 893 size_t nqs = ARRAY_SIZE(cd->qs); in process_qs() 895 for (i = 0; i < nqs; i++, qs++) in process_qs() 896 process_q(cd, qs); in process_qs() 968 struct kfifo *qs = cd->qs; in qthread_fn() local 969 size_t nqs = ARRAY_SIZE(cd->qs); in qthread_fn() [all …]
|
/linux-6.8/drivers/scsi/elx/efct/ |
D | efct_hw_queues.c | 199 struct sli4_queue *qs[SLI4_MAX_CQ_SET_COUNT]; in efct_hw_new_cq_set() local 217 qs[i] = cq->queue; in efct_hw_new_cq_set() 222 if (sli_cq_alloc_set(sli4, qs, num_cqs, entry_count, assefct)) { in efct_hw_new_cq_set() 315 struct sli4_queue *qs[SLI4_MAX_RQ_SET_COUNT * 2] = { NULL }; in efct_hw_new_rq_set() local 342 qs[q_count] = rq->hdr; in efct_hw_new_rq_set() 349 qs[q_count + 1] = rq->data; in efct_hw_new_rq_set() 354 if (sli_fc_rq_set_alloc(&hw->sli, num_rq_pairs, qs, in efct_hw_new_rq_set()
|
/linux-6.8/drivers/s390/cio/ |
D | qdio.h | 96 " .insn rsy,0xeb000000008a,%[qs],%[ccq],0(%[state])" in do_sqbs() 97 : [ccq] "+&d" (_ccq), [qs] "+&d" (_queuestart) in do_sqbs() 115 " .insn rrf,0xb99c0000,%[qs],%[state],%[ccq],0" in do_eqbs() 116 : [ccq] "+&d" (_ccq), [qs] "+&d" (_queuestart), in do_eqbs()
|
/linux-6.8/drivers/net/dsa/ocelot/ |
D | ocelot_ext.c | 47 [QS] = "qs",
|
/linux-6.8/drivers/net/ethernet/marvell/octeontx2/nic/ |
D | otx2_ethtool.c | 382 struct otx2_qset *qs = &pfvf->qset; in otx2_get_ringparam() local 385 ring->rx_pending = qs->rqe_cnt ? qs->rqe_cnt : Q_COUNT(Q_SIZE_256); in otx2_get_ringparam() 387 ring->tx_pending = qs->sqe_cnt ? qs->sqe_cnt : Q_COUNT(Q_SIZE_4K); in otx2_get_ringparam() 402 struct otx2_qset *qs = &pfvf->qset; in otx2_set_ringparam() local 440 if (tx_count == qs->sqe_cnt && rx_count == qs->rqe_cnt && in otx2_set_ringparam() 448 qs->sqe_cnt = tx_count; in otx2_set_ringparam() 449 qs->rqe_cnt = rx_count; in otx2_set_ringparam()
|