Searched refs:hr_cq (Results 1 – 6 of 6) sorted by relevance
| /linux/drivers/infiniband/hw/hns/ |
| H A D | hns_roce_cq.c | 112 static int alloc_cqn(struct hns_roce_dev *hr_dev, struct hns_roce_cq *hr_cq, in alloc_cqn() argument 131 hr_cq->cqn = (id << CQ_BANKID_SHIFT) | bankid; in alloc_cqn() 159 struct hns_roce_cq *hr_cq, in hns_roce_create_cqc() argument 172 hr_dev->hw->write_cqc(hr_dev, hr_cq, mailbox->buf, mtts, dma_handle); in hns_roce_create_cqc() 175 hr_cq->cqn); in hns_roce_create_cqc() 179 hr_cq->cqn, ret); in hns_roce_create_cqc() 186 static int alloc_cqc(struct hns_roce_dev *hr_dev, struct hns_roce_cq *hr_cq) in alloc_cqc() argument 193 ret = hns_roce_mtr_find(hr_dev, &hr_cq->mtr, 0, mtts, ARRAY_SIZE(mtts)); in alloc_cqc() 200 ret = hns_roce_table_get(hr_dev, &cq_table->table, hr_cq->cqn); in alloc_cqc() 203 hr_cq->cqn, ret); in alloc_cqc() [all …]
|
| H A D | hns_roce_restrack.c | 13 struct hns_roce_cq *hr_cq = to_hr_cq(ib_cq); in hns_roce_fill_res_cq_entry() local 20 if (rdma_nl_put_driver_u32(msg, "cq_depth", hr_cq->cq_depth)) in hns_roce_fill_res_cq_entry() 23 if (rdma_nl_put_driver_u32(msg, "cons_index", hr_cq->cons_index)) in hns_roce_fill_res_cq_entry() 26 if (rdma_nl_put_driver_u32(msg, "cqe_size", hr_cq->cqe_size)) in hns_roce_fill_res_cq_entry() 29 if (rdma_nl_put_driver_u32(msg, "arm_sn", hr_cq->arm_sn)) in hns_roce_fill_res_cq_entry() 45 struct hns_roce_cq *hr_cq = to_hr_cq(ib_cq); in hns_roce_fill_res_cq_entry_raw() local 52 ret = hr_dev->hw->query_cqc(hr_dev, hr_cq->cqn, &context); in hns_roce_fill_res_cq_entry_raw()
|
| H A D | hns_roce_hw_v2.c | 917 struct hns_roce_cq *hr_cq = to_hr_cq(ibcq); in handle_drain_completion() local 930 spin_lock_irqsave(&hr_cq->lock, flags); in handle_drain_completion() 931 triggered = hr_cq->is_armed; in handle_drain_completion() 932 hr_cq->is_armed = 1; in handle_drain_completion() 933 spin_unlock_irqrestore(&hr_cq->lock, flags); in handle_drain_completion() 2910 struct hns_roce_cq *hr_cq; in free_mr_init_cq() local 2915 hr_cq = kzalloc_obj(*hr_cq); in free_mr_init_cq() 2916 if (!hr_cq) in free_mr_init_cq() 2919 cq = &hr_cq->ib_cq; in free_mr_init_cq() 2924 kfree(hr_cq); in free_mr_init_cq() [all …]
|
| H A D | hns_roce_main.c | 1112 struct hns_roce_cq *hr_cq = to_hr_cq(cq); in check_and_get_armed_cq() local 1115 spin_lock_irqsave(&hr_cq->lock, flags); in check_and_get_armed_cq() 1117 if (!hr_cq->is_armed) { in check_and_get_armed_cq() 1118 hr_cq->is_armed = 1; in check_and_get_armed_cq() 1119 list_add_tail(&hr_cq->node, cq_list); in check_and_get_armed_cq() 1122 spin_unlock_irqrestore(&hr_cq->lock, flags); in check_and_get_armed_cq() 1128 struct hns_roce_cq *hr_cq; in hns_roce_handle_device_err() local 1148 list_for_each_entry(hr_cq, &cq_list, node) in hns_roce_handle_device_err() 1149 hns_roce_cq_completion(hr_dev, hr_cq->cqn); in hns_roce_handle_device_err()
|
| H A D | hns_roce_qp.c | 1584 struct hns_roce_cq *hr_cq; in hns_roce_wq_overflow() local 1591 hr_cq = to_hr_cq(ib_cq); in hns_roce_wq_overflow() 1592 spin_lock(&hr_cq->lock); in hns_roce_wq_overflow() 1594 spin_unlock(&hr_cq->lock); in hns_roce_wq_overflow()
|
| H A D | hns_roce_device.h | 939 struct hns_roce_cq *hr_cq, void *mb_buf, u64 *mtts,
|