Home
last modified time | relevance | path

Searched refs:to_mcq (Results 1 – 17 of 17) sorted by relevance

/linux/drivers/infiniband/hw/mlx5/
H A Dcq.c612 struct mlx5_ib_cq *cq = to_mcq(ibcq); in mlx5_ib_poll_cq()
650 struct mlx5_ib_cq *cq = to_mcq(ibcq); in mlx5_ib_arm_cq()
666 uar_page, to_mcq(ibcq)->mcq.cons_index); in mlx5_ib_arm_cq()
960 struct mlx5_ib_cq *cq = to_mcq(ibcq); in mlx5_ib_create_cq()
1064 struct mlx5_ib_cq *mcq = to_mcq(cq); in mlx5_ib_pre_destroy_cq()
1071 destroy_cq_kernel(to_mdev(cq->device), to_mcq(cq)); in mlx5_ib_post_destroy_cq()
1083 destroy_cq_user(to_mcq(cq), udata); in mlx5_ib_destroy_cq()
1158 struct mlx5_ib_cq *mcq = to_mcq(cq); in mlx5_ib_modify_cq()
1288 struct mlx5_ib_cq *cq = to_mcq(ibcq); in mlx5_ib_resize_cq()
1441 cq = to_mcq(ibcq); in mlx5_ib_get_cqe_size()
[all …]
H A Dqp.c1590 to_mcq(init_attr->send_cq)); in create_raw_packet_qp()
1613 to_mcq(init_attr->recv_cq)); in create_raw_packet_qp()
2033 MLX5_SET(qpc, qpc, cqn_rcv, to_mcq(devr->c0)->mcq.cqn); in create_xrc_tgt_qp()
2034 MLX5_SET(qpc, qpc, cqn_snd, to_mcq(devr->c0)->mcq.cqn); in create_xrc_tgt_qp()
2117 ts_format = get_qp_ts_format(dev, to_mcq(init_attr->send_cq), in create_dci()
2118 to_mcq(init_attr->recv_cq)); in create_dci()
2176 to_mcq(init_attr->send_cq)->mcq.cqn); in create_dci()
2180 to_mcq(init_attr->recv_cq)->mcq.cqn); in create_dci()
2285 ts_format = get_qp_ts_format(dev, to_mcq(init_attr->send_cq), in create_user_qp()
2286 to_mcq(init_attr->recv_cq)); in create_user_qp()
[all …]
H A Drestrack.c163 struct mlx5_ib_cq *cq = to_mcq(ibcq); in fill_res_cq_entry_raw()
H A Dsrq.c281 in.cqn = to_mcq(init_attr->ext.cq)->mcq.cqn; in mlx5_ib_create_srq()
283 in.cqn = to_mcq(dev->devr.c0)->mcq.cqn; in mlx5_ib_create_srq()
H A Dmlx5_ib.h1264 static inline struct mlx5_ib_cq *to_mcq(struct ib_cq *ibcq) in to_mcq() function
H A Dwr.c38 cq = to_mcq(ib_cq); in mlx5r_wq_overflow()
H A Dmain.c2948 send_mcq = to_mcq(mqp->ibqp.send_cq); in mlx5_ib_handle_internal_error()
2965 recv_mcq = to_mcq(mqp->ibqp.recv_cq); in mlx5_ib_handle_internal_error()
H A Ddevx.c646 to_mcq(uobj->object)->mcq.cqn) == in devx_is_valid_obj_id()
/linux/drivers/infiniband/hw/mthca/
H A Dmthca_provider.c502 to_mcq(init_attr->send_cq), in mthca_create_qp()
503 to_mcq(init_attr->recv_cq), in mthca_create_qp()
527 to_mcq(init_attr->send_cq), in mthca_create_qp()
528 to_mcq(init_attr->recv_cq), in mthca_create_qp()
613 cq = to_mcq(ibcq); in mthca_create_cq()
701 struct mthca_cq *cq = to_mcq(ibcq); in mthca_resize_cq()
786 to_mcq(cq)->arm_db_index); in mthca_destroy_cq()
790 to_mcq(cq)->set_ci_db_index); in mthca_destroy_cq()
792 mthca_free_cq(to_mdev(cq->device), to_mcq(cq)); in mthca_destroy_cq()
H A Dmthca_provider.h301 static inline struct mthca_cq *to_mcq(struct ib_cq *ibcq) in to_mcq() function
H A Dmthca_cq.c658 struct mthca_cq *cq = to_mcq(ibcq); in mthca_poll_cq()
728 to_mcq(cq)->cqn; in mthca_tavor_arm_cq()
738 struct mthca_cq *cq = to_mcq(ibcq); in mthca_arbel_arm_cq()
H A Dmthca_qp.c735 qp_context->cqn_snd = cpu_to_be32(to_mcq(ibqp->send_cq)->cqn); in __mthca_modify_qp()
774 qp_context->cqn_rcv = cpu_to_be32(to_mcq(ibqp->recv_cq)->cqn); in __mthca_modify_qp()
835 mthca_cq_clean(dev, to_mcq(qp->ibqp.recv_cq), qp->qpn, in __mthca_modify_qp()
838 mthca_cq_clean(dev, to_mcq(qp->ibqp.send_cq), qp->qpn, NULL); in __mthca_modify_qp()
1451 send_cq = to_mcq(qp->ibqp.send_cq); in mthca_free_qp()
1452 recv_cq = to_mcq(qp->ibqp.recv_cq); in mthca_free_qp()
1578 cq = to_mcq(ib_cq); in mthca_wq_overflow()
/linux/drivers/infiniband/hw/mlx4/
H A Dcq.c95 struct mlx4_ib_cq *mcq = to_mcq(cq); in mlx4_ib_modify_cq()
186 struct mlx4_ib_cq *cq = to_mcq(ibcq); in mlx4_ib_create_cq()
387 struct mlx4_ib_cq *cq = to_mcq(ibcq); in mlx4_ib_resize_cq()
485 struct mlx4_ib_cq *mcq = to_mcq(cq); in mlx4_ib_destroy_cq()
884 struct mlx4_ib_cq *cq = to_mcq(ibcq); in mlx4_ib_poll_cq()
911 mlx4_cq_arm(&to_mcq(ibcq)->mcq, in mlx4_ib_arm_cq()
H A Dqp.c962 mlx4_ib_lock_cqs(to_mcq(init_attr->send_cq), in create_rq()
963 to_mcq(init_attr->recv_cq)); in create_rq()
971 mcq = to_mcq(init_attr->send_cq); in create_rq()
973 mcq = to_mcq(init_attr->recv_cq); in create_rq()
975 mlx4_ib_unlock_cqs(to_mcq(init_attr->send_cq), in create_rq()
976 to_mcq(init_attr->recv_cq)); in create_rq()
1240 mlx4_ib_lock_cqs(to_mcq(init_attr->send_cq), in create_qp_common()
1241 to_mcq(init_attr->recv_cq)); in create_qp_common()
1249 mcq = to_mcq(init_attr->send_cq); in create_qp_common()
1251 mcq = to_mcq(init_attr->recv_cq); in create_qp_common()
[all …]
H A Dsrq.c181 to_mcq(init_attr->ext.cq)->mcq.cqn : 0; in mlx4_ib_create_srq()
H A Dmlx4_ib.h693 static inline struct mlx4_ib_cq *to_mcq(struct ib_cq *ibcq) in to_mcq() function
H A Dmain.c3086 send_mcq = to_mcq(mqp->ibqp.send_cq); in mlx4_ib_handle_catas_error()
3104 recv_mcq = to_mcq(mqp->ibqp.recv_cq); in mlx4_ib_handle_catas_error()