| /linux/drivers/infiniband/hw/irdma/ |
| H A D | ig3rdma_hw.c | 116 dev->hw_attrs.uk_attrs.hw_rev = IRDMA_GEN_3; in ig3rdma_init_hw() 117 dev->hw_attrs.uk_attrs.max_hw_wq_frags = IG3RDMA_MAX_WQ_FRAGMENT_COUNT; in ig3rdma_init_hw() 118 dev->hw_attrs.uk_attrs.max_hw_read_sges = IG3RDMA_MAX_SGE_RD; in ig3rdma_init_hw() 119 dev->hw_attrs.uk_attrs.max_hw_sq_chunk = IRDMA_MAX_QUANTA_PER_WR; in ig3rdma_init_hw() 122 dev->hw_attrs.uk_attrs.feature_flags |= IRDMA_FEATURE_64_BYTE_CQE; in ig3rdma_init_hw() 123 dev->hw_attrs.uk_attrs.feature_flags |= IRDMA_FEATURE_CQE_TIMESTAMPING; in ig3rdma_init_hw() 125 dev->hw_attrs.uk_attrs.feature_flags |= IRDMA_FEATURE_SRQ; in ig3rdma_init_hw() 126 dev->hw_attrs.uk_attrs.feature_flags |= IRDMA_FEATURE_RTS_AE | in ig3rdma_init_hw() 133 dev->hw_attrs.uk_attrs.min_hw_wq_size = IG3RDMA_MIN_WQ_SIZE; in ig3rdma_init_hw() 134 dev->hw_attrs.uk_attrs.max_hw_srq_quanta = IRDMA_SRQ_MAX_QUANTA; in ig3rdma_init_hw() [all …]
|
| H A D | uk.c | 137 avail_quanta = qp->uk_attrs->max_hw_sq_chunk - in irdma_qp_get_next_send_wqe() 139 qp->uk_attrs->max_hw_sq_chunk); in irdma_qp_get_next_send_wqe() 163 if (qp->uk_attrs->hw_rev == IRDMA_GEN_1 && quanta == 1 && in irdma_qp_get_next_send_wqe() 287 if (qp->uk_attrs->hw_rev >= IRDMA_GEN_2 && !(frag_cnt & 0x01) && in irdma_uk_rdma_write() 291 if (qp->uk_attrs->hw_rev == IRDMA_GEN_2) in irdma_uk_rdma_write() 448 if (srq->uk_attrs->hw_rev >= IRDMA_GEN_2 && !(info->num_sges & 0x01) && in irdma_uk_srq_post_receive() 452 if (srq->uk_attrs->hw_rev == IRDMA_GEN_2) in irdma_uk_srq_post_receive() 521 if (qp->uk_attrs->hw_rev >= IRDMA_GEN_2 && in irdma_uk_rdma_read() 525 if (qp->uk_attrs->hw_rev == IRDMA_GEN_2) in irdma_uk_rdma_read() 610 if (qp->uk_attrs->hw_rev >= IRDMA_GEN_2 && !(frag_cnt & 0x01) && in irdma_uk_send() [all …]
|
| H A D | i40iw_hw.c | 243 dev->hw_attrs.uk_attrs.max_hw_wq_frags = I40IW_MAX_WQ_FRAGMENT_COUNT; in i40iw_init_hw() 244 dev->hw_attrs.uk_attrs.max_hw_read_sges = I40IW_MAX_SGE_RD; in i40iw_init_hw() 246 dev->hw_attrs.uk_attrs.max_hw_inline = I40IW_MAX_INLINE_DATA_SIZE; in i40iw_init_hw() 251 dev->hw_attrs.uk_attrs.max_hw_rq_quanta = I40IW_QP_SW_MAX_RQ_QUANTA; in i40iw_init_hw() 252 dev->hw_attrs.uk_attrs.max_hw_wq_quanta = I40IW_QP_SW_MAX_WQ_QUANTA; in i40iw_init_hw() 253 dev->hw_attrs.uk_attrs.max_hw_sq_chunk = I40IW_MAX_QUANTA_PER_WR; in i40iw_init_hw() 259 dev->hw_attrs.uk_attrs.min_hw_wq_size = I40IW_MIN_WQ_SIZE; in i40iw_init_hw()
|
| H A D | icrdma_hw.c | 71 if (dev->hw_attrs.uk_attrs.hw_rev != IRDMA_GEN_1) in icrdma_ena_irq() 84 if (dev->hw_attrs.uk_attrs.hw_rev != IRDMA_GEN_1) in icrdma_disable_irq() 201 dev->hw_attrs.uk_attrs.min_hw_wq_size = ICRDMA_MIN_WQ_SIZE; in icrdma_init_hw() 202 dev->hw_attrs.uk_attrs.max_hw_sq_chunk = IRDMA_MAX_QUANTA_PER_WR; in icrdma_init_hw() 203 dev->hw_attrs.uk_attrs.feature_flags |= IRDMA_FEATURE_RTS_AE | in icrdma_init_hw()
|
| H A D | user.h | 457 struct irdma_uk_attrs *uk_attrs; member 473 struct irdma_uk_attrs *uk_attrs; member 497 struct irdma_uk_attrs *uk_attrs; member 547 struct irdma_uk_attrs *uk_attrs; member 589 void irdma_get_wqe_shift(struct irdma_uk_attrs *uk_attrs, u32 sge, 591 int irdma_get_sqdepth(struct irdma_uk_attrs *uk_attrs, u32 sq_size, u8 shift, 593 int irdma_get_rqdepth(struct irdma_uk_attrs *uk_attrs, u32 rq_size, u8 shift, 595 int irdma_get_srqdepth(struct irdma_uk_attrs *uk_attrs, u32 srq_size, u8 shift,
|
| H A D | verbs.c | 30 if (hw_attrs->uk_attrs.hw_rev < IRDMA_GEN_3) in irdma_query_device() 40 props->max_send_sge = hw_attrs->uk_attrs.max_hw_wq_frags; in irdma_query_device() 41 props->max_recv_sge = hw_attrs->uk_attrs.max_hw_wq_frags; in irdma_query_device() 45 if (hw_attrs->uk_attrs.hw_rev >= IRDMA_GEN_3) in irdma_query_device() 48 props->max_sge_rd = hw_attrs->uk_attrs.max_hw_read_sges; in irdma_query_device() 63 props->max_srq_sge = hw_attrs->uk_attrs.max_hw_wq_frags; in irdma_query_device() 64 if (hw_attrs->uk_attrs.feature_flags & IRDMA_FEATURE_ATOMIC_OPS) in irdma_query_device() 69 if (hw_attrs->uk_attrs.hw_rev >= IRDMA_GEN_3) { in irdma_query_device() 74 if (hw_attrs->uk_attrs.hw_rev >= IRDMA_GEN_3) in irdma_query_device() 294 struct irdma_uk_attrs *uk_attrs = &iwdev->rf->sc_dev.hw_attrs.uk_attrs; in irdma_alloc_ucontext() local [all …]
|
| H A D | puda.c | 98 if (qp->qp_uk.uk_attrs->hw_rev == IRDMA_GEN_1) { in irdma_puda_post_recvbuf() 236 if (cq->dev->hw_attrs.uk_attrs.hw_rev >= IRDMA_GEN_2) in irdma_puda_poll_info() 279 if (cq->dev->hw_attrs.uk_attrs.hw_rev >= IRDMA_GEN_2) in irdma_puda_poll_info() 305 if (cq->dev->hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_1) { in irdma_puda_poll_info() 451 if (qp->dev->hw_attrs.uk_attrs.hw_rev >= IRDMA_GEN_2) { in irdma_puda_send() 537 if (rsrc->dev->hw_attrs.uk_attrs.hw_rev >= IRDMA_GEN_2) { in irdma_puda_send_buf() 579 if (qp->dev->hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_1) in irdma_puda_qp_setctx() 676 ukqp->uk_attrs = &qp->dev->hw_attrs.uk_attrs; in irdma_puda_qp_create() 1059 if (dev->hw_attrs.uk_attrs.hw_rev >= IRDMA_GEN_2) { in irdma_puda_create_rsrc() 1124 if (qp->dev->hw_attrs.uk_attrs.hw_rev >= IRDMA_GEN_2) { in irdma_ilq_putback_rcvbuf() [all …]
|
| H A D | ctrl.c | 77 if (vsi->dev->hw_attrs.uk_attrs.hw_rev >= IRDMA_GEN_3) { in irdma_set_qos_info() 94 if (vsi->dev->hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_1) in irdma_set_qos_info() 390 info->pd->dev->hw_attrs.uk_attrs.max_hw_wq_frags || in irdma_sc_qp_init() 392 info->pd->dev->hw_attrs.uk_attrs.max_hw_wq_frags) in irdma_sc_qp_init() 425 if (qp->qp_uk.uk_attrs->hw_rev == IRDMA_GEN_1 && qp->pd->abi_ver > 4) in irdma_sc_qp_init() 1145 if (qp->dev->hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_2) in irdma_sc_qp_setctx_roce() 1303 if (dev->hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_1) in irdma_sc_qp_setctx() 1332 if (dev->hw_attrs.uk_attrs.hw_rev >= IRDMA_GEN_2) in irdma_sc_qp_setctx() 1373 if (dev->hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_1) { in irdma_sc_qp_setctx() 1815 if (qp->qp_uk.uk_attrs->hw_rev == IRDMA_GEN_1) { in irdma_sc_send_lsmm() [all …]
|
| H A D | virtchnl.c | 70 dev->hw_attrs.uk_attrs.hw_rev = info->hw_rev; in irdma_sc_vchnl_init() 87 dev->hw_attrs.uk_attrs.hw_rev = dev->vc_caps.hw_rev; in irdma_sc_vchnl_init() 502 if (dev->hw_attrs.uk_attrs.hw_rev >= IRDMA_GEN_3) { in irdma_vchnl_req_get_hmc_fcn() 516 if (dev->hw_attrs.uk_attrs.hw_rev >= IRDMA_GEN_3) { in irdma_vchnl_req_get_hmc_fcn()
|
| H A D | irdma.h | 142 struct irdma_uk_attrs uk_attrs; member
|
| H A D | pble.c | 196 if (dev->hw_attrs.uk_attrs.hw_rev >= IRDMA_GEN_3) in irdma_get_type() 290 dev->hw_attrs.uk_attrs.hw_rev >= IRDMA_GEN_3) { in add_pble_prm()
|
| H A D | i40iw_if.c | 79 rf->sc_dev.hw_attrs.uk_attrs.hw_rev = IRDMA_GEN_1; in i40iw_fill_device_info()
|
| H A D | ig3rdma_if.c | 62 *rdma_ver = rf->sc_dev.hw_attrs.uk_attrs.hw_rev; in ig3rdma_vchnl_init()
|
| H A D | utils.c | 555 if (dev->hw_attrs.uk_attrs.hw_rev >= IRDMA_GEN_3) in irdma_cleanup_pending_cqp_op() 1127 if (rf->sc_dev.hw_attrs.uk_attrs.hw_rev < IRDMA_GEN_3) in irdma_free_gsi_qp_rsrc() 1525 if (buf->vsi->dev->hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_1) in irdma_ieq_update_tcpip_info() 1601 if (buf->vsi->dev->hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_1) in irdma_puda_get_tcpip_info() 1643 if (sc_vsi->dev->hw_attrs.uk_attrs.hw_rev >= IRDMA_GEN_2) in irdma_hw_stats_timeout()
|
| H A D | icrdma_if.c | 215 rf->sc_dev.hw_attrs.uk_attrs.hw_rev = IRDMA_GEN_2; in icrdma_fill_device_info()
|
| H A D | hw.c | 210 if (qp->qp_uk.uk_attrs->hw_rev >= IRDMA_GEN_3) { in irdma_set_flush_fields() 1792 if (iwdev->rf->sc_dev.hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_1) in irdma_rt_deinit_hw() 1969 if (iwdev->rf->sc_dev.hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_1) in irdma_rt_init_hw() 2147 rf->max_cqe = rf->sc_dev.hw_attrs.uk_attrs.max_hw_cq_size; in irdma_initialize_hw_rsrc()
|
| H A D | hmc.c | 232 if (dev->hw_attrs.uk_attrs.hw_rev >= IRDMA_GEN_3 && in irdma_sc_create_hmc_obj()
|
| H A D | cm.c | 3158 if (iwdev->rf->sc_dev.hw_attrs.uk_attrs.hw_rev >= IRDMA_GEN_2) { in irdma_receive_ilq() 3751 if (dev->hw_attrs.uk_attrs.feature_flags & IRDMA_FEATURE_RTS_AE) { in irdma_accept() 4170 if (dev->hw_attrs.uk_attrs.feature_flags & IRDMA_FEATURE_RTS_AE) { in irdma_cm_event_connected()
|