| /linux/drivers/net/ethernet/qlogic/netxen/ |
| H A D | netxen_nic_ctx.c | 270 size_t rq_size, rsp_size; in nx_fw_cmd_create_rx_ctx() local 280 rq_size = in nx_fw_cmd_create_rx_ctx() 285 addr = dma_alloc_coherent(&adapter->pdev->dev, rq_size, in nx_fw_cmd_create_rx_ctx() 350 cmd.req.arg3 = rq_size; in nx_fw_cmd_create_rx_ctx() 394 dma_free_coherent(&adapter->pdev->dev, rq_size, prq, hostrq_phys_addr); in nx_fw_cmd_create_rx_ctx() 424 size_t rq_size, rsp_size; in nx_fw_cmd_create_tx_ctx() local 433 rq_size = SIZEOF_HOSTRQ_TX(nx_hostrq_tx_ctx_t); in nx_fw_cmd_create_tx_ctx() 434 rq_addr = dma_alloc_coherent(&adapter->pdev->dev, rq_size, in nx_fw_cmd_create_tx_ctx() 476 cmd.req.arg3 = rq_size; in nx_fw_cmd_create_tx_ctx() 500 dma_free_coherent(&adapter->pdev->dev, rq_size, rq_addr, rq_phys_addr); in nx_fw_cmd_create_tx_ctx()
|
| /linux/drivers/infiniband/hw/irdma/ |
| H A D | puda.c | 127 rsrc->rx_wqe_idx = ((rsrc->rx_wqe_idx + 1) % rsrc->rq_size); in irdma_puda_replenish_rq() 649 u32 sq_size, rq_size; in irdma_puda_qp_create() local 653 rq_size = rsrc->rq_size * IRDMA_QP_WQE_MIN_SIZE; in irdma_puda_qp_create() 654 rsrc->qpmem.size = ALIGN((sq_size + rq_size + (IRDMA_SHADOW_AREA_SIZE << 3) + IRDMA_QP_CTX_SIZE), in irdma_puda_qp_create() 665 qp->hw_rq_size = irdma_get_encoded_wqe_size(rsrc->rq_size, IRDMA_QUEUE_TYPE_SQ_RQ); in irdma_puda_qp_create() 675 ukqp->shadow_area = ukqp->rq_base[rsrc->rq_size].elem; in irdma_puda_qp_create() 677 qp->shadow_area_pa = qp->rq_pa + rq_size; in irdma_puda_qp_create() 685 ukqp->rq_size = rsrc->rq_size; in irdma_puda_qp_create() 688 IRDMA_RING_INIT(ukqp->rq_ring, ukqp->rq_size); in irdma_puda_qp_create() 1009 rqwridsize = info->rq_size * 8; in irdma_puda_create_rsrc() [all …]
|
| H A D | puda.h | 91 u32 rq_size; member 114 u32 rq_size; member
|
| H A D | user.h | 508 u32 rq_size; member 555 u32 rq_size; member 593 int irdma_get_rqdepth(struct irdma_uk_attrs *uk_attrs, u32 rq_size, u8 shift,
|
| H A D | uk.c | 1514 int irdma_get_rqdepth(struct irdma_uk_attrs *uk_attrs, u32 rq_size, u8 shift, in irdma_get_rqdepth() argument 1519 irdma_round_up_wq(((u64)rq_size << shift) + IRDMA_RQ_RSVD); in irdma_get_rqdepth() 1689 status = irdma_get_rqdepth(ukinfo->uk_attrs, ukinfo->rq_size, in irdma_uk_calc_depth_shift_rq() 1738 qp->rq_size = info->rq_size; in irdma_uk_qp_init() 1742 IRDMA_RING_INIT(qp->rq_ring, qp->rq_size); in irdma_uk_qp_init()
|
| /linux/drivers/scsi/ |
| H A D | stex.c | 346 u16 rq_size; member 360 u16 rq_size; member 411 struct req_msg *req = hba->dma_mem + hba->req_head * hba->rq_size; in stex_alloc_req() 422 hba->req_head * hba->rq_size + sizeof(struct st_msg_header)); in stex_ss_alloc_req() 548 addr = hba->dma_handle + hba->req_head * hba->rq_size; in stex_ss_send_cmd() 1062 h->req_sz = cpu_to_le16(hba->rq_size); in stex_common_handshake() 1074 status_phys = hba->dma_handle + (hba->rq_count+1) * hba->rq_size; in stex_common_handshake() 1153 h->req_sz = cpu_to_le16(hba->rq_size); in stex_ss_handshake() 1544 .rq_size = 1048, 1557 .rq_size = 1048, [all …]
|
| /linux/drivers/scsi/bnx2i/ |
| H A D | bnx2i_init.c | 65 unsigned int rq_size = BNX2I_RQ_WQES_DEFAULT; variable 66 module_param(rq_size, int, 0664); 67 MODULE_PARM_DESC(rq_size, "Configure RQ size");
|
| H A D | bnx2i.h | 796 extern unsigned int rq_size;
|
| /linux/drivers/net/ethernet/qlogic/qlcnic/ |
| H A D | qlcnic_ctx.c | 250 size_t rq_size, rsp_size; in qlcnic_82xx_fw_cmd_create_rx_ctx() local 260 rq_size = SIZEOF_HOSTRQ_RX(struct qlcnic_hostrq_rx_ctx, nrds_rings, in qlcnic_82xx_fw_cmd_create_rx_ctx() 265 addr = dma_alloc_coherent(&adapter->pdev->dev, rq_size, in qlcnic_82xx_fw_cmd_create_rx_ctx() 343 cmd.req.arg[3] = rq_size; in qlcnic_82xx_fw_cmd_create_rx_ctx() 387 dma_free_coherent(&adapter->pdev->dev, rq_size, prq, hostrq_phys_addr); in qlcnic_82xx_fw_cmd_create_rx_ctx() 426 size_t rq_size, rsp_size; in qlcnic_82xx_fw_cmd_create_tx_ctx() local 435 rq_size = SIZEOF_HOSTRQ_TX(struct qlcnic_hostrq_tx_ctx); in qlcnic_82xx_fw_cmd_create_tx_ctx() 436 rq_addr = dma_alloc_coherent(&adapter->pdev->dev, rq_size, in qlcnic_82xx_fw_cmd_create_tx_ctx() 489 cmd.req.arg[3] = rq_size; in qlcnic_82xx_fw_cmd_create_tx_ctx() 518 dma_free_coherent(&adapter->pdev->dev, rq_size, rq_addr, rq_phys_addr); in qlcnic_82xx_fw_cmd_create_tx_ctx()
|
| /linux/drivers/infiniband/ulp/srpt/ |
| H A D | ib_srpt.c | 1856 WARN_ON(ch->rq_size < 1); in srpt_create_ch_ib() 1864 ch->cq = ib_cq_pool_get(sdev->device, ch->rq_size + sq_size, -1, in srpt_create_ch_ib() 1869 ch->rq_size + sq_size, ch->cq); in srpt_create_ch_ib() 1872 ch->cq_size = ch->rq_size + sq_size; in srpt_create_ch_ib() 1895 qp_init->cap.max_recv_wr = ch->rq_size; in srpt_create_ch_ib() 1933 for (i = 0; i < ch->rq_size; i++) in srpt_create_ch_ib() 2178 ch->sport->sdev, ch->rq_size, in srpt_release_channel_work() 2184 sdev, ch->rq_size, in srpt_release_channel_work() 2301 ch->rq_size = min(MAX_SRPT_RQ_SIZE, sdev->device->attrs.max_qp_wr); in srpt_cm_req_recv() 2312 srpt_alloc_ioctx_ring(ch->sport->sdev, ch->rq_size, in srpt_cm_req_recv() [all …]
|
| H A D | ib_srpt.h | 309 int rq_size; member
|
| /linux/include/uapi/rdma/ |
| H A D | cxgb4-abi.h | 83 __u32 rq_size; member
|
| /linux/drivers/infiniband/hw/efa/ |
| H A D | efa_verbs.c | 514 qp->rq_cpu_addr, qp->rq_size, in efa_destroy_qp() 517 qp->rq_size, DMA_TO_DEVICE); in efa_destroy_qp() 582 if (qp->rq_size) { in qp_mmap_entries_setup() 598 address, qp->rq_size, in qp_mmap_entries_setup() 604 resp->rq_mmap_size = qp->rq_size; in qp_mmap_entries_setup() 767 qp->rq_size = PAGE_ALIGN(create_qp_params.rq_ring_size_in_bytes); in efa_create_qp() 768 if (qp->rq_size) { in efa_create_qp() 770 qp->rq_size, DMA_TO_DEVICE); in efa_create_qp() 778 qp->rq_cpu_addr, qp->rq_size, &qp->rq_dma_addr); in efa_create_qp() 831 if (qp->rq_size) in efa_create_qp() [all …]
|
| H A D | efa.h | 117 size_t rq_size; member
|
| /linux/drivers/infiniband/hw/erdma/ |
| H A D | erdma_qp.c | 261 memset(qp->kern_qp.rwr_tbl, 0, qp->attrs.rq_size * sizeof(u64)); in erdma_reset_qp() 263 memset(qp->kern_qp.rq_buf, 0, qp->attrs.rq_size << RQE_SHIFT); in erdma_reset_qp() 706 qp->attrs.rq_size, RQE_SHIFT); in erdma_post_recv_one() 724 qp->kern_qp.rwr_tbl[qp->kern_qp.rq_pi & (qp->attrs.rq_size - 1)] = in erdma_post_recv_one()
|
| H A D | erdma_verbs.c | 55 ilog2(qp->attrs.rq_size)) | in create_qp_cmd() 524 qp->attrs.rq_size << RQE_SHIFT, in free_kernel_qp() 551 kqp->rwr_tbl = vmalloc_array(qp->attrs.rq_size, sizeof(u64)); in init_kernel_qp() 566 size = qp->attrs.rq_size << RQE_SHIFT; in init_kernel_qp() 946 qp->attrs.rq_size * RQE_SIZE)) in init_user_qp() 959 qp->attrs.rq_size << RQE_SHIFT, 0, va + rq_offset, in init_user_qp() 1039 qp->attrs.rq_size = roundup_pow_of_two(attrs->cap.max_recv_wr); in erdma_create_qp() 1055 uresp.num_rqe = qp->attrs.rq_size; in erdma_create_qp() 1856 qp_attr->cap.max_recv_wr = qp->attrs.rq_size; in erdma_query_qp()
|
| H A D | erdma_cq.c | 171 depth = qp->attrs.rq_size; in erdma_poll_one_cqe()
|
| H A D | erdma_verbs.h | 285 u32 rq_size; member
|
| /linux/drivers/vfio/pci/mlx5/ |
| H A D | cmd.c | 1413 u64 rq_size) in mlx5vf_alloc_qp_recv_resources() argument 1415 unsigned int npages = DIV_ROUND_UP_ULL(rq_size, PAGE_SIZE); in mlx5vf_alloc_qp_recv_resources() 1507 u64 rq_size = SZ_2M; in mlx5vf_start_page_tracker() local 1526 if (rq_size < 4ULL * max_msg_size) in mlx5vf_start_page_tracker() 1527 rq_size = 4ULL * max_msg_size; in mlx5vf_start_page_tracker() 1540 max_recv_wr = DIV_ROUND_UP_ULL(rq_size, max_msg_size); in mlx5vf_start_page_tracker() 1564 rq_size); in mlx5vf_start_page_tracker()
|
| /linux/drivers/infiniband/sw/siw/ |
| H A D | siw_verbs.c | 415 qp->attrs.rq_size = 0; in siw_create_qp() 429 qp->attrs.rq_size = num_rqe; in siw_create_qp() 531 qp_attr->cap.max_recv_wr = qp->attrs.rq_size; in siw_query_qp() 1015 if (qp->srq || qp->attrs.rq_size == 0) { in siw_post_receive() 1075 u32 idx = qp->rq_put % qp->attrs.rq_size; in siw_post_receive()
|
| H A D | siw_qp.c | 1267 while (qp->attrs.rq_size) { in siw_rq_flush() 1269 &qp->recvq[qp->rq_get % qp->attrs.rq_size]; in siw_rq_flush()
|
| H A D | siw.h | 258 u32 rq_size; member
|
| /linux/block/ |
| H A D | blk-mq.c | 3602 size_t rq_size, left; in blk_mq_alloc_rqs() local 3611 rq_size = round_up(sizeof(struct request) + set->cmd_size, in blk_mq_alloc_rqs() 3613 left = rq_size * depth; in blk_mq_alloc_rqs() 3632 if (order_to_size(this_order) < rq_size) in blk_mq_alloc_rqs() 3648 entries_per_page = order_to_size(this_order) / rq_size; in blk_mq_alloc_rqs() 3650 left -= to_do * rq_size; in blk_mq_alloc_rqs() 3660 p += rq_size; in blk_mq_alloc_rqs()
|
| /linux/drivers/net/ethernet/microsoft/mana/ |
| H A D | mana_en.c | 2653 u32 cq_size, rq_size; in mana_create_rxq() local 2677 err = mana_alloc_rx_wqe(apc, rxq, &rq_size, &cq_size); in mana_create_rxq() 2681 rq_size = MANA_PAGE_ALIGN(rq_size); in mana_create_rxq() 2688 spec.queue_size = rq_size; in mana_create_rxq()
|
| /linux/drivers/infiniband/hw/bnxt_re/ |
| H A D | roce_hsi.h | 483 __le32 rq_size; member 675 __le32 rq_size; member 831 __le32 rq_size; member 1778 __le32 rq_size; member
|