Lines Matching +refs:get +refs:addr +refs:attrs

122 	struct sockaddr_storage addr;  member
152 .get = param_get_int,
273 ib_dma_unmap_page(ndev->device, sge->addr, in nvmet_rdma_free_inline_pages()
302 sge->addr = ib_dma_map_page(ndev->device, in nvmet_rdma_alloc_inline_pages()
304 if (ib_dma_mapping_error(ndev->device, sge->addr)) in nvmet_rdma_alloc_inline_pages()
315 ib_dma_unmap_page(ndev->device, sge->addr, in nvmet_rdma_alloc_inline_pages()
331 c->sge[0].addr = ib_dma_map_single(ndev->device, c->nvme_cmd, in nvmet_rdma_alloc_cmd()
333 if (ib_dma_mapping_error(ndev->device, c->sge[0].addr)) in nvmet_rdma_alloc_cmd()
351 ib_dma_unmap_single(ndev->device, c->sge[0].addr, in nvmet_rdma_alloc_cmd()
365 ib_dma_unmap_single(ndev->device, c->sge[0].addr, in nvmet_rdma_free_cmd()
415 r->send_sge.addr = ib_dma_map_single(ndev->device, r->req.cqe, in nvmet_rdma_alloc_rsp()
417 if (ib_dma_mapping_error(ndev->device, r->send_sge.addr)) in nvmet_rdma_alloc_rsp()
448 ib_dma_unmap_single(ndev->device, r->send_sge.addr, in nvmet_rdma_free_rsp()
509 cmd->sge[0].addr, cmd->sge[0].length, in nvmet_rdma_post_recv()
635 static int nvmet_rdma_rw_ctx_init(struct nvmet_rdma_rsp *rsp, u64 addr, u32 key, in nvmet_rdma_rw_ctx_init() argument
646 addr, key, nvmet_data_dir(req)); in nvmet_rdma_rw_ctx_init()
649 req->sg, req->sg_cnt, 0, addr, key, in nvmet_rdma_rw_ctx_init()
746 rsp->send_sge.addr, rsp->send_sge.length, in nvmet_rdma_queue_response()
858 u64 off = le64_to_cpu(sgl->addr); in nvmet_rdma_map_sgl_inline()
885 u64 addr = le64_to_cpu(sgl->addr); in nvmet_rdma_map_sgl_keyed() local
903 ret = nvmet_rdma_rw_ctx_init(rsp, addr, key, &sig_attrs); in nvmet_rdma_map_sgl_keyed()
984 cmd->cmd->sge[0].addr, cmd->cmd->sge[0].length, in nvmet_rdma_handle_command()
987 cmd->send_sge.addr, cmd->send_sge.length, in nvmet_rdma_handle_command()
1142 if (!ndev->device->attrs.max_srq_wr || !ndev->device->attrs.max_srq) { in nvmet_rdma_init_srqs()
1151 ndev->srq_size = min(ndev->device->attrs.max_srq_wr, in nvmet_rdma_init_srqs()
1154 ndev->device->attrs.max_srq); in nvmet_rdma_init_srqs()
1214 inline_sge_count = max(cm_id->device->attrs.max_sge_rd, in nvmet_rdma_find_get_device()
1215 cm_id->device->attrs.max_recv_sge) - 1; in nvmet_rdma_find_get_device()
1226 if (nport->pi_enable && !(cm_id->device->attrs.kernel_cap_flags & in nvmet_rdma_find_get_device()
1292 qp_attr.cap.max_send_sge = max(ndev->device->attrs.max_sge_rd, in nvmet_rdma_create_queue_ib()
1293 ndev->device->attrs.max_send_sge); in nvmet_rdma_create_queue_ib()
1555 queue->dev->device->attrs.max_qp_init_rd_atom); in nvmet_rdma_cm_accept()
1869 struct sockaddr *addr = (struct sockaddr *)&port->addr; in nvmet_rdma_enable_port() local
1890 ret = rdma_bind_addr(cm_id, addr); in nvmet_rdma_enable_port()
1892 pr_err("binding CM ID to %pISpcs failed (%d)\n", addr, ret); in nvmet_rdma_enable_port()
1898 pr_err("listening to %pISpcs failed (%d)\n", addr, ret); in nvmet_rdma_enable_port()
1960 nport->disc_addr.trsvcid, &port->addr); in nvmet_rdma_add_port()
1973 (struct sockaddr *)&port->addr); in nvmet_rdma_add_port()
1997 if (inet_addr_is_any((struct sockaddr *)&cm_id->route.addr.src_addr)) { in nvmet_rdma_disc_port_addr()
2001 struct sockaddr *addr = (void *)&req_cm_id->route.addr.src_addr; in nvmet_rdma_disc_port_addr() local
2003 sprintf(traddr, "%pISc", addr); in nvmet_rdma_disc_port_addr()