Lines Matching +refs:get +refs:addr +refs:attrs

122 	struct sockaddr_storage addr;  member
152 .get = param_get_int,
266 ib_dma_unmap_page(ndev->device, sge->addr, in nvmet_rdma_free_inline_pages()
295 sge->addr = ib_dma_map_page(ndev->device, in nvmet_rdma_alloc_inline_pages()
297 if (ib_dma_mapping_error(ndev->device, sge->addr)) in nvmet_rdma_alloc_inline_pages()
308 ib_dma_unmap_page(ndev->device, sge->addr, in nvmet_rdma_alloc_inline_pages()
324 c->sge[0].addr = ib_dma_map_single(ndev->device, c->nvme_cmd, in nvmet_rdma_alloc_cmd()
326 if (ib_dma_mapping_error(ndev->device, c->sge[0].addr)) in nvmet_rdma_alloc_cmd()
344 ib_dma_unmap_single(ndev->device, c->sge[0].addr, in nvmet_rdma_alloc_cmd()
358 ib_dma_unmap_single(ndev->device, c->sge[0].addr, in nvmet_rdma_free_cmd()
408 r->send_sge.addr = ib_dma_map_single(ndev->device, r->req.cqe, in nvmet_rdma_alloc_rsp()
410 if (ib_dma_mapping_error(ndev->device, r->send_sge.addr)) in nvmet_rdma_alloc_rsp()
442 ib_dma_unmap_single(ndev->device, r->send_sge.addr, in nvmet_rdma_free_rsp()
500 cmd->sge[0].addr, cmd->sge[0].length, in nvmet_rdma_post_recv()
626 static int nvmet_rdma_rw_ctx_init(struct nvmet_rdma_rsp *rsp, u64 addr, u32 key, in nvmet_rdma_rw_ctx_init() argument
637 addr, key, nvmet_data_dir(req)); in nvmet_rdma_rw_ctx_init()
640 req->sg, req->sg_cnt, 0, addr, key, in nvmet_rdma_rw_ctx_init()
737 rsp->send_sge.addr, rsp->send_sge.length, in nvmet_rdma_queue_response()
849 u64 off = le64_to_cpu(sgl->addr); in nvmet_rdma_map_sgl_inline()
876 u64 addr = le64_to_cpu(sgl->addr); in nvmet_rdma_map_sgl_keyed() local
894 ret = nvmet_rdma_rw_ctx_init(rsp, addr, key, &sig_attrs); in nvmet_rdma_map_sgl_keyed()
973 cmd->cmd->sge[0].addr, cmd->cmd->sge[0].length, in nvmet_rdma_handle_command()
976 cmd->send_sge.addr, cmd->send_sge.length, in nvmet_rdma_handle_command()
1145 if (!ndev->device->attrs.max_srq_wr || !ndev->device->attrs.max_srq) { in nvmet_rdma_init_srqs()
1154 ndev->srq_size = min(ndev->device->attrs.max_srq_wr, in nvmet_rdma_init_srqs()
1157 ndev->device->attrs.max_srq); in nvmet_rdma_init_srqs()
1217 inline_sge_count = max(cm_id->device->attrs.max_sge_rd, in nvmet_rdma_find_get_device()
1218 cm_id->device->attrs.max_recv_sge) - 1; in nvmet_rdma_find_get_device()
1229 if (nport->pi_enable && !(cm_id->device->attrs.kernel_cap_flags & in nvmet_rdma_find_get_device()
1295 qp_attr.cap.max_send_sge = max(ndev->device->attrs.max_sge_rd, in nvmet_rdma_create_queue_ib()
1296 ndev->device->attrs.max_send_sge); in nvmet_rdma_create_queue_ib()
1556 queue->dev->device->attrs.max_qp_init_rd_atom); in nvmet_rdma_cm_accept()
1866 struct sockaddr *addr = (struct sockaddr *)&port->addr; in nvmet_rdma_enable_port() local
1887 ret = rdma_bind_addr(cm_id, addr); in nvmet_rdma_enable_port()
1889 pr_err("binding CM ID to %pISpcs failed (%d)\n", addr, ret); in nvmet_rdma_enable_port()
1895 pr_err("listening to %pISpcs failed (%d)\n", addr, ret); in nvmet_rdma_enable_port()
1965 nport->disc_addr.trsvcid, &port->addr); in nvmet_rdma_add_port()
1978 (struct sockaddr *)&port->addr); in nvmet_rdma_add_port()
2002 if (inet_addr_is_any((struct sockaddr *)&cm_id->route.addr.src_addr)) { in nvmet_rdma_disc_port_addr()
2006 struct sockaddr *addr = (void *)&req_cm_id->route.addr.src_addr; in nvmet_rdma_disc_port_addr() local
2008 sprintf(traddr, "%pISc", addr); in nvmet_rdma_disc_port_addr()
2022 (struct sockaddr *)&queue->cm_id->route.addr.dst_addr); in nvmet_rdma_host_port_addr()