/linux/net/sunrpc/xprtrdma/ |
H A D | svc_rdma_rw.c | 365 struct ib_send_wr *first_wr; in svc_rdma_post_chunk_ctxt() local 376 first_wr = NULL; in svc_rdma_post_chunk_ctxt() 382 first_wr = rdma_rw_ctx_wrs(&ctxt->rw_ctx, rdma->sc_qp, in svc_rdma_post_chunk_ctxt() 383 rdma->sc_port_num, cqe, first_wr); in svc_rdma_post_chunk_ctxt() 391 ret = ib_post_send(rdma->sc_qp, first_wr, &bad_wr); in svc_rdma_post_chunk_ctxt() 409 if (bad_wr != first_wr) in svc_rdma_post_chunk_ctxt() 686 struct ib_send_wr *first_wr; in svc_rdma_prepare_reply_chunk() local 702 first_wr = sctxt->sc_wr_chain; in svc_rdma_prepare_reply_chunk() 708 first_wr = rdma_rw_ctx_wrs(&rwc->rw_ctx, rdma->sc_qp, in svc_rdma_prepare_reply_chunk() 709 rdma->sc_port_num, cqe, first_wr); in svc_rdma_prepare_reply_chunk() [all...] |
H A D | svc_rdma_sendto.c | 351 struct ib_send_wr *first_wr = ctxt->sc_wr_chain; in svc_rdma_post_send() local 353 const struct ib_send_wr *bad_wr = first_wr; in svc_rdma_post_send() 385 ret = ib_post_send(rdma->sc_qp, first_wr, &bad_wr); in svc_rdma_post_send() 393 if (bad_wr == first_wr) { in svc_rdma_post_send()
|
/linux/drivers/infiniband/core/ |
H A D | rw.c | 500 struct ib_send_wr *first_wr, *last_wr; in rdma_rw_ctx_wrs() local 513 first_wr = &ctx->reg[0].inv_wr; in rdma_rw_ctx_wrs() 515 first_wr = &ctx->reg[0].reg_wr.wr; in rdma_rw_ctx_wrs() 519 first_wr = &ctx->map.wrs[0].wr; in rdma_rw_ctx_wrs() 523 first_wr = &ctx->single.wr.wr; in rdma_rw_ctx_wrs() 537 return first_wr; in rdma_rw_ctx_wrs() 558 struct ib_send_wr *first_wr; in rdma_rw_ctx_post() local 560 first_wr = rdma_rw_ctx_wrs(ctx, qp, port_num, cqe, chain_wr); in rdma_rw_ctx_post() 561 return ib_post_send(qp, first_wr, NULL); in rdma_rw_ctx_post()
|
/linux/drivers/infiniband/ulp/iser/ |
H A D | iser_verbs.c | 860 struct ib_send_wr *first_wr; in iser_post_send() local 875 first_wr = &tx_desc->inv_wr; in iser_post_send() 877 first_wr = &tx_desc->reg_wr.wr; in iser_post_send() 879 first_wr = wr; in iser_post_send() 881 ret = ib_post_send(ib_conn->qp, first_wr, NULL); in iser_post_send()
|
/linux/drivers/nvme/target/ |
H A D | rdma.c | 714 struct ib_send_wr *first_wr; in nvmet_rdma_queue_response() local 725 first_wr = rdma_rw_ctx_wrs(&rsp->rw, cm_id->qp, in nvmet_rdma_queue_response() 728 first_wr = rdma_rw_ctx_wrs(&rsp->rw, cm_id->qp, in nvmet_rdma_queue_response() 731 first_wr = &rsp->send_wr; in nvmet_rdma_queue_response() 740 if (unlikely(ib_post_send(cm_id->qp, first_wr, NULL))) { in nvmet_rdma_queue_response()
|
/linux/fs/smb/server/ |
H A D | transport_rdma.c | 1344 struct ib_send_wr *first_wr; in smb_direct_rdma_xmit() local 1439 first_wr = NULL; in smb_direct_rdma_xmit() 1441 first_wr = rdma_rw_ctx_wrs(&msg->rw_ctx, t->qp, t->qp->port, in smb_direct_rdma_xmit() 1442 &msg->cqe, first_wr); in smb_direct_rdma_xmit() 1445 ret = ib_post_send(t->qp, first_wr, NULL); in smb_direct_rdma_xmit()
|
/linux/drivers/infiniband/ulp/srpt/ |
H A D | ib_srpt.c | 2807 struct ib_send_wr *first_wr = NULL; in srpt_write_pending() local 2832 first_wr = rdma_rw_ctx_wrs(&ctx->rw, ch->qp, ch->sport->port, in srpt_write_pending() 2833 cqe, first_wr); in srpt_write_pending() 2837 ret = ib_post_send(ch->qp, first_wr, NULL); in srpt_write_pending() 2875 struct ib_send_wr send_wr, *first_wr = &send_wr; in srpt_queue_response() local 2906 first_wr = rdma_rw_ctx_wrs(&ctx->rw, ch->qp, in srpt_queue_response() 2907 ch->sport->port, NULL, first_wr); in srpt_queue_response() 2945 ret = ib_post_send(ch->qp, first_wr, NULL); in srpt_queue_response()
|