Lines Matching full:pdu
105 void *pdu; member
152 void *pdu; member
275 return req->pdu; in nvme_tcp_req_cmd_pdu()
280 /* use the pdu space in the back for the data pdu */ in nvme_tcp_req_data_pdu()
281 return req->pdu + sizeof(struct nvme_tcp_cmd_pdu) - in nvme_tcp_req_data_pdu()
478 void *pdu, size_t len) in nvme_tcp_hdgst() argument
482 sg_init_one(&sg, pdu, len); in nvme_tcp_hdgst()
483 ahash_request_set_crypt(hash, &sg, pdu + len, len); in nvme_tcp_hdgst()
488 void *pdu, size_t pdu_len) in nvme_tcp_verify_hdgst() argument
490 struct nvme_tcp_hdr *hdr = pdu; in nvme_tcp_verify_hdgst()
501 recv_digest = *(__le32 *)(pdu + hdr->hlen); in nvme_tcp_verify_hdgst()
502 nvme_tcp_hdgst(queue->rcv_hash, pdu, pdu_len); in nvme_tcp_verify_hdgst()
503 exp_digest = *(__le32 *)(pdu + hdr->hlen); in nvme_tcp_verify_hdgst()
514 static int nvme_tcp_check_ddgst(struct nvme_tcp_queue *queue, void *pdu) in nvme_tcp_check_ddgst() argument
516 struct nvme_tcp_hdr *hdr = pdu; in nvme_tcp_check_ddgst()
539 page_frag_free(req->pdu); in nvme_tcp_exit_request()
548 struct nvme_tcp_cmd_pdu *pdu; in nvme_tcp_init_request() local
553 req->pdu = page_frag_alloc(&queue->pf_cache, in nvme_tcp_init_request()
556 if (!req->pdu) in nvme_tcp_init_request()
559 pdu = req->pdu; in nvme_tcp_init_request()
562 nvme_req(rq)->cmd = &pdu->cmd; in nvme_tcp_init_request()
640 struct nvme_tcp_data_pdu *pdu) in nvme_tcp_handle_c2h_data() argument
644 rq = nvme_find_rq(nvme_tcp_tagset(queue), pdu->command_id); in nvme_tcp_handle_c2h_data()
648 pdu->command_id, nvme_tcp_queue_id(queue)); in nvme_tcp_handle_c2h_data()
659 queue->data_remaining = le32_to_cpu(pdu->data_length); in nvme_tcp_handle_c2h_data()
661 if (pdu->hdr.flags & NVME_TCP_F_DATA_SUCCESS && in nvme_tcp_handle_c2h_data()
662 unlikely(!(pdu->hdr.flags & NVME_TCP_F_DATA_LAST))) { in nvme_tcp_handle_c2h_data()
664 "queue %d tag %#x SUCCESS set but not last PDU\n", in nvme_tcp_handle_c2h_data()
674 struct nvme_tcp_rsp_pdu *pdu) in nvme_tcp_handle_comp() argument
676 struct nvme_completion *cqe = &pdu->cqe; in nvme_tcp_handle_comp()
730 struct nvme_tcp_r2t_pdu *pdu) in nvme_tcp_handle_r2t() argument
734 u32 r2t_length = le32_to_cpu(pdu->r2t_length); in nvme_tcp_handle_r2t()
735 u32 r2t_offset = le32_to_cpu(pdu->r2t_offset); in nvme_tcp_handle_r2t()
737 rq = nvme_find_rq(nvme_tcp_tagset(queue), pdu->command_id); in nvme_tcp_handle_r2t()
741 pdu->command_id, nvme_tcp_queue_id(queue)); in nvme_tcp_handle_r2t()
770 req->ttag = pdu->ttag; in nvme_tcp_handle_r2t()
779 struct nvme_tcp_term_pdu *pdu) in nvme_tcp_handle_c2h_term() argument
783 u32 plen = le32_to_cpu(pdu->hdr.plen); in nvme_tcp_handle_c2h_term()
786 [NVME_TCP_FES_INVALID_PDU_HDR] = "Invalid PDU Header Field", in nvme_tcp_handle_c2h_term()
787 [NVME_TCP_FES_PDU_SEQ_ERR] = "PDU Sequence Error", in nvme_tcp_handle_c2h_term()
797 "Received a malformed C2HTermReq PDU (plen = %u)\n", in nvme_tcp_handle_c2h_term()
802 fes = le16_to_cpu(pdu->fes); in nvme_tcp_handle_c2h_term()
816 char *pdu = queue->pdu; in nvme_tcp_recv_pdu() local
821 &pdu[queue->pdu_offset], rcv_len); in nvme_tcp_recv_pdu()
832 hdr = queue->pdu; in nvme_tcp_recv_pdu()
838 "pdu type %d has unexpected header length (%d)\n", in nvme_tcp_recv_pdu()
848 nvme_tcp_handle_c2h_term(queue, (void *)queue->pdu); in nvme_tcp_recv_pdu()
853 ret = nvme_tcp_verify_hdgst(queue, queue->pdu, hdr->hlen); in nvme_tcp_recv_pdu()
860 ret = nvme_tcp_check_ddgst(queue, queue->pdu); in nvme_tcp_recv_pdu()
867 return nvme_tcp_handle_c2h_data(queue, (void *)queue->pdu); in nvme_tcp_recv_pdu()
870 return nvme_tcp_handle_comp(queue, (void *)queue->pdu); in nvme_tcp_recv_pdu()
873 return nvme_tcp_handle_r2t(queue, (void *)queue->pdu); in nvme_tcp_recv_pdu()
880 "unsupported pdu type (%d)\n", hdr->type); in nvme_tcp_recv_pdu()
895 struct nvme_tcp_data_pdu *pdu = (void *)queue->pdu; in nvme_tcp_recv_data() local
897 nvme_cid_to_rq(nvme_tcp_tagset(queue), pdu->command_id); in nvme_tcp_recv_data()
951 if (pdu->hdr.flags & NVME_TCP_F_DATA_SUCCESS) { in nvme_tcp_recv_data()
966 struct nvme_tcp_data_pdu *pdu = (void *)queue->pdu; in nvme_tcp_recv_ddgst() local
984 pdu->command_id); in nvme_tcp_recv_ddgst()
995 if (pdu->hdr.flags & NVME_TCP_F_DATA_SUCCESS) { in nvme_tcp_recv_ddgst()
997 pdu->command_id); in nvme_tcp_recv_ddgst()
1161 /* fully successful last send in current PDU */ in nvme_tcp_try_send_data()
1183 struct nvme_tcp_cmd_pdu *pdu = nvme_tcp_req_cmd_pdu(req); in nvme_tcp_try_send_cmd_pdu() local
1188 int len = sizeof(*pdu) + hdgst - req->offset; in nvme_tcp_try_send_cmd_pdu()
1197 nvme_tcp_hdgst(queue->snd_hash, pdu, sizeof(*pdu)); in nvme_tcp_try_send_cmd_pdu()
1199 bvec_set_virt(&bvec, (void *)pdu + req->offset, len); in nvme_tcp_try_send_cmd_pdu()
1224 struct nvme_tcp_data_pdu *pdu = nvme_tcp_req_data_pdu(req); in nvme_tcp_try_send_data_pdu() local
1228 int len = sizeof(*pdu) - req->offset + hdgst; in nvme_tcp_try_send_data_pdu()
1232 nvme_tcp_hdgst(queue->snd_hash, pdu, sizeof(*pdu)); in nvme_tcp_try_send_data_pdu()
1237 bvec_set_virt(&bvec, (void *)pdu + req->offset, len); in nvme_tcp_try_send_data_pdu()
1426 page_frag_free(async->pdu); in nvme_tcp_free_async_req()
1435 async->pdu = page_frag_alloc(&queue->pf_cache, in nvme_tcp_alloc_async_req()
1438 if (!async->pdu) in nvme_tcp_alloc_async_req()
1465 kfree(queue->pdu); in nvme_tcp_free_queue()
1548 pr_err("queue %d: bad pdu length returned %d\n", in nvme_tcp_init_connection()
1881 queue->pdu = kmalloc(rcv_pdu_size, GFP_KERNEL); in nvme_tcp_alloc_queue()
1882 if (!queue->pdu) { in nvme_tcp_alloc_queue()
1916 kfree(queue->pdu); in nvme_tcp_alloc_queue()
2038 if (to_tcp_ctrl(ctrl)->async_req.pdu) { in nvme_tcp_free_admin_queue()
2041 to_tcp_ctrl(ctrl)->async_req.pdu = NULL; in nvme_tcp_free_admin_queue()
2619 struct nvme_tcp_cmd_pdu *pdu = ctrl->async_req.pdu; in nvme_tcp_submit_async_event() local
2620 struct nvme_command *cmd = &pdu->cmd; in nvme_tcp_submit_async_event()
2623 memset(pdu, 0, sizeof(*pdu)); in nvme_tcp_submit_async_event()
2624 pdu->hdr.type = nvme_tcp_cmd; in nvme_tcp_submit_async_event()
2626 pdu->hdr.flags |= NVME_TCP_F_HDGST; in nvme_tcp_submit_async_event()
2627 pdu->hdr.hlen = sizeof(*pdu); in nvme_tcp_submit_async_event()
2628 pdu->hdr.plen = cpu_to_le32(pdu->hdr.hlen + hdgst); in nvme_tcp_submit_async_event()
2656 struct nvme_tcp_cmd_pdu *pdu = nvme_tcp_req_cmd_pdu(req); in nvme_tcp_timeout() local
2657 struct nvme_command *cmd = &pdu->cmd; in nvme_tcp_timeout()
2662 rq->tag, nvme_cid(rq), pdu->hdr.type, cmd->common.opcode, in nvme_tcp_timeout()
2695 struct nvme_tcp_cmd_pdu *pdu = nvme_tcp_req_cmd_pdu(req); in nvme_tcp_map_data() local
2696 struct nvme_command *c = &pdu->cmd; in nvme_tcp_map_data()
2715 struct nvme_tcp_cmd_pdu *pdu = nvme_tcp_req_cmd_pdu(req); in nvme_tcp_setup_cmd_pdu() local
2741 pdu->hdr.type = nvme_tcp_cmd; in nvme_tcp_setup_cmd_pdu()
2742 pdu->hdr.flags = 0; in nvme_tcp_setup_cmd_pdu()
2744 pdu->hdr.flags |= NVME_TCP_F_HDGST; in nvme_tcp_setup_cmd_pdu()
2746 pdu->hdr.flags |= NVME_TCP_F_DDGST; in nvme_tcp_setup_cmd_pdu()
2749 pdu->hdr.hlen = sizeof(*pdu); in nvme_tcp_setup_cmd_pdu()
2750 pdu->hdr.pdo = req->pdu_len ? pdu->hdr.hlen + hdgst : 0; in nvme_tcp_setup_cmd_pdu()
2751 pdu->hdr.plen = in nvme_tcp_setup_cmd_pdu()
2752 cpu_to_le32(pdu->hdr.hlen + hdgst + req->pdu_len + ddgst); in nvme_tcp_setup_cmd_pdu()