Home
last modified time | relevance | path

Searched refs:cpl (Results 1 – 25 of 72) sorted by relevance

123

/src/sys/dev/cxgbe/tom/
H A Dt4_connect.c76 const struct cpl_act_establish *cpl = (const void *)(rss + 1); in do_act_establish() local
77 u_int tid = GET_TID(cpl); in do_act_establish()
78 u_int atid = G_TID_TID(ntohl(cpl->tos_atid)); in do_act_establish()
103 send_reset(sc, toep, be32toh(cpl->snd_isn)); in do_act_establish()
107 make_established(toep, be32toh(cpl->snd_isn) - 1, in do_act_establish()
108 be32toh(cpl->rcv_isn) - 1, cpl->tcp_opt); in do_act_establish()
156 const struct cpl_act_open_rpl *cpl = (const void *)(rss + 1); in do_act_open_rpl() local
157 u_int atid = G_TID_TID(G_AOPEN_ATID(be32toh(cpl->atid_status))); in do_act_open_rpl()
158 u_int status = G_AOPEN_STATUS(be32toh(cpl->atid_status)); in do_act_open_rpl()
172 release_tid(sc, GET_TID(cpl), toep->ctrlq); in do_act_open_rpl()
[all …]
H A Dt4_listen.c884 const struct cpl_pass_open_rpl *cpl = (const void *)(rss + 1); in do_pass_open_rpl() local
885 int stid = GET_TID(cpl); in do_pass_open_rpl()
886 unsigned int status = cpl->status; in do_pass_open_rpl()
890 unsigned int opcode = G_CPL_OPCODE(be32toh(OPCODE_TID(cpl))); in do_pass_open_rpl()
960 const struct cpl_close_listsvr_rpl *cpl = (const void *)(rss + 1); in do_close_server_rpl() local
961 int stid = GET_TID(cpl); in do_close_server_rpl()
962 unsigned int status = cpl->status; in do_close_server_rpl()
966 unsigned int opcode = G_CPL_OPCODE(be32toh(OPCODE_TID(cpl))); in do_close_server_rpl()
1027 const struct cpl_abort_req_rss *cpl = (const void *)(rss + 1); in do_abort_req_synqe() local
1028 unsigned int tid = GET_TID(cpl); in do_abort_req_synqe()
[all …]
H A Dt4_cpl_io.c989 struct cpl_tx_data_iso *cpl; in write_iscsi_tx_data_iso() local
1009 cpl = (struct cpl_tx_data_iso *)dst; in write_iscsi_tx_data_iso()
1010 cpl->op_to_scsi = htonl(V_CPL_TX_DATA_ISO_OP(CPL_TX_DATA_ISO) | in write_iscsi_tx_data_iso()
1018 cpl->ahs_len = 0; in write_iscsi_tx_data_iso()
1019 cpl->mpdu = htons(DIV_ROUND_UP(mss, 4)); in write_iscsi_tx_data_iso()
1020 cpl->burst_size = htonl(DIV_ROUND_UP(burst_size, 4)); in write_iscsi_tx_data_iso()
1021 cpl->len = htonl(len); in write_iscsi_tx_data_iso()
1022 cpl->reserved2_seglen_offset = htonl(0); in write_iscsi_tx_data_iso()
1023 cpl->datasn_offset = htonl(0); in write_iscsi_tx_data_iso()
1024 cpl->buffer_offset = htonl(0); in write_iscsi_tx_data_iso()
[all …]
H A Dt4_tls.c396 write_tlstx_cpl(struct cpl_tx_tls_sfo *cpl, struct toepcb *toep, in write_tlstx_cpl() argument
405 cpl->op_to_seg_len = htobe32(V_CPL_TX_TLS_SFO_OPCODE(CPL_TX_TLS_SFO) | in write_tlstx_cpl()
408 cpl->pld_len = htobe32(plen); in write_tlstx_cpl()
410 cpl->type_protover = htobe32(V_CPL_TX_TLS_SFO_TYPE(rec_type)); in write_tlstx_cpl()
411 cpl->seqno_numivs = htobe32(tls_ofld->scmd0.seqno_numivs | in write_tlstx_cpl()
413 cpl->ivgen_hdrlen = htobe32(tls_ofld->scmd0.ivgen_hdrlen); in write_tlstx_cpl()
414 cpl->scmd1 = htobe64(seqno); in write_tlstx_cpl()
499 struct cpl_tx_tls_sfo *cpl; in t4_push_ktls() local
685 cpl = (struct cpl_tx_tls_sfo *)(txwr + 1); in t4_push_ktls()
693 write_tlstx_cpl(cpl, toep, thdr, tls_size, in t4_push_ktls()
[all …]
/src/sys/dev/nvme/
H A Dnvme_qpair.c120 get_status_string(const struct nvme_completion *cpl, char *buf, size_t len) in get_status_string() argument
125 nvme_sc_sbuf(cpl, &sb); in get_status_string()
133 struct nvme_completion *cpl) in nvme_qpair_print_completion() argument
138 crd = NVME_STATUS_GET_CRD(cpl->status); in nvme_qpair_print_completion()
139 m = NVME_STATUS_GET_M(cpl->status); in nvme_qpair_print_completion()
140 dnr = NVME_STATUS_GET_DNR(cpl->status); in nvme_qpair_print_completion()
141 p = NVME_STATUS_GET_P(cpl->status); in nvme_qpair_print_completion()
145 get_status_string(cpl, buf, sizeof(buf)), crd, m, dnr, p, in nvme_qpair_print_completion()
146 cpl->sqid, cpl->cid, cpl->cdw0); in nvme_qpair_print_completion()
150 nvme_completion_is_retry(const struct nvme_completion *cpl) in nvme_completion_is_retry() argument
[all …]
H A Dnvme_ctrlr.c469 if (nvme_completion_is_error(&status.cpl)) { in nvme_ctrlr_identify()
499 if (nvme_completion_is_error(&status.cpl)) { in nvme_ctrlr_set_num_qpairs()
509 sq_allocated = (status.cpl.cdw0 & 0xFFFF) + 1; in nvme_ctrlr_set_num_qpairs()
510 cq_allocated = (status.cpl.cdw0 >> 16) + 1; in nvme_ctrlr_set_num_qpairs()
539 if (nvme_completion_is_error(&status.cpl)) { in nvme_ctrlr_create_qpairs()
548 if (nvme_completion_is_error(&status.cpl)) { in nvme_ctrlr_create_qpairs()
570 if (nvme_completion_is_error(&status.cpl)) { in nvme_ctrlr_delete_qpairs()
579 if (nvme_completion_is_error(&status.cpl)) { in nvme_ctrlr_delete_qpairs()
686 nvme_ctrlr_async_event_cb(void *arg, const struct nvme_completion *cpl) in nvme_ctrlr_async_event_cb() argument
690 if (nvme_completion_is_error(cpl)) { in nvme_ctrlr_async_event_cb()
[all …]
H A Dnvme.c134 nvme_completion_poll_cb(void *arg, const struct nvme_completion *cpl) in nvme_completion_poll_cb() argument
143 memcpy(&status->cpl, cpl, sizeof(*cpl)); in nvme_completion_poll_cb()
H A Dnvme_test.c160 nvme_ns_io_test_cb(void *arg, const struct nvme_completion *cpl) in nvme_ns_io_test_cb() argument
167 if (nvme_completion_is_error(cpl)) { in nvme_ns_io_test_cb()
202 struct nvme_completion cpl; in nvme_ns_io_test() local
214 memset(&cpl, 0, sizeof(cpl)); in nvme_ns_io_test()
216 nvme_ns_io_test_cb(tth, &cpl); in nvme_ns_io_test()
H A Dnvme_util.c215 nvme_sc_sbuf(const struct nvme_completion *cpl, struct sbuf *sb) in nvme_sc_sbuf() argument
220 status = le16toh(cpl->status); in nvme_sc_sbuf()
259 nvme_cpl_sbuf(const struct nvme_completion *cpl, struct sbuf *sb) in nvme_cpl_sbuf() argument
263 status = le16toh(cpl->status); in nvme_cpl_sbuf()
264 nvme_sc_sbuf(cpl, sb); in nvme_cpl_sbuf()
/src/sys/dev/nvmf/controller/
H A Dnvmft_qpair.c194 struct nvme_completion cpl; in _nvmft_send_response() local
199 memcpy(&cpl, cqe, sizeof(cpl)); in _nvmft_send_response()
211 cpl.sqhd = htole16(qp->sqhd); in _nvmft_send_response()
213 cpl.sqhd = 0; in _nvmft_send_response()
216 rc = nvmf_allocate_response(nq, &cpl, M_WAITOK); in _nvmft_send_response()
240 const struct nvme_completion *cpl = cqe; in nvmft_send_response() local
243 KASSERT(BIT_ISSET(NUM_CIDS, cpl->cid, qp->cids), in nvmft_send_response()
244 ("%s: CID %u not busy", __func__, cpl->cid)); in nvmft_send_response()
246 BIT_CLR_ATOMIC(NUM_CIDS, cpl->cid, qp->cids); in nvmft_send_response()
253 struct nvme_completion *cpl = cqe; in nvmft_init_cqe() local
[all …]
/src/sys/dev/cxgbe/cxgbei/
H A Dcxgbei.c196 struct cpl_iscsi_hdr *cpl = mtod(m, struct cpl_iscsi_hdr *); in do_rx_iscsi_hdr() local
197 u_int tid = GET_TID(cpl); in do_rx_iscsi_hdr()
201 uint16_t len_ddp = be16toh(cpl->pdu_len_ddp); in do_rx_iscsi_hdr()
202 uint16_t len = be16toh(cpl->len); in do_rx_iscsi_hdr()
205 MPASS(m->m_pkthdr.len == len + sizeof(*cpl)); in do_rx_iscsi_hdr()
210 m_copydata(m, sizeof(*cpl), ISCSI_BHS_SIZE, (caddr_t)ip->ip_bhs); in do_rx_iscsi_hdr()
213 icp->icp_seq = ntohl(cpl->seq); in do_rx_iscsi_hdr()
233 struct cpl_iscsi_data *cpl = mtod(m, struct cpl_iscsi_data *); in do_rx_iscsi_data() local
234 u_int tid = GET_TID(cpl); in do_rx_iscsi_data()
240 MPASS(m->m_pkthdr.len == be16toh(cpl->len) + sizeof(*cpl)); in do_rx_iscsi_data()
[all …]
/src/sys/dev/cxgbe/crypto/
H A Dt6_kern_tls.c154 struct cpl_act_open_req *cpl; in mk_ktls_act_open_req() local
159 cpl = (struct cpl_act_open_req *)cpl6; in mk_ktls_act_open_req()
163 OPCODE_TID(cpl) = htobe32(MK_OPCODE_TID(CPL_ACT_OPEN_REQ, in mk_ktls_act_open_req()
165 inp_4tuple_get(inp, &cpl->local_ip, &cpl->local_port, in mk_ktls_act_open_req()
166 &cpl->peer_ip, &cpl->peer_port); in mk_ktls_act_open_req()
171 cpl->opt0 = htobe64(options); in mk_ktls_act_open_req()
176 cpl->opt2 = htobe32(options); in mk_ktls_act_open_req()
185 struct cpl_act_open_req6 *cpl; in mk_ktls_act_open_req6() local
190 cpl = (struct cpl_act_open_req6 *)cpl6; in mk_ktls_act_open_req6()
194 OPCODE_TID(cpl) = htobe32(MK_OPCODE_TID(CPL_ACT_OPEN_REQ6, in mk_ktls_act_open_req6()
[all …]
H A Dt7_kern_tls.c962 write_lso_cpl(void *cpl, struct mbuf *m0, uint16_t mss, uint16_t eh_type, in write_lso_cpl() argument
981 lso = cpl; in write_lso_cpl()
994 struct cpl_tx_tls_ack *cpl; in write_tx_tls_ack() local
998 cpl = dst; in write_tx_tls_ack()
999 cpl->op_to_Rsvd2 = htobe32(V_CPL_TX_TLS_ACK_OPCODE(CPL_TX_TLS_ACK) | in write_tx_tls_ack()
1004 cpl->PldLen = htobe32(V_CPL_TX_TLS_ACK_PLDLEN(32 + 16 + hash_len)); in write_tx_tls_ack()
1005 cpl->Rsvd3 = 0; in write_tx_tls_ack()
1007 return (cpl + 1); in write_tx_tls_ack()
1015 struct cpl_fw6_pld *cpl; in write_fw6_pld() local
1023 cpl = (void *)(rss + 1); in write_fw6_pld()
[all …]
/src/sys/dev/cxgb/
H A Dcxgb_offload.h94 #define M_GETHDR_OFLD(qset, ctrl, cpl) \ argument
95 m_gethdr_ofld(qset, ctrl, sizeof(*cpl), (void **)&cpl)
97 m_gethdr_ofld(int qset, int ctrl, int cpllen, void **cpl) in m_gethdr_ofld() argument
108 *cpl = (void *)(oh + 1); in m_gethdr_ofld()
/src/sys/dev/ufshci/
H A Dufshci.c51 ufshci_completion_poll_cb(void *arg, const struct ufshci_completion *cpl, in ufshci_completion_poll_cb() argument
61 memcpy(&status->cpl.response_upiu, &cpl->response_upiu, cpl->size); in ufshci_completion_poll_cb()
H A Dufshci_req_queue.c181 struct ufshci_completion cpl; in ufshci_req_queue_manual_complete_request() local
184 memset(&cpl, 0, sizeof(cpl)); in ufshci_req_queue_manual_complete_request()
185 cpl.response_upiu.header.response = rc; in ufshci_req_queue_manual_complete_request()
187 &cpl.response_upiu); in ufshci_req_queue_manual_complete_request()
195 req->cb_fn(req->cb_arg, &cpl, error); in ufshci_req_queue_manual_complete_request()
244 struct ufshci_completion cpl; in ufshci_req_queue_complete_tracker() local
252 cpl.size = tr->response_size; in ufshci_req_queue_complete_tracker()
254 memcpy(&cpl.response_upiu, in ufshci_req_queue_complete_tracker()
255 (void *)hwq->utmrd[tr->slot_num].response_upiu, cpl.size); in ufshci_req_queue_complete_tracker()
262 memcpy(&cpl.response_upiu, (void *)tr->ucd->response_upiu, in ufshci_req_queue_complete_tracker()
[all …]
/src/sbin/nvmecontrol/
H A Dns.c431 if (nvme_completion_is_error(&pt.cpl)) in nsactive()
476 if (nvme_completion_is_error(&pt.cpl)) in nsallocated()
520 if (nvme_completion_is_error(&pt.cpl)) in nscontrollers()
621 if (nvme_completion_is_error(&pt.cpl)) { in nscreate()
623 get_res_str(NVMEV(NVME_STATUS_SC, pt.cpl.status))); in nscreate()
625 printf("namespace %d created\n", pt.cpl.cdw0); in nscreate()
673 if (nvme_completion_is_error(&pt.cpl)) { in nsdelete()
675 get_res_str(NVMEV(NVME_STATUS_SC, pt.cpl.status))); in nsdelete()
737 if (nvme_completion_is_error(&pt.cpl)) in nsattach()
758 if (nvme_completion_is_error(&pt.cpl)) { in nsattach()
[all …]
H A Dpower.c116 if (nvme_completion_is_error(&pt.cpl)) in power_set()
132 if (nvme_completion_is_error(&pt.cpl)) in power_show()
135 printf("Current Power State is %d\n", pt.cpl.cdw0 & 0x1F); in power_show()
136 printf("Current Workload Hint is %d\n", pt.cpl.cdw0 >> 5); in power_show()
H A Dselftest.c68 if (NVME_STATUS_GET_SCT(pt.cpl.status) == NVME_SCT_COMMAND_SPECIFIC && in selftest_op()
69 NVME_STATUS_GET_SC(pt.cpl.status) == NVME_SC_SELF_TEST_IN_PROGRESS) in selftest_op()
71 else if (nvme_completion_is_error(&pt.cpl)) in selftest_op()
H A Dnvmecontrol.c116 if (nvme_completion_is_error(&pt.cpl)) in read_controller_data()
140 if (nvme_completion_is_error(&pt.cpl)) in read_namespace_data()
164 if (nvme_completion_is_error(&pt.cpl)) in read_active_namespaces()
/src/lib/libnvmf/
H A Dnvmf_controller.c71 struct nvme_completion *cpl = cqe; in nvmf_init_cqe() local
74 memset(cpl, 0, sizeof(*cpl)); in nvmf_init_cqe()
75 cpl->cid = cmd->cid; in nvmf_init_cqe()
76 cpl->status = htole16(status); in nvmf_init_cqe()
83 struct nvme_completion cpl; in nvmf_simple_response() local
88 nvmf_init_cqe(&cpl, nc, status); in nvmf_simple_response()
89 return (nvmf_allocate_response(nc->nc_qpair, &cpl)); in nvmf_simple_response()
/src/usr.sbin/bhyve/amd64/
H A Dtask_switch.c271 sup_paging.cpl = 0; /* implicit supervisor mode */ in read_tss_descriptor()
314 int cpl, dpl, rpl; in validate_seg_desc() local
369 sup_paging.cpl = 0; /* implicit supervisor mode */ in validate_seg_desc()
396 cpl = cs & SEL_RPL_MASK; in validate_seg_desc()
400 if (stackseg && (rpl != cpl || dpl != cpl)) { in validate_seg_desc()
407 if ((conforming && (cpl < dpl)) || in validate_seg_desc()
408 (!conforming && (cpl != dpl))) { in validate_seg_desc()
424 if (!conforming && (rpl > dpl || cpl > dpl)) { in validate_seg_desc()
588 ts->paging.cpl = tss->tss_cs & SEL_RPL_MASK; in tss32_restore()
673 if (vie_alignment_check(paging->cpl, bytes, cr0, rflags, gla)) { in push_errcode()
[all …]
/src/sys/dev/cxgbe/
H A Dt4_sge.c401 const struct cpl_set_tcb_rpl *cpl = (const void *)(rss + 1); in set_tcb_rpl_handler() local
407 tid = GET_TID(cpl); in set_tcb_rpl_handler()
416 cookie = G_COOKIE(cpl->cookie); in set_tcb_rpl_handler()
441 const struct cpl_act_open_rpl *cpl = (const void *)(rss + 1); in act_open_rpl_handler() local
442 u_int cookie = G_TID_COOKIE(G_AOPEN_ATID(be32toh(cpl->atid_status))); in act_open_rpl_handler()
470 const struct cpl_fw4_ack *cpl = (const void *)(rss + 1); in fw4_ack_handler() local
471 unsigned int tid = G_CPL_FW4_ACK_FLOWID(be32toh(OPCODE_TID(cpl))); in fw4_ack_handler()
486 const struct cpl_fw6_pld *cpl; in fw6_pld_handler() local
490 cpl = mtod(m, const void *); in fw6_pld_handler()
492 cpl = (const void *)(rss + 1); in fw6_pld_handler()
[all …]
H A Dt4_netmap.c980 struct cpl_tx_pkt_core *cpl; in cxgbe_nm_tx() local
994 cpl = (void *)(wr + 1); in cxgbe_nm_tx()
1001 cpl->ctrl0 = nm_txq->cpl_ctrl0; in cxgbe_nm_tx()
1002 cpl->pack = 0; in cxgbe_nm_tx()
1003 cpl->len = htobe16(slot->len); in cxgbe_nm_tx()
1004 cpl->ctrl1 = nm_txcsum ? 0 : in cxgbe_nm_tx()
1007 usgl = (void *)(cpl + 1); in cxgbe_nm_tx()
1014 cpl = (void *)(usgl + 1); in cxgbe_nm_tx()
1308 unwrap_nm_fw6_msg(const struct cpl_fw6_msg *cpl) in unwrap_nm_fw6_msg() argument
1311 MPASS(cpl->type == FW_TYPE_RSSCPL || cpl->type == FW6_TYPE_RSSCPL); in unwrap_nm_fw6_msg()
[all …]
H A Dt4_filter.c1508 const struct cpl_act_open_rpl *cpl = (const void *)(rss + 1); in t4_hashfilter_ao_rpl() local
1509 u_int atid = G_TID_TID(G_AOPEN_ATID(be32toh(cpl->atid_status))); in t4_hashfilter_ao_rpl()
1510 u_int status = G_AOPEN_STATUS(be32toh(cpl->atid_status)); in t4_hashfilter_ao_rpl()
1520 f->tid = GET_TID(cpl); in t4_hashfilter_ao_rpl()
1538 release_tid(sc, GET_TID(cpl), &sc->sge.ctrlq[0]); in t4_hashfilter_ao_rpl()
1592 const struct cpl_abort_rpl_rss *cpl = (const void *)(rss + 1); in t4_del_hashfilter_rpl() local
1593 unsigned int tid = GET_TID(cpl); in t4_del_hashfilter_rpl()
1604 if (cpl->status == 0) { in t4_del_hashfilter_rpl()
1717 uint64_t ftuple, struct cpl_act_open_req6 *cpl) in mk_act_open_req6() argument
1719 struct cpl_t5_act_open_req6 *cpl5 = (void *)cpl; in mk_act_open_req6()
[all …]

123