/linux/drivers/infiniband/sw/rxe/ |
H A D | rxe_cq.c | 12 int cqe, int comp_vector) in rxe_cq_chk_attr() 42 int rxe_cq_from_init(struct rxe_dev *rxe, struct rxe_cq *cq, int cqe, in rxe_cq_from_init() 69 int rxe_cq_resize_queue(struct rxe_cq *cq, int cqe, in rxe_cq_resize_queue() 85 int rxe_cq_post(struct rxe_cq *cq, struct rxe_cqe *cqe, int solicited) in rxe_cq_post()
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | en_rx.c | 90 struct mlx5_cqe64 *cqe) in mlx5e_read_enhanced_title_slot() 138 struct mlx5_cqe64 *cqe = mlx5_cqwq_get_wqe(wq, ci); in mlx5e_cqes_update_owner() local 146 struct mlx5_cqe64 *cqe = mlx5_cqwq_get_wqe(wq, ci); in mlx5e_cqes_update_owner() local 196 struct mlx5_cqe64 *cqe, in mlx5e_decompress_enhanced_cqe() 999 struct mlx5_cqe64 *cqe; in mlx5e_poll_ico_cq() local 1148 static void mlx5e_lro_update_tcp_hdr(struct mlx5_cqe64 *cqe, struct tcphdr *tcp) in mlx5e_lro_update_tcp_hdr() 1165 struct mlx5_cqe64 *cqe, in mlx5e_lro_update_hdr() 1266 static void mlx5e_shampo_update_fin_psh_flags(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe, in mlx5e_shampo_update_fin_psh_flags() 1279 struct mlx5_cqe64 *cqe, bool match) in mlx5e_shampo_update_ipv4_tcp_hdr() 1303 struct mlx5_cqe64 *cqe, bool match) in mlx5e_shampo_update_ipv6_tcp_hdr() [all …]
|
/linux/drivers/infiniband/hw/mlx4/ |
H A D | cq.c | 81 struct mlx4_cqe *cqe = get_cqe(cq, n & cq->ibcq.cqe); in get_sw_cqe() local 133 static void mlx4_ib_free_cq_buf(struct mlx4_ib_dev *dev, struct mlx4_ib_cq_buf *buf, int cqe) in mlx4_ib_free_cq_buf() 140 struct ib_umem **umem, u64 buf_addr, int cqe) in mlx4_ib_get_cq_umem() 361 struct mlx4_cqe *cqe, *new_cqe; in mlx4_ib_cq_resize_copy_cqes() local 505 static void dump_cqe(void *cqe) in dump_cqe() 515 static void mlx4_ib_handle_error_cqe(struct mlx4_err_cqe *cqe, in mlx4_ib_handle_error_cqe() 586 unsigned tail, struct mlx4_cqe *cqe, int is_eth) in use_tunnel_data() 665 struct mlx4_cqe *cqe; in mlx4_ib_poll_one() local 924 struct mlx4_cqe *cqe, *dest; in __mlx4_ib_cq_clean() local
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/en_accel/ |
H A D | macsec.h | 32 static inline bool mlx5e_macsec_is_rx_flow(struct mlx5_cqe64 *cqe) in mlx5e_macsec_is_rx_flow() 46 static inline bool mlx5e_macsec_is_rx_flow(struct mlx5_cqe64 *cqe) { return false; } in mlx5e_macsec_is_rx_flow() 49 struct mlx5_cqe64 *cqe) in mlx5e_macsec_offload_handle_rx_skb()
|
H A D | ipsec_rxtx.h | 75 static inline bool mlx5_ipsec_is_rx_flow(struct mlx5_cqe64 *cqe) in mlx5_ipsec_is_rx_flow() 153 static inline bool mlx5_ipsec_is_rx_flow(struct mlx5_cqe64 *cqe) { return false; } in mlx5_ipsec_is_rx_flow()
|
/linux/drivers/scsi/qedi/ |
H A D | qedi_fw.c | 31 union iscsi_cqe *cqe, in qedi_process_logout_resp() 82 union iscsi_cqe *cqe, in qedi_process_text_resp() 178 union iscsi_cqe *cqe, in qedi_process_tmf_resp() 258 union iscsi_cqe *cqe, in qedi_process_login_resp() 320 struct iscsi_cqe_unsolicited *cqe, in qedi_get_rq_bdq_buf() 359 struct iscsi_cqe_unsolicited *cqe, in qedi_put_rq_bdq_buf() 397 struct iscsi_cqe_unsolicited *cqe, in qedi_unsol_pdu_adjust_bdq() 409 union iscsi_cqe *cqe, in qedi_process_nopin_mesg() 480 union iscsi_cqe *cqe, in qedi_process_async_mesg() 534 union iscsi_cqe *cqe, in qedi_process_reject_mesg() [all …]
|
/linux/drivers/infiniband/hw/erdma/ |
H A D | erdma_cq.c | 11 __be32 *cqe = get_queue_entry(cq->kern_cq.qbuf, cq->kern_cq.ci, in get_next_valid_cqe() local 108 static void erdma_process_ud_cqe(struct erdma_cqe *cqe, struct ib_wc *wc) in erdma_process_ud_cqe() 131 struct erdma_cqe *cqe; in erdma_poll_one_cqe() local 228 struct erdma_cqe *cqe, *dst_cqe; in erdma_remove_cqes_of_qp() local
|
/linux/drivers/net/ethernet/mellanox/mlxsw/ |
H A D | pci_hw.h | 286 static inline u16 mlxsw_pci_cqe2_mirror_cong_get(const char *cqe) in mlxsw_pci_cqe2_mirror_cong_get() 343 static inline u64 mlxsw_pci_cqe2_time_stamp_get(const char *cqe) in mlxsw_pci_cqe2_time_stamp_get() 351 static inline u8 mlxsw_pci_cqe2_time_stamp_sec_get(const char *cqe) in mlxsw_pci_cqe2_time_stamp_sec_get() 358 static inline u32 mlxsw_pci_cqe2_time_stamp_nsec_get(const char *cqe) in mlxsw_pci_cqe2_time_stamp_nsec_get()
|
/linux/drivers/scsi/bnx2i/ |
H A D | bnx2i_hwi.c | 1338 struct cqe *cqe) in bnx2i_process_scsi_cmd_resp() 1435 struct cqe *cqe) in bnx2i_process_login_resp() 1503 struct cqe *cqe) in bnx2i_process_text_resp() 1564 struct cqe *cqe) in bnx2i_process_tmf_resp() 1603 struct cqe *cqe) in bnx2i_process_logout_resp() 1649 struct cqe *cqe) in bnx2i_process_nopin_local_cmpl() 1690 struct cqe *cqe) in bnx2i_process_nopin_mesg() 1742 struct cqe *cqe) in bnx2i_process_async_mesg() 1792 struct cqe *cqe) in bnx2i_process_reject_mesg() 1829 struct cqe *cqe) in bnx2i_process_cmd_cleanup_resp() [all …]
|
/linux/drivers/net/ethernet/marvell/octeontx2/nic/ |
H A D | otx2_txrx.c | 107 struct nix_cqe_tx_s *cqe, in otx2_xdp_snd_pkt_handler() 128 struct nix_cqe_tx_s *cqe, in otx2_snd_pkt_handler() 225 struct nix_cqe_rx_s *cqe, struct sk_buff *skb) in otx2_set_rxhash() 246 static void otx2_free_rcv_seg(struct otx2_nic *pfvf, struct nix_cqe_rx_s *cqe, in otx2_free_rcv_seg() 267 struct nix_cqe_rx_s *cqe, int qidx) in otx2_check_rcv_errors() 333 struct nix_cqe_rx_s *cqe, bool *need_xdp_flush) in otx2_rcv_pkt_handler() 398 struct nix_cqe_rx_s *cqe; in otx2_rx_napi_handler() local 469 struct nix_cqe_tx_s *cqe; in otx2_tx_napi_handler() local 1278 struct nix_cqe_rx_s *cqe; in otx2_cleanup_rx_cqes() local 1323 struct nix_cqe_tx_s *cqe; in otx2_cleanup_tx_cqes() local [all …]
|
/linux/drivers/infiniband/hw/mlx5/ |
H A D | cq.c | 84 void *cqe = get_cqe(cq, n & cq->ibcq.cqe); in get_sw_cqe() local 120 static void handle_good_req(struct ib_wc *wc, struct mlx5_cqe64 *cqe, in handle_good_req() 169 static void handle_responder(struct ib_wc *wc, struct mlx5_cqe64 *cqe, in handle_responder() 273 static void dump_cqe(struct mlx5_ib_dev *dev, struct mlx5_err_cqe *cqe, in dump_cqe() 283 struct mlx5_err_cqe *cqe, in mlx5_handle_error_cqe() 365 static void get_sig_err_item(struct mlx5_sig_err_cqe *cqe, in get_sig_err_item() 462 void *cqe; in mlx5_poll_one() local 877 void *cqe; in init_cq_frag_buf() local 1094 void *cqe, *dest; in __mlx5_ib_cq_clean() local
|
/linux/drivers/infiniband/hw/mthca/ |
H A D | mthca_cq.c | 174 static inline struct mthca_cqe *cqe_sw(struct mthca_cqe *cqe) in cqe_sw() 184 static inline void set_cqe_hw(struct mthca_cqe *cqe) in set_cqe_hw() 191 __be32 *cqe = cqe_ptr; in dump_cqe() local 264 static inline int is_recv_cqe(struct mthca_cqe *cqe) in is_recv_cqe() 276 struct mthca_cqe *cqe; in mthca_cq_clean() local 366 void mthca_free_cq_buf(struct mthca_dev *dev, struct mthca_cq_buf *buf, int cqe) in mthca_free_cq_buf() 374 struct mthca_err_cqe *cqe, in handle_error_cqe() 485 struct mthca_cqe *cqe; in mthca_poll_one() local
|
/linux/net/sunrpc/xprtrdma/ |
H A D | frwr_ops.c | 361 struct ib_cqe *cqe = wc->wr_cqe; in frwr_wc_fastreg() local 456 struct ib_cqe *cqe = wc->wr_cqe; in frwr_wc_localinv() local 475 struct ib_cqe *cqe = wc->wr_cqe; in frwr_wc_localinv_wake() local 572 struct ib_cqe *cqe = wc->wr_cqe; in frwr_wc_localinv_done() local
|
H A D | svc_rdma_rw.c | 257 struct ib_cqe *cqe = wc->wr_cqe; in svc_rdma_reply_done() local 286 struct ib_cqe *cqe = wc->wr_cqe; in svc_rdma_write_done() local 320 struct ib_cqe *cqe = wc->wr_cqe; in svc_rdma_wc_read_done() local 368 struct ib_cqe *cqe; in svc_rdma_post_chunk_ctxt() local 688 struct ib_cqe *cqe; in svc_rdma_prepare_reply_chunk() local
|
/linux/drivers/net/ethernet/qlogic/qede/ |
H A D | qede_fp.c | 650 struct eth_fast_path_rx_tpa_start_cqe *cqe) in qede_set_gro_params() 833 struct eth_fast_path_rx_tpa_start_cqe *cqe) in qede_tpa_start() 959 struct eth_fast_path_rx_tpa_cont_cqe *cqe) in qede_tpa_cont() 974 struct eth_fast_path_rx_tpa_end_cqe *cqe) in qede_tpa_end() 1064 static bool qede_pkt_is_ip_fragmented(struct eth_fast_path_rx_reg_cqe *cqe, in qede_pkt_is_ip_fragmented() 1084 struct eth_fast_path_rx_reg_cqe *cqe, in qede_rx_xdp() 1172 struct eth_fast_path_rx_reg_cqe *cqe, in qede_rx_build_jumbo() 1226 union eth_rx_cqe *cqe, in qede_rx_process_tpa_cqe() 1251 union eth_rx_cqe *cqe; in qede_rx_process_cqe() local
|
H A D | qede_ptp.h | 23 union eth_rx_cqe *cqe, in qede_ptp_record_rx_ts()
|
/linux/drivers/infiniband/hw/mana/ |
H A D | cq.c | 187 static inline void handle_ud_sq_cqe(struct mana_ib_qp *qp, struct gdma_comp *cqe) in handle_ud_sq_cqe() 203 static inline void handle_ud_rq_cqe(struct mana_ib_qp *qp, struct gdma_comp *cqe) in handle_ud_rq_cqe() 221 static void mana_handle_cqe(struct mana_ib_dev *mdev, struct gdma_comp *cqe) in mana_handle_cqe()
|
/linux/drivers/infiniband/ulp/iser/ |
H A D | iser_memory.c | 237 struct ib_cqe *cqe, struct ib_send_wr *next_wr) in iser_inv_rkey() 254 struct ib_cqe *cqe = &iser_task->iser_conn->ib_conn.reg_cqe; in iser_reg_sig_mr() local 311 struct ib_cqe *cqe = &iser_task->iser_conn->ib_conn.reg_cqe; in iser_fast_reg_mr() local
|
H A D | iscsi_iser.h | 244 struct ib_cqe cqe; member 271 struct ib_cqe cqe; member 291 struct ib_cqe cqe; member 552 iser_rx(struct ib_cqe *cqe) in iser_rx() 558 iser_tx(struct ib_cqe *cqe) in iser_tx() 564 iser_login(struct ib_cqe *cqe) in iser_login()
|
/linux/drivers/scsi/qedf/ |
H A D | qedf_io.c | 1121 void qedf_scsi_completion(struct qedf_ctx *qedf, struct fcoe_cqe *cqe, in qedf_scsi_completion() 1432 void qedf_process_warning_compl(struct qedf_ctx *qedf, struct fcoe_cqe *cqe, in qedf_process_warning_compl() 1502 void qedf_process_error_detect(struct qedf_ctx *qedf, struct fcoe_cqe *cqe, in qedf_process_error_detect() 1957 void qedf_process_abts_compl(struct qedf_ctx *qedf, struct fcoe_cqe *cqe, in qedf_process_abts_compl() 2278 void qedf_process_cleanup_compl(struct qedf_ctx *qedf, struct fcoe_cqe *cqe, in qedf_process_cleanup_compl() 2476 void qedf_process_tmf_compl(struct qedf_ctx *qedf, struct fcoe_cqe *cqe, in qedf_process_tmf_compl() 2490 struct fcoe_cqe *cqe) in qedf_process_unsol_compl()
|
/linux/drivers/infiniband/hw/bnxt_re/ |
H A D | qplib_fp.c | 1569 struct cq_req *cqe = (struct cq_req *)hw_cqe; in __clean_cq() local 1579 struct cq_res_rc *cqe = (struct cq_res_rc *)hw_cqe; in __clean_cq() local 2384 struct bnxt_qplib_cqe *cqe; in __flush_sq() local 2425 struct bnxt_qplib_cqe *cqe; in __flush_rq() local 2614 struct bnxt_qplib_cqe *cqe; in bnxt_qplib_cq_process_req() local 2740 struct bnxt_qplib_cqe *cqe; in bnxt_qplib_cq_process_res_rc() local 2820 struct bnxt_qplib_cqe *cqe; in bnxt_qplib_cq_process_res_ud() local 2920 struct bnxt_qplib_cqe *cqe; in bnxt_qplib_cq_process_res_raweth_qp1() local 3012 struct bnxt_qplib_cqe *cqe; in bnxt_qplib_cq_process_terminal() local 3124 struct bnxt_qplib_cqe *cqe, in bnxt_qplib_process_flush_list() [all …]
|
/linux/tools/testing/vsock/ |
H A D | vsock_uring_test.c | 63 struct io_uring_cqe *cqe; in vsock_io_uring_client() local 144 struct io_uring_cqe *cqe; in vsock_io_uring_server() local
|
/linux/tools/testing/selftests/drivers/net/hw/ |
H A D | iou-zcrx.c | 213 static void process_accept(struct io_uring *ring, struct io_uring_cqe *cqe) in process_accept() 227 static void process_recvzc(struct io_uring *ring, struct io_uring_cqe *cqe) in process_recvzc() 275 struct io_uring_cqe *cqe; in server_loop() local
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/steering/hws/ |
H A D | send.c | 349 struct mlx5_cqe64 *cqe) in hws_send_engine_dump_error_cqe() 461 struct mlx5_cqe64 *cqe) in hws_send_engine_update_rule() 520 struct mlx5_cqe64 *cqe, in hws_send_engine_update() 606 struct mlx5_cqe64 *cqe; in hws_send_engine_poll_cq() local 890 struct mlx5_cqe64 *cqe; in hws_send_ring_alloc_cq() local 1232 struct mlx5_cqe64 cqe; in hws_send_wqe_fw() local
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/ |
H A D | rx.c | 249 struct mlx5_cqe64 *cqe, in mlx5e_xsk_skb_from_cqe_mpwrq_linear() 306 struct mlx5_cqe64 *cqe, in mlx5e_xsk_skb_from_cqe_linear()
|