Searched refs:compl_desc (Results 1 – 7 of 7) sorted by relevance
| /linux/drivers/net/ethernet/google/gve/ |
| H A D | gve_rx_dqo.c | 437 const struct gve_rx_compl_desc_dqo *compl_desc, in gve_rx_skb_hash() argument 447 skb_set_hash(skb, le32_to_cpu(compl_desc->hash), hash_type); in gve_rx_skb_hash() 489 if (!(ctx->compl_desc->ts_sub_nsecs_low & GVE_DQO_RX_HWTSTAMP_VALID)) in gve_xdp_rx_timestamp() 493 le32_to_cpu(ctx->compl_desc->ts)); in gve_xdp_rx_timestamp() 708 const struct gve_rx_compl_desc_dqo *compl_desc, in gve_rx_xsk_dqo() argument 713 int buf_len = compl_desc->packet_len; in gve_rx_xsk_dqo() 723 gve_xdp->compl_desc = compl_desc; in gve_rx_xsk_dqo() 779 const struct gve_rx_compl_desc_dqo *compl_desc, in gve_rx_dqo() argument 782 const u16 buffer_id = le16_to_cpu(compl_desc->buf_id); in gve_rx_dqo() 783 const bool hbo = compl_desc->header_buffer_overflow; in gve_rx_dqo() [all …]
|
| H A D | gve_tx_dqo.c | 1397 struct gve_tx_compl_desc *compl_desc = in gve_clean_tx_done_dqo() local 1401 if (compl_desc->generation == tx->dqo_compl.cur_gen_bit) in gve_clean_tx_done_dqo() 1410 type = compl_desc->type; in gve_clean_tx_done_dqo() 1414 u16 tx_head = le16_to_cpu(compl_desc->tx_head); in gve_clean_tx_done_dqo() 1418 u16 compl_tag = le16_to_cpu(compl_desc->completion_tag); in gve_clean_tx_done_dqo() 1432 u16 compl_tag = le16_to_cpu(compl_desc->completion_tag); in gve_clean_tx_done_dqo() 1438 u16 compl_tag = le16_to_cpu(compl_desc->completion_tag); in gve_clean_tx_done_dqo() 1474 struct gve_tx_compl_desc *compl_desc; in gve_tx_poll_dqo() local 1493 compl_desc = &tx->dqo.compl_ring[tx->dqo_compl.head]; in gve_tx_poll_dqo() 1494 return compl_desc->generation != tx->dqo_compl.cur_gen_bit; in gve_tx_poll_dqo() [all …]
|
| H A D | gve.h | 212 const struct gve_rx_compl_desc_dqo *compl_desc; member
|
| /linux/drivers/dma/ioat/ |
| H A D | prep.c | 157 struct ioat_ring_ent *compl_desc; in __ioat_prep_xor_lock() local 230 compl_desc = ioat_get_ring_ent(ioat_chan, idx + i); in __ioat_prep_xor_lock() 231 compl_desc->txd.flags = flags & DMA_PREP_INTERRUPT; in __ioat_prep_xor_lock() 232 hw = compl_desc->hw; in __ioat_prep_xor_lock() 238 dump_desc_dbg(ioat_chan, compl_desc); in __ioat_prep_xor_lock() 241 return &compl_desc->txd; in __ioat_prep_xor_lock() 346 struct ioat_ring_ent *compl_desc; in __ioat_prep_pq_lock() local 443 compl_desc = desc; in __ioat_prep_pq_lock() 446 compl_desc = ioat_get_ring_ent(ioat_chan, idx + i); in __ioat_prep_pq_lock() 447 compl_desc->txd.flags = flags & DMA_PREP_INTERRUPT; in __ioat_prep_pq_lock() [all …]
|
| /linux/drivers/crypto/ccree/ |
| H A D | cc_request_mgr.c | 29 struct cc_hw_desc compl_desc; member 170 hw_desc_init(&req_mgr_h->compl_desc); in cc_req_mgr_init() 171 set_din_const(&req_mgr_h->compl_desc, 0, sizeof(u32)); in cc_req_mgr_init() 172 set_dout_dlli(&req_mgr_h->compl_desc, req_mgr_h->dummy_comp_buff_dma, in cc_req_mgr_init() 174 set_flow_mode(&req_mgr_h->compl_desc, BYPASS); in cc_req_mgr_init() 175 set_queue_last_ind(drvdata, &req_mgr_h->compl_desc); in cc_req_mgr_init() 313 enqueue_seq(drvdata, &req_mgr_h->compl_desc, 1); in cc_do_send_request()
|
| /linux/drivers/usb/dwc2/ |
| H A D | gadget.c | 1020 hs_ep->compl_desc = 0; in dwc2_gadget_start_isoc_ddma() 2189 desc_sts = hs_ep->desc_list[hs_ep->compl_desc].status; in dwc2_gadget_complete_isoc_request_ddma() 2224 hs_ep->compl_desc++; in dwc2_gadget_complete_isoc_request_ddma() 2225 if (hs_ep->compl_desc > (MAX_DMA_DESC_NUM_HS_ISOC - 1)) in dwc2_gadget_complete_isoc_request_ddma() 2226 hs_ep->compl_desc = 0; in dwc2_gadget_complete_isoc_request_ddma() 2227 desc_sts = hs_ep->desc_list[hs_ep->compl_desc].status; in dwc2_gadget_complete_isoc_request_ddma() 2250 hs_ep->compl_desc = 0; in dwc2_gadget_handle_isoc_bna() 4125 hs_ep->compl_desc = 0; in dwc2_hsotg_ep_enable()
|
| H A D | core.h | 156 unsigned int compl_desc; member
|