Lines Matching refs:rf

16 	struct irdma_pci_f *rf = iwdev->rf;  in irdma_query_device()  local
17 struct pci_dev *pcidev = iwdev->rf->pcidev; in irdma_query_device()
18 struct irdma_hw_attrs *hw_attrs = &rf->sc_dev.hw_attrs; in irdma_query_device()
26 props->fw_ver = (u64)irdma_fw_major_ver(&rf->sc_dev) << 32 | in irdma_query_device()
27 irdma_fw_minor_ver(&rf->sc_dev); in irdma_query_device()
34 props->hw_ver = rf->pcidev->revision; in irdma_query_device()
37 props->max_qp = rf->max_qp - rf->used_qps; in irdma_query_device()
41 props->max_cq = rf->max_cq - rf->used_cqs; in irdma_query_device()
42 props->max_cqe = rf->max_cqe - 1; in irdma_query_device()
43 props->max_mr = rf->max_mr - rf->used_mrs; in irdma_query_device()
46 props->max_pd = rf->max_pd - rf->used_pds; in irdma_query_device()
55 props->max_ah = rf->max_ah; in irdma_query_device()
56 props->max_mcast_grp = rf->max_mcg; in irdma_query_device()
58 props->max_total_mcast_qp_attach = rf->max_qp * IRDMA_MAX_MGS_PER_CTX; in irdma_query_device()
60 props->max_srq = rf->max_srq - rf->used_srqs; in irdma_query_device()
119 props->max_msg_sz = iwdev->rf->sc_dev.hw_attrs.max_hw_outbound_msg_size; in irdma_query_port()
141 pfn = ((uintptr_t)ucontext->iwdev->rf->sc_dev.hw_regs[IRDMA_DB_ADDR_OFFSET] + in irdma_mmap_legacy()
142 pci_resource_start(ucontext->iwdev->rf->pcidev, 0)) >> PAGE_SHIFT; in irdma_mmap_legacy()
212 pci_resource_start(ucontext->iwdev->rf->pcidev, 0)) >> PAGE_SHIFT; in irdma_mmap()
250 cqp_request = irdma_alloc_and_get_cqp_request(&iwdev->rf->cqp, true); in irdma_alloc_push_page()
262 cqp_info->in.u.manage_push_page.cqp = &iwdev->rf->cqp.sc_cqp; in irdma_alloc_push_page()
265 status = irdma_handle_cqp_op(iwdev->rf, cqp_request); in irdma_alloc_push_page()
267 iwdev->rf->sc_dev.hw_attrs.max_hw_device_pages) { in irdma_alloc_push_page()
272 irdma_put_cqp_request(&iwdev->rf->cqp, cqp_request); in irdma_alloc_push_page()
293 struct irdma_uk_attrs *uk_attrs = &iwdev->rf->sc_dev.hw_attrs.uk_attrs; in irdma_alloc_ucontext()
321 uresp.max_qps = iwdev->rf->max_qp; in irdma_alloc_ucontext()
322 uresp.max_pds = iwdev->rf->sc_dev.hw_attrs.max_hw_pds; in irdma_alloc_ucontext()
323 uresp.wq_size = iwdev->rf->sc_dev.hw_attrs.max_qp_wr * 2; in irdma_alloc_ucontext()
329 u64 bar_off = (uintptr_t)iwdev->rf->sc_dev.hw_regs[IRDMA_DB_ADDR_OFFSET]; in irdma_alloc_ucontext()
398 struct irdma_sc_dev *dev = &iwdev->rf->sc_dev; in irdma_alloc_pd()
399 struct irdma_pci_f *rf = iwdev->rf; in irdma_alloc_pd() local
408 err = irdma_alloc_rsrc(rf, rf->allocated_pds, rf->max_pd, &pd_id, in irdma_alloc_pd()
409 &rf->next_pd); in irdma_alloc_pd()
431 irdma_free_rsrc(rf, rf->allocated_pds, pd_id); in irdma_alloc_pd()
446 irdma_free_rsrc(iwdev->rf, iwdev->rf->allocated_pds, iwpd->sc_pd.pd_id); in irdma_dealloc_pd()
509 bar_off = (uintptr_t)iwdev->rf->sc_dev.hw_regs[IRDMA_DB_ADDR_OFFSET]; in irdma_setup_push_mmap_entries()
562 irdma_cqp_qp_destroy_cmd(&iwdev->rf->sc_dev, &iwqp->sc_qp); in irdma_destroy_qp()
567 iwdev->rf->hwqp1_rsvd = false; in irdma_destroy_qp()
731 mem->va = dma_alloc_coherent(iwdev->rf->hw.device, mem->size, in irdma_setup_kmode_qp()
762 struct irdma_pci_f *rf = iwqp->iwdev->rf; in irdma_cqp_create_qp_cmd() local
768 cqp_request = irdma_alloc_and_get_cqp_request(&rf->cqp, true); in irdma_cqp_create_qp_cmd()
783 status = irdma_handle_cqp_op(rf, cqp_request); in irdma_cqp_create_qp_cmd()
784 irdma_put_cqp_request(&rf->cqp, cqp_request); in irdma_cqp_create_qp_cmd()
793 struct irdma_sc_dev *dev = &iwdev->rf->sc_dev; in irdma_roce_fill_and_set_qpctx_info()
836 struct irdma_sc_dev *dev = &iwdev->rf->sc_dev; in irdma_iw_fill_and_set_qpctx_info()
864 struct irdma_sc_dev *dev = &iwdev->rf->sc_dev; in irdma_validate_qp_attrs()
901 struct irdma_pci_f *rf = iwdev->rf; in irdma_setup_gsi_qp_rsrc() local
905 if (rf->rdma_ver <= IRDMA_GEN_2) { in irdma_setup_gsi_qp_rsrc()
910 spin_lock_irqsave(&rf->rsrc_lock, flags); in irdma_setup_gsi_qp_rsrc()
911 if (!rf->hwqp1_rsvd) { in irdma_setup_gsi_qp_rsrc()
913 rf->hwqp1_rsvd = true; in irdma_setup_gsi_qp_rsrc()
914 spin_unlock_irqrestore(&rf->rsrc_lock, flags); in irdma_setup_gsi_qp_rsrc()
916 spin_unlock_irqrestore(&rf->rsrc_lock, flags); in irdma_setup_gsi_qp_rsrc()
917 ret = irdma_alloc_rsrc(rf, rf->allocated_qps, rf->max_qp, in irdma_setup_gsi_qp_rsrc()
918 qp_num, &rf->next_qp); in irdma_setup_gsi_qp_rsrc()
923 ret = irdma_vchnl_req_add_vport(&rf->sc_dev, iwdev->vport_id, *qp_num, in irdma_setup_gsi_qp_rsrc()
927 irdma_free_rsrc(rf, rf->allocated_qps, *qp_num); in irdma_setup_gsi_qp_rsrc()
929 spin_lock_irqsave(&rf->rsrc_lock, flags); in irdma_setup_gsi_qp_rsrc()
930 rf->hwqp1_rsvd = false; in irdma_setup_gsi_qp_rsrc()
931 spin_unlock_irqrestore(&rf->rsrc_lock, flags); in irdma_setup_gsi_qp_rsrc()
954 struct irdma_pci_f *rf = iwdev->rf; in irdma_create_qp() local
960 struct irdma_sc_dev *dev = &rf->sc_dev; in irdma_create_qp()
1018 err_code = irdma_alloc_rsrc(rf, rf->allocated_qps, rf->max_qp, in irdma_create_qp()
1019 &qp_num, &rf->next_qp); in irdma_create_qp()
1089 irdma_cqp_qp_destroy_cmd(&rf->sc_dev, &iwqp->sc_qp); in irdma_create_qp()
1107 rf->qp_table[qp_num] = iwqp; in irdma_create_qp()
1282 struct irdma_sc_dev *dev = &iwdev->rf->sc_dev; in irdma_modify_qp_roce()
1424 irdma_add_arp(iwdev->rf, local_ip, udp_info->ipv4, in irdma_modify_qp_roce()
1645 struct irdma_sc_dev *dev = &iwdev->rf->sc_dev; in irdma_modify_qp()
1882 static void irdma_srq_free_rsrc(struct irdma_pci_f *rf, struct irdma_srq *iwsrq) in irdma_srq_free_rsrc() argument
1887 dma_free_coherent(rf->sc_dev.hw->device, iwsrq->kmem.size, in irdma_srq_free_rsrc()
1892 irdma_free_rsrc(rf, rf->allocated_srqs, srq->srq_uk.srq_id); in irdma_srq_free_rsrc()
1900 static void irdma_cq_free_rsrc(struct irdma_pci_f *rf, struct irdma_cq *iwcq) in irdma_cq_free_rsrc() argument
1905 dma_free_coherent(rf->sc_dev.hw->device, iwcq->kmem.size, in irdma_cq_free_rsrc()
1908 dma_free_coherent(rf->sc_dev.hw->device, in irdma_cq_free_rsrc()
1914 irdma_free_rsrc(rf, rf->allocated_cqs, cq->cq_uk.cq_id); in irdma_cq_free_rsrc()
1969 irdma_srq_wq_destroy(iwdev->rf, srq); in irdma_destroy_srq()
1970 irdma_srq_free_rsrc(iwdev->rf, iwsrq); in irdma_destroy_srq()
1999 irdma_cq_wq_destroy(iwdev->rf, cq); in irdma_destroy_cq()
2004 irdma_cq_free_rsrc(iwdev->rf, iwcq); in irdma_destroy_cq()
2029 struct irdma_pci_f *rf; in irdma_resize_cq() local
2035 rf = iwdev->rf; in irdma_resize_cq()
2037 if (!(rf->sc_dev.hw_attrs.uk_attrs.feature_flags & in irdma_resize_cq()
2044 if (entries > rf->max_cqe) in irdma_resize_cq()
2114 cqp_request = irdma_alloc_and_get_cqp_request(&rf->cqp, true); in irdma_resize_cq()
2131 ret = irdma_handle_cqp_op(rf, cqp_request); in irdma_resize_cq()
2132 irdma_put_cqp_request(&rf->cqp, cqp_request); in irdma_resize_cq()
2198 struct irdma_pci_f *rf = iwdev->rf; in irdma_modify_srq() local
2215 cqp_request = irdma_alloc_and_get_cqp_request(&rf->cqp, true); in irdma_modify_srq()
2230 status = irdma_handle_cqp_op(rf, cqp_request); in irdma_modify_srq()
2231 irdma_put_cqp_request(&rf->cqp, cqp_request); in irdma_modify_srq()
2300 mem->va = dma_alloc_coherent(iwdev->rf->hw.device, mem->size, in irdma_setup_kmode_srq()
2332 struct irdma_pci_f *rf = iwdev->rf; in irdma_create_srq() local
2339 uk_attrs = &rf->sc_dev.hw_attrs.uk_attrs; in irdma_create_srq()
2351 err_code = irdma_alloc_rsrc(rf, rf->allocated_srqs, rf->max_srq, in irdma_create_srq()
2352 &iwsrq->srq_num, &rf->next_srq); in irdma_create_srq()
2391 cqp_request = irdma_alloc_and_get_cqp_request(&rf->cqp, true); in irdma_create_srq()
2402 err_code = irdma_handle_cqp_op(rf, cqp_request); in irdma_create_srq()
2403 irdma_put_cqp_request(&rf->cqp, cqp_request); in irdma_create_srq()
2422 irdma_srq_wq_destroy(rf, &iwsrq->sc_srq); in irdma_create_srq()
2426 dma_free_coherent(rf->hw.device, iwsrq->kmem.size, in irdma_create_srq()
2429 irdma_free_rsrc(rf, rf->allocated_srqs, iwsrq->srq_num); in irdma_create_srq()
2473 struct irdma_pci_f *rf = iwdev->rf; in irdma_create_cq() local
2477 struct irdma_sc_dev *dev = &rf->sc_dev; in irdma_create_cq()
2495 err_code = irdma_alloc_rsrc(rf, rf->allocated_cqs, rf->max_cq, &cq_num, in irdma_create_cq()
2496 &rf->next_cq); in irdma_create_cq()
2514 if (attr->comp_vector < rf->ceqs_count) in irdma_create_cq()
2550 if (rf->sc_dev.hw_attrs.uk_attrs.feature_flags & in irdma_create_cq()
2579 if (entries < 1 || entries > rf->max_cqe) { in irdma_create_cq()
2631 cqp_request = irdma_alloc_and_get_cqp_request(&rf->cqp, true); in irdma_create_cq()
2643 err_code = irdma_handle_cqp_op(rf, cqp_request); in irdma_create_cq()
2644 irdma_put_cqp_request(&rf->cqp, cqp_request); in irdma_create_cq()
2661 rf->cq_table[cq_num] = iwcq; in irdma_create_cq()
2666 irdma_cq_wq_destroy(rf, cq); in irdma_create_cq()
2668 irdma_cq_free_rsrc(rf, iwcq); in irdma_create_cq()
2708 stag_idx = (stag & iwdev->rf->mr_stagmask) >> IRDMA_CQPSQ_STAG_IDX_S; in irdma_free_stag()
2709 irdma_free_rsrc(iwdev->rf, iwdev->rf->allocated_mrs, stag_idx); in irdma_free_stag()
2729 driver_key = random & ~iwdev->rf->mr_stagmask; in irdma_create_stag()
2730 next_stag_index = (random & iwdev->rf->mr_stagmask) >> 8; in irdma_create_stag()
2731 next_stag_index %= iwdev->rf->max_mr; in irdma_create_stag()
2733 ret = irdma_alloc_rsrc(iwdev->rf, iwdev->rf->allocated_mrs, in irdma_create_stag()
2734 iwdev->rf->max_mr, &stag_index, in irdma_create_stag()
2854 static int irdma_setup_pbles(struct irdma_pci_f *rf, struct irdma_mr *iwmr, in irdma_setup_pbles() argument
2865 status = irdma_get_pble(rf->pble_rsrc, palloc, iwmr->page_cnt, in irdma_setup_pbles()
2910 err = irdma_setup_pbles(iwdev->rf, iwmr, lvl); in irdma_handle_q_mem()
2978 irdma_free_pble(iwdev->rf->pble_rsrc, palloc); in irdma_handle_q_mem()
2998 cqp_request = irdma_alloc_and_get_cqp_request(&iwdev->rf->cqp, true); in irdma_hw_alloc_mw()
3014 cqp_info->in.u.mw_alloc.dev = &iwdev->rf->sc_dev; in irdma_hw_alloc_mw()
3016 status = irdma_handle_cqp_op(iwdev->rf, cqp_request); in irdma_hw_alloc_mw()
3017 irdma_put_cqp_request(&iwdev->rf->cqp, cqp_request); in irdma_hw_alloc_mw()
3064 cqp_request = irdma_alloc_and_get_cqp_request(&iwdev->rf->cqp, true); in irdma_dealloc_mw()
3076 cqp_info->in.u.dealloc_stag.dev = &iwdev->rf->sc_dev; in irdma_dealloc_mw()
3078 irdma_handle_cqp_op(iwdev->rf, cqp_request); in irdma_dealloc_mw()
3079 irdma_put_cqp_request(&iwdev->rf->cqp, cqp_request); in irdma_dealloc_mw()
3100 cqp_request = irdma_alloc_and_get_cqp_request(&iwdev->rf->cqp, true); in irdma_hw_alloc_stag()
3115 cqp_info->in.u.alloc_stag.dev = &iwdev->rf->sc_dev; in irdma_hw_alloc_stag()
3117 status = irdma_handle_cqp_op(iwdev->rf, cqp_request); in irdma_hw_alloc_stag()
3118 irdma_put_cqp_request(&iwdev->rf->cqp, cqp_request); in irdma_hw_alloc_stag()
3164 err_code = irdma_get_pble(iwdev->rf->pble_rsrc, palloc, iwmr->page_cnt, in irdma_alloc_mr()
3177 irdma_free_pble(iwdev->rf->pble_rsrc, palloc); in irdma_alloc_mr()
3250 cqp_request = irdma_alloc_and_get_cqp_request(&iwdev->rf->cqp, true); in irdma_hwreg_mr()
3262 iwdev->rf->sc_dev.hw_attrs.uk_attrs.hw_rev); in irdma_hwreg_mr()
3263 if (iwdev->rf->sc_dev.hw_attrs.uk_attrs.feature_flags & IRDMA_FEATURE_ATOMIC_OPS) in irdma_hwreg_mr()
3287 cqp_info->in.u.mr_reg_non_shared.dev = &iwdev->rf->sc_dev; in irdma_hwreg_mr()
3289 ret = irdma_handle_cqp_op(iwdev->rf, cqp_request); in irdma_hwreg_mr()
3290 irdma_put_cqp_request(&iwdev->rf->cqp, cqp_request); in irdma_hwreg_mr()
3309 err = irdma_setup_pbles(iwdev->rf, iwmr, lvl); in irdma_reg_user_mr_type_mem()
3317 irdma_free_pble(iwdev->rf->pble_rsrc, &iwpbl->pble_alloc); in irdma_reg_user_mr_type_mem()
3346 irdma_free_pble(iwdev->rf->pble_rsrc, &iwpbl->pble_alloc); in irdma_reg_user_mr_type_mem()
3373 iwdev->rf->sc_dev.hw_attrs.page_size_cap : SZ_4K; in irdma_alloc_iwmr()
3474 if (iwdev->rf->sc_dev.hw_attrs.uk_attrs.feature_flags & IRDMA_FEATURE_CQ_RESIZE) in irdma_reg_user_mr_type_cq()
3520 if (len > iwdev->rf->sc_dev.hw_attrs.max_mr_size) in irdma_reg_user_mr()
3596 if (len > iwdev->rf->sc_dev.hw_attrs.max_mr_size) in irdma_reg_user_mr_dmabuf()
3644 cqp_request = irdma_alloc_and_get_cqp_request(&iwdev->rf->cqp, true); in irdma_hwdereg_mr()
3659 cqp_info->in.u.dealloc_stag.dev = &iwdev->rf->sc_dev; in irdma_hwdereg_mr()
3661 status = irdma_handle_cqp_op(iwdev->rf, cqp_request); in irdma_hwdereg_mr()
3662 irdma_put_cqp_request(&iwdev->rf->cqp, cqp_request); in irdma_hwdereg_mr()
3697 iwdev->rf->sc_dev.hw_attrs.page_size_cap, in irdma_rereg_mr_trans()
3744 if (len > iwdev->rf->sc_dev.hw_attrs.max_mr_size) in irdma_rereg_user_mr()
3764 irdma_free_pble(iwdev->rf->pble_rsrc, in irdma_rereg_user_mr()
3921 irdma_free_pble(iwdev->rf->pble_rsrc, &iwpbl->pble_alloc); in irdma_dereg_mr()
3952 dev = &iwqp->iwdev->rf->sc_dev; in irdma_post_send()
4620 irdma_fw_major_ver(&iwdev->rf->sc_dev), in irdma_get_dev_fw_str()
4621 irdma_fw_minor_ver(&iwdev->rf->sc_dev)); in irdma_get_dev_fw_str()
4633 struct irdma_sc_dev *dev = &iwdev->rf->sc_dev; in irdma_alloc_hw_port_stats()
4656 if (iwdev->rf->rdma_ver >= IRDMA_GEN_2) in irdma_get_hw_stats()
4657 irdma_cqp_gather_stats_cmd(&iwdev->rf->sc_dev, iwdev->vsi.pestat, true); in irdma_get_hw_stats()
4659 irdma_cqp_gather_stats_gen1(&iwdev->rf->sc_dev, iwdev->vsi.pestat); in irdma_get_hw_stats()
4689 static void mcast_list_add(struct irdma_pci_f *rf, in mcast_list_add() argument
4692 list_add(&new_elem->list, &rf->mc_qht_list.list); in mcast_list_add()
4710 static struct mc_table_list *mcast_list_lookup_ip(struct irdma_pci_f *rf, in mcast_list_lookup_ip() argument
4716 list_for_each_safe (pos, q, &rf->mc_qht_list.list) { in mcast_list_lookup_ip()
4741 cqp_request = irdma_alloc_and_get_cqp_request(&iwdev->rf->cqp, true); in irdma_mcast_cqp_op()
4750 cqp_info->in.u.mc_create.cqp = &iwdev->rf->cqp.sc_cqp; in irdma_mcast_cqp_op()
4751 status = irdma_handle_cqp_op(iwdev->rf, cqp_request); in irdma_mcast_cqp_op()
4752 irdma_put_cqp_request(&iwdev->rf->cqp, cqp_request); in irdma_mcast_cqp_op()
4800 struct irdma_pci_f *rf = iwdev->rf; in irdma_attach_mcast() local
4834 spin_lock_irqsave(&rf->qh_list_lock, flags); in irdma_attach_mcast()
4835 mc_qht_elem = mcast_list_lookup_ip(rf, ip_addr); in irdma_attach_mcast()
4839 spin_unlock_irqrestore(&rf->qh_list_lock, flags); in irdma_attach_mcast()
4847 ret = irdma_alloc_rsrc(rf, rf->allocated_mcgs, rf->max_mcg, in irdma_attach_mcast()
4848 &mgn, &rf->next_mcg); in irdma_attach_mcast()
4858 dma_mem_mc->va = dma_alloc_coherent(rf->hw.device, in irdma_attach_mcast()
4863 irdma_free_rsrc(rf, rf->allocated_mcgs, mgn); in irdma_attach_mcast()
4875 mc_qht_elem->mc_grp_ctx.hmc_fcn_id = iwdev->rf->sc_dev.hmc_fn_id; in irdma_attach_mcast()
4880 spin_lock_irqsave(&rf->qh_list_lock, flags); in irdma_attach_mcast()
4881 mcast_list_add(rf, mc_qht_elem); in irdma_attach_mcast()
4885 spin_unlock_irqrestore(&rf->qh_list_lock, flags); in irdma_attach_mcast()
4893 spin_unlock_irqrestore(&rf->qh_list_lock, flags); in irdma_attach_mcast()
4915 dma_free_coherent(rf->hw.device, in irdma_attach_mcast()
4920 irdma_free_rsrc(rf, rf->allocated_mcgs, in irdma_attach_mcast()
4940 struct irdma_pci_f *rf = iwdev->rf; in irdma_detach_mcast() local
4955 spin_lock_irqsave(&rf->qh_list_lock, flags); in irdma_detach_mcast()
4956 mc_qht_elem = mcast_list_lookup_ip(rf, ip_addr); in irdma_detach_mcast()
4958 spin_unlock_irqrestore(&rf->qh_list_lock, flags); in irdma_detach_mcast()
4968 spin_unlock_irqrestore(&rf->qh_list_lock, flags); in irdma_detach_mcast()
4974 spin_lock_irqsave(&rf->qh_list_lock, flags); in irdma_detach_mcast()
4975 mcast_list_add(rf, mc_qht_elem); in irdma_detach_mcast()
4976 spin_unlock_irqrestore(&rf->qh_list_lock, flags); in irdma_detach_mcast()
4980 dma_free_coherent(rf->hw.device, in irdma_detach_mcast()
4985 irdma_free_rsrc(rf, rf->allocated_mcgs, in irdma_detach_mcast()
4989 spin_unlock_irqrestore(&rf->qh_list_lock, flags); in irdma_detach_mcast()
5004 struct irdma_pci_f *rf = iwdev->rf; in irdma_create_hw_ah() local
5007 err = irdma_alloc_rsrc(rf, rf->allocated_ahs, rf->max_ah, &ah->sc_ah.ah_info.ah_idx, in irdma_create_hw_ah()
5008 &rf->next_ah); in irdma_create_hw_ah()
5012 err = irdma_ah_cqp_op(rf, &ah->sc_ah, IRDMA_OP_AH_CREATE, sleep, in irdma_create_hw_ah()
5024 irdma_cqp_ce_handler(rf, &rf->ccq.sc_cq); in irdma_create_hw_ah()
5037 irdma_free_rsrc(iwdev->rf, iwdev->rf->allocated_ahs, ah->sc_ah.ah_info.ah_idx); in irdma_create_hw_ah()
5049 struct irdma_pci_f *rf = iwdev->rf; in irdma_setup_ah() local
5059 irdma_sc_init_ah(&rf->sc_dev, sc_ah); in irdma_setup_ah()
5106 ah_info->dst_arpindex = irdma_add_arp(iwdev->rf, ah_info->dest_ip_addr, in irdma_setup_ah()
5143 hash_for_each_possible(iwdev->rf->ah_hash_tbl, ah, list, key) { in irdma_ah_exists()
5169 mutex_lock(&iwdev->rf->ah_tbl_lock); in irdma_destroy_ah()
5171 mutex_unlock(&iwdev->rf->ah_tbl_lock); in irdma_destroy_ah()
5176 mutex_unlock(&iwdev->rf->ah_tbl_lock); in irdma_destroy_ah()
5179 irdma_ah_cqp_op(iwdev->rf, &ah->sc_ah, IRDMA_OP_AH_DESTROY, in irdma_destroy_ah()
5182 irdma_free_rsrc(iwdev->rf, iwdev->rf->allocated_ahs, in irdma_destroy_ah()
5213 mutex_lock(&iwdev->rf->ah_tbl_lock); in irdma_create_user_ah()
5217 mutex_unlock(&iwdev->rf->ah_tbl_lock); in irdma_create_user_ah()
5229 hash_add(iwdev->rf->ah_hash_tbl, &parent_ah->list, key); in irdma_create_user_ah()
5233 mutex_unlock(&iwdev->rf->ah_tbl_lock); in irdma_create_user_ah()
5415 struct pci_dev *pcidev = iwdev->rf->pcidev; in irdma_init_rdma_device()
5423 iwdev->ibdev.num_comp_vectors = iwdev->rf->ceqs_count; in irdma_init_rdma_device()
5426 if (iwdev->rf->rdma_ver == IRDMA_GEN_1) in irdma_init_rdma_device()
5428 if (iwdev->rf->rdma_ver >= IRDMA_GEN_3) in irdma_init_rdma_device()
5472 dma_set_max_seg_size(iwdev->rf->hw.device, UINT_MAX); in irdma_ib_register_device()
5473 ret = ib_register_device(&iwdev->ibdev, "irdma%d", iwdev->rf->hw.device); in irdma_ib_register_device()
5502 irdma_ctrl_deinit_hw(iwdev->rf); in irdma_ib_dealloc_device()
5503 if (iwdev->rf->vchnl_wq) in irdma_ib_dealloc_device()
5504 destroy_workqueue(iwdev->rf->vchnl_wq); in irdma_ib_dealloc_device()