Searched refs:mmkey (Results 1 – 6 of 6) sorted by relevance
| /linux/drivers/infiniband/hw/mlx5/ ! |
| H A D | odp.c | 219 mlx5r_deref_wait_odp_mkey(&mr->mmkey); in free_implicit_child_mr_work() 228 mlx5r_deref_odp_mkey(&imr->mmkey); in free_implicit_child_mr_work() 244 if (!refcount_inc_not_zero(&imr->mmkey.usecount)) in destroy_unused_implicit_child_mr() 251 mlx5r_deref_odp_mkey(&imr->mmkey); in destroy_unused_implicit_child_mr() 257 mlx5_base_mkey(mr->mmkey.key)); in destroy_unused_implicit_child_mr() 493 mr->ibmr.lkey = mr->mmkey.key; in implicit_get_child_mr() 494 mr->ibmr.rkey = mr->mmkey.key; in implicit_get_child_mr() 503 refcount_set(&mr->mmkey.usecount, 2); in implicit_get_child_mr() 527 refcount_inc(&ret->mmkey.usecount); in implicit_get_child_mr() 532 ret = xa_store(&dev->odp_mkeys, mlx5_base_mkey(mr->mmkey.key), in implicit_get_child_mr() [all …]
|
| H A D | mr.c | 133 WARN_ON(xa_load(&dev->odp_mkeys, mlx5_base_mkey(mr->mmkey.key))); in destroy_mkey() 135 return mlx5_core_destroy_mkey(dev->mdev, mr->mmkey.key); in destroy_mkey() 755 err = create_cache_mkey(ent, &mr->mmkey.key); in _mlx5_mr_cache_alloc() 764 mr->mmkey.key = pop_mkey_locked(ent); in _mlx5_mr_cache_alloc() 768 mr->mmkey.cache_ent = ent; in _mlx5_mr_cache_alloc() 769 mr->mmkey.type = MLX5_MKEY_MR; in _mlx5_mr_cache_alloc() 770 mr->mmkey.rb_key = ent->rb_key; in _mlx5_mr_cache_alloc() 771 mr->mmkey.cacheable = true; in _mlx5_mr_cache_alloc() 772 init_waitqueue_head(&mr->mmkey.wait); in _mlx5_mr_cache_alloc() 1079 err = mlx5_ib_create_mkey(dev, &mr->mmkey, in, inlen); in mlx5_ib_get_dma_mr() [all …]
|
| H A D | mlx5_ib.h | 689 struct mlx5_ib_mkey mmkey; member 756 struct mlx5_ib_mkey mmkey; member 1664 struct mlx5_ib_mkey *mmkey) in mlx5r_store_odp_mkey() argument 1666 refcount_set(&mmkey->usecount, 1); in mlx5r_store_odp_mkey() 1668 return xa_err(xa_store(&dev->odp_mkeys, mlx5_base_mkey(mmkey->key), in mlx5r_store_odp_mkey() 1669 mmkey, GFP_KERNEL)); in mlx5r_store_odp_mkey() 1673 static inline void mlx5r_deref_odp_mkey(struct mlx5_ib_mkey *mmkey) in mlx5r_deref_odp_mkey() argument 1675 if (refcount_dec_and_test(&mmkey->usecount)) in mlx5r_deref_odp_mkey() 1676 wake_up(&mmkey->wait); in mlx5r_deref_odp_mkey() 1680 static inline void mlx5r_deref_wait_odp_mkey(struct mlx5_ib_mkey *mmkey) in mlx5r_deref_wait_odp_mkey() argument [all …]
|
| H A D | umr.c | 437 mlx5_mkey_variant(mr->mmkey.key)); in mlx5r_umr_revoke_mr() 439 return mlx5r_umr_post_send_wait(dev, mr->mmkey.key, &wqe, false); in mlx5r_umr_revoke_mr() 476 mlx5_mkey_variant(mr->mmkey.key)); in mlx5r_umr_rereg_pd_access() 478 err = mlx5r_umr_post_send_wait(dev, mr->mmkey.key, &wqe, false); in mlx5r_umr_rereg_pd_access() 614 MLX5_SET(mkc, mkey_seg, mkey_7_0, mlx5_mkey_variant(mr->mmkey.key)); in mlx5r_umr_set_update_xlt_mkey_seg() 748 err = mlx5r_umr_post_send_wait(dev, mr->mmkey.key, &wqe, in _mlx5r_umr_update_mr_pas() 795 err = mlx5r_umr_post_send_wait(dev, mr->mmkey.key, &wqe, true); in _mlx5r_umr_update_mr_pas() 925 err = mlx5r_umr_post_send_wait(dev, mr->mmkey.key, &wqe, true); in mlx5r_umr_update_xlt() 964 mlx5_mkey_variant(mr->mmkey.key)); in mlx5r_umr_update_mr_page_shift() 966 err = mlx5r_umr_post_send_wait(dev, mr->mmkey.key, &wqe, false); in mlx5r_umr_update_mr_page_shift()
|
| H A D | wr.c | 164 int size = (mr->mmkey.ndescs + mr->meta_ndescs) * mr->desc_size; in set_reg_umr_seg() 193 int ndescs = ALIGN(mr->mmkey.ndescs + mr->meta_ndescs, 8) >> 1; in set_reg_mkey_seg() 221 int bcount = mr->desc_size * (mr->mmkey.ndescs + mr->meta_ndescs); in set_reg_data_seg() 643 int mr_list_size = (mr->mmkey.ndescs + mr->meta_ndescs) * mr->desc_size; in set_reg_wr() 890 pa_pi_mr.mmkey.ndescs = mr->mmkey.ndescs; in handle_reg_mr_integrity()
|
| H A D | restrack.c | 129 mlx5_mkey_to_idx(mr->mmkey.key)); in fill_res_mr_entry_raw()
|