Home
last modified time | relevance | path

Searched refs:xa_load (Results 1 – 25 of 244) sorted by relevance

12345678910

/linux/lib/
H A Dtest_xarray.c63 XA_BUG_ON(xa, xa_load(xa, index) != NULL); in xa_erase_index()
153 void *entry = xa_load(xa, j); in check_xa_load()
164 void *entry = xa_load(xa, j); in check_xa_load()
344 XA_BUG_ON(xa, xa_load(xa, 1) != NULL); in check_xa_shrink()
349 XA_BUG_ON(xa, xa_load(xa, 0) != xa_mk_value(0)); in check_xa_shrink()
356 XA_BUG_ON(xa, xa_load(xa, max) != xa_mk_value(0)); in check_xa_shrink()
357 XA_BUG_ON(xa, xa_load(xa, max + 1) != NULL); in check_xa_shrink()
366 XA_BUG_ON(xa, xa_load(xa, max + 1) != NULL); in check_xa_shrink()
379 XA_BUG_ON(xa, xa_load(xa, i - 1) != NULL); in check_insert()
380 XA_BUG_ON(xa, xa_load(xa, i + 1) != NULL); in check_insert()
[all …]
H A Dref_tracker.c388 if (!xa_load(&debugfs_dentries, index) || in ref_tracker_debugfs_show()
432 dentry = xa_load(&debugfs_dentries, (unsigned long)dir); in ref_tracker_dir_debugfs()
465 symlink = xa_load(&debugfs_symlinks, (unsigned long)dir); in ref_tracker_dir_symlink()
466 dentry = xa_load(&debugfs_dentries, (unsigned long)dir); in ref_tracker_dir_symlink()
/linux/drivers/gpu/drm/imagination/
H A Dpvr_free_list.h143 free_list = pvr_free_list_get(xa_load(&pvr_file->free_list_handles, handle)); in pvr_free_list_lookup()
172 free_list = xa_load(&pvr_dev->free_list_ids, id); in pvr_free_list_lookup_id()
H A Dpvr_context.h165 ctx = pvr_context_get(xa_load(&pvr_file->ctx_handles, handle)); in pvr_context_lookup()
195 ctx = xa_load(&pvr_dev->ctx_ids, id); in pvr_context_lookup_id()
H A Dpvr_hwrt.h105 hwrt = xa_load(&pvr_file->hwrt_handles, handle); in pvr_hwrt_dataset_lookup()
/linux/drivers/net/ethernet/intel/ice/
H A Dice_adapter.c102 adapter = xa_load(&ice_adapters, index); in ice_adapter_get()
138 adapter = xa_load(&ice_adapters, index); in ice_adapter_put()
H A Dice_eswitch.c185 repr = xa_load(&pf->eswitch.reprs, *repr_id); in ice_eswitch_update_repr()
588 struct ice_repr *repr = xa_load(&pf->eswitch.reprs, vf->repr_id); in ice_eswitch_detach_vf()
606 struct ice_repr *repr = xa_load(&pf->eswitch.reprs, sf->repr_id); in ice_eswitch_detach_sf()
630 repr = xa_load(&eswitch->reprs, le16_to_cpu(desc->src_vsi)); in ice_eswitch_get_target()
H A Dice_eswitch_br.c312 port = xa_load(&bridge->ports, vsi_idx); in ice_esw_br_port_vlan_lookup()
318 vlan = xa_load(&port->vlans, vid); in ice_esw_br_port_vlan_lookup()
742 port = xa_load(&bridge->ports, vsi_idx); in ice_eswitch_br_port_vlan_add()
753 vlan = xa_load(&port->vlans, vid); in ice_eswitch_br_port_vlan_add()
777 port = xa_load(&bridge->ports, vsi_idx); in ice_eswitch_br_port_vlan_del()
781 vlan = xa_load(&port->vlans, vid); in ice_eswitch_br_port_vlan_del()
/linux/Documentation/translations/zh_CN/core-api/
H A Dxarray.rst63 然后你可以用xa_store()来设置条目,用xa_load()来获取条目。xa_store将用新的条目覆盖任
153 * xa_load()
195 如果你想利用锁来保护你存储在XArray中的数据结构,你可以在调用xa_load()之前调用xa_lock(),然后在
/linux/drivers/iommu/iommufd/
H A Ddevice.c82 igroup = xa_load(&ictx->groups, id); in iommufd_get_group()
362 attach = xa_load(&igroup->pasid_attach, pasid); in iommufd_group_device_num()
409 return !xa_load(&igroup->pasid_attach, pasid); in iommufd_group_first_attach()
445 attach = xa_load(&idev->igroup->pasid_attach, pasid); in iommufd_device_is_attached()
446 return xa_load(&attach->device_array, idev->obj.id); in iommufd_device_is_attached()
470 attach = xa_load(&igroup->pasid_attach, IOMMU_NO_PASID); in iommufd_hwpt_pasid_compat()
695 attach = xa_load(&igroup->pasid_attach, pasid); in iommufd_hw_pagetable_detach()
742 attach = xa_load(&igroup->pasid_attach, IOMMU_NO_PASID); in iommufd_group_remove_reserved_iova()
759 attach = xa_load(&igroup->pasid_attach, IOMMU_NO_PASID); in iommufd_group_do_replace_reserved_iova()
795 attach = xa_load(&igroup->pasid_attach, pasid); in iommufd_device_do_replace()
/linux/drivers/firmware/arm_scmi/
H A Dperf.c584 opp = xa_load(&dom->opps_by_lvl, min_perf); in scmi_perf_limits_set()
592 opp = xa_load(&dom->opps_by_lvl, max_perf); in scmi_perf_limits_set()
663 opp = xa_load(&dom->opps_by_idx, *min_perf); in scmi_perf_limits_get()
669 opp = xa_load(&dom->opps_by_idx, *max_perf); in scmi_perf_limits_get()
733 opp = xa_load(&dom->opps_by_lvl, level); in scmi_perf_level_set()
796 opp = xa_load(&dom->opps_by_idx, *level); in scmi_perf_level_get()
979 opp = xa_load(&dom->opps_by_idx, level); in scmi_dvfs_freq_get()
1125 opp = xa_load(&dom->opps_by_lvl, index); in scmi_perf_xlate_opp_to_freq()
1131 opp = xa_load(&dom->opps_by_idx, index); in scmi_perf_xlate_opp_to_freq()
/linux/virt/lib/
H A Dirqbypass.c111 consumer = xa_load(&consumers, index); in irq_bypass_register_producer()
177 producer = xa_load(&producers, index); in irq_bypass_register_consumer()
/linux/fs/xfs/libxfs/
H A Dxfs_group.c40 xg = xa_load(&mp->m_groups[type].xa, index); in xfs_group_get()
81 xg = xa_load(&mp->m_groups[type].xa, index); in xfs_group_grab()
/linux/drivers/infiniband/hw/hns/
H A Dhns_roce_bond.c52 struct hns_roce_die_info *die_info = xa_load(&roce_bond_xa, bus_num); in hns_roce_get_bond_grp()
281 struct hns_roce_die_info *die_info = xa_load(&roce_bond_xa, bus_num); in alloc_bond_id()
306 struct hns_roce_die_info *die_info = xa_load(&roce_bond_xa, bus_num); in remove_bond_id()
854 if (xa_load(&roce_bond_xa, bus_num)) in hns_roce_alloc_bond_grp()
956 die_info = xa_load(&roce_bond_xa, bus_num); in hns_roce_bond_suspend()
989 die_info = xa_load(&roce_bond_xa, bus_num); in hns_roce_bond_resume()
/linux/drivers/gpu/drm/tegra/
H A Duapi.c173 context = xa_load(&fpriv->contexts, args->context); in tegra_drm_ioctl_channel_close()
203 context = xa_load(&fpriv->contexts, args->context); in tegra_drm_ioctl_channel_map()
284 context = xa_load(&fpriv->contexts, args->context); in tegra_drm_ioctl_channel_unmap()
/linux/drivers/net/ethernet/mellanox/mlx5/core/steering/sws/
H A Ddr_domain.c81 recalc_cs_ft = xa_load(&dmn->csum_fts_xa, vport_num); in mlx5dr_domain_get_recalc_cs_ft_addr()
329 vport_caps = xa_load(&caps->vports.vports_caps_xa, vport); in mlx5dr_domain_get_vport_cap()
546 peer = xa_load(&dmn->peer_dmn_xa, peer_vhca_id); in mlx5dr_domain_set_peer()
552 peer = xa_load(&dmn->peer_dmn_xa, peer_vhca_id); in mlx5dr_domain_set_peer()
/linux/drivers/net/ethernet/mellanox/mlx5/core/
H A Dpagealloc.c106 root = xa_load(&dev->priv.page_root_xa, function); in page_root_per_function()
177 root = xa_load(&dev->priv.page_root_xa, function); in find_fw_page()
259 root = xa_load(&dev->priv.page_root_xa, fwp->function); in free_fwp()
443 root = xa_load(&dev->priv.page_root_xa, function); in release_all_pages()
504 root = xa_load(&dev->priv.page_root_xa, get_function(func_id, ec_function)); in reclaim_pages_cmd()
H A Deq.c815 irq = xa_load(&table->comp_irqs, vecidx); in comp_irq_release_pci()
863 irq = xa_load(&table->comp_irqs, vecidx); in comp_irq_release_sf()
1018 irq = xa_load(&table->comp_irqs, vecidx); in create_comp_eq()
1064 eq = xa_load(&table->comp_eqs, vecidx); in mlx5_comp_eqn_get()
1095 eq = xa_load(&table->comp_eqs, vector); in mlx5_comp_irqn_get()
1112 eq = xa_load(&table->comp_eqs, vector); in mlx5_comp_irq_get_affinity_mask()
/linux/drivers/infiniband/hw/cxgb4/
H A Dev.c127 qhp = xa_load(&dev->qps, CQE_QPID(err_cqe)); in c4iw_ev_dispatch()
228 chp = xa_load(&dev->cqs, qid); in c4iw_ev_handler()
/linux/drivers/gpu/drm/xe/
H A Dxe_tlb_inval_job.c212 xe_assert(gt_to_xe(job->q->gt), !xa_load(&job->dep.drm.dependencies, 0)); in xe_tlb_inval_job_alloc_dep()
249 xa_load(&job->dep.drm.dependencies, 0) == in xe_tlb_inval_job_push()
/linux/drivers/net/ethernet/mellanox/mlx5/core/en/
H A Dmapping.c118 mi = xa_load(&ctx->xarray, index); in mapping_remove()
141 mi = xa_load(&ctx->xarray, index); in mapping_find()
/linux/include/linux/
H A Dphy_link_topology.h55 pdn = xa_load(&topo->phys, phyindex); in phy_link_topo_get_phy()
/linux/drivers/cxl/core/
H A Dcdat.c154 dent = xa_load(dsmas_xa, dslbis->handle); in cdat_dslbis_handler()
669 perf_ctx = xa_load(usp_xa, index); in cxl_endpoint_gather_bandwidth()
795 us_ctx = xa_load(res_xa, us_index); in DEFINE_FREE()
879 hb_ctx = xa_load(hb_xa, hb_index); in cxl_rp_gather_bandwidth()
928 mw_ctx = xa_load(mw_xa, mw_index); in cxl_hb_gather_bandwidth()
/linux/block/
H A Dblk-mq-sched.c462 ctx = xa_load(elv_tbl, q->id); in blk_mq_free_sched_res_batch()
587 ctx = xa_load(elv_tbl, q->id); in blk_mq_alloc_sched_res_batch()
604 ctx = xa_load(elv_tbl, q->id); in blk_mq_alloc_sched_res_batch()
/linux/drivers/net/ethernet/mellanox/mlx5/core/steering/hws/
H A Dvport.c75 entry = xa_load(&ctx->vports.vport_gvmi_xa, vport); in mlx5hws_vport_get_gvmi()

12345678910