Home
last modified time | relevance | path

Searched refs:xa_load (Results 1 – 25 of 216) sorted by relevance

123456789

/linux/lib/
H A Dtest_xarray.c63 XA_BUG_ON(xa, xa_load(xa, index) != NULL); in xa_erase_index()
153 void *entry = xa_load(xa, j); in check_xa_load()
164 void *entry = xa_load(xa, j); in check_xa_load()
344 XA_BUG_ON(xa, xa_load(xa, 1) != NULL); in check_xa_shrink()
349 XA_BUG_ON(xa, xa_load(xa, 0) != xa_mk_value(0)); in check_xa_shrink()
356 XA_BUG_ON(xa, xa_load(xa, max) != xa_mk_value(0)); in check_xa_shrink()
357 XA_BUG_ON(xa, xa_load(xa, max + 1) != NULL); in check_xa_shrink()
366 XA_BUG_ON(xa, xa_load(xa, max + 1) != NULL); in check_xa_shrink()
379 XA_BUG_ON(xa, xa_load(xa, i - 1) != NULL); in check_insert()
380 XA_BUG_ON(xa, xa_load(xa, i + 1) != NULL); in check_insert()
[all …]
H A Dref_tracker.c388 if (!xa_load(&debugfs_dentries, index) || in ref_tracker_debugfs_show()
432 dentry = xa_load(&debugfs_dentries, (unsigned long)dir); in ref_tracker_dir_debugfs()
465 symlink = xa_load(&debugfs_symlinks, (unsigned long)dir); in ref_tracker_dir_symlink()
466 dentry = xa_load(&debugfs_dentries, (unsigned long)dir); in ref_tracker_dir_symlink()
/linux/drivers/gpu/drm/imagination/
H A Dpvr_free_list.h143 free_list = pvr_free_list_get(xa_load(&pvr_file->free_list_handles, handle)); in pvr_free_list_lookup()
172 free_list = xa_load(&pvr_dev->free_list_ids, id); in pvr_free_list_lookup_id()
H A Dpvr_hwrt.h105 hwrt = xa_load(&pvr_file->hwrt_handles, handle); in pvr_hwrt_dataset_lookup()
/linux/drivers/net/ethernet/intel/ice/
H A Dice_adapter.c102 adapter = xa_load(&ice_adapters, index); in ice_adapter_get()
138 adapter = xa_load(&ice_adapters, index); in ice_adapter_put()
H A Dice_eswitch.c185 repr = xa_load(&pf->eswitch.reprs, *repr_id); in ice_eswitch_update_repr()
588 struct ice_repr *repr = xa_load(&pf->eswitch.reprs, vf->repr_id); in ice_eswitch_detach_vf()
606 struct ice_repr *repr = xa_load(&pf->eswitch.reprs, sf->repr_id); in ice_eswitch_detach_sf()
630 repr = xa_load(&eswitch->reprs, le16_to_cpu(desc->src_vsi)); in ice_eswitch_get_target()
H A Dice_eswitch_br.c312 port = xa_load(&bridge->ports, vsi_idx); in ice_esw_br_port_vlan_lookup()
318 vlan = xa_load(&port->vlans, vid); in ice_esw_br_port_vlan_lookup()
742 port = xa_load(&bridge->ports, vsi_idx); in ice_eswitch_br_port_vlan_add()
753 vlan = xa_load(&port->vlans, vid); in ice_eswitch_br_port_vlan_add()
777 port = xa_load(&bridge->ports, vsi_idx); in ice_eswitch_br_port_vlan_del()
781 vlan = xa_load(&port->vlans, vid); in ice_eswitch_br_port_vlan_del()
/linux/Documentation/translations/zh_CN/core-api/
H A Dxarray.rst63 然后你可以用xa_store()来设置条目,用xa_load()来获取条目。xa_store将用新的条目覆盖任
153 * xa_load()
195 如果你想利用锁来保护你存储在XArray中的数据结构,你可以在调用xa_load()之前调用xa_lock(),然后在
/linux/drivers/iommu/iommufd/
H A Ddevice.c82 igroup = xa_load(&ictx->groups, id); in iommufd_get_group()
362 attach = xa_load(&igroup->pasid_attach, pasid); in iommufd_group_device_num()
409 return !xa_load(&igroup->pasid_attach, pasid); in iommufd_group_first_attach()
445 attach = xa_load(&idev->igroup->pasid_attach, pasid); in iommufd_device_is_attached()
446 return xa_load(&attach->device_array, idev->obj.id); in iommufd_device_is_attached()
470 attach = xa_load(&igroup->pasid_attach, IOMMU_NO_PASID); in iommufd_hwpt_pasid_compat()
695 attach = xa_load(&igroup->pasid_attach, pasid); in iommufd_hw_pagetable_detach()
742 attach = xa_load(&igroup->pasid_attach, IOMMU_NO_PASID); in iommufd_group_remove_reserved_iova()
759 attach = xa_load(&igroup->pasid_attach, IOMMU_NO_PASID); in iommufd_group_do_replace_reserved_iova()
795 attach = xa_load(&igroup->pasid_attach, pasid); in iommufd_device_do_replace()
/linux/virt/lib/
H A Dirqbypass.c111 consumer = xa_load(&consumers, index); in irq_bypass_register_producer()
177 producer = xa_load(&producers, index); in irq_bypass_register_consumer()
/linux/drivers/infiniband/hw/hns/
H A Dhns_roce_bond.c52 struct hns_roce_die_info *die_info = xa_load(&roce_bond_xa, bus_num); in hns_roce_get_bond_grp()
281 struct hns_roce_die_info *die_info = xa_load(&roce_bond_xa, bus_num); in alloc_bond_id()
306 struct hns_roce_die_info *die_info = xa_load(&roce_bond_xa, bus_num); in remove_bond_id()
854 if (xa_load(&roce_bond_xa, bus_num)) in hns_roce_alloc_bond_grp()
956 die_info = xa_load(&roce_bond_xa, bus_num); in hns_roce_bond_suspend()
989 die_info = xa_load(&roce_bond_xa, bus_num); in hns_roce_bond_resume()
/linux/drivers/gpu/drm/tegra/
H A Duapi.c173 context = xa_load(&fpriv->contexts, args->context); in tegra_drm_ioctl_channel_close()
203 context = xa_load(&fpriv->contexts, args->context); in tegra_drm_ioctl_channel_map()
284 context = xa_load(&fpriv->contexts, args->context); in tegra_drm_ioctl_channel_unmap()
/linux/drivers/net/ethernet/mellanox/mlx5/core/steering/sws/
H A Ddr_domain.c81 recalc_cs_ft = xa_load(&dmn->csum_fts_xa, vport_num); in mlx5dr_domain_get_recalc_cs_ft_addr()
329 vport_caps = xa_load(&caps->vports.vports_caps_xa, vport); in mlx5dr_domain_get_vport_cap()
546 peer = xa_load(&dmn->peer_dmn_xa, peer_vhca_id); in mlx5dr_domain_set_peer()
552 peer = xa_load(&dmn->peer_dmn_xa, peer_vhca_id); in mlx5dr_domain_set_peer()
/linux/drivers/net/ethernet/mellanox/mlx5/core/
H A Dpagealloc.c106 root = xa_load(&dev->priv.page_root_xa, function); in page_root_per_function()
177 root = xa_load(&dev->priv.page_root_xa, function); in find_fw_page()
259 root = xa_load(&dev->priv.page_root_xa, fwp->function); in free_fwp()
443 root = xa_load(&dev->priv.page_root_xa, function); in release_all_pages()
504 root = xa_load(&dev->priv.page_root_xa, get_function(func_id, ec_function)); in reclaim_pages_cmd()
H A Deq.c815 irq = xa_load(&table->comp_irqs, vecidx); in comp_irq_release_pci()
863 irq = xa_load(&table->comp_irqs, vecidx); in comp_irq_release_sf()
1018 irq = xa_load(&table->comp_irqs, vecidx); in create_comp_eq()
1064 eq = xa_load(&table->comp_eqs, vecidx); in mlx5_comp_eqn_get()
1095 eq = xa_load(&table->comp_eqs, vector); in mlx5_comp_irqn_get()
1112 eq = xa_load(&table->comp_eqs, vector); in mlx5_comp_irq_get_affinity_mask()
/linux/drivers/infiniband/hw/cxgb4/
H A Dev.c127 qhp = xa_load(&dev->qps, CQE_QPID(err_cqe)); in c4iw_ev_dispatch()
228 chp = xa_load(&dev->cqs, qid); in c4iw_ev_handler()
/linux/drivers/gpu/drm/xe/
H A Dxe_tlb_inval_job.c212 xe_assert(gt_to_xe(job->q->gt), !xa_load(&job->dep.drm.dependencies, 0)); in xe_tlb_inval_job_alloc_dep()
249 xa_load(&job->dep.drm.dependencies, 0) == in xe_tlb_inval_job_push()
/linux/drivers/net/ethernet/mellanox/mlx5/core/en/
H A Dmapping.c118 mi = xa_load(&ctx->xarray, index); in mapping_remove()
141 mi = xa_load(&ctx->xarray, index); in mapping_find()
/linux/include/linux/
H A Dphy_link_topology.h55 pdn = xa_load(&topo->phys, phyindex); in phy_link_topo_get_phy()
/linux/drivers/cxl/core/
H A Dcdat.c154 dent = xa_load(dsmas_xa, dslbis->handle); in cdat_dslbis_handler()
669 perf_ctx = xa_load(usp_xa, index); in cxl_endpoint_gather_bandwidth()
795 us_ctx = xa_load(res_xa, us_index); in DEFINE_FREE()
879 hb_ctx = xa_load(hb_xa, hb_index); in cxl_rp_gather_bandwidth()
928 mw_ctx = xa_load(mw_xa, mw_index); in cxl_hb_gather_bandwidth()
/linux/block/
H A Dblk-mq-sched.c462 ctx = xa_load(elv_tbl, q->id); in blk_mq_free_sched_res_batch()
587 ctx = xa_load(elv_tbl, q->id); in blk_mq_alloc_sched_res_batch()
604 ctx = xa_load(elv_tbl, q->id); in blk_mq_alloc_sched_res_batch()
/linux/mm/
H A Dreadahead.c259 struct folio *folio = xa_load(&mapping->i_pages, index + i); in page_cache_ra_unbounded()
786 struct folio *folio = xa_load(&mapping->i_pages, index); in readahead_expand()
815 struct folio *folio = xa_load(&mapping->i_pages, index); in readahead_expand()
/linux/block/partitions/
H A Dcore.c320 if (xa_load(&disk->part_tbl, partno)) in add_partition()
464 part = xa_load(&disk->part_tbl, partno); in bdev_del_partition()
497 part = xa_load(&disk->part_tbl, partno); in bdev_resize_partition()
/linux/drivers/base/
H A Dauxiliary_sysfs.c102 struct auxiliary_irq_info *info __free(kfree) = xa_load(&auxdev->sysfs.irqs, irq); in auxiliary_device_sysfs_irq_remove()
/linux/drivers/infiniband/core/
H A Ddevice.c324 device = xa_load(&devices, index); in ib_device_get_by_index()
415 struct ib_client *client = xa_load(&clients, index); in ib_device_rename()
675 if (xa_load(&devices, device->index) == device) in ib_dealloc_device()
770 client_data = xa_load(&device->client_data, client_id); in remove_client_context()
772 client = xa_load(&clients, client_id); in remove_client_context()
952 cdev = xa_load(&device->compat_devs, rnet->id); in add_one_compat_dev()
1807 if (xa_load(&clients, highest_client_id - 1)) in remove_client_id()
1942 struct ib_client *client = xa_load(&clients, index); in __ib_get_client_nl_info()
2231 if (xa_load(&devices, ib_dev->index) != ib_dev) in ib_device_set_netdev()
2667 struct ib_client *client = xa_load(&clients, index); in ib_get_net_dev_by_params()

123456789