| /linux/Documentation/translations/zh_CN/core-api/ |
| H A D | xarray.rst | 63 然后你可以用xa_store()来设置条目,用xa_load()来获取条目。xa_store将用新的条目覆盖任 64 何条目,并返回存储在该索引的上一个条目。你可以使用xa_erase()来代替调用xa_store()的 83 有时你需要确保对xa_store()的后续调用将不需要分配内存。xa_reserve()函数将在指定索引处存储 121 使用xa_store()、xa_cmpxchg()或xa_insert()也将标记该条目为正在分配。与普通的XArray不同,存储 ``NULL`` 135 xa_store(), xa_cmpxchg(), xa_alloc(), xa_reserve()和xa_insert()函数接受一个gfp_t参数,以 163 * xa_store()
|
| /linux/tools/testing/radix-tree/ |
| H A D | iteration_check_2.c | 43 xa_store(xa, i, xa_mk_value(i), GFP_KERNEL); in throbber() 64 xa_store(&array, 100, xa_mk_value(100), GFP_KERNEL); in iteration_test2()
|
| /linux/drivers/net/ethernet/intel/ice/ |
| H A D | ice_adapter.c | 117 xa_store(&ice_adapters, index, adapter, GFP_KERNEL); in ice_adapter_get()
|
| /linux/drivers/base/ |
| H A D | auxiliary_sysfs.c | 83 xa_store(&auxdev->sysfs.irqs, irq, no_free_ptr(info), GFP_KERNEL); in auxiliary_device_sysfs_irq_add()
|
| H A D | arch_topology.c | 901 xa_store(&hetero_cpu, hetero_id, in parse_acpi_topology()
|
| /linux/drivers/cxl/core/ |
| H A D | cdat.c | 676 ptr = xa_store(usp_xa, index, c, GFP_KERNEL); in cxl_endpoint_gather_bandwidth() 803 ptr = xa_store(res_xa, us_index, n, GFP_KERNEL); in DEFINE_FREE() 886 ptr = xa_store(hb_xa, hb_index, n, GFP_KERNEL); in cxl_rp_gather_bandwidth() 935 ptr = xa_store(mw_xa, mw_index, n, GFP_KERNEL); in cxl_hb_gather_bandwidth()
|
| /linux/drivers/gpu/drm/etnaviv/ |
| H A D | etnaviv_sched.c | 129 xa_store(&gpu->user_fences, submit->out_fence_id, in etnaviv_sched_push_job()
|
| /linux/fs/smb/server/mgmt/ |
| H A D | tree_connect.c | 83 ret = xa_err(xa_store(&sess->tree_conns, tree_conn->id, tree_conn, in ksmbd_tree_conn_connect()
|
| /linux/lib/ |
| H A D | test_xarray.c | 42 return xa_store(xa, index, xa_mk_index(index), gfp); in xa_store_index() 96 XA_BUG_ON(xa, xa_err(xa_store(xa, 1, xa_mk_value(0), GFP_KERNEL)) != 0); in check_xa_err() 667 xa_store(xa, 3, xa, GFP_KERNEL); in check_multi_store() 1013 XA_BUG_ON(xa, xa_store(xa, 3, xa_mk_index(3), GFP_KERNEL) != NULL); in check_xa_alloc_2() 1014 XA_BUG_ON(xa, xa_store(xa, 4, xa_mk_index(4), GFP_KERNEL) != NULL); in check_xa_alloc_2() 1015 XA_BUG_ON(xa, xa_store(xa, 4, NULL, GFP_KERNEL) != xa_mk_index(4)); in check_xa_alloc_2() 1972 XA_BUG_ON(xa, xa_store(xa, 0, name + i, GFP_KERNEL) != NULL); in check_align_2() 1978 XA_BUG_ON(xa, xa_store(xa, 0, name + i, 0) != NULL); in check_align_2() 2215 xa_store(xa, ULONG_MAX, xa, GFP_KERNEL); in check_destroy()
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/steering/sws/ |
| H A D | dr_domain.c | 88 ret = xa_err(xa_store(&dmn->csum_fts_xa, vport_num, in mlx5dr_domain_get_recalc_cs_ft_addr() 550 WARN_ON(xa_err(xa_store(&dmn->peer_dmn_xa, peer_vhca_id, peer_dmn, GFP_KERNEL))); in mlx5dr_domain_set_peer()
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/steering/hws/ |
| H A D | context.c | 253 if (xa_err(xa_store(&ctx->peer_ctx_xa, peer_vhca_id, peer_ctx, GFP_KERNEL))) in mlx5hws_context_set_peer()
|
| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_reg_sr.c | 101 ret = xa_err(xa_store(&sr->xa, idx, pentry, GFP_KERNEL)); in xe_reg_sr_add()
|
| H A D | xe_tlb_inval_job.c | 253 ptr = xa_store(&job->dep.drm.dependencies, 0, fence, in xe_tlb_inval_job_push()
|
| /linux/arch/x86/kernel/cpu/sgx/ |
| H A D | virt.c | 54 ret = xa_err(xa_store(&vepc->page_array, index, epc_page, GFP_KERNEL)); in __sgx_vepc_fault()
|
| /linux/arch/arm64/kernel/ |
| H A D | hibernate.c | 230 ret = xa_store(&mte_pages, pfn, tag_storage, GFP_KERNEL); in save_tags()
|
| /linux/net/qrtr/ |
| H A D | ns.c | 106 if (xa_store(&nodes, node_id, node, GFP_KERNEL)) { in server_match() 273 old = xa_store(&node->servers, port, srv, GFP_KERNEL); in server_del()
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/ |
| H A D | eq.c | 854 return xa_err(xa_store(&table->comp_irqs, vecidx, irq, GFP_KERNEL)); in comp_irq_request_pci() 905 return xa_err(xa_store(&table->comp_irqs, vecidx, irq, GFP_KERNEL)); in comp_irq_request_sf() 1035 err = xa_err(xa_store(&table->comp_eqs, vecidx, eq, GFP_KERNEL)); in create_comp_eq()
|
| H A D | fs_counters.c | 295 err = xa_err(xa_store(&fc_stats->counters, id, counter, GFP_KERNEL)); in mlx5_fc_create()
|
| /linux/io_uring/ |
| H A D | tctx.c | 126 ret = xa_err(xa_store(&tctx->xa, (unsigned long)ctx, in io_tctx_install_node()
|
| /linux/net/psp/ |
| H A D | psp_main.c | 130 xa_store(&psp_devs, psd->id, NULL, GFP_KERNEL); in psp_dev_unregister()
|
| /linux/drivers/tee/ |
| H A D | tee_heap.c | 282 return WARN_ON(xa_is_err(xa_store(&tee_dma_heap, id, h, GFP_KERNEL))); in alloc_dma_heap()
|
| /linux/drivers/infiniband/core/ |
| H A D | device.c | 725 ret = xa_err(xa_store(&device->client_data, client->client_id, NULL, in add_client_context() 982 ret = xa_err(xa_store(&device->compat_devs, rnet->id, in add_one_compat_dev() 1136 ret = xa_err(xa_store(&rdma_nets, rnet->id, NULL, GFP_KERNEL)); in rdma_dev_exit_net() 2023 rc = xa_store(&device->client_data, client->client_id, data, in ib_set_client_data()
|
| /linux/drivers/iommu/ |
| H A D | iommu.c | 3661 WARN_ON(xa_is_err(xa_store(&group->pasid_array, in iommu_attach_device_pasid() 3762 WARN_ON(xa_is_err(xa_store(&group->pasid_array, in iommu_replace_device_pasid() 3899 WARN_ON(xa_is_err(xa_store(&group->pasid_array, in iommu_attach_group_handle() 3962 curr = xa_store(&group->pasid_array, IOMMU_NO_PASID, entry, GFP_KERNEL); in iommu_replace_group_handle()
|
| /linux/drivers/iommu/iommufd/ |
| H A D | ioas.c | 428 rc = xa_err(xa_store(ioas_list, index, ioas, GFP_KERNEL)); in iommufd_take_all_iova_rwsem()
|
| /linux/drivers/gpu/drm/vmwgfx/ |
| H A D | vmwgfx_bo.c | 849 return xa_err(xa_store(&vbo->detached_resources, (unsigned long)res, res, GFP_KERNEL)); in vmw_bo_add_detached_resource()
|