Home
last modified time | relevance | path

Searched refs:xa_store (Results 1 – 25 of 79) sorted by relevance

1234

/linux/Documentation/translations/zh_CN/core-api/
H A Dxarray.rst63 然后你可以用xa_store()来设置条目,用xa_load()来获取条目。xa_store将用新的条目覆盖任
64 何条目,并返回存储在该索引的上一个条目。你可以使用xa_erase()来代替调用xa_store()的
83 有时你需要确保对xa_store()的后续调用将不需要分配内存。xa_reserve()函数将在指定索引处存储
121 使用xa_store()、xa_cmpxchg()或xa_insert()也将标记该条目为正在分配。与普通的XArray不同,存储 ``NULL``
135 xa_store(), xa_cmpxchg(), xa_alloc(), xa_reserve()和xa_insert()函数接受一个gfp_t参数,以
163 * xa_store()
/linux/tools/testing/radix-tree/
H A Diteration_check_2.c43 xa_store(xa, i, xa_mk_value(i), GFP_KERNEL); in throbber()
64 xa_store(&array, 100, xa_mk_value(100), GFP_KERNEL); in iteration_test2()
/linux/drivers/base/
H A Dauxiliary_sysfs.c83 xa_store(&auxdev->sysfs.irqs, irq, no_free_ptr(info), GFP_KERNEL); in auxiliary_device_sysfs_irq_add()
/linux/drivers/net/ethernet/intel/ice/
H A Dice_adapter.c117 xa_store(&ice_adapters, index, adapter, GFP_KERNEL); in ice_adapter_get()
/linux/drivers/cxl/core/
H A Dcdat.c676 ptr = xa_store(usp_xa, index, c, GFP_KERNEL); in cxl_endpoint_gather_bandwidth()
803 ptr = xa_store(res_xa, us_index, n, GFP_KERNEL); in DEFINE_FREE()
886 ptr = xa_store(hb_xa, hb_index, n, GFP_KERNEL); in cxl_rp_gather_bandwidth()
935 ptr = xa_store(mw_xa, mw_index, n, GFP_KERNEL); in cxl_hb_gather_bandwidth()
/linux/lib/
H A Dtest_xarray.c42 return xa_store(xa, index, xa_mk_index(index), gfp); in xa_store_index()
96 XA_BUG_ON(xa, xa_err(xa_store(xa, 1, xa_mk_value(0), GFP_KERNEL)) != 0); in check_xa_err()
667 xa_store(xa, 3, xa, GFP_KERNEL); in check_multi_store()
1013 XA_BUG_ON(xa, xa_store(xa, 3, xa_mk_index(3), GFP_KERNEL) != NULL); in check_xa_alloc_2()
1014 XA_BUG_ON(xa, xa_store(xa, 4, xa_mk_index(4), GFP_KERNEL) != NULL); in check_xa_alloc_2()
1015 XA_BUG_ON(xa, xa_store(xa, 4, NULL, GFP_KERNEL) != xa_mk_index(4)); in check_xa_alloc_2()
1972 XA_BUG_ON(xa, xa_store(xa, 0, name + i, GFP_KERNEL) != NULL); in check_align_2()
1978 XA_BUG_ON(xa, xa_store(xa, 0, name + i, 0) != NULL); in check_align_2()
2215 xa_store(xa, ULONG_MAX, xa, GFP_KERNEL); in check_destroy()
H A Dtest_hmm.c236 entry = xa_store(&dmirror->pt, pfn, entry, GFP_ATOMIC); in dmirror_do_fault()
822 entry = xa_store(&dmirror->pt, addr >> PAGE_SHIFT, entry, GFP_ATOMIC); in dmirror_atomic_map()
873 entry = xa_store(&dmirror->pt, pfn + i, entry, GFP_ATOMIC); in dmirror_migrate_finalize_and_map()
/linux/drivers/gpu/drm/xe/
H A Dxe_reg_sr.c99 ret = xa_err(xa_store(&sr->xa, idx, pentry, GFP_KERNEL)); in xe_reg_sr_add()
H A Dxe_tlb_inval_job.c253 ptr = xa_store(&job->dep.drm.dependencies, 0, fence, in xe_tlb_inval_job_push()
/linux/fs/smb/server/mgmt/
H A Dtree_connect.c83 ret = xa_err(xa_store(&sess->tree_conns, tree_conn->id, tree_conn, in ksmbd_tree_conn_connect()
H A Duser_session.c335 old = xa_store(&sess->rpc_handle_list, id, entry, KSMBD_DEFAULT_GFP); in ksmbd_session_rpc_open()
440 return xa_err(xa_store(&conn->sessions, sess->id, sess, KSMBD_DEFAULT_GFP)); in ksmbd_session_register()
/linux/drivers/net/ethernet/mellanox/mlx5/core/steering/sws/
H A Ddr_domain.c88 ret = xa_err(xa_store(&dmn->csum_fts_xa, vport_num, in mlx5dr_domain_get_recalc_cs_ft_addr()
550 WARN_ON(xa_err(xa_store(&dmn->peer_dmn_xa, peer_vhca_id, peer_dmn, GFP_KERNEL))); in mlx5dr_domain_set_peer()
/linux/drivers/net/ethernet/mellanox/mlx5/core/steering/hws/
H A Dcontext.c253 if (xa_err(xa_store(&ctx->peer_ctx_xa, peer_vhca_id, peer_ctx, GFP_KERNEL))) in mlx5hws_context_set_peer()
/linux/arch/arm64/kernel/
H A Dhibernate.c230 ret = xa_store(&mte_pages, pfn, tag_storage, GFP_KERNEL); in save_tags()
/linux/arch/x86/kernel/cpu/sgx/
H A Dvirt.c54 ret = xa_err(xa_store(&vepc->page_array, index, epc_page, GFP_KERNEL)); in __sgx_vepc_fault()
/linux/drivers/gpu/drm/
H A Ddrm_drv.c199 entry = xa_store(drm_minor_get_xa(type), minor->index, minor, GFP_KERNEL); in drm_minor_register()
223 xa_store(drm_minor_get_xa(type), minor->index, NULL, GFP_KERNEL); in drm_minor_unregister()
/linux/net/qrtr/
H A Dns.c88 if (xa_store(&nodes, node_id, node, GFP_KERNEL)) { in node_get()
246 old = xa_store(&node->servers, port, srv, GFP_KERNEL); in server_add()
/linux/drivers/net/ethernet/mellanox/mlx5/core/
H A Deq.c854 return xa_err(xa_store(&table->comp_irqs, vecidx, irq, GFP_KERNEL)); in comp_irq_request_pci()
905 return xa_err(xa_store(&table->comp_irqs, vecidx, irq, GFP_KERNEL)); in comp_irq_request_sf()
1035 err = xa_err(xa_store(&table->comp_eqs, vecidx, eq, GFP_KERNEL)); in create_comp_eq()
H A Dfs_counters.c295 err = xa_err(xa_store(&fc_stats->counters, id, counter, GFP_KERNEL)); in mlx5_fc_create()
/linux/net/psp/
H A Dpsp_main.c130 xa_store(&psp_devs, psd->id, NULL, GFP_KERNEL); in psp_dev_unregister()
/linux/io_uring/
H A Dtctx.c148 ret = xa_err(xa_store(&tctx->xa, (unsigned long)ctx, in __io_uring_add_tctx_node()
/linux/drivers/tee/
H A Dtee_heap.c282 return WARN_ON(xa_is_err(xa_store(&tee_dma_heap, id, h, GFP_KERNEL))); in alloc_dma_heap()
/linux/drivers/infiniband/core/
H A Ddevice.c724 ret = xa_err(xa_store(&device->client_data, client->client_id, NULL, in add_client_context()
981 ret = xa_err(xa_store(&device->compat_devs, rnet->id, in add_one_compat_dev()
1135 ret = xa_err(xa_store(&rdma_nets, rnet->id, NULL, GFP_KERNEL)); in rdma_dev_exit_net()
2022 rc = xa_store(&device->client_data, client->client_id, data, in ib_set_client_data()
/linux/drivers/iommu/
H A Diommu.c3617 WARN_ON(xa_is_err(xa_store(&group->pasid_array, in iommu_attach_device_pasid()
3718 WARN_ON(xa_is_err(xa_store(&group->pasid_array, in iommu_replace_device_pasid()
3855 WARN_ON(xa_is_err(xa_store(&group->pasid_array, in iommu_attach_group_handle()
3918 curr = xa_store(&group->pasid_array, IOMMU_NO_PASID, entry, GFP_KERNEL); in iommu_replace_group_handle()
/linux/drivers/iommu/iommufd/
H A Dio_pagetable.c1194 xa_store(&iopt->domains, iopt->next_domain_id, domain, GFP_KERNEL); in iopt_table_add_domain()
1267 xa_store(&iopt->domains, index, iter_domain, GFP_KERNEL); in iopt_table_remove_domain()

1234