| /linux/drivers/infiniband/hw/mlx4/ |
| H A D | cm.c | 370 if (xa_err(item)) in alloc_rej_tmout() 371 ret = xa_err(item); in alloc_rej_tmout() 392 old, xa_err(old)); in alloc_rej_tmout() 394 return xa_err(old); in alloc_rej_tmout() 414 if (!item || xa_err(item)) { in lookup_rej_tmout_slave() 416 rem_pv_cm_id, xa_err(item)); in lookup_rej_tmout_slave() 417 slave = !item ? -ENOENT : xa_err(item); in lookup_rej_tmout_slave()
|
| /linux/rust/helpers/ |
| H A D | xarray.c | 7 return xa_err(entry); in rust_helper_xa_err()
|
| /linux/drivers/iommu/amd/ |
| H A D | nested.c | 78 res = ERR_PTR(xa_err(res)); in gdom_info_load_or_alloc_locked() 279 if (WARN_ON(!curr || xa_err(curr))) in nested_domain_free()
|
| /linux/Documentation/translations/zh_CN/core-api/ |
| H A D | xarray.rst | 140 个特殊的值,可以用xa_err()把它变成一个错误值。如果你不需要确切地知道哪个错误发生,使用xa_is_err() 215 err = xa_err(__xa_store(&foo->array, index, entry, GFP_KERNEL));
|
| /linux/include/linux/ |
| H A D | xarray.h | 223 static inline int xa_err(void *entry) in xa_err() function 1088 return xa_err(xa_cmpxchg(xa, index, NULL, XA_ZERO_ENTRY, gfp)); in xa_reserve() 1106 return xa_err(xa_cmpxchg_bh(xa, index, NULL, XA_ZERO_ENTRY, gfp)); in xa_reserve_bh() 1124 return xa_err(xa_cmpxchg_irq(xa, index, NULL, XA_ZERO_ENTRY, gfp)); in xa_reserve_irq() 1437 return xa_err(xas->xa_node); in xas_error()
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/sf/dev/ |
| H A D | dev.c | 137 goto xa_err; in mlx5_sf_dev_add() 140 xa_err: in mlx5_sf_dev_add()
|
| /linux/drivers/cxl/core/ |
| H A D | cdat.c | 678 return xa_err(ptr); in cxl_endpoint_gather_bandwidth() 805 return ERR_PTR(xa_err(ptr)); in DEFINE_FREE() 888 return ERR_PTR(xa_err(ptr)); in cxl_rp_gather_bandwidth() 937 return ERR_PTR(xa_err(ptr)); in cxl_hb_gather_bandwidth()
|
| /linux/fs/smb/server/mgmt/ |
| H A D | tree_connect.c | 83 ret = xa_err(xa_store(&sess->tree_conns, tree_conn->id, tree_conn, in ksmbd_tree_conn_connect()
|
| /linux/rust/kernel/ |
| H A D | xarray.rs | 250 // SAFETY: `__xa_store` returns the old entry at this index on success or `xa_err` if an in store() 252 let errno = unsafe { bindings::xa_err(old) }; in store()
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/steering/sws/ |
| H A D | dr_domain.c | 88 ret = xa_err(xa_store(&dmn->csum_fts_xa, vport_num, in mlx5dr_domain_get_recalc_cs_ft_addr() 550 WARN_ON(xa_err(xa_store(&dmn->peer_dmn_xa, peer_vhca_id, peer_dmn, GFP_KERNEL))); in mlx5dr_domain_set_peer()
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/steering/hws/ |
| H A D | context.c | 253 if (xa_err(xa_store(&ctx->peer_ctx_xa, peer_vhca_id, peer_ctx, GFP_KERNEL))) in mlx5hws_context_set_peer()
|
| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_reg_sr.c | 101 ret = xa_err(xa_store(&sr->xa, idx, pentry, GFP_KERNEL)); in xe_reg_sr_add()
|
| /linux/lib/ |
| H A D | test_xarray.c | 88 XA_BUG_ON(xa, xa_err(xa_store_index(xa, 0, GFP_NOWAIT)) != 0); in check_xa_err() 89 XA_BUG_ON(xa, xa_err(xa_erase(xa, 0)) != 0); in check_xa_err() 92 XA_BUG_ON(xa, xa_err(xa_store_index(xa, 1, GFP_NOWAIT)) != -ENOMEM); in check_xa_err() 93 XA_BUG_ON(xa, xa_err(xa_store_index(xa, 1, GFP_NOWAIT)) != -ENOMEM); in check_xa_err() 95 XA_BUG_ON(xa, xa_err(xa_store_index(xa, 1, GFP_KERNEL)) != 0); in check_xa_err() 96 XA_BUG_ON(xa, xa_err(xa_store(xa, 1, xa_mk_value(0), GFP_KERNEL)) != 0); in check_xa_err() 97 XA_BUG_ON(xa, xa_err(xa_erase(xa, 1)) != 0); in check_xa_err()
|
| H A D | maple_tree.c | 3818 return xa_err(mas->node); in mas_alloc_cyclic() 4532 return xa_err(mas->node); in mas_empty_area() 4593 return xa_err(mas->node); in mas_empty_area_rev() 4901 ret = xa_err(mas->node); in mas_store_gfp() 4963 int ret = xa_err(mas->node); in mas_preallocate() 5760 ret = xa_err(ms.node); in mtree_insert_range() 5812 ret = xa_err(mas.node); in mtree_alloc_range() 5894 ret = xa_err(mas.node); in mtree_alloc_rrange() 6158 ret = xa_err(mas.node); in __mt_dup() 6198 ret = xa_err(mas.node); in mtree_dup()
|
| /linux/arch/x86/kernel/cpu/sgx/ |
| H A D | virt.c | 54 ret = xa_err(xa_store(&vepc->page_array, index, epc_page, GFP_KERNEL)); in __sgx_vepc_fault()
|
| /linux/arch/arm64/kernel/ |
| H A D | hibernate.c | 233 return xa_err(ret); in save_tags()
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/ |
| H A D | eq.c | 854 return xa_err(xa_store(&table->comp_irqs, vecidx, irq, GFP_KERNEL)); in comp_irq_request_pci() 905 return xa_err(xa_store(&table->comp_irqs, vecidx, irq, GFP_KERNEL)); in comp_irq_request_sf() 1035 err = xa_err(xa_store(&table->comp_eqs, vecidx, eq, GFP_KERNEL)); in create_comp_eq()
|
| H A D | fs_counters.c | 295 err = xa_err(xa_store(&fc_stats->counters, id, counter, GFP_KERNEL)); in mlx5_fc_create()
|
| /linux/drivers/block/ |
| H A D | brd.c | 114 return ERR_PTR(xa_err(ret)); in brd_insert_page()
|
| /linux/io_uring/ |
| H A D | tctx.c | 126 ret = xa_err(xa_store(&tctx->xa, (unsigned long)ctx, in io_tctx_install_node()
|
| /linux/drivers/iommu/iommufd/ |
| H A D | viommu.c | 223 rc = xa_err(curr) ?: -EEXIST; in iommufd_vdevice_alloc_ioctl()
|
| H A D | ioas.c | 428 rc = xa_err(xa_store(ioas_list, index, ioas, GFP_KERNEL)); in iommufd_take_all_iova_rwsem()
|
| /linux/net/devlink/ |
| H A D | region.c | 298 err = xa_err(__xa_store(&devlink->snapshot_ids, id, xa_mk_value(count), in __devlink_snapshot_id_increment() 372 err = xa_err(__xa_store(&devlink->snapshot_ids, id, xa_mk_value(0), in __devlink_snapshot_id_insert()
|
| /linux/net/shaper/ |
| H A D | shaper.c | 380 if (xa_err(prev)) { in net_shaper_pre_insert() 383 ret = xa_err(prev); in net_shaper_pre_insert()
|
| /linux/drivers/nvme/target/ |
| H A D | passthru.c | 613 ret = xa_err(old); in nvmet_passthru_ctrl_enable()
|