| /linux/drivers/infiniband/hw/mlx4/ |
| H A D | cm.c | 370 if (xa_err(item)) in alloc_rej_tmout() 371 ret = xa_err(item); in alloc_rej_tmout() 392 old, xa_err(old)); in alloc_rej_tmout() 394 return xa_err(old); in alloc_rej_tmout() 414 if (!item || xa_err(item)) { in lookup_rej_tmout_slave() 416 rem_pv_cm_id, xa_err(item)); in lookup_rej_tmout_slave() 417 slave = !item ? -ENOENT : xa_err(item); in lookup_rej_tmout_slave()
|
| /linux/rust/helpers/ |
| H A D | xarray.c | 7 return xa_err(entry); in rust_helper_xa_err()
|
| /linux/Documentation/translations/zh_CN/core-api/ |
| H A D | xarray.rst | 140 个特殊的值,可以用xa_err()把它变成一个错误值。如果你不需要确切地知道哪个错误发生,使用xa_is_err() 215 err = xa_err(__xa_store(&foo->array, index, entry, GFP_KERNEL));
|
| /linux/include/linux/ |
| H A D | xarray.h | 223 static inline int xa_err(void *entry) in xa_err() function 1088 return xa_err(xa_cmpxchg(xa, index, NULL, XA_ZERO_ENTRY, gfp)); in xa_reserve() 1106 return xa_err(xa_cmpxchg_bh(xa, index, NULL, XA_ZERO_ENTRY, gfp)); in xa_reserve_bh() 1124 return xa_err(xa_cmpxchg_irq(xa, index, NULL, XA_ZERO_ENTRY, gfp)); in xa_reserve_irq() 1437 return xa_err(xas->xa_node); in xas_error()
|
| /linux/lib/ |
| H A D | alloc_tag.c | 519 ret = ERR_PTR(xa_err(mas.node)); in reserve_module_tags() 528 ret = ERR_PTR(xa_err(mas.node)); in reserve_module_tags() 534 ret = ERR_PTR(xa_err(mas.node)); in reserve_module_tags()
|
| H A D | test_xarray.c | 88 XA_BUG_ON(xa, xa_err(xa_store_index(xa, 0, GFP_NOWAIT)) != 0); in check_xa_err() 89 XA_BUG_ON(xa, xa_err(xa_erase(xa, 0)) != 0); in check_xa_err() 92 XA_BUG_ON(xa, xa_err(xa_store_index(xa, 1, GFP_NOWAIT)) != -ENOMEM); in check_xa_err() 93 XA_BUG_ON(xa, xa_err(xa_store_index(xa, 1, GFP_NOWAIT)) != -ENOMEM); in check_xa_err() 95 XA_BUG_ON(xa, xa_err(xa_store_index(xa, 1, GFP_KERNEL)) != 0); in check_xa_err() 96 XA_BUG_ON(xa, xa_err(xa_store(xa, 1, xa_mk_value(0), GFP_KERNEL)) != 0); in check_xa_err() 97 XA_BUG_ON(xa, xa_err(xa_erase(xa, 1)) != 0); in check_xa_err()
|
| H A D | test_hmm.c | 238 return xa_err(entry); in dmirror_do_fault() 823 return xa_err(entry); in dmirror_atomic_map() 874 return xa_err(entry); in dmirror_migrate_finalize_and_map()
|
| H A D | maple_tree.c | 4051 return xa_err(mas->node); in mas_alloc_cyclic() 4765 return xa_err(mas->node); in mas_empty_area() 4826 return xa_err(mas->node); in mas_empty_area_rev() 5134 ret = xa_err(mas->node); in mas_store_gfp() 5196 int ret = xa_err(mas->node); in mas_preallocate() 5993 ret = xa_err(ms.node); in mtree_insert_range() 6045 ret = xa_err(mas.node); in mtree_alloc_range() 6127 ret = xa_err(mas.node); in mtree_alloc_rrange() 6384 ret = xa_err(mas.node); in __mt_dup() 6424 ret = xa_err(mas.node); in mtree_dup()
|
| /linux/drivers/cxl/core/ |
| H A D | cdat.c | 678 return xa_err(ptr); in cxl_endpoint_gather_bandwidth() 805 return ERR_PTR(xa_err(ptr)); in DEFINE_FREE() 888 return ERR_PTR(xa_err(ptr)); in cxl_rp_gather_bandwidth() 937 return ERR_PTR(xa_err(ptr)); in cxl_hb_gather_bandwidth()
|
| /linux/rust/kernel/ |
| H A D | xarray.rs | 251 let errno = unsafe { bindings::xa_err(old) }; in store()
|
| /linux/arch/arm64/kernel/ |
| H A D | hibernate.c | 233 return xa_err(ret); in save_tags()
|
| /linux/io_uring/ |
| H A D | tctx.c | 132 ret = xa_err(xa_store(&tctx->xa, (unsigned long)ctx, in __io_uring_add_tctx_node()
|
| /linux/drivers/infiniband/hw/mlx5/ |
| H A D | qpc.c | 236 err = xa_err(xa_store_irq(&dev->qp_table.dct_xa, qp->qpn, dct, GFP_KERNEL)); in mlx5_core_create_dct() 310 return xa_err(tmp) ?: -EINVAL; in mlx5_core_destroy_dct()
|
| H A D | srq_cmd.c | 653 err = xa_err(xa_store_irq(&table->array, srq->srqn, srq, GFP_KERNEL)); in mlx5_cmd_create_srq() 674 return xa_err(tmp) ?: -EINVAL; in mlx5_cmd_destroy_srq()
|
| /linux/drivers/xen/ |
| H A D | grant-dma-ops.c | 62 ret = xa_err(__xa_store(&xen_grant_dma_devices, (unsigned long)dev, data, in store_xen_grant_dma_data()
|
| /linux/mm/ |
| H A D | memremap.c | 180 error = xa_err(xa_store_range(&pgmap_array, PHYS_PFN(range->start), in pagemap_range()
|
| /linux/drivers/iommu/iommufd/ |
| H A D | viommu.c | 223 rc = xa_err(curr) ?: -EEXIST; in iommufd_vdevice_alloc_ioctl()
|
| H A D | device.c | 121 return ERR_PTR(xa_err(igroup)); in iommufd_get_group() 617 rc = xa_err(attach); in iommufd_hw_pagetable_attach()
|
| H A D | ioas.c | 428 rc = xa_err(xa_store(ioas_list, index, ioas, GFP_KERNEL)); in iommufd_take_all_iova_rwsem()
|
| /linux/kernel/liveupdate/ |
| H A D | kexec_handover.c | 142 return ERR_PTR(xa_err(res)); in xa_load_or_alloc() 201 err = xa_err(physxa); in __kho_preserve_order()
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/ |
| H A D | fs_counters.c | 273 err = xa_err(xa_store(&fc_stats->counters, id, counter, GFP_KERNEL)); in mlx5_fc_create()
|
| /linux/drivers/infiniband/hw/hns/ |
| H A D | hns_roce_srq.c | 125 ret = xa_err(xa_store_irq(&srq_table->xa, srq->srqn, srq, GFP_KERNEL)); in alloc_srqc()
|
| H A D | hns_roce_cq.c | 204 ret = xa_err(xa_store_irq(&cq_table->array, hr_cq->cqn, hr_cq, GFP_KERNEL)); in alloc_cqc()
|
| /linux/drivers/nvme/target/ |
| H A D | passthru.c | 612 ret = xa_err(old); in nvmet_passthru_ctrl_enable()
|
| /linux/net/qrtr/ |
| H A D | ns.c | 250 srv->service, srv->instance, xa_err(old)); in server_add()
|