| /linux/include/linux/ |
| H A D | xarray.h | 178 #define XA_ZERO_ENTRY xa_mk_internal(257) macro 191 return unlikely(entry == XA_ZERO_ENTRY); in xa_is_zero() 1088 return xa_err(xa_cmpxchg(xa, index, NULL, XA_ZERO_ENTRY, gfp)); in xa_reserve() 1106 return xa_err(xa_cmpxchg_bh(xa, index, NULL, XA_ZERO_ENTRY, gfp)); in xa_reserve_bh() 1124 return xa_err(xa_cmpxchg_irq(xa, index, NULL, XA_ZERO_ENTRY, gfp)); in xa_reserve_irq() 1138 xa_cmpxchg(xa, index, XA_ZERO_ENTRY, NULL, 0); in xa_release()
|
| /linux/drivers/iommu/iommufd/ |
| H A D | main.c | 56 rc = xa_alloc(&ictx->objects, &obj->id, XA_ZERO_ENTRY, xa_limit_31b, in _iommufd_object_alloc() 112 WARN_ON(old != XA_ZERO_ENTRY); in iommufd_object_finalize() 124 WARN_ON(old != XA_ZERO_ENTRY); in iommufd_object_abort() 258 xas_store(&xas, (flags & REMOVE_OBJ_TOMBSTONE) ? XA_ZERO_ENTRY : NULL); in iommufd_object_remove()
|
| H A D | device.c | 615 XA_ZERO_ENTRY, GFP_KERNEL); in iommufd_hw_pagetable_attach() 632 rc = xa_insert(&attach->device_array, idev->obj.id, XA_ZERO_ENTRY, in iommufd_hw_pagetable_attach()
|
| /linux/drivers/infiniband/hw/mlx5/ |
| H A D | qpc.c | 308 tmp = xa_cmpxchg_irq(&table->dct_xa, dct->mqp.qpn, dct, XA_ZERO_ENTRY, GFP_KERNEL); in mlx5_core_destroy_dct() 314 xa_cmpxchg_irq(&table->dct_xa, dct->mqp.qpn, XA_ZERO_ENTRY, dct, 0); in mlx5_core_destroy_dct()
|
| H A D | srq_cmd.c | 672 tmp = xa_cmpxchg_irq(&table->array, srq->srqn, srq, XA_ZERO_ENTRY, 0); in mlx5_cmd_destroy_srq() 683 xa_cmpxchg_irq(&table->array, srq->srqn, XA_ZERO_ENTRY, srq, 0); in mlx5_cmd_destroy_srq()
|
| /linux/lib/ |
| H A D | xarray.c | 660 entry = XA_ZERO_ENTRY; in xas_create() 1700 entry = XA_ZERO_ENTRY; in __xa_store() 1813 entry = XA_ZERO_ENTRY; in __xa_insert() 1996 entry = XA_ZERO_ENTRY; in __xa_alloc()
|
| H A D | test_maple_tree.c | 418 XA_ZERO_ENTRY, GFP_KERNEL)); in check_find() 430 MT_BUG_ON(mt, entry != XA_ZERO_ENTRY); in check_find() 446 MT_BUG_ON(mt, entry != XA_ZERO_ENTRY); in check_find() 462 MT_BUG_ON(mt, entry != XA_ZERO_ENTRY); in check_find() 532 MT_BUG_ON(mt, entry != XA_ZERO_ENTRY); in check_find() 758 mtree_store_range(mt, MTREE_ALLOC_MAX, ULONG_MAX, XA_ZERO_ENTRY, in check_alloc_rev_range() 933 mtree_store_range(mt, MTREE_ALLOC_MAX, ULONG_MAX, XA_ZERO_ENTRY, in check_alloc_range()
|
| H A D | test_xarray.c | 502 XA_BUG_ON(xa, xa_cmpxchg(xa, 12345678, XA_ZERO_ENTRY, in check_reserve()
|
| /linux/drivers/infiniband/core/ |
| H A D | ucma.c | 542 uevent->conn_req_ctx, XA_ZERO_ENTRY, in ucma_cleanup_ctx_events() 590 WARN_ON(xa_cmpxchg(&ctx_table, ctx->id, XA_ZERO_ENTRY, NULL, in ucma_destroy_private_ctx() 614 if (__xa_cmpxchg(&ctx_table, ctx->id, ctx, XA_ZERO_ENTRY, in ucma_destroy_id() 1919 WARN_ON(xa_cmpxchg(&ctx_table, ctx->id, ctx, XA_ZERO_ENTRY, in ucma_close()
|
| /linux/drivers/iommu/generic_pt/ |
| H A D | kunit_iommu_pt.h | 381 XA_ZERO_ENTRY, in test_random_map()
|
| /linux/drivers/infiniband/sw/rxe/ |
| H A D | rxe_mr.c | 170 xas_store(&xas, XA_ZERO_ENTRY); in rxe_mr_alloc()
|
| /linux/drivers/iommu/ |
| H A D | iommu.c | 3514 ret = xa_insert(&group->pasid_array, pasid, XA_ZERO_ENTRY, GFP_KERNEL); in iommu_attach_device_pasid() 3580 XA_ZERO_ENTRY, GFP_KERNEL); in iommu_replace_device_pasid() 3742 IOMMU_NO_PASID, XA_ZERO_ENTRY, GFP_KERNEL); in iommu_attach_group_handle()
|
| /linux/mm/ |
| H A D | mmap.c | 1853 mas_store(&vmi.mas, XA_ZERO_ENTRY); in dup_mmap()
|
| /linux/net/ipv4/ |
| H A D | tcp.c | 2427 __xa_cmpxchg(&sk->sk_user_frags, p->tokens[i], XA_ZERO_ENTRY, in tcp_xa_pool_commit_locked() 2463 XA_ZERO_ENTRY, xa_limit_31b, GFP_KERNEL); in tcp_xa_pool_refill()
|