| /linux/kernel/dma/ |
| H A D | debug.h | 24 int nelems, int dir); 43 int nelems, int direction); 47 int nelems, int direction); 75 int nelems, int dir) in debug_dma_unmap_sg() argument 104 int nelems, int direction) in debug_dma_sync_sg_for_cpu() argument 110 int nelems, int direction) in debug_dma_sync_sg_for_device() argument
|
| H A D | mapping.c | 410 int nelems, enum dma_data_direction dir) in __dma_sync_sg_for_cpu() argument 416 dma_direct_sync_sg_for_cpu(dev, sg, nelems, dir); in __dma_sync_sg_for_cpu() 418 iommu_dma_sync_sg_for_cpu(dev, sg, nelems, dir); in __dma_sync_sg_for_cpu() 420 ops->sync_sg_for_cpu(dev, sg, nelems, dir); in __dma_sync_sg_for_cpu() 421 trace_dma_sync_sg_for_cpu(dev, sg, nelems, dir); in __dma_sync_sg_for_cpu() 422 debug_dma_sync_sg_for_cpu(dev, sg, nelems, dir); in __dma_sync_sg_for_cpu() 427 int nelems, enum dma_data_direction dir) in __dma_sync_sg_for_device() argument 433 dma_direct_sync_sg_for_device(dev, sg, nelems, dir); in __dma_sync_sg_for_device() 435 iommu_dma_sync_sg_for_device(dev, sg, nelems, dir); in __dma_sync_sg_for_device() 437 ops->sync_sg_for_device(dev, sg, nelems, dir); in __dma_sync_sg_for_device() [all …]
|
| H A D | dummy.c | 30 int nelems, enum dma_data_direction dir, in dma_dummy_map_sg() argument 37 int nelems, enum dma_data_direction dir, in dma_dummy_unmap_sg() argument
|
| H A D | debug.c | 1375 int nelems, int dir) in debug_dma_unmap_sg() argument 1383 for_each_sg(sglist, s, nelems, i) { in debug_dma_unmap_sg() 1392 .sg_call_ents = nelems, in debug_dma_unmap_sg() 1508 int nelems, int direction) in debug_dma_sync_sg_for_cpu() argument 1516 for_each_sg(sg, s, nelems, i) { in debug_dma_sync_sg_for_cpu() 1525 .sg_call_ents = nelems, in debug_dma_sync_sg_for_cpu() 1539 int nelems, int direction) in debug_dma_sync_sg_for_device() argument 1547 for_each_sg(sg, s, nelems, i) { in debug_dma_sync_sg_for_device() 1556 .sg_call_ents = nelems, in debug_dma_sync_sg_for_device()
|
| /linux/drivers/xen/ |
| H A D | swiotlb-xen.c | 346 xen_swiotlb_unmap_sg(struct device *hwdev, struct scatterlist *sgl, int nelems, in xen_swiotlb_unmap_sg() argument 354 for_each_sg(sgl, sg, nelems, i) in xen_swiotlb_unmap_sg() 361 xen_swiotlb_map_sg(struct device *dev, struct scatterlist *sgl, int nelems, in xen_swiotlb_map_sg() argument 369 for_each_sg(sgl, sg, nelems, i) { in xen_swiotlb_map_sg() 377 return nelems; in xen_swiotlb_map_sg() 386 int nelems, enum dma_data_direction dir) in xen_swiotlb_sync_sg_for_cpu() argument 391 for_each_sg(sgl, sg, nelems, i) { in xen_swiotlb_sync_sg_for_cpu() 399 int nelems, enum dma_data_direction dir) in xen_swiotlb_sync_sg_for_device() argument 404 for_each_sg(sgl, sg, nelems, i) { in xen_swiotlb_sync_sg_for_device()
|
| /linux/arch/powerpc/kernel/ |
| H A D | dma-iommu.c | 132 int nelems, enum dma_data_direction direction, in dma_iommu_map_sg() argument 135 return ppc_iommu_map_sg(dev, get_iommu_table_base(dev), sglist, nelems, in dma_iommu_map_sg() 140 int nelems, enum dma_data_direction direction, in dma_iommu_unmap_sg() argument 143 ppc_iommu_unmap_sg(get_iommu_table_base(dev), sglist, nelems, in dma_iommu_unmap_sg()
|
| H A D | iommu.c | 473 struct scatterlist *sglist, int nelems, in ppc_iommu_map_sg() argument 486 if ((nelems == 0) || !tbl) in ppc_iommu_map_sg() 491 incount = nelems; in ppc_iommu_map_sg() 497 DBG("sg mapping %d elements:\n", nelems); in ppc_iommu_map_sg() 500 for_each_sg(sglist, s, nelems, i) { in ppc_iommu_map_sg() 598 for_each_sg(sglist, s, nelems, i) { in ppc_iommu_map_sg() 616 int nelems, enum dma_data_direction direction, in ppc_iommu_unmap_sg() argument 627 while (nelems--) { in ppc_iommu_unmap_sg()
|
| /linux/arch/sparc/kernel/ |
| H A D | iommu.c | 443 int nelems, enum dma_data_direction direction, in dma_4u_map_sg() argument 460 if (nelems == 0 || !iommu) in dma_4u_map_sg() 478 incount = nelems; in dma_4u_map_sg() 487 for_each_sg(sglist, s, nelems, i) { in dma_4u_map_sg() 565 for_each_sg(sglist, s, nelems, i) { in dma_4u_map_sg() 616 int nelems, enum dma_data_direction direction, in dma_4u_unmap_sg() argument 634 while (nelems--) { in dma_4u_unmap_sg() 706 struct scatterlist *sglist, int nelems, in dma_4u_sync_sg_for_cpu() argument 738 for_each_sg(sglist, sg, nelems, i) { in dma_4u_sync_sg_for_cpu()
|
| H A D | pci_sun4v.c | 479 int nelems, enum dma_data_direction direction, in dma_4v_map_sg() argument 498 if (nelems == 0 || !iommu) in dma_4v_map_sg() 511 incount = nelems; in dma_4v_map_sg() 532 for_each_sg(sglist, s, nelems, i) { in dma_4v_map_sg() 613 for_each_sg(sglist, s, nelems, i) { in dma_4v_map_sg() 634 int nelems, enum dma_data_direction direction, in dma_4v_unmap_sg() argument 655 while (nelems--) { in dma_4v_unmap_sg()
|
| /linux/include/linux/ |
| H A D | dma-mapping.h | 414 int nelems, enum dma_data_direction dir); 416 int nelems, enum dma_data_direction dir); 440 struct scatterlist *sg, int nelems, enum dma_data_direction dir) in dma_sync_sg_for_cpu() argument 443 __dma_sync_sg_for_cpu(dev, sg, nelems, dir); in dma_sync_sg_for_cpu() 447 struct scatterlist *sg, int nelems, enum dma_data_direction dir) in dma_sync_sg_for_device() argument 450 __dma_sync_sg_for_device(dev, sg, nelems, dir); in dma_sync_sg_for_device() 472 struct scatterlist *sg, int nelems, enum dma_data_direction dir) in dma_sync_sg_for_cpu() argument 476 struct scatterlist *sg, int nelems, enum dma_data_direction dir) in dma_sync_sg_for_device() argument
|
| /linux/net/netfilter/ |
| H A D | nft_set_rbtree.c | 580 u32 nelems = atomic_read(&set->nelems) - set->ndeact; in nft_array_elems() local 586 return nelems; in nft_array_elems() 593 return (nelems / 2) + 2; in nft_array_elems() 603 u32 shrinked_max_intervals, nelems = nft_array_elems(set); in nft_array_may_resize() local 630 nelems = 0; in nft_array_may_resize() 642 nelems < shrinked_max_intervals) { in nft_array_may_resize() 647 if (nelems > new_max_intervals) { in nft_array_may_resize() 660 if (WARN_ON_ONCE(nelems > new_max_intervals)) in nft_array_may_resize()
|
| H A D | nft_dynset.c | 64 if (!atomic_add_unless(&set->nelems, 1, set->size)) in nft_dynset_new() 85 atomic_dec(&set->nelems); in nft_dynset_new()
|
| H A D | nf_tables_api.c | 458 return a->set == b->set && a->bound == b->bound && a->nelems < NFT_MAX_SET_NELEMS; in nft_trans_collapse_set_elem_allowed() 465 unsigned int nelems, old_nelems = tail->nelems; in nft_trans_collapse_set_elem() local 477 if (WARN_ON_ONCE(trans->nelems != 1)) in nft_trans_collapse_set_elem() 480 if (check_add_overflow(old_nelems, trans->nelems, &nelems)) in nft_trans_collapse_set_elem() 486 new_trans = krealloc(tail, struct_size(tail, elems, nelems), in nft_trans_collapse_set_elem() 498 new_trans->nelems = nelems; in nft_trans_collapse_set_elem() 4868 unsigned int nelems; in nf_tables_fill_set_info() local 4882 nelems = nft_set_userspace_size(set->ops, atomic_read(&set->nelems)); in nf_tables_fill_set_info() 4883 return nla_put_be32(skb, NFTA_SET_COUNT, htonl(nelems)); in nf_tables_fill_set_info() 5719 atomic_read(&set->nelems) > 0)) { in nf_tables_delset() [all …]
|
| /linux/arch/powerpc/kvm/ |
| H A D | guest-state-buffer.c | 45 gsb->hdr->nelems = cpu_to_be32(0); in kvmppc_gsb_new() 78 u32 nelems = kvmppc_gsb_nelems(gsb); in kvmppc_gsb_put() local 84 kvmppc_gsb_header(gsb)->nelems = cpu_to_be32(nelems + 1); in kvmppc_gsb_put()
|
| /linux/arch/x86/kernel/ |
| H A D | amd_gart_64.c | 324 int nelems, struct scatterlist *sout, in __dma_map_cont() argument 335 for_each_sg(start, s, nelems, i) { in __dma_map_cont() 362 dma_map_cont(struct device *dev, struct scatterlist *start, int nelems, in dma_map_cont() argument 366 BUG_ON(nelems != 1); in dma_map_cont() 371 return __dma_map_cont(dev, start, nelems, sout, pages); in dma_map_cont()
|
| /linux/lib/ |
| H A D | rhashtable.c | 399 unsigned int nelems = atomic_read(&ht->nelems); in rhashtable_shrink() 402 if (nelems) in rhashtable_shrink() 403 size = roundup_pow_of_two(nelems * 3 / 2); in rhashtable_shrink() 625 atomic_inc(&ht->nelems); in rhashtable_try_insert() 1048 /* Cap total entries at 2^31 to avoid nelems overflow. */ in rhashtable_init_noprof() 1082 atomic_set(&ht->nelems, 0); in rhashtable_init_noprof() 397 unsigned int nelems = atomic_read(&ht->nelems); rhashtable_shrink() local
|
| /linux/arch/powerpc/include/asm/ |
| H A D | iommu.h | 262 struct scatterlist *sglist, int nelems, 268 int nelems,
|
| /linux/include/uapi/linux/ |
| H A D | btf.h | 115 __u32 nelems; 114 __u32 nelems; global() member
|
| /linux/tools/include/uapi/linux/ |
| H A D | btf.h | 115 __u32 nelems; 114 __u32 nelems; global() member
|
| /linux/arch/powerpc/platforms/pseries/ |
| H A D | vio.c | 555 int nelems, enum dma_data_direction direction, in vio_dma_iommu_map_sg() argument 564 for_each_sg(sglist, sgl, nelems, count) in vio_dma_iommu_map_sg() 570 ret = ppc_iommu_map_sg(dev, tbl, sglist, nelems, dma_get_mask(dev), in vio_dma_iommu_map_sg() 589 struct scatterlist *sglist, int nelems, in vio_dma_iommu_unmap_sg() argument 599 for_each_sg(sglist, sgl, nelems, count) in vio_dma_iommu_unmap_sg() 602 ppc_iommu_unmap_sg(tbl, sglist, nelems, direction, attrs); in vio_dma_iommu_unmap_sg()
|
| /linux/kernel/bpf/ |
| H A D | btf.c | 2078 u32 i, size, nelems = 1, id = 0; in __btf_resolve_size() local 2110 if (nelems && array->nelems > U32_MAX / nelems) in __btf_resolve_size() 2112 nelems *= array->nelems; in __btf_resolve_size() 2125 if (nelems && size > U32_MAX / nelems) in __btf_resolve_size() 2128 *type_size = nelems * size; in __btf_resolve_size() 2130 *total_nelems = nelems; in __btf_resolve_size() 3140 if (array->nelems && elem_size > U32_MAX / array->nelems) { in btf_array_resolve() 3146 env_stack_pop_resolved(env, elem_type_id, elem_size * array->nelems); in btf_array_resolve() 3157 array->type, array->index_type, array->nelems); in btf_array_log() 3196 for (i = 0; i < array->nelems; i++) { in __btf_array_show() [all …]
|
| /linux/tools/lib/bpf/ |
| H A D | btf_relocate.c | 104 int nelems) in search_btf_name_size() argument 107 int high = nelems - 1; in search_btf_name_size()
|
| /linux/drivers/iommu/ |
| H A D | dma-iommu.c | 1120 int nelems, enum dma_data_direction dir) in iommu_dma_sync_sg_for_cpu() argument 1126 for_each_sg(sgl, sg, nelems, i) in iommu_dma_sync_sg_for_cpu() 1130 for_each_sg(sgl, sg, nelems, i) in iommu_dma_sync_sg_for_cpu() 1135 int nelems, enum dma_data_direction dir) in iommu_dma_sync_sg_for_device() argument 1141 for_each_sg(sgl, sg, nelems, i) in iommu_dma_sync_sg_for_device() 1146 for_each_sg(sgl, sg, nelems, i) in iommu_dma_sync_sg_for_device()
|
| /linux/tools/testing/selftests/bpf/ |
| H A D | btf_helpers.c | 121 arr->type, arr->index_type, arr->nelems); in fprintf_btf_type_raw()
|
| /linux/tools/bpf/bpftool/ |
| H A D | btf.c | 171 jsonw_uint_field(w, "nr_elems", arr->nelems); in dump_btf_type() 174 arr->type, arr->index_type, arr->nelems); in dump_btf_type() 717 hash = hasher(hash, arr->nelems); in btf_type_disambig_hash()
|