/linux/kernel/dma/ |
H A D | debug.h | 25 int nelems, int dir); 52 int nelems, int direction); 56 int nelems, int direction); 78 int nelems, int dir) in debug_dma_unmap_sg() argument 120 int nelems, int direction) in debug_dma_sync_sg_for_cpu() argument 126 int nelems, int direction) in debug_dma_sync_sg_for_device() argument
|
H A D | mapping.c | 398 int nelems, enum dma_data_direction dir) in __dma_sync_sg_for_cpu() argument 404 dma_direct_sync_sg_for_cpu(dev, sg, nelems, dir); in __dma_sync_sg_for_cpu() 406 iommu_dma_sync_sg_for_cpu(dev, sg, nelems, dir); in __dma_sync_sg_for_cpu() 408 ops->sync_sg_for_cpu(dev, sg, nelems, dir); in __dma_sync_sg_for_cpu() 409 trace_dma_sync_sg_for_cpu(dev, sg, nelems, dir); in __dma_sync_sg_for_cpu() 410 debug_dma_sync_sg_for_cpu(dev, sg, nelems, dir); in __dma_sync_sg_for_cpu() 415 int nelems, enum dma_data_direction dir) in __dma_sync_sg_for_device() argument 421 dma_direct_sync_sg_for_device(dev, sg, nelems, dir); in __dma_sync_sg_for_device() 423 iommu_dma_sync_sg_for_device(dev, sg, nelems, dir); in __dma_sync_sg_for_device() 425 ops->sync_sg_for_device(dev, sg, nelems, dir); in __dma_sync_sg_for_device() [all …]
|
H A D | dummy.c | 31 int nelems, enum dma_data_direction dir, in dma_dummy_map_sg() argument 38 int nelems, enum dma_data_direction dir, in dma_dummy_unmap_sg() argument
|
H A D | debug.c | 1350 int nelems, int dir) in debug_dma_unmap_sg() argument 1358 for_each_sg(sglist, s, nelems, i) { in debug_dma_unmap_sg() 1367 .sg_call_ents = nelems, in debug_dma_unmap_sg() 1524 int nelems, int direction) in debug_dma_sync_sg_for_cpu() argument 1532 for_each_sg(sg, s, nelems, i) { in debug_dma_sync_sg_for_cpu() 1541 .sg_call_ents = nelems, in debug_dma_sync_sg_for_cpu() 1555 int nelems, int direction) in debug_dma_sync_sg_for_device() argument 1563 for_each_sg(sg, s, nelems, i) { in debug_dma_sync_sg_for_device() 1572 .sg_call_ents = nelems, in debug_dma_sync_sg_for_device()
|
/linux/drivers/xen/ |
H A D | swiotlb-xen.c | 326 xen_swiotlb_unmap_sg(struct device *hwdev, struct scatterlist *sgl, int nelems, in xen_swiotlb_unmap_sg() argument 334 for_each_sg(sgl, sg, nelems, i) in xen_swiotlb_unmap_sg() 341 xen_swiotlb_map_sg(struct device *dev, struct scatterlist *sgl, int nelems, in xen_swiotlb_map_sg() argument 349 for_each_sg(sgl, sg, nelems, i) { in xen_swiotlb_map_sg() 357 return nelems; in xen_swiotlb_map_sg() 366 int nelems, enum dma_data_direction dir) in xen_swiotlb_sync_sg_for_cpu() argument 371 for_each_sg(sgl, sg, nelems, i) { in xen_swiotlb_sync_sg_for_cpu() 379 int nelems, enum dma_data_direction dir) in xen_swiotlb_sync_sg_for_device() argument 384 for_each_sg(sgl, sg, nelems, i) { in xen_swiotlb_sync_sg_for_device()
|
/linux/arch/sparc/kernel/ |
H A D | iommu.c | 433 int nelems, enum dma_data_direction direction, in dma_4u_map_sg() argument 450 if (nelems == 0 || !iommu) in dma_4u_map_sg() 468 incount = nelems; in dma_4u_map_sg() 477 for_each_sg(sglist, s, nelems, i) { in dma_4u_map_sg() 555 for_each_sg(sglist, s, nelems, i) { in dma_4u_map_sg() 606 int nelems, enum dma_data_direction direction, in dma_4u_unmap_sg() argument 624 while (nelems--) { in dma_4u_unmap_sg() 696 struct scatterlist *sglist, int nelems, in dma_4u_sync_sg_for_cpu() argument 728 for_each_sg(sglist, sg, nelems, i) { in dma_4u_sync_sg_for_cpu()
|
/linux/include/linux/ |
H A D | dma-mapping.h | 286 int nelems, enum dma_data_direction dir); 288 int nelems, enum dma_data_direction dir); 312 struct scatterlist *sg, int nelems, enum dma_data_direction dir) in dma_sync_sg_for_cpu() argument 315 __dma_sync_sg_for_cpu(dev, sg, nelems, dir); in dma_sync_sg_for_cpu() 319 struct scatterlist *sg, int nelems, enum dma_data_direction dir) in dma_sync_sg_for_device() argument 322 __dma_sync_sg_for_device(dev, sg, nelems, dir); in dma_sync_sg_for_device() 343 struct scatterlist *sg, int nelems, enum dma_data_direction dir) in dma_sync_sg_for_cpu() argument 347 struct scatterlist *sg, int nelems, enum dma_data_direction dir) in dma_sync_sg_for_device() argument
|
H A D | iommu-dma.h | 65 int nelems, enum dma_data_direction dir); 67 int nelems, enum dma_data_direction dir);
|
H A D | rhashtable.h | 186 return atomic_read(&ht->nelems) > (tbl->size / 4 * 3) && in rht_grow_above_75() 199 return atomic_read(&ht->nelems) < (tbl->size * 3 / 10) && in rht_shrink_below_30() 211 return atomic_read(&ht->nelems) > tbl->size && in rht_grow_above_100() 223 return atomic_read(&ht->nelems) >= ht->max_elems; in rht_grow_above_max() 796 atomic_inc(&ht->nelems); in __rhashtable_insert_fast() 1066 atomic_dec(&ht->nelems); in __rhashtable_remove_fast_one()
|
/linux/arch/powerpc/kvm/ |
H A D | guest-state-buffer.c | 45 gsb->hdr->nelems = cpu_to_be32(0); in kvmppc_gsb_new() 78 u32 nelems = kvmppc_gsb_nelems(gsb); in kvmppc_gsb_put() local 84 kvmppc_gsb_header(gsb)->nelems = cpu_to_be32(nelems + 1); in kvmppc_gsb_put()
|
/linux/arch/powerpc/kernel/ |
H A D | iommu.c | 472 struct scatterlist *sglist, int nelems, in ppc_iommu_map_sg() argument 485 if ((nelems == 0) || !tbl) in ppc_iommu_map_sg() 490 incount = nelems; in ppc_iommu_map_sg() 496 DBG("sg mapping %d elements:\n", nelems); in ppc_iommu_map_sg() 499 for_each_sg(sglist, s, nelems, i) { in ppc_iommu_map_sg() 597 for_each_sg(sglist, s, nelems, i) { in ppc_iommu_map_sg() 615 int nelems, enum dma_data_direction direction, in ppc_iommu_unmap_sg() argument 626 while (nelems--) { in ppc_iommu_unmap_sg()
|
/linux/lib/ |
H A D | rhashtable.c | 397 unsigned int nelems = atomic_read(&ht->nelems); in rhashtable_shrink() local 400 if (nelems) in rhashtable_shrink() 401 size = roundup_pow_of_two(nelems * 3 / 2); in rhashtable_shrink() 587 atomic_inc(&ht->nelems); in rhashtable_insert_one() 1077 atomic_set(&ht->nelems, 0); in rhashtable_init_noprof()
|
/linux/arch/powerpc/include/asm/ |
H A D | iommu.h | 262 struct scatterlist *sglist, int nelems, 268 int nelems,
|
H A D | guest-state-buffer.h | 208 __be32 nelems; member 385 return be32_to_cpu(kvmppc_gsb_header(gsb)->nelems); in kvmppc_gsb_nelems() 396 kvmppc_gsb_header(gsb)->nelems = cpu_to_be32(0); in kvmppc_gsb_reset()
|
/linux/tools/bpf/bpftool/ |
H A D | btf_dumper.c | 217 if (!arr->nelems) in is_str_array() 234 end_s = s + arr->nelems; in is_str_array() 266 for (i = 0; i < arr->nelems; i++) { in btf_dumper_array() 656 BTF_PRINT_ARG("[%d]", array->nelems); in __btf_dumper_type_only()
|
/linux/include/uapi/linux/ |
H A D | btf.h | 114 __u32 nelems; member
|
/linux/tools/include/uapi/linux/ |
H A D | btf.h | 114 __u32 nelems; member
|
/linux/arch/powerpc/platforms/pseries/ |
H A D | vio.c | 552 int nelems, enum dma_data_direction direction, in vio_dma_iommu_map_sg() argument 561 for_each_sg(sglist, sgl, nelems, count) in vio_dma_iommu_map_sg() 567 ret = ppc_iommu_map_sg(dev, tbl, sglist, nelems, dma_get_mask(dev), in vio_dma_iommu_map_sg() 586 struct scatterlist *sglist, int nelems, in vio_dma_iommu_unmap_sg() argument 596 for_each_sg(sglist, sgl, nelems, count) in vio_dma_iommu_unmap_sg() 599 ppc_iommu_unmap_sg(tbl, sglist, nelems, direction, attrs); in vio_dma_iommu_unmap_sg()
|
/linux/tools/lib/bpf/ |
H A D | btf.c | 382 a->nelems = bswap_32(a->nelems); in btf_bswap_type_rest() 771 __u32 nelems = 1; in btf__resolve_size() local 801 if (nelems && array->nelems > UINT32_MAX / nelems) in btf__resolve_size() 803 nelems *= array->nelems; in btf__resolve_size() 816 if (nelems && size > UINT32_MAX / nelems) in btf__resolve_size() 819 return nelems * size; in btf__resolve_size() 2166 a->nelems = nr_elems; in btf__add_array() 3975 h = hash_combine(h, info->nelems); in btf_hash_array() 3997 info1->nelems == info2->nelems; in btf_equal_array() 4010 return btf_array(t1)->nelems == btf_array(t2)->nelems; in btf_compat_array()
|
H A D | btf_relocate.c | 104 int nelems) in search_btf_name_size() argument 107 int high = nelems - 1; in search_btf_name_size()
|
/linux/kernel/bpf/ |
H A D | btf.c | 1944 u32 i, size, nelems = 1, id = 0; in __btf_resolve_size() local 1976 if (nelems && array->nelems > U32_MAX / nelems) in __btf_resolve_size() 1978 nelems *= array->nelems; in __btf_resolve_size() 1991 if (nelems && size > U32_MAX / nelems) in __btf_resolve_size() 1994 *type_size = nelems * size; in __btf_resolve_size() 1996 *total_nelems = nelems; in __btf_resolve_size() 3006 if (array->nelems && elem_size > U32_MAX / array->nelems) { in btf_array_resolve() 3012 env_stack_pop_resolved(env, elem_type_id, elem_size * array->nelems); in btf_array_resolve() 3023 array->type, array->index_type, array->nelems); in btf_array_log() 3062 for (i = 0; i < array->nelems; i++) { in __btf_array_show() [all …]
|
/linux/drivers/iommu/ |
H A D | dma-iommu.c | 1133 int nelems, enum dma_data_direction dir) in iommu_dma_sync_sg_for_cpu() argument 1139 for_each_sg(sgl, sg, nelems, i) in iommu_dma_sync_sg_for_cpu() 1143 for_each_sg(sgl, sg, nelems, i) in iommu_dma_sync_sg_for_cpu() 1148 int nelems, enum dma_data_direction dir) in iommu_dma_sync_sg_for_device() argument 1154 for_each_sg(sgl, sg, nelems, i) in iommu_dma_sync_sg_for_device() 1159 for_each_sg(sgl, sg, nelems, i) in iommu_dma_sync_sg_for_device()
|
/linux/net/netfilter/ |
H A D | nft_dynset.c | 56 if (!atomic_add_unless(&set->nelems, 1, set->size)) in nft_dynset_new() 77 atomic_dec(&set->nelems); in nft_dynset_new()
|
/linux/tools/testing/selftests/bpf/ |
H A D | btf_helpers.c | 121 arr->type, arr->index_type, arr->nelems); in fprintf_btf_type_raw()
|
/linux/net/ipv6/ |
H A D | proc.c | 47 atomic_read(&net->ipv6.fqdir->rhashtable.nelems), in sockstat6_seq_show()
|