| /linux/kernel/dma/ |
| H A D | debug.h | 24 int nelems, int dir); 43 int nelems, int direction); 47 int nelems, int direction); 75 int nelems, int dir) in debug_dma_unmap_sg() argument 104 int nelems, int direction) in debug_dma_sync_sg_for_cpu() argument 110 int nelems, int direction) in debug_dma_sync_sg_for_device() argument
|
| H A D | dummy.c | 30 int nelems, enum dma_data_direction dir, in dma_dummy_map_sg() argument 37 int nelems, enum dma_data_direction dir, in dma_dummy_unmap_sg() argument
|
| H A D | mapping.c | 419 int nelems, enum dma_data_direction dir) in __dma_sync_sg_for_cpu() argument 425 dma_direct_sync_sg_for_cpu(dev, sg, nelems, dir); in __dma_sync_sg_for_cpu() 427 iommu_dma_sync_sg_for_cpu(dev, sg, nelems, dir); in __dma_sync_sg_for_cpu() 429 ops->sync_sg_for_cpu(dev, sg, nelems, dir); in __dma_sync_sg_for_cpu() 430 trace_dma_sync_sg_for_cpu(dev, sg, nelems, dir); in __dma_sync_sg_for_cpu() 431 debug_dma_sync_sg_for_cpu(dev, sg, nelems, dir); in __dma_sync_sg_for_cpu() 436 int nelems, enum dma_data_direction dir) in __dma_sync_sg_for_device() argument 442 dma_direct_sync_sg_for_device(dev, sg, nelems, dir); in __dma_sync_sg_for_device() 444 iommu_dma_sync_sg_for_device(dev, sg, nelems, dir); in __dma_sync_sg_for_device() 446 ops->sync_sg_for_device(dev, sg, nelems, dir); in __dma_sync_sg_for_device() [all …]
|
| H A D | debug.c | 1375 int nelems, int dir) in debug_dma_unmap_sg() argument 1383 for_each_sg(sglist, s, nelems, i) { in debug_dma_unmap_sg() 1392 .sg_call_ents = nelems, in debug_dma_unmap_sg() 1508 int nelems, int direction) in debug_dma_sync_sg_for_cpu() argument 1516 for_each_sg(sg, s, nelems, i) { in debug_dma_sync_sg_for_cpu() 1525 .sg_call_ents = nelems, in debug_dma_sync_sg_for_cpu() 1539 int nelems, int direction) in debug_dma_sync_sg_for_device() argument 1547 for_each_sg(sg, s, nelems, i) { in debug_dma_sync_sg_for_device() 1556 .sg_call_ents = nelems, in debug_dma_sync_sg_for_device()
|
| /linux/arch/powerpc/kernel/ |
| H A D | dma-iommu.c | 132 int nelems, enum dma_data_direction direction, in dma_iommu_map_sg() argument 135 return ppc_iommu_map_sg(dev, get_iommu_table_base(dev), sglist, nelems, in dma_iommu_map_sg() 140 int nelems, enum dma_data_direction direction, in dma_iommu_unmap_sg() argument 143 ppc_iommu_unmap_sg(get_iommu_table_base(dev), sglist, nelems, in dma_iommu_unmap_sg()
|
| H A D | iommu.c | 473 struct scatterlist *sglist, int nelems, in ppc_iommu_map_sg() argument 486 if ((nelems == 0) || !tbl) in ppc_iommu_map_sg() 491 incount = nelems; in ppc_iommu_map_sg() 497 DBG("sg mapping %d elements:\n", nelems); in ppc_iommu_map_sg() 500 for_each_sg(sglist, s, nelems, i) { in ppc_iommu_map_sg() 598 for_each_sg(sglist, s, nelems, i) { in ppc_iommu_map_sg() 616 int nelems, enum dma_data_direction direction, in ppc_iommu_unmap_sg() argument 627 while (nelems--) { in ppc_iommu_unmap_sg()
|
| /linux/arch/sparc/kernel/ |
| H A D | iommu.c | 443 int nelems, enum dma_data_direction direction, in dma_4u_map_sg() argument 460 if (nelems == 0 || !iommu) in dma_4u_map_sg() 478 incount = nelems; in dma_4u_map_sg() 487 for_each_sg(sglist, s, nelems, i) { in dma_4u_map_sg() 565 for_each_sg(sglist, s, nelems, i) { in dma_4u_map_sg() 616 int nelems, enum dma_data_direction direction, in dma_4u_unmap_sg() argument 634 while (nelems--) { in dma_4u_unmap_sg() 706 struct scatterlist *sglist, int nelems, in dma_4u_sync_sg_for_cpu() argument 738 for_each_sg(sglist, sg, nelems, i) { in dma_4u_sync_sg_for_cpu()
|
| H A D | pci_sun4v.c | 479 int nelems, enum dma_data_direction direction, in dma_4v_map_sg() argument 498 if (nelems == 0 || !iommu) in dma_4v_map_sg() 511 incount = nelems; in dma_4v_map_sg() 532 for_each_sg(sglist, s, nelems, i) { in dma_4v_map_sg() 613 for_each_sg(sglist, s, nelems, i) { in dma_4v_map_sg() 634 int nelems, enum dma_data_direction direction, in dma_4v_unmap_sg() argument 655 while (nelems--) { in dma_4v_unmap_sg()
|
| /linux/include/linux/ |
| H A D | dma-mapping.h | 424 int nelems, enum dma_data_direction dir); 426 int nelems, enum dma_data_direction dir); 450 struct scatterlist *sg, int nelems, enum dma_data_direction dir) in dma_sync_sg_for_cpu() argument 453 __dma_sync_sg_for_cpu(dev, sg, nelems, dir); in dma_sync_sg_for_cpu() 457 struct scatterlist *sg, int nelems, enum dma_data_direction dir) in dma_sync_sg_for_device() argument 460 __dma_sync_sg_for_device(dev, sg, nelems, dir); in dma_sync_sg_for_device() 482 struct scatterlist *sg, int nelems, enum dma_data_direction dir) in dma_sync_sg_for_cpu() argument 486 struct scatterlist *sg, int nelems, enum dma_data_direction dir) in dma_sync_sg_for_device() argument
|
| /linux/net/netfilter/ |
| H A D | nft_set_rbtree.c | 580 u32 nelems = atomic_read(&set->nelems) - set->ndeact; in nft_array_elems() local 586 return nelems; in nft_array_elems() 593 return (nelems / 2) + 2; in nft_array_elems() 603 u32 shrinked_max_intervals, nelems = nft_array_elems(set); in nft_array_may_resize() local 630 nelems = 0; in nft_array_may_resize() 642 nelems < shrinked_max_intervals) { in nft_array_may_resize() 647 if (nelems > new_max_intervals) { in nft_array_may_resize() 660 if (WARN_ON_ONCE(nelems > new_max_intervals)) in nft_array_may_resize()
|
| H A D | nft_dynset.c | 64 if (!atomic_add_unless(&set->nelems, 1, set->size)) in nft_dynset_new() 85 atomic_dec(&set->nelems); in nft_dynset_new()
|
| H A D | nf_tables_api.c | 488 return a->set == b->set && a->bound == b->bound && a->nelems < NFT_MAX_SET_NELEMS; in nft_trans_collapse_set_elem() 495 unsigned int nelems, old_nelems = tail->nelems; in nft_trans_collapse_set_elem() 507 if (WARN_ON_ONCE(trans->nelems != 1)) in nft_trans_try_collapse() 510 if (check_add_overflow(old_nelems, trans->nelems, &nelems)) in nft_trans_try_collapse() 516 new_trans = krealloc(tail, struct_size(tail, elems, nelems), in nft_trans_try_collapse() 528 new_trans->nelems = nelems; 4942 unsigned int nelems; in nf_tables_fill_set() 465 unsigned int nelems, old_nelems = tail->nelems; nft_trans_collapse_set_elem() local 4823 unsigned int nelems; nf_tables_fill_set_info() local 6492 int rem, err = 0, nelems = 0; nf_tables_getsetelem() local 7413 unsigned int max = nft_set_maxsize(set), nelems; nft_add_set_elem() local [all...] |
| /linux/arch/powerpc/kvm/ |
| H A D | guest-state-buffer.c | 45 gsb->hdr->nelems = cpu_to_be32(0); in kvmppc_gsb_new() 78 u32 nelems = kvmppc_gsb_nelems(gsb); in kvmppc_gsb_put() local 84 kvmppc_gsb_header(gsb)->nelems = cpu_to_be32(nelems + 1); in kvmppc_gsb_put()
|
| /linux/arch/x86/kernel/ |
| H A D | amd_gart_64.c | 324 int nelems, struct scatterlist *sout, in __dma_map_cont() argument 335 for_each_sg(start, s, nelems, i) { in __dma_map_cont() 362 dma_map_cont(struct device *dev, struct scatterlist *start, int nelems, in dma_map_cont() argument 366 BUG_ON(nelems != 1); in dma_map_cont() 371 return __dma_map_cont(dev, start, nelems, sout, pages); in dma_map_cont()
|
| /linux/arch/powerpc/include/asm/ |
| H A D | iommu.h | 262 struct scatterlist *sglist, int nelems, 268 int nelems,
|
| /linux/tools/include/uapi/linux/ |
| H A D | btf.h | 127 __u32 nelems; 114 __u32 nelems; global() member
|
| /linux/arch/powerpc/platforms/pseries/ |
| H A D | vio.c | 555 int nelems, enum dma_data_direction direction, in vio_dma_iommu_map_sg() argument 564 for_each_sg(sglist, sgl, nelems, count) in vio_dma_iommu_map_sg() 570 ret = ppc_iommu_map_sg(dev, tbl, sglist, nelems, dma_get_mask(dev), in vio_dma_iommu_map_sg() 589 struct scatterlist *sglist, int nelems, in vio_dma_iommu_unmap_sg() argument 599 for_each_sg(sglist, sgl, nelems, count) in vio_dma_iommu_unmap_sg() 602 ppc_iommu_unmap_sg(tbl, sglist, nelems, direction, attrs); in vio_dma_iommu_unmap_sg()
|
| /linux/kernel/bpf/ |
| H A D | btf.c | 2084 u32 i, size, nelems = 1, id = 0; in __btf_resolve_size() local 2116 if (nelems && array->nelems > U32_MAX / nelems) in __btf_resolve_size() 2118 nelems *= array->nelems; in __btf_resolve_size() 2131 if (nelems && size > U32_MAX / nelems) in __btf_resolve_size() 2134 *type_size = nelems * size; in __btf_resolve_size() 2136 *total_nelems = nelems; in __btf_resolve_size() 3146 if (array->nelems && elem_size > U32_MAX / array->nelems) { in btf_array_resolve() 3152 env_stack_pop_resolved(env, elem_type_id, elem_size * array->nelems); in btf_array_resolve() 3163 array->type, array->index_type, array->nelems); in btf_array_log() 3202 for (i = 0; i < array->nelems; i++) { in __btf_array_show() [all …]
|
| /linux/tools/lib/bpf/ |
| H A D | btf_relocate.c | 104 int nelems) in search_btf_name_size() argument 107 int high = nelems - 1; in search_btf_name_size()
|
| /linux/drivers/iommu/ |
| H A D | dma-iommu.c | 1135 int nelems, enum dma_data_direction dir) in iommu_dma_sync_sg_for_cpu() argument 1141 for_each_sg(sgl, sg, nelems, i) in iommu_dma_sync_sg_for_cpu() 1145 for_each_sg(sgl, sg, nelems, i) in iommu_dma_sync_sg_for_cpu() 1152 int nelems, enum dma_data_direction dir) in iommu_dma_sync_sg_for_device() argument 1158 for_each_sg(sgl, sg, nelems, i) in iommu_dma_sync_sg_for_device() 1163 for_each_sg(sgl, sg, nelems, i) in iommu_dma_sync_sg_for_device()
|
| /linux/tools/testing/selftests/bpf/ |
| H A D | btf_helpers.c | 121 arr->type, arr->index_type, arr->nelems); in fprintf_btf_type_raw()
|
| /linux/tools/bpf/bpftool/ |
| H A D | btf.c | 172 jsonw_uint_field(w, "nr_elems", arr->nelems); in dump_btf_type() 175 arr->type, arr->index_type, arr->nelems); in dump_btf_type() 718 hash = hasher(hash, arr->nelems); in btf_type_disambig_hash()
|
| /linux/lib/ |
| H A D | test_rhashtable.c | 203 total, atomic_read(&ht->nelems), entries, chain_len); in test_bucket_stats() 205 if (total != atomic_read(&ht->nelems) || total != entries) in test_bucket_stats()
|
| /linux/include/linux/mtd/ |
| H A D | rawnand.h | 965 unsigned int nelems; member 973 .nelems = sizeof((struct nand_op_parser_pattern_elem[]) { __VA_ARGS__ }) / \
|
| /linux/include/net/netfilter/ |
| H A D | nf_tables.h | 555 * @nelems: number of elements 590 atomic_t nelems; member 1807 unsigned int nelems; 1808 struct nft_trans_one_elem elems[] __counted_by(nelems); 1794 unsigned int nelems; global() member
|