/linux/drivers/net/ethernet/mellanox/mlx5/core/en/ |
H A D | rqt.c | 145 rss_rqns = kvmalloc_array(indir->actual_table_size, sizeof(*rss_rqns), GFP_KERNEL); in mlx5e_rqt_init_indir() 150 rss_vhca_ids = kvmalloc_array(indir->actual_table_size, sizeof(*rss_vhca_ids), in mlx5e_rqt_init_indir() 243 rss_rqns = kvmalloc_array(indir->actual_table_size, sizeof(*rss_rqns), GFP_KERNEL); in mlx5e_rqt_redirect_indir() 248 rss_vhca_ids = kvmalloc_array(indir->actual_table_size, sizeof(*rss_vhca_ids), in mlx5e_rqt_redirect_indir()
|
/linux/scripts/coccinelle/api/ |
H A D | kfree_mismatch.cocci | 58 kvmalloc_array\)(...) 124 * kvmalloc_array\)(...)@a 134 kvmalloc_array\)(...)
|
/linux/drivers/dma-buf/ |
H A D | udmabuf.c | 118 pages = kvmalloc_array(ubuf->pagecount, sizeof(*pages), GFP_KERNEL); in vmap_udmabuf() 210 ubuf->folios = kvmalloc_array(pgcnt, sizeof(*ubuf->folios), GFP_KERNEL); in init_udmabuf() 218 ubuf->pinned_folios = kvmalloc_array(pgcnt, in init_udmabuf() 411 folios = kvmalloc_array(max_nr_folios, sizeof(*folios), GFP_KERNEL); in udmabuf_create()
|
H A D | st-dma-fence-chain.c | 119 fc->chains = kvmalloc_array(count, sizeof(*fc->chains), in fence_chains_init() 124 fc->fences = kvmalloc_array(count, sizeof(*fc->fences), in fence_chains_init()
|
/linux/io_uring/ |
H A D | alloc_cache.c | 25 cache->entries = kvmalloc_array(max_nr, sizeof(void *), GFP_KERNEL); in io_alloc_cache_init()
|
H A D | memmap.c | 59 pages = kvmalloc_array(nr_pages, sizeof(struct page *), GFP_KERNEL); in io_pin_pages() 164 pages = kvmalloc_array(mr->nr_pages, sizeof(*pages), gfp); in io_region_allocate_pages()
|
/linux/lib/ |
H A D | bucket_locks.c | 34 tlocks = kvmalloc_array(size, sizeof(spinlock_t), gfp); in __alloc_bucket_spinlocks()
|
H A D | test_kho.c | 73 folios_info = kvmalloc_array(state->nr_folios, sizeof(*folios_info), in kho_test_save_data() 181 folios = kvmalloc_array(max_nr, sizeof(*state->folios), GFP_KERNEL); in kho_test_save()
|
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
H A D | mem.c | 194 if (!(mem->mem = kvmalloc_array(size, sizeof(*mem->mem), GFP_KERNEL))) in nvkm_mem_new_host() 196 if (!(mem->dma = kvmalloc_array(size, sizeof(*mem->dma), GFP_KERNEL))) in nvkm_mem_new_host()
|
/linux/kernel/dma/ |
H A D | remap.c | 48 pages = kvmalloc_array(count, sizeof(struct page *), GFP_KERNEL); in dma_common_contiguous_remap()
|
/linux/drivers/gpu/drm/nouveau/dispnv50/ |
H A D | lut.c | 41 in = kvmalloc_array(1024, sizeof(*in), GFP_KERNEL); in nv50_lut_load()
|
/linux/drivers/gpu/drm/ |
H A D | drm_gpusvm.c | 742 pfns = kvmalloc_array(npages, sizeof(*pfns), GFP_KERNEL); in drm_gpusvm_check_pages() 1276 pfns = kvmalloc_array(npages, sizeof(*pfns), GFP_KERNEL); in drm_gpusvm_range_get_pages() 1329 range->dma_addr = kvmalloc_array(npages, in drm_gpusvm_range_get_pages() 1501 pfns = kvmalloc_array(npages, sizeof(*pfns), GFP_KERNEL); in drm_gpusvm_range_evict()
|
H A D | drm_gem.c | 640 pages = kvmalloc_array(npages, sizeof(struct page *), GFP_KERNEL); in drm_gem_get_pages() 788 objs = kvmalloc_array(count, sizeof(struct drm_gem_object *), in drm_gem_objects_lookup() 795 handles = kvmalloc_array(count, sizeof(u32), GFP_KERNEL); in drm_gem_objects_lookup()
|
/linux/drivers/gpu/drm/etnaviv/ |
H A D | etnaviv_gem_prime.c | 129 etnaviv_obj->pages = kvmalloc_array(npages, sizeof(struct page *), GFP_KERNEL); in etnaviv_gem_prime_import_sg_table()
|
/linux/drivers/xen/ |
H A D | gntdev.c | 151 add->grants = kvmalloc_array(count, sizeof(add->grants[0]), in gntdev_alloc_map() 153 add->map_ops = kvmalloc_array(count, sizeof(add->map_ops[0]), in gntdev_alloc_map() 155 add->unmap_ops = kvmalloc_array(count, sizeof(add->unmap_ops[0]), in gntdev_alloc_map() 167 add->kmap_ops = kvmalloc_array(count, sizeof(add->kmap_ops[0]), in gntdev_alloc_map() 169 add->kunmap_ops = kvmalloc_array(count, sizeof(add->kunmap_ops[0]), in gntdev_alloc_map()
|
/linux/tools/virtio/ringtest/ |
H A D | ptr_ring.c | 61 #define kvmalloc_array kmalloc_array macro
|
/linux/drivers/net/ethernet/wangxun/txgbe/ |
H A D | txgbe_hw.c | 89 eeprom_ptrs = kvmalloc_array(TXGBE_EEPROM_LAST_WORD, sizeof(u16), in txgbe_calc_eeprom_checksum()
|
/linux/net/core/ |
H A D | devmem.c | 228 binding->tx_vec = kvmalloc_array(dmabuf->size / PAGE_SIZE, in net_devmem_bind_dmabuf() 276 owner->area.niovs = kvmalloc_array(owner->area.num_niovs, in net_devmem_bind_dmabuf()
|
/linux/drivers/net/ethernet/wangxun/ngbe/ |
H A D | ngbe_ethtool.c | 84 temp_ring = kvmalloc_array(i, sizeof(struct wx_ring), GFP_KERNEL); in ngbe_set_ringparam()
|
/linux/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_hmm.c | 184 pfns = kvmalloc_array(npages, sizeof(*pfns), GFP_KERNEL); in amdgpu_hmm_range_get_pages()
|
H A D | amdgpu_bo_list.c | 192 info = kvmalloc_array(in->bo_number, info_size, GFP_KERNEL); in amdgpu_bo_create_list_entry_array()
|
/linux/drivers/staging/media/ipu3/ |
H A D | ipu3-dmamap.c | 42 pages = kvmalloc_array(count, sizeof(*pages), GFP_KERNEL); in imgu_dmamap_alloc_buffer()
|
/linux/drivers/gpu/drm/virtio/ |
H A D | virtgpu_object.c | 178 *ents = kvmalloc_array(*nents, in virtio_gpu_object_shmem_init()
|
/linux/drivers/gpu/drm/xe/ |
H A D | xe_hmm.c | 254 pfns = kvmalloc_array(npages, sizeof(*pfns), GFP_KERNEL); in xe_hmm_userptr_populate_range()
|
/linux/kernel/module/ |
H A D | decompress.c | 21 new_pages = kvmalloc_array(info->max_pages + extent, in module_extend_max_pages()
|