| /linux/lib/ |
| H A D | bucket_locks.c | 34 tlocks = kvmalloc_objs(spinlock_t, size, gfp); in __alloc_bucket_spinlocks()
|
| H A D | test_kho.c | 214 folios = kvmalloc_objs(*state->folios, max_nr, GFP_KERNEL); in kho_test_save()
|
| /linux/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
| H A D | mem.c | 194 if (!(mem->mem = kvmalloc_objs(*mem->mem, size, GFP_KERNEL))) in nvkm_mem_new_host() 196 if (!(mem->dma = kvmalloc_objs(*mem->dma, size, GFP_KERNEL))) in nvkm_mem_new_host()
|
| /linux/kernel/dma/ |
| H A D | remap.c | 48 pages = kvmalloc_objs(struct page *, count, GFP_KERNEL); in dma_common_contiguous_remap()
|
| /linux/drivers/gpu/drm/nouveau/dispnv50/ |
| H A D | lut.c | 41 in = kvmalloc_objs(*in, 1024, GFP_KERNEL); in nv50_lut_load()
|
| /linux/drivers/gpu/drm/v3d/ |
| H A D | v3d_submit.c | 338 kvmalloc_objs(struct v3d_submit_outsync, count, in v3d_get_multisync_post_deps() 488 query_info->queries = kvmalloc_objs(struct v3d_timestamp_query, in v3d_get_cpu_timestamp_query_params() 546 query_info->queries = kvmalloc_objs(struct v3d_timestamp_query, in v3d_get_cpu_reset_timestamp_params() 602 query_info->queries = kvmalloc_objs(struct v3d_timestamp_query, in v3d_get_cpu_copy_query_results_params() 728 kvmalloc_objs(struct v3d_performance_query, reset.count, in v3d_get_cpu_reset_performance_params() 769 kvmalloc_objs(struct v3d_performance_query, copy.count, in v3d_get_cpu_copy_performance_query_params()
|
| /linux/drivers/dma-buf/ |
| H A D | udmabuf.c | 118 pages = kvmalloc_objs(*pages, ubuf->pagecount, GFP_KERNEL); in vmap_udmabuf() 210 ubuf->folios = kvmalloc_objs(*ubuf->folios, pgcnt, GFP_KERNEL); in init_udmabuf() 218 ubuf->pinned_folios = kvmalloc_objs(*ubuf->pinned_folios, pgcnt, in init_udmabuf()
|
| H A D | st-dma-fence-chain.c | 119 fc->chains = kvmalloc_objs(*fc->chains, count, GFP_KERNEL | __GFP_ZERO); in fence_chains_init() 123 fc->fences = kvmalloc_objs(*fc->fences, count, GFP_KERNEL | __GFP_ZERO); in fence_chains_init()
|
| /linux/io_uring/ |
| H A D | memmap.c | 59 pages = kvmalloc_objs(struct page *, nr_pages, GFP_KERNEL_ACCOUNT); in io_pin_pages() 161 pages = kvmalloc_objs(*pages, mr->nr_pages, gfp); in io_region_allocate_pages()
|
| /linux/drivers/gpu/drm/etnaviv/ |
| H A D | etnaviv_gem_submit.c | 471 bos = kvmalloc_objs(*bos, args->nr_bos, GFP_KERNEL); in etnaviv_ioctl_gem_submit() 472 relocs = kvmalloc_objs(*relocs, args->nr_relocs, GFP_KERNEL); in etnaviv_ioctl_gem_submit() 473 pmrs = kvmalloc_objs(*pmrs, args->nr_pmrs, GFP_KERNEL); in etnaviv_ioctl_gem_submit()
|
| H A D | etnaviv_gem_prime.c | 129 etnaviv_obj->pages = kvmalloc_objs(struct page *, npages, GFP_KERNEL); in etnaviv_gem_prime_import_sg_table()
|
| /linux/drivers/xen/ |
| H A D | gntdev.c | 148 add->grants = kvmalloc_objs(add->grants[0], count, GFP_KERNEL); in gntdev_alloc_map() 149 add->map_ops = kvmalloc_objs(add->map_ops[0], count, GFP_KERNEL); in gntdev_alloc_map() 150 add->unmap_ops = kvmalloc_objs(add->unmap_ops[0], count, GFP_KERNEL); in gntdev_alloc_map() 161 add->kmap_ops = kvmalloc_objs(add->kmap_ops[0], count, in gntdev_alloc_map() 163 add->kunmap_ops = kvmalloc_objs(add->kunmap_ops[0], count, in gntdev_alloc_map()
|
| /linux/net/core/ |
| H A D | devmem.c | 244 binding->tx_vec = kvmalloc_objs(struct net_iov *, in net_devmem_bind_dmabuf() 292 owner->area.niovs = kvmalloc_objs(*owner->area.niovs, in net_devmem_bind_dmabuf()
|
| /linux/drivers/net/ethernet/wangxun/ngbe/ |
| H A D | ngbe_ethtool.c | 84 temp_ring = kvmalloc_objs(struct wx_ring, i, GFP_KERNEL); in ngbe_set_ringparam()
|
| /linux/drivers/gpu/drm/i915/gt/ |
| H A D | shmem_utils.c | 66 pages = kvmalloc_objs(*pages, n_pages, GFP_KERNEL); in shmem_pin_map()
|
| /linux/drivers/accel/ivpu/ |
| H A D | ivpu_gem_userptr.c | 80 pages = kvmalloc_objs(*pages, nr_pages, GFP_KERNEL); in ivpu_create_userptr_dmabuf()
|
| /linux/drivers/gpu/drm/vmwgfx/ |
| H A D | vmwgfx_blit.c | 589 src_pages = kvmalloc_objs(struct page *, src->ttm->num_pages, in vmw_bo_cpu_blit() 599 dst_pages = kvmalloc_objs(struct page *, dst->ttm->num_pages, in vmw_bo_cpu_blit()
|
| /linux/drivers/accel/rocket/ |
| H A D | rocket_job.c | 74 rjob->tasks = kvmalloc_objs(*rjob->tasks, job->task_count, GFP_KERNEL); in rocket_copy_tasks() 613 jobs = kvmalloc_objs(*jobs, args->job_count, GFP_KERNEL); in rocket_ioctl_submit()
|
| /linux/drivers/staging/media/ipu3/ |
| H A D | ipu3-dmamap.c | 42 pages = kvmalloc_objs(*pages, count, GFP_KERNEL); in imgu_dmamap_alloc_buffer()
|
| /linux/drivers/gpu/drm/xen/ |
| H A D | xen_drm_front_gem.c | 50 xen_obj->pages = kvmalloc_objs(struct page *, xen_obj->num_pages, in gem_alloc_pages_array()
|
| /linux/drivers/gpu/drm/virtio/ |
| H A D | virtgpu_object.c | 180 *ents = kvmalloc_objs(struct virtio_gpu_mem_entry, *nents, GFP_KERNEL); in virtio_gpu_object_shmem_init()
|
| H A D | virtgpu_prime.c | 167 *ents = kvmalloc_objs(struct virtio_gpu_mem_entry, sgt->nents, in virtgpu_dma_buf_import_sgt()
|
| /linux/drivers/accel/amdxdna/ |
| H A D | amdxdna_ubuf.c | 192 ubuf->pages = kvmalloc_objs(*ubuf->pages, ubuf->nr_pages, GFP_KERNEL); in amdxdna_get_ubuf()
|
| /linux/drivers/usb/host/ |
| H A D | xhci-sideband.c | 31 pages = kvmalloc_objs(struct page *, n_pages, GFP_KERNEL); in xhci_ring_to_sgtable()
|
| /linux/drivers/gpu/drm/i915/gem/ |
| H A D | i915_gem_userptr.c | 261 pvec = kvmalloc_objs(struct page *, num_pages, GFP_KERNEL); in i915_gem_object_userptr_submit_init()
|