Home
last modified time | relevance | path

Searched refs:num_pages (Results 1 – 25 of 221) sorted by relevance

123456789

/linux/tools/testing/selftests/mm/
H A Dmremap_dontunmap.c45 unsigned long num_pages = 1; in kernel_support_for_mremap_dontunmap() local
46 void *source_mapping = mmap(NULL, num_pages * page_size, PROT_NONE, in kernel_support_for_mremap_dontunmap()
53 mremap(source_mapping, num_pages * page_size, num_pages * page_size, in kernel_support_for_mremap_dontunmap()
58 BUG_ON(munmap(dest_mapping, num_pages * page_size) == -1, in kernel_support_for_mremap_dontunmap()
62 BUG_ON(munmap(source_mapping, num_pages * page_size) == -1, in kernel_support_for_mremap_dontunmap()
78 unsigned long num_pages = size / page_size; in check_region_contains_byte() local
82 for (i = 0; i < num_pages; ++i) { in check_region_contains_byte()
97 unsigned long num_pages = 5; in mremap_dontunmap_simple() local
100 mmap(NULL, num_pages * page_size, PROT_READ | PROT_WRITE, in mremap_dontunmap_simple()
104 memset(source_mapping, 'a', num_pages * page_size); in mremap_dontunmap_simple()
[all …]
/linux/drivers/gpu/drm/vmwgfx/
H A Dvmwgfx_gmr.c39 unsigned long num_pages, in vmw_gmr2_bind() argument
47 uint32_t remap_num = num_pages / VMW_PPN_PER_REMAP + ((num_pages % VMW_PPN_PER_REMAP) > 0); in vmw_gmr2_bind()
48 uint32_t remap_size = VMW_PPN_SIZE * num_pages + (sizeof(remap_cmd) + sizeof(*cmd)) * remap_num; in vmw_gmr2_bind()
58 define_cmd.numPages = num_pages; in vmw_gmr2_bind()
73 while (num_pages > 0) { in vmw_gmr2_bind()
74 unsigned long nr = min_t(unsigned long, num_pages, VMW_PPN_PER_REMAP); in vmw_gmr2_bind()
94 num_pages -= nr; in vmw_gmr2_bind()
128 unsigned long num_pages, in vmw_gmr_bind() argument
141 return vmw_gmr2_bind(dev_priv, &data_iter, num_pages, gmr_id); in vmw_gmr_bind()
H A Dvmwgfx_ttm_buffer.c70 return ++(viter->i) < viter->num_pages; in __vmw_piter_non_sg_next()
107 viter->num_pages = vsgt->num_pages; in vmw_piter_start()
184 vsgt->num_pages = vmw_tt->dma_ttm.num_pages; in vmw_ttm_map_dma()
196 vsgt->pages, vsgt->num_pages, 0, in vmw_ttm_map_dma()
197 (unsigned long)vsgt->num_pages << PAGE_SHIFT, in vmw_ttm_map_dma()
296 ttm->num_pages, vmw_be->gmr_id); in vmw_ttm_bind()
301 vmw_mob_create(ttm->num_pages); in vmw_ttm_bind()
307 &vmw_be->vsgt, ttm->num_pages, in vmw_ttm_bind()
373 ttm->num_pages); in vmw_ttm_populate()
/linux/drivers/gpu/drm/ttm/
H A Dttm_tt.c120 ttm->pages = kvcalloc(ttm->num_pages, sizeof(void*), GFP_KERNEL); in ttm_tt_alloc_page_directory()
129 ttm->pages = kvcalloc(ttm->num_pages, sizeof(*ttm->pages) + in ttm_dma_tt_alloc_page_directory()
134 ttm->dma_address = (void *)(ttm->pages + ttm->num_pages); in ttm_dma_tt_alloc_page_directory()
140 ttm->dma_address = kvzalloc_objs(*ttm->dma_address, ttm->num_pages); in ttm_sg_tt_alloc_page_directory()
159 ttm->num_pages = (PAGE_ALIGN(bo->base.size) >> PAGE_SHIFT) + extra_pages; in ttm_tt_init_fields()
241 for (i = 0; i < ttm->num_pages; ++i) { in ttm_tt_swapin()
325 loff_t size = (loff_t)ttm->num_pages << PAGE_SHIFT; in ttm_tt_swapout()
341 for (i = 0; i < ttm->num_pages; ++i) { in ttm_tt_swapout()
361 return ttm->num_pages; in ttm_tt_swapout()
382 atomic_long_add(ttm->num_pages, &ttm_pages_allocated); in ttm_tt_populate()
[all …]
H A Dttm_pool.c231 unsigned int num_pages = alloc->pages - alloc->caching_divide; in ttm_pool_apply_caching() local
233 if (!num_pages) in ttm_pool_apply_caching()
240 return set_pages_array_wc(alloc->caching_divide, num_pages); in ttm_pool_apply_caching()
242 return set_pages_array_uc(alloc->caching_divide, num_pages); in ttm_pool_apply_caching()
274 unsigned int num_pages) in ttm_pool_unmap() argument
280 dma_unmap_page(pool->dev, dma_addr, (long)num_pages << PAGE_SHIFT, in ttm_pool_unmap()
287 unsigned int i, num_pages = 1 << pt->order; in ttm_pool_type_give() local
289 for (i = 0; i < num_pages; ++i) { in ttm_pool_type_give()
384 unsigned int num_pages; in ttm_pool_shrink() local
396 num_pages = 1 << pt->order; in ttm_pool_shrink()
[all …]
/linux/drivers/xen/
H A Dxen-front-pgdir-shbuf.c160 return DIV_ROUND_UP(buf->num_pages, XEN_NUM_GREFS_PER_PAGE); in get_num_pages_dir()
187 buf->num_grefs = get_num_pages_dir(buf) + buf->num_pages; in guest_calc_num_grefs()
208 unmap_ops = kzalloc_objs(*unmap_ops, buf->num_pages); in backend_unmap()
212 for (i = 0; i < buf->num_pages; i++) { in backend_unmap()
221 buf->num_pages); in backend_unmap()
223 for (i = 0; i < buf->num_pages; i++) { in backend_unmap()
252 map_ops = kzalloc_objs(*map_ops, buf->num_pages); in backend_map()
257 buf->num_pages); in backend_map()
269 grefs_left = buf->num_pages; in backend_map()
293 ret = gnttab_map_refs(map_ops, NULL, buf->pages, buf->num_pages); in backend_map()
[all …]
/linux/drivers/hv/
H A Dhv_proc.c19 int hv_call_deposit_pages(int node, u64 partition_id, u32 num_pages) in hv_call_deposit_pages() argument
32 if (num_pages > HV_DEPOSIT_MAX) in hv_call_deposit_pages()
34 if (!num_pages) in hv_call_deposit_pages()
52 while (num_pages) { in hv_call_deposit_pages()
54 order = 31 - __builtin_clz(num_pages); in hv_call_deposit_pages()
70 num_pages -= counts[i]; in hv_call_deposit_pages()
116 u32 num_pages = 1; in hv_deposit_memory_node() local
122 num_pages = HV_MAX_CONTIGUOUS_ALLOCATION_PAGES; in hv_deposit_memory_node()
126 num_pages = HV_MAX_CONTIGUOUS_ALLOCATION_PAGES; in hv_deposit_memory_node()
140 return hv_call_deposit_pages(node, partition_id, num_pages); in hv_deposit_memory_node()
H A Dhv_balloon.c276 __u32 num_pages; member
443 __u32 num_pages; member
1185 int num_pages = range_array->finfo.page_cnt; in free_balloon_pages() local
1190 for (i = 0; i < num_pages; i++) { in free_balloon_pages()
1201 unsigned int num_pages, in alloc_balloon_pages() argument
1208 for (i = 0; i < num_pages / alloc_unit; i++) { in alloc_balloon_pages()
1253 unsigned int num_pages = dm_device.balloon_wrk.num_pages; in balloon_up() local
1273 if (avail_pages < num_pages || avail_pages - num_pages < floor) { in balloon_up()
1275 avail_pages < num_pages ? "Not enough memory." : in balloon_up()
1278 num_pages = avail_pages > floor ? (avail_pages - floor) : 0; in balloon_up()
[all …]
/linux/drivers/gpu/drm/xen/
H A Dxen_drm_front_gem.c30 size_t num_pages; member
49 xen_obj->num_pages = DIV_ROUND_UP(buf_size, PAGE_SIZE); in gem_alloc_pages_array()
50 xen_obj->pages = kvmalloc_objs(struct page *, xen_obj->num_pages); in gem_alloc_pages_array()
93 ret = vm_map_pages(vma, xen_obj->pages, xen_obj->num_pages); in xen_drm_front_gem_object_mmap()
159 ret = xen_alloc_unpopulated_pages(xen_obj->num_pages, in gem_create()
163 xen_obj->num_pages, ret); in gem_create()
175 xen_obj->num_pages = DIV_ROUND_UP(size, PAGE_SIZE); in gem_create()
212 xen_free_unpopulated_pages(xen_obj->num_pages, in xen_drm_front_gem_free_object_unlocked()
240 xen_obj->pages, xen_obj->num_pages); in xen_drm_front_gem_get_sg_table()
265 xen_obj->num_pages); in xen_drm_front_gem_import_sg_table()
[all …]
/linux/tools/testing/scatterlist/
H A Dmain.c10 unsigned num_pages; member
40 printf("%u input PFNs:", test->num_pages); in fail()
41 for (i = 0; i < test->num_pages; i++) in fail()
87 int left_pages = test->pfn_app ? test->num_pages : 0; in main()
92 set_pages(pages, test->pfn, test->num_pages); in main()
96 &append, pages, test->num_pages, 0, test->size, in main()
100 &append.sgt, pages, test->num_pages, 0, in main()
109 set_pages(pages, test->pfn_app, test->num_pages); in main()
111 &append, pages, test->num_pages, 0, test->size, in main()
/linux/net/ceph/
H A Dpagevec.c13 void ceph_put_page_vector(struct page **pages, int num_pages, bool dirty) in ceph_put_page_vector() argument
17 for (i = 0; i < num_pages; i++) { in ceph_put_page_vector()
26 void ceph_release_page_vector(struct page **pages, int num_pages) in ceph_release_page_vector() argument
30 for (i = 0; i < num_pages; i++) in ceph_release_page_vector()
39 struct page **ceph_alloc_page_vector(int num_pages, gfp_t flags) in ceph_alloc_page_vector() argument
44 pages = kmalloc_objs(*pages, num_pages, flags); in ceph_alloc_page_vector()
47 for (i = 0; i < num_pages; i++) { in ceph_alloc_page_vector()
/linux/drivers/media/common/videobuf2/
H A Dvideobuf2-dma-sg.c49 unsigned int num_pages; member
107 int num_pages; in vb2_dma_sg_alloc() local
121 buf->num_pages = size >> PAGE_SHIFT; in vb2_dma_sg_alloc()
129 buf->pages = kvzalloc_objs(struct page *, buf->num_pages); in vb2_dma_sg_alloc()
138 buf->num_pages, 0, size, GFP_KERNEL); in vb2_dma_sg_alloc()
162 __func__, buf->num_pages); in vb2_dma_sg_alloc()
169 num_pages = buf->num_pages; in vb2_dma_sg_alloc()
170 while (num_pages--) in vb2_dma_sg_alloc()
171 __free_page(buf->pages[num_pages]); in vb2_dma_sg_alloc()
183 int i = buf->num_pages; in vb2_dma_sg_put()
[all …]
/linux/arch/x86/platform/efi/
H A Dmemmap.c129 end = start + (md->num_pages << EFI_PAGE_SHIFT) - 1; in efi_memmap_split_count()
194 end = md->phys_addr + (md->num_pages << EFI_PAGE_SHIFT) - 1; in efi_memmap_insert()
203 md->num_pages = (m_end - md->phys_addr + 1) >> in efi_memmap_insert()
210 md->num_pages = (end - md->phys_addr + 1) >> in efi_memmap_insert()
216 md->num_pages = (m_start - md->phys_addr) >> in efi_memmap_insert()
224 md->num_pages = (m_end - m_start + 1) >> in efi_memmap_insert()
231 md->num_pages = (end - m_end) >> in efi_memmap_insert()
238 md->num_pages = (m_start - md->phys_addr) >> in efi_memmap_insert()
245 md->num_pages = (end - md->phys_addr + 1) >> in efi_memmap_insert()
H A Defi_32.c43 size = md->num_pages << PAGE_SHIFT; in efi_map_region()
51 set_memory_uc((unsigned long)va, md->num_pages); in efi_map_region()
82 int __init efi_setup_page_tables(unsigned long pa_memmap, unsigned num_pages) in efi_setup_page_tables() argument
139 set_memory_x(md->virt_addr, md->num_pages); in efi_runtime_update_mappings()
/linux/drivers/infiniband/sw/siw/
H A Dsiw_mem.c44 int i, num_pages = umem->num_pages; in siw_umem_release() local
49 for (i = 0; num_pages > 0; i++) { in siw_umem_release()
51 num_pages -= PAGES_PER_CHUNK; in siw_umem_release()
341 int num_pages, num_chunks, i, rv = 0; in siw_umem_get() local
347 num_pages = PAGE_ALIGN(start + len - first_page_va) >> PAGE_SHIFT; in siw_umem_get()
348 num_chunks = (num_pages >> CHUNK_SHIFT) + 1; in siw_umem_get()
376 for (i = 0; num_pages > 0; i++) { in siw_umem_get()
377 int nents = min_t(int, num_pages, PAGES_PER_CHUNK); in siw_umem_get()
388 umem->num_pages++; in siw_umem_get()
389 num_pages--; in siw_umem_get()
/linux/drivers/tee/
H A Dtee_shm.c46 unpin_user_pages(shm->pages, shm->num_pages); in release_registered_pages()
48 shm_put_kernel_pages(shm->pages, shm->num_pages); in release_registered_pages()
356 size_t num_pages, in tee_dyn_shm_alloc_helper() argument
386 shm->num_pages = nr_pages; in tee_dyn_shm_alloc_helper()
425 size_t num_pages, off; in register_shm_helper() local
453 num_pages = iov_iter_npages(iter, INT_MAX); in register_shm_helper()
454 if (!num_pages) { in register_shm_helper()
459 shm->pages = kzalloc_objs(*shm->pages, num_pages); in register_shm_helper()
465 len = iov_iter_extract_pages(iter, &shm->pages, LONG_MAX, num_pages, 0, in register_shm_helper()
470 } else if (DIV_ROUND_UP(len + off, PAGE_SIZE) != num_pages) { in register_shm_helper()
[all …]
/linux/drivers/gpu/drm/i915/gem/
H A Di915_gem_userptr.c93 const unsigned long num_pages = obj->base.size >> PAGE_SHIFT; in i915_gem_object_userptr_drop_ref() local
95 unpin_user_pages(pvec, num_pages); in i915_gem_object_userptr_drop_ref()
105 unsigned int num_pages; /* limited by sg_alloc_table_from_pages_segment */ in i915_gem_userptr_get_pages() local
108 if (overflows_type(obj->base.size >> PAGE_SHIFT, num_pages)) in i915_gem_userptr_get_pages()
111 num_pages = obj->base.size >> PAGE_SHIFT; in i915_gem_userptr_get_pages()
125 ret = sg_alloc_table_from_pages_segment(st, pvec, num_pages, 0, in i915_gem_userptr_get_pages()
126 num_pages << PAGE_SHIFT, in i915_gem_userptr_get_pages()
236 const unsigned long num_pages = obj->base.size >> PAGE_SHIFT; in i915_gem_object_userptr_submit_init() local
261 pvec = kvmalloc_objs(struct page *, num_pages); in i915_gem_object_userptr_submit_init()
269 while (pinned < num_pages) { in i915_gem_object_userptr_submit_init()
[all …]
/linux/drivers/gpu/drm/gma500/
H A Dmmu.c479 uint32_t num_pages, uint32_t desired_tile_stride, in psb_mmu_flush_ptes() argument
497 rows = num_pages / desired_tile_stride; in psb_mmu_flush_ptes()
499 desired_tile_stride = num_pages; in psb_mmu_flush_ptes()
527 unsigned long address, uint32_t num_pages) in psb_mmu_remove_pfn_sequence() argument
538 end = addr + (num_pages << PAGE_SHIFT); in psb_mmu_remove_pfn_sequence()
555 psb_mmu_flush_ptes(pd, f_address, num_pages, 1, 1); in psb_mmu_remove_pfn_sequence()
566 uint32_t num_pages, uint32_t desired_tile_stride, in psb_mmu_remove_pages() argument
580 rows = num_pages / desired_tile_stride; in psb_mmu_remove_pages()
582 desired_tile_stride = num_pages; in psb_mmu_remove_pages()
612 psb_mmu_flush_ptes(pd, f_address, num_pages, in psb_mmu_remove_pages()
[all …]
/linux/arch/riscv/include/asm/
H A Dset_memory.h21 int num_pages)) in set_kernel_memory() argument
25 int num_pages = PAGE_ALIGN(end - start) >> PAGE_SHIFT; in set_kernel_memory() local
27 return set_memory(start, num_pages); in set_kernel_memory()
37 int num_pages)) in set_kernel_memory() argument
/linux/drivers/gpu/drm/xe/
H A Dxe_shrinker.c142 unsigned long num_pages; in xe_shrinker_count() local
145 num_pages = ttm_backup_bytes_avail() >> PAGE_SHIFT; in xe_shrinker_count()
149 num_pages = min_t(unsigned long, num_pages, shrinker->shrinkable_pages); in xe_shrinker_count()
151 num_pages = 0; in xe_shrinker_count()
153 num_pages += shrinker->purgeable_pages; in xe_shrinker_count()
156 return num_pages ? num_pages : SHRINK_EMPTY; in xe_shrinker_count()
/linux/drivers/gpu/drm/imagination/
H A Dpvr_free_list.c223 struct sg_table *sgt, u32 offset, u32 num_pages) in pvr_free_list_insert_pages_locked() argument
250 num_pages--; in pvr_free_list_insert_pages_locked()
251 if (!num_pages) in pvr_free_list_insert_pages_locked()
255 if (!num_pages) in pvr_free_list_insert_pages_locked()
280 free_list_node->num_pages; in pvr_free_list_insert_node_locked()
289 offset, free_list_node->num_pages); in pvr_free_list_insert_node_locked()
291 free_list->current_pages += free_list_node->num_pages; in pvr_free_list_insert_node_locked()
297 pvr_free_list_grow(struct pvr_free_list *free_list, u32 num_pages) in pvr_free_list_grow() argument
305 if (num_pages & FREE_LIST_ALIGNMENT) { in pvr_free_list_grow()
316 free_list_node->num_pages = num_pages; in pvr_free_list_grow()
[all …]
/linux/drivers/net/ethernet/huawei/hinic3/
H A Dhinic3_queue_common.c18 qpages->num_pages = max(q_depth / elem_per_page, 1); in hinic3_queue_pages_init()
38 __queue_pages_free(hwdev, qpages, qpages->num_pages); in hinic3_queue_pages_free()
47 qpages->pages = kzalloc_objs(qpages->pages[0], qpages->num_pages); in hinic3_queue_pages_alloc()
54 for (pg_idx = 0; pg_idx < qpages->num_pages; pg_idx++) { in hinic3_queue_pages_alloc()
/linux/drivers/gpu/drm/radeon/
H A Dradeon_ttm.c141 unsigned num_pages; in radeon_move_blit() local
178 num_pages = PFN_UP(new_mem->size) * (PAGE_SIZE / RADEON_GPU_PAGE_SIZE); in radeon_move_blit()
179 fence = radeon_copy(rdev, old_start, new_start, num_pages, bo->base.resv); in radeon_move_blit()
340 unsigned long end = gtt->userptr + (u64)ttm->num_pages * PAGE_SIZE; in radeon_ttm_tt_pin_userptr()
348 unsigned num_pages = ttm->num_pages - pinned; in radeon_ttm_tt_pin_userptr() local
352 r = get_user_pages(userptr, num_pages, write ? FOLL_WRITE : 0, in radeon_ttm_tt_pin_userptr()
359 } while (pinned < ttm->num_pages); in radeon_ttm_tt_pin_userptr()
361 r = sg_alloc_table_from_pages(ttm->sg, ttm->pages, ttm->num_pages, 0, in radeon_ttm_tt_pin_userptr()
362 (u64)ttm->num_pages << PAGE_SHIFT, in radeon_ttm_tt_pin_userptr()
372 ttm->num_pages); in radeon_ttm_tt_pin_userptr()
[all …]
/linux/drivers/gpu/drm/
H A Ddrm_cache.c67 unsigned long num_pages) in drm_cache_flush_clflush() argument
72 for (i = 0; i < num_pages; i++) in drm_cache_flush_clflush()
87 drm_clflush_pages(struct page *pages[], unsigned long num_pages) in drm_clflush_pages() argument
92 drm_cache_flush_clflush(pages, num_pages); in drm_clflush_pages()
101 for (i = 0; i < num_pages; i++) { in drm_clflush_pages()
/linux/drivers/gpu/drm/ttm/tests/
H A Dttm_tt_test.c47 int num_pages = params->size >> PAGE_SHIFT; in ttm_tt_init_basic() local
58 KUNIT_ASSERT_EQ(test, tt->num_pages, num_pages + extra_pages); in ttm_tt_init_basic()
73 int num_pages = (size + SZ_4K) >> PAGE_SHIFT; in ttm_tt_init_misaligned() local
87 KUNIT_ASSERT_EQ(test, tt->num_pages, num_pages); in ttm_tt_init_misaligned()
344 int err, num_pages; in ttm_tt_swapin_basic() local
358 num_pages = ttm_tt_swapout(devs->ttm_dev, tt, GFP_KERNEL); in ttm_tt_swapin_basic()
359 KUNIT_ASSERT_EQ(test, num_pages, expected_num_pages); in ttm_tt_swapin_basic()

123456789