Home
last modified time | relevance | path

Searched refs:nr_pages (Results 1 – 25 of 290) sorted by relevance

12345678910>>...12

/linux/io_uring/
H A Dmemmap.c18 static bool io_mem_alloc_compound(struct page **pages, int nr_pages, in io_mem_alloc_compound() argument
34 for (i = 0; i < nr_pages; i++) in io_mem_alloc_compound()
42 unsigned long start, end, nr_pages; in io_pin_pages() local
53 nr_pages = end - start; in io_pin_pages()
54 if (WARN_ON_ONCE(!nr_pages)) in io_pin_pages()
56 if (WARN_ON_ONCE(nr_pages > INT_MAX)) in io_pin_pages()
59 pages = kvmalloc_objs(struct page *, nr_pages, GFP_KERNEL_ACCOUNT); in io_pin_pages()
63 ret = pin_user_pages_fast(uaddr, nr_pages, FOLL_WRITE | FOLL_LONGTERM, in io_pin_pages()
66 if (ret == nr_pages) { in io_pin_pages()
67 *npages = nr_pages; in io_pin_pages()
[all …]
/linux/mm/
H A Dmemory_hotplug.c60 unsigned long nr_pages = PFN_UP(memory_block_memmap_size()); in memory_block_memmap_on_memory_pages() local
69 return pageblock_align(nr_pages); in memory_block_memmap_on_memory_pages()
70 return nr_pages; in memory_block_memmap_on_memory_pages()
319 static int check_pfn_span(unsigned long pfn, unsigned long nr_pages) in check_pfn_span() argument
328 if (!IS_ALIGNED(pfn | nr_pages, PAGES_PER_SUBSECTION)) in check_pfn_span()
381 int __add_pages(int nid, unsigned long pfn, unsigned long nr_pages, in __add_pages() argument
384 const unsigned long end_pfn = pfn + nr_pages; in __add_pages()
392 VM_BUG_ON(!mhp_range_allowed(PFN_PHYS(pfn), nr_pages * PAGE_SIZE, false)); in __add_pages()
399 || vmem_altmap_offset(altmap) > nr_pages) { in __add_pages()
406 if (check_pfn_span(pfn, nr_pages)) { in __add_pages()
[all …]
H A Dpercpu-km.c55 const int nr_pages = pcpu_group_sizes[0] >> PAGE_SHIFT; in pcpu_create_chunk() local
65 pages = alloc_pages(gfp, order_base_2(nr_pages)); in pcpu_create_chunk()
71 for (i = 0; i < nr_pages; i++) in pcpu_create_chunk()
78 pcpu_chunk_populated(chunk, 0, nr_pages); in pcpu_create_chunk()
89 const int nr_pages = pcpu_group_sizes[0] >> PAGE_SHIFT; in pcpu_destroy_chunk() local
98 __free_pages(chunk->data, order_base_2(nr_pages)); in pcpu_destroy_chunk()
109 size_t nr_pages, alloc_pages; in pcpu_verify_alloc_info() local
117 nr_pages = (ai->groups[0].nr_units * ai->unit_size) >> PAGE_SHIFT; in pcpu_verify_alloc_info()
118 alloc_pages = roundup_pow_of_two(nr_pages); in pcpu_verify_alloc_info()
120 if (alloc_pages > nr_pages) in pcpu_verify_alloc_info()
[all …]
H A Dhugetlb_cgroup.c189 unsigned int nr_pages; in hugetlb_cgroup_move_parent() local
203 nr_pages = folio_nr_pages(folio); in hugetlb_cgroup_move_parent()
207 page_counter_charge(&parent->hugepage[idx], nr_pages); in hugetlb_cgroup_move_parent()
211 page_counter_cancel(counter, nr_pages); in hugetlb_cgroup_move_parent()
253 static int __hugetlb_cgroup_charge_cgroup(int idx, unsigned long nr_pages, in __hugetlb_cgroup_charge_cgroup() argument
274 nr_pages, &counter)) { in __hugetlb_cgroup_charge_cgroup()
290 int hugetlb_cgroup_charge_cgroup(int idx, unsigned long nr_pages, in hugetlb_cgroup_charge_cgroup() argument
293 return __hugetlb_cgroup_charge_cgroup(idx, nr_pages, ptr, false); in hugetlb_cgroup_charge_cgroup()
296 int hugetlb_cgroup_charge_cgroup_rsvd(int idx, unsigned long nr_pages, in hugetlb_cgroup_charge_cgroup_rsvd() argument
299 return __hugetlb_cgroup_charge_cgroup(idx, nr_pages, ptr, true); in hugetlb_cgroup_charge_cgroup_rsvd()
[all …]
H A Dcma.h84 void cma_sysfs_account_success_pages(struct cma *cma, unsigned long nr_pages);
85 void cma_sysfs_account_fail_pages(struct cma *cma, unsigned long nr_pages);
86 void cma_sysfs_account_release_pages(struct cma *cma, unsigned long nr_pages);
89 unsigned long nr_pages) {}; in cma_sysfs_account_success_pages() argument
91 unsigned long nr_pages) {}; in cma_sysfs_account_fail_pages() argument
93 unsigned long nr_pages) {}; in cma_sysfs_account_release_pages() argument
H A Drmap.c934 if (pvmw.pte && ptes != pvmw.nr_pages) in folio_referenced_one()
1002 if (ptes == pvmw.nr_pages) { in folio_referenced_one()
1228 unsigned long nr_pages; member
1237 .nr_pages = state->nr_pages, in mapping_wrprotect_range_one()
1250 pgoff_t pgoff_start, unsigned long nr_pages,
1274 unsigned long pfn, unsigned long nr_pages) in mapping_wrprotect_range() argument
1280 .nr_pages = nr_pages, in mapping_wrprotect_range()
1291 __rmap_walk_file(/* folio = */NULL, mapping, pgoff, nr_pages, &rwc, in mapping_wrprotect_range()
1310 int pfn_mkclean_range(unsigned long pfn, unsigned long nr_pages, pgoff_t pgoff, in pfn_mkclean_range() argument
1315 .nr_pages = nr_pages, in pfn_mkclean_range()
[all …]
H A Dmemcontrol.c138 static void memcg_uncharge(struct mem_cgroup *memcg, unsigned int nr_pages);
144 unsigned int nr_pages; in obj_cgroup_release() local
169 nr_pages = nr_bytes >> PAGE_SHIFT; in obj_cgroup_release()
171 if (nr_pages) { in obj_cgroup_release()
175 mod_memcg_state(memcg, MEMCG_KMEM, -nr_pages); in obj_cgroup_release()
176 memcg1_account_kmem(memcg, -nr_pages); in obj_cgroup_release()
178 memcg_uncharge(memcg, nr_pages); in obj_cgroup_release()
1484 int zid, long nr_pages) in mem_cgroup_update_lru_size() argument
1496 if (nr_pages < 0) in mem_cgroup_update_lru_size()
1497 *lru_size += nr_pages; in mem_cgroup_update_lru_size()
[all …]
H A Dmigrate.c1640 int rc, nr_pages; in migrate_hugetlbs() local
1650 nr_pages = folio_nr_pages(folio); in migrate_hugetlbs()
1663 stats->nr_failed_pages += nr_pages; in migrate_hugetlbs()
1685 stats->nr_failed_pages += nr_pages + nr_retry_pages; in migrate_hugetlbs()
1689 nr_retry_pages += nr_pages; in migrate_hugetlbs()
1692 stats->nr_succeeded += nr_pages; in migrate_hugetlbs()
1702 stats->nr_failed_pages += nr_pages; in migrate_hugetlbs()
1729 int nr_pages; in migrate_folios_move() local
1736 nr_pages = folio_nr_pages(folio); in migrate_folios_move()
1753 *nr_retry_pages += nr_pages; in migrate_folios_move()
[all …]
/linux/include/linux/
H A Dhugetlb_cgroup.h127 extern int hugetlb_cgroup_charge_cgroup(int idx, unsigned long nr_pages,
129 extern int hugetlb_cgroup_charge_cgroup_rsvd(int idx, unsigned long nr_pages,
131 extern void hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages,
134 extern void hugetlb_cgroup_commit_charge_rsvd(int idx, unsigned long nr_pages,
137 extern void hugetlb_cgroup_uncharge_folio(int idx, unsigned long nr_pages,
139 extern void hugetlb_cgroup_uncharge_folio_rsvd(int idx, unsigned long nr_pages,
142 extern void hugetlb_cgroup_uncharge_cgroup(int idx, unsigned long nr_pages,
144 extern void hugetlb_cgroup_uncharge_cgroup_rsvd(int idx, unsigned long nr_pages,
152 unsigned long nr_pages,
162 unsigned long nr_pages, in hugetlb_cgroup_uncharge_file_region() argument
[all …]
/linux/fs/iomap/
H A Dswapfile.c16 unsigned long nr_pages; /* number of pages collected */ member
30 unsigned long nr_pages; in iomap_swapfile_add_extent() local
37 if (unlikely(isi->nr_pages >= isi->sis->max)) in iomap_swapfile_add_extent()
39 max_pages = isi->sis->max - isi->nr_pages; in iomap_swapfile_add_extent()
52 nr_pages = next_ppage - first_ppage; in iomap_swapfile_add_extent()
53 nr_pages = min(nr_pages, max_pages); in iomap_swapfile_add_extent()
69 error = add_swap_extent(isi->sis, isi->nr_pages, nr_pages, first_ppage); in iomap_swapfile_add_extent()
73 isi->nr_pages += nr_pages; in iomap_swapfile_add_extent()
182 if (isi.nr_pages == 0) { in iomap_swapfile_activate()
188 sis->max = isi.nr_pages; in iomap_swapfile_activate()
[all …]
/linux/drivers/xen/
H A Dballoon.c390 static enum bp_state increase_reservation(unsigned long nr_pages) in increase_reservation() argument
396 if (nr_pages > ARRAY_SIZE(frame_list)) in increase_reservation()
397 nr_pages = ARRAY_SIZE(frame_list); in increase_reservation()
400 for (i = 0; i < nr_pages; i++) { in increase_reservation()
402 nr_pages = i; in increase_reservation()
410 rc = xenmem_reservation_increase(nr_pages, frame_list); in increase_reservation()
433 static enum bp_state decrease_reservation(unsigned long nr_pages, gfp_t gfp) in decrease_reservation() argument
441 if (nr_pages > ARRAY_SIZE(frame_list)) in decrease_reservation()
442 nr_pages = ARRAY_SIZE(frame_list); in decrease_reservation()
444 for (i = 0; i < nr_pages; i++) { in decrease_reservation()
[all …]
H A Dunpopulated-alloc.c37 static int fill_list(unsigned int nr_pages) in fill_list() argument
42 unsigned int i, alloc_pages = round_up(nr_pages, PAGES_PER_SECTION); in fill_list()
161 int xen_alloc_unpopulated_pages(unsigned int nr_pages, struct page **pages) in xen_alloc_unpopulated_pages() argument
172 return xen_alloc_ballooned_pages(nr_pages, pages); in xen_alloc_unpopulated_pages()
175 if (list_count < nr_pages) { in xen_alloc_unpopulated_pages()
176 ret = fill_list(nr_pages - list_count); in xen_alloc_unpopulated_pages()
181 for (i = 0; i < nr_pages; i++) { in xen_alloc_unpopulated_pages()
217 void xen_free_unpopulated_pages(unsigned int nr_pages, struct page **pages) in xen_free_unpopulated_pages() argument
222 xen_free_ballooned_pages(nr_pages, pages); in xen_free_unpopulated_pages()
227 for (i = 0; i < nr_pages; i++) { in xen_free_unpopulated_pages()
/linux/arch/arm64/kvm/hyp/nvhe/
H A Dsetup.c38 unsigned long nr_pages; in divide_memory_pool() local
42 nr_pages = pkvm_selftest_pages(); in divide_memory_pool()
43 selftest_base = hyp_early_alloc_contig(nr_pages); in divide_memory_pool()
44 if (nr_pages && !selftest_base) in divide_memory_pool()
47 nr_pages = hyp_vmemmap_pages(sizeof(struct hyp_page)); in divide_memory_pool()
48 vmemmap_base = hyp_early_alloc_contig(nr_pages); in divide_memory_pool()
52 nr_pages = hyp_vm_table_pages(); in divide_memory_pool()
53 vm_table_base = hyp_early_alloc_contig(nr_pages); in divide_memory_pool()
57 nr_pages = hyp_s1_pgtable_pages(); in divide_memory_pool()
58 hyp_pgt_base = hyp_early_alloc_contig(nr_pages); in divide_memory_pool()
[all …]
/linux/drivers/hv/
H A Dmshv_regions.c150 if (page_offset + page_count > region->nr_pages) in mshv_region_process_range()
175 struct mshv_mem_region *mshv_region_create(u64 guest_pfn, u64 nr_pages, in mshv_region_create() argument
180 region = vzalloc(sizeof(*region) + sizeof(struct page *) * nr_pages); in mshv_region_create()
184 region->nr_pages = nr_pages; in mshv_region_create()
219 0, region->nr_pages, in mshv_region_share()
242 0, region->nr_pages, in mshv_region_unshare()
274 0, region->nr_pages); in mshv_region_map()
289 mshv_region_invalidate_pages(region, 0, region->nr_pages); in mshv_region_invalidate()
294 u64 done_count, nr_pages; in mshv_region_pin() local
299 for (done_count = 0; done_count < region->nr_pages; done_count += ret) { in mshv_region_pin()
[all …]
/linux/net/rds/
H A Dinfo.c163 unsigned long nr_pages = 0; in rds_info_getsockopt() local
187 nr_pages = (PAGE_ALIGN(start + len) - (start & PAGE_MASK)) in rds_info_getsockopt()
190 pages = kmalloc_objs(struct page *, nr_pages); in rds_info_getsockopt()
195 ret = pin_user_pages_fast(start, nr_pages, FOLL_WRITE, pages); in rds_info_getsockopt()
196 if (ret != nr_pages) { in rds_info_getsockopt()
198 nr_pages = ret; in rds_info_getsockopt()
200 nr_pages = 0; in rds_info_getsockopt()
205 rdsdebug("len %d nr_pages %lu\n", len, nr_pages); in rds_info_getsockopt()
238 unpin_user_pages(pages, nr_pages); in rds_info_getsockopt()
H A Drdma.c154 static int rds_pin_pages(unsigned long user_addr, unsigned int nr_pages, in rds_pin_pages() argument
163 ret = pin_user_pages_fast(user_addr, nr_pages, gup_flags, pages); in rds_pin_pages()
164 if (ret >= 0 && ret < nr_pages) { in rds_pin_pages()
178 unsigned int nr_pages; in __rds_rdma_map() local
213 nr_pages = rds_pages_in_vec(&args->vec); in __rds_rdma_map()
214 if (nr_pages == 0) { in __rds_rdma_map()
220 * To account for unaligned mr regions, subtract one from nr_pages in __rds_rdma_map()
222 if ((nr_pages - 1) > (RDS_MAX_MSG_SIZE >> PAGE_SHIFT)) { in __rds_rdma_map()
227 rdsdebug("RDS: get_mr addr %llx len %llu nr_pages %u\n", in __rds_rdma_map()
228 args->vec.addr, args->vec.bytes, nr_pages); in __rds_rdma_map()
535 unsigned int nr_pages; rds_rdma_pages() local
563 unsigned int nr_pages; rds_rdma_extra_size() local
615 int nr_pages; rds_cmsg_rdma_args() local
[all...]
/linux/kernel/events/
H A Dring_buffer.c179 if (rb->nr_pages) { in __perf_output_begin()
243 handle->page = (offset >> page_shift) & (rb->nr_pages - 1); in __perf_output_begin()
339 if (!rb->nr_pages) in ring_buffer_init()
680 pgoff_t pgoff, int nr_pages, long watermark, int flags) in rb_alloc_aux() argument
697 if (nr_pages <= 0) in rb_alloc_aux()
708 (unsigned long)nr_pages << (PAGE_SHIFT - 1)); in rb_alloc_aux()
723 max_order = ilog2(nr_pages); in rb_alloc_aux()
731 if (get_order((unsigned long)nr_pages * sizeof(void *)) > MAX_PAGE_ORDER) in rb_alloc_aux()
733 rb->aux_pages = kcalloc_node(nr_pages, sizeof(void *), GFP_KERNEL, in rb_alloc_aux()
739 for (rb->aux_nr_pages = 0; rb->aux_nr_pages < nr_pages;) { in rb_alloc_aux()
[all …]
/linux/mm/kasan/
H A Dshadow.c220 nr_shadow_pages = mem_data->nr_pages >> KASAN_SHADOW_SCALE_SHIFT; in kasan_mem_notifier()
226 if (WARN_ON(mem_data->nr_pages % KASAN_GRANULE_SIZE) || in kasan_mem_notifier()
327 static void ___free_pages_bulk(struct page **pages, int nr_pages) in ___free_pages_bulk() argument
331 for (i = 0; i < nr_pages; i++) { in ___free_pages_bulk()
339 static int ___alloc_pages_bulk(struct page **pages, int nr_pages, gfp_t gfp_mask) in ___alloc_pages_bulk() argument
341 unsigned long nr_populated, nr_total = nr_pages; in ___alloc_pages_bulk()
344 while (nr_pages) { in ___alloc_pages_bulk()
345 nr_populated = alloc_pages_bulk(gfp_mask, nr_pages, pages); in ___alloc_pages_bulk()
347 ___free_pages_bulk(page_array, nr_total - nr_pages); in ___alloc_pages_bulk()
351 nr_pages -= nr_populated; in ___alloc_pages_bulk()
[all …]
/linux/drivers/misc/bcm-vk/
H A Dbcm_vk_sg.c60 dma->nr_pages = last - first + 1; in bcm_vk_dma_alloc()
63 dma->pages = kmalloc_objs(struct page *, dma->nr_pages); in bcm_vk_dma_alloc()
68 data, vkdata->size, dma->nr_pages); in bcm_vk_dma_alloc()
74 dma->nr_pages, in bcm_vk_dma_alloc()
77 if (err != dma->nr_pages) { in bcm_vk_dma_alloc()
78 dma->nr_pages = (err >= 0) ? err : 0; in bcm_vk_dma_alloc()
80 err, dma->nr_pages); in bcm_vk_dma_alloc()
85 dma->sglen = (dma->nr_pages * sizeof(*sgdata)) + in bcm_vk_dma_alloc()
115 for (i = 1; i < dma->nr_pages; i++) { in bcm_vk_dma_alloc()
248 for (i = 0; i < dma->nr_pages; i++) in bcm_vk_dma_free()
/linux/drivers/accel/amdxdna/
H A Damdxdna_ubuf.c19 u64 nr_pages; member
34 ret = sg_alloc_table_from_pages(sg, ubuf->pages, ubuf->nr_pages, 0, in amdxdna_ubuf_map()
35 ubuf->nr_pages << PAGE_SHIFT, GFP_KERNEL); in amdxdna_ubuf_map()
65 unpin_user_pages(ubuf->pages, ubuf->nr_pages); in amdxdna_ubuf_release()
67 atomic64_sub(ubuf->nr_pages, &ubuf->mm->pinned_vm); in amdxdna_ubuf_release()
106 kva = vmap(ubuf->pages, ubuf->nr_pages, VM_MAP, PAGE_KERNEL); in amdxdna_ubuf_vmap()
177 ubuf->nr_pages = exp_info.size >> PAGE_SHIFT; in amdxdna_get_ubuf()
179 new_pinned = atomic64_add_return(ubuf->nr_pages, &ubuf->mm->pinned_vm); in amdxdna_get_ubuf()
187 ubuf->pages = kvmalloc_objs(*ubuf->pages, ubuf->nr_pages); in amdxdna_get_ubuf()
226 atomic64_sub(ubuf->nr_pages, &ubuf->mm->pinned_vm); in amdxdna_get_ubuf()
/linux/drivers/virt/acrn/
H A Dmm.c159 int nr_pages, i, order, nr_regions = 0; in acrn_vm_ram_map() local
172 nr_pages = memmap->len >> PAGE_SHIFT; in acrn_vm_ram_map()
173 if (!nr_pages) in acrn_vm_ram_map()
187 for (i = 0; i < nr_pages; i++) { in acrn_vm_ram_map()
237 pages = vzalloc(array_size(nr_pages, sizeof(*pages))); in acrn_vm_ram_map()
243 nr_pages, FOLL_WRITE | FOLL_LONGTERM, in acrn_vm_ram_map()
248 } else if (pinned != nr_pages) { in acrn_vm_ram_map()
254 remap_vaddr = vmap(pages, nr_pages, VM_MAP, PAGE_KERNEL); in acrn_vm_ram_map()
265 region_mapping->npages = nr_pages; in acrn_vm_ram_map()
280 for (i = 0; i < nr_pages; i += 1 << order) { in acrn_vm_ram_map()
[all …]
/linux/drivers/net/ethernet/ibm/ehea/
H A Dehea_qmr.c93 int i, nr_pages; in hw_queue_dtor() local
100 nr_pages = queue->queue_length / queue->pagesize; in hw_queue_dtor()
102 for (i = 0; i < nr_pages; i += pages_per_kpage) in hw_queue_dtor()
134 ret = hw_queue_ctor(&cq->hw_queue, cq->attr.nr_pages, in ehea_create_cq()
139 for (counter = 0; counter < cq->attr.nr_pages; counter++) { in ehea_create_cq()
152 cq, hret, counter, cq->attr.nr_pages); in ehea_create_cq()
156 if (counter == (cq->attr.nr_pages - 1)) { in ehea_create_cq()
255 ret = hw_queue_ctor(&eq->hw_queue, eq->attr.nr_pages, in ehea_create_eq()
262 for (i = 0; i < eq->attr.nr_pages; i++) { in ehea_create_eq()
276 if (i == (eq->attr.nr_pages - 1)) { in ehea_create_eq()
[all …]
/linux/drivers/firmware/efi/libstub/
H A Dmem.c119 unsigned long nr_pages; in efi_free() local
124 nr_pages = round_up(size, EFI_ALLOC_ALIGN) / EFI_PAGE_SIZE; in efi_free()
125 efi_bs_call(free_pages, addr, nr_pages); in efi_free()
148 unsigned long nr_pages; in efi_low_alloc_above() local
165 nr_pages = size / EFI_PAGE_SIZE; in efi_low_alloc_above()
183 if (desc->num_pages < nr_pages) in efi_low_alloc_above()
197 EFI_LOADER_DATA, nr_pages, &start); in efi_low_alloc_above()
/linux/fs/
H A Dfs-writeback.c39 long nr_pages; member
1018 static long wb_split_bdi_pages(struct bdi_writeback *wb, long nr_pages) in wb_split_bdi_pages() argument
1023 if (nr_pages == LONG_MAX) in wb_split_bdi_pages()
1032 return nr_pages; in wb_split_bdi_pages()
1034 return DIV_ROUND_UP_ULL((u64)nr_pages * this_bw, tot_bw); in wb_split_bdi_pages()
1063 long nr_pages; in bdi_split_work_to_wbs() local
1078 nr_pages = wb_split_bdi_pages(wb, base_work->nr_pages); in bdi_split_work_to_wbs()
1083 work->nr_pages = nr_pages; in bdi_split_work_to_wbs()
1102 work->nr_pages = nr_pages; in bdi_split_work_to_wbs()
1180 work->nr_pages = dirty; in cgroup_writeback_by_id()
[all …]
/linux/arch/arm/mach-rpc/include/mach/
H A Duncompress.h20 unsigned long nr_pages; member
116 unsigned int nr_pages = 0, page_size = PAGE_SIZE; in arch_decomp_setup() local
130 nr_pages += (t->u.mem.size / PAGE_SIZE); in arch_decomp_setup()
134 nr_pages = params->nr_pages; in arch_decomp_setup()
179 if (nr_pages * page_size < 4096*1024) error("<4M of mem\n"); in arch_decomp_setup()

12345678910>>...12