| /linux/io_uring/ |
| H A D | memmap.c | 18 static bool io_mem_alloc_compound(struct page **pages, int nr_pages, in io_mem_alloc_compound() argument 34 for (i = 0; i < nr_pages; i++) in io_mem_alloc_compound() 42 unsigned long start, end, nr_pages; in io_pin_pages() local 53 nr_pages = end - start; in io_pin_pages() 54 if (WARN_ON_ONCE(!nr_pages)) in io_pin_pages() 56 if (WARN_ON_ONCE(nr_pages > INT_MAX)) in io_pin_pages() 59 pages = kvmalloc_objs(struct page *, nr_pages, GFP_KERNEL_ACCOUNT); in io_pin_pages() 63 ret = pin_user_pages_fast(uaddr, nr_pages, FOLL_WRITE | FOLL_LONGTERM, in io_pin_pages() 66 if (ret == nr_pages) { in io_pin_pages() 67 *npages = nr_pages; in io_pin_pages() [all …]
|
| /linux/include/linux/ |
| H A D | hugetlb_cgroup.h | 127 extern int hugetlb_cgroup_charge_cgroup(int idx, unsigned long nr_pages, 129 extern int hugetlb_cgroup_charge_cgroup_rsvd(int idx, unsigned long nr_pages, 131 extern void hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages, 134 extern void hugetlb_cgroup_commit_charge_rsvd(int idx, unsigned long nr_pages, 137 extern void hugetlb_cgroup_uncharge_folio(int idx, unsigned long nr_pages, 139 extern void hugetlb_cgroup_uncharge_folio_rsvd(int idx, unsigned long nr_pages, 142 extern void hugetlb_cgroup_uncharge_cgroup(int idx, unsigned long nr_pages, 144 extern void hugetlb_cgroup_uncharge_cgroup_rsvd(int idx, unsigned long nr_pages, 152 unsigned long nr_pages, 162 unsigned long nr_pages, in hugetlb_cgroup_uncharge_file_region() argument [all …]
|
| H A D | mm_inline.h | 40 long nr_pages) in __update_lru_size() argument 45 WARN_ON_ONCE(nr_pages != (int)nr_pages); in __update_lru_size() 47 mod_lruvec_state(lruvec, NR_LRU_BASE + lru, nr_pages); in __update_lru_size() 49 NR_ZONE_LRU_BASE + lru, nr_pages); in __update_lru_size() 54 long nr_pages) in update_lru_size() argument 56 __update_lru_size(lruvec, lru, zid, nr_pages); in update_lru_size() 58 mem_cgroup_update_lru_size(lruvec, lru, zid, nr_pages); in update_lru_size() 188 WRITE_ONCE(lrugen->nr_pages[old_gen][type][zone], in lru_gen_update_size() 189 lrugen->nr_pages[old_gen][type][zone] - delta); in lru_gen_update_size() 191 WRITE_ONCE(lrugen->nr_pages[new_gen][type][zone], in lru_gen_update_size() [all …]
|
| /linux/mm/ |
| H A D | memory_hotplug.c | 60 unsigned long nr_pages = PFN_UP(memory_block_memmap_size()); in memory_block_memmap_on_memory_pages() local 69 return pageblock_align(nr_pages); in memory_block_memmap_on_memory_pages() 70 return nr_pages; in memory_block_memmap_on_memory_pages() 319 static int check_pfn_span(unsigned long pfn, unsigned long nr_pages) in check_pfn_span() argument 336 if (!IS_ALIGNED(pfn | nr_pages, min_align)) in check_pfn_span() 389 int __add_pages(int nid, unsigned long pfn, unsigned long nr_pages, in __add_pages() argument 392 const unsigned long end_pfn = pfn + nr_pages; in __add_pages() 400 VM_BUG_ON(!mhp_range_allowed(PFN_PHYS(pfn), nr_pages * PAGE_SIZE, false)); in __add_pages() 407 || vmem_altmap_offset(altmap) > nr_pages) { in __add_pages() 414 if (check_pfn_span(pfn, nr_pages)) { in __add_pages() 549 remove_pfn_range_from_zone(struct zone * zone,unsigned long start_pfn,unsigned long nr_pages) remove_pfn_range_from_zone() argument 593 __remove_pages(unsigned long pfn,unsigned long nr_pages,struct vmem_altmap * altmap) __remove_pages() argument 658 online_pages_range(unsigned long start_pfn,unsigned long nr_pages) online_pages_range() argument 704 resize_zone_range(struct zone * zone,unsigned long start_pfn,unsigned long nr_pages) resize_zone_range() argument 715 resize_pgdat_range(struct pglist_data * pgdat,unsigned long start_pfn,unsigned long nr_pages) resize_pgdat_range() argument 749 move_pfn_range_to_zone(struct zone * zone,unsigned long start_pfn,unsigned long nr_pages,struct vmem_altmap * altmap,int migratetype,bool isolate_pageblock) move_pfn_range_to_zone() argument 844 auto_movable_can_online_movable(int nid,struct memory_group * group,unsigned long nr_pages) auto_movable_can_online_movable() argument 899 default_kernel_zone_for_pfn(int nid,unsigned long start_pfn,unsigned long nr_pages) default_kernel_zone_for_pfn() argument 967 auto_movable_zone_for_pfn(int nid,struct memory_group * group,unsigned long pfn,unsigned long nr_pages) auto_movable_zone_for_pfn() argument 1026 default_zone_for_pfn(int nid,unsigned long start_pfn,unsigned long nr_pages) default_zone_for_pfn() argument 1051 zone_for_pfn_range(int online_type,int nid,struct memory_group * group,unsigned long start_pfn,unsigned long nr_pages) zone_for_pfn_range() argument 1070 adjust_present_page_count(struct page * page,struct memory_group * group,long nr_pages) adjust_present_page_count() argument 1090 mhp_init_memmap_on_memory(unsigned long pfn,unsigned long nr_pages,struct zone * zone) mhp_init_memmap_on_memory() argument 1121 mhp_deinit_memmap_on_memory(unsigned long pfn,unsigned long nr_pages) mhp_deinit_memmap_on_memory() argument 1144 online_pages(unsigned long pfn,unsigned long nr_pages,struct zone * zone,struct memory_group * group) online_pages() argument 1883 count_system_ram_pages_cb(unsigned long start_pfn,unsigned long nr_pages,void * data) count_system_ram_pages_cb() argument 1894 offline_pages(unsigned long start_pfn,unsigned long nr_pages,struct zone * zone,struct memory_group * group) offline_pages() argument [all...] |
| H A D | gup_test.c | 12 unsigned long nr_pages, unsigned int gup_test_flags) in put_back_pages() 19 for (i = 0; i < nr_pages; i++) in put_back_pages() 26 unpin_user_pages(pages, nr_pages); in put_back_pages() 30 unpin_user_pages(pages, nr_pages); in put_back_pages() 32 for (i = 0; i < nr_pages; i++) in put_back_pages() 41 unsigned long nr_pages) in verify_dma_pinned() 50 for (i = 0; i < nr_pages; i++) { in verify_dma_pinned() 71 unsigned long nr_pages) in dump_pages_test() 81 if (gup->which_pages[i] > nr_pages) { in dump_pages_test() 105 unsigned long i, nr_pages, add in __gup_test_ioctl() 11 put_back_pages(unsigned int cmd,struct page ** pages,unsigned long nr_pages,unsigned int gup_test_flags) put_back_pages() argument 40 verify_dma_pinned(unsigned int cmd,struct page ** pages,unsigned long nr_pages) verify_dma_pinned() argument 70 dump_pages_test(struct gup_test * gup,struct page ** pages,unsigned long nr_pages) dump_pages_test() argument 104 unsigned long i, nr_pages, addr, next; __gup_test_ioctl() local 224 long nr_pages, cur_pages, addr, remaining_pages; pin_longterm_test_start() local [all...] |
| H A D | percpu-km.c | 55 const int nr_pages = pcpu_group_sizes[0] >> PAGE_SHIFT; in pcpu_create_chunk() local 65 pages = alloc_pages(gfp, order_base_2(nr_pages)); in pcpu_create_chunk() 71 for (i = 0; i < nr_pages; i++) in pcpu_create_chunk() 78 pcpu_chunk_populated(chunk, 0, nr_pages); in pcpu_create_chunk() 89 const int nr_pages = pcpu_group_sizes[0] >> PAGE_SHIFT; in pcpu_destroy_chunk() local 98 __free_pages(chunk->data, order_base_2(nr_pages)); in pcpu_destroy_chunk() 109 size_t nr_pages, alloc_pages; in pcpu_verify_alloc_info() local 117 nr_pages = (ai->groups[0].nr_units * ai->unit_size) >> PAGE_SHIFT; in pcpu_verify_alloc_info() 118 alloc_pages = roundup_pow_of_two(nr_pages); in pcpu_verify_alloc_info() 120 if (alloc_pages > nr_pages) in pcpu_verify_alloc_info() [all …]
|
| H A D | hugetlb_cgroup.c | 189 unsigned int nr_pages; in hugetlb_cgroup_move_parent() local 203 nr_pages = folio_nr_pages(folio); in hugetlb_cgroup_move_parent() 207 page_counter_charge(&parent->hugepage[idx], nr_pages); in hugetlb_cgroup_move_parent() 211 page_counter_cancel(counter, nr_pages); in hugetlb_cgroup_move_parent() 253 static int __hugetlb_cgroup_charge_cgroup(int idx, unsigned long nr_pages, in __hugetlb_cgroup_charge_cgroup() argument 274 nr_pages, &counter)) { in __hugetlb_cgroup_charge_cgroup() 290 int hugetlb_cgroup_charge_cgroup(int idx, unsigned long nr_pages, in hugetlb_cgroup_charge_cgroup() argument 293 return __hugetlb_cgroup_charge_cgroup(idx, nr_pages, ptr, false); in hugetlb_cgroup_charge_cgroup() 296 int hugetlb_cgroup_charge_cgroup_rsvd(int idx, unsigned long nr_pages, in hugetlb_cgroup_charge_cgroup_rsvd() argument 299 return __hugetlb_cgroup_charge_cgroup(idx, nr_pages, ptr, true); in hugetlb_cgroup_charge_cgroup_rsvd() [all …]
|
| H A D | mmu_gather.c | 62 unsigned int nr_pages = 1; in tlb_flush_rmap_batch() local 66 nr_pages = encoded_nr_pages(pages[++i]); in tlb_flush_rmap_batch() 68 folio_remove_rmap_ptes(page_folio(page), page, nr_pages, in tlb_flush_rmap_batch() 106 unsigned int nr, nr_pages; in __tlb_batch_free_encoded_pages() local 127 for (nr = 0, nr_pages = 0; in __tlb_batch_free_encoded_pages() 128 nr < batch->nr && nr_pages < MAX_NR_FOLIOS_PER_FREE; in __tlb_batch_free_encoded_pages() 132 nr_pages += encoded_nr_pages(pages[++nr]); in __tlb_batch_free_encoded_pages() 134 nr_pages++; in __tlb_batch_free_encoded_pages() 167 struct page *page, unsigned int nr_pages, bool delay_rmap, in __tlb_remove_folio_pages_size() argument 177 VM_WARN_ON_ONCE(nr_pages != 1 && page_size != PAGE_SIZE); in __tlb_remove_folio_pages_size() [all …]
|
| H A D | memcontrol.c | 138 static void memcg_uncharge(struct mem_cgroup *memcg, unsigned int nr_pages); 144 unsigned int nr_pages; in obj_cgroup_release() local 169 nr_pages = nr_bytes >> PAGE_SHIFT; in obj_cgroup_release() 171 if (nr_pages) { in obj_cgroup_release() 175 mod_memcg_state(memcg, MEMCG_KMEM, -nr_pages); in obj_cgroup_release() 176 memcg1_account_kmem(memcg, -nr_pages); in obj_cgroup_release() 178 memcg_uncharge(memcg, nr_pages); in obj_cgroup_release() 1291 int zid, int nr_pages) in mem_cgroup_update_lru_size() argument 1303 if (nr_pages < 0) in mem_cgroup_update_lru_size() 1304 *lru_size += nr_pages; in mem_cgroup_update_lru_size() [all …]
|
| H A D | cma.h | 84 void cma_sysfs_account_success_pages(struct cma *cma, unsigned long nr_pages); 85 void cma_sysfs_account_fail_pages(struct cma *cma, unsigned long nr_pages); 86 void cma_sysfs_account_release_pages(struct cma *cma, unsigned long nr_pages); 89 unsigned long nr_pages) {}; in cma_sysfs_account_success_pages() argument 91 unsigned long nr_pages) {}; in cma_sysfs_account_fail_pages() argument 93 unsigned long nr_pages) {}; in cma_sysfs_account_release_pages() argument
|
| H A D | rmap.c | 934 if (pvmw.pte && ptes != pvmw.nr_pages) in folio_referenced_one() 1001 if (ptes == pvmw.nr_pages) { in folio_referenced_one() 1222 unsigned long nr_pages; member 1231 .nr_pages = state->nr_pages, in mapping_wrprotect_range_one() 1244 pgoff_t pgoff_start, unsigned long nr_pages, 1268 unsigned long pfn, unsigned long nr_pages) in mapping_wrprotect_range() argument 1274 .nr_pages = nr_pages, in mapping_wrprotect_range() 1285 __rmap_walk_file(/* folio = */NULL, mapping, pgoff, nr_pages, &rwc, in mapping_wrprotect_range() 1304 int pfn_mkclean_range(unsigned long pfn, unsigned long nr_pages, pgoff_t pgoff, in pfn_mkclean_range() argument 1309 .nr_pages = nr_pages, in pfn_mkclean_range() [all …]
|
| /linux/fs/iomap/ |
| H A D | swapfile.c | 16 unsigned long nr_pages; /* number of pages collected */ member 30 unsigned long nr_pages; in iomap_swapfile_add_extent() local 37 if (unlikely(isi->nr_pages >= isi->sis->max)) in iomap_swapfile_add_extent() 39 max_pages = isi->sis->max - isi->nr_pages; in iomap_swapfile_add_extent() 52 nr_pages = next_ppage - first_ppage; in iomap_swapfile_add_extent() 53 nr_pages = min(nr_pages, max_pages); in iomap_swapfile_add_extent() 69 error = add_swap_extent(isi->sis, isi->nr_pages, nr_pages, first_ppage); in iomap_swapfile_add_extent() 73 isi->nr_pages += nr_pages; in iomap_swapfile_add_extent() 182 if (isi.nr_pages == 0) { in iomap_swapfile_activate() 188 sis->max = isi.nr_pages; in iomap_swapfile_activate() [all …]
|
| /linux/drivers/xen/ |
| H A D | balloon.c | 390 static enum bp_state increase_reservation(unsigned long nr_pages) in increase_reservation() argument 396 if (nr_pages > ARRAY_SIZE(frame_list)) in increase_reservation() 397 nr_pages = ARRAY_SIZE(frame_list); in increase_reservation() 400 for (i = 0; i < nr_pages; i++) { in increase_reservation() 402 nr_pages = i; in increase_reservation() 410 rc = xenmem_reservation_increase(nr_pages, frame_list); in increase_reservation() 433 static enum bp_state decrease_reservation(unsigned long nr_pages, gfp_t gfp) in decrease_reservation() argument 441 if (nr_pages > ARRAY_SIZE(frame_list)) in decrease_reservation() 442 nr_pages = ARRAY_SIZE(frame_list); in decrease_reservation() 444 for (i = 0; i < nr_pages; i++) { in decrease_reservation() [all …]
|
| H A D | unpopulated-alloc.c | 37 static int fill_list(unsigned int nr_pages) in fill_list() argument 42 unsigned int i, alloc_pages = round_up(nr_pages, PAGES_PER_SECTION); in fill_list() 161 int xen_alloc_unpopulated_pages(unsigned int nr_pages, struct page **pages) in xen_alloc_unpopulated_pages() argument 172 return xen_alloc_ballooned_pages(nr_pages, pages); in xen_alloc_unpopulated_pages() 175 if (list_count < nr_pages) { in xen_alloc_unpopulated_pages() 176 ret = fill_list(nr_pages - list_count); in xen_alloc_unpopulated_pages() 181 for (i = 0; i < nr_pages; i++) { in xen_alloc_unpopulated_pages() 217 void xen_free_unpopulated_pages(unsigned int nr_pages, struct page **pages) in xen_free_unpopulated_pages() argument 222 xen_free_ballooned_pages(nr_pages, pages); in xen_free_unpopulated_pages() 227 for (i = 0; i < nr_pages; i++) { in xen_free_unpopulated_pages()
|
| /linux/drivers/hv/ |
| H A D | mshv_regions.c | 150 if (page_offset + page_count > region->nr_pages) in mshv_region_process_range() 175 struct mshv_mem_region *mshv_region_create(u64 guest_pfn, u64 nr_pages, in mshv_region_create() argument 180 region = vzalloc(sizeof(*region) + sizeof(struct page *) * nr_pages); in mshv_region_create() 184 region->nr_pages = nr_pages; in mshv_region_create() 219 0, region->nr_pages, in mshv_region_share() 242 0, region->nr_pages, in mshv_region_unshare() 274 0, region->nr_pages); in mshv_region_map() 289 mshv_region_invalidate_pages(region, 0, region->nr_pages); in mshv_region_invalidate() 294 u64 done_count, nr_pages; in mshv_region_pin() local 299 for (done_count = 0; done_count < region->nr_pages; done_count += ret) { in mshv_region_pin() [all …]
|
| /linux/drivers/net/ethernet/broadcom/bnge/ |
| H A D | bnge_rmem.c | 44 for (i = 0; i < rmem->nr_pages; i++) { in bnge_free_ring() 55 size_t pg_tbl_size = rmem->nr_pages * 8; in bnge_free_ring() 78 if ((rmem->nr_pages > 1 || rmem->depth > 0) && !rmem->pg_tbl) { in bnge_alloc_ring() 79 size_t pg_tbl_size = rmem->nr_pages * 8; in bnge_alloc_ring() 90 for (i = 0; i < rmem->nr_pages; i++) { in bnge_alloc_ring() 104 if (rmem->nr_pages > 1 || rmem->depth > 0) { in bnge_alloc_ring() 105 if (i == rmem->nr_pages - 2 && in bnge_alloc_ring() 108 else if (i == rmem->nr_pages - 1 && in bnge_alloc_ring() 152 ctx_pg->nr_pages = DIV_ROUND_UP(mem_size, BNGE_PAGE_SIZE); in bnge_alloc_ctx_pg_tbls() 153 if (ctx_pg->nr_pages > MAX_CTX_TOTAL_PAGES) { in bnge_alloc_ctx_pg_tbls() [all …]
|
| /linux/arch/arm64/kvm/hyp/include/nvhe/ |
| H A D | mem_protect.h | 38 int __pkvm_host_donate_hyp(u64 pfn, u64 nr_pages); 39 int __pkvm_hyp_donate_host(u64 pfn, u64 nr_pages); 40 int __pkvm_host_share_ffa(u64 pfn, u64 nr_pages); 41 int __pkvm_host_unshare_ffa(u64 pfn, u64 nr_pages); 42 int __pkvm_host_share_guest(u64 pfn, u64 gfn, u64 nr_pages, struct pkvm_hyp_vcpu *vcpu, 44 int __pkvm_host_unshare_guest(u64 gfn, u64 nr_pages, struct pkvm_hyp_vm *hyp_vm); 46 int __pkvm_host_wrprotect_guest(u64 gfn, u64 nr_pages, struct pkvm_hyp_vm *hyp_vm); 47 int __pkvm_host_test_clear_young_guest(u64 gfn, u64 nr_pages, bool mkold, struct pkvm_hyp_vm *vm);
|
| /linux/net/rds/ |
| H A D | info.c | 163 unsigned long nr_pages = 0; in rds_info_getsockopt() local 187 nr_pages = (PAGE_ALIGN(start + len) - (start & PAGE_MASK)) in rds_info_getsockopt() 190 pages = kmalloc_objs(struct page *, nr_pages); in rds_info_getsockopt() 195 ret = pin_user_pages_fast(start, nr_pages, FOLL_WRITE, pages); in rds_info_getsockopt() 196 if (ret != nr_pages) { in rds_info_getsockopt() 198 nr_pages = ret; in rds_info_getsockopt() 200 nr_pages = 0; in rds_info_getsockopt() 205 rdsdebug("len %d nr_pages %lu\n", len, nr_pages); in rds_info_getsockopt() 238 unpin_user_pages(pages, nr_pages); in rds_info_getsockopt()
|
| H A D | rdma.c | 154 static int rds_pin_pages(unsigned long user_addr, unsigned int nr_pages, in rds_pin_pages() argument 163 ret = pin_user_pages_fast(user_addr, nr_pages, gup_flags, pages); in rds_pin_pages() 164 if (ret >= 0 && ret < nr_pages) { in rds_pin_pages() 178 unsigned int nr_pages; in __rds_rdma_map() local 213 nr_pages = rds_pages_in_vec(&args->vec); in __rds_rdma_map() 214 if (nr_pages == 0) { in __rds_rdma_map() 222 if ((nr_pages - 1) > (RDS_MAX_MSG_SIZE >> PAGE_SHIFT)) { in __rds_rdma_map() 228 args->vec.addr, args->vec.bytes, nr_pages); in __rds_rdma_map() 231 pages = kzalloc_objs(struct page *, nr_pages); in __rds_rdma_map() 265 ret = rds_pin_pages(args->vec.addr, nr_pages, pages, 1); in __rds_rdma_map() [all …]
|
| /linux/kernel/events/ |
| H A D | ring_buffer.c | 179 if (rb->nr_pages) { in __perf_output_begin() 243 handle->page = (offset >> page_shift) & (rb->nr_pages - 1); in __perf_output_begin() 339 if (!rb->nr_pages) in ring_buffer_init() 678 pgoff_t pgoff, int nr_pages, long watermark, int flags) in rb_alloc_aux() argument 695 if (nr_pages <= 0) in rb_alloc_aux() 706 (unsigned long)nr_pages << (PAGE_SHIFT - 1)); in rb_alloc_aux() 721 max_order = ilog2(nr_pages); in rb_alloc_aux() 729 if (get_order((unsigned long)nr_pages * sizeof(void *)) > MAX_PAGE_ORDER) in rb_alloc_aux() 731 rb->aux_pages = kcalloc_node(nr_pages, sizeof(void *), GFP_KERNEL, in rb_alloc_aux() 737 for (rb->aux_nr_pages = 0; rb->aux_nr_pages < nr_pages;) { in rb_alloc_aux() [all …]
|
| /linux/include/xen/ |
| H A D | xen.h | 73 int xen_alloc_unpopulated_pages(unsigned int nr_pages, struct page **pages); 74 void xen_free_unpopulated_pages(unsigned int nr_pages, struct page **pages); 80 static inline int xen_alloc_unpopulated_pages(unsigned int nr_pages, in xen_alloc_unpopulated_pages() 83 return xen_alloc_ballooned_pages(nr_pages, pages); in xen_free_unpopulated_pages() argument 85 static inline void xen_free_unpopulated_pages(unsigned int nr_pages, in xen_free_unpopulated_pages() 88 xen_free_ballooned_pages(nr_pages, pages); 78 xen_alloc_unpopulated_pages(unsigned int nr_pages,struct page ** pages) xen_alloc_unpopulated_pages() argument
|
| /linux/drivers/firmware/efi/libstub/ |
| H A D | relocate.c | 28 unsigned long nr_pages; in efi_low_alloc_above() local 45 nr_pages = size / EFI_PAGE_SIZE; in efi_low_alloc_above() 63 if (desc->num_pages < nr_pages) in efi_low_alloc_above() 77 EFI_LOADER_DATA, nr_pages, &start); in efi_low_alloc_above() 121 unsigned long nr_pages; in efi_relocate_kernel() local 139 nr_pages = round_up(alloc_size, EFI_ALLOC_ALIGN) / EFI_PAGE_SIZE; in efi_relocate_kernel() 141 EFI_LOADER_DATA, nr_pages, &efi_addr); in efi_relocate_kernel()
|
| /linux/mm/kasan/ |
| H A D | shadow.c | 220 nr_shadow_pages = mem_data->nr_pages >> KASAN_SHADOW_SCALE_SHIFT; in kasan_mem_notifier() 226 if (WARN_ON(mem_data->nr_pages % KASAN_GRANULE_SIZE) || in kasan_mem_notifier() 327 static void ___free_pages_bulk(struct page **pages, int nr_pages) in ___free_pages_bulk() argument 331 for (i = 0; i < nr_pages; i++) { in ___free_pages_bulk() 339 static int ___alloc_pages_bulk(struct page **pages, int nr_pages, gfp_t gfp_mask) in ___alloc_pages_bulk() argument 341 unsigned long nr_populated, nr_total = nr_pages; in ___alloc_pages_bulk() 344 while (nr_pages) { in ___alloc_pages_bulk() 345 nr_populated = alloc_pages_bulk(gfp_mask, nr_pages, pages); in ___alloc_pages_bulk() 347 ___free_pages_bulk(page_array, nr_total - nr_pages); in ___alloc_pages_bulk() 351 nr_pages -= nr_populated; in ___alloc_pages_bulk() [all …]
|
| /linux/drivers/misc/bcm-vk/ |
| H A D | bcm_vk_sg.c | 60 dma->nr_pages = last - first + 1; in bcm_vk_dma_alloc() 63 dma->pages = kmalloc_objs(struct page *, dma->nr_pages); in bcm_vk_dma_alloc() 68 data, vkdata->size, dma->nr_pages); in bcm_vk_dma_alloc() 74 dma->nr_pages, in bcm_vk_dma_alloc() 77 if (err != dma->nr_pages) { in bcm_vk_dma_alloc() 78 dma->nr_pages = (err >= 0) ? err : 0; in bcm_vk_dma_alloc() 80 err, dma->nr_pages); in bcm_vk_dma_alloc() 85 dma->sglen = (dma->nr_pages * sizeof(*sgdata)) + in bcm_vk_dma_alloc() 115 for (i = 1; i < dma->nr_pages; i++) { in bcm_vk_dma_alloc() 248 for (i = 0; i < dma->nr_pages; i++) in bcm_vk_dma_free()
|
| /linux/drivers/accel/amdxdna/ |
| H A D | amdxdna_ubuf.c | 19 u64 nr_pages; member 35 ret = sg_alloc_table_from_pages(sg, ubuf->pages, ubuf->nr_pages, 0, in amdxdna_ubuf_map() 36 ubuf->nr_pages << PAGE_SHIFT, GFP_KERNEL); in amdxdna_ubuf_map() 72 unpin_user_pages(ubuf->pages, ubuf->nr_pages); in amdxdna_ubuf_release() 74 atomic64_sub(ubuf->nr_pages, &ubuf->mm->pinned_vm); in amdxdna_ubuf_release() 113 kva = vmap(ubuf->pages, ubuf->nr_pages, VM_MAP, PAGE_KERNEL); in amdxdna_ubuf_vmap() 186 ubuf->nr_pages = exp_info.size >> PAGE_SHIFT; in amdxdna_get_ubuf() 188 new_pinned = atomic64_add_return(ubuf->nr_pages, &ubuf->mm->pinned_vm); in amdxdna_get_ubuf() 196 ubuf->pages = kvmalloc_objs(*ubuf->pages, ubuf->nr_pages); in amdxdna_get_ubuf() 235 atomic64_sub(ubuf->nr_pages, &ubuf->mm->pinned_vm); in amdxdna_get_ubuf()
|