Home
last modified time | relevance | path

Searched refs:nr_pages (Results 1 – 25 of 307) sorted by relevance

12345678910>>...13

/linux/mm/
H A Dmemory_hotplug.c59 unsigned long nr_pages = PFN_UP(memory_block_memmap_size()); in memory_block_memmap_on_memory_pages() local
68 return pageblock_align(nr_pages); in memory_block_memmap_on_memory_pages()
69 return nr_pages; in memory_block_memmap_on_memory_pages()
318 static int check_pfn_span(unsigned long pfn, unsigned long nr_pages) in check_pfn_span() argument
335 if (!IS_ALIGNED(pfn | nr_pages, min_align)) in check_pfn_span()
388 int __add_pages(int nid, unsigned long pfn, unsigned long nr_pages, in __add_pages() argument
391 const unsigned long end_pfn = pfn + nr_pages; in __add_pages()
399 VM_BUG_ON(!mhp_range_allowed(PFN_PHYS(pfn), nr_pages * PAGE_SIZE, false)); in __add_pages()
406 || vmem_altmap_offset(altmap) > nr_pages) { in __add_pages()
413 if (check_pfn_span(pfn, nr_pages)) { in __add_pages()
[all …]
H A Dpage_counter.c54 void page_counter_cancel(struct page_counter *counter, unsigned long nr_pages) in page_counter_cancel() argument
58 new = atomic_long_sub_return(nr_pages, &counter->usage); in page_counter_cancel()
61 new, nr_pages)) { in page_counter_cancel()
76 void page_counter_charge(struct page_counter *counter, unsigned long nr_pages) in page_counter_charge() argument
84 new = atomic_long_add_return(nr_pages, &c->usage); in page_counter_charge()
119 unsigned long nr_pages, in page_counter_try_charge() argument
141 new = atomic_long_add_return(nr_pages, &c->usage); in page_counter_try_charge()
143 atomic_long_sub(nr_pages, &c->usage); in page_counter_try_charge()
167 page_counter_cancel(c, nr_pages); in page_counter_try_charge()
177 void page_counter_uncharge(struct page_counter *counter, unsigned long nr_pages) in page_counter_uncharge() argument
[all …]
H A Dpercpu-km.c55 const int nr_pages = pcpu_group_sizes[0] >> PAGE_SHIFT; in pcpu_create_chunk() local
65 pages = alloc_pages(gfp, order_base_2(nr_pages)); in pcpu_create_chunk()
71 for (i = 0; i < nr_pages; i++) in pcpu_create_chunk()
78 pcpu_chunk_populated(chunk, 0, nr_pages); in pcpu_create_chunk()
89 const int nr_pages = pcpu_group_sizes[0] >> PAGE_SHIFT; in pcpu_destroy_chunk() local
98 __free_pages(chunk->data, order_base_2(nr_pages)); in pcpu_destroy_chunk()
109 size_t nr_pages, alloc_pages; in pcpu_verify_alloc_info() local
117 nr_pages = (ai->groups[0].nr_units * ai->unit_size) >> PAGE_SHIFT; in pcpu_verify_alloc_info()
118 alloc_pages = roundup_pow_of_two(nr_pages); in pcpu_verify_alloc_info()
120 if (alloc_pages > nr_pages) in pcpu_verify_alloc_info()
[all …]
H A Dgup_test.c11 unsigned long nr_pages, unsigned int gup_test_flags) in put_back_pages() argument
18 for (i = 0; i < nr_pages; i++) in put_back_pages()
25 unpin_user_pages(pages, nr_pages); in put_back_pages()
29 unpin_user_pages(pages, nr_pages); in put_back_pages()
31 for (i = 0; i < nr_pages; i++) in put_back_pages()
40 unsigned long nr_pages) in verify_dma_pinned() argument
49 for (i = 0; i < nr_pages; i++) { in verify_dma_pinned()
70 unsigned long nr_pages) in dump_pages_test() argument
80 if (gup->which_pages[i] > nr_pages) { in dump_pages_test()
104 unsigned long i, nr_pages, addr, next; in __gup_test_ioctl() local
[all …]
H A Dhugetlb_cgroup.c200 unsigned int nr_pages; in hugetlb_cgroup_move_parent() local
214 nr_pages = folio_nr_pages(folio); in hugetlb_cgroup_move_parent()
218 page_counter_charge(&parent->hugepage[idx], nr_pages); in hugetlb_cgroup_move_parent()
222 page_counter_cancel(counter, nr_pages); in hugetlb_cgroup_move_parent()
264 static int __hugetlb_cgroup_charge_cgroup(int idx, unsigned long nr_pages,
285 nr_pages, &counter)) { in __hugetlb_cgroup_charge_cgroup()
301 int hugetlb_cgroup_charge_cgroup(int idx, unsigned long nr_pages,
304 return __hugetlb_cgroup_charge_cgroup(idx, nr_pages, ptr, false); in hugetlb_cgroup_charge_cgroup()
307 int hugetlb_cgroup_charge_cgroup_rsvd(int idx, unsigned long nr_pages,
310 return __hugetlb_cgroup_charge_cgroup(idx, nr_pages, pt in hugetlb_cgroup_charge_cgroup_rsvd()
265 __hugetlb_cgroup_charge_cgroup(int idx,unsigned long nr_pages,struct hugetlb_cgroup ** ptr,bool rsvd) __hugetlb_cgroup_charge_cgroup() argument
302 hugetlb_cgroup_charge_cgroup(int idx,unsigned long nr_pages,struct hugetlb_cgroup ** ptr) hugetlb_cgroup_charge_cgroup() argument
308 hugetlb_cgroup_charge_cgroup_rsvd(int idx,unsigned long nr_pages,struct hugetlb_cgroup ** ptr) hugetlb_cgroup_charge_cgroup_rsvd() argument
315 __hugetlb_cgroup_commit_charge(int idx,unsigned long nr_pages,struct hugetlb_cgroup * h_cg,struct folio * folio,bool rsvd) __hugetlb_cgroup_commit_charge() argument
336 hugetlb_cgroup_commit_charge(int idx,unsigned long nr_pages,struct hugetlb_cgroup * h_cg,struct folio * folio) hugetlb_cgroup_commit_charge() argument
343 hugetlb_cgroup_commit_charge_rsvd(int idx,unsigned long nr_pages,struct hugetlb_cgroup * h_cg,struct folio * folio) hugetlb_cgroup_commit_charge_rsvd() argument
353 __hugetlb_cgroup_uncharge_folio(int idx,unsigned long nr_pages,struct folio * folio,bool rsvd) __hugetlb_cgroup_uncharge_folio() argument
385 hugetlb_cgroup_uncharge_folio(int idx,unsigned long nr_pages,struct folio * folio) hugetlb_cgroup_uncharge_folio() argument
391 hugetlb_cgroup_uncharge_folio_rsvd(int idx,unsigned long nr_pages,struct folio * folio) hugetlb_cgroup_uncharge_folio_rsvd() argument
397 __hugetlb_cgroup_uncharge_cgroup(int idx,unsigned long nr_pages,struct hugetlb_cgroup * h_cg,bool rsvd) __hugetlb_cgroup_uncharge_cgroup() argument
412 hugetlb_cgroup_uncharge_cgroup(int idx,unsigned long nr_pages,struct hugetlb_cgroup * h_cg) hugetlb_cgroup_uncharge_cgroup() argument
418 hugetlb_cgroup_uncharge_cgroup_rsvd(int idx,unsigned long nr_pages,struct hugetlb_cgroup * h_cg) hugetlb_cgroup_uncharge_cgroup_rsvd() argument
438 hugetlb_cgroup_uncharge_file_region(struct resv_map * resv,struct file_region * rg,unsigned long nr_pages,bool region_del) hugetlb_cgroup_uncharge_file_region() argument
600 unsigned long nr_pages; hugetlb_cgroup_write() local
[all...]
H A Dmmu_gather.c60 unsigned int nr_pages = 1; in tlb_flush_rmap_batch() local
64 nr_pages = encoded_nr_pages(pages[++i]); in tlb_flush_rmap_batch()
66 folio_remove_rmap_ptes(page_folio(page), page, nr_pages, in tlb_flush_rmap_batch()
104 unsigned int nr, nr_pages; in __tlb_batch_free_encoded_pages() local
111 * Make sure we cover page + nr_pages, and don't leave in __tlb_batch_free_encoded_pages()
112 * nr_pages behind when capping the number of entries. in __tlb_batch_free_encoded_pages()
125 for (nr = 0, nr_pages = 0; in __tlb_batch_free_encoded_pages()
126 nr < batch->nr && nr_pages < MAX_NR_FOLIOS_PER_FREE; in __tlb_batch_free_encoded_pages()
130 nr_pages += encoded_nr_pages(pages[++nr]); in __tlb_batch_free_encoded_pages()
132 nr_pages in __tlb_batch_free_encoded_pages()
165 __tlb_remove_folio_pages_size(struct mmu_gather * tlb,struct page * page,unsigned int nr_pages,bool delay_rmap,int page_size) __tlb_remove_folio_pages_size() argument
206 __tlb_remove_folio_pages(struct mmu_gather * tlb,struct page * page,unsigned int nr_pages,bool delay_rmap) __tlb_remove_folio_pages() argument
[all...]
H A Dmemcontrol-v1.h11 unsigned int nr_pages);
14 unsigned int nr_pages) in try_charge() argument
19 return try_charge_memcg(memcg, gfp_mask, nr_pages); in try_charge()
105 void memcg1_account_kmem(struct mem_cgroup *memcg, int nr_pages);
110 bool memcg1_charge_skmem(struct mem_cgroup *memcg, unsigned int nr_pages,
112 static inline void memcg1_uncharge_skmem(struct mem_cgroup *memcg, unsigned int nr_pages) in memcg1_uncharge_skmem() argument
114 page_counter_uncharge(&memcg->tcpmem, nr_pages); in memcg1_uncharge_skmem()
145 static inline void memcg1_account_kmem(struct mem_cgroup *memcg, int nr_pages) {} in memcg1_account_kmem() argument
147 static inline bool memcg1_charge_skmem(struct mem_cgroup *memcg, unsigned int nr_pages, in memcg1_charge_skmem() argument
149 static inline void memcg1_uncharge_skmem(struct mem_cgroup *memcg, unsigned int nr_pages) {} in memcg1_uncharge_skmem() argument
H A Dcma.h47 void cma_sysfs_account_success_pages(struct cma *cma, unsigned long nr_pages);
48 void cma_sysfs_account_fail_pages(struct cma *cma, unsigned long nr_pages);
49 void cma_sysfs_account_release_pages(struct cma *cma, unsigned long nr_pages);
52 unsigned long nr_pages) {}; in cma_sysfs_account_success_pages() argument
54 unsigned long nr_pages) {}; in cma_sysfs_account_fail_pages() argument
56 unsigned long nr_pages) {}; in cma_sysfs_account_release_pages() argument
H A Dmemcontrol.c133 unsigned int nr_pages);
139 unsigned int nr_pages; in obj_cgroup_release() local
164 nr_pages = nr_bytes >> PAGE_SHIFT; in obj_cgroup_release()
166 if (nr_pages) in obj_cgroup_release()
167 obj_cgroup_uncharge_pages(objcg, nr_pages); in obj_cgroup_release()
1282 int zid, int nr_pages) in mem_cgroup_update_lru_size() argument
1294 if (nr_pages < 0) in mem_cgroup_update_lru_size()
1295 *lru_size += nr_pages; in mem_cgroup_update_lru_size()
1300 __func__, lruvec, lru, nr_pages, size)) { in mem_cgroup_update_lru_size()
1305 if (nr_pages > 0) in mem_cgroup_update_lru_size()
[all …]
H A Dpage_ext.c193 unsigned long nr_pages; in alloc_node_page_ext() local
195 nr_pages = NODE_DATA(nid)->node_spanned_pages; in alloc_node_page_ext()
196 if (!nr_pages) in alloc_node_page_ext()
206 nr_pages += MAX_ORDER_NR_PAGES; in alloc_node_page_ext()
208 table_size = page_ext_size * nr_pages; in alloc_node_page_ext()
372 unsigned long nr_pages, in online_page_ext() argument
379 end = SECTION_ALIGN_UP(start_pfn + nr_pages); in online_page_ext()
405 unsigned long nr_pages) in offline_page_ext() argument
410 end = SECTION_ALIGN_UP(start_pfn + nr_pages); in offline_page_ext()
439 mn->nr_pages, mn->status_change_nid); in page_ext_callback()
[all …]
/linux/include/linux/
H A Dhugetlb_cgroup.h127 extern int hugetlb_cgroup_charge_cgroup(int idx, unsigned long nr_pages,
129 extern int hugetlb_cgroup_charge_cgroup_rsvd(int idx, unsigned long nr_pages,
131 extern void hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages,
134 extern void hugetlb_cgroup_commit_charge_rsvd(int idx, unsigned long nr_pages,
137 extern void hugetlb_cgroup_uncharge_folio(int idx, unsigned long nr_pages,
139 extern void hugetlb_cgroup_uncharge_folio_rsvd(int idx, unsigned long nr_pages,
142 extern void hugetlb_cgroup_uncharge_cgroup(int idx, unsigned long nr_pages,
144 extern void hugetlb_cgroup_uncharge_cgroup_rsvd(int idx, unsigned long nr_pages,
152 unsigned long nr_pages,
162 unsigned long nr_pages, in hugetlb_cgroup_uncharge_file_region() argument
[all …]
H A Dpage_counter.h68 void page_counter_cancel(struct page_counter *counter, unsigned long nr_pages);
69 void page_counter_charge(struct page_counter *counter, unsigned long nr_pages);
71 unsigned long nr_pages,
73 void page_counter_uncharge(struct page_counter *counter, unsigned long nr_pages);
74 void page_counter_set_min(struct page_counter *counter, unsigned long nr_pages);
75 void page_counter_set_low(struct page_counter *counter, unsigned long nr_pages);
78 unsigned long nr_pages) in page_counter_set_high() argument
80 WRITE_ONCE(counter->high, nr_pages); in page_counter_set_high()
83 int page_counter_set_max(struct page_counter *counter, unsigned long nr_pages);
85 unsigned long *nr_pages);
H A Dmemory_hotplug.h123 long nr_pages);
125 extern int mhp_init_memmap_on_memory(unsigned long pfn, unsigned long nr_pages,
127 extern void mhp_deinit_memmap_on_memory(unsigned long pfn, unsigned long nr_pages);
128 extern int online_pages(unsigned long pfn, unsigned long nr_pages,
155 extern void __remove_pages(unsigned long start_pfn, unsigned long nr_pages,
159 extern int __add_pages(int nid, unsigned long start_pfn, unsigned long nr_pages,
164 unsigned long nr_pages, struct mhp_params *params)
166 return __add_pages(nid, start_pfn, nr_pages, params); in add_pages() argument
169 int add_pages(int nid, unsigned long start_pfn, unsigned long nr_pages, in add_pages()
280 extern int offline_pages(unsigned long start_pfn, unsigned long nr_pages,
291 offline_pages(unsigned long start_pfn,unsigned long nr_pages,struct zone * zone,struct memory_group * group) offline_pages() argument
[all...]
/linux/fs/iomap/
H A Dswapfile.c19 unsigned long nr_pages; /* number of pages collected */ member
33 unsigned long nr_pages; in iomap_swapfile_add_extent() local
40 if (unlikely(isi->nr_pages >= isi->sis->max)) in iomap_swapfile_add_extent()
42 max_pages = isi->sis->max - isi->nr_pages; in iomap_swapfile_add_extent()
55 nr_pages = next_ppage - first_ppage; in iomap_swapfile_add_extent()
56 nr_pages = min(nr_pages, max_pages); in iomap_swapfile_add_extent()
72 error = add_swap_extent(isi->sis, isi->nr_pages, nr_pages, first_ppage); in iomap_swapfile_add_extent()
76 isi->nr_pages in iomap_swapfile_add_extent()
[all...]
/linux/drivers/xen/
H A Dballoon.c386 static enum bp_state increase_reservation(unsigned long nr_pages) in increase_reservation() argument
392 if (nr_pages > ARRAY_SIZE(frame_list)) in increase_reservation()
393 nr_pages = ARRAY_SIZE(frame_list); in increase_reservation()
396 for (i = 0; i < nr_pages; i++) { in increase_reservation()
398 nr_pages = i; in increase_reservation()
406 rc = xenmem_reservation_increase(nr_pages, frame_list); in increase_reservation()
429 static enum bp_state decrease_reservation(unsigned long nr_pages, gfp_t gfp) in decrease_reservation() argument
437 if (nr_pages > ARRAY_SIZE(frame_list)) in decrease_reservation()
438 nr_pages = ARRAY_SIZE(frame_list); in decrease_reservation()
440 for (i = 0; i < nr_pages; in decrease_reservation()
571 add_ballooned_pages(unsigned int nr_pages) add_ballooned_pages() argument
604 xen_alloc_ballooned_pages(unsigned int nr_pages,struct page ** pages) xen_alloc_ballooned_pages() argument
657 xen_free_ballooned_pages(unsigned int nr_pages,struct page ** pages) xen_free_ballooned_pages() argument
[all...]
H A Dunpopulated-alloc.c34 static int fill_list(unsigned int nr_pages) in fill_list() argument
39 unsigned int i, alloc_pages = round_up(nr_pages, PAGES_PER_SECTION); in fill_list()
158 int xen_alloc_unpopulated_pages(unsigned int nr_pages, struct page **pages) in xen_alloc_unpopulated_pages() argument
169 return xen_alloc_ballooned_pages(nr_pages, pages); in xen_alloc_unpopulated_pages()
172 if (list_count < nr_pages) { in xen_alloc_unpopulated_pages()
173 ret = fill_list(nr_pages - list_count); in xen_alloc_unpopulated_pages()
178 for (i = 0; i < nr_pages; i++) { in xen_alloc_unpopulated_pages()
214 void xen_free_unpopulated_pages(unsigned int nr_pages, struct page **pages) in xen_free_unpopulated_pages() argument
219 xen_free_ballooned_pages(nr_pages, pages); in xen_free_unpopulated_pages()
224 for (i = 0; i < nr_pages; i++) { in xen_free_unpopulated_pages()
/linux/net/rds/
H A Dinfo.c163 unsigned long nr_pages = 0; in rds_info_getsockopt() local
187 nr_pages = (PAGE_ALIGN(start + len) - (start & PAGE_MASK)) in rds_info_getsockopt()
190 pages = kmalloc_array(nr_pages, sizeof(struct page *), GFP_KERNEL); in rds_info_getsockopt()
195 ret = pin_user_pages_fast(start, nr_pages, FOLL_WRITE, pages); in rds_info_getsockopt()
196 if (ret != nr_pages) { in rds_info_getsockopt()
198 nr_pages = ret; in rds_info_getsockopt()
200 nr_pages = 0; in rds_info_getsockopt()
205 rdsdebug("len %d nr_pages %lu\n", len, nr_pages); in rds_info_getsockopt()
238 unpin_user_pages(pages, nr_pages); in rds_info_getsockopt()
H A Drdma.c154 static int rds_pin_pages(unsigned long user_addr, unsigned int nr_pages, in rds_pin_pages() argument
163 ret = pin_user_pages_fast(user_addr, nr_pages, gup_flags, pages); in rds_pin_pages()
164 if (ret >= 0 && ret < nr_pages) { in rds_pin_pages()
178 unsigned int nr_pages; in __rds_rdma_map() local
213 nr_pages = rds_pages_in_vec(&args->vec); in __rds_rdma_map()
214 if (nr_pages == 0) { in __rds_rdma_map()
222 if ((nr_pages - 1) > (RDS_MAX_MSG_SIZE >> PAGE_SHIFT)) { in __rds_rdma_map()
228 args->vec.addr, args->vec.bytes, nr_pages); in __rds_rdma_map()
231 pages = kcalloc(nr_pages, sizeof(struct page *), GFP_KERNEL); in __rds_rdma_map()
265 ret = rds_pin_pages(args->vec.addr, nr_pages, pages, 1); in __rds_rdma_map()
[all …]
/linux/include/xen/
H A Dxen.h65 int xen_alloc_unpopulated_pages(unsigned int nr_pages, struct page **pages);
66 void xen_free_unpopulated_pages(unsigned int nr_pages, struct page **pages);
71 static inline int xen_alloc_unpopulated_pages(unsigned int nr_pages, in xen_alloc_unpopulated_pages() argument
74 return xen_alloc_ballooned_pages(nr_pages, pages); in xen_alloc_unpopulated_pages()
76 static inline void xen_free_unpopulated_pages(unsigned int nr_pages, in xen_free_unpopulated_pages() argument
79 xen_free_ballooned_pages(nr_pages, pages); in xen_free_unpopulated_pages()
/linux/drivers/firmware/efi/libstub/
H A Drelocate.c28 unsigned long nr_pages; in efi_low_alloc_above() local
45 nr_pages = size / EFI_PAGE_SIZE; in efi_low_alloc_above()
63 if (desc->num_pages < nr_pages) in efi_low_alloc_above()
77 EFI_LOADER_DATA, nr_pages, &start); in efi_low_alloc_above()
121 unsigned long nr_pages; in efi_relocate_kernel() local
139 nr_pages = round_up(alloc_size, EFI_ALLOC_ALIGN) / EFI_PAGE_SIZE; in efi_relocate_kernel()
141 EFI_LOADER_DATA, nr_pages, &efi_addr); in efi_relocate_kernel()
/linux/fs/crypto/
H A Dbio.c123 unsigned int nr_pages; in fscrypt_zeroout_range() local
137 nr_pages = min_t(u64, ARRAY_SIZE(pages), in fscrypt_zeroout_range()
147 for (i = 0; i < nr_pages; i++) { in fscrypt_zeroout_range()
153 nr_pages = i; in fscrypt_zeroout_range()
154 if (WARN_ON_ONCE(nr_pages <= 0)) in fscrypt_zeroout_range()
158 bio = bio_alloc(inode->i_sb->s_bdev, nr_pages, REQ_OP_WRITE, GFP_NOFS); in fscrypt_zeroout_range()
184 } while (i != nr_pages && du_remaining != 0); in fscrypt_zeroout_range()
194 for (i = 0; i < nr_pages; i++) in fscrypt_zeroout_range()
/linux/arch/powerpc/platforms/powernv/
H A Dmemtrace.c93 const unsigned long nr_pages = PHYS_PFN(size); in memtrace_clear_range()
101 page = alloc_contig_pages(nr_pages, GFP_KERNEL | __GFP_THISNODE | in memtrace_clear_range()
112 (unsigned long)pfn_to_kaddr(start_pfn + nr_pages), in memtrace_alloc_node()
119 for (pfn = start_pfn; pfn < start_pfn + nr_pages; pfn++) in memtrace_alloc_node()
197 const unsigned long nr_pages = PHYS_PFN(size); in memtrace_init_debugfs()
206 for (pfn = start_pfn; pfn < start_pfn + nr_pages; pfn++) in memtrace_init_debugfs()
209 free_contig_range(start_pfn, nr_pages); in memtrace_init_debugfs()
92 memtrace_clear_range(unsigned long start_pfn,unsigned long nr_pages) memtrace_clear_range() argument
113 const unsigned long nr_pages = PHYS_PFN(size); memtrace_alloc_node() local
216 const unsigned long nr_pages = PHYS_PFN(size); memtrace_free() local
/linux/drivers/misc/bcm-vk/
H A Dbcm_vk_sg.c60 dma->nr_pages = last - first + 1; in bcm_vk_dma_alloc()
63 dma->pages = kmalloc_array(dma->nr_pages, in bcm_vk_dma_alloc()
70 data, vkdata->size, dma->nr_pages); in bcm_vk_dma_alloc()
76 dma->nr_pages, in bcm_vk_dma_alloc()
79 if (err != dma->nr_pages) { in bcm_vk_dma_alloc()
80 dma->nr_pages = (err >= 0) ? err : 0; in bcm_vk_dma_alloc()
82 err, dma->nr_pages); in bcm_vk_dma_alloc()
87 dma->sglen = (dma->nr_pages * sizeof(*sgdata)) + in bcm_vk_dma_alloc()
117 for (i = 1; i < dma->nr_pages; i++) { in bcm_vk_dma_alloc()
250 for (i = 0; i < dma->nr_pages; i++) in bcm_vk_dma_free()
/linux/io_uring/
H A Drsrc.c35 int __io_account_mem(struct user_struct *user, unsigned long nr_pages) in __io_account_mem() argument
39 if (!nr_pages) in __io_account_mem()
47 new_pages = cur_pages + nr_pages; in __io_account_mem()
55 static void io_unaccount_mem(struct io_ring_ctx *ctx, unsigned long nr_pages) in io_unaccount_mem() argument
58 __io_unaccount_mem(ctx->user, nr_pages); in io_unaccount_mem()
61 atomic64_sub(nr_pages, &ctx->mm_account->pinned_vm); in io_unaccount_mem()
64 static int io_account_mem(struct io_ring_ctx *ctx, unsigned long nr_pages) in io_account_mem() argument
69 ret = __io_account_mem(ctx->user, nr_pages); in io_account_mem()
75 atomic64_add(nr_pages, &ctx->mm_account->pinned_vm); in io_account_mem()
564 int nr_pages, struct page *hpage) in headpage_already_acct() argument
[all …]
/linux/drivers/net/ethernet/ibm/ehea/
H A Dehea_qmr.c93 int i, nr_pages; in hw_queue_dtor() local
100 nr_pages = queue->queue_length / queue->pagesize; in hw_queue_dtor()
102 for (i = 0; i < nr_pages; i += pages_per_kpage) in hw_queue_dtor()
134 ret = hw_queue_ctor(&cq->hw_queue, cq->attr.nr_pages, in ehea_create_cq()
139 for (counter = 0; counter < cq->attr.nr_pages; counter++) { in ehea_create_cq()
152 cq, hret, counter, cq->attr.nr_pages); in ehea_create_cq()
156 if (counter == (cq->attr.nr_pages - 1)) { in ehea_create_cq()
255 ret = hw_queue_ctor(&eq->hw_queue, eq->attr.nr_pages, in ehea_create_eq()
262 for (i = 0; i < eq->attr.nr_pages; i++) { in ehea_create_eq()
276 if (i == (eq->attr.nr_pages - 1)) { in ehea_create_eq()
[all …]

12345678910>>...13