Home
last modified time | relevance | path

Searched refs:alloc_pages (Results 1 – 25 of 97) sorted by relevance

1234

/linux/mm/
H A Dpercpu-km.c65 pages = alloc_pages(gfp, order_base_2(nr_pages)); in pcpu_create_chunk()
109 size_t nr_pages, alloc_pages; in pcpu_verify_alloc_info() local
118 alloc_pages = roundup_pow_of_two(nr_pages); in pcpu_verify_alloc_info()
120 if (alloc_pages > nr_pages) in pcpu_verify_alloc_info()
122 alloc_pages - nr_pages); in pcpu_verify_alloc_info()
/linux/drivers/xen/
H A Dunpopulated-alloc.c42 unsigned int i, alloc_pages = round_up(nr_pages, PAGES_PER_SECTION); in fill_list() local
56 alloc_pages * PAGE_SIZE, mhp_range.start, mhp_range.end, in fill_list()
114 for (i = 0; i < alloc_pages; i++) { in fill_list()
131 for (i = 0; i < alloc_pages; i++) { in fill_list()
/linux/tools/testing/selftests/bpf/progs/
H A Dverifier_arena_large.c190 __noinline int alloc_pages(int page_cnt, int pages_atonce, bool first_pass, in alloc_pages()
230 err = alloc_pages(PAGE_CNT, 1, true, PAGE_CNT, 2); in big_alloc2()
245 err = alloc_pages(PAGE_CNT / 2, 1, false, PAGE_CNT, 4); in big_alloc2()
262 err = alloc_pages(PAGE_CNT / 4, 2, false, PAGE_CNT, 6); in big_alloc2()
177 __noinline int alloc_pages(int page_cnt, int pages_atonce, bool first_pass, alloc_pages() function
/linux/lib/tests/
H A Dfortify_kunit.c280 #define TEST_vmalloc(checker, expected_pages, alloc_pages) do { \ argument
283 vmalloc((alloc_pages) * PAGE_SIZE), vfree(p)); \
285 vzalloc((alloc_pages) * PAGE_SIZE), vfree(p)); \
287 __vmalloc((alloc_pages) * PAGE_SIZE, gfp), vfree(p)); \
292 #define TEST_kvmalloc(checker, expected_pages, alloc_pages) do { \ argument
298 kvmalloc((alloc_pages) * PAGE_SIZE, gfp), \
301 kvmalloc_node((alloc_pages) * PAGE_SIZE, gfp, NUMA_NO_NODE), \
304 kvzalloc((alloc_pages) * PAGE_SIZE, gfp), \
307 kvzalloc_node((alloc_pages) * PAGE_SIZE, gfp, NUMA_NO_NODE), \
310 kvcalloc(1, (alloc_pages) * PAGE_SIZE, gfp), \
[all …]
/linux/arch/riscv/kernel/
H A Dunaligned_access_speed.c154 bufs[cpu] = alloc_pages(GFP_KERNEL, MISALIGNED_BUFFER_ORDER); in check_unaligned_access_speed_all_cpus()
256 buf = alloc_pages(GFP_KERNEL, MISALIGNED_BUFFER_ORDER); in riscv_online_cpu()
297 page = alloc_pages(GFP_KERNEL, MISALIGNED_BUFFER_ORDER); in check_vector_unaligned_access()
/linux/drivers/infiniband/hw/vmw_pvrdma/
H A Dpvrdma_misc.c53 u64 npages, bool alloc_pages) in pvrdma_page_dir_init() argument
83 if (alloc_pages) { in pvrdma_page_dir_init()
/linux/drivers/net/ethernet/qlogic/qed/
H A Dqed_chain.c273 goto alloc_pages; in qed_chain_alloc_pbl()
288 alloc_pages: in qed_chain_alloc_pbl()
/linux/mm/kmsan/
H A Dhooks.c164 shadow = alloc_pages(gfp_mask, 1); in kmsan_ioremap_page_range()
165 origin = alloc_pages(gfp_mask, 1); in kmsan_ioremap_page_range()
H A Dkmsan_test.c388 page = alloc_pages(GFP_KERNEL, 0); in test_uninit_page()
401 page = alloc_pages(GFP_KERNEL | __GFP_ZERO, order); in test_uaf_pages_helper()
/linux/Documentation/translations/zh_CN/core-api/
H A Dmemory-allocation.rst21 你可以用 alloc_pages 直接向页面分配器请求页面。也可以使用更专业的分配器,
/linux/lib/
H A Dtest_meminit.c69 page = alloc_pages(GFP_KERNEL, order); in do_alloc_pages_order()
76 page = alloc_pages(GFP_KERNEL, order); in do_alloc_pages_order()
/linux/include/linux/
H A Dgfp.h346 #define alloc_pages(...) alloc_hooks(alloc_pages_noprof(__VA_ARGS__)) macro
351 #define alloc_page(gfp_mask) alloc_pages(gfp_mask, 0)
/linux/kernel/bpf/
H A Darena.c546 long alloc_pages; in arena_alloc_pages() local
566 alloc_pages = min(page_cnt, KMALLOC_MAX_CACHE_SIZE / sizeof(struct page *)); in arena_alloc_pages()
567 pages = kmalloc_nolock(alloc_pages * sizeof(struct page *), __GFP_ACCOUNT, NUMA_NO_NODE); in arena_alloc_pages()
594 long this_batch = min(remaining, alloc_pages); in arena_alloc_pages()
/linux/arch/powerpc/include/asm/book3s/64/
H A Dpgalloc.h31 page = alloc_pages(pgtable_gfp_flags(mm, PGALLOC_GFP | __GFP_RETRY_MAYFAIL), in radix__pgd_alloc()
/linux/drivers/gpu/drm/i915/gem/
H A Di915_gem_internal.c74 page = alloc_pages(gfp | (order ? QUIET : MAYFAIL), in i915_gem_object_get_pages_internal()
/linux/drivers/staging/media/ipu7/
H A Dipu7-dma.c74 pages[i] = alloc_pages(gfp, order); in __alloc_buffer()
76 pages[i] = alloc_pages(gfp, --order); in __alloc_buffer()
/linux/drivers/media/pci/intel/ipu6/
H A Dipu6-dma.c74 pages[i] = alloc_pages(gfp, order); in __alloc_buffer()
76 pages[i] = alloc_pages(gfp, --order); in __alloc_buffer()
/linux/fs/ramfs/
H A Dfile-nommu.c84 pages = alloc_pages(gfp, order); in ramfs_nommu_expand_for_mapping()
/linux/arch/x86/platform/efi/
H A Dmemmap.c25 struct page *p = alloc_pages(GFP_KERNEL, order); in __efi_memmap_alloc_late()
/linux/drivers/staging/media/ipu3/
H A Dipu3-dmamap.c58 page = alloc_pages((order_mask - order_size) ? in imgu_dmamap_alloc_buffer()
/linux/mm/kasan/
H A Dkasan_test_c.c357 pages = alloc_pages(GFP_KERNEL, order); in page_alloc_oob_right()
371 pages = alloc_pages(GFP_KERNEL, order); in page_alloc_uaf()
1930 p_page = alloc_pages(GFP_KERNEL, 1); in vmap_tags()
1971 page = alloc_pages(GFP_KERNEL, 1); in vm_map_ram_tags()
2014 pages = alloc_pages(GFP_KERNEL, order); in match_all_not_assigned()
/linux/kernel/dma/
H A Dpool.c96 page = alloc_pages(gfp | __GFP_NOWARN, order); in atomic_pool_expand()
/linux/drivers/crypto/ccp/
H A Dsev-dev.c453 page = alloc_pages(gfp_mask, order); in __snp_alloc_firmware_pages()
691 page = alloc_pages(GFP_KERNEL_ACCOUNT, get_order(desc->len)); in snp_map_cmd_buf_desc()
1143 page = alloc_pages(GFP_KERNEL | __GFP_ZERO, order); in snp_alloc_hv_fixed_pages()
1545 page = alloc_pages(GFP_KERNEL, get_order(NV_LENGTH)); in __sev_platform_init_handle_init_ex_path()
1996 p = alloc_pages(GFP_KERNEL, order); in sev_update_firmware()
/linux/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_gart.c133 p = alloc_pages(gfp_flags, order); in amdgpu_gart_table_ram_alloc()
/linux/io_uring/
H A Dmemmap.c30 page = alloc_pages(gfp, order); in io_mem_alloc_compound()

1234