| /linux/mm/ |
| H A D | percpu-km.c | 65 pages = alloc_pages(gfp, order_base_2(nr_pages)); in pcpu_create_chunk() 109 size_t nr_pages, alloc_pages; in pcpu_verify_alloc_info() local 118 alloc_pages = roundup_pow_of_two(nr_pages); in pcpu_verify_alloc_info() 120 if (alloc_pages > nr_pages) in pcpu_verify_alloc_info() 122 alloc_pages - nr_pages); in pcpu_verify_alloc_info()
|
| H A D | Kconfig.debug | 30 the patterns before alloc_pages(). Additionally, this option cannot 151 the patterns before alloc_pages. The filling of the memory helps
|
| /linux/drivers/xen/ |
| H A D | unpopulated-alloc.c | 39 unsigned int i, alloc_pages = round_up(nr_pages, PAGES_PER_SECTION); in fill_list() local 53 alloc_pages * PAGE_SIZE, mhp_range.start, mhp_range.end, in fill_list() 111 for (i = 0; i < alloc_pages; i++) { in fill_list() 128 for (i = 0; i < alloc_pages; i++) { in fill_list()
|
| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | verifier_arena_large.c | 177 __noinline int alloc_pages(int page_cnt, int pages_atonce, bool first_pass, in alloc_pages() function 217 err = alloc_pages(PAGE_CNT, 1, true, PAGE_CNT, 2); in big_alloc2() 232 err = alloc_pages(PAGE_CNT / 2, 1, false, PAGE_CNT, 4); in big_alloc2() 249 err = alloc_pages(PAGE_CNT / 4, 2, false, PAGE_CNT, 6); in big_alloc2()
|
| /linux/lib/tests/ |
| H A D | fortify_kunit.c | 280 #define TEST_vmalloc(checker, expected_pages, alloc_pages) do { \ argument 283 vmalloc((alloc_pages) * PAGE_SIZE), vfree(p)); \ 285 vzalloc((alloc_pages) * PAGE_SIZE), vfree(p)); \ 287 __vmalloc((alloc_pages) * PAGE_SIZE, gfp), vfree(p)); \ 292 #define TEST_kvmalloc(checker, expected_pages, alloc_pages) do { \ argument 298 kvmalloc((alloc_pages) * PAGE_SIZE, gfp), \ 301 kvmalloc_node((alloc_pages) * PAGE_SIZE, gfp, NUMA_NO_NODE), \ 304 kvzalloc((alloc_pages) * PAGE_SIZE, gfp), \ 307 kvzalloc_node((alloc_pages) * PAGE_SIZE, gfp, NUMA_NO_NODE), \ 310 kvcalloc(1, (alloc_pages) * PAGE_SIZE, gfp), \ [all …]
|
| /linux/arch/riscv/kernel/ |
| H A D | unaligned_access_speed.c | 154 bufs[cpu] = alloc_pages(GFP_KERNEL, MISALIGNED_BUFFER_ORDER); in check_unaligned_access_speed_all_cpus() 256 buf = alloc_pages(GFP_KERNEL, MISALIGNED_BUFFER_ORDER); in riscv_online_cpu() 297 page = alloc_pages(GFP_KERNEL, MISALIGNED_BUFFER_ORDER); in check_vector_unaligned_access()
|
| /linux/drivers/infiniband/hw/vmw_pvrdma/ |
| H A D | pvrdma_misc.c | 53 u64 npages, bool alloc_pages) in pvrdma_page_dir_init() argument 83 if (alloc_pages) { in pvrdma_page_dir_init()
|
| H A D | pvrdma.h | 541 u64 npages, bool alloc_pages);
|
| /linux/drivers/net/ethernet/qlogic/qed/ |
| H A D | qed_chain.c | 273 goto alloc_pages; in qed_chain_alloc_pbl() 288 alloc_pages: in qed_chain_alloc_pbl()
|
| /linux/mm/kmsan/ |
| H A D | hooks.c | 164 shadow = alloc_pages(gfp_mask, 1); in kmsan_ioremap_page_range() 165 origin = alloc_pages(gfp_mask, 1); in kmsan_ioremap_page_range()
|
| /linux/Documentation/translations/zh_CN/core-api/ |
| H A D | memory-allocation.rst | 21 你可以用 alloc_pages 直接向页面分配器请求页面。也可以使用更专业的分配器,
|
| /linux/lib/ |
| H A D | test_meminit.c | 69 page = alloc_pages(GFP_KERNEL, order); in do_alloc_pages_order() 76 page = alloc_pages(GFP_KERNEL, order); in do_alloc_pages_order()
|
| H A D | test_lockup.c | 310 page = alloc_pages(alloc_pages_gfp, alloc_pages_order); in test_alloc_pages()
|
| /linux/include/linux/ |
| H A D | gfp.h | 342 #define alloc_pages(...) alloc_hooks(alloc_pages_noprof(__VA_ARGS__)) macro 347 #define alloc_page(gfp_mask) alloc_pages(gfp_mask, 0)
|
| /linux/arch/powerpc/include/asm/book3s/64/ |
| H A D | pgalloc.h | 31 page = alloc_pages(pgtable_gfp_flags(mm, PGALLOC_GFP | __GFP_RETRY_MAYFAIL), in radix__pgd_alloc()
|
| /linux/fs/ramfs/ |
| H A D | file-nommu.c | 84 pages = alloc_pages(gfp, order); in ramfs_nommu_expand_for_mapping()
|
| /linux/drivers/staging/media/ipu3/ |
| H A D | ipu3-dmamap.c | 58 page = alloc_pages((order_mask - order_size) ? in imgu_dmamap_alloc_buffer()
|
| /linux/arch/x86/platform/efi/ |
| H A D | memmap.c | 25 struct page *p = alloc_pages(GFP_KERNEL, order); in __efi_memmap_alloc_late()
|
| /linux/kernel/dma/ |
| H A D | pool.c | 96 page = alloc_pages(gfp | __GFP_NOWARN, order); in atomic_pool_expand()
|
| /linux/mm/kasan/ |
| H A D | kasan_test_c.c | 357 pages = alloc_pages(GFP_KERNEL, order); in page_alloc_oob_right() 371 pages = alloc_pages(GFP_KERNEL, order); in page_alloc_uaf() 1910 p_page = alloc_pages(GFP_KERNEL, 1); in vmap_tags() 1951 page = alloc_pages(GFP_KERNEL, 1); in vm_map_ram_tags() 1994 pages = alloc_pages(GFP_KERNEL, order); in match_all_not_assigned()
|
| /linux/drivers/crypto/ccp/ |
| H A D | sev-dev.c | 460 page = alloc_pages(gfp_mask, order); in __snp_alloc_firmware_pages() 698 page = alloc_pages(GFP_KERNEL_ACCOUNT, get_order(desc->len)); in snp_map_cmd_buf_desc() 1150 page = alloc_pages(GFP_KERNEL | __GFP_ZERO, order); in snp_alloc_hv_fixed_pages() 1551 page = alloc_pages(GFP_KERNEL, get_order(NV_LENGTH)); in __sev_platform_init_handle_init_ex_path() 2002 p = alloc_pages(GFP_KERNEL, order); in sev_update_firmware()
|
| /linux/io_uring/ |
| H A D | memmap.c | 30 page = alloc_pages(gfp, order); in io_mem_alloc_compound()
|
| /linux/arch/x86/hyperv/ |
| H A D | hv_crash.c | 556 page = alloc_pages(flags32, order); in hv_crash_trampoline_setup()
|
| /linux/arch/arm/mm/ |
| H A D | dma-mapping.c | 147 page = alloc_pages(gfp, order); in __dma_alloc_buffer() 904 pages[i] = alloc_pages(gfp | __GFP_NORETRY, order); in __iommu_alloc_buffer() 912 pages[i] = alloc_pages(gfp, 0); in __iommu_alloc_buffer()
|
| /linux/kernel/trace/ |
| H A D | trace_sched_switch.c | 212 page = alloc_pages(GFP_KERNEL, order); in allocate_cmdlines_buffer()
|