/linux/mm/ |
H A D | percpu-km.c | 65 pages = alloc_pages(gfp, order_base_2(nr_pages)); in pcpu_create_chunk() 109 size_t nr_pages, alloc_pages; in pcpu_verify_alloc_info() local 118 alloc_pages = roundup_pow_of_two(nr_pages); in pcpu_verify_alloc_info() 120 if (alloc_pages > nr_pages) in pcpu_verify_alloc_info() 122 alloc_pages - nr_pages); in pcpu_verify_alloc_info()
|
H A D | Kconfig.debug | 30 the patterns before alloc_pages(). Additionally, this option cannot 151 the patterns before alloc_pages. The filling of the memory helps
|
/linux/drivers/xen/ |
H A D | unpopulated-alloc.c | 39 unsigned int i, alloc_pages = round_up(nr_pages, PAGES_PER_SECTION); in fill_list() local 53 alloc_pages * PAGE_SIZE, mhp_range.start, mhp_range.end, in fill_list() 111 for (i = 0; i < alloc_pages; i++) { in fill_list() 128 for (i = 0; i < alloc_pages; i++) { in fill_list()
|
/linux/lib/ |
H A D | fortify_kunit.c | 263 #define TEST_vmalloc(checker, expected_pages, alloc_pages) do { \ argument 266 vmalloc((alloc_pages) * PAGE_SIZE), vfree(p)); \ 268 vzalloc((alloc_pages) * PAGE_SIZE), vfree(p)); \ 270 __vmalloc((alloc_pages) * PAGE_SIZE, gfp), vfree(p)); \ 275 #define TEST_kvmalloc(checker, expected_pages, alloc_pages) do { \ argument 281 kvmalloc((alloc_pages) * PAGE_SIZE, gfp), \ 284 kvmalloc_node((alloc_pages) * PAGE_SIZE, gfp, NUMA_NO_NODE), \ 287 kvzalloc((alloc_pages) * PAGE_SIZE, gfp), \ 290 kvzalloc_node((alloc_pages) * PAGE_SIZE, gfp, NUMA_NO_NODE), \ 293 kvcalloc(1, (alloc_pages) * PAGE_SIZE, gfp), \ [all …]
|
H A D | test_meminit.c | 69 page = alloc_pages(GFP_KERNEL, order); in do_alloc_pages_order() 76 page = alloc_pages(GFP_KERNEL, order); in do_alloc_pages_order()
|
H A D | test_lockup.c | 310 page = alloc_pages(alloc_pages_gfp, alloc_pages_order); in test_alloc_pages()
|
/linux/arch/riscv/kernel/ |
H A D | unaligned_access_speed.c | 198 buf = alloc_pages(GFP_KERNEL, MISALIGNED_BUFFER_ORDER); in riscv_online_cpu() 237 bufs[cpu] = alloc_pages(GFP_KERNEL, MISALIGNED_BUFFER_ORDER); in check_unaligned_access_speed_all_cpus() 290 page = alloc_pages(GFP_KERNEL, MISALIGNED_BUFFER_ORDER); in check_vector_unaligned_access()
|
/linux/rust/helpers/ |
H A D | page.c | 8 return alloc_pages(gfp_mask, order); in rust_helper_alloc_pages()
|
/linux/drivers/infiniband/hw/vmw_pvrdma/ |
H A D | pvrdma_misc.c | 53 u64 npages, bool alloc_pages) in pvrdma_page_dir_init() argument 83 if (alloc_pages) { in pvrdma_page_dir_init()
|
H A D | pvrdma.h | 541 u64 npages, bool alloc_pages);
|
/linux/drivers/net/ethernet/qlogic/qed/ |
H A D | qed_chain.c | 273 goto alloc_pages; in qed_chain_alloc_pbl() 288 alloc_pages: in qed_chain_alloc_pbl()
|
/linux/mm/kmsan/ |
H A D | hooks.c | 164 shadow = alloc_pages(gfp_mask, 1); in kmsan_ioremap_page_range() 165 origin = alloc_pages(gfp_mask, 1); in kmsan_ioremap_page_range()
|
/linux/drivers/iommu/ |
H A D | iommu-pages.h | 60 page = alloc_pages(gfp | __GFP_ZERO, order); in __iommu_alloc_pages()
|
/linux/Documentation/translations/zh_CN/core-api/ |
H A D | memory-allocation.rst | 21 你可以用 alloc_pages 直接向页面分配器请求页面。也可以使用更专业的分配器,
|
/linux/arch/powerpc/include/asm/book3s/64/ |
H A D | pgalloc.h | 31 page = alloc_pages(pgtable_gfp_flags(mm, PGALLOC_GFP | __GFP_RETRY_MAYFAIL), in radix__pgd_alloc()
|
/linux/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_internal.c | 75 page = alloc_pages(gfp | (order ? QUIET : MAYFAIL), in i915_gem_object_get_pages_internal()
|
/linux/drivers/media/pci/intel/ipu6/ |
H A D | ipu6-dma.c | 74 pages[i] = alloc_pages(gfp, order); in __alloc_buffer() 76 pages[i] = alloc_pages(gfp, --order); in __alloc_buffer()
|
/linux/fs/ramfs/ |
H A D | file-nommu.c | 84 pages = alloc_pages(gfp, order); in ramfs_nommu_expand_for_mapping()
|
/linux/arch/x86/kernel/ |
H A D | kvmclock.c | 215 p = alloc_pages(GFP_KERNEL, order); in kvmclock_init_mem()
|
/linux/drivers/staging/media/ipu3/ |
H A D | ipu3-dmamap.c | 58 page = alloc_pages((order_mask - order_size) ? in imgu_dmamap_alloc_buffer()
|
/linux/arch/x86/platform/efi/ |
H A D | memmap.c | 25 struct page *p = alloc_pages(GFP_KERNEL, order); in __efi_memmap_alloc_late()
|
/linux/mm/kasan/ |
H A D | kasan_test_c.c | 322 pages = alloc_pages(GFP_KERNEL, order); in page_alloc_oob_right() 336 pages = alloc_pages(GFP_KERNEL, order); in page_alloc_uaf() 1785 p_page = alloc_pages(GFP_KERNEL, 1); in vmap_tags() 1826 page = alloc_pages(GFP_KERNEL, 1); in vm_map_ram_tags() 1869 pages = alloc_pages(GFP_KERNEL, order); in match_all_not_assigned()
|
/linux/kernel/dma/ |
H A D | pool.c | 96 page = alloc_pages(gfp, order); in atomic_pool_expand()
|
/linux/drivers/crypto/ccp/ |
H A D | sev-dev.c | 436 page = alloc_pages(gfp_mask, order); in __snp_alloc_firmware_pages() 674 page = alloc_pages(GFP_KERNEL_ACCOUNT, get_order(desc->len)); in snp_map_cmd_buf_desc() 1228 page = alloc_pages(GFP_KERNEL, get_order(NV_LENGTH)); in __sev_platform_init_handle_init_ex_path() 1616 p = alloc_pages(GFP_KERNEL, order); in sev_update_firmware()
|
/linux/kernel/trace/ |
H A D | trace_sched_switch.c | 212 page = alloc_pages(GFP_KERNEL, order); in allocate_cmdlines_buffer()
|