/linux/tools/testing/selftests/mm/ |
H A D | transhuge-stress.c | 76 len -= len % HPAGE_SIZE; in main() 77 ptr = mmap(NULL, len + HPAGE_SIZE, PROT_RW, mmap_flags, backing_fd, 0); in main() 80 ptr += HPAGE_SIZE - (uintptr_t)ptr % HPAGE_SIZE; in main() 98 for (p = ptr; p < ptr + len; p += HPAGE_SIZE) { in main() 122 if (madvise(p, HPAGE_SIZE - psize(), MADV_DONTNEED)) in main()
|
H A D | ksm_tests.c | 561 len -= len % HPAGE_SIZE; in ksm_merge_hugepages_time() 562 map_ptr_orig = mmap(NULL, len + HPAGE_SIZE, PROT_READ | PROT_WRITE, in ksm_merge_hugepages_time() 564 map_ptr = map_ptr_orig + HPAGE_SIZE - (uintptr_t)map_ptr_orig % HPAGE_SIZE; in ksm_merge_hugepages_time() 578 for (void *p = map_ptr; p < map_ptr + len; p += HPAGE_SIZE) { in ksm_merge_hugepages_time() 609 munmap(map_ptr_orig, len + HPAGE_SIZE); in ksm_merge_hugepages_time() 614 munmap(map_ptr_orig, len + HPAGE_SIZE); in ksm_merge_hugepages_time()
|
H A D | protection_keys.c | 693 size = ALIGN_UP(size, HPAGE_SIZE * 2); in malloc_pkey_anon_huge() 700 ptr = ALIGN_PTR_UP(ptr, HPAGE_SIZE); in malloc_pkey_anon_huge() 702 ret = madvise(ptr, HPAGE_SIZE, MADV_HUGEPAGE); in malloc_pkey_anon_huge() 704 ret = madvise(ptr, HPAGE_SIZE, MADV_WILLNEED); in malloc_pkey_anon_huge() 706 memset(ptr, 0, HPAGE_SIZE); in malloc_pkey_anon_huge() 735 hpagesz_kb = HPAGE_SIZE / 1024; in setup_hugetlbfs() 772 size = ALIGN_UP(size, HPAGE_SIZE * 2); in malloc_pkey_hugetlb() 910 if (size >= HPAGE_SIZE) in test_pkey_alloc_free_attach_pkey0() 911 size = HPAGE_SIZE; in test_pkey_alloc_free_attach_pkey0() 1294 if (size >= HPAGE_SIZE) in test_mprotect_with_pkey_0() [all …]
|
H A D | vm_util.h | 62 #define HPAGE_SIZE (1 << HPAGE_SHIFT) 61 #define HPAGE_SIZE global() macro
|
H A D | vm_util.c | 270 if (mmap(ptr, HPAGE_SIZE, PROT_READ | PROT_WRITE, in allocate_transhuge() 275 if (madvise(ptr, HPAGE_SIZE, MADV_HUGEPAGE)) in allocate_transhuge()
|
H A D | pkey-powerpc.h | 34 #define HPAGE_SIZE (1UL << 24) macro
|
/linux/arch/x86/kernel/ |
H A D | vmlinux.lds.S | 60 #define X86_ALIGN_RODATA_BEGIN . = ALIGN(HPAGE_SIZE); 63 . = ALIGN(HPAGE_SIZE); \ 429 . = ALIGN(HPAGE_SIZE); 433 . = ALIGN(HPAGE_SIZE);
|
/linux/arch/arc/include/asm/ |
H A D | hugepage.h | 17 #define HPAGE_SIZE (_AC(1, UL) << HPAGE_SHIFT) macro 18 #define HPAGE_MASK (~(HPAGE_SIZE - 1))
|
/linux/arch/riscv/include/asm/ |
H A D | page.h | 18 #define HPAGE_SIZE (_AC(1, UL) << HPAGE_SHIFT) macro 19 #define HPAGE_MASK (~(HPAGE_SIZE - 1))
|
/linux/arch/sparc/mm/ |
H A D | hugetlbpage.c | 162 size = HPAGE_SIZE; in huge_tte_to_size() 251 if (size == HPAGE_SIZE) in __set_huge_pte_at() 292 if (size == HPAGE_SIZE) in huge_ptep_get_and_clear()
|
H A D | fault_64.c | 468 mm_rss -= (mm->context.thp_pte_count * (HPAGE_SIZE / PAGE_SIZE)); in do_sparc64_fault()
|
H A D | tsb.c | 562 mm_rss -= saved_thp_pte_count * (HPAGE_SIZE / PAGE_SIZE); in init_new_context()
|
/linux/arch/s390/include/asm/ |
H A D | page.h | 23 #define HPAGE_SIZE (1UL << HPAGE_SHIFT) macro 24 #define HPAGE_MASK (~(HPAGE_SIZE - 1))
|
/linux/arch/mips/include/asm/ |
H A D | hugetlb.h | 79 flush_tlb_range(vma, addr, addr + HPAGE_SIZE); in huge_ptep_set_access_flags()
|
/linux/arch/loongarch/include/asm/ |
H A D | hugetlb.h | 83 flush_tlb_range(vma, addr, addr + HPAGE_SIZE); in huge_ptep_set_access_flags()
|
/linux/arch/s390/kvm/ |
H A D | pv.c | 190 vlen = ALIGN(virt * ((npages * PAGE_SIZE) / HPAGE_SIZE), PAGE_SIZE); in kvm_s390_pv_alloc_vm() 737 #define DUMP_BUFF_LEN HPAGE_SIZE 785 if (!IS_ALIGNED(*gaddr, HPAGE_SIZE)) in kvm_s390_pv_dump_stor_state() 825 uvcb.gaddr += HPAGE_SIZE; in kvm_s390_pv_dump_stor_state()
|
/linux/arch/riscv/mm/ |
H A D | hugetlbpage.c | 404 if (size == HPAGE_SIZE) in __hugetlb_valid_size()
|
/linux/arch/x86/mm/ |
H A D | init_32.c | 777 unsigned long size = (((unsigned long)__init_end + HPAGE_SIZE) & HPAGE_MASK) - start; in mark_nxdata_nx()
|
/linux/arch/s390/mm/ |
H A D | gmap.c | 1014 rc = HPAGE_SIZE; in gmap_protect_one() 2009 gmap_call_notifier(gmap, gaddr, gaddr + HPAGE_SIZE - 1); in pmdp_notify_gmap() 2448 end = start + HPAGE_SIZE; in __s390_enable_skey_hugetlb()
|
/linux/arch/mips/mm/ |
H A D | tlbex.c | 704 small_sequence = (HPAGE_SIZE >> 7) < 0x10000; in build_huge_update_entries() 708 uasm_i_lui(p, tmp, HPAGE_SIZE >> (7 + 16)); in build_huge_update_entries() 714 UASM_i_ADDIU(p, pte, pte, HPAGE_SIZE >> 7); in build_huge_update_entries()
|
/linux/fs/proc/ |
H A D | task_mmu.c | 2430 if (end != start + HPAGE_SIZE) { in pagemap_scan_thp_entry() 2589 if (end != start + HPAGE_SIZE) { in pagemap_scan_hugetlb_entry()
|
/linux/drivers/iommu/iommufd/ |
H A D | io_pagetable.c | 136 max_alignment = HPAGE_SIZE; in iopt_alloc_iova()
|
/linux/mm/ |
H A D | hugetlb.c | 4489 default_hstate_idx = hstate_index(size_to_hstate(HPAGE_SIZE)); in hugetlb_init() 4538 return size == HPAGE_SIZE; in __init()
|
/linux/kernel/sched/ |
H A D | fair.c | 3469 end = ALIGN(start + (pages << PAGE_SHIFT), HPAGE_SIZE); in task_numa_work()
|