| /linux/arch/riscv/mm/ |
| H A D | kasan_init.c | 97 if (pud_none(pudp_get(pudp)) && IS_ALIGNED(vaddr, PUD_SIZE) && in kasan_populate_pud() 98 (next - vaddr) >= PUD_SIZE) { in kasan_populate_pud() 99 phys_addr = memblock_phys_alloc(PUD_SIZE, PUD_SIZE); in kasan_populate_pud() 102 memset(__va(phys_addr), KASAN_SHADOW_INIT, PUD_SIZE); in kasan_populate_pud() 181 if (IS_ALIGNED(vaddr, PUD_SIZE) && (next - vaddr) >= PUD_SIZE) { in kasan_early_clear_pud() 252 if (pud_none(pudp_get(pudp)) && IS_ALIGNED(vaddr, PUD_SIZE) && in kasan_early_populate_pud() 253 (next - vaddr) >= PUD_SIZE) { in kasan_early_populate_pud()
|
| H A D | tlbflush.c | 190 else if (stride_size >= PUD_SIZE) in flush_tlb_range() 191 stride_size = PUD_SIZE; in flush_tlb_range() 221 start, end - start, PUD_SIZE); in flush_pud_tlb_range()
|
| H A D | init.c | 601 if (sz == PUD_SIZE) { in create_pud_mapping() 713 !(pa & (PUD_SIZE - 1)) && !(va & (PUD_SIZE - 1)) && size >= PUD_SIZE) in best_map_size() 714 return PUD_SIZE; in best_map_size() 826 PUD_SIZE, PAGE_TABLE); in set_satp_mode() 1023 nr_pos = (PUD_SIZE - kernel_size) / PMD_SIZE; in setup_vm() 1077 BUG_ON(PUD_SIZE - (kernel_map.virt_addr & (PUD_SIZE - 1)) < kernel_map.size); in setup_vm() 1096 (uintptr_t)fixmap_pmd, PUD_SIZE, PAGE_TABLE); in setup_vm() 1107 (uintptr_t)trampoline_pmd, PUD_SIZE, PAGE_TABLE); in setup_vm() 1616 free_vmemmap_storage(pud_page(pud), PUD_SIZE, altmap); in remove_pud_mapping()
|
| /linux/include/asm-generic/ |
| H A D | pgtable-nopud.h | 20 #define PUD_SIZE (1UL << PUD_SHIFT) macro 21 #define PUD_MASK (~(PUD_SIZE-1))
|
| /linux/arch/powerpc/include/asm/nohash/64/ |
| H A D | pgtable-4k.h | 36 #define PUD_SIZE (1UL << PUD_SHIFT) macro 37 #define PUD_MASK (~(PUD_SIZE-1))
|
| /linux/arch/x86/include/asm/ |
| H A D | pgtable_64_types.h | 84 #define PUD_SIZE (_AC(1, UL) << PUD_SHIFT) macro 85 #define PUD_MASK (~(PUD_SIZE - 1))
|
| /linux/arch/um/include/asm/ |
| H A D | pgtable-4level.h | 23 #define PUD_SIZE (1UL << PUD_SHIFT) macro 24 #define PUD_MASK (~(PUD_SIZE-1))
|
| /linux/mm/ |
| H A D | page_table_check.c | 177 page_table_check_clear(pud_pfn(pud), PUD_SIZE >> PAGE_SHIFT); in __page_table_check_pud_clear() 249 unsigned long stride = PUD_SIZE >> PAGE_SHIFT; in __page_table_check_puds_set() 256 __page_table_check_pud_clear(mm, addr + PUD_SIZE * i, *(pudp + i)); in __page_table_check_puds_set()
|
| H A D | hugetlb.c | 4755 unsigned long ceil = floor + PUD_SIZE; in hugetlb_split() 6784 unsigned long s_end = sbase + PUD_SIZE; in page_table_shareable() 6808 unsigned long end = start + PUD_SIZE; in want_pmd_share() 6834 unsigned long v_start = ALIGN(vma->vm_start, PUD_SIZE), in adjust_range_if_pmd_sharing_possible() 6835 v_end = ALIGN_DOWN(vma->vm_end, PUD_SIZE); in adjust_range_if_pmd_sharing_possible() 6847 *start = ALIGN_DOWN(*start, PUD_SIZE); in adjust_range_if_pmd_sharing_possible() 6850 *end = ALIGN(*end, PUD_SIZE); in adjust_range_if_pmd_sharing_possible() 7013 if (sz == PUD_SIZE) { in huge_pte_alloc() 7058 if (sz == PUD_SIZE) in huge_pte_offset() 7081 if (hp_size == PUD_SIZE) in hugetlb_mask_last_page() [all …]
|
| /linux/arch/riscv/kvm/ |
| H A D | mmu.c | 258 WARN_ON(size != PAGE_SIZE && size != PMD_SIZE && size != PUD_SIZE); in kvm_age_gfn() 278 WARN_ON(size != PAGE_SIZE && size != PMD_SIZE && size != PUD_SIZE); in kvm_test_age_gfn() 379 size = PUD_SIZE; in get_hva_mapping_size() 471 if (vma_pagesize == PMD_SIZE || vma_pagesize == PUD_SIZE) in kvm_riscv_mmu_map() 485 if (vma_pagesize != PUD_SIZE && in kvm_riscv_mmu_map()
|
| /linux/arch/s390/boot/ |
| H A D | vmem.c | 163 IS_ALIGNED(addr, PUD_SIZE) && end - addr >= PUD_SIZE) { in kasan_pud_populate_zero_shadow() 307 !IS_ALIGNED(addr, PUD_SIZE) || (size < PUD_SIZE)) in try_get_large_pud_pa() 311 if (!IS_ALIGNED(pa, PUD_SIZE)) in try_get_large_pud_pa()
|
| /linux/arch/x86/mm/ |
| H A D | kasan_init_64.c | 82 ((end - addr) == PUD_SIZE) && in kasan_populate_pud() 83 IS_ALIGNED(addr, PUD_SIZE)) { in kasan_populate_pud() 84 p = early_alloc(PUD_SIZE, nid, false); in kasan_populate_pud() 87 memblock_free(p, PUD_SIZE); in kasan_populate_pud()
|
| H A D | kaslr.c | 166 vaddr = round_up(vaddr + 1, PUD_SIZE); in kernel_randomize_memory()
|
| /linux/arch/arc/include/asm/ |
| H A D | pgtable-levels.h | 75 #define PUD_SIZE BIT(PUD_SHIFT) macro 76 #define PUD_MASK (~(PUD_SIZE - 1))
|
| /linux/arch/s390/mm/ |
| H A D | vmem.c | 331 if (IS_ALIGNED(addr, PUD_SIZE) && in modify_pud_table() 332 IS_ALIGNED(next, PUD_SIZE)) { in modify_pud_table() 334 vmem_free_pages(pud_deref(*pud), get_order(PUD_SIZE), altmap); in modify_pud_table() 343 if (IS_ALIGNED(addr, PUD_SIZE) && in modify_pud_table() 344 IS_ALIGNED(next, PUD_SIZE) && in modify_pud_table()
|
| /linux/arch/x86/mm/pat/ |
| H A D | set_memory.c | 1511 if (start & (PUD_SIZE - 1)) { in unmap_pud_range() 1512 unsigned long next_page = (start + PUD_SIZE) & PUD_MASK; in unmap_pud_range() 1524 while (end - start >= PUD_SIZE) { in unmap_pud_range() 1529 unmap_pmd_range(pud, start, start + PUD_SIZE); in unmap_pud_range() 1531 start += PUD_SIZE; in unmap_pud_range() 1678 if (start & (PUD_SIZE - 1)) { in populate_pud() 1680 unsigned long next_page = (start + PUD_SIZE) & PUD_MASK; in populate_pud() 1713 while (boot_cpu_has(X86_FEATURE_GBPAGES) && end - start >= PUD_SIZE) { in populate_pud() 1717 start += PUD_SIZE; in populate_pud() 1718 cpa->pfn += PUD_SIZE >> PAGE_SHIFT; in populate_pud() [all …]
|
| /linux/arch/powerpc/mm/book3s64/ |
| H A D | radix_hugetlbpage.c | 38 if (end - start >= PUD_SIZE) in radix__flush_hugetlb_tlb_range()
|
| /linux/arch/riscv/include/asm/ |
| H A D | pgtable-64.h | 36 #define PUD_SIZE (_AC(1, UL) << PUD_SHIFT) macro 37 #define PUD_MASK (~(PUD_SIZE - 1))
|
| /linux/arch/powerpc/include/asm/nohash/ |
| H A D | pgtable.h | 78 else if (sz < PUD_SIZE) in pte_update() 81 pdsize = PUD_SIZE; in pte_update()
|
| /linux/arch/arm64/kvm/ |
| H A D | mmu.c | 101 n += DIV_ROUND_UP(range, PUD_SIZE); in kvm_mmu_split_nr_page_tables() 1450 if ((hva & (PUD_SIZE - 1)) == (pa & (PUD_SIZE - 1)) && in get_vma_page_shift() 1451 ALIGN_DOWN(hva, PUD_SIZE) >= vma->vm_start && in get_vma_page_shift() 1452 ALIGN(hva, PUD_SIZE) <= vma->vm_end) in get_vma_page_shift() 1742 s2vi->max_map_size = PUD_SIZE; in kvm_s2_resolve_vma_size() 1749 if (fault_supports_stage2_huge_mapping(s2fd->memslot, s2fd->hva, PUD_SIZE)) in kvm_s2_resolve_vma_size() 1773 max_map_size = min(s2vi->max_map_size, PUD_SIZE); in kvm_s2_resolve_vma_size() 1786 if (max_map_size >= PMD_SIZE && max_map_size < PUD_SIZE) in kvm_s2_resolve_vma_size()
|
| /linux/arch/x86/boot/compressed/ |
| H A D | kaslr.c | 209 if (memparse(p, &p) != PUD_SIZE) { in parse_gb_huge_pages() 494 pud_start = ALIGN(region->start, PUD_SIZE); in process_gb_huge_pages() 495 pud_end = ALIGN_DOWN(region->start + region->size, PUD_SIZE); in process_gb_huge_pages()
|
| /linux/arch/x86/xen/ |
| H A D | xen-head.S | 152 ELFNOTE(Xen, XEN_ELFNOTE_INIT_P2M, .quad (PUD_SIZE * PTRS_PER_PUD))
|
| /linux/arch/powerpc/include/asm/book3s/64/ |
| H A D | hash.h | 203 nb = SZ_16G / PUD_SIZE; in hash__pte_update()
|
| /linux/arch/x86/kernel/ |
| H A D | head64.c | 230 BUILD_BUG_ON(MODULES_LEN + KERNEL_IMAGE_SIZE > 2*PUD_SIZE); in x86_64_start_kernel()
|
| /linux/arch/loongarch/include/asm/ |
| H A D | pgtable.h | 42 #define PUD_SIZE (1UL << PUD_SHIFT) macro 43 #define PUD_MASK (~(PUD_SIZE-1))
|