/linux/arch/x86/include/asm/ |
H A D | page_types.h | 21 #define PHYSICAL_PUD_PAGE_MASK (((signed long)PUD_MASK) & __PHYSICAL_MASK)
|
H A D | pgtable_64_types.h | 102 #define PUD_MASK (~(PUD_SIZE - 1)) macro
|
/linux/include/asm-generic/ |
H A D | pgtable-nopud.h | 21 #define PUD_MASK (~(PUD_SIZE-1)) macro
|
/linux/arch/powerpc/include/asm/nohash/64/ |
H A D | pgtable-4k.h | 37 #define PUD_MASK (~(PUD_SIZE-1)) macro
|
/linux/arch/x86/mm/ |
H A D | kaslr.c | 140 entropy = (rand % (entropy + 1)) & PUD_MASK; in kernel_randomize_memory()
|
H A D | init_64.c | 611 paddr_next = (paddr & PUD_MASK) + PUD_SIZE; in phys_pud_init() 615 !e820__mapped_any(paddr & PUD_MASK, paddr_next, in phys_pud_init() 617 !e820__mapped_any(paddr & PUD_MASK, paddr_next, in phys_pud_init() 619 !e820__mapped_any(paddr & PUD_MASK, paddr_next, in phys_pud_init()
|
H A D | pti.c | 337 WARN_ON_ONCE(addr & ~PUD_MASK); in pti_clone_pgtable()
|
/linux/arch/arc/include/asm/ |
H A D | pgtable-levels.h | 76 #define PUD_MASK (~(PUD_SIZE - 1)) macro
|
/linux/arch/s390/mm/ |
H A D | pageattr.c | 256 new = __pud(pud_val(new) & PUD_MASK); in modify_pud_page() 279 need_split |= !!(addr & ~PUD_MASK); in walk_pud_level()
|
H A D | hugetlbpage.c | 134 paddr = rste & PUD_MASK; in clear_huge_pte_skeys()
|
H A D | vmem.c | 363 try_free_pmd_table(pud, addr & PUD_MASK); in modify_pud_table()
|
/linux/arch/x86/platform/efi/ |
H A D | efi_64.c | 133 BUILD_BUG_ON((EFI_VA_START & ~PUD_MASK) != 0); in efi_sync_low_kernel_mappings() 134 BUILD_BUG_ON((EFI_VA_END & ~PUD_MASK) != 0); in efi_sync_low_kernel_mappings()
|
/linux/arch/riscv/include/asm/ |
H A D | pgtable-64.h | 37 #define PUD_MASK (~(PUD_SIZE - 1)) macro
|
/linux/arch/arm64/include/asm/ |
H A D | pgtable-hwdef.h | 61 #define PUD_MASK (~(PUD_SIZE-1)) macro
|
/linux/arch/x86/mm/pat/ |
H A D | set_memory.c | 820 offset = virt_addr & ~PUD_MASK; in slow_virt_to_phys() 1118 lpaddr = address & PUD_MASK; in __split_large_page() 1304 unsigned long next_page = (start + PUD_SIZE) & PUD_MASK; in unmap_pud_range() 1468 unsigned long next_page = (start + PUD_SIZE) & PUD_MASK; in populate_pud()
|
/linux/arch/arm64/mm/ |
H A D | mmu.c | 361 ((addr | next | phys) & ~PUD_MASK) == 0 && in alloc_init_pud() 1044 if (!pgtable_range_aligned(start, end, floor, ceiling, PUD_MASK)) in free_empty_pmd_table() 1206 VM_BUG_ON(phys & ~PUD_MASK); in pud_set_huge()
|
/linux/arch/riscv/mm/ |
H A D | pageattr.c | 148 vaddr <= (vaddr & PUD_MASK) && end >= next) in __split_linear_mapping_pud()
|
/linux/mm/ |
H A D | hmm.c | 450 pfn = pud_pfn(pud) + ((addr & ~PUD_MASK) >> PAGE_SHIFT); in hmm_vma_walk_pud()
|
H A D | mremap.c | 428 mask = PUD_MASK; in get_extent()
|
H A D | hugetlb.c | 5088 if (addr & ~PUD_MASK) { in hugetlb_total_pages() 5094 unsigned long floor = addr & PUD_MASK; in hugetlb_acct_memory() 5524 tlb_flush_pmd_range(tlb, address & PUD_MASK, PUD_SIZE); in copy_hugetlb_page_range() 7104 unsigned long sbase = saddr & PUD_MASK; in hugetlb_reserve_pages() 7128 unsigned long start = addr & PUD_MASK; in hugetlb_reserve_pages()
|
H A D | gup.c | 634 pfn += (addr & ~PUD_MASK) >> PAGE_SHIFT; in follow_huge_pud() 3015 fault_pfn = pud_pfn(orig) + ((addr & ~PUD_MASK) >> PAGE_SHIFT); in gup_fast_devmap_pmd_leaf()
|
H A D | debug_vm_pgtable.c | 1168 args->fixed_pud_pfn = __phys_to_pfn(phys & PUD_MASK); in init_fixed_pfns()
|
H A D | memory.c | 215 start &= PUD_MASK; in free_pmd_range() 219 ceiling &= PUD_MASK; in free_pmd_range() 6430 pud_pfn(pud), PUD_MASK, pud_write(pud), in follow_pfnmap_start()
|
H A D | vmscan.c | 3534 pmd = pmd_offset(pud, start & PUD_MASK); in walk_pmd_range() 3592 if (i < PTRS_PER_PMD && get_next_vma(PUD_MASK, PMD_SIZE, args, &start, &end)) in walk_pmd_range() 3620 end = (addr | ~PUD_MASK) + 1; in walk_pud_range()
|
/linux/arch/powerpc/kvm/ |
H A D | book3s_64_mmu_radix.c | 662 unsigned long hgpa = gpa & PUD_MASK; in kvmppc_create_pte() 1395 gpa = (gpa & PUD_MASK) + PUD_SIZE; in debugfs_radix_read()
|