Home
last modified time | relevance | path

Searched refs:PUD_SIZE (Results 1 – 25 of 37) sorted by relevance

12

/linux/arch/riscv/mm/
H A Dkasan_init.c97 if (pud_none(pudp_get(pudp)) && IS_ALIGNED(vaddr, PUD_SIZE) && in kasan_populate_pud()
98 (next - vaddr) >= PUD_SIZE) { in kasan_populate_pud()
99 phys_addr = memblock_phys_alloc(PUD_SIZE, PUD_SIZE); in kasan_populate_pud()
102 memset(__va(phys_addr), KASAN_SHADOW_INIT, PUD_SIZE); in kasan_populate_pud()
181 if (IS_ALIGNED(vaddr, PUD_SIZE) && (next - vaddr) >= PUD_SIZE) { in kasan_early_clear_pud()
252 if (pud_none(pudp_get(pudp)) && IS_ALIGNED(vaddr, PUD_SIZE) && in kasan_early_populate_pud()
253 (next - vaddr) >= PUD_SIZE) { in kasan_early_populate_pud()
H A Dtlbflush.c190 else if (stride_size >= PUD_SIZE) in flush_tlb_range()
191 stride_size = PUD_SIZE; in flush_tlb_range()
221 start, end - start, PUD_SIZE); in flush_pud_tlb_range()
H A Dinit.c601 if (sz == PUD_SIZE) { in create_pud_mapping()
713 !(pa & (PUD_SIZE - 1)) && !(va & (PUD_SIZE - 1)) && size >= PUD_SIZE) in best_map_size()
714 return PUD_SIZE; in best_map_size()
826 PUD_SIZE, PAGE_TABLE); in set_satp_mode()
1023 nr_pos = (PUD_SIZE - kernel_size) / PMD_SIZE; in setup_vm()
1077 BUG_ON(PUD_SIZE - (kernel_map.virt_addr & (PUD_SIZE - 1)) < kernel_map.size); in setup_vm()
1096 (uintptr_t)fixmap_pmd, PUD_SIZE, PAGE_TABLE); in setup_vm()
1107 (uintptr_t)trampoline_pmd, PUD_SIZE, PAGE_TABLE); in setup_vm()
1616 free_vmemmap_storage(pud_page(pud), PUD_SIZE, altmap); in remove_pud_mapping()
/linux/include/asm-generic/
H A Dpgtable-nopud.h20 #define PUD_SIZE (1UL << PUD_SHIFT) macro
21 #define PUD_MASK (~(PUD_SIZE-1))
/linux/arch/powerpc/include/asm/nohash/64/
H A Dpgtable-4k.h36 #define PUD_SIZE (1UL << PUD_SHIFT) macro
37 #define PUD_MASK (~(PUD_SIZE-1))
/linux/arch/x86/include/asm/
H A Dpgtable_64_types.h84 #define PUD_SIZE (_AC(1, UL) << PUD_SHIFT) macro
85 #define PUD_MASK (~(PUD_SIZE - 1))
/linux/arch/um/include/asm/
H A Dpgtable-4level.h23 #define PUD_SIZE (1UL << PUD_SHIFT) macro
24 #define PUD_MASK (~(PUD_SIZE-1))
/linux/mm/
H A Dpage_table_check.c177 page_table_check_clear(pud_pfn(pud), PUD_SIZE >> PAGE_SHIFT); in __page_table_check_pud_clear()
249 unsigned long stride = PUD_SIZE >> PAGE_SHIFT; in __page_table_check_puds_set()
256 __page_table_check_pud_clear(mm, addr + PUD_SIZE * i, *(pudp + i)); in __page_table_check_puds_set()
H A Dhugetlb.c4755 unsigned long ceil = floor + PUD_SIZE; in hugetlb_split()
6784 unsigned long s_end = sbase + PUD_SIZE; in page_table_shareable()
6808 unsigned long end = start + PUD_SIZE; in want_pmd_share()
6834 unsigned long v_start = ALIGN(vma->vm_start, PUD_SIZE), in adjust_range_if_pmd_sharing_possible()
6835 v_end = ALIGN_DOWN(vma->vm_end, PUD_SIZE); in adjust_range_if_pmd_sharing_possible()
6847 *start = ALIGN_DOWN(*start, PUD_SIZE); in adjust_range_if_pmd_sharing_possible()
6850 *end = ALIGN(*end, PUD_SIZE); in adjust_range_if_pmd_sharing_possible()
7013 if (sz == PUD_SIZE) { in huge_pte_alloc()
7058 if (sz == PUD_SIZE) in huge_pte_offset()
7081 if (hp_size == PUD_SIZE) in hugetlb_mask_last_page()
[all …]
/linux/arch/riscv/kvm/
H A Dmmu.c258 WARN_ON(size != PAGE_SIZE && size != PMD_SIZE && size != PUD_SIZE); in kvm_age_gfn()
278 WARN_ON(size != PAGE_SIZE && size != PMD_SIZE && size != PUD_SIZE); in kvm_test_age_gfn()
379 size = PUD_SIZE; in get_hva_mapping_size()
471 if (vma_pagesize == PMD_SIZE || vma_pagesize == PUD_SIZE) in kvm_riscv_mmu_map()
485 if (vma_pagesize != PUD_SIZE && in kvm_riscv_mmu_map()
/linux/arch/s390/boot/
H A Dvmem.c163 IS_ALIGNED(addr, PUD_SIZE) && end - addr >= PUD_SIZE) { in kasan_pud_populate_zero_shadow()
307 !IS_ALIGNED(addr, PUD_SIZE) || (size < PUD_SIZE)) in try_get_large_pud_pa()
311 if (!IS_ALIGNED(pa, PUD_SIZE)) in try_get_large_pud_pa()
/linux/arch/x86/mm/
H A Dkasan_init_64.c82 ((end - addr) == PUD_SIZE) && in kasan_populate_pud()
83 IS_ALIGNED(addr, PUD_SIZE)) { in kasan_populate_pud()
84 p = early_alloc(PUD_SIZE, nid, false); in kasan_populate_pud()
87 memblock_free(p, PUD_SIZE); in kasan_populate_pud()
H A Dkaslr.c166 vaddr = round_up(vaddr + 1, PUD_SIZE); in kernel_randomize_memory()
/linux/arch/arc/include/asm/
H A Dpgtable-levels.h75 #define PUD_SIZE BIT(PUD_SHIFT) macro
76 #define PUD_MASK (~(PUD_SIZE - 1))
/linux/arch/s390/mm/
H A Dvmem.c331 if (IS_ALIGNED(addr, PUD_SIZE) && in modify_pud_table()
332 IS_ALIGNED(next, PUD_SIZE)) { in modify_pud_table()
334 vmem_free_pages(pud_deref(*pud), get_order(PUD_SIZE), altmap); in modify_pud_table()
343 if (IS_ALIGNED(addr, PUD_SIZE) && in modify_pud_table()
344 IS_ALIGNED(next, PUD_SIZE) && in modify_pud_table()
/linux/arch/x86/mm/pat/
H A Dset_memory.c1511 if (start & (PUD_SIZE - 1)) { in unmap_pud_range()
1512 unsigned long next_page = (start + PUD_SIZE) & PUD_MASK; in unmap_pud_range()
1524 while (end - start >= PUD_SIZE) { in unmap_pud_range()
1529 unmap_pmd_range(pud, start, start + PUD_SIZE); in unmap_pud_range()
1531 start += PUD_SIZE; in unmap_pud_range()
1678 if (start & (PUD_SIZE - 1)) { in populate_pud()
1680 unsigned long next_page = (start + PUD_SIZE) & PUD_MASK; in populate_pud()
1713 while (boot_cpu_has(X86_FEATURE_GBPAGES) && end - start >= PUD_SIZE) { in populate_pud()
1717 start += PUD_SIZE; in populate_pud()
1718 cpa->pfn += PUD_SIZE >> PAGE_SHIFT; in populate_pud()
[all …]
/linux/arch/powerpc/mm/book3s64/
H A Dradix_hugetlbpage.c38 if (end - start >= PUD_SIZE) in radix__flush_hugetlb_tlb_range()
/linux/arch/riscv/include/asm/
H A Dpgtable-64.h36 #define PUD_SIZE (_AC(1, UL) << PUD_SHIFT) macro
37 #define PUD_MASK (~(PUD_SIZE - 1))
/linux/arch/powerpc/include/asm/nohash/
H A Dpgtable.h78 else if (sz < PUD_SIZE) in pte_update()
81 pdsize = PUD_SIZE; in pte_update()
/linux/arch/arm64/kvm/
H A Dmmu.c101 n += DIV_ROUND_UP(range, PUD_SIZE); in kvm_mmu_split_nr_page_tables()
1450 if ((hva & (PUD_SIZE - 1)) == (pa & (PUD_SIZE - 1)) && in get_vma_page_shift()
1451 ALIGN_DOWN(hva, PUD_SIZE) >= vma->vm_start && in get_vma_page_shift()
1452 ALIGN(hva, PUD_SIZE) <= vma->vm_end) in get_vma_page_shift()
1742 s2vi->max_map_size = PUD_SIZE; in kvm_s2_resolve_vma_size()
1749 if (fault_supports_stage2_huge_mapping(s2fd->memslot, s2fd->hva, PUD_SIZE)) in kvm_s2_resolve_vma_size()
1773 max_map_size = min(s2vi->max_map_size, PUD_SIZE); in kvm_s2_resolve_vma_size()
1786 if (max_map_size >= PMD_SIZE && max_map_size < PUD_SIZE) in kvm_s2_resolve_vma_size()
/linux/arch/x86/boot/compressed/
H A Dkaslr.c209 if (memparse(p, &p) != PUD_SIZE) { in parse_gb_huge_pages()
494 pud_start = ALIGN(region->start, PUD_SIZE); in process_gb_huge_pages()
495 pud_end = ALIGN_DOWN(region->start + region->size, PUD_SIZE); in process_gb_huge_pages()
/linux/arch/x86/xen/
H A Dxen-head.S152 ELFNOTE(Xen, XEN_ELFNOTE_INIT_P2M, .quad (PUD_SIZE * PTRS_PER_PUD))
/linux/arch/powerpc/include/asm/book3s/64/
H A Dhash.h203 nb = SZ_16G / PUD_SIZE; in hash__pte_update()
/linux/arch/x86/kernel/
H A Dhead64.c230 BUILD_BUG_ON(MODULES_LEN + KERNEL_IMAGE_SIZE > 2*PUD_SIZE); in x86_64_start_kernel()
/linux/arch/loongarch/include/asm/
H A Dpgtable.h42 #define PUD_SIZE (1UL << PUD_SHIFT) macro
43 #define PUD_MASK (~(PUD_SIZE-1))

12