Home
last modified time | relevance | path

Searched refs:PTRS_PER_PMD (Results 1 – 25 of 48) sorted by relevance

12

/linux/arch/x86/boot/startup/
H A Dmap_kernel.c14 extern pmd_t early_dynamic_pgts[EARLY_DYNAMIC_PAGE_TABLES][PTRS_PER_PMD];
64 early_snp_set_memory_shared(paddr, paddr, PTRS_PER_PMD); in sme_postprocess_startup()
90 pmd_t (*early_pgts)[PTRS_PER_PMD] = rip_rel_ptr(early_dynamic_pgts); in __startup_64()
182 pmd[idx % PTRS_PER_PMD] = pmd_entry + i * PMD_SIZE; in __startup_64()
213 for (; i < PTRS_PER_PMD; i++) in __startup_64()
/linux/arch/x86/kernel/
H A Dhead64.c49 extern pmd_t early_dynamic_pgts[EARLY_DYNAMIC_PAGE_TABLES][PTRS_PER_PMD];
141 memset(pmd_p, 0, sizeof(*pmd_p) * PTRS_PER_PMD); in __early_make_pgtable()
H A Despfix_64.c90 #define ESPFIX_PMD_CLONES PTRS_PER_PMD
/linux/arch/loongarch/include/asm/
H A Dpgtable.h61 #define PTRS_PER_PMD (PAGE_SIZE >> PTRLOG) macro
112 …min(PTRS_PER_PGD * PTRS_PER_PUD * PTRS_PER_PMD * PTRS_PER_PTE * PAGE_SIZE, (1UL << cpu_vabits)) - …
116 …min(PTRS_PER_PGD * PTRS_PER_PUD * PTRS_PER_PMD * PTRS_PER_PTE * PAGE_SIZE, (1UL << cpu_vabits) / 2…
198 extern pmd_t invalid_pmd_table[PTRS_PER_PMD];
/linux/arch/loongarch/mm/
H A Dinit.c85 for (; (k < PTRS_PER_PMD) && (vaddr < end); pmd++, k++) { in fixrange_init()
247 pmd_t invalid_pmd_table[PTRS_PER_PMD] __page_aligned_bss;
H A Dpgtable.c78 end = p + PTRS_PER_PMD; in pmd_init()
/linux/drivers/hv/
H A Dmshv_regions.c20 #define MSHV_MAP_FAULT_IN_PAGES PTRS_PER_PMD
45 !IS_ALIGNED(gfn, PTRS_PER_PMD) || in mshv_chunk_stride()
46 !IS_ALIGNED(page_count, PTRS_PER_PMD)) in mshv_chunk_stride()
/linux/arch/sh/include/asm/
H A Dpgtable-3level.h26 #define PTRS_PER_PMD ((1 << PGDIR_SHIFT) / PMD_SIZE) macro
/linux/arch/sh/mm/
H A Dpgtable.c29 PTRS_PER_PMD * (1<<PTE_MAGNITUDE), in pgtable_cache_init()
/linux/include/asm-generic/
H A Dpgtable-nopmd.h21 #define PTRS_PER_PMD 1 macro
/linux/arch/powerpc/include/asm/nohash/64/
H A Dpgtable-4k.h25 #define PTRS_PER_PMD (1 << PMD_INDEX_SIZE) macro
/linux/arch/powerpc/mm/kasan/
H A Dinit_book3s_64.c69 for (i = 0; i < PTRS_PER_PMD; i++) in kasan_init()
H A Dinit_book3e_64.c96 for (i = 0; i < PTRS_PER_PMD; i++) in kasan_early_init()
/linux/arch/riscv/include/asm/
H A Dpgtable-64.h70 #define PTRS_PER_PMD (PAGE_SIZE / sizeof(pmd_t)) macro
251 #define pmd_index(addr) (((addr) >> PMD_SHIFT) & (PTRS_PER_PMD - 1))
/linux/arch/mips/mm/
H A Dpgtable-64.c52 end = p + PTRS_PER_PMD; in pmd_init()
/linux/arch/x86/include/asm/
H A Dpgtable_64_types.h75 #define PTRS_PER_PMD 512 macro
/linux/arch/arm/mm/
H A Didmap.c42 PTRS_PER_PMD * sizeof(pmd_t)); in idmap_add_pmd()
H A Dpgd.c79 clean_dcache_area(new_pmd, PTRS_PER_PMD * sizeof(pmd_t)); in pgd_alloc()
H A Dkasan_init.c29 pmd_t tmp_pmd_table[PTRS_PER_PMD] __page_aligned_bss;
/linux/arch/um/include/asm/
H A Dpgtable-4level.h39 #define PTRS_PER_PMD 512 macro
/linux/arch/x86/xen/
H A Dmmu_pv.c109 static pmd_t level2_ident_pgt[PTRS_PER_PMD] __page_aligned_bss;
608 nr = last ? pmd_index(limit) + 1 : PTRS_PER_PMD; in xen_pmd_walk()
630 if (PTRS_PER_PMD > 1) in xen_pud_walk()
1075 for (; vaddr <= vaddr_end && (pmd < (level2_kernel_pgt + PTRS_PER_PMD)); in xen_cleanhighmap()
1148 for (i = 0; i < PTRS_PER_PMD; i++) { in xen_cleanmfnmap_pud()
1825 for (i = 0; i < PTRS_PER_PMD; ++i) { in xen_setup_kernel_pagetable()
1999 for (idx_pt = 0; idx_pt < min(n_pt, PTRS_PER_PMD); in xen_relocate_p2m()
2018 n_pt -= PTRS_PER_PMD; in xen_relocate_p2m()
/linux/arch/x86/power/
H A Dhibernate_32.c101 for (pmd_idx = 0; pmd_idx < PTRS_PER_PMD; pmd++, pmd_idx++) { in resume_physical_mapping_init()
/linux/arch/arc/include/asm/
H A Dpgtable-levels.h83 #define PTRS_PER_PMD BIT(PUD_SHIFT - PMD_SHIFT) macro
/linux/arch/powerpc/include/asm/book3s/64/
H A Dhash-64k.h211 return *(char **)(pmdp + PTRS_PER_PMD); in get_hpte_slot_array()
/linux/arch/riscv/mm/
H A Dkasan_init.c57 p = memblock_alloc_or_panic(PTRS_PER_PMD * sizeof(pmd_t), PAGE_SIZE); in kasan_populate_pmd()
333 for (i = 0; i < PTRS_PER_PMD; ++i) in kasan_early_init()

12