1 /* 2 * Copyright 2003 PathScale Inc 3 * Derived from include/asm-i386/pgtable.h 4 * Licensed under the GPL 5 */ 6 7 #ifndef __UM_PGTABLE_3LEVEL_H 8 #define __UM_PGTABLE_3LEVEL_H 9 10 #define __ARCH_USE_5LEVEL_HACK 11 #include <asm-generic/pgtable-nopud.h> 12 13 /* PGDIR_SHIFT determines what a third-level page table entry can map */ 14 15 #ifdef CONFIG_64BIT 16 #define PGDIR_SHIFT 30 17 #else 18 #define PGDIR_SHIFT 31 19 #endif 20 #define PGDIR_SIZE (1UL << PGDIR_SHIFT) 21 #define PGDIR_MASK (~(PGDIR_SIZE-1)) 22 23 /* PMD_SHIFT determines the size of the area a second-level page table can 24 * map 25 */ 26 27 #define PMD_SHIFT 21 28 #define PMD_SIZE (1UL << PMD_SHIFT) 29 #define PMD_MASK (~(PMD_SIZE-1)) 30 31 /* 32 * entries per page directory level 33 */ 34 35 #define PTRS_PER_PTE 512 36 #ifdef CONFIG_64BIT 37 #define PTRS_PER_PMD 512 38 #define PTRS_PER_PGD 512 39 #else 40 #define PTRS_PER_PMD 1024 41 #define PTRS_PER_PGD 1024 42 #endif 43 44 #define USER_PTRS_PER_PGD ((TASK_SIZE + (PGDIR_SIZE - 1)) / PGDIR_SIZE) 45 #define FIRST_USER_ADDRESS 0UL 46 47 #define pte_ERROR(e) \ 48 printk("%s:%d: bad pte %p(%016lx).\n", __FILE__, __LINE__, &(e), \ 49 pte_val(e)) 50 #define pmd_ERROR(e) \ 51 printk("%s:%d: bad pmd %p(%016lx).\n", __FILE__, __LINE__, &(e), \ 52 pmd_val(e)) 53 #define pgd_ERROR(e) \ 54 printk("%s:%d: bad pgd %p(%016lx).\n", __FILE__, __LINE__, &(e), \ 55 pgd_val(e)) 56 57 #define pud_none(x) (!(pud_val(x) & ~_PAGE_NEWPAGE)) 58 #define pud_bad(x) ((pud_val(x) & (~PAGE_MASK & ~_PAGE_USER)) != _KERNPG_TABLE) 59 #define pud_present(x) (pud_val(x) & _PAGE_PRESENT) 60 #define pud_populate(mm, pud, pmd) \ 61 set_pud(pud, __pud(_PAGE_TABLE + __pa(pmd))) 62 63 #ifdef CONFIG_64BIT 64 #define set_pud(pudptr, pudval) set_64bit((u64 *) (pudptr), pud_val(pudval)) 65 #else 66 #define set_pud(pudptr, pudval) (*(pudptr) = (pudval)) 67 #endif 68 69 static inline int pgd_newpage(pgd_t pgd) 70 { 71 return(pgd_val(pgd) & _PAGE_NEWPAGE); 72 } 73 74 static inline void pgd_mkuptodate(pgd_t pgd) { pgd_val(pgd) &= ~_PAGE_NEWPAGE; } 75 76 #ifdef CONFIG_64BIT 77 #define set_pmd(pmdptr, pmdval) set_64bit((u64 *) (pmdptr), pmd_val(pmdval)) 78 #else 79 #define set_pmd(pmdptr, pmdval) (*(pmdptr) = (pmdval)) 80 #endif 81 82 struct mm_struct; 83 extern pmd_t *pmd_alloc_one(struct mm_struct *mm, unsigned long address); 84 85 static inline void pud_clear (pud_t *pud) 86 { 87 set_pud(pud, __pud(_PAGE_NEWPAGE)); 88 } 89 90 #define pud_page(pud) phys_to_page(pud_val(pud) & PAGE_MASK) 91 #define pud_page_vaddr(pud) ((unsigned long) __va(pud_val(pud) & PAGE_MASK)) 92 93 /* Find an entry in the second-level page table.. */ 94 #define pmd_offset(pud, address) ((pmd_t *) pud_page_vaddr(*(pud)) + \ 95 pmd_index(address)) 96 97 static inline unsigned long pte_pfn(pte_t pte) 98 { 99 return phys_to_pfn(pte_val(pte)); 100 } 101 102 static inline pte_t pfn_pte(unsigned long page_nr, pgprot_t pgprot) 103 { 104 pte_t pte; 105 phys_t phys = pfn_to_phys(page_nr); 106 107 pte_set_val(pte, phys, pgprot); 108 return pte; 109 } 110 111 static inline pmd_t pfn_pmd(unsigned long page_nr, pgprot_t pgprot) 112 { 113 return __pmd((page_nr << PAGE_SHIFT) | pgprot_val(pgprot)); 114 } 115 116 #endif 117 118