1 #ifndef _ASM_SPARC64_HUGETLB_H 2 #define _ASM_SPARC64_HUGETLB_H 3 4 #include <asm/page.h> 5 #include <asm-generic/hugetlb.h> 6 7 8 void set_huge_pte_at(struct mm_struct *mm, unsigned long addr, 9 pte_t *ptep, pte_t pte); 10 11 pte_t huge_ptep_get_and_clear(struct mm_struct *mm, unsigned long addr, 12 pte_t *ptep); 13 14 static inline int is_hugepage_only_range(struct mm_struct *mm, 15 unsigned long addr, 16 unsigned long len) { 17 return 0; 18 } 19 20 /* 21 * If the arch doesn't supply something else, assume that hugepage 22 * size aligned regions are ok without further preparation. 23 */ 24 static inline int prepare_hugepage_range(struct file *file, 25 unsigned long addr, unsigned long len) 26 { 27 struct hstate *h = hstate_file(file); 28 29 if (len & ~huge_page_mask(h)) 30 return -EINVAL; 31 if (addr & ~huge_page_mask(h)) 32 return -EINVAL; 33 return 0; 34 } 35 36 static inline void huge_ptep_clear_flush(struct vm_area_struct *vma, 37 unsigned long addr, pte_t *ptep) 38 { 39 } 40 41 static inline int huge_pte_none(pte_t pte) 42 { 43 return pte_none(pte); 44 } 45 46 static inline pte_t huge_pte_wrprotect(pte_t pte) 47 { 48 return pte_wrprotect(pte); 49 } 50 51 static inline void huge_ptep_set_wrprotect(struct mm_struct *mm, 52 unsigned long addr, pte_t *ptep) 53 { 54 pte_t old_pte = *ptep; 55 set_huge_pte_at(mm, addr, ptep, pte_wrprotect(old_pte)); 56 } 57 58 static inline int huge_ptep_set_access_flags(struct vm_area_struct *vma, 59 unsigned long addr, pte_t *ptep, 60 pte_t pte, int dirty) 61 { 62 int changed = !pte_same(*ptep, pte); 63 if (changed) { 64 set_huge_pte_at(vma->vm_mm, addr, ptep, pte); 65 flush_tlb_page(vma, addr); 66 } 67 return changed; 68 } 69 70 static inline pte_t huge_ptep_get(pte_t *ptep) 71 { 72 return *ptep; 73 } 74 75 static inline void arch_clear_hugepage_flags(struct page *page) 76 { 77 } 78 79 void hugetlb_free_pgd_range(struct mmu_gather *tlb, unsigned long addr, 80 unsigned long end, unsigned long floor, 81 unsigned long ceiling); 82 83 #endif /* _ASM_SPARC64_HUGETLB_H */ 84