1 /* SPDX-License-Identifier: GPL-2.0-only */
2 /*
3 * arch/arm/include/asm/tlb.h
4 *
5 * Copyright (C) 2002 Russell King
6 *
7 * Experimentation shows that on a StrongARM, it appears to be faster
8 * to use the "invalidate whole tlb" rather than "invalidate single
9 * tlb" for this.
10 *
11 * This appears true for both the process fork+exit case, as well as
12 * the munmap-large-area case.
13 */
14 #ifndef __ASMARM_TLB_H
15 #define __ASMARM_TLB_H
16
17 #include <asm/cacheflush.h>
18
19 #ifndef CONFIG_MMU
20
21 #include <linux/pagemap.h>
22
23 #define tlb_flush(tlb) ((void) tlb)
24
25 #include <asm-generic/tlb.h>
26
27 #else /* !CONFIG_MMU */
28
29 #include <linux/swap.h>
30 #include <asm/tlbflush.h>
31
__tlb_remove_table(void * _table)32 static inline void __tlb_remove_table(void *_table)
33 {
34 free_page_and_swap_cache((struct page *)_table);
35 }
36
37 #include <asm-generic/tlb.h>
38
39 static inline void
__pte_free_tlb(struct mmu_gather * tlb,pgtable_t pte,unsigned long addr)40 __pte_free_tlb(struct mmu_gather *tlb, pgtable_t pte, unsigned long addr)
41 {
42 struct ptdesc *ptdesc = page_ptdesc(pte);
43
44 pagetable_pte_dtor(ptdesc);
45
46 #ifndef CONFIG_ARM_LPAE
47 /*
48 * With the classic ARM MMU, a pte page has two corresponding pmd
49 * entries, each covering 1MB.
50 */
51 addr = (addr & PMD_MASK) + SZ_1M;
52 __tlb_adjust_range(tlb, addr - PAGE_SIZE, 2 * PAGE_SIZE);
53 #endif
54
55 tlb_remove_ptdesc(tlb, ptdesc);
56 }
57
58 static inline void
__pmd_free_tlb(struct mmu_gather * tlb,pmd_t * pmdp,unsigned long addr)59 __pmd_free_tlb(struct mmu_gather *tlb, pmd_t *pmdp, unsigned long addr)
60 {
61 #ifdef CONFIG_ARM_LPAE
62 struct ptdesc *ptdesc = virt_to_ptdesc(pmdp);
63
64 pagetable_pmd_dtor(ptdesc);
65 tlb_remove_ptdesc(tlb, ptdesc);
66 #endif
67 }
68
69 #endif /* CONFIG_MMU */
70 #endif
71