xref: /linux/arch/sh/include/asm/cacheflush.h (revision 621cde16e49b3ecf7d59a8106a20aaebfb4a59a9)
1 /* SPDX-License-Identifier: GPL-2.0 */
2 #ifndef __ASM_SH_CACHEFLUSH_H
3 #define __ASM_SH_CACHEFLUSH_H
4 
5 #include <linux/mm.h>
6 
7 /*
8  * Cache flushing:
9  *
10  *  - flush_cache_all() flushes entire cache
11  *  - flush_cache_mm(mm) flushes the specified mm context's cache lines
12  *  - flush_cache_dup mm(mm) handles cache flushing when forking
13  *  - flush_cache_page(mm, vmaddr, pfn) flushes a single page
14  *  - flush_cache_range(vma, start, end) flushes a range of pages
15  *
16  *  - flush_dcache_folio(folio) flushes(wback&invalidates) a folio for dcache
17  *  - flush_icache_range(start, end) flushes(invalidates) a range for icache
18  *  - flush_icache_pages(vma, pg, nr) flushes(invalidates) pages for icache
19  *  - flush_cache_sigtramp(vaddr) flushes the signal trampoline
20  */
21 extern void (*local_flush_cache_all)(void *args);
22 extern void (*local_flush_cache_mm)(void *args);
23 extern void (*local_flush_cache_dup_mm)(void *args);
24 extern void (*local_flush_cache_page)(void *args);
25 extern void (*local_flush_cache_range)(void *args);
26 extern void (*local_flush_dcache_folio)(void *args);
27 extern void (*local_flush_icache_range)(void *args);
28 extern void (*local_flush_icache_folio)(void *args);
29 extern void (*local_flush_cache_sigtramp)(void *args);
30 
cache_noop(void * args)31 static inline void cache_noop(void *args) { }
32 
33 extern void (*__flush_wback_region)(void *start, int size);
34 extern void (*__flush_purge_region)(void *start, int size);
35 extern void (*__flush_invalidate_region)(void *start, int size);
36 
37 extern void flush_cache_all(void);
38 extern void flush_cache_mm(struct mm_struct *mm);
39 extern void flush_cache_dup_mm(struct mm_struct *mm);
40 extern void flush_cache_page(struct vm_area_struct *vma,
41 				unsigned long addr, unsigned long pfn);
42 extern void flush_cache_range(struct vm_area_struct *vma,
43 				 unsigned long start, unsigned long end);
44 #define ARCH_IMPLEMENTS_FLUSH_DCACHE_PAGE 1
45 void flush_dcache_folio(struct folio *folio);
46 #define flush_dcache_folio flush_dcache_folio
flush_dcache_page(struct page * page)47 static inline void flush_dcache_page(struct page *page)
48 {
49 	flush_dcache_folio(page_folio(page));
50 }
51 
52 extern void flush_icache_range(unsigned long start, unsigned long end);
53 #define flush_icache_user_range flush_icache_range
54 void flush_icache_pages(struct vm_area_struct *vma, struct page *page,
55 		unsigned int nr);
56 #define flush_icache_pages flush_icache_pages
57 extern void flush_cache_sigtramp(unsigned long address);
58 
59 struct flusher_data {
60 	struct vm_area_struct *vma;
61 	unsigned long addr1, addr2;
62 };
63 
64 #define ARCH_HAS_FLUSH_ANON_PAGE
65 extern void __flush_anon_page(struct page *page, unsigned long);
66 
flush_anon_page(struct vm_area_struct * vma,struct page * page,unsigned long vmaddr)67 static inline void flush_anon_page(struct vm_area_struct *vma,
68 				   struct page *page, unsigned long vmaddr)
69 {
70 	if (boot_cpu_data.dcache.n_aliases && PageAnon(page))
71 		__flush_anon_page(page, vmaddr);
72 }
73 
74 #define ARCH_IMPLEMENTS_FLUSH_KERNEL_VMAP_RANGE 1
flush_kernel_vmap_range(void * addr,int size)75 static inline void flush_kernel_vmap_range(void *addr, int size)
76 {
77 	__flush_wback_region(addr, size);
78 }
invalidate_kernel_vmap_range(void * addr,int size)79 static inline void invalidate_kernel_vmap_range(void *addr, int size)
80 {
81 	__flush_invalidate_region(addr, size);
82 }
83 
84 extern void copy_to_user_page(struct vm_area_struct *vma,
85 	struct page *page, unsigned long vaddr, void *dst, const void *src,
86 	unsigned long len);
87 
88 extern void copy_from_user_page(struct vm_area_struct *vma,
89 	struct page *page, unsigned long vaddr, void *dst, const void *src,
90 	unsigned long len);
91 
92 #define flush_cache_vmap(start, end)		local_flush_cache_all(NULL)
93 #define flush_cache_vmap_early(start, end)	do { } while (0)
94 #define flush_cache_vunmap(start, end)		local_flush_cache_all(NULL)
95 
96 #define flush_dcache_mmap_lock(mapping)		do { } while (0)
97 #define flush_dcache_mmap_unlock(mapping)	do { } while (0)
98 
99 void kmap_coherent_init(void);
100 void *kmap_coherent(struct page *page, unsigned long addr);
101 void kunmap_coherent(void *kvaddr);
102 
103 #define PG_dcache_clean	PG_arch_1
104 
105 void cpu_cache_init(void);
106 
107 void __weak l2_cache_init(void);
108 
109 void __weak j2_cache_init(void);
110 void __weak sh2_cache_init(void);
111 void __weak sh2a_cache_init(void);
112 void __weak sh3_cache_init(void);
113 void __weak shx3_cache_init(void);
114 void __weak sh4_cache_init(void);
115 void __weak sh7705_cache_init(void);
116 
117 void __weak sh4__flush_region_init(void);
118 
sh_cacheop_vaddr(void * vaddr)119 static inline void *sh_cacheop_vaddr(void *vaddr)
120 {
121 	if (__in_29bit_mode())
122 		vaddr = (void *)CAC_ADDR((unsigned long)vaddr);
123 	return vaddr;
124 }
125 
126 #endif /* __ASM_SH_CACHEFLUSH_H */
127