xref: /linux/include/linux/cache.h (revision ae22a94997b8a03dcb3c922857c203246711f9d4)
1 /* SPDX-License-Identifier: GPL-2.0 */
2 #ifndef __LINUX_CACHE_H
3 #define __LINUX_CACHE_H
4 
5 #include <uapi/linux/kernel.h>
6 #include <asm/cache.h>
7 
8 #ifndef L1_CACHE_ALIGN
9 #define L1_CACHE_ALIGN(x) __ALIGN_KERNEL(x, L1_CACHE_BYTES)
10 #endif
11 
12 #ifndef SMP_CACHE_BYTES
13 #define SMP_CACHE_BYTES L1_CACHE_BYTES
14 #endif
15 
16 /*
17  * __read_mostly is used to keep rarely changing variables out of frequently
18  * updated cachelines. Its use should be reserved for data that is used
19  * frequently in hot paths. Performance traces can help decide when to use
20  * this. You want __read_mostly data to be tightly packed, so that in the
21  * best case multiple frequently read variables for a hot path will be next
22  * to each other in order to reduce the number of cachelines needed to
23  * execute a critical path. We should be mindful and selective of its use.
24  * ie: if you're going to use it please supply a *good* justification in your
25  * commit log
26  */
27 #ifndef __read_mostly
28 #define __read_mostly
29 #endif
30 
31 /*
32  * __ro_after_init is used to mark things that are read-only after init (i.e.
33  * after mark_rodata_ro() has been called). These are effectively read-only,
34  * but may get written to during init, so can't live in .rodata (via "const").
35  */
36 #ifndef __ro_after_init
37 #define __ro_after_init __section(".data..ro_after_init")
38 #endif
39 
40 #ifndef ____cacheline_aligned
41 #define ____cacheline_aligned __attribute__((__aligned__(SMP_CACHE_BYTES)))
42 #endif
43 
44 #ifndef ____cacheline_aligned_in_smp
45 #ifdef CONFIG_SMP
46 #define ____cacheline_aligned_in_smp ____cacheline_aligned
47 #else
48 #define ____cacheline_aligned_in_smp
49 #endif /* CONFIG_SMP */
50 #endif
51 
52 #ifndef __cacheline_aligned
53 #define __cacheline_aligned					\
54   __attribute__((__aligned__(SMP_CACHE_BYTES),			\
55 		 __section__(".data..cacheline_aligned")))
56 #endif /* __cacheline_aligned */
57 
58 #ifndef __cacheline_aligned_in_smp
59 #ifdef CONFIG_SMP
60 #define __cacheline_aligned_in_smp __cacheline_aligned
61 #else
62 #define __cacheline_aligned_in_smp
63 #endif /* CONFIG_SMP */
64 #endif
65 
66 /*
67  * The maximum alignment needed for some critical structures
68  * These could be inter-node cacheline sizes/L3 cacheline
69  * size etc.  Define this in asm/cache.h for your arch
70  */
71 #ifndef INTERNODE_CACHE_SHIFT
72 #define INTERNODE_CACHE_SHIFT L1_CACHE_SHIFT
73 #endif
74 
75 #if !defined(____cacheline_internodealigned_in_smp)
76 #if defined(CONFIG_SMP)
77 #define ____cacheline_internodealigned_in_smp \
78 	__attribute__((__aligned__(1 << (INTERNODE_CACHE_SHIFT))))
79 #else
80 #define ____cacheline_internodealigned_in_smp
81 #endif
82 #endif
83 
84 #ifndef CONFIG_ARCH_HAS_CACHE_LINE_SIZE
85 #define cache_line_size()	L1_CACHE_BYTES
86 #endif
87 
88 #ifndef __cacheline_group_begin
89 #define __cacheline_group_begin(GROUP) \
90 	__u8 __cacheline_group_begin__##GROUP[0]
91 #endif
92 
93 #ifndef __cacheline_group_end
94 #define __cacheline_group_end(GROUP) \
95 	__u8 __cacheline_group_end__##GROUP[0]
96 #endif
97 
98 #ifndef CACHELINE_ASSERT_GROUP_MEMBER
99 #define CACHELINE_ASSERT_GROUP_MEMBER(TYPE, GROUP, MEMBER) \
100 	BUILD_BUG_ON(!(offsetof(TYPE, MEMBER) >= \
101 		       offsetofend(TYPE, __cacheline_group_begin__##GROUP) && \
102 		       offsetofend(TYPE, MEMBER) <= \
103 		       offsetof(TYPE, __cacheline_group_end__##GROUP)))
104 #endif
105 
106 #ifndef CACHELINE_ASSERT_GROUP_SIZE
107 #define CACHELINE_ASSERT_GROUP_SIZE(TYPE, GROUP, SIZE) \
108 	BUILD_BUG_ON(offsetof(TYPE, __cacheline_group_end__##GROUP) - \
109 		     offsetofend(TYPE, __cacheline_group_begin__##GROUP) > \
110 		     SIZE)
111 #endif
112 
113 /*
114  * Helper to add padding within a struct to ensure data fall into separate
115  * cachelines.
116  */
117 #if defined(CONFIG_SMP)
118 struct cacheline_padding {
119 	char x[0];
120 } ____cacheline_internodealigned_in_smp;
121 #define CACHELINE_PADDING(name)		struct cacheline_padding name
122 #else
123 #define CACHELINE_PADDING(name)
124 #endif
125 
126 #ifdef ARCH_DMA_MINALIGN
127 #define ARCH_HAS_DMA_MINALIGN
128 #else
129 #define ARCH_DMA_MINALIGN __alignof__(unsigned long long)
130 #endif
131 
132 #endif /* __LINUX_CACHE_H */
133