1b2441318SGreg Kroah-Hartman /* SPDX-License-Identifier: GPL-2.0 */ 21da177e4SLinus Torvalds #ifndef __LINUX_CACHE_H 31da177e4SLinus Torvalds #define __LINUX_CACHE_H 41da177e4SLinus Torvalds 5c28aa1f0SJoe Perches #include <uapi/linux/kernel.h> 61da177e4SLinus Torvalds #include <asm/cache.h> 71da177e4SLinus Torvalds 81da177e4SLinus Torvalds #ifndef L1_CACHE_ALIGN 9c28aa1f0SJoe Perches #define L1_CACHE_ALIGN(x) __ALIGN_KERNEL(x, L1_CACHE_BYTES) 101da177e4SLinus Torvalds #endif 111da177e4SLinus Torvalds 121da177e4SLinus Torvalds #ifndef SMP_CACHE_BYTES 131da177e4SLinus Torvalds #define SMP_CACHE_BYTES L1_CACHE_BYTES 141da177e4SLinus Torvalds #endif 151da177e4SLinus Torvalds 16*2cb13decSAlexander Lobakin /** 17*2cb13decSAlexander Lobakin * SMP_CACHE_ALIGN - align a value to the L2 cacheline size 18*2cb13decSAlexander Lobakin * @x: value to align 19*2cb13decSAlexander Lobakin * 20*2cb13decSAlexander Lobakin * On some architectures, L2 ("SMP") CL size is bigger than L1, and sometimes, 21*2cb13decSAlexander Lobakin * this needs to be accounted. 22*2cb13decSAlexander Lobakin * 23*2cb13decSAlexander Lobakin * Return: aligned value. 24*2cb13decSAlexander Lobakin */ 25*2cb13decSAlexander Lobakin #ifndef SMP_CACHE_ALIGN 26*2cb13decSAlexander Lobakin #define SMP_CACHE_ALIGN(x) ALIGN(x, SMP_CACHE_BYTES) 27*2cb13decSAlexander Lobakin #endif 28*2cb13decSAlexander Lobakin 29*2cb13decSAlexander Lobakin /* 30*2cb13decSAlexander Lobakin * ``__aligned_largest`` aligns a field to the value most optimal for the 31*2cb13decSAlexander Lobakin * target architecture to perform memory operations. Get the actual value 32*2cb13decSAlexander Lobakin * to be able to use it anywhere else. 33*2cb13decSAlexander Lobakin */ 34*2cb13decSAlexander Lobakin #ifndef __LARGEST_ALIGN 35*2cb13decSAlexander Lobakin #define __LARGEST_ALIGN sizeof(struct { long x; } __aligned_largest) 36*2cb13decSAlexander Lobakin #endif 37*2cb13decSAlexander Lobakin 38*2cb13decSAlexander Lobakin #ifndef LARGEST_ALIGN 39*2cb13decSAlexander Lobakin #define LARGEST_ALIGN(x) ALIGN(x, __LARGEST_ALIGN) 40*2cb13decSAlexander Lobakin #endif 41*2cb13decSAlexander Lobakin 42c74ba8b3SKees Cook /* 43c74ba8b3SKees Cook * __read_mostly is used to keep rarely changing variables out of frequently 444fa72523SLuis Chamberlain * updated cachelines. Its use should be reserved for data that is used 454fa72523SLuis Chamberlain * frequently in hot paths. Performance traces can help decide when to use 464fa72523SLuis Chamberlain * this. You want __read_mostly data to be tightly packed, so that in the 474fa72523SLuis Chamberlain * best case multiple frequently read variables for a hot path will be next 484fa72523SLuis Chamberlain * to each other in order to reduce the number of cachelines needed to 494fa72523SLuis Chamberlain * execute a critical path. We should be mindful and selective of its use. 504fa72523SLuis Chamberlain * ie: if you're going to use it please supply a *good* justification in your 514fa72523SLuis Chamberlain * commit log 52c74ba8b3SKees Cook */ 53804f1594SKyle McMartin #ifndef __read_mostly 546c036527SChristoph Lameter #define __read_mostly 556c036527SChristoph Lameter #endif 566c036527SChristoph Lameter 57c74ba8b3SKees Cook /* 58c74ba8b3SKees Cook * __ro_after_init is used to mark things that are read-only after init (i.e. 59c74ba8b3SKees Cook * after mark_rodata_ro() has been called). These are effectively read-only, 60c74ba8b3SKees Cook * but may get written to during init, so can't live in .rodata (via "const"). 61c74ba8b3SKees Cook */ 62c74ba8b3SKees Cook #ifndef __ro_after_init 6333def849SJoe Perches #define __ro_after_init __section(".data..ro_after_init") 64c74ba8b3SKees Cook #endif 65c74ba8b3SKees Cook 661da177e4SLinus Torvalds #ifndef ____cacheline_aligned 671da177e4SLinus Torvalds #define ____cacheline_aligned __attribute__((__aligned__(SMP_CACHE_BYTES))) 681da177e4SLinus Torvalds #endif 691da177e4SLinus Torvalds 701da177e4SLinus Torvalds #ifndef ____cacheline_aligned_in_smp 711da177e4SLinus Torvalds #ifdef CONFIG_SMP 721da177e4SLinus Torvalds #define ____cacheline_aligned_in_smp ____cacheline_aligned 731da177e4SLinus Torvalds #else 741da177e4SLinus Torvalds #define ____cacheline_aligned_in_smp 751da177e4SLinus Torvalds #endif /* CONFIG_SMP */ 761da177e4SLinus Torvalds #endif 771da177e4SLinus Torvalds 781da177e4SLinus Torvalds #ifndef __cacheline_aligned 791da177e4SLinus Torvalds #define __cacheline_aligned \ 801da177e4SLinus Torvalds __attribute__((__aligned__(SMP_CACHE_BYTES), \ 814af57b78STim Abbott __section__(".data..cacheline_aligned"))) 821da177e4SLinus Torvalds #endif /* __cacheline_aligned */ 831da177e4SLinus Torvalds 841da177e4SLinus Torvalds #ifndef __cacheline_aligned_in_smp 851da177e4SLinus Torvalds #ifdef CONFIG_SMP 861da177e4SLinus Torvalds #define __cacheline_aligned_in_smp __cacheline_aligned 871da177e4SLinus Torvalds #else 881da177e4SLinus Torvalds #define __cacheline_aligned_in_smp 891da177e4SLinus Torvalds #endif /* CONFIG_SMP */ 901da177e4SLinus Torvalds #endif 911da177e4SLinus Torvalds 9222fc6eccSRavikiran G Thirumalai /* 9322fc6eccSRavikiran G Thirumalai * The maximum alignment needed for some critical structures 9422fc6eccSRavikiran G Thirumalai * These could be inter-node cacheline sizes/L3 cacheline 9522fc6eccSRavikiran G Thirumalai * size etc. Define this in asm/cache.h for your arch 9622fc6eccSRavikiran G Thirumalai */ 9722fc6eccSRavikiran G Thirumalai #ifndef INTERNODE_CACHE_SHIFT 9822fc6eccSRavikiran G Thirumalai #define INTERNODE_CACHE_SHIFT L1_CACHE_SHIFT 9922fc6eccSRavikiran G Thirumalai #endif 10022fc6eccSRavikiran G Thirumalai 10122fc6eccSRavikiran G Thirumalai #if !defined(____cacheline_internodealigned_in_smp) 1021da177e4SLinus Torvalds #if defined(CONFIG_SMP) 10322fc6eccSRavikiran G Thirumalai #define ____cacheline_internodealigned_in_smp \ 10422fc6eccSRavikiran G Thirumalai __attribute__((__aligned__(1 << (INTERNODE_CACHE_SHIFT)))) 1051da177e4SLinus Torvalds #else 10622fc6eccSRavikiran G Thirumalai #define ____cacheline_internodealigned_in_smp 1071da177e4SLinus Torvalds #endif 1081da177e4SLinus Torvalds #endif 1091da177e4SLinus Torvalds 1101b27d05bSPekka Enberg #ifndef CONFIG_ARCH_HAS_CACHE_LINE_SIZE 1111b27d05bSPekka Enberg #define cache_line_size() L1_CACHE_BYTES 1121b27d05bSPekka Enberg #endif 1131b27d05bSPekka Enberg 114aeb9ce05SCoco Li #ifndef __cacheline_group_begin 115aeb9ce05SCoco Li #define __cacheline_group_begin(GROUP) \ 116aeb9ce05SCoco Li __u8 __cacheline_group_begin__##GROUP[0] 117aeb9ce05SCoco Li #endif 118aeb9ce05SCoco Li 119aeb9ce05SCoco Li #ifndef __cacheline_group_end 120aeb9ce05SCoco Li #define __cacheline_group_end(GROUP) \ 121aeb9ce05SCoco Li __u8 __cacheline_group_end__##GROUP[0] 122aeb9ce05SCoco Li #endif 123aeb9ce05SCoco Li 124*2cb13decSAlexander Lobakin /** 125*2cb13decSAlexander Lobakin * __cacheline_group_begin_aligned - declare an aligned group start 126*2cb13decSAlexander Lobakin * @GROUP: name of the group 127*2cb13decSAlexander Lobakin * @...: optional group alignment 128*2cb13decSAlexander Lobakin * 129*2cb13decSAlexander Lobakin * The following block inside a struct: 130*2cb13decSAlexander Lobakin * 131*2cb13decSAlexander Lobakin * __cacheline_group_begin_aligned(grp); 132*2cb13decSAlexander Lobakin * field a; 133*2cb13decSAlexander Lobakin * field b; 134*2cb13decSAlexander Lobakin * __cacheline_group_end_aligned(grp); 135*2cb13decSAlexander Lobakin * 136*2cb13decSAlexander Lobakin * will always be aligned to either the specified alignment or 137*2cb13decSAlexander Lobakin * ``SMP_CACHE_BYTES``. 138*2cb13decSAlexander Lobakin */ 139*2cb13decSAlexander Lobakin #define __cacheline_group_begin_aligned(GROUP, ...) \ 140*2cb13decSAlexander Lobakin __cacheline_group_begin(GROUP) \ 141*2cb13decSAlexander Lobakin __aligned((__VA_ARGS__ + 0) ? : SMP_CACHE_BYTES) 142*2cb13decSAlexander Lobakin 143*2cb13decSAlexander Lobakin /** 144*2cb13decSAlexander Lobakin * __cacheline_group_end_aligned - declare an aligned group end 145*2cb13decSAlexander Lobakin * @GROUP: name of the group 146*2cb13decSAlexander Lobakin * @...: optional alignment (same as was in __cacheline_group_begin_aligned()) 147*2cb13decSAlexander Lobakin * 148*2cb13decSAlexander Lobakin * Note that the end marker is aligned to sizeof(long) to allow more precise 149*2cb13decSAlexander Lobakin * size assertion. It also declares a padding at the end to avoid next field 150*2cb13decSAlexander Lobakin * falling into this cacheline. 151*2cb13decSAlexander Lobakin */ 152*2cb13decSAlexander Lobakin #define __cacheline_group_end_aligned(GROUP, ...) \ 153*2cb13decSAlexander Lobakin __cacheline_group_end(GROUP) __aligned(sizeof(long)); \ 154*2cb13decSAlexander Lobakin struct { } __cacheline_group_pad__##GROUP \ 155*2cb13decSAlexander Lobakin __aligned((__VA_ARGS__ + 0) ? : SMP_CACHE_BYTES) 156*2cb13decSAlexander Lobakin 157aeb9ce05SCoco Li #ifndef CACHELINE_ASSERT_GROUP_MEMBER 158aeb9ce05SCoco Li #define CACHELINE_ASSERT_GROUP_MEMBER(TYPE, GROUP, MEMBER) \ 159aeb9ce05SCoco Li BUILD_BUG_ON(!(offsetof(TYPE, MEMBER) >= \ 160aeb9ce05SCoco Li offsetofend(TYPE, __cacheline_group_begin__##GROUP) && \ 161aeb9ce05SCoco Li offsetofend(TYPE, MEMBER) <= \ 162aeb9ce05SCoco Li offsetof(TYPE, __cacheline_group_end__##GROUP))) 163aeb9ce05SCoco Li #endif 164aeb9ce05SCoco Li 165aeb9ce05SCoco Li #ifndef CACHELINE_ASSERT_GROUP_SIZE 166aeb9ce05SCoco Li #define CACHELINE_ASSERT_GROUP_SIZE(TYPE, GROUP, SIZE) \ 167aeb9ce05SCoco Li BUILD_BUG_ON(offsetof(TYPE, __cacheline_group_end__##GROUP) - \ 168aeb9ce05SCoco Li offsetofend(TYPE, __cacheline_group_begin__##GROUP) > \ 169aeb9ce05SCoco Li SIZE) 170aeb9ce05SCoco Li #endif 171aeb9ce05SCoco Li 172e6ad640bSShakeel Butt /* 173e6ad640bSShakeel Butt * Helper to add padding within a struct to ensure data fall into separate 174e6ad640bSShakeel Butt * cachelines. 175e6ad640bSShakeel Butt */ 176e6ad640bSShakeel Butt #if defined(CONFIG_SMP) 177e6ad640bSShakeel Butt struct cacheline_padding { 178e6ad640bSShakeel Butt char x[0]; 179e6ad640bSShakeel Butt } ____cacheline_internodealigned_in_smp; 180e6ad640bSShakeel Butt #define CACHELINE_PADDING(name) struct cacheline_padding name 181e6ad640bSShakeel Butt #else 182e6ad640bSShakeel Butt #define CACHELINE_PADDING(name) 183e6ad640bSShakeel Butt #endif 184e6ad640bSShakeel Butt 1854ab5f8ecSCatalin Marinas #ifdef ARCH_DMA_MINALIGN 1864ab5f8ecSCatalin Marinas #define ARCH_HAS_DMA_MINALIGN 1874ab5f8ecSCatalin Marinas #else 1884ab5f8ecSCatalin Marinas #define ARCH_DMA_MINALIGN __alignof__(unsigned long long) 1894ab5f8ecSCatalin Marinas #endif 1904ab5f8ecSCatalin Marinas 1911da177e4SLinus Torvalds #endif /* __LINUX_CACHE_H */ 192