xref: /linux/include/linux/shrinker.h (revision 0ea5c948cb64bab5bc7a5516774eb8536f05aa0d)
1 /* SPDX-License-Identifier: GPL-2.0 */
2 #ifndef _LINUX_SHRINKER_H
3 #define _LINUX_SHRINKER_H
4 
5 #include <linux/atomic.h>
6 #include <linux/types.h>
7 #include <linux/refcount.h>
8 #include <linux/completion.h>
9 
10 #define SHRINKER_UNIT_BITS	BITS_PER_LONG
11 
12 /*
13  * Bitmap and deferred work of shrinker::id corresponding to memcg-aware
14  * shrinkers, which have elements charged to the memcg.
15  */
16 struct shrinker_info_unit {
17 	atomic_long_t nr_deferred[SHRINKER_UNIT_BITS];
18 	DECLARE_BITMAP(map, SHRINKER_UNIT_BITS);
19 };
20 
21 struct shrinker_info {
22 	struct rcu_head rcu;
23 	int map_nr_max;
24 	struct shrinker_info_unit *unit[];
25 };
26 
27 /*
28  * This struct is used to pass information from page reclaim to the shrinkers.
29  * We consolidate the values for easier extension later.
30  *
31  * The 'gfpmask' refers to the allocation we are currently trying to
32  * fulfil.
33  */
34 struct shrink_control {
35 	gfp_t gfp_mask;
36 
37 	/* current node being shrunk (for NUMA aware shrinkers) */
38 	int nid;
39 
40 	/*
41 	 * How many objects scan_objects should scan and try to reclaim.
42 	 * This is reset before every call, so it is safe for callees
43 	 * to modify.
44 	 */
45 	unsigned long nr_to_scan;
46 
47 	/*
48 	 * How many objects did scan_objects process?
49 	 * This defaults to nr_to_scan before every call, but the callee
50 	 * should track its actual progress.
51 	 */
52 	unsigned long nr_scanned;
53 
54 	/* current memcg being shrunk (for memcg aware shrinkers) */
55 	struct mem_cgroup *memcg;
56 };
57 
58 #define SHRINK_STOP (~0UL)
59 #define SHRINK_EMPTY (~0UL - 1)
60 /*
61  * A callback you can register to apply pressure to ageable caches.
62  *
63  * @count_objects should return the number of freeable items in the cache. If
64  * there are no objects to free, it should return SHRINK_EMPTY, while 0 is
65  * returned in cases of the number of freeable items cannot be determined
66  * or shrinker should skip this cache for this time (e.g., their number
67  * is below shrinkable limit). No deadlock checks should be done during the
68  * count callback - the shrinker relies on aggregating scan counts that couldn't
69  * be executed due to potential deadlocks to be run at a later call when the
70  * deadlock condition is no longer pending.
71  *
72  * @scan_objects will only be called if @count_objects returned a non-zero
73  * value for the number of freeable objects. The callout should scan the cache
74  * and attempt to free items from the cache. It should then return the number
75  * of objects freed during the scan, or SHRINK_STOP if progress cannot be made
76  * due to potential deadlocks. If SHRINK_STOP is returned, then no further
77  * attempts to call the @scan_objects will be made from the current reclaim
78  * context.
79  *
80  * @flags determine the shrinker abilities, like numa awareness
81  */
82 struct shrinker {
83 	unsigned long (*count_objects)(struct shrinker *,
84 				       struct shrink_control *sc);
85 	unsigned long (*scan_objects)(struct shrinker *,
86 				      struct shrink_control *sc);
87 
88 	long batch;	/* reclaim batch size, 0 = default */
89 	int seeks;	/* seeks to recreate an obj */
90 	unsigned flags;
91 
92 	/*
93 	 * The reference count of this shrinker. Registered shrinker have an
94 	 * initial refcount of 1, then the lookup operations are now allowed
95 	 * to use it via shrinker_try_get(). Later in the unregistration step,
96 	 * the initial refcount will be discarded, and will free the shrinker
97 	 * asynchronously via RCU after its refcount reaches 0.
98 	 */
99 	refcount_t refcount;
100 	struct completion done;	/* use to wait for refcount to reach 0 */
101 	struct rcu_head rcu;
102 
103 	void *private_data;
104 
105 	/* These are for internal use */
106 	struct list_head list;
107 #ifdef CONFIG_MEMCG
108 	/* ID in shrinker_idr */
109 	int id;
110 #endif
111 #ifdef CONFIG_SHRINKER_DEBUG
112 	int debugfs_id;
113 	const char *name;
114 	struct dentry *debugfs_entry;
115 #endif
116 	/* objs pending delete, per node */
117 	atomic_long_t *nr_deferred;
118 };
119 #define DEFAULT_SEEKS 2 /* A good number if you don't know better. */
120 
121 /* Internal flags */
122 #define SHRINKER_REGISTERED	BIT(0)
123 #define SHRINKER_ALLOCATED	BIT(1)
124 
125 /* Flags for users to use */
126 #define SHRINKER_NUMA_AWARE	BIT(2)
127 #define SHRINKER_MEMCG_AWARE	BIT(3)
128 /*
129  * It just makes sense when the shrinker is also MEMCG_AWARE for now,
130  * non-MEMCG_AWARE shrinker should not have this flag set.
131  */
132 #define SHRINKER_NONSLAB	BIT(4)
133 
134 __printf(2, 3)
135 struct shrinker *shrinker_alloc(unsigned int flags, const char *fmt, ...);
136 void shrinker_register(struct shrinker *shrinker);
137 void shrinker_free(struct shrinker *shrinker);
138 
shrinker_try_get(struct shrinker * shrinker)139 static inline bool shrinker_try_get(struct shrinker *shrinker)
140 {
141 	return refcount_inc_not_zero(&shrinker->refcount);
142 }
143 
shrinker_put(struct shrinker * shrinker)144 static inline void shrinker_put(struct shrinker *shrinker)
145 {
146 	if (refcount_dec_and_test(&shrinker->refcount))
147 		complete(&shrinker->done);
148 }
149 
150 #ifdef CONFIG_SHRINKER_DEBUG
151 extern int __printf(2, 3) shrinker_debugfs_rename(struct shrinker *shrinker,
152 						  const char *fmt, ...);
153 #else /* CONFIG_SHRINKER_DEBUG */
154 static inline __printf(2, 3)
shrinker_debugfs_rename(struct shrinker * shrinker,const char * fmt,...)155 int shrinker_debugfs_rename(struct shrinker *shrinker, const char *fmt, ...)
156 {
157 	return 0;
158 }
159 #endif /* CONFIG_SHRINKER_DEBUG */
160 #endif /* _LINUX_SHRINKER_H */
161