xref: /linux/fs/xfs/scrub/xfile.c (revision c532de5a67a70f8533d495f8f2aaa9a0491c3ad0)
1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3  * Copyright (C) 2018-2023 Oracle.  All Rights Reserved.
4  * Author: Darrick J. Wong <djwong@kernel.org>
5  */
6 #include "xfs.h"
7 #include "xfs_fs.h"
8 #include "xfs_shared.h"
9 #include "xfs_format.h"
10 #include "xfs_log_format.h"
11 #include "xfs_trans_resv.h"
12 #include "xfs_mount.h"
13 #include "scrub/scrub.h"
14 #include "scrub/xfile.h"
15 #include "scrub/xfarray.h"
16 #include "scrub/trace.h"
17 #include <linux/shmem_fs.h>
18 
19 /*
20  * Swappable Temporary Memory
21  * ==========================
22  *
23  * Online checking sometimes needs to be able to stage a large amount of data
24  * in memory.  This information might not fit in the available memory and it
25  * doesn't all need to be accessible at all times.  In other words, we want an
26  * indexed data buffer to store data that can be paged out.
27  *
28  * When CONFIG_TMPFS=y, shmemfs is enough of a filesystem to meet those
29  * requirements.  Therefore, the xfile mechanism uses an unlinked shmem file to
30  * store our staging data.  This file is not installed in the file descriptor
31  * table so that user programs cannot access the data, which means that the
32  * xfile must be freed with xfile_destroy.
33  *
34  * xfiles assume that the caller will handle all required concurrency
35  * management; standard vfs locks (freezer and inode) are not taken.  Reads
36  * and writes are satisfied directly from the page cache.
37  */
38 
39 /*
40  * xfiles must not be exposed to userspace and require upper layers to
41  * coordinate access to the one handle returned by the constructor, so
42  * establish a separate lock class for xfiles to avoid confusing lockdep.
43  */
44 static struct lock_class_key xfile_i_mutex_key;
45 
46 /*
47  * Create an xfile of the given size.  The description will be used in the
48  * trace output.
49  */
50 int
51 xfile_create(
52 	const char		*description,
53 	loff_t			isize,
54 	struct xfile		**xfilep)
55 {
56 	struct inode		*inode;
57 	struct xfile		*xf;
58 	int			error;
59 
60 	xf = kmalloc(sizeof(struct xfile), XCHK_GFP_FLAGS);
61 	if (!xf)
62 		return -ENOMEM;
63 
64 	xf->file = shmem_kernel_file_setup(description, isize, VM_NORESERVE);
65 	if (IS_ERR(xf->file)) {
66 		error = PTR_ERR(xf->file);
67 		goto out_xfile;
68 	}
69 
70 	inode = file_inode(xf->file);
71 	lockdep_set_class(&inode->i_rwsem, &xfile_i_mutex_key);
72 
73 	/*
74 	 * We don't want to bother with kmapping data during repair, so don't
75 	 * allow highmem pages to back this mapping.
76 	 */
77 	mapping_set_gfp_mask(inode->i_mapping, GFP_KERNEL);
78 
79 	trace_xfile_create(xf);
80 
81 	*xfilep = xf;
82 	return 0;
83 out_xfile:
84 	kfree(xf);
85 	return error;
86 }
87 
88 /* Close the file and release all resources. */
89 void
90 xfile_destroy(
91 	struct xfile		*xf)
92 {
93 	struct inode		*inode = file_inode(xf->file);
94 
95 	trace_xfile_destroy(xf);
96 
97 	lockdep_set_class(&inode->i_rwsem, &inode->i_sb->s_type->i_mutex_key);
98 	fput(xf->file);
99 	kfree(xf);
100 }
101 
102 /*
103  * Load an object.  Since we're treating this file as "memory", any error or
104  * short IO is treated as a failure to allocate memory.
105  */
106 int
107 xfile_load(
108 	struct xfile		*xf,
109 	void			*buf,
110 	size_t			count,
111 	loff_t			pos)
112 {
113 	struct inode		*inode = file_inode(xf->file);
114 	unsigned int		pflags;
115 
116 	if (count > MAX_RW_COUNT)
117 		return -ENOMEM;
118 	if (inode->i_sb->s_maxbytes - pos < count)
119 		return -ENOMEM;
120 
121 	trace_xfile_load(xf, pos, count);
122 
123 	pflags = memalloc_nofs_save();
124 	while (count > 0) {
125 		struct folio	*folio;
126 		unsigned int	len;
127 		unsigned int	offset;
128 
129 		if (shmem_get_folio(inode, pos >> PAGE_SHIFT, 0, &folio,
130 				SGP_READ) < 0)
131 			break;
132 		if (!folio) {
133 			/*
134 			 * No data stored at this offset, just zero the output
135 			 * buffer until the next page boundary.
136 			 */
137 			len = min_t(ssize_t, count,
138 				PAGE_SIZE - offset_in_page(pos));
139 			memset(buf, 0, len);
140 		} else {
141 			if (filemap_check_wb_err(inode->i_mapping, 0)) {
142 				folio_unlock(folio);
143 				folio_put(folio);
144 				break;
145 			}
146 
147 			offset = offset_in_folio(folio, pos);
148 			len = min_t(ssize_t, count, folio_size(folio) - offset);
149 			memcpy(buf, folio_address(folio) + offset, len);
150 
151 			folio_unlock(folio);
152 			folio_put(folio);
153 		}
154 		count -= len;
155 		pos += len;
156 		buf += len;
157 	}
158 	memalloc_nofs_restore(pflags);
159 
160 	if (count)
161 		return -ENOMEM;
162 	return 0;
163 }
164 
165 /*
166  * Store an object.  Since we're treating this file as "memory", any error or
167  * short IO is treated as a failure to allocate memory.
168  */
169 int
170 xfile_store(
171 	struct xfile		*xf,
172 	const void		*buf,
173 	size_t			count,
174 	loff_t			pos)
175 {
176 	struct inode		*inode = file_inode(xf->file);
177 	unsigned int		pflags;
178 
179 	if (count > MAX_RW_COUNT)
180 		return -ENOMEM;
181 	if (inode->i_sb->s_maxbytes - pos < count)
182 		return -ENOMEM;
183 
184 	trace_xfile_store(xf, pos, count);
185 
186 	/*
187 	 * Increase the file size first so that shmem_get_folio(..., SGP_CACHE),
188 	 * actually allocates a folio instead of erroring out.
189 	 */
190 	if (pos + count > i_size_read(inode))
191 		i_size_write(inode, pos + count);
192 
193 	pflags = memalloc_nofs_save();
194 	while (count > 0) {
195 		struct folio	*folio;
196 		unsigned int	len;
197 		unsigned int	offset;
198 
199 		if (shmem_get_folio(inode, pos >> PAGE_SHIFT, 0, &folio,
200 				SGP_CACHE) < 0)
201 			break;
202 		if (filemap_check_wb_err(inode->i_mapping, 0)) {
203 			folio_unlock(folio);
204 			folio_put(folio);
205 			break;
206 		}
207 
208 		offset = offset_in_folio(folio, pos);
209 		len = min_t(ssize_t, count, folio_size(folio) - offset);
210 		memcpy(folio_address(folio) + offset, buf, len);
211 
212 		folio_mark_dirty(folio);
213 		folio_unlock(folio);
214 		folio_put(folio);
215 
216 		count -= len;
217 		pos += len;
218 		buf += len;
219 	}
220 	memalloc_nofs_restore(pflags);
221 
222 	if (count)
223 		return -ENOMEM;
224 	return 0;
225 }
226 
227 /* Find the next written area in the xfile data for a given offset. */
228 loff_t
229 xfile_seek_data(
230 	struct xfile		*xf,
231 	loff_t			pos)
232 {
233 	loff_t			ret;
234 
235 	ret = vfs_llseek(xf->file, pos, SEEK_DATA);
236 	trace_xfile_seek_data(xf, pos, ret);
237 	return ret;
238 }
239 
240 /*
241  * Grab the (locked) folio for a memory object.  The object cannot span a folio
242  * boundary.  Returns the locked folio if successful, NULL if there was no
243  * folio or it didn't cover the range requested, or an ERR_PTR on failure.
244  */
245 struct folio *
246 xfile_get_folio(
247 	struct xfile		*xf,
248 	loff_t			pos,
249 	size_t			len,
250 	unsigned int		flags)
251 {
252 	struct inode		*inode = file_inode(xf->file);
253 	struct folio		*folio = NULL;
254 	unsigned int		pflags;
255 	int			error;
256 
257 	if (inode->i_sb->s_maxbytes - pos < len)
258 		return ERR_PTR(-ENOMEM);
259 
260 	trace_xfile_get_folio(xf, pos, len);
261 
262 	/*
263 	 * Increase the file size first so that shmem_get_folio(..., SGP_CACHE),
264 	 * actually allocates a folio instead of erroring out.
265 	 */
266 	if ((flags & XFILE_ALLOC) && pos + len > i_size_read(inode))
267 		i_size_write(inode, pos + len);
268 
269 	pflags = memalloc_nofs_save();
270 	error = shmem_get_folio(inode, pos >> PAGE_SHIFT, 0, &folio,
271 			(flags & XFILE_ALLOC) ? SGP_CACHE : SGP_READ);
272 	memalloc_nofs_restore(pflags);
273 	if (error)
274 		return ERR_PTR(error);
275 
276 	if (!folio)
277 		return NULL;
278 
279 	if (len > folio_size(folio) - offset_in_folio(folio, pos)) {
280 		folio_unlock(folio);
281 		folio_put(folio);
282 		return NULL;
283 	}
284 
285 	if (filemap_check_wb_err(inode->i_mapping, 0)) {
286 		folio_unlock(folio);
287 		folio_put(folio);
288 		return ERR_PTR(-EIO);
289 	}
290 
291 	/*
292 	 * Mark the folio dirty so that it won't be reclaimed once we drop the
293 	 * (potentially last) reference in xfile_put_folio.
294 	 */
295 	if (flags & XFILE_ALLOC)
296 		folio_mark_dirty(folio);
297 	return folio;
298 }
299 
300 /*
301  * Release the (locked) folio for a memory object.
302  */
303 void
304 xfile_put_folio(
305 	struct xfile		*xf,
306 	struct folio		*folio)
307 {
308 	trace_xfile_put_folio(xf, folio_pos(folio), folio_size(folio));
309 
310 	folio_unlock(folio);
311 	folio_put(folio);
312 }
313 
314 /* Discard the page cache that's backing a range of the xfile. */
315 void
316 xfile_discard(
317 	struct xfile		*xf,
318 	loff_t			pos,
319 	u64			count)
320 {
321 	trace_xfile_discard(xf, pos, count);
322 
323 	shmem_truncate_range(file_inode(xf->file), pos, pos + count - 1);
324 }
325