xref: /linux/fs/overlayfs/copy_up.c (revision 7f71507851fc7764b36a3221839607d3a45c2025)
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  *
4  * Copyright (C) 2011 Novell Inc.
5  */
6 
7 #include <linux/module.h>
8 #include <linux/fs.h>
9 #include <linux/slab.h>
10 #include <linux/file.h>
11 #include <linux/fileattr.h>
12 #include <linux/splice.h>
13 #include <linux/xattr.h>
14 #include <linux/security.h>
15 #include <linux/uaccess.h>
16 #include <linux/sched/signal.h>
17 #include <linux/cred.h>
18 #include <linux/namei.h>
19 #include <linux/ratelimit.h>
20 #include <linux/exportfs.h>
21 #include "overlayfs.h"
22 
23 #define OVL_COPY_UP_CHUNK_SIZE (1 << 20)
24 
25 static int ovl_ccup_set(const char *buf, const struct kernel_param *param)
26 {
27 	pr_warn("\"check_copy_up\" module option is obsolete\n");
28 	return 0;
29 }
30 
31 static int ovl_ccup_get(char *buf, const struct kernel_param *param)
32 {
33 	return sprintf(buf, "N\n");
34 }
35 
36 module_param_call(check_copy_up, ovl_ccup_set, ovl_ccup_get, NULL, 0644);
37 MODULE_PARM_DESC(check_copy_up, "Obsolete; does nothing");
38 
39 static bool ovl_must_copy_xattr(const char *name)
40 {
41 	return !strcmp(name, XATTR_POSIX_ACL_ACCESS) ||
42 	       !strcmp(name, XATTR_POSIX_ACL_DEFAULT) ||
43 	       !strncmp(name, XATTR_SECURITY_PREFIX, XATTR_SECURITY_PREFIX_LEN);
44 }
45 
46 static int ovl_copy_acl(struct ovl_fs *ofs, const struct path *path,
47 			struct dentry *dentry, const char *acl_name)
48 {
49 	int err;
50 	struct posix_acl *clone, *real_acl = NULL;
51 
52 	real_acl = ovl_get_acl_path(path, acl_name, false);
53 	if (!real_acl)
54 		return 0;
55 
56 	if (IS_ERR(real_acl)) {
57 		err = PTR_ERR(real_acl);
58 		if (err == -ENODATA || err == -EOPNOTSUPP)
59 			return 0;
60 		return err;
61 	}
62 
63 	clone = posix_acl_clone(real_acl, GFP_KERNEL);
64 	posix_acl_release(real_acl); /* release original acl */
65 	if (!clone)
66 		return -ENOMEM;
67 
68 	err = ovl_do_set_acl(ofs, dentry, acl_name, clone);
69 
70 	/* release cloned acl */
71 	posix_acl_release(clone);
72 	return err;
73 }
74 
75 int ovl_copy_xattr(struct super_block *sb, const struct path *oldpath, struct dentry *new)
76 {
77 	struct dentry *old = oldpath->dentry;
78 	ssize_t list_size, size, value_size = 0;
79 	char *buf, *name, *value = NULL;
80 	int error = 0;
81 	size_t slen;
82 
83 	if (!old->d_inode->i_op->listxattr || !new->d_inode->i_op->listxattr)
84 		return 0;
85 
86 	list_size = vfs_listxattr(old, NULL, 0);
87 	if (list_size <= 0) {
88 		if (list_size == -EOPNOTSUPP)
89 			return 0;
90 		return list_size;
91 	}
92 
93 	buf = kvzalloc(list_size, GFP_KERNEL);
94 	if (!buf)
95 		return -ENOMEM;
96 
97 	list_size = vfs_listxattr(old, buf, list_size);
98 	if (list_size <= 0) {
99 		error = list_size;
100 		goto out;
101 	}
102 
103 	for (name = buf; list_size; name += slen) {
104 		slen = strnlen(name, list_size) + 1;
105 
106 		/* underlying fs providing us with an broken xattr list? */
107 		if (WARN_ON(slen > list_size)) {
108 			error = -EIO;
109 			break;
110 		}
111 		list_size -= slen;
112 
113 		if (ovl_is_private_xattr(sb, name))
114 			continue;
115 
116 		error = security_inode_copy_up_xattr(old, name);
117 		if (error == -ECANCELED) {
118 			error = 0;
119 			continue; /* Discard */
120 		}
121 		if (error < 0 && error != -EOPNOTSUPP)
122 			break;
123 
124 		if (is_posix_acl_xattr(name)) {
125 			error = ovl_copy_acl(OVL_FS(sb), oldpath, new, name);
126 			if (!error)
127 				continue;
128 			/* POSIX ACLs must be copied. */
129 			break;
130 		}
131 
132 retry:
133 		size = ovl_do_getxattr(oldpath, name, value, value_size);
134 		if (size == -ERANGE)
135 			size = ovl_do_getxattr(oldpath, name, NULL, 0);
136 
137 		if (size < 0) {
138 			error = size;
139 			break;
140 		}
141 
142 		if (size > value_size) {
143 			void *new;
144 
145 			new = kvmalloc(size, GFP_KERNEL);
146 			if (!new) {
147 				error = -ENOMEM;
148 				break;
149 			}
150 			kvfree(value);
151 			value = new;
152 			value_size = size;
153 			goto retry;
154 		}
155 
156 		error = ovl_do_setxattr(OVL_FS(sb), new, name, value, size, 0);
157 		if (error) {
158 			if (error != -EOPNOTSUPP || ovl_must_copy_xattr(name))
159 				break;
160 
161 			/* Ignore failure to copy unknown xattrs */
162 			error = 0;
163 		}
164 	}
165 	kvfree(value);
166 out:
167 	kvfree(buf);
168 	return error;
169 }
170 
171 static int ovl_copy_fileattr(struct inode *inode, const struct path *old,
172 			     const struct path *new)
173 {
174 	struct fileattr oldfa = { .flags_valid = true };
175 	struct fileattr newfa = { .flags_valid = true };
176 	int err;
177 
178 	err = ovl_real_fileattr_get(old, &oldfa);
179 	if (err) {
180 		/* Ntfs-3g returns -EINVAL for "no fileattr support" */
181 		if (err == -ENOTTY || err == -EINVAL)
182 			return 0;
183 		pr_warn("failed to retrieve lower fileattr (%pd2, err=%i)\n",
184 			old->dentry, err);
185 		return err;
186 	}
187 
188 	/*
189 	 * We cannot set immutable and append-only flags on upper inode,
190 	 * because we would not be able to link upper inode to upper dir
191 	 * not set overlay private xattr on upper inode.
192 	 * Store these flags in overlay.protattr xattr instead.
193 	 */
194 	if (oldfa.flags & OVL_PROT_FS_FLAGS_MASK) {
195 		err = ovl_set_protattr(inode, new->dentry, &oldfa);
196 		if (err == -EPERM)
197 			pr_warn_once("copying fileattr: no xattr on upper\n");
198 		else if (err)
199 			return err;
200 	}
201 
202 	/* Don't bother copying flags if none are set */
203 	if (!(oldfa.flags & OVL_COPY_FS_FLAGS_MASK))
204 		return 0;
205 
206 	err = ovl_real_fileattr_get(new, &newfa);
207 	if (err) {
208 		/*
209 		 * Returning an error if upper doesn't support fileattr will
210 		 * result in a regression, so revert to the old behavior.
211 		 */
212 		if (err == -ENOTTY || err == -EINVAL) {
213 			pr_warn_once("copying fileattr: no support on upper\n");
214 			return 0;
215 		}
216 		pr_warn("failed to retrieve upper fileattr (%pd2, err=%i)\n",
217 			new->dentry, err);
218 		return err;
219 	}
220 
221 	BUILD_BUG_ON(OVL_COPY_FS_FLAGS_MASK & ~FS_COMMON_FL);
222 	newfa.flags &= ~OVL_COPY_FS_FLAGS_MASK;
223 	newfa.flags |= (oldfa.flags & OVL_COPY_FS_FLAGS_MASK);
224 
225 	BUILD_BUG_ON(OVL_COPY_FSX_FLAGS_MASK & ~FS_XFLAG_COMMON);
226 	newfa.fsx_xflags &= ~OVL_COPY_FSX_FLAGS_MASK;
227 	newfa.fsx_xflags |= (oldfa.fsx_xflags & OVL_COPY_FSX_FLAGS_MASK);
228 
229 	return ovl_real_fileattr_set(new, &newfa);
230 }
231 
232 static int ovl_verify_area(loff_t pos, loff_t pos2, loff_t len, loff_t totlen)
233 {
234 	loff_t tmp;
235 
236 	if (pos != pos2)
237 		return -EIO;
238 	if (pos < 0 || len < 0 || totlen < 0)
239 		return -EIO;
240 	if (check_add_overflow(pos, len, &tmp))
241 		return -EIO;
242 	return 0;
243 }
244 
245 static int ovl_sync_file(struct path *path)
246 {
247 	struct file *new_file;
248 	int err;
249 
250 	new_file = ovl_path_open(path, O_LARGEFILE | O_RDONLY);
251 	if (IS_ERR(new_file))
252 		return PTR_ERR(new_file);
253 
254 	err = vfs_fsync(new_file, 0);
255 	fput(new_file);
256 
257 	return err;
258 }
259 
260 static int ovl_copy_up_file(struct ovl_fs *ofs, struct dentry *dentry,
261 			    struct file *new_file, loff_t len,
262 			    bool datasync)
263 {
264 	struct path datapath;
265 	struct file *old_file;
266 	loff_t old_pos = 0;
267 	loff_t new_pos = 0;
268 	loff_t cloned;
269 	loff_t data_pos = -1;
270 	loff_t hole_len;
271 	bool skip_hole = false;
272 	int error = 0;
273 
274 	ovl_path_lowerdata(dentry, &datapath);
275 	if (WARN_ON_ONCE(datapath.dentry == NULL) ||
276 	    WARN_ON_ONCE(len < 0))
277 		return -EIO;
278 
279 	old_file = ovl_path_open(&datapath, O_LARGEFILE | O_RDONLY);
280 	if (IS_ERR(old_file))
281 		return PTR_ERR(old_file);
282 
283 	/* Try to use clone_file_range to clone up within the same fs */
284 	cloned = vfs_clone_file_range(old_file, 0, new_file, 0, len, 0);
285 	if (cloned == len)
286 		goto out_fput;
287 
288 	/* Couldn't clone, so now we try to copy the data */
289 	error = rw_verify_area(READ, old_file, &old_pos, len);
290 	if (!error)
291 		error = rw_verify_area(WRITE, new_file, &new_pos, len);
292 	if (error)
293 		goto out_fput;
294 
295 	/* Check if lower fs supports seek operation */
296 	if (old_file->f_mode & FMODE_LSEEK)
297 		skip_hole = true;
298 
299 	while (len) {
300 		size_t this_len = OVL_COPY_UP_CHUNK_SIZE;
301 		ssize_t bytes;
302 
303 		if (len < this_len)
304 			this_len = len;
305 
306 		if (signal_pending_state(TASK_KILLABLE, current)) {
307 			error = -EINTR;
308 			break;
309 		}
310 
311 		/*
312 		 * Fill zero for hole will cost unnecessary disk space
313 		 * and meanwhile slow down the copy-up speed, so we do
314 		 * an optimization for hole during copy-up, it relies
315 		 * on SEEK_DATA implementation in lower fs so if lower
316 		 * fs does not support it, copy-up will behave as before.
317 		 *
318 		 * Detail logic of hole detection as below:
319 		 * When we detect next data position is larger than current
320 		 * position we will skip that hole, otherwise we copy
321 		 * data in the size of OVL_COPY_UP_CHUNK_SIZE. Actually,
322 		 * it may not recognize all kind of holes and sometimes
323 		 * only skips partial of hole area. However, it will be
324 		 * enough for most of the use cases.
325 		 *
326 		 * We do not hold upper sb_writers throughout the loop to avert
327 		 * lockdep warning with llseek of lower file in nested overlay:
328 		 * - upper sb_writers
329 		 * -- lower ovl_inode_lock (ovl_llseek)
330 		 */
331 		if (skip_hole && data_pos < old_pos) {
332 			data_pos = vfs_llseek(old_file, old_pos, SEEK_DATA);
333 			if (data_pos > old_pos) {
334 				hole_len = data_pos - old_pos;
335 				len -= hole_len;
336 				old_pos = new_pos = data_pos;
337 				continue;
338 			} else if (data_pos == -ENXIO) {
339 				break;
340 			} else if (data_pos < 0) {
341 				skip_hole = false;
342 			}
343 		}
344 
345 		error = ovl_verify_area(old_pos, new_pos, this_len, len);
346 		if (error)
347 			break;
348 
349 		bytes = do_splice_direct(old_file, &old_pos,
350 					 new_file, &new_pos,
351 					 this_len, SPLICE_F_MOVE);
352 		if (bytes <= 0) {
353 			error = bytes;
354 			break;
355 		}
356 		WARN_ON(old_pos != new_pos);
357 
358 		len -= bytes;
359 	}
360 	/* call fsync once, either now or later along with metadata */
361 	if (!error && ovl_should_sync(ofs) && datasync)
362 		error = vfs_fsync(new_file, 0);
363 out_fput:
364 	fput(old_file);
365 	return error;
366 }
367 
368 static int ovl_set_size(struct ovl_fs *ofs,
369 			struct dentry *upperdentry, struct kstat *stat)
370 {
371 	struct iattr attr = {
372 		.ia_valid = ATTR_SIZE,
373 		.ia_size = stat->size,
374 	};
375 
376 	return ovl_do_notify_change(ofs, upperdentry, &attr);
377 }
378 
379 static int ovl_set_timestamps(struct ovl_fs *ofs, struct dentry *upperdentry,
380 			      struct kstat *stat)
381 {
382 	struct iattr attr = {
383 		.ia_valid =
384 		     ATTR_ATIME | ATTR_MTIME | ATTR_ATIME_SET | ATTR_MTIME_SET | ATTR_CTIME,
385 		.ia_atime = stat->atime,
386 		.ia_mtime = stat->mtime,
387 	};
388 
389 	return ovl_do_notify_change(ofs, upperdentry, &attr);
390 }
391 
392 int ovl_set_attr(struct ovl_fs *ofs, struct dentry *upperdentry,
393 		 struct kstat *stat)
394 {
395 	int err = 0;
396 
397 	if (!S_ISLNK(stat->mode)) {
398 		struct iattr attr = {
399 			.ia_valid = ATTR_MODE,
400 			.ia_mode = stat->mode,
401 		};
402 		err = ovl_do_notify_change(ofs, upperdentry, &attr);
403 	}
404 	if (!err) {
405 		struct iattr attr = {
406 			.ia_valid = ATTR_UID | ATTR_GID,
407 			.ia_vfsuid = VFSUIDT_INIT(stat->uid),
408 			.ia_vfsgid = VFSGIDT_INIT(stat->gid),
409 		};
410 		err = ovl_do_notify_change(ofs, upperdentry, &attr);
411 	}
412 	if (!err)
413 		ovl_set_timestamps(ofs, upperdentry, stat);
414 
415 	return err;
416 }
417 
418 struct ovl_fh *ovl_encode_real_fh(struct ovl_fs *ofs, struct dentry *real,
419 				  bool is_upper)
420 {
421 	struct ovl_fh *fh;
422 	int fh_type, dwords;
423 	int buflen = MAX_HANDLE_SZ;
424 	uuid_t *uuid = &real->d_sb->s_uuid;
425 	int err;
426 
427 	/* Make sure the real fid stays 32bit aligned */
428 	BUILD_BUG_ON(OVL_FH_FID_OFFSET % 4);
429 	BUILD_BUG_ON(MAX_HANDLE_SZ + OVL_FH_FID_OFFSET > 255);
430 
431 	fh = kzalloc(buflen + OVL_FH_FID_OFFSET, GFP_KERNEL);
432 	if (!fh)
433 		return ERR_PTR(-ENOMEM);
434 
435 	/*
436 	 * We encode a non-connectable file handle for non-dir, because we
437 	 * only need to find the lower inode number and we don't want to pay
438 	 * the price or reconnecting the dentry.
439 	 */
440 	dwords = buflen >> 2;
441 	fh_type = exportfs_encode_fh(real, (void *)fh->fb.fid, &dwords, 0);
442 	buflen = (dwords << 2);
443 
444 	err = -EIO;
445 	if (WARN_ON(fh_type < 0) ||
446 	    WARN_ON(buflen > MAX_HANDLE_SZ) ||
447 	    WARN_ON(fh_type == FILEID_INVALID))
448 		goto out_err;
449 
450 	fh->fb.version = OVL_FH_VERSION;
451 	fh->fb.magic = OVL_FH_MAGIC;
452 	fh->fb.type = fh_type;
453 	fh->fb.flags = OVL_FH_FLAG_CPU_ENDIAN;
454 	/*
455 	 * When we will want to decode an overlay dentry from this handle
456 	 * and all layers are on the same fs, if we get a disconncted real
457 	 * dentry when we decode fid, the only way to tell if we should assign
458 	 * it to upperdentry or to lowerstack is by checking this flag.
459 	 */
460 	if (is_upper)
461 		fh->fb.flags |= OVL_FH_FLAG_PATH_UPPER;
462 	fh->fb.len = sizeof(fh->fb) + buflen;
463 	if (ovl_origin_uuid(ofs))
464 		fh->fb.uuid = *uuid;
465 
466 	return fh;
467 
468 out_err:
469 	kfree(fh);
470 	return ERR_PTR(err);
471 }
472 
473 struct ovl_fh *ovl_get_origin_fh(struct ovl_fs *ofs, struct dentry *origin)
474 {
475 	/*
476 	 * When lower layer doesn't support export operations store a 'null' fh,
477 	 * so we can use the overlay.origin xattr to distignuish between a copy
478 	 * up and a pure upper inode.
479 	 */
480 	if (!ovl_can_decode_fh(origin->d_sb))
481 		return NULL;
482 
483 	return ovl_encode_real_fh(ofs, origin, false);
484 }
485 
486 int ovl_set_origin_fh(struct ovl_fs *ofs, const struct ovl_fh *fh,
487 		      struct dentry *upper)
488 {
489 	int err;
490 
491 	/*
492 	 * Do not fail when upper doesn't support xattrs.
493 	 */
494 	err = ovl_check_setxattr(ofs, upper, OVL_XATTR_ORIGIN, fh->buf,
495 				 fh ? fh->fb.len : 0, 0);
496 
497 	/* Ignore -EPERM from setting "user.*" on symlink/special */
498 	return err == -EPERM ? 0 : err;
499 }
500 
501 /* Store file handle of @upper dir in @index dir entry */
502 static int ovl_set_upper_fh(struct ovl_fs *ofs, struct dentry *upper,
503 			    struct dentry *index)
504 {
505 	const struct ovl_fh *fh;
506 	int err;
507 
508 	fh = ovl_encode_real_fh(ofs, upper, true);
509 	if (IS_ERR(fh))
510 		return PTR_ERR(fh);
511 
512 	err = ovl_setxattr(ofs, index, OVL_XATTR_UPPER, fh->buf, fh->fb.len);
513 
514 	kfree(fh);
515 	return err;
516 }
517 
518 /*
519  * Create and install index entry.
520  *
521  * Caller must hold i_mutex on indexdir.
522  */
523 static int ovl_create_index(struct dentry *dentry, const struct ovl_fh *fh,
524 			    struct dentry *upper)
525 {
526 	struct ovl_fs *ofs = OVL_FS(dentry->d_sb);
527 	struct dentry *indexdir = ovl_indexdir(dentry->d_sb);
528 	struct inode *dir = d_inode(indexdir);
529 	struct dentry *index = NULL;
530 	struct dentry *temp = NULL;
531 	struct qstr name = { };
532 	int err;
533 
534 	/*
535 	 * For now this is only used for creating index entry for directories,
536 	 * because non-dir are copied up directly to index and then hardlinked
537 	 * to upper dir.
538 	 *
539 	 * TODO: implement create index for non-dir, so we can call it when
540 	 * encoding file handle for non-dir in case index does not exist.
541 	 */
542 	if (WARN_ON(!d_is_dir(dentry)))
543 		return -EIO;
544 
545 	/* Directory not expected to be indexed before copy up */
546 	if (WARN_ON(ovl_test_flag(OVL_INDEX, d_inode(dentry))))
547 		return -EIO;
548 
549 	err = ovl_get_index_name_fh(fh, &name);
550 	if (err)
551 		return err;
552 
553 	temp = ovl_create_temp(ofs, indexdir, OVL_CATTR(S_IFDIR | 0));
554 	err = PTR_ERR(temp);
555 	if (IS_ERR(temp))
556 		goto free_name;
557 
558 	err = ovl_set_upper_fh(ofs, upper, temp);
559 	if (err)
560 		goto out;
561 
562 	index = ovl_lookup_upper(ofs, name.name, indexdir, name.len);
563 	if (IS_ERR(index)) {
564 		err = PTR_ERR(index);
565 	} else {
566 		err = ovl_do_rename(ofs, dir, temp, dir, index, 0);
567 		dput(index);
568 	}
569 out:
570 	if (err)
571 		ovl_cleanup(ofs, dir, temp);
572 	dput(temp);
573 free_name:
574 	kfree(name.name);
575 	return err;
576 }
577 
578 struct ovl_copy_up_ctx {
579 	struct dentry *parent;
580 	struct dentry *dentry;
581 	struct path lowerpath;
582 	struct kstat stat;
583 	struct kstat pstat;
584 	const char *link;
585 	struct dentry *destdir;
586 	struct qstr destname;
587 	struct dentry *workdir;
588 	const struct ovl_fh *origin_fh;
589 	bool origin;
590 	bool indexed;
591 	bool metacopy;
592 	bool metacopy_digest;
593 	bool metadata_fsync;
594 };
595 
596 static int ovl_link_up(struct ovl_copy_up_ctx *c)
597 {
598 	int err;
599 	struct dentry *upper;
600 	struct dentry *upperdir = ovl_dentry_upper(c->parent);
601 	struct ovl_fs *ofs = OVL_FS(c->dentry->d_sb);
602 	struct inode *udir = d_inode(upperdir);
603 
604 	ovl_start_write(c->dentry);
605 
606 	/* Mark parent "impure" because it may now contain non-pure upper */
607 	err = ovl_set_impure(c->parent, upperdir);
608 	if (err)
609 		goto out;
610 
611 	err = ovl_set_nlink_lower(c->dentry);
612 	if (err)
613 		goto out;
614 
615 	inode_lock_nested(udir, I_MUTEX_PARENT);
616 	upper = ovl_lookup_upper(ofs, c->dentry->d_name.name, upperdir,
617 				 c->dentry->d_name.len);
618 	err = PTR_ERR(upper);
619 	if (!IS_ERR(upper)) {
620 		err = ovl_do_link(ofs, ovl_dentry_upper(c->dentry), udir, upper);
621 		dput(upper);
622 
623 		if (!err) {
624 			/* Restore timestamps on parent (best effort) */
625 			ovl_set_timestamps(ofs, upperdir, &c->pstat);
626 			ovl_dentry_set_upper_alias(c->dentry);
627 			ovl_dentry_update_reval(c->dentry, upper);
628 		}
629 	}
630 	inode_unlock(udir);
631 	if (err)
632 		goto out;
633 
634 	err = ovl_set_nlink_upper(c->dentry);
635 
636 out:
637 	ovl_end_write(c->dentry);
638 	return err;
639 }
640 
641 static int ovl_copy_up_data(struct ovl_copy_up_ctx *c, const struct path *temp)
642 {
643 	struct ovl_fs *ofs = OVL_FS(c->dentry->d_sb);
644 	struct file *new_file;
645 	int err;
646 
647 	if (!S_ISREG(c->stat.mode) || c->metacopy || !c->stat.size)
648 		return 0;
649 
650 	new_file = ovl_path_open(temp, O_LARGEFILE | O_WRONLY);
651 	if (IS_ERR(new_file))
652 		return PTR_ERR(new_file);
653 
654 	err = ovl_copy_up_file(ofs, c->dentry, new_file, c->stat.size,
655 			       !c->metadata_fsync);
656 	fput(new_file);
657 
658 	return err;
659 }
660 
661 static int ovl_copy_up_metadata(struct ovl_copy_up_ctx *c, struct dentry *temp)
662 {
663 	struct ovl_fs *ofs = OVL_FS(c->dentry->d_sb);
664 	struct inode *inode = d_inode(c->dentry);
665 	struct path upperpath = { .mnt = ovl_upper_mnt(ofs), .dentry = temp };
666 	int err;
667 
668 	err = ovl_copy_xattr(c->dentry->d_sb, &c->lowerpath, temp);
669 	if (err)
670 		return err;
671 
672 	if (inode->i_flags & OVL_COPY_I_FLAGS_MASK &&
673 	    (S_ISREG(c->stat.mode) || S_ISDIR(c->stat.mode))) {
674 		/*
675 		 * Copy the fileattr inode flags that are the source of already
676 		 * copied i_flags
677 		 */
678 		err = ovl_copy_fileattr(inode, &c->lowerpath, &upperpath);
679 		if (err)
680 			return err;
681 	}
682 
683 	/*
684 	 * Store identifier of lower inode in upper inode xattr to
685 	 * allow lookup of the copy up origin inode.
686 	 *
687 	 * Don't set origin when we are breaking the association with a lower
688 	 * hard link.
689 	 */
690 	if (c->origin) {
691 		err = ovl_set_origin_fh(ofs, c->origin_fh, temp);
692 		if (err)
693 			return err;
694 	}
695 
696 	if (c->metacopy) {
697 		struct path lowerdatapath;
698 		struct ovl_metacopy metacopy_data = OVL_METACOPY_INIT;
699 
700 		ovl_path_lowerdata(c->dentry, &lowerdatapath);
701 		if (WARN_ON_ONCE(lowerdatapath.dentry == NULL))
702 			return -EIO;
703 		err = ovl_get_verity_digest(ofs, &lowerdatapath, &metacopy_data);
704 		if (err)
705 			return err;
706 
707 		if (metacopy_data.digest_algo)
708 			c->metacopy_digest = true;
709 
710 		err = ovl_set_metacopy_xattr(ofs, temp, &metacopy_data);
711 		if (err)
712 			return err;
713 	}
714 
715 	inode_lock(temp->d_inode);
716 	if (S_ISREG(c->stat.mode))
717 		err = ovl_set_size(ofs, temp, &c->stat);
718 	if (!err)
719 		err = ovl_set_attr(ofs, temp, &c->stat);
720 	inode_unlock(temp->d_inode);
721 
722 	/* fsync metadata before moving it into upper dir */
723 	if (!err && ovl_should_sync(ofs) && c->metadata_fsync)
724 		err = ovl_sync_file(&upperpath);
725 
726 	return err;
727 }
728 
729 struct ovl_cu_creds {
730 	const struct cred *old;
731 	struct cred *new;
732 };
733 
734 static int ovl_prep_cu_creds(struct dentry *dentry, struct ovl_cu_creds *cc)
735 {
736 	int err;
737 
738 	cc->old = cc->new = NULL;
739 	err = security_inode_copy_up(dentry, &cc->new);
740 	if (err < 0)
741 		return err;
742 
743 	if (cc->new)
744 		cc->old = override_creds(cc->new);
745 
746 	return 0;
747 }
748 
749 static void ovl_revert_cu_creds(struct ovl_cu_creds *cc)
750 {
751 	if (cc->new) {
752 		revert_creds(cc->old);
753 		put_cred(cc->new);
754 	}
755 }
756 
757 /*
758  * Copyup using workdir to prepare temp file.  Used when copying up directories,
759  * special files or when upper fs doesn't support O_TMPFILE.
760  */
761 static int ovl_copy_up_workdir(struct ovl_copy_up_ctx *c)
762 {
763 	struct ovl_fs *ofs = OVL_FS(c->dentry->d_sb);
764 	struct inode *inode;
765 	struct inode *udir = d_inode(c->destdir), *wdir = d_inode(c->workdir);
766 	struct path path = { .mnt = ovl_upper_mnt(ofs) };
767 	struct dentry *temp, *upper, *trap;
768 	struct ovl_cu_creds cc;
769 	int err;
770 	struct ovl_cattr cattr = {
771 		/* Can't properly set mode on creation because of the umask */
772 		.mode = c->stat.mode & S_IFMT,
773 		.rdev = c->stat.rdev,
774 		.link = c->link
775 	};
776 
777 	err = ovl_prep_cu_creds(c->dentry, &cc);
778 	if (err)
779 		return err;
780 
781 	ovl_start_write(c->dentry);
782 	inode_lock(wdir);
783 	temp = ovl_create_temp(ofs, c->workdir, &cattr);
784 	inode_unlock(wdir);
785 	ovl_end_write(c->dentry);
786 	ovl_revert_cu_creds(&cc);
787 
788 	if (IS_ERR(temp))
789 		return PTR_ERR(temp);
790 
791 	/*
792 	 * Copy up data first and then xattrs. Writing data after
793 	 * xattrs will remove security.capability xattr automatically.
794 	 */
795 	path.dentry = temp;
796 	err = ovl_copy_up_data(c, &path);
797 	/*
798 	 * We cannot hold lock_rename() throughout this helper, because of
799 	 * lock ordering with sb_writers, which shouldn't be held when calling
800 	 * ovl_copy_up_data(), so lock workdir and destdir and make sure that
801 	 * temp wasn't moved before copy up completion or cleanup.
802 	 */
803 	ovl_start_write(c->dentry);
804 	trap = lock_rename(c->workdir, c->destdir);
805 	if (trap || temp->d_parent != c->workdir) {
806 		/* temp or workdir moved underneath us? abort without cleanup */
807 		dput(temp);
808 		err = -EIO;
809 		if (IS_ERR(trap))
810 			goto out;
811 		goto unlock;
812 	} else if (err) {
813 		goto cleanup;
814 	}
815 
816 	err = ovl_copy_up_metadata(c, temp);
817 	if (err)
818 		goto cleanup;
819 
820 	if (S_ISDIR(c->stat.mode) && c->indexed) {
821 		err = ovl_create_index(c->dentry, c->origin_fh, temp);
822 		if (err)
823 			goto cleanup;
824 	}
825 
826 	upper = ovl_lookup_upper(ofs, c->destname.name, c->destdir,
827 				 c->destname.len);
828 	err = PTR_ERR(upper);
829 	if (IS_ERR(upper))
830 		goto cleanup;
831 
832 	err = ovl_do_rename(ofs, wdir, temp, udir, upper, 0);
833 	dput(upper);
834 	if (err)
835 		goto cleanup;
836 
837 	inode = d_inode(c->dentry);
838 	if (c->metacopy_digest)
839 		ovl_set_flag(OVL_HAS_DIGEST, inode);
840 	else
841 		ovl_clear_flag(OVL_HAS_DIGEST, inode);
842 	ovl_clear_flag(OVL_VERIFIED_DIGEST, inode);
843 
844 	if (!c->metacopy)
845 		ovl_set_upperdata(inode);
846 	ovl_inode_update(inode, temp);
847 	if (S_ISDIR(inode->i_mode))
848 		ovl_set_flag(OVL_WHITEOUTS, inode);
849 unlock:
850 	unlock_rename(c->workdir, c->destdir);
851 out:
852 	ovl_end_write(c->dentry);
853 
854 	return err;
855 
856 cleanup:
857 	ovl_cleanup(ofs, wdir, temp);
858 	dput(temp);
859 	goto unlock;
860 }
861 
862 /* Copyup using O_TMPFILE which does not require cross dir locking */
863 static int ovl_copy_up_tmpfile(struct ovl_copy_up_ctx *c)
864 {
865 	struct ovl_fs *ofs = OVL_FS(c->dentry->d_sb);
866 	struct inode *udir = d_inode(c->destdir);
867 	struct dentry *temp, *upper;
868 	struct file *tmpfile;
869 	struct ovl_cu_creds cc;
870 	int err;
871 
872 	err = ovl_prep_cu_creds(c->dentry, &cc);
873 	if (err)
874 		return err;
875 
876 	ovl_start_write(c->dentry);
877 	tmpfile = ovl_do_tmpfile(ofs, c->workdir, c->stat.mode);
878 	ovl_end_write(c->dentry);
879 	ovl_revert_cu_creds(&cc);
880 	if (IS_ERR(tmpfile))
881 		return PTR_ERR(tmpfile);
882 
883 	temp = tmpfile->f_path.dentry;
884 	if (!c->metacopy && c->stat.size) {
885 		err = ovl_copy_up_file(ofs, c->dentry, tmpfile, c->stat.size,
886 				       !c->metadata_fsync);
887 		if (err)
888 			goto out_fput;
889 	}
890 
891 	ovl_start_write(c->dentry);
892 
893 	err = ovl_copy_up_metadata(c, temp);
894 	if (err)
895 		goto out;
896 
897 	inode_lock_nested(udir, I_MUTEX_PARENT);
898 
899 	upper = ovl_lookup_upper(ofs, c->destname.name, c->destdir,
900 				 c->destname.len);
901 	err = PTR_ERR(upper);
902 	if (!IS_ERR(upper)) {
903 		err = ovl_do_link(ofs, temp, udir, upper);
904 		dput(upper);
905 	}
906 	inode_unlock(udir);
907 
908 	if (err)
909 		goto out;
910 
911 	if (c->metacopy_digest)
912 		ovl_set_flag(OVL_HAS_DIGEST, d_inode(c->dentry));
913 	else
914 		ovl_clear_flag(OVL_HAS_DIGEST, d_inode(c->dentry));
915 	ovl_clear_flag(OVL_VERIFIED_DIGEST, d_inode(c->dentry));
916 
917 	if (!c->metacopy)
918 		ovl_set_upperdata(d_inode(c->dentry));
919 	ovl_inode_update(d_inode(c->dentry), dget(temp));
920 
921 out:
922 	ovl_end_write(c->dentry);
923 out_fput:
924 	fput(tmpfile);
925 	return err;
926 }
927 
928 /*
929  * Copy up a single dentry
930  *
931  * All renames start with copy up of source if necessary.  The actual
932  * rename will only proceed once the copy up was successful.  Copy up uses
933  * upper parent i_mutex for exclusion.  Since rename can change d_parent it
934  * is possible that the copy up will lock the old parent.  At that point
935  * the file will have already been copied up anyway.
936  */
937 static int ovl_do_copy_up(struct ovl_copy_up_ctx *c)
938 {
939 	int err;
940 	struct ovl_fs *ofs = OVL_FS(c->dentry->d_sb);
941 	struct dentry *origin = c->lowerpath.dentry;
942 	struct ovl_fh *fh = NULL;
943 	bool to_index = false;
944 
945 	/*
946 	 * Indexed non-dir is copied up directly to the index entry and then
947 	 * hardlinked to upper dir. Indexed dir is copied up to indexdir,
948 	 * then index entry is created and then copied up dir installed.
949 	 * Copying dir up to indexdir instead of workdir simplifies locking.
950 	 */
951 	if (ovl_need_index(c->dentry)) {
952 		c->indexed = true;
953 		if (S_ISDIR(c->stat.mode))
954 			c->workdir = ovl_indexdir(c->dentry->d_sb);
955 		else
956 			to_index = true;
957 	}
958 
959 	if (S_ISDIR(c->stat.mode) || c->stat.nlink == 1 || to_index) {
960 		fh = ovl_get_origin_fh(ofs, origin);
961 		if (IS_ERR(fh))
962 			return PTR_ERR(fh);
963 
964 		/* origin_fh may be NULL */
965 		c->origin_fh = fh;
966 		c->origin = true;
967 	}
968 
969 	if (to_index) {
970 		c->destdir = ovl_indexdir(c->dentry->d_sb);
971 		err = ovl_get_index_name(ofs, origin, &c->destname);
972 		if (err)
973 			goto out_free_fh;
974 	} else if (WARN_ON(!c->parent)) {
975 		/* Disconnected dentry must be copied up to index dir */
976 		err = -EIO;
977 		goto out_free_fh;
978 	} else {
979 		/*
980 		 * c->dentry->d_name is stabilzed by ovl_copy_up_start(),
981 		 * because if we got here, it means that c->dentry has no upper
982 		 * alias and changing ->d_name means going through ovl_rename()
983 		 * that will call ovl_copy_up() on source and target dentry.
984 		 */
985 		c->destname = c->dentry->d_name;
986 		/*
987 		 * Mark parent "impure" because it may now contain non-pure
988 		 * upper
989 		 */
990 		ovl_start_write(c->dentry);
991 		err = ovl_set_impure(c->parent, c->destdir);
992 		ovl_end_write(c->dentry);
993 		if (err)
994 			goto out_free_fh;
995 	}
996 
997 	/* Should we copyup with O_TMPFILE or with workdir? */
998 	if (S_ISREG(c->stat.mode) && ofs->tmpfile)
999 		err = ovl_copy_up_tmpfile(c);
1000 	else
1001 		err = ovl_copy_up_workdir(c);
1002 	if (err)
1003 		goto out;
1004 
1005 	if (c->indexed)
1006 		ovl_set_flag(OVL_INDEX, d_inode(c->dentry));
1007 
1008 	ovl_start_write(c->dentry);
1009 	if (to_index) {
1010 		/* Initialize nlink for copy up of disconnected dentry */
1011 		err = ovl_set_nlink_upper(c->dentry);
1012 	} else {
1013 		struct inode *udir = d_inode(c->destdir);
1014 
1015 		/* Restore timestamps on parent (best effort) */
1016 		inode_lock(udir);
1017 		ovl_set_timestamps(ofs, c->destdir, &c->pstat);
1018 		inode_unlock(udir);
1019 
1020 		ovl_dentry_set_upper_alias(c->dentry);
1021 		ovl_dentry_update_reval(c->dentry, ovl_dentry_upper(c->dentry));
1022 	}
1023 	ovl_end_write(c->dentry);
1024 
1025 out:
1026 	if (to_index)
1027 		kfree(c->destname.name);
1028 out_free_fh:
1029 	kfree(fh);
1030 	return err;
1031 }
1032 
1033 static bool ovl_need_meta_copy_up(struct dentry *dentry, umode_t mode,
1034 				  int flags)
1035 {
1036 	struct ovl_fs *ofs = OVL_FS(dentry->d_sb);
1037 
1038 	if (!ofs->config.metacopy)
1039 		return false;
1040 
1041 	if (!S_ISREG(mode))
1042 		return false;
1043 
1044 	if (flags && ((OPEN_FMODE(flags) & FMODE_WRITE) || (flags & O_TRUNC)))
1045 		return false;
1046 
1047 	/* Fall back to full copy if no fsverity on source data and we require verity */
1048 	if (ofs->config.verity_mode == OVL_VERITY_REQUIRE) {
1049 		struct path lowerdata;
1050 
1051 		ovl_path_lowerdata(dentry, &lowerdata);
1052 
1053 		if (WARN_ON_ONCE(lowerdata.dentry == NULL) ||
1054 		    ovl_ensure_verity_loaded(&lowerdata) ||
1055 		    !fsverity_active(d_inode(lowerdata.dentry))) {
1056 			return false;
1057 		}
1058 	}
1059 
1060 	return true;
1061 }
1062 
1063 static ssize_t ovl_getxattr_value(const struct path *path, char *name, char **value)
1064 {
1065 	ssize_t res;
1066 	char *buf;
1067 
1068 	res = ovl_do_getxattr(path, name, NULL, 0);
1069 	if (res == -ENODATA || res == -EOPNOTSUPP)
1070 		res = 0;
1071 
1072 	if (res > 0) {
1073 		buf = kzalloc(res, GFP_KERNEL);
1074 		if (!buf)
1075 			return -ENOMEM;
1076 
1077 		res = ovl_do_getxattr(path, name, buf, res);
1078 		if (res < 0)
1079 			kfree(buf);
1080 		else
1081 			*value = buf;
1082 	}
1083 	return res;
1084 }
1085 
1086 /* Copy up data of an inode which was copied up metadata only in the past. */
1087 static int ovl_copy_up_meta_inode_data(struct ovl_copy_up_ctx *c)
1088 {
1089 	struct ovl_fs *ofs = OVL_FS(c->dentry->d_sb);
1090 	struct path upperpath;
1091 	int err;
1092 	char *capability = NULL;
1093 	ssize_t cap_size;
1094 
1095 	ovl_path_upper(c->dentry, &upperpath);
1096 	if (WARN_ON(upperpath.dentry == NULL))
1097 		return -EIO;
1098 
1099 	if (c->stat.size) {
1100 		err = cap_size = ovl_getxattr_value(&upperpath, XATTR_NAME_CAPS,
1101 						    &capability);
1102 		if (cap_size < 0)
1103 			goto out;
1104 	}
1105 
1106 	err = ovl_copy_up_data(c, &upperpath);
1107 	if (err)
1108 		goto out_free;
1109 
1110 	/*
1111 	 * Writing to upper file will clear security.capability xattr. We
1112 	 * don't want that to happen for normal copy-up operation.
1113 	 */
1114 	ovl_start_write(c->dentry);
1115 	if (capability) {
1116 		err = ovl_do_setxattr(ofs, upperpath.dentry, XATTR_NAME_CAPS,
1117 				      capability, cap_size, 0);
1118 	}
1119 	if (!err) {
1120 		err = ovl_removexattr(ofs, upperpath.dentry,
1121 				      OVL_XATTR_METACOPY);
1122 	}
1123 	ovl_end_write(c->dentry);
1124 	if (err)
1125 		goto out_free;
1126 
1127 	ovl_clear_flag(OVL_HAS_DIGEST, d_inode(c->dentry));
1128 	ovl_clear_flag(OVL_VERIFIED_DIGEST, d_inode(c->dentry));
1129 	ovl_set_upperdata(d_inode(c->dentry));
1130 out_free:
1131 	kfree(capability);
1132 out:
1133 	return err;
1134 }
1135 
1136 static int ovl_copy_up_one(struct dentry *parent, struct dentry *dentry,
1137 			   int flags)
1138 {
1139 	int err;
1140 	DEFINE_DELAYED_CALL(done);
1141 	struct path parentpath;
1142 	struct ovl_copy_up_ctx ctx = {
1143 		.parent = parent,
1144 		.dentry = dentry,
1145 		.workdir = ovl_workdir(dentry),
1146 	};
1147 
1148 	if (WARN_ON(!ctx.workdir))
1149 		return -EROFS;
1150 
1151 	ovl_path_lower(dentry, &ctx.lowerpath);
1152 	err = vfs_getattr(&ctx.lowerpath, &ctx.stat,
1153 			  STATX_BASIC_STATS, AT_STATX_SYNC_AS_STAT);
1154 	if (err)
1155 		return err;
1156 
1157 	if (!kuid_has_mapping(current_user_ns(), ctx.stat.uid) ||
1158 	    !kgid_has_mapping(current_user_ns(), ctx.stat.gid))
1159 		return -EOVERFLOW;
1160 
1161 	/*
1162 	 * With metacopy disabled, we fsync after final metadata copyup, for
1163 	 * both regular files and directories to get atomic copyup semantics
1164 	 * on filesystems that do not use strict metadata ordering (e.g. ubifs).
1165 	 *
1166 	 * With metacopy enabled we want to avoid fsync on all meta copyup
1167 	 * that will hurt performance of workloads such as chown -R, so we
1168 	 * only fsync on data copyup as legacy behavior.
1169 	 */
1170 	ctx.metadata_fsync = !OVL_FS(dentry->d_sb)->config.metacopy &&
1171 			     (S_ISREG(ctx.stat.mode) || S_ISDIR(ctx.stat.mode));
1172 	ctx.metacopy = ovl_need_meta_copy_up(dentry, ctx.stat.mode, flags);
1173 
1174 	if (parent) {
1175 		ovl_path_upper(parent, &parentpath);
1176 		ctx.destdir = parentpath.dentry;
1177 
1178 		err = vfs_getattr(&parentpath, &ctx.pstat,
1179 				  STATX_ATIME | STATX_MTIME,
1180 				  AT_STATX_SYNC_AS_STAT);
1181 		if (err)
1182 			return err;
1183 	}
1184 
1185 	/* maybe truncate regular file. this has no effect on dirs */
1186 	if (flags & O_TRUNC)
1187 		ctx.stat.size = 0;
1188 
1189 	if (S_ISLNK(ctx.stat.mode)) {
1190 		ctx.link = vfs_get_link(ctx.lowerpath.dentry, &done);
1191 		if (IS_ERR(ctx.link))
1192 			return PTR_ERR(ctx.link);
1193 	}
1194 
1195 	err = ovl_copy_up_start(dentry, flags);
1196 	/* err < 0: interrupted, err > 0: raced with another copy-up */
1197 	if (unlikely(err)) {
1198 		if (err > 0)
1199 			err = 0;
1200 	} else {
1201 		if (!ovl_dentry_upper(dentry))
1202 			err = ovl_do_copy_up(&ctx);
1203 		if (!err && parent && !ovl_dentry_has_upper_alias(dentry))
1204 			err = ovl_link_up(&ctx);
1205 		if (!err && ovl_dentry_needs_data_copy_up_locked(dentry, flags))
1206 			err = ovl_copy_up_meta_inode_data(&ctx);
1207 		ovl_copy_up_end(dentry);
1208 	}
1209 	do_delayed_call(&done);
1210 
1211 	return err;
1212 }
1213 
1214 static int ovl_copy_up_flags(struct dentry *dentry, int flags)
1215 {
1216 	int err = 0;
1217 	const struct cred *old_cred;
1218 	bool disconnected = (dentry->d_flags & DCACHE_DISCONNECTED);
1219 
1220 	/*
1221 	 * With NFS export, copy up can get called for a disconnected non-dir.
1222 	 * In this case, we will copy up lower inode to index dir without
1223 	 * linking it to upper dir.
1224 	 */
1225 	if (WARN_ON(disconnected && d_is_dir(dentry)))
1226 		return -EIO;
1227 
1228 	/*
1229 	 * We may not need lowerdata if we are only doing metacopy up, but it is
1230 	 * not very important to optimize this case, so do lazy lowerdata lookup
1231 	 * before any copy up, so we can do it before taking ovl_inode_lock().
1232 	 */
1233 	err = ovl_verify_lowerdata(dentry);
1234 	if (err)
1235 		return err;
1236 
1237 	old_cred = ovl_override_creds(dentry->d_sb);
1238 	while (!err) {
1239 		struct dentry *next;
1240 		struct dentry *parent = NULL;
1241 
1242 		if (ovl_already_copied_up(dentry, flags))
1243 			break;
1244 
1245 		next = dget(dentry);
1246 		/* find the topmost dentry not yet copied up */
1247 		for (; !disconnected;) {
1248 			parent = dget_parent(next);
1249 
1250 			if (ovl_dentry_upper(parent))
1251 				break;
1252 
1253 			dput(next);
1254 			next = parent;
1255 		}
1256 
1257 		err = ovl_copy_up_one(parent, next, flags);
1258 
1259 		dput(parent);
1260 		dput(next);
1261 	}
1262 	ovl_revert_creds(old_cred);
1263 
1264 	return err;
1265 }
1266 
1267 static bool ovl_open_need_copy_up(struct dentry *dentry, int flags)
1268 {
1269 	/* Copy up of disconnected dentry does not set upper alias */
1270 	if (ovl_already_copied_up(dentry, flags))
1271 		return false;
1272 
1273 	if (special_file(d_inode(dentry)->i_mode))
1274 		return false;
1275 
1276 	if (!ovl_open_flags_need_copy_up(flags))
1277 		return false;
1278 
1279 	return true;
1280 }
1281 
1282 int ovl_maybe_copy_up(struct dentry *dentry, int flags)
1283 {
1284 	if (!ovl_open_need_copy_up(dentry, flags))
1285 		return 0;
1286 
1287 	return ovl_copy_up_flags(dentry, flags);
1288 }
1289 
1290 int ovl_copy_up_with_data(struct dentry *dentry)
1291 {
1292 	return ovl_copy_up_flags(dentry, O_WRONLY);
1293 }
1294 
1295 int ovl_copy_up(struct dentry *dentry)
1296 {
1297 	return ovl_copy_up_flags(dentry, 0);
1298 }
1299