xref: /linux/fs/overlayfs/copy_up.c (revision 55f1b540d893da740a81200450014c45a8103f54)
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  *
4  * Copyright (C) 2011 Novell Inc.
5  */
6 
7 #include <linux/module.h>
8 #include <linux/fs.h>
9 #include <linux/slab.h>
10 #include <linux/file.h>
11 #include <linux/fileattr.h>
12 #include <linux/splice.h>
13 #include <linux/xattr.h>
14 #include <linux/security.h>
15 #include <linux/uaccess.h>
16 #include <linux/sched/signal.h>
17 #include <linux/cred.h>
18 #include <linux/namei.h>
19 #include <linux/fdtable.h>
20 #include <linux/ratelimit.h>
21 #include <linux/exportfs.h>
22 #include "overlayfs.h"
23 
24 #define OVL_COPY_UP_CHUNK_SIZE (1 << 20)
25 
26 static int ovl_ccup_set(const char *buf, const struct kernel_param *param)
27 {
28 	pr_warn("\"check_copy_up\" module option is obsolete\n");
29 	return 0;
30 }
31 
32 static int ovl_ccup_get(char *buf, const struct kernel_param *param)
33 {
34 	return sprintf(buf, "N\n");
35 }
36 
37 module_param_call(check_copy_up, ovl_ccup_set, ovl_ccup_get, NULL, 0644);
38 MODULE_PARM_DESC(check_copy_up, "Obsolete; does nothing");
39 
40 static bool ovl_must_copy_xattr(const char *name)
41 {
42 	return !strcmp(name, XATTR_POSIX_ACL_ACCESS) ||
43 	       !strcmp(name, XATTR_POSIX_ACL_DEFAULT) ||
44 	       !strncmp(name, XATTR_SECURITY_PREFIX, XATTR_SECURITY_PREFIX_LEN);
45 }
46 
47 static int ovl_copy_acl(struct ovl_fs *ofs, const struct path *path,
48 			struct dentry *dentry, const char *acl_name)
49 {
50 	int err;
51 	struct posix_acl *clone, *real_acl = NULL;
52 
53 	real_acl = ovl_get_acl_path(path, acl_name, false);
54 	if (!real_acl)
55 		return 0;
56 
57 	if (IS_ERR(real_acl)) {
58 		err = PTR_ERR(real_acl);
59 		if (err == -ENODATA || err == -EOPNOTSUPP)
60 			return 0;
61 		return err;
62 	}
63 
64 	clone = posix_acl_clone(real_acl, GFP_KERNEL);
65 	posix_acl_release(real_acl); /* release original acl */
66 	if (!clone)
67 		return -ENOMEM;
68 
69 	err = ovl_do_set_acl(ofs, dentry, acl_name, clone);
70 
71 	/* release cloned acl */
72 	posix_acl_release(clone);
73 	return err;
74 }
75 
76 int ovl_copy_xattr(struct super_block *sb, const struct path *oldpath, struct dentry *new)
77 {
78 	struct dentry *old = oldpath->dentry;
79 	ssize_t list_size, size, value_size = 0;
80 	char *buf, *name, *value = NULL;
81 	int error = 0;
82 	size_t slen;
83 
84 	if (!old->d_inode->i_op->listxattr || !new->d_inode->i_op->listxattr)
85 		return 0;
86 
87 	list_size = vfs_listxattr(old, NULL, 0);
88 	if (list_size <= 0) {
89 		if (list_size == -EOPNOTSUPP)
90 			return 0;
91 		return list_size;
92 	}
93 
94 	buf = kvzalloc(list_size, GFP_KERNEL);
95 	if (!buf)
96 		return -ENOMEM;
97 
98 	list_size = vfs_listxattr(old, buf, list_size);
99 	if (list_size <= 0) {
100 		error = list_size;
101 		goto out;
102 	}
103 
104 	for (name = buf; list_size; name += slen) {
105 		slen = strnlen(name, list_size) + 1;
106 
107 		/* underlying fs providing us with an broken xattr list? */
108 		if (WARN_ON(slen > list_size)) {
109 			error = -EIO;
110 			break;
111 		}
112 		list_size -= slen;
113 
114 		if (ovl_is_private_xattr(sb, name))
115 			continue;
116 
117 		error = security_inode_copy_up_xattr(old, name);
118 		if (error == -ECANCELED) {
119 			error = 0;
120 			continue; /* Discard */
121 		}
122 		if (error < 0 && error != -EOPNOTSUPP)
123 			break;
124 
125 		if (is_posix_acl_xattr(name)) {
126 			error = ovl_copy_acl(OVL_FS(sb), oldpath, new, name);
127 			if (!error)
128 				continue;
129 			/* POSIX ACLs must be copied. */
130 			break;
131 		}
132 
133 retry:
134 		size = ovl_do_getxattr(oldpath, name, value, value_size);
135 		if (size == -ERANGE)
136 			size = ovl_do_getxattr(oldpath, name, NULL, 0);
137 
138 		if (size < 0) {
139 			error = size;
140 			break;
141 		}
142 
143 		if (size > value_size) {
144 			void *new;
145 
146 			new = kvmalloc(size, GFP_KERNEL);
147 			if (!new) {
148 				error = -ENOMEM;
149 				break;
150 			}
151 			kvfree(value);
152 			value = new;
153 			value_size = size;
154 			goto retry;
155 		}
156 
157 		error = ovl_do_setxattr(OVL_FS(sb), new, name, value, size, 0);
158 		if (error) {
159 			if (error != -EOPNOTSUPP || ovl_must_copy_xattr(name))
160 				break;
161 
162 			/* Ignore failure to copy unknown xattrs */
163 			error = 0;
164 		}
165 	}
166 	kvfree(value);
167 out:
168 	kvfree(buf);
169 	return error;
170 }
171 
172 static int ovl_copy_fileattr(struct inode *inode, const struct path *old,
173 			     const struct path *new)
174 {
175 	struct fileattr oldfa = { .flags_valid = true };
176 	struct fileattr newfa = { .flags_valid = true };
177 	int err;
178 
179 	err = ovl_real_fileattr_get(old, &oldfa);
180 	if (err) {
181 		/* Ntfs-3g returns -EINVAL for "no fileattr support" */
182 		if (err == -ENOTTY || err == -EINVAL)
183 			return 0;
184 		pr_warn("failed to retrieve lower fileattr (%pd2, err=%i)\n",
185 			old->dentry, err);
186 		return err;
187 	}
188 
189 	/*
190 	 * We cannot set immutable and append-only flags on upper inode,
191 	 * because we would not be able to link upper inode to upper dir
192 	 * not set overlay private xattr on upper inode.
193 	 * Store these flags in overlay.protattr xattr instead.
194 	 */
195 	if (oldfa.flags & OVL_PROT_FS_FLAGS_MASK) {
196 		err = ovl_set_protattr(inode, new->dentry, &oldfa);
197 		if (err == -EPERM)
198 			pr_warn_once("copying fileattr: no xattr on upper\n");
199 		else if (err)
200 			return err;
201 	}
202 
203 	/* Don't bother copying flags if none are set */
204 	if (!(oldfa.flags & OVL_COPY_FS_FLAGS_MASK))
205 		return 0;
206 
207 	err = ovl_real_fileattr_get(new, &newfa);
208 	if (err) {
209 		/*
210 		 * Returning an error if upper doesn't support fileattr will
211 		 * result in a regression, so revert to the old behavior.
212 		 */
213 		if (err == -ENOTTY || err == -EINVAL) {
214 			pr_warn_once("copying fileattr: no support on upper\n");
215 			return 0;
216 		}
217 		pr_warn("failed to retrieve upper fileattr (%pd2, err=%i)\n",
218 			new->dentry, err);
219 		return err;
220 	}
221 
222 	BUILD_BUG_ON(OVL_COPY_FS_FLAGS_MASK & ~FS_COMMON_FL);
223 	newfa.flags &= ~OVL_COPY_FS_FLAGS_MASK;
224 	newfa.flags |= (oldfa.flags & OVL_COPY_FS_FLAGS_MASK);
225 
226 	BUILD_BUG_ON(OVL_COPY_FSX_FLAGS_MASK & ~FS_XFLAG_COMMON);
227 	newfa.fsx_xflags &= ~OVL_COPY_FSX_FLAGS_MASK;
228 	newfa.fsx_xflags |= (oldfa.fsx_xflags & OVL_COPY_FSX_FLAGS_MASK);
229 
230 	return ovl_real_fileattr_set(new, &newfa);
231 }
232 
233 static int ovl_verify_area(loff_t pos, loff_t pos2, loff_t len, loff_t totlen)
234 {
235 	loff_t tmp;
236 
237 	if (pos != pos2)
238 		return -EIO;
239 	if (pos < 0 || len < 0 || totlen < 0)
240 		return -EIO;
241 	if (check_add_overflow(pos, len, &tmp))
242 		return -EIO;
243 	return 0;
244 }
245 
246 static int ovl_sync_file(struct path *path)
247 {
248 	struct file *new_file;
249 	int err;
250 
251 	new_file = ovl_path_open(path, O_LARGEFILE | O_RDONLY);
252 	if (IS_ERR(new_file))
253 		return PTR_ERR(new_file);
254 
255 	err = vfs_fsync(new_file, 0);
256 	fput(new_file);
257 
258 	return err;
259 }
260 
261 static int ovl_copy_up_file(struct ovl_fs *ofs, struct dentry *dentry,
262 			    struct file *new_file, loff_t len,
263 			    bool datasync)
264 {
265 	struct path datapath;
266 	struct file *old_file;
267 	loff_t old_pos = 0;
268 	loff_t new_pos = 0;
269 	loff_t cloned;
270 	loff_t data_pos = -1;
271 	loff_t hole_len;
272 	bool skip_hole = false;
273 	int error = 0;
274 
275 	ovl_path_lowerdata(dentry, &datapath);
276 	if (WARN_ON_ONCE(datapath.dentry == NULL) ||
277 	    WARN_ON_ONCE(len < 0))
278 		return -EIO;
279 
280 	old_file = ovl_path_open(&datapath, O_LARGEFILE | O_RDONLY);
281 	if (IS_ERR(old_file))
282 		return PTR_ERR(old_file);
283 
284 	/* Try to use clone_file_range to clone up within the same fs */
285 	cloned = vfs_clone_file_range(old_file, 0, new_file, 0, len, 0);
286 	if (cloned == len)
287 		goto out_fput;
288 
289 	/* Couldn't clone, so now we try to copy the data */
290 	error = rw_verify_area(READ, old_file, &old_pos, len);
291 	if (!error)
292 		error = rw_verify_area(WRITE, new_file, &new_pos, len);
293 	if (error)
294 		goto out_fput;
295 
296 	/* Check if lower fs supports seek operation */
297 	if (old_file->f_mode & FMODE_LSEEK)
298 		skip_hole = true;
299 
300 	while (len) {
301 		size_t this_len = OVL_COPY_UP_CHUNK_SIZE;
302 		ssize_t bytes;
303 
304 		if (len < this_len)
305 			this_len = len;
306 
307 		if (signal_pending_state(TASK_KILLABLE, current)) {
308 			error = -EINTR;
309 			break;
310 		}
311 
312 		/*
313 		 * Fill zero for hole will cost unnecessary disk space
314 		 * and meanwhile slow down the copy-up speed, so we do
315 		 * an optimization for hole during copy-up, it relies
316 		 * on SEEK_DATA implementation in lower fs so if lower
317 		 * fs does not support it, copy-up will behave as before.
318 		 *
319 		 * Detail logic of hole detection as below:
320 		 * When we detect next data position is larger than current
321 		 * position we will skip that hole, otherwise we copy
322 		 * data in the size of OVL_COPY_UP_CHUNK_SIZE. Actually,
323 		 * it may not recognize all kind of holes and sometimes
324 		 * only skips partial of hole area. However, it will be
325 		 * enough for most of the use cases.
326 		 *
327 		 * We do not hold upper sb_writers throughout the loop to avert
328 		 * lockdep warning with llseek of lower file in nested overlay:
329 		 * - upper sb_writers
330 		 * -- lower ovl_inode_lock (ovl_llseek)
331 		 */
332 		if (skip_hole && data_pos < old_pos) {
333 			data_pos = vfs_llseek(old_file, old_pos, SEEK_DATA);
334 			if (data_pos > old_pos) {
335 				hole_len = data_pos - old_pos;
336 				len -= hole_len;
337 				old_pos = new_pos = data_pos;
338 				continue;
339 			} else if (data_pos == -ENXIO) {
340 				break;
341 			} else if (data_pos < 0) {
342 				skip_hole = false;
343 			}
344 		}
345 
346 		error = ovl_verify_area(old_pos, new_pos, this_len, len);
347 		if (error)
348 			break;
349 
350 		bytes = do_splice_direct(old_file, &old_pos,
351 					 new_file, &new_pos,
352 					 this_len, SPLICE_F_MOVE);
353 		if (bytes <= 0) {
354 			error = bytes;
355 			break;
356 		}
357 		WARN_ON(old_pos != new_pos);
358 
359 		len -= bytes;
360 	}
361 	/* call fsync once, either now or later along with metadata */
362 	if (!error && ovl_should_sync(ofs) && datasync)
363 		error = vfs_fsync(new_file, 0);
364 out_fput:
365 	fput(old_file);
366 	return error;
367 }
368 
369 static int ovl_set_size(struct ovl_fs *ofs,
370 			struct dentry *upperdentry, struct kstat *stat)
371 {
372 	struct iattr attr = {
373 		.ia_valid = ATTR_SIZE,
374 		.ia_size = stat->size,
375 	};
376 
377 	return ovl_do_notify_change(ofs, upperdentry, &attr);
378 }
379 
380 static int ovl_set_timestamps(struct ovl_fs *ofs, struct dentry *upperdentry,
381 			      struct kstat *stat)
382 {
383 	struct iattr attr = {
384 		.ia_valid =
385 		     ATTR_ATIME | ATTR_MTIME | ATTR_ATIME_SET | ATTR_MTIME_SET | ATTR_CTIME,
386 		.ia_atime = stat->atime,
387 		.ia_mtime = stat->mtime,
388 	};
389 
390 	return ovl_do_notify_change(ofs, upperdentry, &attr);
391 }
392 
393 int ovl_set_attr(struct ovl_fs *ofs, struct dentry *upperdentry,
394 		 struct kstat *stat)
395 {
396 	int err = 0;
397 
398 	if (!S_ISLNK(stat->mode)) {
399 		struct iattr attr = {
400 			.ia_valid = ATTR_MODE,
401 			.ia_mode = stat->mode,
402 		};
403 		err = ovl_do_notify_change(ofs, upperdentry, &attr);
404 	}
405 	if (!err) {
406 		struct iattr attr = {
407 			.ia_valid = ATTR_UID | ATTR_GID,
408 			.ia_vfsuid = VFSUIDT_INIT(stat->uid),
409 			.ia_vfsgid = VFSGIDT_INIT(stat->gid),
410 		};
411 		err = ovl_do_notify_change(ofs, upperdentry, &attr);
412 	}
413 	if (!err)
414 		ovl_set_timestamps(ofs, upperdentry, stat);
415 
416 	return err;
417 }
418 
419 struct ovl_fh *ovl_encode_real_fh(struct ovl_fs *ofs, struct dentry *real,
420 				  bool is_upper)
421 {
422 	struct ovl_fh *fh;
423 	int fh_type, dwords;
424 	int buflen = MAX_HANDLE_SZ;
425 	uuid_t *uuid = &real->d_sb->s_uuid;
426 	int err;
427 
428 	/* Make sure the real fid stays 32bit aligned */
429 	BUILD_BUG_ON(OVL_FH_FID_OFFSET % 4);
430 	BUILD_BUG_ON(MAX_HANDLE_SZ + OVL_FH_FID_OFFSET > 255);
431 
432 	fh = kzalloc(buflen + OVL_FH_FID_OFFSET, GFP_KERNEL);
433 	if (!fh)
434 		return ERR_PTR(-ENOMEM);
435 
436 	/*
437 	 * We encode a non-connectable file handle for non-dir, because we
438 	 * only need to find the lower inode number and we don't want to pay
439 	 * the price or reconnecting the dentry.
440 	 */
441 	dwords = buflen >> 2;
442 	fh_type = exportfs_encode_fh(real, (void *)fh->fb.fid, &dwords, 0);
443 	buflen = (dwords << 2);
444 
445 	err = -EIO;
446 	if (WARN_ON(fh_type < 0) ||
447 	    WARN_ON(buflen > MAX_HANDLE_SZ) ||
448 	    WARN_ON(fh_type == FILEID_INVALID))
449 		goto out_err;
450 
451 	fh->fb.version = OVL_FH_VERSION;
452 	fh->fb.magic = OVL_FH_MAGIC;
453 	fh->fb.type = fh_type;
454 	fh->fb.flags = OVL_FH_FLAG_CPU_ENDIAN;
455 	/*
456 	 * When we will want to decode an overlay dentry from this handle
457 	 * and all layers are on the same fs, if we get a disconncted real
458 	 * dentry when we decode fid, the only way to tell if we should assign
459 	 * it to upperdentry or to lowerstack is by checking this flag.
460 	 */
461 	if (is_upper)
462 		fh->fb.flags |= OVL_FH_FLAG_PATH_UPPER;
463 	fh->fb.len = sizeof(fh->fb) + buflen;
464 	if (ovl_origin_uuid(ofs))
465 		fh->fb.uuid = *uuid;
466 
467 	return fh;
468 
469 out_err:
470 	kfree(fh);
471 	return ERR_PTR(err);
472 }
473 
474 struct ovl_fh *ovl_get_origin_fh(struct ovl_fs *ofs, struct dentry *origin)
475 {
476 	/*
477 	 * When lower layer doesn't support export operations store a 'null' fh,
478 	 * so we can use the overlay.origin xattr to distignuish between a copy
479 	 * up and a pure upper inode.
480 	 */
481 	if (!ovl_can_decode_fh(origin->d_sb))
482 		return NULL;
483 
484 	return ovl_encode_real_fh(ofs, origin, false);
485 }
486 
487 int ovl_set_origin_fh(struct ovl_fs *ofs, const struct ovl_fh *fh,
488 		      struct dentry *upper)
489 {
490 	int err;
491 
492 	/*
493 	 * Do not fail when upper doesn't support xattrs.
494 	 */
495 	err = ovl_check_setxattr(ofs, upper, OVL_XATTR_ORIGIN, fh->buf,
496 				 fh ? fh->fb.len : 0, 0);
497 
498 	/* Ignore -EPERM from setting "user.*" on symlink/special */
499 	return err == -EPERM ? 0 : err;
500 }
501 
502 /* Store file handle of @upper dir in @index dir entry */
503 static int ovl_set_upper_fh(struct ovl_fs *ofs, struct dentry *upper,
504 			    struct dentry *index)
505 {
506 	const struct ovl_fh *fh;
507 	int err;
508 
509 	fh = ovl_encode_real_fh(ofs, upper, true);
510 	if (IS_ERR(fh))
511 		return PTR_ERR(fh);
512 
513 	err = ovl_setxattr(ofs, index, OVL_XATTR_UPPER, fh->buf, fh->fb.len);
514 
515 	kfree(fh);
516 	return err;
517 }
518 
519 /*
520  * Create and install index entry.
521  *
522  * Caller must hold i_mutex on indexdir.
523  */
524 static int ovl_create_index(struct dentry *dentry, const struct ovl_fh *fh,
525 			    struct dentry *upper)
526 {
527 	struct ovl_fs *ofs = OVL_FS(dentry->d_sb);
528 	struct dentry *indexdir = ovl_indexdir(dentry->d_sb);
529 	struct inode *dir = d_inode(indexdir);
530 	struct dentry *index = NULL;
531 	struct dentry *temp = NULL;
532 	struct qstr name = { };
533 	int err;
534 
535 	/*
536 	 * For now this is only used for creating index entry for directories,
537 	 * because non-dir are copied up directly to index and then hardlinked
538 	 * to upper dir.
539 	 *
540 	 * TODO: implement create index for non-dir, so we can call it when
541 	 * encoding file handle for non-dir in case index does not exist.
542 	 */
543 	if (WARN_ON(!d_is_dir(dentry)))
544 		return -EIO;
545 
546 	/* Directory not expected to be indexed before copy up */
547 	if (WARN_ON(ovl_test_flag(OVL_INDEX, d_inode(dentry))))
548 		return -EIO;
549 
550 	err = ovl_get_index_name_fh(fh, &name);
551 	if (err)
552 		return err;
553 
554 	temp = ovl_create_temp(ofs, indexdir, OVL_CATTR(S_IFDIR | 0));
555 	err = PTR_ERR(temp);
556 	if (IS_ERR(temp))
557 		goto free_name;
558 
559 	err = ovl_set_upper_fh(ofs, upper, temp);
560 	if (err)
561 		goto out;
562 
563 	index = ovl_lookup_upper(ofs, name.name, indexdir, name.len);
564 	if (IS_ERR(index)) {
565 		err = PTR_ERR(index);
566 	} else {
567 		err = ovl_do_rename(ofs, dir, temp, dir, index, 0);
568 		dput(index);
569 	}
570 out:
571 	if (err)
572 		ovl_cleanup(ofs, dir, temp);
573 	dput(temp);
574 free_name:
575 	kfree(name.name);
576 	return err;
577 }
578 
579 struct ovl_copy_up_ctx {
580 	struct dentry *parent;
581 	struct dentry *dentry;
582 	struct path lowerpath;
583 	struct kstat stat;
584 	struct kstat pstat;
585 	const char *link;
586 	struct dentry *destdir;
587 	struct qstr destname;
588 	struct dentry *workdir;
589 	const struct ovl_fh *origin_fh;
590 	bool origin;
591 	bool indexed;
592 	bool metacopy;
593 	bool metacopy_digest;
594 	bool metadata_fsync;
595 };
596 
597 static int ovl_link_up(struct ovl_copy_up_ctx *c)
598 {
599 	int err;
600 	struct dentry *upper;
601 	struct dentry *upperdir = ovl_dentry_upper(c->parent);
602 	struct ovl_fs *ofs = OVL_FS(c->dentry->d_sb);
603 	struct inode *udir = d_inode(upperdir);
604 
605 	ovl_start_write(c->dentry);
606 
607 	/* Mark parent "impure" because it may now contain non-pure upper */
608 	err = ovl_set_impure(c->parent, upperdir);
609 	if (err)
610 		goto out;
611 
612 	err = ovl_set_nlink_lower(c->dentry);
613 	if (err)
614 		goto out;
615 
616 	inode_lock_nested(udir, I_MUTEX_PARENT);
617 	upper = ovl_lookup_upper(ofs, c->dentry->d_name.name, upperdir,
618 				 c->dentry->d_name.len);
619 	err = PTR_ERR(upper);
620 	if (!IS_ERR(upper)) {
621 		err = ovl_do_link(ofs, ovl_dentry_upper(c->dentry), udir, upper);
622 		dput(upper);
623 
624 		if (!err) {
625 			/* Restore timestamps on parent (best effort) */
626 			ovl_set_timestamps(ofs, upperdir, &c->pstat);
627 			ovl_dentry_set_upper_alias(c->dentry);
628 			ovl_dentry_update_reval(c->dentry, upper);
629 		}
630 	}
631 	inode_unlock(udir);
632 	if (err)
633 		goto out;
634 
635 	err = ovl_set_nlink_upper(c->dentry);
636 
637 out:
638 	ovl_end_write(c->dentry);
639 	return err;
640 }
641 
642 static int ovl_copy_up_data(struct ovl_copy_up_ctx *c, const struct path *temp)
643 {
644 	struct ovl_fs *ofs = OVL_FS(c->dentry->d_sb);
645 	struct file *new_file;
646 	int err;
647 
648 	if (!S_ISREG(c->stat.mode) || c->metacopy || !c->stat.size)
649 		return 0;
650 
651 	new_file = ovl_path_open(temp, O_LARGEFILE | O_WRONLY);
652 	if (IS_ERR(new_file))
653 		return PTR_ERR(new_file);
654 
655 	err = ovl_copy_up_file(ofs, c->dentry, new_file, c->stat.size,
656 			       !c->metadata_fsync);
657 	fput(new_file);
658 
659 	return err;
660 }
661 
662 static int ovl_copy_up_metadata(struct ovl_copy_up_ctx *c, struct dentry *temp)
663 {
664 	struct ovl_fs *ofs = OVL_FS(c->dentry->d_sb);
665 	struct inode *inode = d_inode(c->dentry);
666 	struct path upperpath = { .mnt = ovl_upper_mnt(ofs), .dentry = temp };
667 	int err;
668 
669 	err = ovl_copy_xattr(c->dentry->d_sb, &c->lowerpath, temp);
670 	if (err)
671 		return err;
672 
673 	if (inode->i_flags & OVL_COPY_I_FLAGS_MASK &&
674 	    (S_ISREG(c->stat.mode) || S_ISDIR(c->stat.mode))) {
675 		/*
676 		 * Copy the fileattr inode flags that are the source of already
677 		 * copied i_flags
678 		 */
679 		err = ovl_copy_fileattr(inode, &c->lowerpath, &upperpath);
680 		if (err)
681 			return err;
682 	}
683 
684 	/*
685 	 * Store identifier of lower inode in upper inode xattr to
686 	 * allow lookup of the copy up origin inode.
687 	 *
688 	 * Don't set origin when we are breaking the association with a lower
689 	 * hard link.
690 	 */
691 	if (c->origin) {
692 		err = ovl_set_origin_fh(ofs, c->origin_fh, temp);
693 		if (err)
694 			return err;
695 	}
696 
697 	if (c->metacopy) {
698 		struct path lowerdatapath;
699 		struct ovl_metacopy metacopy_data = OVL_METACOPY_INIT;
700 
701 		ovl_path_lowerdata(c->dentry, &lowerdatapath);
702 		if (WARN_ON_ONCE(lowerdatapath.dentry == NULL))
703 			return -EIO;
704 		err = ovl_get_verity_digest(ofs, &lowerdatapath, &metacopy_data);
705 		if (err)
706 			return err;
707 
708 		if (metacopy_data.digest_algo)
709 			c->metacopy_digest = true;
710 
711 		err = ovl_set_metacopy_xattr(ofs, temp, &metacopy_data);
712 		if (err)
713 			return err;
714 	}
715 
716 	inode_lock(temp->d_inode);
717 	if (S_ISREG(c->stat.mode))
718 		err = ovl_set_size(ofs, temp, &c->stat);
719 	if (!err)
720 		err = ovl_set_attr(ofs, temp, &c->stat);
721 	inode_unlock(temp->d_inode);
722 
723 	/* fsync metadata before moving it into upper dir */
724 	if (!err && ovl_should_sync(ofs) && c->metadata_fsync)
725 		err = ovl_sync_file(&upperpath);
726 
727 	return err;
728 }
729 
730 struct ovl_cu_creds {
731 	const struct cred *old;
732 	struct cred *new;
733 };
734 
735 static int ovl_prep_cu_creds(struct dentry *dentry, struct ovl_cu_creds *cc)
736 {
737 	int err;
738 
739 	cc->old = cc->new = NULL;
740 	err = security_inode_copy_up(dentry, &cc->new);
741 	if (err < 0)
742 		return err;
743 
744 	if (cc->new)
745 		cc->old = override_creds(cc->new);
746 
747 	return 0;
748 }
749 
750 static void ovl_revert_cu_creds(struct ovl_cu_creds *cc)
751 {
752 	if (cc->new) {
753 		revert_creds(cc->old);
754 		put_cred(cc->new);
755 	}
756 }
757 
758 /*
759  * Copyup using workdir to prepare temp file.  Used when copying up directories,
760  * special files or when upper fs doesn't support O_TMPFILE.
761  */
762 static int ovl_copy_up_workdir(struct ovl_copy_up_ctx *c)
763 {
764 	struct ovl_fs *ofs = OVL_FS(c->dentry->d_sb);
765 	struct inode *inode;
766 	struct inode *udir = d_inode(c->destdir), *wdir = d_inode(c->workdir);
767 	struct path path = { .mnt = ovl_upper_mnt(ofs) };
768 	struct dentry *temp, *upper, *trap;
769 	struct ovl_cu_creds cc;
770 	int err;
771 	struct ovl_cattr cattr = {
772 		/* Can't properly set mode on creation because of the umask */
773 		.mode = c->stat.mode & S_IFMT,
774 		.rdev = c->stat.rdev,
775 		.link = c->link
776 	};
777 
778 	err = ovl_prep_cu_creds(c->dentry, &cc);
779 	if (err)
780 		return err;
781 
782 	ovl_start_write(c->dentry);
783 	inode_lock(wdir);
784 	temp = ovl_create_temp(ofs, c->workdir, &cattr);
785 	inode_unlock(wdir);
786 	ovl_end_write(c->dentry);
787 	ovl_revert_cu_creds(&cc);
788 
789 	if (IS_ERR(temp))
790 		return PTR_ERR(temp);
791 
792 	/*
793 	 * Copy up data first and then xattrs. Writing data after
794 	 * xattrs will remove security.capability xattr automatically.
795 	 */
796 	path.dentry = temp;
797 	err = ovl_copy_up_data(c, &path);
798 	/*
799 	 * We cannot hold lock_rename() throughout this helper, because of
800 	 * lock ordering with sb_writers, which shouldn't be held when calling
801 	 * ovl_copy_up_data(), so lock workdir and destdir and make sure that
802 	 * temp wasn't moved before copy up completion or cleanup.
803 	 */
804 	ovl_start_write(c->dentry);
805 	trap = lock_rename(c->workdir, c->destdir);
806 	if (trap || temp->d_parent != c->workdir) {
807 		/* temp or workdir moved underneath us? abort without cleanup */
808 		dput(temp);
809 		err = -EIO;
810 		if (IS_ERR(trap))
811 			goto out;
812 		goto unlock;
813 	} else if (err) {
814 		goto cleanup;
815 	}
816 
817 	err = ovl_copy_up_metadata(c, temp);
818 	if (err)
819 		goto cleanup;
820 
821 	if (S_ISDIR(c->stat.mode) && c->indexed) {
822 		err = ovl_create_index(c->dentry, c->origin_fh, temp);
823 		if (err)
824 			goto cleanup;
825 	}
826 
827 	upper = ovl_lookup_upper(ofs, c->destname.name, c->destdir,
828 				 c->destname.len);
829 	err = PTR_ERR(upper);
830 	if (IS_ERR(upper))
831 		goto cleanup;
832 
833 	err = ovl_do_rename(ofs, wdir, temp, udir, upper, 0);
834 	dput(upper);
835 	if (err)
836 		goto cleanup;
837 
838 	inode = d_inode(c->dentry);
839 	if (c->metacopy_digest)
840 		ovl_set_flag(OVL_HAS_DIGEST, inode);
841 	else
842 		ovl_clear_flag(OVL_HAS_DIGEST, inode);
843 	ovl_clear_flag(OVL_VERIFIED_DIGEST, inode);
844 
845 	if (!c->metacopy)
846 		ovl_set_upperdata(inode);
847 	ovl_inode_update(inode, temp);
848 	if (S_ISDIR(inode->i_mode))
849 		ovl_set_flag(OVL_WHITEOUTS, inode);
850 unlock:
851 	unlock_rename(c->workdir, c->destdir);
852 out:
853 	ovl_end_write(c->dentry);
854 
855 	return err;
856 
857 cleanup:
858 	ovl_cleanup(ofs, wdir, temp);
859 	dput(temp);
860 	goto unlock;
861 }
862 
863 /* Copyup using O_TMPFILE which does not require cross dir locking */
864 static int ovl_copy_up_tmpfile(struct ovl_copy_up_ctx *c)
865 {
866 	struct ovl_fs *ofs = OVL_FS(c->dentry->d_sb);
867 	struct inode *udir = d_inode(c->destdir);
868 	struct dentry *temp, *upper;
869 	struct file *tmpfile;
870 	struct ovl_cu_creds cc;
871 	int err;
872 
873 	err = ovl_prep_cu_creds(c->dentry, &cc);
874 	if (err)
875 		return err;
876 
877 	ovl_start_write(c->dentry);
878 	tmpfile = ovl_do_tmpfile(ofs, c->workdir, c->stat.mode);
879 	ovl_end_write(c->dentry);
880 	ovl_revert_cu_creds(&cc);
881 	if (IS_ERR(tmpfile))
882 		return PTR_ERR(tmpfile);
883 
884 	temp = tmpfile->f_path.dentry;
885 	if (!c->metacopy && c->stat.size) {
886 		err = ovl_copy_up_file(ofs, c->dentry, tmpfile, c->stat.size,
887 				       !c->metadata_fsync);
888 		if (err)
889 			goto out_fput;
890 	}
891 
892 	ovl_start_write(c->dentry);
893 
894 	err = ovl_copy_up_metadata(c, temp);
895 	if (err)
896 		goto out;
897 
898 	inode_lock_nested(udir, I_MUTEX_PARENT);
899 
900 	upper = ovl_lookup_upper(ofs, c->destname.name, c->destdir,
901 				 c->destname.len);
902 	err = PTR_ERR(upper);
903 	if (!IS_ERR(upper)) {
904 		err = ovl_do_link(ofs, temp, udir, upper);
905 		dput(upper);
906 	}
907 	inode_unlock(udir);
908 
909 	if (err)
910 		goto out;
911 
912 	if (c->metacopy_digest)
913 		ovl_set_flag(OVL_HAS_DIGEST, d_inode(c->dentry));
914 	else
915 		ovl_clear_flag(OVL_HAS_DIGEST, d_inode(c->dentry));
916 	ovl_clear_flag(OVL_VERIFIED_DIGEST, d_inode(c->dentry));
917 
918 	if (!c->metacopy)
919 		ovl_set_upperdata(d_inode(c->dentry));
920 	ovl_inode_update(d_inode(c->dentry), dget(temp));
921 
922 out:
923 	ovl_end_write(c->dentry);
924 out_fput:
925 	fput(tmpfile);
926 	return err;
927 }
928 
929 /*
930  * Copy up a single dentry
931  *
932  * All renames start with copy up of source if necessary.  The actual
933  * rename will only proceed once the copy up was successful.  Copy up uses
934  * upper parent i_mutex for exclusion.  Since rename can change d_parent it
935  * is possible that the copy up will lock the old parent.  At that point
936  * the file will have already been copied up anyway.
937  */
938 static int ovl_do_copy_up(struct ovl_copy_up_ctx *c)
939 {
940 	int err;
941 	struct ovl_fs *ofs = OVL_FS(c->dentry->d_sb);
942 	struct dentry *origin = c->lowerpath.dentry;
943 	struct ovl_fh *fh = NULL;
944 	bool to_index = false;
945 
946 	/*
947 	 * Indexed non-dir is copied up directly to the index entry and then
948 	 * hardlinked to upper dir. Indexed dir is copied up to indexdir,
949 	 * then index entry is created and then copied up dir installed.
950 	 * Copying dir up to indexdir instead of workdir simplifies locking.
951 	 */
952 	if (ovl_need_index(c->dentry)) {
953 		c->indexed = true;
954 		if (S_ISDIR(c->stat.mode))
955 			c->workdir = ovl_indexdir(c->dentry->d_sb);
956 		else
957 			to_index = true;
958 	}
959 
960 	if (S_ISDIR(c->stat.mode) || c->stat.nlink == 1 || to_index) {
961 		fh = ovl_get_origin_fh(ofs, origin);
962 		if (IS_ERR(fh))
963 			return PTR_ERR(fh);
964 
965 		/* origin_fh may be NULL */
966 		c->origin_fh = fh;
967 		c->origin = true;
968 	}
969 
970 	if (to_index) {
971 		c->destdir = ovl_indexdir(c->dentry->d_sb);
972 		err = ovl_get_index_name(ofs, origin, &c->destname);
973 		if (err)
974 			goto out_free_fh;
975 	} else if (WARN_ON(!c->parent)) {
976 		/* Disconnected dentry must be copied up to index dir */
977 		err = -EIO;
978 		goto out_free_fh;
979 	} else {
980 		/*
981 		 * c->dentry->d_name is stabilzed by ovl_copy_up_start(),
982 		 * because if we got here, it means that c->dentry has no upper
983 		 * alias and changing ->d_name means going through ovl_rename()
984 		 * that will call ovl_copy_up() on source and target dentry.
985 		 */
986 		c->destname = c->dentry->d_name;
987 		/*
988 		 * Mark parent "impure" because it may now contain non-pure
989 		 * upper
990 		 */
991 		ovl_start_write(c->dentry);
992 		err = ovl_set_impure(c->parent, c->destdir);
993 		ovl_end_write(c->dentry);
994 		if (err)
995 			goto out_free_fh;
996 	}
997 
998 	/* Should we copyup with O_TMPFILE or with workdir? */
999 	if (S_ISREG(c->stat.mode) && ofs->tmpfile)
1000 		err = ovl_copy_up_tmpfile(c);
1001 	else
1002 		err = ovl_copy_up_workdir(c);
1003 	if (err)
1004 		goto out;
1005 
1006 	if (c->indexed)
1007 		ovl_set_flag(OVL_INDEX, d_inode(c->dentry));
1008 
1009 	ovl_start_write(c->dentry);
1010 	if (to_index) {
1011 		/* Initialize nlink for copy up of disconnected dentry */
1012 		err = ovl_set_nlink_upper(c->dentry);
1013 	} else {
1014 		struct inode *udir = d_inode(c->destdir);
1015 
1016 		/* Restore timestamps on parent (best effort) */
1017 		inode_lock(udir);
1018 		ovl_set_timestamps(ofs, c->destdir, &c->pstat);
1019 		inode_unlock(udir);
1020 
1021 		ovl_dentry_set_upper_alias(c->dentry);
1022 		ovl_dentry_update_reval(c->dentry, ovl_dentry_upper(c->dentry));
1023 	}
1024 	ovl_end_write(c->dentry);
1025 
1026 out:
1027 	if (to_index)
1028 		kfree(c->destname.name);
1029 out_free_fh:
1030 	kfree(fh);
1031 	return err;
1032 }
1033 
1034 static bool ovl_need_meta_copy_up(struct dentry *dentry, umode_t mode,
1035 				  int flags)
1036 {
1037 	struct ovl_fs *ofs = OVL_FS(dentry->d_sb);
1038 
1039 	if (!ofs->config.metacopy)
1040 		return false;
1041 
1042 	if (!S_ISREG(mode))
1043 		return false;
1044 
1045 	if (flags && ((OPEN_FMODE(flags) & FMODE_WRITE) || (flags & O_TRUNC)))
1046 		return false;
1047 
1048 	/* Fall back to full copy if no fsverity on source data and we require verity */
1049 	if (ofs->config.verity_mode == OVL_VERITY_REQUIRE) {
1050 		struct path lowerdata;
1051 
1052 		ovl_path_lowerdata(dentry, &lowerdata);
1053 
1054 		if (WARN_ON_ONCE(lowerdata.dentry == NULL) ||
1055 		    ovl_ensure_verity_loaded(&lowerdata) ||
1056 		    !fsverity_active(d_inode(lowerdata.dentry))) {
1057 			return false;
1058 		}
1059 	}
1060 
1061 	return true;
1062 }
1063 
1064 static ssize_t ovl_getxattr_value(const struct path *path, char *name, char **value)
1065 {
1066 	ssize_t res;
1067 	char *buf;
1068 
1069 	res = ovl_do_getxattr(path, name, NULL, 0);
1070 	if (res == -ENODATA || res == -EOPNOTSUPP)
1071 		res = 0;
1072 
1073 	if (res > 0) {
1074 		buf = kzalloc(res, GFP_KERNEL);
1075 		if (!buf)
1076 			return -ENOMEM;
1077 
1078 		res = ovl_do_getxattr(path, name, buf, res);
1079 		if (res < 0)
1080 			kfree(buf);
1081 		else
1082 			*value = buf;
1083 	}
1084 	return res;
1085 }
1086 
1087 /* Copy up data of an inode which was copied up metadata only in the past. */
1088 static int ovl_copy_up_meta_inode_data(struct ovl_copy_up_ctx *c)
1089 {
1090 	struct ovl_fs *ofs = OVL_FS(c->dentry->d_sb);
1091 	struct path upperpath;
1092 	int err;
1093 	char *capability = NULL;
1094 	ssize_t cap_size;
1095 
1096 	ovl_path_upper(c->dentry, &upperpath);
1097 	if (WARN_ON(upperpath.dentry == NULL))
1098 		return -EIO;
1099 
1100 	if (c->stat.size) {
1101 		err = cap_size = ovl_getxattr_value(&upperpath, XATTR_NAME_CAPS,
1102 						    &capability);
1103 		if (cap_size < 0)
1104 			goto out;
1105 	}
1106 
1107 	err = ovl_copy_up_data(c, &upperpath);
1108 	if (err)
1109 		goto out_free;
1110 
1111 	/*
1112 	 * Writing to upper file will clear security.capability xattr. We
1113 	 * don't want that to happen for normal copy-up operation.
1114 	 */
1115 	ovl_start_write(c->dentry);
1116 	if (capability) {
1117 		err = ovl_do_setxattr(ofs, upperpath.dentry, XATTR_NAME_CAPS,
1118 				      capability, cap_size, 0);
1119 	}
1120 	if (!err) {
1121 		err = ovl_removexattr(ofs, upperpath.dentry,
1122 				      OVL_XATTR_METACOPY);
1123 	}
1124 	ovl_end_write(c->dentry);
1125 	if (err)
1126 		goto out_free;
1127 
1128 	ovl_clear_flag(OVL_HAS_DIGEST, d_inode(c->dentry));
1129 	ovl_clear_flag(OVL_VERIFIED_DIGEST, d_inode(c->dentry));
1130 	ovl_set_upperdata(d_inode(c->dentry));
1131 out_free:
1132 	kfree(capability);
1133 out:
1134 	return err;
1135 }
1136 
1137 static int ovl_copy_up_one(struct dentry *parent, struct dentry *dentry,
1138 			   int flags)
1139 {
1140 	int err;
1141 	DEFINE_DELAYED_CALL(done);
1142 	struct path parentpath;
1143 	struct ovl_copy_up_ctx ctx = {
1144 		.parent = parent,
1145 		.dentry = dentry,
1146 		.workdir = ovl_workdir(dentry),
1147 	};
1148 
1149 	if (WARN_ON(!ctx.workdir))
1150 		return -EROFS;
1151 
1152 	ovl_path_lower(dentry, &ctx.lowerpath);
1153 	err = vfs_getattr(&ctx.lowerpath, &ctx.stat,
1154 			  STATX_BASIC_STATS, AT_STATX_SYNC_AS_STAT);
1155 	if (err)
1156 		return err;
1157 
1158 	if (!kuid_has_mapping(current_user_ns(), ctx.stat.uid) ||
1159 	    !kgid_has_mapping(current_user_ns(), ctx.stat.gid))
1160 		return -EOVERFLOW;
1161 
1162 	/*
1163 	 * With metacopy disabled, we fsync after final metadata copyup, for
1164 	 * both regular files and directories to get atomic copyup semantics
1165 	 * on filesystems that do not use strict metadata ordering (e.g. ubifs).
1166 	 *
1167 	 * With metacopy enabled we want to avoid fsync on all meta copyup
1168 	 * that will hurt performance of workloads such as chown -R, so we
1169 	 * only fsync on data copyup as legacy behavior.
1170 	 */
1171 	ctx.metadata_fsync = !OVL_FS(dentry->d_sb)->config.metacopy &&
1172 			     (S_ISREG(ctx.stat.mode) || S_ISDIR(ctx.stat.mode));
1173 	ctx.metacopy = ovl_need_meta_copy_up(dentry, ctx.stat.mode, flags);
1174 
1175 	if (parent) {
1176 		ovl_path_upper(parent, &parentpath);
1177 		ctx.destdir = parentpath.dentry;
1178 
1179 		err = vfs_getattr(&parentpath, &ctx.pstat,
1180 				  STATX_ATIME | STATX_MTIME,
1181 				  AT_STATX_SYNC_AS_STAT);
1182 		if (err)
1183 			return err;
1184 	}
1185 
1186 	/* maybe truncate regular file. this has no effect on dirs */
1187 	if (flags & O_TRUNC)
1188 		ctx.stat.size = 0;
1189 
1190 	if (S_ISLNK(ctx.stat.mode)) {
1191 		ctx.link = vfs_get_link(ctx.lowerpath.dentry, &done);
1192 		if (IS_ERR(ctx.link))
1193 			return PTR_ERR(ctx.link);
1194 	}
1195 
1196 	err = ovl_copy_up_start(dentry, flags);
1197 	/* err < 0: interrupted, err > 0: raced with another copy-up */
1198 	if (unlikely(err)) {
1199 		if (err > 0)
1200 			err = 0;
1201 	} else {
1202 		if (!ovl_dentry_upper(dentry))
1203 			err = ovl_do_copy_up(&ctx);
1204 		if (!err && parent && !ovl_dentry_has_upper_alias(dentry))
1205 			err = ovl_link_up(&ctx);
1206 		if (!err && ovl_dentry_needs_data_copy_up_locked(dentry, flags))
1207 			err = ovl_copy_up_meta_inode_data(&ctx);
1208 		ovl_copy_up_end(dentry);
1209 	}
1210 	do_delayed_call(&done);
1211 
1212 	return err;
1213 }
1214 
1215 static int ovl_copy_up_flags(struct dentry *dentry, int flags)
1216 {
1217 	int err = 0;
1218 	const struct cred *old_cred;
1219 	bool disconnected = (dentry->d_flags & DCACHE_DISCONNECTED);
1220 
1221 	/*
1222 	 * With NFS export, copy up can get called for a disconnected non-dir.
1223 	 * In this case, we will copy up lower inode to index dir without
1224 	 * linking it to upper dir.
1225 	 */
1226 	if (WARN_ON(disconnected && d_is_dir(dentry)))
1227 		return -EIO;
1228 
1229 	/*
1230 	 * We may not need lowerdata if we are only doing metacopy up, but it is
1231 	 * not very important to optimize this case, so do lazy lowerdata lookup
1232 	 * before any copy up, so we can do it before taking ovl_inode_lock().
1233 	 */
1234 	err = ovl_verify_lowerdata(dentry);
1235 	if (err)
1236 		return err;
1237 
1238 	old_cred = ovl_override_creds(dentry->d_sb);
1239 	while (!err) {
1240 		struct dentry *next;
1241 		struct dentry *parent = NULL;
1242 
1243 		if (ovl_already_copied_up(dentry, flags))
1244 			break;
1245 
1246 		next = dget(dentry);
1247 		/* find the topmost dentry not yet copied up */
1248 		for (; !disconnected;) {
1249 			parent = dget_parent(next);
1250 
1251 			if (ovl_dentry_upper(parent))
1252 				break;
1253 
1254 			dput(next);
1255 			next = parent;
1256 		}
1257 
1258 		err = ovl_copy_up_one(parent, next, flags);
1259 
1260 		dput(parent);
1261 		dput(next);
1262 	}
1263 	revert_creds(old_cred);
1264 
1265 	return err;
1266 }
1267 
1268 static bool ovl_open_need_copy_up(struct dentry *dentry, int flags)
1269 {
1270 	/* Copy up of disconnected dentry does not set upper alias */
1271 	if (ovl_already_copied_up(dentry, flags))
1272 		return false;
1273 
1274 	if (special_file(d_inode(dentry)->i_mode))
1275 		return false;
1276 
1277 	if (!ovl_open_flags_need_copy_up(flags))
1278 		return false;
1279 
1280 	return true;
1281 }
1282 
1283 int ovl_maybe_copy_up(struct dentry *dentry, int flags)
1284 {
1285 	if (!ovl_open_need_copy_up(dentry, flags))
1286 		return 0;
1287 
1288 	return ovl_copy_up_flags(dentry, flags);
1289 }
1290 
1291 int ovl_copy_up_with_data(struct dentry *dentry)
1292 {
1293 	return ovl_copy_up_flags(dentry, O_WRONLY);
1294 }
1295 
1296 int ovl_copy_up(struct dentry *dentry)
1297 {
1298 	return ovl_copy_up_flags(dentry, 0);
1299 }
1300