xref: /linux/fs/stat.c (revision c1f3caff2450048ab6c053e5b23698b58f286159)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  *  linux/fs/stat.c
4  *
5  *  Copyright (C) 1991, 1992  Linus Torvalds
6  */
7 
8 #include <linux/export.h>
9 #include <linux/mm.h>
10 #include <linux/errno.h>
11 #include <linux/file.h>
12 #include <linux/highuid.h>
13 #include <linux/fs.h>
14 #include <linux/namei.h>
15 #include <linux/security.h>
16 #include <linux/cred.h>
17 #include <linux/syscalls.h>
18 #include <linux/pagemap.h>
19 #include <linux/compat.h>
20 
21 #include <linux/uaccess.h>
22 #include <asm/unistd.h>
23 
24 #include "internal.h"
25 #include "mount.h"
26 
27 /**
28  * generic_fillattr - Fill in the basic attributes from the inode struct
29  * @mnt_userns:	user namespace of the mount the inode was found from
30  * @inode:	Inode to use as the source
31  * @stat:	Where to fill in the attributes
32  *
33  * Fill in the basic attributes in the kstat structure from data that's to be
34  * found on the VFS inode structure.  This is the default if no getattr inode
35  * operation is supplied.
36  *
37  * If the inode has been found through an idmapped mount the user namespace of
38  * the vfsmount must be passed through @mnt_userns. This function will then
39  * take care to map the inode according to @mnt_userns before filling in the
40  * uid and gid filds. On non-idmapped mounts or if permission checking is to be
41  * performed on the raw inode simply passs init_user_ns.
42  */
43 void generic_fillattr(struct user_namespace *mnt_userns, struct inode *inode,
44 		      struct kstat *stat)
45 {
46 	stat->dev = inode->i_sb->s_dev;
47 	stat->ino = inode->i_ino;
48 	stat->mode = inode->i_mode;
49 	stat->nlink = inode->i_nlink;
50 	stat->uid = i_uid_into_mnt(mnt_userns, inode);
51 	stat->gid = i_gid_into_mnt(mnt_userns, inode);
52 	stat->rdev = inode->i_rdev;
53 	stat->size = i_size_read(inode);
54 	stat->atime = inode->i_atime;
55 	stat->mtime = inode->i_mtime;
56 	stat->ctime = inode->i_ctime;
57 	stat->blksize = i_blocksize(inode);
58 	stat->blocks = inode->i_blocks;
59 }
60 EXPORT_SYMBOL(generic_fillattr);
61 
62 /**
63  * generic_fill_statx_attr - Fill in the statx attributes from the inode flags
64  * @inode:	Inode to use as the source
65  * @stat:	Where to fill in the attribute flags
66  *
67  * Fill in the STATX_ATTR_* flags in the kstat structure for properties of the
68  * inode that are published on i_flags and enforced by the VFS.
69  */
70 void generic_fill_statx_attr(struct inode *inode, struct kstat *stat)
71 {
72 	if (inode->i_flags & S_IMMUTABLE)
73 		stat->attributes |= STATX_ATTR_IMMUTABLE;
74 	if (inode->i_flags & S_APPEND)
75 		stat->attributes |= STATX_ATTR_APPEND;
76 	stat->attributes_mask |= KSTAT_ATTR_VFS_FLAGS;
77 }
78 EXPORT_SYMBOL(generic_fill_statx_attr);
79 
80 /**
81  * vfs_getattr_nosec - getattr without security checks
82  * @path: file to get attributes from
83  * @stat: structure to return attributes in
84  * @request_mask: STATX_xxx flags indicating what the caller wants
85  * @query_flags: Query mode (AT_STATX_SYNC_TYPE)
86  *
87  * Get attributes without calling security_inode_getattr.
88  *
89  * Currently the only caller other than vfs_getattr is internal to the
90  * filehandle lookup code, which uses only the inode number and returns no
91  * attributes to any user.  Any other code probably wants vfs_getattr.
92  */
93 int vfs_getattr_nosec(const struct path *path, struct kstat *stat,
94 		      u32 request_mask, unsigned int query_flags)
95 {
96 	struct user_namespace *mnt_userns;
97 	struct inode *inode = d_backing_inode(path->dentry);
98 
99 	memset(stat, 0, sizeof(*stat));
100 	stat->result_mask |= STATX_BASIC_STATS;
101 	query_flags &= AT_STATX_SYNC_TYPE;
102 
103 	/* allow the fs to override these if it really wants to */
104 	/* SB_NOATIME means filesystem supplies dummy atime value */
105 	if (inode->i_sb->s_flags & SB_NOATIME)
106 		stat->result_mask &= ~STATX_ATIME;
107 
108 	/*
109 	 * Note: If you add another clause to set an attribute flag, please
110 	 * update attributes_mask below.
111 	 */
112 	if (IS_AUTOMOUNT(inode))
113 		stat->attributes |= STATX_ATTR_AUTOMOUNT;
114 
115 	if (IS_DAX(inode))
116 		stat->attributes |= STATX_ATTR_DAX;
117 
118 	stat->attributes_mask |= (STATX_ATTR_AUTOMOUNT |
119 				  STATX_ATTR_DAX);
120 
121 	mnt_userns = mnt_user_ns(path->mnt);
122 	if (inode->i_op->getattr)
123 		return inode->i_op->getattr(mnt_userns, path, stat,
124 					    request_mask, query_flags);
125 
126 	generic_fillattr(mnt_userns, inode, stat);
127 	return 0;
128 }
129 EXPORT_SYMBOL(vfs_getattr_nosec);
130 
131 /*
132  * vfs_getattr - Get the enhanced basic attributes of a file
133  * @path: The file of interest
134  * @stat: Where to return the statistics
135  * @request_mask: STATX_xxx flags indicating what the caller wants
136  * @query_flags: Query mode (AT_STATX_SYNC_TYPE)
137  *
138  * Ask the filesystem for a file's attributes.  The caller must indicate in
139  * request_mask and query_flags to indicate what they want.
140  *
141  * If the file is remote, the filesystem can be forced to update the attributes
142  * from the backing store by passing AT_STATX_FORCE_SYNC in query_flags or can
143  * suppress the update by passing AT_STATX_DONT_SYNC.
144  *
145  * Bits must have been set in request_mask to indicate which attributes the
146  * caller wants retrieving.  Any such attribute not requested may be returned
147  * anyway, but the value may be approximate, and, if remote, may not have been
148  * synchronised with the server.
149  *
150  * 0 will be returned on success, and a -ve error code if unsuccessful.
151  */
152 int vfs_getattr(const struct path *path, struct kstat *stat,
153 		u32 request_mask, unsigned int query_flags)
154 {
155 	int retval;
156 
157 	retval = security_inode_getattr(path);
158 	if (retval)
159 		return retval;
160 	return vfs_getattr_nosec(path, stat, request_mask, query_flags);
161 }
162 EXPORT_SYMBOL(vfs_getattr);
163 
164 /**
165  * vfs_fstat - Get the basic attributes by file descriptor
166  * @fd: The file descriptor referring to the file of interest
167  * @stat: The result structure to fill in.
168  *
169  * This function is a wrapper around vfs_getattr().  The main difference is
170  * that it uses a file descriptor to determine the file location.
171  *
172  * 0 will be returned on success, and a -ve error code if unsuccessful.
173  */
174 int vfs_fstat(int fd, struct kstat *stat)
175 {
176 	struct fd f;
177 	int error;
178 
179 	f = fdget_raw(fd);
180 	if (!f.file)
181 		return -EBADF;
182 	error = vfs_getattr(&f.file->f_path, stat, STATX_BASIC_STATS, 0);
183 	fdput(f);
184 	return error;
185 }
186 
187 /**
188  * vfs_statx - Get basic and extra attributes by filename
189  * @dfd: A file descriptor representing the base dir for a relative filename
190  * @filename: The name of the file of interest
191  * @flags: Flags to control the query
192  * @stat: The result structure to fill in.
193  * @request_mask: STATX_xxx flags indicating what the caller wants
194  *
195  * This function is a wrapper around vfs_getattr().  The main difference is
196  * that it uses a filename and base directory to determine the file location.
197  * Additionally, the use of AT_SYMLINK_NOFOLLOW in flags will prevent a symlink
198  * at the given name from being referenced.
199  *
200  * 0 will be returned on success, and a -ve error code if unsuccessful.
201  */
202 static int vfs_statx(int dfd, const char __user *filename, int flags,
203 	      struct kstat *stat, u32 request_mask)
204 {
205 	struct path path;
206 	unsigned lookup_flags = 0;
207 	int error;
208 
209 	if (flags & ~(AT_SYMLINK_NOFOLLOW | AT_NO_AUTOMOUNT | AT_EMPTY_PATH |
210 		      AT_STATX_SYNC_TYPE))
211 		return -EINVAL;
212 
213 	if (!(flags & AT_SYMLINK_NOFOLLOW))
214 		lookup_flags |= LOOKUP_FOLLOW;
215 	if (!(flags & AT_NO_AUTOMOUNT))
216 		lookup_flags |= LOOKUP_AUTOMOUNT;
217 	if (flags & AT_EMPTY_PATH)
218 		lookup_flags |= LOOKUP_EMPTY;
219 
220 retry:
221 	error = user_path_at(dfd, filename, lookup_flags, &path);
222 	if (error)
223 		goto out;
224 
225 	error = vfs_getattr(&path, stat, request_mask, flags);
226 	stat->mnt_id = real_mount(path.mnt)->mnt_id;
227 	stat->result_mask |= STATX_MNT_ID;
228 	if (path.mnt->mnt_root == path.dentry)
229 		stat->attributes |= STATX_ATTR_MOUNT_ROOT;
230 	stat->attributes_mask |= STATX_ATTR_MOUNT_ROOT;
231 	path_put(&path);
232 	if (retry_estale(error, lookup_flags)) {
233 		lookup_flags |= LOOKUP_REVAL;
234 		goto retry;
235 	}
236 out:
237 	return error;
238 }
239 
240 int vfs_fstatat(int dfd, const char __user *filename,
241 			      struct kstat *stat, int flags)
242 {
243 	return vfs_statx(dfd, filename, flags | AT_NO_AUTOMOUNT,
244 			 stat, STATX_BASIC_STATS);
245 }
246 
247 #ifdef __ARCH_WANT_OLD_STAT
248 
249 /*
250  * For backward compatibility?  Maybe this should be moved
251  * into arch/i386 instead?
252  */
253 static int cp_old_stat(struct kstat *stat, struct __old_kernel_stat __user * statbuf)
254 {
255 	static int warncount = 5;
256 	struct __old_kernel_stat tmp;
257 
258 	if (warncount > 0) {
259 		warncount--;
260 		printk(KERN_WARNING "VFS: Warning: %s using old stat() call. Recompile your binary.\n",
261 			current->comm);
262 	} else if (warncount < 0) {
263 		/* it's laughable, but... */
264 		warncount = 0;
265 	}
266 
267 	memset(&tmp, 0, sizeof(struct __old_kernel_stat));
268 	tmp.st_dev = old_encode_dev(stat->dev);
269 	tmp.st_ino = stat->ino;
270 	if (sizeof(tmp.st_ino) < sizeof(stat->ino) && tmp.st_ino != stat->ino)
271 		return -EOVERFLOW;
272 	tmp.st_mode = stat->mode;
273 	tmp.st_nlink = stat->nlink;
274 	if (tmp.st_nlink != stat->nlink)
275 		return -EOVERFLOW;
276 	SET_UID(tmp.st_uid, from_kuid_munged(current_user_ns(), stat->uid));
277 	SET_GID(tmp.st_gid, from_kgid_munged(current_user_ns(), stat->gid));
278 	tmp.st_rdev = old_encode_dev(stat->rdev);
279 #if BITS_PER_LONG == 32
280 	if (stat->size > MAX_NON_LFS)
281 		return -EOVERFLOW;
282 #endif
283 	tmp.st_size = stat->size;
284 	tmp.st_atime = stat->atime.tv_sec;
285 	tmp.st_mtime = stat->mtime.tv_sec;
286 	tmp.st_ctime = stat->ctime.tv_sec;
287 	return copy_to_user(statbuf,&tmp,sizeof(tmp)) ? -EFAULT : 0;
288 }
289 
290 SYSCALL_DEFINE2(stat, const char __user *, filename,
291 		struct __old_kernel_stat __user *, statbuf)
292 {
293 	struct kstat stat;
294 	int error;
295 
296 	error = vfs_stat(filename, &stat);
297 	if (error)
298 		return error;
299 
300 	return cp_old_stat(&stat, statbuf);
301 }
302 
303 SYSCALL_DEFINE2(lstat, const char __user *, filename,
304 		struct __old_kernel_stat __user *, statbuf)
305 {
306 	struct kstat stat;
307 	int error;
308 
309 	error = vfs_lstat(filename, &stat);
310 	if (error)
311 		return error;
312 
313 	return cp_old_stat(&stat, statbuf);
314 }
315 
316 SYSCALL_DEFINE2(fstat, unsigned int, fd, struct __old_kernel_stat __user *, statbuf)
317 {
318 	struct kstat stat;
319 	int error = vfs_fstat(fd, &stat);
320 
321 	if (!error)
322 		error = cp_old_stat(&stat, statbuf);
323 
324 	return error;
325 }
326 
327 #endif /* __ARCH_WANT_OLD_STAT */
328 
329 #ifdef __ARCH_WANT_NEW_STAT
330 
331 #if BITS_PER_LONG == 32
332 #  define choose_32_64(a,b) a
333 #else
334 #  define choose_32_64(a,b) b
335 #endif
336 
337 #define valid_dev(x)  choose_32_64(old_valid_dev(x),true)
338 #define encode_dev(x) choose_32_64(old_encode_dev,new_encode_dev)(x)
339 
340 #ifndef INIT_STRUCT_STAT_PADDING
341 #  define INIT_STRUCT_STAT_PADDING(st) memset(&st, 0, sizeof(st))
342 #endif
343 
344 static int cp_new_stat(struct kstat *stat, struct stat __user *statbuf)
345 {
346 	struct stat tmp;
347 
348 	if (!valid_dev(stat->dev) || !valid_dev(stat->rdev))
349 		return -EOVERFLOW;
350 #if BITS_PER_LONG == 32
351 	if (stat->size > MAX_NON_LFS)
352 		return -EOVERFLOW;
353 #endif
354 
355 	INIT_STRUCT_STAT_PADDING(tmp);
356 	tmp.st_dev = encode_dev(stat->dev);
357 	tmp.st_ino = stat->ino;
358 	if (sizeof(tmp.st_ino) < sizeof(stat->ino) && tmp.st_ino != stat->ino)
359 		return -EOVERFLOW;
360 	tmp.st_mode = stat->mode;
361 	tmp.st_nlink = stat->nlink;
362 	if (tmp.st_nlink != stat->nlink)
363 		return -EOVERFLOW;
364 	SET_UID(tmp.st_uid, from_kuid_munged(current_user_ns(), stat->uid));
365 	SET_GID(tmp.st_gid, from_kgid_munged(current_user_ns(), stat->gid));
366 	tmp.st_rdev = encode_dev(stat->rdev);
367 	tmp.st_size = stat->size;
368 	tmp.st_atime = stat->atime.tv_sec;
369 	tmp.st_mtime = stat->mtime.tv_sec;
370 	tmp.st_ctime = stat->ctime.tv_sec;
371 #ifdef STAT_HAVE_NSEC
372 	tmp.st_atime_nsec = stat->atime.tv_nsec;
373 	tmp.st_mtime_nsec = stat->mtime.tv_nsec;
374 	tmp.st_ctime_nsec = stat->ctime.tv_nsec;
375 #endif
376 	tmp.st_blocks = stat->blocks;
377 	tmp.st_blksize = stat->blksize;
378 	return copy_to_user(statbuf,&tmp,sizeof(tmp)) ? -EFAULT : 0;
379 }
380 
381 SYSCALL_DEFINE2(newstat, const char __user *, filename,
382 		struct stat __user *, statbuf)
383 {
384 	struct kstat stat;
385 	int error = vfs_stat(filename, &stat);
386 
387 	if (error)
388 		return error;
389 	return cp_new_stat(&stat, statbuf);
390 }
391 
392 SYSCALL_DEFINE2(newlstat, const char __user *, filename,
393 		struct stat __user *, statbuf)
394 {
395 	struct kstat stat;
396 	int error;
397 
398 	error = vfs_lstat(filename, &stat);
399 	if (error)
400 		return error;
401 
402 	return cp_new_stat(&stat, statbuf);
403 }
404 
405 #if !defined(__ARCH_WANT_STAT64) || defined(__ARCH_WANT_SYS_NEWFSTATAT)
406 SYSCALL_DEFINE4(newfstatat, int, dfd, const char __user *, filename,
407 		struct stat __user *, statbuf, int, flag)
408 {
409 	struct kstat stat;
410 	int error;
411 
412 	error = vfs_fstatat(dfd, filename, &stat, flag);
413 	if (error)
414 		return error;
415 	return cp_new_stat(&stat, statbuf);
416 }
417 #endif
418 
419 SYSCALL_DEFINE2(newfstat, unsigned int, fd, struct stat __user *, statbuf)
420 {
421 	struct kstat stat;
422 	int error = vfs_fstat(fd, &stat);
423 
424 	if (!error)
425 		error = cp_new_stat(&stat, statbuf);
426 
427 	return error;
428 }
429 #endif
430 
431 static int do_readlinkat(int dfd, const char __user *pathname,
432 			 char __user *buf, int bufsiz)
433 {
434 	struct path path;
435 	int error;
436 	int empty = 0;
437 	unsigned int lookup_flags = LOOKUP_EMPTY;
438 
439 	if (bufsiz <= 0)
440 		return -EINVAL;
441 
442 retry:
443 	error = user_path_at_empty(dfd, pathname, lookup_flags, &path, &empty);
444 	if (!error) {
445 		struct inode *inode = d_backing_inode(path.dentry);
446 
447 		error = empty ? -ENOENT : -EINVAL;
448 		/*
449 		 * AFS mountpoints allow readlink(2) but are not symlinks
450 		 */
451 		if (d_is_symlink(path.dentry) || inode->i_op->readlink) {
452 			error = security_inode_readlink(path.dentry);
453 			if (!error) {
454 				touch_atime(&path);
455 				error = vfs_readlink(path.dentry, buf, bufsiz);
456 			}
457 		}
458 		path_put(&path);
459 		if (retry_estale(error, lookup_flags)) {
460 			lookup_flags |= LOOKUP_REVAL;
461 			goto retry;
462 		}
463 	}
464 	return error;
465 }
466 
467 SYSCALL_DEFINE4(readlinkat, int, dfd, const char __user *, pathname,
468 		char __user *, buf, int, bufsiz)
469 {
470 	return do_readlinkat(dfd, pathname, buf, bufsiz);
471 }
472 
473 SYSCALL_DEFINE3(readlink, const char __user *, path, char __user *, buf,
474 		int, bufsiz)
475 {
476 	return do_readlinkat(AT_FDCWD, path, buf, bufsiz);
477 }
478 
479 
480 /* ---------- LFS-64 ----------- */
481 #if defined(__ARCH_WANT_STAT64) || defined(__ARCH_WANT_COMPAT_STAT64)
482 
483 #ifndef INIT_STRUCT_STAT64_PADDING
484 #  define INIT_STRUCT_STAT64_PADDING(st) memset(&st, 0, sizeof(st))
485 #endif
486 
487 static long cp_new_stat64(struct kstat *stat, struct stat64 __user *statbuf)
488 {
489 	struct stat64 tmp;
490 
491 	INIT_STRUCT_STAT64_PADDING(tmp);
492 #ifdef CONFIG_MIPS
493 	/* mips has weird padding, so we don't get 64 bits there */
494 	tmp.st_dev = new_encode_dev(stat->dev);
495 	tmp.st_rdev = new_encode_dev(stat->rdev);
496 #else
497 	tmp.st_dev = huge_encode_dev(stat->dev);
498 	tmp.st_rdev = huge_encode_dev(stat->rdev);
499 #endif
500 	tmp.st_ino = stat->ino;
501 	if (sizeof(tmp.st_ino) < sizeof(stat->ino) && tmp.st_ino != stat->ino)
502 		return -EOVERFLOW;
503 #ifdef STAT64_HAS_BROKEN_ST_INO
504 	tmp.__st_ino = stat->ino;
505 #endif
506 	tmp.st_mode = stat->mode;
507 	tmp.st_nlink = stat->nlink;
508 	tmp.st_uid = from_kuid_munged(current_user_ns(), stat->uid);
509 	tmp.st_gid = from_kgid_munged(current_user_ns(), stat->gid);
510 	tmp.st_atime = stat->atime.tv_sec;
511 	tmp.st_atime_nsec = stat->atime.tv_nsec;
512 	tmp.st_mtime = stat->mtime.tv_sec;
513 	tmp.st_mtime_nsec = stat->mtime.tv_nsec;
514 	tmp.st_ctime = stat->ctime.tv_sec;
515 	tmp.st_ctime_nsec = stat->ctime.tv_nsec;
516 	tmp.st_size = stat->size;
517 	tmp.st_blocks = stat->blocks;
518 	tmp.st_blksize = stat->blksize;
519 	return copy_to_user(statbuf,&tmp,sizeof(tmp)) ? -EFAULT : 0;
520 }
521 
522 SYSCALL_DEFINE2(stat64, const char __user *, filename,
523 		struct stat64 __user *, statbuf)
524 {
525 	struct kstat stat;
526 	int error = vfs_stat(filename, &stat);
527 
528 	if (!error)
529 		error = cp_new_stat64(&stat, statbuf);
530 
531 	return error;
532 }
533 
534 SYSCALL_DEFINE2(lstat64, const char __user *, filename,
535 		struct stat64 __user *, statbuf)
536 {
537 	struct kstat stat;
538 	int error = vfs_lstat(filename, &stat);
539 
540 	if (!error)
541 		error = cp_new_stat64(&stat, statbuf);
542 
543 	return error;
544 }
545 
546 SYSCALL_DEFINE2(fstat64, unsigned long, fd, struct stat64 __user *, statbuf)
547 {
548 	struct kstat stat;
549 	int error = vfs_fstat(fd, &stat);
550 
551 	if (!error)
552 		error = cp_new_stat64(&stat, statbuf);
553 
554 	return error;
555 }
556 
557 SYSCALL_DEFINE4(fstatat64, int, dfd, const char __user *, filename,
558 		struct stat64 __user *, statbuf, int, flag)
559 {
560 	struct kstat stat;
561 	int error;
562 
563 	error = vfs_fstatat(dfd, filename, &stat, flag);
564 	if (error)
565 		return error;
566 	return cp_new_stat64(&stat, statbuf);
567 }
568 #endif /* __ARCH_WANT_STAT64 || __ARCH_WANT_COMPAT_STAT64 */
569 
570 static noinline_for_stack int
571 cp_statx(const struct kstat *stat, struct statx __user *buffer)
572 {
573 	struct statx tmp;
574 
575 	memset(&tmp, 0, sizeof(tmp));
576 
577 	tmp.stx_mask = stat->result_mask;
578 	tmp.stx_blksize = stat->blksize;
579 	tmp.stx_attributes = stat->attributes;
580 	tmp.stx_nlink = stat->nlink;
581 	tmp.stx_uid = from_kuid_munged(current_user_ns(), stat->uid);
582 	tmp.stx_gid = from_kgid_munged(current_user_ns(), stat->gid);
583 	tmp.stx_mode = stat->mode;
584 	tmp.stx_ino = stat->ino;
585 	tmp.stx_size = stat->size;
586 	tmp.stx_blocks = stat->blocks;
587 	tmp.stx_attributes_mask = stat->attributes_mask;
588 	tmp.stx_atime.tv_sec = stat->atime.tv_sec;
589 	tmp.stx_atime.tv_nsec = stat->atime.tv_nsec;
590 	tmp.stx_btime.tv_sec = stat->btime.tv_sec;
591 	tmp.stx_btime.tv_nsec = stat->btime.tv_nsec;
592 	tmp.stx_ctime.tv_sec = stat->ctime.tv_sec;
593 	tmp.stx_ctime.tv_nsec = stat->ctime.tv_nsec;
594 	tmp.stx_mtime.tv_sec = stat->mtime.tv_sec;
595 	tmp.stx_mtime.tv_nsec = stat->mtime.tv_nsec;
596 	tmp.stx_rdev_major = MAJOR(stat->rdev);
597 	tmp.stx_rdev_minor = MINOR(stat->rdev);
598 	tmp.stx_dev_major = MAJOR(stat->dev);
599 	tmp.stx_dev_minor = MINOR(stat->dev);
600 	tmp.stx_mnt_id = stat->mnt_id;
601 
602 	return copy_to_user(buffer, &tmp, sizeof(tmp)) ? -EFAULT : 0;
603 }
604 
605 int do_statx(int dfd, const char __user *filename, unsigned flags,
606 	     unsigned int mask, struct statx __user *buffer)
607 {
608 	struct kstat stat;
609 	int error;
610 
611 	if (mask & STATX__RESERVED)
612 		return -EINVAL;
613 	if ((flags & AT_STATX_SYNC_TYPE) == AT_STATX_SYNC_TYPE)
614 		return -EINVAL;
615 
616 	error = vfs_statx(dfd, filename, flags, &stat, mask);
617 	if (error)
618 		return error;
619 
620 	return cp_statx(&stat, buffer);
621 }
622 
623 /**
624  * sys_statx - System call to get enhanced stats
625  * @dfd: Base directory to pathwalk from *or* fd to stat.
626  * @filename: File to stat or "" with AT_EMPTY_PATH
627  * @flags: AT_* flags to control pathwalk.
628  * @mask: Parts of statx struct actually required.
629  * @buffer: Result buffer.
630  *
631  * Note that fstat() can be emulated by setting dfd to the fd of interest,
632  * supplying "" as the filename and setting AT_EMPTY_PATH in the flags.
633  */
634 SYSCALL_DEFINE5(statx,
635 		int, dfd, const char __user *, filename, unsigned, flags,
636 		unsigned int, mask,
637 		struct statx __user *, buffer)
638 {
639 	return do_statx(dfd, filename, flags, mask, buffer);
640 }
641 
642 #ifdef CONFIG_COMPAT
643 static int cp_compat_stat(struct kstat *stat, struct compat_stat __user *ubuf)
644 {
645 	struct compat_stat tmp;
646 
647 	if (!old_valid_dev(stat->dev) || !old_valid_dev(stat->rdev))
648 		return -EOVERFLOW;
649 
650 	memset(&tmp, 0, sizeof(tmp));
651 	tmp.st_dev = old_encode_dev(stat->dev);
652 	tmp.st_ino = stat->ino;
653 	if (sizeof(tmp.st_ino) < sizeof(stat->ino) && tmp.st_ino != stat->ino)
654 		return -EOVERFLOW;
655 	tmp.st_mode = stat->mode;
656 	tmp.st_nlink = stat->nlink;
657 	if (tmp.st_nlink != stat->nlink)
658 		return -EOVERFLOW;
659 	SET_UID(tmp.st_uid, from_kuid_munged(current_user_ns(), stat->uid));
660 	SET_GID(tmp.st_gid, from_kgid_munged(current_user_ns(), stat->gid));
661 	tmp.st_rdev = old_encode_dev(stat->rdev);
662 	if ((u64) stat->size > MAX_NON_LFS)
663 		return -EOVERFLOW;
664 	tmp.st_size = stat->size;
665 	tmp.st_atime = stat->atime.tv_sec;
666 	tmp.st_atime_nsec = stat->atime.tv_nsec;
667 	tmp.st_mtime = stat->mtime.tv_sec;
668 	tmp.st_mtime_nsec = stat->mtime.tv_nsec;
669 	tmp.st_ctime = stat->ctime.tv_sec;
670 	tmp.st_ctime_nsec = stat->ctime.tv_nsec;
671 	tmp.st_blocks = stat->blocks;
672 	tmp.st_blksize = stat->blksize;
673 	return copy_to_user(ubuf, &tmp, sizeof(tmp)) ? -EFAULT : 0;
674 }
675 
676 COMPAT_SYSCALL_DEFINE2(newstat, const char __user *, filename,
677 		       struct compat_stat __user *, statbuf)
678 {
679 	struct kstat stat;
680 	int error;
681 
682 	error = vfs_stat(filename, &stat);
683 	if (error)
684 		return error;
685 	return cp_compat_stat(&stat, statbuf);
686 }
687 
688 COMPAT_SYSCALL_DEFINE2(newlstat, const char __user *, filename,
689 		       struct compat_stat __user *, statbuf)
690 {
691 	struct kstat stat;
692 	int error;
693 
694 	error = vfs_lstat(filename, &stat);
695 	if (error)
696 		return error;
697 	return cp_compat_stat(&stat, statbuf);
698 }
699 
700 #ifndef __ARCH_WANT_STAT64
701 COMPAT_SYSCALL_DEFINE4(newfstatat, unsigned int, dfd,
702 		       const char __user *, filename,
703 		       struct compat_stat __user *, statbuf, int, flag)
704 {
705 	struct kstat stat;
706 	int error;
707 
708 	error = vfs_fstatat(dfd, filename, &stat, flag);
709 	if (error)
710 		return error;
711 	return cp_compat_stat(&stat, statbuf);
712 }
713 #endif
714 
715 COMPAT_SYSCALL_DEFINE2(newfstat, unsigned int, fd,
716 		       struct compat_stat __user *, statbuf)
717 {
718 	struct kstat stat;
719 	int error = vfs_fstat(fd, &stat);
720 
721 	if (!error)
722 		error = cp_compat_stat(&stat, statbuf);
723 	return error;
724 }
725 #endif
726 
727 /* Caller is here responsible for sufficient locking (ie. inode->i_lock) */
728 void __inode_add_bytes(struct inode *inode, loff_t bytes)
729 {
730 	inode->i_blocks += bytes >> 9;
731 	bytes &= 511;
732 	inode->i_bytes += bytes;
733 	if (inode->i_bytes >= 512) {
734 		inode->i_blocks++;
735 		inode->i_bytes -= 512;
736 	}
737 }
738 EXPORT_SYMBOL(__inode_add_bytes);
739 
740 void inode_add_bytes(struct inode *inode, loff_t bytes)
741 {
742 	spin_lock(&inode->i_lock);
743 	__inode_add_bytes(inode, bytes);
744 	spin_unlock(&inode->i_lock);
745 }
746 
747 EXPORT_SYMBOL(inode_add_bytes);
748 
749 void __inode_sub_bytes(struct inode *inode, loff_t bytes)
750 {
751 	inode->i_blocks -= bytes >> 9;
752 	bytes &= 511;
753 	if (inode->i_bytes < bytes) {
754 		inode->i_blocks--;
755 		inode->i_bytes += 512;
756 	}
757 	inode->i_bytes -= bytes;
758 }
759 
760 EXPORT_SYMBOL(__inode_sub_bytes);
761 
762 void inode_sub_bytes(struct inode *inode, loff_t bytes)
763 {
764 	spin_lock(&inode->i_lock);
765 	__inode_sub_bytes(inode, bytes);
766 	spin_unlock(&inode->i_lock);
767 }
768 
769 EXPORT_SYMBOL(inode_sub_bytes);
770 
771 loff_t inode_get_bytes(struct inode *inode)
772 {
773 	loff_t ret;
774 
775 	spin_lock(&inode->i_lock);
776 	ret = __inode_get_bytes(inode);
777 	spin_unlock(&inode->i_lock);
778 	return ret;
779 }
780 
781 EXPORT_SYMBOL(inode_get_bytes);
782 
783 void inode_set_bytes(struct inode *inode, loff_t bytes)
784 {
785 	/* Caller is here responsible for sufficient locking
786 	 * (ie. inode->i_lock) */
787 	inode->i_blocks = bytes >> 9;
788 	inode->i_bytes = bytes & 511;
789 }
790 
791 EXPORT_SYMBOL(inode_set_bytes);
792