at v6.15-rc3 971 lines 27 kB view raw
1// SPDX-License-Identifier: GPL-2.0 2/* 3 * linux/fs/stat.c 4 * 5 * Copyright (C) 1991, 1992 Linus Torvalds 6 */ 7 8#include <linux/blkdev.h> 9#include <linux/export.h> 10#include <linux/mm.h> 11#include <linux/errno.h> 12#include <linux/file.h> 13#include <linux/highuid.h> 14#include <linux/fs.h> 15#include <linux/namei.h> 16#include <linux/security.h> 17#include <linux/cred.h> 18#include <linux/syscalls.h> 19#include <linux/pagemap.h> 20#include <linux/compat.h> 21#include <linux/iversion.h> 22 23#include <linux/uaccess.h> 24#include <asm/unistd.h> 25 26#include <trace/events/timestamp.h> 27 28#include "internal.h" 29#include "mount.h" 30 31/** 32 * fill_mg_cmtime - Fill in the mtime and ctime and flag ctime as QUERIED 33 * @stat: where to store the resulting values 34 * @request_mask: STATX_* values requested 35 * @inode: inode from which to grab the c/mtime 36 * 37 * Given @inode, grab the ctime and mtime out if it and store the result 38 * in @stat. When fetching the value, flag it as QUERIED (if not already) 39 * so the next write will record a distinct timestamp. 40 * 41 * NB: The QUERIED flag is tracked in the ctime, but we set it there even 42 * if only the mtime was requested, as that ensures that the next mtime 43 * change will be distinct. 44 */ 45void fill_mg_cmtime(struct kstat *stat, u32 request_mask, struct inode *inode) 46{ 47 atomic_t *pcn = (atomic_t *)&inode->i_ctime_nsec; 48 49 /* If neither time was requested, then don't report them */ 50 if (!(request_mask & (STATX_CTIME|STATX_MTIME))) { 51 stat->result_mask &= ~(STATX_CTIME|STATX_MTIME); 52 return; 53 } 54 55 stat->mtime = inode_get_mtime(inode); 56 stat->ctime.tv_sec = inode->i_ctime_sec; 57 stat->ctime.tv_nsec = (u32)atomic_read(pcn); 58 if (!(stat->ctime.tv_nsec & I_CTIME_QUERIED)) 59 stat->ctime.tv_nsec = ((u32)atomic_fetch_or(I_CTIME_QUERIED, pcn)); 60 stat->ctime.tv_nsec &= ~I_CTIME_QUERIED; 61 trace_fill_mg_cmtime(inode, &stat->ctime, &stat->mtime); 62} 63EXPORT_SYMBOL(fill_mg_cmtime); 64 65/** 66 * generic_fillattr - Fill in the basic attributes from the inode struct 67 * @idmap: idmap of the mount the inode was found from 68 * @request_mask: statx request_mask 69 * @inode: Inode to use as the source 70 * @stat: Where to fill in the attributes 71 * 72 * Fill in the basic attributes in the kstat structure from data that's to be 73 * found on the VFS inode structure. This is the default if no getattr inode 74 * operation is supplied. 75 * 76 * If the inode has been found through an idmapped mount the idmap of 77 * the vfsmount must be passed through @idmap. This function will then 78 * take care to map the inode according to @idmap before filling in the 79 * uid and gid filds. On non-idmapped mounts or if permission checking is to be 80 * performed on the raw inode simply pass @nop_mnt_idmap. 81 */ 82void generic_fillattr(struct mnt_idmap *idmap, u32 request_mask, 83 struct inode *inode, struct kstat *stat) 84{ 85 vfsuid_t vfsuid = i_uid_into_vfsuid(idmap, inode); 86 vfsgid_t vfsgid = i_gid_into_vfsgid(idmap, inode); 87 88 stat->dev = inode->i_sb->s_dev; 89 stat->ino = inode->i_ino; 90 stat->mode = inode->i_mode; 91 stat->nlink = inode->i_nlink; 92 stat->uid = vfsuid_into_kuid(vfsuid); 93 stat->gid = vfsgid_into_kgid(vfsgid); 94 stat->rdev = inode->i_rdev; 95 stat->size = i_size_read(inode); 96 stat->atime = inode_get_atime(inode); 97 98 if (is_mgtime(inode)) { 99 fill_mg_cmtime(stat, request_mask, inode); 100 } else { 101 stat->ctime = inode_get_ctime(inode); 102 stat->mtime = inode_get_mtime(inode); 103 } 104 105 stat->blksize = i_blocksize(inode); 106 stat->blocks = inode->i_blocks; 107 108 if ((request_mask & STATX_CHANGE_COOKIE) && IS_I_VERSION(inode)) { 109 stat->result_mask |= STATX_CHANGE_COOKIE; 110 stat->change_cookie = inode_query_iversion(inode); 111 } 112 113} 114EXPORT_SYMBOL(generic_fillattr); 115 116/** 117 * generic_fill_statx_attr - Fill in the statx attributes from the inode flags 118 * @inode: Inode to use as the source 119 * @stat: Where to fill in the attribute flags 120 * 121 * Fill in the STATX_ATTR_* flags in the kstat structure for properties of the 122 * inode that are published on i_flags and enforced by the VFS. 123 */ 124void generic_fill_statx_attr(struct inode *inode, struct kstat *stat) 125{ 126 if (inode->i_flags & S_IMMUTABLE) 127 stat->attributes |= STATX_ATTR_IMMUTABLE; 128 if (inode->i_flags & S_APPEND) 129 stat->attributes |= STATX_ATTR_APPEND; 130 stat->attributes_mask |= KSTAT_ATTR_VFS_FLAGS; 131} 132EXPORT_SYMBOL(generic_fill_statx_attr); 133 134/** 135 * generic_fill_statx_atomic_writes - Fill in atomic writes statx attributes 136 * @stat: Where to fill in the attribute flags 137 * @unit_min: Minimum supported atomic write length in bytes 138 * @unit_max: Maximum supported atomic write length in bytes 139 * 140 * Fill in the STATX{_ATTR}_WRITE_ATOMIC flags in the kstat structure from 141 * atomic write unit_min and unit_max values. 142 */ 143void generic_fill_statx_atomic_writes(struct kstat *stat, 144 unsigned int unit_min, 145 unsigned int unit_max) 146{ 147 /* Confirm that the request type is known */ 148 stat->result_mask |= STATX_WRITE_ATOMIC; 149 150 /* Confirm that the file attribute type is known */ 151 stat->attributes_mask |= STATX_ATTR_WRITE_ATOMIC; 152 153 if (unit_min) { 154 stat->atomic_write_unit_min = unit_min; 155 stat->atomic_write_unit_max = unit_max; 156 /* Initially only allow 1x segment */ 157 stat->atomic_write_segments_max = 1; 158 159 /* Confirm atomic writes are actually supported */ 160 stat->attributes |= STATX_ATTR_WRITE_ATOMIC; 161 } 162} 163EXPORT_SYMBOL_GPL(generic_fill_statx_atomic_writes); 164 165/** 166 * vfs_getattr_nosec - getattr without security checks 167 * @path: file to get attributes from 168 * @stat: structure to return attributes in 169 * @request_mask: STATX_xxx flags indicating what the caller wants 170 * @query_flags: Query mode (AT_STATX_SYNC_TYPE) 171 * 172 * Get attributes without calling security_inode_getattr. 173 * 174 * Currently the only caller other than vfs_getattr is internal to the 175 * filehandle lookup code, which uses only the inode number and returns no 176 * attributes to any user. Any other code probably wants vfs_getattr. 177 */ 178int vfs_getattr_nosec(const struct path *path, struct kstat *stat, 179 u32 request_mask, unsigned int query_flags) 180{ 181 struct mnt_idmap *idmap; 182 struct inode *inode = d_backing_inode(path->dentry); 183 184 memset(stat, 0, sizeof(*stat)); 185 stat->result_mask |= STATX_BASIC_STATS; 186 query_flags &= AT_STATX_SYNC_TYPE; 187 188 /* allow the fs to override these if it really wants to */ 189 /* SB_NOATIME means filesystem supplies dummy atime value */ 190 if (inode->i_sb->s_flags & SB_NOATIME) 191 stat->result_mask &= ~STATX_ATIME; 192 193 /* 194 * Note: If you add another clause to set an attribute flag, please 195 * update attributes_mask below. 196 */ 197 if (IS_AUTOMOUNT(inode)) 198 stat->attributes |= STATX_ATTR_AUTOMOUNT; 199 200 if (IS_DAX(inode)) 201 stat->attributes |= STATX_ATTR_DAX; 202 203 stat->attributes_mask |= (STATX_ATTR_AUTOMOUNT | 204 STATX_ATTR_DAX); 205 206 idmap = mnt_idmap(path->mnt); 207 if (inode->i_op->getattr) { 208 int ret; 209 210 ret = inode->i_op->getattr(idmap, path, stat, request_mask, 211 query_flags); 212 if (ret) 213 return ret; 214 } else { 215 generic_fillattr(idmap, request_mask, inode, stat); 216 } 217 218 /* 219 * If this is a block device inode, override the filesystem attributes 220 * with the block device specific parameters that need to be obtained 221 * from the bdev backing inode. 222 */ 223 if (S_ISBLK(stat->mode)) 224 bdev_statx(path, stat, request_mask); 225 226 return 0; 227} 228EXPORT_SYMBOL(vfs_getattr_nosec); 229 230/* 231 * vfs_getattr - Get the enhanced basic attributes of a file 232 * @path: The file of interest 233 * @stat: Where to return the statistics 234 * @request_mask: STATX_xxx flags indicating what the caller wants 235 * @query_flags: Query mode (AT_STATX_SYNC_TYPE) 236 * 237 * Ask the filesystem for a file's attributes. The caller must indicate in 238 * request_mask and query_flags to indicate what they want. 239 * 240 * If the file is remote, the filesystem can be forced to update the attributes 241 * from the backing store by passing AT_STATX_FORCE_SYNC in query_flags or can 242 * suppress the update by passing AT_STATX_DONT_SYNC. 243 * 244 * Bits must have been set in request_mask to indicate which attributes the 245 * caller wants retrieving. Any such attribute not requested may be returned 246 * anyway, but the value may be approximate, and, if remote, may not have been 247 * synchronised with the server. 248 * 249 * 0 will be returned on success, and a -ve error code if unsuccessful. 250 */ 251int vfs_getattr(const struct path *path, struct kstat *stat, 252 u32 request_mask, unsigned int query_flags) 253{ 254 int retval; 255 256 retval = security_inode_getattr(path); 257 if (retval) 258 return retval; 259 return vfs_getattr_nosec(path, stat, request_mask, query_flags); 260} 261EXPORT_SYMBOL(vfs_getattr); 262 263/** 264 * vfs_fstat - Get the basic attributes by file descriptor 265 * @fd: The file descriptor referring to the file of interest 266 * @stat: The result structure to fill in. 267 * 268 * This function is a wrapper around vfs_getattr(). The main difference is 269 * that it uses a file descriptor to determine the file location. 270 * 271 * 0 will be returned on success, and a -ve error code if unsuccessful. 272 */ 273int vfs_fstat(int fd, struct kstat *stat) 274{ 275 CLASS(fd_raw, f)(fd); 276 if (fd_empty(f)) 277 return -EBADF; 278 return vfs_getattr(&fd_file(f)->f_path, stat, STATX_BASIC_STATS, 0); 279} 280 281static int statx_lookup_flags(int flags) 282{ 283 int lookup_flags = 0; 284 285 if (!(flags & AT_SYMLINK_NOFOLLOW)) 286 lookup_flags |= LOOKUP_FOLLOW; 287 if (!(flags & AT_NO_AUTOMOUNT)) 288 lookup_flags |= LOOKUP_AUTOMOUNT; 289 290 return lookup_flags; 291} 292 293static int vfs_statx_path(struct path *path, int flags, struct kstat *stat, 294 u32 request_mask) 295{ 296 int error = vfs_getattr(path, stat, request_mask, flags); 297 if (error) 298 return error; 299 300 if (request_mask & STATX_MNT_ID_UNIQUE) { 301 stat->mnt_id = real_mount(path->mnt)->mnt_id_unique; 302 stat->result_mask |= STATX_MNT_ID_UNIQUE; 303 } else { 304 stat->mnt_id = real_mount(path->mnt)->mnt_id; 305 stat->result_mask |= STATX_MNT_ID; 306 } 307 308 if (path_mounted(path)) 309 stat->attributes |= STATX_ATTR_MOUNT_ROOT; 310 stat->attributes_mask |= STATX_ATTR_MOUNT_ROOT; 311 return 0; 312} 313 314static int vfs_statx_fd(int fd, int flags, struct kstat *stat, 315 u32 request_mask) 316{ 317 CLASS(fd_raw, f)(fd); 318 if (fd_empty(f)) 319 return -EBADF; 320 return vfs_statx_path(&fd_file(f)->f_path, flags, stat, request_mask); 321} 322 323/** 324 * vfs_statx - Get basic and extra attributes by filename 325 * @dfd: A file descriptor representing the base dir for a relative filename 326 * @filename: The name of the file of interest 327 * @flags: Flags to control the query 328 * @stat: The result structure to fill in. 329 * @request_mask: STATX_xxx flags indicating what the caller wants 330 * 331 * This function is a wrapper around vfs_getattr(). The main difference is 332 * that it uses a filename and base directory to determine the file location. 333 * Additionally, the use of AT_SYMLINK_NOFOLLOW in flags will prevent a symlink 334 * at the given name from being referenced. 335 * 336 * 0 will be returned on success, and a -ve error code if unsuccessful. 337 */ 338static int vfs_statx(int dfd, struct filename *filename, int flags, 339 struct kstat *stat, u32 request_mask) 340{ 341 struct path path; 342 unsigned int lookup_flags = statx_lookup_flags(flags); 343 int error; 344 345 if (flags & ~(AT_SYMLINK_NOFOLLOW | AT_NO_AUTOMOUNT | AT_EMPTY_PATH | 346 AT_STATX_SYNC_TYPE)) 347 return -EINVAL; 348 349retry: 350 error = filename_lookup(dfd, filename, lookup_flags, &path, NULL); 351 if (error) 352 return error; 353 error = vfs_statx_path(&path, flags, stat, request_mask); 354 path_put(&path); 355 if (retry_estale(error, lookup_flags)) { 356 lookup_flags |= LOOKUP_REVAL; 357 goto retry; 358 } 359 return error; 360} 361 362int vfs_fstatat(int dfd, const char __user *filename, 363 struct kstat *stat, int flags) 364{ 365 int ret; 366 int statx_flags = flags | AT_NO_AUTOMOUNT; 367 struct filename *name = getname_maybe_null(filename, flags); 368 369 if (!name && dfd >= 0) 370 return vfs_fstat(dfd, stat); 371 372 ret = vfs_statx(dfd, name, statx_flags, stat, STATX_BASIC_STATS); 373 putname(name); 374 375 return ret; 376} 377 378#ifdef __ARCH_WANT_OLD_STAT 379 380/* 381 * For backward compatibility? Maybe this should be moved 382 * into arch/i386 instead? 383 */ 384static int cp_old_stat(struct kstat *stat, struct __old_kernel_stat __user * statbuf) 385{ 386 static int warncount = 5; 387 struct __old_kernel_stat tmp; 388 389 if (warncount > 0) { 390 warncount--; 391 printk(KERN_WARNING "VFS: Warning: %s using old stat() call. Recompile your binary.\n", 392 current->comm); 393 } else if (warncount < 0) { 394 /* it's laughable, but... */ 395 warncount = 0; 396 } 397 398 memset(&tmp, 0, sizeof(struct __old_kernel_stat)); 399 tmp.st_dev = old_encode_dev(stat->dev); 400 tmp.st_ino = stat->ino; 401 if (sizeof(tmp.st_ino) < sizeof(stat->ino) && tmp.st_ino != stat->ino) 402 return -EOVERFLOW; 403 tmp.st_mode = stat->mode; 404 tmp.st_nlink = stat->nlink; 405 if (tmp.st_nlink != stat->nlink) 406 return -EOVERFLOW; 407 SET_UID(tmp.st_uid, from_kuid_munged(current_user_ns(), stat->uid)); 408 SET_GID(tmp.st_gid, from_kgid_munged(current_user_ns(), stat->gid)); 409 tmp.st_rdev = old_encode_dev(stat->rdev); 410#if BITS_PER_LONG == 32 411 if (stat->size > MAX_NON_LFS) 412 return -EOVERFLOW; 413#endif 414 tmp.st_size = stat->size; 415 tmp.st_atime = stat->atime.tv_sec; 416 tmp.st_mtime = stat->mtime.tv_sec; 417 tmp.st_ctime = stat->ctime.tv_sec; 418 return copy_to_user(statbuf,&tmp,sizeof(tmp)) ? -EFAULT : 0; 419} 420 421SYSCALL_DEFINE2(stat, const char __user *, filename, 422 struct __old_kernel_stat __user *, statbuf) 423{ 424 struct kstat stat; 425 int error; 426 427 error = vfs_stat(filename, &stat); 428 if (error) 429 return error; 430 431 return cp_old_stat(&stat, statbuf); 432} 433 434SYSCALL_DEFINE2(lstat, const char __user *, filename, 435 struct __old_kernel_stat __user *, statbuf) 436{ 437 struct kstat stat; 438 int error; 439 440 error = vfs_lstat(filename, &stat); 441 if (error) 442 return error; 443 444 return cp_old_stat(&stat, statbuf); 445} 446 447SYSCALL_DEFINE2(fstat, unsigned int, fd, struct __old_kernel_stat __user *, statbuf) 448{ 449 struct kstat stat; 450 int error = vfs_fstat(fd, &stat); 451 452 if (!error) 453 error = cp_old_stat(&stat, statbuf); 454 455 return error; 456} 457 458#endif /* __ARCH_WANT_OLD_STAT */ 459 460#ifdef __ARCH_WANT_NEW_STAT 461 462#ifndef INIT_STRUCT_STAT_PADDING 463# define INIT_STRUCT_STAT_PADDING(st) memset(&st, 0, sizeof(st)) 464#endif 465 466static int cp_new_stat(struct kstat *stat, struct stat __user *statbuf) 467{ 468 struct stat tmp; 469 470 if (sizeof(tmp.st_dev) < 4 && !old_valid_dev(stat->dev)) 471 return -EOVERFLOW; 472 if (sizeof(tmp.st_rdev) < 4 && !old_valid_dev(stat->rdev)) 473 return -EOVERFLOW; 474#if BITS_PER_LONG == 32 475 if (stat->size > MAX_NON_LFS) 476 return -EOVERFLOW; 477#endif 478 479 INIT_STRUCT_STAT_PADDING(tmp); 480 tmp.st_dev = new_encode_dev(stat->dev); 481 tmp.st_ino = stat->ino; 482 if (sizeof(tmp.st_ino) < sizeof(stat->ino) && tmp.st_ino != stat->ino) 483 return -EOVERFLOW; 484 tmp.st_mode = stat->mode; 485 tmp.st_nlink = stat->nlink; 486 if (tmp.st_nlink != stat->nlink) 487 return -EOVERFLOW; 488 SET_UID(tmp.st_uid, from_kuid_munged(current_user_ns(), stat->uid)); 489 SET_GID(tmp.st_gid, from_kgid_munged(current_user_ns(), stat->gid)); 490 tmp.st_rdev = new_encode_dev(stat->rdev); 491 tmp.st_size = stat->size; 492 tmp.st_atime = stat->atime.tv_sec; 493 tmp.st_mtime = stat->mtime.tv_sec; 494 tmp.st_ctime = stat->ctime.tv_sec; 495#ifdef STAT_HAVE_NSEC 496 tmp.st_atime_nsec = stat->atime.tv_nsec; 497 tmp.st_mtime_nsec = stat->mtime.tv_nsec; 498 tmp.st_ctime_nsec = stat->ctime.tv_nsec; 499#endif 500 tmp.st_blocks = stat->blocks; 501 tmp.st_blksize = stat->blksize; 502 return copy_to_user(statbuf,&tmp,sizeof(tmp)) ? -EFAULT : 0; 503} 504 505SYSCALL_DEFINE2(newstat, const char __user *, filename, 506 struct stat __user *, statbuf) 507{ 508 struct kstat stat; 509 int error = vfs_stat(filename, &stat); 510 511 if (error) 512 return error; 513 return cp_new_stat(&stat, statbuf); 514} 515 516SYSCALL_DEFINE2(newlstat, const char __user *, filename, 517 struct stat __user *, statbuf) 518{ 519 struct kstat stat; 520 int error; 521 522 error = vfs_lstat(filename, &stat); 523 if (error) 524 return error; 525 526 return cp_new_stat(&stat, statbuf); 527} 528 529#if !defined(__ARCH_WANT_STAT64) || defined(__ARCH_WANT_SYS_NEWFSTATAT) 530SYSCALL_DEFINE4(newfstatat, int, dfd, const char __user *, filename, 531 struct stat __user *, statbuf, int, flag) 532{ 533 struct kstat stat; 534 int error; 535 536 error = vfs_fstatat(dfd, filename, &stat, flag); 537 if (error) 538 return error; 539 return cp_new_stat(&stat, statbuf); 540} 541#endif 542 543SYSCALL_DEFINE2(newfstat, unsigned int, fd, struct stat __user *, statbuf) 544{ 545 struct kstat stat; 546 int error = vfs_fstat(fd, &stat); 547 548 if (!error) 549 error = cp_new_stat(&stat, statbuf); 550 551 return error; 552} 553#endif 554 555static int do_readlinkat(int dfd, const char __user *pathname, 556 char __user *buf, int bufsiz) 557{ 558 struct path path; 559 struct filename *name; 560 int error; 561 unsigned int lookup_flags = LOOKUP_EMPTY; 562 563 if (bufsiz <= 0) 564 return -EINVAL; 565 566retry: 567 name = getname_flags(pathname, lookup_flags); 568 error = filename_lookup(dfd, name, lookup_flags, &path, NULL); 569 if (unlikely(error)) { 570 putname(name); 571 return error; 572 } 573 574 /* 575 * AFS mountpoints allow readlink(2) but are not symlinks 576 */ 577 if (d_is_symlink(path.dentry) || 578 d_backing_inode(path.dentry)->i_op->readlink) { 579 error = security_inode_readlink(path.dentry); 580 if (!error) { 581 touch_atime(&path); 582 error = vfs_readlink(path.dentry, buf, bufsiz); 583 } 584 } else { 585 error = (name->name[0] == '\0') ? -ENOENT : -EINVAL; 586 } 587 path_put(&path); 588 putname(name); 589 if (retry_estale(error, lookup_flags)) { 590 lookup_flags |= LOOKUP_REVAL; 591 goto retry; 592 } 593 return error; 594} 595 596SYSCALL_DEFINE4(readlinkat, int, dfd, const char __user *, pathname, 597 char __user *, buf, int, bufsiz) 598{ 599 return do_readlinkat(dfd, pathname, buf, bufsiz); 600} 601 602SYSCALL_DEFINE3(readlink, const char __user *, path, char __user *, buf, 603 int, bufsiz) 604{ 605 return do_readlinkat(AT_FDCWD, path, buf, bufsiz); 606} 607 608 609/* ---------- LFS-64 ----------- */ 610#if defined(__ARCH_WANT_STAT64) || defined(__ARCH_WANT_COMPAT_STAT64) 611 612#ifndef INIT_STRUCT_STAT64_PADDING 613# define INIT_STRUCT_STAT64_PADDING(st) memset(&st, 0, sizeof(st)) 614#endif 615 616static long cp_new_stat64(struct kstat *stat, struct stat64 __user *statbuf) 617{ 618 struct stat64 tmp; 619 620 INIT_STRUCT_STAT64_PADDING(tmp); 621#ifdef CONFIG_MIPS 622 /* mips has weird padding, so we don't get 64 bits there */ 623 tmp.st_dev = new_encode_dev(stat->dev); 624 tmp.st_rdev = new_encode_dev(stat->rdev); 625#else 626 tmp.st_dev = huge_encode_dev(stat->dev); 627 tmp.st_rdev = huge_encode_dev(stat->rdev); 628#endif 629 tmp.st_ino = stat->ino; 630 if (sizeof(tmp.st_ino) < sizeof(stat->ino) && tmp.st_ino != stat->ino) 631 return -EOVERFLOW; 632#ifdef STAT64_HAS_BROKEN_ST_INO 633 tmp.__st_ino = stat->ino; 634#endif 635 tmp.st_mode = stat->mode; 636 tmp.st_nlink = stat->nlink; 637 tmp.st_uid = from_kuid_munged(current_user_ns(), stat->uid); 638 tmp.st_gid = from_kgid_munged(current_user_ns(), stat->gid); 639 tmp.st_atime = stat->atime.tv_sec; 640 tmp.st_atime_nsec = stat->atime.tv_nsec; 641 tmp.st_mtime = stat->mtime.tv_sec; 642 tmp.st_mtime_nsec = stat->mtime.tv_nsec; 643 tmp.st_ctime = stat->ctime.tv_sec; 644 tmp.st_ctime_nsec = stat->ctime.tv_nsec; 645 tmp.st_size = stat->size; 646 tmp.st_blocks = stat->blocks; 647 tmp.st_blksize = stat->blksize; 648 return copy_to_user(statbuf,&tmp,sizeof(tmp)) ? -EFAULT : 0; 649} 650 651SYSCALL_DEFINE2(stat64, const char __user *, filename, 652 struct stat64 __user *, statbuf) 653{ 654 struct kstat stat; 655 int error = vfs_stat(filename, &stat); 656 657 if (!error) 658 error = cp_new_stat64(&stat, statbuf); 659 660 return error; 661} 662 663SYSCALL_DEFINE2(lstat64, const char __user *, filename, 664 struct stat64 __user *, statbuf) 665{ 666 struct kstat stat; 667 int error = vfs_lstat(filename, &stat); 668 669 if (!error) 670 error = cp_new_stat64(&stat, statbuf); 671 672 return error; 673} 674 675SYSCALL_DEFINE2(fstat64, unsigned long, fd, struct stat64 __user *, statbuf) 676{ 677 struct kstat stat; 678 int error = vfs_fstat(fd, &stat); 679 680 if (!error) 681 error = cp_new_stat64(&stat, statbuf); 682 683 return error; 684} 685 686SYSCALL_DEFINE4(fstatat64, int, dfd, const char __user *, filename, 687 struct stat64 __user *, statbuf, int, flag) 688{ 689 struct kstat stat; 690 int error; 691 692 error = vfs_fstatat(dfd, filename, &stat, flag); 693 if (error) 694 return error; 695 return cp_new_stat64(&stat, statbuf); 696} 697#endif /* __ARCH_WANT_STAT64 || __ARCH_WANT_COMPAT_STAT64 */ 698 699static noinline_for_stack int 700cp_statx(const struct kstat *stat, struct statx __user *buffer) 701{ 702 struct statx tmp; 703 704 memset(&tmp, 0, sizeof(tmp)); 705 706 /* STATX_CHANGE_COOKIE is kernel-only for now */ 707 tmp.stx_mask = stat->result_mask & ~STATX_CHANGE_COOKIE; 708 tmp.stx_blksize = stat->blksize; 709 /* STATX_ATTR_CHANGE_MONOTONIC is kernel-only for now */ 710 tmp.stx_attributes = stat->attributes & ~STATX_ATTR_CHANGE_MONOTONIC; 711 tmp.stx_nlink = stat->nlink; 712 tmp.stx_uid = from_kuid_munged(current_user_ns(), stat->uid); 713 tmp.stx_gid = from_kgid_munged(current_user_ns(), stat->gid); 714 tmp.stx_mode = stat->mode; 715 tmp.stx_ino = stat->ino; 716 tmp.stx_size = stat->size; 717 tmp.stx_blocks = stat->blocks; 718 tmp.stx_attributes_mask = stat->attributes_mask; 719 tmp.stx_atime.tv_sec = stat->atime.tv_sec; 720 tmp.stx_atime.tv_nsec = stat->atime.tv_nsec; 721 tmp.stx_btime.tv_sec = stat->btime.tv_sec; 722 tmp.stx_btime.tv_nsec = stat->btime.tv_nsec; 723 tmp.stx_ctime.tv_sec = stat->ctime.tv_sec; 724 tmp.stx_ctime.tv_nsec = stat->ctime.tv_nsec; 725 tmp.stx_mtime.tv_sec = stat->mtime.tv_sec; 726 tmp.stx_mtime.tv_nsec = stat->mtime.tv_nsec; 727 tmp.stx_rdev_major = MAJOR(stat->rdev); 728 tmp.stx_rdev_minor = MINOR(stat->rdev); 729 tmp.stx_dev_major = MAJOR(stat->dev); 730 tmp.stx_dev_minor = MINOR(stat->dev); 731 tmp.stx_mnt_id = stat->mnt_id; 732 tmp.stx_dio_mem_align = stat->dio_mem_align; 733 tmp.stx_dio_offset_align = stat->dio_offset_align; 734 tmp.stx_dio_read_offset_align = stat->dio_read_offset_align; 735 tmp.stx_subvol = stat->subvol; 736 tmp.stx_atomic_write_unit_min = stat->atomic_write_unit_min; 737 tmp.stx_atomic_write_unit_max = stat->atomic_write_unit_max; 738 tmp.stx_atomic_write_segments_max = stat->atomic_write_segments_max; 739 740 return copy_to_user(buffer, &tmp, sizeof(tmp)) ? -EFAULT : 0; 741} 742 743int do_statx(int dfd, struct filename *filename, unsigned int flags, 744 unsigned int mask, struct statx __user *buffer) 745{ 746 struct kstat stat; 747 int error; 748 749 if (mask & STATX__RESERVED) 750 return -EINVAL; 751 if ((flags & AT_STATX_SYNC_TYPE) == AT_STATX_SYNC_TYPE) 752 return -EINVAL; 753 754 /* 755 * STATX_CHANGE_COOKIE is kernel-only for now. Ignore requests 756 * from userland. 757 */ 758 mask &= ~STATX_CHANGE_COOKIE; 759 760 error = vfs_statx(dfd, filename, flags, &stat, mask); 761 if (error) 762 return error; 763 764 return cp_statx(&stat, buffer); 765} 766 767int do_statx_fd(int fd, unsigned int flags, unsigned int mask, 768 struct statx __user *buffer) 769{ 770 struct kstat stat; 771 int error; 772 773 if (mask & STATX__RESERVED) 774 return -EINVAL; 775 if ((flags & AT_STATX_SYNC_TYPE) == AT_STATX_SYNC_TYPE) 776 return -EINVAL; 777 778 /* 779 * STATX_CHANGE_COOKIE is kernel-only for now. Ignore requests 780 * from userland. 781 */ 782 mask &= ~STATX_CHANGE_COOKIE; 783 784 error = vfs_statx_fd(fd, flags, &stat, mask); 785 if (error) 786 return error; 787 788 return cp_statx(&stat, buffer); 789} 790 791/** 792 * sys_statx - System call to get enhanced stats 793 * @dfd: Base directory to pathwalk from *or* fd to stat. 794 * @filename: File to stat or either NULL or "" with AT_EMPTY_PATH 795 * @flags: AT_* flags to control pathwalk. 796 * @mask: Parts of statx struct actually required. 797 * @buffer: Result buffer. 798 * 799 * Note that fstat() can be emulated by setting dfd to the fd of interest, 800 * supplying "" (or preferably NULL) as the filename and setting AT_EMPTY_PATH 801 * in the flags. 802 */ 803SYSCALL_DEFINE5(statx, 804 int, dfd, const char __user *, filename, unsigned, flags, 805 unsigned int, mask, 806 struct statx __user *, buffer) 807{ 808 int ret; 809 struct filename *name = getname_maybe_null(filename, flags); 810 811 if (!name && dfd >= 0) 812 return do_statx_fd(dfd, flags & ~AT_NO_AUTOMOUNT, mask, buffer); 813 814 ret = do_statx(dfd, name, flags, mask, buffer); 815 putname(name); 816 817 return ret; 818} 819 820#if defined(CONFIG_COMPAT) && defined(__ARCH_WANT_COMPAT_STAT) 821static int cp_compat_stat(struct kstat *stat, struct compat_stat __user *ubuf) 822{ 823 struct compat_stat tmp; 824 825 if (sizeof(tmp.st_dev) < 4 && !old_valid_dev(stat->dev)) 826 return -EOVERFLOW; 827 if (sizeof(tmp.st_rdev) < 4 && !old_valid_dev(stat->rdev)) 828 return -EOVERFLOW; 829 830 memset(&tmp, 0, sizeof(tmp)); 831 tmp.st_dev = new_encode_dev(stat->dev); 832 tmp.st_ino = stat->ino; 833 if (sizeof(tmp.st_ino) < sizeof(stat->ino) && tmp.st_ino != stat->ino) 834 return -EOVERFLOW; 835 tmp.st_mode = stat->mode; 836 tmp.st_nlink = stat->nlink; 837 if (tmp.st_nlink != stat->nlink) 838 return -EOVERFLOW; 839 SET_UID(tmp.st_uid, from_kuid_munged(current_user_ns(), stat->uid)); 840 SET_GID(tmp.st_gid, from_kgid_munged(current_user_ns(), stat->gid)); 841 tmp.st_rdev = new_encode_dev(stat->rdev); 842 if ((u64) stat->size > MAX_NON_LFS) 843 return -EOVERFLOW; 844 tmp.st_size = stat->size; 845 tmp.st_atime = stat->atime.tv_sec; 846 tmp.st_atime_nsec = stat->atime.tv_nsec; 847 tmp.st_mtime = stat->mtime.tv_sec; 848 tmp.st_mtime_nsec = stat->mtime.tv_nsec; 849 tmp.st_ctime = stat->ctime.tv_sec; 850 tmp.st_ctime_nsec = stat->ctime.tv_nsec; 851 tmp.st_blocks = stat->blocks; 852 tmp.st_blksize = stat->blksize; 853 return copy_to_user(ubuf, &tmp, sizeof(tmp)) ? -EFAULT : 0; 854} 855 856COMPAT_SYSCALL_DEFINE2(newstat, const char __user *, filename, 857 struct compat_stat __user *, statbuf) 858{ 859 struct kstat stat; 860 int error; 861 862 error = vfs_stat(filename, &stat); 863 if (error) 864 return error; 865 return cp_compat_stat(&stat, statbuf); 866} 867 868COMPAT_SYSCALL_DEFINE2(newlstat, const char __user *, filename, 869 struct compat_stat __user *, statbuf) 870{ 871 struct kstat stat; 872 int error; 873 874 error = vfs_lstat(filename, &stat); 875 if (error) 876 return error; 877 return cp_compat_stat(&stat, statbuf); 878} 879 880#ifndef __ARCH_WANT_STAT64 881COMPAT_SYSCALL_DEFINE4(newfstatat, unsigned int, dfd, 882 const char __user *, filename, 883 struct compat_stat __user *, statbuf, int, flag) 884{ 885 struct kstat stat; 886 int error; 887 888 error = vfs_fstatat(dfd, filename, &stat, flag); 889 if (error) 890 return error; 891 return cp_compat_stat(&stat, statbuf); 892} 893#endif 894 895COMPAT_SYSCALL_DEFINE2(newfstat, unsigned int, fd, 896 struct compat_stat __user *, statbuf) 897{ 898 struct kstat stat; 899 int error = vfs_fstat(fd, &stat); 900 901 if (!error) 902 error = cp_compat_stat(&stat, statbuf); 903 return error; 904} 905#endif 906 907/* Caller is here responsible for sufficient locking (ie. inode->i_lock) */ 908void __inode_add_bytes(struct inode *inode, loff_t bytes) 909{ 910 inode->i_blocks += bytes >> 9; 911 bytes &= 511; 912 inode->i_bytes += bytes; 913 if (inode->i_bytes >= 512) { 914 inode->i_blocks++; 915 inode->i_bytes -= 512; 916 } 917} 918EXPORT_SYMBOL(__inode_add_bytes); 919 920void inode_add_bytes(struct inode *inode, loff_t bytes) 921{ 922 spin_lock(&inode->i_lock); 923 __inode_add_bytes(inode, bytes); 924 spin_unlock(&inode->i_lock); 925} 926 927EXPORT_SYMBOL(inode_add_bytes); 928 929void __inode_sub_bytes(struct inode *inode, loff_t bytes) 930{ 931 inode->i_blocks -= bytes >> 9; 932 bytes &= 511; 933 if (inode->i_bytes < bytes) { 934 inode->i_blocks--; 935 inode->i_bytes += 512; 936 } 937 inode->i_bytes -= bytes; 938} 939 940EXPORT_SYMBOL(__inode_sub_bytes); 941 942void inode_sub_bytes(struct inode *inode, loff_t bytes) 943{ 944 spin_lock(&inode->i_lock); 945 __inode_sub_bytes(inode, bytes); 946 spin_unlock(&inode->i_lock); 947} 948 949EXPORT_SYMBOL(inode_sub_bytes); 950 951loff_t inode_get_bytes(struct inode *inode) 952{ 953 loff_t ret; 954 955 spin_lock(&inode->i_lock); 956 ret = __inode_get_bytes(inode); 957 spin_unlock(&inode->i_lock); 958 return ret; 959} 960 961EXPORT_SYMBOL(inode_get_bytes); 962 963void inode_set_bytes(struct inode *inode, loff_t bytes) 964{ 965 /* Caller is here responsible for sufficient locking 966 * (ie. inode->i_lock) */ 967 inode->i_blocks = bytes >> 9; 968 inode->i_bytes = bytes & 511; 969} 970 971EXPORT_SYMBOL(inode_set_bytes);