1 // SPDX-License-Identifier: LGPL-2.1 2 /* 3 * 4 * vfs operations that deal with files 5 * 6 * Copyright (C) International Business Machines Corp., 2002,2010 7 * Author(s): Steve French (sfrench@us.ibm.com) 8 * Jeremy Allison (jra@samba.org) 9 * 10 */ 11 #include <linux/fs.h> 12 #include <linux/filelock.h> 13 #include <linux/backing-dev.h> 14 #include <linux/stat.h> 15 #include <linux/fcntl.h> 16 #include <linux/pagemap.h> 17 #include <linux/pagevec.h> 18 #include <linux/writeback.h> 19 #include <linux/task_io_accounting_ops.h> 20 #include <linux/delay.h> 21 #include <linux/mount.h> 22 #include <linux/slab.h> 23 #include <linux/swap.h> 24 #include <linux/mm.h> 25 #include <asm/div64.h> 26 #include "cifsfs.h" 27 #include "cifspdu.h" 28 #include "cifsglob.h" 29 #include "cifsproto.h" 30 #include "smb2proto.h" 31 #include "cifs_unicode.h" 32 #include "cifs_debug.h" 33 #include "cifs_fs_sb.h" 34 #include "fscache.h" 35 #include "smbdirect.h" 36 #include "fs_context.h" 37 #include "cifs_ioctl.h" 38 #include "cached_dir.h" 39 #include <trace/events/netfs.h> 40 41 static int cifs_reopen_file(struct cifsFileInfo *cfile, bool can_flush); 42 43 /* 44 * Prepare a subrequest to upload to the server. We need to allocate credits 45 * so that we know the maximum amount of data that we can include in it. 46 */ 47 static void cifs_prepare_write(struct netfs_io_subrequest *subreq) 48 { 49 struct cifs_io_subrequest *wdata = 50 container_of(subreq, struct cifs_io_subrequest, subreq); 51 struct cifs_io_request *req = wdata->req; 52 struct TCP_Server_Info *server; 53 struct cifsFileInfo *open_file = req->cfile; 54 size_t wsize = req->rreq.wsize; 55 int rc; 56 57 if (!wdata->have_xid) { 58 wdata->xid = get_xid(); 59 wdata->have_xid = true; 60 } 61 62 server = cifs_pick_channel(tlink_tcon(open_file->tlink)->ses); 63 wdata->server = server; 64 65 retry: 66 if (open_file->invalidHandle) { 67 rc = cifs_reopen_file(open_file, false); 68 if (rc < 0) { 69 if (rc == -EAGAIN) 70 goto retry; 71 subreq->error = rc; 72 return netfs_prepare_write_failed(subreq); 73 } 74 } 75 76 rc = server->ops->wait_mtu_credits(server, wsize, &wdata->subreq.max_len, 77 &wdata->credits); 78 if (rc < 0) { 79 subreq->error = rc; 80 return netfs_prepare_write_failed(subreq); 81 } 82 83 wdata->credits.rreq_debug_id = subreq->rreq->debug_id; 84 wdata->credits.rreq_debug_index = subreq->debug_index; 85 wdata->credits.in_flight_check = 1; 86 trace_smb3_rw_credits(wdata->rreq->debug_id, 87 wdata->subreq.debug_index, 88 wdata->credits.value, 89 server->credits, server->in_flight, 90 wdata->credits.value, 91 cifs_trace_rw_credits_write_prepare); 92 93 #ifdef CONFIG_CIFS_SMB_DIRECT 94 if (server->smbd_conn) 95 subreq->max_nr_segs = server->smbd_conn->max_frmr_depth; 96 #endif 97 } 98 99 /* 100 * Issue a subrequest to upload to the server. 101 */ 102 static void cifs_issue_write(struct netfs_io_subrequest *subreq) 103 { 104 struct cifs_io_subrequest *wdata = 105 container_of(subreq, struct cifs_io_subrequest, subreq); 106 struct cifs_sb_info *sbi = CIFS_SB(subreq->rreq->inode->i_sb); 107 int rc; 108 109 if (cifs_forced_shutdown(sbi)) { 110 rc = -EIO; 111 goto fail; 112 } 113 114 rc = adjust_credits(wdata->server, wdata, cifs_trace_rw_credits_issue_write_adjust); 115 if (rc) 116 goto fail; 117 118 rc = -EAGAIN; 119 if (wdata->req->cfile->invalidHandle) 120 goto fail; 121 122 wdata->server->ops->async_writev(wdata); 123 out: 124 return; 125 126 fail: 127 if (rc == -EAGAIN) 128 trace_netfs_sreq(subreq, netfs_sreq_trace_retry); 129 else 130 trace_netfs_sreq(subreq, netfs_sreq_trace_fail); 131 add_credits_and_wake_if(wdata->server, &wdata->credits, 0); 132 cifs_write_subrequest_terminated(wdata, rc, false); 133 goto out; 134 } 135 136 static void cifs_netfs_invalidate_cache(struct netfs_io_request *wreq) 137 { 138 cifs_invalidate_cache(wreq->inode, 0); 139 } 140 141 /* 142 * Split the read up according to how many credits we can get for each piece. 143 * It's okay to sleep here if we need to wait for more credit to become 144 * available. 145 * 146 * We also choose the server and allocate an operation ID to be cleaned up 147 * later. 148 */ 149 static bool cifs_clamp_length(struct netfs_io_subrequest *subreq) 150 { 151 struct netfs_io_request *rreq = subreq->rreq; 152 struct cifs_io_subrequest *rdata = container_of(subreq, struct cifs_io_subrequest, subreq); 153 struct cifs_io_request *req = container_of(subreq->rreq, struct cifs_io_request, rreq); 154 struct TCP_Server_Info *server = req->server; 155 struct cifs_sb_info *cifs_sb = CIFS_SB(rreq->inode->i_sb); 156 size_t rsize = 0; 157 int rc; 158 159 rdata->xid = get_xid(); 160 rdata->have_xid = true; 161 rdata->server = server; 162 163 if (cifs_sb->ctx->rsize == 0) 164 cifs_sb->ctx->rsize = 165 server->ops->negotiate_rsize(tlink_tcon(req->cfile->tlink), 166 cifs_sb->ctx); 167 168 169 rc = server->ops->wait_mtu_credits(server, cifs_sb->ctx->rsize, &rsize, 170 &rdata->credits); 171 if (rc) { 172 subreq->error = rc; 173 return false; 174 } 175 176 rdata->credits.in_flight_check = 1; 177 rdata->credits.rreq_debug_id = rreq->debug_id; 178 rdata->credits.rreq_debug_index = subreq->debug_index; 179 180 trace_smb3_rw_credits(rdata->rreq->debug_id, 181 rdata->subreq.debug_index, 182 rdata->credits.value, 183 server->credits, server->in_flight, 0, 184 cifs_trace_rw_credits_read_submit); 185 186 subreq->len = min_t(size_t, subreq->len, rsize); 187 188 #ifdef CONFIG_CIFS_SMB_DIRECT 189 if (server->smbd_conn) 190 subreq->max_nr_segs = server->smbd_conn->max_frmr_depth; 191 #endif 192 return true; 193 } 194 195 /* 196 * Issue a read operation on behalf of the netfs helper functions. We're asked 197 * to make a read of a certain size at a point in the file. We are permitted 198 * to only read a portion of that, but as long as we read something, the netfs 199 * helper will call us again so that we can issue another read. 200 */ 201 static void cifs_req_issue_read(struct netfs_io_subrequest *subreq) 202 { 203 struct netfs_io_request *rreq = subreq->rreq; 204 struct cifs_io_subrequest *rdata = container_of(subreq, struct cifs_io_subrequest, subreq); 205 struct cifs_io_request *req = container_of(subreq->rreq, struct cifs_io_request, rreq); 206 int rc = 0; 207 208 cifs_dbg(FYI, "%s: op=%08x[%x] mapping=%p len=%zu/%zu\n", 209 __func__, rreq->debug_id, subreq->debug_index, rreq->mapping, 210 subreq->transferred, subreq->len); 211 212 if (req->cfile->invalidHandle) { 213 do { 214 rc = cifs_reopen_file(req->cfile, true); 215 } while (rc == -EAGAIN); 216 if (rc) 217 goto out; 218 } 219 220 if (subreq->rreq->origin != NETFS_DIO_READ) 221 __set_bit(NETFS_SREQ_CLEAR_TAIL, &subreq->flags); 222 223 rc = rdata->server->ops->async_readv(rdata); 224 out: 225 if (rc) 226 netfs_subreq_terminated(subreq, rc, false); 227 } 228 229 /* 230 * Writeback calls this when it finds a folio that needs uploading. This isn't 231 * called if writeback only has copy-to-cache to deal with. 232 */ 233 static void cifs_begin_writeback(struct netfs_io_request *wreq) 234 { 235 struct cifs_io_request *req = container_of(wreq, struct cifs_io_request, rreq); 236 int ret; 237 238 ret = cifs_get_writable_file(CIFS_I(wreq->inode), FIND_WR_ANY, &req->cfile); 239 if (ret) { 240 cifs_dbg(VFS, "No writable handle in writepages ret=%d\n", ret); 241 return; 242 } 243 244 wreq->io_streams[0].avail = true; 245 } 246 247 /* 248 * Initialise a request. 249 */ 250 static int cifs_init_request(struct netfs_io_request *rreq, struct file *file) 251 { 252 struct cifs_io_request *req = container_of(rreq, struct cifs_io_request, rreq); 253 struct cifs_sb_info *cifs_sb = CIFS_SB(rreq->inode->i_sb); 254 struct cifsFileInfo *open_file = NULL; 255 256 rreq->rsize = cifs_sb->ctx->rsize; 257 rreq->wsize = cifs_sb->ctx->wsize; 258 req->pid = current->tgid; // Ummm... This may be a workqueue 259 260 if (file) { 261 open_file = file->private_data; 262 rreq->netfs_priv = file->private_data; 263 req->cfile = cifsFileInfo_get(open_file); 264 req->server = cifs_pick_channel(tlink_tcon(req->cfile->tlink)->ses); 265 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_RWPIDFORWARD) 266 req->pid = req->cfile->pid; 267 } else if (rreq->origin != NETFS_WRITEBACK) { 268 WARN_ON_ONCE(1); 269 return -EIO; 270 } 271 272 return 0; 273 } 274 275 /* 276 * Completion of a request operation. 277 */ 278 static void cifs_rreq_done(struct netfs_io_request *rreq) 279 { 280 struct timespec64 atime, mtime; 281 struct inode *inode = rreq->inode; 282 283 /* we do not want atime to be less than mtime, it broke some apps */ 284 atime = inode_set_atime_to_ts(inode, current_time(inode)); 285 mtime = inode_get_mtime(inode); 286 if (timespec64_compare(&atime, &mtime)) 287 inode_set_atime_to_ts(inode, inode_get_mtime(inode)); 288 } 289 290 static void cifs_post_modify(struct inode *inode) 291 { 292 /* Indication to update ctime and mtime as close is deferred */ 293 set_bit(CIFS_INO_MODIFIED_ATTR, &CIFS_I(inode)->flags); 294 } 295 296 static void cifs_free_request(struct netfs_io_request *rreq) 297 { 298 struct cifs_io_request *req = container_of(rreq, struct cifs_io_request, rreq); 299 300 if (req->cfile) 301 cifsFileInfo_put(req->cfile); 302 } 303 304 static void cifs_free_subrequest(struct netfs_io_subrequest *subreq) 305 { 306 struct cifs_io_subrequest *rdata = 307 container_of(subreq, struct cifs_io_subrequest, subreq); 308 int rc = subreq->error; 309 310 if (rdata->subreq.source == NETFS_DOWNLOAD_FROM_SERVER) { 311 #ifdef CONFIG_CIFS_SMB_DIRECT 312 if (rdata->mr) { 313 smbd_deregister_mr(rdata->mr); 314 rdata->mr = NULL; 315 } 316 #endif 317 } 318 319 if (rdata->credits.value != 0) 320 trace_smb3_rw_credits(rdata->rreq->debug_id, 321 rdata->subreq.debug_index, 322 rdata->credits.value, 323 rdata->server ? rdata->server->credits : 0, 324 rdata->server ? rdata->server->in_flight : 0, 325 -rdata->credits.value, 326 cifs_trace_rw_credits_free_subreq); 327 328 add_credits_and_wake_if(rdata->server, &rdata->credits, 0); 329 if (rdata->have_xid) 330 free_xid(rdata->xid); 331 } 332 333 const struct netfs_request_ops cifs_req_ops = { 334 .request_pool = &cifs_io_request_pool, 335 .subrequest_pool = &cifs_io_subrequest_pool, 336 .init_request = cifs_init_request, 337 .free_request = cifs_free_request, 338 .free_subrequest = cifs_free_subrequest, 339 .clamp_length = cifs_clamp_length, 340 .issue_read = cifs_req_issue_read, 341 .done = cifs_rreq_done, 342 .post_modify = cifs_post_modify, 343 .begin_writeback = cifs_begin_writeback, 344 .prepare_write = cifs_prepare_write, 345 .issue_write = cifs_issue_write, 346 .invalidate_cache = cifs_netfs_invalidate_cache, 347 }; 348 349 /* 350 * Mark as invalid, all open files on tree connections since they 351 * were closed when session to server was lost. 352 */ 353 void 354 cifs_mark_open_files_invalid(struct cifs_tcon *tcon) 355 { 356 struct cifsFileInfo *open_file = NULL; 357 struct list_head *tmp; 358 struct list_head *tmp1; 359 360 /* only send once per connect */ 361 spin_lock(&tcon->tc_lock); 362 if (tcon->need_reconnect) 363 tcon->status = TID_NEED_RECON; 364 365 if (tcon->status != TID_NEED_RECON) { 366 spin_unlock(&tcon->tc_lock); 367 return; 368 } 369 tcon->status = TID_IN_FILES_INVALIDATE; 370 spin_unlock(&tcon->tc_lock); 371 372 /* list all files open on tree connection and mark them invalid */ 373 spin_lock(&tcon->open_file_lock); 374 list_for_each_safe(tmp, tmp1, &tcon->openFileList) { 375 open_file = list_entry(tmp, struct cifsFileInfo, tlist); 376 open_file->invalidHandle = true; 377 open_file->oplock_break_cancelled = true; 378 } 379 spin_unlock(&tcon->open_file_lock); 380 381 invalidate_all_cached_dirs(tcon); 382 spin_lock(&tcon->tc_lock); 383 if (tcon->status == TID_IN_FILES_INVALIDATE) 384 tcon->status = TID_NEED_TCON; 385 spin_unlock(&tcon->tc_lock); 386 387 /* 388 * BB Add call to invalidate_inodes(sb) for all superblocks mounted 389 * to this tcon. 390 */ 391 } 392 393 static inline int cifs_convert_flags(unsigned int flags, int rdwr_for_fscache) 394 { 395 if ((flags & O_ACCMODE) == O_RDONLY) 396 return GENERIC_READ; 397 else if ((flags & O_ACCMODE) == O_WRONLY) 398 return rdwr_for_fscache == 1 ? (GENERIC_READ | GENERIC_WRITE) : GENERIC_WRITE; 399 else if ((flags & O_ACCMODE) == O_RDWR) { 400 /* GENERIC_ALL is too much permission to request 401 can cause unnecessary access denied on create */ 402 /* return GENERIC_ALL; */ 403 return (GENERIC_READ | GENERIC_WRITE); 404 } 405 406 return (READ_CONTROL | FILE_WRITE_ATTRIBUTES | FILE_READ_ATTRIBUTES | 407 FILE_WRITE_EA | FILE_APPEND_DATA | FILE_WRITE_DATA | 408 FILE_READ_DATA); 409 } 410 411 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY 412 static u32 cifs_posix_convert_flags(unsigned int flags) 413 { 414 u32 posix_flags = 0; 415 416 if ((flags & O_ACCMODE) == O_RDONLY) 417 posix_flags = SMB_O_RDONLY; 418 else if ((flags & O_ACCMODE) == O_WRONLY) 419 posix_flags = SMB_O_WRONLY; 420 else if ((flags & O_ACCMODE) == O_RDWR) 421 posix_flags = SMB_O_RDWR; 422 423 if (flags & O_CREAT) { 424 posix_flags |= SMB_O_CREAT; 425 if (flags & O_EXCL) 426 posix_flags |= SMB_O_EXCL; 427 } else if (flags & O_EXCL) 428 cifs_dbg(FYI, "Application %s pid %d has incorrectly set O_EXCL flag but not O_CREAT on file open. Ignoring O_EXCL\n", 429 current->comm, current->tgid); 430 431 if (flags & O_TRUNC) 432 posix_flags |= SMB_O_TRUNC; 433 /* be safe and imply O_SYNC for O_DSYNC */ 434 if (flags & O_DSYNC) 435 posix_flags |= SMB_O_SYNC; 436 if (flags & O_DIRECTORY) 437 posix_flags |= SMB_O_DIRECTORY; 438 if (flags & O_NOFOLLOW) 439 posix_flags |= SMB_O_NOFOLLOW; 440 if (flags & O_DIRECT) 441 posix_flags |= SMB_O_DIRECT; 442 443 return posix_flags; 444 } 445 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */ 446 447 static inline int cifs_get_disposition(unsigned int flags) 448 { 449 if ((flags & (O_CREAT | O_EXCL)) == (O_CREAT | O_EXCL)) 450 return FILE_CREATE; 451 else if ((flags & (O_CREAT | O_TRUNC)) == (O_CREAT | O_TRUNC)) 452 return FILE_OVERWRITE_IF; 453 else if ((flags & O_CREAT) == O_CREAT) 454 return FILE_OPEN_IF; 455 else if ((flags & O_TRUNC) == O_TRUNC) 456 return FILE_OVERWRITE; 457 else 458 return FILE_OPEN; 459 } 460 461 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY 462 int cifs_posix_open(const char *full_path, struct inode **pinode, 463 struct super_block *sb, int mode, unsigned int f_flags, 464 __u32 *poplock, __u16 *pnetfid, unsigned int xid) 465 { 466 int rc; 467 FILE_UNIX_BASIC_INFO *presp_data; 468 __u32 posix_flags = 0; 469 struct cifs_sb_info *cifs_sb = CIFS_SB(sb); 470 struct cifs_fattr fattr; 471 struct tcon_link *tlink; 472 struct cifs_tcon *tcon; 473 474 cifs_dbg(FYI, "posix open %s\n", full_path); 475 476 presp_data = kzalloc(sizeof(FILE_UNIX_BASIC_INFO), GFP_KERNEL); 477 if (presp_data == NULL) 478 return -ENOMEM; 479 480 tlink = cifs_sb_tlink(cifs_sb); 481 if (IS_ERR(tlink)) { 482 rc = PTR_ERR(tlink); 483 goto posix_open_ret; 484 } 485 486 tcon = tlink_tcon(tlink); 487 mode &= ~current_umask(); 488 489 posix_flags = cifs_posix_convert_flags(f_flags); 490 rc = CIFSPOSIXCreate(xid, tcon, posix_flags, mode, pnetfid, presp_data, 491 poplock, full_path, cifs_sb->local_nls, 492 cifs_remap(cifs_sb)); 493 cifs_put_tlink(tlink); 494 495 if (rc) 496 goto posix_open_ret; 497 498 if (presp_data->Type == cpu_to_le32(-1)) 499 goto posix_open_ret; /* open ok, caller does qpathinfo */ 500 501 if (!pinode) 502 goto posix_open_ret; /* caller does not need info */ 503 504 cifs_unix_basic_to_fattr(&fattr, presp_data, cifs_sb); 505 506 /* get new inode and set it up */ 507 if (*pinode == NULL) { 508 cifs_fill_uniqueid(sb, &fattr); 509 *pinode = cifs_iget(sb, &fattr); 510 if (!*pinode) { 511 rc = -ENOMEM; 512 goto posix_open_ret; 513 } 514 } else { 515 cifs_revalidate_mapping(*pinode); 516 rc = cifs_fattr_to_inode(*pinode, &fattr, false); 517 } 518 519 posix_open_ret: 520 kfree(presp_data); 521 return rc; 522 } 523 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */ 524 525 static int cifs_nt_open(const char *full_path, struct inode *inode, struct cifs_sb_info *cifs_sb, 526 struct cifs_tcon *tcon, unsigned int f_flags, __u32 *oplock, 527 struct cifs_fid *fid, unsigned int xid, struct cifs_open_info_data *buf) 528 { 529 int rc; 530 int desired_access; 531 int disposition; 532 int create_options = CREATE_NOT_DIR; 533 struct TCP_Server_Info *server = tcon->ses->server; 534 struct cifs_open_parms oparms; 535 int rdwr_for_fscache = 0; 536 537 if (!server->ops->open) 538 return -ENOSYS; 539 540 /* If we're caching, we need to be able to fill in around partial writes. */ 541 if (cifs_fscache_enabled(inode) && (f_flags & O_ACCMODE) == O_WRONLY) 542 rdwr_for_fscache = 1; 543 544 desired_access = cifs_convert_flags(f_flags, rdwr_for_fscache); 545 546 /********************************************************************* 547 * open flag mapping table: 548 * 549 * POSIX Flag CIFS Disposition 550 * ---------- ---------------- 551 * O_CREAT FILE_OPEN_IF 552 * O_CREAT | O_EXCL FILE_CREATE 553 * O_CREAT | O_TRUNC FILE_OVERWRITE_IF 554 * O_TRUNC FILE_OVERWRITE 555 * none of the above FILE_OPEN 556 * 557 * Note that there is not a direct match between disposition 558 * FILE_SUPERSEDE (ie create whether or not file exists although 559 * O_CREAT | O_TRUNC is similar but truncates the existing 560 * file rather than creating a new file as FILE_SUPERSEDE does 561 * (which uses the attributes / metadata passed in on open call) 562 *? 563 *? O_SYNC is a reasonable match to CIFS writethrough flag 564 *? and the read write flags match reasonably. O_LARGEFILE 565 *? is irrelevant because largefile support is always used 566 *? by this client. Flags O_APPEND, O_DIRECT, O_DIRECTORY, 567 * O_FASYNC, O_NOFOLLOW, O_NONBLOCK need further investigation 568 *********************************************************************/ 569 570 disposition = cifs_get_disposition(f_flags); 571 572 /* BB pass O_SYNC flag through on file attributes .. BB */ 573 574 /* O_SYNC also has bit for O_DSYNC so following check picks up either */ 575 if (f_flags & O_SYNC) 576 create_options |= CREATE_WRITE_THROUGH; 577 578 if (f_flags & O_DIRECT) 579 create_options |= CREATE_NO_BUFFER; 580 581 retry_open: 582 oparms = (struct cifs_open_parms) { 583 .tcon = tcon, 584 .cifs_sb = cifs_sb, 585 .desired_access = desired_access, 586 .create_options = cifs_create_options(cifs_sb, create_options), 587 .disposition = disposition, 588 .path = full_path, 589 .fid = fid, 590 }; 591 592 rc = server->ops->open(xid, &oparms, oplock, buf); 593 if (rc) { 594 if (rc == -EACCES && rdwr_for_fscache == 1) { 595 desired_access = cifs_convert_flags(f_flags, 0); 596 rdwr_for_fscache = 2; 597 goto retry_open; 598 } 599 return rc; 600 } 601 if (rdwr_for_fscache == 2) 602 cifs_invalidate_cache(inode, FSCACHE_INVAL_DIO_WRITE); 603 604 /* TODO: Add support for calling posix query info but with passing in fid */ 605 if (tcon->unix_ext) 606 rc = cifs_get_inode_info_unix(&inode, full_path, inode->i_sb, 607 xid); 608 else 609 rc = cifs_get_inode_info(&inode, full_path, buf, inode->i_sb, 610 xid, fid); 611 612 if (rc) { 613 server->ops->close(xid, tcon, fid); 614 if (rc == -ESTALE) 615 rc = -EOPENSTALE; 616 } 617 618 return rc; 619 } 620 621 static bool 622 cifs_has_mand_locks(struct cifsInodeInfo *cinode) 623 { 624 struct cifs_fid_locks *cur; 625 bool has_locks = false; 626 627 down_read(&cinode->lock_sem); 628 list_for_each_entry(cur, &cinode->llist, llist) { 629 if (!list_empty(&cur->locks)) { 630 has_locks = true; 631 break; 632 } 633 } 634 up_read(&cinode->lock_sem); 635 return has_locks; 636 } 637 638 void 639 cifs_down_write(struct rw_semaphore *sem) 640 { 641 while (!down_write_trylock(sem)) 642 msleep(10); 643 } 644 645 static void cifsFileInfo_put_work(struct work_struct *work); 646 void serverclose_work(struct work_struct *work); 647 648 struct cifsFileInfo *cifs_new_fileinfo(struct cifs_fid *fid, struct file *file, 649 struct tcon_link *tlink, __u32 oplock, 650 const char *symlink_target) 651 { 652 struct dentry *dentry = file_dentry(file); 653 struct inode *inode = d_inode(dentry); 654 struct cifsInodeInfo *cinode = CIFS_I(inode); 655 struct cifsFileInfo *cfile; 656 struct cifs_fid_locks *fdlocks; 657 struct cifs_tcon *tcon = tlink_tcon(tlink); 658 struct TCP_Server_Info *server = tcon->ses->server; 659 660 cfile = kzalloc(sizeof(struct cifsFileInfo), GFP_KERNEL); 661 if (cfile == NULL) 662 return cfile; 663 664 fdlocks = kzalloc(sizeof(struct cifs_fid_locks), GFP_KERNEL); 665 if (!fdlocks) { 666 kfree(cfile); 667 return NULL; 668 } 669 670 if (symlink_target) { 671 cfile->symlink_target = kstrdup(symlink_target, GFP_KERNEL); 672 if (!cfile->symlink_target) { 673 kfree(fdlocks); 674 kfree(cfile); 675 return NULL; 676 } 677 } 678 679 INIT_LIST_HEAD(&fdlocks->locks); 680 fdlocks->cfile = cfile; 681 cfile->llist = fdlocks; 682 683 cfile->count = 1; 684 cfile->pid = current->tgid; 685 cfile->uid = current_fsuid(); 686 cfile->dentry = dget(dentry); 687 cfile->f_flags = file->f_flags; 688 cfile->invalidHandle = false; 689 cfile->deferred_close_scheduled = false; 690 cfile->tlink = cifs_get_tlink(tlink); 691 INIT_WORK(&cfile->oplock_break, cifs_oplock_break); 692 INIT_WORK(&cfile->put, cifsFileInfo_put_work); 693 INIT_WORK(&cfile->serverclose, serverclose_work); 694 INIT_DELAYED_WORK(&cfile->deferred, smb2_deferred_work_close); 695 mutex_init(&cfile->fh_mutex); 696 spin_lock_init(&cfile->file_info_lock); 697 698 cifs_sb_active(inode->i_sb); 699 700 /* 701 * If the server returned a read oplock and we have mandatory brlocks, 702 * set oplock level to None. 703 */ 704 if (server->ops->is_read_op(oplock) && cifs_has_mand_locks(cinode)) { 705 cifs_dbg(FYI, "Reset oplock val from read to None due to mand locks\n"); 706 oplock = 0; 707 } 708 709 cifs_down_write(&cinode->lock_sem); 710 list_add(&fdlocks->llist, &cinode->llist); 711 up_write(&cinode->lock_sem); 712 713 spin_lock(&tcon->open_file_lock); 714 if (fid->pending_open->oplock != CIFS_OPLOCK_NO_CHANGE && oplock) 715 oplock = fid->pending_open->oplock; 716 list_del(&fid->pending_open->olist); 717 718 fid->purge_cache = false; 719 server->ops->set_fid(cfile, fid, oplock); 720 721 list_add(&cfile->tlist, &tcon->openFileList); 722 atomic_inc(&tcon->num_local_opens); 723 724 /* if readable file instance put first in list*/ 725 spin_lock(&cinode->open_file_lock); 726 if (file->f_mode & FMODE_READ) 727 list_add(&cfile->flist, &cinode->openFileList); 728 else 729 list_add_tail(&cfile->flist, &cinode->openFileList); 730 spin_unlock(&cinode->open_file_lock); 731 spin_unlock(&tcon->open_file_lock); 732 733 if (fid->purge_cache) 734 cifs_zap_mapping(inode); 735 736 file->private_data = cfile; 737 return cfile; 738 } 739 740 struct cifsFileInfo * 741 cifsFileInfo_get(struct cifsFileInfo *cifs_file) 742 { 743 spin_lock(&cifs_file->file_info_lock); 744 cifsFileInfo_get_locked(cifs_file); 745 spin_unlock(&cifs_file->file_info_lock); 746 return cifs_file; 747 } 748 749 static void cifsFileInfo_put_final(struct cifsFileInfo *cifs_file) 750 { 751 struct inode *inode = d_inode(cifs_file->dentry); 752 struct cifsInodeInfo *cifsi = CIFS_I(inode); 753 struct cifsLockInfo *li, *tmp; 754 struct super_block *sb = inode->i_sb; 755 756 /* 757 * Delete any outstanding lock records. We'll lose them when the file 758 * is closed anyway. 759 */ 760 cifs_down_write(&cifsi->lock_sem); 761 list_for_each_entry_safe(li, tmp, &cifs_file->llist->locks, llist) { 762 list_del(&li->llist); 763 cifs_del_lock_waiters(li); 764 kfree(li); 765 } 766 list_del(&cifs_file->llist->llist); 767 kfree(cifs_file->llist); 768 up_write(&cifsi->lock_sem); 769 770 cifs_put_tlink(cifs_file->tlink); 771 dput(cifs_file->dentry); 772 cifs_sb_deactive(sb); 773 kfree(cifs_file->symlink_target); 774 kfree(cifs_file); 775 } 776 777 static void cifsFileInfo_put_work(struct work_struct *work) 778 { 779 struct cifsFileInfo *cifs_file = container_of(work, 780 struct cifsFileInfo, put); 781 782 cifsFileInfo_put_final(cifs_file); 783 } 784 785 void serverclose_work(struct work_struct *work) 786 { 787 struct cifsFileInfo *cifs_file = container_of(work, 788 struct cifsFileInfo, serverclose); 789 790 struct cifs_tcon *tcon = tlink_tcon(cifs_file->tlink); 791 792 struct TCP_Server_Info *server = tcon->ses->server; 793 int rc = 0; 794 int retries = 0; 795 int MAX_RETRIES = 4; 796 797 do { 798 if (server->ops->close_getattr) 799 rc = server->ops->close_getattr(0, tcon, cifs_file); 800 else if (server->ops->close) 801 rc = server->ops->close(0, tcon, &cifs_file->fid); 802 803 if (rc == -EBUSY || rc == -EAGAIN) { 804 retries++; 805 msleep(250); 806 } 807 } while ((rc == -EBUSY || rc == -EAGAIN) && (retries < MAX_RETRIES) 808 ); 809 810 if (retries == MAX_RETRIES) 811 pr_warn("Serverclose failed %d times, giving up\n", MAX_RETRIES); 812 813 if (cifs_file->offload) 814 queue_work(fileinfo_put_wq, &cifs_file->put); 815 else 816 cifsFileInfo_put_final(cifs_file); 817 } 818 819 /** 820 * cifsFileInfo_put - release a reference of file priv data 821 * 822 * Always potentially wait for oplock handler. See _cifsFileInfo_put(). 823 * 824 * @cifs_file: cifs/smb3 specific info (eg refcounts) for an open file 825 */ 826 void cifsFileInfo_put(struct cifsFileInfo *cifs_file) 827 { 828 _cifsFileInfo_put(cifs_file, true, true); 829 } 830 831 /** 832 * _cifsFileInfo_put - release a reference of file priv data 833 * 834 * This may involve closing the filehandle @cifs_file out on the 835 * server. Must be called without holding tcon->open_file_lock, 836 * cinode->open_file_lock and cifs_file->file_info_lock. 837 * 838 * If @wait_for_oplock_handler is true and we are releasing the last 839 * reference, wait for any running oplock break handler of the file 840 * and cancel any pending one. 841 * 842 * @cifs_file: cifs/smb3 specific info (eg refcounts) for an open file 843 * @wait_oplock_handler: must be false if called from oplock_break_handler 844 * @offload: not offloaded on close and oplock breaks 845 * 846 */ 847 void _cifsFileInfo_put(struct cifsFileInfo *cifs_file, 848 bool wait_oplock_handler, bool offload) 849 { 850 struct inode *inode = d_inode(cifs_file->dentry); 851 struct cifs_tcon *tcon = tlink_tcon(cifs_file->tlink); 852 struct TCP_Server_Info *server = tcon->ses->server; 853 struct cifsInodeInfo *cifsi = CIFS_I(inode); 854 struct super_block *sb = inode->i_sb; 855 struct cifs_sb_info *cifs_sb = CIFS_SB(sb); 856 struct cifs_fid fid = {}; 857 struct cifs_pending_open open; 858 bool oplock_break_cancelled; 859 bool serverclose_offloaded = false; 860 861 spin_lock(&tcon->open_file_lock); 862 spin_lock(&cifsi->open_file_lock); 863 spin_lock(&cifs_file->file_info_lock); 864 865 cifs_file->offload = offload; 866 if (--cifs_file->count > 0) { 867 spin_unlock(&cifs_file->file_info_lock); 868 spin_unlock(&cifsi->open_file_lock); 869 spin_unlock(&tcon->open_file_lock); 870 return; 871 } 872 spin_unlock(&cifs_file->file_info_lock); 873 874 if (server->ops->get_lease_key) 875 server->ops->get_lease_key(inode, &fid); 876 877 /* store open in pending opens to make sure we don't miss lease break */ 878 cifs_add_pending_open_locked(&fid, cifs_file->tlink, &open); 879 880 /* remove it from the lists */ 881 list_del(&cifs_file->flist); 882 list_del(&cifs_file->tlist); 883 atomic_dec(&tcon->num_local_opens); 884 885 if (list_empty(&cifsi->openFileList)) { 886 cifs_dbg(FYI, "closing last open instance for inode %p\n", 887 d_inode(cifs_file->dentry)); 888 /* 889 * In strict cache mode we need invalidate mapping on the last 890 * close because it may cause a error when we open this file 891 * again and get at least level II oplock. 892 */ 893 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_STRICT_IO) 894 set_bit(CIFS_INO_INVALID_MAPPING, &cifsi->flags); 895 cifs_set_oplock_level(cifsi, 0); 896 } 897 898 spin_unlock(&cifsi->open_file_lock); 899 spin_unlock(&tcon->open_file_lock); 900 901 oplock_break_cancelled = wait_oplock_handler ? 902 cancel_work_sync(&cifs_file->oplock_break) : false; 903 904 if (!tcon->need_reconnect && !cifs_file->invalidHandle) { 905 struct TCP_Server_Info *server = tcon->ses->server; 906 unsigned int xid; 907 int rc = 0; 908 909 xid = get_xid(); 910 if (server->ops->close_getattr) 911 rc = server->ops->close_getattr(xid, tcon, cifs_file); 912 else if (server->ops->close) 913 rc = server->ops->close(xid, tcon, &cifs_file->fid); 914 _free_xid(xid); 915 916 if (rc == -EBUSY || rc == -EAGAIN) { 917 // Server close failed, hence offloading it as an async op 918 queue_work(serverclose_wq, &cifs_file->serverclose); 919 serverclose_offloaded = true; 920 } 921 } 922 923 if (oplock_break_cancelled) 924 cifs_done_oplock_break(cifsi); 925 926 cifs_del_pending_open(&open); 927 928 // if serverclose has been offloaded to wq (on failure), it will 929 // handle offloading put as well. If serverclose not offloaded, 930 // we need to handle offloading put here. 931 if (!serverclose_offloaded) { 932 if (offload) 933 queue_work(fileinfo_put_wq, &cifs_file->put); 934 else 935 cifsFileInfo_put_final(cifs_file); 936 } 937 } 938 939 int cifs_open(struct inode *inode, struct file *file) 940 941 { 942 int rc = -EACCES; 943 unsigned int xid; 944 __u32 oplock; 945 struct cifs_sb_info *cifs_sb; 946 struct TCP_Server_Info *server; 947 struct cifs_tcon *tcon; 948 struct tcon_link *tlink; 949 struct cifsFileInfo *cfile = NULL; 950 void *page; 951 const char *full_path; 952 bool posix_open_ok = false; 953 struct cifs_fid fid = {}; 954 struct cifs_pending_open open; 955 struct cifs_open_info_data data = {}; 956 957 xid = get_xid(); 958 959 cifs_sb = CIFS_SB(inode->i_sb); 960 if (unlikely(cifs_forced_shutdown(cifs_sb))) { 961 free_xid(xid); 962 return -EIO; 963 } 964 965 tlink = cifs_sb_tlink(cifs_sb); 966 if (IS_ERR(tlink)) { 967 free_xid(xid); 968 return PTR_ERR(tlink); 969 } 970 tcon = tlink_tcon(tlink); 971 server = tcon->ses->server; 972 973 page = alloc_dentry_path(); 974 full_path = build_path_from_dentry(file_dentry(file), page); 975 if (IS_ERR(full_path)) { 976 rc = PTR_ERR(full_path); 977 goto out; 978 } 979 980 cifs_dbg(FYI, "inode = 0x%p file flags are 0x%x for %s\n", 981 inode, file->f_flags, full_path); 982 983 if (file->f_flags & O_DIRECT && 984 cifs_sb->mnt_cifs_flags & CIFS_MOUNT_STRICT_IO) { 985 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NO_BRL) 986 file->f_op = &cifs_file_direct_nobrl_ops; 987 else 988 file->f_op = &cifs_file_direct_ops; 989 } 990 991 /* Get the cached handle as SMB2 close is deferred */ 992 rc = cifs_get_readable_path(tcon, full_path, &cfile); 993 if (rc == 0) { 994 if (file->f_flags == cfile->f_flags) { 995 file->private_data = cfile; 996 spin_lock(&CIFS_I(inode)->deferred_lock); 997 cifs_del_deferred_close(cfile); 998 spin_unlock(&CIFS_I(inode)->deferred_lock); 999 goto use_cache; 1000 } else { 1001 _cifsFileInfo_put(cfile, true, false); 1002 } 1003 } 1004 1005 if (server->oplocks) 1006 oplock = REQ_OPLOCK; 1007 else 1008 oplock = 0; 1009 1010 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY 1011 if (!tcon->broken_posix_open && tcon->unix_ext && 1012 cap_unix(tcon->ses) && (CIFS_UNIX_POSIX_PATH_OPS_CAP & 1013 le64_to_cpu(tcon->fsUnixInfo.Capability))) { 1014 /* can not refresh inode info since size could be stale */ 1015 rc = cifs_posix_open(full_path, &inode, inode->i_sb, 1016 cifs_sb->ctx->file_mode /* ignored */, 1017 file->f_flags, &oplock, &fid.netfid, xid); 1018 if (rc == 0) { 1019 cifs_dbg(FYI, "posix open succeeded\n"); 1020 posix_open_ok = true; 1021 } else if ((rc == -EINVAL) || (rc == -EOPNOTSUPP)) { 1022 if (tcon->ses->serverNOS) 1023 cifs_dbg(VFS, "server %s of type %s returned unexpected error on SMB posix open, disabling posix open support. Check if server update available.\n", 1024 tcon->ses->ip_addr, 1025 tcon->ses->serverNOS); 1026 tcon->broken_posix_open = true; 1027 } else if ((rc != -EIO) && (rc != -EREMOTE) && 1028 (rc != -EOPNOTSUPP)) /* path not found or net err */ 1029 goto out; 1030 /* 1031 * Else fallthrough to retry open the old way on network i/o 1032 * or DFS errors. 1033 */ 1034 } 1035 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */ 1036 1037 if (server->ops->get_lease_key) 1038 server->ops->get_lease_key(inode, &fid); 1039 1040 cifs_add_pending_open(&fid, tlink, &open); 1041 1042 if (!posix_open_ok) { 1043 if (server->ops->get_lease_key) 1044 server->ops->get_lease_key(inode, &fid); 1045 1046 rc = cifs_nt_open(full_path, inode, cifs_sb, tcon, file->f_flags, &oplock, &fid, 1047 xid, &data); 1048 if (rc) { 1049 cifs_del_pending_open(&open); 1050 goto out; 1051 } 1052 } 1053 1054 cfile = cifs_new_fileinfo(&fid, file, tlink, oplock, data.symlink_target); 1055 if (cfile == NULL) { 1056 if (server->ops->close) 1057 server->ops->close(xid, tcon, &fid); 1058 cifs_del_pending_open(&open); 1059 rc = -ENOMEM; 1060 goto out; 1061 } 1062 1063 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY 1064 if ((oplock & CIFS_CREATE_ACTION) && !posix_open_ok && tcon->unix_ext) { 1065 /* 1066 * Time to set mode which we can not set earlier due to 1067 * problems creating new read-only files. 1068 */ 1069 struct cifs_unix_set_info_args args = { 1070 .mode = inode->i_mode, 1071 .uid = INVALID_UID, /* no change */ 1072 .gid = INVALID_GID, /* no change */ 1073 .ctime = NO_CHANGE_64, 1074 .atime = NO_CHANGE_64, 1075 .mtime = NO_CHANGE_64, 1076 .device = 0, 1077 }; 1078 CIFSSMBUnixSetFileInfo(xid, tcon, &args, fid.netfid, 1079 cfile->pid); 1080 } 1081 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */ 1082 1083 use_cache: 1084 fscache_use_cookie(cifs_inode_cookie(file_inode(file)), 1085 file->f_mode & FMODE_WRITE); 1086 if (!(file->f_flags & O_DIRECT)) 1087 goto out; 1088 if ((file->f_flags & (O_ACCMODE | O_APPEND)) == O_RDONLY) 1089 goto out; 1090 cifs_invalidate_cache(file_inode(file), FSCACHE_INVAL_DIO_WRITE); 1091 1092 out: 1093 free_dentry_path(page); 1094 free_xid(xid); 1095 cifs_put_tlink(tlink); 1096 cifs_free_open_info(&data); 1097 return rc; 1098 } 1099 1100 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY 1101 static int cifs_push_posix_locks(struct cifsFileInfo *cfile); 1102 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */ 1103 1104 /* 1105 * Try to reacquire byte range locks that were released when session 1106 * to server was lost. 1107 */ 1108 static int 1109 cifs_relock_file(struct cifsFileInfo *cfile) 1110 { 1111 struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry)); 1112 struct cifs_tcon *tcon = tlink_tcon(cfile->tlink); 1113 int rc = 0; 1114 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY 1115 struct cifs_sb_info *cifs_sb = CIFS_SB(cfile->dentry->d_sb); 1116 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */ 1117 1118 down_read_nested(&cinode->lock_sem, SINGLE_DEPTH_NESTING); 1119 if (cinode->can_cache_brlcks) { 1120 /* can cache locks - no need to relock */ 1121 up_read(&cinode->lock_sem); 1122 return rc; 1123 } 1124 1125 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY 1126 if (cap_unix(tcon->ses) && 1127 (CIFS_UNIX_FCNTL_CAP & le64_to_cpu(tcon->fsUnixInfo.Capability)) && 1128 ((cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOPOSIXBRL) == 0)) 1129 rc = cifs_push_posix_locks(cfile); 1130 else 1131 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */ 1132 rc = tcon->ses->server->ops->push_mand_locks(cfile); 1133 1134 up_read(&cinode->lock_sem); 1135 return rc; 1136 } 1137 1138 static int 1139 cifs_reopen_file(struct cifsFileInfo *cfile, bool can_flush) 1140 { 1141 int rc = -EACCES; 1142 unsigned int xid; 1143 __u32 oplock; 1144 struct cifs_sb_info *cifs_sb; 1145 struct cifs_tcon *tcon; 1146 struct TCP_Server_Info *server; 1147 struct cifsInodeInfo *cinode; 1148 struct inode *inode; 1149 void *page; 1150 const char *full_path; 1151 int desired_access; 1152 int disposition = FILE_OPEN; 1153 int create_options = CREATE_NOT_DIR; 1154 struct cifs_open_parms oparms; 1155 int rdwr_for_fscache = 0; 1156 1157 xid = get_xid(); 1158 mutex_lock(&cfile->fh_mutex); 1159 if (!cfile->invalidHandle) { 1160 mutex_unlock(&cfile->fh_mutex); 1161 free_xid(xid); 1162 return 0; 1163 } 1164 1165 inode = d_inode(cfile->dentry); 1166 cifs_sb = CIFS_SB(inode->i_sb); 1167 tcon = tlink_tcon(cfile->tlink); 1168 server = tcon->ses->server; 1169 1170 /* 1171 * Can not grab rename sem here because various ops, including those 1172 * that already have the rename sem can end up causing writepage to get 1173 * called and if the server was down that means we end up here, and we 1174 * can never tell if the caller already has the rename_sem. 1175 */ 1176 page = alloc_dentry_path(); 1177 full_path = build_path_from_dentry(cfile->dentry, page); 1178 if (IS_ERR(full_path)) { 1179 mutex_unlock(&cfile->fh_mutex); 1180 free_dentry_path(page); 1181 free_xid(xid); 1182 return PTR_ERR(full_path); 1183 } 1184 1185 cifs_dbg(FYI, "inode = 0x%p file flags 0x%x for %s\n", 1186 inode, cfile->f_flags, full_path); 1187 1188 if (tcon->ses->server->oplocks) 1189 oplock = REQ_OPLOCK; 1190 else 1191 oplock = 0; 1192 1193 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY 1194 if (tcon->unix_ext && cap_unix(tcon->ses) && 1195 (CIFS_UNIX_POSIX_PATH_OPS_CAP & 1196 le64_to_cpu(tcon->fsUnixInfo.Capability))) { 1197 /* 1198 * O_CREAT, O_EXCL and O_TRUNC already had their effect on the 1199 * original open. Must mask them off for a reopen. 1200 */ 1201 unsigned int oflags = cfile->f_flags & 1202 ~(O_CREAT | O_EXCL | O_TRUNC); 1203 1204 rc = cifs_posix_open(full_path, NULL, inode->i_sb, 1205 cifs_sb->ctx->file_mode /* ignored */, 1206 oflags, &oplock, &cfile->fid.netfid, xid); 1207 if (rc == 0) { 1208 cifs_dbg(FYI, "posix reopen succeeded\n"); 1209 oparms.reconnect = true; 1210 goto reopen_success; 1211 } 1212 /* 1213 * fallthrough to retry open the old way on errors, especially 1214 * in the reconnect path it is important to retry hard 1215 */ 1216 } 1217 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */ 1218 1219 /* If we're caching, we need to be able to fill in around partial writes. */ 1220 if (cifs_fscache_enabled(inode) && (cfile->f_flags & O_ACCMODE) == O_WRONLY) 1221 rdwr_for_fscache = 1; 1222 1223 desired_access = cifs_convert_flags(cfile->f_flags, rdwr_for_fscache); 1224 1225 /* O_SYNC also has bit for O_DSYNC so following check picks up either */ 1226 if (cfile->f_flags & O_SYNC) 1227 create_options |= CREATE_WRITE_THROUGH; 1228 1229 if (cfile->f_flags & O_DIRECT) 1230 create_options |= CREATE_NO_BUFFER; 1231 1232 if (server->ops->get_lease_key) 1233 server->ops->get_lease_key(inode, &cfile->fid); 1234 1235 retry_open: 1236 oparms = (struct cifs_open_parms) { 1237 .tcon = tcon, 1238 .cifs_sb = cifs_sb, 1239 .desired_access = desired_access, 1240 .create_options = cifs_create_options(cifs_sb, create_options), 1241 .disposition = disposition, 1242 .path = full_path, 1243 .fid = &cfile->fid, 1244 .reconnect = true, 1245 }; 1246 1247 /* 1248 * Can not refresh inode by passing in file_info buf to be returned by 1249 * ops->open and then calling get_inode_info with returned buf since 1250 * file might have write behind data that needs to be flushed and server 1251 * version of file size can be stale. If we knew for sure that inode was 1252 * not dirty locally we could do this. 1253 */ 1254 rc = server->ops->open(xid, &oparms, &oplock, NULL); 1255 if (rc == -ENOENT && oparms.reconnect == false) { 1256 /* durable handle timeout is expired - open the file again */ 1257 rc = server->ops->open(xid, &oparms, &oplock, NULL); 1258 /* indicate that we need to relock the file */ 1259 oparms.reconnect = true; 1260 } 1261 if (rc == -EACCES && rdwr_for_fscache == 1) { 1262 desired_access = cifs_convert_flags(cfile->f_flags, 0); 1263 rdwr_for_fscache = 2; 1264 goto retry_open; 1265 } 1266 1267 if (rc) { 1268 mutex_unlock(&cfile->fh_mutex); 1269 cifs_dbg(FYI, "cifs_reopen returned 0x%x\n", rc); 1270 cifs_dbg(FYI, "oplock: %d\n", oplock); 1271 goto reopen_error_exit; 1272 } 1273 1274 if (rdwr_for_fscache == 2) 1275 cifs_invalidate_cache(inode, FSCACHE_INVAL_DIO_WRITE); 1276 1277 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY 1278 reopen_success: 1279 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */ 1280 cfile->invalidHandle = false; 1281 mutex_unlock(&cfile->fh_mutex); 1282 cinode = CIFS_I(inode); 1283 1284 if (can_flush) { 1285 rc = filemap_write_and_wait(inode->i_mapping); 1286 if (!is_interrupt_error(rc)) 1287 mapping_set_error(inode->i_mapping, rc); 1288 1289 if (tcon->posix_extensions) { 1290 rc = smb311_posix_get_inode_info(&inode, full_path, 1291 NULL, inode->i_sb, xid); 1292 } else if (tcon->unix_ext) { 1293 rc = cifs_get_inode_info_unix(&inode, full_path, 1294 inode->i_sb, xid); 1295 } else { 1296 rc = cifs_get_inode_info(&inode, full_path, NULL, 1297 inode->i_sb, xid, NULL); 1298 } 1299 } 1300 /* 1301 * Else we are writing out data to server already and could deadlock if 1302 * we tried to flush data, and since we do not know if we have data that 1303 * would invalidate the current end of file on the server we can not go 1304 * to the server to get the new inode info. 1305 */ 1306 1307 /* 1308 * If the server returned a read oplock and we have mandatory brlocks, 1309 * set oplock level to None. 1310 */ 1311 if (server->ops->is_read_op(oplock) && cifs_has_mand_locks(cinode)) { 1312 cifs_dbg(FYI, "Reset oplock val from read to None due to mand locks\n"); 1313 oplock = 0; 1314 } 1315 1316 server->ops->set_fid(cfile, &cfile->fid, oplock); 1317 if (oparms.reconnect) 1318 cifs_relock_file(cfile); 1319 1320 reopen_error_exit: 1321 free_dentry_path(page); 1322 free_xid(xid); 1323 return rc; 1324 } 1325 1326 void smb2_deferred_work_close(struct work_struct *work) 1327 { 1328 struct cifsFileInfo *cfile = container_of(work, 1329 struct cifsFileInfo, deferred.work); 1330 1331 spin_lock(&CIFS_I(d_inode(cfile->dentry))->deferred_lock); 1332 cifs_del_deferred_close(cfile); 1333 cfile->deferred_close_scheduled = false; 1334 spin_unlock(&CIFS_I(d_inode(cfile->dentry))->deferred_lock); 1335 _cifsFileInfo_put(cfile, true, false); 1336 } 1337 1338 static bool 1339 smb2_can_defer_close(struct inode *inode, struct cifs_deferred_close *dclose) 1340 { 1341 struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb); 1342 struct cifsInodeInfo *cinode = CIFS_I(inode); 1343 1344 return (cifs_sb->ctx->closetimeo && cinode->lease_granted && dclose && 1345 (cinode->oplock == CIFS_CACHE_RHW_FLG || 1346 cinode->oplock == CIFS_CACHE_RH_FLG) && 1347 !test_bit(CIFS_INO_CLOSE_ON_LOCK, &cinode->flags)); 1348 1349 } 1350 1351 int cifs_close(struct inode *inode, struct file *file) 1352 { 1353 struct cifsFileInfo *cfile; 1354 struct cifsInodeInfo *cinode = CIFS_I(inode); 1355 struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb); 1356 struct cifs_deferred_close *dclose; 1357 1358 cifs_fscache_unuse_inode_cookie(inode, file->f_mode & FMODE_WRITE); 1359 1360 if (file->private_data != NULL) { 1361 cfile = file->private_data; 1362 file->private_data = NULL; 1363 dclose = kmalloc(sizeof(struct cifs_deferred_close), GFP_KERNEL); 1364 if ((cfile->status_file_deleted == false) && 1365 (smb2_can_defer_close(inode, dclose))) { 1366 if (test_and_clear_bit(CIFS_INO_MODIFIED_ATTR, &cinode->flags)) { 1367 inode_set_mtime_to_ts(inode, 1368 inode_set_ctime_current(inode)); 1369 } 1370 spin_lock(&cinode->deferred_lock); 1371 cifs_add_deferred_close(cfile, dclose); 1372 if (cfile->deferred_close_scheduled && 1373 delayed_work_pending(&cfile->deferred)) { 1374 /* 1375 * If there is no pending work, mod_delayed_work queues new work. 1376 * So, Increase the ref count to avoid use-after-free. 1377 */ 1378 if (!mod_delayed_work(deferredclose_wq, 1379 &cfile->deferred, cifs_sb->ctx->closetimeo)) 1380 cifsFileInfo_get(cfile); 1381 } else { 1382 /* Deferred close for files */ 1383 queue_delayed_work(deferredclose_wq, 1384 &cfile->deferred, cifs_sb->ctx->closetimeo); 1385 cfile->deferred_close_scheduled = true; 1386 spin_unlock(&cinode->deferred_lock); 1387 return 0; 1388 } 1389 spin_unlock(&cinode->deferred_lock); 1390 _cifsFileInfo_put(cfile, true, false); 1391 } else { 1392 _cifsFileInfo_put(cfile, true, false); 1393 kfree(dclose); 1394 } 1395 } 1396 1397 /* return code from the ->release op is always ignored */ 1398 return 0; 1399 } 1400 1401 void 1402 cifs_reopen_persistent_handles(struct cifs_tcon *tcon) 1403 { 1404 struct cifsFileInfo *open_file, *tmp; 1405 struct list_head tmp_list; 1406 1407 if (!tcon->use_persistent || !tcon->need_reopen_files) 1408 return; 1409 1410 tcon->need_reopen_files = false; 1411 1412 cifs_dbg(FYI, "Reopen persistent handles\n"); 1413 INIT_LIST_HEAD(&tmp_list); 1414 1415 /* list all files open on tree connection, reopen resilient handles */ 1416 spin_lock(&tcon->open_file_lock); 1417 list_for_each_entry(open_file, &tcon->openFileList, tlist) { 1418 if (!open_file->invalidHandle) 1419 continue; 1420 cifsFileInfo_get(open_file); 1421 list_add_tail(&open_file->rlist, &tmp_list); 1422 } 1423 spin_unlock(&tcon->open_file_lock); 1424 1425 list_for_each_entry_safe(open_file, tmp, &tmp_list, rlist) { 1426 if (cifs_reopen_file(open_file, false /* do not flush */)) 1427 tcon->need_reopen_files = true; 1428 list_del_init(&open_file->rlist); 1429 cifsFileInfo_put(open_file); 1430 } 1431 } 1432 1433 int cifs_closedir(struct inode *inode, struct file *file) 1434 { 1435 int rc = 0; 1436 unsigned int xid; 1437 struct cifsFileInfo *cfile = file->private_data; 1438 struct cifs_tcon *tcon; 1439 struct TCP_Server_Info *server; 1440 char *buf; 1441 1442 cifs_dbg(FYI, "Closedir inode = 0x%p\n", inode); 1443 1444 if (cfile == NULL) 1445 return rc; 1446 1447 xid = get_xid(); 1448 tcon = tlink_tcon(cfile->tlink); 1449 server = tcon->ses->server; 1450 1451 cifs_dbg(FYI, "Freeing private data in close dir\n"); 1452 spin_lock(&cfile->file_info_lock); 1453 if (server->ops->dir_needs_close(cfile)) { 1454 cfile->invalidHandle = true; 1455 spin_unlock(&cfile->file_info_lock); 1456 if (server->ops->close_dir) 1457 rc = server->ops->close_dir(xid, tcon, &cfile->fid); 1458 else 1459 rc = -ENOSYS; 1460 cifs_dbg(FYI, "Closing uncompleted readdir with rc %d\n", rc); 1461 /* not much we can do if it fails anyway, ignore rc */ 1462 rc = 0; 1463 } else 1464 spin_unlock(&cfile->file_info_lock); 1465 1466 buf = cfile->srch_inf.ntwrk_buf_start; 1467 if (buf) { 1468 cifs_dbg(FYI, "closedir free smb buf in srch struct\n"); 1469 cfile->srch_inf.ntwrk_buf_start = NULL; 1470 if (cfile->srch_inf.smallBuf) 1471 cifs_small_buf_release(buf); 1472 else 1473 cifs_buf_release(buf); 1474 } 1475 1476 cifs_put_tlink(cfile->tlink); 1477 kfree(file->private_data); 1478 file->private_data = NULL; 1479 /* BB can we lock the filestruct while this is going on? */ 1480 free_xid(xid); 1481 return rc; 1482 } 1483 1484 static struct cifsLockInfo * 1485 cifs_lock_init(__u64 offset, __u64 length, __u8 type, __u16 flags) 1486 { 1487 struct cifsLockInfo *lock = 1488 kmalloc(sizeof(struct cifsLockInfo), GFP_KERNEL); 1489 if (!lock) 1490 return lock; 1491 lock->offset = offset; 1492 lock->length = length; 1493 lock->type = type; 1494 lock->pid = current->tgid; 1495 lock->flags = flags; 1496 INIT_LIST_HEAD(&lock->blist); 1497 init_waitqueue_head(&lock->block_q); 1498 return lock; 1499 } 1500 1501 void 1502 cifs_del_lock_waiters(struct cifsLockInfo *lock) 1503 { 1504 struct cifsLockInfo *li, *tmp; 1505 list_for_each_entry_safe(li, tmp, &lock->blist, blist) { 1506 list_del_init(&li->blist); 1507 wake_up(&li->block_q); 1508 } 1509 } 1510 1511 #define CIFS_LOCK_OP 0 1512 #define CIFS_READ_OP 1 1513 #define CIFS_WRITE_OP 2 1514 1515 /* @rw_check : 0 - no op, 1 - read, 2 - write */ 1516 static bool 1517 cifs_find_fid_lock_conflict(struct cifs_fid_locks *fdlocks, __u64 offset, 1518 __u64 length, __u8 type, __u16 flags, 1519 struct cifsFileInfo *cfile, 1520 struct cifsLockInfo **conf_lock, int rw_check) 1521 { 1522 struct cifsLockInfo *li; 1523 struct cifsFileInfo *cur_cfile = fdlocks->cfile; 1524 struct TCP_Server_Info *server = tlink_tcon(cfile->tlink)->ses->server; 1525 1526 list_for_each_entry(li, &fdlocks->locks, llist) { 1527 if (offset + length <= li->offset || 1528 offset >= li->offset + li->length) 1529 continue; 1530 if (rw_check != CIFS_LOCK_OP && current->tgid == li->pid && 1531 server->ops->compare_fids(cfile, cur_cfile)) { 1532 /* shared lock prevents write op through the same fid */ 1533 if (!(li->type & server->vals->shared_lock_type) || 1534 rw_check != CIFS_WRITE_OP) 1535 continue; 1536 } 1537 if ((type & server->vals->shared_lock_type) && 1538 ((server->ops->compare_fids(cfile, cur_cfile) && 1539 current->tgid == li->pid) || type == li->type)) 1540 continue; 1541 if (rw_check == CIFS_LOCK_OP && 1542 (flags & FL_OFDLCK) && (li->flags & FL_OFDLCK) && 1543 server->ops->compare_fids(cfile, cur_cfile)) 1544 continue; 1545 if (conf_lock) 1546 *conf_lock = li; 1547 return true; 1548 } 1549 return false; 1550 } 1551 1552 bool 1553 cifs_find_lock_conflict(struct cifsFileInfo *cfile, __u64 offset, __u64 length, 1554 __u8 type, __u16 flags, 1555 struct cifsLockInfo **conf_lock, int rw_check) 1556 { 1557 bool rc = false; 1558 struct cifs_fid_locks *cur; 1559 struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry)); 1560 1561 list_for_each_entry(cur, &cinode->llist, llist) { 1562 rc = cifs_find_fid_lock_conflict(cur, offset, length, type, 1563 flags, cfile, conf_lock, 1564 rw_check); 1565 if (rc) 1566 break; 1567 } 1568 1569 return rc; 1570 } 1571 1572 /* 1573 * Check if there is another lock that prevents us to set the lock (mandatory 1574 * style). If such a lock exists, update the flock structure with its 1575 * properties. Otherwise, set the flock type to F_UNLCK if we can cache brlocks 1576 * or leave it the same if we can't. Returns 0 if we don't need to request to 1577 * the server or 1 otherwise. 1578 */ 1579 static int 1580 cifs_lock_test(struct cifsFileInfo *cfile, __u64 offset, __u64 length, 1581 __u8 type, struct file_lock *flock) 1582 { 1583 int rc = 0; 1584 struct cifsLockInfo *conf_lock; 1585 struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry)); 1586 struct TCP_Server_Info *server = tlink_tcon(cfile->tlink)->ses->server; 1587 bool exist; 1588 1589 down_read(&cinode->lock_sem); 1590 1591 exist = cifs_find_lock_conflict(cfile, offset, length, type, 1592 flock->c.flc_flags, &conf_lock, 1593 CIFS_LOCK_OP); 1594 if (exist) { 1595 flock->fl_start = conf_lock->offset; 1596 flock->fl_end = conf_lock->offset + conf_lock->length - 1; 1597 flock->c.flc_pid = conf_lock->pid; 1598 if (conf_lock->type & server->vals->shared_lock_type) 1599 flock->c.flc_type = F_RDLCK; 1600 else 1601 flock->c.flc_type = F_WRLCK; 1602 } else if (!cinode->can_cache_brlcks) 1603 rc = 1; 1604 else 1605 flock->c.flc_type = F_UNLCK; 1606 1607 up_read(&cinode->lock_sem); 1608 return rc; 1609 } 1610 1611 static void 1612 cifs_lock_add(struct cifsFileInfo *cfile, struct cifsLockInfo *lock) 1613 { 1614 struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry)); 1615 cifs_down_write(&cinode->lock_sem); 1616 list_add_tail(&lock->llist, &cfile->llist->locks); 1617 up_write(&cinode->lock_sem); 1618 } 1619 1620 /* 1621 * Set the byte-range lock (mandatory style). Returns: 1622 * 1) 0, if we set the lock and don't need to request to the server; 1623 * 2) 1, if no locks prevent us but we need to request to the server; 1624 * 3) -EACCES, if there is a lock that prevents us and wait is false. 1625 */ 1626 static int 1627 cifs_lock_add_if(struct cifsFileInfo *cfile, struct cifsLockInfo *lock, 1628 bool wait) 1629 { 1630 struct cifsLockInfo *conf_lock; 1631 struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry)); 1632 bool exist; 1633 int rc = 0; 1634 1635 try_again: 1636 exist = false; 1637 cifs_down_write(&cinode->lock_sem); 1638 1639 exist = cifs_find_lock_conflict(cfile, lock->offset, lock->length, 1640 lock->type, lock->flags, &conf_lock, 1641 CIFS_LOCK_OP); 1642 if (!exist && cinode->can_cache_brlcks) { 1643 list_add_tail(&lock->llist, &cfile->llist->locks); 1644 up_write(&cinode->lock_sem); 1645 return rc; 1646 } 1647 1648 if (!exist) 1649 rc = 1; 1650 else if (!wait) 1651 rc = -EACCES; 1652 else { 1653 list_add_tail(&lock->blist, &conf_lock->blist); 1654 up_write(&cinode->lock_sem); 1655 rc = wait_event_interruptible(lock->block_q, 1656 (lock->blist.prev == &lock->blist) && 1657 (lock->blist.next == &lock->blist)); 1658 if (!rc) 1659 goto try_again; 1660 cifs_down_write(&cinode->lock_sem); 1661 list_del_init(&lock->blist); 1662 } 1663 1664 up_write(&cinode->lock_sem); 1665 return rc; 1666 } 1667 1668 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY 1669 /* 1670 * Check if there is another lock that prevents us to set the lock (posix 1671 * style). If such a lock exists, update the flock structure with its 1672 * properties. Otherwise, set the flock type to F_UNLCK if we can cache brlocks 1673 * or leave it the same if we can't. Returns 0 if we don't need to request to 1674 * the server or 1 otherwise. 1675 */ 1676 static int 1677 cifs_posix_lock_test(struct file *file, struct file_lock *flock) 1678 { 1679 int rc = 0; 1680 struct cifsInodeInfo *cinode = CIFS_I(file_inode(file)); 1681 unsigned char saved_type = flock->c.flc_type; 1682 1683 if ((flock->c.flc_flags & FL_POSIX) == 0) 1684 return 1; 1685 1686 down_read(&cinode->lock_sem); 1687 posix_test_lock(file, flock); 1688 1689 if (lock_is_unlock(flock) && !cinode->can_cache_brlcks) { 1690 flock->c.flc_type = saved_type; 1691 rc = 1; 1692 } 1693 1694 up_read(&cinode->lock_sem); 1695 return rc; 1696 } 1697 1698 /* 1699 * Set the byte-range lock (posix style). Returns: 1700 * 1) <0, if the error occurs while setting the lock; 1701 * 2) 0, if we set the lock and don't need to request to the server; 1702 * 3) FILE_LOCK_DEFERRED, if we will wait for some other file_lock; 1703 * 4) FILE_LOCK_DEFERRED + 1, if we need to request to the server. 1704 */ 1705 static int 1706 cifs_posix_lock_set(struct file *file, struct file_lock *flock) 1707 { 1708 struct cifsInodeInfo *cinode = CIFS_I(file_inode(file)); 1709 int rc = FILE_LOCK_DEFERRED + 1; 1710 1711 if ((flock->c.flc_flags & FL_POSIX) == 0) 1712 return rc; 1713 1714 cifs_down_write(&cinode->lock_sem); 1715 if (!cinode->can_cache_brlcks) { 1716 up_write(&cinode->lock_sem); 1717 return rc; 1718 } 1719 1720 rc = posix_lock_file(file, flock, NULL); 1721 up_write(&cinode->lock_sem); 1722 return rc; 1723 } 1724 1725 int 1726 cifs_push_mandatory_locks(struct cifsFileInfo *cfile) 1727 { 1728 unsigned int xid; 1729 int rc = 0, stored_rc; 1730 struct cifsLockInfo *li, *tmp; 1731 struct cifs_tcon *tcon; 1732 unsigned int num, max_num, max_buf; 1733 LOCKING_ANDX_RANGE *buf, *cur; 1734 static const int types[] = { 1735 LOCKING_ANDX_LARGE_FILES, 1736 LOCKING_ANDX_SHARED_LOCK | LOCKING_ANDX_LARGE_FILES 1737 }; 1738 int i; 1739 1740 xid = get_xid(); 1741 tcon = tlink_tcon(cfile->tlink); 1742 1743 /* 1744 * Accessing maxBuf is racy with cifs_reconnect - need to store value 1745 * and check it before using. 1746 */ 1747 max_buf = tcon->ses->server->maxBuf; 1748 if (max_buf < (sizeof(struct smb_hdr) + sizeof(LOCKING_ANDX_RANGE))) { 1749 free_xid(xid); 1750 return -EINVAL; 1751 } 1752 1753 BUILD_BUG_ON(sizeof(struct smb_hdr) + sizeof(LOCKING_ANDX_RANGE) > 1754 PAGE_SIZE); 1755 max_buf = min_t(unsigned int, max_buf - sizeof(struct smb_hdr), 1756 PAGE_SIZE); 1757 max_num = (max_buf - sizeof(struct smb_hdr)) / 1758 sizeof(LOCKING_ANDX_RANGE); 1759 buf = kcalloc(max_num, sizeof(LOCKING_ANDX_RANGE), GFP_KERNEL); 1760 if (!buf) { 1761 free_xid(xid); 1762 return -ENOMEM; 1763 } 1764 1765 for (i = 0; i < 2; i++) { 1766 cur = buf; 1767 num = 0; 1768 list_for_each_entry_safe(li, tmp, &cfile->llist->locks, llist) { 1769 if (li->type != types[i]) 1770 continue; 1771 cur->Pid = cpu_to_le16(li->pid); 1772 cur->LengthLow = cpu_to_le32((u32)li->length); 1773 cur->LengthHigh = cpu_to_le32((u32)(li->length>>32)); 1774 cur->OffsetLow = cpu_to_le32((u32)li->offset); 1775 cur->OffsetHigh = cpu_to_le32((u32)(li->offset>>32)); 1776 if (++num == max_num) { 1777 stored_rc = cifs_lockv(xid, tcon, 1778 cfile->fid.netfid, 1779 (__u8)li->type, 0, num, 1780 buf); 1781 if (stored_rc) 1782 rc = stored_rc; 1783 cur = buf; 1784 num = 0; 1785 } else 1786 cur++; 1787 } 1788 1789 if (num) { 1790 stored_rc = cifs_lockv(xid, tcon, cfile->fid.netfid, 1791 (__u8)types[i], 0, num, buf); 1792 if (stored_rc) 1793 rc = stored_rc; 1794 } 1795 } 1796 1797 kfree(buf); 1798 free_xid(xid); 1799 return rc; 1800 } 1801 1802 static __u32 1803 hash_lockowner(fl_owner_t owner) 1804 { 1805 return cifs_lock_secret ^ hash32_ptr((const void *)owner); 1806 } 1807 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */ 1808 1809 struct lock_to_push { 1810 struct list_head llist; 1811 __u64 offset; 1812 __u64 length; 1813 __u32 pid; 1814 __u16 netfid; 1815 __u8 type; 1816 }; 1817 1818 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY 1819 static int 1820 cifs_push_posix_locks(struct cifsFileInfo *cfile) 1821 { 1822 struct inode *inode = d_inode(cfile->dentry); 1823 struct cifs_tcon *tcon = tlink_tcon(cfile->tlink); 1824 struct file_lock *flock; 1825 struct file_lock_context *flctx = locks_inode_context(inode); 1826 unsigned int count = 0, i; 1827 int rc = 0, xid, type; 1828 struct list_head locks_to_send, *el; 1829 struct lock_to_push *lck, *tmp; 1830 __u64 length; 1831 1832 xid = get_xid(); 1833 1834 if (!flctx) 1835 goto out; 1836 1837 spin_lock(&flctx->flc_lock); 1838 list_for_each(el, &flctx->flc_posix) { 1839 count++; 1840 } 1841 spin_unlock(&flctx->flc_lock); 1842 1843 INIT_LIST_HEAD(&locks_to_send); 1844 1845 /* 1846 * Allocating count locks is enough because no FL_POSIX locks can be 1847 * added to the list while we are holding cinode->lock_sem that 1848 * protects locking operations of this inode. 1849 */ 1850 for (i = 0; i < count; i++) { 1851 lck = kmalloc(sizeof(struct lock_to_push), GFP_KERNEL); 1852 if (!lck) { 1853 rc = -ENOMEM; 1854 goto err_out; 1855 } 1856 list_add_tail(&lck->llist, &locks_to_send); 1857 } 1858 1859 el = locks_to_send.next; 1860 spin_lock(&flctx->flc_lock); 1861 for_each_file_lock(flock, &flctx->flc_posix) { 1862 unsigned char ftype = flock->c.flc_type; 1863 1864 if (el == &locks_to_send) { 1865 /* 1866 * The list ended. We don't have enough allocated 1867 * structures - something is really wrong. 1868 */ 1869 cifs_dbg(VFS, "Can't push all brlocks!\n"); 1870 break; 1871 } 1872 length = cifs_flock_len(flock); 1873 if (ftype == F_RDLCK || ftype == F_SHLCK) 1874 type = CIFS_RDLCK; 1875 else 1876 type = CIFS_WRLCK; 1877 lck = list_entry(el, struct lock_to_push, llist); 1878 lck->pid = hash_lockowner(flock->c.flc_owner); 1879 lck->netfid = cfile->fid.netfid; 1880 lck->length = length; 1881 lck->type = type; 1882 lck->offset = flock->fl_start; 1883 } 1884 spin_unlock(&flctx->flc_lock); 1885 1886 list_for_each_entry_safe(lck, tmp, &locks_to_send, llist) { 1887 int stored_rc; 1888 1889 stored_rc = CIFSSMBPosixLock(xid, tcon, lck->netfid, lck->pid, 1890 lck->offset, lck->length, NULL, 1891 lck->type, 0); 1892 if (stored_rc) 1893 rc = stored_rc; 1894 list_del(&lck->llist); 1895 kfree(lck); 1896 } 1897 1898 out: 1899 free_xid(xid); 1900 return rc; 1901 err_out: 1902 list_for_each_entry_safe(lck, tmp, &locks_to_send, llist) { 1903 list_del(&lck->llist); 1904 kfree(lck); 1905 } 1906 goto out; 1907 } 1908 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */ 1909 1910 static int 1911 cifs_push_locks(struct cifsFileInfo *cfile) 1912 { 1913 struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry)); 1914 struct cifs_tcon *tcon = tlink_tcon(cfile->tlink); 1915 int rc = 0; 1916 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY 1917 struct cifs_sb_info *cifs_sb = CIFS_SB(cfile->dentry->d_sb); 1918 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */ 1919 1920 /* we are going to update can_cache_brlcks here - need a write access */ 1921 cifs_down_write(&cinode->lock_sem); 1922 if (!cinode->can_cache_brlcks) { 1923 up_write(&cinode->lock_sem); 1924 return rc; 1925 } 1926 1927 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY 1928 if (cap_unix(tcon->ses) && 1929 (CIFS_UNIX_FCNTL_CAP & le64_to_cpu(tcon->fsUnixInfo.Capability)) && 1930 ((cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOPOSIXBRL) == 0)) 1931 rc = cifs_push_posix_locks(cfile); 1932 else 1933 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */ 1934 rc = tcon->ses->server->ops->push_mand_locks(cfile); 1935 1936 cinode->can_cache_brlcks = false; 1937 up_write(&cinode->lock_sem); 1938 return rc; 1939 } 1940 1941 static void 1942 cifs_read_flock(struct file_lock *flock, __u32 *type, int *lock, int *unlock, 1943 bool *wait_flag, struct TCP_Server_Info *server) 1944 { 1945 if (flock->c.flc_flags & FL_POSIX) 1946 cifs_dbg(FYI, "Posix\n"); 1947 if (flock->c.flc_flags & FL_FLOCK) 1948 cifs_dbg(FYI, "Flock\n"); 1949 if (flock->c.flc_flags & FL_SLEEP) { 1950 cifs_dbg(FYI, "Blocking lock\n"); 1951 *wait_flag = true; 1952 } 1953 if (flock->c.flc_flags & FL_ACCESS) 1954 cifs_dbg(FYI, "Process suspended by mandatory locking - not implemented yet\n"); 1955 if (flock->c.flc_flags & FL_LEASE) 1956 cifs_dbg(FYI, "Lease on file - not implemented yet\n"); 1957 if (flock->c.flc_flags & 1958 (~(FL_POSIX | FL_FLOCK | FL_SLEEP | 1959 FL_ACCESS | FL_LEASE | FL_CLOSE | FL_OFDLCK))) 1960 cifs_dbg(FYI, "Unknown lock flags 0x%x\n", 1961 flock->c.flc_flags); 1962 1963 *type = server->vals->large_lock_type; 1964 if (lock_is_write(flock)) { 1965 cifs_dbg(FYI, "F_WRLCK\n"); 1966 *type |= server->vals->exclusive_lock_type; 1967 *lock = 1; 1968 } else if (lock_is_unlock(flock)) { 1969 cifs_dbg(FYI, "F_UNLCK\n"); 1970 *type |= server->vals->unlock_lock_type; 1971 *unlock = 1; 1972 /* Check if unlock includes more than one lock range */ 1973 } else if (lock_is_read(flock)) { 1974 cifs_dbg(FYI, "F_RDLCK\n"); 1975 *type |= server->vals->shared_lock_type; 1976 *lock = 1; 1977 } else if (flock->c.flc_type == F_EXLCK) { 1978 cifs_dbg(FYI, "F_EXLCK\n"); 1979 *type |= server->vals->exclusive_lock_type; 1980 *lock = 1; 1981 } else if (flock->c.flc_type == F_SHLCK) { 1982 cifs_dbg(FYI, "F_SHLCK\n"); 1983 *type |= server->vals->shared_lock_type; 1984 *lock = 1; 1985 } else 1986 cifs_dbg(FYI, "Unknown type of lock\n"); 1987 } 1988 1989 static int 1990 cifs_getlk(struct file *file, struct file_lock *flock, __u32 type, 1991 bool wait_flag, bool posix_lck, unsigned int xid) 1992 { 1993 int rc = 0; 1994 __u64 length = cifs_flock_len(flock); 1995 struct cifsFileInfo *cfile = (struct cifsFileInfo *)file->private_data; 1996 struct cifs_tcon *tcon = tlink_tcon(cfile->tlink); 1997 struct TCP_Server_Info *server = tcon->ses->server; 1998 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY 1999 __u16 netfid = cfile->fid.netfid; 2000 2001 if (posix_lck) { 2002 int posix_lock_type; 2003 2004 rc = cifs_posix_lock_test(file, flock); 2005 if (!rc) 2006 return rc; 2007 2008 if (type & server->vals->shared_lock_type) 2009 posix_lock_type = CIFS_RDLCK; 2010 else 2011 posix_lock_type = CIFS_WRLCK; 2012 rc = CIFSSMBPosixLock(xid, tcon, netfid, 2013 hash_lockowner(flock->c.flc_owner), 2014 flock->fl_start, length, flock, 2015 posix_lock_type, wait_flag); 2016 return rc; 2017 } 2018 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */ 2019 2020 rc = cifs_lock_test(cfile, flock->fl_start, length, type, flock); 2021 if (!rc) 2022 return rc; 2023 2024 /* BB we could chain these into one lock request BB */ 2025 rc = server->ops->mand_lock(xid, cfile, flock->fl_start, length, type, 2026 1, 0, false); 2027 if (rc == 0) { 2028 rc = server->ops->mand_lock(xid, cfile, flock->fl_start, length, 2029 type, 0, 1, false); 2030 flock->c.flc_type = F_UNLCK; 2031 if (rc != 0) 2032 cifs_dbg(VFS, "Error unlocking previously locked range %d during test of lock\n", 2033 rc); 2034 return 0; 2035 } 2036 2037 if (type & server->vals->shared_lock_type) { 2038 flock->c.flc_type = F_WRLCK; 2039 return 0; 2040 } 2041 2042 type &= ~server->vals->exclusive_lock_type; 2043 2044 rc = server->ops->mand_lock(xid, cfile, flock->fl_start, length, 2045 type | server->vals->shared_lock_type, 2046 1, 0, false); 2047 if (rc == 0) { 2048 rc = server->ops->mand_lock(xid, cfile, flock->fl_start, length, 2049 type | server->vals->shared_lock_type, 0, 1, false); 2050 flock->c.flc_type = F_RDLCK; 2051 if (rc != 0) 2052 cifs_dbg(VFS, "Error unlocking previously locked range %d during test of lock\n", 2053 rc); 2054 } else 2055 flock->c.flc_type = F_WRLCK; 2056 2057 return 0; 2058 } 2059 2060 void 2061 cifs_move_llist(struct list_head *source, struct list_head *dest) 2062 { 2063 struct list_head *li, *tmp; 2064 list_for_each_safe(li, tmp, source) 2065 list_move(li, dest); 2066 } 2067 2068 void 2069 cifs_free_llist(struct list_head *llist) 2070 { 2071 struct cifsLockInfo *li, *tmp; 2072 list_for_each_entry_safe(li, tmp, llist, llist) { 2073 cifs_del_lock_waiters(li); 2074 list_del(&li->llist); 2075 kfree(li); 2076 } 2077 } 2078 2079 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY 2080 int 2081 cifs_unlock_range(struct cifsFileInfo *cfile, struct file_lock *flock, 2082 unsigned int xid) 2083 { 2084 int rc = 0, stored_rc; 2085 static const int types[] = { 2086 LOCKING_ANDX_LARGE_FILES, 2087 LOCKING_ANDX_SHARED_LOCK | LOCKING_ANDX_LARGE_FILES 2088 }; 2089 unsigned int i; 2090 unsigned int max_num, num, max_buf; 2091 LOCKING_ANDX_RANGE *buf, *cur; 2092 struct cifs_tcon *tcon = tlink_tcon(cfile->tlink); 2093 struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry)); 2094 struct cifsLockInfo *li, *tmp; 2095 __u64 length = cifs_flock_len(flock); 2096 struct list_head tmp_llist; 2097 2098 INIT_LIST_HEAD(&tmp_llist); 2099 2100 /* 2101 * Accessing maxBuf is racy with cifs_reconnect - need to store value 2102 * and check it before using. 2103 */ 2104 max_buf = tcon->ses->server->maxBuf; 2105 if (max_buf < (sizeof(struct smb_hdr) + sizeof(LOCKING_ANDX_RANGE))) 2106 return -EINVAL; 2107 2108 BUILD_BUG_ON(sizeof(struct smb_hdr) + sizeof(LOCKING_ANDX_RANGE) > 2109 PAGE_SIZE); 2110 max_buf = min_t(unsigned int, max_buf - sizeof(struct smb_hdr), 2111 PAGE_SIZE); 2112 max_num = (max_buf - sizeof(struct smb_hdr)) / 2113 sizeof(LOCKING_ANDX_RANGE); 2114 buf = kcalloc(max_num, sizeof(LOCKING_ANDX_RANGE), GFP_KERNEL); 2115 if (!buf) 2116 return -ENOMEM; 2117 2118 cifs_down_write(&cinode->lock_sem); 2119 for (i = 0; i < 2; i++) { 2120 cur = buf; 2121 num = 0; 2122 list_for_each_entry_safe(li, tmp, &cfile->llist->locks, llist) { 2123 if (flock->fl_start > li->offset || 2124 (flock->fl_start + length) < 2125 (li->offset + li->length)) 2126 continue; 2127 if (current->tgid != li->pid) 2128 continue; 2129 if (types[i] != li->type) 2130 continue; 2131 if (cinode->can_cache_brlcks) { 2132 /* 2133 * We can cache brlock requests - simply remove 2134 * a lock from the file's list. 2135 */ 2136 list_del(&li->llist); 2137 cifs_del_lock_waiters(li); 2138 kfree(li); 2139 continue; 2140 } 2141 cur->Pid = cpu_to_le16(li->pid); 2142 cur->LengthLow = cpu_to_le32((u32)li->length); 2143 cur->LengthHigh = cpu_to_le32((u32)(li->length>>32)); 2144 cur->OffsetLow = cpu_to_le32((u32)li->offset); 2145 cur->OffsetHigh = cpu_to_le32((u32)(li->offset>>32)); 2146 /* 2147 * We need to save a lock here to let us add it again to 2148 * the file's list if the unlock range request fails on 2149 * the server. 2150 */ 2151 list_move(&li->llist, &tmp_llist); 2152 if (++num == max_num) { 2153 stored_rc = cifs_lockv(xid, tcon, 2154 cfile->fid.netfid, 2155 li->type, num, 0, buf); 2156 if (stored_rc) { 2157 /* 2158 * We failed on the unlock range 2159 * request - add all locks from the tmp 2160 * list to the head of the file's list. 2161 */ 2162 cifs_move_llist(&tmp_llist, 2163 &cfile->llist->locks); 2164 rc = stored_rc; 2165 } else 2166 /* 2167 * The unlock range request succeed - 2168 * free the tmp list. 2169 */ 2170 cifs_free_llist(&tmp_llist); 2171 cur = buf; 2172 num = 0; 2173 } else 2174 cur++; 2175 } 2176 if (num) { 2177 stored_rc = cifs_lockv(xid, tcon, cfile->fid.netfid, 2178 types[i], num, 0, buf); 2179 if (stored_rc) { 2180 cifs_move_llist(&tmp_llist, 2181 &cfile->llist->locks); 2182 rc = stored_rc; 2183 } else 2184 cifs_free_llist(&tmp_llist); 2185 } 2186 } 2187 2188 up_write(&cinode->lock_sem); 2189 kfree(buf); 2190 return rc; 2191 } 2192 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */ 2193 2194 static int 2195 cifs_setlk(struct file *file, struct file_lock *flock, __u32 type, 2196 bool wait_flag, bool posix_lck, int lock, int unlock, 2197 unsigned int xid) 2198 { 2199 int rc = 0; 2200 __u64 length = cifs_flock_len(flock); 2201 struct cifsFileInfo *cfile = (struct cifsFileInfo *)file->private_data; 2202 struct cifs_tcon *tcon = tlink_tcon(cfile->tlink); 2203 struct TCP_Server_Info *server = tcon->ses->server; 2204 struct inode *inode = d_inode(cfile->dentry); 2205 2206 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY 2207 if (posix_lck) { 2208 int posix_lock_type; 2209 2210 rc = cifs_posix_lock_set(file, flock); 2211 if (rc <= FILE_LOCK_DEFERRED) 2212 return rc; 2213 2214 if (type & server->vals->shared_lock_type) 2215 posix_lock_type = CIFS_RDLCK; 2216 else 2217 posix_lock_type = CIFS_WRLCK; 2218 2219 if (unlock == 1) 2220 posix_lock_type = CIFS_UNLCK; 2221 2222 rc = CIFSSMBPosixLock(xid, tcon, cfile->fid.netfid, 2223 hash_lockowner(flock->c.flc_owner), 2224 flock->fl_start, length, 2225 NULL, posix_lock_type, wait_flag); 2226 goto out; 2227 } 2228 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */ 2229 if (lock) { 2230 struct cifsLockInfo *lock; 2231 2232 lock = cifs_lock_init(flock->fl_start, length, type, 2233 flock->c.flc_flags); 2234 if (!lock) 2235 return -ENOMEM; 2236 2237 rc = cifs_lock_add_if(cfile, lock, wait_flag); 2238 if (rc < 0) { 2239 kfree(lock); 2240 return rc; 2241 } 2242 if (!rc) 2243 goto out; 2244 2245 /* 2246 * Windows 7 server can delay breaking lease from read to None 2247 * if we set a byte-range lock on a file - break it explicitly 2248 * before sending the lock to the server to be sure the next 2249 * read won't conflict with non-overlapted locks due to 2250 * pagereading. 2251 */ 2252 if (!CIFS_CACHE_WRITE(CIFS_I(inode)) && 2253 CIFS_CACHE_READ(CIFS_I(inode))) { 2254 cifs_zap_mapping(inode); 2255 cifs_dbg(FYI, "Set no oplock for inode=%p due to mand locks\n", 2256 inode); 2257 CIFS_I(inode)->oplock = 0; 2258 } 2259 2260 rc = server->ops->mand_lock(xid, cfile, flock->fl_start, length, 2261 type, 1, 0, wait_flag); 2262 if (rc) { 2263 kfree(lock); 2264 return rc; 2265 } 2266 2267 cifs_lock_add(cfile, lock); 2268 } else if (unlock) 2269 rc = server->ops->mand_unlock_range(cfile, flock, xid); 2270 2271 out: 2272 if ((flock->c.flc_flags & FL_POSIX) || (flock->c.flc_flags & FL_FLOCK)) { 2273 /* 2274 * If this is a request to remove all locks because we 2275 * are closing the file, it doesn't matter if the 2276 * unlocking failed as both cifs.ko and the SMB server 2277 * remove the lock on file close 2278 */ 2279 if (rc) { 2280 cifs_dbg(VFS, "%s failed rc=%d\n", __func__, rc); 2281 if (!(flock->c.flc_flags & FL_CLOSE)) 2282 return rc; 2283 } 2284 rc = locks_lock_file_wait(file, flock); 2285 } 2286 return rc; 2287 } 2288 2289 int cifs_flock(struct file *file, int cmd, struct file_lock *fl) 2290 { 2291 int rc, xid; 2292 int lock = 0, unlock = 0; 2293 bool wait_flag = false; 2294 bool posix_lck = false; 2295 struct cifs_sb_info *cifs_sb; 2296 struct cifs_tcon *tcon; 2297 struct cifsFileInfo *cfile; 2298 __u32 type; 2299 2300 xid = get_xid(); 2301 2302 if (!(fl->c.flc_flags & FL_FLOCK)) { 2303 rc = -ENOLCK; 2304 free_xid(xid); 2305 return rc; 2306 } 2307 2308 cfile = (struct cifsFileInfo *)file->private_data; 2309 tcon = tlink_tcon(cfile->tlink); 2310 2311 cifs_read_flock(fl, &type, &lock, &unlock, &wait_flag, 2312 tcon->ses->server); 2313 cifs_sb = CIFS_FILE_SB(file); 2314 2315 if (cap_unix(tcon->ses) && 2316 (CIFS_UNIX_FCNTL_CAP & le64_to_cpu(tcon->fsUnixInfo.Capability)) && 2317 ((cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOPOSIXBRL) == 0)) 2318 posix_lck = true; 2319 2320 if (!lock && !unlock) { 2321 /* 2322 * if no lock or unlock then nothing to do since we do not 2323 * know what it is 2324 */ 2325 rc = -EOPNOTSUPP; 2326 free_xid(xid); 2327 return rc; 2328 } 2329 2330 rc = cifs_setlk(file, fl, type, wait_flag, posix_lck, lock, unlock, 2331 xid); 2332 free_xid(xid); 2333 return rc; 2334 2335 2336 } 2337 2338 int cifs_lock(struct file *file, int cmd, struct file_lock *flock) 2339 { 2340 int rc, xid; 2341 int lock = 0, unlock = 0; 2342 bool wait_flag = false; 2343 bool posix_lck = false; 2344 struct cifs_sb_info *cifs_sb; 2345 struct cifs_tcon *tcon; 2346 struct cifsFileInfo *cfile; 2347 __u32 type; 2348 2349 rc = -EACCES; 2350 xid = get_xid(); 2351 2352 cifs_dbg(FYI, "%s: %pD2 cmd=0x%x type=0x%x flags=0x%x r=%lld:%lld\n", __func__, file, cmd, 2353 flock->c.flc_flags, flock->c.flc_type, 2354 (long long)flock->fl_start, 2355 (long long)flock->fl_end); 2356 2357 cfile = (struct cifsFileInfo *)file->private_data; 2358 tcon = tlink_tcon(cfile->tlink); 2359 2360 cifs_read_flock(flock, &type, &lock, &unlock, &wait_flag, 2361 tcon->ses->server); 2362 cifs_sb = CIFS_FILE_SB(file); 2363 set_bit(CIFS_INO_CLOSE_ON_LOCK, &CIFS_I(d_inode(cfile->dentry))->flags); 2364 2365 if (cap_unix(tcon->ses) && 2366 (CIFS_UNIX_FCNTL_CAP & le64_to_cpu(tcon->fsUnixInfo.Capability)) && 2367 ((cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOPOSIXBRL) == 0)) 2368 posix_lck = true; 2369 /* 2370 * BB add code here to normalize offset and length to account for 2371 * negative length which we can not accept over the wire. 2372 */ 2373 if (IS_GETLK(cmd)) { 2374 rc = cifs_getlk(file, flock, type, wait_flag, posix_lck, xid); 2375 free_xid(xid); 2376 return rc; 2377 } 2378 2379 if (!lock && !unlock) { 2380 /* 2381 * if no lock or unlock then nothing to do since we do not 2382 * know what it is 2383 */ 2384 free_xid(xid); 2385 return -EOPNOTSUPP; 2386 } 2387 2388 rc = cifs_setlk(file, flock, type, wait_flag, posix_lck, lock, unlock, 2389 xid); 2390 free_xid(xid); 2391 return rc; 2392 } 2393 2394 void cifs_write_subrequest_terminated(struct cifs_io_subrequest *wdata, ssize_t result, 2395 bool was_async) 2396 { 2397 struct netfs_io_request *wreq = wdata->rreq; 2398 struct netfs_inode *ictx = netfs_inode(wreq->inode); 2399 loff_t wrend; 2400 2401 if (result > 0) { 2402 wrend = wdata->subreq.start + wdata->subreq.transferred + result; 2403 2404 if (wrend > ictx->zero_point && 2405 (wdata->rreq->origin == NETFS_UNBUFFERED_WRITE || 2406 wdata->rreq->origin == NETFS_DIO_WRITE)) 2407 ictx->zero_point = wrend; 2408 if (wrend > ictx->remote_i_size) 2409 netfs_resize_file(ictx, wrend, true); 2410 } 2411 2412 netfs_write_subrequest_terminated(&wdata->subreq, result, was_async); 2413 } 2414 2415 struct cifsFileInfo *find_readable_file(struct cifsInodeInfo *cifs_inode, 2416 bool fsuid_only) 2417 { 2418 struct cifsFileInfo *open_file = NULL; 2419 struct cifs_sb_info *cifs_sb = CIFS_SB(cifs_inode->netfs.inode.i_sb); 2420 2421 /* only filter by fsuid on multiuser mounts */ 2422 if (!(cifs_sb->mnt_cifs_flags & CIFS_MOUNT_MULTIUSER)) 2423 fsuid_only = false; 2424 2425 spin_lock(&cifs_inode->open_file_lock); 2426 /* we could simply get the first_list_entry since write-only entries 2427 are always at the end of the list but since the first entry might 2428 have a close pending, we go through the whole list */ 2429 list_for_each_entry(open_file, &cifs_inode->openFileList, flist) { 2430 if (fsuid_only && !uid_eq(open_file->uid, current_fsuid())) 2431 continue; 2432 if (OPEN_FMODE(open_file->f_flags) & FMODE_READ) { 2433 if ((!open_file->invalidHandle)) { 2434 /* found a good file */ 2435 /* lock it so it will not be closed on us */ 2436 cifsFileInfo_get(open_file); 2437 spin_unlock(&cifs_inode->open_file_lock); 2438 return open_file; 2439 } /* else might as well continue, and look for 2440 another, or simply have the caller reopen it 2441 again rather than trying to fix this handle */ 2442 } else /* write only file */ 2443 break; /* write only files are last so must be done */ 2444 } 2445 spin_unlock(&cifs_inode->open_file_lock); 2446 return NULL; 2447 } 2448 2449 /* Return -EBADF if no handle is found and general rc otherwise */ 2450 int 2451 cifs_get_writable_file(struct cifsInodeInfo *cifs_inode, int flags, 2452 struct cifsFileInfo **ret_file) 2453 { 2454 struct cifsFileInfo *open_file, *inv_file = NULL; 2455 struct cifs_sb_info *cifs_sb; 2456 bool any_available = false; 2457 int rc = -EBADF; 2458 unsigned int refind = 0; 2459 bool fsuid_only = flags & FIND_WR_FSUID_ONLY; 2460 bool with_delete = flags & FIND_WR_WITH_DELETE; 2461 *ret_file = NULL; 2462 2463 /* 2464 * Having a null inode here (because mapping->host was set to zero by 2465 * the VFS or MM) should not happen but we had reports of on oops (due 2466 * to it being zero) during stress testcases so we need to check for it 2467 */ 2468 2469 if (cifs_inode == NULL) { 2470 cifs_dbg(VFS, "Null inode passed to cifs_writeable_file\n"); 2471 dump_stack(); 2472 return rc; 2473 } 2474 2475 cifs_sb = CIFS_SB(cifs_inode->netfs.inode.i_sb); 2476 2477 /* only filter by fsuid on multiuser mounts */ 2478 if (!(cifs_sb->mnt_cifs_flags & CIFS_MOUNT_MULTIUSER)) 2479 fsuid_only = false; 2480 2481 spin_lock(&cifs_inode->open_file_lock); 2482 refind_writable: 2483 if (refind > MAX_REOPEN_ATT) { 2484 spin_unlock(&cifs_inode->open_file_lock); 2485 return rc; 2486 } 2487 list_for_each_entry(open_file, &cifs_inode->openFileList, flist) { 2488 if (!any_available && open_file->pid != current->tgid) 2489 continue; 2490 if (fsuid_only && !uid_eq(open_file->uid, current_fsuid())) 2491 continue; 2492 if (with_delete && !(open_file->fid.access & DELETE)) 2493 continue; 2494 if (OPEN_FMODE(open_file->f_flags) & FMODE_WRITE) { 2495 if (!open_file->invalidHandle) { 2496 /* found a good writable file */ 2497 cifsFileInfo_get(open_file); 2498 spin_unlock(&cifs_inode->open_file_lock); 2499 *ret_file = open_file; 2500 return 0; 2501 } else { 2502 if (!inv_file) 2503 inv_file = open_file; 2504 } 2505 } 2506 } 2507 /* couldn't find useable FH with same pid, try any available */ 2508 if (!any_available) { 2509 any_available = true; 2510 goto refind_writable; 2511 } 2512 2513 if (inv_file) { 2514 any_available = false; 2515 cifsFileInfo_get(inv_file); 2516 } 2517 2518 spin_unlock(&cifs_inode->open_file_lock); 2519 2520 if (inv_file) { 2521 rc = cifs_reopen_file(inv_file, false); 2522 if (!rc) { 2523 *ret_file = inv_file; 2524 return 0; 2525 } 2526 2527 spin_lock(&cifs_inode->open_file_lock); 2528 list_move_tail(&inv_file->flist, &cifs_inode->openFileList); 2529 spin_unlock(&cifs_inode->open_file_lock); 2530 cifsFileInfo_put(inv_file); 2531 ++refind; 2532 inv_file = NULL; 2533 spin_lock(&cifs_inode->open_file_lock); 2534 goto refind_writable; 2535 } 2536 2537 return rc; 2538 } 2539 2540 struct cifsFileInfo * 2541 find_writable_file(struct cifsInodeInfo *cifs_inode, int flags) 2542 { 2543 struct cifsFileInfo *cfile; 2544 int rc; 2545 2546 rc = cifs_get_writable_file(cifs_inode, flags, &cfile); 2547 if (rc) 2548 cifs_dbg(FYI, "Couldn't find writable handle rc=%d\n", rc); 2549 2550 return cfile; 2551 } 2552 2553 int 2554 cifs_get_writable_path(struct cifs_tcon *tcon, const char *name, 2555 int flags, 2556 struct cifsFileInfo **ret_file) 2557 { 2558 struct cifsFileInfo *cfile; 2559 void *page = alloc_dentry_path(); 2560 2561 *ret_file = NULL; 2562 2563 spin_lock(&tcon->open_file_lock); 2564 list_for_each_entry(cfile, &tcon->openFileList, tlist) { 2565 struct cifsInodeInfo *cinode; 2566 const char *full_path = build_path_from_dentry(cfile->dentry, page); 2567 if (IS_ERR(full_path)) { 2568 spin_unlock(&tcon->open_file_lock); 2569 free_dentry_path(page); 2570 return PTR_ERR(full_path); 2571 } 2572 if (strcmp(full_path, name)) 2573 continue; 2574 2575 cinode = CIFS_I(d_inode(cfile->dentry)); 2576 spin_unlock(&tcon->open_file_lock); 2577 free_dentry_path(page); 2578 return cifs_get_writable_file(cinode, flags, ret_file); 2579 } 2580 2581 spin_unlock(&tcon->open_file_lock); 2582 free_dentry_path(page); 2583 return -ENOENT; 2584 } 2585 2586 int 2587 cifs_get_readable_path(struct cifs_tcon *tcon, const char *name, 2588 struct cifsFileInfo **ret_file) 2589 { 2590 struct cifsFileInfo *cfile; 2591 void *page = alloc_dentry_path(); 2592 2593 *ret_file = NULL; 2594 2595 spin_lock(&tcon->open_file_lock); 2596 list_for_each_entry(cfile, &tcon->openFileList, tlist) { 2597 struct cifsInodeInfo *cinode; 2598 const char *full_path = build_path_from_dentry(cfile->dentry, page); 2599 if (IS_ERR(full_path)) { 2600 spin_unlock(&tcon->open_file_lock); 2601 free_dentry_path(page); 2602 return PTR_ERR(full_path); 2603 } 2604 if (strcmp(full_path, name)) 2605 continue; 2606 2607 cinode = CIFS_I(d_inode(cfile->dentry)); 2608 spin_unlock(&tcon->open_file_lock); 2609 free_dentry_path(page); 2610 *ret_file = find_readable_file(cinode, 0); 2611 return *ret_file ? 0 : -ENOENT; 2612 } 2613 2614 spin_unlock(&tcon->open_file_lock); 2615 free_dentry_path(page); 2616 return -ENOENT; 2617 } 2618 2619 /* 2620 * Flush data on a strict file. 2621 */ 2622 int cifs_strict_fsync(struct file *file, loff_t start, loff_t end, 2623 int datasync) 2624 { 2625 unsigned int xid; 2626 int rc = 0; 2627 struct cifs_tcon *tcon; 2628 struct TCP_Server_Info *server; 2629 struct cifsFileInfo *smbfile = file->private_data; 2630 struct inode *inode = file_inode(file); 2631 struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb); 2632 2633 rc = file_write_and_wait_range(file, start, end); 2634 if (rc) { 2635 trace_cifs_fsync_err(inode->i_ino, rc); 2636 return rc; 2637 } 2638 2639 xid = get_xid(); 2640 2641 cifs_dbg(FYI, "Sync file - name: %pD datasync: 0x%x\n", 2642 file, datasync); 2643 2644 if (!CIFS_CACHE_READ(CIFS_I(inode))) { 2645 rc = cifs_zap_mapping(inode); 2646 if (rc) { 2647 cifs_dbg(FYI, "rc: %d during invalidate phase\n", rc); 2648 rc = 0; /* don't care about it in fsync */ 2649 } 2650 } 2651 2652 tcon = tlink_tcon(smbfile->tlink); 2653 if (!(cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOSSYNC)) { 2654 server = tcon->ses->server; 2655 if (server->ops->flush == NULL) { 2656 rc = -ENOSYS; 2657 goto strict_fsync_exit; 2658 } 2659 2660 if ((OPEN_FMODE(smbfile->f_flags) & FMODE_WRITE) == 0) { 2661 smbfile = find_writable_file(CIFS_I(inode), FIND_WR_ANY); 2662 if (smbfile) { 2663 rc = server->ops->flush(xid, tcon, &smbfile->fid); 2664 cifsFileInfo_put(smbfile); 2665 } else 2666 cifs_dbg(FYI, "ignore fsync for file not open for write\n"); 2667 } else 2668 rc = server->ops->flush(xid, tcon, &smbfile->fid); 2669 } 2670 2671 strict_fsync_exit: 2672 free_xid(xid); 2673 return rc; 2674 } 2675 2676 /* 2677 * Flush data on a non-strict data. 2678 */ 2679 int cifs_fsync(struct file *file, loff_t start, loff_t end, int datasync) 2680 { 2681 unsigned int xid; 2682 int rc = 0; 2683 struct cifs_tcon *tcon; 2684 struct TCP_Server_Info *server; 2685 struct cifsFileInfo *smbfile = file->private_data; 2686 struct inode *inode = file_inode(file); 2687 struct cifs_sb_info *cifs_sb = CIFS_FILE_SB(file); 2688 2689 rc = file_write_and_wait_range(file, start, end); 2690 if (rc) { 2691 trace_cifs_fsync_err(file_inode(file)->i_ino, rc); 2692 return rc; 2693 } 2694 2695 xid = get_xid(); 2696 2697 cifs_dbg(FYI, "Sync file - name: %pD datasync: 0x%x\n", 2698 file, datasync); 2699 2700 tcon = tlink_tcon(smbfile->tlink); 2701 if (!(cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOSSYNC)) { 2702 server = tcon->ses->server; 2703 if (server->ops->flush == NULL) { 2704 rc = -ENOSYS; 2705 goto fsync_exit; 2706 } 2707 2708 if ((OPEN_FMODE(smbfile->f_flags) & FMODE_WRITE) == 0) { 2709 smbfile = find_writable_file(CIFS_I(inode), FIND_WR_ANY); 2710 if (smbfile) { 2711 rc = server->ops->flush(xid, tcon, &smbfile->fid); 2712 cifsFileInfo_put(smbfile); 2713 } else 2714 cifs_dbg(FYI, "ignore fsync for file not open for write\n"); 2715 } else 2716 rc = server->ops->flush(xid, tcon, &smbfile->fid); 2717 } 2718 2719 fsync_exit: 2720 free_xid(xid); 2721 return rc; 2722 } 2723 2724 /* 2725 * As file closes, flush all cached write data for this inode checking 2726 * for write behind errors. 2727 */ 2728 int cifs_flush(struct file *file, fl_owner_t id) 2729 { 2730 struct inode *inode = file_inode(file); 2731 int rc = 0; 2732 2733 if (file->f_mode & FMODE_WRITE) 2734 rc = filemap_write_and_wait(inode->i_mapping); 2735 2736 cifs_dbg(FYI, "Flush inode %p file %p rc %d\n", inode, file, rc); 2737 if (rc) { 2738 /* get more nuanced writeback errors */ 2739 rc = filemap_check_wb_err(file->f_mapping, 0); 2740 trace_cifs_flush_err(inode->i_ino, rc); 2741 } 2742 return rc; 2743 } 2744 2745 static ssize_t 2746 cifs_writev(struct kiocb *iocb, struct iov_iter *from) 2747 { 2748 struct file *file = iocb->ki_filp; 2749 struct cifsFileInfo *cfile = (struct cifsFileInfo *)file->private_data; 2750 struct inode *inode = file->f_mapping->host; 2751 struct cifsInodeInfo *cinode = CIFS_I(inode); 2752 struct TCP_Server_Info *server = tlink_tcon(cfile->tlink)->ses->server; 2753 ssize_t rc; 2754 2755 rc = netfs_start_io_write(inode); 2756 if (rc < 0) 2757 return rc; 2758 2759 /* 2760 * We need to hold the sem to be sure nobody modifies lock list 2761 * with a brlock that prevents writing. 2762 */ 2763 down_read(&cinode->lock_sem); 2764 2765 rc = generic_write_checks(iocb, from); 2766 if (rc <= 0) 2767 goto out; 2768 2769 if (!cifs_find_lock_conflict(cfile, iocb->ki_pos, iov_iter_count(from), 2770 server->vals->exclusive_lock_type, 0, 2771 NULL, CIFS_WRITE_OP)) 2772 rc = netfs_buffered_write_iter_locked(iocb, from, NULL); 2773 else 2774 rc = -EACCES; 2775 out: 2776 up_read(&cinode->lock_sem); 2777 netfs_end_io_write(inode); 2778 if (rc > 0) 2779 rc = generic_write_sync(iocb, rc); 2780 return rc; 2781 } 2782 2783 ssize_t 2784 cifs_strict_writev(struct kiocb *iocb, struct iov_iter *from) 2785 { 2786 struct inode *inode = file_inode(iocb->ki_filp); 2787 struct cifsInodeInfo *cinode = CIFS_I(inode); 2788 struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb); 2789 struct cifsFileInfo *cfile = (struct cifsFileInfo *) 2790 iocb->ki_filp->private_data; 2791 struct cifs_tcon *tcon = tlink_tcon(cfile->tlink); 2792 ssize_t written; 2793 2794 written = cifs_get_writer(cinode); 2795 if (written) 2796 return written; 2797 2798 if (CIFS_CACHE_WRITE(cinode)) { 2799 if (cap_unix(tcon->ses) && 2800 (CIFS_UNIX_FCNTL_CAP & le64_to_cpu(tcon->fsUnixInfo.Capability)) && 2801 ((cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOPOSIXBRL) == 0)) { 2802 written = netfs_file_write_iter(iocb, from); 2803 goto out; 2804 } 2805 written = cifs_writev(iocb, from); 2806 goto out; 2807 } 2808 /* 2809 * For non-oplocked files in strict cache mode we need to write the data 2810 * to the server exactly from the pos to pos+len-1 rather than flush all 2811 * affected pages because it may cause a error with mandatory locks on 2812 * these pages but not on the region from pos to ppos+len-1. 2813 */ 2814 written = netfs_file_write_iter(iocb, from); 2815 if (CIFS_CACHE_READ(cinode)) { 2816 /* 2817 * We have read level caching and we have just sent a write 2818 * request to the server thus making data in the cache stale. 2819 * Zap the cache and set oplock/lease level to NONE to avoid 2820 * reading stale data from the cache. All subsequent read 2821 * operations will read new data from the server. 2822 */ 2823 cifs_zap_mapping(inode); 2824 cifs_dbg(FYI, "Set Oplock/Lease to NONE for inode=%p after write\n", 2825 inode); 2826 cinode->oplock = 0; 2827 } 2828 out: 2829 cifs_put_writer(cinode); 2830 return written; 2831 } 2832 2833 ssize_t cifs_loose_read_iter(struct kiocb *iocb, struct iov_iter *iter) 2834 { 2835 ssize_t rc; 2836 struct inode *inode = file_inode(iocb->ki_filp); 2837 2838 if (iocb->ki_flags & IOCB_DIRECT) 2839 return netfs_unbuffered_read_iter(iocb, iter); 2840 2841 rc = cifs_revalidate_mapping(inode); 2842 if (rc) 2843 return rc; 2844 2845 return netfs_file_read_iter(iocb, iter); 2846 } 2847 2848 ssize_t cifs_file_write_iter(struct kiocb *iocb, struct iov_iter *from) 2849 { 2850 struct inode *inode = file_inode(iocb->ki_filp); 2851 struct cifsInodeInfo *cinode = CIFS_I(inode); 2852 ssize_t written; 2853 int rc; 2854 2855 if (iocb->ki_filp->f_flags & O_DIRECT) { 2856 written = netfs_unbuffered_write_iter(iocb, from); 2857 if (written > 0 && CIFS_CACHE_READ(cinode)) { 2858 cifs_zap_mapping(inode); 2859 cifs_dbg(FYI, 2860 "Set no oplock for inode=%p after a write operation\n", 2861 inode); 2862 cinode->oplock = 0; 2863 } 2864 return written; 2865 } 2866 2867 written = cifs_get_writer(cinode); 2868 if (written) 2869 return written; 2870 2871 written = netfs_file_write_iter(iocb, from); 2872 2873 if (!CIFS_CACHE_WRITE(CIFS_I(inode))) { 2874 rc = filemap_fdatawrite(inode->i_mapping); 2875 if (rc) 2876 cifs_dbg(FYI, "cifs_file_write_iter: %d rc on %p inode\n", 2877 rc, inode); 2878 } 2879 2880 cifs_put_writer(cinode); 2881 return written; 2882 } 2883 2884 ssize_t 2885 cifs_strict_readv(struct kiocb *iocb, struct iov_iter *to) 2886 { 2887 struct inode *inode = file_inode(iocb->ki_filp); 2888 struct cifsInodeInfo *cinode = CIFS_I(inode); 2889 struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb); 2890 struct cifsFileInfo *cfile = (struct cifsFileInfo *) 2891 iocb->ki_filp->private_data; 2892 struct cifs_tcon *tcon = tlink_tcon(cfile->tlink); 2893 int rc = -EACCES; 2894 2895 /* 2896 * In strict cache mode we need to read from the server all the time 2897 * if we don't have level II oplock because the server can delay mtime 2898 * change - so we can't make a decision about inode invalidating. 2899 * And we can also fail with pagereading if there are mandatory locks 2900 * on pages affected by this read but not on the region from pos to 2901 * pos+len-1. 2902 */ 2903 if (!CIFS_CACHE_READ(cinode)) 2904 return netfs_unbuffered_read_iter(iocb, to); 2905 2906 if (cap_unix(tcon->ses) && 2907 (CIFS_UNIX_FCNTL_CAP & le64_to_cpu(tcon->fsUnixInfo.Capability)) && 2908 ((cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOPOSIXBRL) == 0)) { 2909 if (iocb->ki_flags & IOCB_DIRECT) 2910 return netfs_unbuffered_read_iter(iocb, to); 2911 return netfs_buffered_read_iter(iocb, to); 2912 } 2913 2914 /* 2915 * We need to hold the sem to be sure nobody modifies lock list 2916 * with a brlock that prevents reading. 2917 */ 2918 if (iocb->ki_flags & IOCB_DIRECT) { 2919 rc = netfs_start_io_direct(inode); 2920 if (rc < 0) 2921 goto out; 2922 rc = -EACCES; 2923 down_read(&cinode->lock_sem); 2924 if (!cifs_find_lock_conflict( 2925 cfile, iocb->ki_pos, iov_iter_count(to), 2926 tcon->ses->server->vals->shared_lock_type, 2927 0, NULL, CIFS_READ_OP)) 2928 rc = netfs_unbuffered_read_iter_locked(iocb, to); 2929 up_read(&cinode->lock_sem); 2930 netfs_end_io_direct(inode); 2931 } else { 2932 rc = netfs_start_io_read(inode); 2933 if (rc < 0) 2934 goto out; 2935 rc = -EACCES; 2936 down_read(&cinode->lock_sem); 2937 if (!cifs_find_lock_conflict( 2938 cfile, iocb->ki_pos, iov_iter_count(to), 2939 tcon->ses->server->vals->shared_lock_type, 2940 0, NULL, CIFS_READ_OP)) 2941 rc = filemap_read(iocb, to, 0); 2942 up_read(&cinode->lock_sem); 2943 netfs_end_io_read(inode); 2944 } 2945 out: 2946 return rc; 2947 } 2948 2949 static vm_fault_t cifs_page_mkwrite(struct vm_fault *vmf) 2950 { 2951 return netfs_page_mkwrite(vmf, NULL); 2952 } 2953 2954 static const struct vm_operations_struct cifs_file_vm_ops = { 2955 .fault = filemap_fault, 2956 .map_pages = filemap_map_pages, 2957 .page_mkwrite = cifs_page_mkwrite, 2958 }; 2959 2960 int cifs_file_strict_mmap(struct file *file, struct vm_area_struct *vma) 2961 { 2962 int xid, rc = 0; 2963 struct inode *inode = file_inode(file); 2964 2965 xid = get_xid(); 2966 2967 if (!CIFS_CACHE_READ(CIFS_I(inode))) 2968 rc = cifs_zap_mapping(inode); 2969 if (!rc) 2970 rc = generic_file_mmap(file, vma); 2971 if (!rc) 2972 vma->vm_ops = &cifs_file_vm_ops; 2973 2974 free_xid(xid); 2975 return rc; 2976 } 2977 2978 int cifs_file_mmap(struct file *file, struct vm_area_struct *vma) 2979 { 2980 int rc, xid; 2981 2982 xid = get_xid(); 2983 2984 rc = cifs_revalidate_file(file); 2985 if (rc) 2986 cifs_dbg(FYI, "Validation prior to mmap failed, error=%d\n", 2987 rc); 2988 if (!rc) 2989 rc = generic_file_mmap(file, vma); 2990 if (!rc) 2991 vma->vm_ops = &cifs_file_vm_ops; 2992 2993 free_xid(xid); 2994 return rc; 2995 } 2996 2997 static int is_inode_writable(struct cifsInodeInfo *cifs_inode) 2998 { 2999 struct cifsFileInfo *open_file; 3000 3001 spin_lock(&cifs_inode->open_file_lock); 3002 list_for_each_entry(open_file, &cifs_inode->openFileList, flist) { 3003 if (OPEN_FMODE(open_file->f_flags) & FMODE_WRITE) { 3004 spin_unlock(&cifs_inode->open_file_lock); 3005 return 1; 3006 } 3007 } 3008 spin_unlock(&cifs_inode->open_file_lock); 3009 return 0; 3010 } 3011 3012 /* We do not want to update the file size from server for inodes 3013 open for write - to avoid races with writepage extending 3014 the file - in the future we could consider allowing 3015 refreshing the inode only on increases in the file size 3016 but this is tricky to do without racing with writebehind 3017 page caching in the current Linux kernel design */ 3018 bool is_size_safe_to_change(struct cifsInodeInfo *cifsInode, __u64 end_of_file, 3019 bool from_readdir) 3020 { 3021 if (!cifsInode) 3022 return true; 3023 3024 if (is_inode_writable(cifsInode) || 3025 ((cifsInode->oplock & CIFS_CACHE_RW_FLG) != 0 && from_readdir)) { 3026 /* This inode is open for write at least once */ 3027 struct cifs_sb_info *cifs_sb; 3028 3029 cifs_sb = CIFS_SB(cifsInode->netfs.inode.i_sb); 3030 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_DIRECT_IO) { 3031 /* since no page cache to corrupt on directio 3032 we can change size safely */ 3033 return true; 3034 } 3035 3036 if (i_size_read(&cifsInode->netfs.inode) < end_of_file) 3037 return true; 3038 3039 return false; 3040 } else 3041 return true; 3042 } 3043 3044 void cifs_oplock_break(struct work_struct *work) 3045 { 3046 struct cifsFileInfo *cfile = container_of(work, struct cifsFileInfo, 3047 oplock_break); 3048 struct inode *inode = d_inode(cfile->dentry); 3049 struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb); 3050 struct cifsInodeInfo *cinode = CIFS_I(inode); 3051 struct cifs_tcon *tcon; 3052 struct TCP_Server_Info *server; 3053 struct tcon_link *tlink; 3054 int rc = 0; 3055 bool purge_cache = false, oplock_break_cancelled; 3056 __u64 persistent_fid, volatile_fid; 3057 __u16 net_fid; 3058 3059 wait_on_bit(&cinode->flags, CIFS_INODE_PENDING_WRITERS, 3060 TASK_UNINTERRUPTIBLE); 3061 3062 tlink = cifs_sb_tlink(cifs_sb); 3063 if (IS_ERR(tlink)) 3064 goto out; 3065 tcon = tlink_tcon(tlink); 3066 server = tcon->ses->server; 3067 3068 server->ops->downgrade_oplock(server, cinode, cfile->oplock_level, 3069 cfile->oplock_epoch, &purge_cache); 3070 3071 if (!CIFS_CACHE_WRITE(cinode) && CIFS_CACHE_READ(cinode) && 3072 cifs_has_mand_locks(cinode)) { 3073 cifs_dbg(FYI, "Reset oplock to None for inode=%p due to mand locks\n", 3074 inode); 3075 cinode->oplock = 0; 3076 } 3077 3078 if (inode && S_ISREG(inode->i_mode)) { 3079 if (CIFS_CACHE_READ(cinode)) 3080 break_lease(inode, O_RDONLY); 3081 else 3082 break_lease(inode, O_WRONLY); 3083 rc = filemap_fdatawrite(inode->i_mapping); 3084 if (!CIFS_CACHE_READ(cinode) || purge_cache) { 3085 rc = filemap_fdatawait(inode->i_mapping); 3086 mapping_set_error(inode->i_mapping, rc); 3087 cifs_zap_mapping(inode); 3088 } 3089 cifs_dbg(FYI, "Oplock flush inode %p rc %d\n", inode, rc); 3090 if (CIFS_CACHE_WRITE(cinode)) 3091 goto oplock_break_ack; 3092 } 3093 3094 rc = cifs_push_locks(cfile); 3095 if (rc) 3096 cifs_dbg(VFS, "Push locks rc = %d\n", rc); 3097 3098 oplock_break_ack: 3099 /* 3100 * When oplock break is received and there are no active 3101 * file handles but cached, then schedule deferred close immediately. 3102 * So, new open will not use cached handle. 3103 */ 3104 3105 if (!CIFS_CACHE_HANDLE(cinode) && !list_empty(&cinode->deferred_closes)) 3106 cifs_close_deferred_file(cinode); 3107 3108 persistent_fid = cfile->fid.persistent_fid; 3109 volatile_fid = cfile->fid.volatile_fid; 3110 net_fid = cfile->fid.netfid; 3111 oplock_break_cancelled = cfile->oplock_break_cancelled; 3112 3113 _cifsFileInfo_put(cfile, false /* do not wait for ourself */, false); 3114 /* 3115 * MS-SMB2 3.2.5.19.1 and 3.2.5.19.2 (and MS-CIFS 3.2.5.42) do not require 3116 * an acknowledgment to be sent when the file has already been closed. 3117 */ 3118 spin_lock(&cinode->open_file_lock); 3119 /* check list empty since can race with kill_sb calling tree disconnect */ 3120 if (!oplock_break_cancelled && !list_empty(&cinode->openFileList)) { 3121 spin_unlock(&cinode->open_file_lock); 3122 rc = server->ops->oplock_response(tcon, persistent_fid, 3123 volatile_fid, net_fid, cinode); 3124 cifs_dbg(FYI, "Oplock release rc = %d\n", rc); 3125 } else 3126 spin_unlock(&cinode->open_file_lock); 3127 3128 cifs_put_tlink(tlink); 3129 out: 3130 cifs_done_oplock_break(cinode); 3131 } 3132 3133 static int cifs_swap_activate(struct swap_info_struct *sis, 3134 struct file *swap_file, sector_t *span) 3135 { 3136 struct cifsFileInfo *cfile = swap_file->private_data; 3137 struct inode *inode = swap_file->f_mapping->host; 3138 unsigned long blocks; 3139 long long isize; 3140 3141 cifs_dbg(FYI, "swap activate\n"); 3142 3143 if (!swap_file->f_mapping->a_ops->swap_rw) 3144 /* Cannot support swap */ 3145 return -EINVAL; 3146 3147 spin_lock(&inode->i_lock); 3148 blocks = inode->i_blocks; 3149 isize = inode->i_size; 3150 spin_unlock(&inode->i_lock); 3151 if (blocks*512 < isize) { 3152 pr_warn("swap activate: swapfile has holes\n"); 3153 return -EINVAL; 3154 } 3155 *span = sis->pages; 3156 3157 pr_warn_once("Swap support over SMB3 is experimental\n"); 3158 3159 /* 3160 * TODO: consider adding ACL (or documenting how) to prevent other 3161 * users (on this or other systems) from reading it 3162 */ 3163 3164 3165 /* TODO: add sk_set_memalloc(inet) or similar */ 3166 3167 if (cfile) 3168 cfile->swapfile = true; 3169 /* 3170 * TODO: Since file already open, we can't open with DENY_ALL here 3171 * but we could add call to grab a byte range lock to prevent others 3172 * from reading or writing the file 3173 */ 3174 3175 sis->flags |= SWP_FS_OPS; 3176 return add_swap_extent(sis, 0, sis->max, 0); 3177 } 3178 3179 static void cifs_swap_deactivate(struct file *file) 3180 { 3181 struct cifsFileInfo *cfile = file->private_data; 3182 3183 cifs_dbg(FYI, "swap deactivate\n"); 3184 3185 /* TODO: undo sk_set_memalloc(inet) will eventually be needed */ 3186 3187 if (cfile) 3188 cfile->swapfile = false; 3189 3190 /* do we need to unpin (or unlock) the file */ 3191 } 3192 3193 /** 3194 * cifs_swap_rw - SMB3 address space operation for swap I/O 3195 * @iocb: target I/O control block 3196 * @iter: I/O buffer 3197 * 3198 * Perform IO to the swap-file. This is much like direct IO. 3199 */ 3200 static int cifs_swap_rw(struct kiocb *iocb, struct iov_iter *iter) 3201 { 3202 ssize_t ret; 3203 3204 if (iov_iter_rw(iter) == READ) 3205 ret = netfs_unbuffered_read_iter_locked(iocb, iter); 3206 else 3207 ret = netfs_unbuffered_write_iter_locked(iocb, iter, NULL); 3208 if (ret < 0) 3209 return ret; 3210 return 0; 3211 } 3212 3213 const struct address_space_operations cifs_addr_ops = { 3214 .read_folio = netfs_read_folio, 3215 .readahead = netfs_readahead, 3216 .writepages = netfs_writepages, 3217 .dirty_folio = netfs_dirty_folio, 3218 .release_folio = netfs_release_folio, 3219 .direct_IO = noop_direct_IO, 3220 .invalidate_folio = netfs_invalidate_folio, 3221 .migrate_folio = filemap_migrate_folio, 3222 /* 3223 * TODO: investigate and if useful we could add an is_dirty_writeback 3224 * helper if needed 3225 */ 3226 .swap_activate = cifs_swap_activate, 3227 .swap_deactivate = cifs_swap_deactivate, 3228 .swap_rw = cifs_swap_rw, 3229 }; 3230 3231 /* 3232 * cifs_readahead requires the server to support a buffer large enough to 3233 * contain the header plus one complete page of data. Otherwise, we need 3234 * to leave cifs_readahead out of the address space operations. 3235 */ 3236 const struct address_space_operations cifs_addr_ops_smallbuf = { 3237 .read_folio = netfs_read_folio, 3238 .writepages = netfs_writepages, 3239 .dirty_folio = netfs_dirty_folio, 3240 .release_folio = netfs_release_folio, 3241 .invalidate_folio = netfs_invalidate_folio, 3242 .migrate_folio = filemap_migrate_folio, 3243 }; 3244