xref: /linux/fs/smb/client/file.c (revision 60cb1da6ed4a62ec8331e25ad4be87115cd28feb)
1 // SPDX-License-Identifier: LGPL-2.1
2 /*
3  *
4  *   vfs operations that deal with files
5  *
6  *   Copyright (C) International Business Machines  Corp., 2002,2010
7  *   Author(s): Steve French (sfrench@us.ibm.com)
8  *              Jeremy Allison (jra@samba.org)
9  *
10  */
11 #include <linux/fs.h>
12 #include <linux/filelock.h>
13 #include <linux/backing-dev.h>
14 #include <linux/stat.h>
15 #include <linux/fcntl.h>
16 #include <linux/pagemap.h>
17 #include <linux/pagevec.h>
18 #include <linux/writeback.h>
19 #include <linux/task_io_accounting_ops.h>
20 #include <linux/delay.h>
21 #include <linux/mount.h>
22 #include <linux/slab.h>
23 #include <linux/swap.h>
24 #include <linux/mm.h>
25 #include <asm/div64.h>
26 #include "cifsfs.h"
27 #include "cifspdu.h"
28 #include "cifsglob.h"
29 #include "cifsproto.h"
30 #include "smb2proto.h"
31 #include "cifs_unicode.h"
32 #include "cifs_debug.h"
33 #include "cifs_fs_sb.h"
34 #include "fscache.h"
35 #include "smbdirect.h"
36 #include "fs_context.h"
37 #include "cifs_ioctl.h"
38 #include "cached_dir.h"
39 #include <trace/events/netfs.h>
40 
41 static int cifs_reopen_file(struct cifsFileInfo *cfile, bool can_flush);
42 
43 /*
44  * Prepare a subrequest to upload to the server.  We need to allocate credits
45  * so that we know the maximum amount of data that we can include in it.
46  */
47 static void cifs_prepare_write(struct netfs_io_subrequest *subreq)
48 {
49 	struct cifs_io_subrequest *wdata =
50 		container_of(subreq, struct cifs_io_subrequest, subreq);
51 	struct cifs_io_request *req = wdata->req;
52 	struct TCP_Server_Info *server;
53 	struct cifsFileInfo *open_file = req->cfile;
54 	size_t wsize = req->rreq.wsize;
55 	int rc;
56 
57 	if (!wdata->have_xid) {
58 		wdata->xid = get_xid();
59 		wdata->have_xid = true;
60 	}
61 
62 	server = cifs_pick_channel(tlink_tcon(open_file->tlink)->ses);
63 	wdata->server = server;
64 
65 retry:
66 	if (open_file->invalidHandle) {
67 		rc = cifs_reopen_file(open_file, false);
68 		if (rc < 0) {
69 			if (rc == -EAGAIN)
70 				goto retry;
71 			subreq->error = rc;
72 			return netfs_prepare_write_failed(subreq);
73 		}
74 	}
75 
76 	rc = server->ops->wait_mtu_credits(server, wsize, &wdata->subreq.max_len,
77 					   &wdata->credits);
78 	if (rc < 0) {
79 		subreq->error = rc;
80 		return netfs_prepare_write_failed(subreq);
81 	}
82 
83 	wdata->credits.rreq_debug_id = subreq->rreq->debug_id;
84 	wdata->credits.rreq_debug_index = subreq->debug_index;
85 	wdata->credits.in_flight_check = 1;
86 	trace_smb3_rw_credits(wdata->rreq->debug_id,
87 			      wdata->subreq.debug_index,
88 			      wdata->credits.value,
89 			      server->credits, server->in_flight,
90 			      wdata->credits.value,
91 			      cifs_trace_rw_credits_write_prepare);
92 
93 #ifdef CONFIG_CIFS_SMB_DIRECT
94 	if (server->smbd_conn)
95 		subreq->max_nr_segs = server->smbd_conn->max_frmr_depth;
96 #endif
97 }
98 
99 /*
100  * Issue a subrequest to upload to the server.
101  */
102 static void cifs_issue_write(struct netfs_io_subrequest *subreq)
103 {
104 	struct cifs_io_subrequest *wdata =
105 		container_of(subreq, struct cifs_io_subrequest, subreq);
106 	struct cifs_sb_info *sbi = CIFS_SB(subreq->rreq->inode->i_sb);
107 	int rc;
108 
109 	if (cifs_forced_shutdown(sbi)) {
110 		rc = -EIO;
111 		goto fail;
112 	}
113 
114 	rc = adjust_credits(wdata->server, wdata, cifs_trace_rw_credits_issue_write_adjust);
115 	if (rc)
116 		goto fail;
117 
118 	rc = -EAGAIN;
119 	if (wdata->req->cfile->invalidHandle)
120 		goto fail;
121 
122 	wdata->server->ops->async_writev(wdata);
123 out:
124 	return;
125 
126 fail:
127 	if (rc == -EAGAIN)
128 		trace_netfs_sreq(subreq, netfs_sreq_trace_retry);
129 	else
130 		trace_netfs_sreq(subreq, netfs_sreq_trace_fail);
131 	add_credits_and_wake_if(wdata->server, &wdata->credits, 0);
132 	cifs_write_subrequest_terminated(wdata, rc, false);
133 	goto out;
134 }
135 
136 static void cifs_netfs_invalidate_cache(struct netfs_io_request *wreq)
137 {
138 	cifs_invalidate_cache(wreq->inode, 0);
139 }
140 
141 /*
142  * Split the read up according to how many credits we can get for each piece.
143  * It's okay to sleep here if we need to wait for more credit to become
144  * available.
145  *
146  * We also choose the server and allocate an operation ID to be cleaned up
147  * later.
148  */
149 static bool cifs_clamp_length(struct netfs_io_subrequest *subreq)
150 {
151 	struct netfs_io_request *rreq = subreq->rreq;
152 	struct cifs_io_subrequest *rdata = container_of(subreq, struct cifs_io_subrequest, subreq);
153 	struct cifs_io_request *req = container_of(subreq->rreq, struct cifs_io_request, rreq);
154 	struct TCP_Server_Info *server = req->server;
155 	struct cifs_sb_info *cifs_sb = CIFS_SB(rreq->inode->i_sb);
156 	size_t rsize = 0;
157 	int rc;
158 
159 	rdata->xid = get_xid();
160 	rdata->have_xid = true;
161 	rdata->server = server;
162 
163 	if (cifs_sb->ctx->rsize == 0)
164 		cifs_sb->ctx->rsize =
165 			server->ops->negotiate_rsize(tlink_tcon(req->cfile->tlink),
166 						     cifs_sb->ctx);
167 
168 
169 	rc = server->ops->wait_mtu_credits(server, cifs_sb->ctx->rsize, &rsize,
170 					   &rdata->credits);
171 	if (rc) {
172 		subreq->error = rc;
173 		return false;
174 	}
175 
176 	rdata->credits.in_flight_check = 1;
177 	rdata->credits.rreq_debug_id = rreq->debug_id;
178 	rdata->credits.rreq_debug_index = subreq->debug_index;
179 
180 	trace_smb3_rw_credits(rdata->rreq->debug_id,
181 			      rdata->subreq.debug_index,
182 			      rdata->credits.value,
183 			      server->credits, server->in_flight, 0,
184 			      cifs_trace_rw_credits_read_submit);
185 
186 	subreq->len = min_t(size_t, subreq->len, rsize);
187 
188 #ifdef CONFIG_CIFS_SMB_DIRECT
189 	if (server->smbd_conn)
190 		subreq->max_nr_segs = server->smbd_conn->max_frmr_depth;
191 #endif
192 	return true;
193 }
194 
195 /*
196  * Issue a read operation on behalf of the netfs helper functions.  We're asked
197  * to make a read of a certain size at a point in the file.  We are permitted
198  * to only read a portion of that, but as long as we read something, the netfs
199  * helper will call us again so that we can issue another read.
200  */
201 static void cifs_req_issue_read(struct netfs_io_subrequest *subreq)
202 {
203 	struct netfs_io_request *rreq = subreq->rreq;
204 	struct cifs_io_subrequest *rdata = container_of(subreq, struct cifs_io_subrequest, subreq);
205 	struct cifs_io_request *req = container_of(subreq->rreq, struct cifs_io_request, rreq);
206 	int rc = 0;
207 
208 	cifs_dbg(FYI, "%s: op=%08x[%x] mapping=%p len=%zu/%zu\n",
209 		 __func__, rreq->debug_id, subreq->debug_index, rreq->mapping,
210 		 subreq->transferred, subreq->len);
211 
212 	if (req->cfile->invalidHandle) {
213 		do {
214 			rc = cifs_reopen_file(req->cfile, true);
215 		} while (rc == -EAGAIN);
216 		if (rc)
217 			goto out;
218 	}
219 
220 	if (subreq->rreq->origin != NETFS_DIO_READ)
221 		__set_bit(NETFS_SREQ_CLEAR_TAIL, &subreq->flags);
222 
223 	rc = rdata->server->ops->async_readv(rdata);
224 out:
225 	if (rc)
226 		netfs_subreq_terminated(subreq, rc, false);
227 }
228 
229 /*
230  * Writeback calls this when it finds a folio that needs uploading.  This isn't
231  * called if writeback only has copy-to-cache to deal with.
232  */
233 static void cifs_begin_writeback(struct netfs_io_request *wreq)
234 {
235 	struct cifs_io_request *req = container_of(wreq, struct cifs_io_request, rreq);
236 	int ret;
237 
238 	ret = cifs_get_writable_file(CIFS_I(wreq->inode), FIND_WR_ANY, &req->cfile);
239 	if (ret) {
240 		cifs_dbg(VFS, "No writable handle in writepages ret=%d\n", ret);
241 		return;
242 	}
243 
244 	wreq->io_streams[0].avail = true;
245 }
246 
247 /*
248  * Initialise a request.
249  */
250 static int cifs_init_request(struct netfs_io_request *rreq, struct file *file)
251 {
252 	struct cifs_io_request *req = container_of(rreq, struct cifs_io_request, rreq);
253 	struct cifs_sb_info *cifs_sb = CIFS_SB(rreq->inode->i_sb);
254 	struct cifsFileInfo *open_file = NULL;
255 
256 	rreq->rsize = cifs_sb->ctx->rsize;
257 	rreq->wsize = cifs_sb->ctx->wsize;
258 	req->pid = current->tgid; // Ummm...  This may be a workqueue
259 
260 	if (file) {
261 		open_file = file->private_data;
262 		rreq->netfs_priv = file->private_data;
263 		req->cfile = cifsFileInfo_get(open_file);
264 		req->server = cifs_pick_channel(tlink_tcon(req->cfile->tlink)->ses);
265 		if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_RWPIDFORWARD)
266 			req->pid = req->cfile->pid;
267 	} else if (rreq->origin != NETFS_WRITEBACK) {
268 		WARN_ON_ONCE(1);
269 		return -EIO;
270 	}
271 
272 	return 0;
273 }
274 
275 /*
276  * Completion of a request operation.
277  */
278 static void cifs_rreq_done(struct netfs_io_request *rreq)
279 {
280 	struct timespec64 atime, mtime;
281 	struct inode *inode = rreq->inode;
282 
283 	/* we do not want atime to be less than mtime, it broke some apps */
284 	atime = inode_set_atime_to_ts(inode, current_time(inode));
285 	mtime = inode_get_mtime(inode);
286 	if (timespec64_compare(&atime, &mtime))
287 		inode_set_atime_to_ts(inode, inode_get_mtime(inode));
288 }
289 
290 static void cifs_post_modify(struct inode *inode)
291 {
292 	/* Indication to update ctime and mtime as close is deferred */
293 	set_bit(CIFS_INO_MODIFIED_ATTR, &CIFS_I(inode)->flags);
294 }
295 
296 static void cifs_free_request(struct netfs_io_request *rreq)
297 {
298 	struct cifs_io_request *req = container_of(rreq, struct cifs_io_request, rreq);
299 
300 	if (req->cfile)
301 		cifsFileInfo_put(req->cfile);
302 }
303 
304 static void cifs_free_subrequest(struct netfs_io_subrequest *subreq)
305 {
306 	struct cifs_io_subrequest *rdata =
307 		container_of(subreq, struct cifs_io_subrequest, subreq);
308 	int rc = subreq->error;
309 
310 	if (rdata->subreq.source == NETFS_DOWNLOAD_FROM_SERVER) {
311 #ifdef CONFIG_CIFS_SMB_DIRECT
312 		if (rdata->mr) {
313 			smbd_deregister_mr(rdata->mr);
314 			rdata->mr = NULL;
315 		}
316 #endif
317 	}
318 
319 	if (rdata->credits.value != 0)
320 		trace_smb3_rw_credits(rdata->rreq->debug_id,
321 				      rdata->subreq.debug_index,
322 				      rdata->credits.value,
323 				      rdata->server ? rdata->server->credits : 0,
324 				      rdata->server ? rdata->server->in_flight : 0,
325 				      -rdata->credits.value,
326 				      cifs_trace_rw_credits_free_subreq);
327 
328 	add_credits_and_wake_if(rdata->server, &rdata->credits, 0);
329 	if (rdata->have_xid)
330 		free_xid(rdata->xid);
331 }
332 
333 const struct netfs_request_ops cifs_req_ops = {
334 	.request_pool		= &cifs_io_request_pool,
335 	.subrequest_pool	= &cifs_io_subrequest_pool,
336 	.init_request		= cifs_init_request,
337 	.free_request		= cifs_free_request,
338 	.free_subrequest	= cifs_free_subrequest,
339 	.clamp_length		= cifs_clamp_length,
340 	.issue_read		= cifs_req_issue_read,
341 	.done			= cifs_rreq_done,
342 	.post_modify		= cifs_post_modify,
343 	.begin_writeback	= cifs_begin_writeback,
344 	.prepare_write		= cifs_prepare_write,
345 	.issue_write		= cifs_issue_write,
346 	.invalidate_cache	= cifs_netfs_invalidate_cache,
347 };
348 
349 /*
350  * Mark as invalid, all open files on tree connections since they
351  * were closed when session to server was lost.
352  */
353 void
354 cifs_mark_open_files_invalid(struct cifs_tcon *tcon)
355 {
356 	struct cifsFileInfo *open_file = NULL;
357 	struct list_head *tmp;
358 	struct list_head *tmp1;
359 
360 	/* only send once per connect */
361 	spin_lock(&tcon->tc_lock);
362 	if (tcon->need_reconnect)
363 		tcon->status = TID_NEED_RECON;
364 
365 	if (tcon->status != TID_NEED_RECON) {
366 		spin_unlock(&tcon->tc_lock);
367 		return;
368 	}
369 	tcon->status = TID_IN_FILES_INVALIDATE;
370 	spin_unlock(&tcon->tc_lock);
371 
372 	/* list all files open on tree connection and mark them invalid */
373 	spin_lock(&tcon->open_file_lock);
374 	list_for_each_safe(tmp, tmp1, &tcon->openFileList) {
375 		open_file = list_entry(tmp, struct cifsFileInfo, tlist);
376 		open_file->invalidHandle = true;
377 		open_file->oplock_break_cancelled = true;
378 	}
379 	spin_unlock(&tcon->open_file_lock);
380 
381 	invalidate_all_cached_dirs(tcon);
382 	spin_lock(&tcon->tc_lock);
383 	if (tcon->status == TID_IN_FILES_INVALIDATE)
384 		tcon->status = TID_NEED_TCON;
385 	spin_unlock(&tcon->tc_lock);
386 
387 	/*
388 	 * BB Add call to invalidate_inodes(sb) for all superblocks mounted
389 	 * to this tcon.
390 	 */
391 }
392 
393 static inline int cifs_convert_flags(unsigned int flags, int rdwr_for_fscache)
394 {
395 	if ((flags & O_ACCMODE) == O_RDONLY)
396 		return GENERIC_READ;
397 	else if ((flags & O_ACCMODE) == O_WRONLY)
398 		return rdwr_for_fscache == 1 ? (GENERIC_READ | GENERIC_WRITE) : GENERIC_WRITE;
399 	else if ((flags & O_ACCMODE) == O_RDWR) {
400 		/* GENERIC_ALL is too much permission to request
401 		   can cause unnecessary access denied on create */
402 		/* return GENERIC_ALL; */
403 		return (GENERIC_READ | GENERIC_WRITE);
404 	}
405 
406 	return (READ_CONTROL | FILE_WRITE_ATTRIBUTES | FILE_READ_ATTRIBUTES |
407 		FILE_WRITE_EA | FILE_APPEND_DATA | FILE_WRITE_DATA |
408 		FILE_READ_DATA);
409 }
410 
411 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
412 static u32 cifs_posix_convert_flags(unsigned int flags)
413 {
414 	u32 posix_flags = 0;
415 
416 	if ((flags & O_ACCMODE) == O_RDONLY)
417 		posix_flags = SMB_O_RDONLY;
418 	else if ((flags & O_ACCMODE) == O_WRONLY)
419 		posix_flags = SMB_O_WRONLY;
420 	else if ((flags & O_ACCMODE) == O_RDWR)
421 		posix_flags = SMB_O_RDWR;
422 
423 	if (flags & O_CREAT) {
424 		posix_flags |= SMB_O_CREAT;
425 		if (flags & O_EXCL)
426 			posix_flags |= SMB_O_EXCL;
427 	} else if (flags & O_EXCL)
428 		cifs_dbg(FYI, "Application %s pid %d has incorrectly set O_EXCL flag but not O_CREAT on file open. Ignoring O_EXCL\n",
429 			 current->comm, current->tgid);
430 
431 	if (flags & O_TRUNC)
432 		posix_flags |= SMB_O_TRUNC;
433 	/* be safe and imply O_SYNC for O_DSYNC */
434 	if (flags & O_DSYNC)
435 		posix_flags |= SMB_O_SYNC;
436 	if (flags & O_DIRECTORY)
437 		posix_flags |= SMB_O_DIRECTORY;
438 	if (flags & O_NOFOLLOW)
439 		posix_flags |= SMB_O_NOFOLLOW;
440 	if (flags & O_DIRECT)
441 		posix_flags |= SMB_O_DIRECT;
442 
443 	return posix_flags;
444 }
445 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */
446 
447 static inline int cifs_get_disposition(unsigned int flags)
448 {
449 	if ((flags & (O_CREAT | O_EXCL)) == (O_CREAT | O_EXCL))
450 		return FILE_CREATE;
451 	else if ((flags & (O_CREAT | O_TRUNC)) == (O_CREAT | O_TRUNC))
452 		return FILE_OVERWRITE_IF;
453 	else if ((flags & O_CREAT) == O_CREAT)
454 		return FILE_OPEN_IF;
455 	else if ((flags & O_TRUNC) == O_TRUNC)
456 		return FILE_OVERWRITE;
457 	else
458 		return FILE_OPEN;
459 }
460 
461 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
462 int cifs_posix_open(const char *full_path, struct inode **pinode,
463 			struct super_block *sb, int mode, unsigned int f_flags,
464 			__u32 *poplock, __u16 *pnetfid, unsigned int xid)
465 {
466 	int rc;
467 	FILE_UNIX_BASIC_INFO *presp_data;
468 	__u32 posix_flags = 0;
469 	struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
470 	struct cifs_fattr fattr;
471 	struct tcon_link *tlink;
472 	struct cifs_tcon *tcon;
473 
474 	cifs_dbg(FYI, "posix open %s\n", full_path);
475 
476 	presp_data = kzalloc(sizeof(FILE_UNIX_BASIC_INFO), GFP_KERNEL);
477 	if (presp_data == NULL)
478 		return -ENOMEM;
479 
480 	tlink = cifs_sb_tlink(cifs_sb);
481 	if (IS_ERR(tlink)) {
482 		rc = PTR_ERR(tlink);
483 		goto posix_open_ret;
484 	}
485 
486 	tcon = tlink_tcon(tlink);
487 	mode &= ~current_umask();
488 
489 	posix_flags = cifs_posix_convert_flags(f_flags);
490 	rc = CIFSPOSIXCreate(xid, tcon, posix_flags, mode, pnetfid, presp_data,
491 			     poplock, full_path, cifs_sb->local_nls,
492 			     cifs_remap(cifs_sb));
493 	cifs_put_tlink(tlink);
494 
495 	if (rc)
496 		goto posix_open_ret;
497 
498 	if (presp_data->Type == cpu_to_le32(-1))
499 		goto posix_open_ret; /* open ok, caller does qpathinfo */
500 
501 	if (!pinode)
502 		goto posix_open_ret; /* caller does not need info */
503 
504 	cifs_unix_basic_to_fattr(&fattr, presp_data, cifs_sb);
505 
506 	/* get new inode and set it up */
507 	if (*pinode == NULL) {
508 		cifs_fill_uniqueid(sb, &fattr);
509 		*pinode = cifs_iget(sb, &fattr);
510 		if (!*pinode) {
511 			rc = -ENOMEM;
512 			goto posix_open_ret;
513 		}
514 	} else {
515 		cifs_revalidate_mapping(*pinode);
516 		rc = cifs_fattr_to_inode(*pinode, &fattr, false);
517 	}
518 
519 posix_open_ret:
520 	kfree(presp_data);
521 	return rc;
522 }
523 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */
524 
525 static int cifs_nt_open(const char *full_path, struct inode *inode, struct cifs_sb_info *cifs_sb,
526 			struct cifs_tcon *tcon, unsigned int f_flags, __u32 *oplock,
527 			struct cifs_fid *fid, unsigned int xid, struct cifs_open_info_data *buf)
528 {
529 	int rc;
530 	int desired_access;
531 	int disposition;
532 	int create_options = CREATE_NOT_DIR;
533 	struct TCP_Server_Info *server = tcon->ses->server;
534 	struct cifs_open_parms oparms;
535 	int rdwr_for_fscache = 0;
536 
537 	if (!server->ops->open)
538 		return -ENOSYS;
539 
540 	/* If we're caching, we need to be able to fill in around partial writes. */
541 	if (cifs_fscache_enabled(inode) && (f_flags & O_ACCMODE) == O_WRONLY)
542 		rdwr_for_fscache = 1;
543 
544 	desired_access = cifs_convert_flags(f_flags, rdwr_for_fscache);
545 
546 /*********************************************************************
547  *  open flag mapping table:
548  *
549  *	POSIX Flag            CIFS Disposition
550  *	----------            ----------------
551  *	O_CREAT               FILE_OPEN_IF
552  *	O_CREAT | O_EXCL      FILE_CREATE
553  *	O_CREAT | O_TRUNC     FILE_OVERWRITE_IF
554  *	O_TRUNC               FILE_OVERWRITE
555  *	none of the above     FILE_OPEN
556  *
557  *	Note that there is not a direct match between disposition
558  *	FILE_SUPERSEDE (ie create whether or not file exists although
559  *	O_CREAT | O_TRUNC is similar but truncates the existing
560  *	file rather than creating a new file as FILE_SUPERSEDE does
561  *	(which uses the attributes / metadata passed in on open call)
562  *?
563  *?  O_SYNC is a reasonable match to CIFS writethrough flag
564  *?  and the read write flags match reasonably.  O_LARGEFILE
565  *?  is irrelevant because largefile support is always used
566  *?  by this client. Flags O_APPEND, O_DIRECT, O_DIRECTORY,
567  *	 O_FASYNC, O_NOFOLLOW, O_NONBLOCK need further investigation
568  *********************************************************************/
569 
570 	disposition = cifs_get_disposition(f_flags);
571 
572 	/* BB pass O_SYNC flag through on file attributes .. BB */
573 
574 	/* O_SYNC also has bit for O_DSYNC so following check picks up either */
575 	if (f_flags & O_SYNC)
576 		create_options |= CREATE_WRITE_THROUGH;
577 
578 	if (f_flags & O_DIRECT)
579 		create_options |= CREATE_NO_BUFFER;
580 
581 retry_open:
582 	oparms = (struct cifs_open_parms) {
583 		.tcon = tcon,
584 		.cifs_sb = cifs_sb,
585 		.desired_access = desired_access,
586 		.create_options = cifs_create_options(cifs_sb, create_options),
587 		.disposition = disposition,
588 		.path = full_path,
589 		.fid = fid,
590 	};
591 
592 	rc = server->ops->open(xid, &oparms, oplock, buf);
593 	if (rc) {
594 		if (rc == -EACCES && rdwr_for_fscache == 1) {
595 			desired_access = cifs_convert_flags(f_flags, 0);
596 			rdwr_for_fscache = 2;
597 			goto retry_open;
598 		}
599 		return rc;
600 	}
601 	if (rdwr_for_fscache == 2)
602 		cifs_invalidate_cache(inode, FSCACHE_INVAL_DIO_WRITE);
603 
604 	/* TODO: Add support for calling posix query info but with passing in fid */
605 	if (tcon->unix_ext)
606 		rc = cifs_get_inode_info_unix(&inode, full_path, inode->i_sb,
607 					      xid);
608 	else
609 		rc = cifs_get_inode_info(&inode, full_path, buf, inode->i_sb,
610 					 xid, fid);
611 
612 	if (rc) {
613 		server->ops->close(xid, tcon, fid);
614 		if (rc == -ESTALE)
615 			rc = -EOPENSTALE;
616 	}
617 
618 	return rc;
619 }
620 
621 static bool
622 cifs_has_mand_locks(struct cifsInodeInfo *cinode)
623 {
624 	struct cifs_fid_locks *cur;
625 	bool has_locks = false;
626 
627 	down_read(&cinode->lock_sem);
628 	list_for_each_entry(cur, &cinode->llist, llist) {
629 		if (!list_empty(&cur->locks)) {
630 			has_locks = true;
631 			break;
632 		}
633 	}
634 	up_read(&cinode->lock_sem);
635 	return has_locks;
636 }
637 
638 void
639 cifs_down_write(struct rw_semaphore *sem)
640 {
641 	while (!down_write_trylock(sem))
642 		msleep(10);
643 }
644 
645 static void cifsFileInfo_put_work(struct work_struct *work);
646 void serverclose_work(struct work_struct *work);
647 
648 struct cifsFileInfo *cifs_new_fileinfo(struct cifs_fid *fid, struct file *file,
649 				       struct tcon_link *tlink, __u32 oplock,
650 				       const char *symlink_target)
651 {
652 	struct dentry *dentry = file_dentry(file);
653 	struct inode *inode = d_inode(dentry);
654 	struct cifsInodeInfo *cinode = CIFS_I(inode);
655 	struct cifsFileInfo *cfile;
656 	struct cifs_fid_locks *fdlocks;
657 	struct cifs_tcon *tcon = tlink_tcon(tlink);
658 	struct TCP_Server_Info *server = tcon->ses->server;
659 
660 	cfile = kzalloc(sizeof(struct cifsFileInfo), GFP_KERNEL);
661 	if (cfile == NULL)
662 		return cfile;
663 
664 	fdlocks = kzalloc(sizeof(struct cifs_fid_locks), GFP_KERNEL);
665 	if (!fdlocks) {
666 		kfree(cfile);
667 		return NULL;
668 	}
669 
670 	if (symlink_target) {
671 		cfile->symlink_target = kstrdup(symlink_target, GFP_KERNEL);
672 		if (!cfile->symlink_target) {
673 			kfree(fdlocks);
674 			kfree(cfile);
675 			return NULL;
676 		}
677 	}
678 
679 	INIT_LIST_HEAD(&fdlocks->locks);
680 	fdlocks->cfile = cfile;
681 	cfile->llist = fdlocks;
682 
683 	cfile->count = 1;
684 	cfile->pid = current->tgid;
685 	cfile->uid = current_fsuid();
686 	cfile->dentry = dget(dentry);
687 	cfile->f_flags = file->f_flags;
688 	cfile->invalidHandle = false;
689 	cfile->deferred_close_scheduled = false;
690 	cfile->tlink = cifs_get_tlink(tlink);
691 	INIT_WORK(&cfile->oplock_break, cifs_oplock_break);
692 	INIT_WORK(&cfile->put, cifsFileInfo_put_work);
693 	INIT_WORK(&cfile->serverclose, serverclose_work);
694 	INIT_DELAYED_WORK(&cfile->deferred, smb2_deferred_work_close);
695 	mutex_init(&cfile->fh_mutex);
696 	spin_lock_init(&cfile->file_info_lock);
697 
698 	cifs_sb_active(inode->i_sb);
699 
700 	/*
701 	 * If the server returned a read oplock and we have mandatory brlocks,
702 	 * set oplock level to None.
703 	 */
704 	if (server->ops->is_read_op(oplock) && cifs_has_mand_locks(cinode)) {
705 		cifs_dbg(FYI, "Reset oplock val from read to None due to mand locks\n");
706 		oplock = 0;
707 	}
708 
709 	cifs_down_write(&cinode->lock_sem);
710 	list_add(&fdlocks->llist, &cinode->llist);
711 	up_write(&cinode->lock_sem);
712 
713 	spin_lock(&tcon->open_file_lock);
714 	if (fid->pending_open->oplock != CIFS_OPLOCK_NO_CHANGE && oplock)
715 		oplock = fid->pending_open->oplock;
716 	list_del(&fid->pending_open->olist);
717 
718 	fid->purge_cache = false;
719 	server->ops->set_fid(cfile, fid, oplock);
720 
721 	list_add(&cfile->tlist, &tcon->openFileList);
722 	atomic_inc(&tcon->num_local_opens);
723 
724 	/* if readable file instance put first in list*/
725 	spin_lock(&cinode->open_file_lock);
726 	if (file->f_mode & FMODE_READ)
727 		list_add(&cfile->flist, &cinode->openFileList);
728 	else
729 		list_add_tail(&cfile->flist, &cinode->openFileList);
730 	spin_unlock(&cinode->open_file_lock);
731 	spin_unlock(&tcon->open_file_lock);
732 
733 	if (fid->purge_cache)
734 		cifs_zap_mapping(inode);
735 
736 	file->private_data = cfile;
737 	return cfile;
738 }
739 
740 struct cifsFileInfo *
741 cifsFileInfo_get(struct cifsFileInfo *cifs_file)
742 {
743 	spin_lock(&cifs_file->file_info_lock);
744 	cifsFileInfo_get_locked(cifs_file);
745 	spin_unlock(&cifs_file->file_info_lock);
746 	return cifs_file;
747 }
748 
749 static void cifsFileInfo_put_final(struct cifsFileInfo *cifs_file)
750 {
751 	struct inode *inode = d_inode(cifs_file->dentry);
752 	struct cifsInodeInfo *cifsi = CIFS_I(inode);
753 	struct cifsLockInfo *li, *tmp;
754 	struct super_block *sb = inode->i_sb;
755 
756 	/*
757 	 * Delete any outstanding lock records. We'll lose them when the file
758 	 * is closed anyway.
759 	 */
760 	cifs_down_write(&cifsi->lock_sem);
761 	list_for_each_entry_safe(li, tmp, &cifs_file->llist->locks, llist) {
762 		list_del(&li->llist);
763 		cifs_del_lock_waiters(li);
764 		kfree(li);
765 	}
766 	list_del(&cifs_file->llist->llist);
767 	kfree(cifs_file->llist);
768 	up_write(&cifsi->lock_sem);
769 
770 	cifs_put_tlink(cifs_file->tlink);
771 	dput(cifs_file->dentry);
772 	cifs_sb_deactive(sb);
773 	kfree(cifs_file->symlink_target);
774 	kfree(cifs_file);
775 }
776 
777 static void cifsFileInfo_put_work(struct work_struct *work)
778 {
779 	struct cifsFileInfo *cifs_file = container_of(work,
780 			struct cifsFileInfo, put);
781 
782 	cifsFileInfo_put_final(cifs_file);
783 }
784 
785 void serverclose_work(struct work_struct *work)
786 {
787 	struct cifsFileInfo *cifs_file = container_of(work,
788 			struct cifsFileInfo, serverclose);
789 
790 	struct cifs_tcon *tcon = tlink_tcon(cifs_file->tlink);
791 
792 	struct TCP_Server_Info *server = tcon->ses->server;
793 	int rc = 0;
794 	int retries = 0;
795 	int MAX_RETRIES = 4;
796 
797 	do {
798 		if (server->ops->close_getattr)
799 			rc = server->ops->close_getattr(0, tcon, cifs_file);
800 		else if (server->ops->close)
801 			rc = server->ops->close(0, tcon, &cifs_file->fid);
802 
803 		if (rc == -EBUSY || rc == -EAGAIN) {
804 			retries++;
805 			msleep(250);
806 		}
807 	} while ((rc == -EBUSY || rc == -EAGAIN) && (retries < MAX_RETRIES)
808 	);
809 
810 	if (retries == MAX_RETRIES)
811 		pr_warn("Serverclose failed %d times, giving up\n", MAX_RETRIES);
812 
813 	if (cifs_file->offload)
814 		queue_work(fileinfo_put_wq, &cifs_file->put);
815 	else
816 		cifsFileInfo_put_final(cifs_file);
817 }
818 
819 /**
820  * cifsFileInfo_put - release a reference of file priv data
821  *
822  * Always potentially wait for oplock handler. See _cifsFileInfo_put().
823  *
824  * @cifs_file:	cifs/smb3 specific info (eg refcounts) for an open file
825  */
826 void cifsFileInfo_put(struct cifsFileInfo *cifs_file)
827 {
828 	_cifsFileInfo_put(cifs_file, true, true);
829 }
830 
831 /**
832  * _cifsFileInfo_put - release a reference of file priv data
833  *
834  * This may involve closing the filehandle @cifs_file out on the
835  * server. Must be called without holding tcon->open_file_lock,
836  * cinode->open_file_lock and cifs_file->file_info_lock.
837  *
838  * If @wait_for_oplock_handler is true and we are releasing the last
839  * reference, wait for any running oplock break handler of the file
840  * and cancel any pending one.
841  *
842  * @cifs_file:	cifs/smb3 specific info (eg refcounts) for an open file
843  * @wait_oplock_handler: must be false if called from oplock_break_handler
844  * @offload:	not offloaded on close and oplock breaks
845  *
846  */
847 void _cifsFileInfo_put(struct cifsFileInfo *cifs_file,
848 		       bool wait_oplock_handler, bool offload)
849 {
850 	struct inode *inode = d_inode(cifs_file->dentry);
851 	struct cifs_tcon *tcon = tlink_tcon(cifs_file->tlink);
852 	struct TCP_Server_Info *server = tcon->ses->server;
853 	struct cifsInodeInfo *cifsi = CIFS_I(inode);
854 	struct super_block *sb = inode->i_sb;
855 	struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
856 	struct cifs_fid fid = {};
857 	struct cifs_pending_open open;
858 	bool oplock_break_cancelled;
859 	bool serverclose_offloaded = false;
860 
861 	spin_lock(&tcon->open_file_lock);
862 	spin_lock(&cifsi->open_file_lock);
863 	spin_lock(&cifs_file->file_info_lock);
864 
865 	cifs_file->offload = offload;
866 	if (--cifs_file->count > 0) {
867 		spin_unlock(&cifs_file->file_info_lock);
868 		spin_unlock(&cifsi->open_file_lock);
869 		spin_unlock(&tcon->open_file_lock);
870 		return;
871 	}
872 	spin_unlock(&cifs_file->file_info_lock);
873 
874 	if (server->ops->get_lease_key)
875 		server->ops->get_lease_key(inode, &fid);
876 
877 	/* store open in pending opens to make sure we don't miss lease break */
878 	cifs_add_pending_open_locked(&fid, cifs_file->tlink, &open);
879 
880 	/* remove it from the lists */
881 	list_del(&cifs_file->flist);
882 	list_del(&cifs_file->tlist);
883 	atomic_dec(&tcon->num_local_opens);
884 
885 	if (list_empty(&cifsi->openFileList)) {
886 		cifs_dbg(FYI, "closing last open instance for inode %p\n",
887 			 d_inode(cifs_file->dentry));
888 		/*
889 		 * In strict cache mode we need invalidate mapping on the last
890 		 * close  because it may cause a error when we open this file
891 		 * again and get at least level II oplock.
892 		 */
893 		if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_STRICT_IO)
894 			set_bit(CIFS_INO_INVALID_MAPPING, &cifsi->flags);
895 		cifs_set_oplock_level(cifsi, 0);
896 	}
897 
898 	spin_unlock(&cifsi->open_file_lock);
899 	spin_unlock(&tcon->open_file_lock);
900 
901 	oplock_break_cancelled = wait_oplock_handler ?
902 		cancel_work_sync(&cifs_file->oplock_break) : false;
903 
904 	if (!tcon->need_reconnect && !cifs_file->invalidHandle) {
905 		struct TCP_Server_Info *server = tcon->ses->server;
906 		unsigned int xid;
907 		int rc = 0;
908 
909 		xid = get_xid();
910 		if (server->ops->close_getattr)
911 			rc = server->ops->close_getattr(xid, tcon, cifs_file);
912 		else if (server->ops->close)
913 			rc = server->ops->close(xid, tcon, &cifs_file->fid);
914 		_free_xid(xid);
915 
916 		if (rc == -EBUSY || rc == -EAGAIN) {
917 			// Server close failed, hence offloading it as an async op
918 			queue_work(serverclose_wq, &cifs_file->serverclose);
919 			serverclose_offloaded = true;
920 		}
921 	}
922 
923 	if (oplock_break_cancelled)
924 		cifs_done_oplock_break(cifsi);
925 
926 	cifs_del_pending_open(&open);
927 
928 	// if serverclose has been offloaded to wq (on failure), it will
929 	// handle offloading put as well. If serverclose not offloaded,
930 	// we need to handle offloading put here.
931 	if (!serverclose_offloaded) {
932 		if (offload)
933 			queue_work(fileinfo_put_wq, &cifs_file->put);
934 		else
935 			cifsFileInfo_put_final(cifs_file);
936 	}
937 }
938 
939 int cifs_open(struct inode *inode, struct file *file)
940 
941 {
942 	int rc = -EACCES;
943 	unsigned int xid;
944 	__u32 oplock;
945 	struct cifs_sb_info *cifs_sb;
946 	struct TCP_Server_Info *server;
947 	struct cifs_tcon *tcon;
948 	struct tcon_link *tlink;
949 	struct cifsFileInfo *cfile = NULL;
950 	void *page;
951 	const char *full_path;
952 	bool posix_open_ok = false;
953 	struct cifs_fid fid = {};
954 	struct cifs_pending_open open;
955 	struct cifs_open_info_data data = {};
956 
957 	xid = get_xid();
958 
959 	cifs_sb = CIFS_SB(inode->i_sb);
960 	if (unlikely(cifs_forced_shutdown(cifs_sb))) {
961 		free_xid(xid);
962 		return -EIO;
963 	}
964 
965 	tlink = cifs_sb_tlink(cifs_sb);
966 	if (IS_ERR(tlink)) {
967 		free_xid(xid);
968 		return PTR_ERR(tlink);
969 	}
970 	tcon = tlink_tcon(tlink);
971 	server = tcon->ses->server;
972 
973 	page = alloc_dentry_path();
974 	full_path = build_path_from_dentry(file_dentry(file), page);
975 	if (IS_ERR(full_path)) {
976 		rc = PTR_ERR(full_path);
977 		goto out;
978 	}
979 
980 	cifs_dbg(FYI, "inode = 0x%p file flags are 0x%x for %s\n",
981 		 inode, file->f_flags, full_path);
982 
983 	if (file->f_flags & O_DIRECT &&
984 	    cifs_sb->mnt_cifs_flags & CIFS_MOUNT_STRICT_IO) {
985 		if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NO_BRL)
986 			file->f_op = &cifs_file_direct_nobrl_ops;
987 		else
988 			file->f_op = &cifs_file_direct_ops;
989 	}
990 
991 	/* Get the cached handle as SMB2 close is deferred */
992 	rc = cifs_get_readable_path(tcon, full_path, &cfile);
993 	if (rc == 0) {
994 		if (file->f_flags == cfile->f_flags) {
995 			file->private_data = cfile;
996 			spin_lock(&CIFS_I(inode)->deferred_lock);
997 			cifs_del_deferred_close(cfile);
998 			spin_unlock(&CIFS_I(inode)->deferred_lock);
999 			goto use_cache;
1000 		} else {
1001 			_cifsFileInfo_put(cfile, true, false);
1002 		}
1003 	}
1004 
1005 	if (server->oplocks)
1006 		oplock = REQ_OPLOCK;
1007 	else
1008 		oplock = 0;
1009 
1010 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
1011 	if (!tcon->broken_posix_open && tcon->unix_ext &&
1012 	    cap_unix(tcon->ses) && (CIFS_UNIX_POSIX_PATH_OPS_CAP &
1013 				le64_to_cpu(tcon->fsUnixInfo.Capability))) {
1014 		/* can not refresh inode info since size could be stale */
1015 		rc = cifs_posix_open(full_path, &inode, inode->i_sb,
1016 				cifs_sb->ctx->file_mode /* ignored */,
1017 				file->f_flags, &oplock, &fid.netfid, xid);
1018 		if (rc == 0) {
1019 			cifs_dbg(FYI, "posix open succeeded\n");
1020 			posix_open_ok = true;
1021 		} else if ((rc == -EINVAL) || (rc == -EOPNOTSUPP)) {
1022 			if (tcon->ses->serverNOS)
1023 				cifs_dbg(VFS, "server %s of type %s returned unexpected error on SMB posix open, disabling posix open support. Check if server update available.\n",
1024 					 tcon->ses->ip_addr,
1025 					 tcon->ses->serverNOS);
1026 			tcon->broken_posix_open = true;
1027 		} else if ((rc != -EIO) && (rc != -EREMOTE) &&
1028 			 (rc != -EOPNOTSUPP)) /* path not found or net err */
1029 			goto out;
1030 		/*
1031 		 * Else fallthrough to retry open the old way on network i/o
1032 		 * or DFS errors.
1033 		 */
1034 	}
1035 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */
1036 
1037 	if (server->ops->get_lease_key)
1038 		server->ops->get_lease_key(inode, &fid);
1039 
1040 	cifs_add_pending_open(&fid, tlink, &open);
1041 
1042 	if (!posix_open_ok) {
1043 		if (server->ops->get_lease_key)
1044 			server->ops->get_lease_key(inode, &fid);
1045 
1046 		rc = cifs_nt_open(full_path, inode, cifs_sb, tcon, file->f_flags, &oplock, &fid,
1047 				  xid, &data);
1048 		if (rc) {
1049 			cifs_del_pending_open(&open);
1050 			goto out;
1051 		}
1052 	}
1053 
1054 	cfile = cifs_new_fileinfo(&fid, file, tlink, oplock, data.symlink_target);
1055 	if (cfile == NULL) {
1056 		if (server->ops->close)
1057 			server->ops->close(xid, tcon, &fid);
1058 		cifs_del_pending_open(&open);
1059 		rc = -ENOMEM;
1060 		goto out;
1061 	}
1062 
1063 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
1064 	if ((oplock & CIFS_CREATE_ACTION) && !posix_open_ok && tcon->unix_ext) {
1065 		/*
1066 		 * Time to set mode which we can not set earlier due to
1067 		 * problems creating new read-only files.
1068 		 */
1069 		struct cifs_unix_set_info_args args = {
1070 			.mode	= inode->i_mode,
1071 			.uid	= INVALID_UID, /* no change */
1072 			.gid	= INVALID_GID, /* no change */
1073 			.ctime	= NO_CHANGE_64,
1074 			.atime	= NO_CHANGE_64,
1075 			.mtime	= NO_CHANGE_64,
1076 			.device	= 0,
1077 		};
1078 		CIFSSMBUnixSetFileInfo(xid, tcon, &args, fid.netfid,
1079 				       cfile->pid);
1080 	}
1081 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */
1082 
1083 use_cache:
1084 	fscache_use_cookie(cifs_inode_cookie(file_inode(file)),
1085 			   file->f_mode & FMODE_WRITE);
1086 	if (!(file->f_flags & O_DIRECT))
1087 		goto out;
1088 	if ((file->f_flags & (O_ACCMODE | O_APPEND)) == O_RDONLY)
1089 		goto out;
1090 	cifs_invalidate_cache(file_inode(file), FSCACHE_INVAL_DIO_WRITE);
1091 
1092 out:
1093 	free_dentry_path(page);
1094 	free_xid(xid);
1095 	cifs_put_tlink(tlink);
1096 	cifs_free_open_info(&data);
1097 	return rc;
1098 }
1099 
1100 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
1101 static int cifs_push_posix_locks(struct cifsFileInfo *cfile);
1102 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */
1103 
1104 /*
1105  * Try to reacquire byte range locks that were released when session
1106  * to server was lost.
1107  */
1108 static int
1109 cifs_relock_file(struct cifsFileInfo *cfile)
1110 {
1111 	struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry));
1112 	struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
1113 	int rc = 0;
1114 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
1115 	struct cifs_sb_info *cifs_sb = CIFS_SB(cfile->dentry->d_sb);
1116 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */
1117 
1118 	down_read_nested(&cinode->lock_sem, SINGLE_DEPTH_NESTING);
1119 	if (cinode->can_cache_brlcks) {
1120 		/* can cache locks - no need to relock */
1121 		up_read(&cinode->lock_sem);
1122 		return rc;
1123 	}
1124 
1125 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
1126 	if (cap_unix(tcon->ses) &&
1127 	    (CIFS_UNIX_FCNTL_CAP & le64_to_cpu(tcon->fsUnixInfo.Capability)) &&
1128 	    ((cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOPOSIXBRL) == 0))
1129 		rc = cifs_push_posix_locks(cfile);
1130 	else
1131 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */
1132 		rc = tcon->ses->server->ops->push_mand_locks(cfile);
1133 
1134 	up_read(&cinode->lock_sem);
1135 	return rc;
1136 }
1137 
1138 static int
1139 cifs_reopen_file(struct cifsFileInfo *cfile, bool can_flush)
1140 {
1141 	int rc = -EACCES;
1142 	unsigned int xid;
1143 	__u32 oplock;
1144 	struct cifs_sb_info *cifs_sb;
1145 	struct cifs_tcon *tcon;
1146 	struct TCP_Server_Info *server;
1147 	struct cifsInodeInfo *cinode;
1148 	struct inode *inode;
1149 	void *page;
1150 	const char *full_path;
1151 	int desired_access;
1152 	int disposition = FILE_OPEN;
1153 	int create_options = CREATE_NOT_DIR;
1154 	struct cifs_open_parms oparms;
1155 	int rdwr_for_fscache = 0;
1156 
1157 	xid = get_xid();
1158 	mutex_lock(&cfile->fh_mutex);
1159 	if (!cfile->invalidHandle) {
1160 		mutex_unlock(&cfile->fh_mutex);
1161 		free_xid(xid);
1162 		return 0;
1163 	}
1164 
1165 	inode = d_inode(cfile->dentry);
1166 	cifs_sb = CIFS_SB(inode->i_sb);
1167 	tcon = tlink_tcon(cfile->tlink);
1168 	server = tcon->ses->server;
1169 
1170 	/*
1171 	 * Can not grab rename sem here because various ops, including those
1172 	 * that already have the rename sem can end up causing writepage to get
1173 	 * called and if the server was down that means we end up here, and we
1174 	 * can never tell if the caller already has the rename_sem.
1175 	 */
1176 	page = alloc_dentry_path();
1177 	full_path = build_path_from_dentry(cfile->dentry, page);
1178 	if (IS_ERR(full_path)) {
1179 		mutex_unlock(&cfile->fh_mutex);
1180 		free_dentry_path(page);
1181 		free_xid(xid);
1182 		return PTR_ERR(full_path);
1183 	}
1184 
1185 	cifs_dbg(FYI, "inode = 0x%p file flags 0x%x for %s\n",
1186 		 inode, cfile->f_flags, full_path);
1187 
1188 	if (tcon->ses->server->oplocks)
1189 		oplock = REQ_OPLOCK;
1190 	else
1191 		oplock = 0;
1192 
1193 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
1194 	if (tcon->unix_ext && cap_unix(tcon->ses) &&
1195 	    (CIFS_UNIX_POSIX_PATH_OPS_CAP &
1196 				le64_to_cpu(tcon->fsUnixInfo.Capability))) {
1197 		/*
1198 		 * O_CREAT, O_EXCL and O_TRUNC already had their effect on the
1199 		 * original open. Must mask them off for a reopen.
1200 		 */
1201 		unsigned int oflags = cfile->f_flags &
1202 						~(O_CREAT | O_EXCL | O_TRUNC);
1203 
1204 		rc = cifs_posix_open(full_path, NULL, inode->i_sb,
1205 				     cifs_sb->ctx->file_mode /* ignored */,
1206 				     oflags, &oplock, &cfile->fid.netfid, xid);
1207 		if (rc == 0) {
1208 			cifs_dbg(FYI, "posix reopen succeeded\n");
1209 			oparms.reconnect = true;
1210 			goto reopen_success;
1211 		}
1212 		/*
1213 		 * fallthrough to retry open the old way on errors, especially
1214 		 * in the reconnect path it is important to retry hard
1215 		 */
1216 	}
1217 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */
1218 
1219 	/* If we're caching, we need to be able to fill in around partial writes. */
1220 	if (cifs_fscache_enabled(inode) && (cfile->f_flags & O_ACCMODE) == O_WRONLY)
1221 		rdwr_for_fscache = 1;
1222 
1223 	desired_access = cifs_convert_flags(cfile->f_flags, rdwr_for_fscache);
1224 
1225 	/* O_SYNC also has bit for O_DSYNC so following check picks up either */
1226 	if (cfile->f_flags & O_SYNC)
1227 		create_options |= CREATE_WRITE_THROUGH;
1228 
1229 	if (cfile->f_flags & O_DIRECT)
1230 		create_options |= CREATE_NO_BUFFER;
1231 
1232 	if (server->ops->get_lease_key)
1233 		server->ops->get_lease_key(inode, &cfile->fid);
1234 
1235 retry_open:
1236 	oparms = (struct cifs_open_parms) {
1237 		.tcon = tcon,
1238 		.cifs_sb = cifs_sb,
1239 		.desired_access = desired_access,
1240 		.create_options = cifs_create_options(cifs_sb, create_options),
1241 		.disposition = disposition,
1242 		.path = full_path,
1243 		.fid = &cfile->fid,
1244 		.reconnect = true,
1245 	};
1246 
1247 	/*
1248 	 * Can not refresh inode by passing in file_info buf to be returned by
1249 	 * ops->open and then calling get_inode_info with returned buf since
1250 	 * file might have write behind data that needs to be flushed and server
1251 	 * version of file size can be stale. If we knew for sure that inode was
1252 	 * not dirty locally we could do this.
1253 	 */
1254 	rc = server->ops->open(xid, &oparms, &oplock, NULL);
1255 	if (rc == -ENOENT && oparms.reconnect == false) {
1256 		/* durable handle timeout is expired - open the file again */
1257 		rc = server->ops->open(xid, &oparms, &oplock, NULL);
1258 		/* indicate that we need to relock the file */
1259 		oparms.reconnect = true;
1260 	}
1261 	if (rc == -EACCES && rdwr_for_fscache == 1) {
1262 		desired_access = cifs_convert_flags(cfile->f_flags, 0);
1263 		rdwr_for_fscache = 2;
1264 		goto retry_open;
1265 	}
1266 
1267 	if (rc) {
1268 		mutex_unlock(&cfile->fh_mutex);
1269 		cifs_dbg(FYI, "cifs_reopen returned 0x%x\n", rc);
1270 		cifs_dbg(FYI, "oplock: %d\n", oplock);
1271 		goto reopen_error_exit;
1272 	}
1273 
1274 	if (rdwr_for_fscache == 2)
1275 		cifs_invalidate_cache(inode, FSCACHE_INVAL_DIO_WRITE);
1276 
1277 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
1278 reopen_success:
1279 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */
1280 	cfile->invalidHandle = false;
1281 	mutex_unlock(&cfile->fh_mutex);
1282 	cinode = CIFS_I(inode);
1283 
1284 	if (can_flush) {
1285 		rc = filemap_write_and_wait(inode->i_mapping);
1286 		if (!is_interrupt_error(rc))
1287 			mapping_set_error(inode->i_mapping, rc);
1288 
1289 		if (tcon->posix_extensions) {
1290 			rc = smb311_posix_get_inode_info(&inode, full_path,
1291 							 NULL, inode->i_sb, xid);
1292 		} else if (tcon->unix_ext) {
1293 			rc = cifs_get_inode_info_unix(&inode, full_path,
1294 						      inode->i_sb, xid);
1295 		} else {
1296 			rc = cifs_get_inode_info(&inode, full_path, NULL,
1297 						 inode->i_sb, xid, NULL);
1298 		}
1299 	}
1300 	/*
1301 	 * Else we are writing out data to server already and could deadlock if
1302 	 * we tried to flush data, and since we do not know if we have data that
1303 	 * would invalidate the current end of file on the server we can not go
1304 	 * to the server to get the new inode info.
1305 	 */
1306 
1307 	/*
1308 	 * If the server returned a read oplock and we have mandatory brlocks,
1309 	 * set oplock level to None.
1310 	 */
1311 	if (server->ops->is_read_op(oplock) && cifs_has_mand_locks(cinode)) {
1312 		cifs_dbg(FYI, "Reset oplock val from read to None due to mand locks\n");
1313 		oplock = 0;
1314 	}
1315 
1316 	server->ops->set_fid(cfile, &cfile->fid, oplock);
1317 	if (oparms.reconnect)
1318 		cifs_relock_file(cfile);
1319 
1320 reopen_error_exit:
1321 	free_dentry_path(page);
1322 	free_xid(xid);
1323 	return rc;
1324 }
1325 
1326 void smb2_deferred_work_close(struct work_struct *work)
1327 {
1328 	struct cifsFileInfo *cfile = container_of(work,
1329 			struct cifsFileInfo, deferred.work);
1330 
1331 	spin_lock(&CIFS_I(d_inode(cfile->dentry))->deferred_lock);
1332 	cifs_del_deferred_close(cfile);
1333 	cfile->deferred_close_scheduled = false;
1334 	spin_unlock(&CIFS_I(d_inode(cfile->dentry))->deferred_lock);
1335 	_cifsFileInfo_put(cfile, true, false);
1336 }
1337 
1338 static bool
1339 smb2_can_defer_close(struct inode *inode, struct cifs_deferred_close *dclose)
1340 {
1341 	struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb);
1342 	struct cifsInodeInfo *cinode = CIFS_I(inode);
1343 
1344 	return (cifs_sb->ctx->closetimeo && cinode->lease_granted && dclose &&
1345 			(cinode->oplock == CIFS_CACHE_RHW_FLG ||
1346 			 cinode->oplock == CIFS_CACHE_RH_FLG) &&
1347 			!test_bit(CIFS_INO_CLOSE_ON_LOCK, &cinode->flags));
1348 
1349 }
1350 
1351 int cifs_close(struct inode *inode, struct file *file)
1352 {
1353 	struct cifsFileInfo *cfile;
1354 	struct cifsInodeInfo *cinode = CIFS_I(inode);
1355 	struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb);
1356 	struct cifs_deferred_close *dclose;
1357 
1358 	cifs_fscache_unuse_inode_cookie(inode, file->f_mode & FMODE_WRITE);
1359 
1360 	if (file->private_data != NULL) {
1361 		cfile = file->private_data;
1362 		file->private_data = NULL;
1363 		dclose = kmalloc(sizeof(struct cifs_deferred_close), GFP_KERNEL);
1364 		if ((cfile->status_file_deleted == false) &&
1365 		    (smb2_can_defer_close(inode, dclose))) {
1366 			if (test_and_clear_bit(CIFS_INO_MODIFIED_ATTR, &cinode->flags)) {
1367 				inode_set_mtime_to_ts(inode,
1368 						      inode_set_ctime_current(inode));
1369 			}
1370 			spin_lock(&cinode->deferred_lock);
1371 			cifs_add_deferred_close(cfile, dclose);
1372 			if (cfile->deferred_close_scheduled &&
1373 			    delayed_work_pending(&cfile->deferred)) {
1374 				/*
1375 				 * If there is no pending work, mod_delayed_work queues new work.
1376 				 * So, Increase the ref count to avoid use-after-free.
1377 				 */
1378 				if (!mod_delayed_work(deferredclose_wq,
1379 						&cfile->deferred, cifs_sb->ctx->closetimeo))
1380 					cifsFileInfo_get(cfile);
1381 			} else {
1382 				/* Deferred close for files */
1383 				queue_delayed_work(deferredclose_wq,
1384 						&cfile->deferred, cifs_sb->ctx->closetimeo);
1385 				cfile->deferred_close_scheduled = true;
1386 				spin_unlock(&cinode->deferred_lock);
1387 				return 0;
1388 			}
1389 			spin_unlock(&cinode->deferred_lock);
1390 			_cifsFileInfo_put(cfile, true, false);
1391 		} else {
1392 			_cifsFileInfo_put(cfile, true, false);
1393 			kfree(dclose);
1394 		}
1395 	}
1396 
1397 	/* return code from the ->release op is always ignored */
1398 	return 0;
1399 }
1400 
1401 void
1402 cifs_reopen_persistent_handles(struct cifs_tcon *tcon)
1403 {
1404 	struct cifsFileInfo *open_file, *tmp;
1405 	struct list_head tmp_list;
1406 
1407 	if (!tcon->use_persistent || !tcon->need_reopen_files)
1408 		return;
1409 
1410 	tcon->need_reopen_files = false;
1411 
1412 	cifs_dbg(FYI, "Reopen persistent handles\n");
1413 	INIT_LIST_HEAD(&tmp_list);
1414 
1415 	/* list all files open on tree connection, reopen resilient handles  */
1416 	spin_lock(&tcon->open_file_lock);
1417 	list_for_each_entry(open_file, &tcon->openFileList, tlist) {
1418 		if (!open_file->invalidHandle)
1419 			continue;
1420 		cifsFileInfo_get(open_file);
1421 		list_add_tail(&open_file->rlist, &tmp_list);
1422 	}
1423 	spin_unlock(&tcon->open_file_lock);
1424 
1425 	list_for_each_entry_safe(open_file, tmp, &tmp_list, rlist) {
1426 		if (cifs_reopen_file(open_file, false /* do not flush */))
1427 			tcon->need_reopen_files = true;
1428 		list_del_init(&open_file->rlist);
1429 		cifsFileInfo_put(open_file);
1430 	}
1431 }
1432 
1433 int cifs_closedir(struct inode *inode, struct file *file)
1434 {
1435 	int rc = 0;
1436 	unsigned int xid;
1437 	struct cifsFileInfo *cfile = file->private_data;
1438 	struct cifs_tcon *tcon;
1439 	struct TCP_Server_Info *server;
1440 	char *buf;
1441 
1442 	cifs_dbg(FYI, "Closedir inode = 0x%p\n", inode);
1443 
1444 	if (cfile == NULL)
1445 		return rc;
1446 
1447 	xid = get_xid();
1448 	tcon = tlink_tcon(cfile->tlink);
1449 	server = tcon->ses->server;
1450 
1451 	cifs_dbg(FYI, "Freeing private data in close dir\n");
1452 	spin_lock(&cfile->file_info_lock);
1453 	if (server->ops->dir_needs_close(cfile)) {
1454 		cfile->invalidHandle = true;
1455 		spin_unlock(&cfile->file_info_lock);
1456 		if (server->ops->close_dir)
1457 			rc = server->ops->close_dir(xid, tcon, &cfile->fid);
1458 		else
1459 			rc = -ENOSYS;
1460 		cifs_dbg(FYI, "Closing uncompleted readdir with rc %d\n", rc);
1461 		/* not much we can do if it fails anyway, ignore rc */
1462 		rc = 0;
1463 	} else
1464 		spin_unlock(&cfile->file_info_lock);
1465 
1466 	buf = cfile->srch_inf.ntwrk_buf_start;
1467 	if (buf) {
1468 		cifs_dbg(FYI, "closedir free smb buf in srch struct\n");
1469 		cfile->srch_inf.ntwrk_buf_start = NULL;
1470 		if (cfile->srch_inf.smallBuf)
1471 			cifs_small_buf_release(buf);
1472 		else
1473 			cifs_buf_release(buf);
1474 	}
1475 
1476 	cifs_put_tlink(cfile->tlink);
1477 	kfree(file->private_data);
1478 	file->private_data = NULL;
1479 	/* BB can we lock the filestruct while this is going on? */
1480 	free_xid(xid);
1481 	return rc;
1482 }
1483 
1484 static struct cifsLockInfo *
1485 cifs_lock_init(__u64 offset, __u64 length, __u8 type, __u16 flags)
1486 {
1487 	struct cifsLockInfo *lock =
1488 		kmalloc(sizeof(struct cifsLockInfo), GFP_KERNEL);
1489 	if (!lock)
1490 		return lock;
1491 	lock->offset = offset;
1492 	lock->length = length;
1493 	lock->type = type;
1494 	lock->pid = current->tgid;
1495 	lock->flags = flags;
1496 	INIT_LIST_HEAD(&lock->blist);
1497 	init_waitqueue_head(&lock->block_q);
1498 	return lock;
1499 }
1500 
1501 void
1502 cifs_del_lock_waiters(struct cifsLockInfo *lock)
1503 {
1504 	struct cifsLockInfo *li, *tmp;
1505 	list_for_each_entry_safe(li, tmp, &lock->blist, blist) {
1506 		list_del_init(&li->blist);
1507 		wake_up(&li->block_q);
1508 	}
1509 }
1510 
1511 #define CIFS_LOCK_OP	0
1512 #define CIFS_READ_OP	1
1513 #define CIFS_WRITE_OP	2
1514 
1515 /* @rw_check : 0 - no op, 1 - read, 2 - write */
1516 static bool
1517 cifs_find_fid_lock_conflict(struct cifs_fid_locks *fdlocks, __u64 offset,
1518 			    __u64 length, __u8 type, __u16 flags,
1519 			    struct cifsFileInfo *cfile,
1520 			    struct cifsLockInfo **conf_lock, int rw_check)
1521 {
1522 	struct cifsLockInfo *li;
1523 	struct cifsFileInfo *cur_cfile = fdlocks->cfile;
1524 	struct TCP_Server_Info *server = tlink_tcon(cfile->tlink)->ses->server;
1525 
1526 	list_for_each_entry(li, &fdlocks->locks, llist) {
1527 		if (offset + length <= li->offset ||
1528 		    offset >= li->offset + li->length)
1529 			continue;
1530 		if (rw_check != CIFS_LOCK_OP && current->tgid == li->pid &&
1531 		    server->ops->compare_fids(cfile, cur_cfile)) {
1532 			/* shared lock prevents write op through the same fid */
1533 			if (!(li->type & server->vals->shared_lock_type) ||
1534 			    rw_check != CIFS_WRITE_OP)
1535 				continue;
1536 		}
1537 		if ((type & server->vals->shared_lock_type) &&
1538 		    ((server->ops->compare_fids(cfile, cur_cfile) &&
1539 		     current->tgid == li->pid) || type == li->type))
1540 			continue;
1541 		if (rw_check == CIFS_LOCK_OP &&
1542 		    (flags & FL_OFDLCK) && (li->flags & FL_OFDLCK) &&
1543 		    server->ops->compare_fids(cfile, cur_cfile))
1544 			continue;
1545 		if (conf_lock)
1546 			*conf_lock = li;
1547 		return true;
1548 	}
1549 	return false;
1550 }
1551 
1552 bool
1553 cifs_find_lock_conflict(struct cifsFileInfo *cfile, __u64 offset, __u64 length,
1554 			__u8 type, __u16 flags,
1555 			struct cifsLockInfo **conf_lock, int rw_check)
1556 {
1557 	bool rc = false;
1558 	struct cifs_fid_locks *cur;
1559 	struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry));
1560 
1561 	list_for_each_entry(cur, &cinode->llist, llist) {
1562 		rc = cifs_find_fid_lock_conflict(cur, offset, length, type,
1563 						 flags, cfile, conf_lock,
1564 						 rw_check);
1565 		if (rc)
1566 			break;
1567 	}
1568 
1569 	return rc;
1570 }
1571 
1572 /*
1573  * Check if there is another lock that prevents us to set the lock (mandatory
1574  * style). If such a lock exists, update the flock structure with its
1575  * properties. Otherwise, set the flock type to F_UNLCK if we can cache brlocks
1576  * or leave it the same if we can't. Returns 0 if we don't need to request to
1577  * the server or 1 otherwise.
1578  */
1579 static int
1580 cifs_lock_test(struct cifsFileInfo *cfile, __u64 offset, __u64 length,
1581 	       __u8 type, struct file_lock *flock)
1582 {
1583 	int rc = 0;
1584 	struct cifsLockInfo *conf_lock;
1585 	struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry));
1586 	struct TCP_Server_Info *server = tlink_tcon(cfile->tlink)->ses->server;
1587 	bool exist;
1588 
1589 	down_read(&cinode->lock_sem);
1590 
1591 	exist = cifs_find_lock_conflict(cfile, offset, length, type,
1592 					flock->c.flc_flags, &conf_lock,
1593 					CIFS_LOCK_OP);
1594 	if (exist) {
1595 		flock->fl_start = conf_lock->offset;
1596 		flock->fl_end = conf_lock->offset + conf_lock->length - 1;
1597 		flock->c.flc_pid = conf_lock->pid;
1598 		if (conf_lock->type & server->vals->shared_lock_type)
1599 			flock->c.flc_type = F_RDLCK;
1600 		else
1601 			flock->c.flc_type = F_WRLCK;
1602 	} else if (!cinode->can_cache_brlcks)
1603 		rc = 1;
1604 	else
1605 		flock->c.flc_type = F_UNLCK;
1606 
1607 	up_read(&cinode->lock_sem);
1608 	return rc;
1609 }
1610 
1611 static void
1612 cifs_lock_add(struct cifsFileInfo *cfile, struct cifsLockInfo *lock)
1613 {
1614 	struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry));
1615 	cifs_down_write(&cinode->lock_sem);
1616 	list_add_tail(&lock->llist, &cfile->llist->locks);
1617 	up_write(&cinode->lock_sem);
1618 }
1619 
1620 /*
1621  * Set the byte-range lock (mandatory style). Returns:
1622  * 1) 0, if we set the lock and don't need to request to the server;
1623  * 2) 1, if no locks prevent us but we need to request to the server;
1624  * 3) -EACCES, if there is a lock that prevents us and wait is false.
1625  */
1626 static int
1627 cifs_lock_add_if(struct cifsFileInfo *cfile, struct cifsLockInfo *lock,
1628 		 bool wait)
1629 {
1630 	struct cifsLockInfo *conf_lock;
1631 	struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry));
1632 	bool exist;
1633 	int rc = 0;
1634 
1635 try_again:
1636 	exist = false;
1637 	cifs_down_write(&cinode->lock_sem);
1638 
1639 	exist = cifs_find_lock_conflict(cfile, lock->offset, lock->length,
1640 					lock->type, lock->flags, &conf_lock,
1641 					CIFS_LOCK_OP);
1642 	if (!exist && cinode->can_cache_brlcks) {
1643 		list_add_tail(&lock->llist, &cfile->llist->locks);
1644 		up_write(&cinode->lock_sem);
1645 		return rc;
1646 	}
1647 
1648 	if (!exist)
1649 		rc = 1;
1650 	else if (!wait)
1651 		rc = -EACCES;
1652 	else {
1653 		list_add_tail(&lock->blist, &conf_lock->blist);
1654 		up_write(&cinode->lock_sem);
1655 		rc = wait_event_interruptible(lock->block_q,
1656 					(lock->blist.prev == &lock->blist) &&
1657 					(lock->blist.next == &lock->blist));
1658 		if (!rc)
1659 			goto try_again;
1660 		cifs_down_write(&cinode->lock_sem);
1661 		list_del_init(&lock->blist);
1662 	}
1663 
1664 	up_write(&cinode->lock_sem);
1665 	return rc;
1666 }
1667 
1668 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
1669 /*
1670  * Check if there is another lock that prevents us to set the lock (posix
1671  * style). If such a lock exists, update the flock structure with its
1672  * properties. Otherwise, set the flock type to F_UNLCK if we can cache brlocks
1673  * or leave it the same if we can't. Returns 0 if we don't need to request to
1674  * the server or 1 otherwise.
1675  */
1676 static int
1677 cifs_posix_lock_test(struct file *file, struct file_lock *flock)
1678 {
1679 	int rc = 0;
1680 	struct cifsInodeInfo *cinode = CIFS_I(file_inode(file));
1681 	unsigned char saved_type = flock->c.flc_type;
1682 
1683 	if ((flock->c.flc_flags & FL_POSIX) == 0)
1684 		return 1;
1685 
1686 	down_read(&cinode->lock_sem);
1687 	posix_test_lock(file, flock);
1688 
1689 	if (lock_is_unlock(flock) && !cinode->can_cache_brlcks) {
1690 		flock->c.flc_type = saved_type;
1691 		rc = 1;
1692 	}
1693 
1694 	up_read(&cinode->lock_sem);
1695 	return rc;
1696 }
1697 
1698 /*
1699  * Set the byte-range lock (posix style). Returns:
1700  * 1) <0, if the error occurs while setting the lock;
1701  * 2) 0, if we set the lock and don't need to request to the server;
1702  * 3) FILE_LOCK_DEFERRED, if we will wait for some other file_lock;
1703  * 4) FILE_LOCK_DEFERRED + 1, if we need to request to the server.
1704  */
1705 static int
1706 cifs_posix_lock_set(struct file *file, struct file_lock *flock)
1707 {
1708 	struct cifsInodeInfo *cinode = CIFS_I(file_inode(file));
1709 	int rc = FILE_LOCK_DEFERRED + 1;
1710 
1711 	if ((flock->c.flc_flags & FL_POSIX) == 0)
1712 		return rc;
1713 
1714 	cifs_down_write(&cinode->lock_sem);
1715 	if (!cinode->can_cache_brlcks) {
1716 		up_write(&cinode->lock_sem);
1717 		return rc;
1718 	}
1719 
1720 	rc = posix_lock_file(file, flock, NULL);
1721 	up_write(&cinode->lock_sem);
1722 	return rc;
1723 }
1724 
1725 int
1726 cifs_push_mandatory_locks(struct cifsFileInfo *cfile)
1727 {
1728 	unsigned int xid;
1729 	int rc = 0, stored_rc;
1730 	struct cifsLockInfo *li, *tmp;
1731 	struct cifs_tcon *tcon;
1732 	unsigned int num, max_num, max_buf;
1733 	LOCKING_ANDX_RANGE *buf, *cur;
1734 	static const int types[] = {
1735 		LOCKING_ANDX_LARGE_FILES,
1736 		LOCKING_ANDX_SHARED_LOCK | LOCKING_ANDX_LARGE_FILES
1737 	};
1738 	int i;
1739 
1740 	xid = get_xid();
1741 	tcon = tlink_tcon(cfile->tlink);
1742 
1743 	/*
1744 	 * Accessing maxBuf is racy with cifs_reconnect - need to store value
1745 	 * and check it before using.
1746 	 */
1747 	max_buf = tcon->ses->server->maxBuf;
1748 	if (max_buf < (sizeof(struct smb_hdr) + sizeof(LOCKING_ANDX_RANGE))) {
1749 		free_xid(xid);
1750 		return -EINVAL;
1751 	}
1752 
1753 	BUILD_BUG_ON(sizeof(struct smb_hdr) + sizeof(LOCKING_ANDX_RANGE) >
1754 		     PAGE_SIZE);
1755 	max_buf = min_t(unsigned int, max_buf - sizeof(struct smb_hdr),
1756 			PAGE_SIZE);
1757 	max_num = (max_buf - sizeof(struct smb_hdr)) /
1758 						sizeof(LOCKING_ANDX_RANGE);
1759 	buf = kcalloc(max_num, sizeof(LOCKING_ANDX_RANGE), GFP_KERNEL);
1760 	if (!buf) {
1761 		free_xid(xid);
1762 		return -ENOMEM;
1763 	}
1764 
1765 	for (i = 0; i < 2; i++) {
1766 		cur = buf;
1767 		num = 0;
1768 		list_for_each_entry_safe(li, tmp, &cfile->llist->locks, llist) {
1769 			if (li->type != types[i])
1770 				continue;
1771 			cur->Pid = cpu_to_le16(li->pid);
1772 			cur->LengthLow = cpu_to_le32((u32)li->length);
1773 			cur->LengthHigh = cpu_to_le32((u32)(li->length>>32));
1774 			cur->OffsetLow = cpu_to_le32((u32)li->offset);
1775 			cur->OffsetHigh = cpu_to_le32((u32)(li->offset>>32));
1776 			if (++num == max_num) {
1777 				stored_rc = cifs_lockv(xid, tcon,
1778 						       cfile->fid.netfid,
1779 						       (__u8)li->type, 0, num,
1780 						       buf);
1781 				if (stored_rc)
1782 					rc = stored_rc;
1783 				cur = buf;
1784 				num = 0;
1785 			} else
1786 				cur++;
1787 		}
1788 
1789 		if (num) {
1790 			stored_rc = cifs_lockv(xid, tcon, cfile->fid.netfid,
1791 					       (__u8)types[i], 0, num, buf);
1792 			if (stored_rc)
1793 				rc = stored_rc;
1794 		}
1795 	}
1796 
1797 	kfree(buf);
1798 	free_xid(xid);
1799 	return rc;
1800 }
1801 
1802 static __u32
1803 hash_lockowner(fl_owner_t owner)
1804 {
1805 	return cifs_lock_secret ^ hash32_ptr((const void *)owner);
1806 }
1807 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */
1808 
1809 struct lock_to_push {
1810 	struct list_head llist;
1811 	__u64 offset;
1812 	__u64 length;
1813 	__u32 pid;
1814 	__u16 netfid;
1815 	__u8 type;
1816 };
1817 
1818 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
1819 static int
1820 cifs_push_posix_locks(struct cifsFileInfo *cfile)
1821 {
1822 	struct inode *inode = d_inode(cfile->dentry);
1823 	struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
1824 	struct file_lock *flock;
1825 	struct file_lock_context *flctx = locks_inode_context(inode);
1826 	unsigned int count = 0, i;
1827 	int rc = 0, xid, type;
1828 	struct list_head locks_to_send, *el;
1829 	struct lock_to_push *lck, *tmp;
1830 	__u64 length;
1831 
1832 	xid = get_xid();
1833 
1834 	if (!flctx)
1835 		goto out;
1836 
1837 	spin_lock(&flctx->flc_lock);
1838 	list_for_each(el, &flctx->flc_posix) {
1839 		count++;
1840 	}
1841 	spin_unlock(&flctx->flc_lock);
1842 
1843 	INIT_LIST_HEAD(&locks_to_send);
1844 
1845 	/*
1846 	 * Allocating count locks is enough because no FL_POSIX locks can be
1847 	 * added to the list while we are holding cinode->lock_sem that
1848 	 * protects locking operations of this inode.
1849 	 */
1850 	for (i = 0; i < count; i++) {
1851 		lck = kmalloc(sizeof(struct lock_to_push), GFP_KERNEL);
1852 		if (!lck) {
1853 			rc = -ENOMEM;
1854 			goto err_out;
1855 		}
1856 		list_add_tail(&lck->llist, &locks_to_send);
1857 	}
1858 
1859 	el = locks_to_send.next;
1860 	spin_lock(&flctx->flc_lock);
1861 	for_each_file_lock(flock, &flctx->flc_posix) {
1862 		unsigned char ftype = flock->c.flc_type;
1863 
1864 		if (el == &locks_to_send) {
1865 			/*
1866 			 * The list ended. We don't have enough allocated
1867 			 * structures - something is really wrong.
1868 			 */
1869 			cifs_dbg(VFS, "Can't push all brlocks!\n");
1870 			break;
1871 		}
1872 		length = cifs_flock_len(flock);
1873 		if (ftype == F_RDLCK || ftype == F_SHLCK)
1874 			type = CIFS_RDLCK;
1875 		else
1876 			type = CIFS_WRLCK;
1877 		lck = list_entry(el, struct lock_to_push, llist);
1878 		lck->pid = hash_lockowner(flock->c.flc_owner);
1879 		lck->netfid = cfile->fid.netfid;
1880 		lck->length = length;
1881 		lck->type = type;
1882 		lck->offset = flock->fl_start;
1883 	}
1884 	spin_unlock(&flctx->flc_lock);
1885 
1886 	list_for_each_entry_safe(lck, tmp, &locks_to_send, llist) {
1887 		int stored_rc;
1888 
1889 		stored_rc = CIFSSMBPosixLock(xid, tcon, lck->netfid, lck->pid,
1890 					     lck->offset, lck->length, NULL,
1891 					     lck->type, 0);
1892 		if (stored_rc)
1893 			rc = stored_rc;
1894 		list_del(&lck->llist);
1895 		kfree(lck);
1896 	}
1897 
1898 out:
1899 	free_xid(xid);
1900 	return rc;
1901 err_out:
1902 	list_for_each_entry_safe(lck, tmp, &locks_to_send, llist) {
1903 		list_del(&lck->llist);
1904 		kfree(lck);
1905 	}
1906 	goto out;
1907 }
1908 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */
1909 
1910 static int
1911 cifs_push_locks(struct cifsFileInfo *cfile)
1912 {
1913 	struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry));
1914 	struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
1915 	int rc = 0;
1916 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
1917 	struct cifs_sb_info *cifs_sb = CIFS_SB(cfile->dentry->d_sb);
1918 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */
1919 
1920 	/* we are going to update can_cache_brlcks here - need a write access */
1921 	cifs_down_write(&cinode->lock_sem);
1922 	if (!cinode->can_cache_brlcks) {
1923 		up_write(&cinode->lock_sem);
1924 		return rc;
1925 	}
1926 
1927 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
1928 	if (cap_unix(tcon->ses) &&
1929 	    (CIFS_UNIX_FCNTL_CAP & le64_to_cpu(tcon->fsUnixInfo.Capability)) &&
1930 	    ((cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOPOSIXBRL) == 0))
1931 		rc = cifs_push_posix_locks(cfile);
1932 	else
1933 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */
1934 		rc = tcon->ses->server->ops->push_mand_locks(cfile);
1935 
1936 	cinode->can_cache_brlcks = false;
1937 	up_write(&cinode->lock_sem);
1938 	return rc;
1939 }
1940 
1941 static void
1942 cifs_read_flock(struct file_lock *flock, __u32 *type, int *lock, int *unlock,
1943 		bool *wait_flag, struct TCP_Server_Info *server)
1944 {
1945 	if (flock->c.flc_flags & FL_POSIX)
1946 		cifs_dbg(FYI, "Posix\n");
1947 	if (flock->c.flc_flags & FL_FLOCK)
1948 		cifs_dbg(FYI, "Flock\n");
1949 	if (flock->c.flc_flags & FL_SLEEP) {
1950 		cifs_dbg(FYI, "Blocking lock\n");
1951 		*wait_flag = true;
1952 	}
1953 	if (flock->c.flc_flags & FL_ACCESS)
1954 		cifs_dbg(FYI, "Process suspended by mandatory locking - not implemented yet\n");
1955 	if (flock->c.flc_flags & FL_LEASE)
1956 		cifs_dbg(FYI, "Lease on file - not implemented yet\n");
1957 	if (flock->c.flc_flags &
1958 	    (~(FL_POSIX | FL_FLOCK | FL_SLEEP |
1959 	       FL_ACCESS | FL_LEASE | FL_CLOSE | FL_OFDLCK)))
1960 		cifs_dbg(FYI, "Unknown lock flags 0x%x\n",
1961 		         flock->c.flc_flags);
1962 
1963 	*type = server->vals->large_lock_type;
1964 	if (lock_is_write(flock)) {
1965 		cifs_dbg(FYI, "F_WRLCK\n");
1966 		*type |= server->vals->exclusive_lock_type;
1967 		*lock = 1;
1968 	} else if (lock_is_unlock(flock)) {
1969 		cifs_dbg(FYI, "F_UNLCK\n");
1970 		*type |= server->vals->unlock_lock_type;
1971 		*unlock = 1;
1972 		/* Check if unlock includes more than one lock range */
1973 	} else if (lock_is_read(flock)) {
1974 		cifs_dbg(FYI, "F_RDLCK\n");
1975 		*type |= server->vals->shared_lock_type;
1976 		*lock = 1;
1977 	} else if (flock->c.flc_type == F_EXLCK) {
1978 		cifs_dbg(FYI, "F_EXLCK\n");
1979 		*type |= server->vals->exclusive_lock_type;
1980 		*lock = 1;
1981 	} else if (flock->c.flc_type == F_SHLCK) {
1982 		cifs_dbg(FYI, "F_SHLCK\n");
1983 		*type |= server->vals->shared_lock_type;
1984 		*lock = 1;
1985 	} else
1986 		cifs_dbg(FYI, "Unknown type of lock\n");
1987 }
1988 
1989 static int
1990 cifs_getlk(struct file *file, struct file_lock *flock, __u32 type,
1991 	   bool wait_flag, bool posix_lck, unsigned int xid)
1992 {
1993 	int rc = 0;
1994 	__u64 length = cifs_flock_len(flock);
1995 	struct cifsFileInfo *cfile = (struct cifsFileInfo *)file->private_data;
1996 	struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
1997 	struct TCP_Server_Info *server = tcon->ses->server;
1998 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
1999 	__u16 netfid = cfile->fid.netfid;
2000 
2001 	if (posix_lck) {
2002 		int posix_lock_type;
2003 
2004 		rc = cifs_posix_lock_test(file, flock);
2005 		if (!rc)
2006 			return rc;
2007 
2008 		if (type & server->vals->shared_lock_type)
2009 			posix_lock_type = CIFS_RDLCK;
2010 		else
2011 			posix_lock_type = CIFS_WRLCK;
2012 		rc = CIFSSMBPosixLock(xid, tcon, netfid,
2013 				      hash_lockowner(flock->c.flc_owner),
2014 				      flock->fl_start, length, flock,
2015 				      posix_lock_type, wait_flag);
2016 		return rc;
2017 	}
2018 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */
2019 
2020 	rc = cifs_lock_test(cfile, flock->fl_start, length, type, flock);
2021 	if (!rc)
2022 		return rc;
2023 
2024 	/* BB we could chain these into one lock request BB */
2025 	rc = server->ops->mand_lock(xid, cfile, flock->fl_start, length, type,
2026 				    1, 0, false);
2027 	if (rc == 0) {
2028 		rc = server->ops->mand_lock(xid, cfile, flock->fl_start, length,
2029 					    type, 0, 1, false);
2030 		flock->c.flc_type = F_UNLCK;
2031 		if (rc != 0)
2032 			cifs_dbg(VFS, "Error unlocking previously locked range %d during test of lock\n",
2033 				 rc);
2034 		return 0;
2035 	}
2036 
2037 	if (type & server->vals->shared_lock_type) {
2038 		flock->c.flc_type = F_WRLCK;
2039 		return 0;
2040 	}
2041 
2042 	type &= ~server->vals->exclusive_lock_type;
2043 
2044 	rc = server->ops->mand_lock(xid, cfile, flock->fl_start, length,
2045 				    type | server->vals->shared_lock_type,
2046 				    1, 0, false);
2047 	if (rc == 0) {
2048 		rc = server->ops->mand_lock(xid, cfile, flock->fl_start, length,
2049 			type | server->vals->shared_lock_type, 0, 1, false);
2050 		flock->c.flc_type = F_RDLCK;
2051 		if (rc != 0)
2052 			cifs_dbg(VFS, "Error unlocking previously locked range %d during test of lock\n",
2053 				 rc);
2054 	} else
2055 		flock->c.flc_type = F_WRLCK;
2056 
2057 	return 0;
2058 }
2059 
2060 void
2061 cifs_move_llist(struct list_head *source, struct list_head *dest)
2062 {
2063 	struct list_head *li, *tmp;
2064 	list_for_each_safe(li, tmp, source)
2065 		list_move(li, dest);
2066 }
2067 
2068 void
2069 cifs_free_llist(struct list_head *llist)
2070 {
2071 	struct cifsLockInfo *li, *tmp;
2072 	list_for_each_entry_safe(li, tmp, llist, llist) {
2073 		cifs_del_lock_waiters(li);
2074 		list_del(&li->llist);
2075 		kfree(li);
2076 	}
2077 }
2078 
2079 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
2080 int
2081 cifs_unlock_range(struct cifsFileInfo *cfile, struct file_lock *flock,
2082 		  unsigned int xid)
2083 {
2084 	int rc = 0, stored_rc;
2085 	static const int types[] = {
2086 		LOCKING_ANDX_LARGE_FILES,
2087 		LOCKING_ANDX_SHARED_LOCK | LOCKING_ANDX_LARGE_FILES
2088 	};
2089 	unsigned int i;
2090 	unsigned int max_num, num, max_buf;
2091 	LOCKING_ANDX_RANGE *buf, *cur;
2092 	struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
2093 	struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry));
2094 	struct cifsLockInfo *li, *tmp;
2095 	__u64 length = cifs_flock_len(flock);
2096 	struct list_head tmp_llist;
2097 
2098 	INIT_LIST_HEAD(&tmp_llist);
2099 
2100 	/*
2101 	 * Accessing maxBuf is racy with cifs_reconnect - need to store value
2102 	 * and check it before using.
2103 	 */
2104 	max_buf = tcon->ses->server->maxBuf;
2105 	if (max_buf < (sizeof(struct smb_hdr) + sizeof(LOCKING_ANDX_RANGE)))
2106 		return -EINVAL;
2107 
2108 	BUILD_BUG_ON(sizeof(struct smb_hdr) + sizeof(LOCKING_ANDX_RANGE) >
2109 		     PAGE_SIZE);
2110 	max_buf = min_t(unsigned int, max_buf - sizeof(struct smb_hdr),
2111 			PAGE_SIZE);
2112 	max_num = (max_buf - sizeof(struct smb_hdr)) /
2113 						sizeof(LOCKING_ANDX_RANGE);
2114 	buf = kcalloc(max_num, sizeof(LOCKING_ANDX_RANGE), GFP_KERNEL);
2115 	if (!buf)
2116 		return -ENOMEM;
2117 
2118 	cifs_down_write(&cinode->lock_sem);
2119 	for (i = 0; i < 2; i++) {
2120 		cur = buf;
2121 		num = 0;
2122 		list_for_each_entry_safe(li, tmp, &cfile->llist->locks, llist) {
2123 			if (flock->fl_start > li->offset ||
2124 			    (flock->fl_start + length) <
2125 			    (li->offset + li->length))
2126 				continue;
2127 			if (current->tgid != li->pid)
2128 				continue;
2129 			if (types[i] != li->type)
2130 				continue;
2131 			if (cinode->can_cache_brlcks) {
2132 				/*
2133 				 * We can cache brlock requests - simply remove
2134 				 * a lock from the file's list.
2135 				 */
2136 				list_del(&li->llist);
2137 				cifs_del_lock_waiters(li);
2138 				kfree(li);
2139 				continue;
2140 			}
2141 			cur->Pid = cpu_to_le16(li->pid);
2142 			cur->LengthLow = cpu_to_le32((u32)li->length);
2143 			cur->LengthHigh = cpu_to_le32((u32)(li->length>>32));
2144 			cur->OffsetLow = cpu_to_le32((u32)li->offset);
2145 			cur->OffsetHigh = cpu_to_le32((u32)(li->offset>>32));
2146 			/*
2147 			 * We need to save a lock here to let us add it again to
2148 			 * the file's list if the unlock range request fails on
2149 			 * the server.
2150 			 */
2151 			list_move(&li->llist, &tmp_llist);
2152 			if (++num == max_num) {
2153 				stored_rc = cifs_lockv(xid, tcon,
2154 						       cfile->fid.netfid,
2155 						       li->type, num, 0, buf);
2156 				if (stored_rc) {
2157 					/*
2158 					 * We failed on the unlock range
2159 					 * request - add all locks from the tmp
2160 					 * list to the head of the file's list.
2161 					 */
2162 					cifs_move_llist(&tmp_llist,
2163 							&cfile->llist->locks);
2164 					rc = stored_rc;
2165 				} else
2166 					/*
2167 					 * The unlock range request succeed -
2168 					 * free the tmp list.
2169 					 */
2170 					cifs_free_llist(&tmp_llist);
2171 				cur = buf;
2172 				num = 0;
2173 			} else
2174 				cur++;
2175 		}
2176 		if (num) {
2177 			stored_rc = cifs_lockv(xid, tcon, cfile->fid.netfid,
2178 					       types[i], num, 0, buf);
2179 			if (stored_rc) {
2180 				cifs_move_llist(&tmp_llist,
2181 						&cfile->llist->locks);
2182 				rc = stored_rc;
2183 			} else
2184 				cifs_free_llist(&tmp_llist);
2185 		}
2186 	}
2187 
2188 	up_write(&cinode->lock_sem);
2189 	kfree(buf);
2190 	return rc;
2191 }
2192 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */
2193 
2194 static int
2195 cifs_setlk(struct file *file, struct file_lock *flock, __u32 type,
2196 	   bool wait_flag, bool posix_lck, int lock, int unlock,
2197 	   unsigned int xid)
2198 {
2199 	int rc = 0;
2200 	__u64 length = cifs_flock_len(flock);
2201 	struct cifsFileInfo *cfile = (struct cifsFileInfo *)file->private_data;
2202 	struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
2203 	struct TCP_Server_Info *server = tcon->ses->server;
2204 	struct inode *inode = d_inode(cfile->dentry);
2205 
2206 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
2207 	if (posix_lck) {
2208 		int posix_lock_type;
2209 
2210 		rc = cifs_posix_lock_set(file, flock);
2211 		if (rc <= FILE_LOCK_DEFERRED)
2212 			return rc;
2213 
2214 		if (type & server->vals->shared_lock_type)
2215 			posix_lock_type = CIFS_RDLCK;
2216 		else
2217 			posix_lock_type = CIFS_WRLCK;
2218 
2219 		if (unlock == 1)
2220 			posix_lock_type = CIFS_UNLCK;
2221 
2222 		rc = CIFSSMBPosixLock(xid, tcon, cfile->fid.netfid,
2223 				      hash_lockowner(flock->c.flc_owner),
2224 				      flock->fl_start, length,
2225 				      NULL, posix_lock_type, wait_flag);
2226 		goto out;
2227 	}
2228 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */
2229 	if (lock) {
2230 		struct cifsLockInfo *lock;
2231 
2232 		lock = cifs_lock_init(flock->fl_start, length, type,
2233 				      flock->c.flc_flags);
2234 		if (!lock)
2235 			return -ENOMEM;
2236 
2237 		rc = cifs_lock_add_if(cfile, lock, wait_flag);
2238 		if (rc < 0) {
2239 			kfree(lock);
2240 			return rc;
2241 		}
2242 		if (!rc)
2243 			goto out;
2244 
2245 		/*
2246 		 * Windows 7 server can delay breaking lease from read to None
2247 		 * if we set a byte-range lock on a file - break it explicitly
2248 		 * before sending the lock to the server to be sure the next
2249 		 * read won't conflict with non-overlapted locks due to
2250 		 * pagereading.
2251 		 */
2252 		if (!CIFS_CACHE_WRITE(CIFS_I(inode)) &&
2253 					CIFS_CACHE_READ(CIFS_I(inode))) {
2254 			cifs_zap_mapping(inode);
2255 			cifs_dbg(FYI, "Set no oplock for inode=%p due to mand locks\n",
2256 				 inode);
2257 			CIFS_I(inode)->oplock = 0;
2258 		}
2259 
2260 		rc = server->ops->mand_lock(xid, cfile, flock->fl_start, length,
2261 					    type, 1, 0, wait_flag);
2262 		if (rc) {
2263 			kfree(lock);
2264 			return rc;
2265 		}
2266 
2267 		cifs_lock_add(cfile, lock);
2268 	} else if (unlock)
2269 		rc = server->ops->mand_unlock_range(cfile, flock, xid);
2270 
2271 out:
2272 	if ((flock->c.flc_flags & FL_POSIX) || (flock->c.flc_flags & FL_FLOCK)) {
2273 		/*
2274 		 * If this is a request to remove all locks because we
2275 		 * are closing the file, it doesn't matter if the
2276 		 * unlocking failed as both cifs.ko and the SMB server
2277 		 * remove the lock on file close
2278 		 */
2279 		if (rc) {
2280 			cifs_dbg(VFS, "%s failed rc=%d\n", __func__, rc);
2281 			if (!(flock->c.flc_flags & FL_CLOSE))
2282 				return rc;
2283 		}
2284 		rc = locks_lock_file_wait(file, flock);
2285 	}
2286 	return rc;
2287 }
2288 
2289 int cifs_flock(struct file *file, int cmd, struct file_lock *fl)
2290 {
2291 	int rc, xid;
2292 	int lock = 0, unlock = 0;
2293 	bool wait_flag = false;
2294 	bool posix_lck = false;
2295 	struct cifs_sb_info *cifs_sb;
2296 	struct cifs_tcon *tcon;
2297 	struct cifsFileInfo *cfile;
2298 	__u32 type;
2299 
2300 	xid = get_xid();
2301 
2302 	if (!(fl->c.flc_flags & FL_FLOCK)) {
2303 		rc = -ENOLCK;
2304 		free_xid(xid);
2305 		return rc;
2306 	}
2307 
2308 	cfile = (struct cifsFileInfo *)file->private_data;
2309 	tcon = tlink_tcon(cfile->tlink);
2310 
2311 	cifs_read_flock(fl, &type, &lock, &unlock, &wait_flag,
2312 			tcon->ses->server);
2313 	cifs_sb = CIFS_FILE_SB(file);
2314 
2315 	if (cap_unix(tcon->ses) &&
2316 	    (CIFS_UNIX_FCNTL_CAP & le64_to_cpu(tcon->fsUnixInfo.Capability)) &&
2317 	    ((cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOPOSIXBRL) == 0))
2318 		posix_lck = true;
2319 
2320 	if (!lock && !unlock) {
2321 		/*
2322 		 * if no lock or unlock then nothing to do since we do not
2323 		 * know what it is
2324 		 */
2325 		rc = -EOPNOTSUPP;
2326 		free_xid(xid);
2327 		return rc;
2328 	}
2329 
2330 	rc = cifs_setlk(file, fl, type, wait_flag, posix_lck, lock, unlock,
2331 			xid);
2332 	free_xid(xid);
2333 	return rc;
2334 
2335 
2336 }
2337 
2338 int cifs_lock(struct file *file, int cmd, struct file_lock *flock)
2339 {
2340 	int rc, xid;
2341 	int lock = 0, unlock = 0;
2342 	bool wait_flag = false;
2343 	bool posix_lck = false;
2344 	struct cifs_sb_info *cifs_sb;
2345 	struct cifs_tcon *tcon;
2346 	struct cifsFileInfo *cfile;
2347 	__u32 type;
2348 
2349 	rc = -EACCES;
2350 	xid = get_xid();
2351 
2352 	cifs_dbg(FYI, "%s: %pD2 cmd=0x%x type=0x%x flags=0x%x r=%lld:%lld\n", __func__, file, cmd,
2353 		 flock->c.flc_flags, flock->c.flc_type,
2354 		 (long long)flock->fl_start,
2355 		 (long long)flock->fl_end);
2356 
2357 	cfile = (struct cifsFileInfo *)file->private_data;
2358 	tcon = tlink_tcon(cfile->tlink);
2359 
2360 	cifs_read_flock(flock, &type, &lock, &unlock, &wait_flag,
2361 			tcon->ses->server);
2362 	cifs_sb = CIFS_FILE_SB(file);
2363 	set_bit(CIFS_INO_CLOSE_ON_LOCK, &CIFS_I(d_inode(cfile->dentry))->flags);
2364 
2365 	if (cap_unix(tcon->ses) &&
2366 	    (CIFS_UNIX_FCNTL_CAP & le64_to_cpu(tcon->fsUnixInfo.Capability)) &&
2367 	    ((cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOPOSIXBRL) == 0))
2368 		posix_lck = true;
2369 	/*
2370 	 * BB add code here to normalize offset and length to account for
2371 	 * negative length which we can not accept over the wire.
2372 	 */
2373 	if (IS_GETLK(cmd)) {
2374 		rc = cifs_getlk(file, flock, type, wait_flag, posix_lck, xid);
2375 		free_xid(xid);
2376 		return rc;
2377 	}
2378 
2379 	if (!lock && !unlock) {
2380 		/*
2381 		 * if no lock or unlock then nothing to do since we do not
2382 		 * know what it is
2383 		 */
2384 		free_xid(xid);
2385 		return -EOPNOTSUPP;
2386 	}
2387 
2388 	rc = cifs_setlk(file, flock, type, wait_flag, posix_lck, lock, unlock,
2389 			xid);
2390 	free_xid(xid);
2391 	return rc;
2392 }
2393 
2394 void cifs_write_subrequest_terminated(struct cifs_io_subrequest *wdata, ssize_t result,
2395 				      bool was_async)
2396 {
2397 	struct netfs_io_request *wreq = wdata->rreq;
2398 	struct netfs_inode *ictx = netfs_inode(wreq->inode);
2399 	loff_t wrend;
2400 
2401 	if (result > 0) {
2402 		wrend = wdata->subreq.start + wdata->subreq.transferred + result;
2403 
2404 		if (wrend > ictx->zero_point &&
2405 		    (wdata->rreq->origin == NETFS_UNBUFFERED_WRITE ||
2406 		     wdata->rreq->origin == NETFS_DIO_WRITE))
2407 			ictx->zero_point = wrend;
2408 		if (wrend > ictx->remote_i_size)
2409 			netfs_resize_file(ictx, wrend, true);
2410 	}
2411 
2412 	netfs_write_subrequest_terminated(&wdata->subreq, result, was_async);
2413 }
2414 
2415 struct cifsFileInfo *find_readable_file(struct cifsInodeInfo *cifs_inode,
2416 					bool fsuid_only)
2417 {
2418 	struct cifsFileInfo *open_file = NULL;
2419 	struct cifs_sb_info *cifs_sb = CIFS_SB(cifs_inode->netfs.inode.i_sb);
2420 
2421 	/* only filter by fsuid on multiuser mounts */
2422 	if (!(cifs_sb->mnt_cifs_flags & CIFS_MOUNT_MULTIUSER))
2423 		fsuid_only = false;
2424 
2425 	spin_lock(&cifs_inode->open_file_lock);
2426 	/* we could simply get the first_list_entry since write-only entries
2427 	   are always at the end of the list but since the first entry might
2428 	   have a close pending, we go through the whole list */
2429 	list_for_each_entry(open_file, &cifs_inode->openFileList, flist) {
2430 		if (fsuid_only && !uid_eq(open_file->uid, current_fsuid()))
2431 			continue;
2432 		if (OPEN_FMODE(open_file->f_flags) & FMODE_READ) {
2433 			if ((!open_file->invalidHandle)) {
2434 				/* found a good file */
2435 				/* lock it so it will not be closed on us */
2436 				cifsFileInfo_get(open_file);
2437 				spin_unlock(&cifs_inode->open_file_lock);
2438 				return open_file;
2439 			} /* else might as well continue, and look for
2440 			     another, or simply have the caller reopen it
2441 			     again rather than trying to fix this handle */
2442 		} else /* write only file */
2443 			break; /* write only files are last so must be done */
2444 	}
2445 	spin_unlock(&cifs_inode->open_file_lock);
2446 	return NULL;
2447 }
2448 
2449 /* Return -EBADF if no handle is found and general rc otherwise */
2450 int
2451 cifs_get_writable_file(struct cifsInodeInfo *cifs_inode, int flags,
2452 		       struct cifsFileInfo **ret_file)
2453 {
2454 	struct cifsFileInfo *open_file, *inv_file = NULL;
2455 	struct cifs_sb_info *cifs_sb;
2456 	bool any_available = false;
2457 	int rc = -EBADF;
2458 	unsigned int refind = 0;
2459 	bool fsuid_only = flags & FIND_WR_FSUID_ONLY;
2460 	bool with_delete = flags & FIND_WR_WITH_DELETE;
2461 	*ret_file = NULL;
2462 
2463 	/*
2464 	 * Having a null inode here (because mapping->host was set to zero by
2465 	 * the VFS or MM) should not happen but we had reports of on oops (due
2466 	 * to it being zero) during stress testcases so we need to check for it
2467 	 */
2468 
2469 	if (cifs_inode == NULL) {
2470 		cifs_dbg(VFS, "Null inode passed to cifs_writeable_file\n");
2471 		dump_stack();
2472 		return rc;
2473 	}
2474 
2475 	cifs_sb = CIFS_SB(cifs_inode->netfs.inode.i_sb);
2476 
2477 	/* only filter by fsuid on multiuser mounts */
2478 	if (!(cifs_sb->mnt_cifs_flags & CIFS_MOUNT_MULTIUSER))
2479 		fsuid_only = false;
2480 
2481 	spin_lock(&cifs_inode->open_file_lock);
2482 refind_writable:
2483 	if (refind > MAX_REOPEN_ATT) {
2484 		spin_unlock(&cifs_inode->open_file_lock);
2485 		return rc;
2486 	}
2487 	list_for_each_entry(open_file, &cifs_inode->openFileList, flist) {
2488 		if (!any_available && open_file->pid != current->tgid)
2489 			continue;
2490 		if (fsuid_only && !uid_eq(open_file->uid, current_fsuid()))
2491 			continue;
2492 		if (with_delete && !(open_file->fid.access & DELETE))
2493 			continue;
2494 		if (OPEN_FMODE(open_file->f_flags) & FMODE_WRITE) {
2495 			if (!open_file->invalidHandle) {
2496 				/* found a good writable file */
2497 				cifsFileInfo_get(open_file);
2498 				spin_unlock(&cifs_inode->open_file_lock);
2499 				*ret_file = open_file;
2500 				return 0;
2501 			} else {
2502 				if (!inv_file)
2503 					inv_file = open_file;
2504 			}
2505 		}
2506 	}
2507 	/* couldn't find useable FH with same pid, try any available */
2508 	if (!any_available) {
2509 		any_available = true;
2510 		goto refind_writable;
2511 	}
2512 
2513 	if (inv_file) {
2514 		any_available = false;
2515 		cifsFileInfo_get(inv_file);
2516 	}
2517 
2518 	spin_unlock(&cifs_inode->open_file_lock);
2519 
2520 	if (inv_file) {
2521 		rc = cifs_reopen_file(inv_file, false);
2522 		if (!rc) {
2523 			*ret_file = inv_file;
2524 			return 0;
2525 		}
2526 
2527 		spin_lock(&cifs_inode->open_file_lock);
2528 		list_move_tail(&inv_file->flist, &cifs_inode->openFileList);
2529 		spin_unlock(&cifs_inode->open_file_lock);
2530 		cifsFileInfo_put(inv_file);
2531 		++refind;
2532 		inv_file = NULL;
2533 		spin_lock(&cifs_inode->open_file_lock);
2534 		goto refind_writable;
2535 	}
2536 
2537 	return rc;
2538 }
2539 
2540 struct cifsFileInfo *
2541 find_writable_file(struct cifsInodeInfo *cifs_inode, int flags)
2542 {
2543 	struct cifsFileInfo *cfile;
2544 	int rc;
2545 
2546 	rc = cifs_get_writable_file(cifs_inode, flags, &cfile);
2547 	if (rc)
2548 		cifs_dbg(FYI, "Couldn't find writable handle rc=%d\n", rc);
2549 
2550 	return cfile;
2551 }
2552 
2553 int
2554 cifs_get_writable_path(struct cifs_tcon *tcon, const char *name,
2555 		       int flags,
2556 		       struct cifsFileInfo **ret_file)
2557 {
2558 	struct cifsFileInfo *cfile;
2559 	void *page = alloc_dentry_path();
2560 
2561 	*ret_file = NULL;
2562 
2563 	spin_lock(&tcon->open_file_lock);
2564 	list_for_each_entry(cfile, &tcon->openFileList, tlist) {
2565 		struct cifsInodeInfo *cinode;
2566 		const char *full_path = build_path_from_dentry(cfile->dentry, page);
2567 		if (IS_ERR(full_path)) {
2568 			spin_unlock(&tcon->open_file_lock);
2569 			free_dentry_path(page);
2570 			return PTR_ERR(full_path);
2571 		}
2572 		if (strcmp(full_path, name))
2573 			continue;
2574 
2575 		cinode = CIFS_I(d_inode(cfile->dentry));
2576 		spin_unlock(&tcon->open_file_lock);
2577 		free_dentry_path(page);
2578 		return cifs_get_writable_file(cinode, flags, ret_file);
2579 	}
2580 
2581 	spin_unlock(&tcon->open_file_lock);
2582 	free_dentry_path(page);
2583 	return -ENOENT;
2584 }
2585 
2586 int
2587 cifs_get_readable_path(struct cifs_tcon *tcon, const char *name,
2588 		       struct cifsFileInfo **ret_file)
2589 {
2590 	struct cifsFileInfo *cfile;
2591 	void *page = alloc_dentry_path();
2592 
2593 	*ret_file = NULL;
2594 
2595 	spin_lock(&tcon->open_file_lock);
2596 	list_for_each_entry(cfile, &tcon->openFileList, tlist) {
2597 		struct cifsInodeInfo *cinode;
2598 		const char *full_path = build_path_from_dentry(cfile->dentry, page);
2599 		if (IS_ERR(full_path)) {
2600 			spin_unlock(&tcon->open_file_lock);
2601 			free_dentry_path(page);
2602 			return PTR_ERR(full_path);
2603 		}
2604 		if (strcmp(full_path, name))
2605 			continue;
2606 
2607 		cinode = CIFS_I(d_inode(cfile->dentry));
2608 		spin_unlock(&tcon->open_file_lock);
2609 		free_dentry_path(page);
2610 		*ret_file = find_readable_file(cinode, 0);
2611 		return *ret_file ? 0 : -ENOENT;
2612 	}
2613 
2614 	spin_unlock(&tcon->open_file_lock);
2615 	free_dentry_path(page);
2616 	return -ENOENT;
2617 }
2618 
2619 /*
2620  * Flush data on a strict file.
2621  */
2622 int cifs_strict_fsync(struct file *file, loff_t start, loff_t end,
2623 		      int datasync)
2624 {
2625 	unsigned int xid;
2626 	int rc = 0;
2627 	struct cifs_tcon *tcon;
2628 	struct TCP_Server_Info *server;
2629 	struct cifsFileInfo *smbfile = file->private_data;
2630 	struct inode *inode = file_inode(file);
2631 	struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb);
2632 
2633 	rc = file_write_and_wait_range(file, start, end);
2634 	if (rc) {
2635 		trace_cifs_fsync_err(inode->i_ino, rc);
2636 		return rc;
2637 	}
2638 
2639 	xid = get_xid();
2640 
2641 	cifs_dbg(FYI, "Sync file - name: %pD datasync: 0x%x\n",
2642 		 file, datasync);
2643 
2644 	if (!CIFS_CACHE_READ(CIFS_I(inode))) {
2645 		rc = cifs_zap_mapping(inode);
2646 		if (rc) {
2647 			cifs_dbg(FYI, "rc: %d during invalidate phase\n", rc);
2648 			rc = 0; /* don't care about it in fsync */
2649 		}
2650 	}
2651 
2652 	tcon = tlink_tcon(smbfile->tlink);
2653 	if (!(cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOSSYNC)) {
2654 		server = tcon->ses->server;
2655 		if (server->ops->flush == NULL) {
2656 			rc = -ENOSYS;
2657 			goto strict_fsync_exit;
2658 		}
2659 
2660 		if ((OPEN_FMODE(smbfile->f_flags) & FMODE_WRITE) == 0) {
2661 			smbfile = find_writable_file(CIFS_I(inode), FIND_WR_ANY);
2662 			if (smbfile) {
2663 				rc = server->ops->flush(xid, tcon, &smbfile->fid);
2664 				cifsFileInfo_put(smbfile);
2665 			} else
2666 				cifs_dbg(FYI, "ignore fsync for file not open for write\n");
2667 		} else
2668 			rc = server->ops->flush(xid, tcon, &smbfile->fid);
2669 	}
2670 
2671 strict_fsync_exit:
2672 	free_xid(xid);
2673 	return rc;
2674 }
2675 
2676 /*
2677  * Flush data on a non-strict data.
2678  */
2679 int cifs_fsync(struct file *file, loff_t start, loff_t end, int datasync)
2680 {
2681 	unsigned int xid;
2682 	int rc = 0;
2683 	struct cifs_tcon *tcon;
2684 	struct TCP_Server_Info *server;
2685 	struct cifsFileInfo *smbfile = file->private_data;
2686 	struct inode *inode = file_inode(file);
2687 	struct cifs_sb_info *cifs_sb = CIFS_FILE_SB(file);
2688 
2689 	rc = file_write_and_wait_range(file, start, end);
2690 	if (rc) {
2691 		trace_cifs_fsync_err(file_inode(file)->i_ino, rc);
2692 		return rc;
2693 	}
2694 
2695 	xid = get_xid();
2696 
2697 	cifs_dbg(FYI, "Sync file - name: %pD datasync: 0x%x\n",
2698 		 file, datasync);
2699 
2700 	tcon = tlink_tcon(smbfile->tlink);
2701 	if (!(cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOSSYNC)) {
2702 		server = tcon->ses->server;
2703 		if (server->ops->flush == NULL) {
2704 			rc = -ENOSYS;
2705 			goto fsync_exit;
2706 		}
2707 
2708 		if ((OPEN_FMODE(smbfile->f_flags) & FMODE_WRITE) == 0) {
2709 			smbfile = find_writable_file(CIFS_I(inode), FIND_WR_ANY);
2710 			if (smbfile) {
2711 				rc = server->ops->flush(xid, tcon, &smbfile->fid);
2712 				cifsFileInfo_put(smbfile);
2713 			} else
2714 				cifs_dbg(FYI, "ignore fsync for file not open for write\n");
2715 		} else
2716 			rc = server->ops->flush(xid, tcon, &smbfile->fid);
2717 	}
2718 
2719 fsync_exit:
2720 	free_xid(xid);
2721 	return rc;
2722 }
2723 
2724 /*
2725  * As file closes, flush all cached write data for this inode checking
2726  * for write behind errors.
2727  */
2728 int cifs_flush(struct file *file, fl_owner_t id)
2729 {
2730 	struct inode *inode = file_inode(file);
2731 	int rc = 0;
2732 
2733 	if (file->f_mode & FMODE_WRITE)
2734 		rc = filemap_write_and_wait(inode->i_mapping);
2735 
2736 	cifs_dbg(FYI, "Flush inode %p file %p rc %d\n", inode, file, rc);
2737 	if (rc) {
2738 		/* get more nuanced writeback errors */
2739 		rc = filemap_check_wb_err(file->f_mapping, 0);
2740 		trace_cifs_flush_err(inode->i_ino, rc);
2741 	}
2742 	return rc;
2743 }
2744 
2745 static ssize_t
2746 cifs_writev(struct kiocb *iocb, struct iov_iter *from)
2747 {
2748 	struct file *file = iocb->ki_filp;
2749 	struct cifsFileInfo *cfile = (struct cifsFileInfo *)file->private_data;
2750 	struct inode *inode = file->f_mapping->host;
2751 	struct cifsInodeInfo *cinode = CIFS_I(inode);
2752 	struct TCP_Server_Info *server = tlink_tcon(cfile->tlink)->ses->server;
2753 	ssize_t rc;
2754 
2755 	rc = netfs_start_io_write(inode);
2756 	if (rc < 0)
2757 		return rc;
2758 
2759 	/*
2760 	 * We need to hold the sem to be sure nobody modifies lock list
2761 	 * with a brlock that prevents writing.
2762 	 */
2763 	down_read(&cinode->lock_sem);
2764 
2765 	rc = generic_write_checks(iocb, from);
2766 	if (rc <= 0)
2767 		goto out;
2768 
2769 	if (!cifs_find_lock_conflict(cfile, iocb->ki_pos, iov_iter_count(from),
2770 				     server->vals->exclusive_lock_type, 0,
2771 				     NULL, CIFS_WRITE_OP))
2772 		rc = netfs_buffered_write_iter_locked(iocb, from, NULL);
2773 	else
2774 		rc = -EACCES;
2775 out:
2776 	up_read(&cinode->lock_sem);
2777 	netfs_end_io_write(inode);
2778 	if (rc > 0)
2779 		rc = generic_write_sync(iocb, rc);
2780 	return rc;
2781 }
2782 
2783 ssize_t
2784 cifs_strict_writev(struct kiocb *iocb, struct iov_iter *from)
2785 {
2786 	struct inode *inode = file_inode(iocb->ki_filp);
2787 	struct cifsInodeInfo *cinode = CIFS_I(inode);
2788 	struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb);
2789 	struct cifsFileInfo *cfile = (struct cifsFileInfo *)
2790 						iocb->ki_filp->private_data;
2791 	struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
2792 	ssize_t written;
2793 
2794 	written = cifs_get_writer(cinode);
2795 	if (written)
2796 		return written;
2797 
2798 	if (CIFS_CACHE_WRITE(cinode)) {
2799 		if (cap_unix(tcon->ses) &&
2800 		    (CIFS_UNIX_FCNTL_CAP & le64_to_cpu(tcon->fsUnixInfo.Capability)) &&
2801 		    ((cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOPOSIXBRL) == 0)) {
2802 			written = netfs_file_write_iter(iocb, from);
2803 			goto out;
2804 		}
2805 		written = cifs_writev(iocb, from);
2806 		goto out;
2807 	}
2808 	/*
2809 	 * For non-oplocked files in strict cache mode we need to write the data
2810 	 * to the server exactly from the pos to pos+len-1 rather than flush all
2811 	 * affected pages because it may cause a error with mandatory locks on
2812 	 * these pages but not on the region from pos to ppos+len-1.
2813 	 */
2814 	written = netfs_file_write_iter(iocb, from);
2815 	if (CIFS_CACHE_READ(cinode)) {
2816 		/*
2817 		 * We have read level caching and we have just sent a write
2818 		 * request to the server thus making data in the cache stale.
2819 		 * Zap the cache and set oplock/lease level to NONE to avoid
2820 		 * reading stale data from the cache. All subsequent read
2821 		 * operations will read new data from the server.
2822 		 */
2823 		cifs_zap_mapping(inode);
2824 		cifs_dbg(FYI, "Set Oplock/Lease to NONE for inode=%p after write\n",
2825 			 inode);
2826 		cinode->oplock = 0;
2827 	}
2828 out:
2829 	cifs_put_writer(cinode);
2830 	return written;
2831 }
2832 
2833 ssize_t cifs_loose_read_iter(struct kiocb *iocb, struct iov_iter *iter)
2834 {
2835 	ssize_t rc;
2836 	struct inode *inode = file_inode(iocb->ki_filp);
2837 
2838 	if (iocb->ki_flags & IOCB_DIRECT)
2839 		return netfs_unbuffered_read_iter(iocb, iter);
2840 
2841 	rc = cifs_revalidate_mapping(inode);
2842 	if (rc)
2843 		return rc;
2844 
2845 	return netfs_file_read_iter(iocb, iter);
2846 }
2847 
2848 ssize_t cifs_file_write_iter(struct kiocb *iocb, struct iov_iter *from)
2849 {
2850 	struct inode *inode = file_inode(iocb->ki_filp);
2851 	struct cifsInodeInfo *cinode = CIFS_I(inode);
2852 	ssize_t written;
2853 	int rc;
2854 
2855 	if (iocb->ki_filp->f_flags & O_DIRECT) {
2856 		written = netfs_unbuffered_write_iter(iocb, from);
2857 		if (written > 0 && CIFS_CACHE_READ(cinode)) {
2858 			cifs_zap_mapping(inode);
2859 			cifs_dbg(FYI,
2860 				 "Set no oplock for inode=%p after a write operation\n",
2861 				 inode);
2862 			cinode->oplock = 0;
2863 		}
2864 		return written;
2865 	}
2866 
2867 	written = cifs_get_writer(cinode);
2868 	if (written)
2869 		return written;
2870 
2871 	written = netfs_file_write_iter(iocb, from);
2872 
2873 	if (!CIFS_CACHE_WRITE(CIFS_I(inode))) {
2874 		rc = filemap_fdatawrite(inode->i_mapping);
2875 		if (rc)
2876 			cifs_dbg(FYI, "cifs_file_write_iter: %d rc on %p inode\n",
2877 				 rc, inode);
2878 	}
2879 
2880 	cifs_put_writer(cinode);
2881 	return written;
2882 }
2883 
2884 ssize_t
2885 cifs_strict_readv(struct kiocb *iocb, struct iov_iter *to)
2886 {
2887 	struct inode *inode = file_inode(iocb->ki_filp);
2888 	struct cifsInodeInfo *cinode = CIFS_I(inode);
2889 	struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb);
2890 	struct cifsFileInfo *cfile = (struct cifsFileInfo *)
2891 						iocb->ki_filp->private_data;
2892 	struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
2893 	int rc = -EACCES;
2894 
2895 	/*
2896 	 * In strict cache mode we need to read from the server all the time
2897 	 * if we don't have level II oplock because the server can delay mtime
2898 	 * change - so we can't make a decision about inode invalidating.
2899 	 * And we can also fail with pagereading if there are mandatory locks
2900 	 * on pages affected by this read but not on the region from pos to
2901 	 * pos+len-1.
2902 	 */
2903 	if (!CIFS_CACHE_READ(cinode))
2904 		return netfs_unbuffered_read_iter(iocb, to);
2905 
2906 	if (cap_unix(tcon->ses) &&
2907 	    (CIFS_UNIX_FCNTL_CAP & le64_to_cpu(tcon->fsUnixInfo.Capability)) &&
2908 	    ((cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOPOSIXBRL) == 0)) {
2909 		if (iocb->ki_flags & IOCB_DIRECT)
2910 			return netfs_unbuffered_read_iter(iocb, to);
2911 		return netfs_buffered_read_iter(iocb, to);
2912 	}
2913 
2914 	/*
2915 	 * We need to hold the sem to be sure nobody modifies lock list
2916 	 * with a brlock that prevents reading.
2917 	 */
2918 	if (iocb->ki_flags & IOCB_DIRECT) {
2919 		rc = netfs_start_io_direct(inode);
2920 		if (rc < 0)
2921 			goto out;
2922 		rc = -EACCES;
2923 		down_read(&cinode->lock_sem);
2924 		if (!cifs_find_lock_conflict(
2925 			    cfile, iocb->ki_pos, iov_iter_count(to),
2926 			    tcon->ses->server->vals->shared_lock_type,
2927 			    0, NULL, CIFS_READ_OP))
2928 			rc = netfs_unbuffered_read_iter_locked(iocb, to);
2929 		up_read(&cinode->lock_sem);
2930 		netfs_end_io_direct(inode);
2931 	} else {
2932 		rc = netfs_start_io_read(inode);
2933 		if (rc < 0)
2934 			goto out;
2935 		rc = -EACCES;
2936 		down_read(&cinode->lock_sem);
2937 		if (!cifs_find_lock_conflict(
2938 			    cfile, iocb->ki_pos, iov_iter_count(to),
2939 			    tcon->ses->server->vals->shared_lock_type,
2940 			    0, NULL, CIFS_READ_OP))
2941 			rc = filemap_read(iocb, to, 0);
2942 		up_read(&cinode->lock_sem);
2943 		netfs_end_io_read(inode);
2944 	}
2945 out:
2946 	return rc;
2947 }
2948 
2949 static vm_fault_t cifs_page_mkwrite(struct vm_fault *vmf)
2950 {
2951 	return netfs_page_mkwrite(vmf, NULL);
2952 }
2953 
2954 static const struct vm_operations_struct cifs_file_vm_ops = {
2955 	.fault = filemap_fault,
2956 	.map_pages = filemap_map_pages,
2957 	.page_mkwrite = cifs_page_mkwrite,
2958 };
2959 
2960 int cifs_file_strict_mmap(struct file *file, struct vm_area_struct *vma)
2961 {
2962 	int xid, rc = 0;
2963 	struct inode *inode = file_inode(file);
2964 
2965 	xid = get_xid();
2966 
2967 	if (!CIFS_CACHE_READ(CIFS_I(inode)))
2968 		rc = cifs_zap_mapping(inode);
2969 	if (!rc)
2970 		rc = generic_file_mmap(file, vma);
2971 	if (!rc)
2972 		vma->vm_ops = &cifs_file_vm_ops;
2973 
2974 	free_xid(xid);
2975 	return rc;
2976 }
2977 
2978 int cifs_file_mmap(struct file *file, struct vm_area_struct *vma)
2979 {
2980 	int rc, xid;
2981 
2982 	xid = get_xid();
2983 
2984 	rc = cifs_revalidate_file(file);
2985 	if (rc)
2986 		cifs_dbg(FYI, "Validation prior to mmap failed, error=%d\n",
2987 			 rc);
2988 	if (!rc)
2989 		rc = generic_file_mmap(file, vma);
2990 	if (!rc)
2991 		vma->vm_ops = &cifs_file_vm_ops;
2992 
2993 	free_xid(xid);
2994 	return rc;
2995 }
2996 
2997 static int is_inode_writable(struct cifsInodeInfo *cifs_inode)
2998 {
2999 	struct cifsFileInfo *open_file;
3000 
3001 	spin_lock(&cifs_inode->open_file_lock);
3002 	list_for_each_entry(open_file, &cifs_inode->openFileList, flist) {
3003 		if (OPEN_FMODE(open_file->f_flags) & FMODE_WRITE) {
3004 			spin_unlock(&cifs_inode->open_file_lock);
3005 			return 1;
3006 		}
3007 	}
3008 	spin_unlock(&cifs_inode->open_file_lock);
3009 	return 0;
3010 }
3011 
3012 /* We do not want to update the file size from server for inodes
3013    open for write - to avoid races with writepage extending
3014    the file - in the future we could consider allowing
3015    refreshing the inode only on increases in the file size
3016    but this is tricky to do without racing with writebehind
3017    page caching in the current Linux kernel design */
3018 bool is_size_safe_to_change(struct cifsInodeInfo *cifsInode, __u64 end_of_file,
3019 			    bool from_readdir)
3020 {
3021 	if (!cifsInode)
3022 		return true;
3023 
3024 	if (is_inode_writable(cifsInode) ||
3025 		((cifsInode->oplock & CIFS_CACHE_RW_FLG) != 0 && from_readdir)) {
3026 		/* This inode is open for write at least once */
3027 		struct cifs_sb_info *cifs_sb;
3028 
3029 		cifs_sb = CIFS_SB(cifsInode->netfs.inode.i_sb);
3030 		if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_DIRECT_IO) {
3031 			/* since no page cache to corrupt on directio
3032 			we can change size safely */
3033 			return true;
3034 		}
3035 
3036 		if (i_size_read(&cifsInode->netfs.inode) < end_of_file)
3037 			return true;
3038 
3039 		return false;
3040 	} else
3041 		return true;
3042 }
3043 
3044 void cifs_oplock_break(struct work_struct *work)
3045 {
3046 	struct cifsFileInfo *cfile = container_of(work, struct cifsFileInfo,
3047 						  oplock_break);
3048 	struct inode *inode = d_inode(cfile->dentry);
3049 	struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb);
3050 	struct cifsInodeInfo *cinode = CIFS_I(inode);
3051 	struct cifs_tcon *tcon;
3052 	struct TCP_Server_Info *server;
3053 	struct tcon_link *tlink;
3054 	int rc = 0;
3055 	bool purge_cache = false, oplock_break_cancelled;
3056 	__u64 persistent_fid, volatile_fid;
3057 	__u16 net_fid;
3058 
3059 	wait_on_bit(&cinode->flags, CIFS_INODE_PENDING_WRITERS,
3060 			TASK_UNINTERRUPTIBLE);
3061 
3062 	tlink = cifs_sb_tlink(cifs_sb);
3063 	if (IS_ERR(tlink))
3064 		goto out;
3065 	tcon = tlink_tcon(tlink);
3066 	server = tcon->ses->server;
3067 
3068 	server->ops->downgrade_oplock(server, cinode, cfile->oplock_level,
3069 				      cfile->oplock_epoch, &purge_cache);
3070 
3071 	if (!CIFS_CACHE_WRITE(cinode) && CIFS_CACHE_READ(cinode) &&
3072 						cifs_has_mand_locks(cinode)) {
3073 		cifs_dbg(FYI, "Reset oplock to None for inode=%p due to mand locks\n",
3074 			 inode);
3075 		cinode->oplock = 0;
3076 	}
3077 
3078 	if (inode && S_ISREG(inode->i_mode)) {
3079 		if (CIFS_CACHE_READ(cinode))
3080 			break_lease(inode, O_RDONLY);
3081 		else
3082 			break_lease(inode, O_WRONLY);
3083 		rc = filemap_fdatawrite(inode->i_mapping);
3084 		if (!CIFS_CACHE_READ(cinode) || purge_cache) {
3085 			rc = filemap_fdatawait(inode->i_mapping);
3086 			mapping_set_error(inode->i_mapping, rc);
3087 			cifs_zap_mapping(inode);
3088 		}
3089 		cifs_dbg(FYI, "Oplock flush inode %p rc %d\n", inode, rc);
3090 		if (CIFS_CACHE_WRITE(cinode))
3091 			goto oplock_break_ack;
3092 	}
3093 
3094 	rc = cifs_push_locks(cfile);
3095 	if (rc)
3096 		cifs_dbg(VFS, "Push locks rc = %d\n", rc);
3097 
3098 oplock_break_ack:
3099 	/*
3100 	 * When oplock break is received and there are no active
3101 	 * file handles but cached, then schedule deferred close immediately.
3102 	 * So, new open will not use cached handle.
3103 	 */
3104 
3105 	if (!CIFS_CACHE_HANDLE(cinode) && !list_empty(&cinode->deferred_closes))
3106 		cifs_close_deferred_file(cinode);
3107 
3108 	persistent_fid = cfile->fid.persistent_fid;
3109 	volatile_fid = cfile->fid.volatile_fid;
3110 	net_fid = cfile->fid.netfid;
3111 	oplock_break_cancelled = cfile->oplock_break_cancelled;
3112 
3113 	_cifsFileInfo_put(cfile, false /* do not wait for ourself */, false);
3114 	/*
3115 	 * MS-SMB2 3.2.5.19.1 and 3.2.5.19.2 (and MS-CIFS 3.2.5.42) do not require
3116 	 * an acknowledgment to be sent when the file has already been closed.
3117 	 */
3118 	spin_lock(&cinode->open_file_lock);
3119 	/* check list empty since can race with kill_sb calling tree disconnect */
3120 	if (!oplock_break_cancelled && !list_empty(&cinode->openFileList)) {
3121 		spin_unlock(&cinode->open_file_lock);
3122 		rc = server->ops->oplock_response(tcon, persistent_fid,
3123 						  volatile_fid, net_fid, cinode);
3124 		cifs_dbg(FYI, "Oplock release rc = %d\n", rc);
3125 	} else
3126 		spin_unlock(&cinode->open_file_lock);
3127 
3128 	cifs_put_tlink(tlink);
3129 out:
3130 	cifs_done_oplock_break(cinode);
3131 }
3132 
3133 static int cifs_swap_activate(struct swap_info_struct *sis,
3134 			      struct file *swap_file, sector_t *span)
3135 {
3136 	struct cifsFileInfo *cfile = swap_file->private_data;
3137 	struct inode *inode = swap_file->f_mapping->host;
3138 	unsigned long blocks;
3139 	long long isize;
3140 
3141 	cifs_dbg(FYI, "swap activate\n");
3142 
3143 	if (!swap_file->f_mapping->a_ops->swap_rw)
3144 		/* Cannot support swap */
3145 		return -EINVAL;
3146 
3147 	spin_lock(&inode->i_lock);
3148 	blocks = inode->i_blocks;
3149 	isize = inode->i_size;
3150 	spin_unlock(&inode->i_lock);
3151 	if (blocks*512 < isize) {
3152 		pr_warn("swap activate: swapfile has holes\n");
3153 		return -EINVAL;
3154 	}
3155 	*span = sis->pages;
3156 
3157 	pr_warn_once("Swap support over SMB3 is experimental\n");
3158 
3159 	/*
3160 	 * TODO: consider adding ACL (or documenting how) to prevent other
3161 	 * users (on this or other systems) from reading it
3162 	 */
3163 
3164 
3165 	/* TODO: add sk_set_memalloc(inet) or similar */
3166 
3167 	if (cfile)
3168 		cfile->swapfile = true;
3169 	/*
3170 	 * TODO: Since file already open, we can't open with DENY_ALL here
3171 	 * but we could add call to grab a byte range lock to prevent others
3172 	 * from reading or writing the file
3173 	 */
3174 
3175 	sis->flags |= SWP_FS_OPS;
3176 	return add_swap_extent(sis, 0, sis->max, 0);
3177 }
3178 
3179 static void cifs_swap_deactivate(struct file *file)
3180 {
3181 	struct cifsFileInfo *cfile = file->private_data;
3182 
3183 	cifs_dbg(FYI, "swap deactivate\n");
3184 
3185 	/* TODO: undo sk_set_memalloc(inet) will eventually be needed */
3186 
3187 	if (cfile)
3188 		cfile->swapfile = false;
3189 
3190 	/* do we need to unpin (or unlock) the file */
3191 }
3192 
3193 /**
3194  * cifs_swap_rw - SMB3 address space operation for swap I/O
3195  * @iocb: target I/O control block
3196  * @iter: I/O buffer
3197  *
3198  * Perform IO to the swap-file.  This is much like direct IO.
3199  */
3200 static int cifs_swap_rw(struct kiocb *iocb, struct iov_iter *iter)
3201 {
3202 	ssize_t ret;
3203 
3204 	if (iov_iter_rw(iter) == READ)
3205 		ret = netfs_unbuffered_read_iter_locked(iocb, iter);
3206 	else
3207 		ret = netfs_unbuffered_write_iter_locked(iocb, iter, NULL);
3208 	if (ret < 0)
3209 		return ret;
3210 	return 0;
3211 }
3212 
3213 const struct address_space_operations cifs_addr_ops = {
3214 	.read_folio	= netfs_read_folio,
3215 	.readahead	= netfs_readahead,
3216 	.writepages	= netfs_writepages,
3217 	.dirty_folio	= netfs_dirty_folio,
3218 	.release_folio	= netfs_release_folio,
3219 	.direct_IO	= noop_direct_IO,
3220 	.invalidate_folio = netfs_invalidate_folio,
3221 	.migrate_folio	= filemap_migrate_folio,
3222 	/*
3223 	 * TODO: investigate and if useful we could add an is_dirty_writeback
3224 	 * helper if needed
3225 	 */
3226 	.swap_activate	= cifs_swap_activate,
3227 	.swap_deactivate = cifs_swap_deactivate,
3228 	.swap_rw = cifs_swap_rw,
3229 };
3230 
3231 /*
3232  * cifs_readahead requires the server to support a buffer large enough to
3233  * contain the header plus one complete page of data.  Otherwise, we need
3234  * to leave cifs_readahead out of the address space operations.
3235  */
3236 const struct address_space_operations cifs_addr_ops_smallbuf = {
3237 	.read_folio	= netfs_read_folio,
3238 	.writepages	= netfs_writepages,
3239 	.dirty_folio	= netfs_dirty_folio,
3240 	.release_folio	= netfs_release_folio,
3241 	.invalidate_folio = netfs_invalidate_folio,
3242 	.migrate_folio	= filemap_migrate_folio,
3243 };
3244