xref: /linux/fs/netfs/locking.c (revision 2697b79a469b68e3ad3640f55284359c1396278d)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * I/O and data path helper functionality.
4  *
5  * Borrowed from NFS Copyright (c) 2016 Trond Myklebust
6  */
7 
8 #include <linux/kernel.h>
9 #include <linux/netfs.h>
10 #include "internal.h"
11 
12 /*
13  * inode_dio_wait_interruptible - wait for outstanding DIO requests to finish
14  * @inode: inode to wait for
15  *
16  * Waits for all pending direct I/O requests to finish so that we can
17  * proceed with a truncate or equivalent operation.
18  *
19  * Must be called under a lock that serializes taking new references
20  * to i_dio_count, usually by inode->i_mutex.
21  */
22 static int inode_dio_wait_interruptible(struct inode *inode)
23 {
24 	if (!atomic_read(&inode->i_dio_count))
25 		return 0;
26 
27 	wait_queue_head_t *wq = bit_waitqueue(&inode->i_state, __I_DIO_WAKEUP);
28 	DEFINE_WAIT_BIT(q, &inode->i_state, __I_DIO_WAKEUP);
29 
30 	for (;;) {
31 		prepare_to_wait(wq, &q.wq_entry, TASK_INTERRUPTIBLE);
32 		if (!atomic_read(&inode->i_dio_count))
33 			break;
34 		if (signal_pending(current))
35 			break;
36 		schedule();
37 	}
38 	finish_wait(wq, &q.wq_entry);
39 
40 	return atomic_read(&inode->i_dio_count) ? -ERESTARTSYS : 0;
41 }
42 
43 /* Call with exclusively locked inode->i_rwsem */
44 static int netfs_block_o_direct(struct netfs_inode *ictx)
45 {
46 	if (!test_bit(NETFS_ICTX_ODIRECT, &ictx->flags))
47 		return 0;
48 	clear_bit(NETFS_ICTX_ODIRECT, &ictx->flags);
49 	return inode_dio_wait_interruptible(&ictx->inode);
50 }
51 
52 /**
53  * netfs_start_io_read - declare the file is being used for buffered reads
54  * @inode: file inode
55  *
56  * Declare that a buffered read operation is about to start, and ensure
57  * that we block all direct I/O.
58  * On exit, the function ensures that the NETFS_ICTX_ODIRECT flag is unset,
59  * and holds a shared lock on inode->i_rwsem to ensure that the flag
60  * cannot be changed.
61  * In practice, this means that buffered read operations are allowed to
62  * execute in parallel, thanks to the shared lock, whereas direct I/O
63  * operations need to wait to grab an exclusive lock in order to set
64  * NETFS_ICTX_ODIRECT.
65  * Note that buffered writes and truncates both take a write lock on
66  * inode->i_rwsem, meaning that those are serialised w.r.t. the reads.
67  */
68 int netfs_start_io_read(struct inode *inode)
69 	__acquires(inode->i_rwsem)
70 {
71 	struct netfs_inode *ictx = netfs_inode(inode);
72 
73 	/* Be an optimist! */
74 	if (down_read_interruptible(&inode->i_rwsem) < 0)
75 		return -ERESTARTSYS;
76 	if (test_bit(NETFS_ICTX_ODIRECT, &ictx->flags) == 0)
77 		return 0;
78 	up_read(&inode->i_rwsem);
79 
80 	/* Slow path.... */
81 	if (down_write_killable(&inode->i_rwsem) < 0)
82 		return -ERESTARTSYS;
83 	if (netfs_block_o_direct(ictx) < 0) {
84 		up_write(&inode->i_rwsem);
85 		return -ERESTARTSYS;
86 	}
87 	downgrade_write(&inode->i_rwsem);
88 	return 0;
89 }
90 EXPORT_SYMBOL(netfs_start_io_read);
91 
92 /**
93  * netfs_end_io_read - declare that the buffered read operation is done
94  * @inode: file inode
95  *
96  * Declare that a buffered read operation is done, and release the shared
97  * lock on inode->i_rwsem.
98  */
99 void netfs_end_io_read(struct inode *inode)
100 	__releases(inode->i_rwsem)
101 {
102 	up_read(&inode->i_rwsem);
103 }
104 EXPORT_SYMBOL(netfs_end_io_read);
105 
106 /**
107  * netfs_start_io_write - declare the file is being used for buffered writes
108  * @inode: file inode
109  *
110  * Declare that a buffered read operation is about to start, and ensure
111  * that we block all direct I/O.
112  */
113 int netfs_start_io_write(struct inode *inode)
114 	__acquires(inode->i_rwsem)
115 {
116 	struct netfs_inode *ictx = netfs_inode(inode);
117 
118 	if (down_write_killable(&inode->i_rwsem) < 0)
119 		return -ERESTARTSYS;
120 	if (netfs_block_o_direct(ictx) < 0) {
121 		up_write(&inode->i_rwsem);
122 		return -ERESTARTSYS;
123 	}
124 	return 0;
125 }
126 EXPORT_SYMBOL(netfs_start_io_write);
127 
128 /**
129  * netfs_end_io_write - declare that the buffered write operation is done
130  * @inode: file inode
131  *
132  * Declare that a buffered write operation is done, and release the
133  * lock on inode->i_rwsem.
134  */
135 void netfs_end_io_write(struct inode *inode)
136 	__releases(inode->i_rwsem)
137 {
138 	up_write(&inode->i_rwsem);
139 }
140 EXPORT_SYMBOL(netfs_end_io_write);
141 
142 /* Call with exclusively locked inode->i_rwsem */
143 static int netfs_block_buffered(struct inode *inode)
144 {
145 	struct netfs_inode *ictx = netfs_inode(inode);
146 	int ret;
147 
148 	if (!test_bit(NETFS_ICTX_ODIRECT, &ictx->flags)) {
149 		set_bit(NETFS_ICTX_ODIRECT, &ictx->flags);
150 		if (inode->i_mapping->nrpages != 0) {
151 			unmap_mapping_range(inode->i_mapping, 0, 0, 0);
152 			ret = filemap_fdatawait(inode->i_mapping);
153 			if (ret < 0) {
154 				clear_bit(NETFS_ICTX_ODIRECT, &ictx->flags);
155 				return ret;
156 			}
157 		}
158 	}
159 	return 0;
160 }
161 
162 /**
163  * netfs_start_io_direct - declare the file is being used for direct i/o
164  * @inode: file inode
165  *
166  * Declare that a direct I/O operation is about to start, and ensure
167  * that we block all buffered I/O.
168  * On exit, the function ensures that the NETFS_ICTX_ODIRECT flag is set,
169  * and holds a shared lock on inode->i_rwsem to ensure that the flag
170  * cannot be changed.
171  * In practice, this means that direct I/O operations are allowed to
172  * execute in parallel, thanks to the shared lock, whereas buffered I/O
173  * operations need to wait to grab an exclusive lock in order to clear
174  * NETFS_ICTX_ODIRECT.
175  * Note that buffered writes and truncates both take a write lock on
176  * inode->i_rwsem, meaning that those are serialised w.r.t. O_DIRECT.
177  */
178 int netfs_start_io_direct(struct inode *inode)
179 	__acquires(inode->i_rwsem)
180 {
181 	struct netfs_inode *ictx = netfs_inode(inode);
182 	int ret;
183 
184 	/* Be an optimist! */
185 	if (down_read_interruptible(&inode->i_rwsem) < 0)
186 		return -ERESTARTSYS;
187 	if (test_bit(NETFS_ICTX_ODIRECT, &ictx->flags) != 0)
188 		return 0;
189 	up_read(&inode->i_rwsem);
190 
191 	/* Slow path.... */
192 	if (down_write_killable(&inode->i_rwsem) < 0)
193 		return -ERESTARTSYS;
194 	ret = netfs_block_buffered(inode);
195 	if (ret < 0) {
196 		up_write(&inode->i_rwsem);
197 		return ret;
198 	}
199 	downgrade_write(&inode->i_rwsem);
200 	return 0;
201 }
202 EXPORT_SYMBOL(netfs_start_io_direct);
203 
204 /**
205  * netfs_end_io_direct - declare that the direct i/o operation is done
206  * @inode: file inode
207  *
208  * Declare that a direct I/O operation is done, and release the shared
209  * lock on inode->i_rwsem.
210  */
211 void netfs_end_io_direct(struct inode *inode)
212 	__releases(inode->i_rwsem)
213 {
214 	up_read(&inode->i_rwsem);
215 }
216 EXPORT_SYMBOL(netfs_end_io_direct);
217