xref: /linux/lib/iov_iter.c (revision 73b25505ce043b561028e5571d84dc82aa53c2b4)
1 // SPDX-License-Identifier: GPL-2.0-only
2 #include <crypto/hash.h>
3 #include <linux/export.h>
4 #include <linux/bvec.h>
5 #include <linux/fault-inject-usercopy.h>
6 #include <linux/uio.h>
7 #include <linux/pagemap.h>
8 #include <linux/highmem.h>
9 #include <linux/slab.h>
10 #include <linux/vmalloc.h>
11 #include <linux/splice.h>
12 #include <linux/compat.h>
13 #include <net/checksum.h>
14 #include <linux/scatterlist.h>
15 #include <linux/instrumented.h>
16 
17 /* covers ubuf and kbuf alike */
18 #define iterate_buf(i, n, base, len, off, __p, STEP) {		\
19 	size_t __maybe_unused off = 0;				\
20 	len = n;						\
21 	base = __p + i->iov_offset;				\
22 	len -= (STEP);						\
23 	i->iov_offset += len;					\
24 	n = len;						\
25 }
26 
27 /* covers iovec and kvec alike */
28 #define iterate_iovec(i, n, base, len, off, __p, STEP) {	\
29 	size_t off = 0;						\
30 	size_t skip = i->iov_offset;				\
31 	do {							\
32 		len = min(n, __p->iov_len - skip);		\
33 		if (likely(len)) {				\
34 			base = __p->iov_base + skip;		\
35 			len -= (STEP);				\
36 			off += len;				\
37 			skip += len;				\
38 			n -= len;				\
39 			if (skip < __p->iov_len)		\
40 				break;				\
41 		}						\
42 		__p++;						\
43 		skip = 0;					\
44 	} while (n);						\
45 	i->iov_offset = skip;					\
46 	n = off;						\
47 }
48 
49 #define iterate_bvec(i, n, base, len, off, p, STEP) {		\
50 	size_t off = 0;						\
51 	unsigned skip = i->iov_offset;				\
52 	while (n) {						\
53 		unsigned offset = p->bv_offset + skip;		\
54 		unsigned left;					\
55 		void *kaddr = kmap_local_page(p->bv_page +	\
56 					offset / PAGE_SIZE);	\
57 		base = kaddr + offset % PAGE_SIZE;		\
58 		len = min(min(n, (size_t)(p->bv_len - skip)),	\
59 		     (size_t)(PAGE_SIZE - offset % PAGE_SIZE));	\
60 		left = (STEP);					\
61 		kunmap_local(kaddr);				\
62 		len -= left;					\
63 		off += len;					\
64 		skip += len;					\
65 		if (skip == p->bv_len) {			\
66 			skip = 0;				\
67 			p++;					\
68 		}						\
69 		n -= len;					\
70 		if (left)					\
71 			break;					\
72 	}							\
73 	i->iov_offset = skip;					\
74 	n = off;						\
75 }
76 
77 #define iterate_xarray(i, n, base, len, __off, STEP) {		\
78 	__label__ __out;					\
79 	size_t __off = 0;					\
80 	struct folio *folio;					\
81 	loff_t start = i->xarray_start + i->iov_offset;		\
82 	pgoff_t index = start / PAGE_SIZE;			\
83 	XA_STATE(xas, i->xarray, index);			\
84 								\
85 	len = PAGE_SIZE - offset_in_page(start);		\
86 	rcu_read_lock();					\
87 	xas_for_each(&xas, folio, ULONG_MAX) {			\
88 		unsigned left;					\
89 		size_t offset;					\
90 		if (xas_retry(&xas, folio))			\
91 			continue;				\
92 		if (WARN_ON(xa_is_value(folio)))		\
93 			break;					\
94 		if (WARN_ON(folio_test_hugetlb(folio)))		\
95 			break;					\
96 		offset = offset_in_folio(folio, start + __off);	\
97 		while (offset < folio_size(folio)) {		\
98 			base = kmap_local_folio(folio, offset);	\
99 			len = min(n, len);			\
100 			left = (STEP);				\
101 			kunmap_local(base);			\
102 			len -= left;				\
103 			__off += len;				\
104 			n -= len;				\
105 			if (left || n == 0)			\
106 				goto __out;			\
107 			offset += len;				\
108 			len = PAGE_SIZE;			\
109 		}						\
110 	}							\
111 __out:								\
112 	rcu_read_unlock();					\
113 	i->iov_offset += __off;					\
114 	n = __off;						\
115 }
116 
117 #define __iterate_and_advance(i, n, base, len, off, I, K) {	\
118 	if (unlikely(i->count < n))				\
119 		n = i->count;					\
120 	if (likely(n)) {					\
121 		if (likely(iter_is_ubuf(i))) {			\
122 			void __user *base;			\
123 			size_t len;				\
124 			iterate_buf(i, n, base, len, off,	\
125 						i->ubuf, (I)) 	\
126 		} else if (likely(iter_is_iovec(i))) {		\
127 			const struct iovec *iov = iter_iov(i);	\
128 			void __user *base;			\
129 			size_t len;				\
130 			iterate_iovec(i, n, base, len, off,	\
131 						iov, (I))	\
132 			i->nr_segs -= iov - iter_iov(i);	\
133 			i->__iov = iov;				\
134 		} else if (iov_iter_is_bvec(i)) {		\
135 			const struct bio_vec *bvec = i->bvec;	\
136 			void *base;				\
137 			size_t len;				\
138 			iterate_bvec(i, n, base, len, off,	\
139 						bvec, (K))	\
140 			i->nr_segs -= bvec - i->bvec;		\
141 			i->bvec = bvec;				\
142 		} else if (iov_iter_is_kvec(i)) {		\
143 			const struct kvec *kvec = i->kvec;	\
144 			void *base;				\
145 			size_t len;				\
146 			iterate_iovec(i, n, base, len, off,	\
147 						kvec, (K))	\
148 			i->nr_segs -= kvec - i->kvec;		\
149 			i->kvec = kvec;				\
150 		} else if (iov_iter_is_xarray(i)) {		\
151 			void *base;				\
152 			size_t len;				\
153 			iterate_xarray(i, n, base, len, off,	\
154 							(K))	\
155 		}						\
156 		i->count -= n;					\
157 	}							\
158 }
159 #define iterate_and_advance(i, n, base, len, off, I, K) \
160 	__iterate_and_advance(i, n, base, len, off, I, ((void)(K),0))
161 
162 static int copyout(void __user *to, const void *from, size_t n)
163 {
164 	if (should_fail_usercopy())
165 		return n;
166 	if (access_ok(to, n)) {
167 		instrument_copy_to_user(to, from, n);
168 		n = raw_copy_to_user(to, from, n);
169 	}
170 	return n;
171 }
172 
173 static int copyout_nofault(void __user *to, const void *from, size_t n)
174 {
175 	long res;
176 
177 	if (should_fail_usercopy())
178 		return n;
179 
180 	res = copy_to_user_nofault(to, from, n);
181 
182 	return res < 0 ? n : res;
183 }
184 
185 static int copyin(void *to, const void __user *from, size_t n)
186 {
187 	size_t res = n;
188 
189 	if (should_fail_usercopy())
190 		return n;
191 	if (access_ok(from, n)) {
192 		instrument_copy_from_user_before(to, from, n);
193 		res = raw_copy_from_user(to, from, n);
194 		instrument_copy_from_user_after(to, from, n, res);
195 	}
196 	return res;
197 }
198 
199 /*
200  * fault_in_iov_iter_readable - fault in iov iterator for reading
201  * @i: iterator
202  * @size: maximum length
203  *
204  * Fault in one or more iovecs of the given iov_iter, to a maximum length of
205  * @size.  For each iovec, fault in each page that constitutes the iovec.
206  *
207  * Returns the number of bytes not faulted in (like copy_to_user() and
208  * copy_from_user()).
209  *
210  * Always returns 0 for non-userspace iterators.
211  */
212 size_t fault_in_iov_iter_readable(const struct iov_iter *i, size_t size)
213 {
214 	if (iter_is_ubuf(i)) {
215 		size_t n = min(size, iov_iter_count(i));
216 		n -= fault_in_readable(i->ubuf + i->iov_offset, n);
217 		return size - n;
218 	} else if (iter_is_iovec(i)) {
219 		size_t count = min(size, iov_iter_count(i));
220 		const struct iovec *p;
221 		size_t skip;
222 
223 		size -= count;
224 		for (p = iter_iov(i), skip = i->iov_offset; count; p++, skip = 0) {
225 			size_t len = min(count, p->iov_len - skip);
226 			size_t ret;
227 
228 			if (unlikely(!len))
229 				continue;
230 			ret = fault_in_readable(p->iov_base + skip, len);
231 			count -= len - ret;
232 			if (ret)
233 				break;
234 		}
235 		return count + size;
236 	}
237 	return 0;
238 }
239 EXPORT_SYMBOL(fault_in_iov_iter_readable);
240 
241 /*
242  * fault_in_iov_iter_writeable - fault in iov iterator for writing
243  * @i: iterator
244  * @size: maximum length
245  *
246  * Faults in the iterator using get_user_pages(), i.e., without triggering
247  * hardware page faults.  This is primarily useful when we already know that
248  * some or all of the pages in @i aren't in memory.
249  *
250  * Returns the number of bytes not faulted in, like copy_to_user() and
251  * copy_from_user().
252  *
253  * Always returns 0 for non-user-space iterators.
254  */
255 size_t fault_in_iov_iter_writeable(const struct iov_iter *i, size_t size)
256 {
257 	if (iter_is_ubuf(i)) {
258 		size_t n = min(size, iov_iter_count(i));
259 		n -= fault_in_safe_writeable(i->ubuf + i->iov_offset, n);
260 		return size - n;
261 	} else if (iter_is_iovec(i)) {
262 		size_t count = min(size, iov_iter_count(i));
263 		const struct iovec *p;
264 		size_t skip;
265 
266 		size -= count;
267 		for (p = iter_iov(i), skip = i->iov_offset; count; p++, skip = 0) {
268 			size_t len = min(count, p->iov_len - skip);
269 			size_t ret;
270 
271 			if (unlikely(!len))
272 				continue;
273 			ret = fault_in_safe_writeable(p->iov_base + skip, len);
274 			count -= len - ret;
275 			if (ret)
276 				break;
277 		}
278 		return count + size;
279 	}
280 	return 0;
281 }
282 EXPORT_SYMBOL(fault_in_iov_iter_writeable);
283 
284 void iov_iter_init(struct iov_iter *i, unsigned int direction,
285 			const struct iovec *iov, unsigned long nr_segs,
286 			size_t count)
287 {
288 	WARN_ON(direction & ~(READ | WRITE));
289 	*i = (struct iov_iter) {
290 		.iter_type = ITER_IOVEC,
291 		.copy_mc = false,
292 		.nofault = false,
293 		.user_backed = true,
294 		.data_source = direction,
295 		.__iov = iov,
296 		.nr_segs = nr_segs,
297 		.iov_offset = 0,
298 		.count = count
299 	};
300 }
301 EXPORT_SYMBOL(iov_iter_init);
302 
303 static __wsum csum_and_memcpy(void *to, const void *from, size_t len,
304 			      __wsum sum, size_t off)
305 {
306 	__wsum next = csum_partial_copy_nocheck(from, to, len);
307 	return csum_block_add(sum, next, off);
308 }
309 
310 size_t _copy_to_iter(const void *addr, size_t bytes, struct iov_iter *i)
311 {
312 	if (WARN_ON_ONCE(i->data_source))
313 		return 0;
314 	if (user_backed_iter(i))
315 		might_fault();
316 	iterate_and_advance(i, bytes, base, len, off,
317 		copyout(base, addr + off, len),
318 		memcpy(base, addr + off, len)
319 	)
320 
321 	return bytes;
322 }
323 EXPORT_SYMBOL(_copy_to_iter);
324 
325 #ifdef CONFIG_ARCH_HAS_COPY_MC
326 static int copyout_mc(void __user *to, const void *from, size_t n)
327 {
328 	if (access_ok(to, n)) {
329 		instrument_copy_to_user(to, from, n);
330 		n = copy_mc_to_user((__force void *) to, from, n);
331 	}
332 	return n;
333 }
334 
335 /**
336  * _copy_mc_to_iter - copy to iter with source memory error exception handling
337  * @addr: source kernel address
338  * @bytes: total transfer length
339  * @i: destination iterator
340  *
341  * The pmem driver deploys this for the dax operation
342  * (dax_copy_to_iter()) for dax reads (bypass page-cache and the
343  * block-layer). Upon #MC read(2) aborts and returns EIO or the bytes
344  * successfully copied.
345  *
346  * The main differences between this and typical _copy_to_iter().
347  *
348  * * Typical tail/residue handling after a fault retries the copy
349  *   byte-by-byte until the fault happens again. Re-triggering machine
350  *   checks is potentially fatal so the implementation uses source
351  *   alignment and poison alignment assumptions to avoid re-triggering
352  *   hardware exceptions.
353  *
354  * * ITER_KVEC and ITER_BVEC can return short copies.  Compare to
355  *   copy_to_iter() where only ITER_IOVEC attempts might return a short copy.
356  *
357  * Return: number of bytes copied (may be %0)
358  */
359 size_t _copy_mc_to_iter(const void *addr, size_t bytes, struct iov_iter *i)
360 {
361 	if (WARN_ON_ONCE(i->data_source))
362 		return 0;
363 	if (user_backed_iter(i))
364 		might_fault();
365 	__iterate_and_advance(i, bytes, base, len, off,
366 		copyout_mc(base, addr + off, len),
367 		copy_mc_to_kernel(base, addr + off, len)
368 	)
369 
370 	return bytes;
371 }
372 EXPORT_SYMBOL_GPL(_copy_mc_to_iter);
373 #endif /* CONFIG_ARCH_HAS_COPY_MC */
374 
375 static void *memcpy_from_iter(struct iov_iter *i, void *to, const void *from,
376 				 size_t size)
377 {
378 	if (iov_iter_is_copy_mc(i))
379 		return (void *)copy_mc_to_kernel(to, from, size);
380 	return memcpy(to, from, size);
381 }
382 
383 size_t _copy_from_iter(void *addr, size_t bytes, struct iov_iter *i)
384 {
385 	if (WARN_ON_ONCE(!i->data_source))
386 		return 0;
387 
388 	if (user_backed_iter(i))
389 		might_fault();
390 	iterate_and_advance(i, bytes, base, len, off,
391 		copyin(addr + off, base, len),
392 		memcpy_from_iter(i, addr + off, base, len)
393 	)
394 
395 	return bytes;
396 }
397 EXPORT_SYMBOL(_copy_from_iter);
398 
399 size_t _copy_from_iter_nocache(void *addr, size_t bytes, struct iov_iter *i)
400 {
401 	if (WARN_ON_ONCE(!i->data_source))
402 		return 0;
403 
404 	iterate_and_advance(i, bytes, base, len, off,
405 		__copy_from_user_inatomic_nocache(addr + off, base, len),
406 		memcpy(addr + off, base, len)
407 	)
408 
409 	return bytes;
410 }
411 EXPORT_SYMBOL(_copy_from_iter_nocache);
412 
413 #ifdef CONFIG_ARCH_HAS_UACCESS_FLUSHCACHE
414 /**
415  * _copy_from_iter_flushcache - write destination through cpu cache
416  * @addr: destination kernel address
417  * @bytes: total transfer length
418  * @i: source iterator
419  *
420  * The pmem driver arranges for filesystem-dax to use this facility via
421  * dax_copy_from_iter() for ensuring that writes to persistent memory
422  * are flushed through the CPU cache. It is differentiated from
423  * _copy_from_iter_nocache() in that guarantees all data is flushed for
424  * all iterator types. The _copy_from_iter_nocache() only attempts to
425  * bypass the cache for the ITER_IOVEC case, and on some archs may use
426  * instructions that strand dirty-data in the cache.
427  *
428  * Return: number of bytes copied (may be %0)
429  */
430 size_t _copy_from_iter_flushcache(void *addr, size_t bytes, struct iov_iter *i)
431 {
432 	if (WARN_ON_ONCE(!i->data_source))
433 		return 0;
434 
435 	iterate_and_advance(i, bytes, base, len, off,
436 		__copy_from_user_flushcache(addr + off, base, len),
437 		memcpy_flushcache(addr + off, base, len)
438 	)
439 
440 	return bytes;
441 }
442 EXPORT_SYMBOL_GPL(_copy_from_iter_flushcache);
443 #endif
444 
445 static inline bool page_copy_sane(struct page *page, size_t offset, size_t n)
446 {
447 	struct page *head;
448 	size_t v = n + offset;
449 
450 	/*
451 	 * The general case needs to access the page order in order
452 	 * to compute the page size.
453 	 * However, we mostly deal with order-0 pages and thus can
454 	 * avoid a possible cache line miss for requests that fit all
455 	 * page orders.
456 	 */
457 	if (n <= v && v <= PAGE_SIZE)
458 		return true;
459 
460 	head = compound_head(page);
461 	v += (page - head) << PAGE_SHIFT;
462 
463 	if (WARN_ON(n > v || v > page_size(head)))
464 		return false;
465 	return true;
466 }
467 
468 size_t copy_page_to_iter(struct page *page, size_t offset, size_t bytes,
469 			 struct iov_iter *i)
470 {
471 	size_t res = 0;
472 	if (!page_copy_sane(page, offset, bytes))
473 		return 0;
474 	if (WARN_ON_ONCE(i->data_source))
475 		return 0;
476 	page += offset / PAGE_SIZE; // first subpage
477 	offset %= PAGE_SIZE;
478 	while (1) {
479 		void *kaddr = kmap_local_page(page);
480 		size_t n = min(bytes, (size_t)PAGE_SIZE - offset);
481 		n = _copy_to_iter(kaddr + offset, n, i);
482 		kunmap_local(kaddr);
483 		res += n;
484 		bytes -= n;
485 		if (!bytes || !n)
486 			break;
487 		offset += n;
488 		if (offset == PAGE_SIZE) {
489 			page++;
490 			offset = 0;
491 		}
492 	}
493 	return res;
494 }
495 EXPORT_SYMBOL(copy_page_to_iter);
496 
497 size_t copy_page_to_iter_nofault(struct page *page, unsigned offset, size_t bytes,
498 				 struct iov_iter *i)
499 {
500 	size_t res = 0;
501 
502 	if (!page_copy_sane(page, offset, bytes))
503 		return 0;
504 	if (WARN_ON_ONCE(i->data_source))
505 		return 0;
506 	page += offset / PAGE_SIZE; // first subpage
507 	offset %= PAGE_SIZE;
508 	while (1) {
509 		void *kaddr = kmap_local_page(page);
510 		size_t n = min(bytes, (size_t)PAGE_SIZE - offset);
511 
512 		iterate_and_advance(i, n, base, len, off,
513 			copyout_nofault(base, kaddr + offset + off, len),
514 			memcpy(base, kaddr + offset + off, len)
515 		)
516 		kunmap_local(kaddr);
517 		res += n;
518 		bytes -= n;
519 		if (!bytes || !n)
520 			break;
521 		offset += n;
522 		if (offset == PAGE_SIZE) {
523 			page++;
524 			offset = 0;
525 		}
526 	}
527 	return res;
528 }
529 EXPORT_SYMBOL(copy_page_to_iter_nofault);
530 
531 size_t copy_page_from_iter(struct page *page, size_t offset, size_t bytes,
532 			 struct iov_iter *i)
533 {
534 	size_t res = 0;
535 	if (!page_copy_sane(page, offset, bytes))
536 		return 0;
537 	page += offset / PAGE_SIZE; // first subpage
538 	offset %= PAGE_SIZE;
539 	while (1) {
540 		void *kaddr = kmap_local_page(page);
541 		size_t n = min(bytes, (size_t)PAGE_SIZE - offset);
542 		n = _copy_from_iter(kaddr + offset, n, i);
543 		kunmap_local(kaddr);
544 		res += n;
545 		bytes -= n;
546 		if (!bytes || !n)
547 			break;
548 		offset += n;
549 		if (offset == PAGE_SIZE) {
550 			page++;
551 			offset = 0;
552 		}
553 	}
554 	return res;
555 }
556 EXPORT_SYMBOL(copy_page_from_iter);
557 
558 size_t iov_iter_zero(size_t bytes, struct iov_iter *i)
559 {
560 	iterate_and_advance(i, bytes, base, len, count,
561 		clear_user(base, len),
562 		memset(base, 0, len)
563 	)
564 
565 	return bytes;
566 }
567 EXPORT_SYMBOL(iov_iter_zero);
568 
569 size_t copy_page_from_iter_atomic(struct page *page, size_t offset,
570 		size_t bytes, struct iov_iter *i)
571 {
572 	size_t n, copied = 0;
573 
574 	if (!page_copy_sane(page, offset, bytes))
575 		return 0;
576 	if (WARN_ON_ONCE(!i->data_source))
577 		return 0;
578 
579 	do {
580 		char *p;
581 
582 		n = bytes - copied;
583 		if (PageHighMem(page)) {
584 			page += offset / PAGE_SIZE;
585 			offset %= PAGE_SIZE;
586 			n = min_t(size_t, n, PAGE_SIZE - offset);
587 		}
588 
589 		p = kmap_atomic(page) + offset;
590 		iterate_and_advance(i, n, base, len, off,
591 			copyin(p + off, base, len),
592 			memcpy_from_iter(i, p + off, base, len)
593 		)
594 		kunmap_atomic(p);
595 		copied += n;
596 		offset += n;
597 	} while (PageHighMem(page) && copied != bytes && n > 0);
598 
599 	return copied;
600 }
601 EXPORT_SYMBOL(copy_page_from_iter_atomic);
602 
603 static void iov_iter_bvec_advance(struct iov_iter *i, size_t size)
604 {
605 	const struct bio_vec *bvec, *end;
606 
607 	if (!i->count)
608 		return;
609 	i->count -= size;
610 
611 	size += i->iov_offset;
612 
613 	for (bvec = i->bvec, end = bvec + i->nr_segs; bvec < end; bvec++) {
614 		if (likely(size < bvec->bv_len))
615 			break;
616 		size -= bvec->bv_len;
617 	}
618 	i->iov_offset = size;
619 	i->nr_segs -= bvec - i->bvec;
620 	i->bvec = bvec;
621 }
622 
623 static void iov_iter_iovec_advance(struct iov_iter *i, size_t size)
624 {
625 	const struct iovec *iov, *end;
626 
627 	if (!i->count)
628 		return;
629 	i->count -= size;
630 
631 	size += i->iov_offset; // from beginning of current segment
632 	for (iov = iter_iov(i), end = iov + i->nr_segs; iov < end; iov++) {
633 		if (likely(size < iov->iov_len))
634 			break;
635 		size -= iov->iov_len;
636 	}
637 	i->iov_offset = size;
638 	i->nr_segs -= iov - iter_iov(i);
639 	i->__iov = iov;
640 }
641 
642 void iov_iter_advance(struct iov_iter *i, size_t size)
643 {
644 	if (unlikely(i->count < size))
645 		size = i->count;
646 	if (likely(iter_is_ubuf(i)) || unlikely(iov_iter_is_xarray(i))) {
647 		i->iov_offset += size;
648 		i->count -= size;
649 	} else if (likely(iter_is_iovec(i) || iov_iter_is_kvec(i))) {
650 		/* iovec and kvec have identical layouts */
651 		iov_iter_iovec_advance(i, size);
652 	} else if (iov_iter_is_bvec(i)) {
653 		iov_iter_bvec_advance(i, size);
654 	} else if (iov_iter_is_discard(i)) {
655 		i->count -= size;
656 	}
657 }
658 EXPORT_SYMBOL(iov_iter_advance);
659 
660 void iov_iter_revert(struct iov_iter *i, size_t unroll)
661 {
662 	if (!unroll)
663 		return;
664 	if (WARN_ON(unroll > MAX_RW_COUNT))
665 		return;
666 	i->count += unroll;
667 	if (unlikely(iov_iter_is_discard(i)))
668 		return;
669 	if (unroll <= i->iov_offset) {
670 		i->iov_offset -= unroll;
671 		return;
672 	}
673 	unroll -= i->iov_offset;
674 	if (iov_iter_is_xarray(i) || iter_is_ubuf(i)) {
675 		BUG(); /* We should never go beyond the start of the specified
676 			* range since we might then be straying into pages that
677 			* aren't pinned.
678 			*/
679 	} else if (iov_iter_is_bvec(i)) {
680 		const struct bio_vec *bvec = i->bvec;
681 		while (1) {
682 			size_t n = (--bvec)->bv_len;
683 			i->nr_segs++;
684 			if (unroll <= n) {
685 				i->bvec = bvec;
686 				i->iov_offset = n - unroll;
687 				return;
688 			}
689 			unroll -= n;
690 		}
691 	} else { /* same logics for iovec and kvec */
692 		const struct iovec *iov = iter_iov(i);
693 		while (1) {
694 			size_t n = (--iov)->iov_len;
695 			i->nr_segs++;
696 			if (unroll <= n) {
697 				i->__iov = iov;
698 				i->iov_offset = n - unroll;
699 				return;
700 			}
701 			unroll -= n;
702 		}
703 	}
704 }
705 EXPORT_SYMBOL(iov_iter_revert);
706 
707 /*
708  * Return the count of just the current iov_iter segment.
709  */
710 size_t iov_iter_single_seg_count(const struct iov_iter *i)
711 {
712 	if (i->nr_segs > 1) {
713 		if (likely(iter_is_iovec(i) || iov_iter_is_kvec(i)))
714 			return min(i->count, iter_iov(i)->iov_len - i->iov_offset);
715 		if (iov_iter_is_bvec(i))
716 			return min(i->count, i->bvec->bv_len - i->iov_offset);
717 	}
718 	return i->count;
719 }
720 EXPORT_SYMBOL(iov_iter_single_seg_count);
721 
722 void iov_iter_kvec(struct iov_iter *i, unsigned int direction,
723 			const struct kvec *kvec, unsigned long nr_segs,
724 			size_t count)
725 {
726 	WARN_ON(direction & ~(READ | WRITE));
727 	*i = (struct iov_iter){
728 		.iter_type = ITER_KVEC,
729 		.copy_mc = false,
730 		.data_source = direction,
731 		.kvec = kvec,
732 		.nr_segs = nr_segs,
733 		.iov_offset = 0,
734 		.count = count
735 	};
736 }
737 EXPORT_SYMBOL(iov_iter_kvec);
738 
739 void iov_iter_bvec(struct iov_iter *i, unsigned int direction,
740 			const struct bio_vec *bvec, unsigned long nr_segs,
741 			size_t count)
742 {
743 	WARN_ON(direction & ~(READ | WRITE));
744 	*i = (struct iov_iter){
745 		.iter_type = ITER_BVEC,
746 		.copy_mc = false,
747 		.data_source = direction,
748 		.bvec = bvec,
749 		.nr_segs = nr_segs,
750 		.iov_offset = 0,
751 		.count = count
752 	};
753 }
754 EXPORT_SYMBOL(iov_iter_bvec);
755 
756 /**
757  * iov_iter_xarray - Initialise an I/O iterator to use the pages in an xarray
758  * @i: The iterator to initialise.
759  * @direction: The direction of the transfer.
760  * @xarray: The xarray to access.
761  * @start: The start file position.
762  * @count: The size of the I/O buffer in bytes.
763  *
764  * Set up an I/O iterator to either draw data out of the pages attached to an
765  * inode or to inject data into those pages.  The pages *must* be prevented
766  * from evaporation, either by taking a ref on them or locking them by the
767  * caller.
768  */
769 void iov_iter_xarray(struct iov_iter *i, unsigned int direction,
770 		     struct xarray *xarray, loff_t start, size_t count)
771 {
772 	BUG_ON(direction & ~1);
773 	*i = (struct iov_iter) {
774 		.iter_type = ITER_XARRAY,
775 		.copy_mc = false,
776 		.data_source = direction,
777 		.xarray = xarray,
778 		.xarray_start = start,
779 		.count = count,
780 		.iov_offset = 0
781 	};
782 }
783 EXPORT_SYMBOL(iov_iter_xarray);
784 
785 /**
786  * iov_iter_discard - Initialise an I/O iterator that discards data
787  * @i: The iterator to initialise.
788  * @direction: The direction of the transfer.
789  * @count: The size of the I/O buffer in bytes.
790  *
791  * Set up an I/O iterator that just discards everything that's written to it.
792  * It's only available as a READ iterator.
793  */
794 void iov_iter_discard(struct iov_iter *i, unsigned int direction, size_t count)
795 {
796 	BUG_ON(direction != READ);
797 	*i = (struct iov_iter){
798 		.iter_type = ITER_DISCARD,
799 		.copy_mc = false,
800 		.data_source = false,
801 		.count = count,
802 		.iov_offset = 0
803 	};
804 }
805 EXPORT_SYMBOL(iov_iter_discard);
806 
807 static bool iov_iter_aligned_iovec(const struct iov_iter *i, unsigned addr_mask,
808 				   unsigned len_mask)
809 {
810 	size_t size = i->count;
811 	size_t skip = i->iov_offset;
812 	unsigned k;
813 
814 	for (k = 0; k < i->nr_segs; k++, skip = 0) {
815 		const struct iovec *iov = iter_iov(i) + k;
816 		size_t len = iov->iov_len - skip;
817 
818 		if (len > size)
819 			len = size;
820 		if (len & len_mask)
821 			return false;
822 		if ((unsigned long)(iov->iov_base + skip) & addr_mask)
823 			return false;
824 
825 		size -= len;
826 		if (!size)
827 			break;
828 	}
829 	return true;
830 }
831 
832 static bool iov_iter_aligned_bvec(const struct iov_iter *i, unsigned addr_mask,
833 				  unsigned len_mask)
834 {
835 	size_t size = i->count;
836 	unsigned skip = i->iov_offset;
837 	unsigned k;
838 
839 	for (k = 0; k < i->nr_segs; k++, skip = 0) {
840 		size_t len = i->bvec[k].bv_len - skip;
841 
842 		if (len > size)
843 			len = size;
844 		if (len & len_mask)
845 			return false;
846 		if ((unsigned long)(i->bvec[k].bv_offset + skip) & addr_mask)
847 			return false;
848 
849 		size -= len;
850 		if (!size)
851 			break;
852 	}
853 	return true;
854 }
855 
856 /**
857  * iov_iter_is_aligned() - Check if the addresses and lengths of each segments
858  * 	are aligned to the parameters.
859  *
860  * @i: &struct iov_iter to restore
861  * @addr_mask: bit mask to check against the iov element's addresses
862  * @len_mask: bit mask to check against the iov element's lengths
863  *
864  * Return: false if any addresses or lengths intersect with the provided masks
865  */
866 bool iov_iter_is_aligned(const struct iov_iter *i, unsigned addr_mask,
867 			 unsigned len_mask)
868 {
869 	if (likely(iter_is_ubuf(i))) {
870 		if (i->count & len_mask)
871 			return false;
872 		if ((unsigned long)(i->ubuf + i->iov_offset) & addr_mask)
873 			return false;
874 		return true;
875 	}
876 
877 	if (likely(iter_is_iovec(i) || iov_iter_is_kvec(i)))
878 		return iov_iter_aligned_iovec(i, addr_mask, len_mask);
879 
880 	if (iov_iter_is_bvec(i))
881 		return iov_iter_aligned_bvec(i, addr_mask, len_mask);
882 
883 	if (iov_iter_is_xarray(i)) {
884 		if (i->count & len_mask)
885 			return false;
886 		if ((i->xarray_start + i->iov_offset) & addr_mask)
887 			return false;
888 	}
889 
890 	return true;
891 }
892 EXPORT_SYMBOL_GPL(iov_iter_is_aligned);
893 
894 static unsigned long iov_iter_alignment_iovec(const struct iov_iter *i)
895 {
896 	unsigned long res = 0;
897 	size_t size = i->count;
898 	size_t skip = i->iov_offset;
899 	unsigned k;
900 
901 	for (k = 0; k < i->nr_segs; k++, skip = 0) {
902 		const struct iovec *iov = iter_iov(i) + k;
903 		size_t len = iov->iov_len - skip;
904 		if (len) {
905 			res |= (unsigned long)iov->iov_base + skip;
906 			if (len > size)
907 				len = size;
908 			res |= len;
909 			size -= len;
910 			if (!size)
911 				break;
912 		}
913 	}
914 	return res;
915 }
916 
917 static unsigned long iov_iter_alignment_bvec(const struct iov_iter *i)
918 {
919 	unsigned res = 0;
920 	size_t size = i->count;
921 	unsigned skip = i->iov_offset;
922 	unsigned k;
923 
924 	for (k = 0; k < i->nr_segs; k++, skip = 0) {
925 		size_t len = i->bvec[k].bv_len - skip;
926 		res |= (unsigned long)i->bvec[k].bv_offset + skip;
927 		if (len > size)
928 			len = size;
929 		res |= len;
930 		size -= len;
931 		if (!size)
932 			break;
933 	}
934 	return res;
935 }
936 
937 unsigned long iov_iter_alignment(const struct iov_iter *i)
938 {
939 	if (likely(iter_is_ubuf(i))) {
940 		size_t size = i->count;
941 		if (size)
942 			return ((unsigned long)i->ubuf + i->iov_offset) | size;
943 		return 0;
944 	}
945 
946 	/* iovec and kvec have identical layouts */
947 	if (likely(iter_is_iovec(i) || iov_iter_is_kvec(i)))
948 		return iov_iter_alignment_iovec(i);
949 
950 	if (iov_iter_is_bvec(i))
951 		return iov_iter_alignment_bvec(i);
952 
953 	if (iov_iter_is_xarray(i))
954 		return (i->xarray_start + i->iov_offset) | i->count;
955 
956 	return 0;
957 }
958 EXPORT_SYMBOL(iov_iter_alignment);
959 
960 unsigned long iov_iter_gap_alignment(const struct iov_iter *i)
961 {
962 	unsigned long res = 0;
963 	unsigned long v = 0;
964 	size_t size = i->count;
965 	unsigned k;
966 
967 	if (iter_is_ubuf(i))
968 		return 0;
969 
970 	if (WARN_ON(!iter_is_iovec(i)))
971 		return ~0U;
972 
973 	for (k = 0; k < i->nr_segs; k++) {
974 		const struct iovec *iov = iter_iov(i) + k;
975 		if (iov->iov_len) {
976 			unsigned long base = (unsigned long)iov->iov_base;
977 			if (v) // if not the first one
978 				res |= base | v; // this start | previous end
979 			v = base + iov->iov_len;
980 			if (size <= iov->iov_len)
981 				break;
982 			size -= iov->iov_len;
983 		}
984 	}
985 	return res;
986 }
987 EXPORT_SYMBOL(iov_iter_gap_alignment);
988 
989 static int want_pages_array(struct page ***res, size_t size,
990 			    size_t start, unsigned int maxpages)
991 {
992 	unsigned int count = DIV_ROUND_UP(size + start, PAGE_SIZE);
993 
994 	if (count > maxpages)
995 		count = maxpages;
996 	WARN_ON(!count);	// caller should've prevented that
997 	if (!*res) {
998 		*res = kvmalloc_array(count, sizeof(struct page *), GFP_KERNEL);
999 		if (!*res)
1000 			return 0;
1001 	}
1002 	return count;
1003 }
1004 
1005 static ssize_t iter_xarray_populate_pages(struct page **pages, struct xarray *xa,
1006 					  pgoff_t index, unsigned int nr_pages)
1007 {
1008 	XA_STATE(xas, xa, index);
1009 	struct page *page;
1010 	unsigned int ret = 0;
1011 
1012 	rcu_read_lock();
1013 	for (page = xas_load(&xas); page; page = xas_next(&xas)) {
1014 		if (xas_retry(&xas, page))
1015 			continue;
1016 
1017 		/* Has the page moved or been split? */
1018 		if (unlikely(page != xas_reload(&xas))) {
1019 			xas_reset(&xas);
1020 			continue;
1021 		}
1022 
1023 		pages[ret] = find_subpage(page, xas.xa_index);
1024 		get_page(pages[ret]);
1025 		if (++ret == nr_pages)
1026 			break;
1027 	}
1028 	rcu_read_unlock();
1029 	return ret;
1030 }
1031 
1032 static ssize_t iter_xarray_get_pages(struct iov_iter *i,
1033 				     struct page ***pages, size_t maxsize,
1034 				     unsigned maxpages, size_t *_start_offset)
1035 {
1036 	unsigned nr, offset, count;
1037 	pgoff_t index;
1038 	loff_t pos;
1039 
1040 	pos = i->xarray_start + i->iov_offset;
1041 	index = pos >> PAGE_SHIFT;
1042 	offset = pos & ~PAGE_MASK;
1043 	*_start_offset = offset;
1044 
1045 	count = want_pages_array(pages, maxsize, offset, maxpages);
1046 	if (!count)
1047 		return -ENOMEM;
1048 	nr = iter_xarray_populate_pages(*pages, i->xarray, index, count);
1049 	if (nr == 0)
1050 		return 0;
1051 
1052 	maxsize = min_t(size_t, nr * PAGE_SIZE - offset, maxsize);
1053 	i->iov_offset += maxsize;
1054 	i->count -= maxsize;
1055 	return maxsize;
1056 }
1057 
1058 /* must be done on non-empty ITER_UBUF or ITER_IOVEC one */
1059 static unsigned long first_iovec_segment(const struct iov_iter *i, size_t *size)
1060 {
1061 	size_t skip;
1062 	long k;
1063 
1064 	if (iter_is_ubuf(i))
1065 		return (unsigned long)i->ubuf + i->iov_offset;
1066 
1067 	for (k = 0, skip = i->iov_offset; k < i->nr_segs; k++, skip = 0) {
1068 		const struct iovec *iov = iter_iov(i) + k;
1069 		size_t len = iov->iov_len - skip;
1070 
1071 		if (unlikely(!len))
1072 			continue;
1073 		if (*size > len)
1074 			*size = len;
1075 		return (unsigned long)iov->iov_base + skip;
1076 	}
1077 	BUG(); // if it had been empty, we wouldn't get called
1078 }
1079 
1080 /* must be done on non-empty ITER_BVEC one */
1081 static struct page *first_bvec_segment(const struct iov_iter *i,
1082 				       size_t *size, size_t *start)
1083 {
1084 	struct page *page;
1085 	size_t skip = i->iov_offset, len;
1086 
1087 	len = i->bvec->bv_len - skip;
1088 	if (*size > len)
1089 		*size = len;
1090 	skip += i->bvec->bv_offset;
1091 	page = i->bvec->bv_page + skip / PAGE_SIZE;
1092 	*start = skip % PAGE_SIZE;
1093 	return page;
1094 }
1095 
1096 static ssize_t __iov_iter_get_pages_alloc(struct iov_iter *i,
1097 		   struct page ***pages, size_t maxsize,
1098 		   unsigned int maxpages, size_t *start)
1099 {
1100 	unsigned int n, gup_flags = 0;
1101 
1102 	if (maxsize > i->count)
1103 		maxsize = i->count;
1104 	if (!maxsize)
1105 		return 0;
1106 	if (maxsize > MAX_RW_COUNT)
1107 		maxsize = MAX_RW_COUNT;
1108 
1109 	if (likely(user_backed_iter(i))) {
1110 		unsigned long addr;
1111 		int res;
1112 
1113 		if (iov_iter_rw(i) != WRITE)
1114 			gup_flags |= FOLL_WRITE;
1115 		if (i->nofault)
1116 			gup_flags |= FOLL_NOFAULT;
1117 
1118 		addr = first_iovec_segment(i, &maxsize);
1119 		*start = addr % PAGE_SIZE;
1120 		addr &= PAGE_MASK;
1121 		n = want_pages_array(pages, maxsize, *start, maxpages);
1122 		if (!n)
1123 			return -ENOMEM;
1124 		res = get_user_pages_fast(addr, n, gup_flags, *pages);
1125 		if (unlikely(res <= 0))
1126 			return res;
1127 		maxsize = min_t(size_t, maxsize, res * PAGE_SIZE - *start);
1128 		iov_iter_advance(i, maxsize);
1129 		return maxsize;
1130 	}
1131 	if (iov_iter_is_bvec(i)) {
1132 		struct page **p;
1133 		struct page *page;
1134 
1135 		page = first_bvec_segment(i, &maxsize, start);
1136 		n = want_pages_array(pages, maxsize, *start, maxpages);
1137 		if (!n)
1138 			return -ENOMEM;
1139 		p = *pages;
1140 		for (int k = 0; k < n; k++)
1141 			get_page(p[k] = page + k);
1142 		maxsize = min_t(size_t, maxsize, n * PAGE_SIZE - *start);
1143 		i->count -= maxsize;
1144 		i->iov_offset += maxsize;
1145 		if (i->iov_offset == i->bvec->bv_len) {
1146 			i->iov_offset = 0;
1147 			i->bvec++;
1148 			i->nr_segs--;
1149 		}
1150 		return maxsize;
1151 	}
1152 	if (iov_iter_is_xarray(i))
1153 		return iter_xarray_get_pages(i, pages, maxsize, maxpages, start);
1154 	return -EFAULT;
1155 }
1156 
1157 ssize_t iov_iter_get_pages2(struct iov_iter *i, struct page **pages,
1158 		size_t maxsize, unsigned maxpages, size_t *start)
1159 {
1160 	if (!maxpages)
1161 		return 0;
1162 	BUG_ON(!pages);
1163 
1164 	return __iov_iter_get_pages_alloc(i, &pages, maxsize, maxpages, start);
1165 }
1166 EXPORT_SYMBOL(iov_iter_get_pages2);
1167 
1168 ssize_t iov_iter_get_pages_alloc2(struct iov_iter *i,
1169 		struct page ***pages, size_t maxsize, size_t *start)
1170 {
1171 	ssize_t len;
1172 
1173 	*pages = NULL;
1174 
1175 	len = __iov_iter_get_pages_alloc(i, pages, maxsize, ~0U, start);
1176 	if (len <= 0) {
1177 		kvfree(*pages);
1178 		*pages = NULL;
1179 	}
1180 	return len;
1181 }
1182 EXPORT_SYMBOL(iov_iter_get_pages_alloc2);
1183 
1184 size_t csum_and_copy_from_iter(void *addr, size_t bytes, __wsum *csum,
1185 			       struct iov_iter *i)
1186 {
1187 	__wsum sum, next;
1188 	sum = *csum;
1189 	if (WARN_ON_ONCE(!i->data_source))
1190 		return 0;
1191 
1192 	iterate_and_advance(i, bytes, base, len, off, ({
1193 		next = csum_and_copy_from_user(base, addr + off, len);
1194 		sum = csum_block_add(sum, next, off);
1195 		next ? 0 : len;
1196 	}), ({
1197 		sum = csum_and_memcpy(addr + off, base, len, sum, off);
1198 	})
1199 	)
1200 	*csum = sum;
1201 	return bytes;
1202 }
1203 EXPORT_SYMBOL(csum_and_copy_from_iter);
1204 
1205 size_t csum_and_copy_to_iter(const void *addr, size_t bytes, void *_csstate,
1206 			     struct iov_iter *i)
1207 {
1208 	struct csum_state *csstate = _csstate;
1209 	__wsum sum, next;
1210 
1211 	if (WARN_ON_ONCE(i->data_source))
1212 		return 0;
1213 	if (unlikely(iov_iter_is_discard(i))) {
1214 		// can't use csum_memcpy() for that one - data is not copied
1215 		csstate->csum = csum_block_add(csstate->csum,
1216 					       csum_partial(addr, bytes, 0),
1217 					       csstate->off);
1218 		csstate->off += bytes;
1219 		return bytes;
1220 	}
1221 
1222 	sum = csum_shift(csstate->csum, csstate->off);
1223 	iterate_and_advance(i, bytes, base, len, off, ({
1224 		next = csum_and_copy_to_user(addr + off, base, len);
1225 		sum = csum_block_add(sum, next, off);
1226 		next ? 0 : len;
1227 	}), ({
1228 		sum = csum_and_memcpy(base, addr + off, len, sum, off);
1229 	})
1230 	)
1231 	csstate->csum = csum_shift(sum, csstate->off);
1232 	csstate->off += bytes;
1233 	return bytes;
1234 }
1235 EXPORT_SYMBOL(csum_and_copy_to_iter);
1236 
1237 size_t hash_and_copy_to_iter(const void *addr, size_t bytes, void *hashp,
1238 		struct iov_iter *i)
1239 {
1240 #ifdef CONFIG_CRYPTO_HASH
1241 	struct ahash_request *hash = hashp;
1242 	struct scatterlist sg;
1243 	size_t copied;
1244 
1245 	copied = copy_to_iter(addr, bytes, i);
1246 	sg_init_one(&sg, addr, copied);
1247 	ahash_request_set_crypt(hash, &sg, NULL, copied);
1248 	crypto_ahash_update(hash);
1249 	return copied;
1250 #else
1251 	return 0;
1252 #endif
1253 }
1254 EXPORT_SYMBOL(hash_and_copy_to_iter);
1255 
1256 static int iov_npages(const struct iov_iter *i, int maxpages)
1257 {
1258 	size_t skip = i->iov_offset, size = i->count;
1259 	const struct iovec *p;
1260 	int npages = 0;
1261 
1262 	for (p = iter_iov(i); size; skip = 0, p++) {
1263 		unsigned offs = offset_in_page(p->iov_base + skip);
1264 		size_t len = min(p->iov_len - skip, size);
1265 
1266 		if (len) {
1267 			size -= len;
1268 			npages += DIV_ROUND_UP(offs + len, PAGE_SIZE);
1269 			if (unlikely(npages > maxpages))
1270 				return maxpages;
1271 		}
1272 	}
1273 	return npages;
1274 }
1275 
1276 static int bvec_npages(const struct iov_iter *i, int maxpages)
1277 {
1278 	size_t skip = i->iov_offset, size = i->count;
1279 	const struct bio_vec *p;
1280 	int npages = 0;
1281 
1282 	for (p = i->bvec; size; skip = 0, p++) {
1283 		unsigned offs = (p->bv_offset + skip) % PAGE_SIZE;
1284 		size_t len = min(p->bv_len - skip, size);
1285 
1286 		size -= len;
1287 		npages += DIV_ROUND_UP(offs + len, PAGE_SIZE);
1288 		if (unlikely(npages > maxpages))
1289 			return maxpages;
1290 	}
1291 	return npages;
1292 }
1293 
1294 int iov_iter_npages(const struct iov_iter *i, int maxpages)
1295 {
1296 	if (unlikely(!i->count))
1297 		return 0;
1298 	if (likely(iter_is_ubuf(i))) {
1299 		unsigned offs = offset_in_page(i->ubuf + i->iov_offset);
1300 		int npages = DIV_ROUND_UP(offs + i->count, PAGE_SIZE);
1301 		return min(npages, maxpages);
1302 	}
1303 	/* iovec and kvec have identical layouts */
1304 	if (likely(iter_is_iovec(i) || iov_iter_is_kvec(i)))
1305 		return iov_npages(i, maxpages);
1306 	if (iov_iter_is_bvec(i))
1307 		return bvec_npages(i, maxpages);
1308 	if (iov_iter_is_xarray(i)) {
1309 		unsigned offset = (i->xarray_start + i->iov_offset) % PAGE_SIZE;
1310 		int npages = DIV_ROUND_UP(offset + i->count, PAGE_SIZE);
1311 		return min(npages, maxpages);
1312 	}
1313 	return 0;
1314 }
1315 EXPORT_SYMBOL(iov_iter_npages);
1316 
1317 const void *dup_iter(struct iov_iter *new, struct iov_iter *old, gfp_t flags)
1318 {
1319 	*new = *old;
1320 	if (iov_iter_is_bvec(new))
1321 		return new->bvec = kmemdup(new->bvec,
1322 				    new->nr_segs * sizeof(struct bio_vec),
1323 				    flags);
1324 	else if (iov_iter_is_kvec(new) || iter_is_iovec(new))
1325 		/* iovec and kvec have identical layout */
1326 		return new->__iov = kmemdup(new->__iov,
1327 				   new->nr_segs * sizeof(struct iovec),
1328 				   flags);
1329 	return NULL;
1330 }
1331 EXPORT_SYMBOL(dup_iter);
1332 
1333 static __noclone int copy_compat_iovec_from_user(struct iovec *iov,
1334 		const struct iovec __user *uvec, unsigned long nr_segs)
1335 {
1336 	const struct compat_iovec __user *uiov =
1337 		(const struct compat_iovec __user *)uvec;
1338 	int ret = -EFAULT, i;
1339 
1340 	if (!user_access_begin(uiov, nr_segs * sizeof(*uiov)))
1341 		return -EFAULT;
1342 
1343 	for (i = 0; i < nr_segs; i++) {
1344 		compat_uptr_t buf;
1345 		compat_ssize_t len;
1346 
1347 		unsafe_get_user(len, &uiov[i].iov_len, uaccess_end);
1348 		unsafe_get_user(buf, &uiov[i].iov_base, uaccess_end);
1349 
1350 		/* check for compat_size_t not fitting in compat_ssize_t .. */
1351 		if (len < 0) {
1352 			ret = -EINVAL;
1353 			goto uaccess_end;
1354 		}
1355 		iov[i].iov_base = compat_ptr(buf);
1356 		iov[i].iov_len = len;
1357 	}
1358 
1359 	ret = 0;
1360 uaccess_end:
1361 	user_access_end();
1362 	return ret;
1363 }
1364 
1365 static __noclone int copy_iovec_from_user(struct iovec *iov,
1366 		const struct iovec __user *uiov, unsigned long nr_segs)
1367 {
1368 	int ret = -EFAULT;
1369 
1370 	if (!user_access_begin(uiov, nr_segs * sizeof(*uiov)))
1371 		return -EFAULT;
1372 
1373 	do {
1374 		void __user *buf;
1375 		ssize_t len;
1376 
1377 		unsafe_get_user(len, &uiov->iov_len, uaccess_end);
1378 		unsafe_get_user(buf, &uiov->iov_base, uaccess_end);
1379 
1380 		/* check for size_t not fitting in ssize_t .. */
1381 		if (unlikely(len < 0)) {
1382 			ret = -EINVAL;
1383 			goto uaccess_end;
1384 		}
1385 		iov->iov_base = buf;
1386 		iov->iov_len = len;
1387 
1388 		uiov++; iov++;
1389 	} while (--nr_segs);
1390 
1391 	ret = 0;
1392 uaccess_end:
1393 	user_access_end();
1394 	return ret;
1395 }
1396 
1397 struct iovec *iovec_from_user(const struct iovec __user *uvec,
1398 		unsigned long nr_segs, unsigned long fast_segs,
1399 		struct iovec *fast_iov, bool compat)
1400 {
1401 	struct iovec *iov = fast_iov;
1402 	int ret;
1403 
1404 	/*
1405 	 * SuS says "The readv() function *may* fail if the iovcnt argument was
1406 	 * less than or equal to 0, or greater than {IOV_MAX}.  Linux has
1407 	 * traditionally returned zero for zero segments, so...
1408 	 */
1409 	if (nr_segs == 0)
1410 		return iov;
1411 	if (nr_segs > UIO_MAXIOV)
1412 		return ERR_PTR(-EINVAL);
1413 	if (nr_segs > fast_segs) {
1414 		iov = kmalloc_array(nr_segs, sizeof(struct iovec), GFP_KERNEL);
1415 		if (!iov)
1416 			return ERR_PTR(-ENOMEM);
1417 	}
1418 
1419 	if (unlikely(compat))
1420 		ret = copy_compat_iovec_from_user(iov, uvec, nr_segs);
1421 	else
1422 		ret = copy_iovec_from_user(iov, uvec, nr_segs);
1423 	if (ret) {
1424 		if (iov != fast_iov)
1425 			kfree(iov);
1426 		return ERR_PTR(ret);
1427 	}
1428 
1429 	return iov;
1430 }
1431 
1432 /*
1433  * Single segment iovec supplied by the user, import it as ITER_UBUF.
1434  */
1435 static ssize_t __import_iovec_ubuf(int type, const struct iovec __user *uvec,
1436 				   struct iovec **iovp, struct iov_iter *i,
1437 				   bool compat)
1438 {
1439 	struct iovec *iov = *iovp;
1440 	ssize_t ret;
1441 
1442 	if (compat)
1443 		ret = copy_compat_iovec_from_user(iov, uvec, 1);
1444 	else
1445 		ret = copy_iovec_from_user(iov, uvec, 1);
1446 	if (unlikely(ret))
1447 		return ret;
1448 
1449 	ret = import_ubuf(type, iov->iov_base, iov->iov_len, i);
1450 	if (unlikely(ret))
1451 		return ret;
1452 	*iovp = NULL;
1453 	return i->count;
1454 }
1455 
1456 ssize_t __import_iovec(int type, const struct iovec __user *uvec,
1457 		 unsigned nr_segs, unsigned fast_segs, struct iovec **iovp,
1458 		 struct iov_iter *i, bool compat)
1459 {
1460 	ssize_t total_len = 0;
1461 	unsigned long seg;
1462 	struct iovec *iov;
1463 
1464 	if (nr_segs == 1)
1465 		return __import_iovec_ubuf(type, uvec, iovp, i, compat);
1466 
1467 	iov = iovec_from_user(uvec, nr_segs, fast_segs, *iovp, compat);
1468 	if (IS_ERR(iov)) {
1469 		*iovp = NULL;
1470 		return PTR_ERR(iov);
1471 	}
1472 
1473 	/*
1474 	 * According to the Single Unix Specification we should return EINVAL if
1475 	 * an element length is < 0 when cast to ssize_t or if the total length
1476 	 * would overflow the ssize_t return value of the system call.
1477 	 *
1478 	 * Linux caps all read/write calls to MAX_RW_COUNT, and avoids the
1479 	 * overflow case.
1480 	 */
1481 	for (seg = 0; seg < nr_segs; seg++) {
1482 		ssize_t len = (ssize_t)iov[seg].iov_len;
1483 
1484 		if (!access_ok(iov[seg].iov_base, len)) {
1485 			if (iov != *iovp)
1486 				kfree(iov);
1487 			*iovp = NULL;
1488 			return -EFAULT;
1489 		}
1490 
1491 		if (len > MAX_RW_COUNT - total_len) {
1492 			len = MAX_RW_COUNT - total_len;
1493 			iov[seg].iov_len = len;
1494 		}
1495 		total_len += len;
1496 	}
1497 
1498 	iov_iter_init(i, type, iov, nr_segs, total_len);
1499 	if (iov == *iovp)
1500 		*iovp = NULL;
1501 	else
1502 		*iovp = iov;
1503 	return total_len;
1504 }
1505 
1506 /**
1507  * import_iovec() - Copy an array of &struct iovec from userspace
1508  *     into the kernel, check that it is valid, and initialize a new
1509  *     &struct iov_iter iterator to access it.
1510  *
1511  * @type: One of %READ or %WRITE.
1512  * @uvec: Pointer to the userspace array.
1513  * @nr_segs: Number of elements in userspace array.
1514  * @fast_segs: Number of elements in @iov.
1515  * @iovp: (input and output parameter) Pointer to pointer to (usually small
1516  *     on-stack) kernel array.
1517  * @i: Pointer to iterator that will be initialized on success.
1518  *
1519  * If the array pointed to by *@iov is large enough to hold all @nr_segs,
1520  * then this function places %NULL in *@iov on return. Otherwise, a new
1521  * array will be allocated and the result placed in *@iov. This means that
1522  * the caller may call kfree() on *@iov regardless of whether the small
1523  * on-stack array was used or not (and regardless of whether this function
1524  * returns an error or not).
1525  *
1526  * Return: Negative error code on error, bytes imported on success
1527  */
1528 ssize_t import_iovec(int type, const struct iovec __user *uvec,
1529 		 unsigned nr_segs, unsigned fast_segs,
1530 		 struct iovec **iovp, struct iov_iter *i)
1531 {
1532 	return __import_iovec(type, uvec, nr_segs, fast_segs, iovp, i,
1533 			      in_compat_syscall());
1534 }
1535 EXPORT_SYMBOL(import_iovec);
1536 
1537 int import_single_range(int rw, void __user *buf, size_t len,
1538 		 struct iovec *iov, struct iov_iter *i)
1539 {
1540 	if (len > MAX_RW_COUNT)
1541 		len = MAX_RW_COUNT;
1542 	if (unlikely(!access_ok(buf, len)))
1543 		return -EFAULT;
1544 
1545 	iov_iter_ubuf(i, rw, buf, len);
1546 	return 0;
1547 }
1548 EXPORT_SYMBOL(import_single_range);
1549 
1550 int import_ubuf(int rw, void __user *buf, size_t len, struct iov_iter *i)
1551 {
1552 	if (len > MAX_RW_COUNT)
1553 		len = MAX_RW_COUNT;
1554 	if (unlikely(!access_ok(buf, len)))
1555 		return -EFAULT;
1556 
1557 	iov_iter_ubuf(i, rw, buf, len);
1558 	return 0;
1559 }
1560 EXPORT_SYMBOL_GPL(import_ubuf);
1561 
1562 /**
1563  * iov_iter_restore() - Restore a &struct iov_iter to the same state as when
1564  *     iov_iter_save_state() was called.
1565  *
1566  * @i: &struct iov_iter to restore
1567  * @state: state to restore from
1568  *
1569  * Used after iov_iter_save_state() to bring restore @i, if operations may
1570  * have advanced it.
1571  *
1572  * Note: only works on ITER_IOVEC, ITER_BVEC, and ITER_KVEC
1573  */
1574 void iov_iter_restore(struct iov_iter *i, struct iov_iter_state *state)
1575 {
1576 	if (WARN_ON_ONCE(!iov_iter_is_bvec(i) && !iter_is_iovec(i) &&
1577 			 !iter_is_ubuf(i)) && !iov_iter_is_kvec(i))
1578 		return;
1579 	i->iov_offset = state->iov_offset;
1580 	i->count = state->count;
1581 	if (iter_is_ubuf(i))
1582 		return;
1583 	/*
1584 	 * For the *vec iters, nr_segs + iov is constant - if we increment
1585 	 * the vec, then we also decrement the nr_segs count. Hence we don't
1586 	 * need to track both of these, just one is enough and we can deduct
1587 	 * the other from that. ITER_KVEC and ITER_IOVEC are the same struct
1588 	 * size, so we can just increment the iov pointer as they are unionzed.
1589 	 * ITER_BVEC _may_ be the same size on some archs, but on others it is
1590 	 * not. Be safe and handle it separately.
1591 	 */
1592 	BUILD_BUG_ON(sizeof(struct iovec) != sizeof(struct kvec));
1593 	if (iov_iter_is_bvec(i))
1594 		i->bvec -= state->nr_segs - i->nr_segs;
1595 	else
1596 		i->__iov -= state->nr_segs - i->nr_segs;
1597 	i->nr_segs = state->nr_segs;
1598 }
1599 
1600 /*
1601  * Extract a list of contiguous pages from an ITER_XARRAY iterator.  This does not
1602  * get references on the pages, nor does it get a pin on them.
1603  */
1604 static ssize_t iov_iter_extract_xarray_pages(struct iov_iter *i,
1605 					     struct page ***pages, size_t maxsize,
1606 					     unsigned int maxpages,
1607 					     iov_iter_extraction_t extraction_flags,
1608 					     size_t *offset0)
1609 {
1610 	struct page *page, **p;
1611 	unsigned int nr = 0, offset;
1612 	loff_t pos = i->xarray_start + i->iov_offset;
1613 	pgoff_t index = pos >> PAGE_SHIFT;
1614 	XA_STATE(xas, i->xarray, index);
1615 
1616 	offset = pos & ~PAGE_MASK;
1617 	*offset0 = offset;
1618 
1619 	maxpages = want_pages_array(pages, maxsize, offset, maxpages);
1620 	if (!maxpages)
1621 		return -ENOMEM;
1622 	p = *pages;
1623 
1624 	rcu_read_lock();
1625 	for (page = xas_load(&xas); page; page = xas_next(&xas)) {
1626 		if (xas_retry(&xas, page))
1627 			continue;
1628 
1629 		/* Has the page moved or been split? */
1630 		if (unlikely(page != xas_reload(&xas))) {
1631 			xas_reset(&xas);
1632 			continue;
1633 		}
1634 
1635 		p[nr++] = find_subpage(page, xas.xa_index);
1636 		if (nr == maxpages)
1637 			break;
1638 	}
1639 	rcu_read_unlock();
1640 
1641 	maxsize = min_t(size_t, nr * PAGE_SIZE - offset, maxsize);
1642 	iov_iter_advance(i, maxsize);
1643 	return maxsize;
1644 }
1645 
1646 /*
1647  * Extract a list of contiguous pages from an ITER_BVEC iterator.  This does
1648  * not get references on the pages, nor does it get a pin on them.
1649  */
1650 static ssize_t iov_iter_extract_bvec_pages(struct iov_iter *i,
1651 					   struct page ***pages, size_t maxsize,
1652 					   unsigned int maxpages,
1653 					   iov_iter_extraction_t extraction_flags,
1654 					   size_t *offset0)
1655 {
1656 	struct page **p, *page;
1657 	size_t skip = i->iov_offset, offset, size;
1658 	int k;
1659 
1660 	for (;;) {
1661 		if (i->nr_segs == 0)
1662 			return 0;
1663 		size = min(maxsize, i->bvec->bv_len - skip);
1664 		if (size)
1665 			break;
1666 		i->iov_offset = 0;
1667 		i->nr_segs--;
1668 		i->bvec++;
1669 		skip = 0;
1670 	}
1671 
1672 	skip += i->bvec->bv_offset;
1673 	page = i->bvec->bv_page + skip / PAGE_SIZE;
1674 	offset = skip % PAGE_SIZE;
1675 	*offset0 = offset;
1676 
1677 	maxpages = want_pages_array(pages, size, offset, maxpages);
1678 	if (!maxpages)
1679 		return -ENOMEM;
1680 	p = *pages;
1681 	for (k = 0; k < maxpages; k++)
1682 		p[k] = page + k;
1683 
1684 	size = min_t(size_t, size, maxpages * PAGE_SIZE - offset);
1685 	iov_iter_advance(i, size);
1686 	return size;
1687 }
1688 
1689 /*
1690  * Extract a list of virtually contiguous pages from an ITER_KVEC iterator.
1691  * This does not get references on the pages, nor does it get a pin on them.
1692  */
1693 static ssize_t iov_iter_extract_kvec_pages(struct iov_iter *i,
1694 					   struct page ***pages, size_t maxsize,
1695 					   unsigned int maxpages,
1696 					   iov_iter_extraction_t extraction_flags,
1697 					   size_t *offset0)
1698 {
1699 	struct page **p, *page;
1700 	const void *kaddr;
1701 	size_t skip = i->iov_offset, offset, len, size;
1702 	int k;
1703 
1704 	for (;;) {
1705 		if (i->nr_segs == 0)
1706 			return 0;
1707 		size = min(maxsize, i->kvec->iov_len - skip);
1708 		if (size)
1709 			break;
1710 		i->iov_offset = 0;
1711 		i->nr_segs--;
1712 		i->kvec++;
1713 		skip = 0;
1714 	}
1715 
1716 	kaddr = i->kvec->iov_base + skip;
1717 	offset = (unsigned long)kaddr & ~PAGE_MASK;
1718 	*offset0 = offset;
1719 
1720 	maxpages = want_pages_array(pages, size, offset, maxpages);
1721 	if (!maxpages)
1722 		return -ENOMEM;
1723 	p = *pages;
1724 
1725 	kaddr -= offset;
1726 	len = offset + size;
1727 	for (k = 0; k < maxpages; k++) {
1728 		size_t seg = min_t(size_t, len, PAGE_SIZE);
1729 
1730 		if (is_vmalloc_or_module_addr(kaddr))
1731 			page = vmalloc_to_page(kaddr);
1732 		else
1733 			page = virt_to_page(kaddr);
1734 
1735 		p[k] = page;
1736 		len -= seg;
1737 		kaddr += PAGE_SIZE;
1738 	}
1739 
1740 	size = min_t(size_t, size, maxpages * PAGE_SIZE - offset);
1741 	iov_iter_advance(i, size);
1742 	return size;
1743 }
1744 
1745 /*
1746  * Extract a list of contiguous pages from a user iterator and get a pin on
1747  * each of them.  This should only be used if the iterator is user-backed
1748  * (IOBUF/UBUF).
1749  *
1750  * It does not get refs on the pages, but the pages must be unpinned by the
1751  * caller once the transfer is complete.
1752  *
1753  * This is safe to be used where background IO/DMA *is* going to be modifying
1754  * the buffer; using a pin rather than a ref makes forces fork() to give the
1755  * child a copy of the page.
1756  */
1757 static ssize_t iov_iter_extract_user_pages(struct iov_iter *i,
1758 					   struct page ***pages,
1759 					   size_t maxsize,
1760 					   unsigned int maxpages,
1761 					   iov_iter_extraction_t extraction_flags,
1762 					   size_t *offset0)
1763 {
1764 	unsigned long addr;
1765 	unsigned int gup_flags = 0;
1766 	size_t offset;
1767 	int res;
1768 
1769 	if (i->data_source == ITER_DEST)
1770 		gup_flags |= FOLL_WRITE;
1771 	if (extraction_flags & ITER_ALLOW_P2PDMA)
1772 		gup_flags |= FOLL_PCI_P2PDMA;
1773 	if (i->nofault)
1774 		gup_flags |= FOLL_NOFAULT;
1775 
1776 	addr = first_iovec_segment(i, &maxsize);
1777 	*offset0 = offset = addr % PAGE_SIZE;
1778 	addr &= PAGE_MASK;
1779 	maxpages = want_pages_array(pages, maxsize, offset, maxpages);
1780 	if (!maxpages)
1781 		return -ENOMEM;
1782 	res = pin_user_pages_fast(addr, maxpages, gup_flags, *pages);
1783 	if (unlikely(res <= 0))
1784 		return res;
1785 	maxsize = min_t(size_t, maxsize, res * PAGE_SIZE - offset);
1786 	iov_iter_advance(i, maxsize);
1787 	return maxsize;
1788 }
1789 
1790 /**
1791  * iov_iter_extract_pages - Extract a list of contiguous pages from an iterator
1792  * @i: The iterator to extract from
1793  * @pages: Where to return the list of pages
1794  * @maxsize: The maximum amount of iterator to extract
1795  * @maxpages: The maximum size of the list of pages
1796  * @extraction_flags: Flags to qualify request
1797  * @offset0: Where to return the starting offset into (*@pages)[0]
1798  *
1799  * Extract a list of contiguous pages from the current point of the iterator,
1800  * advancing the iterator.  The maximum number of pages and the maximum amount
1801  * of page contents can be set.
1802  *
1803  * If *@pages is NULL, a page list will be allocated to the required size and
1804  * *@pages will be set to its base.  If *@pages is not NULL, it will be assumed
1805  * that the caller allocated a page list at least @maxpages in size and this
1806  * will be filled in.
1807  *
1808  * @extraction_flags can have ITER_ALLOW_P2PDMA set to request peer-to-peer DMA
1809  * be allowed on the pages extracted.
1810  *
1811  * The iov_iter_extract_will_pin() function can be used to query how cleanup
1812  * should be performed.
1813  *
1814  * Extra refs or pins on the pages may be obtained as follows:
1815  *
1816  *  (*) If the iterator is user-backed (ITER_IOVEC/ITER_UBUF), pins will be
1817  *      added to the pages, but refs will not be taken.
1818  *      iov_iter_extract_will_pin() will return true.
1819  *
1820  *  (*) If the iterator is ITER_KVEC, ITER_BVEC or ITER_XARRAY, the pages are
1821  *      merely listed; no extra refs or pins are obtained.
1822  *      iov_iter_extract_will_pin() will return 0.
1823  *
1824  * Note also:
1825  *
1826  *  (*) Use with ITER_DISCARD is not supported as that has no content.
1827  *
1828  * On success, the function sets *@pages to the new pagelist, if allocated, and
1829  * sets *offset0 to the offset into the first page.
1830  *
1831  * It may also return -ENOMEM and -EFAULT.
1832  */
1833 ssize_t iov_iter_extract_pages(struct iov_iter *i,
1834 			       struct page ***pages,
1835 			       size_t maxsize,
1836 			       unsigned int maxpages,
1837 			       iov_iter_extraction_t extraction_flags,
1838 			       size_t *offset0)
1839 {
1840 	maxsize = min_t(size_t, min_t(size_t, maxsize, i->count), MAX_RW_COUNT);
1841 	if (!maxsize)
1842 		return 0;
1843 
1844 	if (likely(user_backed_iter(i)))
1845 		return iov_iter_extract_user_pages(i, pages, maxsize,
1846 						   maxpages, extraction_flags,
1847 						   offset0);
1848 	if (iov_iter_is_kvec(i))
1849 		return iov_iter_extract_kvec_pages(i, pages, maxsize,
1850 						   maxpages, extraction_flags,
1851 						   offset0);
1852 	if (iov_iter_is_bvec(i))
1853 		return iov_iter_extract_bvec_pages(i, pages, maxsize,
1854 						   maxpages, extraction_flags,
1855 						   offset0);
1856 	if (iov_iter_is_xarray(i))
1857 		return iov_iter_extract_xarray_pages(i, pages, maxsize,
1858 						     maxpages, extraction_flags,
1859 						     offset0);
1860 	return -EFAULT;
1861 }
1862 EXPORT_SYMBOL_GPL(iov_iter_extract_pages);
1863