xref: /linux/arch/x86/lib/usercopy_32.c (revision ca55b2fef3a9373fcfc30f82fd26bc7fccbda732)
1 /*
2  * User address space access functions.
3  * The non inlined parts of asm-i386/uaccess.h are here.
4  *
5  * Copyright 1997 Andi Kleen <ak@muc.de>
6  * Copyright 1997 Linus Torvalds
7  */
8 #include <linux/mm.h>
9 #include <linux/highmem.h>
10 #include <linux/blkdev.h>
11 #include <linux/module.h>
12 #include <linux/backing-dev.h>
13 #include <linux/interrupt.h>
14 #include <asm/uaccess.h>
15 #include <asm/mmx.h>
16 #include <asm/asm.h>
17 
18 #ifdef CONFIG_X86_INTEL_USERCOPY
19 /*
20  * Alignment at which movsl is preferred for bulk memory copies.
21  */
22 struct movsl_mask movsl_mask __read_mostly;
23 #endif
24 
25 static inline int __movsl_is_ok(unsigned long a1, unsigned long a2, unsigned long n)
26 {
27 #ifdef CONFIG_X86_INTEL_USERCOPY
28 	if (n >= 64 && ((a1 ^ a2) & movsl_mask.mask))
29 		return 0;
30 #endif
31 	return 1;
32 }
33 #define movsl_is_ok(a1, a2, n) \
34 	__movsl_is_ok((unsigned long)(a1), (unsigned long)(a2), (n))
35 
36 /*
37  * Zero Userspace
38  */
39 
40 #define __do_clear_user(addr,size)					\
41 do {									\
42 	int __d0;							\
43 	might_fault();							\
44 	__asm__ __volatile__(						\
45 		ASM_STAC "\n"						\
46 		"0:	rep; stosl\n"					\
47 		"	movl %2,%0\n"					\
48 		"1:	rep; stosb\n"					\
49 		"2: " ASM_CLAC "\n"					\
50 		".section .fixup,\"ax\"\n"				\
51 		"3:	lea 0(%2,%0,4),%0\n"				\
52 		"	jmp 2b\n"					\
53 		".previous\n"						\
54 		_ASM_EXTABLE(0b,3b)					\
55 		_ASM_EXTABLE(1b,2b)					\
56 		: "=&c"(size), "=&D" (__d0)				\
57 		: "r"(size & 3), "0"(size / 4), "1"(addr), "a"(0));	\
58 } while (0)
59 
60 /**
61  * clear_user: - Zero a block of memory in user space.
62  * @to:   Destination address, in user space.
63  * @n:    Number of bytes to zero.
64  *
65  * Zero a block of memory in user space.
66  *
67  * Returns number of bytes that could not be cleared.
68  * On success, this will be zero.
69  */
70 unsigned long
71 clear_user(void __user *to, unsigned long n)
72 {
73 	might_fault();
74 	if (access_ok(VERIFY_WRITE, to, n))
75 		__do_clear_user(to, n);
76 	return n;
77 }
78 EXPORT_SYMBOL(clear_user);
79 
80 /**
81  * __clear_user: - Zero a block of memory in user space, with less checking.
82  * @to:   Destination address, in user space.
83  * @n:    Number of bytes to zero.
84  *
85  * Zero a block of memory in user space.  Caller must check
86  * the specified block with access_ok() before calling this function.
87  *
88  * Returns number of bytes that could not be cleared.
89  * On success, this will be zero.
90  */
91 unsigned long
92 __clear_user(void __user *to, unsigned long n)
93 {
94 	__do_clear_user(to, n);
95 	return n;
96 }
97 EXPORT_SYMBOL(__clear_user);
98 
99 #ifdef CONFIG_X86_INTEL_USERCOPY
100 static unsigned long
101 __copy_user_intel(void __user *to, const void *from, unsigned long size)
102 {
103 	int d0, d1;
104 	__asm__ __volatile__(
105 		       "       .align 2,0x90\n"
106 		       "1:     movl 32(%4), %%eax\n"
107 		       "       cmpl $67, %0\n"
108 		       "       jbe 3f\n"
109 		       "2:     movl 64(%4), %%eax\n"
110 		       "       .align 2,0x90\n"
111 		       "3:     movl 0(%4), %%eax\n"
112 		       "4:     movl 4(%4), %%edx\n"
113 		       "5:     movl %%eax, 0(%3)\n"
114 		       "6:     movl %%edx, 4(%3)\n"
115 		       "7:     movl 8(%4), %%eax\n"
116 		       "8:     movl 12(%4),%%edx\n"
117 		       "9:     movl %%eax, 8(%3)\n"
118 		       "10:    movl %%edx, 12(%3)\n"
119 		       "11:    movl 16(%4), %%eax\n"
120 		       "12:    movl 20(%4), %%edx\n"
121 		       "13:    movl %%eax, 16(%3)\n"
122 		       "14:    movl %%edx, 20(%3)\n"
123 		       "15:    movl 24(%4), %%eax\n"
124 		       "16:    movl 28(%4), %%edx\n"
125 		       "17:    movl %%eax, 24(%3)\n"
126 		       "18:    movl %%edx, 28(%3)\n"
127 		       "19:    movl 32(%4), %%eax\n"
128 		       "20:    movl 36(%4), %%edx\n"
129 		       "21:    movl %%eax, 32(%3)\n"
130 		       "22:    movl %%edx, 36(%3)\n"
131 		       "23:    movl 40(%4), %%eax\n"
132 		       "24:    movl 44(%4), %%edx\n"
133 		       "25:    movl %%eax, 40(%3)\n"
134 		       "26:    movl %%edx, 44(%3)\n"
135 		       "27:    movl 48(%4), %%eax\n"
136 		       "28:    movl 52(%4), %%edx\n"
137 		       "29:    movl %%eax, 48(%3)\n"
138 		       "30:    movl %%edx, 52(%3)\n"
139 		       "31:    movl 56(%4), %%eax\n"
140 		       "32:    movl 60(%4), %%edx\n"
141 		       "33:    movl %%eax, 56(%3)\n"
142 		       "34:    movl %%edx, 60(%3)\n"
143 		       "       addl $-64, %0\n"
144 		       "       addl $64, %4\n"
145 		       "       addl $64, %3\n"
146 		       "       cmpl $63, %0\n"
147 		       "       ja  1b\n"
148 		       "35:    movl  %0, %%eax\n"
149 		       "       shrl  $2, %0\n"
150 		       "       andl  $3, %%eax\n"
151 		       "       cld\n"
152 		       "99:    rep; movsl\n"
153 		       "36:    movl %%eax, %0\n"
154 		       "37:    rep; movsb\n"
155 		       "100:\n"
156 		       ".section .fixup,\"ax\"\n"
157 		       "101:   lea 0(%%eax,%0,4),%0\n"
158 		       "       jmp 100b\n"
159 		       ".previous\n"
160 		       _ASM_EXTABLE(1b,100b)
161 		       _ASM_EXTABLE(2b,100b)
162 		       _ASM_EXTABLE(3b,100b)
163 		       _ASM_EXTABLE(4b,100b)
164 		       _ASM_EXTABLE(5b,100b)
165 		       _ASM_EXTABLE(6b,100b)
166 		       _ASM_EXTABLE(7b,100b)
167 		       _ASM_EXTABLE(8b,100b)
168 		       _ASM_EXTABLE(9b,100b)
169 		       _ASM_EXTABLE(10b,100b)
170 		       _ASM_EXTABLE(11b,100b)
171 		       _ASM_EXTABLE(12b,100b)
172 		       _ASM_EXTABLE(13b,100b)
173 		       _ASM_EXTABLE(14b,100b)
174 		       _ASM_EXTABLE(15b,100b)
175 		       _ASM_EXTABLE(16b,100b)
176 		       _ASM_EXTABLE(17b,100b)
177 		       _ASM_EXTABLE(18b,100b)
178 		       _ASM_EXTABLE(19b,100b)
179 		       _ASM_EXTABLE(20b,100b)
180 		       _ASM_EXTABLE(21b,100b)
181 		       _ASM_EXTABLE(22b,100b)
182 		       _ASM_EXTABLE(23b,100b)
183 		       _ASM_EXTABLE(24b,100b)
184 		       _ASM_EXTABLE(25b,100b)
185 		       _ASM_EXTABLE(26b,100b)
186 		       _ASM_EXTABLE(27b,100b)
187 		       _ASM_EXTABLE(28b,100b)
188 		       _ASM_EXTABLE(29b,100b)
189 		       _ASM_EXTABLE(30b,100b)
190 		       _ASM_EXTABLE(31b,100b)
191 		       _ASM_EXTABLE(32b,100b)
192 		       _ASM_EXTABLE(33b,100b)
193 		       _ASM_EXTABLE(34b,100b)
194 		       _ASM_EXTABLE(35b,100b)
195 		       _ASM_EXTABLE(36b,100b)
196 		       _ASM_EXTABLE(37b,100b)
197 		       _ASM_EXTABLE(99b,101b)
198 		       : "=&c"(size), "=&D" (d0), "=&S" (d1)
199 		       :  "1"(to), "2"(from), "0"(size)
200 		       : "eax", "edx", "memory");
201 	return size;
202 }
203 
204 static unsigned long
205 __copy_user_zeroing_intel(void *to, const void __user *from, unsigned long size)
206 {
207 	int d0, d1;
208 	__asm__ __volatile__(
209 		       "        .align 2,0x90\n"
210 		       "0:      movl 32(%4), %%eax\n"
211 		       "        cmpl $67, %0\n"
212 		       "        jbe 2f\n"
213 		       "1:      movl 64(%4), %%eax\n"
214 		       "        .align 2,0x90\n"
215 		       "2:      movl 0(%4), %%eax\n"
216 		       "21:     movl 4(%4), %%edx\n"
217 		       "        movl %%eax, 0(%3)\n"
218 		       "        movl %%edx, 4(%3)\n"
219 		       "3:      movl 8(%4), %%eax\n"
220 		       "31:     movl 12(%4),%%edx\n"
221 		       "        movl %%eax, 8(%3)\n"
222 		       "        movl %%edx, 12(%3)\n"
223 		       "4:      movl 16(%4), %%eax\n"
224 		       "41:     movl 20(%4), %%edx\n"
225 		       "        movl %%eax, 16(%3)\n"
226 		       "        movl %%edx, 20(%3)\n"
227 		       "10:     movl 24(%4), %%eax\n"
228 		       "51:     movl 28(%4), %%edx\n"
229 		       "        movl %%eax, 24(%3)\n"
230 		       "        movl %%edx, 28(%3)\n"
231 		       "11:     movl 32(%4), %%eax\n"
232 		       "61:     movl 36(%4), %%edx\n"
233 		       "        movl %%eax, 32(%3)\n"
234 		       "        movl %%edx, 36(%3)\n"
235 		       "12:     movl 40(%4), %%eax\n"
236 		       "71:     movl 44(%4), %%edx\n"
237 		       "        movl %%eax, 40(%3)\n"
238 		       "        movl %%edx, 44(%3)\n"
239 		       "13:     movl 48(%4), %%eax\n"
240 		       "81:     movl 52(%4), %%edx\n"
241 		       "        movl %%eax, 48(%3)\n"
242 		       "        movl %%edx, 52(%3)\n"
243 		       "14:     movl 56(%4), %%eax\n"
244 		       "91:     movl 60(%4), %%edx\n"
245 		       "        movl %%eax, 56(%3)\n"
246 		       "        movl %%edx, 60(%3)\n"
247 		       "        addl $-64, %0\n"
248 		       "        addl $64, %4\n"
249 		       "        addl $64, %3\n"
250 		       "        cmpl $63, %0\n"
251 		       "        ja  0b\n"
252 		       "5:      movl  %0, %%eax\n"
253 		       "        shrl  $2, %0\n"
254 		       "        andl $3, %%eax\n"
255 		       "        cld\n"
256 		       "6:      rep; movsl\n"
257 		       "        movl %%eax,%0\n"
258 		       "7:      rep; movsb\n"
259 		       "8:\n"
260 		       ".section .fixup,\"ax\"\n"
261 		       "9:      lea 0(%%eax,%0,4),%0\n"
262 		       "16:     pushl %0\n"
263 		       "        pushl %%eax\n"
264 		       "        xorl %%eax,%%eax\n"
265 		       "        rep; stosb\n"
266 		       "        popl %%eax\n"
267 		       "        popl %0\n"
268 		       "        jmp 8b\n"
269 		       ".previous\n"
270 		       _ASM_EXTABLE(0b,16b)
271 		       _ASM_EXTABLE(1b,16b)
272 		       _ASM_EXTABLE(2b,16b)
273 		       _ASM_EXTABLE(21b,16b)
274 		       _ASM_EXTABLE(3b,16b)
275 		       _ASM_EXTABLE(31b,16b)
276 		       _ASM_EXTABLE(4b,16b)
277 		       _ASM_EXTABLE(41b,16b)
278 		       _ASM_EXTABLE(10b,16b)
279 		       _ASM_EXTABLE(51b,16b)
280 		       _ASM_EXTABLE(11b,16b)
281 		       _ASM_EXTABLE(61b,16b)
282 		       _ASM_EXTABLE(12b,16b)
283 		       _ASM_EXTABLE(71b,16b)
284 		       _ASM_EXTABLE(13b,16b)
285 		       _ASM_EXTABLE(81b,16b)
286 		       _ASM_EXTABLE(14b,16b)
287 		       _ASM_EXTABLE(91b,16b)
288 		       _ASM_EXTABLE(6b,9b)
289 		       _ASM_EXTABLE(7b,16b)
290 		       : "=&c"(size), "=&D" (d0), "=&S" (d1)
291 		       :  "1"(to), "2"(from), "0"(size)
292 		       : "eax", "edx", "memory");
293 	return size;
294 }
295 
296 /*
297  * Non Temporal Hint version of __copy_user_zeroing_intel.  It is cache aware.
298  * hyoshiok@miraclelinux.com
299  */
300 
301 static unsigned long __copy_user_zeroing_intel_nocache(void *to,
302 				const void __user *from, unsigned long size)
303 {
304 	int d0, d1;
305 
306 	__asm__ __volatile__(
307 	       "        .align 2,0x90\n"
308 	       "0:      movl 32(%4), %%eax\n"
309 	       "        cmpl $67, %0\n"
310 	       "        jbe 2f\n"
311 	       "1:      movl 64(%4), %%eax\n"
312 	       "        .align 2,0x90\n"
313 	       "2:      movl 0(%4), %%eax\n"
314 	       "21:     movl 4(%4), %%edx\n"
315 	       "        movnti %%eax, 0(%3)\n"
316 	       "        movnti %%edx, 4(%3)\n"
317 	       "3:      movl 8(%4), %%eax\n"
318 	       "31:     movl 12(%4),%%edx\n"
319 	       "        movnti %%eax, 8(%3)\n"
320 	       "        movnti %%edx, 12(%3)\n"
321 	       "4:      movl 16(%4), %%eax\n"
322 	       "41:     movl 20(%4), %%edx\n"
323 	       "        movnti %%eax, 16(%3)\n"
324 	       "        movnti %%edx, 20(%3)\n"
325 	       "10:     movl 24(%4), %%eax\n"
326 	       "51:     movl 28(%4), %%edx\n"
327 	       "        movnti %%eax, 24(%3)\n"
328 	       "        movnti %%edx, 28(%3)\n"
329 	       "11:     movl 32(%4), %%eax\n"
330 	       "61:     movl 36(%4), %%edx\n"
331 	       "        movnti %%eax, 32(%3)\n"
332 	       "        movnti %%edx, 36(%3)\n"
333 	       "12:     movl 40(%4), %%eax\n"
334 	       "71:     movl 44(%4), %%edx\n"
335 	       "        movnti %%eax, 40(%3)\n"
336 	       "        movnti %%edx, 44(%3)\n"
337 	       "13:     movl 48(%4), %%eax\n"
338 	       "81:     movl 52(%4), %%edx\n"
339 	       "        movnti %%eax, 48(%3)\n"
340 	       "        movnti %%edx, 52(%3)\n"
341 	       "14:     movl 56(%4), %%eax\n"
342 	       "91:     movl 60(%4), %%edx\n"
343 	       "        movnti %%eax, 56(%3)\n"
344 	       "        movnti %%edx, 60(%3)\n"
345 	       "        addl $-64, %0\n"
346 	       "        addl $64, %4\n"
347 	       "        addl $64, %3\n"
348 	       "        cmpl $63, %0\n"
349 	       "        ja  0b\n"
350 	       "        sfence \n"
351 	       "5:      movl  %0, %%eax\n"
352 	       "        shrl  $2, %0\n"
353 	       "        andl $3, %%eax\n"
354 	       "        cld\n"
355 	       "6:      rep; movsl\n"
356 	       "        movl %%eax,%0\n"
357 	       "7:      rep; movsb\n"
358 	       "8:\n"
359 	       ".section .fixup,\"ax\"\n"
360 	       "9:      lea 0(%%eax,%0,4),%0\n"
361 	       "16:     pushl %0\n"
362 	       "        pushl %%eax\n"
363 	       "        xorl %%eax,%%eax\n"
364 	       "        rep; stosb\n"
365 	       "        popl %%eax\n"
366 	       "        popl %0\n"
367 	       "        jmp 8b\n"
368 	       ".previous\n"
369 	       _ASM_EXTABLE(0b,16b)
370 	       _ASM_EXTABLE(1b,16b)
371 	       _ASM_EXTABLE(2b,16b)
372 	       _ASM_EXTABLE(21b,16b)
373 	       _ASM_EXTABLE(3b,16b)
374 	       _ASM_EXTABLE(31b,16b)
375 	       _ASM_EXTABLE(4b,16b)
376 	       _ASM_EXTABLE(41b,16b)
377 	       _ASM_EXTABLE(10b,16b)
378 	       _ASM_EXTABLE(51b,16b)
379 	       _ASM_EXTABLE(11b,16b)
380 	       _ASM_EXTABLE(61b,16b)
381 	       _ASM_EXTABLE(12b,16b)
382 	       _ASM_EXTABLE(71b,16b)
383 	       _ASM_EXTABLE(13b,16b)
384 	       _ASM_EXTABLE(81b,16b)
385 	       _ASM_EXTABLE(14b,16b)
386 	       _ASM_EXTABLE(91b,16b)
387 	       _ASM_EXTABLE(6b,9b)
388 	       _ASM_EXTABLE(7b,16b)
389 	       : "=&c"(size), "=&D" (d0), "=&S" (d1)
390 	       :  "1"(to), "2"(from), "0"(size)
391 	       : "eax", "edx", "memory");
392 	return size;
393 }
394 
395 static unsigned long __copy_user_intel_nocache(void *to,
396 				const void __user *from, unsigned long size)
397 {
398 	int d0, d1;
399 
400 	__asm__ __volatile__(
401 	       "        .align 2,0x90\n"
402 	       "0:      movl 32(%4), %%eax\n"
403 	       "        cmpl $67, %0\n"
404 	       "        jbe 2f\n"
405 	       "1:      movl 64(%4), %%eax\n"
406 	       "        .align 2,0x90\n"
407 	       "2:      movl 0(%4), %%eax\n"
408 	       "21:     movl 4(%4), %%edx\n"
409 	       "        movnti %%eax, 0(%3)\n"
410 	       "        movnti %%edx, 4(%3)\n"
411 	       "3:      movl 8(%4), %%eax\n"
412 	       "31:     movl 12(%4),%%edx\n"
413 	       "        movnti %%eax, 8(%3)\n"
414 	       "        movnti %%edx, 12(%3)\n"
415 	       "4:      movl 16(%4), %%eax\n"
416 	       "41:     movl 20(%4), %%edx\n"
417 	       "        movnti %%eax, 16(%3)\n"
418 	       "        movnti %%edx, 20(%3)\n"
419 	       "10:     movl 24(%4), %%eax\n"
420 	       "51:     movl 28(%4), %%edx\n"
421 	       "        movnti %%eax, 24(%3)\n"
422 	       "        movnti %%edx, 28(%3)\n"
423 	       "11:     movl 32(%4), %%eax\n"
424 	       "61:     movl 36(%4), %%edx\n"
425 	       "        movnti %%eax, 32(%3)\n"
426 	       "        movnti %%edx, 36(%3)\n"
427 	       "12:     movl 40(%4), %%eax\n"
428 	       "71:     movl 44(%4), %%edx\n"
429 	       "        movnti %%eax, 40(%3)\n"
430 	       "        movnti %%edx, 44(%3)\n"
431 	       "13:     movl 48(%4), %%eax\n"
432 	       "81:     movl 52(%4), %%edx\n"
433 	       "        movnti %%eax, 48(%3)\n"
434 	       "        movnti %%edx, 52(%3)\n"
435 	       "14:     movl 56(%4), %%eax\n"
436 	       "91:     movl 60(%4), %%edx\n"
437 	       "        movnti %%eax, 56(%3)\n"
438 	       "        movnti %%edx, 60(%3)\n"
439 	       "        addl $-64, %0\n"
440 	       "        addl $64, %4\n"
441 	       "        addl $64, %3\n"
442 	       "        cmpl $63, %0\n"
443 	       "        ja  0b\n"
444 	       "        sfence \n"
445 	       "5:      movl  %0, %%eax\n"
446 	       "        shrl  $2, %0\n"
447 	       "        andl $3, %%eax\n"
448 	       "        cld\n"
449 	       "6:      rep; movsl\n"
450 	       "        movl %%eax,%0\n"
451 	       "7:      rep; movsb\n"
452 	       "8:\n"
453 	       ".section .fixup,\"ax\"\n"
454 	       "9:      lea 0(%%eax,%0,4),%0\n"
455 	       "16:     jmp 8b\n"
456 	       ".previous\n"
457 	       _ASM_EXTABLE(0b,16b)
458 	       _ASM_EXTABLE(1b,16b)
459 	       _ASM_EXTABLE(2b,16b)
460 	       _ASM_EXTABLE(21b,16b)
461 	       _ASM_EXTABLE(3b,16b)
462 	       _ASM_EXTABLE(31b,16b)
463 	       _ASM_EXTABLE(4b,16b)
464 	       _ASM_EXTABLE(41b,16b)
465 	       _ASM_EXTABLE(10b,16b)
466 	       _ASM_EXTABLE(51b,16b)
467 	       _ASM_EXTABLE(11b,16b)
468 	       _ASM_EXTABLE(61b,16b)
469 	       _ASM_EXTABLE(12b,16b)
470 	       _ASM_EXTABLE(71b,16b)
471 	       _ASM_EXTABLE(13b,16b)
472 	       _ASM_EXTABLE(81b,16b)
473 	       _ASM_EXTABLE(14b,16b)
474 	       _ASM_EXTABLE(91b,16b)
475 	       _ASM_EXTABLE(6b,9b)
476 	       _ASM_EXTABLE(7b,16b)
477 	       : "=&c"(size), "=&D" (d0), "=&S" (d1)
478 	       :  "1"(to), "2"(from), "0"(size)
479 	       : "eax", "edx", "memory");
480 	return size;
481 }
482 
483 #else
484 
485 /*
486  * Leave these declared but undefined.  They should not be any references to
487  * them
488  */
489 unsigned long __copy_user_zeroing_intel(void *to, const void __user *from,
490 					unsigned long size);
491 unsigned long __copy_user_intel(void __user *to, const void *from,
492 					unsigned long size);
493 unsigned long __copy_user_zeroing_intel_nocache(void *to,
494 				const void __user *from, unsigned long size);
495 #endif /* CONFIG_X86_INTEL_USERCOPY */
496 
497 /* Generic arbitrary sized copy.  */
498 #define __copy_user(to, from, size)					\
499 do {									\
500 	int __d0, __d1, __d2;						\
501 	__asm__ __volatile__(						\
502 		"	cmp  $7,%0\n"					\
503 		"	jbe  1f\n"					\
504 		"	movl %1,%0\n"					\
505 		"	negl %0\n"					\
506 		"	andl $7,%0\n"					\
507 		"	subl %0,%3\n"					\
508 		"4:	rep; movsb\n"					\
509 		"	movl %3,%0\n"					\
510 		"	shrl $2,%0\n"					\
511 		"	andl $3,%3\n"					\
512 		"	.align 2,0x90\n"				\
513 		"0:	rep; movsl\n"					\
514 		"	movl %3,%0\n"					\
515 		"1:	rep; movsb\n"					\
516 		"2:\n"							\
517 		".section .fixup,\"ax\"\n"				\
518 		"5:	addl %3,%0\n"					\
519 		"	jmp 2b\n"					\
520 		"3:	lea 0(%3,%0,4),%0\n"				\
521 		"	jmp 2b\n"					\
522 		".previous\n"						\
523 		_ASM_EXTABLE(4b,5b)					\
524 		_ASM_EXTABLE(0b,3b)					\
525 		_ASM_EXTABLE(1b,2b)					\
526 		: "=&c"(size), "=&D" (__d0), "=&S" (__d1), "=r"(__d2)	\
527 		: "3"(size), "0"(size), "1"(to), "2"(from)		\
528 		: "memory");						\
529 } while (0)
530 
531 #define __copy_user_zeroing(to, from, size)				\
532 do {									\
533 	int __d0, __d1, __d2;						\
534 	__asm__ __volatile__(						\
535 		"	cmp  $7,%0\n"					\
536 		"	jbe  1f\n"					\
537 		"	movl %1,%0\n"					\
538 		"	negl %0\n"					\
539 		"	andl $7,%0\n"					\
540 		"	subl %0,%3\n"					\
541 		"4:	rep; movsb\n"					\
542 		"	movl %3,%0\n"					\
543 		"	shrl $2,%0\n"					\
544 		"	andl $3,%3\n"					\
545 		"	.align 2,0x90\n"				\
546 		"0:	rep; movsl\n"					\
547 		"	movl %3,%0\n"					\
548 		"1:	rep; movsb\n"					\
549 		"2:\n"							\
550 		".section .fixup,\"ax\"\n"				\
551 		"5:	addl %3,%0\n"					\
552 		"	jmp 6f\n"					\
553 		"3:	lea 0(%3,%0,4),%0\n"				\
554 		"6:	pushl %0\n"					\
555 		"	pushl %%eax\n"					\
556 		"	xorl %%eax,%%eax\n"				\
557 		"	rep; stosb\n"					\
558 		"	popl %%eax\n"					\
559 		"	popl %0\n"					\
560 		"	jmp 2b\n"					\
561 		".previous\n"						\
562 		_ASM_EXTABLE(4b,5b)					\
563 		_ASM_EXTABLE(0b,3b)					\
564 		_ASM_EXTABLE(1b,6b)					\
565 		: "=&c"(size), "=&D" (__d0), "=&S" (__d1), "=r"(__d2)	\
566 		: "3"(size), "0"(size), "1"(to), "2"(from)		\
567 		: "memory");						\
568 } while (0)
569 
570 unsigned long __copy_to_user_ll(void __user *to, const void *from,
571 				unsigned long n)
572 {
573 	stac();
574 	if (movsl_is_ok(to, from, n))
575 		__copy_user(to, from, n);
576 	else
577 		n = __copy_user_intel(to, from, n);
578 	clac();
579 	return n;
580 }
581 EXPORT_SYMBOL(__copy_to_user_ll);
582 
583 unsigned long __copy_from_user_ll(void *to, const void __user *from,
584 					unsigned long n)
585 {
586 	stac();
587 	if (movsl_is_ok(to, from, n))
588 		__copy_user_zeroing(to, from, n);
589 	else
590 		n = __copy_user_zeroing_intel(to, from, n);
591 	clac();
592 	return n;
593 }
594 EXPORT_SYMBOL(__copy_from_user_ll);
595 
596 unsigned long __copy_from_user_ll_nozero(void *to, const void __user *from,
597 					 unsigned long n)
598 {
599 	stac();
600 	if (movsl_is_ok(to, from, n))
601 		__copy_user(to, from, n);
602 	else
603 		n = __copy_user_intel((void __user *)to,
604 				      (const void *)from, n);
605 	clac();
606 	return n;
607 }
608 EXPORT_SYMBOL(__copy_from_user_ll_nozero);
609 
610 unsigned long __copy_from_user_ll_nocache(void *to, const void __user *from,
611 					unsigned long n)
612 {
613 	stac();
614 #ifdef CONFIG_X86_INTEL_USERCOPY
615 	if (n > 64 && cpu_has_xmm2)
616 		n = __copy_user_zeroing_intel_nocache(to, from, n);
617 	else
618 		__copy_user_zeroing(to, from, n);
619 #else
620 	__copy_user_zeroing(to, from, n);
621 #endif
622 	clac();
623 	return n;
624 }
625 EXPORT_SYMBOL(__copy_from_user_ll_nocache);
626 
627 unsigned long __copy_from_user_ll_nocache_nozero(void *to, const void __user *from,
628 					unsigned long n)
629 {
630 	stac();
631 #ifdef CONFIG_X86_INTEL_USERCOPY
632 	if (n > 64 && cpu_has_xmm2)
633 		n = __copy_user_intel_nocache(to, from, n);
634 	else
635 		__copy_user(to, from, n);
636 #else
637 	__copy_user(to, from, n);
638 #endif
639 	clac();
640 	return n;
641 }
642 EXPORT_SYMBOL(__copy_from_user_ll_nocache_nozero);
643 
644 /**
645  * copy_to_user: - Copy a block of data into user space.
646  * @to:   Destination address, in user space.
647  * @from: Source address, in kernel space.
648  * @n:    Number of bytes to copy.
649  *
650  * Context: User context only. This function may sleep if pagefaults are
651  *          enabled.
652  *
653  * Copy data from kernel space to user space.
654  *
655  * Returns number of bytes that could not be copied.
656  * On success, this will be zero.
657  */
658 unsigned long _copy_to_user(void __user *to, const void *from, unsigned n)
659 {
660 	if (access_ok(VERIFY_WRITE, to, n))
661 		n = __copy_to_user(to, from, n);
662 	return n;
663 }
664 EXPORT_SYMBOL(_copy_to_user);
665 
666 /**
667  * copy_from_user: - Copy a block of data from user space.
668  * @to:   Destination address, in kernel space.
669  * @from: Source address, in user space.
670  * @n:    Number of bytes to copy.
671  *
672  * Context: User context only. This function may sleep if pagefaults are
673  *          enabled.
674  *
675  * Copy data from user space to kernel space.
676  *
677  * Returns number of bytes that could not be copied.
678  * On success, this will be zero.
679  *
680  * If some data could not be copied, this function will pad the copied
681  * data to the requested size using zero bytes.
682  */
683 unsigned long _copy_from_user(void *to, const void __user *from, unsigned n)
684 {
685 	if (access_ok(VERIFY_READ, from, n))
686 		n = __copy_from_user(to, from, n);
687 	else
688 		memset(to, 0, n);
689 	return n;
690 }
691 EXPORT_SYMBOL(_copy_from_user);
692