Lines Matching full:n
20 static inline int __movsl_is_ok(unsigned long a1, unsigned long a2, unsigned long n) in __movsl_is_ok() argument
23 if (n >= 64 && ((a1 ^ a2) & movsl_mask.mask)) in __movsl_is_ok()
28 #define movsl_is_ok(a1, a2, n) \ argument
29 __movsl_is_ok((unsigned long)(a1), (unsigned long)(a2), (n))
40 ASM_STAC "\n" \
41 "0: rep; stosl\n" \
42 " movl %2,%0\n" \
43 "1: rep; stosb\n" \
44 "2: " ASM_CLAC "\n" \
54 * @n: Number of bytes to zero.
62 clear_user(void __user *to, unsigned long n) in clear_user() argument
65 if (access_ok(to, n)) in clear_user()
66 __do_clear_user(to, n); in clear_user()
67 return n; in clear_user()
74 * @n: Number of bytes to zero.
83 __clear_user(void __user *to, unsigned long n) in __clear_user() argument
85 __do_clear_user(to, n); in __clear_user()
86 return n; in __clear_user()
96 " .align 2,0x90\n" in __copy_user_intel()
97 "1: movl 32(%4), %%eax\n" in __copy_user_intel()
98 " cmpl $67, %0\n" in __copy_user_intel()
99 " jbe 3f\n" in __copy_user_intel()
100 "2: movl 64(%4), %%eax\n" in __copy_user_intel()
101 " .align 2,0x90\n" in __copy_user_intel()
102 "3: movl 0(%4), %%eax\n" in __copy_user_intel()
103 "4: movl 4(%4), %%edx\n" in __copy_user_intel()
104 "5: movl %%eax, 0(%3)\n" in __copy_user_intel()
105 "6: movl %%edx, 4(%3)\n" in __copy_user_intel()
106 "7: movl 8(%4), %%eax\n" in __copy_user_intel()
107 "8: movl 12(%4),%%edx\n" in __copy_user_intel()
108 "9: movl %%eax, 8(%3)\n" in __copy_user_intel()
109 "10: movl %%edx, 12(%3)\n" in __copy_user_intel()
110 "11: movl 16(%4), %%eax\n" in __copy_user_intel()
111 "12: movl 20(%4), %%edx\n" in __copy_user_intel()
112 "13: movl %%eax, 16(%3)\n" in __copy_user_intel()
113 "14: movl %%edx, 20(%3)\n" in __copy_user_intel()
114 "15: movl 24(%4), %%eax\n" in __copy_user_intel()
115 "16: movl 28(%4), %%edx\n" in __copy_user_intel()
116 "17: movl %%eax, 24(%3)\n" in __copy_user_intel()
117 "18: movl %%edx, 28(%3)\n" in __copy_user_intel()
118 "19: movl 32(%4), %%eax\n" in __copy_user_intel()
119 "20: movl 36(%4), %%edx\n" in __copy_user_intel()
120 "21: movl %%eax, 32(%3)\n" in __copy_user_intel()
121 "22: movl %%edx, 36(%3)\n" in __copy_user_intel()
122 "23: movl 40(%4), %%eax\n" in __copy_user_intel()
123 "24: movl 44(%4), %%edx\n" in __copy_user_intel()
124 "25: movl %%eax, 40(%3)\n" in __copy_user_intel()
125 "26: movl %%edx, 44(%3)\n" in __copy_user_intel()
126 "27: movl 48(%4), %%eax\n" in __copy_user_intel()
127 "28: movl 52(%4), %%edx\n" in __copy_user_intel()
128 "29: movl %%eax, 48(%3)\n" in __copy_user_intel()
129 "30: movl %%edx, 52(%3)\n" in __copy_user_intel()
130 "31: movl 56(%4), %%eax\n" in __copy_user_intel()
131 "32: movl 60(%4), %%edx\n" in __copy_user_intel()
132 "33: movl %%eax, 56(%3)\n" in __copy_user_intel()
133 "34: movl %%edx, 60(%3)\n" in __copy_user_intel()
134 " addl $-64, %0\n" in __copy_user_intel()
135 " addl $64, %4\n" in __copy_user_intel()
136 " addl $64, %3\n" in __copy_user_intel()
137 " cmpl $63, %0\n" in __copy_user_intel()
138 " ja 1b\n" in __copy_user_intel()
139 "35: movl %0, %%eax\n" in __copy_user_intel()
140 " shrl $2, %0\n" in __copy_user_intel()
141 " andl $3, %%eax\n" in __copy_user_intel()
142 " cld\n" in __copy_user_intel()
143 "99: rep; movsl\n" in __copy_user_intel()
144 "36: movl %%eax, %0\n" in __copy_user_intel()
145 "37: rep; movsb\n" in __copy_user_intel()
146 "100:\n" in __copy_user_intel()
197 " .align 2,0x90\n" in __copy_user_intel_nocache()
198 "0: movl 32(%4), %%eax\n" in __copy_user_intel_nocache()
199 " cmpl $67, %0\n" in __copy_user_intel_nocache()
200 " jbe 2f\n" in __copy_user_intel_nocache()
201 "1: movl 64(%4), %%eax\n" in __copy_user_intel_nocache()
202 " .align 2,0x90\n" in __copy_user_intel_nocache()
203 "2: movl 0(%4), %%eax\n" in __copy_user_intel_nocache()
204 "21: movl 4(%4), %%edx\n" in __copy_user_intel_nocache()
205 " movnti %%eax, 0(%3)\n" in __copy_user_intel_nocache()
206 " movnti %%edx, 4(%3)\n" in __copy_user_intel_nocache()
207 "3: movl 8(%4), %%eax\n" in __copy_user_intel_nocache()
208 "31: movl 12(%4),%%edx\n" in __copy_user_intel_nocache()
209 " movnti %%eax, 8(%3)\n" in __copy_user_intel_nocache()
210 " movnti %%edx, 12(%3)\n" in __copy_user_intel_nocache()
211 "4: movl 16(%4), %%eax\n" in __copy_user_intel_nocache()
212 "41: movl 20(%4), %%edx\n" in __copy_user_intel_nocache()
213 " movnti %%eax, 16(%3)\n" in __copy_user_intel_nocache()
214 " movnti %%edx, 20(%3)\n" in __copy_user_intel_nocache()
215 "10: movl 24(%4), %%eax\n" in __copy_user_intel_nocache()
216 "51: movl 28(%4), %%edx\n" in __copy_user_intel_nocache()
217 " movnti %%eax, 24(%3)\n" in __copy_user_intel_nocache()
218 " movnti %%edx, 28(%3)\n" in __copy_user_intel_nocache()
219 "11: movl 32(%4), %%eax\n" in __copy_user_intel_nocache()
220 "61: movl 36(%4), %%edx\n" in __copy_user_intel_nocache()
221 " movnti %%eax, 32(%3)\n" in __copy_user_intel_nocache()
222 " movnti %%edx, 36(%3)\n" in __copy_user_intel_nocache()
223 "12: movl 40(%4), %%eax\n" in __copy_user_intel_nocache()
224 "71: movl 44(%4), %%edx\n" in __copy_user_intel_nocache()
225 " movnti %%eax, 40(%3)\n" in __copy_user_intel_nocache()
226 " movnti %%edx, 44(%3)\n" in __copy_user_intel_nocache()
227 "13: movl 48(%4), %%eax\n" in __copy_user_intel_nocache()
228 "81: movl 52(%4), %%edx\n" in __copy_user_intel_nocache()
229 " movnti %%eax, 48(%3)\n" in __copy_user_intel_nocache()
230 " movnti %%edx, 52(%3)\n" in __copy_user_intel_nocache()
231 "14: movl 56(%4), %%eax\n" in __copy_user_intel_nocache()
232 "91: movl 60(%4), %%edx\n" in __copy_user_intel_nocache()
233 " movnti %%eax, 56(%3)\n" in __copy_user_intel_nocache()
234 " movnti %%edx, 60(%3)\n" in __copy_user_intel_nocache()
235 " addl $-64, %0\n" in __copy_user_intel_nocache()
236 " addl $64, %4\n" in __copy_user_intel_nocache()
237 " addl $64, %3\n" in __copy_user_intel_nocache()
238 " cmpl $63, %0\n" in __copy_user_intel_nocache()
239 " ja 0b\n" in __copy_user_intel_nocache()
240 " sfence \n" in __copy_user_intel_nocache()
241 "5: movl %0, %%eax\n" in __copy_user_intel_nocache()
242 " shrl $2, %0\n" in __copy_user_intel_nocache()
243 " andl $3, %%eax\n" in __copy_user_intel_nocache()
244 " cld\n" in __copy_user_intel_nocache()
245 "6: rep; movsl\n" in __copy_user_intel_nocache()
246 " movl %%eax,%0\n" in __copy_user_intel_nocache()
247 "7: rep; movsb\n" in __copy_user_intel_nocache()
248 "8:\n" in __copy_user_intel_nocache()
290 " cmp $7,%0\n" \
291 " jbe 1f\n" \
292 " movl %1,%0\n" \
293 " negl %0\n" \
294 " andl $7,%0\n" \
295 " subl %0,%3\n" \
296 "4: rep; movsb\n" \
297 " movl %3,%0\n" \
298 " shrl $2,%0\n" \
299 " andl $3,%3\n" \
300 " .align 2,0x90\n" \
301 "0: rep; movsl\n" \
302 " movl %3,%0\n" \
303 "1: rep; movsb\n" \
304 "2:\n" \
313 unsigned long __copy_user_ll(void *to, const void *from, unsigned long n) in __copy_user_ll() argument
316 if (movsl_is_ok(to, from, n)) in __copy_user_ll()
317 __copy_user(to, from, n); in __copy_user_ll()
319 n = __copy_user_intel(to, from, n); in __copy_user_ll()
321 return n; in __copy_user_ll()
326 unsigned long n) in __copy_from_user_ll_nocache_nozero() argument
330 if (n > 64 && static_cpu_has(X86_FEATURE_XMM2)) in __copy_from_user_ll_nocache_nozero()
331 n = __copy_user_intel_nocache(to, from, n); in __copy_from_user_ll_nocache_nozero()
333 __copy_user(to, from, n); in __copy_from_user_ll_nocache_nozero()
335 __copy_user(to, from, n); in __copy_from_user_ll_nocache_nozero()
338 return n; in __copy_from_user_ll_nocache_nozero()