Lines Matching +full:- +full:30
1 /* SPDX-License-Identifier: GPL-2.0-only */
11 #include <asm/asm-offsets.h>
41 * copy_user_generic_unrolled - memory copy with exception handling.
84 30: shll $6,%ecx
88 _ASM_EXTABLE_CPY(1b, 30b)
89 _ASM_EXTABLE_CPY(2b, 30b)
90 _ASM_EXTABLE_CPY(3b, 30b)
91 _ASM_EXTABLE_CPY(4b, 30b)
92 _ASM_EXTABLE_CPY(5b, 30b)
93 _ASM_EXTABLE_CPY(6b, 30b)
94 _ASM_EXTABLE_CPY(7b, 30b)
95 _ASM_EXTABLE_CPY(8b, 30b)
96 _ASM_EXTABLE_CPY(9b, 30b)
97 _ASM_EXTABLE_CPY(10b, 30b)
98 _ASM_EXTABLE_CPY(11b, 30b)
99 _ASM_EXTABLE_CPY(12b, 30b)
100 _ASM_EXTABLE_CPY(13b, 30b)
101 _ASM_EXTABLE_CPY(14b, 30b)
102 _ASM_EXTABLE_CPY(15b, 30b)
103 _ASM_EXTABLE_CPY(16b, 30b)
264 * copy_user_nocache - Uncached memory copy with exception handling
269 * - Require 8-byte alignment when size is 8 bytes or larger.
270 * - Require 4-byte alignment when size is 4 bytes.
275 /* If size is less than 8 bytes, go to 4-byte copy */
279 /* If destination is not 8-byte aligned, "cache" copy to align it */
282 /* Set 4x8-byte copy count and remainder */
288 /* Perform 4x8-byte nocache loop-copy */
311 /* Set 8-byte copy count and remainder */
318 /* Perform 8-byte nocache loop-copy */
332 /* If destination is not 4-byte aligned, go to byte copy: */
337 /* Set 4-byte copy count (1 or 0) and remainder */
343 /* Perform 4-byte nocache copy: */
344 30: movl (%rsi),%r8d
353 /* Perform byte "cache" loop-copy for the remainder */
405 _ASM_EXTABLE_CPY(30b, .L_fixup_4b_copy)