/Linux-v5.10/arch/arm/crypto/ |
D | sha1-armv4-large.S | 77 mov r5,r5,ror#30 78 mov r6,r6,ror#30 79 mov r7,r7,ror#30 @ [6] 85 add r7,r8,r7,ror#2 @ E+=K_00_19 90 add r7,r7,r3,ror#27 @ E+=ROR(A,27) 94 add r7,r8,r7,ror#2 @ E+=K_00_19 96 add r7,r7,r3,ror#27 @ E+=ROR(A,27) 101 and r10,r4,r10,ror#2 103 eor r10,r10,r6,ror#2 @ F_00_19(B,C,D) 110 add r6,r8,r6,ror#2 @ E+=K_00_19 [all …]
|
D | sha256-core.S_shipped | 127 eor r0,r8,r8,ror#5 129 eor r0,r0,r8,ror#19 @ Sigma1(e) 144 eor r0,r8,r8,ror#5 146 eor r0,r0,r8,ror#19 @ Sigma1(e) 152 add r11,r11,r0,ror#6 @ h+=Sigma1(e) 156 eor r0,r4,r4,ror#11 174 eor r0,r0,r4,ror#20 @ Sigma0(a) 178 add r11,r11,r0,ror#2 @ h+=Sigma0(a) 185 eor r0,r7,r7,ror#5 187 eor r0,r0,r7,ror#19 @ Sigma1(e) [all …]
|
D | chacha-scalar-core.S | 91 add \a1, \a1, \b1, ror #brot 92 add \a2, \a2, \b2, ror #brot 93 eor \d1, \a1, \d1, ror #drot 94 eor \d2, \a2, \d2, ror #drot 98 add \c1, \c1, \d1, ror #16 99 add \c2, \c2, \d2, ror #16 100 eor \b1, \c1, \b1, ror #brot 101 eor \b2, \c2, \b2, ror #brot 105 add \a1, \a1, \b1, ror #20 106 add \a2, \a2, \b2, ror #20 [all …]
|
D | aes-cipher-core.S | 60 eor \out0, \out0, t0, ror #24 82 eor \out1, \out1, t1, ror #24 83 eor \out0, \out0, t2, ror #16 85 eor \out1, \out1, \t3, ror #16 86 eor \out0, \out0, t0, ror #8 87 eor \out1, \out1, \t4, ror #8
|
/Linux-v5.10/arch/arm64/crypto/ |
D | sha512-core.S_shipped | 99 ror x16,x24,#14 101 eor x6,x24,x24,ror#23 107 eor x16,x16,x6,ror#18 // Sigma1(e) 108 ror x6,x20,#28 110 eor x17,x20,x20,ror#5 115 eor x17,x6,x17,ror#34 // Sigma0(a) 124 ror x16,x23,#14 126 eor x7,x23,x23,ror#23 132 eor x16,x16,x7,ror#18 // Sigma1(e) 133 ror x7,x27,#28 [all …]
|
D | sha256-core.S_shipped | 113 ror w16,w24,#6 115 eor w6,w24,w24,ror#14 121 eor w16,w16,w6,ror#11 // Sigma1(e) 122 ror w6,w20,#2 124 eor w17,w20,w20,ror#9 129 eor w17,w6,w17,ror#13 // Sigma0(a) 138 ror w16,w23,#6 140 eor w7,w23,w23,ror#14 146 eor w16,w16,w7,ror#11 // Sigma1(e) 147 ror w7,w27,#2 [all …]
|
D | chacha-neon-core.S | 265 ror a12, a12, #16 267 ror a13, a13, #16 269 ror a14, a14, #16 271 ror a15, a15, #16 301 ror a4, a4, #20 303 ror a5, a5, #20 305 ror a6, a6, #20 307 ror a7, a7, #20 332 ror a12, a12, #24 334 ror a13, a13, #24 [all …]
|
D | aes-cipher-core.S | 66 eor \out0, \out0, w14, ror #24 67 eor \out1, \out1, w15, ror #24 68 eor \out0, \out0, w16, ror #16 69 eor \out1, \out1, w17, ror #16 70 eor \out0, \out0, \t0, ror #8 71 eor \out1, \out1, \t1, ror #8
|
/Linux-v5.10/arch/arm/lib/ |
D | bswapsdi2.S | 19 eor r3, r0, r0, ror #16 22 eor r0, r3, r0, ror #8 28 eor r3, ip, ip, ror #16 29 eor r1, r0, r0, ror #16 34 eor r1, r1, r0, ror #8 35 eor r0, r3, ip, ror #8
|
D | csumpartial.S | 39 movne sum, sum, ror #8 74 movne r0, r0, ror #8 @ rotate checksum by 8 bits 105 movne sum, sum, ror #8
|
D | io-readsw-armv4.S | 99 _BE_ONLY_( mov ip, ip, ror #8 ) 122 _BE_ONLY_( movne ip, ip, ror #8 )
|
/Linux-v5.10/arch/x86/crypto/ |
D | sha256-ssse3-asm.S | 152 ror $(25-11), y0 # y0 = e >> (25-11) 155 ror $(22-13), y1 # y1 = a >> (22-13) 158 ror $(11-6), y0 # y0 = (e >> (11-6)) ^ (e >> (25-6)) 165 ror $(13-2), y1 # y1 = (a >> (13-2)) ^ (a >> (22-2)) 169 ror $6, y0 # y0 = S1 = (e>>6) & (e>>11) ^ (e>>25) 172 ror $2, y1 # y1 = S0 = (a>>2) ^ (a>>13) ^ (a>>22) 186 por XTMP2, XTMP1 # XTMP1 = W[-15] ror 7 195 ror $(25-11), y0 # y0 = e >> (25-11) 198 ror $(22-13), y1 # y1 = a >> (22-13) 201 ror $(11-6), y0 # y0 = (e >> (11-6)) ^ (e >> (25-6)) [all …]
|
D | sha512-ssse3-asm.S | 124 ror $23, tmp0 # 41 # tmp = e ror 23 126 xor e_64, tmp0 # tmp = (e ror 23) ^ e 130 ror $4, tmp0 # 18 # tmp = ((e ror 23) ^ e) ror 4 131 xor e_64, tmp0 # tmp = (((e ror 23) ^ e) ror 4) ^ e 134 ror $14, tmp0 # 14 # tmp = ((((e ror23)^e)ror4)^e)ror14 = S1(e) 142 ror $5, tmp0 # 39 # tmp = a ror 5 143 xor a_64, tmp0 # tmp = (a ror 5) ^ a 145 ror $6, tmp0 # 34 # tmp = ((a ror 5) ^ a) ror 6 146 xor a_64, tmp0 # tmp = (((a ror 5) ^ a) ror 6) ^ a 148 ror $28, tmp0 # 28 # tmp = ((((a ror5)^a)ror6)^a)ror28 = S0(a) [all …]
|
D | twofish-i586-asm_32.S | 74 ror $16, b ## D;\ 77 ror $16, a ## D;\ 84 ror $15, b ## D;\ 112 ror $16, b ## D;\ 115 ror $16, a ## D;\ 122 ror $16, b ## D;\ 131 ror $1, c ## D;\ 149 ror $16, a ## D;\ 152 ror $16, b ## D;\ 159 ror $15, a ## D;\ [all …]
|
D | twofish-x86_64-asm_64.S | 72 ror $16, b ## D;\ 75 ror $16, a ## D;\ 82 ror $15, b ## D;\ 110 ror $16, b ## D;\ 113 ror $16, a ## D;\ 128 ror $1, c ## D;\ 145 ror $16, a ## D;\ 148 ror $16, b ## D;\ 155 ror $15, a ## D;\ 181 ror $16, b ## D;\ [all …]
|
D | sha512-avx-asm.S | 121 # shld is faster than ror on Sandybridge 130 RORQ tmp0, 23 # 41 # tmp = e ror 23 132 xor e_64, tmp0 # tmp = (e ror 23) ^ e 136 RORQ tmp0, 4 # 18 # tmp = ((e ror 23) ^ e) ror 4 137 xor e_64, tmp0 # tmp = (((e ror 23) ^ e) ror 4) ^ e 148 RORQ tmp0, 5 # 39 # tmp = a ror 5 149 xor a_64, tmp0 # tmp = (a ror 5) ^ a 151 RORQ tmp0, 6 # 34 # tmp = ((a ror 5) ^ a) ror 6 152 xor a_64, tmp0 # tmp = (((a ror 5) ^ a) ror 6) ^ a
|
D | sha512-avx2-asm.S | 178 # Calculate w[t-15] ror 1 181 vpor YTMP2, YTMP3, YTMP3 # YTMP3 = W[-15] ror 1 224 # Calculate w[t-15] ror 8 227 vpor YTMP2, YTMP1, YTMP1 # YTMP1 = W[-15] ror 8 229 vpxor YTMP4, YTMP3, YTMP3 # YTMP3 = W[-15] ror 1 ^ W[-15] >> 7 291 vpor YTMP1, YTMP3, YTMP3 # YTMP3 = W[-2] ror 19 {BABA} 292 vpxor YTMP3, YTMP4, YTMP4 # YTMP4 = W[-2] ror 19 ^ W[-2] >> 6 {BABA} 295 vpor YTMP1, YTMP3, YTMP3 # YTMP3 = W[-2] ror 61 {BABA} 296 vpxor YTMP3, YTMP4, YTMP4 # YTMP4 = s1 = (W[-2] ror 19) ^ 297 # (W[-2] ror 61) ^ (W[-2] >> 6) {BABA} [all …]
|
D | sha256-avx2-asm.S | 195 vpor XTMP2, XTMP3, XTMP3 # XTMP3 = W[-15] ror 7 235 vpxor XTMP2, XTMP3, XTMP3 # XTMP3 = W[-15] ror 7 ^ W[-15] ror 18 263 vpsrlq $19, XTMP2, XTMP3 # XTMP3 = W[-2] ror 19 {xBxA} 271 vpsrlq $17, XTMP2, XTMP2 # XTMP2 = W[-2] ror 17 {xBxA} 322 vpsrlq $19, XTMP2, XTMP3 # XTMP3 = W[-2] ror 19 {xDxC} 328 vpsrlq $17, XTMP2, XTMP2 # XTMP2 = W[-2] ror 17 {xDxC}
|
/Linux-v5.10/arch/arc/lib/ |
D | strlen.S | 18 ror r5,r4 28 ror r5,r4
|
D | strchr-700.S | 30 ror r4,r3 55 ror r4,r3
|
D | strcpy-700.S | 26 ror r12,r8
|
D | strcmp-archs.S | 17 ror r11, r12
|
/Linux-v5.10/arch/arm64/kvm/hyp/nvhe/ |
D | hyp-init.S | 66 ror x0, x0, #24 68 ror x0, x0, #4
|
/Linux-v5.10/arch/arm64/include/asm/ |
D | kvm_mmu.h | 68 ror \reg, \reg, #1 /* rotate to the first tag bit */ 71 ror \reg, \reg, #63 /* rotate back */
|
D | asm-uaccess.h | 31 ror \tmp2, \tmp2, #16
|