/Linux-v5.10/arch/arm64/crypto/ |
D | chacha-neon-core.S | 42 ld1 {v12.4s}, [x10] 46 add v0.4s, v0.4s, v1.4s 47 eor v3.16b, v3.16b, v0.16b 51 add v2.4s, v2.4s, v3.4s 52 eor v4.16b, v1.16b, v2.16b 53 shl v1.4s, v4.4s, #12 54 sri v1.4s, v4.4s, #20 57 add v0.4s, v0.4s, v1.4s 58 eor v3.16b, v3.16b, v0.16b 59 tbl v3.16b, {v3.16b}, v12.16b [all …]
|
D | sm3-ce-core.S | 11 .irp b, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12 12 .set .Lv\b\().4s, \b 44 sm3ss1 v5.4s, v8.4s, \t0\().4s, v9.4s 45 shl \t1\().4s, \t0\().4s, #1 46 sri \t1\().4s, \t0\().4s, #31 47 sm3tt1\ab v8.4s, v5.4s, v10.4s, \i 48 sm3tt2\ab v9.4s, v5.4s, \s0\().4s, \i 53 ext \s4\().16b, \s1\().16b, \s2\().16b, #12 54 ext v6.16b, \s0\().16b, \s1\().16b, #12 55 ext v7.16b, \s2\().16b, \s3\().16b, #8 [all …]
|
D | sha256-core.S_shipped | 42 // on Cortex-A53 (or by 4 cycles per round). 83 b.ne .Lv8_entry 85 b.ne .Lneon_entry 95 sub sp,sp,#4*4 98 ldp w22,w23,[x0,#2*4] 99 ldp w24,w25,[x0,#4*4] 101 ldp w26,w27,[x0,#6*4] 106 ldp w3,w4,[x1],#2*4 107 ldr w19,[x30],#4 // *K++ 120 eor w19,w20,w21 // a^b, b^c in next round [all …]
|
D | aes-ce-core.S | 13 ld1 {v0.16b}, [x2] 14 ld1 {v1.4s}, [x0], #16 18 mov v3.16b, v1.16b 19 b 2f 20 0: mov v2.16b, v1.16b 21 ld1 {v3.4s}, [x0], #16 22 1: aese v0.16b, v2.16b 23 aesmc v0.16b, v0.16b 24 2: ld1 {v1.4s}, [x0], #16 25 aese v0.16b, v3.16b [all …]
|
D | aes-ce-ccm-core.S | 20 ld1 {v0.16b}, [x0] /* load mac */ 23 eor v1.16b, v1.16b, v1.16b 27 ins v1.b[0], w7 28 ext v1.16b, v1.16b, v1.16b, #1 /* rotate in the input bytes */ 30 cbnz w8, 0b 31 eor v0.16b, v0.16b, v1.16b 32 1: ld1 {v3.4s}, [x4] /* load first round key */ 39 mov v5.16b, v3.16b 40 b 4f 41 2: mov v4.16b, v3.16b [all …]
|
D | sha512-ce-core.S | 15 .irp b,0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19 16 .set .Lq\b, \b 17 .set .Lv\b\().2d, \b 40 .align 4 88 ext v6.16b, v\i2\().16b, v\i3\().16b, #8 89 ext v5.16b, v5.16b, v5.16b, #8 90 ext v7.16b, v\i1\().16b, v\i2\().16b, #8 93 ext v5.16b, v\in3\().16b, v\in4\().16b, #8 119 /* load first 4 round constants */ 128 CPU_LE( rev64 v12.16b, v12.16b ) [all …]
|
D | ghash-ce-core.S | 71 ext t3.8b, \ad\().8b, \ad\().8b, #1 // A1 72 ext t5.8b, \ad\().8b, \ad\().8b, #2 // A2 73 ext t7.8b, \ad\().8b, \ad\().8b, #3 // A3 79 tbl t3.16b, {\ad\().16b}, perm1.16b // A1 80 tbl t5.16b, {\ad\().16b}, perm2.16b // A2 81 tbl t7.16b, {\ad\().16b}, perm3.16b // A3 87 __pmull_p8_tail \rq, \ad\().8b, SHASH.8b, 8b,, sh1, sh2, sh3, sh4 91 __pmull_p8_tail \rq, \ad\().8b, SHASH2.8b, 8b,, ss1, ss2, ss3, ss4 95 __pmull_p8_tail \rq, \ad\().16b, SHASH.16b, 16b, 2, sh1, sh2, sh3, sh4 99 pmull\t t3.8h, t3.\nb, \bd // F = A1*B [all …]
|
D | sha2-ce-core.S | 30 mov dg2v.16b, dg0v.16b 32 add t1.4s, v\s0\().4s, \rc\().4s 33 sha256h dg0q, dg1q, t0.4s 34 sha256h2 dg1q, dg2q, t0.4s 37 add t0.4s, v\s0\().4s, \rc\().4s 39 sha256h dg0q, dg1q, t1.4s 40 sha256h2 dg1q, dg2q, t1.4s 45 sha256su0 v\s0\().4s, v\s1\().4s 47 sha256su1 v\s0\().4s, v\s2\().4s, v\s3\().4s 54 .align 4 [all …]
|
/Linux-v5.10/arch/x86/lib/ |
D | usercopy_32.c | 47 "3: lea 0(%2,%0,4),%0\n" \ 48 " jmp 2b\n" \ 50 _ASM_EXTABLE_UA(0b, 3b) \ 51 _ASM_EXTABLE_UA(1b, 2b) \ 53 : "r"(size & 3), "0"(size / 4), "1"(addr), "a"(0)); \ 102 "1: movl 32(%4), %%eax\n" in __copy_user_intel() 105 "2: movl 64(%4), %%eax\n" in __copy_user_intel() 107 "3: movl 0(%4), %%eax\n" in __copy_user_intel() 108 "4: movl 4(%4), %%edx\n" in __copy_user_intel() 110 "6: movl %%edx, 4(%3)\n" in __copy_user_intel() [all …]
|
D | copy_user_64.S | 33 jnz 100b 40 _ASM_EXTABLE_CPY(100b, 103b) 41 _ASM_EXTABLE_CPY(101b, 103b) 69 4: movq 3*8(%rsi),%r11 74 9: movq 4*8(%rsi),%r8 78 13: movq %r8,4*8(%rdi) 85 jnz 1b 96 jnz 18b 105 jnz 21b 120 _ASM_EXTABLE_CPY(1b, 30b) [all …]
|
/Linux-v5.10/arch/arc/include/asm/ |
D | uaccess.h | 63 case 4: __arc_get_user_one(*(k), u, "ld", __ret); break; \ 80 " .align 4\n" \ 85 " j 2b\n" \ 88 " .align 4\n" \ 89 " .word 1b,3b\n" \ 98 "4: ld %R1,[%2, 4]\n" \ 101 " .align 4\n" \ 107 " j 2b\n" \ 110 " .align 4\n" \ 111 " .word 1b,3b\n" \ [all …]
|
/Linux-v5.10/Documentation/userspace-api/media/v4l/ |
D | pixfmt-rgb.rst | 41 - 4 50 - 4 59 - 4 68 - 4 84 - b\ :sub:`1` 85 - b\ :sub:`0` 96 - b\ :sub:`3` 97 - b\ :sub:`2` 98 - b\ :sub:`1` 99 - b\ :sub:`0` [all …]
|
/Linux-v5.10/drivers/staging/comedi/drivers/ni_routing/ni_route_values/ |
D | ni_eseries.c | 54 [B(NI_PFI(0))] = { 55 [B(NI_AI_StartTrigger)] = I(NI_PFI_OUTPUT_AI_START1), 57 [B(NI_PFI(1))] = { 58 [B(NI_AI_ReferenceTrigger)] = I(NI_PFI_OUTPUT_AI_START2), 60 [B(NI_PFI(2))] = { 61 [B(NI_AI_ConvertClock)] = I(NI_PFI_OUTPUT_AI_CONVERT), 63 [B(NI_PFI(3))] = { 64 [B(NI_CtrSource(1))] = I(NI_PFI_OUTPUT_G_SRC1), 66 [B(NI_PFI(4))] = { 67 [B(NI_CtrGate(1))] = I(NI_PFI_OUTPUT_G_GATE1), [all …]
|
D | ni_660x.c | 49 [B(NI_PFI(8))] = { 50 [B(NI_CtrInternalOutput(7))] = I(1), 52 [B(NI_PFI(10))] = { 53 [B(NI_CtrGate(7))] = I(1), 55 [B(NI_PFI(11))] = { 56 [B(NI_CtrSource(7))] = I(1), 58 [B(NI_PFI(12))] = { 59 [B(NI_CtrInternalOutput(6))] = I(1), 61 [B(NI_PFI(14))] = { 62 [B(NI_CtrGate(6))] = I(1), [all …]
|
D | ni_mseries.c | 57 [B(NI_PFI(0))] = { 58 [B(TRIGGER_LINE(0))] = I(18), 59 [B(TRIGGER_LINE(1))] = I(19), 60 [B(TRIGGER_LINE(2))] = I(20), 61 [B(TRIGGER_LINE(3))] = I(21), 62 [B(TRIGGER_LINE(4))] = I(22), 63 [B(TRIGGER_LINE(5))] = I(23), 64 [B(TRIGGER_LINE(6))] = I(24), 65 [B(TRIGGER_LINE(7))] = I(25), 66 [B(NI_CtrSource(0))] = I(9), [all …]
|
/Linux-v5.10/arch/powerpc/lib/ |
D | copy_32.S | 16 lwz r7,4(r4); \ 20 stw r7,4(r6); \ 27 lwz r7,4(r4); \ 34 8 ## n ## 4: \ 35 stw r7,4(r6); \ 46 b 104f; \ 49 b 105f; \ 50 EX_TABLE(8 ## n ## 0b,9 ## n ## 0b); \ 51 EX_TABLE(8 ## n ## 1b,9 ## n ## 0b); \ 52 EX_TABLE(8 ## n ## 2b,9 ## n ## 0b); \ [all …]
|
/Linux-v5.10/arch/m68k/lib/ |
D | uaccess.c | 21 " jne 1b\n" in __generic_copy_from_user() 23 " jeq 4f\n" in __generic_copy_from_user() 26 "4: btst #0,%5\n" in __generic_copy_from_user() 28 "5: "MOVES".b (%1)+,%3\n" in __generic_copy_from_user() 29 " move.b %3,(%2)+\n" in __generic_copy_from_user() 38 " jeq 6b\n" in __generic_copy_from_user() 40 " jra 6b\n" in __generic_copy_from_user() 44 " .align 4\n" in __generic_copy_from_user() 45 " .long 1b,10b\n" in __generic_copy_from_user() 46 " .long 3b,30b\n" in __generic_copy_from_user() [all …]
|
/Linux-v5.10/arch/s390/lib/ |
D | uaccess.c | 113 "6: jz 4f\n" in copy_from_user_mvcos() 117 " j 0b\n" in copy_from_user_mvcos() 118 "2: la %4,4095(%1)\n"/* %4 = ptr + 4095 */ in copy_from_user_mvcos() 119 " nr %4,%3\n" /* %4 = (ptr + 4095) & -4096 */ in copy_from_user_mvcos() 120 " slgr %4,%1\n" in copy_from_user_mvcos() 121 " clgr %0,%4\n" /* copy crosses next page boundary? */ in copy_from_user_mvcos() 123 "3: .insn ss,0xc80000000000,0(%4,%2),0(%1),0\n" in copy_from_user_mvcos() 124 "7: slgr %0,%4\n" in copy_from_user_mvcos() 126 "4: slgr %0,%0\n" in copy_from_user_mvcos() 128 EX_TABLE(0b,2b) EX_TABLE(3b,5b) EX_TABLE(6b,2b) EX_TABLE(7b,5b) in copy_from_user_mvcos() [all …]
|
/Linux-v5.10/arch/x86/crypto/ |
D | twofish-x86_64-asm_64.S | 14 #define b_offset 4 61 * b input register containing b 64 * operations on a and b are interleaved to increase performance 66 #define encrypt_round(a,b,c,d,round)\ argument 67 movzx b ## B, %edi;\ 68 mov s1(%r11,%rdi,4),%r8d;\ 69 movzx a ## B, %edi;\ 70 mov s2(%r11,%rdi,4),%r9d;\ 71 movzx b ## H, %edi;\ 72 ror $16, b ## D;\ [all …]
|
D | twofish-i586-asm_32.S | 17 #define ctx 4 /* Twofish context structure */ 20 #define b_offset 4 62 * b input register containing b 65 * operations on a and b are interleaved to increase performance 67 #define encrypt_round(a,b,c,d,round)\ argument 69 movzx b ## B, %edi;\ 70 mov s1(%ebp,%edi,4),d ## D;\ 71 movzx a ## B, %edi;\ 72 mov s2(%ebp,%edi,4),%esi;\ 73 movzx b ## H, %edi;\ [all …]
|
/Linux-v5.10/arch/parisc/lib/ |
D | lusercopy.S | 63 2: b $lclu_done 66 ASM_EXCEPTIONTABLE_ENTRY(1b,2b) 92 b $lslen_done 95 3: b $lslen_done 98 ASM_EXCEPTIONTABLE_ENTRY(1b,3b) 99 ASM_EXCEPTIONTABLE_ENTRY(2b,3b) 183 b .Lalign_loop64 186 ASM_EXCEPTIONTABLE_ENTRY(20b,.Lcopy_done) 187 ASM_EXCEPTIONTABLE_ENTRY(21b,.Lcopy_done) 205 ASM_EXCEPTIONTABLE_ENTRY(10b,.Lcopy_done) [all …]
|
/Linux-v5.10/arch/powerpc/boot/ |
D | string.S | 18 bne 1b 31 bdnzf 2,1b /* dec ctr, branch if ctr != 0 && !cr0.eq */ 40 bne 1b 45 bne 1b 55 bne 1b 68 beq 1b 81 bdnzt eq,1b 89 bne 1b 97 addi r6,r3,-4 98 cmplwi 0,r5,4 [all …]
|
/Linux-v5.10/arch/arm/mm/ |
D | abort-lv4t.S | 37 /* 0 */ b .data_arm_lateldrhpost @ ldrh rd, [rn], #m/rm 38 /* 1 */ b .data_arm_lateldrhpre @ ldrh rd, [rn, #m/rm] 39 /* 2 */ b .data_unknown 40 /* 3 */ b .data_unknown 41 /* 4 */ b .data_arm_lateldrpostconst @ ldr rd, [rn], #m 42 /* 5 */ b .data_arm_lateldrpreconst @ ldr rd, [rn, #m] 43 /* 6 */ b .data_arm_lateldrpostreg @ ldr rd, [rn], rm 44 /* 7 */ b .data_arm_lateldrprereg @ ldr rd, [rn, rm] 45 /* 8 */ b .data_arm_ldmstm @ ldm*a rn, <rlist> 46 /* 9 */ b .data_arm_ldmstm @ ldm*b rn, <rlist> [all …]
|
/Linux-v5.10/arch/arm/crypto/ |
D | sha256-core.S_shipped | 113 sub sp,sp,#16*4 @ alloca(X[16]) 116 ldr r2,[r1],#4 123 @ ldr r2,[r1],#4 @ 0 125 str r1,[sp,#17*4] @ make room for r1 128 add r4,r4,r12 @ h+=Maj(a,b,c) from the past 135 add r4,r4,r12 @ h+=Maj(a,b,c) from the past 139 ldrb r12,[r1],#4 142 str r1,[sp,#17*4] @ make room for r1 148 ldr r12,[r14],#4 @ *K256++ 150 str r2,[sp,#0*4] [all …]
|
/Linux-v5.10/arch/mips/include/asm/ |
D | unaligned-emul.h | 19 "4:\tli\t%1, %3\n\t" \ 23 STR(PTR)"\t1b, 4b\n\t" \ 24 STR(PTR)"\t2b, 4b\n\t" \ 40 "4:\tli\t%1, %3\n\t" \ 44 STR(PTR)"\t1b, 4b\n\t" \ 45 STR(PTR)"\t2b, 4b\n\t" \ 65 "4:"type##_lbu("$1", "3(%2)")"\n\t" \ 77 STR(PTR)"\t1b, 11b\n\t" \ 78 STR(PTR)"\t2b, 11b\n\t" \ 79 STR(PTR)"\t3b, 11b\n\t" \ [all …]
|