| /Linux-v5.4/arch/x86/crypto/ |
| D | glue_helper-asm-avx.S | 8 #define load_8way(src, x0, x1, x2, x3, x4, x5, x6, x7) \ argument 15 vmovdqu (6*16)(src), x6; \ 18 #define store_8way(dst, x0, x1, x2, x3, x4, x5, x6, x7) \ argument 25 vmovdqu x6, (6*16)(dst); \ 28 #define store_cbc_8way(src, dst, x0, x1, x2, x3, x4, x5, x6, x7) \ argument 34 vpxor (5*16)(src), x6, x6; \ 36 store_8way(dst, x0, x1, x2, x3, x4, x5, x6, x7); 44 #define load_ctr_8way(iv, bswap, x0, x1, x2, x3, x4, x5, x6, x7, t0, t1, t2) \ argument 65 vpshufb t1, x7, x6; \ 72 #define store_ctr_8way(src, dst, x0, x1, x2, x3, x4, x5, x6, x7) \ argument [all …]
|
| D | glue_helper-asm-avx2.S | 8 #define load_16way(src, x0, x1, x2, x3, x4, x5, x6, x7) \ argument 15 vmovdqu (6*32)(src), x6; \ 18 #define store_16way(dst, x0, x1, x2, x3, x4, x5, x6, x7) \ argument 25 vmovdqu x6, (6*32)(dst); \ 28 #define store_cbc_16way(src, dst, x0, x1, x2, x3, x4, x5, x6, x7, t0) \ argument 37 vpxor (5*32+16)(src), x6, x6; \ 39 store_16way(dst, x0, x1, x2, x3, x4, x5, x6, x7); 55 #define load_ctr_16way(iv, bswap, x0, x1, x2, x3, x4, x5, x6, x7, t0, t0x, t1, \ argument 81 vpshufb t1, t2, x6; \ 88 #define store_ctr_16way(src, dst, x0, x1, x2, x3, x4, x5, x6, x7) \ argument [all …]
|
| D | camellia-aesni-avx-asm_64.S | 51 #define roundsm16(x0, x1, x2, x3, x4, x5, x6, x7, t0, t1, t2, t3, t4, t5, t6, \ argument 69 vpshufb t4, x6, x6; \ 84 filter_8bit(x6, t2, t3, t7, t6); \ 96 vaesenclast t4, x6, x6; \ 104 filter_8bit(x6, t0, t1, t7, t6); \ 136 vpxor x6, x1, x1; \ 142 vpxor x0, x6, x6; \ 148 vpxor x6, x3, x3; \ 152 vpxor x1, x6, x6; \ 169 vpxor t1, x6, x6; \ [all …]
|
| D | camellia-aesni-avx2-asm_64.S | 63 #define roundsm32(x0, x1, x2, x3, x4, x5, x6, x7, t0, t1, t2, t3, t4, t5, t6, \ argument 79 vpshufb t4, x6, x6; \ 92 filter_8bit(x6, t2, t3, t7, t4); \ 94 vextracti128 $1, x6, t2##_x; \ 114 vaesenclast t4##_x, x6##_x, x6##_x; \ 116 vinserti128 $1, t2##_x, x6, x6; \ 140 filter_8bit(x6, t0, t1, t7, t6); \ 162 vpxor x6, x1, x1; \ 174 vpxor x0, x6, x6; \ 184 vpxor x6, x3, x3; \ [all …]
|
| D | chacha-ssse3-x86_64.S | 296 # x2 += x6, x14 = rotl32(x14 ^ x2, 16) 323 # x10 += x14, x6 = rotl32(x6 ^ x10, 12) 350 # x2 += x6, x14 = rotl32(x14 ^ x2, 8) 377 # x10 += x14, x6 = rotl32(x6 ^ x10, 7) 398 # x1 += x6, x12 = rotl32(x12 ^ x1, 16) 424 # x11 += x12, x6 = rotl32(x6 ^ x11, 12) 452 # x1 += x6, x12 = rotl32(x12 ^ x1, 8) 478 # x11 += x12, x6 = rotl32(x6 ^ x11, 7) 529 # x6[0-3] += s1[2]
|
| D | chacha-avx512vl-x86_64.S | 519 # x2 += x6, x14 = rotl32(x14 ^ x2, 16) 536 # x10 += x14, x6 = rotl32(x6 ^ x10, 12) 553 # x2 += x6, x14 = rotl32(x14 ^ x2, 8) 570 # x10 += x14, x6 = rotl32(x6 ^ x10, 7) 583 # x1 += x6, x12 = rotl32(x12 ^ x1, 16) 600 # x11 += x12, x6 = rotl32(x6 ^ x11, 12) 617 # x1 += x6, x12 = rotl32(x12 ^ x1, 8) 634 # x11 += x12, x6 = rotl32(x6 ^ x11, 7)
|
| /Linux-v5.4/arch/arm64/crypto/ |
| D | aes-neonbs-core.S | 113 .macro mul_gf16_2, x0, x1, x2, x3, x4, x5, x6, x7, \ 125 eor \t0, \x4, \x6 127 mul_gf4_n_gf4 \t0, \t1, \y0, \y1, \t3, \x6, \x7, \y2, \y3, \t2 132 eor \x6, \x6, \t0 137 .macro inv_gf256, x0, x1, x2, x3, x4, x5, x6, x7, \ 139 eor \t3, \x4, \x6 142 eor \s1, \x7, \x6 166 and \s1, \x6, \x2 185 mul_gf16_2 \x0, \x1, \x2, \x3, \x4, \x5, \x6, \x7, \ 227 .macro add_round_key, x0, x1, x2, x3, x4, x5, x6, x7 [all …]
|
| D | aes-modes.S | 133 enc_prepare w8, x6, x7 134 encrypt_block v4, w8, x6, x7, w9 135 enc_switch_key w3, x2, x6 140 enc_prepare w3, x2, x6 147 encrypt_block v0, w3, x2, x6, w7 149 encrypt_block v1, w3, x2, x6, w7 151 encrypt_block v2, w3, x2, x6, w7 153 encrypt_block v3, w3, x2, x6, w7 163 encrypt_block v4, w3, x2, x6, w7 180 enc_prepare w8, x6, x7 [all …]
|
| D | aes-ce-ccm-core.S | 35 add x6, x4, #16 42 ld1 {v5.4s}, [x6], #16 /* load 2nd round key */ 45 4: ld1 {v3.4s}, [x6], #16 /* load next round key */ 48 5: ld1 {v4.4s}, [x6], #16 /* load next round key */ 52 ld1 {v5.4s}, [x6], #16 /* load next round key */ 127 ldr x8, [x6, #8] /* load lower ctr */ 131 ld1 {v1.8b}, [x6] /* load upper ctr */ 181 str x8, [x6, #8] /* store lsb end of ctr (BE) */
|
| D | sha512-core.S_shipped | 101 eor x6,x24,x24,ror#23 107 eor x16,x16,x6,ror#18 // Sigma1(e) 108 ror x6,x20,#28 115 eor x17,x6,x17,ror#34 // Sigma0(a) 122 ldp x5,x6,[x1],#2*8 169 rev x6,x6 // 3 178 add x24,x24,x6 // h+=X[i] 369 str x6,[sp,#24] 372 eor x6,x21,x21,ror#23 378 eor x16,x16,x6,ror#18 // Sigma1(e) [all …]
|
| /Linux-v5.4/arch/arm64/lib/ |
| D | crc32.S | 26 ldp x5, x6, [x8] 30 CPU_BE( rev x6, x6 ) 51 crc32\c\()x w8, w8, x6 57 ldp x5, x6, [x1, #-16] 61 CPU_BE( rev x6, x6 ) 65 crc32\c\()x w0, w0, x6
|
| D | copy_page.S | 30 ldp x6, x7, [x1, #32] 50 stnp x6, x7, [x0, #32] 51 ldp x6, x7, [x1, #32] 70 stnp x6, x7, [x0, #32]
|
| /Linux-v5.4/arch/arm/crypto/ |
| D | aes-neonbs-core.S | 173 .macro mul_gf16_2, x0, x1, x2, x3, x4, x5, x6, x7, \ 185 veor \t0, \x4, \x6 187 mul_gf4_n_gf4 \t0, \t1, \y0, \y1, \t3, \x6, \x7, \y2, \y3, \t2 192 veor \x6, \x6, \t0 197 .macro inv_gf256, x0, x1, x2, x3, x4, x5, x6, x7, \ 199 veor \t3, \x4, \x6 202 veor \s1, \x7, \x6 226 vand \s1, \x6, \x2 245 mul_gf16_2 \x0, \x1, \x2, \x3, \x4, \x5, \x6, \x7, \ 265 .macro shift_rows, x0, x1, x2, x3, x4, x5, x6, x7, \ [all …]
|
| /Linux-v5.4/arch/arm64/kernel/ |
| D | head.S | 314 mrs_s x6, SYS_ID_AA64MMFR2_EL1 315 and x6, x6, #(0xf << ID_AA64MMFR2_LVA_SHIFT) 317 cbnz x6, 1f 321 adr_l x6, vabits_actual 322 str x5, [x6] 324 dc ivac, x6 // Invalidate potentially stale cache line 343 adr_l x6, idmap_t0sz 344 str x5, [x6] 346 dc ivac, x6 // Invalidate potentially stale cache line 364 create_table_entry x0, x3, EXTRA_SHIFT, x4, x5, x6 [all …]
|
| D | hibernate-asm.S | 73 break_before_make_ttbr_switch x5, x0, x6, x8 87 copy_page x0, x1, x2, x3, x4, x5, x6, x7, x8, x9 104 break_before_make_ttbr_switch x25, x21, x6, x8
|
| D | entry-ftrace.S | 200 stp x6, x7, [sp, #48] 210 ldp x6, x7, [sp, #48]
|
| /Linux-v5.4/arch/arm64/kvm/ |
| D | hyp-init.S | 88 tcr_compute_pa_size x4, #TCR_EL2_PS_SHIFT, x5, x6 145 ldr x6, =SCTLR_ELx_FLAGS 146 bic x5, x5, x6 // Clear SCTL_M and etc
|
| /Linux-v5.4/arch/arm64/boot/dts/xilinx/ |
| D | zynqmp-zcu102-rev1.0.dts | 26 reg = <0x20 0x6>; 30 reg = <0xd0 0x6>;
|
| /Linux-v5.4/arch/arm64/xen/ |
| D | hypercall.S | 104 uaccess_ttbr0_enable x6, x7, x8 110 uaccess_ttbr0_disable x6, x7
|
| /Linux-v5.4/arch/arm64/mm/ |
| D | proc.S | 58 mrs x6, cpacr_el1 72 stp x6, x7, [x0, #32] 88 ldp x6, x8, [x0, #32] 95 msr cpacr_el1, x6 232 end_pgdp .req x6 463 tcr_compute_pa_size x10, #TCR_IPS_SHIFT, x5, x6
|
| /Linux-v5.4/Documentation/devicetree/bindings/gpio/ |
| D | brcm,brcmstb-gpio.txt | 66 interrupts = <0x6>; 78 interrupts = <0x6>; 79 interrupts-extended = <&irq0_aon_intc 0x6>,
|
| /Linux-v5.4/Documentation/input/devices/ |
| D | alps.rst | 110 byte 1: 0 x6 x5 x4 x3 x2 x1 x0 122 byte 1: 0 x6 x5 x4 x3 x2 x1 x0 140 byte 1: 0 x6 x5 x4 x3 x2 x1 x0 164 byte 1: 0 x10 x9 x8 x7 x6 x5 x4 180 byte 1: 0 x8 x7 x6 x5 x4 x3 x2 193 byte 1: 0 x6 x5 x4 x3 x2 x1 x0 208 byte 1: 0 x10 x9 x8 x7 x6 x5 x4 220 byte 0: 0 1 x7 x6 x5 x4 x3 x2 256 byte 1: 0 x6 x5 x4 x3 x2 x1 x0 268 byte 4: 0 x9 x8 x7 x6 x5 x4 x3 [all …]
|
| /Linux-v5.4/arch/arm64/boot/dts/freescale/ |
| D | imx8mq-zii-ultra.dtsi | 539 MX8MQ_IOMUXC_NAND_DATA04_GPIO3_IO10 0x6 540 MX8MQ_IOMUXC_NAND_DATA05_GPIO3_IO11 0x6 541 MX8MQ_IOMUXC_NAND_DATA06_GPIO3_IO12 0x6 542 MX8MQ_IOMUXC_NAND_DATA07_GPIO3_IO13 0x6 584 MX8MQ_IOMUXC_GPIO1_IO03_GPIO1_IO3 0x6 591 MX8MQ_IOMUXC_GPIO1_IO06_GPIO1_IO6 0x6
|
| /Linux-v5.4/arch/sparc/lib/ |
| D | NG2memcpy.S | 86 #define FREG_FROB(x0, x1, x2, x3, x4, x5, x6, x7, x8) \ argument 92 faligndata %x5, %x6, %f10; \ 93 faligndata %x6, %x7, %f12; \ 123 #define FREG_MOVE_7(x0, x1, x2, x3, x4, x5, x6) \ argument 130 fsrc2 %x6, %f12; 131 #define FREG_MOVE_8(x0, x1, x2, x3, x4, x5, x6, x7) \ argument 138 fsrc2 %x6, %f12; \ 167 #define FREG_LOAD_7(base, x0, x1, x2, x3, x4, x5, x6) \ argument 174 EX_LD_FP(LOAD(ldd, base + 0x30, %x6), NG2_retl_o2_plus_g1);
|
| /Linux-v5.4/arch/powerpc/boot/dts/ |
| D | sbc8548.dts | 27 0x6 0x0 0xec000000 0x04000000>; /*64MB Flash*/ 87 reg = <0x6 0x0 0x04000000>;
|