/Linux-v5.10/arch/mips/kernel/ |
D | genex.S | 36 mfc0 k1, CP0_CAUSE 37 andi k1, k1, 0x7c 39 dsll k1, k1, 1 41 PTR_L k0, exception_handlers(k1) 56 mfc0 k1, CP0_CAUSE 58 andi k1, k1, 0x7c 62 beq k1, k0, handle_vced 64 beq k1, k0, handle_vcei 66 dsll k1, k1, 1 69 PTR_L k0, exception_handlers(k1) [all …]
|
D | bmips_vec.S | 42 li k1, CKSEG1 43 or k0, k1 58 mfc0 k1, $22, 3 59 srl k1, 16 60 andi k1, 0x8000 61 or k1, CKSEG1 | BMIPS_RELO_VECTOR_CONTROL_0 62 or k0, k1 63 li k1, 0xa0080000 64 sw k1, 0(k0) 70 li k1, CKSEG1 [all …]
|
/Linux-v5.10/arch/mips/kvm/ |
D | entry.c | 48 #define K1 27 macro 224 /* k0/k1 not being used in host kernel context */ in kvm_mips_build_vcpu_run() 225 UASM_i_ADDIU(&p, K1, SP, -(int)sizeof(struct pt_regs)); in kvm_mips_build_vcpu_run() 229 UASM_i_SW(&p, i, offsetof(struct pt_regs, regs[i]), K1); in kvm_mips_build_vcpu_run() 234 UASM_i_SW(&p, V0, offsetof(struct pt_regs, cp0_status), K1); in kvm_mips_build_vcpu_run() 237 kvm_mips_build_save_scratch(&p, V1, K1); in kvm_mips_build_vcpu_run() 243 UASM_i_ADDIU(&p, K1, A0, offsetof(struct kvm_vcpu, arch)); in kvm_mips_build_vcpu_run() 249 UASM_i_SW(&p, SP, offsetof(struct kvm_vcpu_arch, host_stack), K1); in kvm_mips_build_vcpu_run() 252 UASM_i_SW(&p, GP, offsetof(struct kvm_vcpu_arch, host_gp), K1); in kvm_mips_build_vcpu_run() 263 UASM_i_LW(&p, K0, offsetof(struct kvm_vcpu_arch, guest_ebase), K1); in kvm_mips_build_vcpu_run() [all …]
|
/Linux-v5.10/include/linux/ |
D | btree-128.h | 22 static inline void *btree_lookup128(struct btree_head128 *head, u64 k1, u64 k2) in btree_lookup128() argument 24 u64 key[2] = {k1, k2}; in btree_lookup128() 29 u64 *k1, u64 *k2) in btree_get_prev128() argument 31 u64 key[2] = {*k1, *k2}; in btree_get_prev128() 36 *k1 = key[0]; in btree_get_prev128() 41 static inline int btree_insert128(struct btree_head128 *head, u64 k1, u64 k2, in btree_insert128() argument 44 u64 key[2] = {k1, k2}; in btree_insert128() 49 static inline int btree_update128(struct btree_head128 *head, u64 k1, u64 k2, in btree_update128() argument 52 u64 key[2] = {k1, k2}; in btree_update128() 57 static inline void *btree_remove128(struct btree_head128 *head, u64 k1, u64 k2) in btree_remove128() argument [all …]
|
/Linux-v5.10/arch/mips/cavium-octeon/executive/ |
D | cvmx-boot-vector.c | 18 dmtc0 $k1, $31, 3 # Save $k1 to KScratch2 21 mfc0 $k1, $15, 1 # Ebase 25 andi $k1, 0x3ff # mask out core ID 28 sll $k1, 5 38 daddu $k0, $k0, $k1 42 dmfc0 $k1, $31, 3 # Restore $k1 from KScratch2 66 4: 40bbf803 dmtc0 k1,c0_kscratch2 69 c: 401b7801 mfc0 k1,c0_ebase 72 14: 337b03ff andi k1,k1,0x3ff 75 1c: 001bd940 sll k1,k1,0x5 [all …]
|
/Linux-v5.10/arch/arm/crypto/ |
D | nh-neon-core.S | 33 K1 .req q5 49 .macro _nh_stride k0, k1, k2, k3 59 vadd.u32 T1, T3, \k1 78 vld1.32 {K0,K1}, [KEY]! 88 _nh_stride K0, K1, K2, K3 89 _nh_stride K1, K2, K3, K0 90 _nh_stride K2, K3, K0, K1 91 _nh_stride K3, K0, K1, K2 98 _nh_stride K0, K1, K2, K3 102 _nh_stride K1, K2, K3, K0 [all …]
|
D | sha1-ce-core.S | 17 k1 .req q1 68 vld1.32 {k0-k1}, [ip, :128]! 94 add_update c, 0, k1, 8, 9, 10, 11 96 add_update p, 1, k1, 9, 10, 11, 8 97 add_update p, 0, k1, 10, 11, 8, 9 98 add_update p, 1, k1, 11, 8, 9, 10 99 add_update p, 0, k1, 8, 9, 10, 11
|
/Linux-v5.10/arch/sh/boards/mach-hp6xx/ |
D | pm_wakeup.S | 12 * k1 scratch 17 #define k1 r1 macro 21 mov #-126, k1 23 mov.b k0, @k1 25 mov.l 5f, k1 27 mov.w k0, @k1 29 mov.l 4f, k1 30 jmp @k1
|
/Linux-v5.10/arch/mips/mm/ |
D | cex-oct.S | 31 PTR_LA k1, cache_err_dcache 33 PTR_ADDU k1, k0, k1 /* k1 = &cache_err_dcache[core_id] */ 36 sd k0, (k1) 40 mfc0 k1, CP0_STATUS 41 andi k1, k1, ST0_EXL 42 beqz k1, 1f
|
D | cex-sb1.S | 35 * (0x170-0x17f) are used to preserve k0, k1, and ra. 42 * save/restore k0 and k1 from low memory (Useg is direct 49 sd k1,0x178($0) 59 mfc0 k1,C0_ERRCTL 60 bgtz k1,attempt_recovery 61 sll k0,k1,1 98 li k1,1 << 26 /* ICACHE_EXTERNAL */ 99 and k1,k0 100 bnez k1,unrecoverable 118 ld k1,0x178($0) [all …]
|
D | tlbex.c | 282 #define K1 27 macro 367 r.r2 = K1; in build_get_work_registers() 380 UASM_i_LA(p, K1, (long)&handler_reg_save); in build_get_work_registers() 381 UASM_i_ADDU(p, K0, K0, K1); in build_get_work_registers() 389 r.r1 = K1; in build_get_work_registers() 424 uasm_i_lui(&p, K1, uasm_rel_hi(pgdc)); /* cp0 delay */ in build_r3000_tlb_refill_handler() 425 uasm_i_lw(&p, K1, uasm_rel_lo(pgdc), K1); in build_r3000_tlb_refill_handler() 428 uasm_i_addu(&p, K1, K1, K0); in build_r3000_tlb_refill_handler() 430 uasm_i_lw(&p, K1, 0, K1); /* cp0 delay */ in build_r3000_tlb_refill_handler() 432 uasm_i_addu(&p, K1, K1, K0); in build_r3000_tlb_refill_handler() [all …]
|
/Linux-v5.10/arch/mips/include/asm/ |
D | stackframe.h | 113 * stores the new value in k1, leaving sp unaffected. 121 lui k1, %hi(kernelsp) 123 lui k1, %highest(kernelsp) 124 daddiu k1, %higher(kernelsp) 125 dsll k1, 16 126 daddiu k1, %hi(kernelsp) 127 dsll k1, 16 130 LONG_ADDU k1, k0 136 LONG_L sp, %lo(kernelsp)(k1) 138 LONG_L k1, %lo(kernelsp)(k1) [all …]
|
/Linux-v5.10/arch/arm64/crypto/ |
D | nh-neon-core.S | 22 K1 .req v5 34 .macro _nh_stride k0, k1, k2, k3 44 add T1.4s, T3.4s, \k1\().4s 67 ld1 {K0.4s,K1.4s}, [KEY], #32 77 _nh_stride K0, K1, K2, K3 78 _nh_stride K1, K2, K3, K0 79 _nh_stride K2, K3, K0, K1 80 _nh_stride K3, K0, K1, K2 87 _nh_stride K0, K1, K2, K3 91 _nh_stride K1, K2, K3, K0 [all …]
|
D | sha1-ce-core.S | 15 k1 .req v1 77 loadrc k1.4s, 0x6ed9eba1, w6 105 add_update c, ev, k1, 8, 9, 10, 11 107 add_update p, od, k1, 9, 10, 11, 8 108 add_update p, ev, k1, 10, 11, 8, 9 109 add_update p, od, k1, 11, 8, 9, 10 110 add_update p, ev, k1, 8, 9, 10, 11
|
/Linux-v5.10/arch/x86/crypto/ |
D | nh-sse2-x86_64.S | 17 #define K1 %xmm5 macro 33 .macro _nh_stride k0, k1, k2, k3, offset 45 paddd \k1, T1 77 movdqu 0x10(KEY), K1 88 _nh_stride K0, K1, K2, K3, 0x00 89 _nh_stride K1, K2, K3, K0, 0x10 90 _nh_stride K2, K3, K0, K1, 0x20 91 _nh_stride K3, K0, K1, K2, 0x30 100 _nh_stride K0, K1, K2, K3, 0x00 104 _nh_stride K1, K2, K3, K0, 0x10 [all …]
|
D | nh-avx2-x86_64.S | 18 #define K1 %ymm5 macro 39 .macro _nh_2xstride k0, k1, k2, k3 43 vpaddd \k1, T3, T1 75 vmovdqu 0x10(KEY), K1 88 _nh_2xstride K0, K1, K2, K3 92 vmovdqu 0x30(KEY), K1 93 _nh_2xstride K2, K3, K0, K1 111 _nh_2xstride K0, K1, K2, K3 117 vmovdqa K3, K1 126 _nh_2xstride K0, K1, K2, K3
|
/Linux-v5.10/arch/sh/kernel/cpu/sh3/ |
D | entry.S | 35 * ldc k1, ssr ! delay slot 69 #define k1 r1 macro 83 * k1 scratch 258 mov #0xfffffff0, k1 259 extu.b k1, k1 260 not k1, k1 261 and k1, k2 ! Mask original SR value 304 ! k1 trashed 315 stc spc, k1 317 cmp/hs k0, k1 ! test k1 (saved PC) >= k0 (saved r0) [all …]
|
/Linux-v5.10/crypto/ |
D | rmd128.c | 26 #define K1 RMD_K1 macro 62 ROUND(aa, bb, cc, dd, F1, K1, in[0], 11); in rmd128_transform() 63 ROUND(dd, aa, bb, cc, F1, K1, in[1], 14); in rmd128_transform() 64 ROUND(cc, dd, aa, bb, F1, K1, in[2], 15); in rmd128_transform() 65 ROUND(bb, cc, dd, aa, F1, K1, in[3], 12); in rmd128_transform() 66 ROUND(aa, bb, cc, dd, F1, K1, in[4], 5); in rmd128_transform() 67 ROUND(dd, aa, bb, cc, F1, K1, in[5], 8); in rmd128_transform() 68 ROUND(cc, dd, aa, bb, F1, K1, in[6], 7); in rmd128_transform() 69 ROUND(bb, cc, dd, aa, F1, K1, in[7], 9); in rmd128_transform() 70 ROUND(aa, bb, cc, dd, F1, K1, in[8], 11); in rmd128_transform() [all …]
|
D | rmd256.c | 26 #define K1 RMD_K1 macro 62 ROUND(aa, bb, cc, dd, F1, K1, in[0], 11); in rmd256_transform() 63 ROUND(dd, aa, bb, cc, F1, K1, in[1], 14); in rmd256_transform() 64 ROUND(cc, dd, aa, bb, F1, K1, in[2], 15); in rmd256_transform() 65 ROUND(bb, cc, dd, aa, F1, K1, in[3], 12); in rmd256_transform() 66 ROUND(aa, bb, cc, dd, F1, K1, in[4], 5); in rmd256_transform() 67 ROUND(dd, aa, bb, cc, F1, K1, in[5], 8); in rmd256_transform() 68 ROUND(cc, dd, aa, bb, F1, K1, in[6], 7); in rmd256_transform() 69 ROUND(bb, cc, dd, aa, F1, K1, in[7], 9); in rmd256_transform() 70 ROUND(aa, bb, cc, dd, F1, K1, in[8], 11); in rmd256_transform() [all …]
|
D | rmd160.c | 26 #define K1 RMD_K1 macro 68 ROUND(aa, bb, cc, dd, ee, F1, K1, in[0], 11); in rmd160_transform() 69 ROUND(ee, aa, bb, cc, dd, F1, K1, in[1], 14); in rmd160_transform() 70 ROUND(dd, ee, aa, bb, cc, F1, K1, in[2], 15); in rmd160_transform() 71 ROUND(cc, dd, ee, aa, bb, F1, K1, in[3], 12); in rmd160_transform() 72 ROUND(bb, cc, dd, ee, aa, F1, K1, in[4], 5); in rmd160_transform() 73 ROUND(aa, bb, cc, dd, ee, F1, K1, in[5], 8); in rmd160_transform() 74 ROUND(ee, aa, bb, cc, dd, F1, K1, in[6], 7); in rmd160_transform() 75 ROUND(dd, ee, aa, bb, cc, F1, K1, in[7], 9); in rmd160_transform() 76 ROUND(cc, dd, ee, aa, bb, F1, K1, in[8], 11); in rmd160_transform() [all …]
|
D | rmd320.c | 26 #define K1 RMD_K1 macro 68 ROUND(aa, bb, cc, dd, ee, F1, K1, in[0], 11); in rmd320_transform() 69 ROUND(ee, aa, bb, cc, dd, F1, K1, in[1], 14); in rmd320_transform() 70 ROUND(dd, ee, aa, bb, cc, F1, K1, in[2], 15); in rmd320_transform() 71 ROUND(cc, dd, ee, aa, bb, F1, K1, in[3], 12); in rmd320_transform() 72 ROUND(bb, cc, dd, ee, aa, F1, K1, in[4], 5); in rmd320_transform() 73 ROUND(aa, bb, cc, dd, ee, F1, K1, in[5], 8); in rmd320_transform() 74 ROUND(ee, aa, bb, cc, dd, F1, K1, in[6], 7); in rmd320_transform() 75 ROUND(dd, ee, aa, bb, cc, F1, K1, in[7], 9); in rmd320_transform() 76 ROUND(cc, dd, ee, aa, bb, F1, K1, in[8], 11); in rmd320_transform() [all …]
|
/Linux-v5.10/fs/ext4/ |
D | hash.c | 46 #define K1 0 macro 58 ROUND(F, a, b, c, d, in[0] + K1, 3); in half_md4_transform() 59 ROUND(F, d, a, b, c, in[1] + K1, 7); in half_md4_transform() 60 ROUND(F, c, d, a, b, in[2] + K1, 11); in half_md4_transform() 61 ROUND(F, b, c, d, a, in[3] + K1, 19); in half_md4_transform() 62 ROUND(F, a, b, c, d, in[4] + K1, 3); in half_md4_transform() 63 ROUND(F, d, a, b, c, in[5] + K1, 7); in half_md4_transform() 64 ROUND(F, c, d, a, b, in[6] + K1, 11); in half_md4_transform() 65 ROUND(F, b, c, d, a, in[7] + K1, 19); in half_md4_transform() 95 #undef K1
|
/Linux-v5.10/lib/raid6/ |
D | avx512.c | 71 "vpcmpgtb %%zmm4,%%zmm1,%%k1\n\t" in raid6_avx5121_gen_syndrome() 72 "vpmovm2b %%k1,%%zmm5\n\t" in raid6_avx5121_gen_syndrome() 82 asm volatile("vpcmpgtb %%zmm4,%%zmm1,%%k1\n\t" in raid6_avx5121_gen_syndrome() 83 "vpmovm2b %%k1,%%zmm5\n\t" in raid6_avx5121_gen_syndrome() 126 "vpcmpgtb %%zmm4,%%zmm5,%%k1\n\t" in raid6_avx5121_xor_syndrome() 127 "vpmovm2b %%k1,%%zmm5\n\t" in raid6_avx5121_xor_syndrome() 140 "vpcmpgtb %%zmm4,%%zmm5,%%k1\n\t" in raid6_avx5121_xor_syndrome() 141 "vpmovm2b %%k1,%%zmm5\n\t" in raid6_avx5121_xor_syndrome() 201 "vpcmpgtb %%zmm4,%%zmm1,%%k1\n\t" in raid6_avx5122_gen_syndrome() 203 "vpmovm2b %%k1,%%zmm5\n\t" in raid6_avx5122_gen_syndrome() [all …]
|
/Linux-v5.10/fs/xfs/libxfs/ |
D | xfs_alloc_btree.c | 263 union xfs_btree_key *k1, in xfs_bnobt_diff_two_keys() argument 266 return (int64_t)be32_to_cpu(k1->alloc.ar_startblock) - in xfs_bnobt_diff_two_keys() 273 union xfs_btree_key *k1, in xfs_cntbt_diff_two_keys() argument 278 diff = be32_to_cpu(k1->alloc.ar_blockcount) - in xfs_cntbt_diff_two_keys() 283 return be32_to_cpu(k1->alloc.ar_startblock) - in xfs_cntbt_diff_two_keys() 383 union xfs_btree_key *k1, in xfs_bnobt_keys_inorder() argument 386 return be32_to_cpu(k1->alloc.ar_startblock) < in xfs_bnobt_keys_inorder() 404 union xfs_btree_key *k1, in xfs_cntbt_keys_inorder() argument 407 return be32_to_cpu(k1->alloc.ar_blockcount) < in xfs_cntbt_keys_inorder() 409 (k1->alloc.ar_blockcount == k2->alloc.ar_blockcount && in xfs_cntbt_keys_inorder() [all …]
|
/Linux-v5.10/arch/mips/netlogic/common/ |
D | reset.S | 145 * We use scratch reg 6/7 to save k0/k1 and check for NMI first. 161 dmtc0 k1, $22, 7 163 li k1, 0x80000 164 and k1, k0, k1 165 beqz k1, 1f /* go to real reset entry */ 167 li k1, CKSEG1ADDR(RESET_DATA_PHYS) /* NMI */ 168 ld k0, BOOT_NMI_HANDLER(k1)
|