/Linux-v5.15/arch/mips/kernel/ |
D | genex.S | 36 mfc0 k1, CP0_CAUSE 37 andi k1, k1, 0x7c 39 dsll k1, k1, 1 41 PTR_L k0, exception_handlers(k1) 56 mfc0 k1, CP0_CAUSE 58 andi k1, k1, 0x7c 62 beq k1, k0, handle_vced 64 beq k1, k0, handle_vcei 66 dsll k1, k1, 1 69 PTR_L k0, exception_handlers(k1) [all …]
|
D | bmips_vec.S | 42 li k1, CKSEG1 43 or k0, k1 58 mfc0 k1, $22, 3 59 srl k1, 16 60 andi k1, 0x8000 61 or k1, CKSEG1 | BMIPS_RELO_VECTOR_CONTROL_0 62 or k0, k1 63 li k1, 0xa0080000 64 sw k1, 0(k0) 70 li k1, CKSEG1 [all …]
|
D | octeon_switch.S | 421 * safely modify v1,k0, k1,$10-$15, and $24. It will 438 v3mulu k1, $0, $0 441 sd k1, PT_MTP+8(sp) /* PT_MTP+8 has P1 */ 442 ori k1, $0, 1 443 v3mulu k1, k1, $0 446 sd k1, PT_MPL(sp) /* PT_MPL has MPL0 */ 447 v3mulu k1, $0, $0 450 sd k1, PT_MPL+16(sp) /* PT_MPL+16 has MPL2 */ 510 ld k1, PT_MTP+16(sp) /* P2 */ 516 mtp2 k1 /* P2 */
|
/Linux-v5.15/arch/mips/kvm/ |
D | entry.c | 48 #define K1 27 macro 224 /* k0/k1 not being used in host kernel context */ in kvm_mips_build_vcpu_run() 225 UASM_i_ADDIU(&p, K1, SP, -(int)sizeof(struct pt_regs)); in kvm_mips_build_vcpu_run() 229 UASM_i_SW(&p, i, offsetof(struct pt_regs, regs[i]), K1); in kvm_mips_build_vcpu_run() 234 UASM_i_SW(&p, V0, offsetof(struct pt_regs, cp0_status), K1); in kvm_mips_build_vcpu_run() 237 kvm_mips_build_save_scratch(&p, V1, K1); in kvm_mips_build_vcpu_run() 243 UASM_i_ADDIU(&p, K1, A0, offsetof(struct kvm_vcpu, arch)); in kvm_mips_build_vcpu_run() 249 UASM_i_SW(&p, SP, offsetof(struct kvm_vcpu_arch, host_stack), K1); in kvm_mips_build_vcpu_run() 252 UASM_i_SW(&p, GP, offsetof(struct kvm_vcpu_arch, host_gp), K1); in kvm_mips_build_vcpu_run() 263 UASM_i_LW(&p, K0, offsetof(struct kvm_vcpu_arch, guest_ebase), K1); in kvm_mips_build_vcpu_run() [all …]
|
/Linux-v5.15/include/linux/ |
D | btree-128.h | 22 static inline void *btree_lookup128(struct btree_head128 *head, u64 k1, u64 k2) in btree_lookup128() argument 24 u64 key[2] = {k1, k2}; in btree_lookup128() 29 u64 *k1, u64 *k2) in btree_get_prev128() argument 31 u64 key[2] = {*k1, *k2}; in btree_get_prev128() 36 *k1 = key[0]; in btree_get_prev128() 41 static inline int btree_insert128(struct btree_head128 *head, u64 k1, u64 k2, in btree_insert128() argument 44 u64 key[2] = {k1, k2}; in btree_insert128() 49 static inline int btree_update128(struct btree_head128 *head, u64 k1, u64 k2, in btree_update128() argument 52 u64 key[2] = {k1, k2}; in btree_update128() 57 static inline void *btree_remove128(struct btree_head128 *head, u64 k1, u64 k2) in btree_remove128() argument [all …]
|
/Linux-v5.15/arch/mips/cavium-octeon/executive/ |
D | cvmx-boot-vector.c | 18 dmtc0 $k1, $31, 3 # Save $k1 to KScratch2 21 mfc0 $k1, $15, 1 # Ebase 25 andi $k1, 0x3ff # mask out core ID 28 sll $k1, 5 38 daddu $k0, $k0, $k1 42 dmfc0 $k1, $31, 3 # Restore $k1 from KScratch2 66 4: 40bbf803 dmtc0 k1,c0_kscratch2 69 c: 401b7801 mfc0 k1,c0_ebase 72 14: 337b03ff andi k1,k1,0x3ff 75 1c: 001bd940 sll k1,k1,0x5 [all …]
|
/Linux-v5.15/arch/arm/crypto/ |
D | nh-neon-core.S | 33 K1 .req q5 49 .macro _nh_stride k0, k1, k2, k3 59 vadd.u32 T1, T3, \k1 78 vld1.32 {K0,K1}, [KEY]! 88 _nh_stride K0, K1, K2, K3 89 _nh_stride K1, K2, K3, K0 90 _nh_stride K2, K3, K0, K1 91 _nh_stride K3, K0, K1, K2 98 _nh_stride K0, K1, K2, K3 102 _nh_stride K1, K2, K3, K0 [all …]
|
D | sha1-ce-core.S | 17 k1 .req q1 68 vld1.32 {k0-k1}, [ip, :128]! 94 add_update c, 0, k1, 8, 9, 10, 11 96 add_update p, 1, k1, 9, 10, 11, 8 97 add_update p, 0, k1, 10, 11, 8, 9 98 add_update p, 1, k1, 11, 8, 9, 10 99 add_update p, 0, k1, 8, 9, 10, 11
|
/Linux-v5.15/arch/sh/boards/mach-hp6xx/ |
D | pm_wakeup.S | 12 * k1 scratch 17 #define k1 r1 macro 21 mov #-126, k1 23 mov.b k0, @k1 25 mov.l 5f, k1 27 mov.w k0, @k1 29 mov.l 4f, k1 30 jmp @k1
|
/Linux-v5.15/arch/mips/mm/ |
D | cex-oct.S | 31 PTR_LA k1, cache_err_dcache 33 PTR_ADDU k1, k0, k1 /* k1 = &cache_err_dcache[core_id] */ 36 sd k0, (k1) 40 mfc0 k1, CP0_STATUS 41 andi k1, k1, ST0_EXL 42 beqz k1, 1f
|
D | cex-sb1.S | 35 * (0x170-0x17f) are used to preserve k0, k1, and ra. 42 * save/restore k0 and k1 from low memory (Useg is direct 49 sd k1,0x178($0) 59 mfc0 k1,C0_ERRCTL 60 bgtz k1,attempt_recovery 61 sll k0,k1,1 98 li k1,1 << 26 /* ICACHE_EXTERNAL */ 99 and k1,k0 100 bnez k1,unrecoverable 118 ld k1,0x178($0) [all …]
|
D | tlbex.c | 282 #define K1 27 macro 367 r.r2 = K1; in build_get_work_registers() 380 UASM_i_LA(p, K1, (long)&handler_reg_save); in build_get_work_registers() 381 UASM_i_ADDU(p, K0, K0, K1); in build_get_work_registers() 389 r.r1 = K1; in build_get_work_registers() 424 uasm_i_lui(&p, K1, uasm_rel_hi(pgdc)); /* cp0 delay */ in build_r3000_tlb_refill_handler() 425 uasm_i_lw(&p, K1, uasm_rel_lo(pgdc), K1); in build_r3000_tlb_refill_handler() 428 uasm_i_addu(&p, K1, K1, K0); in build_r3000_tlb_refill_handler() 430 uasm_i_lw(&p, K1, 0, K1); /* cp0 delay */ in build_r3000_tlb_refill_handler() 432 uasm_i_addu(&p, K1, K1, K0); in build_r3000_tlb_refill_handler() [all …]
|
/Linux-v5.15/arch/mips/include/asm/ |
D | stackframe.h | 113 * stores the new value in k1, leaving sp unaffected. 121 lui k1, %hi(kernelsp) 123 lui k1, %highest(kernelsp) 124 daddiu k1, %higher(kernelsp) 125 dsll k1, 16 126 daddiu k1, %hi(kernelsp) 127 dsll k1, 16 130 LONG_ADDU k1, k0 136 LONG_L sp, %lo(kernelsp)(k1) 138 LONG_L k1, %lo(kernelsp)(k1) [all …]
|
/Linux-v5.15/arch/arm64/crypto/ |
D | nh-neon-core.S | 22 K1 .req v5 34 .macro _nh_stride k0, k1, k2, k3 44 add T1.4s, T3.4s, \k1\().4s 67 ld1 {K0.4s,K1.4s}, [KEY], #32 77 _nh_stride K0, K1, K2, K3 78 _nh_stride K1, K2, K3, K0 79 _nh_stride K2, K3, K0, K1 80 _nh_stride K3, K0, K1, K2 87 _nh_stride K0, K1, K2, K3 91 _nh_stride K1, K2, K3, K0 [all …]
|
D | sha1-ce-core.S | 15 k1 .req v1 71 loadrc k1.4s, 0x6ed9eba1, w6 99 add_update c, ev, k1, 8, 9, 10, 11 101 add_update p, od, k1, 9, 10, 11, 8 102 add_update p, ev, k1, 10, 11, 8, 9 103 add_update p, od, k1, 11, 8, 9, 10 104 add_update p, ev, k1, 8, 9, 10, 11
|
/Linux-v5.15/arch/x86/crypto/ |
D | nh-sse2-x86_64.S | 17 #define K1 %xmm5 macro 33 .macro _nh_stride k0, k1, k2, k3, offset 45 paddd \k1, T1 77 movdqu 0x10(KEY), K1 88 _nh_stride K0, K1, K2, K3, 0x00 89 _nh_stride K1, K2, K3, K0, 0x10 90 _nh_stride K2, K3, K0, K1, 0x20 91 _nh_stride K3, K0, K1, K2, 0x30 100 _nh_stride K0, K1, K2, K3, 0x00 104 _nh_stride K1, K2, K3, K0, 0x10 [all …]
|
D | nh-avx2-x86_64.S | 18 #define K1 %ymm5 macro 39 .macro _nh_2xstride k0, k1, k2, k3 43 vpaddd \k1, T3, T1 75 vmovdqu 0x10(KEY), K1 88 _nh_2xstride K0, K1, K2, K3 92 vmovdqu 0x30(KEY), K1 93 _nh_2xstride K2, K3, K0, K1 111 _nh_2xstride K0, K1, K2, K3 117 vmovdqa K3, K1 126 _nh_2xstride K0, K1, K2, K3
|
/Linux-v5.15/arch/sh/kernel/cpu/sh3/ |
D | entry.S | 34 * ldc k1, ssr ! delay slot 68 #define k1 r1 macro 82 * k1 scratch 257 mov #0xfffffff0, k1 258 extu.b k1, k1 259 not k1, k1 260 and k1, k2 ! Mask original SR value 303 ! k1 trashed 314 stc spc, k1 316 cmp/hs k0, k1 ! test k1 (saved PC) >= k0 (saved r0) [all …]
|
/Linux-v5.15/fs/ext4/ |
D | hash.c | 46 #define K1 0 macro 58 ROUND(F, a, b, c, d, in[0] + K1, 3); in half_md4_transform() 59 ROUND(F, d, a, b, c, in[1] + K1, 7); in half_md4_transform() 60 ROUND(F, c, d, a, b, in[2] + K1, 11); in half_md4_transform() 61 ROUND(F, b, c, d, a, in[3] + K1, 19); in half_md4_transform() 62 ROUND(F, a, b, c, d, in[4] + K1, 3); in half_md4_transform() 63 ROUND(F, d, a, b, c, in[5] + K1, 7); in half_md4_transform() 64 ROUND(F, c, d, a, b, in[6] + K1, 11); in half_md4_transform() 65 ROUND(F, b, c, d, a, in[7] + K1, 19); in half_md4_transform() 95 #undef K1
|
/Linux-v5.15/lib/raid6/ |
D | avx512.c | 71 "vpcmpgtb %%zmm4,%%zmm1,%%k1\n\t" in raid6_avx5121_gen_syndrome() 72 "vpmovm2b %%k1,%%zmm5\n\t" in raid6_avx5121_gen_syndrome() 82 asm volatile("vpcmpgtb %%zmm4,%%zmm1,%%k1\n\t" in raid6_avx5121_gen_syndrome() 83 "vpmovm2b %%k1,%%zmm5\n\t" in raid6_avx5121_gen_syndrome() 126 "vpcmpgtb %%zmm4,%%zmm5,%%k1\n\t" in raid6_avx5121_xor_syndrome() 127 "vpmovm2b %%k1,%%zmm5\n\t" in raid6_avx5121_xor_syndrome() 140 "vpcmpgtb %%zmm4,%%zmm5,%%k1\n\t" in raid6_avx5121_xor_syndrome() 141 "vpmovm2b %%k1,%%zmm5\n\t" in raid6_avx5121_xor_syndrome() 201 "vpcmpgtb %%zmm4,%%zmm1,%%k1\n\t" in raid6_avx5122_gen_syndrome() 203 "vpmovm2b %%k1,%%zmm5\n\t" in raid6_avx5122_gen_syndrome() [all …]
|
/Linux-v5.15/crypto/ |
D | rmd160.c | 26 #define K1 RMD_K1 macro 68 ROUND(aa, bb, cc, dd, ee, F1, K1, in[0], 11); in rmd160_transform() 69 ROUND(ee, aa, bb, cc, dd, F1, K1, in[1], 14); in rmd160_transform() 70 ROUND(dd, ee, aa, bb, cc, F1, K1, in[2], 15); in rmd160_transform() 71 ROUND(cc, dd, ee, aa, bb, F1, K1, in[3], 12); in rmd160_transform() 72 ROUND(bb, cc, dd, ee, aa, F1, K1, in[4], 5); in rmd160_transform() 73 ROUND(aa, bb, cc, dd, ee, F1, K1, in[5], 8); in rmd160_transform() 74 ROUND(ee, aa, bb, cc, dd, F1, K1, in[6], 7); in rmd160_transform() 75 ROUND(dd, ee, aa, bb, cc, F1, K1, in[7], 9); in rmd160_transform() 76 ROUND(cc, dd, ee, aa, bb, F1, K1, in[8], 11); in rmd160_transform() [all …]
|
D | khazad.c | 763 u64 K2, K1; in khazad_setkey() local 767 K1 = ((u64)be32_to_cpu(key[2]) << 32) | be32_to_cpu(key[3]); in khazad_setkey() 771 ctx->E[r] = T0[(int)(K1 >> 56) ] ^ in khazad_setkey() 772 T1[(int)(K1 >> 48) & 0xff] ^ in khazad_setkey() 773 T2[(int)(K1 >> 40) & 0xff] ^ in khazad_setkey() 774 T3[(int)(K1 >> 32) & 0xff] ^ in khazad_setkey() 775 T4[(int)(K1 >> 24) & 0xff] ^ in khazad_setkey() 776 T5[(int)(K1 >> 16) & 0xff] ^ in khazad_setkey() 777 T6[(int)(K1 >> 8) & 0xff] ^ in khazad_setkey() 778 T7[(int)(K1 ) & 0xff] ^ in khazad_setkey() [all …]
|
/Linux-v5.15/arch/mips/netlogic/common/ |
D | reset.S | 145 * We use scratch reg 6/7 to save k0/k1 and check for NMI first. 161 dmtc0 k1, $22, 7 163 li k1, 0x80000 164 and k1, k0, k1 165 beqz k1, 1f /* go to real reset entry */ 167 li k1, CKSEG1ADDR(RESET_DATA_PHYS) /* NMI */ 168 ld k0, BOOT_NMI_HANDLER(k1)
|
/Linux-v5.15/fs/xfs/libxfs/ |
D | xfs_alloc_btree.c | 260 const union xfs_btree_key *k1, in xfs_bnobt_diff_two_keys() argument 263 return (int64_t)be32_to_cpu(k1->alloc.ar_startblock) - in xfs_bnobt_diff_two_keys() 270 const union xfs_btree_key *k1, in xfs_cntbt_diff_two_keys() argument 275 diff = be32_to_cpu(k1->alloc.ar_blockcount) - in xfs_cntbt_diff_two_keys() 280 return be32_to_cpu(k1->alloc.ar_startblock) - in xfs_cntbt_diff_two_keys() 380 const union xfs_btree_key *k1, in xfs_bnobt_keys_inorder() argument 383 return be32_to_cpu(k1->alloc.ar_startblock) < in xfs_bnobt_keys_inorder() 401 const union xfs_btree_key *k1, in xfs_cntbt_keys_inorder() argument 404 return be32_to_cpu(k1->alloc.ar_blockcount) < in xfs_cntbt_keys_inorder() 406 (k1->alloc.ar_blockcount == k2->alloc.ar_blockcount && in xfs_cntbt_keys_inorder() [all …]
|
/Linux-v5.15/arch/sh/kernel/cpu/shmobile/ |
D | sleep.S | 19 * k1 scratch 22 #define k1 r1 macro 249 sts pr, k1 251 and k0, k1 254 ldc k1, vbr 257 mov.l @(SH_SLEEP_SR, k1), k0
|