/Linux-v6.1/lib/crypto/ |
D | utils.c | 13 * XOR @len bytes from @src1 and @src2 together, writing the result to @dst 17 void __crypto_xor(u8 *dst, const u8 *src1, const u8 *src2, unsigned int len) in __crypto_xor() argument 24 ((unsigned long)dst ^ (unsigned long)src2)) & in __crypto_xor() 36 *dst++ = *src1++ ^ *src2++; in __crypto_xor() 44 get_unaligned((u64 *)src2); in __crypto_xor() 47 *(u64 *)dst = *(u64 *)src1 ^ *(u64 *)src2; in __crypto_xor() 51 src2 += 8; in __crypto_xor() 58 get_unaligned((u32 *)src2); in __crypto_xor() 61 *(u32 *)dst = *(u32 *)src1 ^ *(u32 *)src2; in __crypto_xor() 65 src2 += 4; in __crypto_xor() [all …]
|
/Linux-v6.1/arch/powerpc/include/asm/ |
D | kvm_fpu.h | 18 extern void fps_fadds(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2); 19 extern void fps_fdivs(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2); 20 extern void fps_fmuls(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2); 21 extern void fps_fsubs(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2); 23 extern void fps_fmadds(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2, 25 extern void fps_fmsubs(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2, 27 extern void fps_fnmadds(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2, 29 extern void fps_fnmsubs(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2, 31 extern void fps_fsel(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2, 37 u64 *dst, u64 *src1, u64 *src2); [all …]
|
/Linux-v6.1/include/linux/ |
D | bitmap.h | 43 * bitmap_and(dst, src1, src2, nbits) *dst = *src1 & *src2 44 * bitmap_or(dst, src1, src2, nbits) *dst = *src1 | *src2 45 * bitmap_xor(dst, src1, src2, nbits) *dst = *src1 ^ *src2 46 * bitmap_andnot(dst, src1, src2, nbits) *dst = *src1 & ~(*src2) 48 * bitmap_equal(src1, src2, nbits) Are *src1 and *src2 equal? 49 * bitmap_intersects(src1, src2, nbits) Do *src1 and *src2 overlap? 50 * bitmap_subset(src1, src2, nbits) Is *src1 a subset of *src2? 54 * bitmap_weight_and(src1, src2, nbits) Hamming Weight of and'ed bitmap 141 const unsigned long *src2, 321 const unsigned long *src2, unsigned int nbits) in bitmap_and() argument [all …]
|
D | linkmode.h | 36 const unsigned long *src2) in linkmode_andnot() argument 38 return bitmap_andnot(dst, src1, src2, __ETHTOOL_LINK_MODE_MASK_NBITS); in linkmode_andnot() 75 const unsigned long *src2) in linkmode_equal() argument 77 return bitmap_equal(src1, src2, __ETHTOOL_LINK_MODE_MASK_NBITS); in linkmode_equal() 81 const unsigned long *src2) in linkmode_intersects() argument 83 return bitmap_intersects(src1, src2, __ETHTOOL_LINK_MODE_MASK_NBITS); in linkmode_intersects() 87 const unsigned long *src2) in linkmode_subset() argument 89 return bitmap_subset(src1, src2, __ETHTOOL_LINK_MODE_MASK_NBITS); in linkmode_subset()
|
D | nodemask.h | 29 * void nodes_and(dst, src1, src2) dst = src1 & src2 [intersection] 30 * void nodes_or(dst, src1, src2) dst = src1 | src2 [union] 31 * void nodes_xor(dst, src1, src2) dst = src1 ^ src2 32 * void nodes_andnot(dst, src1, src2) dst = src1 & ~src2 162 #define nodes_and(dst, src1, src2) \ argument 163 __nodes_and(&(dst), &(src1), &(src2), MAX_NUMNODES) 170 #define nodes_or(dst, src1, src2) \ argument 171 __nodes_or(&(dst), &(src1), &(src2), MAX_NUMNODES) 178 #define nodes_xor(dst, src1, src2) \ argument 179 __nodes_xor(&(dst), &(src1), &(src2), MAX_NUMNODES) [all …]
|
/Linux-v6.1/arch/arm64/lib/ |
D | memcmp.S | 21 #define src2 x1 macro 40 ldr data2, [src2], 8 48 ldr data2, [src2, limit] 53 ldr data2, [src2], 8 67 /* Align src1 and adjust src2 with bytes not yet done. */ 71 sub src2, src2, tmp1 79 ldp data2, data2h, [src2], 16 95 add src2, src2, limit 97 ldp data2, data2h, [src2] 122 ldr data2w, [src2], 4 [all …]
|
D | strncmp.S | 25 #define src2 x1 macro 63 eor tmp1, src1, src2 76 ldr data2, [src2], #8 165 bic src2, src2, #7 168 ldr data2, [src2], #8 187 ldrb data2w, [src2], #1 206 ldrb data2w, [src2], #1 218 src2 | x x x x x a a a a a a a a b b b | c c c c c . . . 227 Align SRC2 down to 16 bytes. This way we can read 16 bytes at a 228 time from SRC2. The comparison happens in 3 steps. After each step [all …]
|
D | strcmp.S | 24 #define src2 x1 macro 57 sub off2, src2, src1 114 neg shift, src2, lsl 3 /* Bits to alignment -64. */ 123 checking to make sure that we don't access beyond the end of SRC2. */ 127 ldrb data2w, [src2], 1 135 neg shift, src2, lsl 3 136 bic src2, src2, 7 137 ldr data3, [src2], 8 148 sub off1, src2, src1
|
/Linux-v6.1/tools/include/linux/ |
D | bitmap.h | 72 const unsigned long *src2, unsigned int nbits) in bitmap_or() argument 75 *dst = *src1 | *src2; in bitmap_or() 77 __bitmap_or(dst, src1, src2, nbits); in bitmap_or() 146 * @src2: operand 2 150 const unsigned long *src2, unsigned int nbits) in bitmap_and() argument 153 return (*dst = *src1 & *src2 & BITMAP_LAST_WORD_MASK(nbits)) != 0; in bitmap_and() 154 return __bitmap_and(dst, src1, src2, nbits); in bitmap_and() 166 const unsigned long *src2, unsigned int nbits) in bitmap_equal() argument 169 return !((*src1 ^ *src2) & BITMAP_LAST_WORD_MASK(nbits)); in bitmap_equal() 172 return !memcmp(src1, src2, nbits / 8); in bitmap_equal() [all …]
|
/Linux-v6.1/arch/ia64/lib/ |
D | copy_page.S | 30 #define src2 r21 macro 60 adds src2=8,in1 73 (p[0]) ld8 t2[0]=[src2],16 79 (p[0]) ld8 t4[0]=[src2],16 84 (p[0]) ld8 t6[0]=[src2],16 89 (p[0]) ld8 t8[0]=[src2],16
|
D | memcpy.S | 31 # define src2 r17 macro 189 adds src2=7,src // src2 = (src + 7) 192 and src2=-8,src2 // src2 = (src + 7) & ~7 196 ld8 t1=[src2] // t1 = 1st source word if src is 8-byte aligned, 2nd otherwise 227 and src2=-8,src // align source pointer 254 (p6) ld8 val[1]=[src2],8 // prime the pump... 279 (p[0]) ld8 val[0]=[src2],8; \
|
D | copy_user.S | 68 #define src2 r26 macro 130 and src2=0x7,src1 // src offset 160 sub t1=8,src2 161 mov t2=src2 171 cmp.leu p14,p15=src2,dst2 172 sub t1=dst2,src2 175 (p14) sub word1=8,src2 // (8 - src offset) 369 adds src2=8,src1 378 (p16) ld8 val2[0]=[src2],16
|
/Linux-v6.1/arch/arc/kernel/ |
D | disasm.c | 117 state->src2 = FIELD_s9(state->words[0]); in disasm_instr() 140 state->src2 = state->words[1]; in disasm_instr() 142 state->src2 = get_reg(state->wb_reg, regs, cregs); in disasm_instr() 224 state->src2 = FIELD_C(state->words[0]); in disasm_instr() 225 if (state->src2 == REG_LIMM) { in disasm_instr() 227 state->src2 = state->words[1]; in disasm_instr() 229 state->src2 = get_reg(state->src2, regs, in disasm_instr() 286 state->src2 = get_reg(FIELD_S_C(state->words[0]), regs, cregs); in disasm_instr() 329 state->src2 = FIELD_S_u7(state->words[0]); in disasm_instr() 347 state->src2 = FIELD_S_u6(state->words[0]); in disasm_instr() [all …]
|
D | unaligned.c | 139 set_reg(state->wb_reg, state->src1 + state->src2, regs, cregs); in fixup_load() 142 state->src2 = 0; in fixup_load() 146 get32_unaligned_check(val, state->src1 + state->src2); in fixup_load() 148 get16_unaligned_check(val, state->src1 + state->src2); in fixup_load() 167 set_reg(state->wb_reg, state->src2 + state->src3, regs, cregs); in fixup_store() 173 set_reg(state->wb_reg, state->src2 + (state->src3 << 1), in fixup_store() 176 set_reg(state->wb_reg, state->src2 + (state->src3 << 2), in fixup_store() 185 put32_unaligned_check(state->src1, state->src2 + state->src3); in fixup_store() 187 put16_unaligned_check(state->src1, state->src2 + state->src3); in fixup_store()
|
/Linux-v6.1/arch/m68k/math-emu/ |
D | multi_arith.h | 115 struct fp_ext *src2) in fp_submant() argument 119 : "g,d" (src2->lowmant), "0,0" (src1->lowmant)); in fp_submant() 121 : "d" (src2->mant.m32[1]), "0" (src1->mant.m32[1])); in fp_submant() 123 : "d" (src2->mant.m32[0]), "0" (src1->mant.m32[0])); in fp_submant() 126 #define fp_mul64(desth, destl, src1, src2) ({ \ argument 128 : "dm" (src1), "0" (src2)); \ 133 #define fp_add64(dest1, dest2, src1, src2) ({ \ argument 135 : "dm,d" (src2), "0,0" (dest2)); \ 166 struct fp_ext *src2) in fp_multiplymant() argument 170 fp_mul64(dest->m32[0], dest->m32[1], src1->mant.m32[0], src2->mant.m32[0]); in fp_multiplymant() [all …]
|
D | fp_log.c | 31 struct fp_ext tmp, src2; in fp_fsqrt() local 64 fp_copy_ext(&src2, dest); in fp_fsqrt() 89 fp_copy_ext(&tmp, &src2); in fp_fsqrt()
|
/Linux-v6.1/arch/sparc/kernel/ |
D | visemul.c | 55 /* 000101000 - four 16-bit compare; set rd if src1 > src2 */ 58 /* 000101100 - two 32-bit compare; set rd if src1 > src2 */ 61 /* 000100000 - four 16-bit compare; set rd if src1 <= src2 */ 64 /* 000100100 - two 32-bit compare; set rd if src1 <= src2 */ 67 /* 000100010 - four 16-bit compare; set rd if src1 != src2 */ 70 /* 000100110 - two 32-bit compare; set rd if src1 != src2 */ 73 /* 000101010 - four 16-bit compare; set rd if src1 == src2 */ 76 /* 000101110 - two 32-bit compare; set rd if src1 == src2 */ 604 s16 src2 = (rs2 >> (byte * 16)) & 0xffff; in pmul() local 605 u32 prod = src1 * src2; in pmul() [all …]
|
/Linux-v6.1/arch/x86/include/asm/ |
D | mpspec.h | 95 #define physids_and(dst, src1, src2) \ argument 96 bitmap_and((dst).mask, (src1).mask, (src2).mask, MAX_LOCAL_APIC) 98 #define physids_or(dst, src1, src2) \ argument 99 bitmap_or((dst).mask, (src1).mask, (src2).mask, MAX_LOCAL_APIC)
|
/Linux-v6.1/arch/sparc/lib/ |
D | xor.S | 404 ENTRY(xor_niagara_3) /* %o0=bytes, %o1=dest, %o2=src1, %o3=src2 */ 417 ldda [%l7 + 0x00] %asi, %g2 /* %g2/%g3 = src2 + 0x00 */ 418 ldda [%l7 + 0x10] %asi, %l0 /* %l0/%l1 = src2 + 0x10 */ 428 ldda [%l7 + 0x20] %asi, %g2 /* %g2/%g3 = src2 + 0x20 */ 437 ldda [%l7 + 0x30] %asi, %l0 /* %l0/%l1 = src2 + 0x30 */ 466 ENTRY(xor_niagara_4) /* %o0=bytes, %o1=dest, %o2=src1, %o3=src2, %o4=src3 */ 480 ldda [%l7 + 0x00] %asi, %i4 /* %i4/%i5 = src2 + 0x00 */ 488 ldda [%l7 + 0x10] %asi, %i4 /* %i4/%i5 = src2 + 0x10 */ 501 ldda [%l7 + 0x20] %asi, %i4 /* %i4/%i5 = src2 + 0x20 */ 514 ldda [%l7 + 0x30] %asi, %i4 /* %i4/%i5 = src2 + 0x30 */ [all …]
|
/Linux-v6.1/drivers/media/platform/st/sti/bdisp/ |
D | bdisp-reg.h | 174 #define BLT_INS_S2_OFF 0x00000000 /* src2 disabled */ 175 #define BLT_INS_S2_MEM 0x00000008 /* src2 fetched from memory */ 176 #define BLT_INS_S2_CF 0x00000018 /* src2 color fill */ 196 #define BLT_ACK_BYPASS_S2S3 0x00000007 /* Bypass src2 and src3 */
|
/Linux-v6.1/include/crypto/ |
D | algapi.h | 150 void __crypto_xor(u8 *dst, const u8 *src1, const u8 *src2, unsigned int size); 171 static inline void crypto_xor_cpy(u8 *dst, const u8 *src1, const u8 *src2, in crypto_xor_cpy() argument 179 unsigned long *s2 = (unsigned long *)src2; in crypto_xor_cpy() 188 __crypto_xor(dst, src1, src2, size); in crypto_xor_cpy()
|
/Linux-v6.1/arch/powerpc/mm/book3s64/ |
D | slice.c | 401 const struct slice_mask *src2) in slice_or_mask() argument 403 dst->low_slices = src1->low_slices | src2->low_slices; in slice_or_mask() 406 bitmap_or(dst->high_slices, src1->high_slices, src2->high_slices, SLICE_NUM_HIGH); in slice_or_mask() 411 const struct slice_mask *src2) in slice_andnot_mask() argument 413 dst->low_slices = src1->low_slices & ~src2->low_slices; in slice_andnot_mask() 416 bitmap_andnot(dst->high_slices, src1->high_slices, src2->high_slices, SLICE_NUM_HIGH); in slice_andnot_mask()
|
/Linux-v6.1/Documentation/devicetree/bindings/sound/ |
D | renesas,rsnd.txt | 39 Playback: [MEM] -> [SRC2] -> [DVC0] -> [SSIU0/SSI0] -> [codec] 49 &src2 &ctu02 &ssi2 150 [MEM] -> [SRC2] -> [CTU03] -+
|
/Linux-v6.1/arch/parisc/math-emu/ |
D | sgl_float.h | 28 #define Sgl_and_signs( src1dst, src2) \ argument 29 Sall(src1dst) = (Sall(src2)|~((unsigned int)1<<31)) & Sall(src1dst) 30 #define Sgl_or_signs( src1dst, src2) \ argument 31 Sall(src1dst) = (Sall(src2)&((unsigned int)1<<31)) | Sall(src1dst)
|
/Linux-v6.1/arch/x86/crypto/ |
D | curve25519-x86_64.c | 163 /* Compute the raw multiplication: tmp <- src1 * src2 */ in fmul() 165 /* Compute src1[0] * src2 */ in fmul() 180 /* Compute src1[1] * src2 */ in fmul() 201 /* Compute src1[2] * src2 */ in fmul() 222 /* Compute src1[3] * src2 */ in fmul() 300 /* Compute src1[0] * src2 */ in fmul2() 315 /* Compute src1[1] * src2 */ in fmul2() 336 /* Compute src1[2] * src2 */ in fmul2() 357 /* Compute src1[3] * src2 */ in fmul2() 383 /* Compute src1[0] * src2 */ in fmul2() [all …]
|