Home
last modified time | relevance | path

Searched full:src2 (Results 1 – 25 of 86) sorted by relevance

1234

/Linux-v6.1/lib/crypto/
Dutils.c13 * XOR @len bytes from @src1 and @src2 together, writing the result to @dst
17 void __crypto_xor(u8 *dst, const u8 *src1, const u8 *src2, unsigned int len) in __crypto_xor() argument
24 ((unsigned long)dst ^ (unsigned long)src2)) & in __crypto_xor()
36 *dst++ = *src1++ ^ *src2++; in __crypto_xor()
44 get_unaligned((u64 *)src2); in __crypto_xor()
47 *(u64 *)dst = *(u64 *)src1 ^ *(u64 *)src2; in __crypto_xor()
51 src2 += 8; in __crypto_xor()
58 get_unaligned((u32 *)src2); in __crypto_xor()
61 *(u32 *)dst = *(u32 *)src1 ^ *(u32 *)src2; in __crypto_xor()
65 src2 += 4; in __crypto_xor()
[all …]
/Linux-v6.1/arch/powerpc/include/asm/
Dkvm_fpu.h18 extern void fps_fadds(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2);
19 extern void fps_fdivs(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2);
20 extern void fps_fmuls(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2);
21 extern void fps_fsubs(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2);
23 extern void fps_fmadds(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2,
25 extern void fps_fmsubs(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2,
27 extern void fps_fnmadds(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2,
29 extern void fps_fnmsubs(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2,
31 extern void fps_fsel(u64 *fpscr, u32 *dst, u32 *src1, u32 *src2,
37 u64 *dst, u64 *src1, u64 *src2);
[all …]
/Linux-v6.1/include/linux/
Dbitmap.h43 * bitmap_and(dst, src1, src2, nbits) *dst = *src1 & *src2
44 * bitmap_or(dst, src1, src2, nbits) *dst = *src1 | *src2
45 * bitmap_xor(dst, src1, src2, nbits) *dst = *src1 ^ *src2
46 * bitmap_andnot(dst, src1, src2, nbits) *dst = *src1 & ~(*src2)
48 * bitmap_equal(src1, src2, nbits) Are *src1 and *src2 equal?
49 * bitmap_intersects(src1, src2, nbits) Do *src1 and *src2 overlap?
50 * bitmap_subset(src1, src2, nbits) Is *src1 a subset of *src2?
54 * bitmap_weight_and(src1, src2, nbits) Hamming Weight of and'ed bitmap
141 const unsigned long *src2,
321 const unsigned long *src2, unsigned int nbits) in bitmap_and() argument
[all …]
Dlinkmode.h36 const unsigned long *src2) in linkmode_andnot() argument
38 return bitmap_andnot(dst, src1, src2, __ETHTOOL_LINK_MODE_MASK_NBITS); in linkmode_andnot()
75 const unsigned long *src2) in linkmode_equal() argument
77 return bitmap_equal(src1, src2, __ETHTOOL_LINK_MODE_MASK_NBITS); in linkmode_equal()
81 const unsigned long *src2) in linkmode_intersects() argument
83 return bitmap_intersects(src1, src2, __ETHTOOL_LINK_MODE_MASK_NBITS); in linkmode_intersects()
87 const unsigned long *src2) in linkmode_subset() argument
89 return bitmap_subset(src1, src2, __ETHTOOL_LINK_MODE_MASK_NBITS); in linkmode_subset()
Dnodemask.h29 * void nodes_and(dst, src1, src2) dst = src1 & src2 [intersection]
30 * void nodes_or(dst, src1, src2) dst = src1 | src2 [union]
31 * void nodes_xor(dst, src1, src2) dst = src1 ^ src2
32 * void nodes_andnot(dst, src1, src2) dst = src1 & ~src2
162 #define nodes_and(dst, src1, src2) \ argument
163 __nodes_and(&(dst), &(src1), &(src2), MAX_NUMNODES)
170 #define nodes_or(dst, src1, src2) \ argument
171 __nodes_or(&(dst), &(src1), &(src2), MAX_NUMNODES)
178 #define nodes_xor(dst, src1, src2) \ argument
179 __nodes_xor(&(dst), &(src1), &(src2), MAX_NUMNODES)
[all …]
/Linux-v6.1/arch/arm64/lib/
Dmemcmp.S21 #define src2 x1 macro
40 ldr data2, [src2], 8
48 ldr data2, [src2, limit]
53 ldr data2, [src2], 8
67 /* Align src1 and adjust src2 with bytes not yet done. */
71 sub src2, src2, tmp1
79 ldp data2, data2h, [src2], 16
95 add src2, src2, limit
97 ldp data2, data2h, [src2]
122 ldr data2w, [src2], 4
[all …]
Dstrncmp.S25 #define src2 x1 macro
63 eor tmp1, src1, src2
76 ldr data2, [src2], #8
165 bic src2, src2, #7
168 ldr data2, [src2], #8
187 ldrb data2w, [src2], #1
206 ldrb data2w, [src2], #1
218 src2 | x x x x x a a a a a a a a b b b | c c c c c . . .
227 Align SRC2 down to 16 bytes. This way we can read 16 bytes at a
228 time from SRC2. The comparison happens in 3 steps. After each step
[all …]
Dstrcmp.S24 #define src2 x1 macro
57 sub off2, src2, src1
114 neg shift, src2, lsl 3 /* Bits to alignment -64. */
123 checking to make sure that we don't access beyond the end of SRC2. */
127 ldrb data2w, [src2], 1
135 neg shift, src2, lsl 3
136 bic src2, src2, 7
137 ldr data3, [src2], 8
148 sub off1, src2, src1
/Linux-v6.1/tools/include/linux/
Dbitmap.h72 const unsigned long *src2, unsigned int nbits) in bitmap_or() argument
75 *dst = *src1 | *src2; in bitmap_or()
77 __bitmap_or(dst, src1, src2, nbits); in bitmap_or()
146 * @src2: operand 2
150 const unsigned long *src2, unsigned int nbits) in bitmap_and() argument
153 return (*dst = *src1 & *src2 & BITMAP_LAST_WORD_MASK(nbits)) != 0; in bitmap_and()
154 return __bitmap_and(dst, src1, src2, nbits); in bitmap_and()
166 const unsigned long *src2, unsigned int nbits) in bitmap_equal() argument
169 return !((*src1 ^ *src2) & BITMAP_LAST_WORD_MASK(nbits)); in bitmap_equal()
172 return !memcmp(src1, src2, nbits / 8); in bitmap_equal()
[all …]
/Linux-v6.1/arch/ia64/lib/
Dcopy_page.S30 #define src2 r21 macro
60 adds src2=8,in1
73 (p[0]) ld8 t2[0]=[src2],16
79 (p[0]) ld8 t4[0]=[src2],16
84 (p[0]) ld8 t6[0]=[src2],16
89 (p[0]) ld8 t8[0]=[src2],16
Dmemcpy.S31 # define src2 r17 macro
189 adds src2=7,src // src2 = (src + 7)
192 and src2=-8,src2 // src2 = (src + 7) & ~7
196 ld8 t1=[src2] // t1 = 1st source word if src is 8-byte aligned, 2nd otherwise
227 and src2=-8,src // align source pointer
254 (p6) ld8 val[1]=[src2],8 // prime the pump...
279 (p[0]) ld8 val[0]=[src2],8; \
Dcopy_user.S68 #define src2 r26 macro
130 and src2=0x7,src1 // src offset
160 sub t1=8,src2
161 mov t2=src2
171 cmp.leu p14,p15=src2,dst2
172 sub t1=dst2,src2
175 (p14) sub word1=8,src2 // (8 - src offset)
369 adds src2=8,src1
378 (p16) ld8 val2[0]=[src2],16
/Linux-v6.1/arch/arc/kernel/
Ddisasm.c117 state->src2 = FIELD_s9(state->words[0]); in disasm_instr()
140 state->src2 = state->words[1]; in disasm_instr()
142 state->src2 = get_reg(state->wb_reg, regs, cregs); in disasm_instr()
224 state->src2 = FIELD_C(state->words[0]); in disasm_instr()
225 if (state->src2 == REG_LIMM) { in disasm_instr()
227 state->src2 = state->words[1]; in disasm_instr()
229 state->src2 = get_reg(state->src2, regs, in disasm_instr()
286 state->src2 = get_reg(FIELD_S_C(state->words[0]), regs, cregs); in disasm_instr()
329 state->src2 = FIELD_S_u7(state->words[0]); in disasm_instr()
347 state->src2 = FIELD_S_u6(state->words[0]); in disasm_instr()
[all …]
Dunaligned.c139 set_reg(state->wb_reg, state->src1 + state->src2, regs, cregs); in fixup_load()
142 state->src2 = 0; in fixup_load()
146 get32_unaligned_check(val, state->src1 + state->src2); in fixup_load()
148 get16_unaligned_check(val, state->src1 + state->src2); in fixup_load()
167 set_reg(state->wb_reg, state->src2 + state->src3, regs, cregs); in fixup_store()
173 set_reg(state->wb_reg, state->src2 + (state->src3 << 1), in fixup_store()
176 set_reg(state->wb_reg, state->src2 + (state->src3 << 2), in fixup_store()
185 put32_unaligned_check(state->src1, state->src2 + state->src3); in fixup_store()
187 put16_unaligned_check(state->src1, state->src2 + state->src3); in fixup_store()
/Linux-v6.1/arch/m68k/math-emu/
Dmulti_arith.h115 struct fp_ext *src2) in fp_submant() argument
119 : "g,d" (src2->lowmant), "0,0" (src1->lowmant)); in fp_submant()
121 : "d" (src2->mant.m32[1]), "0" (src1->mant.m32[1])); in fp_submant()
123 : "d" (src2->mant.m32[0]), "0" (src1->mant.m32[0])); in fp_submant()
126 #define fp_mul64(desth, destl, src1, src2) ({ \ argument
128 : "dm" (src1), "0" (src2)); \
133 #define fp_add64(dest1, dest2, src1, src2) ({ \ argument
135 : "dm,d" (src2), "0,0" (dest2)); \
166 struct fp_ext *src2) in fp_multiplymant() argument
170 fp_mul64(dest->m32[0], dest->m32[1], src1->mant.m32[0], src2->mant.m32[0]); in fp_multiplymant()
[all …]
Dfp_log.c31 struct fp_ext tmp, src2; in fp_fsqrt() local
64 fp_copy_ext(&src2, dest); in fp_fsqrt()
89 fp_copy_ext(&tmp, &src2); in fp_fsqrt()
/Linux-v6.1/arch/sparc/kernel/
Dvisemul.c55 /* 000101000 - four 16-bit compare; set rd if src1 > src2 */
58 /* 000101100 - two 32-bit compare; set rd if src1 > src2 */
61 /* 000100000 - four 16-bit compare; set rd if src1 <= src2 */
64 /* 000100100 - two 32-bit compare; set rd if src1 <= src2 */
67 /* 000100010 - four 16-bit compare; set rd if src1 != src2 */
70 /* 000100110 - two 32-bit compare; set rd if src1 != src2 */
73 /* 000101010 - four 16-bit compare; set rd if src1 == src2 */
76 /* 000101110 - two 32-bit compare; set rd if src1 == src2 */
604 s16 src2 = (rs2 >> (byte * 16)) & 0xffff; in pmul() local
605 u32 prod = src1 * src2; in pmul()
[all …]
/Linux-v6.1/arch/x86/include/asm/
Dmpspec.h95 #define physids_and(dst, src1, src2) \ argument
96 bitmap_and((dst).mask, (src1).mask, (src2).mask, MAX_LOCAL_APIC)
98 #define physids_or(dst, src1, src2) \ argument
99 bitmap_or((dst).mask, (src1).mask, (src2).mask, MAX_LOCAL_APIC)
/Linux-v6.1/arch/sparc/lib/
Dxor.S404 ENTRY(xor_niagara_3) /* %o0=bytes, %o1=dest, %o2=src1, %o3=src2 */
417 ldda [%l7 + 0x00] %asi, %g2 /* %g2/%g3 = src2 + 0x00 */
418 ldda [%l7 + 0x10] %asi, %l0 /* %l0/%l1 = src2 + 0x10 */
428 ldda [%l7 + 0x20] %asi, %g2 /* %g2/%g3 = src2 + 0x20 */
437 ldda [%l7 + 0x30] %asi, %l0 /* %l0/%l1 = src2 + 0x30 */
466 ENTRY(xor_niagara_4) /* %o0=bytes, %o1=dest, %o2=src1, %o3=src2, %o4=src3 */
480 ldda [%l7 + 0x00] %asi, %i4 /* %i4/%i5 = src2 + 0x00 */
488 ldda [%l7 + 0x10] %asi, %i4 /* %i4/%i5 = src2 + 0x10 */
501 ldda [%l7 + 0x20] %asi, %i4 /* %i4/%i5 = src2 + 0x20 */
514 ldda [%l7 + 0x30] %asi, %i4 /* %i4/%i5 = src2 + 0x30 */
[all …]
/Linux-v6.1/drivers/media/platform/st/sti/bdisp/
Dbdisp-reg.h174 #define BLT_INS_S2_OFF 0x00000000 /* src2 disabled */
175 #define BLT_INS_S2_MEM 0x00000008 /* src2 fetched from memory */
176 #define BLT_INS_S2_CF 0x00000018 /* src2 color fill */
196 #define BLT_ACK_BYPASS_S2S3 0x00000007 /* Bypass src2 and src3 */
/Linux-v6.1/include/crypto/
Dalgapi.h150 void __crypto_xor(u8 *dst, const u8 *src1, const u8 *src2, unsigned int size);
171 static inline void crypto_xor_cpy(u8 *dst, const u8 *src1, const u8 *src2, in crypto_xor_cpy() argument
179 unsigned long *s2 = (unsigned long *)src2; in crypto_xor_cpy()
188 __crypto_xor(dst, src1, src2, size); in crypto_xor_cpy()
/Linux-v6.1/arch/powerpc/mm/book3s64/
Dslice.c401 const struct slice_mask *src2) in slice_or_mask() argument
403 dst->low_slices = src1->low_slices | src2->low_slices; in slice_or_mask()
406 bitmap_or(dst->high_slices, src1->high_slices, src2->high_slices, SLICE_NUM_HIGH); in slice_or_mask()
411 const struct slice_mask *src2) in slice_andnot_mask() argument
413 dst->low_slices = src1->low_slices & ~src2->low_slices; in slice_andnot_mask()
416 bitmap_andnot(dst->high_slices, src1->high_slices, src2->high_slices, SLICE_NUM_HIGH); in slice_andnot_mask()
/Linux-v6.1/Documentation/devicetree/bindings/sound/
Drenesas,rsnd.txt39 Playback: [MEM] -> [SRC2] -> [DVC0] -> [SSIU0/SSI0] -> [codec]
49 &src2 &ctu02 &ssi2
150 [MEM] -> [SRC2] -> [CTU03] -+
/Linux-v6.1/arch/parisc/math-emu/
Dsgl_float.h28 #define Sgl_and_signs( src1dst, src2) \ argument
29 Sall(src1dst) = (Sall(src2)|~((unsigned int)1<<31)) & Sall(src1dst)
30 #define Sgl_or_signs( src1dst, src2) \ argument
31 Sall(src1dst) = (Sall(src2)&((unsigned int)1<<31)) | Sall(src1dst)
/Linux-v6.1/arch/x86/crypto/
Dcurve25519-x86_64.c163 /* Compute the raw multiplication: tmp <- src1 * src2 */ in fmul()
165 /* Compute src1[0] * src2 */ in fmul()
180 /* Compute src1[1] * src2 */ in fmul()
201 /* Compute src1[2] * src2 */ in fmul()
222 /* Compute src1[3] * src2 */ in fmul()
300 /* Compute src1[0] * src2 */ in fmul2()
315 /* Compute src1[1] * src2 */ in fmul2()
336 /* Compute src1[2] * src2 */ in fmul2()
357 /* Compute src1[3] * src2 */ in fmul2()
383 /* Compute src1[0] * src2 */ in fmul2()
[all …]

1234