/cmsis-nn-latest/Source/NNSupportFunctions/ |
D | arm_nn_mat_mult_nt_t_s8.c | 278 res00 = SMLAD(val3, val2, res00); in arm_nn_mat_mult_nt_t_s8() 280 res00 = SMLAD(val0, val1, res00); in arm_nn_mat_mult_nt_t_s8() 282 res01 = SMLAD(val3, val5, res01); in arm_nn_mat_mult_nt_t_s8() 283 res01 = SMLAD(val0, val4, res01); in arm_nn_mat_mult_nt_t_s8() 289 res10 = SMLAD(val3, val2, res10); in arm_nn_mat_mult_nt_t_s8() 290 res11 = SMLAD(val3, val5, res11); in arm_nn_mat_mult_nt_t_s8() 291 res10 = SMLAD(val0, val1, res10); in arm_nn_mat_mult_nt_t_s8() 293 res11 = SMLAD(val0, val4, res11); in arm_nn_mat_mult_nt_t_s8() 303 res00 = SMLAD(val3, val2, res00); in arm_nn_mat_mult_nt_t_s8() 305 res00 = SMLAD(val0, val1, res00); in arm_nn_mat_mult_nt_t_s8() [all …]
|
D | arm_nn_mat_mult_nt_t_s4.c | 417 res00 = SMLAD(rhs_low0, lhs_low, res00); in arm_nn_mat_mult_nt_t_s4() 418 res00 = SMLAD(rhs_high0, lhs_high, res00); in arm_nn_mat_mult_nt_t_s4() 419 res01 = SMLAD(rhs_low1, lhs_low, res01); in arm_nn_mat_mult_nt_t_s4() 420 res01 = SMLAD(rhs_high1, lhs_high, res01); in arm_nn_mat_mult_nt_t_s4() 427 res10 = SMLAD(rhs_low0, lhs_low, res10); in arm_nn_mat_mult_nt_t_s4() 428 res11 = SMLAD(rhs_low1, lhs_low, res11); in arm_nn_mat_mult_nt_t_s4() 429 res10 = SMLAD(rhs_high0, lhs_high, res10); in arm_nn_mat_mult_nt_t_s4() 430 res11 = SMLAD(rhs_high1, lhs_high, res11); in arm_nn_mat_mult_nt_t_s4() 441 res00 = SMLAD(rhs_low0, lhs_low, res00); in arm_nn_mat_mult_nt_t_s4() 442 res00 = SMLAD(rhs_high0, lhs_high, res00); in arm_nn_mat_mult_nt_t_s4() [all …]
|
D | arm_nn_vec_mat_mult_t_svdf_s8.c | 194 acc_0 = SMLAD(ker_1, vec_1, acc_0); in arm_nn_vec_mat_mult_t_svdf_s8() 195 acc_0 = SMLAD(ker_0, vec_0, acc_0); in arm_nn_vec_mat_mult_t_svdf_s8() 199 acc_1 = SMLAD(ker_1, vec_1, acc_1); in arm_nn_vec_mat_mult_t_svdf_s8() 200 acc_1 = SMLAD(ker_0, vec_0, acc_1); in arm_nn_vec_mat_mult_t_svdf_s8() 209 acc_0 = SMLAD(ker_1, vec_1, acc_0); in arm_nn_vec_mat_mult_t_svdf_s8() 210 acc_0 = SMLAD(ker_0, vec_0, acc_0); in arm_nn_vec_mat_mult_t_svdf_s8() 214 acc_1 = SMLAD(ker_1, vec_1, acc_1); in arm_nn_vec_mat_mult_t_svdf_s8() 215 acc_1 = SMLAD(ker_0, vec_0, acc_1); in arm_nn_vec_mat_mult_t_svdf_s8() 224 acc_0 = SMLAD(ker_1, vec_1, acc_0); in arm_nn_vec_mat_mult_t_svdf_s8() 225 acc_0 = SMLAD(ker_0, vec_0, acc_0); in arm_nn_vec_mat_mult_t_svdf_s8() [all …]
|
D | arm_nn_mat_mult_nt_t_s8_s32.c | 223 dst_ptr[0] = SMLAD(lhs000, rhs_val00, dst_ptr[0]); in arm_nn_mat_mult_nt_t_s8_s32() 224 dst_ptr[0] = SMLAD(lhs001, rhs_val01, dst_ptr[0]); in arm_nn_mat_mult_nt_t_s8_s32() 225 dst_ptr[dst_idx_col_offset] = SMLAD(lhs100, rhs_val00, dst_ptr[dst_idx_col_offset]); in arm_nn_mat_mult_nt_t_s8_s32() 226 dst_ptr[dst_idx_col_offset] = SMLAD(lhs101, rhs_val01, dst_ptr[dst_idx_col_offset]); in arm_nn_mat_mult_nt_t_s8_s32() 230 dst_ptr[0] = SMLAD(lhs010, rhs_val00, dst_ptr[0]); in arm_nn_mat_mult_nt_t_s8_s32() 231 dst_ptr[0] = SMLAD(lhs011, rhs_val01, dst_ptr[0]); in arm_nn_mat_mult_nt_t_s8_s32() 232 dst_ptr[dst_idx_col_offset] = SMLAD(lhs110, rhs_val00, dst_ptr[dst_idx_col_offset]); in arm_nn_mat_mult_nt_t_s8_s32() 233 dst_ptr[dst_idx_col_offset] = SMLAD(lhs111, rhs_val01, dst_ptr[dst_idx_col_offset]); in arm_nn_mat_mult_nt_t_s8_s32() 258 dst_ptr[0] = SMLAD(lhs00, rhs_val00, dst_ptr[0]); in arm_nn_mat_mult_nt_t_s8_s32() 259 dst_ptr[0] = SMLAD(lhs01, rhs_val01, dst_ptr[0]); in arm_nn_mat_mult_nt_t_s8_s32() [all …]
|
D | arm_nn_vec_mat_mult_t_s16.c | 244 acc_0 = SMLAD(ker_0, vec_part_0, acc_0); in arm_nn_vec_mat_mult_t_s16() 245 acc_0 = SMLAD(ker_1, vec_part_1, acc_0); in arm_nn_vec_mat_mult_t_s16() 249 acc_1 = SMLAD(ker_0, vec_part_0, acc_1); in arm_nn_vec_mat_mult_t_s16() 250 acc_1 = SMLAD(ker_1, vec_part_1, acc_1); in arm_nn_vec_mat_mult_t_s16() 299 acc_0 = SMLAD(ker_0, vec, acc_0); in arm_nn_vec_mat_mult_t_s16() 302 acc_0 = SMLAD(ker_1, vec, acc_0); in arm_nn_vec_mat_mult_t_s16()
|
D | arm_nn_vec_mat_mul_result_acc_s16.c | 244 acc_0 = SMLAD(ker_0, vec_part_0, acc_0); in arm_nn_vec_mat_mul_result_acc_s16() 245 acc_0 = SMLAD(ker_1, vec_part_1, acc_0); in arm_nn_vec_mat_mul_result_acc_s16() 249 acc_1 = SMLAD(ker_0, vec_part_0, acc_1); in arm_nn_vec_mat_mul_result_acc_s16() 250 acc_1 = SMLAD(ker_1, vec_part_1, acc_1); in arm_nn_vec_mat_mul_result_acc_s16() 298 acc_0 = SMLAD(ker_0, vec, acc_0); in arm_nn_vec_mat_mul_result_acc_s16() 301 acc_0 = SMLAD(ker_1, vec, acc_0); in arm_nn_vec_mat_mul_result_acc_s16()
|
D | arm_nn_vec_mat_mul_result_acc_s8_s16.c | 178 acc_0 = SMLAD(ker_1, vec_1, acc_0); in arm_nn_vec_mat_mul_result_acc_s8_s16() 179 acc_0 = SMLAD(ker_0, vec_0, acc_0); in arm_nn_vec_mat_mul_result_acc_s8_s16() 185 acc_1 = SMLAD(ker_1, vec_1, acc_1); in arm_nn_vec_mat_mul_result_acc_s8_s16() 186 acc_1 = SMLAD(ker_0, vec_0, acc_1); in arm_nn_vec_mat_mul_result_acc_s8_s16() 234 acc_0 = SMLAD(ker_1, vec_1, acc_0); in arm_nn_vec_mat_mul_result_acc_s8_s16() 235 acc_0 = SMLAD(ker_0, vec_0, acc_0); in arm_nn_vec_mat_mul_result_acc_s8_s16()
|
D | arm_nn_vec_mat_mult_t_s8.c | 236 acc_0 = SMLAD(ker_1, vec_1, acc_0); in arm_nn_vec_mat_mult_t_s8() 237 acc_0 = SMLAD(ker_0, vec_0, acc_0); in arm_nn_vec_mat_mult_t_s8() 243 acc_1 = SMLAD(ker_1, vec_1, acc_1); in arm_nn_vec_mat_mult_t_s8() 244 acc_1 = SMLAD(ker_0, vec_0, acc_1); in arm_nn_vec_mat_mult_t_s8() 295 acc_0 = SMLAD(ker_1, vec_1, acc_0); in arm_nn_vec_mat_mult_t_s8() 296 acc_0 = SMLAD(ker_0, vec_0, acc_0); in arm_nn_vec_mat_mult_t_s8() 573 acc_0 = SMLAD(ker_1, vec_1, acc_0); in arm_nn_vec_mat_mult_t_s8() 574 acc_0 = SMLAD(ker_0, vec_0, acc_0); in arm_nn_vec_mat_mult_t_s8() 580 acc_1 = SMLAD(ker_1, vec_1, acc_1); in arm_nn_vec_mat_mult_t_s8() 581 acc_1 = SMLAD(ker_0, vec_0, acc_1); in arm_nn_vec_mat_mult_t_s8() [all …]
|
D | arm_nn_vec_mat_mult_t_s4.c | 329 res0 = SMLAD(lhs_low, rhs_low0, res0); in arm_nn_vec_mat_mult_t_s4() 330 res0 = SMLAD(lhs_high, rhs_high0, res0); in arm_nn_vec_mat_mult_t_s4() 331 res1 = SMLAD(lhs_low, rhs_low1, res1); in arm_nn_vec_mat_mult_t_s4() 332 res1 = SMLAD(lhs_high, rhs_high1, res1); in arm_nn_vec_mat_mult_t_s4() 423 res0 = SMLAD(lhs_low, rhs_low0, res0); in arm_nn_vec_mat_mult_t_s4() 424 res0 = SMLAD(lhs_high, rhs_high0, res0); in arm_nn_vec_mat_mult_t_s4() 425 res1 = SMLAD(lhs_low, rhs_low1, res1); in arm_nn_vec_mat_mult_t_s4() 426 res1 = SMLAD(lhs_high, rhs_high1, res1); in arm_nn_vec_mat_mult_t_s4() 663 res0 = SMLAD(lhs_low, rhs_high0, res0); in arm_nn_vec_mat_mult_t_s4() 664 res0 = SMLAD(lhs_high, rhs_low0, res0); in arm_nn_vec_mat_mult_t_s4()
|
/cmsis-nn-latest/Source/ConvolutionFunctions/ |
D | arm_nn_mat_mult_kernel_row_offset_s8_s16.c | 98 ch_0_out_0 = SMLAD(a01, b0, ch_0_out_0); in arm_nn_mat_mult_kernel_row_offset_s8_s16() 99 ch_0_out_1 = SMLAD(a01, b1, ch_0_out_1); in arm_nn_mat_mult_kernel_row_offset_s8_s16() 100 ch_1_out_0 = SMLAD(a11, b0, ch_1_out_0); in arm_nn_mat_mult_kernel_row_offset_s8_s16() 101 ch_1_out_1 = SMLAD(a11, b1, ch_1_out_1); in arm_nn_mat_mult_kernel_row_offset_s8_s16() 106 ch_0_out_0 = SMLAD(a02, b0, ch_0_out_0); in arm_nn_mat_mult_kernel_row_offset_s8_s16() 107 ch_0_out_1 = SMLAD(a02, b1, ch_0_out_1); in arm_nn_mat_mult_kernel_row_offset_s8_s16() 108 ch_1_out_0 = SMLAD(a12, b0, ch_1_out_0); in arm_nn_mat_mult_kernel_row_offset_s8_s16() 109 ch_1_out_1 = SMLAD(a12, b1, ch_1_out_1); in arm_nn_mat_mult_kernel_row_offset_s8_s16() 193 ch_0_out_0 = SMLAD(a01, b0, ch_0_out_0); in arm_nn_mat_mult_kernel_row_offset_s8_s16() 194 ch_0_out_1 = SMLAD(a01, b1, ch_0_out_1); in arm_nn_mat_mult_kernel_row_offset_s8_s16() [all …]
|
D | arm_nn_mat_mult_kernel_s8_s16.c | 95 ch_0_out_0 = SMLAD(a01, b0, ch_0_out_0); in arm_nn_mat_mult_kernel_s8_s16() 96 ch_0_out_1 = SMLAD(a01, b1, ch_0_out_1); in arm_nn_mat_mult_kernel_s8_s16() 97 ch_1_out_0 = SMLAD(a11, b0, ch_1_out_0); in arm_nn_mat_mult_kernel_s8_s16() 98 ch_1_out_1 = SMLAD(a11, b1, ch_1_out_1); in arm_nn_mat_mult_kernel_s8_s16() 103 ch_0_out_0 = SMLAD(a02, b0, ch_0_out_0); in arm_nn_mat_mult_kernel_s8_s16() 104 ch_0_out_1 = SMLAD(a02, b1, ch_0_out_1); in arm_nn_mat_mult_kernel_s8_s16() 105 ch_1_out_0 = SMLAD(a12, b0, ch_1_out_0); in arm_nn_mat_mult_kernel_s8_s16() 106 ch_1_out_1 = SMLAD(a12, b1, ch_1_out_1); in arm_nn_mat_mult_kernel_s8_s16() 188 ch_0_out_0 = SMLAD(a01, b0, ch_0_out_0); in arm_nn_mat_mult_kernel_s8_s16() 189 ch_0_out_1 = SMLAD(a01, b1, ch_0_out_1); in arm_nn_mat_mult_kernel_s8_s16() [all …]
|
D | arm_nn_mat_mult_kernel_s4_s16.c | 102 ch_0_out_0 = SMLAD(a01, b0, ch_0_out_0); in arm_nn_mat_mult_kernel_s4_s16() 103 ch_0_out_1 = SMLAD(a01, b1, ch_0_out_1); in arm_nn_mat_mult_kernel_s4_s16() 104 ch_1_out_0 = SMLAD(a11, b0, ch_1_out_0); in arm_nn_mat_mult_kernel_s4_s16() 105 ch_1_out_1 = SMLAD(a11, b1, ch_1_out_1); in arm_nn_mat_mult_kernel_s4_s16() 110 ch_0_out_0 = SMLAD(a02, b0, ch_0_out_0); in arm_nn_mat_mult_kernel_s4_s16() 111 ch_0_out_1 = SMLAD(a02, b1, ch_0_out_1); in arm_nn_mat_mult_kernel_s4_s16() 112 ch_1_out_0 = SMLAD(a12, b0, ch_1_out_0); in arm_nn_mat_mult_kernel_s4_s16() 113 ch_1_out_1 = SMLAD(a12, b1, ch_1_out_1); in arm_nn_mat_mult_kernel_s4_s16() 248 ch_0_out_0 = SMLAD(a01, b0, ch_0_out_0); in arm_nn_mat_mult_kernel_s4_s16() 249 ch_0_out_1 = SMLAD(a01, b1, ch_0_out_1); in arm_nn_mat_mult_kernel_s4_s16() [all …]
|
D | arm_nn_mat_mult_kernel_s16.c | 101 ch_0_out_0 = SMLAD(a01, b0, ch_0_out_0); in arm_nn_mat_mult_kernel_s16() 102 ch_0_out_1 = SMLAD(a01, b1, ch_0_out_1); in arm_nn_mat_mult_kernel_s16() 103 ch_1_out_0 = SMLAD(a11, b0, ch_1_out_0); in arm_nn_mat_mult_kernel_s16() 104 ch_1_out_1 = SMLAD(a11, b1, ch_1_out_1); in arm_nn_mat_mult_kernel_s16() 109 ch_0_out_0 = SMLAD(a02, b0, ch_0_out_0); in arm_nn_mat_mult_kernel_s16() 110 ch_0_out_1 = SMLAD(a02, b1, ch_0_out_1); in arm_nn_mat_mult_kernel_s16() 111 ch_1_out_0 = SMLAD(a12, b0, ch_1_out_0); in arm_nn_mat_mult_kernel_s16() 112 ch_1_out_1 = SMLAD(a12, b1, ch_1_out_1); in arm_nn_mat_mult_kernel_s16() 259 ch_0_out_0 = SMLAD(a01, b0, ch_0_out_0); in arm_nn_mat_mult_kernel_s16() 260 ch_0_out_1 = SMLAD(a01, b1, ch_0_out_1); in arm_nn_mat_mult_kernel_s16() [all …]
|
D | arm_depthwise_conv_s4_opt.c | 360 sum = SMLAD(op_c, op_b, sum); in arm_depthwise_conv_s4_opt() 364 sum_2 = SMLAD(op_a, op_b, sum_2); in arm_depthwise_conv_s4_opt() 372 sum_3 = SMLAD(op_c, op_b, sum_3); in arm_depthwise_conv_s4_opt() 375 sum_4 = SMLAD(op_a, op_b, sum_4); in arm_depthwise_conv_s4_opt() 560 sum = SMLAD(op_c, op_b, sum); in arm_depthwise_conv_s4_opt() 564 sum_2 = SMLAD(op_a, op_b, sum_2); in arm_depthwise_conv_s4_opt() 572 sum_3 = SMLAD(op_c, op_b, sum_3); in arm_depthwise_conv_s4_opt() 575 sum_4 = SMLAD(op_a, op_b, sum_4); in arm_depthwise_conv_s4_opt()
|
D | arm_depthwise_conv_s8_opt.c | 314 sum = SMLAD(op_c, op_b, sum); in arm_depthwise_conv_s8_opt() 317 sum_2 = SMLAD(op_a, op_b, sum_2); in arm_depthwise_conv_s8_opt() 325 sum_3 = SMLAD(op_c, op_b, sum_3); in arm_depthwise_conv_s8_opt() 328 sum_4 = SMLAD(op_a, op_b, sum_4); in arm_depthwise_conv_s8_opt()
|
D | arm_convolve_s16.c | 229 sum = SMLAD(ker_a1, ip_b1, sum); in arm_convolve_s16() 231 sum = SMLAD(ker_a2, ip_b2, sum); in arm_convolve_s16()
|
D | arm_depthwise_conv_fast_s16.c | 324 sum_1 = SMLAD(col_c, row_c, sum_1); in arm_depthwise_conv_fast_s16() 327 sum_2 = SMLAD(col_a, row_c, sum_2); in arm_depthwise_conv_fast_s16() 335 sum_3 = SMLAD(col_c, row_c, sum_3); in arm_depthwise_conv_fast_s16() 338 sum_4 = SMLAD(col_a, row_c, sum_4); in arm_depthwise_conv_fast_s16()
|
D | arm_convolve_s4.c | 280 sum = SMLAD(ker_a1, ip_b1, sum); in arm_convolve_s4() 282 sum = SMLAD(ker_a2, ip_b2, sum); in arm_convolve_s4()
|
D | arm_convolve_s8.c | 297 sum = SMLAD(ker_a1, ip_b1, sum); in arm_convolve_s8() 299 sum = SMLAD(ker_a2, ip_b2, sum); in arm_convolve_s8()
|
/cmsis-nn-latest/Include/Internal/ |
D | arm_nn_compiler.h | 207 #define SMLAD __smlad macro 249 __STATIC_FORCEINLINE uint32_t SMLAD(uint32_t op1, uint32_t op2, uint32_t op3) in SMLAD() function
|
/cmsis-nn-latest/Source/SVDFunctions/ |
D | arm_svdf_s8.c | 168 sum = SMLAD(r1_1, r2_1, sum); in arm_svdf_s8() 169 sum = SMLAD(r1_2, r2_2, sum); in arm_svdf_s8()
|
D | arm_svdf_state_s16_s8.c | 154 sum = SMLAD(r1, r2, sum); in arm_svdf_state_s16_s8()
|