Home
last modified time | relevance | path

Searched refs:xabs (Results 1 – 21 of 21) sorted by relevance

/picolibc-latest/newlib/libm/machine/spu/headers/
Dasinf4.h80 vector float xabs, x2, x4, x6; in _asinf4() local
92 xabs = spu_andc(x, (vector float)msb); in _asinf4()
115 r1 = spu_madd(_divf4(p, q), spu_mul(xabs, x2), xabs); in _asinf4()
119 r_hi = spu_madd(spu_splats(-0.0012624911f), xabs, in _asinf4()
121 r_hi = spu_madd(r_hi, xabs, spu_splats(-0.0170881256f)); in _asinf4()
122 r_hi = spu_madd(r_hi, xabs, spu_splats(0.0308918810f)); in _asinf4()
124 r_lo = spu_madd(spu_splats(-0.0501743046f), xabs, in _asinf4()
126 r_lo = spu_madd(r_lo, xabs, spu_splats(-0.2145988016f)); in _asinf4()
127 r_lo = spu_madd(r_lo, xabs, pi_over_2); in _asinf4()
131 r2 = spu_nmsub(r, _sqrtf4(spu_sub(spu_splats(1.0f), xabs)), in _asinf4()
[all …]
Dlgammad2.h157 vec_double2 sum, xabs; in _lgammad2() local
166 xabs = spu_andc(x, sign_maskd); in _lgammad2()
167 xf = spu_roundtf(xabs); in _lgammad2()
179 mresult = spu_madd(xabs, spu_splats(ZETA_06_DIV_06), spu_splats(ZETA_05_DIV_05)); in _lgammad2()
180 mresult = spu_madd(xabs, mresult, spu_splats(ZETA_04_DIV_04)); in _lgammad2()
181 mresult = spu_madd(xabs, mresult, spu_splats(ZETA_03_DIV_03)); in _lgammad2()
182 mresult = spu_madd(xabs, mresult, spu_splats(ZETA_02_DIV_02)); in _lgammad2()
183 mresult = spu_mul(xabs, spu_mul(xabs, mresult)); in _lgammad2()
184 mresult = spu_sub(mresult, spu_add(_logd2(xabs), spu_mul(xabs, spu_splats(EULER_MASCHERONI)))); in _lgammad2()
201 xstirling = spu_sel(xabs, spu_add(xabs, spu_splats(6.0)), (vec_ullong2)isshifted); in _lgammad2()
[all …]
Datanhf4.h92 vec_float4 xabs, xsqu; in _atanhf4() local
97 xabs = spu_andc(x, sign_mask); in _atanhf4()
104 fresult = spu_sub(_logf4(spu_add(onef, xabs)), _logf4(spu_sub(onef, xabs))); in _atanhf4()
114 mresult = spu_mul(xabs, mresult); in _atanhf4()
119 use_form = spu_cmpgt(xabs, switch_approx); in _atanhf4()
126 … result = (vec_float4)spu_add((vec_uint4)result, spu_and(one, spu_cmpgt(xabs, spu_splats(0.0f)))); in _atanhf4()
127 … result = (vec_float4)spu_add((vec_uint4)result, spu_and(one, spu_cmpgt(xabs, spu_splats(0.25f)))); in _atanhf4()
132 result = spu_sel(result, huge, spu_cmpeq(xabs, onef)); in _atanhf4()
Dasinhf4.h120 vec_float4 xabs, xsqu; in _asinhf4() local
128 xabs = spu_andc(x, sign_mask); in _asinhf4()
130 islarge = spu_cmpgt(xabs, largef); in _asinhf4()
137 vec_float4 logarg = spu_add(xabs, _sqrtf4(spu_madd(xabs, xabs, onef))); in _asinhf4()
138 logarg = spu_sel(logarg, xabs, islarge); in _asinhf4()
160 mresult = spu_mul(xabs, mresult); in _asinhf4()
165 use_form = spu_cmpgt(xabs, switch_approx); in _asinhf4()
172 truncadd = spu_sel(truncadd, twou, spu_cmpgt(xabs, trunc_part2)); in _asinhf4()
Datand2.h72 vector double xabs, x1; in _atand2() local
76 xabs = spu_andc(x, signbit); in _atand2()
77 gt1 = spu_cmpgt(xabs, oned); in _atand2()
83 x1 = spu_sel(xabs, _recipd2(xabs), gt1); in _atand2()
Dtanhf4.h74 vector float xabs; in _tanhf4() local
82 xabs = spu_andc(x, signbit); in _tanhf4()
88 gttaylor = spu_cmpgt(xabs, spu_splats(0.25f)); in _tanhf4()
100 tresult = spu_mul(xabs, tresult); in _tanhf4()
108 e = _expf4(spu_mul(xabs, twof)); in _tanhf4()
123 result = spu_sel(result, onef, spu_cmpgt(xabs, spu_splats(9.125f))); in _tanhf4()
Dasinhd2.h102 vec_double2 xabs, xsqu; in _asinhd2() local
110 xabs = spu_andc(x, sign_mask); in _asinhd2()
113 xf = spu_roundtf(xabs); in _asinhd2()
121 fresult = spu_add(xabs, fresult); in _asinhd2()
138 mresult = spu_mul(xabs, mresult); in _asinhd2()
151 xabshigh = (vec_uint4)spu_shuffle(xabs, xabs, dup_even); in _asinhd2()
Dlogbf4.h75 vec_uint4 xabs; in _logbf4() local
79 xabs = spu_andc((vec_uint4)x, sign_mask); in _logbf4()
85 exp_unbias = spu_sel(exp_unbias, (vec_float4)lminf, (vec_uint4)spu_cmpeq(xabs, lzero)); in _logbf4()
88 exp_unbias = spu_sel(exp_unbias, x, (vec_uint4)spu_cmpgt(xabs, linf)); in _logbf4()
91 exp_unbias = spu_sel(exp_unbias, (vec_float4)linf, (vec_uint4)spu_cmpeq(xabs, linf)); in _logbf4()
Derfd2.h77 vec_double2 xabs, xsqu, xsign; in _erfd2() local
81 xabs = spu_andc(x, sign_mask); in _erfd2()
87 TAYLOR_ERF(xabs, xsqu, tresult); in _erfd2()
93 CONTFRAC_ERFC(xabs, xsqu, presult); in _erfd2()
100 vec_float4 xf = spu_roundtf(xabs); in _erfd2()
Datanhd2.h101 vec_double2 xabs, xsqu; in _atanhd2() local
108 xabs = spu_andc(x, sign_mask); in _atanhd2()
111 xf = spu_roundtf(xabs); in _atanhd2()
118 fresult = spu_sub(_logd2(spu_add(oned, xabs)), _logd2(spu_sub(oned, xabs))); in _atanhd2()
133 mresult = spu_mul(xabs, mresult); in _atanhd2()
Dtanhd2.h97 vector double xabs; in _tanhd2() local
105 xabs = spu_andc(x, signbit); in _tanhd2()
111 gttaylor = spu_cmpgt(xabs, spu_splats(0.25)); in _tanhd2()
128 tresult = spu_mul(xabs, tresult); in _tanhd2()
136 e = _expd2(spu_mul(xabs, twod)); in _tanhd2()
149 result = spu_sel(result, oned, spu_cmpeq(xabs, infd)); in _tanhd2()
Derfcd2.h82 vec_double2 xabs, xsqu, xsign; in _erfcd2() local
87 xabs = spu_andc(x, sign_mask); in _erfcd2()
95 TAYLOR_ERF(xabs, xsqu, tresult); in _erfcd2()
103 CONTFRAC_ERFC(xabs, xsqu, presult); in _erfcd2()
111 vec_float4 xf = spu_roundtf(xabs); in _erfcd2()
Dlgammaf4.h94 vec_float4 xabs = spu_andc(x, sign_maskf); in _lgammaf4() local
95 vec_float4 ln_x = _logf4(xabs); in _lgammaf4()
96 vec_float4 inv_x = _recipf4(xabs); in _lgammaf4()
99 vec_uint4 isnaninf = spu_cmpgt((vec_uint4)xabs, 0x7F7FFFFF); in _lgammaf4()
382 vec_uchar16 gt_r1start = (vec_uchar16)spu_cmpgt(xabs, r1start); in _lgammaf4()
383 vec_uchar16 gt_r2start = (vec_uchar16)spu_cmpgt(xabs, r2start); in _lgammaf4()
384 vec_uchar16 gt_r3start = (vec_uchar16)spu_cmpgt(xabs, r3start); in _lgammaf4()
385 vec_uchar16 gt_r4start = (vec_uchar16)spu_cmpgt(xabs, r4start); in _lgammaf4()
386 vec_uchar16 gt_r5start = (vec_uchar16)spu_cmpgt(xabs, r5start); in _lgammaf4()
387 vec_uchar16 gt_r6start = (vec_uchar16)spu_cmpgt(xabs, r6start); in _lgammaf4()
[all …]
Datanf.h68 float xabs; in _atanf() local
75 xabs = (x < 0.0f) ? -x : x; in _atanf()
77 if (xabs >= 1.0f) { in _atanf()
Dtgammaf4.h105 vector float xabs; in _tgammaf4() local
120 xabs = spu_andc(x, signbit); in _tgammaf4()
129 xabs = spu_sel(spu_add(xabs, onef), xabs, gt1); in _tgammaf4()
130 xtrunc = _truncf4(xabs); in _tgammaf4()
136 xscaled = spu_add(onef, spu_sub(xabs, xtrunc)); in _tgammaf4()
185 xinv = _recipf4(xabs); in _tgammaf4()
Dtgammad2.h159 vector double xabs; in _tgammad2() local
175 xabs = spu_andc(x, signbit); in _tgammad2()
184 xabs = spu_sel(spu_add(xabs, oned), xabs, gt1); in _tgammad2()
185 xtrunc = _truncd2(xabs); in _tgammad2()
191 xscaled = spu_add(oned, spu_sub(xabs, xtrunc)); in _tgammad2()
240 xinv = _recipd2(xabs); in _tgammad2()
Dpowd2.h73 vector double xabs, xsign; in _powd2() local
78 xabs = spu_andc(x, signmask); in _powd2()
83 out = _exp2d2(spu_mul(y, _log2d2(xabs))); in _powd2()
Dcoshf4.h91 vec_uint4 xabs = spu_andc((vec_uint4)x,sign_mask); in _coshf4() local
92 vec_uint4 use_exp = spu_cmpgt(xabs,threshold); in _coshf4()
Dsinhf4.h91 vec_uint4 xabs = spu_andc((vec_uint4)x,sign_mask); in _sinhf4() local
92 vec_uint4 use_exp = spu_cmpgt(xabs,threshold); in _sinhf4()
Derfcf4.h72 vec_float4 xabs = spu_andc(x, sign_maskf); in _erfcf4() local
393 vec_float4 xappr = spu_sub(xabs, xoffset); in _erfcf4()
423 CONTFRAC_ERFCF4(xabs, xsqu, presult); in _erfcf4()
426 result = spu_sel(presult, result, spu_cmpgt(spu_splats(4.3f), xabs)); in _erfcf4()
429 result = spu_sel(result, zerof, spu_cmpgt(xabs, clamp)); in _erfcf4()
Derff4.h65 vec_float4 xabs = spu_andc(x, sign_maskf); in _erff4() local
363 vec_float4 xappr = spu_sub(xabs, xoffset); in _erff4()
394 result = spu_sel(result, zerof, spu_cmpeq(xabs, zerof)); in _erff4()
397 result = spu_sel(result, onef, spu_cmpgt(xabs, clamp)); in _erff4()