Searched refs:x0 (Results 51 – 75 of 135) sorted by path
123456
71 between 0x0...0x3;
68 between 0x0...0x3;
74 between 0x0...0x7F;
75 between 0x0...0x7F;
73 between 0x0...0x7F;
73 between 0x0...0x7f;
71 between 0x0...0x7f;
123 / uint32_t x1, x0;136 / x0 = LO(x);170 / x0 = LO(dt);270 cmpl %edi, %esi / y0, x0274 subl %edi,%esi / y0, x0279 xorl %edi, %edi / x0331 subl %eax,%esi / t0, x0336 xorl %ebp, %ebp / x0407 / uint32_t x1, x0;420 / x0 = LO(x);[all …]
83 #define SET3(x, x0, x1, x2) ((x)[0] = (x0), (x)[1] = (x1), (x)[2] = (x2)) argument85 #define SEED(x0, x1, x2) (SET3(x, x0, x1, x2), SET3(a, A0, A1, A2), c = C) argument
73 x0 = hx ^ sx; in fmodl()101 x0 = (x0 << 1) | (x1 >> 31); in fmodl()108 x0 = iu | (x0 & im); in fmodl()145 x0 = (x0 << 1) | (x1 >> 31); in fmodl()169 x0 = x0 + x0 + ((x1 & is) != 0); in fmodl()186 x0 = z0; in fmodl()212 x0 = z0; in fmodl()231 x0 = x0 + x0 + ((x1 & is) != 0); in fmodl()258 x0 = 0; in fmodl()265 x1 = x0; in fmodl()[all …]
102 x0 = (x0 << 1) | (x1 >> 31); in fmodquol()109 x0 = iu | (x0 & im); in fmodquol()150 x0 = (x0 << 1) | (x1 >> 31); in fmodquol()174 x0 = x0 + x0 + ((x1 & is) != 0); in fmodquol()200 x0 = z0; in fmodquol()226 x0 = z0; in fmodquol()249 x0 = x0 + x0 + ((x1 & is) != 0); in fmodquol()270 x0 >>= k; in fmodquol()275 x0 = 0; in fmodquol()281 x1 = x0; in fmodquol()[all …]
47 mp_invert(MINT *x1, MINT *x0, MINT *c) in mp_invert() argument68 _mp_move(x0, &u3); in mp_invert()69 _mp_move(x0, &x0_prime); in mp_invert()
102 movlhps %xmm1,%xmm0 / xmm0: 0 x1 0 x0104 shufps $0x88,%xmm2,%xmm0 / xmm0: x3 x2 x1 x0
93 x0 = fabs(*y); in __vatan2()152 x0 *= twom3; in __vatan2()170 x0 *= two110; in __vatan2()172 hx = HI(&x0); in __vatan2()182 xh = x0; in __vatan2()184 z0 = ((y0 - t0 * xh) - t0 * (x0 - xh)) / (x0 + y0 * t0); in __vatan2()417 x0 = z0 * z0; in __vatan2()421 t0 = ah0 + (z0 + (al0 + (z0 * x0) * (p1 + x0 * in __vatan2()422 (p2 + x0 * (p3 + x0 * p4))))); in __vatan2()449 t0 = ah0 + (z0 + (al0 + (z0 * x0) * (p1 + x0 * in __vatan2()[all …]
82 x0 = *y; in __vatan2f()89 x0 = -x0; in __vatan2f()106 x0 = *x; in __vatan2f()114 x0 = -x0; in __vatan2f()127 ah0 = x0 + y0; in __vatan2f()132 ah0 = y0 / x0; in __vatan2f()156 x0 *= twop24; /* scale possibly subnormal x */ in __vatan2f()158 hx = *(int*)&x0; in __vatan2f()418 dx0 = x0; in __vatan2f()469 t0 = (y0 - x0 * (double)base0) / in __vatan2f()[all …]
220 z0 = x0 * x0; in __vcos()257 z0 = x0 * x0; in __vcos()288 z0 = x0 * x0; in __vcos()315 z0 = x0 * x0; in __vcos()342 z0 = x0 * x0; in __vcos()369 z0 = x0 * x0; in __vcos()392 z0 = x0 * x0; in __vcos()411 z0 = x0 * x0; in __vcos()469 z0 = x0 * x0; in __vcos()481 z0 = x0 * x0; in __vcos()[all …]
239 x0 = (x0 - t0) + y0; in __vlibm_vcos_big_ultra3()242 z0 = x0 * x0; in __vlibm_vcos_big_ultra3()288 z0 = x0 * x0; in __vlibm_vcos_big_ultra3()329 z0 = x0 * x0; in __vlibm_vcos_big_ultra3()371 z0 = x0 * x0; in __vlibm_vcos_big_ultra3()407 z0 = x0 * x0; in __vlibm_vcos_big_ultra3()449 z0 = x0 * x0; in __vlibm_vcos_big_ultra3()486 z0 = x0 * x0; in __vlibm_vcos_big_ultra3()524 z0 = x0 * x0; in __vlibm_vcos_big_ultra3()626 z0 = x0 * x0; in __vlibm_vcos_big_ultra3()[all …]