Lines Matching refs:m64_f32
133 float m64_f32[2]; member
4400 res.m64_f32[0] = a.m64_f32[0] - b.m64_f32[0]; in vsub_f32()
4401 res.m64_f32[1] = a.m64_f32[1] - b.m64_f32[1]; in vsub_f32()
6136 res.m64_f32[0] = (a.m64_f32[0] > b.m64_f32[0]) ? a.m64_f32[0] : b.m64_f32[0]; in vmax_f32()
6137 res.m64_f32[1] = (a.m64_f32[1] > b.m64_f32[1]) ? a.m64_f32[1] : b.m64_f32[1]; in vmax_f32()
6226 res.m64_f32[0] = (a.m64_f32[0] < b.m64_f32[0]) ? a.m64_f32[0] : b.m64_f32[0]; in vmin_f32()
6227 res.m64_f32[1] = (a.m64_f32[1] < b.m64_f32[1]) ? a.m64_f32[1] : b.m64_f32[1]; in vmin_f32()
6675 res.m64_f32[0] = (a.m64_f32[0] < a.m64_f32[1]) ? a.m64_f32[1] : a.m64_f32[0]; in _NEON2SSE_PERFORMANCE_WARNING()
6676 res.m64_f32[1] = (b.m64_f32[0] < b.m64_f32[1]) ? b.m64_f32[1] : b.m64_f32[0]; in _NEON2SSE_PERFORMANCE_WARNING()
6764 res.m64_f32[0] = (a.m64_f32[0] > a.m64_f32[1]) ? a.m64_f32[1] : a.m64_f32[0]; in _NEON2SSE_PERFORMANCE_WARNING()
6765 res.m64_f32[1] = (b.m64_f32[0] > b.m64_f32[1]) ? b.m64_f32[1] : b.m64_f32[0]; in _NEON2SSE_PERFORMANCE_WARNING()
6944 res.m64_f32[0] = (3 - a.m64_f32[0] * b.m64_f32[0]) / 2; in vrsqrts_f32()
6945 res.m64_f32[1] = (3 - a.m64_f32[1] * b.m64_f32[1]) / 2; in vrsqrts_f32()
9313 res.m64_f32[0] = *(ptr);
9314 res.m64_f32[1] = *(ptr + 1);
9434 res.m64_f32[lane] = *(ptr);
9550 res.m64_f32[0] = *(ptr);
9551 res.m64_f32[1] = res.m64_f32[0];
9672 *(ptr) = val.m64_f32[0];
9673 *(ptr + 1) = val.m64_f32[1];
9770 *(ptr) = val.m64_f32[lane];
9905 v.val[0].m64_f32[0] = *(ptr);
9906 v.val[0].m64_f32[1] = *(ptr + 2);
9907 v.val[1].m64_f32[0] = *(ptr + 1);
9908 v.val[1].m64_f32[1] = *(ptr + 3);
10194 v.val[0].m64_f32[0] = *(ptr);
10195 v.val[0].m64_f32[1] = *(ptr + 3);
10197 v.val[1].m64_f32[0] = *(ptr + 1);
10198 v.val[1].m64_f32[1] = *(ptr + 4);
10200 v.val[2].m64_f32[0] = *(ptr + 2);
10201 v.val[2].m64_f32[1] = *(ptr + 5);
10413 res.val[0].m64_f32[0] = *(ptr);
10414 res.val[0].m64_f32[1] = *(ptr + 4);
10415 res.val[1].m64_f32[0] = *(ptr + 1);
10416 res.val[1].m64_f32[1] = *(ptr + 5);
10417 res.val[2].m64_f32[0] = *(ptr + 2);
10418 res.val[2].m64_f32[1] = *(ptr + 6);
10419 res.val[3].m64_f32[0] = *(ptr + 3);
10420 res.val[3].m64_f32[1] = *(ptr + 7);
10491 v.val[0].m64_f32[0] = *(ptr); //0,0
10492 v.val[0].m64_f32[1] = *(ptr); //0,0
10493 v.val[1].m64_f32[0] = *(ptr + 1); //1,1
10494 v.val[1].m64_f32[1] = *(ptr + 1); //1,1
10584 v.val[i].m64_f32[0] = *(ptr + i);
10585 v.val[i].m64_f32[1] = *(ptr + i);
10681 v.val[i].m64_f32[0] = *(ptr + i);
10682 v.val[i].m64_f32[1] = *(ptr + i);
11183 *(ptr) = val.val[0].m64_f32[0];
11184 *(ptr + 1) = val.val[1].m64_f32[0];
11185 *(ptr + 2) = val.val[0].m64_f32[1];
11186 *(ptr + 3) = val.val[1].m64_f32[1];
11412 *(ptr) = val.val[0].m64_f32[0];
11413 *(ptr + 1) = val.val[1].m64_f32[0];
11414 *(ptr + 2) = val.val[2].m64_f32[0];
11415 *(ptr + 3) = val.val[0].m64_f32[1];
11416 *(ptr + 4) = val.val[1].m64_f32[1];
11417 *(ptr + 5) = val.val[2].m64_f32[1];
11598 *(ptr) = val.val[0].m64_f32[0];
11599 *(ptr + 1) = val.val[1].m64_f32[0];
11600 *(ptr + 2) = val.val[2].m64_f32[0];
11601 *(ptr + 3) = val.val[3].m64_f32[0];
11602 *(ptr + 4) = val.val[0].m64_f32[1];
11603 *(ptr + 5) = val.val[1].m64_f32[1];
11604 *(ptr + 6) = val.val[2].m64_f32[1];
11605 *(ptr + 7) = val.val[3].m64_f32[1];
11692 *(ptr) = val.val[0].m64_f32[lane];
11693 *(ptr + 1) = val.val[1].m64_f32[lane];
11785 *(ptr) = val.val[0].m64_f32[lane];
11786 *(ptr + 1) = val.val[1].m64_f32[lane];
11787 *(ptr + 2) = val.val[2].m64_f32[lane];
11883 *(ptr) = val.val[0].m64_f32[lane];
11884 *(ptr + 1) = val.val[1].m64_f32[lane];
11885 *(ptr + 2) = val.val[2].m64_f32[lane];
11886 *(ptr + 3) = val.val[3].m64_f32[lane];
11925 #define vget_lane_f32(vec, lane) vec.m64_f32[lane]
12250 res.m64_f32[0] = value; in vdup_n_f32()
12251 res.m64_f32[1] = value; in vdup_n_f32()
12438 res.m64_f32[0] = vec.m64_f32[lane];
12439 res.m64_f32[1] = res.m64_f32[0];
12488 #define vdupq_lane_f32(vec, lane) _mm_load1_ps((vec.m64_f32 + lane))
12757 res.m64_u32[0] = (uint32_t) (a.m64_f32[0] * convconst);
12758 res.m64_u32[1] = (uint32_t) (a.m64_f32[1] * convconst);
12799 res.m64_f32[0] = (float) a.m64_i32[0]; in vcvt_f32_s32()
12800 res.m64_f32[1] = (float) a.m64_i32[1]; in vcvt_f32_s32()
12808 res.m64_f32[0] = (float) a.m64_u32[0]; in vcvt_f32_u32()
12809 res.m64_f32[1] = (float) a.m64_u32[1]; in vcvt_f32_u32()
12840 res.m64_f32[0] = a.m64_i32[0] * convconst;
12841 res.m64_f32[1] = a.m64_i32[1] * convconst;
12851 res.m64_f32[0] = a.m64_u32[0] * convconst;
12852 res.m64_f32[1] = a.m64_u32[1] * convconst;
14175 res.m64_f32[0] = a.m64_f32[0] - b.m64_f32[0] * c; in vmls_n_f32()
14176 res.m64_f32[1] = a.m64_f32[1] - b.m64_f32[1] * c; in vmls_n_f32()
14336 res.m64_f32[0] = a.m64_f32[0];
14337 res.m64_f32[1] = a.m64_f32[1];
14339 res.m64_f32[0] = a.m64_f32[1];
14340 res.m64_f32[1] = b.m64_f32[0];
14437 res.m64_f32[0] = vec.m64_f32[1]; in vrev64_f32()
14438 res.m64_f32[1] = vec.m64_f32[0]; in vrev64_f32()
15703 val.val[0].m64_f32[0] = a.m64_f32[0]; in vtrn_f32()
15704 val.val[0].m64_f32[1] = b.m64_f32[0]; in vtrn_f32()
15705 val.val[1].m64_f32[0] = a.m64_f32[1]; in vtrn_f32()
15706 val.val[1].m64_f32[1] = b.m64_f32[1]; in vtrn_f32()