/external/eigen/unsupported/test/ |
D | openglsupport.cpp | 183 Vector2d vd2; vd2.setRandom(); Vector3d vd23; vd23 << vd2, 0; in test_openglsupport() local 184 VERIFY_MATRIX(glTranslate(vd2), Projective3d(Translation3d(vd23)).matrix()); in test_openglsupport() 202 Vector2d vd2; vd2.setRandom(); Vector3d vd23; vd23 << vd2, 1; in test_openglsupport() local 203 VERIFY_MATRIX(glScale(vd2), Projective3d(Scaling(vd23)).matrix()); in test_openglsupport()
|
/external/llvm-project/clang/test/CodeGen/SystemZ/ |
D | builtins-systemz-zvector3-constrained.c | 34 vector double vd2; in test_core() local 59 vd2 = vd; in test_core() 60 vd += vec_revb(vec_insert(d, vec_revb(vd2), 0)); in test_core()
|
D | builtins-systemz-zvector2-constrained.c | 36 vector double vd2; in test_core() local 61 vd2 = vd; in test_core() 62 vd = vec_insert(d, vd2, 0); in test_core()
|
/external/llvm/test/CodeGen/PowerPC/ |
D | vsx-minmax.ll | 12 @vd2 = common global <2 x double> zeroinitializer, align 16 43 store <2 x double> %12, <2 x double>* @vd2, align 16 86 ;vector double vd1, vd2; 94 ; vd2 = vec_min(vd, vd);
|
/external/llvm-project/llvm/test/CodeGen/PowerPC/ |
D | vsx-minmax.ll | 12 @vd2 = common global <2 x double> zeroinitializer, align 16 43 store <2 x double> %12, <2 x double>* @vd2, align 16 86 ;vector double vd1, vd2; 94 ; vd2 = vec_min(vd, vd);
|
/external/XNNPACK/src/f32-sigmoid/gen/ |
D | avx2-rr1-p5-nr2fma-x24.c | 92 const __m256 vd2 = _mm256_add_ps(ve2, vone); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x24() local 96 __m256 vr2 = _mm256_rcp_ps(vd2); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x24() 100 vr2 = _mm256_fmadd_ps(_mm256_fnmadd_ps(vr2, vd2, vone), vr2, vr2); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x24() 104 vr2 = _mm256_fmadd_ps(_mm256_fnmadd_ps(vr2, vd2, vone), vr2, vr2); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x24()
|
D | avx2-rr1-p5-nr2fma-x32.c | 104 const __m256 vd2 = _mm256_add_ps(ve2, vone); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x32() local 109 __m256 vr2 = _mm256_rcp_ps(vd2); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x32() 114 vr2 = _mm256_fmadd_ps(_mm256_fnmadd_ps(vr2, vd2, vone), vr2, vr2); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x32() 119 vr2 = _mm256_fmadd_ps(_mm256_fnmadd_ps(vr2, vd2, vone), vr2, vr2); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x32()
|
D | avx-rr2-p5-nr2-x24.c | 104 const __m256 vd2 = _mm256_add_ps(ve2, vone); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x24() local 108 __m256 vr2 = _mm256_rcp_ps(vd2); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x24() 114 vr2 = _mm256_mul_ps(vr2, _mm256_sub_ps(vtwo, _mm256_mul_ps(vr2, vd2))); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x24() 115 vr2 = _mm256_mul_ps(vr2, _mm256_sub_ps(vtwo, _mm256_mul_ps(vr2, vd2))); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x24()
|
D | avx512f-rr2-lut32-p2-perm2-scalef-nr1fma-x48.c | 94 const __m512 vd2 = _mm512_add_ps(ve2, vone); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_nr1fma_x48() local 98 __m512 vr2 = _mm512_rcp14_ps(vd2); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_nr1fma_x48() 102 vr2 = _mm512_fmadd_ps(_mm512_fnmadd_ps(vr2, vd2, vone), vr2, vr2); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_nr1fma_x48()
|
D | avx512f-rr1-p5-scalef-nr1fma-x48.c | 85 const __m512 vd2 = _mm512_add_ps(ve2, vone); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_nr1fma_x48() local 89 __m512 vr2 = _mm512_rcp14_ps(vd2); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_nr1fma_x48() 93 vr2 = _mm512_fmadd_ps(_mm512_fnmadd_ps(vr2, vd2, vone), vr2, vr2); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_nr1fma_x48()
|
D | avx512f-rr1-lut16-p3-perm-scalef-nr1fma-x48.c | 88 const __m512 vd2 = _mm512_add_ps(ve2, vone); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_nr1fma_x48() local 92 __m512 vr2 = _mm512_rcp14_ps(vd2); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_nr1fma_x48() 96 vr2 = _mm512_fmadd_ps(_mm512_fnmadd_ps(vr2, vd2, vone), vr2, vr2); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_nr1fma_x48()
|
D | avx2-rr1-p5-nr1fma-x24.c | 92 const __m256 vd2 = _mm256_add_ps(ve2, vone); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr1fma_x24() local 96 __m256 vr2 = _mm256_rcp_ps(vd2); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr1fma_x24() 100 vr2 = _mm256_fmadd_ps(_mm256_fnmadd_ps(vr2, vd2, vone), vr2, vr2); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr1fma_x24()
|
D | avx512f-rr1-p5-scalef-nr1fma-x64.c | 96 const __m512 vd2 = _mm512_add_ps(ve2, vone); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_nr1fma_x64() local 101 __m512 vr2 = _mm512_rcp14_ps(vd2); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_nr1fma_x64() 106 vr2 = _mm512_fmadd_ps(_mm512_fnmadd_ps(vr2, vd2, vone), vr2, vr2); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_nr1fma_x64()
|
D | avx2-rr1-p5-nr2fma-x40.c | 116 const __m256 vd2 = _mm256_add_ps(ve2, vone); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x40() local 122 __m256 vr2 = _mm256_rcp_ps(vd2); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x40() 128 vr2 = _mm256_fmadd_ps(_mm256_fnmadd_ps(vr2, vd2, vone), vr2, vr2); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x40() 134 vr2 = _mm256_fmadd_ps(_mm256_fnmadd_ps(vr2, vd2, vone), vr2, vr2); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x40()
|
D | scalar-lut2048-p1-div-x4.c | 97 const float vd2 = vy2 + vone; in xnn_f32_sigmoid_ukernel__scalar_lut2048_p1_div_x4() local 102 float vf2 = vy2 / vd2; in xnn_f32_sigmoid_ukernel__scalar_lut2048_p1_div_x4()
|
D | scalar-lut64-p2-div-x4.c | 102 const float vd2 = vy2 + vone; in xnn_f32_sigmoid_ukernel__scalar_lut64_p2_div_x4() local 107 float vf2 = vy2 / vd2; in xnn_f32_sigmoid_ukernel__scalar_lut64_p2_div_x4()
|
D | scalar-p5-div-x4.c | 108 const float vd2 = ve2 + vone; in xnn_f32_sigmoid_ukernel__scalar_p5_div_x4() local 113 float vf2 = ve2 / vd2; in xnn_f32_sigmoid_ukernel__scalar_p5_div_x4()
|
D | avx-rr2-p5-nr2-x32.c | 119 const __m256 vd2 = _mm256_add_ps(ve2, vone); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x32() local 124 __m256 vr2 = _mm256_rcp_ps(vd2); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x32() 131 vr2 = _mm256_mul_ps(vr2, _mm256_sub_ps(vtwo, _mm256_mul_ps(vr2, vd2))); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x32() 132 vr2 = _mm256_mul_ps(vr2, _mm256_sub_ps(vtwo, _mm256_mul_ps(vr2, vd2))); in xnn_f32_sigmoid_ukernel__avx_rr2_p5_nr2_x32()
|
D | avx512f-rr2-lut32-p2-perm2-scalef-nr1fma-x64.c | 105 const __m512 vd2 = _mm512_add_ps(ve2, vone); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_nr1fma_x64() local 110 __m512 vr2 = _mm512_rcp14_ps(vd2); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_nr1fma_x64() 115 vr2 = _mm512_fmadd_ps(_mm512_fnmadd_ps(vr2, vd2, vone), vr2, vr2); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_nr1fma_x64()
|
D | avx512f-rr1-lut16-p3-perm-scalef-nr1fma-x64.c | 99 const __m512 vd2 = _mm512_add_ps(ve2, vone); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_nr1fma_x64() local 104 __m512 vr2 = _mm512_rcp14_ps(vd2); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_nr1fma_x64() 109 vr2 = _mm512_fmadd_ps(_mm512_fnmadd_ps(vr2, vd2, vone), vr2, vr2); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_nr1fma_x64()
|
D | avx2-rr1-p5-nr2fma-x48.c | 128 const __m256 vd2 = _mm256_add_ps(ve2, vone); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x48() local 135 __m256 vr2 = _mm256_rcp_ps(vd2); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x48() 142 vr2 = _mm256_fmadd_ps(_mm256_fnmadd_ps(vr2, vd2, vone), vr2, vr2); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x48() 149 vr2 = _mm256_fmadd_ps(_mm256_fnmadd_ps(vr2, vd2, vone), vr2, vr2); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x48()
|
/external/XNNPACK/src/f32-ibilinear/gen/ |
D | scalar-c4.c | 82 const float vd2 = vb2 - vt2; in xnn_f32_ibilinear_ukernel__scalar_c4() local 87 const float vo2 = vt2 + vd2 * valphav; in xnn_f32_ibilinear_ukernel__scalar_c4()
|
/external/deqp-deps/glslang/Test/baseResults/ |
D | hlsl.intrinsics.double.frag.out | 5 0:5 Function Definition: @PixelShaderFunction(d1;d1;d1;vd2;vd3;vd4;u1;u1; ( temp float) 62 0:5 Function Call: @PixelShaderFunction(d1;d1;d1;vd2;vd3;vd4;u1;u1; ( temp float) 89 0:5 Function Definition: @PixelShaderFunction(d1;d1;d1;vd2;vd3;vd4;u1;u1; ( temp float) 146 0:5 Function Call: @PixelShaderFunction(d1;d1;d1;vd2;vd3;vd4;u1;u1; ( temp float) 178 Name 26 "@PixelShaderFunction(d1;d1;d1;vd2;vd3;vd4;u1;u1;" 312 …89: 16(float) FunctionCall 26(@PixelShaderFunction(d1;d1;d1;vd2;vd3;vd4;u1;u1;) 73(param) 75(par… 316 26(@PixelShaderFunction(d1;d1;d1;vd2;vd3;vd4;u1;u1;): 16(float) Function None 17
|
/external/angle/third_party/vulkan-deps/glslang/src/Test/baseResults/ |
D | hlsl.intrinsics.double.frag.out | 5 0:5 Function Definition: @PixelShaderFunction(d1;d1;d1;vd2;vd3;vd4;u1;u1; ( temp float) 62 0:5 Function Call: @PixelShaderFunction(d1;d1;d1;vd2;vd3;vd4;u1;u1; ( temp float) 89 0:5 Function Definition: @PixelShaderFunction(d1;d1;d1;vd2;vd3;vd4;u1;u1; ( temp float) 146 0:5 Function Call: @PixelShaderFunction(d1;d1;d1;vd2;vd3;vd4;u1;u1; ( temp float) 178 Name 26 "@PixelShaderFunction(d1;d1;d1;vd2;vd3;vd4;u1;u1;" 312 …89: 16(float) FunctionCall 26(@PixelShaderFunction(d1;d1;d1;vd2;vd3;vd4;u1;u1;) 73(param) 75(par… 316 26(@PixelShaderFunction(d1;d1;d1;vd2;vd3;vd4;u1;u1;): 16(float) Function None 17
|
/external/XNNPACK/src/f32-ibilinear-chw/gen/ |
D | scalar-p4.c | 92 const float vd2 = vb2 - vt2; in xnn_f32_ibilinear_chw_ukernel__scalar_p4() local 97 const float vo2 = vt2 + vd2 * valphav2; in xnn_f32_ibilinear_chw_ukernel__scalar_p4()
|