/third_party/gn/src/gn/ |
D | immutable_vector_unittest.cc | 16 ImmutableVector<int> vec1 = {100, 42}; in TEST() local 17 EXPECT_FALSE(vec1.empty()); in TEST() 18 EXPECT_EQ(2u, vec1.size()); in TEST() 19 EXPECT_EQ(100, vec1.front()); in TEST() 20 EXPECT_EQ(42, vec1.back()); in TEST() 21 EXPECT_EQ(100, vec1[0]); in TEST() 22 EXPECT_EQ(42, vec1[1]); in TEST() 23 EXPECT_TRUE(vec1.begin()); in TEST() 24 EXPECT_TRUE(vec1.end()); in TEST() 25 EXPECT_NE(vec1.begin(), vec1.end()); in TEST() [all …]
|
/third_party/cmsis/CMSIS/DSP/Source/QuaternionMathFunctions/ |
D | arm_quaternion_inverse_f32.c | 64 f32x4_t vec1,vec2; in arm_quaternion_inverse_f32() local 70 vec1 = vld1q(pInputQuaternions); in arm_quaternion_inverse_f32() 71 vec2 = vmulq(vec1,vec1); in arm_quaternion_inverse_f32() 75 vec1 = vmulq_n_f32(vec1, 1.0f / squaredSum); in arm_quaternion_inverse_f32() 76 vec1 = vsetq_lane_f32(-vgetq_lane(vec1, 0),vec1,0); in arm_quaternion_inverse_f32() 77 vec1 = vnegq_f32(vec1); in arm_quaternion_inverse_f32() 79 vst1q(pInverseQuaternions, vec1); in arm_quaternion_inverse_f32()
|
D | arm_quaternion_conjugate_f32.c | 62 f32x4_t vec1; in arm_quaternion_conjugate_f32() local 66 vec1 = vld1q(pInputQuaternions); in arm_quaternion_conjugate_f32() 69 vec1 = vsetq_lane_f32(-vgetq_lane(vec1, 0),vec1,0); in arm_quaternion_conjugate_f32() 70 vec1 = vnegq_f32(vec1); in arm_quaternion_conjugate_f32() 72 vst1q(pConjugateQuaternions, vec1); in arm_quaternion_conjugate_f32()
|
D | arm_quaternion_normalize_f32.c | 63 f32x4_t vec1,vec2; in arm_quaternion_normalize_f32() local 68 vec1 = vld1q(pInputQuaternions); in arm_quaternion_normalize_f32() 69 vec2 = vmulq(vec1,vec1); in arm_quaternion_normalize_f32() 72 vec1 = vmulq_n_f32(vec1, 1.0f / norm); in arm_quaternion_normalize_f32() 73 vst1q(pNormalizedQuaternions, vec1); in arm_quaternion_normalize_f32()
|
D | arm_quaternion2rotation_f32.c | 85 f32x4_t vec0,vec1, vec2 ,vec3; in arm_quaternion2rotation_f32() local 95 vec1 = vmulq(vec0,vec0); in arm_quaternion2rotation_f32() 114 vec0 = vsetq_lane(vgetq_lane(vec1,0) + vgetq_lane(vec1,1),vec0,0); in arm_quaternion2rotation_f32() 115 vec0 = vsetq_lane(vgetq_lane(vec0,0) - vgetq_lane(vec1,2),vec0,0); in arm_quaternion2rotation_f32() 116 vec0 = vsetq_lane(vgetq_lane(vec0,0) - vgetq_lane(vec1,3),vec0,0); in arm_quaternion2rotation_f32() 124 tmp1 = vgetq_lane(vec1,0) - vgetq_lane(vec1,1); in arm_quaternion2rotation_f32() 125 tmp2 = vgetq_lane(vec1,2) - vgetq_lane(vec1,3); in arm_quaternion2rotation_f32()
|
D | arm_quaternion_norm_f32.c | 64 f32x4_t vec1; in arm_quaternion_norm_f32() local 69 vec1 = vld1q(pInputQuaternions); in arm_quaternion_norm_f32() 70 vec1 = vmulq(vec1,vec1); in arm_quaternion_norm_f32() 71 squaredSum = vecAddAcrossF32Mve(vec1); in arm_quaternion_norm_f32()
|
/third_party/flutter/skia/third_party/externals/libpng/mips/ |
D | filter_msa_intrinsics.c | 664 v8i16 vec0, vec1, vec2; in png_read_filter_row_paeth4_msa() local 696 ILVR_B2_SH(src2, src6, src1, src6, vec0, vec1); in png_read_filter_row_paeth4_msa() 697 HSUB_UB2_SH(vec0, vec1, vec0, vec1); in png_read_filter_row_paeth4_msa() 698 vec2 = vec0 + vec1; in png_read_filter_row_paeth4_msa() 699 ADD_ABS_H3_SH(vec0, vec1, vec2, vec0, vec1, vec2); in png_read_filter_row_paeth4_msa() 700 CMP_AND_SELECT(vec0, vec1, vec2, src1, src2, src6, src10); in png_read_filter_row_paeth4_msa() 701 ILVR_B2_SH(src3, src7, src10, src7, vec0, vec1); in png_read_filter_row_paeth4_msa() 702 HSUB_UB2_SH(vec0, vec1, vec0, vec1); in png_read_filter_row_paeth4_msa() 703 vec2 = vec0 + vec1; in png_read_filter_row_paeth4_msa() 704 ADD_ABS_H3_SH(vec0, vec1, vec2, vec0, vec1, vec2); in png_read_filter_row_paeth4_msa() [all …]
|
/third_party/libpng/mips/ |
D | filter_msa_intrinsics.c | 664 v8i16 vec0, vec1, vec2; in png_read_filter_row_paeth4_msa() local 696 ILVR_B2_SH(src2, src6, src1, src6, vec0, vec1); in png_read_filter_row_paeth4_msa() 697 HSUB_UB2_SH(vec0, vec1, vec0, vec1); in png_read_filter_row_paeth4_msa() 698 vec2 = vec0 + vec1; in png_read_filter_row_paeth4_msa() 699 ADD_ABS_H3_SH(vec0, vec1, vec2, vec0, vec1, vec2); in png_read_filter_row_paeth4_msa() 700 CMP_AND_SELECT(vec0, vec1, vec2, src1, src2, src6, src10); in png_read_filter_row_paeth4_msa() 701 ILVR_B2_SH(src3, src7, src10, src7, vec0, vec1); in png_read_filter_row_paeth4_msa() 702 HSUB_UB2_SH(vec0, vec1, vec0, vec1); in png_read_filter_row_paeth4_msa() 703 vec2 = vec0 + vec1; in png_read_filter_row_paeth4_msa() 704 ADD_ABS_H3_SH(vec0, vec1, vec2, vec0, vec1, vec2); in png_read_filter_row_paeth4_msa() [all …]
|
/third_party/skia/third_party/externals/libpng/mips/ |
D | filter_msa_intrinsics.c | 664 v8i16 vec0, vec1, vec2; in png_read_filter_row_paeth4_msa() local 696 ILVR_B2_SH(src2, src6, src1, src6, vec0, vec1); in png_read_filter_row_paeth4_msa() 697 HSUB_UB2_SH(vec0, vec1, vec0, vec1); in png_read_filter_row_paeth4_msa() 698 vec2 = vec0 + vec1; in png_read_filter_row_paeth4_msa() 699 ADD_ABS_H3_SH(vec0, vec1, vec2, vec0, vec1, vec2); in png_read_filter_row_paeth4_msa() 700 CMP_AND_SELECT(vec0, vec1, vec2, src1, src2, src6, src10); in png_read_filter_row_paeth4_msa() 701 ILVR_B2_SH(src3, src7, src10, src7, vec0, vec1); in png_read_filter_row_paeth4_msa() 702 HSUB_UB2_SH(vec0, vec1, vec0, vec1); in png_read_filter_row_paeth4_msa() 703 vec2 = vec0 + vec1; in png_read_filter_row_paeth4_msa() 704 ADD_ABS_H3_SH(vec0, vec1, vec2, vec0, vec1, vec2); in png_read_filter_row_paeth4_msa() [all …]
|
/third_party/cmsis/CMSIS/DSP/Source/MatrixFunctions/ |
D | arm_mat_inverse_f32.c | 456 f32x4_t vec1, vec2; in arm_mat_inverse_f32() local 461 vec1 = vldrwq_f32(pInT1); in arm_mat_inverse_f32() 463 vec1 = vfmsq_f32(vec1, tmpV, vec2); in arm_mat_inverse_f32() 464 vstrwq_f32(pInT1, vec1); in arm_mat_inverse_f32() 479 f32x4_t vec1, vec2; in arm_mat_inverse_f32() local 482 vec1 = vldrwq_f32(pInT1); in arm_mat_inverse_f32() 484 vec1 = vfmsq_f32(vec1, tmpV, vec2); in arm_mat_inverse_f32() 485 vstrwq_p_f32(pInT1, vec1, p0); in arm_mat_inverse_f32() 492 f32x4_t vec1, vec2; in arm_mat_inverse_f32() local 498 vec1 = vldrwq_f32(pOutT1); in arm_mat_inverse_f32() [all …]
|
D | arm_mat_inverse_f16.c | 438 f16x8_t vec1, vec2; in arm_mat_inverse_f16() local 443 vec1 = vldrhq_f16(pInT1); in arm_mat_inverse_f16() 445 vec1 = vfmsq_f16(vec1, tmpV, vec2); in arm_mat_inverse_f16() 446 vstrhq_f16(pInT1, vec1); in arm_mat_inverse_f16() 461 f16x8_t vec1, vec2; in arm_mat_inverse_f16() local 464 vec1 = vldrhq_f16(pInT1); in arm_mat_inverse_f16() 466 vec1 = vfmsq_f16(vec1, tmpV, vec2); in arm_mat_inverse_f16() 467 vstrhq_p_f16(pInT1, vec1, p0); in arm_mat_inverse_f16() 474 f16x8_t vec1, vec2; in arm_mat_inverse_f16() local 480 vec1 = vldrhq_f16(pOutT1); in arm_mat_inverse_f16() [all …]
|
/third_party/cmsis/CMSIS/DSP/Source/BasicMathFunctions/ |
D | arm_abs_f32.c | 74 f32x4_t vec1; in arm_abs_f32() local 86 vec1 = vld1q(pSrc); in arm_abs_f32() 87 res = vabsq(vec1); in arm_abs_f32() 106 vec1 = vld1q(pSrc); in arm_abs_f32() 107 vstrwq_p(pDst, vabsq(vec1), p0); in arm_abs_f32() 121 f32x4_t vec1; in arm_abs_f32() local 132 vec1 = vld1q_f32(pSrc); in arm_abs_f32() 133 res = vabsq_f32(vec1); in arm_abs_f32()
|
D | arm_abs_f16.c | 74 f16x8_t vec1; in arm_abs_f16() local 86 vec1 = vld1q(pSrc); in arm_abs_f16() 87 res = vabsq(vec1); in arm_abs_f16() 106 vec1 = vld1q(pSrc); in arm_abs_f16() 107 vstrhq_p(pDst, vabsq(vec1), p0); in arm_abs_f16() 122 f16x8_t vec1; in arm_abs_f16() local 133 vec1 = vld1q_f16(pSrc); in arm_abs_f16() 134 res = vabsq_f16(vec1); in arm_abs_f16()
|
D | arm_negate_f32.c | 72 f32x4_t vec1; in arm_negate_f32() local 83 vec1 = vld1q(pSrc); in arm_negate_f32() 84 res = vnegq(vec1); in arm_negate_f32() 101 vec1 = vld1q((float32_t const *) pSrc); in arm_negate_f32() 102 vstrwq_p(pDst, vnegq(vec1), p0); in arm_negate_f32() 116 f32x4_t vec1; in arm_negate_f32() local 127 vec1 = vld1q_f32(pSrc); in arm_negate_f32() 128 res = vnegq_f32(vec1); in arm_negate_f32()
|
D | arm_offset_f32.c | 75 f32x4_t vec1; in arm_offset_f32() local 85 vec1 = vld1q(pSrc); in arm_offset_f32() 86 res = vaddq(vec1,offset); in arm_offset_f32() 103 vec1 = vld1q((float32_t const *) pSrc); in arm_offset_f32() 104 vstrwq_p(pDst, vaddq(vec1, offset), p0); in arm_offset_f32() 120 f32x4_t vec1; in arm_offset_f32() local 131 vec1 = vld1q_f32(pSrc); in arm_offset_f32() 132 res = vaddq_f32(vec1,vdupq_n_f32(offset)); in arm_offset_f32()
|
D | arm_add_f32.c | 73 f32x4_t vec1; in arm_add_f32() local 85 vec1 = vld1q(pSrcA); in arm_add_f32() 87 res = vaddq(vec1, vec2); in arm_add_f32() 106 vec1 = vld1q(pSrcA); in arm_add_f32() 108 vstrwq_p(pDst, vaddq(vec1,vec2), p0); in arm_add_f32() 123 f32x4_t vec1; in arm_add_f32() local 135 vec1 = vld1q_f32(pSrcA); in arm_add_f32() 137 res = vaddq_f32(vec1, vec2); in arm_add_f32()
|
D | arm_mult_f32.c | 73 f32x4_t vec1; in arm_mult_f32() local 84 vec1 = vld1q(pSrcA); in arm_mult_f32() 86 res = vmulq(vec1, vec2); in arm_mult_f32() 104 vec1 = vld1q(pSrcA); in arm_mult_f32() 106 vstrwq_p(pDst, vmulq(vec1,vec2), p0); in arm_mult_f32() 121 f32x4_t vec1; in arm_mult_f32() local 133 vec1 = vld1q_f32(pSrcA); in arm_mult_f32() 135 res = vmulq_f32(vec1, vec2); in arm_mult_f32()
|
D | arm_sub_f32.c | 73 f32x4_t vec1; in arm_sub_f32() local 85 vec1 = vld1q(pSrcA); in arm_sub_f32() 87 res = vsubq(vec1, vec2); in arm_sub_f32() 106 vec1 = vld1q(pSrcA); in arm_sub_f32() 108 vstrwq_p(pDst, vsubq(vec1,vec2), p0); in arm_sub_f32() 123 f32x4_t vec1; in arm_sub_f32() local 135 vec1 = vld1q_f32(pSrcA); in arm_sub_f32() 137 res = vsubq_f32(vec1, vec2); in arm_sub_f32()
|
D | arm_scale_f32.c | 88 f32x4_t vec1; in arm_scale_f32() local 99 vec1 = vld1q(pSrc); in arm_scale_f32() 100 res = vmulq(vec1,scale); in arm_scale_f32() 117 vec1 = vld1q((float32_t const *) pSrc); in arm_scale_f32() 118 vstrwq_p(pDst, vmulq(vec1, scale), p0); in arm_scale_f32() 133 f32x4_t vec1; in arm_scale_f32() local 144 vec1 = vld1q_f32(pSrc); in arm_scale_f32() 145 res = vmulq_f32(vec1, vdupq_n_f32(scale)); in arm_scale_f32()
|
/third_party/glslang/Test/baseResults/ |
D | hlsl.matpack-1.frag.out | 11 0:24 vec1: direct index for structure ( temp 4-component vector of float) 12 …umn_major) temp 4X4 matrix of float mat2, temp 4-component vector of float vec1, temp float foo}) 13 …vec1, temp float foo} g_MyBuffer1, layout( row_major std140) uniform structure{layout( column_maj… 19 …umn_major) temp 4X4 matrix of float mat2, temp 4-component vector of float vec1, temp float foo}) 20 …vec1, temp float foo} g_MyBuffer1, layout( row_major std140) uniform structure{layout( column_maj… 26 0:25 vec1: direct index for structure ( temp 4-component vector of float) 27 …cture{layout( column_major) temp 4X4 matrix of float mat1, temp 4-component vector of float vec1}) 28 …vec1, temp float foo} g_MyBuffer1, layout( row_major std140) uniform structure{layout( column_maj… 34 …cture{layout( column_major) temp 4X4 matrix of float mat1, temp 4-component vector of float vec1}) 35 …vec1, temp float foo} g_MyBuffer1, layout( row_major std140) uniform structure{layout( column_maj… [all …]
|
/third_party/boost/libs/variant/test/ |
D | recursive_variant_test.cpp | 102 std::vector<var1_t> vec1; in test_recursive_variant() local 103 vec1.push_back(3); in test_recursive_variant() 104 vec1.push_back(5); in test_recursive_variant() 105 vec1.push_back(vec1); in test_recursive_variant() 106 vec1.push_back(7); in test_recursive_variant() 108 var1_t var1(vec1); in test_recursive_variant() 114 std::vector<var1_t> vec1_copy = vec1; in test_recursive_variant() 202 vec5.push_back(vec1); in test_recursive_variant() 241 std::vector<var1_t> vec1; in test_recursive_variant_over() local 242 vec1.push_back(3); in test_recursive_variant_over() [all …]
|
/third_party/ffmpeg/libavcodec/ppc/ |
D | audiodsp.c | 42 register vec_s16 vec1; in scalarproduct_int16_altivec() local 47 vec1 = vec_unaligned_load(v1); in scalarproduct_int16_altivec() 48 t = vec_msum(vec1, vec_ld(0, v2), zero_s32v); in scalarproduct_int16_altivec() 67 register vec_s16 vec1; in scalarproduct_int16_vsx() local 72 vec1 = vec_vsx_ld(0, v1); in scalarproduct_int16_vsx() 73 t = vec_msum(vec1, vec_ld(0, v2), zero_s32v); in scalarproduct_int16_vsx()
|
/third_party/mesa3d/src/intel/compiler/ |
D | brw_clip_unfilled.c | 106 vec1(brw_null_reg()), in cull_direction() 143 vec1(brw_null_reg()), in copy_bfc() 204 vec1(brw_null_reg()), in compute_offset() 209 brw_SEL(p, vec1(off), in compute_offset() 213 brw_MUL(p, vec1(off), vec1(off), brw_imm_f(c->key.offset_factor)); in compute_offset() 214 brw_ADD(p, vec1(off), vec1(off), brw_imm_f(c->key.offset_units)); in compute_offset() 217 vec1(brw_null_reg()), in compute_offset() 219 vec1(off), in compute_offset() 221 brw_SEL(p, vec1(off), vec1(off), brw_imm_f(c->key.offset_clamp)); in compute_offset() 233 vec1(brw_null_reg()), in merge_edgeflags() [all …]
|
/third_party/skia/third_party/externals/spirv-tools/test/util/ |
D | small_vector_test.cpp | 60 SmallVector<uint32_t, 6> vec1 = {0, 1, 2, 3}; in TEST() local 61 SmallVector<uint32_t, 6> vec2(vec1); in TEST() 70 EXPECT_EQ(vec1, vec2); in TEST() 74 SmallVector<uint32_t, 2> vec1 = {0, 1, 2, 3}; in TEST() local 75 SmallVector<uint32_t, 2> vec2(vec1); in TEST() 84 EXPECT_EQ(vec1, vec2); in TEST() 88 std::vector<uint32_t> vec1 = {0, 1, 2, 3}; in TEST() local 89 SmallVector<uint32_t, 6> vec2(vec1); in TEST() 98 EXPECT_EQ(vec1, vec2); in TEST() 102 std::vector<uint32_t> vec1 = {0, 1, 2, 3}; in TEST() local [all …]
|
/third_party/skia/third_party/externals/swiftshader/third_party/SPIRV-Tools/test/util/ |
D | small_vector_test.cpp | 60 SmallVector<uint32_t, 6> vec1 = {0, 1, 2, 3}; in TEST() local 61 SmallVector<uint32_t, 6> vec2(vec1); in TEST() 70 EXPECT_EQ(vec1, vec2); in TEST() 74 SmallVector<uint32_t, 2> vec1 = {0, 1, 2, 3}; in TEST() local 75 SmallVector<uint32_t, 2> vec2(vec1); in TEST() 84 EXPECT_EQ(vec1, vec2); in TEST() 88 std::vector<uint32_t> vec1 = {0, 1, 2, 3}; in TEST() local 89 SmallVector<uint32_t, 6> vec2(vec1); in TEST() 98 EXPECT_EQ(vec1, vec2); in TEST() 102 std::vector<uint32_t> vec1 = {0, 1, 2, 3}; in TEST() local [all …]
|