/external/rust/crates/libz-sys/src/zlib-ng/arch/power/ |
D | adler32_power8.c | 81 vector unsigned int vsum1, vsum2; in adler32_power8() local 96 vsum2 = vec_msum(vbuf, v_mul, v_zeros); in adler32_power8() 101 vs2 = vec_add(vsum2, vs2); in adler32_power8() 109 vs2 = vec_sumsu(vs2, vsum2); in adler32_power8() 131 vsum2 = vec_msum(vbuf, v_mul, v_zeros); in adler32_power8() 136 vs2 = vec_add(vsum2, vs2); in adler32_power8() 144 vs2 = vec_sumsu(vs2, vsum2); in adler32_power8()
|
/external/XNNPACK/src/f32-gavgpool-cw/ |
D | neon-x4.c | 38 float32x4_t vsum2 = vmovq_n_f32(0.0f); in xnn_f32_gavgpool_cw_ukernel__neon_x4() local 49 vsum2 = vaddq_f32(vsum2, vi2); in xnn_f32_gavgpool_cw_ukernel__neon_x4() 67 vsum2 = vaddq_f32(vsum2, vi2); in xnn_f32_gavgpool_cw_ukernel__neon_x4() 75 const float32x4_t vsum23 = vpaddq_f32(vsum2, vsum3); in xnn_f32_gavgpool_cw_ukernel__neon_x4() 80 const float32x4_t vsum23 = vcombine_f32(vadd_f32(vget_low_f32(vsum2), vget_high_f32(vsum2)), in xnn_f32_gavgpool_cw_ukernel__neon_x4()
|
D | wasmsimd-arm-x4.c | 38 v128_t vsum2 = vsum0; in xnn_f32_gavgpool_cw_ukernel__wasmsimd_arm_x4() local 53 vsum2 = wasm_f32x4_add(vsum2, vi2); in xnn_f32_gavgpool_cw_ukernel__wasmsimd_arm_x4() 70 vsum2 = wasm_f32x4_add(vsum2, vi2); in xnn_f32_gavgpool_cw_ukernel__wasmsimd_arm_x4() 77 … v128_t vsum23 = wasm_f32x4_add(wasm_v32x4_shuffle(vsum2, vsum3, 0, 2, 4, 6), wasm_v32x4_shuffle(v… in xnn_f32_gavgpool_cw_ukernel__wasmsimd_arm_x4()
|
D | sse-x4.c | 38 __m128 vsum2 = _mm_setzero_ps(); in xnn_f32_gavgpool_cw_ukernel__sse_x4() local 53 vsum2 = _mm_add_ps(vsum2, vi2); in xnn_f32_gavgpool_cw_ukernel__sse_x4() 70 vsum2 = _mm_add_ps(vsum2, vi2); in xnn_f32_gavgpool_cw_ukernel__sse_x4() 77 const __m128 vsum23 = _mm_add_ps(_mm_unpacklo_ps(vsum2, vsum3), _mm_unpackhi_ps(vsum2, vsum3)); in xnn_f32_gavgpool_cw_ukernel__sse_x4()
|
D | wasmsimd-x86-x4.c | 38 v128_t vsum2 = vsum0; in xnn_f32_gavgpool_cw_ukernel__wasmsimd_x86_x4() local 53 vsum2 = wasm_f32x4_add(vsum2, vi2); in xnn_f32_gavgpool_cw_ukernel__wasmsimd_x86_x4() 70 vsum2 = wasm_f32x4_add(vsum2, vi2); in xnn_f32_gavgpool_cw_ukernel__wasmsimd_x86_x4() 77 … v128_t vsum23 = wasm_f32x4_add(wasm_v32x4_shuffle(vsum2, vsum3, 0, 2, 4, 6), wasm_v32x4_shuffle(v… in xnn_f32_gavgpool_cw_ukernel__wasmsimd_x86_x4()
|
D | scalar-x1.c | 32 float vsum2 = 0.f; in xnn_f32_gavgpool_cw_ukernel__scalar_x1() local 38 vsum2 += i0[2]; in xnn_f32_gavgpool_cw_ukernel__scalar_x1() 50 float vout = ( (vsum0 + vsum1) + (vsum2 + vsum3) ) * vmultiplier; in xnn_f32_gavgpool_cw_ukernel__scalar_x1()
|
/external/llvm/test/CodeGen/X86/ |
D | peephole-multiple-folds.ll | 15 %vsum2 = phi <8 x float> [ %vsum2.next, %loopbody ], [ zeroinitializer, %entry ] 19 …%vsum2.next = tail call <8 x float> @llvm.x86.fma.vfmadd.ps.256(<8 x float> %m2, <8 x float> zeroi… 25 %r = fadd <8 x float> %vsum1.next, %vsum2.next
|
/external/llvm-project/llvm/test/CodeGen/X86/ |
D | peephole-multiple-folds.ll | 15 %vsum2 = phi <8 x float> [ %vsum2.next, %loopbody ], [ zeroinitializer, %entry ] 19 …%vsum2.next = tail call <8 x float> @llvm.x86.fma.vfmadd.ps.256(<8 x float> %m2, <8 x float> zeroi… 25 %r = fadd <8 x float> %vsum1.next, %vsum2.next
|
/external/rust/crates/libz-sys/src/zlib-ng/arch/x86/ |
D | adler32_ssse3.c | 81 __m128i vsum2 = _mm_madd_epi16(v_short_sum2, dot3v); in adler32_ssse3() local 83 vsum2 = _mm_add_epi32(vsum2, vs2); in adler32_ssse3() 84 vs2 = _mm_add_epi32(vsum2, vs1_0); in adler32_ssse3()
|
D | adler32_avx.c | 79 __m256i vsum2 = _mm256_madd_epi16(v_short_sum2, dot3v); in adler32_avx2() local 81 vsum2 = _mm256_add_epi32(vsum2, vs2); in adler32_avx2() 82 vs2 = _mm256_add_epi32(vsum2, vs1_0); in adler32_avx2()
|
/external/arm-optimized-routines/networking/aarch64/ |
D | chksum_simd.c | 80 uint64x2_t vsum2 = { 0, 0 }; in __chksum_aarch64_simd() local 92 vsum2 = vpadalq_u32(vsum2, vtmp2); in __chksum_aarch64_simd() 99 vsum0 = vpadalq_u32(vsum0, vreinterpretq_u32_u64(vsum2)); in __chksum_aarch64_simd()
|
/external/llvm-project/libc/AOR_v20.02/networking/aarch64/ |
D | chksum_simd.c | 81 uint64x2_t vsum2 = { 0, 0 }; in __chksum_aarch64_simd() local 93 vsum2 = vpadalq_u32(vsum2, vtmp2); in __chksum_aarch64_simd() 100 vsum0 = vpadalq_u32(vsum0, vreinterpretq_u32_u64(vsum2)); in __chksum_aarch64_simd()
|
/external/arm-optimized-routines/networking/arm/ |
D | chksum_simd.c | 54 uint64x2_t vsum2 = { 0, 0 }; in __chksum_arm_simd() local 65 vsum2 = vpadalq_u32(vsum2, vtmp2); in __chksum_arm_simd() 72 vsum0 = vpadalq_u32(vsum0, vreinterpretq_u32_u64(vsum2)); in __chksum_arm_simd()
|
/external/llvm-project/libc/AOR_v20.02/networking/arm/ |
D | chksum_simd.c | 55 uint64x2_t vsum2 = { 0, 0 }; in __chksum_arm_simd() local 66 vsum2 = vpadalq_u32(vsum2, vtmp2); in __chksum_arm_simd() 73 vsum0 = vpadalq_u32(vsum0, vreinterpretq_u32_u64(vsum2)); in __chksum_arm_simd()
|
/external/libaom/libaom/av1/encoder/x86/ |
D | temporal_filter_sse2.c | 87 static int32_t xx_mask_and_hadd(__m128i vsum1, __m128i vsum2, int i) { in xx_mask_and_hadd() argument 91 vecb = _mm_and_si128(vsum2, *(__m128i *)sse_bytemask_2x4[i][1]); in xx_mask_and_hadd() 147 __m128i vsum2 = _mm_setzero_si128(); in apply_temporal_filter_planewise() local 152 vsum2 = _mm_add_epi32(vsrc[i][1], vsum2); in apply_temporal_filter_planewise() 173 acc_5x5_sse[row][col + i] = xx_mask_and_hadd(vsum1, vsum2, i); in apply_temporal_filter_planewise()
|
/external/llvm-project/llvm/test/CodeGen/NVPTX/ |
D | match.ll | 73 %vsum2 = add i32 %v3, %v4 74 %vsum3 = add i32 %vsum1, %vsum2 109 %vsum2 = add i64 %v3, %v4 110 %vsum3 = add i64 %vsum1, %vsum2
|