/external/libgav1/libgav1/src/dsp/ |
D | loop_restoration.cc | 278 uint16_t* const* sum3, uint16_t* const* sum5, in BoxSum() argument 300 (*sum5)[x] = sum + source0; in BoxSum() 305 ++sum5; in BoxSum() 365 const uint16_t* const sum5[5], const uint32_t* const square_sum5[5], in BoxFilterPreProcess5() 374 b += sum5[dy][x]; in BoxFilterPreProcess5() 473 uint16_t* const sum5[5], in BoxFilterPass1() 479 BoxFilterPreProcess5<bitdepth>(sum5, square_sum5, width, scale, sgr_buffer, in BoxFilterPass1() 514 uint16_t* const sum3[4], uint16_t* const sum5[5], in BoxFilter() 523 BoxFilterPreProcess5<bitdepth>(sum5, square_sum5, width, scales[0], in BoxFilter() 562 uint16_t *sum3[4], *sum5[5], *ma343[4], *ma444[3], *ma565[2]; in BoxFilterProcess() local [all …]
|
D | common.h | 51 alignas(kMaxAlignment) uint16_t sum5[5 * kSgrStride]; member
|
/external/libgav1/libgav1/src/dsp/arm/ |
D | loop_restoration_neon.cc | 885 const uint16x8_t sum5 = vaddq_u16(sum4, sum); in Sum565() local 886 return vaddw_u8(sum5, in Sum565() 893 const uint32x4_t sum5 = vaddq_u32(sum4, sum); in Sum565W() local 894 return vaddw_u16(sum5, src[1]); in Sum565W() 907 const ptrdiff_t sum_stride, uint16_t* sum3, uint16_t* sum5, in BoxSum() argument 924 vst1q_u16(sum5, row5); in BoxSum() 932 sum5 += 8; in BoxSum() 1147 uint8x16_t s[2][2], uint16_t* const sum5[5], uint32_t* const square_sum5[5], in BoxFilterPreProcess5Lo() 1161 vst1q_u16(sum5[3], s5[3]); in BoxFilterPreProcess5Lo() 1162 vst1q_u16(sum5[4], s5[4]); in BoxFilterPreProcess5Lo() [all …]
|
/external/libgav1/libgav1/src/dsp/x86/ |
D | loop_restoration_sse4.cc | 1018 const __m128i sum5 = _mm_add_epi16(sum4, sum); in Sum565Lo() local 1019 return VaddwLo8(sum5, src[1]); in Sum565Lo() 1025 const __m128i sum5 = _mm_add_epi16(sum4, sum); in Sum565Hi() local 1026 return VaddwHi8(sum5, src[1]); in Sum565Hi() 1032 const __m128i sum5 = _mm_add_epi32(sum4, sum); in Sum565WLo() local 1033 return VaddwLo16(sum5, src[1]); in Sum565WLo() 1039 const __m128i sum5 = _mm_add_epi32(sum4, sum); in Sum565WHi() local 1040 return VaddwHi16(sum5, src[1]); in Sum565WHi() 1052 const ptrdiff_t sum_width, uint16_t* sum3, uint16_t* sum5, in BoxSum() argument 1070 StoreAligned32U16(sum5, row5); in BoxSum() [all …]
|
D | loop_restoration_10bit_sse4.cc | 868 const __m128i sum5 = _mm_add_epi16(sum4, sum); in Sum565Lo() local 869 return VaddwLo8(sum5, src[1]); in Sum565Lo() 875 const __m128i sum5 = _mm_add_epi16(sum4, sum); in Sum565Hi() local 876 return VaddwHi8(sum5, src[1]); in Sum565Hi() 882 const __m128i sum5 = _mm_add_epi32(sum4, sum); in Sum565() local 883 return _mm_add_epi32(sum5, src[1]); in Sum565() 896 const ptrdiff_t sum_width, uint16_t* sum3, uint16_t* sum5, in BoxSum() argument 918 StoreAligned32U16(sum5, row5); in BoxSum() 931 sum5 += 16; in BoxSum() 937 sum5 += sum_stride - sum_width; in BoxSum() [all …]
|
D | loop_restoration_avx2.cc | 1210 const __m256i sum5 = _mm256_add_epi16(sum4, sum); in Sum565Lo() local 1211 return VaddwLo8(sum5, src[1]); in Sum565Lo() 1217 const __m256i sum5 = _mm256_add_epi16(sum4, sum); in Sum565Hi() local 1218 return VaddwHi8(sum5, src[1]); in Sum565Hi() 1224 const __m256i sum5 = _mm256_add_epi32(sum4, sum); in Sum565WLo() local 1225 return VaddwLo16(sum5, src[1]); in Sum565WLo() 1231 const __m256i sum5 = _mm256_add_epi32(sum4, sum); in Sum565WHi() local 1232 return VaddwHi16(sum5, src[1]); in Sum565WHi() 1244 const ptrdiff_t sum_width, uint16_t* sum3, uint16_t* sum5, in BoxSum() argument 1256 StoreAligned16(sum5, s5); in BoxSum() [all …]
|
D | loop_restoration_10bit_avx2.cc | 1108 const __m256i sum5 = _mm256_add_epi16(sum4, sum); in Sum565Lo() local 1109 return VaddwLo8(sum5, src[1]); in Sum565Lo() 1115 const __m256i sum5 = _mm256_add_epi16(sum4, sum); in Sum565Hi() local 1116 return VaddwHi8(sum5, src[1]); in Sum565Hi() 1122 const __m256i sum5 = _mm256_add_epi32(sum4, sum); in Sum565() local 1123 return _mm256_add_epi32(sum5, src[1]); in Sum565() 1136 const ptrdiff_t sum_width, uint16_t* sum3, uint16_t* sum5, in BoxSum() argument 1152 StoreAligned16(sum5, s5); in BoxSum() 1158 sum5 += 8; in BoxSum() 1181 StoreAligned64(sum5, row5); in BoxSum() [all …]
|
/external/llvm/test/Analysis/BasicAA/ |
D | pure-const-dce.ll | 19 ; CHECK-NEXT: %sum5 = add i32 %sum4, %tmp5 20 ; CHECK-NEXT: %sum6 = add i32 %sum5, %tmp7 41 %sum5 = add i32 %sum4, %tmp6 ; <i32> [#uses=1] 42 %sum6 = add i32 %sum5, %tmp7 ; <i32> [#uses=1]
|
D | featuretest.ll | 114 %sum5.cast = zext i5 %j to i64 ; <i64> [#uses=1] 115 %P1 = getelementptr i16, i16* %row2col, i64 %sum5.cast
|
/external/llvm-project/llvm/test/Analysis/BasicAA/ |
D | pure-const-dce.ll | 19 ; CHECK-NEXT: %sum5 = add i32 %sum4, %tmp5 20 ; CHECK-NEXT: %sum6 = add i32 %sum5, %tmp7 41 %sum5 = add i32 %sum4, %tmp6 ; <i32> [#uses=1] 42 %sum6 = add i32 %sum5, %tmp7 ; <i32> [#uses=1]
|
D | featuretest.ll | 180 %sum5.cast = zext i5 %j to i64 ; <i64> [#uses=1] 181 %P1 = getelementptr i16, i16* %row2col, i64 %sum5.cast
|
/external/llvm-project/llvm/test/Transforms/Coroutines/ |
D | coro-retcon-resume-values2.ll | 15 %sum5 = call i32 @add(i32 %sum4, i32 %value2) 16 call void @print(i32 %sum5) 80 ; CHECK-NEXT: %sum5 = call i32 @add(i32 %sum4, i32 %1) 81 ; CHECK-NEXT: call void @print(i32 %sum5)
|
/external/llvm-project/llvm/test/Transforms/LoopVectorize/X86/ |
D | mul_slm_16bit.ll | 68 %sum5 = add i32 %sum4, %mul7 70 %add4 = add i32 %add, %sum5 137 %sum5 = add i32 %sum4, %mul7 139 %add4 = add i32 %add, %sum5
|
/external/llvm-project/llvm/test/CodeGen/X86/ |
D | avx2-nontemporal.ll | 91 %sum5 = add i32 %sum4, %v5 92 %sum6 = add i32 %sum5, %v6 93 ret i32 %sum5
|
D | avx512-nontemporal.ll | 69 %sum5 = add i32 %sum4, %v5 70 %sum6 = add i32 %sum5, %v6
|
D | nontemporal.ll | 180 %sum5 = add i32 %sum4, %v5 181 %sum6 = add i32 %sum5, %v6
|
/external/compiler-rt/lib/builtins/ |
D | fp_lib.h | 183 const __uint128_t sum5 = (__uint128_t)product12 + in wideMultiply() local 200 (sum5 << 32) + in wideMultiply()
|
/external/llvm-project/compiler-rt/lib/builtins/ |
D | fp_lib.h | 185 const __uint128_t sum5 = (__uint128_t)product12 + (__uint128_t)product21; in wideMultiply() local 194 (sum5 << 32) + (sum6 << 64); in wideMultiply()
|
/external/llvm/test/CodeGen/X86/ |
D | optimize-max-0.ll | 174 %.sum5 = add i32 %47, %.sum3 175 %48 = getelementptr i8, i8* %j, i32 %.sum5 399 %.sum5 = add i32 %47, %.sum3 400 %48 = getelementptr i8, i8* %j, i32 %.sum5
|
/external/llvm-project/llvm/test/Analysis/Delinearization/ |
D | gcd_multiply_expr.ll | 95 %arrayidx.sum5 = add i32 %add.i, %inc.3.i 96 %arrayidx.4.i = getelementptr inbounds i8, i8* %3, i32 %arrayidx.sum5
|
/external/llvm/test/Analysis/Delinearization/ |
D | gcd_multiply_expr.ll | 94 %arrayidx.sum5 = add i32 %add.i, %inc.3.i 95 %arrayidx.4.i = getelementptr inbounds i8, i8* %3, i32 %arrayidx.sum5
|
/external/llvm-project/polly/test/ScopInfo/ |
D | multiple-types-non-power-of-two.ll | 147 %sum5 = add i8 %sum4, %load.i64.val.trunc 148 %sum6 = add i8 %sum5, %load.i120.val.trunc
|
/external/llvm-project/llvm/test/Analysis/ScalarEvolution/ |
D | avoid-smax-1.ll | 174 %.sum5 = add i32 %55, %.sum3 ; <i32> [#uses=1] 175 %56 = getelementptr i8, i8* %j, i32 %.sum5 ; <i8*> [#uses=1]
|
/external/llvm/test/Analysis/ScalarEvolution/ |
D | avoid-smax-1.ll | 173 %.sum5 = add i32 %55, %.sum3 ; <i32> [#uses=1] 174 %56 = getelementptr i8, i8* %j, i32 %.sum5 ; <i8*> [#uses=1]
|
/external/libaom/libaom/av1/common/arm/ |
D | jnt_convolve_neon.c | 207 uint32x4_t sum4, sum5, sum6, sum7; in compute_avg_8x4() local 233 sum5 = vmull_n_u16(vget_low_u16(res3), fwd_offset); in compute_avg_8x4() 234 sum5 = vmlal_n_u16(sum5, vget_low_u16(d3), bck_offset); in compute_avg_8x4() 236 sum5 = vshrq_n_u32(sum5, DIST_PRECISION_BITS); in compute_avg_8x4() 250 dst5 = vsubq_s32(vreinterpretq_s32_u32(sum5), sub_const_vec); in compute_avg_8x4()
|