Home
last modified time | relevance | path

Searched refs:StoreUnaligned16 (Results 1 – 13 of 13) sorted by relevance

/external/libgav1/libgav1/src/dsp/x86/
Dintrapred_cfl_sse4.cc76 StoreUnaligned16(dst, res); in CflIntraPredictor_SSE4_1()
81 StoreUnaligned16(dst + 16, res); in CflIntraPredictor_SSE4_1()
200 StoreUnaligned16(luma_ptr, samples0); in CflSubsampler444_8xH_SSE4_1()
213 StoreUnaligned16(luma_ptr, samples1); in CflSubsampler444_8xH_SSE4_1()
223 StoreUnaligned16(luma_ptr, samples1); in CflSubsampler444_8xH_SSE4_1()
227 StoreUnaligned16(luma_ptr, samples1); in CflSubsampler444_8xH_SSE4_1()
246 StoreUnaligned16(luma_ptr, _mm_sub_epi16(samples, averages)); in CflSubsampler444_8xH_SSE4_1()
328 StoreUnaligned16(luma_ptr, samples0); in CflSubsampler444_SSE4_1()
329 StoreUnaligned16(luma_ptr + 8, samples1); in CflSubsampler444_SSE4_1()
355 StoreUnaligned16(luma_ptr + 16, samples2); in CflSubsampler444_SSE4_1()
[all …]
Dintra_edge_sse4.cc74 StoreUnaligned16(dest, result); in ComputeKernel1Store12()
114 StoreUnaligned16(dest, result); in ComputeKernel2Store12()
228 StoreUnaligned16(pixel_buffer - 1, result_lo); in IntraEdgeUpsampler_SSE4_1()
240 StoreUnaligned16(pixel_buffer + 15, result_hi); in IntraEdgeUpsampler_SSE4_1()
Dwarp_sse4.cc98 StoreUnaligned16(intermediate_result_row, sum); in HorizontalFilter()
130 StoreUnaligned16(static_cast<int16_t*>(dst_row), sum); in WriteVerticalFilter()
161 StoreUnaligned16(static_cast<int16_t*>(dst_row), sum); in WriteVerticalFilter()
234 StoreUnaligned16(dst_row, sum); in WarpRegion1()
Dinverse_transform_sse4.cc49 StoreUnaligned16(&dst[i * stride + idx], s[i]); in StoreDst()
50 StoreUnaligned16(&dst[(i + 1) * stride + idx], s[i + 1]); in StoreDst()
51 StoreUnaligned16(&dst[(i + 2) * stride + idx], s[i + 2]); in StoreDst()
52 StoreUnaligned16(&dst[(i + 3) * stride + idx], s[i + 3]); in StoreDst()
261 StoreUnaligned16(dst, xy_shifted); in DctDcOnly()
290 StoreUnaligned16(&dst[i], xy); in DctDcOnlyColumn()
1281 StoreUnaligned16(dst, _mm_packs_epi32(b, b1)); in Adst8DcOnly()
1587 StoreUnaligned16(&dst[i * 8], _mm_packs_epi32(b, b1)); in Adst16DcOnly()
1644 StoreUnaligned16(&dst[i * step], _mm_packs_epi32(b, b_hi)); in Identity4_SSE4_1()
1653 StoreUnaligned16(&dst[i * step], b); in Identity4_SSE4_1()
[all …]
Dconvolve_sse4.cc242 StoreUnaligned16(&dest16[x], v_sum); in FilterHorizontal()
470 StoreUnaligned16(dst16 + x + y * dst_stride, sum); in Filter2DVertical()
530 StoreUnaligned16(dst16, sum); in Filter2DVertical4xH()
830 StoreUnaligned16(dst16 + x + y * dst_stride, results); in FilterVertical()
887 StoreUnaligned16(dst16, results); in FilterVertical4xH()
934 StoreUnaligned16(dst16, results); in FilterVertical4xH()
993 StoreUnaligned16(dst16, results); in FilterVertical4xH()
1064 StoreUnaligned16(dst16, results); in FilterVertical4xH()
1451 StoreUnaligned16(&dest[x], v_dest_lo); in ConvolveCompoundCopy_SSE4()
1452 StoreUnaligned16(&dest[x + 8], v_dest_hi); in ConvolveCompoundCopy_SSE4()
[all …]
Dloop_filter_sse4.cc1097 StoreUnaligned16(dst - 8 + 0 * stride, x0); in Vertical14()
1098 StoreUnaligned16(dst - 8 + 1 * stride, x1); in Vertical14()
1099 StoreUnaligned16(dst - 8 + 2 * stride, x2); in Vertical14()
1100 StoreUnaligned16(dst - 8 + 3 * stride, x3); in Vertical14()
1846 StoreUnaligned16(dst - 4 + 0 * stride, x0); in Vertical8()
1847 StoreUnaligned16(dst - 4 + 1 * stride, x1); in Vertical8()
1848 StoreUnaligned16(dst - 4 + 2 * stride, x2); in Vertical8()
1849 StoreUnaligned16(dst - 4 + 3 * stride, x3); in Vertical8()
2181 StoreUnaligned16(dst - 8 + 0 * stride, x0); in Vertical14()
2182 StoreUnaligned16(dst - 8 + 1 * stride, x1); in Vertical14()
[all …]
Dobmc_sse4.cc177 StoreUnaligned16(pred, _mm_packus_epi16(result_lo, result_hi)); in OverlapBlendFromLeft_SSE4_1()
294 StoreUnaligned16(pred + x, _mm_packus_epi16(result_lo, result_hi)); in OverlapBlendFromTop_SSE4_1()
Daverage_blend_sse4.cc68 StoreUnaligned16(dest + x, _mm_packus_epi16(res0, res1)); in AverageBlendLargeRow()
Dintrapred_sse4.cc241 StoreUnaligned16(dst, dc_dup); in DcStore16xH_SSE4_1()
244 StoreUnaligned16(dst, dc_dup); in DcStore16xH_SSE4_1()
255 StoreUnaligned16(dst, dc_dup); in DcStore32xH_SSE4_1()
256 StoreUnaligned16(dst + 16, dc_dup); in DcStore32xH_SSE4_1()
259 StoreUnaligned16(dst, dc_dup); in DcStore32xH_SSE4_1()
260 StoreUnaligned16(dst + 16, dc_dup); in DcStore32xH_SSE4_1()
271 StoreUnaligned16(dst, dc_dup); in DcStore64xH_SSE4_1()
272 StoreUnaligned16(dst + 16, dc_dup); in DcStore64xH_SSE4_1()
273 StoreUnaligned16(dst + 32, dc_dup); in DcStore64xH_SSE4_1()
274 StoreUnaligned16(dst + 48, dc_dup); in DcStore64xH_SSE4_1()
[all …]
Dcommon_sse4.h207 inline void StoreUnaligned16(void* a, const __m128i v) { in StoreUnaligned16() function
Ddistance_weighted_blend_sse4.cc147 StoreUnaligned16(dst + x, _mm_packus_epi16(res_lo, res_hi)); in DistanceWeightedBlendLarge_SSE4_1()
Dintrapred_smooth_sse4.cc131 StoreUnaligned16(dest, _mm_packus_epi16(pred1, pred2)); in WriteSmoothDirectionalSum16()
/external/libgav1/libgav1/src/utils/
Dentropy_decoder.cc337 inline void StoreUnaligned16(void* a, const __m128i v) { in StoreUnaligned16() function
377 StoreUnaligned16(cdf, cdf_vec); in UpdateCdf7To9()
408 StoreUnaligned16(cdf + 2, cdf_vec); in UpdateCdf11()
419 StoreUnaligned16(cdf + 2, cdf_vec); in UpdateCdf11()
440 StoreUnaligned16(cdf, cdf_vec0); in UpdateCdf13()
449 StoreUnaligned16(cdf + 4, cdf_vec1); in UpdateCdf13()
470 StoreUnaligned16(cdf, cdf_vec0); in UpdateCdf16()
480 StoreUnaligned16(cdf + 8, cdf_vec1); in UpdateCdf16()