/external/libaom/libaom/aom_dsp/x86/ |
D | quantize_x86.h | 113 return _mm_max_epi16(eob0, eob1); in scan_for_eob() 119 eob = _mm_max_epi16(eob, eob_shuffled); in accumulate_eob() 121 eob = _mm_max_epi16(eob, eob_shuffled); in accumulate_eob() 123 eob = _mm_max_epi16(eob, eob_shuffled); in accumulate_eob() 158 temp_mask = _mm_max_epi16(mask0, mask1); in update_mask1() 161 *mask = _mm_max_epi16(temp_mask, *mask); in update_mask1() 194 mask1 = _mm_max_epi16(mask0, mask); in calculate_non_zero_count() 196 mask0 = _mm_max_epi16(mask0, mask1); in calculate_non_zero_count() 198 mask0 = _mm_max_epi16(mask0, mask1); in calculate_non_zero_count()
|
D | avg_intrin_sse2.c | 28 absdiff0 = _mm_max_epi16(diff, negdiff); in aom_minmax_8x8_sse2() 34 absdiff = _mm_max_epi16(diff, negdiff); in aom_minmax_8x8_sse2() 35 maxabsdiff = _mm_max_epi16(absdiff0, absdiff); in aom_minmax_8x8_sse2() 42 absdiff = _mm_max_epi16(diff, negdiff); in aom_minmax_8x8_sse2() 43 maxabsdiff = _mm_max_epi16(maxabsdiff, absdiff); in aom_minmax_8x8_sse2() 50 absdiff = _mm_max_epi16(diff, negdiff); in aom_minmax_8x8_sse2() 51 maxabsdiff = _mm_max_epi16(maxabsdiff, absdiff); in aom_minmax_8x8_sse2() 58 absdiff = _mm_max_epi16(diff, negdiff); in aom_minmax_8x8_sse2() 59 maxabsdiff = _mm_max_epi16(maxabsdiff, absdiff); in aom_minmax_8x8_sse2() 66 absdiff = _mm_max_epi16(diff, negdiff); in aom_minmax_8x8_sse2() [all …]
|
D | highbd_convolve_ssse3.c | 108 res_16bit0 = _mm_max_epi16(res_16bit0, zero); in av1_highbd_convolve_y_sr_ssse3() 112 res_16bit1 = _mm_max_epi16(res_16bit1, zero); in av1_highbd_convolve_y_sr_ssse3() 120 res_a_round0 = _mm_max_epi16(res_a_round0, zero); in av1_highbd_convolve_y_sr_ssse3() 124 res_a_round1 = _mm_max_epi16(res_a_round1, zero); in av1_highbd_convolve_y_sr_ssse3() 132 res_a_round0 = _mm_max_epi16(res_a_round0, zero); in av1_highbd_convolve_y_sr_ssse3() 136 res_a_round1 = _mm_max_epi16(res_a_round1, zero); in av1_highbd_convolve_y_sr_ssse3() 239 res = _mm_max_epi16(res, zero); in av1_highbd_convolve_x_sr_ssse3()
|
D | highbd_loopfilter_sse2.c | 21 *pixel = _mm_max_epi16(*pixel, *min); in pixel_clamp() 101 max = _mm_max_epi16(max, abs_diff16(p[i], p[i - 1])); in highbd_filter_mask_dual() 102 max = _mm_max_epi16(max, abs_diff16(q[i], q[i - 1])); in highbd_filter_mask_dual() 137 max01 = _mm_max_epi16(*abs_p1p0, abs_q1q0); in highbd_hev_filter_mask_x_sse2() 146 max = _mm_max_epi16(max, max01); in highbd_hev_filter_mask_x_sse2() 149 max = _mm_max_epi16(max, abs_diff16(pq[i], pq[i - 1])); in highbd_hev_filter_mask_x_sse2() 151 max = _mm_max_epi16(max, _mm_srli_si128(max, 8)); in highbd_hev_filter_mask_x_sse2() 160 __m128i max = _mm_max_epi16(abs_diff16(pq[start], pq[0]), in flat_mask_internal() 164 max = _mm_max_epi16(max, abs_diff16(pq[i], pq[0])); in flat_mask_internal() 166 max = _mm_max_epi16(max, _mm_srli_si128(max, 8)); in flat_mask_internal() [all …]
|
D | highbd_convolve_sse2.c | 85 resReg23_45 = _mm_max_epi16(resReg23_45, _mm_setzero_si128()); in aom_highbd_filter_block1d4_v4_sse2() 87 resReg34_56 = _mm_max_epi16(resReg34_56, _mm_setzero_si128()); in aom_highbd_filter_block1d4_v4_sse2() 145 srcRegFilt32b1_1 = _mm_max_epi16(srcRegFilt32b1_1, _mm_setzero_si128()); in aom_highbd_filter_block1d4_h4_sse2() 244 resReg23_45 = _mm_max_epi16(resReg23_45, _mm_setzero_si128()); in aom_highbd_filter_block1d8_v4_sse2() 246 resReg34_56 = _mm_max_epi16(resReg34_56, _mm_setzero_si128()); in aom_highbd_filter_block1d8_v4_sse2() 322 srcRegFilt32b1_1 = _mm_max_epi16(srcRegFilt32b1_1, _mm_setzero_si128()); in aom_highbd_filter_block1d8_h4_sse2()
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
D | highbd_loopfilter_sse2.c | 113 flat = _mm_max_epi16(abs_p1p0, abs_q1q0); in vpx_highbd_lpf_horizontal_16_sse2() 122 work = _mm_max_epi16( in vpx_highbd_lpf_horizontal_16_sse2() 125 mask = _mm_max_epi16(work, mask); in vpx_highbd_lpf_horizontal_16_sse2() 126 work = _mm_max_epi16( in vpx_highbd_lpf_horizontal_16_sse2() 129 mask = _mm_max_epi16(work, mask); in vpx_highbd_lpf_horizontal_16_sse2() 130 work = _mm_max_epi16( in vpx_highbd_lpf_horizontal_16_sse2() 133 mask = _mm_max_epi16(work, mask); in vpx_highbd_lpf_horizontal_16_sse2() 186 flat = _mm_max_epi16( in vpx_highbd_lpf_horizontal_16_sse2() 189 work = _mm_max_epi16( in vpx_highbd_lpf_horizontal_16_sse2() 192 flat = _mm_max_epi16(work, flat); in vpx_highbd_lpf_horizontal_16_sse2() [all …]
|
D | avg_intrin_sse2.c | 27 absdiff0 = _mm_max_epi16(diff, negdiff); in vpx_minmax_8x8_sse2() 33 absdiff = _mm_max_epi16(diff, negdiff); in vpx_minmax_8x8_sse2() 34 maxabsdiff = _mm_max_epi16(absdiff0, absdiff); in vpx_minmax_8x8_sse2() 41 absdiff = _mm_max_epi16(diff, negdiff); in vpx_minmax_8x8_sse2() 42 maxabsdiff = _mm_max_epi16(maxabsdiff, absdiff); in vpx_minmax_8x8_sse2() 49 absdiff = _mm_max_epi16(diff, negdiff); in vpx_minmax_8x8_sse2() 50 maxabsdiff = _mm_max_epi16(maxabsdiff, absdiff); in vpx_minmax_8x8_sse2() 57 absdiff = _mm_max_epi16(diff, negdiff); in vpx_minmax_8x8_sse2() 58 maxabsdiff = _mm_max_epi16(maxabsdiff, absdiff); in vpx_minmax_8x8_sse2() 65 absdiff = _mm_max_epi16(diff, negdiff); in vpx_minmax_8x8_sse2() [all …]
|
D | quantize_sse2.h | 82 return _mm_max_epi16(eob0, eob1); in scan_for_eob() 88 eob = _mm_max_epi16(eob, eob_shuffled); in accumulate_eob() 90 eob = _mm_max_epi16(eob, eob_shuffled); in accumulate_eob() 92 eob = _mm_max_epi16(eob, eob_shuffled); in accumulate_eob()
|
D | highbd_idct4x4_add_sse2.c | 107 max_input = _mm_max_epi16(io_short[0], io_short[1]); in vpx_highbd_idct4x4_16_add_sse2() 109 max_input = _mm_max_epi16(max_input, _mm_srli_si128(max_input, 8)); in vpx_highbd_idct4x4_16_add_sse2() 111 max_input = _mm_max_epi16(max_input, _mm_srli_si128(max_input, 4)); in vpx_highbd_idct4x4_16_add_sse2() 113 max_input = _mm_max_epi16(max_input, _mm_srli_si128(max_input, 2)); in vpx_highbd_idct4x4_16_add_sse2()
|
/external/libvpx/libvpx/vp9/encoder/x86/ |
D | vp9_quantize_sse2.c | 110 eob = _mm_max_epi16(eob, eob1); in vp9_quantize_fp_sse2() 185 eob0 = _mm_max_epi16(eob0, eob1); in vp9_quantize_fp_sse2() 186 eob = _mm_max_epi16(eob, eob0); in vp9_quantize_fp_sse2() 195 eob = _mm_max_epi16(eob, eob_shuffled); in vp9_quantize_fp_sse2() 197 eob = _mm_max_epi16(eob, eob_shuffled); in vp9_quantize_fp_sse2() 199 eob = _mm_max_epi16(eob, eob_shuffled); in vp9_quantize_fp_sse2()
|
/external/libaom/libaom/av1/common/x86/ |
D | reconinter_ssse3.c | 48 _mm_max_epi16(x0, _mm_add_epi16(diff, xmask_base)), in av1_build_compound_diffwtd_mask_highbd_ssse3() 66 _mm_max_epi16(x0, _mm_add_epi16(diff, xmask_base)), in av1_build_compound_diffwtd_mask_highbd_ssse3() 86 _mm_max_epi16(x0, _mm_add_epi16(diff, xmask_base)), in av1_build_compound_diffwtd_mask_highbd_ssse3() 104 _mm_max_epi16(x0, _mm_add_epi16(diff, xmask_base)), in av1_build_compound_diffwtd_mask_highbd_ssse3()
|
D | highbd_convolve_2d_ssse3.c | 161 res_16bit0 = _mm_max_epi16(res_16bit0, zero); in av1_highbd_convolve_2d_sr_ssse3() 165 res_16bit1 = _mm_max_epi16(res_16bit1, zero); in av1_highbd_convolve_2d_sr_ssse3() 173 res_a_round0 = _mm_max_epi16(res_a_round0, zero); in av1_highbd_convolve_2d_sr_ssse3() 177 res_a_round1 = _mm_max_epi16(res_a_round1, zero); in av1_highbd_convolve_2d_sr_ssse3() 185 res_a_round0 = _mm_max_epi16(res_a_round0, zero); in av1_highbd_convolve_2d_sr_ssse3() 189 res_a_round1 = _mm_max_epi16(res_a_round1, zero); in av1_highbd_convolve_2d_sr_ssse3()
|
/external/XNNPACK/src/qs8-vaddc/gen/ |
D | minmax-sse41-mul32-ld32-x32.c | 80 vout01234567 = _mm_max_epi16(vout01234567, voutput_min); in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x32() 81 vout89ABCDEF = _mm_max_epi16(vout89ABCDEF, voutput_min); in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x32() 82 voutGHIJKLMN = _mm_max_epi16(voutGHIJKLMN, voutput_min); in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x32() 83 voutOPQRSTUV = _mm_max_epi16(voutOPQRSTUV, voutput_min); in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x32() 113 vout01234567 = _mm_max_epi16(vout01234567, voutput_min); in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x32()
|
D | minmax-xop-mul32-ld32-x32.c | 85 vout01234567 = _mm_max_epi16(vout01234567, voutput_min); in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x32() 86 vout89ABCDEF = _mm_max_epi16(vout89ABCDEF, voutput_min); in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x32() 87 voutGHIJKLMN = _mm_max_epi16(voutGHIJKLMN, voutput_min); in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x32() 88 voutOPQRSTUV = _mm_max_epi16(voutOPQRSTUV, voutput_min); in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x32() 118 vout01234567 = _mm_max_epi16(vout01234567, voutput_min); in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x32()
|
D | minmax-xop-mul32-ld32-x24.c | 76 vout01234567 = _mm_max_epi16(vout01234567, voutput_min); in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x24() 77 vout89ABCDEF = _mm_max_epi16(vout89ABCDEF, voutput_min); in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x24() 78 voutGHIJKLMN = _mm_max_epi16(voutGHIJKLMN, voutput_min); in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x24() 107 vout01234567 = _mm_max_epi16(vout01234567, voutput_min); in xnn_qs8_vaddc_minmax_ukernel__xop_mul32_ld32_x24()
|
D | minmax-sse2-mul16-ld64-x32.c | 99 vout01234567 = _mm_max_epi16(vout01234567, voutput_min); in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x32() 100 vout89ABCDEF = _mm_max_epi16(vout89ABCDEF, voutput_min); in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x32() 101 voutGHIJKLMN = _mm_max_epi16(voutGHIJKLMN, voutput_min); in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x32() 102 voutOPQRSTUV = _mm_max_epi16(voutOPQRSTUV, voutput_min); in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x32() 140 vout01234567 = _mm_max_epi16(vout01234567, voutput_min); in xnn_qs8_vaddc_minmax_ukernel__sse2_mul16_ld64_x32()
|
D | minmax-sse41-mul16-ld64-x32.c | 95 vout01234567 = _mm_max_epi16(vout01234567, voutput_min); in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x32() 96 vout89ABCDEF = _mm_max_epi16(vout89ABCDEF, voutput_min); in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x32() 97 voutGHIJKLMN = _mm_max_epi16(voutGHIJKLMN, voutput_min); in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x32() 98 voutOPQRSTUV = _mm_max_epi16(voutOPQRSTUV, voutput_min); in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x32() 135 vout01234567 = _mm_max_epi16(vout01234567, voutput_min); in xnn_qs8_vaddc_minmax_ukernel__sse41_mul16_ld64_x32()
|
D | minmax-sse41-mul32-ld32-x24.c | 71 vout01234567 = _mm_max_epi16(vout01234567, voutput_min); in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x24() 72 vout89ABCDEF = _mm_max_epi16(vout89ABCDEF, voutput_min); in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x24() 73 voutGHIJKLMN = _mm_max_epi16(voutGHIJKLMN, voutput_min); in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x24() 102 vout01234567 = _mm_max_epi16(vout01234567, voutput_min); in xnn_qs8_vaddc_minmax_ukernel__sse41_mul32_ld32_x24()
|
/external/libaom/libaom/av1/encoder/x86/ |
D | av1_quantize_sse2.c | 130 const __m128i eob2 = _mm_max_epi16(eob0, eob1); in quantize() 131 *eob = _mm_max_epi16(*eob, eob2); in quantize() 182 eob = _mm_max_epi16(eob, eob_shuffled); in av1_quantize_fp_sse2() 184 eob = _mm_max_epi16(eob, eob_shuffled); in av1_quantize_fp_sse2() 186 eob = _mm_max_epi16(eob, eob_shuffled); in av1_quantize_fp_sse2()
|
D | av1_highbd_quantize_sse4.c | 99 *eob = _mm_max_epi16(*eob, iscanIdx); in find_eob() 106 *eob = _mm_max_epi16(*eob, eob_shuffled); in get_accumulated_eob() 108 *eob = _mm_max_epi16(*eob, eob_shuffled); in get_accumulated_eob() 110 *eob = _mm_max_epi16(*eob, eob_shuffled); in get_accumulated_eob()
|
/external/XNNPACK/src/qs8-vadd/gen/ |
D | minmax-sse41-mul32-ld32-x32.c | 96 vout01234567 = _mm_max_epi16(vout01234567, voutput_min); in xnn_qs8_vadd_minmax_ukernel__sse41_mul32_ld32_x32() 97 vout89ABCDEF = _mm_max_epi16(vout89ABCDEF, voutput_min); in xnn_qs8_vadd_minmax_ukernel__sse41_mul32_ld32_x32() 98 voutGHIJKLMN = _mm_max_epi16(voutGHIJKLMN, voutput_min); in xnn_qs8_vadd_minmax_ukernel__sse41_mul32_ld32_x32() 99 voutOPQRSTUV = _mm_max_epi16(voutOPQRSTUV, voutput_min); in xnn_qs8_vadd_minmax_ukernel__sse41_mul32_ld32_x32() 135 vout01234567 = _mm_max_epi16(vout01234567, voutput_min); in xnn_qs8_vadd_minmax_ukernel__sse41_mul32_ld32_x32()
|
D | minmax-xop-mul32-ld32-x32.c | 101 vout01234567 = _mm_max_epi16(vout01234567, voutput_min); in xnn_qs8_vadd_minmax_ukernel__xop_mul32_ld32_x32() 102 vout89ABCDEF = _mm_max_epi16(vout89ABCDEF, voutput_min); in xnn_qs8_vadd_minmax_ukernel__xop_mul32_ld32_x32() 103 voutGHIJKLMN = _mm_max_epi16(voutGHIJKLMN, voutput_min); in xnn_qs8_vadd_minmax_ukernel__xop_mul32_ld32_x32() 104 voutOPQRSTUV = _mm_max_epi16(voutOPQRSTUV, voutput_min); in xnn_qs8_vadd_minmax_ukernel__xop_mul32_ld32_x32() 140 vout01234567 = _mm_max_epi16(vout01234567, voutput_min); in xnn_qs8_vadd_minmax_ukernel__xop_mul32_ld32_x32()
|
D | minmax-sse41-mul32-ld32-x24.c | 83 vout01234567 = _mm_max_epi16(vout01234567, voutput_min); in xnn_qs8_vadd_minmax_ukernel__sse41_mul32_ld32_x24() 84 vout89ABCDEF = _mm_max_epi16(vout89ABCDEF, voutput_min); in xnn_qs8_vadd_minmax_ukernel__sse41_mul32_ld32_x24() 85 voutGHIJKLMN = _mm_max_epi16(voutGHIJKLMN, voutput_min); in xnn_qs8_vadd_minmax_ukernel__sse41_mul32_ld32_x24() 120 vout01234567 = _mm_max_epi16(vout01234567, voutput_min); in xnn_qs8_vadd_minmax_ukernel__sse41_mul32_ld32_x24()
|
/external/libvpx/libvpx/vp8/encoder/x86/ |
D | vp8_quantize_sse2.c | 208 x0 = _mm_max_epi16(x0, x1); in vp8_fast_quantize_b_sse2() 213 x0 = _mm_max_epi16(x0, x1); in vp8_fast_quantize_b_sse2() 218 x0 = _mm_max_epi16(x0, x1); in vp8_fast_quantize_b_sse2() 223 x0 = _mm_max_epi16(x0, x1); in vp8_fast_quantize_b_sse2()
|
/external/XNNPACK/src/qs8-requantization/ |
D | fp32-sse2.c | 76 const __m128i xy_clamped = _mm_max_epi16(_mm_min_epi16(xy_packed, vqmax), vqmin); in xnn_qs8_requantize_fp32__sse2() 77 const __m128i zw_clamped = _mm_max_epi16(_mm_min_epi16(zw_packed, vqmax), vqmin); in xnn_qs8_requantize_fp32__sse2()
|