Home
last modified time | relevance | path

Searched refs:_mm_loadl_epi64 (Results 1 – 25 of 165) sorted by relevance

1234567

/external/libmpeg2/common/x86/
Dimpeg2_inter_pred_sse42_intr.c143 src_r0 = _mm_loadl_epi64((__m128i *)src); in impeg2_copy_mb_sse42()
144 src_r1 = _mm_loadl_epi64((__m128i *)(src + src_wd)); in impeg2_copy_mb_sse42()
145 src_r2 = _mm_loadl_epi64((__m128i *)(src + 2 * src_wd)); in impeg2_copy_mb_sse42()
146 src_r3 = _mm_loadl_epi64((__m128i *)(src + 3 * src_wd)); in impeg2_copy_mb_sse42()
157 src_r0 = _mm_loadl_epi64((__m128i *)src); in impeg2_copy_mb_sse42()
158 src_r1 = _mm_loadl_epi64((__m128i *)(src + src_wd)); in impeg2_copy_mb_sse42()
159 src_r2 = _mm_loadl_epi64((__m128i *)(src + 2 * src_wd)); in impeg2_copy_mb_sse42()
160 src_r3 = _mm_loadl_epi64((__m128i *)(src + 3 * src_wd)); in impeg2_copy_mb_sse42()
173 src_r0 = _mm_loadl_epi64((__m128i *)src); in impeg2_copy_mb_sse42()
174 src_r1 = _mm_loadl_epi64((__m128i *)(src + src_wd)); in impeg2_copy_mb_sse42()
[all …]
/external/libhevc/common/x86/
Dihevc_inter_pred_filters_sse42_intr.c130 src0_16x8b = _mm_loadl_epi64((__m128i *)(pu1_src)); /* row =0 */ in ihevc_inter_pred_luma_copy_w16out_sse42()
131 src1_16x8b = _mm_loadl_epi64((__m128i *)(pu1_src + 1 * src_strd)); /* row =1 */ in ihevc_inter_pred_luma_copy_w16out_sse42()
132 src2_16x8b = _mm_loadl_epi64((__m128i *)(pu1_src + 2 * src_strd)); /* row =2 */ in ihevc_inter_pred_luma_copy_w16out_sse42()
133 src3_16x8b = _mm_loadl_epi64((__m128i *)(pu1_src + 3 * src_strd)); /* row =3 */ in ihevc_inter_pred_luma_copy_w16out_sse42()
166 src0_16x8b = _mm_loadl_epi64((__m128i *)(pu1_src)); /* row =0 */ in ihevc_inter_pred_luma_copy_w16out_sse42()
167 src1_16x8b = _mm_loadl_epi64((__m128i *)(pu1_src + 1 * src_strd)); /* row =1 */ in ihevc_inter_pred_luma_copy_w16out_sse42()
168 src2_16x8b = _mm_loadl_epi64((__m128i *)(pu1_src + 2 * src_strd)); /* row =2 */ in ihevc_inter_pred_luma_copy_w16out_sse42()
169 src3_16x8b = _mm_loadl_epi64((__m128i *)(pu1_src + 3 * src_strd)); /* row =3 */ in ihevc_inter_pred_luma_copy_w16out_sse42()
288 src0_16x8b = _mm_loadl_epi64((__m128i *)(pu1_src)); /* row =0 */ in ihevc_inter_pred_chroma_copy_sse42()
289 src1_16x8b = _mm_loadl_epi64((__m128i *)(pu1_src + 1 * src_strd)); /* row =1 */ in ihevc_inter_pred_chroma_copy_sse42()
[all …]
Dihevc_inter_pred_filters_ssse3_intr.c168 src0_16x8b = _mm_loadl_epi64((__m128i *)(pu1_src)); /* row =0 */ in ihevc_inter_pred_luma_copy_ssse3()
169 src1_16x8b = _mm_loadl_epi64((__m128i *)(pu1_src + 1 * src_strd)); /* row =1 */ in ihevc_inter_pred_luma_copy_ssse3()
170 src2_16x8b = _mm_loadl_epi64((__m128i *)(pu1_src + 2 * src_strd)); /* row =2 */ in ihevc_inter_pred_luma_copy_ssse3()
171 src3_16x8b = _mm_loadl_epi64((__m128i *)(pu1_src + 3 * src_strd)); /* row =3 */ in ihevc_inter_pred_luma_copy_ssse3()
195 src0_16x8b = _mm_loadl_epi64((__m128i *)(pu1_src)); /* row =0 */ in ihevc_inter_pred_luma_copy_ssse3()
196 src1_16x8b = _mm_loadl_epi64((__m128i *)(pu1_src + 1 * src_strd)); /* row =1 */ in ihevc_inter_pred_luma_copy_ssse3()
197 src2_16x8b = _mm_loadl_epi64((__m128i *)(pu1_src + 2 * src_strd)); /* row =2 */ in ihevc_inter_pred_luma_copy_ssse3()
198 src3_16x8b = _mm_loadl_epi64((__m128i *)(pu1_src + 3 * src_strd)); /* row =3 */ in ihevc_inter_pred_luma_copy_ssse3()
292 src_temp1_16x8b = _mm_loadl_epi64((__m128i *)pi1_coeff); in ihevc_inter_pred_luma_horz_ssse3()
453 res_temp7_8x16b = _mm_loadl_epi64((__m128i *)(pu1_dst + offset)); in ihevc_inter_pred_luma_horz_ssse3()
[all …]
Dihevc_deblk_ssse3_intr.c144 src_row0_8x16b = _mm_loadl_epi64((__m128i *)(pu1_src - 4)); in ihevc_deblk_luma_vert_ssse3()
145 src_row3_8x16b = _mm_loadl_epi64((__m128i *)((pu1_src - 4) + 3 * src_strd)); in ihevc_deblk_luma_vert_ssse3()
241 src_row1_8x16b = _mm_loadl_epi64((__m128i *)((pu1_src - 4) + src_strd)); in ihevc_deblk_luma_vert_ssse3()
242 src_row2_8x16b = _mm_loadl_epi64((__m128i *)((pu1_src - 4) + 2 * src_strd)); in ihevc_deblk_luma_vert_ssse3()
295 temp_str0_16x8b = _mm_loadl_epi64((__m128i *)(&mask)); in ihevc_deblk_luma_vert_ssse3()
296 const2tc_8x16b = _mm_loadl_epi64((__m128i *)(&tc2)); in ihevc_deblk_luma_vert_ssse3()
429 mask_pq_8x16b = _mm_loadl_epi64((__m128i *)(&mask1)); in ihevc_deblk_luma_vert_ssse3()
459 tmp2_const_8x16b = _mm_loadl_epi64((__m128i *)(shuffle0)); in ihevc_deblk_luma_vert_ssse3()
584 src_q0_8x16b = _mm_loadl_epi64((__m128i *)(pu1_src)); in ihevc_deblk_luma_horz_ssse3()
585 src_q1_8x16b = _mm_loadl_epi64((__m128i *)(pu1_src + src_strd)); in ihevc_deblk_luma_horz_ssse3()
[all …]
/external/XNNPACK/src/q8-avgpool/
Dmp9p8q-sse2.c36 const __m128i vright_shift = _mm_loadl_epi64((const __m128i*) params->sse2.right_shift); in xnn_q8_avgpool_ukernel_mp9p8q__sse2()
52 const __m128i vi0 = _mm_loadl_epi64((const __m128i*) i0); i0 += 8; in xnn_q8_avgpool_ukernel_mp9p8q__sse2()
53 const __m128i vi1 = _mm_loadl_epi64((const __m128i*) i1); i1 += 8; in xnn_q8_avgpool_ukernel_mp9p8q__sse2()
54 const __m128i vi2 = _mm_loadl_epi64((const __m128i*) i2); i2 += 8; in xnn_q8_avgpool_ukernel_mp9p8q__sse2()
55 const __m128i vi3 = _mm_loadl_epi64((const __m128i*) i3); i3 += 8; in xnn_q8_avgpool_ukernel_mp9p8q__sse2()
56 const __m128i vi4 = _mm_loadl_epi64((const __m128i*) i4); i4 += 8; in xnn_q8_avgpool_ukernel_mp9p8q__sse2()
57 const __m128i vi5 = _mm_loadl_epi64((const __m128i*) i5); i5 += 8; in xnn_q8_avgpool_ukernel_mp9p8q__sse2()
58 const __m128i vi6 = _mm_loadl_epi64((const __m128i*) i6); i6 += 8; in xnn_q8_avgpool_ukernel_mp9p8q__sse2()
59 const __m128i vi7 = _mm_loadl_epi64((const __m128i*) i7); i7 += 8; in xnn_q8_avgpool_ukernel_mp9p8q__sse2()
60 const __m128i vi8 = _mm_loadl_epi64((const __m128i*) i8); i8 += 8; in xnn_q8_avgpool_ukernel_mp9p8q__sse2()
[all …]
Dup9-sse2.c36 const __m128i vright_shift = _mm_loadl_epi64((const __m128i*) params->sse2.right_shift); in xnn_q8_avgpool_ukernel_up9__sse2()
76 const __m128i vi0 = _mm_loadl_epi64((const __m128i*) i0); i0 += 8; in xnn_q8_avgpool_ukernel_up9__sse2()
77 const __m128i vi1 = _mm_loadl_epi64((const __m128i*) i1); i1 += 8; in xnn_q8_avgpool_ukernel_up9__sse2()
78 const __m128i vi2 = _mm_loadl_epi64((const __m128i*) i2); i2 += 8; in xnn_q8_avgpool_ukernel_up9__sse2()
79 const __m128i vi3 = _mm_loadl_epi64((const __m128i*) i3); i3 += 8; in xnn_q8_avgpool_ukernel_up9__sse2()
80 const __m128i vi4 = _mm_loadl_epi64((const __m128i*) i4); i4 += 8; in xnn_q8_avgpool_ukernel_up9__sse2()
81 const __m128i vi5 = _mm_loadl_epi64((const __m128i*) i5); i5 += 8; in xnn_q8_avgpool_ukernel_up9__sse2()
82 const __m128i vi6 = _mm_loadl_epi64((const __m128i*) i6); i6 += 8; in xnn_q8_avgpool_ukernel_up9__sse2()
83 const __m128i vi7 = _mm_loadl_epi64((const __m128i*) i7); i7 += 8; in xnn_q8_avgpool_ukernel_up9__sse2()
84 const __m128i vi8 = _mm_loadl_epi64((const __m128i*) i8); i8 += 8; in xnn_q8_avgpool_ukernel_up9__sse2()
[all …]
/external/flac/libFLAC/
Dlpc_intrin_sse41.c70 xmm0 = _mm_loadl_epi64((const __m128i*)(qlp_coeff+0)); // 0 0 q[1] q[0] in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41()
71 xmm1 = _mm_loadl_epi64((const __m128i*)(qlp_coeff+2)); // 0 0 q[3] q[2] in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41()
72 xmm2 = _mm_loadl_epi64((const __m128i*)(qlp_coeff+4)); // 0 0 q[5] q[4] in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41()
73 xmm3 = _mm_loadl_epi64((const __m128i*)(qlp_coeff+6)); // 0 0 q[7] q[6] in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41()
74 xmm4 = _mm_loadl_epi64((const __m128i*)(qlp_coeff+8)); // 0 0 q[9] q[8] in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41()
75 xmm5 = _mm_loadl_epi64((const __m128i*)(qlp_coeff+10)); // 0 0 q[11] q[10] in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41()
88 xmm7 = _mm_loadl_epi64((const __m128i*)(data+i-12)); // 0 0 d[i-11] d[i-12] in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41()
94 xmm6 = _mm_loadl_epi64((const __m128i*)(data+i-10)); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41()
101 xmm6 = _mm_loadl_epi64((const __m128i*)(data+i-8)); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41()
108 xmm6 = _mm_loadl_epi64((const __m128i*)(data+i-6)); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41()
[all …]
Dlpc_intrin_sse2.c423 xmm0 = _mm_loadl_epi64((const __m128i*)(qlp_coeff+0)); // 0 0 q[1] q[0] in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2()
424 xmm1 = _mm_loadl_epi64((const __m128i*)(qlp_coeff+2)); // 0 0 q[3] q[2] in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2()
425 xmm2 = _mm_loadl_epi64((const __m128i*)(qlp_coeff+4)); // 0 0 q[5] q[4] in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2()
426 xmm3 = _mm_loadl_epi64((const __m128i*)(qlp_coeff+6)); // 0 0 q[7] q[6] in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2()
427 xmm4 = _mm_loadl_epi64((const __m128i*)(qlp_coeff+8)); // 0 0 q[9] q[8] in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2()
428 xmm5 = _mm_loadl_epi64((const __m128i*)(qlp_coeff+10)); // 0 0 q[11] q[10] in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2()
441 xmm7 = _mm_loadl_epi64((const __m128i*)(data+i-12)); // 0 0 d[i-11] d[i-12] in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2()
447 xmm6 = _mm_loadl_epi64((const __m128i*)(data+i-10)); in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2()
454 xmm6 = _mm_loadl_epi64((const __m128i*)(data+i-8)); in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2()
461 xmm6 = _mm_loadl_epi64((const __m128i*)(data+i-6)); in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2()
[all …]
/external/XNNPACK/src/q8-gavgpool/
Dmp7p7q-sse2.c44 const __m128i vi0 = _mm_loadl_epi64((const __m128i*) i0); i0 += 8; in xnn_q8_gavgpool_ukernel_mp7p7q__sse2()
45 const __m128i vi1 = _mm_loadl_epi64((const __m128i*) i1); i1 += 8; in xnn_q8_gavgpool_ukernel_mp7p7q__sse2()
46 const __m128i vi2 = _mm_loadl_epi64((const __m128i*) i2); i2 += 8; in xnn_q8_gavgpool_ukernel_mp7p7q__sse2()
47 const __m128i vi3 = _mm_loadl_epi64((const __m128i*) i3); i3 += 8; in xnn_q8_gavgpool_ukernel_mp7p7q__sse2()
48 const __m128i vi4 = _mm_loadl_epi64((const __m128i*) i4); i4 += 8; in xnn_q8_gavgpool_ukernel_mp7p7q__sse2()
49 const __m128i vi5 = _mm_loadl_epi64((const __m128i*) i5); i5 += 8; in xnn_q8_gavgpool_ukernel_mp7p7q__sse2()
50 const __m128i vi6 = _mm_loadl_epi64((const __m128i*) i6); i6 += 8; in xnn_q8_gavgpool_ukernel_mp7p7q__sse2()
86 const __m128i vi0 = _mm_loadl_epi64((const __m128i*) i0); i0 += 8; in xnn_q8_gavgpool_ukernel_mp7p7q__sse2()
87 const __m128i vi1 = _mm_loadl_epi64((const __m128i*) i1); i1 += 8; in xnn_q8_gavgpool_ukernel_mp7p7q__sse2()
88 const __m128i vi2 = _mm_loadl_epi64((const __m128i*) i2); i2 += 8; in xnn_q8_gavgpool_ukernel_mp7p7q__sse2()
[all …]
Dup7-sse2.c59 const __m128i vright_shift = _mm_loadl_epi64((const __m128i*) params->sse2.right_shift); in xnn_q8_gavgpool_ukernel_up7__sse2()
62 const __m128i vi0 = _mm_loadl_epi64((const __m128i*) i0); i0 += 8; in xnn_q8_gavgpool_ukernel_up7__sse2()
63 const __m128i vi1 = _mm_loadl_epi64((const __m128i*) i1); i1 += 8; in xnn_q8_gavgpool_ukernel_up7__sse2()
64 const __m128i vi2 = _mm_loadl_epi64((const __m128i*) i2); i2 += 8; in xnn_q8_gavgpool_ukernel_up7__sse2()
65 const __m128i vi3 = _mm_loadl_epi64((const __m128i*) i3); i3 += 8; in xnn_q8_gavgpool_ukernel_up7__sse2()
66 const __m128i vi4 = _mm_loadl_epi64((const __m128i*) i4); i4 += 8; in xnn_q8_gavgpool_ukernel_up7__sse2()
67 const __m128i vi5 = _mm_loadl_epi64((const __m128i*) i5); i5 += 8; in xnn_q8_gavgpool_ukernel_up7__sse2()
68 const __m128i vi6 = _mm_loadl_epi64((const __m128i*) i6); i6 += 8; in xnn_q8_gavgpool_ukernel_up7__sse2()
131 const __m128i vi0 = _mm_loadl_epi64((const __m128i*) i0); in xnn_q8_gavgpool_ukernel_up7__sse2()
132 const __m128i vi1 = _mm_loadl_epi64((const __m128i*) i1); in xnn_q8_gavgpool_ukernel_up7__sse2()
[all …]
/external/libavc/common/x86/
Dih264_iquant_itrans_recon_dc_ssse3.c136 …predload_r = _mm_loadl_epi64((__m128i *) (&pu1_pred[0])); //p00 p01 p02 p03 0 0 0 0 0 0 0 0 -- all… in ih264_iquant_itrans_recon_4x4_dc_ssse3()
138 …predload_r = _mm_loadl_epi64((__m128i *) (&pu1_pred[pred_strd])); //p10 p11 p12 p13 0 0 0 0 0 0 0 … in ih264_iquant_itrans_recon_4x4_dc_ssse3()
140 …predload_r = _mm_loadl_epi64((__m128i *) (&pu1_pred[2*pred_strd])); //p20 p21 p22 p23 0 0 0 0 0 0 … in ih264_iquant_itrans_recon_4x4_dc_ssse3()
142 …predload_r = _mm_loadl_epi64((__m128i *) (&pu1_pred[3*pred_strd])); //p30 p31 p32 p33 0 0 0 0 0 0 … in ih264_iquant_itrans_recon_4x4_dc_ssse3()
258 …predload_r = _mm_loadl_epi64((__m128i *)(&pu1_pred[0])); //p0 p1 p2 p3 p4 p5 p6 p7 0 0 0 0 0 0 0 0… in ih264_iquant_itrans_recon_8x8_dc_ssse3()
261 …predload_r = _mm_loadl_epi64((__m128i *)(&pu1_pred[pred_strd])); //p0 p1 p2 p3 p4 p5 p6 p7 0 0 0 0… in ih264_iquant_itrans_recon_8x8_dc_ssse3()
264 predload_r = _mm_loadl_epi64( in ih264_iquant_itrans_recon_8x8_dc_ssse3()
268 predload_r = _mm_loadl_epi64( in ih264_iquant_itrans_recon_8x8_dc_ssse3()
272 predload_r = _mm_loadl_epi64( in ih264_iquant_itrans_recon_8x8_dc_ssse3()
276 predload_r = _mm_loadl_epi64( in ih264_iquant_itrans_recon_8x8_dc_ssse3()
[all …]
Dih264_weighted_pred_sse42.c101 y0_0_16x8b = _mm_loadl_epi64((__m128i *)pu1_src1); in ih264_default_weighted_pred_luma_sse42()
102 y0_1_16x8b = _mm_loadl_epi64((__m128i *)(pu1_src1 + src_strd1)); in ih264_default_weighted_pred_luma_sse42()
103 y0_2_16x8b = _mm_loadl_epi64( in ih264_default_weighted_pred_luma_sse42()
105 y0_3_16x8b = _mm_loadl_epi64((__m128i *)(pu1_src1 + src_strd1 * 3)); in ih264_default_weighted_pred_luma_sse42()
107 y1_0_16x8b = _mm_loadl_epi64((__m128i *)pu1_src2); in ih264_default_weighted_pred_luma_sse42()
108 y1_1_16x8b = _mm_loadl_epi64((__m128i *)(pu1_src2 + src_strd2)); in ih264_default_weighted_pred_luma_sse42()
109 y1_2_16x8b = _mm_loadl_epi64( in ih264_default_weighted_pred_luma_sse42()
111 y1_3_16x8b = _mm_loadl_epi64((__m128i *)(pu1_src2 + src_strd2 * 3)); in ih264_default_weighted_pred_luma_sse42()
134 y0_0_16x8b = _mm_loadl_epi64((__m128i *)pu1_src1); in ih264_default_weighted_pred_luma_sse42()
135 y0_1_16x8b = _mm_loadl_epi64((__m128i *)(pu1_src1 + src_strd1)); in ih264_default_weighted_pred_luma_sse42()
[all …]
/external/libaom/libaom/aom_dsp/x86/
Dhighbd_subtract_sse2.c32 u0 = _mm_loadl_epi64((__m128i const *)(src + 0 * src_stride)); in subtract_4x4()
33 u1 = _mm_loadl_epi64((__m128i const *)(src + 1 * src_stride)); in subtract_4x4()
34 u2 = _mm_loadl_epi64((__m128i const *)(src + 2 * src_stride)); in subtract_4x4()
35 u3 = _mm_loadl_epi64((__m128i const *)(src + 3 * src_stride)); in subtract_4x4()
37 v0 = _mm_loadl_epi64((__m128i const *)(pred + 0 * pred_stride)); in subtract_4x4()
38 v1 = _mm_loadl_epi64((__m128i const *)(pred + 1 * pred_stride)); in subtract_4x4()
39 v2 = _mm_loadl_epi64((__m128i const *)(pred + 2 * pred_stride)); in subtract_4x4()
40 v3 = _mm_loadl_epi64((__m128i const *)(pred + 3 * pred_stride)); in subtract_4x4()
64 u0 = _mm_loadl_epi64((__m128i const *)(src + 0 * src_stride)); in subtract_4x8()
65 u1 = _mm_loadl_epi64((__m128i const *)(src + 1 * src_stride)); in subtract_4x8()
[all …]
Davg_intrin_sse2.c24 s0 = _mm_unpacklo_epi8(_mm_loadl_epi64((const __m128i *)(s)), u0); in aom_minmax_8x8_sse2()
25 d0 = _mm_unpacklo_epi8(_mm_loadl_epi64((const __m128i *)(d)), u0); in aom_minmax_8x8_sse2()
30 s0 = _mm_unpacklo_epi8(_mm_loadl_epi64((const __m128i *)(s + p)), u0); in aom_minmax_8x8_sse2()
31 d0 = _mm_unpacklo_epi8(_mm_loadl_epi64((const __m128i *)(d + dp)), u0); in aom_minmax_8x8_sse2()
38 s0 = _mm_unpacklo_epi8(_mm_loadl_epi64((const __m128i *)(s + 2 * p)), u0); in aom_minmax_8x8_sse2()
39 d0 = _mm_unpacklo_epi8(_mm_loadl_epi64((const __m128i *)(d + 2 * dp)), u0); in aom_minmax_8x8_sse2()
46 s0 = _mm_unpacklo_epi8(_mm_loadl_epi64((const __m128i *)(s + 3 * p)), u0); in aom_minmax_8x8_sse2()
47 d0 = _mm_unpacklo_epi8(_mm_loadl_epi64((const __m128i *)(d + 3 * dp)), u0); in aom_minmax_8x8_sse2()
54 s0 = _mm_unpacklo_epi8(_mm_loadl_epi64((const __m128i *)(s + 4 * p)), u0); in aom_minmax_8x8_sse2()
55 d0 = _mm_unpacklo_epi8(_mm_loadl_epi64((const __m128i *)(d + 4 * dp)), u0); in aom_minmax_8x8_sse2()
[all …]
Dmasked_sad_intrin_ssse3.c151 _mm_loadl_epi64((const __m128i *)src_ptr), in aom_masked_sad8xh_ssse3()
152 _mm_loadl_epi64((const __m128i *)&src_ptr[src_stride])); in aom_masked_sad8xh_ssse3()
153 const __m128i a0 = _mm_loadl_epi64((const __m128i *)a_ptr); in aom_masked_sad8xh_ssse3()
154 const __m128i a1 = _mm_loadl_epi64((const __m128i *)&a_ptr[a_stride]); in aom_masked_sad8xh_ssse3()
155 const __m128i b0 = _mm_loadl_epi64((const __m128i *)b_ptr); in aom_masked_sad8xh_ssse3()
156 const __m128i b1 = _mm_loadl_epi64((const __m128i *)&b_ptr[b_stride]); in aom_masked_sad8xh_ssse3()
158 _mm_unpacklo_epi64(_mm_loadl_epi64((const __m128i *)m_ptr), in aom_masked_sad8xh_ssse3()
159 _mm_loadl_epi64((const __m128i *)&m_ptr[m_stride])); in aom_masked_sad8xh_ssse3()
307 _mm_loadl_epi64((const __m128i *)&m_ptr[x]), _mm_setzero_si128()); in highbd_masked_sad_ssse3()
361 _mm_loadl_epi64((const __m128i *)src_ptr), in aom_highbd_masked_sad4xh_ssse3()
[all …]
Dloopfilter_sse2.c333 __m128i limit = _mm_unpacklo_epi32(_mm_loadl_epi64((const __m128i *)_blimit), in aom_lpf_horizontal_4_sse2()
334 _mm_loadl_epi64((const __m128i *)_limit)); in aom_lpf_horizontal_4_sse2()
336 _mm_unpacklo_epi8(_mm_loadl_epi64((const __m128i *)_thresh), zero); in aom_lpf_horizontal_4_sse2()
361 __m128i limit = _mm_unpacklo_epi32(_mm_loadl_epi64((const __m128i *)_blimit), in aom_lpf_vertical_4_sse2()
362 _mm_loadl_epi64((const __m128i *)_limit)); in aom_lpf_vertical_4_sse2()
364 _mm_unpacklo_epi8(_mm_loadl_epi64((const __m128i *)_thresh), zero); in aom_lpf_vertical_4_sse2()
368 x0 = _mm_loadl_epi64((__m128i *)(s - 2 + 0 * p)); in aom_lpf_vertical_4_sse2()
369 x1 = _mm_loadl_epi64((__m128i *)(s - 2 + 1 * p)); in aom_lpf_vertical_4_sse2()
370 x2 = _mm_loadl_epi64((__m128i *)(s - 2 + 2 * p)); in aom_lpf_vertical_4_sse2()
371 x3 = _mm_loadl_epi64((__m128i *)(s - 2 + 3 * p)); in aom_lpf_vertical_4_sse2()
[all …]
/external/XNNPACK/src/q8-dwconv/
Dup8x9-sse2.c46 const __m128i vi0 = _mm_loadl_epi64((const __m128i*) i0); i0 += 8; in xnn_q8_dwconv_ukernel_up8x9__sse2()
48 const __m128i vk0 = _mm_loadl_epi64((const __m128i*) ((uintptr_t) w + 32)); in xnn_q8_dwconv_ukernel_up8x9__sse2()
55 const __m128i vi1 = _mm_loadl_epi64((const __m128i*) i1); i1 += 8; in xnn_q8_dwconv_ukernel_up8x9__sse2()
57 const __m128i vk1 = _mm_loadl_epi64((const __m128i*) ((uintptr_t) w + 40)); in xnn_q8_dwconv_ukernel_up8x9__sse2()
64 const __m128i vi2 = _mm_loadl_epi64((const __m128i*) i2); i2 += 8; in xnn_q8_dwconv_ukernel_up8x9__sse2()
66 const __m128i vk2 = _mm_loadl_epi64((const __m128i*) ((uintptr_t) w + 48)); in xnn_q8_dwconv_ukernel_up8x9__sse2()
73 const __m128i vi3 = _mm_loadl_epi64((const __m128i*) i3); i3 += 8; in xnn_q8_dwconv_ukernel_up8x9__sse2()
75 const __m128i vk3 = _mm_loadl_epi64((const __m128i*) ((uintptr_t) w + 56)); in xnn_q8_dwconv_ukernel_up8x9__sse2()
82 const __m128i vi4 = _mm_loadl_epi64((const __m128i*) i4); i4 += 8; in xnn_q8_dwconv_ukernel_up8x9__sse2()
84 const __m128i vk4 = _mm_loadl_epi64((const __m128i*) ((uintptr_t) w + 64)); in xnn_q8_dwconv_ukernel_up8x9__sse2()
[all …]
/external/libhevc/decoder/x86/
Dihevcd_it_rec_dc_ssse3_intr.c98 m_temp_reg_0 = _mm_loadl_epi64((__m128i *)(pu1_pred)); in ihevcd_itrans_recon_dc_luma_ssse3()
99 m_temp_reg_1 = _mm_loadl_epi64((__m128i *)(pu1_pred + pred_strd)); in ihevcd_itrans_recon_dc_luma_ssse3()
100 m_temp_reg_2 = _mm_loadl_epi64((__m128i *)(pu1_pred + 2 * pred_strd)); in ihevcd_itrans_recon_dc_luma_ssse3()
101 m_temp_reg_3 = _mm_loadl_epi64((__m128i *)(pu1_pred + 3 * pred_strd)); in ihevcd_itrans_recon_dc_luma_ssse3()
141 m_temp_reg_0 = _mm_loadl_epi64((__m128i *)pu1_pred); in ihevcd_itrans_recon_dc_luma_ssse3()
142 m_temp_reg_1 = _mm_loadl_epi64((__m128i *)(pu1_pred + pred_strd)); in ihevcd_itrans_recon_dc_luma_ssse3()
143 m_temp_reg_2 = _mm_loadl_epi64((__m128i *)(pu1_pred + 2 * pred_strd)); in ihevcd_itrans_recon_dc_luma_ssse3()
144 m_temp_reg_3 = _mm_loadl_epi64((__m128i *)(pu1_pred + 3 * pred_strd)); in ihevcd_itrans_recon_dc_luma_ssse3()
244 m_temp_reg_0 = _mm_loadl_epi64((__m128i *)(pu1_pred)); in ihevcd_itrans_recon_dc_chroma_ssse3()
245 m_temp_reg_1 = _mm_loadl_epi64((__m128i *)(pu1_pred + pred_strd)); in ihevcd_itrans_recon_dc_chroma_ssse3()
[all …]
Dihevcd_it_rec_dc_sse42_intr.c97 m_temp_reg_0 = _mm_loadl_epi64((__m128i *)(pu1_pred)); in ihevcd_itrans_recon_dc_luma_sse42()
98 m_temp_reg_1 = _mm_loadl_epi64((__m128i *)(pu1_pred + pred_strd)); in ihevcd_itrans_recon_dc_luma_sse42()
99 m_temp_reg_2 = _mm_loadl_epi64((__m128i *)(pu1_pred + 2 * pred_strd)); in ihevcd_itrans_recon_dc_luma_sse42()
100 m_temp_reg_3 = _mm_loadl_epi64((__m128i *)(pu1_pred + 3 * pred_strd)); in ihevcd_itrans_recon_dc_luma_sse42()
140 m_temp_reg_0 = _mm_loadl_epi64((__m128i *)pu1_pred); in ihevcd_itrans_recon_dc_luma_sse42()
141 m_temp_reg_1 = _mm_loadl_epi64((__m128i *)(pu1_pred + pred_strd)); in ihevcd_itrans_recon_dc_luma_sse42()
142 m_temp_reg_2 = _mm_loadl_epi64((__m128i *)(pu1_pred + 2 * pred_strd)); in ihevcd_itrans_recon_dc_luma_sse42()
143 m_temp_reg_3 = _mm_loadl_epi64((__m128i *)(pu1_pred + 3 * pred_strd)); in ihevcd_itrans_recon_dc_luma_sse42()
244 m_temp_reg_0 = _mm_loadl_epi64((__m128i *)(pu1_pred)); in ihevcd_itrans_recon_dc_chroma_sse42()
245 m_temp_reg_1 = _mm_loadl_epi64((__m128i *)(pu1_pred + pred_strd)); in ihevcd_itrans_recon_dc_chroma_sse42()
[all …]
/external/libvpx/libvpx/vpx_dsp/x86/
Davg_intrin_sse2.c23 s0 = _mm_unpacklo_epi8(_mm_loadl_epi64((const __m128i *)(s)), u0); in vpx_minmax_8x8_sse2()
24 d0 = _mm_unpacklo_epi8(_mm_loadl_epi64((const __m128i *)(d)), u0); in vpx_minmax_8x8_sse2()
29 s0 = _mm_unpacklo_epi8(_mm_loadl_epi64((const __m128i *)(s + p)), u0); in vpx_minmax_8x8_sse2()
30 d0 = _mm_unpacklo_epi8(_mm_loadl_epi64((const __m128i *)(d + dp)), u0); in vpx_minmax_8x8_sse2()
37 s0 = _mm_unpacklo_epi8(_mm_loadl_epi64((const __m128i *)(s + 2 * p)), u0); in vpx_minmax_8x8_sse2()
38 d0 = _mm_unpacklo_epi8(_mm_loadl_epi64((const __m128i *)(d + 2 * dp)), u0); in vpx_minmax_8x8_sse2()
45 s0 = _mm_unpacklo_epi8(_mm_loadl_epi64((const __m128i *)(s + 3 * p)), u0); in vpx_minmax_8x8_sse2()
46 d0 = _mm_unpacklo_epi8(_mm_loadl_epi64((const __m128i *)(d + 3 * dp)), u0); in vpx_minmax_8x8_sse2()
53 s0 = _mm_unpacklo_epi8(_mm_loadl_epi64((const __m128i *)(s + 4 * p)), u0); in vpx_minmax_8x8_sse2()
54 d0 = _mm_unpacklo_epi8(_mm_loadl_epi64((const __m128i *)(d + 4 * dp)), u0); in vpx_minmax_8x8_sse2()
[all …]
Dloopfilter_sse2.c111 _mm_unpacklo_epi64(_mm_loadl_epi64((const __m128i *)blimit), in vpx_lpf_horizontal_4_sse2()
112 _mm_loadl_epi64((const __m128i *)limit)); in vpx_lpf_horizontal_4_sse2()
114 _mm_unpacklo_epi8(_mm_loadl_epi64((const __m128i *)thresh), zero); in vpx_lpf_horizontal_4_sse2()
119 p3p2 = _mm_unpacklo_epi64(_mm_loadl_epi64((__m128i *)(s - 3 * pitch)), in vpx_lpf_horizontal_4_sse2()
120 _mm_loadl_epi64((__m128i *)(s - 4 * pitch))); in vpx_lpf_horizontal_4_sse2()
121 q1p1 = _mm_unpacklo_epi64(_mm_loadl_epi64((__m128i *)(s - 2 * pitch)), in vpx_lpf_horizontal_4_sse2()
122 _mm_loadl_epi64((__m128i *)(s + 1 * pitch))); in vpx_lpf_horizontal_4_sse2()
123 q0p0 = _mm_unpacklo_epi64(_mm_loadl_epi64((__m128i *)(s - 1 * pitch)), in vpx_lpf_horizontal_4_sse2()
124 _mm_loadl_epi64((__m128i *)(s + 0 * pitch))); in vpx_lpf_horizontal_4_sse2()
125 q3q2 = _mm_unpacklo_epi64(_mm_loadl_epi64((__m128i *)(s + 2 * pitch)), in vpx_lpf_horizontal_4_sse2()
[all …]
/external/webp/src/dsp/
Denc_sse2.c57 in0 = _mm_loadl_epi64((const __m128i*)&in[0]); in ITransform_SSE2()
58 in1 = _mm_loadl_epi64((const __m128i*)&in[4]); in ITransform_SSE2()
59 in2 = _mm_loadl_epi64((const __m128i*)&in[8]); in ITransform_SSE2()
60 in3 = _mm_loadl_epi64((const __m128i*)&in[12]); in ITransform_SSE2()
66 const __m128i inB0 = _mm_loadl_epi64((const __m128i*)&in[16]); in ITransform_SSE2()
67 const __m128i inB1 = _mm_loadl_epi64((const __m128i*)&in[20]); in ITransform_SSE2()
68 const __m128i inB2 = _mm_loadl_epi64((const __m128i*)&in[24]); in ITransform_SSE2()
69 const __m128i inB3 = _mm_loadl_epi64((const __m128i*)&in[28]); in ITransform_SSE2()
153 ref0 = _mm_loadl_epi64((const __m128i*)&ref[0 * BPS]); in ITransform_SSE2()
154 ref1 = _mm_loadl_epi64((const __m128i*)&ref[1 * BPS]); in ITransform_SSE2()
[all …]
/external/libaom/libaom/av1/common/x86/
Dconvolve_sse2.c171 src6 = _mm_loadl_epi64((__m128i *)(data + 6 * src_stride)); in av1_convolve_y_sr_sse2()
173 _mm_loadl_epi64((__m128i *)(data + 0 * src_stride)), in av1_convolve_y_sr_sse2()
174 _mm_loadl_epi64((__m128i *)(data + 1 * src_stride))); in av1_convolve_y_sr_sse2()
176 _mm_loadl_epi64((__m128i *)(data + 1 * src_stride)), in av1_convolve_y_sr_sse2()
177 _mm_loadl_epi64((__m128i *)(data + 2 * src_stride))); in av1_convolve_y_sr_sse2()
179 _mm_loadl_epi64((__m128i *)(data + 2 * src_stride)), in av1_convolve_y_sr_sse2()
180 _mm_loadl_epi64((__m128i *)(data + 3 * src_stride))); in av1_convolve_y_sr_sse2()
182 _mm_loadl_epi64((__m128i *)(data + 3 * src_stride)), in av1_convolve_y_sr_sse2()
183 _mm_loadl_epi64((__m128i *)(data + 4 * src_stride))); in av1_convolve_y_sr_sse2()
185 _mm_loadl_epi64((__m128i *)(data + 4 * src_stride)), in av1_convolve_y_sr_sse2()
[all …]
/external/XNNPACK/src/q8-gemm/
D4x4c2-sse2.c66 const __m128i va0 = _mm_loadl_epi64((const __m128i*) a0); in xnn_q8_gemm_ukernel_4x4c2__sse2()
69 const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1); in xnn_q8_gemm_ukernel_4x4c2__sse2()
72 const __m128i va2 = _mm_loadl_epi64((const __m128i*) a2); in xnn_q8_gemm_ukernel_4x4c2__sse2()
75 const __m128i va3 = _mm_loadl_epi64((const __m128i*) a3); in xnn_q8_gemm_ukernel_4x4c2__sse2()
79 const __m128i vb0 = _mm_loadl_epi64((const __m128i*) w); in xnn_q8_gemm_ukernel_4x4c2__sse2()
91 const __m128i vb1 = _mm_loadl_epi64((const __m128i*) ((uintptr_t) w + 8)); in xnn_q8_gemm_ukernel_4x4c2__sse2()
103 const __m128i vb2 = _mm_loadl_epi64((const __m128i*) ((uintptr_t) w + 16)); in xnn_q8_gemm_ukernel_4x4c2__sse2()
115 const __m128i vb3 = _mm_loadl_epi64((const __m128i*) ((uintptr_t) w + 24)); in xnn_q8_gemm_ukernel_4x4c2__sse2()
131 const __m128i va0 = _mm_loadl_epi64((const __m128i*) a0); in xnn_q8_gemm_ukernel_4x4c2__sse2()
134 const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1); in xnn_q8_gemm_ukernel_4x4c2__sse2()
[all …]
/external/XNNPACK/src/q8-igemm/
D4x4c2-sse2.c83 const __m128i va0 = _mm_loadl_epi64((const __m128i*) a0); in xnn_q8_igemm_ukernel_4x4c2__sse2()
86 const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1); in xnn_q8_igemm_ukernel_4x4c2__sse2()
89 const __m128i va2 = _mm_loadl_epi64((const __m128i*) a2); in xnn_q8_igemm_ukernel_4x4c2__sse2()
92 const __m128i va3 = _mm_loadl_epi64((const __m128i*) a3); in xnn_q8_igemm_ukernel_4x4c2__sse2()
96 const __m128i vb0 = _mm_loadl_epi64((const __m128i*) w); in xnn_q8_igemm_ukernel_4x4c2__sse2()
103 const __m128i vb1 = _mm_loadl_epi64((const __m128i*) ((uintptr_t) w + 8)); in xnn_q8_igemm_ukernel_4x4c2__sse2()
110 const __m128i vb2 = _mm_loadl_epi64((const __m128i*) ((uintptr_t) w + 16)); in xnn_q8_igemm_ukernel_4x4c2__sse2()
117 const __m128i vb3 = _mm_loadl_epi64((const __m128i*) ((uintptr_t) w + 24)); in xnn_q8_igemm_ukernel_4x4c2__sse2()
129 const __m128i va0 = _mm_loadl_epi64((const __m128i*) a0); in xnn_q8_igemm_ukernel_4x4c2__sse2()
131 const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1); in xnn_q8_igemm_ukernel_4x4c2__sse2()
[all …]

1234567