• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 /*
2  *  Copyright (c) 2017 The WebM project authors. All Rights Reserved.
3  *
4  *  Use of this source code is governed by a BSD-style license
5  *  that can be found in the LICENSE file in the root of the source
6  *  tree. An additional intellectual property rights grant can be found
7  *  in the file PATENTS.  All contributing project authors may
8  *  be found in the AUTHORS file in the root of the source tree.
9  */
10 
11 #include <emmintrin.h>  // SSE2
12 
13 #include "./vpx_config.h"
14 #include "./vpx_dsp_rtcd.h"
15 #include "vpx/vpx_integer.h"
16 
17 // -----------------------------------------------------------------------------
18 
vpx_highbd_h_predictor_4x4_sse2(uint16_t * dst,ptrdiff_t stride,const uint16_t * above,const uint16_t * left,int bd)19 void vpx_highbd_h_predictor_4x4_sse2(uint16_t *dst, ptrdiff_t stride,
20                                      const uint16_t *above,
21                                      const uint16_t *left, int bd) {
22   const __m128i left_u16 = _mm_loadl_epi64((const __m128i *)left);
23   const __m128i row0 = _mm_shufflelo_epi16(left_u16, 0x0);
24   const __m128i row1 = _mm_shufflelo_epi16(left_u16, 0x55);
25   const __m128i row2 = _mm_shufflelo_epi16(left_u16, 0xaa);
26   const __m128i row3 = _mm_shufflelo_epi16(left_u16, 0xff);
27   (void)above;
28   (void)bd;
29   _mm_storel_epi64((__m128i *)dst, row0);
30   dst += stride;
31   _mm_storel_epi64((__m128i *)dst, row1);
32   dst += stride;
33   _mm_storel_epi64((__m128i *)dst, row2);
34   dst += stride;
35   _mm_storel_epi64((__m128i *)dst, row3);
36 }
37 
vpx_highbd_h_predictor_8x8_sse2(uint16_t * dst,ptrdiff_t stride,const uint16_t * above,const uint16_t * left,int bd)38 void vpx_highbd_h_predictor_8x8_sse2(uint16_t *dst, ptrdiff_t stride,
39                                      const uint16_t *above,
40                                      const uint16_t *left, int bd) {
41   const __m128i left_u16 = _mm_load_si128((const __m128i *)left);
42   const __m128i row0 = _mm_shufflelo_epi16(left_u16, 0x0);
43   const __m128i row1 = _mm_shufflelo_epi16(left_u16, 0x55);
44   const __m128i row2 = _mm_shufflelo_epi16(left_u16, 0xaa);
45   const __m128i row3 = _mm_shufflelo_epi16(left_u16, 0xff);
46   const __m128i row4 = _mm_shufflehi_epi16(left_u16, 0x0);
47   const __m128i row5 = _mm_shufflehi_epi16(left_u16, 0x55);
48   const __m128i row6 = _mm_shufflehi_epi16(left_u16, 0xaa);
49   const __m128i row7 = _mm_shufflehi_epi16(left_u16, 0xff);
50   (void)above;
51   (void)bd;
52   _mm_store_si128((__m128i *)dst, _mm_unpacklo_epi64(row0, row0));
53   dst += stride;
54   _mm_store_si128((__m128i *)dst, _mm_unpacklo_epi64(row1, row1));
55   dst += stride;
56   _mm_store_si128((__m128i *)dst, _mm_unpacklo_epi64(row2, row2));
57   dst += stride;
58   _mm_store_si128((__m128i *)dst, _mm_unpacklo_epi64(row3, row3));
59   dst += stride;
60   _mm_store_si128((__m128i *)dst, _mm_unpackhi_epi64(row4, row4));
61   dst += stride;
62   _mm_store_si128((__m128i *)dst, _mm_unpackhi_epi64(row5, row5));
63   dst += stride;
64   _mm_store_si128((__m128i *)dst, _mm_unpackhi_epi64(row6, row6));
65   dst += stride;
66   _mm_store_si128((__m128i *)dst, _mm_unpackhi_epi64(row7, row7));
67 }
68 
h_store_16_unpacklo(uint16_t ** dst,const ptrdiff_t stride,const __m128i * row)69 static INLINE void h_store_16_unpacklo(uint16_t **dst, const ptrdiff_t stride,
70                                        const __m128i *row) {
71   const __m128i val = _mm_unpacklo_epi64(*row, *row);
72   _mm_store_si128((__m128i *)*dst, val);
73   _mm_store_si128((__m128i *)(*dst + 8), val);
74   *dst += stride;
75 }
76 
h_store_16_unpackhi(uint16_t ** dst,const ptrdiff_t stride,const __m128i * row)77 static INLINE void h_store_16_unpackhi(uint16_t **dst, const ptrdiff_t stride,
78                                        const __m128i *row) {
79   const __m128i val = _mm_unpackhi_epi64(*row, *row);
80   _mm_store_si128((__m128i *)(*dst), val);
81   _mm_store_si128((__m128i *)(*dst + 8), val);
82   *dst += stride;
83 }
84 
vpx_highbd_h_predictor_16x16_sse2(uint16_t * dst,ptrdiff_t stride,const uint16_t * above,const uint16_t * left,int bd)85 void vpx_highbd_h_predictor_16x16_sse2(uint16_t *dst, ptrdiff_t stride,
86                                        const uint16_t *above,
87                                        const uint16_t *left, int bd) {
88   int i;
89   (void)above;
90   (void)bd;
91 
92   for (i = 0; i < 2; i++, left += 8) {
93     const __m128i left_u16 = _mm_load_si128((const __m128i *)left);
94     const __m128i row0 = _mm_shufflelo_epi16(left_u16, 0x0);
95     const __m128i row1 = _mm_shufflelo_epi16(left_u16, 0x55);
96     const __m128i row2 = _mm_shufflelo_epi16(left_u16, 0xaa);
97     const __m128i row3 = _mm_shufflelo_epi16(left_u16, 0xff);
98     const __m128i row4 = _mm_shufflehi_epi16(left_u16, 0x0);
99     const __m128i row5 = _mm_shufflehi_epi16(left_u16, 0x55);
100     const __m128i row6 = _mm_shufflehi_epi16(left_u16, 0xaa);
101     const __m128i row7 = _mm_shufflehi_epi16(left_u16, 0xff);
102     h_store_16_unpacklo(&dst, stride, &row0);
103     h_store_16_unpacklo(&dst, stride, &row1);
104     h_store_16_unpacklo(&dst, stride, &row2);
105     h_store_16_unpacklo(&dst, stride, &row3);
106     h_store_16_unpackhi(&dst, stride, &row4);
107     h_store_16_unpackhi(&dst, stride, &row5);
108     h_store_16_unpackhi(&dst, stride, &row6);
109     h_store_16_unpackhi(&dst, stride, &row7);
110   }
111 }
112 
h_store_32_unpacklo(uint16_t ** dst,const ptrdiff_t stride,const __m128i * row)113 static INLINE void h_store_32_unpacklo(uint16_t **dst, const ptrdiff_t stride,
114                                        const __m128i *row) {
115   const __m128i val = _mm_unpacklo_epi64(*row, *row);
116   _mm_store_si128((__m128i *)(*dst), val);
117   _mm_store_si128((__m128i *)(*dst + 8), val);
118   _mm_store_si128((__m128i *)(*dst + 16), val);
119   _mm_store_si128((__m128i *)(*dst + 24), val);
120   *dst += stride;
121 }
122 
h_store_32_unpackhi(uint16_t ** dst,const ptrdiff_t stride,const __m128i * row)123 static INLINE void h_store_32_unpackhi(uint16_t **dst, const ptrdiff_t stride,
124                                        const __m128i *row) {
125   const __m128i val = _mm_unpackhi_epi64(*row, *row);
126   _mm_store_si128((__m128i *)(*dst), val);
127   _mm_store_si128((__m128i *)(*dst + 8), val);
128   _mm_store_si128((__m128i *)(*dst + 16), val);
129   _mm_store_si128((__m128i *)(*dst + 24), val);
130   *dst += stride;
131 }
132 
vpx_highbd_h_predictor_32x32_sse2(uint16_t * dst,ptrdiff_t stride,const uint16_t * above,const uint16_t * left,int bd)133 void vpx_highbd_h_predictor_32x32_sse2(uint16_t *dst, ptrdiff_t stride,
134                                        const uint16_t *above,
135                                        const uint16_t *left, int bd) {
136   int i;
137   (void)above;
138   (void)bd;
139 
140   for (i = 0; i < 4; i++, left += 8) {
141     const __m128i left_u16 = _mm_load_si128((const __m128i *)left);
142     const __m128i row0 = _mm_shufflelo_epi16(left_u16, 0x0);
143     const __m128i row1 = _mm_shufflelo_epi16(left_u16, 0x55);
144     const __m128i row2 = _mm_shufflelo_epi16(left_u16, 0xaa);
145     const __m128i row3 = _mm_shufflelo_epi16(left_u16, 0xff);
146     const __m128i row4 = _mm_shufflehi_epi16(left_u16, 0x0);
147     const __m128i row5 = _mm_shufflehi_epi16(left_u16, 0x55);
148     const __m128i row6 = _mm_shufflehi_epi16(left_u16, 0xaa);
149     const __m128i row7 = _mm_shufflehi_epi16(left_u16, 0xff);
150     h_store_32_unpacklo(&dst, stride, &row0);
151     h_store_32_unpacklo(&dst, stride, &row1);
152     h_store_32_unpacklo(&dst, stride, &row2);
153     h_store_32_unpacklo(&dst, stride, &row3);
154     h_store_32_unpackhi(&dst, stride, &row4);
155     h_store_32_unpackhi(&dst, stride, &row5);
156     h_store_32_unpackhi(&dst, stride, &row6);
157     h_store_32_unpackhi(&dst, stride, &row7);
158   }
159 }
160 
161 //------------------------------------------------------------------------------
162 // DC 4x4
163 
dc_sum_4(const uint16_t * ref)164 static INLINE __m128i dc_sum_4(const uint16_t *ref) {
165   const __m128i _dcba = _mm_loadl_epi64((const __m128i *)ref);
166   const __m128i _xxdc = _mm_shufflelo_epi16(_dcba, 0xe);
167   const __m128i a = _mm_add_epi16(_dcba, _xxdc);
168   return _mm_add_epi16(a, _mm_shufflelo_epi16(a, 0x1));
169 }
170 
dc_store_4x4(uint16_t * dst,ptrdiff_t stride,const __m128i * dc)171 static INLINE void dc_store_4x4(uint16_t *dst, ptrdiff_t stride,
172                                 const __m128i *dc) {
173   const __m128i dc_dup = _mm_shufflelo_epi16(*dc, 0x0);
174   int i;
175   for (i = 0; i < 4; ++i, dst += stride) {
176     _mm_storel_epi64((__m128i *)dst, dc_dup);
177   }
178 }
179 
vpx_highbd_dc_left_predictor_4x4_sse2(uint16_t * dst,ptrdiff_t stride,const uint16_t * above,const uint16_t * left,int bd)180 void vpx_highbd_dc_left_predictor_4x4_sse2(uint16_t *dst, ptrdiff_t stride,
181                                            const uint16_t *above,
182                                            const uint16_t *left, int bd) {
183   const __m128i two = _mm_cvtsi32_si128(2);
184   const __m128i sum = dc_sum_4(left);
185   const __m128i dc = _mm_srli_epi16(_mm_add_epi16(sum, two), 2);
186   (void)above;
187   (void)bd;
188   dc_store_4x4(dst, stride, &dc);
189 }
190 
vpx_highbd_dc_top_predictor_4x4_sse2(uint16_t * dst,ptrdiff_t stride,const uint16_t * above,const uint16_t * left,int bd)191 void vpx_highbd_dc_top_predictor_4x4_sse2(uint16_t *dst, ptrdiff_t stride,
192                                           const uint16_t *above,
193                                           const uint16_t *left, int bd) {
194   const __m128i two = _mm_cvtsi32_si128(2);
195   const __m128i sum = dc_sum_4(above);
196   const __m128i dc = _mm_srli_epi16(_mm_add_epi16(sum, two), 2);
197   (void)left;
198   (void)bd;
199   dc_store_4x4(dst, stride, &dc);
200 }
201 
vpx_highbd_dc_128_predictor_4x4_sse2(uint16_t * dst,ptrdiff_t stride,const uint16_t * above,const uint16_t * left,int bd)202 void vpx_highbd_dc_128_predictor_4x4_sse2(uint16_t *dst, ptrdiff_t stride,
203                                           const uint16_t *above,
204                                           const uint16_t *left, int bd) {
205   const __m128i dc = _mm_cvtsi32_si128(1 << (bd - 1));
206   const __m128i dc_dup = _mm_shufflelo_epi16(dc, 0x0);
207   (void)above;
208   (void)left;
209   dc_store_4x4(dst, stride, &dc_dup);
210 }
211 
212 //------------------------------------------------------------------------------
213 // DC 8x8
214 
dc_sum_8(const uint16_t * ref)215 static INLINE __m128i dc_sum_8(const uint16_t *ref) {
216   const __m128i ref_u16 = _mm_load_si128((const __m128i *)ref);
217   const __m128i _dcba = _mm_add_epi16(ref_u16, _mm_srli_si128(ref_u16, 8));
218   const __m128i _xxdc = _mm_shufflelo_epi16(_dcba, 0xe);
219   const __m128i a = _mm_add_epi16(_dcba, _xxdc);
220 
221   return _mm_add_epi16(a, _mm_shufflelo_epi16(a, 0x1));
222 }
223 
dc_store_8x8(uint16_t * dst,ptrdiff_t stride,const __m128i * dc)224 static INLINE void dc_store_8x8(uint16_t *dst, ptrdiff_t stride,
225                                 const __m128i *dc) {
226   const __m128i dc_dup_lo = _mm_shufflelo_epi16(*dc, 0);
227   const __m128i dc_dup = _mm_unpacklo_epi64(dc_dup_lo, dc_dup_lo);
228   int i;
229   for (i = 0; i < 8; ++i, dst += stride) {
230     _mm_store_si128((__m128i *)dst, dc_dup);
231   }
232 }
233 
vpx_highbd_dc_left_predictor_8x8_sse2(uint16_t * dst,ptrdiff_t stride,const uint16_t * above,const uint16_t * left,int bd)234 void vpx_highbd_dc_left_predictor_8x8_sse2(uint16_t *dst, ptrdiff_t stride,
235                                            const uint16_t *above,
236                                            const uint16_t *left, int bd) {
237   const __m128i four = _mm_cvtsi32_si128(4);
238   const __m128i sum = dc_sum_8(left);
239   const __m128i dc = _mm_srli_epi16(_mm_add_epi16(sum, four), 3);
240   (void)above;
241   (void)bd;
242   dc_store_8x8(dst, stride, &dc);
243 }
244 
vpx_highbd_dc_top_predictor_8x8_sse2(uint16_t * dst,ptrdiff_t stride,const uint16_t * above,const uint16_t * left,int bd)245 void vpx_highbd_dc_top_predictor_8x8_sse2(uint16_t *dst, ptrdiff_t stride,
246                                           const uint16_t *above,
247                                           const uint16_t *left, int bd) {
248   const __m128i four = _mm_cvtsi32_si128(4);
249   const __m128i sum = dc_sum_8(above);
250   const __m128i dc = _mm_srli_epi16(_mm_add_epi16(sum, four), 3);
251   (void)left;
252   (void)bd;
253   dc_store_8x8(dst, stride, &dc);
254 }
255 
vpx_highbd_dc_128_predictor_8x8_sse2(uint16_t * dst,ptrdiff_t stride,const uint16_t * above,const uint16_t * left,int bd)256 void vpx_highbd_dc_128_predictor_8x8_sse2(uint16_t *dst, ptrdiff_t stride,
257                                           const uint16_t *above,
258                                           const uint16_t *left, int bd) {
259   const __m128i dc = _mm_cvtsi32_si128(1 << (bd - 1));
260   const __m128i dc_dup = _mm_shufflelo_epi16(dc, 0x0);
261   (void)above;
262   (void)left;
263   dc_store_8x8(dst, stride, &dc_dup);
264 }
265 
266 //------------------------------------------------------------------------------
267 // DC 16x16
268 
dc_sum_16(const uint16_t * ref)269 static INLINE __m128i dc_sum_16(const uint16_t *ref) {
270   const __m128i sum_lo = dc_sum_8(ref);
271   const __m128i sum_hi = dc_sum_8(ref + 8);
272   return _mm_add_epi16(sum_lo, sum_hi);
273 }
274 
dc_store_16x16(uint16_t * dst,ptrdiff_t stride,const __m128i * dc)275 static INLINE void dc_store_16x16(uint16_t *dst, ptrdiff_t stride,
276                                   const __m128i *dc) {
277   const __m128i dc_dup_lo = _mm_shufflelo_epi16(*dc, 0);
278   const __m128i dc_dup = _mm_unpacklo_epi64(dc_dup_lo, dc_dup_lo);
279   int i;
280   for (i = 0; i < 16; ++i, dst += stride) {
281     _mm_store_si128((__m128i *)dst, dc_dup);
282     _mm_store_si128((__m128i *)(dst + 8), dc_dup);
283   }
284 }
285 
vpx_highbd_dc_left_predictor_16x16_sse2(uint16_t * dst,ptrdiff_t stride,const uint16_t * above,const uint16_t * left,int bd)286 void vpx_highbd_dc_left_predictor_16x16_sse2(uint16_t *dst, ptrdiff_t stride,
287                                              const uint16_t *above,
288                                              const uint16_t *left, int bd) {
289   const __m128i eight = _mm_cvtsi32_si128(8);
290   const __m128i sum = dc_sum_16(left);
291   const __m128i dc = _mm_srli_epi16(_mm_add_epi16(sum, eight), 4);
292   (void)above;
293   (void)bd;
294   dc_store_16x16(dst, stride, &dc);
295 }
296 
vpx_highbd_dc_top_predictor_16x16_sse2(uint16_t * dst,ptrdiff_t stride,const uint16_t * above,const uint16_t * left,int bd)297 void vpx_highbd_dc_top_predictor_16x16_sse2(uint16_t *dst, ptrdiff_t stride,
298                                             const uint16_t *above,
299                                             const uint16_t *left, int bd) {
300   const __m128i eight = _mm_cvtsi32_si128(8);
301   const __m128i sum = dc_sum_16(above);
302   const __m128i dc = _mm_srli_epi16(_mm_add_epi16(sum, eight), 4);
303   (void)left;
304   (void)bd;
305   dc_store_16x16(dst, stride, &dc);
306 }
307 
vpx_highbd_dc_128_predictor_16x16_sse2(uint16_t * dst,ptrdiff_t stride,const uint16_t * above,const uint16_t * left,int bd)308 void vpx_highbd_dc_128_predictor_16x16_sse2(uint16_t *dst, ptrdiff_t stride,
309                                             const uint16_t *above,
310                                             const uint16_t *left, int bd) {
311   const __m128i dc = _mm_cvtsi32_si128(1 << (bd - 1));
312   const __m128i dc_dup = _mm_shufflelo_epi16(dc, 0x0);
313   (void)above;
314   (void)left;
315   dc_store_16x16(dst, stride, &dc_dup);
316 }
317 
318 //------------------------------------------------------------------------------
319 // DC 32x32
320 
dc_sum_32(const uint16_t * ref)321 static INLINE __m128i dc_sum_32(const uint16_t *ref) {
322   const __m128i zero = _mm_setzero_si128();
323   const __m128i sum_a = dc_sum_16(ref);
324   const __m128i sum_b = dc_sum_16(ref + 16);
325   // 12 bit bd will outrange, so expand to 32 bit before adding final total
326   return _mm_add_epi32(_mm_unpacklo_epi16(sum_a, zero),
327                        _mm_unpacklo_epi16(sum_b, zero));
328 }
329 
dc_store_32x32(uint16_t * dst,ptrdiff_t stride,const __m128i * dc)330 static INLINE void dc_store_32x32(uint16_t *dst, ptrdiff_t stride,
331                                   const __m128i *dc) {
332   const __m128i dc_dup_lo = _mm_shufflelo_epi16(*dc, 0);
333   const __m128i dc_dup = _mm_unpacklo_epi64(dc_dup_lo, dc_dup_lo);
334   int i;
335   for (i = 0; i < 32; ++i, dst += stride) {
336     _mm_store_si128((__m128i *)dst, dc_dup);
337     _mm_store_si128((__m128i *)(dst + 8), dc_dup);
338     _mm_store_si128((__m128i *)(dst + 16), dc_dup);
339     _mm_store_si128((__m128i *)(dst + 24), dc_dup);
340   }
341 }
342 
vpx_highbd_dc_left_predictor_32x32_sse2(uint16_t * dst,ptrdiff_t stride,const uint16_t * above,const uint16_t * left,int bd)343 void vpx_highbd_dc_left_predictor_32x32_sse2(uint16_t *dst, ptrdiff_t stride,
344                                              const uint16_t *above,
345                                              const uint16_t *left, int bd) {
346   const __m128i sixteen = _mm_cvtsi32_si128(16);
347   const __m128i sum = dc_sum_32(left);
348   const __m128i dc = _mm_srli_epi32(_mm_add_epi32(sum, sixteen), 5);
349   (void)above;
350   (void)bd;
351   dc_store_32x32(dst, stride, &dc);
352 }
353 
vpx_highbd_dc_top_predictor_32x32_sse2(uint16_t * dst,ptrdiff_t stride,const uint16_t * above,const uint16_t * left,int bd)354 void vpx_highbd_dc_top_predictor_32x32_sse2(uint16_t *dst, ptrdiff_t stride,
355                                             const uint16_t *above,
356                                             const uint16_t *left, int bd) {
357   const __m128i sixteen = _mm_cvtsi32_si128(16);
358   const __m128i sum = dc_sum_32(above);
359   const __m128i dc = _mm_srli_epi32(_mm_add_epi32(sum, sixteen), 5);
360   (void)left;
361   (void)bd;
362   dc_store_32x32(dst, stride, &dc);
363 }
364 
vpx_highbd_dc_128_predictor_32x32_sse2(uint16_t * dst,ptrdiff_t stride,const uint16_t * above,const uint16_t * left,int bd)365 void vpx_highbd_dc_128_predictor_32x32_sse2(uint16_t *dst, ptrdiff_t stride,
366                                             const uint16_t *above,
367                                             const uint16_t *left, int bd) {
368   const __m128i dc = _mm_cvtsi32_si128(1 << (bd - 1));
369   const __m128i dc_dup = _mm_shufflelo_epi16(dc, 0x0);
370   (void)above;
371   (void)left;
372   dc_store_32x32(dst, stride, &dc_dup);
373 }
374 
375 // -----------------------------------------------------------------------------
376 /*
377 ; ------------------------------------------
378 ; input: x, y, z, result
379 ;
380 ; trick from pascal
381 ; (x+2y+z+2)>>2 can be calculated as:
382 ; result = avg(x,z)
383 ; result -= xor(x,z) & 1
384 ; result = avg(result,y)
385 ; ------------------------------------------
386 */
avg3_epu16(const __m128i * x,const __m128i * y,const __m128i * z)387 static INLINE __m128i avg3_epu16(const __m128i *x, const __m128i *y,
388                                  const __m128i *z) {
389   const __m128i one = _mm_set1_epi16(1);
390   const __m128i a = _mm_avg_epu16(*x, *z);
391   const __m128i b =
392       _mm_subs_epu16(a, _mm_and_si128(_mm_xor_si128(*x, *z), one));
393   return _mm_avg_epu16(b, *y);
394 }
395 
vpx_highbd_d117_predictor_4x4_sse2(uint16_t * dst,ptrdiff_t stride,const uint16_t * above,const uint16_t * left,int bd)396 void vpx_highbd_d117_predictor_4x4_sse2(uint16_t *dst, ptrdiff_t stride,
397                                         const uint16_t *above,
398                                         const uint16_t *left, int bd) {
399   const int I = left[0];
400   const int J = left[1];
401   const int K = left[2];
402   const __m128i XXXXABCD = _mm_loadu_si128((const __m128i *)(above - 4));
403   const __m128i KXXXABCD = _mm_insert_epi16(XXXXABCD, K, 0);
404   const __m128i KJXXABCD = _mm_insert_epi16(KXXXABCD, J, 1);
405   const __m128i KJIXABCD = _mm_insert_epi16(KJXXABCD, I, 2);
406   const __m128i JIXABCD0 = _mm_srli_si128(KJIXABCD, 2);
407   const __m128i IXABCD00 = _mm_srli_si128(KJIXABCD, 4);
408   const __m128i avg2 = _mm_avg_epu16(KJIXABCD, JIXABCD0);
409   const __m128i avg3 = avg3_epu16(&KJIXABCD, &JIXABCD0, &IXABCD00);
410   const __m128i row0 = _mm_srli_si128(avg2, 6);
411   const __m128i row1 = _mm_srli_si128(avg3, 4);
412   const __m128i row2 = _mm_srli_si128(avg2, 4);
413   const __m128i row3 = _mm_srli_si128(avg3, 2);
414   (void)bd;
415   _mm_storel_epi64((__m128i *)dst, row0);
416   dst += stride;
417   _mm_storel_epi64((__m128i *)dst, row1);
418   dst += stride;
419   _mm_storel_epi64((__m128i *)dst, row2);
420   dst += stride;
421   _mm_storel_epi64((__m128i *)dst, row3);
422 
423   dst -= stride;
424   dst[0] = _mm_extract_epi16(avg3, 1);
425   dst[stride] = _mm_extract_epi16(avg3, 0);
426 }
427 
vpx_highbd_d135_predictor_4x4_sse2(uint16_t * dst,ptrdiff_t stride,const uint16_t * above,const uint16_t * left,int bd)428 void vpx_highbd_d135_predictor_4x4_sse2(uint16_t *dst, ptrdiff_t stride,
429                                         const uint16_t *above,
430                                         const uint16_t *left, int bd) {
431   const int I = left[0];
432   const int J = left[1];
433   const int K = left[2];
434   const int L = left[3];
435   const __m128i XXXXABCD = _mm_loadu_si128((const __m128i *)(above - 4));
436   const __m128i KXXXABCD = _mm_insert_epi16(XXXXABCD, K, 0);
437   const __m128i KJXXABCD = _mm_insert_epi16(KXXXABCD, J, 1);
438   const __m128i KJIXABCD = _mm_insert_epi16(KJXXABCD, I, 2);
439   const __m128i JIXABCD0 = _mm_srli_si128(KJIXABCD, 2);
440   const __m128i LKJIXABC = _mm_insert_epi16(_mm_slli_si128(KJIXABCD, 2), L, 0);
441   const __m128i avg3 = avg3_epu16(&JIXABCD0, &KJIXABCD, &LKJIXABC);
442   const __m128i row0 = _mm_srli_si128(avg3, 6);
443   const __m128i row1 = _mm_srli_si128(avg3, 4);
444   const __m128i row2 = _mm_srli_si128(avg3, 2);
445   const __m128i row3 = avg3;
446   (void)bd;
447   _mm_storel_epi64((__m128i *)dst, row0);
448   dst += stride;
449   _mm_storel_epi64((__m128i *)dst, row1);
450   dst += stride;
451   _mm_storel_epi64((__m128i *)dst, row2);
452   dst += stride;
453   _mm_storel_epi64((__m128i *)dst, row3);
454 }
455 
vpx_highbd_d153_predictor_4x4_sse2(uint16_t * dst,ptrdiff_t stride,const uint16_t * above,const uint16_t * left,int bd)456 void vpx_highbd_d153_predictor_4x4_sse2(uint16_t *dst, ptrdiff_t stride,
457                                         const uint16_t *above,
458                                         const uint16_t *left, int bd) {
459   const int I = left[0];
460   const int J = left[1];
461   const int K = left[2];
462   const int L = left[3];
463   const __m128i XXXXXABC = _mm_castps_si128(
464       _mm_loadh_pi(_mm_setzero_ps(), (const __m64 *)(above - 1)));
465   const __m128i LXXXXABC = _mm_insert_epi16(XXXXXABC, L, 0);
466   const __m128i LKXXXABC = _mm_insert_epi16(LXXXXABC, K, 1);
467   const __m128i LKJXXABC = _mm_insert_epi16(LKXXXABC, J, 2);
468   const __m128i LKJIXABC = _mm_insert_epi16(LKJXXABC, I, 3);
469   const __m128i KJIXABC0 = _mm_srli_si128(LKJIXABC, 2);
470   const __m128i JIXABC00 = _mm_srli_si128(LKJIXABC, 4);
471   const __m128i avg3 = avg3_epu16(&LKJIXABC, &KJIXABC0, &JIXABC00);
472   const __m128i avg2 = _mm_avg_epu16(LKJIXABC, KJIXABC0);
473   const __m128i row3 = _mm_unpacklo_epi16(avg2, avg3);
474   const __m128i row2 = _mm_srli_si128(row3, 4);
475   const __m128i row1 = _mm_srli_si128(row3, 8);
476   const __m128i row0 = _mm_srli_si128(avg3, 4);
477   (void)bd;
478   _mm_storel_epi64((__m128i *)dst, row0);
479   dst[0] = _mm_extract_epi16(avg2, 3);
480   dst += stride;
481   _mm_storel_epi64((__m128i *)dst, row1);
482   dst += stride;
483   _mm_storel_epi64((__m128i *)dst, row2);
484   dst += stride;
485   _mm_storel_epi64((__m128i *)dst, row3);
486 }
487 
vpx_highbd_d207_predictor_4x4_sse2(uint16_t * dst,ptrdiff_t stride,const uint16_t * above,const uint16_t * left,int bd)488 void vpx_highbd_d207_predictor_4x4_sse2(uint16_t *dst, ptrdiff_t stride,
489                                         const uint16_t *above,
490                                         const uint16_t *left, int bd) {
491   const __m128i IJKL0000 = _mm_load_si128((const __m128i *)left);
492   const __m128i LLLL0000 = _mm_shufflelo_epi16(IJKL0000, 0xff);
493   const __m128i IJKLLLLL = _mm_unpacklo_epi64(IJKL0000, LLLL0000);
494   const __m128i JKLLLLL0 = _mm_srli_si128(IJKLLLLL, 2);
495   const __m128i KLLLLL00 = _mm_srli_si128(IJKLLLLL, 4);
496   const __m128i avg3 = avg3_epu16(&IJKLLLLL, &JKLLLLL0, &KLLLLL00);
497   const __m128i avg2 = _mm_avg_epu16(IJKLLLLL, JKLLLLL0);
498   const __m128i row0 = _mm_unpacklo_epi16(avg2, avg3);
499   const __m128i row1 = _mm_srli_si128(row0, 4);
500   const __m128i row2 = _mm_srli_si128(row0, 8);
501   const __m128i row3 = LLLL0000;
502   (void)above;
503   (void)bd;
504   _mm_storel_epi64((__m128i *)dst, row0);
505   dst += stride;
506   _mm_storel_epi64((__m128i *)dst, row1);
507   dst += stride;
508   _mm_storel_epi64((__m128i *)dst, row2);
509   dst += stride;
510   _mm_storel_epi64((__m128i *)dst, row3);
511 }
512 
vpx_highbd_d63_predictor_4x4_sse2(uint16_t * dst,ptrdiff_t stride,const uint16_t * above,const uint16_t * left,int bd)513 void vpx_highbd_d63_predictor_4x4_sse2(uint16_t *dst, ptrdiff_t stride,
514                                        const uint16_t *above,
515                                        const uint16_t *left, int bd) {
516   const __m128i ABCDEFGH = _mm_loadu_si128((const __m128i *)above);
517   const __m128i BCDEFGH0 = _mm_srli_si128(ABCDEFGH, 2);
518   const __m128i CDEFGH00 = _mm_srli_si128(ABCDEFGH, 4);
519   const __m128i avg3 = avg3_epu16(&ABCDEFGH, &BCDEFGH0, &CDEFGH00);
520   const __m128i avg2 = _mm_avg_epu16(ABCDEFGH, BCDEFGH0);
521   const __m128i row0 = avg2;
522   const __m128i row1 = avg3;
523   const __m128i row2 = _mm_srli_si128(avg2, 2);
524   const __m128i row3 = _mm_srli_si128(avg3, 2);
525   (void)left;
526   (void)bd;
527   _mm_storel_epi64((__m128i *)dst, row0);
528   dst += stride;
529   _mm_storel_epi64((__m128i *)dst, row1);
530   dst += stride;
531   _mm_storel_epi64((__m128i *)dst, row2);
532   dst += stride;
533   _mm_storel_epi64((__m128i *)dst, row3);
534 }
535