• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1;
2; jdmrgext.asm - merged upsampling/color conversion (64-bit AVX2)
3;
4; Copyright 2009, 2012 Pierre Ossman <ossman@cendio.se> for Cendio AB
5; Copyright (C) 2009, 2012, 2016, D. R. Commander.
6; Copyright (C) 2015, Intel Corporation.
7;
8; Based on the x86 SIMD extension for IJG JPEG library
9; Copyright (C) 1999-2006, MIYASAKA Masaru.
10; For conditions of distribution and use, see copyright notice in jsimdext.inc
11;
12; This file should be assembled with NASM (Netwide Assembler),
13; can *not* be assembled with Microsoft's MASM or any compatible
14; assembler (including Borland's Turbo Assembler).
15; NASM is available from http://nasm.sourceforge.net/ or
16; http://sourceforge.net/project/showfiles.php?group_id=6208
17;
18; [TAB8]
19
20%include "jcolsamp.inc"
21
22; --------------------------------------------------------------------------
23;
24; Upsample and color convert for the case of 2:1 horizontal and 1:1 vertical.
25;
26; GLOBAL(void)
27; jsimd_h2v1_merged_upsample_avx2(JDIMENSION output_width,
28;                                 JSAMPIMAGE input_buf,
29;                                 JDIMENSION in_row_group_ctr,
30;                                 JSAMPARRAY output_buf);
31;
32
33; r10d = JDIMENSION output_width
34; r11 = JSAMPIMAGE input_buf
35; r12d = JDIMENSION in_row_group_ctr
36; r13 = JSAMPARRAY output_buf
37
38%define wk(i)   rbp - (WK_NUM - (i)) * SIZEOF_YMMWORD  ; ymmword wk[WK_NUM]
39%define WK_NUM  3
40
41    align       32
42    GLOBAL_FUNCTION(jsimd_h2v1_merged_upsample_avx2)
43
44EXTN(jsimd_h2v1_merged_upsample_avx2):
45    push        rbp
46    mov         rax, rsp                     ; rax = original rbp
47    sub         rsp, byte 4
48    and         rsp, byte (-SIZEOF_YMMWORD)  ; align to 256 bits
49    mov         [rsp], rax
50    mov         rbp, rsp                     ; rbp = aligned rbp
51    lea         rsp, [wk(0)]
52    collect_args 4
53    push        rbx
54
55    mov         ecx, r10d               ; col
56    test        rcx, rcx
57    jz          near .return
58
59    push        rcx
60
61    mov         rdi, r11
62    mov         ecx, r12d
63    mov         rsi, JSAMPARRAY [rdi+0*SIZEOF_JSAMPARRAY]
64    mov         rbx, JSAMPARRAY [rdi+1*SIZEOF_JSAMPARRAY]
65    mov         rdx, JSAMPARRAY [rdi+2*SIZEOF_JSAMPARRAY]
66    mov         rdi, r13
67    mov         rsi, JSAMPROW [rsi+rcx*SIZEOF_JSAMPROW]  ; inptr0
68    mov         rbx, JSAMPROW [rbx+rcx*SIZEOF_JSAMPROW]  ; inptr1
69    mov         rdx, JSAMPROW [rdx+rcx*SIZEOF_JSAMPROW]  ; inptr2
70    mov         rdi, JSAMPROW [rdi]                      ; outptr
71
72    pop         rcx                     ; col
73
74.columnloop:
75
76    vmovdqu     ymm6, YMMWORD [rbx]     ; ymm6=Cb(0123456789ABCDEFGHIJKLMNOPQRSTUV)
77    vmovdqu     ymm7, YMMWORD [rdx]     ; ymm7=Cr(0123456789ABCDEFGHIJKLMNOPQRSTUV)
78
79    vpxor       ymm1, ymm1, ymm1        ; ymm1=(all 0's)
80    vpcmpeqw    ymm3, ymm3, ymm3
81    vpsllw      ymm3, ymm3, 7           ; ymm3={0xFF80 0xFF80 0xFF80 0xFF80 ..}
82
83    vpermq      ymm6, ymm6, 0xd8        ; ymm6=Cb(01234567GHIJKLMN89ABCDEFOPQRSTUV)
84    vpermq      ymm7, ymm7, 0xd8        ; ymm7=Cr(01234567GHIJKLMN89ABCDEFOPQRSTUV)
85    vpunpcklbw  ymm4, ymm6, ymm1        ; ymm4=Cb(0123456789ABCDEF)=CbL
86    vpunpckhbw  ymm6, ymm6, ymm1        ; ymm6=Cb(GHIJKLMNOPQRSTUV)=CbH
87    vpunpcklbw  ymm0, ymm7, ymm1        ; ymm0=Cr(0123456789ABCDEF)=CrL
88    vpunpckhbw  ymm7, ymm7, ymm1        ; ymm7=Cr(GHIJKLMNOPQRSTUV)=CrH
89
90    vpaddw      ymm5, ymm6, ymm3
91    vpaddw      ymm2, ymm4, ymm3
92    vpaddw      ymm1, ymm7, ymm3
93    vpaddw      ymm3, ymm0, ymm3
94
95    ; (Original)
96    ; R = Y                + 1.40200 * Cr
97    ; G = Y - 0.34414 * Cb - 0.71414 * Cr
98    ; B = Y + 1.77200 * Cb
99    ;
100    ; (This implementation)
101    ; R = Y                + 0.40200 * Cr + Cr
102    ; G = Y - 0.34414 * Cb + 0.28586 * Cr - Cr
103    ; B = Y - 0.22800 * Cb + Cb + Cb
104
105    vpaddw      ymm6, ymm5, ymm5             ; ymm6=2*CbH
106    vpaddw      ymm4, ymm2, ymm2             ; ymm4=2*CbL
107    vpaddw      ymm7, ymm1, ymm1             ; ymm7=2*CrH
108    vpaddw      ymm0, ymm3, ymm3             ; ymm0=2*CrL
109
110    vpmulhw     ymm6, ymm6, [rel PW_MF0228]  ; ymm6=(2*CbH * -FIX(0.22800))
111    vpmulhw     ymm4, ymm4, [rel PW_MF0228]  ; ymm4=(2*CbL * -FIX(0.22800))
112    vpmulhw     ymm7, ymm7, [rel PW_F0402]   ; ymm7=(2*CrH * FIX(0.40200))
113    vpmulhw     ymm0, ymm0, [rel PW_F0402]   ; ymm0=(2*CrL * FIX(0.40200))
114
115    vpaddw      ymm6, ymm6, [rel PW_ONE]
116    vpaddw      ymm4, ymm4, [rel PW_ONE]
117    vpsraw      ymm6, ymm6, 1                ; ymm6=(CbH * -FIX(0.22800))
118    vpsraw      ymm4, ymm4, 1                ; ymm4=(CbL * -FIX(0.22800))
119    vpaddw      ymm7, ymm7, [rel PW_ONE]
120    vpaddw      ymm0, ymm0, [rel PW_ONE]
121    vpsraw      ymm7, ymm7, 1                ; ymm7=(CrH * FIX(0.40200))
122    vpsraw      ymm0, ymm0, 1                ; ymm0=(CrL * FIX(0.40200))
123
124    vpaddw      ymm6, ymm6, ymm5
125    vpaddw      ymm4, ymm4, ymm2
126    vpaddw      ymm6, ymm6, ymm5             ; ymm6=(CbH * FIX(1.77200))=(B-Y)H
127    vpaddw      ymm4, ymm4, ymm2             ; ymm4=(CbL * FIX(1.77200))=(B-Y)L
128    vpaddw      ymm7, ymm7, ymm1             ; ymm7=(CrH * FIX(1.40200))=(R-Y)H
129    vpaddw      ymm0, ymm0, ymm3             ; ymm0=(CrL * FIX(1.40200))=(R-Y)L
130
131    vmovdqa     YMMWORD [wk(0)], ymm6        ; wk(0)=(B-Y)H
132    vmovdqa     YMMWORD [wk(1)], ymm7        ; wk(1)=(R-Y)H
133
134    vpunpckhwd  ymm6, ymm5, ymm1
135    vpunpcklwd  ymm5, ymm5, ymm1
136    vpmaddwd    ymm5, ymm5, [rel PW_MF0344_F0285]
137    vpmaddwd    ymm6, ymm6, [rel PW_MF0344_F0285]
138    vpunpckhwd  ymm7, ymm2, ymm3
139    vpunpcklwd  ymm2, ymm2, ymm3
140    vpmaddwd    ymm2, ymm2, [rel PW_MF0344_F0285]
141    vpmaddwd    ymm7, ymm7, [rel PW_MF0344_F0285]
142
143    vpaddd      ymm5, ymm5, [rel PD_ONEHALF]
144    vpaddd      ymm6, ymm6, [rel PD_ONEHALF]
145    vpsrad      ymm5, ymm5, SCALEBITS
146    vpsrad      ymm6, ymm6, SCALEBITS
147    vpaddd      ymm2, ymm2, [rel PD_ONEHALF]
148    vpaddd      ymm7, ymm7, [rel PD_ONEHALF]
149    vpsrad      ymm2, ymm2, SCALEBITS
150    vpsrad      ymm7, ymm7, SCALEBITS
151
152    vpackssdw   ymm5, ymm5, ymm6        ; ymm5=CbH*-FIX(0.344)+CrH*FIX(0.285)
153    vpackssdw   ymm2, ymm2, ymm7        ; ymm2=CbL*-FIX(0.344)+CrL*FIX(0.285)
154    vpsubw      ymm5, ymm5, ymm1        ; ymm5=CbH*-FIX(0.344)+CrH*-FIX(0.714)=(G-Y)H
155    vpsubw      ymm2, ymm2, ymm3        ; ymm2=CbL*-FIX(0.344)+CrL*-FIX(0.714)=(G-Y)L
156
157    vmovdqa     YMMWORD [wk(2)], ymm5   ; wk(2)=(G-Y)H
158
159    mov         al, 2                   ; Yctr
160    jmp         short .Yloop_1st
161
162.Yloop_2nd:
163    vmovdqa     ymm0, YMMWORD [wk(1)]   ; ymm0=(R-Y)H
164    vmovdqa     ymm2, YMMWORD [wk(2)]   ; ymm2=(G-Y)H
165    vmovdqa     ymm4, YMMWORD [wk(0)]   ; ymm4=(B-Y)H
166
167.Yloop_1st:
168    vmovdqu     ymm7, YMMWORD [rsi]     ; ymm7=Y(0123456789ABCDEFGHIJKLMNOPQRSTUV)
169
170    vpcmpeqw    ymm6, ymm6, ymm6
171    vpsrlw      ymm6, ymm6, BYTE_BIT    ; ymm6={0xFF 0x00 0xFF 0x00 ..}
172    vpand       ymm6, ymm6, ymm7        ; ymm6=Y(02468ACEGIKMOQSU)=YE
173    vpsrlw      ymm7, ymm7, BYTE_BIT    ; ymm7=Y(13579BDFHJLNPRTV)=YO
174
175    vmovdqa     ymm1, ymm0              ; ymm1=ymm0=(R-Y)(L/H)
176    vmovdqa     ymm3, ymm2              ; ymm3=ymm2=(G-Y)(L/H)
177    vmovdqa     ymm5, ymm4              ; ymm5=ymm4=(B-Y)(L/H)
178
179    vpaddw      ymm0, ymm0, ymm6        ; ymm0=((R-Y)+YE)=RE=R(02468ACEGIKMOQSU)
180    vpaddw      ymm1, ymm1, ymm7        ; ymm1=((R-Y)+YO)=RO=R(13579BDFHJLNPRTV)
181    vpackuswb   ymm0, ymm0, ymm0        ; ymm0=R(02468ACE********GIKMOQSU********)
182    vpackuswb   ymm1, ymm1, ymm1        ; ymm1=R(13579BDF********HJLNPRTV********)
183
184    vpaddw      ymm2, ymm2, ymm6        ; ymm2=((G-Y)+YE)=GE=G(02468ACEGIKMOQSU)
185    vpaddw      ymm3, ymm3, ymm7        ; ymm3=((G-Y)+YO)=GO=G(13579BDFHJLNPRTV)
186    vpackuswb   ymm2, ymm2, ymm2        ; ymm2=G(02468ACE********GIKMOQSU********)
187    vpackuswb   ymm3, ymm3, ymm3        ; ymm3=G(13579BDF********HJLNPRTV********)
188
189    vpaddw      ymm4, ymm4, ymm6        ; ymm4=((B-Y)+YE)=BE=B(02468ACEGIKMOQSU)
190    vpaddw      ymm5, ymm5, ymm7        ; ymm5=((B-Y)+YO)=BO=B(13579BDFHJLNPRTV)
191    vpackuswb   ymm4, ymm4, ymm4        ; ymm4=B(02468ACE********GIKMOQSU********)
192    vpackuswb   ymm5, ymm5, ymm5        ; ymm5=B(13579BDF********HJLNPRTV********)
193
194%if RGB_PIXELSIZE == 3  ; ---------------
195
196    ; ymmA=(00 02 04 06 08 0A 0C 0E ** 0G 0I 0K 0M 0O 0Q 0S 0U **)
197    ; ymmB=(01 03 05 07 09 0B 0D 0F ** 0H 0J 0L 0N 0P 0R 0T 0V **)
198    ; ymmC=(10 12 14 16 18 1A 1C 1E ** 1G 1I 1K 1M 1O 1Q 1S 1U **)
199    ; ymmD=(11 13 15 17 19 1B 1D 1F ** 1H 1J 1L 1N 1P 1R 1T 1V **)
200    ; ymmE=(20 22 24 26 28 2A 2C 2E ** 2G 2I 2K 2M 2O 2Q 2S 2U **)
201    ; ymmF=(21 23 25 27 29 2B 2D 2F ** 2H 2J 2L 2N 2P 2R 2T 2V **)
202    ; ymmG=(** ** ** ** ** ** ** ** ** ** ** ** ** ** ** ** ** **)
203    ; ymmH=(** ** ** ** ** ** ** ** ** ** ** ** ** ** ** ** ** **)
204
205    vpunpcklbw  ymmA, ymmA, ymmC        ; ymmA=(00 10 02 12 04 14 06 16 08 18 0A 1A 0C 1C 0E 1E
206                                        ;       0G 1G 0I 1I 0K 1K 0M 1M 0O 1O 0Q 1Q 0S 1S 0U 1U)
207    vpunpcklbw  ymmE, ymmE, ymmB        ; ymmE=(20 01 22 03 24 05 26 07 28 09 2A 0B 2C 0D 2E 0F
208                                        ;       2G 0H 2I 0J 2K 0L 2M 0N 2O 0P 2Q 0R 2S 0T 2U 0V)
209    vpunpcklbw  ymmD, ymmD, ymmF        ; ymmD=(11 21 13 23 15 25 17 27 19 29 1B 2B 1D 2D 1F 2F
210                                        ;       1H 2H 1J 2J 1L 2L 1N 2N 1P 2P 1R 2R 1T 2T 1V 2V)
211
212    vpsrldq     ymmH, ymmA, 2           ; ymmH=(02 12 04 14 06 16 08 18 0A 1A 0C 1C 0E 1E 0G 1G
213                                        ;       0I 1I 0K 1K 0M 1M 0O 1O 0Q 1Q 0S 1S 0U 1U -- --)
214    vpunpckhwd  ymmG, ymmA, ymmE        ; ymmG=(08 18 28 09 0A 1A 2A 0B 0C 1C 2C 0D 0E 1E 2E 0F
215                                        ;       0O 1O 2O 0P 0Q 1Q 2Q 0R 0S 1S 2S 0T 0U 1U 2U 0V)
216    vpunpcklwd  ymmA, ymmA, ymmE        ; ymmA=(00 10 20 01 02 12 22 03 04 14 24 05 06 16 26 07
217                                        ;       0G 1G 2G 0H 0I 1I 2I 0J 0K 1K 2K 0L 0M 1M 2M 0N)
218
219    vpsrldq     ymmE, ymmE, 2           ; ymmE=(22 03 24 05 26 07 28 09 2A 0B 2C 0D 2E 0F 2G 0H
220                                        ;       2I 0J 2K 0L 2M 0N 2O 0P 2Q 0R 2S 0T 2U 0V -- --)
221
222    vpsrldq     ymmB, ymmD, 2           ; ymmB=(13 23 15 25 17 27 19 29 1B 2B 1D 2D 1F 2F 1H 2H
223                                        ;       1J 2J 1L 2L 1N 2N 1P 2P 1R 2R 1T 2T 1V 2V -- --)
224    vpunpckhwd  ymmC, ymmD, ymmH        ; ymmC=(19 29 0A 1A 1B 2B 0C 1C 1D 2D 0E 1E 1F 2F 0G 1G
225                                        ;       1P 2P 0Q 1Q 1R 2R 0S 1S 1T 2T 0U 1U 1V 2V -- --)
226    vpunpcklwd  ymmD, ymmD, ymmH        ; ymmD=(11 21 02 12 13 23 04 14 15 25 06 16 17 27 08 18
227                                        ;       1H 2H 0I 1I 1J 2J 0K 1K 1L 2L 0M 1M 1N 2N 0O 1O)
228
229    vpunpckhwd  ymmF, ymmE, ymmB        ; ymmF=(2A 0B 1B 2B 2C 0D 1D 2D 2E 0F 1F 2F 2G 0H 1H 2H
230                                        ;       2Q 0R 1R 2R 2S 0T 1T 2T 2U 0V 1V 2V -- -- -- --)
231    vpunpcklwd  ymmE, ymmE, ymmB        ; ymmE=(22 03 13 23 24 05 15 25 26 07 17 27 28 09 19 29
232                                        ;       2I 0J 1J 2J 2K 0L 1L 2L 2M 0N 1N 2N 2O 0P 1P 2P)
233
234    vpshufd     ymmH, ymmA, 0x4E        ; ymmH=(04 14 24 05 06 16 26 07 00 10 20 01 02 12 22 03
235                                        ;       0K 1K 2K 0L 0M 1M 2M 0N 0G 1G 2G 0H 0I 1I 2I 0J)
236    vpunpckldq  ymmA, ymmA, ymmD        ; ymmA=(00 10 20 01 11 21 02 12 02 12 22 03 13 23 04 14
237                                        ;       0G 1G 2G 0H 1H 2H 0I 1I 0I 1I 2I 0J 1J 2J 0K 1K)
238    vpunpckhdq  ymmD, ymmD, ymmE        ; ymmD=(15 25 06 16 26 07 17 27 17 27 08 18 28 09 19 29
239                                        ;       1L 2L 0M 1M 2M 0N 1N 2N 1N 2N 0O 1O 2O 0P 1P 2P)
240    vpunpckldq  ymmE, ymmE, ymmH        ; ymmE=(22 03 13 23 04 14 24 05 24 05 15 25 06 16 26 07
241                                        ;       2I 0J 1J 2J 0K 1K 2K 0L 2K 0L 1L 2L 0M 1M 2M 0N)
242
243    vpshufd     ymmH, ymmG, 0x4E        ; ymmH=(0C 1C 2C 0D 0E 1E 2E 0F 08 18 28 09 0A 1A 2A 0B
244                                        ;       0S 1S 2S 0T 0U 1U 2U 0V 0O 1O 2O 0P 0Q 1Q 2Q 0R)
245    vpunpckldq  ymmG, ymmG, ymmC        ; ymmG=(08 18 28 09 19 29 0A 1A 0A 1A 2A 0B 1B 2B 0C 1C
246                                        ;       0O 1O 2O 0P 1P 2P 0Q 1Q 0Q 1Q 2Q 0R 1R 2R 0S 1S)
247    vpunpckhdq  ymmC, ymmC, ymmF        ; ymmC=(1D 2D 0E 1E 2E 0F 1F 2F 1F 2F 0G 1G 2G 0H 1H 2H
248                                        ;       1T 2T 0U 1U 2U 0V 1V 2V 1V 2V -- -- -- -- -- --)
249    vpunpckldq  ymmF, ymmF, ymmH        ; ymmF=(2A 0B 1B 2B 0C 1C 2C 0D 2C 0D 1D 2D 0E 1E 2E 0F
250                                        ;       2Q 0R 1R 2R 0S 1S 2S 0T 2S 0T 1T 2T 0U 1U 2U 0V)
251
252    vpunpcklqdq ymmH, ymmA, ymmE        ; ymmH=(00 10 20 01 11 21 02 12 22 03 13 23 04 14 24 05
253                                        ;       0G 1G 2G 0H 1H 2H 0I 1I 2I 0J 1J 2J 0K 1K 2K 0L)
254    vpunpcklqdq ymmG, ymmD, ymmG        ; ymmG=(15 25 06 16 26 07 17 27 08 18 28 09 19 29 0A 1A
255                                        ;       1L 2L 0M 1M 2M 0N 1N 2N 0O 1O 2O 0P 1P 2P 0Q 1Q)
256    vpunpcklqdq ymmC, ymmF, ymmC        ; ymmC=(2A 0B 1B 2B 0C 1C 2C 0D 1D 2D 0E 1E 2E 0F 1F 2F
257                                        ;       2Q 0R 1R 2R 0S 1S 2S 0T 1T 2T 0U 1U 2U 0V 1V 2V)
258
259    vperm2i128  ymmA, ymmH, ymmG, 0x20  ; ymmA=(00 10 20 01 11 21 02 12 22 03 13 23 04 14 24 05
260                                        ;       15 25 06 16 26 07 17 27 08 18 28 09 19 29 0A 1A)
261    vperm2i128  ymmD, ymmC, ymmH, 0x30  ; ymmD=(2A 0B 1B 2B 0C 1C 2C 0D 1D 2D 0E 1E 2E 0F 1F 2F
262                                        ;       0G 1G 2G 0H 1H 2H 0I 1I 2I 0J 1J 2J 0K 1K 2K 0L)
263    vperm2i128  ymmF, ymmG, ymmC, 0x31  ; ymmF=(1L 2L 0M 1M 2M 0N 1N 2N 0O 1O 2O 0P 1P 2P 0Q 1Q
264                                        ;       2Q 0R 1R 2R 0S 1S 2S 0T 1T 2T 0U 1U 2U 0V 1V 2V)
265
266    cmp         rcx, byte SIZEOF_YMMWORD
267    jb          short .column_st64
268
269    test        rdi, SIZEOF_YMMWORD-1
270    jnz         short .out1
271    ; --(aligned)-------------------
272    vmovntdq    YMMWORD [rdi+0*SIZEOF_YMMWORD], ymmA
273    vmovntdq    YMMWORD [rdi+1*SIZEOF_YMMWORD], ymmD
274    vmovntdq    YMMWORD [rdi+2*SIZEOF_YMMWORD], ymmF
275    jmp         short .out0
276.out1:  ; --(unaligned)-----------------
277    vmovdqu     YMMWORD [rdi+0*SIZEOF_YMMWORD], ymmA
278    vmovdqu     YMMWORD [rdi+1*SIZEOF_YMMWORD], ymmD
279    vmovdqu     YMMWORD [rdi+2*SIZEOF_YMMWORD], ymmF
280.out0:
281    add         rdi, byte RGB_PIXELSIZE*SIZEOF_YMMWORD  ; outptr
282    sub         rcx, byte SIZEOF_YMMWORD
283    jz          near .endcolumn
284
285    add         rsi, byte SIZEOF_YMMWORD  ; inptr0
286    dec         al                        ; Yctr
287    jnz         near .Yloop_2nd
288
289    add         rbx, byte SIZEOF_YMMWORD  ; inptr1
290    add         rdx, byte SIZEOF_YMMWORD  ; inptr2
291    jmp         near .columnloop
292
293.column_st64:
294    lea         rcx, [rcx+rcx*2]            ; imul ecx, RGB_PIXELSIZE
295    cmp         rcx, byte 2*SIZEOF_YMMWORD
296    jb          short .column_st32
297    vmovdqu     YMMWORD [rdi+0*SIZEOF_YMMWORD], ymmA
298    vmovdqu     YMMWORD [rdi+1*SIZEOF_YMMWORD], ymmD
299    add         rdi, byte 2*SIZEOF_YMMWORD  ; outptr
300    vmovdqa     ymmA, ymmF
301    sub         rcx, byte 2*SIZEOF_YMMWORD
302    jmp         short .column_st31
303.column_st32:
304    cmp         rcx, byte SIZEOF_YMMWORD
305    jb          short .column_st31
306    vmovdqu     YMMWORD [rdi+0*SIZEOF_YMMWORD], ymmA
307    add         rdi, byte SIZEOF_YMMWORD    ; outptr
308    vmovdqa     ymmA, ymmD
309    sub         rcx, byte SIZEOF_YMMWORD
310    jmp         short .column_st31
311.column_st31:
312    cmp         rcx, byte SIZEOF_XMMWORD
313    jb          short .column_st15
314    vmovdqu     XMMWORD [rdi+0*SIZEOF_XMMWORD], xmmA
315    add         rdi, byte SIZEOF_XMMWORD    ; outptr
316    vperm2i128  ymmA, ymmA, ymmA, 1
317    sub         rcx, byte SIZEOF_XMMWORD
318.column_st15:
319    ; Store the lower 8 bytes of xmmA to the output when it has enough
320    ; space.
321    cmp         rcx, byte SIZEOF_MMWORD
322    jb          short .column_st7
323    vmovq       XMM_MMWORD [rdi], xmmA
324    add         rdi, byte SIZEOF_MMWORD
325    sub         rcx, byte SIZEOF_MMWORD
326    vpsrldq     xmmA, xmmA, SIZEOF_MMWORD
327.column_st7:
328    ; Store the lower 4 bytes of xmmA to the output when it has enough
329    ; space.
330    cmp         rcx, byte SIZEOF_DWORD
331    jb          short .column_st3
332    vmovd       XMM_DWORD [rdi], xmmA
333    add         rdi, byte SIZEOF_DWORD
334    sub         rcx, byte SIZEOF_DWORD
335    vpsrldq     xmmA, xmmA, SIZEOF_DWORD
336.column_st3:
337    ; Store the lower 2 bytes of rax to the output when it has enough
338    ; space.
339    vmovd       eax, xmmA
340    cmp         rcx, byte SIZEOF_WORD
341    jb          short .column_st1
342    mov         WORD [rdi], ax
343    add         rdi, byte SIZEOF_WORD
344    sub         rcx, byte SIZEOF_WORD
345    shr         rax, 16
346.column_st1:
347    ; Store the lower 1 byte of rax to the output when it has enough
348    ; space.
349    test        rcx, rcx
350    jz          short .endcolumn
351    mov         BYTE [rdi], al
352
353%else  ; RGB_PIXELSIZE == 4 ; -----------
354
355%ifdef RGBX_FILLER_0XFF
356    vpcmpeqb    ymm6, ymm6, ymm6        ; ymm6=XE=X(02468ACE********GIKMOQSU********)
357    vpcmpeqb    ymm7, ymm7, ymm7        ; ymm7=XO=X(13579BDF********HJLNPRTV********)
358%else
359    vpxor       ymm6, ymm6, ymm6        ; ymm6=XE=X(02468ACE********GIKMOQSU********)
360    vpxor       ymm7, ymm7, ymm7        ; ymm7=XO=X(13579BDF********HJLNPRTV********)
361%endif
362    ; ymmA=(00 02 04 06 08 0A 0C 0E ** 0G 0I 0K 0M 0O 0Q 0S 0U **)
363    ; ymmB=(01 03 05 07 09 0B 0D 0F ** 0H 0J 0L 0N 0P 0R 0T 0V **)
364    ; ymmC=(10 12 14 16 18 1A 1C 1E ** 1G 1I 1K 1M 1O 1Q 1S 1U **)
365    ; ymmD=(11 13 15 17 19 1B 1D 1F ** 1H 1J 1L 1N 1P 1R 1T 1V **)
366    ; ymmE=(20 22 24 26 28 2A 2C 2E ** 2G 2I 2K 2M 2O 2Q 2S 2U **)
367    ; ymmF=(21 23 25 27 29 2B 2D 2F ** 2H 2J 2L 2N 2P 2R 2T 2V **)
368    ; ymmG=(30 32 34 36 38 3A 3C 3E ** 3G 3I 3K 3M 3O 3Q 3S 3U **)
369    ; ymmH=(31 33 35 37 39 3B 3D 3F ** 3H 3J 3L 3N 3P 3R 3T 3V **)
370
371    vpunpcklbw  ymmA, ymmA, ymmC        ; ymmA=(00 10 02 12 04 14 06 16 08 18 0A 1A 0C 1C 0E 1E
372                                        ;       0G 1G 0I 1I 0K 1K 0M 1M 0O 1O 0Q 1Q 0S 1S 0U 1U)
373    vpunpcklbw  ymmE, ymmE, ymmG        ; ymmE=(20 30 22 32 24 34 26 36 28 38 2A 3A 2C 3C 2E 3E
374                                        ;       2G 3G 2I 3I 2K 3K 2M 3M 2O 3O 2Q 3Q 2S 3S 2U 3U)
375    vpunpcklbw  ymmB, ymmB, ymmD        ; ymmB=(01 11 03 13 05 15 07 17 09 19 0B 1B 0D 1D 0F 1F
376                                        ;       0H 1H 0J 1J 0L 1L 0N 1N 0P 1P 0R 1R 0T 1T 0V 1V)
377    vpunpcklbw  ymmF, ymmF, ymmH        ; ymmF=(21 31 23 33 25 35 27 37 29 39 2B 3B 2D 3D 2F 3F
378                                        ;       2H 3H 2J 3J 2L 3L 2N 3N 2P 3P 2R 3R 2T 3T 2V 3V)
379
380    vpunpckhwd  ymmC, ymmA, ymmE        ; ymmC=(08 18 28 38 0A 1A 2A 3A 0C 1C 2C 3C 0E 1E 2E 3E
381                                        ;       0O 1O 2O 3O 0Q 1Q 2Q 3Q 0S 1S 2S 3S 0U 1U 2U 3U)
382    vpunpcklwd  ymmA, ymmA, ymmE        ; ymmA=(00 10 20 30 02 12 22 32 04 14 24 34 06 16 26 36
383                                        ;       0G 1G 2G 3G 0I 1I 2I 3I 0K 1K 2K 3K 0M 1M 2M 3M)
384    vpunpckhwd  ymmG, ymmB, ymmF        ; ymmG=(09 19 29 39 0B 1B 2B 3B 0D 1D 2D 3D 0F 1F 2F 3F
385                                        ;       0P 1P 2P 3P 0R 1R 2R 3R 0T 1T 2T 3T 0V 1V 2V 3V)
386    vpunpcklwd  ymmB, ymmB, ymmF        ; ymmB=(01 11 21 31 03 13 23 33 05 15 25 35 07 17 27 37
387                                        ;       0H 1H 2H 3H 0J 1J 2J 3J 0L 1L 2L 3L 0N 1N 2N 3N)
388
389    vpunpckhdq  ymmE, ymmA, ymmB        ; ymmE=(04 14 24 34 05 15 25 35 06 16 26 36 07 17 27 37
390                                        ;       0K 1K 2K 3K 0L 1L 2L 3L 0M 1M 2M 3M 0N 1N 2N 3N)
391    vpunpckldq  ymmB, ymmA, ymmB        ; ymmB=(00 10 20 30 01 11 21 31 02 12 22 32 03 13 23 33
392                                        ;       0G 1G 2G 3G 0H 1H 2H 3H 0I 1I 2I 3I 0J 1J 2J 3J)
393    vpunpckhdq  ymmF, ymmC, ymmG        ; ymmF=(0C 1C 2C 3C 0D 1D 2D 3D 0E 1E 2E 3E 0F 1F 2F 3F
394                                        ;       0S 1S 2S 3S 0T 1T 2T 3T 0U 1U 2U 3U 0V 1V 2V 3V)
395    vpunpckldq  ymmG, ymmC, ymmG        ; ymmG=(08 18 28 38 09 19 29 39 0A 1A 2A 3A 0B 1B 2B 3B
396                                        ;       0O 1O 2O 3O 0P 1P 2P 3P 0Q 1Q 2Q 3Q 0R 1R 2R 3R)
397
398    vperm2i128  ymmA, ymmB, ymmE, 0x20  ; ymmA=(00 10 20 30 01 11 21 31 02 12 22 32 03 13 23 33
399                                        ;       04 14 24 34 05 15 25 35 06 16 26 36 07 17 27 37)
400    vperm2i128  ymmD, ymmG, ymmF, 0x20  ; ymmD=(08 18 28 38 09 19 29 39 0A 1A 2A 3A 0B 1B 2B 3B
401                                        ;       0C 1C 2C 3C 0D 1D 2D 3D 0E 1E 2E 3E 0F 1F 2F 3F)
402    vperm2i128  ymmC, ymmB, ymmE, 0x31  ; ymmC=(0G 1G 2G 3G 0H 1H 2H 3H 0I 1I 2I 3I 0J 1J 2J 3J
403                                        ;       0K 1K 2K 3K 0L 1L 2L 3L 0M 1M 2M 3M 0N 1N 2N 3N)
404    vperm2i128  ymmH, ymmG, ymmF, 0x31  ; ymmH=(0O 1O 2O 3O 0P 1P 2P 3P 0Q 1Q 2Q 3Q 0R 1R 2R 3R
405                                        ;       0S 1S 2S 3S 0T 1T 2T 3T 0U 1U 2U 3U 0V 1V 2V 3V)
406
407    cmp         rcx, byte SIZEOF_YMMWORD
408    jb          short .column_st64
409
410    test        rdi, SIZEOF_YMMWORD-1
411    jnz         short .out1
412    ; --(aligned)-------------------
413    vmovntdq    YMMWORD [rdi+0*SIZEOF_YMMWORD], ymmA
414    vmovntdq    YMMWORD [rdi+1*SIZEOF_YMMWORD], ymmD
415    vmovntdq    YMMWORD [rdi+2*SIZEOF_YMMWORD], ymmC
416    vmovntdq    YMMWORD [rdi+3*SIZEOF_YMMWORD], ymmH
417    jmp         short .out0
418.out1:  ; --(unaligned)-----------------
419    vmovdqu     YMMWORD [rdi+0*SIZEOF_YMMWORD], ymmA
420    vmovdqu     YMMWORD [rdi+1*SIZEOF_YMMWORD], ymmD
421    vmovdqu     YMMWORD [rdi+2*SIZEOF_YMMWORD], ymmC
422    vmovdqu     YMMWORD [rdi+3*SIZEOF_YMMWORD], ymmH
423.out0:
424    add         rdi, RGB_PIXELSIZE*SIZEOF_YMMWORD  ; outptr
425    sub         rcx, byte SIZEOF_YMMWORD
426    jz          near .endcolumn
427
428    add         rsi, byte SIZEOF_YMMWORD  ; inptr0
429    dec         al
430    jnz         near .Yloop_2nd
431
432    add         rbx, byte SIZEOF_YMMWORD  ; inptr1
433    add         rdx, byte SIZEOF_YMMWORD  ; inptr2
434    jmp         near .columnloop
435
436.column_st64:
437    cmp         rcx, byte SIZEOF_YMMWORD/2
438    jb          short .column_st32
439    vmovdqu     YMMWORD [rdi+0*SIZEOF_YMMWORD], ymmA
440    vmovdqu     YMMWORD [rdi+1*SIZEOF_YMMWORD], ymmD
441    add         rdi, byte 2*SIZEOF_YMMWORD  ; outptr
442    vmovdqa     ymmA, ymmC
443    vmovdqa     ymmD, ymmH
444    sub         rcx, byte SIZEOF_YMMWORD/2
445.column_st32:
446    cmp         rcx, byte SIZEOF_YMMWORD/4
447    jb          short .column_st16
448    vmovdqu     YMMWORD [rdi+0*SIZEOF_YMMWORD], ymmA
449    add         rdi, byte SIZEOF_YMMWORD    ; outptr
450    vmovdqa     ymmA, ymmD
451    sub         rcx, byte SIZEOF_YMMWORD/4
452.column_st16:
453    cmp         rcx, byte SIZEOF_YMMWORD/8
454    jb          short .column_st15
455    vmovdqu     XMMWORD [rdi+0*SIZEOF_XMMWORD], xmmA
456    add         rdi, byte SIZEOF_XMMWORD    ; outptr
457    vperm2i128  ymmA, ymmA, ymmA, 1
458    sub         rcx, byte SIZEOF_YMMWORD/8
459.column_st15:
460    ; Store two pixels (8 bytes) of ymmA to the output when it has enough
461    ; space.
462    cmp         rcx, byte SIZEOF_YMMWORD/16
463    jb          short .column_st7
464    vmovq       MMWORD [rdi], xmmA
465    add         rdi, byte SIZEOF_YMMWORD/16*4
466    sub         rcx, byte SIZEOF_YMMWORD/16
467    vpsrldq     xmmA, SIZEOF_YMMWORD/16*4
468.column_st7:
469    ; Store one pixel (4 bytes) of ymmA to the output when it has enough
470    ; space.
471    test        rcx, rcx
472    jz          short .endcolumn
473    vmovd       XMM_DWORD [rdi], xmmA
474
475%endif  ; RGB_PIXELSIZE ; ---------------
476
477.endcolumn:
478    sfence                              ; flush the write buffer
479
480.return:
481    pop         rbx
482    vzeroupper
483    uncollect_args 4
484    mov         rsp, rbp                ; rsp <- aligned rbp
485    pop         rsp                     ; rsp <- original rbp
486    pop         rbp
487    ret
488
489; --------------------------------------------------------------------------
490;
491; Upsample and color convert for the case of 2:1 horizontal and 2:1 vertical.
492;
493; GLOBAL(void)
494; jsimd_h2v2_merged_upsample_avx2(JDIMENSION output_width,
495;                                 JSAMPIMAGE input_buf,
496;                                 JDIMENSION in_row_group_ctr,
497;                                 JSAMPARRAY output_buf);
498;
499
500; r10d = JDIMENSION output_width
501; r11 = JSAMPIMAGE input_buf
502; r12d = JDIMENSION in_row_group_ctr
503; r13 = JSAMPARRAY output_buf
504
505    align       32
506    GLOBAL_FUNCTION(jsimd_h2v2_merged_upsample_avx2)
507
508EXTN(jsimd_h2v2_merged_upsample_avx2):
509    push        rbp
510    mov         rax, rsp
511    mov         rbp, rsp
512    collect_args 4
513    push        rbx
514
515    mov         eax, r10d
516
517    mov         rdi, r11
518    mov         ecx, r12d
519    mov         rsi, JSAMPARRAY [rdi+0*SIZEOF_JSAMPARRAY]
520    mov         rbx, JSAMPARRAY [rdi+1*SIZEOF_JSAMPARRAY]
521    mov         rdx, JSAMPARRAY [rdi+2*SIZEOF_JSAMPARRAY]
522    mov         rdi, r13
523    lea         rsi, [rsi+rcx*SIZEOF_JSAMPROW]
524
525    push        rdx                     ; inptr2
526    push        rbx                     ; inptr1
527    push        rsi                     ; inptr00
528    mov         rbx, rsp
529
530    push        rdi
531    push        rcx
532    push        rax
533
534    %ifdef WIN64
535    mov         r8, rcx
536    mov         r9, rdi
537    mov         rcx, rax
538    mov         rdx, rbx
539    %else
540    mov         rdx, rcx
541    mov         rcx, rdi
542    mov         rdi, rax
543    mov         rsi, rbx
544    %endif
545
546    call        EXTN(jsimd_h2v1_merged_upsample_avx2)
547
548    pop         rax
549    pop         rcx
550    pop         rdi
551    pop         rsi
552    pop         rbx
553    pop         rdx
554
555    add         rdi, byte SIZEOF_JSAMPROW  ; outptr1
556    add         rsi, byte SIZEOF_JSAMPROW  ; inptr01
557
558    push        rdx                     ; inptr2
559    push        rbx                     ; inptr1
560    push        rsi                     ; inptr00
561    mov         rbx, rsp
562
563    push        rdi
564    push        rcx
565    push        rax
566
567    %ifdef WIN64
568    mov         r8, rcx
569    mov         r9, rdi
570    mov         rcx, rax
571    mov         rdx, rbx
572    %else
573    mov         rdx, rcx
574    mov         rcx, rdi
575    mov         rdi, rax
576    mov         rsi, rbx
577    %endif
578
579    call        EXTN(jsimd_h2v1_merged_upsample_avx2)
580
581    pop         rax
582    pop         rcx
583    pop         rdi
584    pop         rsi
585    pop         rbx
586    pop         rdx
587
588    pop         rbx
589    uncollect_args 4
590    pop         rbp
591    ret
592
593; For some reason, the OS X linker does not honor the request to align the
594; segment unless we do this.
595    align       32
596