OLD | NEW |
| (Empty) |
1 ; | |
2 ; jiss2int.asm - accurate integer IDCT (SSE2) | |
3 ; | |
4 ; Copyright 2009 Pierre Ossman <ossman@cendio.se> for Cendio AB | |
5 ; | |
6 ; Based on | |
7 ; x86 SIMD extension for IJG JPEG library | |
8 ; Copyright (C) 1999-2006, MIYASAKA Masaru. | |
9 ; For conditions of distribution and use, see copyright notice in jsimdext.inc | |
10 ; | |
11 ; This file should be assembled with NASM (Netwide Assembler), | |
12 ; can *not* be assembled with Microsoft's MASM or any compatible | |
13 ; assembler (including Borland's Turbo Assembler). | |
14 ; NASM is available from http://nasm.sourceforge.net/ or | |
15 ; http://sourceforge.net/project/showfiles.php?group_id=6208 | |
16 ; | |
17 ; This file contains a slow-but-accurate integer implementation of the | |
18 ; inverse DCT (Discrete Cosine Transform). The following code is based | |
19 ; directly on the IJG's original jidctint.c; see the jidctint.c for | |
20 ; more details. | |
21 ; | |
22 ; [TAB8] | |
23 | |
24 %include "jsimdext.inc" | |
25 %include "jdct.inc" | |
26 | |
27 ; -------------------------------------------------------------------------- | |
28 | |
29 %define CONST_BITS 13 | |
30 %define PASS1_BITS 2 | |
31 | |
32 %define DESCALE_P1 (CONST_BITS-PASS1_BITS) | |
33 %define DESCALE_P2 (CONST_BITS+PASS1_BITS+3) | |
34 | |
35 %if CONST_BITS == 13 | |
36 F_0_298 equ 2446 ; FIX(0.298631336) | |
37 F_0_390 equ 3196 ; FIX(0.390180644) | |
38 F_0_541 equ 4433 ; FIX(0.541196100) | |
39 F_0_765 equ 6270 ; FIX(0.765366865) | |
40 F_0_899 equ 7373 ; FIX(0.899976223) | |
41 F_1_175 equ 9633 ; FIX(1.175875602) | |
42 F_1_501 equ 12299 ; FIX(1.501321110) | |
43 F_1_847 equ 15137 ; FIX(1.847759065) | |
44 F_1_961 equ 16069 ; FIX(1.961570560) | |
45 F_2_053 equ 16819 ; FIX(2.053119869) | |
46 F_2_562 equ 20995 ; FIX(2.562915447) | |
47 F_3_072 equ 25172 ; FIX(3.072711026) | |
48 %else | |
49 ; NASM cannot do compile-time arithmetic on floating-point constants. | |
50 %define DESCALE(x,n) (((x)+(1<<((n)-1)))>>(n)) | |
51 F_0_298 equ DESCALE( 320652955,30-CONST_BITS) ; FIX(0.298631336) | |
52 F_0_390 equ DESCALE( 418953276,30-CONST_BITS) ; FIX(0.390180644) | |
53 F_0_541 equ DESCALE( 581104887,30-CONST_BITS) ; FIX(0.541196100) | |
54 F_0_765 equ DESCALE( 821806413,30-CONST_BITS) ; FIX(0.765366865) | |
55 F_0_899 equ DESCALE( 966342111,30-CONST_BITS) ; FIX(0.899976223) | |
56 F_1_175 equ DESCALE(1262586813,30-CONST_BITS) ; FIX(1.175875602) | |
57 F_1_501 equ DESCALE(1612031267,30-CONST_BITS) ; FIX(1.501321110) | |
58 F_1_847 equ DESCALE(1984016188,30-CONST_BITS) ; FIX(1.847759065) | |
59 F_1_961 equ DESCALE(2106220350,30-CONST_BITS) ; FIX(1.961570560) | |
60 F_2_053 equ DESCALE(2204520673,30-CONST_BITS) ; FIX(2.053119869) | |
61 F_2_562 equ DESCALE(2751909506,30-CONST_BITS) ; FIX(2.562915447) | |
62 F_3_072 equ DESCALE(3299298341,30-CONST_BITS) ; FIX(3.072711026) | |
63 %endif | |
64 | |
65 ; -------------------------------------------------------------------------- | |
66 SECTION SEG_CONST | |
67 | |
68 alignz 16 | |
69 global EXTN(jconst_idct_islow_sse2) PRIVATE | |
70 | |
71 EXTN(jconst_idct_islow_sse2): | |
72 | |
73 PW_F130_F054 times 4 dw (F_0_541+F_0_765), F_0_541 | |
74 PW_F054_MF130 times 4 dw F_0_541, (F_0_541-F_1_847) | |
75 PW_MF078_F117 times 4 dw (F_1_175-F_1_961), F_1_175 | |
76 PW_F117_F078 times 4 dw F_1_175, (F_1_175-F_0_390) | |
77 PW_MF060_MF089 times 4 dw (F_0_298-F_0_899),-F_0_899 | |
78 PW_MF089_F060 times 4 dw -F_0_899, (F_1_501-F_0_899) | |
79 PW_MF050_MF256 times 4 dw (F_2_053-F_2_562),-F_2_562 | |
80 PW_MF256_F050 times 4 dw -F_2_562, (F_3_072-F_2_562) | |
81 PD_DESCALE_P1 times 4 dd 1 << (DESCALE_P1-1) | |
82 PD_DESCALE_P2 times 4 dd 1 << (DESCALE_P2-1) | |
83 PB_CENTERJSAMP times 16 db CENTERJSAMPLE | |
84 | |
85 alignz 16 | |
86 | |
87 ; -------------------------------------------------------------------------- | |
88 SECTION SEG_TEXT | |
89 BITS 32 | |
90 ; | |
91 ; Perform dequantization and inverse DCT on one block of coefficients. | |
92 ; | |
93 ; GLOBAL(void) | |
94 ; jsimd_idct_islow_sse2 (void * dct_table, JCOEFPTR coef_block, | |
95 ; JSAMPARRAY output_buf, JDIMENSION output_col) | |
96 ; | |
97 | |
98 %define dct_table(b) (b)+8 ; jpeg_component_info * compptr | |
99 %define coef_block(b) (b)+12 ; JCOEFPTR coef_block | |
100 %define output_buf(b) (b)+16 ; JSAMPARRAY output_buf | |
101 %define output_col(b) (b)+20 ; JDIMENSION output_col | |
102 | |
103 %define original_ebp ebp+0 | |
104 %define wk(i) ebp-(WK_NUM-(i))*SIZEOF_XMMWORD ; xmmword wk[WK_NUM] | |
105 %define WK_NUM 12 | |
106 | |
107 align 16 | |
108 global EXTN(jsimd_idct_islow_sse2) PRIVATE | |
109 | |
110 EXTN(jsimd_idct_islow_sse2): | |
111 push ebp | |
112 mov eax,esp ; eax = original ebp | |
113 sub esp, byte 4 | |
114 and esp, byte (-SIZEOF_XMMWORD) ; align to 128 bits | |
115 mov [esp],eax | |
116 mov ebp,esp ; ebp = aligned ebp | |
117 lea esp, [wk(0)] | |
118 pushpic ebx | |
119 ; push ecx ; unused | |
120 ; push edx ; need not be preserved | |
121 push esi | |
122 push edi | |
123 | |
124 get_GOT ebx ; get GOT address | |
125 | |
126 ; ---- Pass 1: process columns from input. | |
127 | |
128 ; mov eax, [original_ebp] | |
129 mov edx, POINTER [dct_table(eax)] ; quantptr | |
130 mov esi, JCOEFPTR [coef_block(eax)] ; inptr | |
131 | |
132 %ifndef NO_ZERO_COLUMN_TEST_ISLOW_SSE2 | |
133 mov eax, DWORD [DWBLOCK(1,0,esi,SIZEOF_JCOEF)] | |
134 or eax, DWORD [DWBLOCK(2,0,esi,SIZEOF_JCOEF)] | |
135 jnz near .columnDCT | |
136 | |
137 movdqa xmm0, XMMWORD [XMMBLOCK(1,0,esi,SIZEOF_JCOEF)] | |
138 movdqa xmm1, XMMWORD [XMMBLOCK(2,0,esi,SIZEOF_JCOEF)] | |
139 por xmm0, XMMWORD [XMMBLOCK(3,0,esi,SIZEOF_JCOEF)] | |
140 por xmm1, XMMWORD [XMMBLOCK(4,0,esi,SIZEOF_JCOEF)] | |
141 por xmm0, XMMWORD [XMMBLOCK(5,0,esi,SIZEOF_JCOEF)] | |
142 por xmm1, XMMWORD [XMMBLOCK(6,0,esi,SIZEOF_JCOEF)] | |
143 por xmm0, XMMWORD [XMMBLOCK(7,0,esi,SIZEOF_JCOEF)] | |
144 por xmm1,xmm0 | |
145 packsswb xmm1,xmm1 | |
146 packsswb xmm1,xmm1 | |
147 movd eax,xmm1 | |
148 test eax,eax | |
149 jnz short .columnDCT | |
150 | |
151 ; -- AC terms all zero | |
152 | |
153 movdqa xmm5, XMMWORD [XMMBLOCK(0,0,esi,SIZEOF_JCOEF)] | |
154 pmullw xmm5, XMMWORD [XMMBLOCK(0,0,edx,SIZEOF_ISLOW_MULT_TYPE)] | |
155 | |
156 psllw xmm5,PASS1_BITS | |
157 | |
158 movdqa xmm4,xmm5 ; xmm5=in0=(00 01 02 03 04 05 06 07) | |
159 punpcklwd xmm5,xmm5 ; xmm5=(00 00 01 01 02 02 03 03) | |
160 punpckhwd xmm4,xmm4 ; xmm4=(04 04 05 05 06 06 07 07) | |
161 | |
162 pshufd xmm7,xmm5,0x00 ; xmm7=col0=(00 00 00 00 00 00 00 00) | |
163 pshufd xmm6,xmm5,0x55 ; xmm6=col1=(01 01 01 01 01 01 01 01) | |
164 pshufd xmm1,xmm5,0xAA ; xmm1=col2=(02 02 02 02 02 02 02 02) | |
165 pshufd xmm5,xmm5,0xFF ; xmm5=col3=(03 03 03 03 03 03 03 03) | |
166 pshufd xmm0,xmm4,0x00 ; xmm0=col4=(04 04 04 04 04 04 04 04) | |
167 pshufd xmm3,xmm4,0x55 ; xmm3=col5=(05 05 05 05 05 05 05 05) | |
168 pshufd xmm2,xmm4,0xAA ; xmm2=col6=(06 06 06 06 06 06 06 06) | |
169 pshufd xmm4,xmm4,0xFF ; xmm4=col7=(07 07 07 07 07 07 07 07) | |
170 | |
171 movdqa XMMWORD [wk(8)], xmm6 ; wk(8)=col1 | |
172 movdqa XMMWORD [wk(9)], xmm5 ; wk(9)=col3 | |
173 movdqa XMMWORD [wk(10)], xmm3 ; wk(10)=col5 | |
174 movdqa XMMWORD [wk(11)], xmm4 ; wk(11)=col7 | |
175 jmp near .column_end | |
176 alignx 16,7 | |
177 %endif | |
178 .columnDCT: | |
179 | |
180 ; -- Even part | |
181 | |
182 movdqa xmm0, XMMWORD [XMMBLOCK(0,0,esi,SIZEOF_JCOEF)] | |
183 movdqa xmm1, XMMWORD [XMMBLOCK(2,0,esi,SIZEOF_JCOEF)] | |
184 pmullw xmm0, XMMWORD [XMMBLOCK(0,0,edx,SIZEOF_ISLOW_MULT_TYPE)] | |
185 pmullw xmm1, XMMWORD [XMMBLOCK(2,0,edx,SIZEOF_ISLOW_MULT_TYPE)] | |
186 movdqa xmm2, XMMWORD [XMMBLOCK(4,0,esi,SIZEOF_JCOEF)] | |
187 movdqa xmm3, XMMWORD [XMMBLOCK(6,0,esi,SIZEOF_JCOEF)] | |
188 pmullw xmm2, XMMWORD [XMMBLOCK(4,0,edx,SIZEOF_ISLOW_MULT_TYPE)] | |
189 pmullw xmm3, XMMWORD [XMMBLOCK(6,0,edx,SIZEOF_ISLOW_MULT_TYPE)] | |
190 | |
191 ; (Original) | |
192 ; z1 = (z2 + z3) * 0.541196100; | |
193 ; tmp2 = z1 + z3 * -1.847759065; | |
194 ; tmp3 = z1 + z2 * 0.765366865; | |
195 ; | |
196 ; (This implementation) | |
197 ; tmp2 = z2 * 0.541196100 + z3 * (0.541196100 - 1.847759065); | |
198 ; tmp3 = z2 * (0.541196100 + 0.765366865) + z3 * 0.541196100; | |
199 | |
200 movdqa xmm4,xmm1 ; xmm1=in2=z2 | |
201 movdqa xmm5,xmm1 | |
202 punpcklwd xmm4,xmm3 ; xmm3=in6=z3 | |
203 punpckhwd xmm5,xmm3 | |
204 movdqa xmm1,xmm4 | |
205 movdqa xmm3,xmm5 | |
206 pmaddwd xmm4,[GOTOFF(ebx,PW_F130_F054)] ; xmm4=tmp3L | |
207 pmaddwd xmm5,[GOTOFF(ebx,PW_F130_F054)] ; xmm5=tmp3H | |
208 pmaddwd xmm1,[GOTOFF(ebx,PW_F054_MF130)] ; xmm1=tmp2L | |
209 pmaddwd xmm3,[GOTOFF(ebx,PW_F054_MF130)] ; xmm3=tmp2H | |
210 | |
211 movdqa xmm6,xmm0 | |
212 paddw xmm0,xmm2 ; xmm0=in0+in4 | |
213 psubw xmm6,xmm2 ; xmm6=in0-in4 | |
214 | |
215 pxor xmm7,xmm7 | |
216 pxor xmm2,xmm2 | |
217 punpcklwd xmm7,xmm0 ; xmm7=tmp0L | |
218 punpckhwd xmm2,xmm0 ; xmm2=tmp0H | |
219 psrad xmm7,(16-CONST_BITS) ; psrad xmm7,16 & pslld xmm7,CONST_BITS | |
220 psrad xmm2,(16-CONST_BITS) ; psrad xmm2,16 & pslld xmm2,CONST_BITS | |
221 | |
222 movdqa xmm0,xmm7 | |
223 paddd xmm7,xmm4 ; xmm7=tmp10L | |
224 psubd xmm0,xmm4 ; xmm0=tmp13L | |
225 movdqa xmm4,xmm2 | |
226 paddd xmm2,xmm5 ; xmm2=tmp10H | |
227 psubd xmm4,xmm5 ; xmm4=tmp13H | |
228 | |
229 movdqa XMMWORD [wk(0)], xmm7 ; wk(0)=tmp10L | |
230 movdqa XMMWORD [wk(1)], xmm2 ; wk(1)=tmp10H | |
231 movdqa XMMWORD [wk(2)], xmm0 ; wk(2)=tmp13L | |
232 movdqa XMMWORD [wk(3)], xmm4 ; wk(3)=tmp13H | |
233 | |
234 pxor xmm5,xmm5 | |
235 pxor xmm7,xmm7 | |
236 punpcklwd xmm5,xmm6 ; xmm5=tmp1L | |
237 punpckhwd xmm7,xmm6 ; xmm7=tmp1H | |
238 psrad xmm5,(16-CONST_BITS) ; psrad xmm5,16 & pslld xmm5,CONST_BITS | |
239 psrad xmm7,(16-CONST_BITS) ; psrad xmm7,16 & pslld xmm7,CONST_BITS | |
240 | |
241 movdqa xmm2,xmm5 | |
242 paddd xmm5,xmm1 ; xmm5=tmp11L | |
243 psubd xmm2,xmm1 ; xmm2=tmp12L | |
244 movdqa xmm0,xmm7 | |
245 paddd xmm7,xmm3 ; xmm7=tmp11H | |
246 psubd xmm0,xmm3 ; xmm0=tmp12H | |
247 | |
248 movdqa XMMWORD [wk(4)], xmm5 ; wk(4)=tmp11L | |
249 movdqa XMMWORD [wk(5)], xmm7 ; wk(5)=tmp11H | |
250 movdqa XMMWORD [wk(6)], xmm2 ; wk(6)=tmp12L | |
251 movdqa XMMWORD [wk(7)], xmm0 ; wk(7)=tmp12H | |
252 | |
253 ; -- Odd part | |
254 | |
255 movdqa xmm4, XMMWORD [XMMBLOCK(1,0,esi,SIZEOF_JCOEF)] | |
256 movdqa xmm6, XMMWORD [XMMBLOCK(3,0,esi,SIZEOF_JCOEF)] | |
257 pmullw xmm4, XMMWORD [XMMBLOCK(1,0,edx,SIZEOF_ISLOW_MULT_TYPE)] | |
258 pmullw xmm6, XMMWORD [XMMBLOCK(3,0,edx,SIZEOF_ISLOW_MULT_TYPE)] | |
259 movdqa xmm1, XMMWORD [XMMBLOCK(5,0,esi,SIZEOF_JCOEF)] | |
260 movdqa xmm3, XMMWORD [XMMBLOCK(7,0,esi,SIZEOF_JCOEF)] | |
261 pmullw xmm1, XMMWORD [XMMBLOCK(5,0,edx,SIZEOF_ISLOW_MULT_TYPE)] | |
262 pmullw xmm3, XMMWORD [XMMBLOCK(7,0,edx,SIZEOF_ISLOW_MULT_TYPE)] | |
263 | |
264 movdqa xmm5,xmm6 | |
265 movdqa xmm7,xmm4 | |
266 paddw xmm5,xmm3 ; xmm5=z3 | |
267 paddw xmm7,xmm1 ; xmm7=z4 | |
268 | |
269 ; (Original) | |
270 ; z5 = (z3 + z4) * 1.175875602; | |
271 ; z3 = z3 * -1.961570560; z4 = z4 * -0.390180644; | |
272 ; z3 += z5; z4 += z5; | |
273 ; | |
274 ; (This implementation) | |
275 ; z3 = z3 * (1.175875602 - 1.961570560) + z4 * 1.175875602; | |
276 ; z4 = z3 * 1.175875602 + z4 * (1.175875602 - 0.390180644); | |
277 | |
278 movdqa xmm2,xmm5 | |
279 movdqa xmm0,xmm5 | |
280 punpcklwd xmm2,xmm7 | |
281 punpckhwd xmm0,xmm7 | |
282 movdqa xmm5,xmm2 | |
283 movdqa xmm7,xmm0 | |
284 pmaddwd xmm2,[GOTOFF(ebx,PW_MF078_F117)] ; xmm2=z3L | |
285 pmaddwd xmm0,[GOTOFF(ebx,PW_MF078_F117)] ; xmm0=z3H | |
286 pmaddwd xmm5,[GOTOFF(ebx,PW_F117_F078)] ; xmm5=z4L | |
287 pmaddwd xmm7,[GOTOFF(ebx,PW_F117_F078)] ; xmm7=z4H | |
288 | |
289 movdqa XMMWORD [wk(10)], xmm2 ; wk(10)=z3L | |
290 movdqa XMMWORD [wk(11)], xmm0 ; wk(11)=z3H | |
291 | |
292 ; (Original) | |
293 ; z1 = tmp0 + tmp3; z2 = tmp1 + tmp2; | |
294 ; tmp0 = tmp0 * 0.298631336; tmp1 = tmp1 * 2.053119869; | |
295 ; tmp2 = tmp2 * 3.072711026; tmp3 = tmp3 * 1.501321110; | |
296 ; z1 = z1 * -0.899976223; z2 = z2 * -2.562915447; | |
297 ; tmp0 += z1 + z3; tmp1 += z2 + z4; | |
298 ; tmp2 += z2 + z3; tmp3 += z1 + z4; | |
299 ; | |
300 ; (This implementation) | |
301 ; tmp0 = tmp0 * (0.298631336 - 0.899976223) + tmp3 * -0.899976223; | |
302 ; tmp1 = tmp1 * (2.053119869 - 2.562915447) + tmp2 * -2.562915447; | |
303 ; tmp2 = tmp1 * -2.562915447 + tmp2 * (3.072711026 - 2.562915447); | |
304 ; tmp3 = tmp0 * -0.899976223 + tmp3 * (1.501321110 - 0.899976223); | |
305 ; tmp0 += z3; tmp1 += z4; | |
306 ; tmp2 += z3; tmp3 += z4; | |
307 | |
308 movdqa xmm2,xmm3 | |
309 movdqa xmm0,xmm3 | |
310 punpcklwd xmm2,xmm4 | |
311 punpckhwd xmm0,xmm4 | |
312 movdqa xmm3,xmm2 | |
313 movdqa xmm4,xmm0 | |
314 pmaddwd xmm2,[GOTOFF(ebx,PW_MF060_MF089)] ; xmm2=tmp0L | |
315 pmaddwd xmm0,[GOTOFF(ebx,PW_MF060_MF089)] ; xmm0=tmp0H | |
316 pmaddwd xmm3,[GOTOFF(ebx,PW_MF089_F060)] ; xmm3=tmp3L | |
317 pmaddwd xmm4,[GOTOFF(ebx,PW_MF089_F060)] ; xmm4=tmp3H | |
318 | |
319 paddd xmm2, XMMWORD [wk(10)] ; xmm2=tmp0L | |
320 paddd xmm0, XMMWORD [wk(11)] ; xmm0=tmp0H | |
321 paddd xmm3,xmm5 ; xmm3=tmp3L | |
322 paddd xmm4,xmm7 ; xmm4=tmp3H | |
323 | |
324 movdqa XMMWORD [wk(8)], xmm2 ; wk(8)=tmp0L | |
325 movdqa XMMWORD [wk(9)], xmm0 ; wk(9)=tmp0H | |
326 | |
327 movdqa xmm2,xmm1 | |
328 movdqa xmm0,xmm1 | |
329 punpcklwd xmm2,xmm6 | |
330 punpckhwd xmm0,xmm6 | |
331 movdqa xmm1,xmm2 | |
332 movdqa xmm6,xmm0 | |
333 pmaddwd xmm2,[GOTOFF(ebx,PW_MF050_MF256)] ; xmm2=tmp1L | |
334 pmaddwd xmm0,[GOTOFF(ebx,PW_MF050_MF256)] ; xmm0=tmp1H | |
335 pmaddwd xmm1,[GOTOFF(ebx,PW_MF256_F050)] ; xmm1=tmp2L | |
336 pmaddwd xmm6,[GOTOFF(ebx,PW_MF256_F050)] ; xmm6=tmp2H | |
337 | |
338 paddd xmm2,xmm5 ; xmm2=tmp1L | |
339 paddd xmm0,xmm7 ; xmm0=tmp1H | |
340 paddd xmm1, XMMWORD [wk(10)] ; xmm1=tmp2L | |
341 paddd xmm6, XMMWORD [wk(11)] ; xmm6=tmp2H | |
342 | |
343 movdqa XMMWORD [wk(10)], xmm2 ; wk(10)=tmp1L | |
344 movdqa XMMWORD [wk(11)], xmm0 ; wk(11)=tmp1H | |
345 | |
346 ; -- Final output stage | |
347 | |
348 movdqa xmm5, XMMWORD [wk(0)] ; xmm5=tmp10L | |
349 movdqa xmm7, XMMWORD [wk(1)] ; xmm7=tmp10H | |
350 | |
351 movdqa xmm2,xmm5 | |
352 movdqa xmm0,xmm7 | |
353 paddd xmm5,xmm3 ; xmm5=data0L | |
354 paddd xmm7,xmm4 ; xmm7=data0H | |
355 psubd xmm2,xmm3 ; xmm2=data7L | |
356 psubd xmm0,xmm4 ; xmm0=data7H | |
357 | |
358 movdqa xmm3,[GOTOFF(ebx,PD_DESCALE_P1)] ; xmm3=[PD_DESCALE_P1] | |
359 | |
360 paddd xmm5,xmm3 | |
361 paddd xmm7,xmm3 | |
362 psrad xmm5,DESCALE_P1 | |
363 psrad xmm7,DESCALE_P1 | |
364 paddd xmm2,xmm3 | |
365 paddd xmm0,xmm3 | |
366 psrad xmm2,DESCALE_P1 | |
367 psrad xmm0,DESCALE_P1 | |
368 | |
369 packssdw xmm5,xmm7 ; xmm5=data0=(00 01 02 03 04 05 06 07) | |
370 packssdw xmm2,xmm0 ; xmm2=data7=(70 71 72 73 74 75 76 77) | |
371 | |
372 movdqa xmm4, XMMWORD [wk(4)] ; xmm4=tmp11L | |
373 movdqa xmm3, XMMWORD [wk(5)] ; xmm3=tmp11H | |
374 | |
375 movdqa xmm7,xmm4 | |
376 movdqa xmm0,xmm3 | |
377 paddd xmm4,xmm1 ; xmm4=data1L | |
378 paddd xmm3,xmm6 ; xmm3=data1H | |
379 psubd xmm7,xmm1 ; xmm7=data6L | |
380 psubd xmm0,xmm6 ; xmm0=data6H | |
381 | |
382 movdqa xmm1,[GOTOFF(ebx,PD_DESCALE_P1)] ; xmm1=[PD_DESCALE_P1] | |
383 | |
384 paddd xmm4,xmm1 | |
385 paddd xmm3,xmm1 | |
386 psrad xmm4,DESCALE_P1 | |
387 psrad xmm3,DESCALE_P1 | |
388 paddd xmm7,xmm1 | |
389 paddd xmm0,xmm1 | |
390 psrad xmm7,DESCALE_P1 | |
391 psrad xmm0,DESCALE_P1 | |
392 | |
393 packssdw xmm4,xmm3 ; xmm4=data1=(10 11 12 13 14 15 16 17) | |
394 packssdw xmm7,xmm0 ; xmm7=data6=(60 61 62 63 64 65 66 67) | |
395 | |
396 movdqa xmm6,xmm5 ; transpose coefficients(phase 1) | |
397 punpcklwd xmm5,xmm4 ; xmm5=(00 10 01 11 02 12 03 13) | |
398 punpckhwd xmm6,xmm4 ; xmm6=(04 14 05 15 06 16 07 17) | |
399 movdqa xmm1,xmm7 ; transpose coefficients(phase 1) | |
400 punpcklwd xmm7,xmm2 ; xmm7=(60 70 61 71 62 72 63 73) | |
401 punpckhwd xmm1,xmm2 ; xmm1=(64 74 65 75 66 76 67 77) | |
402 | |
403 movdqa xmm3, XMMWORD [wk(6)] ; xmm3=tmp12L | |
404 movdqa xmm0, XMMWORD [wk(7)] ; xmm0=tmp12H | |
405 movdqa xmm4, XMMWORD [wk(10)] ; xmm4=tmp1L | |
406 movdqa xmm2, XMMWORD [wk(11)] ; xmm2=tmp1H | |
407 | |
408 movdqa XMMWORD [wk(0)], xmm5 ; wk(0)=(00 10 01 11 02 12 03 13) | |
409 movdqa XMMWORD [wk(1)], xmm6 ; wk(1)=(04 14 05 15 06 16 07 17) | |
410 movdqa XMMWORD [wk(4)], xmm7 ; wk(4)=(60 70 61 71 62 72 63 73) | |
411 movdqa XMMWORD [wk(5)], xmm1 ; wk(5)=(64 74 65 75 66 76 67 77) | |
412 | |
413 movdqa xmm5,xmm3 | |
414 movdqa xmm6,xmm0 | |
415 paddd xmm3,xmm4 ; xmm3=data2L | |
416 paddd xmm0,xmm2 ; xmm0=data2H | |
417 psubd xmm5,xmm4 ; xmm5=data5L | |
418 psubd xmm6,xmm2 ; xmm6=data5H | |
419 | |
420 movdqa xmm7,[GOTOFF(ebx,PD_DESCALE_P1)] ; xmm7=[PD_DESCALE_P1] | |
421 | |
422 paddd xmm3,xmm7 | |
423 paddd xmm0,xmm7 | |
424 psrad xmm3,DESCALE_P1 | |
425 psrad xmm0,DESCALE_P1 | |
426 paddd xmm5,xmm7 | |
427 paddd xmm6,xmm7 | |
428 psrad xmm5,DESCALE_P1 | |
429 psrad xmm6,DESCALE_P1 | |
430 | |
431 packssdw xmm3,xmm0 ; xmm3=data2=(20 21 22 23 24 25 26 27) | |
432 packssdw xmm5,xmm6 ; xmm5=data5=(50 51 52 53 54 55 56 57) | |
433 | |
434 movdqa xmm1, XMMWORD [wk(2)] ; xmm1=tmp13L | |
435 movdqa xmm4, XMMWORD [wk(3)] ; xmm4=tmp13H | |
436 movdqa xmm2, XMMWORD [wk(8)] ; xmm2=tmp0L | |
437 movdqa xmm7, XMMWORD [wk(9)] ; xmm7=tmp0H | |
438 | |
439 movdqa xmm0,xmm1 | |
440 movdqa xmm6,xmm4 | |
441 paddd xmm1,xmm2 ; xmm1=data3L | |
442 paddd xmm4,xmm7 ; xmm4=data3H | |
443 psubd xmm0,xmm2 ; xmm0=data4L | |
444 psubd xmm6,xmm7 ; xmm6=data4H | |
445 | |
446 movdqa xmm2,[GOTOFF(ebx,PD_DESCALE_P1)] ; xmm2=[PD_DESCALE_P1] | |
447 | |
448 paddd xmm1,xmm2 | |
449 paddd xmm4,xmm2 | |
450 psrad xmm1,DESCALE_P1 | |
451 psrad xmm4,DESCALE_P1 | |
452 paddd xmm0,xmm2 | |
453 paddd xmm6,xmm2 | |
454 psrad xmm0,DESCALE_P1 | |
455 psrad xmm6,DESCALE_P1 | |
456 | |
457 packssdw xmm1,xmm4 ; xmm1=data3=(30 31 32 33 34 35 36 37) | |
458 packssdw xmm0,xmm6 ; xmm0=data4=(40 41 42 43 44 45 46 47) | |
459 | |
460 movdqa xmm7, XMMWORD [wk(0)] ; xmm7=(00 10 01 11 02 12 03 13) | |
461 movdqa xmm2, XMMWORD [wk(1)] ; xmm2=(04 14 05 15 06 16 07 17) | |
462 | |
463 movdqa xmm4,xmm3 ; transpose coefficients(phase 1) | |
464 punpcklwd xmm3,xmm1 ; xmm3=(20 30 21 31 22 32 23 33) | |
465 punpckhwd xmm4,xmm1 ; xmm4=(24 34 25 35 26 36 27 37) | |
466 movdqa xmm6,xmm0 ; transpose coefficients(phase 1) | |
467 punpcklwd xmm0,xmm5 ; xmm0=(40 50 41 51 42 52 43 53) | |
468 punpckhwd xmm6,xmm5 ; xmm6=(44 54 45 55 46 56 47 57) | |
469 | |
470 movdqa xmm1,xmm7 ; transpose coefficients(phase 2) | |
471 punpckldq xmm7,xmm3 ; xmm7=(00 10 20 30 01 11 21 31) | |
472 punpckhdq xmm1,xmm3 ; xmm1=(02 12 22 32 03 13 23 33) | |
473 movdqa xmm5,xmm2 ; transpose coefficients(phase 2) | |
474 punpckldq xmm2,xmm4 ; xmm2=(04 14 24 34 05 15 25 35) | |
475 punpckhdq xmm5,xmm4 ; xmm5=(06 16 26 36 07 17 27 37) | |
476 | |
477 movdqa xmm3, XMMWORD [wk(4)] ; xmm3=(60 70 61 71 62 72 63 73) | |
478 movdqa xmm4, XMMWORD [wk(5)] ; xmm4=(64 74 65 75 66 76 67 77) | |
479 | |
480 movdqa XMMWORD [wk(6)], xmm2 ; wk(6)=(04 14 24 34 05 15 25 35) | |
481 movdqa XMMWORD [wk(7)], xmm5 ; wk(7)=(06 16 26 36 07 17 27 37) | |
482 | |
483 movdqa xmm2,xmm0 ; transpose coefficients(phase 2) | |
484 punpckldq xmm0,xmm3 ; xmm0=(40 50 60 70 41 51 61 71) | |
485 punpckhdq xmm2,xmm3 ; xmm2=(42 52 62 72 43 53 63 73) | |
486 movdqa xmm5,xmm6 ; transpose coefficients(phase 2) | |
487 punpckldq xmm6,xmm4 ; xmm6=(44 54 64 74 45 55 65 75) | |
488 punpckhdq xmm5,xmm4 ; xmm5=(46 56 66 76 47 57 67 77) | |
489 | |
490 movdqa xmm3,xmm7 ; transpose coefficients(phase 3) | |
491 punpcklqdq xmm7,xmm0 ; xmm7=col0=(00 10 20 30 40 50 60 70) | |
492 punpckhqdq xmm3,xmm0 ; xmm3=col1=(01 11 21 31 41 51 61 71) | |
493 movdqa xmm4,xmm1 ; transpose coefficients(phase 3) | |
494 punpcklqdq xmm1,xmm2 ; xmm1=col2=(02 12 22 32 42 52 62 72) | |
495 punpckhqdq xmm4,xmm2 ; xmm4=col3=(03 13 23 33 43 53 63 73) | |
496 | |
497 movdqa xmm0, XMMWORD [wk(6)] ; xmm0=(04 14 24 34 05 15 25 35) | |
498 movdqa xmm2, XMMWORD [wk(7)] ; xmm2=(06 16 26 36 07 17 27 37) | |
499 | |
500 movdqa XMMWORD [wk(8)], xmm3 ; wk(8)=col1 | |
501 movdqa XMMWORD [wk(9)], xmm4 ; wk(9)=col3 | |
502 | |
503 movdqa xmm3,xmm0 ; transpose coefficients(phase 3) | |
504 punpcklqdq xmm0,xmm6 ; xmm0=col4=(04 14 24 34 44 54 64 74) | |
505 punpckhqdq xmm3,xmm6 ; xmm3=col5=(05 15 25 35 45 55 65 75) | |
506 movdqa xmm4,xmm2 ; transpose coefficients(phase 3) | |
507 punpcklqdq xmm2,xmm5 ; xmm2=col6=(06 16 26 36 46 56 66 76) | |
508 punpckhqdq xmm4,xmm5 ; xmm4=col7=(07 17 27 37 47 57 67 77) | |
509 | |
510 movdqa XMMWORD [wk(10)], xmm3 ; wk(10)=col5 | |
511 movdqa XMMWORD [wk(11)], xmm4 ; wk(11)=col7 | |
512 .column_end: | |
513 | |
514 ; -- Prefetch the next coefficient block | |
515 | |
516 prefetchnta [esi + DCTSIZE2*SIZEOF_JCOEF + 0*32] | |
517 prefetchnta [esi + DCTSIZE2*SIZEOF_JCOEF + 1*32] | |
518 prefetchnta [esi + DCTSIZE2*SIZEOF_JCOEF + 2*32] | |
519 prefetchnta [esi + DCTSIZE2*SIZEOF_JCOEF + 3*32] | |
520 | |
521 ; ---- Pass 2: process rows from work array, store into output array. | |
522 | |
523 mov eax, [original_ebp] | |
524 mov edi, JSAMPARRAY [output_buf(eax)] ; (JSAMPROW *) | |
525 mov eax, JDIMENSION [output_col(eax)] | |
526 | |
527 ; -- Even part | |
528 | |
529 ; xmm7=col0, xmm1=col2, xmm0=col4, xmm2=col6 | |
530 | |
531 ; (Original) | |
532 ; z1 = (z2 + z3) * 0.541196100; | |
533 ; tmp2 = z1 + z3 * -1.847759065; | |
534 ; tmp3 = z1 + z2 * 0.765366865; | |
535 ; | |
536 ; (This implementation) | |
537 ; tmp2 = z2 * 0.541196100 + z3 * (0.541196100 - 1.847759065); | |
538 ; tmp3 = z2 * (0.541196100 + 0.765366865) + z3 * 0.541196100; | |
539 | |
540 movdqa xmm6,xmm1 ; xmm1=in2=z2 | |
541 movdqa xmm5,xmm1 | |
542 punpcklwd xmm6,xmm2 ; xmm2=in6=z3 | |
543 punpckhwd xmm5,xmm2 | |
544 movdqa xmm1,xmm6 | |
545 movdqa xmm2,xmm5 | |
546 pmaddwd xmm6,[GOTOFF(ebx,PW_F130_F054)] ; xmm6=tmp3L | |
547 pmaddwd xmm5,[GOTOFF(ebx,PW_F130_F054)] ; xmm5=tmp3H | |
548 pmaddwd xmm1,[GOTOFF(ebx,PW_F054_MF130)] ; xmm1=tmp2L | |
549 pmaddwd xmm2,[GOTOFF(ebx,PW_F054_MF130)] ; xmm2=tmp2H | |
550 | |
551 movdqa xmm3,xmm7 | |
552 paddw xmm7,xmm0 ; xmm7=in0+in4 | |
553 psubw xmm3,xmm0 ; xmm3=in0-in4 | |
554 | |
555 pxor xmm4,xmm4 | |
556 pxor xmm0,xmm0 | |
557 punpcklwd xmm4,xmm7 ; xmm4=tmp0L | |
558 punpckhwd xmm0,xmm7 ; xmm0=tmp0H | |
559 psrad xmm4,(16-CONST_BITS) ; psrad xmm4,16 & pslld xmm4,CONST_BITS | |
560 psrad xmm0,(16-CONST_BITS) ; psrad xmm0,16 & pslld xmm0,CONST_BITS | |
561 | |
562 movdqa xmm7,xmm4 | |
563 paddd xmm4,xmm6 ; xmm4=tmp10L | |
564 psubd xmm7,xmm6 ; xmm7=tmp13L | |
565 movdqa xmm6,xmm0 | |
566 paddd xmm0,xmm5 ; xmm0=tmp10H | |
567 psubd xmm6,xmm5 ; xmm6=tmp13H | |
568 | |
569 movdqa XMMWORD [wk(0)], xmm4 ; wk(0)=tmp10L | |
570 movdqa XMMWORD [wk(1)], xmm0 ; wk(1)=tmp10H | |
571 movdqa XMMWORD [wk(2)], xmm7 ; wk(2)=tmp13L | |
572 movdqa XMMWORD [wk(3)], xmm6 ; wk(3)=tmp13H | |
573 | |
574 pxor xmm5,xmm5 | |
575 pxor xmm4,xmm4 | |
576 punpcklwd xmm5,xmm3 ; xmm5=tmp1L | |
577 punpckhwd xmm4,xmm3 ; xmm4=tmp1H | |
578 psrad xmm5,(16-CONST_BITS) ; psrad xmm5,16 & pslld xmm5,CONST_BITS | |
579 psrad xmm4,(16-CONST_BITS) ; psrad xmm4,16 & pslld xmm4,CONST_BITS | |
580 | |
581 movdqa xmm0,xmm5 | |
582 paddd xmm5,xmm1 ; xmm5=tmp11L | |
583 psubd xmm0,xmm1 ; xmm0=tmp12L | |
584 movdqa xmm7,xmm4 | |
585 paddd xmm4,xmm2 ; xmm4=tmp11H | |
586 psubd xmm7,xmm2 ; xmm7=tmp12H | |
587 | |
588 movdqa XMMWORD [wk(4)], xmm5 ; wk(4)=tmp11L | |
589 movdqa XMMWORD [wk(5)], xmm4 ; wk(5)=tmp11H | |
590 movdqa XMMWORD [wk(6)], xmm0 ; wk(6)=tmp12L | |
591 movdqa XMMWORD [wk(7)], xmm7 ; wk(7)=tmp12H | |
592 | |
593 ; -- Odd part | |
594 | |
595 movdqa xmm6, XMMWORD [wk(9)] ; xmm6=col3 | |
596 movdqa xmm3, XMMWORD [wk(8)] ; xmm3=col1 | |
597 movdqa xmm1, XMMWORD [wk(11)] ; xmm1=col7 | |
598 movdqa xmm2, XMMWORD [wk(10)] ; xmm2=col5 | |
599 | |
600 movdqa xmm5,xmm6 | |
601 movdqa xmm4,xmm3 | |
602 paddw xmm5,xmm1 ; xmm5=z3 | |
603 paddw xmm4,xmm2 ; xmm4=z4 | |
604 | |
605 ; (Original) | |
606 ; z5 = (z3 + z4) * 1.175875602; | |
607 ; z3 = z3 * -1.961570560; z4 = z4 * -0.390180644; | |
608 ; z3 += z5; z4 += z5; | |
609 ; | |
610 ; (This implementation) | |
611 ; z3 = z3 * (1.175875602 - 1.961570560) + z4 * 1.175875602; | |
612 ; z4 = z3 * 1.175875602 + z4 * (1.175875602 - 0.390180644); | |
613 | |
614 movdqa xmm0,xmm5 | |
615 movdqa xmm7,xmm5 | |
616 punpcklwd xmm0,xmm4 | |
617 punpckhwd xmm7,xmm4 | |
618 movdqa xmm5,xmm0 | |
619 movdqa xmm4,xmm7 | |
620 pmaddwd xmm0,[GOTOFF(ebx,PW_MF078_F117)] ; xmm0=z3L | |
621 pmaddwd xmm7,[GOTOFF(ebx,PW_MF078_F117)] ; xmm7=z3H | |
622 pmaddwd xmm5,[GOTOFF(ebx,PW_F117_F078)] ; xmm5=z4L | |
623 pmaddwd xmm4,[GOTOFF(ebx,PW_F117_F078)] ; xmm4=z4H | |
624 | |
625 movdqa XMMWORD [wk(10)], xmm0 ; wk(10)=z3L | |
626 movdqa XMMWORD [wk(11)], xmm7 ; wk(11)=z3H | |
627 | |
628 ; (Original) | |
629 ; z1 = tmp0 + tmp3; z2 = tmp1 + tmp2; | |
630 ; tmp0 = tmp0 * 0.298631336; tmp1 = tmp1 * 2.053119869; | |
631 ; tmp2 = tmp2 * 3.072711026; tmp3 = tmp3 * 1.501321110; | |
632 ; z1 = z1 * -0.899976223; z2 = z2 * -2.562915447; | |
633 ; tmp0 += z1 + z3; tmp1 += z2 + z4; | |
634 ; tmp2 += z2 + z3; tmp3 += z1 + z4; | |
635 ; | |
636 ; (This implementation) | |
637 ; tmp0 = tmp0 * (0.298631336 - 0.899976223) + tmp3 * -0.899976223; | |
638 ; tmp1 = tmp1 * (2.053119869 - 2.562915447) + tmp2 * -2.562915447; | |
639 ; tmp2 = tmp1 * -2.562915447 + tmp2 * (3.072711026 - 2.562915447); | |
640 ; tmp3 = tmp0 * -0.899976223 + tmp3 * (1.501321110 - 0.899976223); | |
641 ; tmp0 += z3; tmp1 += z4; | |
642 ; tmp2 += z3; tmp3 += z4; | |
643 | |
644 movdqa xmm0,xmm1 | |
645 movdqa xmm7,xmm1 | |
646 punpcklwd xmm0,xmm3 | |
647 punpckhwd xmm7,xmm3 | |
648 movdqa xmm1,xmm0 | |
649 movdqa xmm3,xmm7 | |
650 pmaddwd xmm0,[GOTOFF(ebx,PW_MF060_MF089)] ; xmm0=tmp0L | |
651 pmaddwd xmm7,[GOTOFF(ebx,PW_MF060_MF089)] ; xmm7=tmp0H | |
652 pmaddwd xmm1,[GOTOFF(ebx,PW_MF089_F060)] ; xmm1=tmp3L | |
653 pmaddwd xmm3,[GOTOFF(ebx,PW_MF089_F060)] ; xmm3=tmp3H | |
654 | |
655 paddd xmm0, XMMWORD [wk(10)] ; xmm0=tmp0L | |
656 paddd xmm7, XMMWORD [wk(11)] ; xmm7=tmp0H | |
657 paddd xmm1,xmm5 ; xmm1=tmp3L | |
658 paddd xmm3,xmm4 ; xmm3=tmp3H | |
659 | |
660 movdqa XMMWORD [wk(8)], xmm0 ; wk(8)=tmp0L | |
661 movdqa XMMWORD [wk(9)], xmm7 ; wk(9)=tmp0H | |
662 | |
663 movdqa xmm0,xmm2 | |
664 movdqa xmm7,xmm2 | |
665 punpcklwd xmm0,xmm6 | |
666 punpckhwd xmm7,xmm6 | |
667 movdqa xmm2,xmm0 | |
668 movdqa xmm6,xmm7 | |
669 pmaddwd xmm0,[GOTOFF(ebx,PW_MF050_MF256)] ; xmm0=tmp1L | |
670 pmaddwd xmm7,[GOTOFF(ebx,PW_MF050_MF256)] ; xmm7=tmp1H | |
671 pmaddwd xmm2,[GOTOFF(ebx,PW_MF256_F050)] ; xmm2=tmp2L | |
672 pmaddwd xmm6,[GOTOFF(ebx,PW_MF256_F050)] ; xmm6=tmp2H | |
673 | |
674 paddd xmm0,xmm5 ; xmm0=tmp1L | |
675 paddd xmm7,xmm4 ; xmm7=tmp1H | |
676 paddd xmm2, XMMWORD [wk(10)] ; xmm2=tmp2L | |
677 paddd xmm6, XMMWORD [wk(11)] ; xmm6=tmp2H | |
678 | |
679 movdqa XMMWORD [wk(10)], xmm0 ; wk(10)=tmp1L | |
680 movdqa XMMWORD [wk(11)], xmm7 ; wk(11)=tmp1H | |
681 | |
682 ; -- Final output stage | |
683 | |
684 movdqa xmm5, XMMWORD [wk(0)] ; xmm5=tmp10L | |
685 movdqa xmm4, XMMWORD [wk(1)] ; xmm4=tmp10H | |
686 | |
687 movdqa xmm0,xmm5 | |
688 movdqa xmm7,xmm4 | |
689 paddd xmm5,xmm1 ; xmm5=data0L | |
690 paddd xmm4,xmm3 ; xmm4=data0H | |
691 psubd xmm0,xmm1 ; xmm0=data7L | |
692 psubd xmm7,xmm3 ; xmm7=data7H | |
693 | |
694 movdqa xmm1,[GOTOFF(ebx,PD_DESCALE_P2)] ; xmm1=[PD_DESCALE_P2] | |
695 | |
696 paddd xmm5,xmm1 | |
697 paddd xmm4,xmm1 | |
698 psrad xmm5,DESCALE_P2 | |
699 psrad xmm4,DESCALE_P2 | |
700 paddd xmm0,xmm1 | |
701 paddd xmm7,xmm1 | |
702 psrad xmm0,DESCALE_P2 | |
703 psrad xmm7,DESCALE_P2 | |
704 | |
705 packssdw xmm5,xmm4 ; xmm5=data0=(00 10 20 30 40 50 60 70) | |
706 packssdw xmm0,xmm7 ; xmm0=data7=(07 17 27 37 47 57 67 77) | |
707 | |
708 movdqa xmm3, XMMWORD [wk(4)] ; xmm3=tmp11L | |
709 movdqa xmm1, XMMWORD [wk(5)] ; xmm1=tmp11H | |
710 | |
711 movdqa xmm4,xmm3 | |
712 movdqa xmm7,xmm1 | |
713 paddd xmm3,xmm2 ; xmm3=data1L | |
714 paddd xmm1,xmm6 ; xmm1=data1H | |
715 psubd xmm4,xmm2 ; xmm4=data6L | |
716 psubd xmm7,xmm6 ; xmm7=data6H | |
717 | |
718 movdqa xmm2,[GOTOFF(ebx,PD_DESCALE_P2)] ; xmm2=[PD_DESCALE_P2] | |
719 | |
720 paddd xmm3,xmm2 | |
721 paddd xmm1,xmm2 | |
722 psrad xmm3,DESCALE_P2 | |
723 psrad xmm1,DESCALE_P2 | |
724 paddd xmm4,xmm2 | |
725 paddd xmm7,xmm2 | |
726 psrad xmm4,DESCALE_P2 | |
727 psrad xmm7,DESCALE_P2 | |
728 | |
729 packssdw xmm3,xmm1 ; xmm3=data1=(01 11 21 31 41 51 61 71) | |
730 packssdw xmm4,xmm7 ; xmm4=data6=(06 16 26 36 46 56 66 76) | |
731 | |
732 packsswb xmm5,xmm4 ; xmm5=(00 10 20 30 40 50 60 70 06 16 26
36 46 56 66 76) | |
733 packsswb xmm3,xmm0 ; xmm3=(01 11 21 31 41 51 61 71 07 17 27
37 47 57 67 77) | |
734 | |
735 movdqa xmm6, XMMWORD [wk(6)] ; xmm6=tmp12L | |
736 movdqa xmm2, XMMWORD [wk(7)] ; xmm2=tmp12H | |
737 movdqa xmm1, XMMWORD [wk(10)] ; xmm1=tmp1L | |
738 movdqa xmm7, XMMWORD [wk(11)] ; xmm7=tmp1H | |
739 | |
740 movdqa XMMWORD [wk(0)], xmm5 ; wk(0)=(00 10 20 30 40 50 60 70 06 16 2
6 36 46 56 66 76) | |
741 movdqa XMMWORD [wk(1)], xmm3 ; wk(1)=(01 11 21 31 41 51 61 71 07 17 2
7 37 47 57 67 77) | |
742 | |
743 movdqa xmm4,xmm6 | |
744 movdqa xmm0,xmm2 | |
745 paddd xmm6,xmm1 ; xmm6=data2L | |
746 paddd xmm2,xmm7 ; xmm2=data2H | |
747 psubd xmm4,xmm1 ; xmm4=data5L | |
748 psubd xmm0,xmm7 ; xmm0=data5H | |
749 | |
750 movdqa xmm5,[GOTOFF(ebx,PD_DESCALE_P2)] ; xmm5=[PD_DESCALE_P2] | |
751 | |
752 paddd xmm6,xmm5 | |
753 paddd xmm2,xmm5 | |
754 psrad xmm6,DESCALE_P2 | |
755 psrad xmm2,DESCALE_P2 | |
756 paddd xmm4,xmm5 | |
757 paddd xmm0,xmm5 | |
758 psrad xmm4,DESCALE_P2 | |
759 psrad xmm0,DESCALE_P2 | |
760 | |
761 packssdw xmm6,xmm2 ; xmm6=data2=(02 12 22 32 42 52 62 72) | |
762 packssdw xmm4,xmm0 ; xmm4=data5=(05 15 25 35 45 55 65 75) | |
763 | |
764 movdqa xmm3, XMMWORD [wk(2)] ; xmm3=tmp13L | |
765 movdqa xmm1, XMMWORD [wk(3)] ; xmm1=tmp13H | |
766 movdqa xmm7, XMMWORD [wk(8)] ; xmm7=tmp0L | |
767 movdqa xmm5, XMMWORD [wk(9)] ; xmm5=tmp0H | |
768 | |
769 movdqa xmm2,xmm3 | |
770 movdqa xmm0,xmm1 | |
771 paddd xmm3,xmm7 ; xmm3=data3L | |
772 paddd xmm1,xmm5 ; xmm1=data3H | |
773 psubd xmm2,xmm7 ; xmm2=data4L | |
774 psubd xmm0,xmm5 ; xmm0=data4H | |
775 | |
776 movdqa xmm7,[GOTOFF(ebx,PD_DESCALE_P2)] ; xmm7=[PD_DESCALE_P2] | |
777 | |
778 paddd xmm3,xmm7 | |
779 paddd xmm1,xmm7 | |
780 psrad xmm3,DESCALE_P2 | |
781 psrad xmm1,DESCALE_P2 | |
782 paddd xmm2,xmm7 | |
783 paddd xmm0,xmm7 | |
784 psrad xmm2,DESCALE_P2 | |
785 psrad xmm0,DESCALE_P2 | |
786 | |
787 movdqa xmm5,[GOTOFF(ebx,PB_CENTERJSAMP)] ; xmm5=[PB_CENTERJSAMP] | |
788 | |
789 packssdw xmm3,xmm1 ; xmm3=data3=(03 13 23 33 43 53 63 73) | |
790 packssdw xmm2,xmm0 ; xmm2=data4=(04 14 24 34 44 54 64 74) | |
791 | |
792 movdqa xmm7, XMMWORD [wk(0)] ; xmm7=(00 10 20 30 40 50 60 70 06 16 26
36 46 56 66 76) | |
793 movdqa xmm1, XMMWORD [wk(1)] ; xmm1=(01 11 21 31 41 51 61 71 07 17 27
37 47 57 67 77) | |
794 | |
795 packsswb xmm6,xmm2 ; xmm6=(02 12 22 32 42 52 62 72 04 14 24
34 44 54 64 74) | |
796 packsswb xmm3,xmm4 ; xmm3=(03 13 23 33 43 53 63 73 05 15 25
35 45 55 65 75) | |
797 | |
798 paddb xmm7,xmm5 | |
799 paddb xmm1,xmm5 | |
800 paddb xmm6,xmm5 | |
801 paddb xmm3,xmm5 | |
802 | |
803 movdqa xmm0,xmm7 ; transpose coefficients(phase 1) | |
804 punpcklbw xmm7,xmm1 ; xmm7=(00 01 10 11 20 21 30 31 40 41 50 51 60 6
1 70 71) | |
805 punpckhbw xmm0,xmm1 ; xmm0=(06 07 16 17 26 27 36 37 46 47 56 57 66 6
7 76 77) | |
806 movdqa xmm2,xmm6 ; transpose coefficients(phase 1) | |
807 punpcklbw xmm6,xmm3 ; xmm6=(02 03 12 13 22 23 32 33 42 43 52 53 62 6
3 72 73) | |
808 punpckhbw xmm2,xmm3 ; xmm2=(04 05 14 15 24 25 34 35 44 45 54 55 64 6
5 74 75) | |
809 | |
810 movdqa xmm4,xmm7 ; transpose coefficients(phase 2) | |
811 punpcklwd xmm7,xmm6 ; xmm7=(00 01 02 03 10 11 12 13 20 21 22 23 30 3
1 32 33) | |
812 punpckhwd xmm4,xmm6 ; xmm4=(40 41 42 43 50 51 52 53 60 61 62 63 70 7
1 72 73) | |
813 movdqa xmm5,xmm2 ; transpose coefficients(phase 2) | |
814 punpcklwd xmm2,xmm0 ; xmm2=(04 05 06 07 14 15 16 17 24 25 26 27 34 3
5 36 37) | |
815 punpckhwd xmm5,xmm0 ; xmm5=(44 45 46 47 54 55 56 57 64 65 66 67 74 7
5 76 77) | |
816 | |
817 movdqa xmm1,xmm7 ; transpose coefficients(phase 3) | |
818 punpckldq xmm7,xmm2 ; xmm7=(00 01 02 03 04 05 06 07 10 11 12 13 14 1
5 16 17) | |
819 punpckhdq xmm1,xmm2 ; xmm1=(20 21 22 23 24 25 26 27 30 31 32 33 34 3
5 36 37) | |
820 movdqa xmm3,xmm4 ; transpose coefficients(phase 3) | |
821 punpckldq xmm4,xmm5 ; xmm4=(40 41 42 43 44 45 46 47 50 51 52 53 54 5
5 56 57) | |
822 punpckhdq xmm3,xmm5 ; xmm3=(60 61 62 63 64 65 66 67 70 71 72 73 74 7
5 76 77) | |
823 | |
824 pshufd xmm6,xmm7,0x4E ; xmm6=(10 11 12 13 14 15 16 17 00 01 02 03 04 0
5 06 07) | |
825 pshufd xmm0,xmm1,0x4E ; xmm0=(30 31 32 33 34 35 36 37 20 21 22 23 24 2
5 26 27) | |
826 pshufd xmm2,xmm4,0x4E ; xmm2=(50 51 52 53 54 55 56 57 40 41 42 43 44 4
5 46 47) | |
827 pshufd xmm5,xmm3,0x4E ; xmm5=(70 71 72 73 74 75 76 77 60 61 62 63 64 6
5 66 67) | |
828 | |
829 mov edx, JSAMPROW [edi+0*SIZEOF_JSAMPROW] | |
830 mov esi, JSAMPROW [edi+2*SIZEOF_JSAMPROW] | |
831 movq XMM_MMWORD [edx+eax*SIZEOF_JSAMPLE], xmm7 | |
832 movq XMM_MMWORD [esi+eax*SIZEOF_JSAMPLE], xmm1 | |
833 mov edx, JSAMPROW [edi+4*SIZEOF_JSAMPROW] | |
834 mov esi, JSAMPROW [edi+6*SIZEOF_JSAMPROW] | |
835 movq XMM_MMWORD [edx+eax*SIZEOF_JSAMPLE], xmm4 | |
836 movq XMM_MMWORD [esi+eax*SIZEOF_JSAMPLE], xmm3 | |
837 | |
838 mov edx, JSAMPROW [edi+1*SIZEOF_JSAMPROW] | |
839 mov esi, JSAMPROW [edi+3*SIZEOF_JSAMPROW] | |
840 movq XMM_MMWORD [edx+eax*SIZEOF_JSAMPLE], xmm6 | |
841 movq XMM_MMWORD [esi+eax*SIZEOF_JSAMPLE], xmm0 | |
842 mov edx, JSAMPROW [edi+5*SIZEOF_JSAMPROW] | |
843 mov esi, JSAMPROW [edi+7*SIZEOF_JSAMPROW] | |
844 movq XMM_MMWORD [edx+eax*SIZEOF_JSAMPLE], xmm2 | |
845 movq XMM_MMWORD [esi+eax*SIZEOF_JSAMPLE], xmm5 | |
846 | |
847 pop edi | |
848 pop esi | |
849 ; pop edx ; need not be preserved | |
850 ; pop ecx ; unused | |
851 poppic ebx | |
852 mov esp,ebp ; esp <- aligned ebp | |
853 pop esp ; esp <- original ebp | |
854 pop ebp | |
855 ret | |
856 | |
857 ; For some reason, the OS X linker does not honor the request to align the | |
858 ; segment unless we do this. | |
859 align 16 | |
OLD | NEW |