| Index: simd/jdmrgext-sse2-64.asm
 | 
| diff --git a/simd/jdmrgext-sse2-64.asm b/simd/jdmrgext-sse2-64.asm
 | 
| new file mode 100644
 | 
| index 0000000000000000000000000000000000000000..989d7f17c24e4bae1c4102d8ec8889fe5a390faf
 | 
| --- /dev/null
 | 
| +++ b/simd/jdmrgext-sse2-64.asm
 | 
| @@ -0,0 +1,538 @@
 | 
| +;
 | 
| +; jdmrgext.asm - merged upsampling/color conversion (64-bit SSE2)
 | 
| +;
 | 
| +; Copyright 2009, 2012 Pierre Ossman <ossman@cendio.se> for Cendio AB
 | 
| +; Copyright 2009, 2012 D. R. Commander
 | 
| +;
 | 
| +; Based on
 | 
| +; x86 SIMD extension for IJG JPEG library
 | 
| +; Copyright (C) 1999-2006, MIYASAKA Masaru.
 | 
| +; For conditions of distribution and use, see copyright notice in jsimdext.inc
 | 
| +;
 | 
| +; This file should be assembled with NASM (Netwide Assembler),
 | 
| +; can *not* be assembled with Microsoft's MASM or any compatible
 | 
| +; assembler (including Borland's Turbo Assembler).
 | 
| +; NASM is available from http://nasm.sourceforge.net/ or
 | 
| +; http://sourceforge.net/project/showfiles.php?group_id=6208
 | 
| +;
 | 
| +; [TAB8]
 | 
| +
 | 
| +%include "jcolsamp.inc"
 | 
| +
 | 
| +; --------------------------------------------------------------------------
 | 
| +;
 | 
| +; Upsample and color convert for the case of 2:1 horizontal and 1:1 vertical.
 | 
| +;
 | 
| +; GLOBAL(void)
 | 
| +; jsimd_h2v1_merged_upsample_sse2 (JDIMENSION output_width,
 | 
| +;                                  JSAMPIMAGE input_buf,
 | 
| +;                                  JDIMENSION in_row_group_ctr,
 | 
| +;                                  JSAMPARRAY output_buf);
 | 
| +;
 | 
| +
 | 
| +; r10 = JDIMENSION output_width
 | 
| +; r11 = JSAMPIMAGE input_buf
 | 
| +; r12 = JDIMENSION in_row_group_ctr
 | 
| +; r13 = JSAMPARRAY output_buf
 | 
| +
 | 
| +%define wk(i)           rbp-(WK_NUM-(i))*SIZEOF_XMMWORD ; xmmword wk[WK_NUM]
 | 
| +%define WK_NUM          3
 | 
| +
 | 
| +        align   16
 | 
| +        global  EXTN(jsimd_h2v1_merged_upsample_sse2)
 | 
| +
 | 
| +EXTN(jsimd_h2v1_merged_upsample_sse2):
 | 
| +        push    rbp
 | 
| +        mov     rax,rsp                         ; rax = original rbp
 | 
| +        sub     rsp, byte 4
 | 
| +        and     rsp, byte (-SIZEOF_XMMWORD)     ; align to 128 bits
 | 
| +        mov     [rsp],rax
 | 
| +        mov     rbp,rsp                         ; rbp = aligned rbp
 | 
| +        lea     rsp, [wk(0)]
 | 
| +        collect_args
 | 
| +        push    rbx
 | 
| +
 | 
| +        mov     ecx, r10d        ; col
 | 
| +        test    rcx,rcx
 | 
| +        jz      near .return
 | 
| +
 | 
| +        push    rcx
 | 
| +
 | 
| +        mov     rdi, r11
 | 
| +        mov     ecx, r12d
 | 
| +        mov     rsi, JSAMPARRAY [rdi+0*SIZEOF_JSAMPARRAY]
 | 
| +        mov     rbx, JSAMPARRAY [rdi+1*SIZEOF_JSAMPARRAY]
 | 
| +        mov     rdx, JSAMPARRAY [rdi+2*SIZEOF_JSAMPARRAY]
 | 
| +        mov     rdi, r13
 | 
| +        mov     rsi, JSAMPROW [rsi+rcx*SIZEOF_JSAMPROW]         ; inptr0
 | 
| +        mov     rbx, JSAMPROW [rbx+rcx*SIZEOF_JSAMPROW]         ; inptr1
 | 
| +        mov     rdx, JSAMPROW [rdx+rcx*SIZEOF_JSAMPROW]         ; inptr2
 | 
| +        mov     rdi, JSAMPROW [rdi]                             ; outptr
 | 
| +
 | 
| +        pop     rcx                     ; col
 | 
| +
 | 
| +.columnloop:
 | 
| +
 | 
| +        movdqa    xmm6, XMMWORD [rbx]   ; xmm6=Cb(0123456789ABCDEF)
 | 
| +        movdqa    xmm7, XMMWORD [rdx]   ; xmm7=Cr(0123456789ABCDEF)
 | 
| +
 | 
| +        pxor      xmm1,xmm1             ; xmm1=(all 0's)
 | 
| +        pcmpeqw   xmm3,xmm3
 | 
| +        psllw     xmm3,7                ; xmm3={0xFF80 0xFF80 0xFF80 0xFF80 ..}
 | 
| +
 | 
| +        movdqa    xmm4,xmm6
 | 
| +        punpckhbw xmm6,xmm1             ; xmm6=Cb(89ABCDEF)=CbH
 | 
| +        punpcklbw xmm4,xmm1             ; xmm4=Cb(01234567)=CbL
 | 
| +        movdqa    xmm0,xmm7
 | 
| +        punpckhbw xmm7,xmm1             ; xmm7=Cr(89ABCDEF)=CrH
 | 
| +        punpcklbw xmm0,xmm1             ; xmm0=Cr(01234567)=CrL
 | 
| +
 | 
| +        paddw     xmm6,xmm3
 | 
| +        paddw     xmm4,xmm3
 | 
| +        paddw     xmm7,xmm3
 | 
| +        paddw     xmm0,xmm3
 | 
| +
 | 
| +        ; (Original)
 | 
| +        ; R = Y                + 1.40200 * Cr
 | 
| +        ; G = Y - 0.34414 * Cb - 0.71414 * Cr
 | 
| +        ; B = Y + 1.77200 * Cb
 | 
| +        ;
 | 
| +        ; (This implementation)
 | 
| +        ; R = Y                + 0.40200 * Cr + Cr
 | 
| +        ; G = Y - 0.34414 * Cb + 0.28586 * Cr - Cr
 | 
| +        ; B = Y - 0.22800 * Cb + Cb + Cb
 | 
| +
 | 
| +        movdqa  xmm5,xmm6               ; xmm5=CbH
 | 
| +        movdqa  xmm2,xmm4               ; xmm2=CbL
 | 
| +        paddw   xmm6,xmm6               ; xmm6=2*CbH
 | 
| +        paddw   xmm4,xmm4               ; xmm4=2*CbL
 | 
| +        movdqa  xmm1,xmm7               ; xmm1=CrH
 | 
| +        movdqa  xmm3,xmm0               ; xmm3=CrL
 | 
| +        paddw   xmm7,xmm7               ; xmm7=2*CrH
 | 
| +        paddw   xmm0,xmm0               ; xmm0=2*CrL
 | 
| +
 | 
| +        pmulhw  xmm6,[rel PW_MF0228]    ; xmm6=(2*CbH * -FIX(0.22800))
 | 
| +        pmulhw  xmm4,[rel PW_MF0228]    ; xmm4=(2*CbL * -FIX(0.22800))
 | 
| +        pmulhw  xmm7,[rel PW_F0402]     ; xmm7=(2*CrH * FIX(0.40200))
 | 
| +        pmulhw  xmm0,[rel PW_F0402]     ; xmm0=(2*CrL * FIX(0.40200))
 | 
| +
 | 
| +        paddw   xmm6,[rel PW_ONE]
 | 
| +        paddw   xmm4,[rel PW_ONE]
 | 
| +        psraw   xmm6,1                  ; xmm6=(CbH * -FIX(0.22800))
 | 
| +        psraw   xmm4,1                  ; xmm4=(CbL * -FIX(0.22800))
 | 
| +        paddw   xmm7,[rel PW_ONE]
 | 
| +        paddw   xmm0,[rel PW_ONE]
 | 
| +        psraw   xmm7,1                  ; xmm7=(CrH * FIX(0.40200))
 | 
| +        psraw   xmm0,1                  ; xmm0=(CrL * FIX(0.40200))
 | 
| +
 | 
| +        paddw   xmm6,xmm5
 | 
| +        paddw   xmm4,xmm2
 | 
| +        paddw   xmm6,xmm5               ; xmm6=(CbH * FIX(1.77200))=(B-Y)H
 | 
| +        paddw   xmm4,xmm2               ; xmm4=(CbL * FIX(1.77200))=(B-Y)L
 | 
| +        paddw   xmm7,xmm1               ; xmm7=(CrH * FIX(1.40200))=(R-Y)H
 | 
| +        paddw   xmm0,xmm3               ; xmm0=(CrL * FIX(1.40200))=(R-Y)L
 | 
| +
 | 
| +        movdqa  XMMWORD [wk(0)], xmm6   ; wk(0)=(B-Y)H
 | 
| +        movdqa  XMMWORD [wk(1)], xmm7   ; wk(1)=(R-Y)H
 | 
| +
 | 
| +        movdqa    xmm6,xmm5
 | 
| +        movdqa    xmm7,xmm2
 | 
| +        punpcklwd xmm5,xmm1
 | 
| +        punpckhwd xmm6,xmm1
 | 
| +        pmaddwd   xmm5,[rel PW_MF0344_F0285]
 | 
| +        pmaddwd   xmm6,[rel PW_MF0344_F0285]
 | 
| +        punpcklwd xmm2,xmm3
 | 
| +        punpckhwd xmm7,xmm3
 | 
| +        pmaddwd   xmm2,[rel PW_MF0344_F0285]
 | 
| +        pmaddwd   xmm7,[rel PW_MF0344_F0285]
 | 
| +
 | 
| +        paddd     xmm5,[rel PD_ONEHALF]
 | 
| +        paddd     xmm6,[rel PD_ONEHALF]
 | 
| +        psrad     xmm5,SCALEBITS
 | 
| +        psrad     xmm6,SCALEBITS
 | 
| +        paddd     xmm2,[rel PD_ONEHALF]
 | 
| +        paddd     xmm7,[rel PD_ONEHALF]
 | 
| +        psrad     xmm2,SCALEBITS
 | 
| +        psrad     xmm7,SCALEBITS
 | 
| +
 | 
| +        packssdw  xmm5,xmm6     ; xmm5=CbH*-FIX(0.344)+CrH*FIX(0.285)
 | 
| +        packssdw  xmm2,xmm7     ; xmm2=CbL*-FIX(0.344)+CrL*FIX(0.285)
 | 
| +        psubw     xmm5,xmm1     ; xmm5=CbH*-FIX(0.344)+CrH*-FIX(0.714)=(G-Y)H
 | 
| +        psubw     xmm2,xmm3     ; xmm2=CbL*-FIX(0.344)+CrL*-FIX(0.714)=(G-Y)L
 | 
| +
 | 
| +        movdqa  XMMWORD [wk(2)], xmm5   ; wk(2)=(G-Y)H
 | 
| +
 | 
| +        mov     al,2                    ; Yctr
 | 
| +        jmp     short .Yloop_1st
 | 
| +
 | 
| +.Yloop_2nd:
 | 
| +        movdqa  xmm0, XMMWORD [wk(1)]   ; xmm0=(R-Y)H
 | 
| +        movdqa  xmm2, XMMWORD [wk(2)]   ; xmm2=(G-Y)H
 | 
| +        movdqa  xmm4, XMMWORD [wk(0)]   ; xmm4=(B-Y)H
 | 
| +
 | 
| +.Yloop_1st:
 | 
| +        movdqa  xmm7, XMMWORD [rsi]     ; xmm7=Y(0123456789ABCDEF)
 | 
| +
 | 
| +        pcmpeqw xmm6,xmm6
 | 
| +        psrlw   xmm6,BYTE_BIT           ; xmm6={0xFF 0x00 0xFF 0x00 ..}
 | 
| +        pand    xmm6,xmm7               ; xmm6=Y(02468ACE)=YE
 | 
| +        psrlw   xmm7,BYTE_BIT           ; xmm7=Y(13579BDF)=YO
 | 
| +
 | 
| +        movdqa  xmm1,xmm0               ; xmm1=xmm0=(R-Y)(L/H)
 | 
| +        movdqa  xmm3,xmm2               ; xmm3=xmm2=(G-Y)(L/H)
 | 
| +        movdqa  xmm5,xmm4               ; xmm5=xmm4=(B-Y)(L/H)
 | 
| +
 | 
| +        paddw     xmm0,xmm6             ; xmm0=((R-Y)+YE)=RE=R(02468ACE)
 | 
| +        paddw     xmm1,xmm7             ; xmm1=((R-Y)+YO)=RO=R(13579BDF)
 | 
| +        packuswb  xmm0,xmm0             ; xmm0=R(02468ACE********)
 | 
| +        packuswb  xmm1,xmm1             ; xmm1=R(13579BDF********)
 | 
| +
 | 
| +        paddw     xmm2,xmm6             ; xmm2=((G-Y)+YE)=GE=G(02468ACE)
 | 
| +        paddw     xmm3,xmm7             ; xmm3=((G-Y)+YO)=GO=G(13579BDF)
 | 
| +        packuswb  xmm2,xmm2             ; xmm2=G(02468ACE********)
 | 
| +        packuswb  xmm3,xmm3             ; xmm3=G(13579BDF********)
 | 
| +
 | 
| +        paddw     xmm4,xmm6             ; xmm4=((B-Y)+YE)=BE=B(02468ACE)
 | 
| +        paddw     xmm5,xmm7             ; xmm5=((B-Y)+YO)=BO=B(13579BDF)
 | 
| +        packuswb  xmm4,xmm4             ; xmm4=B(02468ACE********)
 | 
| +        packuswb  xmm5,xmm5             ; xmm5=B(13579BDF********)
 | 
| +
 | 
| +%if RGB_PIXELSIZE == 3 ; ---------------
 | 
| +
 | 
| +        ; xmmA=(00 02 04 06 08 0A 0C 0E **), xmmB=(01 03 05 07 09 0B 0D 0F **)
 | 
| +        ; xmmC=(10 12 14 16 18 1A 1C 1E **), xmmD=(11 13 15 17 19 1B 1D 1F **)
 | 
| +        ; xmmE=(20 22 24 26 28 2A 2C 2E **), xmmF=(21 23 25 27 29 2B 2D 2F **)
 | 
| +        ; xmmG=(** ** ** ** ** ** ** ** **), xmmH=(** ** ** ** ** ** ** ** **)
 | 
| +
 | 
| +        punpcklbw xmmA,xmmC     ; xmmA=(00 10 02 12 04 14 06 16 08 18 0A 1A 0C 1C 0E 1E)
 | 
| +        punpcklbw xmmE,xmmB     ; xmmE=(20 01 22 03 24 05 26 07 28 09 2A 0B 2C 0D 2E 0F)
 | 
| +        punpcklbw xmmD,xmmF     ; xmmD=(11 21 13 23 15 25 17 27 19 29 1B 2B 1D 2D 1F 2F)
 | 
| +
 | 
| +        movdqa    xmmG,xmmA
 | 
| +        movdqa    xmmH,xmmA
 | 
| +        punpcklwd xmmA,xmmE     ; xmmA=(00 10 20 01 02 12 22 03 04 14 24 05 06 16 26 07)
 | 
| +        punpckhwd xmmG,xmmE     ; xmmG=(08 18 28 09 0A 1A 2A 0B 0C 1C 2C 0D 0E 1E 2E 0F)
 | 
| +
 | 
| +        psrldq    xmmH,2        ; xmmH=(02 12 04 14 06 16 08 18 0A 1A 0C 1C 0E 1E -- --)
 | 
| +        psrldq    xmmE,2        ; xmmE=(22 03 24 05 26 07 28 09 2A 0B 2C 0D 2E 0F -- --)
 | 
| +
 | 
| +        movdqa    xmmC,xmmD
 | 
| +        movdqa    xmmB,xmmD
 | 
| +        punpcklwd xmmD,xmmH     ; xmmD=(11 21 02 12 13 23 04 14 15 25 06 16 17 27 08 18)
 | 
| +        punpckhwd xmmC,xmmH     ; xmmC=(19 29 0A 1A 1B 2B 0C 1C 1D 2D 0E 1E 1F 2F -- --)
 | 
| +
 | 
| +        psrldq    xmmB,2        ; xmmB=(13 23 15 25 17 27 19 29 1B 2B 1D 2D 1F 2F -- --)
 | 
| +
 | 
| +        movdqa    xmmF,xmmE
 | 
| +        punpcklwd xmmE,xmmB     ; xmmE=(22 03 13 23 24 05 15 25 26 07 17 27 28 09 19 29)
 | 
| +        punpckhwd xmmF,xmmB     ; xmmF=(2A 0B 1B 2B 2C 0D 1D 2D 2E 0F 1F 2F -- -- -- --)
 | 
| +
 | 
| +        pshufd    xmmH,xmmA,0x4E; xmmH=(04 14 24 05 06 16 26 07 00 10 20 01 02 12 22 03)
 | 
| +        movdqa    xmmB,xmmE
 | 
| +        punpckldq xmmA,xmmD     ; xmmA=(00 10 20 01 11 21 02 12 02 12 22 03 13 23 04 14)
 | 
| +        punpckldq xmmE,xmmH     ; xmmE=(22 03 13 23 04 14 24 05 24 05 15 25 06 16 26 07)
 | 
| +        punpckhdq xmmD,xmmB     ; xmmD=(15 25 06 16 26 07 17 27 17 27 08 18 28 09 19 29)
 | 
| +
 | 
| +        pshufd    xmmH,xmmG,0x4E; xmmH=(0C 1C 2C 0D 0E 1E 2E 0F 08 18 28 09 0A 1A 2A 0B)
 | 
| +        movdqa    xmmB,xmmF
 | 
| +        punpckldq xmmG,xmmC     ; xmmG=(08 18 28 09 19 29 0A 1A 0A 1A 2A 0B 1B 2B 0C 1C)
 | 
| +        punpckldq xmmF,xmmH     ; xmmF=(2A 0B 1B 2B 0C 1C 2C 0D 2C 0D 1D 2D 0E 1E 2E 0F)
 | 
| +        punpckhdq xmmC,xmmB     ; xmmC=(1D 2D 0E 1E 2E 0F 1F 2F 1F 2F -- -- -- -- -- --)
 | 
| +
 | 
| +        punpcklqdq xmmA,xmmE    ; xmmA=(00 10 20 01 11 21 02 12 22 03 13 23 04 14 24 05)
 | 
| +        punpcklqdq xmmD,xmmG    ; xmmD=(15 25 06 16 26 07 17 27 08 18 28 09 19 29 0A 1A)
 | 
| +        punpcklqdq xmmF,xmmC    ; xmmF=(2A 0B 1B 2B 0C 1C 2C 0D 1D 2D 0E 1E 2E 0F 1F 2F)
 | 
| +
 | 
| +        cmp     rcx, byte SIZEOF_XMMWORD
 | 
| +        jb      short .column_st32
 | 
| +
 | 
| +        test    rdi, SIZEOF_XMMWORD-1
 | 
| +        jnz     short .out1
 | 
| +        ; --(aligned)-------------------
 | 
| +        movntdq XMMWORD [rdi+0*SIZEOF_XMMWORD], xmmA
 | 
| +        movntdq XMMWORD [rdi+1*SIZEOF_XMMWORD], xmmD
 | 
| +        movntdq XMMWORD [rdi+2*SIZEOF_XMMWORD], xmmF
 | 
| +        jmp     short .out0
 | 
| +.out1:  ; --(unaligned)-----------------
 | 
| +        movdqu  XMMWORD [rdi+0*SIZEOF_XMMWORD], xmmA
 | 
| +        movdqu  XMMWORD [rdi+1*SIZEOF_XMMWORD], xmmD
 | 
| +        movdqu  XMMWORD [rdi+2*SIZEOF_XMMWORD], xmmF
 | 
| +.out0:
 | 
| +        add     rdi, byte RGB_PIXELSIZE*SIZEOF_XMMWORD  ; outptr
 | 
| +        sub     rcx, byte SIZEOF_XMMWORD
 | 
| +        jz      near .endcolumn
 | 
| +
 | 
| +        add     rsi, byte SIZEOF_XMMWORD        ; inptr0
 | 
| +        dec     al                      ; Yctr
 | 
| +        jnz     near .Yloop_2nd
 | 
| +
 | 
| +        add     rbx, byte SIZEOF_XMMWORD        ; inptr1
 | 
| +        add     rdx, byte SIZEOF_XMMWORD        ; inptr2
 | 
| +        jmp     near .columnloop
 | 
| +
 | 
| +.column_st32:
 | 
| +        lea     rcx, [rcx+rcx*2]                ; imul ecx, RGB_PIXELSIZE
 | 
| +        cmp     rcx, byte 2*SIZEOF_XMMWORD
 | 
| +        jb      short .column_st16
 | 
| +        movdqu  XMMWORD [rdi+0*SIZEOF_XMMWORD], xmmA
 | 
| +        movdqu  XMMWORD [rdi+1*SIZEOF_XMMWORD], xmmD
 | 
| +        add     rdi, byte 2*SIZEOF_XMMWORD      ; outptr
 | 
| +        movdqa  xmmA,xmmF
 | 
| +        sub     rcx, byte 2*SIZEOF_XMMWORD
 | 
| +        jmp     short .column_st15
 | 
| +.column_st16:
 | 
| +        cmp     rcx, byte SIZEOF_XMMWORD
 | 
| +        jb      short .column_st15
 | 
| +        movdqu  XMMWORD [rdi+0*SIZEOF_XMMWORD], xmmA
 | 
| +        add     rdi, byte SIZEOF_XMMWORD        ; outptr
 | 
| +        movdqa  xmmA,xmmD
 | 
| +        sub     rcx, byte SIZEOF_XMMWORD
 | 
| +.column_st15:
 | 
| +        ; Store the lower 8 bytes of xmmA to the output when it has enough
 | 
| +        ; space.
 | 
| +        cmp     rcx, byte SIZEOF_MMWORD
 | 
| +        jb      short .column_st7
 | 
| +        movq    XMM_MMWORD [rdi], xmmA
 | 
| +        add     rdi, byte SIZEOF_MMWORD
 | 
| +        sub     rcx, byte SIZEOF_MMWORD
 | 
| +        psrldq  xmmA, SIZEOF_MMWORD
 | 
| +.column_st7:
 | 
| +        ; Store the lower 4 bytes of xmmA to the output when it has enough
 | 
| +        ; space.
 | 
| +        cmp     rcx, byte SIZEOF_DWORD
 | 
| +        jb      short .column_st3
 | 
| +        movd    XMM_DWORD [rdi], xmmA
 | 
| +        add     rdi, byte SIZEOF_DWORD
 | 
| +        sub     rcx, byte SIZEOF_DWORD
 | 
| +        psrldq  xmmA, SIZEOF_DWORD
 | 
| +.column_st3:
 | 
| +        ; Store the lower 2 bytes of rax to the output when it has enough
 | 
| +        ; space.
 | 
| +        movd    eax, xmmA
 | 
| +        cmp     rcx, byte SIZEOF_WORD
 | 
| +        jb      short .column_st1
 | 
| +        mov     WORD [rdi], ax
 | 
| +        add     rdi, byte SIZEOF_WORD
 | 
| +        sub     rcx, byte SIZEOF_WORD
 | 
| +        shr     rax, 16
 | 
| +.column_st1:
 | 
| +        ; Store the lower 1 byte of rax to the output when it has enough
 | 
| +        ; space.
 | 
| +        test    rcx, rcx
 | 
| +        jz      short .endcolumn
 | 
| +        mov     BYTE [rdi], al
 | 
| +
 | 
| +%else ; RGB_PIXELSIZE == 4 ; -----------
 | 
| +
 | 
| +%ifdef RGBX_FILLER_0XFF
 | 
| +        pcmpeqb   xmm6,xmm6             ; xmm6=XE=X(02468ACE********)
 | 
| +        pcmpeqb   xmm7,xmm7             ; xmm7=XO=X(13579BDF********)
 | 
| +%else
 | 
| +        pxor      xmm6,xmm6             ; xmm6=XE=X(02468ACE********)
 | 
| +        pxor      xmm7,xmm7             ; xmm7=XO=X(13579BDF********)
 | 
| +%endif
 | 
| +        ; xmmA=(00 02 04 06 08 0A 0C 0E **), xmmB=(01 03 05 07 09 0B 0D 0F **)
 | 
| +        ; xmmC=(10 12 14 16 18 1A 1C 1E **), xmmD=(11 13 15 17 19 1B 1D 1F **)
 | 
| +        ; xmmE=(20 22 24 26 28 2A 2C 2E **), xmmF=(21 23 25 27 29 2B 2D 2F **)
 | 
| +        ; xmmG=(30 32 34 36 38 3A 3C 3E **), xmmH=(31 33 35 37 39 3B 3D 3F **)
 | 
| +
 | 
| +        punpcklbw xmmA,xmmC     ; xmmA=(00 10 02 12 04 14 06 16 08 18 0A 1A 0C 1C 0E 1E)
 | 
| +        punpcklbw xmmE,xmmG     ; xmmE=(20 30 22 32 24 34 26 36 28 38 2A 3A 2C 3C 2E 3E)
 | 
| +        punpcklbw xmmB,xmmD     ; xmmB=(01 11 03 13 05 15 07 17 09 19 0B 1B 0D 1D 0F 1F)
 | 
| +        punpcklbw xmmF,xmmH     ; xmmF=(21 31 23 33 25 35 27 37 29 39 2B 3B 2D 3D 2F 3F)
 | 
| +
 | 
| +        movdqa    xmmC,xmmA
 | 
| +        punpcklwd xmmA,xmmE     ; xmmA=(00 10 20 30 02 12 22 32 04 14 24 34 06 16 26 36)
 | 
| +        punpckhwd xmmC,xmmE     ; xmmC=(08 18 28 38 0A 1A 2A 3A 0C 1C 2C 3C 0E 1E 2E 3E)
 | 
| +        movdqa    xmmG,xmmB
 | 
| +        punpcklwd xmmB,xmmF     ; xmmB=(01 11 21 31 03 13 23 33 05 15 25 35 07 17 27 37)
 | 
| +        punpckhwd xmmG,xmmF     ; xmmG=(09 19 29 39 0B 1B 2B 3B 0D 1D 2D 3D 0F 1F 2F 3F)
 | 
| +
 | 
| +        movdqa    xmmD,xmmA
 | 
| +        punpckldq xmmA,xmmB     ; xmmA=(00 10 20 30 01 11 21 31 02 12 22 32 03 13 23 33)
 | 
| +        punpckhdq xmmD,xmmB     ; xmmD=(04 14 24 34 05 15 25 35 06 16 26 36 07 17 27 37)
 | 
| +        movdqa    xmmH,xmmC
 | 
| +        punpckldq xmmC,xmmG     ; xmmC=(08 18 28 38 09 19 29 39 0A 1A 2A 3A 0B 1B 2B 3B)
 | 
| +        punpckhdq xmmH,xmmG     ; xmmH=(0C 1C 2C 3C 0D 1D 2D 3D 0E 1E 2E 3E 0F 1F 2F 3F)
 | 
| +
 | 
| +        cmp     rcx, byte SIZEOF_XMMWORD
 | 
| +        jb      short .column_st32
 | 
| +
 | 
| +        test    rdi, SIZEOF_XMMWORD-1
 | 
| +        jnz     short .out1
 | 
| +        ; --(aligned)-------------------
 | 
| +        movntdq XMMWORD [rdi+0*SIZEOF_XMMWORD], xmmA
 | 
| +        movntdq XMMWORD [rdi+1*SIZEOF_XMMWORD], xmmD
 | 
| +        movntdq XMMWORD [rdi+2*SIZEOF_XMMWORD], xmmC
 | 
| +        movntdq XMMWORD [rdi+3*SIZEOF_XMMWORD], xmmH
 | 
| +        jmp     short .out0
 | 
| +.out1:  ; --(unaligned)-----------------
 | 
| +        movdqu  XMMWORD [rdi+0*SIZEOF_XMMWORD], xmmA
 | 
| +        movdqu  XMMWORD [rdi+1*SIZEOF_XMMWORD], xmmD
 | 
| +        movdqu  XMMWORD [rdi+2*SIZEOF_XMMWORD], xmmC
 | 
| +        movdqu  XMMWORD [rdi+3*SIZEOF_XMMWORD], xmmH
 | 
| +.out0:
 | 
| +        add     rdi, byte RGB_PIXELSIZE*SIZEOF_XMMWORD  ; outptr
 | 
| +        sub     rcx, byte SIZEOF_XMMWORD
 | 
| +        jz      near .endcolumn
 | 
| +
 | 
| +        add     rsi, byte SIZEOF_XMMWORD        ; inptr0
 | 
| +        dec     al                      ; Yctr
 | 
| +        jnz     near .Yloop_2nd
 | 
| +
 | 
| +        add     rbx, byte SIZEOF_XMMWORD        ; inptr1
 | 
| +        add     rdx, byte SIZEOF_XMMWORD        ; inptr2
 | 
| +        jmp     near .columnloop
 | 
| +
 | 
| +.column_st32:
 | 
| +        cmp     rcx, byte SIZEOF_XMMWORD/2
 | 
| +        jb      short .column_st16
 | 
| +        movdqu  XMMWORD [rdi+0*SIZEOF_XMMWORD], xmmA
 | 
| +        movdqu  XMMWORD [rdi+1*SIZEOF_XMMWORD], xmmD
 | 
| +        add     rdi, byte 2*SIZEOF_XMMWORD      ; outptr
 | 
| +        movdqa  xmmA,xmmC
 | 
| +        movdqa  xmmD,xmmH
 | 
| +        sub     rcx, byte SIZEOF_XMMWORD/2
 | 
| +.column_st16:
 | 
| +        cmp     rcx, byte SIZEOF_XMMWORD/4
 | 
| +        jb      short .column_st15
 | 
| +        movdqu  XMMWORD [rdi+0*SIZEOF_XMMWORD], xmmA
 | 
| +        add     rdi, byte SIZEOF_XMMWORD        ; outptr
 | 
| +        movdqa  xmmA,xmmD
 | 
| +        sub     rcx, byte SIZEOF_XMMWORD/4
 | 
| +.column_st15:
 | 
| +        ; Store two pixels (8 bytes) of xmmA to the output when it has enough
 | 
| +        ; space.
 | 
| +        cmp     rcx, byte SIZEOF_XMMWORD/8
 | 
| +        jb      short .column_st7
 | 
| +        movq    XMM_MMWORD [rdi], xmmA
 | 
| +        add     rdi, byte SIZEOF_XMMWORD/8*4
 | 
| +        sub     rcx, byte SIZEOF_XMMWORD/8
 | 
| +        psrldq  xmmA, SIZEOF_XMMWORD/8*4
 | 
| +.column_st7:
 | 
| +        ; Store one pixel (4 bytes) of xmmA to the output when it has enough
 | 
| +        ; space.
 | 
| +        test    rcx, rcx
 | 
| +        jz      short .endcolumn
 | 
| +        movd    XMM_DWORD [rdi], xmmA
 | 
| +
 | 
| +%endif ; RGB_PIXELSIZE ; ---------------
 | 
| +
 | 
| +.endcolumn:
 | 
| +        sfence          ; flush the write buffer
 | 
| +
 | 
| +.return:
 | 
| +        pop     rbx
 | 
| +        uncollect_args
 | 
| +        mov     rsp,rbp         ; rsp <- aligned rbp
 | 
| +        pop     rsp             ; rsp <- original rbp
 | 
| +        pop     rbp
 | 
| +        ret
 | 
| +
 | 
| +; --------------------------------------------------------------------------
 | 
| +;
 | 
| +; Upsample and color convert for the case of 2:1 horizontal and 2:1 vertical.
 | 
| +;
 | 
| +; GLOBAL(void)
 | 
| +; jsimd_h2v2_merged_upsample_sse2 (JDIMENSION output_width,
 | 
| +;                                  JSAMPIMAGE input_buf,
 | 
| +;                                  JDIMENSION in_row_group_ctr,
 | 
| +;                                  JSAMPARRAY output_buf);
 | 
| +;
 | 
| +
 | 
| +; r10 = JDIMENSION output_width
 | 
| +; r11 = JSAMPIMAGE input_buf
 | 
| +; r12 = JDIMENSION in_row_group_ctr
 | 
| +; r13 = JSAMPARRAY output_buf
 | 
| +
 | 
| +        align   16
 | 
| +        global  EXTN(jsimd_h2v2_merged_upsample_sse2)
 | 
| +
 | 
| +EXTN(jsimd_h2v2_merged_upsample_sse2):
 | 
| +        push    rbp
 | 
| +        mov     rax,rsp
 | 
| +        mov     rbp,rsp
 | 
| +        collect_args
 | 
| +        push    rbx
 | 
| +
 | 
| +        mov     eax, r10d
 | 
| +
 | 
| +        mov     rdi, r11
 | 
| +        mov     ecx, r12d
 | 
| +        mov     rsi, JSAMPARRAY [rdi+0*SIZEOF_JSAMPARRAY]
 | 
| +        mov     rbx, JSAMPARRAY [rdi+1*SIZEOF_JSAMPARRAY]
 | 
| +        mov     rdx, JSAMPARRAY [rdi+2*SIZEOF_JSAMPARRAY]
 | 
| +        mov     rdi, r13
 | 
| +        lea     rsi, [rsi+rcx*SIZEOF_JSAMPROW]
 | 
| +
 | 
| +        push    rdx                     ; inptr2
 | 
| +        push    rbx                     ; inptr1
 | 
| +        push    rsi                     ; inptr00
 | 
| +        mov     rbx,rsp
 | 
| +
 | 
| +        push    rdi
 | 
| +        push    rcx
 | 
| +        push    rax
 | 
| +
 | 
| +        %ifdef WIN64
 | 
| +        mov r8, rcx
 | 
| +        mov r9, rdi
 | 
| +        mov rcx, rax
 | 
| +        mov rdx, rbx
 | 
| +        %else
 | 
| +        mov rdx, rcx
 | 
| +        mov rcx, rdi
 | 
| +        mov     rdi, rax
 | 
| +        mov rsi, rbx
 | 
| +        %endif
 | 
| +
 | 
| +        call    EXTN(jsimd_h2v1_merged_upsample_sse2)
 | 
| +
 | 
| +        pop rax
 | 
| +        pop rcx
 | 
| +        pop rdi
 | 
| +        pop rsi
 | 
| +        pop rbx
 | 
| +        pop rdx
 | 
| +
 | 
| +        add     rdi, byte SIZEOF_JSAMPROW       ; outptr1
 | 
| +        add     rsi, byte SIZEOF_JSAMPROW       ; inptr01
 | 
| +
 | 
| +        push    rdx                     ; inptr2
 | 
| +        push    rbx                     ; inptr1
 | 
| +        push    rsi                     ; inptr00
 | 
| +        mov     rbx,rsp
 | 
| +
 | 
| +        push    rdi
 | 
| +        push    rcx
 | 
| +        push    rax
 | 
| +
 | 
| +        %ifdef WIN64
 | 
| +        mov r8, rcx
 | 
| +        mov r9, rdi
 | 
| +        mov rcx, rax
 | 
| +        mov rdx, rbx
 | 
| +        %else
 | 
| +        mov rdx, rcx
 | 
| +        mov rcx, rdi
 | 
| +        mov     rdi, rax
 | 
| +        mov rsi, rbx
 | 
| +        %endif
 | 
| +
 | 
| +        call    EXTN(jsimd_h2v1_merged_upsample_sse2)
 | 
| +
 | 
| +        pop rax
 | 
| +        pop rcx
 | 
| +        pop rdi
 | 
| +        pop rsi
 | 
| +        pop rbx
 | 
| +        pop rdx
 | 
| +
 | 
| +        pop     rbx
 | 
| +        uncollect_args
 | 
| +        pop     rbp
 | 
| +        ret
 | 
| +
 | 
| +; For some reason, the OS X linker does not honor the request to align the
 | 
| +; segment unless we do this.
 | 
| +        align   16
 | 
| 
 |