Index: source/libvpx/vpx_dsp/x86/sad_ssse3.asm |
diff --git a/source/libvpx/vp8/common/x86/sad_ssse3.asm b/source/libvpx/vpx_dsp/x86/sad_ssse3.asm |
similarity index 64% |
rename from source/libvpx/vp8/common/x86/sad_ssse3.asm |
rename to source/libvpx/vpx_dsp/x86/sad_ssse3.asm |
index 278fc0640edc451ae7ca1833997f6616b0bf9ab6..49f204fa04b9848d09e839836224ad40abbff3e5 100644 |
--- a/source/libvpx/vp8/common/x86/sad_ssse3.asm |
+++ b/source/libvpx/vpx_dsp/x86/sad_ssse3.asm |
@@ -146,14 +146,14 @@ |
%endmacro |
-;void int vp8_sad16x16x3_ssse3( |
+;void int vpx_sad16x16x3_ssse3( |
; unsigned char *src_ptr, |
; int src_stride, |
; unsigned char *ref_ptr, |
; int ref_stride, |
; int *results) |
-global sym(vp8_sad16x16x3_ssse3) PRIVATE |
-sym(vp8_sad16x16x3_ssse3): |
+global sym(vpx_sad16x16x3_ssse3) PRIVATE |
+sym(vpx_sad16x16x3_ssse3): |
push rbp |
mov rbp, rsp |
SHADOW_ARGS_TO_STACK 5 |
@@ -169,31 +169,31 @@ sym(vp8_sad16x16x3_ssse3): |
mov rdx, 0xf |
and rdx, rdi |
- jmp .vp8_sad16x16x3_ssse3_skiptable |
-.vp8_sad16x16x3_ssse3_jumptable: |
- dd .vp8_sad16x16x3_ssse3_aligned_by_0 - .vp8_sad16x16x3_ssse3_do_jump |
- dd .vp8_sad16x16x3_ssse3_aligned_by_1 - .vp8_sad16x16x3_ssse3_do_jump |
- dd .vp8_sad16x16x3_ssse3_aligned_by_2 - .vp8_sad16x16x3_ssse3_do_jump |
- dd .vp8_sad16x16x3_ssse3_aligned_by_3 - .vp8_sad16x16x3_ssse3_do_jump |
- dd .vp8_sad16x16x3_ssse3_aligned_by_4 - .vp8_sad16x16x3_ssse3_do_jump |
- dd .vp8_sad16x16x3_ssse3_aligned_by_5 - .vp8_sad16x16x3_ssse3_do_jump |
- dd .vp8_sad16x16x3_ssse3_aligned_by_6 - .vp8_sad16x16x3_ssse3_do_jump |
- dd .vp8_sad16x16x3_ssse3_aligned_by_7 - .vp8_sad16x16x3_ssse3_do_jump |
- dd .vp8_sad16x16x3_ssse3_aligned_by_8 - .vp8_sad16x16x3_ssse3_do_jump |
- dd .vp8_sad16x16x3_ssse3_aligned_by_9 - .vp8_sad16x16x3_ssse3_do_jump |
- dd .vp8_sad16x16x3_ssse3_aligned_by_10 - .vp8_sad16x16x3_ssse3_do_jump |
- dd .vp8_sad16x16x3_ssse3_aligned_by_11 - .vp8_sad16x16x3_ssse3_do_jump |
- dd .vp8_sad16x16x3_ssse3_aligned_by_12 - .vp8_sad16x16x3_ssse3_do_jump |
- dd .vp8_sad16x16x3_ssse3_aligned_by_13 - .vp8_sad16x16x3_ssse3_do_jump |
- dd .vp8_sad16x16x3_ssse3_aligned_by_14 - .vp8_sad16x16x3_ssse3_do_jump |
- dd .vp8_sad16x16x3_ssse3_aligned_by_15 - .vp8_sad16x16x3_ssse3_do_jump |
-.vp8_sad16x16x3_ssse3_skiptable: |
- |
- call .vp8_sad16x16x3_ssse3_do_jump |
-.vp8_sad16x16x3_ssse3_do_jump: |
+ jmp .vpx_sad16x16x3_ssse3_skiptable |
+.vpx_sad16x16x3_ssse3_jumptable: |
+ dd .vpx_sad16x16x3_ssse3_aligned_by_0 - .vpx_sad16x16x3_ssse3_do_jump |
+ dd .vpx_sad16x16x3_ssse3_aligned_by_1 - .vpx_sad16x16x3_ssse3_do_jump |
+ dd .vpx_sad16x16x3_ssse3_aligned_by_2 - .vpx_sad16x16x3_ssse3_do_jump |
+ dd .vpx_sad16x16x3_ssse3_aligned_by_3 - .vpx_sad16x16x3_ssse3_do_jump |
+ dd .vpx_sad16x16x3_ssse3_aligned_by_4 - .vpx_sad16x16x3_ssse3_do_jump |
+ dd .vpx_sad16x16x3_ssse3_aligned_by_5 - .vpx_sad16x16x3_ssse3_do_jump |
+ dd .vpx_sad16x16x3_ssse3_aligned_by_6 - .vpx_sad16x16x3_ssse3_do_jump |
+ dd .vpx_sad16x16x3_ssse3_aligned_by_7 - .vpx_sad16x16x3_ssse3_do_jump |
+ dd .vpx_sad16x16x3_ssse3_aligned_by_8 - .vpx_sad16x16x3_ssse3_do_jump |
+ dd .vpx_sad16x16x3_ssse3_aligned_by_9 - .vpx_sad16x16x3_ssse3_do_jump |
+ dd .vpx_sad16x16x3_ssse3_aligned_by_10 - .vpx_sad16x16x3_ssse3_do_jump |
+ dd .vpx_sad16x16x3_ssse3_aligned_by_11 - .vpx_sad16x16x3_ssse3_do_jump |
+ dd .vpx_sad16x16x3_ssse3_aligned_by_12 - .vpx_sad16x16x3_ssse3_do_jump |
+ dd .vpx_sad16x16x3_ssse3_aligned_by_13 - .vpx_sad16x16x3_ssse3_do_jump |
+ dd .vpx_sad16x16x3_ssse3_aligned_by_14 - .vpx_sad16x16x3_ssse3_do_jump |
+ dd .vpx_sad16x16x3_ssse3_aligned_by_15 - .vpx_sad16x16x3_ssse3_do_jump |
+.vpx_sad16x16x3_ssse3_skiptable: |
+ |
+ call .vpx_sad16x16x3_ssse3_do_jump |
+.vpx_sad16x16x3_ssse3_do_jump: |
pop rcx ; get the address of do_jump |
- mov rax, .vp8_sad16x16x3_ssse3_jumptable - .vp8_sad16x16x3_ssse3_do_jump |
- add rax, rcx ; get the absolute address of vp8_sad16x16x3_ssse3_jumptable |
+ mov rax, .vpx_sad16x16x3_ssse3_jumptable - .vpx_sad16x16x3_ssse3_do_jump |
+ add rax, rcx ; get the absolute address of vpx_sad16x16x3_ssse3_jumptable |
movsxd rax, dword [rax + 4*rdx] ; get the 32 bit offset from the jumptable |
add rcx, rax |
@@ -203,23 +203,23 @@ sym(vp8_sad16x16x3_ssse3): |
jmp rcx |
- PROCESS_16X16X3_OFFSET 0, .vp8_sad16x16x3_ssse3 |
- PROCESS_16X16X3_OFFSET 1, .vp8_sad16x16x3_ssse3 |
- PROCESS_16X16X3_OFFSET 2, .vp8_sad16x16x3_ssse3 |
- PROCESS_16X16X3_OFFSET 3, .vp8_sad16x16x3_ssse3 |
- PROCESS_16X16X3_OFFSET 4, .vp8_sad16x16x3_ssse3 |
- PROCESS_16X16X3_OFFSET 5, .vp8_sad16x16x3_ssse3 |
- PROCESS_16X16X3_OFFSET 6, .vp8_sad16x16x3_ssse3 |
- PROCESS_16X16X3_OFFSET 7, .vp8_sad16x16x3_ssse3 |
- PROCESS_16X16X3_OFFSET 8, .vp8_sad16x16x3_ssse3 |
- PROCESS_16X16X3_OFFSET 9, .vp8_sad16x16x3_ssse3 |
- PROCESS_16X16X3_OFFSET 10, .vp8_sad16x16x3_ssse3 |
- PROCESS_16X16X3_OFFSET 11, .vp8_sad16x16x3_ssse3 |
- PROCESS_16X16X3_OFFSET 12, .vp8_sad16x16x3_ssse3 |
- PROCESS_16X16X3_OFFSET 13, .vp8_sad16x16x3_ssse3 |
- PROCESS_16X16X3_OFFSET 14, .vp8_sad16x16x3_ssse3 |
- |
-.vp8_sad16x16x3_ssse3_aligned_by_15: |
+ PROCESS_16X16X3_OFFSET 0, .vpx_sad16x16x3_ssse3 |
+ PROCESS_16X16X3_OFFSET 1, .vpx_sad16x16x3_ssse3 |
+ PROCESS_16X16X3_OFFSET 2, .vpx_sad16x16x3_ssse3 |
+ PROCESS_16X16X3_OFFSET 3, .vpx_sad16x16x3_ssse3 |
+ PROCESS_16X16X3_OFFSET 4, .vpx_sad16x16x3_ssse3 |
+ PROCESS_16X16X3_OFFSET 5, .vpx_sad16x16x3_ssse3 |
+ PROCESS_16X16X3_OFFSET 6, .vpx_sad16x16x3_ssse3 |
+ PROCESS_16X16X3_OFFSET 7, .vpx_sad16x16x3_ssse3 |
+ PROCESS_16X16X3_OFFSET 8, .vpx_sad16x16x3_ssse3 |
+ PROCESS_16X16X3_OFFSET 9, .vpx_sad16x16x3_ssse3 |
+ PROCESS_16X16X3_OFFSET 10, .vpx_sad16x16x3_ssse3 |
+ PROCESS_16X16X3_OFFSET 11, .vpx_sad16x16x3_ssse3 |
+ PROCESS_16X16X3_OFFSET 12, .vpx_sad16x16x3_ssse3 |
+ PROCESS_16X16X3_OFFSET 13, .vpx_sad16x16x3_ssse3 |
+ PROCESS_16X16X3_OFFSET 14, .vpx_sad16x16x3_ssse3 |
+ |
+.vpx_sad16x16x3_ssse3_aligned_by_15: |
PROCESS_16X2X3 1 |
PROCESS_16X2X3 0 |
PROCESS_16X2X3 0 |
@@ -229,7 +229,7 @@ sym(vp8_sad16x16x3_ssse3): |
PROCESS_16X2X3 0 |
PROCESS_16X2X3 0 |
-.vp8_sad16x16x3_ssse3_store_off: |
+.vpx_sad16x16x3_ssse3_store_off: |
mov rdi, arg(4) ;Results |
movq xmm0, xmm5 |
@@ -259,14 +259,14 @@ sym(vp8_sad16x16x3_ssse3): |
pop rbp |
ret |
-;void int vp8_sad16x8x3_ssse3( |
+;void int vpx_sad16x8x3_ssse3( |
; unsigned char *src_ptr, |
; int src_stride, |
; unsigned char *ref_ptr, |
; int ref_stride, |
; int *results) |
-global sym(vp8_sad16x8x3_ssse3) PRIVATE |
-sym(vp8_sad16x8x3_ssse3): |
+global sym(vpx_sad16x8x3_ssse3) PRIVATE |
+sym(vpx_sad16x8x3_ssse3): |
push rbp |
mov rbp, rsp |
SHADOW_ARGS_TO_STACK 5 |
@@ -282,31 +282,31 @@ sym(vp8_sad16x8x3_ssse3): |
mov rdx, 0xf |
and rdx, rdi |
- jmp .vp8_sad16x8x3_ssse3_skiptable |
-.vp8_sad16x8x3_ssse3_jumptable: |
- dd .vp8_sad16x8x3_ssse3_aligned_by_0 - .vp8_sad16x8x3_ssse3_do_jump |
- dd .vp8_sad16x8x3_ssse3_aligned_by_1 - .vp8_sad16x8x3_ssse3_do_jump |
- dd .vp8_sad16x8x3_ssse3_aligned_by_2 - .vp8_sad16x8x3_ssse3_do_jump |
- dd .vp8_sad16x8x3_ssse3_aligned_by_3 - .vp8_sad16x8x3_ssse3_do_jump |
- dd .vp8_sad16x8x3_ssse3_aligned_by_4 - .vp8_sad16x8x3_ssse3_do_jump |
- dd .vp8_sad16x8x3_ssse3_aligned_by_5 - .vp8_sad16x8x3_ssse3_do_jump |
- dd .vp8_sad16x8x3_ssse3_aligned_by_6 - .vp8_sad16x8x3_ssse3_do_jump |
- dd .vp8_sad16x8x3_ssse3_aligned_by_7 - .vp8_sad16x8x3_ssse3_do_jump |
- dd .vp8_sad16x8x3_ssse3_aligned_by_8 - .vp8_sad16x8x3_ssse3_do_jump |
- dd .vp8_sad16x8x3_ssse3_aligned_by_9 - .vp8_sad16x8x3_ssse3_do_jump |
- dd .vp8_sad16x8x3_ssse3_aligned_by_10 - .vp8_sad16x8x3_ssse3_do_jump |
- dd .vp8_sad16x8x3_ssse3_aligned_by_11 - .vp8_sad16x8x3_ssse3_do_jump |
- dd .vp8_sad16x8x3_ssse3_aligned_by_12 - .vp8_sad16x8x3_ssse3_do_jump |
- dd .vp8_sad16x8x3_ssse3_aligned_by_13 - .vp8_sad16x8x3_ssse3_do_jump |
- dd .vp8_sad16x8x3_ssse3_aligned_by_14 - .vp8_sad16x8x3_ssse3_do_jump |
- dd .vp8_sad16x8x3_ssse3_aligned_by_15 - .vp8_sad16x8x3_ssse3_do_jump |
-.vp8_sad16x8x3_ssse3_skiptable: |
- |
- call .vp8_sad16x8x3_ssse3_do_jump |
-.vp8_sad16x8x3_ssse3_do_jump: |
+ jmp .vpx_sad16x8x3_ssse3_skiptable |
+.vpx_sad16x8x3_ssse3_jumptable: |
+ dd .vpx_sad16x8x3_ssse3_aligned_by_0 - .vpx_sad16x8x3_ssse3_do_jump |
+ dd .vpx_sad16x8x3_ssse3_aligned_by_1 - .vpx_sad16x8x3_ssse3_do_jump |
+ dd .vpx_sad16x8x3_ssse3_aligned_by_2 - .vpx_sad16x8x3_ssse3_do_jump |
+ dd .vpx_sad16x8x3_ssse3_aligned_by_3 - .vpx_sad16x8x3_ssse3_do_jump |
+ dd .vpx_sad16x8x3_ssse3_aligned_by_4 - .vpx_sad16x8x3_ssse3_do_jump |
+ dd .vpx_sad16x8x3_ssse3_aligned_by_5 - .vpx_sad16x8x3_ssse3_do_jump |
+ dd .vpx_sad16x8x3_ssse3_aligned_by_6 - .vpx_sad16x8x3_ssse3_do_jump |
+ dd .vpx_sad16x8x3_ssse3_aligned_by_7 - .vpx_sad16x8x3_ssse3_do_jump |
+ dd .vpx_sad16x8x3_ssse3_aligned_by_8 - .vpx_sad16x8x3_ssse3_do_jump |
+ dd .vpx_sad16x8x3_ssse3_aligned_by_9 - .vpx_sad16x8x3_ssse3_do_jump |
+ dd .vpx_sad16x8x3_ssse3_aligned_by_10 - .vpx_sad16x8x3_ssse3_do_jump |
+ dd .vpx_sad16x8x3_ssse3_aligned_by_11 - .vpx_sad16x8x3_ssse3_do_jump |
+ dd .vpx_sad16x8x3_ssse3_aligned_by_12 - .vpx_sad16x8x3_ssse3_do_jump |
+ dd .vpx_sad16x8x3_ssse3_aligned_by_13 - .vpx_sad16x8x3_ssse3_do_jump |
+ dd .vpx_sad16x8x3_ssse3_aligned_by_14 - .vpx_sad16x8x3_ssse3_do_jump |
+ dd .vpx_sad16x8x3_ssse3_aligned_by_15 - .vpx_sad16x8x3_ssse3_do_jump |
+.vpx_sad16x8x3_ssse3_skiptable: |
+ |
+ call .vpx_sad16x8x3_ssse3_do_jump |
+.vpx_sad16x8x3_ssse3_do_jump: |
pop rcx ; get the address of do_jump |
- mov rax, .vp8_sad16x8x3_ssse3_jumptable - .vp8_sad16x8x3_ssse3_do_jump |
- add rax, rcx ; get the absolute address of vp8_sad16x8x3_ssse3_jumptable |
+ mov rax, .vpx_sad16x8x3_ssse3_jumptable - .vpx_sad16x8x3_ssse3_do_jump |
+ add rax, rcx ; get the absolute address of vpx_sad16x8x3_ssse3_jumptable |
movsxd rax, dword [rax + 4*rdx] ; get the 32 bit offset from the jumptable |
add rcx, rax |
@@ -316,30 +316,30 @@ sym(vp8_sad16x8x3_ssse3): |
jmp rcx |
- PROCESS_16X8X3_OFFSET 0, .vp8_sad16x8x3_ssse3 |
- PROCESS_16X8X3_OFFSET 1, .vp8_sad16x8x3_ssse3 |
- PROCESS_16X8X3_OFFSET 2, .vp8_sad16x8x3_ssse3 |
- PROCESS_16X8X3_OFFSET 3, .vp8_sad16x8x3_ssse3 |
- PROCESS_16X8X3_OFFSET 4, .vp8_sad16x8x3_ssse3 |
- PROCESS_16X8X3_OFFSET 5, .vp8_sad16x8x3_ssse3 |
- PROCESS_16X8X3_OFFSET 6, .vp8_sad16x8x3_ssse3 |
- PROCESS_16X8X3_OFFSET 7, .vp8_sad16x8x3_ssse3 |
- PROCESS_16X8X3_OFFSET 8, .vp8_sad16x8x3_ssse3 |
- PROCESS_16X8X3_OFFSET 9, .vp8_sad16x8x3_ssse3 |
- PROCESS_16X8X3_OFFSET 10, .vp8_sad16x8x3_ssse3 |
- PROCESS_16X8X3_OFFSET 11, .vp8_sad16x8x3_ssse3 |
- PROCESS_16X8X3_OFFSET 12, .vp8_sad16x8x3_ssse3 |
- PROCESS_16X8X3_OFFSET 13, .vp8_sad16x8x3_ssse3 |
- PROCESS_16X8X3_OFFSET 14, .vp8_sad16x8x3_ssse3 |
- |
-.vp8_sad16x8x3_ssse3_aligned_by_15: |
+ PROCESS_16X8X3_OFFSET 0, .vpx_sad16x8x3_ssse3 |
+ PROCESS_16X8X3_OFFSET 1, .vpx_sad16x8x3_ssse3 |
+ PROCESS_16X8X3_OFFSET 2, .vpx_sad16x8x3_ssse3 |
+ PROCESS_16X8X3_OFFSET 3, .vpx_sad16x8x3_ssse3 |
+ PROCESS_16X8X3_OFFSET 4, .vpx_sad16x8x3_ssse3 |
+ PROCESS_16X8X3_OFFSET 5, .vpx_sad16x8x3_ssse3 |
+ PROCESS_16X8X3_OFFSET 6, .vpx_sad16x8x3_ssse3 |
+ PROCESS_16X8X3_OFFSET 7, .vpx_sad16x8x3_ssse3 |
+ PROCESS_16X8X3_OFFSET 8, .vpx_sad16x8x3_ssse3 |
+ PROCESS_16X8X3_OFFSET 9, .vpx_sad16x8x3_ssse3 |
+ PROCESS_16X8X3_OFFSET 10, .vpx_sad16x8x3_ssse3 |
+ PROCESS_16X8X3_OFFSET 11, .vpx_sad16x8x3_ssse3 |
+ PROCESS_16X8X3_OFFSET 12, .vpx_sad16x8x3_ssse3 |
+ PROCESS_16X8X3_OFFSET 13, .vpx_sad16x8x3_ssse3 |
+ PROCESS_16X8X3_OFFSET 14, .vpx_sad16x8x3_ssse3 |
+ |
+.vpx_sad16x8x3_ssse3_aligned_by_15: |
PROCESS_16X2X3 1 |
PROCESS_16X2X3 0 |
PROCESS_16X2X3 0 |
PROCESS_16X2X3 0 |
-.vp8_sad16x8x3_ssse3_store_off: |
+.vpx_sad16x8x3_ssse3_store_off: |
mov rdi, arg(4) ;Results |
movq xmm0, xmm5 |