OLD | NEW |
1 ; | 1 ; |
2 ; Copyright (c) 2010 The WebM project authors. All Rights Reserved. | 2 ; Copyright (c) 2010 The WebM project authors. All Rights Reserved. |
3 ; | 3 ; |
4 ; Use of this source code is governed by a BSD-style license | 4 ; Use of this source code is governed by a BSD-style license |
5 ; that can be found in the LICENSE file in the root of the source | 5 ; that can be found in the LICENSE file in the root of the source |
6 ; tree. An additional intellectual property rights grant can be found | 6 ; tree. An additional intellectual property rights grant can be found |
7 ; in the file PATENTS. All contributing project authors may | 7 ; in the file PATENTS. All contributing project authors may |
8 ; be found in the AUTHORS file in the root of the source tree. | 8 ; be found in the AUTHORS file in the root of the source tree. |
9 ; | 9 ; |
10 | 10 |
(...skipping 132 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
143 %define limit r9 | 143 %define limit r9 |
144 %define thresh r10 | 144 %define thresh r10 |
145 | 145 |
146 %define spp rax | 146 %define spp rax |
147 %define stride3 r11 | 147 %define stride3 r11 |
148 %define stride5 r12 | 148 %define stride5 r12 |
149 %define stride7 r13 | 149 %define stride7 r13 |
150 | 150 |
151 push rbp | 151 push rbp |
152 mov rbp, rsp | 152 mov rbp, rsp |
| 153 SAVE_XMM 11 |
153 push r12 | 154 push r12 |
154 push r13 | 155 push r13 |
155 mov thresh, arg(4) | 156 mov thresh, arg(4) |
156 %else | 157 %else |
157 %define src rdi ; src_ptr | 158 %define src rdi ; src_ptr |
158 %define stride rsi ; src_pixel_step | 159 %define stride rsi ; src_pixel_step |
159 %define blimit rdx | 160 %define blimit rdx |
160 %define limit rcx | 161 %define limit rcx |
161 %define thresh r8 | 162 %define thresh r8 |
162 | 163 |
(...skipping 88 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
251 movdqa xmm8, i13 | 252 movdqa xmm8, i13 |
252 LF_FILTER xmm0, xmm1, xmm3, xmm8, xmm4, xmm2 | 253 LF_FILTER xmm0, xmm1, xmm3, xmm8, xmm4, xmm2 |
253 movdqa i10, xmm0 | 254 movdqa i10, xmm0 |
254 movdqa i11, xmm1 | 255 movdqa i11, xmm1 |
255 movdqa i12, xmm3 | 256 movdqa i12, xmm3 |
256 movdqa i13, xmm8 | 257 movdqa i13, xmm8 |
257 | 258 |
258 %ifidn __OUTPUT_FORMAT__,x64 | 259 %ifidn __OUTPUT_FORMAT__,x64 |
259 pop r13 | 260 pop r13 |
260 pop r12 | 261 pop r12 |
| 262 RESTORE_XMM |
261 pop rbp | 263 pop rbp |
262 %endif | 264 %endif |
263 | 265 |
264 ret | 266 ret |
265 | 267 |
266 | 268 |
267 ;void vp8_loop_filter_bv_y_sse2 | 269 ;void vp8_loop_filter_bv_y_sse2 |
268 ;( | 270 ;( |
269 ; unsigned char *src_ptr, | 271 ; unsigned char *src_ptr, |
270 ; int src_pixel_step, | 272 ; int src_pixel_step, |
(...skipping 533 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
804 times 16 db 0x80 | 806 times 16 db 0x80 |
805 align 16 | 807 align 16 |
806 t1: | 808 t1: |
807 times 16 db 0x01 | 809 times 16 db 0x01 |
808 align 16 | 810 align 16 |
809 t3: | 811 t3: |
810 times 16 db 0x03 | 812 times 16 db 0x03 |
811 align 16 | 813 align 16 |
812 t4: | 814 t4: |
813 times 16 db 0x04 | 815 times 16 db 0x04 |
OLD | NEW |