| Index: source/libvpx/vp8/common/arm/armv6/vp8_variance8x8_armv6.asm
|
| diff --git a/source/libvpx/vp8/common/arm/armv6/vp8_variance8x8_armv6.asm b/source/libvpx/vp8/common/arm/armv6/vp8_variance8x8_armv6.asm
|
| deleted file mode 100644
|
| index 915ee499309146a05efd912542ff5d788dff53a3..0000000000000000000000000000000000000000
|
| --- a/source/libvpx/vp8/common/arm/armv6/vp8_variance8x8_armv6.asm
|
| +++ /dev/null
|
| @@ -1,101 +0,0 @@
|
| -;
|
| -; Copyright (c) 2011 The WebM project authors. All Rights Reserved.
|
| -;
|
| -; Use of this source code is governed by a BSD-style license
|
| -; that can be found in the LICENSE file in the root of the source
|
| -; tree. An additional intellectual property rights grant can be found
|
| -; in the file PATENTS. All contributing project authors may
|
| -; be found in the AUTHORS file in the root of the source tree.
|
| -;
|
| -
|
| -
|
| - EXPORT |vp8_variance8x8_armv6|
|
| -
|
| - ARM
|
| -
|
| - AREA ||.text||, CODE, READONLY, ALIGN=2
|
| -
|
| -; r0 unsigned char *src_ptr
|
| -; r1 int source_stride
|
| -; r2 unsigned char *ref_ptr
|
| -; r3 int recon_stride
|
| -; stack unsigned int *sse
|
| -|vp8_variance8x8_armv6| PROC
|
| -
|
| - push {r4-r10, lr}
|
| -
|
| - pld [r0, r1, lsl #0]
|
| - pld [r2, r3, lsl #0]
|
| -
|
| - mov r12, #8 ; set loop counter to 8 (=block height)
|
| - mov r4, #0 ; initialize sum = 0
|
| - mov r5, #0 ; initialize sse = 0
|
| -
|
| -loop
|
| - ; 1st 4 pixels
|
| - ldr r6, [r0, #0x0] ; load 4 src pixels
|
| - ldr r7, [r2, #0x0] ; load 4 ref pixels
|
| -
|
| - mov lr, #0 ; constant zero
|
| -
|
| - usub8 r8, r6, r7 ; calculate difference
|
| - pld [r0, r1, lsl #1]
|
| - sel r10, r8, lr ; select bytes with positive difference
|
| - usub8 r9, r7, r6 ; calculate difference with reversed operands
|
| - pld [r2, r3, lsl #1]
|
| - sel r8, r9, lr ; select bytes with negative difference
|
| -
|
| - ; calculate partial sums
|
| - usad8 r6, r10, lr ; calculate sum of positive differences
|
| - usad8 r7, r8, lr ; calculate sum of negative differences
|
| - orr r8, r8, r10 ; differences of all 4 pixels
|
| - ; calculate total sum
|
| - add r4, r4, r6 ; add positive differences to sum
|
| - sub r4, r4, r7 ; subtract negative differences from sum
|
| -
|
| - ; calculate sse
|
| - uxtb16 r7, r8 ; byte (two pixels) to halfwords
|
| - uxtb16 r10, r8, ror #8 ; another two pixels to halfwords
|
| - smlad r5, r7, r7, r5 ; dual signed multiply, add and accumulate (1)
|
| -
|
| - ; 2nd 4 pixels
|
| - ldr r6, [r0, #0x4] ; load 4 src pixels
|
| - ldr r7, [r2, #0x4] ; load 4 ref pixels
|
| - smlad r5, r10, r10, r5 ; dual signed multiply, add and accumulate (2)
|
| -
|
| - usub8 r8, r6, r7 ; calculate difference
|
| - add r0, r0, r1 ; set src_ptr to next row
|
| - sel r10, r8, lr ; select bytes with positive difference
|
| - usub8 r9, r7, r6 ; calculate difference with reversed operands
|
| - add r2, r2, r3 ; set dst_ptr to next row
|
| - sel r8, r9, lr ; select bytes with negative difference
|
| -
|
| - ; calculate partial sums
|
| - usad8 r6, r10, lr ; calculate sum of positive differences
|
| - usad8 r7, r8, lr ; calculate sum of negative differences
|
| - orr r8, r8, r10 ; differences of all 4 pixels
|
| -
|
| - ; calculate total sum
|
| - add r4, r4, r6 ; add positive differences to sum
|
| - sub r4, r4, r7 ; subtract negative differences from sum
|
| -
|
| - ; calculate sse
|
| - uxtb16 r7, r8 ; byte (two pixels) to halfwords
|
| - uxtb16 r10, r8, ror #8 ; another two pixels to halfwords
|
| - smlad r5, r7, r7, r5 ; dual signed multiply, add and accumulate (1)
|
| - subs r12, r12, #1 ; next row
|
| - smlad r5, r10, r10, r5 ; dual signed multiply, add and accumulate (2)
|
| -
|
| - bne loop
|
| -
|
| - ; return stuff
|
| - ldr r8, [sp, #32] ; get address of sse
|
| - mul r1, r4, r4 ; sum * sum
|
| - str r5, [r8] ; store sse
|
| - sub r0, r5, r1, ASR #6 ; return (sse - ((sum * sum) >> 6))
|
| -
|
| - pop {r4-r10, pc}
|
| -
|
| - ENDP
|
| -
|
| - END
|
|
|