Index: libvpx/source/libvpx/vp8/encoder/arm/armv6/vp8_variance_halfpixvar16x16_v_armv6.asm |
diff --git a/libvpx/source/libvpx/vp8/encoder/arm/armv6/vp8_variance_halfpixvar16x16_v_armv6.asm b/libvpx/source/libvpx/vp8/encoder/arm/armv6/vp8_variance_halfpixvar16x16_v_armv6.asm |
index 22a50eb002e8c4e158598b786a2e6973b03b5110..9e0a035482cff60a0b48c926941551315942b4d2 100644 |
--- a/libvpx/source/libvpx/vp8/encoder/arm/armv6/vp8_variance_halfpixvar16x16_v_armv6.asm |
+++ b/libvpx/source/libvpx/vp8/encoder/arm/armv6/vp8_variance_halfpixvar16x16_v_armv6.asm |
@@ -25,10 +25,6 @@ |
|vp8_variance_halfpixvar16x16_v_armv6| PROC |
stmfd sp!, {r4-r12, lr} |
- |
- pld [r0, r1, lsl #0] |
- pld [r2, r3, lsl #0] |
- |
mov r8, #0 ; initialize sum = 0 |
ldr r10, c80808080 |
mov r11, #0 ; initialize sse = 0 |
@@ -47,10 +43,8 @@ loop |
eor r4, r4, r10 |
usub8 r6, r4, r5 ; calculate difference |
- pld [r0, r1, lsl #1] |
sel r7, r6, lr ; select bytes with positive difference |
usub8 r6, r5, r4 ; calculate difference with reversed operands |
- pld [r2, r3, lsl #1] |
sel r6, r6, lr ; select bytes with negative difference |
; calculate partial sums |