| Index: source/libvpx/vp9/common/arm/neon/vp9_avg_neon_asm.asm
|
| ===================================================================
|
| --- source/libvpx/vp9/common/arm/neon/vp9_avg_neon_asm.asm (revision 0)
|
| +++ source/libvpx/vp9/common/arm/neon/vp9_avg_neon_asm.asm (working copy)
|
| @@ -0,0 +1,116 @@
|
| +;
|
| +; Copyright (c) 2013 The WebM project authors. All Rights Reserved.
|
| +;
|
| +; Use of this source code is governed by a BSD-style license
|
| +; that can be found in the LICENSE file in the root of the source
|
| +; tree. An additional intellectual property rights grant can be found
|
| +; in the file PATENTS. All contributing project authors may
|
| +; be found in the AUTHORS file in the root of the source tree.
|
| +;
|
| +
|
| + EXPORT |vp9_convolve_avg_neon|
|
| + ARM
|
| + REQUIRE8
|
| + PRESERVE8
|
| +
|
| + AREA ||.text||, CODE, READONLY, ALIGN=2
|
| +
|
| +|vp9_convolve_avg_neon| PROC
|
| + push {r4-r6, lr}
|
| + ldrd r4, r5, [sp, #32]
|
| + mov r6, r2
|
| +
|
| + cmp r4, #32
|
| + bgt avg64
|
| + beq avg32
|
| + cmp r4, #8
|
| + bgt avg16
|
| + beq avg8
|
| + b avg4
|
| +
|
| +avg64
|
| + sub lr, r1, #32
|
| + sub r4, r3, #32
|
| +avg64_h
|
| + pld [r0, r1, lsl #1]
|
| + vld1.8 {q0-q1}, [r0]!
|
| + vld1.8 {q2-q3}, [r0], lr
|
| + pld [r2, r3]
|
| + vld1.8 {q8-q9}, [r6@128]!
|
| + vld1.8 {q10-q11}, [r6@128], r4
|
| + vrhadd.u8 q0, q0, q8
|
| + vrhadd.u8 q1, q1, q9
|
| + vrhadd.u8 q2, q2, q10
|
| + vrhadd.u8 q3, q3, q11
|
| + vst1.8 {q0-q1}, [r2@128]!
|
| + vst1.8 {q2-q3}, [r2@128], r4
|
| + subs r5, r5, #1
|
| + bgt avg64_h
|
| + pop {r4-r6, pc}
|
| +
|
| +avg32
|
| + vld1.8 {q0-q1}, [r0], r1
|
| + vld1.8 {q2-q3}, [r0], r1
|
| + vld1.8 {q8-q9}, [r6@128], r3
|
| + vld1.8 {q10-q11}, [r6@128], r3
|
| + pld [r0]
|
| + vrhadd.u8 q0, q0, q8
|
| + pld [r0, r1]
|
| + vrhadd.u8 q1, q1, q9
|
| + pld [r6]
|
| + vrhadd.u8 q2, q2, q10
|
| + pld [r6, r3]
|
| + vrhadd.u8 q3, q3, q11
|
| + vst1.8 {q0-q1}, [r2@128], r3
|
| + vst1.8 {q2-q3}, [r2@128], r3
|
| + subs r5, r5, #2
|
| + bgt avg32
|
| + pop {r4-r6, pc}
|
| +
|
| +avg16
|
| + vld1.8 {q0}, [r0], r1
|
| + vld1.8 {q1}, [r0], r1
|
| + vld1.8 {q2}, [r6@128], r3
|
| + vld1.8 {q3}, [r6@128], r3
|
| + pld [r0]
|
| + pld [r0, r1]
|
| + vrhadd.u8 q0, q0, q2
|
| + pld [r6]
|
| + pld [r6, r3]
|
| + vrhadd.u8 q1, q1, q3
|
| + vst1.8 {q0}, [r2@128], r3
|
| + vst1.8 {q1}, [r2@128], r3
|
| + subs r5, r5, #2
|
| + bgt avg16
|
| + pop {r4-r6, pc}
|
| +
|
| +avg8
|
| + vld1.8 {d0}, [r0], r1
|
| + vld1.8 {d1}, [r0], r1
|
| + vld1.8 {d2}, [r6@64], r3
|
| + vld1.8 {d3}, [r6@64], r3
|
| + pld [r0]
|
| + pld [r0, r1]
|
| + vrhadd.u8 q0, q0, q1
|
| + pld [r6]
|
| + pld [r6, r3]
|
| + vst1.8 {d0}, [r2@64], r3
|
| + vst1.8 {d1}, [r2@64], r3
|
| + subs r5, r5, #2
|
| + bgt avg8
|
| + pop {r4-r6, pc}
|
| +
|
| +avg4
|
| + vld1.32 {d0[0]}, [r0], r1
|
| + vld1.32 {d0[1]}, [r0], r1
|
| + vld1.32 {d2[0]}, [r6@32], r3
|
| + vld1.32 {d2[1]}, [r6@32], r3
|
| + vrhadd.u8 d0, d0, d2
|
| + vst1.32 {d0[0]}, [r2@32], r3
|
| + vst1.32 {d0[1]}, [r2@32], r3
|
| + subs r5, r5, #2
|
| + bgt avg4
|
| + pop {r4-r6, pc}
|
| + ENDP
|
| +
|
| + END
|
|
|