| Index: source/libvpx/vp8/common/ppc/idctllm_altivec.asm
|
| diff --git a/source/libvpx/vp8/common/ppc/idctllm_altivec.asm b/source/libvpx/vp8/common/ppc/idctllm_altivec.asm
|
| deleted file mode 100644
|
| index 117d9cfc8e868ca1305172db4f5ed539f9e18acf..0000000000000000000000000000000000000000
|
| --- a/source/libvpx/vp8/common/ppc/idctllm_altivec.asm
|
| +++ /dev/null
|
| @@ -1,189 +0,0 @@
|
| -;
|
| -; Copyright (c) 2010 The WebM project authors. All Rights Reserved.
|
| -;
|
| -; Use of this source code is governed by a BSD-style license
|
| -; that can be found in the LICENSE file in the root of the source
|
| -; tree. An additional intellectual property rights grant can be found
|
| -; in the file PATENTS. All contributing project authors may
|
| -; be found in the AUTHORS file in the root of the source tree.
|
| -;
|
| -
|
| -
|
| - .globl short_idct4x4llm_ppc
|
| -
|
| -.macro load_c V, LABEL, OFF, R0, R1
|
| - lis \R0, \LABEL@ha
|
| - la \R1, \LABEL@l(\R0)
|
| - lvx \V, \OFF, \R1
|
| -.endm
|
| -
|
| -;# r3 short *input
|
| -;# r4 short *output
|
| -;# r5 int pitch
|
| - .align 2
|
| -short_idct4x4llm_ppc:
|
| - mfspr r11, 256 ;# get old VRSAVE
|
| - oris r12, r11, 0xfff8
|
| - mtspr 256, r12 ;# set VRSAVE
|
| -
|
| - load_c v8, sinpi8sqrt2, 0, r9, r10
|
| - load_c v9, cospi8sqrt2minus1, 0, r9, r10
|
| - load_c v10, hi_hi, 0, r9, r10
|
| - load_c v11, lo_lo, 0, r9, r10
|
| - load_c v12, shift_16, 0, r9, r10
|
| -
|
| - li r10, 16
|
| - lvx v0, 0, r3 ;# input ip[0], ip[ 4]
|
| - lvx v1, r10, r3 ;# input ip[8], ip[12]
|
| -
|
| - ;# first pass
|
| - vupkhsh v2, v0
|
| - vupkhsh v3, v1
|
| - vaddsws v6, v2, v3 ;# a1 = ip[0]+ip[8]
|
| - vsubsws v7, v2, v3 ;# b1 = ip[0]-ip[8]
|
| -
|
| - vupklsh v0, v0
|
| - vmulosh v4, v0, v8
|
| - vsraw v4, v4, v12
|
| - vaddsws v4, v4, v0 ;# ip[ 4] * sin(pi/8) * sqrt(2)
|
| -
|
| - vupklsh v1, v1
|
| - vmulosh v5, v1, v9
|
| - vsraw v5, v5, v12 ;# ip[12] * cos(pi/8) * sqrt(2)
|
| - vaddsws v5, v5, v1
|
| -
|
| - vsubsws v4, v4, v5 ;# c1
|
| -
|
| - vmulosh v3, v1, v8
|
| - vsraw v3, v3, v12
|
| - vaddsws v3, v3, v1 ;# ip[12] * sin(pi/8) * sqrt(2)
|
| -
|
| - vmulosh v5, v0, v9
|
| - vsraw v5, v5, v12 ;# ip[ 4] * cos(pi/8) * sqrt(2)
|
| - vaddsws v5, v5, v0
|
| -
|
| - vaddsws v3, v3, v5 ;# d1
|
| -
|
| - vaddsws v0, v6, v3 ;# a1 + d1
|
| - vsubsws v3, v6, v3 ;# a1 - d1
|
| -
|
| - vaddsws v1, v7, v4 ;# b1 + c1
|
| - vsubsws v2, v7, v4 ;# b1 - c1
|
| -
|
| - ;# transpose input
|
| - vmrghw v4, v0, v1 ;# a0 b0 a1 b1
|
| - vmrghw v5, v2, v3 ;# c0 d0 c1 d1
|
| -
|
| - vmrglw v6, v0, v1 ;# a2 b2 a3 b3
|
| - vmrglw v7, v2, v3 ;# c2 d2 c3 d3
|
| -
|
| - vperm v0, v4, v5, v10 ;# a0 b0 c0 d0
|
| - vperm v1, v4, v5, v11 ;# a1 b1 c1 d1
|
| -
|
| - vperm v2, v6, v7, v10 ;# a2 b2 c2 d2
|
| - vperm v3, v6, v7, v11 ;# a3 b3 c3 d3
|
| -
|
| - ;# second pass
|
| - vaddsws v6, v0, v2 ;# a1 = ip[0]+ip[8]
|
| - vsubsws v7, v0, v2 ;# b1 = ip[0]-ip[8]
|
| -
|
| - vmulosh v4, v1, v8
|
| - vsraw v4, v4, v12
|
| - vaddsws v4, v4, v1 ;# ip[ 4] * sin(pi/8) * sqrt(2)
|
| -
|
| - vmulosh v5, v3, v9
|
| - vsraw v5, v5, v12 ;# ip[12] * cos(pi/8) * sqrt(2)
|
| - vaddsws v5, v5, v3
|
| -
|
| - vsubsws v4, v4, v5 ;# c1
|
| -
|
| - vmulosh v2, v3, v8
|
| - vsraw v2, v2, v12
|
| - vaddsws v2, v2, v3 ;# ip[12] * sin(pi/8) * sqrt(2)
|
| -
|
| - vmulosh v5, v1, v9
|
| - vsraw v5, v5, v12 ;# ip[ 4] * cos(pi/8) * sqrt(2)
|
| - vaddsws v5, v5, v1
|
| -
|
| - vaddsws v3, v2, v5 ;# d1
|
| -
|
| - vaddsws v0, v6, v3 ;# a1 + d1
|
| - vsubsws v3, v6, v3 ;# a1 - d1
|
| -
|
| - vaddsws v1, v7, v4 ;# b1 + c1
|
| - vsubsws v2, v7, v4 ;# b1 - c1
|
| -
|
| - vspltish v6, 4
|
| - vspltish v7, 3
|
| -
|
| - vpkswss v0, v0, v1
|
| - vpkswss v1, v2, v3
|
| -
|
| - vaddshs v0, v0, v6
|
| - vaddshs v1, v1, v6
|
| -
|
| - vsrah v0, v0, v7
|
| - vsrah v1, v1, v7
|
| -
|
| - ;# transpose output
|
| - vmrghh v2, v0, v1 ;# a0 c0 a1 c1 a2 c2 a3 c3
|
| - vmrglh v3, v0, v1 ;# b0 d0 b1 d1 b2 d2 b3 d3
|
| -
|
| - vmrghh v0, v2, v3 ;# a0 b0 c0 d0 a1 b1 c1 d1
|
| - vmrglh v1, v2, v3 ;# a2 b2 c2 d2 a3 b3 c3 d3
|
| -
|
| - stwu r1,-416(r1) ;# create space on the stack
|
| -
|
| - stvx v0, 0, r1
|
| - lwz r6, 0(r1)
|
| - stw r6, 0(r4)
|
| - lwz r6, 4(r1)
|
| - stw r6, 4(r4)
|
| -
|
| - add r4, r4, r5
|
| -
|
| - lwz r6, 8(r1)
|
| - stw r6, 0(r4)
|
| - lwz r6, 12(r1)
|
| - stw r6, 4(r4)
|
| -
|
| - add r4, r4, r5
|
| -
|
| - stvx v1, 0, r1
|
| - lwz r6, 0(r1)
|
| - stw r6, 0(r4)
|
| - lwz r6, 4(r1)
|
| - stw r6, 4(r4)
|
| -
|
| - add r4, r4, r5
|
| -
|
| - lwz r6, 8(r1)
|
| - stw r6, 0(r4)
|
| - lwz r6, 12(r1)
|
| - stw r6, 4(r4)
|
| -
|
| - addi r1, r1, 416 ;# recover stack
|
| -
|
| - mtspr 256, r11 ;# reset old VRSAVE
|
| -
|
| - blr
|
| -
|
| - .align 4
|
| -sinpi8sqrt2:
|
| - .short 35468, 35468, 35468, 35468, 35468, 35468, 35468, 35468
|
| -
|
| - .align 4
|
| -cospi8sqrt2minus1:
|
| - .short 20091, 20091, 20091, 20091, 20091, 20091, 20091, 20091
|
| -
|
| - .align 4
|
| -shift_16:
|
| - .long 16, 16, 16, 16
|
| -
|
| - .align 4
|
| -hi_hi:
|
| - .byte 0, 1, 2, 3, 4, 5, 6, 7, 16, 17, 18, 19, 20, 21, 22, 23
|
| -
|
| - .align 4
|
| -lo_lo:
|
| - .byte 8, 9, 10, 11, 12, 13, 14, 15, 24, 25, 26, 27, 28, 29, 30, 31
|
|
|