| OLD | NEW |
| 1 /* | 1 /* |
| 2 * Copyright (c) 2010 The WebM project authors. All Rights Reserved. | 2 * Copyright (c) 2010 The WebM project authors. All Rights Reserved. |
| 3 * | 3 * |
| 4 * Use of this source code is governed by a BSD-style license | 4 * Use of this source code is governed by a BSD-style license |
| 5 * that can be found in the LICENSE file in the root of the source | 5 * that can be found in the LICENSE file in the root of the source |
| 6 * tree. An additional intellectual property rights grant can be found | 6 * tree. An additional intellectual property rights grant can be found |
| 7 * in the file PATENTS. All contributing project authors may | 7 * in the file PATENTS. All contributing project authors may |
| 8 * be found in the AUTHORS file in the root of the source tree. | 8 * be found in the AUTHORS file in the root of the source tree. |
| 9 */ | 9 */ |
| 10 | 10 |
| (...skipping 30 matching lines...) Expand all Loading... |
| 41 void (*vp8_short_fdct8x4)(short *input, short *output, int pitch); | 41 void (*vp8_short_fdct8x4)(short *input, short *output, int pitch); |
| 42 void (*vp8_fast_fdct4x4)(short *input, short *output, int pitch); | 42 void (*vp8_fast_fdct4x4)(short *input, short *output, int pitch); |
| 43 void (*vp8_fast_fdct8x4)(short *input, short *output, int pitch); | 43 void (*vp8_fast_fdct8x4)(short *input, short *output, int pitch); |
| 44 void (*short_walsh4x4)(short *input, short *output, int pitch); | 44 void (*short_walsh4x4)(short *input, short *output, int pitch); |
| 45 | 45 |
| 46 void (*vp8_subtract_b)(BLOCK *be, BLOCKD *bd, int pitch); | 46 void (*vp8_subtract_b)(BLOCK *be, BLOCKD *bd, int pitch); |
| 47 void (*vp8_subtract_mby)(short *diff, unsigned char *src, unsigned char *pred, i
nt stride); | 47 void (*vp8_subtract_mby)(short *diff, unsigned char *src, unsigned char *pred, i
nt stride); |
| 48 void (*vp8_subtract_mbuv)(short *diff, unsigned char *usrc, unsigned char *vsrc,
unsigned char *pred, int stride); | 48 void (*vp8_subtract_mbuv)(short *diff, unsigned char *usrc, unsigned char *vsrc,
unsigned char *pred, int stride); |
| 49 void (*vp8_fast_quantize_b)(BLOCK *b, BLOCKD *d); | 49 void (*vp8_fast_quantize_b)(BLOCK *b, BLOCKD *d); |
| 50 | 50 |
| 51 unsigned int (*vp8_get16x16pred_error)(unsigned char *src_ptr, int src_stride, u
nsigned char *ref_ptr, int ref_stride); | |
| 52 unsigned int (*vp8_get8x8var)(unsigned char *src_ptr, int source_stride, unsign
ed char *ref_ptr, int recon_stride, unsigned int *SSE, int *Sum); | |
| 53 unsigned int (*vp8_get16x16var)(unsigned char *src_ptr, int source_stride, unsi
gned char *ref_ptr, int recon_stride, unsigned int *SSE, int *Sum); | |
| 54 unsigned int (*vp8_get4x4sse_cs)(unsigned char *src_ptr, int source_stride, uns
igned char *ref_ptr, int recon_stride); | 51 unsigned int (*vp8_get4x4sse_cs)(unsigned char *src_ptr, int source_stride, uns
igned char *ref_ptr, int recon_stride); |
| 55 | 52 |
| 56 // c imports | 53 // c imports |
| 57 extern int block_error_c(short *coeff, short *dqcoeff); | 54 extern int block_error_c(short *coeff, short *dqcoeff); |
| 58 extern int vp8_mbblock_error_c(MACROBLOCK *mb, int dc); | 55 extern int vp8_mbblock_error_c(MACROBLOCK *mb, int dc); |
| 59 | 56 |
| 60 extern int vp8_mbuverror_c(MACROBLOCK *mb); | 57 extern int vp8_mbuverror_c(MACROBLOCK *mb); |
| 61 extern unsigned int vp8_get8x8var_c(unsigned char *src_ptr, int source_stride,
unsigned char *ref_ptr, int recon_stride, unsigned int *SSE, int *Sum); | 58 extern unsigned int vp8_get8x8var_c(unsigned char *src_ptr, int source_stride,
unsigned char *ref_ptr, int recon_stride, unsigned int *SSE, int *Sum); |
| 62 extern void short_fdct4x4_c(short *input, short *output, int pitch); | 59 extern void short_fdct4x4_c(short *input, short *output, int pitch); |
| 63 extern void short_fdct8x4_c(short *input, short *output, int pitch); | 60 extern void short_fdct8x4_c(short *input, short *output, int pitch); |
| (...skipping 17 matching lines...) Expand all Loading... |
| 81 extern variance_function variance4x4_c; | 78 extern variance_function variance4x4_c; |
| 82 extern variance_function mse16x16_c; | 79 extern variance_function mse16x16_c; |
| 83 | 80 |
| 84 extern sub_pixel_variance_function sub_pixel_variance4x4_c; | 81 extern sub_pixel_variance_function sub_pixel_variance4x4_c; |
| 85 extern sub_pixel_variance_function sub_pixel_variance8x8_c; | 82 extern sub_pixel_variance_function sub_pixel_variance8x8_c; |
| 86 extern sub_pixel_variance_function sub_pixel_variance8x16_c; | 83 extern sub_pixel_variance_function sub_pixel_variance8x16_c; |
| 87 extern sub_pixel_variance_function sub_pixel_variance16x8_c; | 84 extern sub_pixel_variance_function sub_pixel_variance16x8_c; |
| 88 extern sub_pixel_variance_function sub_pixel_variance16x16_c; | 85 extern sub_pixel_variance_function sub_pixel_variance16x16_c; |
| 89 | 86 |
| 90 extern unsigned int vp8_get_mb_ss_c(short *); | 87 extern unsigned int vp8_get_mb_ss_c(short *); |
| 91 extern unsigned int vp8_get16x16pred_error_c(unsigned char *src_ptr, int src_str
ide, unsigned char *ref_ptr, int ref_stride); | |
| 92 extern unsigned int vp8_get8x8var_c(unsigned char *src_ptr, int source_stride,
unsigned char *ref_ptr, int recon_stride, unsigned int *SSE, int *Sum); | |
| 93 extern unsigned int vp8_get16x16var_c(unsigned char *src_ptr, int source_stride
, unsigned char *ref_ptr, int recon_stride, unsigned int *SSE, int *Sum); | |
| 94 extern unsigned int vp8_get4x4sse_cs_c(unsigned char *src_ptr, int source_strid
e, unsigned char *ref_ptr, int recon_stride); | 88 extern unsigned int vp8_get4x4sse_cs_c(unsigned char *src_ptr, int source_strid
e, unsigned char *ref_ptr, int recon_stride); |
| 95 | 89 |
| 96 // ppc | 90 // ppc |
| 97 extern int vp8_block_error_ppc(short *coeff, short *dqcoeff); | 91 extern int vp8_block_error_ppc(short *coeff, short *dqcoeff); |
| 98 | 92 |
| 99 extern void vp8_short_fdct4x4_ppc(short *input, short *output, int pitch); | 93 extern void vp8_short_fdct4x4_ppc(short *input, short *output, int pitch); |
| 100 extern void vp8_short_fdct8x4_ppc(short *input, short *output, int pitch); | 94 extern void vp8_short_fdct8x4_ppc(short *input, short *output, int pitch); |
| 101 | 95 |
| 102 extern void vp8_subtract_mby_ppc(short *diff, unsigned char *src, unsigned char
*pred, int stride); | 96 extern void vp8_subtract_mby_ppc(short *diff, unsigned char *src, unsigned char
*pred, int stride); |
| 103 extern void vp8_subtract_mbuv_ppc(short *diff, unsigned char *usrc, unsigned cha
r *vsrc, unsigned char *pred, int stride); | 97 extern void vp8_subtract_mbuv_ppc(short *diff, unsigned char *usrc, unsigned cha
r *vsrc, unsigned char *pred, int stride); |
| (...skipping 38 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 142 vp8_variance16x16 = vp8_variance16x16_ppc; | 136 vp8_variance16x16 = vp8_variance16x16_ppc; |
| 143 vp8_mse16x16 = vp8_mse16x16_ppc; | 137 vp8_mse16x16 = vp8_mse16x16_ppc; |
| 144 | 138 |
| 145 vp8_sub_pixel_variance4x4 = vp8_sub_pixel_variance4x4_ppc; | 139 vp8_sub_pixel_variance4x4 = vp8_sub_pixel_variance4x4_ppc; |
| 146 vp8_sub_pixel_variance8x8 = vp8_sub_pixel_variance8x8_ppc; | 140 vp8_sub_pixel_variance8x8 = vp8_sub_pixel_variance8x8_ppc; |
| 147 vp8_sub_pixel_variance8x16 = vp8_sub_pixel_variance8x16_ppc; | 141 vp8_sub_pixel_variance8x16 = vp8_sub_pixel_variance8x16_ppc; |
| 148 vp8_sub_pixel_variance16x8 = vp8_sub_pixel_variance16x8_ppc; | 142 vp8_sub_pixel_variance16x8 = vp8_sub_pixel_variance16x8_ppc; |
| 149 vp8_sub_pixel_variance16x16 = vp8_sub_pixel_variance16x16_ppc; | 143 vp8_sub_pixel_variance16x16 = vp8_sub_pixel_variance16x16_ppc; |
| 150 | 144 |
| 151 vp8_get_mb_ss = vp8_get_mb_ss_c; | 145 vp8_get_mb_ss = vp8_get_mb_ss_c; |
| 152 vp8_get16x16pred_error = vp8_get16x16pred_error_c; | |
| 153 vp8_get8x8var = vp8_get8x8var_ppc; | |
| 154 vp8_get16x16var = vp8_get16x16var_ppc; | |
| 155 vp8_get4x4sse_cs = vp8_get4x4sse_cs_c; | 146 vp8_get4x4sse_cs = vp8_get4x4sse_cs_c; |
| 156 | 147 |
| 157 vp8_sad16x16 = vp8_sad16x16_ppc; | 148 vp8_sad16x16 = vp8_sad16x16_ppc; |
| 158 vp8_sad16x8 = vp8_sad16x8_ppc; | 149 vp8_sad16x8 = vp8_sad16x8_ppc; |
| 159 vp8_sad8x16 = vp8_sad8x16_ppc; | 150 vp8_sad8x16 = vp8_sad8x16_ppc; |
| 160 vp8_sad8x8 = vp8_sad8x8_ppc; | 151 vp8_sad8x8 = vp8_sad8x8_ppc; |
| 161 vp8_sad4x4 = vp8_sad4x4_ppc; | 152 vp8_sad4x4 = vp8_sad4x4_ppc; |
| 162 | 153 |
| 163 vp8_block_error = vp8_block_error_ppc; | 154 vp8_block_error = vp8_block_error_ppc; |
| 164 vp8_mbblock_error = vp8_mbblock_error_c; | 155 vp8_mbblock_error = vp8_mbblock_error_c; |
| 165 | 156 |
| 166 vp8_subtract_b = vp8_subtract_b_c; | 157 vp8_subtract_b = vp8_subtract_b_c; |
| 167 vp8_subtract_mby = vp8_subtract_mby_ppc; | 158 vp8_subtract_mby = vp8_subtract_mby_ppc; |
| 168 vp8_subtract_mbuv = vp8_subtract_mbuv_ppc; | 159 vp8_subtract_mbuv = vp8_subtract_mbuv_ppc; |
| 169 } | 160 } |
| OLD | NEW |