| OLD | NEW |
| 1 #ifndef VPX_DSP_RTCD_H_ | 1 #ifndef VPX_DSP_RTCD_H_ |
| 2 #define VPX_DSP_RTCD_H_ | 2 #define VPX_DSP_RTCD_H_ |
| 3 | 3 |
| 4 #ifdef RTCD_C | 4 #ifdef RTCD_C |
| 5 #define RTCD_EXTERN | 5 #define RTCD_EXTERN |
| 6 #else | 6 #else |
| 7 #define RTCD_EXTERN extern | 7 #define RTCD_EXTERN extern |
| 8 #endif | 8 #endif |
| 9 | 9 |
| 10 /* | 10 /* |
| (...skipping 389 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 400 RTCD_EXTERN void (*vpx_int_pro_row)(int16_t *hbuf, const uint8_t *ref, const int
ref_stride, const int height); | 400 RTCD_EXTERN void (*vpx_int_pro_row)(int16_t *hbuf, const uint8_t *ref, const int
ref_stride, const int height); |
| 401 | 401 |
| 402 void vpx_iwht4x4_16_add_c(const tran_low_t *input, uint8_t *dest, int dest_strid
e); | 402 void vpx_iwht4x4_16_add_c(const tran_low_t *input, uint8_t *dest, int dest_strid
e); |
| 403 void vpx_iwht4x4_16_add_sse2(const tran_low_t *input, uint8_t *dest, int dest_st
ride); | 403 void vpx_iwht4x4_16_add_sse2(const tran_low_t *input, uint8_t *dest, int dest_st
ride); |
| 404 RTCD_EXTERN void (*vpx_iwht4x4_16_add)(const tran_low_t *input, uint8_t *dest, i
nt dest_stride); | 404 RTCD_EXTERN void (*vpx_iwht4x4_16_add)(const tran_low_t *input, uint8_t *dest, i
nt dest_stride); |
| 405 | 405 |
| 406 void vpx_iwht4x4_1_add_c(const tran_low_t *input, uint8_t *dest, int dest_stride
); | 406 void vpx_iwht4x4_1_add_c(const tran_low_t *input, uint8_t *dest, int dest_stride
); |
| 407 #define vpx_iwht4x4_1_add vpx_iwht4x4_1_add_c | 407 #define vpx_iwht4x4_1_add vpx_iwht4x4_1_add_c |
| 408 | 408 |
| 409 void vpx_lpf_horizontal_4_c(uint8_t *s, int pitch, const uint8_t *blimit, const
uint8_t *limit, const uint8_t *thresh); | 409 void vpx_lpf_horizontal_4_c(uint8_t *s, int pitch, const uint8_t *blimit, const
uint8_t *limit, const uint8_t *thresh); |
| 410 void vpx_lpf_horizontal_4_mmx(uint8_t *s, int pitch, const uint8_t *blimit, cons
t uint8_t *limit, const uint8_t *thresh); | 410 void vpx_lpf_horizontal_4_sse2(uint8_t *s, int pitch, const uint8_t *blimit, con
st uint8_t *limit, const uint8_t *thresh); |
| 411 RTCD_EXTERN void (*vpx_lpf_horizontal_4)(uint8_t *s, int pitch, const uint8_t *b
limit, const uint8_t *limit, const uint8_t *thresh); | 411 RTCD_EXTERN void (*vpx_lpf_horizontal_4)(uint8_t *s, int pitch, const uint8_t *b
limit, const uint8_t *limit, const uint8_t *thresh); |
| 412 | 412 |
| 413 void vpx_lpf_horizontal_4_dual_c(uint8_t *s, int pitch, const uint8_t *blimit0,
const uint8_t *limit0, const uint8_t *thresh0, const uint8_t *blimit1, const uin
t8_t *limit1, const uint8_t *thresh1); | 413 void vpx_lpf_horizontal_4_dual_c(uint8_t *s, int pitch, const uint8_t *blimit0,
const uint8_t *limit0, const uint8_t *thresh0, const uint8_t *blimit1, const uin
t8_t *limit1, const uint8_t *thresh1); |
| 414 void vpx_lpf_horizontal_4_dual_sse2(uint8_t *s, int pitch, const uint8_t *blimit
0, const uint8_t *limit0, const uint8_t *thresh0, const uint8_t *blimit1, const
uint8_t *limit1, const uint8_t *thresh1); | 414 void vpx_lpf_horizontal_4_dual_sse2(uint8_t *s, int pitch, const uint8_t *blimit
0, const uint8_t *limit0, const uint8_t *thresh0, const uint8_t *blimit1, const
uint8_t *limit1, const uint8_t *thresh1); |
| 415 RTCD_EXTERN void (*vpx_lpf_horizontal_4_dual)(uint8_t *s, int pitch, const uint8
_t *blimit0, const uint8_t *limit0, const uint8_t *thresh0, const uint8_t *blimi
t1, const uint8_t *limit1, const uint8_t *thresh1); | 415 RTCD_EXTERN void (*vpx_lpf_horizontal_4_dual)(uint8_t *s, int pitch, const uint8
_t *blimit0, const uint8_t *limit0, const uint8_t *thresh0, const uint8_t *blimi
t1, const uint8_t *limit1, const uint8_t *thresh1); |
| 416 | 416 |
| 417 void vpx_lpf_horizontal_8_c(uint8_t *s, int pitch, const uint8_t *blimit, const
uint8_t *limit, const uint8_t *thresh); | 417 void vpx_lpf_horizontal_8_c(uint8_t *s, int pitch, const uint8_t *blimit, const
uint8_t *limit, const uint8_t *thresh); |
| 418 void vpx_lpf_horizontal_8_sse2(uint8_t *s, int pitch, const uint8_t *blimit, con
st uint8_t *limit, const uint8_t *thresh); | 418 void vpx_lpf_horizontal_8_sse2(uint8_t *s, int pitch, const uint8_t *blimit, con
st uint8_t *limit, const uint8_t *thresh); |
| 419 RTCD_EXTERN void (*vpx_lpf_horizontal_8)(uint8_t *s, int pitch, const uint8_t *b
limit, const uint8_t *limit, const uint8_t *thresh); | 419 RTCD_EXTERN void (*vpx_lpf_horizontal_8)(uint8_t *s, int pitch, const uint8_t *b
limit, const uint8_t *limit, const uint8_t *thresh); |
| 420 | 420 |
| (...skipping 13 matching lines...) Expand all Loading... |
| 434 | 434 |
| 435 void vpx_lpf_vertical_16_c(uint8_t *s, int pitch, const uint8_t *blimit, const u
int8_t *limit, const uint8_t *thresh); | 435 void vpx_lpf_vertical_16_c(uint8_t *s, int pitch, const uint8_t *blimit, const u
int8_t *limit, const uint8_t *thresh); |
| 436 void vpx_lpf_vertical_16_sse2(uint8_t *s, int pitch, const uint8_t *blimit, cons
t uint8_t *limit, const uint8_t *thresh); | 436 void vpx_lpf_vertical_16_sse2(uint8_t *s, int pitch, const uint8_t *blimit, cons
t uint8_t *limit, const uint8_t *thresh); |
| 437 RTCD_EXTERN void (*vpx_lpf_vertical_16)(uint8_t *s, int pitch, const uint8_t *bl
imit, const uint8_t *limit, const uint8_t *thresh); | 437 RTCD_EXTERN void (*vpx_lpf_vertical_16)(uint8_t *s, int pitch, const uint8_t *bl
imit, const uint8_t *limit, const uint8_t *thresh); |
| 438 | 438 |
| 439 void vpx_lpf_vertical_16_dual_c(uint8_t *s, int pitch, const uint8_t *blimit, co
nst uint8_t *limit, const uint8_t *thresh); | 439 void vpx_lpf_vertical_16_dual_c(uint8_t *s, int pitch, const uint8_t *blimit, co
nst uint8_t *limit, const uint8_t *thresh); |
| 440 void vpx_lpf_vertical_16_dual_sse2(uint8_t *s, int pitch, const uint8_t *blimit,
const uint8_t *limit, const uint8_t *thresh); | 440 void vpx_lpf_vertical_16_dual_sse2(uint8_t *s, int pitch, const uint8_t *blimit,
const uint8_t *limit, const uint8_t *thresh); |
| 441 RTCD_EXTERN void (*vpx_lpf_vertical_16_dual)(uint8_t *s, int pitch, const uint8_
t *blimit, const uint8_t *limit, const uint8_t *thresh); | 441 RTCD_EXTERN void (*vpx_lpf_vertical_16_dual)(uint8_t *s, int pitch, const uint8_
t *blimit, const uint8_t *limit, const uint8_t *thresh); |
| 442 | 442 |
| 443 void vpx_lpf_vertical_4_c(uint8_t *s, int pitch, const uint8_t *blimit, const ui
nt8_t *limit, const uint8_t *thresh); | 443 void vpx_lpf_vertical_4_c(uint8_t *s, int pitch, const uint8_t *blimit, const ui
nt8_t *limit, const uint8_t *thresh); |
| 444 void vpx_lpf_vertical_4_mmx(uint8_t *s, int pitch, const uint8_t *blimit, const
uint8_t *limit, const uint8_t *thresh); | 444 void vpx_lpf_vertical_4_sse2(uint8_t *s, int pitch, const uint8_t *blimit, const
uint8_t *limit, const uint8_t *thresh); |
| 445 RTCD_EXTERN void (*vpx_lpf_vertical_4)(uint8_t *s, int pitch, const uint8_t *bli
mit, const uint8_t *limit, const uint8_t *thresh); | 445 RTCD_EXTERN void (*vpx_lpf_vertical_4)(uint8_t *s, int pitch, const uint8_t *bli
mit, const uint8_t *limit, const uint8_t *thresh); |
| 446 | 446 |
| 447 void vpx_lpf_vertical_4_dual_c(uint8_t *s, int pitch, const uint8_t *blimit0, co
nst uint8_t *limit0, const uint8_t *thresh0, const uint8_t *blimit1, const uint8
_t *limit1, const uint8_t *thresh1); | 447 void vpx_lpf_vertical_4_dual_c(uint8_t *s, int pitch, const uint8_t *blimit0, co
nst uint8_t *limit0, const uint8_t *thresh0, const uint8_t *blimit1, const uint8
_t *limit1, const uint8_t *thresh1); |
| 448 void vpx_lpf_vertical_4_dual_sse2(uint8_t *s, int pitch, const uint8_t *blimit0,
const uint8_t *limit0, const uint8_t *thresh0, const uint8_t *blimit1, const ui
nt8_t *limit1, const uint8_t *thresh1); | 448 void vpx_lpf_vertical_4_dual_sse2(uint8_t *s, int pitch, const uint8_t *blimit0,
const uint8_t *limit0, const uint8_t *thresh0, const uint8_t *blimit1, const ui
nt8_t *limit1, const uint8_t *thresh1); |
| 449 RTCD_EXTERN void (*vpx_lpf_vertical_4_dual)(uint8_t *s, int pitch, const uint8_t
*blimit0, const uint8_t *limit0, const uint8_t *thresh0, const uint8_t *blimit1
, const uint8_t *limit1, const uint8_t *thresh1); | 449 RTCD_EXTERN void (*vpx_lpf_vertical_4_dual)(uint8_t *s, int pitch, const uint8_t
*blimit0, const uint8_t *limit0, const uint8_t *thresh0, const uint8_t *blimit1
, const uint8_t *limit1, const uint8_t *thresh1); |
| 450 | 450 |
| 451 void vpx_lpf_vertical_8_c(uint8_t *s, int pitch, const uint8_t *blimit, const ui
nt8_t *limit, const uint8_t *thresh); | 451 void vpx_lpf_vertical_8_c(uint8_t *s, int pitch, const uint8_t *blimit, const ui
nt8_t *limit, const uint8_t *thresh); |
| 452 void vpx_lpf_vertical_8_sse2(uint8_t *s, int pitch, const uint8_t *blimit, const
uint8_t *limit, const uint8_t *thresh); | 452 void vpx_lpf_vertical_8_sse2(uint8_t *s, int pitch, const uint8_t *blimit, const
uint8_t *limit, const uint8_t *thresh); |
| 453 RTCD_EXTERN void (*vpx_lpf_vertical_8)(uint8_t *s, int pitch, const uint8_t *bli
mit, const uint8_t *limit, const uint8_t *thresh); | 453 RTCD_EXTERN void (*vpx_lpf_vertical_8)(uint8_t *s, int pitch, const uint8_t *bli
mit, const uint8_t *limit, const uint8_t *thresh); |
| 454 | 454 |
| (...skipping 310 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 765 uint32_t vpx_sub_pixel_avg_variance32x32_ssse3(const uint8_t *src_ptr, int sourc
e_stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uin
t32_t *sse, const uint8_t *second_pred); | 765 uint32_t vpx_sub_pixel_avg_variance32x32_ssse3(const uint8_t *src_ptr, int sourc
e_stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uin
t32_t *sse, const uint8_t *second_pred); |
| 766 uint32_t vpx_sub_pixel_avg_variance32x32_avx2(const uint8_t *src_ptr, int source
_stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint
32_t *sse, const uint8_t *second_pred); | 766 uint32_t vpx_sub_pixel_avg_variance32x32_avx2(const uint8_t *src_ptr, int source
_stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint
32_t *sse, const uint8_t *second_pred); |
| 767 RTCD_EXTERN uint32_t (*vpx_sub_pixel_avg_variance32x32)(const uint8_t *src_ptr,
int source_stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_st
ride, uint32_t *sse, const uint8_t *second_pred); | 767 RTCD_EXTERN uint32_t (*vpx_sub_pixel_avg_variance32x32)(const uint8_t *src_ptr,
int source_stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_st
ride, uint32_t *sse, const uint8_t *second_pred); |
| 768 | 768 |
| 769 uint32_t vpx_sub_pixel_avg_variance32x64_c(const uint8_t *src_ptr, int source_st
ride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_
t *sse, const uint8_t *second_pred); | 769 uint32_t vpx_sub_pixel_avg_variance32x64_c(const uint8_t *src_ptr, int source_st
ride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_
t *sse, const uint8_t *second_pred); |
| 770 uint32_t vpx_sub_pixel_avg_variance32x64_sse2(const uint8_t *src_ptr, int source
_stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint
32_t *sse, const uint8_t *second_pred); | 770 uint32_t vpx_sub_pixel_avg_variance32x64_sse2(const uint8_t *src_ptr, int source
_stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint
32_t *sse, const uint8_t *second_pred); |
| 771 uint32_t vpx_sub_pixel_avg_variance32x64_ssse3(const uint8_t *src_ptr, int sourc
e_stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uin
t32_t *sse, const uint8_t *second_pred); | 771 uint32_t vpx_sub_pixel_avg_variance32x64_ssse3(const uint8_t *src_ptr, int sourc
e_stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uin
t32_t *sse, const uint8_t *second_pred); |
| 772 RTCD_EXTERN uint32_t (*vpx_sub_pixel_avg_variance32x64)(const uint8_t *src_ptr,
int source_stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_st
ride, uint32_t *sse, const uint8_t *second_pred); | 772 RTCD_EXTERN uint32_t (*vpx_sub_pixel_avg_variance32x64)(const uint8_t *src_ptr,
int source_stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_st
ride, uint32_t *sse, const uint8_t *second_pred); |
| 773 | 773 |
| 774 uint32_t vpx_sub_pixel_avg_variance4x4_c(const uint8_t *src_ptr, int source_stri
de, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_t
*sse, const uint8_t *second_pred); | 774 uint32_t vpx_sub_pixel_avg_variance4x4_c(const uint8_t *src_ptr, int source_stri
de, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_t
*sse, const uint8_t *second_pred); |
| 775 uint32_t vpx_sub_pixel_avg_variance4x4_sse(const uint8_t *src_ptr, int source_st
ride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_
t *sse, const uint8_t *second_pred); | 775 uint32_t vpx_sub_pixel_avg_variance4x4_sse2(const uint8_t *src_ptr, int source_s
tride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32
_t *sse, const uint8_t *second_pred); |
| 776 uint32_t vpx_sub_pixel_avg_variance4x4_ssse3(const uint8_t *src_ptr, int source_
stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint3
2_t *sse, const uint8_t *second_pred); | 776 uint32_t vpx_sub_pixel_avg_variance4x4_ssse3(const uint8_t *src_ptr, int source_
stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint3
2_t *sse, const uint8_t *second_pred); |
| 777 RTCD_EXTERN uint32_t (*vpx_sub_pixel_avg_variance4x4)(const uint8_t *src_ptr, in
t source_stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stri
de, uint32_t *sse, const uint8_t *second_pred); | 777 RTCD_EXTERN uint32_t (*vpx_sub_pixel_avg_variance4x4)(const uint8_t *src_ptr, in
t source_stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stri
de, uint32_t *sse, const uint8_t *second_pred); |
| 778 | 778 |
| 779 uint32_t vpx_sub_pixel_avg_variance4x8_c(const uint8_t *src_ptr, int source_stri
de, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_t
*sse, const uint8_t *second_pred); | 779 uint32_t vpx_sub_pixel_avg_variance4x8_c(const uint8_t *src_ptr, int source_stri
de, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_t
*sse, const uint8_t *second_pred); |
| 780 uint32_t vpx_sub_pixel_avg_variance4x8_sse(const uint8_t *src_ptr, int source_st
ride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_
t *sse, const uint8_t *second_pred); | 780 uint32_t vpx_sub_pixel_avg_variance4x8_sse2(const uint8_t *src_ptr, int source_s
tride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32
_t *sse, const uint8_t *second_pred); |
| 781 uint32_t vpx_sub_pixel_avg_variance4x8_ssse3(const uint8_t *src_ptr, int source_
stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint3
2_t *sse, const uint8_t *second_pred); | 781 uint32_t vpx_sub_pixel_avg_variance4x8_ssse3(const uint8_t *src_ptr, int source_
stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint3
2_t *sse, const uint8_t *second_pred); |
| 782 RTCD_EXTERN uint32_t (*vpx_sub_pixel_avg_variance4x8)(const uint8_t *src_ptr, in
t source_stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stri
de, uint32_t *sse, const uint8_t *second_pred); | 782 RTCD_EXTERN uint32_t (*vpx_sub_pixel_avg_variance4x8)(const uint8_t *src_ptr, in
t source_stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stri
de, uint32_t *sse, const uint8_t *second_pred); |
| 783 | 783 |
| 784 uint32_t vpx_sub_pixel_avg_variance64x32_c(const uint8_t *src_ptr, int source_st
ride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_
t *sse, const uint8_t *second_pred); | 784 uint32_t vpx_sub_pixel_avg_variance64x32_c(const uint8_t *src_ptr, int source_st
ride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_
t *sse, const uint8_t *second_pred); |
| 785 uint32_t vpx_sub_pixel_avg_variance64x32_sse2(const uint8_t *src_ptr, int source
_stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint
32_t *sse, const uint8_t *second_pred); | 785 uint32_t vpx_sub_pixel_avg_variance64x32_sse2(const uint8_t *src_ptr, int source
_stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint
32_t *sse, const uint8_t *second_pred); |
| 786 uint32_t vpx_sub_pixel_avg_variance64x32_ssse3(const uint8_t *src_ptr, int sourc
e_stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uin
t32_t *sse, const uint8_t *second_pred); | 786 uint32_t vpx_sub_pixel_avg_variance64x32_ssse3(const uint8_t *src_ptr, int sourc
e_stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uin
t32_t *sse, const uint8_t *second_pred); |
| 787 RTCD_EXTERN uint32_t (*vpx_sub_pixel_avg_variance64x32)(const uint8_t *src_ptr,
int source_stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_st
ride, uint32_t *sse, const uint8_t *second_pred); | 787 RTCD_EXTERN uint32_t (*vpx_sub_pixel_avg_variance64x32)(const uint8_t *src_ptr,
int source_stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_st
ride, uint32_t *sse, const uint8_t *second_pred); |
| 788 | 788 |
| 789 uint32_t vpx_sub_pixel_avg_variance64x64_c(const uint8_t *src_ptr, int source_st
ride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_
t *sse, const uint8_t *second_pred); | 789 uint32_t vpx_sub_pixel_avg_variance64x64_c(const uint8_t *src_ptr, int source_st
ride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_
t *sse, const uint8_t *second_pred); |
| 790 uint32_t vpx_sub_pixel_avg_variance64x64_sse2(const uint8_t *src_ptr, int source
_stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint
32_t *sse, const uint8_t *second_pred); | 790 uint32_t vpx_sub_pixel_avg_variance64x64_sse2(const uint8_t *src_ptr, int source
_stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint
32_t *sse, const uint8_t *second_pred); |
| (...skipping 44 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 835 uint32_t vpx_sub_pixel_variance32x32_avx2(const uint8_t *src_ptr, int source_str
ide, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_t
*sse); | 835 uint32_t vpx_sub_pixel_variance32x32_avx2(const uint8_t *src_ptr, int source_str
ide, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_t
*sse); |
| 836 RTCD_EXTERN uint32_t (*vpx_sub_pixel_variance32x32)(const uint8_t *src_ptr, int
source_stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride
, uint32_t *sse); | 836 RTCD_EXTERN uint32_t (*vpx_sub_pixel_variance32x32)(const uint8_t *src_ptr, int
source_stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride
, uint32_t *sse); |
| 837 | 837 |
| 838 uint32_t vpx_sub_pixel_variance32x64_c(const uint8_t *src_ptr, int source_stride
, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_t *s
se); | 838 uint32_t vpx_sub_pixel_variance32x64_c(const uint8_t *src_ptr, int source_stride
, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_t *s
se); |
| 839 uint32_t vpx_sub_pixel_variance32x64_sse2(const uint8_t *src_ptr, int source_str
ide, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_t
*sse); | 839 uint32_t vpx_sub_pixel_variance32x64_sse2(const uint8_t *src_ptr, int source_str
ide, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_t
*sse); |
| 840 uint32_t vpx_sub_pixel_variance32x64_ssse3(const uint8_t *src_ptr, int source_st
ride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_
t *sse); | 840 uint32_t vpx_sub_pixel_variance32x64_ssse3(const uint8_t *src_ptr, int source_st
ride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_
t *sse); |
| 841 RTCD_EXTERN uint32_t (*vpx_sub_pixel_variance32x64)(const uint8_t *src_ptr, int
source_stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride
, uint32_t *sse); | 841 RTCD_EXTERN uint32_t (*vpx_sub_pixel_variance32x64)(const uint8_t *src_ptr, int
source_stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride
, uint32_t *sse); |
| 842 | 842 |
| 843 uint32_t vpx_sub_pixel_variance4x4_c(const uint8_t *src_ptr, int source_stride,
int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_t *sse
); | 843 uint32_t vpx_sub_pixel_variance4x4_c(const uint8_t *src_ptr, int source_stride,
int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_t *sse
); |
| 844 uint32_t vpx_sub_pixel_variance4x4_mmx(const uint8_t *src_ptr, int source_stride
, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_t *s
se); | 844 uint32_t vpx_sub_pixel_variance4x4_mmx(const uint8_t *src_ptr, int source_stride
, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_t *s
se); |
| 845 uint32_t vpx_sub_pixel_variance4x4_sse(const uint8_t *src_ptr, int source_stride
, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_t *s
se); | 845 uint32_t vpx_sub_pixel_variance4x4_sse2(const uint8_t *src_ptr, int source_strid
e, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_t *
sse); |
| 846 uint32_t vpx_sub_pixel_variance4x4_ssse3(const uint8_t *src_ptr, int source_stri
de, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_t
*sse); | 846 uint32_t vpx_sub_pixel_variance4x4_ssse3(const uint8_t *src_ptr, int source_stri
de, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_t
*sse); |
| 847 RTCD_EXTERN uint32_t (*vpx_sub_pixel_variance4x4)(const uint8_t *src_ptr, int so
urce_stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride,
uint32_t *sse); | 847 RTCD_EXTERN uint32_t (*vpx_sub_pixel_variance4x4)(const uint8_t *src_ptr, int so
urce_stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride,
uint32_t *sse); |
| 848 | 848 |
| 849 uint32_t vpx_sub_pixel_variance4x8_c(const uint8_t *src_ptr, int source_stride,
int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_t *sse
); | 849 uint32_t vpx_sub_pixel_variance4x8_c(const uint8_t *src_ptr, int source_stride,
int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_t *sse
); |
| 850 uint32_t vpx_sub_pixel_variance4x8_sse(const uint8_t *src_ptr, int source_stride
, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_t *s
se); | 850 uint32_t vpx_sub_pixel_variance4x8_sse2(const uint8_t *src_ptr, int source_strid
e, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_t *
sse); |
| 851 uint32_t vpx_sub_pixel_variance4x8_ssse3(const uint8_t *src_ptr, int source_stri
de, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_t
*sse); | 851 uint32_t vpx_sub_pixel_variance4x8_ssse3(const uint8_t *src_ptr, int source_stri
de, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_t
*sse); |
| 852 RTCD_EXTERN uint32_t (*vpx_sub_pixel_variance4x8)(const uint8_t *src_ptr, int so
urce_stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride,
uint32_t *sse); | 852 RTCD_EXTERN uint32_t (*vpx_sub_pixel_variance4x8)(const uint8_t *src_ptr, int so
urce_stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride,
uint32_t *sse); |
| 853 | 853 |
| 854 uint32_t vpx_sub_pixel_variance64x32_c(const uint8_t *src_ptr, int source_stride
, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_t *s
se); | 854 uint32_t vpx_sub_pixel_variance64x32_c(const uint8_t *src_ptr, int source_stride
, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_t *s
se); |
| 855 uint32_t vpx_sub_pixel_variance64x32_sse2(const uint8_t *src_ptr, int source_str
ide, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_t
*sse); | 855 uint32_t vpx_sub_pixel_variance64x32_sse2(const uint8_t *src_ptr, int source_str
ide, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_t
*sse); |
| 856 uint32_t vpx_sub_pixel_variance64x32_ssse3(const uint8_t *src_ptr, int source_st
ride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_
t *sse); | 856 uint32_t vpx_sub_pixel_variance64x32_ssse3(const uint8_t *src_ptr, int source_st
ride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_
t *sse); |
| 857 RTCD_EXTERN uint32_t (*vpx_sub_pixel_variance64x32)(const uint8_t *src_ptr, int
source_stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride
, uint32_t *sse); | 857 RTCD_EXTERN uint32_t (*vpx_sub_pixel_variance64x32)(const uint8_t *src_ptr, int
source_stride, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride
, uint32_t *sse); |
| 858 | 858 |
| 859 uint32_t vpx_sub_pixel_variance64x64_c(const uint8_t *src_ptr, int source_stride
, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_t *s
se); | 859 uint32_t vpx_sub_pixel_variance64x64_c(const uint8_t *src_ptr, int source_stride
, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_t *s
se); |
| 860 uint32_t vpx_sub_pixel_variance64x64_sse2(const uint8_t *src_ptr, int source_str
ide, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_t
*sse); | 860 uint32_t vpx_sub_pixel_variance64x64_sse2(const uint8_t *src_ptr, int source_str
ide, int xoffset, int yoffset, const uint8_t *ref_ptr, int ref_stride, uint32_t
*sse); |
| (...skipping 299 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 1160 if (flags & HAS_SSE2) vpx_idct8x8_1_add = vpx_idct8x8_1_add_sse2; | 1160 if (flags & HAS_SSE2) vpx_idct8x8_1_add = vpx_idct8x8_1_add_sse2; |
| 1161 vpx_idct8x8_64_add = vpx_idct8x8_64_add_c; | 1161 vpx_idct8x8_64_add = vpx_idct8x8_64_add_c; |
| 1162 if (flags & HAS_SSE2) vpx_idct8x8_64_add = vpx_idct8x8_64_add_sse2; | 1162 if (flags & HAS_SSE2) vpx_idct8x8_64_add = vpx_idct8x8_64_add_sse2; |
| 1163 vpx_int_pro_col = vpx_int_pro_col_c; | 1163 vpx_int_pro_col = vpx_int_pro_col_c; |
| 1164 if (flags & HAS_SSE2) vpx_int_pro_col = vpx_int_pro_col_sse2; | 1164 if (flags & HAS_SSE2) vpx_int_pro_col = vpx_int_pro_col_sse2; |
| 1165 vpx_int_pro_row = vpx_int_pro_row_c; | 1165 vpx_int_pro_row = vpx_int_pro_row_c; |
| 1166 if (flags & HAS_SSE2) vpx_int_pro_row = vpx_int_pro_row_sse2; | 1166 if (flags & HAS_SSE2) vpx_int_pro_row = vpx_int_pro_row_sse2; |
| 1167 vpx_iwht4x4_16_add = vpx_iwht4x4_16_add_c; | 1167 vpx_iwht4x4_16_add = vpx_iwht4x4_16_add_c; |
| 1168 if (flags & HAS_SSE2) vpx_iwht4x4_16_add = vpx_iwht4x4_16_add_sse2; | 1168 if (flags & HAS_SSE2) vpx_iwht4x4_16_add = vpx_iwht4x4_16_add_sse2; |
| 1169 vpx_lpf_horizontal_4 = vpx_lpf_horizontal_4_c; | 1169 vpx_lpf_horizontal_4 = vpx_lpf_horizontal_4_c; |
| 1170 if (flags & HAS_MMX) vpx_lpf_horizontal_4 = vpx_lpf_horizontal_4_mmx; | 1170 if (flags & HAS_SSE2) vpx_lpf_horizontal_4 = vpx_lpf_horizontal_4_sse2; |
| 1171 vpx_lpf_horizontal_4_dual = vpx_lpf_horizontal_4_dual_c; | 1171 vpx_lpf_horizontal_4_dual = vpx_lpf_horizontal_4_dual_c; |
| 1172 if (flags & HAS_SSE2) vpx_lpf_horizontal_4_dual = vpx_lpf_horizontal_4_dual_
sse2; | 1172 if (flags & HAS_SSE2) vpx_lpf_horizontal_4_dual = vpx_lpf_horizontal_4_dual_
sse2; |
| 1173 vpx_lpf_horizontal_8 = vpx_lpf_horizontal_8_c; | 1173 vpx_lpf_horizontal_8 = vpx_lpf_horizontal_8_c; |
| 1174 if (flags & HAS_SSE2) vpx_lpf_horizontal_8 = vpx_lpf_horizontal_8_sse2; | 1174 if (flags & HAS_SSE2) vpx_lpf_horizontal_8 = vpx_lpf_horizontal_8_sse2; |
| 1175 vpx_lpf_horizontal_8_dual = vpx_lpf_horizontal_8_dual_c; | 1175 vpx_lpf_horizontal_8_dual = vpx_lpf_horizontal_8_dual_c; |
| 1176 if (flags & HAS_SSE2) vpx_lpf_horizontal_8_dual = vpx_lpf_horizontal_8_dual_
sse2; | 1176 if (flags & HAS_SSE2) vpx_lpf_horizontal_8_dual = vpx_lpf_horizontal_8_dual_
sse2; |
| 1177 vpx_lpf_horizontal_edge_16 = vpx_lpf_horizontal_edge_16_c; | 1177 vpx_lpf_horizontal_edge_16 = vpx_lpf_horizontal_edge_16_c; |
| 1178 if (flags & HAS_SSE2) vpx_lpf_horizontal_edge_16 = vpx_lpf_horizontal_edge_1
6_sse2; | 1178 if (flags & HAS_SSE2) vpx_lpf_horizontal_edge_16 = vpx_lpf_horizontal_edge_1
6_sse2; |
| 1179 if (flags & HAS_AVX2) vpx_lpf_horizontal_edge_16 = vpx_lpf_horizontal_edge_1
6_avx2; | 1179 if (flags & HAS_AVX2) vpx_lpf_horizontal_edge_16 = vpx_lpf_horizontal_edge_1
6_avx2; |
| 1180 vpx_lpf_horizontal_edge_8 = vpx_lpf_horizontal_edge_8_c; | 1180 vpx_lpf_horizontal_edge_8 = vpx_lpf_horizontal_edge_8_c; |
| 1181 if (flags & HAS_SSE2) vpx_lpf_horizontal_edge_8 = vpx_lpf_horizontal_edge_8_
sse2; | 1181 if (flags & HAS_SSE2) vpx_lpf_horizontal_edge_8 = vpx_lpf_horizontal_edge_8_
sse2; |
| 1182 if (flags & HAS_AVX2) vpx_lpf_horizontal_edge_8 = vpx_lpf_horizontal_edge_8_
avx2; | 1182 if (flags & HAS_AVX2) vpx_lpf_horizontal_edge_8 = vpx_lpf_horizontal_edge_8_
avx2; |
| 1183 vpx_lpf_vertical_16 = vpx_lpf_vertical_16_c; | 1183 vpx_lpf_vertical_16 = vpx_lpf_vertical_16_c; |
| 1184 if (flags & HAS_SSE2) vpx_lpf_vertical_16 = vpx_lpf_vertical_16_sse2; | 1184 if (flags & HAS_SSE2) vpx_lpf_vertical_16 = vpx_lpf_vertical_16_sse2; |
| 1185 vpx_lpf_vertical_16_dual = vpx_lpf_vertical_16_dual_c; | 1185 vpx_lpf_vertical_16_dual = vpx_lpf_vertical_16_dual_c; |
| 1186 if (flags & HAS_SSE2) vpx_lpf_vertical_16_dual = vpx_lpf_vertical_16_dual_ss
e2; | 1186 if (flags & HAS_SSE2) vpx_lpf_vertical_16_dual = vpx_lpf_vertical_16_dual_ss
e2; |
| 1187 vpx_lpf_vertical_4 = vpx_lpf_vertical_4_c; | 1187 vpx_lpf_vertical_4 = vpx_lpf_vertical_4_c; |
| 1188 if (flags & HAS_MMX) vpx_lpf_vertical_4 = vpx_lpf_vertical_4_mmx; | 1188 if (flags & HAS_SSE2) vpx_lpf_vertical_4 = vpx_lpf_vertical_4_sse2; |
| 1189 vpx_lpf_vertical_4_dual = vpx_lpf_vertical_4_dual_c; | 1189 vpx_lpf_vertical_4_dual = vpx_lpf_vertical_4_dual_c; |
| 1190 if (flags & HAS_SSE2) vpx_lpf_vertical_4_dual = vpx_lpf_vertical_4_dual_sse2
; | 1190 if (flags & HAS_SSE2) vpx_lpf_vertical_4_dual = vpx_lpf_vertical_4_dual_sse2
; |
| 1191 vpx_lpf_vertical_8 = vpx_lpf_vertical_8_c; | 1191 vpx_lpf_vertical_8 = vpx_lpf_vertical_8_c; |
| 1192 if (flags & HAS_SSE2) vpx_lpf_vertical_8 = vpx_lpf_vertical_8_sse2; | 1192 if (flags & HAS_SSE2) vpx_lpf_vertical_8 = vpx_lpf_vertical_8_sse2; |
| 1193 vpx_lpf_vertical_8_dual = vpx_lpf_vertical_8_dual_c; | 1193 vpx_lpf_vertical_8_dual = vpx_lpf_vertical_8_dual_c; |
| 1194 if (flags & HAS_SSE2) vpx_lpf_vertical_8_dual = vpx_lpf_vertical_8_dual_sse2
; | 1194 if (flags & HAS_SSE2) vpx_lpf_vertical_8_dual = vpx_lpf_vertical_8_dual_sse2
; |
| 1195 vpx_minmax_8x8 = vpx_minmax_8x8_c; | 1195 vpx_minmax_8x8 = vpx_minmax_8x8_c; |
| 1196 if (flags & HAS_SSE2) vpx_minmax_8x8 = vpx_minmax_8x8_sse2; | 1196 if (flags & HAS_SSE2) vpx_minmax_8x8 = vpx_minmax_8x8_sse2; |
| 1197 vpx_mse16x16 = vpx_mse16x16_c; | 1197 vpx_mse16x16 = vpx_mse16x16_c; |
| 1198 if (flags & HAS_SSE2) vpx_mse16x16 = vpx_mse16x16_sse2; | 1198 if (flags & HAS_SSE2) vpx_mse16x16 = vpx_mse16x16_sse2; |
| (...skipping 138 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 1337 if (flags & HAS_SSE2) vpx_sub_pixel_avg_variance32x16 = vpx_sub_pixel_avg_va
riance32x16_sse2; | 1337 if (flags & HAS_SSE2) vpx_sub_pixel_avg_variance32x16 = vpx_sub_pixel_avg_va
riance32x16_sse2; |
| 1338 if (flags & HAS_SSSE3) vpx_sub_pixel_avg_variance32x16 = vpx_sub_pixel_avg_v
ariance32x16_ssse3; | 1338 if (flags & HAS_SSSE3) vpx_sub_pixel_avg_variance32x16 = vpx_sub_pixel_avg_v
ariance32x16_ssse3; |
| 1339 vpx_sub_pixel_avg_variance32x32 = vpx_sub_pixel_avg_variance32x32_c; | 1339 vpx_sub_pixel_avg_variance32x32 = vpx_sub_pixel_avg_variance32x32_c; |
| 1340 if (flags & HAS_SSE2) vpx_sub_pixel_avg_variance32x32 = vpx_sub_pixel_avg_va
riance32x32_sse2; | 1340 if (flags & HAS_SSE2) vpx_sub_pixel_avg_variance32x32 = vpx_sub_pixel_avg_va
riance32x32_sse2; |
| 1341 if (flags & HAS_SSSE3) vpx_sub_pixel_avg_variance32x32 = vpx_sub_pixel_avg_v
ariance32x32_ssse3; | 1341 if (flags & HAS_SSSE3) vpx_sub_pixel_avg_variance32x32 = vpx_sub_pixel_avg_v
ariance32x32_ssse3; |
| 1342 if (flags & HAS_AVX2) vpx_sub_pixel_avg_variance32x32 = vpx_sub_pixel_avg_va
riance32x32_avx2; | 1342 if (flags & HAS_AVX2) vpx_sub_pixel_avg_variance32x32 = vpx_sub_pixel_avg_va
riance32x32_avx2; |
| 1343 vpx_sub_pixel_avg_variance32x64 = vpx_sub_pixel_avg_variance32x64_c; | 1343 vpx_sub_pixel_avg_variance32x64 = vpx_sub_pixel_avg_variance32x64_c; |
| 1344 if (flags & HAS_SSE2) vpx_sub_pixel_avg_variance32x64 = vpx_sub_pixel_avg_va
riance32x64_sse2; | 1344 if (flags & HAS_SSE2) vpx_sub_pixel_avg_variance32x64 = vpx_sub_pixel_avg_va
riance32x64_sse2; |
| 1345 if (flags & HAS_SSSE3) vpx_sub_pixel_avg_variance32x64 = vpx_sub_pixel_avg_v
ariance32x64_ssse3; | 1345 if (flags & HAS_SSSE3) vpx_sub_pixel_avg_variance32x64 = vpx_sub_pixel_avg_v
ariance32x64_ssse3; |
| 1346 vpx_sub_pixel_avg_variance4x4 = vpx_sub_pixel_avg_variance4x4_c; | 1346 vpx_sub_pixel_avg_variance4x4 = vpx_sub_pixel_avg_variance4x4_c; |
| 1347 if (flags & HAS_SSE) vpx_sub_pixel_avg_variance4x4 = vpx_sub_pixel_avg_varia
nce4x4_sse; | 1347 if (flags & HAS_SSE2) vpx_sub_pixel_avg_variance4x4 = vpx_sub_pixel_avg_vari
ance4x4_sse2; |
| 1348 if (flags & HAS_SSSE3) vpx_sub_pixel_avg_variance4x4 = vpx_sub_pixel_avg_var
iance4x4_ssse3; | 1348 if (flags & HAS_SSSE3) vpx_sub_pixel_avg_variance4x4 = vpx_sub_pixel_avg_var
iance4x4_ssse3; |
| 1349 vpx_sub_pixel_avg_variance4x8 = vpx_sub_pixel_avg_variance4x8_c; | 1349 vpx_sub_pixel_avg_variance4x8 = vpx_sub_pixel_avg_variance4x8_c; |
| 1350 if (flags & HAS_SSE) vpx_sub_pixel_avg_variance4x8 = vpx_sub_pixel_avg_varia
nce4x8_sse; | 1350 if (flags & HAS_SSE2) vpx_sub_pixel_avg_variance4x8 = vpx_sub_pixel_avg_vari
ance4x8_sse2; |
| 1351 if (flags & HAS_SSSE3) vpx_sub_pixel_avg_variance4x8 = vpx_sub_pixel_avg_var
iance4x8_ssse3; | 1351 if (flags & HAS_SSSE3) vpx_sub_pixel_avg_variance4x8 = vpx_sub_pixel_avg_var
iance4x8_ssse3; |
| 1352 vpx_sub_pixel_avg_variance64x32 = vpx_sub_pixel_avg_variance64x32_c; | 1352 vpx_sub_pixel_avg_variance64x32 = vpx_sub_pixel_avg_variance64x32_c; |
| 1353 if (flags & HAS_SSE2) vpx_sub_pixel_avg_variance64x32 = vpx_sub_pixel_avg_va
riance64x32_sse2; | 1353 if (flags & HAS_SSE2) vpx_sub_pixel_avg_variance64x32 = vpx_sub_pixel_avg_va
riance64x32_sse2; |
| 1354 if (flags & HAS_SSSE3) vpx_sub_pixel_avg_variance64x32 = vpx_sub_pixel_avg_v
ariance64x32_ssse3; | 1354 if (flags & HAS_SSSE3) vpx_sub_pixel_avg_variance64x32 = vpx_sub_pixel_avg_v
ariance64x32_ssse3; |
| 1355 vpx_sub_pixel_avg_variance64x64 = vpx_sub_pixel_avg_variance64x64_c; | 1355 vpx_sub_pixel_avg_variance64x64 = vpx_sub_pixel_avg_variance64x64_c; |
| 1356 if (flags & HAS_SSE2) vpx_sub_pixel_avg_variance64x64 = vpx_sub_pixel_avg_va
riance64x64_sse2; | 1356 if (flags & HAS_SSE2) vpx_sub_pixel_avg_variance64x64 = vpx_sub_pixel_avg_va
riance64x64_sse2; |
| 1357 if (flags & HAS_SSSE3) vpx_sub_pixel_avg_variance64x64 = vpx_sub_pixel_avg_v
ariance64x64_ssse3; | 1357 if (flags & HAS_SSSE3) vpx_sub_pixel_avg_variance64x64 = vpx_sub_pixel_avg_v
ariance64x64_ssse3; |
| 1358 if (flags & HAS_AVX2) vpx_sub_pixel_avg_variance64x64 = vpx_sub_pixel_avg_va
riance64x64_avx2; | 1358 if (flags & HAS_AVX2) vpx_sub_pixel_avg_variance64x64 = vpx_sub_pixel_avg_va
riance64x64_avx2; |
| 1359 vpx_sub_pixel_avg_variance8x16 = vpx_sub_pixel_avg_variance8x16_c; | 1359 vpx_sub_pixel_avg_variance8x16 = vpx_sub_pixel_avg_variance8x16_c; |
| 1360 if (flags & HAS_SSE2) vpx_sub_pixel_avg_variance8x16 = vpx_sub_pixel_avg_var
iance8x16_sse2; | 1360 if (flags & HAS_SSE2) vpx_sub_pixel_avg_variance8x16 = vpx_sub_pixel_avg_var
iance8x16_sse2; |
| (...skipping 20 matching lines...) Expand all Loading... |
| 1381 if (flags & HAS_SSSE3) vpx_sub_pixel_variance32x16 = vpx_sub_pixel_variance3
2x16_ssse3; | 1381 if (flags & HAS_SSSE3) vpx_sub_pixel_variance32x16 = vpx_sub_pixel_variance3
2x16_ssse3; |
| 1382 vpx_sub_pixel_variance32x32 = vpx_sub_pixel_variance32x32_c; | 1382 vpx_sub_pixel_variance32x32 = vpx_sub_pixel_variance32x32_c; |
| 1383 if (flags & HAS_SSE2) vpx_sub_pixel_variance32x32 = vpx_sub_pixel_variance32
x32_sse2; | 1383 if (flags & HAS_SSE2) vpx_sub_pixel_variance32x32 = vpx_sub_pixel_variance32
x32_sse2; |
| 1384 if (flags & HAS_SSSE3) vpx_sub_pixel_variance32x32 = vpx_sub_pixel_variance3
2x32_ssse3; | 1384 if (flags & HAS_SSSE3) vpx_sub_pixel_variance32x32 = vpx_sub_pixel_variance3
2x32_ssse3; |
| 1385 if (flags & HAS_AVX2) vpx_sub_pixel_variance32x32 = vpx_sub_pixel_variance32
x32_avx2; | 1385 if (flags & HAS_AVX2) vpx_sub_pixel_variance32x32 = vpx_sub_pixel_variance32
x32_avx2; |
| 1386 vpx_sub_pixel_variance32x64 = vpx_sub_pixel_variance32x64_c; | 1386 vpx_sub_pixel_variance32x64 = vpx_sub_pixel_variance32x64_c; |
| 1387 if (flags & HAS_SSE2) vpx_sub_pixel_variance32x64 = vpx_sub_pixel_variance32
x64_sse2; | 1387 if (flags & HAS_SSE2) vpx_sub_pixel_variance32x64 = vpx_sub_pixel_variance32
x64_sse2; |
| 1388 if (flags & HAS_SSSE3) vpx_sub_pixel_variance32x64 = vpx_sub_pixel_variance3
2x64_ssse3; | 1388 if (flags & HAS_SSSE3) vpx_sub_pixel_variance32x64 = vpx_sub_pixel_variance3
2x64_ssse3; |
| 1389 vpx_sub_pixel_variance4x4 = vpx_sub_pixel_variance4x4_c; | 1389 vpx_sub_pixel_variance4x4 = vpx_sub_pixel_variance4x4_c; |
| 1390 if (flags & HAS_MMX) vpx_sub_pixel_variance4x4 = vpx_sub_pixel_variance4x4_m
mx; | 1390 if (flags & HAS_MMX) vpx_sub_pixel_variance4x4 = vpx_sub_pixel_variance4x4_m
mx; |
| 1391 if (flags & HAS_SSE) vpx_sub_pixel_variance4x4 = vpx_sub_pixel_variance4x4_s
se; | 1391 if (flags & HAS_SSE2) vpx_sub_pixel_variance4x4 = vpx_sub_pixel_variance4x4_
sse2; |
| 1392 if (flags & HAS_SSSE3) vpx_sub_pixel_variance4x4 = vpx_sub_pixel_variance4x4
_ssse3; | 1392 if (flags & HAS_SSSE3) vpx_sub_pixel_variance4x4 = vpx_sub_pixel_variance4x4
_ssse3; |
| 1393 vpx_sub_pixel_variance4x8 = vpx_sub_pixel_variance4x8_c; | 1393 vpx_sub_pixel_variance4x8 = vpx_sub_pixel_variance4x8_c; |
| 1394 if (flags & HAS_SSE) vpx_sub_pixel_variance4x8 = vpx_sub_pixel_variance4x8_s
se; | 1394 if (flags & HAS_SSE2) vpx_sub_pixel_variance4x8 = vpx_sub_pixel_variance4x8_
sse2; |
| 1395 if (flags & HAS_SSSE3) vpx_sub_pixel_variance4x8 = vpx_sub_pixel_variance4x8
_ssse3; | 1395 if (flags & HAS_SSSE3) vpx_sub_pixel_variance4x8 = vpx_sub_pixel_variance4x8
_ssse3; |
| 1396 vpx_sub_pixel_variance64x32 = vpx_sub_pixel_variance64x32_c; | 1396 vpx_sub_pixel_variance64x32 = vpx_sub_pixel_variance64x32_c; |
| 1397 if (flags & HAS_SSE2) vpx_sub_pixel_variance64x32 = vpx_sub_pixel_variance64
x32_sse2; | 1397 if (flags & HAS_SSE2) vpx_sub_pixel_variance64x32 = vpx_sub_pixel_variance64
x32_sse2; |
| 1398 if (flags & HAS_SSSE3) vpx_sub_pixel_variance64x32 = vpx_sub_pixel_variance6
4x32_ssse3; | 1398 if (flags & HAS_SSSE3) vpx_sub_pixel_variance64x32 = vpx_sub_pixel_variance6
4x32_ssse3; |
| 1399 vpx_sub_pixel_variance64x64 = vpx_sub_pixel_variance64x64_c; | 1399 vpx_sub_pixel_variance64x64 = vpx_sub_pixel_variance64x64_c; |
| 1400 if (flags & HAS_SSE2) vpx_sub_pixel_variance64x64 = vpx_sub_pixel_variance64
x64_sse2; | 1400 if (flags & HAS_SSE2) vpx_sub_pixel_variance64x64 = vpx_sub_pixel_variance64
x64_sse2; |
| 1401 if (flags & HAS_SSSE3) vpx_sub_pixel_variance64x64 = vpx_sub_pixel_variance6
4x64_ssse3; | 1401 if (flags & HAS_SSSE3) vpx_sub_pixel_variance64x64 = vpx_sub_pixel_variance6
4x64_ssse3; |
| 1402 if (flags & HAS_AVX2) vpx_sub_pixel_variance64x64 = vpx_sub_pixel_variance64
x64_avx2; | 1402 if (flags & HAS_AVX2) vpx_sub_pixel_variance64x64 = vpx_sub_pixel_variance64
x64_avx2; |
| 1403 vpx_sub_pixel_variance8x16 = vpx_sub_pixel_variance8x16_c; | 1403 vpx_sub_pixel_variance8x16 = vpx_sub_pixel_variance8x16_c; |
| 1404 if (flags & HAS_MMX) vpx_sub_pixel_variance8x16 = vpx_sub_pixel_variance8x16
_mmx; | 1404 if (flags & HAS_MMX) vpx_sub_pixel_variance8x16 = vpx_sub_pixel_variance8x16
_mmx; |
| (...skipping 67 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 1472 vpx_vector_var = vpx_vector_var_c; | 1472 vpx_vector_var = vpx_vector_var_c; |
| 1473 if (flags & HAS_SSE2) vpx_vector_var = vpx_vector_var_sse2; | 1473 if (flags & HAS_SSE2) vpx_vector_var = vpx_vector_var_sse2; |
| 1474 } | 1474 } |
| 1475 #endif | 1475 #endif |
| 1476 | 1476 |
| 1477 #ifdef __cplusplus | 1477 #ifdef __cplusplus |
| 1478 } // extern "C" | 1478 } // extern "C" |
| 1479 #endif | 1479 #endif |
| 1480 | 1480 |
| 1481 #endif | 1481 #endif |
| OLD | NEW |