Index: source/config/mac/ia32/vp8_rtcd.h |
=================================================================== |
--- source/config/mac/ia32/vp8_rtcd.h (revision 261575) |
+++ source/config/mac/ia32/vp8_rtcd.h (working copy) |
@@ -492,337 +492,239 @@ |
if (flags & HAS_MMX) vp8_bilinear_predict16x16 = vp8_bilinear_predict16x16_mmx; |
if (flags & HAS_SSE2) vp8_bilinear_predict16x16 = vp8_bilinear_predict16x16_sse2; |
if (flags & HAS_SSSE3) vp8_bilinear_predict16x16 = vp8_bilinear_predict16x16_ssse3; |
- |
vp8_bilinear_predict4x4 = vp8_bilinear_predict4x4_c; |
if (flags & HAS_MMX) vp8_bilinear_predict4x4 = vp8_bilinear_predict4x4_mmx; |
- |
vp8_bilinear_predict8x4 = vp8_bilinear_predict8x4_c; |
if (flags & HAS_MMX) vp8_bilinear_predict8x4 = vp8_bilinear_predict8x4_mmx; |
- |
vp8_bilinear_predict8x8 = vp8_bilinear_predict8x8_c; |
if (flags & HAS_MMX) vp8_bilinear_predict8x8 = vp8_bilinear_predict8x8_mmx; |
if (flags & HAS_SSE2) vp8_bilinear_predict8x8 = vp8_bilinear_predict8x8_sse2; |
if (flags & HAS_SSSE3) vp8_bilinear_predict8x8 = vp8_bilinear_predict8x8_ssse3; |
- |
- |
- |
- |
vp8_block_error = vp8_block_error_c; |
if (flags & HAS_MMX) vp8_block_error = vp8_block_error_mmx; |
if (flags & HAS_SSE2) vp8_block_error = vp8_block_error_xmm; |
- |
vp8_build_intra_predictors_mbuv_s = vp8_build_intra_predictors_mbuv_s_c; |
if (flags & HAS_SSE2) vp8_build_intra_predictors_mbuv_s = vp8_build_intra_predictors_mbuv_s_sse2; |
if (flags & HAS_SSSE3) vp8_build_intra_predictors_mbuv_s = vp8_build_intra_predictors_mbuv_s_ssse3; |
- |
vp8_build_intra_predictors_mby_s = vp8_build_intra_predictors_mby_s_c; |
if (flags & HAS_SSE2) vp8_build_intra_predictors_mby_s = vp8_build_intra_predictors_mby_s_sse2; |
if (flags & HAS_SSSE3) vp8_build_intra_predictors_mby_s = vp8_build_intra_predictors_mby_s_ssse3; |
- |
vp8_clear_system_state = vp8_clear_system_state_c; |
if (flags & HAS_MMX) vp8_clear_system_state = vpx_reset_mmx_state; |
- |
vp8_copy32xn = vp8_copy32xn_c; |
if (flags & HAS_SSE2) vp8_copy32xn = vp8_copy32xn_sse2; |
if (flags & HAS_SSE3) vp8_copy32xn = vp8_copy32xn_sse3; |
- |
vp8_copy_mem16x16 = vp8_copy_mem16x16_c; |
if (flags & HAS_MMX) vp8_copy_mem16x16 = vp8_copy_mem16x16_mmx; |
if (flags & HAS_SSE2) vp8_copy_mem16x16 = vp8_copy_mem16x16_sse2; |
- |
vp8_copy_mem8x4 = vp8_copy_mem8x4_c; |
if (flags & HAS_MMX) vp8_copy_mem8x4 = vp8_copy_mem8x4_mmx; |
- |
vp8_copy_mem8x8 = vp8_copy_mem8x8_c; |
if (flags & HAS_MMX) vp8_copy_mem8x8 = vp8_copy_mem8x8_mmx; |
- |
vp8_dc_only_idct_add = vp8_dc_only_idct_add_c; |
if (flags & HAS_MMX) vp8_dc_only_idct_add = vp8_dc_only_idct_add_mmx; |
- |
vp8_denoiser_filter = vp8_denoiser_filter_c; |
if (flags & HAS_SSE2) vp8_denoiser_filter = vp8_denoiser_filter_sse2; |
- |
vp8_dequant_idct_add = vp8_dequant_idct_add_c; |
if (flags & HAS_MMX) vp8_dequant_idct_add = vp8_dequant_idct_add_mmx; |
- |
vp8_dequant_idct_add_uv_block = vp8_dequant_idct_add_uv_block_c; |
if (flags & HAS_MMX) vp8_dequant_idct_add_uv_block = vp8_dequant_idct_add_uv_block_mmx; |
if (flags & HAS_SSE2) vp8_dequant_idct_add_uv_block = vp8_dequant_idct_add_uv_block_sse2; |
- |
vp8_dequant_idct_add_y_block = vp8_dequant_idct_add_y_block_c; |
if (flags & HAS_MMX) vp8_dequant_idct_add_y_block = vp8_dequant_idct_add_y_block_mmx; |
if (flags & HAS_SSE2) vp8_dequant_idct_add_y_block = vp8_dequant_idct_add_y_block_sse2; |
- |
vp8_dequantize_b = vp8_dequantize_b_c; |
if (flags & HAS_MMX) vp8_dequantize_b = vp8_dequantize_b_mmx; |
- |
vp8_diamond_search_sad = vp8_diamond_search_sad_c; |
if (flags & HAS_SSE3) vp8_diamond_search_sad = vp8_diamond_search_sadx4; |
- |
vp8_fast_quantize_b = vp8_fast_quantize_b_c; |
if (flags & HAS_SSE2) vp8_fast_quantize_b = vp8_fast_quantize_b_sse2; |
if (flags & HAS_SSSE3) vp8_fast_quantize_b = vp8_fast_quantize_b_ssse3; |
- |
- |
vp8_filter_by_weight16x16 = vp8_filter_by_weight16x16_c; |
if (flags & HAS_SSE2) vp8_filter_by_weight16x16 = vp8_filter_by_weight16x16_sse2; |
- |
- |
vp8_filter_by_weight8x8 = vp8_filter_by_weight8x8_c; |
if (flags & HAS_SSE2) vp8_filter_by_weight8x8 = vp8_filter_by_weight8x8_sse2; |
- |
vp8_full_search_sad = vp8_full_search_sad_c; |
if (flags & HAS_SSE3) vp8_full_search_sad = vp8_full_search_sadx3; |
if (flags & HAS_SSE4_1) vp8_full_search_sad = vp8_full_search_sadx8; |
- |
vp8_get4x4sse_cs = vp8_get4x4sse_cs_c; |
if (flags & HAS_MMX) vp8_get4x4sse_cs = vp8_get4x4sse_cs_mmx; |
- |
vp8_get_mb_ss = vp8_get_mb_ss_c; |
if (flags & HAS_MMX) vp8_get_mb_ss = vp8_get_mb_ss_mmx; |
if (flags & HAS_SSE2) vp8_get_mb_ss = vp8_get_mb_ss_sse2; |
- |
- |
vp8_loop_filter_bh = vp8_loop_filter_bh_c; |
if (flags & HAS_MMX) vp8_loop_filter_bh = vp8_loop_filter_bh_mmx; |
if (flags & HAS_SSE2) vp8_loop_filter_bh = vp8_loop_filter_bh_sse2; |
- |
vp8_loop_filter_bv = vp8_loop_filter_bv_c; |
if (flags & HAS_MMX) vp8_loop_filter_bv = vp8_loop_filter_bv_mmx; |
if (flags & HAS_SSE2) vp8_loop_filter_bv = vp8_loop_filter_bv_sse2; |
- |
vp8_loop_filter_mbh = vp8_loop_filter_mbh_c; |
if (flags & HAS_MMX) vp8_loop_filter_mbh = vp8_loop_filter_mbh_mmx; |
if (flags & HAS_SSE2) vp8_loop_filter_mbh = vp8_loop_filter_mbh_sse2; |
- |
vp8_loop_filter_mbv = vp8_loop_filter_mbv_c; |
if (flags & HAS_MMX) vp8_loop_filter_mbv = vp8_loop_filter_mbv_mmx; |
if (flags & HAS_SSE2) vp8_loop_filter_mbv = vp8_loop_filter_mbv_sse2; |
- |
vp8_loop_filter_simple_bh = vp8_loop_filter_bhs_c; |
if (flags & HAS_MMX) vp8_loop_filter_simple_bh = vp8_loop_filter_bhs_mmx; |
if (flags & HAS_SSE2) vp8_loop_filter_simple_bh = vp8_loop_filter_bhs_sse2; |
- |
vp8_loop_filter_simple_bv = vp8_loop_filter_bvs_c; |
if (flags & HAS_MMX) vp8_loop_filter_simple_bv = vp8_loop_filter_bvs_mmx; |
if (flags & HAS_SSE2) vp8_loop_filter_simple_bv = vp8_loop_filter_bvs_sse2; |
- |
vp8_loop_filter_simple_mbh = vp8_loop_filter_simple_horizontal_edge_c; |
if (flags & HAS_MMX) vp8_loop_filter_simple_mbh = vp8_loop_filter_simple_horizontal_edge_mmx; |
if (flags & HAS_SSE2) vp8_loop_filter_simple_mbh = vp8_loop_filter_simple_horizontal_edge_sse2; |
- |
vp8_loop_filter_simple_mbv = vp8_loop_filter_simple_vertical_edge_c; |
if (flags & HAS_MMX) vp8_loop_filter_simple_mbv = vp8_loop_filter_simple_vertical_edge_mmx; |
if (flags & HAS_SSE2) vp8_loop_filter_simple_mbv = vp8_loop_filter_simple_vertical_edge_sse2; |
- |
vp8_mbblock_error = vp8_mbblock_error_c; |
if (flags & HAS_MMX) vp8_mbblock_error = vp8_mbblock_error_mmx; |
if (flags & HAS_SSE2) vp8_mbblock_error = vp8_mbblock_error_xmm; |
- |
vp8_mbpost_proc_across_ip = vp8_mbpost_proc_across_ip_c; |
if (flags & HAS_SSE2) vp8_mbpost_proc_across_ip = vp8_mbpost_proc_across_ip_xmm; |
- |
vp8_mbpost_proc_down = vp8_mbpost_proc_down_c; |
if (flags & HAS_MMX) vp8_mbpost_proc_down = vp8_mbpost_proc_down_mmx; |
if (flags & HAS_SSE2) vp8_mbpost_proc_down = vp8_mbpost_proc_down_xmm; |
- |
vp8_mbuverror = vp8_mbuverror_c; |
if (flags & HAS_MMX) vp8_mbuverror = vp8_mbuverror_mmx; |
if (flags & HAS_SSE2) vp8_mbuverror = vp8_mbuverror_xmm; |
- |
vp8_mse16x16 = vp8_mse16x16_c; |
if (flags & HAS_MMX) vp8_mse16x16 = vp8_mse16x16_mmx; |
if (flags & HAS_SSE2) vp8_mse16x16 = vp8_mse16x16_wmt; |
- |
vp8_plane_add_noise = vp8_plane_add_noise_c; |
if (flags & HAS_MMX) vp8_plane_add_noise = vp8_plane_add_noise_mmx; |
if (flags & HAS_SSE2) vp8_plane_add_noise = vp8_plane_add_noise_wmt; |
- |
vp8_post_proc_down_and_across_mb_row = vp8_post_proc_down_and_across_mb_row_c; |
if (flags & HAS_SSE2) vp8_post_proc_down_and_across_mb_row = vp8_post_proc_down_and_across_mb_row_sse2; |
- |
- |
- |
- |
vp8_refining_search_sad = vp8_refining_search_sad_c; |
if (flags & HAS_SSE3) vp8_refining_search_sad = vp8_refining_search_sadx4; |
- |
vp8_regular_quantize_b = vp8_regular_quantize_b_c; |
if (flags & HAS_SSE2) vp8_regular_quantize_b = vp8_regular_quantize_b_sse2; |
- |
- |
vp8_sad16x16 = vp8_sad16x16_c; |
if (flags & HAS_MMX) vp8_sad16x16 = vp8_sad16x16_mmx; |
if (flags & HAS_SSE2) vp8_sad16x16 = vp8_sad16x16_wmt; |
if (flags & HAS_SSE3) vp8_sad16x16 = vp8_sad16x16_sse3; |
- |
vp8_sad16x16x3 = vp8_sad16x16x3_c; |
if (flags & HAS_SSE3) vp8_sad16x16x3 = vp8_sad16x16x3_sse3; |
if (flags & HAS_SSSE3) vp8_sad16x16x3 = vp8_sad16x16x3_ssse3; |
- |
vp8_sad16x16x4d = vp8_sad16x16x4d_c; |
if (flags & HAS_SSE3) vp8_sad16x16x4d = vp8_sad16x16x4d_sse3; |
- |
vp8_sad16x16x8 = vp8_sad16x16x8_c; |
if (flags & HAS_SSE4_1) vp8_sad16x16x8 = vp8_sad16x16x8_sse4; |
- |
vp8_sad16x8 = vp8_sad16x8_c; |
if (flags & HAS_MMX) vp8_sad16x8 = vp8_sad16x8_mmx; |
if (flags & HAS_SSE2) vp8_sad16x8 = vp8_sad16x8_wmt; |
- |
vp8_sad16x8x3 = vp8_sad16x8x3_c; |
if (flags & HAS_SSE3) vp8_sad16x8x3 = vp8_sad16x8x3_sse3; |
if (flags & HAS_SSSE3) vp8_sad16x8x3 = vp8_sad16x8x3_ssse3; |
- |
vp8_sad16x8x4d = vp8_sad16x8x4d_c; |
if (flags & HAS_SSE3) vp8_sad16x8x4d = vp8_sad16x8x4d_sse3; |
- |
vp8_sad16x8x8 = vp8_sad16x8x8_c; |
if (flags & HAS_SSE4_1) vp8_sad16x8x8 = vp8_sad16x8x8_sse4; |
- |
vp8_sad4x4 = vp8_sad4x4_c; |
if (flags & HAS_MMX) vp8_sad4x4 = vp8_sad4x4_mmx; |
if (flags & HAS_SSE2) vp8_sad4x4 = vp8_sad4x4_wmt; |
- |
vp8_sad4x4x3 = vp8_sad4x4x3_c; |
if (flags & HAS_SSE3) vp8_sad4x4x3 = vp8_sad4x4x3_sse3; |
- |
vp8_sad4x4x4d = vp8_sad4x4x4d_c; |
if (flags & HAS_SSE3) vp8_sad4x4x4d = vp8_sad4x4x4d_sse3; |
- |
vp8_sad4x4x8 = vp8_sad4x4x8_c; |
if (flags & HAS_SSE4_1) vp8_sad4x4x8 = vp8_sad4x4x8_sse4; |
- |
vp8_sad8x16 = vp8_sad8x16_c; |
if (flags & HAS_MMX) vp8_sad8x16 = vp8_sad8x16_mmx; |
if (flags & HAS_SSE2) vp8_sad8x16 = vp8_sad8x16_wmt; |
- |
vp8_sad8x16x3 = vp8_sad8x16x3_c; |
if (flags & HAS_SSE3) vp8_sad8x16x3 = vp8_sad8x16x3_sse3; |
- |
vp8_sad8x16x4d = vp8_sad8x16x4d_c; |
if (flags & HAS_SSE3) vp8_sad8x16x4d = vp8_sad8x16x4d_sse3; |
- |
vp8_sad8x16x8 = vp8_sad8x16x8_c; |
if (flags & HAS_SSE4_1) vp8_sad8x16x8 = vp8_sad8x16x8_sse4; |
- |
vp8_sad8x8 = vp8_sad8x8_c; |
if (flags & HAS_MMX) vp8_sad8x8 = vp8_sad8x8_mmx; |
if (flags & HAS_SSE2) vp8_sad8x8 = vp8_sad8x8_wmt; |
- |
vp8_sad8x8x3 = vp8_sad8x8x3_c; |
if (flags & HAS_SSE3) vp8_sad8x8x3 = vp8_sad8x8x3_sse3; |
- |
vp8_sad8x8x4d = vp8_sad8x8x4d_c; |
if (flags & HAS_SSE3) vp8_sad8x8x4d = vp8_sad8x8x4d_sse3; |
- |
vp8_sad8x8x8 = vp8_sad8x8x8_c; |
if (flags & HAS_SSE4_1) vp8_sad8x8x8 = vp8_sad8x8x8_sse4; |
- |
vp8_short_fdct4x4 = vp8_short_fdct4x4_c; |
if (flags & HAS_MMX) vp8_short_fdct4x4 = vp8_short_fdct4x4_mmx; |
if (flags & HAS_SSE2) vp8_short_fdct4x4 = vp8_short_fdct4x4_sse2; |
- |
vp8_short_fdct8x4 = vp8_short_fdct8x4_c; |
if (flags & HAS_MMX) vp8_short_fdct8x4 = vp8_short_fdct8x4_mmx; |
if (flags & HAS_SSE2) vp8_short_fdct8x4 = vp8_short_fdct8x4_sse2; |
- |
vp8_short_idct4x4llm = vp8_short_idct4x4llm_c; |
if (flags & HAS_MMX) vp8_short_idct4x4llm = vp8_short_idct4x4llm_mmx; |
- |
vp8_short_inv_walsh4x4 = vp8_short_inv_walsh4x4_c; |
if (flags & HAS_MMX) vp8_short_inv_walsh4x4 = vp8_short_inv_walsh4x4_mmx; |
if (flags & HAS_SSE2) vp8_short_inv_walsh4x4 = vp8_short_inv_walsh4x4_sse2; |
- |
- |
vp8_short_walsh4x4 = vp8_short_walsh4x4_c; |
if (flags & HAS_SSE2) vp8_short_walsh4x4 = vp8_short_walsh4x4_sse2; |
- |
vp8_sixtap_predict16x16 = vp8_sixtap_predict16x16_c; |
if (flags & HAS_MMX) vp8_sixtap_predict16x16 = vp8_sixtap_predict16x16_mmx; |
if (flags & HAS_SSE2) vp8_sixtap_predict16x16 = vp8_sixtap_predict16x16_sse2; |
if (flags & HAS_SSSE3) vp8_sixtap_predict16x16 = vp8_sixtap_predict16x16_ssse3; |
- |
vp8_sixtap_predict4x4 = vp8_sixtap_predict4x4_c; |
if (flags & HAS_MMX) vp8_sixtap_predict4x4 = vp8_sixtap_predict4x4_mmx; |
if (flags & HAS_SSSE3) vp8_sixtap_predict4x4 = vp8_sixtap_predict4x4_ssse3; |
- |
vp8_sixtap_predict8x4 = vp8_sixtap_predict8x4_c; |
if (flags & HAS_MMX) vp8_sixtap_predict8x4 = vp8_sixtap_predict8x4_mmx; |
if (flags & HAS_SSE2) vp8_sixtap_predict8x4 = vp8_sixtap_predict8x4_sse2; |
if (flags & HAS_SSSE3) vp8_sixtap_predict8x4 = vp8_sixtap_predict8x4_ssse3; |
- |
vp8_sixtap_predict8x8 = vp8_sixtap_predict8x8_c; |
if (flags & HAS_MMX) vp8_sixtap_predict8x8 = vp8_sixtap_predict8x8_mmx; |
if (flags & HAS_SSE2) vp8_sixtap_predict8x8 = vp8_sixtap_predict8x8_sse2; |
if (flags & HAS_SSSE3) vp8_sixtap_predict8x8 = vp8_sixtap_predict8x8_ssse3; |
- |
vp8_sub_pixel_mse16x16 = vp8_sub_pixel_mse16x16_c; |
if (flags & HAS_MMX) vp8_sub_pixel_mse16x16 = vp8_sub_pixel_mse16x16_mmx; |
if (flags & HAS_SSE2) vp8_sub_pixel_mse16x16 = vp8_sub_pixel_mse16x16_wmt; |
- |
vp8_sub_pixel_variance16x16 = vp8_sub_pixel_variance16x16_c; |
if (flags & HAS_MMX) vp8_sub_pixel_variance16x16 = vp8_sub_pixel_variance16x16_mmx; |
if (flags & HAS_SSE2) vp8_sub_pixel_variance16x16 = vp8_sub_pixel_variance16x16_wmt; |
if (flags & HAS_SSSE3) vp8_sub_pixel_variance16x16 = vp8_sub_pixel_variance16x16_ssse3; |
- |
vp8_sub_pixel_variance16x8 = vp8_sub_pixel_variance16x8_c; |
if (flags & HAS_MMX) vp8_sub_pixel_variance16x8 = vp8_sub_pixel_variance16x8_mmx; |
if (flags & HAS_SSE2) vp8_sub_pixel_variance16x8 = vp8_sub_pixel_variance16x8_wmt; |
if (flags & HAS_SSSE3) vp8_sub_pixel_variance16x8 = vp8_sub_pixel_variance16x8_ssse3; |
- |
vp8_sub_pixel_variance4x4 = vp8_sub_pixel_variance4x4_c; |
if (flags & HAS_MMX) vp8_sub_pixel_variance4x4 = vp8_sub_pixel_variance4x4_mmx; |
if (flags & HAS_SSE2) vp8_sub_pixel_variance4x4 = vp8_sub_pixel_variance4x4_wmt; |
- |
vp8_sub_pixel_variance8x16 = vp8_sub_pixel_variance8x16_c; |
if (flags & HAS_MMX) vp8_sub_pixel_variance8x16 = vp8_sub_pixel_variance8x16_mmx; |
if (flags & HAS_SSE2) vp8_sub_pixel_variance8x16 = vp8_sub_pixel_variance8x16_wmt; |
- |
vp8_sub_pixel_variance8x8 = vp8_sub_pixel_variance8x8_c; |
if (flags & HAS_MMX) vp8_sub_pixel_variance8x8 = vp8_sub_pixel_variance8x8_mmx; |
if (flags & HAS_SSE2) vp8_sub_pixel_variance8x8 = vp8_sub_pixel_variance8x8_wmt; |
- |
vp8_subtract_b = vp8_subtract_b_c; |
if (flags & HAS_MMX) vp8_subtract_b = vp8_subtract_b_mmx; |
if (flags & HAS_SSE2) vp8_subtract_b = vp8_subtract_b_sse2; |
- |
vp8_subtract_mbuv = vp8_subtract_mbuv_c; |
if (flags & HAS_MMX) vp8_subtract_mbuv = vp8_subtract_mbuv_mmx; |
if (flags & HAS_SSE2) vp8_subtract_mbuv = vp8_subtract_mbuv_sse2; |
- |
vp8_subtract_mby = vp8_subtract_mby_c; |
if (flags & HAS_MMX) vp8_subtract_mby = vp8_subtract_mby_mmx; |
if (flags & HAS_SSE2) vp8_subtract_mby = vp8_subtract_mby_sse2; |
- |
vp8_variance16x16 = vp8_variance16x16_c; |
if (flags & HAS_MMX) vp8_variance16x16 = vp8_variance16x16_mmx; |
if (flags & HAS_SSE2) vp8_variance16x16 = vp8_variance16x16_wmt; |
- |
vp8_variance16x8 = vp8_variance16x8_c; |
if (flags & HAS_MMX) vp8_variance16x8 = vp8_variance16x8_mmx; |
if (flags & HAS_SSE2) vp8_variance16x8 = vp8_variance16x8_wmt; |
- |
vp8_variance4x4 = vp8_variance4x4_c; |
if (flags & HAS_MMX) vp8_variance4x4 = vp8_variance4x4_mmx; |
if (flags & HAS_SSE2) vp8_variance4x4 = vp8_variance4x4_wmt; |
- |
vp8_variance8x16 = vp8_variance8x16_c; |
if (flags & HAS_MMX) vp8_variance8x16 = vp8_variance8x16_mmx; |
if (flags & HAS_SSE2) vp8_variance8x16 = vp8_variance8x16_wmt; |
- |
vp8_variance8x8 = vp8_variance8x8_c; |
if (flags & HAS_MMX) vp8_variance8x8 = vp8_variance8x8_mmx; |
if (flags & HAS_SSE2) vp8_variance8x8 = vp8_variance8x8_wmt; |
- |
vp8_variance_halfpixvar16x16_h = vp8_variance_halfpixvar16x16_h_c; |
if (flags & HAS_MMX) vp8_variance_halfpixvar16x16_h = vp8_variance_halfpixvar16x16_h_mmx; |
if (flags & HAS_SSE2) vp8_variance_halfpixvar16x16_h = vp8_variance_halfpixvar16x16_h_wmt; |
- |
vp8_variance_halfpixvar16x16_hv = vp8_variance_halfpixvar16x16_hv_c; |
if (flags & HAS_MMX) vp8_variance_halfpixvar16x16_hv = vp8_variance_halfpixvar16x16_hv_mmx; |
if (flags & HAS_SSE2) vp8_variance_halfpixvar16x16_hv = vp8_variance_halfpixvar16x16_hv_wmt; |
- |
vp8_variance_halfpixvar16x16_v = vp8_variance_halfpixvar16x16_v_c; |
if (flags & HAS_MMX) vp8_variance_halfpixvar16x16_v = vp8_variance_halfpixvar16x16_v_mmx; |
if (flags & HAS_SSE2) vp8_variance_halfpixvar16x16_v = vp8_variance_halfpixvar16x16_v_wmt; |