Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(87)

Side by Side Diff: source/config/linux/ia32/vp8_rtcd.h

Issue 232133009: libvpx: Pull from upstream (Closed) Base URL: svn://svn.chromium.org/chrome/trunk/deps/third_party/libvpx/
Patch Set: Created 6 years, 8 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View unified diff | Download patch | Annotate | Revision Log
« no previous file with comments | « source/config/linux/generic/vpx_scale_rtcd.h ('k') | source/config/linux/ia32/vp9_rtcd.h » ('j') | no next file with comments »
Toggle Intra-line Diffs ('i') | Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
OLDNEW
1 #ifndef VP8_RTCD_H_ 1 #ifndef VP8_RTCD_H_
2 #define VP8_RTCD_H_ 2 #define VP8_RTCD_H_
3 3
4 #ifdef RTCD_C 4 #ifdef RTCD_C
5 #define RTCD_EXTERN 5 #define RTCD_EXTERN
6 #else 6 #else
7 #define RTCD_EXTERN extern 7 #define RTCD_EXTERN extern
8 #endif 8 #endif
9 9
10 #ifdef __cplusplus 10 #ifdef __cplusplus
(...skipping 474 matching lines...) Expand 10 before | Expand all | Expand 10 after
485 static void setup_rtcd_internal(void) 485 static void setup_rtcd_internal(void)
486 { 486 {
487 int flags = x86_simd_caps(); 487 int flags = x86_simd_caps();
488 488
489 (void)flags; 489 (void)flags;
490 490
491 vp8_bilinear_predict16x16 = vp8_bilinear_predict16x16_c; 491 vp8_bilinear_predict16x16 = vp8_bilinear_predict16x16_c;
492 if (flags & HAS_MMX) vp8_bilinear_predict16x16 = vp8_bilinear_predict16x16_m mx; 492 if (flags & HAS_MMX) vp8_bilinear_predict16x16 = vp8_bilinear_predict16x16_m mx;
493 if (flags & HAS_SSE2) vp8_bilinear_predict16x16 = vp8_bilinear_predict16x16_ sse2; 493 if (flags & HAS_SSE2) vp8_bilinear_predict16x16 = vp8_bilinear_predict16x16_ sse2;
494 if (flags & HAS_SSSE3) vp8_bilinear_predict16x16 = vp8_bilinear_predict16x16 _ssse3; 494 if (flags & HAS_SSSE3) vp8_bilinear_predict16x16 = vp8_bilinear_predict16x16 _ssse3;
495
496 vp8_bilinear_predict4x4 = vp8_bilinear_predict4x4_c; 495 vp8_bilinear_predict4x4 = vp8_bilinear_predict4x4_c;
497 if (flags & HAS_MMX) vp8_bilinear_predict4x4 = vp8_bilinear_predict4x4_mmx; 496 if (flags & HAS_MMX) vp8_bilinear_predict4x4 = vp8_bilinear_predict4x4_mmx;
498
499 vp8_bilinear_predict8x4 = vp8_bilinear_predict8x4_c; 497 vp8_bilinear_predict8x4 = vp8_bilinear_predict8x4_c;
500 if (flags & HAS_MMX) vp8_bilinear_predict8x4 = vp8_bilinear_predict8x4_mmx; 498 if (flags & HAS_MMX) vp8_bilinear_predict8x4 = vp8_bilinear_predict8x4_mmx;
501
502 vp8_bilinear_predict8x8 = vp8_bilinear_predict8x8_c; 499 vp8_bilinear_predict8x8 = vp8_bilinear_predict8x8_c;
503 if (flags & HAS_MMX) vp8_bilinear_predict8x8 = vp8_bilinear_predict8x8_mmx; 500 if (flags & HAS_MMX) vp8_bilinear_predict8x8 = vp8_bilinear_predict8x8_mmx;
504 if (flags & HAS_SSE2) vp8_bilinear_predict8x8 = vp8_bilinear_predict8x8_sse2 ; 501 if (flags & HAS_SSE2) vp8_bilinear_predict8x8 = vp8_bilinear_predict8x8_sse2 ;
505 if (flags & HAS_SSSE3) vp8_bilinear_predict8x8 = vp8_bilinear_predict8x8_sss e3; 502 if (flags & HAS_SSSE3) vp8_bilinear_predict8x8 = vp8_bilinear_predict8x8_sss e3;
506
507
508
509
510 vp8_block_error = vp8_block_error_c; 503 vp8_block_error = vp8_block_error_c;
511 if (flags & HAS_MMX) vp8_block_error = vp8_block_error_mmx; 504 if (flags & HAS_MMX) vp8_block_error = vp8_block_error_mmx;
512 if (flags & HAS_SSE2) vp8_block_error = vp8_block_error_xmm; 505 if (flags & HAS_SSE2) vp8_block_error = vp8_block_error_xmm;
513
514 vp8_build_intra_predictors_mbuv_s = vp8_build_intra_predictors_mbuv_s_c; 506 vp8_build_intra_predictors_mbuv_s = vp8_build_intra_predictors_mbuv_s_c;
515 if (flags & HAS_SSE2) vp8_build_intra_predictors_mbuv_s = vp8_build_intra_pr edictors_mbuv_s_sse2; 507 if (flags & HAS_SSE2) vp8_build_intra_predictors_mbuv_s = vp8_build_intra_pr edictors_mbuv_s_sse2;
516 if (flags & HAS_SSSE3) vp8_build_intra_predictors_mbuv_s = vp8_build_intra_p redictors_mbuv_s_ssse3; 508 if (flags & HAS_SSSE3) vp8_build_intra_predictors_mbuv_s = vp8_build_intra_p redictors_mbuv_s_ssse3;
517
518 vp8_build_intra_predictors_mby_s = vp8_build_intra_predictors_mby_s_c; 509 vp8_build_intra_predictors_mby_s = vp8_build_intra_predictors_mby_s_c;
519 if (flags & HAS_SSE2) vp8_build_intra_predictors_mby_s = vp8_build_intra_pre dictors_mby_s_sse2; 510 if (flags & HAS_SSE2) vp8_build_intra_predictors_mby_s = vp8_build_intra_pre dictors_mby_s_sse2;
520 if (flags & HAS_SSSE3) vp8_build_intra_predictors_mby_s = vp8_build_intra_pr edictors_mby_s_ssse3; 511 if (flags & HAS_SSSE3) vp8_build_intra_predictors_mby_s = vp8_build_intra_pr edictors_mby_s_ssse3;
521
522 vp8_clear_system_state = vp8_clear_system_state_c; 512 vp8_clear_system_state = vp8_clear_system_state_c;
523 if (flags & HAS_MMX) vp8_clear_system_state = vpx_reset_mmx_state; 513 if (flags & HAS_MMX) vp8_clear_system_state = vpx_reset_mmx_state;
524
525 vp8_copy32xn = vp8_copy32xn_c; 514 vp8_copy32xn = vp8_copy32xn_c;
526 if (flags & HAS_SSE2) vp8_copy32xn = vp8_copy32xn_sse2; 515 if (flags & HAS_SSE2) vp8_copy32xn = vp8_copy32xn_sse2;
527 if (flags & HAS_SSE3) vp8_copy32xn = vp8_copy32xn_sse3; 516 if (flags & HAS_SSE3) vp8_copy32xn = vp8_copy32xn_sse3;
528
529 vp8_copy_mem16x16 = vp8_copy_mem16x16_c; 517 vp8_copy_mem16x16 = vp8_copy_mem16x16_c;
530 if (flags & HAS_MMX) vp8_copy_mem16x16 = vp8_copy_mem16x16_mmx; 518 if (flags & HAS_MMX) vp8_copy_mem16x16 = vp8_copy_mem16x16_mmx;
531 if (flags & HAS_SSE2) vp8_copy_mem16x16 = vp8_copy_mem16x16_sse2; 519 if (flags & HAS_SSE2) vp8_copy_mem16x16 = vp8_copy_mem16x16_sse2;
532
533 vp8_copy_mem8x4 = vp8_copy_mem8x4_c; 520 vp8_copy_mem8x4 = vp8_copy_mem8x4_c;
534 if (flags & HAS_MMX) vp8_copy_mem8x4 = vp8_copy_mem8x4_mmx; 521 if (flags & HAS_MMX) vp8_copy_mem8x4 = vp8_copy_mem8x4_mmx;
535
536 vp8_copy_mem8x8 = vp8_copy_mem8x8_c; 522 vp8_copy_mem8x8 = vp8_copy_mem8x8_c;
537 if (flags & HAS_MMX) vp8_copy_mem8x8 = vp8_copy_mem8x8_mmx; 523 if (flags & HAS_MMX) vp8_copy_mem8x8 = vp8_copy_mem8x8_mmx;
538
539 vp8_dc_only_idct_add = vp8_dc_only_idct_add_c; 524 vp8_dc_only_idct_add = vp8_dc_only_idct_add_c;
540 if (flags & HAS_MMX) vp8_dc_only_idct_add = vp8_dc_only_idct_add_mmx; 525 if (flags & HAS_MMX) vp8_dc_only_idct_add = vp8_dc_only_idct_add_mmx;
541
542 vp8_denoiser_filter = vp8_denoiser_filter_c; 526 vp8_denoiser_filter = vp8_denoiser_filter_c;
543 if (flags & HAS_SSE2) vp8_denoiser_filter = vp8_denoiser_filter_sse2; 527 if (flags & HAS_SSE2) vp8_denoiser_filter = vp8_denoiser_filter_sse2;
544
545 vp8_dequant_idct_add = vp8_dequant_idct_add_c; 528 vp8_dequant_idct_add = vp8_dequant_idct_add_c;
546 if (flags & HAS_MMX) vp8_dequant_idct_add = vp8_dequant_idct_add_mmx; 529 if (flags & HAS_MMX) vp8_dequant_idct_add = vp8_dequant_idct_add_mmx;
547
548 vp8_dequant_idct_add_uv_block = vp8_dequant_idct_add_uv_block_c; 530 vp8_dequant_idct_add_uv_block = vp8_dequant_idct_add_uv_block_c;
549 if (flags & HAS_MMX) vp8_dequant_idct_add_uv_block = vp8_dequant_idct_add_uv _block_mmx; 531 if (flags & HAS_MMX) vp8_dequant_idct_add_uv_block = vp8_dequant_idct_add_uv _block_mmx;
550 if (flags & HAS_SSE2) vp8_dequant_idct_add_uv_block = vp8_dequant_idct_add_u v_block_sse2; 532 if (flags & HAS_SSE2) vp8_dequant_idct_add_uv_block = vp8_dequant_idct_add_u v_block_sse2;
551
552 vp8_dequant_idct_add_y_block = vp8_dequant_idct_add_y_block_c; 533 vp8_dequant_idct_add_y_block = vp8_dequant_idct_add_y_block_c;
553 if (flags & HAS_MMX) vp8_dequant_idct_add_y_block = vp8_dequant_idct_add_y_b lock_mmx; 534 if (flags & HAS_MMX) vp8_dequant_idct_add_y_block = vp8_dequant_idct_add_y_b lock_mmx;
554 if (flags & HAS_SSE2) vp8_dequant_idct_add_y_block = vp8_dequant_idct_add_y_ block_sse2; 535 if (flags & HAS_SSE2) vp8_dequant_idct_add_y_block = vp8_dequant_idct_add_y_ block_sse2;
555
556 vp8_dequantize_b = vp8_dequantize_b_c; 536 vp8_dequantize_b = vp8_dequantize_b_c;
557 if (flags & HAS_MMX) vp8_dequantize_b = vp8_dequantize_b_mmx; 537 if (flags & HAS_MMX) vp8_dequantize_b = vp8_dequantize_b_mmx;
558
559 vp8_diamond_search_sad = vp8_diamond_search_sad_c; 538 vp8_diamond_search_sad = vp8_diamond_search_sad_c;
560 if (flags & HAS_SSE3) vp8_diamond_search_sad = vp8_diamond_search_sadx4; 539 if (flags & HAS_SSE3) vp8_diamond_search_sad = vp8_diamond_search_sadx4;
561
562 vp8_fast_quantize_b = vp8_fast_quantize_b_c; 540 vp8_fast_quantize_b = vp8_fast_quantize_b_c;
563 if (flags & HAS_SSE2) vp8_fast_quantize_b = vp8_fast_quantize_b_sse2; 541 if (flags & HAS_SSE2) vp8_fast_quantize_b = vp8_fast_quantize_b_sse2;
564 if (flags & HAS_SSSE3) vp8_fast_quantize_b = vp8_fast_quantize_b_ssse3; 542 if (flags & HAS_SSSE3) vp8_fast_quantize_b = vp8_fast_quantize_b_ssse3;
565
566
567 vp8_filter_by_weight16x16 = vp8_filter_by_weight16x16_c; 543 vp8_filter_by_weight16x16 = vp8_filter_by_weight16x16_c;
568 if (flags & HAS_SSE2) vp8_filter_by_weight16x16 = vp8_filter_by_weight16x16_ sse2; 544 if (flags & HAS_SSE2) vp8_filter_by_weight16x16 = vp8_filter_by_weight16x16_ sse2;
569
570
571 vp8_filter_by_weight8x8 = vp8_filter_by_weight8x8_c; 545 vp8_filter_by_weight8x8 = vp8_filter_by_weight8x8_c;
572 if (flags & HAS_SSE2) vp8_filter_by_weight8x8 = vp8_filter_by_weight8x8_sse2 ; 546 if (flags & HAS_SSE2) vp8_filter_by_weight8x8 = vp8_filter_by_weight8x8_sse2 ;
573
574 vp8_full_search_sad = vp8_full_search_sad_c; 547 vp8_full_search_sad = vp8_full_search_sad_c;
575 if (flags & HAS_SSE3) vp8_full_search_sad = vp8_full_search_sadx3; 548 if (flags & HAS_SSE3) vp8_full_search_sad = vp8_full_search_sadx3;
576 if (flags & HAS_SSE4_1) vp8_full_search_sad = vp8_full_search_sadx8; 549 if (flags & HAS_SSE4_1) vp8_full_search_sad = vp8_full_search_sadx8;
577
578 vp8_get4x4sse_cs = vp8_get4x4sse_cs_c; 550 vp8_get4x4sse_cs = vp8_get4x4sse_cs_c;
579 if (flags & HAS_MMX) vp8_get4x4sse_cs = vp8_get4x4sse_cs_mmx; 551 if (flags & HAS_MMX) vp8_get4x4sse_cs = vp8_get4x4sse_cs_mmx;
580
581 vp8_get_mb_ss = vp8_get_mb_ss_c; 552 vp8_get_mb_ss = vp8_get_mb_ss_c;
582 if (flags & HAS_MMX) vp8_get_mb_ss = vp8_get_mb_ss_mmx; 553 if (flags & HAS_MMX) vp8_get_mb_ss = vp8_get_mb_ss_mmx;
583 if (flags & HAS_SSE2) vp8_get_mb_ss = vp8_get_mb_ss_sse2; 554 if (flags & HAS_SSE2) vp8_get_mb_ss = vp8_get_mb_ss_sse2;
584
585
586 vp8_loop_filter_bh = vp8_loop_filter_bh_c; 555 vp8_loop_filter_bh = vp8_loop_filter_bh_c;
587 if (flags & HAS_MMX) vp8_loop_filter_bh = vp8_loop_filter_bh_mmx; 556 if (flags & HAS_MMX) vp8_loop_filter_bh = vp8_loop_filter_bh_mmx;
588 if (flags & HAS_SSE2) vp8_loop_filter_bh = vp8_loop_filter_bh_sse2; 557 if (flags & HAS_SSE2) vp8_loop_filter_bh = vp8_loop_filter_bh_sse2;
589
590 vp8_loop_filter_bv = vp8_loop_filter_bv_c; 558 vp8_loop_filter_bv = vp8_loop_filter_bv_c;
591 if (flags & HAS_MMX) vp8_loop_filter_bv = vp8_loop_filter_bv_mmx; 559 if (flags & HAS_MMX) vp8_loop_filter_bv = vp8_loop_filter_bv_mmx;
592 if (flags & HAS_SSE2) vp8_loop_filter_bv = vp8_loop_filter_bv_sse2; 560 if (flags & HAS_SSE2) vp8_loop_filter_bv = vp8_loop_filter_bv_sse2;
593
594 vp8_loop_filter_mbh = vp8_loop_filter_mbh_c; 561 vp8_loop_filter_mbh = vp8_loop_filter_mbh_c;
595 if (flags & HAS_MMX) vp8_loop_filter_mbh = vp8_loop_filter_mbh_mmx; 562 if (flags & HAS_MMX) vp8_loop_filter_mbh = vp8_loop_filter_mbh_mmx;
596 if (flags & HAS_SSE2) vp8_loop_filter_mbh = vp8_loop_filter_mbh_sse2; 563 if (flags & HAS_SSE2) vp8_loop_filter_mbh = vp8_loop_filter_mbh_sse2;
597
598 vp8_loop_filter_mbv = vp8_loop_filter_mbv_c; 564 vp8_loop_filter_mbv = vp8_loop_filter_mbv_c;
599 if (flags & HAS_MMX) vp8_loop_filter_mbv = vp8_loop_filter_mbv_mmx; 565 if (flags & HAS_MMX) vp8_loop_filter_mbv = vp8_loop_filter_mbv_mmx;
600 if (flags & HAS_SSE2) vp8_loop_filter_mbv = vp8_loop_filter_mbv_sse2; 566 if (flags & HAS_SSE2) vp8_loop_filter_mbv = vp8_loop_filter_mbv_sse2;
601
602 vp8_loop_filter_simple_bh = vp8_loop_filter_bhs_c; 567 vp8_loop_filter_simple_bh = vp8_loop_filter_bhs_c;
603 if (flags & HAS_MMX) vp8_loop_filter_simple_bh = vp8_loop_filter_bhs_mmx; 568 if (flags & HAS_MMX) vp8_loop_filter_simple_bh = vp8_loop_filter_bhs_mmx;
604 if (flags & HAS_SSE2) vp8_loop_filter_simple_bh = vp8_loop_filter_bhs_sse2; 569 if (flags & HAS_SSE2) vp8_loop_filter_simple_bh = vp8_loop_filter_bhs_sse2;
605
606 vp8_loop_filter_simple_bv = vp8_loop_filter_bvs_c; 570 vp8_loop_filter_simple_bv = vp8_loop_filter_bvs_c;
607 if (flags & HAS_MMX) vp8_loop_filter_simple_bv = vp8_loop_filter_bvs_mmx; 571 if (flags & HAS_MMX) vp8_loop_filter_simple_bv = vp8_loop_filter_bvs_mmx;
608 if (flags & HAS_SSE2) vp8_loop_filter_simple_bv = vp8_loop_filter_bvs_sse2; 572 if (flags & HAS_SSE2) vp8_loop_filter_simple_bv = vp8_loop_filter_bvs_sse2;
609
610 vp8_loop_filter_simple_mbh = vp8_loop_filter_simple_horizontal_edge_c; 573 vp8_loop_filter_simple_mbh = vp8_loop_filter_simple_horizontal_edge_c;
611 if (flags & HAS_MMX) vp8_loop_filter_simple_mbh = vp8_loop_filter_simple_hor izontal_edge_mmx; 574 if (flags & HAS_MMX) vp8_loop_filter_simple_mbh = vp8_loop_filter_simple_hor izontal_edge_mmx;
612 if (flags & HAS_SSE2) vp8_loop_filter_simple_mbh = vp8_loop_filter_simple_ho rizontal_edge_sse2; 575 if (flags & HAS_SSE2) vp8_loop_filter_simple_mbh = vp8_loop_filter_simple_ho rizontal_edge_sse2;
613
614 vp8_loop_filter_simple_mbv = vp8_loop_filter_simple_vertical_edge_c; 576 vp8_loop_filter_simple_mbv = vp8_loop_filter_simple_vertical_edge_c;
615 if (flags & HAS_MMX) vp8_loop_filter_simple_mbv = vp8_loop_filter_simple_ver tical_edge_mmx; 577 if (flags & HAS_MMX) vp8_loop_filter_simple_mbv = vp8_loop_filter_simple_ver tical_edge_mmx;
616 if (flags & HAS_SSE2) vp8_loop_filter_simple_mbv = vp8_loop_filter_simple_ve rtical_edge_sse2; 578 if (flags & HAS_SSE2) vp8_loop_filter_simple_mbv = vp8_loop_filter_simple_ve rtical_edge_sse2;
617
618 vp8_mbblock_error = vp8_mbblock_error_c; 579 vp8_mbblock_error = vp8_mbblock_error_c;
619 if (flags & HAS_MMX) vp8_mbblock_error = vp8_mbblock_error_mmx; 580 if (flags & HAS_MMX) vp8_mbblock_error = vp8_mbblock_error_mmx;
620 if (flags & HAS_SSE2) vp8_mbblock_error = vp8_mbblock_error_xmm; 581 if (flags & HAS_SSE2) vp8_mbblock_error = vp8_mbblock_error_xmm;
621
622 vp8_mbpost_proc_across_ip = vp8_mbpost_proc_across_ip_c; 582 vp8_mbpost_proc_across_ip = vp8_mbpost_proc_across_ip_c;
623 if (flags & HAS_SSE2) vp8_mbpost_proc_across_ip = vp8_mbpost_proc_across_ip_ xmm; 583 if (flags & HAS_SSE2) vp8_mbpost_proc_across_ip = vp8_mbpost_proc_across_ip_ xmm;
624
625 vp8_mbpost_proc_down = vp8_mbpost_proc_down_c; 584 vp8_mbpost_proc_down = vp8_mbpost_proc_down_c;
626 if (flags & HAS_MMX) vp8_mbpost_proc_down = vp8_mbpost_proc_down_mmx; 585 if (flags & HAS_MMX) vp8_mbpost_proc_down = vp8_mbpost_proc_down_mmx;
627 if (flags & HAS_SSE2) vp8_mbpost_proc_down = vp8_mbpost_proc_down_xmm; 586 if (flags & HAS_SSE2) vp8_mbpost_proc_down = vp8_mbpost_proc_down_xmm;
628
629 vp8_mbuverror = vp8_mbuverror_c; 587 vp8_mbuverror = vp8_mbuverror_c;
630 if (flags & HAS_MMX) vp8_mbuverror = vp8_mbuverror_mmx; 588 if (flags & HAS_MMX) vp8_mbuverror = vp8_mbuverror_mmx;
631 if (flags & HAS_SSE2) vp8_mbuverror = vp8_mbuverror_xmm; 589 if (flags & HAS_SSE2) vp8_mbuverror = vp8_mbuverror_xmm;
632
633 vp8_mse16x16 = vp8_mse16x16_c; 590 vp8_mse16x16 = vp8_mse16x16_c;
634 if (flags & HAS_MMX) vp8_mse16x16 = vp8_mse16x16_mmx; 591 if (flags & HAS_MMX) vp8_mse16x16 = vp8_mse16x16_mmx;
635 if (flags & HAS_SSE2) vp8_mse16x16 = vp8_mse16x16_wmt; 592 if (flags & HAS_SSE2) vp8_mse16x16 = vp8_mse16x16_wmt;
636
637 vp8_plane_add_noise = vp8_plane_add_noise_c; 593 vp8_plane_add_noise = vp8_plane_add_noise_c;
638 if (flags & HAS_MMX) vp8_plane_add_noise = vp8_plane_add_noise_mmx; 594 if (flags & HAS_MMX) vp8_plane_add_noise = vp8_plane_add_noise_mmx;
639 if (flags & HAS_SSE2) vp8_plane_add_noise = vp8_plane_add_noise_wmt; 595 if (flags & HAS_SSE2) vp8_plane_add_noise = vp8_plane_add_noise_wmt;
640
641 vp8_post_proc_down_and_across_mb_row = vp8_post_proc_down_and_across_mb_row_ c; 596 vp8_post_proc_down_and_across_mb_row = vp8_post_proc_down_and_across_mb_row_ c;
642 if (flags & HAS_SSE2) vp8_post_proc_down_and_across_mb_row = vp8_post_proc_d own_and_across_mb_row_sse2; 597 if (flags & HAS_SSE2) vp8_post_proc_down_and_across_mb_row = vp8_post_proc_d own_and_across_mb_row_sse2;
643
644
645
646
647 vp8_refining_search_sad = vp8_refining_search_sad_c; 598 vp8_refining_search_sad = vp8_refining_search_sad_c;
648 if (flags & HAS_SSE3) vp8_refining_search_sad = vp8_refining_search_sadx4; 599 if (flags & HAS_SSE3) vp8_refining_search_sad = vp8_refining_search_sadx4;
649
650 vp8_regular_quantize_b = vp8_regular_quantize_b_c; 600 vp8_regular_quantize_b = vp8_regular_quantize_b_c;
651 if (flags & HAS_SSE2) vp8_regular_quantize_b = vp8_regular_quantize_b_sse2; 601 if (flags & HAS_SSE2) vp8_regular_quantize_b = vp8_regular_quantize_b_sse2;
652
653
654 vp8_sad16x16 = vp8_sad16x16_c; 602 vp8_sad16x16 = vp8_sad16x16_c;
655 if (flags & HAS_MMX) vp8_sad16x16 = vp8_sad16x16_mmx; 603 if (flags & HAS_MMX) vp8_sad16x16 = vp8_sad16x16_mmx;
656 if (flags & HAS_SSE2) vp8_sad16x16 = vp8_sad16x16_wmt; 604 if (flags & HAS_SSE2) vp8_sad16x16 = vp8_sad16x16_wmt;
657 if (flags & HAS_SSE3) vp8_sad16x16 = vp8_sad16x16_sse3; 605 if (flags & HAS_SSE3) vp8_sad16x16 = vp8_sad16x16_sse3;
658
659 vp8_sad16x16x3 = vp8_sad16x16x3_c; 606 vp8_sad16x16x3 = vp8_sad16x16x3_c;
660 if (flags & HAS_SSE3) vp8_sad16x16x3 = vp8_sad16x16x3_sse3; 607 if (flags & HAS_SSE3) vp8_sad16x16x3 = vp8_sad16x16x3_sse3;
661 if (flags & HAS_SSSE3) vp8_sad16x16x3 = vp8_sad16x16x3_ssse3; 608 if (flags & HAS_SSSE3) vp8_sad16x16x3 = vp8_sad16x16x3_ssse3;
662
663 vp8_sad16x16x4d = vp8_sad16x16x4d_c; 609 vp8_sad16x16x4d = vp8_sad16x16x4d_c;
664 if (flags & HAS_SSE3) vp8_sad16x16x4d = vp8_sad16x16x4d_sse3; 610 if (flags & HAS_SSE3) vp8_sad16x16x4d = vp8_sad16x16x4d_sse3;
665
666 vp8_sad16x16x8 = vp8_sad16x16x8_c; 611 vp8_sad16x16x8 = vp8_sad16x16x8_c;
667 if (flags & HAS_SSE4_1) vp8_sad16x16x8 = vp8_sad16x16x8_sse4; 612 if (flags & HAS_SSE4_1) vp8_sad16x16x8 = vp8_sad16x16x8_sse4;
668
669 vp8_sad16x8 = vp8_sad16x8_c; 613 vp8_sad16x8 = vp8_sad16x8_c;
670 if (flags & HAS_MMX) vp8_sad16x8 = vp8_sad16x8_mmx; 614 if (flags & HAS_MMX) vp8_sad16x8 = vp8_sad16x8_mmx;
671 if (flags & HAS_SSE2) vp8_sad16x8 = vp8_sad16x8_wmt; 615 if (flags & HAS_SSE2) vp8_sad16x8 = vp8_sad16x8_wmt;
672
673 vp8_sad16x8x3 = vp8_sad16x8x3_c; 616 vp8_sad16x8x3 = vp8_sad16x8x3_c;
674 if (flags & HAS_SSE3) vp8_sad16x8x3 = vp8_sad16x8x3_sse3; 617 if (flags & HAS_SSE3) vp8_sad16x8x3 = vp8_sad16x8x3_sse3;
675 if (flags & HAS_SSSE3) vp8_sad16x8x3 = vp8_sad16x8x3_ssse3; 618 if (flags & HAS_SSSE3) vp8_sad16x8x3 = vp8_sad16x8x3_ssse3;
676
677 vp8_sad16x8x4d = vp8_sad16x8x4d_c; 619 vp8_sad16x8x4d = vp8_sad16x8x4d_c;
678 if (flags & HAS_SSE3) vp8_sad16x8x4d = vp8_sad16x8x4d_sse3; 620 if (flags & HAS_SSE3) vp8_sad16x8x4d = vp8_sad16x8x4d_sse3;
679
680 vp8_sad16x8x8 = vp8_sad16x8x8_c; 621 vp8_sad16x8x8 = vp8_sad16x8x8_c;
681 if (flags & HAS_SSE4_1) vp8_sad16x8x8 = vp8_sad16x8x8_sse4; 622 if (flags & HAS_SSE4_1) vp8_sad16x8x8 = vp8_sad16x8x8_sse4;
682
683 vp8_sad4x4 = vp8_sad4x4_c; 623 vp8_sad4x4 = vp8_sad4x4_c;
684 if (flags & HAS_MMX) vp8_sad4x4 = vp8_sad4x4_mmx; 624 if (flags & HAS_MMX) vp8_sad4x4 = vp8_sad4x4_mmx;
685 if (flags & HAS_SSE2) vp8_sad4x4 = vp8_sad4x4_wmt; 625 if (flags & HAS_SSE2) vp8_sad4x4 = vp8_sad4x4_wmt;
686
687 vp8_sad4x4x3 = vp8_sad4x4x3_c; 626 vp8_sad4x4x3 = vp8_sad4x4x3_c;
688 if (flags & HAS_SSE3) vp8_sad4x4x3 = vp8_sad4x4x3_sse3; 627 if (flags & HAS_SSE3) vp8_sad4x4x3 = vp8_sad4x4x3_sse3;
689
690 vp8_sad4x4x4d = vp8_sad4x4x4d_c; 628 vp8_sad4x4x4d = vp8_sad4x4x4d_c;
691 if (flags & HAS_SSE3) vp8_sad4x4x4d = vp8_sad4x4x4d_sse3; 629 if (flags & HAS_SSE3) vp8_sad4x4x4d = vp8_sad4x4x4d_sse3;
692
693 vp8_sad4x4x8 = vp8_sad4x4x8_c; 630 vp8_sad4x4x8 = vp8_sad4x4x8_c;
694 if (flags & HAS_SSE4_1) vp8_sad4x4x8 = vp8_sad4x4x8_sse4; 631 if (flags & HAS_SSE4_1) vp8_sad4x4x8 = vp8_sad4x4x8_sse4;
695
696 vp8_sad8x16 = vp8_sad8x16_c; 632 vp8_sad8x16 = vp8_sad8x16_c;
697 if (flags & HAS_MMX) vp8_sad8x16 = vp8_sad8x16_mmx; 633 if (flags & HAS_MMX) vp8_sad8x16 = vp8_sad8x16_mmx;
698 if (flags & HAS_SSE2) vp8_sad8x16 = vp8_sad8x16_wmt; 634 if (flags & HAS_SSE2) vp8_sad8x16 = vp8_sad8x16_wmt;
699
700 vp8_sad8x16x3 = vp8_sad8x16x3_c; 635 vp8_sad8x16x3 = vp8_sad8x16x3_c;
701 if (flags & HAS_SSE3) vp8_sad8x16x3 = vp8_sad8x16x3_sse3; 636 if (flags & HAS_SSE3) vp8_sad8x16x3 = vp8_sad8x16x3_sse3;
702
703 vp8_sad8x16x4d = vp8_sad8x16x4d_c; 637 vp8_sad8x16x4d = vp8_sad8x16x4d_c;
704 if (flags & HAS_SSE3) vp8_sad8x16x4d = vp8_sad8x16x4d_sse3; 638 if (flags & HAS_SSE3) vp8_sad8x16x4d = vp8_sad8x16x4d_sse3;
705
706 vp8_sad8x16x8 = vp8_sad8x16x8_c; 639 vp8_sad8x16x8 = vp8_sad8x16x8_c;
707 if (flags & HAS_SSE4_1) vp8_sad8x16x8 = vp8_sad8x16x8_sse4; 640 if (flags & HAS_SSE4_1) vp8_sad8x16x8 = vp8_sad8x16x8_sse4;
708
709 vp8_sad8x8 = vp8_sad8x8_c; 641 vp8_sad8x8 = vp8_sad8x8_c;
710 if (flags & HAS_MMX) vp8_sad8x8 = vp8_sad8x8_mmx; 642 if (flags & HAS_MMX) vp8_sad8x8 = vp8_sad8x8_mmx;
711 if (flags & HAS_SSE2) vp8_sad8x8 = vp8_sad8x8_wmt; 643 if (flags & HAS_SSE2) vp8_sad8x8 = vp8_sad8x8_wmt;
712
713 vp8_sad8x8x3 = vp8_sad8x8x3_c; 644 vp8_sad8x8x3 = vp8_sad8x8x3_c;
714 if (flags & HAS_SSE3) vp8_sad8x8x3 = vp8_sad8x8x3_sse3; 645 if (flags & HAS_SSE3) vp8_sad8x8x3 = vp8_sad8x8x3_sse3;
715
716 vp8_sad8x8x4d = vp8_sad8x8x4d_c; 646 vp8_sad8x8x4d = vp8_sad8x8x4d_c;
717 if (flags & HAS_SSE3) vp8_sad8x8x4d = vp8_sad8x8x4d_sse3; 647 if (flags & HAS_SSE3) vp8_sad8x8x4d = vp8_sad8x8x4d_sse3;
718
719 vp8_sad8x8x8 = vp8_sad8x8x8_c; 648 vp8_sad8x8x8 = vp8_sad8x8x8_c;
720 if (flags & HAS_SSE4_1) vp8_sad8x8x8 = vp8_sad8x8x8_sse4; 649 if (flags & HAS_SSE4_1) vp8_sad8x8x8 = vp8_sad8x8x8_sse4;
721
722 vp8_short_fdct4x4 = vp8_short_fdct4x4_c; 650 vp8_short_fdct4x4 = vp8_short_fdct4x4_c;
723 if (flags & HAS_MMX) vp8_short_fdct4x4 = vp8_short_fdct4x4_mmx; 651 if (flags & HAS_MMX) vp8_short_fdct4x4 = vp8_short_fdct4x4_mmx;
724 if (flags & HAS_SSE2) vp8_short_fdct4x4 = vp8_short_fdct4x4_sse2; 652 if (flags & HAS_SSE2) vp8_short_fdct4x4 = vp8_short_fdct4x4_sse2;
725
726 vp8_short_fdct8x4 = vp8_short_fdct8x4_c; 653 vp8_short_fdct8x4 = vp8_short_fdct8x4_c;
727 if (flags & HAS_MMX) vp8_short_fdct8x4 = vp8_short_fdct8x4_mmx; 654 if (flags & HAS_MMX) vp8_short_fdct8x4 = vp8_short_fdct8x4_mmx;
728 if (flags & HAS_SSE2) vp8_short_fdct8x4 = vp8_short_fdct8x4_sse2; 655 if (flags & HAS_SSE2) vp8_short_fdct8x4 = vp8_short_fdct8x4_sse2;
729
730 vp8_short_idct4x4llm = vp8_short_idct4x4llm_c; 656 vp8_short_idct4x4llm = vp8_short_idct4x4llm_c;
731 if (flags & HAS_MMX) vp8_short_idct4x4llm = vp8_short_idct4x4llm_mmx; 657 if (flags & HAS_MMX) vp8_short_idct4x4llm = vp8_short_idct4x4llm_mmx;
732
733 vp8_short_inv_walsh4x4 = vp8_short_inv_walsh4x4_c; 658 vp8_short_inv_walsh4x4 = vp8_short_inv_walsh4x4_c;
734 if (flags & HAS_MMX) vp8_short_inv_walsh4x4 = vp8_short_inv_walsh4x4_mmx; 659 if (flags & HAS_MMX) vp8_short_inv_walsh4x4 = vp8_short_inv_walsh4x4_mmx;
735 if (flags & HAS_SSE2) vp8_short_inv_walsh4x4 = vp8_short_inv_walsh4x4_sse2; 660 if (flags & HAS_SSE2) vp8_short_inv_walsh4x4 = vp8_short_inv_walsh4x4_sse2;
736
737
738 vp8_short_walsh4x4 = vp8_short_walsh4x4_c; 661 vp8_short_walsh4x4 = vp8_short_walsh4x4_c;
739 if (flags & HAS_SSE2) vp8_short_walsh4x4 = vp8_short_walsh4x4_sse2; 662 if (flags & HAS_SSE2) vp8_short_walsh4x4 = vp8_short_walsh4x4_sse2;
740
741 vp8_sixtap_predict16x16 = vp8_sixtap_predict16x16_c; 663 vp8_sixtap_predict16x16 = vp8_sixtap_predict16x16_c;
742 if (flags & HAS_MMX) vp8_sixtap_predict16x16 = vp8_sixtap_predict16x16_mmx; 664 if (flags & HAS_MMX) vp8_sixtap_predict16x16 = vp8_sixtap_predict16x16_mmx;
743 if (flags & HAS_SSE2) vp8_sixtap_predict16x16 = vp8_sixtap_predict16x16_sse2 ; 665 if (flags & HAS_SSE2) vp8_sixtap_predict16x16 = vp8_sixtap_predict16x16_sse2 ;
744 if (flags & HAS_SSSE3) vp8_sixtap_predict16x16 = vp8_sixtap_predict16x16_sss e3; 666 if (flags & HAS_SSSE3) vp8_sixtap_predict16x16 = vp8_sixtap_predict16x16_sss e3;
745
746 vp8_sixtap_predict4x4 = vp8_sixtap_predict4x4_c; 667 vp8_sixtap_predict4x4 = vp8_sixtap_predict4x4_c;
747 if (flags & HAS_MMX) vp8_sixtap_predict4x4 = vp8_sixtap_predict4x4_mmx; 668 if (flags & HAS_MMX) vp8_sixtap_predict4x4 = vp8_sixtap_predict4x4_mmx;
748 if (flags & HAS_SSSE3) vp8_sixtap_predict4x4 = vp8_sixtap_predict4x4_ssse3; 669 if (flags & HAS_SSSE3) vp8_sixtap_predict4x4 = vp8_sixtap_predict4x4_ssse3;
749
750 vp8_sixtap_predict8x4 = vp8_sixtap_predict8x4_c; 670 vp8_sixtap_predict8x4 = vp8_sixtap_predict8x4_c;
751 if (flags & HAS_MMX) vp8_sixtap_predict8x4 = vp8_sixtap_predict8x4_mmx; 671 if (flags & HAS_MMX) vp8_sixtap_predict8x4 = vp8_sixtap_predict8x4_mmx;
752 if (flags & HAS_SSE2) vp8_sixtap_predict8x4 = vp8_sixtap_predict8x4_sse2; 672 if (flags & HAS_SSE2) vp8_sixtap_predict8x4 = vp8_sixtap_predict8x4_sse2;
753 if (flags & HAS_SSSE3) vp8_sixtap_predict8x4 = vp8_sixtap_predict8x4_ssse3; 673 if (flags & HAS_SSSE3) vp8_sixtap_predict8x4 = vp8_sixtap_predict8x4_ssse3;
754
755 vp8_sixtap_predict8x8 = vp8_sixtap_predict8x8_c; 674 vp8_sixtap_predict8x8 = vp8_sixtap_predict8x8_c;
756 if (flags & HAS_MMX) vp8_sixtap_predict8x8 = vp8_sixtap_predict8x8_mmx; 675 if (flags & HAS_MMX) vp8_sixtap_predict8x8 = vp8_sixtap_predict8x8_mmx;
757 if (flags & HAS_SSE2) vp8_sixtap_predict8x8 = vp8_sixtap_predict8x8_sse2; 676 if (flags & HAS_SSE2) vp8_sixtap_predict8x8 = vp8_sixtap_predict8x8_sse2;
758 if (flags & HAS_SSSE3) vp8_sixtap_predict8x8 = vp8_sixtap_predict8x8_ssse3; 677 if (flags & HAS_SSSE3) vp8_sixtap_predict8x8 = vp8_sixtap_predict8x8_ssse3;
759
760 vp8_sub_pixel_mse16x16 = vp8_sub_pixel_mse16x16_c; 678 vp8_sub_pixel_mse16x16 = vp8_sub_pixel_mse16x16_c;
761 if (flags & HAS_MMX) vp8_sub_pixel_mse16x16 = vp8_sub_pixel_mse16x16_mmx; 679 if (flags & HAS_MMX) vp8_sub_pixel_mse16x16 = vp8_sub_pixel_mse16x16_mmx;
762 if (flags & HAS_SSE2) vp8_sub_pixel_mse16x16 = vp8_sub_pixel_mse16x16_wmt; 680 if (flags & HAS_SSE2) vp8_sub_pixel_mse16x16 = vp8_sub_pixel_mse16x16_wmt;
763
764 vp8_sub_pixel_variance16x16 = vp8_sub_pixel_variance16x16_c; 681 vp8_sub_pixel_variance16x16 = vp8_sub_pixel_variance16x16_c;
765 if (flags & HAS_MMX) vp8_sub_pixel_variance16x16 = vp8_sub_pixel_variance16x 16_mmx; 682 if (flags & HAS_MMX) vp8_sub_pixel_variance16x16 = vp8_sub_pixel_variance16x 16_mmx;
766 if (flags & HAS_SSE2) vp8_sub_pixel_variance16x16 = vp8_sub_pixel_variance16 x16_wmt; 683 if (flags & HAS_SSE2) vp8_sub_pixel_variance16x16 = vp8_sub_pixel_variance16 x16_wmt;
767 if (flags & HAS_SSSE3) vp8_sub_pixel_variance16x16 = vp8_sub_pixel_variance1 6x16_ssse3; 684 if (flags & HAS_SSSE3) vp8_sub_pixel_variance16x16 = vp8_sub_pixel_variance1 6x16_ssse3;
768
769 vp8_sub_pixel_variance16x8 = vp8_sub_pixel_variance16x8_c; 685 vp8_sub_pixel_variance16x8 = vp8_sub_pixel_variance16x8_c;
770 if (flags & HAS_MMX) vp8_sub_pixel_variance16x8 = vp8_sub_pixel_variance16x8 _mmx; 686 if (flags & HAS_MMX) vp8_sub_pixel_variance16x8 = vp8_sub_pixel_variance16x8 _mmx;
771 if (flags & HAS_SSE2) vp8_sub_pixel_variance16x8 = vp8_sub_pixel_variance16x 8_wmt; 687 if (flags & HAS_SSE2) vp8_sub_pixel_variance16x8 = vp8_sub_pixel_variance16x 8_wmt;
772 if (flags & HAS_SSSE3) vp8_sub_pixel_variance16x8 = vp8_sub_pixel_variance16 x8_ssse3; 688 if (flags & HAS_SSSE3) vp8_sub_pixel_variance16x8 = vp8_sub_pixel_variance16 x8_ssse3;
773
774 vp8_sub_pixel_variance4x4 = vp8_sub_pixel_variance4x4_c; 689 vp8_sub_pixel_variance4x4 = vp8_sub_pixel_variance4x4_c;
775 if (flags & HAS_MMX) vp8_sub_pixel_variance4x4 = vp8_sub_pixel_variance4x4_m mx; 690 if (flags & HAS_MMX) vp8_sub_pixel_variance4x4 = vp8_sub_pixel_variance4x4_m mx;
776 if (flags & HAS_SSE2) vp8_sub_pixel_variance4x4 = vp8_sub_pixel_variance4x4_ wmt; 691 if (flags & HAS_SSE2) vp8_sub_pixel_variance4x4 = vp8_sub_pixel_variance4x4_ wmt;
777
778 vp8_sub_pixel_variance8x16 = vp8_sub_pixel_variance8x16_c; 692 vp8_sub_pixel_variance8x16 = vp8_sub_pixel_variance8x16_c;
779 if (flags & HAS_MMX) vp8_sub_pixel_variance8x16 = vp8_sub_pixel_variance8x16 _mmx; 693 if (flags & HAS_MMX) vp8_sub_pixel_variance8x16 = vp8_sub_pixel_variance8x16 _mmx;
780 if (flags & HAS_SSE2) vp8_sub_pixel_variance8x16 = vp8_sub_pixel_variance8x1 6_wmt; 694 if (flags & HAS_SSE2) vp8_sub_pixel_variance8x16 = vp8_sub_pixel_variance8x1 6_wmt;
781
782 vp8_sub_pixel_variance8x8 = vp8_sub_pixel_variance8x8_c; 695 vp8_sub_pixel_variance8x8 = vp8_sub_pixel_variance8x8_c;
783 if (flags & HAS_MMX) vp8_sub_pixel_variance8x8 = vp8_sub_pixel_variance8x8_m mx; 696 if (flags & HAS_MMX) vp8_sub_pixel_variance8x8 = vp8_sub_pixel_variance8x8_m mx;
784 if (flags & HAS_SSE2) vp8_sub_pixel_variance8x8 = vp8_sub_pixel_variance8x8_ wmt; 697 if (flags & HAS_SSE2) vp8_sub_pixel_variance8x8 = vp8_sub_pixel_variance8x8_ wmt;
785
786 vp8_subtract_b = vp8_subtract_b_c; 698 vp8_subtract_b = vp8_subtract_b_c;
787 if (flags & HAS_MMX) vp8_subtract_b = vp8_subtract_b_mmx; 699 if (flags & HAS_MMX) vp8_subtract_b = vp8_subtract_b_mmx;
788 if (flags & HAS_SSE2) vp8_subtract_b = vp8_subtract_b_sse2; 700 if (flags & HAS_SSE2) vp8_subtract_b = vp8_subtract_b_sse2;
789
790 vp8_subtract_mbuv = vp8_subtract_mbuv_c; 701 vp8_subtract_mbuv = vp8_subtract_mbuv_c;
791 if (flags & HAS_MMX) vp8_subtract_mbuv = vp8_subtract_mbuv_mmx; 702 if (flags & HAS_MMX) vp8_subtract_mbuv = vp8_subtract_mbuv_mmx;
792 if (flags & HAS_SSE2) vp8_subtract_mbuv = vp8_subtract_mbuv_sse2; 703 if (flags & HAS_SSE2) vp8_subtract_mbuv = vp8_subtract_mbuv_sse2;
793
794 vp8_subtract_mby = vp8_subtract_mby_c; 704 vp8_subtract_mby = vp8_subtract_mby_c;
795 if (flags & HAS_MMX) vp8_subtract_mby = vp8_subtract_mby_mmx; 705 if (flags & HAS_MMX) vp8_subtract_mby = vp8_subtract_mby_mmx;
796 if (flags & HAS_SSE2) vp8_subtract_mby = vp8_subtract_mby_sse2; 706 if (flags & HAS_SSE2) vp8_subtract_mby = vp8_subtract_mby_sse2;
797
798 vp8_variance16x16 = vp8_variance16x16_c; 707 vp8_variance16x16 = vp8_variance16x16_c;
799 if (flags & HAS_MMX) vp8_variance16x16 = vp8_variance16x16_mmx; 708 if (flags & HAS_MMX) vp8_variance16x16 = vp8_variance16x16_mmx;
800 if (flags & HAS_SSE2) vp8_variance16x16 = vp8_variance16x16_wmt; 709 if (flags & HAS_SSE2) vp8_variance16x16 = vp8_variance16x16_wmt;
801
802 vp8_variance16x8 = vp8_variance16x8_c; 710 vp8_variance16x8 = vp8_variance16x8_c;
803 if (flags & HAS_MMX) vp8_variance16x8 = vp8_variance16x8_mmx; 711 if (flags & HAS_MMX) vp8_variance16x8 = vp8_variance16x8_mmx;
804 if (flags & HAS_SSE2) vp8_variance16x8 = vp8_variance16x8_wmt; 712 if (flags & HAS_SSE2) vp8_variance16x8 = vp8_variance16x8_wmt;
805
806 vp8_variance4x4 = vp8_variance4x4_c; 713 vp8_variance4x4 = vp8_variance4x4_c;
807 if (flags & HAS_MMX) vp8_variance4x4 = vp8_variance4x4_mmx; 714 if (flags & HAS_MMX) vp8_variance4x4 = vp8_variance4x4_mmx;
808 if (flags & HAS_SSE2) vp8_variance4x4 = vp8_variance4x4_wmt; 715 if (flags & HAS_SSE2) vp8_variance4x4 = vp8_variance4x4_wmt;
809
810 vp8_variance8x16 = vp8_variance8x16_c; 716 vp8_variance8x16 = vp8_variance8x16_c;
811 if (flags & HAS_MMX) vp8_variance8x16 = vp8_variance8x16_mmx; 717 if (flags & HAS_MMX) vp8_variance8x16 = vp8_variance8x16_mmx;
812 if (flags & HAS_SSE2) vp8_variance8x16 = vp8_variance8x16_wmt; 718 if (flags & HAS_SSE2) vp8_variance8x16 = vp8_variance8x16_wmt;
813
814 vp8_variance8x8 = vp8_variance8x8_c; 719 vp8_variance8x8 = vp8_variance8x8_c;
815 if (flags & HAS_MMX) vp8_variance8x8 = vp8_variance8x8_mmx; 720 if (flags & HAS_MMX) vp8_variance8x8 = vp8_variance8x8_mmx;
816 if (flags & HAS_SSE2) vp8_variance8x8 = vp8_variance8x8_wmt; 721 if (flags & HAS_SSE2) vp8_variance8x8 = vp8_variance8x8_wmt;
817
818 vp8_variance_halfpixvar16x16_h = vp8_variance_halfpixvar16x16_h_c; 722 vp8_variance_halfpixvar16x16_h = vp8_variance_halfpixvar16x16_h_c;
819 if (flags & HAS_MMX) vp8_variance_halfpixvar16x16_h = vp8_variance_halfpixva r16x16_h_mmx; 723 if (flags & HAS_MMX) vp8_variance_halfpixvar16x16_h = vp8_variance_halfpixva r16x16_h_mmx;
820 if (flags & HAS_SSE2) vp8_variance_halfpixvar16x16_h = vp8_variance_halfpixv ar16x16_h_wmt; 724 if (flags & HAS_SSE2) vp8_variance_halfpixvar16x16_h = vp8_variance_halfpixv ar16x16_h_wmt;
821
822 vp8_variance_halfpixvar16x16_hv = vp8_variance_halfpixvar16x16_hv_c; 725 vp8_variance_halfpixvar16x16_hv = vp8_variance_halfpixvar16x16_hv_c;
823 if (flags & HAS_MMX) vp8_variance_halfpixvar16x16_hv = vp8_variance_halfpixv ar16x16_hv_mmx; 726 if (flags & HAS_MMX) vp8_variance_halfpixvar16x16_hv = vp8_variance_halfpixv ar16x16_hv_mmx;
824 if (flags & HAS_SSE2) vp8_variance_halfpixvar16x16_hv = vp8_variance_halfpix var16x16_hv_wmt; 727 if (flags & HAS_SSE2) vp8_variance_halfpixvar16x16_hv = vp8_variance_halfpix var16x16_hv_wmt;
825
826 vp8_variance_halfpixvar16x16_v = vp8_variance_halfpixvar16x16_v_c; 728 vp8_variance_halfpixvar16x16_v = vp8_variance_halfpixvar16x16_v_c;
827 if (flags & HAS_MMX) vp8_variance_halfpixvar16x16_v = vp8_variance_halfpixva r16x16_v_mmx; 729 if (flags & HAS_MMX) vp8_variance_halfpixvar16x16_v = vp8_variance_halfpixva r16x16_v_mmx;
828 if (flags & HAS_SSE2) vp8_variance_halfpixvar16x16_v = vp8_variance_halfpixv ar16x16_v_wmt; 730 if (flags & HAS_SSE2) vp8_variance_halfpixvar16x16_v = vp8_variance_halfpixv ar16x16_v_wmt;
829 } 731 }
830 #endif 732 #endif
831 733
832 #ifdef __cplusplus 734 #ifdef __cplusplus
833 } // extern "C" 735 } // extern "C"
834 #endif 736 #endif
835 737
836 #endif 738 #endif
OLDNEW
« no previous file with comments | « source/config/linux/generic/vpx_scale_rtcd.h ('k') | source/config/linux/ia32/vp9_rtcd.h » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698