OLD | NEW |
| (Empty) |
1 /* | |
2 * Copyright (c) 2010 The WebM project authors. All Rights Reserved. | |
3 * | |
4 * Use of this source code is governed by a BSD-style license | |
5 * that can be found in the LICENSE file in the root of the source | |
6 * tree. An additional intellectual property rights grant can be found | |
7 * in the file PATENTS. All contributing project authors may | |
8 * be found in the AUTHORS file in the root of the source tree. | |
9 */ | |
10 | |
11 #include "./vpx_config.h" | |
12 #include "vp9/encoder/vp9_variance.h" | |
13 #include "vpx_ports/mem.h" | |
14 | |
15 unsigned int vp9_get8x8var_mmx(const uint8_t *src, int src_stride, | |
16 const uint8_t *ref, int ref_stride, | |
17 unsigned int *sse, int *sum); | |
18 | |
19 unsigned int vp9_get4x4var_mmx(const uint8_t *src, int src_stride, | |
20 const uint8_t *ref, int ref_stride, | |
21 unsigned int *SSE, int *sum); | |
22 | |
23 unsigned int vp9_variance4x4_mmx(const uint8_t *src, int src_stride, | |
24 const uint8_t *ref, int ref_stride, | |
25 unsigned int *sse) { | |
26 int sum; | |
27 vp9_get4x4var_mmx(src, src_stride, ref, ref_stride, sse, &sum); | |
28 return *sse - (((unsigned int)sum * sum) >> 4); | |
29 } | |
30 | |
31 unsigned int vp9_variance8x8_mmx(const uint8_t *src, int src_stride, | |
32 const uint8_t *ref, int ref_stride, | |
33 unsigned int *sse) { | |
34 int sum; | |
35 vp9_get8x8var_mmx(src, src_stride, ref, ref_stride, sse, &sum); | |
36 return *sse - (((unsigned int)sum * sum) >> 6); | |
37 } | |
38 | |
39 unsigned int vp9_mse16x16_mmx(const uint8_t *src, int src_stride, | |
40 const uint8_t *ref, int ref_stride, | |
41 unsigned int *sse) { | |
42 unsigned int sse0, sse1, sse2, sse3; | |
43 int sum0, sum1, sum2, sum3; | |
44 | |
45 vp9_get8x8var_mmx(src, src_stride, ref, ref_stride, &sse0, &sum0); | |
46 vp9_get8x8var_mmx(src + 8, src_stride, ref + 8, ref_stride, &sse1, &sum1); | |
47 vp9_get8x8var_mmx(src + 8 * src_stride, src_stride, | |
48 ref + 8 * ref_stride, ref_stride, &sse2, &sum2); | |
49 vp9_get8x8var_mmx(src + 8 * src_stride + 8, src_stride, | |
50 ref + 8 * ref_stride + 8, ref_stride, &sse3, &sum3); | |
51 | |
52 *sse = sse0 + sse1 + sse2 + sse3; | |
53 return *sse; | |
54 } | |
55 | |
56 | |
57 unsigned int vp9_variance16x16_mmx(const uint8_t *src, int src_stride, | |
58 const uint8_t *ref, int ref_stride, | |
59 unsigned int *sse) { | |
60 unsigned int sse0, sse1, sse2, sse3; | |
61 int sum0, sum1, sum2, sum3, sum; | |
62 | |
63 vp9_get8x8var_mmx(src, src_stride, ref, ref_stride, &sse0, &sum0); | |
64 vp9_get8x8var_mmx(src + 8, src_stride, ref + 8, ref_stride, &sse1, &sum1); | |
65 vp9_get8x8var_mmx(src + 8 * src_stride, src_stride, | |
66 ref + 8 * ref_stride, ref_stride, &sse2, &sum2); | |
67 vp9_get8x8var_mmx(src + 8 * src_stride + 8, src_stride, | |
68 ref + 8 * ref_stride + 8, ref_stride, &sse3, &sum3); | |
69 | |
70 *sse = sse0 + sse1 + sse2 + sse3; | |
71 sum = sum0 + sum1 + sum2 + sum3; | |
72 return *sse - (((unsigned int)sum * sum) >> 8); | |
73 } | |
74 | |
75 unsigned int vp9_variance16x8_mmx(const uint8_t *src, int src_stride, | |
76 const uint8_t *ref, int ref_stride, | |
77 unsigned int *sse) { | |
78 unsigned int sse0, sse1; | |
79 int sum0, sum1, sum; | |
80 | |
81 vp9_get8x8var_mmx(src, src_stride, ref, ref_stride, &sse0, &sum0); | |
82 vp9_get8x8var_mmx(src + 8, src_stride, ref + 8, ref_stride, &sse1, &sum1); | |
83 | |
84 *sse = sse0 + sse1; | |
85 sum = sum0 + sum1; | |
86 return *sse - (((unsigned int)sum * sum) >> 7); | |
87 } | |
88 | |
89 | |
90 unsigned int vp9_variance8x16_mmx(const uint8_t *src, int src_stride, | |
91 const uint8_t *ref, int ref_stride, | |
92 unsigned int *sse) { | |
93 unsigned int sse0, sse1; | |
94 int sum0, sum1, sum; | |
95 | |
96 vp9_get8x8var_mmx(src, src_stride, ref, ref_stride, &sse0, &sum0); | |
97 vp9_get8x8var_mmx(src + 8 * src_stride, src_stride, | |
98 ref + 8 * ref_stride, ref_stride, &sse1, &sum1); | |
99 | |
100 *sse = sse0 + sse1; | |
101 sum = sum0 + sum1; | |
102 return *sse - (((unsigned int)sum * sum) >> 7); | |
103 } | |
OLD | NEW |