| Index: source/libvpx/vp9/common/mips/msa/vp9_loopfilter_msa.h
|
| diff --git a/source/libvpx/vp9/common/mips/msa/vp9_loopfilter_msa.h b/source/libvpx/vp9/common/mips/msa/vp9_loopfilter_msa.h
|
| new file mode 100644
|
| index 0000000000000000000000000000000000000000..0643e41a5208b2d733274db335c9bfe1827a82f5
|
| --- /dev/null
|
| +++ b/source/libvpx/vp9/common/mips/msa/vp9_loopfilter_msa.h
|
| @@ -0,0 +1,246 @@
|
| +/*
|
| + * Copyright (c) 2015 The WebM project authors. All Rights Reserved.
|
| + *
|
| + * Use of this source code is governed by a BSD-style license
|
| + * that can be found in the LICENSE file in the root of the source
|
| + * tree. An additional intellectual property rights grant can be found
|
| + * in the file PATENTS. All contributing project authors may
|
| + * be found in the AUTHORS file in the root of the source tree.
|
| + */
|
| +
|
| +#ifndef VP9_COMMON_MIPS_MSA_VP9_LOOPFILTER_MSA_H_
|
| +#define VP9_COMMON_MIPS_MSA_VP9_LOOPFILTER_MSA_H_
|
| +
|
| +#include "vp9/common/mips/msa/vp9_macros_msa.h"
|
| +
|
| +#define VP9_LPF_FILTER4_8W(p1_in, p0_in, q0_in, q1_in, mask_in, hev_in, \
|
| + p1_out, p0_out, q0_out, q1_out) { \
|
| + v16i8 p1_m, p0_m, q0_m, q1_m, q0_sub_p0, filt_sign; \
|
| + v16i8 filt, filt1, filt2, cnst4b, cnst3b; \
|
| + v8i16 q0_sub_p0_r, filt_r, cnst3h; \
|
| + \
|
| + p1_m = (v16i8)__msa_xori_b(p1_in, 0x80); \
|
| + p0_m = (v16i8)__msa_xori_b(p0_in, 0x80); \
|
| + q0_m = (v16i8)__msa_xori_b(q0_in, 0x80); \
|
| + q1_m = (v16i8)__msa_xori_b(q1_in, 0x80); \
|
| + \
|
| + filt = __msa_subs_s_b(p1_m, q1_m); \
|
| + filt = filt & (v16i8)hev_in; \
|
| + q0_sub_p0 = q0_m - p0_m; \
|
| + filt_sign = __msa_clti_s_b(filt, 0); \
|
| + \
|
| + cnst3h = __msa_ldi_h(3); \
|
| + q0_sub_p0_r = (v8i16)__msa_ilvr_b(q0_sub_p0, q0_sub_p0); \
|
| + q0_sub_p0_r = __msa_dotp_s_h((v16i8)q0_sub_p0_r, (v16i8)cnst3h); \
|
| + filt_r = (v8i16)__msa_ilvr_b(filt_sign, filt); \
|
| + filt_r += q0_sub_p0_r; \
|
| + filt_r = __msa_sat_s_h(filt_r, 7); \
|
| + \
|
| + /* combine left and right part */ \
|
| + filt = __msa_pckev_b((v16i8)filt_r, (v16i8)filt_r); \
|
| + \
|
| + filt = filt & (v16i8)mask_in; \
|
| + cnst4b = __msa_ldi_b(4); \
|
| + filt1 = __msa_adds_s_b(filt, cnst4b); \
|
| + filt1 >>= 3; \
|
| + \
|
| + cnst3b = __msa_ldi_b(3); \
|
| + filt2 = __msa_adds_s_b(filt, cnst3b); \
|
| + filt2 >>= 3; \
|
| + \
|
| + q0_m = __msa_subs_s_b(q0_m, filt1); \
|
| + q0_out = __msa_xori_b((v16u8)q0_m, 0x80); \
|
| + p0_m = __msa_adds_s_b(p0_m, filt2); \
|
| + p0_out = __msa_xori_b((v16u8)p0_m, 0x80); \
|
| + \
|
| + filt = __msa_srari_b(filt1, 1); \
|
| + hev_in = __msa_xori_b((v16u8)hev_in, 0xff); \
|
| + filt = filt & (v16i8)hev_in; \
|
| + \
|
| + q1_m = __msa_subs_s_b(q1_m, filt); \
|
| + q1_out = __msa_xori_b((v16u8)q1_m, 0x80); \
|
| + p1_m = __msa_adds_s_b(p1_m, filt); \
|
| + p1_out = __msa_xori_b((v16u8)p1_m, 0x80); \
|
| +}
|
| +
|
| +#define VP9_LPF_FILTER4_4W(p1_in, p0_in, q0_in, q1_in, mask_in, hev_in, \
|
| + p1_out, p0_out, q0_out, q1_out) { \
|
| + v16i8 p1_m, p0_m, q0_m, q1_m, q0_sub_p0, filt_sign; \
|
| + v16i8 filt, filt1, filt2, cnst4b, cnst3b; \
|
| + v8i16 q0_sub_p0_r, q0_sub_p0_l, filt_l, filt_r, cnst3h; \
|
| + \
|
| + p1_m = (v16i8)__msa_xori_b(p1_in, 0x80); \
|
| + p0_m = (v16i8)__msa_xori_b(p0_in, 0x80); \
|
| + q0_m = (v16i8)__msa_xori_b(q0_in, 0x80); \
|
| + q1_m = (v16i8)__msa_xori_b(q1_in, 0x80); \
|
| + \
|
| + filt = __msa_subs_s_b(p1_m, q1_m); \
|
| + \
|
| + filt = filt & (v16i8)hev_in; \
|
| + \
|
| + q0_sub_p0 = q0_m - p0_m; \
|
| + filt_sign = __msa_clti_s_b(filt, 0); \
|
| + \
|
| + cnst3h = __msa_ldi_h(3); \
|
| + q0_sub_p0_r = (v8i16)__msa_ilvr_b(q0_sub_p0, q0_sub_p0); \
|
| + q0_sub_p0_r = __msa_dotp_s_h((v16i8)q0_sub_p0_r, (v16i8)cnst3h); \
|
| + filt_r = (v8i16)__msa_ilvr_b(filt_sign, filt); \
|
| + filt_r += q0_sub_p0_r; \
|
| + filt_r = __msa_sat_s_h(filt_r, 7); \
|
| + \
|
| + q0_sub_p0_l = (v8i16)__msa_ilvl_b(q0_sub_p0, q0_sub_p0); \
|
| + q0_sub_p0_l = __msa_dotp_s_h((v16i8)q0_sub_p0_l, (v16i8)cnst3h); \
|
| + filt_l = (v8i16)__msa_ilvl_b(filt_sign, filt); \
|
| + filt_l += q0_sub_p0_l; \
|
| + filt_l = __msa_sat_s_h(filt_l, 7); \
|
| + \
|
| + filt = __msa_pckev_b((v16i8)filt_l, (v16i8)filt_r); \
|
| + filt = filt & (v16i8)mask_in; \
|
| + \
|
| + cnst4b = __msa_ldi_b(4); \
|
| + filt1 = __msa_adds_s_b(filt, cnst4b); \
|
| + filt1 >>= 3; \
|
| + \
|
| + cnst3b = __msa_ldi_b(3); \
|
| + filt2 = __msa_adds_s_b(filt, cnst3b); \
|
| + filt2 >>= 3; \
|
| + \
|
| + q0_m = __msa_subs_s_b(q0_m, filt1); \
|
| + q0_out = __msa_xori_b((v16u8)q0_m, 0x80); \
|
| + p0_m = __msa_adds_s_b(p0_m, filt2); \
|
| + p0_out = __msa_xori_b((v16u8)p0_m, 0x80); \
|
| + \
|
| + filt = __msa_srari_b(filt1, 1); \
|
| + hev_in = __msa_xori_b((v16u8)hev_in, 0xff); \
|
| + filt = filt & (v16i8)hev_in; \
|
| + \
|
| + q1_m = __msa_subs_s_b(q1_m, filt); \
|
| + q1_out = __msa_xori_b((v16u8)q1_m, 0x80); \
|
| + p1_m = __msa_adds_s_b(p1_m, filt); \
|
| + p1_out = __msa_xori_b((v16u8)p1_m, 0x80); \
|
| +}
|
| +
|
| +#define VP9_FLAT4(p3_in, p2_in, p0_in, q0_in, q2_in, q3_in, flat_out) { \
|
| + v16u8 tmp, p2_a_sub_p0, q2_a_sub_q0, p3_a_sub_p0, q3_a_sub_q0; \
|
| + v16u8 zero_in = { 0 }; \
|
| + \
|
| + tmp = __msa_ori_b(zero_in, 1); \
|
| + p2_a_sub_p0 = __msa_asub_u_b(p2_in, p0_in); \
|
| + q2_a_sub_q0 = __msa_asub_u_b(q2_in, q0_in); \
|
| + p3_a_sub_p0 = __msa_asub_u_b(p3_in, p0_in); \
|
| + q3_a_sub_q0 = __msa_asub_u_b(q3_in, q0_in); \
|
| + \
|
| + p2_a_sub_p0 = __msa_max_u_b(p2_a_sub_p0, q2_a_sub_q0); \
|
| + flat_out = __msa_max_u_b(p2_a_sub_p0, flat_out); \
|
| + p3_a_sub_p0 = __msa_max_u_b(p3_a_sub_p0, q3_a_sub_q0); \
|
| + flat_out = __msa_max_u_b(p3_a_sub_p0, flat_out); \
|
| + \
|
| + flat_out = (tmp < (v16u8)flat_out); \
|
| + flat_out = __msa_xori_b(flat_out, 0xff); \
|
| + flat_out = flat_out & (mask); \
|
| +}
|
| +
|
| +#define VP9_FLAT5(p7_in, p6_in, p5_in, p4_in, p0_in, q0_in, q4_in, \
|
| + q5_in, q6_in, q7_in, flat_in, flat2_out) { \
|
| + v16u8 tmp, zero_in = { 0 }; \
|
| + v16u8 p4_a_sub_p0, q4_a_sub_q0, p5_a_sub_p0, q5_a_sub_q0; \
|
| + v16u8 p6_a_sub_p0, q6_a_sub_q0, p7_a_sub_p0, q7_a_sub_q0; \
|
| + \
|
| + tmp = __msa_ori_b(zero_in, 1); \
|
| + p4_a_sub_p0 = __msa_asub_u_b(p4_in, p0_in); \
|
| + q4_a_sub_q0 = __msa_asub_u_b(q4_in, q0_in); \
|
| + p5_a_sub_p0 = __msa_asub_u_b(p5_in, p0_in); \
|
| + q5_a_sub_q0 = __msa_asub_u_b(q5_in, q0_in); \
|
| + p6_a_sub_p0 = __msa_asub_u_b(p6_in, p0_in); \
|
| + q6_a_sub_q0 = __msa_asub_u_b(q6_in, q0_in); \
|
| + p7_a_sub_p0 = __msa_asub_u_b(p7_in, p0_in); \
|
| + q7_a_sub_q0 = __msa_asub_u_b(q7_in, q0_in); \
|
| + \
|
| + p4_a_sub_p0 = __msa_max_u_b(p4_a_sub_p0, q4_a_sub_q0); \
|
| + flat2_out = __msa_max_u_b(p5_a_sub_p0, q5_a_sub_q0); \
|
| + flat2_out = __msa_max_u_b(p4_a_sub_p0, flat2_out); \
|
| + p6_a_sub_p0 = __msa_max_u_b(p6_a_sub_p0, q6_a_sub_q0); \
|
| + flat2_out = __msa_max_u_b(p6_a_sub_p0, flat2_out); \
|
| + p7_a_sub_p0 = __msa_max_u_b(p7_a_sub_p0, q7_a_sub_q0); \
|
| + flat2_out = __msa_max_u_b(p7_a_sub_p0, flat2_out); \
|
| + \
|
| + flat2_out = (tmp < (v16u8)flat2_out); \
|
| + flat2_out = __msa_xori_b(flat2_out, 0xff); \
|
| + flat2_out = flat2_out & flat_in; \
|
| +}
|
| +
|
| +#define VP9_FILTER8(p3_in, p2_in, p1_in, p0_in, \
|
| + q0_in, q1_in, q2_in, q3_in, \
|
| + p2_filt8_out, p1_filt8_out, p0_filt8_out, \
|
| + q0_filt8_out, q1_filt8_out, q2_filt8_out) { \
|
| + v8u16 tmp0, tmp1, tmp2; \
|
| + \
|
| + tmp2 = p2_in + p1_in + p0_in; \
|
| + tmp0 = p3_in << 1; \
|
| + \
|
| + tmp0 = tmp0 + tmp2 + q0_in; \
|
| + tmp1 = tmp0 + p3_in + p2_in; \
|
| + p2_filt8_out = (v8i16)__msa_srari_h((v8i16)tmp1, 3); \
|
| + \
|
| + tmp1 = tmp0 + p1_in + q1_in; \
|
| + p1_filt8_out = (v8i16)__msa_srari_h((v8i16)tmp1, 3); \
|
| + \
|
| + tmp1 = q2_in + q1_in + q0_in; \
|
| + tmp2 = tmp2 + tmp1; \
|
| + tmp0 = tmp2 + (p0_in); \
|
| + tmp0 = tmp0 + (p3_in); \
|
| + p0_filt8_out = (v8i16)__msa_srari_h((v8i16)tmp0, 3); \
|
| + \
|
| + tmp0 = q2_in + q3_in; \
|
| + tmp0 = p0_in + tmp1 + tmp0; \
|
| + tmp1 = q3_in + q3_in; \
|
| + tmp1 = tmp1 + tmp0; \
|
| + q2_filt8_out = (v8i16)__msa_srari_h((v8i16)tmp1, 3); \
|
| + \
|
| + tmp0 = tmp2 + q3_in; \
|
| + tmp1 = tmp0 + q0_in; \
|
| + q0_filt8_out = (v8i16)__msa_srari_h((v8i16)tmp1, 3); \
|
| + \
|
| + tmp1 = tmp0 - p2_in; \
|
| + tmp0 = q1_in + q3_in; \
|
| + tmp1 = tmp0 + tmp1; \
|
| + q1_filt8_out = (v8i16)__msa_srari_h((v8i16)tmp1, 3); \
|
| +}
|
| +
|
| +#define LPF_MASK_HEV(p3_in, p2_in, p1_in, p0_in, \
|
| + q0_in, q1_in, q2_in, q3_in, \
|
| + limit_in, b_limit_in, thresh_in, \
|
| + hev_out, mask_out, flat_out) { \
|
| + v16u8 p3_asub_p2_m, p2_asub_p1_m, p1_asub_p0_m, q1_asub_q0_m; \
|
| + v16u8 p1_asub_q1_m, p0_asub_q0_m, q3_asub_q2_m, q2_asub_q1_m; \
|
| + \
|
| + /* absolute subtraction of pixel values */ \
|
| + p3_asub_p2_m = __msa_asub_u_b(p3_in, p2_in); \
|
| + p2_asub_p1_m = __msa_asub_u_b(p2_in, p1_in); \
|
| + p1_asub_p0_m = __msa_asub_u_b(p1_in, p0_in); \
|
| + q1_asub_q0_m = __msa_asub_u_b(q1_in, q0_in); \
|
| + q2_asub_q1_m = __msa_asub_u_b(q2_in, q1_in); \
|
| + q3_asub_q2_m = __msa_asub_u_b(q3_in, q2_in); \
|
| + p0_asub_q0_m = __msa_asub_u_b(p0_in, q0_in); \
|
| + p1_asub_q1_m = __msa_asub_u_b(p1_in, q1_in); \
|
| + \
|
| + /* calculation of hev */ \
|
| + flat_out = __msa_max_u_b(p1_asub_p0_m, q1_asub_q0_m); \
|
| + hev_out = thresh_in < (v16u8)flat_out; \
|
| + \
|
| + /* calculation of mask */ \
|
| + p0_asub_q0_m = __msa_adds_u_b(p0_asub_q0_m, p0_asub_q0_m); \
|
| + p1_asub_q1_m >>= 1; \
|
| + p0_asub_q0_m = __msa_adds_u_b(p0_asub_q0_m, p1_asub_q1_m); \
|
| + \
|
| + mask_out = b_limit_in < p0_asub_q0_m; \
|
| + mask_out = __msa_max_u_b(flat_out, mask_out); \
|
| + p3_asub_p2_m = __msa_max_u_b(p3_asub_p2_m, p2_asub_p1_m); \
|
| + mask_out = __msa_max_u_b(p3_asub_p2_m, mask_out); \
|
| + q2_asub_q1_m = __msa_max_u_b(q2_asub_q1_m, q3_asub_q2_m); \
|
| + mask_out = __msa_max_u_b(q2_asub_q1_m, mask_out); \
|
| + \
|
| + mask_out = limit_in < (v16u8)mask_out; \
|
| + mask_out = __msa_xori_b(mask_out, 0xff); \
|
| +}
|
| +#endif /* VP9_COMMON_MIPS_MSA_VP9_LOOPFILTER_MSA_H_ */
|
|
|