Chromium Code Reviews| Index: src/opts/SkXfermode_opts_arm_neon.cpp |
| diff --git a/src/opts/SkXfermode_opts_arm_neon.cpp b/src/opts/SkXfermode_opts_arm_neon.cpp |
| new file mode 100644 |
| index 0000000000000000000000000000000000000000..f4ff18c37e614da8dadb1ea18b6d0b58312cd3e8 |
| --- /dev/null |
| +++ b/src/opts/SkXfermode_opts_arm_neon.cpp |
| @@ -0,0 +1,698 @@ |
| +#include "SkXfermode.h" |
| +#include "SkXfermode_proccoeff.h" |
| +#include "SkColorPriv.h" |
| + |
| +#include <arm_neon.h> |
| +#include "SkColor_opts_neon.h" |
| +#include "SkXfermode_opts_arm_neon.h" |
| + |
| +#define SkAlphaMulAlpha(a, b) SkMulDiv255Round(a, b) |
| + |
| + |
| +//////////////////////////////////////////////////////////////////////////////// |
| +// NEONized skia functions |
| +//////////////////////////////////////////////////////////////////////////////// |
| + |
| +static inline uint8x8_t SkAlphaMulAlpha_neon8(uint8x8_t color, uint8x8_t alpha) { |
| + uint16x8_t tmp; |
| + uint8x8_t ret; |
| + |
| + tmp = vmull_u8(color, alpha); |
| + tmp = vaddq_u16(tmp, vdupq_n_u16(128)); |
| + tmp = vaddq_u16(tmp, vshrq_n_u16(tmp, 8)); |
| + |
| + ret = vshrn_n_u16(tmp, 8); |
| + |
| + return ret; |
| +} |
| + |
| +static inline uint16x8_t SkAlphaMulAlpha_neon8_16(uint8x8_t color, uint8x8_t alpha) { |
| + uint16x8_t ret; |
| + |
| + ret = vmull_u8(color, alpha); |
| + ret = vaddq_u16(ret, vdupq_n_u16(128)); |
| + ret = vaddq_u16(ret, vshrq_n_u16(ret, 8)); |
| + |
| + ret = vshrq_n_u16(ret, 8); |
| + |
| + return ret; |
| +} |
| + |
| +static inline uint8x8_t SkDiv255Round_neon8_32_8(int32x4_t p1, int32x4_t p2) { |
| + uint16x8_t tmp; |
| + |
| + tmp = vcombine_u16(vmovn_u32(vreinterpretq_u32_s32(p1)), |
| + vmovn_u32(vreinterpretq_u32_s32(p2))); |
| + |
| + tmp += vdupq_n_u16(128); |
| + tmp += vshrq_n_u16(tmp, 8); |
| + |
| + return vshrn_n_u16(tmp, 8); |
| +} |
| + |
| +static inline uint16x8_t SkDiv255Round_neon8_16_16(uint16x8_t prod) { |
| + prod += vdupq_n_u16(128); |
| + prod += vshrq_n_u16(prod, 8); |
| + |
| + return vshrq_n_u16(prod, 8); |
| +} |
| + |
| +static inline uint8x8_t clamp_div255round_simd8_32(int32x4_t val1, int32x4_t val2) { |
| + uint8x8_t ret; |
| + uint32x4_t cmp1, cmp2; |
| + uint16x8_t cmp16; |
| + uint8x8_t cmp8, cmp8_1; |
| + |
| + // Test if <= 0 |
| + cmp1 = vcleq_s32(val1, vdupq_n_s32(0)); |
| + cmp2 = vcleq_s32(val2, vdupq_n_s32(0)); |
| + cmp16 = vcombine_u16(vmovn_u32(cmp1), vmovn_u32(cmp2)); |
| + cmp8_1 = vmovn_u16(cmp16); |
| + |
| + // Init to zero |
| + ret = vdup_n_u8(0); |
| + |
| + // Test if >= 255*255 |
| + cmp1 = vcgeq_s32(val1, vdupq_n_s32(255*255)); |
| + cmp2 = vcgeq_s32(val2, vdupq_n_s32(255*255)); |
| + cmp16 = vcombine_u16(vmovn_u32(cmp1), vmovn_u32(cmp2)); |
| + cmp8 = vmovn_u16(cmp16); |
| + |
| + // Insert 255 where true |
| + ret = vbsl_u8(cmp8, vdup_n_u8(255), ret); |
| + |
| + // Calc SkDiv255Round |
| + uint8x8_t div = SkDiv255Round_neon8_32_8(val1, val2); |
| + |
| + // Insert where false and previous test false |
| + cmp8 = cmp8 | cmp8_1; |
| + ret = vbsl_u8(cmp8, ret, div); |
| + |
| + // Return the final combination |
| + return ret; |
| +} |
| + |
| +//////////////////////////////////////////////////////////////////////////////// |
| +// 8 pixels modeprocs |
| +//////////////////////////////////////////////////////////////////////////////// |
| + |
| +uint8x8x4_t dstover_modeproc_neon8(uint8x8x4_t src, uint8x8x4_t dst) { |
| + uint8x8x4_t ret; |
| + uint16x8_t src_scale; |
| + |
| + src_scale = vsubw_u8(vdupq_n_u16(256), dst.val[NEON_A]); |
| + |
| + ret.val[NEON_A] = dst.val[NEON_A] + SkAlphaMul_neon8(src.val[NEON_A], src_scale); |
| + ret.val[NEON_R] = dst.val[NEON_R] + SkAlphaMul_neon8(src.val[NEON_R], src_scale); |
| + ret.val[NEON_G] = dst.val[NEON_G] + SkAlphaMul_neon8(src.val[NEON_G], src_scale); |
| + ret.val[NEON_B] = dst.val[NEON_B] + SkAlphaMul_neon8(src.val[NEON_B], src_scale); |
| + |
| + return ret; |
| +} |
| + |
| +uint8x8x4_t srcin_modeproc_neon8(uint8x8x4_t src, uint8x8x4_t dst) { |
| + uint8x8x4_t ret; |
| + uint16x8_t scale; |
| + |
| + scale = SkAlpha255To256_neon8(dst.val[NEON_A]); |
| + |
| + ret.val[NEON_A] = SkAlphaMul_neon8(src.val[NEON_A], scale); |
| + ret.val[NEON_R] = SkAlphaMul_neon8(src.val[NEON_R], scale); |
| + ret.val[NEON_G] = SkAlphaMul_neon8(src.val[NEON_G], scale); |
| + ret.val[NEON_B] = SkAlphaMul_neon8(src.val[NEON_B], scale); |
| + |
| + return ret; |
| +} |
| + |
| +uint8x8x4_t dstin_modeproc_neon8(uint8x8x4_t src, uint8x8x4_t dst) { |
| + uint8x8x4_t ret; |
| + uint16x8_t scale; |
| + |
| + scale = SkAlpha255To256_neon8(src.val[NEON_A]); |
| + |
| + ret = SkAlphaMulQ_neon8(dst, scale); |
| + |
| + return ret; |
| +} |
| + |
| +uint8x8x4_t srcout_modeproc_neon8(uint8x8x4_t src, uint8x8x4_t dst) { |
| + uint8x8x4_t ret; |
| + uint16x8_t scale = vsubw_u8(vdupq_n_u16(256), dst.val[NEON_A]); |
| + |
| + ret = SkAlphaMulQ_neon8(src, scale); |
| + |
| + return ret; |
| +} |
| + |
| +uint8x8x4_t dstout_modeproc_neon8(uint8x8x4_t src, uint8x8x4_t dst) { |
| + uint8x8x4_t ret; |
| + uint16x8_t scale = vsubw_u8(vdupq_n_u16(256), src.val[NEON_A]); |
| + |
| + ret = SkAlphaMulQ_neon8(dst, scale); |
| + |
| + return ret; |
| +} |
| + |
| +uint8x8x4_t srcatop_modeproc_neon8(uint8x8x4_t src, uint8x8x4_t dst) { |
| + uint8x8x4_t ret; |
| + uint8x8_t isa; |
| + |
| + isa = vsub_u8(vdup_n_u8(255), src.val[NEON_A]); |
| + |
| + ret.val[NEON_A] = dst.val[NEON_A]; |
| + ret.val[NEON_R] = SkAlphaMulAlpha_neon8(src.val[NEON_R], dst.val[NEON_A]) |
| + + SkAlphaMulAlpha_neon8(dst.val[NEON_R], isa); |
| + ret.val[NEON_G] = SkAlphaMulAlpha_neon8(src.val[NEON_G], dst.val[NEON_A]) |
| + + SkAlphaMulAlpha_neon8(dst.val[NEON_G], isa); |
| + ret.val[NEON_B] = SkAlphaMulAlpha_neon8(src.val[NEON_B], dst.val[NEON_A]) |
| + + SkAlphaMulAlpha_neon8(dst.val[NEON_B], isa); |
| + |
| + return ret; |
| +} |
| + |
| +uint8x8x4_t dstatop_modeproc_neon8(uint8x8x4_t src, uint8x8x4_t dst) { |
| + uint8x8x4_t ret; |
| + uint8x8_t ida; |
| + |
| + ida = vsub_u8(vdup_n_u8(255), dst.val[NEON_A]); |
| + |
| + ret.val[NEON_A] = src.val[NEON_A]; |
| + ret.val[NEON_R] = SkAlphaMulAlpha_neon8(src.val[NEON_R], ida) |
| + + SkAlphaMulAlpha_neon8(dst.val[NEON_R], src.val[NEON_A]); |
| + ret.val[NEON_G] = SkAlphaMulAlpha_neon8(src.val[NEON_G], ida) |
| + + SkAlphaMulAlpha_neon8(dst.val[NEON_G], src.val[NEON_A]); |
| + ret.val[NEON_B] = SkAlphaMulAlpha_neon8(src.val[NEON_B], ida) |
| + + SkAlphaMulAlpha_neon8(dst.val[NEON_B], src.val[NEON_A]); |
| + |
| + return ret; |
| +} |
| + |
| +uint8x8x4_t xor_modeproc_neon8(uint8x8x4_t src, uint8x8x4_t dst) { |
| + uint8x8x4_t ret; |
| + uint8x8_t isa, ida; |
| + uint16x8_t tmp_wide, tmp_wide2; |
| + |
| + isa = vsub_u8(vdup_n_u8(255), src.val[NEON_A]); |
| + ida = vsub_u8(vdup_n_u8(255), dst.val[NEON_A]); |
| + |
| + // First calc alpha |
| + tmp_wide = vmovl_u8(src.val[NEON_A]); |
| + tmp_wide = vaddw_u8(tmp_wide, dst.val[NEON_A]); |
| + tmp_wide2 = vshll_n_u8(SkAlphaMulAlpha_neon8(src.val[NEON_A], dst.val[NEON_A]), 1); |
| + tmp_wide = vsubq_u16(tmp_wide, tmp_wide2); |
| + ret.val[NEON_A] = vmovn_u16(tmp_wide); |
| + |
| + // Then colors |
| + ret.val[NEON_R] = SkAlphaMulAlpha_neon8(src.val[NEON_R], ida) |
| + + SkAlphaMulAlpha_neon8(dst.val[NEON_R], isa); |
| + ret.val[NEON_G] = SkAlphaMulAlpha_neon8(src.val[NEON_G], ida) |
| + + SkAlphaMulAlpha_neon8(dst.val[NEON_G], isa); |
| + ret.val[NEON_B] = SkAlphaMulAlpha_neon8(src.val[NEON_B], ida) |
| + + SkAlphaMulAlpha_neon8(dst.val[NEON_B], isa); |
| + |
| + return ret; |
| +} |
| + |
| +uint8x8x4_t plus_modeproc_neon8(uint8x8x4_t src, uint8x8x4_t dst) { |
| + uint8x8x4_t ret; |
| + |
| + ret.val[NEON_A] = vqadd_u8(src.val[NEON_A], dst.val[NEON_A]); |
| + ret.val[NEON_R] = vqadd_u8(src.val[NEON_R], dst.val[NEON_R]); |
| + ret.val[NEON_G] = vqadd_u8(src.val[NEON_G], dst.val[NEON_G]); |
| + ret.val[NEON_B] = vqadd_u8(src.val[NEON_B], dst.val[NEON_B]); |
| + |
| + return ret; |
| +} |
| + |
| +uint8x8x4_t modulate_modeproc_neon8(uint8x8x4_t src, uint8x8x4_t dst) { |
| + uint8x8x4_t ret; |
| + |
| + ret.val[NEON_A] = SkAlphaMulAlpha_neon8(src.val[NEON_A], dst.val[NEON_A]); |
| + ret.val[NEON_R] = SkAlphaMulAlpha_neon8(src.val[NEON_R], dst.val[NEON_R]); |
| + ret.val[NEON_G] = SkAlphaMulAlpha_neon8(src.val[NEON_G], dst.val[NEON_G]); |
| + ret.val[NEON_B] = SkAlphaMulAlpha_neon8(src.val[NEON_B], dst.val[NEON_B]); |
| + |
| + return ret; |
| +} |
| + |
| +static inline uint8x8_t srcover_color(uint8x8_t a, uint8x8_t b) { |
| + uint16x8_t tmp; |
| + |
| + tmp = vaddl_u8(a, b); |
| + tmp -= SkAlphaMulAlpha_neon8_16(a, b); |
| + |
| + return vmovn_u16(tmp); |
| +} |
| + |
| +uint8x8x4_t screen_modeproc_neon8(uint8x8x4_t src, uint8x8x4_t dst) { |
| + uint8x8x4_t ret; |
| + |
| + ret.val[NEON_A] = srcover_color(src.val[NEON_A], dst.val[NEON_A]); |
| + ret.val[NEON_R] = srcover_color(src.val[NEON_R], dst.val[NEON_R]); |
| + ret.val[NEON_G] = srcover_color(src.val[NEON_G], dst.val[NEON_G]); |
| + ret.val[NEON_B] = srcover_color(src.val[NEON_B], dst.val[NEON_B]); |
| + |
| + return ret; |
| +} |
| + |
| +template <bool overlay> |
| +static inline uint8x8_t overlay_hardlight_color(uint8x8_t sc, uint8x8_t dc, |
| + uint8x8_t sa, uint8x8_t da) { |
| + /* |
| + * In the end we're gonna use (rc + tmp) with a different rc |
| + * coming from an alternative. |
| + * The whole value (rc + tmp) can always be expressed as |
| + * VAL = COM - SUB in the if case |
| + * VAL = COM + SUB - sa*da in the else case |
| + * |
| + * with COM = 255 * (sc + dc) |
| + * and SUB = sc*da + dc*sa - 2*dc*sc |
| + */ |
| + |
| + // Prepare common subexpressions |
| + uint16x8_t const255 = vdupq_n_u16(255); |
| + uint16x8_t sc_plus_dc = vaddl_u8(sc, dc); |
| + uint16x8_t scda = vmull_u8(sc, da); |
| + uint16x8_t dcsa = vmull_u8(dc, sa); |
| + uint16x8_t sada = vmull_u8(sa, da); |
| + |
| + // Prepare non common subexpressions |
| + uint16x8_t dc2, sc2; |
| + uint32x4_t scdc2_1, scdc2_2; |
| + if (overlay) { |
| + dc2 = vshll_n_u8(dc, 1); |
| + scdc2_1 = vmull_u16(vget_low_u16(dc2), vget_low_u16(vmovl_u8(sc))); |
| + scdc2_2 = vmull_u16(vget_high_u16(dc2), vget_high_u16(vmovl_u8(sc))); |
| + } else { |
| + sc2 = vshll_n_u8(sc, 1); |
| + scdc2_1 = vmull_u16(vget_low_u16(sc2), vget_low_u16(vmovl_u8(dc))); |
| + scdc2_2 = vmull_u16(vget_high_u16(sc2), vget_high_u16(vmovl_u8(dc))); |
| + } |
| + |
| + // Calc COM |
| + int32x4_t com1, com2; |
| + com1 = vreinterpretq_s32_u32( |
| + vmull_u16(vget_low_u16(const255), vget_low_u16(sc_plus_dc))); |
| + com2 = vreinterpretq_s32_u32( |
| + vmull_u16(vget_high_u16(const255), vget_high_u16(sc_plus_dc))); |
| + |
| + // Calc SUB |
| + int32x4_t sub1, sub2; |
| + sub1 = vreinterpretq_s32_u32(vaddl_u16(vget_low_u16(scda), vget_low_u16(dcsa))); |
| + sub2 = vreinterpretq_s32_u32(vaddl_u16(vget_high_u16(scda), vget_high_u16(dcsa))); |
| + sub1 = vsubq_s32(sub1, vreinterpretq_s32_u32(scdc2_1)); |
| + sub2 = vsubq_s32(sub2, vreinterpretq_s32_u32(scdc2_2)); |
| + |
| + // Compare 2*dc <= da |
| + uint16x8_t cmp; |
| + |
| + if (overlay) { |
| + cmp = vcleq_u16(dc2, vmovl_u8(da)); |
| + } else { |
| + cmp = vcleq_u16(sc2, vmovl_u8(sa)); |
| + } |
| + |
| + // Prepare variables |
| + int32x4_t val1_1, val1_2; |
| + int32x4_t val2_1, val2_2; |
| + uint32x4_t cmp1, cmp2; |
| + |
| + cmp1 = vmovl_u16(vget_low_u16(cmp)); |
| + cmp1 |= vshlq_n_u32(cmp1, 16); |
| + cmp2 = vmovl_u16(vget_high_u16(cmp)); |
| + cmp2 |= vshlq_n_u32(cmp2, 16); |
| + |
| + // Calc COM - SUB |
| + val1_1 = com1 - sub1; |
| + val1_2 = com2 - sub2; |
| + |
| + // Calc COM + SUB - sa*da |
| + val2_1 = com1 + sub1; |
| + val2_2 = com2 + sub2; |
| + |
| + val2_1 = vsubq_s32(val2_1, vreinterpretq_s32_u32(vmovl_u16(vget_low_u16(sada)))); |
| + val2_2 = vsubq_s32(val2_2, vreinterpretq_s32_u32(vmovl_u16(vget_high_u16(sada)))); |
| + |
| + // Insert where needed |
| + val1_1 = vbslq_s32(cmp1, val1_1, val2_1); |
| + val1_2 = vbslq_s32(cmp2, val1_2, val2_2); |
| + |
| + // Call the clamp_div255round function |
| + return clamp_div255round_simd8_32(val1_1, val1_2); |
| +} |
| + |
| +static inline uint8x8_t overlay_color(uint8x8_t sc, uint8x8_t dc, |
| + uint8x8_t sa, uint8x8_t da) { |
| + return overlay_hardlight_color<true>(sc, dc, sa, da); |
| +} |
| + |
| +uint8x8x4_t overlay_modeproc_neon8(uint8x8x4_t src, uint8x8x4_t dst) { |
| + uint8x8x4_t ret; |
| + |
| + ret.val[NEON_A] = srcover_color(src.val[NEON_A], dst.val[NEON_A]); |
| + ret.val[NEON_R] = overlay_color(src.val[NEON_R], dst.val[NEON_R], |
| + src.val[NEON_A], dst.val[NEON_A]); |
| + ret.val[NEON_G] = overlay_color(src.val[NEON_G], dst.val[NEON_G], |
| + src.val[NEON_A], dst.val[NEON_A]); |
| + ret.val[NEON_B] = overlay_color(src.val[NEON_B], dst.val[NEON_B], |
| + src.val[NEON_A], dst.val[NEON_A]); |
| + |
| + return ret; |
| +} |
| + |
| +template <bool lighten> |
| +static inline uint8x8_t lighten_darken_color(uint8x8_t sc, uint8x8_t dc, |
| + uint8x8_t sa, uint8x8_t da) { |
| + uint16x8_t sd, ds, cmp, tmp, tmp2; |
| + |
| + // Prepare |
| + sd = vmull_u8(sc, da); |
| + ds = vmull_u8(dc, sa); |
| + |
| + // Do test |
| + if (lighten) { |
| + cmp = vcgtq_u16(sd, ds); |
| + } else { |
| + cmp = vcltq_u16(sd, ds); |
| + } |
| + |
| + // Assign if |
| + tmp = vaddl_u8(sc, dc); |
| + tmp2 = tmp; |
| + tmp -= SkDiv255Round_neon8_16_16(ds); |
| + |
| + // Calc else |
| + tmp2 -= SkDiv255Round_neon8_16_16(sd); |
| + |
| + // Insert where needed |
| + tmp = vbslq_u16(cmp, tmp, tmp2); |
| + |
| + return vmovn_u16(tmp); |
| +} |
| + |
| +static inline uint8x8_t darken_color(uint8x8_t sc, uint8x8_t dc, |
| + uint8x8_t sa, uint8x8_t da) { |
| + return lighten_darken_color<false>(sc, dc, sa, da); |
| +} |
| + |
| +uint8x8x4_t darken_modeproc_neon8(uint8x8x4_t src, uint8x8x4_t dst) { |
| + uint8x8x4_t ret; |
| + |
| + ret.val[NEON_A] = srcover_color(src.val[NEON_A], dst.val[NEON_A]); |
| + ret.val[NEON_R] = darken_color(src.val[NEON_R], dst.val[NEON_R], |
| + src.val[NEON_A], dst.val[NEON_A]); |
| + ret.val[NEON_G] = darken_color(src.val[NEON_G], dst.val[NEON_G], |
| + src.val[NEON_A], dst.val[NEON_A]); |
| + ret.val[NEON_B] = darken_color(src.val[NEON_B], dst.val[NEON_B], |
| + src.val[NEON_A], dst.val[NEON_A]); |
| + |
| + return ret; |
| +} |
| + |
| +static inline uint8x8_t lighten_color(uint8x8_t sc, uint8x8_t dc, |
| + uint8x8_t sa, uint8x8_t da) { |
| + return lighten_darken_color<true>(sc, dc, sa, da); |
| +} |
| + |
| +uint8x8x4_t lighten_modeproc_neon8(uint8x8x4_t src, uint8x8x4_t dst) { |
| + uint8x8x4_t ret; |
| + |
| + ret.val[NEON_A] = srcover_color(src.val[NEON_A], dst.val[NEON_A]); |
| + ret.val[NEON_R] = lighten_color(src.val[NEON_R], dst.val[NEON_R], |
| + src.val[NEON_A], dst.val[NEON_A]); |
| + ret.val[NEON_G] = lighten_color(src.val[NEON_G], dst.val[NEON_G], |
| + src.val[NEON_A], dst.val[NEON_A]); |
| + ret.val[NEON_B] = lighten_color(src.val[NEON_B], dst.val[NEON_B], |
| + src.val[NEON_A], dst.val[NEON_A]); |
| + |
| + return ret; |
| +} |
| + |
| +static inline uint8x8_t hardlight_color(uint8x8_t sc, uint8x8_t dc, |
| + uint8x8_t sa, uint8x8_t da) { |
| + return overlay_hardlight_color<false>(sc, dc, sa, da); |
| +} |
| + |
| +uint8x8x4_t hardlight_modeproc_neon8(uint8x8x4_t src, uint8x8x4_t dst) { |
| + uint8x8x4_t ret; |
| + |
| + ret.val[NEON_A] = srcover_color(src.val[NEON_A], dst.val[NEON_A]); |
| + ret.val[NEON_R] = hardlight_color(src.val[NEON_R], dst.val[NEON_R], |
| + src.val[NEON_A], dst.val[NEON_A]); |
| + ret.val[NEON_G] = hardlight_color(src.val[NEON_G], dst.val[NEON_G], |
| + src.val[NEON_A], dst.val[NEON_A]); |
| + ret.val[NEON_B] = hardlight_color(src.val[NEON_B], dst.val[NEON_B], |
| + src.val[NEON_A], dst.val[NEON_A]); |
| + |
| + return ret; |
| +} |
| + |
| +static inline uint8x8_t difference_color(uint8x8_t sc, uint8x8_t dc, |
| + uint8x8_t sa, uint8x8_t da) { |
| + uint16x8_t sd, ds, tmp; |
| + int16x8_t val; |
| + |
| + sd = vmull_u8(sc, da); |
| + ds = vmull_u8(dc, sa); |
| + |
| + tmp = vminq_u16(sd, ds); |
| + tmp = SkDiv255Round_neon8_16_16(tmp); |
| + tmp = vshlq_n_u16(tmp, 1); |
| + |
| + val = vreinterpretq_s16_u16(vaddl_u8(sc, dc)); |
| + |
| + val -= vreinterpretq_s16_u16(tmp); |
| + |
| + val = vmaxq_s16(val, vdupq_n_s16(0)); |
| + val = vminq_s16(val, vdupq_n_s16(255)); |
| + |
| + return vmovn_u16(vreinterpretq_u16_s16(val)); |
| +} |
| + |
| +uint8x8x4_t difference_modeproc_neon8(uint8x8x4_t src, uint8x8x4_t dst) { |
| + uint8x8x4_t ret; |
| + |
| + ret.val[NEON_A] = srcover_color(src.val[NEON_A], dst.val[NEON_A]); |
| + ret.val[NEON_R] = difference_color(src.val[NEON_R], dst.val[NEON_R], |
| + src.val[NEON_A], dst.val[NEON_A]); |
| + ret.val[NEON_G] = difference_color(src.val[NEON_G], dst.val[NEON_G], |
| + src.val[NEON_A], dst.val[NEON_A]); |
| + ret.val[NEON_B] = difference_color(src.val[NEON_B], dst.val[NEON_B], |
| + src.val[NEON_A], dst.val[NEON_A]); |
| + |
| + return ret; |
| +} |
| + |
| +static inline uint8x8_t exclusion_color(uint8x8_t sc, uint8x8_t dc, |
| + uint8x8_t sa, uint8x8_t da) { |
| + /* The equation can be simplified to 255(sc + dc) - 2 * sc * dc */ |
| + |
| + uint16x8_t sc_plus_dc, scdc, const255; |
| + int32x4_t term1_1, term1_2, term2_1, term2_2; |
| + |
| + /* Calc (sc + dc) and (sc * dc) */ |
| + sc_plus_dc = vaddl_u8(sc, dc); |
| + scdc = vmull_u8(sc, dc); |
| + |
| + /* Prepare constants */ |
| + const255 = vdupq_n_u16(255); |
| + |
| + /* Calc the first term */ |
| + term1_1 = vreinterpretq_s32_u32( |
| + vmull_u16(vget_low_u16(const255), vget_low_u16(sc_plus_dc))); |
| + term1_2 = vreinterpretq_s32_u32( |
| + vmull_u16(vget_high_u16(const255), vget_high_u16(sc_plus_dc))); |
| + |
| + /* Calc the second term */ |
| + term2_1 = vreinterpretq_s32_u32(vshll_n_u16(vget_low_u16(scdc), 1)); |
| + term2_2 = vreinterpretq_s32_u32(vshll_n_u16(vget_high_u16(scdc), 1)); |
| + |
| + return clamp_div255round_simd8_32(term1_1 - term2_1, term1_2 - term2_2); |
| +} |
| + |
| +uint8x8x4_t exclusion_modeproc_neon8(uint8x8x4_t src, uint8x8x4_t dst) { |
| + uint8x8x4_t ret; |
| + |
| + ret.val[NEON_A] = srcover_color(src.val[NEON_A], dst.val[NEON_A]); |
| + ret.val[NEON_R] = exclusion_color(src.val[NEON_R], dst.val[NEON_R], |
| + src.val[NEON_A], dst.val[NEON_A]); |
| + ret.val[NEON_G] = exclusion_color(src.val[NEON_G], dst.val[NEON_G], |
| + src.val[NEON_A], dst.val[NEON_A]); |
| + ret.val[NEON_B] = exclusion_color(src.val[NEON_B], dst.val[NEON_B], |
| + src.val[NEON_A], dst.val[NEON_A]); |
| + |
| + return ret; |
| +} |
| + |
| +static inline uint8x8_t blendfunc_multiply_color(uint8x8_t sc, uint8x8_t dc, |
| + uint8x8_t sa, uint8x8_t da) { |
| + uint32x4_t val1, val2; |
| + uint16x8_t scdc, t1, t2; |
| + |
| + t1 = vmull_u8(sc, vdup_n_u8(255) - da); |
| + t2 = vmull_u8(dc, vdup_n_u8(255) - sa); |
| + scdc = vmull_u8(sc, dc); |
| + |
| + val1 = vaddl_u16(vget_low_u16(t1), vget_low_u16(t2)); |
| + val2 = vaddl_u16(vget_high_u16(t1), vget_high_u16(t2)); |
| + |
| + val1 = vaddw_u16(val1, vget_low_u16(scdc)); |
| + val2 = vaddw_u16(val2, vget_high_u16(scdc)); |
| + |
| + return clamp_div255round_simd8_32( |
| + vreinterpretq_s32_u32(val1), vreinterpretq_s32_u32(val2)); |
| +} |
| + |
| +uint8x8x4_t multiply_modeproc_neon8(uint8x8x4_t src, uint8x8x4_t dst) { |
| + uint8x8x4_t ret; |
| + |
| + ret.val[NEON_A] = srcover_color(src.val[NEON_A], dst.val[NEON_A]); |
| + ret.val[NEON_R] = blendfunc_multiply_color(src.val[NEON_R], dst.val[NEON_R], |
| + src.val[NEON_A], dst.val[NEON_A]); |
| + ret.val[NEON_G] = blendfunc_multiply_color(src.val[NEON_G], dst.val[NEON_G], |
| + src.val[NEON_A], dst.val[NEON_A]); |
| + ret.val[NEON_B] = blendfunc_multiply_color(src.val[NEON_B], dst.val[NEON_B], |
| + src.val[NEON_A], dst.val[NEON_A]); |
| + |
| + return ret; |
| +} |
| + |
| +//////////////////////////////////////////////////////////////////////////////// |
| + |
| +typedef uint8x8x4_t (*SkXfermodeProcSIMD)(uint8x8x4_t src, uint8x8x4_t dst); |
| + |
| +extern SkXfermodeProcSIMD gNEONXfermodeProcs[]; |
| + |
| +SkNEONProcCoeffXfermode::SkNEONProcCoeffXfermode(SkFlattenableReadBuffer& buffer) |
| + : INHERITED(buffer) { |
| + fProcSIMD = reinterpret_cast<void*>(gNEONXfermodeProcs[this->getMode()]); |
| +} |
| + |
| +void SkNEONProcCoeffXfermode::xfer32(SkPMColor dst[], const SkPMColor src[], |
| + int count, const SkAlpha aa[]) const { |
| + SkASSERT(dst && src && count >= 0); |
| + |
| + SkXfermodeProc proc = this->getProc(); |
| + SkXfermodeProcSIMD procSIMD = reinterpret_cast<SkXfermodeProcSIMD>(fProcSIMD); |
| + |
| + if (NULL == aa) { |
| + // Unrolled NEON code |
| + while (count >= 8) { |
| + uint8x8x4_t vsrc, vdst, vres; |
| + |
| +#if (__GNUC__ == 4) && (__GNUC_MINOR__ > 6) |
|
djsollen
2013/10/17 18:34:18
what happens to this code when we go to version 5.
|
| + asm volatile ( |
| + "vld4.u8 %h[vsrc], [%[src]]! \t\n" |
| + "vld4.u8 %h[vdst], [%[dst]] \t\n" |
| + : [vsrc] "=w" (vsrc), [vdst] "=w" (vdst), [src] "+&r" (src) |
| + : [dst] "r" (dst) |
| + : |
| + ); |
| +#else |
| + register uint8x8_t d0 asm("d0"); |
| + register uint8x8_t d1 asm("d1"); |
| + register uint8x8_t d2 asm("d2"); |
| + register uint8x8_t d3 asm("d3"); |
| + register uint8x8_t d4 asm("d4"); |
| + register uint8x8_t d5 asm("d5"); |
| + register uint8x8_t d6 asm("d6"); |
| + register uint8x8_t d7 asm("d7"); |
| + |
| + asm volatile ( |
| + "vld4.u8 {d0-d3},[%[src]]!;" |
| + "vld4.u8 {d4-d7},[%[dst]];" |
| + : "=w" (d0), "=w" (d1), "=w" (d2), "=w" (d3), |
| + "=w" (d4), "=w" (d5), "=w" (d6), "=w" (d7), |
| + [src] "+&r" (src) |
| + : [dst] "r" (dst) |
| + : |
| + ); |
| + vsrc.val[0] = d0; vdst.val[0] = d4; |
| + vsrc.val[1] = d1; vdst.val[1] = d5; |
| + vsrc.val[2] = d2; vdst.val[2] = d6; |
| + vsrc.val[3] = d3; vdst.val[3] = d7; |
| +#endif |
| + |
| + vres = procSIMD(vsrc, vdst); |
| + |
| + vst4_u8((uint8_t*)dst, vres); |
| + |
| + count -= 8; |
| + dst += 8; |
| + } |
| + // Leftovers |
| + for (int i = 0; i < count; i++) { |
| + dst[i] = proc(src[i], dst[i]); |
| + } |
| + } else { |
| + for (int i = count - 1; i >= 0; --i) { |
| + unsigned a = aa[i]; |
| + if (0 != a) { |
| + SkPMColor dstC = dst[i]; |
| + SkPMColor C = proc(src[i], dstC); |
| + if (a != 0xFF) { |
| + C = SkFourByteInterp(C, dstC, a); |
| + } |
| + dst[i] = C; |
| + } |
| + } |
| + } |
| +} |
| + |
| +#ifdef SK_DEVELOPER |
| +void SkNEONProcCoeffXfermode::toString(SkString* str) const { |
| + this->INHERITED::toString(str); |
| +} |
| +#endif |
| + |
| +//////////////////////////////////////////////////////////////////////////////// |
| + |
| +SkXfermodeProcSIMD gNEONXfermodeProcs[] = { |
| + NULL, // kClear_Mode |
| + NULL, // kSrc_Mode |
| + NULL, // kDst_Mode |
| + NULL, // kSrcOver_Mode |
| + dstover_modeproc_neon8, |
| + srcin_modeproc_neon8, |
| + dstin_modeproc_neon8, |
| + srcout_modeproc_neon8, |
| + dstout_modeproc_neon8, |
| + srcatop_modeproc_neon8, |
| + dstatop_modeproc_neon8, |
| + xor_modeproc_neon8, |
| + plus_modeproc_neon8, |
| + modulate_modeproc_neon8, |
| + screen_modeproc_neon8, |
| + |
| + overlay_modeproc_neon8, |
| + darken_modeproc_neon8, |
| + lighten_modeproc_neon8, |
| + NULL, // kColorDodge_Mode |
| + NULL, // kColorBurn_Mode |
| + hardlight_modeproc_neon8, |
| + NULL, // kSoftLight_Mode |
| + difference_modeproc_neon8, |
| + exclusion_modeproc_neon8, |
| + multiply_modeproc_neon8, |
| + |
| + NULL, // kHue_Mode |
| + NULL, // kSaturation_Mode |
| + NULL, // kColor_Mode |
| + NULL, // kLuminosity_Mode |
| +}; |
| + |
| +SK_COMPILE_ASSERT( |
| + SK_ARRAY_COUNT(gNEONXfermodeProcs) == SkXfermode::kLastMode + 1, |
| + mode_count_arm |
| +); |
| + |
| +SkProcCoeffXfermode* SkPlatformXfermodeFactory_impl_neon(const ProcCoeff& rec, |
| + SkXfermode::Mode mode) { |
| + |
| + void* procSIMD = reinterpret_cast<void*>(gNEONXfermodeProcs[mode]); |
| + |
| + if (procSIMD != NULL) { |
| + return SkNEW_ARGS(SkNEONProcCoeffXfermode, (rec, mode, procSIMD)); |
| + } |
| + return NULL; |
| +} |