Chromium Code Reviews| Index: src/opts/SkBlitMask_opts_arm_neon.cpp |
| diff --git a/src/opts/SkBlitMask_opts_arm_neon.cpp b/src/opts/SkBlitMask_opts_arm_neon.cpp |
| new file mode 100644 |
| index 0000000000000000000000000000000000000000..e2945321bd97a54941302be85987e4b6a774025e |
| --- /dev/null |
| +++ b/src/opts/SkBlitMask_opts_arm_neon.cpp |
| @@ -0,0 +1,279 @@ |
| + |
| +#include "SkBlitMask.h" |
| +#include "SkColor_opts_neon.h" |
| + |
| +static void D32_A8_Black_neon(void* SK_RESTRICT dst, size_t dstRB, |
|
mtklein
2013/11/26 14:44:01
Typically we make static methods all_lowercase_lik
kevin.petit.not.used.account
2013/11/26 16:23:32
I did that for consistency with all the static fun
|
| + const void* SK_RESTRICT maskPtr, size_t maskRB, |
| + SkColor, int width, int height) { |
| + SkPMColor* SK_RESTRICT device = (SkPMColor*)dst; |
| + const uint8_t* SK_RESTRICT mask = (const uint8_t*)maskPtr; |
| + |
| + maskRB -= width; |
| + dstRB -= (width << 2); |
| + do { |
| + int w = width; |
| + while (w >= 8) { |
| + uint8x8_t vmask = vld1_u8(mask); |
| + uint16x8_t vscale = vsubw_u8(vdupq_n_u16(256), vmask); |
| + uint8x8x4_t vdevice = vld4_u8((uint8_t*)device); |
| + |
| + vdevice = SkAlphaMulQ_neon8(vdevice, vscale); |
| + vdevice.val[NEON_A] += vmask; |
| + |
| + vst4_u8((uint8_t*)device, vdevice); |
| + |
| + mask += 8; |
| + device += 8; |
| + w -= 8; |
| + } |
| + while (w-- > 0) { |
| + unsigned aa = *mask++; |
| + *device = (aa << SK_A32_SHIFT) |
| + + SkAlphaMulQ(*device, SkAlpha255To256(255 - aa)); |
| + device += 1; |
| + }; |
| + device = (uint32_t*)((char*)device + dstRB); |
| + mask += maskRB; |
| + } while (--height != 0); |
| +} |
| + |
| +static void D32_A8_Opaque_neon(void* SK_RESTRICT dst, size_t dstRB, |
|
mtklein
2013/11/26 14:44:01
Can you use templates to share the bulk of the cod
kevin.petit.not.used.account
2013/11/26 16:23:32
Done. You're right, that was a bit silly to not ha
|
| + const void* SK_RESTRICT maskPtr, size_t maskRB, |
| + SkColor color, int width, int height) { |
| + SkPMColor pmc = SkPreMultiplyColor(color); |
| + SkPMColor* SK_RESTRICT device = (SkPMColor*)dst; |
| + const uint8_t* SK_RESTRICT mask = (const uint8_t*)maskPtr; |
| + uint8x8x4_t vpmc; |
| + |
| + maskRB -= width; |
| + dstRB -= (width << 2); |
| + if (width >= 8) { |
| + vpmc.val[NEON_A] = vdup_n_u8(SkGetPackedA32(pmc)); |
| + vpmc.val[NEON_R] = vdup_n_u8(SkGetPackedR32(pmc)); |
| + vpmc.val[NEON_G] = vdup_n_u8(SkGetPackedG32(pmc)); |
| + vpmc.val[NEON_B] = vdup_n_u8(SkGetPackedB32(pmc)); |
| + } |
| + do { |
| + int w = width; |
| + while (w >= 8) { |
| + uint8x8_t vmask = vld1_u8(mask); |
| + uint16x8_t vmask256 = SkAlpha255To256_neon8(vmask); |
| + uint16x8_t vscale = vsubw_u8(vdupq_n_u16(256), vmask); |
| + uint8x8x4_t vdev = vld4_u8((uint8_t*)device); |
| + |
| + vdev.val[NEON_A] = SkAlphaMul_neon8(vpmc.val[NEON_A], vmask256) |
| + + SkAlphaMul_neon8(vdev.val[NEON_A], vscale); |
| + vdev.val[NEON_R] = SkAlphaMul_neon8(vpmc.val[NEON_R], vmask256) |
| + + SkAlphaMul_neon8(vdev.val[NEON_R], vscale); |
| + vdev.val[NEON_G] = SkAlphaMul_neon8(vpmc.val[NEON_G], vmask256) |
| + + SkAlphaMul_neon8(vdev.val[NEON_G], vscale); |
| + vdev.val[NEON_B] = SkAlphaMul_neon8(vpmc.val[NEON_B], vmask256) |
| + + SkAlphaMul_neon8(vdev.val[NEON_B], vscale); |
| + |
| + vst4_u8((uint8_t*)device, vdev); |
| + |
| + mask += 8; |
| + device += 8; |
| + w -= 8; |
| + } |
| + while (w-- > 0) { |
| + unsigned aa = *mask++; |
| + *device = SkAlphaMulQ(pmc, SkAlpha255To256(aa)) |
| + + SkAlphaMulQ(*device, SkAlpha255To256(255 - aa)); |
| + device += 1; |
| + }; |
| + device = (uint32_t*)((char*)device + dstRB); |
| + mask += maskRB; |
| + } while (--height != 0); |
| +} |
| + |
| +static void D32_A8_Color_neon(void* SK_RESTRICT dst, size_t dstRB, |
| + const void* SK_RESTRICT maskPtr, size_t maskRB, |
| + SkColor color, int width, int height) { |
| + SkPMColor pmc = SkPreMultiplyColor(color); |
| + size_t dstOffset = dstRB - (width << 2); |
| + size_t maskOffset = maskRB - width; |
| + SkPMColor* SK_RESTRICT device = (SkPMColor *)dst; |
| + const uint8_t* SK_RESTRICT mask = (const uint8_t*)maskPtr; |
| + uint8x8x4_t vpmc; |
| + |
| + if (width >= 8) { |
| + vpmc.val[NEON_A] = vdup_n_u8(SkGetPackedA32(pmc)); |
| + vpmc.val[NEON_R] = vdup_n_u8(SkGetPackedR32(pmc)); |
| + vpmc.val[NEON_G] = vdup_n_u8(SkGetPackedG32(pmc)); |
| + vpmc.val[NEON_B] = vdup_n_u8(SkGetPackedB32(pmc)); |
| + } |
| + do { |
| + int w = width; |
| + while (w >= 8) { |
| + uint8x8_t vmask = vld1_u8(mask); |
| + uint16x8_t vscale, vmask256 = SkAlpha255To256_neon8(vmask); |
| + uint8x8x4_t vdev = vld4_u8((uint8_t*)device); |
| + vscale = vsubw_u8(vdupq_n_u16(256), |
| + SkAlphaMul_neon8(vpmc.val[NEON_A], vmask256)); |
| + |
| + vdev.val[NEON_A] = SkAlphaMul_neon8(vpmc.val[NEON_A], vmask256) |
| + + SkAlphaMul_neon8(vdev.val[NEON_A], vscale); |
| + vdev.val[NEON_R] = SkAlphaMul_neon8(vpmc.val[NEON_R], vmask256) |
| + + SkAlphaMul_neon8(vdev.val[NEON_R], vscale); |
| + vdev.val[NEON_G] = SkAlphaMul_neon8(vpmc.val[NEON_G], vmask256) |
| + + SkAlphaMul_neon8(vdev.val[NEON_G], vscale); |
| + vdev.val[NEON_B] = SkAlphaMul_neon8(vpmc.val[NEON_B], vmask256) |
| + + SkAlphaMul_neon8(vdev.val[NEON_B], vscale); |
| + |
| + vst4_u8((uint8_t*)device, vdev); |
| + |
| + mask += 8; |
| + device += 8; |
| + w -= 8; |
| + } |
| + while (w--) { |
| + unsigned aa = *mask++; |
| + *device = SkBlendARGB32(pmc, *device, aa); |
| + device += 1; |
| + }; |
| + device = (uint32_t*)((char*)device + dstOffset); |
| + mask += maskOffset; |
| + } while (--height != 0); |
| +} |
| + |
| +SkBlitMask::ColorProc D32_A8_Factory_neon(SkColor color) { |
| + if (SK_ColorBLACK == color) { |
| + return D32_A8_Black_neon; |
| + } else if (0xFF == SkColorGetA(color)) { |
| + return D32_A8_Opaque_neon; |
| + } else { |
| + return D32_A8_Color_neon; |
| + } |
| +} |
| + |
| +//////////////////////////////////////////////////////////////////////////////// |
| + |
| +void SkBlitLCD16OpaqueRow_neon(SkPMColor dst[], const uint16_t src[], |
| + SkColor color, int width, |
| + SkPMColor opaqueDst) { |
| + int colR = SkColorGetR(color); |
| + int colG = SkColorGetG(color); |
| + int colB = SkColorGetB(color); |
| + |
| + uint8x8_t vcolR, vcolG, vcolB; |
| + uint8x8_t vopqDstA, vopqDstR, vopqDstG, vopqDstB; |
| + |
| + if (width >= 8) { |
| + vcolR = vdup_n_u8(colR); |
| + vcolG = vdup_n_u8(colG); |
| + vcolB = vdup_n_u8(colB); |
| + vopqDstA = vdup_n_u8(SkGetPackedA32(opaqueDst)); |
| + vopqDstR = vdup_n_u8(SkGetPackedR32(opaqueDst)); |
| + vopqDstG = vdup_n_u8(SkGetPackedG32(opaqueDst)); |
| + vopqDstB = vdup_n_u8(SkGetPackedB32(opaqueDst)); |
| + } |
| + |
| + while (width >= 8) { |
| + uint8x8x4_t vdst; |
| + uint16x8_t vmask; |
| + uint16x8_t vmaskR, vmaskG, vmaskB; |
| + uint8x8_t vsel_trans, vsel_opq; |
| + |
| + vdst = vld4_u8((uint8_t*)dst); |
| + vmask = vld1q_u16(src); |
| + |
| + // Prepare compare masks |
| + vsel_trans = vmovn_u16(vceqq_u16(vmask, vdupq_n_u16(0))); |
| + vsel_opq = vmovn_u16(vceqq_u16(vmask, vdupq_n_u16(0xFFFF))); |
| + |
| + // Get all the color masks on 5 bits |
| + vmaskR = vshrq_n_u16(vmask, SK_R16_SHIFT); |
| + vmaskG = vshrq_n_u16(vshlq_n_u16(vmask, SK_R16_BITS), |
| + SK_B16_BITS + SK_R16_BITS + 1); |
| + vmaskB = vmask & vdupq_n_u16(SK_B16_MASK); |
| + |
| + // Upscale to 0..32 |
| + vmaskR = vmaskR + vshrq_n_u16(vmaskR, 4); |
| + vmaskG = vmaskG + vshrq_n_u16(vmaskG, 4); |
| + vmaskB = vmaskB + vshrq_n_u16(vmaskB, 4); |
| + |
| + vdst.val[NEON_A] = vbsl_u8(vsel_trans, vdst.val[NEON_A], vdup_n_u8(0xFF)); |
| + vdst.val[NEON_A] = vbsl_u8(vsel_opq, vopqDstA, vdst.val[NEON_A]); |
| + |
| + vdst.val[NEON_R] = SkBlend32_neon8(vcolR, vdst.val[NEON_R], vmaskR); |
| + vdst.val[NEON_G] = SkBlend32_neon8(vcolG, vdst.val[NEON_G], vmaskG); |
| + vdst.val[NEON_B] = SkBlend32_neon8(vcolB, vdst.val[NEON_B], vmaskB); |
| + |
| + vdst.val[NEON_R] = vbsl_u8(vsel_opq, vopqDstR, vdst.val[NEON_R]); |
| + vdst.val[NEON_G] = vbsl_u8(vsel_opq, vopqDstG, vdst.val[NEON_G]); |
| + vdst.val[NEON_B] = vbsl_u8(vsel_opq, vopqDstB, vdst.val[NEON_B]); |
| + |
| + vst4_u8((uint8_t*)dst, vdst); |
| + |
| + dst += 8; |
| + src += 8; |
| + width -= 8; |
| + } |
| + |
| + // Leftovers |
| + for (int i = 0; i < width; i++) { |
| + dst[i] = SkBlendLCD16Opaque(colR, colG, colB, dst[i], src[i], |
| + opaqueDst); |
| + } |
| +} |
| + |
| +void SkBlitLCD16Row_neon(SkPMColor dst[], const uint16_t src[], |
| + SkColor color, int width, SkPMColor) { |
| + int colA = SkColorGetA(color); |
| + int colR = SkColorGetR(color); |
| + int colG = SkColorGetG(color); |
| + int colB = SkColorGetB(color); |
| + |
| + colA = SkAlpha255To256(colA); |
| + |
| + uint8x8_t vcolR, vcolG, vcolB; |
| + uint16x8_t vcolA; |
| + |
| + if (width >= 8) { |
| + vcolA = vdupq_n_u16(colA); |
| + vcolR = vdup_n_u8(colR); |
| + vcolG = vdup_n_u8(colG); |
| + vcolB = vdup_n_u8(colB); |
| + } |
| + |
| + while (width >= 8) { |
| + uint8x8x4_t vdst; |
| + uint16x8_t vmask; |
| + uint16x8_t vmaskR, vmaskG, vmaskB; |
| + |
| + vdst = vld4_u8((uint8_t*)dst); |
| + vmask = vld1q_u16(src); |
| + |
| + // Get all the color masks on 5 bits |
| + vmaskR = vshrq_n_u16(vmask, SK_R16_SHIFT); |
| + vmaskG = vshrq_n_u16(vshlq_n_u16(vmask, SK_R16_BITS), |
| + SK_B16_BITS + SK_R16_BITS + 1); |
| + vmaskB = vmask & vdupq_n_u16(SK_B16_MASK); |
| + |
| + // Upscale to 0..32 |
| + vmaskR = vmaskR + vshrq_n_u16(vmaskR, 4); |
| + vmaskG = vmaskG + vshrq_n_u16(vmaskG, 4); |
| + vmaskB = vmaskB + vshrq_n_u16(vmaskB, 4); |
| + |
| + vmaskR = vshrq_n_u16(vmaskR * vcolA, 8); |
| + vmaskG = vshrq_n_u16(vmaskG * vcolA, 8); |
| + vmaskB = vshrq_n_u16(vmaskB * vcolA, 8); |
| + |
| + vdst.val[NEON_A] = vdup_n_u8(0xFF); |
| + vdst.val[NEON_R] = SkBlend32_neon8(vcolR, vdst.val[NEON_R], vmaskR); |
| + vdst.val[NEON_G] = SkBlend32_neon8(vcolG, vdst.val[NEON_G], vmaskG); |
| + vdst.val[NEON_B] = SkBlend32_neon8(vcolB, vdst.val[NEON_B], vmaskB); |
| + |
| + vst4_u8((uint8_t*)dst, vdst); |
| + |
| + dst += 8; |
| + src += 8; |
| + width -= 8; |
| + } |
| + |
| + for (int i = 0; i < width; i++) { |
| + dst[i] = SkBlendLCD16(colA, colR, colG, colB, dst[i], src[i]); |
| + } |
| +} |
| + |