Index: src/core/Sk4pxXfermode.h |
diff --git a/src/core/Sk4pxXfermode.h b/src/core/Sk4pxXfermode.h |
index e8610edaba55ed1da08daac55c34add64a613cd5..b4ebd850e3bfee445521be2f312260d3d21dfe94 100644 |
--- a/src/core/Sk4pxXfermode.h |
+++ b/src/core/Sk4pxXfermode.h |
@@ -21,53 +21,41 @@ namespace { |
}; \ |
inline Sk4px Name::Xfer(const Sk4px& s, const Sk4px& d) |
-XFERMODE(Clear) { return Sk4px((SkPMColor)0); } |
+XFERMODE(Clear) { return Sk4px::DupPMColor(0); } |
XFERMODE(Src) { return s; } |
XFERMODE(Dst) { return d; } |
-XFERMODE(SrcIn) { return s.fastMulDiv255Round(d.alphas() ); } |
-XFERMODE(SrcOut) { return s.fastMulDiv255Round(d.alphas().inv()); } |
-XFERMODE(SrcOver) { return s + d.fastMulDiv255Round(s.alphas().inv()); } |
+XFERMODE(SrcIn) { return s.approxMulDiv255(d.alphas() ); } |
+XFERMODE(SrcOut) { return s.approxMulDiv255(d.alphas().inv()); } |
+XFERMODE(SrcOver) { return s + d.approxMulDiv255(s.alphas().inv()); } |
XFERMODE(DstIn) { return SrcIn ::Xfer(d,s); } |
XFERMODE(DstOut) { return SrcOut ::Xfer(d,s); } |
XFERMODE(DstOver) { return SrcOver::Xfer(d,s); } |
// [ S * Da + (1 - Sa) * D] |
-XFERMODE(SrcATop) { |
- return Sk4px::Wide(s.mulWiden(d.alphas()) + d.mulWiden(s.alphas().inv())) |
- .div255RoundNarrow(); |
-} |
+XFERMODE(SrcATop) { return (s * d.alphas() + d * s.alphas().inv()).div255(); } |
XFERMODE(DstATop) { return SrcATop::Xfer(d,s); } |
//[ S * (1 - Da) + (1 - Sa) * D ] |
-XFERMODE(Xor) { |
- return Sk4px::Wide(s.mulWiden(d.alphas().inv()) + d.mulWiden(s.alphas().inv())) |
- .div255RoundNarrow(); |
-} |
+XFERMODE(Xor) { return (s * d.alphas().inv() + d * s.alphas().inv()).div255(); } |
// [S + D ] |
XFERMODE(Plus) { return s.saturatedAdd(d); } |
// [S * D ] |
-XFERMODE(Modulate) { return s.fastMulDiv255Round(d); } |
+XFERMODE(Modulate) { return s.approxMulDiv255(d); } |
// [S + D - S * D] |
XFERMODE(Screen) { |
// Doing the math as S + (1-S)*D or S + (D - S*D) means the add and subtract can be done |
// in 8-bit space without overflow. S + (1-S)*D is a touch faster because inv() is cheap. |
- return s + d.fastMulDiv255Round(s.inv()); |
-} |
-XFERMODE(Multiply) { |
- return Sk4px::Wide(s.mulWiden(d.alphas().inv()) + |
- d.mulWiden(s.alphas().inv()) + |
- s.mulWiden(d)) |
- .div255RoundNarrow(); |
+ return s + d.approxMulDiv255(s.inv()); |
} |
+XFERMODE(Multiply) { return (s * d.alphas().inv() + d * s.alphas().inv() + s*d).div255(); } |
// [ Sa + Da - Sa*Da, Sc + Dc - 2*min(Sc*Da, Dc*Sa) ] (And notice Sa*Da == min(Sa*Da, Da*Sa).) |
XFERMODE(Difference) { |
- auto m = Sk4px::Wide(Sk16h::Min(s.mulWiden(d.alphas()), d.mulWiden(s.alphas()))) |
- .div255RoundNarrow(); |
+ auto m = Sk4px::Wide::Min(s * d.alphas(), d * s.alphas()).div255(); |
// There's no chance of underflow, and if we subtract m before adding s+d, no overflow. |
return (s - m) + (d - m.zeroAlphas()); |
} |
// [ Sa + Da - Sa*Da, Sc + Dc - 2*Sc*Dc ] |
XFERMODE(Exclusion) { |
- auto p = s.fastMulDiv255Round(d); |
+ auto p = s.approxMulDiv255(d); |
// There's no chance of underflow, and if we subtract p before adding src+dst, no overflow. |
return (s - p) + (d - p.zeroAlphas()); |
} |
@@ -77,20 +65,19 @@ XFERMODE(Exclusion) { |
// A reasonable fallback mode for doing AA is to simply apply the transfermode first, |
// then linearly interpolate the AA. |
template <typename Mode> |
-static Sk4px xfer_aa(const Sk4px& s, const Sk4px& d, const Sk16b& aa) { |
- Sk4px noAA = Mode::Xfer(s, d); |
- return Sk4px::Wide(noAA.mulWiden(aa) + d.mulWiden(Sk4px(aa).inv())) |
- .div255RoundNarrow(); |
+static Sk4px xfer_aa(const Sk4px& s, const Sk4px& d, const Sk4px& aa) { |
+ Sk4px bw = Mode::Xfer(s, d); |
+ return (bw * aa + d * aa.inv()).div255(); |
} |
// For some transfermodes we specialize AA, either for correctness or performance. |
#ifndef SK_NO_SPECIALIZED_AA_XFERMODES |
#define XFERMODE_AA(Name) \ |
- template <> Sk4px xfer_aa<Name>(const Sk4px& s, const Sk4px& d, const Sk16b& aa) |
+ template <> Sk4px xfer_aa<Name>(const Sk4px& s, const Sk4px& d, const Sk4px& aa) |
// Plus' clamp needs to happen after AA. skia:3852 |
XFERMODE_AA(Plus) { // [ clamp( (1-AA)D + (AA)(S+D) ) == clamp(D + AA*S) ] |
- return d.saturatedAdd(s.fastMulDiv255Round(aa)); |
+ return d.saturatedAdd(s.approxMulDiv255(aa)); |
} |
#undef XFERMODE_AA |
@@ -110,7 +97,7 @@ public: |
}); |
} else { |
Sk4px::MapDstSrcAlpha(n, dst, src, aa, |
- [&](const Sk4px& dst4, const Sk4px& src4, const Sk16b& alpha) { |
+ [&](const Sk4px& dst4, const Sk4px& src4, const Sk4px& alpha) { |
return xfer_aa<ProcType>(src4, dst4, alpha); |
}); |
} |