| Index: src/opts/Sk4x_neon.h
|
| diff --git a/src/opts/Sk4x_neon.h b/src/opts/Sk4x_neon.h
|
| index 41f6f4688563c98ef9f164d68d578c1a72f31959..892fcb9d41d4df6c8a664ff9ef90a16483469db0 100644
|
| --- a/src/opts/Sk4x_neon.h
|
| +++ b/src/opts/Sk4x_neon.h
|
| @@ -99,26 +99,7 @@ M(Sk4i) greaterThanEqual(const Sk4f& o) const { return vreinterpretq_s32_u32(vcg
|
| M(Sk4f) Min(const Sk4f& a, const Sk4f& b) { return vminq_f32(a.fVec, b.fVec); }
|
| M(Sk4f) Max(const Sk4f& a, const Sk4f& b) { return vmaxq_f32(a.fVec, b.fVec); }
|
|
|
| -// These shuffle operations are implemented more efficiently with SSE.
|
| -// NEON has efficient zip, unzip, and transpose, but it is more costly to
|
| -// exploit zip and unzip in order to shuffle.
|
| -M(Sk4f) zwxy() const {
|
| - float32x4x2_t zip = vzipq_f32(fVec, vdupq_n_f32(0.0));
|
| - return vuzpq_f32(zip.val[1], zip.val[0]).val[0];
|
| -}
|
| -// Note that XYAB and ZWCD share code. If both are needed, they could be
|
| -// implemented more efficiently together. Also, ABXY and CDZW are available
|
| -// as well.
|
| -M(Sk4f) XYAB(const Sk4f& xyzw, const Sk4f& abcd) {
|
| - float32x4x2_t xayb_zcwd = vzipq_f32(xyzw.fVec, abcd.fVec);
|
| - float32x4x2_t axby_czdw = vzipq_f32(abcd.fVec, xyzw.fVec);
|
| - return vuzpq_f32(xayb_zcwd.val[0], axby_czdw.val[0]).val[0];
|
| -}
|
| -M(Sk4f) ZWCD(const Sk4f& xyzw, const Sk4f& abcd) {
|
| - float32x4x2_t xayb_zcwd = vzipq_f32(xyzw.fVec, abcd.fVec);
|
| - float32x4x2_t axby_czdw = vzipq_f32(abcd.fVec, xyzw.fVec);
|
| - return vuzpq_f32(xayb_zcwd.val[1], axby_czdw.val[1]).val[0];
|
| -}
|
| +M(Sk4f) badc() const { return vrev64q_f32(fVec); }
|
|
|
| // Sk4i Methods
|
| #undef M
|
| @@ -174,26 +155,7 @@ M(Sk4i) multiply(const Sk4i& o) const { return vmulq_s32(fVec, o.fVec); }
|
| M(Sk4i) Min(const Sk4i& a, const Sk4i& b) { return vminq_s32(a.fVec, b.fVec); }
|
| M(Sk4i) Max(const Sk4i& a, const Sk4i& b) { return vmaxq_s32(a.fVec, b.fVec); }
|
|
|
| -// These shuffle operations are implemented more efficiently with SSE.
|
| -// NEON has efficient zip, unzip, and transpose, but it is more costly to
|
| -// exploit zip and unzip in order to shuffle.
|
| -M(Sk4i) zwxy() const {
|
| - int32x4x2_t zip = vzipq_s32(fVec, vdupq_n_s32(0.0));
|
| - return vuzpq_s32(zip.val[1], zip.val[0]).val[0];
|
| -}
|
| -// Note that XYAB and ZWCD share code. If both are needed, they could be
|
| -// implemented more efficiently together. Also, ABXY and CDZW are available
|
| -// as well.
|
| -M(Sk4i) XYAB(const Sk4i& xyzw, const Sk4i& abcd) {
|
| - int32x4x2_t xayb_zcwd = vzipq_s32(xyzw.fVec, abcd.fVec);
|
| - int32x4x2_t axby_czdw = vzipq_s32(abcd.fVec, xyzw.fVec);
|
| - return vuzpq_s32(xayb_zcwd.val[0], axby_czdw.val[0]).val[0];
|
| -}
|
| -M(Sk4i) ZWCD(const Sk4i& xyzw, const Sk4i& abcd) {
|
| - int32x4x2_t xayb_zcwd = vzipq_s32(xyzw.fVec, abcd.fVec);
|
| - int32x4x2_t axby_czdw = vzipq_s32(abcd.fVec, xyzw.fVec);
|
| - return vuzpq_s32(xayb_zcwd.val[1], axby_czdw.val[1]).val[0];
|
| -}
|
| +M(Sk4i) badc() const { return vrev64q_s32(fVec); }
|
|
|
| #undef M
|
|
|
|
|