Index: patched-ffmpeg-mt/libavcodec/sh4/qpel.c |
=================================================================== |
--- patched-ffmpeg-mt/libavcodec/sh4/qpel.c (revision 41250) |
+++ patched-ffmpeg-mt/libavcodec/sh4/qpel.c (working copy) |
@@ -26,7 +26,7 @@ |
static inline void OPNAME ## _pixels4_l2_aligned(uint8_t *dst, const uint8_t *src1, const uint8_t *src2, int dst_stride, int src_stride1, int src_stride2, int h) \ |
{\ |
do {\ |
- OP(LP(dst ),rnd_avg32(LP(src1 ),LP(src2 )) ); \ |
+ OP(LP(dst ),rnd_avg32(LPC(src1 ),LPC(src2 )) ); \ |
src1+=src_stride1; \ |
src2+=src_stride2; \ |
dst+=dst_stride; \ |
@@ -36,7 +36,7 @@ |
static inline void OPNAME ## _pixels4_l2_aligned2(uint8_t *dst, const uint8_t *src1, const uint8_t *src2, int dst_stride, int src_stride1, int src_stride2, int h) \ |
{\ |
do {\ |
- OP(LP(dst ),rnd_avg32(AV_RN32(src1 ),LP(src2 )) ); \ |
+ OP(LP(dst ),rnd_avg32(AV_RN32(src1 ),LPC(src2 )) ); \ |
src1+=src_stride1; \ |
src2+=src_stride2; \ |
dst+=dst_stride; \ |
@@ -46,10 +46,10 @@ |
static inline void OPNAME ## _no_rnd_pixels16_l2_aligned2(uint8_t *dst, const uint8_t *src1, const uint8_t *src2, int dst_stride, int src_stride1, int src_stride2, int h) \ |
{\ |
do {\ |
- OP(LP(dst ),no_rnd_avg32(AV_RN32(src1 ),LP(src2 )) ); \ |
- OP(LP(dst+4),no_rnd_avg32(AV_RN32(src1+4),LP(src2+4)) ); \ |
- OP(LP(dst+8),no_rnd_avg32(AV_RN32(src1+8),LP(src2+8)) ); \ |
- OP(LP(dst+12),no_rnd_avg32(AV_RN32(src1+12),LP(src2+12)) ); \ |
+ OP(LP(dst ),no_rnd_avg32(AV_RN32(src1 ),LPC(src2 )) ); \ |
+ OP(LP(dst+4),no_rnd_avg32(AV_RN32(src1+4),LPC(src2+4)) ); \ |
+ OP(LP(dst+8),no_rnd_avg32(AV_RN32(src1+8),LPC(src2+8)) ); \ |
+ OP(LP(dst+12),no_rnd_avg32(AV_RN32(src1+12),LPC(src2+12)) ); \ |
src1+=src_stride1; \ |
src2+=src_stride2; \ |
dst+=dst_stride; \ |
@@ -59,10 +59,10 @@ |
static inline void OPNAME ## _pixels16_l2_aligned2(uint8_t *dst, const uint8_t *src1, const uint8_t *src2, int dst_stride, int src_stride1, int src_stride2, int h) \ |
{\ |
do {\ |
- OP(LP(dst ),rnd_avg32(AV_RN32(src1 ),LP(src2 )) ); \ |
- OP(LP(dst+4),rnd_avg32(AV_RN32(src1+4),LP(src2+4)) ); \ |
- OP(LP(dst+8),rnd_avg32(AV_RN32(src1+8),LP(src2+8)) ); \ |
- OP(LP(dst+12),rnd_avg32(AV_RN32(src1+12),LP(src2+12)) ); \ |
+ OP(LP(dst ),rnd_avg32(AV_RN32(src1 ),LPC(src2 )) ); \ |
+ OP(LP(dst+4),rnd_avg32(AV_RN32(src1+4),LPC(src2+4)) ); \ |
+ OP(LP(dst+8),rnd_avg32(AV_RN32(src1+8),LPC(src2+8)) ); \ |
+ OP(LP(dst+12),rnd_avg32(AV_RN32(src1+12),LPC(src2+12)) ); \ |
src1+=src_stride1; \ |
src2+=src_stride2; \ |
dst+=dst_stride; \ |
@@ -72,8 +72,8 @@ |
static inline void OPNAME ## _no_rnd_pixels8_l2_aligned2(uint8_t *dst, const uint8_t *src1, const uint8_t *src2, int dst_stride, int src_stride1, int src_stride2, int h) \ |
{\ |
do { /* onlye src2 aligned */\ |
- OP(LP(dst ),no_rnd_avg32(AV_RN32(src1 ),LP(src2 )) ); \ |
- OP(LP(dst+4),no_rnd_avg32(AV_RN32(src1+4),LP(src2+4)) ); \ |
+ OP(LP(dst ),no_rnd_avg32(AV_RN32(src1 ),LPC(src2 )) ); \ |
+ OP(LP(dst+4),no_rnd_avg32(AV_RN32(src1+4),LPC(src2+4)) ); \ |
src1+=src_stride1; \ |
src2+=src_stride2; \ |
dst+=dst_stride; \ |
@@ -83,8 +83,8 @@ |
static inline void OPNAME ## _pixels8_l2_aligned2(uint8_t *dst, const uint8_t *src1, const uint8_t *src2, int dst_stride, int src_stride1, int src_stride2, int h) \ |
{\ |
do {\ |
- OP(LP(dst ),rnd_avg32(AV_RN32(src1 ),LP(src2 )) ); \ |
- OP(LP(dst+4),rnd_avg32(AV_RN32(src1+4),LP(src2+4)) ); \ |
+ OP(LP(dst ),rnd_avg32(AV_RN32(src1 ),LPC(src2 )) ); \ |
+ OP(LP(dst+4),rnd_avg32(AV_RN32(src1+4),LPC(src2+4)) ); \ |
src1+=src_stride1; \ |
src2+=src_stride2; \ |
dst+=dst_stride; \ |
@@ -94,8 +94,8 @@ |
static inline void OPNAME ## _no_rnd_pixels8_l2_aligned(uint8_t *dst, const uint8_t *src1, const uint8_t *src2, int dst_stride, int src_stride1, int src_stride2, int h) \ |
{\ |
do {\ |
- OP(LP(dst ),no_rnd_avg32(LP(src1 ),LP(src2 )) ); \ |
- OP(LP(dst+4),no_rnd_avg32(LP(src1+4),LP(src2+4)) ); \ |
+ OP(LP(dst ),no_rnd_avg32(LPC(src1 ),LPC(src2 )) ); \ |
+ OP(LP(dst+4),no_rnd_avg32(LPC(src1+4),LPC(src2+4)) ); \ |
src1+=src_stride1; \ |
src2+=src_stride2; \ |
dst+=dst_stride; \ |
@@ -105,8 +105,8 @@ |
static inline void OPNAME ## _pixels8_l2_aligned(uint8_t *dst, const uint8_t *src1, const uint8_t *src2, int dst_stride, int src_stride1, int src_stride2, int h) \ |
{\ |
do {\ |
- OP(LP(dst ),rnd_avg32(LP(src1 ),LP(src2 )) ); \ |
- OP(LP(dst+4),rnd_avg32(LP(src1+4),LP(src2+4)) ); \ |
+ OP(LP(dst ),rnd_avg32(LPC(src1 ),LPC(src2 )) ); \ |
+ OP(LP(dst+4),rnd_avg32(LPC(src1+4),LPC(src2+4)) ); \ |
src1+=src_stride1; \ |
src2+=src_stride2; \ |
dst+=dst_stride; \ |
@@ -116,10 +116,10 @@ |
static inline void OPNAME ## _no_rnd_pixels16_l2_aligned(uint8_t *dst, const uint8_t *src1, const uint8_t *src2, int dst_stride, int src_stride1, int src_stride2, int h) \ |
{\ |
do {\ |
- OP(LP(dst ),no_rnd_avg32(LP(src1 ),LP(src2 )) ); \ |
- OP(LP(dst+4),no_rnd_avg32(LP(src1+4),LP(src2+4)) ); \ |
- OP(LP(dst+8),no_rnd_avg32(LP(src1+8),LP(src2+8)) ); \ |
- OP(LP(dst+12),no_rnd_avg32(LP(src1+12),LP(src2+12)) ); \ |
+ OP(LP(dst ),no_rnd_avg32(LPC(src1 ),LPC(src2 )) ); \ |
+ OP(LP(dst+4),no_rnd_avg32(LPC(src1+4),LPC(src2+4)) ); \ |
+ OP(LP(dst+8),no_rnd_avg32(LPC(src1+8),LPC(src2+8)) ); \ |
+ OP(LP(dst+12),no_rnd_avg32(LPC(src1+12),LPC(src2+12)) ); \ |
src1+=src_stride1; \ |
src2+=src_stride2; \ |
dst+=dst_stride; \ |
@@ -129,10 +129,10 @@ |
static inline void OPNAME ## _pixels16_l2_aligned(uint8_t *dst, const uint8_t *src1, const uint8_t *src2, int dst_stride, int src_stride1, int src_stride2, int h) \ |
{\ |
do {\ |
- OP(LP(dst ),rnd_avg32(LP(src1 ),LP(src2 )) ); \ |
- OP(LP(dst+4),rnd_avg32(LP(src1+4),LP(src2+4)) ); \ |
- OP(LP(dst+8),rnd_avg32(LP(src1+8),LP(src2+8)) ); \ |
- OP(LP(dst+12),rnd_avg32(LP(src1+12),LP(src2+12)) ); \ |
+ OP(LP(dst ),rnd_avg32(LPC(src1 ),LPC(src2 )) ); \ |
+ OP(LP(dst+4),rnd_avg32(LPC(src1+4),LPC(src2+4)) ); \ |
+ OP(LP(dst+8),rnd_avg32(LPC(src1+8),LPC(src2+8)) ); \ |
+ OP(LP(dst+12),rnd_avg32(LPC(src1+12),LPC(src2+12)) ); \ |
src1+=src_stride1; \ |
src2+=src_stride2; \ |
dst+=dst_stride; \ |
@@ -154,11 +154,11 @@ |
static inline void OPNAME ## _pixels8_l4_aligned(uint8_t *dst, const uint8_t *src1, uint8_t *src2, uint8_t *src3, uint8_t *src4,int dst_stride, int src_stride1, int src_stride2,int src_stride3,int src_stride4, int h){\ |
do { \ |
uint32_t a0,a1,a2,a3; \ |
- UNPACK(a0,a1,LP(src1),LP(src2)); \ |
- UNPACK(a2,a3,LP(src3),LP(src4)); \ |
+ UNPACK(a0,a1,LPC(src1),LPC(src2)); \ |
+ UNPACK(a2,a3,LPC(src3),LPC(src4)); \ |
OP(LP(dst),rnd_PACK(a0,a1,a2,a3)); \ |
- UNPACK(a0,a1,LP(src1+4),LP(src2+4)); \ |
- UNPACK(a2,a3,LP(src3+4),LP(src4+4)); \ |
+ UNPACK(a0,a1,LPC(src1+4),LPC(src2+4)); \ |
+ UNPACK(a2,a3,LPC(src3+4),LPC(src4+4)); \ |
OP(LP(dst+4),rnd_PACK(a0,a1,a2,a3)); \ |
src1+=src_stride1;\ |
src2+=src_stride2;\ |
@@ -171,11 +171,11 @@ |
static inline void OPNAME ## _no_rnd_pixels8_l4_aligned(uint8_t *dst, const uint8_t *src1, uint8_t *src2, uint8_t *src3, uint8_t *src4,int dst_stride, int src_stride1, int src_stride2,int src_stride3,int src_stride4, int h){\ |
do { \ |
uint32_t a0,a1,a2,a3; \ |
- UNPACK(a0,a1,LP(src1),LP(src2)); \ |
- UNPACK(a2,a3,LP(src3),LP(src4)); \ |
+ UNPACK(a0,a1,LPC(src1),LPC(src2)); \ |
+ UNPACK(a2,a3,LPC(src3),LPC(src4)); \ |
OP(LP(dst),no_rnd_PACK(a0,a1,a2,a3)); \ |
- UNPACK(a0,a1,LP(src1+4),LP(src2+4)); \ |
- UNPACK(a2,a3,LP(src3+4),LP(src4+4)); \ |
+ UNPACK(a0,a1,LPC(src1+4),LPC(src2+4)); \ |
+ UNPACK(a2,a3,LPC(src3+4),LPC(src4+4)); \ |
OP(LP(dst+4),no_rnd_PACK(a0,a1,a2,a3)); \ |
src1+=src_stride1;\ |
src2+=src_stride2;\ |
@@ -188,11 +188,11 @@ |
static inline void OPNAME ## _pixels8_l4_aligned0(uint8_t *dst, const uint8_t *src1, uint8_t *src2, uint8_t *src3, uint8_t *src4,int dst_stride, int src_stride1, int src_stride2,int src_stride3,int src_stride4, int h){\ |
do { \ |
uint32_t a0,a1,a2,a3; /* src1 only not aligned */\ |
- UNPACK(a0,a1,AV_RN32(src1),LP(src2)); \ |
- UNPACK(a2,a3,LP(src3),LP(src4)); \ |
+ UNPACK(a0,a1,AV_RN32(src1),LPC(src2)); \ |
+ UNPACK(a2,a3,LPC(src3),LPC(src4)); \ |
OP(LP(dst),rnd_PACK(a0,a1,a2,a3)); \ |
- UNPACK(a0,a1,AV_RN32(src1+4),LP(src2+4)); \ |
- UNPACK(a2,a3,LP(src3+4),LP(src4+4)); \ |
+ UNPACK(a0,a1,AV_RN32(src1+4),LPC(src2+4)); \ |
+ UNPACK(a2,a3,LPC(src3+4),LPC(src4+4)); \ |
OP(LP(dst+4),rnd_PACK(a0,a1,a2,a3)); \ |
src1+=src_stride1;\ |
src2+=src_stride2;\ |
@@ -205,11 +205,11 @@ |
static inline void OPNAME ## _no_rnd_pixels8_l4_aligned0(uint8_t *dst, const uint8_t *src1, uint8_t *src2, uint8_t *src3, uint8_t *src4,int dst_stride, int src_stride1, int src_stride2,int src_stride3,int src_stride4, int h){\ |
do { \ |
uint32_t a0,a1,a2,a3; \ |
- UNPACK(a0,a1,AV_RN32(src1),LP(src2)); \ |
- UNPACK(a2,a3,LP(src3),LP(src4)); \ |
+ UNPACK(a0,a1,AV_RN32(src1),LPC(src2)); \ |
+ UNPACK(a2,a3,LPC(src3),LPC(src4)); \ |
OP(LP(dst),no_rnd_PACK(a0,a1,a2,a3)); \ |
- UNPACK(a0,a1,AV_RN32(src1+4),LP(src2+4)); \ |
- UNPACK(a2,a3,LP(src3+4),LP(src4+4)); \ |
+ UNPACK(a0,a1,AV_RN32(src1+4),LPC(src2+4)); \ |
+ UNPACK(a2,a3,LPC(src3+4),LPC(src4+4)); \ |
OP(LP(dst+4),no_rnd_PACK(a0,a1,a2,a3)); \ |
src1+=src_stride1;\ |
src2+=src_stride2;\ |
@@ -222,17 +222,17 @@ |
static inline void OPNAME ## _pixels16_l4_aligned(uint8_t *dst, const uint8_t *src1, uint8_t *src2, uint8_t *src3, uint8_t *src4,int dst_stride, int src_stride1, int src_stride2,int src_stride3,int src_stride4, int h){\ |
do { \ |
uint32_t a0,a1,a2,a3; \ |
- UNPACK(a0,a1,LP(src1),LP(src2)); \ |
- UNPACK(a2,a3,LP(src3),LP(src4)); \ |
+ UNPACK(a0,a1,LPC(src1),LPC(src2)); \ |
+ UNPACK(a2,a3,LPC(src3),LPC(src4)); \ |
OP(LP(dst),rnd_PACK(a0,a1,a2,a3)); \ |
- UNPACK(a0,a1,LP(src1+4),LP(src2+4)); \ |
- UNPACK(a2,a3,LP(src3+4),LP(src4+4)); \ |
+ UNPACK(a0,a1,LPC(src1+4),LPC(src2+4)); \ |
+ UNPACK(a2,a3,LPC(src3+4),LPC(src4+4)); \ |
OP(LP(dst+8),rnd_PACK(a0,a1,a2,a3)); \ |
- UNPACK(a0,a1,LP(src1+8),LP(src2+8)); \ |
- UNPACK(a2,a3,LP(src3+8),LP(src4+8)); \ |
+ UNPACK(a0,a1,LPC(src1+8),LPC(src2+8)); \ |
+ UNPACK(a2,a3,LPC(src3+8),LPC(src4+8)); \ |
OP(LP(dst+8),rnd_PACK(a0,a1,a2,a3)); \ |
- UNPACK(a0,a1,LP(src1+12),LP(src2+12)); \ |
- UNPACK(a2,a3,LP(src3+12),LP(src4+12)); \ |
+ UNPACK(a0,a1,LPC(src1+12),LPC(src2+12)); \ |
+ UNPACK(a2,a3,LPC(src3+12),LPC(src4+12)); \ |
OP(LP(dst+12),rnd_PACK(a0,a1,a2,a3)); \ |
src1+=src_stride1;\ |
src2+=src_stride2;\ |
@@ -245,17 +245,17 @@ |
static inline void OPNAME ## _no_rnd_pixels16_l4_aligned(uint8_t *dst, const uint8_t *src1, uint8_t *src2, uint8_t *src3, uint8_t *src4,int dst_stride, int src_stride1, int src_stride2,int src_stride3,int src_stride4, int h){\ |
do { \ |
uint32_t a0,a1,a2,a3; \ |
- UNPACK(a0,a1,LP(src1),LP(src2)); \ |
- UNPACK(a2,a3,LP(src3),LP(src4)); \ |
+ UNPACK(a0,a1,LPC(src1),LPC(src2)); \ |
+ UNPACK(a2,a3,LPC(src3),LPC(src4)); \ |
OP(LP(dst),no_rnd_PACK(a0,a1,a2,a3)); \ |
- UNPACK(a0,a1,LP(src1+4),LP(src2+4)); \ |
- UNPACK(a2,a3,LP(src3+4),LP(src4+4)); \ |
+ UNPACK(a0,a1,LPC(src1+4),LPC(src2+4)); \ |
+ UNPACK(a2,a3,LPC(src3+4),LPC(src4+4)); \ |
OP(LP(dst+4),no_rnd_PACK(a0,a1,a2,a3)); \ |
- UNPACK(a0,a1,LP(src1+8),LP(src2+8)); \ |
- UNPACK(a2,a3,LP(src3+8),LP(src4+8)); \ |
+ UNPACK(a0,a1,LPC(src1+8),LPC(src2+8)); \ |
+ UNPACK(a2,a3,LPC(src3+8),LPC(src4+8)); \ |
OP(LP(dst+8),no_rnd_PACK(a0,a1,a2,a3)); \ |
- UNPACK(a0,a1,LP(src1+12),LP(src2+12)); \ |
- UNPACK(a2,a3,LP(src3+12),LP(src4+12)); \ |
+ UNPACK(a0,a1,LPC(src1+12),LPC(src2+12)); \ |
+ UNPACK(a2,a3,LPC(src3+12),LPC(src4+12)); \ |
OP(LP(dst+12),no_rnd_PACK(a0,a1,a2,a3)); \ |
src1+=src_stride1;\ |
src2+=src_stride2;\ |
@@ -268,17 +268,17 @@ |
static inline void OPNAME ## _pixels16_l4_aligned0(uint8_t *dst, const uint8_t *src1, uint8_t *src2, uint8_t *src3, uint8_t *src4,int dst_stride, int src_stride1, int src_stride2,int src_stride3,int src_stride4, int h){\ |
do { /* src1 is unaligned */\ |
uint32_t a0,a1,a2,a3; \ |
- UNPACK(a0,a1,AV_RN32(src1),LP(src2)); \ |
- UNPACK(a2,a3,LP(src3),LP(src4)); \ |
+ UNPACK(a0,a1,AV_RN32(src1),LPC(src2)); \ |
+ UNPACK(a2,a3,LPC(src3),LPC(src4)); \ |
OP(LP(dst),rnd_PACK(a0,a1,a2,a3)); \ |
- UNPACK(a0,a1,AV_RN32(src1+4),LP(src2+4)); \ |
- UNPACK(a2,a3,LP(src3+4),LP(src4+4)); \ |
+ UNPACK(a0,a1,AV_RN32(src1+4),LPC(src2+4)); \ |
+ UNPACK(a2,a3,LPC(src3+4),LPC(src4+4)); \ |
OP(LP(dst+8),rnd_PACK(a0,a1,a2,a3)); \ |
- UNPACK(a0,a1,AV_RN32(src1+8),LP(src2+8)); \ |
- UNPACK(a2,a3,LP(src3+8),LP(src4+8)); \ |
+ UNPACK(a0,a1,AV_RN32(src1+8),LPC(src2+8)); \ |
+ UNPACK(a2,a3,LPC(src3+8),LPC(src4+8)); \ |
OP(LP(dst+8),rnd_PACK(a0,a1,a2,a3)); \ |
- UNPACK(a0,a1,AV_RN32(src1+12),LP(src2+12)); \ |
- UNPACK(a2,a3,LP(src3+12),LP(src4+12)); \ |
+ UNPACK(a0,a1,AV_RN32(src1+12),LPC(src2+12)); \ |
+ UNPACK(a2,a3,LPC(src3+12),LPC(src4+12)); \ |
OP(LP(dst+12),rnd_PACK(a0,a1,a2,a3)); \ |
src1+=src_stride1;\ |
src2+=src_stride2;\ |
@@ -291,17 +291,17 @@ |
static inline void OPNAME ## _no_rnd_pixels16_l4_aligned0(uint8_t *dst, const uint8_t *src1, uint8_t *src2, uint8_t *src3, uint8_t *src4,int dst_stride, int src_stride1, int src_stride2,int src_stride3,int src_stride4, int h){\ |
do { \ |
uint32_t a0,a1,a2,a3; \ |
- UNPACK(a0,a1,AV_RN32(src1),LP(src2)); \ |
- UNPACK(a2,a3,LP(src3),LP(src4)); \ |
+ UNPACK(a0,a1,AV_RN32(src1),LPC(src2)); \ |
+ UNPACK(a2,a3,LPC(src3),LPC(src4)); \ |
OP(LP(dst),no_rnd_PACK(a0,a1,a2,a3)); \ |
- UNPACK(a0,a1,AV_RN32(src1+4),LP(src2+4)); \ |
- UNPACK(a2,a3,LP(src3+4),LP(src4+4)); \ |
+ UNPACK(a0,a1,AV_RN32(src1+4),LPC(src2+4)); \ |
+ UNPACK(a2,a3,LPC(src3+4),LPC(src4+4)); \ |
OP(LP(dst+4),no_rnd_PACK(a0,a1,a2,a3)); \ |
- UNPACK(a0,a1,AV_RN32(src1+8),LP(src2+8)); \ |
- UNPACK(a2,a3,LP(src3+8),LP(src4+8)); \ |
+ UNPACK(a0,a1,AV_RN32(src1+8),LPC(src2+8)); \ |
+ UNPACK(a2,a3,LPC(src3+8),LPC(src4+8)); \ |
OP(LP(dst+8),no_rnd_PACK(a0,a1,a2,a3)); \ |
- UNPACK(a0,a1,AV_RN32(src1+12),LP(src2+12)); \ |
- UNPACK(a2,a3,LP(src3+12),LP(src4+12)); \ |
+ UNPACK(a0,a1,AV_RN32(src1+12),LPC(src2+12)); \ |
+ UNPACK(a2,a3,LPC(src3+12),LPC(src4+12)); \ |
OP(LP(dst+12),no_rnd_PACK(a0,a1,a2,a3)); \ |
src1+=src_stride1;\ |
src2+=src_stride2;\ |