[FFmpeg-devel] [PATCH 1/7] postproc/postprocess: Remove obsolete MMX(EXT)/3Dnow functions
Paul B Mahol
onemda at gmail.com
Tue Nov 8 17:43:02 EET 2022
On 11/8/22, Andreas Rheinhardt <andreas.rheinhardt at outlook.com> wrote:
> Andreas Rheinhardt:
>> postprocess.c currently has C, MMX, MMXEXT, 3DNow as well as
>> SSE2 versions of its internal functions. But given that only
>> ancient 32-bit x86 CPUs don't support SSE2, the MMX, MMXEXT
>> and 3DNow versions are obsolete and are therefore removed by
>> this commit. This saves about 56KB here.
>>
>> (The SSE2 version in particular is not really complete,
>> so that it often falls back to MMXEXT (which means that
>> there were some identical (apart from the name) MMXEXT
>> and SSE2 functions; this duplication no longer exists
>> with this commit.)
>>
>> Signed-off-by: Andreas Rheinhardt <andreas.rheinhardt at outlook.com>
>> ---
>> The PP_CPU_CAPS_MMX(2)/3DNOW could now be deprecated.
>>
>> libpostproc/postprocess.c | 69 ++--
>> libpostproc/postprocess_template.c | 521 ++---------------------------
>> 2 files changed, 57 insertions(+), 533 deletions(-)
>>
>> diff --git a/libpostproc/postprocess.c b/libpostproc/postprocess.c
>> index 383c691cb4..0586e458b4 100644
>> --- a/libpostproc/postprocess.c
>> +++ b/libpostproc/postprocess.c
>> @@ -26,28 +26,27 @@
>> */
>>
>> /*
>> - C MMX MMX2 3DNow AltiVec
>> -isVertDC Ec Ec Ec
>> -isVertMinMaxOk Ec Ec Ec
>> -doVertLowPass E e e Ec
>> -doVertDefFilter Ec Ec e e Ec
>> -isHorizDC Ec Ec Ec
>> -isHorizMinMaxOk a E Ec
>> -doHorizLowPass E e e Ec
>> -doHorizDefFilter Ec Ec e e Ec
>> -do_a_deblock Ec E Ec E
>> -deRing E e e* Ecp
>> -Vertical RKAlgo1 E a a
>> -Horizontal RKAlgo1 a a
>> -Vertical X1# a E E
>> -Horizontal X1# a E E
>> -LinIpolDeinterlace e E E*
>> -CubicIpolDeinterlace a e e*
>> -LinBlendDeinterlace e E E*
>> + C MMX MMX2 AltiVec
>> +isVertDC Ec Ec Ec
>> +isVertMinMaxOk Ec Ec Ec
>> +doVertLowPass E e Ec
>> +doVertDefFilter Ec Ec e Ec
>> +isHorizDC Ec Ec Ec
>> +isHorizMinMaxOk a E Ec
>> +doHorizLowPass E e Ec
>> +doHorizDefFilter Ec Ec e Ec
>> +do_a_deblock Ec E Ec
>> +deRing E e Ecp
>> +Vertical RKAlgo1 E a
>> +Horizontal RKAlgo1 a
>> +Vertical X1# a E
>> +Horizontal X1# a E
>> +LinIpolDeinterlace e E
>> +CubicIpolDeinterlace a e
>> +LinBlendDeinterlace e E
>> MedianDeinterlace# E Ec Ec
>> -TempDeNoiser# E e e Ec
>> +TempDeNoiser# E e Ec
>>
>> -* I do not have a 3DNow! CPU -> it is untested, but no one said it does
>> not work so it seems to work
>> # more or less selfinvented filters so the exactness is not too
>> meaningful
>> E = Exact implementation
>> e = almost exact implementation (slightly different rounding,...)
>> @@ -83,7 +82,6 @@ try to unroll inner for(x=0 ... loop to avoid these damn
>> if(x ... checks
>> #include <stdlib.h>
>> #include <string.h>
>> //#undef HAVE_MMXEXT_INLINE
>> -//#define HAVE_AMD3DNOW_INLINE
>> //#undef HAVE_MMX_INLINE
>> //#undef ARCH_X86
>> //#define DEBUG_BRIGHTNESS
>> @@ -494,7 +492,7 @@ static av_always_inline void do_a_deblock_C(uint8_t
>> *src, int step,
>> }
>> }
>>
>> -//Note: we have C, MMX, MMX2, 3DNOW version there is no 3DNOW+MMX2 one
>> +//Note: we have C and SSE2 version (which uses MMX(EXT) when
>> advantageous)
>> //Plain C versions
>> //we always compile C for testing which needs bitexactness
>> #define TEMPLATE_PP_C 1
>> @@ -508,27 +506,12 @@ static av_always_inline void do_a_deblock_C(uint8_t
>> *src, int step,
>>
>> #if ARCH_X86 && HAVE_INLINE_ASM
>> # if CONFIG_RUNTIME_CPUDETECT
>> -# define TEMPLATE_PP_MMX 1
>> -# include "postprocess_template.c"
>> -# define TEMPLATE_PP_MMXEXT 1
>> -# include "postprocess_template.c"
>> -# define TEMPLATE_PP_3DNOW 1
>> -# include "postprocess_template.c"
>> # define TEMPLATE_PP_SSE2 1
>> # include "postprocess_template.c"
>> # else
>> # if HAVE_SSE2_INLINE
>> # define TEMPLATE_PP_SSE2 1
>> # include "postprocess_template.c"
>> -# elif HAVE_MMXEXT_INLINE
>> -# define TEMPLATE_PP_MMXEXT 1
>> -# include "postprocess_template.c"
>> -# elif HAVE_AMD3DNOW_INLINE
>> -# define TEMPLATE_PP_3DNOW 1
>> -# include "postprocess_template.c"
>> -# elif HAVE_MMX_INLINE
>> -# define TEMPLATE_PP_MMX 1
>> -# include "postprocess_template.c"
>> # endif
>> # endif
>> #endif
>> @@ -549,21 +532,12 @@ static inline void postProcess(const uint8_t src[],
>> int srcStride, uint8_t dst[]
>> #if ARCH_X86 && HAVE_INLINE_ASM
>> // ordered per speed fastest first
>> if (c->cpuCaps & AV_CPU_FLAG_SSE2) pp =
>> postProcess_SSE2;
>> - else if (c->cpuCaps & AV_CPU_FLAG_MMXEXT) pp =
>> postProcess_MMX2;
>> - else if (c->cpuCaps & AV_CPU_FLAG_3DNOW) pp =
>> postProcess_3DNow;
>> - else if (c->cpuCaps & AV_CPU_FLAG_MMX) pp =
>> postProcess_MMX;
>> #elif HAVE_ALTIVEC
>> if (c->cpuCaps & AV_CPU_FLAG_ALTIVEC) pp =
>> postProcess_altivec;
>> #endif
>> #else /* CONFIG_RUNTIME_CPUDETECT */
>> #if HAVE_SSE2_INLINE
>> pp = postProcess_SSE2;
>> -#elif HAVE_MMXEXT_INLINE
>> - pp = postProcess_MMX2;
>> -#elif HAVE_AMD3DNOW_INLINE
>> - pp = postProcess_3DNow;
>> -#elif HAVE_MMX_INLINE
>> - pp = postProcess_MMX;
>> #elif HAVE_ALTIVEC
>> pp = postProcess_altivec;
>> #endif
>> @@ -877,9 +851,6 @@ av_cold pp_context *pp_get_context(int width, int
>> height, int cpuCaps){
>> c->cpuCaps = av_get_cpu_flags();
>> } else {
>> c->cpuCaps = 0;
>> - if (cpuCaps & PP_CPU_CAPS_MMX) c->cpuCaps |=
>> AV_CPU_FLAG_MMX;
>> - if (cpuCaps & PP_CPU_CAPS_MMX2) c->cpuCaps |=
>> AV_CPU_FLAG_MMXEXT;
>> - if (cpuCaps & PP_CPU_CAPS_3DNOW) c->cpuCaps |=
>> AV_CPU_FLAG_3DNOW;
>> if (cpuCaps & PP_CPU_CAPS_ALTIVEC) c->cpuCaps |=
>> AV_CPU_FLAG_ALTIVEC;
>> }
>>
>> diff --git a/libpostproc/postprocess_template.c
>> b/libpostproc/postprocess_template.c
>> index 9f76b7c587..bcf7bdad66 100644
>> --- a/libpostproc/postprocess_template.c
>> +++ b/libpostproc/postprocess_template.c
>> @@ -20,7 +20,7 @@
>>
>> /**
>> * @file
>> - * mmx/mmx2/3dnow postprocess code.
>> + * mmx/mmx2/sse2 postprocess code.
>> */
>>
>> #include "libavutil/mem_internal.h"
>> @@ -57,14 +57,6 @@
>> # define TEMPLATE_PP_MMXEXT 0
>> #endif
>>
>> -#ifdef TEMPLATE_PP_3DNOW
>> -# undef TEMPLATE_PP_MMX
>> -# define TEMPLATE_PP_MMX 1
>> -# define RENAME(a) a ## _3DNow
>> -#else
>> -# define TEMPLATE_PP_3DNOW 0
>> -#endif
>> -
>> #ifdef TEMPLATE_PP_SSE2
>> # undef TEMPLATE_PP_MMX
>> # define TEMPLATE_PP_MMX 1
>> @@ -82,30 +74,19 @@
>>
>> #if TEMPLATE_PP_MMXEXT
>> #define REAL_PAVGB(a,b) "pavgb " #a ", " #b " \n\t"
>> -#elif TEMPLATE_PP_3DNOW
>> -#define REAL_PAVGB(a,b) "pavgusb " #a ", " #b " \n\t"
>> #endif
>> #define PAVGB(a,b) REAL_PAVGB(a,b)
>>
>> #if TEMPLATE_PP_MMXEXT
>> #define PMINUB(a,b,t) "pminub " #a ", " #b " \n\t"
>> -#elif TEMPLATE_PP_MMX
>> -#define PMINUB(b,a,t) \
>> - "movq " #a ", " #t " \n\t"\
>> - "psubusb " #b ", " #t " \n\t"\
>> - "psubb " #t ", " #a " \n\t"
>> #endif
>>
>> #if TEMPLATE_PP_MMXEXT
>> #define PMAXUB(a,b) "pmaxub " #a ", " #b " \n\t"
>> -#elif TEMPLATE_PP_MMX
>> -#define PMAXUB(a,b) \
>> - "psubusb " #a ", " #b " \n\t"\
>> - "paddb " #a ", " #b " \n\t"
>> #endif
>>
>> //FIXME? |255-0| = 1 (should not be a problem ...)
>> -#if TEMPLATE_PP_MMX
>> +#if TEMPLATE_PP_MMXEXT
>> /**
>> * Check if the middle 8x8 Block in the given 8x16 block is flat
>> */
>> @@ -185,20 +166,8 @@ static inline int RENAME(vertClassify)(const uint8_t
>> src[], int stride, PPContex
>> "psubusb %%mm3, %%mm4 \n\t"
>>
>> " \n\t"
>> -#if TEMPLATE_PP_MMXEXT
>> "pxor %%mm7, %%mm7 \n\t"
>> "psadbw %%mm7, %%mm0 \n\t"
>> -#else
>> - "movq %%mm0, %%mm1 \n\t"
>> - "psrlw $8, %%mm0 \n\t"
>> - "paddb %%mm1, %%mm0 \n\t"
>> - "movq %%mm0, %%mm1 \n\t"
>> - "psrlq $16, %%mm0 \n\t"
>> - "paddb %%mm1, %%mm0 \n\t"
>> - "movq %%mm0, %%mm1 \n\t"
>> - "psrlq $32, %%mm0 \n\t"
>> - "paddb %%mm1, %%mm0 \n\t"
>> -#endif
>> "movq %4, %%mm7 \n\t" // QP,..., QP
>> "paddusb %%mm7, %%mm7 \n\t" // 2QP ... 2QP
>> "psubusb %%mm7, %%mm4 \n\t" // Diff <= 2QP ->
>> 0
>> @@ -219,7 +188,7 @@ static inline int RENAME(vertClassify)(const uint8_t
>> src[], int stride, PPContex
>> return 2;
>> }
>> }
>> -#endif //TEMPLATE_PP_MMX
>> +#endif //TEMPLATE_PP_MMXEXT
>>
>> /**
>> * Do a vertical low pass filter on the 8x16 block (only write to the 8x8
>> block in the middle)
>> @@ -228,7 +197,7 @@ static inline int RENAME(vertClassify)(const uint8_t
>> src[], int stride, PPContex
>> #if !TEMPLATE_PP_ALTIVEC
>> static inline void RENAME(doVertLowPass)(uint8_t *src, int stride,
>> PPContext *c)
>> {
>> -#if TEMPLATE_PP_MMXEXT || TEMPLATE_PP_3DNOW
>> +#if TEMPLATE_PP_MMXEXT
>> src+= stride*3;
>> __asm__ volatile( //"movv %0 %1 %2\n\t"
>> "movq %2, %%mm0 \n\t" // QP,..., QP
>> @@ -355,7 +324,7 @@ static inline void RENAME(doVertLowPass)(uint8_t *src,
>> int stride, PPContext *c)
>> : "r" (src), "r" ((x86_reg)stride), "m" (c->pQPb)
>> : "%"FF_REG_a, "%"FF_REG_c
>> );
>> -#else //TEMPLATE_PP_MMXEXT || TEMPLATE_PP_3DNOW
>> +#else //TEMPLATE_PP_MMXEXT
>> const int l1= stride;
>> const int l2= stride + l1;
>> const int l3= stride + l2;
>> @@ -394,7 +363,7 @@ static inline void RENAME(doVertLowPass)(uint8_t *src,
>> int stride, PPContext *c)
>>
>> src++;
>> }
>> -#endif //TEMPLATE_PP_MMXEXT || TEMPLATE_PP_3DNOW
>> +#endif //TEMPLATE_PP_MMXEXT
>> }
>> #endif //TEMPLATE_PP_ALTIVEC
>>
>> @@ -407,7 +376,7 @@ static inline void RENAME(doVertLowPass)(uint8_t *src,
>> int stride, PPContext *c)
>> */
>> static inline void RENAME(vertX1Filter)(uint8_t *src, int stride,
>> PPContext *co)
>> {
>> -#if TEMPLATE_PP_MMXEXT || TEMPLATE_PP_3DNOW
>> +#if TEMPLATE_PP_MMXEXT
>> src+= stride*3;
>>
>> __asm__ volatile(
>> @@ -494,7 +463,7 @@ static inline void RENAME(vertX1Filter)(uint8_t *src,
>> int stride, PPContext *co)
>> NAMED_CONSTRAINTS_ADD(b01)
>> : "%"FF_REG_a, "%"FF_REG_c
>> );
>> -#else //TEMPLATE_PP_MMXEXT || TEMPLATE_PP_3DNOW
>> +#else //TEMPLATE_PP_MMXEXT
>>
>> const int l1= stride;
>> const int l2= stride + l1;
>> @@ -528,13 +497,13 @@ static inline void RENAME(vertX1Filter)(uint8_t
>> *src, int stride, PPContext *co)
>> }
>> src++;
>> }
>> -#endif //TEMPLATE_PP_MMXEXT || TEMPLATE_PP_3DNOW
>> +#endif //TEMPLATE_PP_MMXEXT
>> }
>>
>> #if !TEMPLATE_PP_ALTIVEC
>> static inline void RENAME(doVertDefFilter)(uint8_t src[], int stride,
>> PPContext *c)
>> {
>> -#if TEMPLATE_PP_MMXEXT || TEMPLATE_PP_3DNOW
>> +#if TEMPLATE_PP_MMXEXT
>> /*
>> uint8_t tmp[16];
>> const int l1= stride;
>> @@ -816,239 +785,7 @@ static inline void RENAME(doVertDefFilter)(uint8_t
>> src[], int stride, PPContext
>> }
>> }
>> */
>> -#elif TEMPLATE_PP_MMX
>> - DECLARE_ALIGNED(8, uint64_t, tmp)[4]; // make space for 4 8-byte
>> vars
>> - src+= stride*4;
>> - __asm__ volatile(
>> - "pxor %%mm7, %%mm7 \n\t"
>> -// 0 1 2 3 4 5 6 7
>> -// %0 %0+%1 %0+2%1 eax+2%1 %0+4%1 eax+4%1 edx+%1 edx+2%1
>> -// %0 eax eax+%1 eax+2%1 %0+4%1 edx edx+%1 edx+2%1
>> -
>> - "movq (%0), %%mm0 \n\t"
>> - "movq %%mm0, %%mm1 \n\t"
>> - "punpcklbw %%mm7, %%mm0 \n\t" // low part of line
>> 0
>> - "punpckhbw %%mm7, %%mm1 \n\t" // high part of
>> line 0
>> -
>> - "movq (%0, %1), %%mm2 \n\t"
>> - "lea (%0, %1, 2), %%"FF_REG_a" \n\t"
>> - "movq %%mm2, %%mm3 \n\t"
>> - "punpcklbw %%mm7, %%mm2 \n\t" // low part of line
>> 1
>> - "punpckhbw %%mm7, %%mm3 \n\t" // high part of
>> line 1
>> -
>> - "movq (%%"FF_REG_a"), %%mm4 \n\t"
>> - "movq %%mm4, %%mm5 \n\t"
>> - "punpcklbw %%mm7, %%mm4 \n\t" // low part of line
>> 2
>> - "punpckhbw %%mm7, %%mm5 \n\t" // high part of
>> line 2
>> -
>> - "paddw %%mm0, %%mm0 \n\t" // 2L0
>> - "paddw %%mm1, %%mm1 \n\t" // 2H0
>> - "psubw %%mm4, %%mm2 \n\t" // L1 - L2
>> - "psubw %%mm5, %%mm3 \n\t" // H1 - H2
>> - "psubw %%mm2, %%mm0 \n\t" // 2L0 - L1 + L2
>> - "psubw %%mm3, %%mm1 \n\t" // 2H0 - H1 + H2
>> -
>> - "psllw $2, %%mm2 \n\t" // 4L1 - 4L2
>> - "psllw $2, %%mm3 \n\t" // 4H1 - 4H2
>> - "psubw %%mm2, %%mm0 \n\t" // 2L0 - 5L1 + 5L2
>> - "psubw %%mm3, %%mm1 \n\t" // 2H0 - 5H1 + 5H2
>> -
>> - "movq (%%"FF_REG_a", %1), %%mm2 \n\t"
>> - "movq %%mm2, %%mm3 \n\t"
>> - "punpcklbw %%mm7, %%mm2 \n\t" // L3
>> - "punpckhbw %%mm7, %%mm3 \n\t" // H3
>> -
>> - "psubw %%mm2, %%mm0 \n\t" // 2L0 - 5L1 + 5L2
>> - L3
>> - "psubw %%mm3, %%mm1 \n\t" // 2H0 - 5H1 + 5H2
>> - H3
>> - "psubw %%mm2, %%mm0 \n\t" // 2L0 - 5L1 + 5L2
>> - 2L3
>> - "psubw %%mm3, %%mm1 \n\t" // 2H0 - 5H1 + 5H2
>> - 2H3
>> - "movq %%mm0, (%3) \n\t" // 2L0 - 5L1 + 5L2
>> - 2L3
>> - "movq %%mm1, 8(%3) \n\t" // 2H0 - 5H1 + 5H2
>> - 2H3
>> -
>> - "movq (%%"FF_REG_a", %1, 2), %%mm0 \n\t"
>> - "movq %%mm0, %%mm1 \n\t"
>> - "punpcklbw %%mm7, %%mm0 \n\t" // L4
>> - "punpckhbw %%mm7, %%mm1 \n\t" // H4
>> -
>> - "psubw %%mm0, %%mm2 \n\t" // L3 - L4
>> - "psubw %%mm1, %%mm3 \n\t" // H3 - H4
>> - "movq %%mm2, 16(%3) \n\t" // L3 - L4
>> - "movq %%mm3, 24(%3) \n\t" // H3 - H4
>> - "paddw %%mm4, %%mm4 \n\t" // 2L2
>> - "paddw %%mm5, %%mm5 \n\t" // 2H2
>> - "psubw %%mm2, %%mm4 \n\t" // 2L2 - L3 + L4
>> - "psubw %%mm3, %%mm5 \n\t" // 2H2 - H3 + H4
>> -
>> - "lea (%%"FF_REG_a", %1), %0 \n\t"
>> - "psllw $2, %%mm2 \n\t" // 4L3 - 4L4
>> - "psllw $2, %%mm3 \n\t" // 4H3 - 4H4
>> - "psubw %%mm2, %%mm4 \n\t" // 2L2 - 5L3 + 5L4
>> - "psubw %%mm3, %%mm5 \n\t" // 2H2 - 5H3 + 5H4
>> -//50 opcodes so far
>> - "movq (%0, %1, 2), %%mm2 \n\t"
>> - "movq %%mm2, %%mm3 \n\t"
>> - "punpcklbw %%mm7, %%mm2 \n\t" // L5
>> - "punpckhbw %%mm7, %%mm3 \n\t" // H5
>> - "psubw %%mm2, %%mm4 \n\t" // 2L2 - 5L3 + 5L4
>> - L5
>> - "psubw %%mm3, %%mm5 \n\t" // 2H2 - 5H3 + 5H4
>> - H5
>> - "psubw %%mm2, %%mm4 \n\t" // 2L2 - 5L3 + 5L4
>> - 2L5
>> - "psubw %%mm3, %%mm5 \n\t" // 2H2 - 5H3 + 5H4
>> - 2H5
>> -
>> - "movq (%%"FF_REG_a", %1, 4), %%mm6 \n\t"
>> - "punpcklbw %%mm7, %%mm6 \n\t" // L6
>> - "psubw %%mm6, %%mm2 \n\t" // L5 - L6
>> - "movq (%%"FF_REG_a", %1, 4), %%mm6 \n\t"
>> - "punpckhbw %%mm7, %%mm6 \n\t" // H6
>> - "psubw %%mm6, %%mm3 \n\t" // H5 - H6
>> -
>> - "paddw %%mm0, %%mm0 \n\t" // 2L4
>> - "paddw %%mm1, %%mm1 \n\t" // 2H4
>> - "psubw %%mm2, %%mm0 \n\t" // 2L4 - L5 + L6
>> - "psubw %%mm3, %%mm1 \n\t" // 2H4 - H5 + H6
>> -
>> - "psllw $2, %%mm2 \n\t" // 4L5 - 4L6
>> - "psllw $2, %%mm3 \n\t" // 4H5 - 4H6
>> - "psubw %%mm2, %%mm0 \n\t" // 2L4 - 5L5 + 5L6
>> - "psubw %%mm3, %%mm1 \n\t" // 2H4 - 5H5 + 5H6
>> -
>> - "movq (%0, %1, 4), %%mm2 \n\t"
>> - "movq %%mm2, %%mm3 \n\t"
>> - "punpcklbw %%mm7, %%mm2 \n\t" // L7
>> - "punpckhbw %%mm7, %%mm3 \n\t" // H7
>> -
>> - "paddw %%mm2, %%mm2 \n\t" // 2L7
>> - "paddw %%mm3, %%mm3 \n\t" // 2H7
>> - "psubw %%mm2, %%mm0 \n\t" // 2L4 - 5L5 + 5L6
>> - 2L7
>> - "psubw %%mm3, %%mm1 \n\t" // 2H4 - 5H5 + 5H6
>> - 2H7
>> -
>> - "movq (%3), %%mm2 \n\t" // 2L0 - 5L1 + 5L2
>> - 2L3
>> - "movq 8(%3), %%mm3 \n\t" // 2H0 - 5H1 + 5H2
>> - 2H3
>> -
>> -#if TEMPLATE_PP_MMXEXT
>> - "movq %%mm7, %%mm6 \n\t" // 0
>> - "psubw %%mm0, %%mm6 \n\t"
>> - "pmaxsw %%mm6, %%mm0 \n\t" // |2L4 - 5L5 + 5L6
>> - 2L7|
>> - "movq %%mm7, %%mm6 \n\t" // 0
>> - "psubw %%mm1, %%mm6 \n\t"
>> - "pmaxsw %%mm6, %%mm1 \n\t" // |2H4 - 5H5 + 5H6
>> - 2H7|
>> - "movq %%mm7, %%mm6 \n\t" // 0
>> - "psubw %%mm2, %%mm6 \n\t"
>> - "pmaxsw %%mm6, %%mm2 \n\t" // |2L0 - 5L1 + 5L2
>> - 2L3|
>> - "movq %%mm7, %%mm6 \n\t" // 0
>> - "psubw %%mm3, %%mm6 \n\t"
>> - "pmaxsw %%mm6, %%mm3 \n\t" // |2H0 - 5H1 + 5H2
>> - 2H3|
>> -#else
>> - "movq %%mm7, %%mm6 \n\t" // 0
>> - "pcmpgtw %%mm0, %%mm6 \n\t"
>> - "pxor %%mm6, %%mm0 \n\t"
>> - "psubw %%mm6, %%mm0 \n\t" // |2L4 - 5L5 + 5L6
>> - 2L7|
>> - "movq %%mm7, %%mm6 \n\t" // 0
>> - "pcmpgtw %%mm1, %%mm6 \n\t"
>> - "pxor %%mm6, %%mm1 \n\t"
>> - "psubw %%mm6, %%mm1 \n\t" // |2H4 - 5H5 + 5H6
>> - 2H7|
>> - "movq %%mm7, %%mm6 \n\t" // 0
>> - "pcmpgtw %%mm2, %%mm6 \n\t"
>> - "pxor %%mm6, %%mm2 \n\t"
>> - "psubw %%mm6, %%mm2 \n\t" // |2L0 - 5L1 + 5L2
>> - 2L3|
>> - "movq %%mm7, %%mm6 \n\t" // 0
>> - "pcmpgtw %%mm3, %%mm6 \n\t"
>> - "pxor %%mm6, %%mm3 \n\t"
>> - "psubw %%mm6, %%mm3 \n\t" // |2H0 - 5H1 + 5H2
>> - 2H3|
>> -#endif
>> -
>> -#if TEMPLATE_PP_MMXEXT
>> - "pminsw %%mm2, %%mm0 \n\t"
>> - "pminsw %%mm3, %%mm1 \n\t"
>> -#else
>> - "movq %%mm0, %%mm6 \n\t"
>> - "psubusw %%mm2, %%mm6 \n\t"
>> - "psubw %%mm6, %%mm0 \n\t"
>> - "movq %%mm1, %%mm6 \n\t"
>> - "psubusw %%mm3, %%mm6 \n\t"
>> - "psubw %%mm6, %%mm1 \n\t"
>> -#endif
>> -
>> - "movd %2, %%mm2 \n\t" // QP
>> - "punpcklbw %%mm7, %%mm2 \n\t"
>> -
>> - "movq %%mm7, %%mm6 \n\t" // 0
>> - "pcmpgtw %%mm4, %%mm6 \n\t" // sign(2L2 - 5L3 +
>> 5L4 - 2L5)
>> - "pxor %%mm6, %%mm4 \n\t"
>> - "psubw %%mm6, %%mm4 \n\t" // |2L2 - 5L3 + 5L4
>> - 2L5|
>> - "pcmpgtw %%mm5, %%mm7 \n\t" // sign(2H2 - 5H3 +
>> 5H4 - 2H5)
>> - "pxor %%mm7, %%mm5 \n\t"
>> - "psubw %%mm7, %%mm5 \n\t" // |2H2 - 5H3 + 5H4
>> - 2H5|
>> -// 100 opcodes
>> - "psllw $3, %%mm2 \n\t" // 8QP
>> - "movq %%mm2, %%mm3 \n\t" // 8QP
>> - "pcmpgtw %%mm4, %%mm2 \n\t"
>> - "pcmpgtw %%mm5, %%mm3 \n\t"
>> - "pand %%mm2, %%mm4 \n\t"
>> - "pand %%mm3, %%mm5 \n\t"
>> -
>> -
>> - "psubusw %%mm0, %%mm4 \n\t" // hd
>> - "psubusw %%mm1, %%mm5 \n\t" // ld
>> -
>> -
>> - "movq "MANGLE(w05)", %%mm2 \n\t" // 5
>> - "pmullw %%mm2, %%mm4 \n\t"
>> - "pmullw %%mm2, %%mm5 \n\t"
>> - "movq "MANGLE(w20)", %%mm2 \n\t" // 32
>> - "paddw %%mm2, %%mm4 \n\t"
>> - "paddw %%mm2, %%mm5 \n\t"
>> - "psrlw $6, %%mm4 \n\t"
>> - "psrlw $6, %%mm5 \n\t"
>> -
>> - "movq 16(%3), %%mm0 \n\t" // L3 - L4
>> - "movq 24(%3), %%mm1 \n\t" // H3 - H4
>> -
>> - "pxor %%mm2, %%mm2 \n\t"
>> - "pxor %%mm3, %%mm3 \n\t"
>> -
>> - "pcmpgtw %%mm0, %%mm2 \n\t" // sign (L3-L4)
>> - "pcmpgtw %%mm1, %%mm3 \n\t" // sign (H3-H4)
>> - "pxor %%mm2, %%mm0 \n\t"
>> - "pxor %%mm3, %%mm1 \n\t"
>> - "psubw %%mm2, %%mm0 \n\t" // |L3-L4|
>> - "psubw %%mm3, %%mm1 \n\t" // |H3-H4|
>> - "psrlw $1, %%mm0 \n\t" // |L3 - L4|/2
>> - "psrlw $1, %%mm1 \n\t" // |H3 - H4|/2
>> -
>> - "pxor %%mm6, %%mm2 \n\t"
>> - "pxor %%mm7, %%mm3 \n\t"
>> - "pand %%mm2, %%mm4 \n\t"
>> - "pand %%mm3, %%mm5 \n\t"
>> -
>> -#if TEMPLATE_PP_MMXEXT
>> - "pminsw %%mm0, %%mm4 \n\t"
>> - "pminsw %%mm1, %%mm5 \n\t"
>> -#else
>> - "movq %%mm4, %%mm2 \n\t"
>> - "psubusw %%mm0, %%mm2 \n\t"
>> - "psubw %%mm2, %%mm4 \n\t"
>> - "movq %%mm5, %%mm2 \n\t"
>> - "psubusw %%mm1, %%mm2 \n\t"
>> - "psubw %%mm2, %%mm5 \n\t"
>> -#endif
>> - "pxor %%mm6, %%mm4 \n\t"
>> - "pxor %%mm7, %%mm5 \n\t"
>> - "psubw %%mm6, %%mm4 \n\t"
>> - "psubw %%mm7, %%mm5 \n\t"
>> - "packsswb %%mm5, %%mm4 \n\t"
>> - "movq (%0), %%mm0 \n\t"
>> - "paddb %%mm4, %%mm0 \n\t"
>> - "movq %%mm0, (%0) \n\t"
>> - "movq (%0, %1), %%mm0 \n\t"
>> - "psubb %%mm4, %%mm0 \n\t"
>> - "movq %%mm0, (%0, %1) \n\t"
>> -
>> - : "+r" (src)
>> - : "r" ((x86_reg)stride), "m" (c->pQPb), "r"(tmp)
>> - NAMED_CONSTRAINTS_ADD(w05,w20)
>> - : "%"FF_REG_a
>> - );
>> -#else //TEMPLATE_PP_MMXEXT || TEMPLATE_PP_3DNOW
>> +#else //TEMPLATE_PP_MMXEXT
>> const int l1= stride;
>> const int l2= stride + l1;
>> const int l3= stride + l2;
>> @@ -1086,14 +823,14 @@ static inline void RENAME(doVertDefFilter)(uint8_t
>> src[], int stride, PPContext
>> }
>> src++;
>> }
>> -#endif //TEMPLATE_PP_MMXEXT || TEMPLATE_PP_3DNOW
>> +#endif //TEMPLATE_PP_MMXEXT
>> }
>> #endif //TEMPLATE_PP_ALTIVEC
>>
>> #if !TEMPLATE_PP_ALTIVEC
>> static inline void RENAME(dering)(uint8_t src[], int stride, PPContext
>> *c)
>> {
>> -#if HAVE_7REGS && (TEMPLATE_PP_MMXEXT || TEMPLATE_PP_3DNOW)
>> +#if HAVE_7REGS && TEMPLATE_PP_MMXEXT
>> DECLARE_ALIGNED(8, uint64_t, tmp)[3];
>> __asm__ volatile(
>> "pxor %%mm6, %%mm6 \n\t"
>> @@ -1113,20 +850,10 @@ static inline void RENAME(dering)(uint8_t src[],
>> int stride, PPContext *c)
>>
>> #undef REAL_FIND_MIN_MAX
>> #undef FIND_MIN_MAX
>> -#if TEMPLATE_PP_MMXEXT
>> #define REAL_FIND_MIN_MAX(addr)\
>> "movq " #addr ", %%mm0 \n\t"\
>> "pminub %%mm0, %%mm7 \n\t"\
>> "pmaxub %%mm0, %%mm6 \n\t"
>> -#else
>> -#define REAL_FIND_MIN_MAX(addr)\
>> - "movq " #addr ", %%mm0 \n\t"\
>> - "movq %%mm7, %%mm1 \n\t"\
>> - "psubusb %%mm0, %%mm6 \n\t"\
>> - "paddb %%mm0, %%mm6 \n\t"\
>> - "psubusb %%mm0, %%mm1 \n\t"\
>> - "psubb %%mm1, %%mm7 \n\t"
>> -#endif
>> #define FIND_MIN_MAX(addr) REAL_FIND_MIN_MAX(addr)
>>
>> FIND_MIN_MAX((%%FF_REGa))
>> @@ -1140,49 +867,20 @@ FIND_MIN_MAX((%0, %1, 8))
>>
>> "movq %%mm7, %%mm4 \n\t"
>> "psrlq $8, %%mm7 \n\t"
>> -#if TEMPLATE_PP_MMXEXT
>> "pminub %%mm4, %%mm7 \n\t" // min of pixels
>> "pshufw $0xF9, %%mm7, %%mm4 \n\t"
>> "pminub %%mm4, %%mm7 \n\t" // min of pixels
>> "pshufw $0xFE, %%mm7, %%mm4 \n\t"
>> "pminub %%mm4, %%mm7 \n\t"
>> -#else
>> - "movq %%mm7, %%mm1 \n\t"
>> - "psubusb %%mm4, %%mm1 \n\t"
>> - "psubb %%mm1, %%mm7 \n\t"
>> - "movq %%mm7, %%mm4 \n\t"
>> - "psrlq $16, %%mm7 \n\t"
>> - "movq %%mm7, %%mm1 \n\t"
>> - "psubusb %%mm4, %%mm1 \n\t"
>> - "psubb %%mm1, %%mm7 \n\t"
>> - "movq %%mm7, %%mm4 \n\t"
>> - "psrlq $32, %%mm7 \n\t"
>> - "movq %%mm7, %%mm1 \n\t"
>> - "psubusb %%mm4, %%mm1 \n\t"
>> - "psubb %%mm1, %%mm7 \n\t"
>> -#endif
>>
>>
>> "movq %%mm6, %%mm4 \n\t"
>> "psrlq $8, %%mm6 \n\t"
>> -#if TEMPLATE_PP_MMXEXT
>> "pmaxub %%mm4, %%mm6 \n\t" // max of pixels
>> "pshufw $0xF9, %%mm6, %%mm4 \n\t"
>> "pmaxub %%mm4, %%mm6 \n\t"
>> "pshufw $0xFE, %%mm6, %%mm4 \n\t"
>> "pmaxub %%mm4, %%mm6 \n\t"
>> -#else
>> - "psubusb %%mm4, %%mm6 \n\t"
>> - "paddb %%mm4, %%mm6 \n\t"
>> - "movq %%mm6, %%mm4 \n\t"
>> - "psrlq $16, %%mm6 \n\t"
>> - "psubusb %%mm4, %%mm6 \n\t"
>> - "paddb %%mm4, %%mm6 \n\t"
>> - "movq %%mm6, %%mm4 \n\t"
>> - "psrlq $32, %%mm6 \n\t"
>> - "psubusb %%mm4, %%mm6 \n\t"
>> - "paddb %%mm4, %%mm6 \n\t"
>> -#endif
>> "movq %%mm6, %%mm0 \n\t" // max
>> "psubb %%mm7, %%mm6 \n\t" // max - min
>> "push %%"FF_REG_a" \n\t"
>> @@ -1320,7 +1018,7 @@ DERING_CORE((%0, %1, 8) ,(%%FF_REGd, %1,
>> 4),%%mm2,%%mm4,%%mm0,%%mm3,%%mm5,
>> NAMED_CONSTRAINTS_ADD(deringThreshold,b00,b02,b08)
>> : "%"FF_REG_a, "%"FF_REG_d
>> );
>> -#else // HAVE_7REGS && (TEMPLATE_PP_MMXEXT || TEMPLATE_PP_3DNOW)
>> +#else // HAVE_7REGS && TEMPLATE_PP_MMXEXT
>> int y;
>> int min=255;
>> int max=0;
>> @@ -1438,7 +1136,7 @@ DERING_CORE((%0, %1, 8) ,(%%FF_REGd, %1,
>> 4),%%mm2,%%mm4,%%mm0,%%mm3,%%mm5,
>> // src[0] = src[7]=src[stride*7]=src[stride*7 + 7]=255;
>> }
>> #endif
>> -#endif //TEMPLATE_PP_MMXEXT || TEMPLATE_PP_3DNOW
>> +#endif //TEMPLATE_PP_MMXEXT
>> }
>> #endif //TEMPLATE_PP_ALTIVEC
>>
>> @@ -1450,7 +1148,7 @@ DERING_CORE((%0, %1, 8) ,(%%FF_REGd, %1,
>> 4),%%mm2,%%mm4,%%mm0,%%mm3,%%mm5,
>> */
>> static inline void RENAME(deInterlaceInterpolateLinear)(uint8_t src[],
>> int stride)
>> {
>> -#if TEMPLATE_PP_MMXEXT || TEMPLATE_PP_3DNOW
>> +#if TEMPLATE_PP_MMXEXT
>> src+= 4*stride;
>> __asm__ volatile(
>> "lea (%0, %1), %%"FF_REG_a" \n\t"
>> @@ -1503,14 +1201,13 @@ static inline void
>> RENAME(deInterlaceInterpolateLinear)(uint8_t src[], int strid
>> */
>> static inline void RENAME(deInterlaceInterpolateCubic)(uint8_t src[], int
>> stride)
>> {
>> -#if TEMPLATE_PP_SSE2 || TEMPLATE_PP_MMXEXT || TEMPLATE_PP_3DNOW
>> +#if TEMPLATE_PP_SSE2
>> src+= stride*3;
>> __asm__ volatile(
>> "lea (%0, %1), %%"FF_REG_a" \n\t"
>> "lea (%%"FF_REG_a", %1, 4), %%"FF_REG_d"\n\t"
>> "lea (%%"FF_REG_d", %1, 4), %%"FF_REG_c"\n\t"
>> "add %1, %%"FF_REG_c" \n\t"
>> -#if TEMPLATE_PP_SSE2
>> "pxor %%xmm7, %%xmm7 \n\t"
>> #define REAL_DEINT_CUBIC(a,b,c,d,e)\
>> "movq " #a ", %%xmm0 \n\t"\
>> @@ -1526,33 +1223,6 @@ static inline void
>> RENAME(deInterlaceInterpolateCubic)(uint8_t src[], int stride
>> "psubw %%xmm0, %%xmm1 \n\t"\
>> "packuswb %%xmm1, %%xmm1 \n\t"\
>> "movlps %%xmm1, " #c " \n\t"
>> -#else //TEMPLATE_PP_SSE2
>> - "pxor %%mm7, %%mm7 \n\t"
>> -// 0 1 2 3 4 5 6 7 8
>> 9 10
>> -// %0 eax eax+%1 eax+2%1 %0+4%1 edx edx+%1 edx+2%1
>> %0+8%1 edx+4%1 ecx
>> -
>> -#define REAL_DEINT_CUBIC(a,b,c,d,e)\
>> - "movq " #a ", %%mm0 \n\t"\
>> - "movq " #b ", %%mm1 \n\t"\
>> - "movq " #d ", %%mm2 \n\t"\
>> - "movq " #e ", %%mm3 \n\t"\
>> - PAVGB(%%mm2, %%mm1) /* (b+d) /2 */\
>> - PAVGB(%%mm3, %%mm0) /* (a+e) /2 */\
>> - "movq %%mm0, %%mm2 \n\t"\
>> - "punpcklbw %%mm7, %%mm0 \n\t"\
>> - "punpckhbw %%mm7, %%mm2 \n\t"\
>> - "movq %%mm1, %%mm3 \n\t"\
>> - "punpcklbw %%mm7, %%mm1 \n\t"\
>> - "punpckhbw %%mm7, %%mm3 \n\t"\
>> - "psubw %%mm1, %%mm0 \n\t" /* L(a+e -
>> (b+d))/2 */\
>> - "psubw %%mm3, %%mm2 \n\t" /* H(a+e -
>> (b+d))/2 */\
>> - "psraw $3, %%mm0 \n\t" /* L(a+e -
>> (b+d))/16 */\
>> - "psraw $3, %%mm2 \n\t" /* H(a+e -
>> (b+d))/16 */\
>> - "psubw %%mm0, %%mm1 \n\t" /* L(9b + 9d - a
>> - e)/16 */\
>> - "psubw %%mm2, %%mm3 \n\t" /* H(9b + 9d - a
>> - e)/16 */\
>> - "packuswb %%mm3, %%mm1 \n\t"\
>> - "movq %%mm1, " #c " \n\t"
>> -#endif //TEMPLATE_PP_SSE2
>> #define DEINT_CUBIC(a,b,c,d,e) REAL_DEINT_CUBIC(a,b,c,d,e)
>>
>> DEINT_CUBIC((%0) , (%%FF_REGa, %1), (%%FF_REGa, %1, 2), (%0,
>> %1, 4) , (%%FF_REGd, %1))
>> @@ -1562,13 +1232,11 @@ DEINT_CUBIC((%%FF_REGd, %1), (%0, %1, 8) ,
>> (%%FF_REGd, %1, 4), (%%FF_REGc)
>>
>> : : "r" (src), "r" ((x86_reg)stride)
>> :
>> -#if TEMPLATE_PP_SSE2
>> XMM_CLOBBERS("%xmm0", "%xmm1", "%xmm2", "%xmm3", "%xmm7",)
>> -#endif
>> "%"FF_REG_a, "%"FF_REG_d, "%"FF_REG_c
>> );
>> #undef REAL_DEINT_CUBIC
>> -#else //TEMPLATE_PP_SSE2 || TEMPLATE_PP_MMXEXT || TEMPLATE_PP_3DNOW
>> +#else //TEMPLATE_PP_SSE2
>> int x;
>> src+= stride*3;
>> for(x=0; x<8; x++){
>> @@ -1578,7 +1246,7 @@ DEINT_CUBIC((%%FF_REGd, %1), (%0, %1, 8) ,
>> (%%FF_REGd, %1, 4), (%%FF_REGc)
>> src[stride*9] = av_clip_uint8((-src[stride*6] + 9*src[stride*8] +
>> 9*src[stride*10] - src[stride*12])>>4);
>> src++;
>> }
>> -#endif //TEMPLATE_PP_SSE2 || TEMPLATE_PP_MMXEXT || TEMPLATE_PP_3DNOW
>> +#endif //TEMPLATE_PP_SSE2
>> }
>>
>> /**
>> @@ -1590,7 +1258,7 @@ DEINT_CUBIC((%%FF_REGd, %1), (%0, %1, 8) ,
>> (%%FF_REGd, %1, 4), (%%FF_REGc)
>> */
>> static inline void RENAME(deInterlaceFF)(uint8_t src[], int stride,
>> uint8_t *tmp)
>> {
>> -#if TEMPLATE_PP_MMXEXT || TEMPLATE_PP_3DNOW
>> +#if TEMPLATE_PP_MMXEXT
>> src+= stride*4;
>> __asm__ volatile(
>> "lea (%0, %1), %%"FF_REG_a" \n\t"
>> @@ -1639,7 +1307,7 @@ DEINT_FF((%%FF_REGd, %1), (%%FF_REGd, %1, 2), (%0,
>> %1, 8) , (%%FF_REGd, %1, 4
>> : : "r" (src), "r" ((x86_reg)stride), "r"(tmp)
>> : "%"FF_REG_a, "%"FF_REG_d
>> );
>> -#else //TEMPLATE_PP_MMXEXT || TEMPLATE_PP_3DNOW
>> +#else //TEMPLATE_PP_MMXEXT
>> int x;
>> src+= stride*4;
>> for(x=0; x<8; x++){
>> @@ -1657,7 +1325,7 @@ DEINT_FF((%%FF_REGd, %1), (%%FF_REGd, %1, 2), (%0,
>> %1, 8) , (%%FF_REGd, %1, 4
>>
>> src++;
>> }
>> -#endif //TEMPLATE_PP_MMXEXT || TEMPLATE_PP_3DNOW
>> +#endif //TEMPLATE_PP_MMXEXT
>> }
>>
>> /**
>> @@ -1669,7 +1337,7 @@ DEINT_FF((%%FF_REGd, %1), (%%FF_REGd, %1, 2), (%0,
>> %1, 8) , (%%FF_REGd, %1, 4
>> */
>> static inline void RENAME(deInterlaceL5)(uint8_t src[], int stride,
>> uint8_t *tmp, uint8_t *tmp2)
>> {
>> -#if (TEMPLATE_PP_MMXEXT || TEMPLATE_PP_3DNOW) && HAVE_6REGS
>> +#if TEMPLATE_PP_MMXEXT && HAVE_6REGS
>> src+= stride*4;
>> __asm__ volatile(
>> "lea (%0, %1), %%"FF_REG_a" \n\t"
>> @@ -1729,7 +1397,7 @@ DEINT_L5(%%mm1, %%mm0, (%%FF_REGd, %1, 2), (%0, %1,
>> 8) , (%%FF_REGd, %1, 4
>> : : "r" (src), "r" ((x86_reg)stride), "r"(tmp), "r"(tmp2)
>> : "%"FF_REG_a, "%"FF_REG_d
>> );
>> -#else //(TEMPLATE_PP_MMXEXT || TEMPLATE_PP_3DNOW) && HAVE_6REGS
>> +#else //TEMPLATE_PP_MMXEXT && HAVE_6REGS
>> int x;
>> src+= stride*4;
>> for(x=0; x<8; x++){
>> @@ -1758,7 +1426,7 @@ DEINT_L5(%%mm1, %%mm0, (%%FF_REGd, %1, 2), (%0, %1,
>> 8) , (%%FF_REGd, %1, 4
>>
>> src++;
>> }
>> -#endif //(TEMPLATE_PP_MMXEXT || TEMPLATE_PP_3DNOW) && HAVE_6REGS
>> +#endif // TEMPLATE_PP_MMXEXT && HAVE_6REGS
>> }
>>
>> /**
>> @@ -1770,7 +1438,7 @@ DEINT_L5(%%mm1, %%mm0, (%%FF_REGd, %1, 2), (%0, %1,
>> 8) , (%%FF_REGd, %1, 4
>> */
>> static inline void RENAME(deInterlaceBlendLinear)(uint8_t src[], int
>> stride, uint8_t *tmp)
>> {
>> -#if TEMPLATE_PP_MMXEXT || TEMPLATE_PP_3DNOW
>> +#if TEMPLATE_PP_MMXEXT
>> src+= 4*stride;
>> __asm__ volatile(
>> "lea (%0, %1), %%"FF_REG_a" \n\t"
>> @@ -1817,7 +1485,7 @@ static inline void
>> RENAME(deInterlaceBlendLinear)(uint8_t src[], int stride, uin
>> : : "r" (src), "r" ((x86_reg)stride), "r" (tmp)
>> : "%"FF_REG_a, "%"FF_REG_d
>> );
>> -#else //TEMPLATE_PP_MMXEXT || TEMPLATE_PP_3DNOW
>> +#else //TEMPLATE_PP_MMXEXT
>> int a, b, c, x;
>> src+= 4*stride;
>>
>> @@ -1860,7 +1528,7 @@ static inline void
>> RENAME(deInterlaceBlendLinear)(uint8_t src[], int stride, uin
>> src += 4;
>> tmp += 4;
>> }
>> -#endif //TEMPLATE_PP_MMXEXT || TEMPLATE_PP_3DNOW
>> +#endif //TEMPLATE_PP_MMXEXT
>> }
>>
>> /**
>> @@ -1871,9 +1539,8 @@ static inline void
>> RENAME(deInterlaceBlendLinear)(uint8_t src[], int stride, uin
>> */
>> static inline void RENAME(deInterlaceMedian)(uint8_t src[], int stride)
>> {
>> -#if TEMPLATE_PP_MMX
>> - src+= 4*stride;
>> #if TEMPLATE_PP_MMXEXT
>> + src+= 4*stride;
>> __asm__ volatile(
>> "lea (%0, %1), %%"FF_REG_a" \n\t"
>> "lea (%%"FF_REG_a", %1, 4), %%"FF_REG_d"\n\t"
>> @@ -1922,48 +1589,6 @@ static inline void
>> RENAME(deInterlaceMedian)(uint8_t src[], int stride)
>> : "%"FF_REG_a, "%"FF_REG_d
>> );
>>
>> -#else // MMX without MMX2
>> - __asm__ volatile(
>> - "lea (%0, %1), %%"FF_REG_a" \n\t"
>> - "lea (%%"FF_REG_a", %1, 4), %%"FF_REG_d"\n\t"
>> -// 0 1 2 3 4 5 6 7 8
>> 9
>> -// %0 eax eax+%1 eax+2%1 %0+4%1 edx edx+%1 edx+2%1
>> %0+8%1 edx+4%1
>> - "pxor %%mm7, %%mm7 \n\t"
>> -
>> -#define REAL_MEDIAN(a,b,c)\
>> - "movq " #a ", %%mm0 \n\t"\
>> - "movq " #b ", %%mm2 \n\t"\
>> - "movq " #c ", %%mm1 \n\t"\
>> - "movq %%mm0, %%mm3 \n\t"\
>> - "movq %%mm1, %%mm4 \n\t"\
>> - "movq %%mm2, %%mm5 \n\t"\
>> - "psubusb %%mm1, %%mm3 \n\t"\
>> - "psubusb %%mm2, %%mm4 \n\t"\
>> - "psubusb %%mm0, %%mm5 \n\t"\
>> - "pcmpeqb %%mm7, %%mm3 \n\t"\
>> - "pcmpeqb %%mm7, %%mm4 \n\t"\
>> - "pcmpeqb %%mm7, %%mm5 \n\t"\
>> - "movq %%mm3, %%mm6 \n\t"\
>> - "pxor %%mm4, %%mm3 \n\t"\
>> - "pxor %%mm5, %%mm4 \n\t"\
>> - "pxor %%mm6, %%mm5 \n\t"\
>> - "por %%mm3, %%mm1 \n\t"\
>> - "por %%mm4, %%mm2 \n\t"\
>> - "por %%mm5, %%mm0 \n\t"\
>> - "pand %%mm2, %%mm0 \n\t"\
>> - "pand %%mm1, %%mm0 \n\t"\
>> - "movq %%mm0, " #b " \n\t"
>> -#define MEDIAN(a,b,c) REAL_MEDIAN(a,b,c)
>> -
>> -MEDIAN((%0) , (%%FF_REGa) , (%%FF_REGa, %1))
>> -MEDIAN((%%FF_REGa, %1), (%%FF_REGa, %1, 2), (%0, %1, 4))
>> -MEDIAN((%0, %1, 4) , (%%FF_REGd) , (%%FF_REGd, %1))
>> -MEDIAN((%%FF_REGd, %1), (%%FF_REGd, %1, 2), (%0, %1, 8))
>> -
>> - : : "r" (src), "r" ((x86_reg)stride)
>> - : "%"FF_REG_a, "%"FF_REG_d
>> - );
>> -#endif //TEMPLATE_PP_MMXEXT
>> #else //TEMPLATE_PP_MMX
>> int x, y;
>> src+= 4*stride;
>> @@ -2165,7 +1790,7 @@ static inline void RENAME(tempNoiseReducer)(uint8_t
>> *src, int stride,
>>
>> #define FAST_L2_DIFF
>> //#define L1_DIFF //u should change the thresholds too if u try that one
>> -#if (TEMPLATE_PP_MMXEXT || TEMPLATE_PP_3DNOW) && HAVE_6REGS
>> +#if TEMPLATE_PP_MMXEXT && HAVE_6REGS
>> __asm__ volatile(
>> "lea (%2, %2, 2), %%"FF_REG_a" \n\t" // 3*stride
>> "lea (%2, %2, 4), %%"FF_REG_d" \n\t" // 5*stride
>> @@ -2454,7 +2079,7 @@ L2_DIFF_CORE((%0, %%FF_REGc) , (%1, %%FF_REGc))
>> NAMED_CONSTRAINTS_ADD(b80)
>> : "%"FF_REG_a, "%"FF_REG_d, "%"FF_REG_c, "memory"
>> );
>> -#else //(TEMPLATE_PP_MMXEXT || TEMPLATE_PP_3DNOW) && HAVE_6REGS
>> +#else //TEMPLATE_PP_MMXEXT && HAVE_6REGS
>> {
>> int y;
>> int d=0;
>> @@ -2537,11 +2162,11 @@ Switch between
>> }
>> }
>> }
>> -#endif //(TEMPLATE_PP_MMXEXT || TEMPLATE_PP_3DNOW) && HAVE_6REGS
>> +#endif //TEMPLATE_PP_MMXEXT && HAVE_6REGS
>> }
>> #endif //TEMPLATE_PP_ALTIVEC
>>
>> -#if TEMPLATE_PP_MMX
>> +#if TEMPLATE_PP_MMXEXT
>> /**
>> * accurate deblock filter
>> */
>> @@ -2945,7 +2570,6 @@ static av_always_inline void
>> RENAME(do_a_deblock)(uint8_t *src, int step, int st
>> "movq (%4), %%mm2 \n\t" // 2L0 - 5L1 +
>> 5L2 - 2L3
>> "movq 8(%4), %%mm3 \n\t" // 2H0 - 5H1 +
>> 5H2 - 2H3
>>
>> -#if TEMPLATE_PP_MMXEXT
>> "movq %%mm7, %%mm6 \n\t" // 0
>> "psubw %%mm0, %%mm6 \n\t"
>> "pmaxsw %%mm6, %%mm0 \n\t" // |2L4 - 5L5 +
>> 5L6 - 2L7|
>> @@ -2958,36 +2582,9 @@ static av_always_inline void
>> RENAME(do_a_deblock)(uint8_t *src, int step, int st
>> "movq %%mm7, %%mm6 \n\t" // 0
>> "psubw %%mm3, %%mm6 \n\t"
>> "pmaxsw %%mm6, %%mm3 \n\t" // |2H0 - 5H1 +
>> 5H2 - 2H3|
>> -#else
>> - "movq %%mm7, %%mm6 \n\t" // 0
>> - "pcmpgtw %%mm0, %%mm6 \n\t"
>> - "pxor %%mm6, %%mm0 \n\t"
>> - "psubw %%mm6, %%mm0 \n\t" // |2L4 - 5L5 +
>> 5L6 - 2L7|
>> - "movq %%mm7, %%mm6 \n\t" // 0
>> - "pcmpgtw %%mm1, %%mm6 \n\t"
>> - "pxor %%mm6, %%mm1 \n\t"
>> - "psubw %%mm6, %%mm1 \n\t" // |2H4 - 5H5 +
>> 5H6 - 2H7|
>> - "movq %%mm7, %%mm6 \n\t" // 0
>> - "pcmpgtw %%mm2, %%mm6 \n\t"
>> - "pxor %%mm6, %%mm2 \n\t"
>> - "psubw %%mm6, %%mm2 \n\t" // |2L0 - 5L1 +
>> 5L2 - 2L3|
>> - "movq %%mm7, %%mm6 \n\t" // 0
>> - "pcmpgtw %%mm3, %%mm6 \n\t"
>> - "pxor %%mm6, %%mm3 \n\t"
>> - "psubw %%mm6, %%mm3 \n\t" // |2H0 - 5H1 +
>> 5H2 - 2H3|
>> -#endif
>>
>> -#if TEMPLATE_PP_MMXEXT
>> "pminsw %%mm2, %%mm0 \n\t"
>> "pminsw %%mm3, %%mm1 \n\t"
>> -#else
>> - "movq %%mm0, %%mm6 \n\t"
>> - "psubusw %%mm2, %%mm6 \n\t"
>> - "psubw %%mm6, %%mm0 \n\t"
>> - "movq %%mm1, %%mm6 \n\t"
>> - "psubusw %%mm3, %%mm6 \n\t"
>> - "psubw %%mm6, %%mm1 \n\t"
>> -#endif
>>
>> "movd %2, %%mm2 \n\t" // QP
>> "punpcklbw %%mm7, %%mm2 \n\t"
>> @@ -3041,17 +2638,8 @@ static av_always_inline void
>> RENAME(do_a_deblock)(uint8_t *src, int step, int st
>> "pand %%mm2, %%mm4 \n\t"
>> "pand %%mm3, %%mm5 \n\t"
>>
>> -#if TEMPLATE_PP_MMXEXT
>> "pminsw %%mm0, %%mm4 \n\t"
>> "pminsw %%mm1, %%mm5 \n\t"
>> -#else
>> - "movq %%mm4, %%mm2 \n\t"
>> - "psubusw %%mm0, %%mm2 \n\t"
>> - "psubw %%mm2, %%mm4 \n\t"
>> - "movq %%mm5, %%mm2 \n\t"
>> - "psubusw %%mm1, %%mm2 \n\t"
>> - "psubw %%mm2, %%mm5 \n\t"
>> -#endif
>> "pxor %%mm6, %%mm4 \n\t"
>> "pxor %%mm7, %%mm5 \n\t"
>> "psubw %%mm6, %%mm4 \n\t"
>> @@ -3088,18 +2676,14 @@ static void RENAME(postProcess)(const uint8_t
>> src[], int srcStride, uint8_t dst[
>> static inline void RENAME(blockCopy)(uint8_t dst[], int dstStride, const
>> uint8_t src[], int srcStride,
>> int levelFix, int64_t
>> *packedOffsetAndScale)
>> {
>> -#if !TEMPLATE_PP_MMX || !HAVE_6REGS
>> - int i;
>> -#endif
>> if(levelFix){
>> -#if TEMPLATE_PP_MMX && HAVE_6REGS
>> +#if TEMPLATE_PP_MMXEXT && HAVE_6REGS
>> __asm__ volatile(
>> "movq (%%"FF_REG_a"), %%mm2 \n\t" // packedYOffset
>> "movq 8(%%"FF_REG_a"), %%mm3 \n\t" // packedYScale
>> "lea (%2,%4), %%"FF_REG_a" \n\t"
>> "lea (%3,%5), %%"FF_REG_d" \n\t"
>> "pxor %%mm4, %%mm4 \n\t"
>> -#if TEMPLATE_PP_MMXEXT
>> #define REAL_SCALED_CPY(src1, src2, dst1, dst2)
>> \
>> "movq " #src1 ", %%mm0 \n\t"\
>> "movq " #src1 ", %%mm5 \n\t"\
>> @@ -3122,34 +2706,6 @@ static inline void RENAME(blockCopy)(uint8_t dst[],
>> int dstStride, const uint8_t
>> "movq %%mm0, " #dst1 " \n\t"\
>> "movq %%mm1, " #dst2 " \n\t"\
>>
>> -#else //TEMPLATE_PP_MMXEXT
>> -#define REAL_SCALED_CPY(src1, src2, dst1, dst2)
>> \
>> - "movq " #src1 ", %%mm0 \n\t"\
>> - "movq " #src1 ", %%mm5 \n\t"\
>> - "punpcklbw %%mm4, %%mm0 \n\t"\
>> - "punpckhbw %%mm4, %%mm5 \n\t"\
>> - "psubw %%mm2, %%mm0 \n\t"\
>> - "psubw %%mm2, %%mm5 \n\t"\
>> - "movq " #src2 ", %%mm1 \n\t"\
>> - "psllw $6, %%mm0 \n\t"\
>> - "psllw $6, %%mm5 \n\t"\
>> - "pmulhw %%mm3, %%mm0 \n\t"\
>> - "movq " #src2 ", %%mm6 \n\t"\
>> - "pmulhw %%mm3, %%mm5 \n\t"\
>> - "punpcklbw %%mm4, %%mm1 \n\t"\
>> - "punpckhbw %%mm4, %%mm6 \n\t"\
>> - "psubw %%mm2, %%mm1 \n\t"\
>> - "psubw %%mm2, %%mm6 \n\t"\
>> - "psllw $6, %%mm1 \n\t"\
>> - "psllw $6, %%mm6 \n\t"\
>> - "pmulhw %%mm3, %%mm1 \n\t"\
>> - "pmulhw %%mm3, %%mm6 \n\t"\
>> - "packuswb %%mm5, %%mm0 \n\t"\
>> - "packuswb %%mm6, %%mm1 \n\t"\
>> - "movq %%mm0, " #dst1 " \n\t"\
>> - "movq %%mm1, " #dst2 " \n\t"\
>> -
>> -#endif //TEMPLATE_PP_MMXEXT
>> #define SCALED_CPY(src1, src2, dst1, dst2)\
>> REAL_SCALED_CPY(src1, src2, dst1, dst2)
>>
>> @@ -3170,7 +2726,7 @@ SCALED_CPY((%%FF_REGa, %4), (%%FF_REGa, %4, 2),
>> (%%FF_REGd, %5), (%%FF_REGd, %5,
>> : "%"FF_REG_d
>> );
>> #else //TEMPLATE_PP_MMX && HAVE_6REGS
>> - for(i=0; i<8; i++)
>> + for (int i = 0; i < 8; i++)
>> memcpy( &(dst[dstStride*i]),
>> &(src[srcStride*i]), BLOCK_SIZE);
>> #endif //TEMPLATE_PP_MMX && HAVE_6REGS
>> @@ -3203,7 +2759,7 @@ SIMPLE_CPY((%%FF_REGa, %2), (%%FF_REGa, %2, 2),
>> (%%FF_REGd, %3), (%%FF_REGd, %3,
>> : "%"FF_REG_a, "%"FF_REG_d
>> );
>> #else //TEMPLATE_PP_MMX && HAVE_6REGS
>> - for(i=0; i<8; i++)
>> + for (int i = 0; i < 8; i++)
>> memcpy( &(dst[dstStride*i]),
>> &(src[srcStride*i]), BLOCK_SIZE);
>> #endif //TEMPLATE_PP_MMX && HAVE_6REGS
>> @@ -3696,9 +3252,7 @@ static void RENAME(postProcess)(const uint8_t src[],
>> int srcStride, uint8_t dst[
>> }
>> }
>> }
>> -#if TEMPLATE_PP_3DNOW
>> - __asm__ volatile("femms");
>> -#elif TEMPLATE_PP_MMX
>> +#if TEMPLATE_PP_MMX
>> __asm__ volatile("emms");
>> #endif
>>
>> @@ -3734,5 +3288,4 @@ static void RENAME(postProcess)(const uint8_t src[],
>> int srcStride, uint8_t dst[
>> #undef TEMPLATE_PP_ALTIVEC
>> #undef TEMPLATE_PP_MMX
>> #undef TEMPLATE_PP_MMXEXT
>> -#undef TEMPLATE_PP_3DNOW
>> #undef TEMPLATE_PP_SSE2
>
> Will apply this patchset tomorrow unless there are objections.
>
LGTM
> - Andreas
>
> _______________________________________________
> ffmpeg-devel mailing list
> ffmpeg-devel at ffmpeg.org
> https://ffmpeg.org/mailman/listinfo/ffmpeg-devel
>
> To unsubscribe, visit link above, or email
> ffmpeg-devel-request at ffmpeg.org with subject "unsubscribe".
>
More information about the ffmpeg-devel
mailing list