[FFmpeg-devel] [RFC] use av_get_cpu_flags for real runtime CPU detection in swscale
Måns Rullgård
mans
Wed Sep 8 18:24:02 CEST 2010
Janne Grunau <janne-ffmpeg at jannau.net> writes:
> Hi,
>
> attached patch implements runtime CPU detection in libswscale. One minor
> problem is that it changes behaviour on existing code which sets
> individual flags but obviously not the new SWS_CPU_CAPS_FORCE. I think
> it's acceptable since the flags have no effect with
> !CONFIG_RUNTIME_CPUDETECT.
>
> Janne
> diff --git a/swscale.h b/swscale.h
> index 4e11c9a..ca63796 100644
> --- a/swscale.h
> +++ b/swscale.h
> @@ -30,7 +30,7 @@
> #include "libavutil/avutil.h"
>
> #define LIBSWSCALE_VERSION_MAJOR 0
> -#define LIBSWSCALE_VERSION_MINOR 11
> +#define LIBSWSCALE_VERSION_MINOR 12
> #define LIBSWSCALE_VERSION_MICRO 0
>
> #define LIBSWSCALE_VERSION_INT AV_VERSION_INT(LIBSWSCALE_VERSION_MAJOR, \
> @@ -93,6 +93,7 @@ const char *swscale_license(void);
> #define SWS_CPU_CAPS_ALTIVEC 0x10000000
> #define SWS_CPU_CAPS_BFIN 0x01000000
> #define SWS_CPU_CAPS_SSE2 0x02000000
> +#define SWS_CPU_CAPS_FORCE 0x00100000
What does the force flag mean?
> #define SWS_MAX_REDUCE_CUTOFF 0.002
>
> diff --git a/utils.c b/utils.c
> index e9400f8..9375489 100644
> --- a/utils.c
> +++ b/utils.c
> @@ -44,6 +44,7 @@
> #include "libavutil/avutil.h"
> #include "libavutil/bswap.h"
> #include "libavutil/pixdesc.h"
> +#include "libavutil/cpu.h"
>
> unsigned swscale_version(void)
> {
> @@ -722,7 +723,26 @@ static int handle_jpeg(enum PixelFormat *format)
>
> static int update_flags_cpu(int flags)
> {
> -#if !CONFIG_RUNTIME_CPUDETECT //ensure that the flags match the compiled variant if cpudetect is off
> +#if CONFIG_RUNTIME_CPUDETECT
> + int cpuflags;
> +
> + if (!(flags & SWS_CPU_CAPS_FORCE)) {
> + flags &= ~(SWS_CPU_CAPS_MMX|SWS_CPU_CAPS_MMX2|SWS_CPU_CAPS_3DNOW|SWS_CPU_CAPS_ALTIVEC|SWS_CPU_CAPS_BFIN);
> +
> + cpuflags = av_get_cpu_flags();
> +
> + if (ARCH_X86 && cpuflags & AV_CPU_FLAG_MMX)
> + flags |= SWS_CPU_CAPS_MMX;
> + if (ARCH_X86 && cpuflags & AV_CPU_FLAG_MMX2)
> + flags |= SWS_CPU_CAPS_MMX2;
> + if (ARCH_X86 && cpuflags & AV_CPU_FLAG_3DNOW)
> + flags |= SWS_CPU_CAPS_3DNOW;;
> + if (ARCH_X86 && cpuflags & AV_CPU_FLAG_SSE2)
> + flags |= SWS_CPU_CAPS_SSE2;
> + if (ARCH_PPC && cpuflags & AV_CPU_FLAG_ALTIVEC)
> + flags |= SWS_CPU_CAPS_ALTIVEC;
Why not change libswscale to use the AV_CPU_FLAG_ values directly?
That would avoid this mess entirely.
--
M?ns Rullg?rd
mans at mansr.com
More information about the ffmpeg-devel
mailing list