[FFmpeg-devel] [PATCH V2 2/3] avcodec: add av1 hardware acceleration decoder

James Almer jamrial at gmail.com
Wed Sep 2 18:18:48 EEST 2020


On 9/2/2020 4:31 AM, Fei Wang wrote:
> This av1 decoder is now only used for av1 hardware acceleration
> decoder. Consider it can be extend to a local decoder like hevc
> or vp9 in the future, so define its name as "av1" and put it into
> external libraries codec list.
> 
> Signed-off-by: Fei Wang <fei.w.wang at intel.com>
> ---
>  Changelog              |   1 +
>  configure              |   1 +
>  libavcodec/Makefile    |   1 +
>  libavcodec/allcodecs.c |   1 +
>  libavcodec/av1dec.c    | 728 +++++++++++++++++++++++++++++++++++++++++
>  libavcodec/av1dec.h    |  70 ++++
>  libavcodec/version.h   |   4 +-
>  7 files changed, 804 insertions(+), 2 deletions(-)
>  create mode 100644 libavcodec/av1dec.c
>  create mode 100644 libavcodec/av1dec.h
> 
> diff --git a/Changelog b/Changelog
> index 7467e73306..dd5d9ca50b 100644
> --- a/Changelog
> +++ b/Changelog
> @@ -15,6 +15,7 @@ version <next>:
>  - Argonaut Games ASF muxer
>  - AV1 Low overhead bitstream format demuxer
>  - RPZA video encoder
> +- AV1 decoder (Hardware acceleration used only)
>  
>  
>  version 4.3:
> diff --git a/configure b/configure
> index 564072053b..a694d5a31d 100755
> --- a/configure
> +++ b/configure
> @@ -2685,6 +2685,7 @@ atrac3al_decoder_select="mdct"
>  atrac3p_decoder_select="mdct sinewin"
>  atrac3pal_decoder_select="mdct sinewin"
>  atrac9_decoder_select="mdct"
> +av1_decoder_select="cbs_av1"
>  avrn_decoder_select="exif jpegtables"
>  bink_decoder_select="blockdsp hpeldsp"
>  binkaudio_dct_decoder_select="mdct rdft dct sinewin wma_freqs"
> diff --git a/libavcodec/Makefile b/libavcodec/Makefile
> index ca2e8a2530..8da66e2a84 100644
> --- a/libavcodec/Makefile
> +++ b/libavcodec/Makefile
> @@ -765,6 +765,7 @@ OBJS-$(CONFIG_ZLIB_DECODER)            += lcldec.o
>  OBJS-$(CONFIG_ZLIB_ENCODER)            += lclenc.o
>  OBJS-$(CONFIG_ZMBV_DECODER)            += zmbv.o
>  OBJS-$(CONFIG_ZMBV_ENCODER)            += zmbvenc.o
> +OBJS-$(CONFIG_AV1_DECODER)             += av1dec.o
>  
>  # (AD)PCM decoders/encoders
>  OBJS-$(CONFIG_PCM_ALAW_DECODER)           += pcm.o
> diff --git a/libavcodec/allcodecs.c b/libavcodec/allcodecs.c
> index 729d2fd9ad..030a4c5511 100644
> --- a/libavcodec/allcodecs.c
> +++ b/libavcodec/allcodecs.c
> @@ -762,6 +762,7 @@ extern AVCodec ff_idf_decoder;
>   * above is available */
>  extern AVCodec ff_aac_mf_encoder;
>  extern AVCodec ff_ac3_mf_encoder;
> +extern AVCodec ff_av1_decoder;
>  extern AVCodec ff_h263_v4l2m2m_encoder;
>  extern AVCodec ff_libaom_av1_decoder;
>  extern AVCodec ff_libopenh264_encoder;
> diff --git a/libavcodec/av1dec.c b/libavcodec/av1dec.c
> new file mode 100644
> index 0000000000..6d898eda97
> --- /dev/null
> +++ b/libavcodec/av1dec.c
> @@ -0,0 +1,728 @@
> +/*
> + * AV1video decoder
> + *
> + * This file is part of FFmpeg.
> + *
> + * FFmpeg is free software; you can redistribute it and/or
> + * modify it under the terms of the GNU Lesser General Public
> + * License as published by the Free Software Foundation; either
> + * version 2.1 of the License, or (at your option) any later version.
> + *
> + * FFmpeg is distributed in the hope that it will be useful,
> + * but WITHOUT ANY WARRANTY; without even the implied warranty of
> + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
> + * Lesser General Public License for more details.
> + *
> + * You should have received a copy of the GNU Lesser General Public
> + * License along with FFmpeg; if not, write to the Free Software
> + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
> + */
> +
> +
> +#include "get_bits.h"
> +#include "hwconfig.h"
> +#include "internal.h"
> +#include "profiles.h"
> +#include "libavutil/pixdesc.h"
> +#include "av1dec.h"
> +
> +static void setup_past_independence(AV1Frame *f)
> +{
> +    f->loop_filter_delta_enabled = 1;
> +
> +    f->loop_filter_ref_deltas[AV1_REF_FRAME_INTRA] = 1;
> +    f->loop_filter_ref_deltas[AV1_REF_FRAME_LAST] = 0;
> +    f->loop_filter_ref_deltas[AV1_REF_FRAME_LAST2] = 0;
> +    f->loop_filter_ref_deltas[AV1_REF_FRAME_LAST3] = 0;
> +    f->loop_filter_ref_deltas[AV1_REF_FRAME_GOLDEN] = -1;
> +    f->loop_filter_ref_deltas[AV1_REF_FRAME_BWDREF] = 0;
> +    f->loop_filter_ref_deltas[AV1_REF_FRAME_ALTREF2] = -1;
> +    f->loop_filter_ref_deltas[AV1_REF_FRAME_ALTREF] = -1;
> +
> +    f->loop_filter_mode_deltas[0] = 0;
> +    f->loop_filter_mode_deltas[1] = 0;
> +}
> +
> +static void load_previous_and_update(AV1DecContext *s)
> +{
> +    uint8_t primary_frame, prev_frame;
> +
> +    primary_frame = s->raw_frame_header->primary_ref_frame;
> +    prev_frame = s->raw_frame_header->ref_frame_idx[primary_frame];
> +    memcpy(s->cur_frame.loop_filter_ref_deltas,
> +           s->ref[prev_frame].loop_filter_ref_deltas,
> +           AV1_NUM_REF_FRAMES * sizeof(int8_t));
> +    memcpy(s->cur_frame.loop_filter_mode_deltas,
> +           s->ref[prev_frame].loop_filter_mode_deltas,
> +           2 * sizeof(int8_t));
> +
> +    if (s->raw_frame_header->loop_filter_delta_update) {
> +        for (int i = 0; i < AV1_NUM_REF_FRAMES; i++) {
> +            if (s->raw_frame_header->update_ref_delta[i])
> +                s->cur_frame.loop_filter_ref_deltas[i] =
> +                    s->raw_frame_header->loop_filter_ref_deltas[i];
> +        }
> +
> +        for (int i = 0; i < 2; i++) {
> +            if (s->raw_frame_header->update_mode_delta[i])
> +                s->cur_frame.loop_filter_mode_deltas[i] =
> +                    s->raw_frame_header->loop_filter_mode_deltas[i];
> +        }
> +    }
> +
> +    s->cur_frame.loop_filter_delta_enabled =
> +        s->raw_frame_header->loop_filter_delta_enabled;
> +}
> +
> +static uint32_t inverse_recenter(int r, uint32_t v)
> +{
> +    if (v > 2 * r)
> +        return v;
> +    else if (v & 1)
> +        return r - ((v + 1) >> 1);
> +    else
> +        return r + (v >> 1);
> +}
> +
> +static uint32_t decode_unsigned_subexp_with_ref(uint32_t sub_exp,
> +                                                int mx, int r)
> +{
> +    if ((r << 1) <= mx) {
> +        return inverse_recenter(r, sub_exp);
> +    } else {
> +        return mx - 1 - inverse_recenter(mx - 1 - r, sub_exp);
> +    }
> +}
> +
> +static int32_t decode_signed_subexp_with_ref(uint32_t sub_exp, int low,
> +                                             int high, int r)
> +{
> +    int32_t x = decode_unsigned_subexp_with_ref(sub_exp, high - low, r - low);
> +    return x + low;
> +}
> +
> +static void read_global_param(AV1DecContext *s, int type, int ref, int idx)
> +{
> +    uint8_t primary_frame, prev_frame;
> +    uint32_t abs_bits, prec_bits, round, prec_diff, sub, mx;
> +    int32_t r;
> +
> +    primary_frame = s->raw_frame_header->primary_ref_frame;
> +    prev_frame = s->raw_frame_header->ref_frame_idx[primary_frame];
> +    abs_bits = AV1_GM_ABS_ALPHA_BITS;
> +    prec_bits = AV1_GM_ALPHA_PREC_BITS;
> +
> +    if (idx < 2) {
> +        if (type == AV1_WARP_MODEL_TRANSLATION) {
> +            abs_bits = AV1_GM_ABS_TRANS_ONLY_BITS -
> +                !s->raw_frame_header->allow_high_precision_mv;
> +            prec_bits = AV1_GM_TRANS_ONLY_PREC_BITS -
> +                !s->raw_frame_header->allow_high_precision_mv;
> +        } else {
> +            abs_bits = AV1_GM_ABS_TRANS_BITS;
> +            prec_bits = AV1_GM_TRANS_PREC_BITS;
> +        }
> +    }
> +    round = (idx % 3) == 2 ? (1 << AV1_WARPEDMODEL_PREC_BITS) : 0;
> +    prec_diff = AV1_WARPEDMODEL_PREC_BITS - prec_bits;
> +    sub = (idx % 3) == 2 ? (1 << prec_bits) : 0;
> +    mx = 1 << abs_bits;
> +    r = (s->ref[prev_frame].gm_params[ref][idx] >> prec_diff) - sub;
> +
> +    s->cur_frame.gm_params[ref][idx] =
> +        (decode_signed_subexp_with_ref(s->raw_frame_header->gm_params[ref][idx],
> +                                       -mx, mx + 1, r) << prec_diff) + round;
> +}
> +
> +/**
> +* update gm type/params, since cbs already implemented part of this funcation,
> +* so we don't need to full implement spec.
> +*/
> +static void global_motion_params(AV1DecContext *s)
> +{
> +    const AV1RawFrameHeader *header = s->raw_frame_header;
> +    int type, ref;
> +
> +    for (ref = AV1_REF_FRAME_LAST; ref <= AV1_REF_FRAME_ALTREF; ref++) {
> +        s->cur_frame.gm_type[ref] = AV1_WARP_MODEL_IDENTITY;
> +        for (int i = 0; i < 6; i++)
> +            s->cur_frame.gm_params[ref][i] = (i % 3 == 2) ?
> +                                             1 << AV1_WARPEDMODEL_PREC_BITS : 0;
> +    }
> +    if (header->frame_type == AV1_FRAME_KEY ||
> +        header->frame_type == AV1_FRAME_INTRA_ONLY)
> +        return;
> +
> +    for (ref = AV1_REF_FRAME_LAST; ref <= AV1_REF_FRAME_ALTREF; ref++) {
> +        if (header->is_global[ref]) {
> +            if (header->is_rot_zoom[ref]) {
> +                type = AV1_WARP_MODEL_ROTZOOM;
> +            } else {
> +                type = header->is_translation[ref] ? AV1_WARP_MODEL_TRANSLATION
> +                                                   : AV1_WARP_MODEL_AFFINE;
> +            }
> +        } else {
> +            type = AV1_WARP_MODEL_IDENTITY;
> +        }
> +        s->cur_frame.gm_type[ref] = type;
> +
> +        if (type >= AV1_WARP_MODEL_ROTZOOM) {
> +            read_global_param(s, type, ref, 2);
> +            read_global_param(s, type, ref, 3);
> +            if (type == AV1_WARP_MODEL_AFFINE) {
> +                read_global_param(s, type, ref, 4);
> +                read_global_param(s, type, ref, 5);
> +            } else {
> +                s->cur_frame.gm_params[ref][4] = -s->cur_frame.gm_params[ref][3];
> +                s->cur_frame.gm_params[ref][5] = s->cur_frame.gm_params[ref][2];
> +            }
> +        }
> +        if (type >= AV1_WARP_MODEL_TRANSLATION) {
> +            read_global_param(s, type, ref, 0);
> +            read_global_param(s, type, ref, 1);
> +        }
> +    }
> +}
> +
> +static int get_tiles_info(AVCodecContext *avctx, AV1RawTileGroup *tile_group)
> +{
> +    AV1DecContext *s = avctx->priv_data;
> +    GetBitContext gb;
> +    uint16_t tile_num, tile_row, tile_col;
> +    uint32_t size = 0, size_bytes = 0, offset = 0;
> +    int ret = 0;
> +
> +    if ((ret = init_get_bits8(&gb,
> +                              tile_group->tile_data.data,
> +                              tile_group->tile_data.data_size)) < 0) {
> +        av_log(avctx, AV_LOG_ERROR, "Failed to initialize bitstream reader.\n");
> +        return ret;
> +    }
> +
> +    s->tg_start = tile_group->tg_start;
> +    s->tg_end = tile_group->tg_end;
> +
> +    for (tile_num = tile_group->tg_start; tile_num <= tile_group->tg_end; tile_num++) {
> +        tile_row = tile_num / s->raw_frame_header->tile_cols;
> +        tile_col = tile_num % s->raw_frame_header->tile_cols;
> +
> +        if (tile_num == tile_group->tg_end) {
> +            s->tile_group_info[tile_num].tile_size = get_bits_left(&gb) / 8;
> +            s->tile_group_info[tile_num].tile_offset = offset;
> +            s->tile_group_info[tile_num].tile_row = tile_row;
> +            s->tile_group_info[tile_num].tile_column = tile_col;
> +            return 0;
> +        }
> +        size_bytes = s->raw_frame_header->tile_size_bytes_minus1 + 1;
> +        size = get_bits_le(&gb, size_bytes * 8) + 1;
> +        skip_bits(&gb, size * 8);
> +
> +        offset += size_bytes;
> +
> +        s->tile_group_info[tile_num].tile_size = size;
> +        s->tile_group_info[tile_num].tile_offset = offset;
> +        s->tile_group_info[tile_num].tile_row = tile_row;
> +        s->tile_group_info[tile_num].tile_column = tile_col;
> +
> +        offset += size;
> +    }
> +
> +    return 0;
> +
> +}
> +
> +static int get_pixel_format(AVCodecContext *avctx)
> +{
> +    AV1DecContext *s = avctx->priv_data;
> +    const AV1RawSequenceHeader *seq = s->raw_seq;
> +    uint8_t bit_depth;
> +    int ret = 0;
> +    enum AVPixelFormat pix_fmt = AV_PIX_FMT_NONE;
> +#define HWACCEL_MAX (0)
> +    enum AVPixelFormat pix_fmts[HWACCEL_MAX + 1], *fmtp = pix_fmts;
> +
> +    if (seq->seq_profile == 2 && seq->color_config.high_bitdepth)
> +        bit_depth = seq->color_config.twelve_bit ? 12 : 10;
> +    else if (seq->seq_profile <= 2)
> +        bit_depth = seq->color_config.high_bitdepth ? 10 : 8;
> +    else {
> +        av_log(avctx, AV_LOG_ERROR,
> +               "Unknow av1 profile %d.\n", seq->seq_profile);
> +        return -1;
> +    }
> +
> +    if (!seq->color_config.mono_chrome) {
> +        // 4:4:4 x:0 y:0, 4:2:2 x:1 y:0, 4:2:0 x:1 y:1
> +        if (seq->color_config.subsampling_x == 0 &&
> +            seq->color_config.subsampling_y == 0) {
> +            if (bit_depth == 8)
> +                pix_fmt = AV_PIX_FMT_YUV444P;
> +            else if (bit_depth == 10)
> +                pix_fmt = AV_PIX_FMT_YUV444P10;
> +            else if (bit_depth == 12)
> +                pix_fmt = AV_PIX_FMT_YUV444P12;
> +            else
> +                av_log(avctx, AV_LOG_WARNING, "Unknow av1 pixel format.\n");
> +        } else if (seq->color_config.subsampling_x == 1 &&
> +                   seq->color_config.subsampling_y == 0) {
> +            if (bit_depth == 8)
> +                pix_fmt = AV_PIX_FMT_YUV422P;
> +            else if (bit_depth == 10)
> +                pix_fmt = AV_PIX_FMT_YUV422P10;
> +            else if (bit_depth == 12)
> +                pix_fmt = AV_PIX_FMT_YUV422P12;
> +            else
> +                av_log(avctx, AV_LOG_WARNING, "Unknow av1 pixel format.\n");
> +        } else if (seq->color_config.subsampling_x == 1 &&
> +                   seq->color_config.subsampling_y == 1) {
> +            if (bit_depth == 8)
> +                pix_fmt = AV_PIX_FMT_YUV420P;
> +            else if (bit_depth == 10)
> +                pix_fmt = AV_PIX_FMT_YUV420P10;
> +            else if (bit_depth == 12)
> +                pix_fmt = AV_PIX_FMT_YUV420P12;
> +            else
> +                av_log(avctx, AV_LOG_WARNING, "Unknow av1 pixel format.\n");
> +        }
> +    } else {
> +        if (seq->color_config.subsampling_x == 1 &&
> +            seq->color_config.subsampling_y == 1)
> +            pix_fmt = AV_PIX_FMT_YUV440P;
> +        else
> +            av_log(avctx, AV_LOG_WARNING, "Unknow av1 pixel format.\n");
> +    }
> +
> +    av_log(avctx, AV_LOG_DEBUG, "Av1 decode get format: %s.\n",
> +           av_get_pix_fmt_name(pix_fmt));
> +
> +    if (pix_fmt == AV_PIX_FMT_NONE)
> +        return -1;
> +    s->pix_fmt = pix_fmt;
> +
> +    *fmtp = AV_PIX_FMT_NONE;
> +    avctx->sw_pix_fmt = s->pix_fmt;
> +    ret = ff_thread_get_format(avctx, pix_fmts);
> +    if (ret < 0)
> +        return ret;
> +
> +    /**
> +     * check if the HW accel is inited correctly. If not, abort. Since now the
> +     * av1 decoder doesn't support native decode, if it will be implemented in
> +     * the feature, please remove this check.
> +     */
> +    if (!avctx->hwaccel) {
> +        av_log(avctx, AV_LOG_ERROR, "Your platform doesn't suppport"
> +               " hardware acceleration AV1 decode.\n");
> +        abort();

Forcing a crash seems a bit harsh. You should return AVERROR(ENOSYS)
instead.

> +    }
> +
> +    return 0;
> +}
> +
> +static void av1_frame_unref(AVCodecContext *avctx, AV1Frame *f)
> +{
> +    ff_thread_release_buffer(avctx, &f->tf);
> +    av_buffer_unref(&f->hwaccel_priv_buf);
> +    f->hwaccel_picture_private = NULL;
> +}
> +
> +static int av1_frame_ref(AVCodecContext *avctx, AV1Frame *dst, AV1Frame *src)
> +{
> +    int ret = 0;
> +
> +    ret = ff_thread_ref_frame(&dst->tf, &src->tf);
> +    if (ret < 0)
> +        return ret;
> +
> +    if (src->hwaccel_picture_private) {
> +        dst->hwaccel_priv_buf = av_buffer_ref(src->hwaccel_priv_buf);
> +        if (!dst->hwaccel_priv_buf)
> +            goto fail;
> +        dst->hwaccel_picture_private = dst->hwaccel_priv_buf->data;
> +    }
> +
> +    dst->loop_filter_delta_enabled = src->loop_filter_delta_enabled;
> +    memcpy(dst->loop_filter_ref_deltas,
> +           src->loop_filter_ref_deltas,
> +           AV1_NUM_REF_FRAMES * sizeof(int8_t));
> +    memcpy(dst->loop_filter_mode_deltas,
> +           src->loop_filter_mode_deltas,
> +           2 * sizeof(int8_t));
> +    memcpy(dst->gm_type,
> +           src->gm_type,
> +           AV1_NUM_REF_FRAMES * sizeof(uint8_t));
> +    memcpy(dst->gm_params,
> +           src->gm_params,
> +           AV1_NUM_REF_FRAMES * 6 * sizeof(int32_t));
> +
> +    return 0;
> +
> +fail:
> +    av1_frame_unref(avctx, dst);
> +    return AVERROR(ENOMEM);
> +}
> +
> +static av_cold int av1_decode_free(AVCodecContext *avctx)
> +{
> +    AV1DecContext *s = avctx->priv_data;
> +
> +    for (int i = 0; i < FF_ARRAY_ELEMS(s->ref); i++) {
> +        if (s->ref[i].tf.f->buf[0])
> +            av1_frame_unref(avctx, &s->ref[i]);
> +        av_frame_free(&s->ref[i].tf.f);
> +    }
> +    if (s->cur_frame.tf.f->buf[0])
> +        av1_frame_unref(avctx, &s->cur_frame);
> +    av_frame_free(&s->cur_frame.tf.f);
> +
> +    if (s->seq_ref) {
> +        av_buffer_unref(&s->seq_ref);
> +        s->raw_seq = NULL;
> +    }
> +
> +    if (s->header_ref) {
> +        av_buffer_unref(&s->header_ref);
> +        s->raw_frame_header = NULL;
> +    }
> +
> +    ff_cbs_fragment_free(&s->current_obu);
> +    ff_cbs_close(&s->cbc);
> +
> +    return 0;
> +}
> +
> +static av_cold int av1_decode_init(AVCodecContext *avctx)
> +{
> +    AV1DecContext *s = avctx->priv_data;
> +    int ret = 0;
> +
> +    for (int i = 0; i < FF_ARRAY_ELEMS(s->ref); i++) {
> +        s->ref[i].tf.f = av_frame_alloc();
> +        if (!s->ref[i].tf.f) {
> +            av1_decode_free(avctx);
> +            av_log(avctx, AV_LOG_ERROR,
> +                   "Failed to allocate reference frame buffer %d.\n", i);
> +            return AVERROR(ENOMEM);
> +        }
> +    }
> +
> +    s->cur_frame.tf.f = av_frame_alloc();
> +    if (!s->cur_frame.tf.f) {
> +        av1_decode_free(avctx);
> +        av_log(avctx, AV_LOG_ERROR,
> +               "Failed to allocate current frame buffer.\n");
> +        return AVERROR(ENOMEM);
> +    }
> +
> +    s->avctx = avctx;
> +    s->pix_fmt = AV_PIX_FMT_NONE;
> +    s->seq_ref = NULL;
> +    s->header_ref = NULL;
> +
> +    ret = ff_cbs_init(&s->cbc, AV_CODEC_ID_AV1, avctx);
> +    if (ret < 0)
> +        return ret;

You should parse avctx->extradata with ff_cbs_read(), if any, and
initialize avctx with values taken from the filled AV1RawSequenceHeader.
Profile, level, color info, timing info, the like.

And use (CodedBitstreamAV1Context *)(s->cbc->priv_data)->sequence_header
instead of looping through all the OBUs potentially present in extradata.

> +    return 0;
> +}
> +
> +static int av1_frame_alloc(AVCodecContext *avctx, AV1Frame *f)
> +{
> +    int ret;
> +    if ((ret = ff_thread_get_buffer(avctx, &f->tf, AV_GET_BUFFER_FLAG_REF)) < 0)

ff_thread_get_buffer() will take some avctx values and use them to
initialize the frame. As is, avctx has whatever values the caller put in
it before calling avcodec_open2() (for example those from the
container), or taken from the sequence header in extradata once you
implement the above. This will not work for frame size changes.

What you need to do here is update avctx with props taken from the
current AV1RawFrameHeader before calling get_buffer(). Width, height and
sample_aspect_ratio (derived from width/height and render_width/height)
should be enough. Use ff_set_dimensions() and ff_set_sar() for it.

You should also fill frame->pict_type from
s->cur_frame_header->frame_type, but that can be done at any time within
this function.

> +        return ret;
> +
> +    if (avctx->hwaccel) {
> +        const AVHWAccel *hwaccel = avctx->hwaccel;
> +        if (hwaccel->frame_priv_data_size) {
> +            f->hwaccel_priv_buf =
> +                av_buffer_allocz(hwaccel->frame_priv_data_size);
> +            if (!f->hwaccel_priv_buf)
> +                goto fail;
> +            f->hwaccel_picture_private = f->hwaccel_priv_buf->data;
> +        }
> +    }
> +    return 0;
> +
> +fail:
> +    av1_frame_unref(avctx, f);
> +    return AVERROR(ENOMEM);
> +}
> +
> +static int set_output_frame(AVFrame *srcframe, AVFrame *frame, AVPacket *pkt)
> +{
> +    int ret = 0;
> +    if ((ret = av_frame_ref(frame, srcframe)) < 0) {
> +        return ret;
> +    }
> +
> +    frame->pts = pkt->pts;
> +    frame->pkt_dts = pkt->dts;
> +
> +    return 0;
> +}
> +
> +static int update_reference_list (AVCodecContext *avctx)
> +{
> +    AV1DecContext *s = avctx->priv_data;
> +    AV1RawFrameHeader *header= s->raw_frame_header;
> +    int ret = 0;
> +
> +    for (int i = 0; i < AV1_NUM_REF_FRAMES; i++) {
> +        if (header->refresh_frame_flags & (1 << i)) {
> +            if (s->ref[i].tf.f->buf[0])
> +                av1_frame_unref(avctx, &s->ref[i]);
> +            if ((ret = av1_frame_ref(avctx, &s->ref[i], &s->cur_frame)) < 0) {
> +                av_log(avctx, AV_LOG_DEBUG, "Ref frame error:%d.\n", ret);
> +                return ret;
> +            }
> +        }
> +    }
> +    return 0;
> +}
> +
> +static int init_tile_data(AV1DecContext *s)
> +{
> +    s->tile_num =
> +        s->raw_frame_header->tile_cols * s->raw_frame_header->tile_rows;
> +    s->tile_group_info = av_mallocz_array(s->tile_num, sizeof(TileGroupInfo));
> +    if (!s->tile_group_info) {
> +        return AVERROR(ENOMEM);
> +    }
> +    return 0;
> +}
> +
> +static int free_tile_data(AV1DecContext *s)
> +{
> +    s->tile_num = 0;
> +    if (s->tile_group_info)
> +        av_freep(&s->tile_group_info);
> +    return 0;
> +}

You're allocating and freeing this buffer on every frame. You could
allocate it once, reuse it for every frame, and resizing it if it needs
to be bigger (very unlikely for frames to have different amount of tiles
within the same video sequence, but in theory possible).

> +
> +static int get_current_frame(AVCodecContext *avctx)
> +{
> +    AV1DecContext *s = avctx->priv_data;
> +    int ret = 0;
> +
> +    if (s->cur_frame.tf.f->buf[0])
> +        av1_frame_unref(avctx, &s->cur_frame);
> +
> +    ret = av1_frame_alloc(avctx, &s->cur_frame);
> +    if (ret < 0) {
> +        av_log(avctx, AV_LOG_ERROR,
> +               "Failed to allocate space for current frame.\n");
> +        return ret;
> +    }
> +
> +    ret = init_tile_data(s);
> +    if (ret < 0) {
> +        av_log(avctx, AV_LOG_ERROR, "Failed to init tile data.\n");
> +        return ret;
> +    }
> +
> +    if (s->raw_frame_header->primary_ref_frame == AV1_PRIMARY_REF_NONE)
> +        setup_past_independence(&s->cur_frame);
> +    else
> +        load_previous_and_update(s);
> +
> +    global_motion_params(s);
> +
> +    return ret;
> +}
> +
> +static int av1_decode_frame(AVCodecContext *avctx, void *frame,
> +                            int *got_frame, AVPacket *pkt)
> +{
> +    AV1DecContext *s = avctx->priv_data;
> +    AV1RawTileGroup *raw_tile_group = NULL;
> +    int ret = 0;
> +
> +    ret = ff_cbs_read_packet(s->cbc, &s->current_obu, pkt);
> +    if (ret < 0) {
> +        av_log(avctx, AV_LOG_ERROR, "Failed to read packet.\n");
> +        goto end;
> +    }
> +    av_log(avctx, AV_LOG_DEBUG, "Total obu for this frame:%d.\n",
> +           s->current_obu.nb_units);
> +
> +    for (int i = 0; i < s->current_obu.nb_units; i++) {
> +        CodedBitstreamUnit *unit = &s->current_obu.units[i];
> +        AV1RawOBU *obu = unit->content;
> +        av_log(avctx, AV_LOG_DEBUG, "Obu idx:%d, obu type:%d.\n", i, unit->type);
> +
> +        switch (unit->type) {
> +        case AV1_OBU_SEQUENCE_HEADER:
> +            if (s->seq_ref) {

No need to check s->seq_ref. Just call av_buffer_unref()
unconditionally. It will unreference it if non-NULL, and do nothing if
it's already NULL.

> +                av_buffer_unref(&s->seq_ref);
> +                s->raw_seq = NULL;

Not really necessary. It will be set again a few lines below.

> +            }
> +            s->seq_ref = av_buffer_ref(unit->content_ref);

Unchecked av_buffer_ref() call.

> +            s->raw_seq = &obu->obu.sequence_header;
> +            if (s->pix_fmt == AV_PIX_FMT_NONE) {
> +                ret = get_pixel_format(avctx);
> +                if (ret < 0) {
> +                    av_log(avctx, AV_LOG_ERROR,
> +                           "Failed to get pixel format.\n");
> +                    goto end;
> +                }
> +            }
> +
> +            if (avctx->hwaccel && avctx->hwaccel->decode_params) {
> +                ret = avctx->hwaccel->decode_params(avctx, unit->type, unit->data,
> +                                                    unit->data_size);
> +                if (ret < 0) {
> +                    av_log(avctx, AV_LOG_ERROR, "HW accel decode params fail.\n");
> +                    return ret;
> +                }
> +            }
> +            break;
> +        case AV1_OBU_FRAME_HEADER:
> +        case AV1_OBU_REDUNDANT_FRAME_HEADER:

You should not parse a redundant frame header if a frame header was
parsed before it and there are still tile groups that remain to be
parsed for it.

> +            if (obu->obu.frame_header.show_existing_frame) {
> +                if (set_output_frame(s->ref[obu->obu.frame_header.frame_to_show_map_idx].tf.f,
> +                                     frame, pkt) < 0) {
> +                    av_log(avctx, AV_LOG_DEBUG, "Set output frame error:%d.\n",
> +                           ret);
> +                    goto end;
> +                }
> +                *got_frame = 1;
> +                goto end;
> +            } else {
> +                if (s->header_ref) {
> +                    av_buffer_unref(&s->header_ref);
> +                    s->raw_frame_header = NULL;

Same as with s->seq_ref.

> +                }
> +                s->header_ref = av_buffer_ref(unit->content_ref);

Unchecked call.

> +                s->raw_frame_header = &obu->obu.frame_header;
> +                ret = get_current_frame(avctx);
> +                if (ret < 0) {
> +                    av_log(avctx, AV_LOG_DEBUG, "Get current frame error:%d.\n",
> +                           ret);
> +                    goto end;
> +                }
> +
> +                if (avctx->hwaccel) {
> +                    ret = avctx->hwaccel->start_frame(avctx, unit->data,
> +                                                      unit->data_size);
> +                    if (ret < 0) {
> +                        av_log(avctx, AV_LOG_ERROR, "HW accel start frame fail.\n");
> +                        return ret;
> +                    }
> +                }
> +            }
> +            break;
> +        case AV1_OBU_FRAME:
> +        case AV1_OBU_TILE_GROUP:
> +            if (unit->type == AV1_OBU_FRAME) {
> +                if (s->header_ref) {
> +                    av_buffer_unref(&s->header_ref);
> +                    s->raw_frame_header = NULL;
> +                }
> +                s->header_ref = av_buffer_ref(unit->content_ref);
> +                s->raw_frame_header = &obu->obu.frame.header;
> +                raw_tile_group = &obu->obu.frame.tile_group;
> +                ret = get_current_frame(avctx);
> +                if (ret < 0) {
> +                    av_log(avctx, AV_LOG_DEBUG, "Get current frame error:%d.\n",
> +                           ret);
> +                    goto end;
> +                }
> +
> +                if (avctx->hwaccel) {
> +                    ret = avctx->hwaccel->start_frame(avctx, unit->data,
> +                                                      unit->data_size);
> +                    if (ret < 0) {
> +                        av_log(avctx, AV_LOG_ERROR, "HW accel start frame fail.\n");
> +                        return ret;
> +                    }
> +                }
> +            } else {
> +                raw_tile_group = &obu->obu.tile_group;
> +            }
> +            get_tiles_info(avctx, raw_tile_group);
> +
> +            if (avctx->hwaccel) {
> +                ret = avctx->hwaccel->decode_slice(avctx,
> +                                                   raw_tile_group->tile_data.data,
> +                                                   raw_tile_group->tile_data.data_size);
> +                if (ret < 0) {
> +                    av_log(avctx, AV_LOG_ERROR,
> +                           "HW accel decode slice fail.\n");
> +                    return ret;
> +                }
> +            }
> +            break;
> +        case AV1_OBU_TILE_LIST:
> +        case AV1_OBU_TEMPORAL_DELIMITER:
> +        case AV1_OBU_PADDING:
> +        case AV1_OBU_METADATA:
> +            break;
> +        default:
> +            av_log(avctx, AV_LOG_DEBUG,
> +                   "Unknow obu type: %d (%zd bits).\n",
> +                   unit->type, unit->data_size);
> +        }
> +
> +        if (raw_tile_group && (s->tile_num == raw_tile_group->tg_end + 1)) {
> +            if (avctx->hwaccel) {
> +                ret = avctx->hwaccel->end_frame(avctx);
> +                if (ret < 0) {
> +                    av_log(avctx, AV_LOG_ERROR, "HW accel end frame fail.\n");
> +                    return ret;
> +                }
> +            }
> +
> +            ret = update_reference_list(avctx);
> +            if (ret < 0) {
> +                av_log(avctx, AV_LOG_ERROR, "Failed to update reference list.\n");
> +                return ret;
> +            }
> +
> +            if (s->raw_frame_header->show_frame) {
> +                if (set_output_frame(s->cur_frame.tf.f, frame, pkt) < 0) {
> +                    av_log(avctx, AV_LOG_DEBUG, "Set output frame error:%d.\n",
> +                    ret);
> +                    goto end;
> +                }
> +                *got_frame = 1;

You need to set frame->key_frame to 1 if s->raw_frame_header->frame_type
== AV1_FRAME_KEY and s->raw_frame_header->show_frame == 1.

> +            }
> +            free_tile_data(s);
> +            raw_tile_group = NULL;
> +        }
> +    }

You're duplicating a bit of code in the loop above. You could merge it
even more. Frame header and Frame are both the same up to the point the
Tile Group bits would start for the latter.

You could do something like:

/*****/
switch (unit->type) {
case AV1_OBU_REDUNDANT_FRAME_HEADER:
    if (s->cur_frame_header)
        break; // In the middle of a frame, so ignore it
// fall-through
case AV1_OBU_FRAME:
case AV1_OBU_FRAME_HEADER:
    /* s->cur_frame_header_buf set here */

    if (unit->type == AV1_OBU_FRAME)
        s->cur_frame_header = &obu->obu.frame.header;
    else
        s->cur_frame_header = &obu->obu.frame_header;

    /* Frame header handling code here */

    if (unit->type != AV1_OBU_FRAME)
        break; // Keep going only for Frame OBUs
// fall-through
case AV1_OBU_TILE_GROUP:
    if (!s->cur_frame_header) { // Sanity check
        ret = AVERROR_INVALIDDATA;
        goto end;
    }
    if (unit->type == AV1_OBU_FRAME)
        raw_tile_group = &obu->obu.frame.tile_group;
    else
        raw_tile_group = &obu->obu.tile_group;

    /* Tile group handling code here */

    break;
}

if (raw_tile_group && (s->tile_num == raw_tile_group->tg_end + 1)) {
    /* Finish frame, output it if show_frame == 1 */

    s->cur_frame_header = NULL;
    raw_tile_group = NULL;
}
/*****/

> +
> +end:
> +    free_tile_data(s);
> +    ff_cbs_fragment_reset(&s->current_obu);
> +    return ret;
> +}
> +
> +static void av1_decode_flush(AVCodecContext *avctx)
> +{
> +    AV1DecContext *s = avctx->priv_data;
> +
> +    for (int i = 0; i < FF_ARRAY_ELEMS(s->ref); i++)
> +        av1_frame_unref(avctx, &s->ref[i]);
> +
> +    av1_frame_unref(avctx, &s->cur_frame);
> +}
> +
> +AVCodec ff_av1_decoder = {
> +    .name                  = "av1",
> +    .long_name             = NULL_IF_CONFIG_SMALL("Alliance for Open Media AV1"),
> +    .type                  = AVMEDIA_TYPE_VIDEO,
> +    .id                    = AV_CODEC_ID_AV1,
> +    .priv_data_size        = sizeof(AV1DecContext),
> +    .init                  = av1_decode_init,
> +    .close                 = av1_decode_free,
> +    .decode                = av1_decode_frame,
> +    .capabilities          = AV_CODEC_CAP_DR1,
> +    .caps_internal         = FF_CODEC_CAP_INIT_THREADSAFE |
> +                             FF_CODEC_CAP_INIT_CLEANUP |
> +                             FF_CODEC_CAP_SETS_PKT_DTS,
> +    .flush                 = av1_decode_flush,
> +    .profiles              = NULL_IF_CONFIG_SMALL(ff_av1_profiles),
> +    .hw_configs            = (const AVCodecHWConfigInternal * []) {
> +        NULL
> +    },
> +};
> diff --git a/libavcodec/av1dec.h b/libavcodec/av1dec.h
> new file mode 100644
> index 0000000000..ab0bda4d3a
> --- /dev/null
> +++ b/libavcodec/av1dec.h
> @@ -0,0 +1,70 @@
> +/*
> + * AV1 video decoder
> + * *
> + * This file is part of FFmpeg.
> + *
> + * FFmpeg is free software; you can redistribute it and/or
> + * modify it under the terms of the GNU Lesser General Public
> + * License as published by the Free Software Foundation; either
> + * version 2.1 of the License, or (at your option) any later version.
> + *
> + * FFmpeg is distributed in the hope that it will be useful,
> + * but WITHOUT ANY WARRANTY; without even the implied warranty of
> + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
> + * Lesser General Public License for more details.
> + *
> + * You should have received a copy of the GNU Lesser General Public
> + * License along with FFmpeg; if not, write to the Free Software
> + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
> + */
> +
> +#ifndef AVCODEC_AV1DEC_H
> +#define AVCODEC_AV1DEC_H
> +
> +#include "thread.h"
> +#include "cbs.h"
> +#include "cbs_av1.h"
> +
> +typedef struct AV1Frame {
> +    ThreadFrame tf;
> +
> +    AVBufferRef *hwaccel_priv_buf;
> +    void *hwaccel_picture_private;
> +
> +    uint8_t loop_filter_delta_enabled;
> +    int8_t  loop_filter_ref_deltas[AV1_NUM_REF_FRAMES];
> +    int8_t  loop_filter_mode_deltas[2];
> +    uint8_t gm_type[AV1_NUM_REF_FRAMES];
> +    int32_t gm_params[AV1_NUM_REF_FRAMES][6];
> +} AV1Frame;
> +
> +typedef struct TileGroupInfo {
> +    uint32_t tile_offset;
> +    uint32_t tile_size;
> +    uint16_t tile_row;
> +    uint16_t tile_column;
> +} TileGroupInfo;
> +
> +typedef struct AV1DecContext {
> +    const AVClass *class;
> +    AVCodecContext *avctx;
> +
> +    enum AVPixelFormat pix_fmt;
> +    CodedBitstreamContext *cbc;
> +    CodedBitstreamFragment current_obu;
> +
> +    AVBufferRef *seq_ref;
> +    AV1RawSequenceHeader *raw_seq;
> +    AVBufferRef *header_ref;
> +    AV1RawFrameHeader *raw_frame_header;
> +    TileGroupInfo *tile_group_info;
> +    uint16_t tile_num;
> +    uint16_t tg_start;
> +    uint16_t tg_end;
> +
> +    AV1Frame ref[AV1_NUM_REF_FRAMES];
> +    AV1Frame cur_frame;
> +
> +} AV1DecContext;
> +
> +#endif /* AVCODEC_AV1DEC_H */
> diff --git a/libavcodec/version.h b/libavcodec/version.h
> index 91e0564570..739caf5c6e 100644
> --- a/libavcodec/version.h
> +++ b/libavcodec/version.h
> @@ -28,8 +28,8 @@
>  #include "libavutil/version.h"
>  
>  #define LIBAVCODEC_VERSION_MAJOR  58
> -#define LIBAVCODEC_VERSION_MINOR 101
> -#define LIBAVCODEC_VERSION_MICRO 101
> +#define LIBAVCODEC_VERSION_MINOR 102
> +#define LIBAVCODEC_VERSION_MICRO 100
>  
>  #define LIBAVCODEC_VERSION_INT  AV_VERSION_INT(LIBAVCODEC_VERSION_MAJOR, \
>                                                 LIBAVCODEC_VERSION_MINOR, \
> 



More information about the ffmpeg-devel mailing list