[FFmpeg-devel] [GSOC][PATCH 1/3] lavc/cfhd:3d transform decoding for both progressive and interlaced
Gagandeep Singh
deepgagan231197 at gmail.com
Tue Aug 14 10:34:59 EEST 2018
On Tue, Aug 14, 2018 at 1:03 PM Gagandeep Singh <deepgagan231197 at gmail.com>
wrote:
> ---
> libavcodec/cfhd.c | 511 ++++++++++++++++++++++++++++++++++++++++------
> libavcodec/cfhd.h | 13 +-
> 2 files changed, 454 insertions(+), 70 deletions(-)
>
> diff --git a/libavcodec/cfhd.c b/libavcodec/cfhd.c
> index 76838f0869..3906d7a86d 100644
> --- a/libavcodec/cfhd.c
> +++ b/libavcodec/cfhd.c
> @@ -41,12 +41,15 @@
> #define ALPHA_COMPAND_GAIN 9400
>
> enum CFHDParam {
> + TransformType = 10,
> ChannelCount = 12,
> SubbandCount = 14,
> + Pframe = 19,
> ImageWidth = 20,
> ImageHeight = 21,
> LowpassPrecision = 35,
> SubbandNumber = 48,
> + EncodingMethod = 52,
> Quantization = 53,
> ChannelNumber = 62,
> SampleFlags = 68,
> @@ -64,6 +67,7 @@ static av_cold int cfhd_init(AVCodecContext *avctx)
>
> avctx->bits_per_raw_sample = 10;
> s->avctx = avctx;
> + s->progressive = 0;
>
> return ff_cfhd_init_vlcs(s);
> }
> @@ -84,6 +88,10 @@ static void init_peak_table_defaults(CFHDContext *s)
>
> static void init_frame_defaults(CFHDContext *s)
> {
> + s->sample_type = 0;
> + s->transform_type = 0;
> + s->pframe = 0;
> + s->first_wavelet = 0;
> s->coded_width = 0;
> s->coded_height = 0;
> s->cropped_height = 0;
> @@ -97,14 +105,15 @@ static void init_frame_defaults(CFHDContext *s)
> s->pshift = 1;
> s->codebook = 0;
> s->difference_coding = 0;
> - s->progressive = 0;
> init_plane_defaults(s);
> init_peak_table_defaults(s);
> }
>
> /* TODO: merge with VLC tables or use LUT */
> -static inline int dequant_and_decompand(int level, int quantisation, int
> codebook)
> +static inline int dequant_and_decompand(int level, int quantisation, int
> codebook, int lossless)
> {
> + if (lossless)
> + return level;
> if (codebook == 0 || codebook == 1) {
> int64_t abslevel = abs(level);
> if (level < 264)
> @@ -193,16 +202,21 @@ static inline void filter(int16_t *output, ptrdiff_t
> out_stride,
> }
> }
>
> -static inline void interlaced_vertical_filter(int16_t *output, int16_t
> *low, int16_t *high,
> - int width, int linesize, int plane)
> +static inline void inverse_temporal_filter(int16_t *output, int16_t *low,
> int16_t *high,
> + int width, int linesize, int
> temporal_for_highpass)
> {
> int i;
> int16_t even, odd;
> for (i = 0; i < width; i++) {
> even = (low[i] - high[i])/2;
> odd = (low[i] + high[i])/2;
> - output[i] = av_clip_uintp2(even, 10);
> - output[i + linesize] = av_clip_uintp2(odd, 10);
> + if (!temporal_for_highpass) {
> + output[i] = av_clip_uintp2(even, 10);
> + output[i + linesize] = av_clip_uintp2(odd, 10);
> + } else {
> + low[i] = even;
> + high[i] = odd;
> + }
> }
> }
> static void horiz_filter(int16_t *output, int16_t *low, int16_t *high,
> @@ -231,9 +245,12 @@ static void free_buffers(CFHDContext *s)
> for (i = 0; i < FF_ARRAY_ELEMS(s->plane); i++) {
> av_freep(&s->plane[i].idwt_buf);
> av_freep(&s->plane[i].idwt_tmp);
> -
> - for (j = 0; j < 9; j++)
> - s->plane[i].subband[j] = NULL;
> + if (s->transform_type == 0)
> + for (j = 0; j < 9; j++)
> + s->plane[i].subband[j] = NULL;
> + else
> + for (j = 0; j < 17; j++)
> + s->plane[i].subband[j] = NULL;
>
> for (j = 0; j < 8; j++)
> s->plane[i].l_h[j] = NULL;
> @@ -247,7 +264,7 @@ static int alloc_buffers(AVCodecContext *avctx)
> CFHDContext *s = avctx->priv_data;
> int i, j, ret, planes;
> int chroma_x_shift, chroma_y_shift;
> - unsigned k;
> + unsigned k, t;
>
> if ((ret = ff_set_dimensions(avctx, s->coded_width, s->coded_height))
> < 0)
> return ret;
> @@ -261,6 +278,7 @@ static int alloc_buffers(AVCodecContext *avctx)
>
> for (i = 0; i < planes; i++) {
> int w8, h8, w4, h4, w2, h2;
> + int16_t *frame2;
> int width = i ? avctx->width >> chroma_x_shift : avctx->width;
> int height = i ? avctx->height >> chroma_y_shift : avctx->height;
> ptrdiff_t stride = FFALIGN(width / 8, 8) * 8;
> @@ -277,28 +295,68 @@ static int alloc_buffers(AVCodecContext *avctx)
> w2 = w4 * 2;
> h2 = h4 * 2;
>
> - s->plane[i].idwt_buf =
> - av_mallocz_array(height * stride,
> sizeof(*s->plane[i].idwt_buf));
> - s->plane[i].idwt_tmp =
> - av_malloc_array(height * stride,
> sizeof(*s->plane[i].idwt_tmp));
> - if (!s->plane[i].idwt_buf || !s->plane[i].idwt_tmp)
> - return AVERROR(ENOMEM);
> -
> - s->plane[i].subband[0] = s->plane[i].idwt_buf;
> - s->plane[i].subband[1] = s->plane[i].idwt_buf + 2 * w8 * h8;
> - s->plane[i].subband[2] = s->plane[i].idwt_buf + 1 * w8 * h8;
> - s->plane[i].subband[3] = s->plane[i].idwt_buf + 3 * w8 * h8;
> - s->plane[i].subband[4] = s->plane[i].idwt_buf + 2 * w4 * h4;
> - s->plane[i].subband[5] = s->plane[i].idwt_buf + 1 * w4 * h4;
> - s->plane[i].subband[6] = s->plane[i].idwt_buf + 3 * w4 * h4;
> - s->plane[i].subband[7] = s->plane[i].idwt_buf + 2 * w2 * h2;
> - s->plane[i].subband[8] = s->plane[i].idwt_buf + 1 * w2 * h2;
> - s->plane[i].subband[9] = s->plane[i].idwt_buf + 3 * w2 * h2;
> -
> - for (j = 0; j < DWT_LEVELS; j++) {
> - for (k = 0; k < FF_ARRAY_ELEMS(s->plane[i].band[j]); k++) {
> - s->plane[i].band[j][k].a_width = w8 << j;
> - s->plane[i].band[j][k].a_height = h8 << j;
> + if (s->transform_type == 0) {
> + s->plane[i].idwt_buf =
> + av_mallocz_array(height * stride,
> sizeof(*s->plane[i].idwt_buf));
> + s->plane[i].idwt_tmp =
> + av_malloc_array(height * stride,
> sizeof(*s->plane[i].idwt_tmp));
> + if (!s->plane[i].idwt_buf || !s->plane[i].idwt_tmp)
> + return AVERROR(ENOMEM);
> + } else if (s->transform_type == 2) {
> + s->plane[i].idwt_buf =
> + av_mallocz_array(2 * height * stride,
> sizeof(*s->plane[i].idwt_buf));
> + s->plane[i].idwt_tmp =
> + av_malloc_array(2 * height * stride,
> sizeof(*s->plane[i].idwt_tmp));
> + if (!s->plane[i].idwt_buf || !s->plane[i].idwt_tmp)
> + return AVERROR(ENOMEM);
> + }
> +
> + if (s->transform_type == 0) {
> + s->plane[i].subband[0] = s->plane[i].idwt_buf;
> + s->plane[i].subband[1] = s->plane[i].idwt_buf + 2 * w8 * h8;
> + s->plane[i].subband[2] = s->plane[i].idwt_buf + 1 * w8 * h8;
> + s->plane[i].subband[3] = s->plane[i].idwt_buf + 3 * w8 * h8;
> + s->plane[i].subband[4] = s->plane[i].idwt_buf + 2 * w4 * h4;
> + s->plane[i].subband[5] = s->plane[i].idwt_buf + 1 * w4 * h4;
> + s->plane[i].subband[6] = s->plane[i].idwt_buf + 3 * w4 * h4;
> + s->plane[i].subband[7] = s->plane[i].idwt_buf + 2 * w2 * h2;
> + s->plane[i].subband[8] = s->plane[i].idwt_buf + 1 * w2 * h2;
> + s->plane[i].subband[9] = s->plane[i].idwt_buf + 3 * w2 * h2;
> + } else if (s->transform_type == 2) {
> + s->plane[i].subband[0] = s->plane[i].idwt_buf;
> + s->plane[i].subband[1] = s->plane[i].idwt_buf + 2 * w8 * h8;
> + s->plane[i].subband[2] = s->plane[i].idwt_buf + 1 * w8 * h8;
> + s->plane[i].subband[3] = s->plane[i].idwt_buf + 3 * w8 * h8;
> + s->plane[i].subband[4] = s->plane[i].idwt_buf + 2 * w4 * h4;
> + s->plane[i].subband[5] = s->plane[i].idwt_buf + 1 * w4 * h4;
> + s->plane[i].subband[6] = s->plane[i].idwt_buf + 3 * w4 * h4;
> + frame2 =
> + s->plane[i].subband[7] = s->plane[i].idwt_buf + 4 * w2 * h2;
> + s->plane[i].subband[8] = frame2 + 2 * w4 * h4;
> + s->plane[i].subband[9] = frame2 + 1 * w4 * h4;
> + s->plane[i].subband[10] = frame2 + 3 * w4 * h4;
> + s->plane[i].subband[11] = frame2 + 2 * w2 * h2;
> + s->plane[i].subband[12] = frame2 + 1 * w2 * h2;
> + s->plane[i].subband[13] = frame2 + 3 * w2 * h2;
> + s->plane[i].subband[14] = s->plane[i].idwt_buf + 2 * w2 * h2;
> + s->plane[i].subband[15] = s->plane[i].idwt_buf + 1 * w2 * h2;
> + s->plane[i].subband[16] = s->plane[i].idwt_buf + 3 * w2 * h2;
> + }
> +
> + if (s->transform_type == 0) {
> + for (j = 0; j < DWT_LEVELS - 3; j++) {
> + for (k = 0; k < FF_ARRAY_ELEMS(s->plane[i].band[j]); k++)
> {
> + s->plane[i].band[j][k].a_width = w8 << j;
> + s->plane[i].band[j][k].a_height = h8 << j;
> + }
> + }
> + } else if (s->transform_type == 2) {
> + for (j = 0; j < DWT_LEVELS; j++) {
> + t = j < 1 ? 0 : (j < 3 ? 1 : 2);
> + for (k = 0; k < FF_ARRAY_ELEMS(s->plane[i].band[0]); k++)
> {
> + s->plane[i].band[j][k].a_width = w8 << t;
> + s->plane[i].band[j][k].a_height = h8 << t;
> + }
> }
> }
>
> @@ -311,6 +369,11 @@ static int alloc_buffers(AVCodecContext *avctx)
> // s->plane[i].l_h[5] = ll1;
> s->plane[i].l_h[6] = s->plane[i].idwt_tmp;
> s->plane[i].l_h[7] = s->plane[i].idwt_tmp + 2 * w2 * h2;
> + if (s->transform_type == 2) {
> + frame2 = s->plane[i].idwt_tmp + 4 * w2 * h2;
> + s->plane[i].l_h[8] = frame2;
> + s->plane[i].l_h[9] = frame2 + 2 * w2 * h2;
> + }
> }
>
> s->a_height = s->coded_height;
> @@ -349,6 +412,9 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
> } else if (tag == SampleFlags) {
> av_log(avctx, AV_LOG_DEBUG, "Progressive?%"PRIu16"\n", data);
> s->progressive = data & 0x0001;
> + } else if (tag == Pframe) {
> + s->pframe = 1;
> + av_log(avctx, AV_LOG_DEBUG, "Frame type %"PRIu16"\n", data);
> } else if (tag == ImageWidth) {
> av_log(avctx, AV_LOG_DEBUG, "Width %"PRIu16"\n", data);
> s->coded_width = data;
> @@ -373,7 +439,7 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
> }
> } else if (tag == SubbandCount) {
> av_log(avctx, AV_LOG_DEBUG, "Subband Count: %"PRIu16"\n",
> data);
> - if (data != SUBBAND_COUNT) {
> + if (data != 10 && data != 17) {
> av_log(avctx, AV_LOG_ERROR, "Subband Count of %"PRIu16"
> is unsupported\n", data);
> ret = AVERROR_PATCHWELCOME;
> break;
> @@ -405,7 +471,7 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
> } else if (tag == 51) {
> av_log(avctx, AV_LOG_DEBUG, "Subband number actual
> %"PRIu16"\n", data);
> s->subband_num_actual = data;
> - if (s->subband_num_actual >= 10) {
> + if (s->subband_num_actual >= 17 && s->subband_num_actual !=
> 255) {
> av_log(avctx, AV_LOG_ERROR, "Invalid subband number
> actual\n");
> ret = AVERROR(EINVAL);
> break;
> @@ -420,9 +486,15 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
> s->prescale_shift[1] = (data >> 3) & 0x7;
> s->prescale_shift[2] = (data >> 6) & 0x7;
> av_log(avctx, AV_LOG_DEBUG, "Prescale shift (VC-5): %x\n",
> data);
> + } else if (tag == EncodingMethod) {
> + s->encode_method = data;
> + av_log(avctx, AV_LOG_DEBUG, "Encode Method for Subband %d :
> %x\n",s->subband_num_actual, data);
> } else if (tag == 27) {
> av_log(avctx, AV_LOG_DEBUG, "Lowpass width %"PRIu16"\n",
> data);
> - if (data < 3 || data >
> s->plane[s->channel_num].band[0][0].a_width) {
> + if (s->coded_width == 0){
> + s->coded_width = data << 3;
> + }
> + if (data < 3) {
> av_log(avctx, AV_LOG_ERROR, "Invalid lowpass width\n");
> ret = AVERROR(EINVAL);
> break;
> @@ -431,20 +503,21 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
> s->plane[s->channel_num].band[0][0].stride = data;
> } else if (tag == 28) {
> av_log(avctx, AV_LOG_DEBUG, "Lowpass height %"PRIu16"\n",
> data);
> - if (data < 3 || data >
> s->plane[s->channel_num].band[0][0].a_height) {
> + if (s->coded_height == 0)
> + s->coded_height = data << 3;
> + if (data < 3) {
> av_log(avctx, AV_LOG_ERROR, "Invalid lowpass height\n");
> ret = AVERROR(EINVAL);
> break;
> }
> s->plane[s->channel_num].band[0][0].height = data;
> - } else if (tag == 1)
> + } else if (tag == 1) {
> + s->sample_type = data;
> + if (data == 2)
> + s->pframe = 1;
> av_log(avctx, AV_LOG_DEBUG, "Sample type? %"PRIu16"\n", data);
> - else if (tag == 10) {
> - if (data != 0) {
> - avpriv_report_missing_feature(avctx, "Transform type of
> %"PRIu16, data);
> - ret = AVERROR_PATCHWELCOME;
> - break;
> - }
> + } else if (tag == 10) {
> + s->transform_type = data;
> av_log(avctx, AV_LOG_DEBUG, "Transform-type? %"PRIu16"\n",
> data);
> } else if (abstag >= 0x4000 && abstag <= 0x40ff) {
> if (abstag == 0x4001)
> @@ -551,8 +624,8 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
> av_log(avctx, AV_LOG_DEBUG, "Unknown tag %i data %x\n", tag,
> data);
>
> /* Some kind of end of header tag */
> - if (tag == 4 && data == 0x1a4a && s->coded_width &&
> s->coded_height &&
> - s->coded_format != AV_PIX_FMT_NONE) {
> + if (((tag == 4 && (data == 0x1a4a || data == 0xf0f)) ||
> s->sample_type == 1) && s->coded_width && s->coded_height &&
> + s->coded_format != AV_PIX_FMT_NONE && s->sample_type != 3 &&
> s->sample_type != 6) {
> if (s->a_width != s->coded_width || s->a_height !=
> s->coded_height ||
> s->a_format != s->coded_format) {
> free_buffers(s);
> @@ -619,8 +692,7 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
>
> av_log(avctx, AV_LOG_DEBUG, "Lowpass coefficients %d\n",
> lowpass_width * lowpass_height);
> }
> -
> - if (tag == 55 && s->subband_num_actual != 255 && s->a_width &&
> s->a_height) {
> + if ((tag == 55 || tag == 82) && s->a_width && s->a_height) {
> int highpass_height =
> s->plane[s->channel_num].band[s->level][s->subband_num].height;
> int highpass_width =
> s->plane[s->channel_num].band[s->level][s->subband_num].width;
> int highpass_a_width =
> s->plane[s->channel_num].band[s->level][s->subband_num].a_width;
> @@ -638,18 +710,24 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
> }
>
> if (highpass_height > highpass_a_height || highpass_width >
> highpass_a_width || a_expected < highpass_height *
> (uint64_t)highpass_stride) {
> - av_log(avctx, AV_LOG_ERROR, "Too many highpass
> coefficients\n");
> - ret = AVERROR(EINVAL);
> - goto end;
> + if (s->subband_num_actual != 255) {
> + av_log(avctx, AV_LOG_ERROR, "Too many highpass
> coefficients\n");
> + ret = AVERROR(EINVAL);
> + goto end;
> + }
> }
> expected = highpass_height * highpass_stride;
>
> av_log(avctx, AV_LOG_DEBUG, "Start subband coeffs plane %i
> level %i codebook %i expected %i\n", s->channel_num, s->level, s->codebook,
> expected);
>
> init_get_bits(&s->gb, gb.buffer,
> bytestream2_get_bytes_left(&gb) * 8);
> + if (s->subband_num_actual == 255) {
> + expected = 0;
> + goto finish;
> + }
> {
> OPEN_READER(re, &s->gb);
> - if (!s->codebook) {
> + if (!s->codebook && !(s->transform_type == 2 &&
> s->subband_num_actual == 7)) {
> while (1) {
> UPDATE_CACHE(re, &s->gb);
> GET_RL_VLC(level, run, re, &s->gb,
> s->table_9_rl_vlc,
> @@ -664,9 +742,14 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
> if (count > expected)
> break;
>
> - coeff = dequant_and_decompand(level,
> s->quantisation, 0);
> + coeff = dequant_and_decompand(level,
> s->quantisation, 0, (s->sample_type == 2 || s->sample_type == 3) &&
> s->pframe && s->subband_num_actual == 7 && s->encode_method == 5);
> for (i = 0; i < run; i++)
> - *coeff_data++ = coeff;
> + if (tag != 82)
> + *coeff_data++ = coeff;
> + else {
> + *coeff_data |= coeff << 8;
> + *coeff_data++ *= s->quantisation;
> + }
> }
> if (s->peak.level)
> peak_table(coeff_data - expected, &s->peak,
> expected);
> @@ -688,9 +771,14 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
> if (count > expected)
> break;
>
> - coeff = dequant_and_decompand(level,
> s->quantisation, s->codebook);
> + coeff = dequant_and_decompand(level,
> s->quantisation, s->codebook, (s->sample_type == 2 || s->sample_type == 3)
> && s->pframe && s->subband_num_actual == 7 && s->encode_method == 5);
> for (i = 0; i < run; i++)
> - *coeff_data++ = coeff;
> + if (tag != 82)
> + *coeff_data++ = coeff;
> + else {
> + *coeff_data |= coeff << 8;
> + *coeff_data++ *= s->quantisation;
> + }
> }
> if (s->peak.level)
> peak_table(coeff_data - expected, &s->peak,
> expected);
> @@ -701,12 +789,12 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
> CLOSE_READER(re, &s->gb);
> }
>
> - if (count > expected) {
> + if (count > expected && s->subband_num_actual != 255) {
> av_log(avctx, AV_LOG_ERROR, "Escape codeword not found,
> probably corrupt data\n");
> ret = AVERROR(EINVAL);
> goto end;
> }
> -
> + finish:
> bytes = FFALIGN(AV_CEIL_RSHIFT(get_bits_count(&s->gb), 3), 4);
> if (bytes > bytestream2_get_bytes_left(&gb)) {
> av_log(avctx, AV_LOG_ERROR, "Bitstream overread error\n");
> @@ -726,14 +814,15 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
> }
> }
> }
> -
> - if (!s->a_width || !s->a_height || s->a_format == AV_PIX_FMT_NONE ||
> - s->coded_width || s->coded_height || s->coded_format !=
> AV_PIX_FMT_NONE) {
> + //disabled to run mountain sample file
> +#if 0
> + if ((!s->a_width || !s->a_height || s->a_format == AV_PIX_FMT_NONE ||
> + s->coded_width || s->coded_height || s->coded_format !=
> AV_PIX_FMT_NONE) && s->sample_type != 1) {
> av_log(avctx, AV_LOG_ERROR, "Invalid dimensions\n");
> ret = AVERROR(EINVAL);
> goto end;
> }
> -
> +#endif
> if (!got_buffer) {
> av_log(avctx, AV_LOG_ERROR, "No end of header tag found\n");
> ret = AVERROR(EINVAL);
> @@ -741,8 +830,9 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
> }
>
> planes = av_pix_fmt_count_planes(avctx->pix_fmt);
> + if (s->transform_type == 0 && s->sample_type != 1) {
> for (plane = 0; plane < planes && !ret; plane++) {
> - /* level 1 */
> + /* level 1 */
> int lowpass_height = s->plane[plane].band[0][0].height;
> int lowpass_width = s->plane[plane].band[0][0].width;
> int highpass_stride = s->plane[plane].band[0][1].stride;
> @@ -921,14 +1011,305 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
> low = s->plane[plane].l_h[6];
> high = s->plane[plane].l_h[7];
> for (i = 0; i < lowpass_height; i++) {
> - interlaced_vertical_filter(dst, low, high, lowpass_width
> * 2, pic->linesize[act_plane]/2, act_plane);
> + inverse_temporal_filter(dst, low, high, lowpass_width *
> 2, pic->linesize[act_plane]/2, 0);
> low += lowpass_width * 2;
> high += lowpass_width * 2;
> dst += pic->linesize[act_plane];
> }
> }
> }
> + //this is the serial version on ip sample decoding so buffers
> allocated using alloc_buffers() are not freed,
> + //so the stored decoded coefficients data is used for generating the
> second frame once empty packet is passed in sample_type = 1
> + } else if (s->transform_type == 2 && s->sample_type != 1) {
> + for (plane = 0; plane < planes && !ret; plane++) {
> + int lowpass_height = s->plane[plane].band[0][0].height;
> + int lowpass_width = s->plane[plane].band[0][0].width;
> + int highpass_stride = s->plane[plane].band[0][1].stride;
> + int act_plane = plane == 1 ? 2 : plane == 2 ? 1 : plane;
> + int16_t *low, *high, *output, *dst;
> +
> + if (lowpass_height > s->plane[plane].band[0][0].a_height ||
> lowpass_width > s->plane[plane].band[0][0].a_width ||
> + !highpass_stride || s->plane[plane].band[0][1].width >
> s->plane[plane].band[0][1].a_width) {
> + av_log(avctx, AV_LOG_ERROR, "Invalid plane dimensions\n");
> + ret = AVERROR(EINVAL);
> + goto end;
> + }
> +
> + av_log(avctx, AV_LOG_DEBUG, "Decoding level 1 plane %i %i %i
> %i\n", plane, lowpass_height, lowpass_width, highpass_stride);
> +
> + low = s->plane[plane].subband[0];
> + high = s->plane[plane].subband[2];
> + output = s->plane[plane].l_h[0];
> + for (i = 0; i < lowpass_width; i++) {
> + vert_filter(output, lowpass_width, low, lowpass_width,
> high, highpass_stride, lowpass_height);
> + low++;
> + high++;
> + output++;
> + }
> +
> + low = s->plane[plane].subband[1];
> + high = s->plane[plane].subband[3];
> + output = s->plane[plane].l_h[1];
> + for (i = 0; i < lowpass_width; i++) {
> + vert_filter(output, lowpass_width, low, highpass_stride,
> high, highpass_stride, lowpass_height);
> + low++;
> + high++;
> + output++;
> + }
> +
> + low = s->plane[plane].l_h[0];
> + high = s->plane[plane].l_h[1];
> + output = s->plane[plane].subband[0];
> + for (i = 0; i < lowpass_height * 2; i++) {
> + horiz_filter(output, low, high, lowpass_width);
> + low += lowpass_width;
> + high += lowpass_width;
> + output += lowpass_width * 2;
> + }
> + if (s->bpc == 12) {
> + output = s->plane[plane].subband[0];
> + for (i = 0; i < lowpass_height * 2; i++) {
> + for (j = 0; j < lowpass_width * 2; j++)
> + output[j] *= 4;
> +
> + output += lowpass_width * 2;
> + }
> + }
> +
> + lowpass_height = s->plane[plane].band[1][1].height;
> + lowpass_width = s->plane[plane].band[1][1].width;
> + highpass_stride = s->plane[plane].band[1][1].stride;
> +
> + if (lowpass_height > s->plane[plane].band[1][1].a_height ||
> lowpass_width > s->plane[plane].band[1][1].a_width ||
> + !highpass_stride || s->plane[plane].band[1][1].width >
> s->plane[plane].band[1][1].a_width) {
> + av_log(avctx, AV_LOG_ERROR, "Invalid plane dimensions\n");
> + ret = AVERROR(EINVAL);
> + goto end;
> + }
> +
> + av_log(avctx, AV_LOG_DEBUG, "Level 2 lowpass plane %i %i %i
> %i\n", plane, lowpass_height, lowpass_width, highpass_stride);
> +
> + low = s->plane[plane].subband[0];
> + high = s->plane[plane].subband[5];
> + output = s->plane[plane].l_h[3];
> + for (i = 0; i < lowpass_width; i++) {
> + vert_filter(output, lowpass_width, low, lowpass_width,
> high, highpass_stride, lowpass_height);
> + low++;
> + high++;
> + output++;
> + }
>
> + low = s->plane[plane].subband[4];
> + high = s->plane[plane].subband[6];
> + output = s->plane[plane].l_h[4];
> + for (i = 0; i < lowpass_width; i++) {
> + vert_filter(output, lowpass_width, low, highpass_stride,
> high, highpass_stride, lowpass_height);
> + low++;
> + high++;
> + output++;
> + }
> +
> + low = s->plane[plane].l_h[3];
> + high = s->plane[plane].l_h[4];
> + output = s->plane[plane].subband[0];
> + for (i = 0; i < lowpass_height * 2; i++) {
> + horiz_filter(output, low, high, lowpass_width);
> + low += lowpass_width;
> + high += lowpass_width;
> + output += lowpass_width * 2;
> + }
> +
> + output = s->plane[plane].subband[0];
> + for (i = 0; i < lowpass_height * 2; i++) {
> + for (j = 0; j < lowpass_width * 2; j++)
> + output[j] *= 4;
> + output += lowpass_width * 2;
> + }
> +
> + low = s->plane[plane].subband[7];
> + high = s->plane[plane].subband[9];
> + output = s->plane[plane].l_h[3];
> + for (i = 0; i < lowpass_width; i++) {
> + vert_filter(output, lowpass_width, low, lowpass_width,
> high, highpass_stride, lowpass_height);
> + low++;
> + high++;
> + output++;
> + }
> +
> + low = s->plane[plane].subband[8];
> + high = s->plane[plane].subband[10];
> + output = s->plane[plane].l_h[4];
> + for (i = 0; i < lowpass_width; i++) {
> + vert_filter(output, lowpass_width, low, highpass_stride,
> high, highpass_stride, lowpass_height);
> + low++;
> + high++;
> + output++;
> + }
> +
> + low = s->plane[plane].l_h[3];
> + high = s->plane[plane].l_h[4];
> + output = s->plane[plane].subband[7];
> + for (i = 0; i < lowpass_height * 2; i++) {
> + horiz_filter(output, low, high, lowpass_width);
> + low += lowpass_width;
> + high += lowpass_width;
> + output += lowpass_width * 2;
> + }
> +
> + lowpass_height = s->plane[plane].band[4][1].height;
> + lowpass_width = s->plane[plane].band[4][1].width;
> + highpass_stride = s->plane[plane].band[4][1].stride;
> + av_log(avctx, AV_LOG_DEBUG, "temporal level %i %i %i %i\n",
> plane, lowpass_height, lowpass_width, highpass_stride);
> +
> + if (lowpass_height > s->plane[plane].band[4][1].a_height ||
> lowpass_width > s->plane[plane].band[4][1].a_width ||
> + !highpass_stride || s->plane[plane].band[4][1].width >
> s->plane[plane].band[4][1].a_width) {
> + av_log(avctx, AV_LOG_ERROR, "Invalid plane dimensions\n");
> + ret = AVERROR(EINVAL);
> + goto end;
> + }
> +
> + low = s->plane[plane].subband[0];
> + high = s->plane[plane].subband[7];
> + output = s->plane[plane].subband[0];
> + for (i = 0; i < lowpass_height; i++) {
> + inverse_temporal_filter(output, low, high, lowpass_width,
> 4 * lowpass_width * lowpass_height, 1);
> + low += lowpass_width;
> + high += lowpass_width;
> + }
> + if (s->progressive) {
> + low = s->plane[plane].subband[0];
> + high = s->plane[plane].subband[15];
> + output = s->plane[plane].l_h[6];
> + for (i = 0; i < lowpass_width; i++) {
> + vert_filter(output, lowpass_width, low,
> lowpass_width, high, highpass_stride, lowpass_height);
> + low++;
> + high++;
> + output++;
> + }
> +
> + low = s->plane[plane].subband[14];
> + high = s->plane[plane].subband[16];
> + output = s->plane[plane].l_h[7];
> + for (i = 0; i < lowpass_width; i++) {
> + vert_filter(output, lowpass_width, low,
> highpass_stride, high, highpass_stride, lowpass_height);
> + low++;
> + high++;
> + output++;
> + }
> +
> + low = s->plane[plane].subband[7];
> + high = s->plane[plane].subband[12];
> + output = s->plane[plane].l_h[8];
> + for (i = 0; i < lowpass_width; i++) {
> + vert_filter(output, lowpass_width, low,
> lowpass_width, high, highpass_stride, lowpass_height);
> + low++;
> + high++;
> + output++;
> + }
> +
> + low = s->plane[plane].subband[11];
> + high = s->plane[plane].subband[13];
> + output = s->plane[plane].l_h[9];
> + for (i = 0; i < lowpass_width; i++) {
> + vert_filter(output, lowpass_width, low,
> highpass_stride, high, highpass_stride, lowpass_height);
> + low++;
> + high++;
> + output++;
> + }
> +
> + dst = (int16_t *)pic->data[act_plane];
> + low = s->plane[plane].l_h[6];
> + high = s->plane[plane].l_h[7];
> + for (i = 0; i < lowpass_height * 2; i++) {
> + horiz_filter(dst, low, high, lowpass_width);
> + low += lowpass_width;
> + high += lowpass_width;
> + dst += pic->linesize[act_plane] / 2;
> + }
> + } else {
> + pic->interlaced_frame = 1;
> + low = s->plane[plane].subband[0];
> + high = s->plane[plane].subband[14];
> + output = s->plane[plane].l_h[6];
> + for (i = 0; i < lowpass_height; i++) {
> + horiz_filter(output, low, high, lowpass_width);
> + low += lowpass_width;
> + high += lowpass_width;
> + output += lowpass_width * 2;
> + }
> +
> + low = s->plane[plane].subband[15];
> + high = s->plane[plane].subband[16];
> + output = s->plane[plane].l_h[7];
> + for (i = 0; i < lowpass_height; i++) {
> + horiz_filter(output, low, high, lowpass_width);
> + low += lowpass_width;
> + high += lowpass_width;
> + output += lowpass_width * 2;
> + }
> +
> + low = s->plane[plane].subband[7];
> + high = s->plane[plane].subband[11];
> + output = s->plane[plane].l_h[8];
> + for (i = 0; i < lowpass_height; i++) {
> + horiz_filter(output, low, high, lowpass_width);
> + low += lowpass_width;
> + high += lowpass_width;
> + output += lowpass_width * 2;
> + }
> +
> + low = s->plane[plane].subband[12];
> + high = s->plane[plane].subband[13];
> + output = s->plane[plane].l_h[9];
> + for (i = 0; i < lowpass_height; i++) {
> + horiz_filter(output, low, high, lowpass_width);
> + low += lowpass_width;
> + high += lowpass_width;
> + output += lowpass_width * 2;
> + }
> +
> +
> + dst = (int16_t *)pic->data[act_plane];
> + low = s->plane[plane].l_h[6];
> + high = s->plane[plane].l_h[7];
> + for (i = 0; i < lowpass_height; i++) {
> + inverse_temporal_filter(dst, low, high, lowpass_width
> * 2, pic->linesize[act_plane]/2, 0);
> + low += lowpass_width * 2;
> + high += lowpass_width * 2;
> + dst += pic->linesize[act_plane];
> + }
> + }
> + }
> + } else if (s->sample_type == 1) {
> + int16_t *low, *high, *dst;
> + int lowpass_height, lowpass_width;
> + for (plane = 0; plane < planes && !ret; plane++) {
> + int act_plane = plane == 1 ? 2 : plane == 2 ? 1 : plane;
> + lowpass_height = s->plane[plane].band[4][1].height;
> + lowpass_width = s->plane[plane].band[4][1].width;
> + if (s->progressive) {
> + dst = (int16_t *)pic->data[act_plane];
> + low = s->plane[plane].l_h[8];
> + high = s->plane[plane].l_h[9];
> + for (i = 0; i < lowpass_height * 2; i++) {
> + horiz_filter(dst, low, high, lowpass_width);
> + low += lowpass_width;
> + high += lowpass_width;
> + dst += pic->linesize[act_plane] / 2;
> + }
> + } else {
> + dst = (int16_t *)pic->data[act_plane];
> + low = s->plane[plane].l_h[8];
> + high = s->plane[plane].l_h[9];
> + for (i = 0; i < lowpass_height; i++) {
> + inverse_temporal_filter(dst, low, high, lowpass_width
> * 2, pic->linesize[act_plane]/2, 0);
> + low += lowpass_width * 2;
> + high += lowpass_width * 2;
> + dst += pic->linesize[act_plane];
> + }
> + }
> + }
> + }
>
> end:
> if (ret < 0)
> @@ -961,6 +1342,6 @@ AVCodec ff_cfhd_decoder = {
> .init = cfhd_init,
> .close = cfhd_close,
> .decode = cfhd_decode,
> - .capabilities = AV_CODEC_CAP_DR1 | AV_CODEC_CAP_FRAME_THREADS,
> - .caps_internal = FF_CODEC_CAP_INIT_THREADSAFE |
> FF_CODEC_CAP_INIT_CLEANUP,
> + .capabilities = AV_CODEC_CAP_DR1,
> + .caps_internal = FF_CODEC_CAP_INIT_CLEANUP,
> };
> diff --git a/libavcodec/cfhd.h b/libavcodec/cfhd.h
> index 4f2c82d8bc..047c0f2028 100644
> --- a/libavcodec/cfhd.h
> +++ b/libavcodec/cfhd.h
> @@ -31,15 +31,14 @@
> #include "vlc.h"
>
> #define VLC_BITS 9
> -#define SUBBAND_COUNT 10
> -
> +#define SUBBAND_COUNT 17
> typedef struct CFHD_RL_VLC_ELEM {
> int16_t level;
> int8_t len;
> uint16_t run;
> } CFHD_RL_VLC_ELEM;
>
> -#define DWT_LEVELS 3
> +#define DWT_LEVELS 6
>
> typedef struct SubBand {
> int level;
> @@ -64,7 +63,7 @@ typedef struct Plane {
>
> /* TODO: merge this into SubBand structure */
> int16_t *subband[SUBBAND_COUNT];
> - int16_t *l_h[8];
> + int16_t *l_h[10];
>
> SubBand band[DWT_LEVELS][4];
> } Plane;
> @@ -83,9 +82,13 @@ typedef struct CFHDContext {
>
> CFHD_RL_VLC_ELEM table_18_rl_vlc[4572];
> VLC vlc_18;
> -
> GetBitContext gb;
>
> + int sample_type;
> + int transform_type;
> + int encode_method;
> + int first_wavelet;
> + int pframe;
> int coded_width;
> int coded_height;
> int cropped_height;
> --
> 2.17.1
>
>
> From d6ac8a4ee2f08a6d3cab7e7500d740c661062c9f Mon Sep 17 00:00:00 2001
> From: Gagandeep Singh <deepgagan231197 at gmail.com>
> Date: Tue, 14 Aug 2018 00:20:06 +0530
> Subject: [GSOC][FFmpeg-devel][PATCH 2/3] lavc/cfhd:corrected decompanding
> for table 9 used in cfhd
>
> ---
> libavcodec/cfhd.c | 25 ++++++++++++++++++++++++-
> 1 file changed, 24 insertions(+), 1 deletion(-)
>
> diff --git a/libavcodec/cfhd.c b/libavcodec/cfhd.c
> index 3906d7a86d..2c538f0bbd 100644
> --- a/libavcodec/cfhd.c
> +++ b/libavcodec/cfhd.c
> @@ -114,7 +114,30 @@ static inline int dequant_and_decompand(int level,
> int quantisation, int codeboo
> {
> if (lossless)
> return level;
> - if (codebook == 0 || codebook == 1) {
> + if (codebook == 0) {
> + if (level >= 40 && level < 264) {
> + if (level >= 54) {
> + level -= 54;
> + level <<= 2;
> + level += 54;
> + }
> + level -= 40;
> + level <<= 2;
> + level += 40;
> + } else if (level <= -40) {
> + level = -level;
> + if (level >= 54) {
> + level -= 54;
> + level <<= 2;
> + level += 54;
> + }
> + level -= 40;
> + level <<= 2;
> + level += 40;
> + level = -level;
> + }
> + return level * quantisation;
> + } else if (codebook == 1) {
> int64_t abslevel = abs(level);
> if (level < 264)
> return (abslevel + ((768 * abslevel * abslevel * abslevel) /
> (255 * 255 * 255))) *
> --
> 2.17.1
>
>
> From fa23549c61a6d8413cdc79c570376c53795a6ff1 Mon Sep 17 00:00:00 2001
> From: Gagandeep Singh <deepgagan231197 at gmail.com>
> Date: Tue, 14 Aug 2018 12:43:20 +0530
> Subject: [GSOC][FFmpeg-devel][PATCH 3/3] lavc/cfhd:frame threading support
> for 3d transform
> progressive and interlaced samples
>
> ---
> libavcodec/cfhd.c | 378 ++++++++++++++++++++++++++++------------------
> libavcodec/cfhd.h | 8 +-
> 2 files changed, 242 insertions(+), 144 deletions(-)
>
> diff --git a/libavcodec/cfhd.c b/libavcodec/cfhd.c
> index 2c538f0bbd..7c298056ca 100644
> --- a/libavcodec/cfhd.c
> +++ b/libavcodec/cfhd.c
> @@ -63,13 +63,23 @@ enum CFHDParam {
>
> static av_cold int cfhd_init(AVCodecContext *avctx)
> {
> + int ret;
> +
> CFHDContext *s = avctx->priv_data;
> + if (!avctx->internal->is_copy) {
> + avctx->internal->allocate_progress = 1;
> + ret = ff_cfhd_init_vlcs(s);
> + } else
> + ret = 0;
>
> avctx->bits_per_raw_sample = 10;
> s->avctx = avctx;
> s->progressive = 0;
> + s->i_frame.f = av_frame_alloc();
> + s->p_frame.f = av_frame_alloc();
>
> - return ff_cfhd_init_vlcs(s);
> +
> + return ret;
> }
>
> static void init_plane_defaults(CFHDContext *s)
> @@ -268,15 +278,18 @@ static void free_buffers(CFHDContext *s)
> for (i = 0; i < FF_ARRAY_ELEMS(s->plane); i++) {
> av_freep(&s->plane[i].idwt_buf);
> av_freep(&s->plane[i].idwt_tmp);
> - if (s->transform_type == 0)
> + if (s->transform_type == 0) {
> for (j = 0; j < 9; j++)
> s->plane[i].subband[j] = NULL;
> - else
> + for (j = 0; j < 8; j++)
> + s->plane[i].l_h[j] = NULL;
> + }
> + else {
> for (j = 0; j < 17; j++)
> s->plane[i].subband[j] = NULL;
> -
> - for (j = 0; j < 8; j++)
> - s->plane[i].l_h[j] = NULL;
> + for (j = 0; j < 12; j++)
> + s->plane[i].l_h[j] = NULL;
> + }
> }
> s->a_height = 0;
> s->a_width = 0;
> @@ -394,8 +407,10 @@ static int alloc_buffers(AVCodecContext *avctx)
> s->plane[i].l_h[7] = s->plane[i].idwt_tmp + 2 * w2 * h2;
> if (s->transform_type == 2) {
> frame2 = s->plane[i].idwt_tmp + 4 * w2 * h2;
> - s->plane[i].l_h[8] = frame2;
> - s->plane[i].l_h[9] = frame2 + 2 * w2 * h2;
> + s->plane[i].l_h[8] = frame2;
> + s->plane[i].l_h[9] = frame2 + 2 * w4 * h4;
> + s->plane[i].l_h[10] = frame2;
> + s->plane[i].l_h[11] = frame2 + 2 * w2 * h2;
> }
> }
>
> @@ -406,14 +421,28 @@ static int alloc_buffers(AVCodecContext *avctx)
> return 0;
> }
>
> +static int update_thread_context(AVCodecContext *dst, const
> AVCodecContext *src)
> +{
> + CFHDContext *csrc = src->priv_data;
> + CFHDContext *cdst = dst->priv_data;
> + cdst->transform_type = csrc->transform_type;
> + if (csrc->sample_type != 1 && csrc->transform_type != 0) {
> + cdst->progressive = csrc->progressive;
> + cdst->picture = &csrc->p_frame;
> + cdst->connection = &csrc->i_frame;
> + cdst->buffers = csrc->plane;
> + }
> +
> + return 0;
> +}
> +
> static int cfhd_decode(AVCodecContext *avctx, void *data, int *got_frame,
> AVPacket *avpkt)
> {
> CFHDContext *s = avctx->priv_data;
> GetByteContext gb;
> ThreadFrame frame = { .f = data };
> - AVFrame *pic = data;
> - int ret = 0, i, j, planes, plane, got_buffer = 0;
> + int ret = 0, i, j, planes, plane, got_buffer = 0, progress1 = 1,
> progress2 = 1;
> int16_t *coeff_data;
>
> s->coded_format = AV_PIX_FMT_YUV422P10;
> @@ -537,7 +566,9 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
> } else if (tag == 1) {
> s->sample_type = data;
> if (data == 2)
> - s->pframe = 1;
> + s->pframe = 1;
> + else if (data == 1)
> + s->transform_type = 2;
> av_log(avctx, AV_LOG_DEBUG, "Sample type? %"PRIu16"\n", data);
> } else if (tag == 10) {
> s->transform_type = data;
> @@ -657,21 +688,54 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
> return ret;
> }
> }
> - ret = ff_set_dimensions(avctx, s->coded_width,
> s->coded_height);
> - if (ret < 0)
> - return ret;
> - if (s->cropped_height)
> - avctx->height = s->cropped_height;
> - frame.f->width =
> - frame.f->height = 0;
> -
> - if ((ret = ff_thread_get_buffer(avctx, &frame, 0)) < 0)
> - return ret;
> -
> + if (s->transform_type == 2) {
> + if (s->sample_type != 1) {
> + s->picture = &s->i_frame;
> + s->connection = &s->p_frame;
> + s->buffers = s->plane;
> + }
> + ret = ff_set_dimensions(avctx, s->coded_width,
> s->coded_height);
> + if (ret < 0)
> + return ret;
> + if (s->sample_type != 1) {
> + if (s->i_frame.f->data[0])
> + ff_thread_release_buffer(avctx, &s->i_frame);
> + if (s->p_frame.f->data[0])
> + ff_thread_release_buffer(avctx, &s->p_frame);
> + av_frame_copy_props(s->i_frame.f, frame.f);
> + av_frame_copy_props(s->p_frame.f, frame.f);
> + if (s->cropped_height)
> + avctx->height = s->cropped_height;
> + s->picture->f->width =
> + s->picture->f->height = 0;
> + s->connection->f->width =
> + s->connection->f->height = 0;
> + if ((ret = ff_thread_get_buffer(avctx, s->picture,
> 0)) < 0)
> + return ret;
> + if ((ret = ff_thread_get_buffer(avctx, s->connection,
> 0)) < 0)
> + return ret;
> + }
> + } else {
> + s->picture = &s->i_frame;
> + s->buffers = s->plane;
> + if (s->picture->f->data[0])
> + ff_thread_release_buffer(avctx, s->picture);
> + av_frame_copy_props(s->i_frame.f, frame.f);
> + ret = ff_set_dimensions(avctx, s->coded_width,
> s->coded_height);
> + if (ret < 0)
> + return ret;
> + if (s->cropped_height)
> + avctx->height = s->cropped_height;
> + s->picture->f->width =
> + s->picture->f->height = 0;
> + if ((ret = ff_thread_get_buffer(avctx, s->picture, 0)) <
> 0)
> + return ret;
> + }
> s->coded_width = 0;
> s->coded_height = 0;
> s->coded_format = AV_PIX_FMT_NONE;
> got_buffer = 1;
> + ff_thread_finish_setup(avctx);
> }
> coeff_data =
> s->plane[s->channel_num].subband[s->subband_num_actual];
>
> @@ -835,6 +899,8 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
> &coeff_data[(highpass_height - 1) *
> highpass_stride],
> highpass_stride * sizeof(*coeff_data));
> }
> + if (s->transform_type == 2 && s->subband_num_actual == 10)
> + ff_thread_report_progress(s->picture, progress1 += 1, 0);
> }
> }
> //disabled to run mountain sample file
> @@ -975,7 +1041,6 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
> ret = AVERROR(EINVAL);
> goto end;
> }
> -
> av_log(avctx, AV_LOG_DEBUG, "Level 3 plane %i %i %i %i\n", plane,
> lowpass_height, lowpass_width, highpass_stride);
> if (s->progressive) {
> low = s->plane[plane].subband[0];
> @@ -998,18 +1063,18 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
> output++;
> }
>
> - dst = (int16_t *)pic->data[act_plane];
> + dst = (int16_t *)s->picture->f->data[act_plane];
> low = s->plane[plane].l_h[6];
> high = s->plane[plane].l_h[7];
> for (i = 0; i < lowpass_height * 2; i++) {
> horiz_filter_clip(dst, low, high, lowpass_width, s->bpc);
> low += lowpass_width;
> high += lowpass_width;
> - dst += pic->linesize[act_plane] / 2;
> + dst += s->picture->f->linesize[act_plane] / 2;
> }
> } else {
> - av_log(avctx, AV_LOG_DEBUG, "interlaced frame ? %d",
> pic->interlaced_frame);
> - pic->interlaced_frame = 1;
> + av_log(avctx, AV_LOG_DEBUG, "interlaced frame ? %d",
> s->picture->f->interlaced_frame);
> + s->picture->f->interlaced_frame = 1;
> low = s->plane[plane].subband[0];
> high = s->plane[plane].subband[7];
> output = s->plane[plane].l_h[6];
> @@ -1030,23 +1095,23 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
> output += lowpass_width * 2;
> }
>
> - dst = (int16_t *)pic->data[act_plane];
> + dst = (int16_t *)s->picture->f->data[act_plane];
> low = s->plane[plane].l_h[6];
> high = s->plane[plane].l_h[7];
> for (i = 0; i < lowpass_height; i++) {
> - inverse_temporal_filter(dst, low, high, lowpass_width *
> 2, pic->linesize[act_plane]/2, 0);
> + inverse_temporal_filter(dst, low, high, lowpass_width *
> 2, s->picture->f->linesize[act_plane]/2, 0);
> low += lowpass_width * 2;
> high += lowpass_width * 2;
> - dst += pic->linesize[act_plane];
> + dst += s->picture->f->linesize[act_plane];
> }
> }
> }
> - //this is the serial version on ip sample decoding so buffers
> allocated using alloc_buffers() are not freed,
> - //so the stored decoded coefficients data is used for generating the
> second frame once empty packet is passed in sample_type = 1
> + av_frame_ref(frame.f, s->picture->f);
> + ff_thread_report_progress(s->picture, INT_MAX, 0);
> } else if (s->transform_type == 2 && s->sample_type != 1) {
> for (plane = 0; plane < planes && !ret; plane++) {
> - int lowpass_height = s->plane[plane].band[0][0].height;
> - int lowpass_width = s->plane[plane].band[0][0].width;
> + int lowpass_height = s->plane[plane].band[0][1].height;
> + int lowpass_width = s->plane[plane].band[0][1].width;
> int highpass_stride = s->plane[plane].band[0][1].stride;
> int act_plane = plane == 1 ? 2 : plane == 2 ? 1 : plane;
> int16_t *low, *high, *output, *dst;
> @@ -1058,8 +1123,6 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
> goto end;
> }
>
> - av_log(avctx, AV_LOG_DEBUG, "Decoding level 1 plane %i %i %i
> %i\n", plane, lowpass_height, lowpass_width, highpass_stride);
> -
> low = s->plane[plane].subband[0];
> high = s->plane[plane].subband[2];
> output = s->plane[plane].l_h[0];
> @@ -1110,8 +1173,6 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
> goto end;
> }
>
> - av_log(avctx, AV_LOG_DEBUG, "Level 2 lowpass plane %i %i %i
> %i\n", plane, lowpass_height, lowpass_width, highpass_stride);
> -
> low = s->plane[plane].subband[0];
> high = s->plane[plane].subband[5];
> output = s->plane[plane].l_h[3];
> @@ -1149,40 +1210,9 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
> output += lowpass_width * 2;
> }
>
> - low = s->plane[plane].subband[7];
> - high = s->plane[plane].subband[9];
> - output = s->plane[plane].l_h[3];
> - for (i = 0; i < lowpass_width; i++) {
> - vert_filter(output, lowpass_width, low, lowpass_width,
> high, highpass_stride, lowpass_height);
> - low++;
> - high++;
> - output++;
> - }
> -
> - low = s->plane[plane].subband[8];
> - high = s->plane[plane].subband[10];
> - output = s->plane[plane].l_h[4];
> - for (i = 0; i < lowpass_width; i++) {
> - vert_filter(output, lowpass_width, low, highpass_stride,
> high, highpass_stride, lowpass_height);
> - low++;
> - high++;
> - output++;
> - }
> -
> - low = s->plane[plane].l_h[3];
> - high = s->plane[plane].l_h[4];
> - output = s->plane[plane].subband[7];
> - for (i = 0; i < lowpass_height * 2; i++) {
> - horiz_filter(output, low, high, lowpass_width);
> - low += lowpass_width;
> - high += lowpass_width;
> - output += lowpass_width * 2;
> - }
> -
> lowpass_height = s->plane[plane].band[4][1].height;
> lowpass_width = s->plane[plane].band[4][1].width;
> highpass_stride = s->plane[plane].band[4][1].stride;
> - av_log(avctx, AV_LOG_DEBUG, "temporal level %i %i %i %i\n",
> plane, lowpass_height, lowpass_width, highpass_stride);
>
> if (lowpass_height > s->plane[plane].band[4][1].a_height ||
> lowpass_width > s->plane[plane].band[4][1].a_width ||
> !highpass_stride || s->plane[plane].band[4][1].width >
> s->plane[plane].band[4][1].a_width) {
> @@ -1190,7 +1220,7 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
> ret = AVERROR(EINVAL);
> goto end;
> }
> -
> + ff_thread_await_progress(s->connection, progress2 += 1, 0);
> low = s->plane[plane].subband[0];
> high = s->plane[plane].subband[7];
> output = s->plane[plane].subband[0];
> @@ -1199,6 +1229,7 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
> low += lowpass_width;
> high += lowpass_width;
> }
> + ff_thread_report_progress(s->picture, progress1 += 1, 0);
> if (s->progressive) {
> low = s->plane[plane].subband[0];
> high = s->plane[plane].subband[15];
> @@ -1220,37 +1251,17 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
> output++;
> }
>
> - low = s->plane[plane].subband[7];
> - high = s->plane[plane].subband[12];
> - output = s->plane[plane].l_h[8];
> - for (i = 0; i < lowpass_width; i++) {
> - vert_filter(output, lowpass_width, low,
> lowpass_width, high, highpass_stride, lowpass_height);
> - low++;
> - high++;
> - output++;
> - }
> -
> - low = s->plane[plane].subband[11];
> - high = s->plane[plane].subband[13];
> - output = s->plane[plane].l_h[9];
> - for (i = 0; i < lowpass_width; i++) {
> - vert_filter(output, lowpass_width, low,
> highpass_stride, high, highpass_stride, lowpass_height);
> - low++;
> - high++;
> - output++;
> - }
> -
> - dst = (int16_t *)pic->data[act_plane];
> + dst = (int16_t *)s->picture->f->data[act_plane];
> low = s->plane[plane].l_h[6];
> high = s->plane[plane].l_h[7];
> for (i = 0; i < lowpass_height * 2; i++) {
> - horiz_filter(dst, low, high, lowpass_width);
> + horiz_filter_clip(dst, low, high, lowpass_width,
> s->bpc);
> low += lowpass_width;
> high += lowpass_width;
> - dst += pic->linesize[act_plane] / 2;
> + dst += s->picture->f->linesize[act_plane] / 2;
> }
> } else {
> - pic->interlaced_frame = 1;
> + s->picture->f->interlaced_frame = 1;
> low = s->plane[plane].subband[0];
> high = s->plane[plane].subband[14];
> output = s->plane[plane].l_h[6];
> @@ -1271,67 +1282,137 @@ static int cfhd_decode(AVCodecContext *avctx,
> void *data, int *got_frame,
> output += lowpass_width * 2;
> }
>
> - low = s->plane[plane].subband[7];
> - high = s->plane[plane].subband[11];
> - output = s->plane[plane].l_h[8];
> - for (i = 0; i < lowpass_height; i++) {
> - horiz_filter(output, low, high, lowpass_width);
> - low += lowpass_width;
> - high += lowpass_width;
> - output += lowpass_width * 2;
> - }
> -
> - low = s->plane[plane].subband[12];
> - high = s->plane[plane].subband[13];
> - output = s->plane[plane].l_h[9];
> - for (i = 0; i < lowpass_height; i++) {
> - horiz_filter(output, low, high, lowpass_width);
> - low += lowpass_width;
> - high += lowpass_width;
>
discard this thread, really sorry for my skills in sending patches
Gagandeep Singh
More information about the ffmpeg-devel
mailing list