[FFmpeg-devel] [GSOC][PATCH 3/3] lavc/cfhd:frame threading support for 3d transform progressive and interlaced samples

Gagandeep Singh deepgagan231197 at gmail.com
Tue Aug 14 10:46:56 EEST 2018


On Tue, Aug 14, 2018 at 12:58 PM Gagandeep Singh <deepgagan231197 at gmail.com>
wrote:

> ---
>  libavcodec/cfhd.c | 378 ++++++++++++++++++++++++++++------------------
>  libavcodec/cfhd.h |   8 +-
>  2 files changed, 242 insertions(+), 144 deletions(-)
>
> diff --git a/libavcodec/cfhd.c b/libavcodec/cfhd.c
> index 2c538f0bbd..7c298056ca 100644
> --- a/libavcodec/cfhd.c
> +++ b/libavcodec/cfhd.c
> @@ -63,13 +63,23 @@ enum CFHDParam {
>
>  static av_cold int cfhd_init(AVCodecContext *avctx)
>  {
> +    int ret;
> +
>      CFHDContext *s = avctx->priv_data;
> +    if (!avctx->internal->is_copy) {
> +        avctx->internal->allocate_progress = 1;
> +        ret = ff_cfhd_init_vlcs(s);
> +    } else
> +        ret = 0;
>
>      avctx->bits_per_raw_sample = 10;
>      s->avctx                   = avctx;
>      s->progressive             = 0;
> +    s->i_frame.f = av_frame_alloc();
> +    s->p_frame.f = av_frame_alloc();
>
> -    return ff_cfhd_init_vlcs(s);
> +
> +    return ret;
>  }
>
>  static void init_plane_defaults(CFHDContext *s)
> @@ -268,15 +278,18 @@ static void free_buffers(CFHDContext *s)
>      for (i = 0; i < FF_ARRAY_ELEMS(s->plane); i++) {
>          av_freep(&s->plane[i].idwt_buf);
>          av_freep(&s->plane[i].idwt_tmp);
> -        if (s->transform_type == 0)
> +        if (s->transform_type == 0) {
>              for (j = 0; j < 9; j++)
>                  s->plane[i].subband[j] = NULL;
> -        else
> +            for (j = 0; j < 8; j++)
> +                s->plane[i].l_h[j] = NULL;
> +        }
> +        else {
>              for (j = 0; j < 17; j++)
>                  s->plane[i].subband[j] = NULL;
> -
> -        for (j = 0; j < 8; j++)
> -            s->plane[i].l_h[j] = NULL;
> +            for (j = 0; j < 12; j++)
> +                s->plane[i].l_h[j] = NULL;
> +        }
>      }
>      s->a_height = 0;
>      s->a_width  = 0;
> @@ -394,8 +407,10 @@ static int alloc_buffers(AVCodecContext *avctx)
>          s->plane[i].l_h[7] = s->plane[i].idwt_tmp + 2 * w2 * h2;
>          if (s->transform_type == 2) {
>              frame2 = s->plane[i].idwt_tmp + 4 * w2 * h2;
> -            s->plane[i].l_h[8] = frame2;
> -            s->plane[i].l_h[9] = frame2 + 2 * w2 * h2;
> +            s->plane[i].l_h[8]  = frame2;
> +            s->plane[i].l_h[9]  = frame2 + 2 * w4 * h4;
> +            s->plane[i].l_h[10] = frame2;
> +            s->plane[i].l_h[11] = frame2 + 2 * w2 * h2;
>              }
>      }
>
> @@ -406,14 +421,28 @@ static int alloc_buffers(AVCodecContext *avctx)
>      return 0;
>  }
>
> +static int update_thread_context(AVCodecContext *dst, const
> AVCodecContext *src)
> +{
> +    CFHDContext *csrc = src->priv_data;
> +    CFHDContext *cdst = dst->priv_data;
> +    cdst->transform_type = csrc->transform_type;
> +    if (csrc->sample_type != 1 && csrc->transform_type != 0) {
> +        cdst->progressive = csrc->progressive;
> +        cdst->picture = &csrc->p_frame;
> +        cdst->connection = &csrc->i_frame;
> +        cdst->buffers = csrc->plane;
> +    }
> +
> +    return 0;
> +}
> +
>  static int cfhd_decode(AVCodecContext *avctx, void *data, int *got_frame,
>                         AVPacket *avpkt)
>  {
>      CFHDContext *s = avctx->priv_data;
>      GetByteContext gb;
>      ThreadFrame frame = { .f = data };
> -    AVFrame *pic = data;
> -    int ret = 0, i, j, planes, plane, got_buffer = 0;
> +    int ret = 0, i, j, planes, plane, got_buffer = 0, progress1 = 1,
> progress2 = 1;
>      int16_t *coeff_data;
>
>      s->coded_format = AV_PIX_FMT_YUV422P10;
> @@ -537,7 +566,9 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
>          } else if (tag == 1) {
>              s->sample_type = data;
>              if (data == 2)
> -                s->pframe = 1;
> +                s->pframe  = 1;
> +            else if (data == 1)
> +                s->transform_type = 2;
>              av_log(avctx, AV_LOG_DEBUG, "Sample type? %"PRIu16"\n", data);
>          } else if (tag == 10) {
>              s->transform_type = data;
> @@ -657,21 +688,54 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
>                      return ret;
>                  }
>              }
> -            ret = ff_set_dimensions(avctx, s->coded_width,
> s->coded_height);
> -            if (ret < 0)
> -                return ret;
> -            if (s->cropped_height)
> -                avctx->height = s->cropped_height;
> -            frame.f->width =
> -            frame.f->height = 0;
> -
> -            if ((ret = ff_thread_get_buffer(avctx, &frame, 0)) < 0)
> -                return ret;
> -
> +            if (s->transform_type == 2) {
> +                if (s->sample_type != 1) {
> +                    s->picture = &s->i_frame;
> +                    s->connection = &s->p_frame;
> +                    s->buffers = s->plane;
> +                }
> +                ret = ff_set_dimensions(avctx, s->coded_width,
> s->coded_height);
> +                if (ret < 0)
> +                    return ret;
> +                if (s->sample_type != 1) {
> +                    if (s->i_frame.f->data[0])
> +                        ff_thread_release_buffer(avctx, &s->i_frame);
> +                    if (s->p_frame.f->data[0])
> +                        ff_thread_release_buffer(avctx, &s->p_frame);
> +                    av_frame_copy_props(s->i_frame.f, frame.f);
> +                    av_frame_copy_props(s->p_frame.f, frame.f);
> +                    if (s->cropped_height)
> +                        avctx->height = s->cropped_height;
> +                    s->picture->f->width =
> +                    s->picture->f->height = 0;
> +                    s->connection->f->width =
> +                    s->connection->f->height = 0;
> +                    if ((ret = ff_thread_get_buffer(avctx, s->picture,
> 0)) < 0)
> +                        return ret;
> +                    if ((ret = ff_thread_get_buffer(avctx, s->connection,
> 0)) < 0)
> +                        return ret;
> +                }
> +            } else {
> +                s->picture = &s->i_frame;
> +                s->buffers = s->plane;
> +                if (s->picture->f->data[0])
> +                    ff_thread_release_buffer(avctx, s->picture);
> +                av_frame_copy_props(s->i_frame.f, frame.f);
> +                ret = ff_set_dimensions(avctx, s->coded_width,
> s->coded_height);
> +                if (ret < 0)
> +                    return ret;
> +                if (s->cropped_height)
> +                    avctx->height = s->cropped_height;
> +                s->picture->f->width =
> +                s->picture->f->height = 0;
> +                if ((ret = ff_thread_get_buffer(avctx, s->picture, 0)) <
> 0)
> +                    return ret;
> +            }
>              s->coded_width = 0;
>              s->coded_height = 0;
>              s->coded_format = AV_PIX_FMT_NONE;
>              got_buffer = 1;
> +            ff_thread_finish_setup(avctx);
>          }
>          coeff_data =
> s->plane[s->channel_num].subband[s->subband_num_actual];
>
> @@ -835,6 +899,8 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
>                         &coeff_data[(highpass_height - 1) *
> highpass_stride],
>                         highpass_stride * sizeof(*coeff_data));
>              }
> +            if (s->transform_type == 2 && s->subband_num_actual == 10)
> +                ff_thread_report_progress(s->picture, progress1 += 1, 0);
>          }
>      }
>      //disabled to run mountain sample file
> @@ -975,7 +1041,6 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
>              ret = AVERROR(EINVAL);
>              goto end;
>          }
> -
>          av_log(avctx, AV_LOG_DEBUG, "Level 3 plane %i %i %i %i\n", plane,
> lowpass_height, lowpass_width, highpass_stride);
>          if (s->progressive) {
>              low    = s->plane[plane].subband[0];
> @@ -998,18 +1063,18 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
>                  output++;
>              }
>
> -            dst = (int16_t *)pic->data[act_plane];
> +            dst = (int16_t *)s->picture->f->data[act_plane];
>              low  = s->plane[plane].l_h[6];
>              high = s->plane[plane].l_h[7];
>              for (i = 0; i < lowpass_height * 2; i++) {
>                  horiz_filter_clip(dst, low, high, lowpass_width, s->bpc);
>                  low  += lowpass_width;
>                  high += lowpass_width;
> -                dst  += pic->linesize[act_plane] / 2;
> +                dst  += s->picture->f->linesize[act_plane] / 2;
>              }
>          } else {
> -            av_log(avctx, AV_LOG_DEBUG, "interlaced frame ? %d",
> pic->interlaced_frame);
> -            pic->interlaced_frame = 1;
> +            av_log(avctx, AV_LOG_DEBUG, "interlaced frame ? %d",
> s->picture->f->interlaced_frame);
> +            s->picture->f->interlaced_frame = 1;
>              low    = s->plane[plane].subband[0];
>              high   = s->plane[plane].subband[7];
>              output = s->plane[plane].l_h[6];
> @@ -1030,23 +1095,23 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
>                  output += lowpass_width * 2;
>              }
>
> -            dst  = (int16_t *)pic->data[act_plane];
> +            dst  = (int16_t *)s->picture->f->data[act_plane];
>              low  = s->plane[plane].l_h[6];
>              high = s->plane[plane].l_h[7];
>              for (i = 0; i < lowpass_height; i++) {
> -                inverse_temporal_filter(dst, low, high, lowpass_width *
> 2,  pic->linesize[act_plane]/2, 0);
> +                inverse_temporal_filter(dst, low, high, lowpass_width *
> 2,  s->picture->f->linesize[act_plane]/2, 0);
>                  low  += lowpass_width * 2;
>                  high += lowpass_width * 2;
> -                dst  += pic->linesize[act_plane];
> +                dst  += s->picture->f->linesize[act_plane];
>              }
>          }
>      }
> -    //this is the serial version on ip sample decoding so buffers
> allocated using alloc_buffers() are not freed,
> -    //so the stored decoded coefficients data is used for generating the
> second frame once empty packet is passed in sample_type = 1
> +    av_frame_ref(frame.f, s->picture->f);
> +    ff_thread_report_progress(s->picture, INT_MAX, 0);
>      } else if (s->transform_type == 2 && s->sample_type != 1) {
>          for (plane = 0; plane < planes && !ret; plane++) {
> -            int lowpass_height  = s->plane[plane].band[0][0].height;
> -            int lowpass_width   = s->plane[plane].band[0][0].width;
> +            int lowpass_height  = s->plane[plane].band[0][1].height;
> +            int lowpass_width   = s->plane[plane].band[0][1].width;
>              int highpass_stride = s->plane[plane].band[0][1].stride;
>              int act_plane = plane == 1 ? 2 : plane == 2 ? 1 : plane;
>              int16_t *low, *high, *output, *dst;
> @@ -1058,8 +1123,6 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
>                  goto end;
>              }
>
> -            av_log(avctx, AV_LOG_DEBUG, "Decoding level 1 plane %i %i %i
> %i\n", plane, lowpass_height, lowpass_width, highpass_stride);
> -
>              low    = s->plane[plane].subband[0];
>              high   = s->plane[plane].subband[2];
>              output = s->plane[plane].l_h[0];
> @@ -1110,8 +1173,6 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
>                  goto end;
>              }
>
> -            av_log(avctx, AV_LOG_DEBUG, "Level 2 lowpass plane %i %i %i
> %i\n", plane, lowpass_height, lowpass_width, highpass_stride);
> -
>              low    = s->plane[plane].subband[0];
>              high   = s->plane[plane].subband[5];
>              output = s->plane[plane].l_h[3];
> @@ -1149,40 +1210,9 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
>                  output += lowpass_width * 2;
>              }
>
> -            low    = s->plane[plane].subband[7];
> -            high   = s->plane[plane].subband[9];
> -            output = s->plane[plane].l_h[3];
> -            for (i = 0; i < lowpass_width; i++) {
> -                vert_filter(output, lowpass_width, low, lowpass_width,
> high, highpass_stride, lowpass_height);
> -                low++;
> -                high++;
> -                output++;
> -            }
> -
> -            low    = s->plane[plane].subband[8];
> -            high   = s->plane[plane].subband[10];
> -            output = s->plane[plane].l_h[4];
> -            for (i = 0; i < lowpass_width; i++) {
> -                vert_filter(output, lowpass_width, low, highpass_stride,
> high, highpass_stride, lowpass_height);
> -                low++;
> -                high++;
> -                output++;
> -            }
> -
> -            low    = s->plane[plane].l_h[3];
> -            high   = s->plane[plane].l_h[4];
> -            output = s->plane[plane].subband[7];
> -            for (i = 0; i < lowpass_height * 2; i++) {
> -                horiz_filter(output, low, high, lowpass_width);
> -                low    += lowpass_width;
> -                high   += lowpass_width;
> -                output += lowpass_width * 2;
> -            }
> -
>              lowpass_height  = s->plane[plane].band[4][1].height;
>              lowpass_width   = s->plane[plane].band[4][1].width;
>              highpass_stride = s->plane[plane].band[4][1].stride;
> -            av_log(avctx, AV_LOG_DEBUG, "temporal level %i %i %i %i\n",
> plane, lowpass_height, lowpass_width, highpass_stride);
>
>              if (lowpass_height > s->plane[plane].band[4][1].a_height ||
> lowpass_width > s->plane[plane].band[4][1].a_width ||
>                  !highpass_stride || s->plane[plane].band[4][1].width >
> s->plane[plane].band[4][1].a_width) {
> @@ -1190,7 +1220,7 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
>                  ret = AVERROR(EINVAL);
>                  goto end;
>              }
> -
> +            ff_thread_await_progress(s->connection, progress2 += 1, 0);
>              low    = s->plane[plane].subband[0];
>              high   = s->plane[plane].subband[7];
>              output = s->plane[plane].subband[0];
> @@ -1199,6 +1229,7 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
>                  low    += lowpass_width;
>                  high   += lowpass_width;
>              }
> +            ff_thread_report_progress(s->picture, progress1 += 1, 0);
>              if (s->progressive) {
>                  low    = s->plane[plane].subband[0];
>                  high   = s->plane[plane].subband[15];
> @@ -1220,37 +1251,17 @@ static int cfhd_decode(AVCodecContext *avctx, void
> *data, int *got_frame,
>                      output++;
>                  }
>
> -                low    = s->plane[plane].subband[7];
> -                high   = s->plane[plane].subband[12];
> -                output = s->plane[plane].l_h[8];
> -                for (i = 0; i < lowpass_width; i++) {
> -                    vert_filter(output, lowpass_width, low,
> lowpass_width, high, highpass_stride, lowpass_height);
> -                    low++;
> -                    high++;
> -                    output++;
> -                }
> -
> -                low    = s->plane[plane].subband[11];
> -                high   = s->plane[plane].subband[13];
> -                output = s->plane[plane].l_h[9];
> -                for (i = 0; i < lowpass_width; i++) {
> -                    vert_filter(output, lowpass_width, low,
> highpass_stride, high, highpass_stride, lowpass_height);
> -                    low++;
> -                    high++;
> -                    output++;
> -                }
> -
> -                dst = (int16_t *)pic->data[act_plane];
> +                dst = (int16_t *)s->picture->f->data[act_plane];
>                  low  = s->plane[plane].l_h[6];
>                  high = s->plane[plane].l_h[7];
>                  for (i = 0; i < lowpass_height * 2; i++) {
> -                    horiz_filter(dst, low, high, lowpass_width);
> +                    horiz_filter_clip(dst, low, high, lowpass_width,
> s->bpc);
>                      low  += lowpass_width;
>                      high += lowpass_width;
> -                    dst  += pic->linesize[act_plane] / 2;
> +                    dst  += s->picture->f->linesize[act_plane] / 2;
>                  }
>              } else {
> -                pic->interlaced_frame = 1;
> +                s->picture->f->interlaced_frame = 1;
>                  low    = s->plane[plane].subband[0];
>                  high   = s->plane[plane].subband[14];
>                  output = s->plane[plane].l_h[6];
> @@ -1271,67 +1282,137 @@ static int cfhd_decode(AVCodecContext *avctx,
> void *data, int *got_frame,
>                      output += lowpass_width * 2;
>                  }
>
> -                low    = s->plane[plane].subband[7];
> -                high   = s->plane[plane].subband[11];
> -                output = s->plane[plane].l_h[8];
> -                for (i = 0; i < lowpass_height; i++) {
> -                    horiz_filter(output, low, high, lowpass_width);
> -                    low    += lowpass_width;
> -                    high   += lowpass_width;
> -                    output += lowpass_width * 2;
> -                }
> -
> -                low    = s->plane[plane].subband[12];
> -                high   = s->plane[plane].subband[13];
> -                output = s->plane[plane].l_h[9];
> -                for (i = 0; i < lowpass_height; i++) {
> -                    horiz_filter(output, low, high, lowpass_width);
> -                    low    += lowpass_width;
> -                    high   += lowpass_width;
> -                    output += lowpass_width * 2;
> -                }
> -
> -
> -                dst  = (int16_t *)pic->data[act_plane];
> +                dst  = (int16_t *)s->picture->f->data[act_plane];
>                  low  = s->plane[plane].l_h[6];
>                  high = s->plane[plane].l_h[7];
>                  for (i = 0; i < lowpass_height; i++) {
> -                    inverse_temporal_filter(dst, low, high, lowpass_width
> * 2,  pic->linesize[act_plane]/2, 0);
> +                    inverse_temporal_filter(dst, low, high, lowpass_width
> * 2,  s->picture->f->linesize[act_plane]/2, 0);
>                      low  += lowpass_width * 2;
>                      high += lowpass_width * 2;
> -                    dst  += pic->linesize[act_plane];
> +                    dst  += s->picture->f->linesize[act_plane];
>                  }
>              }
>          }
> +        ff_thread_report_progress(s->picture, INT_MAX, 0);
> +        ff_thread_await_progress(s->connection, INT_MAX, 0);
> +        av_frame_ref(frame.f, s->picture->f);
>      } else if (s->sample_type == 1) {
> -        int16_t *low, *high, *dst;
> -        int lowpass_height, lowpass_width;
> +        int16_t *low, *high, *dst, *output;
> +        int lowpass_height, lowpass_width, highpass_stride, act_plane;
> +        progress1 = 1, progress2 = 1;
>          for (plane = 0; plane < planes && !ret; plane++) {
> -            int act_plane = plane == 1 ? 2 : plane == 2 ? 1 : plane;
> -            lowpass_height  = s->plane[plane].band[4][1].height;
> -            lowpass_width   = s->plane[plane].band[4][1].width;
> +            ff_thread_await_progress(s->connection, progress1 += 1, 0);
> +            // highpass inverse for temporal
> +            lowpass_height  = s->buffers[plane].band[1][1].a_height;
> +            lowpass_width   = s->buffers[plane].band[1][1].a_width;
> +            highpass_stride = s->buffers[plane].band[1][1].a_width;
> +
> +            low    = s->buffers[plane].subband[7];
> +            high   = s->buffers[plane].subband[9];
> +            output = s->buffers[plane].l_h[8];
> +            for (i = 0; i < lowpass_width; i++) {
> +                vert_filter(output, lowpass_width, low, lowpass_width,
> high, highpass_stride, lowpass_height);
> +                low++;
> +                high++;
> +                output++;
> +            }
> +
> +            low    = s->buffers[plane].subband[8];
> +            high   = s->buffers[plane].subband[10];
> +            output = s->buffers[plane].l_h[9];
> +            for (i = 0; i < lowpass_width; i++) {
> +                vert_filter(output, lowpass_width, low, highpass_stride,
> high, highpass_stride, lowpass_height);
> +                low++;
> +                high++;
> +                output++;
> +            }
> +
> +            low    = s->buffers[plane].l_h[8];
> +            high   = s->buffers[plane].l_h[9];
> +            output = s->buffers[plane].subband[7];
> +            for (i = 0; i < lowpass_height * 2; i++) {
> +                horiz_filter(output, low, high, lowpass_width);
> +                low    += lowpass_width;
> +                high   += lowpass_width;
> +                output += lowpass_width * 2;
> +            }
> +            ff_thread_report_progress(s->picture, progress2 += 1, 0);
> +        }
> +        for (plane = 0; plane < planes && !ret; plane++) {
> +            ff_thread_await_progress(s->connection, progress1 += 1, 0);
> +
> +            act_plane = plane == 1 ? 2 : plane == 2 ? 1 : plane;
> +            lowpass_height  = s->buffers[plane].band[4][1].a_height;
> +            lowpass_width   = s->buffers[plane].band[4][1].a_width;
> +            highpass_stride = s->buffers[plane].band[4][1].a_width;
> +
>              if (s->progressive) {
> -                dst = (int16_t *)pic->data[act_plane];
> -                low  = s->plane[plane].l_h[8];
> -                high = s->plane[plane].l_h[9];
> +                low    = s->buffers[plane].subband[7];
> +                high   = s->buffers[plane].subband[12];
> +                output = s->buffers[plane].l_h[10];
> +                for (i = 0; i < lowpass_width; i++) {
> +                    vert_filter(output, lowpass_width, low,
> lowpass_width, high, highpass_stride, lowpass_height);
> +                    low++;
> +                    high++;
> +                    output++;
> +                }
> +
> +                low    = s->buffers[plane].subband[11];
> +                high   = s->buffers[plane].subband[13];
> +                output = s->buffers[plane].l_h[11];
> +                for (i = 0; i < lowpass_width; i++) {
> +                    vert_filter(output, lowpass_width, low,
> highpass_stride, high, highpass_stride, lowpass_height);
> +                    low++;
> +                    high++;
> +                    output++;
> +                }
> +
> +                dst = (int16_t *)s->picture->f->data[act_plane];
> +                low  = s->buffers[plane].l_h[10];
> +                high = s->buffers[plane].l_h[11];
>                  for (i = 0; i < lowpass_height * 2; i++) {
> -                    horiz_filter(dst, low, high, lowpass_width);
> +                    horiz_filter_clip(dst, low, high, lowpass_width,
> s->bpc);
>                      low  += lowpass_width;
>                      high += lowpass_width;
> -                    dst  += pic->linesize[act_plane] / 2;
> +                    dst  += s->picture->f->linesize[act_plane] / 2;
>                  }
>              } else {
> -                dst  = (int16_t *)pic->data[act_plane];
> -                low  = s->plane[plane].l_h[8];
> -                high = s->plane[plane].l_h[9];
> +                av_log(avctx, AV_LOG_DEBUG, "interlaced frame ? %d",
> s->picture->f->interlaced_frame);
> +                s->picture->f->interlaced_frame = 1;
> +                low    = s->buffers[plane].subband[7];
> +                high   = s->buffers[plane].subband[11];
> +                output = s->buffers[plane].l_h[10];
> +                for (i = 0; i < lowpass_height; i++) {
> +                    horiz_filter(output, low, high, lowpass_width);
> +                    low    += lowpass_width;
> +                    high   += lowpass_width;
> +                    output += lowpass_width * 2;
> +                }
> +
> +                low    = s->buffers[plane].subband[12];
> +                high   = s->buffers[plane].subband[13];
> +                output = s->buffers[plane].l_h[11];
>                  for (i = 0; i < lowpass_height; i++) {
> -                    inverse_temporal_filter(dst, low, high, lowpass_width
> * 2,  pic->linesize[act_plane]/2, 0);
> +                    horiz_filter(output, low, high, lowpass_width);
> +                    low    += lowpass_width;
> +                    high   += lowpass_width;
> +                    output += lowpass_width * 2;
> +                }
> +
> +                dst  = (int16_t *)s->picture->f->data[act_plane];
> +                low  = s->buffers[plane].l_h[10];
> +                high = s->buffers[plane].l_h[11];
> +                for (i = 0; i < lowpass_height; i++) {
> +                    inverse_temporal_filter(dst, low, high, lowpass_width
> * 2,  s->picture->f->linesize[act_plane]/2, 0);
>                      low  += lowpass_width * 2;
>                      high += lowpass_width * 2;
> -                    dst  += pic->linesize[act_plane];
> +                    dst  += s->picture->f->linesize[act_plane];
>                  }
>              }
>          }
> +        ff_thread_report_progress(s->picture, INT_MAX, 0);
> +        ff_thread_await_progress(s->connection, INT_MAX, 0);
> +        av_frame_ref(frame.f, s->picture->f);
>      }
>
>  end:
> @@ -1352,19 +1433,30 @@ static av_cold int cfhd_close(AVCodecContext
> *avctx)
>          ff_free_vlc(&s->vlc_9);
>          ff_free_vlc(&s->vlc_18);
>      }
> +    if (s->i_frame.f && s->i_frame.f->data[0])
> +        ff_thread_release_buffer(avctx, &s->i_frame);
> +    if (s->p_frame.f && s->p_frame.f->data[0])
> +        ff_thread_release_buffer(avctx, &s->p_frame);
> +
> +    if (s->i_frame.f)
> +        av_frame_free(&s->i_frame.f);
> +    if (s->p_frame.f)
> +        av_frame_free(&s->p_frame.f);
>
>      return 0;
>  }
>
>  AVCodec ff_cfhd_decoder = {
> -    .name             = "cfhd",
> -    .long_name        = NULL_IF_CONFIG_SMALL("Cineform HD"),
> -    .type             = AVMEDIA_TYPE_VIDEO,
> -    .id               = AV_CODEC_ID_CFHD,
> -    .priv_data_size   = sizeof(CFHDContext),
> -    .init             = cfhd_init,
> -    .close            = cfhd_close,
> -    .decode           = cfhd_decode,
> -    .capabilities     = AV_CODEC_CAP_DR1,
> -    .caps_internal    = FF_CODEC_CAP_INIT_CLEANUP,
> +    .name                  = "cfhd",
> +    .long_name             = NULL_IF_CONFIG_SMALL("Cineform HD"),
> +    .type                  = AVMEDIA_TYPE_VIDEO,
> +    .id                    = AV_CODEC_ID_CFHD,
> +    .priv_data_size        = sizeof(CFHDContext),
> +    .init                  = cfhd_init,
> +    .close                 = cfhd_close,
> +    .decode                = cfhd_decode,
> +    .init_thread_copy      = ONLY_IF_THREADS_ENABLED(cfhd_init),
> +    .update_thread_context =
> ONLY_IF_THREADS_ENABLED(update_thread_context),
> +    .capabilities          = AV_CODEC_CAP_DR1 |
> AV_CODEC_CAP_FRAME_THREADS,
> +    .caps_internal         = FF_CODEC_CAP_INIT_THREADSAFE |
> FF_CODEC_CAP_INIT_CLEANUP,
>  };
> diff --git a/libavcodec/cfhd.h b/libavcodec/cfhd.h
> index 047c0f2028..d7a2ffe0a7 100644
> --- a/libavcodec/cfhd.h
> +++ b/libavcodec/cfhd.h
> @@ -29,6 +29,7 @@
>  #include "bytestream.h"
>  #include "get_bits.h"
>  #include "vlc.h"
> +#include "thread.h"
>
>  #define VLC_BITS       9
>  #define SUBBAND_COUNT 17
> @@ -63,7 +64,7 @@ typedef struct Plane {
>
>      /* TODO: merge this into SubBand structure */
>      int16_t *subband[SUBBAND_COUNT];
> -    int16_t *l_h[10];
> +    int16_t *l_h[12];
>
>      SubBand band[DWT_LEVELS][4];
>  } Plane;
> @@ -76,6 +77,10 @@ typedef struct Peak {
>
>  typedef struct CFHDContext {
>      AVCodecContext *avctx;
> +    ThreadFrame i_frame;
> +    ThreadFrame p_frame;
> +    ThreadFrame *connection;
> +    ThreadFrame *picture;
>
>      CFHD_RL_VLC_ELEM table_9_rl_vlc[2088];
>      VLC vlc_9;
> @@ -116,6 +121,7 @@ typedef struct CFHDContext {
>
>      uint8_t prescale_shift[3];
>      Plane plane[4];
> +    Plane *buffers;
>      Peak peak;
>  } CFHDContext;
>
> --
> 2.17.1
>
>
Discard this thread please. Really sorry for the inconvenience.

Gagandeep Singh


More information about the ffmpeg-devel mailing list