[FFmpeg-cvslog] Add a QSV decoding example.

Anton Khirnov git at videolan.org
Sun Mar 1 00:46:16 CET 2015


ffmpeg | branch: master | Anton Khirnov <anton at khirnov.net> | Thu Feb 12 00:07:29 2015 +0100| [990e4a6639d0714198583c2812b1074b5751898f] | committer: Anton Khirnov

Add a QSV decoding example.

> http://git.videolan.org/gitweb.cgi/ffmpeg.git/?a=commit;h=990e4a6639d0714198583c2812b1074b5751898f
---

 configure             |    7 +
 doc/Makefile          |    1 +
 doc/examples/qsvdec.c |  484 +++++++++++++++++++++++++++++++++++++++++++++++++
 3 files changed, 492 insertions(+)

diff --git a/configure b/configure
index 4998851..0e182b4 100755
--- a/configure
+++ b/configure
@@ -1137,6 +1137,7 @@ EXAMPLE_LIST="
     filter_audio_example
     metadata_example
     output_example
+    qsvdec_example
     transcode_aac_example
 "
 
@@ -1550,6 +1551,7 @@ HAVE_LIST="
     sdl
     section_data_rel_ro
     threads
+    vaapi_x11
     vdpau_x11
     xlib
 "
@@ -2187,6 +2189,7 @@ avcodec_example_deps="avcodec avutil"
 filter_audio_example_deps="avfilter avutil"
 metadata_example_deps="avformat avutil"
 output_example_deps="avcodec avformat avutil swscale"
+qsvdec_example_deps="avcodec avutil libmfx h264_qsv_decoder vaapi_x11"
 transcode_aac_example_deps="avcodec avformat avresample"
 
 # libraries, in linking order
@@ -4330,6 +4333,10 @@ if enabled x11grab; then
     require Xfixes X11/extensions/Xfixes.h XFixesGetCursorImage -lXfixes
 fi
 
+enabled vaapi && enabled xlib &&
+    check_lib2 "va/va.h va/va_x11.h" vaGetDisplay -lva -lva-x11 &&
+    enable vaapi_x11
+
 enabled vdpau &&
     check_cpp_condition vdpau/vdpau.h "defined VDP_DECODER_PROFILE_MPEG4_PART2_ASP" ||
     disable vdpau
diff --git a/doc/Makefile b/doc/Makefile
index 1c205a5..2f6a5fb 100644
--- a/doc/Makefile
+++ b/doc/Makefile
@@ -19,6 +19,7 @@ DOC_EXAMPLES-$(CONFIG_AVCODEC_EXAMPLE)          += avcodec
 DOC_EXAMPLES-$(CONFIG_FILTER_AUDIO_EXAMPLE)     += filter_audio
 DOC_EXAMPLES-$(CONFIG_METADATA_EXAMPLE)         += metadata
 DOC_EXAMPLES-$(CONFIG_OUTPUT_EXAMPLE)           += output
+DOC_EXAMPLES-$(CONFIG_QSVDEC_EXAMPLE)           += qsvdec
 DOC_EXAMPLES-$(CONFIG_TRANSCODE_AAC_EXAMPLE)    += transcode_aac
 ALL_DOC_EXAMPLES = avcodec filter_audio metadata output transcode_aac
 
diff --git a/doc/examples/qsvdec.c b/doc/examples/qsvdec.c
new file mode 100644
index 0000000..6dbb210
--- /dev/null
+++ b/doc/examples/qsvdec.c
@@ -0,0 +1,484 @@
+/*
+ * Copyright (c) 2015 Anton Khirnov
+ *
+ * Permission is hereby granted, free of charge, to any person obtaining a copy
+ * of this software and associated documentation files (the "Software"), to deal
+ * in the Software without restriction, including without limitation the rights
+ * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
+ * copies of the Software, and to permit persons to whom the Software is
+ * furnished to do so, subject to the following conditions:
+ *
+ * The above copyright notice and this permission notice shall be included in
+ * all copies or substantial portions of the Software.
+ *
+ * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
+ * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
+ * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
+ * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
+ * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
+ * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
+ * THE SOFTWARE.
+ */
+
+/**
+ * @file
+ * Intel QSV-accelerated H.264 decoding example.
+ *
+ * @example qsvdec.c
+ * This example shows how to do QSV-accelerated H.264 decoding with output
+ * frames in the VA-API video surfaces.
+ */
+
+#include "config.h"
+
+#include <stdio.h>
+
+#include <mfx/mfxvideo.h>
+
+#include <va/va.h>
+#include <va/va_x11.h>
+#include <X11/Xlib.h>
+
+#include "libavformat/avformat.h"
+#include "libavformat/avio.h"
+
+#include "libavcodec/avcodec.h"
+#include "libavcodec/qsv.h"
+
+#include "libavutil/error.h"
+#include "libavutil/mem.h"
+
+typedef struct DecodeContext {
+    mfxSession mfx_session;
+    VADisplay va_dpy;
+
+    VASurfaceID *surfaces;
+    mfxMemId    *surface_ids;
+    int         *surface_used;
+    int       nb_surfaces;
+
+    mfxFrameInfo frame_info;
+} DecodeContext;
+
+static mfxStatus frame_alloc(mfxHDL pthis, mfxFrameAllocRequest *req,
+                             mfxFrameAllocResponse *resp)
+{
+    DecodeContext *decode = pthis;
+    int err, i;
+
+    if (decode->surfaces) {
+        fprintf(stderr, "Multiple allocation requests.\n");
+        return MFX_ERR_MEMORY_ALLOC;
+    }
+    if (!(req->Type & MFX_MEMTYPE_VIDEO_MEMORY_DECODER_TARGET)) {
+        fprintf(stderr, "Unsupported surface type: %d\n", req->Type);
+        return MFX_ERR_UNSUPPORTED;
+    }
+    if (req->Info.BitDepthLuma != 8 || req->Info.BitDepthChroma != 8 ||
+        req->Info.Shift || req->Info.FourCC != MFX_FOURCC_NV12 ||
+        req->Info.ChromaFormat != MFX_CHROMAFORMAT_YUV420) {
+        fprintf(stderr, "Unsupported surface properties.\n");
+        return MFX_ERR_UNSUPPORTED;
+    }
+
+    decode->surfaces     = av_malloc_array (req->NumFrameSuggested, sizeof(*decode->surfaces));
+    decode->surface_ids  = av_malloc_array (req->NumFrameSuggested, sizeof(*decode->surface_ids));
+    decode->surface_used = av_mallocz_array(req->NumFrameSuggested, sizeof(*decode->surface_used));
+    if (!decode->surfaces || !decode->surface_ids || !decode->surface_used)
+        goto fail;
+
+    err = vaCreateSurfaces(decode->va_dpy, VA_RT_FORMAT_YUV420,
+                           req->Info.Width, req->Info.Height,
+                           decode->surfaces, req->NumFrameSuggested,
+                           NULL, 0);
+    if (err != VA_STATUS_SUCCESS) {
+        fprintf(stderr, "Error allocating VA surfaces\n");
+        goto fail;
+    }
+    decode->nb_surfaces = req->NumFrameSuggested;
+
+    for (i = 0; i < decode->nb_surfaces; i++)
+        decode->surface_ids[i] = &decode->surfaces[i];
+
+    resp->mids           = decode->surface_ids;
+    resp->NumFrameActual = decode->nb_surfaces;
+
+    decode->frame_info = req->Info;
+
+    return MFX_ERR_NONE;
+fail:
+    av_freep(&decode->surfaces);
+    av_freep(&decode->surface_ids);
+    av_freep(&decode->surface_used);
+
+    return MFX_ERR_MEMORY_ALLOC;
+}
+
+static mfxStatus frame_free(mfxHDL pthis, mfxFrameAllocResponse *resp)
+{
+    DecodeContext *decode = pthis;
+
+    if (decode->surfaces)
+        vaDestroySurfaces(decode->va_dpy, decode->surfaces, decode->nb_surfaces);
+    av_freep(&decode->surfaces);
+    av_freep(&decode->surface_ids);
+    av_freep(&decode->surface_used);
+    decode->nb_surfaces = 0;
+
+    return MFX_ERR_NONE;
+}
+
+static mfxStatus frame_lock(mfxHDL pthis, mfxMemId mid, mfxFrameData *ptr)
+{
+    return MFX_ERR_UNSUPPORTED;
+}
+
+static mfxStatus frame_unlock(mfxHDL pthis, mfxMemId mid, mfxFrameData *ptr)
+{
+    return MFX_ERR_UNSUPPORTED;
+}
+
+static mfxStatus frame_get_hdl(mfxHDL pthis, mfxMemId mid, mfxHDL *hdl)
+{
+    *hdl = mid;
+    return MFX_ERR_NONE;
+}
+
+static void free_buffer(void *opaque, uint8_t *data)
+{
+    int *used = opaque;
+    *used = 0;
+    av_freep(&data);
+}
+
+static int get_buffer(AVCodecContext *avctx, AVFrame *frame, int flags)
+{
+    DecodeContext *decode = avctx->opaque;
+
+    mfxFrameSurface1 *surf;
+    AVBufferRef *surf_buf;
+    int idx;
+
+    for (idx = 0; idx < decode->nb_surfaces; idx++) {
+        if (!decode->surface_used[idx])
+            break;
+    }
+    if (idx == decode->nb_surfaces) {
+        fprintf(stderr, "No free surfaces\n");
+        return AVERROR(ENOMEM);
+    }
+
+    surf = av_mallocz(sizeof(*surf));
+    if (!surf)
+        return AVERROR(ENOMEM);
+    surf_buf = av_buffer_create((uint8_t*)surf, sizeof(*surf), free_buffer,
+                                &decode->surface_used[idx], AV_BUFFER_FLAG_READONLY);
+    if (!surf_buf) {
+        av_freep(&surf);
+        return AVERROR(ENOMEM);
+    }
+
+    surf->Info       = decode->frame_info;
+    surf->Data.MemId = &decode->surfaces[idx];
+
+    frame->buf[0]  = surf_buf;
+    frame->data[3] = (uint8_t*)surf;
+
+    decode->surface_used[idx] = 1;
+
+    return 0;
+}
+
+static int get_format(AVCodecContext *avctx, const enum AVPixelFormat *pix_fmts)
+{
+    while (*pix_fmts != AV_PIX_FMT_NONE) {
+        if (*pix_fmts == AV_PIX_FMT_QSV) {
+            if (!avctx->hwaccel_context) {
+                DecodeContext *decode = avctx->opaque;
+                AVQSVContext *qsv = av_qsv_alloc_context();
+                if (!qsv)
+                    return AV_PIX_FMT_NONE;
+
+                qsv->session   = decode->mfx_session;
+                qsv->iopattern = MFX_IOPATTERN_OUT_VIDEO_MEMORY;
+
+                avctx->hwaccel_context = qsv;
+            }
+
+            return AV_PIX_FMT_QSV;
+        }
+
+        pix_fmts++;
+    }
+
+    fprintf(stderr, "The QSV pixel format not offered in get_format()\n");
+
+    return AV_PIX_FMT_NONE;
+}
+
+static int decode_packet(DecodeContext *decode, AVCodecContext *decoder_ctx,
+                         AVFrame *frame, AVPacket *pkt,
+                         AVIOContext *output_ctx)
+{
+    int ret = 0;
+    int got_frame = 1;
+
+    while (pkt->size > 0 || (!pkt->data && got_frame)) {
+        ret = avcodec_decode_video2(decoder_ctx, frame, &got_frame, pkt);
+        if (ret < 0) {
+            fprintf(stderr, "Error during decoding\n");
+            return ret;
+        }
+
+        pkt->data += ret;
+        pkt->size -= ret;
+
+        /* A real program would do something useful with the decoded frame here.
+         * We just retrieve the raw data and write it to a file, which is rather
+         * useless but pedagogic. */
+        if (got_frame) {
+            mfxFrameSurface1 *surf = (mfxFrameSurface1*)frame->data[3];
+            VASurfaceID    surface = *(VASurfaceID*)surf->Data.MemId;
+
+            VAImageFormat img_fmt = {
+                .fourcc         = VA_FOURCC_NV12,
+                .byte_order     = VA_LSB_FIRST,
+                .bits_per_pixel = 8,
+                .depth          = 8,
+            };
+
+            VAImage img;
+
+            VAStatus err;
+            uint8_t *data;
+            int i, j;
+
+            img.buf      = VA_INVALID_ID;
+            img.image_id = VA_INVALID_ID;
+
+            err = vaCreateImage(decode->va_dpy, &img_fmt,
+                                frame->width, frame->height, &img);
+            if (err != VA_STATUS_SUCCESS) {
+                fprintf(stderr, "Error creating an image: %s\n",
+                        vaErrorStr(err));
+                ret = AVERROR_UNKNOWN;
+                goto fail;
+            }
+
+            err = vaGetImage(decode->va_dpy, surface, 0, 0,
+                             frame->width, frame->height,
+                             img.image_id);
+            if (err != VA_STATUS_SUCCESS) {
+                fprintf(stderr, "Error getting an image: %s\n",
+                        vaErrorStr(err));
+                ret = AVERROR_UNKNOWN;
+                goto fail;
+            }
+
+            err = vaMapBuffer(decode->va_dpy, img.buf, (void**)&data);
+            if (err != VA_STATUS_SUCCESS) {
+                fprintf(stderr, "Error mapping the image buffer: %s\n",
+                        vaErrorStr(err));
+                ret = AVERROR_UNKNOWN;
+                goto fail;
+            }
+
+            for (i = 0; i < img.num_planes; i++)
+                for (j = 0; j < (img.height >> (i > 0)); j++)
+                    avio_write(output_ctx, data + img.offsets[i] + j * img.pitches[i], img.width);
+
+fail:
+            if (img.buf != VA_INVALID_ID)
+                vaUnmapBuffer(decode->va_dpy, img.buf);
+            if (img.image_id != VA_INVALID_ID)
+                vaDestroyImage(decode->va_dpy, img.image_id);
+            av_frame_unref(frame);
+
+            if (ret < 0)
+                return ret;
+        }
+    }
+
+    return 0;
+}
+
+int main(int argc, char **argv)
+{
+    AVFormatContext *input_ctx = NULL;
+    AVStream *video_st = NULL;
+    AVCodecContext *decoder_ctx = NULL;
+    const AVCodec *decoder;
+
+    AVPacket pkt = { 0 };
+    AVFrame *frame = NULL;
+
+    DecodeContext decode = { NULL };
+
+    Display *dpy = NULL;
+    int va_ver_major, va_ver_minor;
+
+    mfxIMPL mfx_impl = MFX_IMPL_AUTO_ANY;
+    mfxVersion mfx_ver = { { 1, 1 } };
+
+    mfxFrameAllocator frame_allocator = {
+        .pthis = &decode,
+        .Alloc = frame_alloc,
+        .Lock  = frame_lock,
+        .Unlock = frame_unlock,
+        .GetHDL = frame_get_hdl,
+        .Free   = frame_free,
+    };
+
+    AVIOContext *output_ctx = NULL;
+
+    int ret, i, err;
+
+    av_register_all();
+
+    if (argc < 3) {
+        fprintf(stderr, "Usage: %s <input file> <output file>\n", argv[0]);
+        return 1;
+    }
+
+    /* open the input file */
+    ret = avformat_open_input(&input_ctx, argv[1], NULL, NULL);
+    if (ret < 0) {
+        fprintf(stderr, "Cannot open input file '%s': ", argv[1]);
+        goto finish;
+    }
+
+    /* find the first H.264 video stream */
+    for (i = 0; i < input_ctx->nb_streams; i++) {
+        AVStream *st = input_ctx->streams[i];
+
+        if (st->codec->codec_id == AV_CODEC_ID_H264 && !video_st)
+            video_st = st;
+        else
+            st->discard = AVDISCARD_ALL;
+    }
+    if (!video_st) {
+        fprintf(stderr, "No H.264 video stream in the input file\n");
+        goto finish;
+    }
+
+    /* initialize VA-API */
+    dpy = XOpenDisplay(NULL);
+    if (!dpy) {
+        fprintf(stderr, "Cannot open the X display\n");
+        goto finish;
+    }
+    decode.va_dpy = vaGetDisplay(dpy);
+    if (!decode.va_dpy) {
+        fprintf(stderr, "Cannot open the VA display\n");
+        goto finish;
+    }
+
+    err = vaInitialize(decode.va_dpy, &va_ver_major, &va_ver_minor);
+    if (err != VA_STATUS_SUCCESS) {
+        fprintf(stderr, "Cannot initialize VA: %s\n", vaErrorStr(err));
+        goto finish;
+    }
+    fprintf(stderr, "Initialized VA v%d.%d\n", va_ver_major, va_ver_minor);
+
+    /* initialize an MFX session */
+    err = MFXInit(mfx_impl, &mfx_ver, &decode.mfx_session);
+    if (err != MFX_ERR_NONE) {
+        fprintf(stderr, "Error initializing an MFX session\n");
+        goto finish;
+    }
+
+    MFXVideoCORE_SetHandle(decode.mfx_session, MFX_HANDLE_VA_DISPLAY, decode.va_dpy);
+    MFXVideoCORE_SetFrameAllocator(decode.mfx_session, &frame_allocator);
+
+    /* initialize the decoder */
+    decoder = avcodec_find_decoder_by_name("h264_qsv");
+    if (!decoder) {
+        fprintf(stderr, "The QSV decoder is not present in libavcodec\n");
+        goto finish;
+    }
+
+    decoder_ctx = avcodec_alloc_context3(decoder);
+    if (!decoder_ctx) {
+        ret = AVERROR(ENOMEM);
+        goto finish;
+    }
+    decoder_ctx->codec_id = AV_CODEC_ID_H264;
+    if (video_st->codec->extradata_size) {
+        decoder_ctx->extradata = av_mallocz(video_st->codec->extradata_size +
+                                            FF_INPUT_BUFFER_PADDING_SIZE);
+        if (!decoder_ctx->extradata) {
+            ret = AVERROR(ENOMEM);
+            goto finish;
+        }
+        memcpy(decoder_ctx->extradata, video_st->codec->extradata,
+               video_st->codec->extradata_size);
+        decoder_ctx->extradata_size = video_st->codec->extradata_size;
+    }
+    decoder_ctx->refcounted_frames = 1;
+
+    decoder_ctx->opaque      = &decode;
+    decoder_ctx->get_buffer2 = get_buffer;
+    decoder_ctx->get_format  = get_format;
+
+    ret = avcodec_open2(decoder_ctx, NULL, NULL);
+    if (ret < 0) {
+        fprintf(stderr, "Error opening the decoder: ");
+        goto finish;
+    }
+
+    /* open the output stream */
+    ret = avio_open(&output_ctx, argv[2], AVIO_FLAG_WRITE);
+    if (ret < 0) {
+        fprintf(stderr, "Error opening the output context: ");
+        goto finish;
+    }
+
+    frame = av_frame_alloc();
+    if (!frame) {
+        ret = AVERROR(ENOMEM);
+        goto finish;
+    }
+
+    /* actual decoding */
+    while (ret >= 0) {
+        ret = av_read_frame(input_ctx, &pkt);
+        if (ret < 0)
+            break;
+
+        if (pkt.stream_index == video_st->index)
+            ret = decode_packet(&decode, decoder_ctx, frame, &pkt, output_ctx);
+
+        av_packet_unref(&pkt);
+    }
+
+    /* flush the decoder */
+    pkt.data = NULL;
+    pkt.size = 0;
+    ret = decode_packet(&decode, decoder_ctx, frame, &pkt, output_ctx);
+
+finish:
+    if (ret < 0) {
+        char buf[1024];
+        av_strerror(ret, buf, sizeof(buf));
+        fprintf(stderr, "%s\n", buf);
+    }
+
+    avformat_close_input(&input_ctx);
+
+    av_frame_free(&frame);
+
+    if (decode.mfx_session)
+        MFXClose(decode.mfx_session);
+    if (decode.va_dpy)
+        vaTerminate(decode.va_dpy);
+    if (dpy)
+        XCloseDisplay(dpy);
+
+    if (decoder_ctx)
+        av_freep(&decoder_ctx->hwaccel_context);
+    avcodec_free_context(&decoder_ctx);
+
+    avio_close(output_ctx);
+
+    return ret;
+}



More information about the ffmpeg-cvslog mailing list