FFmpeg
 All Data Structures Namespaces Files Functions Variables Typedefs Enumerations Enumerator Macros Groups Pages
hwcontext_qsv.c
Go to the documentation of this file.
1 /*
2  * This file is part of FFmpeg.
3  *
4  * FFmpeg is free software; you can redistribute it and/or
5  * modify it under the terms of the GNU Lesser General Public
6  * License as published by the Free Software Foundation; either
7  * version 2.1 of the License, or (at your option) any later version.
8  *
9  * FFmpeg is distributed in the hope that it will be useful,
10  * but WITHOUT ANY WARRANTY; without even the implied warranty of
11  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
12  * Lesser General Public License for more details.
13  *
14  * You should have received a copy of the GNU Lesser General Public
15  * License along with FFmpeg; if not, write to the Free Software
16  * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
17  */
18 
19 #include <stdint.h>
20 #include <string.h>
21 
22 #include <mfx/mfxvideo.h>
23 
24 #include "config.h"
25 
26 #if CONFIG_VAAPI
27 #include "hwcontext_vaapi.h"
28 #endif
29 #if CONFIG_DXVA2
30 #include "hwcontext_dxva2.h"
31 #endif
32 
33 #include "buffer.h"
34 #include "common.h"
35 #include "hwcontext.h"
36 #include "hwcontext_internal.h"
37 #include "hwcontext_qsv.h"
38 #include "mem.h"
39 #include "pixfmt.h"
40 #include "pixdesc.h"
41 #include "time.h"
42 
43 typedef struct QSVDevicePriv {
46 
47 typedef struct QSVDeviceContext {
48  mfxHDL handle;
49  mfxHandleType handle_type;
50  mfxVersion ver;
51  mfxIMPL impl;
52 
56 
57 typedef struct QSVFramesContext {
58  mfxSession session_download;
59  mfxSession session_upload;
60 
62  mfxFrameSurface1 *surfaces_internal;
64 
65  // used in the frame allocator for non-opaque surfaces
66  mfxMemId *mem_ids;
67  // used in the opaque alloc request for opaque surfaces
68  mfxFrameSurface1 **surface_ptrs;
69 
70  mfxExtOpaqueSurfaceAlloc opaque_alloc;
71  mfxExtBuffer *ext_buffers[1];
73 
74 static const struct {
75  mfxHandleType handle_type;
79 #if CONFIG_VAAPI
80  { MFX_HANDLE_VA_DISPLAY, AV_HWDEVICE_TYPE_VAAPI, AV_PIX_FMT_VAAPI },
81 #endif
82 #if CONFIG_DXVA2
83  { MFX_HANDLE_D3D9_DEVICE_MANAGER, AV_HWDEVICE_TYPE_DXVA2, AV_PIX_FMT_DXVA2_VLD },
84 #endif
85  { 0 },
86 };
87 
88 static const struct {
90  uint32_t fourcc;
92  { AV_PIX_FMT_NV12, MFX_FOURCC_NV12 },
93  { AV_PIX_FMT_P010, MFX_FOURCC_P010 },
94  { AV_PIX_FMT_PAL8, MFX_FOURCC_P8 },
95 };
96 
98 {
99  AVQSVDeviceContext *hwctx = ctx->hwctx;
100  QSVDeviceContext *s = ctx->internal->priv;
101 
102  mfxStatus err;
103  int i;
104 
105  for (i = 0; supported_handle_types[i].handle_type; i++) {
106  err = MFXVideoCORE_GetHandle(hwctx->session, supported_handle_types[i].handle_type,
107  &s->handle);
108  if (err == MFX_ERR_NONE) {
109  s->handle_type = supported_handle_types[i].handle_type;
110  s->child_device_type = supported_handle_types[i].device_type;
111  s->child_pix_fmt = supported_handle_types[i].pix_fmt;
112  break;
113  }
114  }
115  if (!s->handle) {
116  av_log(ctx, AV_LOG_VERBOSE, "No supported hw handle could be retrieved "
117  "from the session\n");
118  }
119 
120  err = MFXQueryIMPL(hwctx->session, &s->impl);
121  if (err == MFX_ERR_NONE)
122  err = MFXQueryVersion(hwctx->session, &s->ver);
123  if (err != MFX_ERR_NONE) {
124  av_log(ctx, AV_LOG_ERROR, "Error querying the session attributes\n");
125  return AVERROR_UNKNOWN;
126  }
127 
128  return 0;
129 }
130 
132 {
133  QSVFramesContext *s = ctx->internal->priv;
134 
135  if (s->session_download) {
136  MFXVideoVPP_Close(s->session_download);
137  MFXClose(s->session_download);
138  }
139  s->session_download = NULL;
140 
141  if (s->session_upload) {
142  MFXVideoVPP_Close(s->session_upload);
143  MFXClose(s->session_upload);
144  }
145  s->session_upload = NULL;
146 
147  av_freep(&s->mem_ids);
148  av_freep(&s->surface_ptrs);
151 }
152 
153 static void qsv_pool_release_dummy(void *opaque, uint8_t *data)
154 {
155 }
156 
157 static AVBufferRef *qsv_pool_alloc(void *opaque, int size)
158 {
160  QSVFramesContext *s = ctx->internal->priv;
161  AVQSVFramesContext *hwctx = ctx->hwctx;
162 
163  if (s->nb_surfaces_used < hwctx->nb_surfaces) {
164  s->nb_surfaces_used++;
166  sizeof(*hwctx->surfaces), qsv_pool_release_dummy, NULL, 0);
167  }
168 
169  return NULL;
170 }
171 
173 {
174  AVQSVFramesContext *hwctx = ctx->hwctx;
175  QSVFramesContext *s = ctx->internal->priv;
176  QSVDeviceContext *device_priv = ctx->device_ctx->internal->priv;
177 
178  AVBufferRef *child_device_ref = NULL;
179  AVBufferRef *child_frames_ref = NULL;
180 
181  AVHWDeviceContext *child_device_ctx;
182  AVHWFramesContext *child_frames_ctx;
183 
184  int i, ret = 0;
185 
186  if (!device_priv->handle) {
187  av_log(ctx, AV_LOG_ERROR,
188  "Cannot create a non-opaque internal surface pool without "
189  "a hardware handle\n");
190  return AVERROR(EINVAL);
191  }
192 
193  child_device_ref = av_hwdevice_ctx_alloc(device_priv->child_device_type);
194  if (!child_device_ref)
195  return AVERROR(ENOMEM);
196  child_device_ctx = (AVHWDeviceContext*)child_device_ref->data;
197 
198 #if CONFIG_VAAPI
199  if (child_device_ctx->type == AV_HWDEVICE_TYPE_VAAPI) {
200  AVVAAPIDeviceContext *child_device_hwctx = child_device_ctx->hwctx;
201  child_device_hwctx->display = (VADisplay)device_priv->handle;
202  }
203 #endif
204 #if CONFIG_DXVA2
205  if (child_device_ctx->type == AV_HWDEVICE_TYPE_DXVA2) {
206  AVDXVA2DeviceContext *child_device_hwctx = child_device_ctx->hwctx;
207  child_device_hwctx->devmgr = (IDirect3DDeviceManager9*)device_priv->handle;
208  }
209 #endif
210 
211  ret = av_hwdevice_ctx_init(child_device_ref);
212  if (ret < 0) {
213  av_log(ctx, AV_LOG_ERROR, "Error initializing a child device context\n");
214  goto fail;
215  }
216 
217  child_frames_ref = av_hwframe_ctx_alloc(child_device_ref);
218  if (!child_frames_ref) {
219  ret = AVERROR(ENOMEM);
220  goto fail;
221  }
222  child_frames_ctx = (AVHWFramesContext*)child_frames_ref->data;
223 
224  child_frames_ctx->format = device_priv->child_pix_fmt;
225  child_frames_ctx->sw_format = ctx->sw_format;
226  child_frames_ctx->initial_pool_size = ctx->initial_pool_size;
227  child_frames_ctx->width = ctx->width;
228  child_frames_ctx->height = ctx->height;
229 
230 #if CONFIG_DXVA2
231  if (child_device_ctx->type == AV_HWDEVICE_TYPE_DXVA2) {
232  AVDXVA2FramesContext *child_frames_hwctx = child_frames_ctx->hwctx;
233  if (hwctx->frame_type & MFX_MEMTYPE_VIDEO_MEMORY_PROCESSOR_TARGET)
234  child_frames_hwctx->surface_type = DXVA2_VideoProcessorRenderTarget;
235  else
236  child_frames_hwctx->surface_type = DXVA2_VideoDecoderRenderTarget;
237  }
238 #endif
239 
240  ret = av_hwframe_ctx_init(child_frames_ref);
241  if (ret < 0) {
242  av_log(ctx, AV_LOG_ERROR, "Error initializing a child frames context\n");
243  goto fail;
244  }
245 
246 #if CONFIG_VAAPI
247  if (child_device_ctx->type == AV_HWDEVICE_TYPE_VAAPI) {
248  AVVAAPIFramesContext *child_frames_hwctx = child_frames_ctx->hwctx;
249  for (i = 0; i < ctx->initial_pool_size; i++)
250  s->surfaces_internal[i].Data.MemId = child_frames_hwctx->surface_ids + i;
251  hwctx->frame_type = MFX_MEMTYPE_VIDEO_MEMORY_DECODER_TARGET;
252  }
253 #endif
254 #if CONFIG_DXVA2
255  if (child_device_ctx->type == AV_HWDEVICE_TYPE_DXVA2) {
256  AVDXVA2FramesContext *child_frames_hwctx = child_frames_ctx->hwctx;
257  for (i = 0; i < ctx->initial_pool_size; i++)
258  s->surfaces_internal[i].Data.MemId = (mfxMemId)child_frames_hwctx->surfaces[i];
259  if (child_frames_hwctx->surface_type == DXVA2_VideoProcessorRenderTarget)
260  hwctx->frame_type = MFX_MEMTYPE_VIDEO_MEMORY_PROCESSOR_TARGET;
261  else
262  hwctx->frame_type = MFX_MEMTYPE_VIDEO_MEMORY_DECODER_TARGET;
263  }
264 #endif
265 
266  s->child_frames_ref = child_frames_ref;
267  child_frames_ref = NULL;
268 
269 fail:
270  av_buffer_unref(&child_device_ref);
271  av_buffer_unref(&child_frames_ref);
272  return ret;
273 }
274 
276 {
277  QSVFramesContext *s = ctx->internal->priv;
278  AVQSVFramesContext *frames_hwctx = ctx->hwctx;
279  const AVPixFmtDescriptor *desc;
280 
281  int i, ret = 0;
282 
283  desc = av_pix_fmt_desc_get(ctx->sw_format);
284  if (!desc)
285  return AVERROR_BUG;
286 
287  if (ctx->initial_pool_size <= 0) {
288  av_log(ctx, AV_LOG_ERROR, "QSV requires a fixed frame pool size\n");
289  return AVERROR(EINVAL);
290  }
291 
293  sizeof(*s->surfaces_internal));
294  if (!s->surfaces_internal)
295  return AVERROR(ENOMEM);
296 
297  for (i = 0; i < ctx->initial_pool_size; i++) {
298  mfxFrameSurface1 *surf = &s->surfaces_internal[i];
299 
300  surf->Info.BitDepthLuma = desc->comp[0].depth;
301  surf->Info.BitDepthChroma = desc->comp[0].depth;
302  surf->Info.Shift = desc->comp[0].depth > 8;
303 
304  if (desc->log2_chroma_w && desc->log2_chroma_h)
305  surf->Info.ChromaFormat = MFX_CHROMAFORMAT_YUV420;
306  else if (desc->log2_chroma_w)
307  surf->Info.ChromaFormat = MFX_CHROMAFORMAT_YUV422;
308  else
309  surf->Info.ChromaFormat = MFX_CHROMAFORMAT_YUV444;
310 
311  surf->Info.FourCC = fourcc;
312  surf->Info.Width = ctx->width;
313  surf->Info.CropW = ctx->width;
314  surf->Info.Height = ctx->height;
315  surf->Info.CropH = ctx->height;
316  surf->Info.FrameRateExtN = 25;
317  surf->Info.FrameRateExtD = 1;
318  }
319 
320  if (!(frames_hwctx->frame_type & MFX_MEMTYPE_OPAQUE_FRAME)) {
321  ret = qsv_init_child_ctx(ctx);
322  if (ret < 0)
323  return ret;
324  }
325 
326  ctx->internal->pool_internal = av_buffer_pool_init2(sizeof(mfxFrameSurface1),
327  ctx, qsv_pool_alloc, NULL);
328  if (!ctx->internal->pool_internal)
329  return AVERROR(ENOMEM);
330 
331  frames_hwctx->surfaces = s->surfaces_internal;
332  frames_hwctx->nb_surfaces = ctx->initial_pool_size;
333 
334  return 0;
335 }
336 
337 static mfxStatus frame_alloc(mfxHDL pthis, mfxFrameAllocRequest *req,
338  mfxFrameAllocResponse *resp)
339 {
340  AVHWFramesContext *ctx = pthis;
341  QSVFramesContext *s = ctx->internal->priv;
342  AVQSVFramesContext *hwctx = ctx->hwctx;
343  mfxFrameInfo *i = &req->Info;
344  mfxFrameInfo *i1 = &hwctx->surfaces[0].Info;
345 
346  if (!(req->Type & MFX_MEMTYPE_VIDEO_MEMORY_PROCESSOR_TARGET) ||
347  !(req->Type & (MFX_MEMTYPE_FROM_VPPIN | MFX_MEMTYPE_FROM_VPPOUT)) ||
348  !(req->Type & MFX_MEMTYPE_EXTERNAL_FRAME))
349  return MFX_ERR_UNSUPPORTED;
350  if (i->Width != i1->Width || i->Height != i1->Height ||
351  i->FourCC != i1->FourCC || i->ChromaFormat != i1->ChromaFormat) {
352  av_log(ctx, AV_LOG_ERROR, "Mismatching surface properties in an "
353  "allocation request: %dx%d %d %d vs %dx%d %d %d\n",
354  i->Width, i->Height, i->FourCC, i->ChromaFormat,
355  i1->Width, i1->Height, i1->FourCC, i1->ChromaFormat);
356  return MFX_ERR_UNSUPPORTED;
357  }
358 
359  resp->mids = s->mem_ids;
360  resp->NumFrameActual = hwctx->nb_surfaces;
361 
362  return MFX_ERR_NONE;
363 }
364 
365 static mfxStatus frame_free(mfxHDL pthis, mfxFrameAllocResponse *resp)
366 {
367  return MFX_ERR_NONE;
368 }
369 
370 static mfxStatus frame_lock(mfxHDL pthis, mfxMemId mid, mfxFrameData *ptr)
371 {
372  return MFX_ERR_UNSUPPORTED;
373 }
374 
375 static mfxStatus frame_unlock(mfxHDL pthis, mfxMemId mid, mfxFrameData *ptr)
376 {
377  return MFX_ERR_UNSUPPORTED;
378 }
379 
380 static mfxStatus frame_get_hdl(mfxHDL pthis, mfxMemId mid, mfxHDL *hdl)
381 {
382  *hdl = mid;
383  return MFX_ERR_NONE;
384 }
385 
387  mfxSession *session, int upload)
388 {
389  QSVFramesContext *s = ctx->internal->priv;
390  AVQSVFramesContext *frames_hwctx = ctx->hwctx;
391  QSVDeviceContext *device_priv = ctx->device_ctx->internal->priv;
392  int opaque = !!(frames_hwctx->frame_type & MFX_MEMTYPE_OPAQUE_FRAME);
393 
394  mfxFrameAllocator frame_allocator = {
395  .pthis = ctx,
396  .Alloc = frame_alloc,
397  .Lock = frame_lock,
398  .Unlock = frame_unlock,
399  .GetHDL = frame_get_hdl,
400  .Free = frame_free,
401  };
402 
403  mfxVideoParam par;
404  mfxStatus err;
405 
406  err = MFXInit(device_priv->impl, &device_priv->ver, session);
407  if (err != MFX_ERR_NONE) {
408  av_log(ctx, AV_LOG_ERROR, "Error initializing an internal session\n");
409  return AVERROR_UNKNOWN;
410  }
411 
412  if (device_priv->handle) {
413  err = MFXVideoCORE_SetHandle(*session, device_priv->handle_type,
414  device_priv->handle);
415  if (err != MFX_ERR_NONE)
416  return AVERROR_UNKNOWN;
417  }
418 
419  if (!opaque) {
420  err = MFXVideoCORE_SetFrameAllocator(*session, &frame_allocator);
421  if (err != MFX_ERR_NONE)
422  return AVERROR_UNKNOWN;
423  }
424 
425  memset(&par, 0, sizeof(par));
426 
427  if (opaque) {
428  par.ExtParam = s->ext_buffers;
429  par.NumExtParam = FF_ARRAY_ELEMS(s->ext_buffers);
430  par.IOPattern = upload ? MFX_IOPATTERN_OUT_OPAQUE_MEMORY :
431  MFX_IOPATTERN_IN_OPAQUE_MEMORY;
432  } else {
433  par.IOPattern = upload ? MFX_IOPATTERN_OUT_VIDEO_MEMORY :
434  MFX_IOPATTERN_IN_VIDEO_MEMORY;
435  }
436 
437  par.IOPattern |= upload ? MFX_IOPATTERN_IN_SYSTEM_MEMORY :
438  MFX_IOPATTERN_OUT_SYSTEM_MEMORY;
439  par.AsyncDepth = 1;
440 
441  par.vpp.In = frames_hwctx->surfaces[0].Info;
442 
443  /* Apparently VPP requires the frame rate to be set to some value, otherwise
444  * init will fail (probably for the framerate conversion filter). Since we
445  * are only doing data upload/download here, we just invent an arbitrary
446  * value */
447  par.vpp.In.FrameRateExtN = 25;
448  par.vpp.In.FrameRateExtD = 1;
449  par.vpp.Out = par.vpp.In;
450 
451  err = MFXVideoVPP_Init(*session, &par);
452  if (err != MFX_ERR_NONE) {
453  av_log(ctx, AV_LOG_VERBOSE, "Error opening the internal VPP session."
454  "Surface upload/download will not be possible\n");
455  MFXClose(*session);
456  *session = NULL;
457  }
458 
459  return 0;
460 }
461 
463 {
464  QSVFramesContext *s = ctx->internal->priv;
465  AVQSVFramesContext *frames_hwctx = ctx->hwctx;
466 
467  int opaque = !!(frames_hwctx->frame_type & MFX_MEMTYPE_OPAQUE_FRAME);
468 
469  uint32_t fourcc = 0;
470  int i, ret;
471 
472  for (i = 0; i < FF_ARRAY_ELEMS(supported_pixel_formats); i++) {
473  if (supported_pixel_formats[i].pix_fmt == ctx->sw_format) {
474  fourcc = supported_pixel_formats[i].fourcc;
475  break;
476  }
477  }
478  if (!fourcc) {
479  av_log(ctx, AV_LOG_ERROR, "Unsupported pixel format\n");
480  return AVERROR(ENOSYS);
481  }
482 
483  if (!ctx->pool) {
484  ret = qsv_init_pool(ctx, fourcc);
485  if (ret < 0) {
486  av_log(ctx, AV_LOG_ERROR, "Error creating an internal frame pool\n");
487  return ret;
488  }
489  }
490 
491  if (opaque) {
492  s->surface_ptrs = av_mallocz_array(frames_hwctx->nb_surfaces,
493  sizeof(*s->surface_ptrs));
494  if (!s->surface_ptrs)
495  return AVERROR(ENOMEM);
496 
497  for (i = 0; i < frames_hwctx->nb_surfaces; i++)
498  s->surface_ptrs[i] = frames_hwctx->surfaces + i;
499 
500  s->opaque_alloc.In.Surfaces = s->surface_ptrs;
501  s->opaque_alloc.In.NumSurface = frames_hwctx->nb_surfaces;
502  s->opaque_alloc.In.Type = frames_hwctx->frame_type;
503 
504  s->opaque_alloc.Out = s->opaque_alloc.In;
505 
506  s->opaque_alloc.Header.BufferId = MFX_EXTBUFF_OPAQUE_SURFACE_ALLOCATION;
507  s->opaque_alloc.Header.BufferSz = sizeof(s->opaque_alloc);
508 
509  s->ext_buffers[0] = (mfxExtBuffer*)&s->opaque_alloc;
510  } else {
511  s->mem_ids = av_mallocz_array(frames_hwctx->nb_surfaces, sizeof(*s->mem_ids));
512  if (!s->mem_ids)
513  return AVERROR(ENOMEM);
514 
515  for (i = 0; i < frames_hwctx->nb_surfaces; i++)
516  s->mem_ids[i] = frames_hwctx->surfaces[i].Data.MemId;
517  }
518 
519  ret = qsv_init_internal_session(ctx, &s->session_download, 0);
520  if (ret < 0)
521  return ret;
522 
523  ret = qsv_init_internal_session(ctx, &s->session_upload, 1);
524  if (ret < 0)
525  return ret;
526 
527  return 0;
528 }
529 
531 {
532  frame->buf[0] = av_buffer_pool_get(ctx->pool);
533  if (!frame->buf[0])
534  return AVERROR(ENOMEM);
535 
536  frame->data[3] = frame->buf[0]->data;
537  frame->format = AV_PIX_FMT_QSV;
538  frame->width = ctx->width;
539  frame->height = ctx->height;
540 
541  return 0;
542 }
543 
546  enum AVPixelFormat **formats)
547 {
548  enum AVPixelFormat *fmts;
549 
550  fmts = av_malloc_array(2, sizeof(*fmts));
551  if (!fmts)
552  return AVERROR(ENOMEM);
553 
554  fmts[0] = ctx->sw_format;
555  fmts[1] = AV_PIX_FMT_NONE;
556 
557  *formats = fmts;
558 
559  return 0;
560 }
561 
563  AVFrame *dst, const AVFrame *src, int flags)
564 {
565  QSVFramesContext *s = ctx->internal->priv;
566  mfxFrameSurface1 *surf = (mfxFrameSurface1*)src->data[3];
567  AVHWFramesContext *child_frames_ctx;
568 
569  AVFrame *dummy;
570  int ret = 0;
571 
572  if (!s->child_frames_ref)
573  return AVERROR(ENOSYS);
574  child_frames_ctx = (AVHWFramesContext*)s->child_frames_ref->data;
575 
576  dummy = av_frame_alloc();
577  if (!dummy)
578  return AVERROR(ENOMEM);
579 
580  dummy->buf[0] = av_buffer_ref(src->buf[0]);
581  dummy->hw_frames_ctx = av_buffer_ref(s->child_frames_ref);
582  if (!dummy->buf[0] || !dummy->hw_frames_ctx)
583  goto fail;
584 
585  dummy->format = child_frames_ctx->format;
586  dummy->width = src->width;
587  dummy->height = src->height;
588  dummy->data[3] = surf->Data.MemId;
589 
590  ret = av_hwframe_map(dst, dummy, flags);
591 
592 fail:
594 
595  return ret;
596 }
597 
599  const AVFrame *src)
600 {
601  QSVFramesContext *s = ctx->internal->priv;
602  AVHWFramesContext *child_frames_ctx = (AVHWFramesContext*)s->child_frames_ref->data;
603  int download = !!src->hw_frames_ctx;
604  mfxFrameSurface1 *surf = (mfxFrameSurface1*)(download ? src->data[3] : dst->data[3]);
605 
606  AVFrame *dummy;
607  int ret;
608 
609  dummy = av_frame_alloc();
610  if (!dummy)
611  return AVERROR(ENOMEM);
612 
613  dummy->format = child_frames_ctx->format;
614  dummy->width = src->width;
615  dummy->height = src->height;
616  dummy->buf[0] = download ? src->buf[0] : dst->buf[0];
617  dummy->data[3] = surf->Data.MemId;
618  dummy->hw_frames_ctx = s->child_frames_ref;
619 
620  ret = download ? av_hwframe_transfer_data(dst, dummy, 0) :
621  av_hwframe_transfer_data(dummy, src, 0);
622 
623  dummy->buf[0] = NULL;
624  dummy->data[3] = NULL;
625  dummy->hw_frames_ctx = NULL;
626 
627  av_frame_free(&dummy);
628 
629  return ret;
630 }
631 
633  const AVFrame *src)
634 {
635  QSVFramesContext *s = ctx->internal->priv;
636  mfxFrameSurface1 out = {{ 0 }};
637  mfxFrameSurface1 *in = (mfxFrameSurface1*)src->data[3];
638 
639  mfxSyncPoint sync = NULL;
640  mfxStatus err;
641 
642  if (!s->session_download) {
643  if (s->child_frames_ref)
644  return qsv_transfer_data_child(ctx, dst, src);
645 
646  av_log(ctx, AV_LOG_ERROR, "Surface download not possible\n");
647  return AVERROR(ENOSYS);
648  }
649 
650  out.Info = in->Info;
651  out.Data.PitchLow = dst->linesize[0];
652  out.Data.Y = dst->data[0];
653  out.Data.U = dst->data[1];
654  out.Data.V = dst->data[2];
655  out.Data.A = dst->data[3];
656 
657  do {
658  err = MFXVideoVPP_RunFrameVPPAsync(s->session_download, in, &out, NULL, &sync);
659  if (err == MFX_WRN_DEVICE_BUSY)
660  av_usleep(1);
661  } while (err == MFX_WRN_DEVICE_BUSY);
662 
663  if (err < 0 || !sync) {
664  av_log(ctx, AV_LOG_ERROR, "Error downloading the surface\n");
665  return AVERROR_UNKNOWN;
666  }
667 
668  do {
669  err = MFXVideoCORE_SyncOperation(s->session_download, sync, 1000);
670  } while (err == MFX_WRN_IN_EXECUTION);
671  if (err < 0) {
672  av_log(ctx, AV_LOG_ERROR, "Error synchronizing the operation: %d\n", err);
673  return AVERROR_UNKNOWN;
674  }
675 
676  return 0;
677 }
678 
680  const AVFrame *src)
681 {
682  QSVFramesContext *s = ctx->internal->priv;
683  mfxFrameSurface1 in = {{ 0 }};
684  mfxFrameSurface1 *out = (mfxFrameSurface1*)dst->data[3];
685 
686  mfxSyncPoint sync = NULL;
687  mfxStatus err;
688 
689  if (!s->session_upload) {
690  if (s->child_frames_ref)
691  return qsv_transfer_data_child(ctx, dst, src);
692 
693  av_log(ctx, AV_LOG_ERROR, "Surface upload not possible\n");
694  return AVERROR(ENOSYS);
695  }
696 
697  in.Info = out->Info;
698  in.Data.PitchLow = src->linesize[0];
699  in.Data.Y = src->data[0];
700  in.Data.U = src->data[1];
701  in.Data.V = src->data[2];
702  in.Data.A = src->data[3];
703 
704  do {
705  err = MFXVideoVPP_RunFrameVPPAsync(s->session_upload, &in, out, NULL, &sync);
706  if (err == MFX_WRN_DEVICE_BUSY)
707  av_usleep(1);
708  } while (err == MFX_WRN_DEVICE_BUSY);
709 
710  if (err < 0 || !sync) {
711  av_log(ctx, AV_LOG_ERROR, "Error uploading the surface\n");
712  return AVERROR_UNKNOWN;
713  }
714 
715  do {
716  err = MFXVideoCORE_SyncOperation(s->session_upload, sync, 1000);
717  } while (err == MFX_WRN_IN_EXECUTION);
718  if (err < 0) {
719  av_log(ctx, AV_LOG_ERROR, "Error synchronizing the operation\n");
720  return AVERROR_UNKNOWN;
721  }
722 
723  return 0;
724 }
725 
727  const void *hwconfig,
728  AVHWFramesConstraints *constraints)
729 {
730  int i;
731 
733  sizeof(*constraints->valid_sw_formats));
734  if (!constraints->valid_sw_formats)
735  return AVERROR(ENOMEM);
736 
737  for (i = 0; i < FF_ARRAY_ELEMS(supported_pixel_formats); i++)
738  constraints->valid_sw_formats[i] = supported_pixel_formats[i].pix_fmt;
739  constraints->valid_sw_formats[FF_ARRAY_ELEMS(supported_pixel_formats)] = AV_PIX_FMT_NONE;
740 
741  constraints->valid_hw_formats = av_malloc_array(2, sizeof(*constraints->valid_hw_formats));
742  if (!constraints->valid_hw_formats)
743  return AVERROR(ENOMEM);
744 
745  constraints->valid_hw_formats[0] = AV_PIX_FMT_QSV;
746  constraints->valid_hw_formats[1] = AV_PIX_FMT_NONE;
747 
748  return 0;
749 }
750 
752 {
753  AVQSVDeviceContext *hwctx = ctx->hwctx;
754  QSVDevicePriv *priv = ctx->user_opaque;
755 
756  if (hwctx->session)
757  MFXClose(hwctx->session);
758 
760  av_freep(&priv);
761 }
762 
763 static mfxIMPL choose_implementation(const char *device)
764 {
765  static const struct {
766  const char *name;
767  mfxIMPL impl;
768  } impl_map[] = {
769  { "auto", MFX_IMPL_AUTO },
770  { "sw", MFX_IMPL_SOFTWARE },
771  { "hw", MFX_IMPL_HARDWARE },
772  { "auto_any", MFX_IMPL_AUTO_ANY },
773  { "hw_any", MFX_IMPL_HARDWARE_ANY },
774  { "hw2", MFX_IMPL_HARDWARE2 },
775  { "hw3", MFX_IMPL_HARDWARE3 },
776  { "hw4", MFX_IMPL_HARDWARE4 },
777  };
778 
779  mfxIMPL impl = MFX_IMPL_AUTO_ANY;
780  int i;
781 
782  if (device) {
783  for (i = 0; i < FF_ARRAY_ELEMS(impl_map); i++)
784  if (!strcmp(device, impl_map[i].name)) {
785  impl = impl_map[i].impl;
786  break;
787  }
788  if (i == FF_ARRAY_ELEMS(impl_map))
789  impl = strtol(device, NULL, 0);
790  }
791 
792  return impl;
793 }
794 
795 static int qsv_device_create(AVHWDeviceContext *ctx, const char *device,
796  AVDictionary *opts, int flags)
797 {
798  AVQSVDeviceContext *hwctx = ctx->hwctx;
799  QSVDevicePriv *priv;
800  enum AVHWDeviceType child_device_type;
802 
803  mfxVersion ver = { { 3, 1 } };
804  mfxIMPL impl;
805  mfxHDL handle;
806  mfxHandleType handle_type;
807  mfxStatus err;
808  int ret;
809 
810  priv = av_mallocz(sizeof(*priv));
811  if (!priv)
812  return AVERROR(ENOMEM);
813 
814  ctx->user_opaque = priv;
815  ctx->free = qsv_device_free;
816 
817  e = av_dict_get(opts, "child_device", NULL, 0);
818 
819  if (CONFIG_VAAPI)
820  child_device_type = AV_HWDEVICE_TYPE_VAAPI;
821  else if (CONFIG_DXVA2)
822  child_device_type = AV_HWDEVICE_TYPE_DXVA2;
823  else {
824  av_log(ctx, AV_LOG_ERROR, "No supported child device type is enabled\n");
825  return AVERROR(ENOSYS);
826  }
827 
828  ret = av_hwdevice_ctx_create(&priv->child_device_ctx, child_device_type,
829  e ? e->value : NULL, NULL, 0);
830  if (ret < 0)
831  return ret;
832 
833  {
834  AVHWDeviceContext *child_device_ctx = (AVHWDeviceContext*)priv->child_device_ctx->data;
835 #if CONFIG_VAAPI
836  AVVAAPIDeviceContext *child_device_hwctx = child_device_ctx->hwctx;
837  handle_type = MFX_HANDLE_VA_DISPLAY;
838  handle = (mfxHDL)child_device_hwctx->display;
839 #elif CONFIG_DXVA2
840  AVDXVA2DeviceContext *child_device_hwctx = child_device_ctx->hwctx;
841  handle_type = MFX_HANDLE_D3D9_DEVICE_MANAGER;
842  handle = (mfxHDL)child_device_hwctx->devmgr;
843 #endif
844  }
845 
846  impl = choose_implementation(device);
847 
848  err = MFXInit(impl, &ver, &hwctx->session);
849  if (err != MFX_ERR_NONE) {
850  av_log(ctx, AV_LOG_ERROR, "Error initializing an MFX session\n");
851  return AVERROR_UNKNOWN;
852  }
853 
854  err = MFXVideoCORE_SetHandle(hwctx->session, handle_type, handle);
855  if (err != MFX_ERR_NONE)
856  return AVERROR_UNKNOWN;
857 
858  return 0;
859 }
860 
863  .name = "QSV",
864 
865  .device_hwctx_size = sizeof(AVQSVDeviceContext),
866  .device_priv_size = sizeof(QSVDeviceContext),
867  .frames_hwctx_size = sizeof(AVQSVFramesContext),
868  .frames_priv_size = sizeof(QSVFramesContext),
869 
870  .device_create = qsv_device_create,
871  .device_init = qsv_device_init,
872  .frames_get_constraints = qsv_frames_get_constraints,
873  .frames_init = qsv_frames_init,
874  .frames_uninit = qsv_frames_uninit,
875  .frames_get_buffer = qsv_get_buffer,
876  .transfer_get_formats = qsv_transfer_get_formats,
877  .transfer_data_to = qsv_transfer_data_to,
878  .transfer_data_from = qsv_transfer_data_from,
879  .map_from = qsv_map_from,
880 
881  .pix_fmts = (const enum AVPixelFormat[]){ AV_PIX_FMT_QSV, AV_PIX_FMT_NONE },
882 };
This struct aggregates all the (hardware/vendor-specific) "high-level" state, i.e.
Definition: hwcontext.h:54
#define NULL
Definition: coverity.c:32
const char * s
Definition: avisynth_c.h:768
uint32_t fourcc
Definition: hwcontext_qsv.c:90
static int qsv_init_child_ctx(AVHWFramesContext *ctx)
void av_buffer_unref(AVBufferRef **buf)
Free a given reference and automatically free the buffer if there are no more references to it...
Definition: buffer.c:125
const AVPixFmtDescriptor * av_pix_fmt_desc_get(enum AVPixelFormat pix_fmt)
Definition: pixdesc.c:2333
VAAPI-specific data associated with a frame pool.
This structure describes decoded (raw) audio or video data.
Definition: frame.h:187
ptrdiff_t const GLvoid * data
Definition: opengl_enc.c:101
static int qsv_frames_get_constraints(AVHWDeviceContext *ctx, const void *hwconfig, AVHWFramesConstraints *constraints)
mfxFrameSurface1 * surfaces_internal
Definition: hwcontext_qsv.c:62
int frame_type
A combination of MFX_MEMTYPE_* describing the frame pool.
Definition: hwcontext_qsv.h:49
mfxExtBuffer * ext_buffers[1]
Definition: hwcontext_qsv.c:71
This struct is allocated as AVHWFramesContext.hwctx.
Memory handling functions.
mfxHandleType handle_type
Definition: hwcontext_qsv.c:75
AVBufferRef * buf[AV_NUM_DATA_POINTERS]
AVBuffer references backing the data for this frame.
Definition: frame.h:370
This struct is allocated as AVHWFramesContext.hwctx.
Definition: hwcontext_qsv.h:42
const char * desc
Definition: nvenc.c:60
An API-specific header for AV_HWDEVICE_TYPE_DXVA2.
int width
The allocated dimensions of the frames in this pool.
Definition: hwcontext.h:222
mfxMemId * mem_ids
Definition: hwcontext_qsv.c:66
enum AVPixelFormat format
The pixel format identifying the underlying HW surface type.
Definition: hwcontext.h:202
mfxHandleType handle_type
Definition: hwcontext_qsv.c:49
void * av_mallocz(size_t size)
Allocate a memory block with alignment suitable for all memory accesses (including vectors if availab...
Definition: mem.c:222
int av_usleep(unsigned usec)
Sleep for a period of time.
Definition: time.c:84
mfxVersion ver
Definition: hwcontext_qsv.c:50
#define src
Definition: vp8dsp.c:254
static mfxStatus frame_alloc(mfxHDL pthis, mfxFrameAllocRequest *req, mfxFrameAllocResponse *resp)
uint8_t log2_chroma_w
Amount to shift the luma width right to find the chroma width.
Definition: pixdesc.h:92
static int qsv_init_pool(AVHWFramesContext *ctx, uint32_t fourcc)
API-specific header for AV_HWDEVICE_TYPE_VAAPI.
AVBufferRef * hw_frames_ctx
For hwaccel-format frames, this should be a reference to the AVHWFramesContext describing the frame...
Definition: frame.h:515
#define AV_PIX_FMT_P010
Definition: pixfmt.h:394
static int qsv_transfer_data_child(AVHWFramesContext *ctx, AVFrame *dst, const AVFrame *src)
AVBufferPool * pool_internal
static int qsv_transfer_data_to(AVHWFramesContext *ctx, AVFrame *dst, const AVFrame *src)
enum AVHWDeviceType type
AVComponentDescriptor comp[4]
Parameters that describe how pixels are packed.
Definition: pixdesc.h:117
uint8_t
AVFrame * av_frame_alloc(void)
Allocate an AVFrame and set its fields to default values.
Definition: frame.c:150
8 bits with AV_PIX_FMT_RGB32 palette
Definition: pixfmt.h:73
DWORD surface_type
The surface type (e.g.
enum AVHWDeviceType child_device_type
Definition: hwcontext_qsv.c:53
static int qsv_map_from(AVHWFramesContext *ctx, AVFrame *dst, const AVFrame *src, int flags)
static AVFrame * frame
void * hwctx
The format-specific data, allocated and freed by libavutil along with this context.
Definition: hwcontext.h:85
AVDictionaryEntry * av_dict_get(const AVDictionary *m, const char *key, const AVDictionaryEntry *prev, int flags)
Get a dictionary entry with matching key.
Definition: dict.c:40
int av_hwdevice_ctx_create(AVBufferRef **pdevice_ref, enum AVHWDeviceType type, const char *device, AVDictionary *opts, int flags)
Open a device of the specified type and create an AVHWDeviceContext for it.
Definition: hwcontext.c:502
static int flags
Definition: log.c:57
#define AV_LOG_VERBOSE
Detailed information.
Definition: log.h:192
static mfxStatus frame_unlock(mfxHDL pthis, mfxMemId mid, mfxFrameData *ptr)
ptrdiff_t size
Definition: opengl_enc.c:101
static int qsv_frames_init(AVHWFramesContext *ctx)
#define av_log(a,...)
This struct is allocated as AVHWDeviceContext.hwctx.
Definition: hwcontext_qsv.h:35
static int qsv_get_buffer(AVHWFramesContext *ctx, AVFrame *frame)
int width
width and height of the video frame
Definition: frame.h:239
#define AV_LOG_ERROR
Something went wrong and cannot losslessly be recovered.
Definition: log.h:176
uint8_t log2_chroma_h
Amount to shift the luma height right to find the chroma height.
Definition: pixdesc.h:101
void(* free)(struct AVHWDeviceContext *ctx)
This field may be set by the caller before calling av_hwdevice_ctx_init().
Definition: hwcontext.h:97
#define AVERROR(e)
Definition: error.h:43
static int qsv_transfer_data_from(AVHWFramesContext *ctx, AVFrame *dst, const AVFrame *src)
void av_frame_free(AVFrame **frame)
Free the frame and any dynamically allocated objects in it, e.g.
Definition: frame.c:163
planar YUV 4:2:0, 12bpp, 1 plane for Y and 1 plane for the UV components, which are interleaved (firs...
Definition: pixfmt.h:90
enum AVHWDeviceType device_type
Definition: hwcontext_qsv.c:76
static mfxStatus frame_get_hdl(mfxHDL pthis, mfxMemId mid, mfxHDL *hdl)
IDirect3DDeviceManager9 * devmgr
AVBufferRef * av_buffer_create(uint8_t *data, int size, void(*free)(void *opaque, uint8_t *data), void *opaque, int flags)
Create an AVBuffer from an existing array.
Definition: buffer.c:28
int av_hwframe_ctx_init(AVBufferRef *ref)
Finalize the context before use.
Definition: hwcontext.c:268
enum AVHWDeviceType type
This field identifies the underlying API used for hardware access.
Definition: hwcontext.h:72
static void * av_mallocz_array(size_t nmemb, size_t size)
Definition: mem.h:229
#define fail()
Definition: checkasm.h:89
static int qsv_transfer_get_formats(AVHWFramesContext *ctx, enum AVHWFrameTransferDirection dir, enum AVPixelFormat **formats)
int av_hwframe_transfer_data(AVFrame *dst, const AVFrame *src, int flags)
Copy data to or from a hw surface.
Definition: hwcontext.c:378
int initial_pool_size
Initial size of the frame pool.
Definition: hwcontext.h:192
AVDictionary * opts
Definition: movenc.c:50
static mfxStatus frame_free(mfxHDL pthis, mfxFrameAllocResponse *resp)
AVHWDeviceContext * device_ctx
The parent AVHWDeviceContext.
Definition: hwcontext.h:142
AVFormatContext * ctx
Definition: movenc.c:48
mfxFrameSurface1 ** surface_ptrs
Definition: hwcontext_qsv.c:68
int dummy
Definition: motion.c:64
mfxSession session_download
Definition: hwcontext_qsv.c:58
AVBufferPool * av_buffer_pool_init2(int size, void *opaque, AVBufferRef *(*alloc)(void *opaque, int size), void(*pool_free)(void *opaque))
Allocate and initialize a buffer pool with a more complex allocator.
Definition: buffer.c:218
#define FF_ARRAY_ELEMS(a)
VADisplay display
The VADisplay handle, to be filled by the user.
mfxSession session
Definition: hwcontext_qsv.h:36
static void qsv_frames_uninit(AVHWFramesContext *ctx)
int format
format of the frame, -1 if unknown or unset Values correspond to enum AVPixelFormat for video frames...
Definition: frame.h:251
AVBufferRef * av_hwdevice_ctx_alloc(enum AVHWDeviceType type)
Allocate an AVHWDeviceContext for a given hardware type.
Definition: hwcontext.c:74
static mfxStatus frame_lock(mfxHDL pthis, mfxMemId mid, mfxFrameData *ptr)
const HWContextType ff_hwcontext_type_qsv
This struct describes the constraints on hardware frames attached to a given device with a hardware-s...
Definition: hwcontext.h:373
int linesize[AV_NUM_DATA_POINTERS]
For video, size in bytes of each picture line.
Definition: frame.h:218
mfxFrameSurface1 * surfaces
Definition: hwcontext_qsv.h:43
Descriptor that unambiguously describes how the bits of a pixel are stored in the up to 4 data planes...
Definition: pixdesc.h:81
uint8_t * data
The data buffer.
Definition: buffer.h:89
void * hwctx
The format-specific data, allocated and freed automatically along with this context.
Definition: hwcontext.h:155
uint8_t pi<< 24) CONV_FUNC_GROUP(AV_SAMPLE_FMT_FLT, float, AV_SAMPLE_FMT_U8, uint8_t,(*(constuint8_t *) pi-0x80)*(1.0f/(1<< 7))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_DBL, double, AV_SAMPLE_FMT_U8, uint8_t,(*(constuint8_t *) pi-0x80)*(1.0/(1<< 7))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_U8, uint8_t, AV_SAMPLE_FMT_S16, int16_t,(*(constint16_t *) pi >>8)+0x80) CONV_FUNC_GROUP(AV_SAMPLE_FMT_FLT, float, AV_SAMPLE_FMT_S16, int16_t,*(constint16_t *) pi *(1.0f/(1<< 15))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_DBL, double, AV_SAMPLE_FMT_S16, int16_t,*(constint16_t *) pi *(1.0/(1<< 15))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_U8, uint8_t, AV_SAMPLE_FMT_S32, int32_t,(*(constint32_t *) pi >>24)+0x80) CONV_FUNC_GROUP(AV_SAMPLE_FMT_FLT, float, AV_SAMPLE_FMT_S32, int32_t,*(constint32_t *) pi *(1.0f/(1U<< 31))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_DBL, double, AV_SAMPLE_FMT_S32, int32_t,*(constint32_t *) pi *(1.0/(1U<< 31))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_U8, uint8_t, AV_SAMPLE_FMT_FLT, float, av_clip_uint8(lrintf(*(constfloat *) pi *(1<< 7))+0x80)) CONV_FUNC_GROUP(AV_SAMPLE_FMT_S16, int16_t, AV_SAMPLE_FMT_FLT, float, av_clip_int16(lrintf(*(constfloat *) pi *(1<< 15)))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_S32, int32_t, AV_SAMPLE_FMT_FLT, float, av_clipl_int32(llrintf(*(constfloat *) pi *(1U<< 31)))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_U8, uint8_t, AV_SAMPLE_FMT_DBL, double, av_clip_uint8(lrint(*(constdouble *) pi *(1<< 7))+0x80)) CONV_FUNC_GROUP(AV_SAMPLE_FMT_S16, int16_t, AV_SAMPLE_FMT_DBL, double, av_clip_int16(lrint(*(constdouble *) pi *(1<< 15)))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_S32, int32_t, AV_SAMPLE_FMT_DBL, double, av_clipl_int32(llrint(*(constdouble *) pi *(1U<< 31))))#defineSET_CONV_FUNC_GROUP(ofmt, ifmt) staticvoidset_generic_function(AudioConvert *ac){}voidff_audio_convert_free(AudioConvert **ac){if(!*ac) return;ff_dither_free(&(*ac) ->dc);av_freep(ac);}AudioConvert *ff_audio_convert_alloc(AVAudioResampleContext *avr, enumAVSampleFormatout_fmt, enumAVSampleFormatin_fmt, intchannels, intsample_rate, intapply_map){AudioConvert *ac;intin_planar, out_planar;ac=av_mallocz(sizeof(*ac));if(!ac) returnNULL;ac->avr=avr;ac->out_fmt=out_fmt;ac->in_fmt=in_fmt;ac->channels=channels;ac->apply_map=apply_map;if(avr->dither_method!=AV_RESAMPLE_DITHER_NONE &&av_get_packed_sample_fmt(out_fmt)==AV_SAMPLE_FMT_S16 &&av_get_bytes_per_sample(in_fmt)>2){ac->dc=ff_dither_alloc(avr, out_fmt, in_fmt, channels, sample_rate, apply_map);if(!ac->dc){av_free(ac);returnNULL;}returnac;}in_planar=ff_sample_fmt_is_planar(in_fmt, channels);out_planar=ff_sample_fmt_is_planar(out_fmt, channels);if(in_planar==out_planar){ac->func_type=CONV_FUNC_TYPE_FLAT;ac->planes=in_planar?ac->channels:1;}elseif(in_planar) ac->func_type=CONV_FUNC_TYPE_INTERLEAVE;elseac->func_type=CONV_FUNC_TYPE_DEINTERLEAVE;set_generic_function(ac);if(ARCH_AARCH64) ff_audio_convert_init_aarch64(ac);if(ARCH_ARM) ff_audio_convert_init_arm(ac);if(ARCH_X86) ff_audio_convert_init_x86(ac);returnac;}intff_audio_convert(AudioConvert *ac, AudioData *out, AudioData *in){intuse_generic=1;intlen=in->nb_samples;intp;if(ac->dc){av_log(ac->avr, AV_LOG_TRACE,"%dsamples-audio_convert:%sto%s(dithered)\n", len, av_get_sample_fmt_name(ac->in_fmt), av_get_sample_fmt_name(ac->out_fmt));returnff_convert_dither(ac-> in
#define AVERROR_BUG
Internal bug, also see AVERROR_BUG2.
Definition: error.h:50
int av_hwdevice_ctx_init(AVBufferRef *ref)
Finalize the device context before use.
Definition: hwcontext.c:132
static int qsv_device_create(AVHWDeviceContext *ctx, const char *device, AVDictionary *opts, int flags)
static const struct @240 supported_pixel_formats[]
static int qsv_init_internal_session(AVHWFramesContext *ctx, mfxSession *session, int upload)
This struct describes a set or pool of "hardware" frames (i.e.
Definition: hwcontext.h:117
HW acceleration through QSV, data[3] contains a pointer to the mfxFrameSurface1 structure.
Definition: pixfmt.h:236
static mfxIMPL choose_implementation(const char *device)
refcounted data buffer API
enum AVPixelFormat * valid_hw_formats
A list of possible values for format in the hw_frames_ctx, terminated by AV_PIX_FMT_NONE.
Definition: hwcontext.h:378
mfxExtOpaqueSurfaceAlloc opaque_alloc
Definition: hwcontext_qsv.c:70
HW decoding through DXVA2, Picture.data[3] contains a LPDIRECT3DSURFACE9 pointer. ...
Definition: pixfmt.h:148
AVHWFramesInternal * internal
Private data used internally by libavutil.
Definition: hwcontext.h:127
static AVBufferRef * qsv_pool_alloc(void *opaque, int size)
uint8_t * data[AV_NUM_DATA_POINTERS]
pointer to the picture/channel planes.
Definition: frame.h:201
void * user_opaque
Arbitrary user data, to be used e.g.
Definition: hwcontext.h:102
A reference to a data buffer.
Definition: buffer.h:81
static void qsv_device_free(AVHWDeviceContext *ctx)
common internal and external API header
if(ret< 0)
Definition: vf_mcdeint.c:282
int av_hwframe_map(AVFrame *dst, const AVFrame *src, int flags)
Map a hardware frame.
Definition: hwcontext.c:605
AVBufferRef * av_hwframe_ctx_alloc(AVBufferRef *device_ref_in)
Allocate an AVHWFramesContext tied to a given device context.
Definition: hwcontext.c:182
AVBufferRef * av_buffer_ref(AVBufferRef *buf)
Create a new reference to an AVBuffer.
Definition: buffer.c:93
static int qsv_device_init(AVHWDeviceContext *ctx)
Definition: hwcontext_qsv.c:97
#define AVERROR_UNKNOWN
Unknown error, typically from an external library.
Definition: error.h:71
enum AVPixelFormat child_pix_fmt
Definition: hwcontext_qsv.c:54
IDirect3DSurface9 ** surfaces
The surface pool.
AVHWFrameTransferDirection
Definition: hwcontext.h:335
pixel format definitions
AVBufferPool * pool
A pool from which the frames are allocated by av_hwframe_get_buffer().
Definition: hwcontext.h:183
AVHWDeviceType
Definition: hwcontext.h:27
This struct is allocated as AVHWDeviceContext.hwctx.
char * value
Definition: dict.h:87
enum AVPixelFormat * valid_sw_formats
A list of possible values for sw_format in the hw_frames_ctx, terminated by AV_PIX_FMT_NONE.
Definition: hwcontext.h:385
VAAPI connection details.
int height
Definition: frame.h:239
FILE * out
Definition: movenc.c:54
#define av_freep(p)
VASurfaceID * surface_ids
The surfaces IDs of all surfaces in the pool after creation.
An API-specific header for AV_HWDEVICE_TYPE_QSV.
AVBufferRef * child_device_ctx
Definition: hwcontext_qsv.c:44
AVBufferRef * child_frames_ref
Definition: hwcontext_qsv.c:61
AVBufferRef * av_buffer_pool_get(AVBufferPool *pool)
Allocate a new AVBuffer, reusing an old buffer from the pool when available.
Definition: buffer.c:334
#define av_malloc_array(a, b)
formats
Definition: signature.h:48
AVHWDeviceInternal * internal
Private data used internally by libavutil.
Definition: hwcontext.h:64
int depth
Number of bits in the component.
Definition: pixdesc.h:58
static void qsv_pool_release_dummy(void *opaque, uint8_t *data)
enum AVPixelFormat sw_format
The pixel format identifying the actual data layout of the hardware frames.
Definition: hwcontext.h:215
AVPixelFormat
Pixel format.
Definition: pixfmt.h:60
enum AVPixelFormat pix_fmt
Definition: hwcontext_qsv.c:77
mfxSession session_upload
Definition: hwcontext_qsv.c:59
const char * name
Definition: opengl_enc.c:103
static const struct @239 supported_handle_types[]