On Sun, 2021-07-18 at 11:57 +0800, Linjie Fu wrote: > Hi Haihao, > > On Wed, Jun 9, 2021 at 3:36 PM Haihao Xiang <haihao.xi...@intel.com> wrote: > > > > Include hstack_qsv, vstack_qsv and xstack_qsv, some code is copy and > > pasted from other filters > > > > Example: > > $> ffmpeg -hwaccel qsv -c:v hevc_qsv -i input.h265 -filter_complex > > "[0:v][0:v]hstack_qsv" -f null - > > --- > > configure | 6 + > > libavfilter/Makefile | 3 + > > libavfilter/allfilters.c | 3 + > > libavfilter/vf_stack_qsv.c | 499 +++++++++++++++++++++++++++++++++++++ > > 4 files changed, 511 insertions(+) > > create mode 100644 libavfilter/vf_stack_qsv.c > > > > diff --git a/configure b/configure > > index 6bfd98b384..89adb3a374 100755 > > --- a/configure > > +++ b/configure > > @@ -3705,6 +3705,12 @@ vpp_qsv_filter_select="qsvvpp" > > xfade_opencl_filter_deps="opencl" > > yadif_cuda_filter_deps="ffnvcodec" > > yadif_cuda_filter_deps_any="cuda_nvcc cuda_llvm" > > +hstack_qsv_filter_deps="libmfx" > > +hstack_qsv_filter_select="qsvvpp" > > +vstack_qsv_filter_deps="libmfx" > > +vstack_qsv_filter_select="qsvvpp" > > +xstack_qsv_filter_deps="libmfx" > > +xstack_qsv_filter_select="qsvvpp" > > > > # examples > > avio_list_dir_deps="avformat avutil" > > diff --git a/libavfilter/Makefile b/libavfilter/Makefile > > index bc81033e3f..16cf2c8712 100644 > > --- a/libavfilter/Makefile > > +++ b/libavfilter/Makefile > > @@ -493,6 +493,9 @@ OBJS-$(CONFIG_YAEPBLUR_FILTER) += > > vf_yaepblur.o > > OBJS-$(CONFIG_ZMQ_FILTER) += f_zmq.o > > OBJS-$(CONFIG_ZOOMPAN_FILTER) += vf_zoompan.o > > OBJS-$(CONFIG_ZSCALE_FILTER) += vf_zscale.o > > +OBJS-$(CONFIG_HSTACK_QSV_FILTER) += vf_stack_qsv.o framesync.o > > +OBJS-$(CONFIG_VSTACK_QSV_FILTER) += vf_stack_qsv.o framesync.o > > +OBJS-$(CONFIG_XSTACK_QSV_FILTER) += vf_stack_qsv.o framesync.o > > > > OBJS-$(CONFIG_ALLRGB_FILTER) += vsrc_testsrc.o > > OBJS-$(CONFIG_ALLYUV_FILTER) += vsrc_testsrc.o > > diff --git a/libavfilter/allfilters.c b/libavfilter/allfilters.c > > index c6afef835f..278ccb99a9 100644 > > --- a/libavfilter/allfilters.c > > +++ b/libavfilter/allfilters.c > > @@ -471,6 +471,9 @@ extern const AVFilter ff_vf_yaepblur; > > extern const AVFilter ff_vf_zmq; > > extern const AVFilter ff_vf_zoompan; > > extern const AVFilter ff_vf_zscale; > > +extern const AVFilter ff_vf_hstack_qsv; > > +extern const AVFilter ff_vf_vstack_qsv; > > +extern const AVFilter ff_vf_xstack_qsv; > > > > extern const AVFilter ff_vsrc_allrgb; > > extern const AVFilter ff_vsrc_allyuv; > > diff --git a/libavfilter/vf_stack_qsv.c b/libavfilter/vf_stack_qsv.c > > new file mode 100644 > > index 0000000000..87f611eece > > --- /dev/null > > +++ b/libavfilter/vf_stack_qsv.c > > @@ -0,0 +1,499 @@ > > +/* > > + * This file is part of FFmpeg. > > + * > > + * FFmpeg is free software; you can redistribute it and/or > > + * modify it under the terms of the GNU Lesser General Public > > + * License as published by the Free Software Foundation; either > > + * version 2.1 of the License, or (at your option) any later version. > > + * > > + * FFmpeg is distributed in the hope that it will be useful, > > + * but WITHOUT ANY WARRANTY; without even the implied warranty of > > + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU > > + * Lesser General Public License for more details. > > + * > > + * You should have received a copy of the GNU Lesser General Public > > + * License along with FFmpeg; if not, write to the Free Software > > + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA > > 02110-1301 > > USA > > + */ > > + > > +/** > > + * @file > > + * Hardware accelerated hstack, vstack and xstack filters based on Intel > > Quick Sync Video VPP > > + */ > > + > > +#include "libavutil/opt.h" > > +#include "libavutil/common.h" > > +#include "libavutil/pixdesc.h" > > +#include "libavutil/eval.h" > > +#include "libavutil/hwcontext.h" > > +#include "libavutil/avstring.h" > > +#include "libavutil/avassert.h" > > +#include "libavutil/imgutils.h" > > +#include "libavutil/mathematics.h" > > +#include "libavutil/parseutils.h" > > + > > +#include "internal.h" > > +#include "avfilter.h" > > "avfilter.h" has been included in lots of *.h like "internal.h" and > "qsvvpp.h", hence seems to be redundant here. > > > +#include "filters.h" > > +#include "formats.h" > > +#include "video.h" > > + > > +#include "framesync.h" > > +#include "qsvvpp.h" > > + > > +#define OFFSET(x) offsetof(QSVStackContext, x) > > +#define FLAGS (AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_FILTERING_PARAM) > > + > > +enum { > > + QSV_STACK_H = 0, > > + QSV_STACK_V = 1, > > + QSV_STACK_X = 2 > > +}; > > + > > +typedef struct QSVStackContext { > > + const AVClass *class; > > + QSVVPPContext *qsv; > > + QSVVPPParam qsv_param; > > + mfxExtVPPComposite comp_conf; > > + int mode; > > + FFFrameSync fs; > > + > > + /* Options */ > > + int nb_inputs; > > + int shortest; > > + double scale; > > + char *layout; > > + uint8_t fillcolor[4]; > > + char *fillcolor_str; > > + int fillcolor_enable; > > +} QSVStackContext; > > + > > +static void rgb2yuv(float r, float g, float b, int *y, int *u, int *v, int > > depth) > > +{ > > + *y = ((0.21260*219.0/255.0) * r + (0.71520*219.0/255.0) * g + > > + (0.07220*219.0/255.0) * b) * ((1 << depth) - 1); > > + *u = (-(0.11457*224.0/255.0) * r - (0.38543*224.0/255.0) * g + > > + (0.50000*224.0/255.0) * b + 0.5) * ((1 << depth) - 1); > > + *v = ((0.50000*224.0/255.0) * r - (0.45415*224.0/255.0) * g - > > + (0.04585*224.0/255.0) * b + 0.5) * ((1 << depth) - 1); > > +} > > + > > +static int process_frame(FFFrameSync *fs) > > +{ > > + AVFilterContext *ctx = fs->parent; > > + QSVStackContext *sctx = fs->opaque; > > + AVFrame *frame = NULL; > > + int ret = 0, i; > > + > > + for (i = 0; i < ctx->nb_inputs; i++) { > > + ret = ff_framesync_get_frame(fs, i, &frame, 0); > > + if (ret == 0) > > + ret = ff_qsvvpp_filter_frame(sctx->qsv, ctx->inputs[i], frame); > > + if (ret < 0 && ret != AVERROR(EAGAIN)) > > + break; > > + } > > + > > + if (ret == 0 && sctx->qsv->got_frame == 0) { > > + for (i = 0; i < ctx->nb_inputs; i++) > > + FF_FILTER_FORWARD_WANTED(ctx->outputs[0], ctx->inputs[i]); > > + > > + ret = FFERROR_NOT_READY; > > + } > > + > > + return ret; > > +} > > + > > +static int init_framesync(AVFilterContext *ctx) > > +{ > > + QSVStackContext *sctx = ctx->priv; > > + int ret, i; > > + > > + ret = ff_framesync_init(&sctx->fs, ctx, ctx->nb_inputs); > > + > > + if (ret < 0) > > + return ret; > > + > > + sctx->fs.on_event = process_frame; > > + sctx->fs.opaque = sctx; > > + > > + for (i = 0; i < ctx->nb_inputs; i++) { > > + FFFrameSyncIn *in = &sctx->fs.in[i]; > > + in->before = EXT_STOP; > > + in->after = sctx->shortest ? EXT_STOP : EXT_INFINITY; > > + in->sync = i ? 1 : 2; > > + in->time_base = ctx->inputs[i]->time_base; > > + } > > + > > + return ff_framesync_configure(&sctx->fs); > > +} > > + > > +#define SET_INPUT_STREAM(is, x, y, w, h) do { \ > > + is->DstX = x; \ > > + is->DstY = y; \ > > + is->DstW = w; \ > > + is->DstH = h; \ > > + is->GlobalAlpha = 255; \ > > + is->GlobalAlphaEnable = 1; \ > > + is->PixelAlphaEnable = 0; \ > > + } while (0) > > + > > +static int config_output(AVFilterLink *outlink) > > +{ > > + AVFilterContext *ctx = outlink->src; > > + QSVStackContext *sctx = ctx->priv; > > + AVFilterLink *inlink0 = ctx->inputs[0]; > > + int width, height, i, ret; > > + enum AVPixelFormat in_format; > > + int depth = 8; > > + > > + if (inlink0->format == AV_PIX_FMT_QSV) { > > + if (!inlink0->hw_frames_ctx || !inlink0->hw_frames_ctx->data) > > + return AVERROR(EINVAL); > > + > > + in_format = ((AVHWFramesContext*)inlink0->hw_frames_ctx->data)- > > >sw_format; > > + } else > > + in_format = inlink0->format; > > + > > + sctx->qsv_param.out_sw_format = in_format; > > + > > + for (i = 1; i < sctx->nb_inputs; i++) { > > + AVFilterLink *inlink = ctx->inputs[i]; > > + > > + if (inlink0->format == AV_PIX_FMT_QSV) { > > + AVHWFramesContext *hwfc0 = (AVHWFramesContext *)inlink0- > > >hw_frames_ctx->data; > > + AVHWFramesContext *hwfc = (AVHWFramesContext *)inlink- > > >hw_frames_ctx->data; > > + > > + if (inlink0->format != inlink->format) { > > + av_log(ctx, AV_LOG_ERROR, "Mixing hardware and software > > pixel formats is not supported.\n"); > > + > > + return AVERROR(EINVAL); > > + } else if (hwfc0->device_ctx != hwfc->device_ctx) { > > + av_log(ctx, AV_LOG_ERROR, "Inputs with different underlying > > QSV devices are forbidden.\n"); > > + > > + return AVERROR(EINVAL); > > + } > > + } > > + } > > + > > + if (in_format == AV_PIX_FMT_P010) > > + depth = 10; > > + > > + if (sctx->fillcolor_enable) { > > + int Y, U, V; > > + > > + rgb2yuv(sctx->fillcolor[0] / 255.0, sctx->fillcolor[1] / 255.0, > > + sctx->fillcolor[2] / 255.0, &Y, &U, &V, depth); > > + sctx->comp_conf.Y = Y; > > + sctx->comp_conf.U = U; > > + sctx->comp_conf.V = V; > > + } > > + > > + if (sctx->mode == QSV_STACK_H) { > > + height = inlink0->h * sctx->scale; > > + width = 0; > > + > > + for (i = 0; i < sctx->nb_inputs; i++) { > > + AVFilterLink *inlink = ctx->inputs[i]; > > + mfxVPPCompInputStream *is = &sctx->comp_conf.InputStream[i]; > > + > > + if (inlink0->h != inlink->h) { > > + av_log(ctx, AV_LOG_ERROR, "Input %d height %d does not > > match input %d height %d.\n", i, inlink->h, 0, inlink0->h); > > + return AVERROR(EINVAL); > > + } > > + > > + SET_INPUT_STREAM(is, width, 0, inlink->w * sctx->scale, inlink- > > >h * sctx->scale); > > + width += inlink->w * sctx->scale; > > + } > > + } else if (sctx->mode == QSV_STACK_V) { > > + height = 0; > > + width = inlink0->w * sctx->scale; > > + > > + for (i = 0; i < sctx->nb_inputs; i++) { > > + AVFilterLink *inlink = ctx->inputs[i]; > > + mfxVPPCompInputStream *is = &sctx->comp_conf.InputStream[i]; > > + > > + if (inlink0->w != inlink->w) { > > + av_log(ctx, AV_LOG_ERROR, "Input %d width %d does not match > > input %d width %d.\n", i, inlink->w, 0, inlink0->w); > > + return AVERROR(EINVAL); > > + } > > + > > + SET_INPUT_STREAM(is, 0, height, inlink->w * sctx->scale, > > inlink->h * sctx->scale); > > + height += inlink->h * sctx->scale; > > + } > > + } else { > > + char *arg, *p = sctx->layout, *saveptr = NULL; > > + char *arg2, *p2, *saveptr2 = NULL; > > + char *arg3, *p3, *saveptr3 = NULL; > > + int inw, inh, size, j; > > + > > + width = ctx->inputs[0]->w * sctx->scale; > > + height = ctx->inputs[0]->h * sctx->scale; > > + > > + for (i = 0; i < sctx->nb_inputs; i++) { > > + AVFilterLink *inlink = ctx->inputs[i]; > > + mfxVPPCompInputStream *is = &sctx->comp_conf.InputStream[i]; > > + > > + if (!(arg = av_strtok(p, "|", &saveptr))) > > + return AVERROR(EINVAL); > > + > > + p = NULL; > > + p2 = arg; > > + inw = inh = 0; > > + > > + for (j = 0; j < 2; j++) { > > + if (!(arg2 = av_strtok(p2, "_", &saveptr2))) > > + return AVERROR(EINVAL); > > + > > + p2 = NULL; > > + p3 = arg2; > > + > > + while ((arg3 = av_strtok(p3, "+", &saveptr3))) { > > + p3 = NULL; > > + if (sscanf(arg3, "w%d", &size) == 1) { > > + if (size == i || size < 0 || size >= sctx- > > >nb_inputs) > > + return AVERROR(EINVAL); > > + > > + if (!j) > > + inw += ctx->inputs[size]->w * sctx->scale; > > + else > > + inh += ctx->inputs[size]->w * sctx->scale; > > + } else if (sscanf(arg3, "h%d", &size) == 1) { > > + if (size == i || size < 0 || size >= sctx- > > >nb_inputs) > > + return AVERROR(EINVAL); > > + > > + if (!j) > > + inw += ctx->inputs[size]->h * sctx->scale; > > + else > > + inh += ctx->inputs[size]->h * sctx->scale; > > + } else if (sscanf(arg3, "%d", &size) == 1) { > > + if (size < 0) > > + return AVERROR(EINVAL); > > + > > + if (!j) > > + inw += size; > > + else > > + inh += size; > > + } else { > > + return AVERROR(EINVAL); > > + } > > + } > > + } > > + > > + SET_INPUT_STREAM(is, inw, inh, inlink->w * sctx->scale, inlink- > > >h * sctx->scale); > > + width = FFMAX(width, inlink->w * sctx->scale + inw); > > + height = FFMAX(height, inlink->h * sctx->scale + inh); > > + } > > + > > + } > > + > > + outlink->w = width; > > + outlink->h = height; > > + outlink->frame_rate = inlink0->frame_rate; > > + outlink->time_base = av_inv_q(outlink->frame_rate); > > + outlink->sample_aspect_ratio = inlink0->sample_aspect_ratio; > > + > > + ret = init_framesync(ctx); > > + > > + if (ret < 0) > > + return ret; > > + > > + return ff_qsvvpp_create(ctx, &sctx->qsv, &sctx->qsv_param); > > +} > > + > > +/* > > + * Callback for qsvvpp > > + * @Note: qsvvpp composition does not generate PTS for result frame. > > + * so we assign the PTS from framesync to the output frame. > > + */ > > + > > +static int filter_callback(AVFilterLink *outlink, AVFrame *frame) > > +{ > > + QSVStackContext *sctx = outlink->src->priv; > > + > > + frame->pts = av_rescale_q(sctx->fs.pts, > > + sctx->fs.time_base, outlink->time_base); > > + return ff_filter_frame(outlink, frame); > > +} > > + > > + > > +static int stack_qsv_init(AVFilterContext *ctx) > > +{ > > + QSVStackContext *sctx = ctx->priv; > > + int i, ret; > > + > > + if (!strcmp(ctx->filter->name, "hstack_qsv")) > > + sctx->mode = QSV_STACK_H; > > + else if (!strcmp(ctx->filter->name, "vstack_qsv")) > > + sctx->mode = QSV_STACK_V; > > + else { > > + av_assert0(strcmp(ctx->filter->name, "xstack_qsv") == 0); > > + sctx->mode = QSV_STACK_X; > > + > > + if (strcmp(sctx->fillcolor_str, "none") && > > + av_parse_color(sctx->fillcolor, sctx->fillcolor_str, -1, ctx) > > >= 0) { > > + sctx->fillcolor_enable = 1; > > + } else { > > + sctx->fillcolor_enable = 0; > > + } > > + > > + if (!sctx->layout) { > > + if (sctx->nb_inputs == 2) { > > + sctx->layout = av_strdup("0_0|w0_0"); > > + > > + if (!sctx->layout) > > + return AVERROR(ENOMEM); > > + } else { > > + av_log(ctx, AV_LOG_ERROR, "No layout specified.\n"); > > + > > + return AVERROR(EINVAL); > > + } > > + } > > + } > > + > > + for (i = 0; i < sctx->nb_inputs; i++) { > > + AVFilterPad pad = { 0 }; > > + > > + pad.type = AVMEDIA_TYPE_VIDEO; > > + pad.name = av_asprintf("input%d", i); > > + > > + if (!pad.name) > > + return AVERROR(ENOMEM); > > + > > + if ((ret = ff_insert_inpad(ctx, i, &pad)) < 0) { > > + av_freep(&pad.name); > > + > > + return ret; > > + } > > + } > > + > > + /* fill composite config */ > > + sctx->comp_conf.Header.BufferId = MFX_EXTBUFF_VPP_COMPOSITE; > > + sctx->comp_conf.Header.BufferSz = sizeof(sctx->comp_conf); > > + sctx->comp_conf.NumInputStream = sctx->nb_inputs; > > + sctx->comp_conf.InputStream = av_mallocz_array(sctx->nb_inputs, > > + sizeof(*sctx- > > >comp_conf.InputStream)); > > + if (!sctx->comp_conf.InputStream) > > + return AVERROR(ENOMEM); > > + > > + /* initialize QSVVPP params */ > > + sctx->qsv_param.filter_frame = filter_callback; > > + sctx->qsv_param.ext_buf = av_mallocz(sizeof(*sctx->qsv_param.ext_buf)); > > + > > + if (!sctx->qsv_param.ext_buf) > > + return AVERROR(ENOMEM); > > + > > + sctx->qsv_param.ext_buf[0] = (mfxExtBuffer *)&sctx->comp_conf; > > + sctx->qsv_param.num_ext_buf = 1; > > + sctx->qsv_param.num_crop = 0; > > + > > + return 0; > > +} > > + > > +static av_cold void stack_qsv_uninit(AVFilterContext *ctx) > > +{ > > + QSVStackContext *sctx = ctx->priv; > > + int i; > > + > > + ff_qsvvpp_free(&sctx->qsv); > > + ff_framesync_uninit(&sctx->fs); > > + av_freep(&sctx->comp_conf.InputStream); > > + av_freep(&sctx->qsv_param.ext_buf); > > + > > + for (i = 0; i < ctx->nb_inputs; i++) > > + av_freep(&ctx->input_pads[i].name); > > +} > > + > > +static int stack_qsv_activate(AVFilterContext *ctx) > > +{ > > + QSVStackContext *sctx = ctx->priv; > > + return ff_framesync_activate(&sctx->fs); > > +} > > + > > +static int stack_qsv_query_formats(AVFilterContext *ctx) > > +{ > > + static const enum AVPixelFormat pixel_formats[] = { > > + AV_PIX_FMT_NV12, > > + AV_PIX_FMT_P010, > > + AV_PIX_FMT_QSV, > > + AV_PIX_FMT_NONE, > > + }; > > + AVFilterFormats *pix_fmts = ff_make_format_list(pixel_formats); > > + > > + return ff_set_common_formats(ctx, pix_fmts); > > +} > > + > > +static const AVFilterPad stack_qsv_outputs[] = { > > + { > > + .name = "default", > > + .type = AVMEDIA_TYPE_VIDEO, > > + .config_props = config_output, > > + }, > > + > > + { NULL } > > +}; > > + > > +static const AVOption stack_qsv_options[] = { > > + { "inputs", "set number of inputs", OFFSET(nb_inputs), AV_OPT_TYPE_INT, > > { .i64 = 2 }, 2, 72, .flags = FLAGS }, > > + { "shortest", "force termination when the shortest input terminates", > > OFFSET(shortest), AV_OPT_TYPE_BOOL, { .i64 = 0 }, 0, 1, FLAGS }, > > + { "scale", "set scale factor", OFFSET(scale), AV_OPT_TYPE_DOUBLE, { > > .dbl = 1.0 }, 0.125, 8, FLAGS }, > > + { NULL } > > +}; > > + > > +#define hstack_qsv_options stack_qsv_options > > +AVFILTER_DEFINE_CLASS(hstack_qsv); > > + > > +const AVFilter ff_vf_hstack_qsv = { > > + .name = "hstack_qsv", > > + .description = NULL_IF_CONFIG_SMALL("Quick Sync Video hstack."), > > + .priv_size = sizeof(QSVStackContext), > > + .priv_class = &hstack_qsv_class, > > + .query_formats = stack_qsv_query_formats, > > + .outputs = stack_qsv_outputs, > > + .init = stack_qsv_init, > > + .uninit = stack_qsv_uninit, > > + .activate = stack_qsv_activate, > > + .flags_internal = FF_FILTER_FLAG_HWFRAME_AWARE, > > + .flags = AVFILTER_FLAG_DYNAMIC_INPUTS, > > +}; > > + > > +#define vstack_qsv_options stack_qsv_options > > +AVFILTER_DEFINE_CLASS(vstack_qsv); > > + > > +const AVFilter ff_vf_vstack_qsv = { > > + .name = "vstack_qsv", > > + .description = NULL_IF_CONFIG_SMALL("Quick Sync Video vstack."), > > + .priv_size = sizeof(QSVStackContext), > > + .priv_class = &vstack_qsv_class, > > + .query_formats = stack_qsv_query_formats, > > + .outputs = stack_qsv_outputs, > > + .init = stack_qsv_init, > > + .uninit = stack_qsv_uninit, > > + .activate = stack_qsv_activate, > > + .flags_internal = FF_FILTER_FLAG_HWFRAME_AWARE, > > + .flags = AVFILTER_FLAG_DYNAMIC_INPUTS, > > +}; > > + > > +static const AVOption xstack_qsv_options[] = { > > + { "inputs", "set number of inputs", OFFSET(nb_inputs), AV_OPT_TYPE_INT, > > { .i64 = 2 }, 2, 72, .flags = FLAGS }, > > + { "layout", "set custom layout", OFFSET(layout), AV_OPT_TYPE_STRING, > > {.str = NULL}, 0, 0, .flags = FLAGS }, > > + { "shortest", "force termination when the shortest input terminates", > > OFFSET(shortest), AV_OPT_TYPE_BOOL, { .i64 = 0 }, 0, 1, FLAGS }, > > + { "fill", "set the color for unused pixels", OFFSET(fillcolor_str), > > AV_OPT_TYPE_STRING, {.str = "none"}, .flags = FLAGS }, > > + { "scale", "set scale factor", OFFSET(scale), AV_OPT_TYPE_DOUBLE, { > > .dbl = 1.0 }, 0.125, 8, FLAGS }, > > + { NULL } > > +}; > > + > > +AVFILTER_DEFINE_CLASS(xstack_qsv); > > + > > +const AVFilter ff_vf_xstack_qsv = { > > + .name = "xstack_qsv", > > + .description = NULL_IF_CONFIG_SMALL("Quick Sync Video xstack."), > > + .priv_size = sizeof(QSVStackContext), > > + .priv_class = &xstack_qsv_class, > > + .query_formats = stack_qsv_query_formats, > > + .outputs = stack_qsv_outputs, > > + .init = stack_qsv_init, > > + .uninit = stack_qsv_uninit, > > + .activate = stack_qsv_activate, > > + .flags_internal = FF_FILTER_FLAG_HWFRAME_AWARE, > > + .flags = AVFILTER_FLAG_DYNAMIC_INPUTS, > > +}; > > -- > > 2.25.1 > > Thanks for the stack related patch, functionally works for me at the > last attempt weeks ago. > The performance is explicitly faster than the combination of sets of > overlay_qsv filters on the platform I tested. > > Since qsv stack filters support scale factors (and customized layout > for xstack_qsv), > would you please help to provide some docs for these new qsv filters > to elaborate more like[1], > cmdline examples would help a lot for users, especially for xstack maybe. > > [1] https://ffmpeg.org/ffmpeg-all.html#vstack
Thanks for reviewing & testing the patch. I will update the patch to add some docs. - Haihao _______________________________________________ ffmpeg-devel mailing list ffmpeg-devel@ffmpeg.org https://ffmpeg.org/mailman/listinfo/ffmpeg-devel To unsubscribe, visit link above, or email ffmpeg-devel-requ...@ffmpeg.org with subject "unsubscribe".