This source file includes following definitions.
- eval_expr
- have_alpha_planar
- config_main_input
- config_overlay_input
- process_frame
- init_framesync
- config_output
- filter_callback
- overlay_qsv_init
- overlay_qsv_uninit
- activate
- overlay_qsv_query_formats
#include "libavutil/opt.h"
#include "libavutil/common.h"
#include "libavutil/pixdesc.h"
#include "libavutil/eval.h"
#include "libavutil/hwcontext.h"
#include "libavutil/avstring.h"
#include "libavutil/avassert.h"
#include "libavutil/imgutils.h"
#include "libavutil/mathematics.h"
#include "internal.h"
#include "avfilter.h"
#include "formats.h"
#include "video.h"
#include "framesync.h"
#include "qsvvpp.h"
#define MAIN 0
#define OVERLAY 1
#define OFFSET(x) offsetof(QSVOverlayContext, x)
#define FLAGS (AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_FILTERING_PARAM)
enum var_name {
VAR_MAIN_iW, VAR_MW,
VAR_MAIN_iH, VAR_MH,
VAR_OVERLAY_iW,
VAR_OVERLAY_iH,
VAR_OVERLAY_X, VAR_OX,
VAR_OVERLAY_Y, VAR_OY,
VAR_OVERLAY_W, VAR_OW,
VAR_OVERLAY_H, VAR_OH,
VAR_VARS_NB
};
typedef struct QSVOverlayContext {
const AVClass *class;
FFFrameSync fs;
QSVVPPContext *qsv;
QSVVPPParam qsv_param;
mfxExtVPPComposite comp_conf;
double var_values[VAR_VARS_NB];
char *overlay_ox, *overlay_oy, *overlay_ow, *overlay_oh;
uint16_t overlay_alpha, overlay_pixel_alpha;
} QSVOverlayContext;
static const char *const var_names[] = {
"main_w", "W",
"main_h", "H",
"overlay_iw",
"overlay_ih",
"overlay_x", "x",
"overlay_y", "y",
"overlay_w", "w",
"overlay_h", "h",
NULL
};
static const AVOption overlay_qsv_options[] = {
{ "x", "Overlay x position", OFFSET(overlay_ox), AV_OPT_TYPE_STRING, { .str="0"}, 0, 255, .flags = FLAGS},
{ "y", "Overlay y position", OFFSET(overlay_oy), AV_OPT_TYPE_STRING, { .str="0"}, 0, 255, .flags = FLAGS},
{ "w", "Overlay width", OFFSET(overlay_ow), AV_OPT_TYPE_STRING, { .str="overlay_iw"}, 0, 255, .flags = FLAGS},
{ "h", "Overlay height", OFFSET(overlay_oh), AV_OPT_TYPE_STRING, { .str="overlay_ih*w/overlay_iw"}, 0, 255, .flags = FLAGS},
{ "alpha", "Overlay global alpha", OFFSET(overlay_alpha), AV_OPT_TYPE_INT, { .i64 = 255}, 0, 255, .flags = FLAGS},
{ "eof_action", "Action to take when encountering EOF from secondary input ",
OFFSET(fs.opt_eof_action), AV_OPT_TYPE_INT, { .i64 = EOF_ACTION_REPEAT },
EOF_ACTION_REPEAT, EOF_ACTION_PASS, .flags = FLAGS, "eof_action" },
{ "repeat", "Repeat the previous frame.", 0, AV_OPT_TYPE_CONST, { .i64 = EOF_ACTION_REPEAT }, .flags = FLAGS, "eof_action" },
{ "endall", "End both streams.", 0, AV_OPT_TYPE_CONST, { .i64 = EOF_ACTION_ENDALL }, .flags = FLAGS, "eof_action" },
{ "pass", "Pass through the main input.", 0, AV_OPT_TYPE_CONST, { .i64 = EOF_ACTION_PASS }, .flags = FLAGS, "eof_action" },
{ "shortest", "force termination when the shortest input terminates", OFFSET(fs.opt_shortest), AV_OPT_TYPE_BOOL, { .i64 = 0 }, 0, 1, FLAGS },
{ "repeatlast", "repeat overlay of the last overlay frame", OFFSET(fs.opt_repeatlast), AV_OPT_TYPE_BOOL, {.i64=1}, 0, 1, FLAGS },
{ NULL }
};
FRAMESYNC_DEFINE_CLASS(overlay_qsv, QSVOverlayContext, fs);
static int eval_expr(AVFilterContext *ctx)
{
QSVOverlayContext *vpp = ctx->priv;
double *var_values = vpp->var_values;
int ret = 0;
AVExpr *ox_expr = NULL, *oy_expr = NULL;
AVExpr *ow_expr = NULL, *oh_expr = NULL;
#define PASS_EXPR(e, s) {\
ret = av_expr_parse(&e, s, var_names, NULL, NULL, NULL, NULL, 0, ctx); \
if (ret < 0) {\
av_log(ctx, AV_LOG_ERROR, "Error when passing '%s'.\n", s);\
goto release;\
}\
}
PASS_EXPR(ox_expr, vpp->overlay_ox);
PASS_EXPR(oy_expr, vpp->overlay_oy);
PASS_EXPR(ow_expr, vpp->overlay_ow);
PASS_EXPR(oh_expr, vpp->overlay_oh);
#undef PASS_EXPR
var_values[VAR_OVERLAY_W] =
var_values[VAR_OW] = av_expr_eval(ow_expr, var_values, NULL);
var_values[VAR_OVERLAY_H] =
var_values[VAR_OH] = av_expr_eval(oh_expr, var_values, NULL);
var_values[VAR_OVERLAY_W] =
var_values[VAR_OW] = av_expr_eval(ow_expr, var_values, NULL);
var_values[VAR_OVERLAY_X] =
var_values[VAR_OX] = av_expr_eval(ox_expr, var_values, NULL);
var_values[VAR_OVERLAY_Y] =
var_values[VAR_OY] = av_expr_eval(oy_expr, var_values, NULL);
var_values[VAR_OVERLAY_X] =
var_values[VAR_OX] = av_expr_eval(ox_expr, var_values, NULL);
var_values[VAR_OVERLAY_W] =
var_values[VAR_OW] = av_expr_eval(ow_expr, var_values, NULL);
var_values[VAR_OVERLAY_H] =
var_values[VAR_OH] = av_expr_eval(oh_expr, var_values, NULL);
var_values[VAR_OVERLAY_W] =
var_values[VAR_OW] = av_expr_eval(ow_expr, var_values, NULL);
release:
av_expr_free(ox_expr);
av_expr_free(oy_expr);
av_expr_free(ow_expr);
av_expr_free(oh_expr);
return ret;
}
static int have_alpha_planar(AVFilterLink *link)
{
enum AVPixelFormat pix_fmt = link->format;
const AVPixFmtDescriptor *desc;
AVHWFramesContext *fctx;
if (link->format == AV_PIX_FMT_QSV) {
fctx = (AVHWFramesContext *)link->hw_frames_ctx->data;
pix_fmt = fctx->sw_format;
}
desc = av_pix_fmt_desc_get(pix_fmt);
if (!desc)
return 0;
return !!(desc->flags & AV_PIX_FMT_FLAG_ALPHA);
}
static int config_main_input(AVFilterLink *inlink)
{
AVFilterContext *ctx = inlink->dst;
QSVOverlayContext *vpp = ctx->priv;
mfxVPPCompInputStream *st = &vpp->comp_conf.InputStream[0];
av_log(ctx, AV_LOG_DEBUG, "Input[%d] is of %s.\n", FF_INLINK_IDX(inlink),
av_get_pix_fmt_name(inlink->format));
vpp->var_values[VAR_MAIN_iW] =
vpp->var_values[VAR_MW] = inlink->w;
vpp->var_values[VAR_MAIN_iH] =
vpp->var_values[VAR_MH] = inlink->h;
st->DstX = 0;
st->DstY = 0;
st->DstW = inlink->w;
st->DstH = inlink->h;
st->GlobalAlphaEnable = 0;
st->PixelAlphaEnable = 0;
return 0;
}
static int config_overlay_input(AVFilterLink *inlink)
{
AVFilterContext *ctx = inlink->dst;
QSVOverlayContext *vpp = ctx->priv;
mfxVPPCompInputStream *st = &vpp->comp_conf.InputStream[1];
int ret = 0;
av_log(ctx, AV_LOG_DEBUG, "Input[%d] is of %s.\n", FF_INLINK_IDX(inlink),
av_get_pix_fmt_name(inlink->format));
vpp->var_values[VAR_OVERLAY_iW] = inlink->w;
vpp->var_values[VAR_OVERLAY_iH] = inlink->h;
ret = eval_expr(ctx);
if (ret < 0)
return ret;
st->DstX = vpp->var_values[VAR_OX];
st->DstY = vpp->var_values[VAR_OY];
st->DstW = vpp->var_values[VAR_OW];
st->DstH = vpp->var_values[VAR_OH];
st->GlobalAlpha = vpp->overlay_alpha;
st->GlobalAlphaEnable = (st->GlobalAlpha < 255);
st->PixelAlphaEnable = have_alpha_planar(inlink);
return 0;
}
static int process_frame(FFFrameSync *fs)
{
AVFilterContext *ctx = fs->parent;
QSVOverlayContext *s = fs->opaque;
AVFrame *frame = NULL;
int ret = 0, i;
for (i = 0; i < ctx->nb_inputs; i++) {
ret = ff_framesync_get_frame(fs, i, &frame, 0);
if (ret == 0)
ret = ff_qsvvpp_filter_frame(s->qsv, ctx->inputs[i], frame);
if (ret < 0 && ret != AVERROR(EAGAIN))
break;
}
return ret;
}
static int init_framesync(AVFilterContext *ctx)
{
QSVOverlayContext *s = ctx->priv;
int ret, i;
s->fs.on_event = process_frame;
s->fs.opaque = s;
ret = ff_framesync_init(&s->fs, ctx, ctx->nb_inputs);
if (ret < 0)
return ret;
for (i = 0; i < ctx->nb_inputs; i++) {
FFFrameSyncIn *in = &s->fs.in[i];
in->before = EXT_STOP;
in->after = EXT_INFINITY;
in->sync = i ? 1 : 2;
in->time_base = ctx->inputs[i]->time_base;
}
return ff_framesync_configure(&s->fs);
}
static int config_output(AVFilterLink *outlink)
{
AVFilterContext *ctx = outlink->src;
QSVOverlayContext *vpp = ctx->priv;
AVFilterLink *in0 = ctx->inputs[0];
AVFilterLink *in1 = ctx->inputs[1];
int ret;
av_log(ctx, AV_LOG_DEBUG, "Output is of %s.\n", av_get_pix_fmt_name(outlink->format));
if ((in0->format == AV_PIX_FMT_QSV && in1->format != AV_PIX_FMT_QSV) ||
(in0->format != AV_PIX_FMT_QSV && in1->format == AV_PIX_FMT_QSV)) {
av_log(ctx, AV_LOG_ERROR, "Mixing hardware and software pixel formats is not supported.\n");
return AVERROR(EINVAL);
} else if (in0->format == AV_PIX_FMT_QSV) {
AVHWFramesContext *hw_frame0 = (AVHWFramesContext *)in0->hw_frames_ctx->data;
AVHWFramesContext *hw_frame1 = (AVHWFramesContext *)in1->hw_frames_ctx->data;
if (hw_frame0->device_ctx != hw_frame1->device_ctx) {
av_log(ctx, AV_LOG_ERROR, "Inputs with different underlying QSV devices are forbidden.\n");
return AVERROR(EINVAL);
}
}
outlink->w = vpp->var_values[VAR_MW];
outlink->h = vpp->var_values[VAR_MH];
outlink->frame_rate = in0->frame_rate;
outlink->time_base = av_inv_q(outlink->frame_rate);
ret = init_framesync(ctx);
if (ret < 0)
return ret;
return ff_qsvvpp_create(ctx, &vpp->qsv, &vpp->qsv_param);
}
static int filter_callback(AVFilterLink *outlink, AVFrame *frame)
{
QSVOverlayContext *s = outlink->src->priv;
frame->pts = av_rescale_q(s->fs.pts,
s->fs.time_base, outlink->time_base);
return ff_filter_frame(outlink, frame);
}
static int overlay_qsv_init(AVFilterContext *ctx)
{
QSVOverlayContext *vpp = ctx->priv;
vpp->comp_conf.Header.BufferId = MFX_EXTBUFF_VPP_COMPOSITE;
vpp->comp_conf.Header.BufferSz = sizeof(vpp->comp_conf);
vpp->comp_conf.NumInputStream = ctx->nb_inputs;
vpp->comp_conf.InputStream = av_mallocz_array(ctx->nb_inputs,
sizeof(*vpp->comp_conf.InputStream));
if (!vpp->comp_conf.InputStream)
return AVERROR(ENOMEM);
vpp->qsv_param.filter_frame = filter_callback;
vpp->qsv_param.ext_buf = av_mallocz(sizeof(*vpp->qsv_param.ext_buf));
if (!vpp->qsv_param.ext_buf)
return AVERROR(ENOMEM);
vpp->qsv_param.ext_buf[0] = (mfxExtBuffer *)&vpp->comp_conf;
vpp->qsv_param.num_ext_buf = 1;
vpp->qsv_param.out_sw_format = AV_PIX_FMT_NV12;
vpp->qsv_param.num_crop = 0;
return 0;
}
static av_cold void overlay_qsv_uninit(AVFilterContext *ctx)
{
QSVOverlayContext *vpp = ctx->priv;
ff_qsvvpp_free(&vpp->qsv);
ff_framesync_uninit(&vpp->fs);
av_freep(&vpp->comp_conf.InputStream);
av_freep(&vpp->qsv_param.ext_buf);
}
static int activate(AVFilterContext *ctx)
{
QSVOverlayContext *s = ctx->priv;
return ff_framesync_activate(&s->fs);
}
static int overlay_qsv_query_formats(AVFilterContext *ctx)
{
int i;
int ret;
static const enum AVPixelFormat main_in_fmts[] = {
AV_PIX_FMT_YUV420P,
AV_PIX_FMT_NV12,
AV_PIX_FMT_YUYV422,
AV_PIX_FMT_RGB32,
AV_PIX_FMT_QSV,
AV_PIX_FMT_NONE
};
static const enum AVPixelFormat out_pix_fmts[] = {
AV_PIX_FMT_NV12,
AV_PIX_FMT_QSV,
AV_PIX_FMT_NONE
};
for (i = 0; i < ctx->nb_inputs; i++) {
ret = ff_formats_ref(ff_make_format_list(main_in_fmts), &ctx->inputs[i]->out_formats);
if (ret < 0)
return ret;
}
ret = ff_formats_ref(ff_make_format_list(out_pix_fmts), &ctx->outputs[0]->in_formats);
if (ret < 0)
return ret;
return 0;
}
static const AVFilterPad overlay_qsv_inputs[] = {
{
.name = "main",
.type = AVMEDIA_TYPE_VIDEO,
.config_props = config_main_input,
.needs_fifo = 1,
},
{
.name = "overlay",
.type = AVMEDIA_TYPE_VIDEO,
.config_props = config_overlay_input,
.needs_fifo = 1,
},
{ NULL }
};
static const AVFilterPad overlay_qsv_outputs[] = {
{
.name = "default",
.type = AVMEDIA_TYPE_VIDEO,
.config_props = config_output,
},
{ NULL }
};
AVFilter ff_vf_overlay_qsv = {
.name = "overlay_qsv",
.description = NULL_IF_CONFIG_SMALL("Quick Sync Video overlay."),
.priv_size = sizeof(QSVOverlayContext),
.query_formats = overlay_qsv_query_formats,
.preinit = overlay_qsv_framesync_preinit,
.init = overlay_qsv_init,
.uninit = overlay_qsv_uninit,
.activate = activate,
.inputs = overlay_qsv_inputs,
.outputs = overlay_qsv_outputs,
.priv_class = &overlay_qsv_class,
.flags_internal = FF_FILTER_FLAG_HWFRAME_AWARE,
};