This source file includes following definitions.
- init
- query_formats
- config_input
- config_output
- filter_frame
- uninit
#include "libavutil/avstring.h"
#include "libavutil/imgutils.h"
#include "libavutil/opt.h"
#include "libavutil/pixdesc.h"
#include "avfilter.h"
#include "formats.h"
#include "internal.h"
#include "video.h"
typedef struct DetelecineContext {
const AVClass *class;
int first_field;
char *pattern;
int start_frame;
int init_len;
unsigned int pattern_pos;
unsigned int nskip_fields;
int64_t start_time;
AVRational pts;
AVRational ts_unit;
int occupied;
int nb_planes;
int planeheight[4];
int stride[4];
AVFrame *frame[2];
AVFrame *temp;
} DetelecineContext;
#define OFFSET(x) offsetof(DetelecineContext, x)
#define FLAGS AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_VIDEO_PARAM
static const AVOption detelecine_options[] = {
{"first_field", "select first field", OFFSET(first_field), AV_OPT_TYPE_INT, {.i64=0}, 0, 1, FLAGS, "field"},
{"top", "select top field first", 0, AV_OPT_TYPE_CONST, {.i64=0}, 0, 0, FLAGS, "field"},
{"t", "select top field first", 0, AV_OPT_TYPE_CONST, {.i64=0}, 0, 0, FLAGS, "field"},
{"bottom", "select bottom field first", 0, AV_OPT_TYPE_CONST, {.i64=1}, 0, 0, FLAGS, "field"},
{"b", "select bottom field first", 0, AV_OPT_TYPE_CONST, {.i64=1}, 0, 0, FLAGS, "field"},
{"pattern", "pattern that describe for how many fields a frame is to be displayed", OFFSET(pattern), AV_OPT_TYPE_STRING, {.str="23"}, 0, 0, FLAGS},
{"start_frame", "position of first frame with respect to the pattern if stream is cut", OFFSET(start_frame), AV_OPT_TYPE_INT, {.i64=0}, 0, 13, FLAGS},
{NULL}
};
AVFILTER_DEFINE_CLASS(detelecine);
static av_cold int init(AVFilterContext *ctx)
{
DetelecineContext *s = ctx->priv;
const char *p;
int max = 0;
int sum = 0;
if (!strlen(s->pattern)) {
av_log(ctx, AV_LOG_ERROR, "No pattern provided.\n");
return AVERROR_INVALIDDATA;
}
for (p = s->pattern; *p; p++) {
if (!av_isdigit(*p)) {
av_log(ctx, AV_LOG_ERROR, "Provided pattern includes non-numeric characters.\n");
return AVERROR_INVALIDDATA;
}
sum += *p - '0';
max = FFMAX(*p - '0', max);
s->pts.num += *p - '0';
s->pts.den += 2;
}
if (s->start_frame >= sum) {
av_log(ctx, AV_LOG_ERROR, "Provided start_frame is too big.\n");
return AVERROR_INVALIDDATA;
}
s->nskip_fields = 0;
s->pattern_pos = 0;
s->start_time = AV_NOPTS_VALUE;
s->init_len = 0;
if (s->start_frame != 0) {
int nfields = 0;
for (p = s->pattern; *p; p++) {
nfields += *p - '0';
s->pattern_pos++;
if (nfields >= 2*s->start_frame) {
s->init_len = nfields - 2*s->start_frame;
break;
}
}
}
av_log(ctx, AV_LOG_INFO, "Detelecine pattern %s removes up to %d frames per frame, pts advance factor: %d/%d\n",
s->pattern, (max + 1) / 2, s->pts.num, s->pts.den);
return 0;
}
static int query_formats(AVFilterContext *ctx)
{
AVFilterFormats *pix_fmts = NULL;
int fmt, ret;
for (fmt = 0; av_pix_fmt_desc_get(fmt); fmt++) {
const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(fmt);
if (!(desc->flags & AV_PIX_FMT_FLAG_HWACCEL ||
desc->flags & AV_PIX_FMT_FLAG_PAL ||
desc->flags & AV_PIX_FMT_FLAG_BITSTREAM) &&
(ret = ff_add_format(&pix_fmts, fmt)) < 0)
return ret;
}
return ff_set_common_formats(ctx, pix_fmts);
}
static int config_input(AVFilterLink *inlink)
{
DetelecineContext *s = inlink->dst->priv;
const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(inlink->format);
int ret;
s->temp = ff_get_video_buffer(inlink, inlink->w, inlink->h);
if (!s->temp)
return AVERROR(ENOMEM);
s->frame[0] = ff_get_video_buffer(inlink, inlink->w, inlink->h);
if (!s->frame[0])
return AVERROR(ENOMEM);
s->frame[1] = ff_get_video_buffer(inlink, inlink->w, inlink->h);
if (!s->frame[1])
return AVERROR(ENOMEM);
if ((ret = av_image_fill_linesizes(s->stride, inlink->format, inlink->w)) < 0)
return ret;
s->planeheight[1] = s->planeheight[2] = AV_CEIL_RSHIFT(inlink->h, desc->log2_chroma_h);
s->planeheight[0] = s->planeheight[3] = inlink->h;
s->nb_planes = av_pix_fmt_count_planes(inlink->format);
return 0;
}
static int config_output(AVFilterLink *outlink)
{
AVFilterContext *ctx = outlink->src;
DetelecineContext *s = ctx->priv;
const AVFilterLink *inlink = ctx->inputs[0];
AVRational fps = inlink->frame_rate;
if (!fps.num || !fps.den) {
av_log(ctx, AV_LOG_ERROR, "The input needs a constant frame rate; "
"current rate of %d/%d is invalid\n", fps.num, fps.den);
return AVERROR(EINVAL);
}
fps = av_mul_q(fps, av_inv_q(s->pts));
av_log(ctx, AV_LOG_VERBOSE, "FPS: %d/%d -> %d/%d\n",
inlink->frame_rate.num, inlink->frame_rate.den, fps.num, fps.den);
outlink->frame_rate = fps;
outlink->time_base = av_mul_q(inlink->time_base, s->pts);
av_log(ctx, AV_LOG_VERBOSE, "TB: %d/%d -> %d/%d\n",
inlink->time_base.num, inlink->time_base.den, outlink->time_base.num, outlink->time_base.den);
s->ts_unit = av_inv_q(av_mul_q(fps, outlink->time_base));
return 0;
}
static int filter_frame(AVFilterLink *inlink, AVFrame *inpicref)
{
AVFilterContext *ctx = inlink->dst;
AVFilterLink *outlink = ctx->outputs[0];
DetelecineContext *s = ctx->priv;
int i, len = 0, ret = 0, out = 0;
if (s->start_time == AV_NOPTS_VALUE)
s->start_time = inpicref->pts;
if (s->nskip_fields >= 2) {
s->nskip_fields -= 2;
av_frame_free(&inpicref);
return 0;
} else if (s->nskip_fields >= 1) {
for (i = 0; i < s->nb_planes; i++) {
av_image_copy_plane(s->temp->data[i], s->temp->linesize[i],
inpicref->data[i], inpicref->linesize[i],
s->stride[i],
s->planeheight[i]);
}
s->occupied = 1;
s->nskip_fields--;
av_frame_free(&inpicref);
return 0;
}
if (s->nskip_fields == 0) {
len = s->init_len;
s->init_len = 0;
while(!len && s->pattern[s->pattern_pos]) {
len = s->pattern[s->pattern_pos] - '0';
s->pattern_pos++;
}
if (!s->pattern[s->pattern_pos])
s->pattern_pos = 0;
if(!len) {
av_frame_free(&inpicref);
return 0;
}
if (len == 1 && s->occupied) {
s->occupied = 0;
for (i = 0; i < s->nb_planes; i++)
av_image_copy_plane(s->frame[out]->data[i], s->frame[out]->linesize[i],
s->temp->data[i], s->temp->linesize[i],
s->stride[i],
s->planeheight[i]);
len = 0;
while(!len && s->pattern[s->pattern_pos]) {
len = s->pattern[s->pattern_pos] - '0';
s->pattern_pos++;
}
if (!s->pattern[s->pattern_pos])
s->pattern_pos = 0;
s->occupied = 0;
++out;
}
if (s->occupied) {
for (i = 0; i < s->nb_planes; i++) {
av_image_copy_plane(s->frame[out]->data[i] + s->frame[out]->linesize[i] * s->first_field,
s->frame[out]->linesize[i] * 2,
inpicref->data[i] + inpicref->linesize[i] * s->first_field,
inpicref->linesize[i] * 2,
s->stride[i],
(s->planeheight[i] - s->first_field + 1) / 2);
av_image_copy_plane(s->frame[out]->data[i] + s->frame[out]->linesize[i] * !s->first_field,
s->frame[out]->linesize[i] * 2,
s->temp->data[i] + s->temp->linesize[i] * !s->first_field,
s->temp->linesize[i] * 2,
s->stride[i],
(s->planeheight[i] - !s->first_field + 1) / 2);
}
s->occupied = 0;
if (len <= 2) {
for (i = 0; i < s->nb_planes; i++) {
av_image_copy_plane(s->temp->data[i], s->temp->linesize[i],
inpicref->data[i], inpicref->linesize[i],
s->stride[i],
s->planeheight[i]);
}
s->occupied = 1;
}
++out;
len = (len >= 3) ? len - 3 : 0;
} else {
if (len >= 2) {
for (i = 0; i < s->nb_planes; i++)
av_image_copy_plane(s->frame[out]->data[i], s->frame[out]->linesize[i],
inpicref->data[i], inpicref->linesize[i],
s->stride[i],
s->planeheight[i]);
len -= 2;
++out;
} else if (len == 1) {
for (i = 0; i < s->nb_planes; i++)
av_image_copy_plane(s->frame[out]->data[i], s->frame[out]->linesize[i],
inpicref->data[i], inpicref->linesize[i],
s->stride[i],
s->planeheight[i]);
for (i = 0; i < s->nb_planes; i++) {
av_image_copy_plane(s->temp->data[i], s->temp->linesize[i],
inpicref->data[i], inpicref->linesize[i],
s->stride[i],
s->planeheight[i]);
}
s->occupied = 1;
len--;
++out;
}
}
if (len == 1 && s->occupied)
{
len--;
s->occupied = 0;
}
}
s->nskip_fields = len;
for (i = 0; i < out; ++i) {
AVFrame *frame = av_frame_clone(s->frame[i]);
if (!frame) {
av_frame_free(&inpicref);
return AVERROR(ENOMEM);
}
av_frame_copy_props(frame, inpicref);
frame->pts = ((s->start_time == AV_NOPTS_VALUE) ? 0 : s->start_time) +
av_rescale(outlink->frame_count_in, s->ts_unit.num,
s->ts_unit.den);
ret = ff_filter_frame(outlink, frame);
}
av_frame_free(&inpicref);
return ret;
}
static av_cold void uninit(AVFilterContext *ctx)
{
DetelecineContext *s = ctx->priv;
av_frame_free(&s->temp);
av_frame_free(&s->frame[0]);
av_frame_free(&s->frame[1]);
}
static const AVFilterPad detelecine_inputs[] = {
{
.name = "default",
.type = AVMEDIA_TYPE_VIDEO,
.filter_frame = filter_frame,
.config_props = config_input,
},
{ NULL }
};
static const AVFilterPad detelecine_outputs[] = {
{
.name = "default",
.type = AVMEDIA_TYPE_VIDEO,
.config_props = config_output,
},
{ NULL }
};
AVFilter ff_vf_detelecine = {
.name = "detelecine",
.description = NULL_IF_CONFIG_SMALL("Apply an inverse telecine pattern."),
.priv_size = sizeof(DetelecineContext),
.priv_class = &detelecine_class,
.init = init,
.uninit = uninit,
.query_formats = query_formats,
.inputs = detelecine_inputs,
.outputs = detelecine_outputs,
};