@@ -363,6 +363,7 @@ static int filter_frame(AVFilterLink *inlink, AVFrame *in)
{
AVFilterContext *ctx = inlink->dst;
AVFilterLink *outlink = ctx->outputs[0];
+ FilterLink *outl = ff_filter_link(outlink);
AudioDRCContext *s = ctx->priv;
AVFrame *out;
int ret;
@@ -373,7 +374,7 @@ static int filter_frame(AVFilterLink *inlink, AVFrame *in)
goto fail;
}
- s->var_values[VAR_SN] = outlink->sample_count_in;
+ s->var_values[VAR_SN] = outl->sample_count_in;
s->var_values[VAR_T] = s->var_values[VAR_SN] * (double)1/outlink->sample_rate;
s->in = in;
@@ -355,8 +355,9 @@ static void process_frame(AVFilterContext *ctx,
double *prior, double *prior_band_excit, int track_noise)
{
AVFilterLink *outlink = ctx->outputs[0];
+ FilterLink *outl = ff_filter_link(outlink);
const double *abs_var = dnch->abs_var;
- const double ratio = outlink->frame_count_out ? s->ratio : 1.0;
+ const double ratio = outl->frame_count_out ? s->ratio : 1.0;
const double rratio = 1. - ratio;
const int *bin2band = s->bin2band;
double *noisy_data = dnch->noisy_data;
@@ -38,6 +38,7 @@
#include "audio.h"
#include "avfilter.h"
+#include "filters.h"
#include "internal.h"
typedef struct AShowInfoContext {
@@ -173,6 +174,7 @@ static void dump_unknown(AVFilterContext *ctx, AVFrameSideData *sd)
static int filter_frame(AVFilterLink *inlink, AVFrame *buf)
{
+ FilterLink *inl = ff_filter_link(inlink);
AVFilterContext *ctx = inlink->dst;
AShowInfoContext *s = ctx->priv;
char chlayout_str[128];
@@ -203,7 +205,7 @@ static int filter_frame(AVFilterLink *inlink, AVFrame *buf)
"n:%"PRId64" pts:%s pts_time:%s "
"fmt:%s channels:%d chlayout:%s rate:%d nb_samples:%d "
"checksum:%08"PRIX32" ",
- inlink->frame_count_out,
+ inl->frame_count_out,
av_ts2str(buf->pts), av_ts2timestr(buf->pts, &inlink->time_base),
av_get_sample_fmt_name(buf->format), buf->ch_layout.nb_channels, chlayout_str,
buf->sample_rate, buf->nb_samples,
@@ -725,6 +725,7 @@ static void perform_compression(DynamicAudioNormalizerContext *s, AVFrame *frame
static int analyze_frame(AVFilterContext *ctx, AVFilterLink *outlink, AVFrame **frame)
{
+ FilterLink *outl = ff_filter_link(outlink);
DynamicAudioNormalizerContext *s = ctx->priv;
AVFrame *analyze_frame;
@@ -780,7 +781,7 @@ static int analyze_frame(AVFilterContext *ctx, AVFilterLink *outlink, AVFrame **
analyze_frame = *frame;
}
- s->var_values[VAR_SN] = outlink->sample_count_in;
+ s->var_values[VAR_SN] = outl->sample_count_in;
s->var_values[VAR_T] = s->var_values[VAR_SN] * (double)1/outlink->sample_rate;
if (s->channels_coupled) {
@@ -35,6 +35,7 @@
#include "audio.h"
#include "avfilter.h"
+#include "filters.h"
#include "formats.h"
#include "internal.h"
#include "af_volume.h"
@@ -328,6 +329,7 @@ static int process_command(AVFilterContext *ctx, const char *cmd, const char *ar
static int filter_frame(AVFilterLink *inlink, AVFrame *buf)
{
+ FilterLink *inl = ff_filter_link(inlink);
AVFilterContext *ctx = inlink->dst;
VolumeContext *vol = inlink->dst->priv;
AVFilterLink *outlink = inlink->dst->outputs[0];
@@ -380,7 +382,7 @@ static int filter_frame(AVFilterLink *inlink, AVFrame *buf)
}
vol->var_values[VAR_PTS] = TS2D(buf->pts);
vol->var_values[VAR_T ] = TS2T(buf->pts, inlink->time_base);
- vol->var_values[VAR_N ] = inlink->frame_count_out;
+ vol->var_values[VAR_N ] = inl->frame_count_out;
#if FF_API_FRAME_PKT
FF_DISABLE_DEPRECATION_WARNINGS
@@ -207,10 +207,11 @@ static av_cold int config_props(AVFilterLink *outlink)
static int activate(AVFilterContext *ctx)
{
AVFilterLink *outlink = ctx->outputs[0];
+ FilterLink *outl = ff_filter_link(outlink);
SineContext *sine = ctx->priv;
AVFrame *frame;
double values[VAR_VARS_NB] = {
- [VAR_N] = outlink->frame_count_in,
+ [VAR_N] = outl->frame_count_in,
[VAR_PTS] = sine->pts,
[VAR_T] = sine->pts * av_q2d(outlink->time_base),
[VAR_TB] = av_q2d(outlink->time_base),
@@ -298,6 +298,7 @@ static inline void plot_freq(ShowFreqsContext *s, int ch,
double a, int f, uint8_t fg[4], int *prev_y,
AVFrame *out, AVFilterLink *outlink)
{
+ FilterLink *outl = ff_filter_link(outlink);
const int w = s->w;
const float min = s->minamp;
const float avg = s->avg_data[ch][f];
@@ -337,12 +338,12 @@ static inline void plot_freq(ShowFreqsContext *s, int ch,
switch (s->avg) {
case 0:
- y = s->avg_data[ch][f] = !outlink->frame_count_in ? y : FFMIN(0, y);
+ y = s->avg_data[ch][f] = !outl->frame_count_in ? y : FFMIN(0, y);
break;
case 1:
break;
default:
- s->avg_data[ch][f] = avg + y * (y - avg) / (FFMIN(outlink->frame_count_in + 1, s->avg) * (float)y);
+ s->avg_data[ch][f] = avg + y * (y - avg) / (FFMIN(outl->frame_count_in + 1, s->avg) * (float)y);
y = av_clip(s->avg_data[ch][f], 0, outlink->h - 1);
break;
}
@@ -991,6 +991,7 @@ static int default_filter_frame(AVFilterLink *link, AVFrame *frame)
static int ff_filter_frame_framed(AVFilterLink *link, AVFrame *frame)
{
+ FilterLink *l = ff_filter_link(link);
int (*filter_frame)(AVFilterLink *, AVFrame *);
AVFilterContext *dstctx = link->dst;
AVFilterPad *dst = link->dstpad;
@@ -1012,7 +1013,7 @@ static int ff_filter_frame_framed(AVFilterLink *link, AVFrame *frame)
(dstctx->filter->flags & AVFILTER_FLAG_SUPPORT_TIMELINE_GENERIC))
filter_frame = default_filter_frame;
ret = filter_frame(link, frame);
- link->frame_count_out++;
+ l->frame_count_out++;
return ret;
fail:
@@ -1058,8 +1059,8 @@ int ff_filter_frame(AVFilterLink *link, AVFrame *frame)
}
li->frame_blocked_in = link->frame_wanted_out = 0;
- link->frame_count_in++;
- link->sample_count_in += frame->nb_samples;
+ li->l.frame_count_in++;
+ li->l.sample_count_in += frame->nb_samples;
filter_unblock(link->dst);
ret = ff_framequeue_add(&li->fifo, frame);
if (ret < 0) {
@@ -1163,7 +1164,7 @@ static int ff_filter_frame_to_filter(AVFilterLink *link)
filter_unblock(dst);
/* AVFilterPad.filter_frame() expect frame_count_out to have the value
before the frame; ff_filter_frame_framed() will re-increment it. */
- link->frame_count_out--;
+ li->l.frame_count_out--;
ret = ff_filter_frame_framed(link, frame);
if (ret < 0 && ret != li->status_out) {
link_set_out_status(link, ret, AV_NOPTS_VALUE);
@@ -1444,8 +1445,8 @@ static void consume_update(FilterLinkInternal *li, const AVFrame *frame)
ff_inlink_process_commands(link, frame);
if (link == link->dst->inputs[0])
link->dst->is_disabled = !ff_inlink_evaluate_timeline_at_frame(link, frame);
- link->frame_count_out++;
- link->sample_count_out += frame->nb_samples;
+ li->l.frame_count_out++;
+ li->l.sample_count_out += frame->nb_samples;
}
int ff_inlink_consume_frame(AVFilterLink *link, AVFrame **rframe)
@@ -1552,6 +1553,7 @@ int ff_inlink_process_commands(AVFilterLink *link, const AVFrame *frame)
int ff_inlink_evaluate_timeline_at_frame(AVFilterLink *link, const AVFrame *frame)
{
+ FilterLink *l = ff_filter_link(link);
AVFilterContext *dstctx = link->dst;
int64_t pts = frame->pts;
#if FF_API_FRAME_PKT
@@ -1563,7 +1565,7 @@ FF_ENABLE_DEPRECATION_WARNINGS
if (!dstctx->enable_str)
return 1;
- dstctx->var_values[VAR_N] = link->frame_count_out;
+ dstctx->var_values[VAR_N] = l->frame_count_out;
dstctx->var_values[VAR_T] = pts == AV_NOPTS_VALUE ? NAN : pts * av_q2d(link->time_base);
dstctx->var_values[VAR_W] = link->w;
dstctx->var_values[VAR_H] = link->h;
@@ -600,16 +600,6 @@ struct AVFilterLink {
*/
struct AVFilterGraph *graph;
- /**
- * Number of past frames sent through the link.
- */
- int64_t frame_count_in, frame_count_out;
-
- /**
- * Number of past samples sent through the link.
- */
- int64_t sample_count_in, sample_count_out;
-
/**
* True if a frame is currently wanted on the output of this filter.
* Set when ff_request_frame() is called by the output,
@@ -1402,8 +1402,8 @@ int avfilter_graph_request_oldest(AVFilterGraph *graph)
return AVERROR_EOF;
av_assert1(!oldest->dst->filter->activate);
av_assert1(oldesti->age_index >= 0);
- frame_count = oldest->frame_count_out;
- while (frame_count == oldest->frame_count_out) {
+ frame_count = oldesti->l.frame_count_out;
+ while (frame_count == oldesti->l.frame_count_out) {
r = ff_filter_graph_run_once(graph);
if (r == AVERROR(EAGAIN) &&
!oldest->frame_wanted_out && !oldesti->frame_blocked_in &&
@@ -308,33 +308,33 @@ static int draw_items(AVFilterContext *ctx,
drawtext(out, xpos, ypos, buffer, len, frames > 0 ? frames >= 10 ? frames >= 50 ? s->red : s->yellow : s->green : s->white);
xpos += len * 8;
}
- if ((flags & FLAG_FCIN) && (!(mode & MODE_NOZERO) || l->frame_count_in)) {
- len = snprintf(buffer, sizeof(buffer)-1, " | in: %"PRId64, l->frame_count_in);
+ if ((flags & FLAG_FCIN) && (!(mode & MODE_NOZERO) || fl->frame_count_in)) {
+ len = snprintf(buffer, sizeof(buffer)-1, " | in: %"PRId64, fl->frame_count_in);
drawtext(out, xpos, ypos, buffer, len, s->white);
xpos += len * 8;
}
- if ((flags & FLAG_FCOUT) && (!(mode & MODE_NOZERO) || l->frame_count_out)) {
- len = snprintf(buffer, sizeof(buffer)-1, " | out: %"PRId64, l->frame_count_out);
+ if ((flags & FLAG_FCOUT) && (!(mode & MODE_NOZERO) || fl->frame_count_out)) {
+ len = snprintf(buffer, sizeof(buffer)-1, " | out: %"PRId64, fl->frame_count_out);
drawtext(out, xpos, ypos, buffer, len, s->white);
xpos += len * 8;
}
- if ((flags & FLAG_FC_DELTA) && (!(mode & MODE_NOZERO) || (l->frame_count_in - l->frame_count_out))) {
- len = snprintf(buffer, sizeof(buffer)-1, " | delta: %"PRId64, l->frame_count_in - l->frame_count_out);
+ if ((flags & FLAG_FC_DELTA) && (!(mode & MODE_NOZERO) || (fl->frame_count_in - fl->frame_count_out))) {
+ len = snprintf(buffer, sizeof(buffer)-1, " | delta: %"PRId64, fl->frame_count_in - fl->frame_count_out);
drawtext(out, xpos, ypos, buffer, len, s->white);
xpos += len * 8;
}
- if ((flags & FLAG_SCIN) && (!(mode & MODE_NOZERO) || l->sample_count_in)) {
- len = snprintf(buffer, sizeof(buffer)-1, " | sin: %"PRId64, l->sample_count_in);
+ if ((flags & FLAG_SCIN) && (!(mode & MODE_NOZERO) || fl->sample_count_in)) {
+ len = snprintf(buffer, sizeof(buffer)-1, " | sin: %"PRId64, fl->sample_count_in);
drawtext(out, xpos, ypos, buffer, len, s->white);
xpos += len * 8;
}
- if ((flags & FLAG_SCOUT) && (!(mode & MODE_NOZERO) || l->sample_count_out)) {
- len = snprintf(buffer, sizeof(buffer)-1, " | sout: %"PRId64, l->sample_count_out);
+ if ((flags & FLAG_SCOUT) && (!(mode & MODE_NOZERO) || fl->sample_count_out)) {
+ len = snprintf(buffer, sizeof(buffer)-1, " | sout: %"PRId64, fl->sample_count_out);
drawtext(out, xpos, ypos, buffer, len, s->white);
xpos += len * 8;
}
- if ((flags & FLAG_SC_DELTA) && (!(mode & MODE_NOZERO) || (l->sample_count_in - l->sample_count_out))) {
- len = snprintf(buffer, sizeof(buffer)-1, " | sdelta: %"PRId64, l->sample_count_in - l->sample_count_out);
+ if ((flags & FLAG_SC_DELTA) && (!(mode & MODE_NOZERO) || (fl->sample_count_in - fl->sample_count_out))) {
+ len = snprintf(buffer, sizeof(buffer)-1, " | sdelta: %"PRId64, fl->sample_count_in - fl->sample_count_out);
drawtext(out, xpos, ypos, buffer, len, s->white);
xpos += len * 8;
}
@@ -46,6 +46,7 @@ static int activate(AVFilterContext *ctx)
{
LatencyContext *s = ctx->priv;
AVFilterLink *inlink = ctx->inputs[0];
+ FilterLink *inl = ff_filter_link(inlink);
AVFilterLink *outlink = ctx->outputs[0];
FF_FILTER_FORWARD_STATUS_BACK(outlink, inlink);
@@ -53,14 +54,15 @@ static int activate(AVFilterContext *ctx)
if (!ctx->is_disabled && ctx->inputs[0]->src &&
ctx->inputs[0]->src->nb_inputs > 0) {
AVFilterLink *prevlink = ctx->inputs[0]->src->inputs[0];
+ FilterLink *prevl = ff_filter_link(prevlink);
int64_t delta = 0;
switch (prevlink->type) {
case AVMEDIA_TYPE_AUDIO:
- delta = prevlink->sample_count_in - inlink->sample_count_out;
+ delta = prevl->sample_count_in - inl->sample_count_out;
break;
case AVMEDIA_TYPE_VIDEO:
- delta = prevlink->frame_count_in - inlink->frame_count_out;
+ delta = prevl->frame_count_in - inl->frame_count_out;
break;
}
@@ -144,6 +144,7 @@ static int push_samples(AVFilterContext *ctx, int nb_samples)
static int afilter_frame(AVFilterLink *inlink, AVFrame *frame)
{
+ FilterLink *inl = ff_filter_link(inlink);
AVFilterContext *ctx = inlink->dst;
AVFilterLink *outlink = ctx->outputs[0];
LoopContext *s = ctx->priv;
@@ -158,7 +159,7 @@ static int afilter_frame(AVFilterLink *inlink, AVFrame *frame)
int drain = 0;
if (s->start < 0)
- s->start = inlink->sample_count_out - written;
+ s->start = inl->sample_count_out - written;
ret = av_audio_fifo_write(s->fifo, (void **)frame->extended_data, written);
if (ret < 0)
@@ -374,6 +375,7 @@ static int push_frame(AVFilterContext *ctx)
static int filter_frame(AVFilterLink *inlink, AVFrame *frame)
{
+ FilterLink *inl = ff_filter_link(inlink);
AVFilterContext *ctx = inlink->dst;
AVFilterLink *outlink = ctx->outputs[0];
FilterLink *outl = ff_filter_link(outlink);
@@ -381,7 +383,7 @@ static int filter_frame(AVFilterLink *inlink, AVFrame *frame)
int64_t duration;
int ret = 0;
- if (((s->start >= 0 && inlink->frame_count_out >= s->start) ||
+ if (((s->start >= 0 && inl->frame_count_out >= s->start) ||
(s->time_pts != AV_NOPTS_VALUE &&
frame->pts >= s->time_pts)) &&
s->size > 0 && s->loop != 0) {
@@ -36,6 +36,7 @@
#include "libavformat/avio.h"
#include "avfilter.h"
#include "audio.h"
+#include "filters.h"
#include "internal.h"
#include "video.h"
@@ -303,6 +304,7 @@ static av_cold void uninit(AVFilterContext *ctx)
static int filter_frame(AVFilterLink *inlink, AVFrame *frame)
{
+ FilterLink *inl = ff_filter_link(inlink);
AVFilterContext *ctx = inlink->dst;
AVFilterLink *outlink = ctx->outputs[0];
MetadataContext *s = ctx->priv;
@@ -336,14 +338,14 @@ static int filter_frame(AVFilterLink *inlink, AVFrame *frame)
case METADATA_PRINT:
if (!s->key && e) {
s->print(ctx, "frame:%-4"PRId64" pts:%-7s pts_time:%s\n",
- inlink->frame_count_out, av_ts2str(frame->pts), av_ts2timestr(frame->pts, &inlink->time_base));
+ inl->frame_count_out, av_ts2str(frame->pts), av_ts2timestr(frame->pts, &inlink->time_base));
s->print(ctx, "%s=%s\n", e->key, e->value);
while ((e = av_dict_iterate(*metadata, e)) != NULL) {
s->print(ctx, "%s=%s\n", e->key, e->value);
}
} else if (e && e->value && (!s->value || (e->value && s->compare(s, e->value, s->value)))) {
s->print(ctx, "frame:%-4"PRId64" pts:%-7s pts_time:%s\n",
- inlink->frame_count_out, av_ts2str(frame->pts), av_ts2timestr(frame->pts, &inlink->time_base));
+ inl->frame_count_out, av_ts2str(frame->pts), av_ts2timestr(frame->pts, &inlink->time_base));
s->print(ctx, "%s=%s\n", s->key, e->value);
}
return ff_filter_frame(outlink, frame);
@@ -162,6 +162,7 @@ static int current_segment_finished(AVFilterContext *ctx, AVFrame *frame)
{
SegmentContext *s = ctx->priv;
AVFilterLink *inlink = ctx->inputs[0];
+ FilterLink *inl = ff_filter_link(inlink);
int ret = 0;
if (s->use_timestamps) {
@@ -169,10 +170,10 @@ static int current_segment_finished(AVFilterContext *ctx, AVFrame *frame)
} else {
switch (inlink->type) {
case AVMEDIA_TYPE_VIDEO:
- ret = inlink->frame_count_out - 1 >= s->points[s->current_point];
+ ret = inl->frame_count_out - 1 >= s->points[s->current_point];
break;
case AVMEDIA_TYPE_AUDIO:
- ret = inlink->sample_count_out - frame->nb_samples >= s->points[s->current_point];
+ ret = inl->sample_count_out - frame->nb_samples >= s->points[s->current_point];
break;
}
}
@@ -183,6 +184,7 @@ static int current_segment_finished(AVFilterContext *ctx, AVFrame *frame)
static int activate(AVFilterContext *ctx)
{
AVFilterLink *inlink = ctx->inputs[0];
+ FilterLink *inl = ff_filter_link(inlink);
SegmentContext *s = ctx->priv;
AVFrame *frame = NULL;
int ret, status;
@@ -199,14 +201,14 @@ static int activate(AVFilterContext *ctx)
ret = ff_inlink_consume_frame(inlink, &frame);
break;
case AVMEDIA_TYPE_AUDIO:
- diff = s->points[s->current_point] - inlink->sample_count_out;
+ diff = s->points[s->current_point] - inl->sample_count_out;
while (diff <= 0) {
ff_outlink_set_status(ctx->outputs[s->current_point], AVERROR_EOF, s->last_pts);
s->current_point++;
if (s->current_point >= s->nb_points)
return AVERROR(EINVAL);
- diff = s->points[s->current_point] - inlink->sample_count_out;
+ diff = s->points[s->current_point] - inl->sample_count_out;
}
if (s->use_timestamps) {
max_samples = av_rescale_q(diff, av_make_q(1, inlink->sample_rate), inlink->time_base);
@@ -34,6 +34,7 @@
#include "libavutil/pixdesc.h"
#include "avfilter.h"
#include "audio.h"
+#include "filters.h"
#include "formats.h"
#include "internal.h"
#include "video.h"
@@ -342,6 +343,7 @@ static void select_frame(AVFilterContext *ctx, AVFrame *frame)
{
SelectContext *select = ctx->priv;
AVFilterLink *inlink = ctx->inputs[0];
+ FilterLink *inl = ff_filter_link(inlink);
double res;
if (isnan(select->var_values[VAR_START_PTS]))
@@ -349,7 +351,7 @@ static void select_frame(AVFilterContext *ctx, AVFrame *frame)
if (isnan(select->var_values[VAR_START_T]))
select->var_values[VAR_START_T] = TS2D(frame->pts) * av_q2d(inlink->time_base);
- select->var_values[VAR_N ] = inlink->frame_count_out;
+ select->var_values[VAR_N ] = inl->frame_count_out;
select->var_values[VAR_PTS] = TS2D(frame->pts);
select->var_values[VAR_T ] = TS2D(frame->pts) * av_q2d(inlink->time_base);
#if FF_API_FRAME_PKT
@@ -33,6 +33,7 @@
#include "libavutil/opt.h"
#include "libavutil/parseutils.h"
#include "avfilter.h"
+#include "filters.h"
#include "internal.h"
#include "audio.h"
#include "video.h"
@@ -488,6 +489,7 @@ static av_cold void uninit(AVFilterContext *ctx)
static int filter_frame(AVFilterLink *inlink, AVFrame *ref)
{
+ FilterLink *inl = ff_filter_link(inlink);
AVFilterContext *ctx = inlink->dst;
SendCmdContext *s = ctx->priv;
int64_t ts;
@@ -535,7 +537,7 @@ static int filter_frame(AVFilterLink *inlink, AVFrame *ref)
double end = TS2T(interval->end_ts, AV_TIME_BASE_Q);
double current = TS2T(ref->pts, inlink->time_base);
- var_values[VAR_N] = inlink->frame_count_in;
+ var_values[VAR_N] = inl->frame_count_in;
#if FF_API_FRAME_PKT
FF_DISABLE_DEPRECATION_WARNINGS
var_values[VAR_POS] = ref->pkt_pos == -1 ? NAN : ref->pkt_pos;
@@ -65,11 +65,12 @@ static int process_frame(FFFrameSync *fs)
for (j = 0; j < ctx->nb_inputs; j++) {
for (i = 0; i < s->nb_map; i++) {
+ FilterLink *outl = ff_filter_link(ctx->outputs[i]);
if (s->map[i] == j) {
AVFrame *out;
if (s->is_audio && s->last_pts[j] == in[j]->pts &&
- ctx->outputs[i]->frame_count_in > 0)
+ outl->frame_count_in > 0)
continue;
out = av_frame_clone(in[j]);
if (!out)
@@ -70,6 +70,16 @@ typedef struct FilterLink {
*/
int max_samples;
+ /**
+ * Number of past frames sent through the link.
+ */
+ int64_t frame_count_in, frame_count_out;
+
+ /**
+ * Number of past samples sent through the link.
+ */
+ int64_t sample_count_in, sample_count_out;
+
/**
* Frame rate of the stream on the link, or 1/0 if unknown or variable.
*
@@ -780,12 +780,13 @@ static int qrencode_query_formats(AVFilterContext *ctx)
static int filter_frame(AVFilterLink *inlink, AVFrame *frame)
{
+ FilterLink *inl = ff_filter_link(inlink);
AVFilterContext *ctx = inlink->dst;
AVFilterLink *outlink = ctx->outputs[0];
QREncodeContext *qr = ctx->priv;
int ret;
- V(n) = inlink->frame_count_out;
+ V(n) = inl->frame_count_out;
V(t) = frame->pts == AV_NOPTS_VALUE ?
NAN : frame->pts * av_q2d(inlink->time_base);
V(pict_type) = frame->pict_type;
@@ -28,6 +28,7 @@
#include "libavutil/timestamp.h"
#include "avfilter.h"
#include "bbox.h"
+#include "filters.h"
#include "internal.h"
typedef struct BBoxContext {
@@ -75,6 +76,7 @@ static const enum AVPixelFormat pix_fmts[] = {
static int filter_frame(AVFilterLink *inlink, AVFrame *frame)
{
+ FilterLink *inl = ff_filter_link(inlink);
AVFilterContext *ctx = inlink->dst;
BBoxContext *bbox = ctx->priv;
FFBoundingBox box;
@@ -88,7 +90,7 @@ static int filter_frame(AVFilterLink *inlink, AVFrame *frame)
h = box.y2 - box.y1 + 1;
av_log(ctx, AV_LOG_INFO,
- "n:%"PRId64" pts:%s pts_time:%s", inlink->frame_count_out,
+ "n:%"PRId64" pts:%s pts_time:%s", inl->frame_count_out,
av_ts2str(frame->pts), av_ts2timestr(frame->pts, &inlink->time_base));
if (has_bbox) {
@@ -30,6 +30,7 @@
#include "libavutil/pixdesc.h"
#include "libavutil/timestamp.h"
#include "avfilter.h"
+#include "filters.h"
#include "internal.h"
#include "video.h"
@@ -173,6 +174,7 @@ static int black_counter(AVFilterContext *ctx, void *arg,
static int filter_frame(AVFilterLink *inlink, AVFrame *picref)
{
+ FilterLink *inl = ff_filter_link(inlink);
AVFilterContext *ctx = inlink->dst;
BlackDetectContext *s = ctx->priv;
double picture_black_ratio = 0;
@@ -195,7 +197,7 @@ static int filter_frame(AVFilterLink *inlink, AVFrame *picref)
av_log(ctx, AV_LOG_DEBUG,
"frame:%"PRId64" picture_black_ratio:%f pts:%s t:%s type:%c\n",
- inlink->frame_count_out, picture_black_ratio,
+ inl->frame_count_out, picture_black_ratio,
av_ts2str(picref->pts), av_ts2timestr(picref->pts, &s->time_base),
av_get_picture_type_char(picref->pict_type));
@@ -174,10 +174,11 @@ static int filter_slice(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs)
const uint8_t *top = td->top->data[td->plane];
const uint8_t *bottom = td->bottom->data[td->plane];
uint8_t *dst = td->dst->data[td->plane];
+ FilterLink *inl = ff_filter_link(td->inlink);
double values[VAR_VARS_NB];
SliceParams sliceparam = {.values = &values[0], .starty = slice_start, .e = td->param->e ? td->param->e[jobnr] : NULL};
- values[VAR_N] = td->inlink->frame_count_out;
+ values[VAR_N] = inl->frame_count_out;
values[VAR_T] = td->dst->pts == AV_NOPTS_VALUE ? NAN : td->dst->pts * av_q2d(td->inlink->time_base);
values[VAR_W] = td->w;
values[VAR_H] = td->h;
@@ -32,6 +32,8 @@
#include "libavutil/mem.h"
#include "libavutil/opt.h"
#include "libavutil/pixdesc.h"
+
+#include "filters.h"
#include "internal.h"
#include "video.h"
@@ -198,6 +200,7 @@ static void set_meta(AVDictionary **metadata, const char *key, float d)
static int blockdetect_filter_frame(AVFilterLink *inlink, AVFrame *in)
{
+ FilterLink *inl = ff_filter_link(inlink);
AVFilterContext *ctx = inlink->dst;
BLKContext *s = ctx->priv;
AVFilterLink *outlink = ctx->outputs[0];
@@ -236,7 +239,7 @@ static int blockdetect_filter_frame(AVFilterLink *inlink, AVFrame *in)
set_meta(metadata, "lavfi.block", block);
- s->nb_frames = inlink->frame_count_in;
+ s->nb_frames = inl->frame_count_in;
return ff_filter_frame(outlink, in);
}
@@ -34,6 +34,8 @@
#include "libavutil/opt.h"
#include "libavutil/pixdesc.h"
#include "libavutil/qsort.h"
+
+#include "filters.h"
#include "internal.h"
#include "edge_common.h"
#include "video.h"
@@ -256,6 +258,7 @@ static void set_meta(AVDictionary **metadata, const char *key, float d)
static int blurdetect_filter_frame(AVFilterLink *inlink, AVFrame *in)
{
+ FilterLink *inl = ff_filter_link(inlink);
AVFilterContext *ctx = inlink->dst;
BLRContext *s = ctx->priv;
AVFilterLink *outlink = ctx->outputs[0];
@@ -316,7 +319,7 @@ static int blurdetect_filter_frame(AVFilterLink *inlink, AVFrame *in)
set_meta(metadata, "lavfi.blur", blur);
- s->nb_frames = inlink->frame_count_in;
+ s->nb_frames = inl->frame_count_in;
return ff_filter_frame(outlink, in);
}
@@ -26,6 +26,7 @@
#include <stdio.h>
#include "avfilter.h"
+#include "filters.h"
#include "formats.h"
#include "internal.h"
#include "video.h"
@@ -255,12 +256,13 @@ static int config_output(AVFilterLink *link)
static int filter_frame(AVFilterLink *link, AVFrame *frame)
{
+ FilterLink *l = ff_filter_link(link);
AVFilterContext *ctx = link->dst;
CropContext *s = ctx->priv;
const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(link->format);
int i;
- s->var_values[VAR_N] = link->frame_count_out;
+ s->var_values[VAR_N] = l->frame_count_out;
s->var_values[VAR_T] = frame->pts == AV_NOPTS_VALUE ?
NAN : frame->pts * av_q2d(link->time_base);
#if FF_API_FRAME_PKT
@@ -25,6 +25,7 @@
#include "libavutil/xga_font_data.h"
#include "avfilter.h"
#include "drawutils.h"
+#include "filters.h"
#include "formats.h"
#include "internal.h"
#include "video.h"
@@ -1038,6 +1039,7 @@ static void draw_scope(OscilloscopeContext *s, int x0, int y0, int x1, int y1,
static int oscilloscope_filter_frame(AVFilterLink *inlink, AVFrame *frame)
{
+ FilterLink *inl = ff_filter_link(inlink);
AVFilterContext *ctx = inlink->dst;
OscilloscopeContext *s = ctx->priv;
AVFilterLink *outlink = ctx->outputs[0];
@@ -1047,7 +1049,7 @@ static int oscilloscope_filter_frame(AVFilterLink *inlink, AVFrame *frame)
int i, c;
s->nb_values = 0;
- draw_scope(s, s->x1, s->y1, s->x2, s->y2, frame, s->values, inlink->frame_count_in & 1);
+ draw_scope(s, s->x1, s->y1, s->x2, s->y2, frame, s->values, inl->frame_count_in & 1);
ff_blend_rectangle(&s->draw, &s->dark, frame->data, frame->linesize,
frame->width, frame->height,
s->ox, s->oy, s->width, s->height + 20 * s->statistics);
@@ -33,6 +33,7 @@
#include "libavutil/pixdesc.h"
#include "libavutil/eval.h"
#include "avfilter.h"
+#include "filters.h"
#include "internal.h"
#include "video.h"
static const char * const var_names[] = {
@@ -286,6 +287,7 @@ static int config_input(AVFilterLink *inlink)
static int filter_frame(AVFilterLink *inlink, AVFrame *in)
{
+ FilterLink *inl = ff_filter_link(inlink);
DelogoContext *s = inlink->dst->priv;
AVFilterLink *outlink = inlink->dst->outputs[0];
const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(inlink->format);
@@ -297,7 +299,7 @@ static int filter_frame(AVFilterLink *inlink, AVFrame *in)
AVRational sar;
int ret;
- s->var_values[VAR_N] = inlink->frame_count_out;
+ s->var_values[VAR_N] = inl->frame_count_out;
s->var_values[VAR_T] = TS2T(in->pts, inlink->time_base);
s->x = av_expr_eval(s->x_pexpr, s->var_values, s);
s->y = av_expr_eval(s->y_pexpr, s->var_values, s);
@@ -193,6 +193,7 @@ static int filter_frame(AVFilterLink *inlink, AVFrame *inpicref)
{
AVFilterContext *ctx = inlink->dst;
AVFilterLink *outlink = ctx->outputs[0];
+ FilterLink *outl = ff_filter_link(outlink);
DetelecineContext *s = ctx->priv;
int i, len = 0, ret = 0, out = 0;
@@ -332,7 +333,7 @@ static int filter_frame(AVFilterLink *inlink, AVFrame *inpicref)
av_frame_copy_props(frame, inpicref);
frame->pts = ((s->start_time == AV_NOPTS_VALUE) ? 0 : s->start_time) +
- av_rescale(outlink->frame_count_in, s->ts_unit.num,
+ av_rescale(outl->frame_count_in, s->ts_unit.num,
s->ts_unit.den);
ret = ff_filter_frame(outlink, frame);
}
@@ -1553,6 +1553,7 @@ static int draw_text(AVFilterContext *ctx, AVFrame *frame)
{
DrawTextContext *s = ctx->priv;
AVFilterLink *inlink = ctx->inputs[0];
+ FilterLink *inl = ff_filter_link(inlink);
int x = 0, y = 0, ret;
int shift_x64, shift_y64;
int x64, y64;
@@ -1596,7 +1597,7 @@ static int draw_text(AVFilterContext *ctx, AVFrame *frame)
if (s->tc_opt_string) {
char tcbuf[AV_TIMECODE_STR_SIZE];
- av_timecode_make_string(&s->tc, tcbuf, inlink->frame_count_out);
+ av_timecode_make_string(&s->tc, tcbuf, inl->frame_count_out);
av_bprint_clear(bp);
av_bprintf(bp, "%s%s", s->text, tcbuf);
}
@@ -1828,6 +1829,7 @@ static int draw_text(AVFilterContext *ctx, AVFrame *frame)
static int filter_frame(AVFilterLink *inlink, AVFrame *frame)
{
+ FilterLink *inl = ff_filter_link(inlink);
AVFilterContext *ctx = inlink->dst;
AVFilterLink *outlink = ctx->outputs[0];
DrawTextContext *s = ctx->priv;
@@ -1848,7 +1850,7 @@ static int filter_frame(AVFilterLink *inlink, AVFrame *frame)
}
}
- if (s->reload && !(inlink->frame_count_out % s->reload)) {
+ if (s->reload && !(inl->frame_count_out % s->reload)) {
if ((ret = ff_load_textfile(ctx, (const char *)s->textfile, &s->text, NULL)) < 0) {
av_frame_free(&frame);
return ret;
@@ -1862,7 +1864,7 @@ static int filter_frame(AVFilterLink *inlink, AVFrame *frame)
#endif
}
- s->var_values[VAR_N] = inlink->frame_count_out + s->start_number;
+ s->var_values[VAR_N] = inl->frame_count_out + s->start_number;
s->var_values[VAR_T] = frame->pts == AV_NOPTS_VALUE ?
NAN : frame->pts * av_q2d(inlink->time_base);
@@ -221,6 +221,7 @@ static const enum AVPixelFormat pixel_fmts_eq[] = {
static int filter_frame(AVFilterLink *inlink, AVFrame *in)
{
+ FilterLink *inl = ff_filter_link(inlink);
AVFilterContext *ctx = inlink->dst;
AVFilterLink *outlink = inlink->dst->outputs[0];
EQContext *eq = ctx->priv;
@@ -237,7 +238,7 @@ static int filter_frame(AVFilterLink *inlink, AVFrame *in)
av_frame_copy_props(out, in);
desc = av_pix_fmt_desc_get(inlink->format);
- eq->var_values[VAR_N] = inlink->frame_count_out;
+ eq->var_values[VAR_N] = inl->frame_count_out;
#if FF_API_FRAME_PKT
FF_DISABLE_DEPRECATION_WARNINGS
{
@@ -33,6 +33,7 @@
#include "libavutil/pixdesc.h"
#include "avfilter.h"
#include "drawutils.h"
+#include "filters.h"
#include "formats.h"
#include "internal.h"
#include "video.h"
@@ -442,6 +443,7 @@ static int config_input(AVFilterLink *inlink)
static int filter_frame(AVFilterLink *inlink, AVFrame *frame)
{
+ FilterLink *inl = ff_filter_link(inlink);
AVFilterContext *ctx = inlink->dst;
FadeContext *s = ctx->priv;
@@ -449,7 +451,7 @@ static int filter_frame(AVFilterLink *inlink, AVFrame *frame)
if (s->fade_state == VF_FADE_WAITING) {
s->factor=0;
if (frame->pts >= s->start_time_pts
- && inlink->frame_count_out >= s->start_frame) {
+ && inl->frame_count_out >= s->start_frame) {
// Time to start fading
s->fade_state = VF_FADE_FADING;
@@ -460,15 +462,15 @@ static int filter_frame(AVFilterLink *inlink, AVFrame *frame)
// Save start frame in case we are starting based on time and fading based on frames
if (s->start_time_pts != 0 && s->start_frame == 0) {
- s->start_frame = inlink->frame_count_out;
+ s->start_frame = inl->frame_count_out;
}
}
}
if (s->fade_state == VF_FADE_FADING) {
if (s->duration_pts == 0) {
// Fading based on frame count
- s->factor = (inlink->frame_count_out - s->start_frame) * s->fade_per_frame;
- if (inlink->frame_count_out > s->start_frame + s->nb_frames) {
+ s->factor = (inl->frame_count_out - s->start_frame) * s->fade_per_frame;
+ if (inl->frame_count_out > s->start_frame + s->nb_frames) {
s->fade_state = VF_FADE_DONE;
}
@@ -24,6 +24,7 @@
* FFT domain filtering.
*/
+#include "filters.h"
#include "internal.h"
#include "video.h"
#include "libavutil/common.h"
@@ -284,10 +285,11 @@ static av_cold int initialize(AVFilterContext *ctx)
static void do_eval(FFTFILTContext *s, AVFilterLink *inlink, int plane)
{
+ FilterLink *l = ff_filter_link(inlink);
double values[VAR_VARS_NB];
int i, j;
- values[VAR_N] = inlink->frame_count_out;
+ values[VAR_N] = l->frame_count_out;
values[VAR_W] = s->planewidth[plane];
values[VAR_H] = s->planeheight[plane];
values[VAR_WS] = s->rdft_hlen[plane];
@@ -25,6 +25,7 @@
#include "libavutil/opt.h"
#include "libavutil/pixdesc.h"
#include "avfilter.h"
+#include "filters.h"
#include "formats.h"
#include "internal.h"
#include "video.h"
@@ -113,8 +114,10 @@ static int config_input(AVFilterLink *inlink)
static int filter_frame(AVFilterLink *inlink, AVFrame *in)
{
+ FilterLink *inl = ff_filter_link(inlink);
AVFilterContext *ctx = inlink->dst;
AVFilterLink *outlink = ctx->outputs[0];
+ FilterLink *outl = ff_filter_link(outlink);
FieldHintContext *s = ctx->priv;
AVFrame *out, *top, *bottom;
char buf[1024] = { 0 };
@@ -152,9 +155,9 @@ static int filter_frame(AVFilterLink *inlink, AVFrame *in)
}
switch (s->mode) {
case ABSOLUTE_HINT:
- if (tf > outlink->frame_count_in + 1 || tf < FFMAX(0, outlink->frame_count_in - 1) ||
- bf > outlink->frame_count_in + 1 || bf < FFMAX(0, outlink->frame_count_in - 1)) {
- av_log(ctx, AV_LOG_ERROR, "Out of range frames %"PRId64" and/or %"PRId64" on line %"PRId64" for %"PRId64". input frame.\n", tf, bf, s->line, inlink->frame_count_out);
+ if (tf > outl->frame_count_in + 1 || tf < FFMAX(0, outl->frame_count_in - 1) ||
+ bf > outl->frame_count_in + 1 || bf < FFMAX(0, outl->frame_count_in - 1)) {
+ av_log(ctx, AV_LOG_ERROR, "Out of range frames %"PRId64" and/or %"PRId64" on line %"PRId64" for %"PRId64". input frame.\n", tf, bf, s->line, inl->frame_count_out);
return AVERROR_INVALIDDATA;
}
break;
@@ -162,7 +165,7 @@ static int filter_frame(AVFilterLink *inlink, AVFrame *in)
case RELATIVE_HINT:
if (tf > 1 || tf < -1 ||
bf > 1 || bf < -1) {
- av_log(ctx, AV_LOG_ERROR, "Out of range %"PRId64" and/or %"PRId64" on line %"PRId64" for %"PRId64". input frame.\n", tf, bf, s->line, inlink->frame_count_out);
+ av_log(ctx, AV_LOG_ERROR, "Out of range %"PRId64" and/or %"PRId64" on line %"PRId64" for %"PRId64". input frame.\n", tf, bf, s->line, inl->frame_count_out);
return AVERROR_INVALIDDATA;
}
break;
@@ -175,7 +178,7 @@ static int filter_frame(AVFilterLink *inlink, AVFrame *in)
fseek(s->hint, 0, SEEK_SET);
continue;
}
- av_log(ctx, AV_LOG_ERROR, "Missing entry for %"PRId64". input frame.\n", inlink->frame_count_out);
+ av_log(ctx, AV_LOG_ERROR, "Missing entry for %"PRId64". input frame.\n", inl->frame_count_out);
return AVERROR_INVALIDDATA;
}
}
@@ -187,8 +190,8 @@ static int filter_frame(AVFilterLink *inlink, AVFrame *in)
switch (s->mode) {
case ABSOLUTE_HINT:
- top = s->frame[tf - outlink->frame_count_in + 1];
- bottom = s->frame[bf - outlink->frame_count_in + 1];
+ top = s->frame[tf - outl->frame_count_in + 1];
+ bottom = s->frame[bf - outl->frame_count_in + 1];
break;
case PATTERN_HINT:
case RELATIVE_HINT:
@@ -681,6 +681,7 @@ static int filter_frame(AVFilterLink *inlink, AVFrame *in)
{
AVFilterContext *ctx = inlink->dst;
AVFilterLink *outlink = ctx->outputs[0];
+ FilterLink *outl = ff_filter_link(outlink);
FieldMatchContext *fm = ctx->priv;
int combs[] = { -1, -1, -1, -1, -1 };
int order, field, i, match, interlaced_frame, sc = 0, ret = 0;
@@ -753,7 +754,7 @@ static int filter_frame(AVFilterLink *inlink, AVFrame *in)
/* scene change check */
if (fm->combmatch == COMBMATCH_SC) {
- if (fm->lastn == outlink->frame_count_in - 1) {
+ if (fm->lastn == outl->frame_count_in - 1) {
if (fm->lastscdiff > fm->scthresh)
sc = 1;
} else if (luma_abs_diff(fm->prv, fm->src) > fm->scthresh) {
@@ -761,7 +762,7 @@ static int filter_frame(AVFilterLink *inlink, AVFrame *in)
}
if (!sc) {
- fm->lastn = outlink->frame_count_in;
+ fm->lastn = outl->frame_count_in;
fm->lastscdiff = luma_abs_diff(fm->src, fm->nxt);
sc = fm->lastscdiff > fm->scthresh;
}
@@ -831,7 +832,7 @@ FF_ENABLE_DEPRECATION_WARNINGS
if (interlaced_frame) {
dst->flags |= AV_FRAME_FLAG_INTERLACED;
av_log(ctx, AV_LOG_WARNING, "Frame #%"PRId64" at %s is still interlaced\n",
- outlink->frame_count_in, av_ts2timestr(in->pts, &inlink->time_base));
+ outl->frame_count_in, av_ts2timestr(in->pts, &inlink->time_base));
#if FF_API_INTERLACED_FRAME
FF_DISABLE_DEPRECATION_WARNINGS
dst->top_field_first = field;
@@ -24,6 +24,8 @@
#include "libavutil/mem.h"
#include "libavutil/opt.h"
+
+#include "filters.h"
#include "internal.h"
#include "video.h"
@@ -172,6 +174,7 @@ static float search(FOCContext *foc, int pass, int maxpass, int xmin, int xmax,
static int filter_frame(AVFilterLink *inlink, AVFrame *in)
{
+ FilterLink *inl = ff_filter_link(inlink);
AVFilterContext *ctx = inlink->dst;
FOCContext *foc = ctx->priv;
float best_score;
@@ -208,7 +211,7 @@ static int filter_frame(AVFilterLink *inlink, AVFrame *in)
}
av_log(ctx, AV_LOG_INFO, "Found at n=%"PRId64" pts_time=%f x=%d y=%d with score=%f\n",
- inlink->frame_count_out, TS2D(in->pts) * av_q2d(inlink->time_base),
+ inl->frame_count_out, TS2D(in->pts) * av_q2d(inlink->time_base),
best_x, best_y, best_score);
foc->last_x = best_x;
foc->last_y = best_y;
@@ -64,9 +64,10 @@ static int config_output_props(AVFilterLink *outlink)
static int filter_frame(AVFilterLink *inlink, AVFrame *ref)
{
+ FilterLink *inl = ff_filter_link(inlink);
FrameStepContext *framestep = inlink->dst->priv;
- if (!(inlink->frame_count_out % framestep->frame_step)) {
+ if (!(inl->frame_count_out % framestep->frame_step)) {
return ff_filter_frame(inlink->dst->outputs[0], ref);
} else {
av_frame_free(&ref);
@@ -74,12 +74,14 @@ static int config_output(AVFilterLink *outlink)
static int activate(AVFilterContext *ctx)
{
+ FilterLink *inl0 = ff_filter_link(ctx->inputs[0]);
+ FilterLink *inl1 = ff_filter_link(ctx->inputs[1]);
AVFilterLink *outlink = ctx->outputs[0];
FreezeFramesContext *s = ctx->priv;
AVFrame *frame = NULL;
- int drop = ctx->inputs[0]->frame_count_out >= s->first &&
- ctx->inputs[0]->frame_count_out <= s->last;
- int replace = ctx->inputs[1]->frame_count_out == s->replace;
+ int drop = inl0->frame_count_out >= s->first &&
+ inl0->frame_count_out <= s->last;
+ int replace = inl1->frame_count_out == s->replace;
int ret;
FF_FILTER_FORWARD_STATUS_BACK_ALL(outlink, ctx);
@@ -32,6 +32,8 @@
#include "libavutil/mem.h"
#include "libavutil/opt.h"
#include "libavutil/pixdesc.h"
+
+#include "filters.h"
#include "formats.h"
#include "internal.h"
#include "video.h"
@@ -453,13 +455,14 @@ static int slice_geq_filter(AVFilterContext *ctx, void *arg, int jobnr, int nb_j
static int geq_filter_frame(AVFilterLink *inlink, AVFrame *in)
{
int plane;
+ FilterLink *inl = ff_filter_link(inlink);
AVFilterContext *ctx = inlink->dst;
const int nb_threads = FFMIN(MAX_NB_THREADS, ff_filter_get_nb_threads(ctx));
GEQContext *geq = ctx->priv;
AVFilterLink *outlink = inlink->dst->outputs[0];
AVFrame *out;
- geq->values[VAR_N] = inlink->frame_count_out,
+ geq->values[VAR_N] = inl->frame_count_out,
geq->values[VAR_T] = in->pts == AV_NOPTS_VALUE ? NAN : in->pts * av_q2d(inlink->time_base),
geq->picref = in;
@@ -360,6 +360,7 @@ static void apply_lut10(HueContext *s,
static int filter_frame(AVFilterLink *inlink, AVFrame *inpic)
{
+ FilterLink *inl = ff_filter_link(inlink);
HueContext *hue = inlink->dst->priv;
AVFilterLink *outlink = inlink->dst->outputs[0];
AVFrame *outpic;
@@ -381,7 +382,7 @@ static int filter_frame(AVFilterLink *inlink, AVFrame *inpic)
av_frame_copy_props(outpic, inpic);
}
- hue->var_values[VAR_N] = inlink->frame_count_out;
+ hue->var_values[VAR_N] = inl->frame_count_out;
hue->var_values[VAR_T] = TS2T(inpic->pts, inlink->time_base);
hue->var_values[VAR_PTS] = TS2D(inpic->pts);
@@ -742,6 +742,7 @@ static const AVFrame *ref_frame(const struct pl_frame_mix *mix)
static void update_crops(AVFilterContext *ctx, LibplaceboInput *in,
struct pl_frame *target, double target_pts)
{
+ FilterLink *outl = ff_filter_link(ctx->outputs[0]);
LibplaceboContext *s = ctx->priv;
const AVFrame *ref = ref_frame(&in->mix);
@@ -761,7 +762,7 @@ static void update_crops(AVFilterContext *ctx, LibplaceboInput *in,
av_q2d(in->link->sample_aspect_ratio) : 1.0;
s->var_values[VAR_IN_T] = s->var_values[VAR_T] = image_pts;
s->var_values[VAR_OUT_T] = s->var_values[VAR_OT] = target_pts;
- s->var_values[VAR_N] = ctx->outputs[0]->frame_count_out;
+ s->var_values[VAR_N] = outl->frame_count_out;
/* Clear these explicitly to avoid leaking previous frames' state */
s->var_values[VAR_CROP_W] = s->var_values[VAR_CW] = NAN;
@@ -1000,6 +1001,7 @@ static int libplacebo_activate(AVFilterContext *ctx)
int ret, ok = 0, retry = 0;
LibplaceboContext *s = ctx->priv;
AVFilterLink *outlink = ctx->outputs[0];
+ FilterLink *outl = ff_filter_link(outlink);
int64_t pts, out_pts;
FF_FILTER_FORWARD_STATUS_BACK_ALL(outlink, ctx);
@@ -1012,7 +1014,7 @@ static int libplacebo_activate(AVFilterContext *ctx)
if (ff_outlink_frame_wanted(outlink)) {
if (s->fps.num) {
- out_pts = outlink->frame_count_out;
+ out_pts = outl->frame_count_out;
} else {
/* Determine the PTS of the next frame from any active input */
out_pts = INT64_MAX;
@@ -36,6 +36,7 @@
#include "libavutil/opt.h"
#include "libavutil/timestamp.h"
#include "internal.h"
+#include "filters.h"
#include "drawutils.h"
#include "framesync.h"
#include "video.h"
@@ -881,6 +882,7 @@ static int do_blend(FFFrameSync *fs)
AVFrame *mainpic, *second;
OverlayContext *s = ctx->priv;
AVFilterLink *inlink = ctx->inputs[0];
+ FilterLink *inl = ff_filter_link(inlink);
int ret;
ret = ff_framesync_dualinput_get_writable(fs, &mainpic, &second);
@@ -891,7 +893,7 @@ static int do_blend(FFFrameSync *fs)
if (s->eval_mode == EVAL_MODE_FRAME) {
- s->var_values[VAR_N] = inlink->frame_count_out;
+ s->var_values[VAR_N] = inl->frame_count_out;
s->var_values[VAR_T] = mainpic->pts == AV_NOPTS_VALUE ?
NAN : mainpic->pts * av_q2d(inlink->time_base);
#if FF_API_FRAME_PKT
@@ -236,6 +236,7 @@ static int overlay_cuda_blend(FFFrameSync *fs)
OverlayCUDAContext *ctx = avctx->priv;
AVFilterLink *outlink = avctx->outputs[0];
AVFilterLink *inlink = avctx->inputs[0];
+ FilterLink *inl = ff_filter_link(inlink);
CudaFunctions *cu = ctx->hwctx->internal->cuda_dl;
CUcontext dummy, cuda_ctx = ctx->hwctx->cuda_ctx;
@@ -270,7 +271,7 @@ static int overlay_cuda_blend(FFFrameSync *fs)
}
if (ctx->eval_mode == EVAL_MODE_FRAME) {
- ctx->var_values[VAR_N] = inlink->frame_count_out;
+ ctx->var_values[VAR_N] = inl->frame_count_out;
ctx->var_values[VAR_T] = input_main->pts == AV_NOPTS_VALUE ?
NAN : input_main->pts * av_q2d(inlink->time_base);
@@ -26,6 +26,7 @@
#include "libavutil/pixdesc.h"
#include "libavutil/opt.h"
#include "avfilter.h"
+#include "filters.h"
#include "internal.h"
#include "video.h"
@@ -122,13 +123,15 @@ enum { VAR_W, VAR_H, VAR_IN, VAR_ON, VAR_VARS_
static int calc_persp_luts(AVFilterContext *ctx, AVFilterLink *inlink)
{
+ FilterLink *inl = ff_filter_link(inlink);
PerspectiveContext *s = ctx->priv;
AVFilterLink *outlink = ctx->outputs[0];
+ FilterLink *outl = ff_filter_link(outlink);
double (*ref)[2] = s->ref;
double values[VAR_VARS_NB] = { [VAR_W] = inlink->w, [VAR_H] = inlink->h,
- [VAR_IN] = inlink->frame_count_out + 1,
- [VAR_ON] = outlink->frame_count_in + 1 };
+ [VAR_IN] = inl->frame_count_out + 1,
+ [VAR_ON] = outl->frame_count_in + 1 };
const int h = values[VAR_H];
const int w = values[VAR_W];
double x0, x1, x2, x3, x4, x5, x6, x7, x8, q;
@@ -89,6 +89,7 @@ static int query_formats(AVFilterContext *ctx)
static int filter_frame(AVFilterLink *inlink, AVFrame *frame)
{
+ FilterLink *inl = ff_filter_link(inlink);
AVFilterContext *ctx = inlink->dst;
AVFilterLink *outlink = ctx->outputs[0];
QuircContext *quirc = ctx->priv;
@@ -104,7 +105,7 @@ static int filter_frame(AVFilterLink *inlink, AVFrame *frame)
codes_count = quirc_count(quirc->quirc);
av_log(ctx, AV_LOG_VERBOSE,
- "Found count %d codes in image #%ld\n", codes_count, inlink->frame_count_out);
+ "Found count %d codes in image #%ld\n", codes_count, inl->frame_count_out);
if (codes_count) {
int i, j;
@@ -33,6 +33,7 @@
#include "avfilter.h"
#include "drawutils.h"
+#include "filters.h"
#include "internal.h"
#include "video.h"
@@ -501,6 +502,7 @@ static int filter_slice(AVFilterContext *ctx, void *arg, int job, int nb_jobs)
static int filter_frame(AVFilterLink *inlink, AVFrame *in)
{
+ FilterLink *inl = ff_filter_link(inlink);
AVFilterContext *ctx = inlink->dst;
AVFilterLink *outlink = ctx->outputs[0];
AVFrame *out;
@@ -515,7 +517,7 @@ static int filter_frame(AVFilterLink *inlink, AVFrame *in)
}
av_frame_copy_props(out, in);
- rot->var_values[VAR_N] = inlink->frame_count_out;
+ rot->var_values[VAR_N] = inl->frame_count_out;
rot->var_values[VAR_T] = TS2T(in->pts, inlink->time_base);
rot->angle = res = av_expr_eval(rot->angle_expr, rot->var_values, rot);
@@ -879,6 +879,7 @@ static int scale_field(ScaleContext *scale, AVFrame *dst, AVFrame *src,
static int scale_frame(AVFilterLink *link, AVFrame **frame_in,
AVFrame **frame_out)
{
+ FilterLink *inl = ff_filter_link(link);
AVFilterContext *ctx = link->dst;
ScaleContext *scale = ctx->priv;
AVFilterLink *outlink = ctx->outputs[0];
@@ -938,7 +939,7 @@ static int scale_frame(AVFilterLink *link, AVFrame **frame_in,
}
if (ctx->filter == &ff_vf_scale2ref) {
- scale->var_values[VAR_S2R_MAIN_N] = link->frame_count_out;
+ scale->var_values[VAR_S2R_MAIN_N] = inl->frame_count_out;
scale->var_values[VAR_S2R_MAIN_T] = TS2T(in->pts, link->time_base);
#if FF_API_FRAME_PKT
FF_DISABLE_DEPRECATION_WARNINGS
@@ -946,7 +947,7 @@ FF_DISABLE_DEPRECATION_WARNINGS
FF_ENABLE_DEPRECATION_WARNINGS
#endif
} else {
- scale->var_values[VAR_N] = link->frame_count_out;
+ scale->var_values[VAR_N] = inl->frame_count_out;
scale->var_values[VAR_T] = TS2T(in->pts, link->time_base);
#if FF_API_FRAME_PKT
FF_DISABLE_DEPRECATION_WARNINGS
@@ -1035,6 +1036,8 @@ static int do_scale(FFFrameSync *fs)
if (ref) {
AVFilterLink *reflink = ctx->inputs[1];
+ FilterLink *rl = ff_filter_link(reflink);
+
frame_changed = ref->width != reflink->w ||
ref->height != reflink->h ||
ref->format != reflink->format ||
@@ -1058,7 +1061,7 @@ static int do_scale(FFFrameSync *fs)
}
if (scale->eval_mode == EVAL_MODE_FRAME) {
- scale->var_values[VAR_REF_N] = reflink->frame_count_out;
+ scale->var_values[VAR_REF_N] = rl->frame_count_out;
scale->var_values[VAR_REF_T] = TS2T(ref->pts, reflink->time_base);
#if FF_API_FRAME_PKT
FF_DISABLE_DEPRECATION_WARNINGS
@@ -1097,6 +1100,7 @@ static int filter_frame(AVFilterLink *link, AVFrame *in)
static int filter_frame_ref(AVFilterLink *link, AVFrame *in)
{
+ FilterLink *l = ff_filter_link(link);
ScaleContext *scale = link->dst->priv;
AVFilterLink *outlink = link->dst->outputs[1];
int frame_changed;
@@ -1122,7 +1126,7 @@ static int filter_frame_ref(AVFilterLink *link, AVFrame *in)
}
if (scale->eval_mode == EVAL_MODE_FRAME) {
- scale->var_values[VAR_N] = link->frame_count_out;
+ scale->var_values[VAR_N] = l->frame_count_out;
scale->var_values[VAR_T] = TS2T(in->pts, link->time_base);
#if FF_API_FRAME_PKT
FF_DISABLE_DEPRECATION_WARNINGS
@@ -791,6 +791,7 @@ static int (*const nppscale_process[])(AVFilterContext *ctx, NPPScaleStageContex
static int nppscale_scale(AVFilterLink *link, AVFrame *out, AVFrame *in)
{
+ FilterLink *inl = ff_filter_link(link);
AVFilterContext *ctx = link->dst;
NPPScaleContext *s = ctx->priv;
AVFilterLink *outlink = ctx->outputs[0];
@@ -840,7 +841,7 @@ static int nppscale_scale(AVFilterLink *link, AVFrame *out, AVFrame *in)
}
if (ctx->filter == &ff_vf_scale2ref_npp) {
- s->var_values[VAR_S2R_MAIN_N] = link->frame_count_out;
+ s->var_values[VAR_S2R_MAIN_N] = inl->frame_count_out;
s->var_values[VAR_S2R_MAIN_T] = TS2T(in->pts, link->time_base);
#if FF_API_FRAME_PKT
FF_DISABLE_DEPRECATION_WARNINGS
@@ -848,7 +849,7 @@ FF_DISABLE_DEPRECATION_WARNINGS
FF_ENABLE_DEPRECATION_WARNINGS
#endif
} else {
- s->var_values[VAR_N] = link->frame_count_out;
+ s->var_values[VAR_N] = inl->frame_count_out;
s->var_values[VAR_T] = TS2T(in->pts, link->time_base);
#if FF_API_FRAME_PKT
FF_DISABLE_DEPRECATION_WARNINGS
@@ -947,6 +948,7 @@ fail:
static int nppscale_filter_frame_ref(AVFilterLink *link, AVFrame *in)
{
+ FilterLink *inl = ff_filter_link(link);
NPPScaleContext *scale = link->dst->priv;
AVFilterLink *outlink = link->dst->outputs[1];
int frame_changed;
@@ -968,7 +970,7 @@ static int nppscale_filter_frame_ref(AVFilterLink *link, AVFrame *in)
}
if (scale->eval_mode == EVAL_MODE_FRAME) {
- scale->var_values[VAR_N] = link->frame_count_out;
+ scale->var_values[VAR_N] = inl->frame_count_out;
scale->var_values[VAR_T] = TS2T(in->pts, link->time_base);
#if FF_API_FRAME_PKT
FF_DISABLE_DEPRECATION_WARNINGS
@@ -728,6 +728,7 @@ static void update_sample_stats(int depth, int be, const uint8_t *src, int len,
static int filter_frame(AVFilterLink *inlink, AVFrame *frame)
{
+ FilterLink *inl = ff_filter_link(inlink);
AVFilterContext *ctx = inlink->dst;
ShowInfoContext *s = ctx->priv;
const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(inlink->format);
@@ -761,7 +762,7 @@ static int filter_frame(AVFilterLink *inlink, AVFrame *frame)
"n:%4"PRId64" pts:%7s pts_time:%-7s duration:%7"PRId64
" duration_time:%-7s "
"fmt:%s cl:%s sar:%d/%d s:%dx%d i:%c iskey:%d type:%c ",
- inlink->frame_count_out,
+ inl->frame_count_out,
av_ts2str(frame->pts), av_ts2timestr(frame->pts, &inlink->time_base),
frame->duration, av_ts2timestr(frame->duration, &inlink->time_base),
desc->name, av_chroma_location_name(frame->chroma_location),
@@ -25,6 +25,7 @@
#include "libavutil/opt.h"
#include "avfilter.h"
+#include "filters.h"
#include "formats.h"
#include "internal.h"
#include "video.h"
@@ -78,6 +79,7 @@ enum { VAR_W, VAR_H, VAR_A, VAR_N, VAR_T,
static int filter_frame(AVFilterLink *inlink, AVFrame *in)
{
+ FilterLink *inl = ff_filter_link(inlink);
AVFilterContext *ctx = inlink->dst;
AVFilterLink *outlink = ctx->outputs[0];
SwapRectContext *s = ctx->priv;
@@ -97,7 +99,7 @@ static int filter_frame(AVFilterLink *inlink, AVFrame *in)
var_values[VAR_A] = (float) inlink->w / inlink->h;
var_values[VAR_SAR] = inlink->sample_aspect_ratio.num ? av_q2d(inlink->sample_aspect_ratio) : 1;
var_values[VAR_DAR] = var_values[VAR_A] * var_values[VAR_SAR];
- var_values[VAR_N] = inlink->frame_count_out;
+ var_values[VAR_N] = inl->frame_count_out;
var_values[VAR_T] = in->pts == AV_NOPTS_VALUE ? NAN : in->pts * av_q2d(inlink->time_base);
#if FF_API_FRAME_PKT
FF_DISABLE_DEPRECATION_WARNINGS
@@ -167,6 +167,7 @@ static int filter_frame(AVFilterLink *inlink, AVFrame *inpicref)
{
AVFilterContext *ctx = inlink->dst;
AVFilterLink *outlink = ctx->outputs[0];
+ FilterLink *outl = ff_filter_link(outlink);
TelecineContext *s = ctx->priv;
int i, len, ret = 0, nout = 0;
@@ -281,7 +282,7 @@ FF_ENABLE_DEPRECATION_WARNINGS
else
frame->flags &= ~AV_FRAME_FLAG_TOP_FIELD_FIRST;
frame->pts = ((s->start_time == AV_NOPTS_VALUE) ? 0 : s->start_time) +
- av_rescale(outlink->frame_count_in, s->ts_unit.num,
+ av_rescale(outl->frame_count_in, s->ts_unit.num,
s->ts_unit.den);
ret = ff_filter_frame(outlink, frame);
}
@@ -376,6 +376,7 @@ void copy_picture_field(TInterlaceContext *tinterlace,
static int filter_frame(AVFilterLink *inlink, AVFrame *picref)
{
+ FilterLink *inl = ff_filter_link(inlink);
AVFilterContext *ctx = inlink->dst;
AVFilterLink *outlink = ctx->outputs[0];
FilterLink *l = ff_filter_link(outlink);
@@ -418,12 +419,12 @@ FF_ENABLE_DEPRECATION_WARNINGS
copy_picture_field(tinterlace, out->data, out->linesize,
(const uint8_t **)cur->data, cur->linesize,
inlink->format, inlink->w, inlink->h,
- FIELD_UPPER_AND_LOWER, 1, tinterlace->mode == MODE_MERGEX2 ? (1 + inlink->frame_count_out) & 1 ? FIELD_LOWER : FIELD_UPPER : FIELD_UPPER, tinterlace->flags);
+ FIELD_UPPER_AND_LOWER, 1, tinterlace->mode == MODE_MERGEX2 ? (1 + inl->frame_count_out) & 1 ? FIELD_LOWER : FIELD_UPPER : FIELD_UPPER, tinterlace->flags);
/* write even frame lines into the lower field of the new frame */
copy_picture_field(tinterlace, out->data, out->linesize,
(const uint8_t **)next->data, next->linesize,
inlink->format, inlink->w, inlink->h,
- FIELD_UPPER_AND_LOWER, 1, tinterlace->mode == MODE_MERGEX2 ? (1 + inlink->frame_count_out) & 1 ? FIELD_UPPER : FIELD_LOWER : FIELD_LOWER, tinterlace->flags);
+ FIELD_UPPER_AND_LOWER, 1, tinterlace->mode == MODE_MERGEX2 ? (1 + inl->frame_count_out) & 1 ? FIELD_UPPER : FIELD_LOWER : FIELD_LOWER, tinterlace->flags);
if (tinterlace->mode != MODE_MERGEX2)
av_frame_free(&tinterlace->next);
break;
@@ -445,7 +446,7 @@ FF_ENABLE_DEPRECATION_WARNINGS
out->height = outlink->h;
out->sample_aspect_ratio = av_mul_q(cur->sample_aspect_ratio, av_make_q(2, 1));
- field = (1 + outlink->frame_count_in) & 1 ? FIELD_UPPER : FIELD_LOWER;
+ field = (1 + l->frame_count_in) & 1 ? FIELD_UPPER : FIELD_LOWER;
full = out->color_range == AVCOL_RANGE_JPEG || ff_fmt_is_in(out->format, full_scale_yuvj_pix_fmts);
/* copy upper and lower fields */
copy_picture_field(tinterlace, out->data, out->linesize,
@@ -150,12 +150,13 @@ static double get_natural_factor(const VignetteContext *s, int x, int y)
static void update_context(VignetteContext *s, AVFilterLink *inlink, AVFrame *frame)
{
+ FilterLink *inl = ff_filter_link(inlink);
int x, y;
float *dst = s->fmap;
int dst_linesize = s->fmap_linesize;
if (frame) {
- s->var_values[VAR_N] = inlink->frame_count_out;
+ s->var_values[VAR_N] = inl->frame_count_out;
s->var_values[VAR_T] = TS2T(frame->pts, inlink->time_base);
s->var_values[VAR_PTS] = TS2D(frame->pts);
} else {
@@ -100,12 +100,13 @@ typedef struct ThreadData {
static int weave_slice(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs)
{
AVFilterLink *inlink = ctx->inputs[0];
+ FilterLink *inl = ff_filter_link(inlink);
WeaveContext *s = ctx->priv;
ThreadData *td = arg;
AVFrame *in = td->in;
AVFrame *out = td->out;
- const int weave = (s->double_weave && !(inlink->frame_count_out & 1));
+ const int weave = (s->double_weave && !(inl->frame_count_out & 1));
const int field1 = weave ? s->first_field : (!s->first_field);
const int field2 = weave ? (!s->first_field) : s->first_field;
@@ -156,6 +156,7 @@ static int output_single_frame(AVFilterContext *ctx, AVFrame *in, double *var_va
{
ZPContext *s = ctx->priv;
AVFilterLink *outlink = ctx->outputs[0];
+ FilterLink *outl = ff_filter_link(outlink);
AVFilterLink *inlink = ctx->inputs[0];
int64_t pts = s->frame_count;
int k, x, y, w, h, ret = 0;
@@ -172,7 +173,7 @@ static int output_single_frame(AVFilterContext *ctx, AVFrame *in, double *var_va
var_values[VAR_OUT_TIME] = pts * av_q2d(outlink->time_base);
var_values[VAR_TIME] = var_values[VAR_OT] = var_values[VAR_OUT_TIME];
var_values[VAR_FRAME] = i;
- var_values[VAR_ON] = outlink->frame_count_in;
+ var_values[VAR_ON] = outl->frame_count_in;
*zoom = av_expr_eval(s->zoom_expr, var_values, NULL);
@@ -260,7 +261,9 @@ static int activate(AVFilterContext *ctx)
{
ZPContext *s = ctx->priv;
AVFilterLink *inlink = ctx->inputs[0];
+ FilterLink *inl = ff_filter_link(inlink);
AVFilterLink *outlink = ctx->outputs[0];
+ FilterLink *outl = ff_filter_link(outlink);
int status, ret = 0;
int64_t pts;
@@ -283,8 +286,8 @@ static int activate(AVFilterContext *ctx)
s->var_values[VAR_IN_H] = s->var_values[VAR_IH] = s->in->height;
s->var_values[VAR_OUT_W] = s->var_values[VAR_OW] = s->w;
s->var_values[VAR_OUT_H] = s->var_values[VAR_OH] = s->h;
- s->var_values[VAR_IN] = inlink->frame_count_out - 1;
- s->var_values[VAR_ON] = outlink->frame_count_in;
+ s->var_values[VAR_IN] = inl->frame_count_out - 1;
+ s->var_values[VAR_ON] = outl->frame_count_in;
s->var_values[VAR_PX] = s->x;
s->var_values[VAR_PY] = s->y;
s->var_values[VAR_X] = 0;
@@ -292,12 +292,13 @@ static int config_props(AVFilterLink *outlink)
static int request_frame(AVFilterLink *outlink)
{
+ FilterLink *outl = ff_filter_link(outlink);
MPTestContext *test = outlink->src->priv;
AVFrame *picref;
int w = WIDTH, h = HEIGHT,
cw = AV_CEIL_RSHIFT(w, test->hsub), ch = AV_CEIL_RSHIFT(h, test->vsub);
- uint64_t frame = outlink->frame_count_in / test->max_frames;
- uint64_t mod = outlink->frame_count_in % test->max_frames;
+ uint64_t frame = outl->frame_count_in / test->max_frames;
+ uint64_t mod = outl->frame_count_in % test->max_frames;
enum test_type tt = test->test;
int i;