diff mbox series

[FFmpeg-devel] avfilter: add (a)segment filters

Message ID 20210809222258.10052-1-onemda@gmail.com
State New
Headers show
Series [FFmpeg-devel] avfilter: add (a)segment filters | expand

Checks

Context Check Description
andriy/x86_make success Make finished
andriy/x86_make_fate success Make fate finished
andriy/PPC64_make success Make finished
andriy/PPC64_make_fate success Make fate finished

Commit Message

Paul B Mahol Aug. 9, 2021, 10:22 p.m. UTC
Signed-off-by: Paul B Mahol <onemda@gmail.com>
---
 doc/filters.texi         |  34 ++++
 libavfilter/Makefile     |   2 +
 libavfilter/allfilters.c |   2 +
 libavfilter/f_segment.c  | 335 +++++++++++++++++++++++++++++++++++++++
 4 files changed, 373 insertions(+)
 create mode 100644 libavfilter/f_segment.c

Comments

Nicolas George Aug. 16, 2021, 8:42 a.m. UTC | #1
Paul B Mahol (12021-08-10):
> Signed-off-by: Paul B Mahol <onemda@gmail.com>
> ---
>  doc/filters.texi         |  34 ++++
>  libavfilter/Makefile     |   2 +
>  libavfilter/allfilters.c |   2 +
>  libavfilter/f_segment.c  | 335 +++++++++++++++++++++++++++++++++++++++
>  4 files changed, 373 insertions(+)
>  create mode 100644 libavfilter/f_segment.c
> 
> diff --git a/doc/filters.texi b/doc/filters.texi
> index bdeb3fedfd..b65afbda65 100644
> --- a/doc/filters.texi
> +++ b/doc/filters.texi
> @@ -25665,6 +25665,40 @@ A processing speed faster than what is possible without these filters cannot
>  be achieved.
>  @end table
>  
> +@section segment, asegment
> +
> +Split single input stream into multiple streams.
> +
> +This filter does opposite of concat filters.
> +
> +@code{segment} works on video frames, @code{asegment} on audio samples.
> +
> +This filter accepts the following options:
> +

> +@table @option
> +@item timestamps
> +Timestamps of output segments separated by '|'. The first segment will run
> +from the beginning of the input stream. The last segment will run until
> +the end of the input stream
> +
> +@item frames, samples
> +Exact frame/sample count to split the segments.
> +@end table

Since you adopted my suggestion of + for relative parsing (thank you for
it), it needs to be documented:

"In both cases, prefixing a point with '+' will make it relative to the
previous point."

> +
> +@subsection Examples
> +
> +@itemize
> +
> +@item
> +Split input audio stream into three output audio streams, starting at start of input audio stream
> +and storing that in 1st output audio stream, then following at 60th second and storing than in 2nd
> +output audio stream, and last after 120th second of input audio stream store in 3rd output audio stream:

> +@example
> +asegment=timestamps="60 | 120"
> +@end example

I would suggest to use another number than 120, because 120-60=60 which
makes the explanation ambiguous; 150 would be a good choice.

> +
> +@end itemize
> +
>  @anchor{select}
>  @section select, aselect
>  
> diff --git a/libavfilter/Makefile b/libavfilter/Makefile
> index 49c0c8342b..102ce7beff 100644
> --- a/libavfilter/Makefile
> +++ b/libavfilter/Makefile
> @@ -78,6 +78,7 @@ OBJS-$(CONFIG_AREALTIME_FILTER)              += f_realtime.o
>  OBJS-$(CONFIG_ARESAMPLE_FILTER)              += af_aresample.o
>  OBJS-$(CONFIG_AREVERSE_FILTER)               += f_reverse.o
>  OBJS-$(CONFIG_ARNNDN_FILTER)                 += af_arnndn.o
> +OBJS-$(CONFIG_ASEGMENT_FILTER)               += f_segment.o
>  OBJS-$(CONFIG_ASELECT_FILTER)                += f_select.o
>  OBJS-$(CONFIG_ASENDCMD_FILTER)               += f_sendcmd.o
>  OBJS-$(CONFIG_ASETNSAMPLES_FILTER)           += af_asetnsamples.o
> @@ -404,6 +405,7 @@ OBJS-$(CONFIG_SCALE_VULKAN_FILTER)           += vf_scale_vulkan.o vulkan.o
>  OBJS-$(CONFIG_SCALE2REF_FILTER)              += vf_scale.o scale_eval.o
>  OBJS-$(CONFIG_SCDET_FILTER)                  += vf_scdet.o
>  OBJS-$(CONFIG_SCROLL_FILTER)                 += vf_scroll.o
> +OBJS-$(CONFIG_SEGMENT_FILTER)                += f_segment.o
>  OBJS-$(CONFIG_SELECT_FILTER)                 += f_select.o
>  OBJS-$(CONFIG_SELECTIVECOLOR_FILTER)         += vf_selectivecolor.o
>  OBJS-$(CONFIG_SENDCMD_FILTER)                += f_sendcmd.o
> diff --git a/libavfilter/allfilters.c b/libavfilter/allfilters.c
> index ae74f9c891..73040d2824 100644
> --- a/libavfilter/allfilters.c
> +++ b/libavfilter/allfilters.c
> @@ -71,6 +71,7 @@ extern const AVFilter ff_af_arealtime;
>  extern const AVFilter ff_af_aresample;
>  extern const AVFilter ff_af_areverse;
>  extern const AVFilter ff_af_arnndn;
> +extern const AVFilter ff_af_asegment;
>  extern const AVFilter ff_af_aselect;
>  extern const AVFilter ff_af_asendcmd;
>  extern const AVFilter ff_af_asetnsamples;
> @@ -385,6 +386,7 @@ extern const AVFilter ff_vf_scale_vulkan;
>  extern const AVFilter ff_vf_scale2ref;
>  extern const AVFilter ff_vf_scdet;
>  extern const AVFilter ff_vf_scroll;
> +extern const AVFilter ff_vf_segment;
>  extern const AVFilter ff_vf_select;
>  extern const AVFilter ff_vf_selectivecolor;
>  extern const AVFilter ff_vf_sendcmd;
> diff --git a/libavfilter/f_segment.c b/libavfilter/f_segment.c
> new file mode 100644
> index 0000000000..7e0521b3c1
> --- /dev/null
> +++ b/libavfilter/f_segment.c
> @@ -0,0 +1,335 @@
> +/*
> + * This file is part of FFmpeg.
> + *
> + * FFmpeg is free software; you can redistribute it and/or
> + * modify it under the terms of the GNU Lesser General Public
> + * License as published by the Free Software Foundation; either
> + * version 2.1 of the License, or (at your option) any later version.
> + *
> + * FFmpeg is distributed in the hope that it will be useful,
> + * but WITHOUT ANY WARRANTY; without even the implied warranty of
> + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
> + * Lesser General Public License for more details.
> + *
> + * You should have received a copy of the GNU Lesser General Public
> + * License along with FFmpeg; if not, write to the Free Software
> + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
> + */
> +
> +#include <stdint.h>
> +
> +#include "libavutil/avstring.h"
> +#include "libavutil/channel_layout.h"
> +#include "libavutil/common.h"
> +#include "libavutil/log.h"
> +#include "libavutil/mathematics.h"
> +#include "libavutil/opt.h"
> +#include "libavutil/parseutils.h"
> +#include "libavutil/samplefmt.h"
> +
> +#include "audio.h"
> +#include "avfilter.h"
> +#include "filters.h"
> +#include "internal.h"
> +
> +typedef struct SegmentContext {
> +    const AVClass *class;
> +
> +    char *timestamps_str;
> +    char *points_str;
> +    int use_timestamps;
> +
> +    int current_point;
> +    int nb_points;
> +
> +    int64_t *points;
> +} SegmentContext;
> +
> +static void count_points(char *item_str, int *nb_items)
> +{
> +    char *p;
> +
> +    if (!item_str)
> +        return;
> +
> +    *nb_items = 1;
> +    for (p = item_str; *p; p++) {
> +        if (*p == '|')
> +            (*nb_items)++;
> +    }
> +}
> +
> +static int parse_points(AVFilterContext *ctx, char *item_str, int nb_points, int64_t *points)
> +{
> +    SegmentContext *s = ctx->priv;
> +    char *arg, *p = item_str;
> +    char *saveptr = NULL;
> +    int64_t ref, cur = 0;
> +    int ret = 0;
> +
> +    for (int i = 0; i < nb_points; i++) {
> +        if (!(arg = av_strtok(p, "|", &saveptr)))
> +            return AVERROR(EINVAL);
> +
> +        p = NULL;
> +        ref = 0;
> +        if (*arg == '+') {
> +            ref = cur;
> +            arg++;
> +        }
> +
> +        if (s->use_timestamps) {
> +            ret = av_parse_time(&points[i], arg, s->use_timestamps);
> +        } else {
> +            if (sscanf(arg, "%"SCNd64, &points[i]) != 1)
> +                ret = AVERROR(EINVAL);
> +        }
> +
> +        if (ret < 0) {
> +            av_log(ctx, AV_LOG_ERROR, "Invalid splits supplied: %s\n", arg);

> +            return AVERROR(EINVAL);

return ret

> +        }
> +
> +        cur = points[i];
> +        points[i] += ref;
> +    }
> +
> +    return 0;
> +}
> +
> +static av_cold int init(AVFilterContext *ctx, enum AVMediaType type)
> +{
> +    SegmentContext *s = ctx->priv;
> +    char *split_str;
> +    int ret;
> +
> +    if (s->timestamps_str && s->points_str) {
> +        av_log(ctx, AV_LOG_ERROR, "Both timestamps and counts supplied.\n");
> +        return AVERROR(EINVAL);
> +    } else if (s->timestamps_str) {
> +        s->use_timestamps = 1;
> +        split_str = s->timestamps_str;
> +    } else if (s->points_str) {
> +        split_str = s->points_str;
> +    } else {
> +        av_log(ctx, AV_LOG_ERROR, "Neither timestamps nor durations nor counts supplied.\n");
> +        return AVERROR(EINVAL);
> +    }
> +
> +    count_points(split_str, &s->nb_points);
> +    s->nb_points++;
> +
> +    s->points = av_calloc(s->nb_points, sizeof(*s->points));
> +    if (!s->points)
> +        return AVERROR(ENOMEM);
> +
> +    ret = parse_points(ctx, split_str, s->nb_points - 1, s->points);
> +    if (ret < 0)
> +        return ret;
> +
> +    s->points[s->nb_points - 1] = INT64_MAX;
> +
> +    for (int i = 0; i < s->nb_points; i++) {
> +        AVFilterPad pad = { 0 };
> +
> +        pad.type = type;
> +        pad.name = av_asprintf("output%d", i);
> +        if (!pad.name)
> +            return AVERROR(ENOMEM);
> +
> +        if ((ret = ff_insert_outpad(ctx, i, &pad)) < 0) {
> +            av_freep(&pad.name);
> +            return ret;
> +        }
> +    }
> +
> +    return 0;
> +}
> +
> +static int config_input(AVFilterLink *inlink)
> +{
> +    AVFilterContext *ctx = inlink->dst;
> +    SegmentContext *s = ctx->priv;
> +    AVRational tb = inlink->time_base;
> +
> +    if (s->use_timestamps) {
> +        for (int i = 0; i < s->nb_points - 1; i++)
> +            s->points[i] = av_rescale_q(s->points[i], AV_TIME_BASE_Q, tb);
> +    }
> +
> +    return 0;
> +}
> +
> +static int current_segment_finished(AVFilterContext *ctx, AVFrame *frame)
> +{
> +    SegmentContext *s = ctx->priv;
> +    AVFilterLink *inlink = ctx->inputs[0];
> +    int ret = 0;
> +
> +    if (s->use_timestamps) {
> +        ret = frame->pts >= s->points[s->current_point];
> +    } else {
> +        switch (inlink->type) {
> +        case AVMEDIA_TYPE_VIDEO:
> +            ret = inlink->frame_count_out - 1 >= s->points[s->current_point];
> +            break;
> +        case AVMEDIA_TYPE_AUDIO:
> +            ret = inlink->sample_count_out - frame->nb_samples >= s->points[s->current_point];
> +            break;
> +        }
> +    }
> +
> +    return ret;
> +}
> +
> +static int activate(AVFilterContext *ctx)
> +{
> +    AVFilterLink *inlink = ctx->inputs[0];
> +    SegmentContext *s = ctx->priv;
> +    AVFrame *frame = NULL;
> +    int ret, status;
> +    int max_samples;
> +    int64_t diff;
> +    int64_t pts;
> +
> +    for (int i = s->current_point; i < s->nb_points; i++) {
> +        FF_FILTER_FORWARD_STATUS_BACK_ALL(ctx->outputs[i], ctx);
> +    }
> +
> +    switch (inlink->type) {
> +    case AVMEDIA_TYPE_VIDEO:
> +        ret = ff_inlink_consume_frame(inlink, &frame);
> +        break;
> +    case AVMEDIA_TYPE_AUDIO:
> +        diff = s->points[s->current_point] - inlink->sample_count_out;
> +        if (s->use_timestamps) {
> +            max_samples = av_rescale_q(diff, av_make_q(1, inlink->sample_rate), inlink->time_base);
> +        } else {
> +            max_samples = FFMAX(1, FFMIN(diff, INT_MAX));
> +        }
> +        ret = ff_inlink_consume_samples(inlink, 1, max_samples, &frame);
> +        break;
> +    }
> +

> +    if (ret > 0) {
> +        while (current_segment_finished(ctx, frame)) {
> +            ff_outlink_set_status(ctx->outputs[s->current_point], AVERROR_EOF, frame->pts);
> +            s->current_point++;
> +        }
> +
> +        if (s->current_point >= s->nb_points) {
> +            av_frame_free(&frame);
> +            return AVERROR(EINVAL);
> +        }
> +
> +        ret = ff_filter_frame(ctx->outputs[s->current_point], frame);
> +    }

I wonder if you should subtract the start of the current segment to the
timestamp of the input frame before feeding it to output.

I.e. if the input is [0;600[, output [0;60[, [0;90[, [0;450[ instead of
[0;60[, [60;150[, [150;600[.

> +
> +    if (ret < 0) {
> +        return ret;
> +    } else if (ff_inlink_acknowledge_status(inlink, &status, &pts)) {
> +        for (int i = s->current_point; i < s->nb_points; i++)
> +            ff_outlink_set_status(ctx->outputs[i], status, pts);
> +        return 0;
> +    } else {
> +        for (int i = s->current_point; i < s->nb_points; i++) {
> +            if (ff_outlink_frame_wanted(ctx->outputs[i]))
> +                ff_inlink_request_frame(inlink);
> +        }
> +        return 0;
> +    }
> +}
> +
> +static av_cold void uninit(AVFilterContext *ctx)
> +{
> +    SegmentContext *s = ctx->priv;
> +
> +    av_freep(&s->points);
> +
> +    for (unsigned i = 0; i < ctx->nb_outputs; i++)
> +        av_freep(&ctx->output_pads[i].name);
> +}
> +
> +#define OFFSET(x) offsetof(SegmentContext, x)
> +#define COMMON_OPTS \
> +    { "timestamps", "timestamps of input at which to split input", OFFSET(timestamps_str),  AV_OPT_TYPE_STRING, { .str = NULL }, 0, 0, FLAGS }, \
> +
> +#if CONFIG_SEGMENT_FILTER
> +
> +static av_cold int video_init(AVFilterContext *ctx)
> +{
> +    return init(ctx, AVMEDIA_TYPE_VIDEO);
> +}
> +
> +#define FLAGS AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_FILTERING_PARAM
> +static const AVOption segment_options[] = {
> +    COMMON_OPTS
> +    { "frames", "frames at which to split input", OFFSET(points_str), AV_OPT_TYPE_STRING,  { .str = NULL }, 0, 0, FLAGS },
> +    { NULL }
> +};
> +#undef FLAGS
> +
> +AVFILTER_DEFINE_CLASS(segment);
> +
> +static const AVFilterPad segment_inputs[] = {
> +    {
> +        .name         = "default",
> +        .type         = AVMEDIA_TYPE_VIDEO,
> +        .config_props = config_input,
> +    },
> +    { NULL }
> +};
> +
> +const AVFilter ff_vf_segment = {
> +    .name        = "segment",
> +    .description = NULL_IF_CONFIG_SMALL("Segment video stream."),
> +    .init        = video_init,
> +    .uninit      = uninit,
> +    .priv_size   = sizeof(SegmentContext),
> +    .priv_class  = &segment_class,
> +    .activate    = activate,
> +    .inputs      = segment_inputs,
> +    .outputs     = NULL,
> +    .flags       = AVFILTER_FLAG_DYNAMIC_OUTPUTS,
> +};
> +#endif // CONFIG_SEGMENT_FILTER
> +
> +#if CONFIG_ASEGMENT_FILTER
> +
> +static av_cold int audio_init(AVFilterContext *ctx)
> +{
> +    return init(ctx, AVMEDIA_TYPE_AUDIO);
> +}
> +
> +#define FLAGS AV_OPT_FLAG_AUDIO_PARAM | AV_OPT_FLAG_FILTERING_PARAM
> +static const AVOption asegment_options[] = {
> +    COMMON_OPTS
> +    { "samples", "samples at which to split input", OFFSET(points_str), AV_OPT_TYPE_STRING,  { .str = NULL }, 0, 0, FLAGS },
> +    { NULL }
> +};
> +#undef FLAGS
> +
> +AVFILTER_DEFINE_CLASS(asegment);
> +
> +static const AVFilterPad asegment_inputs[] = {
> +    {
> +        .name         = "default",
> +        .type         = AVMEDIA_TYPE_AUDIO,
> +        .config_props = config_input,
> +    },
> +    { NULL }
> +};
> +
> +const AVFilter ff_af_asegment = {
> +    .name        = "asegment",
> +    .description = NULL_IF_CONFIG_SMALL("Segment audio stream."),
> +    .init        = audio_init,
> +    .uninit      = uninit,
> +    .priv_size   = sizeof(SegmentContext),
> +    .priv_class  = &asegment_class,
> +    .activate    = activate,
> +    .inputs      = asegment_inputs,
> +    .outputs     = NULL,
> +    .flags       = AVFILTER_FLAG_DYNAMIC_OUTPUTS,
> +};
> +#endif // CONFIG_ASEGMENT_FILTER

Apart from these details, LGTM, thanks for your efforts.
diff mbox series

Patch

diff --git a/doc/filters.texi b/doc/filters.texi
index bdeb3fedfd..b65afbda65 100644
--- a/doc/filters.texi
+++ b/doc/filters.texi
@@ -25665,6 +25665,40 @@  A processing speed faster than what is possible without these filters cannot
 be achieved.
 @end table
 
+@section segment, asegment
+
+Split single input stream into multiple streams.
+
+This filter does opposite of concat filters.
+
+@code{segment} works on video frames, @code{asegment} on audio samples.
+
+This filter accepts the following options:
+
+@table @option
+@item timestamps
+Timestamps of output segments separated by '|'. The first segment will run
+from the beginning of the input stream. The last segment will run until
+the end of the input stream
+
+@item frames, samples
+Exact frame/sample count to split the segments.
+@end table
+
+@subsection Examples
+
+@itemize
+
+@item
+Split input audio stream into three output audio streams, starting at start of input audio stream
+and storing that in 1st output audio stream, then following at 60th second and storing than in 2nd
+output audio stream, and last after 120th second of input audio stream store in 3rd output audio stream:
+@example
+asegment=timestamps="60 | 120"
+@end example
+
+@end itemize
+
 @anchor{select}
 @section select, aselect
 
diff --git a/libavfilter/Makefile b/libavfilter/Makefile
index 49c0c8342b..102ce7beff 100644
--- a/libavfilter/Makefile
+++ b/libavfilter/Makefile
@@ -78,6 +78,7 @@  OBJS-$(CONFIG_AREALTIME_FILTER)              += f_realtime.o
 OBJS-$(CONFIG_ARESAMPLE_FILTER)              += af_aresample.o
 OBJS-$(CONFIG_AREVERSE_FILTER)               += f_reverse.o
 OBJS-$(CONFIG_ARNNDN_FILTER)                 += af_arnndn.o
+OBJS-$(CONFIG_ASEGMENT_FILTER)               += f_segment.o
 OBJS-$(CONFIG_ASELECT_FILTER)                += f_select.o
 OBJS-$(CONFIG_ASENDCMD_FILTER)               += f_sendcmd.o
 OBJS-$(CONFIG_ASETNSAMPLES_FILTER)           += af_asetnsamples.o
@@ -404,6 +405,7 @@  OBJS-$(CONFIG_SCALE_VULKAN_FILTER)           += vf_scale_vulkan.o vulkan.o
 OBJS-$(CONFIG_SCALE2REF_FILTER)              += vf_scale.o scale_eval.o
 OBJS-$(CONFIG_SCDET_FILTER)                  += vf_scdet.o
 OBJS-$(CONFIG_SCROLL_FILTER)                 += vf_scroll.o
+OBJS-$(CONFIG_SEGMENT_FILTER)                += f_segment.o
 OBJS-$(CONFIG_SELECT_FILTER)                 += f_select.o
 OBJS-$(CONFIG_SELECTIVECOLOR_FILTER)         += vf_selectivecolor.o
 OBJS-$(CONFIG_SENDCMD_FILTER)                += f_sendcmd.o
diff --git a/libavfilter/allfilters.c b/libavfilter/allfilters.c
index ae74f9c891..73040d2824 100644
--- a/libavfilter/allfilters.c
+++ b/libavfilter/allfilters.c
@@ -71,6 +71,7 @@  extern const AVFilter ff_af_arealtime;
 extern const AVFilter ff_af_aresample;
 extern const AVFilter ff_af_areverse;
 extern const AVFilter ff_af_arnndn;
+extern const AVFilter ff_af_asegment;
 extern const AVFilter ff_af_aselect;
 extern const AVFilter ff_af_asendcmd;
 extern const AVFilter ff_af_asetnsamples;
@@ -385,6 +386,7 @@  extern const AVFilter ff_vf_scale_vulkan;
 extern const AVFilter ff_vf_scale2ref;
 extern const AVFilter ff_vf_scdet;
 extern const AVFilter ff_vf_scroll;
+extern const AVFilter ff_vf_segment;
 extern const AVFilter ff_vf_select;
 extern const AVFilter ff_vf_selectivecolor;
 extern const AVFilter ff_vf_sendcmd;
diff --git a/libavfilter/f_segment.c b/libavfilter/f_segment.c
new file mode 100644
index 0000000000..7e0521b3c1
--- /dev/null
+++ b/libavfilter/f_segment.c
@@ -0,0 +1,335 @@ 
+/*
+ * This file is part of FFmpeg.
+ *
+ * FFmpeg is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU Lesser General Public
+ * License as published by the Free Software Foundation; either
+ * version 2.1 of the License, or (at your option) any later version.
+ *
+ * FFmpeg is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
+ * Lesser General Public License for more details.
+ *
+ * You should have received a copy of the GNU Lesser General Public
+ * License along with FFmpeg; if not, write to the Free Software
+ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
+ */
+
+#include <stdint.h>
+
+#include "libavutil/avstring.h"
+#include "libavutil/channel_layout.h"
+#include "libavutil/common.h"
+#include "libavutil/log.h"
+#include "libavutil/mathematics.h"
+#include "libavutil/opt.h"
+#include "libavutil/parseutils.h"
+#include "libavutil/samplefmt.h"
+
+#include "audio.h"
+#include "avfilter.h"
+#include "filters.h"
+#include "internal.h"
+
+typedef struct SegmentContext {
+    const AVClass *class;
+
+    char *timestamps_str;
+    char *points_str;
+    int use_timestamps;
+
+    int current_point;
+    int nb_points;
+
+    int64_t *points;
+} SegmentContext;
+
+static void count_points(char *item_str, int *nb_items)
+{
+    char *p;
+
+    if (!item_str)
+        return;
+
+    *nb_items = 1;
+    for (p = item_str; *p; p++) {
+        if (*p == '|')
+            (*nb_items)++;
+    }
+}
+
+static int parse_points(AVFilterContext *ctx, char *item_str, int nb_points, int64_t *points)
+{
+    SegmentContext *s = ctx->priv;
+    char *arg, *p = item_str;
+    char *saveptr = NULL;
+    int64_t ref, cur = 0;
+    int ret = 0;
+
+    for (int i = 0; i < nb_points; i++) {
+        if (!(arg = av_strtok(p, "|", &saveptr)))
+            return AVERROR(EINVAL);
+
+        p = NULL;
+        ref = 0;
+        if (*arg == '+') {
+            ref = cur;
+            arg++;
+        }
+
+        if (s->use_timestamps) {
+            ret = av_parse_time(&points[i], arg, s->use_timestamps);
+        } else {
+            if (sscanf(arg, "%"SCNd64, &points[i]) != 1)
+                ret = AVERROR(EINVAL);
+        }
+
+        if (ret < 0) {
+            av_log(ctx, AV_LOG_ERROR, "Invalid splits supplied: %s\n", arg);
+            return AVERROR(EINVAL);
+        }
+
+        cur = points[i];
+        points[i] += ref;
+    }
+
+    return 0;
+}
+
+static av_cold int init(AVFilterContext *ctx, enum AVMediaType type)
+{
+    SegmentContext *s = ctx->priv;
+    char *split_str;
+    int ret;
+
+    if (s->timestamps_str && s->points_str) {
+        av_log(ctx, AV_LOG_ERROR, "Both timestamps and counts supplied.\n");
+        return AVERROR(EINVAL);
+    } else if (s->timestamps_str) {
+        s->use_timestamps = 1;
+        split_str = s->timestamps_str;
+    } else if (s->points_str) {
+        split_str = s->points_str;
+    } else {
+        av_log(ctx, AV_LOG_ERROR, "Neither timestamps nor durations nor counts supplied.\n");
+        return AVERROR(EINVAL);
+    }
+
+    count_points(split_str, &s->nb_points);
+    s->nb_points++;
+
+    s->points = av_calloc(s->nb_points, sizeof(*s->points));
+    if (!s->points)
+        return AVERROR(ENOMEM);
+
+    ret = parse_points(ctx, split_str, s->nb_points - 1, s->points);
+    if (ret < 0)
+        return ret;
+
+    s->points[s->nb_points - 1] = INT64_MAX;
+
+    for (int i = 0; i < s->nb_points; i++) {
+        AVFilterPad pad = { 0 };
+
+        pad.type = type;
+        pad.name = av_asprintf("output%d", i);
+        if (!pad.name)
+            return AVERROR(ENOMEM);
+
+        if ((ret = ff_insert_outpad(ctx, i, &pad)) < 0) {
+            av_freep(&pad.name);
+            return ret;
+        }
+    }
+
+    return 0;
+}
+
+static int config_input(AVFilterLink *inlink)
+{
+    AVFilterContext *ctx = inlink->dst;
+    SegmentContext *s = ctx->priv;
+    AVRational tb = inlink->time_base;
+
+    if (s->use_timestamps) {
+        for (int i = 0; i < s->nb_points - 1; i++)
+            s->points[i] = av_rescale_q(s->points[i], AV_TIME_BASE_Q, tb);
+    }
+
+    return 0;
+}
+
+static int current_segment_finished(AVFilterContext *ctx, AVFrame *frame)
+{
+    SegmentContext *s = ctx->priv;
+    AVFilterLink *inlink = ctx->inputs[0];
+    int ret = 0;
+
+    if (s->use_timestamps) {
+        ret = frame->pts >= s->points[s->current_point];
+    } else {
+        switch (inlink->type) {
+        case AVMEDIA_TYPE_VIDEO:
+            ret = inlink->frame_count_out - 1 >= s->points[s->current_point];
+            break;
+        case AVMEDIA_TYPE_AUDIO:
+            ret = inlink->sample_count_out - frame->nb_samples >= s->points[s->current_point];
+            break;
+        }
+    }
+
+    return ret;
+}
+
+static int activate(AVFilterContext *ctx)
+{
+    AVFilterLink *inlink = ctx->inputs[0];
+    SegmentContext *s = ctx->priv;
+    AVFrame *frame = NULL;
+    int ret, status;
+    int max_samples;
+    int64_t diff;
+    int64_t pts;
+
+    for (int i = s->current_point; i < s->nb_points; i++) {
+        FF_FILTER_FORWARD_STATUS_BACK_ALL(ctx->outputs[i], ctx);
+    }
+
+    switch (inlink->type) {
+    case AVMEDIA_TYPE_VIDEO:
+        ret = ff_inlink_consume_frame(inlink, &frame);
+        break;
+    case AVMEDIA_TYPE_AUDIO:
+        diff = s->points[s->current_point] - inlink->sample_count_out;
+        if (s->use_timestamps) {
+            max_samples = av_rescale_q(diff, av_make_q(1, inlink->sample_rate), inlink->time_base);
+        } else {
+            max_samples = FFMAX(1, FFMIN(diff, INT_MAX));
+        }
+        ret = ff_inlink_consume_samples(inlink, 1, max_samples, &frame);
+        break;
+    }
+
+    if (ret > 0) {
+        while (current_segment_finished(ctx, frame)) {
+            ff_outlink_set_status(ctx->outputs[s->current_point], AVERROR_EOF, frame->pts);
+            s->current_point++;
+        }
+
+        if (s->current_point >= s->nb_points) {
+            av_frame_free(&frame);
+            return AVERROR(EINVAL);
+        }
+
+        ret = ff_filter_frame(ctx->outputs[s->current_point], frame);
+    }
+
+    if (ret < 0) {
+        return ret;
+    } else if (ff_inlink_acknowledge_status(inlink, &status, &pts)) {
+        for (int i = s->current_point; i < s->nb_points; i++)
+            ff_outlink_set_status(ctx->outputs[i], status, pts);
+        return 0;
+    } else {
+        for (int i = s->current_point; i < s->nb_points; i++) {
+            if (ff_outlink_frame_wanted(ctx->outputs[i]))
+                ff_inlink_request_frame(inlink);
+        }
+        return 0;
+    }
+}
+
+static av_cold void uninit(AVFilterContext *ctx)
+{
+    SegmentContext *s = ctx->priv;
+
+    av_freep(&s->points);
+
+    for (unsigned i = 0; i < ctx->nb_outputs; i++)
+        av_freep(&ctx->output_pads[i].name);
+}
+
+#define OFFSET(x) offsetof(SegmentContext, x)
+#define COMMON_OPTS \
+    { "timestamps", "timestamps of input at which to split input", OFFSET(timestamps_str),  AV_OPT_TYPE_STRING, { .str = NULL }, 0, 0, FLAGS }, \
+
+#if CONFIG_SEGMENT_FILTER
+
+static av_cold int video_init(AVFilterContext *ctx)
+{
+    return init(ctx, AVMEDIA_TYPE_VIDEO);
+}
+
+#define FLAGS AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_FILTERING_PARAM
+static const AVOption segment_options[] = {
+    COMMON_OPTS
+    { "frames", "frames at which to split input", OFFSET(points_str), AV_OPT_TYPE_STRING,  { .str = NULL }, 0, 0, FLAGS },
+    { NULL }
+};
+#undef FLAGS
+
+AVFILTER_DEFINE_CLASS(segment);
+
+static const AVFilterPad segment_inputs[] = {
+    {
+        .name         = "default",
+        .type         = AVMEDIA_TYPE_VIDEO,
+        .config_props = config_input,
+    },
+    { NULL }
+};
+
+const AVFilter ff_vf_segment = {
+    .name        = "segment",
+    .description = NULL_IF_CONFIG_SMALL("Segment video stream."),
+    .init        = video_init,
+    .uninit      = uninit,
+    .priv_size   = sizeof(SegmentContext),
+    .priv_class  = &segment_class,
+    .activate    = activate,
+    .inputs      = segment_inputs,
+    .outputs     = NULL,
+    .flags       = AVFILTER_FLAG_DYNAMIC_OUTPUTS,
+};
+#endif // CONFIG_SEGMENT_FILTER
+
+#if CONFIG_ASEGMENT_FILTER
+
+static av_cold int audio_init(AVFilterContext *ctx)
+{
+    return init(ctx, AVMEDIA_TYPE_AUDIO);
+}
+
+#define FLAGS AV_OPT_FLAG_AUDIO_PARAM | AV_OPT_FLAG_FILTERING_PARAM
+static const AVOption asegment_options[] = {
+    COMMON_OPTS
+    { "samples", "samples at which to split input", OFFSET(points_str), AV_OPT_TYPE_STRING,  { .str = NULL }, 0, 0, FLAGS },
+    { NULL }
+};
+#undef FLAGS
+
+AVFILTER_DEFINE_CLASS(asegment);
+
+static const AVFilterPad asegment_inputs[] = {
+    {
+        .name         = "default",
+        .type         = AVMEDIA_TYPE_AUDIO,
+        .config_props = config_input,
+    },
+    { NULL }
+};
+
+const AVFilter ff_af_asegment = {
+    .name        = "asegment",
+    .description = NULL_IF_CONFIG_SMALL("Segment audio stream."),
+    .init        = audio_init,
+    .uninit      = uninit,
+    .priv_size   = sizeof(SegmentContext),
+    .priv_class  = &asegment_class,
+    .activate    = activate,
+    .inputs      = asegment_inputs,
+    .outputs     = NULL,
+    .flags       = AVFILTER_FLAG_DYNAMIC_OUTPUTS,
+};
+#endif // CONFIG_ASEGMENT_FILTER