Paul B Mahol (12021-08-10): > Signed-off-by: Paul B Mahol <one...@gmail.com> > --- > doc/filters.texi | 34 ++++ > libavfilter/Makefile | 2 + > libavfilter/allfilters.c | 2 + > libavfilter/f_segment.c | 335 +++++++++++++++++++++++++++++++++++++++ > 4 files changed, 373 insertions(+) > create mode 100644 libavfilter/f_segment.c > > diff --git a/doc/filters.texi b/doc/filters.texi > index bdeb3fedfd..b65afbda65 100644 > --- a/doc/filters.texi > +++ b/doc/filters.texi > @@ -25665,6 +25665,40 @@ A processing speed faster than what is possible > without these filters cannot > be achieved. > @end table > > +@section segment, asegment > + > +Split single input stream into multiple streams. > + > +This filter does opposite of concat filters. > + > +@code{segment} works on video frames, @code{asegment} on audio samples. > + > +This filter accepts the following options: > +
> +@table @option > +@item timestamps > +Timestamps of output segments separated by '|'. The first segment will run > +from the beginning of the input stream. The last segment will run until > +the end of the input stream > + > +@item frames, samples > +Exact frame/sample count to split the segments. > +@end table Since you adopted my suggestion of + for relative parsing (thank you for it), it needs to be documented: "In both cases, prefixing a point with '+' will make it relative to the previous point." > + > +@subsection Examples > + > +@itemize > + > +@item > +Split input audio stream into three output audio streams, starting at start > of input audio stream > +and storing that in 1st output audio stream, then following at 60th second > and storing than in 2nd > +output audio stream, and last after 120th second of input audio stream store > in 3rd output audio stream: > +@example > +asegment=timestamps="60 | 120" > +@end example I would suggest to use another number than 120, because 120-60=60 which makes the explanation ambiguous; 150 would be a good choice. > + > +@end itemize > + > @anchor{select} > @section select, aselect > > diff --git a/libavfilter/Makefile b/libavfilter/Makefile > index 49c0c8342b..102ce7beff 100644 > --- a/libavfilter/Makefile > +++ b/libavfilter/Makefile > @@ -78,6 +78,7 @@ OBJS-$(CONFIG_AREALTIME_FILTER) += f_realtime.o > OBJS-$(CONFIG_ARESAMPLE_FILTER) += af_aresample.o > OBJS-$(CONFIG_AREVERSE_FILTER) += f_reverse.o > OBJS-$(CONFIG_ARNNDN_FILTER) += af_arnndn.o > +OBJS-$(CONFIG_ASEGMENT_FILTER) += f_segment.o > OBJS-$(CONFIG_ASELECT_FILTER) += f_select.o > OBJS-$(CONFIG_ASENDCMD_FILTER) += f_sendcmd.o > OBJS-$(CONFIG_ASETNSAMPLES_FILTER) += af_asetnsamples.o > @@ -404,6 +405,7 @@ OBJS-$(CONFIG_SCALE_VULKAN_FILTER) += > vf_scale_vulkan.o vulkan.o > OBJS-$(CONFIG_SCALE2REF_FILTER) += vf_scale.o scale_eval.o > OBJS-$(CONFIG_SCDET_FILTER) += vf_scdet.o > OBJS-$(CONFIG_SCROLL_FILTER) += vf_scroll.o > +OBJS-$(CONFIG_SEGMENT_FILTER) += f_segment.o > OBJS-$(CONFIG_SELECT_FILTER) += f_select.o > OBJS-$(CONFIG_SELECTIVECOLOR_FILTER) += vf_selectivecolor.o > OBJS-$(CONFIG_SENDCMD_FILTER) += f_sendcmd.o > diff --git a/libavfilter/allfilters.c b/libavfilter/allfilters.c > index ae74f9c891..73040d2824 100644 > --- a/libavfilter/allfilters.c > +++ b/libavfilter/allfilters.c > @@ -71,6 +71,7 @@ extern const AVFilter ff_af_arealtime; > extern const AVFilter ff_af_aresample; > extern const AVFilter ff_af_areverse; > extern const AVFilter ff_af_arnndn; > +extern const AVFilter ff_af_asegment; > extern const AVFilter ff_af_aselect; > extern const AVFilter ff_af_asendcmd; > extern const AVFilter ff_af_asetnsamples; > @@ -385,6 +386,7 @@ extern const AVFilter ff_vf_scale_vulkan; > extern const AVFilter ff_vf_scale2ref; > extern const AVFilter ff_vf_scdet; > extern const AVFilter ff_vf_scroll; > +extern const AVFilter ff_vf_segment; > extern const AVFilter ff_vf_select; > extern const AVFilter ff_vf_selectivecolor; > extern const AVFilter ff_vf_sendcmd; > diff --git a/libavfilter/f_segment.c b/libavfilter/f_segment.c > new file mode 100644 > index 0000000000..7e0521b3c1 > --- /dev/null > +++ b/libavfilter/f_segment.c > @@ -0,0 +1,335 @@ > +/* > + * This file is part of FFmpeg. > + * > + * FFmpeg is free software; you can redistribute it and/or > + * modify it under the terms of the GNU Lesser General Public > + * License as published by the Free Software Foundation; either > + * version 2.1 of the License, or (at your option) any later version. > + * > + * FFmpeg is distributed in the hope that it will be useful, > + * but WITHOUT ANY WARRANTY; without even the implied warranty of > + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU > + * Lesser General Public License for more details. > + * > + * You should have received a copy of the GNU Lesser General Public > + * License along with FFmpeg; if not, write to the Free Software > + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 > USA > + */ > + > +#include <stdint.h> > + > +#include "libavutil/avstring.h" > +#include "libavutil/channel_layout.h" > +#include "libavutil/common.h" > +#include "libavutil/log.h" > +#include "libavutil/mathematics.h" > +#include "libavutil/opt.h" > +#include "libavutil/parseutils.h" > +#include "libavutil/samplefmt.h" > + > +#include "audio.h" > +#include "avfilter.h" > +#include "filters.h" > +#include "internal.h" > + > +typedef struct SegmentContext { > + const AVClass *class; > + > + char *timestamps_str; > + char *points_str; > + int use_timestamps; > + > + int current_point; > + int nb_points; > + > + int64_t *points; > +} SegmentContext; > + > +static void count_points(char *item_str, int *nb_items) > +{ > + char *p; > + > + if (!item_str) > + return; > + > + *nb_items = 1; > + for (p = item_str; *p; p++) { > + if (*p == '|') > + (*nb_items)++; > + } > +} > + > +static int parse_points(AVFilterContext *ctx, char *item_str, int nb_points, > int64_t *points) > +{ > + SegmentContext *s = ctx->priv; > + char *arg, *p = item_str; > + char *saveptr = NULL; > + int64_t ref, cur = 0; > + int ret = 0; > + > + for (int i = 0; i < nb_points; i++) { > + if (!(arg = av_strtok(p, "|", &saveptr))) > + return AVERROR(EINVAL); > + > + p = NULL; > + ref = 0; > + if (*arg == '+') { > + ref = cur; > + arg++; > + } > + > + if (s->use_timestamps) { > + ret = av_parse_time(&points[i], arg, s->use_timestamps); > + } else { > + if (sscanf(arg, "%"SCNd64, &points[i]) != 1) > + ret = AVERROR(EINVAL); > + } > + > + if (ret < 0) { > + av_log(ctx, AV_LOG_ERROR, "Invalid splits supplied: %s\n", arg); > + return AVERROR(EINVAL); return ret > + } > + > + cur = points[i]; > + points[i] += ref; > + } > + > + return 0; > +} > + > +static av_cold int init(AVFilterContext *ctx, enum AVMediaType type) > +{ > + SegmentContext *s = ctx->priv; > + char *split_str; > + int ret; > + > + if (s->timestamps_str && s->points_str) { > + av_log(ctx, AV_LOG_ERROR, "Both timestamps and counts supplied.\n"); > + return AVERROR(EINVAL); > + } else if (s->timestamps_str) { > + s->use_timestamps = 1; > + split_str = s->timestamps_str; > + } else if (s->points_str) { > + split_str = s->points_str; > + } else { > + av_log(ctx, AV_LOG_ERROR, "Neither timestamps nor durations nor > counts supplied.\n"); > + return AVERROR(EINVAL); > + } > + > + count_points(split_str, &s->nb_points); > + s->nb_points++; > + > + s->points = av_calloc(s->nb_points, sizeof(*s->points)); > + if (!s->points) > + return AVERROR(ENOMEM); > + > + ret = parse_points(ctx, split_str, s->nb_points - 1, s->points); > + if (ret < 0) > + return ret; > + > + s->points[s->nb_points - 1] = INT64_MAX; > + > + for (int i = 0; i < s->nb_points; i++) { > + AVFilterPad pad = { 0 }; > + > + pad.type = type; > + pad.name = av_asprintf("output%d", i); > + if (!pad.name) > + return AVERROR(ENOMEM); > + > + if ((ret = ff_insert_outpad(ctx, i, &pad)) < 0) { > + av_freep(&pad.name); > + return ret; > + } > + } > + > + return 0; > +} > + > +static int config_input(AVFilterLink *inlink) > +{ > + AVFilterContext *ctx = inlink->dst; > + SegmentContext *s = ctx->priv; > + AVRational tb = inlink->time_base; > + > + if (s->use_timestamps) { > + for (int i = 0; i < s->nb_points - 1; i++) > + s->points[i] = av_rescale_q(s->points[i], AV_TIME_BASE_Q, tb); > + } > + > + return 0; > +} > + > +static int current_segment_finished(AVFilterContext *ctx, AVFrame *frame) > +{ > + SegmentContext *s = ctx->priv; > + AVFilterLink *inlink = ctx->inputs[0]; > + int ret = 0; > + > + if (s->use_timestamps) { > + ret = frame->pts >= s->points[s->current_point]; > + } else { > + switch (inlink->type) { > + case AVMEDIA_TYPE_VIDEO: > + ret = inlink->frame_count_out - 1 >= s->points[s->current_point]; > + break; > + case AVMEDIA_TYPE_AUDIO: > + ret = inlink->sample_count_out - frame->nb_samples >= > s->points[s->current_point]; > + break; > + } > + } > + > + return ret; > +} > + > +static int activate(AVFilterContext *ctx) > +{ > + AVFilterLink *inlink = ctx->inputs[0]; > + SegmentContext *s = ctx->priv; > + AVFrame *frame = NULL; > + int ret, status; > + int max_samples; > + int64_t diff; > + int64_t pts; > + > + for (int i = s->current_point; i < s->nb_points; i++) { > + FF_FILTER_FORWARD_STATUS_BACK_ALL(ctx->outputs[i], ctx); > + } > + > + switch (inlink->type) { > + case AVMEDIA_TYPE_VIDEO: > + ret = ff_inlink_consume_frame(inlink, &frame); > + break; > + case AVMEDIA_TYPE_AUDIO: > + diff = s->points[s->current_point] - inlink->sample_count_out; > + if (s->use_timestamps) { > + max_samples = av_rescale_q(diff, av_make_q(1, > inlink->sample_rate), inlink->time_base); > + } else { > + max_samples = FFMAX(1, FFMIN(diff, INT_MAX)); > + } > + ret = ff_inlink_consume_samples(inlink, 1, max_samples, &frame); > + break; > + } > + > + if (ret > 0) { > + while (current_segment_finished(ctx, frame)) { > + ff_outlink_set_status(ctx->outputs[s->current_point], > AVERROR_EOF, frame->pts); > + s->current_point++; > + } > + > + if (s->current_point >= s->nb_points) { > + av_frame_free(&frame); > + return AVERROR(EINVAL); > + } > + > + ret = ff_filter_frame(ctx->outputs[s->current_point], frame); > + } I wonder if you should subtract the start of the current segment to the timestamp of the input frame before feeding it to output. I.e. if the input is [0;600[, output [0;60[, [0;90[, [0;450[ instead of [0;60[, [60;150[, [150;600[. > + > + if (ret < 0) { > + return ret; > + } else if (ff_inlink_acknowledge_status(inlink, &status, &pts)) { > + for (int i = s->current_point; i < s->nb_points; i++) > + ff_outlink_set_status(ctx->outputs[i], status, pts); > + return 0; > + } else { > + for (int i = s->current_point; i < s->nb_points; i++) { > + if (ff_outlink_frame_wanted(ctx->outputs[i])) > + ff_inlink_request_frame(inlink); > + } > + return 0; > + } > +} > + > +static av_cold void uninit(AVFilterContext *ctx) > +{ > + SegmentContext *s = ctx->priv; > + > + av_freep(&s->points); > + > + for (unsigned i = 0; i < ctx->nb_outputs; i++) > + av_freep(&ctx->output_pads[i].name); > +} > + > +#define OFFSET(x) offsetof(SegmentContext, x) > +#define COMMON_OPTS \ > + { "timestamps", "timestamps of input at which to split input", > OFFSET(timestamps_str), AV_OPT_TYPE_STRING, { .str = NULL }, 0, 0, FLAGS }, \ > + > +#if CONFIG_SEGMENT_FILTER > + > +static av_cold int video_init(AVFilterContext *ctx) > +{ > + return init(ctx, AVMEDIA_TYPE_VIDEO); > +} > + > +#define FLAGS AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_FILTERING_PARAM > +static const AVOption segment_options[] = { > + COMMON_OPTS > + { "frames", "frames at which to split input", OFFSET(points_str), > AV_OPT_TYPE_STRING, { .str = NULL }, 0, 0, FLAGS }, > + { NULL } > +}; > +#undef FLAGS > + > +AVFILTER_DEFINE_CLASS(segment); > + > +static const AVFilterPad segment_inputs[] = { > + { > + .name = "default", > + .type = AVMEDIA_TYPE_VIDEO, > + .config_props = config_input, > + }, > + { NULL } > +}; > + > +const AVFilter ff_vf_segment = { > + .name = "segment", > + .description = NULL_IF_CONFIG_SMALL("Segment video stream."), > + .init = video_init, > + .uninit = uninit, > + .priv_size = sizeof(SegmentContext), > + .priv_class = &segment_class, > + .activate = activate, > + .inputs = segment_inputs, > + .outputs = NULL, > + .flags = AVFILTER_FLAG_DYNAMIC_OUTPUTS, > +}; > +#endif // CONFIG_SEGMENT_FILTER > + > +#if CONFIG_ASEGMENT_FILTER > + > +static av_cold int audio_init(AVFilterContext *ctx) > +{ > + return init(ctx, AVMEDIA_TYPE_AUDIO); > +} > + > +#define FLAGS AV_OPT_FLAG_AUDIO_PARAM | AV_OPT_FLAG_FILTERING_PARAM > +static const AVOption asegment_options[] = { > + COMMON_OPTS > + { "samples", "samples at which to split input", OFFSET(points_str), > AV_OPT_TYPE_STRING, { .str = NULL }, 0, 0, FLAGS }, > + { NULL } > +}; > +#undef FLAGS > + > +AVFILTER_DEFINE_CLASS(asegment); > + > +static const AVFilterPad asegment_inputs[] = { > + { > + .name = "default", > + .type = AVMEDIA_TYPE_AUDIO, > + .config_props = config_input, > + }, > + { NULL } > +}; > + > +const AVFilter ff_af_asegment = { > + .name = "asegment", > + .description = NULL_IF_CONFIG_SMALL("Segment audio stream."), > + .init = audio_init, > + .uninit = uninit, > + .priv_size = sizeof(SegmentContext), > + .priv_class = &asegment_class, > + .activate = activate, > + .inputs = asegment_inputs, > + .outputs = NULL, > + .flags = AVFILTER_FLAG_DYNAMIC_OUTPUTS, > +}; > +#endif // CONFIG_ASEGMENT_FILTER Apart from these details, LGTM, thanks for your efforts. -- Nicolas George
signature.asc
Description: PGP signature
_______________________________________________ ffmpeg-devel mailing list ffmpeg-devel@ffmpeg.org https://ffmpeg.org/mailman/listinfo/ffmpeg-devel To unsubscribe, visit link above, or email ffmpeg-devel-requ...@ffmpeg.org with subject "unsubscribe".