/* * This file is part of FFmpeg. * * FFmpeg is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public * License as published by the Free Software Foundation; either * version 2.1 of the License, or (at your option) any later version. * * FFmpeg is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Lesser General Public License for more details. * * You should have received a copy of the GNU Lesser General Public * License along with FFmpeg; if not, write to the Free Software * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA */ #include "config_components.h" #include #include "libavutil/avstring.h" #include "libavutil/log.h" #include "libavutil/mathematics.h" #include "libavutil/mem.h" #include "libavutil/opt.h" #include "libavutil/parseutils.h" #include "avfilter.h" #include "filters.h" #include "internal.h" typedef struct SegmentContext { const AVClass *class; char *timestamps_str; char *points_str; int use_timestamps; int current_point; int nb_points; int64_t last_pts; int64_t *points; } SegmentContext; static void count_points(char *item_str, int *nb_items) { char *p; if (!item_str) return; *nb_items = 1; for (p = item_str; *p; p++) { if (*p == '|') (*nb_items)++; } } static int parse_points(AVFilterContext *ctx, char *item_str, int nb_points, int64_t *points) { SegmentContext *s = ctx->priv; char *arg, *p = item_str; char *saveptr = NULL; int64_t ref, cur = 0; int ret = 0; for (int i = 0; i < nb_points; i++) { if (!(arg = av_strtok(p, "|", &saveptr))) return AVERROR(EINVAL); p = NULL; ref = 0; if (*arg == '+') { ref = cur; arg++; } if (s->use_timestamps) { ret = av_parse_time(&points[i], arg, s->use_timestamps); } else { if (sscanf(arg, "%"SCNd64, &points[i]) != 1) ret = AVERROR(EINVAL); } if (ret < 0) { av_log(ctx, AV_LOG_ERROR, "Invalid splits supplied: %s\n", arg); return ret; } cur = points[i]; points[i] += ref; } return 0; } static av_cold int init(AVFilterContext *ctx, enum AVMediaType type) { SegmentContext *s = ctx->priv; char *split_str; int ret; if (s->timestamps_str && s->points_str) { av_log(ctx, AV_LOG_ERROR, "Both timestamps and counts supplied.\n"); return AVERROR(EINVAL); } else if (s->timestamps_str) { s->use_timestamps = 1; split_str = s->timestamps_str; } else if (s->points_str) { split_str = s->points_str; } else { av_log(ctx, AV_LOG_ERROR, "Neither timestamps nor durations nor counts supplied.\n"); return AVERROR(EINVAL); } count_points(split_str, &s->nb_points); s->nb_points++; s->points = av_calloc(s->nb_points, sizeof(*s->points)); if (!s->points) return AVERROR(ENOMEM); ret = parse_points(ctx, split_str, s->nb_points - 1, s->points); if (ret < 0) return ret; s->points[s->nb_points - 1] = INT64_MAX; for (int i = 0; i < s->nb_points; i++) { AVFilterPad pad = { 0 }; pad.type = type; pad.name = av_asprintf("output%d", i); if (!pad.name) return AVERROR(ENOMEM); if ((ret = ff_append_outpad_free_name(ctx, &pad)) < 0) return ret; } return 0; } static int config_input(AVFilterLink *inlink) { AVFilterContext *ctx = inlink->dst; SegmentContext *s = ctx->priv; AVRational tb = inlink->time_base; if (s->use_timestamps) { for (int i = 0; i < s->nb_points - 1; i++) s->points[i] = av_rescale_q(s->points[i], AV_TIME_BASE_Q, tb); } return 0; } static int current_segment_finished(AVFilterContext *ctx, AVFrame *frame) { SegmentContext *s = ctx->priv; AVFilterLink *inlink = ctx->inputs[0]; int ret = 0; if (s->use_timestamps) { ret = frame->pts >= s->points[s->current_point]; } else { switch (inlink->type) { case AVMEDIA_TYPE_VIDEO: ret = inlink->frame_count_out - 1 >= s->points[s->current_point]; break; case AVMEDIA_TYPE_AUDIO: ret = inlink->sample_count_out - frame->nb_samples >= s->points[s->current_point]; break; } } return ret; } static int activate(AVFilterContext *ctx) { AVFilterLink *inlink = ctx->inputs[0]; SegmentContext *s = ctx->priv; AVFrame *frame = NULL; int ret, status; int64_t max_samples; int64_t diff; int64_t pts; for (int i = s->current_point; i < s->nb_points; i++) { FF_FILTER_FORWARD_STATUS_BACK_ALL(ctx->outputs[i], ctx); } switch (inlink->type) { case AVMEDIA_TYPE_VIDEO: ret = ff_inlink_consume_frame(inlink, &frame); break; case AVMEDIA_TYPE_AUDIO: diff = s->points[s->current_point] - inlink->sample_count_out; while (diff <= 0) { ff_outlink_set_status(ctx->outputs[s->current_point], AVERROR_EOF, s->last_pts); s->current_point++; if (s->current_point >= s->nb_points) return AVERROR(EINVAL); diff = s->points[s->current_point] - inlink->sample_count_out; } if (s->use_timestamps) { max_samples = av_rescale_q(diff, av_make_q(1, inlink->sample_rate), inlink->time_base); } else { max_samples = FFMAX(1, FFMIN(diff, INT_MAX)); } if (max_samples <= 0 || max_samples > INT_MAX) ret = ff_inlink_consume_frame(inlink, &frame); else ret = ff_inlink_consume_samples(inlink, 1, max_samples, &frame); break; default: return AVERROR_BUG; } if (ret > 0) { s->last_pts = frame->pts; while (current_segment_finished(ctx, frame)) { ff_outlink_set_status(ctx->outputs[s->current_point], AVERROR_EOF, frame->pts); s->current_point++; } if (s->current_point >= s->nb_points) { av_frame_free(&frame); return AVERROR(EINVAL); } ret = ff_filter_frame(ctx->outputs[s->current_point], frame); } if (ret < 0) { return ret; } else if (ff_inlink_acknowledge_status(inlink, &status, &pts)) { for (int i = s->current_point; i < s->nb_points; i++) ff_outlink_set_status(ctx->outputs[i], status, pts); return 0; } else { for (int i = s->current_point; i < s->nb_points; i++) { if (ff_outlink_frame_wanted(ctx->outputs[i])) ff_inlink_request_frame(inlink); } return 0; } } static av_cold void uninit(AVFilterContext *ctx) { SegmentContext *s = ctx->priv; av_freep(&s->points); } #define OFFSET(x) offsetof(SegmentContext, x) #define COMMON_OPTS \ { "timestamps", "timestamps of input at which to split input", OFFSET(timestamps_str), AV_OPT_TYPE_STRING, { .str = NULL }, 0, 0, FLAGS }, \ #if CONFIG_SEGMENT_FILTER static av_cold int video_init(AVFilterContext *ctx) { return init(ctx, AVMEDIA_TYPE_VIDEO); } #define FLAGS AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_FILTERING_PARAM static const AVOption segment_options[] = { COMMON_OPTS { "frames", "frames at which to split input", OFFSET(points_str), AV_OPT_TYPE_STRING, { .str = NULL }, 0, 0, FLAGS }, { NULL } }; #undef FLAGS AVFILTER_DEFINE_CLASS(segment); static const AVFilterPad segment_inputs[] = { { .name = "default", .type = AVMEDIA_TYPE_VIDEO, .config_props = config_input, }, }; const AVFilter ff_vf_segment = { .name = "segment", .description = NULL_IF_CONFIG_SMALL("Segment video stream."), .init = video_init, .uninit = uninit, .priv_size = sizeof(SegmentContext), .priv_class = &segment_class, .activate = activate, FILTER_INPUTS(segment_inputs), .outputs = NULL, .flags = AVFILTER_FLAG_DYNAMIC_OUTPUTS | AVFILTER_FLAG_METADATA_ONLY, }; #endif // CONFIG_SEGMENT_FILTER #if CONFIG_ASEGMENT_FILTER static av_cold int audio_init(AVFilterContext *ctx) { return init(ctx, AVMEDIA_TYPE_AUDIO); } #define FLAGS AV_OPT_FLAG_AUDIO_PARAM | AV_OPT_FLAG_FILTERING_PARAM static const AVOption asegment_options[] = { COMMON_OPTS { "samples", "samples at which to split input", OFFSET(points_str), AV_OPT_TYPE_STRING, { .str = NULL }, 0, 0, FLAGS }, { NULL } }; #undef FLAGS AVFILTER_DEFINE_CLASS(asegment); static const AVFilterPad asegment_inputs[] = { { .name = "default", .type = AVMEDIA_TYPE_AUDIO, .config_props = config_input, }, }; const AVFilter ff_af_asegment = { .name = "asegment", .description = NULL_IF_CONFIG_SMALL("Segment audio stream."), .init = audio_init, .uninit = uninit, .priv_size = sizeof(SegmentContext), .priv_class = &asegment_class, .activate = activate, FILTER_INPUTS(asegment_inputs), .outputs = NULL, .flags = AVFILTER_FLAG_DYNAMIC_OUTPUTS | AVFILTER_FLAG_METADATA_ONLY, }; #endif // CONFIG_ASEGMENT_FILTER