| 1 | /* |
| 2 | * Copyright (c) 2013 Stefano Sabatini |
| 3 | * |
| 4 | * This file is part of FFmpeg. |
| 5 | * |
| 6 | * FFmpeg is free software; you can redistribute it and/or |
| 7 | * modify it under the terms of the GNU Lesser General Public |
| 8 | * License as published by the Free Software Foundation; either |
| 9 | * version 2.1 of the License, or (at your option) any later version. |
| 10 | * |
| 11 | * FFmpeg is distributed in the hope that it will be useful, |
| 12 | * but WITHOUT ANY WARRANTY; without even the implied warranty of |
| 13 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
| 14 | * Lesser General Public License for more details. |
| 15 | * |
| 16 | * You should have received a copy of the GNU Lesser General Public |
| 17 | * License along with FFmpeg; if not, write to the Free Software |
| 18 | * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA |
| 19 | */ |
| 20 | |
| 21 | /** |
| 22 | * @file |
| 23 | * audio and video interleaver |
| 24 | */ |
| 25 | |
| 26 | #include "libavutil/avassert.h" |
| 27 | #include "libavutil/avstring.h" |
| 28 | #include "libavutil/opt.h" |
| 29 | #include "avfilter.h" |
| 30 | #include "bufferqueue.h" |
| 31 | #include "formats.h" |
| 32 | #include "internal.h" |
| 33 | #include "audio.h" |
| 34 | #include "video.h" |
| 35 | |
| 36 | typedef struct { |
| 37 | const AVClass *class; |
| 38 | int nb_inputs; |
| 39 | struct FFBufQueue *queues; |
| 40 | } InterleaveContext; |
| 41 | |
| 42 | #define OFFSET(x) offsetof(InterleaveContext, x) |
| 43 | |
| 44 | #define DEFINE_OPTIONS(filt_name, flags_) \ |
| 45 | static const AVOption filt_name##_options[] = { \ |
| 46 | { "nb_inputs", "set number of inputs", OFFSET(nb_inputs), AV_OPT_TYPE_INT, {.i64 = 2}, 1, INT_MAX, .flags = flags_ }, \ |
| 47 | { "n", "set number of inputs", OFFSET(nb_inputs), AV_OPT_TYPE_INT, {.i64 = 2}, 1, INT_MAX, .flags = flags_ }, \ |
| 48 | { NULL } \ |
| 49 | } |
| 50 | |
| 51 | inline static int push_frame(AVFilterContext *ctx) |
| 52 | { |
| 53 | InterleaveContext *s = ctx->priv; |
| 54 | AVFrame *frame; |
| 55 | int i, queue_idx = -1; |
| 56 | int64_t pts_min = INT64_MAX; |
| 57 | |
| 58 | /* look for oldest frame */ |
| 59 | for (i = 0; i < ctx->nb_inputs; i++) { |
| 60 | struct FFBufQueue *q = &s->queues[i]; |
| 61 | |
| 62 | if (!q->available && !ctx->inputs[i]->closed) |
| 63 | return 0; |
| 64 | if (q->available) { |
| 65 | frame = ff_bufqueue_peek(q, 0); |
| 66 | if (frame->pts < pts_min) { |
| 67 | pts_min = frame->pts; |
| 68 | queue_idx = i; |
| 69 | } |
| 70 | } |
| 71 | } |
| 72 | |
| 73 | /* all inputs are closed */ |
| 74 | if (queue_idx < 0) |
| 75 | return AVERROR_EOF; |
| 76 | |
| 77 | frame = ff_bufqueue_get(&s->queues[queue_idx]); |
| 78 | av_log(ctx, AV_LOG_DEBUG, "queue:%d -> frame time:%f\n", |
| 79 | queue_idx, frame->pts * av_q2d(AV_TIME_BASE_Q)); |
| 80 | return ff_filter_frame(ctx->outputs[0], frame); |
| 81 | } |
| 82 | |
| 83 | static int filter_frame(AVFilterLink *inlink, AVFrame *frame) |
| 84 | { |
| 85 | AVFilterContext *ctx = inlink->dst; |
| 86 | InterleaveContext *s = ctx->priv; |
| 87 | unsigned in_no = FF_INLINK_IDX(inlink); |
| 88 | |
| 89 | if (frame->pts == AV_NOPTS_VALUE) { |
| 90 | av_log(ctx, AV_LOG_WARNING, |
| 91 | "NOPTS value for input frame cannot be accepted, frame discarded\n"); |
| 92 | av_frame_free(&frame); |
| 93 | return AVERROR_INVALIDDATA; |
| 94 | } |
| 95 | |
| 96 | /* queue frame */ |
| 97 | frame->pts = av_rescale_q(frame->pts, inlink->time_base, AV_TIME_BASE_Q); |
| 98 | av_log(ctx, AV_LOG_DEBUG, "frame pts:%f -> queue idx:%d available:%d\n", |
| 99 | frame->pts * av_q2d(AV_TIME_BASE_Q), in_no, s->queues[in_no].available); |
| 100 | ff_bufqueue_add(ctx, &s->queues[in_no], frame); |
| 101 | |
| 102 | return push_frame(ctx); |
| 103 | } |
| 104 | |
| 105 | static av_cold int init(AVFilterContext *ctx) |
| 106 | { |
| 107 | InterleaveContext *s = ctx->priv; |
| 108 | const AVFilterPad *outpad = &ctx->filter->outputs[0]; |
| 109 | int i; |
| 110 | |
| 111 | s->queues = av_calloc(s->nb_inputs, sizeof(s->queues[0])); |
| 112 | if (!s->queues) |
| 113 | return AVERROR(ENOMEM); |
| 114 | |
| 115 | for (i = 0; i < s->nb_inputs; i++) { |
| 116 | AVFilterPad inpad = { 0 }; |
| 117 | |
| 118 | inpad.name = av_asprintf("input%d", i); |
| 119 | if (!inpad.name) |
| 120 | return AVERROR(ENOMEM); |
| 121 | inpad.type = outpad->type; |
| 122 | inpad.filter_frame = filter_frame; |
| 123 | |
| 124 | switch (outpad->type) { |
| 125 | case AVMEDIA_TYPE_VIDEO: |
| 126 | inpad.get_video_buffer = ff_null_get_video_buffer; break; |
| 127 | case AVMEDIA_TYPE_AUDIO: |
| 128 | inpad.get_audio_buffer = ff_null_get_audio_buffer; break; |
| 129 | default: |
| 130 | av_assert0(0); |
| 131 | } |
| 132 | ff_insert_inpad(ctx, i, &inpad); |
| 133 | } |
| 134 | |
| 135 | return 0; |
| 136 | } |
| 137 | |
| 138 | static av_cold void uninit(AVFilterContext *ctx) |
| 139 | { |
| 140 | InterleaveContext *s = ctx->priv; |
| 141 | int i; |
| 142 | |
| 143 | for (i = 0; i < ctx->nb_inputs; i++) { |
| 144 | ff_bufqueue_discard_all(&s->queues[i]); |
| 145 | av_freep(&s->queues[i]); |
| 146 | av_freep(&ctx->input_pads[i].name); |
| 147 | } |
| 148 | } |
| 149 | |
| 150 | static int config_output(AVFilterLink *outlink) |
| 151 | { |
| 152 | AVFilterContext *ctx = outlink->src; |
| 153 | AVFilterLink *inlink0 = ctx->inputs[0]; |
| 154 | int i; |
| 155 | |
| 156 | if (outlink->type == AVMEDIA_TYPE_VIDEO) { |
| 157 | outlink->time_base = AV_TIME_BASE_Q; |
| 158 | outlink->w = inlink0->w; |
| 159 | outlink->h = inlink0->h; |
| 160 | outlink->sample_aspect_ratio = inlink0->sample_aspect_ratio; |
| 161 | outlink->format = inlink0->format; |
| 162 | outlink->frame_rate = (AVRational) {1, 0}; |
| 163 | for (i = 1; i < ctx->nb_inputs; i++) { |
| 164 | AVFilterLink *inlink = ctx->inputs[i]; |
| 165 | |
| 166 | if (outlink->w != inlink->w || |
| 167 | outlink->h != inlink->h || |
| 168 | outlink->sample_aspect_ratio.num != inlink->sample_aspect_ratio.num || |
| 169 | outlink->sample_aspect_ratio.den != inlink->sample_aspect_ratio.den) { |
| 170 | av_log(ctx, AV_LOG_ERROR, "Parameters for input link %s " |
| 171 | "(size %dx%d, SAR %d:%d) do not match the corresponding " |
| 172 | "output link parameters (%dx%d, SAR %d:%d)\n", |
| 173 | ctx->input_pads[i].name, inlink->w, inlink->h, |
| 174 | inlink->sample_aspect_ratio.num, |
| 175 | inlink->sample_aspect_ratio.den, |
| 176 | outlink->w, outlink->h, |
| 177 | outlink->sample_aspect_ratio.num, |
| 178 | outlink->sample_aspect_ratio.den); |
| 179 | return AVERROR(EINVAL); |
| 180 | } |
| 181 | } |
| 182 | } |
| 183 | |
| 184 | outlink->flags |= FF_LINK_FLAG_REQUEST_LOOP; |
| 185 | return 0; |
| 186 | } |
| 187 | |
| 188 | static int request_frame(AVFilterLink *outlink) |
| 189 | { |
| 190 | AVFilterContext *ctx = outlink->src; |
| 191 | InterleaveContext *s = ctx->priv; |
| 192 | int i, ret; |
| 193 | |
| 194 | for (i = 0; i < ctx->nb_inputs; i++) { |
| 195 | if (!s->queues[i].available && !ctx->inputs[i]->closed) { |
| 196 | ret = ff_request_frame(ctx->inputs[i]); |
| 197 | if (ret != AVERROR_EOF) |
| 198 | return ret; |
| 199 | } |
| 200 | } |
| 201 | |
| 202 | return push_frame(ctx); |
| 203 | } |
| 204 | |
| 205 | #if CONFIG_INTERLEAVE_FILTER |
| 206 | |
| 207 | DEFINE_OPTIONS(interleave, AV_OPT_FLAG_VIDEO_PARAM|AV_OPT_FLAG_FILTERING_PARAM); |
| 208 | AVFILTER_DEFINE_CLASS(interleave); |
| 209 | |
| 210 | static const AVFilterPad interleave_outputs[] = { |
| 211 | { |
| 212 | .name = "default", |
| 213 | .type = AVMEDIA_TYPE_VIDEO, |
| 214 | .config_props = config_output, |
| 215 | .request_frame = request_frame, |
| 216 | }, |
| 217 | { NULL } |
| 218 | }; |
| 219 | |
| 220 | AVFilter ff_vf_interleave = { |
| 221 | .name = "interleave", |
| 222 | .description = NULL_IF_CONFIG_SMALL("Temporally interleave video inputs."), |
| 223 | .priv_size = sizeof(InterleaveContext), |
| 224 | .init = init, |
| 225 | .uninit = uninit, |
| 226 | .outputs = interleave_outputs, |
| 227 | .priv_class = &interleave_class, |
| 228 | .flags = AVFILTER_FLAG_DYNAMIC_INPUTS, |
| 229 | }; |
| 230 | |
| 231 | #endif |
| 232 | |
| 233 | #if CONFIG_AINTERLEAVE_FILTER |
| 234 | |
| 235 | DEFINE_OPTIONS(ainterleave, AV_OPT_FLAG_AUDIO_PARAM|AV_OPT_FLAG_FILTERING_PARAM); |
| 236 | AVFILTER_DEFINE_CLASS(ainterleave); |
| 237 | |
| 238 | static const AVFilterPad ainterleave_outputs[] = { |
| 239 | { |
| 240 | .name = "default", |
| 241 | .type = AVMEDIA_TYPE_AUDIO, |
| 242 | .config_props = config_output, |
| 243 | .request_frame = request_frame, |
| 244 | }, |
| 245 | { NULL } |
| 246 | }; |
| 247 | |
| 248 | AVFilter ff_af_ainterleave = { |
| 249 | .name = "ainterleave", |
| 250 | .description = NULL_IF_CONFIG_SMALL("Temporally interleave audio inputs."), |
| 251 | .priv_size = sizeof(InterleaveContext), |
| 252 | .init = init, |
| 253 | .uninit = uninit, |
| 254 | .outputs = ainterleave_outputs, |
| 255 | .priv_class = &ainterleave_class, |
| 256 | .flags = AVFILTER_FLAG_DYNAMIC_INPUTS, |
| 257 | }; |
| 258 | |
| 259 | #endif |