Line | Branch | Exec | Source |
---|---|---|---|
1 | /* | ||
2 | * Copyright (c) 2015 Paul B Mahol | ||
3 | * | ||
4 | * This file is part of FFmpeg. | ||
5 | * | ||
6 | * FFmpeg is free software; you can redistribute it and/or | ||
7 | * modify it under the terms of the GNU Lesser General Public | ||
8 | * License as published by the Free Software Foundation; either | ||
9 | * version 2.1 of the License, or (at your option) any later version. | ||
10 | * | ||
11 | * FFmpeg is distributed in the hope that it will be useful, | ||
12 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | ||
13 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU | ||
14 | * Lesser General Public License for more details. | ||
15 | * | ||
16 | * You should have received a copy of the GNU Lesser General Public | ||
17 | * License along with FFmpeg; if not, write to the Free Software | ||
18 | * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA | ||
19 | */ | ||
20 | |||
21 | /** | ||
22 | * @file | ||
23 | * audio to video multimedia aphasemeter filter | ||
24 | */ | ||
25 | |||
26 | #include <float.h> | ||
27 | |||
28 | #include "libavutil/channel_layout.h" | ||
29 | #include "libavutil/intreadwrite.h" | ||
30 | #include "libavutil/opt.h" | ||
31 | #include "libavutil/parseutils.h" | ||
32 | #include "libavutil/timestamp.h" | ||
33 | #include "avfilter.h" | ||
34 | #include "filters.h" | ||
35 | #include "formats.h" | ||
36 | #include "audio.h" | ||
37 | #include "video.h" | ||
38 | |||
39 | typedef struct AudioPhaseMeterContext { | ||
40 | const AVClass *class; | ||
41 | AVFrame *out, *in; | ||
42 | int64_t last_pts; | ||
43 | int do_video; | ||
44 | int do_phasing_detection; | ||
45 | int w, h; | ||
46 | AVRational frame_rate; | ||
47 | int contrast[4]; | ||
48 | uint8_t *mpc_str; | ||
49 | uint8_t mpc[4]; | ||
50 | int draw_median_phase; | ||
51 | int is_mono; | ||
52 | int is_out_phase; | ||
53 | int start_mono_presence; | ||
54 | int start_out_phase_presence; | ||
55 | int nb_samples; | ||
56 | float tolerance; | ||
57 | float angle; | ||
58 | float phase; | ||
59 | AVRational time_base; | ||
60 | int64_t duration; | ||
61 | int64_t frame_end; | ||
62 | int64_t mono_idx[2]; | ||
63 | int64_t out_phase_idx[2]; | ||
64 | } AudioPhaseMeterContext; | ||
65 | |||
66 | #define MAX_DURATION (24*60*60*1000000LL) | ||
67 | #define OFFSET(x) offsetof(AudioPhaseMeterContext, x) | ||
68 | #define FLAGS AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_VIDEO_PARAM | ||
69 | #define get_duration(index) (index[1] - index[0]) | ||
70 | |||
71 | static const AVOption aphasemeter_options[] = { | ||
72 | { "rate", "set video rate", OFFSET(frame_rate), AV_OPT_TYPE_VIDEO_RATE, {.str="25"}, 0, INT_MAX, FLAGS }, | ||
73 | { "r", "set video rate", OFFSET(frame_rate), AV_OPT_TYPE_VIDEO_RATE, {.str="25"}, 0, INT_MAX, FLAGS }, | ||
74 | { "size", "set video size", OFFSET(w), AV_OPT_TYPE_IMAGE_SIZE, {.str="800x400"}, 0, 0, FLAGS }, | ||
75 | { "s", "set video size", OFFSET(w), AV_OPT_TYPE_IMAGE_SIZE, {.str="800x400"}, 0, 0, FLAGS }, | ||
76 | { "rc", "set red contrast", OFFSET(contrast[0]), AV_OPT_TYPE_INT, {.i64=2}, 0, 255, FLAGS }, | ||
77 | { "gc", "set green contrast", OFFSET(contrast[1]), AV_OPT_TYPE_INT, {.i64=7}, 0, 255, FLAGS }, | ||
78 | { "bc", "set blue contrast", OFFSET(contrast[2]), AV_OPT_TYPE_INT, {.i64=1}, 0, 255, FLAGS }, | ||
79 | { "mpc", "set median phase color", OFFSET(mpc_str), AV_OPT_TYPE_STRING, {.str = "none"}, 0, 0, FLAGS }, | ||
80 | { "video", "set video output", OFFSET(do_video), AV_OPT_TYPE_BOOL, {.i64 = 1}, 0, 1, FLAGS }, | ||
81 | { "phasing", "set mono and out-of-phase detection output", OFFSET(do_phasing_detection), AV_OPT_TYPE_BOOL, {.i64 = 0}, 0, 1, FLAGS }, | ||
82 | { "tolerance", "set phase tolerance for mono detection", OFFSET(tolerance), AV_OPT_TYPE_FLOAT, {.dbl = 0.}, 0, 1, FLAGS }, | ||
83 | { "t", "set phase tolerance for mono detection", OFFSET(tolerance), AV_OPT_TYPE_FLOAT, {.dbl = 0.}, 0, 1, FLAGS }, | ||
84 | { "angle", "set angle threshold for out-of-phase detection", OFFSET(angle), AV_OPT_TYPE_FLOAT, {.dbl = 170.}, 90, 180, FLAGS }, | ||
85 | { "a", "set angle threshold for out-of-phase detection", OFFSET(angle), AV_OPT_TYPE_FLOAT, {.dbl = 170.}, 90, 180, FLAGS }, | ||
86 | { "duration", "set minimum mono or out-of-phase duration in seconds", OFFSET(duration), AV_OPT_TYPE_DURATION, {.i64=2000000}, 0, MAX_DURATION, FLAGS }, | ||
87 | { "d", "set minimum mono or out-of-phase duration in seconds", OFFSET(duration), AV_OPT_TYPE_DURATION, {.i64=2000000}, 0, MAX_DURATION, FLAGS }, | ||
88 | { NULL } | ||
89 | }; | ||
90 | |||
91 | AVFILTER_DEFINE_CLASS(aphasemeter); | ||
92 | |||
93 | 2 | static int query_formats(const AVFilterContext *ctx, | |
94 | AVFilterFormatsConfig **cfg_in, | ||
95 | AVFilterFormatsConfig **cfg_out) | ||
96 | { | ||
97 | 2 | const AudioPhaseMeterContext *s = ctx->priv; | |
98 | 2 | AVFilterFormats *formats = NULL; | |
99 | static const enum AVSampleFormat sample_fmts[] = { AV_SAMPLE_FMT_FLT, AV_SAMPLE_FMT_NONE }; | ||
100 | static const enum AVPixelFormat pix_fmts[] = { AV_PIX_FMT_RGBA, AV_PIX_FMT_NONE }; | ||
101 | static const AVChannelLayout layouts[] = { | ||
102 | AV_CHANNEL_LAYOUT_STEREO, | ||
103 | { .nb_channels = 0 }, | ||
104 | }; | ||
105 | int ret; | ||
106 | |||
107 | 2 | formats = ff_make_format_list(sample_fmts); | |
108 |
2/4✓ Branch 1 taken 2 times.
✗ Branch 2 not taken.
✗ Branch 3 not taken.
✓ Branch 4 taken 2 times.
|
4 | if ((ret = ff_formats_ref(formats, &cfg_in[0]->formats)) < 0 || |
109 | 2 | (ret = ff_formats_ref(formats, &cfg_out[0]->formats)) < 0) | |
110 | ✗ | return ret; | |
111 | |||
112 | 2 | ret = ff_set_common_channel_layouts_from_list2(ctx, cfg_in, cfg_out, layouts); | |
113 |
1/2✗ Branch 0 not taken.
✓ Branch 1 taken 2 times.
|
2 | if (ret < 0) |
114 | ✗ | return ret; | |
115 | |||
116 |
1/2✗ Branch 0 not taken.
✓ Branch 1 taken 2 times.
|
2 | if (s->do_video) { |
117 | ✗ | formats = ff_make_format_list(pix_fmts); | |
118 | ✗ | if ((ret = ff_formats_ref(formats, &cfg_out[1]->formats)) < 0) | |
119 | ✗ | return ret; | |
120 | } | ||
121 | |||
122 | 2 | return 0; | |
123 | } | ||
124 | |||
125 | 2 | static int config_input(AVFilterLink *inlink) | |
126 | { | ||
127 | 2 | AVFilterContext *ctx = inlink->dst; | |
128 | 2 | AudioPhaseMeterContext *s = ctx->priv; | |
129 | 2 | s->duration = av_rescale(s->duration, inlink->sample_rate, AV_TIME_BASE); | |
130 | |||
131 |
1/2✗ Branch 0 not taken.
✓ Branch 1 taken 2 times.
|
2 | if (s->do_video) |
132 | ✗ | s->nb_samples = FFMAX(1, av_rescale(inlink->sample_rate, s->frame_rate.den, s->frame_rate.num)); | |
133 | |||
134 | 2 | return 0; | |
135 | } | ||
136 | |||
137 | ✗ | static int config_video_output(AVFilterLink *outlink) | |
138 | { | ||
139 | ✗ | AVFilterContext *ctx = outlink->src; | |
140 | ✗ | AudioPhaseMeterContext *s = ctx->priv; | |
141 | ✗ | FilterLink *l = ff_filter_link(outlink); | |
142 | |||
143 | ✗ | s->last_pts = AV_NOPTS_VALUE; | |
144 | |||
145 | ✗ | outlink->w = s->w; | |
146 | ✗ | outlink->h = s->h; | |
147 | ✗ | outlink->sample_aspect_ratio = (AVRational){1,1}; | |
148 | ✗ | l->frame_rate = s->frame_rate; | |
149 | ✗ | outlink->time_base = av_inv_q(l->frame_rate); | |
150 | |||
151 | ✗ | if (!strcmp(s->mpc_str, "none")) | |
152 | ✗ | s->draw_median_phase = 0; | |
153 | ✗ | else if (av_parse_color(s->mpc, s->mpc_str, -1, ctx) >= 0) | |
154 | ✗ | s->draw_median_phase = 1; | |
155 | else | ||
156 | ✗ | return AVERROR(EINVAL); | |
157 | |||
158 | ✗ | return 0; | |
159 | } | ||
160 | |||
161 | 96000 | static inline int get_x(float phase, int w) | |
162 | { | ||
163 | 96000 | return (phase + 1.f) / 2.f * (w - 1.f); | |
164 | } | ||
165 | |||
166 | 58 | static inline void add_metadata(AVFrame *insamples, const char *key, char *value) | |
167 | { | ||
168 | char buf[128]; | ||
169 | |||
170 | 58 | snprintf(buf, sizeof(buf), "lavfi.aphasemeter.%s", key); | |
171 | 58 | av_dict_set(&insamples->metadata, buf, value, 0); | |
172 | 58 | } | |
173 | |||
174 | ✗ | static inline void update_mono_detection(AudioPhaseMeterContext *s, AVFrame *insamples, int mono_measurement) | |
175 | { | ||
176 | int64_t mono_duration; | ||
177 | ✗ | if (!s->is_mono && mono_measurement) { | |
178 | ✗ | s->is_mono = 1; | |
179 | ✗ | s->start_mono_presence = 1; | |
180 | ✗ | s->mono_idx[0] = insamples->pts; | |
181 | } | ||
182 | ✗ | if (s->is_mono && mono_measurement && s->start_mono_presence) { | |
183 | ✗ | s->mono_idx[1] = s->frame_end; | |
184 | ✗ | mono_duration = get_duration(s->mono_idx); | |
185 | ✗ | if (mono_duration >= s->duration) { | |
186 | ✗ | add_metadata(insamples, "mono_start", av_ts2timestr(s->mono_idx[0], &s->time_base)); | |
187 | ✗ | av_log(s, AV_LOG_INFO, "mono_start: %s\n", av_ts2timestr(s->mono_idx[0], &s->time_base)); | |
188 | ✗ | s->start_mono_presence = 0; | |
189 | } | ||
190 | } | ||
191 | ✗ | if (s->is_mono && !mono_measurement) { | |
192 | ✗ | s->mono_idx[1] = insamples ? insamples->pts : s->frame_end; | |
193 | ✗ | mono_duration = get_duration(s->mono_idx); | |
194 | ✗ | if (mono_duration >= s->duration) { | |
195 | ✗ | if (insamples) { | |
196 | ✗ | add_metadata(insamples, "mono_end", av_ts2timestr(s->mono_idx[1], &s->time_base)); | |
197 | ✗ | add_metadata(insamples, "mono_duration", av_ts2timestr(mono_duration, &s->time_base)); | |
198 | } | ||
199 | ✗ | av_log(s, AV_LOG_INFO, "mono_end: %s | mono_duration: %s\n", av_ts2timestr(s->mono_idx[1], &s->time_base), av_ts2timestr(mono_duration, &s->time_base)); | |
200 | } | ||
201 | ✗ | s->is_mono = 0; | |
202 | } | ||
203 | ✗ | } | |
204 | |||
205 | ✗ | static inline void update_out_phase_detection(AudioPhaseMeterContext *s, AVFrame *insamples, int out_phase_measurement) | |
206 | { | ||
207 | int64_t out_phase_duration; | ||
208 | ✗ | if (!s->is_out_phase && out_phase_measurement) { | |
209 | ✗ | s->is_out_phase = 1; | |
210 | ✗ | s->start_out_phase_presence = 1; | |
211 | ✗ | s->out_phase_idx[0] = insamples->pts; | |
212 | } | ||
213 | ✗ | if (s->is_out_phase && out_phase_measurement && s->start_out_phase_presence) { | |
214 | ✗ | s->out_phase_idx[1] = s->frame_end; | |
215 | ✗ | out_phase_duration = get_duration(s->out_phase_idx); | |
216 | ✗ | if (out_phase_duration >= s->duration) { | |
217 | ✗ | add_metadata(insamples, "out_phase_start", av_ts2timestr(s->out_phase_idx[0], &s->time_base)); | |
218 | ✗ | av_log(s, AV_LOG_INFO, "out_phase_start: %s\n", av_ts2timestr(s->out_phase_idx[0], &s->time_base)); | |
219 | ✗ | s->start_out_phase_presence = 0; | |
220 | } | ||
221 | } | ||
222 | ✗ | if (s->is_out_phase && !out_phase_measurement) { | |
223 | ✗ | s->out_phase_idx[1] = insamples ? insamples->pts : s->frame_end; | |
224 | ✗ | out_phase_duration = get_duration(s->out_phase_idx); | |
225 | ✗ | if (out_phase_duration >= s->duration) { | |
226 | ✗ | if (insamples) { | |
227 | ✗ | add_metadata(insamples, "out_phase_end", av_ts2timestr(s->out_phase_idx[1], &s->time_base)); | |
228 | ✗ | add_metadata(insamples, "out_phase_duration", av_ts2timestr(out_phase_duration, &s->time_base)); | |
229 | } | ||
230 | ✗ | av_log(s, AV_LOG_INFO, "out_phase_end: %s | out_phase_duration: %s\n", av_ts2timestr(s->out_phase_idx[1], &s->time_base), av_ts2timestr(out_phase_duration, &s->time_base)); | |
231 | } | ||
232 | ✗ | s->is_out_phase = 0; | |
233 | } | ||
234 | ✗ | } | |
235 | |||
236 | 58 | static int filter_frame(AVFilterLink *inlink, AVFrame *in) | |
237 | { | ||
238 | 58 | AVFilterContext *ctx = inlink->dst; | |
239 | 58 | AudioPhaseMeterContext *s = ctx->priv; | |
240 |
1/2✗ Branch 0 not taken.
✓ Branch 1 taken 58 times.
|
58 | AVFilterLink *outlink = s->do_video ? ctx->outputs[1] : NULL; |
241 | 58 | AVFilterLink *aoutlink = ctx->outputs[0]; | |
242 | AVDictionary **metadata; | ||
243 | 58 | const int rc = s->contrast[0]; | |
244 | 58 | const int gc = s->contrast[1]; | |
245 | 58 | const int bc = s->contrast[2]; | |
246 | 58 | float fphase = 0; | |
247 | AVFrame *out; | ||
248 | uint8_t *dst; | ||
249 | int i, ret; | ||
250 | int mono_measurement; | ||
251 | int out_phase_measurement; | ||
252 | 58 | float tolerance = 1.0f - s->tolerance; | |
253 | 58 | float angle = cosf(s->angle/180.0f*M_PIf); | |
254 | int64_t new_pts; | ||
255 | |||
256 |
1/6✗ Branch 0 not taken.
✓ Branch 1 taken 58 times.
✗ Branch 2 not taken.
✗ Branch 3 not taken.
✗ Branch 4 not taken.
✗ Branch 5 not taken.
|
58 | if (s->do_video && (!s->out || s->out->width != outlink->w || |
257 | ✗ | s->out->height != outlink->h)) { | |
258 | ✗ | av_frame_free(&s->out); | |
259 | ✗ | s->out = ff_get_video_buffer(outlink, outlink->w, outlink->h); | |
260 | ✗ | if (!s->out) { | |
261 | ✗ | ret = AVERROR(ENOMEM); | |
262 | ✗ | goto fail; | |
263 | } | ||
264 | |||
265 | ✗ | out = s->out; | |
266 | ✗ | for (i = 0; i < outlink->h; i++) | |
267 | ✗ | memset(out->data[0] + i * out->linesize[0], 0, outlink->w * 4); | |
268 |
1/2✗ Branch 0 not taken.
✓ Branch 1 taken 58 times.
|
58 | } else if (s->do_video) { |
269 | ✗ | ret = ff_inlink_make_frame_writable(outlink, &s->out); | |
270 | ✗ | if (ret < 0) | |
271 | ✗ | goto fail; | |
272 | ✗ | out = s->out; | |
273 | ✗ | for (i = outlink->h - 1; i >= 10; i--) | |
274 | ✗ | memmove(out->data[0] + (i ) * out->linesize[0], | |
275 | ✗ | out->data[0] + (i-1) * out->linesize[0], | |
276 | ✗ | outlink->w * 4); | |
277 | ✗ | for (i = 0; i < outlink->w; i++) | |
278 | ✗ | AV_WL32(out->data[0] + i * 4, 0); | |
279 | } | ||
280 | |||
281 |
2/2✓ Branch 0 taken 96000 times.
✓ Branch 1 taken 58 times.
|
96058 | for (i = 0; i < in->nb_samples; i++) { |
282 | 96000 | const float *src = (float *)in->data[0] + i * 2; | |
283 | 96000 | const float f = src[0] * src[1] / (src[0]*src[0] + src[1] * src[1]) * 2; | |
284 |
2/2✓ Branch 0 taken 95999 times.
✓ Branch 1 taken 1 times.
|
96000 | const float phase = isnan(f) ? 1 : f; |
285 | 96000 | const int x = get_x(phase, s->w); | |
286 | |||
287 |
1/2✗ Branch 0 not taken.
✓ Branch 1 taken 96000 times.
|
96000 | if (s->do_video) { |
288 | ✗ | dst = out->data[0] + x * 4; | |
289 | ✗ | dst[0] = FFMIN(255, dst[0] + rc); | |
290 | ✗ | dst[1] = FFMIN(255, dst[1] + gc); | |
291 | ✗ | dst[2] = FFMIN(255, dst[2] + bc); | |
292 | ✗ | dst[3] = 255; | |
293 | } | ||
294 | 96000 | fphase += phase; | |
295 | } | ||
296 | 58 | fphase /= in->nb_samples; | |
297 | 58 | s->phase = fphase; | |
298 | |||
299 |
1/2✗ Branch 0 not taken.
✓ Branch 1 taken 58 times.
|
58 | if (s->do_video) { |
300 | ✗ | if (s->draw_median_phase) { | |
301 | ✗ | dst = out->data[0] + get_x(fphase, s->w) * 4; | |
302 | ✗ | AV_WL32(dst, AV_RL32(s->mpc)); | |
303 | } | ||
304 | |||
305 | ✗ | for (i = 1; i < 10 && i < outlink->h; i++) | |
306 | ✗ | memcpy(out->data[0] + i * out->linesize[0], out->data[0], outlink->w * 4); | |
307 | } | ||
308 | |||
309 | 58 | metadata = &in->metadata; | |
310 |
1/2✓ Branch 0 taken 58 times.
✗ Branch 1 not taken.
|
58 | if (metadata) { |
311 | uint8_t value[128]; | ||
312 | |||
313 | 58 | snprintf(value, sizeof(value), "%f", fphase); | |
314 | 58 | add_metadata(in, "phase", value); | |
315 | } | ||
316 | |||
317 |
1/2✗ Branch 0 not taken.
✓ Branch 1 taken 58 times.
|
58 | if (s->do_phasing_detection) { |
318 | ✗ | s->time_base = inlink->time_base; | |
319 | ✗ | s->frame_end = in->pts + av_rescale_q(in->nb_samples, | |
320 | ✗ | (AVRational){ 1, in->sample_rate }, inlink->time_base); | |
321 | |||
322 | ✗ | mono_measurement = (tolerance - fphase) < FLT_EPSILON; | |
323 | ✗ | out_phase_measurement = (angle - fphase) > FLT_EPSILON; | |
324 | |||
325 | ✗ | update_mono_detection(s, in, mono_measurement); | |
326 | ✗ | update_out_phase_detection(s, in, out_phase_measurement); | |
327 | } | ||
328 | |||
329 |
1/2✗ Branch 0 not taken.
✓ Branch 1 taken 58 times.
|
58 | if (s->do_video) |
330 | ✗ | new_pts = av_rescale_q(in->pts, inlink->time_base, outlink->time_base); | |
331 |
1/4✗ Branch 0 not taken.
✓ Branch 1 taken 58 times.
✗ Branch 2 not taken.
✗ Branch 3 not taken.
|
58 | if (s->do_video && new_pts != s->last_pts) { |
332 | AVFrame *clone; | ||
333 | |||
334 | ✗ | s->out->pts = s->last_pts = new_pts; | |
335 | ✗ | s->out->duration = 1; | |
336 | |||
337 | ✗ | clone = av_frame_clone(s->out); | |
338 | ✗ | if (!clone) { | |
339 | ✗ | ret = AVERROR(ENOMEM); | |
340 | ✗ | goto fail; | |
341 | } | ||
342 | ✗ | ret = ff_filter_frame(outlink, clone); | |
343 | ✗ | if (ret < 0) | |
344 | ✗ | goto fail; | |
345 | } | ||
346 | 58 | s->in = NULL; | |
347 | 58 | return ff_filter_frame(aoutlink, in); | |
348 | ✗ | fail: | |
349 | ✗ | av_frame_free(&in); | |
350 | ✗ | s->in = NULL; | |
351 | ✗ | return ret; | |
352 | } | ||
353 | |||
354 | 120 | static int activate(AVFilterContext *ctx) | |
355 | { | ||
356 | 120 | AVFilterLink *inlink = ctx->inputs[0]; | |
357 | 120 | AVFilterLink *outlink = ctx->outputs[0]; | |
358 | 120 | AudioPhaseMeterContext *s = ctx->priv; | |
359 | int ret; | ||
360 | |||
361 |
1/2✗ Branch 1 not taken.
✓ Branch 2 taken 120 times.
|
120 | FF_FILTER_FORWARD_STATUS_BACK(outlink, inlink); |
362 |
1/2✗ Branch 0 not taken.
✓ Branch 1 taken 120 times.
|
120 | if (s->do_video) |
363 | ✗ | FF_FILTER_FORWARD_STATUS_BACK(ctx->outputs[1], inlink); | |
364 | |||
365 |
1/2✓ Branch 0 taken 120 times.
✗ Branch 1 not taken.
|
120 | if (!s->in) { |
366 |
1/2✗ Branch 0 not taken.
✓ Branch 1 taken 120 times.
|
120 | if (s->nb_samples > 0) |
367 | ✗ | ret = ff_inlink_consume_samples(inlink, s->nb_samples, s->nb_samples, &s->in); | |
368 | else | ||
369 | 120 | ret = ff_inlink_consume_frame(inlink, &s->in); | |
370 |
1/2✗ Branch 0 not taken.
✓ Branch 1 taken 120 times.
|
120 | if (ret < 0) |
371 | ✗ | return ret; | |
372 |
2/2✓ Branch 0 taken 58 times.
✓ Branch 1 taken 62 times.
|
120 | if (ret > 0) |
373 | 58 | return filter_frame(inlink, s->in); | |
374 | } | ||
375 | |||
376 |
4/4✓ Branch 1 taken 2 times.
✓ Branch 2 taken 60 times.
✓ Branch 4 taken 2 times.
✓ Branch 5 taken 2 times.
|
64 | FF_FILTER_FORWARD_STATUS_ALL(inlink, ctx); |
377 |
1/2✓ Branch 1 taken 60 times.
✗ Branch 2 not taken.
|
60 | FF_FILTER_FORWARD_WANTED(outlink, inlink); |
378 | ✗ | if (s->do_video) | |
379 | ✗ | FF_FILTER_FORWARD_WANTED(ctx->outputs[1], inlink); | |
380 | |||
381 | ✗ | return FFERROR_NOT_READY; | |
382 | } | ||
383 | |||
384 | 2 | static av_cold void uninit(AVFilterContext *ctx) | |
385 | { | ||
386 | 2 | AudioPhaseMeterContext *s = ctx->priv; | |
387 | |||
388 |
1/2✗ Branch 0 not taken.
✓ Branch 1 taken 2 times.
|
2 | if (s->do_phasing_detection) { |
389 | ✗ | update_mono_detection(s, NULL, 0); | |
390 | ✗ | update_out_phase_detection(s, NULL, 0); | |
391 | } | ||
392 | 2 | av_frame_free(&s->out); | |
393 | 2 | } | |
394 | |||
395 | 2 | static av_cold int init(AVFilterContext *ctx) | |
396 | { | ||
397 | 2 | AudioPhaseMeterContext *s = ctx->priv; | |
398 | AVFilterPad pad; | ||
399 | int ret; | ||
400 | |||
401 | 2 | pad = (AVFilterPad){ | |
402 | .name = "out0", | ||
403 | .type = AVMEDIA_TYPE_AUDIO, | ||
404 | }; | ||
405 | 2 | ret = ff_append_outpad(ctx, &pad); | |
406 |
1/2✗ Branch 0 not taken.
✓ Branch 1 taken 2 times.
|
2 | if (ret < 0) |
407 | ✗ | return ret; | |
408 | |||
409 |
1/2✗ Branch 0 not taken.
✓ Branch 1 taken 2 times.
|
2 | if (s->do_video) { |
410 | ✗ | pad = (AVFilterPad){ | |
411 | .name = "out1", | ||
412 | .type = AVMEDIA_TYPE_VIDEO, | ||
413 | .config_props = config_video_output, | ||
414 | }; | ||
415 | ✗ | ret = ff_append_outpad(ctx, &pad); | |
416 | ✗ | if (ret < 0) | |
417 | ✗ | return ret; | |
418 | } | ||
419 | |||
420 | 2 | return 0; | |
421 | } | ||
422 | |||
423 | static const AVFilterPad inputs[] = { | ||
424 | { | ||
425 | .name = "default", | ||
426 | .type = AVMEDIA_TYPE_AUDIO, | ||
427 | .config_props = config_input, | ||
428 | }, | ||
429 | }; | ||
430 | |||
431 | const AVFilter ff_avf_aphasemeter = { | ||
432 | .name = "aphasemeter", | ||
433 | .description = NULL_IF_CONFIG_SMALL("Convert input audio to phase meter video output."), | ||
434 | .init = init, | ||
435 | .uninit = uninit, | ||
436 | .priv_size = sizeof(AudioPhaseMeterContext), | ||
437 | FILTER_INPUTS(inputs), | ||
438 | .activate = activate, | ||
439 | .outputs = NULL, | ||
440 | FILTER_QUERY_FUNC2(query_formats), | ||
441 | .priv_class = &aphasemeter_class, | ||
442 | .flags = AVFILTER_FLAG_DYNAMIC_OUTPUTS, | ||
443 | }; | ||
444 |