FFmpeg coverage


Directory: ../../../ffmpeg/
File: src/libavfilter/vf_mcdeint.c
Date: 2024-11-20 23:03:26
Exec Total Coverage
Lines: 98 118 83.1%
Functions: 3 3 100.0%
Branches: 55 69 79.7%

Line Branch Exec Source
1 /*
2 * Copyright (c) 2006 Michael Niedermayer <michaelni@gmx.at>
3 *
4 * This file is part of FFmpeg.
5 *
6 * FFmpeg is free software; you can redistribute it and/or modify
7 * it under the terms of the GNU General Public License as published by
8 * the Free Software Foundation; either version 2 of the License, or
9 * (at your option) any later version.
10 *
11 * FFmpeg is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14 * GNU General Public License for more details.
15 *
16 * You should have received a copy of the GNU General Public License along
17 * with FFmpeg; if not, write to the Free Software Foundation, Inc.,
18 * 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
19 */
20
21 /**
22 * @file
23 * Motion Compensation Deinterlacer
24 * Ported from MPlayer libmpcodecs/vf_mcdeint.c.
25 *
26 * Known Issues:
27 *
28 * The motion estimation is somewhat at the mercy of the input, if the
29 * input frames are created purely based on spatial interpolation then
30 * for example a thin black line or another random and not
31 * interpolateable pattern will cause problems.
32 * Note: completely ignoring the "unavailable" lines during motion
33 * estimation did not look any better, so the most obvious solution
34 * would be to improve tfields or penalize problematic motion vectors.
35 *
36 * If non iterative ME is used then snow currently ignores the OBMC
37 * window and as a result sometimes creates artifacts.
38 *
39 * Only past frames are used, we should ideally use future frames too,
40 * something like filtering the whole movie in forward and then
41 * backward direction seems like an interesting idea but the current
42 * filter framework is FAR from supporting such things.
43 *
44 * Combining the motion compensated image with the input image also is
45 * not as trivial as it seems, simple blindly taking even lines from
46 * one and odd ones from the other does not work at all as ME/MC
47 * sometimes has nothing in the previous frames which matches the
48 * current. The current algorithm has been found by trial and error
49 * and almost certainly can be improved...
50 */
51
52 #include "libavutil/opt.h"
53 #include "libavcodec/avcodec.h"
54 #include "avfilter.h"
55 #include "filters.h"
56 #include "video.h"
57
58 enum MCDeintMode {
59 MODE_FAST = 0,
60 MODE_MEDIUM,
61 MODE_SLOW,
62 MODE_EXTRA_SLOW,
63 MODE_NB,
64 };
65
66 enum MCDeintParity {
67 PARITY_TFF = 0, ///< top field first
68 PARITY_BFF = 1, ///< bottom field first
69 };
70
71 typedef struct MCDeintContext {
72 const AVClass *class;
73 int mode; ///< MCDeintMode
74 int parity; ///< MCDeintParity
75 int qp;
76 AVPacket *pkt;
77 AVFrame *frame_dec;
78 AVCodecContext *enc_ctx;
79 } MCDeintContext;
80
81 #define OFFSET(x) offsetof(MCDeintContext, x)
82 #define FLAGS AV_OPT_FLAG_VIDEO_PARAM|AV_OPT_FLAG_FILTERING_PARAM
83 #define CONST(name, help, val, u) { name, help, 0, AV_OPT_TYPE_CONST, {.i64=val}, INT_MIN, INT_MAX, FLAGS, .unit = u }
84
85 static const AVOption mcdeint_options[] = {
86 { "mode", "set mode", OFFSET(mode), AV_OPT_TYPE_INT, {.i64=MODE_FAST}, 0, MODE_NB-1, FLAGS, .unit="mode" },
87 CONST("fast", NULL, MODE_FAST, "mode"),
88 CONST("medium", NULL, MODE_MEDIUM, "mode"),
89 CONST("slow", NULL, MODE_SLOW, "mode"),
90 CONST("extra_slow", NULL, MODE_EXTRA_SLOW, "mode"),
91
92 { "parity", "set the assumed picture field parity", OFFSET(parity), AV_OPT_TYPE_INT, {.i64=PARITY_BFF}, -1, 1, FLAGS, .unit = "parity" },
93 CONST("tff", "assume top field first", PARITY_TFF, "parity"),
94 CONST("bff", "assume bottom field first", PARITY_BFF, "parity"),
95
96 { "qp", "set qp", OFFSET(qp), AV_OPT_TYPE_INT, {.i64=1}, INT_MIN, INT_MAX, FLAGS },
97 { NULL }
98 };
99
100 AVFILTER_DEFINE_CLASS(mcdeint);
101
102 2 static int config_props(AVFilterLink *inlink)
103 {
104 2 AVFilterContext *ctx = inlink->dst;
105 2 MCDeintContext *mcdeint = ctx->priv;
106 const AVCodec *enc;
107 AVCodecContext *enc_ctx;
108 2 AVDictionary *opts = NULL;
109 int ret;
110
111
1/2
✗ Branch 1 not taken.
✓ Branch 2 taken 2 times.
2 if (!(enc = avcodec_find_encoder(AV_CODEC_ID_SNOW))) {
112 av_log(ctx, AV_LOG_ERROR, "Snow encoder is not enabled in libavcodec\n");
113 return AVERROR(EINVAL);
114 }
115
116 2 mcdeint->pkt = av_packet_alloc();
117
1/2
✗ Branch 0 not taken.
✓ Branch 1 taken 2 times.
2 if (!mcdeint->pkt)
118 return AVERROR(ENOMEM);
119 2 mcdeint->frame_dec = av_frame_alloc();
120
1/2
✗ Branch 0 not taken.
✓ Branch 1 taken 2 times.
2 if (!mcdeint->frame_dec)
121 return AVERROR(ENOMEM);
122 2 mcdeint->enc_ctx = avcodec_alloc_context3(enc);
123
1/2
✗ Branch 0 not taken.
✓ Branch 1 taken 2 times.
2 if (!mcdeint->enc_ctx)
124 return AVERROR(ENOMEM);
125 2 enc_ctx = mcdeint->enc_ctx;
126 2 enc_ctx->width = inlink->w;
127 2 enc_ctx->height = inlink->h;
128 2 enc_ctx->time_base = (AVRational){1,25}; // meaningless
129 2 enc_ctx->gop_size = INT_MAX;
130 2 enc_ctx->max_b_frames = 0;
131 2 enc_ctx->pix_fmt = AV_PIX_FMT_YUV420P;
132 2 enc_ctx->flags = AV_CODEC_FLAG_QSCALE | AV_CODEC_FLAG_LOW_DELAY | AV_CODEC_FLAG_RECON_FRAME;
133 2 enc_ctx->strict_std_compliance = FF_COMPLIANCE_EXPERIMENTAL;
134 2 enc_ctx->global_quality = 1;
135 2 enc_ctx->me_cmp = enc_ctx->me_sub_cmp = FF_CMP_SAD;
136 2 enc_ctx->mb_cmp = FF_CMP_SSE;
137 2 av_dict_set(&opts, "memc_only", "1", 0);
138 2 av_dict_set(&opts, "no_bitstream", "1", 0);
139
140
2/5
✗ Branch 0 not taken.
✗ Branch 1 not taken.
✓ Branch 2 taken 1 times.
✓ Branch 3 taken 1 times.
✗ Branch 4 not taken.
2 switch (mcdeint->mode) {
141 case MODE_EXTRA_SLOW:
142 enc_ctx->refs = 3;
143 case MODE_SLOW:
144 av_dict_set(&opts, "motion_est", "iter", 0);
145 1 case MODE_MEDIUM:
146 1 enc_ctx->flags |= AV_CODEC_FLAG_4MV;
147 1 enc_ctx->dia_size = 2;
148 2 case MODE_FAST:
149 2 enc_ctx->flags |= AV_CODEC_FLAG_QPEL;
150 }
151
152 2 ret = avcodec_open2(enc_ctx, enc, &opts);
153 2 av_dict_free(&opts);
154
1/2
✗ Branch 0 not taken.
✓ Branch 1 taken 2 times.
2 if (ret < 0)
155 return ret;
156
157 2 return 0;
158 }
159
160 4 static av_cold void uninit(AVFilterContext *ctx)
161 {
162 4 MCDeintContext *mcdeint = ctx->priv;
163
164 4 av_packet_free(&mcdeint->pkt);
165 4 avcodec_free_context(&mcdeint->enc_ctx);
166 4 av_frame_free(&mcdeint->frame_dec);
167 4 }
168
169 60 static int filter_frame(AVFilterLink *inlink, AVFrame *inpic)
170 {
171 60 MCDeintContext *mcdeint = inlink->dst->priv;
172 60 AVFilterLink *outlink = inlink->dst->outputs[0];
173 60 AVFrame *outpic, *frame_dec = mcdeint->frame_dec;
174 60 AVPacket *pkt = mcdeint->pkt;
175 int x, y, i, ret;
176
177 60 outpic = ff_get_video_buffer(outlink, outlink->w, outlink->h);
178
1/2
✗ Branch 0 not taken.
✓ Branch 1 taken 60 times.
60 if (!outpic) {
179 av_frame_free(&inpic);
180 return AVERROR(ENOMEM);
181 }
182 60 av_frame_copy_props(outpic, inpic);
183 60 inpic->quality = mcdeint->qp * FF_QP2LAMBDA;
184
185 60 ret = avcodec_send_frame(mcdeint->enc_ctx, inpic);
186
1/2
✗ Branch 0 not taken.
✓ Branch 1 taken 60 times.
60 if (ret < 0) {
187 av_log(mcdeint->enc_ctx, AV_LOG_ERROR, "Error sending a frame for encoding\n");
188 goto end;
189 }
190 60 ret = avcodec_receive_packet(mcdeint->enc_ctx, pkt);
191
1/2
✗ Branch 0 not taken.
✓ Branch 1 taken 60 times.
60 if (ret < 0) {
192 av_log(mcdeint->enc_ctx, AV_LOG_ERROR, "Error receiving a packet from encoding\n");
193 goto end;
194 }
195 60 av_packet_unref(pkt);
196 60 ret = avcodec_receive_frame(mcdeint->enc_ctx, frame_dec);
197
1/2
✗ Branch 0 not taken.
✓ Branch 1 taken 60 times.
60 if (ret < 0) {
198 av_log(mcdeint->enc_ctx, AV_LOG_ERROR, "Error receiving a frame from encoding\n");
199 goto end;
200 }
201
202
2/2
✓ Branch 0 taken 180 times.
✓ Branch 1 taken 60 times.
240 for (i = 0; i < 3; i++) {
203 180 int is_chroma = !!i;
204 180 int w = AV_CEIL_RSHIFT(inlink->w, is_chroma);
205 180 int h = AV_CEIL_RSHIFT(inlink->h, is_chroma);
206 180 int fils = frame_dec->linesize[i];
207 180 int srcs = inpic ->linesize[i];
208 180 int dsts = outpic ->linesize[i];
209
210
2/2
✓ Branch 0 taken 69120 times.
✓ Branch 1 taken 180 times.
69300 for (y = 0; y < h; y++) {
211
2/2
✓ Branch 0 taken 34560 times.
✓ Branch 1 taken 34560 times.
69120 if ((y ^ mcdeint->parity) & 1) {
212
2/2
✓ Branch 0 taken 18662400 times.
✓ Branch 1 taken 34560 times.
18696960 for (x = 0; x < w; x++) {
213 18662400 uint8_t *filp = &frame_dec->data[i][x + y*fils];
214 18662400 uint8_t *srcp = &inpic ->data[i][x + y*srcs];
215 18662400 uint8_t *dstp = &outpic ->data[i][x + y*dsts];
216
217
4/4
✓ Branch 0 taken 18619200 times.
✓ Branch 1 taken 43200 times.
✓ Branch 2 taken 18576000 times.
✓ Branch 3 taken 43200 times.
18662400 if (y > 0 && y < h-1){
218
4/4
✓ Branch 0 taken 18472860 times.
✓ Branch 1 taken 103140 times.
✓ Branch 2 taken 103140 times.
✓ Branch 3 taken 18369720 times.
18576000 int is_edge = x < 3 || x > w-4;
219 18576000 int diff0 = filp[-fils] - srcp[-srcs];
220 18576000 int diff1 = filp[+fils] - srcp[+srcs];
221 18576000 int temp = filp[0];
222
223 #define DELTA(j) av_clip(j, -x, w-1-x)
224
225 #define GET_SCORE_EDGE(j)\
226 FFABS(srcp[-srcs+DELTA(-1+(j))] - srcp[+srcs+DELTA(-1-(j))])+\
227 FFABS(srcp[-srcs+DELTA(j) ] - srcp[+srcs+DELTA( -(j))])+\
228 FFABS(srcp[-srcs+DELTA(1+(j)) ] - srcp[+srcs+DELTA( 1-(j))])
229
230 #define GET_SCORE(j)\
231 FFABS(srcp[-srcs-1+(j)] - srcp[+srcs-1-(j)])+\
232 FFABS(srcp[-srcs +(j)] - srcp[+srcs -(j)])+\
233 FFABS(srcp[-srcs+1+(j)] - srcp[+srcs+1-(j)])
234
235 #define CHECK_EDGE(j)\
236 { int score = GET_SCORE_EDGE(j);\
237 if (score < spatial_score){\
238 spatial_score = score;\
239 diff0 = filp[-fils+DELTA(j)] - srcp[-srcs+DELTA(j)];\
240 diff1 = filp[+fils+DELTA(-(j))] - srcp[+srcs+DELTA(-(j))];\
241
242 #define CHECK(j)\
243 { int score = GET_SCORE(j);\
244 if (score < spatial_score){\
245 spatial_score= score;\
246 diff0 = filp[-fils+(j)] - srcp[-srcs+(j)];\
247 diff1 = filp[+fils-(j)] - srcp[+srcs-(j)];\
248
249
2/2
✓ Branch 0 taken 206280 times.
✓ Branch 1 taken 18369720 times.
18576000 if (is_edge) {
250 206280 int spatial_score = GET_SCORE_EDGE(0) - 1;
251
4/4
✓ Branch 0 taken 1708 times.
✓ Branch 1 taken 204572 times.
✓ Branch 2 taken 534 times.
✓ Branch 3 taken 1174 times.
206280 CHECK_EDGE(-1) CHECK_EDGE(-2) }} }}
252
4/4
✓ Branch 0 taken 1682 times.
✓ Branch 1 taken 204598 times.
✓ Branch 2 taken 550 times.
✓ Branch 3 taken 1132 times.
206280 CHECK_EDGE( 1) CHECK_EDGE( 2) }} }}
253 } else {
254 18369720 int spatial_score = GET_SCORE(0) - 1;
255
4/4
✓ Branch 0 taken 3205620 times.
✓ Branch 1 taken 15164100 times.
✓ Branch 2 taken 1352370 times.
✓ Branch 3 taken 1853250 times.
18369720 CHECK(-1) CHECK(-2) }} }}
256
4/4
✓ Branch 0 taken 2865070 times.
✓ Branch 1 taken 15504650 times.
✓ Branch 2 taken 1101458 times.
✓ Branch 3 taken 1763612 times.
18369720 CHECK( 1) CHECK( 2) }} }}
257 }
258
259
260
2/2
✓ Branch 0 taken 7866255 times.
✓ Branch 1 taken 10709745 times.
18576000 if (diff0 + diff1 > 0)
261 7866255 temp -= (diff0 + diff1 - FFABS(FFABS(diff0) - FFABS(diff1)) / 2) / 2;
262 else
263 10709745 temp -= (diff0 + diff1 + FFABS(FFABS(diff0) - FFABS(diff1)) / 2) / 2;
264
1/2
✗ Branch 0 not taken.
✓ Branch 1 taken 18576000 times.
18576000 *filp = *dstp = temp > 255U ? ~(temp>>31) : temp;
265 } else {
266 86400 *dstp = *filp;
267 }
268 }
269 }
270 }
271
272
2/2
✓ Branch 0 taken 69120 times.
✓ Branch 1 taken 180 times.
69300 for (y = 0; y < h; y++) {
273
2/2
✓ Branch 0 taken 34560 times.
✓ Branch 1 taken 34560 times.
69120 if (!((y ^ mcdeint->parity) & 1)) {
274
2/2
✓ Branch 0 taken 18662400 times.
✓ Branch 1 taken 34560 times.
18696960 for (x = 0; x < w; x++) {
275 18662400 frame_dec->data[i][x + y*fils] =
276 18662400 outpic ->data[i][x + y*dsts] = inpic->data[i][x + y*srcs];
277 }
278 }
279 }
280 }
281 60 mcdeint->parity ^= 1;
282
283 60 end:
284 60 av_packet_unref(pkt);
285 60 av_frame_free(&inpic);
286
1/2
✗ Branch 0 not taken.
✓ Branch 1 taken 60 times.
60 if (ret < 0) {
287 av_frame_free(&outpic);
288 return ret;
289 }
290 60 return ff_filter_frame(outlink, outpic);
291 }
292
293 static const AVFilterPad mcdeint_inputs[] = {
294 {
295 .name = "default",
296 .type = AVMEDIA_TYPE_VIDEO,
297 .filter_frame = filter_frame,
298 .config_props = config_props,
299 },
300 };
301
302 const AVFilter ff_vf_mcdeint = {
303 .name = "mcdeint",
304 .description = NULL_IF_CONFIG_SMALL("Apply motion compensating deinterlacing."),
305 .priv_size = sizeof(MCDeintContext),
306 .uninit = uninit,
307 FILTER_INPUTS(mcdeint_inputs),
308 FILTER_OUTPUTS(ff_video_default_filterpad),
309 FILTER_SINGLE_PIXFMT(AV_PIX_FMT_YUV420P),
310 .priv_class = &mcdeint_class,
311 };
312