Line | Branch | Exec | Source |
---|---|---|---|
1 | /* | ||
2 | * ColorMatrix v2.2 for Avisynth 2.5.x | ||
3 | * | ||
4 | * Copyright (C) 2006-2007 Kevin Stone | ||
5 | * | ||
6 | * ColorMatrix 1.x is Copyright (C) Wilbert Dijkhof | ||
7 | * | ||
8 | * This program is free software; you can redistribute it and/or modify it | ||
9 | * under the terms of the GNU General Public License as published by the | ||
10 | * Free Software Foundation; either version 2 of the License, or (at your | ||
11 | * option) any later version. | ||
12 | * | ||
13 | * This program is distributed in the hope that it will be useful, but | ||
14 | * OUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY | ||
15 | * or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public | ||
16 | * License for more details. | ||
17 | * | ||
18 | * You should have received a copy of the GNU General Public License | ||
19 | * along with this program; if not, write to the Free Software Foundation, | ||
20 | * Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA | ||
21 | */ | ||
22 | |||
23 | /** | ||
24 | * @file | ||
25 | * ColorMatrix 2.0 is based on the original ColorMatrix filter by Wilbert | ||
26 | * Dijkhof. It adds the ability to convert between any of: Rec.709, FCC, | ||
27 | * Rec.601, and SMPTE 240M. It also makes pre and post clipping optional, | ||
28 | * adds an option to use scaled or non-scaled coefficients, and more... | ||
29 | */ | ||
30 | |||
31 | #include <float.h> | ||
32 | #include "avfilter.h" | ||
33 | #include "filters.h" | ||
34 | #include "video.h" | ||
35 | #include "libavutil/opt.h" | ||
36 | #include "libavutil/pixdesc.h" | ||
37 | |||
38 | #define NS(n) ((n) < 0 ? (int)((n)*65536.0-0.5+DBL_EPSILON) : (int)((n)*65536.0+0.5)) | ||
39 | #define CB(n) av_clip_uint8(n) | ||
40 | |||
41 | static const double yuv_coeff_luma[5][3] = { | ||
42 | { +0.7152, +0.0722, +0.2126 }, // Rec.709 (0) | ||
43 | { +0.5900, +0.1100, +0.3000 }, // FCC (1) | ||
44 | { +0.5870, +0.1140, +0.2990 }, // Rec.601 (ITU-R BT.470-2/SMPTE 170M) (2) | ||
45 | { +0.7010, +0.0870, +0.2120 }, // SMPTE 240M (3) | ||
46 | { +0.6780, +0.0593, +0.2627 }, // Rec.2020 (4) | ||
47 | }; | ||
48 | |||
49 | enum ColorMode { | ||
50 | COLOR_MODE_NONE = -1, | ||
51 | COLOR_MODE_BT709, | ||
52 | COLOR_MODE_FCC, | ||
53 | COLOR_MODE_BT601, | ||
54 | COLOR_MODE_SMPTE240M, | ||
55 | COLOR_MODE_BT2020, | ||
56 | COLOR_MODE_COUNT | ||
57 | }; | ||
58 | |||
59 | typedef struct ColorMatrixContext { | ||
60 | const AVClass *class; | ||
61 | int yuv_convert[25][3][3]; | ||
62 | int interlaced; | ||
63 | int source, dest; ///< ColorMode | ||
64 | int mode; | ||
65 | int hsub, vsub; | ||
66 | } ColorMatrixContext; | ||
67 | |||
68 | typedef struct ThreadData { | ||
69 | AVFrame *dst; | ||
70 | const AVFrame *src; | ||
71 | int c2; | ||
72 | int c3; | ||
73 | int c4; | ||
74 | int c5; | ||
75 | int c6; | ||
76 | int c7; | ||
77 | } ThreadData; | ||
78 | |||
79 | #define OFFSET(x) offsetof(ColorMatrixContext, x) | ||
80 | #define FLAGS AV_OPT_FLAG_VIDEO_PARAM|AV_OPT_FLAG_FILTERING_PARAM | ||
81 | |||
82 | static const AVOption colormatrix_options[] = { | ||
83 | { "src", "set source color matrix", OFFSET(source), AV_OPT_TYPE_INT, {.i64=COLOR_MODE_NONE}, COLOR_MODE_NONE, COLOR_MODE_COUNT-1, .flags=FLAGS, .unit="color_mode" }, | ||
84 | { "dst", "set destination color matrix", OFFSET(dest), AV_OPT_TYPE_INT, {.i64=COLOR_MODE_NONE}, COLOR_MODE_NONE, COLOR_MODE_COUNT-1, .flags=FLAGS, .unit="color_mode" }, | ||
85 | { "bt709", "set BT.709 colorspace", 0, AV_OPT_TYPE_CONST, {.i64=COLOR_MODE_BT709}, .flags=FLAGS, .unit="color_mode" }, | ||
86 | { "fcc", "set FCC colorspace ", 0, AV_OPT_TYPE_CONST, {.i64=COLOR_MODE_FCC}, .flags=FLAGS, .unit="color_mode" }, | ||
87 | { "bt601", "set BT.601 colorspace", 0, AV_OPT_TYPE_CONST, {.i64=COLOR_MODE_BT601}, .flags=FLAGS, .unit="color_mode" }, | ||
88 | { "bt470", "set BT.470 colorspace", 0, AV_OPT_TYPE_CONST, {.i64=COLOR_MODE_BT601}, .flags=FLAGS, .unit="color_mode" }, | ||
89 | { "bt470bg", "set BT.470 colorspace", 0, AV_OPT_TYPE_CONST, {.i64=COLOR_MODE_BT601}, .flags=FLAGS, .unit="color_mode" }, | ||
90 | { "smpte170m", "set SMTPE-170M colorspace", 0, AV_OPT_TYPE_CONST, {.i64=COLOR_MODE_BT601}, .flags=FLAGS, .unit="color_mode" }, | ||
91 | { "smpte240m", "set SMPTE-240M colorspace", 0, AV_OPT_TYPE_CONST, {.i64=COLOR_MODE_SMPTE240M}, .flags=FLAGS, .unit="color_mode" }, | ||
92 | { "bt2020", "set BT.2020 colorspace", 0, AV_OPT_TYPE_CONST, {.i64=COLOR_MODE_BT2020}, .flags=FLAGS, .unit="color_mode" }, | ||
93 | { NULL } | ||
94 | }; | ||
95 | |||
96 | AVFILTER_DEFINE_CLASS(colormatrix); | ||
97 | |||
98 | #define ma m[0][0] | ||
99 | #define mb m[0][1] | ||
100 | #define mc m[0][2] | ||
101 | #define md m[1][0] | ||
102 | #define me m[1][1] | ||
103 | #define mf m[1][2] | ||
104 | #define mg m[2][0] | ||
105 | #define mh m[2][1] | ||
106 | #define mi m[2][2] | ||
107 | |||
108 | #define ima im[0][0] | ||
109 | #define imb im[0][1] | ||
110 | #define imc im[0][2] | ||
111 | #define imd im[1][0] | ||
112 | #define ime im[1][1] | ||
113 | #define imf im[1][2] | ||
114 | #define img im[2][0] | ||
115 | #define imh im[2][1] | ||
116 | #define imi im[2][2] | ||
117 | |||
118 | 120 | static void inverse3x3(double im[3][3], double m[3][3]) | |
119 | { | ||
120 | 120 | double det = ma * (me * mi - mf * mh) - mb * (md * mi - mf * mg) + mc * (md * mh - me * mg); | |
121 | 120 | det = 1.0 / det; | |
122 | 120 | ima = det * (me * mi - mf * mh); | |
123 | 120 | imb = det * (mc * mh - mb * mi); | |
124 | 120 | imc = det * (mb * mf - mc * me); | |
125 | 120 | imd = det * (mf * mg - md * mi); | |
126 | 120 | ime = det * (ma * mi - mc * mg); | |
127 | 120 | imf = det * (mc * md - ma * mf); | |
128 | 120 | img = det * (md * mh - me * mg); | |
129 | 120 | imh = det * (mb * mg - ma * mh); | |
130 | 120 | imi = det * (ma * me - mb * md); | |
131 | 120 | } | |
132 | |||
133 | 600 | static void solve_coefficients(double cm[3][3], double rgb[3][3], double yuv[3][3]) | |
134 | { | ||
135 | int i, j; | ||
136 |
2/2✓ Branch 0 taken 1800 times.
✓ Branch 1 taken 600 times.
|
2400 | for (i = 0; i < 3; i++) |
137 |
2/2✓ Branch 0 taken 5400 times.
✓ Branch 1 taken 1800 times.
|
7200 | for (j = 0; j < 3; j++) |
138 | 5400 | cm[i][j] = yuv[i][0] * rgb[0][j] + yuv[i][1] * rgb[1][j] + yuv[i][2] * rgb[2][j]; | |
139 | 600 | } | |
140 | |||
141 | 24 | static void calc_coefficients(AVFilterContext *ctx) | |
142 | { | ||
143 | 24 | ColorMatrixContext *color = ctx->priv; | |
144 | double yuv_coeff[5][3][3]; | ||
145 | double rgb_coeffd[5][3][3]; | ||
146 | double yuv_convertd[25][3][3]; | ||
147 | double bscale, rscale; | ||
148 | 24 | int v = 0; | |
149 | int i, j, k; | ||
150 |
2/2✓ Branch 0 taken 120 times.
✓ Branch 1 taken 24 times.
|
144 | for (i = 0; i < 5; i++) { |
151 | 120 | yuv_coeff[i][0][0] = yuv_coeff_luma[i][0]; | |
152 | 120 | yuv_coeff[i][0][1] = yuv_coeff_luma[i][1]; | |
153 | 120 | yuv_coeff[i][0][2] = yuv_coeff_luma[i][2]; | |
154 | 120 | bscale = 0.5 / (yuv_coeff[i][0][1] - 1.0); | |
155 | 120 | rscale = 0.5 / (yuv_coeff[i][0][2] - 1.0); | |
156 | 120 | yuv_coeff[i][1][0] = bscale * yuv_coeff[i][0][0]; | |
157 | 120 | yuv_coeff[i][1][1] = 0.5; | |
158 | 120 | yuv_coeff[i][1][2] = bscale * yuv_coeff[i][0][2]; | |
159 | 120 | yuv_coeff[i][2][0] = rscale * yuv_coeff[i][0][0]; | |
160 | 120 | yuv_coeff[i][2][1] = rscale * yuv_coeff[i][0][1]; | |
161 | 120 | yuv_coeff[i][2][2] = 0.5; | |
162 | } | ||
163 |
2/2✓ Branch 0 taken 120 times.
✓ Branch 1 taken 24 times.
|
144 | for (i = 0; i < 5; i++) |
164 | 120 | inverse3x3(rgb_coeffd[i], yuv_coeff[i]); | |
165 |
2/2✓ Branch 0 taken 120 times.
✓ Branch 1 taken 24 times.
|
144 | for (i = 0; i < 5; i++) { |
166 |
2/2✓ Branch 0 taken 600 times.
✓ Branch 1 taken 120 times.
|
720 | for (j = 0; j < 5; j++) { |
167 | 600 | solve_coefficients(yuv_convertd[v], rgb_coeffd[i], yuv_coeff[j]); | |
168 |
2/2✓ Branch 0 taken 1800 times.
✓ Branch 1 taken 600 times.
|
2400 | for (k = 0; k < 3; k++) { |
169 |
2/2✓ Branch 0 taken 312 times.
✓ Branch 1 taken 1488 times.
|
1800 | color->yuv_convert[v][k][0] = NS(yuv_convertd[v][k][0]); |
170 |
2/2✓ Branch 0 taken 576 times.
✓ Branch 1 taken 1224 times.
|
1800 | color->yuv_convert[v][k][1] = NS(yuv_convertd[v][k][1]); |
171 |
2/2✓ Branch 0 taken 528 times.
✓ Branch 1 taken 1272 times.
|
1800 | color->yuv_convert[v][k][2] = NS(yuv_convertd[v][k][2]); |
172 | } | ||
173 |
2/4✓ Branch 0 taken 600 times.
✗ Branch 1 not taken.
✓ Branch 2 taken 600 times.
✗ Branch 3 not taken.
|
600 | if (color->yuv_convert[v][0][0] != 65536 || color->yuv_convert[v][1][0] != 0 || |
174 |
1/2✗ Branch 0 not taken.
✓ Branch 1 taken 600 times.
|
600 | color->yuv_convert[v][2][0] != 0) { |
175 | ✗ | av_log(ctx, AV_LOG_ERROR, "error calculating conversion coefficients\n"); | |
176 | } | ||
177 | 600 | v++; | |
178 | } | ||
179 | } | ||
180 | 24 | } | |
181 | |||
182 | static const char * const color_modes[] = {"bt709", "fcc", "bt601", "smpte240m", "bt2020"}; | ||
183 | |||
184 | 24 | static av_cold int init(AVFilterContext *ctx) | |
185 | { | ||
186 | 24 | ColorMatrixContext *color = ctx->priv; | |
187 | |||
188 |
1/2✗ Branch 0 not taken.
✓ Branch 1 taken 24 times.
|
24 | if (color->dest == COLOR_MODE_NONE) { |
189 | ✗ | av_log(ctx, AV_LOG_ERROR, "Unspecified destination color space\n"); | |
190 | ✗ | return AVERROR(EINVAL); | |
191 | } | ||
192 | |||
193 |
1/2✗ Branch 0 not taken.
✓ Branch 1 taken 24 times.
|
24 | if (color->source == color->dest) { |
194 | ✗ | av_log(ctx, AV_LOG_ERROR, "Source and destination color space must not be identical\n"); | |
195 | ✗ | return AVERROR(EINVAL); | |
196 | } | ||
197 | |||
198 | 24 | calc_coefficients(ctx); | |
199 | |||
200 | 24 | return 0; | |
201 | } | ||
202 | |||
203 | ✗ | static int process_slice_uyvy422(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs) | |
204 | { | ||
205 | ✗ | const ThreadData *td = arg; | |
206 | ✗ | const AVFrame *src = td->src; | |
207 | ✗ | AVFrame *dst = td->dst; | |
208 | ✗ | const int height = src->height; | |
209 | ✗ | const int width = src->width*2; | |
210 | ✗ | const int src_pitch = src->linesize[0]; | |
211 | ✗ | const int dst_pitch = dst->linesize[0]; | |
212 | ✗ | const int slice_start = (height * jobnr ) / nb_jobs; | |
213 | ✗ | const int slice_end = (height * (jobnr+1)) / nb_jobs; | |
214 | ✗ | const unsigned char *srcp = src->data[0] + slice_start * src_pitch; | |
215 | ✗ | unsigned char *dstp = dst->data[0] + slice_start * dst_pitch; | |
216 | ✗ | const int c2 = td->c2; | |
217 | ✗ | const int c3 = td->c3; | |
218 | ✗ | const int c4 = td->c4; | |
219 | ✗ | const int c5 = td->c5; | |
220 | ✗ | const int c6 = td->c6; | |
221 | ✗ | const int c7 = td->c7; | |
222 | int x, y; | ||
223 | |||
224 | ✗ | for (y = slice_start; y < slice_end; y++) { | |
225 | ✗ | for (x = 0; x < width; x += 4) { | |
226 | ✗ | const int u = srcp[x + 0] - 128; | |
227 | ✗ | const int v = srcp[x + 2] - 128; | |
228 | ✗ | const int uvval = c2 * u + c3 * v + 1081344; | |
229 | ✗ | dstp[x + 0] = CB((c4 * u + c5 * v + 8421376) >> 16); | |
230 | ✗ | dstp[x + 1] = CB((65536 * (srcp[x + 1] - 16) + uvval) >> 16); | |
231 | ✗ | dstp[x + 2] = CB((c6 * u + c7 * v + 8421376) >> 16); | |
232 | ✗ | dstp[x + 3] = CB((65536 * (srcp[x + 3] - 16) + uvval) >> 16); | |
233 | } | ||
234 | ✗ | srcp += src_pitch; | |
235 | ✗ | dstp += dst_pitch; | |
236 | } | ||
237 | |||
238 | ✗ | return 0; | |
239 | } | ||
240 | |||
241 | ✗ | static int process_slice_yuv444p(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs) | |
242 | { | ||
243 | ✗ | const ThreadData *td = arg; | |
244 | ✗ | const AVFrame *src = td->src; | |
245 | ✗ | AVFrame *dst = td->dst; | |
246 | ✗ | const int height = src->height; | |
247 | ✗ | const int width = src->width; | |
248 | ✗ | const int slice_start = (height * jobnr ) / nb_jobs; | |
249 | ✗ | const int slice_end = (height * (jobnr+1)) / nb_jobs; | |
250 | ✗ | const int src_pitchY = src->linesize[0]; | |
251 | ✗ | const int src_pitchUV = src->linesize[1]; | |
252 | ✗ | const unsigned char *srcpU = src->data[1] + slice_start * src_pitchUV; | |
253 | ✗ | const unsigned char *srcpV = src->data[2] + slice_start * src_pitchUV; | |
254 | ✗ | const unsigned char *srcpY = src->data[0] + slice_start * src_pitchY; | |
255 | ✗ | const int dst_pitchY = dst->linesize[0]; | |
256 | ✗ | const int dst_pitchUV = dst->linesize[1]; | |
257 | ✗ | unsigned char *dstpU = dst->data[1] + slice_start * dst_pitchUV; | |
258 | ✗ | unsigned char *dstpV = dst->data[2] + slice_start * dst_pitchUV; | |
259 | ✗ | unsigned char *dstpY = dst->data[0] + slice_start * dst_pitchY; | |
260 | ✗ | const int c2 = td->c2; | |
261 | ✗ | const int c3 = td->c3; | |
262 | ✗ | const int c4 = td->c4; | |
263 | ✗ | const int c5 = td->c5; | |
264 | ✗ | const int c6 = td->c6; | |
265 | ✗ | const int c7 = td->c7; | |
266 | int x, y; | ||
267 | |||
268 | ✗ | for (y = slice_start; y < slice_end; y++) { | |
269 | ✗ | for (x = 0; x < width; x++) { | |
270 | ✗ | const int u = srcpU[x] - 128; | |
271 | ✗ | const int v = srcpV[x] - 128; | |
272 | ✗ | const int uvval = c2 * u + c3 * v + 1081344; | |
273 | ✗ | dstpY[x] = CB((65536 * (srcpY[x] - 16) + uvval) >> 16); | |
274 | ✗ | dstpU[x] = CB((c4 * u + c5 * v + 8421376) >> 16); | |
275 | ✗ | dstpV[x] = CB((c6 * u + c7 * v + 8421376) >> 16); | |
276 | } | ||
277 | ✗ | srcpY += src_pitchY; | |
278 | ✗ | dstpY += dst_pitchY; | |
279 | ✗ | srcpU += src_pitchUV; | |
280 | ✗ | srcpV += src_pitchUV; | |
281 | ✗ | dstpU += dst_pitchUV; | |
282 | ✗ | dstpV += dst_pitchUV; | |
283 | } | ||
284 | |||
285 | ✗ | return 0; | |
286 | } | ||
287 | |||
288 | ✗ | static int process_slice_yuv422p(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs) | |
289 | { | ||
290 | ✗ | const ThreadData *td = arg; | |
291 | ✗ | const AVFrame *src = td->src; | |
292 | ✗ | AVFrame *dst = td->dst; | |
293 | ✗ | const int height = src->height; | |
294 | ✗ | const int width = src->width; | |
295 | ✗ | const int slice_start = (height * jobnr ) / nb_jobs; | |
296 | ✗ | const int slice_end = (height * (jobnr+1)) / nb_jobs; | |
297 | ✗ | const int src_pitchY = src->linesize[0]; | |
298 | ✗ | const int src_pitchUV = src->linesize[1]; | |
299 | ✗ | const unsigned char *srcpU = src->data[1] + slice_start * src_pitchUV; | |
300 | ✗ | const unsigned char *srcpV = src->data[2] + slice_start * src_pitchUV; | |
301 | ✗ | const unsigned char *srcpY = src->data[0] + slice_start * src_pitchY; | |
302 | ✗ | const int dst_pitchY = dst->linesize[0]; | |
303 | ✗ | const int dst_pitchUV = dst->linesize[1]; | |
304 | ✗ | unsigned char *dstpU = dst->data[1] + slice_start * dst_pitchUV; | |
305 | ✗ | unsigned char *dstpV = dst->data[2] + slice_start * dst_pitchUV; | |
306 | ✗ | unsigned char *dstpY = dst->data[0] + slice_start * dst_pitchY; | |
307 | ✗ | const int c2 = td->c2; | |
308 | ✗ | const int c3 = td->c3; | |
309 | ✗ | const int c4 = td->c4; | |
310 | ✗ | const int c5 = td->c5; | |
311 | ✗ | const int c6 = td->c6; | |
312 | ✗ | const int c7 = td->c7; | |
313 | int x, y; | ||
314 | |||
315 | ✗ | for (y = slice_start; y < slice_end; y++) { | |
316 | ✗ | for (x = 0; x < width; x += 2) { | |
317 | ✗ | const int u = srcpU[x >> 1] - 128; | |
318 | ✗ | const int v = srcpV[x >> 1] - 128; | |
319 | ✗ | const int uvval = c2 * u + c3 * v + 1081344; | |
320 | ✗ | dstpY[x + 0] = CB((65536 * (srcpY[x + 0] - 16) + uvval) >> 16); | |
321 | ✗ | dstpY[x + 1] = CB((65536 * (srcpY[x + 1] - 16) + uvval) >> 16); | |
322 | ✗ | dstpU[x >> 1] = CB((c4 * u + c5 * v + 8421376) >> 16); | |
323 | ✗ | dstpV[x >> 1] = CB((c6 * u + c7 * v + 8421376) >> 16); | |
324 | } | ||
325 | ✗ | srcpY += src_pitchY; | |
326 | ✗ | dstpY += dst_pitchY; | |
327 | ✗ | srcpU += src_pitchUV; | |
328 | ✗ | srcpV += src_pitchUV; | |
329 | ✗ | dstpU += dst_pitchUV; | |
330 | ✗ | dstpV += dst_pitchUV; | |
331 | } | ||
332 | |||
333 | ✗ | return 0; | |
334 | } | ||
335 | |||
336 | 60 | static int process_slice_yuv420p(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs) | |
337 | { | ||
338 | 60 | const ThreadData *td = arg; | |
339 | 60 | const AVFrame *src = td->src; | |
340 | 60 | AVFrame *dst = td->dst; | |
341 | 60 | const int height = FFALIGN(src->height, 2) >> 1; | |
342 | 60 | const int width = src->width; | |
343 | 60 | const int slice_start = ((height * jobnr ) / nb_jobs) << 1; | |
344 | 60 | const int slice_end = ((height * (jobnr+1)) / nb_jobs) << 1; | |
345 | 60 | const int src_pitchY = src->linesize[0]; | |
346 | 60 | const int src_pitchUV = src->linesize[1]; | |
347 | 60 | const int dst_pitchY = dst->linesize[0]; | |
348 | 60 | const int dst_pitchUV = dst->linesize[1]; | |
349 | 60 | const unsigned char *srcpY = src->data[0] + src_pitchY * slice_start; | |
350 | 60 | const unsigned char *srcpU = src->data[1] + src_pitchUV * (slice_start >> 1); | |
351 | 60 | const unsigned char *srcpV = src->data[2] + src_pitchUV * (slice_start >> 1); | |
352 | 60 | const unsigned char *srcpN = src->data[0] + src_pitchY * (slice_start + 1); | |
353 | 60 | unsigned char *dstpU = dst->data[1] + dst_pitchUV * (slice_start >> 1); | |
354 | 60 | unsigned char *dstpV = dst->data[2] + dst_pitchUV * (slice_start >> 1); | |
355 | 60 | unsigned char *dstpY = dst->data[0] + dst_pitchY * slice_start; | |
356 | 60 | unsigned char *dstpN = dst->data[0] + dst_pitchY * (slice_start + 1); | |
357 | 60 | const int c2 = td->c2; | |
358 | 60 | const int c3 = td->c3; | |
359 | 60 | const int c4 = td->c4; | |
360 | 60 | const int c5 = td->c5; | |
361 | 60 | const int c6 = td->c6; | |
362 | 60 | const int c7 = td->c7; | |
363 | int x, y; | ||
364 | |||
365 |
2/2✓ Branch 0 taken 8640 times.
✓ Branch 1 taken 60 times.
|
8700 | for (y = slice_start; y < slice_end; y += 2) { |
366 |
2/2✓ Branch 0 taken 1520640 times.
✓ Branch 1 taken 8640 times.
|
1529280 | for (x = 0; x < width; x += 2) { |
367 | 1520640 | const int u = srcpU[x >> 1] - 128; | |
368 | 1520640 | const int v = srcpV[x >> 1] - 128; | |
369 | 1520640 | const int uvval = c2 * u + c3 * v + 1081344; | |
370 | 1520640 | dstpY[x + 0] = CB((65536 * (srcpY[x + 0] - 16) + uvval) >> 16); | |
371 | 1520640 | dstpY[x + 1] = CB((65536 * (srcpY[x + 1] - 16) + uvval) >> 16); | |
372 | 1520640 | dstpN[x + 0] = CB((65536 * (srcpN[x + 0] - 16) + uvval) >> 16); | |
373 | 1520640 | dstpN[x + 1] = CB((65536 * (srcpN[x + 1] - 16) + uvval) >> 16); | |
374 | 1520640 | dstpU[x >> 1] = CB((c4 * u + c5 * v + 8421376) >> 16); | |
375 | 1520640 | dstpV[x >> 1] = CB((c6 * u + c7 * v + 8421376) >> 16); | |
376 | } | ||
377 | 8640 | srcpY += src_pitchY << 1; | |
378 | 8640 | dstpY += dst_pitchY << 1; | |
379 | 8640 | srcpN += src_pitchY << 1; | |
380 | 8640 | dstpN += dst_pitchY << 1; | |
381 | 8640 | srcpU += src_pitchUV; | |
382 | 8640 | srcpV += src_pitchUV; | |
383 | 8640 | dstpU += dst_pitchUV; | |
384 | 8640 | dstpV += dst_pitchUV; | |
385 | } | ||
386 | |||
387 | 60 | return 0; | |
388 | } | ||
389 | |||
390 | 12 | static int config_input(AVFilterLink *inlink) | |
391 | { | ||
392 | 12 | AVFilterContext *ctx = inlink->dst; | |
393 | 12 | ColorMatrixContext *color = ctx->priv; | |
394 | 12 | const AVPixFmtDescriptor *pix_desc = av_pix_fmt_desc_get(inlink->format); | |
395 | |||
396 | 12 | color->hsub = pix_desc->log2_chroma_w; | |
397 | 12 | color->vsub = pix_desc->log2_chroma_h; | |
398 | |||
399 | 12 | av_log(ctx, AV_LOG_VERBOSE, "%s -> %s\n", | |
400 | 12 | color_modes[color->source], color_modes[color->dest]); | |
401 | |||
402 | 12 | return 0; | |
403 | } | ||
404 | |||
405 | 60 | static int filter_frame(AVFilterLink *link, AVFrame *in) | |
406 | { | ||
407 | 60 | AVFilterContext *ctx = link->dst; | |
408 | 60 | ColorMatrixContext *color = ctx->priv; | |
409 | 60 | AVFilterLink *outlink = ctx->outputs[0]; | |
410 | AVFrame *out; | ||
411 | 60 | ThreadData td = {0}; | |
412 | |||
413 | 60 | out = ff_get_video_buffer(outlink, outlink->w, outlink->h); | |
414 |
1/2✗ Branch 0 not taken.
✓ Branch 1 taken 60 times.
|
60 | if (!out) { |
415 | ✗ | av_frame_free(&in); | |
416 | ✗ | return AVERROR(ENOMEM); | |
417 | } | ||
418 | 60 | av_frame_copy_props(out, in); | |
419 | |||
420 |
1/2✗ Branch 0 not taken.
✓ Branch 1 taken 60 times.
|
60 | if (color->source == COLOR_MODE_NONE) { |
421 | ✗ | enum AVColorSpace cs = in->colorspace; | |
422 | enum ColorMode source; | ||
423 | |||
424 | ✗ | switch(cs) { | |
425 | ✗ | case AVCOL_SPC_BT709 : source = COLOR_MODE_BT709 ; break; | |
426 | ✗ | case AVCOL_SPC_FCC : source = COLOR_MODE_FCC ; break; | |
427 | ✗ | case AVCOL_SPC_SMPTE240M : source = COLOR_MODE_SMPTE240M ; break; | |
428 | ✗ | case AVCOL_SPC_BT470BG : source = COLOR_MODE_BT601 ; break; | |
429 | ✗ | case AVCOL_SPC_SMPTE170M : source = COLOR_MODE_BT601 ; break; | |
430 | ✗ | case AVCOL_SPC_BT2020_NCL: source = COLOR_MODE_BT2020 ; break; | |
431 | ✗ | case AVCOL_SPC_BT2020_CL : source = COLOR_MODE_BT2020 ; break; | |
432 | ✗ | default : | |
433 | ✗ | av_log(ctx, AV_LOG_ERROR, "Input frame does not specify a supported colorspace, and none has been specified as source either\n"); | |
434 | ✗ | av_frame_free(&out); | |
435 | ✗ | return AVERROR(EINVAL); | |
436 | } | ||
437 | ✗ | color->mode = source * 5 + color->dest; | |
438 | } else | ||
439 | 60 | color->mode = color->source * 5 + color->dest; | |
440 | |||
441 |
4/6✓ Branch 0 taken 15 times.
✓ Branch 1 taken 15 times.
✓ Branch 2 taken 15 times.
✓ Branch 3 taken 15 times.
✗ Branch 4 not taken.
✗ Branch 5 not taken.
|
60 | switch(color->dest) { |
442 | 15 | case COLOR_MODE_BT709 : out->colorspace = AVCOL_SPC_BT709 ; break; | |
443 | 15 | case COLOR_MODE_FCC : out->colorspace = AVCOL_SPC_FCC ; break; | |
444 | 15 | case COLOR_MODE_SMPTE240M: out->colorspace = AVCOL_SPC_SMPTE240M ; break; | |
445 | 15 | case COLOR_MODE_BT601 : out->colorspace = AVCOL_SPC_BT470BG ; break; | |
446 | ✗ | case COLOR_MODE_BT2020 : out->colorspace = AVCOL_SPC_BT2020_NCL; break; | |
447 | } | ||
448 | |||
449 | 60 | td.src = in; | |
450 | 60 | td.dst = out; | |
451 | 60 | td.c2 = color->yuv_convert[color->mode][0][1]; | |
452 | 60 | td.c3 = color->yuv_convert[color->mode][0][2]; | |
453 | 60 | td.c4 = color->yuv_convert[color->mode][1][1]; | |
454 | 60 | td.c5 = color->yuv_convert[color->mode][1][2]; | |
455 | 60 | td.c6 = color->yuv_convert[color->mode][2][1]; | |
456 | 60 | td.c7 = color->yuv_convert[color->mode][2][2]; | |
457 | |||
458 |
1/2✗ Branch 0 not taken.
✓ Branch 1 taken 60 times.
|
60 | if (in->format == AV_PIX_FMT_YUV444P) |
459 | ✗ | ff_filter_execute(ctx, process_slice_yuv444p, &td, NULL, | |
460 | ✗ | FFMIN(in->height, ff_filter_get_nb_threads(ctx))); | |
461 |
1/2✗ Branch 0 not taken.
✓ Branch 1 taken 60 times.
|
60 | else if (in->format == AV_PIX_FMT_YUV422P) |
462 | ✗ | ff_filter_execute(ctx, process_slice_yuv422p, &td, NULL, | |
463 | ✗ | FFMIN(in->height, ff_filter_get_nb_threads(ctx))); | |
464 |
1/2✓ Branch 0 taken 60 times.
✗ Branch 1 not taken.
|
60 | else if (in->format == AV_PIX_FMT_YUV420P) |
465 | 60 | ff_filter_execute(ctx, process_slice_yuv420p, &td, NULL, | |
466 |
1/2✓ Branch 0 taken 60 times.
✗ Branch 1 not taken.
|
60 | FFMIN(in->height / 2, ff_filter_get_nb_threads(ctx))); |
467 | else | ||
468 | ✗ | ff_filter_execute(ctx, process_slice_uyvy422, &td, NULL, | |
469 | ✗ | FFMIN(in->height, ff_filter_get_nb_threads(ctx))); | |
470 | |||
471 | 60 | av_frame_free(&in); | |
472 | 60 | return ff_filter_frame(outlink, out); | |
473 | } | ||
474 | |||
475 | static const AVFilterPad colormatrix_inputs[] = { | ||
476 | { | ||
477 | .name = "default", | ||
478 | .type = AVMEDIA_TYPE_VIDEO, | ||
479 | .config_props = config_input, | ||
480 | .filter_frame = filter_frame, | ||
481 | }, | ||
482 | }; | ||
483 | |||
484 | const AVFilter ff_vf_colormatrix = { | ||
485 | .name = "colormatrix", | ||
486 | .description = NULL_IF_CONFIG_SMALL("Convert color matrix."), | ||
487 | .priv_size = sizeof(ColorMatrixContext), | ||
488 | .init = init, | ||
489 | FILTER_INPUTS(colormatrix_inputs), | ||
490 | FILTER_OUTPUTS(ff_video_default_filterpad), | ||
491 | FILTER_PIXFMTS(AV_PIX_FMT_YUV444P, | ||
492 | AV_PIX_FMT_YUV422P, | ||
493 | AV_PIX_FMT_YUV420P, | ||
494 | AV_PIX_FMT_UYVY422), | ||
495 | .priv_class = &colormatrix_class, | ||
496 | .flags = AVFILTER_FLAG_SUPPORT_TIMELINE_GENERIC | AVFILTER_FLAG_SLICE_THREADS, | ||
497 | }; | ||
498 |