GCC Code Coverage Report | |||||||||||||||||||||
|
|||||||||||||||||||||
Line | Branch | Exec | Source |
1 |
/* |
||
2 |
* H.26L/H.264/AVC/JVT/14496-10/... decoder |
||
3 |
* Copyright (c) 2003 Michael Niedermayer <michaelni@gmx.at> |
||
4 |
* |
||
5 |
* This file is part of FFmpeg. |
||
6 |
* |
||
7 |
* FFmpeg is free software; you can redistribute it and/or |
||
8 |
* modify it under the terms of the GNU Lesser General Public |
||
9 |
* License as published by the Free Software Foundation; either |
||
10 |
* version 2.1 of the License, or (at your option) any later version. |
||
11 |
* |
||
12 |
* FFmpeg is distributed in the hope that it will be useful, |
||
13 |
* but WITHOUT ANY WARRANTY; without even the implied warranty of |
||
14 |
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
||
15 |
* Lesser General Public License for more details. |
||
16 |
* |
||
17 |
* You should have received a copy of the GNU Lesser General Public |
||
18 |
* License along with FFmpeg; if not, write to the Free Software |
||
19 |
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA |
||
20 |
*/ |
||
21 |
|||
22 |
/** |
||
23 |
* @file |
||
24 |
* H.264 / AVC / MPEG-4 part10 codec. |
||
25 |
* @author Michael Niedermayer <michaelni@gmx.at> |
||
26 |
*/ |
||
27 |
|||
28 |
#include "libavutil/avassert.h" |
||
29 |
#include "libavutil/display.h" |
||
30 |
#include "libavutil/imgutils.h" |
||
31 |
#include "libavutil/stereo3d.h" |
||
32 |
#include "internal.h" |
||
33 |
#include "cabac.h" |
||
34 |
#include "cabac_functions.h" |
||
35 |
#include "error_resilience.h" |
||
36 |
#include "avcodec.h" |
||
37 |
#include "h264.h" |
||
38 |
#include "h264dec.h" |
||
39 |
#include "h264data.h" |
||
40 |
#include "h264chroma.h" |
||
41 |
#include "h264_mvpred.h" |
||
42 |
#include "h264_ps.h" |
||
43 |
#include "golomb.h" |
||
44 |
#include "mathops.h" |
||
45 |
#include "mpegutils.h" |
||
46 |
#include "mpegvideo.h" |
||
47 |
#include "rectangle.h" |
||
48 |
#include "thread.h" |
||
49 |
|||
50 |
static const uint8_t field_scan[16+1] = { |
||
51 |
0 + 0 * 4, 0 + 1 * 4, 1 + 0 * 4, 0 + 2 * 4, |
||
52 |
0 + 3 * 4, 1 + 1 * 4, 1 + 2 * 4, 1 + 3 * 4, |
||
53 |
2 + 0 * 4, 2 + 1 * 4, 2 + 2 * 4, 2 + 3 * 4, |
||
54 |
3 + 0 * 4, 3 + 1 * 4, 3 + 2 * 4, 3 + 3 * 4, |
||
55 |
}; |
||
56 |
|||
57 |
static const uint8_t field_scan8x8[64+1] = { |
||
58 |
0 + 0 * 8, 0 + 1 * 8, 0 + 2 * 8, 1 + 0 * 8, |
||
59 |
1 + 1 * 8, 0 + 3 * 8, 0 + 4 * 8, 1 + 2 * 8, |
||
60 |
2 + 0 * 8, 1 + 3 * 8, 0 + 5 * 8, 0 + 6 * 8, |
||
61 |
0 + 7 * 8, 1 + 4 * 8, 2 + 1 * 8, 3 + 0 * 8, |
||
62 |
2 + 2 * 8, 1 + 5 * 8, 1 + 6 * 8, 1 + 7 * 8, |
||
63 |
2 + 3 * 8, 3 + 1 * 8, 4 + 0 * 8, 3 + 2 * 8, |
||
64 |
2 + 4 * 8, 2 + 5 * 8, 2 + 6 * 8, 2 + 7 * 8, |
||
65 |
3 + 3 * 8, 4 + 1 * 8, 5 + 0 * 8, 4 + 2 * 8, |
||
66 |
3 + 4 * 8, 3 + 5 * 8, 3 + 6 * 8, 3 + 7 * 8, |
||
67 |
4 + 3 * 8, 5 + 1 * 8, 6 + 0 * 8, 5 + 2 * 8, |
||
68 |
4 + 4 * 8, 4 + 5 * 8, 4 + 6 * 8, 4 + 7 * 8, |
||
69 |
5 + 3 * 8, 6 + 1 * 8, 6 + 2 * 8, 5 + 4 * 8, |
||
70 |
5 + 5 * 8, 5 + 6 * 8, 5 + 7 * 8, 6 + 3 * 8, |
||
71 |
7 + 0 * 8, 7 + 1 * 8, 6 + 4 * 8, 6 + 5 * 8, |
||
72 |
6 + 6 * 8, 6 + 7 * 8, 7 + 2 * 8, 7 + 3 * 8, |
||
73 |
7 + 4 * 8, 7 + 5 * 8, 7 + 6 * 8, 7 + 7 * 8, |
||
74 |
}; |
||
75 |
|||
76 |
static const uint8_t field_scan8x8_cavlc[64+1] = { |
||
77 |
0 + 0 * 8, 1 + 1 * 8, 2 + 0 * 8, 0 + 7 * 8, |
||
78 |
2 + 2 * 8, 2 + 3 * 8, 2 + 4 * 8, 3 + 3 * 8, |
||
79 |
3 + 4 * 8, 4 + 3 * 8, 4 + 4 * 8, 5 + 3 * 8, |
||
80 |
5 + 5 * 8, 7 + 0 * 8, 6 + 6 * 8, 7 + 4 * 8, |
||
81 |
0 + 1 * 8, 0 + 3 * 8, 1 + 3 * 8, 1 + 4 * 8, |
||
82 |
1 + 5 * 8, 3 + 1 * 8, 2 + 5 * 8, 4 + 1 * 8, |
||
83 |
3 + 5 * 8, 5 + 1 * 8, 4 + 5 * 8, 6 + 1 * 8, |
||
84 |
5 + 6 * 8, 7 + 1 * 8, 6 + 7 * 8, 7 + 5 * 8, |
||
85 |
0 + 2 * 8, 0 + 4 * 8, 0 + 5 * 8, 2 + 1 * 8, |
||
86 |
1 + 6 * 8, 4 + 0 * 8, 2 + 6 * 8, 5 + 0 * 8, |
||
87 |
3 + 6 * 8, 6 + 0 * 8, 4 + 6 * 8, 6 + 2 * 8, |
||
88 |
5 + 7 * 8, 6 + 4 * 8, 7 + 2 * 8, 7 + 6 * 8, |
||
89 |
1 + 0 * 8, 1 + 2 * 8, 0 + 6 * 8, 3 + 0 * 8, |
||
90 |
1 + 7 * 8, 3 + 2 * 8, 2 + 7 * 8, 4 + 2 * 8, |
||
91 |
3 + 7 * 8, 5 + 2 * 8, 4 + 7 * 8, 5 + 4 * 8, |
||
92 |
6 + 3 * 8, 6 + 5 * 8, 7 + 3 * 8, 7 + 7 * 8, |
||
93 |
}; |
||
94 |
|||
95 |
// zigzag_scan8x8_cavlc[i] = zigzag_scan8x8[(i/4) + 16*(i%4)] |
||
96 |
static const uint8_t zigzag_scan8x8_cavlc[64+1] = { |
||
97 |
0 + 0 * 8, 1 + 1 * 8, 1 + 2 * 8, 2 + 2 * 8, |
||
98 |
4 + 1 * 8, 0 + 5 * 8, 3 + 3 * 8, 7 + 0 * 8, |
||
99 |
3 + 4 * 8, 1 + 7 * 8, 5 + 3 * 8, 6 + 3 * 8, |
||
100 |
2 + 7 * 8, 6 + 4 * 8, 5 + 6 * 8, 7 + 5 * 8, |
||
101 |
1 + 0 * 8, 2 + 0 * 8, 0 + 3 * 8, 3 + 1 * 8, |
||
102 |
3 + 2 * 8, 0 + 6 * 8, 4 + 2 * 8, 6 + 1 * 8, |
||
103 |
2 + 5 * 8, 2 + 6 * 8, 6 + 2 * 8, 5 + 4 * 8, |
||
104 |
3 + 7 * 8, 7 + 3 * 8, 4 + 7 * 8, 7 + 6 * 8, |
||
105 |
0 + 1 * 8, 3 + 0 * 8, 0 + 4 * 8, 4 + 0 * 8, |
||
106 |
2 + 3 * 8, 1 + 5 * 8, 5 + 1 * 8, 5 + 2 * 8, |
||
107 |
1 + 6 * 8, 3 + 5 * 8, 7 + 1 * 8, 4 + 5 * 8, |
||
108 |
4 + 6 * 8, 7 + 4 * 8, 5 + 7 * 8, 6 + 7 * 8, |
||
109 |
0 + 2 * 8, 2 + 1 * 8, 1 + 3 * 8, 5 + 0 * 8, |
||
110 |
1 + 4 * 8, 2 + 4 * 8, 6 + 0 * 8, 4 + 3 * 8, |
||
111 |
0 + 7 * 8, 4 + 4 * 8, 7 + 2 * 8, 3 + 6 * 8, |
||
112 |
5 + 5 * 8, 6 + 5 * 8, 6 + 6 * 8, 7 + 7 * 8, |
||
113 |
}; |
||
114 |
|||
115 |
26804 |
static void release_unused_pictures(H264Context *h, int remove_current) |
|
116 |
{ |
||
117 |
int i; |
||
118 |
|||
119 |
/* release non reference frames */ |
||
120 |
✓✓ | 991748 |
for (i = 0; i < H264_MAX_PICTURE_COUNT; i++) { |
121 |
✓✓✓✓ ✓✓ |
964944 |
if (h->DPB[i].f->buf[0] && !h->DPB[i].reference && |
122 |
✓✓ | 2589 |
(remove_current || &h->DPB[i] != h->cur_pic_ptr)) { |
123 |
21437 |
ff_h264_unref_picture(h, &h->DPB[i]); |
|
124 |
} |
||
125 |
} |
||
126 |
26804 |
} |
|
127 |
|||
128 |
34216 |
static int alloc_scratch_buffers(H264SliceContext *sl, int linesize) |
|
129 |
{ |
||
130 |
34216 |
const H264Context *h = sl->h264; |
|
131 |
34216 |
int alloc_size = FFALIGN(FFABS(linesize) + 32, 32); |
|
132 |
|||
133 |
34216 |
av_fast_malloc(&sl->bipred_scratchpad, &sl->bipred_scratchpad_allocated, 16 * 6 * alloc_size); |
|
134 |
// edge emu needs blocksize + filter length - 1 |
||
135 |
// (= 21x21 for H.264) |
||
136 |
34216 |
av_fast_malloc(&sl->edge_emu_buffer, &sl->edge_emu_buffer_allocated, alloc_size * 2 * 21); |
|
137 |
|||
138 |
34216 |
av_fast_mallocz(&sl->top_borders[0], &sl->top_borders_allocated[0], |
|
139 |
34216 |
h->mb_width * 16 * 3 * sizeof(uint8_t) * 2); |
|
140 |
34216 |
av_fast_mallocz(&sl->top_borders[1], &sl->top_borders_allocated[1], |
|
141 |
34216 |
h->mb_width * 16 * 3 * sizeof(uint8_t) * 2); |
|
142 |
|||
143 |
✓✗✓✗ |
34216 |
if (!sl->bipred_scratchpad || !sl->edge_emu_buffer || |
144 |
✓✗✗✓ |
34216 |
!sl->top_borders[0] || !sl->top_borders[1]) { |
145 |
av_freep(&sl->bipred_scratchpad); |
||
146 |
av_freep(&sl->edge_emu_buffer); |
||
147 |
av_freep(&sl->top_borders[0]); |
||
148 |
av_freep(&sl->top_borders[1]); |
||
149 |
|||
150 |
sl->bipred_scratchpad_allocated = 0; |
||
151 |
sl->edge_emu_buffer_allocated = 0; |
||
152 |
sl->top_borders_allocated[0] = 0; |
||
153 |
sl->top_borders_allocated[1] = 0; |
||
154 |
return AVERROR(ENOMEM); |
||
155 |
} |
||
156 |
|||
157 |
34216 |
return 0; |
|
158 |
} |
||
159 |
|||
160 |
546 |
static int init_table_pools(H264Context *h) |
|
161 |
{ |
||
162 |
546 |
const int big_mb_num = h->mb_stride * (h->mb_height + 1) + 1; |
|
163 |
546 |
const int mb_array_size = h->mb_stride * h->mb_height; |
|
164 |
546 |
const int b4_stride = h->mb_width * 4 + 1; |
|
165 |
546 |
const int b4_array_size = b4_stride * h->mb_height * 4; |
|
166 |
|||
167 |
546 |
h->qscale_table_pool = av_buffer_pool_init(big_mb_num + h->mb_stride, |
|
168 |
av_buffer_allocz); |
||
169 |
546 |
h->mb_type_pool = av_buffer_pool_init((big_mb_num + h->mb_stride) * |
|
170 |
sizeof(uint32_t), av_buffer_allocz); |
||
171 |
546 |
h->motion_val_pool = av_buffer_pool_init(2 * (b4_array_size + 4) * |
|
172 |
sizeof(int16_t), av_buffer_allocz); |
||
173 |
546 |
h->ref_index_pool = av_buffer_pool_init(4 * mb_array_size, av_buffer_allocz); |
|
174 |
|||
175 |
✓✗✓✗ ✓✗ |
546 |
if (!h->qscale_table_pool || !h->mb_type_pool || !h->motion_val_pool || |
176 |
✗✓ | 546 |
!h->ref_index_pool) { |
177 |
av_buffer_pool_uninit(&h->qscale_table_pool); |
||
178 |
av_buffer_pool_uninit(&h->mb_type_pool); |
||
179 |
av_buffer_pool_uninit(&h->motion_val_pool); |
||
180 |
av_buffer_pool_uninit(&h->ref_index_pool); |
||
181 |
return AVERROR(ENOMEM); |
||
182 |
} |
||
183 |
|||
184 |
546 |
return 0; |
|
185 |
} |
||
186 |
|||
187 |
23472 |
static int alloc_picture(H264Context *h, H264Picture *pic) |
|
188 |
{ |
||
189 |
23472 |
int i, ret = 0; |
|
190 |
|||
191 |
✗✓ | 23472 |
av_assert0(!pic->f->data[0]); |
192 |
|||
193 |
23472 |
pic->tf.f = pic->f; |
|
194 |
23472 |
ret = ff_thread_get_buffer(h->avctx, &pic->tf, pic->reference ? |
|
195 |
AV_GET_BUFFER_FLAG_REF : 0); |
||
196 |
✗✓ | 23472 |
if (ret < 0) |
197 |
goto fail; |
||
198 |
|||
199 |
✗✓ | 23472 |
if (h->avctx->hwaccel) { |
200 |
const AVHWAccel *hwaccel = h->avctx->hwaccel; |
||
201 |
av_assert0(!pic->hwaccel_picture_private); |
||
202 |
if (hwaccel->frame_priv_data_size) { |
||
203 |
pic->hwaccel_priv_buf = av_buffer_allocz(hwaccel->frame_priv_data_size); |
||
204 |
if (!pic->hwaccel_priv_buf) |
||
205 |
return AVERROR(ENOMEM); |
||
206 |
pic->hwaccel_picture_private = pic->hwaccel_priv_buf->data; |
||
207 |
} |
||
208 |
} |
||
209 |
if (CONFIG_GRAY && !h->avctx->hwaccel && h->flags & AV_CODEC_FLAG_GRAY && pic->f->data[2]) { |
||
210 |
int h_chroma_shift, v_chroma_shift; |
||
211 |
av_pix_fmt_get_chroma_sub_sample(pic->f->format, |
||
212 |
&h_chroma_shift, &v_chroma_shift); |
||
213 |
|||
214 |
for(i=0; i<AV_CEIL_RSHIFT(pic->f->height, v_chroma_shift); i++) { |
||
215 |
memset(pic->f->data[1] + pic->f->linesize[1]*i, |
||
216 |
0x80, AV_CEIL_RSHIFT(pic->f->width, h_chroma_shift)); |
||
217 |
memset(pic->f->data[2] + pic->f->linesize[2]*i, |
||
218 |
0x80, AV_CEIL_RSHIFT(pic->f->width, h_chroma_shift)); |
||
219 |
} |
||
220 |
} |
||
221 |
|||
222 |
✓✓ | 23472 |
if (!h->qscale_table_pool) { |
223 |
546 |
ret = init_table_pools(h); |
|
224 |
✗✓ | 546 |
if (ret < 0) |
225 |
goto fail; |
||
226 |
} |
||
227 |
|||
228 |
23472 |
pic->qscale_table_buf = av_buffer_pool_get(h->qscale_table_pool); |
|
229 |
23472 |
pic->mb_type_buf = av_buffer_pool_get(h->mb_type_pool); |
|
230 |
✓✗✗✓ |
23472 |
if (!pic->qscale_table_buf || !pic->mb_type_buf) |
231 |
goto fail; |
||
232 |
|||
233 |
23472 |
pic->mb_type = (uint32_t*)pic->mb_type_buf->data + 2 * h->mb_stride + 1; |
|
234 |
23472 |
pic->qscale_table = pic->qscale_table_buf->data + 2 * h->mb_stride + 1; |
|
235 |
|||
236 |
✓✓ | 70416 |
for (i = 0; i < 2; i++) { |
237 |
46944 |
pic->motion_val_buf[i] = av_buffer_pool_get(h->motion_val_pool); |
|
238 |
46944 |
pic->ref_index_buf[i] = av_buffer_pool_get(h->ref_index_pool); |
|
239 |
✓✗✗✓ |
46944 |
if (!pic->motion_val_buf[i] || !pic->ref_index_buf[i]) |
240 |
goto fail; |
||
241 |
|||
242 |
46944 |
pic->motion_val[i] = (int16_t (*)[2])pic->motion_val_buf[i]->data + 4; |
|
243 |
46944 |
pic->ref_index[i] = pic->ref_index_buf[i]->data; |
|
244 |
} |
||
245 |
|||
246 |
23472 |
pic->pps_buf = av_buffer_ref(h->ps.pps_ref); |
|
247 |
✗✓ | 23472 |
if (!pic->pps_buf) |
248 |
goto fail; |
||
249 |
23472 |
pic->pps = (const PPS*)pic->pps_buf->data; |
|
250 |
|||
251 |
23472 |
pic->mb_width = h->mb_width; |
|
252 |
23472 |
pic->mb_height = h->mb_height; |
|
253 |
23472 |
pic->mb_stride = h->mb_stride; |
|
254 |
|||
255 |
23472 |
return 0; |
|
256 |
fail: |
||
257 |
ff_h264_unref_picture(h, pic); |
||
258 |
return (ret < 0) ? ret : AVERROR(ENOMEM); |
||
259 |
} |
||
260 |
|||
261 |
23472 |
static int find_unused_picture(H264Context *h) |
|
262 |
{ |
||
263 |
int i; |
||
264 |
|||
265 |
✓✗ | 82320 |
for (i = 0; i < H264_MAX_PICTURE_COUNT; i++) { |
266 |
✓✓ | 82320 |
if (!h->DPB[i].f->buf[0]) |
267 |
23472 |
return i; |
|
268 |
} |
||
269 |
return AVERROR_INVALIDDATA; |
||
270 |
} |
||
271 |
|||
272 |
|||
273 |
#define IN_RANGE(a, b, size) (((void*)(a) >= (void*)(b)) && ((void*)(a) < (void*)((b) + (size)))) |
||
274 |
|||
275 |
#define REBASE_PICTURE(pic, new_ctx, old_ctx) \ |
||
276 |
(((pic) && (pic) >= (old_ctx)->DPB && \ |
||
277 |
(pic) < (old_ctx)->DPB + H264_MAX_PICTURE_COUNT) ? \ |
||
278 |
&(new_ctx)->DPB[(pic) - (old_ctx)->DPB] : NULL) |
||
279 |
|||
280 |
129 |
static void copy_picture_range(H264Picture **to, H264Picture **from, int count, |
|
281 |
H264Context *new_base, |
||
282 |
H264Context *old_base) |
||
283 |
{ |
||
284 |
int i; |
||
285 |
|||
286 |
✓✓ | 3655 |
for (i = 0; i < count; i++) { |
287 |
av_assert1(!from[i] || |
||
288 |
IN_RANGE(from[i], old_base, 1) || |
||
289 |
IN_RANGE(from[i], old_base->DPB, H264_MAX_PICTURE_COUNT)); |
||
290 |
✓✓✓✗ ✓✗ |
3526 |
to[i] = REBASE_PICTURE(from[i], new_base, old_base); |
291 |
} |
||
292 |
129 |
} |
|
293 |
|||
294 |
static int h264_slice_header_init(H264Context *h); |
||
295 |
|||
296 |
43 |
int ff_h264_update_thread_context(AVCodecContext *dst, |
|
297 |
const AVCodecContext *src) |
||
298 |
{ |
||
299 |
43 |
H264Context *h = dst->priv_data, *h1 = src->priv_data; |
|
300 |
43 |
int inited = h->context_initialized, err = 0; |
|
301 |
43 |
int need_reinit = 0; |
|
302 |
int i, ret; |
||
303 |
|||
304 |
✗✓ | 43 |
if (dst == src) |
305 |
return 0; |
||
306 |
|||
307 |
✓✓✗✓ |
43 |
if (inited && !h1->ps.sps) |
308 |
return AVERROR_INVALIDDATA; |
||
309 |
|||
310 |
✓✓ | 43 |
if (inited && |
311 |
✓✗ | 35 |
(h->width != h1->width || |
312 |
✓✗ | 35 |
h->height != h1->height || |
313 |
✓✗ | 35 |
h->mb_width != h1->mb_width || |
314 |
✓✗ | 35 |
h->mb_height != h1->mb_height || |
315 |
✓✗ | 35 |
!h->ps.sps || |
316 |
✓✗ | 35 |
h->ps.sps->bit_depth_luma != h1->ps.sps->bit_depth_luma || |
317 |
✓✗ | 35 |
h->ps.sps->chroma_format_idc != h1->ps.sps->chroma_format_idc || |
318 |
✗✓ | 35 |
h->ps.sps->colorspace != h1->ps.sps->colorspace)) { |
319 |
need_reinit = 1; |
||
320 |
} |
||
321 |
|||
322 |
/* copy block_offset since frame_start may not be called */ |
||
323 |
43 |
memcpy(h->block_offset, h1->block_offset, sizeof(h->block_offset)); |
|
324 |
|||
325 |
// SPS/PPS |
||
326 |
✓✓ | 1419 |
for (i = 0; i < FF_ARRAY_ELEMS(h->ps.sps_list); i++) { |
327 |
1376 |
ret = av_buffer_replace(&h->ps.sps_list[i], h1->ps.sps_list[i]); |
|
328 |
✗✓ | 1376 |
if (ret < 0) |
329 |
return ret; |
||
330 |
} |
||
331 |
✓✓ | 11051 |
for (i = 0; i < FF_ARRAY_ELEMS(h->ps.pps_list); i++) { |
332 |
11008 |
ret = av_buffer_replace(&h->ps.pps_list[i], h1->ps.pps_list[i]); |
|
333 |
✗✓ | 11008 |
if (ret < 0) |
334 |
return ret; |
||
335 |
} |
||
336 |
|||
337 |
43 |
ret = av_buffer_replace(&h->ps.pps_ref, h1->ps.pps_ref); |
|
338 |
✗✓ | 43 |
if (ret < 0) |
339 |
return ret; |
||
340 |
43 |
h->ps.pps = NULL; |
|
341 |
43 |
h->ps.sps = NULL; |
|
342 |
✓✗ | 43 |
if (h1->ps.pps_ref) { |
343 |
43 |
h->ps.pps = (const PPS*)h->ps.pps_ref->data; |
|
344 |
43 |
h->ps.sps = h->ps.pps->sps; |
|
345 |
} |
||
346 |
|||
347 |
✓✗✓✓ |
43 |
if (need_reinit || !inited) { |
348 |
8 |
h->width = h1->width; |
|
349 |
8 |
h->height = h1->height; |
|
350 |
8 |
h->mb_height = h1->mb_height; |
|
351 |
8 |
h->mb_width = h1->mb_width; |
|
352 |
8 |
h->mb_num = h1->mb_num; |
|
353 |
8 |
h->mb_stride = h1->mb_stride; |
|
354 |
8 |
h->b_stride = h1->b_stride; |
|
355 |
8 |
h->x264_build = h1->x264_build; |
|
356 |
|||
357 |
✓✗✓✗ |
8 |
if (h->context_initialized || h1->context_initialized) { |
358 |
✗✓ | 8 |
if ((err = h264_slice_header_init(h)) < 0) { |
359 |
av_log(h->avctx, AV_LOG_ERROR, "h264_slice_header_init() failed"); |
||
360 |
return err; |
||
361 |
} |
||
362 |
} |
||
363 |
|||
364 |
/* copy block_offset since frame_start may not be called */ |
||
365 |
8 |
memcpy(h->block_offset, h1->block_offset, sizeof(h->block_offset)); |
|
366 |
} |
||
367 |
|||
368 |
43 |
h->avctx->coded_height = h1->avctx->coded_height; |
|
369 |
43 |
h->avctx->coded_width = h1->avctx->coded_width; |
|
370 |
43 |
h->avctx->width = h1->avctx->width; |
|
371 |
43 |
h->avctx->height = h1->avctx->height; |
|
372 |
43 |
h->width_from_caller = h1->width_from_caller; |
|
373 |
43 |
h->height_from_caller = h1->height_from_caller; |
|
374 |
43 |
h->coded_picture_number = h1->coded_picture_number; |
|
375 |
43 |
h->first_field = h1->first_field; |
|
376 |
43 |
h->picture_structure = h1->picture_structure; |
|
377 |
43 |
h->mb_aff_frame = h1->mb_aff_frame; |
|
378 |
43 |
h->droppable = h1->droppable; |
|
379 |
|||
380 |
✓✓ | 1591 |
for (i = 0; i < H264_MAX_PICTURE_COUNT; i++) { |
381 |
1548 |
ff_h264_unref_picture(h, &h->DPB[i]); |
|
382 |
✓✓✗✓ |
1730 |
if (h1->DPB[i].f->buf[0] && |
383 |
182 |
(ret = ff_h264_ref_picture(h, &h->DPB[i], &h1->DPB[i])) < 0) |
|
384 |
return ret; |
||
385 |
} |
||
386 |
|||
387 |
✓✓✓✗ ✓✗ |
43 |
h->cur_pic_ptr = REBASE_PICTURE(h1->cur_pic_ptr, h, h1); |
388 |
43 |
ff_h264_unref_picture(h, &h->cur_pic); |
|
389 |
✓✗ | 43 |
if (h1->cur_pic.f->buf[0]) { |
390 |
43 |
ret = ff_h264_ref_picture(h, &h->cur_pic, &h1->cur_pic); |
|
391 |
✗✓ | 43 |
if (ret < 0) |
392 |
return ret; |
||
393 |
} |
||
394 |
|||
395 |
43 |
h->enable_er = h1->enable_er; |
|
396 |
43 |
h->workaround_bugs = h1->workaround_bugs; |
|
397 |
43 |
h->droppable = h1->droppable; |
|
398 |
|||
399 |
// extradata/NAL handling |
||
400 |
43 |
h->is_avc = h1->is_avc; |
|
401 |
43 |
h->nal_length_size = h1->nal_length_size; |
|
402 |
|||
403 |
43 |
memcpy(&h->poc, &h1->poc, sizeof(h->poc)); |
|
404 |
|||
405 |
43 |
memcpy(h->short_ref, h1->short_ref, sizeof(h->short_ref)); |
|
406 |
43 |
memcpy(h->long_ref, h1->long_ref, sizeof(h->long_ref)); |
|
407 |
43 |
memcpy(h->delayed_pic, h1->delayed_pic, sizeof(h->delayed_pic)); |
|
408 |
43 |
memcpy(h->last_pocs, h1->last_pocs, sizeof(h->last_pocs)); |
|
409 |
|||
410 |
43 |
h->next_output_pic = h1->next_output_pic; |
|
411 |
43 |
h->next_outputed_poc = h1->next_outputed_poc; |
|
412 |
|||
413 |
43 |
memcpy(h->mmco, h1->mmco, sizeof(h->mmco)); |
|
414 |
43 |
h->nb_mmco = h1->nb_mmco; |
|
415 |
43 |
h->mmco_reset = h1->mmco_reset; |
|
416 |
43 |
h->explicit_ref_marking = h1->explicit_ref_marking; |
|
417 |
43 |
h->long_ref_count = h1->long_ref_count; |
|
418 |
43 |
h->short_ref_count = h1->short_ref_count; |
|
419 |
|||
420 |
43 |
copy_picture_range(h->short_ref, h1->short_ref, 32, h, h1); |
|
421 |
43 |
copy_picture_range(h->long_ref, h1->long_ref, 32, h, h1); |
|
422 |
43 |
copy_picture_range(h->delayed_pic, h1->delayed_pic, |
|
423 |
MAX_DELAYED_PIC_COUNT + 2, h, h1); |
||
424 |
|||
425 |
43 |
h->frame_recovered = h1->frame_recovered; |
|
426 |
|||
427 |
43 |
ret = av_buffer_replace(&h->sei.a53_caption.buf_ref, h1->sei.a53_caption.buf_ref); |
|
428 |
✗✓ | 43 |
if (ret < 0) |
429 |
return ret; |
||
430 |
|||
431 |
✗✓ | 43 |
for (i = 0; i < h->sei.unregistered.nb_buf_ref; i++) |
432 |
av_buffer_unref(&h->sei.unregistered.buf_ref[i]); |
||
433 |
43 |
h->sei.unregistered.nb_buf_ref = 0; |
|
434 |
|||
435 |
✗✓ | 43 |
if (h1->sei.unregistered.nb_buf_ref) { |
436 |
ret = av_reallocp_array(&h->sei.unregistered.buf_ref, |
||
437 |
h1->sei.unregistered.nb_buf_ref, |
||
438 |
sizeof(*h->sei.unregistered.buf_ref)); |
||
439 |
if (ret < 0) |
||
440 |
return ret; |
||
441 |
|||
442 |
for (i = 0; i < h1->sei.unregistered.nb_buf_ref; i++) { |
||
443 |
h->sei.unregistered.buf_ref[i] = av_buffer_ref(h1->sei.unregistered.buf_ref[i]); |
||
444 |
if (!h->sei.unregistered.buf_ref[i]) |
||
445 |
return AVERROR(ENOMEM); |
||
446 |
h->sei.unregistered.nb_buf_ref++; |
||
447 |
} |
||
448 |
} |
||
449 |
43 |
h->sei.unregistered.x264_build = h1->sei.unregistered.x264_build; |
|
450 |
|||
451 |
✓✓ | 43 |
if (!h->cur_pic_ptr) |
452 |
11 |
return 0; |
|
453 |
|||
454 |
✓✓ | 32 |
if (!h->droppable) { |
455 |
19 |
err = ff_h264_execute_ref_pic_marking(h); |
|
456 |
19 |
h->poc.prev_poc_msb = h->poc.poc_msb; |
|
457 |
19 |
h->poc.prev_poc_lsb = h->poc.poc_lsb; |
|
458 |
} |
||
459 |
32 |
h->poc.prev_frame_num_offset = h->poc.frame_num_offset; |
|
460 |
32 |
h->poc.prev_frame_num = h->poc.frame_num; |
|
461 |
|||
462 |
32 |
h->recovery_frame = h1->recovery_frame; |
|
463 |
|||
464 |
32 |
return err; |
|
465 |
} |
||
466 |
|||
467 |
36 |
int ff_h264_update_thread_context_for_user(AVCodecContext *dst, |
|
468 |
const AVCodecContext *src) |
||
469 |
{ |
||
470 |
36 |
H264Context *h = dst->priv_data; |
|
471 |
36 |
const H264Context *h1 = src->priv_data; |
|
472 |
|||
473 |
36 |
h->is_avc = h1->is_avc; |
|
474 |
36 |
h->nal_length_size = h1->nal_length_size; |
|
475 |
|||
476 |
36 |
return 0; |
|
477 |
} |
||
478 |
|||
479 |
23472 |
static int h264_frame_start(H264Context *h) |
|
480 |
{ |
||
481 |
H264Picture *pic; |
||
482 |
int i, ret; |
||
483 |
23472 |
const int pixel_shift = h->pixel_shift; |
|
484 |
|||
485 |
✗✓ | 23472 |
if (!ff_thread_can_start_frame(h->avctx)) { |
486 |
av_log(h->avctx, AV_LOG_ERROR, "Attempt to start a frame outside SETUP state\n"); |
||
487 |
return -1; |
||
488 |
} |
||
489 |
|||
490 |
23472 |
release_unused_pictures(h, 1); |
|
491 |
23472 |
h->cur_pic_ptr = NULL; |
|
492 |
|||
493 |
23472 |
i = find_unused_picture(h); |
|
494 |
✗✓ | 23472 |
if (i < 0) { |
495 |
av_log(h->avctx, AV_LOG_ERROR, "no frame buffer available\n"); |
||
496 |
return i; |
||
497 |
} |
||
498 |
23472 |
pic = &h->DPB[i]; |
|
499 |
|||
500 |
✓✓ | 23472 |
pic->reference = h->droppable ? 0 : h->picture_structure; |
501 |
23472 |
pic->f->coded_picture_number = h->coded_picture_number++; |
|
502 |
23472 |
pic->field_picture = h->picture_structure != PICT_FRAME; |
|
503 |
23472 |
pic->frame_num = h->poc.frame_num; |
|
504 |
/* |
||
505 |
* Zero key_frame here; IDR markings per slice in frame or fields are ORed |
||
506 |
* in later. |
||
507 |
* See decode_nal_units(). |
||
508 |
*/ |
||
509 |
23472 |
pic->f->key_frame = 0; |
|
510 |
23472 |
pic->mmco_reset = 0; |
|
511 |
23472 |
pic->recovered = 0; |
|
512 |
23472 |
pic->invalid_gap = 0; |
|
513 |
23472 |
pic->sei_recovery_frame_cnt = h->sei.recovery_point.recovery_frame_cnt; |
|
514 |
|||
515 |
23472 |
pic->f->pict_type = h->slice_ctx[0].slice_type; |
|
516 |
|||
517 |
23472 |
pic->f->crop_left = h->crop_left; |
|
518 |
23472 |
pic->f->crop_right = h->crop_right; |
|
519 |
23472 |
pic->f->crop_top = h->crop_top; |
|
520 |
23472 |
pic->f->crop_bottom = h->crop_bottom; |
|
521 |
|||
522 |
✗✓ | 23472 |
if ((ret = alloc_picture(h, pic)) < 0) |
523 |
return ret; |
||
524 |
|||
525 |
23472 |
h->cur_pic_ptr = pic; |
|
526 |
23472 |
ff_h264_unref_picture(h, &h->cur_pic); |
|
527 |
if (CONFIG_ERROR_RESILIENCE) { |
||
528 |
23472 |
ff_h264_set_erpic(&h->slice_ctx[0].er.cur_pic, NULL); |
|
529 |
} |
||
530 |
|||
531 |
✗✓ | 23472 |
if ((ret = ff_h264_ref_picture(h, &h->cur_pic, h->cur_pic_ptr)) < 0) |
532 |
return ret; |
||
533 |
|||
534 |
✓✓ | 47244 |
for (i = 0; i < h->nb_slice_ctx; i++) { |
535 |
23772 |
h->slice_ctx[i].linesize = h->cur_pic_ptr->f->linesize[0]; |
|
536 |
23772 |
h->slice_ctx[i].uvlinesize = h->cur_pic_ptr->f->linesize[1]; |
|
537 |
} |
||
538 |
|||
539 |
✓✓ | 23472 |
if (CONFIG_ERROR_RESILIENCE && h->enable_er) { |
540 |
23172 |
ff_er_frame_start(&h->slice_ctx[0].er); |
|
541 |
23172 |
ff_h264_set_erpic(&h->slice_ctx[0].er.last_pic, NULL); |
|
542 |
23172 |
ff_h264_set_erpic(&h->slice_ctx[0].er.next_pic, NULL); |
|
543 |
} |
||
544 |
|||
545 |
✓✓ | 399024 |
for (i = 0; i < 16; i++) { |
546 |
375552 |
h->block_offset[i] = (4 * ((scan8[i] - scan8[0]) & 7) << pixel_shift) + 4 * pic->f->linesize[0] * ((scan8[i] - scan8[0]) >> 3); |
|
547 |
375552 |
h->block_offset[48 + i] = (4 * ((scan8[i] - scan8[0]) & 7) << pixel_shift) + 8 * pic->f->linesize[0] * ((scan8[i] - scan8[0]) >> 3); |
|
548 |
} |
||
549 |
✓✓ | 399024 |
for (i = 0; i < 16; i++) { |
550 |
375552 |
h->block_offset[16 + i] = |
|
551 |
375552 |
h->block_offset[32 + i] = (4 * ((scan8[i] - scan8[0]) & 7) << pixel_shift) + 4 * pic->f->linesize[1] * ((scan8[i] - scan8[0]) >> 3); |
|
552 |
375552 |
h->block_offset[48 + 16 + i] = |
|
553 |
375552 |
h->block_offset[48 + 32 + i] = (4 * ((scan8[i] - scan8[0]) & 7) << pixel_shift) + 8 * pic->f->linesize[1] * ((scan8[i] - scan8[0]) >> 3); |
|
554 |
} |
||
555 |
|||
556 |
/* We mark the current picture as non-reference after allocating it, so |
||
557 |
* that if we break out due to an error it can be released automatically |
||
558 |
* in the next ff_mpv_frame_start(). |
||
559 |
*/ |
||
560 |
23472 |
h->cur_pic_ptr->reference = 0; |
|
561 |
|||
562 |
23472 |
h->cur_pic_ptr->field_poc[0] = h->cur_pic_ptr->field_poc[1] = INT_MAX; |
|
563 |
|||
564 |
23472 |
h->next_output_pic = NULL; |
|
565 |
|||
566 |
23472 |
h->postpone_filter = 0; |
|
567 |
|||
568 |
✓✓✓✓ |
23472 |
h->mb_aff_frame = h->ps.sps->mb_aff && (h->picture_structure == PICT_FRAME); |
569 |
|||
570 |
✓✓ | 23472 |
if (h->sei.unregistered.x264_build >= 0) |
571 |
1829 |
h->x264_build = h->sei.unregistered.x264_build; |
|
572 |
|||
573 |
assert(h->cur_pic_ptr->long_ref == 0); |
||
574 |
|||
575 |
23472 |
return 0; |
|
576 |
} |
||
577 |
|||
578 |
11609103 |
static av_always_inline void backup_mb_border(const H264Context *h, H264SliceContext *sl, |
|
579 |
uint8_t *src_y, |
||
580 |
uint8_t *src_cb, uint8_t *src_cr, |
||
581 |
int linesize, int uvlinesize, |
||
582 |
int simple) |
||
583 |
{ |
||
584 |
uint8_t *top_border; |
||
585 |
11609103 |
int top_idx = 1; |
|
586 |
11609103 |
const int pixel_shift = h->pixel_shift; |
|
587 |
11609103 |
int chroma444 = CHROMA444(h); |
|
588 |
11609103 |
int chroma422 = CHROMA422(h); |
|
589 |
|||
590 |
11609103 |
src_y -= linesize; |
|
591 |
11609103 |
src_cb -= uvlinesize; |
|
592 |
11609103 |
src_cr -= uvlinesize; |
|
593 |
|||
594 |
✓✗✓✓ |
11609103 |
if (!simple && FRAME_MBAFF(h)) { |
595 |
✓✓ | 1814136 |
if (sl->mb_y & 1) { |
596 |
✓✓ | 907068 |
if (!MB_MBAFF(sl)) { |
597 |
665811 |
top_border = sl->top_borders[0][sl->mb_x]; |
|
598 |
665811 |
AV_COPY128(top_border, src_y + 15 * linesize); |
|
599 |
✓✓ | 665811 |
if (pixel_shift) |
600 |
213139 |
AV_COPY128(top_border + 16, src_y + 15 * linesize + 16); |
|
601 |
if (simple || !CONFIG_GRAY || !(h->flags & AV_CODEC_FLAG_GRAY)) { |
||
602 |
✗✓ | 665811 |
if (chroma444) { |
603 |
if (pixel_shift) { |
||
604 |
AV_COPY128(top_border + 32, src_cb + 15 * uvlinesize); |
||
605 |
AV_COPY128(top_border + 48, src_cb + 15 * uvlinesize + 16); |
||
606 |
AV_COPY128(top_border + 64, src_cr + 15 * uvlinesize); |
||
607 |
AV_COPY128(top_border + 80, src_cr + 15 * uvlinesize + 16); |
||
608 |
} else { |
||
609 |
AV_COPY128(top_border + 16, src_cb + 15 * uvlinesize); |
||
610 |
AV_COPY128(top_border + 32, src_cr + 15 * uvlinesize); |
||
611 |
} |
||
612 |
✓✓ | 665811 |
} else if (chroma422) { |
613 |
✓✗ | 106510 |
if (pixel_shift) { |
614 |
106510 |
AV_COPY128(top_border + 32, src_cb + 15 * uvlinesize); |
|
615 |
106510 |
AV_COPY128(top_border + 48, src_cr + 15 * uvlinesize); |
|
616 |
} else { |
||
617 |
AV_COPY64(top_border + 16, src_cb + 15 * uvlinesize); |
||
618 |
AV_COPY64(top_border + 24, src_cr + 15 * uvlinesize); |
||
619 |
} |
||
620 |
} else { |
||
621 |
✓✓ | 559301 |
if (pixel_shift) { |
622 |
106629 |
AV_COPY128(top_border + 32, src_cb + 7 * uvlinesize); |
|
623 |
106629 |
AV_COPY128(top_border + 48, src_cr + 7 * uvlinesize); |
|
624 |
} else { |
||
625 |
452672 |
AV_COPY64(top_border + 16, src_cb + 7 * uvlinesize); |
|
626 |
452672 |
AV_COPY64(top_border + 24, src_cr + 7 * uvlinesize); |
|
627 |
} |
||
628 |
} |
||
629 |
} |
||
630 |
} |
||
631 |
✓✓ | 907068 |
} else if (MB_MBAFF(sl)) { |
632 |
241257 |
top_idx = 0; |
|
633 |
} else |
||
634 |
665811 |
return; |
|
635 |
} |
||
636 |
|||
637 |
10943292 |
top_border = sl->top_borders[top_idx][sl->mb_x]; |
|
638 |
/* There are two lines saved, the line above the top macroblock |
||
639 |
* of a pair, and the line above the bottom macroblock. */ |
||
640 |
10943292 |
AV_COPY128(top_border, src_y + 16 * linesize); |
|
641 |
✓✓ | 10943292 |
if (pixel_shift) |
642 |
1300901 |
AV_COPY128(top_border + 16, src_y + 16 * linesize + 16); |
|
643 |
|||
644 |
if (simple || !CONFIG_GRAY || !(h->flags & AV_CODEC_FLAG_GRAY)) { |
||
645 |
✓✓ | 10943292 |
if (chroma444) { |
646 |
✓✓ | 80328 |
if (pixel_shift) { |
647 |
19800 |
AV_COPY128(top_border + 32, src_cb + 16 * linesize); |
|
648 |
19800 |
AV_COPY128(top_border + 48, src_cb + 16 * linesize + 16); |
|
649 |
19800 |
AV_COPY128(top_border + 64, src_cr + 16 * linesize); |
|
650 |
19800 |
AV_COPY128(top_border + 80, src_cr + 16 * linesize + 16); |
|
651 |
} else { |
||
652 |
60528 |
AV_COPY128(top_border + 16, src_cb + 16 * linesize); |
|
653 |
60528 |
AV_COPY128(top_border + 32, src_cr + 16 * linesize); |
|
654 |
} |
||
655 |
✓✓ | 10862964 |
} else if (chroma422) { |
656 |
✓✓ | 588635 |
if (pixel_shift) { |
657 |
584675 |
AV_COPY128(top_border + 32, src_cb + 16 * uvlinesize); |
|
658 |
584675 |
AV_COPY128(top_border + 48, src_cr + 16 * uvlinesize); |
|
659 |
} else { |
||
660 |
3960 |
AV_COPY64(top_border + 16, src_cb + 16 * uvlinesize); |
|
661 |
3960 |
AV_COPY64(top_border + 24, src_cr + 16 * uvlinesize); |
|
662 |
} |
||
663 |
} else { |
||
664 |
✓✓ | 10274329 |
if (pixel_shift) { |
665 |
696426 |
AV_COPY128(top_border + 32, src_cb + 8 * uvlinesize); |
|
666 |
696426 |
AV_COPY128(top_border + 48, src_cr + 8 * uvlinesize); |
|
667 |
} else { |
||
668 |
9577903 |
AV_COPY64(top_border + 16, src_cb + 8 * uvlinesize); |
|
669 |
9577903 |
AV_COPY64(top_border + 24, src_cr + 8 * uvlinesize); |
|
670 |
} |
||
671 |
} |
||
672 |
} |
||
673 |
} |
||
674 |
|||
675 |
/** |
||
676 |
* Initialize implicit_weight table. |
||
677 |
* @param field 0/1 initialize the weight for interlaced MBAFF |
||
678 |
* -1 initializes the rest |
||
679 |
*/ |
||
680 |
3299 |
static void implicit_weight_table(const H264Context *h, H264SliceContext *sl, int field) |
|
681 |
{ |
||
682 |
int ref0, ref1, i, cur_poc, ref_start, ref_count0, ref_count1; |
||
683 |
|||
684 |
✓✓ | 9897 |
for (i = 0; i < 2; i++) { |
685 |
6598 |
sl->pwt.luma_weight_flag[i] = 0; |
|
686 |
6598 |
sl->pwt.chroma_weight_flag[i] = 0; |
|
687 |
} |
||
688 |
|||
689 |
✓✓ | 3299 |
if (field < 0) { |
690 |
✓✗ | 1589 |
if (h->picture_structure == PICT_FRAME) { |
691 |
1589 |
cur_poc = h->cur_pic_ptr->poc; |
|
692 |
} else { |
||
693 |
cur_poc = h->cur_pic_ptr->field_poc[h->picture_structure - 1]; |
||
694 |
} |
||
695 |
✓✓✓✓ ✓✓ |
1589 |
if (sl->ref_count[0] == 1 && sl->ref_count[1] == 1 && !FRAME_MBAFF(h) && |
696 |
✓✓ | 97 |
sl->ref_list[0][0].poc + (int64_t)sl->ref_list[1][0].poc == 2LL * cur_poc) { |
697 |
34 |
sl->pwt.use_weight = 0; |
|
698 |
34 |
sl->pwt.use_weight_chroma = 0; |
|
699 |
34 |
return; |
|
700 |
} |
||
701 |
1555 |
ref_start = 0; |
|
702 |
1555 |
ref_count0 = sl->ref_count[0]; |
|
703 |
1555 |
ref_count1 = sl->ref_count[1]; |
|
704 |
} else { |
||
705 |
1710 |
cur_poc = h->cur_pic_ptr->field_poc[field]; |
|
706 |
1710 |
ref_start = 16; |
|
707 |
1710 |
ref_count0 = 16 + 2 * sl->ref_count[0]; |
|
708 |
1710 |
ref_count1 = 16 + 2 * sl->ref_count[1]; |
|
709 |
} |
||
710 |
|||
711 |
3265 |
sl->pwt.use_weight = 2; |
|
712 |
3265 |
sl->pwt.use_weight_chroma = 2; |
|
713 |
3265 |
sl->pwt.luma_log2_weight_denom = 5; |
|
714 |
3265 |
sl->pwt.chroma_log2_weight_denom = 5; |
|
715 |
|||
716 |
✓✓ | 14879 |
for (ref0 = ref_start; ref0 < ref_count0; ref0++) { |
717 |
11614 |
int64_t poc0 = sl->ref_list[0][ref0].poc; |
|
718 |
✓✓ | 33317 |
for (ref1 = ref_start; ref1 < ref_count1; ref1++) { |
719 |
21703 |
int w = 32; |
|
720 |
✓✗✓✗ |
21703 |
if (!sl->ref_list[0][ref0].parent->long_ref && !sl->ref_list[1][ref1].parent->long_ref) { |
721 |
21703 |
int poc1 = sl->ref_list[1][ref1].poc; |
|
722 |
21703 |
int td = av_clip_int8(poc1 - poc0); |
|
723 |
✓✓ | 21703 |
if (td) { |
724 |
21637 |
int tb = av_clip_int8(cur_poc - poc0); |
|
725 |
21637 |
int tx = (16384 + (FFABS(td) >> 1)) / td; |
|
726 |
21637 |
int dist_scale_factor = (tb * tx + 32) >> 8; |
|
727 |
✓✓✓✗ |
21637 |
if (dist_scale_factor >= -64 && dist_scale_factor <= 128) |
728 |
21633 |
w = 64 - dist_scale_factor; |
|
729 |
} |
||
730 |
} |
||
731 |
✓✓ | 21703 |
if (field < 0) { |
732 |
5823 |
sl->pwt.implicit_weight[ref0][ref1][0] = |
|
733 |
5823 |
sl->pwt.implicit_weight[ref0][ref1][1] = w; |
|
734 |
} else { |
||
735 |
15880 |
sl->pwt.implicit_weight[ref0][ref1][field] = w; |
|
736 |
} |
||
737 |
} |
||
738 |
} |
||
739 |
} |
||
740 |
|||
741 |
/** |
||
742 |
* initialize scan tables |
||
743 |
*/ |
||
744 |
546 |
static void init_scan_tables(H264Context *h) |
|
745 |
{ |
||
746 |
int i; |
||
747 |
✓✓ | 9282 |
for (i = 0; i < 16; i++) { |
748 |
#define TRANSPOSE(x) ((x) >> 2) | (((x) << 2) & 0xF) |
||
749 |
8736 |
h->zigzag_scan[i] = TRANSPOSE(ff_zigzag_scan[i]); |
|
750 |
8736 |
h->field_scan[i] = TRANSPOSE(field_scan[i]); |
|
751 |
#undef TRANSPOSE |
||
752 |
} |
||
753 |
✓✓ | 35490 |
for (i = 0; i < 64; i++) { |
754 |
#define TRANSPOSE(x) ((x) >> 3) | (((x) & 7) << 3) |
||
755 |
34944 |
h->zigzag_scan8x8[i] = TRANSPOSE(ff_zigzag_direct[i]); |
|
756 |
34944 |
h->zigzag_scan8x8_cavlc[i] = TRANSPOSE(zigzag_scan8x8_cavlc[i]); |
|
757 |
34944 |
h->field_scan8x8[i] = TRANSPOSE(field_scan8x8[i]); |
|
758 |
34944 |
h->field_scan8x8_cavlc[i] = TRANSPOSE(field_scan8x8_cavlc[i]); |
|
759 |
#undef TRANSPOSE |
||
760 |
} |
||
761 |
✓✓ | 546 |
if (h->ps.sps->transform_bypass) { // FIXME same ugly |
762 |
12 |
memcpy(h->zigzag_scan_q0 , ff_zigzag_scan , sizeof(h->zigzag_scan_q0 )); |
|
763 |
12 |
memcpy(h->zigzag_scan8x8_q0 , ff_zigzag_direct , sizeof(h->zigzag_scan8x8_q0 )); |
|
764 |
12 |
memcpy(h->zigzag_scan8x8_cavlc_q0 , zigzag_scan8x8_cavlc , sizeof(h->zigzag_scan8x8_cavlc_q0)); |
|
765 |
12 |
memcpy(h->field_scan_q0 , field_scan , sizeof(h->field_scan_q0 )); |
|
766 |
12 |
memcpy(h->field_scan8x8_q0 , field_scan8x8 , sizeof(h->field_scan8x8_q0 )); |
|
767 |
12 |
memcpy(h->field_scan8x8_cavlc_q0 , field_scan8x8_cavlc , sizeof(h->field_scan8x8_cavlc_q0 )); |
|
768 |
} else { |
||
769 |
534 |
memcpy(h->zigzag_scan_q0 , h->zigzag_scan , sizeof(h->zigzag_scan_q0 )); |
|
770 |
534 |
memcpy(h->zigzag_scan8x8_q0 , h->zigzag_scan8x8 , sizeof(h->zigzag_scan8x8_q0 )); |
|
771 |
534 |
memcpy(h->zigzag_scan8x8_cavlc_q0 , h->zigzag_scan8x8_cavlc , sizeof(h->zigzag_scan8x8_cavlc_q0)); |
|
772 |
534 |
memcpy(h->field_scan_q0 , h->field_scan , sizeof(h->field_scan_q0 )); |
|
773 |
534 |
memcpy(h->field_scan8x8_q0 , h->field_scan8x8 , sizeof(h->field_scan8x8_q0 )); |
|
774 |
534 |
memcpy(h->field_scan8x8_cavlc_q0 , h->field_scan8x8_cavlc , sizeof(h->field_scan8x8_cavlc_q0 )); |
|
775 |
} |
||
776 |
546 |
} |
|
777 |
|||
778 |
26353 |
static enum AVPixelFormat get_pixel_format(H264Context *h, int force_callback) |
|
779 |
{ |
||
780 |
#define HWACCEL_MAX (CONFIG_H264_DXVA2_HWACCEL + \ |
||
781 |
(CONFIG_H264_D3D11VA_HWACCEL * 2) + \ |
||
782 |
CONFIG_H264_NVDEC_HWACCEL + \ |
||
783 |
CONFIG_H264_VAAPI_HWACCEL + \ |
||
784 |
CONFIG_H264_VIDEOTOOLBOX_HWACCEL + \ |
||
785 |
CONFIG_H264_VDPAU_HWACCEL) |
||
786 |
26353 |
enum AVPixelFormat pix_fmts[HWACCEL_MAX + 2], *fmt = pix_fmts; |
|
787 |
26353 |
const enum AVPixelFormat *choices = pix_fmts; |
|
788 |
int i; |
||
789 |
|||
790 |
✓✓✗✗ ✓✗ |
26353 |
switch (h->ps.sps->bit_depth_luma) { |
791 |
155 |
case 9: |
|
792 |
✗✓ | 155 |
if (CHROMA444(h)) { |
793 |
if (h->avctx->colorspace == AVCOL_SPC_RGB) { |
||
794 |
*fmt++ = AV_PIX_FMT_GBRP9; |
||
795 |
} else |
||
796 |
*fmt++ = AV_PIX_FMT_YUV444P9; |
||
797 |
✓✓ | 155 |
} else if (CHROMA422(h)) |
798 |
52 |
*fmt++ = AV_PIX_FMT_YUV422P9; |
|
799 |
else |
||
800 |
103 |
*fmt++ = AV_PIX_FMT_YUV420P9; |
|
801 |
155 |
break; |
|
802 |
355 |
case 10: |
|
803 |
✓✓ | 355 |
if (CHROMA444(h)) { |
804 |
✗✓ | 51 |
if (h->avctx->colorspace == AVCOL_SPC_RGB) { |
805 |
*fmt++ = AV_PIX_FMT_GBRP10; |
||
806 |
} else |
||
807 |
51 |
*fmt++ = AV_PIX_FMT_YUV444P10; |
|
808 |
✓✓ | 304 |
} else if (CHROMA422(h)) |
809 |
164 |
*fmt++ = AV_PIX_FMT_YUV422P10; |
|
810 |
else |
||
811 |
140 |
*fmt++ = AV_PIX_FMT_YUV420P10; |
|
812 |
355 |
break; |
|
813 |
case 12: |
||
814 |
if (CHROMA444(h)) { |
||
815 |
if (h->avctx->colorspace == AVCOL_SPC_RGB) { |
||
816 |
*fmt++ = AV_PIX_FMT_GBRP12; |
||
817 |
} else |
||
818 |
*fmt++ = AV_PIX_FMT_YUV444P12; |
||
819 |
} else if (CHROMA422(h)) |
||
820 |
*fmt++ = AV_PIX_FMT_YUV422P12; |
||
821 |
else |
||
822 |
*fmt++ = AV_PIX_FMT_YUV420P12; |
||
823 |
break; |
||
824 |
case 14: |
||
825 |
if (CHROMA444(h)) { |
||
826 |
if (h->avctx->colorspace == AVCOL_SPC_RGB) { |
||
827 |
*fmt++ = AV_PIX_FMT_GBRP14; |
||
828 |
} else |
||
829 |
*fmt++ = AV_PIX_FMT_YUV444P14; |
||
830 |
} else if (CHROMA422(h)) |
||
831 |
*fmt++ = AV_PIX_FMT_YUV422P14; |
||
832 |
else |
||
833 |
*fmt++ = AV_PIX_FMT_YUV420P14; |
||
834 |
break; |
||
835 |
25843 |
case 8: |
|
836 |
#if CONFIG_H264_VDPAU_HWACCEL |
||
837 |
25843 |
*fmt++ = AV_PIX_FMT_VDPAU; |
|
838 |
#endif |
||
839 |
#if CONFIG_H264_NVDEC_HWACCEL |
||
840 |
*fmt++ = AV_PIX_FMT_CUDA; |
||
841 |
#endif |
||
842 |
✓✓ | 25843 |
if (CHROMA444(h)) { |
843 |
✗✓ | 405 |
if (h->avctx->colorspace == AVCOL_SPC_RGB) |
844 |
*fmt++ = AV_PIX_FMT_GBRP; |
||
845 |
✓✓ | 405 |
else if (h->avctx->color_range == AVCOL_RANGE_JPEG) |
846 |
2 |
*fmt++ = AV_PIX_FMT_YUVJ444P; |
|
847 |
else |
||
848 |
403 |
*fmt++ = AV_PIX_FMT_YUV444P; |
|
849 |
✓✓ | 25438 |
} else if (CHROMA422(h)) { |
850 |
✗✓ | 22 |
if (h->avctx->color_range == AVCOL_RANGE_JPEG) |
851 |
*fmt++ = AV_PIX_FMT_YUVJ422P; |
||
852 |
else |
||
853 |
22 |
*fmt++ = AV_PIX_FMT_YUV422P; |
|
854 |
} else { |
||
855 |
#if CONFIG_H264_DXVA2_HWACCEL |
||
856 |
*fmt++ = AV_PIX_FMT_DXVA2_VLD; |
||
857 |
#endif |
||
858 |
#if CONFIG_H264_D3D11VA_HWACCEL |
||
859 |
*fmt++ = AV_PIX_FMT_D3D11VA_VLD; |
||
860 |
*fmt++ = AV_PIX_FMT_D3D11; |
||
861 |
#endif |
||
862 |
#if CONFIG_H264_VAAPI_HWACCEL |
||
863 |
25416 |
*fmt++ = AV_PIX_FMT_VAAPI; |
|
864 |
#endif |
||
865 |
#if CONFIG_H264_VIDEOTOOLBOX_HWACCEL |
||
866 |
*fmt++ = AV_PIX_FMT_VIDEOTOOLBOX; |
||
867 |
#endif |
||
868 |
✗✓ | 25416 |
if (h->avctx->codec->pix_fmts) |
869 |
choices = h->avctx->codec->pix_fmts; |
||
870 |
✓✓ | 25416 |
else if (h->avctx->color_range == AVCOL_RANGE_JPEG) |
871 |
3 |
*fmt++ = AV_PIX_FMT_YUVJ420P; |
|
872 |
else |
||
873 |
25413 |
*fmt++ = AV_PIX_FMT_YUV420P; |
|
874 |
} |
||
875 |
25843 |
break; |
|
876 |
default: |
||
877 |
av_log(h->avctx, AV_LOG_ERROR, |
||
878 |
"Unsupported bit depth %d\n", h->ps.sps->bit_depth_luma); |
||
879 |
return AVERROR_INVALIDDATA; |
||
880 |
} |
||
881 |
|||
882 |
26353 |
*fmt = AV_PIX_FMT_NONE; |
|
883 |
|||
884 |
✓✓ | 78153 |
for (i=0; choices[i] != AV_PIX_FMT_NONE; i++) |
885 |
✓✓✓✓ |
77612 |
if (choices[i] == h->avctx->pix_fmt && !force_callback) |
886 |
25812 |
return choices[i]; |
|
887 |
541 |
return ff_thread_get_format(h->avctx, choices); |
|
888 |
} |
||
889 |
|||
890 |
/* export coded and cropped frame dimensions to AVCodecContext */ |
||
891 |
26111 |
static void init_dimensions(H264Context *h) |
|
892 |
{ |
||
893 |
26111 |
const SPS *sps = (const SPS*)h->ps.sps; |
|
894 |
26111 |
int cr = sps->crop_right; |
|
895 |
26111 |
int cl = sps->crop_left; |
|
896 |
26111 |
int ct = sps->crop_top; |
|
897 |
26111 |
int cb = sps->crop_bottom; |
|
898 |
26111 |
int width = h->width - (cr + cl); |
|
899 |
26111 |
int height = h->height - (ct + cb); |
|
900 |
✗✓ | 26111 |
av_assert0(sps->crop_right + sps->crop_left < (unsigned)h->width); |
901 |
✗✓ | 26111 |
av_assert0(sps->crop_top + sps->crop_bottom < (unsigned)h->height); |
902 |
|||
903 |
/* handle container cropping */ |
||
904 |
✓✓✓✗ |
26111 |
if (h->width_from_caller > 0 && h->height_from_caller > 0 && |
905 |
✓✓✓✗ |
23892 |
!sps->crop_top && !sps->crop_left && |
906 |
✓✓ | 23891 |
FFALIGN(h->width_from_caller, 16) == FFALIGN(width, 16) && |
907 |
✓✗ | 23886 |
FFALIGN(h->height_from_caller, 16) == FFALIGN(height, 16) && |
908 |
✓✗ | 23886 |
h->width_from_caller <= width && |
909 |
✓✓ | 23886 |
h->height_from_caller <= height) { |
910 |
23885 |
width = h->width_from_caller; |
|
911 |
23885 |
height = h->height_from_caller; |
|
912 |
23885 |
cl = 0; |
|
913 |
23885 |
ct = 0; |
|
914 |
23885 |
cr = h->width - width; |
|
915 |
23885 |
cb = h->height - height; |
|
916 |
} else { |
||
917 |
2226 |
h->width_from_caller = 0; |
|
918 |
2226 |
h->height_from_caller = 0; |
|
919 |
} |
||
920 |
|||
921 |
26111 |
h->avctx->coded_width = h->width; |
|
922 |
26111 |
h->avctx->coded_height = h->height; |
|
923 |
26111 |
h->avctx->width = width; |
|
924 |
26111 |
h->avctx->height = height; |
|
925 |
26111 |
h->crop_right = cr; |
|
926 |
26111 |
h->crop_left = cl; |
|
927 |
26111 |
h->crop_top = ct; |
|
928 |
26111 |
h->crop_bottom = cb; |
|
929 |
26111 |
} |
|
930 |
|||
931 |
546 |
static int h264_slice_header_init(H264Context *h) |
|
932 |
{ |
||
933 |
546 |
const SPS *sps = h->ps.sps; |
|
934 |
int i, ret; |
||
935 |
|||
936 |
✗✓ | 546 |
if (!sps) { |
937 |
ret = AVERROR_INVALIDDATA; |
||
938 |
goto fail; |
||
939 |
} |
||
940 |
|||
941 |
546 |
ff_set_sar(h->avctx, sps->sar); |
|
942 |
546 |
av_pix_fmt_get_chroma_sub_sample(h->avctx->pix_fmt, |
|
943 |
&h->chroma_x_shift, &h->chroma_y_shift); |
||
944 |
|||
945 |
✓✓ | 546 |
if (sps->timing_info_present_flag) { |
946 |
143 |
int64_t den = sps->time_scale; |
|
947 |
✗✓ | 143 |
if (h->x264_build < 44U) |
948 |
den *= 2; |
||
949 |
143 |
av_reduce(&h->avctx->framerate.den, &h->avctx->framerate.num, |
|
950 |
143 |
sps->num_units_in_tick * h->avctx->ticks_per_frame, den, 1 << 30); |
|
951 |
} |
||
952 |
|||
953 |
546 |
ff_h264_free_tables(h); |
|
954 |
|||
955 |
546 |
h->first_field = 0; |
|
956 |
546 |
h->prev_interlaced_frame = 1; |
|
957 |
|||
958 |
546 |
init_scan_tables(h); |
|
959 |
546 |
ret = ff_h264_alloc_tables(h); |
|
960 |
✗✓ | 546 |
if (ret < 0) { |
961 |
av_log(h->avctx, AV_LOG_ERROR, "Could not allocate memory\n"); |
||
962 |
goto fail; |
||
963 |
} |
||
964 |
|||
965 |
✓✗✓✗ |
546 |
if (sps->bit_depth_luma < 8 || sps->bit_depth_luma > 14 || |
966 |
✓✗✗✓ |
546 |
sps->bit_depth_luma == 11 || sps->bit_depth_luma == 13 |
967 |
) { |
||
968 |
av_log(h->avctx, AV_LOG_ERROR, "Unsupported bit depth %d\n", |
||
969 |
sps->bit_depth_luma); |
||
970 |
ret = AVERROR_INVALIDDATA; |
||
971 |
goto fail; |
||
972 |
} |
||
973 |
|||
974 |
546 |
h->cur_bit_depth_luma = |
|
975 |
546 |
h->avctx->bits_per_raw_sample = sps->bit_depth_luma; |
|
976 |
546 |
h->cur_chroma_format_idc = sps->chroma_format_idc; |
|
977 |
546 |
h->pixel_shift = sps->bit_depth_luma > 8; |
|
978 |
546 |
h->chroma_format_idc = sps->chroma_format_idc; |
|
979 |
546 |
h->bit_depth_luma = sps->bit_depth_luma; |
|
980 |
|||
981 |
546 |
ff_h264dsp_init(&h->h264dsp, sps->bit_depth_luma, |
|
982 |
sps->chroma_format_idc); |
||
983 |
546 |
ff_h264chroma_init(&h->h264chroma, sps->bit_depth_chroma); |
|
984 |
546 |
ff_h264qpel_init(&h->h264qpel, sps->bit_depth_luma); |
|
985 |
546 |
ff_h264_pred_init(&h->hpc, h->avctx->codec_id, sps->bit_depth_luma, |
|
986 |
sps->chroma_format_idc); |
||
987 |
546 |
ff_videodsp_init(&h->vdsp, sps->bit_depth_luma); |
|
988 |
|||
989 |
✓✓ | 546 |
if (!HAVE_THREADS || !(h->avctx->active_thread_type & FF_THREAD_SLICE)) { |
990 |
545 |
ret = ff_h264_slice_context_init(h, &h->slice_ctx[0]); |
|
991 |
✗✓ | 545 |
if (ret < 0) { |
992 |
av_log(h->avctx, AV_LOG_ERROR, "context_init() failed.\n"); |
||
993 |
goto fail; |
||
994 |
} |
||
995 |
} else { |
||
996 |
✓✓ | 3 |
for (i = 0; i < h->nb_slice_ctx; i++) { |
997 |
2 |
H264SliceContext *sl = &h->slice_ctx[i]; |
|
998 |
|||
999 |
2 |
sl->h264 = h; |
|
1000 |
2 |
sl->intra4x4_pred_mode = h->intra4x4_pred_mode + i * 8 * 2 * h->mb_stride; |
|
1001 |
2 |
sl->mvd_table[0] = h->mvd_table[0] + i * 8 * 2 * h->mb_stride; |
|
1002 |
2 |
sl->mvd_table[1] = h->mvd_table[1] + i * 8 * 2 * h->mb_stride; |
|
1003 |
|||
1004 |
✗✓ | 2 |
if ((ret = ff_h264_slice_context_init(h, sl)) < 0) { |
1005 |
av_log(h->avctx, AV_LOG_ERROR, "context_init() failed.\n"); |
||
1006 |
goto fail; |
||
1007 |
} |
||
1008 |
} |
||
1009 |
} |
||
1010 |
|||
1011 |
546 |
h->context_initialized = 1; |
|
1012 |
|||
1013 |
546 |
return 0; |
|
1014 |
fail: |
||
1015 |
ff_h264_free_tables(h); |
||
1016 |
h->context_initialized = 0; |
||
1017 |
return ret; |
||
1018 |
} |
||
1019 |
|||
1020 |
51630 |
static enum AVPixelFormat non_j_pixfmt(enum AVPixelFormat a) |
|
1021 |
{ |
||
1022 |
✓✗✗✓ |
51630 |
switch (a) { |
1023 |
2 |
case AV_PIX_FMT_YUVJ420P: return AV_PIX_FMT_YUV420P; |
|
1024 |
case AV_PIX_FMT_YUVJ422P: return AV_PIX_FMT_YUV422P; |
||
1025 |
case AV_PIX_FMT_YUVJ444P: return AV_PIX_FMT_YUV444P; |
||
1026 |
51628 |
default: |
|
1027 |
51628 |
return a; |
|
1028 |
} |
||
1029 |
} |
||
1030 |
|||
1031 |
26111 |
static int h264_init_ps(H264Context *h, const H264SliceContext *sl, int first_slice) |
|
1032 |
{ |
||
1033 |
const SPS *sps; |
||
1034 |
26111 |
int needs_reinit = 0, must_reinit, ret; |
|
1035 |
|||
1036 |
✓✓ | 26111 |
if (first_slice) { |
1037 |
25918 |
av_buffer_unref(&h->ps.pps_ref); |
|
1038 |
25918 |
h->ps.pps = NULL; |
|
1039 |
25918 |
h->ps.pps_ref = av_buffer_ref(h->ps.pps_list[sl->pps_id]); |
|
1040 |
✗✓ | 25918 |
if (!h->ps.pps_ref) |
1041 |
return AVERROR(ENOMEM); |
||
1042 |
25918 |
h->ps.pps = (const PPS*)h->ps.pps_ref->data; |
|
1043 |
} |
||
1044 |
|||
1045 |
✓✓ | 26111 |
if (h->ps.sps != h->ps.pps->sps) { |
1046 |
542 |
h->ps.sps = (const SPS*)h->ps.pps->sps; |
|
1047 |
|||
1048 |
✓✓ | 542 |
if (h->mb_width != h->ps.sps->mb_width || |
1049 |
✓✗ | 6 |
h->mb_height != h->ps.sps->mb_height || |
1050 |
✓✓ | 6 |
h->cur_bit_depth_luma != h->ps.sps->bit_depth_luma || |
1051 |
✓✓ | 5 |
h->cur_chroma_format_idc != h->ps.sps->chroma_format_idc |
1052 |
) |
||
1053 |
538 |
needs_reinit = 1; |
|
1054 |
|||
1055 |
✓✓ | 542 |
if (h->bit_depth_luma != h->ps.sps->bit_depth_luma || |
1056 |
✓✓ | 8 |
h->chroma_format_idc != h->ps.sps->chroma_format_idc) |
1057 |
535 |
needs_reinit = 1; |
|
1058 |
} |
||
1059 |
26111 |
sps = h->ps.sps; |
|
1060 |
|||
1061 |
✓✓ | 51690 |
must_reinit = (h->context_initialized && |
1062 |
✓✓ | 25579 |
( 16*sps->mb_width != h->avctx->coded_width |
1063 |
✓✗ | 25575 |
|| 16*sps->mb_height != h->avctx->coded_height |
1064 |
✓✓ | 25575 |
|| h->cur_bit_depth_luma != sps->bit_depth_luma |
1065 |
✓✓ | 25574 |
|| h->cur_chroma_format_idc != sps->chroma_format_idc |
1066 |
✓✗ | 25573 |
|| h->mb_width != sps->mb_width |
1067 |
✗✓ | 25573 |
|| h->mb_height != sps->mb_height |
1068 |
)); |
||
1069 |
✓✓ | 26111 |
if (h->avctx->pix_fmt == AV_PIX_FMT_NONE |
1070 |
✓✓ | 25815 |
|| (non_j_pixfmt(h->avctx->pix_fmt) != non_j_pixfmt(get_pixel_format(h, 0)))) |
1071 |
299 |
must_reinit = 1; |
|
1072 |
|||
1073 |
✓✓✓✓ |
26111 |
if (first_slice && av_cmp_q(sps->sar, h->avctx->sample_aspect_ratio)) |
1074 |
53 |
must_reinit = 1; |
|
1075 |
|||
1076 |
✓✗ | 26111 |
if (!h->setup_finished) { |
1077 |
26111 |
h->avctx->profile = ff_h264_get_profile(sps); |
|
1078 |
26111 |
h->avctx->level = sps->level_idc; |
|
1079 |
26111 |
h->avctx->refs = sps->ref_frame_count; |
|
1080 |
|||
1081 |
26111 |
h->mb_width = sps->mb_width; |
|
1082 |
26111 |
h->mb_height = sps->mb_height; |
|
1083 |
26111 |
h->mb_num = h->mb_width * h->mb_height; |
|
1084 |
26111 |
h->mb_stride = h->mb_width + 1; |
|
1085 |
|||
1086 |
26111 |
h->b_stride = h->mb_width * 4; |
|
1087 |
|||
1088 |
26111 |
h->chroma_y_shift = sps->chroma_format_idc <= 1; // 400 uses yuv420p |
|
1089 |
|||
1090 |
26111 |
h->width = 16 * h->mb_width; |
|
1091 |
26111 |
h->height = 16 * h->mb_height; |
|
1092 |
|||
1093 |
26111 |
init_dimensions(h); |
|
1094 |
|||
1095 |
✓✓ | 26111 |
if (sps->video_signal_type_present_flag) { |
1096 |
1736 |
h->avctx->color_range = sps->full_range > 0 ? AVCOL_RANGE_JPEG |
|
1097 |
✓✓ | 868 |
: AVCOL_RANGE_MPEG; |
1098 |
✓✓ | 868 |
if (sps->colour_description_present_flag) { |
1099 |
✓✓ | 770 |
if (h->avctx->colorspace != sps->colorspace) |
1100 |
14 |
needs_reinit = 1; |
|
1101 |
770 |
h->avctx->color_primaries = sps->color_primaries; |
|
1102 |
770 |
h->avctx->color_trc = sps->color_trc; |
|
1103 |
770 |
h->avctx->colorspace = sps->colorspace; |
|
1104 |
} |
||
1105 |
} |
||
1106 |
|||
1107 |
✗✓✗✗ |
26111 |
if (h->sei.alternative_transfer.present && |
1108 |
av_color_transfer_name(h->sei.alternative_transfer.preferred_transfer_characteristics) && |
||
1109 |
h->sei.alternative_transfer.preferred_transfer_characteristics != AVCOL_TRC_UNSPECIFIED) { |
||
1110 |
h->avctx->color_trc = h->sei.alternative_transfer.preferred_transfer_characteristics; |
||
1111 |
} |
||
1112 |
} |
||
1113 |
26111 |
h->avctx->chroma_sample_location = sps->chroma_location; |
|
1114 |
|||
1115 |
✓✓✓✓ ✗✓ |
26111 |
if (!h->context_initialized || must_reinit || needs_reinit) { |
1116 |
538 |
int flush_changes = h->context_initialized; |
|
1117 |
538 |
h->context_initialized = 0; |
|
1118 |
✗✓ | 538 |
if (sl != h->slice_ctx) { |
1119 |
av_log(h->avctx, AV_LOG_ERROR, |
||
1120 |
"changing width %d -> %d / height %d -> %d on " |
||
1121 |
"slice %d\n", |
||
1122 |
h->width, h->avctx->coded_width, |
||
1123 |
h->height, h->avctx->coded_height, |
||
1124 |
h->current_slice + 1); |
||
1125 |
return AVERROR_INVALIDDATA; |
||
1126 |
} |
||
1127 |
|||
1128 |
av_assert1(first_slice); |
||
1129 |
|||
1130 |
✓✓ | 538 |
if (flush_changes) |
1131 |
6 |
ff_h264_flush_change(h); |
|
1132 |
|||
1133 |
✗✓ | 538 |
if ((ret = get_pixel_format(h, 1)) < 0) |
1134 |
return ret; |
||
1135 |
538 |
h->avctx->pix_fmt = ret; |
|
1136 |
|||
1137 |
538 |
av_log(h->avctx, AV_LOG_VERBOSE, "Reinit context to %dx%d, " |
|
1138 |
538 |
"pix_fmt: %s\n", h->width, h->height, av_get_pix_fmt_name(h->avctx->pix_fmt)); |
|
1139 |
|||
1140 |
✗✓ | 538 |
if ((ret = h264_slice_header_init(h)) < 0) { |
1141 |
av_log(h->avctx, AV_LOG_ERROR, |
||
1142 |
"h264_slice_header_init() failed\n"); |
||
1143 |
return ret; |
||
1144 |
} |
||
1145 |
} |
||
1146 |
|||
1147 |
26111 |
return 0; |
|
1148 |
} |
||
1149 |
|||
1150 |
22774 |
static int h264_export_frame_props(H264Context *h) |
|
1151 |
{ |
||
1152 |
22774 |
const SPS *sps = h->ps.sps; |
|
1153 |
22774 |
H264Picture *cur = h->cur_pic_ptr; |
|
1154 |
22774 |
AVFrame *out = cur->f; |
|
1155 |
|||
1156 |
22774 |
out->interlaced_frame = 0; |
|
1157 |
22774 |
out->repeat_pict = 0; |
|
1158 |
|||
1159 |
/* Signal interlacing information externally. */ |
||
1160 |
/* Prioritize picture timing SEI information over used |
||
1161 |
* decoding process if it exists. */ |
||
1162 |
✓✓ | 22774 |
if (h->sei.picture_timing.present) { |
1163 |
1358 |
int ret = ff_h264_sei_process_picture_timing(&h->sei.picture_timing, sps, |
|
1164 |
1358 |
h->avctx); |
|
1165 |
✗✓ | 1358 |
if (ret < 0) { |
1166 |
av_log(h->avctx, AV_LOG_ERROR, "Error processing a picture timing SEI\n"); |
||
1167 |
if (h->avctx->err_recognition & AV_EF_EXPLODE) |
||
1168 |
return ret; |
||
1169 |
h->sei.picture_timing.present = 0; |
||
1170 |
} |
||
1171 |
} |
||
1172 |
|||
1173 |
✓✓✓✓ |
24124 |
if (sps->pic_struct_present_flag && h->sei.picture_timing.present) { |
1174 |
1350 |
H264SEIPictureTiming *pt = &h->sei.picture_timing; |
|
1175 |
✓✓✓✓ ✗✗✗ |
1350 |
switch (pt->pic_struct) { |
1176 |
132 |
case H264_SEI_PIC_STRUCT_FRAME: |
|
1177 |
132 |
break; |
|
1178 |
347 |
case H264_SEI_PIC_STRUCT_TOP_FIELD: |
|
1179 |
case H264_SEI_PIC_STRUCT_BOTTOM_FIELD: |
||
1180 |
347 |
out->interlaced_frame = 1; |
|
1181 |
347 |
break; |
|
1182 |
704 |
case H264_SEI_PIC_STRUCT_TOP_BOTTOM: |
|
1183 |
case H264_SEI_PIC_STRUCT_BOTTOM_TOP: |
||
1184 |
✓✓✗✓ |
704 |
if (FIELD_OR_MBAFF_PICTURE(h)) |
1185 |
700 |
out->interlaced_frame = 1; |
|
1186 |
else |
||
1187 |
// try to flag soft telecine progressive |
||
1188 |
4 |
out->interlaced_frame = h->prev_interlaced_frame; |
|
1189 |
704 |
break; |
|
1190 |
167 |
case H264_SEI_PIC_STRUCT_TOP_BOTTOM_TOP: |
|
1191 |
case H264_SEI_PIC_STRUCT_BOTTOM_TOP_BOTTOM: |
||
1192 |
/* Signal the possibility of telecined film externally |
||
1193 |
* (pic_struct 5,6). From these hints, let the applications |
||
1194 |
* decide if they apply deinterlacing. */ |
||
1195 |
167 |
out->repeat_pict = 1; |
|
1196 |
167 |
break; |
|
1197 |
case H264_SEI_PIC_STRUCT_FRAME_DOUBLING: |
||
1198 |
out->repeat_pict = 2; |
||
1199 |
break; |
||
1200 |
case H264_SEI_PIC_STRUCT_FRAME_TRIPLING: |
||
1201 |
out->repeat_pict = 4; |
||
1202 |
break; |
||
1203 |
} |
||
1204 |
|||
1205 |
✓✓ | 1350 |
if ((pt->ct_type & 3) && |
1206 |
✓✓ | 651 |
pt->pic_struct <= H264_SEI_PIC_STRUCT_BOTTOM_TOP) |
1207 |
487 |
out->interlaced_frame = (pt->ct_type & (1 << 1)) != 0; |
|
1208 |
} else { |
||
1209 |
/* Derive interlacing flag from used decoding process. */ |
||
1210 |
✓✓✓✓ |
21424 |
out->interlaced_frame = FIELD_OR_MBAFF_PICTURE(h); |
1211 |
} |
||
1212 |
22774 |
h->prev_interlaced_frame = out->interlaced_frame; |
|
1213 |
|||
1214 |
✓✓ | 22774 |
if (cur->field_poc[0] != cur->field_poc[1]) { |
1215 |
/* Derive top_field_first from field pocs. */ |
||
1216 |
3268 |
out->top_field_first = cur->field_poc[0] < cur->field_poc[1]; |
|
1217 |
} else { |
||
1218 |
✓✓✓✓ |
19506 |
if (sps->pic_struct_present_flag && h->sei.picture_timing.present) { |
1219 |
/* Use picture timing SEI information. Even if it is a |
||
1220 |
* information of a past frame, better than nothing. */ |
||
1221 |
✓✓ | 139 |
if (h->sei.picture_timing.pic_struct == H264_SEI_PIC_STRUCT_TOP_BOTTOM || |
1222 |
✓✓ | 138 |
h->sei.picture_timing.pic_struct == H264_SEI_PIC_STRUCT_TOP_BOTTOM_TOP) |
1223 |
3 |
out->top_field_first = 1; |
|
1224 |
else |
||
1225 |
136 |
out->top_field_first = 0; |
|
1226 |
✓✓ | 19367 |
} else if (out->interlaced_frame) { |
1227 |
/* Default to top field first when pic_struct_present_flag |
||
1228 |
* is not set but interlaced frame detected */ |
||
1229 |
2554 |
out->top_field_first = 1; |
|
1230 |
} else { |
||
1231 |
/* Most likely progressive */ |
||
1232 |
16813 |
out->top_field_first = 0; |
|
1233 |
} |
||
1234 |
} |
||
1235 |
|||
1236 |
✗✓ | 22774 |
if (h->sei.frame_packing.present && |
1237 |
h->sei.frame_packing.arrangement_type <= 6 && |
||
1238 |
h->sei.frame_packing.content_interpretation_type > 0 && |
||
1239 |
h->sei.frame_packing.content_interpretation_type < 3) { |
||
1240 |
H264SEIFramePacking *fp = &h->sei.frame_packing; |
||
1241 |
AVStereo3D *stereo = av_stereo3d_create_side_data(out); |
||
1242 |
if (stereo) { |
||
1243 |
switch (fp->arrangement_type) { |
||
1244 |
case H264_SEI_FPA_TYPE_CHECKERBOARD: |
||
1245 |
stereo->type = AV_STEREO3D_CHECKERBOARD; |
||
1246 |
break; |
||
1247 |
case H264_SEI_FPA_TYPE_INTERLEAVE_COLUMN: |
||
1248 |
stereo->type = AV_STEREO3D_COLUMNS; |
||
1249 |
break; |
||
1250 |
case H264_SEI_FPA_TYPE_INTERLEAVE_ROW: |
||
1251 |
stereo->type = AV_STEREO3D_LINES; |
||
1252 |
break; |
||
1253 |
case H264_SEI_FPA_TYPE_SIDE_BY_SIDE: |
||
1254 |
if (fp->quincunx_sampling_flag) |
||
1255 |
stereo->type = AV_STEREO3D_SIDEBYSIDE_QUINCUNX; |
||
1256 |
else |
||
1257 |
stereo->type = AV_STEREO3D_SIDEBYSIDE; |
||
1258 |
break; |
||
1259 |
case H264_SEI_FPA_TYPE_TOP_BOTTOM: |
||
1260 |
stereo->type = AV_STEREO3D_TOPBOTTOM; |
||
1261 |
break; |
||
1262 |
case H264_SEI_FPA_TYPE_INTERLEAVE_TEMPORAL: |
||
1263 |
stereo->type = AV_STEREO3D_FRAMESEQUENCE; |
||
1264 |
break; |
||
1265 |
case H264_SEI_FPA_TYPE_2D: |
||
1266 |
stereo->type = AV_STEREO3D_2D; |
||
1267 |
break; |
||
1268 |
} |
||
1269 |
|||
1270 |
if (fp->content_interpretation_type == 2) |
||
1271 |
stereo->flags = AV_STEREO3D_FLAG_INVERT; |
||
1272 |
|||
1273 |
if (fp->arrangement_type == H264_SEI_FPA_TYPE_INTERLEAVE_TEMPORAL) { |
||
1274 |
if (fp->current_frame_is_frame0_flag) |
||
1275 |
stereo->view = AV_STEREO3D_VIEW_LEFT; |
||
1276 |
else |
||
1277 |
stereo->view = AV_STEREO3D_VIEW_RIGHT; |
||
1278 |
} |
||
1279 |
} |
||
1280 |
} |
||
1281 |
|||
1282 |
✓✓ | 22774 |
if (h->sei.display_orientation.present && |
1283 |
✓✗ | 1 |
(h->sei.display_orientation.anticlockwise_rotation || |
1284 |
✓✗ | 1 |
h->sei.display_orientation.hflip || |
1285 |
✗✓ | 1 |
h->sei.display_orientation.vflip)) { |
1286 |
H264SEIDisplayOrientation *o = &h->sei.display_orientation; |
||
1287 |
double angle = o->anticlockwise_rotation * 360 / (double) (1 << 16); |
||
1288 |
AVFrameSideData *rotation = av_frame_new_side_data(out, |
||
1289 |
AV_FRAME_DATA_DISPLAYMATRIX, |
||
1290 |
sizeof(int32_t) * 9); |
||
1291 |
if (rotation) { |
||
1292 |
av_display_rotation_set((int32_t *)rotation->data, angle); |
||
1293 |
av_display_matrix_flip((int32_t *)rotation->data, |
||
1294 |
o->hflip, o->vflip); |
||
1295 |
} |
||
1296 |
} |
||
1297 |
|||
1298 |
✓✓ | 22774 |
if (h->sei.afd.present) { |
1299 |
239 |
AVFrameSideData *sd = av_frame_new_side_data(out, AV_FRAME_DATA_AFD, |
|
1300 |
sizeof(uint8_t)); |
||
1301 |
|||
1302 |
✓✗ | 239 |
if (sd) { |
1303 |
239 |
*sd->data = h->sei.afd.active_format_description; |
|
1304 |
239 |
h->sei.afd.present = 0; |
|
1305 |
} |
||
1306 |
} |
||
1307 |
|||
1308 |
✓✓ | 22774 |
if (h->sei.a53_caption.buf_ref) { |
1309 |
1 |
H264SEIA53Caption *a53 = &h->sei.a53_caption; |
|
1310 |
|||
1311 |
1 |
AVFrameSideData *sd = av_frame_new_side_data_from_buf(out, AV_FRAME_DATA_A53_CC, a53->buf_ref); |
|
1312 |
✗✓ | 1 |
if (!sd) |
1313 |
av_buffer_unref(&a53->buf_ref); |
||
1314 |
1 |
a53->buf_ref = NULL; |
|
1315 |
|||
1316 |
1 |
h->avctx->properties |= FF_CODEC_PROPERTY_CLOSED_CAPTIONS; |
|
1317 |
} |
||
1318 |
|||
1319 |
✓✓ | 22948 |
for (int i = 0; i < h->sei.unregistered.nb_buf_ref; i++) { |
1320 |
174 |
H264SEIUnregistered *unreg = &h->sei.unregistered; |
|
1321 |
|||
1322 |
✓✗ | 174 |
if (unreg->buf_ref[i]) { |
1323 |
174 |
AVFrameSideData *sd = av_frame_new_side_data_from_buf(out, |
|
1324 |
AV_FRAME_DATA_SEI_UNREGISTERED, |
||
1325 |
174 |
unreg->buf_ref[i]); |
|
1326 |
✗✓ | 174 |
if (!sd) |
1327 |
av_buffer_unref(&unreg->buf_ref[i]); |
||
1328 |
174 |
unreg->buf_ref[i] = NULL; |
|
1329 |
} |
||
1330 |
} |
||
1331 |
22774 |
h->sei.unregistered.nb_buf_ref = 0; |
|
1332 |
|||
1333 |
✓✓ | 22774 |
if (h->sei.picture_timing.timecode_cnt > 0) { |
1334 |
uint32_t *tc_sd; |
||
1335 |
char tcbuf[AV_TIMECODE_STR_SIZE]; |
||
1336 |
|||
1337 |
651 |
AVFrameSideData *tcside = av_frame_new_side_data(out, |
|
1338 |
AV_FRAME_DATA_S12M_TIMECODE, |
||
1339 |
sizeof(uint32_t)*4); |
||
1340 |
✗✓ | 651 |
if (!tcside) |
1341 |
return AVERROR(ENOMEM); |
||
1342 |
|||
1343 |
651 |
tc_sd = (uint32_t*)tcside->data; |
|
1344 |
651 |
tc_sd[0] = h->sei.picture_timing.timecode_cnt; |
|
1345 |
|||
1346 |
✓✓ | 1603 |
for (int i = 0; i < tc_sd[0]; i++) { |
1347 |
952 |
int drop = h->sei.picture_timing.timecode[i].dropframe; |
|
1348 |
952 |
int hh = h->sei.picture_timing.timecode[i].hours; |
|
1349 |
952 |
int mm = h->sei.picture_timing.timecode[i].minutes; |
|
1350 |
952 |
int ss = h->sei.picture_timing.timecode[i].seconds; |
|
1351 |
952 |
int ff = h->sei.picture_timing.timecode[i].frame; |
|
1352 |
|||
1353 |
952 |
tc_sd[i + 1] = av_timecode_get_smpte(h->avctx->framerate, drop, hh, mm, ss, ff); |
|
1354 |
952 |
av_timecode_make_smpte_tc_string2(tcbuf, h->avctx->framerate, tc_sd[i + 1], 0, 0); |
|
1355 |
952 |
av_dict_set(&out->metadata, "timecode", tcbuf, 0); |
|
1356 |
} |
||
1357 |
651 |
h->sei.picture_timing.timecode_cnt = 0; |
|
1358 |
} |
||
1359 |
|||
1360 |
22774 |
return 0; |
|
1361 |
} |
||
1362 |
|||
1363 |
22774 |
static int h264_select_output_frame(H264Context *h) |
|
1364 |
{ |
||
1365 |
22774 |
const SPS *sps = h->ps.sps; |
|
1366 |
22774 |
H264Picture *out = h->cur_pic_ptr; |
|
1367 |
22774 |
H264Picture *cur = h->cur_pic_ptr; |
|
1368 |
int i, pics, out_of_order, out_idx; |
||
1369 |
|||
1370 |
22774 |
cur->mmco_reset = h->mmco_reset; |
|
1371 |
22774 |
h->mmco_reset = 0; |
|
1372 |
|||
1373 |
✓✓ | 22774 |
if (sps->bitstream_restriction_flag || |
1374 |
✗✓ | 19459 |
h->avctx->strict_std_compliance >= FF_COMPLIANCE_STRICT) { |
1375 |
3315 |
h->avctx->has_b_frames = FFMAX(h->avctx->has_b_frames, sps->num_reorder_frames); |
|
1376 |
} |
||
1377 |
|||
1378 |
372772 |
for (i = 0; 1; i++) { |
|
1379 |
✓✓✓✓ |
372772 |
if(i == MAX_DELAYED_PIC_COUNT || cur->poc < h->last_pocs[i]){ |
1380 |
✓✗ | 22774 |
if(i) |
1381 |
22774 |
h->last_pocs[i-1] = cur->poc; |
|
1382 |
22774 |
break; |
|
1383 |
✓✓ | 349998 |
} else if(i) { |
1384 |
327224 |
h->last_pocs[i-1]= h->last_pocs[i]; |
|
1385 |
} |
||
1386 |
} |
||
1387 |
22774 |
out_of_order = MAX_DELAYED_PIC_COUNT - i; |
|
1388 |
✓✓ | 22774 |
if( cur->f->pict_type == AV_PICTURE_TYPE_B |
1389 |
✓✓✓✓ |
15257 |
|| (h->last_pocs[MAX_DELAYED_PIC_COUNT-2] > INT_MIN && h->last_pocs[MAX_DELAYED_PIC_COUNT-1] - (int64_t)h->last_pocs[MAX_DELAYED_PIC_COUNT-2] > 2)) |
1390 |
12110 |
out_of_order = FFMAX(out_of_order, 1); |
|
1391 |
✗✓ | 22774 |
if (out_of_order == MAX_DELAYED_PIC_COUNT) { |
1392 |
av_log(h->avctx, AV_LOG_VERBOSE, "Invalid POC %d<%d\n", cur->poc, h->last_pocs[0]); |
||
1393 |
for (i = 1; i < MAX_DELAYED_PIC_COUNT; i++) |
||
1394 |
h->last_pocs[i] = INT_MIN; |
||
1395 |
h->last_pocs[0] = cur->poc; |
||
1396 |
cur->mmco_reset = 1; |
||
1397 |
✓✓✓✗ |
22774 |
} else if(h->avctx->has_b_frames < out_of_order && !sps->bitstream_restriction_flag){ |
1398 |
✓✓ | 136 |
int loglevel = h->avctx->frame_number > 1 ? AV_LOG_WARNING : AV_LOG_VERBOSE; |
1399 |
136 |
av_log(h->avctx, loglevel, "Increasing reorder buffer to %d\n", out_of_order); |
|
1400 |
136 |
h->avctx->has_b_frames = out_of_order; |
|
1401 |
} |
||
1402 |
|||
1403 |
22774 |
pics = 0; |
|
1404 |
✓✓ | 51757 |
while (h->delayed_pic[pics]) |
1405 |
28983 |
pics++; |
|
1406 |
|||
1407 |
✗✓ | 22774 |
av_assert0(pics <= MAX_DELAYED_PIC_COUNT); |
1408 |
|||
1409 |
22774 |
h->delayed_pic[pics++] = cur; |
|
1410 |
✓✓ | 22774 |
if (cur->reference == 0) |
1411 |
21314 |
cur->reference = DELAYED_PIC_REF; |
|
1412 |
|||
1413 |
22774 |
out = h->delayed_pic[0]; |
|
1414 |
22774 |
out_idx = 0; |
|
1415 |
✓✓ | 48862 |
for (i = 1; h->delayed_pic[i] && |
1416 |
✓✓ | 26534 |
!h->delayed_pic[i]->f->key_frame && |
1417 |
✓✓ | 26175 |
!h->delayed_pic[i]->mmco_reset; |
1418 |
26088 |
i++) |
|
1419 |
✓✓ | 26088 |
if (h->delayed_pic[i]->poc < out->poc) { |
1420 |
9697 |
out = h->delayed_pic[i]; |
|
1421 |
9697 |
out_idx = i; |
|
1422 |
} |
||
1423 |
✓✓ | 22774 |
if (h->avctx->has_b_frames == 0 && |
1424 |
✓✓✓✓ |
9635 |
(h->delayed_pic[0]->f->key_frame || h->delayed_pic[0]->mmco_reset)) |
1425 |
783 |
h->next_outputed_poc = INT_MIN; |
|
1426 |
22774 |
out_of_order = out->poc < h->next_outputed_poc; |
|
1427 |
|||
1428 |
✓✓✓✓ |
22774 |
if (out_of_order || pics > h->avctx->has_b_frames) { |
1429 |
22289 |
out->reference &= ~DELAYED_PIC_REF; |
|
1430 |
✓✓ | 58780 |
for (i = out_idx; h->delayed_pic[i]; i++) |
1431 |
36491 |
h->delayed_pic[i] = h->delayed_pic[i + 1]; |
|
1432 |
} |
||
1433 |
✓✓✓✓ |
22774 |
if (!out_of_order && pics > h->avctx->has_b_frames) { |
1434 |
22175 |
h->next_output_pic = out; |
|
1435 |
✓✓✓✓ ✓✓✓✓ |
22175 |
if (out_idx == 0 && h->delayed_pic[0] && (h->delayed_pic[0]->f->key_frame || h->delayed_pic[0]->mmco_reset)) { |
1436 |
86 |
h->next_outputed_poc = INT_MIN; |
|
1437 |
} else |
||
1438 |
22089 |
h->next_outputed_poc = out->poc; |
|
1439 |
|||
1440 |
✓✓ | 22175 |
if (out->recovered) { |
1441 |
// We have reached an recovery point and all frames after it in |
||
1442 |
// display order are "recovered". |
||
1443 |
22086 |
h->frame_recovered |= FRAME_RECOVERED_SEI; |
|
1444 |
} |
||
1445 |
22175 |
out->recovered |= !!(h->frame_recovered & FRAME_RECOVERED_SEI); |
|
1446 |
|||
1447 |
✓✓ | 22175 |
if (!out->recovered) { |
1448 |
✓✗ | 78 |
if (!(h->avctx->flags & AV_CODEC_FLAG_OUTPUT_CORRUPT) && |
1449 |
✓✗ | 78 |
!(h->avctx->flags2 & AV_CODEC_FLAG2_SHOW_ALL)) { |
1450 |
78 |
h->next_output_pic = NULL; |
|
1451 |
} else { |
||
1452 |
out->f->flags |= AV_FRAME_FLAG_CORRUPT; |
||
1453 |
} |
||
1454 |
} |
||
1455 |
} else { |
||
1456 |
✓✓ | 599 |
av_log(h->avctx, AV_LOG_DEBUG, "no picture %s\n", out_of_order ? "ooo" : ""); |
1457 |
} |
||
1458 |
|||
1459 |
22774 |
return 0; |
|
1460 |
} |
||
1461 |
|||
1462 |
/* This function is called right after decoding the slice header for a first |
||
1463 |
* slice in a field (or a frame). It decides whether we are decoding a new frame |
||
1464 |
* or a second field in a pair and does the necessary setup. |
||
1465 |
*/ |
||
1466 |
26111 |
static int h264_field_start(H264Context *h, const H264SliceContext *sl, |
|
1467 |
const H2645NAL *nal, int first_slice) |
||
1468 |
{ |
||
1469 |
int i; |
||
1470 |
const SPS *sps; |
||
1471 |
|||
1472 |
int last_pic_structure, last_pic_droppable, ret; |
||
1473 |
|||
1474 |
26111 |
ret = h264_init_ps(h, sl, first_slice); |
|
1475 |
✗✓ | 26111 |
if (ret < 0) |
1476 |
return ret; |
||
1477 |
|||
1478 |
26111 |
sps = h->ps.sps; |
|
1479 |
|||
1480 |
✓✗✓✓ |
26111 |
if (sps && sps->bitstream_restriction_flag && |
1481 |
✓✓ | 3647 |
h->avctx->has_b_frames < sps->num_reorder_frames) { |
1482 |
25 |
h->avctx->has_b_frames = sps->num_reorder_frames; |
|
1483 |
} |
||
1484 |
|||
1485 |
26111 |
last_pic_droppable = h->droppable; |
|
1486 |
26111 |
last_pic_structure = h->picture_structure; |
|
1487 |
26111 |
h->droppable = (nal->ref_idc == 0); |
|
1488 |
26111 |
h->picture_structure = sl->picture_structure; |
|
1489 |
|||
1490 |
26111 |
h->poc.frame_num = sl->frame_num; |
|
1491 |
26111 |
h->poc.poc_lsb = sl->poc_lsb; |
|
1492 |
26111 |
h->poc.delta_poc_bottom = sl->delta_poc_bottom; |
|
1493 |
26111 |
h->poc.delta_poc[0] = sl->delta_poc[0]; |
|
1494 |
26111 |
h->poc.delta_poc[1] = sl->delta_poc[1]; |
|
1495 |
|||
1496 |
/* Shorten frame num gaps so we don't have to allocate reference |
||
1497 |
* frames just to throw them away */ |
||
1498 |
✓✓ | 26111 |
if (h->poc.frame_num != h->poc.prev_frame_num) { |
1499 |
14645 |
int unwrap_prev_frame_num = h->poc.prev_frame_num; |
|
1500 |
14645 |
int max_frame_num = 1 << sps->log2_max_frame_num; |
|
1501 |
|||
1502 |
✓✓ | 14645 |
if (unwrap_prev_frame_num > h->poc.frame_num) |
1503 |
145 |
unwrap_prev_frame_num -= max_frame_num; |
|
1504 |
|||
1505 |
✓✓ | 14645 |
if ((h->poc.frame_num - unwrap_prev_frame_num) > sps->ref_frame_count) { |
1506 |
73 |
unwrap_prev_frame_num = (h->poc.frame_num - sps->ref_frame_count) - 1; |
|
1507 |
✓✓ | 73 |
if (unwrap_prev_frame_num < 0) |
1508 |
2 |
unwrap_prev_frame_num += max_frame_num; |
|
1509 |
|||
1510 |
73 |
h->poc.prev_frame_num = unwrap_prev_frame_num; |
|
1511 |
} |
||
1512 |
} |
||
1513 |
|||
1514 |
/* See if we have a decoded first field looking for a pair... |
||
1515 |
* Here, we're using that to see if we should mark previously |
||
1516 |
* decode frames as "finished". |
||
1517 |
* We have to do that before the "dummy" in-between frame allocation, |
||
1518 |
* since that can modify h->cur_pic_ptr. */ |
||
1519 |
✓✓ | 26111 |
if (h->first_field) { |
1520 |
3332 |
int last_field = last_pic_structure == PICT_BOTTOM_FIELD; |
|
1521 |
✗✓ | 3332 |
av_assert0(h->cur_pic_ptr); |
1522 |
✗✓ | 3332 |
av_assert0(h->cur_pic_ptr->f->buf[0]); |
1523 |
assert(h->cur_pic_ptr->reference != DELAYED_PIC_REF); |
||
1524 |
|||
1525 |
/* Mark old field/frame as completed */ |
||
1526 |
✓✗ | 3332 |
if (h->cur_pic_ptr->tf.owner[last_field] == h->avctx) { |
1527 |
3332 |
ff_thread_report_progress(&h->cur_pic_ptr->tf, INT_MAX, last_field); |
|
1528 |
} |
||
1529 |
|||
1530 |
/* figure out if we have a complementary field pair */ |
||
1531 |
✓✗✗✓ |
3332 |
if (!FIELD_PICTURE(h) || h->picture_structure == last_pic_structure) { |
1532 |
/* Previous field is unmatched. Don't display it, but let it |
||
1533 |
* remain for reference if marked as such. */ |
||
1534 |
if (last_pic_structure != PICT_FRAME) { |
||
1535 |
ff_thread_report_progress(&h->cur_pic_ptr->tf, INT_MAX, |
||
1536 |
last_pic_structure == PICT_TOP_FIELD); |
||
1537 |
} |
||
1538 |
} else { |
||
1539 |
✗✓ | 3332 |
if (h->cur_pic_ptr->frame_num != h->poc.frame_num) { |
1540 |
/* This and previous field were reference, but had |
||
1541 |
* different frame_nums. Consider this field first in |
||
1542 |
* pair. Throw away previous field except for reference |
||
1543 |
* purposes. */ |
||
1544 |
if (last_pic_structure != PICT_FRAME) { |
||
1545 |
ff_thread_report_progress(&h->cur_pic_ptr->tf, INT_MAX, |
||
1546 |
last_pic_structure == PICT_TOP_FIELD); |
||
1547 |
} |
||
1548 |
} else { |
||
1549 |
/* Second field in complementary pair */ |
||
1550 |
✓✓✓✗ |
3332 |
if (!((last_pic_structure == PICT_TOP_FIELD && |
1551 |
✗✓ | 3324 |
h->picture_structure == PICT_BOTTOM_FIELD) || |
1552 |
8 |
(last_pic_structure == PICT_BOTTOM_FIELD && |
|
1553 |
✗✓ | 8 |
h->picture_structure == PICT_TOP_FIELD))) { |
1554 |
av_log(h->avctx, AV_LOG_ERROR, |
||
1555 |
"Invalid field mode combination %d/%d\n", |
||
1556 |
last_pic_structure, h->picture_structure); |
||
1557 |
h->picture_structure = last_pic_structure; |
||
1558 |
h->droppable = last_pic_droppable; |
||
1559 |
return AVERROR_INVALIDDATA; |
||
1560 |
✗✓ | 3332 |
} else if (last_pic_droppable != h->droppable) { |
1561 |
avpriv_request_sample(h->avctx, |
||
1562 |
"Found reference and non-reference fields in the same frame, which"); |
||
1563 |
h->picture_structure = last_pic_structure; |
||
1564 |
h->droppable = last_pic_droppable; |
||
1565 |
return AVERROR_PATCHWELCOME; |
||
1566 |
} |
||
1567 |
} |
||
1568 |
} |
||
1569 |
} |
||
1570 |
|||
1571 |
✓✓✓✗ |
26804 |
while (h->poc.frame_num != h->poc.prev_frame_num && !h->first_field && |
1572 |
✓✓ | 15338 |
h->poc.frame_num != (h->poc.prev_frame_num + 1) % (1 << sps->log2_max_frame_num)) { |
1573 |
✓✓ | 693 |
H264Picture *prev = h->short_ref_count ? h->short_ref[0] : NULL; |
1574 |
693 |
av_log(h->avctx, AV_LOG_DEBUG, "Frame num gap %d %d\n", |
|
1575 |
h->poc.frame_num, h->poc.prev_frame_num); |
||
1576 |
✓✓ | 693 |
if (!sps->gaps_in_frame_num_allowed_flag) |
1577 |
✓✓ | 1972 |
for(i=0; i<FF_ARRAY_ELEMS(h->last_pocs); i++) |
1578 |
1856 |
h->last_pocs[i] = INT_MIN; |
|
1579 |
693 |
ret = h264_frame_start(h); |
|
1580 |
✗✓ | 693 |
if (ret < 0) { |
1581 |
h->first_field = 0; |
||
1582 |
return ret; |
||
1583 |
} |
||
1584 |
|||
1585 |
693 |
h->poc.prev_frame_num++; |
|
1586 |
693 |
h->poc.prev_frame_num %= 1 << sps->log2_max_frame_num; |
|
1587 |
693 |
h->cur_pic_ptr->frame_num = h->poc.prev_frame_num; |
|
1588 |
693 |
h->cur_pic_ptr->invalid_gap = !sps->gaps_in_frame_num_allowed_flag; |
|
1589 |
693 |
ff_thread_report_progress(&h->cur_pic_ptr->tf, INT_MAX, 0); |
|
1590 |
693 |
ff_thread_report_progress(&h->cur_pic_ptr->tf, INT_MAX, 1); |
|
1591 |
|||
1592 |
693 |
h->explicit_ref_marking = 0; |
|
1593 |
693 |
ret = ff_h264_execute_ref_pic_marking(h); |
|
1594 |
✗✓✗✗ |
693 |
if (ret < 0 && (h->avctx->err_recognition & AV_EF_EXPLODE)) |
1595 |
return ret; |
||
1596 |
/* Error concealment: If a ref is missing, copy the previous ref |
||
1597 |
* in its place. |
||
1598 |
* FIXME: Avoiding a memcpy would be nice, but ref handling makes |
||
1599 |
* many assumptions about there being no actual duplicates. |
||
1600 |
* FIXME: This does not copy padding for out-of-frame motion |
||
1601 |
* vectors. Given we are concealing a lost frame, this probably |
||
1602 |
* is not noticeable by comparison, but it should be fixed. */ |
||
1603 |
✓✗ | 693 |
if (h->short_ref_count) { |
1604 |
693 |
int c[4] = { |
|
1605 |
693 |
1<<(h->ps.sps->bit_depth_luma-1), |
|
1606 |
693 |
1<<(h->ps.sps->bit_depth_chroma-1), |
|
1607 |
693 |
1<<(h->ps.sps->bit_depth_chroma-1), |
|
1608 |
-1 |
||
1609 |
}; |
||
1610 |
|||
1611 |
✓✓ | 693 |
if (prev && |
1612 |
✓✗ | 669 |
h->short_ref[0]->f->width == prev->f->width && |
1613 |
✓✗ | 669 |
h->short_ref[0]->f->height == prev->f->height && |
1614 |
✓✗ | 669 |
h->short_ref[0]->f->format == prev->f->format) { |
1615 |
669 |
ff_thread_await_progress(&prev->tf, INT_MAX, 0); |
|
1616 |
✓✓ | 669 |
if (prev->field_picture) |
1617 |
28 |
ff_thread_await_progress(&prev->tf, INT_MAX, 1); |
|
1618 |
669 |
ff_thread_release_buffer(h->avctx, &h->short_ref[0]->tf); |
|
1619 |
669 |
h->short_ref[0]->tf.f = h->short_ref[0]->f; |
|
1620 |
669 |
ret = ff_thread_ref_frame(&h->short_ref[0]->tf, &prev->tf); |
|
1621 |
✗✓ | 669 |
if (ret < 0) |
1622 |
return ret; |
||
1623 |
669 |
h->short_ref[0]->poc = prev->poc + 2U; |
|
1624 |
669 |
ff_thread_report_progress(&h->short_ref[0]->tf, INT_MAX, 0); |
|
1625 |
✓✓ | 669 |
if (h->short_ref[0]->field_picture) |
1626 |
28 |
ff_thread_report_progress(&h->short_ref[0]->tf, INT_MAX, 1); |
|
1627 |
✓✓✓✗ |
24 |
} else if (!h->frame_recovered && !h->avctx->hwaccel) |
1628 |
17 |
ff_color_frame(h->short_ref[0]->f, c); |
|
1629 |
693 |
h->short_ref[0]->frame_num = h->poc.prev_frame_num; |
|
1630 |
} |
||
1631 |
} |
||
1632 |
|||
1633 |
/* See if we have a decoded first field looking for a pair... |
||
1634 |
* We're using that to see whether to continue decoding in that |
||
1635 |
* frame, or to allocate a new one. */ |
||
1636 |
✓✓ | 26111 |
if (h->first_field) { |
1637 |
✗✓ | 3332 |
av_assert0(h->cur_pic_ptr); |
1638 |
✗✓ | 3332 |
av_assert0(h->cur_pic_ptr->f->buf[0]); |
1639 |
assert(h->cur_pic_ptr->reference != DELAYED_PIC_REF); |
||
1640 |
|||
1641 |
/* figure out if we have a complementary field pair */ |
||
1642 |
✓✗✗✓ |
3332 |
if (!FIELD_PICTURE(h) || h->picture_structure == last_pic_structure) { |
1643 |
/* Previous field is unmatched. Don't display it, but let it |
||
1644 |
* remain for reference if marked as such. */ |
||
1645 |
h->missing_fields ++; |
||
1646 |
h->cur_pic_ptr = NULL; |
||
1647 |
h->first_field = FIELD_PICTURE(h); |
||
1648 |
} else { |
||
1649 |
3332 |
h->missing_fields = 0; |
|
1650 |
✗✓ | 3332 |
if (h->cur_pic_ptr->frame_num != h->poc.frame_num) { |
1651 |
ff_thread_report_progress(&h->cur_pic_ptr->tf, INT_MAX, |
||
1652 |
h->picture_structure==PICT_BOTTOM_FIELD); |
||
1653 |
/* This and the previous field had different frame_nums. |
||
1654 |
* Consider this field first in pair. Throw away previous |
||
1655 |
* one except for reference purposes. */ |
||
1656 |
h->first_field = 1; |
||
1657 |
h->cur_pic_ptr = NULL; |
||
1658 |
✗✓ | 3332 |
} else if (h->cur_pic_ptr->reference & DELAYED_PIC_REF) { |
1659 |
/* This frame was already output, we cannot draw into it |
||
1660 |
* anymore. |
||
1661 |
*/ |
||
1662 |
h->first_field = 1; |
||
1663 |
h->cur_pic_ptr = NULL; |
||
1664 |
} else { |
||
1665 |
/* Second field in complementary pair */ |
||
1666 |
3332 |
h->first_field = 0; |
|
1667 |
} |
||
1668 |
} |
||
1669 |
} else { |
||
1670 |
/* Frame or first field in a potentially complementary pair */ |
||
1671 |
22779 |
h->first_field = FIELD_PICTURE(h); |
|
1672 |
} |
||
1673 |
|||
1674 |
✓✓✓✓ |
26111 |
if (!FIELD_PICTURE(h) || h->first_field) { |
1675 |
✗✓ | 22779 |
if (h264_frame_start(h) < 0) { |
1676 |
h->first_field = 0; |
||
1677 |
return AVERROR_INVALIDDATA; |
||
1678 |
} |
||
1679 |
} else { |
||
1680 |
3332 |
int field = h->picture_structure == PICT_BOTTOM_FIELD; |
|
1681 |
3332 |
release_unused_pictures(h, 0); |
|
1682 |
3332 |
h->cur_pic_ptr->tf.owner[field] = h->avctx; |
|
1683 |
} |
||
1684 |
/* Some macroblocks can be accessed before they're available in case |
||
1685 |
* of lost slices, MBAFF or threading. */ |
||
1686 |
✓✓ | 26111 |
if (FIELD_PICTURE(h)) { |
1687 |
✓✓ | 170961 |
for(i = (h->picture_structure == PICT_BOTTOM_FIELD); i<h->mb_height; i++) |
1688 |
164292 |
memset(h->slice_table + i*h->mb_stride, -1, (h->mb_stride - (i+1==h->mb_height)) * sizeof(*h->slice_table)); |
|
1689 |
} else { |
||
1690 |
19442 |
memset(h->slice_table, -1, |
|
1691 |
19442 |
(h->mb_height * h->mb_stride - 1) * sizeof(*h->slice_table)); |
|
1692 |
} |
||
1693 |
|||
1694 |
26111 |
ret = ff_h264_init_poc(h->cur_pic_ptr->field_poc, &h->cur_pic_ptr->poc, |
|
1695 |
h->ps.sps, &h->poc, h->picture_structure, nal->ref_idc); |
||
1696 |
✗✓ | 26111 |
if (ret < 0) |
1697 |
return ret; |
||
1698 |
|||
1699 |
26111 |
memcpy(h->mmco, sl->mmco, sl->nb_mmco * sizeof(*h->mmco)); |
|
1700 |
26111 |
h->nb_mmco = sl->nb_mmco; |
|
1701 |
26111 |
h->explicit_ref_marking = sl->explicit_ref_marking; |
|
1702 |
|||
1703 |
26111 |
h->picture_idr = nal->type == H264_NAL_IDR_SLICE; |
|
1704 |
|||
1705 |
✓✓ | 26111 |
if (h->sei.recovery_point.recovery_frame_cnt >= 0) { |
1706 |
60 |
const int sei_recovery_frame_cnt = h->sei.recovery_point.recovery_frame_cnt; |
|
1707 |
|||
1708 |
✓✓✓✓ |
60 |
if (h->poc.frame_num != sei_recovery_frame_cnt || sl->slice_type_nos != AV_PICTURE_TYPE_I) |
1709 |
47 |
h->valid_recovery_point = 1; |
|
1710 |
|||
1711 |
✓✓ | 60 |
if ( h->recovery_frame < 0 |
1712 |
✗✓ | 1 |
|| av_mod_uintp2(h->recovery_frame - h->poc.frame_num, h->ps.sps->log2_max_frame_num) > sei_recovery_frame_cnt) { |
1713 |
59 |
h->recovery_frame = av_mod_uintp2(h->poc.frame_num + sei_recovery_frame_cnt, h->ps.sps->log2_max_frame_num); |
|
1714 |
|||
1715 |
✓✓ | 59 |
if (!h->valid_recovery_point) |
1716 |
6 |
h->recovery_frame = h->poc.frame_num; |
|
1717 |
} |
||
1718 |
} |
||
1719 |
|||
1720 |
26111 |
h->cur_pic_ptr->f->key_frame |= (nal->type == H264_NAL_IDR_SLICE); |
|
1721 |
|||
1722 |
✓✓ | 26111 |
if (nal->type == H264_NAL_IDR_SLICE || |
1723 |
✓✓✓✓ |
25097 |
(h->recovery_frame == h->poc.frame_num && nal->ref_idc)) { |
1724 |
1061 |
h->recovery_frame = -1; |
|
1725 |
1061 |
h->cur_pic_ptr->recovered = 1; |
|
1726 |
} |
||
1727 |
// If we have an IDR, all frames after it in decoded order are |
||
1728 |
// "recovered". |
||
1729 |
✓✓ | 26111 |
if (nal->type == H264_NAL_IDR_SLICE) |
1730 |
1014 |
h->frame_recovered |= FRAME_RECOVERED_IDR; |
|
1731 |
#if 1 |
||
1732 |
26111 |
h->cur_pic_ptr->recovered |= h->frame_recovered; |
|
1733 |
#else |
||
1734 |
h->cur_pic_ptr->recovered |= !!(h->frame_recovered & FRAME_RECOVERED_IDR); |
||
1735 |
#endif |
||
1736 |
|||
1737 |
/* Set the frame properties/side data. Only done for the second field in |
||
1738 |
* field coded frames, since some SEI information is present for each field |
||
1739 |
* and is merged by the SEI parsing code. */ |
||
1740 |
✓✓✓✓ ✗✓ |
26111 |
if (!FIELD_PICTURE(h) || !h->first_field || h->missing_fields > 1) { |
1741 |
22774 |
ret = h264_export_frame_props(h); |
|
1742 |
✗✓ | 22774 |
if (ret < 0) |
1743 |
return ret; |
||
1744 |
|||
1745 |
22774 |
ret = h264_select_output_frame(h); |
|
1746 |
✗✓ | 22774 |
if (ret < 0) |
1747 |
return ret; |
||
1748 |
} |
||
1749 |
|||
1750 |
26111 |
return 0; |
|
1751 |
} |
||
1752 |
|||
1753 |
34685 |
static int h264_slice_header_parse(const H264Context *h, H264SliceContext *sl, |
|
1754 |
const H2645NAL *nal) |
||
1755 |
{ |
||
1756 |
const SPS *sps; |
||
1757 |
const PPS *pps; |
||
1758 |
int ret; |
||
1759 |
unsigned int slice_type, tmp, i; |
||
1760 |
int field_pic_flag, bottom_field_flag; |
||
1761 |
✓✓✓✓ |
34685 |
int first_slice = sl == h->slice_ctx && !h->current_slice; |
1762 |
int picture_structure; |
||
1763 |
|||
1764 |
✓✓ | 34685 |
if (first_slice) |
1765 |
✗✓ | 26378 |
av_assert0(!h->setup_finished); |
1766 |
|||
1767 |
34685 |
sl->first_mb_addr = get_ue_golomb_long(&sl->gb); |
|
1768 |
|||
1769 |
34685 |
slice_type = get_ue_golomb_31(&sl->gb); |
|
1770 |
✗✓ | 34685 |
if (slice_type > 9) { |
1771 |
av_log(h->avctx, AV_LOG_ERROR, |
||
1772 |
"slice type %d too large at %d\n", |
||
1773 |
slice_type, sl->first_mb_addr); |
||
1774 |
return AVERROR_INVALIDDATA; |
||
1775 |
} |
||
1776 |
✓✓ | 34685 |
if (slice_type > 4) { |
1777 |
21749 |
slice_type -= 5; |
|
1778 |
21749 |
sl->slice_type_fixed = 1; |
|
1779 |
} else |
||
1780 |
12936 |
sl->slice_type_fixed = 0; |
|
1781 |
|||
1782 |
34685 |
slice_type = ff_h264_golomb_to_pict_type[slice_type]; |
|
1783 |
34685 |
sl->slice_type = slice_type; |
|
1784 |
34685 |
sl->slice_type_nos = slice_type & 3; |
|
1785 |
|||
1786 |
✓✓ | 34685 |
if (nal->type == H264_NAL_IDR_SLICE && |
1787 |
✗✓ | 2078 |
sl->slice_type_nos != AV_PICTURE_TYPE_I) { |
1788 |
av_log(h->avctx, AV_LOG_ERROR, "A non-intra slice in an IDR NAL unit.\n"); |
||
1789 |
return AVERROR_INVALIDDATA; |
||
1790 |
} |
||
1791 |
|||
1792 |
34685 |
sl->pps_id = get_ue_golomb(&sl->gb); |
|
1793 |
✗✓ | 34685 |
if (sl->pps_id >= MAX_PPS_COUNT) { |
1794 |
av_log(h->avctx, AV_LOG_ERROR, "pps_id %u out of range\n", sl->pps_id); |
||
1795 |
return AVERROR_INVALIDDATA; |
||
1796 |
} |
||
1797 |
✓✓ | 34685 |
if (!h->ps.pps_list[sl->pps_id]) { |
1798 |
389 |
av_log(h->avctx, AV_LOG_ERROR, |
|
1799 |
"non-existing PPS %u referenced\n", |
||
1800 |
sl->pps_id); |
||
1801 |
389 |
return AVERROR_INVALIDDATA; |
|
1802 |
} |
||
1803 |
34296 |
pps = (const PPS*)h->ps.pps_list[sl->pps_id]->data; |
|
1804 |
34296 |
sps = pps->sps; |
|
1805 |
|||
1806 |
34296 |
sl->frame_num = get_bits(&sl->gb, sps->log2_max_frame_num); |
|
1807 |
✓✓ | 34296 |
if (!first_slice) { |
1808 |
✗✓ | 8298 |
if (h->poc.frame_num != sl->frame_num) { |
1809 |
av_log(h->avctx, AV_LOG_ERROR, "Frame num change from %d to %d\n", |
||
1810 |
h->poc.frame_num, sl->frame_num); |
||
1811 |
return AVERROR_INVALIDDATA; |
||
1812 |
} |
||
1813 |
} |
||
1814 |
|||
1815 |
34296 |
sl->mb_mbaff = 0; |
|
1816 |
|||
1817 |
✓✓ | 34296 |
if (sps->frame_mbs_only_flag) { |
1818 |
20355 |
picture_structure = PICT_FRAME; |
|
1819 |
} else { |
||
1820 |
✗✓✗✗ |
13941 |
if (!sps->direct_8x8_inference_flag && slice_type == AV_PICTURE_TYPE_B) { |
1821 |
av_log(h->avctx, AV_LOG_ERROR, "This stream was generated by a broken encoder, invalid 8x8 inference\n"); |
||
1822 |
return -1; |
||
1823 |
} |
||
1824 |
13941 |
field_pic_flag = get_bits1(&sl->gb); |
|
1825 |
✓✓ | 13941 |
if (field_pic_flag) { |
1826 |
8676 |
bottom_field_flag = get_bits1(&sl->gb); |
|
1827 |
8676 |
picture_structure = PICT_TOP_FIELD + bottom_field_flag; |
|
1828 |
} else { |
||
1829 |
5265 |
picture_structure = PICT_FRAME; |
|
1830 |
} |
||
1831 |
} |
||
1832 |
34296 |
sl->picture_structure = picture_structure; |
|
1833 |
34296 |
sl->mb_field_decoding_flag = picture_structure != PICT_FRAME; |
|
1834 |
|||
1835 |
✓✓ | 34296 |
if (picture_structure == PICT_FRAME) { |
1836 |
25620 |
sl->curr_pic_num = sl->frame_num; |
|
1837 |
25620 |
sl->max_pic_num = 1 << sps->log2_max_frame_num; |
|
1838 |
} else { |
||
1839 |
8676 |
sl->curr_pic_num = 2 * sl->frame_num + 1; |
|
1840 |
8676 |
sl->max_pic_num = 1 << (sps->log2_max_frame_num + 1); |
|
1841 |
} |
||
1842 |
|||
1843 |
✓✓ | 34296 |
if (nal->type == H264_NAL_IDR_SLICE) |
1844 |
2073 |
get_ue_golomb_long(&sl->gb); /* idr_pic_id */ |
|
1845 |
|||
1846 |
✓✓ | 34296 |
if (sps->poc_type == 0) { |
1847 |
27128 |
sl->poc_lsb = get_bits(&sl->gb, sps->log2_max_poc_lsb); |
|
1848 |
|||
1849 |
✓✓✓✓ |
27128 |
if (pps->pic_order_present == 1 && picture_structure == PICT_FRAME) |
1850 |
5830 |
sl->delta_poc_bottom = get_se_golomb(&sl->gb); |
|
1851 |
} |
||
1852 |
|||
1853 |
✓✓✓✓ |
34296 |
if (sps->poc_type == 1 && !sps->delta_pic_order_always_zero_flag) { |
1854 |
2848 |
sl->delta_poc[0] = get_se_golomb(&sl->gb); |
|
1855 |
|||
1856 |
✓✓✓✓ |
2848 |
if (pps->pic_order_present == 1 && picture_structure == PICT_FRAME) |
1857 |
453 |
sl->delta_poc[1] = get_se_golomb(&sl->gb); |
|
1858 |
} |
||
1859 |
|||
1860 |
34296 |
sl->redundant_pic_count = 0; |
|
1861 |
✗✓ | 34296 |
if (pps->redundant_pic_cnt_present) |
1862 |
sl->redundant_pic_count = get_ue_golomb(&sl->gb); |
||
1863 |
|||
1864 |
✓✓ | 34296 |
if (sl->slice_type_nos == AV_PICTURE_TYPE_B) |
1865 |
12076 |
sl->direct_spatial_mv_pred = get_bits1(&sl->gb); |
|
1866 |
|||
1867 |
34296 |
ret = ff_h264_parse_ref_count(&sl->list_count, sl->ref_count, |
|
1868 |
&sl->gb, pps, sl->slice_type_nos, |
||
1869 |
34296 |
picture_structure, h->avctx); |
|
1870 |
✗✓ | 34296 |
if (ret < 0) |
1871 |
return ret; |
||
1872 |
|||
1873 |
✓✓ | 34296 |
if (sl->slice_type_nos != AV_PICTURE_TYPE_I) { |
1874 |
30064 |
ret = ff_h264_decode_ref_pic_list_reordering(sl, h->avctx); |
|
1875 |
✗✓ | 30064 |
if (ret < 0) { |
1876 |
sl->ref_count[1] = sl->ref_count[0] = 0; |
||
1877 |
return ret; |
||
1878 |
} |
||
1879 |
} |
||
1880 |
|||
1881 |
34296 |
sl->pwt.use_weight = 0; |
|
1882 |
✓✓ | 102888 |
for (i = 0; i < 2; i++) { |
1883 |
68592 |
sl->pwt.luma_weight_flag[i] = 0; |
|
1884 |
68592 |
sl->pwt.chroma_weight_flag[i] = 0; |
|
1885 |
} |
||
1886 |
✓✓✓✓ |
34296 |
if ((pps->weighted_pred && sl->slice_type_nos == AV_PICTURE_TYPE_P) || |
1887 |
✓✓ | 32976 |
(pps->weighted_bipred_idc == 1 && |
1888 |
✓✗ | 278 |
sl->slice_type_nos == AV_PICTURE_TYPE_B)) { |
1889 |
1598 |
ret = ff_h264_pred_weight_table(&sl->gb, sps, sl->ref_count, |
|
1890 |
sl->slice_type_nos, &sl->pwt, |
||
1891 |
1598 |
picture_structure, h->avctx); |
|
1892 |
✗✓ | 1598 |
if (ret < 0) |
1893 |
return ret; |
||
1894 |
} |
||
1895 |
|||
1896 |
34296 |
sl->explicit_ref_marking = 0; |
|
1897 |
✓✓ | 34296 |
if (nal->ref_idc) { |
1898 |
22852 |
ret = ff_h264_decode_ref_pic_marking(sl, &sl->gb, nal, h->avctx); |
|
1899 |
✗✓✗✗ |
22852 |
if (ret < 0 && (h->avctx->err_recognition & AV_EF_EXPLODE)) |
1900 |
return AVERROR_INVALIDDATA; |
||
1901 |
} |
||
1902 |
|||
1903 |
✓✓✓✓ |
34296 |
if (sl->slice_type_nos != AV_PICTURE_TYPE_I && pps->cabac) { |
1904 |
15919 |
tmp = get_ue_golomb_31(&sl->gb); |
|
1905 |
✗✓ | 15919 |
if (tmp > 2) { |
1906 |
av_log(h->avctx, AV_LOG_ERROR, "cabac_init_idc %u overflow\n", tmp); |
||
1907 |
return AVERROR_INVALIDDATA; |
||
1908 |
} |
||
1909 |
15919 |
sl->cabac_init_idc = tmp; |
|
1910 |
} |
||
1911 |
|||
1912 |
34296 |
sl->last_qscale_diff = 0; |
|
1913 |
34296 |
tmp = pps->init_qp + (unsigned)get_se_golomb(&sl->gb); |
|
1914 |
✗✓ | 34296 |
if (tmp > 51 + 6 * (sps->bit_depth_luma - 8)) { |
1915 |
av_log(h->avctx, AV_LOG_ERROR, "QP %u out of range\n", tmp); |
||
1916 |
return AVERROR_INVALIDDATA; |
||
1917 |
} |
||
1918 |
34296 |
sl->qscale = tmp; |
|
1919 |
34296 |
sl->chroma_qp[0] = get_chroma_qp(pps, 0, sl->qscale); |
|
1920 |
34296 |
sl->chroma_qp[1] = get_chroma_qp(pps, 1, sl->qscale); |
|
1921 |
// FIXME qscale / qp ... stuff |
||
1922 |
✗✓ | 34296 |
if (sl->slice_type == AV_PICTURE_TYPE_SP) |
1923 |
get_bits1(&sl->gb); /* sp_for_switch_flag */ |
||
1924 |
✓✗ | 34296 |
if (sl->slice_type == AV_PICTURE_TYPE_SP || |
1925 |
✗✓ | 34296 |
sl->slice_type == AV_PICTURE_TYPE_SI) |
1926 |
get_se_golomb(&sl->gb); /* slice_qs_delta */ |
||
1927 |
|||
1928 |
34296 |
sl->deblocking_filter = 1; |
|
1929 |
34296 |
sl->slice_alpha_c0_offset = 0; |
|
1930 |
34296 |
sl->slice_beta_offset = 0; |
|
1931 |
✓✓ | 34296 |
if (pps->deblocking_filter_parameters_present) { |
1932 |
19840 |
tmp = get_ue_golomb_31(&sl->gb); |
|
1933 |
✗✓ | 19840 |
if (tmp > 2) { |
1934 |
av_log(h->avctx, AV_LOG_ERROR, |
||
1935 |
"deblocking_filter_idc %u out of range\n", tmp); |
||
1936 |
return AVERROR_INVALIDDATA; |
||
1937 |
} |
||
1938 |
19840 |
sl->deblocking_filter = tmp; |
|
1939 |
✓✓ | 19840 |
if (sl->deblocking_filter < 2) |
1940 |
17339 |
sl->deblocking_filter ^= 1; // 1<->0 |
|
1941 |
|||
1942 |
✓✓ | 19840 |
if (sl->deblocking_filter) { |
1943 |
12905 |
int slice_alpha_c0_offset_div2 = get_se_golomb(&sl->gb); |
|
1944 |
12905 |
int slice_beta_offset_div2 = get_se_golomb(&sl->gb); |
|
1945 |
✓✗✓✗ |
12905 |
if (slice_alpha_c0_offset_div2 > 6 || |
1946 |
✓✗ | 12905 |
slice_alpha_c0_offset_div2 < -6 || |
1947 |
✗✓ | 12905 |
slice_beta_offset_div2 > 6 || |
1948 |
slice_beta_offset_div2 < -6) { |
||
1949 |
av_log(h->avctx, AV_LOG_ERROR, |
||
1950 |
"deblocking filter parameters %d %d out of range\n", |
||
1951 |
slice_alpha_c0_offset_div2, slice_beta_offset_div2); |
||
1952 |
return AVERROR_INVALIDDATA; |
||
1953 |
} |
||
1954 |
12905 |
sl->slice_alpha_c0_offset = slice_alpha_c0_offset_div2 * 2; |
|
1955 |
12905 |
sl->slice_beta_offset = slice_beta_offset_div2 * 2; |
|
1956 |
} |
||
1957 |
} |
||
1958 |
|||
1959 |
34296 |
return 0; |
|
1960 |
} |
||
1961 |
|||
1962 |
/* do all the per-slice initialization needed before we can start decoding the |
||
1963 |
* actual MBs */ |
||
1964 |
34216 |
static int h264_slice_init(H264Context *h, H264SliceContext *sl, |
|
1965 |
const H2645NAL *nal) |
||
1966 |
{ |
||
1967 |
34216 |
int i, j, ret = 0; |
|
1968 |
|||
1969 |
✓✓✗✓ |
34216 |
if (h->picture_idr && nal->type != H264_NAL_IDR_SLICE) { |
1970 |
av_log(h->avctx, AV_LOG_ERROR, "Invalid mix of IDR and non-IDR slices\n"); |
||
1971 |
return AVERROR_INVALIDDATA; |
||
1972 |
} |
||
1973 |
|||
1974 |
av_assert1(h->mb_num == h->mb_width * h->mb_height); |
||
1975 |
✓✓✓✓ ✓✗ |
34216 |
if (sl->first_mb_addr << FIELD_OR_MBAFF_PICTURE(h) >= h->mb_num || |
1976 |
✗✓ | 34216 |
sl->first_mb_addr >= h->mb_num) { |
1977 |
av_log(h->avctx, AV_LOG_ERROR, "first_mb_in_slice overflow\n"); |
||
1978 |
return AVERROR_INVALIDDATA; |
||
1979 |
} |
||
1980 |
34216 |
sl->resync_mb_x = sl->mb_x = sl->first_mb_addr % h->mb_width; |
|
1981 |
68432 |
sl->resync_mb_y = sl->mb_y = (sl->first_mb_addr / h->mb_width) << |
|
1982 |
✓✓✓✓ |
34216 |
FIELD_OR_MBAFF_PICTURE(h); |
1983 |
✓✓ | 34216 |
if (h->picture_structure == PICT_BOTTOM_FIELD) |
1984 |
4295 |
sl->resync_mb_y = sl->mb_y = sl->mb_y + 1; |
|
1985 |
av_assert1(sl->mb_y < h->mb_height); |
||
1986 |
|||
1987 |
34216 |
ret = ff_h264_build_ref_list(h, sl); |
|
1988 |
✗✓ | 34216 |
if (ret < 0) |
1989 |
return ret; |
||
1990 |
|||
1991 |
✓✓ | 34216 |
if (h->ps.pps->weighted_bipred_idc == 2 && |
1992 |
✓✓ | 2426 |
sl->slice_type_nos == AV_PICTURE_TYPE_B) { |
1993 |
1589 |
implicit_weight_table(h, sl, -1); |
|
1994 |
✓✓ | 1589 |
if (FRAME_MBAFF(h)) { |
1995 |
855 |
implicit_weight_table(h, sl, 0); |
|
1996 |
855 |
implicit_weight_table(h, sl, 1); |
|
1997 |
} |
||
1998 |
} |
||
1999 |
|||
2000 |
✓✓✓✓ |
34216 |
if (sl->slice_type_nos == AV_PICTURE_TYPE_B && !sl->direct_spatial_mv_pred) |
2001 |
7703 |
ff_h264_direct_dist_scale_factor(h, sl); |
|
2002 |
✓✗ | 34216 |
if (!h->setup_finished) |
2003 |
34216 |
ff_h264_direct_ref_list_init(h, sl); |
|
2004 |
|||
2005 |
✓✗ | 34216 |
if (h->avctx->skip_loop_filter >= AVDISCARD_ALL || |
2006 |
✗✓ | 34216 |
(h->avctx->skip_loop_filter >= AVDISCARD_NONKEY && |
2007 |
h->nal_unit_type != H264_NAL_IDR_SLICE) || |
||
2008 |
✗✓ | 34216 |
(h->avctx->skip_loop_filter >= AVDISCARD_NONINTRA && |
2009 |
sl->slice_type_nos != AV_PICTURE_TYPE_I) || |
||
2010 |
✗✓ | 34216 |
(h->avctx->skip_loop_filter >= AVDISCARD_BIDIR && |
2011 |
sl->slice_type_nos == AV_PICTURE_TYPE_B) || |
||
2012 |
✗✓ | 34216 |
(h->avctx->skip_loop_filter >= AVDISCARD_NONREF && |
2013 |
nal->ref_idc == 0)) |
||
2014 |
sl->deblocking_filter = 0; |
||
2015 |
|||
2016 |
✓✓✗✓ |
34216 |
if (sl->deblocking_filter == 1 && h->nb_slice_ctx > 1) { |
2017 |
if (h->avctx->flags2 & AV_CODEC_FLAG2_FAST) { |
||
2018 |
/* Cheat slightly for speed: |
||
2019 |
* Do not bother to deblock across slices. */ |
||
2020 |
sl->deblocking_filter = 2; |
||
2021 |
} else { |
||
2022 |
h->postpone_filter = 1; |
||
2023 |
} |
||
2024 |
} |
||
2025 |
34216 |
sl->qp_thresh = 15 - |
|
2026 |
34216 |
FFMIN(sl->slice_alpha_c0_offset, sl->slice_beta_offset) - |
|
2027 |
34216 |
FFMAX3(0, |
|
2028 |
h->ps.pps->chroma_qp_index_offset[0], |
||
2029 |
34216 |
h->ps.pps->chroma_qp_index_offset[1]) + |
|
2030 |
34216 |
6 * (h->ps.sps->bit_depth_luma - 8); |
|
2031 |
|||
2032 |
34216 |
sl->slice_num = ++h->current_slice; |
|
2033 |
|||
2034 |
✓✗ | 34216 |
if (sl->slice_num) |
2035 |
34216 |
h->slice_row[(sl->slice_num-1)&(MAX_SLICES-1)]= sl->resync_mb_y; |
|
2036 |
✓✓ | 34216 |
if ( h->slice_row[sl->slice_num&(MAX_SLICES-1)] + 3 >= sl->resync_mb_y |
2037 |
✓✓ | 31548 |
&& h->slice_row[sl->slice_num&(MAX_SLICES-1)] <= sl->resync_mb_y |
2038 |
✗✓ | 23735 |
&& sl->slice_num >= MAX_SLICES) { |
2039 |
//in case of ASO this check needs to be updated depending on how we decide to assign slice numbers in this case |
||
2040 |
av_log(h->avctx, AV_LOG_WARNING, "Possibly too many slices (%d >= %d), increase MAX_SLICES and recompile if there are artifacts\n", sl->slice_num, MAX_SLICES); |
||
2041 |
} |
||
2042 |
|||
2043 |
✓✓ | 102648 |
for (j = 0; j < 2; j++) { |
2044 |
int id_list[16]; |
||
2045 |
68432 |
int *ref2frm = h->ref2frm[sl->slice_num & (MAX_SLICES - 1)][j]; |
|
2046 |
✓✓ | 1163344 |
for (i = 0; i < 16; i++) { |
2047 |
1094912 |
id_list[i] = 60; |
|
2048 |
✓✓✓✓ |
1094912 |
if (j < sl->list_count && i < sl->ref_count[j] && |
2049 |
✓✗ | 134229 |
sl->ref_list[j][i].parent->f->buf[0]) { |
2050 |
int k; |
||
2051 |
134229 |
AVBuffer *buf = sl->ref_list[j][i].parent->f->buf[0]->buffer; |
|
2052 |
✓✓ | 459585 |
for (k = 0; k < h->short_ref_count; k++) |
2053 |
✓✓ | 457652 |
if (h->short_ref[k]->f->buf[0]->buffer == buf) { |
2054 |
132296 |
id_list[i] = k; |
|
2055 |
132296 |
break; |
|
2056 |
} |
||
2057 |
✓✓ | 144307 |
for (k = 0; k < h->long_ref_count; k++) |
2058 |
✓✓✓✓ |
11870 |
if (h->long_ref[k] && h->long_ref[k]->f->buf[0]->buffer == buf) { |
2059 |
1792 |
id_list[i] = h->short_ref_count + k; |
|
2060 |
1792 |
break; |
|
2061 |
} |
||
2062 |
} |
||
2063 |
} |
||
2064 |
|||
2065 |
68432 |
ref2frm[0] = |
|
2066 |
68432 |
ref2frm[1] = -1; |
|
2067 |
✓✓ | 1163344 |
for (i = 0; i < 16; i++) |
2068 |
1094912 |
ref2frm[i + 2] = 4 * id_list[i] + (sl->ref_list[j][i].reference & 3); |
|
2069 |
68432 |
ref2frm[18 + 0] = |
|
2070 |
68432 |
ref2frm[18 + 1] = -1; |
|
2071 |
✓✓ | 2258256 |
for (i = 16; i < 48; i++) |
2072 |
2189824 |
ref2frm[i + 4] = 4 * id_list[(i - 16) >> 1] + |
|
2073 |
2189824 |
(sl->ref_list[j][i].reference & 3); |
|
2074 |
} |
||
2075 |
|||
2076 |
✗✓ | 34216 |
if (h->avctx->debug & FF_DEBUG_PICT_INFO) { |
2077 |
av_log(h->avctx, AV_LOG_DEBUG, |
||
2078 |
"slice:%d %s mb:%d %c%s%s frame:%d poc:%d/%d ref:%d/%d qp:%d loop:%d:%d:%d weight:%d%s %s\n", |
||
2079 |
sl->slice_num, |
||
2080 |
(h->picture_structure == PICT_FRAME ? "F" : h->picture_structure == PICT_TOP_FIELD ? "T" : "B"), |
||
2081 |
sl->mb_y * h->mb_width + sl->mb_x, |
||
2082 |
av_get_picture_type_char(sl->slice_type), |
||
2083 |
sl->slice_type_fixed ? " fix" : "", |
||
2084 |
nal->type == H264_NAL_IDR_SLICE ? " IDR" : "", |
||
2085 |
h->poc.frame_num, |
||
2086 |
h->cur_pic_ptr->field_poc[0], |
||
2087 |
h->cur_pic_ptr->field_poc[1], |
||
2088 |
sl->ref_count[0], sl->ref_count[1], |
||
2089 |
sl->qscale, |
||
2090 |
sl->deblocking_filter, |
||
2091 |
sl->slice_alpha_c0_offset, sl->slice_beta_offset, |
||
2092 |
sl->pwt.use_weight, |
||
2093 |
sl->pwt.use_weight == 1 && sl->pwt.use_weight_chroma ? "c" : "", |
||
2094 |
sl->slice_type == AV_PICTURE_TYPE_B ? (sl->direct_spatial_mv_pred ? "SPAT" : "TEMP") : ""); |
||
2095 |
} |
||
2096 |
|||
2097 |
34216 |
return 0; |
|
2098 |
} |
||
2099 |
|||
2100 |
34685 |
int ff_h264_queue_decode_slice(H264Context *h, const H2645NAL *nal) |
|
2101 |
{ |
||
2102 |
34685 |
H264SliceContext *sl = h->slice_ctx + h->nb_slice_ctx_queued; |
|
2103 |
✓✓✓✓ |
34685 |
int first_slice = sl == h->slice_ctx && !h->current_slice; |
2104 |
int ret; |
||
2105 |
|||
2106 |
34685 |
sl->gb = nal->gb; |
|
2107 |
|||
2108 |
34685 |
ret = h264_slice_header_parse(h, sl, nal); |
|
2109 |
✓✓ | 34685 |
if (ret < 0) |
2110 |
389 |
return ret; |
|
2111 |
|||
2112 |
// discard redundant pictures |
||
2113 |
✗✓ | 34296 |
if (sl->redundant_pic_count > 0) { |
2114 |
sl->ref_count[0] = sl->ref_count[1] = 0; |
||
2115 |
return 0; |
||
2116 |
} |
||
2117 |
|||
2118 |
✓✓✗✓ |
34296 |
if (sl->first_mb_addr == 0 || !h->current_slice) { |
2119 |
✗✓ | 26191 |
if (h->setup_finished) { |
2120 |
av_log(h->avctx, AV_LOG_ERROR, "Too many fields\n"); |
||
2121 |
return AVERROR_INVALIDDATA; |
||
2122 |
} |
||
2123 |
} |
||
2124 |
|||
2125 |
✓✓ | 34296 |
if (sl->first_mb_addr == 0) { // FIXME better field boundary detection |
2126 |
✓✓ | 26191 |
if (h->current_slice) { |
2127 |
// this slice starts a new field |
||
2128 |
// first decode any pending queued slices |
||
2129 |
✗✓ | 193 |
if (h->nb_slice_ctx_queued) { |
2130 |
H264SliceContext tmp_ctx; |
||
2131 |
|||
2132 |
ret = ff_h264_execute_decode_slices(h); |
||
2133 |
if (ret < 0 && (h->avctx->err_recognition & AV_EF_EXPLODE)) |
||
2134 |
return ret; |
||
2135 |
|||
2136 |
memcpy(&tmp_ctx, h->slice_ctx, sizeof(tmp_ctx)); |
||
2137 |
memcpy(h->slice_ctx, sl, sizeof(tmp_ctx)); |
||
2138 |
memcpy(sl, &tmp_ctx, sizeof(tmp_ctx)); |
||
2139 |
sl = h->slice_ctx; |
||
2140 |
} |
||
2141 |
|||
2142 |
✓✗✓✗ ✓✗ |
193 |
if (h->cur_pic_ptr && FIELD_PICTURE(h) && h->first_field) { |
2143 |
193 |
ret = ff_h264_field_end(h, h->slice_ctx, 1); |
|
2144 |
✗✓ | 193 |
if (ret < 0) |
2145 |
return ret; |
||
2146 |
} else if (h->cur_pic_ptr && !FIELD_PICTURE(h) && !h->first_field && h->nal_unit_type == H264_NAL_IDR_SLICE) { |
||
2147 |
av_log(h, AV_LOG_WARNING, "Broken frame packetizing\n"); |
||
2148 |
ret = ff_h264_field_end(h, h->slice_ctx, 1); |
||
2149 |
ff_thread_report_progress(&h->cur_pic_ptr->tf, INT_MAX, 0); |
||
2150 |
ff_thread_report_progress(&h->cur_pic_ptr->tf, INT_MAX, 1); |
||
2151 |
h->cur_pic_ptr = NULL; |
||
2152 |
if (ret < 0) |
||
2153 |
return ret; |
||
2154 |
} else |
||
2155 |
return AVERROR_INVALIDDATA; |
||
2156 |
} |
||
2157 |
|||
2158 |
✓✓ | 26191 |
if (!h->first_field) { |
2159 |
✓✓✓✗ |
22859 |
if (h->cur_pic_ptr && !h->droppable) { |
2160 |
299 |
ff_thread_report_progress(&h->cur_pic_ptr->tf, INT_MAX, |
|
2161 |
299 |
h->picture_structure == PICT_BOTTOM_FIELD); |
|
2162 |
} |
||
2163 |
22859 |
h->cur_pic_ptr = NULL; |
|
2164 |
} |
||
2165 |
} |
||
2166 |
|||
2167 |
✓✓ | 34296 |
if (!h->current_slice) |
2168 |
✗✓ | 26191 |
av_assert0(sl == h->slice_ctx); |
2169 |
|||
2170 |
✓✓✓✓ |
34296 |
if (h->current_slice == 0 && !h->first_field) { |
2171 |
22859 |
if ( |
|
2172 |
✓✓✓✗ |
22859 |
(h->avctx->skip_frame >= AVDISCARD_NONREF && !h->nal_ref_idc) || |
2173 |
✓✓✓✓ |
22859 |
(h->avctx->skip_frame >= AVDISCARD_BIDIR && sl->slice_type_nos == AV_PICTURE_TYPE_B) || |
2174 |
✓✓✓✓ |
22815 |
(h->avctx->skip_frame >= AVDISCARD_NONINTRA && sl->slice_type_nos != AV_PICTURE_TYPE_I) || |
2175 |
✓✓✓✗ ✓✗ |
22779 |
(h->avctx->skip_frame >= AVDISCARD_NONKEY && h->nal_unit_type != H264_NAL_IDR_SLICE && h->sei.recovery_point.recovery_frame_cnt < 0) || |
2176 |
✗✓ | 22779 |
h->avctx->skip_frame >= AVDISCARD_ALL) { |
2177 |
80 |
return 0; |
|
2178 |
} |
||
2179 |
} |
||
2180 |
|||
2181 |
✓✓ | 34216 |
if (!first_slice) { |
2182 |
8298 |
const PPS *pps = (const PPS*)h->ps.pps_list[sl->pps_id]->data; |
|
2183 |
|||
2184 |
✓✗ | 8298 |
if (h->ps.pps->sps_id != pps->sps_id || |
2185 |
✗✓ | 8298 |
h->ps.pps->transform_8x8_mode != pps->transform_8x8_mode /*|| |
2186 |
(h->setup_finished && h->ps.pps != pps)*/) { |
||
2187 |
av_log(h->avctx, AV_LOG_ERROR, "PPS changed between slices\n"); |
||
2188 |
return AVERROR_INVALIDDATA; |
||
2189 |
} |
||
2190 |
✗✓ | 8298 |
if (h->ps.sps != pps->sps) { |
2191 |
av_log(h->avctx, AV_LOG_ERROR, |
||
2192 |
"SPS changed in the middle of the frame\n"); |
||
2193 |
return AVERROR_INVALIDDATA; |
||
2194 |
} |
||
2195 |
} |
||
2196 |
|||
2197 |
✓✓ | 34216 |
if (h->current_slice == 0) { |
2198 |
26111 |
ret = h264_field_start(h, sl, nal, first_slice); |
|
2199 |
✗✓ | 26111 |
if (ret < 0) |
2200 |
return ret; |
||
2201 |
} else { |
||
2202 |
✓✗ | 8105 |
if (h->picture_structure != sl->picture_structure || |
2203 |
✗✓ | 8105 |
h->droppable != (nal->ref_idc == 0)) { |
2204 |
av_log(h->avctx, AV_LOG_ERROR, |
||
2205 |
"Changing field mode (%d -> %d) between slices is not allowed\n", |
||
2206 |
h->picture_structure, sl->picture_structure); |
||
2207 |
return AVERROR_INVALIDDATA; |
||
2208 |
✗✓ | 8105 |
} else if (!h->cur_pic_ptr) { |
2209 |
av_log(h->avctx, AV_LOG_ERROR, |
||
2210 |
"unset cur_pic_ptr on slice %d\n", |
||
2211 |
h->current_slice + 1); |
||
2212 |
return AVERROR_INVALIDDATA; |
||
2213 |
} |
||
2214 |
} |
||
2215 |
|||
2216 |
34216 |
ret = h264_slice_init(h, sl, nal); |
|
2217 |
✗✓ | 34216 |
if (ret < 0) |
2218 |
return ret; |
||
2219 |
|||
2220 |
34216 |
h->nb_slice_ctx_queued++; |
|
2221 |
|||
2222 |
34216 |
return 0; |
|
2223 |
} |
||
2224 |
|||
2225 |
int ff_h264_get_slice_type(const H264SliceContext *sl) |
||
2226 |
{ |
||
2227 |
switch (sl->slice_type) { |
||
2228 |
case AV_PICTURE_TYPE_P: |
||
2229 |
return 0; |
||
2230 |
case AV_PICTURE_TYPE_B: |
||
2231 |
return 1; |
||
2232 |
case AV_PICTURE_TYPE_I: |
||
2233 |
return 2; |
||
2234 |
case AV_PICTURE_TYPE_SP: |
||
2235 |
return 3; |
||
2236 |
case AV_PICTURE_TYPE_SI: |
||
2237 |
return 4; |
||
2238 |
default: |
||
2239 |
return AVERROR_INVALIDDATA; |
||
2240 |
} |
||
2241 |
} |
||
2242 |
|||
2243 |
11509805 |
static av_always_inline void fill_filter_caches_inter(const H264Context *h, |
|
2244 |
H264SliceContext *sl, |
||
2245 |
int mb_type, int top_xy, |
||
2246 |
int left_xy[LEFT_MBS], |
||
2247 |
int top_type, |
||
2248 |
int left_type[LEFT_MBS], |
||
2249 |
int mb_xy, int list) |
||
2250 |
{ |
||
2251 |
11509805 |
int b_stride = h->b_stride; |
|
2252 |
11509805 |
int16_t(*mv_dst)[2] = &sl->mv_cache[list][scan8[0]]; |
|
2253 |
11509805 |
int8_t *ref_cache = &sl->ref_cache[list][scan8[0]]; |
|
2254 |
✗✓✗✗ |
11509805 |
if (IS_INTER(mb_type) || IS_DIRECT(mb_type)) { |
2255 |
✓✓ | 11509805 |
if (USES_LIST(top_type, list)) { |
2256 |
9256034 |
const int b_xy = h->mb2b_xy[top_xy] + 3 * b_stride; |
|
2257 |
9256034 |
const int b8_xy = 4 * top_xy + 2; |
|
2258 |
✓✓ | 9256034 |
const int *ref2frm = &h->ref2frm[h->slice_table[top_xy] & (MAX_SLICES - 1)][list][(MB_MBAFF(sl) ? 20 : 2)]; |
2259 |
9256034 |
AV_COPY128(mv_dst - 1 * 8, h->cur_pic.motion_val[list][b_xy + 0]); |
|
2260 |
9256034 |
ref_cache[0 - 1 * 8] = |
|
2261 |
9256034 |
ref_cache[1 - 1 * 8] = ref2frm[h->cur_pic.ref_index[list][b8_xy + 0]]; |
|
2262 |
9256034 |
ref_cache[2 - 1 * 8] = |
|
2263 |
9256034 |
ref_cache[3 - 1 * 8] = ref2frm[h->cur_pic.ref_index[list][b8_xy + 1]]; |
|
2264 |
} else { |
||
2265 |
2253771 |
AV_ZERO128(mv_dst - 1 * 8); |
|
2266 |
2253771 |
AV_WN32A(&ref_cache[0 - 1 * 8], ((LIST_NOT_USED) & 0xFF) * 0x01010101u); |
|
2267 |
} |
||
2268 |
|||
2269 |
✓✓ | 11509805 |
if (!IS_INTERLACED(mb_type ^ left_type[LTOP])) { |
2270 |
✓✓ | 11099142 |
if (USES_LIST(left_type[LTOP], list)) { |
2271 |
9284607 |
const int b_xy = h->mb2b_xy[left_xy[LTOP]] + 3; |
|
2272 |
9284607 |
const int b8_xy = 4 * left_xy[LTOP] + 1; |
|
2273 |
✓✓ | 9284607 |
const int *ref2frm = &h->ref2frm[h->slice_table[left_xy[LTOP]] & (MAX_SLICES - 1)][list][(MB_MBAFF(sl) ? 20 : 2)]; |
2274 |
9284607 |
AV_COPY32(mv_dst - 1 + 0, h->cur_pic.motion_val[list][b_xy + b_stride * 0]); |
|
2275 |
9284607 |
AV_COPY32(mv_dst - 1 + 8, h->cur_pic.motion_val[list][b_xy + b_stride * 1]); |
|
2276 |
9284607 |
AV_COPY32(mv_dst - 1 + 16, h->cur_pic.motion_val[list][b_xy + b_stride * 2]); |
|
2277 |
9284607 |
AV_COPY32(mv_dst - 1 + 24, h->cur_pic.motion_val[list][b_xy + b_stride * 3]); |
|
2278 |
9284607 |
ref_cache[-1 + 0] = |
|
2279 |
9284607 |
ref_cache[-1 + 8] = ref2frm[h->cur_pic.ref_index[list][b8_xy + 2 * 0]]; |
|
2280 |
9284607 |
ref_cache[-1 + 16] = |
|
2281 |
9284607 |
ref_cache[-1 + 24] = ref2frm[h->cur_pic.ref_index[list][b8_xy + 2 * 1]]; |
|
2282 |
} else { |
||
2283 |
1814535 |
AV_ZERO32(mv_dst - 1 + 0); |
|
2284 |
1814535 |
AV_ZERO32(mv_dst - 1 + 8); |
|
2285 |
1814535 |
AV_ZERO32(mv_dst - 1 + 16); |
|
2286 |
1814535 |
AV_ZERO32(mv_dst - 1 + 24); |
|
2287 |
1814535 |
ref_cache[-1 + 0] = |
|
2288 |
1814535 |
ref_cache[-1 + 8] = |
|
2289 |
1814535 |
ref_cache[-1 + 16] = |
|
2290 |
1814535 |
ref_cache[-1 + 24] = LIST_NOT_USED; |
|
2291 |
} |
||
2292 |
} |
||
2293 |
} |
||
2294 |
|||
2295 |
✓✓ | 11509805 |
if (!USES_LIST(mb_type, list)) { |
2296 |
1262990 |
fill_rectangle(mv_dst, 4, 4, 8, pack16to32(0, 0), 4); |
|
2297 |
1262990 |
AV_WN32A(&ref_cache[0 * 8], ((LIST_NOT_USED) & 0xFF) * 0x01010101u); |
|
2298 |
1262990 |
AV_WN32A(&ref_cache[1 * 8], ((LIST_NOT_USED) & 0xFF) * 0x01010101u); |
|
2299 |
1262990 |
AV_WN32A(&ref_cache[2 * 8], ((LIST_NOT_USED) & 0xFF) * 0x01010101u); |
|
2300 |
1262990 |
AV_WN32A(&ref_cache[3 * 8], ((LIST_NOT_USED) & 0xFF) * 0x01010101u); |
|
2301 |
1262990 |
return; |
|
2302 |
} |
||
2303 |
|||
2304 |
{ |
||
2305 |
10246815 |
int8_t *ref = &h->cur_pic.ref_index[list][4 * mb_xy]; |
|
2306 |
✓✓ | 10246815 |
const int *ref2frm = &h->ref2frm[sl->slice_num & (MAX_SLICES - 1)][list][(MB_MBAFF(sl) ? 20 : 2)]; |
2307 |
10246815 |
uint32_t ref01 = (pack16to32(ref2frm[ref[0]], ref2frm[ref[1]]) & 0x00FF00FF) * 0x0101; |
|
2308 |
10246815 |
uint32_t ref23 = (pack16to32(ref2frm[ref[2]], ref2frm[ref[3]]) & 0x00FF00FF) * 0x0101; |
|
2309 |
10246815 |
AV_WN32A(&ref_cache[0 * 8], ref01); |
|
2310 |
10246815 |
AV_WN32A(&ref_cache[1 * 8], ref01); |
|
2311 |
10246815 |
AV_WN32A(&ref_cache[2 * 8], ref23); |
|
2312 |
10246815 |
AV_WN32A(&ref_cache[3 * 8], ref23); |
|
2313 |
} |
||
2314 |
|||
2315 |
{ |
||
2316 |
10246815 |
int16_t(*mv_src)[2] = &h->cur_pic.motion_val[list][4 * sl->mb_x + 4 * sl->mb_y * b_stride]; |
|
2317 |
10246815 |
AV_COPY128(mv_dst + 8 * 0, mv_src + 0 * b_stride); |
|
2318 |
10246815 |
AV_COPY128(mv_dst + 8 * 1, mv_src + 1 * b_stride); |
|
2319 |
10246815 |
AV_COPY128(mv_dst + 8 * 2, mv_src + 2 * b_stride); |
|
2320 |
10246815 |
AV_COPY128(mv_dst + 8 * 3, mv_src + 3 * b_stride); |
|
2321 |
} |
||
2322 |
} |
||
2323 |
|||
2324 |
/** |
||
2325 |
* @return non zero if the loop filter can be skipped |
||
2326 |
*/ |
||
2327 |
11609103 |
static int fill_filter_caches(const H264Context *h, H264SliceContext *sl, int mb_type) |
|
2328 |
{ |
||
2329 |
11609103 |
const int mb_xy = sl->mb_xy; |
|
2330 |
int top_xy, left_xy[LEFT_MBS]; |
||
2331 |
int top_type, left_type[LEFT_MBS]; |
||
2332 |
uint8_t *nnz; |
||
2333 |
uint8_t *nnz_cache; |
||
2334 |
|||
2335 |
11609103 |
top_xy = mb_xy - (h->mb_stride << MB_FIELD(sl)); |
|
2336 |
|||
2337 |
11609103 |
left_xy[LBOT] = left_xy[LTOP] = mb_xy - 1; |
|
2338 |
✓✓ | 11609103 |
if (FRAME_MBAFF(h)) { |
2339 |
1814136 |
const int left_mb_field_flag = IS_INTERLACED(h->cur_pic.mb_type[mb_xy - 1]); |
|
2340 |
1814136 |
const int curr_mb_field_flag = IS_INTERLACED(mb_type); |
|
2341 |
✓✓ | 1814136 |
if (sl->mb_y & 1) { |
2342 |
✓✓ | 907068 |
if (left_mb_field_flag != curr_mb_field_flag) |
2343 |
189266 |
left_xy[LTOP] -= h->mb_stride; |
|
2344 |
} else { |
||
2345 |
✓✓ | 907068 |
if (curr_mb_field_flag) |
2346 |
241257 |
top_xy += h->mb_stride & |
|
2347 |
241257 |
(((h->cur_pic.mb_type[top_xy] >> 7) & 1) - 1); |
|
2348 |
✓✓ | 907068 |
if (left_mb_field_flag != curr_mb_field_flag) |
2349 |
189266 |
left_xy[LBOT] += h->mb_stride; |
|
2350 |
} |
||
2351 |
} |
||
2352 |
|||
2353 |
11609103 |
sl->top_mb_xy = top_xy; |
|
2354 |
11609103 |
sl->left_mb_xy[LTOP] = left_xy[LTOP]; |
|
2355 |
11609103 |
sl->left_mb_xy[LBOT] = left_xy[LBOT]; |
|
2356 |
{ |
||
2357 |
/* For sufficiently low qp, filtering wouldn't do anything. |
||
2358 |
* This is a conservative estimate: could also check beta_offset |
||
2359 |
* and more accurate chroma_qp. */ |
||
2360 |
11609103 |
int qp_thresh = sl->qp_thresh; // FIXME strictly we should store qp_thresh for each mb of a slice |
|
2361 |
11609103 |
int qp = h->cur_pic.qscale_table[mb_xy]; |
|
2362 |
✓✓ | 11609103 |
if (qp <= qp_thresh && |
2363 |
✓✓ | 1458628 |
(left_xy[LTOP] < 0 || |
2364 |
✓✓✓✓ |
1458628 |
((qp + h->cur_pic.qscale_table[left_xy[LTOP]] + 1) >> 1) <= qp_thresh) && |
2365 |
1328998 |
(top_xy < 0 || |
|
2366 |
✓✓ | 1328998 |
((qp + h->cur_pic.qscale_table[top_xy] + 1) >> 1) <= qp_thresh)) { |
2367 |
✓✓ | 1213195 |
if (!FRAME_MBAFF(h)) |
2368 |
990750 |
return 1; |
|
2369 |
✓✓ | 222445 |
if ((left_xy[LTOP] < 0 || |
2370 |
✓✓ | 222351 |
((qp + h->cur_pic.qscale_table[left_xy[LBOT]] + 1) >> 1) <= qp_thresh) && |
2371 |
✓✓ | 222040 |
(top_xy < h->mb_stride || |
2372 |
✓✓ | 207469 |
((qp + h->cur_pic.qscale_table[top_xy - h->mb_stride] + 1) >> 1) <= qp_thresh)) |
2373 |
205646 |
return 1; |
|
2374 |
} |
||
2375 |
} |
||
2376 |
|||
2377 |
10412707 |
top_type = h->cur_pic.mb_type[top_xy]; |
|
2378 |
10412707 |
left_type[LTOP] = h->cur_pic.mb_type[left_xy[LTOP]]; |
|
2379 |
10412707 |
left_type[LBOT] = h->cur_pic.mb_type[left_xy[LBOT]]; |
|
2380 |
✓✓ | 10412707 |
if (sl->deblocking_filter == 2) { |
2381 |
✓✓ | 470376 |
if (h->slice_table[top_xy] != sl->slice_num) |
2382 |
56878 |
top_type = 0; |
|
2383 |
✓✓ | 470376 |
if (h->slice_table[left_xy[LBOT]] != sl->slice_num) |
2384 |
13874 |
left_type[LTOP] = left_type[LBOT] = 0; |
|
2385 |
} else { |
||
2386 |
✓✓ | 9942331 |
if (h->slice_table[top_xy] == 0xFFFF) |
2387 |
454528 |
top_type = 0; |
|
2388 |
✓✓ | 9942331 |
if (h->slice_table[left_xy[LBOT]] == 0xFFFF) |
2389 |
295880 |
left_type[LTOP] = left_type[LBOT] = 0; |
|
2390 |
} |
||
2391 |
10412707 |
sl->top_type = top_type; |
|
2392 |
10412707 |
sl->left_type[LTOP] = left_type[LTOP]; |
|
2393 |
10412707 |
sl->left_type[LBOT] = left_type[LBOT]; |
|
2394 |
|||
2395 |
✓✓ | 10412707 |
if (IS_INTRA(mb_type)) |
2396 |
2890916 |
return 0; |
|
2397 |
|||
2398 |
7521791 |
fill_filter_caches_inter(h, sl, mb_type, top_xy, left_xy, |
|
2399 |
top_type, left_type, mb_xy, 0); |
||
2400 |
✓✓ | 7521791 |
if (sl->list_count == 2) |
2401 |
3988014 |
fill_filter_caches_inter(h, sl, mb_type, top_xy, left_xy, |
|
2402 |
top_type, left_type, mb_xy, 1); |
||
2403 |
|||
2404 |
7521791 |
nnz = h->non_zero_count[mb_xy]; |
|
2405 |
7521791 |
nnz_cache = sl->non_zero_count_cache; |
|
2406 |
7521791 |
AV_COPY32(&nnz_cache[4 + 8 * 1], &nnz[0]); |
|
2407 |
7521791 |
AV_COPY32(&nnz_cache[4 + 8 * 2], &nnz[4]); |
|
2408 |
7521791 |
AV_COPY32(&nnz_cache[4 + 8 * 3], &nnz[8]); |
|
2409 |
7521791 |
AV_COPY32(&nnz_cache[4 + 8 * 4], &nnz[12]); |
|
2410 |
7521791 |
sl->cbp = h->cbp_table[mb_xy]; |
|
2411 |
|||
2412 |
✓✓ | 7521791 |
if (top_type) { |
2413 |
7112311 |
nnz = h->non_zero_count[top_xy]; |
|
2414 |
7112311 |
AV_COPY32(&nnz_cache[4 + 8 * 0], &nnz[3 * 4]); |
|
2415 |
} |
||
2416 |
|||
2417 |
✓✓ | 7521791 |
if (left_type[LTOP]) { |
2418 |
7267756 |
nnz = h->non_zero_count[left_xy[LTOP]]; |
|
2419 |
7267756 |
nnz_cache[3 + 8 * 1] = nnz[3 + 0 * 4]; |
|
2420 |
7267756 |
nnz_cache[3 + 8 * 2] = nnz[3 + 1 * 4]; |
|
2421 |
7267756 |
nnz_cache[3 + 8 * 3] = nnz[3 + 2 * 4]; |
|
2422 |
7267756 |
nnz_cache[3 + 8 * 4] = nnz[3 + 3 * 4]; |
|
2423 |
} |
||
2424 |
|||
2425 |
/* CAVLC 8x8dct requires NNZ values for residual decoding that differ |
||
2426 |
* from what the loop filter needs */ |
||
2427 |
✓✓✓✓ |
7521791 |
if (!CABAC(h) && h->ps.pps->transform_8x8_mode) { |
2428 |
✓✓ | 246512 |
if (IS_8x8DCT(top_type)) { |
2429 |
107880 |
nnz_cache[4 + 8 * 0] = |
|
2430 |
107880 |
nnz_cache[5 + 8 * 0] = (h->cbp_table[top_xy] & 0x4000) >> 12; |
|
2431 |
107880 |
nnz_cache[6 + 8 * 0] = |
|
2432 |
107880 |
nnz_cache[7 + 8 * 0] = (h->cbp_table[top_xy] & 0x8000) >> 12; |
|
2433 |
} |
||
2434 |
✓✓ | 246512 |
if (IS_8x8DCT(left_type[LTOP])) { |
2435 |
110456 |
nnz_cache[3 + 8 * 1] = |
|
2436 |
110456 |
nnz_cache[3 + 8 * 2] = (h->cbp_table[left_xy[LTOP]] & 0x2000) >> 12; // FIXME check MBAFF |
|
2437 |
} |
||
2438 |
✓✓ | 246512 |
if (IS_8x8DCT(left_type[LBOT])) { |
2439 |
110556 |
nnz_cache[3 + 8 * 3] = |
|
2440 |
110556 |
nnz_cache[3 + 8 * 4] = (h->cbp_table[left_xy[LBOT]] & 0x8000) >> 12; // FIXME check MBAFF |
|
2441 |
} |
||
2442 |
|||
2443 |
✓✓ | 246512 |
if (IS_8x8DCT(mb_type)) { |
2444 |
104923 |
nnz_cache[scan8[0]] = |
|
2445 |
104923 |
nnz_cache[scan8[1]] = |
|
2446 |
104923 |
nnz_cache[scan8[2]] = |
|
2447 |
104923 |
nnz_cache[scan8[3]] = (sl->cbp & 0x1000) >> 12; |
|
2448 |
|||
2449 |
104923 |
nnz_cache[scan8[0 + 4]] = |
|
2450 |
104923 |
nnz_cache[scan8[1 + 4]] = |
|
2451 |
104923 |
nnz_cache[scan8[2 + 4]] = |
|
2452 |
104923 |
nnz_cache[scan8[3 + 4]] = (sl->cbp & 0x2000) >> 12; |
|
2453 |
|||
2454 |
104923 |
nnz_cache[scan8[0 + 8]] = |
|
2455 |
104923 |
nnz_cache[scan8[1 + 8]] = |
|
2456 |
104923 |
nnz_cache[scan8[2 + 8]] = |
|
2457 |
104923 |
nnz_cache[scan8[3 + 8]] = (sl->cbp & 0x4000) >> 12; |
|
2458 |
|||
2459 |
104923 |
nnz_cache[scan8[0 + 12]] = |
|
2460 |
104923 |
nnz_cache[scan8[1 + 12]] = |
|
2461 |
104923 |
nnz_cache[scan8[2 + 12]] = |
|
2462 |
104923 |
nnz_cache[scan8[3 + 12]] = (sl->cbp & 0x8000) >> 12; |
|
2463 |
} |
||
2464 |
} |
||
2465 |
|||
2466 |
7521791 |
return 0; |
|
2467 |
} |
||
2468 |
|||
2469 |
365832 |
static void loop_filter(const H264Context *h, H264SliceContext *sl, int start_x, int end_x) |
|
2470 |
{ |
||
2471 |
uint8_t *dest_y, *dest_cb, *dest_cr; |
||
2472 |
int linesize, uvlinesize, mb_x, mb_y; |
||
2473 |
365832 |
const int end_mb_y = sl->mb_y + FRAME_MBAFF(h); |
|
2474 |
365832 |
const int old_slice_type = sl->slice_type; |
|
2475 |
365832 |
const int pixel_shift = h->pixel_shift; |
|
2476 |
365832 |
const int block_h = 16 >> h->chroma_y_shift; |
|
2477 |
|||
2478 |
✗✓ | 365832 |
if (h->postpone_filter) |
2479 |
return; |
||
2480 |
|||
2481 |
✓✓ | 365832 |
if (sl->deblocking_filter) { |
2482 |
✓✓ | 10998705 |
for (mb_x = start_x; mb_x < end_x; mb_x++) |
2483 |
✓✓ | 22311138 |
for (mb_y = end_mb_y - FRAME_MBAFF(h); mb_y <= end_mb_y; mb_y++) { |
2484 |
int mb_xy, mb_type; |
||
2485 |
11609103 |
mb_xy = sl->mb_xy = mb_x + mb_y * h->mb_stride; |
|
2486 |
11609103 |
mb_type = h->cur_pic.mb_type[mb_xy]; |
|
2487 |
|||
2488 |
✓✓ | 11609103 |
if (FRAME_MBAFF(h)) |
2489 |
1814136 |
sl->mb_mbaff = |
|
2490 |
1814136 |
sl->mb_field_decoding_flag = !!IS_INTERLACED(mb_type); |
|
2491 |
|||
2492 |
11609103 |
sl->mb_x = mb_x; |
|
2493 |
11609103 |
sl->mb_y = mb_y; |
|
2494 |
11609103 |
dest_y = h->cur_pic.f->data[0] + |
|
2495 |
11609103 |
((mb_x << pixel_shift) + mb_y * sl->linesize) * 16; |
|
2496 |
23218206 |
dest_cb = h->cur_pic.f->data[1] + |
|
2497 |
✓✓ | 11609103 |
(mb_x << pixel_shift) * (8 << CHROMA444(h)) + |
2498 |
11609103 |
mb_y * sl->uvlinesize * block_h; |
|
2499 |
23218206 |
dest_cr = h->cur_pic.f->data[2] + |
|
2500 |
✓✓ | 11609103 |
(mb_x << pixel_shift) * (8 << CHROMA444(h)) + |
2501 |
11609103 |
mb_y * sl->uvlinesize * block_h; |
|
2502 |
// FIXME simplify above |
||
2503 |
|||
2504 |
✓✓ | 11609103 |
if (MB_FIELD(sl)) { |
2505 |
4418472 |
linesize = sl->mb_linesize = sl->linesize * 2; |
|
2506 |
4418472 |
uvlinesize = sl->mb_uvlinesize = sl->uvlinesize * 2; |
|
2507 |
✓✓ | 4418472 |
if (mb_y & 1) { // FIXME move out of this function? |
2508 |
2206497 |
dest_y -= sl->linesize * 15; |
|
2509 |
2206497 |
dest_cb -= sl->uvlinesize * (block_h - 1); |
|
2510 |
2206497 |
dest_cr -= sl->uvlinesize * (block_h - 1); |
|
2511 |
} |
||
2512 |
} else { |
||
2513 |
7190631 |
linesize = sl->mb_linesize = sl->linesize; |
|
2514 |
7190631 |
uvlinesize = sl->mb_uvlinesize = sl->uvlinesize; |
|
2515 |
} |
||
2516 |
11609103 |
backup_mb_border(h, sl, dest_y, dest_cb, dest_cr, linesize, |
|
2517 |
uvlinesize, 0); |
||
2518 |
✓✓ | 11609103 |
if (fill_filter_caches(h, sl, mb_type)) |
2519 |
1196396 |
continue; |
|
2520 |
10412707 |
sl->chroma_qp[0] = get_chroma_qp(h->ps.pps, 0, h->cur_pic.qscale_table[mb_xy]); |
|
2521 |
10412707 |
sl->chroma_qp[1] = get_chroma_qp(h->ps.pps, 1, h->cur_pic.qscale_table[mb_xy]); |
|
2522 |
|||
2523 |
✓✓ | 10412707 |
if (FRAME_MBAFF(h)) { |
2524 |
1608490 |
ff_h264_filter_mb(h, sl, mb_x, mb_y, dest_y, dest_cb, dest_cr, |
|
2525 |
linesize, uvlinesize); |
||
2526 |
} else { |
||
2527 |
8804217 |
ff_h264_filter_mb_fast(h, sl, mb_x, mb_y, dest_y, dest_cb, |
|
2528 |
dest_cr, linesize, uvlinesize); |
||
2529 |
} |
||
2530 |
} |
||
2531 |
} |
||
2532 |
365832 |
sl->slice_type = old_slice_type; |
|
2533 |
365832 |
sl->mb_x = end_x; |
|
2534 |
365832 |
sl->mb_y = end_mb_y - FRAME_MBAFF(h); |
|
2535 |
365832 |
sl->chroma_qp[0] = get_chroma_qp(h->ps.pps, 0, sl->qscale); |
|
2536 |
365832 |
sl->chroma_qp[1] = get_chroma_qp(h->ps.pps, 1, sl->qscale); |
|
2537 |
} |
||
2538 |
|||
2539 |
22884 |
static void predict_field_decoding_flag(const H264Context *h, H264SliceContext *sl) |
|
2540 |
{ |
||
2541 |
22884 |
const int mb_xy = sl->mb_x + sl->mb_y * h->mb_stride; |
|
2542 |
45768 |
int mb_type = (h->slice_table[mb_xy - 1] == sl->slice_num) ? |
|
2543 |
✗✓ | 45768 |
h->cur_pic.mb_type[mb_xy - 1] : |
2544 |
✓✓ | 22884 |
(h->slice_table[mb_xy - h->mb_stride] == sl->slice_num) ? |
2545 |
22748 |
h->cur_pic.mb_type[mb_xy - h->mb_stride] : 0; |
|
2546 |
22884 |
sl->mb_mbaff = sl->mb_field_decoding_flag = IS_INTERLACED(mb_type) ? 1 : 0; |
|
2547 |
22884 |
} |
|
2548 |
|||
2549 |
/** |
||
2550 |
* Draw edges and report progress for the last MB row. |
||
2551 |
*/ |
||
2552 |
361737 |
static void decode_finish_row(const H264Context *h, H264SliceContext *sl) |
|
2553 |
{ |
||
2554 |
361737 |
int top = 16 * (sl->mb_y >> FIELD_PICTURE(h)); |
|
2555 |
361737 |
int pic_height = 16 * h->mb_height >> FIELD_PICTURE(h); |
|
2556 |
361737 |
int height = 16 << FRAME_MBAFF(h); |
|
2557 |
361737 |
int deblock_border = (16 + 4) << FRAME_MBAFF(h); |
|
2558 |
|||
2559 |
✓✓ | 361737 |
if (sl->deblocking_filter) { |
2560 |
✓✓ | 293587 |
if ((top + height) >= pic_height) |
2561 |
20530 |
height += deblock_border; |
|
2562 |
293587 |
top -= deblock_border; |
|
2563 |
} |
||
2564 |
|||
2565 |
✓✗✓✓ |
361737 |
if (top >= pic_height || (top + height) < 0) |
2566 |
20531 |
return; |
|
2567 |
|||
2568 |
341206 |
height = FFMIN(height, pic_height - top); |
|
2569 |
✓✓ | 341206 |
if (top < 0) { |
2570 |
20531 |
height = top + height; |
|
2571 |
20531 |
top = 0; |
|
2572 |
} |
||
2573 |
|||
2574 |
341206 |
ff_h264_draw_horiz_band(h, sl, top, height); |
|
2575 |
|||
2576 |
✓✓✗✓ |
341206 |
if (h->droppable || sl->h264->slice_ctx[0].er.error_occurred) |
2577 |
122771 |
return; |
|
2578 |
|||
2579 |
218435 |
ff_thread_report_progress(&h->cur_pic_ptr->tf, top + height - 1, |
|
2580 |
218435 |
h->picture_structure == PICT_BOTTOM_FIELD); |
|
2581 |
} |
||
2582 |
|||
2583 |
34216 |
static void er_add_slice(H264SliceContext *sl, |
|
2584 |
int startx, int starty, |
||
2585 |
int endx, int endy, int status) |
||
2586 |
{ |
||
2587 |
✓✓ | 34216 |
if (!sl->h264->enable_er) |
2588 |
1200 |
return; |
|
2589 |
|||
2590 |
if (CONFIG_ERROR_RESILIENCE) { |
||
2591 |
33016 |
ERContext *er = &sl->h264->slice_ctx[0].er; |
|
2592 |
|||
2593 |
33016 |
ff_er_add_slice(er, startx, starty, endx, endy, status); |
|
2594 |
} |
||
2595 |
} |
||
2596 |
|||
2597 |
34216 |
static int decode_slice(struct AVCodecContext *avctx, void *arg) |
|
2598 |
{ |
||
2599 |
34216 |
H264SliceContext *sl = arg; |
|
2600 |
34216 |
const H264Context *h = sl->h264; |
|
2601 |
34216 |
int lf_x_start = sl->mb_x; |
|
2602 |
34216 |
int orig_deblock = sl->deblocking_filter; |
|
2603 |
int ret; |
||
2604 |
|||
2605 |
34216 |
sl->linesize = h->cur_pic_ptr->f->linesize[0]; |
|
2606 |
34216 |
sl->uvlinesize = h->cur_pic_ptr->f->linesize[1]; |
|
2607 |
|||
2608 |
34216 |
ret = alloc_scratch_buffers(sl, sl->linesize); |
|
2609 |
✗✓ | 34216 |
if (ret < 0) |
2610 |
return ret; |
||
2611 |
|||
2612 |
34216 |
sl->mb_skip_run = -1; |
|
2613 |
|||
2614 |
✗✓ | 34216 |
av_assert0(h->block_offset[15] == (4 * ((scan8[15] - scan8[0]) & 7) << h->pixel_shift) + 4 * sl->linesize * ((scan8[15] - scan8[0]) >> 3)); |
2615 |
|||
2616 |
✗✓ | 34216 |
if (h->postpone_filter) |
2617 |
sl->deblocking_filter = 0; |
||
2618 |
|||
2619 |
✓✓✓✓ |
34216 |
sl->is_complex = FRAME_MBAFF(h) || h->picture_structure != PICT_FRAME || |
2620 |
(CONFIG_GRAY && (h->flags & AV_CODEC_FLAG_GRAY)); |
||
2621 |
|||
2622 |
✓✓✓✓ ✓✗ |
34216 |
if (!(h->avctx->active_thread_type & FF_THREAD_SLICE) && h->picture_structure == PICT_FRAME && h->slice_ctx[0].er.error_status_table) { |
2623 |
24420 |
const int start_i = av_clip(sl->resync_mb_x + sl->resync_mb_y * h->mb_width, 0, h->mb_num - 1); |
|
2624 |
✓✓ | 24420 |
if (start_i) { |
2625 |
5278 |
int prev_status = h->slice_ctx[0].er.error_status_table[h->slice_ctx[0].er.mb_index2xy[start_i - 1]]; |
|
2626 |
5278 |
prev_status &= ~ VP_START; |
|
2627 |
✗✓ | 5278 |
if (prev_status != (ER_MV_END | ER_DC_END | ER_AC_END)) |
2628 |
h->slice_ctx[0].er.error_occurred = 1; |
||
2629 |
} |
||
2630 |
} |
||
2631 |
|||
2632 |
✓✓ | 34216 |
if (h->ps.pps->cabac) { |
2633 |
/* realign */ |
||
2634 |
18293 |
align_get_bits(&sl->gb); |
|
2635 |
|||
2636 |
/* init cabac */ |
||
2637 |
36586 |
ret = ff_init_cabac_decoder(&sl->cabac, |
|
2638 |
18293 |
sl->gb.buffer + get_bits_count(&sl->gb) / 8, |
|
2639 |
18293 |
(get_bits_left(&sl->gb) + 7) / 8); |
|
2640 |
✗✓ | 18293 |
if (ret < 0) |
2641 |
return ret; |
||
2642 |
|||
2643 |
18293 |
ff_h264_init_cabac_states(h, sl); |
|
2644 |
|||
2645 |
8532186 |
for (;;) { |
|
2646 |
int ret, eos; |
||
2647 |
✗✓ | 8550479 |
if (sl->mb_x + sl->mb_y * h->mb_width >= sl->next_slice_idx) { |
2648 |
av_log(h->avctx, AV_LOG_ERROR, "Slice overlaps with next at %d\n", |
||
2649 |
sl->next_slice_idx); |
||
2650 |
er_add_slice(sl, sl->resync_mb_x, sl->resync_mb_y, sl->mb_x, |
||
2651 |
sl->mb_y, ER_MB_ERROR); |
||
2652 |
return AVERROR_INVALIDDATA; |
||
2653 |
} |
||
2654 |
|||
2655 |
8550479 |
ret = ff_h264_decode_mb_cabac(h, sl); |
|
2656 |
|||
2657 |
✓✓ | 8550479 |
if (ret >= 0) |
2658 |
8550478 |
ff_h264_hl_decode_mb(h, sl); |
|
2659 |
|||
2660 |
// FIXME optimal? or let mb_decode decode 16x32 ? |
||
2661 |
✓✓✓✓ |
8550479 |
if (ret >= 0 && FRAME_MBAFF(h)) { |
2662 |
798720 |
sl->mb_y++; |
|
2663 |
|||
2664 |
798720 |
ret = ff_h264_decode_mb_cabac(h, sl); |
|
2665 |
|||
2666 |
✓✗ | 798720 |
if (ret >= 0) |
2667 |
798720 |
ff_h264_hl_decode_mb(h, sl); |
|
2668 |
798720 |
sl->mb_y--; |
|
2669 |
} |
||
2670 |
8550479 |
eos = get_cabac_terminate(&sl->cabac); |
|
2671 |
|||
2672 |
✗✓ | 8550479 |
if ((h->workaround_bugs & FF_BUG_TRUNCATED) && |
2673 |
sl->cabac.bytestream > sl->cabac.bytestream_end + 2) { |
||
2674 |
er_add_slice(sl, sl->resync_mb_x, sl->resync_mb_y, sl->mb_x - 1, |
||
2675 |
sl->mb_y, ER_MB_END); |
||
2676 |
if (sl->mb_x >= lf_x_start) |
||
2677 |
loop_filter(h, sl, lf_x_start, sl->mb_x + 1); |
||
2678 |
goto finish; |
||
2679 |
} |
||
2680 |
✓✓ | 8550479 |
if (sl->cabac.bytestream > sl->cabac.bytestream_end + 2 ) |
2681 |
1 |
av_log(h->avctx, AV_LOG_DEBUG, "bytestream overread %"PTRDIFF_SPECIFIER"\n", sl->cabac.bytestream_end - sl->cabac.bytestream); |
|
2682 |
✓✓✓✓ |
8550479 |
if (ret < 0 || sl->cabac.bytestream > sl->cabac.bytestream_end + 4) { |
2683 |
2 |
av_log(h->avctx, AV_LOG_ERROR, |
|
2684 |
"error while decoding MB %d %d, bytestream %"PTRDIFF_SPECIFIER"\n", |
||
2685 |
sl->mb_x, sl->mb_y, |
||
2686 |
2 |
sl->cabac.bytestream_end - sl->cabac.bytestream); |
|
2687 |
2 |
er_add_slice(sl, sl->resync_mb_x, sl->resync_mb_y, sl->mb_x, |
|
2688 |
sl->mb_y, ER_MB_ERROR); |
||
2689 |
2 |
return AVERROR_INVALIDDATA; |
|
2690 |
} |
||
2691 |
|||
2692 |
✓✓ | 8550477 |
if (++sl->mb_x >= h->mb_width) { |
2693 |
209108 |
loop_filter(h, sl, lf_x_start, sl->mb_x); |
|
2694 |
209108 |
sl->mb_x = lf_x_start = 0; |
|
2695 |
209108 |
decode_finish_row(h, sl); |
|
2696 |
209108 |
++sl->mb_y; |
|
2697 |
✓✓✓✓ |
209108 |
if (FIELD_OR_MBAFF_PICTURE(h)) { |
2698 |
76828 |
++sl->mb_y; |
|
2699 |
✓✓✓✓ |
76828 |
if (FRAME_MBAFF(h) && sl->mb_y < h->mb_height) |
2700 |
17906 |
predict_field_decoding_flag(h, sl); |
|
2701 |
} |
||
2702 |
} |
||
2703 |
|||
2704 |
✓✓✗✓ |
8550477 |
if (eos || sl->mb_y >= h->mb_height) { |
2705 |
ff_tlog(h->avctx, "slice end %d %d\n", |
||
2706 |
get_bits_count(&sl->gb), sl->gb.size_in_bits); |
||
2707 |
18291 |
er_add_slice(sl, sl->resync_mb_x, sl->resync_mb_y, sl->mb_x - 1, |
|
2708 |
sl->mb_y, ER_MB_END); |
||
2709 |
✓✓ | 18291 |
if (sl->mb_x > lf_x_start) |
2710 |
1382 |
loop_filter(h, sl, lf_x_start, sl->mb_x); |
|
2711 |
18291 |
goto finish; |
|
2712 |
} |
||
2713 |
} |
||
2714 |
} else { |
||
2715 |
4076548 |
for (;;) { |
|
2716 |
int ret; |
||
2717 |
|||
2718 |
✗✓ | 4092471 |
if (sl->mb_x + sl->mb_y * h->mb_width >= sl->next_slice_idx) { |
2719 |
av_log(h->avctx, AV_LOG_ERROR, "Slice overlaps with next at %d\n", |
||
2720 |
sl->next_slice_idx); |
||
2721 |
er_add_slice(sl, sl->resync_mb_x, sl->resync_mb_y, sl->mb_x, |
||
2722 |
sl->mb_y, ER_MB_ERROR); |
||
2723 |
return AVERROR_INVALIDDATA; |
||
2724 |
} |
||
2725 |
|||
2726 |
4092471 |
ret = ff_h264_decode_mb_cavlc(h, sl); |
|
2727 |
|||
2728 |
✓✗ | 4092471 |
if (ret >= 0) |
2729 |
4092471 |
ff_h264_hl_decode_mb(h, sl); |
|
2730 |
|||
2731 |
// FIXME optimal? or let mb_decode decode 16x32 ? |
||
2732 |
✓✗✓✓ |
4092471 |
if (ret >= 0 && FRAME_MBAFF(h)) { |
2733 |
254472 |
sl->mb_y++; |
|
2734 |
254472 |
ret = ff_h264_decode_mb_cavlc(h, sl); |
|
2735 |
|||
2736 |
✓✗ | 254472 |
if (ret >= 0) |
2737 |
254472 |
ff_h264_hl_decode_mb(h, sl); |
|
2738 |
254472 |
sl->mb_y--; |
|
2739 |
} |
||
2740 |
|||
2741 |
✗✓ | 4092471 |
if (ret < 0) { |
2742 |
av_log(h->avctx, AV_LOG_ERROR, |
||
2743 |
"error while decoding MB %d %d\n", sl->mb_x, sl->mb_y); |
||
2744 |
er_add_slice(sl, sl->resync_mb_x, sl->resync_mb_y, sl->mb_x, |
||
2745 |
sl->mb_y, ER_MB_ERROR); |
||
2746 |
return ret; |
||
2747 |
} |
||
2748 |
|||
2749 |
✓✓ | 4092471 |
if (++sl->mb_x >= h->mb_width) { |
2750 |
152629 |
loop_filter(h, sl, lf_x_start, sl->mb_x); |
|
2751 |
152629 |
sl->mb_x = lf_x_start = 0; |
|
2752 |
152629 |
decode_finish_row(h, sl); |
|
2753 |
152629 |
++sl->mb_y; |
|
2754 |
✓✓✓✓ |
152629 |
if (FIELD_OR_MBAFF_PICTURE(h)) { |
2755 |
32143 |
++sl->mb_y; |
|
2756 |
✓✓✓✓ |
32143 |
if (FRAME_MBAFF(h) && sl->mb_y < h->mb_height) |
2757 |
4978 |
predict_field_decoding_flag(h, sl); |
|
2758 |
} |
||
2759 |
✓✓ | 152629 |
if (sl->mb_y >= h->mb_height) { |
2760 |
ff_tlog(h->avctx, "slice end %d %d\n", |
||
2761 |
get_bits_count(&sl->gb), sl->gb.size_in_bits); |
||
2762 |
|||
2763 |
✗✓ | 12346 |
if ( get_bits_left(&sl->gb) == 0 |
2764 |
|| get_bits_left(&sl->gb) > 0 && !(h->avctx->err_recognition & AV_EF_AGGRESSIVE)) { |
||
2765 |
12346 |
er_add_slice(sl, sl->resync_mb_x, sl->resync_mb_y, |
|
2766 |
12346 |
sl->mb_x - 1, sl->mb_y, ER_MB_END); |
|
2767 |
|||
2768 |
12346 |
goto finish; |
|
2769 |
} else { |
||
2770 |
er_add_slice(sl, sl->resync_mb_x, sl->resync_mb_y, |
||
2771 |
sl->mb_x, sl->mb_y, ER_MB_END); |
||
2772 |
|||
2773 |
return AVERROR_INVALIDDATA; |
||
2774 |
} |
||
2775 |
} |
||
2776 |
} |
||
2777 |
|||
2778 |
✓✓✓✓ |
4080125 |
if (get_bits_left(&sl->gb) <= 0 && sl->mb_skip_run <= 0) { |
2779 |
ff_tlog(h->avctx, "slice end %d %d\n", |
||
2780 |
get_bits_count(&sl->gb), sl->gb.size_in_bits); |
||
2781 |
|||
2782 |
✓✗ | 3577 |
if (get_bits_left(&sl->gb) == 0) { |
2783 |
3577 |
er_add_slice(sl, sl->resync_mb_x, sl->resync_mb_y, |
|
2784 |
3577 |
sl->mb_x - 1, sl->mb_y, ER_MB_END); |
|
2785 |
✓✓ | 3577 |
if (sl->mb_x > lf_x_start) |
2786 |
2713 |
loop_filter(h, sl, lf_x_start, sl->mb_x); |
|
2787 |
|||
2788 |
3577 |
goto finish; |
|
2789 |
} else { |
||
2790 |
er_add_slice(sl, sl->resync_mb_x, sl->resync_mb_y, sl->mb_x, |
||
2791 |
sl->mb_y, ER_MB_ERROR); |
||
2792 |
|||
2793 |
return AVERROR_INVALIDDATA; |
||
2794 |
} |
||
2795 |
} |
||
2796 |
} |
||
2797 |
} |
||
2798 |
|||
2799 |
34214 |
finish: |
|
2800 |
34214 |
sl->deblocking_filter = orig_deblock; |
|
2801 |
34214 |
return 0; |
|
2802 |
} |
||
2803 |
|||
2804 |
/** |
||
2805 |
* Call decode_slice() for each context. |
||
2806 |
* |
||
2807 |
* @param h h264 master context |
||
2808 |
*/ |
||
2809 |
60505 |
int ff_h264_execute_decode_slices(H264Context *h) |
|
2810 |
{ |
||
2811 |
60505 |
AVCodecContext *const avctx = h->avctx; |
|
2812 |
H264SliceContext *sl; |
||
2813 |
60505 |
int context_count = h->nb_slice_ctx_queued; |
|
2814 |
60505 |
int ret = 0; |
|
2815 |
int i, j; |
||
2816 |
|||
2817 |
60505 |
h->slice_ctx[0].next_slice_idx = INT_MAX; |
|
2818 |
|||
2819 |
✓✗✓✓ |
60505 |
if (h->avctx->hwaccel || context_count < 1) |
2820 |
26739 |
return 0; |
|
2821 |
|||
2822 |
✓✗✗✓ |
33766 |
av_assert0(context_count && h->slice_ctx[context_count - 1].mb_y < h->mb_height); |
2823 |
|||
2824 |
✓✓ | 33766 |
if (context_count == 1) { |
2825 |
|||
2826 |
33316 |
h->slice_ctx[0].next_slice_idx = h->mb_width * h->mb_height; |
|
2827 |
33316 |
h->postpone_filter = 0; |
|
2828 |
|||
2829 |
33316 |
ret = decode_slice(avctx, &h->slice_ctx[0]); |
|
2830 |
33316 |
h->mb_y = h->slice_ctx[0].mb_y; |
|
2831 |
✓✓ | 33316 |
if (ret < 0) |
2832 |
2 |
goto finish; |
|
2833 |
} else { |
||
2834 |
✗✓ | 450 |
av_assert0(context_count > 0); |
2835 |
✓✓ | 1350 |
for (i = 0; i < context_count; i++) { |
2836 |
900 |
int next_slice_idx = h->mb_width * h->mb_height; |
|
2837 |
int slice_idx; |
||
2838 |
|||
2839 |
900 |
sl = &h->slice_ctx[i]; |
|
2840 |
if (CONFIG_ERROR_RESILIENCE) { |
||
2841 |
900 |
sl->er.error_count = 0; |
|
2842 |
} |
||
2843 |
|||
2844 |
/* make sure none of those slices overlap */ |
||
2845 |
900 |
slice_idx = sl->mb_y * h->mb_width + sl->mb_x; |
|
2846 |
✓✓ | 2700 |
for (j = 0; j < context_count; j++) { |
2847 |
1800 |
H264SliceContext *sl2 = &h->slice_ctx[j]; |
|
2848 |
1800 |
int slice_idx2 = sl2->mb_y * h->mb_width + sl2->mb_x; |
|
2849 |
|||
2850 |
✓✓✓✓ |
1800 |
if (i == j || slice_idx2 < slice_idx) |
2851 |
1350 |
continue; |
|
2852 |
450 |
next_slice_idx = FFMIN(next_slice_idx, slice_idx2); |
|
2853 |
} |
||
2854 |
900 |
sl->next_slice_idx = next_slice_idx; |
|
2855 |
} |
||
2856 |
|||
2857 |
450 |
avctx->execute(avctx, decode_slice, h->slice_ctx, |
|
2858 |
NULL, context_count, sizeof(h->slice_ctx[0])); |
||
2859 |
|||
2860 |
/* pull back stuff from slices to master context */ |
||
2861 |
450 |
sl = &h->slice_ctx[context_count - 1]; |
|
2862 |
450 |
h->mb_y = sl->mb_y; |
|
2863 |
if (CONFIG_ERROR_RESILIENCE) { |
||
2864 |
✓✓ | 900 |
for (i = 1; i < context_count; i++) |
2865 |
450 |
h->slice_ctx[0].er.error_count += h->slice_ctx[i].er.error_count; |
|
2866 |
} |
||
2867 |
|||
2868 |
✓✗ | 450 |
if (h->postpone_filter) { |
2869 |
h->postpone_filter = 0; |
||
2870 |
|||
2871 |
for (i = 0; i < context_count; i++) { |
||
2872 |
int y_end, x_end; |
||
2873 |
|||
2874 |
sl = &h->slice_ctx[i]; |
||
2875 |
y_end = FFMIN(sl->mb_y + 1, h->mb_height); |
||
2876 |
x_end = (sl->mb_y >= h->mb_height) ? h->mb_width : sl->mb_x; |
||
2877 |
|||
2878 |
for (j = sl->resync_mb_y; j < y_end; j += 1 + FIELD_OR_MBAFF_PICTURE(h)) { |
||
2879 |
sl->mb_y = j; |
||
2880 |
loop_filter(h, sl, j > sl->resync_mb_y ? 0 : sl->resync_mb_x, |
||
2881 |
j == y_end - 1 ? x_end : h->mb_width); |
||
2882 |
} |
||
2883 |
} |
||
2884 |
} |
||
2885 |
} |
||
2886 |
|||
2887 |
450 |
finish: |
|
2888 |
33766 |
h->nb_slice_ctx_queued = 0; |
|
2889 |
33766 |
return ret; |
|
2890 |
} |
Generated by: GCOVR (Version 4.2) |