FFmpeg
vp8.c
Go to the documentation of this file.
1 /*
2  * VP7/VP8 compatible video decoder
3  *
4  * Copyright (C) 2010 David Conrad
5  * Copyright (C) 2010 Ronald S. Bultje
6  * Copyright (C) 2010 Fiona Glaser
7  * Copyright (C) 2012 Daniel Kang
8  * Copyright (C) 2014 Peter Ross
9  *
10  * This file is part of FFmpeg.
11  *
12  * FFmpeg is free software; you can redistribute it and/or
13  * modify it under the terms of the GNU Lesser General Public
14  * License as published by the Free Software Foundation; either
15  * version 2.1 of the License, or (at your option) any later version.
16  *
17  * FFmpeg is distributed in the hope that it will be useful,
18  * but WITHOUT ANY WARRANTY; without even the implied warranty of
19  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
20  * Lesser General Public License for more details.
21  *
22  * You should have received a copy of the GNU Lesser General Public
23  * License along with FFmpeg; if not, write to the Free Software
24  * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
25  */
26 
27 #include "config_components.h"
28 
29 #include "libavutil/mem_internal.h"
30 
31 #include "avcodec.h"
32 #include "codec_internal.h"
33 #include "decode.h"
34 #include "hwaccel_internal.h"
35 #include "hwconfig.h"
36 #include "mathops.h"
37 #include "refstruct.h"
38 #include "thread.h"
39 #include "threadframe.h"
40 #include "vp8.h"
41 #include "vp89_rac.h"
42 #include "vp8data.h"
43 #include "vpx_rac.h"
44 
45 #if ARCH_ARM
46 # include "arm/vp8.h"
47 #endif
48 
49 // fixme: add 1 bit to all the calls to this?
51 {
52  int v;
53 
54  if (!vp89_rac_get(c))
55  return 0;
56 
57  v = vp89_rac_get_uint(c, bits);
58 
59  if (vp89_rac_get(c))
60  v = -v;
61 
62  return v;
63 }
64 
66 {
67  int v = vp89_rac_get_uint(c, 7) << 1;
68  return v + !v;
69 }
70 
71 // DCTextra
72 static int vp8_rac_get_coeff(VPXRangeCoder *c, const uint8_t *prob)
73 {
74  int v = 0;
75 
76  do {
77  v = (v<<1) + vpx_rac_get_prob(c, *prob++);
78  } while (*prob);
79 
80  return v;
81 }
82 
83 static void free_buffers(VP8Context *s)
84 {
85  int i;
86  if (s->thread_data)
87  for (i = 0; i < MAX_THREADS; i++) {
88 #if HAVE_THREADS
89  pthread_cond_destroy(&s->thread_data[i].cond);
90  pthread_mutex_destroy(&s->thread_data[i].lock);
91 #endif
92  av_freep(&s->thread_data[i].filter_strength);
93  }
94  av_freep(&s->thread_data);
95  av_freep(&s->macroblocks_base);
96  av_freep(&s->intra4x4_pred_mode_top);
97  av_freep(&s->top_nnz);
98  av_freep(&s->top_border);
99 
100  s->macroblocks = NULL;
101 }
102 
104 {
105  int ret;
106  if ((ret = ff_thread_get_ext_buffer(s->avctx, &f->tf,
107  ref ? AV_GET_BUFFER_FLAG_REF : 0)) < 0)
108  return ret;
109  f->seg_map = ff_refstruct_allocz(s->mb_width * s->mb_height);
110  if (!f->seg_map) {
111  ret = AVERROR(ENOMEM);
112  goto fail;
113  }
114  ret = ff_hwaccel_frame_priv_alloc(s->avctx, &f->hwaccel_picture_private);
115  if (ret < 0)
116  goto fail;
117 
118  return 0;
119 
120 fail:
121  ff_refstruct_unref(&f->seg_map);
123  return ret;
124 }
125 
127 {
128  ff_refstruct_unref(&f->seg_map);
129  ff_refstruct_unref(&f->hwaccel_picture_private);
131 }
132 
133 #if CONFIG_VP8_DECODER
134 static int vp8_ref_frame(VP8Frame *dst, const VP8Frame *src)
135 {
136  int ret;
137 
138  vp8_release_frame(dst);
139 
140  if ((ret = ff_thread_ref_frame(&dst->tf, &src->tf)) < 0)
141  return ret;
142  ff_refstruct_replace(&dst->seg_map, src->seg_map);
144  src->hwaccel_picture_private);
145 
146  return 0;
147 }
148 #endif /* CONFIG_VP8_DECODER */
149 
150 static void vp8_decode_flush_impl(AVCodecContext *avctx, int free_mem)
151 {
152  VP8Context *s = avctx->priv_data;
153  int i;
154 
155  for (i = 0; i < FF_ARRAY_ELEMS(s->frames); i++)
156  vp8_release_frame(&s->frames[i]);
157  memset(s->framep, 0, sizeof(s->framep));
158 
159  if (free_mem)
160  free_buffers(s);
161 
162  if (FF_HW_HAS_CB(avctx, flush))
163  FF_HW_SIMPLE_CALL(avctx, flush);
164 }
165 
166 static void vp8_decode_flush(AVCodecContext *avctx)
167 {
168  vp8_decode_flush_impl(avctx, 0);
169 }
170 
172 {
173  VP8Frame *frame = NULL;
174  int i;
175 
176  // find a free buffer
177  for (i = 0; i < 5; i++)
178  if (&s->frames[i] != s->framep[VP8_FRAME_CURRENT] &&
179  &s->frames[i] != s->framep[VP8_FRAME_PREVIOUS] &&
180  &s->frames[i] != s->framep[VP8_FRAME_GOLDEN] &&
181  &s->frames[i] != s->framep[VP8_FRAME_ALTREF]) {
182  frame = &s->frames[i];
183  break;
184  }
185  if (i == 5) {
186  av_log(s->avctx, AV_LOG_FATAL, "Ran out of free frames!\n");
187  abort();
188  }
189  if (frame->tf.f->buf[0])
191 
192  return frame;
193 }
194 
196 {
197  enum AVPixelFormat pix_fmts[] = {
198 #if CONFIG_VP8_VAAPI_HWACCEL
200 #endif
201 #if CONFIG_VP8_NVDEC_HWACCEL
203 #endif
206  };
207 
208  return ff_get_format(s->avctx, pix_fmts);
209 }
210 
211 static av_always_inline
212 int update_dimensions(VP8Context *s, int width, int height, int is_vp7)
213 {
214  AVCodecContext *avctx = s->avctx;
215  int i, ret, dim_reset = 0;
216 
217  if (width != s->avctx->width || ((width+15)/16 != s->mb_width || (height+15)/16 != s->mb_height) && s->macroblocks_base ||
218  height != s->avctx->height) {
219  vp8_decode_flush_impl(s->avctx, 1);
220 
221  ret = ff_set_dimensions(s->avctx, width, height);
222  if (ret < 0)
223  return ret;
224 
225  dim_reset = (s->macroblocks_base != NULL);
226  }
227 
228  if ((s->pix_fmt == AV_PIX_FMT_NONE || dim_reset) &&
229  !s->actually_webp && !is_vp7) {
230  s->pix_fmt = get_pixel_format(s);
231  if (s->pix_fmt < 0)
232  return AVERROR(EINVAL);
233  avctx->pix_fmt = s->pix_fmt;
234  }
235 
236  s->mb_width = (s->avctx->coded_width + 15) / 16;
237  s->mb_height = (s->avctx->coded_height + 15) / 16;
238 
239  s->mb_layout = is_vp7 || avctx->active_thread_type == FF_THREAD_SLICE &&
240  avctx->thread_count > 1;
241  if (!s->mb_layout) { // Frame threading and one thread
242  s->macroblocks_base = av_mallocz((s->mb_width + s->mb_height * 2 + 1) *
243  sizeof(*s->macroblocks));
244  s->intra4x4_pred_mode_top = av_mallocz(s->mb_width * 4);
245  } else // Sliced threading
246  s->macroblocks_base = av_mallocz((s->mb_width + 2) * (s->mb_height + 2) *
247  sizeof(*s->macroblocks));
248  s->top_nnz = av_mallocz(s->mb_width * sizeof(*s->top_nnz));
249  s->top_border = av_mallocz((s->mb_width + 1) * sizeof(*s->top_border));
250  s->thread_data = av_mallocz(MAX_THREADS * sizeof(VP8ThreadData));
251 
252  if (!s->macroblocks_base || !s->top_nnz || !s->top_border ||
253  !s->thread_data || (!s->intra4x4_pred_mode_top && !s->mb_layout)) {
254  free_buffers(s);
255  return AVERROR(ENOMEM);
256  }
257 
258  for (i = 0; i < MAX_THREADS; i++) {
259  s->thread_data[i].filter_strength =
260  av_mallocz(s->mb_width * sizeof(*s->thread_data[0].filter_strength));
261  if (!s->thread_data[i].filter_strength) {
262  free_buffers(s);
263  return AVERROR(ENOMEM);
264  }
265 #if HAVE_THREADS
266  ret = pthread_mutex_init(&s->thread_data[i].lock, NULL);
267  if (ret) {
268  free_buffers(s);
269  return AVERROR(ret);
270  }
271  ret = pthread_cond_init(&s->thread_data[i].cond, NULL);
272  if (ret) {
273  free_buffers(s);
274  return AVERROR(ret);
275  }
276 #endif
277  }
278 
279  s->macroblocks = s->macroblocks_base + 1;
280 
281  return 0;
282 }
283 
285 {
287 }
288 
290 {
292 }
293 
294 
296 {
297  VPXRangeCoder *c = &s->c;
298  int i;
299 
300  s->segmentation.update_map = vp89_rac_get(c);
301  s->segmentation.update_feature_data = vp89_rac_get(c);
302 
303  if (s->segmentation.update_feature_data) {
304  s->segmentation.absolute_vals = vp89_rac_get(c);
305 
306  for (i = 0; i < 4; i++)
307  s->segmentation.base_quant[i] = vp8_rac_get_sint(c, 7);
308 
309  for (i = 0; i < 4; i++)
310  s->segmentation.filter_level[i] = vp8_rac_get_sint(c, 6);
311  }
312  if (s->segmentation.update_map)
313  for (i = 0; i < 3; i++)
314  s->prob->segmentid[i] = vp89_rac_get(c) ? vp89_rac_get_uint(c, 8) : 255;
315 }
316 
318 {
319  VPXRangeCoder *c = &s->c;
320  int i;
321 
322  for (i = 0; i < 4; i++) {
323  if (vp89_rac_get(c)) {
324  s->lf_delta.ref[i] = vp89_rac_get_uint(c, 6);
325 
326  if (vp89_rac_get(c))
327  s->lf_delta.ref[i] = -s->lf_delta.ref[i];
328  }
329  }
330 
331  for (i = MODE_I4x4; i <= VP8_MVMODE_SPLIT; i++) {
332  if (vp89_rac_get(c)) {
333  s->lf_delta.mode[i] = vp89_rac_get_uint(c, 6);
334 
335  if (vp89_rac_get(c))
336  s->lf_delta.mode[i] = -s->lf_delta.mode[i];
337  }
338  }
339 }
340 
341 static int setup_partitions(VP8Context *s, const uint8_t *buf, int buf_size)
342 {
343  const uint8_t *sizes = buf;
344  int i;
345  int ret;
346 
347  s->num_coeff_partitions = 1 << vp89_rac_get_uint(&s->c, 2);
348 
349  buf += 3 * (s->num_coeff_partitions - 1);
350  buf_size -= 3 * (s->num_coeff_partitions - 1);
351  if (buf_size < 0)
352  return -1;
353 
354  for (i = 0; i < s->num_coeff_partitions - 1; i++) {
355  int size = AV_RL24(sizes + 3 * i);
356  if (buf_size - size < 0)
357  return -1;
358  s->coeff_partition_size[i] = size;
359 
360  ret = ff_vpx_init_range_decoder(&s->coeff_partition[i], buf, size);
361  if (ret < 0)
362  return ret;
363  buf += size;
364  buf_size -= size;
365  }
366 
367  s->coeff_partition_size[i] = buf_size;
368 
369  return ff_vpx_init_range_decoder(&s->coeff_partition[i], buf, buf_size);
370 }
371 
373 {
374  VPXRangeCoder *c = &s->c;
375 
376  int yac_qi = vp89_rac_get_uint(c, 7);
377  int ydc_qi = vp89_rac_get(c) ? vp89_rac_get_uint(c, 7) : yac_qi;
378  int y2dc_qi = vp89_rac_get(c) ? vp89_rac_get_uint(c, 7) : yac_qi;
379  int y2ac_qi = vp89_rac_get(c) ? vp89_rac_get_uint(c, 7) : yac_qi;
380  int uvdc_qi = vp89_rac_get(c) ? vp89_rac_get_uint(c, 7) : yac_qi;
381  int uvac_qi = vp89_rac_get(c) ? vp89_rac_get_uint(c, 7) : yac_qi;
382 
383  s->qmat[0].luma_qmul[0] = vp7_ydc_qlookup[ydc_qi];
384  s->qmat[0].luma_qmul[1] = vp7_yac_qlookup[yac_qi];
385  s->qmat[0].luma_dc_qmul[0] = vp7_y2dc_qlookup[y2dc_qi];
386  s->qmat[0].luma_dc_qmul[1] = vp7_y2ac_qlookup[y2ac_qi];
387  s->qmat[0].chroma_qmul[0] = FFMIN(vp7_ydc_qlookup[uvdc_qi], 132);
388  s->qmat[0].chroma_qmul[1] = vp7_yac_qlookup[uvac_qi];
389 }
390 
392 {
393  VPXRangeCoder *c = &s->c;
394  int i, base_qi;
395 
396  s->quant.yac_qi = vp89_rac_get_uint(c, 7);
397  s->quant.ydc_delta = vp8_rac_get_sint(c, 4);
398  s->quant.y2dc_delta = vp8_rac_get_sint(c, 4);
399  s->quant.y2ac_delta = vp8_rac_get_sint(c, 4);
400  s->quant.uvdc_delta = vp8_rac_get_sint(c, 4);
401  s->quant.uvac_delta = vp8_rac_get_sint(c, 4);
402 
403  for (i = 0; i < 4; i++) {
404  if (s->segmentation.enabled) {
405  base_qi = s->segmentation.base_quant[i];
406  if (!s->segmentation.absolute_vals)
407  base_qi += s->quant.yac_qi;
408  } else
409  base_qi = s->quant.yac_qi;
410 
411  s->qmat[i].luma_qmul[0] = vp8_dc_qlookup[av_clip_uintp2(base_qi + s->quant.ydc_delta, 7)];
412  s->qmat[i].luma_qmul[1] = vp8_ac_qlookup[av_clip_uintp2(base_qi, 7)];
413  s->qmat[i].luma_dc_qmul[0] = vp8_dc_qlookup[av_clip_uintp2(base_qi + s->quant.y2dc_delta, 7)] * 2;
414  /* 101581>>16 is equivalent to 155/100 */
415  s->qmat[i].luma_dc_qmul[1] = vp8_ac_qlookup[av_clip_uintp2(base_qi + s->quant.y2ac_delta, 7)] * 101581 >> 16;
416  s->qmat[i].chroma_qmul[0] = vp8_dc_qlookup[av_clip_uintp2(base_qi + s->quant.uvdc_delta, 7)];
417  s->qmat[i].chroma_qmul[1] = vp8_ac_qlookup[av_clip_uintp2(base_qi + s->quant.uvac_delta, 7)];
418 
419  s->qmat[i].luma_dc_qmul[1] = FFMAX(s->qmat[i].luma_dc_qmul[1], 8);
420  s->qmat[i].chroma_qmul[0] = FFMIN(s->qmat[i].chroma_qmul[0], 132);
421  }
422 }
423 
424 /**
425  * Determine which buffers golden and altref should be updated with after this frame.
426  * The spec isn't clear here, so I'm going by my understanding of what libvpx does
427  *
428  * Intra frames update all 3 references
429  * Inter frames update VP8_FRAME_PREVIOUS if the update_last flag is set
430  * If the update (golden|altref) flag is set, it's updated with the current frame
431  * if update_last is set, and VP8_FRAME_PREVIOUS otherwise.
432  * If the flag is not set, the number read means:
433  * 0: no update
434  * 1: VP8_FRAME_PREVIOUS
435  * 2: update golden with altref, or update altref with golden
436  */
438 {
439  VPXRangeCoder *c = &s->c;
440 
441  if (update)
442  return VP8_FRAME_CURRENT;
443 
444  switch (vp89_rac_get_uint(c, 2)) {
445  case 1:
446  return VP8_FRAME_PREVIOUS;
447  case 2:
449  }
450  return VP8_FRAME_NONE;
451 }
452 
454 {
455  int i, j;
456  for (i = 0; i < 4; i++)
457  for (j = 0; j < 16; j++)
458  memcpy(s->prob->token[i][j], vp8_token_default_probs[i][vp8_coeff_band[j]],
459  sizeof(s->prob->token[i][j]));
460 }
461 
463 {
464  VPXRangeCoder *c = &s->c;
465  int i, j, k, l, m;
466 
467  for (i = 0; i < 4; i++)
468  for (j = 0; j < 8; j++)
469  for (k = 0; k < 3; k++)
470  for (l = 0; l < NUM_DCT_TOKENS-1; l++)
472  int prob = vp89_rac_get_uint(c, 8);
473  for (m = 0; vp8_coeff_band_indexes[j][m] >= 0; m++)
474  s->prob->token[i][vp8_coeff_band_indexes[j][m]][k][l] = prob;
475  }
476 }
477 
478 #define VP7_MVC_SIZE 17
479 #define VP8_MVC_SIZE 19
480 
482  int mvc_size)
483 {
484  VPXRangeCoder *c = &s->c;
485  int i, j;
486 
487  if (vp89_rac_get(c))
488  for (i = 0; i < 4; i++)
489  s->prob->pred16x16[i] = vp89_rac_get_uint(c, 8);
490  if (vp89_rac_get(c))
491  for (i = 0; i < 3; i++)
492  s->prob->pred8x8c[i] = vp89_rac_get_uint(c, 8);
493 
494  // 17.2 MV probability update
495  for (i = 0; i < 2; i++)
496  for (j = 0; j < mvc_size; j++)
498  s->prob->mvc[i][j] = vp8_rac_get_nn(c);
499 }
500 
501 static void update_refs(VP8Context *s)
502 {
503  VPXRangeCoder *c = &s->c;
504 
505  int update_golden = vp89_rac_get(c);
506  int update_altref = vp89_rac_get(c);
507 
508  s->update_golden = ref_to_update(s, update_golden, VP8_FRAME_GOLDEN);
509  s->update_altref = ref_to_update(s, update_altref, VP8_FRAME_ALTREF);
510 }
511 
512 static void copy_chroma(AVFrame *dst, const AVFrame *src, int width, int height)
513 {
514  int i, j;
515 
516  for (j = 1; j < 3; j++) {
517  for (i = 0; i < height / 2; i++)
518  memcpy(dst->data[j] + i * dst->linesize[j],
519  src->data[j] + i * src->linesize[j], width / 2);
520  }
521 }
522 
523 static void fade(uint8_t *dst, ptrdiff_t dst_linesize,
524  const uint8_t *src, ptrdiff_t src_linesize,
525  int width, int height,
526  int alpha, int beta)
527 {
528  int i, j;
529  for (j = 0; j < height; j++) {
530  const uint8_t *src2 = src + j * src_linesize;
531  uint8_t *dst2 = dst + j * dst_linesize;
532  for (i = 0; i < width; i++) {
533  uint8_t y = src2[i];
534  dst2[i] = av_clip_uint8(y + ((y * beta) >> 8) + alpha);
535  }
536  }
537 }
538 
539 static int vp7_fade_frame(VP8Context *s, int alpha, int beta)
540 {
541  int ret;
542 
543  if (!s->keyframe && (alpha || beta)) {
544  int width = s->mb_width * 16;
545  int height = s->mb_height * 16;
546  const AVFrame *src;
547  AVFrame *dst;
548 
549  if (!s->framep[VP8_FRAME_PREVIOUS] ||
550  !s->framep[VP8_FRAME_GOLDEN]) {
551  av_log(s->avctx, AV_LOG_WARNING, "Discarding interframe without a prior keyframe!\n");
552  return AVERROR_INVALIDDATA;
553  }
554 
555  src =
556  dst = s->framep[VP8_FRAME_PREVIOUS]->tf.f;
557 
558  /* preserve the golden frame, write a new previous frame */
559  if (s->framep[VP8_FRAME_GOLDEN] == s->framep[VP8_FRAME_PREVIOUS]) {
561  if ((ret = vp8_alloc_frame(s, s->framep[VP8_FRAME_PREVIOUS], 1)) < 0)
562  return ret;
563 
564  dst = s->framep[VP8_FRAME_PREVIOUS]->tf.f;
565 
566  copy_chroma(dst, src, width, height);
567  }
568 
569  fade(dst->data[0], dst->linesize[0],
570  src->data[0], src->linesize[0],
571  width, height, alpha, beta);
572  }
573 
574  return 0;
575 }
576 
577 static int vp7_decode_frame_header(VP8Context *s, const uint8_t *buf, int buf_size)
578 {
579  VPXRangeCoder *c = &s->c;
580  int part1_size, hscale, vscale, i, j, ret;
581  int width = s->avctx->width;
582  int height = s->avctx->height;
583  int alpha = 0;
584  int beta = 0;
585  int fade_present = 1;
586 
587  if (buf_size < 4) {
588  return AVERROR_INVALIDDATA;
589  }
590 
591  s->profile = (buf[0] >> 1) & 7;
592  if (s->profile > 1) {
593  avpriv_request_sample(s->avctx, "Unknown profile %d", s->profile);
594  return AVERROR_INVALIDDATA;
595  }
596 
597  s->keyframe = !(buf[0] & 1);
598  s->invisible = 0;
599  part1_size = AV_RL24(buf) >> 4;
600 
601  if (buf_size < 4 - s->profile + part1_size) {
602  av_log(s->avctx, AV_LOG_ERROR, "Buffer size %d is too small, needed : %d\n", buf_size, 4 - s->profile + part1_size);
603  return AVERROR_INVALIDDATA;
604  }
605 
606  buf += 4 - s->profile;
607  buf_size -= 4 - s->profile;
608 
609  memcpy(s->put_pixels_tab, s->vp8dsp.put_vp8_epel_pixels_tab, sizeof(s->put_pixels_tab));
610 
611  ret = ff_vpx_init_range_decoder(c, buf, part1_size);
612  if (ret < 0)
613  return ret;
614  buf += part1_size;
615  buf_size -= part1_size;
616 
617  /* A. Dimension information (keyframes only) */
618  if (s->keyframe) {
619  width = vp89_rac_get_uint(c, 12);
620  height = vp89_rac_get_uint(c, 12);
621  hscale = vp89_rac_get_uint(c, 2);
622  vscale = vp89_rac_get_uint(c, 2);
623  if (hscale || vscale)
624  avpriv_request_sample(s->avctx, "Upscaling");
625 
626  s->update_golden = s->update_altref = VP8_FRAME_CURRENT;
628  memcpy(s->prob->pred16x16, vp8_pred16x16_prob_inter,
629  sizeof(s->prob->pred16x16));
630  memcpy(s->prob->pred8x8c, vp8_pred8x8c_prob_inter,
631  sizeof(s->prob->pred8x8c));
632  for (i = 0; i < 2; i++)
633  memcpy(s->prob->mvc[i], vp7_mv_default_prob[i],
634  sizeof(vp7_mv_default_prob[i]));
635  memset(&s->segmentation, 0, sizeof(s->segmentation));
636  memset(&s->lf_delta, 0, sizeof(s->lf_delta));
637  memcpy(s->prob[0].scan, ff_zigzag_scan, sizeof(s->prob[0].scan));
638  }
639 
640  if (s->keyframe || s->profile > 0)
641  memset(s->inter_dc_pred, 0 , sizeof(s->inter_dc_pred));
642 
643  /* B. Decoding information for all four macroblock-level features */
644  for (i = 0; i < 4; i++) {
645  s->feature_enabled[i] = vp89_rac_get(c);
646  if (s->feature_enabled[i]) {
647  s->feature_present_prob[i] = vp89_rac_get_uint(c, 8);
648 
649  for (j = 0; j < 3; j++)
650  s->feature_index_prob[i][j] =
651  vp89_rac_get(c) ? vp89_rac_get_uint(c, 8) : 255;
652 
653  if (vp7_feature_value_size[s->profile][i])
654  for (j = 0; j < 4; j++)
655  s->feature_value[i][j] =
657  }
658  }
659 
660  s->segmentation.enabled = 0;
661  s->segmentation.update_map = 0;
662  s->lf_delta.enabled = 0;
663 
664  s->num_coeff_partitions = 1;
665  ret = ff_vpx_init_range_decoder(&s->coeff_partition[0], buf, buf_size);
666  if (ret < 0)
667  return ret;
668 
669  if (!s->macroblocks_base || /* first frame */
670  width != s->avctx->width || height != s->avctx->height ||
671  (width + 15) / 16 != s->mb_width || (height + 15) / 16 != s->mb_height) {
672  if ((ret = vp7_update_dimensions(s, width, height)) < 0)
673  return ret;
674  }
675 
676  /* C. Dequantization indices */
677  vp7_get_quants(s);
678 
679  /* D. Golden frame update flag (a Flag) for interframes only */
680  if (!s->keyframe) {
681  s->update_golden = vp89_rac_get(c) ? VP8_FRAME_CURRENT : VP8_FRAME_NONE;
682  s->sign_bias[VP8_FRAME_GOLDEN] = 0;
683  }
684 
685  s->update_last = 1;
686  s->update_probabilities = 1;
687 
688  if (s->profile > 0) {
689  s->update_probabilities = vp89_rac_get(c);
690  if (!s->update_probabilities)
691  s->prob[1] = s->prob[0];
692 
693  if (!s->keyframe)
694  fade_present = vp89_rac_get(c);
695  }
696 
697  if (vpx_rac_is_end(c))
698  return AVERROR_INVALIDDATA;
699  /* E. Fading information for previous frame */
700  if (fade_present && vp89_rac_get(c)) {
701  alpha = (int8_t) vp89_rac_get_uint(c, 8);
702  beta = (int8_t) vp89_rac_get_uint(c, 8);
703  }
704 
705  /* F. Loop filter type */
706  if (!s->profile)
707  s->filter.simple = vp89_rac_get(c);
708 
709  /* G. DCT coefficient ordering specification */
710  if (vp89_rac_get(c))
711  for (i = 1; i < 16; i++)
712  s->prob[0].scan[i] = ff_zigzag_scan[vp89_rac_get_uint(c, 4)];
713 
714  /* H. Loop filter levels */
715  if (s->profile > 0)
716  s->filter.simple = vp89_rac_get(c);
717  s->filter.level = vp89_rac_get_uint(c, 6);
718  s->filter.sharpness = vp89_rac_get_uint(c, 3);
719 
720  /* I. DCT coefficient probability update; 13.3 Token Probability Updates */
722 
723  s->mbskip_enabled = 0;
724 
725  /* J. The remaining frame header data occurs ONLY FOR INTERFRAMES */
726  if (!s->keyframe) {
727  s->prob->intra = vp89_rac_get_uint(c, 8);
728  s->prob->last = vp89_rac_get_uint(c, 8);
730  }
731 
732  if (vpx_rac_is_end(c))
733  return AVERROR_INVALIDDATA;
734 
735  if ((ret = vp7_fade_frame(s, alpha, beta)) < 0)
736  return ret;
737 
738  return 0;
739 }
740 
741 static int vp8_decode_frame_header(VP8Context *s, const uint8_t *buf, int buf_size)
742 {
743  VPXRangeCoder *c = &s->c;
744  int header_size, hscale, vscale, ret;
745  int width = s->avctx->width;
746  int height = s->avctx->height;
747 
748  if (buf_size < 3) {
749  av_log(s->avctx, AV_LOG_ERROR, "Insufficent data (%d) for header\n", buf_size);
750  return AVERROR_INVALIDDATA;
751  }
752 
753  s->keyframe = !(buf[0] & 1);
754  s->profile = (buf[0]>>1) & 7;
755  s->invisible = !(buf[0] & 0x10);
756  header_size = AV_RL24(buf) >> 5;
757  buf += 3;
758  buf_size -= 3;
759 
760  s->header_partition_size = header_size;
761 
762  if (s->profile > 3)
763  av_log(s->avctx, AV_LOG_WARNING, "Unknown profile %d\n", s->profile);
764 
765  if (!s->profile)
766  memcpy(s->put_pixels_tab, s->vp8dsp.put_vp8_epel_pixels_tab,
767  sizeof(s->put_pixels_tab));
768  else // profile 1-3 use bilinear, 4+ aren't defined so whatever
769  memcpy(s->put_pixels_tab, s->vp8dsp.put_vp8_bilinear_pixels_tab,
770  sizeof(s->put_pixels_tab));
771 
772  if (header_size > buf_size - 7 * s->keyframe) {
773  av_log(s->avctx, AV_LOG_ERROR, "Header size larger than data provided\n");
774  return AVERROR_INVALIDDATA;
775  }
776 
777  if (s->keyframe) {
778  if (AV_RL24(buf) != 0x2a019d) {
779  av_log(s->avctx, AV_LOG_ERROR,
780  "Invalid start code 0x%x\n", AV_RL24(buf));
781  return AVERROR_INVALIDDATA;
782  }
783  width = AV_RL16(buf + 3) & 0x3fff;
784  height = AV_RL16(buf + 5) & 0x3fff;
785  hscale = buf[4] >> 6;
786  vscale = buf[6] >> 6;
787  buf += 7;
788  buf_size -= 7;
789 
790  if (hscale || vscale)
791  avpriv_request_sample(s->avctx, "Upscaling");
792 
793  s->update_golden = s->update_altref = VP8_FRAME_CURRENT;
795  memcpy(s->prob->pred16x16, vp8_pred16x16_prob_inter,
796  sizeof(s->prob->pred16x16));
797  memcpy(s->prob->pred8x8c, vp8_pred8x8c_prob_inter,
798  sizeof(s->prob->pred8x8c));
799  memcpy(s->prob->mvc, vp8_mv_default_prob,
800  sizeof(s->prob->mvc));
801  memset(&s->segmentation, 0, sizeof(s->segmentation));
802  memset(&s->lf_delta, 0, sizeof(s->lf_delta));
803  }
804 
805  ret = ff_vpx_init_range_decoder(c, buf, header_size);
806  if (ret < 0)
807  return ret;
808  buf += header_size;
809  buf_size -= header_size;
810 
811  if (s->keyframe) {
812  s->colorspace = vp89_rac_get(c);
813  if (s->colorspace)
814  av_log(s->avctx, AV_LOG_WARNING, "Unspecified colorspace\n");
815  s->fullrange = vp89_rac_get(c);
816  }
817 
818  if ((s->segmentation.enabled = vp89_rac_get(c)))
820  else
821  s->segmentation.update_map = 0; // FIXME: move this to some init function?
822 
823  s->filter.simple = vp89_rac_get(c);
824  s->filter.level = vp89_rac_get_uint(c, 6);
825  s->filter.sharpness = vp89_rac_get_uint(c, 3);
826 
827  if ((s->lf_delta.enabled = vp89_rac_get(c))) {
828  s->lf_delta.update = vp89_rac_get(c);
829  if (s->lf_delta.update)
831  }
832 
833  if (setup_partitions(s, buf, buf_size)) {
834  av_log(s->avctx, AV_LOG_ERROR, "Invalid partitions\n");
835  return AVERROR_INVALIDDATA;
836  }
837 
838  if (!s->macroblocks_base || /* first frame */
839  width != s->avctx->width || height != s->avctx->height ||
840  (width+15)/16 != s->mb_width || (height+15)/16 != s->mb_height)
841  if ((ret = vp8_update_dimensions(s, width, height)) < 0)
842  return ret;
843 
844  vp8_get_quants(s);
845 
846  if (!s->keyframe) {
847  update_refs(s);
848  s->sign_bias[VP8_FRAME_GOLDEN] = vp89_rac_get(c);
849  s->sign_bias[VP8_FRAME_ALTREF] = vp89_rac_get(c);
850  }
851 
852  // if we aren't saving this frame's probabilities for future frames,
853  // make a copy of the current probabilities
854  if (!(s->update_probabilities = vp89_rac_get(c)))
855  s->prob[1] = s->prob[0];
856 
857  s->update_last = s->keyframe || vp89_rac_get(c);
858 
860 
861  if ((s->mbskip_enabled = vp89_rac_get(c)))
862  s->prob->mbskip = vp89_rac_get_uint(c, 8);
863 
864  if (!s->keyframe) {
865  s->prob->intra = vp89_rac_get_uint(c, 8);
866  s->prob->last = vp89_rac_get_uint(c, 8);
867  s->prob->golden = vp89_rac_get_uint(c, 8);
869  }
870 
871  // Record the entropy coder state here so that hwaccels can use it.
872  s->c.code_word = vpx_rac_renorm(&s->c);
873  s->coder_state_at_header_end.input = s->c.buffer - (-s->c.bits / 8);
874  s->coder_state_at_header_end.range = s->c.high;
875  s->coder_state_at_header_end.value = s->c.code_word >> 16;
876  s->coder_state_at_header_end.bit_count = -s->c.bits % 8;
877 
878  return 0;
879 }
880 
881 static av_always_inline
882 void clamp_mv(const VP8mvbounds *s, VP8mv *dst, const VP8mv *src)
883 {
884  dst->x = av_clip(src->x, av_clip(s->mv_min.x, INT16_MIN, INT16_MAX),
885  av_clip(s->mv_max.x, INT16_MIN, INT16_MAX));
886  dst->y = av_clip(src->y, av_clip(s->mv_min.y, INT16_MIN, INT16_MAX),
887  av_clip(s->mv_max.y, INT16_MIN, INT16_MAX));
888 }
889 
890 /**
891  * Motion vector coding, 17.1.
892  */
893 static av_always_inline int read_mv_component(VPXRangeCoder *c, const uint8_t *p, int vp7)
894 {
895  int bit, x = 0;
896 
897  if (vpx_rac_get_prob_branchy(c, p[0])) {
898  int i;
899 
900  for (i = 0; i < 3; i++)
901  x += vpx_rac_get_prob(c, p[9 + i]) << i;
902  for (i = (vp7 ? 7 : 9); i > 3; i--)
903  x += vpx_rac_get_prob(c, p[9 + i]) << i;
904  if (!(x & (vp7 ? 0xF0 : 0xFFF0)) || vpx_rac_get_prob(c, p[12]))
905  x += 8;
906  } else {
907  // small_mvtree
908  const uint8_t *ps = p + 2;
909  bit = vpx_rac_get_prob(c, *ps);
910  ps += 1 + 3 * bit;
911  x += 4 * bit;
912  bit = vpx_rac_get_prob(c, *ps);
913  ps += 1 + bit;
914  x += 2 * bit;
915  x += vpx_rac_get_prob(c, *ps);
916  }
917 
918  return (x && vpx_rac_get_prob(c, p[1])) ? -x : x;
919 }
920 
921 static int vp7_read_mv_component(VPXRangeCoder *c, const uint8_t *p)
922 {
923  return read_mv_component(c, p, 1);
924 }
925 
926 static int vp8_read_mv_component(VPXRangeCoder *c, const uint8_t *p)
927 {
928  return read_mv_component(c, p, 0);
929 }
930 
931 static av_always_inline
932 const uint8_t *get_submv_prob(uint32_t left, uint32_t top, int is_vp7)
933 {
934  if (is_vp7)
935  return vp7_submv_prob;
936 
937  if (left == top)
938  return vp8_submv_prob[4 - !!left];
939  if (!top)
940  return vp8_submv_prob[2];
941  return vp8_submv_prob[1 - !!left];
942 }
943 
944 /**
945  * Split motion vector prediction, 16.4.
946  * @returns the number of motion vectors parsed (2, 4 or 16)
947  */
948 static av_always_inline
950  int layout, int is_vp7)
951 {
952  int part_idx;
953  int n, num;
954  const VP8Macroblock *top_mb;
955  const VP8Macroblock *left_mb = &mb[-1];
956  const uint8_t *mbsplits_left = vp8_mbsplits[left_mb->partitioning];
957  const uint8_t *mbsplits_top, *mbsplits_cur, *firstidx;
958  const VP8mv *top_mv;
959  const VP8mv *left_mv = left_mb->bmv;
960  const VP8mv *cur_mv = mb->bmv;
961 
962  if (!layout) // layout is inlined, s->mb_layout is not
963  top_mb = &mb[2];
964  else
965  top_mb = &mb[-s->mb_width - 1];
966  mbsplits_top = vp8_mbsplits[top_mb->partitioning];
967  top_mv = top_mb->bmv;
968 
972  else
973  part_idx = VP8_SPLITMVMODE_8x8;
974  } else {
975  part_idx = VP8_SPLITMVMODE_4x4;
976  }
977 
978  num = vp8_mbsplit_count[part_idx];
979  mbsplits_cur = vp8_mbsplits[part_idx],
980  firstidx = vp8_mbfirstidx[part_idx];
981  mb->partitioning = part_idx;
982 
983  for (n = 0; n < num; n++) {
984  int k = firstidx[n];
985  uint32_t left, above;
986  const uint8_t *submv_prob;
987 
988  if (!(k & 3))
989  left = AV_RN32A(&left_mv[mbsplits_left[k + 3]]);
990  else
991  left = AV_RN32A(&cur_mv[mbsplits_cur[k - 1]]);
992  if (k <= 3)
993  above = AV_RN32A(&top_mv[mbsplits_top[k + 12]]);
994  else
995  above = AV_RN32A(&cur_mv[mbsplits_cur[k - 4]]);
996 
997  submv_prob = get_submv_prob(left, above, is_vp7);
998 
999  if (vpx_rac_get_prob_branchy(c, submv_prob[0])) {
1000  if (vpx_rac_get_prob_branchy(c, submv_prob[1])) {
1001  if (vpx_rac_get_prob_branchy(c, submv_prob[2])) {
1002  mb->bmv[n].y = mb->mv.y +
1003  read_mv_component(c, s->prob->mvc[0], is_vp7);
1004  mb->bmv[n].x = mb->mv.x +
1005  read_mv_component(c, s->prob->mvc[1], is_vp7);
1006  } else {
1007  AV_ZERO32(&mb->bmv[n]);
1008  }
1009  } else {
1010  AV_WN32A(&mb->bmv[n], above);
1011  }
1012  } else {
1013  AV_WN32A(&mb->bmv[n], left);
1014  }
1015  }
1016 
1017  return num;
1018 }
1019 
1020 /**
1021  * The vp7 reference decoder uses a padding macroblock column (added to right
1022  * edge of the frame) to guard against illegal macroblock offsets. The
1023  * algorithm has bugs that permit offsets to straddle the padding column.
1024  * This function replicates those bugs.
1025  *
1026  * @param[out] edge_x macroblock x address
1027  * @param[out] edge_y macroblock y address
1028  *
1029  * @return macroblock offset legal (boolean)
1030  */
1031 static int vp7_calculate_mb_offset(int mb_x, int mb_y, int mb_width,
1032  int xoffset, int yoffset, int boundary,
1033  int *edge_x, int *edge_y)
1034 {
1035  int vwidth = mb_width + 1;
1036  int new = (mb_y + yoffset) * vwidth + mb_x + xoffset;
1037  if (new < boundary || new % vwidth == vwidth - 1)
1038  return 0;
1039  *edge_y = new / vwidth;
1040  *edge_x = new % vwidth;
1041  return 1;
1042 }
1043 
1044 static const VP8mv *get_bmv_ptr(const VP8Macroblock *mb, int subblock)
1045 {
1046  return &mb->bmv[mb->mode == VP8_MVMODE_SPLIT ? vp8_mbsplits[mb->partitioning][subblock] : 0];
1047 }
1048 
1049 static av_always_inline
1051  int mb_x, int mb_y, int layout)
1052 {
1053  enum { CNT_ZERO, CNT_NEAREST, CNT_NEAR };
1054  enum { VP8_EDGE_TOP, VP8_EDGE_LEFT, VP8_EDGE_TOPLEFT };
1055  int idx = CNT_ZERO;
1056  VP8mv near_mv[3];
1057  uint8_t cnt[3] = { 0 };
1058  VPXRangeCoder *c = &s->c;
1059  int i;
1060 
1061  AV_ZERO32(&near_mv[0]);
1062  AV_ZERO32(&near_mv[1]);
1063  AV_ZERO32(&near_mv[2]);
1064 
1065  for (i = 0; i < VP7_MV_PRED_COUNT; i++) {
1066  const VP7MVPred * pred = &vp7_mv_pred[i];
1067  int edge_x, edge_y;
1068 
1069  if (vp7_calculate_mb_offset(mb_x, mb_y, s->mb_width, pred->xoffset,
1070  pred->yoffset, !s->profile, &edge_x, &edge_y)) {
1071  const VP8Macroblock *edge = (s->mb_layout == 1)
1072  ? s->macroblocks_base + 1 + edge_x +
1073  (s->mb_width + 1) * (edge_y + 1)
1074  : s->macroblocks + edge_x +
1075  (s->mb_height - edge_y - 1) * 2;
1076  uint32_t mv = AV_RN32A(get_bmv_ptr(edge, vp7_mv_pred[i].subblock));
1077  if (mv) {
1078  if (AV_RN32A(&near_mv[CNT_NEAREST])) {
1079  if (mv == AV_RN32A(&near_mv[CNT_NEAREST])) {
1080  idx = CNT_NEAREST;
1081  } else if (AV_RN32A(&near_mv[CNT_NEAR])) {
1082  if (mv != AV_RN32A(&near_mv[CNT_NEAR]))
1083  continue;
1084  idx = CNT_NEAR;
1085  } else {
1086  AV_WN32A(&near_mv[CNT_NEAR], mv);
1087  idx = CNT_NEAR;
1088  }
1089  } else {
1090  AV_WN32A(&near_mv[CNT_NEAREST], mv);
1091  idx = CNT_NEAREST;
1092  }
1093  } else {
1094  idx = CNT_ZERO;
1095  }
1096  } else {
1097  idx = CNT_ZERO;
1098  }
1099  cnt[idx] += vp7_mv_pred[i].score;
1100  }
1101 
1102  mb->partitioning = VP8_SPLITMVMODE_NONE;
1103 
1104  if (vpx_rac_get_prob_branchy(c, vp7_mode_contexts[cnt[CNT_ZERO]][0])) {
1105  mb->mode = VP8_MVMODE_MV;
1106 
1107  if (vpx_rac_get_prob_branchy(c, vp7_mode_contexts[cnt[CNT_NEAREST]][1])) {
1108 
1109  if (vpx_rac_get_prob_branchy(c, vp7_mode_contexts[cnt[CNT_NEAR]][2])) {
1110 
1111  if (cnt[CNT_NEAREST] > cnt[CNT_NEAR])
1112  AV_WN32A(&mb->mv, cnt[CNT_ZERO] > cnt[CNT_NEAREST] ? 0 : AV_RN32A(&near_mv[CNT_NEAREST]));
1113  else
1114  AV_WN32A(&mb->mv, cnt[CNT_ZERO] > cnt[CNT_NEAR] ? 0 : AV_RN32A(&near_mv[CNT_NEAR]));
1115 
1116  if (vpx_rac_get_prob_branchy(c, vp7_mode_contexts[cnt[CNT_NEAR]][3])) {
1117  mb->mode = VP8_MVMODE_SPLIT;
1118  mb->mv = mb->bmv[decode_splitmvs(s, c, mb, layout, IS_VP7) - 1];
1119  } else {
1120  mb->mv.y += vp7_read_mv_component(c, s->prob->mvc[0]);
1121  mb->mv.x += vp7_read_mv_component(c, s->prob->mvc[1]);
1122  mb->bmv[0] = mb->mv;
1123  }
1124  } else {
1125  mb->mv = near_mv[CNT_NEAR];
1126  mb->bmv[0] = mb->mv;
1127  }
1128  } else {
1129  mb->mv = near_mv[CNT_NEAREST];
1130  mb->bmv[0] = mb->mv;
1131  }
1132  } else {
1133  mb->mode = VP8_MVMODE_ZERO;
1134  AV_ZERO32(&mb->mv);
1135  mb->bmv[0] = mb->mv;
1136  }
1137 }
1138 
1139 static av_always_inline
1141  int mb_x, int mb_y, int layout)
1142 {
1143  VP8Macroblock *mb_edge[3] = { 0 /* top */,
1144  mb - 1 /* left */,
1145  0 /* top-left */ };
1146  enum { CNT_ZERO, CNT_NEAREST, CNT_NEAR, CNT_SPLITMV };
1147  enum { VP8_EDGE_TOP, VP8_EDGE_LEFT, VP8_EDGE_TOPLEFT };
1148  int idx = CNT_ZERO;
1149  int cur_sign_bias = s->sign_bias[mb->ref_frame];
1150  const int8_t *sign_bias = s->sign_bias;
1151  VP8mv near_mv[4];
1152  uint8_t cnt[4] = { 0 };
1153  VPXRangeCoder *c = &s->c;
1154 
1155  if (!layout) { // layout is inlined (s->mb_layout is not)
1156  mb_edge[0] = mb + 2;
1157  mb_edge[2] = mb + 1;
1158  } else {
1159  mb_edge[0] = mb - s->mb_width - 1;
1160  mb_edge[2] = mb - s->mb_width - 2;
1161  }
1162 
1163  AV_ZERO32(&near_mv[0]);
1164  AV_ZERO32(&near_mv[1]);
1165  AV_ZERO32(&near_mv[2]);
1166 
1167  /* Process MB on top, left and top-left */
1168 #define MV_EDGE_CHECK(n) \
1169  { \
1170  const VP8Macroblock *edge = mb_edge[n]; \
1171  int edge_ref = edge->ref_frame; \
1172  if (edge_ref != VP8_FRAME_CURRENT) { \
1173  uint32_t mv = AV_RN32A(&edge->mv); \
1174  if (mv) { \
1175  if (cur_sign_bias != sign_bias[edge_ref]) { \
1176  /* SWAR negate of the values in mv. */ \
1177  mv = ~mv; \
1178  mv = ((mv & 0x7fff7fff) + \
1179  0x00010001) ^ (mv & 0x80008000); \
1180  } \
1181  if (!n || mv != AV_RN32A(&near_mv[idx])) \
1182  AV_WN32A(&near_mv[++idx], mv); \
1183  cnt[idx] += 1 + (n != 2); \
1184  } else \
1185  cnt[CNT_ZERO] += 1 + (n != 2); \
1186  } \
1187  }
1188 
1189  MV_EDGE_CHECK(0)
1190  MV_EDGE_CHECK(1)
1191  MV_EDGE_CHECK(2)
1192 
1193  mb->partitioning = VP8_SPLITMVMODE_NONE;
1194  if (vpx_rac_get_prob_branchy(c, vp8_mode_contexts[cnt[CNT_ZERO]][0])) {
1195  mb->mode = VP8_MVMODE_MV;
1196 
1197  /* If we have three distinct MVs, merge first and last if they're the same */
1198  if (cnt[CNT_SPLITMV] &&
1199  AV_RN32A(&near_mv[1 + VP8_EDGE_TOP]) == AV_RN32A(&near_mv[1 + VP8_EDGE_TOPLEFT]))
1200  cnt[CNT_NEAREST] += 1;
1201 
1202  /* Swap near and nearest if necessary */
1203  if (cnt[CNT_NEAR] > cnt[CNT_NEAREST]) {
1204  FFSWAP(uint8_t, cnt[CNT_NEAREST], cnt[CNT_NEAR]);
1205  FFSWAP(VP8mv, near_mv[CNT_NEAREST], near_mv[CNT_NEAR]);
1206  }
1207 
1208  if (vpx_rac_get_prob_branchy(c, vp8_mode_contexts[cnt[CNT_NEAREST]][1])) {
1209  if (vpx_rac_get_prob_branchy(c, vp8_mode_contexts[cnt[CNT_NEAR]][2])) {
1210  /* Choose the best mv out of 0,0 and the nearest mv */
1211  clamp_mv(mv_bounds, &mb->mv, &near_mv[CNT_ZERO + (cnt[CNT_NEAREST] >= cnt[CNT_ZERO])]);
1212  cnt[CNT_SPLITMV] = ((mb_edge[VP8_EDGE_LEFT]->mode == VP8_MVMODE_SPLIT) +
1213  (mb_edge[VP8_EDGE_TOP]->mode == VP8_MVMODE_SPLIT)) * 2 +
1214  (mb_edge[VP8_EDGE_TOPLEFT]->mode == VP8_MVMODE_SPLIT);
1215 
1216  if (vpx_rac_get_prob_branchy(c, vp8_mode_contexts[cnt[CNT_SPLITMV]][3])) {
1217  mb->mode = VP8_MVMODE_SPLIT;
1218  mb->mv = mb->bmv[decode_splitmvs(s, c, mb, layout, IS_VP8) - 1];
1219  } else {
1220  mb->mv.y += vp8_read_mv_component(c, s->prob->mvc[0]);
1221  mb->mv.x += vp8_read_mv_component(c, s->prob->mvc[1]);
1222  mb->bmv[0] = mb->mv;
1223  }
1224  } else {
1225  clamp_mv(mv_bounds, &mb->mv, &near_mv[CNT_NEAR]);
1226  mb->bmv[0] = mb->mv;
1227  }
1228  } else {
1229  clamp_mv(mv_bounds, &mb->mv, &near_mv[CNT_NEAREST]);
1230  mb->bmv[0] = mb->mv;
1231  }
1232  } else {
1233  mb->mode = VP8_MVMODE_ZERO;
1234  AV_ZERO32(&mb->mv);
1235  mb->bmv[0] = mb->mv;
1236  }
1237 }
1238 
1239 static av_always_inline
1241  int mb_x, int keyframe, int layout)
1242 {
1243  uint8_t *intra4x4 = mb->intra4x4_pred_mode_mb;
1244 
1245  if (layout) {
1246  VP8Macroblock *mb_top = mb - s->mb_width - 1;
1247  memcpy(mb->intra4x4_pred_mode_top, mb_top->intra4x4_pred_mode_top, 4);
1248  }
1249  if (keyframe) {
1250  int x, y;
1251  uint8_t *top;
1252  uint8_t *const left = s->intra4x4_pred_mode_left;
1253  if (layout)
1254  top = mb->intra4x4_pred_mode_top;
1255  else
1256  top = s->intra4x4_pred_mode_top + 4 * mb_x;
1257  for (y = 0; y < 4; y++) {
1258  for (x = 0; x < 4; x++) {
1259  const uint8_t *ctx;
1260  ctx = vp8_pred4x4_prob_intra[top[x]][left[y]];
1261  *intra4x4 = vp89_rac_get_tree(c, vp8_pred4x4_tree, ctx);
1262  left[y] = top[x] = *intra4x4;
1263  intra4x4++;
1264  }
1265  }
1266  } else {
1267  int i;
1268  for (i = 0; i < 16; i++)
1269  intra4x4[i] = vp89_rac_get_tree(c, vp8_pred4x4_tree,
1271  }
1272 }
1273 
1274 static av_always_inline
1275 void decode_mb_mode(VP8Context *s, const VP8mvbounds *mv_bounds,
1276  VP8Macroblock *mb, int mb_x, int mb_y,
1277  uint8_t *segment, const uint8_t *ref, int layout, int is_vp7)
1278 {
1279  VPXRangeCoder *c = &s->c;
1280  static const char * const vp7_feature_name[] = { "q-index",
1281  "lf-delta",
1282  "partial-golden-update",
1283  "blit-pitch" };
1284  if (is_vp7) {
1285  int i;
1286  *segment = 0;
1287  for (i = 0; i < 4; i++) {
1288  if (s->feature_enabled[i]) {
1289  if (vpx_rac_get_prob_branchy(c, s->feature_present_prob[i])) {
1291  s->feature_index_prob[i]);
1292  av_log(s->avctx, AV_LOG_WARNING,
1293  "Feature %s present in macroblock (value 0x%x)\n",
1294  vp7_feature_name[i], s->feature_value[i][index]);
1295  }
1296  }
1297  }
1298  } else if (s->segmentation.update_map) {
1299  int bit = vpx_rac_get_prob(c, s->prob->segmentid[0]);
1300  *segment = vpx_rac_get_prob(c, s->prob->segmentid[1+bit]) + 2*bit;
1301  } else if (s->segmentation.enabled)
1302  *segment = ref ? *ref : *segment;
1303  mb->segment = *segment;
1304 
1305  mb->skip = s->mbskip_enabled ? vpx_rac_get_prob(c, s->prob->mbskip) : 0;
1306 
1307  if (s->keyframe) {
1310 
1311  if (mb->mode == MODE_I4x4) {
1312  decode_intra4x4_modes(s, c, mb, mb_x, 1, layout);
1313  } else {
1314  const uint32_t modes = (is_vp7 ? vp7_pred4x4_mode
1315  : vp8_pred4x4_mode)[mb->mode] * 0x01010101u;
1316  if (s->mb_layout)
1317  AV_WN32A(mb->intra4x4_pred_mode_top, modes);
1318  else
1319  AV_WN32A(s->intra4x4_pred_mode_top + 4 * mb_x, modes);
1320  AV_WN32A(s->intra4x4_pred_mode_left, modes);
1321  }
1322 
1323  mb->chroma_pred_mode = vp89_rac_get_tree(c, vp8_pred8x8c_tree,
1325  mb->ref_frame = VP8_FRAME_CURRENT;
1326  } else if (vpx_rac_get_prob_branchy(c, s->prob->intra)) {
1327  // inter MB, 16.2
1328  if (vpx_rac_get_prob_branchy(c, s->prob->last))
1329  mb->ref_frame =
1330  (!is_vp7 && vpx_rac_get_prob(c, s->prob->golden)) ? VP8_FRAME_ALTREF
1331  : VP8_FRAME_GOLDEN;
1332  else
1333  mb->ref_frame = VP8_FRAME_PREVIOUS;
1334  s->ref_count[mb->ref_frame - 1]++;
1335 
1336  // motion vectors, 16.3
1337  if (is_vp7)
1338  vp7_decode_mvs(s, mb, mb_x, mb_y, layout);
1339  else
1340  vp8_decode_mvs(s, mv_bounds, mb, mb_x, mb_y, layout);
1341  } else {
1342  // intra MB, 16.1
1344  s->prob->pred16x16);
1345 
1346  if (mb->mode == MODE_I4x4)
1347  decode_intra4x4_modes(s, c, mb, mb_x, 0, layout);
1348 
1349  mb->chroma_pred_mode = vp89_rac_get_tree(c, vp8_pred8x8c_tree,
1350  s->prob->pred8x8c);
1351  mb->ref_frame = VP8_FRAME_CURRENT;
1352  mb->partitioning = VP8_SPLITMVMODE_NONE;
1353  AV_ZERO32(&mb->bmv[0]);
1354  }
1355 }
1356 
1357 /**
1358  * @param r arithmetic bitstream reader context
1359  * @param block destination for block coefficients
1360  * @param probs probabilities to use when reading trees from the bitstream
1361  * @param i initial coeff index, 0 unless a separate DC block is coded
1362  * @param qmul array holding the dc/ac dequant factor at position 0/1
1363  *
1364  * @return 0 if no coeffs were decoded
1365  * otherwise, the index of the last coeff decoded plus one
1366  */
1367 static av_always_inline
1369  uint8_t probs[16][3][NUM_DCT_TOKENS - 1],
1370  int i, const uint8_t *token_prob, const int16_t qmul[2],
1371  const uint8_t scan[16], int vp7)
1372 {
1373  VPXRangeCoder c = *r;
1374  goto skip_eob;
1375  do {
1376  int coeff;
1377 restart:
1378  if (!vpx_rac_get_prob_branchy(&c, token_prob[0])) // DCT_EOB
1379  break;
1380 
1381 skip_eob:
1382  if (!vpx_rac_get_prob_branchy(&c, token_prob[1])) { // DCT_0
1383  if (++i == 16)
1384  break; // invalid input; blocks should end with EOB
1385  token_prob = probs[i][0];
1386  if (vp7)
1387  goto restart;
1388  goto skip_eob;
1389  }
1390 
1391  if (!vpx_rac_get_prob_branchy(&c, token_prob[2])) { // DCT_1
1392  coeff = 1;
1393  token_prob = probs[i + 1][1];
1394  } else {
1395  if (!vpx_rac_get_prob_branchy(&c, token_prob[3])) { // DCT 2,3,4
1396  coeff = vpx_rac_get_prob_branchy(&c, token_prob[4]);
1397  if (coeff)
1398  coeff += vpx_rac_get_prob(&c, token_prob[5]);
1399  coeff += 2;
1400  } else {
1401  // DCT_CAT*
1402  if (!vpx_rac_get_prob_branchy(&c, token_prob[6])) {
1403  if (!vpx_rac_get_prob_branchy(&c, token_prob[7])) { // DCT_CAT1
1405  } else { // DCT_CAT2
1406  coeff = 7;
1407  coeff += vpx_rac_get_prob(&c, vp8_dct_cat2_prob[0]) << 1;
1409  }
1410  } else { // DCT_CAT3 and up
1411  int a = vpx_rac_get_prob(&c, token_prob[8]);
1412  int b = vpx_rac_get_prob(&c, token_prob[9 + a]);
1413  int cat = (a << 1) + b;
1414  coeff = 3 + (8 << cat);
1416  }
1417  }
1418  token_prob = probs[i + 1][2];
1419  }
1420  block[scan[i]] = (vp89_rac_get(&c) ? -coeff : coeff) * qmul[!!i];
1421  } while (++i < 16);
1422 
1423  *r = c;
1424  return i;
1425 }
1426 
1427 static av_always_inline
1428 int inter_predict_dc(int16_t block[16], int16_t pred[2])
1429 {
1430  int16_t dc = block[0];
1431  int ret = 0;
1432 
1433  if (pred[1] > 3) {
1434  dc += pred[0];
1435  ret = 1;
1436  }
1437 
1438  if (!pred[0] | !dc | ((int32_t)pred[0] ^ (int32_t)dc) >> 31) {
1439  block[0] = pred[0] = dc;
1440  pred[1] = 0;
1441  } else {
1442  if (pred[0] == dc)
1443  pred[1]++;
1444  block[0] = pred[0] = dc;
1445  }
1446 
1447  return ret;
1448 }
1449 
1451  int16_t block[16],
1452  uint8_t probs[16][3][NUM_DCT_TOKENS - 1],
1453  int i, const uint8_t *token_prob,
1454  const int16_t qmul[2],
1455  const uint8_t scan[16])
1456 {
1457  return decode_block_coeffs_internal(r, block, probs, i,
1458  token_prob, qmul, scan, IS_VP7);
1459 }
1460 
1461 #ifndef vp8_decode_block_coeffs_internal
1463  int16_t block[16],
1464  uint8_t probs[16][3][NUM_DCT_TOKENS - 1],
1465  int i, const uint8_t *token_prob,
1466  const int16_t qmul[2])
1467 {
1468  return decode_block_coeffs_internal(r, block, probs, i,
1469  token_prob, qmul, ff_zigzag_scan, IS_VP8);
1470 }
1471 #endif
1472 
1473 /**
1474  * @param c arithmetic bitstream reader context
1475  * @param block destination for block coefficients
1476  * @param probs probabilities to use when reading trees from the bitstream
1477  * @param i initial coeff index, 0 unless a separate DC block is coded
1478  * @param zero_nhood the initial prediction context for number of surrounding
1479  * all-zero blocks (only left/top, so 0-2)
1480  * @param qmul array holding the dc/ac dequant factor at position 0/1
1481  * @param scan scan pattern (VP7 only)
1482  *
1483  * @return 0 if no coeffs were decoded
1484  * otherwise, the index of the last coeff decoded plus one
1485  */
1486 static av_always_inline
1488  uint8_t probs[16][3][NUM_DCT_TOKENS - 1],
1489  int i, int zero_nhood, const int16_t qmul[2],
1490  const uint8_t scan[16], int vp7)
1491 {
1492  const uint8_t *token_prob = probs[i][zero_nhood];
1493  if (!vpx_rac_get_prob_branchy(c, token_prob[0])) // DCT_EOB
1494  return 0;
1495  return vp7 ? vp7_decode_block_coeffs_internal(c, block, probs, i,
1496  token_prob, qmul, scan)
1498  token_prob, qmul);
1499 }
1500 
1501 static av_always_inline
1503  VP8Macroblock *mb, uint8_t t_nnz[9], uint8_t l_nnz[9],
1504  int is_vp7)
1505 {
1506  int i, x, y, luma_start = 0, luma_ctx = 3;
1507  int nnz_pred, nnz, nnz_total = 0;
1508  int segment = mb->segment;
1509  int block_dc = 0;
1510 
1511  if (mb->mode != MODE_I4x4 && (is_vp7 || mb->mode != VP8_MVMODE_SPLIT)) {
1512  nnz_pred = t_nnz[8] + l_nnz[8];
1513 
1514  // decode DC values and do hadamard
1515  nnz = decode_block_coeffs(c, td->block_dc, s->prob->token[1], 0,
1516  nnz_pred, s->qmat[segment].luma_dc_qmul,
1517  ff_zigzag_scan, is_vp7);
1518  l_nnz[8] = t_nnz[8] = !!nnz;
1519 
1520  if (is_vp7 && mb->mode > MODE_I4x4) {
1521  nnz |= inter_predict_dc(td->block_dc,
1522  s->inter_dc_pred[mb->ref_frame - 1]);
1523  }
1524 
1525  if (nnz) {
1526  nnz_total += nnz;
1527  block_dc = 1;
1528  if (nnz == 1)
1529  s->vp8dsp.vp8_luma_dc_wht_dc(td->block, td->block_dc);
1530  else
1531  s->vp8dsp.vp8_luma_dc_wht(td->block, td->block_dc);
1532  }
1533  luma_start = 1;
1534  luma_ctx = 0;
1535  }
1536 
1537  // luma blocks
1538  for (y = 0; y < 4; y++)
1539  for (x = 0; x < 4; x++) {
1540  nnz_pred = l_nnz[y] + t_nnz[x];
1541  nnz = decode_block_coeffs(c, td->block[y][x],
1542  s->prob->token[luma_ctx],
1543  luma_start, nnz_pred,
1544  s->qmat[segment].luma_qmul,
1545  s->prob[0].scan, is_vp7);
1546  /* nnz+block_dc may be one more than the actual last index,
1547  * but we don't care */
1548  td->non_zero_count_cache[y][x] = nnz + block_dc;
1549  t_nnz[x] = l_nnz[y] = !!nnz;
1550  nnz_total += nnz;
1551  }
1552 
1553  // chroma blocks
1554  // TODO: what to do about dimensions? 2nd dim for luma is x,
1555  // but for chroma it's (y<<1)|x
1556  for (i = 4; i < 6; i++)
1557  for (y = 0; y < 2; y++)
1558  for (x = 0; x < 2; x++) {
1559  nnz_pred = l_nnz[i + 2 * y] + t_nnz[i + 2 * x];
1560  nnz = decode_block_coeffs(c, td->block[i][(y << 1) + x],
1561  s->prob->token[2], 0, nnz_pred,
1562  s->qmat[segment].chroma_qmul,
1563  s->prob[0].scan, is_vp7);
1564  td->non_zero_count_cache[i][(y << 1) + x] = nnz;
1565  t_nnz[i + 2 * x] = l_nnz[i + 2 * y] = !!nnz;
1566  nnz_total += nnz;
1567  }
1568 
1569  // if there were no coded coeffs despite the macroblock not being marked skip,
1570  // we MUST not do the inner loop filter and should not do IDCT
1571  // Since skip isn't used for bitstream prediction, just manually set it.
1572  if (!nnz_total)
1573  mb->skip = 1;
1574 }
1575 
1576 static av_always_inline
1577 void backup_mb_border(uint8_t *top_border, const uint8_t *src_y,
1578  const uint8_t *src_cb, const uint8_t *src_cr,
1579  ptrdiff_t linesize, ptrdiff_t uvlinesize, int simple)
1580 {
1581  AV_COPY128(top_border, src_y + 15 * linesize);
1582  if (!simple) {
1583  AV_COPY64(top_border + 16, src_cb + 7 * uvlinesize);
1584  AV_COPY64(top_border + 24, src_cr + 7 * uvlinesize);
1585  }
1586 }
1587 
1588 static av_always_inline
1589 void xchg_mb_border(uint8_t *top_border, uint8_t *src_y, uint8_t *src_cb,
1590  uint8_t *src_cr, ptrdiff_t linesize, ptrdiff_t uvlinesize, int mb_x,
1591  int mb_y, int mb_width, int simple, int xchg)
1592 {
1593  uint8_t *top_border_m1 = top_border - 32; // for TL prediction
1594  src_y -= linesize;
1595  src_cb -= uvlinesize;
1596  src_cr -= uvlinesize;
1597 
1598 #define XCHG(a, b, xchg) \
1599  do { \
1600  if (xchg) \
1601  AV_SWAP64(b, a); \
1602  else \
1603  AV_COPY64(b, a); \
1604  } while (0)
1605 
1606  XCHG(top_border_m1 + 8, src_y - 8, xchg);
1607  XCHG(top_border, src_y, xchg);
1608  XCHG(top_border + 8, src_y + 8, 1);
1609  if (mb_x < mb_width - 1)
1610  XCHG(top_border + 32, src_y + 16, 1);
1611 
1612  // only copy chroma for normal loop filter
1613  // or to initialize the top row to 127
1614  if (!simple || !mb_y) {
1615  XCHG(top_border_m1 + 16, src_cb - 8, xchg);
1616  XCHG(top_border_m1 + 24, src_cr - 8, xchg);
1617  XCHG(top_border + 16, src_cb, 1);
1618  XCHG(top_border + 24, src_cr, 1);
1619  }
1620 }
1621 
1622 static av_always_inline
1623 int check_dc_pred8x8_mode(int mode, int mb_x, int mb_y)
1624 {
1625  if (!mb_x)
1626  return mb_y ? TOP_DC_PRED8x8 : DC_128_PRED8x8;
1627  else
1628  return mb_y ? mode : LEFT_DC_PRED8x8;
1629 }
1630 
1631 static av_always_inline
1632 int check_tm_pred8x8_mode(int mode, int mb_x, int mb_y, int vp7)
1633 {
1634  if (!mb_x)
1635  return mb_y ? VERT_PRED8x8 : (vp7 ? DC_128_PRED8x8 : DC_129_PRED8x8);
1636  else
1637  return mb_y ? mode : HOR_PRED8x8;
1638 }
1639 
1640 static av_always_inline
1641 int check_intra_pred8x8_mode_emuedge(int mode, int mb_x, int mb_y, int vp7)
1642 {
1643  switch (mode) {
1644  case DC_PRED8x8:
1645  return check_dc_pred8x8_mode(mode, mb_x, mb_y);
1646  case VERT_PRED8x8:
1647  return !mb_y ? (vp7 ? DC_128_PRED8x8 : DC_127_PRED8x8) : mode;
1648  case HOR_PRED8x8:
1649  return !mb_x ? (vp7 ? DC_128_PRED8x8 : DC_129_PRED8x8) : mode;
1650  case PLANE_PRED8x8: /* TM */
1651  return check_tm_pred8x8_mode(mode, mb_x, mb_y, vp7);
1652  }
1653  return mode;
1654 }
1655 
1656 static av_always_inline
1657 int check_tm_pred4x4_mode(int mode, int mb_x, int mb_y, int vp7)
1658 {
1659  if (!mb_x) {
1660  return mb_y ? VERT_VP8_PRED : (vp7 ? DC_128_PRED : DC_129_PRED);
1661  } else {
1662  return mb_y ? mode : HOR_VP8_PRED;
1663  }
1664 }
1665 
1666 static av_always_inline
1667 int check_intra_pred4x4_mode_emuedge(int mode, int mb_x, int mb_y,
1668  int *copy_buf, int vp7)
1669 {
1670  switch (mode) {
1671  case VERT_PRED:
1672  if (!mb_x && mb_y) {
1673  *copy_buf = 1;
1674  return mode;
1675  }
1676  /* fall-through */
1677  case DIAG_DOWN_LEFT_PRED:
1678  case VERT_LEFT_PRED:
1679  return !mb_y ? (vp7 ? DC_128_PRED : DC_127_PRED) : mode;
1680  case HOR_PRED:
1681  if (!mb_y) {
1682  *copy_buf = 1;
1683  return mode;
1684  }
1685  /* fall-through */
1686  case HOR_UP_PRED:
1687  return !mb_x ? (vp7 ? DC_128_PRED : DC_129_PRED) : mode;
1688  case TM_VP8_PRED:
1689  return check_tm_pred4x4_mode(mode, mb_x, mb_y, vp7);
1690  case DC_PRED: /* 4x4 DC doesn't use the same "H.264-style" exceptions
1691  * as 16x16/8x8 DC */
1692  case DIAG_DOWN_RIGHT_PRED:
1693  case VERT_RIGHT_PRED:
1694  case HOR_DOWN_PRED:
1695  if (!mb_y || !mb_x)
1696  *copy_buf = 1;
1697  return mode;
1698  }
1699  return mode;
1700 }
1701 
1702 static av_always_inline
1703 void intra_predict(VP8Context *s, VP8ThreadData *td, uint8_t *const dst[3],
1704  VP8Macroblock *mb, int mb_x, int mb_y, int is_vp7)
1705 {
1706  int x, y, mode, nnz;
1707  uint32_t tr;
1708 
1709  /* for the first row, we need to run xchg_mb_border to init the top edge
1710  * to 127 otherwise, skip it if we aren't going to deblock */
1711  if (mb_y && (s->deblock_filter || !mb_y) && td->thread_nr == 0)
1712  xchg_mb_border(s->top_border[mb_x + 1], dst[0], dst[1], dst[2],
1713  s->linesize, s->uvlinesize, mb_x, mb_y, s->mb_width,
1714  s->filter.simple, 1);
1715 
1716  if (mb->mode < MODE_I4x4) {
1717  mode = check_intra_pred8x8_mode_emuedge(mb->mode, mb_x, mb_y, is_vp7);
1718  s->hpc.pred16x16[mode](dst[0], s->linesize);
1719  } else {
1720  uint8_t *ptr = dst[0];
1721  const uint8_t *intra4x4 = mb->intra4x4_pred_mode_mb;
1722  const uint8_t lo = is_vp7 ? 128 : 127;
1723  const uint8_t hi = is_vp7 ? 128 : 129;
1724  const uint8_t tr_top[4] = { lo, lo, lo, lo };
1725 
1726  // all blocks on the right edge of the macroblock use bottom edge
1727  // the top macroblock for their topright edge
1728  const uint8_t *tr_right = ptr - s->linesize + 16;
1729 
1730  // if we're on the right edge of the frame, said edge is extended
1731  // from the top macroblock
1732  if (mb_y && mb_x == s->mb_width - 1) {
1733  tr = tr_right[-1] * 0x01010101u;
1734  tr_right = (uint8_t *) &tr;
1735  }
1736 
1737  if (mb->skip)
1738  AV_ZERO128(td->non_zero_count_cache);
1739 
1740  for (y = 0; y < 4; y++) {
1741  const uint8_t *topright = ptr + 4 - s->linesize;
1742  for (x = 0; x < 4; x++) {
1743  int copy = 0;
1744  ptrdiff_t linesize = s->linesize;
1745  uint8_t *dst = ptr + 4 * x;
1746  LOCAL_ALIGNED(4, uint8_t, copy_dst, [5 * 8]);
1747 
1748  if ((y == 0 || x == 3) && mb_y == 0) {
1749  topright = tr_top;
1750  } else if (x == 3)
1751  topright = tr_right;
1752 
1753  mode = check_intra_pred4x4_mode_emuedge(intra4x4[x], mb_x + x,
1754  mb_y + y, &copy, is_vp7);
1755  if (copy) {
1756  dst = copy_dst + 12;
1757  linesize = 8;
1758  if (!(mb_y + y)) {
1759  copy_dst[3] = lo;
1760  AV_WN32A(copy_dst + 4, lo * 0x01010101U);
1761  } else {
1762  AV_COPY32(copy_dst + 4, ptr + 4 * x - s->linesize);
1763  if (!(mb_x + x)) {
1764  copy_dst[3] = hi;
1765  } else {
1766  copy_dst[3] = ptr[4 * x - s->linesize - 1];
1767  }
1768  }
1769  if (!(mb_x + x)) {
1770  copy_dst[11] =
1771  copy_dst[19] =
1772  copy_dst[27] =
1773  copy_dst[35] = hi;
1774  } else {
1775  copy_dst[11] = ptr[4 * x - 1];
1776  copy_dst[19] = ptr[4 * x + s->linesize - 1];
1777  copy_dst[27] = ptr[4 * x + s->linesize * 2 - 1];
1778  copy_dst[35] = ptr[4 * x + s->linesize * 3 - 1];
1779  }
1780  }
1781  s->hpc.pred4x4[mode](dst, topright, linesize);
1782  if (copy) {
1783  AV_COPY32(ptr + 4 * x, copy_dst + 12);
1784  AV_COPY32(ptr + 4 * x + s->linesize, copy_dst + 20);
1785  AV_COPY32(ptr + 4 * x + s->linesize * 2, copy_dst + 28);
1786  AV_COPY32(ptr + 4 * x + s->linesize * 3, copy_dst + 36);
1787  }
1788 
1789  nnz = td->non_zero_count_cache[y][x];
1790  if (nnz) {
1791  if (nnz == 1)
1792  s->vp8dsp.vp8_idct_dc_add(ptr + 4 * x,
1793  td->block[y][x], s->linesize);
1794  else
1795  s->vp8dsp.vp8_idct_add(ptr + 4 * x,
1796  td->block[y][x], s->linesize);
1797  }
1798  topright += 4;
1799  }
1800 
1801  ptr += 4 * s->linesize;
1802  intra4x4 += 4;
1803  }
1804  }
1805 
1806  mode = check_intra_pred8x8_mode_emuedge(mb->chroma_pred_mode,
1807  mb_x, mb_y, is_vp7);
1808  s->hpc.pred8x8[mode](dst[1], s->uvlinesize);
1809  s->hpc.pred8x8[mode](dst[2], s->uvlinesize);
1810 
1811  if (mb_y && (s->deblock_filter || !mb_y) && td->thread_nr == 0)
1812  xchg_mb_border(s->top_border[mb_x + 1], dst[0], dst[1], dst[2],
1813  s->linesize, s->uvlinesize, mb_x, mb_y, s->mb_width,
1814  s->filter.simple, 0);
1815 }
1816 
1817 static const uint8_t subpel_idx[3][8] = {
1818  { 0, 1, 2, 1, 2, 1, 2, 1 }, // nr. of left extra pixels,
1819  // also function pointer index
1820  { 0, 3, 5, 3, 5, 3, 5, 3 }, // nr. of extra pixels required
1821  { 0, 2, 3, 2, 3, 2, 3, 2 }, // nr. of right extra pixels
1822 };
1823 
1824 /**
1825  * luma MC function
1826  *
1827  * @param s VP8 decoding context
1828  * @param dst target buffer for block data at block position
1829  * @param ref reference picture buffer at origin (0, 0)
1830  * @param mv motion vector (relative to block position) to get pixel data from
1831  * @param x_off horizontal position of block from origin (0, 0)
1832  * @param y_off vertical position of block from origin (0, 0)
1833  * @param block_w width of block (16, 8 or 4)
1834  * @param block_h height of block (always same as block_w)
1835  * @param width width of src/dst plane data
1836  * @param height height of src/dst plane data
1837  * @param linesize size of a single line of plane data, including padding
1838  * @param mc_func motion compensation function pointers (bilinear or sixtap MC)
1839  */
1840 static av_always_inline
1841 void vp8_mc_luma(VP8Context *s, VP8ThreadData *td, uint8_t *dst,
1842  const ThreadFrame *ref, const VP8mv *mv,
1843  int x_off, int y_off, int block_w, int block_h,
1844  int width, int height, ptrdiff_t linesize,
1845  vp8_mc_func mc_func[3][3])
1846 {
1847  const uint8_t *src = ref->f->data[0];
1848 
1849  if (AV_RN32A(mv)) {
1850  ptrdiff_t src_linesize = linesize;
1851 
1852  int mx = (mv->x * 2) & 7, mx_idx = subpel_idx[0][mx];
1853  int my = (mv->y * 2) & 7, my_idx = subpel_idx[0][my];
1854 
1855  x_off += mv->x >> 2;
1856  y_off += mv->y >> 2;
1857 
1858  // edge emulation
1859  ff_thread_await_progress(ref, (3 + y_off + block_h + subpel_idx[2][my]) >> 4, 0);
1860  src += y_off * linesize + x_off;
1861  if (x_off < mx_idx || x_off >= width - block_w - subpel_idx[2][mx] ||
1862  y_off < my_idx || y_off >= height - block_h - subpel_idx[2][my]) {
1863  s->vdsp.emulated_edge_mc(td->edge_emu_buffer,
1864  src - my_idx * linesize - mx_idx,
1865  EDGE_EMU_LINESIZE, linesize,
1866  block_w + subpel_idx[1][mx],
1867  block_h + subpel_idx[1][my],
1868  x_off - mx_idx, y_off - my_idx,
1869  width, height);
1870  src = td->edge_emu_buffer + mx_idx + EDGE_EMU_LINESIZE * my_idx;
1871  src_linesize = EDGE_EMU_LINESIZE;
1872  }
1873  mc_func[my_idx][mx_idx](dst, linesize, src, src_linesize, block_h, mx, my);
1874  } else {
1875  ff_thread_await_progress(ref, (3 + y_off + block_h) >> 4, 0);
1876  mc_func[0][0](dst, linesize, src + y_off * linesize + x_off,
1877  linesize, block_h, 0, 0);
1878  }
1879 }
1880 
1881 /**
1882  * chroma MC function
1883  *
1884  * @param s VP8 decoding context
1885  * @param dst1 target buffer for block data at block position (U plane)
1886  * @param dst2 target buffer for block data at block position (V plane)
1887  * @param ref reference picture buffer at origin (0, 0)
1888  * @param mv motion vector (relative to block position) to get pixel data from
1889  * @param x_off horizontal position of block from origin (0, 0)
1890  * @param y_off vertical position of block from origin (0, 0)
1891  * @param block_w width of block (16, 8 or 4)
1892  * @param block_h height of block (always same as block_w)
1893  * @param width width of src/dst plane data
1894  * @param height height of src/dst plane data
1895  * @param linesize size of a single line of plane data, including padding
1896  * @param mc_func motion compensation function pointers (bilinear or sixtap MC)
1897  */
1898 static av_always_inline
1899 void vp8_mc_chroma(VP8Context *s, VP8ThreadData *td, uint8_t *dst1,
1900  uint8_t *dst2, const ThreadFrame *ref, const VP8mv *mv,
1901  int x_off, int y_off, int block_w, int block_h,
1902  int width, int height, ptrdiff_t linesize,
1903  vp8_mc_func mc_func[3][3])
1904 {
1905  const uint8_t *src1 = ref->f->data[1], *src2 = ref->f->data[2];
1906 
1907  if (AV_RN32A(mv)) {
1908  int mx = mv->x & 7, mx_idx = subpel_idx[0][mx];
1909  int my = mv->y & 7, my_idx = subpel_idx[0][my];
1910 
1911  x_off += mv->x >> 3;
1912  y_off += mv->y >> 3;
1913 
1914  // edge emulation
1915  src1 += y_off * linesize + x_off;
1916  src2 += y_off * linesize + x_off;
1917  ff_thread_await_progress(ref, (3 + y_off + block_h + subpel_idx[2][my]) >> 3, 0);
1918  if (x_off < mx_idx || x_off >= width - block_w - subpel_idx[2][mx] ||
1919  y_off < my_idx || y_off >= height - block_h - subpel_idx[2][my]) {
1920  s->vdsp.emulated_edge_mc(td->edge_emu_buffer,
1921  src1 - my_idx * linesize - mx_idx,
1922  EDGE_EMU_LINESIZE, linesize,
1923  block_w + subpel_idx[1][mx],
1924  block_h + subpel_idx[1][my],
1925  x_off - mx_idx, y_off - my_idx, width, height);
1926  src1 = td->edge_emu_buffer + mx_idx + EDGE_EMU_LINESIZE * my_idx;
1927  mc_func[my_idx][mx_idx](dst1, linesize, src1, EDGE_EMU_LINESIZE, block_h, mx, my);
1928 
1929  s->vdsp.emulated_edge_mc(td->edge_emu_buffer,
1930  src2 - my_idx * linesize - mx_idx,
1931  EDGE_EMU_LINESIZE, linesize,
1932  block_w + subpel_idx[1][mx],
1933  block_h + subpel_idx[1][my],
1934  x_off - mx_idx, y_off - my_idx, width, height);
1935  src2 = td->edge_emu_buffer + mx_idx + EDGE_EMU_LINESIZE * my_idx;
1936  mc_func[my_idx][mx_idx](dst2, linesize, src2, EDGE_EMU_LINESIZE, block_h, mx, my);
1937  } else {
1938  mc_func[my_idx][mx_idx](dst1, linesize, src1, linesize, block_h, mx, my);
1939  mc_func[my_idx][mx_idx](dst2, linesize, src2, linesize, block_h, mx, my);
1940  }
1941  } else {
1942  ff_thread_await_progress(ref, (3 + y_off + block_h) >> 3, 0);
1943  mc_func[0][0](dst1, linesize, src1 + y_off * linesize + x_off, linesize, block_h, 0, 0);
1944  mc_func[0][0](dst2, linesize, src2 + y_off * linesize + x_off, linesize, block_h, 0, 0);
1945  }
1946 }
1947 
1948 static av_always_inline
1949 void vp8_mc_part(VP8Context *s, VP8ThreadData *td, uint8_t *const dst[3],
1950  const ThreadFrame *ref_frame, int x_off, int y_off,
1951  int bx_off, int by_off, int block_w, int block_h,
1952  int width, int height, const VP8mv *mv)
1953 {
1954  VP8mv uvmv = *mv;
1955 
1956  /* Y */
1957  vp8_mc_luma(s, td, dst[0] + by_off * s->linesize + bx_off,
1958  ref_frame, mv, x_off + bx_off, y_off + by_off,
1959  block_w, block_h, width, height, s->linesize,
1960  s->put_pixels_tab[block_w == 8]);
1961 
1962  /* U/V */
1963  if (s->profile == 3) {
1964  /* this block only applies VP8; it is safe to check
1965  * only the profile, as VP7 profile <= 1 */
1966  uvmv.x &= ~7;
1967  uvmv.y &= ~7;
1968  }
1969  x_off >>= 1;
1970  y_off >>= 1;
1971  bx_off >>= 1;
1972  by_off >>= 1;
1973  width >>= 1;
1974  height >>= 1;
1975  block_w >>= 1;
1976  block_h >>= 1;
1977  vp8_mc_chroma(s, td, dst[1] + by_off * s->uvlinesize + bx_off,
1978  dst[2] + by_off * s->uvlinesize + bx_off, ref_frame,
1979  &uvmv, x_off + bx_off, y_off + by_off,
1980  block_w, block_h, width, height, s->uvlinesize,
1981  s->put_pixels_tab[1 + (block_w == 4)]);
1982 }
1983 
1984 /* Fetch pixels for estimated mv 4 macroblocks ahead.
1985  * Optimized for 64-byte cache lines. Inspired by ffh264 prefetch_motion. */
1986 static av_always_inline
1988  int mb_x, int mb_y, int mb_xy, int ref)
1989 {
1990  /* Don't prefetch refs that haven't been used very often this frame. */
1991  if (s->ref_count[ref - 1] > (mb_xy >> 5)) {
1992  int x_off = mb_x << 4, y_off = mb_y << 4;
1993  int mx = (mb->mv.x >> 2) + x_off + 8;
1994  int my = (mb->mv.y >> 2) + y_off;
1995  uint8_t **src = s->framep[ref]->tf.f->data;
1996  int off = mx + (my + (mb_x & 3) * 4) * s->linesize + 64;
1997  /* For threading, a ff_thread_await_progress here might be useful, but
1998  * it actually slows down the decoder. Since a bad prefetch doesn't
1999  * generate bad decoder output, we don't run it here. */
2000  s->vdsp.prefetch(src[0] + off, s->linesize, 4);
2001  off = (mx >> 1) + ((my >> 1) + (mb_x & 7)) * s->uvlinesize + 64;
2002  s->vdsp.prefetch(src[1] + off, src[2] - src[1], 2);
2003  }
2004 }
2005 
2006 /**
2007  * Apply motion vectors to prediction buffer, chapter 18.
2008  */
2009 static av_always_inline
2010 void inter_predict(VP8Context *s, VP8ThreadData *td, uint8_t *const dst[3],
2011  VP8Macroblock *mb, int mb_x, int mb_y)
2012 {
2013  int x_off = mb_x << 4, y_off = mb_y << 4;
2014  int width = 16 * s->mb_width, height = 16 * s->mb_height;
2015  const ThreadFrame *ref = &s->framep[mb->ref_frame]->tf;
2016  const VP8mv *bmv = mb->bmv;
2017 
2018  switch (mb->partitioning) {
2019  case VP8_SPLITMVMODE_NONE:
2020  vp8_mc_part(s, td, dst, ref, x_off, y_off,
2021  0, 0, 16, 16, width, height, &mb->mv);
2022  break;
2023  case VP8_SPLITMVMODE_4x4: {
2024  int x, y;
2025  VP8mv uvmv;
2026 
2027  /* Y */
2028  for (y = 0; y < 4; y++) {
2029  for (x = 0; x < 4; x++) {
2030  vp8_mc_luma(s, td, dst[0] + 4 * y * s->linesize + x * 4,
2031  ref, &bmv[4 * y + x],
2032  4 * x + x_off, 4 * y + y_off, 4, 4,
2033  width, height, s->linesize,
2034  s->put_pixels_tab[2]);
2035  }
2036  }
2037 
2038  /* U/V */
2039  x_off >>= 1;
2040  y_off >>= 1;
2041  width >>= 1;
2042  height >>= 1;
2043  for (y = 0; y < 2; y++) {
2044  for (x = 0; x < 2; x++) {
2045  uvmv.x = mb->bmv[2 * y * 4 + 2 * x ].x +
2046  mb->bmv[2 * y * 4 + 2 * x + 1].x +
2047  mb->bmv[(2 * y + 1) * 4 + 2 * x ].x +
2048  mb->bmv[(2 * y + 1) * 4 + 2 * x + 1].x;
2049  uvmv.y = mb->bmv[2 * y * 4 + 2 * x ].y +
2050  mb->bmv[2 * y * 4 + 2 * x + 1].y +
2051  mb->bmv[(2 * y + 1) * 4 + 2 * x ].y +
2052  mb->bmv[(2 * y + 1) * 4 + 2 * x + 1].y;
2053  uvmv.x = (uvmv.x + 2 + FF_SIGNBIT(uvmv.x)) >> 2;
2054  uvmv.y = (uvmv.y + 2 + FF_SIGNBIT(uvmv.y)) >> 2;
2055  if (s->profile == 3) {
2056  uvmv.x &= ~7;
2057  uvmv.y &= ~7;
2058  }
2059  vp8_mc_chroma(s, td, dst[1] + 4 * y * s->uvlinesize + x * 4,
2060  dst[2] + 4 * y * s->uvlinesize + x * 4, ref,
2061  &uvmv, 4 * x + x_off, 4 * y + y_off, 4, 4,
2062  width, height, s->uvlinesize,
2063  s->put_pixels_tab[2]);
2064  }
2065  }
2066  break;
2067  }
2068  case VP8_SPLITMVMODE_16x8:
2069  vp8_mc_part(s, td, dst, ref, x_off, y_off,
2070  0, 0, 16, 8, width, height, &bmv[0]);
2071  vp8_mc_part(s, td, dst, ref, x_off, y_off,
2072  0, 8, 16, 8, width, height, &bmv[1]);
2073  break;
2074  case VP8_SPLITMVMODE_8x16:
2075  vp8_mc_part(s, td, dst, ref, x_off, y_off,
2076  0, 0, 8, 16, width, height, &bmv[0]);
2077  vp8_mc_part(s, td, dst, ref, x_off, y_off,
2078  8, 0, 8, 16, width, height, &bmv[1]);
2079  break;
2080  case VP8_SPLITMVMODE_8x8:
2081  vp8_mc_part(s, td, dst, ref, x_off, y_off,
2082  0, 0, 8, 8, width, height, &bmv[0]);
2083  vp8_mc_part(s, td, dst, ref, x_off, y_off,
2084  8, 0, 8, 8, width, height, &bmv[1]);
2085  vp8_mc_part(s, td, dst, ref, x_off, y_off,
2086  0, 8, 8, 8, width, height, &bmv[2]);
2087  vp8_mc_part(s, td, dst, ref, x_off, y_off,
2088  8, 8, 8, 8, width, height, &bmv[3]);
2089  break;
2090  }
2091 }
2092 
2093 static av_always_inline
2094 void idct_mb(VP8Context *s, VP8ThreadData *td, uint8_t *const dst[3],
2095  const VP8Macroblock *mb)
2096 {
2097  int x, y, ch;
2098 
2099  if (mb->mode != MODE_I4x4) {
2100  uint8_t *y_dst = dst[0];
2101  for (y = 0; y < 4; y++) {
2102  uint32_t nnz4 = AV_RL32(td->non_zero_count_cache[y]);
2103  if (nnz4) {
2104  if (nnz4 & ~0x01010101) {
2105  for (x = 0; x < 4; x++) {
2106  if ((uint8_t) nnz4 == 1)
2107  s->vp8dsp.vp8_idct_dc_add(y_dst + 4 * x,
2108  td->block[y][x],
2109  s->linesize);
2110  else if ((uint8_t) nnz4 > 1)
2111  s->vp8dsp.vp8_idct_add(y_dst + 4 * x,
2112  td->block[y][x],
2113  s->linesize);
2114  nnz4 >>= 8;
2115  if (!nnz4)
2116  break;
2117  }
2118  } else {
2119  s->vp8dsp.vp8_idct_dc_add4y(y_dst, td->block[y], s->linesize);
2120  }
2121  }
2122  y_dst += 4 * s->linesize;
2123  }
2124  }
2125 
2126  for (ch = 0; ch < 2; ch++) {
2127  uint32_t nnz4 = AV_RL32(td->non_zero_count_cache[4 + ch]);
2128  if (nnz4) {
2129  uint8_t *ch_dst = dst[1 + ch];
2130  if (nnz4 & ~0x01010101) {
2131  for (y = 0; y < 2; y++) {
2132  for (x = 0; x < 2; x++) {
2133  if ((uint8_t) nnz4 == 1)
2134  s->vp8dsp.vp8_idct_dc_add(ch_dst + 4 * x,
2135  td->block[4 + ch][(y << 1) + x],
2136  s->uvlinesize);
2137  else if ((uint8_t) nnz4 > 1)
2138  s->vp8dsp.vp8_idct_add(ch_dst + 4 * x,
2139  td->block[4 + ch][(y << 1) + x],
2140  s->uvlinesize);
2141  nnz4 >>= 8;
2142  if (!nnz4)
2143  goto chroma_idct_end;
2144  }
2145  ch_dst += 4 * s->uvlinesize;
2146  }
2147  } else {
2148  s->vp8dsp.vp8_idct_dc_add4uv(ch_dst, td->block[4 + ch], s->uvlinesize);
2149  }
2150  }
2151 chroma_idct_end:
2152  ;
2153  }
2154 }
2155 
2156 static av_always_inline
2158  VP8FilterStrength *f, int is_vp7)
2159 {
2160  int interior_limit, filter_level;
2161 
2162  if (s->segmentation.enabled) {
2163  filter_level = s->segmentation.filter_level[mb->segment];
2164  if (!s->segmentation.absolute_vals)
2165  filter_level += s->filter.level;
2166  } else
2167  filter_level = s->filter.level;
2168 
2169  if (s->lf_delta.enabled) {
2170  filter_level += s->lf_delta.ref[mb->ref_frame];
2171  filter_level += s->lf_delta.mode[mb->mode];
2172  }
2173 
2174  filter_level = av_clip_uintp2(filter_level, 6);
2175 
2176  interior_limit = filter_level;
2177  if (s->filter.sharpness) {
2178  interior_limit >>= (s->filter.sharpness + 3) >> 2;
2179  interior_limit = FFMIN(interior_limit, 9 - s->filter.sharpness);
2180  }
2181  interior_limit = FFMAX(interior_limit, 1);
2182 
2183  f->filter_level = filter_level;
2184  f->inner_limit = interior_limit;
2185  f->inner_filter = is_vp7 || !mb->skip || mb->mode == MODE_I4x4 ||
2186  mb->mode == VP8_MVMODE_SPLIT;
2187 }
2188 
2189 static av_always_inline
2190 void filter_mb(const VP8Context *s, uint8_t *const dst[3], const VP8FilterStrength *f,
2191  int mb_x, int mb_y, int is_vp7)
2192 {
2193  int mbedge_lim, bedge_lim_y, bedge_lim_uv, hev_thresh;
2194  int filter_level = f->filter_level;
2195  int inner_limit = f->inner_limit;
2196  int inner_filter = f->inner_filter;
2197  ptrdiff_t linesize = s->linesize;
2198  ptrdiff_t uvlinesize = s->uvlinesize;
2199  static const uint8_t hev_thresh_lut[2][64] = {
2200  { 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1,
2201  2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2,
2202  3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3,
2203  3, 3, 3, 3 },
2204  { 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1,
2205  1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1,
2206  2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2,
2207  2, 2, 2, 2 }
2208  };
2209 
2210  if (!filter_level)
2211  return;
2212 
2213  if (is_vp7) {
2214  bedge_lim_y = filter_level;
2215  bedge_lim_uv = filter_level * 2;
2216  mbedge_lim = filter_level + 2;
2217  } else {
2218  bedge_lim_y =
2219  bedge_lim_uv = filter_level * 2 + inner_limit;
2220  mbedge_lim = bedge_lim_y + 4;
2221  }
2222 
2223  hev_thresh = hev_thresh_lut[s->keyframe][filter_level];
2224 
2225  if (mb_x) {
2226  s->vp8dsp.vp8_h_loop_filter16y(dst[0], linesize,
2227  mbedge_lim, inner_limit, hev_thresh);
2228  s->vp8dsp.vp8_h_loop_filter8uv(dst[1], dst[2], uvlinesize,
2229  mbedge_lim, inner_limit, hev_thresh);
2230  }
2231 
2232 #define H_LOOP_FILTER_16Y_INNER(cond) \
2233  if (cond && inner_filter) { \
2234  s->vp8dsp.vp8_h_loop_filter16y_inner(dst[0] + 4, linesize, \
2235  bedge_lim_y, inner_limit, \
2236  hev_thresh); \
2237  s->vp8dsp.vp8_h_loop_filter16y_inner(dst[0] + 8, linesize, \
2238  bedge_lim_y, inner_limit, \
2239  hev_thresh); \
2240  s->vp8dsp.vp8_h_loop_filter16y_inner(dst[0] + 12, linesize, \
2241  bedge_lim_y, inner_limit, \
2242  hev_thresh); \
2243  s->vp8dsp.vp8_h_loop_filter8uv_inner(dst[1] + 4, dst[2] + 4, \
2244  uvlinesize, bedge_lim_uv, \
2245  inner_limit, hev_thresh); \
2246  }
2247 
2248  H_LOOP_FILTER_16Y_INNER(!is_vp7)
2249 
2250  if (mb_y) {
2251  s->vp8dsp.vp8_v_loop_filter16y(dst[0], linesize,
2252  mbedge_lim, inner_limit, hev_thresh);
2253  s->vp8dsp.vp8_v_loop_filter8uv(dst[1], dst[2], uvlinesize,
2254  mbedge_lim, inner_limit, hev_thresh);
2255  }
2256 
2257  if (inner_filter) {
2258  s->vp8dsp.vp8_v_loop_filter16y_inner(dst[0] + 4 * linesize,
2259  linesize, bedge_lim_y,
2260  inner_limit, hev_thresh);
2261  s->vp8dsp.vp8_v_loop_filter16y_inner(dst[0] + 8 * linesize,
2262  linesize, bedge_lim_y,
2263  inner_limit, hev_thresh);
2264  s->vp8dsp.vp8_v_loop_filter16y_inner(dst[0] + 12 * linesize,
2265  linesize, bedge_lim_y,
2266  inner_limit, hev_thresh);
2267  s->vp8dsp.vp8_v_loop_filter8uv_inner(dst[1] + 4 * uvlinesize,
2268  dst[2] + 4 * uvlinesize,
2269  uvlinesize, bedge_lim_uv,
2270  inner_limit, hev_thresh);
2271  }
2272 
2273  H_LOOP_FILTER_16Y_INNER(is_vp7)
2274 }
2275 
2276 static av_always_inline
2277 void filter_mb_simple(const VP8Context *s, uint8_t *dst, const VP8FilterStrength *f,
2278  int mb_x, int mb_y)
2279 {
2280  int mbedge_lim, bedge_lim;
2281  int filter_level = f->filter_level;
2282  int inner_limit = f->inner_limit;
2283  int inner_filter = f->inner_filter;
2284  ptrdiff_t linesize = s->linesize;
2285 
2286  if (!filter_level)
2287  return;
2288 
2289  bedge_lim = 2 * filter_level + inner_limit;
2290  mbedge_lim = bedge_lim + 4;
2291 
2292  if (mb_x)
2293  s->vp8dsp.vp8_h_loop_filter_simple(dst, linesize, mbedge_lim);
2294  if (inner_filter) {
2295  s->vp8dsp.vp8_h_loop_filter_simple(dst + 4, linesize, bedge_lim);
2296  s->vp8dsp.vp8_h_loop_filter_simple(dst + 8, linesize, bedge_lim);
2297  s->vp8dsp.vp8_h_loop_filter_simple(dst + 12, linesize, bedge_lim);
2298  }
2299 
2300  if (mb_y)
2301  s->vp8dsp.vp8_v_loop_filter_simple(dst, linesize, mbedge_lim);
2302  if (inner_filter) {
2303  s->vp8dsp.vp8_v_loop_filter_simple(dst + 4 * linesize, linesize, bedge_lim);
2304  s->vp8dsp.vp8_v_loop_filter_simple(dst + 8 * linesize, linesize, bedge_lim);
2305  s->vp8dsp.vp8_v_loop_filter_simple(dst + 12 * linesize, linesize, bedge_lim);
2306  }
2307 }
2308 
2309 #define MARGIN (16 << 2)
2310 static av_always_inline
2312  const VP8Frame *prev_frame, int is_vp7)
2313 {
2314  VP8Context *s = avctx->priv_data;
2315  int mb_x, mb_y;
2316 
2317  s->mv_bounds.mv_min.y = -MARGIN;
2318  s->mv_bounds.mv_max.y = ((s->mb_height - 1) << 6) + MARGIN;
2319  for (mb_y = 0; mb_y < s->mb_height; mb_y++) {
2320  VP8Macroblock *mb = s->macroblocks_base +
2321  ((s->mb_width + 1) * (mb_y + 1) + 1);
2322  int mb_xy = mb_y * s->mb_width;
2323 
2324  AV_WN32A(s->intra4x4_pred_mode_left, DC_PRED * 0x01010101);
2325 
2326  s->mv_bounds.mv_min.x = -MARGIN;
2327  s->mv_bounds.mv_max.x = ((s->mb_width - 1) << 6) + MARGIN;
2328 
2329  for (mb_x = 0; mb_x < s->mb_width; mb_x++, mb_xy++, mb++) {
2330  if (vpx_rac_is_end(&s->c)) {
2331  return AVERROR_INVALIDDATA;
2332  }
2333  if (mb_y == 0)
2334  AV_WN32A((mb - s->mb_width - 1)->intra4x4_pred_mode_top,
2335  DC_PRED * 0x01010101);
2336  decode_mb_mode(s, &s->mv_bounds, mb, mb_x, mb_y, curframe->seg_map + mb_xy,
2337  prev_frame && prev_frame->seg_map ?
2338  prev_frame->seg_map + mb_xy : NULL, 1, is_vp7);
2339  s->mv_bounds.mv_min.x -= 64;
2340  s->mv_bounds.mv_max.x -= 64;
2341  }
2342  s->mv_bounds.mv_min.y -= 64;
2343  s->mv_bounds.mv_max.y -= 64;
2344  }
2345  return 0;
2346 }
2347 
2348 static int vp7_decode_mv_mb_modes(AVCodecContext *avctx, VP8Frame *cur_frame,
2349  const VP8Frame *prev_frame)
2350 {
2351  return vp78_decode_mv_mb_modes(avctx, cur_frame, prev_frame, IS_VP7);
2352 }
2353 
2354 static int vp8_decode_mv_mb_modes(AVCodecContext *avctx, VP8Frame *cur_frame,
2355  const VP8Frame *prev_frame)
2356 {
2357  return vp78_decode_mv_mb_modes(avctx, cur_frame, prev_frame, IS_VP8);
2358 }
2359 
2360 #if HAVE_THREADS
2361 #define check_thread_pos(td, otd, mb_x_check, mb_y_check) \
2362  do { \
2363  int tmp = (mb_y_check << 16) | (mb_x_check & 0xFFFF); \
2364  if (atomic_load(&otd->thread_mb_pos) < tmp) { \
2365  pthread_mutex_lock(&otd->lock); \
2366  atomic_store(&td->wait_mb_pos, tmp); \
2367  do { \
2368  if (atomic_load(&otd->thread_mb_pos) >= tmp) \
2369  break; \
2370  pthread_cond_wait(&otd->cond, &otd->lock); \
2371  } while (1); \
2372  atomic_store(&td->wait_mb_pos, INT_MAX); \
2373  pthread_mutex_unlock(&otd->lock); \
2374  } \
2375  } while (0)
2376 
2377 #define update_pos(td, mb_y, mb_x) \
2378  do { \
2379  int pos = (mb_y << 16) | (mb_x & 0xFFFF); \
2380  int sliced_threading = (avctx->active_thread_type == FF_THREAD_SLICE) && \
2381  (num_jobs > 1); \
2382  int is_null = !next_td || !prev_td; \
2383  int pos_check = (is_null) ? 1 : \
2384  (next_td != td && pos >= atomic_load(&next_td->wait_mb_pos)) || \
2385  (prev_td != td && pos >= atomic_load(&prev_td->wait_mb_pos)); \
2386  atomic_store(&td->thread_mb_pos, pos); \
2387  if (sliced_threading && pos_check) { \
2388  pthread_mutex_lock(&td->lock); \
2389  pthread_cond_broadcast(&td->cond); \
2390  pthread_mutex_unlock(&td->lock); \
2391  } \
2392  } while (0)
2393 #else
2394 #define check_thread_pos(td, otd, mb_x_check, mb_y_check) while(0)
2395 #define update_pos(td, mb_y, mb_x) while(0)
2396 #endif
2397 
2399  int jobnr, int threadnr, int is_vp7)
2400 {
2401  VP8Context *s = avctx->priv_data;
2402  VP8ThreadData *prev_td, *next_td, *td = &s->thread_data[threadnr];
2403  int mb_y = atomic_load(&td->thread_mb_pos) >> 16;
2404  int mb_x, mb_xy = mb_y * s->mb_width;
2405  int num_jobs = s->num_jobs;
2406  const VP8Frame *prev_frame = s->prev_frame;
2407  VP8Frame *curframe = s->curframe;
2408  VPXRangeCoder *coeff_c = &s->coeff_partition[mb_y & (s->num_coeff_partitions - 1)];
2409 
2410  VP8Macroblock *mb;
2411  uint8_t *dst[3] = {
2412  curframe->tf.f->data[0] + 16 * mb_y * s->linesize,
2413  curframe->tf.f->data[1] + 8 * mb_y * s->uvlinesize,
2414  curframe->tf.f->data[2] + 8 * mb_y * s->uvlinesize
2415  };
2416 
2417  if (vpx_rac_is_end(&s->c))
2418  return AVERROR_INVALIDDATA;
2419 
2420  if (mb_y == 0)
2421  prev_td = td;
2422  else
2423  prev_td = &s->thread_data[(jobnr + num_jobs - 1) % num_jobs];
2424  if (mb_y == s->mb_height - 1)
2425  next_td = td;
2426  else
2427  next_td = &s->thread_data[(jobnr + 1) % num_jobs];
2428  if (s->mb_layout == 1)
2429  mb = s->macroblocks_base + ((s->mb_width + 1) * (mb_y + 1) + 1);
2430  else {
2431  // Make sure the previous frame has read its segmentation map,
2432  // if we re-use the same map.
2433  if (prev_frame && s->segmentation.enabled &&
2434  !s->segmentation.update_map)
2435  ff_thread_await_progress(&prev_frame->tf, mb_y, 0);
2436  mb = s->macroblocks + (s->mb_height - mb_y - 1) * 2;
2437  memset(mb - 1, 0, sizeof(*mb)); // zero left macroblock
2438  AV_WN32A(s->intra4x4_pred_mode_left, DC_PRED * 0x01010101);
2439  }
2440 
2441  if (!is_vp7 || mb_y == 0)
2442  memset(td->left_nnz, 0, sizeof(td->left_nnz));
2443 
2444  td->mv_bounds.mv_min.x = -MARGIN;
2445  td->mv_bounds.mv_max.x = ((s->mb_width - 1) << 6) + MARGIN;
2446 
2447  for (mb_x = 0; mb_x < s->mb_width; mb_x++, mb_xy++, mb++) {
2448  if (vpx_rac_is_end(&s->c))
2449  return AVERROR_INVALIDDATA;
2450  // Wait for previous thread to read mb_x+2, and reach mb_y-1.
2451  if (prev_td != td) {
2452  if (threadnr != 0) {
2453  check_thread_pos(td, prev_td,
2454  mb_x + (is_vp7 ? 2 : 1),
2455  mb_y - (is_vp7 ? 2 : 1));
2456  } else {
2457  check_thread_pos(td, prev_td,
2458  mb_x + (is_vp7 ? 2 : 1) + s->mb_width + 3,
2459  mb_y - (is_vp7 ? 2 : 1));
2460  }
2461  }
2462 
2463  s->vdsp.prefetch(dst[0] + (mb_x & 3) * 4 * s->linesize + 64,
2464  s->linesize, 4);
2465  s->vdsp.prefetch(dst[1] + (mb_x & 7) * s->uvlinesize + 64,
2466  dst[2] - dst[1], 2);
2467 
2468  if (!s->mb_layout)
2469  decode_mb_mode(s, &td->mv_bounds, mb, mb_x, mb_y, curframe->seg_map + mb_xy,
2470  prev_frame && prev_frame->seg_map ?
2471  prev_frame->seg_map + mb_xy : NULL, 0, is_vp7);
2472 
2473  prefetch_motion(s, mb, mb_x, mb_y, mb_xy, VP8_FRAME_PREVIOUS);
2474 
2475  if (!mb->skip) {
2476  if (vpx_rac_is_end(coeff_c))
2477  return AVERROR_INVALIDDATA;
2478  decode_mb_coeffs(s, td, coeff_c, mb, s->top_nnz[mb_x], td->left_nnz, is_vp7);
2479  }
2480 
2481  if (mb->mode <= MODE_I4x4)
2482  intra_predict(s, td, dst, mb, mb_x, mb_y, is_vp7);
2483  else
2484  inter_predict(s, td, dst, mb, mb_x, mb_y);
2485 
2486  prefetch_motion(s, mb, mb_x, mb_y, mb_xy, VP8_FRAME_GOLDEN);
2487 
2488  if (!mb->skip) {
2489  idct_mb(s, td, dst, mb);
2490  } else {
2491  AV_ZERO64(td->left_nnz);
2492  AV_WN64(s->top_nnz[mb_x], 0); // array of 9, so unaligned
2493 
2494  /* Reset DC block predictors if they would exist
2495  * if the mb had coefficients */
2496  if (mb->mode != MODE_I4x4 && mb->mode != VP8_MVMODE_SPLIT) {
2497  td->left_nnz[8] = 0;
2498  s->top_nnz[mb_x][8] = 0;
2499  }
2500  }
2501 
2502  if (s->deblock_filter)
2503  filter_level_for_mb(s, mb, &td->filter_strength[mb_x], is_vp7);
2504 
2505  if (s->deblock_filter && num_jobs != 1 && threadnr == num_jobs - 1) {
2506  if (s->filter.simple)
2507  backup_mb_border(s->top_border[mb_x + 1], dst[0],
2508  NULL, NULL, s->linesize, 0, 1);
2509  else
2510  backup_mb_border(s->top_border[mb_x + 1], dst[0],
2511  dst[1], dst[2], s->linesize, s->uvlinesize, 0);
2512  }
2513 
2514  prefetch_motion(s, mb, mb_x, mb_y, mb_xy, VP8_FRAME_ALTREF);
2515 
2516  dst[0] += 16;
2517  dst[1] += 8;
2518  dst[2] += 8;
2519  td->mv_bounds.mv_min.x -= 64;
2520  td->mv_bounds.mv_max.x -= 64;
2521 
2522  if (mb_x == s->mb_width + 1) {
2523  update_pos(td, mb_y, s->mb_width + 3);
2524  } else {
2525  update_pos(td, mb_y, mb_x);
2526  }
2527  }
2528  return 0;
2529 }
2530 
2531 static int vp7_decode_mb_row_no_filter(AVCodecContext *avctx, void *tdata,
2532  int jobnr, int threadnr)
2533 {
2534  return decode_mb_row_no_filter(avctx, tdata, jobnr, threadnr, 1);
2535 }
2536 
2537 static int vp8_decode_mb_row_no_filter(AVCodecContext *avctx, void *tdata,
2538  int jobnr, int threadnr)
2539 {
2540  return decode_mb_row_no_filter(avctx, tdata, jobnr, threadnr, 0);
2541 }
2542 
2543 static av_always_inline void filter_mb_row(AVCodecContext *avctx, void *tdata,
2544  int jobnr, int threadnr, int is_vp7)
2545 {
2546  VP8Context *s = avctx->priv_data;
2547  VP8ThreadData *td = &s->thread_data[threadnr];
2548  int mb_x, mb_y = atomic_load(&td->thread_mb_pos) >> 16, num_jobs = s->num_jobs;
2549  AVFrame *curframe = s->curframe->tf.f;
2550  VP8Macroblock *mb;
2551  VP8ThreadData *prev_td, *next_td;
2552  uint8_t *dst[3] = {
2553  curframe->data[0] + 16 * mb_y * s->linesize,
2554  curframe->data[1] + 8 * mb_y * s->uvlinesize,
2555  curframe->data[2] + 8 * mb_y * s->uvlinesize
2556  };
2557 
2558  if (s->mb_layout == 1)
2559  mb = s->macroblocks_base + ((s->mb_width + 1) * (mb_y + 1) + 1);
2560  else
2561  mb = s->macroblocks + (s->mb_height - mb_y - 1) * 2;
2562 
2563  if (mb_y == 0)
2564  prev_td = td;
2565  else
2566  prev_td = &s->thread_data[(jobnr + num_jobs - 1) % num_jobs];
2567  if (mb_y == s->mb_height - 1)
2568  next_td = td;
2569  else
2570  next_td = &s->thread_data[(jobnr + 1) % num_jobs];
2571 
2572  for (mb_x = 0; mb_x < s->mb_width; mb_x++, mb++) {
2573  const VP8FilterStrength *f = &td->filter_strength[mb_x];
2574  if (prev_td != td)
2575  check_thread_pos(td, prev_td,
2576  (mb_x + 1) + (s->mb_width + 3), mb_y - 1);
2577  if (next_td != td)
2578  if (next_td != &s->thread_data[0])
2579  check_thread_pos(td, next_td, mb_x + 1, mb_y + 1);
2580 
2581  if (num_jobs == 1) {
2582  if (s->filter.simple)
2583  backup_mb_border(s->top_border[mb_x + 1], dst[0],
2584  NULL, NULL, s->linesize, 0, 1);
2585  else
2586  backup_mb_border(s->top_border[mb_x + 1], dst[0],
2587  dst[1], dst[2], s->linesize, s->uvlinesize, 0);
2588  }
2589 
2590  if (s->filter.simple)
2591  filter_mb_simple(s, dst[0], f, mb_x, mb_y);
2592  else
2593  filter_mb(s, dst, f, mb_x, mb_y, is_vp7);
2594  dst[0] += 16;
2595  dst[1] += 8;
2596  dst[2] += 8;
2597 
2598  update_pos(td, mb_y, (s->mb_width + 3) + mb_x);
2599  }
2600 }
2601 
2602 static void vp7_filter_mb_row(AVCodecContext *avctx, void *tdata,
2603  int jobnr, int threadnr)
2604 {
2605  filter_mb_row(avctx, tdata, jobnr, threadnr, 1);
2606 }
2607 
2608 static void vp8_filter_mb_row(AVCodecContext *avctx, void *tdata,
2609  int jobnr, int threadnr)
2610 {
2611  filter_mb_row(avctx, tdata, jobnr, threadnr, 0);
2612 }
2613 
2614 static av_always_inline
2615 int vp78_decode_mb_row_sliced(AVCodecContext *avctx, void *tdata, int jobnr,
2616  int threadnr, int is_vp7)
2617 {
2618  const VP8Context *s = avctx->priv_data;
2619  VP8ThreadData *td = &s->thread_data[jobnr];
2620  VP8ThreadData *next_td = NULL, *prev_td = NULL;
2621  VP8Frame *curframe = s->curframe;
2622  int mb_y, num_jobs = s->num_jobs;
2623  int ret;
2624 
2625  td->thread_nr = threadnr;
2626  td->mv_bounds.mv_min.y = -MARGIN - 64 * threadnr;
2627  td->mv_bounds.mv_max.y = ((s->mb_height - 1) << 6) + MARGIN - 64 * threadnr;
2628  for (mb_y = jobnr; mb_y < s->mb_height; mb_y += num_jobs) {
2629  atomic_store(&td->thread_mb_pos, mb_y << 16);
2630  ret = s->decode_mb_row_no_filter(avctx, tdata, jobnr, threadnr);
2631  if (ret < 0) {
2632  update_pos(td, s->mb_height, INT_MAX & 0xFFFF);
2633  return ret;
2634  }
2635  if (s->deblock_filter)
2636  s->filter_mb_row(avctx, tdata, jobnr, threadnr);
2637  update_pos(td, mb_y, INT_MAX & 0xFFFF);
2638 
2639  td->mv_bounds.mv_min.y -= 64 * num_jobs;
2640  td->mv_bounds.mv_max.y -= 64 * num_jobs;
2641 
2642  if (avctx->active_thread_type == FF_THREAD_FRAME)
2643  ff_thread_report_progress(&curframe->tf, mb_y, 0);
2644  }
2645 
2646  return 0;
2647 }
2648 
2649 static int vp7_decode_mb_row_sliced(AVCodecContext *avctx, void *tdata,
2650  int jobnr, int threadnr)
2651 {
2652  return vp78_decode_mb_row_sliced(avctx, tdata, jobnr, threadnr, IS_VP7);
2653 }
2654 
2655 static int vp8_decode_mb_row_sliced(AVCodecContext *avctx, void *tdata,
2656  int jobnr, int threadnr)
2657 {
2658  return vp78_decode_mb_row_sliced(avctx, tdata, jobnr, threadnr, IS_VP8);
2659 }
2660 
2661 static av_always_inline
2662 int vp78_decode_frame(AVCodecContext *avctx, AVFrame *rframe, int *got_frame,
2663  const AVPacket *avpkt, int is_vp7)
2664 {
2665  VP8Context *s = avctx->priv_data;
2666  int ret, i, referenced, num_jobs;
2667  enum AVDiscard skip_thresh;
2668  VP8Frame *av_uninit(curframe), *prev_frame;
2669 
2670  if (is_vp7)
2671  ret = vp7_decode_frame_header(s, avpkt->data, avpkt->size);
2672  else
2673  ret = vp8_decode_frame_header(s, avpkt->data, avpkt->size);
2674 
2675  if (ret < 0)
2676  goto err;
2677 
2678  if (!is_vp7 && s->actually_webp) {
2679  // VP8 in WebP is supposed to be intra-only. Enforce this here
2680  // to ensure that output is reproducible with frame-threading.
2681  if (!s->keyframe)
2682  return AVERROR_INVALIDDATA;
2683  // avctx->pix_fmt already set in caller.
2684  } else if (!is_vp7 && s->pix_fmt == AV_PIX_FMT_NONE) {
2685  s->pix_fmt = get_pixel_format(s);
2686  if (s->pix_fmt < 0) {
2687  ret = AVERROR(EINVAL);
2688  goto err;
2689  }
2690  avctx->pix_fmt = s->pix_fmt;
2691  }
2692 
2693  prev_frame = s->framep[VP8_FRAME_CURRENT];
2694 
2695  referenced = s->update_last || s->update_golden == VP8_FRAME_CURRENT ||
2696  s->update_altref == VP8_FRAME_CURRENT;
2697 
2698  skip_thresh = !referenced ? AVDISCARD_NONREF
2699  : !s->keyframe ? AVDISCARD_NONKEY
2700  : AVDISCARD_ALL;
2701 
2702  if (avctx->skip_frame >= skip_thresh) {
2703  s->invisible = 1;
2704  memcpy(&s->next_framep[0], &s->framep[0], sizeof(s->framep[0]) * 4);
2705  goto skip_decode;
2706  }
2707  s->deblock_filter = s->filter.level && avctx->skip_loop_filter < skip_thresh;
2708 
2709  // release no longer referenced frames
2710  for (i = 0; i < 5; i++)
2711  if (s->frames[i].tf.f->buf[0] &&
2712  &s->frames[i] != prev_frame &&
2713  &s->frames[i] != s->framep[VP8_FRAME_PREVIOUS] &&
2714  &s->frames[i] != s->framep[VP8_FRAME_GOLDEN] &&
2715  &s->frames[i] != s->framep[VP8_FRAME_ALTREF])
2716  vp8_release_frame(&s->frames[i]);
2717 
2718  curframe = s->framep[VP8_FRAME_CURRENT] = vp8_find_free_buffer(s);
2719 
2720  if (!s->colorspace)
2721  avctx->colorspace = AVCOL_SPC_BT470BG;
2722  if (s->fullrange)
2723  avctx->color_range = AVCOL_RANGE_JPEG;
2724  else
2725  avctx->color_range = AVCOL_RANGE_MPEG;
2726 
2727  /* Given that arithmetic probabilities are updated every frame, it's quite
2728  * likely that the values we have on a random interframe are complete
2729  * junk if we didn't start decode on a keyframe. So just don't display
2730  * anything rather than junk. */
2731  if (!s->keyframe && (!s->framep[VP8_FRAME_PREVIOUS] ||
2732  !s->framep[VP8_FRAME_GOLDEN] ||
2733  !s->framep[VP8_FRAME_ALTREF])) {
2734  av_log(avctx, AV_LOG_WARNING,
2735  "Discarding interframe without a prior keyframe!\n");
2737  goto err;
2738  }
2739 
2740  if (s->keyframe)
2741  curframe->tf.f->flags |= AV_FRAME_FLAG_KEY;
2742  else
2743  curframe->tf.f->flags &= ~AV_FRAME_FLAG_KEY;
2744  curframe->tf.f->pict_type = s->keyframe ? AV_PICTURE_TYPE_I
2746  if ((ret = vp8_alloc_frame(s, curframe, referenced)) < 0)
2747  goto err;
2748 
2749  // check if golden and altref are swapped
2750  if (s->update_altref != VP8_FRAME_NONE)
2751  s->next_framep[VP8_FRAME_ALTREF] = s->framep[s->update_altref];
2752  else
2753  s->next_framep[VP8_FRAME_ALTREF] = s->framep[VP8_FRAME_ALTREF];
2754 
2755  if (s->update_golden != VP8_FRAME_NONE)
2756  s->next_framep[VP8_FRAME_GOLDEN] = s->framep[s->update_golden];
2757  else
2758  s->next_framep[VP8_FRAME_GOLDEN] = s->framep[VP8_FRAME_GOLDEN];
2759 
2760  if (s->update_last)
2761  s->next_framep[VP8_FRAME_PREVIOUS] = curframe;
2762  else
2763  s->next_framep[VP8_FRAME_PREVIOUS] = s->framep[VP8_FRAME_PREVIOUS];
2764 
2765  s->next_framep[VP8_FRAME_CURRENT] = curframe;
2766 
2767  if (!is_vp7 && !s->actually_webp)
2768  ff_thread_finish_setup(avctx);
2769 
2770  if (avctx->hwaccel) {
2771  const FFHWAccel *hwaccel = ffhwaccel(avctx->hwaccel);
2772  ret = hwaccel->start_frame(avctx, avpkt->data, avpkt->size);
2773  if (ret < 0)
2774  goto err;
2775 
2776  ret = hwaccel->decode_slice(avctx, avpkt->data, avpkt->size);
2777  if (ret < 0)
2778  goto err;
2779 
2780  ret = hwaccel->end_frame(avctx);
2781  if (ret < 0)
2782  goto err;
2783 
2784  } else {
2785  s->linesize = curframe->tf.f->linesize[0];
2786  s->uvlinesize = curframe->tf.f->linesize[1];
2787 
2788  memset(s->top_nnz, 0, s->mb_width * sizeof(*s->top_nnz));
2789  /* Zero macroblock structures for top/top-left prediction
2790  * from outside the frame. */
2791  if (!s->mb_layout)
2792  memset(s->macroblocks + s->mb_height * 2 - 1, 0,
2793  (s->mb_width + 1) * sizeof(*s->macroblocks));
2794  if (!s->mb_layout && s->keyframe)
2795  memset(s->intra4x4_pred_mode_top, DC_PRED, s->mb_width * 4);
2796 
2797  memset(s->ref_count, 0, sizeof(s->ref_count));
2798 
2799  if (s->mb_layout == 1) {
2800  // Make sure the previous frame has read its segmentation map,
2801  // if we re-use the same map.
2802  if (prev_frame && s->segmentation.enabled &&
2803  !s->segmentation.update_map)
2804  ff_thread_await_progress(&prev_frame->tf, 1, 0);
2805  if (is_vp7)
2806  ret = vp7_decode_mv_mb_modes(avctx, curframe, prev_frame);
2807  else
2808  ret = vp8_decode_mv_mb_modes(avctx, curframe, prev_frame);
2809  if (ret < 0)
2810  goto err;
2811  }
2812 
2813  if (avctx->active_thread_type == FF_THREAD_FRAME)
2814  num_jobs = 1;
2815  else
2816  num_jobs = FFMIN(s->num_coeff_partitions, avctx->thread_count);
2817  s->num_jobs = num_jobs;
2818  s->curframe = curframe;
2819  s->prev_frame = prev_frame;
2820  s->mv_bounds.mv_min.y = -MARGIN;
2821  s->mv_bounds.mv_max.y = ((s->mb_height - 1) << 6) + MARGIN;
2822  for (i = 0; i < MAX_THREADS; i++) {
2823  VP8ThreadData *td = &s->thread_data[i];
2824  atomic_init(&td->thread_mb_pos, 0);
2825  atomic_init(&td->wait_mb_pos, INT_MAX);
2826  }
2827  if (is_vp7)
2828  avctx->execute2(avctx, vp7_decode_mb_row_sliced, s->thread_data, NULL,
2829  num_jobs);
2830  else
2831  avctx->execute2(avctx, vp8_decode_mb_row_sliced, s->thread_data, NULL,
2832  num_jobs);
2833  }
2834 
2835  ff_thread_report_progress(&curframe->tf, INT_MAX, 0);
2836  memcpy(&s->framep[0], &s->next_framep[0], sizeof(s->framep[0]) * 4);
2837 
2838 skip_decode:
2839  // if future frames don't use the updated probabilities,
2840  // reset them to the values we saved
2841  if (!s->update_probabilities)
2842  s->prob[0] = s->prob[1];
2843 
2844  if (!s->invisible) {
2845  if ((ret = av_frame_ref(rframe, curframe->tf.f)) < 0)
2846  return ret;
2847  *got_frame = 1;
2848  }
2849 
2850  return avpkt->size;
2851 err:
2852  memcpy(&s->next_framep[0], &s->framep[0], sizeof(s->framep[0]) * 4);
2853  return ret;
2854 }
2855 
2857  int *got_frame, AVPacket *avpkt)
2858 {
2859  return vp78_decode_frame(avctx, frame, got_frame, avpkt, IS_VP8);
2860 }
2861 
2862 #if CONFIG_VP7_DECODER
2863 static int vp7_decode_frame(AVCodecContext *avctx, AVFrame *frame,
2864  int *got_frame, AVPacket *avpkt)
2865 {
2866  return vp78_decode_frame(avctx, frame, got_frame, avpkt, IS_VP7);
2867 }
2868 #endif /* CONFIG_VP7_DECODER */
2869 
2871 {
2872  VP8Context *s = avctx->priv_data;
2873  int i;
2874 
2875  vp8_decode_flush_impl(avctx, 1);
2876  for (i = 0; i < FF_ARRAY_ELEMS(s->frames); i++)
2877  av_frame_free(&s->frames[i].tf.f);
2878 
2879  return 0;
2880 }
2881 
2883 {
2884  int i;
2885  for (i = 0; i < FF_ARRAY_ELEMS(s->frames); i++) {
2886  s->frames[i].tf.f = av_frame_alloc();
2887  if (!s->frames[i].tf.f)
2888  return AVERROR(ENOMEM);
2889  }
2890  return 0;
2891 }
2892 
2893 static av_always_inline
2894 int vp78_decode_init(AVCodecContext *avctx, int is_vp7)
2895 {
2896  VP8Context *s = avctx->priv_data;
2897  int ret;
2898 
2899  s->avctx = avctx;
2900  s->pix_fmt = AV_PIX_FMT_NONE;
2901  avctx->pix_fmt = AV_PIX_FMT_YUV420P;
2902 
2903  ff_videodsp_init(&s->vdsp, 8);
2904 
2905  ff_vp78dsp_init(&s->vp8dsp);
2906  if (CONFIG_VP7_DECODER && is_vp7) {
2907  ff_h264_pred_init(&s->hpc, AV_CODEC_ID_VP7, 8, 1);
2908  ff_vp7dsp_init(&s->vp8dsp);
2909  s->decode_mb_row_no_filter = vp7_decode_mb_row_no_filter;
2910  s->filter_mb_row = vp7_filter_mb_row;
2911  } else if (CONFIG_VP8_DECODER && !is_vp7) {
2912  ff_h264_pred_init(&s->hpc, AV_CODEC_ID_VP8, 8, 1);
2913  ff_vp8dsp_init(&s->vp8dsp);
2914  s->decode_mb_row_no_filter = vp8_decode_mb_row_no_filter;
2915  s->filter_mb_row = vp8_filter_mb_row;
2916  }
2917 
2918  /* does not change for VP8 */
2919  memcpy(s->prob[0].scan, ff_zigzag_scan, sizeof(s->prob[0].scan));
2920 
2921  if ((ret = vp8_init_frames(s)) < 0) {
2922  ff_vp8_decode_free(avctx);
2923  return ret;
2924  }
2925 
2926  return 0;
2927 }
2928 
2929 #if CONFIG_VP7_DECODER
2930 static int vp7_decode_init(AVCodecContext *avctx)
2931 {
2932  return vp78_decode_init(avctx, IS_VP7);
2933 }
2934 #endif /* CONFIG_VP7_DECODER */
2935 
2937 {
2938  return vp78_decode_init(avctx, IS_VP8);
2939 }
2940 
2941 #if CONFIG_VP8_DECODER
2942 #if HAVE_THREADS
2943 #define REBASE(pic) ((pic) ? (pic) - &s_src->frames[0] + &s->frames[0] : NULL)
2944 
2945 static int vp8_decode_update_thread_context(AVCodecContext *dst,
2946  const AVCodecContext *src)
2947 {
2948  VP8Context *s = dst->priv_data, *s_src = src->priv_data;
2949  int i;
2950 
2951  if (s->macroblocks_base &&
2952  (s_src->mb_width != s->mb_width || s_src->mb_height != s->mb_height)) {
2953  free_buffers(s);
2954  s->mb_width = s_src->mb_width;
2955  s->mb_height = s_src->mb_height;
2956  }
2957 
2958  s->pix_fmt = s_src->pix_fmt;
2959  s->prob[0] = s_src->prob[!s_src->update_probabilities];
2960  s->segmentation = s_src->segmentation;
2961  s->lf_delta = s_src->lf_delta;
2962  memcpy(s->sign_bias, s_src->sign_bias, sizeof(s->sign_bias));
2963 
2964  for (i = 0; i < FF_ARRAY_ELEMS(s_src->frames); i++) {
2965  if (s_src->frames[i].tf.f->buf[0]) {
2966  int ret = vp8_ref_frame(&s->frames[i], &s_src->frames[i]);
2967  if (ret < 0)
2968  return ret;
2969  }
2970  }
2971 
2972  s->framep[0] = REBASE(s_src->next_framep[0]);
2973  s->framep[1] = REBASE(s_src->next_framep[1]);
2974  s->framep[2] = REBASE(s_src->next_framep[2]);
2975  s->framep[3] = REBASE(s_src->next_framep[3]);
2976 
2977  return 0;
2978 }
2979 #endif /* HAVE_THREADS */
2980 #endif /* CONFIG_VP8_DECODER */
2981 
2982 #if CONFIG_VP7_DECODER
2983 const FFCodec ff_vp7_decoder = {
2984  .p.name = "vp7",
2985  CODEC_LONG_NAME("On2 VP7"),
2986  .p.type = AVMEDIA_TYPE_VIDEO,
2987  .p.id = AV_CODEC_ID_VP7,
2988  .priv_data_size = sizeof(VP8Context),
2989  .init = vp7_decode_init,
2990  .close = ff_vp8_decode_free,
2991  FF_CODEC_DECODE_CB(vp7_decode_frame),
2992  .p.capabilities = AV_CODEC_CAP_DR1,
2993  .flush = vp8_decode_flush,
2994 };
2995 #endif /* CONFIG_VP7_DECODER */
2996 
2997 #if CONFIG_VP8_DECODER
2998 const FFCodec ff_vp8_decoder = {
2999  .p.name = "vp8",
3000  CODEC_LONG_NAME("On2 VP8"),
3001  .p.type = AVMEDIA_TYPE_VIDEO,
3002  .p.id = AV_CODEC_ID_VP8,
3003  .priv_data_size = sizeof(VP8Context),
3005  .close = ff_vp8_decode_free,
3007  .p.capabilities = AV_CODEC_CAP_DR1 | AV_CODEC_CAP_FRAME_THREADS |
3009  .caps_internal = FF_CODEC_CAP_ALLOCATE_PROGRESS,
3010  .flush = vp8_decode_flush,
3011  UPDATE_THREAD_CONTEXT(vp8_decode_update_thread_context),
3012  .hw_configs = (const AVCodecHWConfigInternal *const []) {
3013 #if CONFIG_VP8_VAAPI_HWACCEL
3014  HWACCEL_VAAPI(vp8),
3015 #endif
3016 #if CONFIG_VP8_NVDEC_HWACCEL
3017  HWACCEL_NVDEC(vp8),
3018 #endif
3019  NULL
3020  },
3021 };
3022 #endif /* CONFIG_VP7_DECODER */
vp8_mode_contexts
static const int vp8_mode_contexts[6][4]
Definition: vp8data.h:118
hwconfig.h
vp8_dct_cat1_prob
static const uint8_t vp8_dct_cat1_prob[]
Definition: vp8data.h:336
decode_mb_mode
static av_always_inline void decode_mb_mode(VP8Context *s, const VP8mvbounds *mv_bounds, VP8Macroblock *mb, int mb_x, int mb_y, uint8_t *segment, const uint8_t *ref, int layout, int is_vp7)
Definition: vp8.c:1275
VP7_MV_PRED_COUNT
#define VP7_MV_PRED_COUNT
Definition: vp8data.h:68
AVCodecContext::hwaccel
const struct AVHWAccel * hwaccel
Hardware accelerator in use.
Definition: avcodec.h:1427
ff_vp8_decode_free
av_cold int ff_vp8_decode_free(AVCodecContext *avctx)
Definition: vp8.c:2870
vp7_pred4x4_mode
static const uint8_t vp7_pred4x4_mode[]
Definition: vp8data.h:33
HOR_PRED8x8
#define HOR_PRED8x8
Definition: h264pred.h:69
decode_block_coeffs_internal
static av_always_inline int decode_block_coeffs_internal(VPXRangeCoder *r, int16_t block[16], uint8_t probs[16][3][NUM_DCT_TOKENS - 1], int i, const uint8_t *token_prob, const int16_t qmul[2], const uint8_t scan[16], int vp7)
Definition: vp8.c:1368
vp8_release_frame
static void vp8_release_frame(VP8Frame *f)
Definition: vp8.c:126
vp7_mv_pred
static const VP7MVPred vp7_mv_pred[VP7_MV_PRED_COUNT]
Definition: vp8data.h:69
AV_LOG_WARNING
#define AV_LOG_WARNING
Something somehow does not look correct.
Definition: log.h:186
AV_PIX_FMT_CUDA
@ AV_PIX_FMT_CUDA
HW acceleration through CUDA.
Definition: pixfmt.h:260
td
#define td
Definition: regdef.h:70
AVPixelFormat
AVPixelFormat
Pixel format.
Definition: pixfmt.h:71
vp8_decode_block_coeffs_internal
static int vp8_decode_block_coeffs_internal(VPXRangeCoder *r, int16_t block[16], uint8_t probs[16][3][NUM_DCT_TOKENS - 1], int i, const uint8_t *token_prob, const int16_t qmul[2])
Definition: vp8.c:1462
vp7_read_mv_component
static int vp7_read_mv_component(VPXRangeCoder *c, const uint8_t *p)
Definition: vp8.c:921
vp7_calculate_mb_offset
static int vp7_calculate_mb_offset(int mb_x, int mb_y, int mb_width, int xoffset, int yoffset, int boundary, int *edge_x, int *edge_y)
The vp7 reference decoder uses a padding macroblock column (added to right edge of the frame) to guar...
Definition: vp8.c:1031
av_clip
#define av_clip
Definition: common.h:98
atomic_store
#define atomic_store(object, desired)
Definition: stdatomic.h:85
backup_mb_border
static av_always_inline void backup_mb_border(uint8_t *top_border, const uint8_t *src_y, const uint8_t *src_cb, const uint8_t *src_cr, ptrdiff_t linesize, ptrdiff_t uvlinesize, int simple)
Definition: vp8.c:1577
VP8Macroblock::partitioning
uint8_t partitioning
Definition: vp8.h:101
VP8_FRAME_CURRENT
@ VP8_FRAME_CURRENT
Definition: vp8.h:44
r
const char * r
Definition: vf_curves.c:126
vp7_filter_mb_row
static void vp7_filter_mb_row(AVCodecContext *avctx, void *tdata, int jobnr, int threadnr)
Definition: vp8.c:2602
AVERROR
Filter the word “frame” indicates either a video frame or a group of audio as stored in an AVFrame structure Format for each input and each output the list of supported formats For video that means pixel format For audio that means channel sample they are references to shared objects When the negotiation mechanism computes the intersection of the formats supported at each end of a all references to both lists are replaced with a reference to the intersection And when a single format is eventually chosen for a link amongst the remaining all references to the list are updated That means that if a filter requires that its input and output have the same format amongst a supported all it has to do is use a reference to the same list of formats query_formats can leave some formats unset and return AVERROR(EAGAIN) to cause the negotiation mechanism toagain later. That can be used by filters with complex requirements to use the format negotiated on one link to set the formats supported on another. Frame references ownership and permissions
DC_PRED8x8
#define DC_PRED8x8
Definition: h264pred.h:68
IS_VP7
#define IS_VP7
Definition: vp8dsp.h:100
AVCodecContext::colorspace
enum AVColorSpace colorspace
YUV colorspace type.
Definition: avcodec.h:685
vp78_decode_init
static av_always_inline int vp78_decode_init(AVCodecContext *avctx, int is_vp7)
Definition: vp8.c:2894
mem_internal.h
DC_128_PRED
@ DC_128_PRED
Definition: vp9.h:58
ff_get_format
int ff_get_format(AVCodecContext *avctx, const enum AVPixelFormat *fmt)
Select the (possibly hardware accelerated) pixel format.
Definition: decode.c:1220
u
#define u(width, name, range_min, range_max)
Definition: cbs_h2645.c:250
check_tm_pred8x8_mode
static av_always_inline int check_tm_pred8x8_mode(int mode, int mb_x, int mb_y, int vp7)
Definition: vp8.c:1632
vp8_submv_prob
static const uint8_t vp8_submv_prob[5][3]
Definition: vp8data.h:153
av_clip_uintp2
#define av_clip_uintp2
Definition: common.h:122
pthread_mutex_init
static av_always_inline int pthread_mutex_init(pthread_mutex_t *mutex, const pthread_mutexattr_t *attr)
Definition: os2threads.h:104
vp7_ydc_qlookup
static const uint16_t vp7_ydc_qlookup[]
Definition: vp8data.h:585
src1
const pixel * src1
Definition: h264pred_template.c:421
HOR_VP8_PRED
#define HOR_VP8_PRED
unaveraged version of HOR_PRED, see
Definition: h264pred.h:63
mv
static const int8_t mv[256][2]
Definition: 4xm.c:80
vp7_decode_mvs
static av_always_inline void vp7_decode_mvs(VP8Context *s, VP8Macroblock *mb, int mb_x, int mb_y, int layout)
Definition: vp8.c:1050
vp7_mv_default_prob
static const uint8_t vp7_mv_default_prob[2][17]
Definition: vp8data.h:551
check_intra_pred4x4_mode_emuedge
static av_always_inline int check_intra_pred4x4_mode_emuedge(int mode, int mb_x, int mb_y, int *copy_buf, int vp7)
Definition: vp8.c:1667
ff_vp8_token_update_probs
const uint8_t ff_vp8_token_update_probs[4][8][3][11]
Definition: vp8data.c:43
av_frame_free
void av_frame_free(AVFrame **frame)
Free the frame and any dynamically allocated objects in it, e.g.
Definition: frame.c:130
check_tm_pred4x4_mode
static av_always_inline int check_tm_pred4x4_mode(int mode, int mb_x, int mb_y, int vp7)
Definition: vp8.c:1657
vp7_y2dc_qlookup
static const uint16_t vp7_y2dc_qlookup[]
Definition: vp8data.h:610
AVFrame
This structure describes decoded (raw) audio or video data.
Definition: frame.h:344
TM_VP8_PRED
@ TM_VP8_PRED
Definition: vp9.h:55
AVCOL_RANGE_JPEG
@ AVCOL_RANGE_JPEG
Full range content.
Definition: pixfmt.h:683
vp8_mc_chroma
static av_always_inline void vp8_mc_chroma(VP8Context *s, VP8ThreadData *td, uint8_t *dst1, uint8_t *dst2, const ThreadFrame *ref, const VP8mv *mv, int x_off, int y_off, int block_w, int block_h, int width, int height, ptrdiff_t linesize, vp8_mc_func mc_func[3][3])
chroma MC function
Definition: vp8.c:1899
AVPacket::data
uint8_t * data
Definition: packet.h:522
inter_predict_dc
static av_always_inline int inter_predict_dc(int16_t block[16], int16_t pred[2])
Definition: vp8.c:1428
DC_PRED
@ DC_PRED
Definition: vp9.h:48
b
#define b
Definition: input.c:41
VP7_MVC_SIZE
#define VP7_MVC_SIZE
Definition: vp8.c:478
VERT_LEFT_PRED
@ VERT_LEFT_PRED
Definition: vp9.h:53
vp8_get_quants
static void vp8_get_quants(VP8Context *s)
Definition: vp8.c:391
FFCodec
Definition: codec_internal.h:127
AV_WN32A
#define AV_WN32A(p, v)
Definition: intreadwrite.h:536
FF_HW_SIMPLE_CALL
#define FF_HW_SIMPLE_CALL(avctx, function)
Definition: hwaccel_internal.h:174
cat
#define cat(a, bpp, b)
Definition: vp9dsp_init.h:32
FFMAX
#define FFMAX(a, b)
Definition: macros.h:47
VP8mvbounds
Definition: vp8.h:115
vp89_rac.h
inter_predict
static av_always_inline void inter_predict(VP8Context *s, VP8ThreadData *td, uint8_t *const dst[3], VP8Macroblock *mb, int mb_x, int mb_y)
Apply motion vectors to prediction buffer, chapter 18.
Definition: vp8.c:2010
VP8_SPLITMVMODE_4x4
@ VP8_SPLITMVMODE_4x4
4x4 blocks of 4x4px each
Definition: vp8.h:80
AVFrame::buf
AVBufferRef * buf[AV_NUM_DATA_POINTERS]
AVBuffer references backing the data for this frame.
Definition: frame.h:557
ff_set_dimensions
int ff_set_dimensions(AVCodecContext *s, int width, int height)
Check that the provided frame dimensions are valid and set them on the codec context.
Definition: utils.c:94
VP8_FRAME_ALTREF
@ VP8_FRAME_ALTREF
Definition: vp8.h:47
VERT_VP8_PRED
#define VERT_VP8_PRED
for VP8, VERT_PRED is the average of
Definition: h264pred.h:60
VPXRangeCoder
Definition: vpx_rac.h:35
thread.h
ff_thread_await_progress
the pkt_dts and pkt_pts fields in AVFrame will work as usual Restrictions on codec whose streams don t reset across will not work because their bitstreams cannot be decoded in parallel *The contents of buffers must not be read before ff_thread_await_progress() has been called on them. reget_buffer() and buffer age optimizations no longer work. *The contents of buffers must not be written to after ff_thread_report_progress() has been called on them. This includes draw_edges(). Porting codecs to frame threading
ThreadFrame::f
AVFrame * f
Definition: threadframe.h:28
AVFrame::data
uint8_t * data[AV_NUM_DATA_POINTERS]
pointer to the picture/channel planes.
Definition: frame.h:365
VP8_MVC_SIZE
#define VP8_MVC_SIZE
Definition: vp8.c:479
vp8_decode_mb_row_no_filter
static int vp8_decode_mb_row_no_filter(AVCodecContext *avctx, void *tdata, int jobnr, int threadnr)
Definition: vp8.c:2537
vp8_rac_get_sint
static int vp8_rac_get_sint(VPXRangeCoder *c, int bits)
Definition: vp8.c:50
vp8_pred8x8c_tree
static const int8_t vp8_pred8x8c_tree[3][2]
Definition: vp8data.h:180
XCHG
#define XCHG(a, b, xchg)
bit
#define bit(string, value)
Definition: cbs_mpeg2.c:56
update_pos
#define update_pos(td, mb_y, mb_x)
Definition: vp8.c:2395
vp8.h
get_bmv_ptr
static const VP8mv * get_bmv_ptr(const VP8Macroblock *mb, int subblock)
Definition: vp8.c:1044
update_dimensions
static av_always_inline int update_dimensions(VP8Context *s, int width, int height, int is_vp7)
Definition: vp8.c:212
AVCOL_SPC_BT470BG
@ AVCOL_SPC_BT470BG
also ITU-R BT601-6 625 / ITU-R BT1358 625 / ITU-R BT1700 625 PAL & SECAM / IEC 61966-2-4 xvYCC601
Definition: pixfmt.h:615
VP8_SPLITMVMODE_8x8
@ VP8_SPLITMVMODE_8x8
2x2 blocks of 8x8px each
Definition: vp8.h:79
FFCodec::p
AVCodec p
The public AVCodec.
Definition: codec_internal.h:131
FFHWAccel
Definition: hwaccel_internal.h:34
DC_127_PRED
@ DC_127_PRED
Definition: vp9.h:59
vp8_mv_update_prob
static const uint8_t vp8_mv_update_prob[2][19]
Definition: vp8data.h:540
AVCodecContext::skip_frame
enum AVDiscard skip_frame
Skip decoding for selected frames.
Definition: avcodec.h:1819
fail
#define fail()
Definition: checkasm.h:179
VERT_PRED
@ VERT_PRED
Definition: vp9.h:46
AVCodecContext::thread_count
int thread_count
thread count is used to decide how many independent tasks should be passed to execute()
Definition: avcodec.h:1582
ff_vp8_decoder
const FFCodec ff_vp8_decoder
VP8mv::y
int16_t y
Definition: vp8.h:86
DIAG_DOWN_RIGHT_PRED
@ DIAG_DOWN_RIGHT_PRED
Definition: vp9.h:50
MAX_THREADS
#define MAX_THREADS
Definition: frame_thread_encoder.c:35
ff_videodsp_init
av_cold void ff_videodsp_init(VideoDSPContext *ctx, int bpc)
Definition: videodsp.c:39
update
static av_always_inline void update(SilenceDetectContext *s, AVFrame *insamples, int is_silence, int current_sample, int64_t nb_samples_notify, AVRational time_base)
Definition: af_silencedetect.c:77
VP8Macroblock::bmv
VP8mv bmv[16]
Definition: vp8.h:107
idct_mb
static av_always_inline void idct_mb(VP8Context *s, VP8ThreadData *td, uint8_t *const dst[3], const VP8Macroblock *mb)
Definition: vp8.c:2094
check_intra_pred8x8_mode_emuedge
static av_always_inline int check_intra_pred8x8_mode_emuedge(int mode, int mb_x, int mb_y, int vp7)
Definition: vp8.c:1641
filter_level_for_mb
static av_always_inline void filter_level_for_mb(const VP8Context *s, const VP8Macroblock *mb, VP8FilterStrength *f, int is_vp7)
Definition: vp8.c:2157
read_mv_component
static av_always_inline int read_mv_component(VPXRangeCoder *c, const uint8_t *p, int vp7)
Motion vector coding, 17.1.
Definition: vp8.c:893
vp8_filter_mb_row
static void vp8_filter_mb_row(AVCodecContext *avctx, void *tdata, int jobnr, int threadnr)
Definition: vp8.c:2608
vp7_get_quants
static void vp7_get_quants(VP8Context *s)
Definition: vp8.c:372
refstruct.h
VP8_SPLITMVMODE_16x8
@ VP8_SPLITMVMODE_16x8
2 16x8 blocks (vertical)
Definition: vp8.h:77
vp8_init_frames
static av_cold int vp8_init_frames(VP8Context *s)
Definition: vp8.c:2882
av_frame_alloc
AVFrame * av_frame_alloc(void)
Allocate an AVFrame and set its fields to default values.
Definition: frame.c:118
vp8_mc_part
static av_always_inline void vp8_mc_part(VP8Context *s, VP8ThreadData *td, uint8_t *const dst[3], const ThreadFrame *ref_frame, int x_off, int y_off, int bx_off, int by_off, int block_w, int block_h, int width, int height, const VP8mv *mv)
Definition: vp8.c:1949
vp8_mc_luma
static av_always_inline void vp8_mc_luma(VP8Context *s, VP8ThreadData *td, uint8_t *dst, const ThreadFrame *ref, const VP8mv *mv, int x_off, int y_off, int block_w, int block_h, int width, int height, ptrdiff_t linesize, vp8_mc_func mc_func[3][3])
luma MC function
Definition: vp8.c:1841
ff_vp7dsp_init
void ff_vp7dsp_init(VP8DSPContext *c)
AV_LOG_ERROR
#define AV_LOG_ERROR
Something went wrong and cannot losslessly be recovered.
Definition: log.h:180
ff_vp8dsp_init
void ff_vp8dsp_init(VP8DSPContext *c)
FF_ARRAY_ELEMS
#define FF_ARRAY_ELEMS(a)
Definition: sinewin_tablegen.c:29
HOR_PRED
@ HOR_PRED
Definition: vp9.h:47
av_cold
#define av_cold
Definition: attributes.h:90
vp8_dct_cat2_prob
static const uint8_t vp8_dct_cat2_prob[]
Definition: vp8data.h:339
AV_FRAME_FLAG_KEY
#define AV_FRAME_FLAG_KEY
A flag to mark frames that are keyframes.
Definition: frame.h:595
ff_thread_report_progress
void ff_thread_report_progress(ThreadFrame *f, int n, int field)
Notify later decoding threads when part of their reference picture is ready.
Definition: pthread_frame.c:573
LOCAL_ALIGNED
#define LOCAL_ALIGNED(a, t, v,...)
Definition: mem_internal.h:133
width
#define width
vp8_pred4x4_mode
static const uint8_t vp8_pred4x4_mode[]
Definition: vp8data.h:40
FF_CODEC_DECODE_CB
#define FF_CODEC_DECODE_CB(func)
Definition: codec_internal.h:287
ff_hwaccel_frame_priv_alloc
int ff_hwaccel_frame_priv_alloc(AVCodecContext *avctx, void **hwaccel_picture_private)
Allocate a hwaccel frame private data if the provided avctx uses a hwaccel method that needs it.
Definition: decode.c:1901
s
#define s(width, name)
Definition: cbs_vp9.c:198
filter_mb_simple
static av_always_inline void filter_mb_simple(const VP8Context *s, uint8_t *dst, const VP8FilterStrength *f, int mb_x, int mb_y)
Definition: vp8.c:2277
vpx_rac_renorm
static av_always_inline unsigned int vpx_rac_renorm(VPXRangeCoder *c)
Definition: vpx_rac.h:58
AV_ZERO64
#define AV_ZERO64(d)
Definition: intreadwrite.h:629
vp8_pred8x8c_prob_inter
static const uint8_t vp8_pred8x8c_prob_inter[3]
Definition: vp8data.h:189
DC_129_PRED8x8
#define DC_129_PRED8x8
Definition: h264pred.h:86
AV_ZERO32
#define AV_ZERO32(d)
Definition: intreadwrite.h:625
AV_GET_BUFFER_FLAG_REF
#define AV_GET_BUFFER_FLAG_REF
The decoder will keep a reference to the frame and may reuse it later.
Definition: avcodec.h:425
VP8Frame::tf
ThreadFrame tf
Definition: vp8.h:153
vp8_pred16x16_tree_intra
static const int8_t vp8_pred16x16_tree_intra[4][2]
Definition: vp8data.h:47
bits
uint8_t bits
Definition: vp3data.h:128
parse_segment_info
static void parse_segment_info(VP8Context *s)
Definition: vp8.c:295
vp8_pred4x4_prob_inter
static const uint8_t vp8_pred4x4_prob_inter[9]
Definition: vp8data.h:192
pix_fmts
static enum AVPixelFormat pix_fmts[]
Definition: libkvazaar.c:304
vp8_mbsplits
static const uint8_t vp8_mbsplits[5][16]
Definition: vp8data.h:127
ctx
AVFormatContext * ctx
Definition: movenc.c:48
vp78_decode_frame
static av_always_inline int vp78_decode_frame(AVCodecContext *avctx, AVFrame *rframe, int *got_frame, const AVPacket *avpkt, int is_vp7)
Definition: vp8.c:2662
decode.h
AV_RL16
uint64_t_TMPL AV_WL64 unsigned int_TMPL AV_WL32 unsigned int_TMPL AV_WL24 unsigned int_TMPL AV_RL16
Definition: bytestream.h:94
vp7_mode_contexts
static const int vp7_mode_contexts[31][4]
Definition: vp8data.h:84
vp78_decode_mv_mb_modes
static av_always_inline int vp78_decode_mv_mb_modes(AVCodecContext *avctx, VP8Frame *curframe, const VP8Frame *prev_frame, int is_vp7)
Definition: vp8.c:2311
AV_PIX_FMT_YUV420P
@ AV_PIX_FMT_YUV420P
planar YUV 4:2:0, 12bpp, (1 Cr & Cb sample per 2x2 Y samples)
Definition: pixfmt.h:73
atomic_load
#define atomic_load(object)
Definition: stdatomic.h:93
VP8_SPLITMVMODE_8x16
@ VP8_SPLITMVMODE_8x16
2 8x16 blocks (horizontal)
Definition: vp8.h:78
CODEC_LONG_NAME
#define CODEC_LONG_NAME(str)
Definition: codec_internal.h:272
VP8Frame::seg_map
uint8_t * seg_map
RefStruct reference.
Definition: vp8.h:154
frame
static AVFrame * frame
Definition: demux_decode.c:54
ff_thread_ref_frame
int ff_thread_ref_frame(ThreadFrame *dst, const ThreadFrame *src)
Definition: utils.c:853
vp8_mv_default_prob
static const uint8_t vp8_mv_default_prob[2][19]
Definition: vp8data.h:562
vp8_coeff_band_indexes
static const int8_t vp8_coeff_band_indexes[8][10]
Definition: vp8data.h:325
TOP_DC_PRED8x8
#define TOP_DC_PRED8x8
Definition: h264pred.h:75
if
if(ret)
Definition: filter_design.txt:179
vp8_pred16x16_prob_inter
static const uint8_t vp8_pred16x16_prob_inter[4]
Definition: vp8data.h:164
vp8_decode_flush_impl
static void vp8_decode_flush_impl(AVCodecContext *avctx, int free_mem)
Definition: vp8.c:150
AV_CODEC_CAP_FRAME_THREADS
#define AV_CODEC_CAP_FRAME_THREADS
Codec supports frame-level multithreading.
Definition: codec.h:110
AVDISCARD_ALL
@ AVDISCARD_ALL
discard all
Definition: defs.h:219
threadframe.h
vp8_rac_get_nn
static int vp8_rac_get_nn(VPXRangeCoder *c)
Definition: vp8.c:65
clamp_mv
static av_always_inline void clamp_mv(const VP8mvbounds *s, VP8mv *dst, const VP8mv *src)
Definition: vp8.c:882
NULL
#define NULL
Definition: coverity.c:32
sizes
static const int sizes[][2]
Definition: img2dec.c:60
AV_COPY128
#define AV_COPY128(d, s)
Definition: intreadwrite.h:605
AVCodecContext::color_range
enum AVColorRange color_range
MPEG vs JPEG YUV range.
Definition: avcodec.h:695
vp7_decode_mb_row_sliced
static int vp7_decode_mb_row_sliced(AVCodecContext *avctx, void *tdata, int jobnr, int threadnr)
Definition: vp8.c:2649
AV_COPY64
#define AV_COPY64(d, s)
Definition: intreadwrite.h:601
hwaccel_internal.h
vp8_update_dimensions
static int vp8_update_dimensions(VP8Context *s, int width, int height)
Definition: vp8.c:289
VP8FilterStrength
Definition: vp8.h:89
NUM_DCT_TOKENS
@ NUM_DCT_TOKENS
Definition: vp8.h:64
AV_PICTURE_TYPE_I
@ AV_PICTURE_TYPE_I
Intra.
Definition: avutil.h:279
vp89_rac_get_uint
static av_unused int vp89_rac_get_uint(VPXRangeCoder *c, int bits)
Definition: vp89_rac.h:41
check_thread_pos
#define check_thread_pos(td, otd, mb_x_check, mb_y_check)
Definition: vp8.c:2394
VP7MVPred
Definition: vp8data.h:61
ff_thread_release_ext_buffer
void ff_thread_release_ext_buffer(ThreadFrame *f)
Unref a ThreadFrame.
Definition: pthread_frame.c:996
mathops.h
flush
void(* flush)(AVBSFContext *ctx)
Definition: dts2pts.c:367
vp8_mc_func
void(* vp8_mc_func)(uint8_t *dst, ptrdiff_t dstStride, const uint8_t *src, ptrdiff_t srcStride, int h, int x, int y)
Definition: vp8dsp.h:33
vp7_yac_qlookup
static const uint16_t vp7_yac_qlookup[]
Definition: vp8data.h:597
vp8_token_default_probs
static const uint8_t vp8_token_default_probs[4][8][3][NUM_DCT_TOKENS - 1]
Definition: vp8data.h:345
ff_refstruct_allocz
static void * ff_refstruct_allocz(size_t size)
Equivalent to ff_refstruct_alloc_ext(size, 0, NULL, NULL)
Definition: refstruct.h:105
VERT_PRED8x8
#define VERT_PRED8x8
Definition: h264pred.h:70
vp8_mbsplit_count
static const uint8_t vp8_mbsplit_count[4]
Definition: vp8data.h:142
UPDATE_THREAD_CONTEXT
#define UPDATE_THREAD_CONTEXT(func)
Definition: codec_internal.h:281
vp8_decode_mvs
static av_always_inline void vp8_decode_mvs(VP8Context *s, const VP8mvbounds *mv_bounds, VP8Macroblock *mb, int mb_x, int mb_y, int layout)
Definition: vp8.c:1140
AV_ZERO128
#define AV_ZERO128(d)
Definition: intreadwrite.h:633
FF_HW_HAS_CB
#define FF_HW_HAS_CB(avctx, function)
Definition: hwaccel_internal.h:177
VP8FrameType
VP8FrameType
Definition: vp8.h:42
decode_mb_row_no_filter
static av_always_inline int decode_mb_row_no_filter(AVCodecContext *avctx, void *tdata, int jobnr, int threadnr, int is_vp7)
Definition: vp8.c:2398
VP8mv
Definition: vp8.h:84
vp7_feature_value_size
static const uint8_t vp7_feature_value_size[2][4]
Definition: vp8data.h:573
index
int index
Definition: gxfenc.c:89
c
Undefined Behavior In the C some operations are like signed integer dereferencing freed accessing outside allocated Undefined Behavior must not occur in a C it is not safe even if the output of undefined operations is unused The unsafety may seem nit picking but Optimizing compilers have in fact optimized code on the assumption that no undefined Behavior occurs Optimizing code based on wrong assumptions can and has in some cases lead to effects beyond the output of computations The signed integer overflow problem in speed critical code Code which is highly optimized and works with signed integers sometimes has the problem that often the output of the computation does not c
Definition: undefined.txt:32
VP8Frame
Definition: vp8.h:152
vp8.h
ff_vp8_decode_init
av_cold int ff_vp8_decode_init(AVCodecContext *avctx)
Definition: vp8.c:2936
VP8_FRAME_GOLDEN
@ VP8_FRAME_GOLDEN
Definition: vp8.h:46
FF_SIGNBIT
#define FF_SIGNBIT(x)
Definition: mathops.h:130
vp8_mbfirstidx
static const uint8_t vp8_mbfirstidx[4][16]
Definition: vp8data.h:135
DC_127_PRED8x8
#define DC_127_PRED8x8
Definition: h264pred.h:85
AVDISCARD_NONKEY
@ AVDISCARD_NONKEY
discard all frames except keyframes
Definition: defs.h:218
f
f
Definition: af_crystalizer.c:121
init
int(* init)(AVBSFContext *ctx)
Definition: dts2pts.c:365
xchg_mb_border
static av_always_inline void xchg_mb_border(uint8_t *top_border, uint8_t *src_y, uint8_t *src_cb, uint8_t *src_cr, ptrdiff_t linesize, ptrdiff_t uvlinesize, int mb_x, int mb_y, int mb_width, int simple, int xchg)
Definition: vp8.c:1589
ff_zigzag_scan
const uint8_t ff_zigzag_scan[16+1]
Definition: mathtables.c:109
AV_CODEC_CAP_DR1
#define AV_CODEC_CAP_DR1
Codec uses get_buffer() or get_encode_buffer() for allocating buffers and supports custom allocators.
Definition: codec.h:52
vp8_pred4x4_tree
static const int8_t vp8_pred4x4_tree[9][2]
Definition: vp8data.h:168
MV_EDGE_CHECK
#define MV_EDGE_CHECK(n)
AVPacket::size
int size
Definition: packet.h:523
dc
Tag MUST be and< 10hcoeff half pel interpolation filter coefficients, hcoeff[0] are the 2 middle coefficients[1] are the next outer ones and so on, resulting in a filter like:...eff[2], hcoeff[1], hcoeff[0], hcoeff[0], hcoeff[1], hcoeff[2] ... the sign of the coefficients is not explicitly stored but alternates after each coeff and coeff[0] is positive, so ...,+,-,+,-,+,+,-,+,-,+,... hcoeff[0] is not explicitly stored but found by subtracting the sum of all stored coefficients with signs from 32 hcoeff[0]=32 - hcoeff[1] - hcoeff[2] - ... a good choice for hcoeff and htaps is htaps=6 hcoeff={40,-10, 2} an alternative which requires more computations at both encoder and decoder side and may or may not be better is htaps=8 hcoeff={42,-14, 6,-2}ref_frames minimum of the number of available reference frames and max_ref_frames for example the first frame after a key frame always has ref_frames=1spatial_decomposition_type wavelet type 0 is a 9/7 symmetric compact integer wavelet 1 is a 5/3 symmetric compact integer wavelet others are reserved stored as delta from last, last is reset to 0 if always_reset||keyframeqlog quality(logarithmic quantizer scale) stored as delta from last, last is reset to 0 if always_reset||keyframemv_scale stored as delta from last, last is reset to 0 if always_reset||keyframe FIXME check that everything works fine if this changes between framesqbias dequantization bias stored as delta from last, last is reset to 0 if always_reset||keyframeblock_max_depth maximum depth of the block tree stored as delta from last, last is reset to 0 if always_reset||keyframequant_table quantization tableHighlevel bitstream structure:==============================--------------------------------------------|Header|--------------------------------------------|------------------------------------|||Block0||||split?||||yes no||||......... intra?||||:Block01 :yes no||||:Block02 :....... ..........||||:Block03 ::y DC ::ref index:||||:Block04 ::cb DC ::motion x :||||......... :cr DC ::motion y :||||....... ..........|||------------------------------------||------------------------------------|||Block1|||...|--------------------------------------------|------------ ------------ ------------|||Y subbands||Cb subbands||Cr subbands||||--- ---||--- ---||--- ---|||||LL0||HL0||||LL0||HL0||||LL0||HL0|||||--- ---||--- ---||--- ---||||--- ---||--- ---||--- ---|||||LH0||HH0||||LH0||HH0||||LH0||HH0|||||--- ---||--- ---||--- ---||||--- ---||--- ---||--- ---|||||HL1||LH1||||HL1||LH1||||HL1||LH1|||||--- ---||--- ---||--- ---||||--- ---||--- ---||--- ---|||||HH1||HL2||||HH1||HL2||||HH1||HL2|||||...||...||...|||------------ ------------ ------------|--------------------------------------------Decoding process:=================------------|||Subbands|------------||||------------|Intra DC||||LL0 subband prediction ------------|\ Dequantization ------------------- \||Reference frames|\ IDWT|------- -------|Motion \|||Frame 0||Frame 1||Compensation . OBMC v -------|------- -------|--------------. \------> Frame n output Frame Frame<----------------------------------/|...|------------------- Range Coder:============Binary Range Coder:------------------- The implemented range coder is an adapted version based upon "Range encoding: an algorithm for removing redundancy from a digitised message." by G. N. N. Martin. The symbols encoded by the Snow range coder are bits(0|1). The associated probabilities are not fix but change depending on the symbol mix seen so far. bit seen|new state ---------+----------------------------------------------- 0|256 - state_transition_table[256 - old_state];1|state_transition_table[old_state];state_transition_table={ 0, 0, 0, 0, 0, 0, 0, 0, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 190, 191, 192, 194, 194, 195, 196, 197, 198, 199, 200, 201, 202, 202, 204, 205, 206, 207, 208, 209, 209, 210, 211, 212, 213, 215, 215, 216, 217, 218, 219, 220, 220, 222, 223, 224, 225, 226, 227, 227, 229, 229, 230, 231, 232, 234, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 248, 0, 0, 0, 0, 0, 0, 0};FIXME Range Coding of integers:------------------------- FIXME Neighboring Blocks:===================left and top are set to the respective blocks unless they are outside of the image in which case they are set to the Null block top-left is set to the top left block unless it is outside of the image in which case it is set to the left block if this block has no larger parent block or it is at the left side of its parent block and the top right block is not outside of the image then the top right block is used for top-right else the top-left block is used Null block y, cb, cr are 128 level, ref, mx and my are 0 Motion Vector Prediction:=========================1. the motion vectors of all the neighboring blocks are scaled to compensate for the difference of reference frames scaled_mv=(mv *(256 *(current_reference+1)/(mv.reference+1))+128)> the median of the scaled top and top right vectors is used as motion vector prediction the used motion vector is the sum of the predictor and(mvx_diff, mvy_diff) *mv_scale Intra DC Prediction block[y][x] dc[1]
Definition: snow.txt:400
copy
static void copy(const float *p1, float *p2, const int length)
Definition: vf_vaguedenoiser.c:185
av_frame_ref
int av_frame_ref(AVFrame *dst, const AVFrame *src)
Set up a new reference to the data described by the source frame.
Definition: frame.c:354
codec_internal.h
vp8_coeff_band
static const uint8_t vp8_coeff_band[16]
Definition: vp8data.h:319
subpel_idx
static const uint8_t subpel_idx[3][8]
Definition: vp8.c:1817
vp7_update_dimensions
static int vp7_update_dimensions(VP8Context *s, int width, int height)
Definition: vp8.c:284
EDGE_EMU_LINESIZE
#define EDGE_EMU_LINESIZE
Definition: vp8.h:146
size
int size
Definition: twinvq_data.h:10344
VERT_RIGHT_PRED
@ VERT_RIGHT_PRED
Definition: vp9.h:51
FF_CODEC_CAP_ALLOCATE_PROGRESS
#define FF_CODEC_CAP_ALLOCATE_PROGRESS
Definition: codec_internal.h:69
free_buffers
static void free_buffers(VP8Context *s)
Definition: vp8.c:83
DC_128_PRED8x8
#define DC_128_PRED8x8
Definition: h264pred.h:76
decode_block_coeffs
static av_always_inline int decode_block_coeffs(VPXRangeCoder *c, int16_t block[16], uint8_t probs[16][3][NUM_DCT_TOKENS - 1], int i, int zero_nhood, const int16_t qmul[2], const uint8_t scan[16], int vp7)
Definition: vp8.c:1487
FF_THREAD_SLICE
#define FF_THREAD_SLICE
Decode more than one part of a single frame at once.
Definition: avcodec.h:1594
ff_vp8_dct_cat_prob
const uint8_t *const ff_vp8_dct_cat_prob[]
Definition: vp8data.c:36
vp8_pred8x8c_prob_intra
static const uint8_t vp8_pred8x8c_prob_intra[3]
Definition: vp8data.h:186
vp8_decode_mv_mb_modes
static int vp8_decode_mv_mb_modes(AVCodecContext *avctx, VP8Frame *cur_frame, const VP8Frame *prev_frame)
Definition: vp8.c:2354
AV_RL24
uint64_t_TMPL AV_WL64 unsigned int_TMPL AV_WL32 unsigned int_TMPL AV_RL24
Definition: bytestream.h:93
AVCodecHWConfigInternal
Definition: hwconfig.h:25
vp8_pred4x4_prob_intra
static const uint8_t vp8_pred4x4_prob_intra[10][10][9]
Definition: vp8data.h:196
ref_frame
static int ref_frame(VVCFrame *dst, const VVCFrame *src)
Definition: vvcdec.c:561
VP8ThreadData
Definition: vp8.h:120
height
#define height
a
The reader does not expect b to be semantically here and if the code is changed by maybe adding a a division or other the signedness will almost certainly be mistaken To avoid this confusion a new type was SUINT is the C unsigned type but it holds a signed int to use the same example SUINT a
Definition: undefined.txt:41
AV_CODEC_CAP_SLICE_THREADS
#define AV_CODEC_CAP_SLICE_THREADS
Codec supports slice-based (or partition-based) multithreading.
Definition: codec.h:114
vp8_decode_frame_header
static int vp8_decode_frame_header(VP8Context *s, const uint8_t *buf, int buf_size)
Definition: vp8.c:741
vp8_mbsplit_prob
static const uint8_t vp8_mbsplit_prob[3]
Definition: vp8data.h:145
setup_partitions
static int setup_partitions(VP8Context *s, const uint8_t *buf, int buf_size)
Definition: vp8.c:341
vp8_pred16x16_tree_inter
static const int8_t vp8_pred16x16_tree_inter[4][2]
Definition: vp8data.h:54
vp7_feature_index_tree
static const int8_t vp7_feature_index_tree[4][2]
Definition: vp8data.h:578
AVCodecContext::skip_loop_filter
enum AVDiscard skip_loop_filter
Skip loop filtering for selected frames.
Definition: avcodec.h:1805
HWACCEL_NVDEC
#define HWACCEL_NVDEC(codec)
Definition: hwconfig.h:68
PLANE_PRED8x8
#define PLANE_PRED8x8
Definition: h264pred.h:71
vpx_rac_is_end
static av_always_inline int vpx_rac_is_end(VPXRangeCoder *c)
returns 1 if the end of the stream has been reached, 0 otherwise.
Definition: vpx_rac.h:51
AV_PIX_FMT_VAAPI
@ AV_PIX_FMT_VAAPI
Hardware acceleration through VA-API, data[3] contains a VASurfaceID.
Definition: pixfmt.h:126
mb
#define mb
Definition: vf_colormatrix.c:99
pthread_cond_destroy
static av_always_inline int pthread_cond_destroy(pthread_cond_t *cond)
Definition: os2threads.h:144
MODE_I4x4
#define MODE_I4x4
Definition: vp8.h:68
FF_THREAD_FRAME
#define FF_THREAD_FRAME
Decode more than one frame at once.
Definition: avcodec.h:1593
H_LOOP_FILTER_16Y_INNER
#define H_LOOP_FILTER_16Y_INNER(cond)
vp78_decode_mb_row_sliced
static av_always_inline int vp78_decode_mb_row_sliced(AVCodecContext *avctx, void *tdata, int jobnr, int threadnr, int is_vp7)
Definition: vp8.c:2615
pthread_mutex_destroy
static av_always_inline int pthread_mutex_destroy(pthread_mutex_t *mutex)
Definition: os2threads.h:112
layout
Filter the word “frame” indicates either a video frame or a group of audio as stored in an AVFrame structure Format for each input and each output the list of supported formats For video that means pixel format For audio that means channel layout
Definition: filter_design.txt:18
AV_CODEC_ID_VP7
@ AV_CODEC_ID_VP7
Definition: codec_id.h:233
ref_to_update
static VP8FrameType ref_to_update(VP8Context *s, int update, VP8FrameType ref)
Determine which buffers golden and altref should be updated with after this frame.
Definition: vp8.c:437
vp8_read_mv_component
static int vp8_read_mv_component(VPXRangeCoder *c, const uint8_t *p)
Definition: vp8.c:926
i
#define i(width, name, range_min, range_max)
Definition: cbs_h2645.c:255
DC_129_PRED
@ DC_129_PRED
Definition: vp9.h:60
modes
static const SiprModeParam modes[MODE_COUNT]
Definition: sipr.c:70
vpx_rac.h
src2
const pixel * src2
Definition: h264pred_template.c:422
vp7_fade_frame
static int vp7_fade_frame(VP8Context *s, int alpha, int beta)
Definition: vp8.c:539
av_always_inline
#define av_always_inline
Definition: attributes.h:49
FFMIN
#define FFMIN(a, b)
Definition: macros.h:49
vpx_rac_get_prob_branchy
static av_always_inline int vpx_rac_get_prob_branchy(VPXRangeCoder *c, int prob)
Definition: vpx_rac.h:99
intra_predict
static av_always_inline void intra_predict(VP8Context *s, VP8ThreadData *td, uint8_t *const dst[3], VP8Macroblock *mb, int mb_x, int mb_y, int is_vp7)
Definition: vp8.c:1703
AV_COPY32
#define AV_COPY32(d, s)
Definition: intreadwrite.h:597
av_mallocz
void * av_mallocz(size_t size)
Allocate a memory block with alignment suitable for all memory accesses (including vectors if availab...
Definition: mem.c:254
AVCodec::name
const char * name
Name of the codec implementation.
Definition: codec.h:194
VP8mv::x
int16_t x
Definition: vp8.h:85
vp78_reset_probability_tables
static void vp78_reset_probability_tables(VP8Context *s)
Definition: vp8.c:453
vp7_decode_frame_header
static int vp7_decode_frame_header(VP8Context *s, const uint8_t *buf, int buf_size)
Definition: vp8.c:577
VP8_FRAME_NONE
@ VP8_FRAME_NONE
Definition: vp8.h:43
profile
int profile
Definition: mxfenc.c:2226
fade
static void fade(uint8_t *dst, ptrdiff_t dst_linesize, const uint8_t *src, ptrdiff_t src_linesize, int width, int height, int alpha, int beta)
Definition: vp8.c:523
decode_intra4x4_modes
static av_always_inline void decode_intra4x4_modes(VP8Context *s, VPXRangeCoder *c, VP8Macroblock *mb, int mb_x, int keyframe, int layout)
Definition: vp8.c:1240
VP8Macroblock
Definition: vp8.h:95
AVCodecContext::pix_fmt
enum AVPixelFormat pix_fmt
Pixel format, see AV_PIX_FMT_xxx.
Definition: avcodec.h:657
ff_thread_get_ext_buffer
int ff_thread_get_ext_buffer(AVCodecContext *avctx, ThreadFrame *f, int flags)
Wrapper around ff_get_buffer() for frame-multithreaded codecs.
Definition: pthread_frame.c:968
AVCOL_RANGE_MPEG
@ AVCOL_RANGE_MPEG
Narrow or limited range content.
Definition: pixfmt.h:666
vp8_decode_mb_row_sliced
static int vp8_decode_mb_row_sliced(AVCodecContext *avctx, void *tdata, int jobnr, int threadnr)
Definition: vp8.c:2655
ff_vp7_decoder
const FFCodec ff_vp7_decoder
VP8_SPLITMVMODE_NONE
@ VP8_SPLITMVMODE_NONE
(only used in prediction) no split MVs
Definition: vp8.h:81
LEFT_DC_PRED8x8
#define LEFT_DC_PRED8x8
Definition: h264pred.h:74
prefetch_motion
static av_always_inline void prefetch_motion(const VP8Context *s, const VP8Macroblock *mb, int mb_x, int mb_y, int mb_xy, int ref)
Definition: vp8.c:1987
avcodec.h
AV_RN32A
#define AV_RN32A(p)
Definition: intreadwrite.h:524
vp89_rac_get_tree
static av_always_inline int vp89_rac_get_tree(VPXRangeCoder *c, const int8_t(*tree)[2], const uint8_t *probs)
Definition: vp89_rac.h:54
decode_mb_coeffs
static av_always_inline void decode_mb_coeffs(VP8Context *s, VP8ThreadData *td, VPXRangeCoder *c, VP8Macroblock *mb, uint8_t t_nnz[9], uint8_t l_nnz[9], int is_vp7)
Definition: vp8.c:1502
av_uninit
#define av_uninit(x)
Definition: attributes.h:154
ret
ret
Definition: filter_design.txt:187
AV_LOG_FATAL
#define AV_LOG_FATAL
Something went wrong and recovery is not possible.
Definition: log.h:174
check_dc_pred8x8_mode
static av_always_inline int check_dc_pred8x8_mode(int mode, int mb_x, int mb_y)
Definition: vp8.c:1623
pred
static const float pred[4]
Definition: siprdata.h:259
vp7_decode_mb_row_no_filter
static int vp7_decode_mb_row_no_filter(AVCodecContext *avctx, void *tdata, int jobnr, int threadnr)
Definition: vp8.c:2531
FFSWAP
#define FFSWAP(type, a, b)
Definition: macros.h:52
vp8_pred16x16_prob_intra
static const uint8_t vp8_pred16x16_prob_intra[4]
Definition: vp8data.h:161
vp8_ac_qlookup
static const uint16_t vp8_ac_qlookup[VP8_MAX_QUANT+1]
Definition: vp8data.h:529
prob
#define prob(name, subs,...)
Definition: cbs_vp9.c:325
hwaccel
static const char * hwaccel
Definition: ffplay.c:356
ff_vpx_init_range_decoder
int ff_vpx_init_range_decoder(VPXRangeCoder *c, const uint8_t *buf, int buf_size)
Definition: vpx_rac.c:42
ff_refstruct_replace
void ff_refstruct_replace(void *dstp, const void *src)
Ensure *dstp refers to the same object as src.
Definition: refstruct.c:160
ff_thread_finish_setup
the pkt_dts and pkt_pts fields in AVFrame will work as usual Restrictions on codec whose streams don t reset across will not work because their bitstreams cannot be decoded in parallel *The contents of buffers must not be read before as well as code calling up to before the decode process starts Call ff_thread_finish_setup() afterwards. If some code can 't be moved
left
Tag MUST be and< 10hcoeff half pel interpolation filter coefficients, hcoeff[0] are the 2 middle coefficients[1] are the next outer ones and so on, resulting in a filter like:...eff[2], hcoeff[1], hcoeff[0], hcoeff[0], hcoeff[1], hcoeff[2] ... the sign of the coefficients is not explicitly stored but alternates after each coeff and coeff[0] is positive, so ...,+,-,+,-,+,+,-,+,-,+,... hcoeff[0] is not explicitly stored but found by subtracting the sum of all stored coefficients with signs from 32 hcoeff[0]=32 - hcoeff[1] - hcoeff[2] - ... a good choice for hcoeff and htaps is htaps=6 hcoeff={40,-10, 2} an alternative which requires more computations at both encoder and decoder side and may or may not be better is htaps=8 hcoeff={42,-14, 6,-2}ref_frames minimum of the number of available reference frames and max_ref_frames for example the first frame after a key frame always has ref_frames=1spatial_decomposition_type wavelet type 0 is a 9/7 symmetric compact integer wavelet 1 is a 5/3 symmetric compact integer wavelet others are reserved stored as delta from last, last is reset to 0 if always_reset||keyframeqlog quality(logarithmic quantizer scale) stored as delta from last, last is reset to 0 if always_reset||keyframemv_scale stored as delta from last, last is reset to 0 if always_reset||keyframe FIXME check that everything works fine if this changes between framesqbias dequantization bias stored as delta from last, last is reset to 0 if always_reset||keyframeblock_max_depth maximum depth of the block tree stored as delta from last, last is reset to 0 if always_reset||keyframequant_table quantization tableHighlevel bitstream structure:==============================--------------------------------------------|Header|--------------------------------------------|------------------------------------|||Block0||||split?||||yes no||||......... intra?||||:Block01 :yes no||||:Block02 :....... ..........||||:Block03 ::y DC ::ref index:||||:Block04 ::cb DC ::motion x :||||......... :cr DC ::motion y :||||....... ..........|||------------------------------------||------------------------------------|||Block1|||...|--------------------------------------------|------------ ------------ ------------|||Y subbands||Cb subbands||Cr subbands||||--- ---||--- ---||--- ---|||||LL0||HL0||||LL0||HL0||||LL0||HL0|||||--- ---||--- ---||--- ---||||--- ---||--- ---||--- ---|||||LH0||HH0||||LH0||HH0||||LH0||HH0|||||--- ---||--- ---||--- ---||||--- ---||--- ---||--- ---|||||HL1||LH1||||HL1||LH1||||HL1||LH1|||||--- ---||--- ---||--- ---||||--- ---||--- ---||--- ---|||||HH1||HL2||||HH1||HL2||||HH1||HL2|||||...||...||...|||------------ ------------ ------------|--------------------------------------------Decoding process:=================------------|||Subbands|------------||||------------|Intra DC||||LL0 subband prediction ------------|\ Dequantization ------------------- \||Reference frames|\ IDWT|------- -------|Motion \|||Frame 0||Frame 1||Compensation . OBMC v -------|------- -------|--------------. \------> Frame n output Frame Frame<----------------------------------/|...|------------------- Range Coder:============Binary Range Coder:------------------- The implemented range coder is an adapted version based upon "Range encoding: an algorithm for removing redundancy from a digitised message." by G. N. N. Martin. The symbols encoded by the Snow range coder are bits(0|1). The associated probabilities are not fix but change depending on the symbol mix seen so far. bit seen|new state ---------+----------------------------------------------- 0|256 - state_transition_table[256 - old_state];1|state_transition_table[old_state];state_transition_table={ 0, 0, 0, 0, 0, 0, 0, 0, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 190, 191, 192, 194, 194, 195, 196, 197, 198, 199, 200, 201, 202, 202, 204, 205, 206, 207, 208, 209, 209, 210, 211, 212, 213, 215, 215, 216, 217, 218, 219, 220, 220, 222, 223, 224, 225, 226, 227, 227, 229, 229, 230, 231, 232, 234, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 248, 0, 0, 0, 0, 0, 0, 0};FIXME Range Coding of integers:------------------------- FIXME Neighboring Blocks:===================left and top are set to the respective blocks unless they are outside of the image in which case they are set to the Null block top-left is set to the top left block unless it is outside of the image in which case it is set to the left block if this block has no larger parent block or it is at the left side of its parent block and the top right block is not outside of the image then the top right block is used for top-right else the top-left block is used Null block y, cb, cr are 128 level, ref, mx and my are 0 Motion Vector Prediction:=========================1. the motion vectors of all the neighboring blocks are scaled to compensate for the difference of reference frames scaled_mv=(mv *(256 *(current_reference+1)/(mv.reference+1))+128)> the median of the scaled left
Definition: snow.txt:386
AV_RL32
uint64_t_TMPL AV_WL64 unsigned int_TMPL AV_RL32
Definition: bytestream.h:92
vp89_rac_get
static av_always_inline int vp89_rac_get(VPXRangeCoder *c)
Definition: vp89_rac.h:36
AVCodecContext
main external API structure.
Definition: avcodec.h:445
AVCodecContext::active_thread_type
int active_thread_type
Which multithreading methods are in use by the codec.
Definition: avcodec.h:1601
VP8Macroblock::intra4x4_pred_mode_top
uint8_t intra4x4_pred_mode_top[4]
Definition: vp8.h:105
decode_splitmvs
static av_always_inline int decode_splitmvs(const VP8Context *s, VPXRangeCoder *c, VP8Macroblock *mb, int layout, int is_vp7)
Split motion vector prediction, 16.4.
Definition: vp8.c:949
ThreadFrame
Definition: threadframe.h:27
ff_h264_pred_init
av_cold void ff_h264_pred_init(H264PredContext *h, int codec_id, const int bit_depth, int chroma_format_idc)
Set the intra prediction function pointers.
Definition: h264pred.c:437
HOR_UP_PRED
@ HOR_UP_PRED
Definition: vp9.h:54
vp8data.h
mode
mode
Definition: ebur128.h:83
AV_PIX_FMT_NONE
@ AV_PIX_FMT_NONE
Definition: pixfmt.h:72
ffhwaccel
static const FFHWAccel * ffhwaccel(const AVHWAccel *codec)
Definition: hwaccel_internal.h:166
VP8Macroblock::mode
uint8_t mode
Definition: vp8.h:99
IS_VP8
#define IS_VP8
Definition: vp8dsp.h:101
VP8_MVMODE_SPLIT
@ VP8_MVMODE_SPLIT
Definition: vp8.h:73
ref
static int ref[MAX_W *MAX_W]
Definition: jpeg2000dwt.c:112
HOR_DOWN_PRED
@ HOR_DOWN_PRED
Definition: vp9.h:52
vp7_decode_block_coeffs_internal
static int vp7_decode_block_coeffs_internal(VPXRangeCoder *r, int16_t block[16], uint8_t probs[16][3][NUM_DCT_TOKENS - 1], int i, const uint8_t *token_prob, const int16_t qmul[2], const uint8_t scan[16])
Definition: vp8.c:1450
filter_mb
static av_always_inline void filter_mb(const VP8Context *s, uint8_t *const dst[3], const VP8FilterStrength *f, int mb_x, int mb_y, int is_vp7)
Definition: vp8.c:2190
segment
Definition: hls.c:76
av_clip_uint8
#define av_clip_uint8
Definition: common.h:104
vp78_update_probability_tables
static void vp78_update_probability_tables(VP8Context *s)
Definition: vp8.c:462
AV_PICTURE_TYPE_P
@ AV_PICTURE_TYPE_P
Predicted.
Definition: avutil.h:280
AVMEDIA_TYPE_VIDEO
@ AVMEDIA_TYPE_VIDEO
Definition: avutil.h:201
vp78_update_pred16x16_pred8x8_mvc_probabilities
static void vp78_update_pred16x16_pred8x8_mvc_probabilities(VP8Context *s, int mvc_size)
Definition: vp8.c:481
avpriv_request_sample
#define avpriv_request_sample(...)
Definition: tableprint_vlc.h:36
update_refs
static void update_refs(VP8Context *s)
Definition: vp8.c:501
vp7_decode_mv_mb_modes
static int vp7_decode_mv_mb_modes(AVCodecContext *avctx, VP8Frame *cur_frame, const VP8Frame *prev_frame)
Definition: vp8.c:2348
update_lf_deltas
static void update_lf_deltas(VP8Context *s)
Definition: vp8.c:317
filter_mb_row
static av_always_inline void filter_mb_row(AVCodecContext *avctx, void *tdata, int jobnr, int threadnr, int is_vp7)
Definition: vp8.c:2543
alpha
static const int16_t alpha[]
Definition: ilbcdata.h:55
AVPacket
This structure stores compressed data.
Definition: packet.h:499
AVCodecContext::priv_data
void * priv_data
Definition: avcodec.h:472
get_pixel_format
static enum AVPixelFormat get_pixel_format(VP8Context *s)
Definition: vp8.c:195
av_freep
#define av_freep(p)
Definition: tableprint_vlc.h:34
ff_vp8_decode_frame
int ff_vp8_decode_frame(AVCodecContext *avctx, AVFrame *frame, int *got_frame, AVPacket *avpkt)
Definition: vp8.c:2856
src
INIT_CLIP pixel * src
Definition: h264pred_template.c:418
VP7MVPred::score
uint8_t score
Definition: vp8data.h:65
VP8Context
Definition: vp8.h:160
HWACCEL_VAAPI
#define HWACCEL_VAAPI(codec)
Definition: hwconfig.h:70
DIAG_DOWN_LEFT_PRED
@ DIAG_DOWN_LEFT_PRED
Definition: vp9.h:49
vp8_find_free_buffer
static VP8Frame * vp8_find_free_buffer(VP8Context *s)
Definition: vp8.c:171
int32_t
int32_t
Definition: audioconvert.c:56
AV_CODEC_ID_VP8
@ AV_CODEC_ID_VP8
Definition: codec_id.h:192
AVFrame::linesize
int linesize[AV_NUM_DATA_POINTERS]
For video, a positive or negative value, which is typically indicating the size in bytes of each pict...
Definition: frame.h:389
coeff
static const double coeff[2][5]
Definition: vf_owdenoise.c:79
block
The exact code depends on how similar the blocks are and how related they are to the block
Definition: filter_design.txt:207
av_log
#define av_log(a,...)
Definition: tableprint_vlc.h:27
AV_WN64
#define AV_WN64(p, v)
Definition: intreadwrite.h:378
VP8_MVMODE_ZERO
@ VP8_MVMODE_ZERO
Definition: vp8.h:71
AVERROR_INVALIDDATA
#define AVERROR_INVALIDDATA
Invalid data found when processing input.
Definition: error.h:61
pthread_cond_init
static av_always_inline int pthread_cond_init(pthread_cond_t *cond, const pthread_condattr_t *attr)
Definition: os2threads.h:133
vp7_y2ac_qlookup
static const uint16_t vp7_y2ac_qlookup[]
Definition: vp8data.h:623
atomic_init
#define atomic_init(obj, value)
Definition: stdatomic.h:33
vp7_submv_prob
static const uint8_t vp7_submv_prob[3]
Definition: vp8data.h:149
AVDiscard
AVDiscard
Definition: defs.h:210
AVDISCARD_NONREF
@ AVDISCARD_NONREF
discard all non reference
Definition: defs.h:215
vp8_rac_get_coeff
static int vp8_rac_get_coeff(VPXRangeCoder *c, const uint8_t *prob)
Definition: vp8.c:72
vp8_dc_qlookup
static const uint8_t vp8_dc_qlookup[VP8_MAX_QUANT+1]
Definition: vp8data.h:518
copy_chroma
static void copy_chroma(AVFrame *dst, const AVFrame *src, int width, int height)
Definition: vp8.c:512
VP8_FRAME_PREVIOUS
@ VP8_FRAME_PREVIOUS
Definition: vp8.h:45
vpx_rac_get_prob
#define vpx_rac_get_prob
Definition: vpx_rac.h:82
vp8_decode_flush
static void vp8_decode_flush(AVCodecContext *avctx)
Definition: vp8.c:166
AVCodecContext::execute2
int(* execute2)(struct AVCodecContext *c, int(*func)(struct AVCodecContext *c2, void *arg, int jobnr, int threadnr), void *arg2, int *ret, int count)
The codec may call this to execute several independent things.
Definition: avcodec.h:1631
VP8_MVMODE_MV
@ VP8_MVMODE_MV
Definition: vp8.h:72
MARGIN
#define MARGIN
Definition: vp8.c:2309
ff_refstruct_unref
void ff_refstruct_unref(void *objp)
Decrement the reference count of the underlying object and automatically free the object if there are...
Definition: refstruct.c:120
vp8_alloc_frame
static int vp8_alloc_frame(VP8Context *s, VP8Frame *f, int ref)
Definition: vp8.c:103
get_submv_prob
static const av_always_inline uint8_t * get_submv_prob(uint32_t left, uint32_t top, int is_vp7)
Definition: vp8.c:932
ff_vp78dsp_init
av_cold void ff_vp78dsp_init(VP8DSPContext *dsp)
Definition: vp8dsp.c:668
VP8Frame::hwaccel_picture_private
void * hwaccel_picture_private
RefStruct reference.
Definition: vp8.h:156