FFmpeg  4.3
svq1enc.c
Go to the documentation of this file.
1 /*
2  * SVQ1 Encoder
3  * Copyright (C) 2004 Mike Melanson <melanson@pcisys.net>
4  *
5  * This file is part of FFmpeg.
6  *
7  * FFmpeg is free software; you can redistribute it and/or
8  * modify it under the terms of the GNU Lesser General Public
9  * License as published by the Free Software Foundation; either
10  * version 2.1 of the License, or (at your option) any later version.
11  *
12  * FFmpeg is distributed in the hope that it will be useful,
13  * but WITHOUT ANY WARRANTY; without even the implied warranty of
14  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
15  * Lesser General Public License for more details.
16  *
17  * You should have received a copy of the GNU Lesser General Public
18  * License along with FFmpeg; if not, write to the Free Software
19  * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
20  */
21 
22 /**
23  * @file
24  * Sorenson Vector Quantizer #1 (SVQ1) video codec.
25  * For more information of the SVQ1 algorithm, visit:
26  * http://www.pcisys.net/~melanson/codecs/
27  */
28 
29 #include "avcodec.h"
30 #include "hpeldsp.h"
31 #include "me_cmp.h"
32 #include "mpegvideo.h"
33 #include "h263.h"
34 #include "internal.h"
35 #include "mpegutils.h"
36 #include "packet_internal.h"
37 #include "svq1.h"
38 #include "svq1enc.h"
39 #include "svq1enc_cb.h"
40 #include "libavutil/avassert.h"
41 
42 
44 {
45  int i;
46 
47  /* frame code */
48  put_bits(&s->pb, 22, 0x20);
49 
50  /* temporal reference (sure hope this is a "don't care") */
51  put_bits(&s->pb, 8, 0x00);
52 
53  /* frame type */
54  put_bits(&s->pb, 2, frame_type - 1);
55 
57  /* no checksum since frame code is 0x20 */
58  /* no embedded string either */
59  /* output 5 unknown bits (2 + 2 + 1) */
60  put_bits(&s->pb, 5, 2); /* 2 needed by quicktime decoder */
61 
64  s->frame_width, s->frame_height);
65  put_bits(&s->pb, 3, i);
66 
67  if (i == 7) {
68  put_bits(&s->pb, 12, s->frame_width);
69  put_bits(&s->pb, 12, s->frame_height);
70  }
71  }
72 
73  /* no checksum or extra data (next 2 bits get 0) */
74  put_bits(&s->pb, 2, 0);
75 }
76 
77 #define QUALITY_THRESHOLD 100
78 #define THRESHOLD_MULTIPLIER 0.6
79 
80 static int ssd_int8_vs_int16_c(const int8_t *pix1, const int16_t *pix2,
81  intptr_t size)
82 {
83  int score = 0, i;
84 
85  for (i = 0; i < size; i++)
86  score += (pix1[i] - pix2[i]) * (pix1[i] - pix2[i]);
87  return score;
88 }
89 
91  uint8_t *decoded, int stride, int level,
92  int threshold, int lambda, int intra)
93 {
94  int count, y, x, i, j, split, best_mean, best_score, best_count;
95  int best_vector[6];
96  int block_sum[7] = { 0, 0, 0, 0, 0, 0 };
97  int w = 2 << (level + 2 >> 1);
98  int h = 2 << (level + 1 >> 1);
99  int size = w * h;
100  int16_t (*block)[256] = s->encoded_block_levels[level];
101  const int8_t *codebook_sum, *codebook;
102  const uint16_t(*mean_vlc)[2];
103  const uint8_t(*multistage_vlc)[2];
104 
105  best_score = 0;
106  // FIXME: Optimize, this does not need to be done multiple times.
107  if (intra) {
108  // level is 5 when encode_block is called from svq1_encode_plane
109  // and always < 4 when called recursively from this function.
110  codebook_sum = level < 4 ? svq1_intra_codebook_sum[level] : NULL;
111  codebook = ff_svq1_intra_codebooks[level];
112  mean_vlc = ff_svq1_intra_mean_vlc;
113  multistage_vlc = ff_svq1_intra_multistage_vlc[level];
114  for (y = 0; y < h; y++) {
115  for (x = 0; x < w; x++) {
116  int v = src[x + y * stride];
117  block[0][x + w * y] = v;
118  best_score += v * v;
119  block_sum[0] += v;
120  }
121  }
122  } else {
123  // level is 5 or < 4, see above for details.
124  codebook_sum = level < 4 ? svq1_inter_codebook_sum[level] : NULL;
125  codebook = ff_svq1_inter_codebooks[level];
126  mean_vlc = ff_svq1_inter_mean_vlc + 256;
127  multistage_vlc = ff_svq1_inter_multistage_vlc[level];
128  for (y = 0; y < h; y++) {
129  for (x = 0; x < w; x++) {
130  int v = src[x + y * stride] - ref[x + y * stride];
131  block[0][x + w * y] = v;
132  best_score += v * v;
133  block_sum[0] += v;
134  }
135  }
136  }
137 
138  best_count = 0;
139  best_score -= (int)((unsigned)block_sum[0] * block_sum[0] >> (level + 3));
140  best_mean = block_sum[0] + (size >> 1) >> (level + 3);
141 
142  if (level < 4) {
143  for (count = 1; count < 7; count++) {
144  int best_vector_score = INT_MAX;
145  int best_vector_sum = -999, best_vector_mean = -999;
146  const int stage = count - 1;
147  const int8_t *vector;
148 
149  for (i = 0; i < 16; i++) {
150  int sum = codebook_sum[stage * 16 + i];
151  int sqr, diff, score;
152 
153  vector = codebook + stage * size * 16 + i * size;
154  sqr = s->ssd_int8_vs_int16(vector, block[stage], size);
155  diff = block_sum[stage] - sum;
156  score = sqr - (diff * (int64_t)diff >> (level + 3)); // FIXME: 64 bits slooow
157  if (score < best_vector_score) {
158  int mean = diff + (size >> 1) >> (level + 3);
159  av_assert2(mean > -300 && mean < 300);
160  mean = av_clip(mean, intra ? 0 : -256, 255);
161  best_vector_score = score;
162  best_vector[stage] = i;
163  best_vector_sum = sum;
164  best_vector_mean = mean;
165  }
166  }
167  av_assert0(best_vector_mean != -999);
168  vector = codebook + stage * size * 16 + best_vector[stage] * size;
169  for (j = 0; j < size; j++)
170  block[stage + 1][j] = block[stage][j] - vector[j];
171  block_sum[stage + 1] = block_sum[stage] - best_vector_sum;
172  best_vector_score += lambda *
173  (+1 + 4 * count +
174  multistage_vlc[1 + count][1]
175  + mean_vlc[best_vector_mean][1]);
176 
177  if (best_vector_score < best_score) {
178  best_score = best_vector_score;
179  best_count = count;
180  best_mean = best_vector_mean;
181  }
182  }
183  }
184 
185  split = 0;
186  if (best_score > threshold && level) {
187  int score = 0;
188  int offset = level & 1 ? stride * h / 2 : w / 2;
189  PutBitContext backup[6];
190 
191  for (i = level - 1; i >= 0; i--)
192  backup[i] = s->reorder_pb[i];
193  score += encode_block(s, src, ref, decoded, stride, level - 1,
194  threshold >> 1, lambda, intra);
195  score += encode_block(s, src + offset, ref + offset, decoded + offset,
196  stride, level - 1, threshold >> 1, lambda, intra);
197  score += lambda;
198 
199  if (score < best_score) {
200  best_score = score;
201  split = 1;
202  } else {
203  for (i = level - 1; i >= 0; i--)
204  s->reorder_pb[i] = backup[i];
205  }
206  }
207  if (level > 0)
208  put_bits(&s->reorder_pb[level], 1, split);
209 
210  if (!split) {
211  av_assert1(best_mean >= 0 && best_mean < 256 || !intra);
212  av_assert1(best_mean >= -256 && best_mean < 256);
213  av_assert1(best_count >= 0 && best_count < 7);
214  av_assert1(level < 4 || best_count == 0);
215 
216  /* output the encoding */
217  put_bits(&s->reorder_pb[level],
218  multistage_vlc[1 + best_count][1],
219  multistage_vlc[1 + best_count][0]);
220  put_bits(&s->reorder_pb[level], mean_vlc[best_mean][1],
221  mean_vlc[best_mean][0]);
222 
223  for (i = 0; i < best_count; i++) {
224  av_assert2(best_vector[i] >= 0 && best_vector[i] < 16);
225  put_bits(&s->reorder_pb[level], 4, best_vector[i]);
226  }
227 
228  for (y = 0; y < h; y++)
229  for (x = 0; x < w; x++)
230  decoded[x + y * stride] = src[x + y * stride] -
231  block[best_count][x + w * y] +
232  best_mean;
233  }
234 
235  return best_score;
236 }
237 
239  s->block_index[0]= s->b8_stride*(s->mb_y*2 ) + s->mb_x*2;
240  s->block_index[1]= s->b8_stride*(s->mb_y*2 ) + 1 + s->mb_x*2;
241  s->block_index[2]= s->b8_stride*(s->mb_y*2 + 1) + s->mb_x*2;
242  s->block_index[3]= s->b8_stride*(s->mb_y*2 + 1) + 1 + s->mb_x*2;
243  s->block_index[4]= s->mb_stride*(s->mb_y + 1) + s->b8_stride*s->mb_height*2 + s->mb_x;
244  s->block_index[5]= s->mb_stride*(s->mb_y + s->mb_height + 2) + s->b8_stride*s->mb_height*2 + s->mb_x;
245 }
246 
247 static int svq1_encode_plane(SVQ1EncContext *s, int plane,
248  unsigned char *src_plane,
249  unsigned char *ref_plane,
250  unsigned char *decoded_plane,
251  int width, int height, int src_stride, int stride)
252 {
253  int x, y;
254  int i;
255  int block_width, block_height;
256  int level;
257  int threshold[6];
258  uint8_t *src = s->scratchbuf + stride * 32;
259  const int lambda = (s->quality * s->quality) >>
260  (2 * FF_LAMBDA_SHIFT);
261 
262  /* figure out the acceptable level thresholds in advance */
263  threshold[5] = QUALITY_THRESHOLD;
264  for (level = 4; level >= 0; level--)
265  threshold[level] = threshold[level + 1] * THRESHOLD_MULTIPLIER;
266 
267  block_width = (width + 15) / 16;
268  block_height = (height + 15) / 16;
269 
270  if (s->pict_type == AV_PICTURE_TYPE_P) {
271  s->m.avctx = s->avctx;
272  s->m.current_picture_ptr = &s->m.current_picture;
273  s->m.last_picture_ptr = &s->m.last_picture;
274  s->m.last_picture.f->data[0] = ref_plane;
275  s->m.linesize =
276  s->m.last_picture.f->linesize[0] =
277  s->m.new_picture.f->linesize[0] =
278  s->m.current_picture.f->linesize[0] = stride;
279  s->m.width = width;
280  s->m.height = height;
281  s->m.mb_width = block_width;
282  s->m.mb_height = block_height;
283  s->m.mb_stride = s->m.mb_width + 1;
284  s->m.b8_stride = 2 * s->m.mb_width + 1;
285  s->m.f_code = 1;
286  s->m.pict_type = s->pict_type;
287  s->m.motion_est = s->motion_est;
288  s->m.me.scene_change_score = 0;
289  // s->m.out_format = FMT_H263;
290  // s->m.unrestricted_mv = 1;
291  s->m.lambda = s->quality;
292  s->m.qscale = s->m.lambda * 139 +
293  FF_LAMBDA_SCALE * 64 >>
294  FF_LAMBDA_SHIFT + 7;
295  s->m.lambda2 = s->m.lambda * s->m.lambda +
296  FF_LAMBDA_SCALE / 2 >>
298 
299  if (!s->motion_val8[plane]) {
300  s->motion_val8[plane] = av_mallocz((s->m.b8_stride *
301  block_height * 2 + 2) *
302  2 * sizeof(int16_t));
303  s->motion_val16[plane] = av_mallocz((s->m.mb_stride *
304  (block_height + 2) + 1) *
305  2 * sizeof(int16_t));
306  if (!s->motion_val8[plane] || !s->motion_val16[plane])
307  return AVERROR(ENOMEM);
308  }
309 
310  s->m.mb_type = s->mb_type;
311 
312  // dummies, to avoid segfaults
313  s->m.current_picture.mb_mean = (uint8_t *)s->dummy;
314  s->m.current_picture.mb_var = (uint16_t *)s->dummy;
315  s->m.current_picture.mc_mb_var = (uint16_t *)s->dummy;
316  s->m.current_picture.mb_type = s->dummy;
317 
318  s->m.current_picture.motion_val[0] = s->motion_val8[plane] + 2;
319  s->m.p_mv_table = s->motion_val16[plane] +
320  s->m.mb_stride + 1;
321  s->m.mecc = s->mecc; // move
322  ff_init_me(&s->m);
323 
324  s->m.me.dia_size = s->avctx->dia_size;
325  s->m.first_slice_line = 1;
326  for (y = 0; y < block_height; y++) {
327  s->m.new_picture.f->data[0] = src - y * 16 * stride; // ugly
328  s->m.mb_y = y;
329 
330  for (i = 0; i < 16 && i + 16 * y < height; i++) {
331  memcpy(&src[i * stride], &src_plane[(i + 16 * y) * src_stride],
332  width);
333  for (x = width; x < 16 * block_width; x++)
334  src[i * stride + x] = src[i * stride + x - 1];
335  }
336  for (; i < 16 && i + 16 * y < 16 * block_height; i++)
337  memcpy(&src[i * stride], &src[(i - 1) * stride],
338  16 * block_width);
339 
340  for (x = 0; x < block_width; x++) {
341  s->m.mb_x = x;
342  init_block_index(&s->m);
343 
344  ff_estimate_p_frame_motion(&s->m, x, y);
345  }
346  s->m.first_slice_line = 0;
347  }
348 
350  ff_fix_long_mvs(&s->m, NULL, 0, s->m.p_mv_table, s->m.f_code,
352  }
353 
354  s->m.first_slice_line = 1;
355  for (y = 0; y < block_height; y++) {
356  for (i = 0; i < 16 && i + 16 * y < height; i++) {
357  memcpy(&src[i * stride], &src_plane[(i + 16 * y) * src_stride],
358  width);
359  for (x = width; x < 16 * block_width; x++)
360  src[i * stride + x] = src[i * stride + x - 1];
361  }
362  for (; i < 16 && i + 16 * y < 16 * block_height; i++)
363  memcpy(&src[i * stride], &src[(i - 1) * stride], 16 * block_width);
364 
365  s->m.mb_y = y;
366  for (x = 0; x < block_width; x++) {
367  uint8_t reorder_buffer[2][6][7 * 32];
368  int count[2][6];
369  int offset = y * 16 * stride + x * 16;
370  uint8_t *decoded = decoded_plane + offset;
371  uint8_t *ref = ref_plane + offset;
372  int score[4] = { 0, 0, 0, 0 }, best;
373  uint8_t *temp = s->scratchbuf;
374 
375  if (s->pb.buf_end - s->pb.buf -
376  (put_bits_count(&s->pb) >> 3) < 3000) { // FIXME: check size
377  av_log(s->avctx, AV_LOG_ERROR, "encoded frame too large\n");
378  return -1;
379  }
380 
381  s->m.mb_x = x;
382  init_block_index(&s->m);
383 
384  if (s->pict_type == AV_PICTURE_TYPE_I ||
385  (s->m.mb_type[x + y * s->m.mb_stride] &
387  for (i = 0; i < 6; i++)
388  init_put_bits(&s->reorder_pb[i], reorder_buffer[0][i],
389  7 * 32);
390  if (s->pict_type == AV_PICTURE_TYPE_P) {
392  put_bits(&s->reorder_pb[5], vlc[1], vlc[0]);
393  score[0] = vlc[1] * lambda;
394  }
395  score[0] += encode_block(s, src + 16 * x, NULL, temp, stride,
396  5, 64, lambda, 1);
397  for (i = 0; i < 6; i++) {
398  count[0][i] = put_bits_count(&s->reorder_pb[i]);
399  flush_put_bits(&s->reorder_pb[i]);
400  }
401  } else
402  score[0] = INT_MAX;
403 
404  best = 0;
405 
406  if (s->pict_type == AV_PICTURE_TYPE_P) {
408  int mx, my, pred_x, pred_y, dxy;
409  int16_t *motion_ptr;
410 
411  motion_ptr = ff_h263_pred_motion(&s->m, 0, 0, &pred_x, &pred_y);
412  if (s->m.mb_type[x + y * s->m.mb_stride] &
414  for (i = 0; i < 6; i++)
415  init_put_bits(&s->reorder_pb[i], reorder_buffer[1][i],
416  7 * 32);
417 
418  put_bits(&s->reorder_pb[5], vlc[1], vlc[0]);
419 
420  s->m.pb = s->reorder_pb[5];
421  mx = motion_ptr[0];
422  my = motion_ptr[1];
423  av_assert1(mx >= -32 && mx <= 31);
424  av_assert1(my >= -32 && my <= 31);
425  av_assert1(pred_x >= -32 && pred_x <= 31);
426  av_assert1(pred_y >= -32 && pred_y <= 31);
427  ff_h263_encode_motion(&s->m.pb, mx - pred_x, 1);
428  ff_h263_encode_motion(&s->m.pb, my - pred_y, 1);
429  s->reorder_pb[5] = s->m.pb;
430  score[1] += lambda * put_bits_count(&s->reorder_pb[5]);
431 
432  dxy = (mx & 1) + 2 * (my & 1);
433 
434  s->hdsp.put_pixels_tab[0][dxy](temp + 16*stride,
435  ref + (mx >> 1) +
436  stride * (my >> 1),
437  stride, 16);
438 
439  score[1] += encode_block(s, src + 16 * x, temp + 16*stride,
440  decoded, stride, 5, 64, lambda, 0);
441  best = score[1] <= score[0];
442 
444  score[2] = s->mecc.sse[0](NULL, src + 16 * x, ref,
445  stride, 16);
446  score[2] += vlc[1] * lambda;
447  if (score[2] < score[best] && mx == 0 && my == 0) {
448  best = 2;
449  s->hdsp.put_pixels_tab[0][0](decoded, ref, stride, 16);
450  put_bits(&s->pb, vlc[1], vlc[0]);
451  }
452  }
453 
454  if (best == 1) {
455  for (i = 0; i < 6; i++) {
456  count[1][i] = put_bits_count(&s->reorder_pb[i]);
457  flush_put_bits(&s->reorder_pb[i]);
458  }
459  } else {
460  motion_ptr[0] =
461  motion_ptr[1] =
462  motion_ptr[2] =
463  motion_ptr[3] =
464  motion_ptr[0 + 2 * s->m.b8_stride] =
465  motion_ptr[1 + 2 * s->m.b8_stride] =
466  motion_ptr[2 + 2 * s->m.b8_stride] =
467  motion_ptr[3 + 2 * s->m.b8_stride] = 0;
468  }
469  }
470 
471  s->rd_total += score[best];
472 
473  if (best != 2)
474  for (i = 5; i >= 0; i--)
475  avpriv_copy_bits(&s->pb, reorder_buffer[best][i],
476  count[best][i]);
477  if (best == 0)
478  s->hdsp.put_pixels_tab[0][0](decoded, temp, stride, 16);
479  }
480  s->m.first_slice_line = 0;
481  }
482  return 0;
483 }
484 
486 {
487  SVQ1EncContext *const s = avctx->priv_data;
488  int i;
489 
490  av_log(avctx, AV_LOG_DEBUG, "RD: %f\n",
491  s->rd_total / (double)(avctx->width * avctx->height *
492  avctx->frame_number));
493 
494  s->m.mb_type = NULL;
495  ff_mpv_common_end(&s->m);
496 
497  av_freep(&s->m.me.scratchpad);
498  av_freep(&s->m.me.map);
499  av_freep(&s->m.me.score_map);
500  av_freep(&s->mb_type);
501  av_freep(&s->dummy);
502  av_freep(&s->scratchbuf);
503 
504  for (i = 0; i < 3; i++) {
505  av_freep(&s->motion_val8[i]);
506  av_freep(&s->motion_val16[i]);
507  }
508 
509  av_frame_free(&s->current_picture);
510  av_frame_free(&s->last_picture);
511 
512  return 0;
513 }
514 
516 {
517  SVQ1EncContext *const s = avctx->priv_data;
518  int ret;
519 
520  if (avctx->width >= 4096 || avctx->height >= 4096) {
521  av_log(avctx, AV_LOG_ERROR, "Dimensions too large, maximum is 4095x4095\n");
522  return AVERROR(EINVAL);
523  }
524 
525  ff_hpeldsp_init(&s->hdsp, avctx->flags);
526  ff_me_cmp_init(&s->mecc, avctx);
527  ff_mpegvideoencdsp_init(&s->m.mpvencdsp, avctx);
528 
529  s->current_picture = av_frame_alloc();
530  s->last_picture = av_frame_alloc();
531  if (!s->current_picture || !s->last_picture) {
532  svq1_encode_end(avctx);
533  return AVERROR(ENOMEM);
534  }
535 
536  s->frame_width = avctx->width;
537  s->frame_height = avctx->height;
538 
539  s->y_block_width = (s->frame_width + 15) / 16;
540  s->y_block_height = (s->frame_height + 15) / 16;
541 
542  s->c_block_width = (s->frame_width / 4 + 15) / 16;
543  s->c_block_height = (s->frame_height / 4 + 15) / 16;
544 
545  s->avctx = avctx;
546  s->m.avctx = avctx;
547 
548  if ((ret = ff_mpv_common_init(&s->m)) < 0) {
549  svq1_encode_end(avctx);
550  return ret;
551  }
552 
553  s->m.picture_structure = PICT_FRAME;
554  s->m.me.temp =
555  s->m.me.scratchpad = av_mallocz((avctx->width + 64) *
556  2 * 16 * 2 * sizeof(uint8_t));
557  s->m.me.map = av_mallocz(ME_MAP_SIZE * sizeof(uint32_t));
558  s->m.me.score_map = av_mallocz(ME_MAP_SIZE * sizeof(uint32_t));
559  s->mb_type = av_mallocz((s->y_block_width + 1) *
560  s->y_block_height * sizeof(int16_t));
561  s->dummy = av_mallocz((s->y_block_width + 1) *
562  s->y_block_height * sizeof(int32_t));
563  s->ssd_int8_vs_int16 = ssd_int8_vs_int16_c;
564 
565  if (!s->m.me.temp || !s->m.me.scratchpad || !s->m.me.map ||
566  !s->m.me.score_map || !s->mb_type || !s->dummy) {
567  svq1_encode_end(avctx);
568  return AVERROR(ENOMEM);
569  }
570 
571  if (ARCH_PPC)
573  if (ARCH_X86)
575 
576  ff_h263_encode_init(&s->m); // mv_penalty
577 
578  return 0;
579 }
580 
582  const AVFrame *pict, int *got_packet)
583 {
584  SVQ1EncContext *const s = avctx->priv_data;
585  int i, ret;
586 
587  if ((ret = ff_alloc_packet2(avctx, pkt, s->y_block_width * s->y_block_height *
589  return ret;
590 
591  if (avctx->pix_fmt != AV_PIX_FMT_YUV410P) {
592  av_log(avctx, AV_LOG_ERROR, "unsupported pixel format\n");
593  return -1;
594  }
595 
596  if (!s->current_picture->data[0]) {
597  if ((ret = ff_get_buffer(avctx, s->current_picture, 0)) < 0) {
598  return ret;
599  }
600  }
601  if (!s->last_picture->data[0]) {
602  ret = ff_get_buffer(avctx, s->last_picture, 0);
603  if (ret < 0)
604  return ret;
605  }
606  if (!s->scratchbuf) {
607  s->scratchbuf = av_malloc_array(s->current_picture->linesize[0], 16 * 3);
608  if (!s->scratchbuf)
609  return AVERROR(ENOMEM);
610  }
611 
612  FFSWAP(AVFrame*, s->current_picture, s->last_picture);
613 
614  init_put_bits(&s->pb, pkt->data, pkt->size);
615 
616  if (avctx->gop_size && (avctx->frame_number % avctx->gop_size))
617  s->pict_type = AV_PICTURE_TYPE_P;
618  else
619  s->pict_type = AV_PICTURE_TYPE_I;
620  s->quality = pict->quality;
621 
622 #if FF_API_CODED_FRAME
624  avctx->coded_frame->pict_type = s->pict_type;
625  avctx->coded_frame->key_frame = s->pict_type == AV_PICTURE_TYPE_I;
627 #endif
628 
629  ff_side_data_set_encoder_stats(pkt, pict->quality, NULL, 0, s->pict_type);
630 
631  svq1_write_header(s, s->pict_type);
632  for (i = 0; i < 3; i++) {
633  int ret = svq1_encode_plane(s, i,
634  pict->data[i],
635  s->last_picture->data[i],
636  s->current_picture->data[i],
637  s->frame_width / (i ? 4 : 1),
638  s->frame_height / (i ? 4 : 1),
639  pict->linesize[i],
640  s->current_picture->linesize[i]);
641  emms_c();
642  if (ret < 0) {
643  int j;
644  for (j = 0; j < i; j++) {
645  av_freep(&s->motion_val8[j]);
646  av_freep(&s->motion_val16[j]);
647  }
648  av_freep(&s->scratchbuf);
649  return -1;
650  }
651  }
652 
653  // avpriv_align_put_bits(&s->pb);
654  while (put_bits_count(&s->pb) & 31)
655  put_bits(&s->pb, 1, 0);
656 
657  flush_put_bits(&s->pb);
658 
659  pkt->size = put_bits_count(&s->pb) / 8;
660  if (s->pict_type == AV_PICTURE_TYPE_I)
662  *got_packet = 1;
663 
664  return 0;
665 }
666 
667 #define OFFSET(x) offsetof(struct SVQ1EncContext, x)
668 #define VE AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_ENCODING_PARAM
669 static const AVOption options[] = {
670  { "motion-est", "Motion estimation algorithm", OFFSET(motion_est), AV_OPT_TYPE_INT, { .i64 = FF_ME_EPZS }, FF_ME_ZERO, FF_ME_XONE, VE, "motion-est"},
671  { "zero", NULL, 0, AV_OPT_TYPE_CONST, { .i64 = FF_ME_ZERO }, 0, 0, FF_MPV_OPT_FLAGS, "motion-est" },
672  { "epzs", NULL, 0, AV_OPT_TYPE_CONST, { .i64 = FF_ME_EPZS }, 0, 0, FF_MPV_OPT_FLAGS, "motion-est" },
673  { "xone", NULL, 0, AV_OPT_TYPE_CONST, { .i64 = FF_ME_XONE }, 0, 0, FF_MPV_OPT_FLAGS, "motion-est" },
674 
675  { NULL },
676 };
677 
678 static const AVClass svq1enc_class = {
679  .class_name = "svq1enc",
680  .item_name = av_default_item_name,
681  .option = options,
682  .version = LIBAVUTIL_VERSION_INT,
683 };
684 
686  .name = "svq1",
687  .long_name = NULL_IF_CONFIG_SMALL("Sorenson Vector Quantizer 1 / Sorenson Video 1 / SVQ1"),
688  .type = AVMEDIA_TYPE_VIDEO,
689  .id = AV_CODEC_ID_SVQ1,
690  .priv_data_size = sizeof(SVQ1EncContext),
691  .priv_class = &svq1enc_class,
693  .encode2 = svq1_encode_frame,
694  .close = svq1_encode_end,
695  .pix_fmts = (const enum AVPixelFormat[]) { AV_PIX_FMT_YUV410P,
696  AV_PIX_FMT_NONE },
697 };
PICT_FRAME
#define PICT_FRAME
Definition: mpegutils.h:39
ff_mpv_common_init
av_cold int ff_mpv_common_init(MpegEncContext *s)
init common structure for both encoder and decoder.
Definition: mpegvideo.c:890
AVCodec
AVCodec.
Definition: codec.h:190
FF_ENABLE_DEPRECATION_WARNINGS
#define FF_ENABLE_DEPRECATION_WARNINGS
Definition: internal.h:85
stride
int stride
Definition: mace.c:144
AVPixelFormat
AVPixelFormat
Pixel format.
Definition: pixfmt.h:64
level
uint8_t level
Definition: svq3.c:209
THRESHOLD_MULTIPLIER
#define THRESHOLD_MULTIPLIER
Definition: svq1enc.c:78
ARCH_X86
#define ARCH_X86
Definition: config.h:38
init
static av_cold int init(AVCodecContext *avctx)
Definition: avrndec.c:35
FF_LAMBDA_SCALE
#define FF_LAMBDA_SCALE
Definition: avutil.h:226
AVERROR
Filter the word “frame” indicates either a video frame or a group of audio as stored in an AVFrame structure Format for each input and each output the list of supported formats For video that means pixel format For audio that means channel sample they are references to shared objects When the negotiation mechanism computes the intersection of the formats supported at each end of a all references to both lists are replaced with a reference to the intersection And when a single format is eventually chosen for a link amongst the remaining all references to the list are updated That means that if a filter requires that its input and output have the same format amongst a supported all it has to do is use a reference to the same list of formats query_formats can leave some formats unset and return AVERROR(EAGAIN) to cause the negotiation mechanism toagain later. That can be used by filters with complex requirements to use the format negotiated on one link to set the formats supported on another. Frame references ownership and permissions
svq1enc_cb.h
svq1_encode_plane
static int svq1_encode_plane(SVQ1EncContext *s, int plane, unsigned char *src_plane, unsigned char *ref_plane, unsigned char *decoded_plane, int width, int height, int src_stride, int stride)
Definition: svq1enc.c:247
FFSWAP
#define FFSWAP(type, a, b)
Definition: common.h:99
ff_fix_long_p_mvs
void ff_fix_long_p_mvs(MpegEncContext *s, int type)
Definition: motion_est.c:1651
FF_ME_EPZS
#define FF_ME_EPZS
Definition: motion_est.h:41
ff_side_data_set_encoder_stats
int ff_side_data_set_encoder_stats(AVPacket *pkt, int quality, int64_t *error, int error_count, int pict_type)
Definition: avpacket.c:728
SVQ1EncContext
Definition: svq1enc.h:34
init_put_bits
static void init_put_bits(PutBitContext *s, uint8_t *buffer, int buffer_size)
Initialize the PutBitContext s.
Definition: put_bits.h:48
ff_svq1_intra_codebooks
const int8_t *const ff_svq1_intra_codebooks[6]
Definition: svq1_cb.h:1519
av_frame_free
void av_frame_free(AVFrame **frame)
Free the frame and any dynamically allocated objects in it, e.g.
Definition: frame.c:203
AVFrame
This structure describes decoded (raw) audio or video data.
Definition: frame.h:300
put_bits
static void put_bits(Jpeg2000EncoderContext *s, int val, int n)
put n times val bit
Definition: j2kenc.c:208
internal.h
AVPacket::data
uint8_t * data
Definition: packet.h:355
AVOption
AVOption.
Definition: opt.h:246
mpegvideo.h
FF_LAMBDA_SHIFT
#define FF_LAMBDA_SHIFT
Definition: avutil.h:225
CANDIDATE_MB_TYPE_INTER
#define CANDIDATE_MB_TYPE_INTER
Definition: mpegutils.h:105
mpegutils.h
AV_PKT_FLAG_KEY
#define AV_PKT_FLAG_KEY
The packet contains a keyframe.
Definition: packet.h:388
svq1_inter_codebook_sum
static const int8_t svq1_inter_codebook_sum[4][16 *6]
Definition: svq1enc_cb.h:32
AVFrame::data
uint8_t * data[AV_NUM_DATA_POINTERS]
pointer to the picture/channel planes.
Definition: frame.h:314
ff_svq1_intra_multistage_vlc
const uint8_t ff_svq1_intra_multistage_vlc[6][8][2]
Definition: svq1_vlc.h:33
ff_mpegvideoencdsp_init
av_cold void ff_mpegvideoencdsp_init(MpegvideoEncDSPContext *c, AVCodecContext *avctx)
Definition: mpegvideoencdsp.c:232
ff_svq1_inter_mean_vlc
const uint16_t ff_svq1_inter_mean_vlc[512][2]
Definition: svq1_vlc.h:136
ff_match_2uint16
int ff_match_2uint16(const uint16_t(*tab)[2], int size, int a, int b)
Return the index into tab at which {a,b} match elements {[0],[1]} of tab.
Definition: utils.c:1796
ff_h263_pred_motion
int16_t * ff_h263_pred_motion(MpegEncContext *s, int block, int dir, int *px, int *py)
Definition: h263.c:307
x
FFmpeg Automated Testing Environment ************************************Introduction Using FATE from your FFmpeg source directory Submitting the results to the FFmpeg result aggregation server Uploading new samples to the fate suite FATE makefile targets and variables Makefile targets Makefile variables Examples Introduction **************FATE is an extended regression suite on the client side and a means for results aggregation and presentation on the server side The first part of this document explains how you can use FATE from your FFmpeg source directory to test your ffmpeg binary The second part describes how you can run FATE to submit the results to FFmpeg’s FATE server In any way you can have a look at the publicly viewable FATE results by visiting this as it can be seen if some test on some platform broke with their recent contribution This usually happens on the platforms the developers could not test on The second part of this document describes how you can run FATE to submit your results to FFmpeg’s FATE server If you want to submit your results be sure to check that your combination of OS and compiler is not already listed on the above mentioned website In the third part you can find a comprehensive listing of FATE makefile targets and variables Using FATE from your FFmpeg source directory **********************************************If you want to run FATE on your machine you need to have the samples in place You can get the samples via the build target fate rsync Use this command from the top level source this will cause FATE to fail NOTE To use a custom wrapper to run the pass ‘ target exec’ to ‘configure’ or set the TARGET_EXEC Make variable Submitting the results to the FFmpeg result aggregation server ****************************************************************To submit your results to the server you should run fate through the shell script ‘tests fate sh’ from the FFmpeg sources This script needs to be invoked with a configuration file as its first argument tests fate sh path to fate_config A configuration file template with comments describing the individual configuration variables can be found at ‘doc fate_config sh template’ Create a configuration that suits your based on the configuration template The ‘slot’ configuration variable can be any string that is not yet but it is suggested that you name it adhering to the following pattern ‘ARCH OS COMPILER COMPILER VERSION’ The configuration file itself will be sourced in a shell therefore all shell features may be used This enables you to setup the environment as you need it for your build For your first test runs the ‘fate_recv’ variable should be empty or commented out This will run everything as normal except that it will omit the submission of the results to the server The following files should be present in $workdir as specified in the configuration it may help to try out the ‘ssh’ command with one or more ‘ v’ options You should get detailed output concerning your SSH configuration and the authentication process The only thing left is to automate the execution of the fate sh script and the synchronisation of the samples directory Uploading new samples to the fate suite *****************************************If you need a sample uploaded send a mail to samples request This is for developers who have an account on the fate suite server If you upload new please make sure they are as small as space on each network bandwidth and so on benefit from smaller test cases Also keep in mind older checkouts use existing sample that means in practice generally do not remove or overwrite files as it likely would break older checkouts or releases Also all needed samples for a commit should be ideally before the push If you need an account for frequently uploading samples or you wish to help others by doing that send a mail to ffmpeg devel rsync vauL Duo x
Definition: fate.txt:150
ff_me_cmp_init
av_cold void ff_me_cmp_init(MECmpContext *c, AVCodecContext *avctx)
Definition: me_cmp.c:1035
AVFrame::key_frame
int key_frame
1 -> keyframe, 0-> not
Definition: frame.h:378
AVCodecContext::flags
int flags
AV_CODEC_FLAG_*.
Definition: avcodec.h:606
init_block_index
static void init_block_index(MpegEncContext *s)
Definition: svq1enc.c:238
ssd_int8_vs_int16_c
static int ssd_int8_vs_int16_c(const int8_t *pix1, const int16_t *pix2, intptr_t size)
Definition: svq1enc.c:80
av_frame_alloc
AVFrame * av_frame_alloc(void)
Allocate an AVFrame and set its fields to default values.
Definition: frame.c:190
svq1_encode_init
static av_cold int svq1_encode_init(AVCodecContext *avctx)
Definition: svq1enc.c:515
ff_mpv_common_end
void ff_mpv_common_end(MpegEncContext *s)
Definition: mpegvideo.c:1131
avassert.h
AV_LOG_ERROR
#define AV_LOG_ERROR
Something went wrong and cannot losslessly be recovered.
Definition: log.h:176
av_cold
#define av_cold
Definition: attributes.h:90
ff_h263_encode_init
void ff_h263_encode_init(MpegEncContext *s)
Definition: ituh263enc.c:761
width
#define width
emms_c
#define emms_c()
Definition: internal.h:55
MAX_MB_BYTES
#define MAX_MB_BYTES
Definition: mpegutils.h:47
s
#define s(width, name)
Definition: cbs_vp9.c:257
av_assert0
#define av_assert0(cond)
assert() equivalent, that is always enabled.
Definition: avassert.h:37
pix_fmts
static enum AVPixelFormat pix_fmts[]
Definition: libkvazaar.c:275
AV_LOG_DEBUG
#define AV_LOG_DEBUG
Stuff which is only useful for libav* developers.
Definition: log.h:197
av_mallocz
void * av_mallocz(size_t size)
Allocate a memory block with alignment suitable for all memory accesses (including vectors if availab...
Definition: mem.c:237
AV_INPUT_BUFFER_MIN_SIZE
#define AV_INPUT_BUFFER_MIN_SIZE
Definition: avcodec.h:222
PutBitContext
Definition: put_bits.h:35
ff_hpeldsp_init
av_cold void ff_hpeldsp_init(HpelDSPContext *c, int flags)
Definition: hpeldsp.c:338
int32_t
int32_t
Definition: audio_convert.c:194
encode_block
static int encode_block(SVQ1EncContext *s, uint8_t *src, uint8_t *ref, uint8_t *decoded, int stride, int level, int threshold, int lambda, int intra)
Definition: svq1enc.c:90
LIBAVUTIL_VERSION_INT
#define LIBAVUTIL_VERSION_INT
Definition: version.h:85
AVClass
Describe the class of an AVClass context structure.
Definition: log.h:67
NULL
#define NULL
Definition: coverity.c:32
SVQ1_BLOCK_INTRA
#define SVQ1_BLOCK_INTRA
Definition: svq1.h:43
VE
#define VE
Definition: svq1enc.c:668
av_default_item_name
const char * av_default_item_name(void *ptr)
Return the context name.
Definition: log.c:235
AV_PICTURE_TYPE_I
@ AV_PICTURE_TYPE_I
Intra.
Definition: avutil.h:274
src
#define src
Definition: vp8dsp.c:254
SVQ1_BLOCK_SKIP
#define SVQ1_BLOCK_SKIP
Definition: svq1.h:40
ff_svq1_encoder
AVCodec ff_svq1_encoder
Definition: svq1enc.c:685
ME_MAP_SIZE
#define ME_MAP_SIZE
Definition: motion_est.h:38
FF_ME_XONE
#define FF_ME_XONE
Definition: motion_est.h:42
svq1_encode_end
static av_cold int svq1_encode_end(AVCodecContext *avctx)
Definition: svq1enc.c:485
avpriv_copy_bits
void avpriv_copy_bits(PutBitContext *pb, const uint8_t *src, int length)
Copy the content of src to the bitstream.
Definition: bitstream.c:64
ff_svq1_inter_codebooks
const int8_t *const ff_svq1_inter_codebooks[6]
Definition: svq1_cb.h:776
AVFrame::pict_type
enum AVPictureType pict_type
Picture type of the frame.
Definition: frame.h:383
ff_get_buffer
int ff_get_buffer(AVCodecContext *avctx, AVFrame *frame, int flags)
Get a buffer for a frame.
Definition: decode.c:1854
AVPacket::size
int size
Definition: packet.h:356
NULL_IF_CONFIG_SMALL
#define NULL_IF_CONFIG_SMALL(x)
Return NULL if CONFIG_SMALL is true, otherwise the argument without modification.
Definition: internal.h:186
AVCodecContext::gop_size
int gop_size
the number of pictures in a group of pictures, or 0 for intra_only
Definition: avcodec.h:721
AVFrame::quality
int quality
quality (between 1 (good) and FF_LAMBDA_MAX (bad))
Definition: frame.h:423
size
int size
Definition: twinvq_data.h:11134
split
static char * split(char *message, char delim)
Definition: af_channelmap.c:81
height
#define height
offset
it s the only field you need to keep assuming you have a context There is some magic you don t need to care about around this just let it vf offset
Definition: writing_filters.txt:86
AVPacket::flags
int flags
A combination of AV_PKT_FLAG values.
Definition: packet.h:361
AV_CODEC_ID_SVQ1
@ AV_CODEC_ID_SVQ1
Definition: codec_id.h:71
svq1enc.h
av_assert2
#define av_assert2(cond)
assert() equivalent, that does lie in speed critical code.
Definition: avassert.h:64
svq1_write_header
static void svq1_write_header(SVQ1EncContext *s, int frame_type)
Definition: svq1enc.c:43
i
#define i(width, name, range_min, range_max)
Definition: cbs_h2645.c:269
put_bits_count
static int put_bits_count(PutBitContext *s)
Definition: put_bits.h:67
frame_type
frame_type
Definition: jpeg2000_parser.c:31
ff_svq1enc_init_x86
void ff_svq1enc_init_x86(SVQ1EncContext *c)
Definition: svq1enc_init.c:32
av_malloc_array
#define av_malloc_array(a, b)
Definition: tableprint_vlc.h:32
av_assert1
#define av_assert1(cond)
assert() equivalent, that does not lie in speed critical code.
Definition: avassert.h:53
uint8_t
uint8_t
Definition: audio_convert.c:194
AVCodec::name
const char * name
Name of the codec implementation.
Definition: codec.h:197
ff_init_me
int ff_init_me(MpegEncContext *s)
Definition: motion_est.c:306
AVCodecContext::height
int height
Definition: avcodec.h:699
AVCodecContext::pix_fmt
enum AVPixelFormat pix_fmt
Pixel format, see AV_PIX_FMT_xxx.
Definition: avcodec.h:736
SVQ1_BLOCK_INTER
#define SVQ1_BLOCK_INTER
Definition: svq1.h:41
svq1_intra_codebook_sum
static const int8_t svq1_intra_codebook_sum[4][16 *6]
Definition: svq1enc_cb.h:59
avcodec.h
ff_svq1_inter_multistage_vlc
const uint8_t ff_svq1_inter_multistage_vlc[6][8][2]
Definition: svq1_vlc.h:50
ret
ret
Definition: filter_design.txt:187
AVClass::class_name
const char * class_name
The name of the class; usually it is the same name as the context structure type to which the AVClass...
Definition: log.h:72
w
FFmpeg Automated Testing Environment ************************************Introduction Using FATE from your FFmpeg source directory Submitting the results to the FFmpeg result aggregation server Uploading new samples to the fate suite FATE makefile targets and variables Makefile targets Makefile variables Examples Introduction **************FATE is an extended regression suite on the client side and a means for results aggregation and presentation on the server side The first part of this document explains how you can use FATE from your FFmpeg source directory to test your ffmpeg binary The second part describes how you can run FATE to submit the results to FFmpeg’s FATE server In any way you can have a look at the publicly viewable FATE results by visiting this as it can be seen if some test on some platform broke with their recent contribution This usually happens on the platforms the developers could not test on The second part of this document describes how you can run FATE to submit your results to FFmpeg’s FATE server If you want to submit your results be sure to check that your combination of OS and compiler is not already listed on the above mentioned website In the third part you can find a comprehensive listing of FATE makefile targets and variables Using FATE from your FFmpeg source directory **********************************************If you want to run FATE on your machine you need to have the samples in place You can get the samples via the build target fate rsync Use this command from the top level source this will cause FATE to fail NOTE To use a custom wrapper to run the pass ‘ target exec’ to ‘configure’ or set the TARGET_EXEC Make variable Submitting the results to the FFmpeg result aggregation server ****************************************************************To submit your results to the server you should run fate through the shell script ‘tests fate sh’ from the FFmpeg sources This script needs to be invoked with a configuration file as its first argument tests fate sh path to fate_config A configuration file template with comments describing the individual configuration variables can be found at ‘doc fate_config sh template’ Create a configuration that suits your based on the configuration template The ‘slot’ configuration variable can be any string that is not yet but it is suggested that you name it adhering to the following pattern ‘ARCH OS COMPILER COMPILER VERSION’ The configuration file itself will be sourced in a shell therefore all shell features may be used This enables you to setup the environment as you need it for your build For your first test runs the ‘fate_recv’ variable should be empty or commented out This will run everything as normal except that it will omit the submission of the results to the server The following files should be present in $workdir as specified in the configuration it may help to try out the ‘ssh’ command with one or more ‘ v’ options You should get detailed output concerning your SSH configuration and the authentication process The only thing left is to automate the execution of the fate sh script and the synchronisation of the samples directory Uploading new samples to the fate suite *****************************************If you need a sample uploaded send a mail to samples request This is for developers who have an account on the fate suite server If you upload new please make sure they are as small as space on each network bandwidth and so on benefit from smaller test cases Also keep in mind older checkouts use existing sample that means in practice generally do not remove or overwrite files as it likely would break older checkouts or releases Also all needed samples for a commit should be ideally before the push If you need an account for frequently uploading samples or you wish to help others by doing that send a mail to ffmpeg devel rsync vauL Duo ug o o w
Definition: fate.txt:150
options
static const AVOption options[]
Definition: svq1enc.c:669
me_cmp.h
FF_ARRAY_ELEMS
#define FF_ARRAY_ELEMS(a)
Definition: sinewin_tablegen_template.c:38
FF_MPV_OPT_FLAGS
#define FF_MPV_OPT_FLAGS
Definition: mpegvideo.h:615
ff_fix_long_mvs
void ff_fix_long_mvs(MpegEncContext *s, uint8_t *field_select_table, int field_select, int16_t(*mv_table)[2], int f_code, int type, int truncate)
Definition: motion_est.c:1700
AVCodecContext::coded_frame
attribute_deprecated AVFrame * coded_frame
the picture in the bitstream
Definition: avcodec.h:1776
AVCodecContext
main external API structure.
Definition: avcodec.h:526
ff_estimate_p_frame_motion
void ff_estimate_p_frame_motion(MpegEncContext *s, int mb_x, int mb_y)
Definition: motion_est.c:885
ff_svq1_block_type_vlc
const uint8_t ff_svq1_block_type_vlc[4][2]
Definition: svq1_vlc.h:27
ARCH_PPC
#define ARCH_PPC
Definition: config.h:29
pkt
static AVPacket pkt
Definition: demuxing_decoding.c:54
ff_svq1_intra_mean_vlc
const uint16_t ff_svq1_intra_mean_vlc[256][2]
Definition: svq1_vlc.h:67
CANDIDATE_MB_TYPE_INTRA
#define CANDIDATE_MB_TYPE_INTRA
Definition: mpegutils.h:104
AV_PIX_FMT_NONE
@ AV_PIX_FMT_NONE
Definition: pixfmt.h:65
AV_OPT_TYPE_INT
@ AV_OPT_TYPE_INT
Definition: opt.h:223
svq1.h
ref
static int ref[MAX_W *MAX_W]
Definition: jpeg2000dwt.c:107
temp
else temp
Definition: vf_mcdeint.c:256
FF_DISABLE_DEPRECATION_WARNINGS
#define FF_DISABLE_DEPRECATION_WARNINGS
Definition: internal.h:84
AV_PICTURE_TYPE_P
@ AV_PICTURE_TYPE_P
Predicted.
Definition: avutil.h:275
AVMEDIA_TYPE_VIDEO
@ AVMEDIA_TYPE_VIDEO
Definition: avutil.h:201
packet_internal.h
AVCodecContext::frame_number
int frame_number
Frame counter, set by libavcodec.
Definition: avcodec.h:1217
flush_put_bits
static void flush_put_bits(PutBitContext *s)
Pad the end of the output stream with zeros.
Definition: put_bits.h:101
diff
static av_always_inline int diff(const uint32_t a, const uint32_t b)
Definition: vf_palettegen.c:136
AVPacket
This structure stores compressed data.
Definition: packet.h:332
AVCodecContext::priv_data
void * priv_data
Definition: avcodec.h:553
av_freep
#define av_freep(p)
Definition: tableprint_vlc.h:35
AVCodecContext::width
int width
picture width / height.
Definition: avcodec.h:699
OFFSET
#define OFFSET(x)
Definition: svq1enc.c:667
hpeldsp.h
AVFrame::linesize
int linesize[AV_NUM_DATA_POINTERS]
For video, size in bytes of each picture line.
Definition: frame.h:331
AV_PIX_FMT_YUV410P
@ AV_PIX_FMT_YUV410P
planar YUV 4:1:0, 9bpp, (1 Cr & Cb sample per 4x4 Y samples)
Definition: pixfmt.h:72
block
The exact code depends on how similar the blocks are and how related they are to the block
Definition: filter_design.txt:207
av_log
#define av_log(a,...)
Definition: tableprint_vlc.h:28
svq1_encode_frame
static int svq1_encode_frame(AVCodecContext *avctx, AVPacket *pkt, const AVFrame *pict, int *got_packet)
Definition: svq1enc.c:581
h
h
Definition: vp9dsp_template.c:2038
ff_svq1enc_init_ppc
av_cold void ff_svq1enc_init_ppc(SVQ1EncContext *c)
Definition: svq1enc_altivec.c:74
svq1enc_class
static const AVClass svq1enc_class
Definition: svq1enc.c:678
ff_alloc_packet2
int ff_alloc_packet2(AVCodecContext *avctx, AVPacket *avpkt, int64_t size, int64_t min_size)
Check AVPacket size and/or allocate data.
Definition: encode.c:32
int
int
Definition: ffmpeg_filter.c:192
ff_svq1_frame_size_table
const uint16_t ff_svq1_frame_size_table[7][2]
Definition: svq1.c:40
AV_OPT_TYPE_CONST
@ AV_OPT_TYPE_CONST
Definition: opt.h:232
MpegEncContext
MpegEncContext.
Definition: mpegvideo.h:81
QUALITY_THRESHOLD
#define QUALITY_THRESHOLD
Definition: svq1enc.c:77
ff_h263_encode_motion
void ff_h263_encode_motion(PutBitContext *pb, int val, int f_code)
Definition: ituh263enc.c:646
FF_ME_ZERO
#define FF_ME_ZERO
Definition: motion_est.h:40
h263.h