Go to the documentation of this file.
65 #define ME_CACHE_SIZE 1024
76 #define PTR_ADD(ptr, off) ((ptr) ? (ptr) + (off) : NULL)
79 uint8_t *
const ref[3], uint8_t *
const ref2[3],
80 int x,
int y,
int ref_index)
85 ((y*
c->uvstride + x) >>
s->chroma_h_shift),
86 ((y*
c->uvstride + x) >>
s->chroma_h_shift),
88 for (
int i = 0;
i < 3;
i++) {
100 const int el =
FFMIN(e, 10);
105 for (
i = 0;
i < el;
i++)
111 for (
i = e - 1;
i >= el;
i--)
138 for (
int i =
log2 - 1;
i >= 0;
i--)
152 for (
int i = 0;
frame->data[
i];
i++) {
158 frame->width =
s->avctx->width;
159 frame->height =
s->avctx->height;
169 int plane_index,
ret;
179 s->spatial_decomposition_type = enc->
pred;
184 for(plane_index=0; plane_index<3; plane_index++){
185 s->plane[plane_index].diag_mc= 1;
186 s->plane[plane_index].htaps= 6;
187 s->plane[plane_index].hcoeff[0]= 40;
188 s->plane[plane_index].hcoeff[1]= -10;
189 s->plane[plane_index].hcoeff[2]= 2;
190 s->plane[plane_index].fast_mc= 1;
200 enc->qdsp.put_qpel_pixels_tab [0][dy+dx/4]=\
201 enc->qdsp.put_no_rnd_qpel_pixels_tab[0][dy+dx/4]=\
202 s->h264qpel.put_h264_qpel_pixels_tab[0][dy+dx/4];\
203 enc->qdsp.put_qpel_pixels_tab [1][dy+dx/4]=\
204 enc->qdsp.put_no_rnd_qpel_pixels_tab[1][dy+dx/4]=\
205 s->h264qpel.put_h264_qpel_pixels_tab[1][dy+dx/4];
270 s->colorspace_type= 0;
274 s->colorspace_type = 1;
287 if (!
s->input_picture)
298 int size=
s->b_width *
s->b_height << 2*
s->block_max_depth;
299 for(
i=0;
i<
s->max_ref_frames;
i++){
302 if (!
s->ref_mvs[
i] || !
s->ref_scores[
i])
316 for (
i = 0;
i <
h;
i++) {
317 for (j = 0; j <
w; j++) {
321 pix += line_size -
w;
333 for (
i = 0;
i <
w;
i++) {
334 for (j = 0; j <
w; j ++) {
338 pix += line_size -
w;
370 #define P_TOPRIGHT P[3]
371 #define P_MEDIAN P[4]
373 #define FLAG_QPEL 1 //must be 1
379 uint8_t p_buffer[1024];
380 uint8_t i_buffer[1024];
381 uint8_t p_state[
sizeof(
s->block_state)];
382 uint8_t i_state[
sizeof(
s->block_state)];
384 uint8_t *pbbak=
s->c.bytestream;
385 uint8_t *pbbak_start=
s->c.bytestream_start;
386 int score, score2, iscore, i_len, p_len, block_s, sum, base_bits;
387 const int w=
s->b_width <<
s->block_max_depth;
388 const int h=
s->b_height <<
s->block_max_depth;
389 const int rem_depth=
s->block_max_depth -
level;
390 const int index= (x + y*
w) << rem_depth;
392 int trx= (x+1)<<rem_depth;
393 int try= (y+1)<<rem_depth;
400 int pl =
left->color[0];
401 int pcb=
left->color[1];
402 int pcr=
left->color[2];
406 const int stride=
s->current_picture->linesize[0];
407 const int uvstride=
s->current_picture->linesize[1];
408 const uint8_t *
const current_data[3] = {
s->input_picture->data[0] + (x + y*
stride)*block_w,
409 PTR_ADD(
s->input_picture->data[1], ((x*block_w)>>
s->chroma_h_shift) + ((y*uvstride*block_w)>>
s->chroma_v_shift)),
410 PTR_ADD(
s->input_picture->data[2], ((x*block_w)>>
s->chroma_h_shift) + ((y*uvstride*block_w)>>
s->chroma_v_shift))};
412 int16_t last_mv[3][2];
414 const int shift= 1+qpel;
419 int ref, best_ref, ref_score, ref_mx, ref_my;
424 set_blocks(
s,
level, x, y, pl, pcb, pcr, 0, 0, 0,
BLOCK_INTRA);
437 last_mv[0][0]=
s->block[
index].mx;
438 last_mv[0][1]=
s->block[
index].my;
439 last_mv[1][0]= right->
mx;
440 last_mv[1][1]= right->
my;
441 last_mv[2][0]= bottom->
mx;
442 last_mv[2][1]= bottom->
my;
457 c->xmin = - x*block_w - 16+3;
458 c->ymin = - y*block_w - 16+3;
459 c->xmax = - (x+1)*block_w + (
w<<(
LOG2_MB_SIZE -
s->block_max_depth)) + 16-3;
460 c->ymax = - (y+1)*block_w + (
h<<(
LOG2_MB_SIZE -
s->block_max_depth)) + 16-3;
489 init_ref(
c, current_data,
s->last_picture[
ref]->data,
NULL, block_w*x, block_w*y, 0);
499 ref_score =
c->sub_motion_search(&enc->
m.
s, &ref_mx, &ref_my, ref_score,
509 if(score > ref_score){
519 base_bits=
get_rac_count(&
s->c) - 8*(
s->c.bytestream -
s->c.bytestream_start);
522 pc.bytestream= p_buffer;
523 memcpy(p_state,
s->block_state,
sizeof(
s->block_state));
525 if(
level!=
s->block_max_depth)
526 put_rac(&pc, &p_state[4 + s_context], 1);
528 if(
s->ref_frames > 1)
529 put_symbol(&pc, &p_state[128 + 1024 + 32*ref_context], best_ref, 0);
531 put_symbol(&pc, &p_state[128 + 32*(mx_context + 16*!!best_ref)],
mx - pmx, 1);
532 put_symbol(&pc, &p_state[128 + 32*(my_context + 16*!!best_ref)],
my - pmy, 1);
533 p_len= pc.bytestream - pc.bytestream_start;
536 block_s= block_w*block_w;
538 l= (sum + block_s/2)/block_s;
539 iscore =
pix_norm1(current_data[0],
stride, block_w) - 2*l*sum + l*l*block_s;
541 if (
s->nb_planes > 2) {
542 block_s= block_w*block_w>>(
s->chroma_h_shift +
s->chroma_v_shift);
543 sum =
pix_sum(current_data[1], uvstride, block_w>>
s->chroma_h_shift, block_w>>
s->chroma_v_shift);
544 cb= (sum + block_s/2)/block_s;
546 sum =
pix_sum(current_data[2], uvstride, block_w>>
s->chroma_h_shift, block_w>>
s->chroma_v_shift);
547 cr= (sum + block_s/2)/block_s;
554 ic.bytestream= i_buffer;
555 memcpy(i_state,
s->block_state,
sizeof(
s->block_state));
556 if(
level!=
s->block_max_depth)
557 put_rac(&ic, &i_state[4 + s_context], 1);
560 if (
s->nb_planes > 2) {
564 i_len= ic.bytestream - ic.bytestream_start;
567 av_assert1(iscore < 255*255*256 + enc->lambda2*10);
573 int varc= iscore >> 8;
574 int vard= score >> 8;
575 if (vard <= 64 || vard < varc)
581 if(
level!=
s->block_max_depth){
582 put_rac(&
s->c, &
s->block_state[4 + s_context], 0);
589 if(score2 < score && score2 < iscore)
595 memcpy(pbbak, i_buffer, i_len);
597 s->c.bytestream_start= pbbak_start;
598 s->c.bytestream= pbbak + i_len;
599 set_blocks(
s,
level, x, y, l,
cb,
cr, pmx, pmy, 0,
BLOCK_INTRA);
600 memcpy(
s->block_state, i_state,
sizeof(
s->block_state));
603 memcpy(pbbak, p_buffer, p_len);
605 s->c.bytestream_start= pbbak_start;
606 s->c.bytestream= pbbak + p_len;
607 set_blocks(
s,
level, x, y, pl, pcb, pcr,
mx,
my, best_ref, 0);
608 memcpy(
s->block_state, p_state,
sizeof(
s->block_state));
614 const int w=
s->b_width <<
s->block_max_depth;
615 const int rem_depth=
s->block_max_depth -
level;
616 const int index= (x + y*
w) << rem_depth;
617 int trx= (x+1)<<rem_depth;
623 int pl =
left->color[0];
624 int pcb=
left->color[1];
625 int pcr=
left->color[2];
633 set_blocks(
s,
level, x, y, pl, pcb, pcr, 0, 0, 0,
BLOCK_INTRA);
637 if(
level!=
s->block_max_depth){
639 put_rac(&
s->c, &
s->block_state[4 + s_context], 1);
641 put_rac(&
s->c, &
s->block_state[4 + s_context], 0);
653 if (
s->nb_planes > 2) {
657 set_blocks(
s,
level, x, y,
b->color[0],
b->color[1],
b->color[2], pmx, pmy, 0,
BLOCK_INTRA);
661 if(
s->ref_frames > 1)
662 put_symbol(&
s->c, &
s->block_state[128 + 1024 + 32*ref_context],
b->ref, 0);
663 put_symbol(&
s->c, &
s->block_state[128 + 32*mx_context],
b->mx - pmx, 1);
664 put_symbol(&
s->c, &
s->block_state[128 + 32*my_context],
b->my - pmy, 1);
665 set_blocks(
s,
level, x, y, pl, pcb, pcr,
b->mx,
b->my,
b->ref, 0);
673 Plane *
p= &
s->plane[plane_index];
674 const int block_size =
MB_SIZE >>
s->block_max_depth;
675 const int block_w = plane_index ? block_size>>
s->chroma_h_shift : block_size;
676 const int block_h = plane_index ? block_size>>
s->chroma_v_shift : block_size;
678 const int obmc_stride= plane_index ? (2*block_size)>>
s->chroma_h_shift : 2*block_size;
679 const int ref_stride=
s->current_picture->linesize[plane_index];
680 const uint8_t *
src =
s->input_picture->data[plane_index];
682 const int b_stride =
s->b_width <<
s->block_max_depth;
683 const int w=
p->width;
684 const int h=
p->height;
685 int index= mb_x + mb_y*b_stride;
694 b->color[plane_index]= 0;
695 memset(
dst, 0, obmc_stride*obmc_stride*
sizeof(
IDWTELEM));
698 int mb_x2= mb_x + (
i &1) - 1;
699 int mb_y2= mb_y + (
i>>1) - 1;
700 int x= block_w*mb_x2 + block_w/2;
701 int y= block_h*mb_y2 + block_h/2;
704 x, y, block_w, block_h,
w,
h, obmc_stride, ref_stride, obmc_stride, mb_x2, mb_y2, 0, 0, plane_index);
706 for(y2=
FFMAX(y, 0); y2<
FFMIN(
h, y+block_h); y2++){
707 for(x2=
FFMAX(x, 0); x2<
FFMIN(
w, x+block_w); x2++){
708 int index= x2-(block_w*mb_x - block_w/2) + (y2-(block_h*mb_y - block_h/2))*obmc_stride;
709 int obmc_v= obmc[
index];
711 if(y<0) obmc_v += obmc[
index + block_h*obmc_stride];
712 if(x<0) obmc_v += obmc[
index + block_w];
713 if(y+block_h>
h) obmc_v += obmc[
index - block_h*obmc_stride];
714 if(x+block_w>
w) obmc_v += obmc[
index - block_w];
719 ab += (
src[x2 + y2*ref_stride] - (d>>
FRAC_BITS)) * obmc_v;
720 aa += obmc_v * obmc_v;
730 const int b_stride =
s->b_width <<
s->block_max_depth;
731 const int b_height =
s->b_height<<
s->block_max_depth;
732 int index= x + y*b_stride;
742 if(x<0 || x>=b_stride || y>=b_height)
768 int plane_index, uint8_t (*obmc_edged)[
MB_SIZE * 2])
771 Plane *
p= &
s->plane[plane_index];
772 const int block_size =
MB_SIZE >>
s->block_max_depth;
773 const int block_w = plane_index ? block_size>>
s->chroma_h_shift : block_size;
774 const int block_h = plane_index ? block_size>>
s->chroma_v_shift : block_size;
775 const int obmc_stride= plane_index ? (2*block_size)>>
s->chroma_h_shift : 2*block_size;
776 const int ref_stride=
s->current_picture->linesize[plane_index];
777 uint8_t *
dst=
s->current_picture->data[plane_index];
778 const uint8_t *
src =
s->input_picture->data[plane_index];
780 uint8_t *cur =
s->scratchbuf;
782 const int b_stride =
s->b_width <<
s->block_max_depth;
783 const int b_height =
s->b_height<<
s->block_max_depth;
784 const int w=
p->width;
785 const int h=
p->height;
789 int sx= block_w*mb_x - block_w/2;
790 int sy= block_h*mb_y - block_h/2;
791 int x0=
FFMAX(0,-sx);
792 int y0=
FFMAX(0,-sy);
793 int x1=
FFMIN(block_w*2,
w-sx);
794 int y1=
FFMIN(block_h*2,
h-sy);
799 ff_snow_pred_block(
s, cur,
tmp, ref_stride, sx, sy, block_w*2, block_h*2, &
s->block[mb_x + mb_y*b_stride], plane_index,
w,
h);
801 for(y=y0; y<y1; y++){
802 const uint8_t *obmc1= obmc_edged[y];
804 uint8_t *cur1 = cur + y*ref_stride;
805 uint8_t *dst1 =
dst + sx + (sy+y)*ref_stride;
806 for(x=x0; x<x1; x++){
807 #if FRAC_BITS >= LOG2_OBMC_MAX
813 if(v&(~255)) v= ~(v>>31);
820 && (mb_x == 0 || mb_x == b_stride-1)
821 && (mb_y == 0 || mb_y == b_height-1)){
831 memcpy(
dst + sx+x0 + (sy+y)*ref_stride, cur + x0 + y*ref_stride, x1-x0);
842 distortion =
ff_w97_32_c(&enc->
m.
s,
src + sx + sy*ref_stride,
dst + sx + sy*ref_stride, ref_stride, 32);
844 distortion =
ff_w53_32_c(&enc->
m.
s,
src + sx + sy*ref_stride,
dst + sx + sy*ref_stride, ref_stride, 32);
848 int off = sx+16*(
i&1) + (sy+16*(
i>>1))*ref_stride;
854 distortion = enc->
m.
s.
me.
me_cmp[0](&enc->
m.
s,
src + sx + sy*ref_stride,
dst + sx + sy*ref_stride, ref_stride, block_w*2);
865 if(mb_x == b_stride-2)
868 return distortion + rate*penalty_factor;
875 Plane *
p= &
s->plane[plane_index];
876 const int block_size =
MB_SIZE >>
s->block_max_depth;
877 const int block_w = plane_index ? block_size>>
s->chroma_h_shift : block_size;
878 const int block_h = plane_index ? block_size>>
s->chroma_v_shift : block_size;
880 const int obmc_stride= plane_index ? (2*block_size)>>
s->chroma_h_shift : 2*block_size;
881 const int ref_stride=
s->current_picture->linesize[plane_index];
882 uint8_t *
dst=
s->current_picture->data[plane_index];
883 const uint8_t *
src =
s->input_picture->data[plane_index];
887 const int b_stride =
s->b_width <<
s->block_max_depth;
888 const int w=
p->width;
889 const int h=
p->height;
897 int mb_x2= mb_x + (
i%3) - 1;
898 int mb_y2= mb_y + (
i/3) - 1;
899 int x= block_w*mb_x2 + block_w/2;
900 int y= block_h*mb_y2 + block_h/2;
903 x, y, block_w, block_h,
w,
h, 0, ref_stride, obmc_stride, mb_x2, mb_y2, 1, 1, plane_index);
906 for(y2= y; y2<0; y2++)
907 memcpy(
dst + x + y2*ref_stride,
src + x + y2*ref_stride, block_w);
908 for(y2=
h; y2<y+block_h; y2++)
909 memcpy(
dst + x + y2*ref_stride,
src + x + y2*ref_stride, block_w);
911 for(y2= y; y2<y+block_h; y2++)
912 memcpy(
dst + x + y2*ref_stride,
src + x + y2*ref_stride, -x);
915 for(y2= y; y2<y+block_h; y2++)
916 memcpy(
dst +
w + y2*ref_stride,
src +
w + y2*ref_stride, x+block_w -
w);
920 distortion += enc->
m.
s.
me.
me_cmp[block_w==8](&enc->
m.
s,
src + x + y*ref_stride,
dst + x + y*ref_stride, ref_stride, block_h);
934 for(
i=merged?4:0;
i<9;
i++){
935 static const int dxy[9][2] = {{0,0},{1,0},{0,1},{1,1},{2,0},{2,1},{-1,2},{0,2},{1,2}};
939 return distortion + rate*penalty_factor;
943 const int w=
b->width;
944 const int h=
b->height;
949 int *runs =
s->run_buffer;
956 int l=0, lt=0, t=0, rt=0;
978 if(px<b->parent->width && py<b->parent->height)
983 runs[run_index++]=
run;
991 max_index= run_index;
992 runs[run_index++]=
run;
994 run= runs[run_index++];
997 if(run_index <= max_index)
1001 if(
s->c.bytestream_end -
s->c.bytestream <
w*40){
1007 int l=0, lt=0, t=0, rt=0;
1029 if(px<b->parent->width && py<b->parent->height)
1038 run= runs[run_index++];
1040 if(run_index <= max_index)
1050 int l2= 2*
FFABS(l) + (l<0);
1051 int t2= 2*
FFABS(t) + (t<0);
1070 uint8_t (*obmc_edged)[
MB_SIZE * 2],
int *best_rd)
1073 const int b_stride=
s->b_width <<
s->block_max_depth;
1101 int mb_x,
int mb_y,
int p0,
int p1,
1102 uint8_t (*obmc_edged)[
MB_SIZE * 2],
int *best_rd)
1105 const int b_stride =
s->b_width <<
s->block_max_depth;
1127 if (rd < *best_rd) {
1137 int p0,
int p1,
int ref,
int *best_rd)
1140 const int b_stride=
s->b_width <<
s->block_max_depth;
1149 backup[0] =
block[0];
1150 backup[1] =
block[1];
1151 backup[2] =
block[b_stride];
1152 backup[3] =
block[b_stride + 1];
1177 block[0]= backup[0];
1178 block[1]= backup[1];
1179 block[b_stride]= backup[2];
1180 block[b_stride+1]= backup[3];
1188 int pass, mb_x, mb_y;
1189 const int b_width =
s->b_width <<
s->block_max_depth;
1190 const int b_height=
s->b_height <<
s->block_max_depth;
1191 const int b_stride= b_width;
1196 uint8_t
state[
sizeof(
s->block_state)];
1197 memcpy(
state,
s->block_state,
sizeof(
s->block_state));
1198 for(mb_y= 0; mb_y<
s->b_height; mb_y++)
1199 for(mb_x= 0; mb_x<
s->b_width; mb_x++)
1202 memcpy(
s->block_state,
state,
sizeof(
s->block_state));
1205 for(pass=0; pass<25; pass++){
1208 for(mb_y= 0; mb_y<b_height; mb_y++){
1209 for(mb_x= 0; mb_x<b_width; mb_x++){
1210 int dia_change,
i, j,
ref;
1211 int best_rd= INT_MAX, ref_rd;
1213 const int index= mb_x + mb_y * b_stride;
1223 const int b_w= (
MB_SIZE >>
s->block_max_depth);
1239 for (y = 0; y < b_w * 2; y++)
1240 memcpy(obmc_edged[y],
ff_obmc_tab[
s->block_max_depth] + y * b_w * 2, b_w * 2);
1242 for(y=0; y<b_w*2; y++)
1243 memset(obmc_edged[y], obmc_edged[y][0] + obmc_edged[y][b_w-1], b_w);
1244 if(mb_x==b_stride-1)
1245 for(y=0; y<b_w*2; y++)
1246 memset(obmc_edged[y]+b_w, obmc_edged[y][b_w] + obmc_edged[y][b_w*2-1], b_w);
1248 for(x=0; x<b_w*2; x++)
1249 obmc_edged[0][x] += obmc_edged[b_w-1][x];
1250 for(y=1; y<b_w; y++)
1251 memcpy(obmc_edged[y], obmc_edged[0], b_w*2);
1253 if(mb_y==b_height-1){
1254 for(x=0; x<b_w*2; x++)
1255 obmc_edged[b_w*2-1][x] += obmc_edged[b_w][x];
1256 for(y=b_w; y<b_w*2-1; y++)
1257 memcpy(obmc_edged[y], obmc_edged[b_w*2-1], b_w*2);
1262 if(mb_x==0 || mb_y==0 || mb_x==b_width-1 || mb_y==b_height-1){
1263 const uint8_t *
src =
s->input_picture->data[0];
1264 uint8_t *
dst=
s->current_picture->data[0];
1265 const int stride=
s->current_picture->linesize[0];
1266 const int block_w=
MB_SIZE >>
s->block_max_depth;
1267 const int block_h=
MB_SIZE >>
s->block_max_depth;
1268 const int sx= block_w*mb_x - block_w/2;
1269 const int sy= block_h*mb_y - block_h/2;
1270 const int w=
s->plane[0].width;
1271 const int h=
s->plane[0].height;
1276 for(y=
h; y<sy+block_h*2; y++)
1279 for(y=sy; y<sy+block_h*2; y++)
1282 if(sx+block_w*2 >
w){
1283 for(y=sy; y<sy+block_h*2; y++)
1289 for(
i=0;
i <
s->nb_planes;
i++)
1302 int16_t (*mvr)[2]= &
s->ref_mvs[
ref][
index];
1311 check_block_inter(enc, mb_x, mb_y, mvr[-b_stride][0], mvr[-b_stride][1], obmc_edged, &best_rd);
1313 check_block_inter(enc, mb_x, mb_y, mvr[-1][0], mvr[-1][1], obmc_edged, &best_rd);
1317 check_block_inter(enc, mb_x, mb_y, mvr[b_stride][0], mvr[b_stride][1], obmc_edged, &best_rd);
1322 int newx =
block->mx;
1323 int newy =
block->my;
1326 for(
i=0;
i < dia_size;
i++){
1328 dia_change |=
check_block_inter(enc, mb_x, mb_y, newx+4*(
i-j), newy+(4*j), obmc_edged, &best_rd);
1329 dia_change |=
check_block_inter(enc, mb_x, mb_y, newx-4*(
i-j), newy-(4*j), obmc_edged, &best_rd);
1330 dia_change |=
check_block_inter(enc, mb_x, mb_y, newx-(4*j), newy+4*(
i-j), obmc_edged, &best_rd);
1331 dia_change |=
check_block_inter(enc, mb_x, mb_y, newx+(4*j), newy-4*(
i-j), obmc_edged, &best_rd);
1337 static const int square[8][2]= {{+1, 0},{-1, 0},{ 0,+1},{ 0,-1},{+1,+1},{-1,-1},{+1,-1},{-1,+1},};
1344 mvr[0][0]=
block->mx;
1345 mvr[0][1]=
block->my;
1346 if(ref_rd > best_rd){
1373 if(
s->block_max_depth == 1){
1375 for(mb_y= 0; mb_y<b_height; mb_y+=2){
1376 for(mb_x= 0; mb_x<b_width; mb_x+=2){
1378 int best_rd, init_rd;
1379 const int index= mb_x + mb_y * b_stride;
1384 b[2]=
b[0]+b_stride;
1400 (
b[0]->
my +
b[1]->
my +
b[2]->
my +
b[3]->
my + 2) >> 2, 0, &best_rd);
1406 if(init_rd != best_rd)
1439 const int w=
b->width;
1440 const int h=
b->height;
1443 int x,y, thres1, thres2;
1461 if((
unsigned)(
i+thres1) > thres2){
1481 if((
unsigned)(
i+thres1) > thres2){
1500 const int w=
b->width;
1501 const int h=
b->height;
1522 const int w=
b->width;
1523 const int h=
b->height;
1526 for(y=
h-1; y>=0; y--){
1527 for(x=
w-1; x>=0; x--){
1546 const int w=
b->width;
1547 const int h=
b->height;
1570 int plane_index,
level, orientation;
1572 for(plane_index=0; plane_index<
FFMIN(
s->nb_planes, 2); plane_index++){
1574 for(orientation=
level ? 1:0; orientation<4; orientation++){
1575 if(orientation==2)
continue;
1576 put_symbol(&
s->c,
s->header_state,
s->plane[plane_index].band[
level][orientation].qlog, 1);
1586 memset(kstate,
MID_STATE,
sizeof(kstate));
1589 if(
s->keyframe ||
s->always_reset){
1591 s->last_spatial_decomposition_type=
1595 s->last_block_max_depth= 0;
1596 for(plane_index=0; plane_index<2; plane_index++){
1597 Plane *
p= &
s->plane[plane_index];
1600 memset(
p->last_hcoeff, 0,
sizeof(
p->last_hcoeff));
1605 put_rac(&
s->c,
s->header_state,
s->always_reset);
1606 put_symbol(&
s->c,
s->header_state,
s->temporal_decomposition_type, 0);
1607 put_symbol(&
s->c,
s->header_state,
s->temporal_decomposition_count, 0);
1608 put_symbol(&
s->c,
s->header_state,
s->spatial_decomposition_count, 0);
1610 if (
s->nb_planes > 2) {
1614 put_rac(&
s->c,
s->header_state,
s->spatial_scalability);
1623 for(plane_index=0; plane_index<
FFMIN(
s->nb_planes, 2); plane_index++){
1624 Plane *
p= &
s->plane[plane_index];
1625 update_mc |=
p->last_htaps !=
p->htaps;
1626 update_mc |=
p->last_diag_mc !=
p->diag_mc;
1627 update_mc |= !!memcmp(
p->last_hcoeff,
p->hcoeff,
sizeof(
p->hcoeff));
1629 put_rac(&
s->c,
s->header_state, update_mc);
1631 for(plane_index=0; plane_index<
FFMIN(
s->nb_planes, 2); plane_index++){
1632 Plane *
p= &
s->plane[plane_index];
1635 for(
i=
p->htaps/2;
i;
i--)
1639 if(
s->last_spatial_decomposition_count !=
s->spatial_decomposition_count){
1641 put_symbol(&
s->c,
s->header_state,
s->spatial_decomposition_count, 0);
1647 put_symbol(&
s->c,
s->header_state,
s->spatial_decomposition_type -
s->last_spatial_decomposition_type, 1);
1649 put_symbol(&
s->c,
s->header_state,
s->mv_scale -
s->last_mv_scale, 1);
1650 put_symbol(&
s->c,
s->header_state,
s->qbias -
s->last_qbias , 1);
1651 put_symbol(&
s->c,
s->header_state,
s->block_max_depth -
s->last_block_max_depth, 1);
1659 for(plane_index=0; plane_index<2; plane_index++){
1660 Plane *
p= &
s->plane[plane_index];
1661 p->last_diag_mc=
p->diag_mc;
1662 p->last_htaps =
p->htaps;
1663 memcpy(
p->last_hcoeff,
p->hcoeff,
sizeof(
p->hcoeff));
1667 s->last_spatial_decomposition_type =
s->spatial_decomposition_type;
1668 s->last_qlog =
s->qlog;
1669 s->last_qbias =
s->qbias;
1670 s->last_mv_scale =
s->mv_scale;
1671 s->last_block_max_depth =
s->block_max_depth;
1672 s->last_spatial_decomposition_count =
s->spatial_decomposition_count;
1686 uint32_t coef_sum= 0;
1687 int level, orientation, delta_qlog;
1690 for(orientation=
level ? 1 : 0; orientation<4; orientation++){
1693 const int w=
b->width;
1694 const int h=
b->height;
1698 const int qdiv= (1<<16)/qmul;
1708 coef_sum+=
abs(buf[x+y*
stride]) * qdiv >> 16;
1715 coef_sum = (uint64_t)coef_sum * coef_sum >> 16;
1730 s->qlog+= delta_qlog;
1737 int level, orientation, x, y;
1741 for(orientation=
level ? 1 : 0; orientation<4; orientation++){
1745 memset(
s->spatial_idwt_buffer, 0,
sizeof(*
s->spatial_idwt_buffer)*
width*
height);
1746 ibuf[
b->width/2 +
b->height/2*
b->stride]= 256*16;
1749 for(x=0; x<
width; x++){
1754 if (orientation == 2)
1757 if (orientation != 1)
1765 const AVFrame *pict,
int *got_packet)
1773 const int width=
s->avctx->width;
1774 const int height=
s->avctx->height;
1775 int level, orientation, plane_index,
i, y,
ret;
1776 uint8_t rc_header_bak[
sizeof(
s->header_state)];
1777 uint8_t rc_block_bak[
sizeof(
s->block_state)];
1785 for(
i=0;
i <
s->nb_planes;
i++){
1786 int hshift=
i ?
s->chroma_h_shift : 0;
1787 int vshift=
i ?
s->chroma_v_shift : 0;
1789 memcpy(&
s->input_picture->data[
i][y *
s->input_picture->linesize[
i]],
1798 pic =
s->input_picture;
1827 if (
s->current_picture->data[0]) {
1828 int w =
s->avctx->width;
1829 int h =
s->avctx->height;
1832 s->current_picture->linesize[0],
w ,
h ,
1834 if (
s->current_picture->data[2]) {
1836 s->current_picture->linesize[1],
w>>
s->chroma_h_shift,
h>>
s->chroma_v_shift,
1839 s->current_picture->linesize[2],
w>>
s->chroma_h_shift,
h>>
s->chroma_v_shift,
1853 int block_width = (
width +15)>>4;
1854 int block_height= (
height+15)>>4;
1855 int stride=
s->current_picture->linesize[0];
1863 mpv-> new_pic =
s->input_picture;
1891 memcpy(rc_header_bak,
s->header_state,
sizeof(
s->header_state));
1892 memcpy(rc_block_bak,
s->block_state,
sizeof(
s->block_state));
1897 s->spatial_decomposition_count= 5;
1899 while( !(
width >>(
s->chroma_h_shift +
s->spatial_decomposition_count))
1900 || !(
height>>(
s->chroma_v_shift +
s->spatial_decomposition_count)))
1901 s->spatial_decomposition_count--;
1903 if (
s->spatial_decomposition_count <= 0) {
1913 if(
s->last_spatial_decomposition_count !=
s->spatial_decomposition_count){
1914 for(plane_index=0; plane_index <
s->nb_planes; plane_index++){
1920 mpv->
misc_bits = 8 * (
s->c.bytestream -
s->c.bytestream_start);
1924 for(plane_index=0; plane_index < s->nb_planes; plane_index++){
1925 Plane *
p= &
s->plane[plane_index];
1933 if(pict->
data[plane_index])
1968 ff_spatial_dwt(
s->spatial_dwt_buffer,
s->temp_dwt_buffer,
w,
h,
w,
s->spatial_decomposition_type,
s->spatial_decomposition_count);
1970 if (enc->
pass1_rc && plane_index==0) {
1972 if (delta_qlog <= INT_MIN)
1977 memcpy(
s->header_state, rc_header_bak,
sizeof(
s->header_state));
1978 memcpy(
s->block_state, rc_block_bak,
sizeof(
s->block_state));
1985 for(orientation=
level ? 1 : 0; orientation<4; orientation++){
2000 for(orientation=
level ? 1 : 0; orientation<4; orientation++){
2007 ff_spatial_idwt(
s->spatial_idwt_buffer,
s->temp_idwt_buffer,
w,
h,
w,
s->spatial_decomposition_type,
s->spatial_decomposition_count);
2011 s->spatial_idwt_buffer[y*
w + x] *= 1 <<
FRAC_BITS;
2021 s->current_picture->data[plane_index][y*
s->current_picture->linesize[plane_index] + x]=
2022 pict->
data[plane_index][y*pict->
linesize[plane_index] + x];
2026 memset(
s->spatial_idwt_buffer, 0,
sizeof(
IDWTELEM)*
w*
h);
2033 if(pict->
data[plane_index])
2036 int d=
s->current_picture->data[plane_index][y*
s->current_picture->linesize[plane_index] + x] - pict->
data[plane_index][y*pict->
linesize[plane_index] + x];
2040 s->avctx->error[plane_index] +=
error;
2051 s->current_picture->pict_type = pic->
pict_type;
2052 s->current_picture->quality = pic->
quality;
2053 enc->
m.
frame_bits = 8 * (
s->c.bytestream -
s->c.bytestream_start);
2055 enc->
m.
total_bits += 8*(
s->c.bytestream -
s->c.bytestream_start);
2071 s->current_picture->pict_type);
2107 #define OFFSET(x) offsetof(SnowEncContext, x)
2108 #define VE AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_ENCODING_PARAM
2115 {
"memc_only",
"Only do ME/MC (I frames -> ref, P frame -> ME+MC).",
OFFSET(memc_only),
AV_OPT_TYPE_BOOL, { .i64 = 0 }, 0, 1,
VE },
2116 {
"no_bitstream",
"Skip final bitstream writeout.",
OFFSET(no_bitstream),
AV_OPT_TYPE_BOOL, { .i64 = 0 }, 0, 1,
VE },
2117 {
"intra_penalty",
"Penalty for intra blocks in block decision",
OFFSET(intra_penalty),
AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX,
VE },
2118 {
"iterative_dia_size",
"Dia size for the iterative ME",
OFFSET(iterative_dia_size),
AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX,
VE },
2119 {
"sc_threshold",
"Scene change threshold",
OFFSET(scenechange_threshold),
AV_OPT_TYPE_INT, { .i64 = 0 }, INT_MIN, INT_MAX,
VE },
2123 {
"rc_eq",
"Set rate control equation. When computing the expression, besides the standard functions "
2124 "defined in the section 'Expression Evaluation', the following functions are available: "
2125 "bits2qp(bits), qp2bits(qp). Also the following constants are available: iTex pTex tex mv "
2126 "fCode iCount mcVar var isI isP isB avgQP qComp avgIITex avgPITex avgPPTex avgBPTex avgTex.",
static void error(const char *err)
#define CODEC_PIXFMTS(...)
static int encode_subband(SnowContext *s, SubBand *b, const IDWTELEM *src, const IDWTELEM *parent, int stride, int orientation)
int misc_bits
cbp, mb_type
static void decorrelate(SnowContext *s, SubBand *b, IDWTELEM *src, int stride, int inverse, int use_median)
static void set_blocks(SnowContext *s, int level, int x, int y, int l, int cb, int cr, int mx, int my, int ref, int type)
#define FF_CODEC_CAP_INIT_CLEANUP
The codec allows calling the close function for deallocation even if the init function returned a fai...
Filter the word “frame” indicates either a video frame or a group of audio as stored in an AVFrame structure Format for each input and each output the list of supported formats For video that means pixel format For audio that means channel sample they are references to shared objects When the negotiation mechanism computes the intersection of the formats supported at each end of a all references to both lists are replaced with a reference to the intersection And when a single format is eventually chosen for a link amongst the remaining all references to the list are updated That means that if a filter requires that its input and output have the same format amongst a supported all it has to do is use a reference to the same list of formats query_formats can leave some formats unset and return AVERROR(EAGAIN) to cause the negotiation mechanism toagain later. That can be used by filters with complex requirements to use the format negotiated on one link to set the formats supported on another. Frame references ownership and permissions
av_cold int ff_me_init(MotionEstContext *c, AVCodecContext *avctx, const MECmpContext *mecc, int mpvenc)
static int ratecontrol_1pass(SnowEncContext *enc, AVFrame *pict)
static double cb(void *priv, double x, double y)
static av_cold int encode_end(AVCodecContext *avctx)
int scenechange_threshold
MpegEncContext c
the common base context
#define AV_CODEC_FLAG_QSCALE
Use fixed qscale.
Motion estimation context.
void ff_me_init_pic(MPVEncContext *const s)
#define AV_CODEC_CAP_ENCODER_RECON_FRAME
The encoder is able to output reconstructed frame data, i.e.
void av_frame_free(AVFrame **frame)
Free the frame and any dynamically allocated objects in it, e.g.
This structure describes decoded (raw) audio or video data.
int64_t pts
Presentation timestamp in time_base units (time when frame should be shown to user).
static void update_last_header_values(SnowContext *s)
enum AVPictureType pict_type
AV_PICTURE_TYPE_I, AV_PICTURE_TYPE_P, AV_PICTURE_TYPE_B, ...
static void iterative_me(SnowEncContext *enc)
static int get_penalty_factor(int lambda, int lambda2, int type)
unsigned int lambda
Lagrange multiplier used in rate distortion.
static int encode_subband_c0run(SnowContext *s, SubBand *b, const IDWTELEM *src, const IDWTELEM *parent, int stride, int orientation)
int b8_stride
2*mb_width+1 used for some 8x8 block arrays to allow simple addressing
#define AV_CODEC_FLAG_PSNR
error[?] variables will be set during encoding.
static int encode_frame(AVCodecContext *avctx, AVPacket *pkt, const AVFrame *pict, int *got_packet)
int64_t mb_var_sum
sum of MB variance for current frame
#define AV_CODEC_FLAG_4MV
4 MV per MB allowed / advanced prediction for H.263.
#define AV_PKT_FLAG_KEY
The packet contains a keyframe.
#define FF_INPUT_BUFFER_MIN_SIZE
Used by some encoders as upper bound for the length of headers.
av_cold void ff_snow_common_end(SnowContext *s)
uint8_t * data[AV_NUM_DATA_POINTERS]
pointer to the picture/channel planes.
void ff_spatial_dwt(DWTELEM *buffer, DWTELEM *temp, int width, int height, int stride, int type, int decomposition_count)
uint8_t type
Bitfield of BLOCK_*.
static av_always_inline int check_4block_inter(SnowEncContext *enc, int mb_x, int mb_y, int p0, int p1, int ref, int *best_rd)
uint8_t ptrdiff_t const uint8_t ptrdiff_t int intptr_t mx
AVCodec p
The public AVCodec.
void ff_spatial_idwt(IDWTELEM *buffer, IDWTELEM *temp, int width, int height, int stride, int type, int decomposition_count)
unsigned me_cache_generation
static void encode_blocks(SnowEncContext *enc, int search)
av_cold void ff_init_range_encoder(RangeCoder *c, uint8_t *buf, int buf_size)
int refs
number of reference frames
ptrdiff_t uvlinesize
line size, for chroma in bytes, may be different from width
av_cold void ff_me_cmp_init(MECmpContext *c, AVCodecContext *avctx)
static av_always_inline int check_block_intra(SnowEncContext *enc, int mb_x, int mb_y, int p[3], uint8_t(*obmc_edged)[MB_SIZE *2], int *best_rd)
int flags
AV_CODEC_FLAG_*.
int av_pix_fmt_get_chroma_sub_sample(enum AVPixelFormat pix_fmt, int *h_shift, int *v_shift)
Utility function to access log2_chroma_w log2_chroma_h from the pixel format AVPixFmtDescriptor.
void ff_snow_pred_block(SnowContext *s, uint8_t *dst, uint8_t *tmp, ptrdiff_t stride, int sx, int sy, int b_w, int b_h, const BlockNode *block, int plane_index, int w, int h)
it s the only field you need to keep assuming you have a context There is some magic you don t need to care about around this just let it vf type
static int get_4block_rd(SnowEncContext *enc, int mb_x, int mb_y, int plane_index)
int ff_encode_add_stats_side_data(AVPacket *pkt, int quality, const int64_t error[], int error_count, enum AVPictureType pict_type)
#define FF_CODEC_ENCODE_CB(func)
AVFrame * av_frame_alloc(void)
Allocate an AVFrame and set its fields to default values.
int ff_snow_common_init_after_header(AVCodecContext *avctx)
#define AV_LOG_ERROR
Something went wrong and cannot losslessly be recovered.
MPVWorkPicture cur_pic
copy of the current picture structure.
#define AV_FRAME_FLAG_KEY
A flag to mark frames that are keyframes.
static int encode_q_branch(SnowEncContext *enc, int level, int x, int y)
int global_quality
Global quality for codecs which cannot change it per frame.
MPVPicture * ptr
RefStruct reference.
#define AV_CEIL_RSHIFT(a, b)
int64_t mc_mb_var_sum
motion compensated MB variance for current frame
#define BLOCK_OPT
Block needs no checks in this round of iterative motion estiation.
RateControlContext rc_context
contains stuff only accessed in ratecontrol.c
static void calculate_visual_weight(SnowContext *s, Plane *p)
#define AV_CODEC_CAP_ENCODER_REORDERED_OPAQUE
This encoder can reorder user opaque values from input AVFrames and return them with corresponding ou...
#define av_assert0(cond)
assert() equivalent, that is always enabled.
int mb_num
number of MBs of a picture
#define AV_LOG_DEBUG
Stuff which is only useful for libav* developers.
static av_always_inline void add_yblock(SnowContext *s, int sliced, slice_buffer *sb, IDWTELEM *dst, uint8_t *dst8, const uint8_t *obmc, int src_x, int src_y, int b_w, int b_h, int w, int h, int dst_stride, int src_stride, int obmc_stride, int b_x, int b_y, int add, int offset_dst, int plane_index)
static int pix_norm1(const uint8_t *pix, int line_size, int w)
av_cold int ff_snow_common_init(AVCodecContext *avctx)
#define PTR_ADD(ptr, off)
static int get_encode_buffer(SnowContext *s, AVFrame *frame)
@ AV_PIX_FMT_YUV420P
planar YUV 4:2:0, 12bpp, (1 Cr & Cb sample per 2x2 Y samples)
uint64_t encoding_error[SNOW_MAX_PLANES]
#define CODEC_LONG_NAME(str)
uint8_t ptrdiff_t const uint8_t ptrdiff_t int intptr_t intptr_t my
#define FFABS(a)
Absolute value, Note, INT_MIN / INT64_MIN result in undefined behavior as they are not representable ...
it s the only field you need to keep assuming you have a context There is some magic you don t need to care about around this just let it vf default minimum maximum flags name is the option keep it simple and lowercase description are in without and describe what they for example set the foo of the bar offset is the offset of the field in your context
#define LIBAVUTIL_VERSION_INT
Describe the class of an AVClass context structure.
unsigned me_cache[ME_CACHE_SIZE]
static int bias(int x, int c)
void(* draw_edges)(uint8_t *buf, ptrdiff_t wrap, int width, int height, int w, int h, int sides)
float ff_rate_estimate_qscale(MPVMainEncContext *const m, int dry_run)
int16_t my
Motion vector component Y, see mv_scale.
static int get_block_rd(SnowEncContext *enc, int mb_x, int mb_y, int plane_index, uint8_t(*obmc_edged)[MB_SIZE *2])
struct AVCodecInternal * internal
Private context used for internal data.
int64_t bit_rate
the average bitrate
int ff_rac_terminate(RangeCoder *c, int version)
Terminates the range coder.
int display_picture_number
#define ROUNDED_DIV(a, b)
const char * av_default_item_name(void *ptr)
Return the context name.
@ AV_PICTURE_TYPE_I
Intra.
unsigned int lambda2
(lambda*lambda) >> FF_LAMBDA_SHIFT
IDWTELEM obmc_scratchpad[MB_SIZE *MB_SIZE *12 *2]
static void correlate(SnowContext *s, SubBand *b, IDWTELEM *src, int stride, int inverse, int use_median)
@ AV_PIX_FMT_GRAY8
Y , 8bpp.
const uint8_t(* ff_h263_get_mv_penalty(void))[MAX_DMV *2+1]
Undefined Behavior In the C some operations are like signed integer dereferencing freed accessing outside allocated Undefined Behavior must not occur in a C it is not safe even if the output of undefined operations is unused The unsafety may seem nit picking but Optimizing compilers have in fact optimized code on the assumption that no undefined Behavior occurs Optimizing code based on wrong assumptions can and has in some cases lead to effects beyond the output of computations The signed integer overflow problem in speed critical code Code which is highly optimized and works with signed integers sometimes has the problem that often the output of the computation does not c
static void init_ref(MotionEstContext *c, const uint8_t *const src[3], uint8_t *const ref[3], uint8_t *const ref2[3], int x, int y, int ref_index)
static void put_symbol(RangeCoder *c, uint8_t *state, int v, int is_signed)
int ff_encode_alloc_frame(AVCodecContext *avctx, AVFrame *frame)
Allocate buffers for a frame.
int quarter_sample
1->qpel, 0->half pel ME/MC
char * stats_out
pass1 encoding statistics output buffer
#define AV_CODEC_FLAG_QPEL
Use qpel MC.
MPVWorkPicture last_pic
copy of the previous picture structure.
enum AVPictureType pict_type
Picture type of the frame.
int(* init)(AVBSFContext *ctx)
#define AV_CODEC_CAP_DR1
Codec uses get_buffer() or get_encode_buffer() for allocating buffers and supports custom allocators.
int gop_size
the number of pictures in a group of pictures, or 0 for intra_only
static void encode_header(SnowContext *s)
static int shift(int a, int b)
uint8_t ptrdiff_t const uint8_t ptrdiff_t int intptr_t intptr_t int int16_t * dst
int quality
quality (between 1 (good) and FF_LAMBDA_MAX (bad))
#define i(width, name, range_min, range_max)
int mb_stride
mb_width+1 used for some arrays to allow simple addressing of left & top MBs without sig11
void ff_build_rac_states(RangeCoder *c, int factor, int max_p)
const uint8_t(* mv_penalty)[MAX_DMV *2+1]
bit amount needed to encode a MV
static int pix_sum(const uint8_t *pix, int line_size, int w, int h)
int frame_bits
bits used for the current frame
const FFCodec ff_snow_encoder
struct AVCodecContext * avctx
#define AV_CODEC_FLAG_PASS2
Use internal 2pass ratecontrol in second pass mode.
The reader does not expect b to be semantically here and if the code is changed by maybe adding a a division or other the signedness will almost certainly be mistaken To avoid this confusion a new type was SUINT is the C unsigned type but it holds a signed int to use the same example SUINT a
it s the only field you need to keep assuming you have a context There is some magic you don t need to care about around this just let it vf offset
int flags
A combination of AV_PKT_FLAG values.
static void encode_q_branch2(SnowContext *s, int level, int x, int y)
const int8_t ff_quant3bA[256]
const uint8_t *const ff_obmc_tab[4]
#define av_assert2(cond)
assert() equivalent, that does lie in speed critical code.
#define ENCODER_EXTRA_BITS
#define AV_CODEC_FLAG_RECON_FRAME
Request the encoder to output reconstructed frames, i.e. frames that would be produced by decoding th...
static void pred_mv(DiracBlock *block, int stride, int x, int y, int ref)
int f_code
forward MV resolution
static int get_block_bits(SnowContext *s, int x, int y, int w)
int ff_get_mb_score(MPVEncContext *s, int mx, int my, int src_index, int ref_index, int size, int h, int add_rate)
int ff_w53_32_c(MPVEncContext *v, const uint8_t *pix1, const uint8_t *pix2, ptrdiff_t line_size, int h)
#define BLOCK_INTRA
Intra block, inter otherwise.
int motion_est
ME algorithm.
#define av_assert1(cond)
assert() equivalent, that does not lie in speed critical code.
static int qscale2qlog(int qscale)
av_cold int ff_rate_control_init(MPVMainEncContext *const m)
it s the only field you need to keep assuming you have a context There is some magic you don t need to care about around this just let it vf default value
int dia_size
ME diamond size & shape.
void av_frame_unref(AVFrame *frame)
Unreference all the buffers referenced by frame and reset the frame fields.
int mb_lmin
minimum MB Lagrange multiplier
const uint8_t ff_qexp[QROOT]
void ff_write_pass1_stats(MPVMainEncContext *const m)
static av_always_inline void predict_plane(SnowContext *s, IDWTELEM *buf, int plane_index, int add)
const char * name
Name of the codec implementation.
const EXTERN uint32_t ff_square_tab[512]
static int get_dc(SnowEncContext *enc, int mb_x, int mb_y, int plane_index)
enum AVPixelFormat pix_fmt
Pixel format, see AV_PIX_FMT_xxx.
@ AVCOL_RANGE_MPEG
Narrow or limited range content.
void * av_calloc(size_t nmemb, size_t size)
int ff_w97_32_c(MPVEncContext *v, const uint8_t *pix1, const uint8_t *pix2, ptrdiff_t line_size, int h)
int64_t frame_num
Frame counter, set by libavcodec.
MpegvideoEncDSPContext mpvencdsp
static const float pred[4]
static float search(FOCContext *foc, int pass, int maxpass, int xmin, int xmax, int ymin, int ymax, int *best_x, int *best_y, float best_score)
const char * class_name
The name of the class; usually it is the same name as the context structure type to which the AVClass...
these buffered frames must be flushed immediately if a new input produces new the filter must not call request_frame to get more It must just process the frame or queue it The task of requesting more frames is left to the filter s request_frame method or the application If a filter has several the filter must be ready for frames arriving randomly on any input any filter with several inputs will most likely require some kind of queuing mechanism It is perfectly acceptable to have a limited queue and to drop frames when the inputs are too unbalanced request_frame For filters that do not use the this method is called when a frame is wanted on an output For a it should directly call filter_frame on the corresponding output For a if there are queued frames already one of these frames should be pushed If the filter should request a frame on one of its repeatedly until at least one frame has been pushed Return or at least make progress towards producing a frame
static av_cold int encode_init(AVCodecContext *avctx)
static const AVOption options[]
AVFrame * recon_frame
When the AV_CODEC_FLAG_RECON_FRAME flag is used.
uint8_t * scratchpad
data area for the ME algo, so that the ME does not need to malloc/free.
Tag MUST be and< 10hcoeff half pel interpolation filter coefficients, hcoeff[0] are the 2 middle coefficients[1] are the next outer ones and so on, resulting in a filter like:...eff[2], hcoeff[1], hcoeff[0], hcoeff[0], hcoeff[1], hcoeff[2] ... the sign of the coefficients is not explicitly stored but alternates after each coeff and coeff[0] is positive, so ...,+,-,+,-,+,+,-,+,-,+,... hcoeff[0] is not explicitly stored but found by subtracting the sum of all stored coefficients with signs from 32 hcoeff[0]=32 - hcoeff[1] - hcoeff[2] - ... a good choice for hcoeff and htaps is htaps=6 hcoeff={40,-10, 2} an alternative which requires more computations at both encoder and decoder side and may or may not be better is htaps=8 hcoeff={42,-14, 6,-2}ref_frames minimum of the number of available reference frames and max_ref_frames for example the first frame after a key frame always has ref_frames=1spatial_decomposition_type wavelet type 0 is a 9/7 symmetric compact integer wavelet 1 is a 5/3 symmetric compact integer wavelet others are reserved stored as delta from last, last is reset to 0 if always_reset||keyframeqlog quality(logarithmic quantizer scale) stored as delta from last, last is reset to 0 if always_reset||keyframemv_scale stored as delta from last, last is reset to 0 if always_reset||keyframe FIXME check that everything works fine if this changes between framesqbias dequantization bias stored as delta from last, last is reset to 0 if always_reset||keyframeblock_max_depth maximum depth of the block tree stored as delta from last, last is reset to 0 if always_reset||keyframequant_table quantization tableHighlevel bitstream structure:==============================--------------------------------------------|Header|--------------------------------------------|------------------------------------|||Block0||||split?||||yes no||||......... intra?||||:Block01 :yes no||||:Block02 :....... ..........||||:Block03 ::y DC ::ref index:||||:Block04 ::cb DC ::motion x :||||......... :cr DC ::motion y :||||....... ..........|||------------------------------------||------------------------------------|||Block1|||...|--------------------------------------------|------------ ------------ ------------|||Y subbands||Cb subbands||Cr subbands||||--- ---||--- ---||--- ---|||||LL0||HL0||||LL0||HL0||||LL0||HL0|||||--- ---||--- ---||--- ---||||--- ---||--- ---||--- ---|||||LH0||HH0||||LH0||HH0||||LH0||HH0|||||--- ---||--- ---||--- ---||||--- ---||--- ---||--- ---|||||HL1||LH1||||HL1||LH1||||HL1||LH1|||||--- ---||--- ---||--- ---||||--- ---||--- ---||--- ---|||||HH1||HL2||||HH1||HL2||||HH1||HL2|||||...||...||...|||------------ ------------ ------------|--------------------------------------------Decoding process:=================------------|||Subbands|------------||||------------|Intra DC||||LL0 subband prediction ------------|\ Dequantization ------------------- \||Reference frames|\ IDWT|------- -------|Motion \|||Frame 0||Frame 1||Compensation . OBMC v -------|------- -------|--------------. \------> Frame n output Frame Frame<----------------------------------/|...|------------------- Range Coder:============Binary Range Coder:------------------- The implemented range coder is an adapted version based upon "Range encoding: an algorithm for removing redundancy from a digitised message." by G. N. N. Martin. The symbols encoded by the Snow range coder are bits(0|1). The associated probabilities are not fix but change depending on the symbol mix seen so far. bit seen|new state ---------+----------------------------------------------- 0|256 - state_transition_table[256 - old_state];1|state_transition_table[old_state];state_transition_table={ 0, 0, 0, 0, 0, 0, 0, 0, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 190, 191, 192, 194, 194, 195, 196, 197, 198, 199, 200, 201, 202, 202, 204, 205, 206, 207, 208, 209, 209, 210, 211, 212, 213, 215, 215, 216, 217, 218, 219, 220, 220, 222, 223, 224, 225, 226, 227, 227, 229, 229, 230, 231, 232, 234, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 248, 0, 0, 0, 0, 0, 0, 0};FIXME Range Coding of integers:------------------------- FIXME Neighboring Blocks:===================left and top are set to the respective blocks unless they are outside of the image in which case they are set to the Null block top-left is set to the top left block unless it is outside of the image in which case it is set to the left block if this block has no larger parent block or it is at the left side of its parent block and the top right block is not outside of the image then the top right block is used for top-right else the top-left block is used Null block y, cb, cr are 128 level, ref, mx and my are 0 Motion Vector Prediction:=========================1. the motion vectors of all the neighboring blocks are scaled to compensate for the difference of reference frames scaled_mv=(mv *(256 *(current_reference+1)/(mv.reference+1))+128)> the median of the scaled left
void ff_snow_reset_contexts(SnowContext *s)
static void encode_qlogs(SnowContext *s)
int av_frame_replace(AVFrame *dst, const AVFrame *src)
Ensure the destination frame refers to the same data described by the source frame,...
main external API structure.
@ AV_OPT_TYPE_INT
Underlying C type is int.
int height
picture size. must be a multiple of 16
static int ref[MAX_W *MAX_W]
static int get_rac_count(RangeCoder *c)
int mb_lmax
maximum MB Lagrange multiplier
static void put_symbol2(RangeCoder *c, uint8_t *state, int v, int log2)
IDirect3DDxgiInterfaceAccess _COM_Outptr_ void ** p
@ AV_PIX_FMT_YUV444P
planar YUV 4:4:4, 24bpp, (1 Cr & Cb sample per 1x1 Y samples)
@ AV_PICTURE_TYPE_P
Predicted.
static av_always_inline int same_block(BlockNode *a, BlockNode *b)
int16_t mx
Motion vector component X, see mv_scale.
int ff_epzs_motion_search(MPVEncContext *s, int *mx_ptr, int *my_ptr, int P[10][2], int src_index, int ref_index, const int16_t(*last_mv)[2], int ref_mv_scale, int size, int h)
This structure stores compressed data.
@ AV_OPT_TYPE_BOOL
Underlying C type is int.
static double cr(void *priv, double x, double y)
int ff_snow_frames_prepare(SnowContext *s)
uint8_t * emu_edge_buffer
static void quantize(SnowContext *s, SubBand *b, IDWTELEM *dst, DWTELEM *src, int stride, int bias)
static void dequantize(SnowContext *s, SubBand *b, IDWTELEM *src, int stride)
av_cold void ff_mpegvideoencdsp_init(MpegvideoEncDSPContext *c, AVCodecContext *avctx)
int width
picture width / height.
static const BlockNode null_block
int linesize[AV_NUM_DATA_POINTERS]
For video, a positive or negative value, which is typically indicating the size in bytes of each pict...
@ AV_PIX_FMT_YUV410P
planar YUV 4:1:0, 9bpp, (1 Cr & Cb sample per 4x4 Y samples)
The exact code depends on how similar the blocks are and how related they are to the block
enum OutputFormat out_format
output format
static const AVClass snowenc_class
@ AV_OPT_TYPE_STRING
Underlying C type is a uint8_t* that is either NULL or points to a C string allocated with the av_mal...
#define FF_QP2LAMBDA
factor to convert from H.263 QP to lambda
static av_always_inline int check_block_inter(SnowEncContext *enc, int mb_x, int mb_y, int p0, int p1, uint8_t(*obmc_edged)[MB_SIZE *2], int *best_rd)
@ AV_OPT_TYPE_CONST
Special option type for declaring named constants.
int ff_snow_alloc_blocks(SnowContext *s)
int ff_alloc_packet(AVCodecContext *avctx, AVPacket *avpkt, int64_t size)
Check AVPacket size and allocate data.
uint8_t ref
Reference frame index.
MPVEncContext s
The main slicecontext.
ptrdiff_t linesize
line size, in bytes, may be different from width
av_cold void ff_hpeldsp_init(HpelDSPContext *c, int flags)
#define AV_CODEC_FLAG_PASS1
Use internal 2pass ratecontrol in first pass mode.
int mb_height
number of MBs horizontally & vertically
av_cold void ff_rate_control_uninit(RateControlContext *rcc)
int unrestricted_mv
mv can point outside of the coded picture