412 |
const MACROBLOCK * pMB, |
const MACROBLOCK * pMB, |
413 |
const uint32_t x_pos, |
const uint32_t x_pos, |
414 |
const uint32_t y_pos, |
const uint32_t y_pos, |
415 |
const uint32_t acpred_flag, |
const uint32_t fcode, |
416 |
const uint32_t cbp, |
const uint32_t cbp, |
417 |
Bitstream * bs, |
Bitstream * bs, |
418 |
const uint32_t quant, |
const uint32_t quant, |
419 |
const uint32_t rounding, |
const uint32_t rounding, |
420 |
const int reduced_resolution, |
const int reduced_resolution) |
|
const int mcsel) |
|
421 |
{ |
{ |
422 |
|
|
423 |
DECLARE_ALIGNED_MATRIX(block, 6, 64, int16_t, CACHE_LINE); |
DECLARE_ALIGNED_MATRIX(block, 6, 64, int16_t, CACHE_LINE); |
429 |
uint32_t i; |
uint32_t i; |
430 |
uint32_t iQuant = pMB->quant; |
uint32_t iQuant = pMB->quant; |
431 |
uint8_t *pY_Cur, *pU_Cur, *pV_Cur; |
uint8_t *pY_Cur, *pU_Cur, *pV_Cur; |
|
uint8_t *pY_Ref, *pU_Ref, *pV_Ref; /* ref for GMC is _not_ pRef itself */ |
|
432 |
|
|
433 |
int uv_dx, uv_dy; |
int uv_dx, uv_dy; |
434 |
VECTOR mv[4]; /* local copy of mvs */ |
VECTOR mv[4]; /* local copy of mvs */ |
449 |
mv[i] = pMB->mvs[i]; |
mv[i] = pMB->mvs[i]; |
450 |
} |
} |
451 |
|
|
|
if (mcsel) { |
|
|
mv[0].x = mv[0].y = mv[1].x = mv[1].y = mv[2].x = mv[2].y = mv[3].x = mv[3].y = 0; |
|
|
/* position in ref is same as the block, set vector to (0,0) */ |
|
|
pY_Ref = dec->gmc.y; |
|
|
pU_Ref = dec->gmc.u; |
|
|
pV_Ref = dec->gmc.v; |
|
|
/* but reference itself isn't. It's warped... */ |
|
|
/* Btw., this is too slow! For GMC it should simply be transfer_16to8add() */ |
|
|
} else { |
|
|
pY_Ref = dec->refn[0].y; |
|
|
pU_Ref = dec->refn[0].u; |
|
|
pV_Ref = dec->refn[0].v; |
|
|
} |
|
|
|
|
452 |
if (pMB->mode == MODE_INTER || pMB->mode == MODE_INTER_Q) { |
if (pMB->mode == MODE_INTER || pMB->mode == MODE_INTER_Q) { |
453 |
|
|
454 |
uv_dx = mv[0].x / (1 + dec->quarterpel); |
uv_dx = mv[0].x / (1 + dec->quarterpel); |
460 |
start_timer(); |
start_timer(); |
461 |
if (reduced_resolution) |
if (reduced_resolution) |
462 |
{ |
{ |
463 |
interpolate32x32_switch(dec->cur.y, pY_Ref , 32*x_pos, 32*y_pos, |
interpolate32x32_switch(dec->cur.y, dec->refn[0].y, 32*x_pos, 32*y_pos, |
464 |
mv[0].x, mv[0].y, stride, rounding); |
mv[0].x, mv[0].y, stride, rounding); |
465 |
interpolate16x16_switch(dec->cur.u, pU_Ref , 16 * x_pos, 16 * y_pos, |
interpolate16x16_switch(dec->cur.u, dec->refn[0].u, 16 * x_pos, 16 * y_pos, |
466 |
uv_dx, uv_dy, stride2, rounding); |
uv_dx, uv_dy, stride2, rounding); |
467 |
interpolate16x16_switch(dec->cur.v, pV_Ref , 16 * x_pos, 16 * y_pos, |
interpolate16x16_switch(dec->cur.v, dec->refn[0].v, 16 * x_pos, 16 * y_pos, |
468 |
uv_dx, uv_dy, stride2, rounding); |
uv_dx, uv_dy, stride2, rounding); |
469 |
|
|
470 |
} |
} |
471 |
else |
else |
472 |
{ |
{ |
473 |
if(dec->quarterpel) { |
if(dec->quarterpel) { |
474 |
interpolate16x16_quarterpel(dec->cur.y, pY_Ref , dec->qtmp.y, dec->qtmp.y + 64, |
interpolate16x16_quarterpel(dec->cur.y, dec->refn[0].y, dec->qtmp.y, dec->qtmp.y + 64, |
475 |
dec->qtmp.y + 128, 16*x_pos, 16*y_pos, |
dec->qtmp.y + 128, 16*x_pos, 16*y_pos, |
476 |
mv[0].x, mv[0].y, stride, rounding); |
mv[0].x, mv[0].y, stride, rounding); |
477 |
} |
} |
478 |
else { |
else { |
479 |
interpolate16x16_switch(dec->cur.y, pY_Ref , 16*x_pos, 16*y_pos, |
interpolate16x16_switch(dec->cur.y, dec->refn[0].y, 16*x_pos, 16*y_pos, |
480 |
mv[0].x, mv[0].y, stride, rounding); |
mv[0].x, mv[0].y, stride, rounding); |
481 |
} |
} |
482 |
|
|
483 |
interpolate8x8_switch(dec->cur.u, pU_Ref , 8 * x_pos, 8 * y_pos, |
interpolate8x8_switch(dec->cur.u, dec->refn[0].u, 8 * x_pos, 8 * y_pos, |
484 |
uv_dx, uv_dy, stride2, rounding); |
uv_dx, uv_dy, stride2, rounding); |
485 |
interpolate8x8_switch(dec->cur.v, pV_Ref , 8 * x_pos, 8 * y_pos, |
interpolate8x8_switch(dec->cur.v, dec->refn[0].v, 8 * x_pos, 8 * y_pos, |
486 |
uv_dx, uv_dy, stride2, rounding); |
uv_dx, uv_dy, stride2, rounding); |
487 |
} |
} |
488 |
stop_comp_timer(); |
stop_comp_timer(); |
507 |
start_timer(); |
start_timer(); |
508 |
if (reduced_resolution) |
if (reduced_resolution) |
509 |
{ |
{ |
510 |
interpolate16x16_switch(dec->cur.y, pY_Ref , 32*x_pos, 32*y_pos, |
interpolate16x16_switch(dec->cur.y, dec->refn[0].y, 32*x_pos, 32*y_pos, |
511 |
mv[0].x, mv[0].y, stride, rounding); |
mv[0].x, mv[0].y, stride, rounding); |
512 |
interpolate16x16_switch(dec->cur.y, pY_Ref , 32*x_pos + 16, 32*y_pos, |
interpolate16x16_switch(dec->cur.y, dec->refn[0].y , 32*x_pos + 16, 32*y_pos, |
513 |
mv[1].x, mv[1].y, stride, rounding); |
mv[1].x, mv[1].y, stride, rounding); |
514 |
interpolate16x16_switch(dec->cur.y, pY_Ref , 32*x_pos, 32*y_pos + 16, |
interpolate16x16_switch(dec->cur.y, dec->refn[0].y , 32*x_pos, 32*y_pos + 16, |
515 |
mv[2].x, mv[2].y, stride, rounding); |
mv[2].x, mv[2].y, stride, rounding); |
516 |
interpolate16x16_switch(dec->cur.y, pY_Ref , 32*x_pos + 16, 32*y_pos + 16, |
interpolate16x16_switch(dec->cur.y, dec->refn[0].y , 32*x_pos + 16, 32*y_pos + 16, |
517 |
mv[3].x, mv[3].y, stride, rounding); |
mv[3].x, mv[3].y, stride, rounding); |
518 |
interpolate16x16_switch(dec->cur.u, pU_Ref , 16 * x_pos, 16 * y_pos, |
interpolate16x16_switch(dec->cur.u, dec->refn[0].u , 16 * x_pos, 16 * y_pos, |
519 |
uv_dx, uv_dy, stride2, rounding); |
uv_dx, uv_dy, stride2, rounding); |
520 |
interpolate16x16_switch(dec->cur.v, pV_Ref , 16 * x_pos, 16 * y_pos, |
interpolate16x16_switch(dec->cur.v, dec->refn[0].v , 16 * x_pos, 16 * y_pos, |
521 |
uv_dx, uv_dy, stride2, rounding); |
uv_dx, uv_dy, stride2, rounding); |
522 |
|
|
523 |
// set_block(pY_Cur, stride, 32, 32, 127); |
// set_block(pY_Cur, stride, 32, 32, 127); |
525 |
else |
else |
526 |
{ |
{ |
527 |
if(dec->quarterpel) { |
if(dec->quarterpel) { |
528 |
interpolate8x8_quarterpel(dec->cur.y, pY_Ref , dec->qtmp.y, dec->qtmp.y + 64, |
interpolate8x8_quarterpel(dec->cur.y, dec->refn[0].y , dec->qtmp.y, dec->qtmp.y + 64, |
529 |
dec->qtmp.y + 128, 16*x_pos, 16*y_pos, |
dec->qtmp.y + 128, 16*x_pos, 16*y_pos, |
530 |
mv[0].x, mv[0].y, stride, rounding); |
mv[0].x, mv[0].y, stride, rounding); |
531 |
interpolate8x8_quarterpel(dec->cur.y, pY_Ref , dec->qtmp.y, dec->qtmp.y + 64, |
interpolate8x8_quarterpel(dec->cur.y, dec->refn[0].y , dec->qtmp.y, dec->qtmp.y + 64, |
532 |
dec->qtmp.y + 128, 16*x_pos + 8, 16*y_pos, |
dec->qtmp.y + 128, 16*x_pos + 8, 16*y_pos, |
533 |
mv[1].x, mv[1].y, stride, rounding); |
mv[1].x, mv[1].y, stride, rounding); |
534 |
interpolate8x8_quarterpel(dec->cur.y, pY_Ref , dec->qtmp.y, dec->qtmp.y + 64, |
interpolate8x8_quarterpel(dec->cur.y, dec->refn[0].y , dec->qtmp.y, dec->qtmp.y + 64, |
535 |
dec->qtmp.y + 128, 16*x_pos, 16*y_pos + 8, |
dec->qtmp.y + 128, 16*x_pos, 16*y_pos + 8, |
536 |
mv[2].x, mv[2].y, stride, rounding); |
mv[2].x, mv[2].y, stride, rounding); |
537 |
interpolate8x8_quarterpel(dec->cur.y, pY_Ref , dec->qtmp.y, dec->qtmp.y + 64, |
interpolate8x8_quarterpel(dec->cur.y, dec->refn[0].y , dec->qtmp.y, dec->qtmp.y + 64, |
538 |
dec->qtmp.y + 128, 16*x_pos + 8, 16*y_pos + 8, |
dec->qtmp.y + 128, 16*x_pos + 8, 16*y_pos + 8, |
539 |
mv[3].x, mv[3].y, stride, rounding); |
mv[3].x, mv[3].y, stride, rounding); |
540 |
} |
} |
541 |
else { |
else { |
542 |
interpolate8x8_switch(dec->cur.y, pY_Ref , 16*x_pos, 16*y_pos, |
interpolate8x8_switch(dec->cur.y, dec->refn[0].y , 16*x_pos, 16*y_pos, |
543 |
mv[0].x, mv[0].y, stride, rounding); |
mv[0].x, mv[0].y, stride, rounding); |
544 |
interpolate8x8_switch(dec->cur.y, pY_Ref , 16*x_pos + 8, 16*y_pos, |
interpolate8x8_switch(dec->cur.y, dec->refn[0].y , 16*x_pos + 8, 16*y_pos, |
545 |
mv[1].x, mv[1].y, stride, rounding); |
mv[1].x, mv[1].y, stride, rounding); |
546 |
interpolate8x8_switch(dec->cur.y, pY_Ref , 16*x_pos, 16*y_pos + 8, |
interpolate8x8_switch(dec->cur.y, dec->refn[0].y , 16*x_pos, 16*y_pos + 8, |
547 |
mv[2].x, mv[2].y, stride, rounding); |
mv[2].x, mv[2].y, stride, rounding); |
548 |
interpolate8x8_switch(dec->cur.y, pY_Ref , 16*x_pos + 8, 16*y_pos + 8, |
interpolate8x8_switch(dec->cur.y, dec->refn[0].y , 16*x_pos + 8, 16*y_pos + 8, |
549 |
mv[3].x, mv[3].y, stride, rounding); |
mv[3].x, mv[3].y, stride, rounding); |
550 |
} |
} |
551 |
|
|
552 |
interpolate8x8_switch(dec->cur.u, pU_Ref , 8 * x_pos, 8 * y_pos, |
interpolate8x8_switch(dec->cur.u, dec->refn[0].u , 8 * x_pos, 8 * y_pos, |
553 |
uv_dx, uv_dy, stride2, rounding); |
uv_dx, uv_dy, stride2, rounding); |
554 |
interpolate8x8_switch(dec->cur.v, pV_Ref , 8 * x_pos, 8 * y_pos, |
interpolate8x8_switch(dec->cur.v, dec->refn[0].v , 8 * x_pos, 8 * y_pos, |
555 |
uv_dx, uv_dy, stride2, rounding); |
uv_dx, uv_dy, stride2, rounding); |
556 |
} |
} |
557 |
stop_comp_timer(); |
stop_comp_timer(); |
621 |
stop_transfer_timer(); |
stop_transfer_timer(); |
622 |
} |
} |
623 |
|
|
624 |
|
static __inline int gmc_sanitize(int value, int quarterpel, int fcode) |
625 |
|
{ |
626 |
|
int length = 1 << (fcode+4); |
627 |
|
|
628 |
|
// if (quarterpel) value *= 2; |
629 |
|
|
630 |
|
if (value < -length) |
631 |
|
return -length; |
632 |
|
else if (value >= length) |
633 |
|
return length-1; |
634 |
|
else return value; |
635 |
|
} |
636 |
|
|
637 |
|
|
638 |
|
static void |
639 |
|
decoder_mbgmc(DECODER * dec, |
640 |
|
MACROBLOCK * const pMB, |
641 |
|
const uint32_t x_pos, |
642 |
|
const uint32_t y_pos, |
643 |
|
const uint32_t fcode, |
644 |
|
const uint32_t cbp, |
645 |
|
Bitstream * bs, |
646 |
|
const uint32_t quant, |
647 |
|
const uint32_t rounding, |
648 |
|
const int reduced_resolution) /* no reduced res support */ |
649 |
|
{ |
650 |
|
|
651 |
|
DECLARE_ALIGNED_MATRIX(block, 6, 64, int16_t, CACHE_LINE); |
652 |
|
DECLARE_ALIGNED_MATRIX(data, 6, 64, int16_t, CACHE_LINE); |
653 |
|
|
654 |
|
const uint32_t stride = dec->edged_width; |
655 |
|
const uint32_t stride2 = stride / 2; |
656 |
|
const uint32_t next_block = stride * (reduced_resolution ? 16 : 8); |
657 |
|
uint32_t i; |
658 |
|
const uint32_t iQuant = pMB->quant; |
659 |
|
uint8_t *const pY_Cur=dec->cur.y + (y_pos << 4) * stride + (x_pos << 4); |
660 |
|
uint8_t *const pU_Cur=dec->cur.u + (y_pos << 3) * stride2 + (x_pos << 3); |
661 |
|
uint8_t *const pV_Cur=dec->cur.v + (y_pos << 3) * stride2 + (x_pos << 3); |
662 |
|
|
663 |
|
pMB->mvs[0] = pMB->mvs[1] = pMB->mvs[2] = pMB->mvs[3] = pMB->amv; |
664 |
|
|
665 |
|
start_timer(); |
666 |
|
|
667 |
|
/* this is where the calculations are done */ |
668 |
|
|
669 |
|
{ |
670 |
|
pMB->amv = generate_GMCimageMB(&dec->gmc_data, &dec->refn[0], x_pos, y_pos, |
671 |
|
stride, stride2, dec->quarterpel, rounding, &dec->cur); |
672 |
|
|
673 |
|
pMB->amv.x = gmc_sanitize(pMB->amv.x, dec->quarterpel, fcode); |
674 |
|
pMB->amv.y = gmc_sanitize(pMB->amv.y, dec->quarterpel, fcode); |
675 |
|
} |
676 |
|
pMB->mvs[0] = pMB->mvs[1] = pMB->mvs[2] = pMB->mvs[3] = pMB->amv; |
677 |
|
|
678 |
|
/* transfer16x16_copy(pY_Cur, dec->gmc.y + (y_pos << 4)*stride + (x_pos << 4), stride); |
679 |
|
transfer8x8_copy(pU_Cur, dec->gmc.u + (y_pos << 3)*stride2 + (x_pos << 3), stride2); |
680 |
|
transfer8x8_copy(pV_Cur, dec->gmc.v + (y_pos << 3)*stride2 + (x_pos << 3), stride2); |
681 |
|
*/ |
682 |
|
|
683 |
|
|
684 |
|
stop_transfer_timer(); |
685 |
|
|
686 |
|
if (!cbp) return; |
687 |
|
|
688 |
|
for (i = 0; i < 6; i++) { |
689 |
|
int direction = dec->alternate_vertical_scan ? 2 : 0; |
690 |
|
|
691 |
|
if (cbp & (1 << (5 - i))) // coded |
692 |
|
{ |
693 |
|
memset(&block[i * 64], 0, 64 * sizeof(int16_t)); // clear |
694 |
|
|
695 |
|
start_timer(); |
696 |
|
get_inter_block(bs, &block[i * 64], direction); |
697 |
|
stop_coding_timer(); |
698 |
|
|
699 |
|
start_timer(); |
700 |
|
if (dec->quant_type == 0) { |
701 |
|
dequant_inter(&data[i * 64], &block[i * 64], iQuant); |
702 |
|
} else { |
703 |
|
dequant4_inter(&data[i * 64], &block[i * 64], iQuant); |
704 |
|
} |
705 |
|
stop_iquant_timer(); |
706 |
|
|
707 |
|
start_timer(); |
708 |
|
idct(&data[i * 64]); |
709 |
|
stop_idct_timer(); |
710 |
|
} |
711 |
|
} |
712 |
|
|
713 |
|
/* interlace + GMC is this possible ??? */ |
714 |
|
/* if (dec->interlacing && pMB->field_dct) { |
715 |
|
next_block = stride; |
716 |
|
stride *= 2; |
717 |
|
} |
718 |
|
*/ |
719 |
|
start_timer(); |
720 |
|
if (cbp & 32) |
721 |
|
transfer_16to8add(pY_Cur, &data[0 * 64], stride); |
722 |
|
if (cbp & 16) |
723 |
|
transfer_16to8add(pY_Cur + 8, &data[1 * 64], stride); |
724 |
|
if (cbp & 8) |
725 |
|
transfer_16to8add(pY_Cur + next_block, &data[2 * 64], stride); |
726 |
|
if (cbp & 4) |
727 |
|
transfer_16to8add(pY_Cur + 8 + next_block, &data[3 * 64], stride); |
728 |
|
if (cbp & 2) |
729 |
|
transfer_16to8add(pU_Cur, &data[4 * 64], stride2); |
730 |
|
if (cbp & 1) |
731 |
|
transfer_16to8add(pV_Cur, &data[5 * 64], stride2); |
732 |
|
stop_transfer_timer(); |
733 |
|
} |
734 |
|
|
735 |
|
|
736 |
void |
void |
737 |
decoder_iframe(DECODER * dec, |
decoder_iframe(DECODER * dec, |
862 |
|
|
863 |
|
|
864 |
|
|
|
static __inline int gmc_sanitize(int value, int quarterpel, int fcode) |
|
|
{ |
|
|
int length = 1 << (fcode+4); |
|
|
|
|
|
if (quarterpel) value *= 2; |
|
|
|
|
|
if (value < -length) |
|
|
return -length; |
|
|
else if (value >= length) |
|
|
return length-1; |
|
|
else return value; |
|
|
} |
|
865 |
|
|
866 |
|
|
867 |
/* for P_VOP set gmc_warp to NULL */ |
/* for P_VOP set gmc_warp to NULL */ |
882 |
uint32_t mb_width = dec->mb_width; |
uint32_t mb_width = dec->mb_width; |
883 |
uint32_t mb_height = dec->mb_height; |
uint32_t mb_height = dec->mb_height; |
884 |
|
|
|
static int framecount=0; |
|
885 |
if (reduced_resolution) |
if (reduced_resolution) |
886 |
{ |
{ |
887 |
mb_width = (dec->width + 31) / 32; |
mb_width = (dec->width + 31) / 32; |
895 |
|
|
896 |
if (gmc_warp) |
if (gmc_warp) |
897 |
{ |
{ |
898 |
char filename[80]; |
|
|
sprintf(filename,"dGMC%05d.pgm",framecount); |
|
899 |
// accuracy: 0==1/2, 1=1/4, 2=1/8, 3=1/16 |
// accuracy: 0==1/2, 1=1/4, 2=1/8, 3=1/16 |
900 |
if ( (dec->sprite_warping_accuracy != 3) || (dec->sprite_warping_points != 2) ) |
if ( (dec->sprite_warping_accuracy != 3) || (dec->sprite_warping_points != 2) ) |
901 |
{ |
{ |
908 |
(2 << dec->sprite_warping_accuracy), gmc_warp, |
(2 << dec->sprite_warping_accuracy), gmc_warp, |
909 |
dec->width, dec->height, &dec->gmc_data); |
dec->width, dec->height, &dec->gmc_data); |
910 |
|
|
911 |
|
/* image warping is done block-based in decoder_mbgmc(), now */ |
912 |
|
/* |
913 |
generate_GMCimage(&dec->gmc_data, &dec->refn[0], |
generate_GMCimage(&dec->gmc_data, &dec->refn[0], |
914 |
mb_width, mb_height, |
mb_width, mb_height, |
915 |
dec->edged_width, dec->edged_width/2, |
dec->edged_width, dec->edged_width/2, |
916 |
fcode, 0, 0, |
fcode, dec->quarterpel, 0, |
917 |
rounding, dec->mbs, &dec->gmc); |
rounding, dec->mbs, &dec->gmc); |
|
|
|
|
/* |
|
|
sprintf(filename,"dGMC%05d.pgm",framecount); |
|
|
image_dump_yuvpgm(&dec->gmc, |
|
|
dec->edged_width, dec->width, dec->height, filename); |
|
|
|
|
|
sprintf(filename,"dREF%05d.pgm",framecount); |
|
|
image_dump_yuvpgm(&dec->refn[0], |
|
|
dec->edged_width, dec->width, dec->height, filename); |
|
|
sprintf(filename,"dCUR%05d.pgm",framecount); |
|
|
image_dump_yuvpgm(&dec->cur, |
|
|
dec->edged_width, dec->width, dec->height, filename); |
|
|
framecount++; |
|
918 |
*/ |
*/ |
919 |
} |
} |
920 |
|
|
1008 |
} |
} |
1009 |
} |
} |
1010 |
|
|
1011 |
if (mb->mode == MODE_INTER || mb->mode == MODE_INTER_Q) { |
if (mcsel) { |
1012 |
|
decoder_mbgmc(dec, mb, x, y, fcode, cbp, bs, quant, |
1013 |
|
rounding, reduced_resolution); |
1014 |
|
continue; |
1015 |
|
|
1016 |
if (mcsel) |
} else if (mb->mode == MODE_INTER || mb->mode == MODE_INTER_Q) { |
|
{ |
|
|
mb->mvs[0] = mb->mvs[1] = mb->mvs[2] = mb->mvs[3] = mb->amv; |
|
|
/* already clipped to fcode */ |
|
1017 |
|
|
1018 |
} else if (dec->interlacing && mb->field_pred) { |
if (dec->interlacing && mb->field_pred) { |
1019 |
get_motion_vector(dec, bs, x, y, 0, &mb->mvs[0], |
get_motion_vector(dec, bs, x, y, 0, &mb->mvs[0], |
1020 |
fcode, bound); |
fcode, bound); |
1021 |
get_motion_vector(dec, bs, x, y, 0, &mb->mvs[1], |
get_motion_vector(dec, bs, x, y, 0, &mb->mvs[1], |
1042 |
continue; |
continue; |
1043 |
} |
} |
1044 |
|
|
1045 |
decoder_mbinter(dec, mb, x, y, acpred_flag, cbp, bs, quant, |
decoder_mbinter(dec, mb, x, y, fcode, cbp, bs, quant, |
1046 |
rounding, reduced_resolution, mcsel); |
rounding, reduced_resolution); |
1047 |
|
|
1048 |
} |
} |
1049 |
else if (gmc_warp) /* a not coded S(GMC)-VOP macroblock */ |
else if (gmc_warp) /* a not coded S(GMC)-VOP macroblock */ |
1050 |
{ |
{ |
1051 |
mb->mode = MODE_NOT_CODED_GMC; |
mb->mode = MODE_NOT_CODED_GMC; |
|
mb->mvs[0] = mb->mvs[1] = mb->mvs[2] = mb->mvs[3] = mb->amv; |
|
1052 |
|
|
1053 |
start_timer(); |
start_timer(); |
1054 |
|
|
1055 |
transfer16x16_copy(dec->cur.y + (16*y)*dec->edged_width + (16*x), |
decoder_mbgmc(dec, mb, x, y, fcode, 0x00, bs, quant, |
1056 |
dec->gmc.y + (16*y)*dec->edged_width + (16*x), |
rounding, reduced_resolution); |
|
dec->edged_width); |
|
|
|
|
|
transfer8x8_copy(dec->cur.u + (8*y)*dec->edged_width/2 + (8*x), |
|
|
dec->gmc.u + (8*y)*dec->edged_width/2 + (8*x), |
|
|
dec->edged_width/2); |
|
|
|
|
|
transfer8x8_copy(dec->cur.v + (8*y)*dec->edged_width/2 + (8*x), |
|
|
dec->gmc.v + (8*y)*dec->edged_width/2 + (8*x), |
|
|
dec->edged_width/2); |
|
1057 |
|
|
1058 |
stop_transfer_timer(); |
stop_transfer_timer(); |
1059 |
|
|
1611 |
mb->mvs[0] = mb->mvs[1] = mb->mvs[2] = mb->mvs[3] = zeromv; |
mb->mvs[0] = mb->mvs[1] = mb->mvs[2] = mb->mvs[3] = zeromv; |
1612 |
|
|
1613 |
// skip if the co-located P_VOP macroblock is not coded |
// skip if the co-located P_VOP macroblock is not coded |
1614 |
// note: gmc+not_coded isn't skipped |
// if not codec in co-located S_VOP macroblock is _not_ automatically skipped |
1615 |
|
|
1616 |
if (last_mb->mode == MODE_NOT_CODED) { |
if (last_mb->mode == MODE_NOT_CODED) { |
1617 |
//DEBUG2("Skip MB in B-frame at (X,Y)=!",x,y); |
//DEBUG2("Skip MB in B-frame at (X,Y)=!",x,y); |