138 |
|
|
139 |
/* Quantize the block */ |
/* Quantize the block */ |
140 |
start_timer(); |
start_timer(); |
141 |
quant[mpeg](&data[0 * 64], &qcoeff[0 * 64], pMB->quant, scaler_lum); |
quant[mpeg](&data[0 * 64], &qcoeff[0 * 64], pMB->quant, scaler_lum, pParam->mpeg_quant_matrices); |
142 |
quant[mpeg](&data[1 * 64], &qcoeff[1 * 64], pMB->quant, scaler_lum); |
quant[mpeg](&data[1 * 64], &qcoeff[1 * 64], pMB->quant, scaler_lum, pParam->mpeg_quant_matrices); |
143 |
quant[mpeg](&data[2 * 64], &qcoeff[2 * 64], pMB->quant, scaler_lum); |
quant[mpeg](&data[2 * 64], &qcoeff[2 * 64], pMB->quant, scaler_lum, pParam->mpeg_quant_matrices); |
144 |
quant[mpeg](&data[3 * 64], &qcoeff[3 * 64], pMB->quant, scaler_lum); |
quant[mpeg](&data[3 * 64], &qcoeff[3 * 64], pMB->quant, scaler_lum, pParam->mpeg_quant_matrices); |
145 |
quant[mpeg](&data[4 * 64], &qcoeff[4 * 64], pMB->quant, scaler_chr); |
quant[mpeg](&data[4 * 64], &qcoeff[4 * 64], pMB->quant, scaler_chr, pParam->mpeg_quant_matrices); |
146 |
quant[mpeg](&data[5 * 64], &qcoeff[5 * 64], pMB->quant, scaler_chr); |
quant[mpeg](&data[5 * 64], &qcoeff[5 * 64], pMB->quant, scaler_chr, pParam->mpeg_quant_matrices); |
147 |
stop_quant_timer(); |
stop_quant_timer(); |
148 |
} |
} |
149 |
|
|
168 |
scaler_chr = get_dc_scaler(iQuant, 0); |
scaler_chr = get_dc_scaler(iQuant, 0); |
169 |
|
|
170 |
start_timer(); |
start_timer(); |
171 |
dequant[mpeg](&qcoeff[0 * 64], &data[0 * 64], iQuant, scaler_lum); |
dequant[mpeg](&qcoeff[0 * 64], &data[0 * 64], iQuant, scaler_lum, pParam->mpeg_quant_matrices); |
172 |
dequant[mpeg](&qcoeff[1 * 64], &data[1 * 64], iQuant, scaler_lum); |
dequant[mpeg](&qcoeff[1 * 64], &data[1 * 64], iQuant, scaler_lum, pParam->mpeg_quant_matrices); |
173 |
dequant[mpeg](&qcoeff[2 * 64], &data[2 * 64], iQuant, scaler_lum); |
dequant[mpeg](&qcoeff[2 * 64], &data[2 * 64], iQuant, scaler_lum, pParam->mpeg_quant_matrices); |
174 |
dequant[mpeg](&qcoeff[3 * 64], &data[3 * 64], iQuant, scaler_lum); |
dequant[mpeg](&qcoeff[3 * 64], &data[3 * 64], iQuant, scaler_lum, pParam->mpeg_quant_matrices); |
175 |
dequant[mpeg](&qcoeff[4 * 64], &data[4 * 64], iQuant, scaler_chr); |
dequant[mpeg](&qcoeff[4 * 64], &data[4 * 64], iQuant, scaler_chr, pParam->mpeg_quant_matrices); |
176 |
dequant[mpeg](&qcoeff[5 * 64], &data[5 * 64], iQuant, scaler_chr); |
dequant[mpeg](&qcoeff[5 * 64], &data[5 * 64], iQuant, scaler_chr, pParam->mpeg_quant_matrices); |
177 |
stop_iquant_timer(); |
stop_iquant_timer(); |
178 |
} |
} |
179 |
|
|
183 |
int Q, |
int Q, |
184 |
const uint16_t * const Zigzag, |
const uint16_t * const Zigzag, |
185 |
const uint16_t * const QuantMatrix, |
const uint16_t * const QuantMatrix, |
186 |
int Non_Zero); |
int Non_Zero, |
187 |
|
int Sum); |
188 |
|
|
189 |
/* Quantize all blocks -- Inter mode */ |
/* Quantize all blocks -- Inter mode */ |
190 |
static __inline uint8_t |
static __inline uint8_t |
215 |
/* Quantize the block */ |
/* Quantize the block */ |
216 |
start_timer(); |
start_timer(); |
217 |
|
|
218 |
sum = quant[mpeg](&qcoeff[i*64], &data[i*64], pMB->quant); |
sum = quant[mpeg](&qcoeff[i*64], &data[i*64], pMB->quant, pParam->mpeg_quant_matrices); |
219 |
|
|
220 |
if(sum && (frame->vop_flags & XVID_VOP_TRELLISQUANT)) { |
if(sum && (pMB->quant > 2) && (frame->vop_flags & XVID_VOP_TRELLISQUANT)) { |
221 |
|
const uint16_t *matrix; |
222 |
const static uint16_t h263matrix[] = |
const static uint16_t h263matrix[] = |
223 |
{ |
{ |
224 |
16, 16, 16, 16, 16, 16, 16, 16, |
16, 16, 16, 16, 16, 16, 16, 16, |
230 |
16, 16, 16, 16, 16, 16, 16, 16, |
16, 16, 16, 16, 16, 16, 16, 16, |
231 |
16, 16, 16, 16, 16, 16, 16, 16 |
16, 16, 16, 16, 16, 16, 16, 16 |
232 |
}; |
}; |
233 |
|
|
234 |
|
matrix = (mpeg)?get_inter_matrix(pParam->mpeg_quant_matrices):h263matrix; |
235 |
sum = dct_quantize_trellis_c(&qcoeff[i*64], &data[i*64], |
sum = dct_quantize_trellis_c(&qcoeff[i*64], &data[i*64], |
236 |
pMB->quant, &scan_tables[0][0], |
pMB->quant, &scan_tables[0][0], |
237 |
(mpeg)?(uint16_t*)get_inter_matrix():h263matrix, |
matrix, |
238 |
63); |
63, |
239 |
|
sum); |
240 |
} |
} |
241 |
stop_quant_timer(); |
stop_quant_timer(); |
242 |
|
|
286 |
mpeg = !!(pParam->vol_flags & XVID_VOL_MPEGQUANT); |
mpeg = !!(pParam->vol_flags & XVID_VOL_MPEGQUANT); |
287 |
|
|
288 |
start_timer(); |
start_timer(); |
289 |
if(cbp & (1 << (5 - 0))) dequant[mpeg](&data[0 * 64], &qcoeff[0 * 64], iQuant); |
if(cbp & (1 << (5 - 0))) dequant[mpeg](&data[0 * 64], &qcoeff[0 * 64], iQuant, pParam->mpeg_quant_matrices); |
290 |
if(cbp & (1 << (5 - 1))) dequant[mpeg](&data[1 * 64], &qcoeff[1 * 64], iQuant); |
if(cbp & (1 << (5 - 1))) dequant[mpeg](&data[1 * 64], &qcoeff[1 * 64], iQuant, pParam->mpeg_quant_matrices); |
291 |
if(cbp & (1 << (5 - 2))) dequant[mpeg](&data[2 * 64], &qcoeff[2 * 64], iQuant); |
if(cbp & (1 << (5 - 2))) dequant[mpeg](&data[2 * 64], &qcoeff[2 * 64], iQuant, pParam->mpeg_quant_matrices); |
292 |
if(cbp & (1 << (5 - 3))) dequant[mpeg](&data[3 * 64], &qcoeff[3 * 64], iQuant); |
if(cbp & (1 << (5 - 3))) dequant[mpeg](&data[3 * 64], &qcoeff[3 * 64], iQuant, pParam->mpeg_quant_matrices); |
293 |
if(cbp & (1 << (5 - 4))) dequant[mpeg](&data[4 * 64], &qcoeff[4 * 64], iQuant); |
if(cbp & (1 << (5 - 4))) dequant[mpeg](&data[4 * 64], &qcoeff[4 * 64], iQuant, pParam->mpeg_quant_matrices); |
294 |
if(cbp & (1 << (5 - 5))) dequant[mpeg](&data[5 * 64], &qcoeff[5 * 64], iQuant); |
if(cbp & (1 << (5 - 5))) dequant[mpeg](&data[5 * 64], &qcoeff[5 * 64], iQuant, pParam->mpeg_quant_matrices); |
295 |
stop_iquant_timer(); |
stop_iquant_timer(); |
296 |
} |
} |
297 |
|
|
374 |
|
|
375 |
transfer_operation_16to8_t *transfer_op = NULL; |
transfer_operation_16to8_t *transfer_op = NULL; |
376 |
|
|
|
if (pMB->field_dct) { |
|
|
next_block = stride; |
|
|
stride *= 2; |
|
|
} |
|
|
|
|
377 |
/* Makes this vars booleans */ |
/* Makes this vars booleans */ |
378 |
vop_reduced = !!(frame->vop_flags & XVID_VOP_REDUCED); |
vop_reduced = !!(frame->vop_flags & XVID_VOP_REDUCED); |
379 |
|
|
382 |
pU_Cur = pCurrent->u + (y_pos << (3+vop_reduced)) * stride2 + (x_pos << (3+vop_reduced)); |
pU_Cur = pCurrent->u + (y_pos << (3+vop_reduced)) * stride2 + (x_pos << (3+vop_reduced)); |
383 |
pV_Cur = pCurrent->v + (y_pos << (3+vop_reduced)) * stride2 + (x_pos << (3+vop_reduced)); |
pV_Cur = pCurrent->v + (y_pos << (3+vop_reduced)) * stride2 + (x_pos << (3+vop_reduced)); |
384 |
|
|
385 |
|
if (pMB->field_dct) { |
386 |
|
next_block = stride; |
387 |
|
stride *= 2; |
388 |
|
} |
389 |
|
|
390 |
/* Block size */ |
/* Block size */ |
391 |
cst = 8<<vop_reduced; |
cst = 8<<vop_reduced; |
392 |
|
|
652 |
* IEEE Transactions on Image Processing, Vol.9, No.8, Aug. 2000. |
* IEEE Transactions on Image Processing, Vol.9, No.8, Aug. 2000. |
653 |
* |
* |
654 |
* we are at stake with a simplified Bellmand-Ford / Dijkstra Single |
* we are at stake with a simplified Bellmand-Ford / Dijkstra Single |
655 |
* Source Shorted Path algo. But due to the underlying graph structure |
* Source Shortest Path algo. But due to the underlying graph structure |
656 |
* ("Trellis"), it can be turned into a dynamic programming algo, |
* ("Trellis"), it can be turned into a dynamic programming algo, |
657 |
* partially saving the explicit graph's nodes representation. And |
* partially saving the explicit graph's nodes representation. And |
658 |
* without using a heap, since the open frontier of the DAG is always |
* without using a heap, since the open frontier of the DAG is always |
659 |
* known, and of fixed sized. |
* known, and of fixed size. |
660 |
*--------------------------------------------------------------------------*/ |
*--------------------------------------------------------------------------*/ |
661 |
|
|
662 |
|
|
783 |
return -1; |
return -1; |
784 |
} |
} |
785 |
|
|
|
static int __inline |
|
|
Compute_Sum(const int16_t *C, int last) |
|
|
{ |
|
|
int sum = 0; |
|
|
|
|
|
while(last--) |
|
|
sum += abs(C[last]); |
|
|
|
|
|
return(sum); |
|
|
} |
|
|
|
|
786 |
/* this routine has been strippen of all debug code */ |
/* this routine has been strippen of all debug code */ |
787 |
static int |
static int |
788 |
dct_quantize_trellis_c(int16_t *const Out, |
dct_quantize_trellis_c(int16_t *const Out, |
790 |
int Q, |
int Q, |
791 |
const uint16_t * const Zigzag, |
const uint16_t * const Zigzag, |
792 |
const uint16_t * const QuantMatrix, |
const uint16_t * const QuantMatrix, |
793 |
int Non_Zero) |
int Non_Zero, |
794 |
|
int Sum) |
795 |
{ |
{ |
796 |
|
|
797 |
/* |
/* Note: We should search last non-zero coeffs on *real* DCT input coeffs |
798 |
* Note: We should search last non-zero coeffs on *real* DCT input coeffs (In[]), |
* (In[]), not quantized one (Out[]). However, it only improves the result |
799 |
* not quantized one (Out[]). However, it only improves the result *very* |
* *very* slightly (~0.01dB), whereas speed drops to crawling level :) |
800 |
* slightly (~0.01dB), whereas speed drops to crawling level :) |
* Well, actually, taking 1 more coeff past Non_Zero into account sometimes |
801 |
* Well, actually, taking 1 more coeff past Non_Zero into account sometimes helps. |
* helps. */ |
|
*/ |
|
802 |
typedef struct { int16_t Run, Level; } NODE; |
typedef struct { int16_t Run, Level; } NODE; |
803 |
|
|
804 |
NODE Nodes[65], Last; |
NODE Nodes[65], Last; |
805 |
uint32_t Run_Costs0[64+1]; |
uint32_t Run_Costs0[64+1]; |
806 |
uint32_t * const Run_Costs = Run_Costs0 + 1; |
uint32_t * const Run_Costs = Run_Costs0 + 1; |
807 |
|
|
808 |
const int Lambda = Trellis_Lambda_Tabs[Q-1]; /* it's 1/lambda, actually */ |
/* it's 1/lambda, actually */ |
809 |
|
const int Lambda = Trellis_Lambda_Tabs[Q-1]; |
810 |
|
|
811 |
int Run_Start = -1; |
int Run_Start = -1; |
812 |
uint32_t Min_Cost = 2<<TL_SHIFT; |
uint32_t Min_Cost = 2<<TL_SHIFT; |
814 |
int Last_Node = -1; |
int Last_Node = -1; |
815 |
uint32_t Last_Cost = 0; |
uint32_t Last_Cost = 0; |
816 |
|
|
817 |
int i, j, sum; |
int i, j; |
818 |
Run_Costs[-1] = 2<<TL_SHIFT; /* source (w/ CBP penalty) */ |
|
819 |
|
/* source (w/ CBP penalty) */ |
820 |
|
Run_Costs[-1] = 2<<TL_SHIFT; |
821 |
|
|
822 |
Non_Zero = Find_Last(Out, Zigzag, Non_Zero); |
Non_Zero = Find_Last(Out, Zigzag, Non_Zero); |
823 |
if (Non_Zero<0) |
if (Non_Zero<0) |
857 |
const uint32_t Cost = Cost_Base + (Code_Len20[Run-1]<<TL_SHIFT); |
const uint32_t Cost = Cost_Base + (Code_Len20[Run-1]<<TL_SHIFT); |
858 |
const uint32_t lCost = Cost_Base + (Code_Len24[Run-1]<<TL_SHIFT); |
const uint32_t lCost = Cost_Base + (Code_Len24[Run-1]<<TL_SHIFT); |
859 |
|
|
860 |
/* |
/* TODO: what about tie-breaks? Should we favor short runs or |
|
* TODO: what about tie-breaks? Should we favor short runs or |
|
861 |
* long runs? Although the error is the same, it would not be |
* long runs? Although the error is the same, it would not be |
862 |
* spread the same way along high and low frequencies... |
* spread the same way along high and low frequencies... */ |
|
*/ |
|
863 |
|
|
864 |
/* (I'd say: favour short runs => hifreq errors (HVS) -- gruel ) */ |
/* Gruel: I'd say, favour short runs => hifreq errors (HVS) */ |
865 |
|
|
866 |
if (Cost<Best_Cost) { |
if (Cost<Best_Cost) { |
867 |
Best_Cost = Cost; |
Best_Cost = Cost; |
876 |
} |
} |
877 |
if (Last_Node==i) |
if (Last_Node==i) |
878 |
Last.Level = Nodes[i].Level; |
Last.Level = Nodes[i].Level; |
879 |
} else { /* "big" levels */ |
} else if (51U>(uint32_t)(Level1+25)) { |
880 |
|
/* "big" levels (not less than ESC3, though) */ |
881 |
const uint8_t *Tbl_L1, *Tbl_L2, *Tbl_L1_Last, *Tbl_L2_Last; |
const uint8_t *Tbl_L1, *Tbl_L2, *Tbl_L1_Last, *Tbl_L2_Last; |
882 |
int Level2; |
int Level2; |
883 |
int dQ1, dQ2; |
int dQ1, dQ2; |
913 |
uint32_t Cost1, Cost2; |
uint32_t Cost1, Cost2; |
914 |
int bLevel; |
int bLevel; |
915 |
|
|
916 |
/* |
/* for sub-optimal (but slightly worth it, speed-wise) search, |
917 |
* for sub-optimal (but slightly worth it, speed-wise) search, uncomment the following: |
* uncomment the following: |
918 |
* if (Cost_Base>=Best_Cost) continue; |
* if (Cost_Base>=Best_Cost) continue; |
919 |
* (? doesn't seem to have any effect -- gruel ) |
* (? doesn't seem to have any effect -- gruel ) */ |
|
*/ |
|
920 |
|
|
921 |
Cost1 = Cost_Base + (Tbl_L1[Run-1]<<TL_SHIFT); |
Cost1 = Cost_Base + (Tbl_L1[Run-1]<<TL_SHIFT); |
922 |
Cost2 = Cost_Base + (Tbl_L2[Run-1]<<TL_SHIFT) + dDist21; |
Cost2 = Cost_Base + (Tbl_L2[Run-1]<<TL_SHIFT) + dDist21; |
951 |
Last_Node = i; |
Last_Node = i; |
952 |
} |
} |
953 |
} /* end of "for Run" */ |
} /* end of "for Run" */ |
954 |
|
} else { |
955 |
|
/* Very very high levels, with no chance of being optimizable |
956 |
|
* => Simply pick best Run. */ |
957 |
|
int Run; |
958 |
|
for(Run=i-Run_Start; Run>0; --Run) { |
959 |
|
/* 30 bits + no distortion */ |
960 |
|
const uint32_t Cost = (30<<TL_SHIFT) + Run_Costs[i-Run]; |
961 |
|
if (Cost<Best_Cost) { |
962 |
|
Best_Cost = Cost; |
963 |
|
Nodes[i].Run = Run; |
964 |
|
Nodes[i].Level = Level1; |
965 |
|
} |
966 |
|
|
967 |
|
if (Cost<Last_Cost) { |
968 |
|
Last_Cost = Cost; |
969 |
|
Last.Run = Run; |
970 |
|
Last.Level = Level1; |
971 |
|
Last_Node = i; |
972 |
|
} |
973 |
} |
} |
974 |
|
} |
975 |
|
|
976 |
|
|
977 |
Run_Costs[i] = Best_Cost; |
Run_Costs[i] = Best_Cost; |
978 |
|
|
980 |
Min_Cost = Best_Cost; |
Min_Cost = Best_Cost; |
981 |
Run_Start = i; |
Run_Start = i; |
982 |
} else { |
} else { |
983 |
/* |
/* as noticed by Michael Niedermayer (michaelni at gmx.at), |
984 |
* as noticed by Michael Niedermayer (michaelni at gmx.at), there's |
* there's a code shorter by 1 bit for a larger run (!), same |
985 |
* a code shorter by 1 bit for a larger run (!), same level. We give |
* level. We give it a chance by not moving the left barrier too |
986 |
* it a chance by not moving the left barrier too much. |
* much. */ |
|
*/ |
|
|
|
|
987 |
while( Run_Costs[Run_Start]>Min_Cost+(1<<TL_SHIFT) ) |
while( Run_Costs[Run_Start]>Min_Cost+(1<<TL_SHIFT) ) |
988 |
Run_Start++; |
Run_Start++; |
989 |
|
|
990 |
/* spread on preceding coeffs the cost incurred by skipping this one */ |
/* spread on preceding coeffs the cost incurred by skipping this |
991 |
|
* one */ |
992 |
for(j=Run_Start; j<i; ++j) Run_Costs[j] += Dist0; |
for(j=Run_Start; j<i; ++j) Run_Costs[j] += Dist0; |
993 |
Min_Cost += Dist0; |
Min_Cost += Dist0; |
994 |
} |
} |
995 |
} |
} |
996 |
|
|
997 |
/* It seems trellis doesn't give good results... just compute the Out sum and |
/* It seems trellis doesn't give good results... just leave the block untouched |
998 |
* quit (even if we did not modify it, upperlayer relies on this data) */ |
* and return the original sum value */ |
999 |
if (Last_Node<0) |
if (Last_Node<0) |
1000 |
return Compute_Sum(Out, Non_Zero); |
return Sum; |
1001 |
|
|
1002 |
/* reconstruct optimal sequence backward with surviving paths */ |
/* reconstruct optimal sequence backward with surviving paths */ |
1003 |
memset(Out, 0x00, 64*sizeof(*Out)); |
memset(Out, 0x00, 64*sizeof(*Out)); |
1004 |
Out[Zigzag[Last_Node]] = Last.Level; |
Out[Zigzag[Last_Node]] = Last.Level; |
1005 |
i = Last_Node - Last.Run; |
i = Last_Node - Last.Run; |
1006 |
sum = 0; |
Sum = abs(Last.Level); |
1007 |
while(i>=0) { |
while(i>=0) { |
1008 |
Out[Zigzag[i]] = Nodes[i].Level; |
Out[Zigzag[i]] = Nodes[i].Level; |
1009 |
sum += abs(Nodes[i].Level); |
Sum += abs(Nodes[i].Level); |
1010 |
i -= Nodes[i].Run; |
i -= Nodes[i].Run; |
1011 |
} |
} |
1012 |
|
|
1013 |
return sum; |
return Sum; |
1014 |
} |
} |
1015 |
|
|
1016 |
/* original version including heavy debugging info */ |
/* original version including heavy debugging info */ |