1 |
/***************************************************************************** |
/****************************************************************************** |
2 |
* |
* * |
3 |
* XVID MPEG-4 VIDEO CODEC |
* This file is part of XviD, a free MPEG-4 video encoder/decoder * |
4 |
* - MB Transfert/Quantization functions - |
* * |
5 |
* |
* XviD is an implementation of a part of one or more MPEG-4 Video tools * |
6 |
* Copyright(C) 2001-2003 Peter Ross <pross@xvid.org> |
* as specified in ISO/IEC 14496-2 standard. Those intending to use this * |
7 |
* 2001-2003 Michael Militzer <isibaar@xvid.org> |
* software module in hardware or software products are advised that its * |
8 |
* 2003 Edouard Gomez <ed.gomez@free.fr> |
* use may infringe existing patents or copyrights, and any such use * |
9 |
* |
* would be at such party's own risk. The original developer of this * |
10 |
* This program is free software ; you can redistribute it and/or modify |
* software module and his/her company, and subsequent editors and their * |
11 |
* it under the terms of the GNU General Public License as published by |
* companies, will have no liability for use of this software or * |
12 |
* the Free Software Foundation ; either version 2 of the License, or |
* modifications or derivatives thereof. * |
13 |
* (at your option) any later version. |
* * |
14 |
* |
* XviD is free software; you can redistribute it and/or modify it * |
15 |
* This program is distributed in the hope that it will be useful, |
* under the terms of the GNU General Public License as published by * |
16 |
* but WITHOUT ANY WARRANTY ; without even the implied warranty of |
* the Free Software Foundation; either version 2 of the License, or * |
17 |
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
* (at your option) any later version. * |
18 |
* GNU General Public License for more details. |
* * |
19 |
* |
* XviD is distributed in the hope that it will be useful, but * |
20 |
* You should have received a copy of the GNU General Public License |
* WITHOUT ANY WARRANTY; without even the implied warranty of * |
21 |
* along with this program ; if not, write to the Free Software |
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * |
22 |
* Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA |
* GNU General Public License for more details. * |
23 |
* |
* * |
24 |
* $Id$ |
* You should have received a copy of the GNU General Public License * |
25 |
* |
* along with this program; if not, write to the Free Software * |
26 |
****************************************************************************/ |
* Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA * |
27 |
|
* * |
28 |
|
******************************************************************************/ |
29 |
|
|
30 |
|
/****************************************************************************** |
31 |
|
* * |
32 |
|
* mbtransquant.c * |
33 |
|
* * |
34 |
|
* Copyright (C) 2001 - Peter Ross <pross@cs.rmit.edu.au> * |
35 |
|
* Copyright (C) 2001 - Michael Militzer <isibaar@xvid.org> * |
36 |
|
* * |
37 |
|
* For more information visit the XviD homepage: http://www.xvid.org * |
38 |
|
* * |
39 |
|
******************************************************************************/ |
40 |
|
|
41 |
|
/****************************************************************************** |
42 |
|
* * |
43 |
|
* Revision history: * |
44 |
|
* * |
45 |
|
* 29.03.2002 interlacing speedup - used transfer strides instead of * |
46 |
|
* manual field-to-frame conversion * |
47 |
|
* 26.03.2002 interlacing support - moved transfers outside loops * |
48 |
|
* 22.12.2001 get_dc_scaler() moved to common.h * |
49 |
|
* 19.11.2001 introduced coefficient thresholding (Isibaar) * |
50 |
|
* 17.11.2001 initial version * |
51 |
|
* * |
52 |
|
******************************************************************************/ |
53 |
|
|
|
#include <stdio.h> |
|
|
#include <stdlib.h> |
|
54 |
#include <string.h> |
#include <string.h> |
55 |
|
|
56 |
#include "../portab.h" |
#include "../portab.h" |
59 |
#include "../global.h" |
#include "../global.h" |
60 |
#include "mem_transfer.h" |
#include "mem_transfer.h" |
61 |
#include "timer.h" |
#include "timer.h" |
|
#include "../bitstream/mbcoding.h" |
|
|
#include "../bitstream/zigzag.h" |
|
62 |
#include "../dct/fdct.h" |
#include "../dct/fdct.h" |
63 |
#include "../dct/idct.h" |
#include "../dct/idct.h" |
64 |
#include "../quant/quant.h" |
#include "../quant/quant_mpeg4.h" |
65 |
|
#include "../quant/quant_h263.h" |
66 |
#include "../encoder.h" |
#include "../encoder.h" |
67 |
|
|
68 |
#include "../image/reduced.h" |
#include "../image/reduced.h" |
|
#include "../quant/quant_matrix.h" |
|
69 |
|
|
70 |
MBFIELDTEST_PTR MBFieldTest; |
MBFIELDTEST_PTR MBFieldTest; |
71 |
|
|
72 |
/* |
#define TOOSMALL_LIMIT 1 /* skip blocks having a coefficient sum below this value */ |
|
* Skip blocks having a coefficient sum below this value. This value will be |
|
|
* corrected according to the MB quantizer to avoid artifacts for quant==1 |
|
|
*/ |
|
|
#define PVOP_TOOSMALL_LIMIT 1 |
|
|
#define BVOP_TOOSMALL_LIMIT 3 |
|
|
|
|
|
/***************************************************************************** |
|
|
* Local functions |
|
|
****************************************************************************/ |
|
|
|
|
|
/* permute block and return field dct choice */ |
|
|
static __inline uint32_t |
|
|
MBDecideFieldDCT(int16_t data[6 * 64]) |
|
|
{ |
|
|
uint32_t field = MBFieldTest(data); |
|
|
|
|
|
if (field) |
|
|
MBFrameToField(data); |
|
|
|
|
|
return field; |
|
|
} |
|
73 |
|
|
|
/* Performs Forward DCT on all blocks */ |
|
74 |
static __inline void |
static __inline void |
75 |
MBfDCT(const MBParam * const pParam, |
MBfDCT(int16_t data[6 * 64]) |
|
const FRAMEINFO * const frame, |
|
|
MACROBLOCK * const pMB, |
|
|
uint32_t x_pos, |
|
|
uint32_t y_pos, |
|
|
int16_t data[6 * 64]) |
|
76 |
{ |
{ |
|
/* Handles interlacing */ |
|
|
start_timer(); |
|
|
pMB->field_dct = 0; |
|
|
if ((frame->vol_flags & XVID_VOL_INTERLACING) && |
|
|
(x_pos>0) && (x_pos<pParam->mb_width-1) && |
|
|
(y_pos>0) && (y_pos<pParam->mb_height-1)) { |
|
|
pMB->field_dct = MBDecideFieldDCT(data); |
|
|
} |
|
|
stop_interlacing_timer(); |
|
|
|
|
|
/* Perform DCT */ |
|
77 |
start_timer(); |
start_timer(); |
78 |
fdct(&data[0 * 64]); |
fdct(&data[0 * 64]); |
79 |
fdct(&data[1 * 64]); |
fdct(&data[1 * 64]); |
84 |
stop_dct_timer(); |
stop_dct_timer(); |
85 |
} |
} |
86 |
|
|
|
/* Performs Inverse DCT on all blocks */ |
|
|
static __inline void |
|
|
MBiDCT(int16_t data[6 * 64], |
|
|
const uint8_t cbp) |
|
|
{ |
|
|
start_timer(); |
|
|
if(cbp & (1 << (5 - 0))) idct(&data[0 * 64]); |
|
|
if(cbp & (1 << (5 - 1))) idct(&data[1 * 64]); |
|
|
if(cbp & (1 << (5 - 2))) idct(&data[2 * 64]); |
|
|
if(cbp & (1 << (5 - 3))) idct(&data[3 * 64]); |
|
|
if(cbp & (1 << (5 - 4))) idct(&data[4 * 64]); |
|
|
if(cbp & (1 << (5 - 5))) idct(&data[5 * 64]); |
|
|
stop_idct_timer(); |
|
|
} |
|
87 |
|
|
88 |
/* Quantize all blocks -- Intra mode */ |
static __inline uint32_t |
89 |
static __inline void |
QuantizeInterBlock( int16_t qcoeff[64], |
90 |
MBQuantIntra(const MBParam * pParam, |
const int16_t data[64], |
91 |
const FRAMEINFO * const frame, |
const uint32_t iQuant, |
92 |
const MACROBLOCK * pMB, |
const uint32_t quant_type) |
|
int16_t qcoeff[6 * 64], |
|
|
int16_t data[6*64]) |
|
|
{ |
|
|
int mpeg; |
|
|
int scaler_lum, scaler_chr; |
|
|
|
|
|
quant_intraFuncPtr const quant[2] = |
|
93 |
{ |
{ |
94 |
quant_h263_intra, |
uint32_t sum; |
|
quant_mpeg_intra |
|
|
}; |
|
|
|
|
|
mpeg = !!(pParam->vol_flags & XVID_VOL_MPEGQUANT); |
|
|
scaler_lum = get_dc_scaler(pMB->quant, 1); |
|
|
scaler_chr = get_dc_scaler(pMB->quant, 0); |
|
95 |
|
|
|
/* Quantize the block */ |
|
96 |
start_timer(); |
start_timer(); |
97 |
quant[mpeg](&data[0 * 64], &qcoeff[0 * 64], pMB->quant, scaler_lum); |
if (quant_type == H263_QUANT) |
98 |
quant[mpeg](&data[1 * 64], &qcoeff[1 * 64], pMB->quant, scaler_lum); |
sum = quant_inter(qcoeff, data, iQuant); |
99 |
quant[mpeg](&data[2 * 64], &qcoeff[2 * 64], pMB->quant, scaler_lum); |
else |
100 |
quant[mpeg](&data[3 * 64], &qcoeff[3 * 64], pMB->quant, scaler_lum); |
sum = quant4_inter(qcoeff, data, iQuant); |
|
quant[mpeg](&data[4 * 64], &qcoeff[4 * 64], pMB->quant, scaler_chr); |
|
|
quant[mpeg](&data[5 * 64], &qcoeff[5 * 64], pMB->quant, scaler_chr); |
|
|
stop_quant_timer(); |
|
|
} |
|
|
|
|
|
/* DeQuantize all blocks -- Intra mode */ |
|
|
static __inline void |
|
|
MBDeQuantIntra(const MBParam * pParam, |
|
|
const int iQuant, |
|
|
int16_t qcoeff[6 * 64], |
|
|
int16_t data[6*64]) |
|
|
{ |
|
|
int mpeg; |
|
|
int scaler_lum, scaler_chr; |
|
|
|
|
|
quant_intraFuncPtr const dequant[2] = |
|
|
{ |
|
|
dequant_h263_intra, |
|
|
dequant_mpeg_intra |
|
|
}; |
|
|
|
|
|
mpeg = !!(pParam->vol_flags & XVID_VOL_MPEGQUANT); |
|
|
scaler_lum = get_dc_scaler(iQuant, 1); |
|
|
scaler_chr = get_dc_scaler(iQuant, 0); |
|
101 |
|
|
102 |
start_timer(); |
stop_quant_timer(); |
103 |
dequant[mpeg](&qcoeff[0 * 64], &data[0 * 64], iQuant, scaler_lum); |
return sum; |
|
dequant[mpeg](&qcoeff[1 * 64], &data[1 * 64], iQuant, scaler_lum); |
|
|
dequant[mpeg](&qcoeff[2 * 64], &data[2 * 64], iQuant, scaler_lum); |
|
|
dequant[mpeg](&qcoeff[3 * 64], &data[3 * 64], iQuant, scaler_lum); |
|
|
dequant[mpeg](&qcoeff[4 * 64], &data[4 * 64], iQuant, scaler_chr); |
|
|
dequant[mpeg](&qcoeff[5 * 64], &data[5 * 64], iQuant, scaler_chr); |
|
|
stop_iquant_timer(); |
|
104 |
} |
} |
105 |
|
|
106 |
static int |
void |
107 |
dct_quantize_trellis_c(int16_t *const Out, |
MBTransQuantIntra(const MBParam * const pParam, |
108 |
const int16_t *const In, |
FRAMEINFO * const frame, |
109 |
int Q, |
MACROBLOCK * const pMB, |
110 |
const uint16_t * const Zigzag, |
const uint32_t x_pos, |
111 |
const uint16_t * const QuantMatrix, |
const uint32_t y_pos, |
|
int Non_Zero); |
|
|
|
|
|
/* Quantize all blocks -- Inter mode */ |
|
|
static __inline uint8_t |
|
|
MBQuantInter(const MBParam * pParam, |
|
|
const FRAMEINFO * const frame, |
|
|
const MACROBLOCK * pMB, |
|
112 |
int16_t data[6 * 64], |
int16_t data[6 * 64], |
113 |
int16_t qcoeff[6 * 64], |
int16_t qcoeff[6 * 64]) |
|
int bvop, |
|
|
int limit) |
|
114 |
{ |
{ |
115 |
|
|
116 |
|
uint32_t stride = pParam->edged_width; |
117 |
|
const uint32_t stride2 = stride / 2; |
118 |
|
uint32_t next_block = stride * ((frame->global_flags & XVID_REDUCED)?16:8); |
119 |
int i; |
int i; |
120 |
uint8_t cbp = 0; |
const uint32_t iQuant = pMB->quant; |
121 |
int sum; |
uint8_t *pY_Cur, *pU_Cur, *pV_Cur; |
122 |
int code_block, mpeg; |
const IMAGE * const pCurrent = &frame->image; |
|
|
|
|
quant_interFuncPtr const quant[2] = |
|
|
{ |
|
|
quant_h263_inter, |
|
|
quant_mpeg_inter |
|
|
}; |
|
|
|
|
|
mpeg = !!(pParam->vol_flags & XVID_VOL_MPEGQUANT); |
|
|
|
|
|
for (i = 0; i < 6; i++) { |
|
123 |
|
|
|
/* Quantize the block */ |
|
124 |
start_timer(); |
start_timer(); |
125 |
|
if ((frame->global_flags & XVID_REDUCED)) |
|
sum = quant[mpeg](&qcoeff[i*64], &data[i*64], pMB->quant); |
|
|
|
|
|
if(sum && (frame->vop_flags & XVID_VOP_TRELLISQUANT)) { |
|
|
const static uint16_t h263matrix[] = |
|
126 |
{ |
{ |
127 |
16, 16, 16, 16, 16, 16, 16, 16, |
pY_Cur = pCurrent->y + (y_pos << 5) * stride + (x_pos << 5); |
128 |
16, 16, 16, 16, 16, 16, 16, 16, |
pU_Cur = pCurrent->u + (y_pos << 4) * stride2 + (x_pos << 4); |
129 |
16, 16, 16, 16, 16, 16, 16, 16, |
pV_Cur = pCurrent->v + (y_pos << 4) * stride2 + (x_pos << 4); |
130 |
16, 16, 16, 16, 16, 16, 16, 16, |
|
131 |
16, 16, 16, 16, 16, 16, 16, 16, |
filter_18x18_to_8x8(&data[0 * 64], pY_Cur, stride); |
132 |
16, 16, 16, 16, 16, 16, 16, 16, |
filter_18x18_to_8x8(&data[1 * 64], pY_Cur + 16, stride); |
133 |
16, 16, 16, 16, 16, 16, 16, 16, |
filter_18x18_to_8x8(&data[2 * 64], pY_Cur + next_block, stride); |
134 |
16, 16, 16, 16, 16, 16, 16, 16 |
filter_18x18_to_8x8(&data[3 * 64], pY_Cur + next_block + 16, stride); |
135 |
}; |
filter_18x18_to_8x8(&data[4 * 64], pU_Cur, stride2); |
136 |
sum = dct_quantize_trellis_c(&qcoeff[i*64], &data[i*64], |
filter_18x18_to_8x8(&data[5 * 64], pV_Cur, stride2); |
|
pMB->quant, &scan_tables[0][0], |
|
|
(mpeg)?(uint16_t*)get_inter_matrix():h263matrix, |
|
|
63); |
|
|
} |
|
|
stop_quant_timer(); |
|
|
|
|
|
/* |
|
|
* We code the block if the sum is higher than the limit and if the first |
|
|
* two AC coefficients in zig zag order are not zero. |
|
|
*/ |
|
|
code_block = 0; |
|
|
if ((sum >= limit) || (qcoeff[i*64+1] != 0) || (qcoeff[i*64+8] != 0)) { |
|
|
code_block = 1; |
|
|
} else { |
|
|
|
|
|
if (bvop && (pMB->mode == MODE_DIRECT || pMB->mode == MODE_DIRECT_NO4V)) { |
|
|
/* dark blocks prevention for direct mode */ |
|
|
if ((qcoeff[i*64] < -1) || (qcoeff[i*64] > 0)) |
|
|
code_block = 1; |
|
137 |
} else { |
} else { |
138 |
/* not direct mode */ |
pY_Cur = pCurrent->y + (y_pos << 4) * stride + (x_pos << 4); |
139 |
if (qcoeff[i*64] != 0) |
pU_Cur = pCurrent->u + (y_pos << 3) * stride2 + (x_pos << 3); |
140 |
code_block = 1; |
pV_Cur = pCurrent->v + (y_pos << 3) * stride2 + (x_pos << 3); |
141 |
} |
|
142 |
} |
transfer_8to16copy(&data[0 * 64], pY_Cur, stride); |
143 |
|
transfer_8to16copy(&data[1 * 64], pY_Cur + 8, stride); |
144 |
/* Set the corresponding cbp bit */ |
transfer_8to16copy(&data[2 * 64], pY_Cur + next_block, stride); |
145 |
cbp |= code_block << (5 - i); |
transfer_8to16copy(&data[3 * 64], pY_Cur + next_block + 8, stride); |
146 |
|
transfer_8to16copy(&data[4 * 64], pU_Cur, stride2); |
147 |
|
transfer_8to16copy(&data[5 * 64], pV_Cur, stride2); |
148 |
} |
} |
149 |
|
stop_transfer_timer(); |
150 |
|
|
151 |
return(cbp); |
/* XXX: rrv+interlacing is buggy */ |
152 |
|
start_timer(); |
153 |
|
pMB->field_dct = 0; |
154 |
|
if ((frame->global_flags & XVID_INTERLACING) && |
155 |
|
(x_pos>0) && (x_pos<pParam->mb_width-1) && |
156 |
|
(y_pos>0) && (y_pos<pParam->mb_height-1)) { |
157 |
|
pMB->field_dct = MBDecideFieldDCT(data); |
158 |
} |
} |
159 |
|
stop_interlacing_timer(); |
160 |
|
|
161 |
/* DeQuantize all blocks -- Inter mode */ |
MBfDCT(data); |
|
static __inline void |
|
|
MBDeQuantInter(const MBParam * pParam, |
|
|
const int iQuant, |
|
|
int16_t data[6 * 64], |
|
|
int16_t qcoeff[6 * 64], |
|
|
const uint8_t cbp) |
|
|
{ |
|
|
int mpeg; |
|
|
|
|
|
quant_interFuncPtr const dequant[2] = |
|
|
{ |
|
|
dequant_h263_inter, |
|
|
dequant_mpeg_inter |
|
|
}; |
|
162 |
|
|
163 |
mpeg = !!(pParam->vol_flags & XVID_VOL_MPEGQUANT); |
for (i = 0; i < 6; i++) { |
164 |
|
const uint32_t iDcScaler = get_dc_scaler(iQuant, i < 4); |
165 |
|
|
166 |
start_timer(); |
start_timer(); |
167 |
if(cbp & (1 << (5 - 0))) dequant[mpeg](&data[0 * 64], &qcoeff[0 * 64], iQuant); |
if (pParam->m_quant_type == H263_QUANT) |
168 |
if(cbp & (1 << (5 - 1))) dequant[mpeg](&data[1 * 64], &qcoeff[1 * 64], iQuant); |
quant_intra(&qcoeff[i * 64], &data[i * 64], iQuant, iDcScaler); |
169 |
if(cbp & (1 << (5 - 2))) dequant[mpeg](&data[2 * 64], &qcoeff[2 * 64], iQuant); |
else |
170 |
if(cbp & (1 << (5 - 3))) dequant[mpeg](&data[3 * 64], &qcoeff[3 * 64], iQuant); |
quant4_intra(&qcoeff[i * 64], &data[i * 64], iQuant, iDcScaler); |
171 |
if(cbp & (1 << (5 - 4))) dequant[mpeg](&data[4 * 64], &qcoeff[4 * 64], iQuant); |
stop_quant_timer(); |
|
if(cbp & (1 << (5 - 5))) dequant[mpeg](&data[5 * 64], &qcoeff[5 * 64], iQuant); |
|
|
stop_iquant_timer(); |
|
|
} |
|
|
|
|
|
typedef void (transfer_operation_8to16_t) (int16_t *Dst, const uint8_t *Src, int BpS); |
|
|
typedef void (transfer_operation_16to8_t) (uint8_t *Dst, const int16_t *Src, int BpS); |
|
|
|
|
172 |
|
|
173 |
static __inline void |
/* speedup: dont decode when encoding only ivops */ |
174 |
MBTrans8to16(const MBParam * const pParam, |
if (pParam->iMaxKeyInterval != 1 || pParam->max_bframes > 0) |
|
const FRAMEINFO * const frame, |
|
|
const MACROBLOCK * const pMB, |
|
|
const uint32_t x_pos, |
|
|
const uint32_t y_pos, |
|
|
int16_t data[6 * 64]) |
|
175 |
{ |
{ |
176 |
uint32_t stride = pParam->edged_width; |
start_timer(); |
177 |
uint32_t stride2 = stride / 2; |
if (pParam->m_quant_type == H263_QUANT) |
178 |
uint32_t next_block = stride * 8; |
dequant_intra(&data[i * 64], &qcoeff[i * 64], iQuant, iDcScaler); |
179 |
int32_t cst; |
else |
180 |
int vop_reduced; |
dequant4_intra(&data[i * 64], &qcoeff[i * 64], iQuant, iDcScaler); |
181 |
uint8_t *pY_Cur, *pU_Cur, *pV_Cur; |
stop_iquant_timer(); |
|
const IMAGE * const pCurrent = &frame->image; |
|
|
transfer_operation_8to16_t * const functions[2] = |
|
|
{ |
|
|
(transfer_operation_8to16_t *)transfer_8to16copy, |
|
|
(transfer_operation_8to16_t *)filter_18x18_to_8x8 |
|
|
}; |
|
|
transfer_operation_8to16_t *transfer_op = NULL; |
|
|
|
|
|
vop_reduced = !!(frame->vop_flags & XVID_VOP_REDUCED); |
|
|
|
|
|
/* Image pointers */ |
|
|
pY_Cur = pCurrent->y + (y_pos << (4+vop_reduced)) * stride + (x_pos << (4+vop_reduced)); |
|
|
pU_Cur = pCurrent->u + (y_pos << (3+vop_reduced)) * stride2 + (x_pos << (3+vop_reduced)); |
|
|
pV_Cur = pCurrent->v + (y_pos << (3+vop_reduced)) * stride2 + (x_pos << (3+vop_reduced)); |
|
|
|
|
|
/* Block size */ |
|
|
cst = 8<<vop_reduced; |
|
|
|
|
|
/* Operation function */ |
|
|
transfer_op = functions[vop_reduced]; |
|
182 |
|
|
|
/* Do the transfer */ |
|
183 |
start_timer(); |
start_timer(); |
184 |
transfer_op(&data[0 * 64], pY_Cur, stride); |
idct(&data[i * 64]); |
185 |
transfer_op(&data[1 * 64], pY_Cur + cst, stride); |
stop_idct_timer(); |
186 |
transfer_op(&data[2 * 64], pY_Cur + next_block, stride); |
} |
|
transfer_op(&data[3 * 64], pY_Cur + next_block + cst, stride); |
|
|
transfer_op(&data[4 * 64], pU_Cur, stride2); |
|
|
transfer_op(&data[5 * 64], pV_Cur, stride2); |
|
|
stop_transfer_timer(); |
|
187 |
} |
} |
188 |
|
|
189 |
static __inline void |
/* speedup: dont decode when encoding only ivops */ |
190 |
MBTrans16to8(const MBParam * const pParam, |
if (pParam->iMaxKeyInterval != 1 || pParam->max_bframes > 0) |
|
const FRAMEINFO * const frame, |
|
|
const MACROBLOCK * const pMB, |
|
|
const uint32_t x_pos, |
|
|
const uint32_t y_pos, |
|
|
int16_t data[6 * 64], |
|
|
const uint32_t add, /* Must be 1 or 0 */ |
|
|
const uint8_t cbp) |
|
|
{ |
|
|
uint8_t *pY_Cur, *pU_Cur, *pV_Cur; |
|
|
uint32_t stride = pParam->edged_width; |
|
|
uint32_t stride2 = stride / 2; |
|
|
uint32_t next_block = stride * 8; |
|
|
uint32_t cst; |
|
|
int vop_reduced; |
|
|
const IMAGE * const pCurrent = &frame->image; |
|
|
|
|
|
/* Array of function pointers, indexed by [vop_reduced<<1+add] */ |
|
|
transfer_operation_16to8_t * const functions[4] = |
|
191 |
{ |
{ |
|
(transfer_operation_16to8_t*)transfer_16to8copy, |
|
|
(transfer_operation_16to8_t*)transfer_16to8add, |
|
|
(transfer_operation_16to8_t*)copy_upsampled_8x8_16to8, |
|
|
(transfer_operation_16to8_t*)add_upsampled_8x8_16to8 |
|
|
}; |
|
|
|
|
|
transfer_operation_16to8_t *transfer_op = NULL; |
|
192 |
|
|
193 |
if (pMB->field_dct) { |
if (pMB->field_dct) { |
194 |
next_block = stride; |
next_block = stride; |
195 |
stride *= 2; |
stride *= 2; |
196 |
} |
} |
197 |
|
|
|
/* Makes this vars booleans */ |
|
|
vop_reduced = !!(frame->vop_flags & XVID_VOP_REDUCED); |
|
|
|
|
|
/* Image pointers */ |
|
|
pY_Cur = pCurrent->y + (y_pos << (4+vop_reduced)) * stride + (x_pos << (4+vop_reduced)); |
|
|
pU_Cur = pCurrent->u + (y_pos << (3+vop_reduced)) * stride2 + (x_pos << (3+vop_reduced)); |
|
|
pV_Cur = pCurrent->v + (y_pos << (3+vop_reduced)) * stride2 + (x_pos << (3+vop_reduced)); |
|
|
|
|
|
/* Block size */ |
|
|
cst = 8<<vop_reduced; |
|
|
|
|
|
/* Operation function */ |
|
|
transfer_op = functions[(vop_reduced<<1) + add]; |
|
|
|
|
|
/* Do the operation */ |
|
198 |
start_timer(); |
start_timer(); |
199 |
if (cbp&32) transfer_op(pY_Cur, &data[0 * 64], stride); |
if ((frame->global_flags & XVID_REDUCED)) { |
200 |
if (cbp&16) transfer_op(pY_Cur + cst, &data[1 * 64], stride); |
copy_upsampled_8x8_16to8(pY_Cur, &data[0 * 64], stride); |
201 |
if (cbp& 8) transfer_op(pY_Cur + next_block, &data[2 * 64], stride); |
copy_upsampled_8x8_16to8(pY_Cur + 16, &data[1 * 64], stride); |
202 |
if (cbp& 4) transfer_op(pY_Cur + next_block + cst, &data[3 * 64], stride); |
copy_upsampled_8x8_16to8(pY_Cur + next_block, &data[2 * 64], stride); |
203 |
if (cbp& 2) transfer_op(pU_Cur, &data[4 * 64], stride2); |
copy_upsampled_8x8_16to8(pY_Cur + next_block + 16, &data[3 * 64], stride); |
204 |
if (cbp& 1) transfer_op(pV_Cur, &data[5 * 64], stride2); |
copy_upsampled_8x8_16to8(pU_Cur, &data[4 * 64], stride2); |
205 |
|
copy_upsampled_8x8_16to8(pV_Cur, &data[5 * 64], stride2); |
206 |
|
} else { |
207 |
|
transfer_16to8copy(pY_Cur, &data[0 * 64], stride); |
208 |
|
transfer_16to8copy(pY_Cur + 8, &data[1 * 64], stride); |
209 |
|
transfer_16to8copy(pY_Cur + next_block, &data[2 * 64], stride); |
210 |
|
transfer_16to8copy(pY_Cur + next_block + 8, &data[3 * 64], stride); |
211 |
|
transfer_16to8copy(pU_Cur, &data[4 * 64], stride2); |
212 |
|
transfer_16to8copy(pV_Cur, &data[5 * 64], stride2); |
213 |
|
} |
214 |
stop_transfer_timer(); |
stop_transfer_timer(); |
215 |
} |
} |
216 |
|
|
217 |
/***************************************************************************** |
} |
|
* Module functions |
|
|
****************************************************************************/ |
|
218 |
|
|
219 |
void |
uint8_t |
220 |
MBTransQuantIntra(const MBParam * const pParam, |
MBTransQuantInter(const MBParam * const pParam, |
221 |
const FRAMEINFO * const frame, |
FRAMEINFO * const frame, |
222 |
MACROBLOCK * const pMB, |
MACROBLOCK * const pMB, |
223 |
const uint32_t x_pos, |
const uint32_t x_pos, |
224 |
const uint32_t y_pos, |
const uint32_t y_pos, |
225 |
int16_t data[6 * 64], |
int16_t data[6 * 64], |
226 |
int16_t qcoeff[6 * 64]) |
int16_t qcoeff[6 * 64]) |
227 |
{ |
{ |
228 |
|
uint32_t stride = pParam->edged_width; |
229 |
|
const uint32_t stride2 = stride / 2; |
230 |
|
uint32_t next_block = stride * ((frame->global_flags & XVID_REDUCED)?16:8); |
231 |
|
int i; |
232 |
|
const uint32_t iQuant = pMB->quant; |
233 |
|
uint8_t *pY_Cur, *pU_Cur, *pV_Cur; |
234 |
|
int cbp = 0; |
235 |
|
uint32_t sum; |
236 |
|
const IMAGE * const pCurrent = &frame->image; |
237 |
|
|
238 |
/* Transfer data */ |
if ((frame->global_flags & XVID_REDUCED)) { |
239 |
MBTrans8to16(pParam, frame, pMB, x_pos, y_pos, data); |
pY_Cur = pCurrent->y + (y_pos << 5) * stride + (x_pos << 5); |
240 |
|
pU_Cur = pCurrent->u + (y_pos << 4) * stride2 + (x_pos << 4); |
241 |
/* Perform DCT (and field decision) */ |
pV_Cur = pCurrent->v + (y_pos << 4) * stride2 + (x_pos << 4); |
242 |
MBfDCT(pParam, frame, pMB, x_pos, y_pos, data); |
} else { |
243 |
|
pY_Cur = pCurrent->y + (y_pos << 4) * stride + (x_pos << 4); |
244 |
/* Quantize the block */ |
pU_Cur = pCurrent->u + (y_pos << 3) * stride2 + (x_pos << 3); |
245 |
MBQuantIntra(pParam, frame, pMB, data, qcoeff); |
pV_Cur = pCurrent->v + (y_pos << 3) * stride2 + (x_pos << 3); |
|
|
|
|
/* DeQuantize the block */ |
|
|
MBDeQuantIntra(pParam, pMB->quant, data, qcoeff); |
|
|
|
|
|
/* Perform inverse DCT*/ |
|
|
MBiDCT(data, 0x3F); |
|
|
|
|
|
/* Transfer back the data -- Don't add data */ |
|
|
MBTrans16to8(pParam, frame, pMB, x_pos, y_pos, data, 0, 0x3F); |
|
246 |
} |
} |
247 |
|
|
248 |
|
start_timer(); |
249 |
|
pMB->field_dct = 0; |
250 |
|
if ((frame->global_flags & XVID_INTERLACING) && |
251 |
|
(x_pos>0) && (x_pos<pParam->mb_width-1) && |
252 |
|
(y_pos>0) && (y_pos<pParam->mb_height-1)) { |
253 |
|
pMB->field_dct = MBDecideFieldDCT(data); |
254 |
|
} |
255 |
|
stop_interlacing_timer(); |
256 |
|
|
257 |
uint8_t |
MBfDCT(data); |
|
MBTransQuantInter(const MBParam * const pParam, |
|
|
const FRAMEINFO * const frame, |
|
|
MACROBLOCK * const pMB, |
|
|
const uint32_t x_pos, |
|
|
const uint32_t y_pos, |
|
|
int16_t data[6 * 64], |
|
|
int16_t qcoeff[6 * 64]) |
|
|
{ |
|
|
uint8_t cbp; |
|
|
uint32_t limit; |
|
258 |
|
|
259 |
/* There is no MBTrans8to16 for Inter block, that's done in motion compensation |
for (i = 0; i < 6; i++) { |
260 |
* already */ |
const uint32_t limit = TOOSMALL_LIMIT + ((iQuant == 1) ? 1 : 0); |
261 |
|
/* |
262 |
|
* no need to transfer 8->16-bit |
263 |
|
* (this is performed already in motion compensation) |
264 |
|
*/ |
265 |
|
|
266 |
/* Perform DCT (and field decision) */ |
sum = QuantizeInterBlock(&qcoeff[i * 64], &data[i * 64], iQuant, pParam->m_quant_type); |
|
MBfDCT(pParam, frame, pMB, x_pos, y_pos, data); |
|
267 |
|
|
268 |
/* Set the limit threshold */ |
if (sum >= limit) { |
|
limit = PVOP_TOOSMALL_LIMIT + ((pMB->quant == 1)? 1 : 0); |
|
269 |
|
|
270 |
if (frame->vop_flags & XVID_VOP_CARTOON) |
start_timer(); |
271 |
limit *= 3; |
if (pParam->m_quant_type == H263_QUANT) |
272 |
|
dequant_inter(&data[i * 64], &qcoeff[i * 64], iQuant); |
273 |
|
else |
274 |
|
dequant4_inter(&data[i * 64], &qcoeff[i * 64], iQuant); |
275 |
|
stop_iquant_timer(); |
276 |
|
|
277 |
/* Quantize the block */ |
cbp |= 1 << (5 - i); |
|
cbp = MBQuantInter(pParam, frame, pMB, data, qcoeff, 0, limit); |
|
278 |
|
|
279 |
/* DeQuantize the block */ |
start_timer(); |
280 |
MBDeQuantInter(pParam, pMB->quant, data, qcoeff, cbp); |
idct(&data[i * 64]); |
281 |
|
stop_idct_timer(); |
282 |
|
} |
283 |
|
} |
284 |
|
|
285 |
/* Perform inverse DCT*/ |
if (pMB->field_dct) { |
286 |
MBiDCT(data, cbp); |
next_block = stride; |
287 |
|
stride *= 2; |
288 |
|
} |
289 |
|
|
290 |
/* Transfer back the data -- Add the data */ |
start_timer(); |
291 |
MBTrans16to8(pParam, frame, pMB, x_pos, y_pos, data, 1, cbp); |
if ((frame->global_flags & XVID_REDUCED)) { |
292 |
|
if (cbp & 32) |
293 |
|
add_upsampled_8x8_16to8(pY_Cur, &data[0 * 64], stride); |
294 |
|
if (cbp & 16) |
295 |
|
add_upsampled_8x8_16to8(pY_Cur + 16, &data[1 * 64], stride); |
296 |
|
if (cbp & 8) |
297 |
|
add_upsampled_8x8_16to8(pY_Cur + next_block, &data[2 * 64], stride); |
298 |
|
if (cbp & 4) |
299 |
|
add_upsampled_8x8_16to8(pY_Cur + 16 + next_block, &data[3 * 64], stride); |
300 |
|
if (cbp & 2) |
301 |
|
add_upsampled_8x8_16to8(pU_Cur, &data[4 * 64], stride2); |
302 |
|
if (cbp & 1) |
303 |
|
add_upsampled_8x8_16to8(pV_Cur, &data[5 * 64], stride2); |
304 |
|
} else { |
305 |
|
if (cbp & 32) |
306 |
|
transfer_16to8add(pY_Cur, &data[0 * 64], stride); |
307 |
|
if (cbp & 16) |
308 |
|
transfer_16to8add(pY_Cur + 8, &data[1 * 64], stride); |
309 |
|
if (cbp & 8) |
310 |
|
transfer_16to8add(pY_Cur + next_block, &data[2 * 64], stride); |
311 |
|
if (cbp & 4) |
312 |
|
transfer_16to8add(pY_Cur + next_block + 8, &data[3 * 64], stride); |
313 |
|
if (cbp & 2) |
314 |
|
transfer_16to8add(pU_Cur, &data[4 * 64], stride2); |
315 |
|
if (cbp & 1) |
316 |
|
transfer_16to8add(pV_Cur, &data[5 * 64], stride2); |
317 |
|
} |
318 |
|
stop_transfer_timer(); |
319 |
|
|
320 |
return(cbp); |
return (uint8_t) cbp; |
321 |
} |
} |
322 |
|
|
323 |
uint8_t |
uint8_t |
324 |
MBTransQuantInterBVOP(const MBParam * pParam, |
MBTransQuantInterBVOP(const MBParam * pParam, |
325 |
FRAMEINFO * frame, |
FRAMEINFO * frame, |
326 |
MACROBLOCK * pMB, |
MACROBLOCK * pMB, |
|
const uint32_t x_pos, |
|
|
const uint32_t y_pos, |
|
327 |
int16_t data[6 * 64], |
int16_t data[6 * 64], |
328 |
int16_t qcoeff[6 * 64]) |
int16_t qcoeff[6 * 64]) |
329 |
{ |
{ |
330 |
uint8_t cbp; |
int cbp = 0; |
331 |
uint32_t limit; |
int i; |
|
|
|
|
/* There is no MBTrans8to16 for Inter block, that's done in motion compensation |
|
|
* already */ |
|
332 |
|
|
333 |
/* Perform DCT (and field decision) */ |
/* there is no MBTrans for Inter block, that's done in motion compensation already */ |
|
MBfDCT(pParam, frame, pMB, x_pos, y_pos, data); |
|
334 |
|
|
335 |
/* Set the limit threshold */ |
start_timer(); |
336 |
limit = BVOP_TOOSMALL_LIMIT; |
pMB->field_dct = 0; |
337 |
|
if ((frame->global_flags & XVID_INTERLACING)) { |
338 |
|
pMB->field_dct = MBDecideFieldDCT(data); |
339 |
|
} |
340 |
|
stop_interlacing_timer(); |
341 |
|
|
342 |
if (frame->vop_flags & XVID_VOP_CARTOON) |
MBfDCT(data); |
|
limit *= 2; |
|
343 |
|
|
344 |
/* Quantize the block */ |
for (i = 0; i < 6; i++) { |
345 |
cbp = MBQuantInter(pParam, frame, pMB, data, qcoeff, 1, limit); |
int codedecision = 0; |
346 |
|
|
347 |
/* |
int sum = QuantizeInterBlock(&qcoeff[i * 64], &data[i * 64], pMB->quant, pParam->m_quant_type); |
|
* History comment: |
|
|
* We don't have to DeQuant, iDCT and Transfer back data for B-frames. |
|
|
* |
|
|
* BUT some plugins require the rebuilt original frame to be passed so we |
|
|
* have to take care of that here |
|
|
*/ |
|
|
if((pParam->plugin_flags & XVID_REQORIGINAL)) { |
|
348 |
|
|
349 |
/* DeQuantize the block */ |
if ((sum > 2) || (qcoeff[i*64+1] != 0) || (qcoeff[i*64+8] != 0) ) codedecision = 1; |
350 |
MBDeQuantInter(pParam, pMB->quant, data, qcoeff, cbp); |
else { |
351 |
|
if (pMB->mode == MODE_DIRECT || pMB->mode == MODE_DIRECT_NO4V) { |
352 |
|
// dark blocks prevention for direct mode |
353 |
|
if ( (qcoeff[i*64] < -1) || (qcoeff[i*64] > 0) ) codedecision = 1; |
354 |
|
} else |
355 |
|
if (qcoeff[i*64] != 0) codedecision = 1; // not direct mode |
356 |
|
} |
357 |
|
|
358 |
/* Perform inverse DCT*/ |
if (codedecision) cbp |= 1 << (5 - i); |
359 |
MBiDCT(data, cbp); |
} |
360 |
|
|
361 |
/* Transfer back the data -- Add the data */ |
/* we don't have to DeQuant, iDCT and Transfer back data for B-frames if we don't reconstruct this frame */ |
362 |
MBTrans16to8(pParam, frame, pMB, x_pos, y_pos, data, 1, cbp); |
/* warning: reconstruction not supported yet */ |
363 |
|
return (uint8_t) cbp; |
364 |
} |
} |
365 |
|
|
366 |
return(cbp); |
/* permute block and return field dct choice */ |
367 |
|
|
368 |
|
static uint32_t |
369 |
|
MBDecideFieldDCT(int16_t data[6 * 64]) |
370 |
|
{ |
371 |
|
const uint32_t field = MBFieldTest(data); |
372 |
|
if (field) MBFrameToField(data); |
373 |
|
|
374 |
|
return field; |
375 |
} |
} |
376 |
|
|
377 |
/* if sum(diff between field lines) < sum(diff between frame lines), use field dct */ |
/* if sum(diff between field lines) < sum(diff between frame lines), use field dct */ |
378 |
|
|
379 |
uint32_t |
uint32_t |
380 |
MBFieldTest_c(int16_t data[6 * 64]) |
MBFieldTest_c(int16_t data[6 * 64]) |
381 |
{ |
{ |
389 |
for (i = 0; i < 7; ++i) { |
for (i = 0; i < 7; ++i) { |
390 |
for (j = 0; j < 8; ++j) { |
for (j = 0; j < 8; ++j) { |
391 |
frame += |
frame += |
392 |
abs(data[0 * 64 + (i + 1) * 8 + j] - data[0 * 64 + i * 8 + j]); |
ABS(data[0 * 64 + (i + 1) * 8 + j] - data[0 * 64 + i * 8 + j]); |
393 |
frame += |
frame += |
394 |
abs(data[1 * 64 + (i + 1) * 8 + j] - data[1 * 64 + i * 8 + j]); |
ABS(data[1 * 64 + (i + 1) * 8 + j] - data[1 * 64 + i * 8 + j]); |
395 |
frame += |
frame += |
396 |
abs(data[2 * 64 + (i + 1) * 8 + j] - data[2 * 64 + i * 8 + j]); |
ABS(data[2 * 64 + (i + 1) * 8 + j] - data[2 * 64 + i * 8 + j]); |
397 |
frame += |
frame += |
398 |
abs(data[3 * 64 + (i + 1) * 8 + j] - data[3 * 64 + i * 8 + j]); |
ABS(data[3 * 64 + (i + 1) * 8 + j] - data[3 * 64 + i * 8 + j]); |
399 |
|
|
400 |
field += |
field += |
401 |
abs(data[blocks[i + 1] + lines[i + 1] + j] - |
ABS(data[blocks[i + 1] + lines[i + 1] + j] - |
402 |
data[blocks[i] + lines[i] + j]); |
data[blocks[i] + lines[i] + j]); |
403 |
field += |
field += |
404 |
abs(data[blocks[i + 1] + lines[i + 1] + 8 + j] - |
ABS(data[blocks[i + 1] + lines[i + 1] + 8 + j] - |
405 |
data[blocks[i] + lines[i] + 8 + j]); |
data[blocks[i] + lines[i] + 8 + j]); |
406 |
field += |
field += |
407 |
abs(data[blocks[i + 1] + 64 + lines[i + 1] + j] - |
ABS(data[blocks[i + 1] + 64 + lines[i + 1] + j] - |
408 |
data[blocks[i] + 64 + lines[i] + j]); |
data[blocks[i] + 64 + lines[i] + j]); |
409 |
field += |
field += |
410 |
abs(data[blocks[i + 1] + 64 + lines[i + 1] + 8 + j] - |
ABS(data[blocks[i + 1] + 64 + lines[i + 1] + 8 + j] - |
411 |
data[blocks[i] + 64 + lines[i] + 8 + j]); |
data[blocks[i] + 64 + lines[i] + 8 + j]); |
412 |
} |
} |
413 |
} |
} |
428 |
|
|
429 |
/* left blocks */ |
/* left blocks */ |
430 |
|
|
431 |
/* 1=2, 2=4, 4=8, 8=1 */ |
// 1=2, 2=4, 4=8, 8=1 |
432 |
MOVLINE(tmp, LINE(0, 1)); |
MOVLINE(tmp, LINE(0, 1)); |
433 |
MOVLINE(LINE(0, 1), LINE(0, 2)); |
MOVLINE(LINE(0, 1), LINE(0, 2)); |
434 |
MOVLINE(LINE(0, 2), LINE(0, 4)); |
MOVLINE(LINE(0, 2), LINE(0, 4)); |
435 |
MOVLINE(LINE(0, 4), LINE(2, 0)); |
MOVLINE(LINE(0, 4), LINE(2, 0)); |
436 |
MOVLINE(LINE(2, 0), tmp); |
MOVLINE(LINE(2, 0), tmp); |
437 |
|
|
438 |
/* 3=6, 6=12, 12=9, 9=3 */ |
// 3=6, 6=12, 12=9, 9=3 |
439 |
MOVLINE(tmp, LINE(0, 3)); |
MOVLINE(tmp, LINE(0, 3)); |
440 |
MOVLINE(LINE(0, 3), LINE(0, 6)); |
MOVLINE(LINE(0, 3), LINE(0, 6)); |
441 |
MOVLINE(LINE(0, 6), LINE(2, 4)); |
MOVLINE(LINE(0, 6), LINE(2, 4)); |
442 |
MOVLINE(LINE(2, 4), LINE(2, 1)); |
MOVLINE(LINE(2, 4), LINE(2, 1)); |
443 |
MOVLINE(LINE(2, 1), tmp); |
MOVLINE(LINE(2, 1), tmp); |
444 |
|
|
445 |
/* 5=10, 10=5 */ |
// 5=10, 10=5 |
446 |
MOVLINE(tmp, LINE(0, 5)); |
MOVLINE(tmp, LINE(0, 5)); |
447 |
MOVLINE(LINE(0, 5), LINE(2, 2)); |
MOVLINE(LINE(0, 5), LINE(2, 2)); |
448 |
MOVLINE(LINE(2, 2), tmp); |
MOVLINE(LINE(2, 2), tmp); |
449 |
|
|
450 |
/* 7=14, 14=13, 13=11, 11=7 */ |
// 7=14, 14=13, 13=11, 11=7 |
451 |
MOVLINE(tmp, LINE(0, 7)); |
MOVLINE(tmp, LINE(0, 7)); |
452 |
MOVLINE(LINE(0, 7), LINE(2, 6)); |
MOVLINE(LINE(0, 7), LINE(2, 6)); |
453 |
MOVLINE(LINE(2, 6), LINE(2, 5)); |
MOVLINE(LINE(2, 6), LINE(2, 5)); |
456 |
|
|
457 |
/* right blocks */ |
/* right blocks */ |
458 |
|
|
459 |
/* 1=2, 2=4, 4=8, 8=1 */ |
// 1=2, 2=4, 4=8, 8=1 |
460 |
MOVLINE(tmp, LINE(1, 1)); |
MOVLINE(tmp, LINE(1, 1)); |
461 |
MOVLINE(LINE(1, 1), LINE(1, 2)); |
MOVLINE(LINE(1, 1), LINE(1, 2)); |
462 |
MOVLINE(LINE(1, 2), LINE(1, 4)); |
MOVLINE(LINE(1, 2), LINE(1, 4)); |
463 |
MOVLINE(LINE(1, 4), LINE(3, 0)); |
MOVLINE(LINE(1, 4), LINE(3, 0)); |
464 |
MOVLINE(LINE(3, 0), tmp); |
MOVLINE(LINE(3, 0), tmp); |
465 |
|
|
466 |
/* 3=6, 6=12, 12=9, 9=3 */ |
// 3=6, 6=12, 12=9, 9=3 |
467 |
MOVLINE(tmp, LINE(1, 3)); |
MOVLINE(tmp, LINE(1, 3)); |
468 |
MOVLINE(LINE(1, 3), LINE(1, 6)); |
MOVLINE(LINE(1, 3), LINE(1, 6)); |
469 |
MOVLINE(LINE(1, 6), LINE(3, 4)); |
MOVLINE(LINE(1, 6), LINE(3, 4)); |
470 |
MOVLINE(LINE(3, 4), LINE(3, 1)); |
MOVLINE(LINE(3, 4), LINE(3, 1)); |
471 |
MOVLINE(LINE(3, 1), tmp); |
MOVLINE(LINE(3, 1), tmp); |
472 |
|
|
473 |
/* 5=10, 10=5 */ |
// 5=10, 10=5 |
474 |
MOVLINE(tmp, LINE(1, 5)); |
MOVLINE(tmp, LINE(1, 5)); |
475 |
MOVLINE(LINE(1, 5), LINE(3, 2)); |
MOVLINE(LINE(1, 5), LINE(3, 2)); |
476 |
MOVLINE(LINE(3, 2), tmp); |
MOVLINE(LINE(3, 2), tmp); |
477 |
|
|
478 |
/* 7=14, 14=13, 13=11, 11=7 */ |
// 7=14, 14=13, 13=11, 11=7 |
479 |
MOVLINE(tmp, LINE(1, 7)); |
MOVLINE(tmp, LINE(1, 7)); |
480 |
MOVLINE(LINE(1, 7), LINE(3, 6)); |
MOVLINE(LINE(1, 7), LINE(3, 6)); |
481 |
MOVLINE(LINE(3, 6), LINE(3, 5)); |
MOVLINE(LINE(3, 6), LINE(3, 5)); |
482 |
MOVLINE(LINE(3, 5), LINE(3, 3)); |
MOVLINE(LINE(3, 5), LINE(3, 3)); |
483 |
MOVLINE(LINE(3, 3), tmp); |
MOVLINE(LINE(3, 3), tmp); |
484 |
} |
} |
|
|
|
|
/***************************************************************************** |
|
|
* Trellis based R-D optimal quantization |
|
|
* |
|
|
* Trellis Quant code (C) 2003 Pascal Massimino skal(at)planet-d.net |
|
|
* |
|
|
****************************************************************************/ |
|
|
|
|
|
/*---------------------------------------------------------------------------- |
|
|
* |
|
|
* Trellis-Based quantization |
|
|
* |
|
|
* So far I understand this paper: |
|
|
* |
|
|
* "Trellis-Based R-D Optimal Quantization in H.263+" |
|
|
* J.Wen, M.Luttrell, J.Villasenor |
|
|
* IEEE Transactions on Image Processing, Vol.9, No.8, Aug. 2000. |
|
|
* |
|
|
* we are at stake with a simplified Bellmand-Ford / Dijkstra Single |
|
|
* Source Shorted Path algo. But due to the underlying graph structure |
|
|
* ("Trellis"), it can be turned into a dynamic programming algo, |
|
|
* partially saving the explicit graph's nodes representation. And |
|
|
* without using a heap, since the open frontier of the DAG is always |
|
|
* known, and of fixed sized. |
|
|
*--------------------------------------------------------------------------*/ |
|
|
|
|
|
|
|
|
|
|
|
/* Codes lengths for relevant levels. */ |
|
|
|
|
|
/* let's factorize: */ |
|
|
static const uint8_t Code_Len0[64] = { |
|
|
30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30, |
|
|
30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30 }; |
|
|
static const uint8_t Code_Len1[64] = { |
|
|
20,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30, |
|
|
30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30 }; |
|
|
static const uint8_t Code_Len2[64] = { |
|
|
19,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30, |
|
|
30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30 }; |
|
|
static const uint8_t Code_Len3[64] = { |
|
|
18,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30, |
|
|
30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30 }; |
|
|
static const uint8_t Code_Len4[64] = { |
|
|
17,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30, |
|
|
30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30 }; |
|
|
static const uint8_t Code_Len5[64] = { |
|
|
16,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30, |
|
|
30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30 }; |
|
|
static const uint8_t Code_Len6[64] = { |
|
|
15,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30, |
|
|
30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30 }; |
|
|
static const uint8_t Code_Len7[64] = { |
|
|
13,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30, |
|
|
30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30 }; |
|
|
static const uint8_t Code_Len8[64] = { |
|
|
11,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30, |
|
|
30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30 }; |
|
|
static const uint8_t Code_Len9[64] = { |
|
|
12,21,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30, |
|
|
30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30 }; |
|
|
static const uint8_t Code_Len10[64] = { |
|
|
12,20,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30, |
|
|
30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30 }; |
|
|
static const uint8_t Code_Len11[64] = { |
|
|
12,19,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30, |
|
|
30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30 }; |
|
|
static const uint8_t Code_Len12[64] = { |
|
|
11,17,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30, |
|
|
30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30 }; |
|
|
static const uint8_t Code_Len13[64] = { |
|
|
11,15,21,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30, |
|
|
30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30 }; |
|
|
static const uint8_t Code_Len14[64] = { |
|
|
10,12,19,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30, |
|
|
30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30 }; |
|
|
static const uint8_t Code_Len15[64] = { |
|
|
10,13,17,19,21,21,21,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30, |
|
|
30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30 }; |
|
|
static const uint8_t Code_Len16[64] = { |
|
|
9,12,13,18,18,19,19,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30, |
|
|
30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30}; |
|
|
static const uint8_t Code_Len17[64] = { |
|
|
8,11,13,14,14,14,15,19,19,19,21,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30, |
|
|
30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30 }; |
|
|
static const uint8_t Code_Len18[64] = { |
|
|
7, 9,11,11,13,13,13,15,15,15,16,22,22,22,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30, |
|
|
30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30 }; |
|
|
static const uint8_t Code_Len19[64] = { |
|
|
5, 7, 9,10,10,11,11,11,11,11,13,14,16,17,17,18,18,18,18,18,18,18,18,20,20,21,21,30,30,30,30,30, |
|
|
30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30 }; |
|
|
static const uint8_t Code_Len20[64] = { |
|
|
3, 4, 5, 6, 6, 6, 7, 7, 7, 7, 8, 8, 8, 9, 9,10,10,10,10,10,10,10,10,12,12,13,13,12,13,14,15,15, |
|
|
15,16,16,16,16,17,17,17,18,18,19,19,19,19,19,19,19,19,21,21,22,22,30,30,30,30,30,30,30,30,30,30 }; |
|
|
|
|
|
/* a few more table for LAST table: */ |
|
|
static const uint8_t Code_Len21[64] = { |
|
|
13,20,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30, |
|
|
30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30}; |
|
|
static const uint8_t Code_Len22[64] = { |
|
|
12,15,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30, |
|
|
30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30}; |
|
|
static const uint8_t Code_Len23[64] = { |
|
|
10,12,15,15,15,16,16,16,16,17,17,17,17,17,17,17,17,18,18,18,18,18,18,18,18,19,19,19,19,20,20,20, |
|
|
20,21,21,21,21,21,21,21,21,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30,30}; |
|
|
static const uint8_t Code_Len24[64] = { |
|
|
5, 7, 7, 7, 7, 8, 8, 8, 8, 9, 9, 9, 9, 9, 9, 9, 9,10,10,10,10,10,10,10,10,11,11,11,11,12,12,12, |
|
|
12,13,13,13,13,13,13,13,13,14,16,16,16,16,17,17,17,17,18,18,18,18,18,18,18,18,19,19,19,19,19,19}; |
|
|
|
|
|
|
|
|
static const uint8_t * const B16_17_Code_Len[24] = { /* levels [1..24] */ |
|
|
Code_Len20,Code_Len19,Code_Len18,Code_Len17, |
|
|
Code_Len16,Code_Len15,Code_Len14,Code_Len13, |
|
|
Code_Len12,Code_Len11,Code_Len10,Code_Len9, |
|
|
Code_Len8, Code_Len7 ,Code_Len6 ,Code_Len5, |
|
|
Code_Len4, Code_Len3, Code_Len3 ,Code_Len2, |
|
|
Code_Len2, Code_Len1, Code_Len1, Code_Len1, |
|
|
}; |
|
|
|
|
|
static const uint8_t * const B16_17_Code_Len_Last[6] = { /* levels [1..6] */ |
|
|
Code_Len24,Code_Len23,Code_Len22,Code_Len21, Code_Len3, Code_Len1, |
|
|
}; |
|
|
|
|
|
/* TL_SHIFT controls the precision of the RD optimizations in trellis |
|
|
* valid range is [10..16]. The bigger, the more trellis is vulnerable |
|
|
* to overflows in cost formulas. |
|
|
* - 10 allows ac values up to 2^11 == 2048 |
|
|
* - 16 allows ac values up to 2^8 == 256 |
|
|
*/ |
|
|
#define TL_SHIFT 11 |
|
|
#define TL(q) ((0xfe00>>(16-TL_SHIFT))/(q*q)) |
|
|
|
|
|
static const int Trellis_Lambda_Tabs[31] = { |
|
|
TL( 1),TL( 2),TL( 3),TL( 4),TL( 5),TL( 6), TL( 7), |
|
|
TL( 8),TL( 9),TL(10),TL(11),TL(12),TL(13),TL(14), TL(15), |
|
|
TL(16),TL(17),TL(18),TL(19),TL(20),TL(21),TL(22), TL(23), |
|
|
TL(24),TL(25),TL(26),TL(27),TL(28),TL(29),TL(30), TL(31) |
|
|
}; |
|
|
#undef TL |
|
|
|
|
|
static int __inline |
|
|
Find_Last(const int16_t *C, const uint16_t *Zigzag, int i) |
|
|
{ |
|
|
while(i>=0) |
|
|
if (C[Zigzag[i]]) |
|
|
return i; |
|
|
else i--; |
|
|
return -1; |
|
|
} |
|
|
|
|
|
static int __inline |
|
|
Compute_Sum(const int16_t *C, int last) |
|
|
{ |
|
|
int sum = 0; |
|
|
|
|
|
while(last--) |
|
|
sum += abs(C[last]); |
|
|
|
|
|
return(sum); |
|
|
} |
|
|
|
|
|
/* this routine has been strippen of all debug code */ |
|
|
static int |
|
|
dct_quantize_trellis_c(int16_t *const Out, |
|
|
const int16_t *const In, |
|
|
int Q, |
|
|
const uint16_t * const Zigzag, |
|
|
const uint16_t * const QuantMatrix, |
|
|
int Non_Zero) |
|
|
{ |
|
|
|
|
|
/* |
|
|
* Note: We should search last non-zero coeffs on *real* DCT input coeffs (In[]), |
|
|
* not quantized one (Out[]). However, it only improves the result *very* |
|
|
* slightly (~0.01dB), whereas speed drops to crawling level :) |
|
|
* Well, actually, taking 1 more coeff past Non_Zero into account sometimes helps. |
|
|
*/ |
|
|
typedef struct { int16_t Run, Level; } NODE; |
|
|
|
|
|
NODE Nodes[65], Last; |
|
|
uint32_t Run_Costs0[64+1]; |
|
|
uint32_t * const Run_Costs = Run_Costs0 + 1; |
|
|
|
|
|
const int Lambda = Trellis_Lambda_Tabs[Q-1]; /* it's 1/lambda, actually */ |
|
|
|
|
|
int Run_Start = -1; |
|
|
uint32_t Min_Cost = 2<<TL_SHIFT; |
|
|
|
|
|
int Last_Node = -1; |
|
|
uint32_t Last_Cost = 0; |
|
|
|
|
|
int i, j, sum; |
|
|
Run_Costs[-1] = 2<<TL_SHIFT; /* source (w/ CBP penalty) */ |
|
|
|
|
|
Non_Zero = Find_Last(Out, Zigzag, Non_Zero); |
|
|
if (Non_Zero<0) |
|
|
return 0; /* Sum is zero if there are only zero coeffs */ |
|
|
|
|
|
for(i=0; i<=Non_Zero; i++) { |
|
|
const int q = ((Q*QuantMatrix[Zigzag[i]])>>4); |
|
|
const int Mult = 2*q; |
|
|
const int Bias = (q-1) | 1; |
|
|
const int Lev0 = Mult + Bias; |
|
|
|
|
|
const int AC = In[Zigzag[i]]; |
|
|
const int Level1 = Out[Zigzag[i]]; |
|
|
const unsigned int Dist0 = Lambda* AC*AC; |
|
|
uint32_t Best_Cost = 0xf0000000; |
|
|
Last_Cost += Dist0; |
|
|
|
|
|
/* very specialized loop for -1,0,+1 */ |
|
|
if ((uint32_t)(Level1+1)<3) { |
|
|
int dQ; |
|
|
int Run; |
|
|
uint32_t Cost0; |
|
|
|
|
|
if (AC<0) { |
|
|
Nodes[i].Level = -1; |
|
|
dQ = Lev0 + AC; |
|
|
} else { |
|
|
Nodes[i].Level = 1; |
|
|
dQ = Lev0 - AC; |
|
|
} |
|
|
Cost0 = Lambda*dQ*dQ; |
|
|
|
|
|
Nodes[i].Run = 1; |
|
|
Best_Cost = (Code_Len20[0]<<TL_SHIFT) + Run_Costs[i-1]+Cost0; |
|
|
for(Run=i-Run_Start; Run>0; --Run) { |
|
|
const uint32_t Cost_Base = Cost0 + Run_Costs[i-Run]; |
|
|
const uint32_t Cost = Cost_Base + (Code_Len20[Run-1]<<TL_SHIFT); |
|
|
const uint32_t lCost = Cost_Base + (Code_Len24[Run-1]<<TL_SHIFT); |
|
|
|
|
|
/* |
|
|
* TODO: what about tie-breaks? Should we favor short runs or |
|
|
* long runs? Although the error is the same, it would not be |
|
|
* spread the same way along high and low frequencies... |
|
|
*/ |
|
|
|
|
|
/* (I'd say: favour short runs => hifreq errors (HVS) -- gruel ) */ |
|
|
|
|
|
if (Cost<Best_Cost) { |
|
|
Best_Cost = Cost; |
|
|
Nodes[i].Run = Run; |
|
|
} |
|
|
|
|
|
if (lCost<Last_Cost) { |
|
|
Last_Cost = lCost; |
|
|
Last.Run = Run; |
|
|
Last_Node = i; |
|
|
} |
|
|
} |
|
|
if (Last_Node==i) |
|
|
Last.Level = Nodes[i].Level; |
|
|
} else { /* "big" levels */ |
|
|
const uint8_t *Tbl_L1, *Tbl_L2, *Tbl_L1_Last, *Tbl_L2_Last; |
|
|
int Level2; |
|
|
int dQ1, dQ2; |
|
|
int Run; |
|
|
uint32_t Dist1,Dist2; |
|
|
int dDist21; |
|
|
|
|
|
if (Level1>1) { |
|
|
dQ1 = Level1*Mult-AC + Bias; |
|
|
dQ2 = dQ1 - Mult; |
|
|
Level2 = Level1-1; |
|
|
Tbl_L1 = (Level1<=24) ? B16_17_Code_Len[Level1-1] : Code_Len0; |
|
|
Tbl_L2 = (Level2<=24) ? B16_17_Code_Len[Level2-1] : Code_Len0; |
|
|
Tbl_L1_Last = (Level1<=6) ? B16_17_Code_Len_Last[Level1-1] : Code_Len0; |
|
|
Tbl_L2_Last = (Level2<=6) ? B16_17_Code_Len_Last[Level2-1] : Code_Len0; |
|
|
} else { /* Level1<-1 */ |
|
|
dQ1 = Level1*Mult-AC - Bias; |
|
|
dQ2 = dQ1 + Mult; |
|
|
Level2 = Level1 + 1; |
|
|
Tbl_L1 = (Level1>=-24) ? B16_17_Code_Len[Level1^-1] : Code_Len0; |
|
|
Tbl_L2 = (Level2>=-24) ? B16_17_Code_Len[Level2^-1] : Code_Len0; |
|
|
Tbl_L1_Last = (Level1>=- 6) ? B16_17_Code_Len_Last[Level1^-1] : Code_Len0; |
|
|
Tbl_L2_Last = (Level2>=- 6) ? B16_17_Code_Len_Last[Level2^-1] : Code_Len0; |
|
|
} |
|
|
|
|
|
Dist1 = Lambda*dQ1*dQ1; |
|
|
Dist2 = Lambda*dQ2*dQ2; |
|
|
dDist21 = Dist2-Dist1; |
|
|
|
|
|
for(Run=i-Run_Start; Run>0; --Run) |
|
|
{ |
|
|
const uint32_t Cost_Base = Dist1 + Run_Costs[i-Run]; |
|
|
uint32_t Cost1, Cost2; |
|
|
int bLevel; |
|
|
|
|
|
/* |
|
|
* for sub-optimal (but slightly worth it, speed-wise) search, uncomment the following: |
|
|
* if (Cost_Base>=Best_Cost) continue; |
|
|
* (? doesn't seem to have any effect -- gruel ) |
|
|
*/ |
|
|
|
|
|
Cost1 = Cost_Base + (Tbl_L1[Run-1]<<TL_SHIFT); |
|
|
Cost2 = Cost_Base + (Tbl_L2[Run-1]<<TL_SHIFT) + dDist21; |
|
|
|
|
|
if (Cost2<Cost1) { |
|
|
Cost1 = Cost2; |
|
|
bLevel = Level2; |
|
|
} else { |
|
|
bLevel = Level1; |
|
|
} |
|
|
|
|
|
if (Cost1<Best_Cost) { |
|
|
Best_Cost = Cost1; |
|
|
Nodes[i].Run = Run; |
|
|
Nodes[i].Level = bLevel; |
|
|
} |
|
|
|
|
|
Cost1 = Cost_Base + (Tbl_L1_Last[Run-1]<<TL_SHIFT); |
|
|
Cost2 = Cost_Base + (Tbl_L2_Last[Run-1]<<TL_SHIFT) + dDist21; |
|
|
|
|
|
if (Cost2<Cost1) { |
|
|
Cost1 = Cost2; |
|
|
bLevel = Level2; |
|
|
} else { |
|
|
bLevel = Level1; |
|
|
} |
|
|
|
|
|
if (Cost1<Last_Cost) { |
|
|
Last_Cost = Cost1; |
|
|
Last.Run = Run; |
|
|
Last.Level = bLevel; |
|
|
Last_Node = i; |
|
|
} |
|
|
} /* end of "for Run" */ |
|
|
|
|
|
} |
|
|
|
|
|
Run_Costs[i] = Best_Cost; |
|
|
|
|
|
if (Best_Cost < Min_Cost + Dist0) { |
|
|
Min_Cost = Best_Cost; |
|
|
Run_Start = i; |
|
|
} else { |
|
|
/* |
|
|
* as noticed by Michael Niedermayer (michaelni at gmx.at), there's |
|
|
* a code shorter by 1 bit for a larger run (!), same level. We give |
|
|
* it a chance by not moving the left barrier too much. |
|
|
*/ |
|
|
|
|
|
while( Run_Costs[Run_Start]>Min_Cost+(1<<TL_SHIFT) ) |
|
|
Run_Start++; |
|
|
|
|
|
/* spread on preceding coeffs the cost incurred by skipping this one */ |
|
|
for(j=Run_Start; j<i; ++j) Run_Costs[j] += Dist0; |
|
|
Min_Cost += Dist0; |
|
|
} |
|
|
} |
|
|
|
|
|
/* It seems trellis doesn't give good results... just compute the Out sum and |
|
|
* quit (even if we did not modify it, upperlayer relies on this data) */ |
|
|
if (Last_Node<0) |
|
|
return Compute_Sum(Out, Non_Zero); |
|
|
|
|
|
/* reconstruct optimal sequence backward with surviving paths */ |
|
|
memset(Out, 0x00, 64*sizeof(*Out)); |
|
|
Out[Zigzag[Last_Node]] = Last.Level; |
|
|
i = Last_Node - Last.Run; |
|
|
sum = 0; |
|
|
while(i>=0) { |
|
|
Out[Zigzag[i]] = Nodes[i].Level; |
|
|
sum += abs(Nodes[i].Level); |
|
|
i -= Nodes[i].Run; |
|
|
} |
|
|
|
|
|
return sum; |
|
|
} |
|
|
|
|
|
/* original version including heavy debugging info */ |
|
|
|
|
|
#ifdef DBGTRELL |
|
|
|
|
|
#define DBG 0 |
|
|
|
|
|
static __inline uint32_t Evaluate_Cost(const int16_t *C, int Mult, int Bias, |
|
|
const uint16_t * Zigzag, int Max, int Lambda) |
|
|
{ |
|
|
#if (DBG>0) |
|
|
const int16_t * const Ref = C + 6*64; |
|
|
int Last = Max; |
|
|
int Bits = 0; |
|
|
int Dist = 0; |
|
|
int i; |
|
|
uint32_t Cost; |
|
|
|
|
|
while(Last>=0 && C[Zigzag[Last]]==0) |
|
|
Last--; |
|
|
|
|
|
if (Last>=0) { |
|
|
int j=0, j0=0; |
|
|
int Run, Level; |
|
|
|
|
|
Bits = 2; /* CBP */ |
|
|
while(j<Last) { |
|
|
while(!C[Zigzag[j]]) |
|
|
j++; |
|
|
if (j==Last) |
|
|
break; |
|
|
Level=C[Zigzag[j]]; |
|
|
Run = j - j0; |
|
|
j0 = ++j; |
|
|
if (Level>=-24 && Level<=24) |
|
|
Bits += B16_17_Code_Len[(Level<0) ? -Level-1 : Level-1][Run]; |
|
|
else |
|
|
Bits += 30; |
|
|
} |
|
|
Level = C[Zigzag[Last]]; |
|
|
Run = j - j0; |
|
|
if (Level>=-6 && Level<=6) |
|
|
Bits += B16_17_Code_Len_Last[(Level<0) ? -Level-1 : Level-1][Run]; |
|
|
else |
|
|
Bits += 30; |
|
|
} |
|
|
|
|
|
for(i=0; i<=Last; ++i) { |
|
|
int V = C[Zigzag[i]]*Mult; |
|
|
if (V>0) |
|
|
V += Bias; |
|
|
else |
|
|
if (V<0) |
|
|
V -= Bias; |
|
|
V -= Ref[Zigzag[i]]; |
|
|
Dist += V*V; |
|
|
} |
|
|
Cost = Lambda*Dist + (Bits<<TL_SHIFT); |
|
|
if (DBG==1) |
|
|
printf( " Last:%2d/%2d Cost = [(Bits=%5.0d) + Lambda*(Dist=%6.0d) = %d ] >>12= %d ", Last,Max, Bits, Dist, Cost, Cost>>12 ); |
|
|
return Cost; |
|
|
|
|
|
#else |
|
|
return 0; |
|
|
#endif |
|
|
} |
|
|
|
|
|
|
|
|
static int |
|
|
dct_quantize_trellis_h263_c(int16_t *const Out, const int16_t *const In, int Q, const uint16_t * const Zigzag, int Non_Zero) |
|
|
{ |
|
|
|
|
|
/* |
|
|
* Note: We should search last non-zero coeffs on *real* DCT input coeffs (In[]), |
|
|
* not quantized one (Out[]). However, it only improves the result *very* |
|
|
* slightly (~0.01dB), whereas speed drops to crawling level :) |
|
|
* Well, actually, taking 1 more coeff past Non_Zero into account sometimes helps. |
|
|
*/ |
|
|
typedef struct { int16_t Run, Level; } NODE; |
|
|
|
|
|
NODE Nodes[65], Last; |
|
|
uint32_t Run_Costs0[64+1]; |
|
|
uint32_t * const Run_Costs = Run_Costs0 + 1; |
|
|
const int Mult = 2*Q; |
|
|
const int Bias = (Q-1) | 1; |
|
|
const int Lev0 = Mult + Bias; |
|
|
const int Lambda = Trellis_Lambda_Tabs[Q-1]; /* it's 1/lambda, actually */ |
|
|
|
|
|
int Run_Start = -1; |
|
|
Run_Costs[-1] = 2<<TL_SHIFT; /* source (w/ CBP penalty) */ |
|
|
uint32_t Min_Cost = 2<<TL_SHIFT; |
|
|
|
|
|
int Last_Node = -1; |
|
|
uint32_t Last_Cost = 0; |
|
|
|
|
|
int i, j; |
|
|
|
|
|
#if (DBG>0) |
|
|
Last.Level = 0; Last.Run = -1; /* just initialize to smthg */ |
|
|
#endif |
|
|
|
|
|
Non_Zero = Find_Last(Out, Zigzag, Non_Zero); |
|
|
if (Non_Zero<0) |
|
|
return -1; |
|
|
|
|
|
for(i=0; i<=Non_Zero; i++) |
|
|
{ |
|
|
const int AC = In[Zigzag[i]]; |
|
|
const int Level1 = Out[Zigzag[i]]; |
|
|
const int Dist0 = Lambda* AC*AC; |
|
|
uint32_t Best_Cost = 0xf0000000; |
|
|
Last_Cost += Dist0; |
|
|
|
|
|
if ((uint32_t)(Level1+1)<3) /* very specialized loop for -1,0,+1 */ |
|
|
{ |
|
|
int dQ; |
|
|
int Run; |
|
|
uint32_t Cost0; |
|
|
|
|
|
if (AC<0) { |
|
|
Nodes[i].Level = -1; |
|
|
dQ = Lev0 + AC; |
|
|
} else { |
|
|
Nodes[i].Level = 1; |
|
|
dQ = Lev0 - AC; |
|
|
} |
|
|
Cost0 = Lambda*dQ*dQ; |
|
|
|
|
|
Nodes[i].Run = 1; |
|
|
Best_Cost = (Code_Len20[0]<<TL_SHIFT) + Run_Costs[i-1]+Cost0; |
|
|
for(Run=i-Run_Start; Run>0; --Run) |
|
|
{ |
|
|
const uint32_t Cost_Base = Cost0 + Run_Costs[i-Run]; |
|
|
const uint32_t Cost = Cost_Base + (Code_Len20[Run-1]<<TL_SHIFT); |
|
|
const uint32_t lCost = Cost_Base + (Code_Len24[Run-1]<<TL_SHIFT); |
|
|
|
|
|
/* |
|
|
* TODO: what about tie-breaks? Should we favor short runs or |
|
|
* long runs? Although the error is the same, it would not be |
|
|
* spread the same way along high and low frequencies... |
|
|
*/ |
|
|
if (Cost<Best_Cost) { |
|
|
Best_Cost = Cost; |
|
|
Nodes[i].Run = Run; |
|
|
} |
|
|
|
|
|
if (lCost<Last_Cost) { |
|
|
Last_Cost = lCost; |
|
|
Last.Run = Run; |
|
|
Last_Node = i; |
|
|
} |
|
|
} |
|
|
if (Last_Node==i) |
|
|
Last.Level = Nodes[i].Level; |
|
|
|
|
|
if (DBG==1) { |
|
|
Run_Costs[i] = Best_Cost; |
|
|
printf( "Costs #%2d: ", i); |
|
|
for(j=-1;j<=Non_Zero;++j) { |
|
|
if (j==Run_Start) printf( " %3.0d|", Run_Costs[j]>>12 ); |
|
|
else if (j>Run_Start && j<i) printf( " %3.0d|", Run_Costs[j]>>12 ); |
|
|
else if (j==i) printf( "(%3.0d)", Run_Costs[j]>>12 ); |
|
|
else printf( " - |" ); |
|
|
} |
|
|
printf( "<%3.0d %2d %d>", Min_Cost>>12, Nodes[i].Level, Nodes[i].Run ); |
|
|
printf( " Last:#%2d {%3.0d %2d %d}", Last_Node, Last_Cost>>12, Last.Level, Last.Run ); |
|
|
printf( " AC:%3.0d Dist0:%3d Dist(%d)=%d", AC, Dist0>>12, Nodes[i].Level, Cost0>>12 ); |
|
|
printf( "\n" ); |
|
|
} |
|
|
} |
|
|
else /* "big" levels */ |
|
|
{ |
|
|
const uint8_t *Tbl_L1, *Tbl_L2, *Tbl_L1_Last, *Tbl_L2_Last; |
|
|
int Level2; |
|
|
int dQ1, dQ2; |
|
|
int Run; |
|
|
uint32_t Dist1,Dist2; |
|
|
int dDist21; |
|
|
|
|
|
if (Level1>1) { |
|
|
dQ1 = Level1*Mult-AC + Bias; |
|
|
dQ2 = dQ1 - Mult; |
|
|
Level2 = Level1-1; |
|
|
Tbl_L1 = (Level1<=24) ? B16_17_Code_Len[Level1-1] : Code_Len0; |
|
|
Tbl_L2 = (Level2<=24) ? B16_17_Code_Len[Level2-1] : Code_Len0; |
|
|
Tbl_L1_Last = (Level1<=6) ? B16_17_Code_Len_Last[Level1-1] : Code_Len0; |
|
|
Tbl_L2_Last = (Level2<=6) ? B16_17_Code_Len_Last[Level2-1] : Code_Len0; |
|
|
} else { /* Level1<-1 */ |
|
|
dQ1 = Level1*Mult-AC - Bias; |
|
|
dQ2 = dQ1 + Mult; |
|
|
Level2 = Level1 + 1; |
|
|
Tbl_L1 = (Level1>=-24) ? B16_17_Code_Len[Level1^-1] : Code_Len0; |
|
|
Tbl_L2 = (Level2>=-24) ? B16_17_Code_Len[Level2^-1] : Code_Len0; |
|
|
Tbl_L1_Last = (Level1>=- 6) ? B16_17_Code_Len_Last[Level1^-1] : Code_Len0; |
|
|
Tbl_L2_Last = (Level2>=- 6) ? B16_17_Code_Len_Last[Level2^-1] : Code_Len0; |
|
|
} |
|
|
Dist1 = Lambda*dQ1*dQ1; |
|
|
Dist2 = Lambda*dQ2*dQ2; |
|
|
dDist21 = Dist2-Dist1; |
|
|
|
|
|
for(Run=i-Run_Start; Run>0; --Run) |
|
|
{ |
|
|
const uint32_t Cost_Base = Dist1 + Run_Costs[i-Run]; |
|
|
uint32_t Cost1, Cost2; |
|
|
int bLevel; |
|
|
|
|
|
/* |
|
|
* for sub-optimal (but slightly worth it, speed-wise) search, uncomment the following: |
|
|
* if (Cost_Base>=Best_Cost) continue; |
|
|
*/ |
|
|
Cost1 = Cost_Base + (Tbl_L1[Run-1]<<TL_SHIFT); |
|
|
Cost2 = Cost_Base + (Tbl_L2[Run-1]<<TL_SHIFT) + dDist21; |
|
|
|
|
|
if (Cost2<Cost1) { |
|
|
Cost1 = Cost2; |
|
|
bLevel = Level2; |
|
|
} else |
|
|
bLevel = Level1; |
|
|
|
|
|
if (Cost1<Best_Cost) { |
|
|
Best_Cost = Cost1; |
|
|
Nodes[i].Run = Run; |
|
|
Nodes[i].Level = bLevel; |
|
|
} |
|
|
|
|
|
Cost1 = Cost_Base + (Tbl_L1_Last[Run-1]<<TL_SHIFT); |
|
|
Cost2 = Cost_Base + (Tbl_L2_Last[Run-1]<<TL_SHIFT) + dDist21; |
|
|
|
|
|
if (Cost2<Cost1) { |
|
|
Cost1 = Cost2; |
|
|
bLevel = Level2; |
|
|
} else |
|
|
bLevel = Level1; |
|
|
|
|
|
if (Cost1<Last_Cost) { |
|
|
Last_Cost = Cost1; |
|
|
Last.Run = Run; |
|
|
Last.Level = bLevel; |
|
|
Last_Node = i; |
|
|
} |
|
|
} /* end of "for Run" */ |
|
|
|
|
|
if (DBG==1) { |
|
|
Run_Costs[i] = Best_Cost; |
|
|
printf( "Costs #%2d: ", i); |
|
|
for(j=-1;j<=Non_Zero;++j) { |
|
|
if (j==Run_Start) printf( " %3.0d|", Run_Costs[j]>>12 ); |
|
|
else if (j>Run_Start && j<i) printf( " %3.0d|", Run_Costs[j]>>12 ); |
|
|
else if (j==i) printf( "(%3.0d)", Run_Costs[j]>>12 ); |
|
|
else printf( " - |" ); |
|
|
} |
|
|
printf( "<%3.0d %2d %d>", Min_Cost>>12, Nodes[i].Level, Nodes[i].Run ); |
|
|
printf( " Last:#%2d {%3.0d %2d %d}", Last_Node, Last_Cost>>12, Last.Level, Last.Run ); |
|
|
printf( " AC:%3.0d Dist0:%3d Dist(%2d):%3d Dist(%2d):%3d", AC, Dist0>>12, Level1, Dist1>>12, Level2, Dist2>>12 ); |
|
|
printf( "\n" ); |
|
|
} |
|
|
} |
|
|
|
|
|
Run_Costs[i] = Best_Cost; |
|
|
|
|
|
if (Best_Cost < Min_Cost + Dist0) { |
|
|
Min_Cost = Best_Cost; |
|
|
Run_Start = i; |
|
|
} |
|
|
else |
|
|
{ |
|
|
/* |
|
|
* as noticed by Michael Niedermayer (michaelni at gmx.at), there's |
|
|
* a code shorter by 1 bit for a larger run (!), same level. We give |
|
|
* it a chance by not moving the left barrier too much. |
|
|
*/ |
|
|
|
|
|
while( Run_Costs[Run_Start]>Min_Cost+(1<<TL_SHIFT) ) |
|
|
Run_Start++; |
|
|
|
|
|
/* spread on preceding coeffs the cost incurred by skipping this one */ |
|
|
for(j=Run_Start; j<i; ++j) Run_Costs[j] += Dist0; |
|
|
Min_Cost += Dist0; |
|
|
} |
|
|
} |
|
|
|
|
|
if (DBG) { |
|
|
Last_Cost = Evaluate_Cost(Out,Mult,Bias, Zigzag,Non_Zero, Lambda); |
|
|
if (DBG==1) { |
|
|
printf( "=> " ); |
|
|
for(i=0; i<=Non_Zero; ++i) printf( "[%3.0d] ", Out[Zigzag[i]] ); |
|
|
printf( "\n" ); |
|
|
} |
|
|
} |
|
|
|
|
|
if (Last_Node<0) |
|
|
return -1; |
|
|
|
|
|
/* reconstruct optimal sequence backward with surviving paths */ |
|
|
memset(Out, 0x00, 64*sizeof(*Out)); |
|
|
Out[Zigzag[Last_Node]] = Last.Level; |
|
|
i = Last_Node - Last.Run; |
|
|
while(i>=0) { |
|
|
Out[Zigzag[i]] = Nodes[i].Level; |
|
|
i -= Nodes[i].Run; |
|
|
} |
|
|
|
|
|
if (DBG) { |
|
|
uint32_t Cost = Evaluate_Cost(Out,Mult,Bias, Zigzag,Non_Zero, Lambda); |
|
|
if (DBG==1) { |
|
|
printf( "<= " ); |
|
|
for(i=0; i<=Last_Node; ++i) printf( "[%3.0d] ", Out[Zigzag[i]] ); |
|
|
printf( "\n--------------------------------\n" ); |
|
|
} |
|
|
if (Cost>Last_Cost) printf( "!!! %u > %u\n", Cost, Last_Cost ); |
|
|
} |
|
|
return Last_Node; |
|
|
} |
|
|
|
|
|
#undef DBG |
|
|
|
|
|
#endif |
|