32 |
* |
* |
33 |
* History: |
* History: |
34 |
* |
* |
35 |
|
* 29.03.2002 interlacing fix - compensated block wasn't being used when |
36 |
|
* reconstructing blocks, thus artifacts |
37 |
|
* interlacing speedup - used transfers to re-interlace |
38 |
|
* interlaced decoding should be as fast as progressive now |
39 |
|
* 26.03.2002 interlacing support - moved transfers outside decode loop |
40 |
* 26.12.2001 decoder_mbinter: dequant/idct moved within if(coded) block |
* 26.12.2001 decoder_mbinter: dequant/idct moved within if(coded) block |
41 |
* 22.12.2001 block based interpolation |
* 22.12.2001 block based interpolation |
42 |
* 01.12.2001 inital version; (c)2001 peter ross <pross@cs.rmit.edu.au> |
* 01.12.2001 inital version; (c)2001 peter ross <pross@cs.rmit.edu.au> |
73 |
{ |
{ |
74 |
DECODER * dec; |
DECODER * dec; |
75 |
|
|
76 |
dec = xvid_malloc(sizeof(DECODER), 16); |
dec = xvid_malloc(sizeof(DECODER), CACHE_LINE); |
77 |
if (dec == NULL) |
if (dec == NULL) |
78 |
{ |
{ |
79 |
return XVID_ERR_MEMORY; |
return XVID_ERR_MEMORY; |
102 |
return XVID_ERR_MEMORY; |
return XVID_ERR_MEMORY; |
103 |
} |
} |
104 |
|
|
105 |
dec->mbs = xvid_malloc(sizeof(MACROBLOCK) * dec->mb_width * dec->mb_height, 16); |
dec->mbs = xvid_malloc(sizeof(MACROBLOCK) * dec->mb_width * dec->mb_height, CACHE_LINE); |
106 |
if (dec->mbs == NULL) |
if (dec->mbs == NULL) |
107 |
{ |
{ |
108 |
image_destroy(&dec->cur, dec->edged_width, dec->edged_height); |
image_destroy(&dec->cur, dec->edged_width, dec->edged_height); |
140 |
|
|
141 |
// decode an intra macroblock |
// decode an intra macroblock |
142 |
|
|
143 |
void decoder_mbintra(DECODER * dec, MACROBLOCK * mb, int x, int y, uint32_t acpred_flag, uint32_t cbp, Bitstream * bs, int quant, int intra_dc_threshold) |
void decoder_mbintra(DECODER * dec, |
144 |
{ |
MACROBLOCK * pMB, |
145 |
uint32_t k; |
const uint32_t x_pos, |
146 |
|
const uint32_t y_pos, |
147 |
|
const uint32_t acpred_flag, |
148 |
|
const uint32_t cbp, |
149 |
|
Bitstream * bs, |
150 |
|
const uint32_t quant, |
151 |
|
const uint32_t intra_dc_threshold) |
152 |
|
{ |
153 |
|
|
154 |
|
DECLARE_ALIGNED_MATRIX(block, 6, 64, int16_t, CACHE_LINE); |
155 |
|
DECLARE_ALIGNED_MATRIX(data, 6, 64, int16_t, CACHE_LINE); |
156 |
|
|
157 |
|
uint32_t stride = dec->edged_width; |
158 |
|
uint32_t stride2 = stride / 2; |
159 |
|
uint32_t next_block = stride * 8; |
160 |
|
uint32_t i; |
161 |
|
uint32_t iQuant = pMB->quant; |
162 |
|
uint8_t *pY_Cur, *pU_Cur, *pV_Cur; |
163 |
|
|
164 |
|
pY_Cur = dec->cur.y + (y_pos << 4) * stride + (x_pos << 4); |
165 |
|
pU_Cur = dec->cur.u + (y_pos << 3) * stride2 + (x_pos << 3); |
166 |
|
pV_Cur = dec->cur.v + (y_pos << 3) * stride2 + (x_pos << 3); |
167 |
|
|
168 |
|
memset(block, 0, 6*64*sizeof(int16_t)); // clear |
169 |
|
|
170 |
for (k = 0; k < 6; k++) |
for (i = 0; i < 6; i++) |
171 |
{ |
{ |
172 |
uint32_t dcscalar; |
uint32_t iDcScaler = get_dc_scaler(iQuant, i < 4); |
|
int16_t block[64]; |
|
|
int16_t data[64]; |
|
173 |
int16_t predictors[8]; |
int16_t predictors[8]; |
174 |
int start_coeff; |
int start_coeff; |
175 |
|
|
|
dcscalar = get_dc_scaler(mb->quant, k < 4); |
|
|
|
|
176 |
start_timer(); |
start_timer(); |
177 |
predict_acdc(dec->mbs, x, y, dec->mb_width, k, block, mb->quant, dcscalar, predictors); |
predict_acdc(dec->mbs, x_pos, y_pos, dec->mb_width, i, &block[i*64], iQuant, iDcScaler, predictors); |
178 |
if (!acpred_flag) |
if (!acpred_flag) |
179 |
{ |
{ |
180 |
mb->acpred_directions[k] = 0; |
pMB->acpred_directions[i] = 0; |
181 |
} |
} |
182 |
stop_prediction_timer(); |
stop_prediction_timer(); |
183 |
|
|
|
memset(block, 0, 64*sizeof(int16_t)); // clear |
|
|
|
|
184 |
if (quant < intra_dc_threshold) |
if (quant < intra_dc_threshold) |
185 |
{ |
{ |
186 |
int dc_size; |
int dc_size; |
187 |
int dc_dif; |
int dc_dif; |
188 |
|
|
189 |
dc_size = k < 4 ? get_dc_size_lum(bs) : get_dc_size_chrom(bs); |
dc_size = i < 4 ? get_dc_size_lum(bs) : get_dc_size_chrom(bs); |
190 |
dc_dif = dc_size ? get_dc_dif(bs, dc_size) : 0 ; |
dc_dif = dc_size ? get_dc_dif(bs, dc_size) : 0 ; |
191 |
|
|
192 |
if (dc_size > 8) |
if (dc_size > 8) |
194 |
BitstreamSkip(bs, 1); // marker |
BitstreamSkip(bs, 1); // marker |
195 |
} |
} |
196 |
|
|
197 |
block[0] = dc_dif; |
block[i*64 + 0] = dc_dif; |
198 |
start_coeff = 1; |
start_coeff = 1; |
199 |
} |
} |
200 |
else |
else |
203 |
} |
} |
204 |
|
|
205 |
start_timer(); |
start_timer(); |
206 |
if (cbp & (1 << (5-k))) // coded |
if (cbp & (1 << (5-i))) // coded |
207 |
{ |
{ |
208 |
get_intra_block(bs, block, mb->acpred_directions[k], start_coeff); |
get_intra_block(bs, &block[i*64], pMB->acpred_directions[i], start_coeff); |
209 |
} |
} |
210 |
stop_coding_timer(); |
stop_coding_timer(); |
211 |
|
|
212 |
start_timer(); |
start_timer(); |
213 |
add_acdc(mb, k, block, dcscalar, predictors); |
add_acdc(pMB, i, &block[i*64], iDcScaler, predictors); |
214 |
stop_prediction_timer(); |
stop_prediction_timer(); |
215 |
|
|
216 |
start_timer(); |
start_timer(); |
217 |
if (dec->quant_type == 0) |
if (dec->quant_type == 0) |
218 |
{ |
{ |
219 |
dequant_intra(data, block, mb->quant, dcscalar); |
dequant_intra(&data[i*64], &block[i*64], iQuant, iDcScaler); |
220 |
} |
} |
221 |
else |
else |
222 |
{ |
{ |
223 |
dequant4_intra(data, block, mb->quant, dcscalar); |
dequant4_intra(&data[i*64], &block[i*64], iQuant, iDcScaler); |
224 |
} |
} |
225 |
stop_iquant_timer(); |
stop_iquant_timer(); |
226 |
|
|
227 |
start_timer(); |
start_timer(); |
228 |
idct(data); |
idct(&data[i*64]); |
229 |
stop_idct_timer(); |
stop_idct_timer(); |
|
|
|
|
start_timer(); |
|
|
if (k < 4) |
|
|
{ |
|
|
transfer_16to8copy(dec->cur.y + (16*y*dec->edged_width) + 16*x + (4*(k&2)*dec->edged_width) + 8*(k&1), data, dec->edged_width); |
|
230 |
} |
} |
231 |
else if (k == 4) |
|
232 |
{ |
if (pMB->field_dct) |
|
transfer_16to8copy(dec->cur.u+ 8*y*(dec->edged_width/2) + 8*x, data, (dec->edged_width/2)); |
|
|
} |
|
|
else // if (k == 5) |
|
233 |
{ |
{ |
234 |
transfer_16to8copy(dec->cur.v + 8*y*(dec->edged_width/2) + 8*x, data, (dec->edged_width/2)); |
next_block = stride; |
235 |
|
stride *= 2; |
236 |
} |
} |
237 |
|
|
238 |
|
start_timer(); |
239 |
|
transfer_16to8copy(pY_Cur, &data[0*64], stride); |
240 |
|
transfer_16to8copy(pY_Cur + 8, &data[1*64], stride); |
241 |
|
transfer_16to8copy(pY_Cur + next_block, &data[2*64], stride); |
242 |
|
transfer_16to8copy(pY_Cur + 8 + next_block, &data[3*64], stride); |
243 |
|
transfer_16to8copy(pU_Cur, &data[4*64], stride2); |
244 |
|
transfer_16to8copy(pV_Cur, &data[5*64], stride2); |
245 |
stop_transfer_timer(); |
stop_transfer_timer(); |
246 |
} |
} |
|
} |
|
247 |
|
|
248 |
|
|
249 |
|
|
257 |
|
|
258 |
// decode an inter macroblock |
// decode an inter macroblock |
259 |
|
|
260 |
void decoder_mbinter(DECODER * dec, MACROBLOCK * mb, int x, int y, uint32_t acpred_flag, uint32_t cbp, Bitstream * bs, int quant, int rounding) |
void decoder_mbinter(DECODER * dec, |
261 |
{ |
const MACROBLOCK * pMB, |
262 |
const uint32_t stride = dec->edged_width; |
const uint32_t x_pos, |
263 |
const uint32_t stride2 = dec->edged_width / 2; |
const uint32_t y_pos, |
264 |
|
const uint32_t acpred_flag, |
265 |
|
const uint32_t cbp, |
266 |
|
Bitstream * bs, |
267 |
|
const uint32_t quant, |
268 |
|
const uint32_t rounding) |
269 |
|
{ |
270 |
|
|
271 |
|
DECLARE_ALIGNED_MATRIX(block,6, 64, int16_t, CACHE_LINE); |
272 |
|
DECLARE_ALIGNED_MATRIX(data, 6, 64, int16_t, CACHE_LINE); |
273 |
|
|
274 |
|
uint32_t stride = dec->edged_width; |
275 |
|
uint32_t stride2 = stride / 2; |
276 |
|
uint32_t next_block = stride * 8; |
277 |
|
uint32_t i; |
278 |
|
uint32_t iQuant = pMB->quant; |
279 |
|
uint8_t *pY_Cur, *pU_Cur, *pV_Cur; |
280 |
int uv_dx, uv_dy; |
int uv_dx, uv_dy; |
|
uint32_t k; |
|
281 |
|
|
282 |
if (mb->mode == MODE_INTER || mb->mode == MODE_INTER_Q) |
pY_Cur = dec->cur.y + (y_pos << 4) * stride + (x_pos << 4); |
283 |
|
pU_Cur = dec->cur.u + (y_pos << 3) * stride2 + (x_pos << 3); |
284 |
|
pV_Cur = dec->cur.v + (y_pos << 3) * stride2 + (x_pos << 3); |
285 |
|
|
286 |
|
if (pMB->mode == MODE_INTER || pMB->mode == MODE_INTER_Q) |
287 |
{ |
{ |
288 |
uv_dx = mb->mvs[0].x; |
uv_dx = pMB->mvs[0].x; |
289 |
uv_dy = mb->mvs[0].y; |
uv_dy = pMB->mvs[0].y; |
290 |
|
|
291 |
uv_dx = (uv_dx & 3) ? (uv_dx >> 1) | 1 : uv_dx / 2; |
uv_dx = (uv_dx & 3) ? (uv_dx >> 1) | 1 : uv_dx / 2; |
292 |
uv_dy = (uv_dy & 3) ? (uv_dy >> 1) | 1 : uv_dy / 2; |
uv_dy = (uv_dy & 3) ? (uv_dy >> 1) | 1 : uv_dy / 2; |
294 |
else |
else |
295 |
{ |
{ |
296 |
int sum; |
int sum; |
297 |
sum = mb->mvs[0].x + mb->mvs[1].x + mb->mvs[2].x + mb->mvs[3].x; |
sum = pMB->mvs[0].x + pMB->mvs[1].x + pMB->mvs[2].x + pMB->mvs[3].x; |
298 |
uv_dx = (sum == 0 ? 0 : SIGN(sum) * (roundtab[ABS(sum) % 16] + (ABS(sum) / 16) * 2) ); |
uv_dx = (sum == 0 ? 0 : SIGN(sum) * (roundtab[ABS(sum) % 16] + (ABS(sum) / 16) * 2) ); |
299 |
|
|
300 |
sum = mb->mvs[0].y + mb->mvs[1].y + mb->mvs[2].y + mb->mvs[3].y; |
sum = pMB->mvs[0].y + pMB->mvs[1].y + pMB->mvs[2].y + pMB->mvs[3].y; |
301 |
uv_dy = (sum == 0 ? 0 : SIGN(sum) * (roundtab[ABS(sum) % 16] + (ABS(sum) / 16) * 2) ); |
uv_dy = (sum == 0 ? 0 : SIGN(sum) * (roundtab[ABS(sum) % 16] + (ABS(sum) / 16) * 2) ); |
302 |
} |
} |
303 |
|
|
304 |
start_timer(); |
start_timer(); |
305 |
interpolate8x8_switch(dec->cur.y, dec->refn.y, 16*x, 16*y , mb->mvs[0].x, mb->mvs[0].y, stride, rounding); |
interpolate8x8_switch(dec->cur.y, dec->refn.y, 16*x_pos, 16*y_pos , pMB->mvs[0].x, pMB->mvs[0].y, stride, rounding); |
306 |
interpolate8x8_switch(dec->cur.y, dec->refn.y, 16*x + 8, 16*y , mb->mvs[1].x, mb->mvs[1].y, stride, rounding); |
interpolate8x8_switch(dec->cur.y, dec->refn.y, 16*x_pos + 8, 16*y_pos , pMB->mvs[1].x, pMB->mvs[1].y, stride, rounding); |
307 |
interpolate8x8_switch(dec->cur.y, dec->refn.y, 16*x, 16*y + 8, mb->mvs[2].x, mb->mvs[2].y, stride, rounding); |
interpolate8x8_switch(dec->cur.y, dec->refn.y, 16*x_pos, 16*y_pos + 8, pMB->mvs[2].x, pMB->mvs[2].y, stride, rounding); |
308 |
interpolate8x8_switch(dec->cur.y, dec->refn.y, 16*x + 8, 16*y + 8, mb->mvs[3].x, mb->mvs[3].y, stride, rounding); |
interpolate8x8_switch(dec->cur.y, dec->refn.y, 16*x_pos + 8, 16*y_pos + 8, pMB->mvs[3].x, pMB->mvs[3].y, stride, rounding); |
309 |
interpolate8x8_switch(dec->cur.u, dec->refn.u, 8*x, 8*y, uv_dx, uv_dy, stride2, rounding); |
interpolate8x8_switch(dec->cur.u, dec->refn.u, 8*x_pos, 8*y_pos, uv_dx, uv_dy, stride2, rounding); |
310 |
interpolate8x8_switch(dec->cur.v, dec->refn.v, 8*x, 8*y, uv_dx, uv_dy, stride2, rounding); |
interpolate8x8_switch(dec->cur.v, dec->refn.v, 8*x_pos, 8*y_pos, uv_dx, uv_dy, stride2, rounding); |
311 |
stop_comp_timer(); |
stop_comp_timer(); |
312 |
|
|
313 |
|
for (i = 0; i < 6; i++) |
|
for (k = 0; k < 6; k++) |
|
314 |
{ |
{ |
315 |
int16_t block[64]; |
if (cbp & (1 << (5-i))) // coded |
|
int16_t data[64]; |
|
|
|
|
|
if (cbp & (1 << (5-k))) // coded |
|
316 |
{ |
{ |
317 |
memset(block, 0, 64 * sizeof(int16_t)); // clear |
memset(&block[i*64], 0, 64 * sizeof(int16_t)); // clear |
318 |
|
|
319 |
start_timer(); |
start_timer(); |
320 |
get_inter_block(bs, block); |
get_inter_block(bs, &block[i*64]); |
321 |
stop_coding_timer(); |
stop_coding_timer(); |
322 |
|
|
323 |
start_timer(); |
start_timer(); |
324 |
if (dec->quant_type == 0) |
if (dec->quant_type == 0) |
325 |
{ |
{ |
326 |
dequant_inter(data, block, mb->quant); |
dequant_inter(&data[i*64], &block[i*64], iQuant); |
327 |
} |
} |
328 |
else |
else |
329 |
{ |
{ |
330 |
dequant4_inter(data, block, mb->quant); |
dequant4_inter(&data[i*64], &block[i*64], iQuant); |
331 |
} |
} |
332 |
stop_iquant_timer(); |
stop_iquant_timer(); |
333 |
|
|
334 |
start_timer(); |
start_timer(); |
335 |
idct(data); |
idct(&data[i*64]); |
336 |
stop_idct_timer(); |
stop_idct_timer(); |
|
|
|
|
start_timer(); |
|
|
if (k < 4) |
|
|
{ |
|
|
transfer_16to8add(dec->cur.y + (16*y + 4*(k&2))*stride + 16*x + 8*(k&1), data, stride); |
|
337 |
} |
} |
|
else if (k == 4) |
|
|
{ |
|
|
transfer_16to8add(dec->cur.u + 8*y*stride2 + 8*x, data, stride2); |
|
338 |
} |
} |
339 |
else // k == 5 |
|
340 |
|
if (pMB->field_dct) |
341 |
{ |
{ |
342 |
transfer_16to8add(dec->cur.v + 8*y*stride2 + 8*x, data, stride2); |
next_block = stride; |
343 |
|
stride *= 2; |
344 |
} |
} |
345 |
|
|
346 |
|
start_timer(); |
347 |
|
if (cbp & 32) |
348 |
|
transfer_16to8add(pY_Cur, &data[0*64], stride); |
349 |
|
if (cbp & 16) |
350 |
|
transfer_16to8add(pY_Cur + 8, &data[1*64], stride); |
351 |
|
if (cbp & 8) |
352 |
|
transfer_16to8add(pY_Cur + next_block, &data[2*64], stride); |
353 |
|
if (cbp & 4) |
354 |
|
transfer_16to8add(pY_Cur + 8 + next_block, &data[3*64], stride); |
355 |
|
if (cbp & 2) |
356 |
|
transfer_16to8add(pU_Cur, &data[4*64], stride2); |
357 |
|
if (cbp & 1) |
358 |
|
transfer_16to8add(pV_Cur, &data[5*64], stride2); |
359 |
stop_transfer_timer(); |
stop_transfer_timer(); |
360 |
} |
} |
|
} |
|
|
} |
|
|
|
|
361 |
|
|
362 |
|
|
363 |
void decoder_iframe(DECODER * dec, Bitstream * bs, int quant, int intra_dc_threshold) |
void decoder_iframe(DECODER * dec, Bitstream * bs, int quant, int intra_dc_threshold) |
364 |
{ |
{ |
365 |
|
|
366 |
uint32_t x, y; |
uint32_t x, y; |
367 |
|
|
368 |
for (y = 0; y < dec->mb_height; y++) |
for (y = 0; y < dec->mb_height; y++) |
406 |
} |
} |
407 |
mb->quant = quant; |
mb->quant = quant; |
408 |
|
|
409 |
|
if (dec->interlacing) |
410 |
|
{ |
411 |
|
mb->field_dct = BitstreamGetBit(bs); |
412 |
|
DEBUG1("deci: field_dct: ", mb->field_dct); |
413 |
|
} |
414 |
|
|
415 |
decoder_mbintra(dec, mb, x, y, acpred_flag, cbp, bs, quant, intra_dc_threshold); |
decoder_mbintra(dec, mb, x, y, acpred_flag, cbp, bs, quant, intra_dc_threshold); |
416 |
} |
} |
417 |
} |
} |
418 |
|
|
419 |
} |
} |
420 |
|
|
421 |
|
|
422 |
void get_motion_vector(DECODER *dec, Bitstream *bs, int x, int y, int k, VECTOR * mv, int fcode) |
void get_motion_vector(DECODER *dec, Bitstream *bs, int x, int y, int k, VECTOR * mv, int fcode) |
423 |
{ |
{ |
424 |
|
|
425 |
int scale_fac = 1 << (fcode - 1); |
int scale_fac = 1 << (fcode - 1); |
426 |
int high = (32 * scale_fac) - 1; |
int high = (32 * scale_fac) - 1; |
427 |
int low = ((-32) * scale_fac); |
int low = ((-32) * scale_fac); |
471 |
|
|
472 |
void decoder_pframe(DECODER * dec, Bitstream * bs, int rounding, int quant, int fcode, int intra_dc_threshold) |
void decoder_pframe(DECODER * dec, Bitstream * bs, int rounding, int quant, int fcode, int intra_dc_threshold) |
473 |
{ |
{ |
474 |
|
|
475 |
uint32_t x, y; |
uint32_t x, y; |
476 |
|
|
477 |
image_swap(&dec->cur, &dec->refn); |
image_swap(&dec->cur, &dec->refn); |
478 |
|
|
479 |
start_timer(); |
start_timer(); |
480 |
image_setedges(&dec->refn, dec->edged_width, dec->edged_height, dec->width, dec->height); |
image_setedges(&dec->refn, dec->edged_width, dec->edged_height, dec->width, dec->height, dec->interlacing); |
481 |
stop_edges_timer(); |
stop_edges_timer(); |
482 |
|
|
483 |
for (y = 0; y < dec->mb_height; y++) |
for (y = 0; y < dec->mb_height; y++) |
530 |
} |
} |
531 |
mb->quant = quant; |
mb->quant = quant; |
532 |
|
|
533 |
|
if (dec->interlacing) |
534 |
|
{ |
535 |
|
mb->field_dct = BitstreamGetBit(bs); |
536 |
|
DEBUG1("decp: field_dct: ", mb->field_dct); |
537 |
|
|
538 |
if (mb->mode == MODE_INTER || mb->mode == MODE_INTER_Q) |
if (mb->mode == MODE_INTER || mb->mode == MODE_INTER_Q) |
539 |
{ |
{ |
540 |
|
mb->field_pred = BitstreamGetBit(bs); |
541 |
|
DEBUG1("decp: field_pred: ", mb->field_pred); |
542 |
|
|
543 |
|
if (mb->field_pred) |
544 |
|
{ |
545 |
|
mb->field_for_top = BitstreamGetBit(bs); |
546 |
|
DEBUG1("decp: field_for_top: ", mb->field_for_top); |
547 |
|
mb->field_for_bot = BitstreamGetBit(bs); |
548 |
|
DEBUG1("decp: field_for_bot: ", mb->field_for_bot); |
549 |
|
} |
550 |
|
} |
551 |
|
} |
552 |
|
|
553 |
|
if (mb->mode == MODE_INTER || mb->mode == MODE_INTER_Q) |
554 |
|
{ |
555 |
|
if (dec->interlacing && mb->field_pred) |
556 |
|
{ |
557 |
|
get_motion_vector(dec, bs, x, y, 0, &mb->mvs[0], fcode); |
558 |
|
get_motion_vector(dec, bs, x, y, 0, &mb->mvs[1], fcode); |
559 |
|
} |
560 |
|
else |
561 |
|
{ |
562 |
get_motion_vector(dec, bs, x, y, 0, &mb->mvs[0], fcode); |
get_motion_vector(dec, bs, x, y, 0, &mb->mvs[0], fcode); |
563 |
mb->mvs[1].x = mb->mvs[2].x = mb->mvs[3].x = mb->mvs[0].x; |
mb->mvs[1].x = mb->mvs[2].x = mb->mvs[3].x = mb->mvs[0].x; |
564 |
mb->mvs[1].y = mb->mvs[2].y = mb->mvs[3].y = mb->mvs[0].y; |
mb->mvs[1].y = mb->mvs[2].y = mb->mvs[3].y = mb->mvs[0].y; |
565 |
} |
} |
566 |
|
} |
567 |
else if (mb->mode == MODE_INTER4V /* || mb->mode == MODE_INTER4V_Q */) |
else if (mb->mode == MODE_INTER4V /* || mb->mode == MODE_INTER4V_Q */) |
568 |
{ |
{ |
569 |
get_motion_vector(dec, bs, x, y, 0, &mb->mvs[0], fcode); |
get_motion_vector(dec, bs, x, y, 0, &mb->mvs[0], fcode); |
624 |
|
|
625 |
int decoder_decode(DECODER * dec, XVID_DEC_FRAME * frame) |
int decoder_decode(DECODER * dec, XVID_DEC_FRAME * frame) |
626 |
{ |
{ |
627 |
|
|
628 |
Bitstream bs; |
Bitstream bs; |
629 |
uint32_t rounding; |
uint32_t rounding; |
630 |
uint32_t quant; |
uint32_t quant; |
668 |
stop_global_timer(); |
stop_global_timer(); |
669 |
|
|
670 |
return XVID_ERR_OK; |
return XVID_ERR_OK; |
671 |
|
|
672 |
} |
} |