42 |
#include "../quant/quant.h" |
#include "../quant/quant.h" |
43 |
#include "../encoder.h" |
#include "../encoder.h" |
44 |
|
|
|
#include "../image/reduced.h" |
|
45 |
#include "../quant/quant_matrix.h" |
#include "../quant/quant_matrix.h" |
46 |
|
|
47 |
MBFIELDTEST_PTR MBFieldTest; |
MBFIELDTEST_PTR MBFieldTest; |
309 |
uint32_t stride = pParam->edged_width; |
uint32_t stride = pParam->edged_width; |
310 |
uint32_t stride2 = stride / 2; |
uint32_t stride2 = stride / 2; |
311 |
uint32_t next_block = stride * 8; |
uint32_t next_block = stride * 8; |
|
int32_t cst; |
|
|
int vop_reduced; |
|
312 |
uint8_t *pY_Cur, *pU_Cur, *pV_Cur; |
uint8_t *pY_Cur, *pU_Cur, *pV_Cur; |
313 |
const IMAGE * const pCurrent = &frame->image; |
const IMAGE * const pCurrent = &frame->image; |
|
transfer_operation_8to16_t * const functions[2] = |
|
|
{ |
|
|
(transfer_operation_8to16_t *)transfer_8to16copy, |
|
|
(transfer_operation_8to16_t *)filter_18x18_to_8x8 |
|
|
}; |
|
|
transfer_operation_8to16_t *transfer_op = NULL; |
|
|
|
|
|
vop_reduced = !!(frame->vop_flags & XVID_VOP_REDUCED); |
|
314 |
|
|
315 |
/* Image pointers */ |
/* Image pointers */ |
316 |
pY_Cur = pCurrent->y + (y_pos << (4+vop_reduced)) * stride + (x_pos << (4+vop_reduced)); |
pY_Cur = pCurrent->y + (y_pos << 4) * stride + (x_pos << 4); |
317 |
pU_Cur = pCurrent->u + (y_pos << (3+vop_reduced)) * stride2 + (x_pos << (3+vop_reduced)); |
pU_Cur = pCurrent->u + (y_pos << 3) * stride2 + (x_pos << 3); |
318 |
pV_Cur = pCurrent->v + (y_pos << (3+vop_reduced)) * stride2 + (x_pos << (3+vop_reduced)); |
pV_Cur = pCurrent->v + (y_pos << 3) * stride2 + (x_pos << 3); |
|
|
|
|
/* Block size */ |
|
|
cst = 8<<vop_reduced; |
|
|
|
|
|
/* Operation function */ |
|
|
transfer_op = functions[vop_reduced]; |
|
319 |
|
|
320 |
/* Do the transfer */ |
/* Do the transfer */ |
321 |
start_timer(); |
start_timer(); |
322 |
transfer_op(&data[0 * 64], pY_Cur, stride); |
transfer_8to16copy(&data[0 * 64], pY_Cur, stride); |
323 |
transfer_op(&data[1 * 64], pY_Cur + cst, stride); |
transfer_8to16copy(&data[1 * 64], pY_Cur + 8, stride); |
324 |
transfer_op(&data[2 * 64], pY_Cur + next_block, stride); |
transfer_8to16copy(&data[2 * 64], pY_Cur + next_block, stride); |
325 |
transfer_op(&data[3 * 64], pY_Cur + next_block + cst, stride); |
transfer_8to16copy(&data[3 * 64], pY_Cur + next_block + 8, stride); |
326 |
transfer_op(&data[4 * 64], pU_Cur, stride2); |
transfer_8to16copy(&data[4 * 64], pU_Cur, stride2); |
327 |
transfer_op(&data[5 * 64], pV_Cur, stride2); |
transfer_8to16copy(&data[5 * 64], pV_Cur, stride2); |
328 |
stop_transfer_timer(); |
stop_transfer_timer(); |
329 |
} |
} |
330 |
|
|
342 |
uint32_t stride = pParam->edged_width; |
uint32_t stride = pParam->edged_width; |
343 |
uint32_t stride2 = stride / 2; |
uint32_t stride2 = stride / 2; |
344 |
uint32_t next_block = stride * 8; |
uint32_t next_block = stride * 8; |
|
uint32_t cst; |
|
|
int vop_reduced; |
|
345 |
const IMAGE * const pCurrent = &frame->image; |
const IMAGE * const pCurrent = &frame->image; |
346 |
|
|
347 |
/* Array of function pointers, indexed by [vop_reduced<<1+add] */ |
/* Array of function pointers, indexed by [add] */ |
348 |
transfer_operation_16to8_t * const functions[4] = |
transfer_operation_16to8_t * const functions[2] = |
349 |
{ |
{ |
350 |
(transfer_operation_16to8_t*)transfer_16to8copy, |
(transfer_operation_16to8_t*)transfer_16to8copy, |
351 |
(transfer_operation_16to8_t*)transfer_16to8add, |
(transfer_operation_16to8_t*)transfer_16to8add, |
|
(transfer_operation_16to8_t*)copy_upsampled_8x8_16to8, |
|
|
(transfer_operation_16to8_t*)add_upsampled_8x8_16to8 |
|
352 |
}; |
}; |
353 |
|
|
354 |
transfer_operation_16to8_t *transfer_op = NULL; |
transfer_operation_16to8_t *transfer_op = NULL; |
355 |
|
|
|
/* Makes this vars booleans */ |
|
|
vop_reduced = !!(frame->vop_flags & XVID_VOP_REDUCED); |
|
|
|
|
356 |
/* Image pointers */ |
/* Image pointers */ |
357 |
pY_Cur = pCurrent->y + (y_pos << (4+vop_reduced)) * stride + (x_pos << (4+vop_reduced)); |
pY_Cur = pCurrent->y + (y_pos << 4) * stride + (x_pos << 4); |
358 |
pU_Cur = pCurrent->u + (y_pos << (3+vop_reduced)) * stride2 + (x_pos << (3+vop_reduced)); |
pU_Cur = pCurrent->u + (y_pos << 3) * stride2 + (x_pos << 3); |
359 |
pV_Cur = pCurrent->v + (y_pos << (3+vop_reduced)) * stride2 + (x_pos << (3+vop_reduced)); |
pV_Cur = pCurrent->v + (y_pos << 3) * stride2 + (x_pos << 3); |
360 |
|
|
361 |
if (pMB->field_dct) { |
if (pMB->field_dct) { |
362 |
next_block = stride; |
next_block = stride; |
363 |
stride *= 2; |
stride *= 2; |
364 |
} |
} |
365 |
|
|
|
/* Block size */ |
|
|
cst = 8<<vop_reduced; |
|
|
|
|
366 |
/* Operation function */ |
/* Operation function */ |
367 |
transfer_op = functions[(vop_reduced<<1) + add]; |
transfer_op = functions[add]; |
368 |
|
|
369 |
/* Do the operation */ |
/* Do the operation */ |
370 |
start_timer(); |
start_timer(); |
371 |
if (cbp&32) transfer_op(pY_Cur, &data[0 * 64], stride); |
if (cbp&32) transfer_op(pY_Cur, &data[0 * 64], stride); |
372 |
if (cbp&16) transfer_op(pY_Cur + cst, &data[1 * 64], stride); |
if (cbp&16) transfer_op(pY_Cur + 8, &data[1 * 64], stride); |
373 |
if (cbp& 8) transfer_op(pY_Cur + next_block, &data[2 * 64], stride); |
if (cbp& 8) transfer_op(pY_Cur + next_block, &data[2 * 64], stride); |
374 |
if (cbp& 4) transfer_op(pY_Cur + next_block + cst, &data[3 * 64], stride); |
if (cbp& 4) transfer_op(pY_Cur + next_block + 8, &data[3 * 64], stride); |
375 |
if (cbp& 2) transfer_op(pU_Cur, &data[4 * 64], stride2); |
if (cbp& 2) transfer_op(pU_Cur, &data[4 * 64], stride2); |
376 |
if (cbp& 1) transfer_op(pV_Cur, &data[5 * 64], stride2); |
if (cbp& 1) transfer_op(pV_Cur, &data[5 * 64], stride2); |
377 |
stop_transfer_timer(); |
stop_transfer_timer(); |