373 |
DECLARE_ALIGNED_MATRIX(block, 1, 64, int16_t, CACHE_LINE); |
DECLARE_ALIGNED_MATRIX(block, 1, 64, int16_t, CACHE_LINE); |
374 |
DECLARE_ALIGNED_MATRIX(data, 6, 64, int16_t, CACHE_LINE); |
DECLARE_ALIGNED_MATRIX(data, 6, 64, int16_t, CACHE_LINE); |
375 |
|
|
376 |
int stride = dec->edged_width, next_block = stride * (reduced_resolution ? 16 : 8); |
int stride = dec->edged_width; |
377 |
|
int next_block = stride * (reduced_resolution ? 16 : 8); |
378 |
const int stride2 = stride/2; |
const int stride2 = stride/2; |
379 |
int i; |
int i; |
380 |
const uint32_t iQuant = pMB->quant; |
const uint32_t iQuant = pMB->quant; |
451 |
{ |
{ |
452 |
uint32_t stride = dec->edged_width; |
uint32_t stride = dec->edged_width; |
453 |
uint32_t stride2 = stride / 2; |
uint32_t stride2 = stride / 2; |
|
uint32_t next_block = stride * (reduced_resolution ? 16 : 8); |
|
454 |
uint32_t i; |
uint32_t i; |
455 |
|
|
456 |
uint8_t *pY_Cur, *pU_Cur, *pV_Cur; |
uint8_t *pY_Cur, *pU_Cur, *pV_Cur; |
963 |
{ |
{ |
964 |
uint32_t stride = dec->edged_width; |
uint32_t stride = dec->edged_width; |
965 |
uint32_t stride2 = stride / 2; |
uint32_t stride2 = stride / 2; |
|
uint32_t next_block = stride * 8; |
|
966 |
int uv_dx, uv_dy; |
int uv_dx, uv_dy; |
967 |
int b_uv_dx, b_uv_dy; |
int b_uv_dx, b_uv_dy; |
968 |
uint8_t *pY_Cur, *pU_Cur, *pV_Cur; |
uint8_t *pY_Cur, *pU_Cur, *pV_Cur; |