78 |
extern XVID_QP_FUNCS xvid_QP_Add_Funcs_mmx; |
extern XVID_QP_FUNCS xvid_QP_Add_Funcs_mmx; |
79 |
#endif |
#endif |
80 |
|
|
81 |
|
#ifdef ARCH_IS_PPC |
82 |
|
extern XVID_QP_FUNCS xvid_QP_Funcs_Altivec_C; |
83 |
|
extern XVID_QP_FUNCS xvid_QP_Add_Funcs_Altivec_C; |
84 |
|
#endif |
85 |
|
|
86 |
extern XVID_QP_FUNCS *xvid_QP_Funcs; /* <- main pointer for enc/dec structure */ |
extern XVID_QP_FUNCS *xvid_QP_Funcs; /* <- main pointer for enc/dec structure */ |
87 |
extern XVID_QP_FUNCS *xvid_QP_Add_Funcs; /* <- main pointer for enc/dec structure */ |
extern XVID_QP_FUNCS *xvid_QP_Add_Funcs; /* <- main pointer for enc/dec structure */ |
88 |
|
|
125 |
|
|
126 |
int32_t x_int, y_int; |
int32_t x_int, y_int; |
127 |
|
|
128 |
const int32_t xRef = x*4 + dx; |
const int32_t xRef = (int)x*4 + dx; |
129 |
const int32_t yRef = y*4 + dy; |
const int32_t yRef = (int)y*4 + dy; |
130 |
|
|
131 |
Ops = xvid_QP_Funcs; |
Ops = xvid_QP_Funcs; |
132 |
quads = (dx&3) | ((dy&3)<<2); |
quads = (dx&3) | ((dy&3)<<2); |
140 |
y_int--; |
y_int--; |
141 |
|
|
142 |
dst = cur + y * stride + x; |
dst = cur + y * stride + x; |
143 |
src = refn + y_int * stride + x_int; |
src = refn + y_int * (int)stride + x_int; |
144 |
|
|
145 |
tmp = refh; /* we need at least a 16 x stride scratch block */ |
tmp = refh; /* we need at least a 16 x stride scratch block */ |
146 |
|
|
228 |
|
|
229 |
int32_t x_int, y_int; |
int32_t x_int, y_int; |
230 |
|
|
231 |
const int32_t xRef = x*4 + dx; |
const int32_t xRef = (int)x*4 + dx; |
232 |
const int32_t yRef = y*4 + dy; |
const int32_t yRef = (int)y*4 + dy; |
233 |
|
|
234 |
Ops = xvid_QP_Add_Funcs; |
Ops = xvid_QP_Add_Funcs; |
235 |
Ops_Copy = xvid_QP_Funcs; |
Ops_Copy = xvid_QP_Funcs; |
244 |
y_int--; |
y_int--; |
245 |
|
|
246 |
dst = cur + y * stride + x; |
dst = cur + y * stride + x; |
247 |
src = refn + y_int * stride + x_int; |
src = refn + y_int * (int)stride + x_int; |
248 |
|
|
249 |
tmp = refh; /* we need at least a 16 x stride scratch block */ |
tmp = refh; /* we need at least a 16 x stride scratch block */ |
250 |
|
|
333 |
|
|
334 |
int32_t x_int, y_int; |
int32_t x_int, y_int; |
335 |
|
|
336 |
const int32_t xRef = x*4 + dx; |
const int32_t xRef = (int)x*4 + dx; |
337 |
const int32_t yRef = y*4 + dy; |
const int32_t yRef = (int)y*4 + dy; |
338 |
|
|
339 |
Ops = xvid_QP_Funcs; |
Ops = xvid_QP_Funcs; |
340 |
quads = (dx&3) | ((dy&3)<<2); |
quads = (dx&3) | ((dy&3)<<2); |
348 |
y_int--; |
y_int--; |
349 |
|
|
350 |
dst = cur + y * stride + x; |
dst = cur + y * stride + x; |
351 |
src = refn + y_int * stride + x_int; |
src = refn + y_int * (int)stride + x_int; |
352 |
|
|
353 |
tmp = refh; /* we need at least a 16 x stride scratch block */ |
tmp = refh; /* we need at least a 16 x stride scratch block */ |
354 |
|
|
433 |
|
|
434 |
int32_t x_int, y_int; |
int32_t x_int, y_int; |
435 |
|
|
436 |
const int32_t xRef = x*4 + dx; |
const int32_t xRef = (int)x*4 + dx; |
437 |
const int32_t yRef = y*4 + dy; |
const int32_t yRef = (int)y*4 + dy; |
438 |
|
|
439 |
Ops = xvid_QP_Funcs; |
Ops = xvid_QP_Funcs; |
440 |
quads = (dx&3) | ((dy&3)<<2); |
quads = (dx&3) | ((dy&3)<<2); |
448 |
y_int--; |
y_int--; |
449 |
|
|
450 |
dst = cur + y * stride + x; |
dst = cur + y * stride + x; |
451 |
src = refn + y_int * stride + x_int; |
src = refn + y_int * (int)stride + x_int; |
452 |
|
|
453 |
tmp = refh; /* we need at least a 16 x stride scratch block */ |
tmp = refh; /* we need at least a 16 x stride scratch block */ |
454 |
|
|
533 |
|
|
534 |
int32_t x_int, y_int; |
int32_t x_int, y_int; |
535 |
|
|
536 |
const int32_t xRef = x*4 + dx; |
const int32_t xRef = (int)x*4 + dx; |
537 |
const int32_t yRef = y*4 + dy; |
const int32_t yRef = (int)y*4 + dy; |
538 |
|
|
539 |
Ops = xvid_QP_Add_Funcs; |
Ops = xvid_QP_Add_Funcs; |
540 |
Ops_Copy = xvid_QP_Funcs; |
Ops_Copy = xvid_QP_Funcs; |
549 |
y_int--; |
y_int--; |
550 |
|
|
551 |
dst = cur + y * stride + x; |
dst = cur + y * stride + x; |
552 |
src = refn + y_int * stride + x_int; |
src = refn + y_int * (int)stride + x_int; |
553 |
|
|
554 |
tmp = refh; /* we need at least a 16 x stride scratch block */ |
tmp = refh; /* we need at least a 16 x stride scratch block */ |
555 |
|
|