120 |
INTERPOLATE8X8 interpolate8x8_halfpel_hv_ia64; |
INTERPOLATE8X8 interpolate8x8_halfpel_hv_ia64; |
121 |
#endif |
#endif |
122 |
|
|
123 |
|
#ifdef ARCH_IS_PPC |
124 |
|
INTERPOLATE8X8 interpolate8x8_halfpel_h_altivec_c; |
125 |
|
INTERPOLATE8X8 interpolate8x8_halfpel_v_altivec_c; |
126 |
|
INTERPOLATE8X8 interpolate8x8_halfpel_hv_altivec_c; |
127 |
|
#endif |
128 |
|
|
129 |
INTERPOLATE8X8_AVG2 interpolate8x8_avg2_c; |
INTERPOLATE8X8_AVG2 interpolate8x8_avg2_c; |
130 |
INTERPOLATE8X8_AVG4 interpolate8x8_avg4_c; |
INTERPOLATE8X8_AVG4 interpolate8x8_avg4_c; |
131 |
|
|
134 |
INTERPOLATE8X8_AVG4 interpolate8x8_avg4_mmx; |
INTERPOLATE8X8_AVG4 interpolate8x8_avg4_mmx; |
135 |
#endif |
#endif |
136 |
|
|
137 |
|
#ifdef ARCH_IS_PPC |
138 |
|
INTERPOLATE8X8_AVG2 interpolate8x8_avg2_altivec_c; |
139 |
|
INTERPOLATE8X8_AVG4 interpolate8x8_avg4_altivec_c; |
140 |
|
#endif |
141 |
|
|
142 |
INTERPOLATE_LOWPASS interpolate8x8_lowpass_h_c; |
INTERPOLATE_LOWPASS interpolate8x8_lowpass_h_c; |
143 |
INTERPOLATE_LOWPASS interpolate8x8_lowpass_v_c; |
INTERPOLATE_LOWPASS interpolate8x8_lowpass_v_c; |
144 |
|
|
156 |
INTERPOLATE8X8_6TAP_LOWPASS interpolate8x8_6tap_lowpass_v_mmx; |
INTERPOLATE8X8_6TAP_LOWPASS interpolate8x8_6tap_lowpass_v_mmx; |
157 |
#endif |
#endif |
158 |
|
|
159 |
|
#ifdef ARCH_IS_PPC |
160 |
|
INTERPOLATE8X8_6TAP_LOWPASS interpolate8x8_6tap_lowpass_h_altivec_c; |
161 |
|
#endif |
162 |
|
|
163 |
static __inline void |
static __inline void |
164 |
interpolate8x8_switch(uint8_t * const cur, |
interpolate8x8_switch(uint8_t * const cur, |
165 |
const uint8_t * const refn, |
const uint8_t * const refn, |
264 |
const uint32_t stride, |
const uint32_t stride, |
265 |
const uint32_t rounding) |
const uint32_t rounding) |
266 |
{ |
{ |
267 |
const int32_t xRef = (int)x*4 + dx; |
const int32_t xRef = x*4 + dx; |
268 |
const int32_t yRef = (int)y*4 + dy; |
const int32_t yRef = y*4 + dy; |
269 |
|
|
270 |
uint8_t *src, *dst; |
uint8_t *src, *dst; |
271 |
uint8_t *halfpel_h, *halfpel_v, *halfpel_hv; |
uint8_t *halfpel_h, *halfpel_v, *halfpel_hv; |
283 |
|
|
284 |
y_frac = yRef - (4*y_int); |
y_frac = yRef - (4*y_int); |
285 |
|
|
286 |
src = refn + y_int * (int)stride + x_int; |
src = refn + y_int * stride + x_int; |
287 |
halfpel_h = refh; |
halfpel_h = refh; |
288 |
halfpel_v = refv; |
halfpel_v = refv; |
289 |
halfpel_hv = refhv; |
halfpel_hv = refhv; |
390 |
const uint32_t stride, |
const uint32_t stride, |
391 |
const uint32_t rounding) |
const uint32_t rounding) |
392 |
{ |
{ |
393 |
const int32_t xRef = (int)x*4 + dx; |
const int32_t xRef = x*4 + dx; |
394 |
const int32_t yRef = (int)y*4 + dy; |
const int32_t yRef = y*4 + dy; |
395 |
|
|
396 |
uint8_t *src, *dst; |
uint8_t *src, *dst; |
397 |
uint8_t *halfpel_h, *halfpel_v, *halfpel_hv; |
uint8_t *halfpel_h, *halfpel_v, *halfpel_hv; |
409 |
|
|
410 |
y_frac = yRef - (4*y_int); |
y_frac = yRef - (4*y_int); |
411 |
|
|
412 |
src = refn + y_int * (int)stride + x_int; |
src = refn + y_int * stride + x_int; |
413 |
halfpel_h = refh; |
halfpel_h = refh; |
414 |
halfpel_v = refv; |
halfpel_v = refv; |
415 |
halfpel_hv = refhv; |
halfpel_hv = refhv; |