249 |
sad16bi = sad16bi_c; |
sad16bi = sad16bi_c; |
250 |
sad8bi = sad8bi_c; |
sad8bi = sad8bi_c; |
251 |
dev16 = dev16_c; |
dev16 = dev16_c; |
252 |
|
sad16v = sad16v_c; |
253 |
|
|
254 |
Halfpel8_Refine = Halfpel8_Refine_c; |
// Halfpel8_Refine = Halfpel8_Refine_c; |
255 |
|
|
256 |
#ifdef ARCH_X86 |
#ifdef ARCH_X86 |
257 |
if ((cpu_flags & XVID_CPU_MMX) > 0) { |
if ((cpu_flags & XVID_CPU_MMX) > 0) { |
308 |
sad16bi = sad16bi_mmx; |
sad16bi = sad16bi_mmx; |
309 |
sad8bi = sad8bi_mmx; |
sad8bi = sad8bi_mmx; |
310 |
dev16 = dev16_mmx; |
dev16 = dev16_mmx; |
311 |
|
sad16v = sad16v_mmx; |
312 |
|
|
313 |
} |
} |
314 |
|
|
347 |
sad16bi = sad16bi_xmm; |
sad16bi = sad16bi_xmm; |
348 |
sad8bi = sad8bi_xmm; |
sad8bi = sad8bi_xmm; |
349 |
dev16 = dev16_xmm; |
dev16 = dev16_xmm; |
350 |
|
sad16v = sad16v_xmm; |
351 |
|
fprintf(stderr,"sad16v=XMM\n"); |
352 |
|
|
353 |
} |
} |
354 |
|
|
395 |
sad16bi = sad16bi_ia64; |
sad16bi = sad16bi_ia64; |
396 |
sad8 = sad8_ia64; |
sad8 = sad8_ia64; |
397 |
dev16 = dev16_ia64; |
dev16 = dev16_ia64; |
398 |
Halfpel8_Refine = Halfpel8_Refine_ia64; |
// Halfpel8_Refine = Halfpel8_Refine_ia64; |
399 |
quant_intra = quant_intra_ia64; |
quant_intra = quant_intra_ia64; |
400 |
dequant_intra = dequant_intra_ia64; |
dequant_intra = dequant_intra_ia64; |
401 |
quant_inter = quant_inter_ia64; |
quant_inter = quant_inter_ia64; |