--- xvid.c 2002/07/21 23:34:08 1.32 +++ xvid.c 2002/09/25 21:26:36 1.33.2.4 @@ -37,7 +37,7 @@ * - 22.12.2001 API change: added xvid_init() - Isibaar * - 16.12.2001 inital version; (c)2001 peter ross * - * $Id: xvid.c,v 1.32 2002/07/21 23:34:08 chl Exp $ + * $Id: xvid.c,v 1.33.2.4 2002/09/25 21:26:36 Isibaar Exp $ * ****************************************************************************/ @@ -245,11 +245,13 @@ /* Functions used in motion estimation algorithms */ calc_cbp = calc_cbp_c; sad16 = sad16_c; - sad16bi = sad16bi_c; sad8 = sad8_c; - dev16 = dev16_c; + sad16bi = sad16bi_c; sad8bi = sad8bi_c; - Halfpel8_Refine = Halfpel8_Refine_c; + dev16 = dev16_c; + sad16v = sad16v_c; + +// Halfpel8_Refine = Halfpel8_Refine_c; #ifdef ARCH_X86 if ((cpu_flags & XVID_CPU_MMX) > 0) { @@ -303,10 +305,22 @@ calc_cbp = calc_cbp_mmx; sad16 = sad16_mmx; sad8 = sad8_mmx; + sad16bi = sad16bi_mmx; + sad8bi = sad8bi_mmx; dev16 = dev16_mmx; + sad16v = sad16v_mmx; + + } + /* these 3dnow functions are faster than mmx, but slower than xmm. */ + if ((cpu_flags & XVID_CPU_3DNOW) > 0) { + + /* ME functions */ + sad16bi = sad16bi_3dn; + sad8bi = sad8bi_3dn; } + if ((cpu_flags & XVID_CPU_MMXEXT) > 0) { /* Inverse DCT */ @@ -329,9 +343,12 @@ /* ME functions */ sad16 = sad16_xmm; - sad16bi = sad16bi_xmm; sad8 = sad8_xmm; + sad16bi = sad16bi_xmm; + sad8bi = sad8bi_xmm; dev16 = dev16_xmm; + sad16v = sad16v_xmm; + fprintf(stderr,"sad16v=XMM\n"); } @@ -378,7 +395,7 @@ sad16bi = sad16bi_ia64; sad8 = sad8_ia64; dev16 = dev16_ia64; - Halfpel8_Refine = Halfpel8_Refine_ia64; +// Halfpel8_Refine = Halfpel8_Refine_ia64; quant_intra = quant_intra_ia64; dequant_intra = dequant_intra_ia64; quant_inter = quant_inter_ia64;