36 |
#include "dct/fdct.h" |
#include "dct/fdct.h" |
37 |
#include "image/colorspace.h" |
#include "image/colorspace.h" |
38 |
#include "image/interpolate8x8.h" |
#include "image/interpolate8x8.h" |
|
#include "image/reduced.h" |
|
39 |
#include "utils/mem_transfer.h" |
#include "utils/mem_transfer.h" |
40 |
#include "utils/mbfunctions.h" |
#include "utils/mbfunctions.h" |
41 |
#include "quant/quant.h" |
#include "quant/quant.h" |
51 |
unsigned int xvid_debug = 0; /* xvid debug mask */ |
unsigned int xvid_debug = 0; /* xvid debug mask */ |
52 |
#endif |
#endif |
53 |
|
|
54 |
#if defined(ARCH_IS_IA32) |
#if defined(ARCH_IS_IA32) && defined(_MSC_VER) |
|
#if defined(_MSC_VER) |
|
55 |
# include <windows.h> |
# include <windows.h> |
56 |
#else |
#elif defined(ARCH_IS_IA32) || defined(ARCH_IS_PPC) |
57 |
# include <signal.h> |
# include <signal.h> |
58 |
# include <setjmp.h> |
# include <setjmp.h> |
59 |
|
|
76 |
* 0 : SIGILL was *not* signalled |
* 0 : SIGILL was *not* signalled |
77 |
* 1 : SIGILL was signalled |
* 1 : SIGILL was signalled |
78 |
*/ |
*/ |
79 |
|
#if defined(ARCH_IS_IA32) && defined(_MSC_VER) |
80 |
int |
static int |
81 |
sigill_check(void (*func)()) |
sigill_check(void (*func)()) |
82 |
{ |
{ |
|
#if defined(_MSC_VER) |
|
83 |
_try { |
_try { |
84 |
func(); |
func(); |
85 |
} |
} _except(EXCEPTION_EXECUTE_HANDLER) { |
|
_except(EXCEPTION_EXECUTE_HANDLER) { |
|
86 |
|
|
87 |
if (_exception_code() == STATUS_ILLEGAL_INSTRUCTION) |
if (_exception_code() == STATUS_ILLEGAL_INSTRUCTION) |
88 |
return 1; |
return(1); |
89 |
} |
} |
90 |
return 0; |
return(0); |
91 |
#else |
} |
92 |
|
#elif defined(ARCH_IS_IA32) || defined(ARCH_IS_PPC) |
93 |
|
static int |
94 |
|
sigill_check(void (*func)()) |
95 |
|
{ |
96 |
void * old_handler; |
void * old_handler; |
97 |
int jmpret; |
int jmpret; |
98 |
|
|
99 |
|
/* Set our SIGILL handler */ |
100 |
old_handler = signal(SIGILL, sigill_handler); |
old_handler = signal(SIGILL, sigill_handler); |
101 |
if (old_handler == SIG_ERR) |
|
102 |
{ |
/* Check for error */ |
103 |
return -1; |
if (old_handler == SIG_ERR) { |
104 |
|
return(-1); |
105 |
} |
} |
106 |
|
|
107 |
|
/* Save stack context, so if func triggers a SIGILL, we can still roll |
108 |
|
* back to a valid CPU state */ |
109 |
jmpret = setjmp(mark); |
jmpret = setjmp(mark); |
110 |
if (jmpret == 0) |
|
111 |
{ |
/* If setjmp returned directly, then its returned value is 0, and we still |
112 |
|
* have to test the passed func. Otherwise it means the stack context has |
113 |
|
* been restored by a longjmp() call, which in our case happens only in the |
114 |
|
* signal handler */ |
115 |
|
if (jmpret == 0) { |
116 |
func(); |
func(); |
117 |
} |
} |
118 |
|
|
119 |
|
/* Restore old signal handler */ |
120 |
signal(SIGILL, old_handler); |
signal(SIGILL, old_handler); |
121 |
|
|
122 |
return jmpret; |
return(jmpret); |
|
#endif |
|
123 |
} |
} |
124 |
#endif |
#endif |
125 |
|
|
141 |
#endif |
#endif |
142 |
|
|
143 |
#if defined(ARCH_IS_PPC) |
#if defined(ARCH_IS_PPC) |
144 |
#if defined(ARCH_IS_PPC_ALTIVEC) |
if (!sigill_check(altivec_trigger)) |
145 |
cpu_flags |= XVID_CPU_ALTIVEC; |
cpu_flags |= XVID_CPU_ALTIVEC; |
146 |
#endif |
#endif |
|
#endif |
|
147 |
|
|
148 |
return cpu_flags; |
return cpu_flags; |
149 |
} |
} |
210 |
transfer_8to16sub = transfer_8to16sub_c; |
transfer_8to16sub = transfer_8to16sub_c; |
211 |
transfer_8to16subro = transfer_8to16subro_c; |
transfer_8to16subro = transfer_8to16subro_c; |
212 |
transfer_8to16sub2 = transfer_8to16sub2_c; |
transfer_8to16sub2 = transfer_8to16sub2_c; |
213 |
|
transfer_8to16sub2ro = transfer_8to16sub2ro_c; |
214 |
transfer_16to8add = transfer_16to8add_c; |
transfer_16to8add = transfer_16to8add_c; |
215 |
transfer8x8_copy = transfer8x8_copy_c; |
transfer8x8_copy = transfer8x8_copy_c; |
216 |
|
|
222 |
interpolate8x8_halfpel_v = interpolate8x8_halfpel_v_c; |
interpolate8x8_halfpel_v = interpolate8x8_halfpel_v_c; |
223 |
interpolate8x8_halfpel_hv = interpolate8x8_halfpel_hv_c; |
interpolate8x8_halfpel_hv = interpolate8x8_halfpel_hv_c; |
224 |
|
|
225 |
|
interpolate8x8_halfpel_add = interpolate8x8_halfpel_add_c; |
226 |
|
interpolate8x8_halfpel_h_add = interpolate8x8_halfpel_h_add_c; |
227 |
|
interpolate8x8_halfpel_v_add = interpolate8x8_halfpel_v_add_c; |
228 |
|
interpolate8x8_halfpel_hv_add = interpolate8x8_halfpel_hv_add_c; |
229 |
|
|
230 |
interpolate16x16_lowpass_h = interpolate16x16_lowpass_h_c; |
interpolate16x16_lowpass_h = interpolate16x16_lowpass_h_c; |
231 |
interpolate16x16_lowpass_v = interpolate16x16_lowpass_v_c; |
interpolate16x16_lowpass_v = interpolate16x16_lowpass_v_c; |
232 |
interpolate16x16_lowpass_hv = interpolate16x16_lowpass_hv_c; |
interpolate16x16_lowpass_hv = interpolate16x16_lowpass_hv_c; |
244 |
/* postprocessing */ |
/* postprocessing */ |
245 |
image_brightness = image_brightness_c; |
image_brightness = image_brightness_c; |
246 |
|
|
|
/* reduced resolution */ |
|
|
copy_upsampled_8x8_16to8 = xvid_Copy_Upsampled_8x8_16To8_C; |
|
|
add_upsampled_8x8_16to8 = xvid_Add_Upsampled_8x8_16To8_C; |
|
|
vfilter_31 = xvid_VFilter_31_C; |
|
|
hfilter_31 = xvid_HFilter_31_C; |
|
|
filter_18x18_to_8x8 = xvid_Filter_18x18_To_8x8_C; |
|
|
filter_diff_18x18_to_8x8 = xvid_Filter_Diff_18x18_To_8x8_C; |
|
|
|
|
247 |
/* Initialize internal colorspace transformation tables */ |
/* Initialize internal colorspace transformation tables */ |
248 |
colorspace_init(); |
colorspace_init(); |
249 |
|
|
299 |
dev16 = dev16_c; |
dev16 = dev16_c; |
300 |
sad16v = sad16v_c; |
sad16v = sad16v_c; |
301 |
sse8_16bit = sse8_16bit_c; |
sse8_16bit = sse8_16bit_c; |
302 |
|
sse8_8bit = sse8_8bit_c; |
303 |
|
|
304 |
#if defined(ARCH_IS_IA32) |
#if defined(ARCH_IS_IA32) |
305 |
|
|
|
if ((cpu_flags & XVID_CPU_ASM)) { |
|
|
vfilter_31 = xvid_VFilter_31_x86; |
|
|
hfilter_31 = xvid_HFilter_31_x86; |
|
|
} |
|
|
|
|
306 |
if ((cpu_flags & XVID_CPU_MMX) || (cpu_flags & XVID_CPU_MMXEXT) || |
if ((cpu_flags & XVID_CPU_MMX) || (cpu_flags & XVID_CPU_MMXEXT) || |
307 |
(cpu_flags & XVID_CPU_3DNOW) || (cpu_flags & XVID_CPU_3DNOWEXT) || |
(cpu_flags & XVID_CPU_3DNOW) || (cpu_flags & XVID_CPU_3DNOWEXT) || |
308 |
(cpu_flags & XVID_CPU_SSE) || (cpu_flags & XVID_CPU_SSE2)) |
(cpu_flags & XVID_CPU_SSE) || (cpu_flags & XVID_CPU_SSE2)) |
349 |
interpolate8x8_halfpel_v = interpolate8x8_halfpel_v_mmx; |
interpolate8x8_halfpel_v = interpolate8x8_halfpel_v_mmx; |
350 |
interpolate8x8_halfpel_hv = interpolate8x8_halfpel_hv_mmx; |
interpolate8x8_halfpel_hv = interpolate8x8_halfpel_hv_mmx; |
351 |
|
|
352 |
|
interpolate8x8_halfpel_add = interpolate8x8_halfpel_add_mmx; |
353 |
|
interpolate8x8_halfpel_h_add = interpolate8x8_halfpel_h_add_mmx; |
354 |
|
interpolate8x8_halfpel_v_add = interpolate8x8_halfpel_v_add_mmx; |
355 |
|
interpolate8x8_halfpel_hv_add = interpolate8x8_halfpel_hv_add_mmx; |
356 |
|
|
357 |
interpolate8x8_6tap_lowpass_h = interpolate8x8_6tap_lowpass_h_mmx; |
interpolate8x8_6tap_lowpass_h = interpolate8x8_6tap_lowpass_h_mmx; |
358 |
interpolate8x8_6tap_lowpass_v = interpolate8x8_6tap_lowpass_v_mmx; |
interpolate8x8_6tap_lowpass_v = interpolate8x8_6tap_lowpass_v_mmx; |
359 |
|
|
363 |
/* postprocessing */ |
/* postprocessing */ |
364 |
image_brightness = image_brightness_mmx; |
image_brightness = image_brightness_mmx; |
365 |
|
|
|
/* reduced resolution */ |
|
|
copy_upsampled_8x8_16to8 = xvid_Copy_Upsampled_8x8_16To8_mmx; |
|
|
add_upsampled_8x8_16to8 = xvid_Add_Upsampled_8x8_16To8_mmx; |
|
|
hfilter_31 = xvid_HFilter_31_mmx; |
|
|
filter_18x18_to_8x8 = xvid_Filter_18x18_To_8x8_mmx; |
|
|
filter_diff_18x18_to_8x8 = xvid_Filter_Diff_18x18_To_8x8_mmx; |
|
|
|
|
366 |
/* image input xxx_to_yv12 related functions */ |
/* image input xxx_to_yv12 related functions */ |
367 |
yv12_to_yv12 = yv12_to_yv12_mmx; |
yv12_to_yv12 = yv12_to_yv12_mmx; |
368 |
bgr_to_yv12 = bgr_to_yv12_mmx; |
bgr_to_yv12 = bgr_to_yv12_mmx; |
388 |
dev16 = dev16_mmx; |
dev16 = dev16_mmx; |
389 |
sad16v = sad16v_mmx; |
sad16v = sad16v_mmx; |
390 |
sse8_16bit = sse8_16bit_mmx; |
sse8_16bit = sse8_16bit_mmx; |
391 |
|
sse8_8bit = sse8_8bit_mmx; |
392 |
} |
} |
393 |
|
|
394 |
/* these 3dnow functions are faster than mmx, but slower than xmm. */ |
/* these 3dnow functions are faster than mmx, but slower than xmm. */ |
416 |
interpolate8x8_halfpel_v = interpolate8x8_halfpel_v_xmm; |
interpolate8x8_halfpel_v = interpolate8x8_halfpel_v_xmm; |
417 |
interpolate8x8_halfpel_hv = interpolate8x8_halfpel_hv_xmm; |
interpolate8x8_halfpel_hv = interpolate8x8_halfpel_hv_xmm; |
418 |
|
|
419 |
/* reduced resolution */ |
interpolate8x8_halfpel_add = interpolate8x8_halfpel_add_xmm; |
420 |
copy_upsampled_8x8_16to8 = xvid_Copy_Upsampled_8x8_16To8_xmm; |
interpolate8x8_halfpel_h_add = interpolate8x8_halfpel_h_add_xmm; |
421 |
add_upsampled_8x8_16to8 = xvid_Add_Upsampled_8x8_16To8_xmm; |
interpolate8x8_halfpel_v_add = interpolate8x8_halfpel_v_add_xmm; |
422 |
|
interpolate8x8_halfpel_hv_add = interpolate8x8_halfpel_hv_add_xmm; |
423 |
|
|
424 |
/* Quantization */ |
/* Quantization */ |
425 |
quant_mpeg_intra = quant_mpeg_intra_xmm; |
quant_mpeg_intra = quant_mpeg_intra_xmm; |
430 |
|
|
431 |
/* Buffer transfer */ |
/* Buffer transfer */ |
432 |
transfer_8to16sub2 = transfer_8to16sub2_xmm; |
transfer_8to16sub2 = transfer_8to16sub2_xmm; |
433 |
|
transfer_8to16sub2ro = transfer_8to16sub2ro_xmm; |
434 |
|
|
435 |
/* Colorspace transformation */ |
/* Colorspace transformation */ |
436 |
yv12_to_yv12 = yv12_to_yv12_xmm; |
yv12_to_yv12 = yv12_to_yv12_xmm; |
456 |
|
|
457 |
if ((cpu_flags & XVID_CPU_3DNOWEXT)) { |
if ((cpu_flags & XVID_CPU_3DNOWEXT)) { |
458 |
|
|
|
/* Inverse DCT */ |
|
|
idct = idct_3dne; |
|
|
|
|
459 |
/* Buffer transfer */ |
/* Buffer transfer */ |
460 |
transfer_8to16copy = transfer_8to16copy_3dne; |
transfer_8to16copy = transfer_8to16copy_3dne; |
461 |
transfer_16to8copy = transfer_16to8copy_3dne; |
transfer_16to8copy = transfer_16to8copy_3dne; |
462 |
transfer_8to16sub = transfer_8to16sub_3dne; |
transfer_8to16sub = transfer_8to16sub_3dne; |
463 |
transfer_8to16subro = transfer_8to16subro_3dne; |
transfer_8to16subro = transfer_8to16subro_3dne; |
|
transfer_8to16sub2 = transfer_8to16sub2_3dne; |
|
464 |
transfer_16to8add = transfer_16to8add_3dne; |
transfer_16to8add = transfer_16to8add_3dne; |
465 |
transfer8x8_copy = transfer8x8_copy_3dne; |
transfer8x8_copy = transfer8x8_copy_3dne; |
466 |
|
|
467 |
|
if ((cpu_flags & XVID_CPU_MMXEXT)) { |
468 |
|
/* Inverse DCT */ |
469 |
|
idct = idct_3dne; |
470 |
|
|
471 |
|
/* Buffer transfer */ |
472 |
|
transfer_8to16sub2 = transfer_8to16sub2_3dne; |
473 |
|
|
474 |
|
/* Interpolation */ |
475 |
|
interpolate8x8_halfpel_h = interpolate8x8_halfpel_h_3dne; |
476 |
|
interpolate8x8_halfpel_v = interpolate8x8_halfpel_v_3dne; |
477 |
|
interpolate8x8_halfpel_hv = interpolate8x8_halfpel_hv_3dne; |
478 |
|
|
479 |
/* Quantization */ |
/* Quantization */ |
480 |
quant_h263_intra = quant_h263_intra_3dne; |
quant_h263_intra = quant_h263_intra_3dne; /* cmov only */ |
481 |
quant_h263_inter = quant_h263_inter_3dne; |
quant_h263_inter = quant_h263_inter_3dne; |
482 |
dequant_mpeg_intra = dequant_mpeg_intra_3dne; |
dequant_mpeg_intra = dequant_mpeg_intra_3dne; /* cmov only */ |
483 |
dequant_mpeg_inter = dequant_mpeg_inter_3dne; |
dequant_mpeg_inter = dequant_mpeg_inter_3dne; |
484 |
dequant_h263_intra = dequant_h263_intra_3dne; |
dequant_h263_intra = dequant_h263_intra_3dne; |
485 |
dequant_h263_inter = dequant_h263_inter_3dne; |
dequant_h263_inter = dequant_h263_inter_3dne; |
486 |
|
|
487 |
/* ME functions */ |
/* ME functions */ |
488 |
calc_cbp = calc_cbp_3dne; |
calc_cbp = calc_cbp_3dne; |
489 |
|
|
490 |
sad16 = sad16_3dne; |
sad16 = sad16_3dne; |
491 |
sad8 = sad8_3dne; |
sad8 = sad8_3dne; |
492 |
sad16bi = sad16bi_3dne; |
sad16bi = sad16bi_3dne; |
493 |
sad8bi = sad8bi_3dne; |
sad8bi = sad8bi_3dne; |
494 |
dev16 = dev16_3dne; |
dev16 = dev16_3dne; |
495 |
|
} |
|
/* Interpolation */ |
|
|
interpolate8x8_halfpel_h = interpolate8x8_halfpel_h_3dne; |
|
|
interpolate8x8_halfpel_v = interpolate8x8_halfpel_v_3dne; |
|
|
interpolate8x8_halfpel_hv = interpolate8x8_halfpel_hv_3dne; |
|
496 |
} |
} |
497 |
|
|
498 |
if ((cpu_flags & XVID_CPU_SSE2)) { |
if ((cpu_flags & XVID_CPU_SSE2)) { |
512 |
/* DCT operators |
/* DCT operators |
513 |
* no iDCT because it's not "Walken matching" */ |
* no iDCT because it's not "Walken matching" */ |
514 |
fdct = fdct_sse2_skal; |
fdct = fdct_sse2_skal; |
515 |
|
|
516 |
|
/* postprocessing */ |
517 |
|
image_brightness = image_brightness_sse2; |
518 |
} |
} |
519 |
#endif /* ARCH_IS_IA32 */ |
#endif /* ARCH_IS_IA32 */ |
520 |
|
|
545 |
#endif |
#endif |
546 |
|
|
547 |
#if defined(ARCH_IS_PPC) |
#if defined(ARCH_IS_PPC) |
548 |
if ((cpu_flags & XVID_CPU_ASM)) |
if ((cpu_flags & XVID_CPU_ALTIVEC)) { |
549 |
{ |
/* sad operators */ |
550 |
calc_cbp = calc_cbp_ppc; |
sad16 = sad16_altivec_c; |
551 |
} |
sad16bi = sad16bi_altivec_c; |
552 |
|
sad8 = sad8_altivec_c; |
553 |
|
dev16 = dev16_altivec_c; |
554 |
|
|
555 |
|
sse8_16bit = sse8_16bit_altivec_c; |
556 |
|
|
557 |
|
/* mem transfer */ |
558 |
|
transfer_8to16copy = transfer_8to16copy_altivec_c; |
559 |
|
transfer_16to8copy = transfer_16to8copy_altivec_c; |
560 |
|
transfer_8to16sub = transfer_8to16sub_altivec_c; |
561 |
|
transfer_8to16subro = transfer_8to16subro_altivec_c; |
562 |
|
transfer_8to16sub2 = transfer_8to16sub2_altivec_c; |
563 |
|
transfer_16to8add = transfer_16to8add_altivec_c; |
564 |
|
transfer8x8_copy = transfer8x8_copy_altivec_c; |
565 |
|
|
566 |
if ((cpu_flags & XVID_CPU_ALTIVEC)) |
/* Inverse DCT */ |
567 |
{ |
idct = idct_altivec_c; |
568 |
calc_cbp = calc_cbp_altivec; |
|
569 |
fdct = fdct_altivec; |
/* Interpolation */ |
570 |
idct = idct_altivec; |
interpolate8x8_halfpel_h = interpolate8x8_halfpel_h_altivec_c; |
571 |
sadInit = sadInit_altivec; |
interpolate8x8_halfpel_v = interpolate8x8_halfpel_v_altivec_c; |
572 |
sad16 = sad16_altivec; |
interpolate8x8_halfpel_hv = interpolate8x8_halfpel_hv_altivec_c; |
573 |
sad8 = sad8_altivec; |
|
574 |
dev16 = dev16_altivec; |
interpolate8x8_avg2 = interpolate8x8_avg2_altivec_c; |
575 |
|
interpolate8x8_avg4 = interpolate8x8_avg4_altivec_c; |
576 |
|
|
577 |
|
interpolate8x8_halfpel_add = interpolate8x8_halfpel_add_altivec_c; |
578 |
|
interpolate8x8_halfpel_h_add = interpolate8x8_halfpel_h_add_altivec_c; |
579 |
|
interpolate8x8_halfpel_v_add = interpolate8x8_halfpel_v_add_altivec_c; |
580 |
|
interpolate8x8_halfpel_hv_add = interpolate8x8_halfpel_hv_add_altivec_c; |
581 |
|
|
582 |
|
/* Colorspace conversion */ |
583 |
|
bgra_to_yv12 = bgra_to_yv12_altivec_c; |
584 |
|
abgr_to_yv12 = abgr_to_yv12_altivec_c; |
585 |
|
rgba_to_yv12 = rgba_to_yv12_altivec_c; |
586 |
|
argb_to_yv12 = argb_to_yv12_altivec_c; |
587 |
|
|
588 |
|
yuyv_to_yv12 = yuyv_to_yv12_altivec_c; |
589 |
|
uyvy_to_yv12 = uyvy_to_yv12_altivec_c; |
590 |
|
|
591 |
|
yv12_to_yuyv = yv12_to_yuyv_altivec_c; |
592 |
|
yv12_to_uyvy = yv12_to_uyvy_altivec_c; |
593 |
|
|
594 |
|
/* Quantization */ |
595 |
|
quant_h263_intra = quant_h263_intra_altivec_c; |
596 |
|
quant_h263_inter = quant_h263_inter_altivec_c; |
597 |
|
dequant_h263_intra = dequant_h263_intra_altivec_c; |
598 |
|
dequant_h263_inter = dequant_h263_inter_altivec_c; |
599 |
|
|
600 |
|
/* Qpel stuff */ |
601 |
|
xvid_QP_Funcs = &xvid_QP_Funcs_Altivec_C; |
602 |
|
xvid_QP_Add_Funcs = &xvid_QP_Add_Funcs_Altivec_C; |
603 |
} |
} |
604 |
#endif |
#endif |
605 |
|
|
607 |
xvid_debug = init->debug; |
xvid_debug = init->debug; |
608 |
#endif |
#endif |
609 |
|
|
610 |
return 0; |
return(0); |
611 |
} |
} |
612 |
|
|
613 |
|
|
618 |
return XVID_ERR_VERSION; |
return XVID_ERR_VERSION; |
619 |
|
|
620 |
info->actual_version = XVID_VERSION; |
info->actual_version = XVID_VERSION; |
621 |
info->build = "xvid-1.0.0"; |
info->build = "xvid-1.1-cvshead"; |
622 |
info->cpu_flags = detect_cpu_flags(); |
info->cpu_flags = detect_cpu_flags(); |
623 |
|
|
624 |
#if defined(_SMP) && defined(WIN32) |
#if defined(_SMP) && defined(WIN32) |