33 #define cpuid(index, eax, ebx, ecx, edx) \
34 ff_cpu_cpuid(index, &eax, &ebx, &ecx, &edx)
36 #define xgetbv(index, eax, edx) \
37 ff_cpu_xgetbv(index, &eax, &edx)
42 #define cpuid(index, eax, ebx, ecx, edx) \
44 "mov %%"FF_REG_b", %%"FF_REG_S" \n\t" \
46 "xchg %%"FF_REG_b", %%"FF_REG_S \
47 : "=a" (eax), "=S" (ebx), "=c" (ecx), "=d" (edx) \
48 : "0" (index), "2"(0))
50 #define xgetbv(index, eax, edx) \
51 __asm__ (".byte 0x0f, 0x01, 0xd0" : "=a"(eax), "=d"(edx) : "c" (index))
53 #define get_eflags(x) \
54 __asm__ volatile ("pushfl \n" \
58 #define set_eflags(x) \
59 __asm__ volatile ("push %0 \n" \
67 #define cpuid_test() 1
71 #define cpuid_test ff_cpu_cpuid_test
75 static int cpuid_test(
void)
82 set_eflags(a ^ 0x200000);
96 int eax, ebx, ecx, edx;
97 int max_std_level, max_ext_level, std_caps = 0, ext_caps = 0;
98 int family = 0, model = 0;
99 union {
int i[3];
char c[12]; } vendor;
104 cpuid(0, max_std_level, vendor.i[0], vendor.i[2], vendor.i[1]);
106 if (max_std_level >= 1) {
107 cpuid(1, eax, ebx, ecx, std_caps);
108 family = ((eax >> 8) & 0xf) + ((eax >> 20) & 0xff);
109 model = ((eax >> 4) & 0xf) + ((eax >> 12) & 0xf0);
110 if (std_caps & (1 << 15))
112 if (std_caps & (1 << 23))
114 if (std_caps & (1 << 25))
117 if (std_caps & (1 << 25))
119 if (std_caps & (1 << 26))
123 if (ecx & 0x00000200 )
125 if (ecx & 0x00080000 )
127 if (ecx & 0x00100000 )
129 if (ecx & 0x01000000 )
133 if ((ecx & 0x18000000) == 0x18000000) {
136 if ((eax & 0x6) == 0x6) {
138 if (ecx & 0x00001000)
145 if (max_std_level >= 7) {
146 cpuid(7, eax, ebx, ecx, edx);
152 if (ebx & 0x00000008) {
154 if (ebx & 0x00000100)
159 cpuid(0x80000000, max_ext_level, ebx, ecx, edx);
161 if (max_ext_level >= 0x80000001) {
162 cpuid(0x80000001, eax, ebx, ecx, ext_caps);
163 if (ext_caps & (1
U << 31))
165 if (ext_caps & (1 << 30))
167 if (ext_caps & (1 << 23))
169 if (ext_caps & (1 << 22))
172 if (!strncmp(vendor.c,
"AuthenticAMD", 12)) {
196 if (ecx & 0x00000800)
198 if (ecx & 0x00010000)
203 if (!strncmp(vendor.c,
"GenuineIntel", 12)) {
204 if (family == 6 && (model == 9 || model == 13 || model == 14)) {
222 if (family == 6 && model == 28)
#define AV_CPU_FLAG_AVX
AVX functions: requires OS support even if YMM registers aren't used.
#define AV_CPU_FLAG_SSE
SSE functions.
#define AV_CPU_FLAG_CMOV
supports cmov instruction
#define AV_CPU_FLAG_FMA3
Haswell FMA3 functions.
#define AV_CPU_FLAG_MMXEXT
SSE integer functions or AMD MMX ext.
#define AV_CPU_FLAG_ATOM
Atom processor, some SSSE3 instructions are slower.
#define AV_CPU_FLAG_AVX2
AVX2 functions: requires OS support even if YMM registers aren't used.
#define AV_CPU_FLAG_SSE2SLOW
SSE2 supported, but usually not faster.
#define AV_CPU_FLAG_XOP
Bulldozer XOP functions.
#define AV_CPU_FLAG_SSE42
Nehalem SSE4.2 functions.
#define AV_CPU_FLAG_SSSE3
Conroe SSSE3 functions.
#define AV_CPU_FLAG_SSE3
Prescott SSE3 functions.
int ff_get_cpu_flags_x86(void)
#define AV_CPU_FLAG_BMI2
Bit Manipulation Instruction Set 2.
#define AV_CPU_FLAG_3DNOW
AMD 3DNOW.
#define AV_CPU_FLAG_SSE3SLOW
SSE3 supported, but usually not faster.
#define AV_CPU_FLAG_AVXSLOW
AVX supported, but slow when using YMM registers (e.g. Bulldozer)
#define AV_CPU_FLAG_BMI1
Bit Manipulation Instruction Set 1.
#define AV_CPU_FLAG_MMX
standard MMX
#define AV_CPU_FLAG_FMA4
Bulldozer FMA4 functions.
#define AV_CPU_FLAG_SSE4
Penryn SSE4.1 functions.
#define AV_CPU_FLAG_AESNI
Advanced Encryption Standard functions.
#define AV_CPU_FLAG_3DNOWEXT
AMD 3DNowExt.
#define AV_CPU_FLAG_SSE2
PIV SSE2 functions.