1 /* 2 * Portions of this file taken from the Linux kernel, 3 * Copyright 1991-2009 Linus Torvalds and contributors 4 * 5 * This program is free software; you can redistribute it and/or modify 6 * it under the terms of the GNU General Public License as published by 7 * the Free Software Foundation; either version 2 of the License, or 8 * (at your option) any later version. 9 * This program is distributed in the hope that it will be useful, 10 * but WITHOUT ANY WARRANTY; without even the implied warranty of 11 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 12 * GNU General Public License for more details. 13 * 14 * You should have received a copy of the GNU General Public License 15 * along with this program; if not, write to the Free Software 16 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA 17 */ 18 19 #include <stdio.h> 20 #include <string.h> 21 #include "cpuid.h" 22 23 const char *cpu_flags_names[] = { 24 CPU_FLAGS(STRUCT_MEMBER_NAMES) 25 }; 26 27 size_t cpu_flags_offset[] = { 28 CPU_FLAGS(STRUCTURE_MEMBER_OFFSETS) 29 }; 30 31 size_t cpu_flags_count = sizeof cpu_flags_names / sizeof *cpu_flags_names; 32 33 struct cpu_dev *cpu_devs[X86_VENDOR_NUM] = { }; 34 35 bool get_cpu_flag_value_from_name(s_cpu *cpu, const char * flag_name) { 36 size_t i; 37 bool cpu_flag_present=false, *flag_value = &cpu_flag_present; 38 39 for (i = 0; i < cpu_flags_count; i++) { 40 if (strcmp(cpu_flags_names[i],flag_name) == 0) { 41 flag_value = (bool *)((char *)&cpu->flags + cpu_flags_offset[i]); 42 } 43 } 44 return *flag_value; 45 } 46 47 48 /* 49 * CPUID functions returning a single datum 50 */ 51 52 /* Probe for the CPUID instruction */ 53 static int have_cpuid_p(void) 54 { 55 return cpu_has_eflag(X86_EFLAGS_ID); 56 } 57 58 static struct cpu_dev amd_cpu_dev = { 59 .c_vendor = "AMD", 60 .c_ident = {"AuthenticAMD"} 61 }; 62 63 static struct cpu_dev intel_cpu_dev = { 64 .c_vendor = "Intel", 65 .c_ident = {"GenuineIntel"} 66 }; 67 68 static struct cpu_dev cyrix_cpu_dev = { 69 .c_vendor = "Cyrix", 70 .c_ident = {"CyrixInstead"} 71 }; 72 73 static struct cpu_dev umc_cpu_dev = { 74 .c_vendor = "UMC", 75 .c_ident = {"UMC UMC UMC"} 76 77 }; 78 79 static struct cpu_dev nexgen_cpu_dev = { 80 .c_vendor = "Nexgen", 81 .c_ident = {"NexGenDriven"} 82 }; 83 84 static struct cpu_dev centaur_cpu_dev = { 85 .c_vendor = "Centaur", 86 .c_ident = {"CentaurHauls"} 87 }; 88 89 static struct cpu_dev rise_cpu_dev = { 90 .c_vendor = "Rise", 91 .c_ident = {"RiseRiseRise"} 92 }; 93 94 static struct cpu_dev transmeta_cpu_dev = { 95 .c_vendor = "Transmeta", 96 .c_ident = {"GenuineTMx86", "TransmetaCPU"} 97 }; 98 99 static struct cpu_dev nsc_cpu_dev = { 100 .c_vendor = "National Semiconductor", 101 .c_ident = {"Geode by NSC"} 102 }; 103 104 static struct cpu_dev unknown_cpu_dev = { 105 .c_vendor = "Unknown Vendor", 106 .c_ident = {"Unknown CPU"} 107 }; 108 109 /* 110 * Read NSC/Cyrix DEVID registers (DIR) to get more detailed info. about the CPU 111 */ 112 void do_cyrix_devid(unsigned char *dir0, unsigned char *dir1) 113 { 114 unsigned char ccr2, ccr3; 115 116 /* we test for DEVID by checking whether CCR3 is writable */ 117 ccr3 = getCx86(CX86_CCR3); 118 setCx86(CX86_CCR3, ccr3 ^ 0x80); 119 getCx86(0xc0); /* dummy to change bus */ 120 121 if (getCx86(CX86_CCR3) == ccr3) { /* no DEVID regs. */ 122 ccr2 = getCx86(CX86_CCR2); 123 setCx86(CX86_CCR2, ccr2 ^ 0x04); 124 getCx86(0xc0); /* dummy */ 125 126 if (getCx86(CX86_CCR2) == ccr2) /* old Cx486SLC/DLC */ 127 *dir0 = 0xfd; 128 else { /* Cx486S A step */ 129 setCx86(CX86_CCR2, ccr2); 130 *dir0 = 0xfe; 131 } 132 } else { 133 setCx86(CX86_CCR3, ccr3); /* restore CCR3 */ 134 135 /* read DIR0 and DIR1 CPU registers */ 136 *dir0 = getCx86(CX86_DIR0); 137 *dir1 = getCx86(CX86_DIR1); 138 } 139 } 140 141 void init_cpu_devs(void) 142 { 143 cpu_devs[X86_VENDOR_INTEL] = &intel_cpu_dev; 144 cpu_devs[X86_VENDOR_CYRIX] = &cyrix_cpu_dev; 145 cpu_devs[X86_VENDOR_AMD] = &amd_cpu_dev; 146 cpu_devs[X86_VENDOR_UMC] = &umc_cpu_dev; 147 cpu_devs[X86_VENDOR_NEXGEN] = &nexgen_cpu_dev; 148 cpu_devs[X86_VENDOR_CENTAUR] = ¢aur_cpu_dev; 149 cpu_devs[X86_VENDOR_RISE] = &rise_cpu_dev; 150 cpu_devs[X86_VENDOR_TRANSMETA] = &transmeta_cpu_dev; 151 cpu_devs[X86_VENDOR_NSC] = &nsc_cpu_dev; 152 cpu_devs[X86_VENDOR_UNKNOWN] = &unknown_cpu_dev; 153 } 154 155 void get_cpu_vendor(struct cpuinfo_x86 *c) 156 { 157 char *v = c->x86_vendor_id; 158 int i; 159 init_cpu_devs(); 160 for (i = 0; i < X86_VENDOR_NUM-1; i++) { 161 if (cpu_devs[i]) { 162 if (!strcmp(v, cpu_devs[i]->c_ident[0]) || 163 (cpu_devs[i]->c_ident[1] && 164 !strcmp(v, cpu_devs[i]->c_ident[1]))) { 165 c->x86_vendor = i; 166 return; 167 } 168 } 169 } 170 171 c->x86_vendor = X86_VENDOR_UNKNOWN; 172 } 173 174 int get_model_name(struct cpuinfo_x86 *c) 175 { 176 unsigned int *v; 177 char *p, *q; 178 179 if (cpuid_eax(0x80000000) < 0x80000004) 180 return 0; 181 182 v = (unsigned int *)c->x86_model_id; 183 cpuid(0x80000002, &v[0], &v[1], &v[2], &v[3]); 184 cpuid(0x80000003, &v[4], &v[5], &v[6], &v[7]); 185 cpuid(0x80000004, &v[8], &v[9], &v[10], &v[11]); 186 c->x86_model_id[48] = 0; 187 188 /* Intel chips right-justify this string for some dumb reason; 189 undo that brain damage */ 190 p = q = &c->x86_model_id[0]; 191 while (*p == ' ') 192 p++; 193 if (p != q) { 194 while (*p) 195 *q++ = *p++; 196 while (q <= &c->x86_model_id[48]) 197 *q++ = '\0'; /* Zero-pad the rest */ 198 } 199 200 return 1; 201 } 202 203 void detect_cache(uint32_t xlvl, struct cpuinfo_x86 *c) 204 { 205 uint32_t eax, ebx, ecx, edx, l2size; 206 /* Detecting L1 cache */ 207 if (xlvl >= 0x80000005) { 208 cpuid(0x80000005, &eax, &ebx, &ecx, &edx); 209 c->x86_l1_data_cache_size = ecx >> 24; 210 c->x86_l1_instruction_cache_size = edx >> 24; 211 } 212 213 /* Detecting L2 cache */ 214 c->x86_l2_cache_size = 0; 215 216 if (xlvl < 0x80000006) /* Some chips just has a large L1. */ 217 return; 218 219 cpuid(0x80000006, &eax, &ebx, &ecx, &edx); 220 l2size = ecx >> 16; 221 222 /* Vendor based fixes */ 223 switch (c->x86_vendor) { 224 case X86_VENDOR_INTEL: 225 /* 226 * Intel PIII Tualatin. This comes in two flavours. 227 * One has 256kb of cache, the other 512. We have no way 228 * to determine which, so we use a boottime override 229 * for the 512kb model, and assume 256 otherwise. 230 */ 231 if ((c->x86 == 6) && (c->x86_model == 11) && (l2size == 0)) 232 l2size = 256; 233 break; 234 case X86_VENDOR_AMD: 235 /* AMD errata T13 (order #21922) */ 236 if ((c->x86 == 6)) { 237 if (c->x86_model == 3 && c->x86_mask == 0) /* Duron Rev A0 */ 238 l2size = 64; 239 if (c->x86_model == 4 && (c->x86_mask == 0 || c->x86_mask == 1)) /* Tbird rev A1/A2 */ 240 l2size = 256; 241 } 242 break; 243 } 244 c->x86_l2_cache_size = l2size; 245 } 246 247 void detect_cyrix(struct cpuinfo_x86 *c) { 248 unsigned char dir0, dir0_msn, dir0_lsn, dir1 = 0; 249 char *buf = c->x86_model_id; 250 char Cx86_cb[] = "?.5x Core/Bus Clock"; 251 const char cyrix_model_mult1[] = "12??43"; 252 const char cyrix_model_mult2[] = "12233445"; 253 const char *p = NULL; 254 255 do_cyrix_devid(&dir0, &dir1); 256 dir0_msn = dir0 >> 4; /* identifies CPU "family" */ 257 dir0_lsn = dir0 & 0xf; /* model or clock multiplier */ 258 c->x86_model = (dir1 >> 4) + 1; 259 c->x86_mask = dir1 & 0xf; 260 switch (dir0_msn) { 261 unsigned char tmp; 262 263 case 0: /* Cx486SLC/DLC/SRx/DRx */ 264 p = Cx486_name[dir0_lsn & 7]; 265 break; 266 267 case 1: /* Cx486S/DX/DX2/DX4 */ 268 p = (dir0_lsn & 8) ? Cx486D_name[dir0_lsn & 5] : Cx486S_name[dir0_lsn & 3]; 269 break; 270 271 case 2: /* 5x86 */ 272 Cx86_cb[2] = cyrix_model_mult1[dir0_lsn & 5]; 273 p = Cx86_cb+2; 274 break; 275 276 case 3: /* 6x86/6x86L */ 277 Cx86_cb[1] = ' '; 278 Cx86_cb[2] = cyrix_model_mult1[dir0_lsn & 5]; 279 if (dir1 > 0x21) { /* 686L */ 280 Cx86_cb[0] = 'L'; 281 p = Cx86_cb; 282 (c->x86_model)++; 283 } else /* 686 */ 284 p = Cx86_cb+1; 285 286 c->coma_bug = 1; 287 break; 288 case 4: 289 c->x86_l1_data_cache_size = 16; /* Yep 16K integrated cache thats it */ 290 if (c->cpuid_level != 2) { /* Media GX */ 291 Cx86_cb[2] = (dir0_lsn & 1) ? '3' : '4'; 292 p = Cx86_cb+2; 293 } 294 break; 295 296 case 5: /* 6x86MX/M II */ 297 if (dir1 > 7) { 298 dir0_msn++; /* M II */ 299 } else { 300 c->coma_bug = 1; /* 6x86MX, it has the bug. */ 301 } 302 303 tmp = (!(dir0_lsn & 7) || dir0_lsn & 1) ? 2 : 0; 304 Cx86_cb[tmp] = cyrix_model_mult2[dir0_lsn & 7]; 305 p = Cx86_cb+tmp; 306 if (((dir1 & 0x0f) > 4) || ((dir1 & 0xf0) == 0x20)) 307 (c->x86_model)++; 308 break; 309 310 case 0xf: /* Cyrix 486 without DEVID registers */ 311 switch (dir0_lsn) { 312 case 0xd: /* either a 486SLC or DLC w/o DEVID */ 313 dir0_msn = 0; 314 p = Cx486_name[(c->hard_math) ? 1 : 0]; 315 break; 316 317 case 0xe: /* a 486S A step */ 318 dir0_msn = 0; 319 p = Cx486S_name[0]; 320 break; 321 } 322 break; 323 324 default: 325 dir0_msn = 7; 326 break; 327 } 328 329 /* If the processor is unknown, we keep the model name we got 330 * from the generic call */ 331 if (dir0_msn < 7) { 332 strcpy(buf, Cx86_model[dir0_msn & 7]); 333 if (p) strcat(buf, p); 334 } 335 } 336 337 void generic_identify(struct cpuinfo_x86 *c) 338 { 339 uint32_t tfms, xlvl; 340 uint32_t eax, ebx, ecx, edx; 341 342 /* Get vendor name */ 343 cpuid(0x00000000, 344 (uint32_t *) & c->cpuid_level, 345 (uint32_t *) & c->x86_vendor_id[0], 346 (uint32_t *) & c->x86_vendor_id[8], 347 (uint32_t *) & c->x86_vendor_id[4]); 348 349 get_cpu_vendor(c); 350 351 /* Intel-defined flags: level 0x00000001 */ 352 if (c->cpuid_level >= 0x00000001) { 353 uint32_t capability, excap; 354 cpuid(0x00000001, &tfms, &ebx, &excap, &capability); 355 c->x86_capability[0] = capability; 356 c->x86_capability[4] = excap; 357 c->x86 = (tfms >> 8) & 15; 358 c->x86_model = (tfms >> 4) & 15; 359 if (c->x86 == 0xf) 360 c->x86 += (tfms >> 20) & 0xff; 361 if (c->x86 >= 0x6) 362 c->x86_model += ((tfms >> 16) & 0xF) << 4; 363 c->x86_mask = tfms & 15; 364 if (cpu_has(c, X86_FEATURE_CLFLSH)) 365 c->x86_clflush_size = ((ebx >> 8) & 0xff) * 8; 366 } else { 367 /* Have CPUID level 0 only - unheard of */ 368 c->x86 = 4; 369 } 370 371 /* AMD-defined flags: level 0x80000001 */ 372 xlvl = cpuid_eax(0x80000000); 373 if ((xlvl & 0xffff0000) == 0x80000000) { 374 if (xlvl >= 0x80000001) { 375 c->x86_capability[1] = cpuid_edx(0x80000001); 376 c->x86_capability[6] = cpuid_ecx(0x80000001); 377 } 378 if (xlvl >= 0x80000004) 379 get_model_name(c); /* Default name */ 380 } 381 382 /* Specific detection code */ 383 switch (c->x86_vendor) { 384 case X86_VENDOR_CYRIX: 385 case X86_VENDOR_NSC: detect_cyrix(c); break; 386 default: break; 387 } 388 389 /* Detecting the number of cores */ 390 switch (c->x86_vendor) { 391 case X86_VENDOR_AMD: 392 if (xlvl >= 0x80000008) { 393 c->x86_num_cores = (cpuid_ecx(0x80000008) & 0xff) + 1; 394 if (c->x86_num_cores & (c->x86_num_cores - 1)) 395 c->x86_num_cores = 1; 396 } 397 break; 398 case X86_VENDOR_INTEL: 399 if (c->cpuid_level >= 0x00000004) { 400 cpuid(0x4, &eax, &ebx, &ecx, &edx); 401 c->x86_num_cores = ((eax & 0xfc000000) >> 26) + 1; 402 } 403 break; 404 default: 405 c->x86_num_cores = 1; 406 break; 407 } 408 409 detect_cache(xlvl, c); 410 } 411 412 /* 413 * Checksum an MP configuration block. 414 */ 415 416 static int mpf_checksum(unsigned char *mp, int len) 417 { 418 int sum = 0; 419 420 while (len--) 421 sum += *mp++; 422 423 return sum & 0xFF; 424 } 425 426 static int smp_scan_config(unsigned long base, unsigned long length) 427 { 428 unsigned long *bp = (unsigned long *)base; 429 struct intel_mp_floating *mpf; 430 431 // printf("Scan SMP from %p for %ld bytes.\n", bp,length); 432 if (sizeof(*mpf) != 16) { 433 printf("Error: MPF size\n"); 434 return 0; 435 } 436 437 while (length > 0) { 438 mpf = (struct intel_mp_floating *)bp; 439 if ((*bp == SMP_MAGIC_IDENT) && 440 (mpf->mpf_length == 1) && 441 !mpf_checksum((unsigned char *)bp, 16) && 442 ((mpf->mpf_specification == 1) 443 || (mpf->mpf_specification == 4))) { 444 return 1; 445 } 446 bp += 4; 447 length -= 16; 448 } 449 return 0; 450 } 451 452 int find_smp_config(void) 453 { 454 // unsigned int address; 455 456 /* 457 * FIXME: Linux assumes you have 640K of base ram.. 458 * this continues the error... 459 * 460 * 1) Scan the bottom 1K for a signature 461 * 2) Scan the top 1K of base RAM 462 * 3) Scan the 64K of bios 463 */ 464 if (smp_scan_config(0x0, 0x400) || 465 smp_scan_config(639 * 0x400, 0x400) || 466 smp_scan_config(0xF0000, 0x10000)) 467 return 1; 468 /* 469 * If it is an SMP machine we should know now, unless the 470 * configuration is in an EISA/MCA bus machine with an 471 * extended bios data area. 472 * 473 * there is a real-mode segmented pointer pointing to the 474 * 4K EBDA area at 0x40E, calculate and scan it here. 475 * 476 * NOTE! There are Linux loaders that will corrupt the EBDA 477 * area, and as such this kind of SMP config may be less 478 * trustworthy, simply because the SMP table may have been 479 * stomped on during early boot. These loaders are buggy and 480 * should be fixed. 481 * 482 * MP1.4 SPEC states to only scan first 1K of 4K EBDA. 483 */ 484 485 // address = get_bios_ebda(); 486 // if (address) 487 // smp_scan_config(address, 0x400); 488 return 0; 489 } 490 491 void set_cpu_flags(struct cpuinfo_x86 *c, s_cpu * cpu) 492 { 493 cpu->flags.fpu = cpu_has(c, X86_FEATURE_FPU); 494 cpu->flags.vme = cpu_has(c, X86_FEATURE_VME); 495 cpu->flags.de = cpu_has(c, X86_FEATURE_DE); 496 cpu->flags.pse = cpu_has(c, X86_FEATURE_PSE); 497 cpu->flags.tsc = cpu_has(c, X86_FEATURE_TSC); 498 cpu->flags.msr = cpu_has(c, X86_FEATURE_MSR); 499 cpu->flags.pae = cpu_has(c, X86_FEATURE_PAE); 500 cpu->flags.mce = cpu_has(c, X86_FEATURE_MCE); 501 cpu->flags.cx8 = cpu_has(c, X86_FEATURE_CX8); 502 cpu->flags.apic = cpu_has(c, X86_FEATURE_APIC); 503 cpu->flags.sep = cpu_has(c, X86_FEATURE_SEP); 504 cpu->flags.mtrr = cpu_has(c, X86_FEATURE_MTRR); 505 cpu->flags.pge = cpu_has(c, X86_FEATURE_PGE); 506 cpu->flags.mca = cpu_has(c, X86_FEATURE_MCA); 507 cpu->flags.cmov = cpu_has(c, X86_FEATURE_CMOV); 508 cpu->flags.pat = cpu_has(c, X86_FEATURE_PAT); 509 cpu->flags.pse_36 = cpu_has(c, X86_FEATURE_PSE36); 510 cpu->flags.psn = cpu_has(c, X86_FEATURE_PN); 511 cpu->flags.clflsh = cpu_has(c, X86_FEATURE_CLFLSH); 512 cpu->flags.dts = cpu_has(c, X86_FEATURE_DTES); 513 cpu->flags.acpi = cpu_has(c, X86_FEATURE_ACPI); 514 cpu->flags.pbe = cpu_has(c, X86_FEATURE_PBE); 515 cpu->flags.mmx = cpu_has(c, X86_FEATURE_MMX); 516 cpu->flags.fxsr = cpu_has(c, X86_FEATURE_FXSR); 517 cpu->flags.sse = cpu_has(c, X86_FEATURE_XMM); 518 cpu->flags.sse2 = cpu_has(c, X86_FEATURE_XMM2); 519 cpu->flags.ss = cpu_has(c, X86_FEATURE_SELFSNOOP); 520 cpu->flags.htt = cpu_has(c, X86_FEATURE_HT); 521 cpu->flags.acc = cpu_has(c, X86_FEATURE_ACC); 522 cpu->flags.syscall = cpu_has(c, X86_FEATURE_SYSCALL); 523 cpu->flags.mp = cpu_has(c, X86_FEATURE_MP); 524 cpu->flags.nx = cpu_has(c, X86_FEATURE_NX); 525 cpu->flags.mmxext = cpu_has(c, X86_FEATURE_MMXEXT); 526 cpu->flags.fxsr_opt = cpu_has(c, X86_FEATURE_FXSR_OPT); 527 cpu->flags.gbpages = cpu_has(c, X86_FEATURE_GBPAGES); 528 cpu->flags.rdtscp = cpu_has(c, X86_FEATURE_RDTSCP); 529 cpu->flags.lm = cpu_has(c, X86_FEATURE_LM); 530 cpu->flags.nowext = cpu_has(c, X86_FEATURE_3DNOWEXT); 531 cpu->flags.now = cpu_has(c, X86_FEATURE_3DNOW); 532 cpu->flags.smp = find_smp_config(); 533 cpu->flags.pni = cpu_has(c, X86_FEATURE_XMM3); 534 cpu->flags.pclmulqd = cpu_has(c, X86_FEATURE_PCLMULQDQ); 535 cpu->flags.dtes64 = cpu_has(c, X86_FEATURE_DTES64); 536 cpu->flags.vmx = cpu_has(c, X86_FEATURE_VMX); 537 cpu->flags.smx = cpu_has(c, X86_FEATURE_SMX); 538 cpu->flags.est = cpu_has(c, X86_FEATURE_EST); 539 cpu->flags.tm2 = cpu_has(c, X86_FEATURE_TM2); 540 cpu->flags.sse3 = cpu_has(c, X86_FEATURE_SSE3); 541 cpu->flags.cid = cpu_has(c, X86_FEATURE_CID); 542 cpu->flags.fma = cpu_has(c, X86_FEATURE_FMA); 543 cpu->flags.cx16 = cpu_has(c, X86_FEATURE_CX16); 544 cpu->flags.xtpr = cpu_has(c, X86_FEATURE_XTPR); 545 cpu->flags.pdcm = cpu_has(c, X86_FEATURE_PDCM); 546 cpu->flags.dca = cpu_has(c, X86_FEATURE_DCA); 547 cpu->flags.xmm4_1 = cpu_has(c, X86_FEATURE_XMM4_1); 548 cpu->flags.xmm4_2 = cpu_has(c, X86_FEATURE_XMM4_2); 549 cpu->flags.x2apic = cpu_has(c, X86_FEATURE_X2APIC); 550 cpu->flags.movbe = cpu_has(c, X86_FEATURE_MOVBE); 551 cpu->flags.popcnt = cpu_has(c, X86_FEATURE_POPCNT); 552 cpu->flags.aes = cpu_has(c, X86_FEATURE_AES); 553 cpu->flags.xsave = cpu_has(c, X86_FEATURE_XSAVE); 554 cpu->flags.osxsave = cpu_has(c, X86_FEATURE_OSXSAVE); 555 cpu->flags.avx = cpu_has(c, X86_FEATURE_AVX); 556 cpu->flags.hypervisor = cpu_has(c, X86_FEATURE_HYPERVISOR); 557 cpu->flags.ace2 = cpu_has(c, X86_FEATURE_ACE2); 558 cpu->flags.ace2_en = cpu_has(c, X86_FEATURE_ACE2_EN); 559 cpu->flags.phe = cpu_has(c, X86_FEATURE_PHE); 560 cpu->flags.phe_en = cpu_has(c, X86_FEATURE_PHE_EN); 561 cpu->flags.pmm = cpu_has(c, X86_FEATURE_PMM); 562 cpu->flags.pmm_en = cpu_has(c, X86_FEATURE_PMM_EN); 563 cpu->flags.extapic = cpu_has(c, X86_FEATURE_EXTAPIC); 564 cpu->flags.cr8_legacy = cpu_has(c, X86_FEATURE_CR8_LEGACY); 565 cpu->flags.abm = cpu_has(c, X86_FEATURE_ABM); 566 cpu->flags.sse4a = cpu_has(c, X86_FEATURE_SSE4A); 567 cpu->flags.misalignsse = cpu_has(c, X86_FEATURE_MISALIGNSSE); 568 cpu->flags.nowprefetch = cpu_has(c, X86_FEATURE_3DNOWPREFETCH); 569 cpu->flags.osvw = cpu_has(c, X86_FEATURE_OSVW); 570 cpu->flags.ibs = cpu_has(c, X86_FEATURE_IBS); 571 cpu->flags.sse5 = cpu_has(c, X86_FEATURE_SSE5); 572 cpu->flags.skinit = cpu_has(c, X86_FEATURE_SKINIT); 573 cpu->flags.wdt = cpu_has(c, X86_FEATURE_WDT); 574 cpu->flags.ida = cpu_has(c, X86_FEATURE_IDA); 575 cpu->flags.arat = cpu_has(c, X86_FEATURE_ARAT); 576 cpu->flags.tpr_shadow = cpu_has(c, X86_FEATURE_TPR_SHADOW); 577 cpu->flags.vnmi = cpu_has(c, X86_FEATURE_VNMI); 578 cpu->flags.flexpriority = cpu_has(c, X86_FEATURE_FLEXPRIORITY); 579 cpu->flags.ept = cpu_has(c, X86_FEATURE_EPT); 580 cpu->flags.vpid = cpu_has(c, X86_FEATURE_VPID); 581 cpu->flags.svm = cpu_has(c, X86_FEATURE_SVM); 582 } 583 584 void set_generic_info(struct cpuinfo_x86 *c, s_cpu * cpu) 585 { 586 cpu->family = c->x86; 587 cpu->vendor_id = c->x86_vendor; 588 cpu->model_id = c->x86_model; 589 cpu->stepping = c->x86_mask; 590 strlcpy(cpu->vendor, cpu_devs[c->x86_vendor]->c_vendor, 591 sizeof(cpu->vendor)); 592 strlcpy(cpu->model, c->x86_model_id, sizeof(cpu->model)); 593 cpu->num_cores = c->x86_num_cores; 594 cpu->l1_data_cache_size = c->x86_l1_data_cache_size; 595 cpu->l1_instruction_cache_size = c->x86_l1_instruction_cache_size; 596 cpu->l2_cache_size = c->x86_l2_cache_size; 597 } 598 599 void detect_cpu(s_cpu * cpu) 600 { 601 struct cpuinfo_x86 c; 602 memset(&c,0,sizeof(c)); 603 c.x86_clflush_size = 32; 604 c.x86_vendor = X86_VENDOR_UNKNOWN; 605 c.cpuid_level = -1; /* CPUID not detected */ 606 c.x86_num_cores = 1; 607 memset(&cpu->flags, 0, sizeof(s_cpu_flags)); 608 609 if (!have_cpuid_p()) 610 return; 611 612 generic_identify(&c); 613 set_generic_info(&c, cpu); 614 set_cpu_flags(&c, cpu); 615 } 616