• Home
  • Raw
  • Download

Lines Matching refs:c

71 static void init_amd_k5(struct cpuinfo_x86 *c)  in init_amd_k5()  argument
83 if (c->x86_model == 9 || c->x86_model == 10) { in init_amd_k5()
90 static void init_amd_k6(struct cpuinfo_x86 *c) in init_amd_k6() argument
96 if (c->x86_model < 6) { in init_amd_k6()
98 if (c->x86_model == 0) { in init_amd_k6()
99 clear_cpu_cap(c, X86_FEATURE_APIC); in init_amd_k6()
100 set_cpu_cap(c, X86_FEATURE_PGE); in init_amd_k6()
105 if (c->x86_model == 6 && c->x86_mask == 1) { in init_amd_k6()
134 if (c->x86_model < 8 || in init_amd_k6()
135 (c->x86_model == 8 && c->x86_mask < 8)) { in init_amd_k6()
154 if ((c->x86_model == 8 && c->x86_mask > 7) || in init_amd_k6()
155 c->x86_model == 9 || c->x86_model == 13) { in init_amd_k6()
176 if (c->x86_model == 10) { in init_amd_k6()
184 static void init_amd_k7(struct cpuinfo_x86 *c) in init_amd_k7() argument
194 if (c->x86_model >= 6 && c->x86_model <= 10) { in init_amd_k7()
195 if (!cpu_has(c, X86_FEATURE_XMM)) { in init_amd_k7()
198 set_cpu_cap(c, X86_FEATURE_XMM); in init_amd_k7()
207 if ((c->x86_model == 8 && c->x86_mask >= 1) || (c->x86_model > 8)) { in init_amd_k7()
217 set_cpu_cap(c, X86_FEATURE_K7); in init_amd_k7()
220 if (!c->cpu_index) in init_amd_k7()
228 if ((c->x86_model == 6) && ((c->x86_mask == 0) || in init_amd_k7()
229 (c->x86_mask == 1))) in init_amd_k7()
233 if ((c->x86_model == 7) && (c->x86_mask == 0)) in init_amd_k7()
243 if (((c->x86_model == 6) && (c->x86_mask >= 2)) || in init_amd_k7()
244 ((c->x86_model == 7) && (c->x86_mask >= 1)) || in init_amd_k7()
245 (c->x86_model > 7)) in init_amd_k7()
246 if (cpu_has(c, X86_FEATURE_MP)) in init_amd_k7()
291 static void amd_get_topology(struct cpuinfo_x86 *c) in amd_get_topology() argument
307 c->compute_unit_id = ebx & 0xff; in amd_get_topology()
309 } else if (cpu_has(c, X86_FEATURE_NODEID_MSR)) { in amd_get_topology()
323 set_cpu_cap(c, X86_FEATURE_AMD_DCM); in amd_get_topology()
324 cores_per_node = c->x86_max_cores / nodes; in amd_get_topology()
331 c->cpu_core_id %= cores_per_node; in amd_get_topology()
332 c->compute_unit_id %= cus_per_node; in amd_get_topology()
341 static void amd_detect_cmp(struct cpuinfo_x86 *c) in amd_detect_cmp() argument
347 bits = c->x86_coreid_bits; in amd_detect_cmp()
349 c->cpu_core_id = c->initial_apicid & ((1 << bits)-1); in amd_detect_cmp()
351 c->phys_proc_id = c->initial_apicid >> bits; in amd_detect_cmp()
353 per_cpu(cpu_llc_id, cpu) = c->phys_proc_id; in amd_detect_cmp()
354 amd_get_topology(c); in amd_detect_cmp()
368 static void srat_detect_node(struct cpuinfo_x86 *c) in srat_detect_node() argument
373 unsigned apicid = c->apicid; in srat_detect_node()
385 x86_cpuinit.fixup_cpu_id(c, node); in srat_detect_node()
407 int ht_nodeid = c->initial_apicid; in srat_detect_node()
420 static void early_init_amd_mc(struct cpuinfo_x86 *c) in early_init_amd_mc() argument
426 if (c->extended_cpuid_level < 0x80000008) in early_init_amd_mc()
431 c->x86_max_cores = (ecx & 0xff) + 1; in early_init_amd_mc()
438 while ((1 << bits) < c->x86_max_cores) in early_init_amd_mc()
442 c->x86_coreid_bits = bits; in early_init_amd_mc()
446 static void bsp_init_amd(struct cpuinfo_x86 *c) in bsp_init_amd() argument
450 if (c->x86 >= 0xf) { in bsp_init_amd()
468 if (cpu_has(c, X86_FEATURE_CONSTANT_TSC)) { in bsp_init_amd()
470 if (c->x86 > 0x10 || in bsp_init_amd()
471 (c->x86 == 0x10 && c->x86_model >= 0x2)) { in bsp_init_amd()
481 if (c->x86 == 0x15) { in bsp_init_amd()
494 static void early_init_amd(struct cpuinfo_x86 *c) in early_init_amd() argument
496 early_init_amd_mc(c); in early_init_amd()
502 if (c->x86_power & (1 << 8)) { in early_init_amd()
503 set_cpu_cap(c, X86_FEATURE_CONSTANT_TSC); in early_init_amd()
504 set_cpu_cap(c, X86_FEATURE_NONSTOP_TSC); in early_init_amd()
510 set_cpu_cap(c, X86_FEATURE_SYSCALL32); in early_init_amd()
513 if (c->x86 == 5) in early_init_amd()
514 if (c->x86_model == 13 || c->x86_model == 9 || in early_init_amd()
515 (c->x86_model == 8 && c->x86_mask >= 8)) in early_init_amd()
516 set_cpu_cap(c, X86_FEATURE_K6_MTRR); in early_init_amd()
520 if (cpu_has_apic && c->x86 >= 0xf) { in early_init_amd()
524 set_cpu_cap(c, X86_FEATURE_EXTD_APICID); in early_init_amd()
533 set_cpu_cap(c, X86_FEATURE_VMMCALL); in early_init_amd()
536 if (c->x86 == 0x16 && c->x86_model <= 0xf) in early_init_amd()
544 static void init_amd_k8(struct cpuinfo_x86 *c) in init_amd_k8() argument
552 set_cpu_cap(c, X86_FEATURE_REP_GOOD); in init_amd_k8()
559 if (c->x86_model < 0x14 && cpu_has(c, X86_FEATURE_LAHF_LM)) { in init_amd_k8()
560 clear_cpu_cap(c, X86_FEATURE_LAHF_LM); in init_amd_k8()
567 if (!c->x86_model_id[0]) in init_amd_k8()
568 strcpy(c->x86_model_id, "Hammer"); in init_amd_k8()
571 static void init_amd_gh(struct cpuinfo_x86 *c) in init_amd_gh() argument
575 if (c == &boot_cpu_data) in init_amd_gh()
602 if (cpu_has_amd_erratum(c, amd_erratum_383)) in init_amd_gh()
603 set_cpu_bug(c, X86_BUG_AMD_TLB_MMATCH); in init_amd_gh()
608 static void init_amd_ln(struct cpuinfo_x86 *c) in init_amd_ln() argument
617 static void init_amd_bd(struct cpuinfo_x86 *c) in init_amd_bd() argument
622 if ((c->x86_model >= 0x10) && (c->x86_model <= 0x1f) && in init_amd_bd()
623 !cpu_has(c, X86_FEATURE_TOPOEXT)) { in init_amd_bd()
628 set_cpu_cap(c, X86_FEATURE_TOPOEXT); in init_amd_bd()
638 if ((c->x86_model >= 0x02) && (c->x86_model < 0x20)) { in init_amd_bd()
646 static void init_amd(struct cpuinfo_x86 *c) in init_amd() argument
658 if (c->x86 == 0xf) in init_amd()
662 early_init_amd(c); in init_amd()
668 clear_cpu_cap(c, 0*32+31); in init_amd()
670 if (c->x86 >= 0x10) in init_amd()
671 set_cpu_cap(c, X86_FEATURE_REP_GOOD); in init_amd()
674 c->apicid = hard_smp_processor_id(); in init_amd()
677 if (c->x86 < 6) in init_amd()
678 clear_cpu_cap(c, X86_FEATURE_MCE); in init_amd()
680 switch (c->x86) { in init_amd()
681 case 4: init_amd_k5(c); break; in init_amd()
682 case 5: init_amd_k6(c); break; in init_amd()
683 case 6: init_amd_k7(c); break; in init_amd()
684 case 0xf: init_amd_k8(c); break; in init_amd()
685 case 0x10: init_amd_gh(c); break; in init_amd()
686 case 0x12: init_amd_ln(c); break; in init_amd()
687 case 0x15: init_amd_bd(c); break; in init_amd()
691 if (c->x86 >= 6) in init_amd()
692 set_cpu_bug(c, X86_BUG_FXSAVE_LEAK); in init_amd()
694 cpu_detect_cache_sizes(c); in init_amd()
697 if (c->extended_cpuid_level >= 0x80000008) { in init_amd()
698 amd_detect_cmp(c); in init_amd()
699 srat_detect_node(c); in init_amd()
703 detect_ht(c); in init_amd()
706 init_amd_cacheinfo(c); in init_amd()
708 if (c->x86 >= 0xf) in init_amd()
709 set_cpu_cap(c, X86_FEATURE_K8); in init_amd()
713 set_cpu_cap(c, X86_FEATURE_MFENCE_RDTSC); in init_amd()
720 if (c->x86 > 0x11) in init_amd()
721 set_cpu_cap(c, X86_FEATURE_ARAT); in init_amd()
723 if (cpu_has_amd_erratum(c, amd_erratum_400)) in init_amd()
724 set_cpu_bug(c, X86_BUG_AMD_APIC_C1E); in init_amd()
726 rdmsr_safe(MSR_AMD64_PATCH_LEVEL, &c->microcode, &dummy); in init_amd()
730 static unsigned int amd_size_cache(struct cpuinfo_x86 *c, unsigned int size) in amd_size_cache() argument
733 if ((c->x86 == 6)) { in amd_size_cache()
735 if (c->x86_model == 3 && c->x86_mask == 0) in amd_size_cache()
738 if (c->x86_model == 4 && in amd_size_cache()
739 (c->x86_mask == 0 || c->x86_mask == 1)) in amd_size_cache()
746 static void cpu_detect_tlb_amd(struct cpuinfo_x86 *c) in cpu_detect_tlb_amd() argument
751 if (c->x86 < 0xf) in cpu_detect_tlb_amd()
754 if (c->extended_cpuid_level < 0x80000006) in cpu_detect_tlb_amd()
766 if (c->x86 == 0xf) { in cpu_detect_tlb_amd()
783 if (c->x86 == 0x15 && c->x86_model <= 0x1f) { in cpu_detect_tlb_amd()