/arch/x86/kernel/cpu/ |
D | amd.c | 95 cpu_has(cpu, X86_FEATURE_OSVW)) { in cpu_has_amd_erratum() 299 if (!cpu_has(c, X86_FEATURE_XMM)) { in init_amd_k7() 347 if (cpu_has(c, X86_FEATURE_MP)) in init_amd_k7() 440 } else if (cpu_has(c, X86_FEATURE_NODEID_MSR)) { in amd_get_topology() 478 if (!cpu_has(c, X86_FEATURE_AMD_PPIN)) in amd_detect_ppin() 614 if (cpu_has(c, X86_FEATURE_CONSTANT_TSC)) { in bsp_init_amd() 641 if (cpu_has(c, X86_FEATURE_MWAITX)) in bsp_init_amd() 697 if (cpu_has(c, X86_FEATURE_SME) || cpu_has(c, X86_FEATURE_SEV)) { in early_detect_mem_encrypt() 812 !cpu_has(c, X86_FEATURE_TOPOEXT)) { in early_init_amd() 823 if (cpu_has(c, X86_FEATURE_TOPOEXT)) in early_init_amd() [all …]
|
D | hygon.c | 56 if (cpu_has(c, X86_FEATURE_TOPOEXT)) in hygon_get_topology_early() 100 } else if (cpu_has(c, X86_FEATURE_NODEID_MSR)) { in hygon_get_topology() 227 if (cpu_has(c, X86_FEATURE_CONSTANT_TSC)) { in bsp_init_hygon() 235 if (cpu_has(c, X86_FEATURE_MWAITX)) in bsp_init_hygon() 342 if (cpu_has(c, X86_FEATURE_XMM2)) { in init_hygon() 362 if (!cpu_has(c, X86_FEATURE_XENPV)) in init_hygon()
|
D | common.c | 265 if (!cpu_has(c, X86_FEATURE_PN) || !disable_x86_serial_nr) in squash_the_stupid_serial_number() 306 if (cpu_has(c, X86_FEATURE_SMEP)) in setup_smep() 324 if (cpu_has(c, X86_FEATURE_SMAP)) { in setup_smap() 341 if (!cpu_has(c, X86_FEATURE_UMIP)) in setup_umip() 479 if (!cpu_has(c, X86_FEATURE_PKU)) in setup_pku() 538 if (!cpu_has(c, df->feature)) in filter_cpuid_features() 757 if (!cpu_has(c, X86_FEATURE_HT)) in detect_ht_early() 760 if (cpu_has(c, X86_FEATURE_CMP_LEGACY)) in detect_ht_early() 763 if (cpu_has(c, X86_FEATURE_XTOPOLOGY)) in detect_ht_early() 866 if (cpu_has(c, X86_FEATURE_SPEC_CTRL)) { in init_speculation_control() [all …]
|
D | intel.c | 167 if (cpu_has(c, X86_FEATURE_HYPERVISOR)) in bad_spectre_microcode() 282 if (c->x86 >= 6 && !cpu_has(c, X86_FEATURE_IA64)) in early_init_intel() 286 if ((cpu_has(c, X86_FEATURE_SPEC_CTRL) || in early_init_intel() 287 cpu_has(c, X86_FEATURE_INTEL_STIBP) || in early_init_intel() 288 cpu_has(c, X86_FEATURE_IBRS) || cpu_has(c, X86_FEATURE_IBPB) || in early_init_intel() 289 cpu_has(c, X86_FEATURE_STIBP)) && bad_spectre_microcode(c)) { in early_init_intel() 421 if (cpu_has(c, X86_FEATURE_TME)) in early_init_intel() 626 if (!cpu_has(c, X86_FEATURE_XTOPOLOGY)) { in init_intel() 646 if (cpu_has(c, X86_FEATURE_XMM2)) in init_intel() 1186 if (!cpu_has(c, X86_FEATURE_CORE_CAPABILITIES)) in cpu_set_core_cap_bits()
|
D | perfctr-watchdog.c | 55 if (cpu_has(&boot_cpu_data, X86_FEATURE_ARCH_PERFMON)) in nmi_perfctr_msr_to_bit() 88 if (cpu_has(&boot_cpu_data, X86_FEATURE_ARCH_PERFMON)) in nmi_evntsel_msr_to_bit()
|
D | proc.c | 86 if (cpu_has(c, X86_FEATURE_TSC)) { in show_cpuinfo() 106 if (cpu_has(c, i) && x86_cap_flags[i] != NULL) in show_cpuinfo() 110 if (cpu_has(c, X86_FEATURE_VMX) && c->vmx_capability[0]) { in show_cpuinfo()
|
D | feat_ctl.c | 120 if (cpu_has(c, X86_FEATURE_VMX) && IS_ENABLED(CONFIG_KVM_INTEL)) { in init_ia32_feat_ctl() 132 if (!cpu_has(c, X86_FEATURE_VMX)) in init_ia32_feat_ctl()
|
D | rdrand.c | 36 if (!cpu_has(c, X86_FEATURE_RDRAND)) in x86_init_rdrand()
|
D | match.c | 54 if (m->feature != X86_FEATURE_ANY && !cpu_has(c, m->feature)) in x86_match_cpu()
|
/arch/x86/include/asm/ |
D | acpi.h | 101 if (cpu_has(c, X86_FEATURE_EST)) in arch_acpi_set_pdc_bits() 104 if (cpu_has(c, X86_FEATURE_ACPI)) in arch_acpi_set_pdc_bits() 110 if (!cpu_has(c, X86_FEATURE_MWAIT)) in arch_acpi_set_pdc_bits()
|
D | cpufeature.h | 124 #define cpu_has(c, bit) \ macro 144 #define boot_cpu_has(bit) cpu_has(&boot_cpu_data, bit) 208 #define cpu_has_bug(c, bit) cpu_has(c, (bit))
|
/arch/x86/kernel/cpu/mce/ |
D | therm_throt.c | 416 if (cpu_has(c, X86_FEATURE_PLN) && int_pln_enable) { in thermal_throttle_add_dev() 424 if (cpu_has(c, X86_FEATURE_PTS)) { in thermal_throttle_add_dev() 443 if (cpu_has(c, X86_FEATURE_PLN) && int_pln_enable) { in thermal_throttle_add_dev() 631 if (!cpu_has(c, X86_FEATURE_ACPI) || !cpu_has(c, X86_FEATURE_ACC)) in intel_thermal_supported() 685 if (cpu_has(c, X86_FEATURE_TM2)) { in intel_init_thermal() 699 if (cpu_has(c, X86_FEATURE_PLN) && !int_pln_enable) in intel_init_thermal() 703 else if (cpu_has(c, X86_FEATURE_PLN) && int_pln_enable) in intel_init_thermal() 711 if (cpu_has(c, X86_FEATURE_PTS)) { in intel_init_thermal() 713 if (cpu_has(c, X86_FEATURE_PLN) && !int_pln_enable) in intel_init_thermal() 718 else if (cpu_has(c, X86_FEATURE_PLN) && int_pln_enable) in intel_init_thermal()
|
D | p5.c | 54 if (!cpu_has(c, X86_FEATURE_MCE)) in intel_p5_mcheck_init()
|
D | core.c | 513 return cpu_has(c, X86_FEATURE_MCE) && cpu_has(c, X86_FEATURE_MCA); in mce_available() 1879 mce_flags.overflow_recov = !!cpu_has(c, X86_FEATURE_OVERFLOW_RECOV); in __mcheck_cpu_init_early() 1880 mce_flags.succor = !!cpu_has(c, X86_FEATURE_SUCCOR); in __mcheck_cpu_init_early() 1881 mce_flags.smca = !!cpu_has(c, X86_FEATURE_SMCA); in __mcheck_cpu_init_early()
|
/arch/x86/kernel/cpu/mtrr/ |
D | if.c | 415 if ((!cpu_has(c, X86_FEATURE_MTRR)) && in mtrr_if_init() 416 (!cpu_has(c, X86_FEATURE_K6_MTRR)) && in mtrr_if_init() 417 (!cpu_has(c, X86_FEATURE_CYRIX_ARR)) && in mtrr_if_init() 418 (!cpu_has(c, X86_FEATURE_CENTAUR_MCR))) in mtrr_if_init()
|
/arch/x86/kernel/cpu/resctrl/ |
D | core.c | 972 if (!cpu_has(c, X86_FEATURE_CQM_LLC)) { in resctrl_cpu_detect() 982 if (cpu_has(c, X86_FEATURE_CQM_OCCUP_LLC) || in resctrl_cpu_detect() 983 cpu_has(c, X86_FEATURE_CQM_MBM_TOTAL) || in resctrl_cpu_detect() 984 cpu_has(c, X86_FEATURE_CQM_MBM_LOCAL)) { in resctrl_cpu_detect()
|
/arch/x86/kernel/ |
D | msr.c | 221 if (!cpu_has(c, X86_FEATURE_MSR)) in msr_open()
|
D | process.c | 804 if (!cpu_has(c, X86_FEATURE_MWAIT) || boot_cpu_has_bug(X86_BUG_MONITOR)) in prefer_mwait_c1_over_halt()
|
D | tsc.c | 1564 int constant_tsc = cpu_has(&cpu_data(cpu), X86_FEATURE_CONSTANT_TSC); in calibrate_delay_is_known()
|
/arch/x86/power/ |
D | cpu.c | 271 if (cpu_has(c, X86_FEATURE_MSR_IA32_FEAT_CTL)) in __restore_processor_state()
|
/arch/x86/kernel/cpu/microcode/ |
D | intel.c | 988 cpu_has(c, X86_FEATURE_IA64)) { in init_intel_microcode()
|
/arch/x86/kvm/svm/ |
D | svm.c | 503 if (cpu_has(&boot_cpu_data, X86_FEATURE_OSVW)) { in svm_hardware_enable()
|
/arch/x86/events/intel/ |
D | core.c | 4951 if (!cpu_has(&boot_cpu_data, X86_FEATURE_ARCH_PERFMON)) { in intel_pmu_init()
|
/arch/x86/kvm/ |
D | x86.c | 10579 if (__cr4_reserved_bits(cpu_has, c) != in kvm_arch_check_processor_compat() 10580 __cr4_reserved_bits(cpu_has, &boot_cpu_data)) in kvm_arch_check_processor_compat()
|