/arch/x86/hyperv/ |
D | hv_init.c | 83 wrmsrl(HV_X64_MSR_VP_ASSIST_PAGE, msr.as_uint64); in hv_cpu_init() 110 wrmsrl(HV_X64_MSR_TSC_EMULATION_STATUS, *(u64 *)&emu_status); in hyperv_stop_tsc_emulation() 158 wrmsrl(HV_X64_MSR_REENLIGHTENMENT_CONTROL, *((u64 *)&re_ctrl)); in set_hv_tscchange_cb() 159 wrmsrl(HV_X64_MSR_TSC_EMULATION_CONTROL, *((u64 *)&emu_ctrl)); in set_hv_tscchange_cb() 174 wrmsrl(HV_X64_MSR_REENLIGHTENMENT_CONTROL, *(u64 *)&re_ctrl); in clear_hv_tscchange_cb() 201 wrmsrl(HV_X64_MSR_VP_ASSIST_PAGE, msr.as_uint64); in hv_cpu_die() 221 wrmsrl(HV_X64_MSR_REENLIGHTENMENT_CONTROL, *((u64 *)&re_ctrl)); in hv_cpu_die() 264 wrmsrl(HV_X64_MSR_HYPERCALL, hypercall_msr.as_uint64); in hv_suspend() 283 wrmsrl(HV_X64_MSR_HYPERCALL, hypercall_msr.as_uint64); in hv_resume() 377 wrmsrl(HV_X64_MSR_GUEST_OS_ID, guest_id); in hyperv_init() [all …]
|
/arch/x86/kernel/cpu/ |
D | tsx.c | 39 wrmsrl(MSR_IA32_TSX_CTRL, tsx); in tsx_disable() 58 wrmsrl(MSR_IA32_TSX_CTRL, tsx); in tsx_enable() 119 wrmsrl(MSR_TSX_FORCE_ABORT, msr); in tsx_clear_cpuid() 123 wrmsrl(MSR_IA32_TSX_CTRL, msr); in tsx_clear_cpuid() 152 wrmsrl(MSR_IA32_MCU_OPT_CTRL, mcu_opt_ctrl); in tsx_dev_mode_disable()
|
D | common.c | 606 wrmsrl(MSR_GS_BASE, cpu_kernelmode_gs_base(cpu)); in load_percpu_segment() 1531 wrmsrl(MSR_FS_BASE, 1); in detect_null_seg_behavior() 1534 wrmsrl(MSR_FS_BASE, old_base); in detect_null_seg_behavior() 1906 wrmsrl(MSR_LSTAR, (unsigned long)entry_SYSCALL_64); in syscall_init() 1909 wrmsrl(MSR_CSTAR, (unsigned long)entry_SYSCALL_compat); in syscall_init() 1921 wrmsrl(MSR_CSTAR, (unsigned long)ignore_sysret); in syscall_init() 1931 wrmsrl(MSR_SYSCALL_MASK, in syscall_init() 2133 wrmsrl(MSR_FS_BASE, 0); in cpu_init() 2134 wrmsrl(MSR_KERNEL_GS_BASE, 0); in cpu_init()
|
D | intel_epb.c | 97 wrmsrl(MSR_IA32_ENERGY_PERF_BIAS, (epb & ~EPB_MASK) | val); in intel_epb_restore()
|
/arch/x86/events/intel/ |
D | uncore_nhmex.c | 202 wrmsrl(NHMEX_U_MSR_PMON_GLOBAL_CTL, NHMEX_U_PMON_GLOBAL_EN_ALL); in nhmex_uncore_msr_init_box() 207 wrmsrl(NHMEX_U_MSR_PMON_GLOBAL_CTL, 0); in nhmex_uncore_msr_exit_box() 221 wrmsrl(msr, config); in nhmex_uncore_msr_disable_box() 236 wrmsrl(msr, config); in nhmex_uncore_msr_enable_box() 242 wrmsrl(event->hw.config_base, 0); in nhmex_uncore_msr_disable_event() 250 wrmsrl(hwc->config_base, NHMEX_PMON_CTL_EN_BIT0); in nhmex_uncore_msr_enable_event() 252 wrmsrl(hwc->config_base, hwc->config | NHMEX_PMON_CTL_EN_BIT22); in nhmex_uncore_msr_enable_event() 254 wrmsrl(hwc->config_base, hwc->config | NHMEX_PMON_CTL_EN_BIT0); in nhmex_uncore_msr_enable_event() 384 wrmsrl(reg1->reg, reg1->config); in nhmex_bbox_msr_enable_event() 385 wrmsrl(reg1->reg + 1, reg2->config); in nhmex_bbox_msr_enable_event() [all …]
|
D | uncore_snb.c | 170 wrmsrl(hwc->config_base, hwc->config | SNB_UNC_CTL_EN); in snb_uncore_msr_enable_event() 172 wrmsrl(hwc->config_base, SNB_UNC_CTL_EN); in snb_uncore_msr_enable_event() 177 wrmsrl(event->hw.config_base, 0); in snb_uncore_msr_disable_event() 183 wrmsrl(SNB_UNC_PERF_GLOBAL_CTL, in snb_uncore_msr_init_box() 190 wrmsrl(SNB_UNC_PERF_GLOBAL_CTL, in snb_uncore_msr_enable_box() 197 wrmsrl(SNB_UNC_PERF_GLOBAL_CTL, 0); in snb_uncore_msr_exit_box() 282 wrmsrl(SKL_UNC_PERF_GLOBAL_CTL, in skl_uncore_msr_init_box() 293 wrmsrl(SKL_UNC_PERF_GLOBAL_CTL, in skl_uncore_msr_enable_box() 300 wrmsrl(SKL_UNC_PERF_GLOBAL_CTL, 0); in skl_uncore_msr_exit_box() 435 wrmsrl(SKL_UNC_PERF_GLOBAL_CTL, SNB_UNC_GLOBAL_CTL_EN); in rkl_uncore_msr_init_box() [all …]
|
D | lbr.c | 147 wrmsrl(MSR_LBR_SELECT, lbr_select); in __intel_pmu_lbr_enable() 165 wrmsrl(MSR_IA32_DEBUGCTLMSR, debugctl); in __intel_pmu_lbr_enable() 168 wrmsrl(MSR_ARCH_LBR_CTL, lbr_select | ARCH_LBR_CTL_LBREN); in __intel_pmu_lbr_enable() 176 wrmsrl(MSR_ARCH_LBR_CTL, 0); in __intel_pmu_lbr_disable() 182 wrmsrl(MSR_IA32_DEBUGCTLMSR, debugctl); in __intel_pmu_lbr_disable() 190 wrmsrl(x86_pmu.lbr_from + i, 0); in intel_pmu_lbr_reset_32() 198 wrmsrl(x86_pmu.lbr_from + i, 0); in intel_pmu_lbr_reset_64() 199 wrmsrl(x86_pmu.lbr_to + i, 0); in intel_pmu_lbr_reset_64() 201 wrmsrl(x86_pmu.lbr_info + i, 0); in intel_pmu_lbr_reset_64() 208 wrmsrl(MSR_ARCH_LBR_DEPTH, x86_pmu.lbr_nr); in intel_pmu_arch_lbr_reset() [all …]
|
D | knc.c | 164 wrmsrl(MSR_KNC_IA32_PERF_GLOBAL_CTRL, val); in knc_pmu_disable_all() 173 wrmsrl(MSR_KNC_IA32_PERF_GLOBAL_CTRL, val); in knc_pmu_enable_all() 210 wrmsrl(MSR_KNC_IA32_PERF_GLOBAL_OVF_CONTROL, ack); in knc_pmu_ack_status()
|
D | p6.c | 145 wrmsrl(MSR_P6_EVNTSEL0, val); in p6_pmu_disable_all() 155 wrmsrl(MSR_P6_EVNTSEL0, val); in p6_pmu_enable_all()
|
D | uncore_discovery.c | 354 wrmsrl(uncore_msr_box_ctl(box), GENERIC_PMON_BOX_CTL_INT); in intel_generic_uncore_msr_init_box() 359 wrmsrl(uncore_msr_box_ctl(box), GENERIC_PMON_BOX_CTL_FRZ); in intel_generic_uncore_msr_disable_box() 364 wrmsrl(uncore_msr_box_ctl(box), 0); in intel_generic_uncore_msr_enable_box() 372 wrmsrl(hwc->config_base, hwc->config); in intel_generic_uncore_msr_enable_event() 380 wrmsrl(hwc->config_base, 0); in intel_generic_uncore_msr_disable_event()
|
D | pt.c | 409 wrmsrl(MSR_IA32_RTIT_CTL, ctl); in pt_config_start() 468 wrmsrl(pt_address_ranges[range].msr_a, filter->msr_a); in pt_config_filters() 473 wrmsrl(pt_address_ranges[range].msr_b, filter->msr_b); in pt_config_filters() 492 wrmsrl(MSR_IA32_RTIT_STATUS, 0); in pt_config() 535 wrmsrl(MSR_IA32_RTIT_CTL, ctl); in pt_config_stop() 624 wrmsrl(MSR_IA32_RTIT_OUTPUT_BASE, reg); in pt_config_buffer() 630 wrmsrl(MSR_IA32_RTIT_OUTPUT_MASK, reg); in pt_config_buffer() 930 wrmsrl(MSR_IA32_RTIT_STATUS, status); in pt_handle_status() 1540 wrmsrl(MSR_IA32_RTIT_CTL, event->hw.config); in intel_pt_handle_vmx()
|
/arch/x86/xen/ |
D | suspend.c | 44 wrmsrl(MSR_IA32_SPEC_CTRL, this_cpu_read(spec_ctrl)); in xen_vcpu_notify_restore() 62 wrmsrl(MSR_IA32_SPEC_CTRL, 0); in xen_vcpu_notify_suspend()
|
/arch/x86/include/asm/ |
D | virtext.h | 121 wrmsrl(MSR_VM_HSAVE_PA, 0); in cpu_svm_disable() 136 wrmsrl(MSR_EFER, efer & ~EFER_SVME); in cpu_svm_disable()
|
D | fsgsbase.h | 73 wrmsrl(MSR_FS_BASE, fsbase); in x86_fsbase_write_cpu()
|
/arch/x86/kernel/cpu/mce/ |
D | inject.c | 465 wrmsrl(MSR_IA32_MCG_STATUS, m.mcgstatus); in prepare_msrs() 469 wrmsrl(MSR_AMD64_SMCA_MCx_DESTAT(b), m.status); in prepare_msrs() 470 wrmsrl(MSR_AMD64_SMCA_MCx_DEADDR(b), m.addr); in prepare_msrs() 472 wrmsrl(MSR_AMD64_SMCA_MCx_STATUS(b), m.status); in prepare_msrs() 473 wrmsrl(MSR_AMD64_SMCA_MCx_ADDR(b), m.addr); in prepare_msrs() 476 wrmsrl(MSR_AMD64_SMCA_MCx_MISC(b), m.misc); in prepare_msrs() 477 wrmsrl(MSR_AMD64_SMCA_MCx_SYND(b), m.synd); in prepare_msrs() 479 wrmsrl(MSR_IA32_MCx_STATUS(b), m.status); in prepare_msrs() 480 wrmsrl(MSR_IA32_MCx_ADDR(b), m.addr); in prepare_msrs() 481 wrmsrl(MSR_IA32_MCx_MISC(b), m.misc); in prepare_msrs()
|
D | intel.c | 171 wrmsrl(MSR_IA32_MCx_CTL2(bank), val); in cmci_toggle_interrupt_mode() 309 wrmsrl(MSR_IA32_MCx_CTL2(i), val); in cmci_discover() 364 wrmsrl(MSR_IA32_MCx_CTL2(bank), val); in __cmci_disable_bank() 458 wrmsrl(MSR_IA32_MCG_EXT_CTL, val | MCG_EXT_CTL_LMCE_EN); in intel_init_lmce() 470 wrmsrl(MSR_IA32_MCG_EXT_CTL, val); in intel_clear_lmce()
|
/arch/x86/power/ |
D | cpu.c | 57 wrmsrl(msr->info.msr_no, msr->info.reg.q); in msr_restore_context() 199 wrmsrl(MSR_IA32_MISC_ENABLE, ctxt->misc_enable); in __restore_processor_state() 209 wrmsrl(MSR_EFER, ctxt->efer); in __restore_processor_state() 232 wrmsrl(MSR_GS_BASE, ctxt->kernelmode_gs_base); in __restore_processor_state() 255 wrmsrl(MSR_FS_BASE, ctxt->fs_base); in __restore_processor_state() 256 wrmsrl(MSR_KERNEL_GS_BASE, ctxt->usermode_gs_base); in __restore_processor_state()
|
/arch/x86/kernel/ |
D | kvm.c | 300 wrmsrl(MSR_KVM_ASYNC_PF_ACK, 1); in DEFINE_IDTENTRY_SYSVEC() 326 wrmsrl(MSR_KVM_STEAL_TIME, (slow_virt_to_phys(st) | KVM_MSR_ENABLED)); in kvm_register_steal_time() 360 wrmsrl(MSR_KVM_ASYNC_PF_INT, HYPERVISOR_CALLBACK_VECTOR); in kvm_guest_cpu_init() 362 wrmsrl(MSR_KVM_ASYNC_PF_EN, pa); in kvm_guest_cpu_init() 375 wrmsrl(MSR_KVM_PV_EOI_EN, pa); in kvm_guest_cpu_init() 387 wrmsrl(MSR_KVM_ASYNC_PF_EN, 0); in kvm_pv_disable_apf() 449 wrmsrl(MSR_KVM_PV_EOI_EN, 0); in kvm_guest_cpu_offline() 685 wrmsrl(MSR_KVM_POLL_CONTROL, 0); in kvm_resume() 1019 wrmsrl(MSR_KVM_POLL_CONTROL, 0); in kvm_disable_host_haltpoll() 1024 wrmsrl(MSR_KVM_POLL_CONTROL, 1); in kvm_enable_host_haltpoll()
|
D | tsc_sync.c | 72 wrmsrl(MSR_IA32_TSC_ADJUST, adj->adjusted); in tsc_verify_tsc_adjust() 144 wrmsrl(MSR_IA32_TSC_ADJUST, 0); in tsc_sanitize_first_cpu() 233 wrmsrl(MSR_IA32_TSC_ADJUST, ref->adjusted); in tsc_store_and_check_tsc_adjust() 529 wrmsrl(MSR_IA32_TSC_ADJUST, cur->adjusted); in check_tsc_sync_target()
|
D | process.c | 287 wrmsrl(MSR_MISC_FEATURES_ENABLES, msrval); in set_cpuid_faulting() 502 wrmsrl(MSR_AMD64_LS_CFG, msr); in amd_set_core_ssb_state() 519 wrmsrl(MSR_AMD64_LS_CFG, msr); in amd_set_core_ssb_state() 529 wrmsrl(MSR_AMD64_LS_CFG, msr); in amd_set_core_ssb_state() 538 wrmsrl(MSR_AMD64_LS_CFG, msr); in amd_set_core_ssb_state() 548 wrmsrl(MSR_AMD64_VIRT_SPEC_CTRL, ssbd_tif_to_spec_ctrl(tifn)); in amd_set_ssb_virt_state() 655 wrmsrl(MSR_IA32_DEBUGCTLMSR, debugctl); in __switch_to_xtra()
|
D | reboot_fixups_32.c | 30 wrmsrl(MSR_DIVIL_SOFT_RESET, 1ULL); in cs5536_warm_reset()
|
D | kvmclock.c | 63 wrmsrl(msr_kvm_wall_clock, slow_virt_to_phys(&wall_clock)); in kvm_get_wallclock() 176 wrmsrl(msr_kvm_system_time, pa); in kvm_register_clock()
|
D | process_64.c | 199 wrmsrl(MSR_KERNEL_GS_BASE, gsbase); in __wrgsbase_inactive() 331 wrmsrl(which == FS ? MSR_FS_BASE : MSR_KERNEL_GS_BASE, in load_seg_legacy() 456 wrmsrl(MSR_KERNEL_GS_BASE, gsbase); in x86_gsbase_write_cpu_inactive()
|
/arch/x86/events/zhaoxin/ |
D | core.c | 257 wrmsrl(MSR_CORE_PERF_GLOBAL_CTRL, 0); in zhaoxin_pmu_disable_all() 262 wrmsrl(MSR_CORE_PERF_GLOBAL_CTRL, x86_pmu.intel_ctrl); in zhaoxin_pmu_enable_all() 276 wrmsrl(MSR_CORE_PERF_GLOBAL_OVF_CTRL, ack); in zhaoxin_pmu_ack_status() 298 wrmsrl(hwc->config_base, ctrl_val); in zhaoxin_pmu_disable_fixed() 335 wrmsrl(hwc->config_base, ctrl_val); in zhaoxin_pmu_enable_fixed()
|
/arch/x86/kernel/fpu/ |
D | xstate.c | 154 wrmsrl(MSR_IA32_XSS, xfeatures_mask_supervisor() | in fpu__init_cpu_xstate() 633 wrmsrl(MSR_IA32_XSS, xfeatures_mask_supervisor()); in get_xsaves_size_no_independent() 642 wrmsrl(MSR_IA32_XSS, xfeatures_mask_supervisor() | mask); in get_xsaves_size_no_independent() 847 wrmsrl(MSR_IA32_XSS, xfeatures_mask_supervisor() | in fpu__resume_cpu()
|