Searched refs:VCPU_SREG_LDTR (Results 1 – 5 of 5) sorted by relevance
/arch/x86/kvm/ |
D | emulate.c | 1620 VCPU_SREG_LDTR)) in get_descriptor_table_ptr() 1805 case VCPU_SREG_LDTR: in __load_segment_descriptor() 2574 ctxt->ops->set_segment(ctxt, selector, &desc, 0, VCPU_SREG_LDTR); in rsm_load_state_32() 2649 ctxt->ops->set_segment(ctxt, selector, &desc, base3, VCPU_SREG_LDTR); in rsm_load_state_64() 3107 tss->ldt = get_segment_selector(ctxt, VCPU_SREG_LDTR); in save_state_to_tss16() 3131 set_segment_selector(ctxt, tss->ldt, VCPU_SREG_LDTR); in load_state_from_tss16() 3143 ret = __load_segment_descriptor(ctxt, tss->ldt, VCPU_SREG_LDTR, cpl, in load_state_from_tss16() 3251 set_segment_selector(ctxt, tss->ldt_selector, VCPU_SREG_LDTR); in load_state_from_tss32() 3276 ret = __load_segment_descriptor(ctxt, tss->ldt_selector, VCPU_SREG_LDTR, in load_state_from_tss32() 3844 return em_store_sreg(ctxt, VCPU_SREG_LDTR); in em_sldt() [all …]
|
D | x86.c | 8751 kvm_get_segment(vcpu, &seg, VCPU_SREG_LDTR); in enter_smm_save_state_32() 8815 kvm_get_segment(vcpu, &seg, VCPU_SREG_LDTR); in enter_smm_save_state_64() 9774 kvm_get_segment(vcpu, &sregs->ldt, VCPU_SREG_LDTR); in __get_sregs() 9980 kvm_set_segment(vcpu, &sregs->ldt, VCPU_SREG_LDTR); in __set_sregs()
|
/arch/x86/include/asm/ |
D | kvm_host.h | 182 VCPU_SREG_LDTR, enumerator
|
/arch/x86/kvm/vmx/ |
D | vmx.c | 3293 if (vmx->rmode.vm86_active && seg != VCPU_SREG_LDTR) { in vmx_get_segment() 3371 if (vmx->rmode.vm86_active && seg != VCPU_SREG_LDTR) { in vmx_set_segment() 3395 if (is_unrestricted_guest(vcpu) && (seg != VCPU_SREG_LDTR)) in vmx_set_segment() 3554 vmx_get_segment(vcpu, &ldtr, VCPU_SREG_LDTR); in ldtr_valid()
|
/arch/x86/kvm/svm/ |
D | svm.c | 1516 case VCPU_SREG_LDTR: return &save->ldtr; in svm_seg()
|