Searched refs:VCPU_SREG_LDTR (Results 1 – 6 of 6) sorted by relevance
/arch/x86/kvm/ |
D | emulate.c | 1517 VCPU_SREG_LDTR)) in get_descriptor_table_ptr() 1702 case VCPU_SREG_LDTR: in __load_segment_descriptor() 2471 ctxt->ops->set_segment(ctxt, selector, &desc, 0, VCPU_SREG_LDTR); in rsm_load_state_32() 2546 ctxt->ops->set_segment(ctxt, selector, &desc, base3, VCPU_SREG_LDTR); in rsm_load_state_64() 3011 tss->ldt = get_segment_selector(ctxt, VCPU_SREG_LDTR); in save_state_to_tss16() 3035 set_segment_selector(ctxt, tss->ldt, VCPU_SREG_LDTR); in load_state_from_tss16() 3047 ret = __load_segment_descriptor(ctxt, tss->ldt, VCPU_SREG_LDTR, cpl, in load_state_from_tss16() 3155 set_segment_selector(ctxt, tss->ldt_selector, VCPU_SREG_LDTR); in load_state_from_tss32() 3180 ret = __load_segment_descriptor(ctxt, tss->ldt_selector, VCPU_SREG_LDTR, in load_state_from_tss32() 3748 return em_store_sreg(ctxt, VCPU_SREG_LDTR); in em_sldt() [all …]
|
D | x86.c | 9336 kvm_get_segment(vcpu, &seg, VCPU_SREG_LDTR); in enter_smm_save_state_32() 9400 kvm_get_segment(vcpu, &seg, VCPU_SREG_LDTR); in enter_smm_save_state_64() 10452 kvm_get_segment(vcpu, &sregs->ldt, VCPU_SREG_LDTR); in __get_sregs_common() 10678 kvm_set_segment(vcpu, &sregs->ldt, VCPU_SREG_LDTR); in __set_sregs_common()
|
/arch/x86/include/asm/ |
D | kvm_host.h | 191 VCPU_SREG_LDTR, enumerator
|
/arch/x86/kvm/vmx/ |
D | vmx.c | 3324 if (vmx->rmode.vm86_active && seg != VCPU_SREG_LDTR) { in vmx_get_segment() 3402 if (vmx->rmode.vm86_active && seg != VCPU_SREG_LDTR) { in __vmx_set_segment() 3426 if (is_unrestricted_guest(vcpu) && (seg != VCPU_SREG_LDTR)) in __vmx_set_segment() 3589 vmx_get_segment(vcpu, &ldtr, VCPU_SREG_LDTR); in ldtr_valid()
|
D | nested.c | 4398 __vmx_set_segment(vcpu, &seg, VCPU_SREG_LDTR); in load_vmcs12_host_state()
|
/arch/x86/kvm/svm/ |
D | svm.c | 1618 case VCPU_SREG_LDTR: return &save01->ldtr; in svm_seg()
|