Symbol: VCPU_SREG_ES
arch/x86/kvm/emulate.c
2634
tss->es = get_segment_selector(ctxt, VCPU_SREG_ES);
arch/x86/kvm/emulate.c
2663
set_segment_selector(ctxt, tss->es, VCPU_SREG_ES);
arch/x86/kvm/emulate.c
2678
ret = __load_segment_descriptor(ctxt, tss->es, VCPU_SREG_ES, cpl,
arch/x86/kvm/emulate.c
2747
tss->es = get_segment_selector(ctxt, VCPU_SREG_ES);
arch/x86/kvm/emulate.c
2782
set_segment_selector(ctxt, tss->es, VCPU_SREG_ES);
arch/x86/kvm/emulate.c
2810
ret = __load_segment_descriptor(ctxt, tss->es, VCPU_SREG_ES, cpl,
arch/x86/kvm/emulate.c
4655
op->addr.mem.seg = VCPU_SREG_ES;
arch/x86/kvm/emulate.c
4734
op->val = VCPU_SREG_ES;
arch/x86/kvm/emulate.c
4913
ctxt->seg_override = VCPU_SREG_ES;
arch/x86/kvm/smm.c
212
enter_smm_save_seg_32(vcpu, &smram->es, &smram->es_sel, VCPU_SREG_ES);
arch/x86/kvm/smm.c
264
enter_smm_save_seg_64(vcpu, &smram->es, VCPU_SREG_ES);
arch/x86/kvm/smm.c
355
kvm_set_segment(vcpu, &ds, VCPU_SREG_ES);
arch/x86/kvm/smm.c
493
rsm_load_seg_32(vcpu, &smstate->es, smstate->es_sel, VCPU_SREG_ES);
arch/x86/kvm/smm.c
555
rsm_load_seg_64(vcpu, &smstate->es, VCPU_SREG_ES);
arch/x86/kvm/svm/svm.c
1559
case VCPU_SREG_ES: return &save->es;
arch/x86/kvm/svm/svm.c
1618
case VCPU_SREG_ES:
arch/x86/kvm/vmx/nested.c
4880
__vmx_set_segment(vcpu, &seg, VCPU_SREG_ES);
arch/x86/kvm/vmx/vmx.c
3200
vmx_get_segment(vcpu, &vmx->rmode.segs[VCPU_SREG_ES], VCPU_SREG_ES);
arch/x86/kvm/vmx/vmx.c
3223
fix_pmode_seg(vcpu, VCPU_SREG_ES, &vmx->rmode.segs[VCPU_SREG_ES]);
arch/x86/kvm/vmx/vmx.c
3277
vmx_get_segment(vcpu, &vmx->rmode.segs[VCPU_SREG_ES], VCPU_SREG_ES);
arch/x86/kvm/vmx/vmx.c
3303
fix_rmode_seg(VCPU_SREG_ES, &vmx->rmode.segs[VCPU_SREG_ES]);
arch/x86/kvm/vmx/vmx.c
4030
if (!rmode_segment_valid(vcpu, VCPU_SREG_ES))
arch/x86/kvm/vmx/vmx.c
4046
if (!data_segment_valid(vcpu, VCPU_SREG_ES))
arch/x86/kvm/vmx/vmx.c
5111
seg_setup(VCPU_SREG_ES);
arch/x86/kvm/x86.c
12141
kvm_get_segment(vcpu, &sregs->es, VCPU_SREG_ES);
arch/x86/kvm/x86.c
12404
kvm_set_segment(vcpu, &sregs->es, VCPU_SREG_ES);