VCPU_SREG_ES
tss->es = get_segment_selector(ctxt, VCPU_SREG_ES);
set_segment_selector(ctxt, tss->es, VCPU_SREG_ES);
ret = __load_segment_descriptor(ctxt, tss->es, VCPU_SREG_ES, cpl,
tss->es = get_segment_selector(ctxt, VCPU_SREG_ES);
set_segment_selector(ctxt, tss->es, VCPU_SREG_ES);
ret = __load_segment_descriptor(ctxt, tss->es, VCPU_SREG_ES, cpl,
op->addr.mem.seg = VCPU_SREG_ES;
op->val = VCPU_SREG_ES;
ctxt->seg_override = VCPU_SREG_ES;
enter_smm_save_seg_32(vcpu, &smram->es, &smram->es_sel, VCPU_SREG_ES);
enter_smm_save_seg_64(vcpu, &smram->es, VCPU_SREG_ES);
kvm_set_segment(vcpu, &ds, VCPU_SREG_ES);
rsm_load_seg_32(vcpu, &smstate->es, smstate->es_sel, VCPU_SREG_ES);
rsm_load_seg_64(vcpu, &smstate->es, VCPU_SREG_ES);
case VCPU_SREG_ES: return &save->es;
case VCPU_SREG_ES:
__vmx_set_segment(vcpu, &seg, VCPU_SREG_ES);
vmx_get_segment(vcpu, &vmx->rmode.segs[VCPU_SREG_ES], VCPU_SREG_ES);
fix_pmode_seg(vcpu, VCPU_SREG_ES, &vmx->rmode.segs[VCPU_SREG_ES]);
vmx_get_segment(vcpu, &vmx->rmode.segs[VCPU_SREG_ES], VCPU_SREG_ES);
fix_rmode_seg(VCPU_SREG_ES, &vmx->rmode.segs[VCPU_SREG_ES]);
if (!rmode_segment_valid(vcpu, VCPU_SREG_ES))
if (!data_segment_valid(vcpu, VCPU_SREG_ES))
seg_setup(VCPU_SREG_ES);
kvm_get_segment(vcpu, &sregs->es, VCPU_SREG_ES);
kvm_set_segment(vcpu, &sregs->es, VCPU_SREG_ES);