VCPU_SREG_CS
if (seg == VCPU_SREG_CS || seg == VCPU_SREG_TR)
case VCPU_SREG_CS:
ctxt->src.val = get_segment_selector(ctxt, VCPU_SREG_CS);
rc = load_segment_descriptor(ctxt, cs, VCPU_SREG_CS);
rc = load_segment_descriptor(ctxt, (u16)cs, VCPU_SREG_CS);
rc = __load_segment_descriptor(ctxt, sel, VCPU_SREG_CS, cpl,
rc = __load_segment_descriptor(ctxt, (u16)cs, VCPU_SREG_CS, cpl,
ops->set_segment(ctxt, cs_sel, &cs, 0, VCPU_SREG_CS);
ops->set_segment(ctxt, cs_sel, &cs, 0, VCPU_SREG_CS);
ops->set_segment(ctxt, cs_sel, &cs, 0, VCPU_SREG_CS);
tss->cs = get_segment_selector(ctxt, VCPU_SREG_CS);
set_segment_selector(ctxt, tss->cs, VCPU_SREG_CS);
ret = __load_segment_descriptor(ctxt, tss->cs, VCPU_SREG_CS, cpl,
tss->cs = get_segment_selector(ctxt, VCPU_SREG_CS);
set_segment_selector(ctxt, tss->cs, VCPU_SREG_CS);
ret = __load_segment_descriptor(ctxt, tss->cs, VCPU_SREG_CS, cpl,
ops->get_segment(ctxt, &old_cs, &old_desc, NULL, VCPU_SREG_CS);
rc = __load_segment_descriptor(ctxt, sel, VCPU_SREG_CS, cpl,
ops->set_segment(ctxt, old_cs, &old_desc, 0, VCPU_SREG_CS);
if (ctxt->modrm_reg == VCPU_SREG_CS || ctxt->modrm_reg > VCPU_SREG_GS)
op->val = VCPU_SREG_CS;
ctxt->ops->get_segment(ctxt, &dummy, &desc, NULL, VCPU_SREG_CS);
ctxt->seg_override = VCPU_SREG_CS;
struct segmented_address addr = { .seg = VCPU_SREG_CS,
if (!ctxt->ops->get_segment(ctxt, &selector, &cs, &base3, VCPU_SREG_CS))
struct segmented_address addr = { .seg = VCPU_SREG_CS,
enter_smm_save_seg_32(vcpu, &smram->cs, &smram->cs_sel, VCPU_SREG_CS);
enter_smm_save_seg_64(vcpu, &smram->cs, VCPU_SREG_CS);
kvm_set_segment(vcpu, &cs, VCPU_SREG_CS);
rsm_load_seg_32(vcpu, &smstate->cs, smstate->cs_sel, VCPU_SREG_CS);
rsm_load_seg_64(vcpu, &smstate->cs, VCPU_SREG_CS);
kvm_set_segment(vcpu, &cs_desc, VCPU_SREG_CS);
case VCPU_SREG_CS: return &save->cs;
svm_get_segment(vcpu, &cs, VCPU_SREG_CS);
VCPU_SREG_CS);
__vmx_set_segment(vcpu, &seg, VCPU_SREG_CS);
if (seg == VCPU_SREG_CS || seg == VCPU_SREG_SS)
vmx_get_segment(vcpu, &vmx->rmode.segs[VCPU_SREG_CS], VCPU_SREG_CS);
fix_pmode_seg(vcpu, VCPU_SREG_CS, &vmx->rmode.segs[VCPU_SREG_CS]);
if (seg == VCPU_SREG_CS)
vmx_get_segment(vcpu, &vmx->rmode.segs[VCPU_SREG_CS], VCPU_SREG_CS);
fix_rmode_seg(VCPU_SREG_CS, &vmx->rmode.segs[VCPU_SREG_CS]);
u32 ar = vmx_read_guest_seg_ar(to_vmx(vcpu), VCPU_SREG_CS);
if (seg == VCPU_SREG_CS)
vmx_get_segment(vcpu, &cs, VCPU_SREG_CS);
vmx_get_segment(vcpu, &cs, VCPU_SREG_CS);
if (!rmode_segment_valid(vcpu, VCPU_SREG_CS))
if (seg == VCPU_SREG_CS)
seg_setup(VCPU_SREG_CS);
kvm_get_segment(vcpu, &sregs->cs, VCPU_SREG_CS);
kvm_set_segment(vcpu, &sregs->cs, VCPU_SREG_CS);
kvm_get_segment(vcpu, &cs, VCPU_SREG_CS);
kvm_set_segment(vcpu, &cs, VCPU_SREG_CS);
return (u32)(get_segment_base(vcpu, VCPU_SREG_CS) +