Symbol: VCPU_SREG_CS
arch/x86/kvm/emulate.c
1563
if (seg == VCPU_SREG_CS || seg == VCPU_SREG_TR)
arch/x86/kvm/emulate.c
1612
case VCPU_SREG_CS:
arch/x86/kvm/emulate.c
2020
ctxt->src.val = get_segment_selector(ctxt, VCPU_SREG_CS);
arch/x86/kvm/emulate.c
2043
rc = load_segment_descriptor(ctxt, cs, VCPU_SREG_CS);
arch/x86/kvm/emulate.c
2113
rc = load_segment_descriptor(ctxt, (u16)cs, VCPU_SREG_CS);
arch/x86/kvm/emulate.c
2158
rc = __load_segment_descriptor(ctxt, sel, VCPU_SREG_CS, cpl,
arch/x86/kvm/emulate.c
2238
rc = __load_segment_descriptor(ctxt, (u16)cs, VCPU_SREG_CS, cpl,
arch/x86/kvm/emulate.c
2377
ops->set_segment(ctxt, cs_sel, &cs, 0, VCPU_SREG_CS);
arch/x86/kvm/emulate.c
2444
ops->set_segment(ctxt, cs_sel, &cs, 0, VCPU_SREG_CS);
arch/x86/kvm/emulate.c
2509
ops->set_segment(ctxt, cs_sel, &cs, 0, VCPU_SREG_CS);
arch/x86/kvm/emulate.c
2635
tss->cs = get_segment_selector(ctxt, VCPU_SREG_CS);
arch/x86/kvm/emulate.c
2664
set_segment_selector(ctxt, tss->cs, VCPU_SREG_CS);
arch/x86/kvm/emulate.c
2682
ret = __load_segment_descriptor(ctxt, tss->cs, VCPU_SREG_CS, cpl,
arch/x86/kvm/emulate.c
2748
tss->cs = get_segment_selector(ctxt, VCPU_SREG_CS);
arch/x86/kvm/emulate.c
2783
set_segment_selector(ctxt, tss->cs, VCPU_SREG_CS);
arch/x86/kvm/emulate.c
2814
ret = __load_segment_descriptor(ctxt, tss->cs, VCPU_SREG_CS, cpl,
arch/x86/kvm/emulate.c
3106
ops->get_segment(ctxt, &old_cs, &old_desc, NULL, VCPU_SREG_CS);
arch/x86/kvm/emulate.c
3109
rc = __load_segment_descriptor(ctxt, sel, VCPU_SREG_CS, cpl,
arch/x86/kvm/emulate.c
3133
ops->set_segment(ctxt, old_cs, &old_desc, 0, VCPU_SREG_CS);
arch/x86/kvm/emulate.c
3355
if (ctxt->modrm_reg == VCPU_SREG_CS || ctxt->modrm_reg > VCPU_SREG_GS)
arch/x86/kvm/emulate.c
4738
op->val = VCPU_SREG_CS;
arch/x86/kvm/emulate.c
4872
ctxt->ops->get_segment(ctxt, &dummy, &desc, NULL, VCPU_SREG_CS);
arch/x86/kvm/emulate.c
4917
ctxt->seg_override = VCPU_SREG_CS;
arch/x86/kvm/emulate.c
719
struct segmented_address addr = { .seg = VCPU_SREG_CS,
arch/x86/kvm/emulate.c
756
if (!ctxt->ops->get_segment(ctxt, &selector, &cs, &base3, VCPU_SREG_CS))
arch/x86/kvm/emulate.c
848
struct segmented_address addr = { .seg = VCPU_SREG_CS,
arch/x86/kvm/smm.c
213
enter_smm_save_seg_32(vcpu, &smram->cs, &smram->cs_sel, VCPU_SREG_CS);
arch/x86/kvm/smm.c
265
enter_smm_save_seg_64(vcpu, &smram->cs, VCPU_SREG_CS);
arch/x86/kvm/smm.c
353
kvm_set_segment(vcpu, &cs, VCPU_SREG_CS);
arch/x86/kvm/smm.c
494
rsm_load_seg_32(vcpu, &smstate->cs, smstate->cs_sel, VCPU_SREG_CS);
arch/x86/kvm/smm.c
556
rsm_load_seg_64(vcpu, &smstate->cs, VCPU_SREG_CS);
arch/x86/kvm/smm.c
611
kvm_set_segment(vcpu, &cs_desc, VCPU_SREG_CS);
arch/x86/kvm/svm/svm.c
1557
case VCPU_SREG_CS: return &save->cs;
arch/x86/kvm/svm/svm.c
1657
svm_get_segment(vcpu, &cs, VCPU_SREG_CS);
arch/x86/kvm/trace.h
931
VCPU_SREG_CS);
arch/x86/kvm/vmx/nested.c
4867
__vmx_set_segment(vcpu, &seg, VCPU_SREG_CS);
arch/x86/kvm/vmx/vmx.c
3183
if (seg == VCPU_SREG_CS || seg == VCPU_SREG_SS)
arch/x86/kvm/vmx/vmx.c
3205
vmx_get_segment(vcpu, &vmx->rmode.segs[VCPU_SREG_CS], VCPU_SREG_CS);
arch/x86/kvm/vmx/vmx.c
3221
fix_pmode_seg(vcpu, VCPU_SREG_CS, &vmx->rmode.segs[VCPU_SREG_CS]);
arch/x86/kvm/vmx/vmx.c
3235
if (seg == VCPU_SREG_CS)
arch/x86/kvm/vmx/vmx.c
3282
vmx_get_segment(vcpu, &vmx->rmode.segs[VCPU_SREG_CS], VCPU_SREG_CS);
arch/x86/kvm/vmx/vmx.c
3302
fix_rmode_seg(VCPU_SREG_CS, &vmx->rmode.segs[VCPU_SREG_CS]);
arch/x86/kvm/vmx/vmx.c
3842
u32 ar = vmx_read_guest_seg_ar(to_vmx(vcpu), VCPU_SREG_CS);
arch/x86/kvm/vmx/vmx.c
3879
if (seg == VCPU_SREG_CS)
arch/x86/kvm/vmx/vmx.c
3898
vmx_get_segment(vcpu, &cs, VCPU_SREG_CS);
arch/x86/kvm/vmx/vmx.c
4008
vmx_get_segment(vcpu, &cs, VCPU_SREG_CS);
arch/x86/kvm/vmx/vmx.c
4024
if (!rmode_segment_valid(vcpu, VCPU_SREG_CS))
arch/x86/kvm/vmx/vmx.c
4137
if (seg == VCPU_SREG_CS)
arch/x86/kvm/vmx/vmx.c
5106
seg_setup(VCPU_SREG_CS);
arch/x86/kvm/x86.c
12139
kvm_get_segment(vcpu, &sregs->cs, VCPU_SREG_CS);
arch/x86/kvm/x86.c
12402
kvm_set_segment(vcpu, &sregs->cs, VCPU_SREG_CS);
arch/x86/kvm/x86.c
13066
kvm_get_segment(vcpu, &cs, VCPU_SREG_CS);
arch/x86/kvm/x86.c
13069
kvm_set_segment(vcpu, &cs, VCPU_SREG_CS);
arch/x86/kvm/x86.c
13769
return (u32)(get_segment_base(vcpu, VCPU_SREG_CS) +