Symbol: VCPU_SREG_SS
arch/x86/kvm/emulate.c
1128
ctxt->modrm_seg = VCPU_SREG_SS;
arch/x86/kvm/emulate.c
1203
ctxt->modrm_seg = VCPU_SREG_SS;
arch/x86/kvm/emulate.c
1566
if (seg == VCPU_SREG_SS) {
arch/x86/kvm/emulate.c
1604
case VCPU_SREG_SS:
arch/x86/kvm/emulate.c
1691
err_vec = (seg == VCPU_SREG_SS) ? SS_VECTOR : NP_VECTOR;
arch/x86/kvm/emulate.c
1746
if (seg == VCPU_SREG_SS && selector == 3 &&
arch/x86/kvm/emulate.c
1811
addr.seg = VCPU_SREG_SS;
arch/x86/kvm/emulate.c
1830
addr.seg = VCPU_SREG_SS;
arch/x86/kvm/emulate.c
1947
if (seg == VCPU_SREG_SS)
arch/x86/kvm/emulate.c
2378
ops->set_segment(ctxt, ss_sel, &ss, 0, VCPU_SREG_SS);
arch/x86/kvm/emulate.c
2445
ops->set_segment(ctxt, ss_sel, &ss, 0, VCPU_SREG_SS);
arch/x86/kvm/emulate.c
2510
ops->set_segment(ctxt, ss_sel, &ss, 0, VCPU_SREG_SS);
arch/x86/kvm/emulate.c
2636
tss->ss = get_segment_selector(ctxt, VCPU_SREG_SS);
arch/x86/kvm/emulate.c
2665
set_segment_selector(ctxt, tss->ss, VCPU_SREG_SS);
arch/x86/kvm/emulate.c
2686
ret = __load_segment_descriptor(ctxt, tss->ss, VCPU_SREG_SS, cpl,
arch/x86/kvm/emulate.c
2749
tss->ss = get_segment_selector(ctxt, VCPU_SREG_SS);
arch/x86/kvm/emulate.c
2784
set_segment_selector(ctxt, tss->ss, VCPU_SREG_SS);
arch/x86/kvm/emulate.c
2818
ret = __load_segment_descriptor(ctxt, tss->ss, VCPU_SREG_SS, cpl,
arch/x86/kvm/emulate.c
3358
if (ctxt->modrm_reg == VCPU_SREG_SS)
arch/x86/kvm/emulate.c
473
ctxt->ops->get_segment(ctxt, &sel, &ss, NULL, VCPU_SREG_SS);
arch/x86/kvm/emulate.c
4742
op->val = VCPU_SREG_SS;
arch/x86/kvm/emulate.c
4921
ctxt->seg_override = VCPU_SREG_SS;
arch/x86/kvm/emulate.c
698
if (addr.seg == VCPU_SREG_SS)
arch/x86/kvm/smm.c
214
enter_smm_save_seg_32(vcpu, &smram->ss, &smram->ss_sel, VCPU_SREG_SS);
arch/x86/kvm/smm.c
266
enter_smm_save_seg_64(vcpu, &smram->ss, VCPU_SREG_SS);
arch/x86/kvm/smm.c
358
kvm_set_segment(vcpu, &ds, VCPU_SREG_SS);
arch/x86/kvm/smm.c
495
rsm_load_seg_32(vcpu, &smstate->ss, smstate->ss_sel, VCPU_SREG_SS);
arch/x86/kvm/smm.c
557
rsm_load_seg_64(vcpu, &smstate->ss, VCPU_SREG_SS);
arch/x86/kvm/svm/svm.c
1562
case VCPU_SREG_SS: return &save->ss;
arch/x86/kvm/svm/svm.c
1631
case VCPU_SREG_SS:
arch/x86/kvm/svm/svm.c
1825
if (seg == VCPU_SREG_SS)
arch/x86/kvm/vmx/nested.c
4882
__vmx_set_segment(vcpu, &seg, VCPU_SREG_SS);
arch/x86/kvm/vmx/nested.c
5334
seg_reg == VCPU_SREG_SS ?
arch/x86/kvm/vmx/vmx.c
3183
if (seg == VCPU_SREG_CS || seg == VCPU_SREG_SS)
arch/x86/kvm/vmx/vmx.c
3204
vmx_get_segment(vcpu, &vmx->rmode.segs[VCPU_SREG_SS], VCPU_SREG_SS);
arch/x86/kvm/vmx/vmx.c
3222
fix_pmode_seg(vcpu, VCPU_SREG_SS, &vmx->rmode.segs[VCPU_SREG_SS]);
arch/x86/kvm/vmx/vmx.c
3281
vmx_get_segment(vcpu, &vmx->rmode.segs[VCPU_SREG_SS], VCPU_SREG_SS);
arch/x86/kvm/vmx/vmx.c
3301
fix_rmode_seg(VCPU_SREG_SS, &vmx->rmode.segs[VCPU_SREG_SS]);
arch/x86/kvm/vmx/vmx.c
3765
ar = vmx_read_guest_seg_ar(vmx, VCPU_SREG_SS);
arch/x86/kvm/vmx/vmx.c
3926
vmx_get_segment(vcpu, &ss, VCPU_SREG_SS);
arch/x86/kvm/vmx/vmx.c
4009
vmx_get_segment(vcpu, &ss, VCPU_SREG_SS);
arch/x86/kvm/vmx/vmx.c
4026
if (!rmode_segment_valid(vcpu, VCPU_SREG_SS))
arch/x86/kvm/vmx/vmx.c
5114
seg_setup(VCPU_SREG_SS);
arch/x86/kvm/x86.c
12144
kvm_get_segment(vcpu, &sregs->ss, VCPU_SREG_SS);
arch/x86/kvm/x86.c
12407
kvm_set_segment(vcpu, &sregs->ss, VCPU_SREG_SS);