VCPU_SREG_SS
ctxt->modrm_seg = VCPU_SREG_SS;
ctxt->modrm_seg = VCPU_SREG_SS;
if (seg == VCPU_SREG_SS) {
case VCPU_SREG_SS:
err_vec = (seg == VCPU_SREG_SS) ? SS_VECTOR : NP_VECTOR;
if (seg == VCPU_SREG_SS && selector == 3 &&
addr.seg = VCPU_SREG_SS;
addr.seg = VCPU_SREG_SS;
if (seg == VCPU_SREG_SS)
ops->set_segment(ctxt, ss_sel, &ss, 0, VCPU_SREG_SS);
ops->set_segment(ctxt, ss_sel, &ss, 0, VCPU_SREG_SS);
ops->set_segment(ctxt, ss_sel, &ss, 0, VCPU_SREG_SS);
tss->ss = get_segment_selector(ctxt, VCPU_SREG_SS);
set_segment_selector(ctxt, tss->ss, VCPU_SREG_SS);
ret = __load_segment_descriptor(ctxt, tss->ss, VCPU_SREG_SS, cpl,
tss->ss = get_segment_selector(ctxt, VCPU_SREG_SS);
set_segment_selector(ctxt, tss->ss, VCPU_SREG_SS);
ret = __load_segment_descriptor(ctxt, tss->ss, VCPU_SREG_SS, cpl,
if (ctxt->modrm_reg == VCPU_SREG_SS)
ctxt->ops->get_segment(ctxt, &sel, &ss, NULL, VCPU_SREG_SS);
op->val = VCPU_SREG_SS;
ctxt->seg_override = VCPU_SREG_SS;
if (addr.seg == VCPU_SREG_SS)
enter_smm_save_seg_32(vcpu, &smram->ss, &smram->ss_sel, VCPU_SREG_SS);
enter_smm_save_seg_64(vcpu, &smram->ss, VCPU_SREG_SS);
kvm_set_segment(vcpu, &ds, VCPU_SREG_SS);
rsm_load_seg_32(vcpu, &smstate->ss, smstate->ss_sel, VCPU_SREG_SS);
rsm_load_seg_64(vcpu, &smstate->ss, VCPU_SREG_SS);
case VCPU_SREG_SS: return &save->ss;
case VCPU_SREG_SS:
if (seg == VCPU_SREG_SS)
__vmx_set_segment(vcpu, &seg, VCPU_SREG_SS);
seg_reg == VCPU_SREG_SS ?
if (seg == VCPU_SREG_CS || seg == VCPU_SREG_SS)
vmx_get_segment(vcpu, &vmx->rmode.segs[VCPU_SREG_SS], VCPU_SREG_SS);
fix_pmode_seg(vcpu, VCPU_SREG_SS, &vmx->rmode.segs[VCPU_SREG_SS]);
vmx_get_segment(vcpu, &vmx->rmode.segs[VCPU_SREG_SS], VCPU_SREG_SS);
fix_rmode_seg(VCPU_SREG_SS, &vmx->rmode.segs[VCPU_SREG_SS]);
ar = vmx_read_guest_seg_ar(vmx, VCPU_SREG_SS);
vmx_get_segment(vcpu, &ss, VCPU_SREG_SS);
vmx_get_segment(vcpu, &ss, VCPU_SREG_SS);
if (!rmode_segment_valid(vcpu, VCPU_SREG_SS))
seg_setup(VCPU_SREG_SS);
kvm_get_segment(vcpu, &sregs->ss, VCPU_SREG_SS);
kvm_set_segment(vcpu, &sregs->ss, VCPU_SREG_SS);