VCPU_REGS_RSP
if (base_reg == VCPU_REGS_RSP || base_reg == VCPU_REGS_RBP)
base_reg == VCPU_REGS_RSP)
addr.ea = reg_read(ctxt, VCPU_REGS_RSP) & stack_mask(ctxt);
addr.ea = reg_read(ctxt, VCPU_REGS_RSP) & stack_mask(ctxt);
assign_masked(reg_rmw(ctxt, VCPU_REGS_RBP), reg_read(ctxt, VCPU_REGS_RSP),
assign_masked(reg_rmw(ctxt, VCPU_REGS_RSP),
reg_read(ctxt, VCPU_REGS_RSP) - frame_size,
assign_masked(reg_rmw(ctxt, VCPU_REGS_RSP), reg_read(ctxt, VCPU_REGS_RBP),
unsigned long old_esp = reg_read(ctxt, VCPU_REGS_RSP);
(reg == VCPU_REGS_RSP) ?
if (reg == VCPU_REGS_RSP) {
*reg_write(ctxt, VCPU_REGS_RSP) = (efer & EFER_LMA) ? msr_data :
*reg_write(ctxt, VCPU_REGS_RSP) = rcx;
tss->sp = reg_read(ctxt, VCPU_REGS_RSP);
*reg_write(ctxt, VCPU_REGS_RSP) = tss->sp;
tss->esp = reg_read(ctxt, VCPU_REGS_RSP);
*reg_write(ctxt, VCPU_REGS_RSP) = tss->esp;
masked_increment(reg_rmw(ctxt, VCPU_REGS_RSP), stack_mask(ctxt), inc);
return kvm_register_read_raw(vcpu, VCPU_REGS_RSP);
kvm_register_write_raw(vcpu, VCPU_REGS_RSP, val);
save->rsp = svm->vcpu.arch.regs[VCPU_REGS_RSP];
svm->vmcb->save.rsp = vcpu->arch.regs[VCPU_REGS_RSP];
vcpu->arch.regs[VCPU_REGS_RSP] = svm->vmcb->save.rsp;
svm->vmcb->save.rsp = vcpu->arch.regs[VCPU_REGS_RSP];
case VCPU_REGS_RSP:
vcpu->arch.regs[VCPU_REGS_RSP] = vmcs_readl(GUEST_RSP);
if (kvm_register_is_dirty(vcpu, VCPU_REGS_RSP))
vmcs_writel(GUEST_RSP, vcpu->arch.regs[VCPU_REGS_RSP]);
(1 << VCPU_REGS_RSP) | \
kvm_register_mark_dirty(vcpu, VCPU_REGS_RSP);