guest_msr
void _kvmppc_restore_tm_pr(struct kvm_vcpu *vcpu, u64 guest_msr);
void _kvmppc_save_tm_pr(struct kvm_vcpu *vcpu, u64 guest_msr);
unsigned long guest_msr = kvmppc_get_msr(vcpu);
(((guest_msr & MSR_TS_MASK) >> (MSR_TS_S_LG - 1))
guest_msr &= ~(MSR_TS_MASK);
kvmppc_set_msr(vcpu, guest_msr);
unsigned long guest_msr = kvmppc_get_msr(vcpu);
guest_msr &= ~(MSR_TS_MASK);
guest_msr |= MSR_TS_S;
kvmppc_set_msr(vcpu, guest_msr);
unsigned long guest_msr = kvmppc_get_msr(vcpu);
(((guest_msr & MSR_TS_MASK) >> (MSR_TS_S_LG - 1))
MSR_TM_ACTIVE(guest_msr)) {
if (guest_msr & MSR_PR)
if (guest_msr & MSR_HV)
ulong guest_msr = kvmppc_get_msr(vcpu);
WARN_ON(guest_msr & MSR_PR);
ulong guest_msr = kvmppc_get_msr(vcpu);
if (guest_msr & MSR_PR) {
if (!MSR_TM_ACTIVE(guest_msr)) {
ulong guest_msr = kvmppc_get_msr(vcpu);
if (guest_msr & MSR_PR) {
if (MSR_TM_ACTIVE(guest_msr) ||
unsigned long guest_msr = vcpu->arch.shregs.msr;
if (MSR_TM_ACTIVE(guest_msr)) {
kvmppc_save_tm_hv(vcpu, guest_msr, true);
unsigned long guest_msr = vcpu->arch.shregs.msr;
if (MSR_TM_ACTIVE(guest_msr)) {
kvmppc_restore_tm_hv(vcpu, guest_msr, true);
ulong guest_msr = kvmppc_get_msr(vcpu);
ulong smsr = guest_msr;
smsr |= (guest_msr & vcpu->arch.guest_owned_ext);
if (!(guest_msr & MSR_PR))
vm = vm_create_with_one_vcpu(&vcpu, guest_msr);