MSR_TM_ACTIVE
if (!MSR_TM_ACTIVE(new->thread.regs->msr)){
if (!MSR_TM_ACTIVE(prev->thread.regs->msr) && prev->thread.load_tm == 0)
if (!MSR_TM_ACTIVE(regs->msr))
if (MSR_TM_ACTIVE(regs->msr))
if (!MSR_TM_ACTIVE(cpumsr) &&
MSR_TM_ACTIVE(current->thread.regs->msr))
if (!MSR_TM_ACTIVE(cpumsr) &&
MSR_TM_ACTIVE(current->thread.regs->msr))
if (!MSR_TM_ACTIVE(cpumsr) &&
MSR_TM_ACTIVE(current->thread.regs->msr))
MSR_TM_ACTIVE(tsk->thread.regs->msr) &&
if (!MSR_TM_ACTIVE(thr->regs->msr))
if (!MSR_TM_ACTIVE(target->thread.regs->msr))
if (!MSR_TM_ACTIVE(target->thread.regs->msr))
if (!MSR_TM_ACTIVE(target->thread.regs->msr))
if (!MSR_TM_ACTIVE(target->thread.regs->msr))
if (!MSR_TM_ACTIVE(target->thread.regs->msr))
if (!MSR_TM_ACTIVE(target->thread.regs->msr))
if (!MSR_TM_ACTIVE(target->thread.regs->msr))
if (!MSR_TM_ACTIVE(target->thread.regs->msr))
if (!MSR_TM_ACTIVE(target->thread.regs->msr))
if (!MSR_TM_ACTIVE(target->thread.regs->msr))
if (!MSR_TM_ACTIVE(target->thread.regs->msr))
if (MSR_TM_ACTIVE(target->thread.regs->msr))
if (!MSR_TM_ACTIVE(target->thread.regs->msr))
if (!MSR_TM_ACTIVE(target->thread.regs->msr))
if (MSR_TM_ACTIVE(target->thread.regs->msr))
if (!MSR_TM_ACTIVE(target->thread.regs->msr))
if (!MSR_TM_ACTIVE(target->thread.regs->msr))
if (MSR_TM_ACTIVE(target->thread.regs->msr))
if (!MSR_TM_ACTIVE(target->thread.regs->msr))
if (!MSR_TM_ACTIVE(target->thread.regs->msr))
if (!MSR_TM_ACTIVE(target->thread.regs->msr))
if (MSR_TM_ACTIVE(regs->msr)) {
if (MSR_TM_ACTIVE(msr_hi<<32)) {
if (MSR_TM_ACTIVE(msr_hi<<32)) {
if (MSR_TM_ACTIVE(msr))
if (MSR_TM_ACTIVE(msr)) {
if (MSR_TM_ACTIVE(msr))
if (MSR_TM_ACTIVE(msr))
BUG_ON(!MSR_TM_ACTIVE(msr));
if (IS_ENABLED(CONFIG_PPC_TRANSACTIONAL_MEM) && MSR_TM_ACTIVE(msr)) {
if (!MSR_TM_ACTIVE(msr))
if (MSR_TM_ACTIVE(msr)) {
MSR_TM_ACTIVE(guest_msr)) {
!MSR_TM_ACTIVE(srr1))
if (!MSR_TM_ACTIVE(guest_msr)) {
if (MSR_TM_ACTIVE(guest_msr) ||
if (MSR_TM_ACTIVE(kvmppc_get_msr(vcpu)) &&
if (MSR_TM_ACTIVE(current->thread.regs->msr)) {
if (!MSR_TM_ACTIVE(l2_regs.msr))
if (MSR_TM_ACTIVE(guest_msr)) {
if (MSR_TM_ACTIVE(guest_msr)) {
if (!MSR_TM_ACTIVE(msr)) {
if (MSR_TM_ACTIVE(msr) || !(vcpu->arch.texasr & TEXASR_FS)) {
if (!(MSR_TM_ACTIVE(kvmppc_get_msr(vcpu)))) {
if (!MSR_TM_ACTIVE(kvmppc_get_msr(vcpu))) {
MSR_TM_ACTIVE(current->thread.regs->msr)) {