kvmppc_get_pc
ulong pc = kvmppc_get_pc(vcpu);
regs->pc = kvmppc_get_pc(vcpu);
kvmppc_get_pc(vcpu), eaddr, vcpu_book3s->sdr1, pteg,
kvmppc_mmu_map_segment(vcpu, kvmppc_get_pc(vcpu));
kvmppc_set_pc(vcpu, kvmppc_get_pc(vcpu) + (is_prefixed ? 8 : 4));
vcpu->arch.tfiar = kvmppc_get_pc(vcpu);
vcpu->arch.tfiar = kvmppc_get_pc(vcpu);
kvmppc_set_pc(vcpu, kvmppc_get_pc(vcpu) + 4);
vcpu->arch.tfhar = kvmppc_get_pc(vcpu) + 4;
vcpu->arch.tfiar = kvmppc_get_pc(vcpu);
vcpu->run->debug.arch.address = kvmppc_get_pc(vcpu);
kvmppc_set_pc(vcpu, kvmppc_get_pc(vcpu) + 4);
vcpu->arch.trap, kvmppc_get_pc(vcpu),
vcpu->arch.fault_dar = kvmppc_get_pc(vcpu);
vcpu->arch.trap, kvmppc_get_pc(vcpu),
vcpu->arch.trap, kvmppc_get_pc(vcpu),
vcpu->arch.fault_dar = kvmppc_get_pc(vcpu);
dt->srr0 = cpu_to_be64(kvmppc_get_pc(vcpu));
pc = kvmppc_get_pc(vcpu);
kvmppc_get_pc(vcpu), ppc_inst_val(last_inst));
__func__, kvmppc_get_pc(vcpu), ppc_inst_val(last_inst));
pc = kvmppc_get_pc(vcpu);
sr = svcpu->sr[kvmppc_get_pc(vcpu) >> SID_SHIFT];
kvmppc_mmu_map_segment(vcpu, kvmppc_get_pc(vcpu));
r = kvmppc_handle_pagefault(vcpu, kvmppc_get_pc(vcpu), exit_nr);
kvmppc_mmu_pte_flush(vcpu, kvmppc_get_pc(vcpu), ~0xFFFUL);
if (kvmppc_mmu_map_segment(vcpu, kvmppc_get_pc(vcpu)) < 0) {
kvmppc_set_pc(vcpu, kvmppc_get_pc(vcpu) - 4);
exit_nr, kvmppc_get_pc(vcpu), shadow_srr1);
kvmppc_mmu_map_segment(vcpu, kvmppc_get_pc(vcpu));
ulong pc = kvmppc_get_pc(vcpu);
ulong pc = kvmppc_get_pc(vcpu);
geaddr = kvmppc_get_pc(vcpu);
vcpu->run->debug.arch.address = kvmppc_get_pc(vcpu);
trace_kvm_ppc_instr(inst, kvmppc_get_pc(vcpu), emulated);
kvmppc_set_pc(vcpu, kvmppc_get_pc(vcpu) + 4);
trace_kvm_ppc_instr(ppc_inst_val(inst), kvmppc_get_pc(vcpu), emulated);
kvmppc_set_pc(vcpu, kvmppc_get_pc(vcpu) + ppc_inst_len(inst));
__entry->pc = kvmppc_get_pc(vcpu);
__entry->pc = kvmppc_get_pc(vcpu);
__entry->pc = kvmppc_get_pc(vcpu);
__entry->pc = kvmppc_get_pc(vcpu);
__entry->pc = kvmppc_get_pc(vcpu);