mftb
#define MFTBL(dest) mftb dest
unsigned long tb = mftb();
unsigned long tb = mftb();
unsigned long __start = mftb(); \
#define MFTBL(dest) mftb dest
return mftb() - tstamp;
return mftb();
return mftb();
tblo = mftb();
choose_random_kstack_offset(mftb());
now = mftb();
start = mftb();
while (mftb() - start < loops)
vc->stolen_tb += mftb() - vcpu->arch.busy_preempt;
now = mftb();
kvmppc_core_start_stolen(vc, mftb());
kvmppc_core_end_stolen(vc, mftb());
kvmppc_remove_runnable(vc, vcpu, mftb());
kvmppc_core_end_stolen(pvc, mftb());
kvmppc_core_end_stolen(pvc, mftb());
kvmppc_remove_runnable(vc, vcpu, mftb());
kvmppc_core_start_stolen(vc, mftb());
vcpu->arch.busy_preempt = mftb();
now = mftb();
kvmppc_remove_runnable(vc, vcpu, mftb());
*tb = mftb();
*tb = mftb();
vcpu->arch.stolen_logged = vcore_stolen_time(vc, mftb());
kvmppc_remove_runnable(vc, v, mftb());
kvmppc_remove_runnable(vc, vcpu, mftb());
tb = mftb();
now = mftb();
u64 tb = mftb() - vc->tb_offset_applied;
if ((mftb() & 0xffffff) < (new_tb & 0xffffff)) {
*tb = mftb();
if ((mftb() & 0xffffff) < (new_tb & 0xffffff)) {
u64 new_tb = mftb() - vc->tb_offset_applied;
if ((mftb() & 0xffffff) < (new_tb & 0xffffff)) {
u64 new_tb = mftb() + vc->tb_offset;
if ((mftb() & 0xffffff) < (new_tb & 0xffffff)) {
if (mftb() & 0x1)
if (soft_invalid || (mftb() & 0x1))
if (mftb() & 0x1)
if (mftb() & 0x1)
slot_offset = mftb() & 0x7;
if (mftb() & 0x1)
long start = mftb();
now = mftb();
h->tb_total += mftb() - h->tb_start;
h->tb_start = mftb();
slot_offset = mftb() & 0x7;
seq_printf(m, "tbr=%ld\n", mftb());
cmd->init_time = mftb();
ms->wcol_tx_timestamp = mftb();
ms->wcol_ticks += mftb() - ms->wcol_tx_timestamp;
ms->wcol_tx_timestamp = mftb();
ms->timestamp = mftb();
ms->wcol_tx_timestamp = mftb();
if (((int)mftb()) - ms->timestamp < 0)
tb_start = mftb();
tb_now = mftb();
#define MFTBL(dest) mftb dest