get_paca
#define get_accounting(tsk) (&get_paca()->accounting)
get_paca()->ftrace_enabled = 0;
get_paca()->ftrace_enabled = 1;
get_paca()->ftrace_enabled = ftrace_enabled;
return get_paca()->ftrace_enabled;
get_paca()->irq_happened |= PACA_IRQ_PMI;
get_paca()->irq_happened &= ~PACA_IRQ_PMI;
if (get_paca()->irq_happened & PACA_IRQ_PMI)
return __lazy_irq_pending(get_paca()->irq_happened);
WARN_ON(!(get_paca()->irq_happened & PACA_IRQ_HARD_DIS));
if (get_paca()->irq_happened & PACA_IRQ_MUST_HARD_MASK)
get_paca()->irq_happened &= ~PACA_IRQ_HARD_DIS;
return &get_paca()->shadow_vcpu;
xirr = get_paca()->kvm_hstate.saved_xirr;
get_paca()->kvm_hstate.saved_xirr = 0;
#define get_lppaca() (get_paca()->lppaca_ptr)
get_paca()->pgd = NULL;
#define get_slb_shadow() (get_paca()->slb_shadow_ptr)
get_paca()->pmcregs_in_use = inuse;
return get_paca()->pmcregs_in_use;
#define LOCK_TOKEN (*(u32 *)(&get_paca()->lock_token))
#define LOCK_TOKEN (*(u32 *)(&get_paca()->paca_index))
#define hard_smp_processor_id() (get_paca()->hw_cpu_id)
get_paca()->canary = canary;
get_paca()->in_nmi = 1;
get_paca()->in_nmi = 0;
mce->cpu = get_paca()->paca_index;
if (get_paca()->in_mce < MAX_MCE_DEPTH) {
if (get_paca()->in_mce < MAX_MCE_DEPTH)
if (get_paca()->kvm_hstate.in_guest)
memcpy(&get_paca()->mm_ctx_low_slices_psize, mm_ctx_low_slices(context),
memcpy(&get_paca()->mm_ctx_high_slices_psize, mm_ctx_high_slices(context),
u64 dscr = get_paca()->dscr_default;
pr_cont("\nPACATMSCRATCH: %016llx ", get_paca()->tm_scratch);
get_paca()->kernel_msr = MSR_KERNEL;
*(unsigned long *)val = get_paca()->dscr_default;
get_paca()->dscr_default = *(unsigned long *)val;
regs->nip, regs->msr, get_paca()->tm_scratch);
BUG_ON(get_paca()->in_nmi == 0);
if (get_paca()->in_nmi > 1)
get_paca()->sprg_vdso = val;
get_paca()->kexec_state = KEXEC_STATE_IRQS_OFF;
get_paca()->kexec_state = KEXEC_STATE_IRQS_OFF;
memcpy(&kexec_paca, get_paca(), sizeof(struct paca_struct));
#define LOCK_TOKEN (*(u32 *)(&get_paca()->lock_token))
#define LOCK_TOKEN (*(u32 *)(&get_paca()->paca_index))
psizes = get_paca()->mm_ctx_low_slices_psize;
psizes = get_paca()->mm_ctx_high_slices_psize;
} else if (get_paca()->vmalloc_sllp !=
get_paca()->vmalloc_sllp =
get_paca()->slb_cache_ptr = 0;
get_paca()->slb_kern_bitmap = (1U << SLB_NUM_BOLTED) - 1;
get_paca()->slb_used_bitmap = get_paca()->slb_kern_bitmap;
assert_slb_presence(true, get_paca()->kstack);
get_paca()->slb_cache_ptr = 0;
get_paca()->slb_kern_bitmap = (1U << SLB_NUM_BOLTED) - 1;
get_paca()->slb_used_bitmap = get_paca()->slb_kern_bitmap;
get_paca()->slb_save_cache_ptr = get_paca()->slb_cache_ptr;
pr_err("SLB RR allocator index %d\n", get_paca()->stab_rr);
pr_err("SLB cache ptr value = %d\n", get_paca()->slb_save_cache_ptr);
n = min_t(int, get_paca()->slb_save_cache_ptr, SLB_CACHE_ENTRIES);
pr_err("%02d EA[0-35]=%9x\n", i, get_paca()->slb_cache[i]);
pr_err("%02d EA[0-35]=%9x\n", i, get_paca()->slb_cache[i]);
unsigned long slbie_data = get_paca()->slb_cache[index];
unsigned long ksp = get_paca()->kstack;
unsigned long slbie_data = get_paca()->slb_cache[index];
get_paca()->slb_cache_ptr = 0;
get_paca()->slb_kern_bitmap = (1U << SLB_NUM_BOLTED) - 1;
unsigned long offset = get_paca()->slb_cache_ptr;
get_paca()->slb_kern_bitmap = (1U << SLB_NUM_BOLTED) - 1;
get_paca()->slb_cache_ptr = 0;
get_paca()->slb_used_bitmap = get_paca()->slb_kern_bitmap;
get_paca()->vmalloc_sllp = SLB_VSID_KERNEL | vmalloc_llp;
get_paca()->stab_rr = SLB_NUM_BOLTED - 1;
get_paca()->slb_kern_bitmap = (1U << SLB_NUM_BOLTED) - 1;
get_paca()->slb_used_bitmap = get_paca()->slb_kern_bitmap;
(get_paca()->kstack & slb_esid_mask(mmu_kernel_ssize)) > PAGE_OFFSET)
create_shadowed_slbe(get_paca()->kstack,
get_paca()->pgd = mm->pgd;
get_paca()->tcd.esel_next = i;
get_paca()->tcd.esel_max = mfspr(SPRN_TLB1CFG) & TLBnCFG_N_ENTRY;
get_paca()->tcd.esel_first = i;
struct paca_struct *paca = get_paca();
struct paca_struct *paca = get_paca();
struct paca_struct *paca = get_paca();
get_paca()->kexec_state = KEXEC_STATE_REAL_MODE;
if (get_paca()->dtl_ridx != be64_to_cpu(get_lppaca()->dtl_idx))