Symbol: save_area
arch/powerpc/include/asm/lppaca.h
143
} save_area[SLB_NUM_BOLTED];
arch/powerpc/kernel/asm-offsets.c
226
OFFSET(SLBSHADOW_STACKVSID, slb_shadow, save_area[SLB_NUM_BOLTED - 1].vsid);
arch/powerpc/kernel/asm-offsets.c
227
OFFSET(SLBSHADOW_STACKESID, slb_shadow, save_area[SLB_NUM_BOLTED - 1].esid);
arch/powerpc/kernel/asm-offsets.c
228
OFFSET(SLBSHADOW_SAVEAREA, slb_shadow, save_area);
arch/powerpc/kvm/book3s_hv_ras.c
52
if ((void *) &slb->save_area[n] > vcpu->arch.slb_shadow.pinned_end)
arch/powerpc/kvm/book3s_hv_ras.c
57
unsigned long rb = be64_to_cpu(slb->save_area[i].esid);
arch/powerpc/kvm/book3s_hv_ras.c
58
unsigned long rs = be64_to_cpu(slb->save_area[i].vsid);
arch/powerpc/mm/book3s64/slb.c
126
: "r" (be64_to_cpu(p->save_area[index].vsid)),
arch/powerpc/mm/book3s64/slb.c
127
"r" (be64_to_cpu(p->save_area[index].esid)));
arch/powerpc/mm/book3s64/slb.c
172
ksp_esid_data = be64_to_cpu(p->save_area[KSTACK_INDEX].esid);
arch/powerpc/mm/book3s64/slb.c
173
ksp_vsid_data = be64_to_cpu(p->save_area[KSTACK_INDEX].vsid);
arch/powerpc/mm/book3s64/slb.c
86
WRITE_ONCE(p->save_area[index].esid, 0);
arch/powerpc/mm/book3s64/slb.c
87
WRITE_ONCE(p->save_area[index].vsid, cpu_to_be64(mk_vsid_data(ea, ssize, flags)));
arch/powerpc/mm/book3s64/slb.c
88
WRITE_ONCE(p->save_area[index].esid, cpu_to_be64(mk_esid_data(ea, ssize, index)));
arch/powerpc/mm/book3s64/slb.c
93
WRITE_ONCE(get_slb_shadow()->save_area[index].esid, cpu_to_be64(index));
arch/powerpc/xmon/xmon.c
2645
esid = be64_to_cpu(p->slb_shadow_ptr->save_area[i].esid);
arch/powerpc/xmon/xmon.c
2646
vsid = be64_to_cpu(p->slb_shadow_ptr->save_area[i].vsid);
arch/s390/include/asm/ipl.h
58
struct save_area;
arch/s390/include/asm/ipl.h
59
struct save_area * __init save_area_alloc(bool is_boot_cpu);
arch/s390/include/asm/ipl.h
60
struct save_area * __init save_area_boot_cpu(void);
arch/s390/include/asm/ipl.h
61
void __init save_area_add_regs(struct save_area *, void *regs);
arch/s390/include/asm/ipl.h
62
void __init save_area_add_vxrs(struct save_area *, __vector128 *vxrs);
arch/s390/include/asm/lowcore.h
102
__u64 save_area[8]; /* 0x0200 */
arch/s390/kernel/asm-offsets.c
119
OFFSET(__LC_SAVE_AREA, lowcore, save_area);
arch/s390/kernel/crash_dump.c
107
void __init save_area_add_vxrs(struct save_area *sa, __vector128 *vxrs)
arch/s390/kernel/crash_dump.c
296
static void *fill_cpu_elf_notes(void *ptr, int cpu, struct save_area *sa)
arch/s390/kernel/crash_dump.c
331
struct save_area *sa = NULL;
arch/s390/kernel/crash_dump.c
472
struct save_area *sa;
arch/s390/kernel/crash_dump.c
570
struct save_area *sa;
arch/s390/kernel/crash_dump.c
63
struct save_area * __init save_area_alloc(bool is_boot_cpu)
arch/s390/kernel/crash_dump.c
65
struct save_area *sa;
arch/s390/kernel/crash_dump.c
79
struct save_area * __init save_area_boot_cpu(void)
arch/s390/kernel/crash_dump.c
81
return list_first_entry_or_null(&dump_save_areas, struct save_area, list);
arch/s390/kernel/crash_dump.c
87
void __init save_area_add_regs(struct save_area *sa, void *regs)
arch/s390/kernel/smp.c
601
struct save_area *sa;
arch/s390/kernel/smp.c
618
struct save_area *sa;
arch/x86/kvm/svm/svm.c
477
return &sd->save_area->host_sev_es_save;
arch/x86/kvm/svm/svm.c
583
if (!sd->save_area)
arch/x86/kvm/svm/svm.c
589
sd->save_area = NULL;
arch/x86/kvm/svm/svm.c
607
sd->save_area = page_address(save_area_page);
arch/x86/kvm/svm/svm.h
351
struct vmcb *save_area;
drivers/s390/char/zcore.c
113
struct save_area *sa;
sound/pci/cs46xx/cs46xx_dsp_task_types.h
138
save_area,
tools/testing/selftests/kvm/include/x86/svm_util.h
22
struct vmcb_save_area *save_area; /* gva */
tools/testing/selftests/kvm/lib/x86/svm.c
40
svm->save_area = (void *)vm_vaddr_alloc_page(vm);
tools/testing/selftests/kvm/lib/x86/svm.c
41
svm->save_area_hva = addr_gva2hva(vm, (uintptr_t)svm->save_area);
tools/testing/selftests/kvm/lib/x86/svm.c
42
svm->save_area_gpa = addr_gva2gpa(vm, (uintptr_t)svm->save_area);