Symbol: MSR_VSX
arch/powerpc/include/asm/switch_to.h
85
msr_check_and_clear(MSR_FP|MSR_VEC|MSR_VSX);
arch/powerpc/kernel/interrupt.c
216
mathflags |= MSR_VEC | MSR_VSX;
arch/powerpc/kernel/process.c
1080
new->thread.regs->msr &= ~(MSR_FP | MSR_VEC | MSR_VSX);
arch/powerpc/kernel/process.c
1134
msr_diff &= MSR_FP | MSR_VEC | MSR_VSX;
arch/powerpc/kernel/process.c
125
newmsr |= MSR_VSX;
arch/powerpc/kernel/process.c
143
newmsr &= ~MSR_VSX;
arch/powerpc/kernel/process.c
1454
{MSR_VSX, "VSX"},
arch/powerpc/kernel/process.c
159
msr &= ~MSR_VSX;
arch/powerpc/kernel/process.c
1809
childregs->msr &= ~(MSR_FP|MSR_VEC|MSR_VSX);
arch/powerpc/kernel/process.c
242
msr &= ~MSR_VSX;
arch/powerpc/kernel/process.c
308
WARN_ON((msr & MSR_VSX) && !((msr & MSR_FP) && (msr & MSR_VEC)));
arch/powerpc/kernel/process.c
321
msr_check_and_set(MSR_FP|MSR_VEC|MSR_VSX);
arch/powerpc/kernel/process.c
323
msr_check_and_clear(MSR_FP|MSR_VEC|MSR_VSX);
arch/powerpc/kernel/process.c
332
cpumsr = msr_check_and_set(MSR_FP|MSR_VEC|MSR_VSX);
arch/powerpc/kernel/process.c
335
(current->thread.regs->msr & (MSR_VSX|MSR_VEC|MSR_FP))) {
arch/powerpc/kernel/process.c
356
if (tsk->thread.regs->msr & (MSR_VSX|MSR_VEC|MSR_FP)) {
arch/powerpc/kernel/process.c
413
msr_all_available |= MSR_VSX;
arch/powerpc/kernel/process.c
437
WARN_ON((usermsr & MSR_VSX) && !((usermsr & MSR_FP) && (usermsr & MSR_VEC)));
arch/powerpc/kernel/process.c
533
if ((!(msr & MSR_VSX)) && should_restore_vsx()) {
arch/powerpc/kernel/process.c
535
new_msr |= MSR_VSX;
arch/powerpc/kernel/process.c
553
if (new_msr & MSR_VSX)
arch/powerpc/kernel/process.c
577
WARN_ON((usermsr & MSR_VSX) && !((usermsr & MSR_FP) && (usermsr & MSR_VEC)));
arch/powerpc/kernel/signal_32.c
1029
(new_msr & MSR_VSX))
arch/powerpc/kernel/signal_32.c
302
msr &= ~MSR_VSX;
arch/powerpc/kernel/signal_32.c
312
msr |= MSR_VSX;
arch/powerpc/kernel/signal_32.c
431
if (msr & MSR_VSX)
arch/powerpc/kernel/signal_32.c
436
msr |= MSR_VSX;
arch/powerpc/kernel/signal_32.c
520
regs_set_return_msr(regs, regs->msr & ~MSR_VSX);
arch/powerpc/kernel/signal_32.c
521
if (msr & MSR_VSX) {
arch/powerpc/kernel/signal_32.c
621
regs_set_return_msr(regs, regs->msr & ~MSR_VSX);
arch/powerpc/kernel/signal_32.c
622
if (msr & MSR_VSX) {
arch/powerpc/kernel/signal_32.c
653
if (msr & MSR_VSX) {
arch/powerpc/kernel/signal_64.c
163
msr &= ~MSR_VSX;
arch/powerpc/kernel/signal_64.c
176
msr |= MSR_VSX;
arch/powerpc/kernel/signal_64.c
237
msr |= tsk->thread.ckpt_regs.msr & (MSR_FP | MSR_VEC | MSR_VSX);
arch/powerpc/kernel/signal_64.c
302
if (msr & MSR_VSX)
arch/powerpc/kernel/signal_64.c
310
msr |= MSR_VSX;
arch/powerpc/kernel/signal_64.c
387
regs_set_return_msr(regs, regs->msr & ~(MSR_FP | MSR_FE0 | MSR_FE1 | MSR_VEC | MSR_VSX));
arch/powerpc/kernel/signal_64.c
418
if ((msr & MSR_VSX) != 0) {
arch/powerpc/kernel/signal_64.c
506
regs_set_return_msr(regs, regs->msr & ~(MSR_FP | MSR_FE0 | MSR_FE1 | MSR_VEC | MSR_VSX));
arch/powerpc/kernel/signal_64.c
552
if (v_regs && ((msr & MSR_VSX) != 0)) {
arch/powerpc/kernel/signal_64.c
678
(new_msr & MSR_VSX))
arch/powerpc/kernel/traps.c
974
msr_mask = MSR_VSX;
arch/powerpc/kvm/book3s_emulate.c
177
kvmppc_giveup_ext(vcpu, MSR_VSX);
arch/powerpc/kvm/book3s_hv.c
5143
msr |= MSR_VSX;
arch/powerpc/kvm/book3s_hv_p9_entry.c
506
msr_needed |= MSR_VSX;
arch/powerpc/kvm/book3s_pr.c
1371
ext_msr = MSR_VSX;
arch/powerpc/kvm/book3s_pr.c
184
kvmppc_giveup_ext(vcpu, MSR_FP | MSR_VEC | MSR_VSX);
arch/powerpc/kvm/book3s_pr.c
1848
kvmppc_giveup_ext(vcpu, MSR_FP | MSR_VEC | MSR_VSX);
arch/powerpc/kvm/book3s_pr.c
368
(MSR_FP | MSR_VEC | MSR_VSX);
arch/powerpc/kvm/book3s_pr.c
391
kvmppc_giveup_ext(vcpu, MSR_VSX);
arch/powerpc/kvm/book3s_pr.c
805
if (msr & MSR_VSX)
arch/powerpc/kvm/book3s_pr.c
835
vcpu->arch.guest_owned_ext &= ~(msr | MSR_VSX);
arch/powerpc/kvm/book3s_pr.c
873
if (msr == MSR_VSX) {
arch/powerpc/kvm/book3s_pr.c
887
msr = MSR_FP | MSR_VEC | MSR_VSX;
arch/powerpc/kvm/emulate_loadstore.c
317
MSR_VSX);
arch/powerpc/kvm/emulate_loadstore.c
42
if (!(kvmppc_get_msr(vcpu) & MSR_VSX)) {
arch/powerpc/kvm/powerpc.c
1193
vcpu->kvm->arch.kvm_ops->giveup_ext(vcpu, MSR_VSX);
arch/powerpc/lib/sstep.c
3450
unsigned long msrbit = MSR_VSX;
arch/powerpc/lib/sstep.c
3520
unsigned long msrbit = MSR_VSX;
arch/powerpc/lib/test_emulate_step.c
72
regs->msr |= MSR_VSX;