Symbol: MSR_VEC
arch/powerpc/include/asm/switch_to.h
64
msr_check_and_clear(MSR_VEC);
arch/powerpc/include/asm/switch_to.h
85
msr_check_and_clear(MSR_FP|MSR_VEC|MSR_VSX);
arch/powerpc/kernel/interrupt.c
216
mathflags |= MSR_VEC | MSR_VSX;
arch/powerpc/kernel/interrupt.c
218
mathflags |= MSR_VEC;
arch/powerpc/kernel/process.c
1080
new->thread.regs->msr &= ~(MSR_FP | MSR_VEC | MSR_VSX);
arch/powerpc/kernel/process.c
1134
msr_diff &= MSR_FP | MSR_VEC | MSR_VSX;
arch/powerpc/kernel/process.c
1140
if (cpu_has_feature(CPU_FTR_ALTIVEC) && msr_diff & MSR_VEC)
arch/powerpc/kernel/process.c
1205
if (usermsr & MSR_VEC)
arch/powerpc/kernel/process.c
1453
{MSR_VEC, "VEC"},
arch/powerpc/kernel/process.c
1809
childregs->msr &= ~(MSR_FP|MSR_VEC|MSR_VSX);
arch/powerpc/kernel/process.c
240
msr &= ~MSR_VEC;
arch/powerpc/kernel/process.c
250
msr_check_and_set(MSR_VEC);
arch/powerpc/kernel/process.c
252
msr_check_and_clear(MSR_VEC);
arch/powerpc/kernel/process.c
262
cpumsr = msr_check_and_set(MSR_VEC);
arch/powerpc/kernel/process.c
264
if (current->thread.regs && (current->thread.regs->msr & MSR_VEC)) {
arch/powerpc/kernel/process.c
289
if (tsk->thread.regs->msr & MSR_VEC) {
arch/powerpc/kernel/process.c
308
WARN_ON((msr & MSR_VSX) && !((msr & MSR_FP) && (msr & MSR_VEC)));
arch/powerpc/kernel/process.c
313
if (msr & MSR_VEC)
arch/powerpc/kernel/process.c
321
msr_check_and_set(MSR_FP|MSR_VEC|MSR_VSX);
arch/powerpc/kernel/process.c
323
msr_check_and_clear(MSR_FP|MSR_VEC|MSR_VSX);
arch/powerpc/kernel/process.c
332
cpumsr = msr_check_and_set(MSR_FP|MSR_VEC|MSR_VSX);
arch/powerpc/kernel/process.c
335
(current->thread.regs->msr & (MSR_VSX|MSR_VEC|MSR_FP))) {
arch/powerpc/kernel/process.c
356
if (tsk->thread.regs->msr & (MSR_VSX|MSR_VEC|MSR_FP)) {
arch/powerpc/kernel/process.c
411
msr_all_available |= MSR_VEC;
arch/powerpc/kernel/process.c
437
WARN_ON((usermsr & MSR_VSX) && !((usermsr & MSR_FP) && (usermsr & MSR_VEC)));
arch/powerpc/kernel/process.c
441
if (usermsr & MSR_VEC)
arch/powerpc/kernel/process.c
530
if ((!(msr & MSR_VEC)) && should_restore_altivec())
arch/powerpc/kernel/process.c
531
new_msr |= MSR_VEC;
arch/powerpc/kernel/process.c
534
if (((msr | new_msr) & (MSR_FP | MSR_VEC)) == (MSR_FP | MSR_VEC))
arch/powerpc/kernel/process.c
550
if (new_msr & MSR_VEC)
arch/powerpc/kernel/process.c
577
WARN_ON((usermsr & MSR_VSX) && !((usermsr & MSR_FP) && (usermsr & MSR_VEC)));
arch/powerpc/kernel/process.c
582
if (usermsr & MSR_VEC)
arch/powerpc/kernel/process.c
970
if ((thr->ckpt_regs.msr & MSR_VEC) == 0)
arch/powerpc/kernel/signal_32.c
283
msr |= MSR_VEC;
arch/powerpc/kernel/signal_32.c
388
if (msr & MSR_VEC)
arch/powerpc/kernel/signal_32.c
400
msr |= MSR_VEC;
arch/powerpc/kernel/signal_32.c
410
if (msr & MSR_VEC)
arch/powerpc/kernel/signal_32.c
498
regs_set_return_msr(regs, regs->msr & ~MSR_VEC);
arch/powerpc/kernel/signal_32.c
499
if (msr & MSR_VEC) {
arch/powerpc/kernel/signal_32.c
598
regs_set_return_msr(regs, regs->msr & ~MSR_VEC);
arch/powerpc/kernel/signal_32.c
599
if (msr & MSR_VEC) {
arch/powerpc/kernel/signal_32.c
643
if (msr & MSR_VEC)
arch/powerpc/kernel/signal_32.c
700
msr_check_and_set(msr & (MSR_FP | MSR_VEC));
arch/powerpc/kernel/signal_32.c
705
if (msr & MSR_VEC) {
arch/powerpc/kernel/signal_32.c
707
regs_set_return_msr(regs, regs->msr | MSR_VEC);
arch/powerpc/kernel/signal_64.c
147
msr |= MSR_VEC;
arch/powerpc/kernel/signal_64.c
237
msr |= tsk->thread.ckpt_regs.msr & (MSR_FP | MSR_VEC | MSR_VSX);
arch/powerpc/kernel/signal_64.c
251
if (msr & MSR_VEC)
arch/powerpc/kernel/signal_64.c
263
msr |= MSR_VEC;
arch/powerpc/kernel/signal_64.c
271
if (msr & MSR_VEC)
arch/powerpc/kernel/signal_64.c
387
regs_set_return_msr(regs, regs->msr & ~(MSR_FP | MSR_FE0 | MSR_FE1 | MSR_VEC | MSR_VSX));
arch/powerpc/kernel/signal_64.c
394
if (v_regs != NULL && (msr & MSR_VEC) != 0) {
arch/powerpc/kernel/signal_64.c
506
regs_set_return_msr(regs, regs->msr & ~(MSR_FP | MSR_FE0 | MSR_FE1 | MSR_VEC | MSR_VSX));
arch/powerpc/kernel/signal_64.c
518
if (v_regs != NULL && tm_v_regs != NULL && (msr & MSR_VEC) != 0) {
arch/powerpc/kernel/signal_64.c
599
msr_check_and_set(msr & (MSR_FP | MSR_VEC));
arch/powerpc/kernel/signal_64.c
604
if (msr & MSR_VEC) {
arch/powerpc/kernel/signal_64.c
606
regs_set_return_msr(regs, regs->msr | MSR_VEC);
arch/powerpc/kernel/traps.c
976
msr_mask = MSR_VEC;
arch/powerpc/kvm/book3s_hv.c
5141
msr |= MSR_VEC;
arch/powerpc/kvm/book3s_hv_p9_entry.c
504
msr_needed |= MSR_VEC;
arch/powerpc/kvm/book3s_pr.c
1367
ext_msr = MSR_VEC;
arch/powerpc/kvm/book3s_pr.c
184
kvmppc_giveup_ext(vcpu, MSR_FP | MSR_VEC | MSR_VSX);
arch/powerpc/kvm/book3s_pr.c
1848
kvmppc_giveup_ext(vcpu, MSR_FP | MSR_VEC | MSR_VSX);
arch/powerpc/kvm/book3s_pr.c
368
(MSR_FP | MSR_VEC | MSR_VSX);
arch/powerpc/kvm/book3s_pr.c
375
else if (ext_diff == MSR_VEC)
arch/powerpc/kvm/book3s_pr.c
806
msr |= MSR_FP | MSR_VEC;
arch/powerpc/kvm/book3s_pr.c
828
if (msr & MSR_VEC) {
arch/powerpc/kvm/book3s_pr.c
829
if (current->thread.regs->msr & MSR_VEC)
arch/powerpc/kvm/book3s_pr.c
887
msr = MSR_FP | MSR_VEC | MSR_VSX;
arch/powerpc/kvm/book3s_pr.c
908
if (msr & MSR_VEC) {
arch/powerpc/kvm/book3s_pr.c
946
if (lost_ext & MSR_VEC) {
arch/powerpc/kvm/booke.c
202
if (!(current->thread.regs->msr & MSR_VEC)) {
arch/powerpc/kvm/booke.c
207
current->thread.regs->msr |= MSR_VEC;
arch/powerpc/kvm/booke.c
221
if (current->thread.regs->msr & MSR_VEC)
arch/powerpc/kvm/emulate_loadstore.c
272
MSR_VEC);
arch/powerpc/kvm/emulate_loadstore.c
54
if (!(kvmppc_get_msr(vcpu) & MSR_VEC)) {
arch/powerpc/kvm/powerpc.c
1210
vcpu->kvm->arch.kvm_ops->giveup_ext(vcpu, MSR_VEC);
arch/powerpc/lib/sstep.c
1034
if (regs->msr & MSR_VEC) {
arch/powerpc/lib/sstep.c
3443
if (!user_mode(regs) && !(regs->msr & MSR_VEC))
arch/powerpc/lib/sstep.c
3457
msrbit = MSR_VEC;
arch/powerpc/lib/sstep.c
3513
if (!user_mode(regs) && !(regs->msr & MSR_VEC))
arch/powerpc/lib/sstep.c
3527
msrbit = MSR_VEC;
arch/powerpc/lib/sstep.c
700
if (regs->msr & MSR_VEC)
arch/powerpc/lib/sstep.c
726
if (regs->msr & MSR_VEC)
arch/powerpc/lib/sstep.c
988
if (regs->msr & MSR_VEC) {
arch/powerpc/lib/test_emulate_step.c
71
regs->msr |= MSR_VEC;