x86_task_fpu
struct fpu *old_fpu = x86_task_fpu(old);
extern struct fpu *x86_task_fpu(struct task_struct *task);
struct fpu *fpu = x86_task_fpu(current);
fpuperm = &x86_task_fpu(current->group_leader)->guest_perm;
struct fpstate *fpstate = x86_task_fpu(current)->fpstate;
struct fpu *fpu = x86_task_fpu(current);
save_fpregs_to_fpstate(x86_task_fpu(current));
WARN_ON_FPU(fpu != x86_task_fpu(current));
struct fpu *src_fpu = x86_task_fpu(current->group_leader);
xstate = get_xsave_addr(&x86_task_fpu(dst)->fpstate->regs.xsave,
fpu = x86_task_fpu(tsk);
if (fpu == x86_task_fpu(current)) {
struct fpu *fpu = x86_task_fpu(current);
WARN_ON_FPU(fpu != x86_task_fpu(current));
fpstate_reset(x86_task_fpu(current));
struct fpu *fpu = x86_task_fpu(current);
struct fpu *fpu = x86_task_fpu(current);
sync_fpstate(x86_task_fpu(target));
struct fpu *fpu = x86_task_fpu(target);
struct fpu *fpu = x86_task_fpu(target);
struct fpu *fpu = x86_task_fpu(target);
__convert_from_fxsr(env, tsk, &x86_task_fpu(tsk)->fpstate->regs.fxsave);
struct fpu *fpu = x86_task_fpu(target);
struct fpu *fpu = x86_task_fpu(target);
if (fpu == x86_task_fpu(current))
WARN_ON_FPU(fpu == x86_task_fpu(current));
struct fpu *fpu = x86_task_fpu(target);
struct fpu *fpu = x86_task_fpu(target);
struct fpstate *fpstate = x86_task_fpu(tsk)->fpstate;
struct fpu *fpu = x86_task_fpu(current);
struct fpu *fpu = x86_task_fpu(tsk);
struct fpu *fpu = x86_task_fpu(current);
if (__get_user(magic2, (__u32 __user *)(fpstate + x86_task_fpu(current)->fpstate->user_size)))
unsigned long frame_size = xstate_sigframe_size(x86_task_fpu(current)->fpstate);
trace_x86_fpu_xstate_check_failed(x86_task_fpu(current));
struct xregs_state *xsave = &x86_task_fpu(tsk)->fpstate->regs.xsave;
fxsave(&x86_task_fpu(tsk)->fpstate->regs.fxsave);
__copy_xstate_to_uabi_buf(to, x86_task_fpu(tsk)->fpstate,
x86_task_fpu(tsk)->fpstate->user_xfeatures,
return copy_uabi_to_xstate(x86_task_fpu(tsk)->fpstate, NULL, ubuf, &tsk->thread.pkru);
if (fpstate->xfd == x86_task_fpu(current)->fpstate->xfd)
struct fpu *fpu = x86_task_fpu(current);
struct fpu *fpu = x86_task_fpu(current->group_leader);
fpu = x86_task_fpu(current->group_leader);
timestamp = READ_ONCE(x86_task_fpu(task)->avx512_timestamp);
fpstate_reset(x86_task_fpu(current));
wrmsrq(MSR_IA32_XFD, x86_task_fpu(current)->fpstate->xfd);
struct fpu *fpu = x86_task_fpu(current->group_leader);
struct fpstate *fpstate = x86_task_fpu(current)->fpstate;
xfd_validate_state(x86_task_fpu(current)->fpstate, mask, true);
fpstate_free(x86_task_fpu(tsk));
struct fpu *fpu = x86_task_fpu(current);
fsize += x86_task_fpu(current->group_leader)->perm.__user_state_size;
mask = x86_task_fpu(current->group_leader)->perm.__state_perm;
struct fpu *fpu = x86_task_fpu(task);
fpstate_init_soft(&x86_task_fpu(current)->fpstate->regs.soft);
struct swregs_state *s387 = &x86_task_fpu(target)->fpstate->regs.soft;
struct swregs_state *s387 = &x86_task_fpu(target)->fpstate->regs.soft;
#define I387 (&x86_task_fpu(current)->fpstate->regs)