arch/sh/include/asm/processor_32.h
108
union thread_xstate *xstate;
arch/sh/kernel/cpu/fpu.c
22
if (!tsk->thread.xstate) {
arch/sh/kernel/cpu/fpu.c
23
tsk->thread.xstate = kmem_cache_alloc(task_xstate_cachep,
arch/sh/kernel/cpu/fpu.c
25
if (!tsk->thread.xstate)
arch/sh/kernel/cpu/fpu.c
30
struct sh_fpu_hard_struct *fp = &tsk->thread.xstate->hardfpu;
arch/sh/kernel/cpu/fpu.c
34
struct sh_fpu_soft_struct *fp = &tsk->thread.xstate->softfpu;
arch/sh/kernel/cpu/sh2a/fpu.c
456
if ((tsk->thread.xstate->hardfpu.fpscr & FPSCR_FPU_ERROR)) {
arch/sh/kernel/cpu/sh2a/fpu.c
458
denormal_to_double (&tsk->thread.xstate->hardfpu,
arch/sh/kernel/cpu/sh2a/fpu.c
473
hx = tsk->thread.xstate->hardfpu.fp_regs[n];
arch/sh/kernel/cpu/sh2a/fpu.c
474
hy = tsk->thread.xstate->hardfpu.fp_regs[m];
arch/sh/kernel/cpu/sh2a/fpu.c
475
fpscr = tsk->thread.xstate->hardfpu.fpscr;
arch/sh/kernel/cpu/sh2a/fpu.c
485
| tsk->thread.xstate->hardfpu.fp_regs[n+1];
arch/sh/kernel/cpu/sh2a/fpu.c
487
| tsk->thread.xstate->hardfpu.fp_regs[m+1];
arch/sh/kernel/cpu/sh2a/fpu.c
492
tsk->thread.xstate->hardfpu.fp_regs[n] = llx >> 32;
arch/sh/kernel/cpu/sh2a/fpu.c
493
tsk->thread.xstate->hardfpu.fp_regs[n+1] = llx & 0xffffffff;
arch/sh/kernel/cpu/sh2a/fpu.c
502
tsk->thread.xstate->hardfpu.fp_regs[n] = hx;
arch/sh/kernel/cpu/sh2a/fpu.c
516
hx = tsk->thread.xstate->hardfpu.fp_regs[n];
arch/sh/kernel/cpu/sh2a/fpu.c
517
hy = tsk->thread.xstate->hardfpu.fp_regs[m];
arch/sh/kernel/cpu/sh2a/fpu.c
518
fpscr = tsk->thread.xstate->hardfpu.fpscr;
arch/sh/kernel/cpu/sh2a/fpu.c
52
: "0" ((char *)(&tsk->thread.xstate->hardfpu.status)),
arch/sh/kernel/cpu/sh2a/fpu.c
528
| tsk->thread.xstate->hardfpu.fp_regs[n+1];
arch/sh/kernel/cpu/sh2a/fpu.c
530
| tsk->thread.xstate->hardfpu.fp_regs[m+1];
arch/sh/kernel/cpu/sh2a/fpu.c
535
tsk->thread.xstate->hardfpu.fp_regs[n] = llx >> 32;
arch/sh/kernel/cpu/sh2a/fpu.c
536
tsk->thread.xstate->hardfpu.fp_regs[n+1] = llx & 0xffffffff;
arch/sh/kernel/cpu/sh2a/fpu.c
545
tsk->thread.xstate->hardfpu.fp_regs[n] = hx;
arch/sh/kernel/cpu/sh2a/fpu.c
563
tsk->thread.xstate->hardfpu.fpscr &=
arch/sh/kernel/cpu/sh2a/fpu.c
84
: "0" (tsk->thread.xstate), "r" (FPSCR_RCHG)
arch/sh/kernel/cpu/sh4/fpu.c
134
:"0" (tsk->thread.xstate), "r" (FPSCR_RCHG)
arch/sh/kernel/cpu/sh4/fpu.c
230
if ((tsk->thread.xstate->hardfpu.fpscr & FPSCR_CAUSE_ERROR))
arch/sh/kernel/cpu/sh4/fpu.c
232
denormal_to_double(&tsk->thread.xstate->hardfpu,
arch/sh/kernel/cpu/sh4/fpu.c
248
hx = tsk->thread.xstate->hardfpu.fp_regs[n];
arch/sh/kernel/cpu/sh4/fpu.c
249
hy = tsk->thread.xstate->hardfpu.fp_regs[m];
arch/sh/kernel/cpu/sh4/fpu.c
250
fpscr = tsk->thread.xstate->hardfpu.fpscr;
arch/sh/kernel/cpu/sh4/fpu.c
260
| tsk->thread.xstate->hardfpu.fp_regs[n + 1];
arch/sh/kernel/cpu/sh4/fpu.c
262
| tsk->thread.xstate->hardfpu.fp_regs[m + 1];
arch/sh/kernel/cpu/sh4/fpu.c
264
tsk->thread.xstate->hardfpu.fp_regs[n] = llx >> 32;
arch/sh/kernel/cpu/sh4/fpu.c
265
tsk->thread.xstate->hardfpu.fp_regs[n + 1] = llx & 0xffffffff;
arch/sh/kernel/cpu/sh4/fpu.c
271
tsk->thread.xstate->hardfpu.fp_regs[n] = hx;
arch/sh/kernel/cpu/sh4/fpu.c
286
hx = tsk->thread.xstate->hardfpu.fp_regs[n];
arch/sh/kernel/cpu/sh4/fpu.c
287
hy = tsk->thread.xstate->hardfpu.fp_regs[m];
arch/sh/kernel/cpu/sh4/fpu.c
288
fpscr = tsk->thread.xstate->hardfpu.fpscr;
arch/sh/kernel/cpu/sh4/fpu.c
298
| tsk->thread.xstate->hardfpu.fp_regs[n + 1];
arch/sh/kernel/cpu/sh4/fpu.c
300
| tsk->thread.xstate->hardfpu.fp_regs[m + 1];
arch/sh/kernel/cpu/sh4/fpu.c
305
tsk->thread.xstate->hardfpu.fp_regs[n] = llx >> 32;
arch/sh/kernel/cpu/sh4/fpu.c
306
tsk->thread.xstate->hardfpu.fp_regs[n + 1] = llx & 0xffffffff;
arch/sh/kernel/cpu/sh4/fpu.c
315
tsk->thread.xstate->hardfpu.fp_regs[n] = hx;
arch/sh/kernel/cpu/sh4/fpu.c
330
hx = tsk->thread.xstate->hardfpu.fp_regs[n];
arch/sh/kernel/cpu/sh4/fpu.c
331
hy = tsk->thread.xstate->hardfpu.fp_regs[m];
arch/sh/kernel/cpu/sh4/fpu.c
332
fpscr = tsk->thread.xstate->hardfpu.fpscr;
arch/sh/kernel/cpu/sh4/fpu.c
342
| tsk->thread.xstate->hardfpu.fp_regs[n + 1];
arch/sh/kernel/cpu/sh4/fpu.c
344
| tsk->thread.xstate->hardfpu.fp_regs[m + 1];
arch/sh/kernel/cpu/sh4/fpu.c
348
tsk->thread.xstate->hardfpu.fp_regs[n] = llx >> 32;
arch/sh/kernel/cpu/sh4/fpu.c
349
tsk->thread.xstate->hardfpu.fp_regs[n + 1] = llx & 0xffffffff;
arch/sh/kernel/cpu/sh4/fpu.c
355
tsk->thread.xstate->hardfpu.fp_regs[n] = hx;
arch/sh/kernel/cpu/sh4/fpu.c
368
hx = tsk->thread.xstate->hardfpu.fp_regs[m];
arch/sh/kernel/cpu/sh4/fpu.c
370
if ((tsk->thread.xstate->hardfpu.fpscr & FPSCR_CAUSE_ERROR)
arch/sh/kernel/cpu/sh4/fpu.c
375
llx = ((long long)tsk->thread.xstate->hardfpu.fp_regs[m] << 32)
arch/sh/kernel/cpu/sh4/fpu.c
376
| tsk->thread.xstate->hardfpu.fp_regs[m + 1];
arch/sh/kernel/cpu/sh4/fpu.c
378
tsk->thread.xstate->hardfpu.fpul = float64_to_float32(llx);
arch/sh/kernel/cpu/sh4/fpu.c
397
int roundingMode = FPSCR_ROUNDING_MODE(tsk->thread.xstate->hardfpu.fpscr);
arch/sh/kernel/cpu/sh4/fpu.c
409
tsk->thread.xstate->hardfpu.fpscr &=
arch/sh/kernel/cpu/sh4/fpu.c
411
tsk->thread.xstate->hardfpu.fpscr |= fpu_exception_flags;
arch/sh/kernel/cpu/sh4/fpu.c
414
tsk->thread.xstate->hardfpu.fpscr |= (fpu_exception_flags >> 10);
arch/sh/kernel/cpu/sh4/fpu.c
418
if ((((tsk->thread.xstate->hardfpu.fpscr & FPSCR_ENABLE_MASK) >> 7) &
arch/sh/kernel/cpu/sh4/fpu.c
84
:"0"((char *)(&tsk->thread.xstate->hardfpu.status)),
arch/sh/kernel/process.c
29
if (src->thread.xstate) {
arch/sh/kernel/process.c
30
dst->thread.xstate = kmem_cache_alloc(task_xstate_cachep,
arch/sh/kernel/process.c
32
if (!dst->thread.xstate)
arch/sh/kernel/process.c
34
memcpy(dst->thread.xstate, src->thread.xstate, xstate_size);
arch/sh/kernel/process.c
42
if (tsk->thread.xstate) {
arch/sh/kernel/process.c
43
kmem_cache_free(task_xstate_cachep, tsk->thread.xstate);
arch/sh/kernel/process.c
44
tsk->thread.xstate = NULL;
arch/sh/kernel/process_32.c
157
prefetch(next_t->xstate);
arch/sh/kernel/ptrace_32.c
177
return membuf_write(&to, target->thread.xstate,
arch/sh/kernel/ptrace_32.c
196
&target->thread.xstate->hardfpu, 0, -1);
arch/sh/kernel/ptrace_32.c
199
&target->thread.xstate->softfpu, 0, -1);
arch/sh/kernel/ptrace_32.c
369
tmp = ((unsigned long *)child->thread.xstate)
arch/sh/kernel/ptrace_32.c
404
((unsigned long *)child->thread.xstate)
arch/sh/kernel/signal_32.c
109
return __copy_to_user(&sc->sc_fpregs[0], &tsk->thread.xstate->hardfpu,
arch/sh/kernel/signal_32.c
85
return __copy_from_user(&tsk->thread.xstate->hardfpu, &sc->sc_fpregs[0],
arch/sh/math-emu/math.c
497
struct sh_fpu_soft_struct *fpu = &(tsk->thread.xstate->softfpu);
arch/x86/include/asm/perf_event.h
593
u64 xstate;
arch/x86/kernel/fpu/core.c
637
struct cet_user_state *xstate;
arch/x86/kernel/fpu/core.c
643
xstate = get_xsave_addr(&x86_task_fpu(dst)->fpstate->regs.xsave,
arch/x86/kernel/fpu/core.c
652
if (WARN_ON_ONCE(!xstate))
arch/x86/kernel/fpu/core.c
655
xstate->user_ssp = (u64)ssp;
arch/x86/kernel/fpu/xstate.c
1125
static void copy_feature(bool from_xstate, struct membuf *to, void *xstate,
arch/x86/kernel/fpu/xstate.c
1128
membuf_write(to, from_xstate ? xstate : init_xstate, size);
arch/x86/kernel/fpu/xstate.c
1443
void xsaves(struct xregs_state *xstate, u64 mask)
arch/x86/kernel/fpu/xstate.c
1450
XSTATE_OP(XSAVES, xstate, (u32)mask, (u32)(mask >> 32), err);
arch/x86/kernel/fpu/xstate.c
1467
void xrstors(struct xregs_state *xstate, u64 mask)
arch/x86/kernel/fpu/xstate.c
1474
XSTATE_OP(XRSTORS, xstate, (u32)mask, (u32)(mask >> 32), err);
arch/x86/kernel/fpu/xstate.c
339
static __init void os_xrstor_booting(struct xregs_state *xstate)
arch/x86/kernel/fpu/xstate.c
347
XSTATE_OP(XRSTORS, xstate, lmask, hmask, err);
arch/x86/kernel/fpu/xstate.c
349
XSTATE_OP(XRSTOR, xstate, lmask, hmask, err);
arch/x86/kernel/fpu/xstate.h
331
struct xregs_state *xstate = ((__force struct xregs_state *)buf);
arch/x86/kernel/fpu/xstate.h
339
XSTATE_OP(XRSTOR, xstate, lmask, hmask, err);
arch/x86/kernel/fpu/xstate.h
351
struct xregs_state *xstate = &fpstate->regs.xsave;
arch/x86/kernel/fpu/xstate.h
360
XSTATE_OP(XRSTORS, xstate, lmask, hmask, err);
arch/x86/kernel/fpu/xstate.h
362
XSTATE_OP(XRSTOR, xstate, lmask, hmask, err);
arch/x86/kvm/x86.c
5825
union fpregs_state *xstate = (union fpregs_state *)guest_xsave->region;
arch/x86/kvm/x86.c
5835
xstate->xsave.header.xfeatures &= ~vcpu->arch.guest_fpu.fpstate->xfd;
drivers/net/ppp/ppp_generic.c
132
unsigned int xstate; /* transmit state bits 68 */
drivers/net/ppp/ppp_generic.c
1848
if ((ppp->xstate & SC_COMP_RUN) && ppp->xc_state &&
drivers/net/ppp/ppp_generic.c
3094
ppp->xstate &= ~SC_COMP_RUN;
drivers/net/ppp/ppp_generic.c
3157
ppp->xstate &= ~SC_COMP_RUN;
drivers/net/ppp/ppp_generic.c
3170
ppp->xstate &= ~SC_COMP_RUN;
drivers/net/ppp/ppp_generic.c
3198
ppp->xstate |= SC_COMP_RUN;
drivers/net/ppp/ppp_generic.c
3212
if (ppp->xc_state && (ppp->xstate & SC_COMP_RUN))
drivers/net/ppp/ppp_generic.c
3223
void *xstate, *rstate;
drivers/net/ppp/ppp_generic.c
3228
ppp->xstate = 0;
drivers/net/ppp/ppp_generic.c
3230
xstate = ppp->xc_state;
drivers/net/ppp/ppp_generic.c
3238
if (xstate) {
drivers/net/ppp/ppp_generic.c
3239
xcomp->comp_free(xstate);
drivers/net/ppp/ppp_generic.c
842
val = ppp->flags | ppp->xstate | ppp->rstate;
tools/testing/selftests/kvm/x86/amx_test.c
149
struct xstate *xstate)
tools/testing/selftests/kvm/x86/amx_test.c
183
xstate->header.xstate_bv = XFEATURE_MASK_XTILE_DATA;
tools/testing/selftests/kvm/x86/amx_test.c
184
__xsavec(xstate, XFEATURE_MASK_XTILE_DATA);
tools/testing/selftests/kvm/x86/amx_test.c
185
GUEST_ASSERT(!(xstate->header.xstate_bv & XFEATURE_MASK_XTILE_DATA));
tools/testing/selftests/kvm/x86/amx_test.c
186
GUEST_ASSERT(xstate->header.xcomp_bv & XFEATURE_MASK_XTILE_DATA);
tools/testing/selftests/kvm/x86/amx_test.c
197
xstate->header.xstate_bv = XFEATURE_MASK_XTILE_DATA;
tools/testing/selftests/kvm/x86/amx_test.c
198
__xsavec(xstate, XFEATURE_MASK_XTILE_DATA);
tools/testing/selftests/kvm/x86/amx_test.c
199
GUEST_ASSERT(!(xstate->header.xstate_bv & XFEATURE_MASK_XTILE_DATA));
tools/testing/selftests/kvm/x86/amx_test.c
200
GUEST_ASSERT((xstate->header.xcomp_bv & XFEATURE_MASK_XTILE_DATA));
tools/testing/selftests/kvm/x86/amx_test.c
239
vm_vaddr_t amx_cfg, tiledata, xstate;
tools/testing/selftests/kvm/x86/amx_test.c
274
xstate = vm_vaddr_alloc_pages(vm, DIV_ROUND_UP(XSAVE_SIZE, PAGE_SIZE));
tools/testing/selftests/kvm/x86/amx_test.c
275
memset(addr_gva2hva(vm, xstate), 0, PAGE_SIZE * DIV_ROUND_UP(XSAVE_SIZE, PAGE_SIZE));
tools/testing/selftests/kvm/x86/amx_test.c
276
vcpu_args_set(vcpu, 3, amx_cfg, tiledata, xstate);
tools/testing/selftests/kvm/x86/amx_test.c
83
static inline void __xsavec(struct xstate *xstate, uint64_t rfbm)
tools/testing/selftests/kvm/x86/amx_test.c
89
: : "D" (xstate), "a" (rfbm_lo), "d" (rfbm_hi)
tools/testing/selftests/x86/xstate.c
101
xsave(xbuf2, xstate.mask);
tools/testing/selftests/x86/xstate.c
124
load_rand_xstate(&xstate, xbuf);
tools/testing/selftests/x86/xstate.c
137
load_rand_xstate(&xstate, xbuf);
tools/testing/selftests/x86/xstate.c
221
xstate.name, iterations, num_threads);
tools/testing/selftests/x86/xstate.c
265
load_rand_xstate(&xstate, xbuf);
tools/testing/selftests/x86/xstate.c
266
load_init_xstate(&xstate, xbuf);
tools/testing/selftests/x86/xstate.c
296
printf("[RUN]\t%s: inject xstate via ptrace().\n", xstate.name);
tools/testing/selftests/x86/xstate.c
298
load_rand_xstate(&xstate, xbuf1);
tools/testing/selftests/x86/xstate.c
387
if (get_fpx_sw_bytes_features(xbuf) & xstate.mask)
tools/testing/selftests/x86/xstate.c
392
if (get_xstatebv(xbuf) & xstate.mask)
tools/testing/selftests/x86/xstate.c
408
set_rand_data(&xstate, xbuf);
tools/testing/selftests/x86/xstate.c
424
printf("[RUN]\t%s: load xstate and raise SIGUSR1\n", xstate.name);
tools/testing/selftests/x86/xstate.c
428
load_rand_xstate(&xstate, stashed_xbuf);
tools/testing/selftests/x86/xstate.c
440
xstate.name);
tools/testing/selftests/x86/xstate.c
469
xstate = get_xstate_info(feature_num);
tools/testing/selftests/x86/xstate.c
470
if (!xstate.size || !xstate.xbuf_offset) {
tools/testing/selftests/x86/xstate.c
472
xstate.size, xstate.xbuf_offset);
tools/testing/selftests/x86/xstate.c
50
static struct xstate_info xstate;
tools/testing/selftests/x86/xstate.c
61
static inline void load_rand_xstate(struct xstate_info *xstate, struct xsave_buffer *xbuf)
tools/testing/selftests/x86/xstate.c
64
set_xstatebv(xbuf, xstate->mask);
tools/testing/selftests/x86/xstate.c
65
set_rand_data(xstate, xbuf);
tools/testing/selftests/x86/xstate.c
66
xrstor(xbuf, xstate->mask);
tools/testing/selftests/x86/xstate.c
69
static inline void load_init_xstate(struct xstate_info *xstate, struct xsave_buffer *xbuf)
tools/testing/selftests/x86/xstate.c
72
xrstor(xbuf, xstate->mask);
tools/testing/selftests/x86/xstate.c
77
memcpy(&xbuf_dst->bytes[xstate.xbuf_offset],
tools/testing/selftests/x86/xstate.c
78
&xbuf_src->bytes[xstate.xbuf_offset],
tools/testing/selftests/x86/xstate.c
79
xstate.size);
tools/testing/selftests/x86/xstate.c
86
ret = memcmp(&xbuf1->bytes[xstate.xbuf_offset],
tools/testing/selftests/x86/xstate.c
87
&xbuf2->bytes[xstate.xbuf_offset],
tools/testing/selftests/x86/xstate.c
88
xstate.size);
tools/testing/selftests/x86/xstate.h
125
struct xstate_info xstate = { };
tools/testing/selftests/x86/xstate.h
130
return xstate;
tools/testing/selftests/x86/xstate.h
133
xstate.name = xfeature_names[xfeature_num];
tools/testing/selftests/x86/xstate.h
134
xstate.num = xfeature_num;
tools/testing/selftests/x86/xstate.h
135
xstate.mask = 1 << xfeature_num;
tools/testing/selftests/x86/xstate.h
139
xstate.size = eax;
tools/testing/selftests/x86/xstate.h
140
xstate.xbuf_offset = ebx;
tools/testing/selftests/x86/xstate.h
141
return xstate;
tools/testing/selftests/x86/xstate.h
178
static inline void set_rand_data(struct xstate_info *xstate, struct xsave_buffer *xbuf)
tools/testing/selftests/x86/xstate.h
180
int *ptr = (int *)&xbuf->bytes[xstate->xbuf_offset];
tools/testing/selftests/x86/xstate.h
190
for (i = 0; i < xstate->size / sizeof(int); i++, ptr++)