Symbol: t
arch/alpha/include/asm/bitops.h
395
unsigned long t, a, r;
arch/alpha/include/asm/bitops.h
397
t = __kernel_cmpbge (x, 0x0101010101010101UL);
arch/alpha/include/asm/bitops.h
398
a = __flsm1_tab[t];
arch/alpha/include/asm/bitops.h
399
t = __kernel_extbl (x, a);
arch/alpha/include/asm/bitops.h
400
r = a*8 + __flsm1_tab[t] + (x != 0);
arch/alpha/kernel/err_impl.h
25
#define SUBPACKET_ANNOTATION(c, t, r, d, a) {NULL, (c), (t), (r), (d), (a)}
arch/alpha/kernel/process.c
339
int elf_core_copy_task_fpregs(struct task_struct *t, elf_fpregset_t *fpu)
arch/alpha/kernel/process.c
341
memcpy(fpu, task_thread_info(t)->fp, 32 * 8);
arch/alpha/kernel/process.c
360
thread_saved_pc(struct task_struct *t)
arch/alpha/kernel/process.c
362
unsigned long base = (unsigned long)task_stack_page(t);
arch/alpha/kernel/process.c
363
unsigned long fp, sp = task_thread_info(t)->pcb.ksp;
arch/alpha/kernel/srmcons.c
70
srmcons_receive_chars(struct timer_list *t)
arch/alpha/kernel/srmcons.c
72
struct srmcons_private *srmconsp = timer_container_of(srmconsp, t,
arch/arc/include/asm/io.h
105
u##t x = __raw_read##f(addr); \
arch/arc/include/asm/io.h
110
u##t x = __raw_read##f(addr); \
arch/arc/include/asm/io.h
155
#define __raw_writesx(t,f) \
arch/arc/include/asm/io.h
159
bool is_aligned = ((unsigned long)ptr % ((t) / 8)) == 0; \
arch/arc/include/asm/io.h
160
const u##t *buf = ptr; \
arch/arc/include/asm/io.h
92
#define __raw_readsx(t,f) \
arch/arc/include/asm/io.h
96
bool is_aligned = ((unsigned long)ptr % ((t) / 8)) == 0; \
arch/arc/include/asm/io.h
97
u##t *buf = ptr; \
arch/arc/kernel/jump_label.c
51
u32 s, S, t;
arch/arc/kernel/jump_label.c
72
t = (u_offset >> 21) & GENMASK(3, 0);
arch/arc/kernel/jump_label.c
77
instruction_r = (S << 6) | t;
arch/arc/kernel/unwind.c
890
#define FRAME_REG(r, t) (((t *)frame)[reg_info[r].offs])
arch/arm/common/bL_switcher.c
272
struct bL_thread *t = arg;
arch/arm/common/bL_switcher.c
278
complete(&t->started);
arch/arm/common/bL_switcher.c
283
wait_event_interruptible(t->wq,
arch/arm/common/bL_switcher.c
284
t->wanted_cluster != -1 ||
arch/arm/common/bL_switcher.c
287
spin_lock(&t->lock);
arch/arm/common/bL_switcher.c
288
cluster = t->wanted_cluster;
arch/arm/common/bL_switcher.c
289
completer = t->completer;
arch/arm/common/bL_switcher.c
290
completer_cookie = t->completer_cookie;
arch/arm/common/bL_switcher.c
291
t->wanted_cluster = -1;
arch/arm/common/bL_switcher.c
292
t->completer = NULL;
arch/arm/common/bL_switcher.c
293
spin_unlock(&t->lock);
arch/arm/common/bL_switcher.c
344
struct bL_thread *t;
arch/arm/common/bL_switcher.c
351
t = &bL_threads[cpu];
arch/arm/common/bL_switcher.c
353
if (IS_ERR(t->task))
arch/arm/common/bL_switcher.c
354
return PTR_ERR(t->task);
arch/arm/common/bL_switcher.c
355
if (!t->task)
arch/arm/common/bL_switcher.c
358
spin_lock(&t->lock);
arch/arm/common/bL_switcher.c
359
if (t->completer) {
arch/arm/common/bL_switcher.c
360
spin_unlock(&t->lock);
arch/arm/common/bL_switcher.c
363
t->completer = completer;
arch/arm/common/bL_switcher.c
364
t->completer_cookie = completer_cookie;
arch/arm/common/bL_switcher.c
365
t->wanted_cluster = new_cluster_id;
arch/arm/common/bL_switcher.c
366
spin_unlock(&t->lock);
arch/arm/common/bL_switcher.c
367
wake_up(&t->wq);
arch/arm/common/bL_switcher.c
575
struct bL_thread *t = &bL_threads[cpu];
arch/arm/common/bL_switcher.c
576
spin_lock_init(&t->lock);
arch/arm/common/bL_switcher.c
577
init_waitqueue_head(&t->wq);
arch/arm/common/bL_switcher.c
578
init_completion(&t->started);
arch/arm/common/bL_switcher.c
579
t->wanted_cluster = -1;
arch/arm/common/bL_switcher.c
580
t->task = bL_switcher_thread_create(cpu, t);
arch/arm/common/bL_switcher.c
603
struct bL_thread *t;
arch/arm/common/bL_switcher.c
628
t = &bL_threads[cpu];
arch/arm/common/bL_switcher.c
629
task = t->task;
arch/arm/common/bL_switcher.c
630
t->task = NULL;
arch/arm/common/bL_switcher.c
638
init_completion(&t->started);
arch/arm/common/bL_switcher.c
639
t->wanted_cluster = bL_switcher_cpu_original_cluster[cpu];
arch/arm/common/bL_switcher.c
640
task = bL_switcher_thread_create(cpu, t);
arch/arm/common/bL_switcher.c
642
wait_for_completion(&t->started);
arch/arm/include/asm/assembler.h
470
.macro usraccoff, instr, reg, ptr, inc, off, cond, abort, t=TUSER()
arch/arm/include/asm/assembler.h
510
.macro usracc, instr, reg, ptr, inc, cond, rept, abort, t=TUSER()
arch/arm/include/asm/domain.h
134
#define TUSER(instr) instr ## t
arch/arm/include/asm/memory.h
249
phys_addr_t t;
arch/arm/include/asm/memory.h
252
__pv_stub(x, t, "add");
arch/arm/include/asm/memory.h
254
__pv_add_carry_stub(x, t);
arch/arm/include/asm/memory.h
256
return t;
arch/arm/include/asm/memory.h
261
unsigned long t;
arch/arm/include/asm/memory.h
269
__pv_stub((unsigned long) x, t, "sub");
arch/arm/include/asm/memory.h
270
return t;
arch/arm/include/asm/vdso/cp15.h
19
#define __read_sysreg(r, w, c, t) ({ \
arch/arm/include/asm/vdso/cp15.h
20
t __val; \
arch/arm/include/asm/vdso/cp15.h
26
#define __write_sysreg(v, r, w, c, t) asm volatile(w " " c : : "r" ((t)(v)))
arch/arm/include/uapi/asm/setup.h
181
#define tag_next(t) ((struct tag *)((__u32 *)(t) + (t)->hdr.size))
arch/arm/include/uapi/asm/setup.h
184
#define for_each_tag(t,base) \
arch/arm/include/uapi/asm/setup.h
185
for (t = base; t->hdr.size; t = tag_next(t))
arch/arm/include/uapi/asm/swab.h
30
__u32 t;
arch/arm/include/uapi/asm/swab.h
39
asm ("eor\t%0, %1, %1, ror #16" : "=r" (t) : "r" (x));
arch/arm/include/uapi/asm/swab.h
42
t = x ^ ((x << 16) | (x >> 16)); /* eor r1,r0,r0,ror #16 */
arch/arm/include/uapi/asm/swab.h
45
t &= ~0x00FF0000; /* bic r1,r1,#0x00FF0000 */
arch/arm/include/uapi/asm/swab.h
46
x ^= (t >> 8); /* eor r0,r0,r1,lsr #8 */
arch/arm/kernel/atags_parse.c
146
struct tagtable *t;
arch/arm/kernel/atags_parse.c
148
for (t = &__tagtable_begin; t < &__tagtable_end; t++)
arch/arm/kernel/atags_parse.c
149
if (tag->hdr.tag == t->tag) {
arch/arm/kernel/atags_parse.c
150
t->parse(tag);
arch/arm/kernel/atags_parse.c
154
return t < &__tagtable_end;
arch/arm/kernel/atags_parse.c
161
static void __init parse_tags(const struct tag *t)
arch/arm/kernel/atags_parse.c
163
for (; t->hdr.size; t = tag_next(t))
arch/arm/kernel/atags_parse.c
164
if (!parse_tag(t))
arch/arm/kernel/atags_parse.c
166
t->hdr.tag);
arch/arm/kernel/io.c
47
unsigned char *t = to;
arch/arm/kernel/io.c
50
*t = readb(from);
arch/arm/kernel/io.c
51
t++;
arch/arm/kernel/ptrace.c
384
struct thread_struct *t = &tsk->thread;
arch/arm/kernel/ptrace.c
387
if (t->debug.hbp[i]) {
arch/arm/kernel/ptrace.c
388
unregister_hw_breakpoint(t->debug.hbp[i]);
arch/arm/kernel/ptrace.c
389
t->debug.hbp[i] = NULL;
arch/arm/kernel/thumbee.c
30
static int thumbee_notifier(struct notifier_block *self, unsigned long cmd, void *t)
arch/arm/kernel/thumbee.c
32
struct thread_info *thread = t;
arch/arm/kernel/xscale-cp0.c
33
static int dsp_do(struct notifier_block *self, unsigned long cmd, void *t)
arch/arm/kernel/xscale-cp0.c
35
struct thread_info *thread = t;
arch/arm/kernel/xscale-cp0.c
58
static int iwmmxt_do(struct notifier_block *self, unsigned long cmd, void *t)
arch/arm/kernel/xscale-cp0.c
60
struct thread_info *thread = t;
arch/arm/mach-omap2/pm-debug.c
177
s64 t;
arch/arm/mach-omap2/pm-debug.c
180
t = sched_clock();
arch/arm/mach-omap2/pm-debug.c
185
pwrdm->timer = t;
arch/arm/mach-omap2/pm-debug.c
54
s64 t;
arch/arm/mach-omap2/pm-debug.c
60
t = sched_clock();
arch/arm/mach-omap2/pm-debug.c
62
pwrdm->state_timer[prev] += t - pwrdm->timer;
arch/arm/mach-omap2/pm-debug.c
64
pwrdm->timer = t;
arch/arm/mach-omap2/usb-tusb6010.c
49
struct gpmc_timings t;
arch/arm/mach-omap2/usb-tusb6010.c
67
gpmc_calc_timings(&t, &tusb_async, &dev_t);
arch/arm/mach-omap2/usb-tusb6010.c
69
return gpmc_cs_set_timings(async_cs, &t, &tusb_async);
arch/arm/mach-omap2/usb-tusb6010.c
75
struct gpmc_timings t;
arch/arm/mach-omap2/usb-tusb6010.c
94
gpmc_calc_timings(&t, &tusb_sync, &dev_t);
arch/arm/mach-omap2/usb-tusb6010.c
96
return gpmc_cs_set_timings(sync_cs, &t, &tusb_sync);
arch/arm/mach-orion5x/common.c
367
void __init tag_fixup_mem32(struct tag *t, char **from)
arch/arm/mach-orion5x/common.c
369
for (; t->hdr.size; t = tag_next(t))
arch/arm/mach-orion5x/common.c
370
if (t->hdr.tag == ATAG_MEM &&
arch/arm/mach-orion5x/common.c
371
(!t->u.mem.size || t->u.mem.size & ~PAGE_MASK ||
arch/arm/mach-orion5x/common.c
372
t->u.mem.start & ~PAGE_MASK)) {
arch/arm/mach-orion5x/common.c
375
t->u.mem.size / 1024, t->u.mem.start);
arch/arm/mach-orion5x/common.c
376
t->hdr.tag = 0;
arch/arm/mach-rpc/include/mach/uncompress.h
115
struct tag *t = (struct tag *)params;
arch/arm/mach-rpc/include/mach/uncompress.h
118
if (t->hdr.tag == ATAG_CORE) {
arch/arm/mach-rpc/include/mach/uncompress.h
119
for (; t->hdr.size; t = tag_next(t)) {
arch/arm/mach-rpc/include/mach/uncompress.h
120
if (t->hdr.tag == ATAG_VIDEOTEXT) {
arch/arm/mach-rpc/include/mach/uncompress.h
121
video_num_rows = t->u.videotext.video_lines;
arch/arm/mach-rpc/include/mach/uncompress.h
122
video_num_cols = t->u.videotext.video_cols;
arch/arm/mach-rpc/include/mach/uncompress.h
123
video_x = t->u.videotext.x;
arch/arm/mach-rpc/include/mach/uncompress.h
124
video_y = t->u.videotext.y;
arch/arm/mach-rpc/include/mach/uncompress.h
125
} else if (t->hdr.tag == ATAG_VIDEOLFB) {
arch/arm/mach-rpc/include/mach/uncompress.h
126
bytes_per_char_h = t->u.videolfb.lfb_depth;
arch/arm/mach-rpc/include/mach/uncompress.h
128
} else if (t->hdr.tag == ATAG_MEM) {
arch/arm/mach-rpc/include/mach/uncompress.h
130
nr_pages += (t->u.mem.size / PAGE_SIZE);
arch/arm/mm/context.c
103
void *t)
arch/arm/mm/context.c
107
struct thread_info *thread = t;
arch/arm/mm/init.c
373
struct task_struct *t, *s;
arch/arm/mm/init.c
375
for_each_process(t) {
arch/arm/mm/init.c
376
if (t->flags & PF_KTHREAD)
arch/arm/mm/init.c
378
for_each_thread(t, s)
arch/arm/mm/mmu.c
704
struct mem_type *t = &mem_types[i];
arch/arm/mm/mmu.c
705
if (t->prot_l1)
arch/arm/mm/mmu.c
706
t->prot_l1 |= PMD_DOMAIN(t->domain);
arch/arm/mm/mmu.c
707
if (t->prot_sect)
arch/arm/mm/mmu.c
708
t->prot_sect |= PMD_DOMAIN(t->domain);
arch/arm/probes/kprobes/test-core.c
505
unsigned n, i, t, t0;
arch/arm/probes/kprobes/test-core.c
511
t = sched_clock() - t0;
arch/arm/probes/kprobes/test-core.c
512
if (t >= 250000000)
arch/arm/probes/kprobes/test-core.c
515
return t / n; /* Time for one iteration in nanoseconds */
arch/arm/probes/uprobes/core.c
162
bool arch_uprobe_xol_was_trapped(struct task_struct *t)
arch/arm/probes/uprobes/core.c
164
if (t->thread.trap_no != UPROBE_TRAP_NR)
arch/arm/vfp/vfpdouble.c
744
struct vfp_double *t = vdn;
arch/arm/vfp/vfpdouble.c
746
vdm = t;
arch/arm/vfp/vfpdouble.c
802
struct vfp_double *t = vdn;
arch/arm/vfp/vfpdouble.c
804
vdm = t;
arch/arm/vfp/vfpsingle.c
789
struct vfp_single *t = vsn;
arch/arm/vfp/vfpsingle.c
791
vsm = t;
arch/arm/vfp/vfpsingle.c
846
struct vfp_single *t = vsn;
arch/arm/vfp/vfpsingle.c
848
vsm = t;
arch/arm64/include/asm/fpsimd.h
349
static inline void vec_init_vq_map(enum vec_type t) { }
arch/arm64/include/asm/fpsimd.h
350
static inline void vec_update_vq_map(enum vec_type t) { }
arch/arm64/include/asm/fpsimd.h
351
static inline int vec_verify_vq_map(enum vec_type t) { return 0; }
arch/arm64/include/asm/pgtable-hwdef.h
158
#define PMD_ATTRINDX(t) (_AT(pmdval_t, (t)) << 2)
arch/arm64/include/asm/pgtable-hwdef.h
195
#define PTE_ATTRINDX(t) (_AT(pteval_t, (t)) << 2)
arch/arm64/include/asm/pgtable-hwdef.h
219
#define PTE_S2_MEMATTR(t) (_AT(pteval_t, (t)) << 2)
arch/arm64/include/asm/processor.h
282
#define task_user_tls(t) \
arch/arm64/include/asm/processor.h
285
if (is_compat_thread(task_thread_info(t))) \
arch/arm64/include/asm/processor.h
286
__tls = &(t)->thread.uw.tp2_value; \
arch/arm64/include/asm/processor.h
288
__tls = &(t)->thread.uw.tp_value; \
arch/arm64/include/asm/processor.h
292
#define task_user_tls(t) (&(t)->thread.uw.tp_value)
arch/arm64/include/asm/sysreg.h
1107
__emit_inst(0xd5200000|(\sreg)|(.L__gpr_num_\rt))
arch/arm64/include/asm/sysreg.h
1111
__emit_inst(0xd5000000|(\sreg)|(.L__gpr_num_\rt))
arch/arm64/kernel/fpsimd.c
1750
void fpsimd_flush_task_state(struct task_struct *t)
arch/arm64/kernel/fpsimd.c
1752
t->thread.fpsimd_cpu = NR_CPUS;
arch/arm64/kernel/fpsimd.c
1753
t->thread.kernel_fpsimd_state = NULL;
arch/arm64/kernel/fpsimd.c
1762
set_tsk_thread_flag(t, TIF_FOREIGN_FPSTATE);
arch/arm64/kernel/fpsimd.c
201
static inline void sme_free(struct task_struct *t) { }
arch/arm64/kernel/probes/uprobes.c
106
return t->thread.fault_code != UPROBE_INV_FAULT_CODE;
arch/arm64/kernel/probes/uprobes.c
99
bool arch_uprobe_xol_was_trapped(struct task_struct *t)
arch/arm64/kernel/ptrace.c
210
struct thread_struct *t = &tsk->thread;
arch/arm64/kernel/ptrace.c
213
if (t->debug.hbp_break[i]) {
arch/arm64/kernel/ptrace.c
214
unregister_hw_breakpoint(t->debug.hbp_break[i]);
arch/arm64/kernel/ptrace.c
215
t->debug.hbp_break[i] = NULL;
arch/arm64/kernel/ptrace.c
220
if (t->debug.hbp_watch[i]) {
arch/arm64/kernel/ptrace.c
221
unregister_hw_breakpoint(t->debug.hbp_watch[i]);
arch/arm64/kernel/ptrace.c
222
t->debug.hbp_watch[i] = NULL;
arch/arm64/kvm/sys_regs.c
169
#define MAPPED_EL2_SYSREG(r, m, t) \
arch/arm64/kvm/sys_regs.c
171
locate_mapped_el2_register(vcpu, r, m, t, loc); \
arch/arm64/kvm/vgic/vgic-its.c
34
#define vgic_its_read_entry_lock(i, g, valp, t) \
arch/arm64/kvm/vgic/vgic-its.c
36
int __sz = vgic_its_get_abi(i)->t##_esz; \
arch/arm64/kvm/vgic/vgic-its.c
51
#define vgic_its_write_entry_lock(i, g, val, t) \
arch/arm64/kvm/vgic/vgic-its.c
53
int __sz = vgic_its_get_abi(i)->t##_esz; \
arch/csky/include/asm/futex.h
110
[t] "=&r" (tmp)
arch/csky/include/asm/futex.h
34
[u] "+m" (*uaddr), [t] "=&r" (tmp) \
arch/csky/kernel/probes/uprobes.c
76
bool arch_uprobe_xol_was_trapped(struct task_struct *t)
arch/csky/kernel/probes/uprobes.c
78
if (t->thread.trap_no != UPROBE_TRAP_NR)
arch/csky/kernel/process.c
73
int elf_core_copy_task_fpregs(struct task_struct *t, elf_fpregset_t *fpu)
arch/loongarch/include/asm/fpu.h
229
static inline void save_lsx(struct task_struct *t)
arch/loongarch/include/asm/fpu.h
232
_save_lsx(&t->thread.fpu);
arch/loongarch/include/asm/fpu.h
235
static inline void restore_lsx(struct task_struct *t)
arch/loongarch/include/asm/fpu.h
238
_restore_lsx(&t->thread.fpu);
arch/loongarch/include/asm/fpu.h
247
static inline void restore_lsx_upper(struct task_struct *t)
arch/loongarch/include/asm/fpu.h
250
_restore_lsx_upper(&t->thread.fpu);
arch/loongarch/include/asm/fpu.h
256
static inline void save_lsx(struct task_struct *t) {}
arch/loongarch/include/asm/fpu.h
257
static inline void restore_lsx(struct task_struct *t) {}
arch/loongarch/include/asm/fpu.h
259
static inline void restore_lsx_upper(struct task_struct *t) {}
arch/loongarch/include/asm/fpu.h
277
static inline void save_lasx(struct task_struct *t)
arch/loongarch/include/asm/fpu.h
280
_save_lasx(&t->thread.fpu);
arch/loongarch/include/asm/fpu.h
283
static inline void restore_lasx(struct task_struct *t)
arch/loongarch/include/asm/fpu.h
286
_restore_lasx(&t->thread.fpu);
arch/loongarch/include/asm/fpu.h
295
static inline void restore_lasx_upper(struct task_struct *t)
arch/loongarch/include/asm/fpu.h
298
_restore_lasx_upper(&t->thread.fpu);
arch/loongarch/include/asm/fpu.h
304
static inline void save_lasx(struct task_struct *t) {}
arch/loongarch/include/asm/fpu.h
305
static inline void restore_lasx(struct task_struct *t) {}
arch/loongarch/include/asm/fpu.h
307
static inline void restore_lasx_upper(struct task_struct *t) {}
arch/loongarch/include/asm/fpu.h
51
static inline void save_lsx(struct task_struct *t);
arch/loongarch/include/asm/fpu.h
52
static inline void restore_lsx(struct task_struct *t);
arch/loongarch/include/asm/fpu.h
56
static inline void save_lasx(struct task_struct *t);
arch/loongarch/include/asm/fpu.h
57
static inline void restore_lasx(struct task_struct *t);
arch/loongarch/include/asm/gpr-num.h
20
.equ .L__gpr_num_$t\num, 12 + \num
arch/loongarch/kernel/hw_breakpoint.c
100
GEN_WRITE_WB_REG_CASES(CSR_CFG_CTRL, CTRL, t, val);
arch/loongarch/kernel/hw_breakpoint.c
101
GEN_WRITE_WB_REG_CASES(CSR_CFG_ASID, ASID, t, val);
arch/loongarch/kernel/hw_breakpoint.c
169
struct thread_struct *t = &tsk->thread;
arch/loongarch/kernel/hw_breakpoint.c
172
if (t->hbp_break[i]) {
arch/loongarch/kernel/hw_breakpoint.c
173
unregister_hw_breakpoint(t->hbp_break[i]);
arch/loongarch/kernel/hw_breakpoint.c
174
t->hbp_break[i] = NULL;
arch/loongarch/kernel/hw_breakpoint.c
179
if (t->hbp_watch[i]) {
arch/loongarch/kernel/hw_breakpoint.c
180
unregister_hw_breakpoint(t->hbp_watch[i]);
arch/loongarch/kernel/hw_breakpoint.c
181
t->hbp_watch[i] = NULL;
arch/loongarch/kernel/hw_breakpoint.c
78
static u64 read_wb_reg(int reg, int n, int t)
arch/loongarch/kernel/hw_breakpoint.c
83
GEN_READ_WB_REG_CASES(CSR_CFG_ADDR, ADDR, t, val);
arch/loongarch/kernel/hw_breakpoint.c
84
GEN_READ_WB_REG_CASES(CSR_CFG_MASK, MASK, t, val);
arch/loongarch/kernel/hw_breakpoint.c
85
GEN_READ_WB_REG_CASES(CSR_CFG_CTRL, CTRL, t, val);
arch/loongarch/kernel/hw_breakpoint.c
86
GEN_READ_WB_REG_CASES(CSR_CFG_ASID, ASID, t, val);
arch/loongarch/kernel/hw_breakpoint.c
95
static void write_wb_reg(int reg, int n, int t, u64 val)
arch/loongarch/kernel/hw_breakpoint.c
98
GEN_WRITE_WB_REG_CASES(CSR_CFG_ADDR, ADDR, t, val);
arch/loongarch/kernel/hw_breakpoint.c
99
GEN_WRITE_WB_REG_CASES(CSR_CFG_MASK, MASK, t, val);
arch/loongarch/kernel/uprobes.c
68
bool arch_uprobe_xol_was_trapped(struct task_struct *t)
arch/loongarch/kernel/uprobes.c
70
if (t->thread.trap_nr != UPROBE_TRAP_NR)
arch/m68k/68000/m68328.h
5
int m68328_hwclk(int set, struct rtc_time *t);
arch/m68k/68000/timers.c
122
int m68328_hwclk(int set, struct rtc_time *t)
arch/m68k/68000/timers.c
126
t->tm_year = 1;
arch/m68k/68000/timers.c
127
t->tm_mon = 0;
arch/m68k/68000/timers.c
128
t->tm_mday = 1;
arch/m68k/68000/timers.c
129
t->tm_hour = (now >> 24) % 24;
arch/m68k/68000/timers.c
130
t->tm_min = (now >> 16) % 60;
arch/m68k/68000/timers.c
131
t->tm_sec = now % 60;
arch/m68k/apollo/config.c
181
int dn_dummy_hwclk(int op, struct rtc_time *t) {
arch/m68k/apollo/config.c
185
t->tm_sec=rtc->second;
arch/m68k/apollo/config.c
186
t->tm_min=rtc->minute;
arch/m68k/apollo/config.c
187
t->tm_hour=rtc->hours;
arch/m68k/apollo/config.c
188
t->tm_mday=rtc->day_of_month;
arch/m68k/apollo/config.c
189
t->tm_wday=rtc->day_of_week;
arch/m68k/apollo/config.c
190
t->tm_mon = rtc->month - 1;
arch/m68k/apollo/config.c
191
t->tm_year=rtc->year;
arch/m68k/apollo/config.c
192
if (t->tm_year < 70)
arch/m68k/apollo/config.c
193
t->tm_year += 100;
arch/m68k/apollo/config.c
195
rtc->second=t->tm_sec;
arch/m68k/apollo/config.c
196
rtc->minute=t->tm_min;
arch/m68k/apollo/config.c
197
rtc->hours=t->tm_hour;
arch/m68k/apollo/config.c
198
rtc->day_of_month=t->tm_mday;
arch/m68k/apollo/config.c
199
if(t->tm_wday!=-1)
arch/m68k/apollo/config.c
200
rtc->day_of_week=t->tm_wday;
arch/m68k/apollo/config.c
201
rtc->month = t->tm_mon + 1;
arch/m68k/apollo/config.c
202
rtc->year = t->tm_year % 100;
arch/m68k/atari/atari.h
14
int atari_mste_hwclk(int op, struct rtc_time *t);
arch/m68k/atari/atari.h
15
int atari_tt_hwclk(int op, struct rtc_time *t);
arch/m68k/atari/time.c
142
int atari_mste_hwclk( int op, struct rtc_time *t )
arch/m68k/atari/time.c
155
val.sec_ones = t->tm_sec % 10;
arch/m68k/atari/time.c
156
val.sec_tens = t->tm_sec / 10;
arch/m68k/atari/time.c
157
val.min_ones = t->tm_min % 10;
arch/m68k/atari/time.c
158
val.min_tens = t->tm_min / 10;
arch/m68k/atari/time.c
159
hour = t->tm_hour;
arch/m68k/atari/time.c
168
val.day_ones = t->tm_mday % 10;
arch/m68k/atari/time.c
169
val.day_tens = t->tm_mday / 10;
arch/m68k/atari/time.c
170
val.mon_ones = (t->tm_mon+1) % 10;
arch/m68k/atari/time.c
171
val.mon_tens = (t->tm_mon+1) / 10;
arch/m68k/atari/time.c
172
year = t->tm_year - 80;
arch/m68k/atari/time.c
175
val.weekday = t->tm_wday;
arch/m68k/atari/time.c
183
t->tm_sec = val.sec_ones + val.sec_tens * 10;
arch/m68k/atari/time.c
184
t->tm_min = val.min_ones + val.min_tens * 10;
arch/m68k/atari/time.c
192
t->tm_hour = hour;
arch/m68k/atari/time.c
193
t->tm_mday = val.day_ones + val.day_tens * 10;
arch/m68k/atari/time.c
194
t->tm_mon = val.mon_ones + val.mon_tens * 10 - 1;
arch/m68k/atari/time.c
195
t->tm_year = val.year_ones + val.year_tens * 10 + 80;
arch/m68k/atari/time.c
196
t->tm_wday = val.weekday;
arch/m68k/atari/time.c
201
int atari_tt_hwclk( int op, struct rtc_time *t )
arch/m68k/atari/time.c
214
sec = t->tm_sec;
arch/m68k/atari/time.c
215
min = t->tm_min;
arch/m68k/atari/time.c
216
hour = t->tm_hour;
arch/m68k/atari/time.c
217
day = t->tm_mday;
arch/m68k/atari/time.c
218
mon = t->tm_mon + 1;
arch/m68k/atari/time.c
219
year = t->tm_year - atari_rtc_year_offset;
arch/m68k/atari/time.c
220
wday = t->tm_wday + (t->tm_wday >= 0);
arch/m68k/atari/time.c
311
t->tm_sec = sec;
arch/m68k/atari/time.c
312
t->tm_min = min;
arch/m68k/atari/time.c
313
t->tm_hour = hour;
arch/m68k/atari/time.c
314
t->tm_mday = day;
arch/m68k/atari/time.c
315
t->tm_mon = mon - 1;
arch/m68k/atari/time.c
316
t->tm_year = year + atari_rtc_year_offset;
arch/m68k/atari/time.c
317
t->tm_wday = wday - 1;
arch/m68k/bvme6000/config.c
283
int bvme6000_hwclk(int op, struct rtc_time *t)
arch/m68k/bvme6000/config.c
292
rtc->t0cr_rtmr = t->tm_year%4;
arch/m68k/bvme6000/config.c
294
rtc->bcd_sec = bin2bcd(t->tm_sec);
arch/m68k/bvme6000/config.c
295
rtc->bcd_min = bin2bcd(t->tm_min);
arch/m68k/bvme6000/config.c
296
rtc->bcd_hr = bin2bcd(t->tm_hour);
arch/m68k/bvme6000/config.c
297
rtc->bcd_dom = bin2bcd(t->tm_mday);
arch/m68k/bvme6000/config.c
298
rtc->bcd_mth = bin2bcd(t->tm_mon + 1);
arch/m68k/bvme6000/config.c
299
rtc->bcd_year = bin2bcd(t->tm_year%100);
arch/m68k/bvme6000/config.c
300
if (t->tm_wday >= 0)
arch/m68k/bvme6000/config.c
301
rtc->bcd_dow = bin2bcd(t->tm_wday+1);
arch/m68k/bvme6000/config.c
302
rtc->t0cr_rtmr = t->tm_year%4 | 0x08;
arch/m68k/bvme6000/config.c
307
t->tm_sec = bcd2bin(rtc->bcd_sec);
arch/m68k/bvme6000/config.c
308
t->tm_min = bcd2bin(rtc->bcd_min);
arch/m68k/bvme6000/config.c
309
t->tm_hour = bcd2bin(rtc->bcd_hr);
arch/m68k/bvme6000/config.c
310
t->tm_mday = bcd2bin(rtc->bcd_dom);
arch/m68k/bvme6000/config.c
311
t->tm_mon = bcd2bin(rtc->bcd_mth)-1;
arch/m68k/bvme6000/config.c
312
t->tm_year = bcd2bin(rtc->bcd_year);
arch/m68k/bvme6000/config.c
313
if (t->tm_year < 70)
arch/m68k/bvme6000/config.c
314
t->tm_year += 100;
arch/m68k/bvme6000/config.c
315
t->tm_wday = bcd2bin(rtc->bcd_dow)-1;
arch/m68k/bvme6000/config.c
316
} while (t->tm_sec != bcd2bin(rtc->bcd_sec));
arch/m68k/hp300/config.c
202
static int hp300_hwclk(int op, struct rtc_time *t)
arch/m68k/hp300/config.c
205
t->tm_sec = hp300_rtc_read(RTC_REG_SEC1) * 10 +
arch/m68k/hp300/config.c
207
t->tm_min = hp300_rtc_read(RTC_REG_MIN1) * 10 +
arch/m68k/hp300/config.c
209
t->tm_hour = (hp300_rtc_read(RTC_REG_HOUR1) & 3) * 10 +
arch/m68k/hp300/config.c
211
t->tm_wday = -1;
arch/m68k/hp300/config.c
212
t->tm_mday = hp300_rtc_read(RTC_REG_DAY1) * 10 +
arch/m68k/hp300/config.c
214
t->tm_mon = hp300_rtc_read(RTC_REG_MON1) * 10 +
arch/m68k/hp300/config.c
216
t->tm_year = hp300_rtc_read(RTC_REG_YEAR1) * 10 +
arch/m68k/hp300/config.c
218
if (t->tm_year <= 69)
arch/m68k/hp300/config.c
219
t->tm_year += 100;
arch/m68k/hp300/config.c
221
hp300_rtc_write(RTC_REG_SEC1, t->tm_sec / 10);
arch/m68k/hp300/config.c
222
hp300_rtc_write(RTC_REG_SEC2, t->tm_sec % 10);
arch/m68k/hp300/config.c
223
hp300_rtc_write(RTC_REG_MIN1, t->tm_min / 10);
arch/m68k/hp300/config.c
224
hp300_rtc_write(RTC_REG_MIN2, t->tm_min % 10);
arch/m68k/hp300/config.c
226
((t->tm_hour / 10) & 3) | RTC_HOUR1_24HMODE);
arch/m68k/hp300/config.c
227
hp300_rtc_write(RTC_REG_HOUR2, t->tm_hour % 10);
arch/m68k/hp300/config.c
228
hp300_rtc_write(RTC_REG_DAY1, t->tm_mday / 10);
arch/m68k/hp300/config.c
229
hp300_rtc_write(RTC_REG_DAY2, t->tm_mday % 10);
arch/m68k/hp300/config.c
230
hp300_rtc_write(RTC_REG_MON1, (t->tm_mon + 1) / 10);
arch/m68k/hp300/config.c
231
hp300_rtc_write(RTC_REG_MON2, (t->tm_mon + 1) % 10);
arch/m68k/hp300/config.c
232
if (t->tm_year >= 100)
arch/m68k/hp300/config.c
233
t->tm_year -= 100;
arch/m68k/hp300/config.c
234
hp300_rtc_write(RTC_REG_YEAR1, t->tm_year / 10);
arch/m68k/hp300/config.c
235
hp300_rtc_write(RTC_REG_YEAR2, t->tm_year % 10);
arch/m68k/ifpsp060/src/fpsp.S
24765
# if it's a fmove out instruction, we don't have to fix a7
arch/m68k/ifpsp060/src/pfpsp.S
14725
# if it's a fmove out instruction, we don't have to fix a7
arch/m68k/include/asm/atomic.h
43
int t, tmp; \
arch/m68k/include/asm/atomic.h
50
: "+m" (*v), "=&d" (t), "=&d" (tmp) \
arch/m68k/include/asm/atomic.h
52
return t; \
arch/m68k/include/asm/atomic.h
58
int t, tmp; \
arch/m68k/include/asm/atomic.h
65
: "+m" (*v), "=&d" (t), "=&d" (tmp) \
arch/m68k/include/asm/atomic.h
76
int t; \
arch/m68k/include/asm/atomic.h
79
t = (v->counter c_op i); \
arch/m68k/include/asm/atomic.h
82
return t; \
arch/m68k/include/asm/atomic.h
89
int t; \
arch/m68k/include/asm/atomic.h
92
t = v->counter; \
arch/m68k/include/asm/atomic.h
96
return t; \
arch/m68k/include/asm/uaccess.h
131
__typeof__(*(ptr)) t; \
arch/m68k/include/asm/uaccess.h
155
(x) = __gu_val.t; \
arch/m68k/kernel/process.c
218
int elf_core_copy_task_fpregs(struct task_struct *t, elf_fpregset_t *fpu)
arch/m68k/mac/mac.h
12
int mac_hwclk(int op, struct rtc_time *t);
arch/m68k/mac/misc.c
610
int mac_hwclk(int op, struct rtc_time *t)
arch/m68k/mac/misc.c
636
t->tm_wday = 0;
arch/m68k/mac/misc.c
638
&t->tm_year, &t->tm_mon, &t->tm_mday,
arch/m68k/mac/misc.c
639
&t->tm_hour, &t->tm_min, &t->tm_sec);
arch/m68k/mac/misc.c
640
pr_debug("%s: read %ptR\n", __func__, t);
arch/m68k/mac/misc.c
642
pr_debug("%s: tried to write %ptR\n", __func__, t);
arch/m68k/mac/misc.c
648
via_set_rtc_time(t);
arch/m68k/mac/misc.c
653
cuda_set_rtc_time(t);
arch/m68k/mac/misc.c
658
pmu_set_rtc_time(t);
arch/m68k/q40/config.c
203
static int q40_hwclk(int op, struct rtc_time *t)
arch/m68k/q40/config.c
209
Q40_RTC_SECS = bin2bcd(t->tm_sec);
arch/m68k/q40/config.c
210
Q40_RTC_MINS = bin2bcd(t->tm_min);
arch/m68k/q40/config.c
211
Q40_RTC_HOUR = bin2bcd(t->tm_hour);
arch/m68k/q40/config.c
212
Q40_RTC_DATE = bin2bcd(t->tm_mday);
arch/m68k/q40/config.c
213
Q40_RTC_MNTH = bin2bcd(t->tm_mon + 1);
arch/m68k/q40/config.c
214
Q40_RTC_YEAR = bin2bcd(t->tm_year%100);
arch/m68k/q40/config.c
215
if (t->tm_wday >= 0)
arch/m68k/q40/config.c
216
Q40_RTC_DOW = bin2bcd(t->tm_wday+1);
arch/m68k/q40/config.c
223
t->tm_year = bcd2bin (Q40_RTC_YEAR);
arch/m68k/q40/config.c
224
t->tm_mon = bcd2bin (Q40_RTC_MNTH)-1;
arch/m68k/q40/config.c
225
t->tm_mday = bcd2bin (Q40_RTC_DATE);
arch/m68k/q40/config.c
226
t->tm_hour = bcd2bin (Q40_RTC_HOUR);
arch/m68k/q40/config.c
227
t->tm_min = bcd2bin (Q40_RTC_MINS);
arch/m68k/q40/config.c
228
t->tm_sec = bcd2bin (Q40_RTC_SECS);
arch/m68k/q40/config.c
232
if (t->tm_year < 70)
arch/m68k/q40/config.c
233
t->tm_year += 100;
arch/m68k/q40/config.c
234
t->tm_wday = bcd2bin(Q40_RTC_DOW)-1;
arch/m68k/sun3/intersil.c
28
int sun3_hwclk(int set, struct rtc_time *t)
arch/m68k/sun3/intersil.c
42
todintersil->hour = t->tm_hour;
arch/m68k/sun3/intersil.c
43
todintersil->minute = t->tm_min;
arch/m68k/sun3/intersil.c
44
todintersil->second = t->tm_sec;
arch/m68k/sun3/intersil.c
45
todintersil->month = t->tm_mon + 1;
arch/m68k/sun3/intersil.c
46
todintersil->day = t->tm_mday;
arch/m68k/sun3/intersil.c
47
todintersil->year = (t->tm_year - 68) % 100;
arch/m68k/sun3/intersil.c
48
todintersil->weekday = t->tm_wday;
arch/m68k/sun3/intersil.c
51
t->tm_sec = todintersil->csec;
arch/m68k/sun3/intersil.c
52
t->tm_hour = todintersil->hour;
arch/m68k/sun3/intersil.c
53
t->tm_min = todintersil->minute;
arch/m68k/sun3/intersil.c
54
t->tm_sec = todintersil->second;
arch/m68k/sun3/intersil.c
55
t->tm_mon = todintersil->month - 1;
arch/m68k/sun3/intersil.c
56
t->tm_mday = todintersil->day;
arch/m68k/sun3/intersil.c
57
t->tm_year = todintersil->year + 68;
arch/m68k/sun3/intersil.c
58
t->tm_wday = todintersil->weekday;
arch/m68k/sun3/intersil.c
59
if (t->tm_year < 70)
arch/m68k/sun3/intersil.c
60
t->tm_year += 100;
arch/m68k/sun3/sun3.h
14
int sun3_hwclk(int set, struct rtc_time *t);
arch/m68k/sun3x/time.c
40
int sun3x_hwclk(int set, struct rtc_time *t)
arch/m68k/sun3x/time.c
50
h->sec = bin2bcd(t->tm_sec);
arch/m68k/sun3x/time.c
51
h->min = bin2bcd(t->tm_min);
arch/m68k/sun3x/time.c
52
h->hour = bin2bcd(t->tm_hour);
arch/m68k/sun3x/time.c
53
h->wday = bin2bcd(t->tm_wday);
arch/m68k/sun3x/time.c
54
h->mday = bin2bcd(t->tm_mday);
arch/m68k/sun3x/time.c
55
h->month = bin2bcd(t->tm_mon + 1);
arch/m68k/sun3x/time.c
56
h->year = bin2bcd(t->tm_year % 100);
arch/m68k/sun3x/time.c
60
t->tm_sec = bcd2bin(h->sec);
arch/m68k/sun3x/time.c
61
t->tm_min = bcd2bin(h->min);
arch/m68k/sun3x/time.c
62
t->tm_hour = bcd2bin(h->hour);
arch/m68k/sun3x/time.c
63
t->tm_wday = bcd2bin(h->wday);
arch/m68k/sun3x/time.c
64
t->tm_mday = bcd2bin(h->mday);
arch/m68k/sun3x/time.c
65
t->tm_mon = bcd2bin(h->month) - 1;
arch/m68k/sun3x/time.c
66
t->tm_year = bcd2bin(h->year);
arch/m68k/sun3x/time.c
68
if (t->tm_year < 70)
arch/m68k/sun3x/time.c
69
t->tm_year += 100;
arch/m68k/sun3x/time.h
5
extern int sun3x_hwclk(int set, struct rtc_time *t);
arch/microblaze/include/asm/mmu.h
43
unsigned long t:1; /* Normal or I/O type */
arch/microblaze/kernel/process.c
136
int elf_core_copy_task_fpregs(struct task_struct *t, elf_fpregset_t *fpu)
arch/microblaze/kernel/ptrace.c
47
struct task_struct *t)
arch/microblaze/kernel/ptrace.c
72
regs = task_pt_regs(t);
arch/microblaze/lib/muldi3.c
10
#define __ll_highpart(t) ((unsigned long) (t) >> (W_TYPE_SIZE / 2))
arch/microblaze/lib/muldi3.c
9
#define __ll_lowpart(t) ((unsigned long) (t) & (__ll_B - 1))
arch/mips/alchemy/common/clock.c
1052
struct clk_aliastable *t = alchemy_clk_aliases;
arch/mips/alchemy/common/clock.c
1108
while (t->base) {
arch/mips/alchemy/common/clock.c
1109
if (t->cputype == ctype)
arch/mips/alchemy/common/clock.c
1110
clk_add_alias(t->alias, NULL, t->base, NULL);
arch/mips/alchemy/common/clock.c
1111
t++;
arch/mips/alchemy/common/clock.c
121
unsigned long t;
arch/mips/alchemy/common/clock.c
129
t = 396000000;
arch/mips/alchemy/common/clock.c
131
t = alchemy_rdsys(AU1000_SYS_CPUPLL) & 0x7f;
arch/mips/alchemy/common/clock.c
133
t &= 0x3f;
arch/mips/alchemy/common/clock.c
134
t *= parent_rate;
arch/mips/alchemy/common/clock.c
137
return t;
arch/mips/alchemy/common/clock.c
339
static struct clk __init *alchemy_clk_setup_lrclk(const char *pn, int t)
arch/mips/alchemy/common/clock.c
351
switch (t) {
arch/mips/alchemy/common/clock.c
703
unsigned long v, t;
arch/mips/alchemy/common/clock.c
706
t = parent_rate / (((v >> sh) & 0xff) + 1);
arch/mips/alchemy/common/clock.c
708
t /= 2;
arch/mips/alchemy/common/clock.c
710
return t;
arch/mips/alchemy/common/time.c
103
t = 0xffffff;
arch/mips/alchemy/common/time.c
104
while ((alchemy_rdsys(AU1000_SYS_CNTRCTRL) & SYS_CNTRL_C1S) && --t)
arch/mips/alchemy/common/time.c
106
if (!t)
arch/mips/alchemy/common/time.c
110
t = 0xffffff;
arch/mips/alchemy/common/time.c
111
while ((alchemy_rdsys(AU1000_SYS_CNTRCTRL) & SYS_CNTRL_C1S) && --t)
arch/mips/alchemy/common/time.c
113
if (!t)
arch/mips/alchemy/common/time.c
149
int t;
arch/mips/alchemy/common/time.c
151
t = alchemy_get_cputype();
arch/mips/alchemy/common/time.c
152
if (t == ALCHEMY_CPU_UNKNOWN ||
arch/mips/alchemy/common/time.c
153
alchemy_time_init(alchemy_m2inttab[t]))
arch/mips/alchemy/common/time.c
78
unsigned long t;
arch/mips/alchemy/common/time.c
95
t = 0xffffff;
arch/mips/alchemy/common/time.c
96
while ((alchemy_rdsys(AU1000_SYS_CNTRCTRL) & SYS_CNTRL_T1S) && --t)
arch/mips/alchemy/common/time.c
98
if (!t)
arch/mips/alchemy/devboards/db1200.c
105
unsigned short t = bcsr_read(BCSR_HEXLEDS);
arch/mips/alchemy/devboards/db1200.c
106
bcsr_write(BCSR_HEXLEDS, ~t);
arch/mips/alchemy/devboards/db1200.c
107
if (bcsr_read(BCSR_HEXLEDS) != t) {
arch/mips/alchemy/devboards/db1200.c
108
bcsr_write(BCSR_HEXLEDS, t);
arch/mips/alchemy/devboards/db1200.c
91
unsigned short t = bcsr_read(BCSR_HEXLEDS);
arch/mips/alchemy/devboards/db1200.c
92
bcsr_write(BCSR_HEXLEDS, ~t);
arch/mips/alchemy/devboards/db1200.c
93
if (bcsr_read(BCSR_HEXLEDS) != t) {
arch/mips/alchemy/devboards/db1200.c
94
bcsr_write(BCSR_HEXLEDS, t);
arch/mips/ath79/clock.c
155
u32 t;
arch/mips/ath79/clock.c
157
t = ath79_reset_rr(AR933X_RESET_REG_BOOTSTRAP);
arch/mips/ath79/clock.c
158
if (t & AR933X_BOOTSTRAP_REF_CLK_40)
arch/mips/ath79/clock.c
176
u32 t;
arch/mips/ath79/clock.c
180
t = (cpu_config >> AR933X_PLL_CPU_CONFIG_REFDIV_SHIFT) &
arch/mips/ath79/clock.c
182
ref_div = t;
arch/mips/ath79/clock.c
187
t = (cpu_config >> AR933X_PLL_CPU_CONFIG_OUTDIV_SHIFT) &
arch/mips/ath79/clock.c
189
if (t == 0)
arch/mips/ath79/clock.c
190
t = 1;
arch/mips/ath79/clock.c
192
out_div = (1 << t);
arch/mips/ath79/clock.c
215
u64 t;
arch/mips/ath79/clock.c
218
t = ref;
arch/mips/ath79/clock.c
219
t *= nint;
arch/mips/ath79/clock.c
220
do_div(t, ref_div);
arch/mips/ath79/clock.c
221
ret = t;
arch/mips/ath79/clock.c
223
t = ref;
arch/mips/ath79/clock.c
224
t *= nfrac;
arch/mips/ath79/clock.c
225
do_div(t, ref_div * frac);
arch/mips/ath79/clock.c
226
ret += t;
arch/mips/ath79/common.c
113
t = ath79_reset_rr(reg);
arch/mips/ath79/common.c
114
ath79_reset_wr(reg, t | mask);
arch/mips/ath79/common.c
123
u32 t;
arch/mips/ath79/common.c
145
t = ath79_reset_rr(reg);
arch/mips/ath79/common.c
146
ath79_reset_wr(reg, t & ~mask);
arch/mips/ath79/common.c
91
u32 t;
arch/mips/ath79/early_printk.c
24
u32 t;
arch/mips/ath79/early_printk.c
27
t = __raw_readl(reg);
arch/mips/ath79/early_printk.c
28
if ((t & val) == val)
arch/mips/ath79/early_printk.c
61
u32 t;
arch/mips/ath79/early_printk.c
92
t = __raw_readl(gpio_base + AR71XX_GPIO_REG_FUNC);
arch/mips/ath79/early_printk.c
93
t |= uart_en;
arch/mips/ath79/early_printk.c
94
__raw_writel(t, gpio_base + AR71XX_GPIO_REG_FUNC);
arch/mips/cavium-octeon/octeon-irq.c
2170
static int octeon_irq_cib_set_type(struct irq_data *data, unsigned int t)
arch/mips/cavium-octeon/octeon-irq.c
2172
irqd_set_trigger_type(data, t);
arch/mips/cavium-octeon/octeon-irq.c
677
static int octeon_irq_ciu_set_type(struct irq_data *data, unsigned int t)
arch/mips/cavium-octeon/octeon-irq.c
679
irqd_set_trigger_type(data, t);
arch/mips/cavium-octeon/octeon-irq.c
681
if (t & IRQ_TYPE_EDGE_BOTH)
arch/mips/cavium-octeon/octeon-irq.c
693
u32 t = irqd_get_trigger_type(data);
arch/mips/cavium-octeon/octeon-irq.c
699
cfg.s.int_type = (t & IRQ_TYPE_EDGE_BOTH) != 0;
arch/mips/cavium-octeon/octeon-irq.c
700
cfg.s.rx_xor = (t & (IRQ_TYPE_LEVEL_LOW | IRQ_TYPE_EDGE_FALLING)) != 0;
arch/mips/cavium-octeon/octeon-irq.c
721
static int octeon_irq_ciu_gpio_set_type(struct irq_data *data, unsigned int t)
arch/mips/cavium-octeon/octeon-irq.c
723
irqd_set_trigger_type(data, t);
arch/mips/cavium-octeon/octeon-irq.c
726
if (t & IRQ_TYPE_EDGE_BOTH)
arch/mips/cavium-octeon/setup.c
665
u64 t;
arch/mips/cavium-octeon/setup.c
729
t = read_c0_cvmctl();
arch/mips/cavium-octeon/setup.c
730
if ((t & (1ull << 27)) == 0) {
arch/mips/include/asm/asmmacro.h
239
insn_if_mips 0x41000000 | (\rt << 16) | (\rs << 11) | (\u << 5) | (\sel)
arch/mips/include/asm/asmmacro.h
240
insn32_if_mm 0x0000000E | (\rt << 21) | (\rs << 16) | (\u << 10) | (\sel << 4)
arch/mips/include/asm/asmmacro.h
244
insn_if_mips 0x41800000 | (\rt << 16) | (\rs << 11) | (\u << 5) | (\sel)
arch/mips/include/asm/asmmacro.h
245
insn32_if_mm 0x00000006 | (\rt << 21) | (\rs << 16) | (\u << 10) | (\sel << 4)
arch/mips/include/asm/msa.h
134
static inline void save_msa(struct task_struct *t)
arch/mips/include/asm/msa.h
137
_save_msa(t);
arch/mips/include/asm/msa.h
140
static inline void restore_msa(struct task_struct *t)
arch/mips/include/asm/msa.h
143
_restore_msa(t);
arch/mips/include/asm/octeon/cvmx-fau.h
237
cvmx_fau_tagwait64_t t;
arch/mips/include/asm/octeon/cvmx-fau.h
241
return result.t;
arch/mips/include/asm/octeon/cvmx-fau.h
261
cvmx_fau_tagwait32_t t;
arch/mips/include/asm/octeon/cvmx-fau.h
266
return result.t;
arch/mips/include/asm/octeon/cvmx-fau.h
285
cvmx_fau_tagwait16_t t;
arch/mips/include/asm/octeon/cvmx-fau.h
290
return result.t;
arch/mips/include/asm/octeon/cvmx-fau.h
308
cvmx_fau_tagwait8_t t;
arch/mips/include/asm/octeon/cvmx-fau.h
312
return result.t;
arch/mips/include/asm/uaccess.h
234
__typeof__(*(addr)) t; \
arch/mips/include/asm/uaccess.h
255
(val) = __gu_tmp.t; \
arch/mips/include/asm/watch.h
15
void mips_install_watch_registers(struct task_struct *t);
arch/mips/kernel/elf.c
311
struct task_struct *t = current;
arch/mips/kernel/elf.c
318
t->thread.fpu.fcr31 = c->fpu_csr31;
arch/mips/kernel/elf.c
322
t->thread.fpu.fcr31 &= ~FPU_CSR_NAN2008;
arch/mips/kernel/elf.c
324
t->thread.fpu.fcr31 &= ~FPU_CSR_ABS2008;
arch/mips/kernel/elf.c
328
t->thread.fpu.fcr31 |= FPU_CSR_NAN2008;
arch/mips/kernel/elf.c
330
t->thread.fpu.fcr31 |= FPU_CSR_ABS2008;
arch/mips/kernel/process.c
531
struct thread_struct *t = &tsk->thread;
arch/mips/kernel/process.c
534
if (t->reg31 == (unsigned long) ret_from_fork)
arch/mips/kernel/process.c
535
return t->reg31;
arch/mips/kernel/process.c
538
return ((unsigned long *)t->reg29)[schedule_mfi.pc_offset];
arch/mips/kernel/process.c
788
struct task_struct *t;
arch/mips/kernel/process.c
824
for_each_thread(task, t) {
arch/mips/kernel/process.c
827
clear_tsk_thread_flag(t, TIF_32BIT_FPREGS);
arch/mips/kernel/process.c
829
set_tsk_thread_flag(t, TIF_32BIT_FPREGS);
arch/mips/kernel/process.c
830
clear_tsk_thread_flag(t, TIF_MSA_CTX_LIVE);
arch/mips/kernel/process.c
835
set_tsk_thread_flag(t, TIF_HYBRID_FPREGS);
arch/mips/kernel/process.c
837
clear_tsk_thread_flag(t, TIF_HYBRID_FPREGS);
arch/mips/kernel/process.c
851
for_each_thread(task, t)
arch/mips/kernel/process.c
852
cpumask_set_cpu(task_cpu(t), &process_cpus);
arch/mips/kernel/vpe-mt.c
114
| (t->index << VPECONF0_XTC_SHIFT));
arch/mips/kernel/vpe-mt.c
208
struct tc *t;
arch/mips/kernel/vpe-mt.c
213
t = list_entry(v->tc.next, struct tc, tc);
arch/mips/kernel/vpe-mt.c
214
if (t != NULL) {
arch/mips/kernel/vpe-mt.c
215
settc(t->index);
arch/mips/kernel/vpe-mt.c
229
struct tc *t;
arch/mips/kernel/vpe-mt.c
232
t = list_entry(v->tc.next, struct tc, tc);
arch/mips/kernel/vpe-mt.c
233
if (t == NULL)
arch/mips/kernel/vpe-mt.c
241
settc(t->index);
arch/mips/kernel/vpe-mt.c
31
struct tc *t;
arch/mips/kernel/vpe-mt.c
330
struct tc *t;
arch/mips/kernel/vpe-mt.c
395
t = alloc_tc(tc);
arch/mips/kernel/vpe-mt.c
396
if (!t) {
arch/mips/kernel/vpe-mt.c
419
list_add(&t->tc, &v->tc);
arch/mips/kernel/vpe-mt.c
446
t->pvpe = v; /* set the parent vpe */
arch/mips/kernel/vpe-mt.c
467
t->pvpe = get_vpe(0); /* set the parent vpe */
arch/mips/kernel/vpe-mt.c
57
t = list_first_entry(&v->tc, struct tc, tc);
arch/mips/kernel/vpe-mt.c
62
settc(t->index);
arch/mips/kernel/vpe-mt.c
72
t->index);
arch/mips/kernel/vpe.c
77
struct tc *res, *t;
arch/mips/kernel/vpe.c
81
list_for_each_entry(t, &vpecontrol.tc_list, list) {
arch/mips/kernel/vpe.c
82
if (t->index == index) {
arch/mips/kernel/vpe.c
83
res = t;
arch/mips/kernel/watch.c
120
unsigned int t;
arch/mips/kernel/watch.c
130
t = read_c0_watchlo0();
arch/mips/kernel/watch.c
132
c->watch_reg_masks[0] = t & MIPS_WATCHLO_IRW;
arch/mips/kernel/watch.c
138
t = read_c0_watchhi0();
arch/mips/kernel/watch.c
139
write_c0_watchhi0(t | MIPS_WATCHHI_MASK);
arch/mips/kernel/watch.c
141
t = read_c0_watchhi0();
arch/mips/kernel/watch.c
142
c->watch_reg_masks[0] |= (t & MIPS_WATCHHI_MASK);
arch/mips/kernel/watch.c
143
if ((t & MIPS_WATCHHI_M) == 0)
arch/mips/kernel/watch.c
148
t = read_c0_watchlo1();
arch/mips/kernel/watch.c
150
c->watch_reg_masks[1] = t & MIPS_WATCHLO_IRW;
arch/mips/kernel/watch.c
154
t = read_c0_watchhi1();
arch/mips/kernel/watch.c
155
write_c0_watchhi1(t | MIPS_WATCHHI_MASK);
arch/mips/kernel/watch.c
157
t = read_c0_watchhi1();
arch/mips/kernel/watch.c
158
c->watch_reg_masks[1] |= (t & MIPS_WATCHHI_MASK);
arch/mips/kernel/watch.c
159
if ((t & MIPS_WATCHHI_M) == 0)
arch/mips/kernel/watch.c
164
t = read_c0_watchlo2();
arch/mips/kernel/watch.c
166
c->watch_reg_masks[2] = t & MIPS_WATCHLO_IRW;
arch/mips/kernel/watch.c
170
t = read_c0_watchhi2();
arch/mips/kernel/watch.c
171
write_c0_watchhi2(t | MIPS_WATCHHI_MASK);
arch/mips/kernel/watch.c
173
t = read_c0_watchhi2();
arch/mips/kernel/watch.c
174
c->watch_reg_masks[2] |= (t & MIPS_WATCHHI_MASK);
arch/mips/kernel/watch.c
175
if ((t & MIPS_WATCHHI_M) == 0)
arch/mips/kernel/watch.c
18
void mips_install_watch_registers(struct task_struct *t)
arch/mips/kernel/watch.c
180
t = read_c0_watchlo3();
arch/mips/kernel/watch.c
182
c->watch_reg_masks[3] = t & MIPS_WATCHLO_IRW;
arch/mips/kernel/watch.c
186
t = read_c0_watchhi3();
arch/mips/kernel/watch.c
187
write_c0_watchhi3(t | MIPS_WATCHHI_MASK);
arch/mips/kernel/watch.c
189
t = read_c0_watchhi3();
arch/mips/kernel/watch.c
190
c->watch_reg_masks[3] |= (t & MIPS_WATCHHI_MASK);
arch/mips/kernel/watch.c
191
if ((t & MIPS_WATCHHI_M) == 0)
arch/mips/kernel/watch.c
196
t = read_c0_watchhi4();
arch/mips/kernel/watch.c
197
if ((t & MIPS_WATCHHI_M) == 0)
arch/mips/kernel/watch.c
20
struct mips3264_watch_reg_state *watches = &t->thread.watch.mips3264;
arch/mips/kernel/watch.c
201
t = read_c0_watchhi5();
arch/mips/kernel/watch.c
202
if ((t & MIPS_WATCHHI_M) == 0)
arch/mips/kernel/watch.c
206
t = read_c0_watchhi6();
arch/mips/kernel/watch.c
207
if ((t & MIPS_WATCHHI_M) == 0)
arch/mips/math-emu/cp1emu.c
1420
union ieee754##p s, union ieee754##p t) \
arch/mips/math-emu/cp1emu.c
1423
s = f1(s, t); \
arch/mips/math-emu/dp_maddf.c
202
t = DPXMULT(lxm, hym);
arch/mips/math-emu/dp_maddf.c
204
at = lrm + (t << 32);
arch/mips/math-emu/dp_maddf.c
208
hrm = hrm + (t >> 32);
arch/mips/math-emu/dp_maddf.c
210
t = DPXMULT(hxm, lym);
arch/mips/math-emu/dp_maddf.c
212
at = lrm + (t << 32);
arch/mips/math-emu/dp_maddf.c
216
hrm = hrm + (t >> 32);
arch/mips/math-emu/dp_maddf.c
302
t = 0;
arch/mips/math-emu/dp_maddf.c
303
while ((hzm >> (62 - t)) == 0)
arch/mips/math-emu/dp_maddf.c
304
t++;
arch/mips/math-emu/dp_maddf.c
306
assert(t <= 62);
arch/mips/math-emu/dp_maddf.c
307
if (t) {
arch/mips/math-emu/dp_maddf.c
308
hzm = hzm << t | lzm >> (64 - t);
arch/mips/math-emu/dp_maddf.c
309
lzm = lzm << t;
arch/mips/math-emu/dp_maddf.c
310
ze -= t;
arch/mips/math-emu/dp_maddf.c
53
u64 t;
arch/mips/math-emu/dp_mul.c
127
t = DPXMULT(lxm, hym);
arch/mips/math-emu/dp_mul.c
129
at = lrm + (t << 32);
arch/mips/math-emu/dp_mul.c
133
hrm = hrm + (t >> 32);
arch/mips/math-emu/dp_mul.c
135
t = DPXMULT(hxm, lym);
arch/mips/math-emu/dp_mul.c
137
at = lrm + (t << 32);
arch/mips/math-emu/dp_mul.c
141
hrm = hrm + (t >> 32);
arch/mips/math-emu/dp_mul.c
23
u64 t;
arch/mips/math-emu/dp_sqrt.c
100
z = t;
arch/mips/math-emu/dp_sqrt.c
101
t.bexp += 0x001;
arch/mips/math-emu/dp_sqrt.c
102
t = ieee754dp_add(t, z);
arch/mips/math-emu/dp_sqrt.c
106
t = ieee754dp_div(z, ieee754dp_add(t, x));
arch/mips/math-emu/dp_sqrt.c
107
t.bexp += 0x001;
arch/mips/math-emu/dp_sqrt.c
108
y = ieee754dp_add(y, t);
arch/mips/math-emu/dp_sqrt.c
117
t = ieee754dp_div(x, y);
arch/mips/math-emu/dp_sqrt.c
119
if (ieee754_csr.sx & IEEE754_INEXACT || t.bits != y.bits) {
arch/mips/math-emu/dp_sqrt.c
123
t.bits -= 1;
arch/mips/math-emu/dp_sqrt.c
134
t.bits += 1;
arch/mips/math-emu/dp_sqrt.c
139
y = ieee754dp_add(y, t);
arch/mips/math-emu/dp_sqrt.c
23
union ieee754dp y, z, t;
arch/mips/math-emu/dp_sqrt.c
92
t = ieee754dp_div(x, y);
arch/mips/math-emu/dp_sqrt.c
93
y = ieee754dp_add(y, t);
arch/mips/math-emu/dp_sqrt.c
99
t = ieee754dp_mul(y, y);
arch/mips/math-emu/sp_mul.c
126
t = lxm * hym; /* 16 * 16 => 32 */
arch/mips/math-emu/sp_mul.c
127
at = lrm + (t << 16);
arch/mips/math-emu/sp_mul.c
130
hrm = hrm + (t >> 16);
arch/mips/math-emu/sp_mul.c
132
t = hxm * lym; /* 16 * 16 => 32 */
arch/mips/math-emu/sp_mul.c
133
at = lrm + (t << 16);
arch/mips/math-emu/sp_mul.c
136
hrm = hrm + (t >> 16);
arch/mips/math-emu/sp_mul.c
23
unsigned int t;
arch/mips/math-emu/sp_sqrt.c
14
int ix, s, q, m, t, i;
arch/mips/math-emu/sp_sqrt.c
78
t = s + r;
arch/mips/math-emu/sp_sqrt.c
79
if (t <= ix) {
arch/mips/math-emu/sp_sqrt.c
80
s = t + r;
arch/mips/math-emu/sp_sqrt.c
81
ix -= t;
arch/mips/mm/cerr-sb1.c
425
uint64_t t;
arch/mips/mm/cerr-sb1.c
434
t = dword & mask_72_64[i];
arch/mips/mm/cerr-sb1.c
435
w = (uint32_t)(t >> 32);
arch/mips/mm/cerr-sb1.c
438
w = (uint32_t)(t & 0xFFFFFFFF);
arch/mips/mm/tlbex.c
1706
int t = scratch >= 0 ? scratch : pte;
arch/mips/mm/tlbex.c
1715
uasm_i_srl(p, t, cur, _PAGE_PRESENT_SHIFT);
arch/mips/mm/tlbex.c
1716
cur = t;
arch/mips/mm/tlbex.c
1718
uasm_i_andi(p, t, cur, 1);
arch/mips/mm/tlbex.c
1719
uasm_il_beqz(p, r, t, lid);
arch/mips/mm/tlbex.c
1720
if (pte == t)
arch/mips/mm/tlbex.c
1726
uasm_i_srl(p, t, cur, _PAGE_PRESENT_SHIFT);
arch/mips/mm/tlbex.c
1727
cur = t;
arch/mips/mm/tlbex.c
1729
uasm_i_andi(p, t, cur,
arch/mips/mm/tlbex.c
1731
uasm_i_xori(p, t, t, _PAGE_PRESENT >> _PAGE_PRESENT_SHIFT);
arch/mips/mm/tlbex.c
1732
uasm_il_bnez(p, r, t, lid);
arch/mips/mm/tlbex.c
1733
if (pte == t)
arch/mips/mm/tlbex.c
1758
int t = scratch >= 0 ? scratch : pte;
arch/mips/mm/tlbex.c
1762
uasm_i_srl(p, t, cur, _PAGE_PRESENT_SHIFT);
arch/mips/mm/tlbex.c
1763
cur = t;
arch/mips/mm/tlbex.c
1765
uasm_i_andi(p, t, cur,
arch/mips/mm/tlbex.c
1767
uasm_i_xori(p, t, t,
arch/mips/mm/tlbex.c
1769
uasm_il_bnez(p, r, t, lid);
arch/mips/mm/tlbex.c
1770
if (pte == t)
arch/mips/mm/tlbex.c
1803
int t = scratch >= 0 ? scratch : pte;
arch/mips/mm/tlbex.c
1804
uasm_i_srl(p, t, pte, _PAGE_WRITE_SHIFT);
arch/mips/mm/tlbex.c
1805
uasm_i_andi(p, t, t, 1);
arch/mips/mm/tlbex.c
1806
uasm_il_beqz(p, r, t, lid);
arch/mips/mm/tlbex.c
1807
if (pte == t)
arch/mips/pci/pci-ar71xx.c
258
u32 t;
arch/mips/pci/pci-ar71xx.c
263
t = __raw_readl(base + AR71XX_RESET_REG_PCI_INT_ENABLE);
arch/mips/pci/pci-ar71xx.c
264
__raw_writel(t | (1 << irq), base + AR71XX_RESET_REG_PCI_INT_ENABLE);
arch/mips/pci/pci-ar71xx.c
275
u32 t;
arch/mips/pci/pci-ar71xx.c
280
t = __raw_readl(base + AR71XX_RESET_REG_PCI_INT_ENABLE);
arch/mips/pci/pci-ar71xx.c
281
__raw_writel(t & ~(1 << irq), base + AR71XX_RESET_REG_PCI_INT_ENABLE);
arch/mips/pci/pci-ar71xx.c
332
u32 t;
arch/mips/pci/pci-ar71xx.c
371
t = PCI_COMMAND_MEMORY | PCI_COMMAND_MASTER | PCI_COMMAND_INVALIDATE
arch/mips/pci/pci-ar71xx.c
373
ar71xx_pci_local_write(apc, PCI_COMMAND, 4, t);
arch/mips/pci/pci-ar71xx.c
71
u32 t;
arch/mips/pci/pci-ar71xx.c
73
t = ar71xx_pci_ble_table[size & 3][where & 3];
arch/mips/pci/pci-ar71xx.c
74
BUG_ON(t == 0xf);
arch/mips/pci/pci-ar71xx.c
75
t <<= (local) ? 20 : 4;
arch/mips/pci/pci-ar71xx.c
77
return t;
arch/mips/pci/pci-ar724x.c
253
u32 t;
arch/mips/pci/pci-ar724x.c
261
t = __raw_readl(base + AR724X_PCI_REG_INT_MASK);
arch/mips/pci/pci-ar724x.c
262
__raw_writel(t | AR724X_PCI_INT_DEV0,
arch/mips/pci/pci-ar724x.c
274
u32 t;
arch/mips/pci/pci-ar724x.c
282
t = __raw_readl(base + AR724X_PCI_REG_INT_MASK);
arch/mips/pci/pci-ar724x.c
283
__raw_writel(t & ~AR724X_PCI_INT_DEV0,
arch/mips/pci/pci-ar724x.c
289
t = __raw_readl(base + AR724X_PCI_REG_INT_STATUS);
arch/mips/pci/pci-ar724x.c
290
__raw_writel(t | AR724X_PCI_INT_DEV0,
arch/mips/pci/pci-rt3883.c
152
u32 t;
arch/mips/pci/pci-rt3883.c
156
t = rt3883_pci_r32(rpc, RT3883_PCI_REG_PCIENA);
arch/mips/pci/pci-rt3883.c
157
rt3883_pci_w32(rpc, t | BIT(d->hwirq), RT3883_PCI_REG_PCIENA);
arch/mips/pci/pci-rt3883.c
165
u32 t;
arch/mips/pci/pci-rt3883.c
169
t = rt3883_pci_r32(rpc, RT3883_PCI_REG_PCIENA);
arch/mips/pci/pci-rt3883.c
170
rt3883_pci_w32(rpc, t & ~BIT(d->hwirq), RT3883_PCI_REG_PCIENA);
arch/mips/pci/pci-rt3883.c
305
u32 t;
arch/mips/pci/pci-rt3883.c
320
t = rt_sysc_r32(RT3883_SYSC_REG_PCIE_CLK_GEN0);
arch/mips/pci/pci-rt3883.c
321
t &= ~BIT(31);
arch/mips/pci/pci-rt3883.c
322
rt_sysc_w32(t, RT3883_SYSC_REG_PCIE_CLK_GEN0);
arch/mips/pci/pci-rt3883.c
324
t = rt_sysc_r32(RT3883_SYSC_REG_PCIE_CLK_GEN1);
arch/mips/pci/pci-rt3883.c
325
t &= 0x80ffffff;
arch/mips/pci/pci-rt3883.c
326
rt_sysc_w32(t, RT3883_SYSC_REG_PCIE_CLK_GEN1);
arch/mips/pci/pci-rt3883.c
328
t = rt_sysc_r32(RT3883_SYSC_REG_PCIE_CLK_GEN1);
arch/mips/pci/pci-rt3883.c
329
t |= 0xa << 24;
arch/mips/pci/pci-rt3883.c
330
rt_sysc_w32(t, RT3883_SYSC_REG_PCIE_CLK_GEN1);
arch/mips/pci/pci-rt3883.c
332
t = rt_sysc_r32(RT3883_SYSC_REG_PCIE_CLK_GEN0);
arch/mips/pci/pci-rt3883.c
333
t |= BIT(31);
arch/mips/pci/pci-rt3883.c
334
rt_sysc_w32(t, RT3883_SYSC_REG_PCIE_CLK_GEN0);
arch/mips/pci/pci-rt3883.c
366
t = (RT3883_P2P_BR_DEVNUM << RT3883_PCICFG_P2P_BR_DEVNUM_S);
arch/mips/pci/pci-rt3883.c
367
rt3883_pci_w32(rpc, t, RT3883_PCI_REG_PCICFG);
arch/mips/pci/pci-rt3883.c
376
t = rt3883_pci_r32(rpc, RT3883_PCI_REG_STATUS(1));
arch/mips/pci/pci-rt3883.c
378
rpc->pcie_ready = t & BIT(0);
arch/mips/pci/pci-rt3883.c
382
t = rt_sysc_r32(RT3883_SYSC_REG_RSTCTRL);
arch/mips/pci/pci-rt3883.c
383
t |= RT3883_RSTCTRL_PCIE;
arch/mips/pci/pci-rt3883.c
384
rt_sysc_w32(t, RT3883_SYSC_REG_RSTCTRL);
arch/mips/pci/pci-rt3883.c
385
t &= ~RT3883_RSTCTRL_PCIE;
arch/mips/pci/pci-rt3883.c
386
rt_sysc_w32(t, RT3883_SYSC_REG_RSTCTRL);
arch/mips/pci/pci-rt3883.c
389
t = rt_sysc_r32(RT3883_SYSC_REG_CLKCFG1);
arch/mips/pci/pci-rt3883.c
390
t &= ~RT3883_CLKCFG1_PCIE_CLK_EN;
arch/mips/pci/pci-rt3883.c
391
rt_sysc_w32(t, RT3883_SYSC_REG_CLKCFG1);
arch/mips/pci/pci-rt3883.c
393
t = rt_sysc_r32(RT3883_SYSC_REG_PCIE_CLK_GEN0);
arch/mips/pci/pci-rt3883.c
394
t &= ~0xf000c080;
arch/mips/pci/pci-rt3883.c
395
rt_sysc_w32(t, RT3883_SYSC_REG_PCIE_CLK_GEN0);
arch/mips/ralink/rt305x.c
29
u32 t;
arch/mips/ralink/rt305x.c
31
t = __raw_readl(RT305X_SYSC_BASE + SYSC_REG_SYSTEM_CONFIG);
arch/mips/ralink/rt305x.c
32
t = (t >> RT5350_SYSCFG0_DRAM_SIZE_SHIFT) &
arch/mips/ralink/rt305x.c
35
switch (t) {
arch/mips/ralink/rt305x.c
52
panic("rt5350: invalid DRAM size: %u", t);
arch/mips/ralink/timer.c
70
u32 t = TMR0CTL_MODE_PERIODIC | TMR0CTL_PRESCALE_VAL;
arch/mips/ralink/timer.c
71
rt_timer_w32(rt, TIMER_REG_TMR0CTL, t);
arch/mips/ralink/timer.c
90
u32 t;
arch/mips/ralink/timer.c
94
t = rt_timer_r32(rt, TIMER_REG_TMR0CTL);
arch/mips/ralink/timer.c
95
t |= TMR0CTL_ENABLE;
arch/mips/ralink/timer.c
96
rt_timer_w32(rt, TIMER_REG_TMR0CTL, t);
arch/mips/sibyte/swarm/rtc_m41t81.c
139
int m41t81_set_time(time64_t t)
arch/mips/sibyte/swarm/rtc_m41t81.c
145
rtc_time64_to_tm(t, &tm);
arch/mips/sibyte/swarm/rtc_xicor1241.c
108
int xicor_set_time(time64_t t)
arch/mips/sibyte/swarm/rtc_xicor1241.c
114
rtc_time64_to_tm(t, &tm);
arch/parisc/include/asm/assembly.h
135
.macro shlw r, sa, t
arch/parisc/include/asm/assembly.h
140
.macro shld r, sa, t
arch/parisc/include/asm/assembly.h
145
.macro shr r, sa, t
arch/parisc/include/asm/assembly.h
150
.macro shrd r, sa, t
arch/parisc/include/asm/assembly.h
157
.macro extru_safe r, p, len, t
arch/parisc/include/asm/assembly.h
167
.macro depi_safe i, p, len, t
arch/parisc/include/asm/assembly.h
177
.macro dep_safe i, p, len, t
arch/parisc/include/asm/compat.h
131
static inline int __is_compat_task(struct task_struct *t)
arch/parisc/include/asm/compat.h
133
return test_tsk_thread_flag(t, TIF_32BIT);
arch/parisc/include/asm/psw.h
73
unsigned int t:1;
arch/parisc/include/asm/uaccess.h
108
(val) = __gu_tmp.t; \
arch/parisc/include/asm/uaccess.h
96
__typeof__(*(ptr)) t; \
arch/parisc/include/asm/unwind.h
59
struct task_struct *t;
arch/parisc/include/asm/unwind.h
74
void unwind_frame_init(struct unwind_frame_info *info, struct task_struct *t,
arch/parisc/include/asm/unwind.h
77
struct task_struct *t);
arch/parisc/kernel/module.c
140
int s, t;
arch/parisc/kernel/module.c
143
t = (as16 << 1) & 0xffff;
arch/parisc/kernel/module.c
145
return (t ^ s ^ (s >> 1)) | (s >> 15);
arch/parisc/kernel/module.c
530
#define r(t) ELF32_R_TYPE(rel[i].r_info)==t ? #t :
arch/parisc/kernel/module.c
675
#define r(t) ELF64_R_TYPE(rel[i].r_info)==t ? #t :
arch/parisc/kernel/ptrace.c
105
pa_psw(task)->t = 0;
arch/parisc/kernel/ptrace.c
117
pa_psw(task)->t = 1;
arch/parisc/kernel/ptrace.c
57
pa_psw(task)->t = 0;
arch/parisc/kernel/traps.c
209
void show_stack(struct task_struct *t, unsigned long *sp, const char *loglvl)
arch/parisc/kernel/traps.c
211
parisc_show_stack(t, NULL, loglvl);
arch/parisc/kernel/unwind.c
309
if ((info->prev_sp - (unsigned long) task_stack_page(info->t))
arch/parisc/kernel/unwind.c
392
void unwind_frame_init(struct unwind_frame_info *info, struct task_struct *t,
arch/parisc/kernel/unwind.c
396
info->t = t;
arch/parisc/kernel/unwind.c
403
t ? (int)t->pid : -1, info->sp, info->ip);
arch/parisc/kernel/unwind.c
406
void unwind_frame_init_from_blocked_task(struct unwind_frame_info *info, struct task_struct *t)
arch/parisc/kernel/unwind.c
408
struct pt_regs *r = &t->thread.regs;
arch/parisc/kernel/unwind.c
417
unwind_frame_init(info, t, r2);
arch/parisc/kernel/unwind.c
462
next_frame->t ? (int)next_frame->t->pid : -1,
arch/parisc/math-emu/fpudispatch.c
1070
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
1073
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
1080
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
1083
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
1100
if (t & 1)
arch/parisc/math-emu/fpudispatch.c
1116
&fpregs[r2],&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
1119
&fpregs[r2],&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
1127
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
1130
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
1137
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
1140
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
1389
u_int rm1, rm2, ra, t; /* operands */
arch/parisc/math-emu/fpudispatch.c
1404
t = extru(ir,fptpos,5) * sizeof(double)/sizeof(u_int);
arch/parisc/math-emu/fpudispatch.c
1405
if (t == 0)
arch/parisc/math-emu/fpudispatch.c
1410
&fpregs[ra], &fpregs[0], &fpregs[t]));
arch/parisc/math-emu/fpudispatch.c
1413
&fpregs[ra], &fpregs[0], &fpregs[t]));
arch/parisc/math-emu/fpudispatch.c
1426
t = ((extru(ir,fptpos,5)<<1)|(extru(ir,fpxtpos,1)));
arch/parisc/math-emu/fpudispatch.c
1427
if (t == 0)
arch/parisc/math-emu/fpudispatch.c
1432
&fpregs[ra], &fpregs[0], &fpregs[t]));
arch/parisc/math-emu/fpudispatch.c
1435
&fpregs[ra], &fpregs[0], &fpregs[t]));
arch/parisc/math-emu/fpudispatch.c
277
u_int r1,r2,t; /* operand register offsets */
arch/parisc/math-emu/fpudispatch.c
293
t = extru(ir,fptpos,5) * sizeof(double)/sizeof(u_int);
arch/parisc/math-emu/fpudispatch.c
294
if (t == 0 && class != 2) /* don't allow fr0 as a dest */
arch/parisc/math-emu/fpudispatch.c
309
t &= ~3; /* force to even reg #s */
arch/parisc/math-emu/fpudispatch.c
311
fpregs[t+3] = fpregs[r1+3];
arch/parisc/math-emu/fpudispatch.c
312
fpregs[t+2] = fpregs[r1+2];
arch/parisc/math-emu/fpudispatch.c
315
fpregs[t+1] = fpregs[r1+1];
arch/parisc/math-emu/fpudispatch.c
318
fpregs[t] = fpregs[r1];
arch/parisc/math-emu/fpudispatch.c
327
t &= ~3; /* force to even reg #s */
arch/parisc/math-emu/fpudispatch.c
329
fpregs[t+3] = fpregs[r1+3];
arch/parisc/math-emu/fpudispatch.c
330
fpregs[t+2] = fpregs[r1+2];
arch/parisc/math-emu/fpudispatch.c
333
fpregs[t+1] = fpregs[r1+1];
arch/parisc/math-emu/fpudispatch.c
337
fpregs[t] = fpregs[r1] & 0x7fffffff;
arch/parisc/math-emu/fpudispatch.c
346
t &= ~3; /* force to even reg #s */
arch/parisc/math-emu/fpudispatch.c
348
fpregs[t+3] = fpregs[r1+3];
arch/parisc/math-emu/fpudispatch.c
349
fpregs[t+2] = fpregs[r1+2];
arch/parisc/math-emu/fpudispatch.c
352
fpregs[t+1] = fpregs[r1+1];
arch/parisc/math-emu/fpudispatch.c
356
fpregs[t] = fpregs[r1] ^ 0x80000000;
arch/parisc/math-emu/fpudispatch.c
365
t &= ~3; /* force to even reg #s */
arch/parisc/math-emu/fpudispatch.c
367
fpregs[t+3] = fpregs[r1+3];
arch/parisc/math-emu/fpudispatch.c
368
fpregs[t+2] = fpregs[r1+2];
arch/parisc/math-emu/fpudispatch.c
371
fpregs[t+1] = fpregs[r1+1];
arch/parisc/math-emu/fpudispatch.c
375
fpregs[t] = fpregs[r1] | 0x80000000;
arch/parisc/math-emu/fpudispatch.c
383
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
386
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
396
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
399
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
428
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
431
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
440
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
443
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
446
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
449
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
456
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
459
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
462
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
465
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
472
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
475
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
478
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
481
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
488
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
491
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
494
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
497
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
504
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
507
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
510
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
513
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
520
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
523
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
526
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
529
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
651
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
654
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
664
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
667
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
677
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
680
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
690
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
693
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
703
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
706
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
724
u_int r1,r2,t; /* operand register offsets */
arch/parisc/math-emu/fpudispatch.c
736
t = ((extru(ir,fptpos,5)<<1)|(extru(ir,fpxtpos,1)));
arch/parisc/math-emu/fpudispatch.c
737
if (t == 0 && class != 2)
arch/parisc/math-emu/fpudispatch.c
751
t &= ~1;
arch/parisc/math-emu/fpudispatch.c
766
fpregs[t+1] = fpregs[r1+1];
arch/parisc/math-emu/fpudispatch.c
769
fpregs[t] = fpregs[r1];
arch/parisc/math-emu/fpudispatch.c
779
fpregs[t+1] = fpregs[r1+1];
arch/parisc/math-emu/fpudispatch.c
782
fpregs[t] = fpregs[r1] & 0x7fffffff;
arch/parisc/math-emu/fpudispatch.c
792
fpregs[t+1] = fpregs[r1+1];
arch/parisc/math-emu/fpudispatch.c
795
fpregs[t] = fpregs[r1] ^ 0x80000000;
arch/parisc/math-emu/fpudispatch.c
805
fpregs[t+1] = fpregs[r1+1];
arch/parisc/math-emu/fpudispatch.c
808
fpregs[t] = fpregs[r1] | 0x80000000;
arch/parisc/math-emu/fpudispatch.c
816
&fpregs[t], status));
arch/parisc/math-emu/fpudispatch.c
819
&fpregs[t], status));
arch/parisc/math-emu/fpudispatch.c
829
&fpregs[t], status));
arch/parisc/math-emu/fpudispatch.c
832
&fpregs[t], status));
arch/parisc/math-emu/fpudispatch.c
846
t &= ~1;
arch/parisc/math-emu/fpudispatch.c
859
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
862
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
871
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
874
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
877
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
880
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
887
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
890
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
893
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
896
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
903
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
906
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
909
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
912
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
919
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
922
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
925
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
928
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
935
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
938
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
941
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
944
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
951
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
954
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
957
&fpregs[t],status));
arch/parisc/math-emu/fpudispatch.c
960
&fpregs[t],status));
arch/parisc/mm/fault.c
229
const char *t = NULL;
arch/parisc/mm/fault.c
232
t = trap_description[code];
arch/parisc/mm/fault.c
234
return t ? t : "Unknown trap";
arch/parisc/mm/init.c
751
unsigned long s, t, e;
arch/parisc/mm/init.c
764
t = (uintptr_t) dereference_function_descriptor(&_sdata);
arch/parisc/mm/init.c
765
BUG_ON(t != e);
arch/parisc/mm/init.c
772
t = (uintptr_t) dereference_function_descriptor(&_edata);
arch/parisc/mm/init.c
774
BUG_ON(t != e);
arch/parisc/net/bpf_jit.h
374
unsigned s, t;
arch/parisc/net/bpf_jit.h
377
t = (as16 << 1) & 0xffff;
arch/parisc/net/bpf_jit.h
379
return (t ^ s ^ (s >> 1)) | (s >> 15);
arch/parisc/net/bpf_jit.h
422
static inline u32 hppa_t6_insn(u8 opcode, u8 r2, u8 r1, u8 c, u8 f, u8 ext6, u16 t)
arch/parisc/net/bpf_jit.h
425
(ext6 << 6) | t);
arch/parisc/net/bpf_jit.h
429
static inline u32 hppa_t7_insn(u8 opcode, u8 r, u8 t, u32 im11)
arch/parisc/net/bpf_jit.h
431
return ((opcode << 26) | (r << 21) | (t << 16) | low_sign_unext(im11, 11));
arch/parisc/net/bpf_jit.h
435
static inline u32 hppa_t10_insn(u8 opcode, u8 r2, u8 r1, u8 c, u8 ext3, u8 cp, u8 t)
arch/parisc/net/bpf_jit.h
438
(ext3 << 10) | (cp << 5) | t);
arch/parisc/net/bpf_jit.h
461
static inline u32 hppa_t21_insn(u8 opcode, u8 r2, u8 r1, u8 ext8, u8 t)
arch/parisc/net/bpf_jit.h
463
return ((opcode << 26) | (r2 << 21) | (r1 << 16) | (ext8 << 5) | t);
arch/powerpc/include/asm/atomic.h
141
int t;
arch/powerpc/include/asm/atomic.h
154
: "=&r" (t)
arch/powerpc/include/asm/atomic.h
158
return t;
arch/powerpc/include/asm/atomic.h
169
int t;
arch/powerpc/include/asm/atomic.h
181
2:" : "=&b" (t)
arch/powerpc/include/asm/atomic.h
185
return t;
arch/powerpc/include/asm/atomic.h
195
s64 t;
arch/powerpc/include/asm/atomic.h
199
__asm__ __volatile__("ld %0,0(%1)" : "=r"(t) : "b"(&v->counter));
arch/powerpc/include/asm/atomic.h
201
__asm__ __volatile__("ld%U1%X1 %0,%1" : "=r"(t) : DS_FORM_CONSTRAINT (v->counter));
arch/powerpc/include/asm/atomic.h
203
return t;
arch/powerpc/include/asm/atomic.h
218
s64 t; \
arch/powerpc/include/asm/atomic.h
225
: "=&r" (t), "+m" (v->counter) \
arch/powerpc/include/asm/atomic.h
234
s64 t; \
arch/powerpc/include/asm/atomic.h
241
: "=&r" (t), "+m" (v->counter) \
arch/powerpc/include/asm/atomic.h
245
return t; \
arch/powerpc/include/asm/atomic.h
252
s64 res, t; \
arch/powerpc/include/asm/atomic.h
259
: "=&r" (res), "=&r" (t), "+m" (v->counter) \
arch/powerpc/include/asm/atomic.h
29
int t;
arch/powerpc/include/asm/atomic.h
300
s64 t;
arch/powerpc/include/asm/atomic.h
307
: "=&r" (t), "+m" (v->counter)
arch/powerpc/include/asm/atomic.h
315
s64 t;
arch/powerpc/include/asm/atomic.h
322
: "=&r" (t), "+m" (v->counter)
arch/powerpc/include/asm/atomic.h
326
return t;
arch/powerpc/include/asm/atomic.h
33
__asm__ __volatile__("lwz %0,0(%1)" : "=r"(t) : "b"(&v->counter));
arch/powerpc/include/asm/atomic.h
331
s64 t;
arch/powerpc/include/asm/atomic.h
338
: "=&r" (t), "+m" (v->counter)
arch/powerpc/include/asm/atomic.h
346
s64 t;
arch/powerpc/include/asm/atomic.h
35
__asm__ __volatile__("lwz%U1%X1 %0,%1" : "=r"(t) : "m<>"(v->counter));
arch/powerpc/include/asm/atomic.h
353
: "=&r" (t), "+m" (v->counter)
arch/powerpc/include/asm/atomic.h
357
return t;
arch/powerpc/include/asm/atomic.h
369
s64 t;
arch/powerpc/include/asm/atomic.h
37
return t;
arch/powerpc/include/asm/atomic.h
380
2:" : "=&r" (t)
arch/powerpc/include/asm/atomic.h
384
return t;
arch/powerpc/include/asm/atomic.h
399
s64 t;
arch/powerpc/include/asm/atomic.h
412
: "=&r" (t)
arch/powerpc/include/asm/atomic.h
416
return t;
arch/powerpc/include/asm/atomic.h
52
int t; \
arch/powerpc/include/asm/atomic.h
59
: "=&r" (t), "+m" (v->counter) \
arch/powerpc/include/asm/atomic.h
67
int t; \
arch/powerpc/include/asm/atomic.h
74
: "=&r" (t), "+m" (v->counter) \
arch/powerpc/include/asm/atomic.h
78
return t; \
arch/powerpc/include/asm/atomic.h
84
int res, t; \
arch/powerpc/include/asm/atomic.h
91
: "=&r" (res), "=&r" (t), "+m" (v->counter) \
arch/powerpc/include/asm/bitops.h
156
unsigned long old, t; \
arch/powerpc/include/asm/bitops.h
165
: "=&r" (old), "=&r" (t) \
arch/powerpc/include/asm/bitops.h
180
unsigned long old, t;
arch/powerpc/include/asm/bitops.h
192
: "=&r" (old), "=&r" (t)
arch/powerpc/include/asm/bitops.h
203
: "=&r" (old), "=&r" (t)
arch/powerpc/include/asm/bitops.h
239
unsigned long old, t;
arch/powerpc/include/asm/bitops.h
247
: "=&r" (old), "=&r" (t)
arch/powerpc/include/asm/cpm1.h
483
#define RCCR_TIMEP(t) (((t) & 0x3F)<<8) /* RISC Timer Period */
arch/powerpc/include/asm/local.h
106
long t;
arch/powerpc/include/asm/local.h
110
t = l->v;
arch/powerpc/include/asm/local.h
114
return t;
arch/powerpc/include/asm/local.h
43
long t; \
arch/powerpc/include/asm/local.h
47
t = (l->v c_op a); \
arch/powerpc/include/asm/local.h
50
return t; \
arch/powerpc/include/asm/local.h
81
long t;
arch/powerpc/include/asm/local.h
85
t = l->v;
arch/powerpc/include/asm/local.h
86
if (t == o)
arch/powerpc/include/asm/local.h
90
return t;
arch/powerpc/include/asm/ppc-opcode.h
326
#define ___PPC_RT(t) ___PPC_RS(t)
arch/powerpc/include/asm/ppc-opcode.h
334
#define __PPC_RT(t) ___PPC_RT(__REG_##t)
arch/powerpc/include/asm/ppc-opcode.h
341
#define __PPC_T_TLB(t) (((t) & 0x3) << 21)
arch/powerpc/include/asm/ppc-opcode.h
352
#define __PPC_CT(t) (((t) & 0x0f) << 21)
arch/powerpc/include/asm/ppc-opcode.h
376
#define PPC_RAW_DARN(t, l) (0x7c0005e6 | ___PPC_RT(t) | (((l) & 0x3) << 16))
arch/powerpc/include/asm/ppc-opcode.h
379
#define PPC_RAW_LQARX(t, a, b, eh) (0x7c000228 | ___PPC_RT(t) | ___PPC_RA(a) | ___PPC_RB(b) | __PPC_EH(eh))
arch/powerpc/include/asm/ppc-opcode.h
380
#define PPC_RAW_LDARX(t, a, b, eh) (0x7c0000a8 | ___PPC_RT(t) | ___PPC_RA(a) | ___PPC_RB(b) | __PPC_EH(eh))
arch/powerpc/include/asm/ppc-opcode.h
381
#define PPC_RAW_LWARX(t, a, b, eh) (0x7c000028 | ___PPC_RT(t) | ___PPC_RA(a) | ___PPC_RB(b) | __PPC_EH(eh))
arch/powerpc/include/asm/ppc-opcode.h
384
#define PPC_RAW_STQCX(t, a, b) (0x7c00016d | ___PPC_RT(t) | ___PPC_RA(a) | ___PPC_RB(b))
arch/powerpc/include/asm/ppc-opcode.h
385
#define PPC_RAW_MADDHD(t, a, b, c) (0x10000030 | ___PPC_RT(t) | ___PPC_RA(a) | ___PPC_RB(b) | ___PPC_RC(c))
arch/powerpc/include/asm/ppc-opcode.h
386
#define PPC_RAW_MADDHDU(t, a, b, c) (0x10000031 | ___PPC_RT(t) | ___PPC_RA(a) | ___PPC_RB(b) | ___PPC_RC(c))
arch/powerpc/include/asm/ppc-opcode.h
387
#define PPC_RAW_MADDLD(t, a, b, c) (0x10000033 | ___PPC_RT(t) | ___PPC_RA(a) | ___PPC_RB(b) | ___PPC_RC(c))
arch/powerpc/include/asm/ppc-opcode.h
401
#define PPC_RAW_TLBILX(t, a, b) (0x7c000024 | __PPC_T_TLB(t) | __PPC_RA0(a) | __PPC_RB(b))
arch/powerpc/include/asm/ppc-opcode.h
414
#define PPC_RAW_ERATILX(t, a, b) (0x7c000066 | __PPC_T_TLB(t) | __PPC_RA0(a) | __PPC_RB(b))
arch/powerpc/include/asm/ppc-opcode.h
416
#define PPC_RAW_ERATSX(t, a, w) (0x7c000126 | __PPC_RS(t) | __PPC_RA0(a) | __PPC_RB(b))
arch/powerpc/include/asm/ppc-opcode.h
417
#define PPC_RAW_ERATSX_DOT(t, a, w) (0x7c000127 | __PPC_RS(t) | __PPC_RA0(a) | __PPC_RB(b))
arch/powerpc/include/asm/ppc-opcode.h
418
#define PPC_RAW_SLBFEE_DOT(t, b) (0x7c0007a7 | __PPC_RT(t) | __PPC_RB(b))
arch/powerpc/include/asm/ppc-opcode.h
419
#define __PPC_RAW_SLBFEE_DOT(t, b) (0x7c0007a7 | ___PPC_RT(t) | ___PPC_RB(b))
arch/powerpc/include/asm/ppc-opcode.h
421
#define PPC_RAW_LBZCIX(t, a, b) (0x7c0006aa | __PPC_RT(t) | __PPC_RA(a) | __PPC_RB(b))
arch/powerpc/include/asm/ppc-opcode.h
435
#define VSX_XX3(t, a, b) (__PPC_XT(t) | __PPC_XA(a) | __PPC_XB(b))
arch/powerpc/include/asm/ppc-opcode.h
438
#define PPC_RAW_MFVRD(a, t) (0x7c000066 | VSX_XX1((t) + 32, a, R0))
arch/powerpc/include/asm/ppc-opcode.h
439
#define PPC_RAW_MTVRD(t, a) (0x7c000166 | VSX_XX1((t) + 32, a, R0))
arch/powerpc/include/asm/ppc-opcode.h
440
#define PPC_RAW_VPMSUMW(t, a, b) (0x10000488 | VSX_XX3((t), a, b))
arch/powerpc/include/asm/ppc-opcode.h
441
#define PPC_RAW_VPMSUMD(t, a, b) (0x100004c8 | VSX_XX3((t), a, b))
arch/powerpc/include/asm/ppc-opcode.h
442
#define PPC_RAW_XXLOR(t, a, b) (0xf0000490 | VSX_XX3((t), a, b))
arch/powerpc/include/asm/ppc-opcode.h
443
#define PPC_RAW_XXSWAPD(t, a) (0xf0000250 | VSX_XX3((t), a, a))
arch/powerpc/include/asm/ppc-opcode.h
444
#define PPC_RAW_XVCPSGNDP(t, a, b) ((0xf0000780 | VSX_XX3((t), (a), (b))))
arch/powerpc/include/asm/ppc-opcode.h
477
#define PPC_RAW_LWZX(t, a, b) (0x7c00002e | ___PPC_RT(t) | ___PPC_RA(a) | ___PPC_RB(b))
arch/powerpc/include/asm/ppc-opcode.h
480
#define PPC_RAW_LFSX(t, a, b) (0x7c00042e | ___PPC_RT(t) | ___PPC_RA(a) | ___PPC_RB(b))
arch/powerpc/include/asm/ppc-opcode.h
482
#define PPC_RAW_LFDX(t, a, b) (0x7c0004ae | ___PPC_RT(t) | ___PPC_RA(a) | ___PPC_RB(b))
arch/powerpc/include/asm/ppc-opcode.h
484
#define PPC_RAW_LVX(t, a, b) (0x7c0000ce | ___PPC_RT(t) | ___PPC_RA(a) | ___PPC_RB(b))
arch/powerpc/include/asm/ppc-opcode.h
486
#define PPC_RAW_ADDE(t, a, b) (0x7c000114 | ___PPC_RT(t) | ___PPC_RA(a) | ___PPC_RB(b))
arch/powerpc/include/asm/ppc-opcode.h
487
#define PPC_RAW_ADDZE(t, a) (0x7c000194 | ___PPC_RT(t) | ___PPC_RA(a))
arch/powerpc/include/asm/ppc-opcode.h
488
#define PPC_RAW_ADDME(t, a) (0x7c0001d4 | ___PPC_RT(t) | ___PPC_RA(a))
arch/powerpc/include/asm/ppc-opcode.h
489
#define PPC_RAW_ADD(t, a, b) (0x7c000214 | ___PPC_RT(t) | ___PPC_RA(a) | ___PPC_RB(b))
arch/powerpc/include/asm/ppc-opcode.h
490
#define PPC_RAW_ADD_DOT(t, a, b) (0x7c000214 | ___PPC_RT(t) | ___PPC_RA(a) | ___PPC_RB(b) | 0x1)
arch/powerpc/include/asm/ppc-opcode.h
491
#define PPC_RAW_ADDC(t, a, b) (0x7c000014 | ___PPC_RT(t) | ___PPC_RA(a) | ___PPC_RB(b))
arch/powerpc/include/asm/ppc-opcode.h
492
#define PPC_RAW_ADDC_DOT(t, a, b) (0x7c000014 | ___PPC_RT(t) | ___PPC_RA(a) | ___PPC_RB(b) | 0x1)
arch/powerpc/include/asm/ppc-opcode.h
497
#define PPC_RAW_MFLR(t) (0x7c0802a6 | ___PPC_RT(t))
arch/powerpc/include/asm/ppc-opcode.h
542
#define PPC_RAW_DIVDE(t, a, b) (0x7c000352 | ___PPC_RT(t) | ___PPC_RA(a) | ___PPC_RB(b))
arch/powerpc/include/asm/ppc-opcode.h
543
#define PPC_RAW_DIVDE_DOT(t, a, b) (0x7c000352 | ___PPC_RT(t) | ___PPC_RA(a) | ___PPC_RB(b) | 0x1)
arch/powerpc/include/asm/ppc-opcode.h
544
#define PPC_RAW_DIVDEU(t, a, b) (0x7c000312 | ___PPC_RT(t) | ___PPC_RA(a) | ___PPC_RB(b))
arch/powerpc/include/asm/ppc-opcode.h
545
#define PPC_RAW_DIVDEU_DOT(t, a, b) (0x7c000312 | ___PPC_RT(t) | ___PPC_RA(a) | ___PPC_RB(b) | 0x1)
arch/powerpc/include/asm/ppc-opcode.h
598
#define PPC_RAW_SETB(t, bfa) (0x7c000100 | ___PPC_RT(t) | ___PPC_RA((bfa) << 2))
arch/powerpc/include/asm/ppc-opcode.h
616
#define PPC_DARN(t, l) stringify_in_c(.long PPC_RAW_DARN(t, l))
arch/powerpc/include/asm/ppc-opcode.h
619
#define PPC_DIVDE(t, a, b) stringify_in_c(.long PPC_RAW_DIVDE(t, a, b))
arch/powerpc/include/asm/ppc-opcode.h
620
#define PPC_DIVDEU(t, a, b) stringify_in_c(.long PPC_RAW_DIVDEU(t, a, b))
arch/powerpc/include/asm/ppc-opcode.h
622
#define PPC_LQARX(t, a, b, eh) stringify_in_c(.long PPC_RAW_LQARX(t, a, b, eh))
arch/powerpc/include/asm/ppc-opcode.h
623
#define PPC_STQCX(t, a, b) stringify_in_c(.long PPC_RAW_STQCX(t, a, b))
arch/powerpc/include/asm/ppc-opcode.h
624
#define PPC_MADDHD(t, a, b, c) stringify_in_c(.long PPC_RAW_MADDHD(t, a, b, c))
arch/powerpc/include/asm/ppc-opcode.h
625
#define PPC_MADDHDU(t, a, b, c) stringify_in_c(.long PPC_RAW_MADDHDU(t, a, b, c))
arch/powerpc/include/asm/ppc-opcode.h
626
#define PPC_MADDLD(t, a, b, c) stringify_in_c(.long PPC_RAW_MADDLD(t, a, b, c))
arch/powerpc/include/asm/ppc-opcode.h
639
#define PPC_TLBILX(t, a, b) stringify_in_c(.long PPC_RAW_TLBILX(t, a, b))
arch/powerpc/include/asm/ppc-opcode.h
657
#define PPC_ERATILX(t, a, b) stringify_in_c(.long PPC_RAW_ERATILX(t, a, b))
arch/powerpc/include/asm/ppc-opcode.h
659
#define PPC_ERATSX(t, a, w) stringify_in_c(.long PPC_RAW_ERATSX(t, a, w))
arch/powerpc/include/asm/ppc-opcode.h
660
#define PPC_ERATSX_DOT(t, a, w) stringify_in_c(.long PPC_RAW_ERATSX_DOT(t, a, w))
arch/powerpc/include/asm/ppc-opcode.h
661
#define PPC_SLBFEE_DOT(t, b) stringify_in_c(.long PPC_RAW_SLBFEE_DOT(t, b))
arch/powerpc/include/asm/ppc-opcode.h
662
#define __PPC_SLBFEE_DOT(t, b) stringify_in_c(.long __PPC_RAW_SLBFEE_DOT(t, b))
arch/powerpc/include/asm/ppc-opcode.h
665
#define LBZCIX(t, a, b) stringify_in_c(.long PPC_RAW_LBZCIX(t, a, b))
arch/powerpc/include/asm/ppc-opcode.h
673
#define MFVRD(a, t) stringify_in_c(.long PPC_RAW_MFVRD(a, t))
arch/powerpc/include/asm/ppc-opcode.h
674
#define MTVRD(t, a) stringify_in_c(.long PPC_RAW_MTVRD(t, a))
arch/powerpc/include/asm/ppc-opcode.h
675
#define VPMSUMW(t, a, b) stringify_in_c(.long PPC_RAW_VPMSUMW(t, a, b))
arch/powerpc/include/asm/ppc-opcode.h
676
#define VPMSUMD(t, a, b) stringify_in_c(.long PPC_RAW_VPMSUMD(t, a, b))
arch/powerpc/include/asm/ppc-opcode.h
677
#define XXLOR(t, a, b) stringify_in_c(.long PPC_RAW_XXLOR(t, a, b))
arch/powerpc/include/asm/ppc-opcode.h
678
#define XXSWAPD(t, a) stringify_in_c(.long PPC_RAW_XXSWAPD(t, a))
arch/powerpc/include/asm/ppc-opcode.h
679
#define XVCPSGNDP(t, a, b) stringify_in_c(.long (PPC_RAW_XVCPSGNDP(t, a, b)))
arch/powerpc/include/asm/sfp-machine.h
37
#define __ll_lowpart(t) ((UWtype) (t) & (__ll_B - 1))
arch/powerpc/include/asm/sfp-machine.h
38
#define __ll_highpart(t) ((UWtype) (t) >> (W_TYPE_SIZE / 2))
arch/powerpc/include/asm/sstep.h
102
#define GETTYPE(t) ((t) & INSTR_TYPE_MASK)
arch/powerpc/include/asm/sstep.h
103
#define GETLENGTH(t) (((t) & PREFIXED) ? 8 : 4)
arch/powerpc/include/asm/sstep.h
105
#define MKOP(t, f, s) ((t) | (f) | SIZE(s))
arch/powerpc/include/asm/switch_to.h
109
static inline void __giveup_spe(struct task_struct *t) { }
arch/powerpc/include/asm/switch_to.h
112
static inline void clear_task_ebb(struct task_struct *t)
arch/powerpc/include/asm/switch_to.h
116
t->thread.ebbrr = 0;
arch/powerpc/include/asm/switch_to.h
117
t->thread.ebbhr = 0;
arch/powerpc/include/asm/switch_to.h
118
t->thread.bescr = 0;
arch/powerpc/include/asm/switch_to.h
119
t->thread.mmcr2 = 0;
arch/powerpc/include/asm/switch_to.h
120
t->thread.mmcr0 = 0;
arch/powerpc/include/asm/switch_to.h
121
t->thread.siar = 0;
arch/powerpc/include/asm/switch_to.h
122
t->thread.sdar = 0;
arch/powerpc/include/asm/switch_to.h
123
t->thread.sier = 0;
arch/powerpc/include/asm/switch_to.h
124
t->thread.used_ebb = 0;
arch/powerpc/include/asm/switch_to.h
131
extern int set_thread_tidr(struct task_struct *t);
arch/powerpc/include/asm/switch_to.h
49
static inline void save_fpu(struct task_struct *t) { }
arch/powerpc/include/asm/switch_to.h
50
static inline void flush_fp_to_thread(struct task_struct *t) { }
arch/powerpc/include/asm/switch_to.h
67
static inline void save_altivec(struct task_struct *t) { }
arch/powerpc/include/asm/switch_to.h
68
static inline void __giveup_altivec(struct task_struct *t) { }
arch/powerpc/kernel/cpu_setup_power.c
140
void __setup_cpu_power7(unsigned long offset, struct cpu_spec *t)
arch/powerpc/kernel/cpu_setup_power.c
142
if (!init_hvmode_206(t))
arch/powerpc/kernel/cpu_setup_power.c
165
void __setup_cpu_power8(unsigned long offset, struct cpu_spec *t)
arch/powerpc/kernel/cpu_setup_power.c
17
static bool init_hvmode_206(struct cpu_spec *t)
arch/powerpc/kernel/cpu_setup_power.c
171
if (!init_hvmode_206(t))
arch/powerpc/kernel/cpu_setup_power.c
204
void __setup_cpu_power9(unsigned long offset, struct cpu_spec *t)
arch/powerpc/kernel/cpu_setup_power.c
209
if (!init_hvmode_206(t))
arch/powerpc/kernel/cpu_setup_power.c
245
void __setup_cpu_power10(unsigned long offset, struct cpu_spec *t)
arch/powerpc/kernel/cpu_setup_power.c
25
t->cpu_features &= ~(CPU_FTR_HVMODE | CPU_FTR_P9_TM_HV_ASSIST);
arch/powerpc/kernel/cpu_setup_power.c
252
if (!init_hvmode_206(t))
arch/powerpc/kernel/cputable.c
103
return t;
arch/powerpc/kernel/cputable.c
133
struct cpu_spec *t = &the_cpu_spec;
arch/powerpc/kernel/cputable.c
137
t = PTRRELOC(t);
arch/powerpc/kernel/cputable.c
141
t->cpu_name = s->cpu_name;
arch/powerpc/kernel/cputable.c
35
struct cpu_spec *t = &the_cpu_spec;
arch/powerpc/kernel/cputable.c
37
t = PTRRELOC(t);
arch/powerpc/kernel/cputable.c
42
memcpy(t, s, sizeof(*t));
arch/powerpc/kernel/cputable.c
50
struct cpu_spec *t = &the_cpu_spec;
arch/powerpc/kernel/cputable.c
53
t = PTRRELOC(t);
arch/powerpc/kernel/cputable.c
54
old = *t;
arch/powerpc/kernel/cputable.c
60
memcpy(t, s, sizeof(*t));
arch/powerpc/kernel/cputable.c
68
t->num_pmcs = old.num_pmcs;
arch/powerpc/kernel/cputable.c
69
t->pmc_type = old.pmc_type;
arch/powerpc/kernel/cputable.c
75
t->cpu_features |= old.cpu_features & CPU_FTR_PMAO_BUG;
arch/powerpc/kernel/cputable.c
80
t->mmu_features |= MMU_FTR_KUAP;
arch/powerpc/kernel/cputable.c
89
*PTRRELOC(&powerpc_base_platform) = t->platform;
arch/powerpc/kernel/cputable.c
98
if (t->cpu_setup) {
arch/powerpc/kernel/cputable.c
99
t->cpu_setup(offset, t);
arch/powerpc/kernel/eeh_event.c
80
struct task_struct *t;
arch/powerpc/kernel/eeh_event.c
83
t = kthread_run(eeh_event_handler, NULL, "eehd");
arch/powerpc/kernel/eeh_event.c
84
if (IS_ERR(t)) {
arch/powerpc/kernel/eeh_event.c
85
ret = PTR_ERR(t);
arch/powerpc/kernel/hw_breakpoint.c
583
struct thread_struct *t = &tsk->thread;
arch/powerpc/kernel/hw_breakpoint.c
586
unregister_hw_breakpoint(t->ptrace_bps[i]);
arch/powerpc/kernel/hw_breakpoint.c
587
t->ptrace_bps[i] = NULL;
arch/powerpc/kernel/process.c
1154
static inline void save_sprs(struct thread_struct *t)
arch/powerpc/kernel/process.c
1158
t->vrsave = mfspr(SPRN_VRSAVE);
arch/powerpc/kernel/process.c
1162
t->spefscr = mfspr(SPRN_SPEFSCR);
arch/powerpc/kernel/process.c
1166
t->dscr = mfspr(SPRN_DSCR);
arch/powerpc/kernel/process.c
1169
t->bescr = mfspr(SPRN_BESCR);
arch/powerpc/kernel/process.c
1170
t->ebbhr = mfspr(SPRN_EBBHR);
arch/powerpc/kernel/process.c
1171
t->ebbrr = mfspr(SPRN_EBBRR);
arch/powerpc/kernel/process.c
1173
t->fscr = mfspr(SPRN_FSCR);
arch/powerpc/kernel/process.c
1181
t->tar = mfspr(SPRN_TAR);
arch/powerpc/kernel/process.c
1185
t->hashkeyr = mfspr(SPRN_HASHKEYR);
arch/powerpc/kernel/process.c
1188
t->dexcr = mfspr(SPRN_DEXCR);
arch/powerpc/kernel/process.c
1660
int set_thread_tidr(struct task_struct *t)
arch/powerpc/kernel/process.c
1665
if (t != current)
arch/powerpc/kernel/process.c
1668
if (t->thread.tidr)
arch/powerpc/kernel/process.c
1671
t->thread.tidr = (u16)task_pid_nr(t);
arch/powerpc/kernel/process.c
1672
mtspr(SPRN_TIDR, t->thread.tidr);
arch/powerpc/kernel/traps.c
886
unsigned int ra, rb, t, i, sel, instr, rc;
arch/powerpc/kernel/traps.c
923
t = (instr >> 21) & 0x1f;
arch/powerpc/kernel/traps.c
925
vdst = (u8 *)&current->thread.vr_state.vr[t];
arch/powerpc/kernel/traps.c
927
vdst = (u8 *)&current->thread.fp_state.fpr[t][0];
arch/powerpc/kernel/uprobes.c
97
bool arch_uprobe_xol_was_trapped(struct task_struct *t)
arch/powerpc/kernel/uprobes.c
99
if (t->thread.trap_nr != UPROBE_TRAP_NR)
arch/powerpc/kvm/book3s_emulate.c
409
ulong b, t;
arch/powerpc/kvm/book3s_emulate.c
413
if (!vcpu->arch.mmu.slbfee(vcpu, b, &t))
arch/powerpc/kvm/book3s_emulate.c
415
kvmppc_set_gpr(vcpu, rt, t);
arch/powerpc/kvm/book3s_emulate.c
426
ulong t, rb_val;
arch/powerpc/kvm/book3s_emulate.c
429
t = vcpu->arch.mmu.slbmfee(vcpu, rb_val);
arch/powerpc/kvm/book3s_emulate.c
430
kvmppc_set_gpr(vcpu, rt, t);
arch/powerpc/kvm/book3s_emulate.c
437
ulong t, rb_val;
arch/powerpc/kvm/book3s_emulate.c
440
t = vcpu->arch.mmu.slbmfev(vcpu, rb_val);
arch/powerpc/kvm/book3s_emulate.c
441
kvmppc_set_gpr(vcpu, rt, t);
arch/powerpc/kvm/book3s_paired_singles.c
592
void (*func)(u64 *t,
arch/powerpc/kvm/book3s_pr.c
799
struct thread_struct *t = &current->thread;
arch/powerpc/kvm/book3s_pr.c
822
if (t->regs->msr & MSR_FP)
arch/powerpc/kvm/book3s_pr.c
824
t->fp_save_area = NULL;
arch/powerpc/kvm/book3s_pr.c
831
t->vr_save_area = NULL;
arch/powerpc/kvm/book3s_pr.c
862
struct thread_struct *t = &current->thread;
arch/powerpc/kvm/book3s_pr.c
904
t->fp_save_area = &vcpu->arch.fp;
arch/powerpc/kvm/book3s_pr.c
914
t->vr_save_area = &vcpu->arch.vr;
arch/powerpc/kvm/book3s_pr.c
919
t->regs->msr |= msr;
arch/powerpc/kvm/booke.c
629
static void kvmppc_watchdog_func(struct timer_list *t)
arch/powerpc/kvm/booke.c
631
struct kvm_vcpu *vcpu = timer_container_of(vcpu, t, arch.wdt_timer);
arch/powerpc/lib/test_emulate_step.c
52
#define TEST_PADDI(t, a, i, pr) \
arch/powerpc/lib/test_emulate_step.c
54
PPC_RAW_ADDI(t, a, i))
arch/powerpc/mm/book3s32/mmu_context.c
69
int init_new_context(struct task_struct *t, struct mm_struct *mm)
arch/powerpc/mm/book3s64/radix_tlb.c
358
struct tlbiel_pid *t = info;
arch/powerpc/mm/book3s64/radix_tlb.c
360
if (t->ric == RIC_FLUSH_TLB)
arch/powerpc/mm/book3s64/radix_tlb.c
361
_tlbiel_pid(t->pid, RIC_FLUSH_TLB);
arch/powerpc/mm/book3s64/radix_tlb.c
362
else if (t->ric == RIC_FLUSH_PWC)
arch/powerpc/mm/book3s64/radix_tlb.c
363
_tlbiel_pid(t->pid, RIC_FLUSH_PWC);
arch/powerpc/mm/book3s64/radix_tlb.c
365
_tlbiel_pid(t->pid, RIC_FLUSH_ALL);
arch/powerpc/mm/book3s64/radix_tlb.c
372
struct tlbiel_pid t = { .pid = pid, .ric = ric };
arch/powerpc/mm/book3s64/radix_tlb.c
374
on_each_cpu_mask(cpus, do_tlbiel_pid, &t, 1);
arch/powerpc/mm/book3s64/radix_tlb.c
496
struct tlbiel_va *t = info;
arch/powerpc/mm/book3s64/radix_tlb.c
498
if (t->ric == RIC_FLUSH_TLB)
arch/powerpc/mm/book3s64/radix_tlb.c
499
_tlbiel_va(t->va, t->pid, t->psize, RIC_FLUSH_TLB);
arch/powerpc/mm/book3s64/radix_tlb.c
500
else if (t->ric == RIC_FLUSH_PWC)
arch/powerpc/mm/book3s64/radix_tlb.c
501
_tlbiel_va(t->va, t->pid, t->psize, RIC_FLUSH_PWC);
arch/powerpc/mm/book3s64/radix_tlb.c
503
_tlbiel_va(t->va, t->pid, t->psize, RIC_FLUSH_ALL);
arch/powerpc/mm/book3s64/radix_tlb.c
511
struct tlbiel_va t = { .va = va, .pid = pid, .psize = psize, .ric = ric };
arch/powerpc/mm/book3s64/radix_tlb.c
512
on_each_cpu_mask(cpus, do_tlbiel_va, &t, 1);
arch/powerpc/mm/book3s64/radix_tlb.c
528
struct tlbiel_va_range *t = info;
arch/powerpc/mm/book3s64/radix_tlb.c
530
_tlbiel_va_range(t->start, t->end, t->pid, t->page_size,
arch/powerpc/mm/book3s64/radix_tlb.c
531
t->psize, t->also_pwc);
arch/powerpc/mm/book3s64/radix_tlb.c
562
struct tlbiel_va_range t = { .start = start, .end = end,
arch/powerpc/mm/book3s64/radix_tlb.c
566
on_each_cpu_mask(cpus, do_tlbiel_va_range, &t, 1);
arch/powerpc/mm/nohash/mmu_context.c
307
int init_new_context(struct task_struct *t, struct mm_struct *mm)
arch/powerpc/perf/req-gen/_begin.h
8
#define CAT2_STR_(t, s) __stringify(t/s)
arch/powerpc/perf/req-gen/_begin.h
9
#define CAT2_STR(t, s) CAT2_STR_(t, s)
arch/powerpc/platforms/chrp/setup.c
101
unsigned int t;
arch/powerpc/platforms/chrp/setup.c
117
t = in_le32(gg2_pci_config_base+
arch/powerpc/platforms/chrp/setup.c
120
if (!(t & 1))
arch/powerpc/platforms/chrp/setup.c
122
switch ((t>>8) & 0x1f) {
arch/powerpc/platforms/chrp/setup.c
146
gg2_memtypes[sdramen ? 1 : ((t>>1) & 3)]);
arch/powerpc/platforms/chrp/setup.c
149
t = in_le32(gg2_pci_config_base+GG2_PCI_CC_CTRL);
arch/powerpc/platforms/chrp/setup.c
151
gg2_cachesizes[(t>>7) & 3],
arch/powerpc/platforms/chrp/setup.c
152
gg2_cachetypes[(t>>2) & 3],
arch/powerpc/platforms/chrp/setup.c
153
gg2_cachemodes[t & 3]);
arch/powerpc/platforms/powermac/low_i2c.c
360
static void kw_i2c_timeout(struct timer_list *t)
arch/powerpc/platforms/powermac/low_i2c.c
362
struct pmac_i2c_host_kw *host = timer_container_of(host, t,
arch/powerpc/platforms/powermac/pfunc_core.c
22
#define LOG_BLOB(t,b,c)
arch/powerpc/platforms/powermac/udbg_adb.c
77
int k, t, on;
arch/powerpc/platforms/powermac/udbg_adb.c
82
t = 0;
arch/powerpc/platforms/powermac/udbg_adb.c
86
if (--t < 0) {
arch/powerpc/platforms/powermac/udbg_adb.c
90
t = 200000;
arch/powerpc/platforms/ps3/repository.c
493
enum ps3_interrupt_type t;
arch/powerpc/platforms/ps3/repository.c
497
repo->dev_index, res_index, &t, &id);
arch/powerpc/platforms/ps3/repository.c
505
if (t == intr_type) {
arch/powerpc/platforms/ps3/repository.c
531
enum ps3_reg_type t;
arch/powerpc/platforms/ps3/repository.c
536
repo->dev_index, res_index, &t, &a, &l);
arch/powerpc/platforms/ps3/repository.c
544
if (t == reg_type) {
arch/powerpc/xmon/xmon.c
2327
#define SWAP(a, b, t) ((t) = (a), (a) = (b), (b) = (t))
arch/powerpc/xmon/xmon.c
2332
int t;
arch/powerpc/xmon/xmon.c
2336
SWAP(val[0], val[1], t);
arch/powerpc/xmon/xmon.c
2339
SWAP(val[0], val[3], t);
arch/powerpc/xmon/xmon.c
2340
SWAP(val[1], val[2], t);
arch/powerpc/xmon/xmon.c
2343
SWAP(val[0], val[7], t);
arch/powerpc/xmon/xmon.c
2344
SWAP(val[1], val[6], t);
arch/powerpc/xmon/xmon.c
2345
SWAP(val[2], val[5], t);
arch/powerpc/xmon/xmon.c
2346
SWAP(val[3], val[4], t);
arch/riscv/include/asm/futex.h
95
: [r] "+r" (ret), [v] "=&r" (val), [u] "+m" (*uaddr), [t] "=&r" (tmp)
arch/riscv/kernel/probes/uprobes.c
83
bool arch_uprobe_xol_was_trapped(struct task_struct *t)
arch/riscv/kernel/probes/uprobes.c
85
if (t->thread.bad_cause != UPROBE_TRAP_NR)
arch/riscv/kernel/usercfi.c
380
int arch_get_shadow_stack_status(struct task_struct *t, unsigned long __user *status)
arch/riscv/kernel/usercfi.c
388
bcfi_status |= (is_shstk_enabled(t) ? PR_SHADOW_STACK_ENABLE : 0);
arch/riscv/kernel/usercfi.c
393
int arch_set_shadow_stack_status(struct task_struct *t, unsigned long status)
arch/riscv/kernel/usercfi.c
406
if (is_shstk_locked(t))
arch/riscv/kernel/usercfi.c
411
if (enable_shstk && !is_shstk_enabled(t)) {
arch/riscv/kernel/usercfi.c
415
if (is_shstk_allocated(t))
arch/riscv/kernel/usercfi.c
422
set_shstk_base(t, addr, size);
arch/riscv/kernel/usercfi.c
423
set_active_shstk(t, addr + size);
arch/riscv/kernel/usercfi.c
441
shstk_release(t);
arch/riscv/kernel/usercfi.c
443
set_shstk_status(t, enable_shstk);
arch/riscv/kernel/usercfi.c
460
int arch_get_indir_br_lp_status(struct task_struct *t, unsigned long __user *status)
arch/riscv/kernel/usercfi.c
468
fcfi_status |= (is_indir_lp_enabled(t) ? PR_INDIR_BR_LP_ENABLE : 0);
arch/riscv/kernel/usercfi.c
473
int arch_set_indir_br_lp_status(struct task_struct *t, unsigned long status)
arch/riscv/kernel/usercfi.c
481
if (is_indir_lp_locked(t))
arch/riscv/kernel/usercfi.c
489
set_indir_lp_status(t, enable_indir_lp);
arch/riscv/kernel/vdso/hwprobe.c
76
struct riscv_hwprobe t = {
arch/riscv/kernel/vdso/hwprobe.c
81
if (!riscv_hwprobe_pair_cmp(&t, p))
arch/riscv/kvm/vcpu_timer.c
105
struct kvm_vcpu_timer *t = &vcpu->arch.timer;
arch/riscv/kvm/vcpu_timer.c
107
return t->timer_next_event(vcpu, ncycles);
arch/riscv/kvm/vcpu_timer.c
113
struct kvm_vcpu_timer *t = container_of(h, struct kvm_vcpu_timer, hrt);
arch/riscv/kvm/vcpu_timer.c
114
struct kvm_vcpu *vcpu = container_of(t, struct kvm_vcpu, arch.timer);
arch/riscv/kvm/vcpu_timer.c
117
if (kvm_riscv_current_cycles(gt) < t->next_cycles) {
arch/riscv/kvm/vcpu_timer.c
118
delta_ns = kvm_riscv_delta_cycles2ns(t->next_cycles, gt, t);
arch/riscv/kvm/vcpu_timer.c
119
hrtimer_forward_now(&t->hrt, ktime_set(0, delta_ns));
arch/riscv/kvm/vcpu_timer.c
123
t->next_set = false;
arch/riscv/kvm/vcpu_timer.c
131
struct kvm_vcpu_timer *t = &vcpu->arch.timer;
arch/riscv/kvm/vcpu_timer.c
134
if (!kvm_riscv_delta_cycles2ns(t->next_cycles, gt, t) ||
arch/riscv/kvm/vcpu_timer.c
143
struct kvm_vcpu_timer *t = &vcpu->arch.timer;
arch/riscv/kvm/vcpu_timer.c
147
if (!t->init_done)
arch/riscv/kvm/vcpu_timer.c
150
delta_ns = kvm_riscv_delta_cycles2ns(t->next_cycles, gt, t);
arch/riscv/kvm/vcpu_timer.c
151
hrtimer_start(&t->hrt, ktime_set(0, delta_ns), HRTIMER_MODE_REL);
arch/riscv/kvm/vcpu_timer.c
152
t->next_set = true;
arch/riscv/kvm/vcpu_timer.c
163
struct kvm_vcpu_timer *t = &vcpu->arch.timer;
arch/riscv/kvm/vcpu_timer.c
184
reg_val = t->next_cycles;
arch/riscv/kvm/vcpu_timer.c
187
reg_val = (t->next_set) ? KVM_RISCV_TIMER_STATE_ON :
arch/riscv/kvm/vcpu_timer.c
203
struct kvm_vcpu_timer *t = &vcpu->arch.timer;
arch/riscv/kvm/vcpu_timer.c
229
t->next_cycles = reg_val;
arch/riscv/kvm/vcpu_timer.c
235
ret = kvm_riscv_vcpu_timer_cancel(t);
arch/riscv/kvm/vcpu_timer.c
247
struct kvm_vcpu_timer *t = &vcpu->arch.timer;
arch/riscv/kvm/vcpu_timer.c
249
if (t->init_done)
arch/riscv/kvm/vcpu_timer.c
25
struct kvm_vcpu_timer *t)
arch/riscv/kvm/vcpu_timer.c
252
t->init_done = true;
arch/riscv/kvm/vcpu_timer.c
253
t->next_set = false;
arch/riscv/kvm/vcpu_timer.c
257
t->sstc_enabled = true;
arch/riscv/kvm/vcpu_timer.c
258
hrtimer_setup(&t->hrt, kvm_riscv_vcpu_vstimer_expired, CLOCK_MONOTONIC,
arch/riscv/kvm/vcpu_timer.c
260
t->timer_next_event = kvm_riscv_vcpu_update_vstimecmp;
arch/riscv/kvm/vcpu_timer.c
262
t->sstc_enabled = false;
arch/riscv/kvm/vcpu_timer.c
263
hrtimer_setup(&t->hrt, kvm_riscv_vcpu_hrtimer_expired, CLOCK_MONOTONIC,
arch/riscv/kvm/vcpu_timer.c
265
t->timer_next_event = kvm_riscv_vcpu_update_hrtimer;
arch/riscv/kvm/vcpu_timer.c
283
struct kvm_vcpu_timer *t = &vcpu->arch.timer;
arch/riscv/kvm/vcpu_timer.c
285
t->next_cycles = -1ULL;
arch/riscv/kvm/vcpu_timer.c
303
struct kvm_vcpu_timer *t = &vcpu->arch.timer;
arch/riscv/kvm/vcpu_timer.c
307
if (!t->sstc_enabled)
arch/riscv/kvm/vcpu_timer.c
312
ncsr_write(CSR_VSTIMECMPH, (u32)(t->next_cycles >> 32));
arch/riscv/kvm/vcpu_timer.c
313
ncsr_write(CSR_VSTIMECMP, (u32)(t->next_cycles));
arch/riscv/kvm/vcpu_timer.c
315
ncsr_write(CSR_VSTIMECMP, t->next_cycles);
arch/riscv/kvm/vcpu_timer.c
319
if (unlikely(!t->init_done))
arch/riscv/kvm/vcpu_timer.c
327
struct kvm_vcpu_timer *t = &vcpu->arch.timer;
arch/riscv/kvm/vcpu_timer.c
329
if (!t->sstc_enabled)
arch/riscv/kvm/vcpu_timer.c
333
t->next_cycles = ncsr_read(CSR_VSTIMECMP);
arch/riscv/kvm/vcpu_timer.c
334
t->next_cycles |= (u64)ncsr_read(CSR_VSTIMECMPH) << 32;
arch/riscv/kvm/vcpu_timer.c
336
t->next_cycles = ncsr_read(CSR_VSTIMECMP);
arch/riscv/kvm/vcpu_timer.c
342
struct kvm_vcpu_timer *t = &vcpu->arch.timer;
arch/riscv/kvm/vcpu_timer.c
344
if (!t->sstc_enabled)
arch/riscv/kvm/vcpu_timer.c
369
if (unlikely(!t->init_done))
arch/riscv/kvm/vcpu_timer.c
45
struct kvm_vcpu_timer *t = container_of(h, struct kvm_vcpu_timer, hrt);
arch/riscv/kvm/vcpu_timer.c
46
struct kvm_vcpu *vcpu = container_of(t, struct kvm_vcpu, arch.timer);
arch/riscv/kvm/vcpu_timer.c
49
if (kvm_riscv_current_cycles(gt) < t->next_cycles) {
arch/riscv/kvm/vcpu_timer.c
50
delta_ns = kvm_riscv_delta_cycles2ns(t->next_cycles, gt, t);
arch/riscv/kvm/vcpu_timer.c
51
hrtimer_forward_now(&t->hrt, ktime_set(0, delta_ns));
arch/riscv/kvm/vcpu_timer.c
55
t->next_set = false;
arch/riscv/kvm/vcpu_timer.c
61
static int kvm_riscv_vcpu_timer_cancel(struct kvm_vcpu_timer *t)
arch/riscv/kvm/vcpu_timer.c
63
if (!t->init_done || !t->next_set)
arch/riscv/kvm/vcpu_timer.c
66
hrtimer_cancel(&t->hrt);
arch/riscv/kvm/vcpu_timer.c
67
t->next_set = false;
arch/riscv/kvm/vcpu_timer.c
86
struct kvm_vcpu_timer *t = &vcpu->arch.timer;
arch/riscv/kvm/vcpu_timer.c
90
if (!t->init_done)
arch/riscv/kvm/vcpu_timer.c
95
delta_ns = kvm_riscv_delta_cycles2ns(ncycles, gt, t);
arch/riscv/kvm/vcpu_timer.c
96
t->next_cycles = ncycles;
arch/riscv/kvm/vcpu_timer.c
97
hrtimer_start(&t->hrt, ktime_set(0, delta_ns), HRTIMER_MODE_REL);
arch/riscv/kvm/vcpu_timer.c
98
t->next_set = true;
arch/riscv/mm/init.c
100
(((t) - (b)) >> LOG2_SZ_1K));
arch/riscv/mm/init.c
103
static inline void print_mlm(char *name, unsigned long b, unsigned long t)
arch/riscv/mm/init.c
105
pr_notice("%12s : 0x%08lx - 0x%08lx (%4ld MB)\n", name, b, t,
arch/riscv/mm/init.c
106
(((t) - (b)) >> LOG2_SZ_1M));
arch/riscv/mm/init.c
109
static inline void print_mlg(char *name, unsigned long b, unsigned long t)
arch/riscv/mm/init.c
111
pr_notice("%12s : 0x%08lx - 0x%08lx (%4ld GB)\n", name, b, t,
arch/riscv/mm/init.c
112
(((t) - (b)) >> LOG2_SZ_1G));
arch/riscv/mm/init.c
116
static inline void print_mlt(char *name, unsigned long b, unsigned long t)
arch/riscv/mm/init.c
118
pr_notice("%12s : 0x%08lx - 0x%08lx (%4ld TB)\n", name, b, t,
arch/riscv/mm/init.c
119
(((t) - (b)) >> LOG2_SZ_1T));
arch/riscv/mm/init.c
122
#define print_mlt(n, b, t) do {} while (0)
arch/riscv/mm/init.c
125
static inline void print_ml(char *name, unsigned long b, unsigned long t)
arch/riscv/mm/init.c
127
unsigned long diff = t - b;
arch/riscv/mm/init.c
130
print_mlt(name, b, t);
arch/riscv/mm/init.c
132
print_mlg(name, b, t);
arch/riscv/mm/init.c
134
print_mlm(name, b, t);
arch/riscv/mm/init.c
136
print_mlk(name, b, t);
arch/riscv/mm/init.c
97
static inline void print_mlk(char *name, unsigned long b, unsigned long t)
arch/riscv/mm/init.c
99
pr_notice("%12s : 0x%08lx - 0x%08lx (%4ld kB)\n", name, b, t,
arch/s390/boot/physmem_info.c
209
enum reserved_range_type t;
arch/s390/boot/physmem_info.c
218
for_each_physmem_reserved_range(t, range, &start, &end) {
arch/s390/boot/physmem_info.c
219
boot_emerg("%016lx %016lx %s\n", start, end, get_rr_type_name(t));
arch/s390/boot/physmem_info.c
259
int t;
arch/s390/boot/physmem_info.c
261
for (t = 0; t < RR_MAX; t++) {
arch/s390/boot/physmem_info.c
262
if (!get_physmem_reserved(t, &res_addr, &res_size))
arch/s390/boot/physmem_info.c
375
enum reserved_range_type t;
arch/s390/boot/physmem_info.c
379
for_each_physmem_reserved_range(t, range, &start, &end) {
arch/s390/boot/physmem_info.c
382
get_rr_type_name(t), start, end, (unsigned long)range,
arch/s390/boot/vmem.c
44
#define POPULATE_MODE_NAME(t) case POPULATE_ ## t: return #t
arch/s390/boot/vmem.c
45
static inline const char *get_populate_mode_name(enum populate_mode t)
arch/s390/boot/vmem.c
47
switch (t) {
arch/s390/crypto/aes_s390.c
834
u8 t[GHASH_DIGEST_SIZE];/* Tag */
arch/s390/crypto/aes_s390.c
908
if (crypto_memneq(tag, param.t, taglen))
arch/s390/crypto/aes_s390.c
911
scatterwalk_map_and_copy(param.t, req->dst, len, taglen, 1);
arch/s390/include/asm/cpu_mf.h
139
unsigned int t:1; /* 2 - Timestamp format */
arch/s390/include/asm/debug.h
183
unsigned int t = tag;
arch/s390/include/asm/debug.h
187
return debug_event_common(id, level, &t, sizeof(unsigned int));
arch/s390/include/asm/debug.h
205
unsigned long t = tag;
arch/s390/include/asm/debug.h
209
return debug_event_common(id, level, &t, sizeof(unsigned long));
arch/s390/include/asm/debug.h
308
unsigned int t = tag;
arch/s390/include/asm/debug.h
312
return debug_exception_common(id, level, &t, sizeof(unsigned int));
arch/s390/include/asm/debug.h
331
unsigned long t = tag;
arch/s390/include/asm/debug.h
335
return debug_exception_common(id, level, &t, sizeof(unsigned long));
arch/s390/include/asm/fcx.h
150
#define TSB_FORMAT(t) ((t)->flags & 7)
arch/s390/include/asm/physmem_info.h
121
#define RR_TYPE_NAME(t) case RR_ ## t: return #t
arch/s390/include/asm/physmem_info.h
122
static inline const char *get_rr_type_name(enum reserved_range_type t)
arch/s390/include/asm/physmem_info.h
124
switch (t) {
arch/s390/include/asm/physmem_info.h
138
#define for_each_physmem_reserved_type_range(t, range, p_start, p_end) \
arch/s390/include/asm/physmem_info.h
139
for (range = &physmem_info.reserved[t], *p_start = range->start, *p_end = range->end; \
arch/s390/include/asm/physmem_info.h
143
static inline struct reserved_range *__physmem_reserved_next(enum reserved_range_type *t,
arch/s390/include/asm/physmem_info.h
147
range = &physmem_info.reserved[*t];
arch/s390/include/asm/physmem_info.h
153
while (++*t < RR_MAX) {
arch/s390/include/asm/physmem_info.h
154
range = &physmem_info.reserved[*t];
arch/s390/include/asm/physmem_info.h
161
#define for_each_physmem_reserved_range(t, range, p_start, p_end) \
arch/s390/include/asm/physmem_info.h
162
for (t = 0, range = __physmem_reserved_next(&t, NULL), \
arch/s390/include/asm/physmem_info.h
164
range; range = __physmem_reserved_next(&t, range), \
arch/s390/include/asm/stacktrace.h
150
#define CALL_TYPECHECK_1(t, a, ...) \
arch/s390/include/asm/stacktrace.h
151
typecheck(t, a)
arch/s390/include/asm/stacktrace.h
152
#define CALL_TYPECHECK_2(t, a, ...) \
arch/s390/include/asm/stacktrace.h
154
typecheck(t, a)
arch/s390/include/asm/stacktrace.h
155
#define CALL_TYPECHECK_3(t, a, ...) \
arch/s390/include/asm/stacktrace.h
157
typecheck(t, a)
arch/s390/include/asm/stacktrace.h
158
#define CALL_TYPECHECK_4(t, a, ...) \
arch/s390/include/asm/stacktrace.h
160
typecheck(t, a)
arch/s390/include/asm/stacktrace.h
161
#define CALL_TYPECHECK_5(t, a, ...) \
arch/s390/include/asm/stacktrace.h
163
typecheck(t, a)
arch/s390/include/asm/stacktrace.h
166
#define CALL_PARM_1(t, a, ...) t
arch/s390/include/asm/stacktrace.h
167
#define CALL_PARM_2(t, a, ...) t, CALL_PARM_1(__VA_ARGS__)
arch/s390/include/asm/stacktrace.h
168
#define CALL_PARM_3(t, a, ...) t, CALL_PARM_2(__VA_ARGS__)
arch/s390/include/asm/stacktrace.h
169
#define CALL_PARM_4(t, a, ...) t, CALL_PARM_3(__VA_ARGS__)
arch/s390/include/asm/stacktrace.h
170
#define CALL_PARM_5(t, a, ...) t, CALL_PARM_4(__VA_ARGS__)
arch/s390/include/asm/stacktrace.h
171
#define CALL_PARM_6(t, a, ...) t, CALL_PARM_5(__VA_ARGS__)
arch/s390/include/asm/sysinfo.h
46
unsigned char t:1;
arch/s390/include/uapi/asm/guarded_storage.h
37
__u8 t : 1;
arch/s390/include/uapi/asm/runtime_instr.h
31
__u32 t : 1;
arch/s390/include/uapi/asm/vtoc.h
159
__u16 t; /* RTA of the first track of free extent */
arch/s390/kernel/sysinfo.c
106
seq_printf(m, "Capacity Transient: %d\n", info->t);
arch/s390/kvm/dat.h
37
#define _CRSTE_TOK(l, t, p) ((union crste) { \
arch/s390/kvm/dat.h
40
.tok.type = (t), \
arch/s390/kvm/dat.h
50
#define _PTE_TOK(t, p) ((union pte) { .tok.i = 1, .tok.type = (t), .tok.par = (p) })
arch/sh/boards/mach-dreamcast/rtc.c
36
time64_t t;
arch/sh/boards/mach-dreamcast/rtc.c
47
t = (u32)(val1 - TWENTY_YEARS);
arch/sh/boards/mach-dreamcast/rtc.c
49
rtc_time64_to_tm(t, tm);
arch/sh/boards/mach-se/7724/setup.c
615
int t = 10000;
arch/sh/boards/mach-se/7724/setup.c
617
while (t--) {
arch/sh/drivers/heartbeat.c
59
static void heartbeat_timer(struct timer_list *t)
arch/sh/drivers/heartbeat.c
61
struct heartbeat_data *hd = timer_container_of(hd, t, timer);
arch/sh/drivers/pci/common.c
100
struct pci_channel *hose = timer_container_of(hose, t, serr_timer);
arch/sh/drivers/pci/common.c
89
static void pcibios_enable_err(struct timer_list *t)
arch/sh/drivers/pci/common.c
91
struct pci_channel *hose = timer_container_of(hose, t, err_timer);
arch/sh/drivers/pci/common.c
98
static void pcibios_enable_serr(struct timer_list *t)
arch/sh/drivers/push-switch.c
26
static void switch_timer(struct timer_list *t)
arch/sh/drivers/push-switch.c
28
struct push_switch *psw = timer_container_of(psw, t, debounce);
arch/sh/kernel/hw_breakpoint.c
265
struct thread_struct *t = &tsk->thread;
arch/sh/kernel/hw_breakpoint.c
268
unregister_hw_breakpoint(t->ptrace_bps[i]);
arch/sh/kernel/hw_breakpoint.c
269
t->ptrace_bps[i] = NULL;
arch/sh/math-emu/math.c
62
{u32 t[2]; FP_PACK_DP(t,f); ((u32*)&r)[0]=t[1]; ((u32*)&r)[1]=t[0];}
arch/sh/math-emu/math.c
64
{u32 t[2]; t[0]=((u32*)&r)[1]; t[1]=((u32*)&r)[0]; FP_UNPACK_DP(f,t);}
arch/sh/math-emu/sfp-util.h
29
#define __ll_lowpart(t) ((UWtype) (t) & (__ll_B - 1))
arch/sh/math-emu/sfp-util.h
30
#define __ll_highpart(t) ((UWtype) (t) >> (W_TYPE_SIZE / 2))
arch/sparc/include/asm/bitext.h
24
int bit_map_string_get(struct bit_map *t, int len, int align);
arch/sparc/include/asm/bitext.h
25
void bit_map_clear(struct bit_map *t, int offset, int len);
arch/sparc/include/asm/bitext.h
26
void bit_map_init(struct bit_map *t, unsigned long *map, int size);
arch/sparc/include/asm/string.h
15
#define memcpy(t, f, n) __builtin_memcpy(t, f, n)
arch/sparc/include/uapi/asm/signal.h
19
#define SUBSIG_BADTRAP(t) (0x80 + (t))
arch/sparc/kernel/cpumap.c
296
static void increment_rover(struct cpuinfo_tree *t, int node_index,
arch/sparc/kernel/cpumap.c
299
struct cpuinfo_node *node = &t->nodes[node_index];
arch/sparc/kernel/cpumap.c
302
top_level = t->nodes[root_index].level;
arch/sparc/kernel/cpumap.c
314
node = &t->nodes[node->parent_index];
arch/sparc/kernel/cpumap.c
318
static int iterate_cpu(struct cpuinfo_tree *t, unsigned int root_index)
arch/sparc/kernel/cpumap.c
340
for (level = t->nodes[root_index].level; level < CPUINFO_LVL_MAX;
arch/sparc/kernel/cpumap.c
342
new_index = t->nodes[index].rover;
arch/sparc/kernel/cpumap.c
344
increment_rover(t, index, root_index, rover_inc_table);
arch/sparc/kernel/kgdb_32.c
49
struct thread_info *t = task_thread_info(p);
arch/sparc/kernel/kgdb_32.c
55
gdb_regs[GDB_G6] = (unsigned long) t;
arch/sparc/kernel/kgdb_32.c
59
gdb_regs[GDB_SP] = t->ksp;
arch/sparc/kernel/kgdb_32.c
62
win = (struct reg_window32 *) t->ksp;
arch/sparc/kernel/kgdb_32.c
73
gdb_regs[GDB_PSR] = t->kpsr;
arch/sparc/kernel/kgdb_32.c
74
gdb_regs[GDB_WIM] = t->kwim;
arch/sparc/kernel/kgdb_32.c
76
gdb_regs[GDB_PC] = t->kpc;
arch/sparc/kernel/kgdb_32.c
77
gdb_regs[GDB_NPC] = t->kpc + 4;
arch/sparc/kernel/kgdb_64.c
47
struct thread_info *t = task_thread_info(p);
arch/sparc/kernel/kgdb_64.c
56
gdb_regs[GDB_G6] = (unsigned long) t;
arch/sparc/kernel/kgdb_64.c
60
gdb_regs[GDB_SP] = t->ksp;
arch/sparc/kernel/kgdb_64.c
63
win = (struct reg_window *) (t->ksp + STACK_BIAS);
arch/sparc/kernel/kgdb_64.c
72
if (t->new_child)
arch/sparc/kernel/kgdb_64.c
80
cwp = __thread_flag_byte_ptr(t)[TI_FLAG_BYTE_CWP];
arch/sparc/kernel/ldc.c
224
unsigned long t;
arch/sparc/kernel/ldc.c
226
t = tx_advance(lp, lp->tx_tail);
arch/sparc/kernel/ldc.c
227
if (t == lp->tx_head)
arch/sparc/kernel/ldc.c
230
*new_tail = t;
arch/sparc/kernel/ldc.c
279
unsigned long h, t;
arch/sparc/kernel/ldc.c
282
t = tx_advance(lp, lp->tx_tail);
arch/sparc/kernel/ldc.c
283
if (t == h)
arch/sparc/kernel/ldc.c
286
*new_tail = t;
arch/sparc/kernel/mdesc.c
1018
u64 t = mdesc_arc_target(hp, a);
arch/sparc/kernel/mdesc.c
1022
name = mdesc_node_name(hp, t);
arch/sparc/kernel/mdesc.c
1026
id = mdesc_get_property(hp, t, "id", NULL);
arch/sparc/kernel/mdesc.c
1205
u64 j, t = mdesc_arc_target(hp, a);
arch/sparc/kernel/mdesc.c
1208
t_name = mdesc_node_name(hp, t);
arch/sparc/kernel/mdesc.c
1210
fill_in_one_cache(c, hp, t);
arch/sparc/kernel/mdesc.c
1214
mdesc_for_each_arc(j, hp, t, MDESC_ARC_TYPE_FWD) {
arch/sparc/kernel/mdesc.c
980
u64 t = mdesc_arc_target(hp, a);
arch/sparc/kernel/mdesc.c
984
name = mdesc_node_name(hp, t);
arch/sparc/kernel/mdesc.c
988
id = mdesc_get_property(hp, t, "id", NULL);
arch/sparc/kernel/process_64.c
400
struct thread_info *t = task_thread_info(tsk);
arch/sparc/kernel/process_64.c
402
if (t->utraps) {
arch/sparc/kernel/process_64.c
403
if (t->utraps[0] < 2)
arch/sparc/kernel/process_64.c
404
kfree (t->utraps);
arch/sparc/kernel/process_64.c
406
t->utraps[0]--;
arch/sparc/kernel/process_64.c
412
struct thread_info *t = current_thread_info();
arch/sparc/kernel/process_64.c
415
mm = t->task->mm;
arch/sparc/kernel/process_64.c
422
t->fpsaved[0] = 0;
arch/sparc/kernel/process_64.c
468
struct thread_info *t)
arch/sparc/kernel/process_64.c
473
t->rwbuf_stkptrs[i] = t->rwbuf_stkptrs[i+1];
arch/sparc/kernel/process_64.c
474
memcpy(&t->reg_window[i], &t->reg_window[i+1],
arch/sparc/kernel/process_64.c
481
struct thread_info *t = current_thread_info();
arch/sparc/kernel/process_64.c
488
struct reg_window *rwin = &t->reg_window[window];
arch/sparc/kernel/process_64.c
492
sp = t->rwbuf_stkptrs[window];
arch/sparc/kernel/process_64.c
500
shift_window_buffer(window, get_thread_wsaved() - 1, t);
arch/sparc/kernel/process_64.c
519
struct thread_info *t = current_thread_info();
arch/sparc/kernel/process_64.c
528
struct reg_window *rwin = &t->reg_window[window];
arch/sparc/kernel/process_64.c
532
orig_sp = sp = t->rwbuf_stkptrs[window];
arch/sparc/kernel/process_64.c
574
struct thread_info *t = task_thread_info(p);
arch/sparc/kernel/process_64.c
586
t->new_child = 1;
arch/sparc/kernel/process_64.c
587
t->ksp = ((unsigned long) child_trap_frame) - STACK_BIAS;
arch/sparc/kernel/process_64.c
588
t->kregs = (struct pt_regs *) (child_trap_frame +
arch/sparc/kernel/process_64.c
590
t->fpsaved[0] = 0;
arch/sparc/kernel/process_64.c
594
__thread_flag_byte_ptr(t)[TI_FLAG_BYTE_CWP] =
arch/sparc/kernel/process_64.c
596
t->kregs->u_regs[UREG_G1] = (unsigned long) args->fn;
arch/sparc/kernel/process_64.c
597
t->kregs->u_regs[UREG_G2] = (unsigned long) args->fn_arg;
arch/sparc/kernel/process_64.c
603
if (t->flags & _TIF_32BIT) {
arch/sparc/kernel/process_64.c
607
t->kregs->u_regs[UREG_FP] = sp;
arch/sparc/kernel/process_64.c
608
__thread_flag_byte_ptr(t)[TI_FLAG_BYTE_CWP] =
arch/sparc/kernel/process_64.c
616
t->kregs->u_regs[UREG_FP] = csp;
arch/sparc/kernel/process_64.c
618
if (t->utraps)
arch/sparc/kernel/process_64.c
619
t->utraps[0]++;
arch/sparc/kernel/process_64.c
626
t->kregs->u_regs[UREG_I0] = 0;
arch/sparc/kernel/process_64.c
629
t->kregs->u_regs[UREG_G1] = 0;
arch/sparc/kernel/process_64.c
634
t->kregs->u_regs[UREG_I0] = current->pid;
arch/sparc/kernel/process_64.c
635
t->kregs->u_regs[UREG_I1] = 1;
arch/sparc/kernel/process_64.c
642
t->kregs->u_regs[UREG_G7] = tls;
arch/sparc/kernel/prom_irqtrans.c
818
struct irq_trans *t = &pci_irq_trans_table[i];
arch/sparc/kernel/prom_irqtrans.c
820
if (!strcmp(model, t->name)) {
arch/sparc/kernel/prom_irqtrans.c
821
t->init(dp);
arch/sparc/kernel/ptrace_64.c
345
struct thread_info *t = task_thread_info(target);
arch/sparc/kernel/ptrace_64.c
351
fprs = t->fpsaved[0];
arch/sparc/kernel/ptrace_64.c
354
membuf_write(&to, t->fpregs, 16 * sizeof(u64));
arch/sparc/kernel/ptrace_64.c
359
membuf_write(&to, t->fpregs + 16, 16 * sizeof(u64));
arch/sparc/kernel/ptrace_64.c
363
membuf_store(&to, t->xfsr[0]);
arch/sparc/kernel/ptrace_64.c
364
membuf_store(&to, t->gsr[0]);
arch/sparc/kernel/ptrace_64.c
680
struct thread_info *t = task_thread_info(target);
arch/sparc/kernel/ptrace_64.c
686
enabled = t->fpsaved[0] & FPRS_FEF;
arch/sparc/kernel/ptrace_64.c
688
membuf_write(&to, t->fpregs, 32 * sizeof(u32));
arch/sparc/kernel/ptrace_64.c
691
membuf_store(&to, (u32)t->xfsr[0]);
arch/sparc/kernel/ptrace_64.c
833
struct thread_info *t = task_thread_info(target);
arch/sparc/kernel/ptrace_64.c
838
membuf_write(&to, t->fpregs, 32 * sizeof(u32));
arch/sparc/kernel/ptrace_64.c
839
if (t->fpsaved[0] & FPRS_FEF)
arch/sparc/kernel/ptrace_64.c
840
membuf_store(&to, (u32)t->xfsr[0]);
arch/sparc/kernel/sigutil_32.c
102
struct thread_info *t = current_thread_info();
arch/sparc/kernel/sigutil_32.c
114
err |= copy_from_user(&t->reg_window[i],
arch/sparc/kernel/sigutil_32.c
117
err |= __get_user(t->rwbuf_stkptrs[i],
arch/sparc/kernel/sigutil_32.c
123
t->w_saved = wsaved;
arch/sparc/kernel/sigutil_32.c
125
if (t->w_saved)
arch/sparc/kernel/sigutil_64.c
76
struct thread_info *t = current_thread_info();
arch/sparc/kernel/sigutil_64.c
88
err |= copy_from_user(&t->reg_window[i],
arch/sparc/kernel/sigutil_64.c
91
err |= __get_user(t->rwbuf_stkptrs[i],
arch/sparc/kernel/smp_64.c
1084
struct tlb_pending_info *t = info;
arch/sparc/kernel/smp_64.c
1086
__flush_tlb_pending(t->ctx, t->nr, t->vaddrs);
arch/sparc/kernel/smp_64.c
210
} t[NUM_ROUNDS];
arch/sparc/kernel/smp_64.c
235
t[i].rt = rt;
arch/sparc/kernel/smp_64.c
236
t[i].master = master_time_stamp;
arch/sparc/kernel/smp_64.c
237
t[i].diff = delta;
arch/sparc/kernel/smp_64.c
238
t[i].lat = adjust_latency/4;
arch/sparc/kernel/smp_64.c
247
t[i].rt, t[i].master, t[i].diff, t[i].lat);
arch/sparc/kernel/stacktrace.c
19
struct task_struct *t;
arch/sparc/kernel/stacktrace.c
32
t = tp->task;
arch/sparc/kernel/stacktrace.c
62
ret_stack = ftrace_graph_get_ret_stack(t,
arch/sparc/kernel/traps_64.c
2826
void notrace init_cur_cpu_trap(struct thread_info *t)
arch/sparc/kernel/traps_64.c
2831
p->thread = t;
arch/sparc/kernel/uprobes.c
306
bool arch_uprobe_xol_was_trapped(struct task_struct *t)
arch/sparc/kernel/viohs.c
805
static void vio_port_timer(struct timer_list *t)
arch/sparc/kernel/viohs.c
807
struct vio_driver_state *vio = timer_container_of(vio, t, timer);
arch/sparc/lib/bitext.c
103
void bit_map_clear(struct bit_map *t, int offset, int len)
arch/sparc/lib/bitext.c
107
if (t->used < len)
arch/sparc/lib/bitext.c
109
spin_lock(&t->lock);
arch/sparc/lib/bitext.c
111
if (test_bit(offset + i, t->map) == 0)
arch/sparc/lib/bitext.c
113
__clear_bit(offset + i, t->map);
arch/sparc/lib/bitext.c
115
if (offset < t->first_free)
arch/sparc/lib/bitext.c
116
t->first_free = offset;
arch/sparc/lib/bitext.c
117
t->used -= len;
arch/sparc/lib/bitext.c
118
spin_unlock(&t->lock);
arch/sparc/lib/bitext.c
121
void bit_map_init(struct bit_map *t, unsigned long *map, int size)
arch/sparc/lib/bitext.c
124
memset(t, 0, sizeof *t);
arch/sparc/lib/bitext.c
125
spin_lock_init(&t->lock);
arch/sparc/lib/bitext.c
126
t->map = map;
arch/sparc/lib/bitext.c
127
t->size = size;
arch/sparc/lib/bitext.c
28
int bit_map_string_get(struct bit_map *t, int len, int align)
arch/sparc/lib/bitext.c
35
if (t->num_colors) {
arch/sparc/lib/bitext.c
38
align = t->num_colors;
arch/sparc/lib/bitext.c
47
if (align < 0 || align >= t->size)
arch/sparc/lib/bitext.c
49
if (len <= 0 || len > t->size)
arch/sparc/lib/bitext.c
53
spin_lock(&t->lock);
arch/sparc/lib/bitext.c
54
if (len < t->last_size)
arch/sparc/lib/bitext.c
55
offset = t->first_free;
arch/sparc/lib/bitext.c
57
offset = t->last_off & ~align1;
arch/sparc/lib/bitext.c
60
off_new = find_next_zero_bit(t->map, t->size, offset);
arch/sparc/lib/bitext.c
64
if (offset >= t->size)
arch/sparc/lib/bitext.c
66
if (count + len > t->size) {
arch/sparc/lib/bitext.c
67
spin_unlock(&t->lock);
arch/sparc/lib/bitext.c
70
t->size, t->used, offset, len, align, count);
arch/sparc/lib/bitext.c
74
if (offset + len > t->size) {
arch/sparc/lib/bitext.c
75
count += t->size - offset;
arch/sparc/lib/bitext.c
81
while (test_bit(offset + i, t->map) == 0) {
arch/sparc/lib/bitext.c
84
bitmap_set(t->map, offset, len);
arch/sparc/lib/bitext.c
85
if (offset == t->first_free)
arch/sparc/lib/bitext.c
86
t->first_free = find_next_zero_bit
arch/sparc/lib/bitext.c
87
(t->map, t->size,
arch/sparc/lib/bitext.c
88
t->first_free + len);
arch/sparc/lib/bitext.c
89
if ((t->last_off = offset + len) >= t->size)
arch/sparc/lib/bitext.c
90
t->last_off = 0;
arch/sparc/lib/bitext.c
91
t->used += len;
arch/sparc/lib/bitext.c
92
t->last_size = len;
arch/sparc/lib/bitext.c
93
spin_unlock(&t->lock);
arch/sparc/lib/bitext.c
98
if ((offset += i + 1) >= t->size)
arch/um/drivers/vector_kern.c
1530
static void vector_timer_expire(struct timer_list *t)
arch/um/drivers/vector_kern.c
1532
struct vector_private *vp = timer_container_of(vp, t, tl);
arch/um/include/shared/os.h
117
flags.t = 1;
arch/um/include/shared/os.h
70
unsigned int t : 1; /* O_TRUNC */
arch/um/include/shared/os.h
77
.t = 0, .a = 0, .e = 0, .cl = 0 })
arch/um/kernel/irq.c
104
enum um_irq_type t)
arch/um/kernel/irq.c
106
struct irq_reg *reg = &entry->reg[t];
arch/um/kernel/irq.c
135
enum um_irq_type t;
arch/um/kernel/irq.c
137
for (t = 0; t < NUM_IRQ_TYPES; t++) {
arch/um/kernel/irq.c
138
struct irq_reg *reg = &entry->reg[t];
arch/um/kernel/irq.c
155
enum um_irq_type t)
arch/um/kernel/irq.c
165
static void sigio_reg_handler(int idx, struct irq_entry *entry, enum um_irq_type t,
arch/um/kernel/irq.c
169
struct irq_reg *reg = &entry->reg[t];
arch/um/kernel/irq.c
177
if (irq_do_timetravel_handler(entry, t))
arch/um/kernel/irq.c
228
enum um_irq_type t;
arch/um/kernel/irq.c
232
for (t = 0; t < NUM_IRQ_TYPES; t++)
arch/um/kernel/irq.c
233
sigio_reg_handler(i, irq_entry, t, regs,
arch/um/kernel/irq.c
570
enum um_irq_type t;
arch/um/kernel/irq.c
573
for (t = 0; t < NUM_IRQ_TYPES; t++) {
arch/um/kernel/irq.c
574
if (!entry->reg[t].events)
arch/um/kernel/irq.c
583
if (entry->reg[t].wakeup ||
arch/um/kernel/irq.c
584
entry->reg[t].irq == SIGIO_WRITE_IRQ
arch/um/kernel/irq.c
586
|| entry->reg[t].timetravel_handler
arch/um/kernel/irq.c
637
enum um_irq_type t;
arch/um/kernel/irq.c
639
for (t = 0; t < NUM_IRQ_TYPES; t++) {
arch/um/kernel/irq.c
640
if (!entry->reg[t].events)
arch/um/kernel/irq.c
643
if (entry->reg[t].irq != d->irq)
arch/um/kernel/irq.c
645
entry->reg[t].wakeup = on;
arch/um/kernel/reboot.c
27
struct task_struct *t;
arch/um/kernel/reboot.c
29
t = find_lock_task_mm(p);
arch/um/kernel/reboot.c
30
if (!t)
arch/um/kernel/reboot.c
32
pid = t->mm->context.id.pid;
arch/um/kernel/reboot.c
33
task_unlock(t);
arch/um/os-Linux/file.c
181
if (flags.t)
arch/um/os-Linux/time.c
47
timer_t *t = &event_high_res_timer[cpu];
arch/um/os-Linux/time.c
51
.sigev_value.sival_ptr = t,
arch/um/os-Linux/time.c
55
if (timer_create(CLOCK_MONOTONIC, &sev, t) == -1)
arch/x86/coco/sev/vc-handle.c
79
struct thread_struct *t = &current->thread;
arch/x86/coco/sev/vc-handle.c
80
struct io_bitmap *iobm = t->io_bitmap;
arch/x86/crypto/camellia_glue.c
775
u64 t = l; \
arch/x86/crypto/camellia_glue.c
777
r = (r << bits) | (t >> (64 - bits)); \
arch/x86/entry/vdso/common/vclock_gettime.c
26
__kernel_old_time_t __vdso_time(__kernel_old_time_t *t)
arch/x86/entry/vdso/common/vclock_gettime.c
28
return __cvdso_time(t);
arch/x86/entry/vdso/common/vclock_gettime.c
31
__kernel_old_time_t time(__kernel_old_time_t *t) __attribute__((weak, alias("__vdso_time")));
arch/x86/events/intel/core.c
6633
EVENT_ATTR_STR(cycles-t, cycles_t, "event=0x3c,in_tx=1");
arch/x86/events/intel/core.c
7130
EVENT_ATTR_STR_HYBRID(cycles-t, cycles_t_adl, "event=0x3c,in_tx=1", hybrid_big);
arch/x86/events/intel/pt.c
639
#define TOPA_ENTRY(t, i) \
arch/x86/events/intel/pt.c
641
? &topa_to_page(t)->table[(t)->last] \
arch/x86/events/intel/pt.c
642
: &topa_to_page(t)->table[(i)])
arch/x86/events/intel/pt.c
643
#define TOPA_ENTRY_SIZE(t, i) (sizes(TOPA_ENTRY((t), (i))->size))
arch/x86/events/intel/pt.c
644
#define TOPA_ENTRY_PAGES(t, i) (1 << TOPA_ENTRY((t), (i))->size)
arch/x86/events/intel/uncore_snb.c
1276
#define for_each_imc_pci_id(x, t) \
arch/x86/events/intel/uncore_snb.c
1277
for (x = (t); (x)->pci_id; x++)
arch/x86/events/intel/uncore_snbep.c
3824
struct intel_uncore_topology *t;
arch/x86/events/intel/uncore_snbep.c
3827
t = &type->topology[die][idx];
arch/x86/events/intel/uncore_snbep.c
3828
t->pmu_idx = idx;
arch/x86/events/intel/uncore_snbep.c
3829
t->iio->segment = segment;
arch/x86/events/intel/uncore_snbep.c
3830
t->iio->pci_bus_no = (cpu_bus_msr >> (idx * BUS_NUM_STRIDE)) & 0xff;
arch/x86/hyperv/ivm.c
553
struct hv_enc_pfn_region *ent, *t;
arch/x86/hyperv/ivm.c
563
list_for_each_entry_safe(ent, t, &hv_list_enc, list) {
arch/x86/include/asm/desc.h
119
#define load_TLS(t, cpu) native_load_tls(t, cpu)
arch/x86/include/asm/desc.h
292
static inline void native_load_tls(struct thread_struct *t, unsigned int cpu)
arch/x86/include/asm/desc.h
298
gdt[GDT_ENTRY_TLS_MIN + i] = t->tls_array[i];
arch/x86/include/asm/e820/api.h
20
extern u64 e820__range_update_table(struct e820_table *t, u64 start, u64 size, enum e820_type old_type, enum e820_type new_type);
arch/x86/include/asm/elf.h
153
static inline void elf_common_init(struct thread_struct *t,
arch/x86/include/asm/elf.h
161
t->fsbase = t->gsbase = 0;
arch/x86/include/asm/elf.h
162
t->fsindex = t->gsindex = 0;
arch/x86/include/asm/elf.h
163
t->ds = t->es = ds;
arch/x86/include/asm/nmi.h
78
#define register_nmi_handler(t, fn, fg, n, init...) \
arch/x86/include/asm/nmi.h
86
__register_nmi_handler((t), &fn##_na); \
arch/x86/include/asm/paravirt.h
243
static inline void load_TLS(struct thread_struct *t, unsigned cpu)
arch/x86/include/asm/paravirt.h
245
PVOP_VCALL2(pv_ops, cpu.load_tls, t, cpu);
arch/x86/include/asm/paravirt_types.h
50
void (*load_tls)(struct thread_struct *t, unsigned int cpu);
arch/x86/include/asm/perf_event_p4.h
88
u32 t = p4_config_unpack_escr((v)); \
arch/x86/include/asm/perf_event_p4.h
89
t = t & P4_ESCR_EVENTMASK_MASK; \
arch/x86/include/asm/perf_event_p4.h
90
t = t >> P4_ESCR_EVENTMASK_SHIFT; \
arch/x86/include/asm/perf_event_p4.h
91
t; \
arch/x86/include/asm/perf_event_p4.h
96
u32 t = p4_config_unpack_escr((v)); \
arch/x86/include/asm/perf_event_p4.h
97
t = t & P4_ESCR_EVENT_MASK; \
arch/x86/include/asm/perf_event_p4.h
98
t = t >> P4_ESCR_EVENT_SHIFT; \
arch/x86/include/asm/perf_event_p4.h
99
t; \
arch/x86/include/asm/ptrace.h
467
# define do_set_thread_area_64(p, s, t) do_arch_prctl_64(p, s, t)
arch/x86/include/asm/ptrace.h
469
# define do_set_thread_area_64(p, s, t) (0)
arch/x86/include/asm/string_32.h
150
#define memcpy(t, f, n) __builtin_memcpy(t, f, n)
arch/x86/include/asm/syscall_wrapper.h
76
#define __SC_COMPAT_CAST(t, a) \
arch/x86/include/asm/syscall_wrapper.h
77
(__typeof(__builtin_choose_expr(__TYPE_IS_L(t), 0, 0U))) \
arch/x86/include/asm/uv/uv_mmrs.h
2287
unsigned long t:1; /* RO */
arch/x86/include/asm/uv/uv_mmrs.h
2301
unsigned long t:1; /* RO */
arch/x86/include/asm/uv/uv_mmrs.h
2315
unsigned long t:1; /* RO */
arch/x86/include/asm/uv/uv_mmrs.h
2329
unsigned long t:1; /* RO */
arch/x86/include/asm/uv/uv_mmrs.h
2343
unsigned long t:1; /* RO */
arch/x86/include/asm/uv/uv_mmrs.h
2390
unsigned long t:1; /* RO */
arch/x86/include/asm/uv/uv_mmrs.h
2404
unsigned long t:1; /* RO */
arch/x86/include/asm/uv/uv_mmrs.h
2418
unsigned long t:1; /* RO */
arch/x86/include/asm/uv/uv_mmrs.h
2432
unsigned long t:1; /* RO */
arch/x86/include/asm/uv/uv_mmrs.h
2446
unsigned long t:1; /* RO */
arch/x86/include/asm/uv/uv_mmrs.h
2493
unsigned long t:1; /* RO */
arch/x86/include/asm/uv/uv_mmrs.h
2507
unsigned long t:1; /* RO */
arch/x86/include/asm/uv/uv_mmrs.h
2521
unsigned long t:1; /* RO */
arch/x86/include/asm/uv/uv_mmrs.h
2535
unsigned long t:1; /* RO */
arch/x86/include/asm/uv/uv_mmrs.h
2549
unsigned long t:1; /* RO */
arch/x86/include/asm/uv/uv_mmrs.h
2596
unsigned long t:1; /* RO */
arch/x86/include/asm/uv/uv_mmrs.h
2610
unsigned long t:1; /* RO */
arch/x86/include/asm/uv/uv_mmrs.h
2624
unsigned long t:1; /* RO */
arch/x86/include/asm/uv/uv_mmrs.h
2638
unsigned long t:1; /* RO */
arch/x86/include/asm/uv/uv_mmrs.h
2652
unsigned long t:1; /* RO */
arch/x86/include/asm/uv/uv_mmrs.h
4475
unsigned long t:1; /* RO */
arch/x86/include/asm/uv/uv_mmrs.h
4489
unsigned long t:1; /* RO */
arch/x86/include/asm/uv/uv_mmrs.h
4503
unsigned long t:1; /* RO */
arch/x86/include/asm/uv/uv_mmrs.h
4517
unsigned long t:1; /* RO */
arch/x86/include/asm/uv/uv_mmrs.h
4531
unsigned long t:1; /* RO */
arch/x86/include/asm/vm86.h
53
#define free_vm86(t) do { \
arch/x86/include/asm/vm86.h
54
struct thread_struct *__t = (t); \
arch/x86/kernel/apm_32.c
851
static unsigned long t;
arch/x86/kernel/apm_32.c
856
if (++t < 5) {
arch/x86/kernel/apm_32.c
858
t = jiffies;
arch/x86/kernel/apm_32.c
934
unsigned int t;
arch/x86/kernel/apm_32.c
936
t = jiffies;
arch/x86/kernel/apm_32.c
940
if (t != jiffies) {
arch/x86/kernel/cpu/mce/amd.c
133
static const char *smca_get_name(enum smca_bank_types t)
arch/x86/kernel/cpu/mce/amd.c
135
if (t >= N_SMCA_BANK_TYPES)
arch/x86/kernel/cpu/mce/amd.c
138
return smca_names[t];
arch/x86/kernel/cpu/mce/core.c
1039
static noinstr int mce_timed_out(u64 *t, const char *msg)
arch/x86/kernel/cpu/mce/core.c
1057
if ((s64)*t < SPINUNIT) {
arch/x86/kernel/cpu/mce/core.c
1066
*t -= SPINUNIT;
arch/x86/kernel/cpu/mce/core.c
1766
static void __start_timer(struct timer_list *t, unsigned long interval)
arch/x86/kernel/cpu/mce/core.c
1773
if (!timer_pending(t) || time_before(when, t->expires))
arch/x86/kernel/cpu/mce/core.c
1774
mod_timer(t, round_jiffies(when));
arch/x86/kernel/cpu/mce/core.c
1791
static void mce_timer_fn(struct timer_list *t)
arch/x86/kernel/cpu/mce/core.c
1796
WARN_ON(cpu_t != t);
arch/x86/kernel/cpu/mce/core.c
1813
__start_timer(t, HZ);
arch/x86/kernel/cpu/mce/core.c
1816
__start_timer(t, iv);
arch/x86/kernel/cpu/mce/core.c
1827
struct timer_list *t = this_cpu_ptr(&mce_timer);
arch/x86/kernel/cpu/mce/core.c
1832
__start_timer(t, HZ);
arch/x86/kernel/cpu/mce/core.c
2099
static void mce_start_timer(struct timer_list *t)
arch/x86/kernel/cpu/mce/core.c
2105
__start_timer(t, iv);
arch/x86/kernel/cpu/mce/core.c
2111
struct timer_list *t = this_cpu_ptr(&mce_timer);
arch/x86/kernel/cpu/mce/core.c
2113
timer_setup(t, mce_timer_fn, TIMER_PINNED);
arch/x86/kernel/cpu/mce/core.c
2118
struct timer_list *t = this_cpu_ptr(&mce_timer);
arch/x86/kernel/cpu/mce/core.c
2120
timer_setup(t, mce_timer_fn, TIMER_PINNED);
arch/x86/kernel/cpu/mce/core.c
2121
mce_start_timer(t);
arch/x86/kernel/cpu/mce/core.c
2793
struct timer_list *t = this_cpu_ptr(&mce_timer);
arch/x86/kernel/cpu/mce/core.c
2798
mce_start_timer(t);
arch/x86/kernel/cpu/mce/core.c
2804
struct timer_list *t = this_cpu_ptr(&mce_timer);
arch/x86/kernel/cpu/mce/core.c
2807
timer_delete_sync(t);
arch/x86/kernel/cpu/mce/genpool.c
32
static bool is_duplicate_mce_record(struct mce_evt_llist *t, struct mce_evt_llist *l)
arch/x86/kernel/cpu/mce/genpool.c
37
err1 = &t->err;
arch/x86/kernel/cpu/mce/genpool.c
59
struct mce_evt_llist *node, *t;
arch/x86/kernel/cpu/mce/genpool.c
66
llist_for_each_entry_safe(node, t, head, llnode) {
arch/x86/kernel/cpu/mce/genpool.c
67
if (!is_duplicate_mce_record(node, t))
arch/x86/kernel/cpu/resctrl/ctrlmondata.c
23
u32 closid, enum resctrl_conf_type t, u32 cfg_val)
arch/x86/kernel/cpu/resctrl/ctrlmondata.c
27
u32 idx = resctrl_get_config_index(closid, t);
arch/x86/kernel/cpu/resctrl/ctrlmondata.c
50
enum resctrl_conf_type t;
arch/x86/kernel/cpu/resctrl/ctrlmondata.c
59
for (t = 0; t < CDP_NUM_TYPES; t++) {
arch/x86/kernel/cpu/resctrl/ctrlmondata.c
60
cfg = &hw_dom->d_resctrl.staged_config[t];
arch/x86/kernel/cpu/resctrl/ctrlmondata.c
64
idx = resctrl_get_config_index(closid, t);
arch/x86/kernel/e820.c
544
__init u64 e820__range_update_table(struct e820_table *t, u64 start, u64 size,
arch/x86/kernel/e820.c
547
return __e820__range_update(t, start, size, old_type, new_type);
arch/x86/kernel/hw_breakpoint.c
473
struct thread_struct *t = &tsk->thread;
arch/x86/kernel/hw_breakpoint.c
476
unregister_hw_breakpoint(t->ptrace_bps[i]);
arch/x86/kernel/hw_breakpoint.c
477
t->ptrace_bps[i] = NULL;
arch/x86/kernel/hw_breakpoint.c
480
t->virtual_dr6 = 0;
arch/x86/kernel/hw_breakpoint.c
481
t->ptrace_dr7 = 0;
arch/x86/kernel/idt.c
197
idt_setup_from_table(gate_desc *idt, const struct idt_data *t, int size, bool sys)
arch/x86/kernel/idt.c
201
for (; size > 0; t++, size--) {
arch/x86/kernel/idt.c
202
idt_init_desc(&desc, t);
arch/x86/kernel/idt.c
203
write_idt_entry(idt, t->vector, &desc);
arch/x86/kernel/idt.c
205
set_bit(t->vector, system_vectors);
arch/x86/kernel/ioport.c
119
t->io_bitmap = iobm;
arch/x86/kernel/ioport.c
181
struct thread_struct *t = &current->thread;
arch/x86/kernel/ioport.c
187
old = t->iopl_emul;
arch/x86/kernel/ioport.c
200
t->iopl_emul = level;
arch/x86/kernel/ioport.c
39
struct thread_struct *t = &tsk->thread;
arch/x86/kernel/ioport.c
41
if (t->iopl_emul == 3 || t->io_bitmap) {
arch/x86/kernel/ioport.c
73
struct thread_struct *t = &current->thread;
arch/x86/kernel/ioport.c
88
iobm = t->io_bitmap;
arch/x86/kernel/process.c
130
struct thread_struct *t = &tsk->thread;
arch/x86/kernel/process.c
135
free_vm86(t);
arch/x86/kernel/process.c
473
struct thread_struct *t = &current->thread;
arch/x86/kernel/process.c
481
if (IS_ENABLED(CONFIG_X86_IOPL_IOPERM) && t->iopl_emul == 3) {
arch/x86/kernel/process.c
484
struct io_bitmap *iobm = t->io_bitmap;
arch/x86/kernel/ptrace.c
633
struct thread_struct *t = &tsk->thread;
arch/x86/kernel/ptrace.c
634
struct perf_event *bp = t->ptrace_bps[nr];
arch/x86/kernel/ptrace.c
655
t->ptrace_bps[nr] = bp;
arch/x86/kernel/tls.c
106
if (t == &current->thread)
arch/x86/kernel/tls.c
107
load_TLS(t, cpu);
arch/x86/kernel/tls.c
24
struct thread_struct *t = &current->thread;
arch/x86/kernel/tls.c
246
struct thread_struct *t = &target->thread;
arch/x86/kernel/tls.c
248
while (n > 0 && desc_empty(&t->tls_array[n - 1]))
arch/x86/kernel/tls.c
28
if (desc_empty(&t->tls_array[idx]))
arch/x86/kernel/tls.c
88
struct thread_struct *t = &p->thread;
arch/x86/kernel/tls.c
89
struct desc_struct *desc = &t->tls_array[idx - GDT_ENTRY_TLS_MIN];
arch/x86/kernel/traps.c
804
struct thread_struct *t = &current->thread;
arch/x86/kernel/traps.c
808
if (!IS_ENABLED(CONFIG_X86_IOPL_IOPERM) || t->iopl_emul != 3)
arch/x86/kernel/traps.c
820
if (!t->iopl_warn && printk_ratelimit()) {
arch/x86/kernel/traps.c
825
t->iopl_warn = 1;
arch/x86/kernel/uprobes.c
1637
bool arch_uprobe_xol_was_trapped(struct task_struct *t)
arch/x86/kernel/uprobes.c
1639
if (t->thread.trap_nr != UPROBE_TRAP_NR)
arch/x86/kvm/fpu.h
10
#define sse128_lo(x) ({ __sse128_u t; t.vec = x; t.as_u64[0]; })
arch/x86/kvm/fpu.h
11
#define sse128_hi(x) ({ __sse128_u t; t.vec = x; t.as_u64[1]; })
arch/x86/kvm/fpu.h
12
#define sse128_l0(x) ({ __sse128_u t; t.vec = x; t.as_u32[0]; })
arch/x86/kvm/fpu.h
13
#define sse128_l1(x) ({ __sse128_u t; t.vec = x; t.as_u32[1]; })
arch/x86/kvm/fpu.h
14
#define sse128_l2(x) ({ __sse128_u t; t.vec = x; t.as_u32[2]; })
arch/x86/kvm/fpu.h
15
#define sse128_l3(x) ({ __sse128_u t; t.vec = x; t.as_u32[3]; })
arch/x86/kvm/fpu.h
16
#define sse128(lo, hi) ({ __sse128_u t; t.as_u64[0] = lo; t.as_u64[1] = hi; t.vec; })
arch/x86/kvm/i8254.c
118
s64 d, t;
arch/x86/kvm/i8254.c
121
t = kpit_elapsed(pit, c, channel);
arch/x86/kvm/i8254.c
122
d = mul_u64_u32_div(t, KVM_PIT_FREQ, NSEC_PER_SEC);
arch/x86/kvm/i8254.c
145
s64 d, t;
arch/x86/kvm/i8254.c
148
t = kpit_elapsed(pit, c, channel);
arch/x86/kvm/i8254.c
149
d = mul_u64_u32_div(t, KVM_PIT_FREQ, NSEC_PER_SEC);
arch/x86/kvm/mtrr.c
56
static bool valid_mtrr_type(unsigned t)
arch/x86/kvm/mtrr.c
58
return t < 8 && (1 << t) & 0x73; /* 0, 1, 4, 5, 6 */
arch/x86/kvm/tss.h
31
u16 t;
arch/x86/kvm/x86.c
2967
static int do_kvmclock_base(s64 *t, u64 *tsc_timestamp)
arch/x86/kvm/x86.c
2981
*t = ns;
arch/x86/kvm/x86.c
2990
static int do_monotonic(s64 *t, u64 *tsc_timestamp)
arch/x86/kvm/x86.c
3004
*t = ns;
arch/x86/kvm/x86.c
8307
#define emulator_try_cmpxchg_user(t, ptr, old, new) \
arch/x86/kvm/x86.c
8308
(__try_cmpxchg_user((t __user *)(ptr), (t *)(old), *(t *)(new), efault ## t))
arch/x86/kvm/xen.c
1571
static void cancel_evtchn_poll(struct timer_list *t)
arch/x86/kvm/xen.c
1573
struct kvm_vcpu *vcpu = timer_container_of(vcpu, t,
arch/x86/lib/insn.c
23
#define leXX_to_cpu(t, r) \
arch/x86/lib/insn.c
25
__typeof__(t) v; \
arch/x86/lib/insn.c
26
switch (sizeof(t)) { \
arch/x86/lib/insn.c
37
#define validate_next(t, insn, n) \
arch/x86/lib/insn.c
38
((insn)->next_byte + sizeof(t) + n <= (insn)->end_kaddr)
arch/x86/lib/insn.c
40
#define __get_next(t, insn) \
arch/x86/lib/insn.c
41
({ t r = get_unaligned((t *)(insn)->next_byte); (insn)->next_byte += sizeof(t); leXX_to_cpu(t, r); })
arch/x86/lib/insn.c
43
#define __peek_nbyte_next(t, insn, n) \
arch/x86/lib/insn.c
44
({ t r = get_unaligned((t *)(insn)->next_byte + n); leXX_to_cpu(t, r); })
arch/x86/lib/insn.c
46
#define get_next(t, insn) \
arch/x86/lib/insn.c
47
({ if (unlikely(!validate_next(t, insn, 0))) goto err_out; __get_next(t, insn); })
arch/x86/lib/insn.c
49
#define peek_nbyte_next(t, insn, n) \
arch/x86/lib/insn.c
50
({ if (unlikely(!validate_next(t, insn, n))) goto err_out; __peek_nbyte_next(t, insn, n); })
arch/x86/lib/insn.c
52
#define peek_next(t, insn) peek_nbyte_next(t, insn, 0)
arch/x86/math-emu/fpu_aux.c
133
FPU_REG t;
arch/x86/math-emu/fpu_aux.c
164
reg_copy(st0_ptr, &t);
arch/x86/math-emu/fpu_aux.c
166
reg_copy(&t, sti_ptr);
arch/x86/math-emu/reg_ld_str.c
755
FPU_REG t;
arch/x86/math-emu/reg_ld_str.c
771
reg_copy(st0_ptr, &t);
arch/x86/math-emu/reg_ld_str.c
772
precision_loss = FPU_round_to_int(&t, st0_tag);
arch/x86/math-emu/reg_ld_str.c
773
((long *)&tll)[0] = t.sigl;
arch/x86/math-emu/reg_ld_str.c
774
((long *)&tll)[1] = t.sigh;
arch/x86/math-emu/reg_ld_str.c
776
((t.sigh & 0x80000000) &&
arch/x86/math-emu/reg_ld_str.c
777
!((t.sigh == 0x80000000) && (t.sigl == 0) && signnegative(&t)))) {
arch/x86/math-emu/reg_ld_str.c
789
if (signnegative(&t))
arch/x86/math-emu/reg_ld_str.c
805
FPU_REG t;
arch/x86/math-emu/reg_ld_str.c
820
reg_copy(st0_ptr, &t);
arch/x86/math-emu/reg_ld_str.c
821
precision_loss = FPU_round_to_int(&t, st0_tag);
arch/x86/math-emu/reg_ld_str.c
822
if (t.sigh ||
arch/x86/math-emu/reg_ld_str.c
823
((t.sigl & 0x80000000) &&
arch/x86/math-emu/reg_ld_str.c
824
!((t.sigl == 0x80000000) && signnegative(&t)))) {
arch/x86/math-emu/reg_ld_str.c
830
t.sigl = 0x80000000;
arch/x86/math-emu/reg_ld_str.c
836
if (signnegative(&t))
arch/x86/math-emu/reg_ld_str.c
837
t.sigl = -(long)t.sigl;
arch/x86/math-emu/reg_ld_str.c
842
FPU_put_user(t.sigl, (unsigned long __user *)d);
arch/x86/math-emu/reg_ld_str.c
851
FPU_REG t;
arch/x86/math-emu/reg_ld_str.c
866
reg_copy(st0_ptr, &t);
arch/x86/math-emu/reg_ld_str.c
867
precision_loss = FPU_round_to_int(&t, st0_tag);
arch/x86/math-emu/reg_ld_str.c
868
if (t.sigh ||
arch/x86/math-emu/reg_ld_str.c
869
((t.sigl & 0xffff8000) &&
arch/x86/math-emu/reg_ld_str.c
870
!((t.sigl == 0x8000) && signnegative(&t)))) {
arch/x86/math-emu/reg_ld_str.c
876
t.sigl = 0x8000;
arch/x86/math-emu/reg_ld_str.c
882
if (signnegative(&t))
arch/x86/math-emu/reg_ld_str.c
883
t.sigl = -t.sigl;
arch/x86/math-emu/reg_ld_str.c
888
FPU_put_user((short)t.sigl, d);
arch/x86/math-emu/reg_ld_str.c
897
FPU_REG t;
arch/x86/math-emu/reg_ld_str.c
915
reg_copy(st0_ptr, &t);
arch/x86/math-emu/reg_ld_str.c
916
precision_loss = FPU_round_to_int(&t, st0_tag);
arch/x86/math-emu/reg_ld_str.c
917
ll = significand(&t);
arch/x86/math-emu/reg_ld_str.c
920
if ((t.sigh > 0x0de0b6b3) ||
arch/x86/math-emu/reg_ld_str.c
921
((t.sigh == 0x0de0b6b3) && (t.sigl > 0xa763ffff))) {
arch/x86/net/bpf_jit_comp.c
4027
enum bpf_text_poke_type t;
arch/x86/net/bpf_jit_comp.c
4040
t = old_addr ? BPF_MOD_JUMP : BPF_MOD_NOP;
arch/x86/net/bpf_jit_comp.c
4042
t, BPF_MOD_JUMP,
arch/x86/net/bpf_jit_comp.c
4053
t = old_bypass_addr ? BPF_MOD_JUMP : BPF_MOD_NOP;
arch/x86/net/bpf_jit_comp.c
4055
t, BPF_MOD_JUMP, old_bypass_addr,
arch/x86/net/bpf_jit_comp.c
4064
t = old_addr ? BPF_MOD_JUMP : BPF_MOD_NOP;
arch/x86/net/bpf_jit_comp.c
4066
t, BPF_MOD_NOP, old_addr, NULL);
arch/x86/platform/uv/uv_time.c
207
u64 *t = &head->cpu[bcpu].expires;
arch/x86/platform/uv/uv_time.c
214
*t = expires;
arch/x86/platform/uv/uv_time.c
221
*t = ULLONG_MAX;
arch/x86/platform/uv/uv_time.c
243
u64 *t = &head->cpu[bcpu].expires;
arch/x86/platform/uv/uv_time.c
249
if ((head->next_cpu == bcpu && uv_read_rtc(NULL) >= *t) || force)
arch/x86/platform/uv/uv_time.c
253
*t = ULLONG_MAX;
arch/x86/um/asm/processor.h
39
#define task_pt_regs(t) (&(t)->thread.regs)
arch/x86/um/tls_32.c
200
struct thread_struct *t = &task->thread;
arch/x86/um/tls_32.c
205
t->arch.tls_array[idx - GDT_ENTRY_TLS_MIN].tls = *info;
arch/x86/um/tls_32.c
206
t->arch.tls_array[idx - GDT_ENTRY_TLS_MIN].present = 1;
arch/x86/um/tls_32.c
207
t->arch.tls_array[idx - GDT_ENTRY_TLS_MIN].flushed = flushed;
arch/x86/um/tls_32.c
234
struct thread_struct *t = &task->thread;
arch/x86/um/tls_32.c
239
if (!t->arch.tls_array[idx - GDT_ENTRY_TLS_MIN].present)
arch/x86/um/tls_32.c
242
*info = t->arch.tls_array[idx - GDT_ENTRY_TLS_MIN].tls;
arch/x86/um/tls_32.c
250
!t->arch.tls_array[idx - GDT_ENTRY_TLS_MIN].flushed)) {
arch/x86/um/tls_32.c
62
struct thread_struct *t = &task->thread;
arch/x86/um/tls_32.c
66
if (!t->arch.tls_array[idx].present)
arch/x86/um/tls_64.c
15
t->thread.regs.regs.gp[FS_BASE / sizeof(unsigned long)] = tls;
arch/x86/um/tls_64.c
9
int arch_set_tls(struct task_struct *t, unsigned long tls)
arch/x86/um/vdso/um_vdso.c
44
__kernel_old_time_t __vdso_time(__kernel_old_time_t *t)
arch/x86/um/vdso/um_vdso.c
50
: "0" (__NR_time), "D" (t) : "cc", "r11", "cx", "memory");
arch/x86/um/vdso/um_vdso.c
54
__kernel_old_time_t time(__kernel_old_time_t *t) __attribute__((weak, alias("__vdso_time")));
arch/x86/xen/enlighten_pv.c
616
static void load_TLS_descriptor(struct thread_struct *t,
arch/x86/xen/enlighten_pv.c
624
if (desc_equal(shadow, &t->tls_array[i]))
arch/x86/xen/enlighten_pv.c
627
*shadow = t->tls_array[i];
arch/x86/xen/enlighten_pv.c
633
MULTI_update_descriptor(mc.mc, maddr.maddr, t->tls_array[i]);
arch/x86/xen/enlighten_pv.c
636
static void xen_load_tls(struct thread_struct *t, unsigned int cpu)
arch/x86/xen/enlighten_pv.c
648
load_TLS_descriptor(t, cpu, 0);
arch/x86/xen/enlighten_pv.c
649
load_TLS_descriptor(t, cpu, 1);
arch/x86/xen/enlighten_pv.c
650
load_TLS_descriptor(t, cpu, 2);
arch/x86/xen/time.c
397
struct vcpu_register_time_memory_area t;
arch/x86/xen/time.c
405
t.addr.v = NULL;
arch/x86/xen/time.c
407
ret = HYPERVISOR_vcpu_op(VCPUOP_register_vcpu_time_memory_area, 0, &t);
arch/x86/xen/time.c
417
struct vcpu_register_time_memory_area t;
arch/x86/xen/time.c
423
t.addr.v = &xen_clock->pvti;
arch/x86/xen/time.c
425
ret = HYPERVISOR_vcpu_op(VCPUOP_register_vcpu_time_memory_area, 0, &t);
arch/x86/xen/time.c
448
struct vcpu_register_time_memory_area t;
arch/x86/xen/time.c
456
t.addr.v = &ti->pvti;
arch/x86/xen/time.c
458
ret = HYPERVISOR_vcpu_op(VCPUOP_register_vcpu_time_memory_area, 0, &t);
arch/x86/xen/time.c
471
t.addr.v = NULL;
arch/x86/xen/time.c
473
0, &t);
arch/xtensa/include/asm/asm-uaccess.h
50
movi \at, __XTENSA_UL_CONST(TASK_SIZE)
arch/xtensa/include/asm/asm-uaccess.h
51
bgeu \as, \at, \error
arch/xtensa/include/asm/asm-uaccess.h
52
sub \at, \at, \as
arch/xtensa/include/asm/asm-uaccess.h
53
bgeu \aa, \at, \error
arch/xtensa/include/asm/asm-uaccess.h
80
user_ok \aa, \as, \at, \error
arch/xtensa/include/asm/asmmacro.h
108
sub \at, \as, \ar
arch/xtensa/include/asm/asmmacro.h
110
addi \at, \at, (1 << \incr_log2) - 1
arch/xtensa/include/asm/asmmacro.h
111
srli \at, \at, \incr_log2
arch/xtensa/include/asm/asmmacro.h
113
loop \at, 99f
arch/xtensa/include/asm/asmmacro.h
53
movi \at, ((\size + \incr - 1) / (\incr))
arch/xtensa/include/asm/asmmacro.h
54
loop \at, 99f
arch/xtensa/include/asm/asmmacro.h
56
addi \at, \ar, \size
arch/xtensa/include/asm/asmmacro.h
70
addi \at, \as, (1 << \incr_log2) - 1
arch/xtensa/include/asm/asmmacro.h
72
extui \at, \at, \incr_log2, \mask_log2
arch/xtensa/include/asm/asmmacro.h
74
srli \at, \at, \incr_log2
arch/xtensa/include/asm/asmmacro.h
77
loop\cond \at, 99f
arch/xtensa/include/asm/asmmacro.h
80
extui \at, \as, \incr_log2, \mask_log2
arch/xtensa/include/asm/asmmacro.h
83
srli \at, \as, \incr_log2
arch/xtensa/include/asm/asmmacro.h
87
b\ncond \at, 99f
arch/xtensa/include/asm/asmmacro.h
91
slli \at, \at, \incr_log2
arch/xtensa/include/asm/asmmacro.h
92
add \at, \ar, \at
arch/xtensa/include/asm/asmmacro.h
94
add \at, \ar, \as
arch/xtensa/include/asm/cacheasm.h
106
__loop_cache_all \ar \at diwbi XCHAL_DCACHE_SIZE \
arch/xtensa/include/asm/cacheasm.h
116
__loop_cache_all \ar \at diwb XCHAL_DCACHE_SIZE \
arch/xtensa/include/asm/cacheasm.h
126
__loop_cache_all \ar \at dii XCHAL_DCACHE_SIZE \
arch/xtensa/include/asm/cacheasm.h
136
__loop_cache_all \ar \at iii XCHAL_ICACHE_SIZE \
arch/xtensa/include/asm/cacheasm.h
147
__loop_cache_range \ar \as \at dhwbi XCHAL_DCACHE_LINEWIDTH
arch/xtensa/include/asm/cacheasm.h
156
__loop_cache_range \ar \as \at dhwb XCHAL_DCACHE_LINEWIDTH
arch/xtensa/include/asm/cacheasm.h
165
__loop_cache_range \ar \as \at dhi XCHAL_DCACHE_LINEWIDTH
arch/xtensa/include/asm/cacheasm.h
174
__loop_cache_range \ar \as \at ihi XCHAL_ICACHE_LINEWIDTH
arch/xtensa/include/asm/cacheasm.h
45
__loopi \ar, \at, \size, (_reps << (\line_width))
arch/xtensa/include/asm/cacheasm.h
51
__endla \ar, \at, _reps << (\line_width)
arch/xtensa/include/asm/cacheasm.h
59
__loop_cache_unroll \ar, \at, \insn, \size, \line_width, \max_immed
arch/xtensa/include/asm/cacheasm.h
66
extui \at, \ar, 0, \line_width
arch/xtensa/include/asm/cacheasm.h
67
add \as, \as, \at
arch/xtensa/include/asm/cacheasm.h
69
__loops \ar, \as, \at, \line_width
arch/xtensa/include/asm/cacheasm.h
71
__endla \ar, \at, (1 << (\line_width))
arch/xtensa/include/asm/cacheasm.h
78
__loop_cache_unroll \ar, \at, \insn, PAGE_SIZE, \line_width, \max_immed
arch/xtensa/include/asm/cacheasm.h
86
__loop_cache_all \ar \at diu XCHAL_DCACHE_SIZE \
arch/xtensa/include/asm/cacheasm.h
96
__loop_cache_all \ar \at iiu XCHAL_ICACHE_SIZE \
arch/xtensa/include/asm/coprocessor.h
104
#define __REG0(cc,abi,t,name,s,a) __REG0_ ## cc (abi,name)
arch/xtensa/include/asm/coprocessor.h
105
#define __REG1(cc,abi,t,name,s,a) __REG1_ ## cc (name)
arch/xtensa/kernel/hw_breakpoint.c
237
struct thread_struct *t = &tsk->thread;
arch/xtensa/kernel/hw_breakpoint.c
240
if (t->ptrace_bp[i]) {
arch/xtensa/kernel/hw_breakpoint.c
241
unregister_hw_breakpoint(t->ptrace_bp[i]);
arch/xtensa/kernel/hw_breakpoint.c
242
t->ptrace_bp[i] = NULL;
arch/xtensa/kernel/hw_breakpoint.c
246
if (t->ptrace_wp[i]) {
arch/xtensa/kernel/hw_breakpoint.c
247
unregister_hw_breakpoint(t->ptrace_wp[i]);
arch/xtensa/kernel/hw_breakpoint.c
248
t->ptrace_wp[i] = NULL;
arch/xtensa/kernel/setup.c
142
tagtable_t *t;
arch/xtensa/kernel/setup.c
156
for (t = &__tagtable_begin; t < &__tagtable_end; t++) {
arch/xtensa/kernel/setup.c
157
if (tag->id == t->tag) {
arch/xtensa/kernel/setup.c
158
t->parse(tag);
arch/xtensa/kernel/setup.c
162
if (t == &__tagtable_end)
arch/xtensa/mm/cache.c
283
unsigned long t = TLBTEMP_BASE_1 + (vaddr & DCACHE_ALIAS_MASK);
arch/xtensa/mm/cache.c
285
__flush_invalidate_dcache_page_alias(t, phys);
arch/xtensa/mm/cache.c
299
unsigned long t = TLBTEMP_BASE_1 + (vaddr & DCACHE_ALIAS_MASK);
arch/xtensa/mm/cache.c
304
__invalidate_icache_page_alias(t, phys);
arch/xtensa/mm/cache.c
326
unsigned long t = TLBTEMP_BASE_1 + (vaddr & DCACHE_ALIAS_MASK);
arch/xtensa/mm/cache.c
328
__flush_invalidate_dcache_page_alias(t, phys);
arch/xtensa/platforms/iss/network.c
340
static void iss_net_timer(struct timer_list *t)
arch/xtensa/platforms/iss/network.c
342
struct iss_net_private *lp = timer_container_of(lp, t, timer);
block/blk-core.c
382
static void blk_rq_timed_out_timer(struct timer_list *t)
block/blk-core.c
384
struct request_queue *q = timer_container_of(q, t, timeout);
block/blk-iolatency.c
651
static void blkiolatency_timer_fn(struct timer_list *t)
block/blk-iolatency.c
653
struct blk_iolatency *blkiolat = timer_container_of(blkiolat, t,
block/blk-settings.c
670
static bool blk_stack_atomic_writes_tail(struct queue_limits *t,
block/blk-settings.c
674
if (t->atomic_write_hw_boundary != b->atomic_write_hw_boundary)
block/blk-settings.c
678
if (t->atomic_write_hw_unit_min > b->atomic_write_hw_unit_max)
block/blk-settings.c
682
if (t->atomic_write_hw_unit_max < b->atomic_write_hw_unit_min)
block/blk-settings.c
685
t->atomic_write_hw_max = min(t->atomic_write_hw_max,
block/blk-settings.c
687
t->atomic_write_hw_unit_min = max(t->atomic_write_hw_unit_min,
block/blk-settings.c
689
t->atomic_write_hw_unit_max = min(t->atomic_write_hw_unit_max,
block/blk-settings.c
694
static void blk_stack_atomic_writes_chunk_sectors(struct queue_limits *t)
block/blk-settings.c
698
if (!t->chunk_sectors)
block/blk-settings.c
706
if (check_shl_overflow(t->chunk_sectors, SECTOR_SHIFT, &chunk_bytes))
block/blk-settings.c
707
chunk_bytes = t->chunk_sectors;
block/blk-settings.c
719
t->atomic_write_hw_unit_max = min(t->atomic_write_hw_unit_max,
block/blk-settings.c
722
t->atomic_write_hw_unit_min = min(t->atomic_write_hw_unit_min,
block/blk-settings.c
723
t->atomic_write_hw_unit_max);
block/blk-settings.c
724
t->atomic_write_hw_max = min(t->atomic_write_hw_max, chunk_bytes);
block/blk-settings.c
728
static bool blk_stack_atomic_writes_head(struct queue_limits *t,
block/blk-settings.c
731
if (!blk_valid_atomic_writes_boundary(t->chunk_sectors,
block/blk-settings.c
735
t->atomic_write_hw_unit_max = b->atomic_write_hw_unit_max;
block/blk-settings.c
736
t->atomic_write_hw_unit_min = b->atomic_write_hw_unit_min;
block/blk-settings.c
737
t->atomic_write_hw_max = b->atomic_write_hw_max;
block/blk-settings.c
738
t->atomic_write_hw_boundary = b->atomic_write_hw_boundary;
block/blk-settings.c
742
static void blk_stack_atomic_writes_limits(struct queue_limits *t,
block/blk-settings.c
755
if (t->atomic_write_hw_max == UINT_MAX) {
block/blk-settings.c
756
if (!blk_stack_atomic_writes_head(t, b))
block/blk-settings.c
759
if (!blk_stack_atomic_writes_tail(t, b))
block/blk-settings.c
762
blk_stack_atomic_writes_chunk_sectors(t);
block/blk-settings.c
766
t->atomic_write_hw_max = 0;
block/blk-settings.c
767
t->atomic_write_hw_unit_max = 0;
block/blk-settings.c
768
t->atomic_write_hw_unit_min = 0;
block/blk-settings.c
769
t->atomic_write_hw_boundary = 0;
block/blk-settings.c
793
int blk_stack_limits(struct queue_limits *t, struct queue_limits *b,
block/blk-settings.c
799
t->features |= (b->features & BLK_FEAT_INHERIT_MASK);
block/blk-settings.c
808
t->features &= ~BLK_FEAT_NOWAIT;
block/blk-settings.c
810
t->features &= ~BLK_FEAT_POLL;
block/blk-settings.c
812
t->flags |= (b->flags & BLK_FLAG_MISALIGNED);
block/blk-settings.c
814
t->max_sectors = min_not_zero(t->max_sectors, b->max_sectors);
block/blk-settings.c
815
t->max_user_sectors = min_not_zero(t->max_user_sectors,
block/blk-settings.c
817
t->max_hw_sectors = min_not_zero(t->max_hw_sectors, b->max_hw_sectors);
block/blk-settings.c
818
t->max_dev_sectors = min_not_zero(t->max_dev_sectors, b->max_dev_sectors);
block/blk-settings.c
819
t->max_write_zeroes_sectors = min(t->max_write_zeroes_sectors,
block/blk-settings.c
821
t->max_user_wzeroes_unmap_sectors =
block/blk-settings.c
822
min(t->max_user_wzeroes_unmap_sectors,
block/blk-settings.c
824
t->max_hw_wzeroes_unmap_sectors =
block/blk-settings.c
825
min(t->max_hw_wzeroes_unmap_sectors,
block/blk-settings.c
828
t->max_hw_zone_append_sectors = min(t->max_hw_zone_append_sectors,
block/blk-settings.c
831
t->seg_boundary_mask = min_not_zero(t->seg_boundary_mask,
block/blk-settings.c
833
t->virt_boundary_mask = min_not_zero(t->virt_boundary_mask,
block/blk-settings.c
836
t->max_segments = min_not_zero(t->max_segments, b->max_segments);
block/blk-settings.c
837
t->max_discard_segments = min_not_zero(t->max_discard_segments,
block/blk-settings.c
839
t->max_integrity_segments = min_not_zero(t->max_integrity_segments,
block/blk-settings.c
842
t->max_segment_size = min_not_zero(t->max_segment_size,
block/blk-settings.c
850
if (t->alignment_offset != alignment) {
block/blk-settings.c
852
top = max(t->physical_block_size, t->io_min)
block/blk-settings.c
853
+ t->alignment_offset;
block/blk-settings.c
858
t->flags |= BLK_FLAG_MISALIGNED;
block/blk-settings.c
863
t->logical_block_size = max(t->logical_block_size,
block/blk-settings.c
866
t->physical_block_size = max(t->physical_block_size,
block/blk-settings.c
869
t->io_min = max(t->io_min, b->io_min);
block/blk-settings.c
870
t->io_opt = lcm_not_zero(t->io_opt, b->io_opt);
block/blk-settings.c
871
t->dma_alignment = max(t->dma_alignment, b->dma_alignment);
block/blk-settings.c
875
t->chunk_sectors = gcd(t->chunk_sectors, b->chunk_sectors);
block/blk-settings.c
878
if (t->physical_block_size & (t->logical_block_size - 1)) {
block/blk-settings.c
879
t->physical_block_size = t->logical_block_size;
block/blk-settings.c
880
t->flags |= BLK_FLAG_MISALIGNED;
block/blk-settings.c
885
if (t->io_min & (t->physical_block_size - 1)) {
block/blk-settings.c
886
t->io_min = t->physical_block_size;
block/blk-settings.c
887
t->flags |= BLK_FLAG_MISALIGNED;
block/blk-settings.c
892
if (t->io_opt & (t->physical_block_size - 1)) {
block/blk-settings.c
893
t->io_opt = 0;
block/blk-settings.c
894
t->flags |= BLK_FLAG_MISALIGNED;
block/blk-settings.c
899
if (t->chunk_sectors % (t->physical_block_size >> SECTOR_SHIFT)) {
block/blk-settings.c
900
t->chunk_sectors = 0;
block/blk-settings.c
901
t->flags |= BLK_FLAG_MISALIGNED;
block/blk-settings.c
906
t->alignment_offset = lcm_not_zero(t->alignment_offset, alignment)
block/blk-settings.c
907
% max(t->physical_block_size, t->io_min);
block/blk-settings.c
910
if (t->alignment_offset & (t->logical_block_size - 1)) {
block/blk-settings.c
911
t->flags |= BLK_FLAG_MISALIGNED;
block/blk-settings.c
915
t->max_sectors = blk_round_down_sectors(t->max_sectors, t->logical_block_size);
block/blk-settings.c
916
t->max_hw_sectors = blk_round_down_sectors(t->max_hw_sectors, t->logical_block_size);
block/blk-settings.c
917
t->max_dev_sectors = blk_round_down_sectors(t->max_dev_sectors, t->logical_block_size);
block/blk-settings.c
923
t->max_discard_sectors = min_not_zero(t->max_discard_sectors,
block/blk-settings.c
925
t->max_hw_discard_sectors = min_not_zero(t->max_hw_discard_sectors,
block/blk-settings.c
927
t->discard_granularity = max(t->discard_granularity,
block/blk-settings.c
929
t->discard_alignment = lcm_not_zero(t->discard_alignment, alignment) %
block/blk-settings.c
930
t->discard_granularity;
block/blk-settings.c
932
t->max_secure_erase_sectors = min_not_zero(t->max_secure_erase_sectors,
block/blk-settings.c
934
t->zone_write_granularity = max(t->zone_write_granularity,
block/blk-settings.c
936
if (!(t->features & BLK_FEAT_ZONED)) {
block/blk-settings.c
937
t->zone_write_granularity = 0;
block/blk-settings.c
938
t->max_zone_append_sectors = 0;
block/blk-settings.c
940
blk_stack_atomic_writes_limits(t, b, start);
block/blk-settings.c
961
void queue_limits_stack_bdev(struct queue_limits *t, struct block_device *bdev,
block/blk-settings.c
964
if (blk_stack_limits(t, bdev_limits(bdev),
block/blk-settings.c
985
bool queue_limits_stack_integrity(struct queue_limits *t,
block/blk-settings.c
988
struct blk_integrity *ti = &t->integrity;
block/blk-stat.c
77
static void blk_stat_timer_fn(struct timer_list *t)
block/blk-stat.c
79
struct blk_stat_callback *cb = timer_container_of(cb, t, timer);
block/blk-throttle.c
1119
static void throtl_pending_timer_fn(struct timer_list *t)
block/blk-throttle.c
1121
struct throtl_service_queue *sq = timer_container_of(sq, t,
block/blk-throttle.c
45
static void throtl_pending_timer_fn(struct timer_list *t);
block/kyber-iosched.c
270
static void kyber_timer_fn(struct timer_list *t)
block/kyber-iosched.c
272
struct kyber_queue_data *kqd = timer_container_of(kqd, t, timer);
crypto/aegis.h
65
const u32 *t = aes_enc_tab;
crypto/aegis.h
68
d0 = t[s[ 0]] ^ rol32(t[s[ 5]], 8) ^ rol32(t[s[10]], 16) ^ rol32(t[s[15]], 24);
crypto/aegis.h
69
d1 = t[s[ 4]] ^ rol32(t[s[ 9]], 8) ^ rol32(t[s[14]], 16) ^ rol32(t[s[ 3]], 24);
crypto/aegis.h
70
d2 = t[s[ 8]] ^ rol32(t[s[13]], 8) ^ rol32(t[s[ 2]], 16) ^ rol32(t[s[ 7]], 24);
crypto/aegis.h
71
d3 = t[s[12]] ^ rol32(t[s[ 1]], 8) ^ rol32(t[s[ 6]], 16) ^ rol32(t[s[11]], 24);
crypto/cast5_generic.c
305
u32 l, r, t;
crypto/cast5_generic.c
327
t = l; l = r; r = t ^ F1(r, Km[0], Kr[0]);
crypto/cast5_generic.c
328
t = l; l = r; r = t ^ F2(r, Km[1], Kr[1]);
crypto/cast5_generic.c
329
t = l; l = r; r = t ^ F3(r, Km[2], Kr[2]);
crypto/cast5_generic.c
330
t = l; l = r; r = t ^ F1(r, Km[3], Kr[3]);
crypto/cast5_generic.c
331
t = l; l = r; r = t ^ F2(r, Km[4], Kr[4]);
crypto/cast5_generic.c
332
t = l; l = r; r = t ^ F3(r, Km[5], Kr[5]);
crypto/cast5_generic.c
333
t = l; l = r; r = t ^ F1(r, Km[6], Kr[6]);
crypto/cast5_generic.c
334
t = l; l = r; r = t ^ F2(r, Km[7], Kr[7]);
crypto/cast5_generic.c
335
t = l; l = r; r = t ^ F3(r, Km[8], Kr[8]);
crypto/cast5_generic.c
336
t = l; l = r; r = t ^ F1(r, Km[9], Kr[9]);
crypto/cast5_generic.c
337
t = l; l = r; r = t ^ F2(r, Km[10], Kr[10]);
crypto/cast5_generic.c
338
t = l; l = r; r = t ^ F3(r, Km[11], Kr[11]);
crypto/cast5_generic.c
340
t = l; l = r; r = t ^ F1(r, Km[12], Kr[12]);
crypto/cast5_generic.c
341
t = l; l = r; r = t ^ F2(r, Km[13], Kr[13]);
crypto/cast5_generic.c
342
t = l; l = r; r = t ^ F3(r, Km[14], Kr[14]);
crypto/cast5_generic.c
343
t = l; l = r; r = t ^ F1(r, Km[15], Kr[15]);
crypto/cast5_generic.c
360
u32 l, r, t;
crypto/cast5_generic.c
372
t = l; l = r; r = t ^ F1(r, Km[15], Kr[15]);
crypto/cast5_generic.c
373
t = l; l = r; r = t ^ F3(r, Km[14], Kr[14]);
crypto/cast5_generic.c
374
t = l; l = r; r = t ^ F2(r, Km[13], Kr[13]);
crypto/cast5_generic.c
375
t = l; l = r; r = t ^ F1(r, Km[12], Kr[12]);
crypto/cast5_generic.c
377
t = l; l = r; r = t ^ F3(r, Km[11], Kr[11]);
crypto/cast5_generic.c
378
t = l; l = r; r = t ^ F2(r, Km[10], Kr[10]);
crypto/cast5_generic.c
379
t = l; l = r; r = t ^ F1(r, Km[9], Kr[9]);
crypto/cast5_generic.c
380
t = l; l = r; r = t ^ F3(r, Km[8], Kr[8]);
crypto/cast5_generic.c
381
t = l; l = r; r = t ^ F2(r, Km[7], Kr[7]);
crypto/cast5_generic.c
382
t = l; l = r; r = t ^ F1(r, Km[6], Kr[6]);
crypto/cast5_generic.c
383
t = l; l = r; r = t ^ F3(r, Km[5], Kr[5]);
crypto/cast5_generic.c
384
t = l; l = r; r = t ^ F2(r, Km[4], Kr[4]);
crypto/cast5_generic.c
385
t = l; l = r; r = t ^ F1(r, Km[3], Kr[3]);
crypto/cast5_generic.c
386
t = l; l = r; r = t ^ F3(r, Km[2], Kr[2]);
crypto/cast5_generic.c
387
t = l; l = r; r = t ^ F2(r, Km[1], Kr[1]);
crypto/cast5_generic.c
388
t = l; l = r; r = t ^ F1(r, Km[0], Kr[0]);
crypto/ecc.c
549
u64 t[ECC_MAX_DIGITS * 2];
crypto/ecc.c
554
vli_umult(t, r + ndigits, c, ndigits);
crypto/ecc.c
556
vli_add(r, r, t, ndigits * 2);
crypto/ecc.c
558
vli_set(t, mod, ndigits);
crypto/ecc.c
559
vli_clear(t + ndigits, ndigits);
crypto/ecc.c
560
while (vli_cmp(r, t, ndigits * 2) >= 0)
crypto/ecc.c
561
vli_sub(r, r, t, ndigits * 2);
crypto/fcrypt.c
60
u32 t = lo & ((1 << n) - 1); \
crypto/fcrypt.c
62
hi = (hi >> n) | (t << (24-n)); \
crypto/lrw.c
146
be128 t = rctx->t;
crypto/lrw.c
177
be128_xor(wdst++, &t, wsrc++);
crypto/lrw.c
181
be128_xor(&t, &t,
crypto/lrw.c
237
memcpy(&rctx->t, req->iv, sizeof(rctx->t));
crypto/lrw.c
240
gf128mul_64k_bbe(&rctx->t, ctx->table);
crypto/lrw.c
53
be128 t;
crypto/xts.c
113
rctx->t = t;
crypto/xts.c
114
gf128mul_x_ble(&t, &t);
crypto/xts.c
116
le128_xor(wdst, &t, wsrc);
crypto/xts.c
118
gf128mul_x_ble(&rctx->t, &t);
crypto/xts.c
123
le128_xor(wdst++, &t, wsrc++);
crypto/xts.c
124
gf128mul_x_ble(&t, &t);
crypto/xts.c
152
le128_xor(&b, &rctx->t, &b);
crypto/xts.c
178
le128_xor(b, &rctx->t, b);
crypto/xts.c
193
le128_xor(b, &rctx->t, b);
crypto/xts.c
256
crypto_cipher_encrypt_one(ctx->tweak, (u8 *)&rctx->t, req->iv);
crypto/xts.c
35
le128 t;
crypto/xts.c
90
le128 t = rctx->t;
drivers/accel/qaic/qaic_timesync.c
130
static void qaic_timesync_timer(struct timer_list *t)
drivers/accel/qaic/qaic_timesync.c
132
struct mqts_dev *mqtsdev = timer_container_of(mqtsdev, t, timer);
drivers/accel/qaic/qaic_timesync.c
169
ret = mod_timer(t, jiffies + msecs_to_jiffies(timesync_delay_ms));
drivers/accessibility/speakup/speakup_dtlk.c
285
u_char *t;
drivers/accessibility/speakup/speakup_dtlk.c
298
t = buf;
drivers/accessibility/speakup/speakup_dtlk.c
300
status.serial_number = t[0] + t[1] * 256;
drivers/accessibility/speakup/speakup_dtlk.c
301
t += 2;
drivers/accessibility/speakup/speakup_dtlk.c
302
for (i = 0; *t != '\r'; t++) {
drivers/accessibility/speakup/speakup_dtlk.c
303
status.rom_version[i] = *t;
drivers/accessibility/speakup/speakup_dtlk.c
308
t++;
drivers/accessibility/speakup/speakup_dtlk.c
309
status.mode = *t++;
drivers/accessibility/speakup/speakup_dtlk.c
310
status.punc_level = *t++;
drivers/accessibility/speakup/speakup_dtlk.c
311
status.formant_freq = *t++;
drivers/accessibility/speakup/speakup_dtlk.c
312
status.pitch = *t++;
drivers/accessibility/speakup/speakup_dtlk.c
313
status.speed = *t++;
drivers/accessibility/speakup/speakup_dtlk.c
314
status.volume = *t++;
drivers/accessibility/speakup/speakup_dtlk.c
315
status.tone = *t++;
drivers/accessibility/speakup/speakup_dtlk.c
316
status.expression = *t++;
drivers/accessibility/speakup/speakup_dtlk.c
317
status.ext_dict_loaded = *t++;
drivers/accessibility/speakup/speakup_dtlk.c
318
status.ext_dict_status = *t++;
drivers/accessibility/speakup/speakup_dtlk.c
319
status.free_ram = *t++;
drivers/accessibility/speakup/speakup_dtlk.c
320
status.articulation = *t++;
drivers/accessibility/speakup/speakup_dtlk.c
321
status.reverb = *t++;
drivers/accessibility/speakup/speakup_dtlk.c
322
status.eob = *t++;
drivers/accessibility/speakup/speakup_ltlk.c
141
unsigned char *t, i;
drivers/accessibility/speakup/speakup_ltlk.c
150
t = buf + 2;
drivers/accessibility/speakup/speakup_ltlk.c
151
for (i = 0; *t != '\r'; t++) {
drivers/accessibility/speakup/speakup_ltlk.c
152
rom_v[i] = *t;
drivers/acpi/acpi_dbg.c
420
struct task_struct *t;
drivers/acpi/acpi_dbg.c
427
t = kthread_create(acpi_aml_thread, NULL, "aml");
drivers/acpi/acpi_dbg.c
428
if (IS_ERR(t)) {
drivers/acpi/acpi_dbg.c
430
return PTR_ERR(t);
drivers/acpi/acpi_dbg.c
434
acpi_aml_io.thread = t;
drivers/acpi/acpi_dbg.c
435
acpi_set_debugger_thread_id((acpi_thread_id)(unsigned long)t);
drivers/acpi/acpi_dbg.c
436
wake_up_process(t);
drivers/acpi/acpica/acmacros.h
377
#define ACPI_SET_DESCRIPTOR_TYPE(d, t) (((union acpi_descriptor *)(void *)(d))->common.descriptor_type = (t))
drivers/acpi/acpica/acutils.h
446
#define acpi_ut_create_internal_object(t) acpi_ut_create_internal_object_dbg (_acpi_module_name,__LINE__,_COMPONENT,t)
drivers/acpi/apei/einj-core.c
237
static int einj_timedout(u64 *t)
drivers/acpi/apei/einj-core.c
239
if ((s64)*t < SLEEP_UNIT_MIN) {
drivers/acpi/apei/einj-core.c
243
*t -= SLEEP_UNIT_MIN;
drivers/acpi/apei/erst.c
118
static int erst_timedout(u64 *t, u64 spin_unit)
drivers/acpi/apei/erst.c
120
if ((s64)*t < spin_unit) {
drivers/acpi/apei/erst.c
124
*t -= spin_unit;
drivers/acpi/apei/ghes.c
1198
static void ghes_poll_func(struct timer_list *t)
drivers/acpi/apei/ghes.c
1200
struct ghes *ghes = timer_container_of(ghes, t, timer);
drivers/acpi/arm64/iort.c
191
struct iort_its_msi_chip *its_msi_chip, *t;
drivers/acpi/arm64/iort.c
194
list_for_each_entry_safe(its_msi_chip, t, &iort_msi_chip_list, list) {
drivers/acpi/ec.c
1178
struct transaction *t;
drivers/acpi/ec.c
1185
t = &q->transaction;
drivers/acpi/ec.c
1186
t->command = ACPI_EC_COMMAND_QUERY;
drivers/acpi/ec.c
1187
t->rdata = pval;
drivers/acpi/ec.c
1188
t->rlen = 1;
drivers/acpi/ec.c
650
static void acpi_ec_spurious_interrupt(struct acpi_ec *ec, struct transaction *t)
drivers/acpi/ec.c
652
if (t->irq_count < ec_storm_threshold)
drivers/acpi/ec.c
653
++t->irq_count;
drivers/acpi/ec.c
656
if (t->irq_count == ec_storm_threshold)
drivers/acpi/ec.c
662
struct transaction *t = ec->curr;
drivers/acpi/ec.c
674
if (!t || !(t->flags & ACPI_EC_COMMAND_POLL)) {
drivers/acpi/ec.c
679
if (!t)
drivers/acpi/ec.c
683
if (t->flags & ACPI_EC_COMMAND_POLL) {
drivers/acpi/ec.c
684
if (t->wlen > t->wi) {
drivers/acpi/ec.c
686
acpi_ec_write_data(ec, t->wdata[t->wi++]);
drivers/acpi/ec.c
688
acpi_ec_spurious_interrupt(ec, t);
drivers/acpi/ec.c
689
} else if (t->rlen > t->ri) {
drivers/acpi/ec.c
691
t->rdata[t->ri++] = acpi_ec_read_data(ec);
drivers/acpi/ec.c
692
if (t->rlen == t->ri) {
drivers/acpi/ec.c
695
if (t->command == ACPI_EC_COMMAND_QUERY)
drivers/acpi/ec.c
700
acpi_ec_spurious_interrupt(ec, t);
drivers/acpi/ec.c
702
} else if (t->wlen == t->wi && !(status & ACPI_EC_FLAG_IBF)) {
drivers/acpi/ec.c
707
acpi_ec_write_cmd(ec, t->command);
drivers/acpi/ec.c
784
struct transaction *t)
drivers/acpi/ec.c
789
if (t->rdata)
drivers/acpi/ec.c
790
memset(t->rdata, 0, t->rlen);
drivers/acpi/ec.c
801
ec->curr = t;
drivers/acpi/ec.c
802
ec_dbg_req("Command(%s) started", acpi_ec_cmd_string(t->command));
drivers/acpi/ec.c
809
if (t->irq_count == ec_storm_threshold)
drivers/acpi/ec.c
811
ec_dbg_req("Command(%s) stopped", acpi_ec_cmd_string(t->command));
drivers/acpi/ec.c
821
static int acpi_ec_transaction(struct acpi_ec *ec, struct transaction *t)
drivers/acpi/ec.c
826
if (!ec || (!t) || (t->wlen && !t->wdata) || (t->rlen && !t->rdata))
drivers/acpi/ec.c
838
status = acpi_ec_transaction_unlocked(ec, t);
drivers/acpi/ec.c
850
struct transaction t = {.command = ACPI_EC_BURST_ENABLE,
drivers/acpi/ec.c
854
return acpi_ec_transaction_unlocked(ec, &t);
drivers/acpi/ec.c
859
struct transaction t = {.command = ACPI_EC_BURST_DISABLE,
drivers/acpi/ec.c
864
acpi_ec_transaction_unlocked(ec, &t) : 0;
drivers/acpi/ec.c
871
struct transaction t = {.command = ACPI_EC_COMMAND_READ,
drivers/acpi/ec.c
875
result = acpi_ec_transaction(ec, &t);
drivers/acpi/ec.c
884
struct transaction t = {.command = ACPI_EC_COMMAND_READ,
drivers/acpi/ec.c
888
result = acpi_ec_transaction_unlocked(ec, &t);
drivers/acpi/ec.c
896
struct transaction t = {.command = ACPI_EC_COMMAND_WRITE,
drivers/acpi/ec.c
900
return acpi_ec_transaction(ec, &t);
drivers/acpi/ec.c
906
struct transaction t = {.command = ACPI_EC_COMMAND_WRITE,
drivers/acpi/ec.c
910
return acpi_ec_transaction_unlocked(ec, &t);
drivers/acpi/ec.c
944
struct transaction t = {.command = command,
drivers/acpi/ec.c
951
return acpi_ec_transaction(first_ec, &t);
drivers/acpi/processor_idle.c
1028
struct acpi_lpi_state *t)
drivers/acpi/processor_idle.c
1030
curr_level->composite_states[curr_level->composite_states_size++] = t;
drivers/acpi/processor_idle.c
1039
struct acpi_lpi_state *p, *t = curr_level->entries;
drivers/acpi/processor_idle.c
1042
for (j = 0; j < state_count; j++, t++) {
drivers/acpi/processor_idle.c
1045
if (!(t->flags & ACPI_LPI_STATE_FLAGS_ENABLED))
drivers/acpi/processor_idle.c
1058
memcpy(flpi, t, sizeof(*t));
drivers/acpi/processor_idle.c
1066
if (t->index <= p->enable_parent_state &&
drivers/acpi/processor_idle.c
1067
combine_lpi_states(p, t, flpi)) {
drivers/acpi/resource.c
765
int irq, p, t;
drivers/acpi/resource.c
784
!acpi_get_override_irq(gsi, &t, &p)) {
drivers/acpi/resource.c
785
u8 trig = t ? ACPI_LEVEL_SENSITIVE : ACPI_EDGE_SENSITIVE;
drivers/acpi/resource.c
790
t ? "level" : "edge",
drivers/acpi/thermal.c
502
int t;
drivers/acpi/thermal.c
513
t = tz->trips.passive.tc1 * (tz->temp_dk -
drivers/acpi/thermal.c
517
if (t > 0)
drivers/acpi/thermal.c
519
else if (t < 0)
drivers/acpi/thermal.c
527
t = acpi_thermal_temp(tz, tz->temp_dk);
drivers/acpi/thermal.c
528
if (t <= trip->temperature)
drivers/android/binder.c
1503
struct binder_transaction *t)
drivers/android/binder.c
1507
BUG_ON(target_thread->transaction_stack != t);
drivers/android/binder.c
1511
t->from = NULL;
drivers/android/binder.c
1578
struct binder_transaction *t)
drivers/android/binder.c
1582
guard(spinlock)(&t->lock);
drivers/android/binder.c
1583
from = t->from;
drivers/android/binder.c
1601
struct binder_transaction *t)
drivers/android/binder.c
1602
__acquires(&t->from->proc->inner_lock)
drivers/android/binder.c
1606
from = binder_get_txn_from(t);
drivers/android/binder.c
1612
if (t->from) {
drivers/android/binder.c
1613
BUG_ON(from != t->from);
drivers/android/binder.c
1632
static void binder_free_txn_fixups(struct binder_transaction *t)
drivers/android/binder.c
1636
list_for_each_entry_safe(fixup, tmp, &t->fd_fixups, fixup_entry) {
drivers/android/binder.c
1645
static void binder_txn_latency_free(struct binder_transaction *t)
drivers/android/binder.c
1649
spin_lock(&t->lock);
drivers/android/binder.c
1650
from_proc = t->from ? t->from->proc->pid : 0;
drivers/android/binder.c
1651
from_thread = t->from ? t->from->pid : 0;
drivers/android/binder.c
1652
to_proc = t->to_proc ? t->to_proc->pid : 0;
drivers/android/binder.c
1653
to_thread = t->to_thread ? t->to_thread->pid : 0;
drivers/android/binder.c
1654
spin_unlock(&t->lock);
drivers/android/binder.c
1656
trace_binder_txn_latency_free(t, from_proc, from_thread, to_proc, to_thread);
drivers/android/binder.c
1659
static void binder_free_transaction(struct binder_transaction *t)
drivers/android/binder.c
1661
struct binder_proc *target_proc = t->to_proc;
drivers/android/binder.c
1671
if (t->buffer)
drivers/android/binder.c
1672
t->buffer->transaction = NULL;
drivers/android/binder.c
1676
binder_txn_latency_free(t);
drivers/android/binder.c
1681
binder_free_txn_fixups(t);
drivers/android/binder.c
1682
kfree(t);
drivers/android/binder.c
1686
static void binder_send_failed_reply(struct binder_transaction *t,
drivers/android/binder.c
1692
BUG_ON(t->flags & TF_ONE_WAY);
drivers/android/binder.c
1694
target_thread = binder_get_txn_from_and_acq_inner(t);
drivers/android/binder.c
1698
t->debug_id,
drivers/android/binder.c
1702
binder_pop_transaction_ilocked(target_thread, t);
drivers/android/binder.c
1721
binder_free_transaction(t);
drivers/android/binder.c
1725
next = t->from_parent;
drivers/android/binder.c
1729
t->debug_id);
drivers/android/binder.c
1731
binder_free_transaction(t);
drivers/android/binder.c
1737
t = next;
drivers/android/binder.c
1740
t->debug_id);
drivers/android/binder.c
1750
static void binder_cleanup_transaction(struct binder_transaction *t,
drivers/android/binder.c
1754
if (t->buffer->target_node && !(t->flags & TF_ONE_WAY)) {
drivers/android/binder.c
1755
binder_send_failed_reply(t, error_code);
drivers/android/binder.c
1759
t->debug_id, reason);
drivers/android/binder.c
1760
binder_free_transaction(t);
drivers/android/binder.c
2221
struct binder_transaction *t,
drivers/android/binder.c
2226
struct binder_proc *target_proc = t->to_proc;
drivers/android/binder.c
2263
trace_binder_transaction_node_to_ref(t, node, &rdata);
drivers/android/binder.c
2274
struct binder_transaction *t,
drivers/android/binder.c
2278
struct binder_proc *target_proc = t->to_proc;
drivers/android/binder.c
2314
trace_binder_transaction_ref_to_node(t, node, &src_rdata);
drivers/android/binder.c
2333
trace_binder_transaction_ref_to_ref(t, node, &src_rdata,
drivers/android/binder.c
2347
struct binder_transaction *t,
drivers/android/binder.c
2352
struct binder_proc *target_proc = t->to_proc;
drivers/android/binder.c
2361
target_allows_fd = t->buffer->target_node->accept_fds;
drivers/android/binder.c
2397
trace_binder_transaction_fd_send(t, fd, fixup->offset);
drivers/android/binder.c
2398
list_add_tail(&fixup->fixup_entry, &t->fd_fixups);
drivers/android/binder.c
2658
struct binder_transaction *t,
drivers/android/binder.c
2691
fda_offset = parent->buffer - t->buffer->user_data +
drivers/android/binder.c
2713
ret = binder_translate_fd(fd, offset, t, thread,
drivers/android/binder.c
2722
struct binder_transaction *t,
drivers/android/binder.c
2731
struct binder_buffer *b = t->buffer;
drivers/android/binder.c
2733
struct binder_proc *target_proc = t->to_proc;
drivers/android/binder.c
2805
binder_find_outdated_transaction_ilocked(struct binder_transaction *t,
drivers/android/binder.c
2816
if (binder_can_update_transaction(t_queued, t))
drivers/android/binder.c
2843
static int binder_proc_transaction(struct binder_transaction *t,
drivers/android/binder.c
2847
struct binder_node *node = t->buffer->target_node;
drivers/android/binder.c
2848
bool oneway = !!(t->flags & TF_ONE_WAY);
drivers/android/binder.c
2881
binder_enqueue_thread_work_ilocked(thread, &t->work);
drivers/android/binder.c
2883
binder_enqueue_work_ilocked(&t->work, &proc->todo);
drivers/android/binder.c
2885
if ((t->flags & TF_UPDATE_TXN) && frozen) {
drivers/android/binder.c
2886
t_outdated = binder_find_outdated_transaction_ilocked(t,
drivers/android/binder.c
2891
t->debug_id, t_outdated->debug_id);
drivers/android/binder.c
2896
binder_enqueue_work_ilocked(&t->work, &node->async_todo);
drivers/android/binder.c
2970
static void binder_set_txn_from_error(struct binder_transaction *t, int id,
drivers/android/binder.c
2973
struct binder_thread *from = binder_get_txn_from_and_acq_inner(t);
drivers/android/binder.c
3000
struct binder_transaction *t,
drivers/android/binder.c
3012
trace_binder_netlink_report(context, t, data_size, error);
drivers/android/binder.c
3024
nla_put_u32(skb, BINDER_A_REPORT_FROM_PID, t->from_pid) ||
drivers/android/binder.c
3025
nla_put_u32(skb, BINDER_A_REPORT_FROM_TID, t->from_tid))
drivers/android/binder.c
3028
if (t->to_proc &&
drivers/android/binder.c
3029
nla_put_u32(skb, BINDER_A_REPORT_TO_PID, t->to_proc->pid))
drivers/android/binder.c
3032
if (t->to_thread &&
drivers/android/binder.c
3033
nla_put_u32(skb, BINDER_A_REPORT_TO_TID, t->to_thread->pid))
drivers/android/binder.c
3036
if (t->is_reply && nla_put_flag(skb, BINDER_A_REPORT_IS_REPLY))
drivers/android/binder.c
3039
if (nla_put_u32(skb, BINDER_A_REPORT_FLAGS, t->flags) ||
drivers/android/binder.c
3040
nla_put_u32(skb, BINDER_A_REPORT_CODE, t->code) ||
drivers/android/binder.c
3061
struct binder_transaction *t;
drivers/android/binder.c
3104
t = kzalloc_obj(*t);
drivers/android/binder.c
3105
if (!t) {
drivers/android/binder.c
3113
INIT_LIST_HEAD(&t->fd_fixups);
drivers/android/binder.c
3115
spin_lock_init(&t->lock);
drivers/android/binder.c
3116
t->debug_id = t_debug_id;
drivers/android/binder.c
3117
t->start_time = t_start_time;
drivers/android/binder.c
3118
t->from_pid = proc->pid;
drivers/android/binder.c
3119
t->from_tid = thread->pid;
drivers/android/binder.c
3120
t->sender_euid = task_euid(proc->tsk);
drivers/android/binder.c
3121
t->code = tr->code;
drivers/android/binder.c
3122
t->flags = tr->flags;
drivers/android/binder.c
3123
t->priority = task_nice(current);
drivers/android/binder.c
3124
t->work.type = BINDER_WORK_TRANSACTION;
drivers/android/binder.c
3125
t->is_async = !reply && (tr->flags & TF_ONE_WAY);
drivers/android/binder.c
3126
t->is_reply = reply;
drivers/android/binder.c
3128
t->from = thread;
drivers/android/binder.c
3317
t->to_proc = target_proc;
drivers/android/binder.c
3318
t->to_thread = target_thread;
drivers/android/binder.c
3337
proc->pid, thread->pid, t->debug_id,
drivers/android/binder.c
3344
proc->pid, thread->pid, t->debug_id,
drivers/android/binder.c
3375
trace_binder_transaction(reply, t, target_node);
drivers/android/binder.c
3377
t->buffer = binder_alloc_new_buf(&target_proc->alloc, tr->data_size,
drivers/android/binder.c
3379
!reply && (t->flags & TF_ONE_WAY));
drivers/android/binder.c
3380
if (IS_ERR(t->buffer)) {
drivers/android/binder.c
3383
ret = PTR_ERR(t->buffer);
drivers/android/binder.c
3390
return_error_param = PTR_ERR(t->buffer);
drivers/android/binder.c
3394
t->buffer = NULL;
drivers/android/binder.c
3404
t->security_ctx = t->buffer->user_data + buf_offset;
drivers/android/binder.c
3406
t->buffer, buf_offset,
drivers/android/binder.c
3409
t->security_ctx = 0;
drivers/android/binder.c
3415
t->buffer->debug_id = t->debug_id;
drivers/android/binder.c
3416
t->buffer->transaction = t;
drivers/android/binder.c
3417
t->buffer->target_node = target_node;
drivers/android/binder.c
3418
t->buffer->clear_on_free = !!(t->flags & TF_CLEAR_BUF);
drivers/android/binder.c
3419
trace_binder_transaction_alloc_buf(t->buffer);
drivers/android/binder.c
3423
t->buffer,
drivers/android/binder.c
3469
t->buffer,
drivers/android/binder.c
3489
t->buffer, user_offset,
drivers/android/binder.c
3500
t->buffer, object_offset, &object);
drivers/android/binder.c
3506
(u64)t->buffer->data_size);
drivers/android/binder.c
3526
ret = binder_translate_binder(fp, t, thread);
drivers/android/binder.c
3530
t->buffer,
drivers/android/binder.c
3546
ret = binder_translate_handle(fp, t, thread);
drivers/android/binder.c
3549
t->buffer,
drivers/android/binder.c
3565
int ret = binder_translate_fd(fp->fd, fd_offset, t,
drivers/android/binder.c
3571
t->buffer,
drivers/android/binder.c
3592
binder_validate_ptr(target_proc, t->buffer,
drivers/android/binder.c
3605
if (!binder_validate_fixup(target_proc, t->buffer,
drivers/android/binder.c
3623
binder_get_object(proc, user_buffer, t->buffer,
drivers/android/binder.c
3637
&user_object.bbo, t,
drivers/android/binder.c
3641
t->buffer,
drivers/android/binder.c
3682
bp->buffer = t->buffer->user_data + sg_buf_offset;
drivers/android/binder.c
3687
ret = binder_fixup_parent(&pf_head, t,
drivers/android/binder.c
3695
t->buffer,
drivers/android/binder.c
3720
t->buffer, user_offset,
drivers/android/binder.c
3731
ret = binder_do_deferred_txn_copies(&target_proc->alloc, t->buffer,
drivers/android/binder.c
3741
if (t->buffer->oneway_spam_suspect) {
drivers/android/binder.c
3743
binder_netlink_report(proc, t, tr->data_size,
drivers/android/binder.c
3757
BUG_ON(t->buffer->async_transaction != 0);
drivers/android/binder.c
3759
binder_enqueue_thread_work_ilocked(target_thread, &t->work);
drivers/android/binder.c
3764
} else if (!(t->flags & TF_ONE_WAY)) {
drivers/android/binder.c
3765
BUG_ON(t->buffer->async_transaction != 0);
drivers/android/binder.c
3775
t->from_parent = thread->transaction_stack;
drivers/android/binder.c
3776
thread->transaction_stack = t;
drivers/android/binder.c
3778
return_error = binder_proc_transaction(t,
drivers/android/binder.c
3782
binder_pop_transaction_ilocked(thread, t);
drivers/android/binder.c
3793
struct binder_transaction t_copy = *t;
drivers/android/binder.c
3796
BUG_ON(t->buffer->async_transaction != 1);
drivers/android/binder.c
3797
return_error = binder_proc_transaction(t, target_proc, NULL);
drivers/android/binder.c
3838
binder_free_txn_fixups(t);
drivers/android/binder.c
3839
trace_binder_transaction_failed_buffer_release(t->buffer);
drivers/android/binder.c
3840
binder_transaction_buffer_release(target_proc, NULL, t->buffer,
drivers/android/binder.c
3845
t->buffer->transaction = NULL;
drivers/android/binder.c
3846
binder_alloc_free_buf(&target_proc->alloc, t->buffer);
drivers/android/binder.c
3856
binder_txn_latency_free(t);
drivers/android/binder.c
3867
binder_netlink_report(proc, t, tr->data_size, return_error);
drivers/android/binder.c
3868
kfree(t);
drivers/android/binder.c
4689
struct binder_transaction *t)
drivers/android/binder.c
4694
list_for_each_entry(fixup, &t->fd_fixups, fixup_entry) {
drivers/android/binder.c
4700
t->debug_id, fd);
drivers/android/binder.c
4706
t->debug_id, fd);
drivers/android/binder.c
4707
trace_binder_transaction_fd_recv(t, fd, fixup->offset);
drivers/android/binder.c
4709
if (binder_alloc_copy_to_buffer(&proc->alloc, t->buffer,
drivers/android/binder.c
4716
list_for_each_entry_safe(fixup, tmp, &t->fd_fixups, fixup_entry) {
drivers/android/binder.c
4725
binder_free_txn_fixups(t);
drivers/android/binder.c
4786
struct binder_transaction *t = NULL;
drivers/android/binder.c
4816
t = container_of(w, struct binder_transaction, work);
drivers/android/binder.c
5029
if (!t)
drivers/android/binder.c
5032
BUG_ON(t->buffer == NULL);
drivers/android/binder.c
5033
if (t->buffer->target_node) {
drivers/android/binder.c
5034
struct binder_node *target_node = t->buffer->target_node;
drivers/android/binder.c
5038
t->saved_priority = task_nice(current);
drivers/android/binder.c
5039
if (t->priority < target_node->min_priority &&
drivers/android/binder.c
5040
!(t->flags & TF_ONE_WAY))
drivers/android/binder.c
5041
binder_set_nice(t->priority);
drivers/android/binder.c
5042
else if (!(t->flags & TF_ONE_WAY) ||
drivers/android/binder.c
5043
t->saved_priority > target_node->min_priority)
drivers/android/binder.c
5051
trd->code = t->code;
drivers/android/binder.c
5052
trd->flags = t->flags;
drivers/android/binder.c
5053
trd->sender_euid = from_kuid(current_user_ns(), t->sender_euid);
drivers/android/binder.c
5055
t_from = binder_get_txn_from(t);
drivers/android/binder.c
5066
ret = binder_apply_fd_fixups(proc, t);
drivers/android/binder.c
5068
struct binder_buffer *buffer = t->buffer;
drivers/android/binder.c
5069
bool oneway = !!(t->flags & TF_ONE_WAY);
drivers/android/binder.c
5070
int tid = t->debug_id;
drivers/android/binder.c
5075
binder_cleanup_transaction(t, "fd fixups failed",
drivers/android/binder.c
5094
trd->data_size = t->buffer->data_size;
drivers/android/binder.c
5095
trd->offsets_size = t->buffer->offsets_size;
drivers/android/binder.c
5096
trd->data.ptr.buffer = t->buffer->user_data;
drivers/android/binder.c
5098
ALIGN(t->buffer->data_size,
drivers/android/binder.c
5101
tr.secctx = t->security_ctx;
drivers/android/binder.c
5102
if (t->security_ctx) {
drivers/android/binder.c
5110
binder_cleanup_transaction(t, "put_user failed",
drivers/android/binder.c
5120
binder_cleanup_transaction(t, "copy_to_user failed",
drivers/android/binder.c
5127
trace_binder_transaction_received(t);
drivers/android/binder.c
5135
t->debug_id, t_from ? t_from->proc->pid : 0,
drivers/android/binder.c
5137
t->buffer->data_size, t->buffer->offsets_size);
drivers/android/binder.c
5141
t->buffer->allow_user_free = 1;
drivers/android/binder.c
5142
if (cmd != BR_REPLY && !(t->flags & TF_ONE_WAY)) {
drivers/android/binder.c
5144
t->to_parent = thread->transaction_stack;
drivers/android/binder.c
5145
t->to_thread = thread;
drivers/android/binder.c
5146
thread->transaction_stack = t;
drivers/android/binder.c
5149
binder_free_transaction(t);
drivers/android/binder.c
5193
struct binder_transaction *t;
drivers/android/binder.c
5195
t = container_of(w, struct binder_transaction, work);
drivers/android/binder.c
5197
binder_cleanup_transaction(t, "process died.",
drivers/android/binder.c
5342
struct binder_transaction *t;
drivers/android/binder.c
5361
t = thread->transaction_stack;
drivers/android/binder.c
5362
if (t) {
drivers/android/binder.c
5363
spin_lock(&t->lock);
drivers/android/binder.c
5364
if (t->to_thread == thread)
drivers/android/binder.c
5365
send_reply = t;
drivers/android/binder.c
5367
__acquire(&t->lock);
drivers/android/binder.c
5371
while (t) {
drivers/android/binder.c
5372
last_t = t;
drivers/android/binder.c
5377
t->debug_id,
drivers/android/binder.c
5378
(t->to_thread == thread) ? "in" : "out");
drivers/android/binder.c
5380
if (t->to_thread == thread) {
drivers/android/binder.c
5382
t->to_proc = NULL;
drivers/android/binder.c
5383
t->to_thread = NULL;
drivers/android/binder.c
5384
if (t->buffer) {
drivers/android/binder.c
5385
t->buffer->transaction = NULL;
drivers/android/binder.c
5386
t->buffer = NULL;
drivers/android/binder.c
5388
t = t->to_parent;
drivers/android/binder.c
5389
} else if (t->from == thread) {
drivers/android/binder.c
5390
t->from = NULL;
drivers/android/binder.c
5391
t = t->from_parent;
drivers/android/binder.c
5395
if (t)
drivers/android/binder.c
5396
spin_lock(&t->lock);
drivers/android/binder.c
5398
__acquire(&t->lock);
drivers/android/binder.c
5401
__release(&t->lock);
drivers/android/binder.c
6392
struct binder_transaction *t)
drivers/android/binder.c
6395
struct binder_buffer *buffer = t->buffer;
drivers/android/binder.c
6398
spin_lock(&t->lock);
drivers/android/binder.c
6399
to_proc = t->to_proc;
drivers/android/binder.c
6402
prefix, t->debug_id, t,
drivers/android/binder.c
6403
t->from_pid,
drivers/android/binder.c
6404
t->from_tid,
drivers/android/binder.c
6406
t->to_thread ? t->to_thread->pid : 0,
drivers/android/binder.c
6407
t->code, t->flags, t->priority, t->is_async, t->is_reply,
drivers/android/binder.c
6408
ktime_ms_delta(current_time, t->start_time));
drivers/android/binder.c
6409
spin_unlock(&t->lock);
drivers/android/binder.c
6438
struct binder_transaction *t;
drivers/android/binder.c
6442
t = container_of(w, struct binder_transaction, work);
drivers/android/binder.c
6444
m, proc, transaction_prefix, t);
drivers/android/binder.c
6493
struct binder_transaction *t;
drivers/android/binder.c
6503
t = thread->transaction_stack;
drivers/android/binder.c
6504
while (t) {
drivers/android/binder.c
6505
if (t->from == thread) {
drivers/android/binder.c
6507
" outgoing transaction", t);
drivers/android/binder.c
6508
t = t->from_parent;
drivers/android/binder.c
6509
} else if (t->to_thread == thread) {
drivers/android/binder.c
6511
" incoming transaction", t);
drivers/android/binder.c
6512
t = t->to_parent;
drivers/android/binder.c
6515
" bad transaction", t);
drivers/android/binder.c
6516
t = NULL;
drivers/android/binder/rust_binder.h
53
struct rb_transaction_layout t;
drivers/android/binder/rust_binder.h
60
static inline size_t rust_binder_transaction_debug_id(rust_binder_transaction t)
drivers/android/binder/rust_binder.h
62
return *(size_t *) (t + RUST_BINDER_LAYOUT.t.debug_id);
drivers/android/binder/rust_binder.h
65
static inline u32 rust_binder_transaction_code(rust_binder_transaction t)
drivers/android/binder/rust_binder.h
67
return *(u32 *) (t + RUST_BINDER_LAYOUT.t.code);
drivers/android/binder/rust_binder.h
70
static inline u32 rust_binder_transaction_flags(rust_binder_transaction t)
drivers/android/binder/rust_binder.h
72
return *(u32 *) (t + RUST_BINDER_LAYOUT.t.flags);
drivers/android/binder/rust_binder.h
76
static inline rust_binder_node rust_binder_transaction_target_node(rust_binder_transaction t)
drivers/android/binder/rust_binder.h
78
void *p = *(void **) (t + RUST_BINDER_LAYOUT.t.target_node);
drivers/android/binder/rust_binder.h
85
static inline rust_binder_process rust_binder_transaction_to_proc(rust_binder_transaction t)
drivers/android/binder/rust_binder.h
87
void *p = *(void **) (t + RUST_BINDER_LAYOUT.t.to_proc);
drivers/android/binder/rust_binder.h
92
static inline struct task_struct *rust_binder_process_task(rust_binder_process t)
drivers/android/binder/rust_binder.h
94
return *(struct task_struct **) (t + RUST_BINDER_LAYOUT.p.task);
drivers/android/binder/rust_binder.h
97
static inline size_t rust_binder_node_debug_id(rust_binder_node t)
drivers/android/binder/rust_binder.h
99
return *(size_t *) (t + RUST_BINDER_LAYOUT.n.debug_id);
drivers/android/binder/rust_binder_events.h
34
TP_PROTO(bool reply, rust_binder_transaction t, struct task_struct *thread),
drivers/android/binder/rust_binder_events.h
35
TP_ARGS(reply, t, thread),
drivers/android/binder/rust_binder_events.h
46
rust_binder_process to = rust_binder_transaction_to_proc(t);
drivers/android/binder/rust_binder_events.h
47
rust_binder_node target_node = rust_binder_transaction_target_node(t);
drivers/android/binder/rust_binder_events.h
49
__entry->debug_id = rust_binder_transaction_debug_id(t);
drivers/android/binder/rust_binder_events.h
54
__entry->code = rust_binder_transaction_code(t);
drivers/android/binder/rust_binder_events.h
55
__entry->flags = rust_binder_transaction_flags(t);
drivers/android/binder_trace.h
107
TP_PROTO(bool reply, struct binder_transaction *t,
drivers/android/binder_trace.h
109
TP_ARGS(reply, t, target_node),
drivers/android/binder_trace.h
120
__entry->debug_id = t->debug_id;
drivers/android/binder_trace.h
122
__entry->to_proc = t->to_proc->pid;
drivers/android/binder_trace.h
123
__entry->to_thread = t->to_thread ? t->to_thread->pid : 0;
drivers/android/binder_trace.h
125
__entry->code = t->code;
drivers/android/binder_trace.h
126
__entry->flags = t->flags;
drivers/android/binder_trace.h
135
TP_PROTO(struct binder_transaction *t),
drivers/android/binder_trace.h
136
TP_ARGS(t),
drivers/android/binder_trace.h
142
__entry->debug_id = t->debug_id;
drivers/android/binder_trace.h
148
TP_PROTO(struct binder_transaction *t, struct binder_node *node,
drivers/android/binder_trace.h
150
TP_ARGS(t, node, rdata),
drivers/android/binder_trace.h
160
__entry->debug_id = t->debug_id;
drivers/android/binder_trace.h
173
TP_PROTO(struct binder_transaction *t, struct binder_node *node,
drivers/android/binder_trace.h
175
TP_ARGS(t, node, rdata),
drivers/android/binder_trace.h
185
__entry->debug_id = t->debug_id;
drivers/android/binder_trace.h
198
TP_PROTO(struct binder_transaction *t, struct binder_node *node,
drivers/android/binder_trace.h
201
TP_ARGS(t, node, src_ref, dest_ref),
drivers/android/binder_trace.h
212
__entry->debug_id = t->debug_id;
drivers/android/binder_trace.h
226
TP_PROTO(struct binder_transaction *t, int fd, size_t offset),
drivers/android/binder_trace.h
227
TP_ARGS(t, fd, offset),
drivers/android/binder_trace.h
235
__entry->debug_id = t->debug_id;
drivers/android/binder_trace.h
244
TP_PROTO(struct binder_transaction *t, int fd, size_t offset),
drivers/android/binder_trace.h
245
TP_ARGS(t, fd, offset),
drivers/android/binder_trace.h
253
__entry->debug_id = t->debug_id;
drivers/android/binder_trace.h
407
struct binder_transaction *t,
drivers/android/binder_trace.h
410
TP_ARGS(context, t, data_size, error),
drivers/android/binder_trace.h
426
__entry->from_pid = t->from_pid;
drivers/android/binder_trace.h
427
__entry->from_tid = t->from_tid;
drivers/android/binder_trace.h
428
__entry->to_pid = t->to_proc ? t->to_proc->pid : 0;
drivers/android/binder_trace.h
429
__entry->to_tid = t->to_thread ? t->to_thread->pid : 0;
drivers/android/binder_trace.h
430
__entry->is_reply = t->is_reply;
drivers/android/binder_trace.h
431
__entry->flags = t->flags;
drivers/android/binder_trace.h
432
__entry->code = t->code;
drivers/android/binder_trace.h
78
TP_PROTO(struct binder_transaction *t,
drivers/android/binder_trace.h
81
TP_ARGS(t, from_proc, from_thread, to_proc, to_thread),
drivers/android/binder_trace.h
92
__entry->debug_id = t->debug_id;
drivers/android/binder_trace.h
97
__entry->code = t->code;
drivers/android/binder_trace.h
98
__entry->flags = t->flags;
drivers/ata/ata_generic.c
129
u16 t;
drivers/ata/ata_generic.c
138
pci_read_config_word(dev, 0x40, &t);
drivers/ata/ata_generic.c
139
if (t != 0)
drivers/ata/ata_generic.c
145
pci_read_config_word(dev, 0x40, &t);
drivers/ata/ata_generic.c
146
if (t) {
drivers/ata/libahci.c
1036
static void ahci_sw_activity_blink(struct timer_list *t)
drivers/ata/libahci.c
1038
struct ahci_em_priv *emp = timer_container_of(emp, t, timer);
drivers/ata/libata-core.c
3350
const struct ata_timing *t;
drivers/ata/libata-core.c
3356
for (t = ata_timing_find_mode(base_mode);
drivers/ata/libata-core.c
3357
t && ata_xfer_mode2shift(t->mode) == xfer_shift; t++) {
drivers/ata/libata-core.c
3363
this_cycle = t->cycle;
drivers/ata/libata-core.c
3366
this_cycle = t->udma;
drivers/ata/libata-core.c
3375
last_mode = t->mode;
drivers/ata/libata-eh.c
877
void ata_eh_fastdrain_timerfn(struct timer_list *t)
drivers/ata/libata-eh.c
879
struct ata_port *ap = timer_container_of(ap, t, fastdrain_timer);
drivers/ata/libata-pata-timings.c
100
const struct ata_timing *t = ata_timing;
drivers/ata/libata-pata-timings.c
102
while (xfer_mode > t->mode)
drivers/ata/libata-pata-timings.c
103
t++;
drivers/ata/libata-pata-timings.c
105
if (xfer_mode == t->mode)
drivers/ata/libata-pata-timings.c
106
return t;
drivers/ata/libata-pata-timings.c
116
struct ata_timing *t, int T, int UT)
drivers/ata/libata-pata-timings.c
129
memcpy(t, s, sizeof(*s));
drivers/ata/libata-pata-timings.c
148
ata_timing_merge(&p, t, t, ATA_TIMING_CYCLE | ATA_TIMING_CYC8B);
drivers/ata/libata-pata-timings.c
155
ata_timing_quantize(t, t, T, UT);
drivers/ata/libata-pata-timings.c
165
ata_timing_merge(&p, t, t, ATA_TIMING_ALL);
drivers/ata/libata-pata-timings.c
172
if (t->act8b + t->rec8b < t->cyc8b) {
drivers/ata/libata-pata-timings.c
173
t->act8b += (t->cyc8b - (t->act8b + t->rec8b)) / 2;
drivers/ata/libata-pata-timings.c
174
t->rec8b = t->cyc8b - t->act8b;
drivers/ata/libata-pata-timings.c
177
if (t->active + t->recover < t->cycle) {
drivers/ata/libata-pata-timings.c
178
t->active += (t->cycle - (t->active + t->recover)) / 2;
drivers/ata/libata-pata-timings.c
179
t->recover = t->cycle - t->active;
drivers/ata/libata-pata-timings.c
187
if (t->active + t->recover > t->cycle)
drivers/ata/libata-pata-timings.c
188
t->cycle = t->active + t->recover;
drivers/ata/libata-pata-timings.c
60
static void ata_timing_quantize(const struct ata_timing *t,
drivers/ata/libata-pata-timings.c
63
q->setup = EZ(t->setup, T);
drivers/ata/libata-pata-timings.c
64
q->act8b = EZ(t->act8b, T);
drivers/ata/libata-pata-timings.c
65
q->rec8b = EZ(t->rec8b, T);
drivers/ata/libata-pata-timings.c
66
q->cyc8b = EZ(t->cyc8b, T);
drivers/ata/libata-pata-timings.c
67
q->active = EZ(t->active, T);
drivers/ata/libata-pata-timings.c
68
q->recover = EZ(t->recover, T);
drivers/ata/libata-pata-timings.c
69
q->dmack_hold = EZ(t->dmack_hold, T);
drivers/ata/libata-pata-timings.c
70
q->cycle = EZ(t->cycle, T);
drivers/ata/libata-pata-timings.c
71
q->udma = EZ(t->udma, UT);
drivers/ata/libata-sata.c
240
unsigned long last_jiffies, t;
drivers/ata/libata-sata.c
244
t = ata_deadline(jiffies, params[2]);
drivers/ata/libata-sata.c
245
if (time_before(t, deadline))
drivers/ata/libata-sata.c
246
deadline = t;
drivers/ata/libata-transport.c
48
struct scsi_transport_template t;
drivers/ata/libata-transport.c
65
#define to_ata_internal(tmpl) container_of(tmpl, struct ata_internal, t)
drivers/ata/libata-transport.c
765
i->t.eh_strategy_handler = ata_scsi_error;
drivers/ata/libata-transport.c
766
i->t.user_scan = ata_scsi_user_scan;
drivers/ata/libata-transport.c
768
i->t.host_attrs.ac.attrs = &i->port_attrs[0];
drivers/ata/libata-transport.c
769
i->t.host_attrs.ac.class = &ata_port_class.class;
drivers/ata/libata-transport.c
770
i->t.host_attrs.ac.match = ata_tport_match;
drivers/ata/libata-transport.c
771
transport_container_register(&i->t.host_attrs);
drivers/ata/libata-transport.c
810
return &i->t;
drivers/ata/libata-transport.c
817
void ata_release_transport(struct scsi_transport_template *t)
drivers/ata/libata-transport.c
819
struct ata_internal *i = to_ata_internal(t);
drivers/ata/libata-transport.c
821
transport_container_unregister(&i->t.host_attrs);
drivers/ata/libata-transport.h
12
void ata_release_transport(struct scsi_transport_template *t);
drivers/ata/libata.h
179
extern void ata_eh_fastdrain_timerfn(struct timer_list *t);
drivers/ata/pata_acpi.c
116
const struct ata_timing *t;
drivers/ata/pata_acpi.c
122
t = ata_timing_find_mode(adev->pio_mode);
drivers/ata/pata_acpi.c
123
acpi->gtm.drive[unit].pio = t->cycle;
drivers/ata/pata_acpi.c
139
const struct ata_timing *t;
drivers/ata/pata_acpi.c
145
t = ata_timing_find_mode(adev->dma_mode);
drivers/ata/pata_acpi.c
147
acpi->gtm.drive[unit].dma = t->udma;
drivers/ata/pata_acpi.c
150
acpi->gtm.drive[unit].dma = t->cycle;
drivers/ata/pata_ali.c
170
static void ali_program_modes(struct ata_port *ap, struct ata_device *adev, struct ata_timing *t, u8 ultra)
drivers/ata/pata_ali.c
180
if (t != NULL) {
drivers/ata/pata_ali.c
181
t->setup = clamp_val(t->setup, 1, 8) & 7;
drivers/ata/pata_ali.c
182
t->act8b = clamp_val(t->act8b, 1, 8) & 7;
drivers/ata/pata_ali.c
183
t->rec8b = clamp_val(t->rec8b, 1, 16) & 15;
drivers/ata/pata_ali.c
184
t->active = clamp_val(t->active, 1, 8) & 7;
drivers/ata/pata_ali.c
185
t->recover = clamp_val(t->recover, 1, 16) & 15;
drivers/ata/pata_ali.c
187
pci_write_config_byte(pdev, cas, t->setup);
drivers/ata/pata_ali.c
188
pci_write_config_byte(pdev, cbt, (t->act8b << 4) | t->rec8b);
drivers/ata/pata_ali.c
189
pci_write_config_byte(pdev, drwt, (t->active << 4) | t->recover);
drivers/ata/pata_ali.c
210
struct ata_timing t;
drivers/ata/pata_ali.c
213
ata_timing_compute(adev, adev->pio_mode, &t, T, 1);
drivers/ata/pata_ali.c
217
ata_timing_merge(&p, &t, &t, ATA_TIMING_SETUP|ATA_TIMING_8BIT);
drivers/ata/pata_ali.c
220
ata_timing_merge(&p, &t, &t, ATA_TIMING_SETUP|ATA_TIMING_8BIT);
drivers/ata/pata_ali.c
227
ali_program_modes(ap, adev, &t, 0);
drivers/ata/pata_ali.c
245
struct ata_timing t;
drivers/ata/pata_ali.c
262
ata_timing_compute(adev, adev->dma_mode, &t, T, 1);
drivers/ata/pata_ali.c
266
ata_timing_merge(&p, &t, &t, ATA_TIMING_SETUP|ATA_TIMING_8BIT);
drivers/ata/pata_ali.c
269
ata_timing_merge(&p, &t, &t, ATA_TIMING_SETUP|ATA_TIMING_8BIT);
drivers/ata/pata_ali.c
272
ali_program_modes(ap, adev, &t, 0);
drivers/ata/pata_amd.c
103
t = at.udma ? (0xc0 | amd_cyc2udma[clamp_val(at.udma, 2, 10)]) : 0x03;
drivers/ata/pata_amd.c
107
t = at.udma ? (0xc0 | amd_cyc2udma[clamp_val(at.udma, 1, 10)]) : 0x03;
drivers/ata/pata_amd.c
111
t = at.udma ? (0xc0 | amd_cyc2udma[clamp_val(at.udma, 1, 15)]) : 0x03;
drivers/ata/pata_amd.c
120
pci_write_config_byte(pdev, offset + 0x10 + (3 - dn), t);
drivers/ata/pata_amd.c
55
u8 t;
drivers/ata/pata_amd.c
85
pci_read_config_byte(pdev, offset + 0x0C, &t);
drivers/ata/pata_amd.c
86
t = (t & ~(3 << ((3 - dn) << 1))) | ((clamp_val(at.setup, 1, 4) - 1) << ((3 - dn) << 1));
drivers/ata/pata_amd.c
87
pci_write_config_byte(pdev, offset + 0x0C , t);
drivers/ata/pata_amd.c
99
t = at.udma ? (0xc0 | (clamp_val(at.udma, 2, 5) - 2)) : 0x03;
drivers/ata/pata_atp867x.c
208
struct ata_timing t, p;
drivers/ata/pata_atp867x.c
215
ata_timing_compute(adev, speed, &t, T, UT);
drivers/ata/pata_atp867x.c
218
ata_timing_merge(&p, &t, &t, ATA_TIMING_8BIT);
drivers/ata/pata_atp867x.c
228
b = atp867x_get_active_clocks_shifted(ap, t.active) |
drivers/ata/pata_atp867x.c
229
atp867x_get_recover_clocks_shifted(ap, t.recover);
drivers/ata/pata_atp867x.c
236
b = atp867x_get_active_clocks_shifted(ap, t.act8b) |
drivers/ata/pata_atp867x.c
237
atp867x_get_recover_clocks_shifted(ap, t.rec8b);
drivers/ata/pata_cmd640.c
103
reg |= t.setup;
drivers/ata/pata_cmd640.c
107
pci_write_config_byte(pdev, arttim + 1, (t.active << 4) | t.recover);
drivers/ata/pata_cmd640.c
114
reg |= t.setup;
drivers/ata/pata_cmd640.c
116
timing->reg58[adev->devno] = (t.active << 4) | t.recover;
drivers/ata/pata_cmd640.c
56
struct ata_timing t;
drivers/ata/pata_cmd640.c
63
if (ata_timing_compute(adev, adev->pio_mode, &t, T, 0) < 0) {
drivers/ata/pata_cmd640.c
73
ata_timing_merge(&p, &t, &t, ATA_TIMING_SETUP);
drivers/ata/pata_cmd640.c
77
if (t.recover > 16) {
drivers/ata/pata_cmd640.c
78
t.active += t.recover - 16;
drivers/ata/pata_cmd640.c
79
t.recover = 16;
drivers/ata/pata_cmd640.c
81
if (t.active > 16)
drivers/ata/pata_cmd640.c
82
t.active = 16;
drivers/ata/pata_cmd640.c
87
if (t.recover > 1)
drivers/ata/pata_cmd640.c
88
t.recover--; /* 640B only */
drivers/ata/pata_cmd640.c
90
t.recover = 15;
drivers/ata/pata_cmd640.c
92
if (t.setup > 4)
drivers/ata/pata_cmd640.c
93
t.setup = 0xC0;
drivers/ata/pata_cmd640.c
95
t.setup = setup_data[t.setup];
drivers/ata/pata_cmd640.c
98
t.active &= 0x0F; /* 0 = 16 */
drivers/ata/pata_cmd64x.c
118
if (ata_timing_compute(adev, mode, &t, T, 0) < 0) {
drivers/ata/pata_cmd64x.c
129
ata_timing_merge(&t, &tp, &t, ATA_TIMING_SETUP);
drivers/ata/pata_cmd64x.c
134
t.active, t.recover, t.setup);
drivers/ata/pata_cmd64x.c
135
if (t.recover > 16) {
drivers/ata/pata_cmd64x.c
136
t.active += t.recover - 16;
drivers/ata/pata_cmd64x.c
137
t.recover = 16;
drivers/ata/pata_cmd64x.c
139
if (t.active > 16)
drivers/ata/pata_cmd64x.c
140
t.active = 16;
drivers/ata/pata_cmd64x.c
145
if (t.recover == 16)
drivers/ata/pata_cmd64x.c
146
t.recover = 0;
drivers/ata/pata_cmd64x.c
147
else if (t.recover > 1)
drivers/ata/pata_cmd64x.c
148
t.recover--;
drivers/ata/pata_cmd64x.c
150
t.recover = 15;
drivers/ata/pata_cmd64x.c
152
if (t.setup > 4)
drivers/ata/pata_cmd64x.c
153
t.setup = 0xC0;
drivers/ata/pata_cmd64x.c
155
t.setup = setup_data[t.setup];
drivers/ata/pata_cmd64x.c
157
t.active &= 0x0F; /* 0 = 16 */
drivers/ata/pata_cmd64x.c
162
reg |= t.setup;
drivers/ata/pata_cmd64x.c
166
pci_write_config_byte(pdev, drwtim, (t.active << 4) | t.recover);
drivers/ata/pata_cmd64x.c
97
struct ata_timing t;
drivers/ata/pata_cypress.c
59
struct ata_timing t;
drivers/ata/pata_cypress.c
64
if (ata_timing_compute(adev, adev->pio_mode, &t, T, 1) < 0) {
drivers/ata/pata_cypress.c
69
time_16 = clamp_val(t.recover - 1, 0, 15) |
drivers/ata/pata_cypress.c
70
(clamp_val(t.active - 1, 0, 15) << 4);
drivers/ata/pata_cypress.c
71
time_8 = clamp_val(t.act8b - 1, 0, 15) |
drivers/ata/pata_cypress.c
72
(clamp_val(t.rec8b - 1, 0, 15) << 4);
drivers/ata/pata_cypress.c
78
addr |= clamp_val(t.setup - 1, 0, 15);
drivers/ata/pata_cypress.c
88
addr |= (clamp_val(t.setup - 1, 0, 15) << 4);
drivers/ata/pata_ep93xx.c
131
struct ata_timing t;
drivers/ata/pata_ep93xx.c
266
const struct ata_timing *t = &drv_data->t;
drivers/ata/pata_ep93xx.c
267
unsigned long t0 = reg ? t->cyc8b : t->cycle;
drivers/ata/pata_ep93xx.c
268
unsigned long t2 = reg ? t->act8b : t->active;
drivers/ata/pata_ep93xx.c
269
unsigned long t2i = reg ? t->rec8b : t->recover;
drivers/ata/pata_ep93xx.c
271
ep93xx_pata_rw_begin(base, addr, t->setup);
drivers/ata/pata_ep93xx.c
300
const struct ata_timing *t = &drv_data->t;
drivers/ata/pata_ep93xx.c
301
unsigned long t0 = reg ? t->cyc8b : t->cycle;
drivers/ata/pata_ep93xx.c
302
unsigned long t2 = reg ? t->act8b : t->active;
drivers/ata/pata_ep93xx.c
303
unsigned long t2i = reg ? t->rec8b : t->recover;
drivers/ata/pata_ep93xx.c
305
ep93xx_pata_rw_begin(base, addr, t->setup);
drivers/ata/pata_ep93xx.c
343
ata_timing_compute(adev, adev->pio_mode, &drv_data->t, T, 0);
drivers/ata/pata_ep93xx.c
345
struct ata_timing t;
drivers/ata/pata_ep93xx.c
346
ata_timing_compute(pair, pair->pio_mode, &t, T, 0);
drivers/ata/pata_ep93xx.c
347
ata_timing_merge(&t, &drv_data->t, &drv_data->t,
drivers/ata/pata_ep93xx.c
867
drv_data->t = *ata_timing_find_mode(XFER_PIO_0);
drivers/ata/pata_hpt366.c
232
u32 mask, reg, t;
drivers/ata/pata_hpt366.c
242
t = hpt36x_find_mode(ap, mode);
drivers/ata/pata_hpt366.c
250
reg = ((reg & ~mask) | (t & mask)) & ~0xc0000000;
drivers/ata/pata_icside.c
191
struct ata_timing t;
drivers/ata/pata_icside.c
198
if (ata_timing_compute(adev, adev->dma_mode, &t, 1000, 1))
drivers/ata/pata_icside.c
205
if (t.active <= 50 && t.recover <= 375 && t.cycle <= 425) {
drivers/ata/pata_icside.c
208
} else if (t.active <= 125 && t.recover <= 375 && t.cycle <= 500) {
drivers/ata/pata_icside.c
211
} else if (t.active <= 200 && t.recover <= 550 && t.cycle <= 750) {
drivers/ata/pata_icside.c
220
t.active, t.recover, t.cycle, iomd_type);
drivers/ata/pata_macio.c
409
const struct pata_macio_timing *t;
drivers/ata/pata_macio.c
422
t = pata_macio_find_timing(priv, adev->pio_mode);
drivers/ata/pata_macio.c
423
if (t == NULL) {
drivers/ata/pata_macio.c
426
t = pata_macio_find_timing(priv, XFER_PIO_0);
drivers/ata/pata_macio.c
428
BUG_ON(t == NULL);
drivers/ata/pata_macio.c
431
priv->treg[adev->devno][0] |= t->reg1;
drivers/ata/pata_macio.c
434
t = pata_macio_find_timing(priv, adev->dma_mode);
drivers/ata/pata_macio.c
435
if (t == NULL || (t->reg1 == 0 && t->reg2 == 0)) {
drivers/ata/pata_macio.c
437
t = pata_macio_find_timing(priv, XFER_MW_DMA_0);
drivers/ata/pata_macio.c
439
BUG_ON(t == NULL);
drivers/ata/pata_macio.c
442
priv->treg[adev->devno][0] |= t->reg1;
drivers/ata/pata_macio.c
443
priv->treg[adev->devno][1] |= t->reg2;
drivers/ata/pata_macio.c
88
#define SYSCLK_TICKS(t) (((t) + IDE_SYSCLK_NS - 1) / IDE_SYSCLK_NS)
drivers/ata/pata_macio.c
89
#define SYSCLK_TICKS_66(t) (((t) + IDE_SYSCLK_66_NS - 1) / IDE_SYSCLK_66_NS)
drivers/ata/pata_mpc52xx.c
298
struct mpc52xx_ata_timings *t = &priv->timings[dev];
drivers/ata/pata_mpc52xx.c
304
t->mdma1 = ((u32)s->t0M << 24) | ((u32)s->td << 16) | ((u32)s->tkw << 8) | s->tm;
drivers/ata/pata_mpc52xx.c
305
t->mdma2 = ((u32)s->th << 24) | ((u32)s->tj << 16) | ((u32)s->tn << 8);
drivers/ata/pata_mpc52xx.c
306
t->using_udma = 0;
drivers/ata/pata_mpc52xx.c
315
struct mpc52xx_ata_timings *t = &priv->timings[dev];
drivers/ata/pata_mpc52xx.c
321
t->udma1 = ((u32)s->t2cyc << 24) | ((u32)s->tcyc << 16) | ((u32)s->tds << 8) | s->tdh;
drivers/ata/pata_mpc52xx.c
322
t->udma2 = ((u32)s->tdvs << 24) | ((u32)s->tdvh << 16) | ((u32)s->tfs << 8) | s->tli;
drivers/ata/pata_mpc52xx.c
323
t->udma3 = ((u32)s->tmli << 24) | ((u32)s->taz << 16) | ((u32)s->tenv << 8) | s->tsr;
drivers/ata/pata_mpc52xx.c
324
t->udma4 = ((u32)s->tss << 24) | ((u32)s->trfs << 16) | ((u32)s->trp << 8) | s->tack;
drivers/ata/pata_mpc52xx.c
325
t->udma5 = (u32)s->tzah << 24;
drivers/ata/pata_mpc52xx.c
326
t->using_udma = 1;
drivers/ata/pata_ns87415.c
58
struct ata_timing t;
drivers/ata/pata_ns87415.c
66
ata_timing_compute(adev, adev->pio_mode, &t, T, 0);
drivers/ata/pata_ns87415.c
68
clocking = 17 - clamp_val(t.active, 2, 17);
drivers/ata/pata_ns87415.c
69
clocking |= (16 - clamp_val(t.recover, 1, 16)) << 4;
drivers/ata/pata_parport/bpck.c
212
int t, s;
drivers/ata/pata_parport/bpck.c
221
t2(2); t = r1()&0xf8;
drivers/ata/pata_parport/bpck.c
223
if ((f7) || (t != o1)) {
drivers/ata/pata_parport/bpck.c
227
if ((t == o1) && ((!f7) || (s == o1))) {
drivers/ata/pata_via.c
248
struct ata_timing t, p;
drivers/ata/pata_via.c
265
ata_timing_compute(adev, mode, &t, T, UT);
drivers/ata/pata_via.c
271
ata_timing_merge(&p, &t, &t, ATA_TIMING_8BIT);
drivers/ata/pata_via.c
282
setup |= (clamp_val(t.setup, 1, 4) - 1) << shift;
drivers/ata/pata_via.c
288
((clamp_val(t.act8b, 1, 16) - 1) << 4) | (clamp_val(t.rec8b, 1, 16) - 1));
drivers/ata/pata_via.c
290
((clamp_val(t.active, 1, 16) - 1) << 4) | (clamp_val(t.recover, 1, 16) - 1));
drivers/ata/pata_via.c
296
ut = t.udma ? (0xe0 | (clamp_val(t.udma, 2, 5) - 2)) : 0x03;
drivers/ata/pata_via.c
299
ut = t.udma ? (0xe8 | (clamp_val(t.udma, 2, 9) - 2)) : 0x0f;
drivers/ata/pata_via.c
302
ut = t.udma ? (0xe0 | (clamp_val(t.udma, 2, 9) - 2)) : 0x07;
drivers/ata/pata_via.c
305
ut = t.udma ? (0xe0 | (clamp_val(t.udma, 2, 9) - 2)) : 0x07;
drivers/ata/pata_via.c
318
if (t.udma) {
drivers/ata/sata_mv.c
3233
u32 t;
drivers/ata/sata_mv.c
3238
t = readl(reg);
drivers/ata/sata_mv.c
3239
writel(t | STOP_PCI_MASTER, reg);
drivers/ata/sata_mv.c
3243
t = readl(reg);
drivers/ata/sata_mv.c
3244
if (PCI_MASTER_EMPTY & t)
drivers/ata/sata_mv.c
3247
if (!(PCI_MASTER_EMPTY & t)) {
drivers/ata/sata_mv.c
3256
writel(t | GLOB_SFT_RST, reg);
drivers/ata/sata_mv.c
3257
t = readl(reg);
drivers/ata/sata_mv.c
3259
} while (!(GLOB_SFT_RST & t) && (i-- > 0));
drivers/ata/sata_mv.c
3261
if (!(GLOB_SFT_RST & t)) {
drivers/ata/sata_mv.c
3270
writel(t & ~(GLOB_SFT_RST | STOP_PCI_MASTER), reg);
drivers/ata/sata_mv.c
3271
t = readl(reg);
drivers/ata/sata_mv.c
3273
} while ((GLOB_SFT_RST & t) && (i-- > 0));
drivers/ata/sata_mv.c
3275
if (GLOB_SFT_RST & t) {
drivers/atm/idt77252.c
1534
tst_timer(struct timer_list *t)
drivers/atm/idt77252.c
1536
struct idt77252_dev *card = timer_container_of(card, t, tst_timer);
drivers/atm/idt77252.c
2074
idt77252_est_timer(struct timer_list *t)
drivers/atm/idt77252.c
2076
struct rate_estimator *est = timer_container_of(est, t, timer);
drivers/atm/iphase.c
703
u32 t;
drivers/atm/iphase.c
723
t = readl(iadev->reg+IPHASE5575_EEPROM_ACCESS);
drivers/atm/iphase.c
724
while (!(t & NVDO))
drivers/atm/iphase.c
725
t = readl(iadev->reg+IPHASE5575_EEPROM_ACCESS);
drivers/atm/iphase.c
740
u32 t;
drivers/atm/iphase.c
753
NVRAM_CLKIN(t);
drivers/atm/iphase.c
754
val |= (t << i);
drivers/atm/iphase.h
1373
u32 t; \
drivers/atm/iphase.h
1374
t = readl(iadev->reg+IPHASE5575_EEPROM_ACCESS); \
drivers/atm/iphase.h
1375
t &= (val); \
drivers/atm/iphase.h
1376
writel(t, iadev->reg+IPHASE5575_EEPROM_ACCESS); \
drivers/atm/iphase.h
1387
u32 t; \
drivers/atm/iphase.h
1388
t = readl(iadev->reg+IPHASE5575_EEPROM_ACCESS); \
drivers/atm/iphase.h
1389
t |= (val); \
drivers/atm/iphase.h
1390
writel(t, iadev->reg+IPHASE5575_EEPROM_ACCESS); \
drivers/atm/lanai.c
1759
static void lanai_timed_poll(struct timer_list *t)
drivers/atm/lanai.c
1761
struct lanai_dev *lanai = timer_container_of(lanai, t, timer);
drivers/atm/lanai.c
477
u32 t;
drivers/atm/lanai.c
478
t = readl(reg_addr(lanai, reg));
drivers/atm/lanai.c
480
(int) reg, t);
drivers/atm/lanai.c
481
return t;
drivers/auxdisplay/line-display.c
125
static void linedisp_scroll(struct timer_list *t)
drivers/auxdisplay/line-display.c
127
struct linedisp *linedisp = timer_container_of(linedisp, t, timer);
drivers/base/arch_topology.c
510
struct device_node *t __free(device_node) =
drivers/base/arch_topology.c
513
if (!t)
drivers/base/arch_topology.c
517
cpu = get_cpu_for_node(t);
drivers/base/arch_topology.c
524
pr_err("%pOF: Can't get CPU for thread\n", t);
drivers/base/power/main.c
537
static void dpm_watchdog_handler(struct timer_list *t)
drivers/base/power/main.c
539
struct dpm_watchdog *wd = timer_container_of(wd, t, timer);
drivers/base/power/wakeup.c
61
static void pm_wakeup_timer_fn(struct timer_list *t);
drivers/base/power/wakeup.c
756
static void pm_wakeup_timer_fn(struct timer_list *t)
drivers/base/power/wakeup.c
758
struct wakeup_source *ws = timer_container_of(ws, t, timer);
drivers/base/regmap/regmap-irq.c
254
const struct regmap_irq_type *t = &irq_data->type;
drivers/base/regmap/regmap-irq.c
256
if ((t->types_supported & type) != type)
drivers/base/regmap/regmap-irq.c
259
reg = t->type_reg_offset / map->reg_stride;
drivers/base/regmap/regmap-irq.c
591
const struct regmap_irq_type *t = &irq_data->type;
drivers/base/regmap/regmap-irq.c
593
if (t->type_reg_mask)
drivers/base/regmap/regmap-irq.c
594
buf[0][idx] &= ~t->type_reg_mask;
drivers/base/regmap/regmap-irq.c
596
buf[0][idx] &= ~(t->type_falling_val |
drivers/base/regmap/regmap-irq.c
597
t->type_rising_val |
drivers/base/regmap/regmap-irq.c
598
t->type_level_low_val |
drivers/base/regmap/regmap-irq.c
599
t->type_level_high_val);
drivers/base/regmap/regmap-irq.c
603
buf[0][idx] |= t->type_falling_val;
drivers/base/regmap/regmap-irq.c
607
buf[0][idx] |= t->type_rising_val;
drivers/base/regmap/regmap-irq.c
611
buf[0][idx] |= (t->type_falling_val |
drivers/base/regmap/regmap-irq.c
612
t->type_rising_val);
drivers/base/regmap/regmap-irq.c
616
buf[0][idx] |= t->type_level_high_val;
drivers/base/regmap/regmap-irq.c
620
buf[0][idx] |= t->type_level_low_val;
drivers/base/regmap/regmap-spi.c
18
struct spi_transfer t[2];
drivers/base/regmap/regmap-spi.c
43
struct spi_transfer t[2] = { { .tx_buf = reg, .len = reg_len, },
drivers/base/regmap/regmap-spi.c
47
spi_message_add_tail(&t[0], &m);
drivers/base/regmap/regmap-spi.c
48
spi_message_add_tail(&t[1], &m);
drivers/base/regmap/regmap-spi.c
64
async->t[0].tx_buf = reg;
drivers/base/regmap/regmap-spi.c
65
async->t[0].len = reg_len;
drivers/base/regmap/regmap-spi.c
66
async->t[1].tx_buf = val;
drivers/base/regmap/regmap-spi.c
67
async->t[1].len = val_len;
drivers/base/regmap/regmap-spi.c
70
spi_message_add_tail(&async->t[0], &async->m);
drivers/base/regmap/regmap-spi.c
72
spi_message_add_tail(&async->t[1], &async->m);
drivers/bcma/sprom.c
124
return t[crc ^ data];
drivers/bcma/sprom.c
90
static const u8 t[] = {
drivers/block/aoe/aoe.h
125
struct aoetgt *t; /* parent target I belong to */
drivers/block/aoe/aoe.h
223
void aoecmd_wreset(struct aoetgt *t);
drivers/block/aoe/aoeblk.c
114
struct aoetgt **t, **te;
drivers/block/aoe/aoeblk.c
129
t = d->targets;
drivers/block/aoe/aoeblk.c
130
te = t + d->ntargets;
drivers/block/aoe/aoeblk.c
131
for (; t < te && *t; t++) {
drivers/block/aoe/aoeblk.c
133
seq_printf(s, "falloc: %ld\n", (*t)->falloc);
drivers/block/aoe/aoeblk.c
135
list_empty(&(*t)->ffree) ? NULL : (*t)->ffree.next);
drivers/block/aoe/aoeblk.c
136
seq_printf(s, "%pm:%d:%d:%d\n", (*t)->addr, (*t)->nout,
drivers/block/aoe/aoeblk.c
137
(*t)->maxout, (*t)->nframes);
drivers/block/aoe/aoeblk.c
138
seq_printf(s, "\tssthresh:%d\n", (*t)->ssthresh);
drivers/block/aoe/aoeblk.c
139
seq_printf(s, "\ttaint:%d\n", (*t)->taint);
drivers/block/aoe/aoeblk.c
140
seq_printf(s, "\tr:%d\n", (*t)->rpkts);
drivers/block/aoe/aoeblk.c
141
seq_printf(s, "\tw:%d\n", (*t)->wpkts);
drivers/block/aoe/aoeblk.c
142
ifp = (*t)->ifs;
drivers/block/aoe/aoeblk.c
143
ife = ifp + ARRAY_SIZE((*t)->ifs);
drivers/block/aoe/aoeblk.c
50
struct aoetgt *t = d->targets[0];
drivers/block/aoe/aoeblk.c
52
if (t == NULL)
drivers/block/aoe/aoeblk.c
54
return sysfs_emit(page, "%pm\n", t->addr);
drivers/block/aoe/aoeblk.c
62
struct aoetgt **t, **te;
drivers/block/aoe/aoeblk.c
69
t = d->targets;
drivers/block/aoe/aoeblk.c
70
te = t + d->ntargets;
drivers/block/aoe/aoeblk.c
71
for (; t < te && *t; t++) {
drivers/block/aoe/aoeblk.c
72
ifp = (*t)->ifs;
drivers/block/aoe/aoecmd.c
1001
if (!t || t->maxout >= t->nframes)
drivers/block/aoe/aoecmd.c
1003
if (t->maxout < t->ssthresh)
drivers/block/aoe/aoecmd.c
1004
t->maxout += 1;
drivers/block/aoe/aoecmd.c
1005
else if (t->nout == t->maxout && t->next_cwnd-- == 0) {
drivers/block/aoe/aoecmd.c
1006
t->maxout += 1;
drivers/block/aoe/aoecmd.c
1007
t->next_cwnd = t->maxout;
drivers/block/aoe/aoecmd.c
1014
struct aoetgt **t, **e;
drivers/block/aoe/aoecmd.c
1016
t = d->targets;
drivers/block/aoe/aoecmd.c
1017
e = t + d->ntargets;
drivers/block/aoe/aoecmd.c
1018
for (; t < e && *t; t++)
drivers/block/aoe/aoecmd.c
1019
if (memcmp((*t)->addr, addr, sizeof((*t)->addr)) == 0)
drivers/block/aoe/aoecmd.c
1020
return *t;
drivers/block/aoe/aoecmd.c
1085
struct aoetgt *t;
drivers/block/aoe/aoecmd.c
1094
t = f->t;
drivers/block/aoe/aoecmd.c
1095
d = t->d;
drivers/block/aoe/aoecmd.c
1144
ifp = getif(t, skb->dev);
drivers/block/aoe/aoecmd.c
1160
ataid_complete(d, t, skb->data);
drivers/block/aoe/aoecmd.c
1171
if (t->taint > 0
drivers/block/aoe/aoecmd.c
1172
&& --t->taint > 0
drivers/block/aoe/aoecmd.c
1173
&& t->nout_probes == 0) {
drivers/block/aoe/aoecmd.c
1176
probe(t);
drivers/block/aoe/aoecmd.c
1177
t->nout_probes++;
drivers/block/aoe/aoecmd.c
1214
actual_id = f->t->d->aoeminor % ncpus;
drivers/block/aoe/aoecmd.c
128
aoehdr_atainit(struct aoedev *d, struct aoetgt *t, struct aoe_hdr *h)
drivers/block/aoe/aoecmd.c
1287
id = f->t->d->aoeminor % ncpus;
drivers/block/aoe/aoecmd.c
132
memcpy(h->src, t->ifp->nd->dev_addr, sizeof h->src);
drivers/block/aoe/aoecmd.c
133
memcpy(h->dst, t->addr, sizeof h->dst);
drivers/block/aoe/aoecmd.c
1330
calc_rttavg(d, f->t, tsince_hr(f));
drivers/block/aoe/aoecmd.c
1331
f->t->nout--;
drivers/block/aoe/aoecmd.c
1333
f->t->nout_probes--;
drivers/block/aoe/aoecmd.c
1385
struct aoetgt *t;
drivers/block/aoe/aoecmd.c
1391
t = *d->tgt;
drivers/block/aoe/aoecmd.c
1399
f->tag = aoehdr_atainit(d, t, h);
drivers/block/aoe/aoecmd.c
1401
t->nout++;
drivers/block/aoe/aoecmd.c
1410
dev_hold(t->ifp->nd);
drivers/block/aoe/aoecmd.c
1411
skb->dev = t->ifp->nd;
drivers/block/aoe/aoecmd.c
1421
dev_put(t->ifp->nd);
drivers/block/aoe/aoecmd.c
1449
struct aoetgt *t, **tt, **te;
drivers/block/aoe/aoecmd.c
1461
t = kzalloc_obj(*t, GFP_ATOMIC);
drivers/block/aoe/aoecmd.c
1462
if (!t)
drivers/block/aoe/aoecmd.c
1464
t->nframes = nframes;
drivers/block/aoe/aoecmd.c
1465
t->d = d;
drivers/block/aoe/aoecmd.c
1466
memcpy(t->addr, addr, sizeof t->addr);
drivers/block/aoe/aoecmd.c
1467
t->ifp = t->ifs;
drivers/block/aoe/aoecmd.c
1468
aoecmd_wreset(t);
drivers/block/aoe/aoecmd.c
1469
t->maxout = t->nframes / 2;
drivers/block/aoe/aoecmd.c
1470
INIT_LIST_HEAD(&t->ffree);
drivers/block/aoe/aoecmd.c
1471
return *tt = t;
drivers/block/aoe/aoecmd.c
1481
struct aoetgt **t, **e;
drivers/block/aoe/aoecmd.c
1484
t = d->targets;
drivers/block/aoe/aoecmd.c
1485
e = t + d->ntargets;
drivers/block/aoe/aoecmd.c
1486
for (; t < e && *t; t++)
drivers/block/aoe/aoecmd.c
1487
if (bcnt == 0 || bcnt > (*t)->minbcnt)
drivers/block/aoe/aoecmd.c
1488
bcnt = (*t)->minbcnt;
drivers/block/aoe/aoecmd.c
1497
setifbcnt(struct aoetgt *t, struct net_device *nd, int bcnt)
drivers/block/aoe/aoecmd.c
1503
d = t->d;
drivers/block/aoe/aoecmd.c
1505
p = t->ifs;
drivers/block/aoe/aoecmd.c
1525
t->minbcnt = minbcnt;
drivers/block/aoe/aoecmd.c
1535
struct aoetgt *t;
drivers/block/aoe/aoecmd.c
156
ifrotate(struct aoetgt *t)
drivers/block/aoe/aoecmd.c
1578
t = gettgt(d, h->src);
drivers/block/aoe/aoecmd.c
1579
if (t) {
drivers/block/aoe/aoecmd.c
1580
t->nframes = n;
drivers/block/aoe/aoecmd.c
1581
if (n < t->maxout)
drivers/block/aoe/aoecmd.c
1582
aoecmd_wreset(t);
drivers/block/aoe/aoecmd.c
1584
t = addtgt(d, h->src, n);
drivers/block/aoe/aoecmd.c
1585
if (!t)
drivers/block/aoe/aoecmd.c
1594
setifbcnt(t, skb->dev, n);
drivers/block/aoe/aoecmd.c
160
ifp = t->ifp;
drivers/block/aoe/aoecmd.c
1612
aoecmd_wreset(struct aoetgt *t)
drivers/block/aoe/aoecmd.c
1614
t->maxout = 1;
drivers/block/aoe/aoecmd.c
1615
t->ssthresh = t->nframes / 2;
drivers/block/aoe/aoecmd.c
1616
t->next_cwnd = t->nframes;
drivers/block/aoe/aoecmd.c
162
if (ifp >= &t->ifs[NAOEIFS] || ifp->nd == NULL)
drivers/block/aoe/aoecmd.c
1622
struct aoetgt **t, **te;
drivers/block/aoe/aoecmd.c
1628
t = d->targets;
drivers/block/aoe/aoecmd.c
1629
te = t + d->ntargets;
drivers/block/aoe/aoecmd.c
163
ifp = t->ifs;
drivers/block/aoe/aoecmd.c
1630
for (; t < te && *t; t++)
drivers/block/aoe/aoecmd.c
1631
aoecmd_wreset(*t);
drivers/block/aoe/aoecmd.c
166
return t->ifp = ifp;
drivers/block/aoe/aoecmd.c
1673
d = f->t->d;
drivers/block/aoe/aoecmd.c
194
struct aoetgt *t;
drivers/block/aoe/aoecmd.c
196
t = f->t;
drivers/block/aoe/aoecmd.c
201
list_add(&f->head, &t->ffree);
drivers/block/aoe/aoecmd.c
205
newtframe(struct aoedev *d, struct aoetgt *t)
drivers/block/aoe/aoecmd.c
211
if (list_empty(&t->ffree)) {
drivers/block/aoe/aoecmd.c
212
if (t->falloc >= NSKBPOOLMAX*2)
drivers/block/aoe/aoecmd.c
217
t->falloc++;
drivers/block/aoe/aoecmd.c
218
f->t = t;
drivers/block/aoe/aoecmd.c
220
pos = t->ffree.next;
drivers/block/aoe/aoecmd.c
252
struct aoetgt *t, **tt;
drivers/block/aoe/aoecmd.c
266
t = *tt;
drivers/block/aoe/aoecmd.c
267
if (!t->taint) {
drivers/block/aoe/aoecmd.c
269
totout += t->nout;
drivers/block/aoe/aoecmd.c
271
if (t->nout < t->maxout
drivers/block/aoe/aoecmd.c
272
&& (use_tainted || !t->taint)
drivers/block/aoe/aoecmd.c
273
&& t->ifp->nd) {
drivers/block/aoe/aoecmd.c
274
f = newtframe(d, t);
drivers/block/aoe/aoecmd.c
276
ifrotate(t);
drivers/block/aoe/aoecmd.c
309
struct aoedev *d = f->t->d;
drivers/block/aoe/aoecmd.c
319
struct aoetgt *t;
drivers/block/aoe/aoecmd.c
334
t = f->t;
drivers/block/aoe/aoecmd.c
335
f->tag = aoehdr_atainit(t->d, t, h);
drivers/block/aoe/aoecmd.c
337
t->nout++;
drivers/block/aoe/aoecmd.c
344
if (t->d->flags & DEVFL_EXT) {
drivers/block/aoe/aoecmd.c
357
t->wpkts++;
drivers/block/aoe/aoecmd.c
359
t->rpkts++;
drivers/block/aoe/aoecmd.c
364
dev_hold(t->ifp->nd);
drivers/block/aoe/aoecmd.c
365
skb->dev = t->ifp->nd;
drivers/block/aoe/aoecmd.c
406
dev_put(f->t->ifp->nd);
drivers/block/aoe/aoecmd.c
459
struct aoetgt *t;
drivers/block/aoe/aoecmd.c
463
t = f->t;
drivers/block/aoe/aoecmd.c
466
if (ifrotate(t) == NULL) {
drivers/block/aoe/aoecmd.c
479
h->src, h->dst, t->nout);
drivers/block/aoe/aoecmd.c
486
memcpy(h->dst, t->addr, sizeof h->dst);
drivers/block/aoe/aoecmd.c
487
memcpy(h->src, t->ifp->nd->dev_addr, sizeof h->src);
drivers/block/aoe/aoecmd.c
489
dev_hold(t->ifp->nd);
drivers/block/aoe/aoecmd.c
490
skb->dev = t->ifp->nd;
drivers/block/aoe/aoecmd.c
493
dev_put(t->ifp->nd);
drivers/block/aoe/aoecmd.c
531
getif(struct aoetgt *t, struct net_device *nd)
drivers/block/aoe/aoecmd.c
535
p = t->ifs;
drivers/block/aoe/aoecmd.c
544
ejectif(struct aoetgt *t, struct aoeif *ifp)
drivers/block/aoe/aoecmd.c
551
e = t->ifs + NAOEIFS - 1;
drivers/block/aoe/aoecmd.c
564
nf = newframe(f->t->d);
drivers/block/aoe/aoecmd.c
567
if (nf->t == f->t) {
drivers/block/aoe/aoecmd.c
585
probe(struct aoetgt *t)
drivers/block/aoe/aoecmd.c
594
d = t->d;
drivers/block/aoe/aoecmd.c
595
f = newtframe(d, t);
drivers/block/aoe/aoecmd.c
599
t->addr,
drivers/block/aoe/aoecmd.c
605
ifrotate(t);
drivers/block/aoe/aoecmd.c
606
f->iter.bi_size = t->d->maxbcnt ? t->d->maxbcnt : DEFAULTBCNT;
drivers/block/aoe/aoecmd.c
627
dev_put(f->t->ifp->nd);
drivers/block/aoe/aoecmd.c
634
long t;
drivers/block/aoe/aoecmd.c
636
t = 2 * d->rttavg >> RTTSCALE;
drivers/block/aoe/aoecmd.c
637
t += 8 * d->rttdev >> RTTDSCALE;
drivers/block/aoe/aoecmd.c
638
if (t == 0)
drivers/block/aoe/aoecmd.c
639
t = 1;
drivers/block/aoe/aoecmd.c
641
return t;
drivers/block/aoe/aoecmd.c
647
struct aoetgt *t;
drivers/block/aoe/aoecmd.c
659
t = f->t;
drivers/block/aoe/aoecmd.c
660
if (t->taint) {
drivers/block/aoe/aoecmd.c
664
if (t->nout_probes == 0
drivers/block/aoe/aoecmd.c
666
probe(t);
drivers/block/aoe/aoecmd.c
667
t->nout_probes++;
drivers/block/aoe/aoecmd.c
673
t = f->t;
drivers/block/aoe/aoecmd.c
678
} else if (tsince_hr(f) < t->taint * rto(d)) {
drivers/block/aoe/aoecmd.c
687
f->t->d->flags |= DEVFL_KICKME;
drivers/block/aoe/aoecmd.c
690
if (t->nout >= t->maxout)
drivers/block/aoe/aoecmd.c
693
t->nout++;
drivers/block/aoe/aoecmd.c
695
t->nout_probes++;
drivers/block/aoe/aoecmd.c
707
scorn(struct aoetgt *t)
drivers/block/aoe/aoecmd.c
711
n = t->taint++;
drivers/block/aoe/aoecmd.c
712
t->taint += t->taint * 2;
drivers/block/aoe/aoecmd.c
713
if (n > t->taint)
drivers/block/aoe/aoecmd.c
714
t->taint = n;
drivers/block/aoe/aoecmd.c
715
if (t->taint > MAX_TAINT)
drivers/block/aoe/aoecmd.c
716
t->taint = MAX_TAINT;
drivers/block/aoe/aoecmd.c
737
struct aoetgt *t;
drivers/block/aoe/aoecmd.c
794
t = f->t;
drivers/block/aoe/aoecmd.c
799
scorn(t); /* avoid this target */
drivers/block/aoe/aoecmd.c
801
if (t->maxout != 1) {
drivers/block/aoe/aoecmd.c
802
t->ssthresh = t->maxout / 2;
drivers/block/aoe/aoecmd.c
803
t->maxout = 1;
drivers/block/aoe/aoecmd.c
807
t->nout_probes--;
drivers/block/aoe/aoecmd.c
809
ifp = getif(t, f->skb->dev);
drivers/block/aoe/aoecmd.c
810
if (ifp && ++ifp->lost > (t->nframes << 1)
drivers/block/aoe/aoecmd.c
811
&& (ifp != t->ifs || t->ifs[1].nd)) {
drivers/block/aoe/aoecmd.c
812
ejectif(t, ifp);
drivers/block/aoe/aoecmd.c
817
t->nout--;
drivers/block/aoe/aoecmd.c
930
ataid_complete(struct aoedev *d, struct aoetgt *t, unsigned char *id)
drivers/block/aoe/aoecmd.c
972
t->addr,
drivers/block/aoe/aoecmd.c
987
calc_rttavg(struct aoedev *d, struct aoetgt *t, int rtt)
drivers/block/aoe/aoedev.c
148
dummy_timer(struct timer_list *t)
drivers/block/aoe/aoedev.c
152
d = timer_container_of(d, t, timer);
drivers/block/aoe/aoedev.c
18
static void freetgt(struct aoedev *d, struct aoetgt *t);
drivers/block/aoe/aoedev.c
191
aoe_failbuf(f->t->d, f->buf);
drivers/block/aoe/aoedev.c
199
struct aoetgt *t, **tt, **te;
drivers/block/aoe/aoedev.c
222
for (; tt < te && (t = *tt); tt++) {
drivers/block/aoe/aoedev.c
223
aoecmd_wreset(t);
drivers/block/aoe/aoedev.c
224
t->nout = 0;
drivers/block/aoe/aoedev.c
274
struct aoetgt **t, **e;
drivers/block/aoe/aoedev.c
295
t = d->targets;
drivers/block/aoe/aoedev.c
296
e = t + d->ntargets;
drivers/block/aoe/aoedev.c
297
for (; t < e && *t; t++)
drivers/block/aoe/aoedev.c
298
freetgt(d, *t);
drivers/block/aoe/aoedev.c
510
freetgt(struct aoedev *d, struct aoetgt *t)
drivers/block/aoe/aoedev.c
516
for (ifp = t->ifs; ifp < &t->ifs[NAOEIFS]; ++ifp) {
drivers/block/aoe/aoedev.c
522
head = &t->ffree;
drivers/block/aoe/aoedev.c
529
kfree(t);
drivers/block/aoe/aoemain.c
21
static void discover_timer(struct timer_list *t)
drivers/block/aoe/aoemain.c
23
mod_timer(t, jiffies + HZ * 60); /* one minute */
drivers/block/drbd/drbd_actlog.c
304
unsigned int t = device->al_tr_number % (device->ldev->md.al_size_4k);
drivers/block/drbd/drbd_actlog.c
307
t = ((t % stripes) * stripe_size_4kB) + t/stripes;
drivers/block/drbd/drbd_actlog.c
310
t *= 8;
drivers/block/drbd/drbd_actlog.c
313
return device->ldev->md.md_offset + device->ldev->md.al_offset + t;
drivers/block/drbd/drbd_int.h
1457
extern void resync_timer_fn(struct timer_list *t);
drivers/block/drbd/drbd_int.h
1458
extern void start_resync_timer_fn(struct timer_list *t);
drivers/block/drbd/drbd_main.c
1688
struct p_trim *t = (struct p_trim*)p;
drivers/block/drbd/drbd_main.c
1689
t->size = cpu_to_be32(req->i.size);
drivers/block/drbd/drbd_main.c
1690
err = __send_command(peer_device->connection, device->vnr, sock, cmd, sizeof(*t), NULL, 0);
drivers/block/drbd/drbd_main.c
3564
static void md_sync_timer_fn(struct timer_list *t)
drivers/block/drbd/drbd_main.c
3566
struct drbd_device *device = timer_container_of(device, t,
drivers/block/drbd/drbd_main.c
55
static void md_sync_timer_fn(struct timer_list *t);
drivers/block/drbd/drbd_nl.c
1150
struct lru_cache *n, *t;
drivers/block/drbd/drbd_nl.c
1160
t = device->act_log;
drivers/block/drbd/drbd_nl.c
1169
if (t) {
drivers/block/drbd/drbd_nl.c
1170
for (i = 0; i < t->nr_elements; i++) {
drivers/block/drbd/drbd_nl.c
1171
e = lc_element_by_index(t, i);
drivers/block/drbd/drbd_nl.c
1186
lc_destroy(t);
drivers/block/drbd/drbd_receiver.c
243
struct drbd_peer_request *peer_req, *t;
drivers/block/drbd/drbd_receiver.c
250
list_for_each_entry_safe(peer_req, t, &work_list, w.list) {
drivers/block/drbd/drbd_receiver.c
263
struct drbd_peer_request *peer_req, *t;
drivers/block/drbd/drbd_receiver.c
274
list_for_each_entry_safe(peer_req, t, &work_list, w.list) {
drivers/block/drbd/drbd_receiver.c
338
long t;
drivers/block/drbd/drbd_receiver.c
340
t = rcu_dereference(connection->net_conf)->ping_timeo * HZ/10;
drivers/block/drbd/drbd_receiver.c
343
t = wait_event_timeout(connection->ping_wait, connection->cstate < C_WF_REPORT_PARAMS, t);
drivers/block/drbd/drbd_receiver.c
345
if (t)
drivers/block/drbd/drbd_receiver.c
5722
long t;
drivers/block/drbd/drbd_receiver.c
5727
t = ping_timeout ? nc->ping_timeo : nc->ping_int;
drivers/block/drbd/drbd_receiver.c
5730
t *= HZ;
drivers/block/drbd/drbd_receiver.c
5732
t /= 10;
drivers/block/drbd/drbd_receiver.c
5734
connection->meta.socket->sk->sk_rcvtimeo = t;
drivers/block/drbd/drbd_receiver.c
5812
long t;
drivers/block/drbd/drbd_receiver.c
5814
t = rcu_dereference(connection->net_conf)->ping_timeo * HZ/10;
drivers/block/drbd/drbd_receiver.c
5817
t = wait_event_timeout(connection->ping_wait,
drivers/block/drbd/drbd_receiver.c
5819
t);
drivers/block/drbd/drbd_receiver.c
5820
if (t)
drivers/block/drbd/drbd_req.c
1701
void request_timer_fn(struct timer_list *t)
drivers/block/drbd/drbd_req.c
1703
struct drbd_device *device = timer_container_of(device, t,
drivers/block/drbd/drbd_req.h
274
extern void request_timer_fn(struct timer_list *t);
drivers/block/drbd/drbd_vli.h
138
#define LEVEL(t,b,v) \
drivers/block/drbd/drbd_vli.h
141
*out = ((in & ((~0ULL) >> (64-t))) >> b) + adj; \
drivers/block/drbd/drbd_vli.h
142
return t; \
drivers/block/drbd/drbd_vli.h
144
adj += 1ULL << (t - b); \
drivers/block/drbd/drbd_vli.h
164
#define LEVEL(t,b,v) do { \
drivers/block/drbd/drbd_vli.h
165
max += 1ULL << (t - b); \
drivers/block/drbd/drbd_vli.h
169
return t; \
drivers/block/drbd/drbd_worker.c
1680
void start_resync_timer_fn(struct timer_list *t)
drivers/block/drbd/drbd_worker.c
1682
struct drbd_device *device = timer_container_of(device, t,
drivers/block/drbd/drbd_worker.c
443
void resync_timer_fn(struct timer_list *t)
drivers/block/drbd/drbd_worker.c
445
struct drbd_device *device = timer_container_of(device, t,
drivers/block/drbd/drbd_worker.c
934
const unsigned long t = device->rs_total;
drivers/block/drbd/drbd_worker.c
936
(t == 0) ? 0 :
drivers/block/drbd/drbd_worker.c
937
(t < 100000) ? ((s*100)/t) : (s/(t/100));
drivers/block/floppy.c
918
static void motor_off_callback(struct timer_list *t)
drivers/block/floppy.c
920
unsigned long nr = t - motor_off_timer;
drivers/block/swim3.c
239
static void scan_timeout(struct timer_list *t);
drivers/block/swim3.c
240
static void seek_timeout(struct timer_list *t);
drivers/block/swim3.c
241
static void settle_timeout(struct timer_list *t);
drivers/block/swim3.c
242
static void xfer_timeout(struct timer_list *t);
drivers/block/swim3.c
362
void (*proc)(struct timer_list *t))
drivers/block/swim3.c
556
static void scan_timeout(struct timer_list *t)
drivers/block/swim3.c
558
struct floppy_state *fs = timer_container_of(fs, t, timeout);
drivers/block/swim3.c
580
static void seek_timeout(struct timer_list *t)
drivers/block/swim3.c
582
struct floppy_state *fs = timer_container_of(fs, t, timeout);
drivers/block/swim3.c
599
static void settle_timeout(struct timer_list *t)
drivers/block/swim3.c
601
struct floppy_state *fs = timer_container_of(fs, t, timeout);
drivers/block/swim3.c
628
static void xfer_timeout(struct timer_list *t)
drivers/block/swim3.c
630
struct floppy_state *fs = timer_container_of(fs, t, timeout);
drivers/bluetooth/bluecard_cs.c
159
static void bluecard_activity_led_timeout(struct timer_list *t)
drivers/bluetooth/bluecard_cs.c
161
struct bluecard_info *info = timer_container_of(info, t, timer);
drivers/bluetooth/btintel.h
197
#define INTEL_CNVX_TOP_PACK_SWAB(t, s) __swab16(((__u16)(((t) << 4) | (s))))
drivers/bluetooth/btnxpuart.c
483
static void ps_timeout_func(struct timer_list *t)
drivers/bluetooth/btnxpuart.c
485
struct ps_data *data = timer_container_of(data, t, ps_timer);
drivers/bluetooth/hci_bcsp.c
692
static void bcsp_timed_event(struct timer_list *t)
drivers/bluetooth/hci_bcsp.c
694
struct bcsp_struct *bcsp = timer_container_of(bcsp, t, tbcsp);
drivers/bluetooth/hci_h5.c
151
static void h5_timed_event(struct timer_list *t)
drivers/bluetooth/hci_h5.c
155
struct h5 *h5 = timer_container_of(h5, t, timer);
drivers/bluetooth/hci_qca.c
477
static void hci_ibs_tx_idle_timeout(struct timer_list *t)
drivers/bluetooth/hci_qca.c
479
struct qca_data *qca = timer_container_of(qca, t, tx_idle_timer);
drivers/bluetooth/hci_qca.c
510
static void hci_ibs_wake_retrans_timeout(struct timer_list *t)
drivers/bluetooth/hci_qca.c
512
struct qca_data *qca = timer_container_of(qca, t, wake_retrans_timer);
drivers/bus/mhi/host/pci_generic.c
1256
static void health_check(struct timer_list *t)
drivers/bus/mhi/host/pci_generic.c
1258
struct mhi_pci_device *mhi_pdev = timer_container_of(mhi_pdev, t,
drivers/bus/moxtet.c
100
if (*t == mdev->id)
drivers/bus/moxtet.c
91
const enum turris_mox_module_id *t;
drivers/bus/moxtet.c
99
for (t = tdrv->id_table; *t; ++t)
drivers/char/dsp56k.c
62
long i, t, m; \
drivers/char/dsp56k.c
66
for (t = 0; t < timeout && !ENABLE; t++) \
drivers/char/dsp56k.c
79
int t; \
drivers/char/dsp56k.c
80
for(t = 0; t < n && !DSP56K_TRANSMIT; t++) \
drivers/char/dsp56k.c
89
int t; \
drivers/char/dsp56k.c
90
for(t = 0; t < n && !DSP56K_RECEIVE; t++) \
drivers/char/dtlk.c
512
unsigned char *t;
drivers/char/dtlk.c
531
t = buf;
drivers/char/dtlk.c
532
status.serial_number = t[0] + t[1] * 256; /* serial number is
drivers/char/dtlk.c
534
t += 2;
drivers/char/dtlk.c
537
while (*t != '\r') {
drivers/char/dtlk.c
538
status.rom_version[i] = *t;
drivers/char/dtlk.c
541
t++;
drivers/char/dtlk.c
544
t++;
drivers/char/dtlk.c
546
status.mode = *t++;
drivers/char/dtlk.c
547
status.punc_level = *t++;
drivers/char/dtlk.c
548
status.formant_freq = *t++;
drivers/char/dtlk.c
549
status.pitch = *t++;
drivers/char/dtlk.c
550
status.speed = *t++;
drivers/char/dtlk.c
551
status.volume = *t++;
drivers/char/dtlk.c
552
status.tone = *t++;
drivers/char/dtlk.c
553
status.expression = *t++;
drivers/char/dtlk.c
554
status.ext_dict_loaded = *t++;
drivers/char/dtlk.c
555
status.ext_dict_status = *t++;
drivers/char/dtlk.c
556
status.free_ram = *t++;
drivers/char/dtlk.c
557
status.articulation = *t++;
drivers/char/dtlk.c
558
status.reverb = *t++;
drivers/char/dtlk.c
559
status.eob = *t++;
drivers/char/hpet.c
124
unsigned long t, mc, base, k;
drivers/char/hpet.c
128
t = devp->hd_ireqfreq;
drivers/char/hpet.c
146
base = mc % t;
drivers/char/hpet.c
147
k = (mc - base + hpetp->hp_delta) / t;
drivers/char/hpet.c
148
write_counter(t * (k + 1) + base,
drivers/char/hpet.c
437
unsigned long g, v, t, m;
drivers/char/hpet.c
497
t = devp->hd_ireqfreq;
drivers/char/hpet.c
519
write_counter(t + m + hpetp->hp_delta, &timer->hpet_compare);
drivers/char/hpet.c
524
write_counter(t, &timer->hpet_compare);
drivers/char/hpet.c
528
write_counter(t + m + hpetp->hp_delta, &timer->hpet_compare);
drivers/char/hpet.c
749
unsigned long t, m, count, i, flags, start;
drivers/char/hpet.c
764
t = read_counter(&timer->hpet_compare);
drivers/char/hpet.c
775
write_counter(t + m + hpetp->hp_delta, &timer->hpet_compare);
drivers/char/hw_random/pic32-rng.c
54
u32 t;
drivers/char/hw_random/pic32-rng.c
58
t = readl(priv->base + RNGRCNT) & RCNT_MASK;
drivers/char/hw_random/pic32-rng.c
59
if (t == 64) {
drivers/char/hw_random/xgene-rng.c
89
static void xgene_rng_expired_timer(struct timer_list *t)
drivers/char/hw_random/xgene-rng.c
91
struct xgene_rng_dev *ctx = timer_container_of(ctx, t, failure_timer);
drivers/char/ipmi/bt-bmc.c
348
static void poll_timer(struct timer_list *t)
drivers/char/ipmi/bt-bmc.c
350
struct bt_bmc *bt_bmc = timer_container_of(bt_bmc, t, poll_timer);
drivers/char/ipmi/ipmi_ipmb.c
395
struct task_struct *t = iidev->thread;
drivers/char/ipmi/ipmi_ipmb.c
401
kthread_stop(t);
drivers/char/ipmi/ipmi_msghandler.c
45
static void smi_work(struct work_struct *t);
drivers/char/ipmi/ipmi_msghandler.c
4855
static void smi_work(struct work_struct *t)
drivers/char/ipmi/ipmi_msghandler.c
4858
struct ipmi_smi *intf = from_work(intf, t, smi_work);
drivers/char/ipmi/ipmi_si_hardcode.c
65
int t;
drivers/char/ipmi/ipmi_si_hardcode.c
73
t = match_string(si_to_str, -1, si_type_str);
drivers/char/ipmi/ipmi_si_hardcode.c
74
if (t < 0) {
drivers/char/ipmi/ipmi_si_hardcode.c
79
p.type = t;
drivers/char/ipmi/ipmi_si_intf.c
1102
static void smi_timeout(struct timer_list *t)
drivers/char/ipmi/ipmi_si_intf.c
1104
struct smi_info *smi_info = timer_container_of(smi_info, t,
drivers/char/ipmi/ipmi_si_intf.c
275
struct timespec64 t;
drivers/char/ipmi/ipmi_si_intf.c
277
ktime_get_ts64(&t);
drivers/char/ipmi/ipmi_si_intf.c
278
dev_dbg(smi_info->io.dev, "**%s: %ptSp\n", msg, &t);
drivers/char/ipmi/ipmi_ssif.c
1083
struct timespec64 t;
drivers/char/ipmi/ipmi_ssif.c
1085
ktime_get_real_ts64(&t);
drivers/char/ipmi/ipmi_ssif.c
1087
msg->data[0], msg->data[1], &t);
drivers/char/ipmi/ipmi_ssif.c
540
static void retry_timeout(struct timer_list *t)
drivers/char/ipmi/ipmi_ssif.c
542
struct ssif_info *ssif_info = timer_container_of(ssif_info, t,
drivers/char/ipmi/ipmi_ssif.c
565
static void watch_timeout(struct timer_list *t)
drivers/char/ipmi/ipmi_ssif.c
567
struct ssif_info *ssif_info = timer_container_of(ssif_info, t,
drivers/char/ipmi/ssif_bmc.c
298
static void response_timeout(struct timer_list *t)
drivers/char/ipmi/ssif_bmc.c
300
struct ssif_bmc_ctx *ssif_bmc = timer_container_of(ssif_bmc, t,
drivers/char/tlclk.c
188
static void switchover_timeout(struct timer_list *t);
drivers/char/tpm/tpm-dev-common.c
92
static void user_reader_timeout(struct timer_list *t)
drivers/char/tpm/tpm-dev-common.c
94
struct file_priv *priv = timer_container_of(priv, t, user_read_timer);
drivers/char/xillybus/xillybus_core.c
1826
long t;
drivers/char/xillybus/xillybus_core.c
1833
t = wait_event_interruptible_timeout(endpoint->ep_wait,
drivers/char/xillybus/xillybus_core.c
1836
if (t <= 0) {
drivers/char/xillybus/xillybus_core.c
1847
long t;
drivers/char/xillybus/xillybus_core.c
1896
t = wait_event_interruptible_timeout(endpoint->ep_wait,
drivers/char/xillybus/xillybus_core.c
1899
if (t <= 0) {
drivers/char/xillybus/xillybus_core.c
636
long t;
drivers/char/xillybus/xillybus_core.c
646
t = wait_event_interruptible_timeout(channel->wr_wait,
drivers/char/xillybus/xillybus_core.c
650
if (t <= 0) {
drivers/clk/at91/sam9x7.c
202
u8 t;
drivers/clk/at91/sam9x7.c
212
.t = PLL_TYPE_FRAC,
drivers/clk/at91/sam9x7.c
225
.t = PLL_TYPE_DIV,
drivers/clk/at91/sam9x7.c
238
.t = PLL_TYPE_FRAC,
drivers/clk/at91/sam9x7.c
247
.t = PLL_TYPE_DIV,
drivers/clk/at91/sam9x7.c
262
.t = PLL_TYPE_FRAC,
drivers/clk/at91/sam9x7.c
273
.t = PLL_TYPE_DIV,
drivers/clk/at91/sam9x7.c
284
.t = PLL_TYPE_DIV,
drivers/clk/at91/sam9x7.c
295
.t = PLL_TYPE_FRAC,
drivers/clk/at91/sam9x7.c
306
.t = PLL_TYPE_DIV,
drivers/clk/at91/sam9x7.c
322
.t = PLL_TYPE_DIV,
drivers/clk/at91/sam9x7.c
788
switch (sam9x7_plls[i][j].t) {
drivers/clk/at91/sama7d65.c
1165
switch (sama7d65_plls[i][j].t) {
drivers/clk/at91/sama7d65.c
202
u8 t;
drivers/clk/at91/sama7d65.c
212
.t = PLL_TYPE_FRAC,
drivers/clk/at91/sama7d65.c
225
.t = PLL_TYPE_DIV,
drivers/clk/at91/sama7d65.c
243
.t = PLL_TYPE_FRAC,
drivers/clk/at91/sama7d65.c
257
.t = PLL_TYPE_DIV,
drivers/clk/at91/sama7d65.c
273
.t = PLL_TYPE_FRAC,
drivers/clk/at91/sama7d65.c
286
.t = PLL_TYPE_DIV,
drivers/clk/at91/sama7d65.c
298
.t = PLL_TYPE_FRAC,
drivers/clk/at91/sama7d65.c
307
.t = PLL_TYPE_DIV,
drivers/clk/at91/sama7d65.c
319
.t = PLL_TYPE_FRAC,
drivers/clk/at91/sama7d65.c
328
.t = PLL_TYPE_DIV,
drivers/clk/at91/sama7d65.c
341
.t = PLL_TYPE_FRAC,
drivers/clk/at91/sama7d65.c
350
.t = PLL_TYPE_DIV,
drivers/clk/at91/sama7d65.c
361
.t = PLL_TYPE_DIV,
drivers/clk/at91/sama7d65.c
374
.t = PLL_TYPE_FRAC,
drivers/clk/at91/sama7d65.c
383
.t = PLL_TYPE_DIV,
drivers/clk/at91/sama7d65.c
396
.t = PLL_TYPE_FRAC,
drivers/clk/at91/sama7d65.c
405
.t = PLL_TYPE_DIV,
drivers/clk/at91/sama7d65.c
418
.t = PLL_TYPE_FRAC,
drivers/clk/at91/sama7d65.c
427
.t = PLL_TYPE_DIV,
drivers/clk/at91/sama7g5.c
1042
switch (sama7g5_plls[i][j].t) {
drivers/clk/at91/sama7g5.c
160
u8 t;
drivers/clk/at91/sama7g5.c
170
.t = PLL_TYPE_FRAC,
drivers/clk/at91/sama7g5.c
183
.t = PLL_TYPE_DIV,
drivers/clk/at91/sama7g5.c
201
.t = PLL_TYPE_FRAC,
drivers/clk/at91/sama7g5.c
215
.t = PLL_TYPE_DIV,
drivers/clk/at91/sama7g5.c
231
.t = PLL_TYPE_FRAC,
drivers/clk/at91/sama7g5.c
244
.t = PLL_TYPE_DIV,
drivers/clk/at91/sama7g5.c
256
.t = PLL_TYPE_FRAC,
drivers/clk/at91/sama7g5.c
265
.t = PLL_TYPE_DIV,
drivers/clk/at91/sama7g5.c
277
.t = PLL_TYPE_FRAC,
drivers/clk/at91/sama7g5.c
285
.t = PLL_TYPE_DIV,
drivers/clk/at91/sama7g5.c
297
.t = PLL_TYPE_FRAC,
drivers/clk/at91/sama7g5.c
306
.t = PLL_TYPE_DIV,
drivers/clk/at91/sama7g5.c
317
.t = PLL_TYPE_DIV,
drivers/clk/at91/sama7g5.c
330
.t = PLL_TYPE_FRAC,
drivers/clk/at91/sama7g5.c
339
.t = PLL_TYPE_DIV,
drivers/clk/clk.c
4588
struct hlist_node *t;
drivers/clk/clk.c
4591
hlist_for_each_entry_safe(child, t, &clk->core->children,
drivers/clk/clk_test.c
2233
const struct clk_leaf_mux_set_rate_parent_determine_rate_test_case *t, char *desc)
drivers/clk/clk_test.c
2235
strcpy(desc, t->desc);
drivers/clk/clk_test.c
2645
const struct clk_register_clk_parent_data_test_case *t, char *desc)
drivers/clk/clk_test.c
2647
strcpy(desc, t->desc);
drivers/clk/imx/clk-pll14xx.c
132
unsigned long prate, struct imx_pll14xx_rate_table *t)
drivers/clk/imx/clk-pll14xx.c
156
t->rate = tt->rate;
drivers/clk/imx/clk-pll14xx.c
157
t->mdiv = tt->mdiv;
drivers/clk/imx/clk-pll14xx.c
158
t->pdiv = tt->pdiv;
drivers/clk/imx/clk-pll14xx.c
159
t->sdiv = tt->sdiv;
drivers/clk/imx/clk-pll14xx.c
160
t->kdiv = tt->kdiv;
drivers/clk/imx/clk-pll14xx.c
180
t->rate = (unsigned int)fout;
drivers/clk/imx/clk-pll14xx.c
181
t->mdiv = mdiv;
drivers/clk/imx/clk-pll14xx.c
182
t->pdiv = pdiv;
drivers/clk/imx/clk-pll14xx.c
183
t->sdiv = sdiv;
drivers/clk/imx/clk-pll14xx.c
184
t->kdiv = kdiv;
drivers/clk/imx/clk-pll14xx.c
202
t->rate = (unsigned int)fout;
drivers/clk/imx/clk-pll14xx.c
203
t->mdiv = mdiv;
drivers/clk/imx/clk-pll14xx.c
204
t->pdiv = pdiv;
drivers/clk/imx/clk-pll14xx.c
205
t->sdiv = sdiv;
drivers/clk/imx/clk-pll14xx.c
206
t->kdiv = kdiv;
drivers/clk/imx/clk-pll14xx.c
215
clk_hw_get_name(&pll->hw), prate, rate, t->rate, t->pdiv, t->sdiv,
drivers/clk/imx/clk-pll14xx.c
216
t->mdiv, t->kdiv);
drivers/clk/imx/clk-pll14xx.c
244
struct imx_pll14xx_rate_table t;
drivers/clk/imx/clk-pll14xx.c
246
imx_pll14xx_calc_settings(pll, req->rate, req->best_parent_rate, &t);
drivers/clk/imx/clk-pll14xx.c
248
req->rate = t.rate;
drivers/clk/pxa/clk-pxa25x.c
176
unsigned int t;
drivers/clk/pxa/clk-pxa25x.c
179
t = clkcfg & (1 << 0);
drivers/clk/pxa/clk-pxa25x.c
180
if (t)
drivers/clk/pxa/clk-pxa25x.c
219
unsigned int l, m, n2, t;
drivers/clk/pxa/clk-pxa25x.c
222
t = clkcfg & (1 << 0);
drivers/clk/pxa/clk-pxa27x.c
205
unsigned int t, ht;
drivers/clk/pxa/clk-pxa27x.c
210
t = clkcfg & (1 << 0);
drivers/clk/pxa/clk-pxa27x.c
302
unsigned int t, ht, osc_forced;
drivers/clk/pxa/clk-pxa27x.c
310
t = clkcfg & (1 << 0);
drivers/clk/pxa/clk-pxa27x.c
313
if (ht || t)
drivers/clk/pxa/clk-pxa3xx.c
316
unsigned int t;
drivers/clk/pxa/clk-pxa3xx.c
323
t = xclkcfg & 0x1;
drivers/clk/pxa/clk-pxa3xx.c
325
if (t)
drivers/clk/pxa/clk-pxa3xx.c
337
unsigned int t, xclkcfg;
drivers/clk/pxa/clk-pxa3xx.c
341
t = xclkcfg & 0x1;
drivers/clk/pxa/clk-pxa3xx.c
343
return t ? (parent_rate / xn) * 2 : parent_rate;
drivers/clk/pxa/clk-pxa3xx.c
354
unsigned int t, xclkcfg;
drivers/clk/pxa/clk-pxa3xx.c
358
t = xclkcfg & 0x1;
drivers/clk/pxa/clk-pxa3xx.c
361
return t ? parent_rate * xl * xn : parent_rate * xl;
drivers/clk/ralink/clk-mtmips.c
410
u32 t;
drivers/clk/ralink/clk-mtmips.c
412
regmap_read(sysc, SYSC_REG_SYSTEM_CONFIG, &t);
drivers/clk/ralink/clk-mtmips.c
413
t = (t >> RT5350_SYSCFG0_CPUCLK_SHIFT) & RT5350_SYSCFG0_CPUCLK_MASK;
drivers/clk/ralink/clk-mtmips.c
415
switch (t) {
drivers/clk/ralink/clk-mtmips.c
441
u32 t;
drivers/clk/ralink/clk-mtmips.c
443
regmap_read(sysc, SYSC_REG_SYSTEM_CONFIG, &t);
drivers/clk/ralink/clk-mtmips.c
444
t = (t >> RT3352_SYSCFG0_CPUCLK_SHIFT) & RT3352_SYSCFG0_CPUCLK_MASK;
drivers/clk/ralink/clk-mtmips.c
446
switch (t) {
drivers/clk/ralink/clk-mtmips.c
461
u32 t;
drivers/clk/ralink/clk-mtmips.c
463
regmap_read(sysc, SYSC_REG_SYSTEM_CONFIG, &t);
drivers/clk/ralink/clk-mtmips.c
464
t = (t >> RT305X_SYSCFG_CPUCLK_SHIFT) & RT305X_SYSCFG_CPUCLK_MASK;
drivers/clk/ralink/clk-mtmips.c
466
switch (t) {
drivers/clk/ralink/clk-mtmips.c
481
u32 t;
drivers/clk/ralink/clk-mtmips.c
483
regmap_read(sysc, SYSC_REG_SYSTEM_CONFIG, &t);
drivers/clk/ralink/clk-mtmips.c
484
t = (t >> RT3883_SYSCFG0_CPUCLK_SHIFT) & RT3883_SYSCFG0_CPUCLK_MASK;
drivers/clk/ralink/clk-mtmips.c
486
switch (t) {
drivers/clk/ralink/clk-mtmips.c
506
u32 t;
drivers/clk/ralink/clk-mtmips.c
508
regmap_read(sysc, SYSC_REG_SYSTEM_CONFIG, &t);
drivers/clk/ralink/clk-mtmips.c
509
ddr2 = t & RT3883_SYSCFG0_DRAM_TYPE_DDR2;
drivers/clk/ralink/clk-mtmips.c
531
u32 t;
drivers/clk/ralink/clk-mtmips.c
533
regmap_read(sysc, SYSC_REG_SYSTEM_CONFIG, &t);
drivers/clk/ralink/clk-mtmips.c
534
t = (t >> RT2880_CONFIG_CPUCLK_SHIFT) & RT2880_CONFIG_CPUCLK_MASK;
drivers/clk/ralink/clk-mtmips.c
536
switch (t) {
drivers/clk/ralink/clk-mtmips.c
552
u64 t;
drivers/clk/ralink/clk-mtmips.c
554
t = ref_rate;
drivers/clk/ralink/clk-mtmips.c
555
t *= mul;
drivers/clk/ralink/clk-mtmips.c
556
t = div_u64(t, div);
drivers/clk/ralink/clk-mtmips.c
558
return t;
drivers/clk/ralink/clk-mtmips.c
568
u32 t;
drivers/clk/ralink/clk-mtmips.c
572
regmap_read(sysc, SYSC_REG_CPLL_CONFIG0, &t);
drivers/clk/ralink/clk-mtmips.c
573
if (t & CPLL_CFG0_BYPASS_REF_CLK) {
drivers/clk/ralink/clk-mtmips.c
575
} else if ((t & CPLL_CFG0_SW_CFG) == 0) {
drivers/clk/ralink/clk-mtmips.c
578
mul = (t >> CPLL_CFG0_PLL_MULT_RATIO_SHIFT) &
drivers/clk/ralink/clk-mtmips.c
581
if (t & CPLL_CFG0_LC_CURFCK)
drivers/clk/ralink/clk-mtmips.c
584
div = (t >> CPLL_CFG0_PLL_DIV_RATIO_SHIFT) &
drivers/clk/ralink/clk-mtmips.c
592
regmap_read(sysc, SYSC_REG_CPLL_CONFIG1, &t);
drivers/clk/ralink/clk-mtmips.c
593
if (t & CPLL_CFG1_CPU_AUX1)
drivers/clk/ralink/clk-mtmips.c
596
if (t & CPLL_CFG1_CPU_AUX0)
drivers/clk/ralink/clk-mtmips.c
607
u32 t;
drivers/clk/ralink/clk-mtmips.c
611
regmap_read(sysc, SYSC_REG_CPU_SYS_CLKCFG, &t);
drivers/clk/ralink/clk-mtmips.c
612
mul = t & CPU_SYS_CLKCFG_CPU_FFRAC_MASK;
drivers/clk/ralink/clk-mtmips.c
613
div = (t >> CPU_SYS_CLKCFG_CPU_FDIV_SHIFT) &
drivers/clk/ralink/clk-mtmips.c
631
u32 t;
drivers/clk/ralink/clk-mtmips.c
635
regmap_read(sysc, SYSC_REG_CPU_SYS_CLKCFG, &t);
drivers/clk/ralink/clk-mtmips.c
636
ocp_ratio = (t >> CPU_SYS_CLKCFG_OCP_RATIO_SHIFT) &
drivers/clk/ralink/clk-mtmips.c
655
u32 t;
drivers/clk/ralink/clk-mtmips.c
657
regmap_read(sysc, SYSC_REG_CLKCFG0, &t);
drivers/clk/ralink/clk-mtmips.c
658
if (t & CLKCFG0_PERI_CLK_SEL)
drivers/clk/ralink/clk-mtmips.c
669
u32 t;
drivers/clk/ralink/clk-mtmips.c
671
regmap_read(sysc, SYSC_REG_SYSTEM_CONFIG, &t);
drivers/clk/ralink/clk-mtmips.c
672
if (t & MT7620_XTAL_FREQ_SEL)
drivers/clk/ralink/clk-mtmips.c
895
u32 t;
drivers/clk/ralink/clk-mtmips.c
905
regmap_read(priv->sysc, SYSC_REG_CPU_SYS_CLKCFG, &t);
drivers/clk/ralink/clk-mtmips.c
906
t &= ~(CLKCFG_FDIV_MASK | CLKCFG_FFRAC_MASK);
drivers/clk/ralink/clk-mtmips.c
907
t |= CLKCFG_FDIV_USB_VAL | CLKCFG_FFRAC_USB_VAL;
drivers/clk/ralink/clk-mtmips.c
908
regmap_write(priv->sysc, SYSC_REG_CPU_SYS_CLKCFG, t);
drivers/clk/samsung/clk.h
194
#define __DIV(_id, cname, pname, o, s, w, f, df, t) \
drivers/clk/samsung/clk.h
204
.table = t, \
drivers/clk/samsung/clk.h
213
#define DIV_T(_id, cname, pname, o, s, w, t) \
drivers/clk/samsung/clk.h
214
__DIV(_id, cname, pname, o, s, w, 0, 0, t)
drivers/clk/tegra/clk-tegra124-emc.c
121
int i, k, t;
drivers/clk/tegra/clk-tegra124-emc.c
130
for (t = k; t < tegra->num_timings; t++) {
drivers/clk/tegra/clk-tegra124-emc.c
131
if (tegra->timings[t].ram_code != ram_code)
drivers/clk/tegra/clk-tegra124-emc.c
135
for (i = k; i < t; i++) {
drivers/clk/tegra/clk-tegra124-emc.c
138
if (timing->rate < req->rate && i != t - 1)
drivers/clocksource/arm_arch_timer.c
559
#define arch_timer_check_ool_workaround(t,a) do { } while(0)
drivers/clocksource/arm_arch_timer_mmio.c
119
static noinstr u64 arch_counter_mmio_get_cnt(struct arch_timer *t)
drivers/clocksource/arm_arch_timer_mmio.c
121
int offset_lo = t->access == VIRT_ACCESS ? CNTVCT_LO : CNTPCT_LO;
drivers/clocksource/arm_arch_timer_mmio.c
125
cnt_hi = __le32_to_cpu((__le32 __force)__raw_readl(t->base + offset_lo + 4));
drivers/clocksource/arm_arch_timer_mmio.c
126
cnt_lo = __le32_to_cpu((__le32 __force)__raw_readl(t->base + offset_lo));
drivers/clocksource/arm_arch_timer_mmio.c
127
tmp_hi = __le32_to_cpu((__le32 __force)__raw_readl(t->base + offset_lo + 4));
drivers/clocksource/timer-atmel-pit.c
78
u32 t;
drivers/clocksource/timer-atmel-pit.c
82
t = pit_read(data->base, AT91_PIT_PIIR);
drivers/clocksource/timer-atmel-pit.c
85
elapsed += PIT_PICNT(t) * data->cycle;
drivers/clocksource/timer-atmel-pit.c
86
elapsed += PIT_CPIV(t);
drivers/clocksource/timer-digicolor.c
48
#define CONTROL(t) ((t)*8)
drivers/clocksource/timer-digicolor.c
49
#define COUNT(t) ((t)*8 + 4)
drivers/clocksource/timer-ti-dm-systimer.c
100
writel_relaxed(BIT(1) | BIT(2), t->base + t->ifctrl);
drivers/clocksource/timer-ti-dm-systimer.c
108
static int __init dmtimer_systimer_type2_reset(struct dmtimer_systimer *t)
drivers/clocksource/timer-ti-dm-systimer.c
110
void __iomem *sysc = t->base + t->sysc;
drivers/clocksource/timer-ti-dm-systimer.c
113
dmtimer_systimer_enable(t);
drivers/clocksource/timer-ti-dm-systimer.c
122
static int __init dmtimer_systimer_reset(struct dmtimer_systimer *t)
drivers/clocksource/timer-ti-dm-systimer.c
126
if (dmtimer_systimer_revision1(t))
drivers/clocksource/timer-ti-dm-systimer.c
127
ret = dmtimer_systimer_type1_reset(t);
drivers/clocksource/timer-ti-dm-systimer.c
129
ret = dmtimer_systimer_type2_reset(t);
drivers/clocksource/timer-ti-dm-systimer.c
323
static int __init dmtimer_systimer_init_clock(struct dmtimer_systimer *t,
drivers/clocksource/timer-ti-dm-systimer.c
352
t->ick = clock;
drivers/clocksource/timer-ti-dm-systimer.c
354
t->fck = clock;
drivers/clocksource/timer-ti-dm-systimer.c
362
struct dmtimer_systimer *t)
drivers/clocksource/timer-ti-dm-systimer.c
371
t->base = of_iomap(np, 0);
drivers/clocksource/timer-ti-dm-systimer.c
372
if (!t->base)
drivers/clocksource/timer-ti-dm-systimer.c
385
error = dmtimer_systimer_init_clock(t, np->parent, "fck", &rate);
drivers/clocksource/timer-ti-dm-systimer.c
389
t->rate = rate;
drivers/clocksource/timer-ti-dm-systimer.c
391
error = dmtimer_systimer_init_clock(t, np->parent, "ick", &rate);
drivers/clocksource/timer-ti-dm-systimer.c
395
if (dmtimer_systimer_revision1(t)) {
drivers/clocksource/timer-ti-dm-systimer.c
396
t->irq_stat = OMAP_TIMER_V1_STAT_OFFSET;
drivers/clocksource/timer-ti-dm-systimer.c
397
t->irq_ena = OMAP_TIMER_V1_INT_EN_OFFSET;
drivers/clocksource/timer-ti-dm-systimer.c
398
t->pend = _OMAP_TIMER_WRITE_PEND_OFFSET;
drivers/clocksource/timer-ti-dm-systimer.c
401
t->irq_stat = OMAP_TIMER_V2_IRQSTATUS;
drivers/clocksource/timer-ti-dm-systimer.c
402
t->irq_ena = OMAP_TIMER_V2_IRQENABLE_SET;
drivers/clocksource/timer-ti-dm-systimer.c
404
t->pend = regbase + _OMAP_TIMER_WRITE_PEND_OFFSET;
drivers/clocksource/timer-ti-dm-systimer.c
407
t->sysc = OMAP_TIMER_OCP_CFG_OFFSET;
drivers/clocksource/timer-ti-dm-systimer.c
408
t->load = regbase + _OMAP_TIMER_LOAD_OFFSET;
drivers/clocksource/timer-ti-dm-systimer.c
409
t->counter = regbase + _OMAP_TIMER_COUNTER_OFFSET;
drivers/clocksource/timer-ti-dm-systimer.c
410
t->ctrl = regbase + _OMAP_TIMER_CTRL_OFFSET;
drivers/clocksource/timer-ti-dm-systimer.c
411
t->wakeup = regbase + _OMAP_TIMER_WAKEUP_EN_OFFSET;
drivers/clocksource/timer-ti-dm-systimer.c
412
t->ifctrl = regbase + _OMAP_TIMER_IF_CTRL_OFFSET;
drivers/clocksource/timer-ti-dm-systimer.c
414
dmtimer_systimer_reset(t);
drivers/clocksource/timer-ti-dm-systimer.c
415
dmtimer_systimer_enable(t);
drivers/clocksource/timer-ti-dm-systimer.c
416
pr_debug("dmtimer rev %08x sysc %08x\n", readl_relaxed(t->base),
drivers/clocksource/timer-ti-dm-systimer.c
417
readl_relaxed(t->base + t->sysc));
drivers/clocksource/timer-ti-dm-systimer.c
422
iounmap(t->base);
drivers/clocksource/timer-ti-dm-systimer.c
437
struct dmtimer_systimer *t = &clkevt->t;
drivers/clocksource/timer-ti-dm-systimer.c
439
writel_relaxed(OMAP_TIMER_INT_OVERFLOW, t->base + t->irq_stat);
drivers/clocksource/timer-ti-dm-systimer.c
449
struct dmtimer_systimer *t = &clkevt->t;
drivers/clocksource/timer-ti-dm-systimer.c
450
void __iomem *pend = t->base + t->pend;
drivers/clocksource/timer-ti-dm-systimer.c
454
writel_relaxed(0xffffffff - cycles, t->base + t->counter);
drivers/clocksource/timer-ti-dm-systimer.c
458
writel_relaxed(OMAP_TIMER_CTRL_ST, t->base + t->ctrl);
drivers/clocksource/timer-ti-dm-systimer.c
466
struct dmtimer_systimer *t = &clkevt->t;
drivers/clocksource/timer-ti-dm-systimer.c
467
void __iomem *ctrl = t->base + t->ctrl;
drivers/clocksource/timer-ti-dm-systimer.c
477
udelay(3500000 / t->rate + 1);
drivers/clocksource/timer-ti-dm-systimer.c
479
writel_relaxed(OMAP_TIMER_INT_OVERFLOW, t->base + t->irq_stat);
drivers/clocksource/timer-ti-dm-systimer.c
487
struct dmtimer_systimer *t = &clkevt->t;
drivers/clocksource/timer-ti-dm-systimer.c
488
void __iomem *pend = t->base + t->pend;
drivers/clocksource/timer-ti-dm-systimer.c
495
writel_relaxed(clkevt->period, t->base + t->load);
drivers/clocksource/timer-ti-dm-systimer.c
499
writel_relaxed(clkevt->period, t->base + t->counter);
drivers/clocksource/timer-ti-dm-systimer.c
504
t->base + t->ctrl);
drivers/clocksource/timer-ti-dm-systimer.c
512
struct dmtimer_systimer *t = &clkevt->t;
drivers/clocksource/timer-ti-dm-systimer.c
514
dmtimer_systimer_disable(t);
drivers/clocksource/timer-ti-dm-systimer.c
515
clk_disable(t->fck);
drivers/clocksource/timer-ti-dm-systimer.c
521
struct dmtimer_systimer *t = &clkevt->t;
drivers/clocksource/timer-ti-dm-systimer.c
524
error = clk_enable(t->fck);
drivers/clocksource/timer-ti-dm-systimer.c
528
dmtimer_systimer_enable(t);
drivers/clocksource/timer-ti-dm-systimer.c
529
writel_relaxed(OMAP_TIMER_INT_OVERFLOW, t->base + t->irq_ena);
drivers/clocksource/timer-ti-dm-systimer.c
530
writel_relaxed(OMAP_TIMER_INT_OVERFLOW, t->base + t->wakeup);
drivers/clocksource/timer-ti-dm-systimer.c
541
struct dmtimer_systimer *t;
drivers/clocksource/timer-ti-dm-systimer.c
544
t = &clkevt->t;
drivers/clocksource/timer-ti-dm-systimer.c
55
struct dmtimer_systimer t;
drivers/clocksource/timer-ti-dm-systimer.c
565
error = dmtimer_systimer_setup(np, &clkevt->t);
drivers/clocksource/timer-ti-dm-systimer.c
569
clkevt->period = 0xffffffff - DIV_ROUND_CLOSEST(t->rate, HZ);
drivers/clocksource/timer-ti-dm-systimer.c
576
writel_relaxed(OMAP_TIMER_CTRL_POSTED, t->base + t->ifctrl);
drivers/clocksource/timer-ti-dm-systimer.c
583
writel_relaxed(OMAP_TIMER_INT_OVERFLOW, t->base + t->irq_ena);
drivers/clocksource/timer-ti-dm-systimer.c
584
writel_relaxed(OMAP_TIMER_INT_OVERFLOW, t->base + t->wakeup);
drivers/clocksource/timer-ti-dm-systimer.c
588
"always-on " : "", t->rate, np->parent);
drivers/clocksource/timer-ti-dm-systimer.c
593
iounmap(t->base);
drivers/clocksource/timer-ti-dm-systimer.c
61
struct dmtimer_systimer t;
drivers/clocksource/timer-ti-dm-systimer.c
615
clockevents_config_and_register(&clkevt->dev, clkevt->t.rate,
drivers/clocksource/timer-ti-dm-systimer.c
66
static bool dmtimer_systimer_revision1(struct dmtimer_systimer *t)
drivers/clocksource/timer-ti-dm-systimer.c
664
struct dmtimer_systimer *t = &clkevt->t;
drivers/clocksource/timer-ti-dm-systimer.c
666
clockevents_config_and_register(dev, t->rate, 3, ULONG_MAX);
drivers/clocksource/timer-ti-dm-systimer.c
675
struct dmtimer_systimer *t = &clkevt->t;
drivers/clocksource/timer-ti-dm-systimer.c
677
if (t->sysc) {
drivers/clocksource/timer-ti-dm-systimer.c
68
u32 tidr = readl_relaxed(t->base);
drivers/clocksource/timer-ti-dm-systimer.c
715
struct dmtimer_systimer *t = &clksrc->t;
drivers/clocksource/timer-ti-dm-systimer.c
717
return (u64)readl_relaxed(t->base + t->counter);
drivers/clocksource/timer-ti-dm-systimer.c
73
static void dmtimer_systimer_enable(struct dmtimer_systimer *t)
drivers/clocksource/timer-ti-dm-systimer.c
730
struct dmtimer_systimer *t = &clksrc->t;
drivers/clocksource/timer-ti-dm-systimer.c
732
clksrc->loadval = readl_relaxed(t->base + t->counter);
drivers/clocksource/timer-ti-dm-systimer.c
733
dmtimer_systimer_disable(t);
drivers/clocksource/timer-ti-dm-systimer.c
734
clk_disable(t->fck);
drivers/clocksource/timer-ti-dm-systimer.c
740
struct dmtimer_systimer *t = &clksrc->t;
drivers/clocksource/timer-ti-dm-systimer.c
743
error = clk_enable(t->fck);
drivers/clocksource/timer-ti-dm-systimer.c
747
dmtimer_systimer_enable(t);
drivers/clocksource/timer-ti-dm-systimer.c
748
writel_relaxed(clksrc->loadval, t->base + t->counter);
drivers/clocksource/timer-ti-dm-systimer.c
750
t->base + t->ctrl);
drivers/clocksource/timer-ti-dm-systimer.c
756
struct dmtimer_systimer *t;
drivers/clocksource/timer-ti-dm-systimer.c
765
t = &clksrc->t;
drivers/clocksource/timer-ti-dm-systimer.c
767
error = dmtimer_systimer_setup(np, t);
drivers/clocksource/timer-ti-dm-systimer.c
77
if (dmtimer_systimer_revision1(t))
drivers/clocksource/timer-ti-dm-systimer.c
783
writel_relaxed(0, t->base + t->counter);
drivers/clocksource/timer-ti-dm-systimer.c
785
t->base + t->ctrl);
drivers/clocksource/timer-ti-dm-systimer.c
792
dmtimer_sched_clock_counter = t->base + t->counter;
drivers/clocksource/timer-ti-dm-systimer.c
793
sched_clock_register(dmtimer_read_sched_clock, 32, t->rate);
drivers/clocksource/timer-ti-dm-systimer.c
796
if (clocksource_register_hz(dev, t->rate))
drivers/clocksource/timer-ti-dm-systimer.c
82
writel_relaxed(val, t->base + t->sysc);
drivers/clocksource/timer-ti-dm-systimer.c
85
static void dmtimer_systimer_disable(struct dmtimer_systimer *t)
drivers/clocksource/timer-ti-dm-systimer.c
87
if (!dmtimer_systimer_revision1(t))
drivers/clocksource/timer-ti-dm-systimer.c
90
writel_relaxed(DMTIMER_TYPE1_DISABLE, t->base + t->sysc);
drivers/clocksource/timer-ti-dm-systimer.c
93
static int __init dmtimer_systimer_type1_reset(struct dmtimer_systimer *t)
drivers/clocksource/timer-ti-dm-systimer.c
95
void __iomem *syss = t->base + OMAP_TIMER_V1_SYS_STAT_OFFSET;
drivers/clocksource/timer-ti-dm-systimer.c
99
dmtimer_systimer_enable(t);
drivers/clocksource/timer-ti-dm.c
516
struct dmtimer *timer = NULL, *t;
drivers/clocksource/timer-ti-dm.c
538
list_for_each_entry(t, &omap_timer_list, node) {
drivers/clocksource/timer-ti-dm.c
539
if (t->reserved)
drivers/clocksource/timer-ti-dm.c
544
if (id == t->pdev->id) {
drivers/clocksource/timer-ti-dm.c
545
timer = t;
drivers/clocksource/timer-ti-dm.c
551
if (cap == (t->capability & cap)) {
drivers/clocksource/timer-ti-dm.c
562
timer = t;
drivers/clocksource/timer-ti-dm.c
566
if (t->capability == cap)
drivers/clocksource/timer-ti-dm.c
571
if (np == t->pdev->dev.of_node) {
drivers/clocksource/timer-ti-dm.c
572
timer = t;
drivers/clocksource/timer-ti-dm.c
579
timer = t;
drivers/clocksource/timer-vt8500.c
37
#define msecs_to_loops(t) (loops_per_jiffy / 1000 * HZ * t)
drivers/comedi/drivers/comedi_test.c
198
static void waveform_ai_timer(struct timer_list *t)
drivers/comedi/drivers/comedi_test.c
200
struct waveform_private *devpriv = timer_container_of(devpriv, t,
drivers/comedi/drivers/comedi_test.c
446
static void waveform_ao_timer(struct timer_list *t)
drivers/comedi/drivers/comedi_test.c
448
struct waveform_private *devpriv = timer_container_of(devpriv, t,
drivers/comedi/drivers/das16.c
518
static void das16_timer_interrupt(struct timer_list *t)
drivers/comedi/drivers/das16.c
520
struct das16_private_struct *devpriv = timer_container_of(devpriv, t,
drivers/comedi/drivers/jr3_pci.c
564
static void jr3_pci_poll_dev(struct timer_list *t)
drivers/comedi/drivers/jr3_pci.c
566
struct jr3_pci_dev_private *devpriv = timer_container_of(devpriv, t,
drivers/comedi/kcomedilib/kcomedilib_main.c
67
unsigned int t = 0;
drivers/comedi/kcomedilib/kcomedilib_main.c
76
while ((t = find_next_bit(destinations[cur],
drivers/comedi/kcomedilib/kcomedilib_main.c
78
t)) < COMEDI_NUM_BOARD_MINORS) {
drivers/comedi/kcomedilib/kcomedilib_main.c
82
if (kcomedilib_to_from[t][f])
drivers/comedi/kcomedilib/kcomedilib_main.c
85
t++;
drivers/cpufreq/bmips-cpufreq.c
45
#define BMIPS(c, t, m, f) { \
drivers/cpufreq/bmips-cpufreq.c
47
.bmips_type = (t), \
drivers/cpufreq/cpufreq.c
644
struct cpufreq_governor *t;
drivers/cpufreq/cpufreq.c
646
for_each_governor(t)
drivers/cpufreq/cpufreq.c
647
if (!strncasecmp(str_governor, t->name, CPUFREQ_NAME_LEN))
drivers/cpufreq/cpufreq.c
648
return t;
drivers/cpufreq/cpufreq.c
655
struct cpufreq_governor *t;
drivers/cpufreq/cpufreq.c
658
t = find_governor(str_governor);
drivers/cpufreq/cpufreq.c
659
if (!t)
drivers/cpufreq/cpufreq.c
662
if (!try_module_get(t->owner))
drivers/cpufreq/cpufreq.c
663
t = NULL;
drivers/cpufreq/cpufreq.c
668
return t;
drivers/cpufreq/cpufreq.c
688
struct cpufreq_governor *t;
drivers/cpufreq/cpufreq.c
690
t = get_governor(str_governor);
drivers/cpufreq/cpufreq.c
691
if (t)
drivers/cpufreq/cpufreq.c
692
return t;
drivers/cpufreq/cpufreq.c
863
struct cpufreq_governor *t;
drivers/cpufreq/cpufreq.c
871
for_each_governor(t) {
drivers/cpufreq/cpufreq.c
875
i += sysfs_emit_at(buf, i, "%s ", t->name);
drivers/cpufreq/longhaul.c
165
u32 t;
drivers/cpufreq/longhaul.c
194
t = inl(acpi_gbl_FADT.xpm_timer_block.address);
drivers/cpufreq/longhaul.c
211
t = inl(acpi_gbl_FADT.xpm_timer_block.address);
drivers/cpufreq/longhaul.c
231
t = inl(acpi_gbl_FADT.xpm_timer_block.address);
drivers/cpufreq/powernv-cpufreq.c
670
static void gpstate_timer_handler(struct timer_list *t)
drivers/cpufreq/powernv-cpufreq.c
672
struct global_pstate_info *gpstates = timer_container_of(gpstates, t,
drivers/crypto/axis/artpec6_crypto.c
2073
static void artpec6_crypto_timeout(struct timer_list *t)
drivers/crypto/axis/artpec6_crypto.c
2075
struct artpec6_crypto *ac = timer_container_of(ac, t, timer);
drivers/crypto/ccp/sev-dev-tsm.c
344
struct sev_tio_status *t = kzalloc_obj(*t);
drivers/crypto/ccp/sev-dev-tsm.c
350
if (!t)
drivers/crypto/ccp/sev-dev-tsm.c
363
memcpy(t, tio_status_page, sizeof(*t));
drivers/crypto/ccp/sev-dev-tsm.c
367
t->tio_en, t->tio_init_done,
drivers/crypto/ccp/sev-dev-tsm.c
368
t->spdm_req_size_min, t->spdm_req_size_max,
drivers/crypto/ccp/sev-dev-tsm.c
369
t->spdm_rsp_size_min, t->spdm_rsp_size_max,
drivers/crypto/ccp/sev-dev-tsm.c
370
t->spdm_scratch_size_min, t->spdm_scratch_size_max,
drivers/crypto/ccp/sev-dev-tsm.c
371
t->spdm_out_size_min, t->spdm_out_size_max,
drivers/crypto/ccp/sev-dev-tsm.c
372
t->devctx_size, t->tdictx_size,
drivers/crypto/ccp/sev-dev-tsm.c
373
t->tio_crypto_alg);
drivers/crypto/ccp/sev-dev-tsm.c
376
sev->tio_status = t;
drivers/crypto/ccp/sev-dev-tsm.c
382
ret, t->tio_en, t->tio_init_done, boot_cpu_has(X86_FEATURE_SEV));
drivers/crypto/ccp/sev-dev-tsm.c
383
kfree(t);
drivers/crypto/hifn_795x.c
1307
struct scatterlist *t;
drivers/crypto/hifn_795x.c
1324
t = &rctx->walk.cache[0];
drivers/crypto/hifn_795x.c
1327
if (t->length && rctx->walk.flags & ASYNC_FLAGS_MISALIGNED) {
drivers/crypto/hifn_795x.c
1328
BUG_ON(!sg_page(t));
drivers/crypto/hifn_795x.c
1329
dpage = sg_page(t);
drivers/crypto/hifn_795x.c
1331
len = t->length;
drivers/crypto/hifn_795x.c
1343
t++;
drivers/crypto/hifn_795x.c
1424
struct scatterlist *dst, *t;
drivers/crypto/hifn_795x.c
1445
t = &w->cache[idx];
drivers/crypto/hifn_795x.c
1493
t->length = copy;
drivers/crypto/hifn_795x.c
1494
t->offset = offset;
drivers/crypto/hifn_795x.c
1655
struct scatterlist *dst, *t;
drivers/crypto/hifn_795x.c
1659
t = &rctx->walk.cache[idx];
drivers/crypto/hifn_795x.c
1665
__func__, sg_page(t), t->length,
drivers/crypto/hifn_795x.c
1668
if (!t->length) {
drivers/crypto/hifn_795x.c
1674
saddr = kmap_atomic(sg_page(t));
drivers/crypto/hifn_795x.c
1676
err = skcipher_get(saddr, &t->length, t->offset,
drivers/crypto/hifn_795x.c
2246
static int hifn_alg_alloc(struct hifn_device *dev, const struct hifn_alg_template *t)
drivers/crypto/hifn_795x.c
2255
alg->alg = t->skcipher;
drivers/crypto/hifn_795x.c
2260
"%s", t->name) >= CRYPTO_MAX_ALG_NAME)
drivers/crypto/hifn_795x.c
2263
"%s-%s", t->drv_name, dev->name) >= CRYPTO_MAX_ALG_NAME)
drivers/crypto/hifn_795x.c
2268
alg->alg.base.cra_blocksize = t->bsize;
drivers/crypto/omap-aes.c
499
static void omap_aes_done_task(struct work_struct *t)
drivers/crypto/omap-aes.c
501
struct omap_aes_dev *dd = from_work(dd, t, done_task);
drivers/crypto/omap-des.c
584
static void omap_des_done_task(struct work_struct *t)
drivers/crypto/omap-des.c
586
struct omap_des_dev *dd = from_work(dd, t, done_task);
drivers/crypto/omap-sham.c
1708
static void omap_sham_done_task(struct work_struct *t)
drivers/crypto/omap-sham.c
1710
struct omap_sham_dev *dd = from_work(dd, t, done_task);
drivers/crypto/tegra/tegra-se-aes.c
967
unsigned int q, t;
drivers/crypto/tegra/tegra-se-aes.c
975
t = rctx->authsize;
drivers/crypto/tegra/tegra-se-aes.c
976
nonce[0] |= (((t - 2) / 2) << 3);
drivers/cxl/core/hdm.c
713
struct cxl_dport **t = &cxlsd->target[0];
drivers/cxl/core/hdm.c
716
*tgt = FIELD_PREP(GENMASK(7, 0), t[0]->port_id);
drivers/cxl/core/hdm.c
718
*tgt |= FIELD_PREP(GENMASK(15, 8), t[1]->port_id);
drivers/cxl/core/hdm.c
720
*tgt |= FIELD_PREP(GENMASK(23, 16), t[2]->port_id);
drivers/cxl/core/hdm.c
722
*tgt |= FIELD_PREP(GENMASK(31, 24), t[3]->port_id);
drivers/cxl/core/hdm.c
724
*tgt |= FIELD_PREP(GENMASK_ULL(39, 32), t[4]->port_id);
drivers/cxl/core/hdm.c
726
*tgt |= FIELD_PREP(GENMASK_ULL(47, 40), t[5]->port_id);
drivers/cxl/core/hdm.c
728
*tgt |= FIELD_PREP(GENMASK_ULL(55, 48), t[6]->port_id);
drivers/cxl/core/hdm.c
730
*tgt |= FIELD_PREP(GENMASK_ULL(63, 56), t[7]->port_id);
drivers/dma-buf/dma-fence-unwrap.c
136
ktime_t t = dma_fence_timestamp(tmp);
drivers/dma-buf/dma-fence-unwrap.c
138
if (ktime_after(t, timestamp))
drivers/dma-buf/dma-fence-unwrap.c
139
timestamp = t;
drivers/dma-buf/dma-fence-unwrap.c
168
ktime_t t = dma_fence_timestamp(tmp);
drivers/dma-buf/dma-fence-unwrap.c
170
if (ktime_after(t, timestamp))
drivers/dma-buf/dma-fence-unwrap.c
171
timestamp = t;
drivers/dma-buf/st-dma-fence.c
419
const struct race_thread *t = arg;
drivers/dma-buf/st-dma-fence.c
436
rcu_assign_pointer(t->fences[t->id], f1);
drivers/dma-buf/st-dma-fence.c
441
f2 = dma_fence_get_rcu_safe(&t->fences[!t->id]);
drivers/dma-buf/st-dma-fence.c
445
if (t->before)
drivers/dma-buf/st-dma-fence.c
455
if (!t->before)
drivers/dma-buf/st-dma-fence.c
465
t->id, pass, miss,
drivers/dma-buf/st-dma-fence.c
466
t->before ? "before" : "after",
drivers/dma-buf/st-dma-fence.c
473
rcu_assign_pointer(t->fences[t->id], NULL);
drivers/dma-buf/st-dma-fence.c
482
__func__, t->id, pass, miss);
drivers/dma-buf/st-dma-fence.c
493
struct race_thread t[2];
drivers/dma-buf/st-dma-fence.c
496
for (i = 0; i < ARRAY_SIZE(t); i++) {
drivers/dma-buf/st-dma-fence.c
497
t[i].fences = f;
drivers/dma-buf/st-dma-fence.c
498
t[i].id = i;
drivers/dma-buf/st-dma-fence.c
499
t[i].before = pass;
drivers/dma-buf/st-dma-fence.c
500
t[i].task = kthread_run(thread_signal_callback, &t[i],
drivers/dma-buf/st-dma-fence.c
502
if (IS_ERR(t[i].task)) {
drivers/dma-buf/st-dma-fence.c
503
ret = PTR_ERR(t[i].task);
drivers/dma-buf/st-dma-fence.c
505
kthread_stop_put(t[i].task);
drivers/dma-buf/st-dma-fence.c
508
get_task_struct(t[i].task);
drivers/dma-buf/st-dma-fence.c
513
for (i = 0; i < ARRAY_SIZE(t); i++) {
drivers/dma-buf/st-dma-fence.c
516
err = kthread_stop_put(t[i].task);
drivers/dma/altera-msgdma.c
683
static void msgdma_tasklet(struct tasklet_struct *t)
drivers/dma/altera-msgdma.c
685
struct msgdma_device *mdev = from_tasklet(mdev, t, irq_tasklet);
drivers/dma/apple-admac.c
715
static void admac_chan_tasklet(struct tasklet_struct *t)
drivers/dma/apple-admac.c
717
struct admac_chan *adchan = from_tasklet(adchan, t, tasklet);
drivers/dma/at_hdmac.c
366
static inline struct at_desc *to_atdma_desc(struct dma_async_tx_descriptor *t)
drivers/dma/at_hdmac.c
368
return container_of(t, struct at_desc, vd.tx);
drivers/dma/at_xdmac.c
1762
static void at_xdmac_tasklet(struct tasklet_struct *t)
drivers/dma/at_xdmac.c
1764
struct at_xdmac_chan *atchan = from_tasklet(atchan, t, tasklet);
drivers/dma/bcm2835-dma.c
199
struct dma_async_tx_descriptor *t)
drivers/dma/bcm2835-dma.c
201
return container_of(t, struct bcm2835_desc, vd.tx);
drivers/dma/dw/core.c
464
static void dw_dma_tasklet(struct tasklet_struct *t)
drivers/dma/dw/core.c
466
struct dw_dma *dw = from_tasklet(dw, t, tasklet);
drivers/dma/ep93xx_dma.c
805
static void ep93xx_dma_tasklet(struct tasklet_struct *t)
drivers/dma/ep93xx_dma.c
807
struct ep93xx_dma_chan *edmac = from_tasklet(edmac, t, tasklet);
drivers/dma/fsl_raid.c
158
static void fsl_re_dequeue(struct tasklet_struct *t)
drivers/dma/fsl_raid.c
160
struct fsl_re_chan *re_chan = from_tasklet(re_chan, t, irqtask);
drivers/dma/fsldma.c
980
static void dma_do_tasklet(struct tasklet_struct *t)
drivers/dma/fsldma.c
982
struct fsldma_chan *chan = from_tasklet(chan, t, tasklet);
drivers/dma/idxd/cdev.c
329
u16 h, t, size;
drivers/dma/idxd/cdev.c
338
t = status.tail;
drivers/dma/idxd/cdev.c
342
while (h != t) {
drivers/dma/idxd/debugfs.c
62
u16 h, t, evl_size, i;
drivers/dma/idxd/debugfs.c
72
t = evl_status.tail;
drivers/dma/idxd/debugfs.c
79
i = t;
drivers/dma/idxd/debugfs.c
82
if (i == t)
drivers/dma/idxd/irq.c
108
struct idxd_desc *d, *t;
drivers/dma/idxd/irq.c
114
llist_for_each_entry_safe(d, t, head, llnode)
drivers/dma/idxd/irq.c
118
list_for_each_entry_safe(d, t, &ie->work_list, list) {
drivers/dma/idxd/irq.c
124
list_for_each_entry_safe(d, t, &flist, list) {
drivers/dma/idxd/irq.c
357
unsigned int h, t;
drivers/dma/idxd/irq.c
371
t = evl_status.tail;
drivers/dma/idxd/irq.c
375
while (h != t) {
drivers/dma/idxd/irq.c
598
struct idxd_desc *desc, *t;
drivers/dma/idxd/irq.c
605
llist_for_each_entry_safe(desc, t, head, llnode) {
drivers/dma/idxd/submit.c
100
struct idxd_desc *d, *t, *found = NULL;
drivers/dma/idxd/submit.c
112
llist_for_each_entry_safe(d, t, head, llnode) {
drivers/dma/idxd/submit.c
139
list_for_each_entry_safe(d, t, &flist, list) {
drivers/dma/img-mdc-dma.c
173
static inline struct mdc_tx_desc *to_mdc_desc(struct dma_async_tx_descriptor *t)
drivers/dma/img-mdc-dma.c
175
struct virt_dma_desc *vdesc = container_of(t, struct virt_dma_desc, tx);
drivers/dma/imx-dma.c
338
static void imxdma_watchdog(struct timer_list *t)
drivers/dma/imx-dma.c
340
struct imxdma_channel *imxdmac = timer_container_of(imxdmac, t,
drivers/dma/imx-dma.c
595
static void imxdma_tasklet(struct tasklet_struct *t)
drivers/dma/imx-dma.c
597
struct imxdma_channel *imxdmac = from_tasklet(imxdmac, t, dma_tasklet);
drivers/dma/imx-sdma.c
313
u32 t : 1;
drivers/dma/imx-sdma.c
834
static struct sdma_desc *to_sdma_desc(struct dma_async_tx_descriptor *t)
drivers/dma/imx-sdma.c
836
return container_of(t, struct sdma_desc, vd.tx);
drivers/dma/ioat/dma.c
693
void ioat_cleanup_event(struct tasklet_struct *t)
drivers/dma/ioat/dma.c
695
struct ioatdma_chan *ioat_chan = from_tasklet(ioat_chan, t, cleanup_task);
drivers/dma/ioat/dma.c
902
void ioat_timer_event(struct timer_list *t)
drivers/dma/ioat/dma.c
904
struct ioatdma_chan *ioat_chan = timer_container_of(ioat_chan, t,
drivers/dma/ioat/dma.h
397
void ioat_cleanup_event(struct tasklet_struct *t);
drivers/dma/ioat/dma.h
398
void ioat_timer_event(struct timer_list *t);
drivers/dma/k3dma.c
298
static void k3_dma_tasklet(struct tasklet_struct *t)
drivers/dma/k3dma.c
300
struct k3_dma_dev *d = from_tasklet(d, t, task);
drivers/dma/mediatek/mtk-cqdma.c
358
static void mtk_cqdma_tasklet_cb(struct tasklet_struct *t)
drivers/dma/mediatek/mtk-cqdma.c
360
struct mtk_cqdma_pchan *pc = from_tasklet(pc, t, tasklet);
drivers/dma/mediatek/mtk-uart-apdma.c
115
(struct dma_async_tx_descriptor *t)
drivers/dma/mediatek/mtk-uart-apdma.c
117
return container_of(t, struct mtk_uart_apdma_desc, vd.tx);
drivers/dma/mmp_pdma.c
1042
static void dma_do_tasklet(struct tasklet_struct *t)
drivers/dma/mmp_pdma.c
1044
struct mmp_pdma_chan *chan = from_tasklet(chan, t, tasklet);
drivers/dma/mmp_tdma.c
349
static void dma_do_tasklet(struct tasklet_struct *t)
drivers/dma/mmp_tdma.c
351
struct mmp_tdma_chan *tdmac = from_tasklet(tdmac, t, tasklet);
drivers/dma/moxart-dma.c
168
struct dma_async_tx_descriptor *t)
drivers/dma/moxart-dma.c
170
return container_of(t, struct moxart_desc, vd.tx);
drivers/dma/mpc512x_dma.c
417
static void mpc_dma_tasklet(struct tasklet_struct *t)
drivers/dma/mpc512x_dma.c
419
struct mpc_dma *mdma = from_tasklet(mdma, t, tasklet);
drivers/dma/mv_xor.c
339
static void mv_xor_tasklet(struct tasklet_struct *t)
drivers/dma/mv_xor.c
341
struct mv_xor_chan *chan = from_tasklet(chan, t, irq_tasklet);
drivers/dma/mv_xor_v2.c
556
static void mv_xor_v2_tasklet(struct tasklet_struct *t)
drivers/dma/mv_xor_v2.c
558
struct mv_xor_v2_device *xor_dev = from_tasklet(xor_dev, t,
drivers/dma/mxs-dma.c
303
static void mxs_dma_tasklet(struct tasklet_struct *t)
drivers/dma/mxs-dma.c
305
struct mxs_dma_chan *mxs_chan = from_tasklet(mxs_chan, t, tasklet);
drivers/dma/nbpfaxi.c
1128
static void nbpf_chan_tasklet(struct tasklet_struct *t)
drivers/dma/nbpfaxi.c
1130
struct nbpf_channel *chan = from_tasklet(chan, t, tasklet);
drivers/dma/pch_dma.c
668
static void pdc_tasklet(struct tasklet_struct *t)
drivers/dma/pch_dma.c
670
struct pch_dma_chan *pd_chan = from_tasklet(pd_chan, t, tasklet);
drivers/dma/pl330.c
1580
static void pl330_dotask(struct tasklet_struct *t)
drivers/dma/pl330.c
1582
struct pl330_dmac *pl330 = from_tasklet(pl330, t, tasks);
drivers/dma/pl330.c
2068
static void pl330_tasklet(struct tasklet_struct *t)
drivers/dma/pl330.c
2070
struct dma_pl330_chan *pch = from_tasklet(pch, t, task);
drivers/dma/pl330.c
252
#define UNTIL(t, s) while (!(_state(t) & (s))) cpu_relax();
drivers/dma/pl330.c
868
#define msecs_to_loops(t) (loops_per_jiffy / 1000 * HZ * t)
drivers/dma/plx_dma.c
244
static void plx_dma_desc_task(struct tasklet_struct *t)
drivers/dma/plx_dma.c
246
struct plx_dma_dev *plxdev = from_tasklet(plxdev, t, desc_task);
drivers/dma/ppc4xx/adma.c
1661
static void ppc440spe_adma_tasklet(struct tasklet_struct *t)
drivers/dma/ppc4xx/adma.c
1663
struct ppc440spe_adma_chan *chan = from_tasklet(chan, t, irq_tasklet);
drivers/dma/qcom/bam_dma.c
1099
static void dma_tasklet(struct tasklet_struct *t)
drivers/dma/qcom/bam_dma.c
1101
struct bam_device *bdev = from_tasklet(bdev, t, task);
drivers/dma/qcom/hidma.c
219
static void hidma_issue_task(struct tasklet_struct *t)
drivers/dma/qcom/hidma.c
221
struct hidma_dev *dmadev = from_tasklet(dmadev, t, task);
drivers/dma/qcom/hidma_ll.c
176
static void hidma_ll_tre_complete(struct tasklet_struct *t)
drivers/dma/qcom/hidma_ll.c
178
struct hidma_lldev *lldev = from_tasklet(lldev, t, task);
drivers/dma/sa11x0-dma.c
326
static void sa11x0_dma_tasklet(struct tasklet_struct *t)
drivers/dma/sa11x0-dma.c
328
struct sa11x0_dma_dev *d = from_tasklet(d, t, task);
drivers/dma/sf-pdma/sf-pdma.c
298
static void sf_pdma_donebh_tasklet(struct tasklet_struct *t)
drivers/dma/sf-pdma/sf-pdma.c
300
struct sf_pdma_chan *chan = from_tasklet(chan, t, done_tasklet);
drivers/dma/sf-pdma/sf-pdma.c
322
static void sf_pdma_errbh_tasklet(struct tasklet_struct *t)
drivers/dma/sf-pdma/sf-pdma.c
324
struct sf_pdma_chan *chan = from_tasklet(chan, t, err_tasklet);
drivers/dma/ste_dma40.c
1594
static void dma_tasklet(struct tasklet_struct *t)
drivers/dma/ste_dma40.c
1596
struct d40_chan *d40c = from_tasklet(d40c, t, tasklet);
drivers/dma/sun6i-dma.c
478
static void sun6i_dma_tasklet(struct tasklet_struct *t)
drivers/dma/sun6i-dma.c
480
struct sun6i_dma_dev *sdev = from_tasklet(sdev, t, task);
drivers/dma/tegra20-apb-dma.c
635
static void tegra_dma_tasklet(struct tasklet_struct *t)
drivers/dma/tegra20-apb-dma.c
637
struct tegra_dma_channel *tdc = from_tasklet(tdc, t, tasklet);
drivers/dma/ti/edma.c
158
#define EDMA_TCC(t) ((t) << 12)
drivers/dma/ti/k3-udma.c
327
static inline struct udma_desc *to_udma_desc(struct dma_async_tx_descriptor *t)
drivers/dma/ti/k3-udma.c
329
return container_of(t, struct udma_desc, vd.tx);
drivers/dma/ti/k3-udma.c
4048
static void udma_vchan_complete(struct tasklet_struct *t)
drivers/dma/ti/k3-udma.c
4050
struct virt_dma_chan *vc = from_tasklet(vc, t, task);
drivers/dma/ti/omap-dma.c
243
static inline struct omap_desc *to_omap_dma_desc(struct dma_async_tx_descriptor *t)
drivers/dma/ti/omap-dma.c
245
return container_of(t, struct omap_desc, vd.tx);
drivers/dma/timb_dma.c
566
static void td_tasklet(struct tasklet_struct *t)
drivers/dma/timb_dma.c
568
struct timb_dma *td = from_tasklet(td, t, tasklet);
drivers/dma/txx9dmac.c
604
static void txx9dmac_chan_tasklet(struct tasklet_struct *t)
drivers/dma/txx9dmac.c
610
dc = from_tasklet(dc, t, tasklet);
drivers/dma/txx9dmac.c
641
static void txx9dmac_tasklet(struct tasklet_struct *t)
drivers/dma/txx9dmac.c
647
struct txx9dmac_dev *ddev = from_tasklet(ddev, t, tasklet);
drivers/dma/virt-dma.c
83
static void vchan_complete(struct tasklet_struct *t)
drivers/dma/virt-dma.c
85
struct virt_dma_chan *vc = from_tasklet(vc, t, task);
drivers/dma/xgene-dma.c
979
static void xgene_dma_tasklet_cb(struct tasklet_struct *t)
drivers/dma/xgene-dma.c
981
struct xgene_dma_chan *chan = from_tasklet(chan, t, tasklet);
drivers/dma/xilinx/xilinx_dma.c
1120
static void xilinx_dma_do_tasklet(struct tasklet_struct *t)
drivers/dma/xilinx/xilinx_dma.c
1122
struct xilinx_dma_chan *chan = from_tasklet(chan, t, tasklet);
drivers/dma/xilinx/xilinx_dpdma.c
1579
static void xilinx_dpdma_chan_err_task(struct tasklet_struct *t)
drivers/dma/xilinx/xilinx_dpdma.c
1581
struct xilinx_dpdma_chan *chan = from_tasklet(chan, t, err_task);
drivers/dma/xilinx/zynqmp_dma.c
765
static void zynqmp_dma_do_tasklet(struct tasklet_struct *t)
drivers/dma/xilinx/zynqmp_dma.c
767
struct zynqmp_dma_chan *chan = from_tasklet(chan, t, tasklet);
drivers/edac/igen6_edac.c
729
u32 t = MAD_INTER_CHANNEL_DDR_TYPE(mad_inter);
drivers/edac/igen6_edac.c
731
switch (t) {
drivers/edac/mce_amd.c
724
static const char *smca_get_long_name(enum smca_bank_types t)
drivers/edac/mce_amd.c
726
if (t >= N_SMCA_BANK_TYPES)
drivers/edac/mce_amd.c
729
return smca_long_names[t];
drivers/edac/sb_edac.c
870
static const char *get_intlv_mode_str(u32 reg, enum type t)
drivers/edac/sb_edac.c
872
if (t == KNIGHTS_LANDING)
drivers/firewire/core-transaction.c
100
struct fw_transaction *t;
drivers/firewire/core-transaction.c
105
t = find_and_pop_transaction_entry(card, iter == transaction);
drivers/firewire/core-transaction.c
106
if (!t)
drivers/firewire/core-transaction.c
110
if (!t->with_tstamp) {
drivers/firewire/core-transaction.c
111
t->callback.without_tstamp(card, rcode, NULL, 0, t->callback_data);
drivers/firewire/core-transaction.c
113
t->callback.with_tstamp(card, rcode, t->packet.timestamp, response_tstamp, NULL, 0,
drivers/firewire/core-transaction.c
114
t->callback_data);
drivers/firewire/core-transaction.c
1140
struct fw_transaction *t = NULL;
drivers/firewire/core-transaction.c
1182
t = find_and_pop_transaction_entry(card,
drivers/firewire/core-transaction.c
1186
trace_async_response_inbound((uintptr_t)t, card->index, p->generation, p->speed, p->ack,
drivers/firewire/core-transaction.c
1189
if (!t) {
drivers/firewire/core-transaction.c
1199
card->driver->cancel_packet(card, &t->packet);
drivers/firewire/core-transaction.c
1201
if (!t->with_tstamp) {
drivers/firewire/core-transaction.c
1202
t->callback.without_tstamp(card, rcode, data, data_length, t->callback_data);
drivers/firewire/core-transaction.c
1204
t->callback.with_tstamp(card, rcode, t->packet.timestamp, p->timestamp, data,
drivers/firewire/core-transaction.c
1205
data_length, t->callback_data);
drivers/firewire/core-transaction.c
159
struct fw_transaction *t = timer_container_of(t, timer, split_timeout_timer);
drivers/firewire/core-transaction.c
160
struct fw_card *card = t->card;
drivers/firewire/core-transaction.c
163
if (list_empty(&t->link))
drivers/firewire/core-transaction.c
165
remove_transaction_entry(card, t);
drivers/firewire/core-transaction.c
168
if (!t->with_tstamp) {
drivers/firewire/core-transaction.c
169
t->callback.without_tstamp(card, RCODE_CANCELLED, NULL, 0, t->callback_data);
drivers/firewire/core-transaction.c
171
t->callback.with_tstamp(card, RCODE_CANCELLED, t->packet.timestamp,
drivers/firewire/core-transaction.c
172
t->split_timeout_cycle, NULL, 0, t->callback_data);
drivers/firewire/core-transaction.c
177
static void start_split_transaction_timeout(struct fw_transaction *t, unsigned int delta)
drivers/firewire/core-transaction.c
179
if (list_empty(&t->link) || WARN_ON(t->is_split_transaction))
drivers/firewire/core-transaction.c
182
t->is_split_transaction = true;
drivers/firewire/core-transaction.c
184
mod_timer(&t->split_timeout_timer, jiffies + delta);
drivers/firewire/core-transaction.c
192
struct fw_transaction *t =
drivers/firewire/core-transaction.c
195
trace_async_request_outbound_complete((uintptr_t)t, card->index, packet->generation,
drivers/firewire/core-transaction.c
200
close_transaction(t, card, RCODE_COMPLETE, packet->timestamp);
drivers/firewire/core-transaction.c
209
t->split_timeout_cycle =
drivers/firewire/core-transaction.c
217
start_split_transaction_timeout(t, delta);
drivers/firewire/core-transaction.c
223
close_transaction(t, card, RCODE_BUSY, packet->timestamp);
drivers/firewire/core-transaction.c
226
close_transaction(t, card, RCODE_DATA_ERROR, packet->timestamp);
drivers/firewire/core-transaction.c
229
close_transaction(t, card, RCODE_TYPE_ERROR, packet->timestamp);
drivers/firewire/core-transaction.c
236
close_transaction(t, card, status, packet->timestamp);
drivers/firewire/core-transaction.c
382
void __fw_send_request(struct fw_card *card, struct fw_transaction *t, int tcode,
drivers/firewire/core-transaction.c
39
static int try_cancel_split_timeout(struct fw_transaction *t)
drivers/firewire/core-transaction.c
41
if (t->is_split_transaction)
drivers/firewire/core-transaction.c
415
t->node_id = destination_id;
drivers/firewire/core-transaction.c
416
t->tlabel = tlabel;
drivers/firewire/core-transaction.c
417
t->card = card;
drivers/firewire/core-transaction.c
418
t->is_split_transaction = false;
drivers/firewire/core-transaction.c
419
timer_setup(&t->split_timeout_timer, split_transaction_timeout_callback, 0);
drivers/firewire/core-transaction.c
42
return timer_delete(&t->split_timeout_timer);
drivers/firewire/core-transaction.c
420
t->callback = callback;
drivers/firewire/core-transaction.c
421
t->with_tstamp = with_tstamp;
drivers/firewire/core-transaction.c
422
t->callback_data = callback_data;
drivers/firewire/core-transaction.c
423
t->packet.callback = transmit_complete_callback;
drivers/firewire/core-transaction.c
429
fw_fill_request(&t->packet, tcode, t->tlabel, destination_id, card->node_id,
drivers/firewire/core-transaction.c
436
list_add_tail(&t->link, &card->transactions.list);
drivers/firewire/core-transaction.c
439
trace_async_request_outbound_initiate((uintptr_t)t, card->index, generation, speed,
drivers/firewire/core-transaction.c
440
t->packet.header, payload,
drivers/firewire/core-transaction.c
443
card->driver->send_request(card, &t->packet);
drivers/firewire/core-transaction.c
485
struct fw_transaction t;
drivers/firewire/core-transaction.c
487
timer_setup_on_stack(&t.split_timeout_timer, NULL, 0);
drivers/firewire/core-transaction.c
490
fw_send_request(card, &t, tcode, destination_id, generation, speed,
drivers/firewire/core-transaction.c
493
timer_destroy_on_stack(&t.split_timeout_timer);
drivers/firewire/core-transaction.c
57
struct fw_transaction *t, *tmp;
drivers/firewire/core-transaction.c
63
list_for_each_entry_safe(t, tmp, &card->transactions.list, link) {
drivers/firewire/core-transaction.c
64
if (try_cancel_split_timeout(t))
drivers/firewire/core-transaction.c
65
list_move(&t->link, &pending_list);
drivers/firewire/core-transaction.c
69
list_for_each_entry_safe(t, tmp, &pending_list, link) {
drivers/firewire/core-transaction.c
70
list_del(&t->link);
drivers/firewire/core-transaction.c
72
if (!t->with_tstamp) {
drivers/firewire/core-transaction.c
73
t->callback.without_tstamp(card, RCODE_CANCELLED, NULL, 0,
drivers/firewire/core-transaction.c
74
t->callback_data);
drivers/firewire/core-transaction.c
76
t->callback.with_tstamp(card, RCODE_CANCELLED, t->packet.timestamp, 0,
drivers/firewire/core-transaction.c
77
NULL, 0, t->callback_data);
drivers/firewire/core-transaction.c
85
struct fw_transaction *iter, *t = NULL; \
drivers/firewire/core-transaction.c
88
t = iter; \
drivers/firewire/core-transaction.c
92
if (t && try_cancel_split_timeout(t)) \
drivers/firewire/core-transaction.c
93
remove_transaction_entry(card, t); \
drivers/firewire/core-transaction.c
94
t; \
drivers/firewire/sbp2.c
257
struct fw_transaction t;
drivers/firewire/sbp2.c
508
fw_send_request(device->card, &orb->t, TCODE_WRITE_BLOCK_REQUEST,
drivers/firewire/sbp2.c
527
if (fw_cancel_transaction(device->card, &orb->t) == 0)
drivers/firewire/sbp2.c
667
struct fw_transaction *t;
drivers/firewire/sbp2.c
670
t = kmalloc_obj(*t, GFP_ATOMIC);
drivers/firewire/sbp2.c
671
if (t == NULL)
drivers/firewire/sbp2.c
674
fw_send_request(device->card, t, TCODE_WRITE_QUADLET_REQUEST,
drivers/firewire/sbp2.c
677
&d, 4, complete_agent_reset_write_no_wait, t);
drivers/firmware/arm_scmi/base.c
104
struct scmi_xfer *t;
drivers/firmware/arm_scmi/base.c
117
ret = ph->xops->xfer_get_init(ph, cmd, 0, size, &t);
drivers/firmware/arm_scmi/base.c
121
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/base.c
123
strscpy(vendor_id, t->rx.buf, size);
drivers/firmware/arm_scmi/base.c
125
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/base.c
144
struct scmi_xfer *t;
drivers/firmware/arm_scmi/base.c
148
0, sizeof(*impl_ver), &t);
drivers/firmware/arm_scmi/base.c
152
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/base.c
154
impl_ver = t->rx.buf;
drivers/firmware/arm_scmi/base.c
158
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/base.c
178
struct scmi_xfer *t;
drivers/firmware/arm_scmi/base.c
185
sizeof(*num_skip), 0, &t);
drivers/firmware/arm_scmi/base.c
189
num_skip = t->tx.buf;
drivers/firmware/arm_scmi/base.c
190
num_ret = t->rx.buf;
drivers/firmware/arm_scmi/base.c
191
list = t->rx.buf + sizeof(*num_ret);
drivers/firmware/arm_scmi/base.c
200
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/base.c
214
if (t->rx.len < (sizeof(u32) * 2)) {
drivers/firmware/arm_scmi/base.c
216
t->rx.len);
drivers/firmware/arm_scmi/base.c
221
real_list_sz = t->rx.len - sizeof(u32);
drivers/firmware/arm_scmi/base.c
243
ph->xops->reset_rx_to_maxsz(ph, t);
drivers/firmware/arm_scmi/base.c
246
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/base.c
268
struct scmi_xfer *t;
drivers/firmware/arm_scmi/base.c
271
sizeof(__le32), sizeof(*agent_info), &t);
drivers/firmware/arm_scmi/base.c
275
put_unaligned_le32(id, t->tx.buf);
drivers/firmware/arm_scmi/base.c
277
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/base.c
279
agent_info = t->rx.buf;
drivers/firmware/arm_scmi/base.c
283
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/base.c
293
struct scmi_xfer *t;
drivers/firmware/arm_scmi/base.c
297
sizeof(*cfg), 0, &t);
drivers/firmware/arm_scmi/base.c
301
cfg = t->tx.buf;
drivers/firmware/arm_scmi/base.c
304
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/base.c
306
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/base.c
69
struct scmi_xfer *t;
drivers/firmware/arm_scmi/base.c
74
0, sizeof(*attr_info), &t);
drivers/firmware/arm_scmi/base.c
78
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/base.c
80
attr_info = t->rx.buf;
drivers/firmware/arm_scmi/base.c
85
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/clock.c
195
struct scmi_xfer *t;
drivers/firmware/arm_scmi/clock.c
199
0, sizeof(*attr), &t);
drivers/firmware/arm_scmi/clock.c
203
attr = t->rx.buf;
drivers/firmware/arm_scmi/clock.c
205
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/clock.c
211
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/clock.c
322
struct scmi_xfer *t;
drivers/firmware/arm_scmi/clock.c
327
sizeof(clk_id), sizeof(perm), &t);
drivers/firmware/arm_scmi/clock.c
331
put_unaligned_le32(clk_id, t->tx.buf);
drivers/firmware/arm_scmi/clock.c
333
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/clock.c
335
perm = get_unaligned_le32(t->rx.buf);
drivers/firmware/arm_scmi/clock.c
342
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/clock.c
352
struct scmi_xfer *t;
drivers/firmware/arm_scmi/clock.c
357
sizeof(clk_id), sizeof(*attr), &t);
drivers/firmware/arm_scmi/clock.c
361
put_unaligned_le32(clk_id, t->tx.buf);
drivers/firmware/arm_scmi/clock.c
362
attr = t->rx.buf;
drivers/firmware/arm_scmi/clock.c
364
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/clock.c
376
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/clock.c
555
struct scmi_xfer *t;
drivers/firmware/arm_scmi/clock.c
558
sizeof(__le32), sizeof(u64), &t);
drivers/firmware/arm_scmi/clock.c
562
put_unaligned_le32(clk_id, t->tx.buf);
drivers/firmware/arm_scmi/clock.c
564
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/clock.c
566
*value = get_unaligned_le64(t->rx.buf);
drivers/firmware/arm_scmi/clock.c
568
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/clock.c
577
struct scmi_xfer *t;
drivers/firmware/arm_scmi/clock.c
589
ret = ph->xops->xfer_get_init(ph, CLOCK_RATE_SET, sizeof(*cfg), 0, &t);
drivers/firmware/arm_scmi/clock.c
597
cfg = t->tx.buf;
drivers/firmware/arm_scmi/clock.c
604
ret = ph->xops->do_xfer_with_response(ph, t);
drivers/firmware/arm_scmi/clock.c
608
resp = t->rx.buf;
drivers/firmware/arm_scmi/clock.c
617
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/clock.c
623
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/clock.c
634
struct scmi_xfer *t;
drivers/firmware/arm_scmi/clock.c
641
sizeof(*cfg), 0, &t);
drivers/firmware/arm_scmi/clock.c
645
t->hdr.poll_completion = atomic;
drivers/firmware/arm_scmi/clock.c
647
cfg = t->tx.buf;
drivers/firmware/arm_scmi/clock.c
651
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/clock.c
653
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/clock.c
662
struct scmi_xfer *t;
drivers/firmware/arm_scmi/clock.c
678
sizeof(*cfg), 0, &t);
drivers/firmware/arm_scmi/clock.c
682
t->hdr.poll_completion = false;
drivers/firmware/arm_scmi/clock.c
684
cfg = t->tx.buf;
drivers/firmware/arm_scmi/clock.c
688
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/clock.c
690
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/clock.c
700
struct scmi_xfer *t;
drivers/firmware/arm_scmi/clock.c
703
sizeof(__le32), sizeof(u32), &t);
drivers/firmware/arm_scmi/clock.c
707
put_unaligned_le32(clk_id, t->tx.buf);
drivers/firmware/arm_scmi/clock.c
709
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/clock.c
711
*parent_id = get_unaligned_le32(t->rx.buf);
drivers/firmware/arm_scmi/clock.c
713
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/clock.c
726
struct scmi_xfer *t;
drivers/firmware/arm_scmi/clock.c
734
sizeof(*cfg), 0, &t);
drivers/firmware/arm_scmi/clock.c
738
t->hdr.poll_completion = atomic;
drivers/firmware/arm_scmi/clock.c
743
cfg = t->tx.buf;
drivers/firmware/arm_scmi/clock.c
751
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/clock.c
753
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/clock.c
799
struct scmi_xfer *t;
drivers/firmware/arm_scmi/clock.c
803
sizeof(*cfg), 0, &t);
drivers/firmware/arm_scmi/clock.c
807
t->hdr.poll_completion = atomic;
drivers/firmware/arm_scmi/clock.c
811
cfg = t->tx.buf;
drivers/firmware/arm_scmi/clock.c
815
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/clock.c
817
struct scmi_msg_resp_clock_config_get *resp = t->rx.buf;
drivers/firmware/arm_scmi/clock.c
829
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/clock.c
840
struct scmi_xfer *t;
drivers/firmware/arm_scmi/clock.c
847
sizeof(clk_id), sizeof(*resp), &t);
drivers/firmware/arm_scmi/clock.c
851
t->hdr.poll_completion = atomic;
drivers/firmware/arm_scmi/clock.c
852
put_unaligned_le32(clk_id, t->tx.buf);
drivers/firmware/arm_scmi/clock.c
853
resp = t->rx.buf;
drivers/firmware/arm_scmi/clock.c
855
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/clock.c
859
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/clock.c
974
struct scmi_xfer *t;
drivers/firmware/arm_scmi/clock.c
977
ret = ph->xops->xfer_get_init(ph, message_id, sizeof(*notify), 0, &t);
drivers/firmware/arm_scmi/clock.c
981
notify = t->tx.buf;
drivers/firmware/arm_scmi/clock.c
985
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/clock.c
987
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/driver.c
1609
struct scmi_xfer *t;
drivers/firmware/arm_scmi/driver.c
1611
ret = xfer_get_init(ph, PROTOCOL_VERSION, 0, sizeof(*version), &t);
drivers/firmware/arm_scmi/driver.c
1615
ret = do_xfer(ph, t);
drivers/firmware/arm_scmi/driver.c
1617
rev_info = t->rx.buf;
drivers/firmware/arm_scmi/driver.c
1621
xfer_put(ph, t);
drivers/firmware/arm_scmi/driver.c
1688
struct scmi_xfer *t;
drivers/firmware/arm_scmi/driver.c
1692
ret = ph->xops->xfer_get_init(ph, cmd_id, txlen, sizeof(*resp), &t);
drivers/firmware/arm_scmi/driver.c
1696
put_unaligned_le32(res_id, t->tx.buf);
drivers/firmware/arm_scmi/driver.c
1698
put_unaligned_le32(*flags, t->tx.buf + sizeof(res_id));
drivers/firmware/arm_scmi/driver.c
1699
resp = t->rx.buf;
drivers/firmware/arm_scmi/driver.c
1701
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/driver.c
1705
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/driver.c
1745
struct scmi_xfer *t;
drivers/firmware/arm_scmi/driver.c
1748
sizeof(__le32), 0, &t);
drivers/firmware/arm_scmi/driver.c
1752
put_unaligned_le32(message_id, t->tx.buf);
drivers/firmware/arm_scmi/driver.c
1753
ret = do_xfer(ph, t);
drivers/firmware/arm_scmi/driver.c
1755
*attributes = get_unaligned_le32(t->rx.buf);
drivers/firmware/arm_scmi/driver.c
1756
xfer_put(ph, t);
drivers/firmware/arm_scmi/driver.c
1779
struct scmi_xfer *t;
drivers/firmware/arm_scmi/driver.c
1802
ret = ph->xops->xfer_get_init(ph, msg_id, tx_size, 0, &i->t);
drivers/firmware/arm_scmi/driver.c
1809
i->msg = i->t->tx.buf;
drivers/firmware/arm_scmi/driver.c
1810
i->resp = i->t->rx.buf;
drivers/firmware/arm_scmi/driver.c
1832
ret = ph->xops->do_xfer(ph, i->t);
drivers/firmware/arm_scmi/driver.c
1836
st->rx_len = i->t->rx.len;
drivers/firmware/arm_scmi/driver.c
1857
ph->xops->reset_rx_to_maxsz(ph, i->t);
drivers/firmware/arm_scmi/driver.c
1866
ph->xops->xfer_put(ph, i->t);
drivers/firmware/arm_scmi/driver.c
1912
struct scmi_xfer *t;
drivers/firmware/arm_scmi/driver.c
1934
sizeof(*info), sizeof(*resp), &t);
drivers/firmware/arm_scmi/driver.c
1938
info = t->tx.buf;
drivers/firmware/arm_scmi/driver.c
1947
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/driver.c
1951
resp = t->rx.buf;
drivers/firmware/arm_scmi/driver.c
1998
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/driver.c
2013
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/driver.c
2088
struct scmi_xfer *t;
drivers/firmware/arm_scmi/driver.c
2098
sizeof(__le32), 0, &t);
drivers/firmware/arm_scmi/driver.c
2102
put_unaligned_le32(pi->proto->supported_version, t->tx.buf);
drivers/firmware/arm_scmi/driver.c
2103
ret = do_xfer(ph, t);
drivers/firmware/arm_scmi/driver.c
2107
xfer_put(ph, t);
drivers/firmware/arm_scmi/perf.c
199
struct scmi_xfer *t;
drivers/firmware/arm_scmi/perf.c
203
sizeof(*attr), &t);
drivers/firmware/arm_scmi/perf.c
207
attr = t->rx.buf;
drivers/firmware/arm_scmi/perf.c
209
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/perf.c
226
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/perf.c
257
struct scmi_xfer *t;
drivers/firmware/arm_scmi/perf.c
261
sizeof(dom_info->id), sizeof(*attr), &t);
drivers/firmware/arm_scmi/perf.c
265
put_unaligned_le32(dom_info->id, t->tx.buf);
drivers/firmware/arm_scmi/perf.c
266
attr = t->rx.buf;
drivers/firmware/arm_scmi/perf.c
268
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/perf.c
318
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/perf.c
528
struct scmi_xfer *t;
drivers/firmware/arm_scmi/perf.c
532
sizeof(*limits), 0, &t);
drivers/firmware/arm_scmi/perf.c
536
limits = t->tx.buf;
drivers/firmware/arm_scmi/perf.c
541
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/perf.c
543
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/perf.c
607
struct scmi_xfer *t;
drivers/firmware/arm_scmi/perf.c
611
sizeof(__le32), 0, &t);
drivers/firmware/arm_scmi/perf.c
615
put_unaligned_le32(domain, t->tx.buf);
drivers/firmware/arm_scmi/perf.c
617
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/perf.c
619
limits = t->rx.buf;
drivers/firmware/arm_scmi/perf.c
625
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/perf.c
683
struct scmi_xfer *t;
drivers/firmware/arm_scmi/perf.c
686
ret = ph->xops->xfer_get_init(ph, PERF_LEVEL_SET, sizeof(*lvl), 0, &t);
drivers/firmware/arm_scmi/perf.c
690
t->hdr.poll_completion = poll;
drivers/firmware/arm_scmi/perf.c
691
lvl = t->tx.buf;
drivers/firmware/arm_scmi/perf.c
695
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/perf.c
697
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/perf.c
747
struct scmi_xfer *t;
drivers/firmware/arm_scmi/perf.c
750
sizeof(u32), sizeof(u32), &t);
drivers/firmware/arm_scmi/perf.c
754
t->hdr.poll_completion = poll;
drivers/firmware/arm_scmi/perf.c
755
put_unaligned_le32(domain, t->tx.buf);
drivers/firmware/arm_scmi/perf.c
757
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/perf.c
759
*level = get_unaligned_le32(t->rx.buf);
drivers/firmware/arm_scmi/perf.c
761
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/perf.c
811
struct scmi_xfer *t;
drivers/firmware/arm_scmi/perf.c
814
ret = ph->xops->xfer_get_init(ph, message_id, sizeof(*notify), 0, &t);
drivers/firmware/arm_scmi/perf.c
818
notify = t->tx.buf;
drivers/firmware/arm_scmi/perf.c
822
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/perf.c
824
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/pinctrl.c
132
struct scmi_xfer *t;
drivers/firmware/arm_scmi/pinctrl.c
135
ret = ph->xops->xfer_get_init(ph, PROTOCOL_ATTRIBUTES, 0, sizeof(*attr), &t);
drivers/firmware/arm_scmi/pinctrl.c
139
attr = t->rx.buf;
drivers/firmware/arm_scmi/pinctrl.c
141
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/pinctrl.c
152
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/pinctrl.c
195
struct scmi_xfer *t;
drivers/firmware/arm_scmi/pinctrl.c
208
sizeof(*rx), &t);
drivers/firmware/arm_scmi/pinctrl.c
212
tx = t->tx.buf;
drivers/firmware/arm_scmi/pinctrl.c
213
rx = t->rx.buf;
drivers/firmware/arm_scmi/pinctrl.c
217
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/pinctrl.c
227
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/pinctrl.c
469
struct scmi_xfer *t;
drivers/firmware/arm_scmi/pinctrl.c
491
chunk * 2 * sizeof(__le32), 0, &t);
drivers/firmware/arm_scmi/pinctrl.c
495
tx = t->tx.buf;
drivers/firmware/arm_scmi/pinctrl.c
509
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/pinctrl.c
511
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/pinctrl.c
528
struct scmi_xfer *t;
drivers/firmware/arm_scmi/pinctrl.c
537
sizeof(*tx), 0, &t);
drivers/firmware/arm_scmi/pinctrl.c
541
tx = t->tx.buf;
drivers/firmware/arm_scmi/pinctrl.c
547
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/pinctrl.c
548
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/pinctrl.c
559
struct scmi_xfer *t;
drivers/firmware/arm_scmi/pinctrl.c
572
ret = ph->xops->xfer_get_init(ph, cmd, sizeof(*tx), 0, &t);
drivers/firmware/arm_scmi/pinctrl.c
576
tx = t->tx.buf;
drivers/firmware/arm_scmi/pinctrl.c
580
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/pinctrl.c
581
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/power.c
115
struct scmi_xfer *t;
drivers/firmware/arm_scmi/power.c
119
sizeof(domain), sizeof(*attr), &t);
drivers/firmware/arm_scmi/power.c
123
put_unaligned_le32(domain, t->tx.buf);
drivers/firmware/arm_scmi/power.c
124
attr = t->rx.buf;
drivers/firmware/arm_scmi/power.c
126
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/power.c
137
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/power.c
157
struct scmi_xfer *t;
drivers/firmware/arm_scmi/power.c
160
ret = ph->xops->xfer_get_init(ph, POWER_STATE_SET, sizeof(*st), 0, &t);
drivers/firmware/arm_scmi/power.c
164
st = t->tx.buf;
drivers/firmware/arm_scmi/power.c
169
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/power.c
171
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/power.c
179
struct scmi_xfer *t;
drivers/firmware/arm_scmi/power.c
181
ret = ph->xops->xfer_get_init(ph, POWER_STATE_GET, sizeof(u32), sizeof(u32), &t);
drivers/firmware/arm_scmi/power.c
185
put_unaligned_le32(domain, t->tx.buf);
drivers/firmware/arm_scmi/power.c
187
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/power.c
189
*state = get_unaligned_le32(t->rx.buf);
drivers/firmware/arm_scmi/power.c
191
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/power.c
223
struct scmi_xfer *t;
drivers/firmware/arm_scmi/power.c
227
sizeof(*notify), 0, &t);
drivers/firmware/arm_scmi/power.c
231
notify = t->tx.buf;
drivers/firmware/arm_scmi/power.c
235
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/power.c
237
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/power.c
81
struct scmi_xfer *t;
drivers/firmware/arm_scmi/power.c
85
0, sizeof(*attr), &t);
drivers/firmware/arm_scmi/power.c
89
attr = t->rx.buf;
drivers/firmware/arm_scmi/power.c
91
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/power.c
99
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/powercap.c
145
struct scmi_xfer *t;
drivers/firmware/arm_scmi/powercap.c
148
sizeof(u32), &t);
drivers/firmware/arm_scmi/powercap.c
152
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/powercap.c
156
attributes = get_unaligned_le32(t->rx.buf);
drivers/firmware/arm_scmi/powercap.c
160
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/powercap.c
199
struct scmi_xfer *t;
drivers/firmware/arm_scmi/powercap.c
204
sizeof(domain), sizeof(*resp), &t);
drivers/firmware/arm_scmi/powercap.c
208
put_unaligned_le32(domain, t->tx.buf);
drivers/firmware/arm_scmi/powercap.c
209
resp = t->rx.buf;
drivers/firmware/arm_scmi/powercap.c
211
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/powercap.c
283
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/powercap.c
319
struct scmi_xfer *t;
drivers/firmware/arm_scmi/powercap.c
322
sizeof(u32), &t);
drivers/firmware/arm_scmi/powercap.c
326
put_unaligned_le32(domain_id, t->tx.buf);
drivers/firmware/arm_scmi/powercap.c
327
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/powercap.c
329
*power_cap = get_unaligned_le32(t->rx.buf);
drivers/firmware/arm_scmi/powercap.c
331
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/powercap.c
370
struct scmi_xfer *t;
drivers/firmware/arm_scmi/powercap.c
374
sizeof(*msg), 0, &t);
drivers/firmware/arm_scmi/powercap.c
378
msg = t->tx.buf;
drivers/firmware/arm_scmi/powercap.c
386
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/powercap.c
388
ret = ph->xops->do_xfer_with_response(ph, t);
drivers/firmware/arm_scmi/powercap.c
392
resp = t->rx.buf;
drivers/firmware/arm_scmi/powercap.c
403
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/powercap.c
470
struct scmi_xfer *t;
drivers/firmware/arm_scmi/powercap.c
473
sizeof(u32), &t);
drivers/firmware/arm_scmi/powercap.c
477
put_unaligned_le32(domain_id, t->tx.buf);
drivers/firmware/arm_scmi/powercap.c
478
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/powercap.c
480
*pai = get_unaligned_le32(t->rx.buf);
drivers/firmware/arm_scmi/powercap.c
482
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/powercap.c
511
struct scmi_xfer *t;
drivers/firmware/arm_scmi/powercap.c
515
sizeof(*msg), 0, &t);
drivers/firmware/arm_scmi/powercap.c
519
msg = t->tx.buf;
drivers/firmware/arm_scmi/powercap.c
524
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/powercap.c
526
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/powercap.c
558
struct scmi_xfer *t;
drivers/firmware/arm_scmi/powercap.c
567
sizeof(u32), sizeof(*resp), &t);
drivers/firmware/arm_scmi/powercap.c
571
resp = t->rx.buf;
drivers/firmware/arm_scmi/powercap.c
572
put_unaligned_le32(domain_id, t->tx.buf);
drivers/firmware/arm_scmi/powercap.c
573
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/powercap.c
579
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/powercap.c
747
struct scmi_xfer *t;
drivers/firmware/arm_scmi/powercap.c
755
sizeof(*notify), 0, &t);
drivers/firmware/arm_scmi/powercap.c
759
notify = t->tx.buf;
drivers/firmware/arm_scmi/powercap.c
789
sizeof(*notify), 0, &t);
drivers/firmware/arm_scmi/powercap.c
793
notify = t->tx.buf;
drivers/firmware/arm_scmi/powercap.c
804
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/powercap.c
806
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/reset.c
117
struct scmi_xfer *t;
drivers/firmware/arm_scmi/reset.c
122
sizeof(domain), sizeof(*attr), &t);
drivers/firmware/arm_scmi/reset.c
126
put_unaligned_le32(domain, t->tx.buf);
drivers/firmware/arm_scmi/reset.c
127
attr = t->rx.buf;
drivers/firmware/arm_scmi/reset.c
129
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/reset.c
143
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/reset.c
193
struct scmi_xfer *t;
drivers/firmware/arm_scmi/reset.c
204
ret = ph->xops->xfer_get_init(ph, RESET, sizeof(*dom), 0, &t);
drivers/firmware/arm_scmi/reset.c
208
dom = t->tx.buf;
drivers/firmware/arm_scmi/reset.c
214
ret = ph->xops->do_xfer_with_response(ph, t);
drivers/firmware/arm_scmi/reset.c
216
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/reset.c
218
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/reset.c
271
struct scmi_xfer *t;
drivers/firmware/arm_scmi/reset.c
274
ret = ph->xops->xfer_get_init(ph, RESET_NOTIFY, sizeof(*cfg), 0, &t);
drivers/firmware/arm_scmi/reset.c
278
cfg = t->tx.buf;
drivers/firmware/arm_scmi/reset.c
282
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/reset.c
284
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/reset.c
77
struct scmi_xfer *t;
drivers/firmware/arm_scmi/reset.c
81
0, sizeof(attr), &t);
drivers/firmware/arm_scmi/reset.c
85
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/reset.c
87
attr = get_unaligned_le32(t->rx.buf);
drivers/firmware/arm_scmi/reset.c
91
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/sensors.c
230
struct scmi_xfer *t;
drivers/firmware/arm_scmi/sensors.c
234
0, sizeof(*attr), &t);
drivers/firmware/arm_scmi/sensors.c
238
attr = t->rx.buf;
drivers/firmware/arm_scmi/sensors.c
240
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/sensors.c
249
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/sensors.c
716
struct scmi_xfer *t;
drivers/firmware/arm_scmi/sensors.c
719
ret = ph->xops->xfer_get_init(ph, message_id, sizeof(*cfg), 0, &t);
drivers/firmware/arm_scmi/sensors.c
723
cfg = t->tx.buf;
drivers/firmware/arm_scmi/sensors.c
727
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/sensors.c
729
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/sensors.c
756
struct scmi_xfer *t;
drivers/firmware/arm_scmi/sensors.c
760
sizeof(*trip), 0, &t);
drivers/firmware/arm_scmi/sensors.c
764
trip = t->tx.buf;
drivers/firmware/arm_scmi/sensors.c
770
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/sensors.c
772
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/sensors.c
780
struct scmi_xfer *t;
drivers/firmware/arm_scmi/sensors.c
787
sizeof(__le32), sizeof(__le32), &t);
drivers/firmware/arm_scmi/sensors.c
791
put_unaligned_le32(sensor_id, t->tx.buf);
drivers/firmware/arm_scmi/sensors.c
792
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/sensors.c
796
*sensor_config = get_unaligned_le64(t->rx.buf);
drivers/firmware/arm_scmi/sensors.c
800
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/sensors.c
808
struct scmi_xfer *t;
drivers/firmware/arm_scmi/sensors.c
816
sizeof(*msg), 0, &t);
drivers/firmware/arm_scmi/sensors.c
820
msg = t->tx.buf;
drivers/firmware/arm_scmi/sensors.c
824
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/sensors.c
831
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/sensors.c
854
struct scmi_xfer *t;
drivers/firmware/arm_scmi/sensors.c
863
sizeof(*sensor), 0, &t);
drivers/firmware/arm_scmi/sensors.c
867
sensor = t->tx.buf;
drivers/firmware/arm_scmi/sensors.c
872
ret = ph->xops->do_xfer_with_response(ph, t);
drivers/firmware/arm_scmi/sensors.c
876
resp = t->rx.buf;
drivers/firmware/arm_scmi/sensors.c
885
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/sensors.c
887
*value = get_unaligned_le64(t->rx.buf);
drivers/firmware/arm_scmi/sensors.c
890
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/sensors.c
922
struct scmi_xfer *t;
drivers/firmware/arm_scmi/sensors.c
936
sizeof(*sensor), 0, &t);
drivers/firmware/arm_scmi/sensors.c
940
sensor = t->tx.buf;
drivers/firmware/arm_scmi/sensors.c
944
ret = ph->xops->do_xfer_with_response(ph, t);
drivers/firmware/arm_scmi/sensors.c
949
resp = t->rx.buf;
drivers/firmware/arm_scmi/sensors.c
960
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/sensors.c
965
resp_readings = t->rx.buf;
drivers/firmware/arm_scmi/sensors.c
972
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/system.c
56
struct scmi_xfer *t;
drivers/firmware/arm_scmi/system.c
60
sizeof(*notify), 0, &t);
drivers/firmware/arm_scmi/system.c
64
notify = t->tx.buf;
drivers/firmware/arm_scmi/system.c
67
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/system.c
69
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
102
attr = t->rx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
104
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
110
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
119
struct scmi_xfer *t;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
125
sizeof(*rtc_notify), 0, &t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
129
rtc_notify = t->tx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
137
sizeof(*button_notify), 0, &t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
141
button_notify = t->tx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
147
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
149
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
239
struct scmi_xfer *t;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
245
ret = ph->xops->xfer_get_init(ph, IMX_BBM_RTC_TIME_SET, sizeof(*cfg), 0, &t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
249
cfg = t->tx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
255
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
257
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
267
struct scmi_xfer *t;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
274
sizeof(u64), &t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
278
cfg = t->tx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
282
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
284
*value = get_unaligned_le64(t->rx.buf);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
286
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
296
struct scmi_xfer *t;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
302
ret = ph->xops->xfer_get_init(ph, IMX_BBM_RTC_ALARM_SET, sizeof(*cfg), 0, &t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
306
cfg = t->tx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
313
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
315
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
322
struct scmi_xfer *t;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
325
ret = ph->xops->xfer_get_init(ph, IMX_BBM_BUTTON_GET, 0, sizeof(u32), &t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
329
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
331
*state = get_unaligned_le32(t->rx.buf);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
333
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
95
struct scmi_xfer *t;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
98
ret = ph->xops->xfer_get_init(ph, PROTOCOL_ATTRIBUTES, 0, sizeof(*attr), &t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
109
struct scmi_xfer *t;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
117
0, &t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
121
in = t->tx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
132
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
134
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
143
struct scmi_xfer *t;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
156
0, &t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
160
put_unaligned_le32(cpuid, t->tx.buf);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
161
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
163
out = t->rx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
169
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
184
struct scmi_xfer *t;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
188
sizeof(*attr), &t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
192
attr = t->rx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
194
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
201
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
211
struct scmi_xfer *t;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
214
ret = ph->xops->xfer_get_init(ph, SCMI_IMX_CPU_ATTRIBUTES, sizeof(u32), 0, &t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
218
put_unaligned_le32(cpuid, t->tx.buf);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
219
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
221
out = t->rx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
228
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
79
struct scmi_xfer *t;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
92
ret = ph->xops->xfer_get_init(ph, msg_id, sizeof(u32), 0, &t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
96
put_unaligned_le32(cpuid, t->tx.buf);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
97
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
99
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-lmm.c
101
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-lmm.c
109
struct scmi_xfer *t;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-lmm.c
122
ret = ph->xops->xfer_get_init(ph, msg_id, sizeof(u32), 0, &t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-lmm.c
126
put_unaligned_le32(lmid, t->tx.buf);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-lmm.c
127
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-lmm.c
129
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-lmm.c
138
struct scmi_xfer *t;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-lmm.c
142
0, &t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-lmm.c
146
in = t->tx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-lmm.c
152
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-lmm.c
154
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-lmm.c
163
struct scmi_xfer *t;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-lmm.c
171
0, &t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-lmm.c
175
in = t->tx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-lmm.c
181
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-lmm.c
183
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-lmm.c
199
struct scmi_xfer *t;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-lmm.c
203
sizeof(*attr), &t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-lmm.c
207
attr = t->rx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-lmm.c
209
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-lmm.c
221
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-lmm.c
80
struct scmi_xfer *t;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-lmm.c
83
ret = ph->xops->xfer_get_init(ph, SCMI_IMX_LMM_ATTRIBUTES, sizeof(u32), 0, &t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-lmm.c
87
put_unaligned_le32(lmid, t->tx.buf);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-lmm.c
88
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-lmm.c
90
out = t->rx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
109
struct scmi_xfer *t;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
113
sizeof(*attr), &t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
117
attr = t->rx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
119
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
128
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
154
struct scmi_xfer *t;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
162
sizeof(*in), 0, &t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
166
in = t->tx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
170
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
172
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
243
struct scmi_xfer *t;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
253
0, &t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
257
put_unaligned_le32(ctrl_id, t->tx.buf);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
258
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
260
out = t->rx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
264
*num * sizeof(__le32) > t->rx.len - sizeof(__le32)) {
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
265
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
273
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
282
struct scmi_xfer *t;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
295
sizeof(*in) + num * sizeof(__le32), 0, &t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
299
in = t->tx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
305
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
307
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
316
struct scmi_xfer *t;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
320
sizeof(*out), &t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
324
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
326
out = t->rx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
334
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
343
struct scmi_xfer *t;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
346
ret = ph->xops->xfer_get_init(ph, SCMI_IMX_MISC_BOARD_INFO, 0, sizeof(*out), &t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
350
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
352
out = t->rx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
358
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
367
struct scmi_xfer *t;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
370
ret = ph->xops->xfer_get_init(ph, SCMI_IMX_MISC_CFG_INFO_GET, 0, sizeof(*out), &t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
374
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
376
out = t->rx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
382
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/voltage.c
268
struct scmi_xfer *t;
drivers/firmware/arm_scmi/voltage.c
274
ret = ph->xops->xfer_get_init(ph, cmd_id, sizeof(__le32), 0, &t);
drivers/firmware/arm_scmi/voltage.c
278
put_unaligned_le32(domain_id, t->tx.buf);
drivers/firmware/arm_scmi/voltage.c
279
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/voltage.c
281
*value = get_unaligned_le32(t->rx.buf);
drivers/firmware/arm_scmi/voltage.c
283
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/voltage.c
291
struct scmi_xfer *t;
drivers/firmware/arm_scmi/voltage.c
299
sizeof(*cmd), 0, &t);
drivers/firmware/arm_scmi/voltage.c
303
cmd = t->tx.buf;
drivers/firmware/arm_scmi/voltage.c
307
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/voltage.c
309
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/voltage.c
326
struct scmi_xfer *t;
drivers/firmware/arm_scmi/voltage.c
335
sizeof(*cmd), 0, &t);
drivers/firmware/arm_scmi/voltage.c
341
cmd = t->tx.buf;
drivers/firmware/arm_scmi/voltage.c
347
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/voltage.c
350
ret = ph->xops->do_xfer_with_response(ph, t);
drivers/firmware/arm_scmi/voltage.c
354
resp = t->rx.buf;
drivers/firmware/arm_scmi/voltage.c
365
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scmi/voltage.c
77
struct scmi_xfer *t;
drivers/firmware/arm_scmi/voltage.c
80
sizeof(__le32), &t);
drivers/firmware/arm_scmi/voltage.c
84
ret = ph->xops->do_xfer(ph, t);
drivers/firmware/arm_scmi/voltage.c
87
NUM_VOLTAGE_DOMAINS(get_unaligned_le32(t->rx.buf));
drivers/firmware/arm_scmi/voltage.c
89
ph->xops->xfer_put(ph, t);
drivers/firmware/arm_scpi.c
360
struct scpi_xfer *t, *match = NULL;
drivers/firmware/arm_scpi.c
377
list_for_each_entry(t, &ch->rx_pending, node)
drivers/firmware/arm_scpi.c
378
if (CMD_XTRACT_UNIQ(t->cmd) == CMD_XTRACT_UNIQ(cmd)) {
drivers/firmware/arm_scpi.c
379
list_del(&t->node);
drivers/firmware/arm_scpi.c
380
match = t;
drivers/firmware/arm_scpi.c
428
struct scpi_xfer *t = msg;
drivers/firmware/arm_scpi.c
432
if (t->tx_buf) {
drivers/firmware/arm_scpi.c
434
memcpy_toio(ch->tx_payload, t->tx_buf, t->tx_len);
drivers/firmware/arm_scpi.c
436
memcpy_toio(mem->payload, t->tx_buf, t->tx_len);
drivers/firmware/arm_scpi.c
439
if (t->rx_buf) {
drivers/firmware/arm_scpi.c
442
t->cmd |= FIELD_PREP(CMD_TOKEN_ID_MASK, ch->token);
drivers/firmware/arm_scpi.c
444
list_add_tail(&t->node, &ch->rx_pending);
drivers/firmware/arm_scpi.c
449
iowrite32(t->cmd, &mem->command);
drivers/firmware/arm_scpi.c
454
struct scpi_xfer *t;
drivers/firmware/arm_scpi.c
461
t = list_first_entry(&ch->xfers_list, struct scpi_xfer, node);
drivers/firmware/arm_scpi.c
462
list_del(&t->node);
drivers/firmware/arm_scpi.c
464
return t;
drivers/firmware/arm_scpi.c
467
static void put_scpi_xfer(struct scpi_xfer *t, struct scpi_chan *ch)
drivers/firmware/arm_scpi.c
470
list_add_tail(&t->node, &ch->xfers_list);
drivers/firmware/dmi-id.c
112
char *t;
drivers/firmware/dmi-id.c
118
t = kmalloc(strlen(c) + 1, GFP_KERNEL);
drivers/firmware/dmi-id.c
119
if (!t)
drivers/firmware/dmi-id.c
121
ascii_filter(t, c);
drivers/firmware/dmi-id.c
122
l = scnprintf(p, left, ":%s%s", f->prefix, t);
drivers/firmware/dmi-id.c
123
kfree(t);
drivers/firmware/efi/libstub/efi-stub-helper.c
495
efi_config_table_t *t = (void *)tables;
drivers/firmware/efi/libstub/efi-stub-helper.c
497
if (efi_guidcmp(t->guid, guid) == 0)
drivers/firmware/efi/libstub/efi-stub-helper.c
498
return efi_table_attr(t, table);
drivers/firmware/iscsi_ibft.c
229
static int ibft_verify_hdr(char *t, struct ibft_hdr *hdr, int id, int length)
drivers/firmware/iscsi_ibft.c
234
"found %d instead!\n", t, id, hdr->id);
drivers/firmware/iscsi_ibft.c
240
"found %d instead!\n", t, length, hdr->length);
drivers/firmware/tegra/bpmp-debugfs.c
694
u32 d, t;
drivers/firmware/tegra/bpmp-debugfs.c
712
err = seqbuf_read_u32(seqbuf, &t);
drivers/firmware/tegra/bpmp-debugfs.c
719
if (t & DEBUGFS_S_ISDIR) {
drivers/firmware/tegra/bpmp-debugfs.c
729
mode = t & DEBUGFS_S_IRUSR ? S_IRUSR : 0;
drivers/firmware/tegra/bpmp-debugfs.c
730
mode |= t & DEBUGFS_S_IWUSR ? S_IWUSR : 0;
drivers/gpib/agilent_82357a/agilent_82357a.c
40
static void agilent_82357a_timeout_handler(struct timer_list *t)
drivers/gpib/agilent_82357a/agilent_82357a.c
42
struct agilent_82357a_priv *a_priv = timer_container_of(a_priv, t,
drivers/gpib/common/gpib_os.c
134
static void pseudo_irq_handler(struct timer_list *t)
drivers/gpib/common/gpib_os.c
136
struct gpib_pseudo_irq *pseudo_irq = timer_container_of(pseudo_irq, t,
drivers/gpib/common/gpib_os.c
86
static void watchdog_timeout(struct timer_list *t)
drivers/gpib/common/gpib_os.c
88
struct gpib_board *board = timer_container_of(board, t, timer);
drivers/gpib/common/iblib.c
574
static void wait_timeout(struct timer_list *t)
drivers/gpib/common/iblib.c
576
struct wait_info *winfo = timer_container_of(winfo, t, timer);
drivers/gpib/ni_usb/ni_usb_gpib.c
94
static void ni_usb_timeout_handler(struct timer_list *t)
drivers/gpib/ni_usb/ni_usb_gpib.c
96
struct ni_usb_priv *ni_priv = timer_container_of(ni_priv, t,
drivers/gpio/gpio-tegra.c
35
#define GPIO_CNF(t, x) (GPIO_REG(t, x) + 0x00)
drivers/gpio/gpio-tegra.c
36
#define GPIO_OE(t, x) (GPIO_REG(t, x) + 0x10)
drivers/gpio/gpio-tegra.c
37
#define GPIO_OUT(t, x) (GPIO_REG(t, x) + 0X20)
drivers/gpio/gpio-tegra.c
38
#define GPIO_IN(t, x) (GPIO_REG(t, x) + 0x30)
drivers/gpio/gpio-tegra.c
39
#define GPIO_INT_STA(t, x) (GPIO_REG(t, x) + 0x40)
drivers/gpio/gpio-tegra.c
40
#define GPIO_INT_ENB(t, x) (GPIO_REG(t, x) + 0x50)
drivers/gpio/gpio-tegra.c
41
#define GPIO_INT_LVL(t, x) (GPIO_REG(t, x) + 0x60)
drivers/gpio/gpio-tegra.c
42
#define GPIO_INT_CLR(t, x) (GPIO_REG(t, x) + 0x70)
drivers/gpio/gpio-tegra.c
43
#define GPIO_DBC_CNT(t, x) (GPIO_REG(t, x) + 0xF0)
drivers/gpio/gpio-tegra.c
46
#define GPIO_MSK_CNF(t, x) (GPIO_REG(t, x) + t->soc->upper_offset + 0x00)
drivers/gpio/gpio-tegra.c
47
#define GPIO_MSK_OE(t, x) (GPIO_REG(t, x) + t->soc->upper_offset + 0x10)
drivers/gpio/gpio-tegra.c
48
#define GPIO_MSK_OUT(t, x) (GPIO_REG(t, x) + t->soc->upper_offset + 0X20)
drivers/gpio/gpio-tegra.c
49
#define GPIO_MSK_DBC_EN(t, x) (GPIO_REG(t, x) + t->soc->upper_offset + 0x30)
drivers/gpio/gpio-tegra.c
50
#define GPIO_MSK_INT_STA(t, x) (GPIO_REG(t, x) + t->soc->upper_offset + 0x40)
drivers/gpio/gpio-tegra.c
51
#define GPIO_MSK_INT_ENB(t, x) (GPIO_REG(t, x) + t->soc->upper_offset + 0x50)
drivers/gpio/gpio-tegra.c
52
#define GPIO_MSK_INT_LVL(t, x) (GPIO_REG(t, x) + t->soc->upper_offset + 0x60)
drivers/gpu/drm/amd/amdgpu/amdgpu_connectors.c
606
struct drm_display_mode *t, *mode;
drivers/gpu/drm/amd/amdgpu/amdgpu_connectors.c
609
list_for_each_entry_safe(mode, t, &connector->probed_modes, head) {
drivers/gpu/drm/amd/amdgpu/amdgpu_connectors.c
619
list_for_each_entry_safe(mode, t, &connector->probed_modes, head) {
drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c
278
static void amdgpu_fence_fallback(struct timer_list *t)
drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c
280
struct amdgpu_ring *ring = timer_container_of(ring, t,
drivers/gpu/drm/amd/amdgpu/amdgpu_gart.c
304
unsigned t;
drivers/gpu/drm/amd/amdgpu/amdgpu_gart.c
317
t = offset / AMDGPU_GPU_PAGE_SIZE;
drivers/gpu/drm/amd/amdgpu/amdgpu_gart.c
323
for (j = 0; j < AMDGPU_GPU_PAGES_IN_CPU_PAGE; j++, t++) {
drivers/gpu/drm/amd/amdgpu/amdgpu_gart.c
325
t, page_base, flags);
drivers/gpu/drm/amd/amdgpu/amdgpu_gart.c
352
unsigned i, j, t;
drivers/gpu/drm/amd/amdgpu/amdgpu_gart.c
358
t = offset / AMDGPU_GPU_PAGE_SIZE;
drivers/gpu/drm/amd/amdgpu/amdgpu_gart.c
362
for (j = 0; j < AMDGPU_GPU_PAGES_IN_CPU_PAGE; j++, t++) {
drivers/gpu/drm/amd/amdgpu/amdgpu_gart.c
363
amdgpu_gmc_set_pte_pde(adev, dst, t, page_base, flags);
drivers/gpu/drm/amd/amdgpu/amdgpu_gart.c
424
unsigned int i, j, t;
drivers/gpu/drm/amd/amdgpu/amdgpu_gart.c
435
t = offset / AMDGPU_GPU_PAGE_SIZE;
drivers/gpu/drm/amd/amdgpu/amdgpu_gart.c
439
for (j = 0; j < AMDGPU_GPU_PAGES_IN_CPU_PAGE; j++, t++) {
drivers/gpu/drm/amd/amdgpu/amdgpu_gart.c
441
amdgpu_gmc_set_pte_pde(adev, dst, t, page_base, flags);
drivers/gpu/drm/amd/amdgpu/amdgpu_gart.c
443
amdgpu_gmc_set_pte_pde(adev, dst, t, page_base, ctrl_flags);
drivers/gpu/drm/amd/amdgpu/amdgpu_ring.h
433
#define amdgpu_ring_test_ib(r, t) ((r)->funcs->test_ib ? (r)->funcs->test_ib((r), (t)) : 0)
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
136
static void amdgpu_mux_resubmit_fallback(struct timer_list *t)
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
138
struct amdgpu_ring_mux *mux = timer_container_of(mux, t,
drivers/gpu/drm/amd/amdgpu/amdgpu_sdma.h
191
#define amdgpu_emit_copy_buffer(adev, ib, s, d, b, t) (adev)->mman.buffer_funcs->emit_copy_buffer((ib), (s), (d), (b), (t))
drivers/gpu/drm/amd/amdgpu/amdgpu_vf_error.h
28
#define AMDGIM_ERROR_CODE(t,c) (((t&0xF)<<12)|(c&0xFFF))
drivers/gpu/drm/amd/amdgpu/sid.h
557
#define DMA_PACKET(cmd, b, t, s, n) ((((cmd) & 0xF) << 28) | \
drivers/gpu/drm/amd/amdgpu/sid.h
559
(((t) & 0x1) << 23) | \
drivers/gpu/drm/amd/amdgpu/umc_v8_10.c
172
uint32_t t = 0;
drivers/gpu/drm/amd/amdgpu/umc_v8_10.c
174
for (t = 0; t < ARRAY_SIZE(umc_v8_10_channelnum_map_colbit_table); t++)
drivers/gpu/drm/amd/amdgpu/umc_v8_10.c
175
if (channel_num == umc_v8_10_channelnum_map_colbit_table[t].channel_num)
drivers/gpu/drm/amd/amdgpu/umc_v8_10.c
176
return umc_v8_10_channelnum_map_colbit_table[t].col_bit;
drivers/gpu/drm/amd/display/dc/bios/command_table_helper.c
126
enum transmitter t)
drivers/gpu/drm/amd/display/dc/bios/command_table_helper.c
128
switch (t) {
drivers/gpu/drm/amd/display/dc/bios/command_table_helper.c
297
uint8_t phy_id_to_atom(enum transmitter t)
drivers/gpu/drm/amd/display/dc/bios/command_table_helper.c
301
switch (t) {
drivers/gpu/drm/amd/display/dc/bios/command_table_helper.h
58
enum transmitter t);
drivers/gpu/drm/amd/display/dc/bios/command_table_helper.h
63
uint8_t phy_id_to_atom(enum transmitter t);
drivers/gpu/drm/amd/display/dc/bios/command_table_helper2.c
150
enum transmitter t)
drivers/gpu/drm/amd/display/dc/bios/command_table_helper2.c
152
switch (t) {
drivers/gpu/drm/amd/display/dc/bios/command_table_helper2.h
53
enum transmitter t);
drivers/gpu/drm/amd/display/dc/bios/command_table_helper_struct.h
51
uint8_t (*transmitter_bp_to_atom)(enum transmitter t);
drivers/gpu/drm/amd/display/dc/bios/command_table_helper_struct.h
58
uint8_t (*phy_id_to_atom)(enum transmitter t);
drivers/gpu/drm/amd/display/dc/dio/dcn31/dcn31_dio_link_encoder.c
70
static uint8_t phy_id_from_transmitter(enum transmitter t)
drivers/gpu/drm/amd/display/dc/dio/dcn31/dcn31_dio_link_encoder.c
74
switch (t) {
drivers/gpu/drm/amd/display/dc/dio/dcn32/dcn32_dio_link_encoder.c
65
static uint8_t phy_id_from_transmitter(enum transmitter t)
drivers/gpu/drm/amd/display/dc/dio/dcn32/dcn32_dio_link_encoder.c
69
switch (t) {
drivers/gpu/drm/amd/display/modules/color/color_gamma.c
1005
t = dc_fixpt_div(dc_fixpt_sub(E1, ks),
drivers/gpu/drm/amd/display/modules/color/color_gamma.c
1008
t = dc_fixpt_zero;
drivers/gpu/drm/amd/display/modules/color/color_gamma.c
1013
t2 = dc_fixpt_mul(t, t);
drivers/gpu/drm/amd/display/modules/color/color_gamma.c
1014
t3 = dc_fixpt_mul(t2, t);
drivers/gpu/drm/amd/display/modules/color/color_gamma.c
1031
dc_fixpt_add(t, dc_fixpt_sub(t3, temp1))));
drivers/gpu/drm/amd/display/modules/color/color_gamma.c
977
struct fixed31_32 t;
drivers/gpu/drm/amd/pm/legacy-dpm/kv_dpm.c
389
data |= ((values->t << local_cac_reg->t_shift) &
drivers/gpu/drm/amd/pm/legacy-dpm/kv_dpm.h
81
u32 t;
drivers/gpu/drm/amd/pm/legacy-dpm/si_dpm.c
1875
u16 v, s32 t, u32 ileakage, u32 *leakage)
drivers/gpu/drm/amd/pm/legacy-dpm/si_dpm.c
1883
temperature = div64_s64(drm_int2fixp(t), 1000);
drivers/gpu/drm/amd/pm/legacy-dpm/si_dpm.c
1904
s32 t,
drivers/gpu/drm/amd/pm/legacy-dpm/si_dpm.c
1908
si_calculate_leakage_for_v_and_t_formula(coeff, v, t, i_leakage, leakage);
drivers/gpu/drm/amd/pm/legacy-dpm/si_dpm.c
2689
s32 t;
drivers/gpu/drm/amd/pm/legacy-dpm/si_dpm.c
2697
t = (1000 * (i * t_step + t0));
drivers/gpu/drm/amd/pm/legacy-dpm/si_dpm.c
2705
t,
drivers/gpu/drm/amd/pm/legacy-dpm/si_dpm.c
3369
static int r600_calculate_at(u32 t, u32 h, u32 fh, u32 fl, u32 *tl, u32 *th)
drivers/gpu/drm/amd/pm/legacy-dpm/si_dpm.c
3378
t1 = (t * (k - 100));
drivers/gpu/drm/amd/pm/legacy-dpm/si_dpm.c
3381
ah = ((a * t) + 5000) / 10000;
drivers/gpu/drm/amd/pm/legacy-dpm/si_dpm.c
3384
*th = t - ah;
drivers/gpu/drm/amd/pm/legacy-dpm/si_dpm.c
3385
*tl = t + al;
drivers/gpu/drm/bridge/tda998x_drv.c
752
static void tda998x_edid_delay_done(struct timer_list *t)
drivers/gpu/drm/bridge/tda998x_drv.c
754
struct tda998x_priv *priv = timer_container_of(priv, t,
drivers/gpu/drm/drm_connector.c
761
struct drm_display_mode *mode, *t;
drivers/gpu/drm/drm_connector.c
782
list_for_each_entry_safe(mode, t, &connector->probed_modes, head)
drivers/gpu/drm/drm_connector.c
785
list_for_each_entry_safe(mode, t, &connector->modes, head)
drivers/gpu/drm/drm_drv.c
136
#define DRM_MINOR_LIMIT(t) ({ \
drivers/gpu/drm/drm_drv.c
137
typeof(t) _t = (t); \
drivers/gpu/drm/drm_edid.c
2981
#define MODE_REFRESH_DIFF(c,t) (abs((c) - (t)))
drivers/gpu/drm/drm_edid.c
2989
struct drm_display_mode *t, *cur_mode, *preferred_mode;
drivers/gpu/drm/drm_edid.c
3004
list_for_each_entry_safe(cur_mode, t, &connector->probed_modes, head) {
drivers/gpu/drm/drm_edid.c
3359
const struct std_timing *t)
drivers/gpu/drm/drm_edid.c
3365
unsigned aspect_ratio = (t->vfreq_aspect & EDID_TIMING_ASPECT_MASK)
drivers/gpu/drm/drm_edid.c
3367
unsigned vfreq = (t->vfreq_aspect & EDID_TIMING_VFREQ_MASK)
drivers/gpu/drm/drm_edid.c
3371
if (bad_std_timing(t->hsize, t->vfreq_aspect))
drivers/gpu/drm/drm_edid.c
3375
hsize = t->hsize * 8 + 248;
drivers/gpu/drm/drm_edid.c
3609
const struct edid *edid, const u8 *t)
drivers/gpu/drm/drm_edid.c
3613
hmin = t[7];
drivers/gpu/drm/drm_edid.c
3615
hmin += ((t[4] & 0x04) ? 255 : 0);
drivers/gpu/drm/drm_edid.c
3616
hmax = t[8];
drivers/gpu/drm/drm_edid.c
3618
hmax += ((t[4] & 0x08) ? 255 : 0);
drivers/gpu/drm/drm_edid.c
3626
const struct edid *edid, const u8 *t)
drivers/gpu/drm/drm_edid.c
3630
vmin = t[5];
drivers/gpu/drm/drm_edid.c
3632
vmin += ((t[4] & 0x01) ? 255 : 0);
drivers/gpu/drm/drm_edid.c
3633
vmax = t[6];
drivers/gpu/drm/drm_edid.c
3635
vmax += ((t[4] & 0x02) ? 255 : 0);
drivers/gpu/drm/drm_edid.c
3642
range_pixel_clock(const struct edid *edid, const u8 *t)
drivers/gpu/drm/drm_edid.c
3645
if (t[9] == 0 || t[9] == 255)
drivers/gpu/drm/drm_edid.c
3649
if (edid->revision >= 4 && t[10] == DRM_EDID_CVT_SUPPORT_FLAG)
drivers/gpu/drm/drm_edid.c
3650
return (t[9] * 10000) - ((t[12] >> 2) * 250);
drivers/gpu/drm/drm_edid.c
3653
return t[9] * 10000 + 5001;
drivers/gpu/drm/drm_edid.c
3662
const u8 *t = (const u8 *)timing;
drivers/gpu/drm/drm_edid.c
3664
if (!mode_in_hsync_range(mode, edid, t))
drivers/gpu/drm/drm_edid.c
3667
if (!mode_in_vsync_range(mode, edid, t))
drivers/gpu/drm/drm_edid.c
3670
max_clock = range_pixel_clock(edid, t);
drivers/gpu/drm/drm_edid.c
3676
if (edid->revision >= 4 && t[10] == DRM_EDID_CVT_SUPPORT_FLAG)
drivers/gpu/drm/drm_edid.c
3677
if (t[13] && mode->hdisplay > 8 * (t[13] + (256 * (t[12]&0x3))))
drivers/gpu/drm/drm_modes.c
1809
struct drm_display_mode *mode, *t;
drivers/gpu/drm/drm_modes.c
1811
list_for_each_entry_safe(mode, t, mode_list, head) {
drivers/gpu/drm/drm_suballoc.c
362
long t;
drivers/gpu/drm/drm_suballoc.c
365
t = dma_fence_wait_any_timeout(fences, count, intr,
drivers/gpu/drm/drm_suballoc.c
371
r = (t > 0) ? 0 : t;
drivers/gpu/drm/drm_syncobj.c
1327
ktime_t t, *tp = NULL;
drivers/gpu/drm/drm_syncobj.c
1351
t = ns_to_ktime(args->deadline_nsec);
drivers/gpu/drm/drm_syncobj.c
1352
tp = &t;
drivers/gpu/drm/drm_syncobj.c
1370
ktime_t t, *tp = NULL;
drivers/gpu/drm/drm_syncobj.c
1395
t = ns_to_ktime(args->deadline_nsec);
drivers/gpu/drm/drm_syncobj.c
1396
tp = &t;
drivers/gpu/drm/drm_vblank.c
1344
struct drm_pending_vblank_event *e, *t;
drivers/gpu/drm/drm_vblank.c
1381
list_for_each_entry_safe(e, t, &dev->vblank_event_list, base.link) {
drivers/gpu/drm/drm_vblank.c
1880
struct drm_pending_vblank_event *e, *t;
drivers/gpu/drm/drm_vblank.c
1888
list_for_each_entry_safe(e, t, &dev->vblank_event_list, base.link) {
drivers/gpu/drm/drm_vblank.c
497
static void vblank_disable_fn(struct timer_list *t)
drivers/gpu/drm/drm_vblank.c
499
struct drm_vblank_crtc *vblank = timer_container_of(vblank, t,
drivers/gpu/drm/exynos/exynos_drm_vidi.c
161
static void vidi_fake_vblank_timer(struct timer_list *t)
drivers/gpu/drm/exynos/exynos_drm_vidi.c
163
struct vidi_context *ctx = timer_container_of(ctx, t, timer);
drivers/gpu/drm/gud/gud_pipe.c
239
static void gud_usb_bulk_timeout(struct timer_list *t)
drivers/gpu/drm/gud/gud_pipe.c
241
struct gud_usb_bulk_context *ctx = timer_container_of(ctx, t, timer);
drivers/gpu/drm/hyperv/hyperv_drm_proto.c
219
unsigned long t;
drivers/gpu/drm/hyperv/hyperv_drm_proto.c
228
t = wait_for_completion_timeout(&hv->wait, VMBUS_VSP_TIMEOUT);
drivers/gpu/drm/hyperv/hyperv_drm_proto.c
229
if (!t) {
drivers/gpu/drm/hyperv/hyperv_drm_proto.c
251
unsigned long t;
drivers/gpu/drm/hyperv/hyperv_drm_proto.c
262
t = wait_for_completion_timeout(&hv->wait, VMBUS_VSP_TIMEOUT);
drivers/gpu/drm/hyperv/hyperv_drm_proto.c
263
if (!t) {
drivers/gpu/drm/hyperv/hyperv_drm_proto.c
376
unsigned long t;
drivers/gpu/drm/hyperv/hyperv_drm_proto.c
388
t = wait_for_completion_timeout(&hv->wait, VMBUS_VSP_TIMEOUT);
drivers/gpu/drm/hyperv/hyperv_drm_proto.c
389
if (!t) {
drivers/gpu/drm/i915/display/intel_bios.c
306
const u16 *t = data_block + ptrs->ptr[i].fp_timing.offset +
drivers/gpu/drm/i915/display/intel_bios.c
309
if (*t != 0xffff)
drivers/gpu/drm/i915/display/intel_casf.c
172
static u32 casf_coeff(struct intel_crtc_state *crtc_state, int t)
drivers/gpu/drm/i915/display/intel_casf.c
177
value = crtc_state->hw.casf_params.coeff[t];
drivers/gpu/drm/i915/display/intel_casf.c
207
int t;
drivers/gpu/drm/i915/display/intel_casf.c
209
t = casf_coeff_tap(i);
drivers/gpu/drm/i915/display/intel_casf.c
210
tmp = casf_coeff(crtc_state, t);
drivers/gpu/drm/i915/display/intel_casf.c
212
t = casf_coeff_tap(i + 1);
drivers/gpu/drm/i915/display/intel_casf.c
213
tmp |= casf_coeff(crtc_state, t) << 16;
drivers/gpu/drm/i915/display/intel_display.h
541
#define assert_transcoder_enabled(d, t) assert_transcoder(d, t, true)
drivers/gpu/drm/i915/display/intel_display.h
542
#define assert_transcoder_disabled(d, t) assert_transcoder(d, t, false)
drivers/gpu/drm/i915/display/intel_psr_regs.h
167
#define EDP_MAX_SU_DISABLE_TIME(t) REG_FIELD_PREP(EDP_MAX_SU_DISABLE_TIME, (t))
drivers/gpu/drm/i915/display/intel_vbt_defs.h
704
struct dot_clock_override_entry_gen2 t[0];
drivers/gpu/drm/i915/display/skl_scaler.c
658
static u16 glk_nearest_filter_coef(int t)
drivers/gpu/drm/i915/display/skl_scaler.c
660
return t == 3 ? 0x0800 : 0x3000;
drivers/gpu/drm/i915/display/skl_scaler.c
712
int t;
drivers/gpu/drm/i915/display/skl_scaler.c
714
t = glk_coef_tap(i);
drivers/gpu/drm/i915/display/skl_scaler.c
715
tmp = glk_nearest_filter_coef(t);
drivers/gpu/drm/i915/display/skl_scaler.c
717
t = glk_coef_tap(i + 1);
drivers/gpu/drm/i915/display/skl_scaler.c
718
tmp |= glk_nearest_filter_coef(t) << 16;
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
143
static int prepare_blit(const struct tiled_blits *t,
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
158
struct intel_gt *gt = t->ce->engine->gt;
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
166
*cs++ = i915_mmio_reg_offset(BLIT_CCTL(t->ce->engine->mmio_base));
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
170
src_pitch = t->width; /* in dwords */
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
181
dst_pitch = t->width; /* in dwords */
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
196
*cs++ = t->height << 16 | t->width;
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
227
src_pitch = t->width * 4;
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
233
dst_pitch = t->width * 4;
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
242
*cs++ = t->height << 16 | t->width;
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
261
static void tiled_blits_destroy_buffers(struct tiled_blits *t)
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
265
for (i = 0; i < ARRAY_SIZE(t->buffers); i++)
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
266
i915_vma_put(t->buffers[i].vma);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
268
i915_vma_put(t->scratch.vma);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
269
i915_vma_put(t->batch);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
273
__create_vma(struct tiled_blits *t, size_t size, bool lmem)
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
275
struct drm_i915_private *i915 = t->ce->vm->i915;
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
286
vma = i915_vma_instance(obj, t->ce->vm, NULL);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
293
static struct i915_vma *create_vma(struct tiled_blits *t, bool lmem)
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
295
return __create_vma(t, PAGE_ALIGN(t->width * t->height * 4), lmem);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
298
static int tiled_blits_create_buffers(struct tiled_blits *t,
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
302
struct drm_i915_private *i915 = t->ce->engine->i915;
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
305
t->width = width;
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
306
t->height = height;
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
308
t->batch = __create_vma(t, PAGE_SIZE, false);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
309
if (IS_ERR(t->batch))
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
310
return PTR_ERR(t->batch);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
312
t->scratch.vma = create_vma(t, false);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
313
if (IS_ERR(t->scratch.vma)) {
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
314
i915_vma_put(t->batch);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
315
return PTR_ERR(t->scratch.vma);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
318
for (i = 0; i < ARRAY_SIZE(t->buffers); i++) {
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
321
vma = create_vma(t, HAS_LMEM(i915) && i % 2);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
323
tiled_blits_destroy_buffers(t);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
327
t->buffers[i].vma = vma;
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
328
t->buffers[i].tiling =
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
335
static void fill_scratch(struct tiled_blits *t, u32 *vaddr, u32 val)
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
339
t->scratch.start_val = val;
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
340
for (i = 0; i < t->width * t->height; i++)
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
343
i915_gem_object_flush_map(t->scratch.vma->obj);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
418
static int verify_buffer(const struct tiled_blits *t,
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
426
x = i915_prandom_u32_max_state(t->width, prng);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
427
y = i915_prandom_u32_max_state(t->height, prng);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
428
p = y * t->width + x;
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
438
p * 4, t->width * 4,
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
474
tiled_blit(struct tiled_blits *t,
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
493
err = i915_vma_pin(t->batch, 0, 0, PIN_USER | PIN_HIGH);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
499
err = prepare_blit(t, dst, src, t->batch->obj);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
503
rq = intel_context_create_request(t->ce);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
509
err = igt_vma_move_to_active_unlocked(t->batch, rq, 0);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
516
i915_vma_offset(t->batch),
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
517
i915_vma_size(t->batch),
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
527
i915_vma_unpin(t->batch);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
539
struct tiled_blits *t;
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
543
t = kzalloc_obj(*t);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
544
if (!t)
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
547
t->ce = intel_context_create(engine);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
548
if (IS_ERR(t->ce)) {
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
549
err = PTR_ERR(t->ce);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
553
t->align = i915_vm_min_alignment(t->ce->vm, INTEL_MEMORY_LOCAL);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
554
t->align = max(t->align,
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
555
i915_vm_min_alignment(t->ce->vm, INTEL_MEMORY_SYSTEM));
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
557
hole_size = 2 * round_up(WIDTH * HEIGHT * 4, t->align);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
559
hole_size += 2 * t->align; /* padding on either side */
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
561
mutex_lock(&t->ce->vm->mutex);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
563
err = drm_mm_insert_node_in_range(&t->ce->vm->mm, &hole,
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
564
hole_size, t->align,
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
570
mutex_unlock(&t->ce->vm->mutex);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
576
t->hole = hole.start + t->align;
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
577
pr_info("Using hole at %llx\n", t->hole);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
579
err = tiled_blits_create_buffers(t, WIDTH, HEIGHT, prng);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
583
return t;
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
586
intel_context_put(t->ce);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
588
kfree(t);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
592
static void tiled_blits_destroy(struct tiled_blits *t)
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
594
tiled_blits_destroy_buffers(t);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
596
intel_context_put(t->ce);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
597
kfree(t);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
600
static int tiled_blits_prepare(struct tiled_blits *t,
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
603
u64 offset = round_up(t->width * t->height * 4, t->align);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
608
map = i915_gem_object_pin_map_unlocked(t->scratch.vma->obj, I915_MAP_WC);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
613
for (i = 0; i < ARRAY_SIZE(t->buffers); i++) {
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
614
fill_scratch(t, map, prandom_u32_state(prng));
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
615
GEM_BUG_ON(verify_buffer(t, &t->scratch, prng));
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
617
err = tiled_blit(t,
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
618
&t->buffers[i], t->hole + offset,
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
619
&t->scratch, t->hole);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
621
err = verify_buffer(t, &t->buffers[i], prng);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
628
i915_gem_object_unpin_map(t->scratch.vma->obj);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
632
static int tiled_blits_bounce(struct tiled_blits *t, struct rnd_state *prng)
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
634
u64 offset = round_up(t->width * t->height * 4, 2 * t->align);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
639
err = tiled_blit(t,
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
640
&t->buffers[1], t->hole + offset / 2,
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
641
&t->buffers[0], t->hole + 2 * offset);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
646
t->buffers[2].tiling = t->buffers[0].tiling;
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
649
err = tiled_blit(t,
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
650
&t->buffers[2], t->hole + t->align,
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
651
&t->buffers[1], t->hole + 3 * offset / 2);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
655
err = verify_buffer(t, &t->buffers[2], prng);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
665
struct tiled_blits *t;
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
668
t = tiled_blits_create(engine, prng);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
669
if (IS_ERR(t))
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
670
return PTR_ERR(t);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
672
err = tiled_blits_prepare(t, prng);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
676
err = tiled_blits_bounce(t, prng);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
681
tiled_blits_destroy(t);
drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c
108
err = igt_live_test_begin(&t, i915, __func__, engine->name);
drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c
1377
struct igt_live_test t;
drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c
1394
err = igt_live_test_begin(&t, i915, __func__, "");
drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c
1484
if (igt_live_test_end(&t))
drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c
166
err = igt_live_test_end(&t);
drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c
1778
struct igt_live_test t;
drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c
1797
err = igt_live_test_begin(&t, i915, __func__, "");
drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c
1894
if (igt_live_test_end(&t))
drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c
383
struct igt_live_test t;
drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c
385
err = igt_live_test_begin(&t, i915, __func__, "");
drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c
39
struct igt_live_test t;
drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c
406
if (igt_live_test_end(&t)) {
drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c
678
struct igt_live_test t;
drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c
693
err = igt_live_test_begin(&t, i915, __func__, engine->name);
drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c
771
if (igt_live_test_end(&t))
drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c
790
struct igt_live_test t;
drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c
817
err = igt_live_test_begin(&t, i915, __func__, "");
drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c
906
if (igt_live_test_end(&t))
drivers/gpu/drm/i915/gt/intel_engine_cs.c
1866
struct tasklet_struct *t = &engine->sched_engine->tasklet;
drivers/gpu/drm/i915/gt/intel_engine_cs.c
1868
if (!t->callback)
drivers/gpu/drm/i915/gt/intel_engine_cs.c
1872
if (tasklet_trylock(t)) {
drivers/gpu/drm/i915/gt/intel_engine_cs.c
1874
if (__tasklet_is_enabled(t))
drivers/gpu/drm/i915/gt/intel_engine_cs.c
1875
t->callback(t);
drivers/gpu/drm/i915/gt/intel_engine_cs.c
1876
tasklet_unlock(t);
drivers/gpu/drm/i915/gt/intel_engine_cs.c
1882
tasklet_unlock_wait(t);
drivers/gpu/drm/i915/gt/intel_engine_cs.c
2064
static const char *repr_timer(const struct timer_list *t)
drivers/gpu/drm/i915/gt/intel_engine_cs.c
2066
if (!READ_ONCE(t->expires))
drivers/gpu/drm/i915/gt/intel_engine_cs.c
2069
if (timer_pending(t))
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
2412
const struct timer_list *t = &engine->execlists.preempt;
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
2417
if (!timer_expired(t))
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
2427
static void execlists_submission_tasklet(struct tasklet_struct *t)
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
2430
from_tasklet(sched_engine, t, tasklet);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
2534
#define execlists_kick(t, member) \
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
2535
__execlists_kick(container_of(t, struct intel_engine_execlists, member))
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
3134
static void nop_submission_tasklet(struct tasklet_struct *t)
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
3137
from_tasklet(sched_engine, t, tasklet);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
3806
static void virtual_submission_tasklet(struct tasklet_struct *t)
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
3809
from_tasklet(sched_engine, t, tasklet);
drivers/gpu/drm/i915/gt/intel_mocs.c
550
#define for_each_mocs(mocs, t, i) \
drivers/gpu/drm/i915/gt/intel_mocs.c
552
i < (t)->n_entries ? (mocs = get_entry_control((t), i)), 1 : 0;\
drivers/gpu/drm/i915/gt/intel_mocs.c
607
#define for_each_l3cc(l3cc, t, i) \
drivers/gpu/drm/i915/gt/intel_mocs.c
609
i < ((t)->n_entries + 1) / 2 ? \
drivers/gpu/drm/i915/gt/intel_mocs.c
610
(l3cc = l3cc_combine(get_entry_l3cc((t), 2 * i), \
drivers/gpu/drm/i915/gt/intel_mocs.c
611
get_entry_l3cc((t), 2 * i + 1))), 1 : \
drivers/gpu/drm/i915/gt/intel_rps.c
1499
unsigned int t, state1, state2;
drivers/gpu/drm/i915/gt/intel_rps.c
1514
t = ips_mch_val(uncore);
drivers/gpu/drm/i915/gt/intel_rps.c
1515
if (t > 80)
drivers/gpu/drm/i915/gt/intel_rps.c
1516
corr = t * 2349 + 135940;
drivers/gpu/drm/i915/gt/intel_rps.c
1517
else if (t >= 50)
drivers/gpu/drm/i915/gt/intel_rps.c
1518
corr = t * 964 + 29317;
drivers/gpu/drm/i915/gt/intel_rps.c
1520
corr = t * 301 + 1004;
drivers/gpu/drm/i915/gt/intel_rps.c
267
u16 t;
drivers/gpu/drm/i915/gt/intel_rps.c
300
cparams[i].t == DIV_ROUND_CLOSEST(mem_freq, 1000)) {
drivers/gpu/drm/i915/gt/intel_rps.c
79
static void rps_timer(struct timer_list *t)
drivers/gpu/drm/i915/gt/intel_rps.c
81
struct intel_rps *rps = timer_container_of(rps, t, timer);
drivers/gpu/drm/i915/gt/mock_engine.c
109
static void hw_delay_complete(struct timer_list *t)
drivers/gpu/drm/i915/gt/mock_engine.c
111
struct mock_engine *engine = timer_container_of(engine, t, hw_delay);
drivers/gpu/drm/i915/gt/selftest_engine_pm.c
264
ktime_t t[2];
drivers/gpu/drm/i915/gt/selftest_engine_pm.c
281
de = intel_engine_get_busy_time(engine, &t[0]);
drivers/gpu/drm/i915/gt/selftest_engine_pm.c
283
de = ktime_sub(intel_engine_get_busy_time(engine, &t[1]), de);
drivers/gpu/drm/i915/gt/selftest_engine_pm.c
285
dt = ktime_sub(t[1], t[0]);
drivers/gpu/drm/i915/gt/selftest_engine_pm.c
320
de = intel_engine_get_busy_time(engine, &t[0]);
drivers/gpu/drm/i915/gt/selftest_engine_pm.c
322
de = ktime_sub(intel_engine_get_busy_time(engine, &t[1]), de);
drivers/gpu/drm/i915/gt/selftest_engine_pm.c
324
dt = ktime_sub(t[1], t[0]);
drivers/gpu/drm/i915/gt/selftest_execlists.c
1586
struct igt_live_test t;
drivers/gpu/drm/i915/gt/selftest_execlists.c
1595
if (igt_live_test_begin(&t, gt->i915, __func__, engine->name)) {
drivers/gpu/drm/i915/gt/selftest_execlists.c
1697
if (igt_live_test_end(&t)) {
drivers/gpu/drm/i915/gt/selftest_execlists.c
1761
struct igt_live_test t;
drivers/gpu/drm/i915/gt/selftest_execlists.c
1767
if (igt_live_test_begin(&t, gt->i915, __func__, engine->name)) {
drivers/gpu/drm/i915/gt/selftest_execlists.c
1808
if (igt_live_test_end(&t)) {
drivers/gpu/drm/i915/gt/selftest_execlists.c
1854
struct igt_live_test t;
drivers/gpu/drm/i915/gt/selftest_execlists.c
186
struct igt_live_test t;
drivers/gpu/drm/i915/gt/selftest_execlists.c
1860
if (igt_live_test_begin(&t, gt->i915, __func__, engine->name)) {
drivers/gpu/drm/i915/gt/selftest_execlists.c
1904
if (igt_live_test_end(&t)) {
drivers/gpu/drm/i915/gt/selftest_execlists.c
195
if (igt_live_test_begin(&t, gt->i915, __func__, engine->name)) {
drivers/gpu/drm/i915/gt/selftest_execlists.c
2062
struct igt_live_test t;
drivers/gpu/drm/i915/gt/selftest_execlists.c
2067
if (igt_live_test_begin(&t, arg->engine->i915,
drivers/gpu/drm/i915/gt/selftest_execlists.c
2098
if (igt_live_test_end(&t))
drivers/gpu/drm/i915/gt/selftest_execlists.c
2106
struct igt_live_test t;
drivers/gpu/drm/i915/gt/selftest_execlists.c
2111
if (igt_live_test_begin(&t, arg->engine->i915,
drivers/gpu/drm/i915/gt/selftest_execlists.c
2169
if (igt_live_test_end(&t))
drivers/gpu/drm/i915/gt/selftest_execlists.c
2177
struct igt_live_test t;
drivers/gpu/drm/i915/gt/selftest_execlists.c
2182
if (igt_live_test_begin(&t, arg->engine->i915,
drivers/gpu/drm/i915/gt/selftest_execlists.c
2264
if (igt_live_test_end(&t))
drivers/gpu/drm/i915/gt/selftest_execlists.c
2578
struct igt_live_test t;
drivers/gpu/drm/i915/gt/selftest_execlists.c
2609
if (igt_live_test_begin(&t, gt->i915, __func__, engine->name)) {
drivers/gpu/drm/i915/gt/selftest_execlists.c
2683
if (igt_live_test_end(&t)) {
drivers/gpu/drm/i915/gt/selftest_execlists.c
2805
struct igt_live_test t;
drivers/gpu/drm/i915/gt/selftest_execlists.c
2809
if (igt_live_test_begin(&t, engine->i915, __func__, engine->name))
drivers/gpu/drm/i915/gt/selftest_execlists.c
2912
if (igt_live_test_end(&t))
drivers/gpu/drm/i915/gt/selftest_execlists.c
2979
struct igt_live_test t;
drivers/gpu/drm/i915/gt/selftest_execlists.c
2988
if (igt_live_test_begin(&t, gt->i915, __func__, engine->name))
drivers/gpu/drm/i915/gt/selftest_execlists.c
3040
if (igt_live_test_end(&t))
drivers/gpu/drm/i915/gt/selftest_execlists.c
308
if (igt_live_test_end(&t))
drivers/gpu/drm/i915/gt/selftest_execlists.c
3285
struct igt_live_test t;
drivers/gpu/drm/i915/gt/selftest_execlists.c
3294
if (igt_live_test_begin(&t, gt->i915, __func__, engine->name)) {
drivers/gpu/drm/i915/gt/selftest_execlists.c
3357
if (igt_live_test_end(&t))
drivers/gpu/drm/i915/gt/selftest_execlists.c
348
struct igt_live_test t;
drivers/gpu/drm/i915/gt/selftest_execlists.c
357
if (igt_live_test_begin(&t, gt->i915, __func__, engine->name)) {
drivers/gpu/drm/i915/gt/selftest_execlists.c
3646
struct igt_live_test t;
drivers/gpu/drm/i915/gt/selftest_execlists.c
3673
if (igt_live_test_begin(&t, smoke.gt->i915, __func__, "all")) {
drivers/gpu/drm/i915/gt/selftest_execlists.c
3695
if (igt_live_test_end(&t))
drivers/gpu/drm/i915/gt/selftest_execlists.c
3723
struct igt_live_test t;
drivers/gpu/drm/i915/gt/selftest_execlists.c
3745
err = igt_live_test_begin(&t, gt->i915, __func__, ve[0]->engine->name);
drivers/gpu/drm/i915/gt/selftest_execlists.c
3818
err = igt_live_test_end(&t);
drivers/gpu/drm/i915/gt/selftest_execlists.c
3917
struct igt_live_test t;
drivers/gpu/drm/i915/gt/selftest_execlists.c
3936
err = igt_live_test_begin(&t, gt->i915, __func__, ve->engine->name);
drivers/gpu/drm/i915/gt/selftest_execlists.c
3981
err = igt_live_test_end(&t);
drivers/gpu/drm/i915/gt/selftest_execlists.c
4196
struct igt_live_test t;
drivers/gpu/drm/i915/gt/selftest_execlists.c
4221
err = igt_live_test_begin(&t, gt->i915, __func__, ve->engine->name);
drivers/gpu/drm/i915/gt/selftest_execlists.c
4285
if (igt_live_test_end(&t))
drivers/gpu/drm/i915/gt/selftest_execlists.c
468
if (igt_live_test_end(&t))
drivers/gpu/drm/i915/gt/selftest_execlists.c
498
struct igt_live_test t;
drivers/gpu/drm/i915/gt/selftest_execlists.c
500
if (igt_live_test_begin(&t, gt->i915, __func__, engine->name)) {
drivers/gpu/drm/i915/gt/selftest_execlists.c
547
if (igt_live_test_end(&t)) {
drivers/gpu/drm/i915/gt/selftest_hangcheck.c
1880
struct tasklet_struct * const t = &engine->sched_engine->tasklet;
drivers/gpu/drm/i915/gt/selftest_hangcheck.c
1886
if (t->func)
drivers/gpu/drm/i915/gt/selftest_hangcheck.c
1887
tasklet_disable(t);
drivers/gpu/drm/i915/gt/selftest_hangcheck.c
1897
if (t->func) {
drivers/gpu/drm/i915/gt/selftest_hangcheck.c
1898
tasklet_enable(t);
drivers/gpu/drm/i915/gt/selftest_hangcheck.c
1899
tasklet_hi_schedule(t);
drivers/gpu/drm/i915/gt/selftest_lrc.c
370
}, *t;
drivers/gpu/drm/i915/gt/selftest_lrc.c
383
for (t = tbl; t->name; t++) {
drivers/gpu/drm/i915/gt/selftest_lrc.c
384
int dw = find_offset(hw, t->reg);
drivers/gpu/drm/i915/gt/selftest_lrc.c
386
if (dw != t->offset) {
drivers/gpu/drm/i915/gt/selftest_lrc.c
389
t->name,
drivers/gpu/drm/i915/gt/selftest_lrc.c
390
t->reg,
drivers/gpu/drm/i915/gt/selftest_lrc.c
392
t->offset);
drivers/gpu/drm/i915/gt/selftest_migrate.c
856
ktime_t t[5];
drivers/gpu/drm/i915/gt/selftest_migrate.c
860
for (pass = 0; pass < ARRAY_SIZE(t); pass++) {
drivers/gpu/drm/i915/gt/selftest_migrate.c
877
t[pass] = ktime_sub(t1, t0);
drivers/gpu/drm/i915/gt/selftest_migrate.c
882
sort(t, ARRAY_SIZE(t), sizeof(*t), wrap_ktime_compare, NULL);
drivers/gpu/drm/i915/gt/selftest_migrate.c
887
t[1] + 2 * t[2] + t[3]) >> 20);
drivers/gpu/drm/i915/gt/selftest_migrate.c
935
ktime_t t[5];
drivers/gpu/drm/i915/gt/selftest_migrate.c
939
for (pass = 0; pass < ARRAY_SIZE(t); pass++) {
drivers/gpu/drm/i915/gt/selftest_migrate.c
960
t[pass] = ktime_sub(t1, t0);
drivers/gpu/drm/i915/gt/selftest_migrate.c
965
sort(t, ARRAY_SIZE(t), sizeof(*t), wrap_ktime_compare, NULL);
drivers/gpu/drm/i915/gt/selftest_migrate.c
970
t[1] + 2 * t[2] + t[3]) >> 20);
drivers/gpu/drm/i915/gt/selftest_reset.c
330
struct tasklet_struct *t = &engine->sched_engine->tasklet;
drivers/gpu/drm/i915/gt/selftest_reset.c
332
if (t->func)
drivers/gpu/drm/i915/gt/selftest_reset.c
333
tasklet_disable(t);
drivers/gpu/drm/i915/gt/selftest_reset.c
357
if (t->func) {
drivers/gpu/drm/i915/gt/selftest_reset.c
358
tasklet_enable(t);
drivers/gpu/drm/i915/gt/selftest_reset.c
359
tasklet_hi_schedule(t);
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
134
static void ct_receive_tasklet_func(struct tasklet_struct *t);
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
1342
static void ct_receive_tasklet_func(struct tasklet_struct *t)
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
1344
struct intel_guc_ct *ct = from_tasklet(ct, t, receive_tasklet);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
1047
static void guc_submission_tasklet(struct tasklet_struct *t)
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
1050
from_tasklet(sched_engine, t, tasklet);
drivers/gpu/drm/i915/gvt/gtt.c
91
#define gtt_init_entry(e, t, p, v) do { \
drivers/gpu/drm/i915/gvt/gtt.c
92
(e)->type = t; \
drivers/gpu/drm/i915/gvt/handlers.c
1147
u8 t)
drivers/gpu/drm/i915/gvt/handlers.c
1149
if ((t & DP_TRAINING_PATTERN_MASK) == DP_TRAINING_PATTERN_1) {
drivers/gpu/drm/i915/gvt/handlers.c
1157
} else if ((t & DP_TRAINING_PATTERN_MASK) ==
drivers/gpu/drm/i915/gvt/handlers.c
1173
} else if ((t & DP_TRAINING_PATTERN_MASK) ==
drivers/gpu/drm/i915/gvt/handlers.c
1234
int t;
drivers/gpu/drm/i915/gvt/handlers.c
1263
for (t = 0; t < 4; t++) {
drivers/gpu/drm/i915/gvt/handlers.c
1264
u32 r = vgpu_vreg(vgpu, offset + 8 + t * 4);
drivers/gpu/drm/i915/gvt/handlers.c
1266
buf[t * 4] = (r >> 24) & 0xff;
drivers/gpu/drm/i915/gvt/handlers.c
1267
buf[t * 4 + 1] = (r >> 16) & 0xff;
drivers/gpu/drm/i915/gvt/handlers.c
1268
buf[t * 4 + 2] = (r >> 8) & 0xff;
drivers/gpu/drm/i915/gvt/handlers.c
1269
buf[t * 4 + 3] = r & 0xff;
drivers/gpu/drm/i915/gvt/handlers.c
1274
for (t = 0; t <= len; t++) {
drivers/gpu/drm/i915/gvt/handlers.c
1275
int p = addr + t;
drivers/gpu/drm/i915/gvt/handlers.c
1277
dpcd->data[p] = buf[t];
drivers/gpu/drm/i915/gvt/handlers.c
1281
buf[t]);
drivers/gpu/drm/i915/gvt/handlers.c
1333
int t;
drivers/gpu/drm/i915/gvt/handlers.c
1335
t = dpcd->data[addr + i - 1];
drivers/gpu/drm/i915/gvt/handlers.c
1336
t <<= (24 - 8 * (i % 4));
drivers/gpu/drm/i915/gvt/handlers.c
1337
ret |= t;
drivers/gpu/drm/i915/i915_reg.h
925
#define DG1_MSTR_TILE(t) REG_BIT(t)
drivers/gpu/drm/i915/i915_request.c
1871
unsigned long t;
drivers/gpu/drm/i915/i915_request.c
1886
t = local_clock();
drivers/gpu/drm/i915/i915_request.c
1889
return t;
drivers/gpu/drm/i915/i915_selftest.h
139
#define igt_timeout(t, fmt, ...) \
drivers/gpu/drm/i915/i915_selftest.h
140
__igt_timeout((t), KERN_DEBUG pr_fmt(fmt), ##__VA_ARGS__)
drivers/gpu/drm/i915/i915_sw_fence.c
428
static void timer_i915_sw_fence_wake(struct timer_list *t)
drivers/gpu/drm/i915/i915_sw_fence.c
430
struct i915_sw_dma_fence_cb_timer *cb = timer_container_of(cb, t,
drivers/gpu/drm/i915/i915_tasklet.h
11
static inline void tasklet_lock(struct tasklet_struct *t)
drivers/gpu/drm/i915/i915_tasklet.h
13
while (!tasklet_trylock(t))
drivers/gpu/drm/i915/i915_tasklet.h
17
static inline bool tasklet_is_locked(const struct tasklet_struct *t)
drivers/gpu/drm/i915/i915_tasklet.h
19
return test_bit(TASKLET_STATE_RUN, &t->state);
drivers/gpu/drm/i915/i915_tasklet.h
22
static inline void __tasklet_disable_sync_once(struct tasklet_struct *t)
drivers/gpu/drm/i915/i915_tasklet.h
24
if (!atomic_fetch_inc(&t->count))
drivers/gpu/drm/i915/i915_tasklet.h
25
tasklet_unlock_spin_wait(t);
drivers/gpu/drm/i915/i915_tasklet.h
28
static inline bool __tasklet_is_enabled(const struct tasklet_struct *t)
drivers/gpu/drm/i915/i915_tasklet.h
30
return !atomic_read(&t->count);
drivers/gpu/drm/i915/i915_tasklet.h
33
static inline bool __tasklet_enable(struct tasklet_struct *t)
drivers/gpu/drm/i915/i915_tasklet.h
35
return atomic_dec_and_test(&t->count);
drivers/gpu/drm/i915/i915_tasklet.h
38
static inline bool __tasklet_is_scheduled(struct tasklet_struct *t)
drivers/gpu/drm/i915/i915_tasklet.h
40
return test_bit(TASKLET_STATE_SCHED, &t->state);
drivers/gpu/drm/i915/i915_timer_util.c
10
if (!timer_active(t))
drivers/gpu/drm/i915/i915_timer_util.c
13
timer_delete(t);
drivers/gpu/drm/i915/i915_timer_util.c
14
WRITE_ONCE(t->expires, 0);
drivers/gpu/drm/i915/i915_timer_util.c
17
void set_timer_ms(struct timer_list *t, unsigned long timeout)
drivers/gpu/drm/i915/i915_timer_util.c
20
cancel_timer(t);
drivers/gpu/drm/i915/i915_timer_util.c
35
mod_timer(t, jiffies + timeout ?: 1);
drivers/gpu/drm/i915/i915_timer_util.c
8
void cancel_timer(struct timer_list *t)
drivers/gpu/drm/i915/i915_timer_util.h
10
void cancel_timer(struct timer_list *t);
drivers/gpu/drm/i915/i915_timer_util.h
11
void set_timer_ms(struct timer_list *t, unsigned long timeout);
drivers/gpu/drm/i915/i915_timer_util.h
13
static inline bool timer_active(const struct timer_list *t)
drivers/gpu/drm/i915/i915_timer_util.h
15
return READ_ONCE(t->expires);
drivers/gpu/drm/i915/i915_timer_util.h
18
static inline bool timer_expired(const struct timer_list *t)
drivers/gpu/drm/i915/i915_timer_util.h
20
return timer_active(t) && !timer_pending(t);
drivers/gpu/drm/i915/intel_wakeref.c
139
static void wakeref_auto_timeout(struct timer_list *t)
drivers/gpu/drm/i915/intel_wakeref.c
141
struct intel_wakeref_auto *wf = timer_container_of(wf, t, timer);
drivers/gpu/drm/i915/selftests/i915_request.c
1042
struct igt_live_test t;
drivers/gpu/drm/i915/selftests/i915_request.c
1062
err = igt_live_test_begin(&t, i915, __func__, engine->name);
drivers/gpu/drm/i915/selftests/i915_request.c
1101
err = igt_live_test_end(&t);
drivers/gpu/drm/i915/selftests/i915_request.c
1196
struct igt_live_test t;
drivers/gpu/drm/i915/selftests/i915_request.c
1210
err = igt_live_test_begin(&t, i915, __func__, "");
drivers/gpu/drm/i915/selftests/i915_request.c
1296
err = igt_live_test_end(&t);
drivers/gpu/drm/i915/selftests/i915_request.c
1325
struct igt_live_test t;
drivers/gpu/drm/i915/selftests/i915_request.c
1340
err = igt_live_test_begin(&t, i915, __func__, "");
drivers/gpu/drm/i915/selftests/i915_request.c
1427
err = igt_live_test_end(&t);
drivers/gpu/drm/i915/selftests/i915_request.c
1635
struct igt_live_test t;
drivers/gpu/drm/i915/selftests/i915_request.c
1639
err = igt_live_test_begin(&t, i915, __func__, name);
drivers/gpu/drm/i915/selftests/i915_request.c
1680
if (igt_live_test_end(&t))
drivers/gpu/drm/i915/selftests/i915_request.c
1815
threads[i].t = &smoke[idx];
drivers/gpu/drm/i915/selftests/i915_request.c
2876
struct igt_live_test t;
drivers/gpu/drm/i915/selftests/i915_request.c
2879
err = igt_live_test_begin(&t, i915, __func__, name);
drivers/gpu/drm/i915/selftests/i915_request.c
2900
if (igt_live_test_end(&t))
drivers/gpu/drm/i915/selftests/i915_request.c
306
struct smoketest *t;
drivers/gpu/drm/i915/selftests/i915_request.c
314
struct smoketest *t = thread->t;
drivers/gpu/drm/i915/selftests/i915_request.c
315
const unsigned int max_batch = min(t->ncontexts, t->max_batch) - 1;
drivers/gpu/drm/i915/selftests/i915_request.c
316
const unsigned int total = 4 * t->ncontexts + 1;
drivers/gpu/drm/i915/selftests/i915_request.c
3203
struct igt_live_test t;
drivers/gpu/drm/i915/selftests/i915_request.c
3207
err = igt_live_test_begin(&t, i915, __func__, name);
drivers/gpu/drm/i915/selftests/i915_request.c
3256
if (igt_live_test_end(&t))
drivers/gpu/drm/i915/selftests/i915_request.c
367
t->contexts[order[n] % t->ncontexts];
drivers/gpu/drm/i915/selftests/i915_request.c
371
ce = i915_gem_context_get_engine(ctx, t->engine->legacy_idx);
drivers/gpu/drm/i915/selftests/i915_request.c
373
rq = t->request_alloc(ce);
drivers/gpu/drm/i915/selftests/i915_request.c
412
t->engine->name);
drivers/gpu/drm/i915/selftests/i915_request.c
415
intel_gt_set_wedged(t->engine->gt);
drivers/gpu/drm/i915/selftests/i915_request.c
446
atomic_long_add(num_fences, &t->num_fences);
drivers/gpu/drm/i915/selftests/i915_request.c
447
atomic_long_add(num_waits, &t->num_waits);
drivers/gpu/drm/i915/selftests/i915_request.c
458
struct smoketest t = {
drivers/gpu/drm/i915/selftests/i915_request.c
479
t.contexts = kzalloc_objs(*t.contexts, t.ncontexts);
drivers/gpu/drm/i915/selftests/i915_request.c
480
if (!t.contexts) {
drivers/gpu/drm/i915/selftests/i915_request.c
485
for (n = 0; n < t.ncontexts; n++) {
drivers/gpu/drm/i915/selftests/i915_request.c
486
t.contexts[n] = mock_context(t.engine->i915, "mock");
drivers/gpu/drm/i915/selftests/i915_request.c
487
if (!t.contexts[n]) {
drivers/gpu/drm/i915/selftests/i915_request.c
504
threads[n].t = &t;
drivers/gpu/drm/i915/selftests/i915_request.c
527
atomic_long_read(&t.num_waits),
drivers/gpu/drm/i915/selftests/i915_request.c
528
atomic_long_read(&t.num_fences),
drivers/gpu/drm/i915/selftests/i915_request.c
532
for (n = 0; n < t.ncontexts; n++) {
drivers/gpu/drm/i915/selftests/i915_request.c
533
if (!t.contexts[n])
drivers/gpu/drm/i915/selftests/i915_request.c
535
mock_context_close(t.contexts[n]);
drivers/gpu/drm/i915/selftests/i915_request.c
537
kfree(t.contexts);
drivers/gpu/drm/i915/selftests/i915_request.c
572
struct igt_live_test t;
drivers/gpu/drm/i915/selftests/i915_request.c
586
err = igt_live_test_begin(&t, i915, __func__, engine->name);
drivers/gpu/drm/i915/selftests/i915_request.c
631
err = igt_live_test_end(&t);
drivers/gpu/drm/i915/selftests/i915_request.c
927
struct igt_live_test t;
drivers/gpu/drm/i915/selftests/i915_request.c
933
err = igt_live_test_begin(&t, i915, __func__, engine->name);
drivers/gpu/drm/i915/selftests/i915_request.c
943
err2 = igt_live_test_end(&t);
drivers/gpu/drm/i915/selftests/i915_vma.c
1002
for (t = types; *t; t++) {
drivers/gpu/drm/i915/selftests/i915_vma.c
1005
.type = *t,
drivers/gpu/drm/i915/selftests/i915_vma.c
1020
plane_info[0].dst_stride = *t == I915_GTT_VIEW_ROTATED ?
drivers/gpu/drm/i915/selftests/i915_vma.c
1029
GEM_BUG_ON(vma->gtt_view.type != *t);
drivers/gpu/drm/i915/selftests/i915_vma.c
1043
if (*t == I915_GTT_VIEW_ROTATED)
drivers/gpu/drm/i915/selftests/i915_vma.c
1075
if (*t == I915_GTT_VIEW_ROTATED)
drivers/gpu/drm/i915/selftests/i915_vma.c
1084
*t == I915_GTT_VIEW_ROTATED ? "Rotated" : "Remapped",
drivers/gpu/drm/i915/selftests/i915_vma.c
579
}, *t;
drivers/gpu/drm/i915/selftests/i915_vma.c
592
for (t = types; *t; t++) {
drivers/gpu/drm/i915/selftests/i915_vma.c
596
.type = *t,
drivers/gpu/drm/i915/selftests/i915_vma.c
988
}, *t;
drivers/gpu/drm/i915/selftests/igt_live_test.c
15
int igt_live_test_begin(struct igt_live_test *t,
drivers/gpu/drm/i915/selftests/igt_live_test.c
26
t->i915 = i915;
drivers/gpu/drm/i915/selftests/igt_live_test.c
27
t->func = func;
drivers/gpu/drm/i915/selftests/igt_live_test.c
28
t->name = name;
drivers/gpu/drm/i915/selftests/igt_live_test.c
40
t->reset_engine[i][id] =
drivers/gpu/drm/i915/selftests/igt_live_test.c
45
t->reset_global = i915_reset_count(&i915->gpu_error);
drivers/gpu/drm/i915/selftests/igt_live_test.c
50
int igt_live_test_end(struct igt_live_test *t)
drivers/gpu/drm/i915/selftests/igt_live_test.c
52
struct drm_i915_private *i915 = t->i915;
drivers/gpu/drm/i915/selftests/igt_live_test.c
61
if (t->reset_global != i915_reset_count(&i915->gpu_error)) {
drivers/gpu/drm/i915/selftests/igt_live_test.c
63
t->func, t->name,
drivers/gpu/drm/i915/selftests/igt_live_test.c
64
i915_reset_count(&i915->gpu_error) - t->reset_global);
drivers/gpu/drm/i915/selftests/igt_live_test.c
70
if (t->reset_engine[i][id] ==
drivers/gpu/drm/i915/selftests/igt_live_test.c
75
t->func, t->name, engine->name,
drivers/gpu/drm/i915/selftests/igt_live_test.c
77
t->reset_engine[i][id]);
drivers/gpu/drm/i915/selftests/igt_live_test.h
30
int igt_live_test_begin(struct igt_live_test *t,
drivers/gpu/drm/i915/selftests/igt_live_test.h
34
int igt_live_test_end(struct igt_live_test *t);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
1248
ktime_t t[5];
drivers/gpu/drm/i915/selftests/intel_memory_region.c
1254
for (pass = 0; pass < ARRAY_SIZE(t); pass++) {
drivers/gpu/drm/i915/selftests/intel_memory_region.c
1262
t[pass] = ktime_sub(t1, t0);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
1265
sort(t, ARRAY_SIZE(t), sizeof(*t), wrap_ktime_compare, NULL);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
1266
if (t[0] <= 0) {
drivers/gpu/drm/i915/selftests/intel_memory_region.c
1273
t[0], t[4]);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
1284
t[1] + 2 * t[2] + t[3]) >> 20);
drivers/gpu/drm/i915/selftests/lib_sw_fence.c
56
static void timed_fence_wake(struct timer_list *t)
drivers/gpu/drm/i915/selftests/lib_sw_fence.c
58
struct timed_fence *tf = timer_container_of(tf, t, timer);
drivers/gpu/drm/imagination/pvr_rogue_meta.h
17
#define META_CR_CTRLREG_BASE(t) (0x04800000U + (0x1000U * (t)))
drivers/gpu/drm/mediatek/mtk_dp.c
2591
static void mtk_dp_debounce_timer(struct timer_list *t)
drivers/gpu/drm/mediatek/mtk_dp.c
2593
struct mtk_dp *mtk_dp = timer_container_of(mtk_dp, t, debounce_timer);
drivers/gpu/drm/mediatek/mtk_dsi.c
663
static s32 mtk_dsi_switch_to_cmd_mode(struct mtk_dsi *dsi, u8 irq_flag, u32 t)
drivers/gpu/drm/mediatek/mtk_dsi.c
668
if (!mtk_dsi_wait_for_irq_done(dsi, irq_flag, t)) {
drivers/gpu/drm/msm/adreno/a5xx_preempt.c
80
static void a5xx_preempt_timer(struct timer_list *t)
drivers/gpu/drm/msm/adreno/a5xx_preempt.c
82
struct a5xx_gpu *a5xx_gpu = timer_container_of(a5xx_gpu, t,
drivers/gpu/drm/msm/adreno/a6xx_preempt.c
88
static void a6xx_preempt_timer(struct timer_list *t)
drivers/gpu/drm/msm/adreno/a6xx_preempt.c
90
struct a6xx_gpu *a6xx_gpu = timer_container_of(a6xx_gpu, t,
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c
2710
static void dpu_encoder_frame_done_timeout(struct timer_list *t)
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c
2712
struct dpu_encoder_virt *dpu_enc = timer_container_of(dpu_enc, t,
drivers/gpu/drm/msm/msm_fence.c
19
static enum hrtimer_restart deadline_timer(struct hrtimer *t)
drivers/gpu/drm/msm/msm_fence.c
21
struct msm_fence_context *fctx = container_of(t,
drivers/gpu/drm/msm/msm_gpu.c
641
static void hangcheck_handler(struct timer_list *t)
drivers/gpu/drm/msm/msm_gpu.c
643
struct msm_gpu *gpu = timer_container_of(gpu, t, hangcheck_timer);
drivers/gpu/drm/msm/msm_io_utils.c
115
static enum hrtimer_restart msm_hrtimer_worktimer(struct hrtimer *t)
drivers/gpu/drm/msm/msm_io_utils.c
117
struct msm_hrtimer_work *work = container_of(t,
drivers/gpu/drm/nouveau/include/nvkm/core/memory.h
106
#define nvkm_fill(t,s,o,a,d,c) do { \
drivers/gpu/drm/nouveau/include/nvkm/core/memory.h
108
u##t __iomem *_m = nvkm_kmap(o); \
drivers/gpu/drm/nouveau/include/nvkm/core/memory.h
112
iowrite##t##_native(_d, &_m[_o++]); \
drivers/gpu/drm/nouveau/include/nvkm/core/memory.h
118
nvkm_wo##t((o), _a, _d); \
drivers/gpu/drm/nouveau/nouveau_fence.c
267
unsigned long t = jiffies, timeout = t + wait;
drivers/gpu/drm/nouveau/nouveau_fence.c
272
t = jiffies;
drivers/gpu/drm/nouveau/nouveau_fence.c
274
if (wait != MAX_SCHEDULE_TIMEOUT && time_after_eq(t, timeout)) {
drivers/gpu/drm/nouveau/nouveau_fence.c
294
return timeout - t;
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1020
const u16 t = outp->info.hasht;
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1022
u32 data = nvbios_outp_match(bios, t, m, ver, hdr, cnt, len, iedt);
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
1024
OUTP_DBG(outp, "missing IEDT for %04x:%04x", t, m);
drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgf108.c
755
const u32 t = timeslice_mode;
drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgf108.c
758
gf100_grctx_patch_wr32(chan, o + 0x20, (t << 28) | (b << 16) | bo);
drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgf117.c
263
const u32 t = timeslice_mode;
drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgf117.c
269
gf100_grctx_patch_wr32(chan, o + 0xc0, (t << 28) | (b << 16) | bo);
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.h
43
#define GPC_UNIT(t, r) (0x500000 + (t) * 0x8000 + (r))
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.h
44
#define PPC_UNIT(t, m, r) (0x503000 + (t) * 0x8000 + (m) * 0x200 + (r))
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.h
45
#define TPC_UNIT(t, m, r) (0x504000 + (t) * 0x8000 + (m) * 0x800 + (r))
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c
346
#define T(t) cfg->timing_10_##t
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c
71
#define T(t) cfg->timing_10_##t
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1940
struct debugfs_blob_wrapper *t)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1953
t->data = p;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1954
t->size = s->size;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1956
dent = debugfs_create_blob(name, 0444, parent, t);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1959
memset(t, 0, sizeof(*t));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1963
i_size_write(d_inode(dent), t->size);
drivers/gpu/drm/omapdrm/dss/dispc.c
3095
struct videomode t = *vm;
drivers/gpu/drm/omapdrm/dss/dispc.c
3097
DSSDBG("channel %d xres %u yres %u\n", channel, t.hactive, t.vactive);
drivers/gpu/drm/omapdrm/dss/dispc.c
3099
if (dispc_mgr_check_timings(dispc, channel, &t)) {
drivers/gpu/drm/omapdrm/dss/dispc.c
3105
_dispc_mgr_set_lcd_timings(dispc, channel, &t);
drivers/gpu/drm/omapdrm/dss/dispc.c
3107
xtot = t.hactive + t.hfront_porch + t.hsync_len + t.hback_porch;
drivers/gpu/drm/omapdrm/dss/dispc.c
3108
ytot = t.vactive + t.vfront_porch + t.vsync_len + t.vback_porch;
drivers/gpu/drm/omapdrm/dss/dispc.c
3115
t.hsync_len, t.hfront_porch, t.hback_porch,
drivers/gpu/drm/omapdrm/dss/dispc.c
3116
t.vsync_len, t.vfront_porch, t.vback_porch);
drivers/gpu/drm/omapdrm/dss/dispc.c
3118
vm_flag_to_int(t.flags, DISPLAY_FLAGS_VSYNC_HIGH, DISPLAY_FLAGS_VSYNC_LOW),
drivers/gpu/drm/omapdrm/dss/dispc.c
3119
vm_flag_to_int(t.flags, DISPLAY_FLAGS_HSYNC_HIGH, DISPLAY_FLAGS_HSYNC_LOW),
drivers/gpu/drm/omapdrm/dss/dispc.c
3120
vm_flag_to_int(t.flags, DISPLAY_FLAGS_PIXDATA_POSEDGE, DISPLAY_FLAGS_PIXDATA_NEGEDGE),
drivers/gpu/drm/omapdrm/dss/dispc.c
3121
vm_flag_to_int(t.flags, DISPLAY_FLAGS_DE_HIGH, DISPLAY_FLAGS_DE_LOW),
drivers/gpu/drm/omapdrm/dss/dispc.c
3122
vm_flag_to_int(t.flags, DISPLAY_FLAGS_SYNC_POSEDGE, DISPLAY_FLAGS_SYNC_NEGEDGE));
drivers/gpu/drm/omapdrm/dss/dispc.c
3126
if (t.flags & DISPLAY_FLAGS_INTERLACED)
drivers/gpu/drm/omapdrm/dss/dispc.c
3127
t.vactive /= 2;
drivers/gpu/drm/omapdrm/dss/dispc.c
3131
!!(t.flags & DISPLAY_FLAGS_DOUBLECLK),
drivers/gpu/drm/omapdrm/dss/dispc.c
3135
dispc_mgr_set_size(dispc, channel, t.hactive, t.vactive);
drivers/gpu/drm/omapdrm/dss/dsi.c
1229
int t = 0;
drivers/gpu/drm/omapdrm/dss/dsi.c
1237
if (++t > 1000) {
drivers/gpu/drm/omapdrm/dss/dsi.c
1300
unsigned int t;
drivers/gpu/drm/omapdrm/dss/dsi.c
1302
for (t = 0; t < dsi->num_lanes_supported; ++t)
drivers/gpu/drm/omapdrm/dss/dsi.c
1303
if (dsi->lanes[t].function == functions[i])
drivers/gpu/drm/omapdrm/dss/dsi.c
1306
if (t == dsi->num_lanes_supported)
drivers/gpu/drm/omapdrm/dss/dsi.c
1309
lane_number = t;
drivers/gpu/drm/omapdrm/dss/dsi.c
1310
polarity = dsi->lanes[t].polarity;
drivers/gpu/drm/omapdrm/dss/dsi.c
141
int t;
drivers/gpu/drm/omapdrm/dss/dsi.c
1423
int t, i;
drivers/gpu/drm/omapdrm/dss/dsi.c
1437
t = 100000;
drivers/gpu/drm/omapdrm/dss/dsi.c
144
t = 100;
drivers/gpu/drm/omapdrm/dss/dsi.c
145
while (t-- > 0) {
drivers/gpu/drm/omapdrm/dss/dsi.c
1453
if (--t == 0) {
drivers/gpu/drm/omapdrm/dss/dsi.c
177
ktime_t t, setup_time, trans_time;
drivers/gpu/drm/omapdrm/dss/dsi.c
184
t = ktime_get();
drivers/gpu/drm/omapdrm/dss/dsi.c
191
trans_time = ktime_sub(t, dsi->perf_start_time);
drivers/gpu/drm/omapdrm/dss/dsi.c
3577
const struct omap_dss_dsi_videomode_timings *t)
drivers/gpu/drm/omapdrm/dss/dsi.c
3579
unsigned long byteclk = t->hsclk / 4;
drivers/gpu/drm/omapdrm/dss/dsi.c
3582
wc = DIV_ROUND_UP(t->hact * t->bitspp, 8);
drivers/gpu/drm/omapdrm/dss/dsi.c
3583
pps = DIV_ROUND_UP(wc + 6, t->ndl); /* pixel packet size */
drivers/gpu/drm/omapdrm/dss/dsi.c
3584
bl = t->hss + t->hsa + t->hse + t->hbp + t->hfp;
drivers/gpu/drm/omapdrm/dss/dsi.c
3593
t->hss, t->hsa, t->hse, t->hbp, pps, t->hfp,
drivers/gpu/drm/omapdrm/dss/dsi.c
3595
TO_DSI_T(t->hss),
drivers/gpu/drm/omapdrm/dss/dsi.c
3596
TO_DSI_T(t->hsa),
drivers/gpu/drm/omapdrm/dss/dsi.c
3597
TO_DSI_T(t->hse),
drivers/gpu/drm/omapdrm/dss/dsi.c
3598
TO_DSI_T(t->hbp),
drivers/gpu/drm/omapdrm/dss/dsi.c
3600
TO_DSI_T(t->hfp),
drivers/gpu/drm/omapdrm/dss/dsi.c
3638
const struct omap_dss_dsi_videomode_timings *t)
drivers/gpu/drm/omapdrm/dss/dsi.c
3641
unsigned long byteclk = t->hsclk / 4;
drivers/gpu/drm/omapdrm/dss/dsi.c
3646
dsi_tput = (u64)byteclk * t->ndl * 8;
drivers/gpu/drm/omapdrm/dss/dsi.c
3647
pck = (u32)div64_u64(dsi_tput, t->bitspp);
drivers/gpu/drm/omapdrm/dss/dsi.c
3648
dsi_hact = DIV_ROUND_UP(DIV_ROUND_UP(t->hact * t->bitspp, 8) + 6, t->ndl);
drivers/gpu/drm/omapdrm/dss/dsi.c
3649
dsi_htot = t->hss + t->hsa + t->hse + t->hbp + dsi_hact + t->hfp;
drivers/gpu/drm/omapdrm/dss/dsi.c
3652
vm.hsync_len = div64_u64((u64)(t->hsa + t->hse) * pck, byteclk);
drivers/gpu/drm/omapdrm/dss/dsi.c
3653
vm.hback_porch = div64_u64((u64)t->hbp * pck, byteclk);
drivers/gpu/drm/omapdrm/dss/dsi.c
3654
vm.hfront_porch = div64_u64((u64)t->hfp * pck, byteclk);
drivers/gpu/drm/omapdrm/dss/dsi.c
3655
vm.hactive = t->hact;
drivers/gpu/drm/omapdrm/dss/dsi.c
3864
int t;
drivers/gpu/drm/omapdrm/dss/dsi.c
3867
t = 1 - hfp;
drivers/gpu/drm/omapdrm/dss/dsi.c
3868
hbp = max(hbp - t, 1);
drivers/gpu/drm/omapdrm/dss/dsi.c
3873
t = 1 - hfp;
drivers/gpu/drm/omapdrm/dss/dsi.c
3874
hsa = max(hsa - t, 1);
drivers/gpu/drm/omapdrm/dss/dsi.c
3922
int t;
drivers/gpu/drm/omapdrm/dss/dsi.c
3925
t = 1 - hfp;
drivers/gpu/drm/omapdrm/dss/dsi.c
3926
hbp = max(hbp - t, 1);
drivers/gpu/drm/omapdrm/dss/dsi.c
3931
t = 1 - hfp;
drivers/gpu/drm/omapdrm/dss/dsi.c
3932
hsa = max(hsa - t, 1);
drivers/gpu/drm/omapdrm/dss/dsi.c
878
int t = 0;
drivers/gpu/drm/omapdrm/dss/dsi.c
890
if (++t > 1000) {
drivers/gpu/drm/omapdrm/dss/hdmi.h
286
u32 t = 0, v;
drivers/gpu/drm/omapdrm/dss/hdmi.h
288
if (t++ > 10000)
drivers/gpu/drm/omapdrm/dss/hdmi4_core.c
121
int t;
drivers/gpu/drm/omapdrm/dss/hdmi4_core.c
129
t = 0;
drivers/gpu/drm/omapdrm/dss/hdmi4_core.c
132
if (t++ > 10000) {
drivers/gpu/drm/omapdrm/dss/pll.c
333
int t;
drivers/gpu/drm/omapdrm/dss/pll.c
336
t = 100;
drivers/gpu/drm/omapdrm/dss/pll.c
337
while (t-- > 0) {
drivers/gpu/drm/omapdrm/dss/pll.c
368
int t = 100;
drivers/gpu/drm/omapdrm/dss/pll.c
370
while (t-- > 0) {
drivers/gpu/drm/omapdrm/dss/venc.c
340
int t = 1000;
drivers/gpu/drm/omapdrm/dss/venc.c
344
if (--t == 0) {
drivers/gpu/drm/panel/panel-tpo-tpg110.c
193
struct spi_transfer t[2];
drivers/gpu/drm/panel/panel-tpo-tpg110.c
198
memset(t, 0, sizeof(t));
drivers/gpu/drm/panel/panel-tpo-tpg110.c
210
t[0].bits_per_word = 8;
drivers/gpu/drm/panel/panel-tpo-tpg110.c
211
t[0].tx_buf = &buf[0];
drivers/gpu/drm/panel/panel-tpo-tpg110.c
212
t[0].len = 1;
drivers/gpu/drm/panel/panel-tpo-tpg110.c
214
t[1].tx_buf = &buf[1];
drivers/gpu/drm/panel/panel-tpo-tpg110.c
215
t[1].len = 1;
drivers/gpu/drm/panel/panel-tpo-tpg110.c
216
t[1].bits_per_word = 8;
drivers/gpu/drm/panel/panel-tpo-tpg110.c
227
t[0].bits_per_word = 7;
drivers/gpu/drm/panel/panel-tpo-tpg110.c
228
t[0].tx_buf = &buf[0];
drivers/gpu/drm/panel/panel-tpo-tpg110.c
229
t[0].len = 1;
drivers/gpu/drm/panel/panel-tpo-tpg110.c
231
t[1].rx_buf = &buf[1];
drivers/gpu/drm/panel/panel-tpo-tpg110.c
232
t[1].len = 1;
drivers/gpu/drm/panel/panel-tpo-tpg110.c
233
t[1].bits_per_word = 8;
drivers/gpu/drm/panel/panel-tpo-tpg110.c
236
spi_message_add_tail(&t[0], &m);
drivers/gpu/drm/panel/panel-tpo-tpg110.c
237
spi_message_add_tail(&t[1], &m);
drivers/gpu/drm/radeon/kv_dpm.h
55
u32 t;
drivers/gpu/drm/radeon/mkregtable.c
131
static void table_offset_add(struct table *t, struct offset *offset)
drivers/gpu/drm/radeon/mkregtable.c
133
list_add_tail(&offset->list, &t->offsets);
drivers/gpu/drm/radeon/mkregtable.c
136
static void table_init(struct table *t)
drivers/gpu/drm/radeon/mkregtable.c
138
INIT_LIST_HEAD(&t->offsets);
drivers/gpu/drm/radeon/mkregtable.c
139
t->offset_max = 0;
drivers/gpu/drm/radeon/mkregtable.c
140
t->nentry = 0;
drivers/gpu/drm/radeon/mkregtable.c
141
t->table = NULL;
drivers/gpu/drm/radeon/mkregtable.c
144
static void table_print(struct table *t)
drivers/gpu/drm/radeon/mkregtable.c
148
nlloop = (t->nentry + 3) / 4;
drivers/gpu/drm/radeon/mkregtable.c
149
c = t->nentry;
drivers/gpu/drm/radeon/mkregtable.c
150
printf("static const unsigned %s_reg_safe_bm[%d] = {\n", t->gpu_prefix,
drivers/gpu/drm/radeon/mkregtable.c
151
t->nentry);
drivers/gpu/drm/radeon/mkregtable.c
162
printf("0x%08X,", t->table[id++]);
drivers/gpu/drm/radeon/mkregtable.c
169
static int table_build(struct table *t)
drivers/gpu/drm/radeon/mkregtable.c
174
t->nentry = ((t->offset_max >> 2) + 31) / 32;
drivers/gpu/drm/radeon/mkregtable.c
175
t->table = (unsigned *)malloc(sizeof(unsigned) * t->nentry);
drivers/gpu/drm/radeon/mkregtable.c
176
if (t->table == NULL)
drivers/gpu/drm/radeon/mkregtable.c
178
memset(t->table, 0xff, sizeof(unsigned) * t->nentry);
drivers/gpu/drm/radeon/mkregtable.c
179
list_for_each_entry(offset, &t->offsets, list) {
drivers/gpu/drm/radeon/mkregtable.c
183
t->table[i] ^= m;
drivers/gpu/drm/radeon/mkregtable.c
189
static int parser_auth(struct table *t, const char *filename)
drivers/gpu/drm/radeon/mkregtable.c
227
t->gpu_prefix = gpu_name;
drivers/gpu/drm/radeon/mkregtable.c
253
table_offset_add(t, offset);
drivers/gpu/drm/radeon/mkregtable.c
254
if (o > t->offset_max)
drivers/gpu/drm/radeon/mkregtable.c
255
t->offset_max = o;
drivers/gpu/drm/radeon/mkregtable.c
260
if (t->offset_max < last_reg)
drivers/gpu/drm/radeon/mkregtable.c
261
t->offset_max = last_reg;
drivers/gpu/drm/radeon/mkregtable.c
262
return table_build(t);
drivers/gpu/drm/radeon/mkregtable.c
267
struct table t;
drivers/gpu/drm/radeon/mkregtable.c
273
table_init(&t);
drivers/gpu/drm/radeon/mkregtable.c
274
if (parser_auth(&t, argv[1])) {
drivers/gpu/drm/radeon/mkregtable.c
278
table_print(&t);
drivers/gpu/drm/radeon/ni_dpm.c
3056
s32 t;
drivers/gpu/drm/radeon/ni_dpm.c
3069
t = (1000 * ((i + 1) * 8));
drivers/gpu/drm/radeon/ni_dpm.c
3071
if (t < ni_pi->cac_data.leakage_minimum_temperature)
drivers/gpu/drm/radeon/ni_dpm.c
3072
t = ni_pi->cac_data.leakage_minimum_temperature;
drivers/gpu/drm/radeon/ni_dpm.c
3077
t,
drivers/gpu/drm/radeon/ni_dpm.c
741
u16 v, s32 t,
drivers/gpu/drm/radeon/ni_dpm.c
749
temperature = div64_s64(drm_int2fixp(t), 1000);
drivers/gpu/drm/radeon/ni_dpm.c
764
s32 t,
drivers/gpu/drm/radeon/ni_dpm.c
768
ni_calculate_leakage_for_v_and_t_formula(coeff, v, t, i_leakage, leakage);
drivers/gpu/drm/radeon/nid.h
1337
#define DMA_PACKET(cmd, t, s, n) ((((cmd) & 0xF) << 28) | \
drivers/gpu/drm/radeon/nid.h
1338
(((t) & 0x1) << 23) | \
drivers/gpu/drm/radeon/r100.c
2106
static void r100_cs_track_texture_print(struct r100_cs_track_texture *t)
drivers/gpu/drm/radeon/r100.c
2108
DRM_DEBUG("pitch %d\n", t->pitch);
drivers/gpu/drm/radeon/r100.c
2109
DRM_DEBUG("use_pitch %d\n", t->use_pitch);
drivers/gpu/drm/radeon/r100.c
2110
DRM_DEBUG("width %d\n", t->width);
drivers/gpu/drm/radeon/r100.c
2111
DRM_DEBUG("width_11 %d\n", t->width_11);
drivers/gpu/drm/radeon/r100.c
2112
DRM_DEBUG("height %d\n", t->height);
drivers/gpu/drm/radeon/r100.c
2113
DRM_DEBUG("height_11 %d\n", t->height_11);
drivers/gpu/drm/radeon/r100.c
2114
DRM_DEBUG("num levels %d\n", t->num_levels);
drivers/gpu/drm/radeon/r100.c
2115
DRM_DEBUG("depth %d\n", t->txdepth);
drivers/gpu/drm/radeon/r100.c
2116
DRM_DEBUG("bpp %d\n", t->cpp);
drivers/gpu/drm/radeon/r100.c
2117
DRM_DEBUG("coordinate type %d\n", t->tex_coord_type);
drivers/gpu/drm/radeon/r100.c
2118
DRM_DEBUG("width round to power of 2 %d\n", t->roundup_w);
drivers/gpu/drm/radeon/r100.c
2119
DRM_DEBUG("height round to power of 2 %d\n", t->roundup_h);
drivers/gpu/drm/radeon/r100.c
2120
DRM_DEBUG("compress format %d\n", t->compress_format);
drivers/gpu/drm/radeon/r600_dpm.c
219
int r600_calculate_at(u32 t, u32 h, u32 fh, u32 fl, u32 *tl, u32 *th)
drivers/gpu/drm/radeon/r600_dpm.c
228
t1 = (t * (k - 100));
drivers/gpu/drm/radeon/r600_dpm.c
231
ah = ((a * t) + 5000) / 10000;
drivers/gpu/drm/radeon/r600_dpm.c
234
*th = t - ah;
drivers/gpu/drm/radeon/r600_dpm.c
235
*tl = t + al;
drivers/gpu/drm/radeon/r600_dpm.c
388
void r600_set_sst(struct radeon_device *rdev, u32 t)
drivers/gpu/drm/radeon/r600_dpm.c
390
WREG32_P(CG_SSP, CG_SST(t), ~CG_SST_MASK);
drivers/gpu/drm/radeon/r600_dpm.c
393
void r600_set_git(struct radeon_device *rdev, u32 t)
drivers/gpu/drm/radeon/r600_dpm.c
395
WREG32_P(CG_GIT, CG_GICST(t), ~CG_GICST_MASK);
drivers/gpu/drm/radeon/r600_dpm.c
403
void r600_set_fct(struct radeon_device *rdev, u32 t)
drivers/gpu/drm/radeon/r600_dpm.c
405
WREG32_P(CG_FC_T, FC_T(t), ~FC_T_MASK);
drivers/gpu/drm/radeon/r600_dpm.h
141
int r600_calculate_at(u32 t, u32 h, u32 fh, u32 fl, u32 *tl, u32 *th);
drivers/gpu/drm/radeon/r600_dpm.h
162
void r600_set_sst(struct radeon_device *rdev, u32 t);
drivers/gpu/drm/radeon/r600_dpm.h
163
void r600_set_git(struct radeon_device *rdev, u32 t);
drivers/gpu/drm/radeon/r600_dpm.h
165
void r600_set_fct(struct radeon_device *rdev, u32 t);
drivers/gpu/drm/radeon/r600d.h
645
#define DMA_PACKET(cmd, t, s, n) ((((cmd) & 0xF) << 28) | \
drivers/gpu/drm/radeon/r600d.h
646
(((t) & 0x1) << 23) | \
drivers/gpu/drm/radeon/radeon_connectors.c
741
struct drm_display_mode *t, *mode;
drivers/gpu/drm/radeon/radeon_connectors.c
744
list_for_each_entry_safe(mode, t, &connector->probed_modes, head) {
drivers/gpu/drm/radeon/radeon_connectors.c
754
list_for_each_entry_safe(mode, t, &connector->probed_modes, head) {
drivers/gpu/drm/radeon/radeon_fence.c
1004
signed long t)
drivers/gpu/drm/radeon/radeon_fence.c
1013
return t;
drivers/gpu/drm/radeon/radeon_fence.c
1015
while (t > 0) {
drivers/gpu/drm/radeon/radeon_fence.c
1029
t = -EDEADLK;
drivers/gpu/drm/radeon/radeon_fence.c
1033
t = schedule_timeout(t);
drivers/gpu/drm/radeon/radeon_fence.c
1035
if (t > 0 && intr && signal_pending(current))
drivers/gpu/drm/radeon/radeon_fence.c
1036
t = -ERESTARTSYS;
drivers/gpu/drm/radeon/radeon_fence.c
1042
return t;
drivers/gpu/drm/radeon/radeon_gart.c
244
unsigned int t, p;
drivers/gpu/drm/radeon/radeon_gart.c
251
t = offset / RADEON_GPU_PAGE_SIZE;
drivers/gpu/drm/radeon/radeon_gart.c
252
p = t / (PAGE_SIZE / RADEON_GPU_PAGE_SIZE);
drivers/gpu/drm/radeon/radeon_gart.c
256
for (j = 0; j < (PAGE_SIZE / RADEON_GPU_PAGE_SIZE); j++, t++) {
drivers/gpu/drm/radeon/radeon_gart.c
257
rdev->gart.pages_entry[t] = rdev->dummy_page.entry;
drivers/gpu/drm/radeon/radeon_gart.c
259
radeon_gart_set_page(rdev, t,
drivers/gpu/drm/radeon/radeon_gart.c
289
unsigned int t, p;
drivers/gpu/drm/radeon/radeon_gart.c
297
t = offset / RADEON_GPU_PAGE_SIZE;
drivers/gpu/drm/radeon/radeon_gart.c
298
p = t / (PAGE_SIZE / RADEON_GPU_PAGE_SIZE);
drivers/gpu/drm/radeon/radeon_gart.c
304
for (j = 0; j < (PAGE_SIZE / RADEON_GPU_PAGE_SIZE); j++, t++) {
drivers/gpu/drm/radeon/radeon_gart.c
306
rdev->gart.pages_entry[t] = page_entry;
drivers/gpu/drm/radeon/radeon_gart.c
308
radeon_gart_set_page(rdev, t, page_entry);
drivers/gpu/drm/radeon/rv770d.h
663
#define DMA_PACKET(cmd, t, s, n) ((((cmd) & 0xF) << 28) | \
drivers/gpu/drm/radeon/rv770d.h
664
(((t) & 0x1) << 23) | \
drivers/gpu/drm/radeon/si_dpm.c
1707
u16 v, s32 t, u32 ileakage, u32 *leakage)
drivers/gpu/drm/radeon/si_dpm.c
1715
temperature = div64_s64(drm_int2fixp(t), 1000);
drivers/gpu/drm/radeon/si_dpm.c
1736
s32 t,
drivers/gpu/drm/radeon/si_dpm.c
1740
si_calculate_leakage_for_v_and_t_formula(coeff, v, t, i_leakage, leakage);
drivers/gpu/drm/radeon/si_dpm.c
2521
s32 t;
drivers/gpu/drm/radeon/si_dpm.c
2529
t = (1000 * (i * t_step + t0));
drivers/gpu/drm/radeon/si_dpm.c
2537
t,
drivers/gpu/drm/radeon/sid.h
1853
#define DMA_PACKET(cmd, b, t, s, n) ((((cmd) & 0xF) << 28) | \
drivers/gpu/drm/radeon/sid.h
1855
(((t) & 0x1) << 23) | \
drivers/gpu/drm/radeon/sumo_dpm.c
891
u32 t = 1;
drivers/gpu/drm/radeon/sumo_dpm.c
895
deep_sleep_cntl |= HS(t > 4095 ? 4095 : t);
drivers/gpu/drm/sitronix/st7571-i2c.c
104
t->ignore_nak = true;
drivers/gpu/drm/sitronix/st7571-i2c.c
107
t, &st7571_i2c_regmap_config);
drivers/gpu/drm/sitronix/st7571-i2c.c
43
struct st7571_i2c_transport *t = context;
drivers/gpu/drm/sitronix/st7571-i2c.c
47
.addr = t->client->addr,
drivers/gpu/drm/sitronix/st7571-i2c.c
48
.flags = t->ignore_nak ? I2C_M_IGNORE_NAK : 0,
drivers/gpu/drm/sitronix/st7571-i2c.c
53
ret = i2c_transfer(t->client->adapter, &msg, 1);
drivers/gpu/drm/sitronix/st7571-i2c.c
61
if (ret < 0 && t->ignore_nak)
drivers/gpu/drm/sitronix/st7571-i2c.c
88
struct st7571_i2c_transport *t;
drivers/gpu/drm/sitronix/st7571-i2c.c
91
t = devm_kzalloc(&client->dev, sizeof(*t), GFP_KERNEL);
drivers/gpu/drm/sitronix/st7571-i2c.c
92
if (!t)
drivers/gpu/drm/sitronix/st7571-i2c.c
95
t->client = client;
drivers/gpu/drm/solomon/ssd130x-spi.c
33
struct ssd130x_spi_transport *t = context;
drivers/gpu/drm/solomon/ssd130x-spi.c
34
struct spi_device *spi = t->spi;
drivers/gpu/drm/solomon/ssd130x-spi.c
38
gpiod_set_value_cansleep(t->dc, 0);
drivers/gpu/drm/solomon/ssd130x-spi.c
41
gpiod_set_value_cansleep(t->dc, 1);
drivers/gpu/drm/solomon/ssd130x-spi.c
64
struct ssd130x_spi_transport *t;
drivers/gpu/drm/solomon/ssd130x-spi.c
75
t = devm_kzalloc(dev, sizeof(*t), GFP_KERNEL);
drivers/gpu/drm/solomon/ssd130x-spi.c
76
if (!t)
drivers/gpu/drm/solomon/ssd130x-spi.c
79
t->spi = spi;
drivers/gpu/drm/solomon/ssd130x-spi.c
80
t->dc = dc;
drivers/gpu/drm/solomon/ssd130x-spi.c
82
regmap = devm_regmap_init(dev, NULL, t, &ssd130x_spi_regmap_config);
drivers/gpu/drm/sun4i/sun4i_hdmi.h
102
#define SUN4I_HDMI_PKT_CTRL_TYPE(n, t) ((t) << (((n) % 4) * 4))
drivers/gpu/drm/tests/drm_atomic_state_test.c
331
static void drm_check_in_clone_mode_desc(const struct drm_clone_mode_test *t,
drivers/gpu/drm/tests/drm_atomic_state_test.c
334
sprintf(desc, "%s", t->name);
drivers/gpu/drm/tests/drm_atomic_state_test.c
337
static void drm_check_valid_clones_desc(const struct drm_clone_mode_test *t,
drivers/gpu/drm/tests/drm_atomic_state_test.c
340
sprintf(desc, "%s", t->name);
drivers/gpu/drm/tests/drm_client_modeset_test.c
175
drm_connector_pick_cmdline_mode_desc(const struct drm_connector_pick_cmdline_mode_test *t,
drivers/gpu/drm/tests/drm_client_modeset_test.c
178
sprintf(desc, "%s", t->cmdline);
drivers/gpu/drm/tests/drm_cmdline_parser_test.c
1016
static void drm_cmdline_tv_option_desc(const struct drm_cmdline_tv_option_test *t,
drivers/gpu/drm/tests/drm_cmdline_parser_test.c
1019
sprintf(desc, "%s", t->name);
drivers/gpu/drm/tests/drm_cmdline_parser_test.c
941
static void drm_cmdline_invalid_desc(const struct drm_cmdline_invalid_test *t,
drivers/gpu/drm/tests/drm_cmdline_parser_test.c
944
sprintf(desc, "%s", t->name);
drivers/gpu/drm/tests/drm_connector_test.c
1139
drm_connector_hdmi_init_formats_yuv420_allowed_desc(const struct drm_connector_hdmi_init_formats_yuv420_allowed_test *t,
drivers/gpu/drm/tests/drm_connector_test.c
1143
t->supported_formats, t->yuv420_allowed);
drivers/gpu/drm/tests/drm_connector_test.c
1323
drm_get_tv_mode_from_name_valid_desc(const struct drm_get_tv_mode_from_name_test *t,
drivers/gpu/drm/tests/drm_connector_test.c
1326
sprintf(desc, "%s", t->name);
drivers/gpu/drm/tests/drm_connector_test.c
1384
drm_hdmi_connector_get_broadcast_rgb_name_valid_desc(const struct drm_hdmi_connector_get_broadcast_rgb_name_test *t,
drivers/gpu/drm/tests/drm_connector_test.c
1387
sprintf(desc, "%s", t->expected_name);
drivers/gpu/drm/tests/drm_connector_test.c
1442
drm_hdmi_connector_get_output_format_name_valid_desc(const struct drm_hdmi_connector_get_output_format_name_test *t,
drivers/gpu/drm/tests/drm_connector_test.c
1445
sprintf(desc, "%s", t->expected_name);
drivers/gpu/drm/tests/drm_dp_mst_helper_test.c
207
static void dp_mst_calc_pbn_div_desc(const struct drm_dp_mst_calc_pbn_div_test *t, char *desc)
drivers/gpu/drm/tests/drm_dp_mst_helper_test.c
209
sprintf(desc, "Link rate %d lane count %d", t->link_rate, t->lane_count);
drivers/gpu/drm/tests/drm_dp_mst_helper_test.c
553
drm_dp_mst_sideband_msg_req_desc(const struct drm_dp_mst_sideband_msg_req_test *t, char *desc)
drivers/gpu/drm/tests/drm_dp_mst_helper_test.c
555
strcpy(desc, t->desc);
drivers/gpu/drm/tests/drm_dp_mst_helper_test.c
63
static void dp_mst_calc_pbn_mode_desc(const struct drm_dp_mst_calc_pbn_mode_test *t, char *desc)
drivers/gpu/drm/tests/drm_dp_mst_helper_test.c
65
sprintf(desc, "Clock %d BPP %d DSC %s", t->clock, t->bpp, t->dsc ? "enabled" : "disabled");
drivers/gpu/drm/tests/drm_format_helper_test.c
1309
static void clip_offset_case_desc(struct clip_offset_case *t, char *desc)
drivers/gpu/drm/tests/drm_format_helper_test.c
1311
strscpy(desc, t->name, KUNIT_PARAM_DESC_SIZE);
drivers/gpu/drm/tests/drm_format_helper_test.c
1652
static void fb_memcpy_case_desc(struct fb_memcpy_case *t, char *desc)
drivers/gpu/drm/tests/drm_format_helper_test.c
1654
snprintf(desc, KUNIT_PARAM_DESC_SIZE, "%s: %p4cc", t->name, &t->format);
drivers/gpu/drm/tests/drm_format_helper_test.c
633
static void convert_xrgb8888_case_desc(struct convert_xrgb8888_case *t,
drivers/gpu/drm/tests/drm_format_helper_test.c
636
strscpy(desc, t->name, KUNIT_PARAM_DESC_SIZE);
drivers/gpu/drm/tests/drm_framebuffer_test.c
414
static void drm_framebuffer_test_to_desc(const struct drm_framebuffer_test *t, char *desc)
drivers/gpu/drm/tests/drm_framebuffer_test.c
416
strscpy(desc, t->name, KUNIT_PARAM_DESC_SIZE);
drivers/gpu/drm/tests/drm_framebuffer_test.c
495
check_src_coords_test_to_desc(const struct drm_framebuffer_check_src_coords_case *t,
drivers/gpu/drm/tests/drm_framebuffer_test.c
498
strscpy(desc, t->name, KUNIT_PARAM_DESC_SIZE);
drivers/gpu/drm/tests/drm_panic_test.c
195
static void drm_test_panic_desc(const struct drm_test_mode *t, char *desc)
drivers/gpu/drm/tests/drm_panic_test.c
198
t->fname, t->width, t->height, &t->format);
drivers/gpu/drm/tests/drm_plane_helper_test.c
130
static void drm_check_plane_state_desc(const struct drm_check_plane_state_test *t,
drivers/gpu/drm/tests/drm_plane_helper_test.c
133
sprintf(desc, "%s", t->name);
drivers/gpu/drm/tests/drm_probe_helper_test.c
191
drm_connector_helper_tv_get_modes_desc(const struct drm_connector_helper_tv_get_modes_test *t,
drivers/gpu/drm/tests/drm_probe_helper_test.c
194
sprintf(desc, "%s", t->name);
drivers/gpu/drm/tests/drm_rect_test.c
327
static void drm_rect_intersect_case_desc(const struct drm_rect_intersect_case *t, char *desc)
drivers/gpu/drm/tests/drm_rect_test.c
331
t->description, DRM_RECT_ARG(&t->r1), DRM_RECT_ARG(&t->r2));
drivers/gpu/drm/tests/drm_rect_test.c
400
static void drm_rect_scale_case_desc(const struct drm_rect_scale_case *t, char *desc)
drivers/gpu/drm/tests/drm_rect_test.c
402
strscpy(desc, t->name, KUNIT_PARAM_DESC_SIZE);
drivers/gpu/drm/tests/drm_rect_test.c
482
static void drm_rect_rotate_case_desc(const struct drm_rect_rotate_case *t, char *desc)
drivers/gpu/drm/tests/drm_rect_test.c
484
strscpy(desc, t->name, KUNIT_PARAM_DESC_SIZE);
drivers/gpu/drm/tests/drm_sysfb_modeset_test.c
124
static void sysfb_build_fourcc_list_case_desc(struct sysfb_build_fourcc_list_case *t, char *desc)
drivers/gpu/drm/tests/drm_sysfb_modeset_test.c
126
strscpy(desc, t->name, KUNIT_PARAM_DESC_SIZE);
drivers/gpu/drm/ttm/tests/ttm_bo_test.c
176
static void signal_for_ttm_bo_reserve(struct timer_list *t)
drivers/gpu/drm/ttm/tests/ttm_bo_test.c
178
struct signal_timer *s_timer = timer_container_of(s_timer, t, timer);
drivers/gpu/drm/ttm/tests/ttm_bo_test.c
51
static void ttm_bo_init_case_desc(const struct ttm_bo_test_case *t,
drivers/gpu/drm/ttm/tests/ttm_bo_test.c
54
strscpy(desc, t->description, KUNIT_PARAM_DESC_SIZE);
drivers/gpu/drm/ttm/tests/ttm_bo_validate_test.c
80
static void ttm_bo_validate_case_desc(const struct ttm_bo_validate_test_case *t,
drivers/gpu/drm/ttm/tests/ttm_bo_validate_test.c
83
strscpy(desc, t->description, KUNIT_PARAM_DESC_SIZE);
drivers/gpu/drm/ttm/tests/ttm_device_test.c
142
static void ttm_device_case_desc(const struct ttm_device_test_case *t, char *desc)
drivers/gpu/drm/ttm/tests/ttm_device_test.c
144
strscpy(desc, t->description, KUNIT_PARAM_DESC_SIZE);
drivers/gpu/drm/ttm/tests/ttm_pool_test.c
126
static void ttm_pool_alloc_case_desc(const struct ttm_pool_test_case *t,
drivers/gpu/drm/ttm/tests/ttm_pool_test.c
129
strscpy(desc, t->description, KUNIT_PARAM_DESC_SIZE);
drivers/gpu/drm/ttm/tests/ttm_resource_test.c
100
static void ttm_resource_case_desc(const struct ttm_resource_test_case *t, char *desc)
drivers/gpu/drm/ttm/tests/ttm_resource_test.c
102
strscpy(desc, t->description, KUNIT_PARAM_DESC_SIZE);
drivers/gpu/drm/ttm/tests/ttm_tt_test.c
30
static void ttm_tt_init_case_desc(const struct ttm_tt_test_case *t,
drivers/gpu/drm/ttm/tests/ttm_tt_test.c
33
strscpy(desc, t->description, KUNIT_PARAM_DESC_SIZE);
drivers/gpu/drm/vboxvideo/vbox_hgsmi.c
51
struct hgsmi_buffer_tail *t;
drivers/gpu/drm/vboxvideo/vbox_hgsmi.c
55
total_size = size + sizeof(*h) + sizeof(*t);
drivers/gpu/drm/vboxvideo/vbox_hgsmi.c
60
t = (struct hgsmi_buffer_tail *)((u8 *)h + sizeof(*h) + size);
drivers/gpu/drm/vboxvideo/vbox_hgsmi.c
68
t->reserved = 0;
drivers/gpu/drm/vboxvideo/vbox_hgsmi.c
69
t->checksum = hgsmi_checksum(offset, h, t);
drivers/gpu/drm/vc4/tests/vc4_test_pv_muxing.c
176
static void vc4_test_pv_muxing_desc(const struct pv_muxing_param *t, char *desc)
drivers/gpu/drm/vc4/tests/vc4_test_pv_muxing.c
178
strscpy(desc, t->name, KUNIT_PARAM_DESC_SIZE);
drivers/gpu/drm/vc4/vc4_bo.c
680
static void vc4_bo_cache_time_timer(struct timer_list *t)
drivers/gpu/drm/vc4/vc4_bo.c
682
struct vc4_dev *vc4 = timer_container_of(vc4, t, bo_cache.time_timer);
drivers/gpu/drm/vc4/vc4_gem.c
328
vc4_hangcheck_elapsed(struct timer_list *t)
drivers/gpu/drm/vc4/vc4_gem.c
330
struct vc4_dev *vc4 = timer_container_of(vc4, t, hangcheck.timer);
drivers/gpu/drm/vgem/vgem_fence.c
62
static void vgem_fence_timeout(struct timer_list *t)
drivers/gpu/drm/vgem/vgem_fence.c
64
struct vgem_fence *fence = timer_container_of(fence, t, timer);
drivers/gpu/drm/vkms/tests/vkms_color_test.c
125
KUNIT_EXPECT_EQ(test, lerp_u16(params->a, params->b, params->t), params->expected);
drivers/gpu/drm/vkms/tests/vkms_color_test.c
36
s64 t;
drivers/gpu/drm/vkms/tests/vkms_format_test.c
255
static void vkms_format_test_yuv_u16_to_argb_u16_case_desc(struct yuv_u16_to_argb_u16_case *t,
drivers/gpu/drm/vkms/tests/vkms_format_test.c
259
drm_get_color_encoding_name(t->encoding), drm_get_color_range_name(t->range));
drivers/gpu/drm/vkms/vkms_composer.c
67
VISIBLE_IF_KUNIT u16 lerp_u16(u16 a, u16 b, s64 t)
drivers/gpu/drm/vkms/vkms_composer.c
72
s64 delta = drm_fixp_mul(b_fp - a_fp, t);
drivers/gpu/drm/vkms/vkms_composer.h
21
u16 lerp_u16(u16 a, u16 b, s64 t);
drivers/gpu/drm/xe/regs/xe_irq_regs.h
16
#define DG1_MSTR_TILE(t) REG_BIT(t)
drivers/gpu/drm/xe/tests/xe_guc_g2g_test.c
302
int i, t;
drivers/gpu/drm/xe/tests/xe_guc_g2g_test.c
313
for (t = 0; t < XE_G2G_TYPE_LIMIT; t++)
drivers/gpu/drm/xe/tests/xe_guc_g2g_test.c
314
guc_g2g_deregister(guc, tile, dev, t);
drivers/gpu/drm/xe/tests/xe_guc_g2g_test.c
569
int t, ret;
drivers/gpu/drm/xe/tests/xe_guc_g2g_test.c
591
for (t = 0; t < XE_G2G_TYPE_LIMIT; t++) {
drivers/gpu/drm/xe/tests/xe_guc_g2g_test.c
592
ret = g2g_register_flat(guc, tile, dev, t, have_dev);
drivers/gpu/drm/xe/tests/xe_rtp_test.c
491
static void rtp_to_sr_desc(const struct rtp_to_sr_test_case *t, char *desc)
drivers/gpu/drm/xe/tests/xe_rtp_test.c
493
strscpy(desc, t->name, KUNIT_PARAM_DESC_SIZE);
drivers/gpu/drm/xe/tests/xe_rtp_test.c
498
static void rtp_desc(const struct rtp_test_case *t, char *desc)
drivers/gpu/drm/xe/tests/xe_rtp_test.c
500
strscpy(desc, t->name, KUNIT_PARAM_DESC_SIZE);
drivers/gpu/drm/xe/xe_bo.c
2353
struct xe_tile *t;
drivers/gpu/drm/xe/xe_bo.c
2363
for_each_tile(t, xe, id) {
drivers/gpu/drm/xe/xe_bo.c
2364
if (t != tile && !(bo->flags & XE_BO_FLAG_GGTTx(t)))
drivers/gpu/drm/xe/xe_bo.c
2368
err = xe_ggtt_insert_bo_at(t->mem.ggtt, bo,
drivers/gpu/drm/xe/xe_bo.c
2372
err = xe_ggtt_insert_bo(t->mem.ggtt, bo, exec);
drivers/gpu/drm/xe/xe_guc.c
522
int t, err;
drivers/gpu/drm/xe/xe_guc.c
548
for (t = 0; t < XE_G2G_TYPE_LIMIT; t++) {
drivers/gpu/drm/xe/xe_guc.c
549
err = guc_g2g_register(guc, far_gt, t, have_dev);
drivers/gpu/drm/xe/xe_guc.c
551
while (--t >= 0)
drivers/gpu/drm/xe/xe_guc.c
552
guc_g2g_deregister(guc, far_tile, far_dev, t);
drivers/gpu/drm/xe/xe_guc.c
573
for (t = 0; t < XE_G2G_TYPE_LIMIT; t++)
drivers/gpu/drm/xe/xe_guc.c
574
guc_g2g_deregister(guc, tile, dev, t);
drivers/gpu/drm/xe/xe_vram.c
135
struct xe_tile *t;
drivers/gpu/drm/xe/xe_vram.c
139
for_each_tile(t, xe, id)
drivers/gpu/drm/xe/xe_vram.c
140
for_each_if(t->id < tile->id)
drivers/gpu/drm/xe/xe_vram.c
141
offset += xe_tile_sriov_vf_lmem(t);
drivers/gpu/drm/xe/xe_wait_user_fence.c
60
unsigned long long t;
drivers/gpu/drm/xe/xe_wait_user_fence.c
82
t = drm_timeout_abs_to_jiffies(args->timeout);
drivers/gpu/drm/xe/xe_wait_user_fence.c
84
t = nsecs_to_jiffies(args->timeout);
drivers/gpu/drm/xe/xe_wait_user_fence.c
92
if (t > MAX_SCHEDULE_TIMEOUT)
drivers/gpu/drm/xe/xe_wait_user_fence.c
95
timeout = t;
drivers/greybus/operation.c
296
static void gb_operation_timeout(struct timer_list *t)
drivers/greybus/operation.c
298
struct gb_operation *operation = timer_container_of(operation, t,
drivers/hid/bpf/hid_bpf_struct_ops.c
112
if (t != state) {
drivers/hid/bpf/hid_bpf_struct_ops.c
133
if (t != state)
drivers/hid/bpf/hid_bpf_struct_ops.c
149
static int hid_bpf_ops_init_member(const struct btf_type *t,
drivers/hid/bpf/hid_bpf_struct_ops.c
160
moff = __btf_member_bit_offset(t, member) / 8;
drivers/hid/bpf/hid_bpf_struct_ops.c
39
static int hid_bpf_ops_check_member(const struct btf_type *t,
drivers/hid/bpf/hid_bpf_struct_ops.c
43
u32 moff = __btf_member_bit_offset(t, member) / 8;
drivers/hid/bpf/hid_bpf_struct_ops.c
85
const struct btf_type *t;
drivers/hid/bpf/hid_bpf_struct_ops.c
89
t = btf_type_by_id(reg->btf, reg->btf_id);
drivers/hid/bpf/hid_bpf_struct_ops.c
98
if (t == state && write_range->struct_name != cur)
drivers/hid/bpf/progs/hid_bpf_async.h
146
err = bpf_timer_init(&elem->t, &hid_bpf_async_ctx_map, CLOCK_MONOTONIC);
drivers/hid/bpf/progs/hid_bpf_async.h
150
err = bpf_timer_set_callback(&elem->t, __start_wq_timer_cb);
drivers/hid/bpf/progs/hid_bpf_async.h
199
err = bpf_timer_set_callback(&elem->t, __start_wq_timer_cb);
drivers/hid/bpf/progs/hid_bpf_async.h
203
err = bpf_timer_start(&elem->t, ms_to_ns(milliseconds), 0);
drivers/hid/bpf/progs/hid_bpf_async.h
27
struct bpf_timer t;
drivers/hid/hid-apple.c
646
static void apple_battery_timer_tick(struct timer_list *t)
drivers/hid/hid-apple.c
648
struct apple_sc *asc = timer_container_of(asc, t, battery_timer);
drivers/hid/hid-appleir.c
168
static void key_up_tick(struct timer_list *t)
drivers/hid/hid-appleir.c
170
struct appleir *appleir = timer_container_of(appleir, t, key_up_timer);
drivers/hid/hid-appletb-kbd.c
167
static void appletb_inactivity_timer(struct timer_list *t)
drivers/hid/hid-appletb-kbd.c
169
struct appletb_kbd *kbd = timer_container_of(kbd, t, inactivity_timer);
drivers/hid/hid-hyperv.c
333
unsigned long t;
drivers/hid/hid-hyperv.c
358
t = wait_for_completion_timeout(&input_dev->wait_event, 5*HZ);
drivers/hid/hid-hyperv.c
359
if (!t) {
drivers/hid/hid-hyperv.c
373
t = wait_for_completion_timeout(&input_dev->wait_event, 5*HZ);
drivers/hid/hid-hyperv.c
374
if (!t) {
drivers/hid/hid-letsketch.c
156
static void letsketch_inrange_timeout(struct timer_list *t)
drivers/hid/hid-letsketch.c
158
struct letsketch_data *data = timer_container_of(data, t,
drivers/hid/hid-magicmouse.c
842
static void magicmouse_battery_timer_tick(struct timer_list *t)
drivers/hid/hid-magicmouse.c
844
struct magicmouse_sc *msc = timer_container_of(msc, t, battery_timer);
drivers/hid/hid-mcp2200.c
105
int len, t;
drivers/hid/hid-mcp2200.c
121
t = wait_for_completion_timeout(&mcp->wait_in_report,
drivers/hid/hid-mcp2200.c
123
if (!t)
drivers/hid/hid-mcp2221.c
178
unsigned long t;
drivers/hid/hid-mcp2221.c
186
t = wait_for_completion_timeout(&mcp->wait_in_report,
drivers/hid/hid-mcp2221.c
188
if (!t)
drivers/hid/hid-multitouch.c
1932
static void mt_expired_timeout(struct timer_list *t)
drivers/hid/hid-multitouch.c
1934
struct mt_device *td = timer_container_of(td, t, release_timer);
drivers/hid/hid-prodikeys.c
228
static void pcmidi_sustained_note_release(struct timer_list *t)
drivers/hid/hid-prodikeys.c
230
struct pcmidi_sustain *pms = timer_container_of(pms, t, timer);
drivers/hid/hid-sony.c
571
static void ghl_magic_poke(struct timer_list *t)
drivers/hid/hid-sony.c
574
struct sony_sc *sc = timer_container_of(sc, t, ghl_poke_timer);
drivers/hid/hid-uclogic-core.c
33
static void uclogic_inrange_timeout(struct timer_list *t)
drivers/hid/hid-uclogic-core.c
35
struct uclogic_drvdata *drvdata = timer_container_of(drvdata, t,
drivers/hid/hid-uclogic-params-test.c
130
static void uclogic_parse_ugee_v2_desc_case_desc(struct uclogic_parse_ugee_v2_desc_case *t,
drivers/hid/hid-uclogic-params-test.c
133
strscpy(desc, t->name, KUNIT_PARAM_DESC_SIZE);
drivers/hid/hid-uclogic-rdesc-test.c
183
static void uclogic_template_case_desc(struct uclogic_template_case *t,
drivers/hid/hid-uclogic-rdesc-test.c
186
strscpy(desc, t->name, KUNIT_PARAM_DESC_SIZE);
drivers/hid/hid-wiimote-core.c
1241
static void wiimote_init_timeout(struct timer_list *t)
drivers/hid/hid-wiimote-core.c
1243
struct wiimote_data *wdata = timer_container_of(wdata, t, timer);
drivers/hid/usbhid/hid-core.c
107
static void hid_retry_timeout(struct timer_list *t)
drivers/hid/usbhid/hid-core.c
109
struct usbhid_device *usbhid = timer_container_of(usbhid, t, io_retry);
drivers/hid/wacom_wac.c
856
unsigned int x, y, distance, t;
drivers/hid/wacom_wac.c
912
t = (data[6] << 3) | ((data[7] & 0xC0) >> 5) | (data[1] & 1);
drivers/hid/wacom_wac.c
914
t >>= 1;
drivers/hid/wacom_wac.c
915
input_report_abs(input, ABS_PRESSURE, t);
drivers/hid/wacom_wac.c
923
input_report_key(input, BTN_TOUCH, t > 10);
drivers/hid/wacom_wac.c
939
t = (data[6] << 3) | ((data[7] >> 5) & 7);
drivers/hid/wacom_wac.c
940
t = (data[7] & 0x20) ? ((t > 900) ? ((t-1) / 2 - 1350) :
drivers/hid/wacom_wac.c
941
((t-1) / 2 + 450)) : (450 - t / 2) ;
drivers/hid/wacom_wac.c
942
input_report_abs(input, ABS_Z, t);
drivers/hid/wacom_wac.c
945
t = (data[6] << 3) | ((data[7] >> 5) & 7);
drivers/hid/wacom_wac.c
947
((t - 1) / 2) : -t / 2);
drivers/hid/wacom_wac.c
959
t = (data[6] << 2) | ((data[7] >> 6) & 3);
drivers/hid/wacom_wac.c
960
input_report_abs(input, ABS_THROTTLE, (data[8] & 0x08) ? -t : t);
drivers/hsi/clients/ssi_protocol.c
454
static void ssip_keep_alive(struct timer_list *t)
drivers/hsi/clients/ssi_protocol.c
456
struct ssi_protocol *ssi = timer_container_of(ssi, t, keep_alive);
drivers/hsi/clients/ssi_protocol.c
481
static void ssip_rx_wd(struct timer_list *t)
drivers/hsi/clients/ssi_protocol.c
483
struct ssi_protocol *ssi = timer_container_of(ssi, t, rx_wd);
drivers/hsi/clients/ssi_protocol.c
490
static void ssip_tx_wd(struct timer_list *t)
drivers/hsi/clients/ssi_protocol.c
492
struct ssi_protocol *ssi = timer_container_of(ssi, t, tx_wd);
drivers/hte/hte-tegra194-test.c
80
static void gpio_timer_cb(struct timer_list *t)
drivers/hte/hte-tegra194-test.c
82
(void)t;
drivers/hv/hv_balloon.c
1721
unsigned long t;
drivers/hv/hv_balloon.c
1756
t = wait_for_completion_timeout(&dm_device.host_event, secs_to_jiffies(5));
drivers/hv/hv_balloon.c
1757
if (t == 0) {
drivers/hv/hv_balloon.c
1814
t = wait_for_completion_timeout(&dm_device.host_event, secs_to_jiffies(5));
drivers/hv/hv_balloon.c
1815
if (t == 0) {
drivers/hwmon/corsair-cpro.c
123
unsigned long t;
drivers/hwmon/corsair-cpro.c
146
t = wait_for_completion_timeout(&ccp->wait_input_report, msecs_to_jiffies(REQ_TIMEOUT));
drivers/hwmon/corsair-cpro.c
147
if (!t)
drivers/hwmon/hih6130.c
137
t = (tmp[0] << 8) + tmp[1];
drivers/hwmon/hih6130.c
138
hih6130->humidity = hih6130_rh_ticks_to_per_cent_mille(t);
drivers/hwmon/hih6130.c
140
t = (tmp[2] << 8) + tmp[3];
drivers/hwmon/hih6130.c
141
hih6130->temperature = hih6130_temp_ticks_to_millicelsius(t);
drivers/hwmon/hih6130.c
86
int t;
drivers/hwmon/hwmon.c
168
long t;
drivers/hwmon/hwmon.c
173
tdata->index, &t);
drivers/hwmon/hwmon.c
177
*temp = t;
drivers/hwmon/lm95234.c
59
static int lm95234_read_temp(struct regmap *regmap, int index, long *t)
drivers/hwmon/lm95234.c
86
*t = DIV_ROUND_CLOSEST(temp * 125, 32);
drivers/hwmon/nct6683.c
418
struct sensor_device_template **t;
drivers/hwmon/nct6683.c
427
t = tg->templates;
drivers/hwmon/nct6683.c
428
for (count = 0; *t; t++, count++)
drivers/hwmon/nct6683.c
452
t = tg->templates;
drivers/hwmon/nct6683.c
453
while (*t) {
drivers/hwmon/nct6683.c
455
(*t)->dev_attr.attr.name, tg->base + i);
drivers/hwmon/nct6683.c
456
if ((*t)->s2) {
drivers/hwmon/nct6683.c
460
a2->nr = (*t)->u.s.nr + i;
drivers/hwmon/nct6683.c
461
a2->index = (*t)->u.s.index;
drivers/hwmon/nct6683.c
463
(*t)->dev_attr.attr.mode;
drivers/hwmon/nct6683.c
464
a2->dev_attr.show = (*t)->dev_attr.show;
drivers/hwmon/nct6683.c
465
a2->dev_attr.store = (*t)->dev_attr.store;
drivers/hwmon/nct6683.c
471
a->index = (*t)->u.index + i;
drivers/hwmon/nct6683.c
473
(*t)->dev_attr.attr.mode;
drivers/hwmon/nct6683.c
474
a->dev_attr.show = (*t)->dev_attr.show;
drivers/hwmon/nct6683.c
475
a->dev_attr.store = (*t)->dev_attr.store;
drivers/hwmon/nct6683.c
480
t++;
drivers/hwmon/nct6775-core.c
1085
struct sensor_device_template **t;
drivers/hwmon/nct6775-core.c
1091
t = tg->templates;
drivers/hwmon/nct6775-core.c
1092
for (count = 0; *t; t++, count++)
drivers/hwmon/nct6775-core.c
1116
t = tg->templates;
drivers/hwmon/nct6775-core.c
1117
while (*t != NULL) {
drivers/hwmon/nct6775-core.c
1119
(*t)->dev_attr.attr.name, tg->base + i);
drivers/hwmon/nct6775-core.c
1120
if ((*t)->s2) {
drivers/hwmon/nct6775-core.c
1124
a2->nr = (*t)->u.s.nr + i;
drivers/hwmon/nct6775-core.c
1125
a2->index = (*t)->u.s.index;
drivers/hwmon/nct6775-core.c
1127
(*t)->dev_attr.attr.mode;
drivers/hwmon/nct6775-core.c
1128
a2->dev_attr.show = (*t)->dev_attr.show;
drivers/hwmon/nct6775-core.c
1129
a2->dev_attr.store = (*t)->dev_attr.store;
drivers/hwmon/nct6775-core.c
1135
a->index = (*t)->u.index + i;
drivers/hwmon/nct6775-core.c
1137
(*t)->dev_attr.attr.mode;
drivers/hwmon/nct6775-core.c
1138
a->dev_attr.show = (*t)->dev_attr.show;
drivers/hwmon/nct6775-core.c
1139
a->dev_attr.store = (*t)->dev_attr.store;
drivers/hwmon/nct6775-core.c
1144
t++;
drivers/hwmon/nct6775-core.c
1428
u8 t = fanmodecfg & 0x0f;
drivers/hwmon/nct6775-core.c
1434
t |= (reg & 0x70) >> 1;
drivers/hwmon/nct6775-core.c
1436
data->target_speed_tolerance[i] = t;
drivers/hwmon/npcm750-pwm-fan.c
323
static void npcm7xx_fan_polling(struct timer_list *t)
drivers/hwmon/npcm750-pwm-fan.c
328
data = timer_container_of(data, t, fan_timer);
drivers/hwmon/pwm-fan.c
80
static void sample_timer(struct timer_list *t)
drivers/hwmon/pwm-fan.c
82
struct pwm_fan_ctx *ctx = timer_container_of(ctx, t, rpm_timer);
drivers/hwtracing/stm/p_sys-t.c
70
#define MIPI_SYST_TYPE(t) ((u32)(MIPI_SYST_TYPE_ ## t))
drivers/hwtracing/stm/p_sys-t.c
82
#define MIPI_SYST_TYPES(t, s) (MIPI_SYST_TYPE(t) | \
drivers/hwtracing/stm/p_sys-t.c
83
MIPI_SYST_SUBTYPE(t ## _ ## s))
drivers/i2c/busses/i2c-at91-master.c
113
filter_width = DIV_ROUND_UP(t->digital_filter_width_ns
drivers/i2c/busses/i2c-at91-master.c
129
cdiv, ckdiv, hold, t->sda_hold_ns, filter_width,
drivers/i2c/busses/i2c-at91-master.c
130
t->digital_filter_width_ns);
drivers/i2c/busses/i2c-at91-master.c
73
struct i2c_timings timings, *t = &timings;
drivers/i2c/busses/i2c-at91-master.c
75
i2c_parse_fw_timings(dev->dev, t, true);
drivers/i2c/busses/i2c-at91-master.c
78
2 * t->bus_freq_hz) - offset);
drivers/i2c/busses/i2c-at91-master.c
95
hold = DIV_ROUND_UP(t->sda_hold_ns
drivers/i2c/busses/i2c-designware-common.c
206
struct i2c_timings *t = &dev->timings;
drivers/i2c/busses/i2c-designware-common.c
214
if (t->bus_freq_hz == supported_speeds[i])
drivers/i2c/busses/i2c-designware-common.c
220
t->bus_freq_hz);
drivers/i2c/busses/i2c-designware-common.c
307
struct i2c_timings *t = &dev->timings;
drivers/i2c/busses/i2c-designware-common.c
319
switch (t->bus_freq_hz) {
drivers/i2c/busses/i2c-designware-common.c
468
struct i2c_timings *t = &dev->timings;
drivers/i2c/busses/i2c-designware-common.c
474
if (acpi_speed && t->bus_freq_hz)
drivers/i2c/busses/i2c-designware-common.c
475
t->bus_freq_hz = min(t->bus_freq_hz, acpi_speed);
drivers/i2c/busses/i2c-designware-common.c
476
else if (acpi_speed || t->bus_freq_hz)
drivers/i2c/busses/i2c-designware-common.c
477
t->bus_freq_hz = max(t->bus_freq_hz, acpi_speed);
drivers/i2c/busses/i2c-designware-common.c
479
t->bus_freq_hz = I2C_MAX_FAST_MODE_FREQ;
drivers/i2c/busses/i2c-designware-common.c
484
struct i2c_timings *t = &dev->timings;
drivers/i2c/busses/i2c-designware-common.c
488
i2c_parse_fw_timings(device, t, false);
drivers/i2c/busses/i2c-designware-common.c
622
struct i2c_timings *t = &dev->timings;
drivers/i2c/busses/i2c-designware-common.c
645
fsleep(DIV_ROUND_CLOSEST_ULL(10 * MICRO, t->bus_freq_hz));
drivers/i2c/busses/i2c-designware-master.c
139
t->bus_freq_hz = I2C_MAX_FAST_MODE_FREQ;
drivers/i2c/busses/i2c-designware-master.c
184
dev_dbg(dev->dev, "Bus speed: %s\n", i2c_freq_mode_string(t->bus_freq_hz));
drivers/i2c/busses/i2c-designware-master.c
38
struct i2c_timings *t = &dev->timings;
drivers/i2c/busses/i2c-designware-master.c
53
sda_falling_time = t->sda_fall_ns ?: 300; /* ns */
drivers/i2c/busses/i2c-designware-master.c
54
scl_falling_time = t->scl_fall_ns ?: 300; /* ns */
drivers/i2c/busses/i2c-designware-master.c
82
if (t->bus_freq_hz == I2C_MAX_FAST_MODE_PLUS_FREQ) {
drivers/i2c/busses/i2c-designware-master.c
926
struct i2c_timings *t = &dev->timings;
drivers/i2c/busses/i2c-designware-master.c
939
switch (t->bus_freq_hz) {
drivers/i2c/busses/i2c-designware-platdrv.c
191
struct i2c_timings *t = &dev->timings;
drivers/i2c/busses/i2c-designware-platdrv.c
197
if (!dev->sda_hold_time && t->sda_hold_ns)
drivers/i2c/busses/i2c-designware-platdrv.c
199
DIV_S64_ROUND_CLOSEST(clk_khz * t->sda_hold_ns, MICRO);
drivers/i2c/busses/i2c-gxp.c
493
writeb(2000000 / drvdata->t.bus_freq_hz,
drivers/i2c/busses/i2c-gxp.c
557
i2c_parse_fw_timings(&pdev->dev, &drvdata->t, true);
drivers/i2c/busses/i2c-gxp.c
79
struct i2c_timings t;
drivers/i2c/busses/i2c-hisi.c
109
struct i2c_timings t;
drivers/i2c/busses/i2c-hisi.c
397
total_cnt = DIV_ROUND_UP_ULL(ctlr->clk_rate_khz * HZ_PER_KHZ, ctlr->t.bus_freq_hz);
drivers/i2c/busses/i2c-hisi.c
403
scl_fall_cnt = NSEC_TO_CYCLES(ctlr->t.scl_fall_ns, ctlr->clk_rate_khz);
drivers/i2c/busses/i2c-hisi.c
405
scl_rise_cnt = NSEC_TO_CYCLES(ctlr->t.scl_rise_ns, ctlr->clk_rate_khz);
drivers/i2c/busses/i2c-hisi.c
419
i2c_parse_fw_timings(ctlr->dev, &ctlr->t, true);
drivers/i2c/busses/i2c-hisi.c
420
ctlr->spk_len = NSEC_TO_CYCLES(ctlr->t.digital_filter_width_ns, ctlr->clk_rate_khz);
drivers/i2c/busses/i2c-hisi.c
422
switch (ctlr->t.bus_freq_hz) {
drivers/i2c/busses/i2c-hisi.c
436
ctlr->t.bus_freq_hz = I2C_MAX_STANDARD_MODE_FREQ;
drivers/i2c/busses/i2c-hisi.c
446
sda_hold_cnt = NSEC_TO_CYCLES(ctlr->t.sda_hold_ns, ctlr->clk_rate_khz);
drivers/i2c/busses/i2c-hisi.c
514
i2c_freq_mode_string(ctlr->t.bus_freq_hz), hw_version);
drivers/i2c/busses/i2c-ibm_iic.c
242
const struct ibm_iic_timings *t = &timings[dev->fast_mode ? 1 : 0];
drivers/i2c/busses/i2c-ibm_iic.c
262
ndelay(t->buf);
drivers/i2c/busses/i2c-ibm_iic.c
267
ndelay(t->hd_sta);
drivers/i2c/busses/i2c-ibm_iic.c
273
ndelay(t->low / 2);
drivers/i2c/busses/i2c-ibm_iic.c
276
ndelay(t->low / 2);
drivers/i2c/busses/i2c-ibm_iic.c
281
ndelay(t->high);
drivers/i2c/busses/i2c-ibm_iic.c
286
ndelay(t->low / 2);
drivers/i2c/busses/i2c-ibm_iic.c
288
ndelay(t->low / 2);
drivers/i2c/busses/i2c-ibm_iic.c
293
ndelay(t->high);
drivers/i2c/busses/i2c-ibm_iic.c
297
ndelay(t->low);
drivers/i2c/busses/i2c-ibm_iic.c
301
ndelay(t->su_sto);
drivers/i2c/busses/i2c-ibm_iic.c
304
ndelay(t->buf);
drivers/i2c/busses/i2c-img-scb.c
832
static void img_i2c_check_timer(struct timer_list *t)
drivers/i2c/busses/i2c-img-scb.c
834
struct img_i2c *i2c = timer_container_of(i2c, t, check_timer);
drivers/i2c/busses/i2c-imx.c
890
static enum hrtimer_restart i2c_imx_slave_timeout(struct hrtimer *t)
drivers/i2c/busses/i2c-imx.c
892
struct imx_i2c_struct *i2c_imx = container_of(t, struct imx_i2c_struct,
drivers/i2c/busses/i2c-ls2x.c
103
i2c_parse_fw_timings(dev, t, false);
drivers/i2c/busses/i2c-ls2x.c
105
if (acpi_speed || t->bus_freq_hz)
drivers/i2c/busses/i2c-ls2x.c
106
t->bus_freq_hz = max(t->bus_freq_hz, acpi_speed);
drivers/i2c/busses/i2c-ls2x.c
108
t->bus_freq_hz = LS2X_I2C_FREQ_STD;
drivers/i2c/busses/i2c-ls2x.c
115
val = LS2X_I2C_PCLK_FREQ / (5 * t->bus_freq_hz) - 1;
drivers/i2c/busses/i2c-ls2x.c
99
struct i2c_timings *t = &priv->i2c_t;
drivers/i2c/busses/i2c-rcar.c
291
struct i2c_timings t = {
drivers/i2c/busses/i2c-rcar.c
299
i2c_parse_fw_timings(dev, &t, false);
drivers/i2c/busses/i2c-rcar.c
326
if (t.bus_freq_hz > I2C_MAX_FAST_MODE_FREQ && priv->devtype >= I2C_RCAR_GEN4)
drivers/i2c/busses/i2c-rcar.c
341
sum = t.scl_fall_ns + t.scl_rise_ns + t.scl_int_delay_ns;
drivers/i2c/busses/i2c-rcar.c
353
scgd = DIV_ROUND_UP(ick, t.bus_freq_hz ?: 1);
drivers/i2c/busses/i2c-rcar.c
361
scl, t.bus_freq_hz, rate, round, cdf, scgd);
drivers/i2c/busses/i2c-rcar.c
381
x = DIV_ROUND_UP(rate, t.bus_freq_hz ?: 1);
drivers/i2c/busses/i2c-rcar.c
395
scl, t.bus_freq_hz, rate, round, cdf, priv->schd, priv->scld, priv->smd);
drivers/i2c/busses/i2c-riic.c
358
struct i2c_timings *t = &riic->i2c_t;
drivers/i2c/busses/i2c-riic.c
364
if (t->bus_freq_hz > max_freq)
drivers/i2c/busses/i2c-riic.c
367
t->bus_freq_hz, max_freq);
drivers/i2c/busses/i2c-riic.c
386
total_ticks = DIV_ROUND_UP(rate, t->bus_freq_hz ?: 1);
drivers/i2c/busses/i2c-riic.c
403
t->bus_freq_hz);
drivers/i2c/busses/i2c-riic.c
421
brl -= t->scl_fall_ns / ns_per_tick;
drivers/i2c/busses/i2c-riic.c
422
brh -= t->scl_rise_ns / ns_per_tick;
drivers/i2c/busses/i2c-riic.c
432
t->scl_fall_ns / ns_per_tick, t->scl_rise_ns / ns_per_tick, cks, brl, brh);
drivers/i2c/busses/i2c-riic.c
449
if (fast_mode_plus && t->bus_freq_hz > I2C_MAX_FAST_MODE_FREQ)
drivers/i2c/busses/i2c-rk3x.c
1255
i2c_parse_fw_timings(&pdev->dev, &i2c->t, true);
drivers/i2c/busses/i2c-rk3x.c
208
struct i2c_timings t;
drivers/i2c/busses/i2c-rk3x.c
567
struct i2c_timings *t,
drivers/i2c/busses/i2c-rk3x.c
586
if (WARN_ON(t->bus_freq_hz > I2C_MAX_FAST_MODE_FREQ))
drivers/i2c/busses/i2c-rk3x.c
587
t->bus_freq_hz = I2C_MAX_FAST_MODE_FREQ;
drivers/i2c/busses/i2c-rk3x.c
590
if (WARN_ON(t->bus_freq_hz < 1000))
drivers/i2c/busses/i2c-rk3x.c
591
t->bus_freq_hz = 1000;
drivers/i2c/busses/i2c-rk3x.c
605
spec = rk3x_i2c_get_spec(t->bus_freq_hz);
drivers/i2c/busses/i2c-rk3x.c
606
min_high_ns = t->scl_rise_ns + spec->min_high_ns;
drivers/i2c/busses/i2c-rk3x.c
617
(t->scl_rise_ns + spec->min_setup_start_ns) * 1000, 875));
drivers/i2c/busses/i2c-rk3x.c
619
(t->scl_rise_ns + spec->min_setup_start_ns + t->sda_fall_ns +
drivers/i2c/busses/i2c-rk3x.c
622
min_low_ns = t->scl_fall_ns + spec->min_low_ns;
drivers/i2c/busses/i2c-rk3x.c
628
scl_rate_khz = t->bus_freq_hz / 1000;
drivers/i2c/busses/i2c-rk3x.c
746
struct i2c_timings *t,
drivers/i2c/busses/i2c-rk3x.c
765
if (WARN_ON(t->bus_freq_hz > I2C_MAX_FAST_MODE_PLUS_FREQ))
drivers/i2c/busses/i2c-rk3x.c
766
t->bus_freq_hz = I2C_MAX_FAST_MODE_PLUS_FREQ;
drivers/i2c/busses/i2c-rk3x.c
769
if (WARN_ON(t->bus_freq_hz < 1000))
drivers/i2c/busses/i2c-rk3x.c
770
t->bus_freq_hz = 1000;
drivers/i2c/busses/i2c-rk3x.c
778
spec = rk3x_i2c_get_spec(t->bus_freq_hz);
drivers/i2c/busses/i2c-rk3x.c
782
scl_rate_khz = t->bus_freq_hz / 1000;
drivers/i2c/busses/i2c-rk3x.c
785
min_high_ns = t->scl_rise_ns + spec->min_high_ns;
drivers/i2c/busses/i2c-rk3x.c
788
min_low_ns = t->scl_fall_ns + spec->min_low_ns;
drivers/i2c/busses/i2c-rk3x.c
845
min_setup_start_ns = t->scl_rise_ns + spec->min_setup_start_ns;
drivers/i2c/busses/i2c-rk3x.c
850
min_setup_stop_ns = t->scl_rise_ns + spec->min_setup_stop_ns;
drivers/i2c/busses/i2c-rk3x.c
877
struct i2c_timings *t = &i2c->t;
drivers/i2c/busses/i2c-rk3x.c
884
ret = i2c->soc_data->calc_timings(clk_rate, t, &calc);
drivers/i2c/busses/i2c-rk3x.c
885
WARN_ONCE(ret != 0, "Could not reach SCL freq %u", t->bus_freq_hz);
drivers/i2c/busses/i2c-rk3x.c
905
HZ_PER_GHZ / t->bus_freq_hz,
drivers/i2c/busses/i2c-rk3x.c
940
if (i2c->soc_data->calc_timings(ndata->new_rate, &i2c->t,
drivers/i2c/busses/i2c-rzv2m.c
106
struct i2c_timings t;
drivers/i2c/busses/i2c-rzv2m.c
109
i2c_parse_fw_timings(dev, &t, true);
drivers/i2c/busses/i2c-rzv2m.c
112
total_pclks = pclk_hz / t.bus_freq_hz;
drivers/i2c/busses/i2c-rzv2m.c
114
trf_ns = t.scl_rise_ns + t.scl_fall_ns;
drivers/i2c/busses/i2c-rzv2m.c
118
switch (t.bus_freq_hz) {
drivers/i2c/busses/i2c-st.c
276
struct st_i2c_timings *t = &i2c_timings[i2c_dev->mode];
drivers/i2c/busses/i2c-st.c
292
val = rate / (2 * t->rate);
drivers/i2c/busses/i2c-st.c
302
val = t->rep_start_hold / ns_per_clk;
drivers/i2c/busses/i2c-st.c
306
val = t->rep_start_setup / ns_per_clk;
drivers/i2c/busses/i2c-st.c
310
val = t->start_hold / ns_per_clk;
drivers/i2c/busses/i2c-st.c
314
val = t->data_setup_time / ns_per_clk;
drivers/i2c/busses/i2c-st.c
318
val = t->stop_setup_time / ns_per_clk;
drivers/i2c/busses/i2c-st.c
322
val = t->bus_free_time / ns_per_clk;
drivers/i2c/busses/i2c-stm32f7.c
668
struct i2c_timings timings, *t = &timings;
drivers/i2c/busses/i2c-stm32f7.c
671
t->bus_freq_hz = I2C_MAX_STANDARD_MODE_FREQ;
drivers/i2c/busses/i2c-stm32f7.c
672
t->scl_rise_ns = i2c_dev->setup.rise_time;
drivers/i2c/busses/i2c-stm32f7.c
673
t->scl_fall_ns = i2c_dev->setup.fall_time;
drivers/i2c/busses/i2c-stm32f7.c
675
i2c_parse_fw_timings(i2c_dev->dev, t, false);
drivers/i2c/busses/i2c-stm32f7.c
677
if (t->bus_freq_hz > I2C_MAX_FAST_MODE_PLUS_FREQ) {
drivers/i2c/busses/i2c-stm32f7.c
679
t->bus_freq_hz, I2C_MAX_FAST_MODE_PLUS_FREQ);
drivers/i2c/busses/i2c-stm32f7.c
683
setup->speed_freq = t->bus_freq_hz;
drivers/i2c/busses/i2c-stm32f7.c
684
i2c_dev->setup.rise_time = t->scl_rise_ns;
drivers/i2c/busses/i2c-stm32f7.c
685
i2c_dev->setup.fall_time = t->scl_fall_ns;
drivers/i2c/busses/i2c-stm32f7.c
686
i2c_dev->dnf_dt = t->digital_filter_width_ns;
drivers/i2c/busses/i2c-stm32f7.c
758
struct stm32f7_i2c_timings *t = &i2c_dev->timing;
drivers/i2c/busses/i2c-stm32f7.c
762
timing |= STM32F7_I2C_TIMINGR_PRESC(t->presc);
drivers/i2c/busses/i2c-stm32f7.c
763
timing |= STM32F7_I2C_TIMINGR_SCLDEL(t->scldel);
drivers/i2c/busses/i2c-stm32f7.c
764
timing |= STM32F7_I2C_TIMINGR_SDADEL(t->sdadel);
drivers/i2c/busses/i2c-stm32f7.c
765
timing |= STM32F7_I2C_TIMINGR_SCLH(t->sclh);
drivers/i2c/busses/i2c-stm32f7.c
766
timing |= STM32F7_I2C_TIMINGR_SCLL(t->scll);
drivers/i2c/busses/i2c-tegra.c
735
struct i2c_timings *t = &i2c_dev->timings;
drivers/i2c/busses/i2c-tegra.c
778
if (WARN_ON(t->bus_freq_hz > max_bus_freq_hz))
drivers/i2c/busses/i2c-tegra.c
779
t->bus_freq_hz = max_bus_freq_hz;
drivers/i2c/busses/i2c-tegra.c
781
if (t->bus_freq_hz <= I2C_MAX_STANDARD_MODE_FREQ) {
drivers/i2c/busses/i2c-tegra.c
786
} else if (t->bus_freq_hz <= I2C_MAX_FAST_MODE_FREQ) {
drivers/i2c/busses/i2c-tegra.c
791
} else if (t->bus_freq_hz <= I2C_MAX_FAST_MODE_PLUS_FREQ) {
drivers/i2c/busses/i2c-tegra.c
843
t->bus_freq_hz * clk_multiplier);
drivers/i2c/i2c-core-base.c
1950
void i2c_parse_fw_timings(struct device *dev, struct i2c_timings *t, bool use_defaults)
drivers/i2c/i2c-core-base.c
1955
i2c_parse_timing(dev, "clock-frequency", &t->bus_freq_hz,
drivers/i2c/i2c-core-base.c
1958
d = t->bus_freq_hz <= I2C_MAX_STANDARD_MODE_FREQ ? 1000 :
drivers/i2c/i2c-core-base.c
1959
t->bus_freq_hz <= I2C_MAX_FAST_MODE_FREQ ? 300 : 120;
drivers/i2c/i2c-core-base.c
1960
i2c_parse_timing(dev, "i2c-scl-rising-time-ns", &t->scl_rise_ns, d, u);
drivers/i2c/i2c-core-base.c
1962
d = t->bus_freq_hz <= I2C_MAX_FAST_MODE_FREQ ? 300 : 120;
drivers/i2c/i2c-core-base.c
1963
i2c_parse_timing(dev, "i2c-scl-falling-time-ns", &t->scl_fall_ns, d, u);
drivers/i2c/i2c-core-base.c
1966
&t->scl_int_delay_ns, 0, u);
drivers/i2c/i2c-core-base.c
1967
i2c_parse_timing(dev, "i2c-sda-falling-time-ns", &t->sda_fall_ns,
drivers/i2c/i2c-core-base.c
1968
t->scl_fall_ns, u);
drivers/i2c/i2c-core-base.c
1969
i2c_parse_timing(dev, "i2c-sda-hold-time-ns", &t->sda_hold_ns, 0, u);
drivers/i2c/i2c-core-base.c
1971
&t->digital_filter_width_ns, 0, u);
drivers/i2c/i2c-core-base.c
1973
&t->analog_filter_cutoff_freq_hz, 0, u);
drivers/i3c/master/i3c-master-cdns.c
247
#define IBIR_THR(t) ((t) << 24)
drivers/i3c/master/i3c-master-cdns.c
248
#define CMDR_THR(t) ((t) << 16)
drivers/i3c/master/i3c-master-cdns.c
249
#define IBI_THR(t) ((t) << 8)
drivers/i3c/master/i3c-master-cdns.c
250
#define CMD_THR(t) (t)
drivers/i3c/master/i3c-master-cdns.c
253
#define RX_THR(t) ((t) << 16)
drivers/i3c/master/i3c-master-cdns.c
254
#define TX_THR(t) (t)
drivers/i3c/master/i3c-master-cdns.c
257
#define SLV_DDR_RX_THR(t) ((t) << 16)
drivers/i3c/master/i3c-master-cdns.c
258
#define SLV_DDR_TX_THR(t) (t)
drivers/i3c/master/renesas-i3c.c
542
struct i2c_timings t;
drivers/i3c/master/renesas-i3c.c
559
i2c_parse_fw_timings(&m->dev, &t, true);
drivers/i3c/master/renesas-i3c.c
599
od_low_ticks -= t.scl_fall_ns / (NSEC_PER_SEC / i3c->rate) + 1;
drivers/i3c/master/renesas-i3c.c
600
od_high_ticks -= t.scl_rise_ns / (NSEC_PER_SEC / i3c->rate) + 1;
drivers/iio/accel/bmc150-accel-core.c
1207
struct bmc150_accel_trigger *t = iio_trigger_get_drvdata(trig);
drivers/iio/accel/bmc150-accel-core.c
1208
struct bmc150_accel_data *data = t->data;
drivers/iio/accel/bmc150-accel-core.c
1213
if (t == &t->data->triggers[BMC150_ACCEL_TRIGGER_DATA_READY])
drivers/iio/accel/bmc150-accel-core.c
1229
struct bmc150_accel_trigger *t = iio_trigger_get_drvdata(trig);
drivers/iio/accel/bmc150-accel-core.c
1230
struct bmc150_accel_data *data = t->data;
drivers/iio/accel/bmc150-accel-core.c
1235
if (t->enabled == state) {
drivers/iio/accel/bmc150-accel-core.c
1240
if (t->setup) {
drivers/iio/accel/bmc150-accel-core.c
1241
ret = t->setup(t, state);
drivers/iio/accel/bmc150-accel-core.c
1248
ret = bmc150_accel_set_interrupt(data, t->intr, state);
drivers/iio/accel/bmc150-accel-core.c
1254
t->enabled = state;
drivers/iio/accel/bmc150-accel-core.c
1385
int (*setup)(struct bmc150_accel_trigger *t, bool state);
drivers/iio/accel/bmc150-accel-core.c
1418
struct bmc150_accel_trigger *t = &data->triggers[i];
drivers/iio/accel/bmc150-accel-core.c
1420
t->indio_trig = devm_iio_trigger_alloc(dev,
drivers/iio/accel/bmc150-accel-core.c
1424
if (!t->indio_trig) {
drivers/iio/accel/bmc150-accel-core.c
1429
t->indio_trig->ops = &bmc150_accel_trigger_ops;
drivers/iio/accel/bmc150-accel-core.c
1430
t->intr = bmc150_accel_triggers[i].intr;
drivers/iio/accel/bmc150-accel-core.c
1431
t->data = data;
drivers/iio/accel/bmc150-accel-core.c
1432
t->setup = bmc150_accel_triggers[i].setup;
drivers/iio/accel/bmc150-accel-core.c
1433
iio_trigger_set_drvdata(t->indio_trig, t);
drivers/iio/accel/bmc150-accel-core.c
1435
ret = iio_trigger_register(t->indio_trig);
drivers/iio/accel/bmc150-accel-core.c
292
static int bmc150_accel_any_motion_setup(struct bmc150_accel_trigger *t,
drivers/iio/accel/bmc150-accel-core.c
296
return bmc150_accel_update_slope(t->data);
drivers/iio/accel/bmc150-accel.h
41
int (*setup)(struct bmc150_accel_trigger *t, bool state);
drivers/iio/accel/ssp_accel_sensor.c
31
u32 t;
drivers/iio/accel/ssp_accel_sensor.c
36
t = ssp_get_sensor_delay(data, SSP_ACCELEROMETER_SENSOR);
drivers/iio/accel/ssp_accel_sensor.c
37
ssp_convert_to_freq(t, val, val2);
drivers/iio/adc/ad4000.c
583
struct spi_transfer t = {
drivers/iio/adc/ad4000.c
591
ret = spi_sync_transfer(st->spi, &t, 1);
drivers/iio/adc/ad4130.c
454
struct spi_transfer t[] = {
drivers/iio/adc/ad4130.c
471
t[1].len = size;
drivers/iio/adc/ad4130.c
473
ret = spi_sync_transfer(st->spi, t, ARRAY_SIZE(t));
drivers/iio/adc/ad7091r8.c
121
struct spi_transfer t[] = {
drivers/iio/adc/ad7091r8.c
137
ret = spi_sync_transfer(spi, t, ARRAY_SIZE(t));
drivers/iio/adc/ad7191.c
91
struct spi_transfer t = {
drivers/iio/adc/ad7191.c
97
spi_message_init_with_transfers(&m, &t, 1);
drivers/iio/adc/ad7280a.c
230
struct spi_transfer t = {
drivers/iio/adc/ad7280a.c
238
ret = spi_sync_transfer(st->spi, &t, 1);
drivers/iio/adc/ad7292.c
124
struct spi_transfer t[] = {
drivers/iio/adc/ad7292.c
141
ret = spi_sync_transfer(st->spi, t, ARRAY_SIZE(t));
drivers/iio/adc/ad7606_spi.c
103
struct spi_transfer t[] = {
drivers/iio/adc/ad7606_spi.c
117
ret = spi_sync_transfer(spi, t, ARRAY_SIZE(t));
drivers/iio/adc/ad7779.c
178
struct spi_transfer t = {
drivers/iio/adc/ad7779.c
187
t.len = 2;
drivers/iio/adc/ad7779.c
189
t.len = 3;
drivers/iio/adc/ad7779.c
191
t.len - 1, 0);
drivers/iio/adc/ad7779.c
194
ret = spi_sync_transfer(st->spi, &t, 1);
drivers/iio/adc/ad7779.c
587
struct spi_transfer t = {
drivers/iio/adc/ad7779.c
594
ret = spi_sync_transfer(st->spi, &t, 1);
drivers/iio/adc/ad7779.c
612
struct spi_transfer t = {
drivers/iio/adc/ad7779.c
624
ret = spi_sync_transfer(st->spi, &t, 1);
drivers/iio/adc/ad_sigma_delta.c
101
spi_message_add_tail(&t, &m);
drivers/iio/adc/ad_sigma_delta.c
125
struct spi_transfer t[] = {
drivers/iio/adc/ad_sigma_delta.c
141
spi_message_add_tail(&t[0], &m);
drivers/iio/adc/ad_sigma_delta.c
143
spi_message_add_tail(&t[1], &m);
drivers/iio/adc/ad_sigma_delta.c
251
struct spi_transfer t[] = {
drivers/iio/adc/ad_sigma_delta.c
300
t[0].tx_buf = data;
drivers/iio/adc/ad_sigma_delta.c
301
spi_message_add_tail(&t[0], &m);
drivers/iio/adc/ad_sigma_delta.c
309
t[1].tx_buf = data + 1;
drivers/iio/adc/ad_sigma_delta.c
312
spi_message_add_tail(&t[1], &m);
drivers/iio/adc/ad_sigma_delta.c
74
struct spi_transfer t = {
drivers/iio/adc/berlin2-adc.c
82
#define BERLIN2_ADC_CHANNEL(n, t) \
drivers/iio/adc/berlin2-adc.c
86
.type = t, \
drivers/iio/adc/dln2-adc.c
473
const struct dln2_adc_demux_table *t;
drivers/iio/adc/dln2-adc.c
484
t = &dln2->demux[i];
drivers/iio/adc/dln2-adc.c
485
memcpy((void *)data.values + t->to,
drivers/iio/adc/dln2-adc.c
486
(void *)dev_data.values + t->from, t->length);
drivers/iio/adc/ep93xx_adc.c
116
u32 t;
drivers/iio/adc/ep93xx_adc.c
118
t = readl_relaxed(priv->base + EP93XX_ADC_RESULT);
drivers/iio/adc/ep93xx_adc.c
119
if (t & EP93XX_ADC_SDR) {
drivers/iio/adc/ep93xx_adc.c
120
*value = sign_extend32(t, 15);
drivers/iio/adc/ltc2496.c
40
struct spi_transfer t = {
drivers/iio/adc/ltc2496.c
49
ret = spi_sync_transfer(st->spi, &t, 1);
drivers/iio/adc/ti-ads124s08.c
200
struct spi_transfer t[] = {
drivers/iio/adc/ti-ads124s08.c
215
ret = spi_sync_transfer(priv->spi, t, ARRAY_SIZE(t));
drivers/iio/adc/ti-ads8688.c
215
struct spi_transfer t[] = {
drivers/iio/adc/ti-ads8688.c
235
ret = spi_sync_transfer(st->spi, t, ARRAY_SIZE(t));
drivers/iio/common/ssp_sensors/ssp_dev.c
168
static void ssp_wdt_timer_func(struct timer_list *t)
drivers/iio/common/ssp_sensors/ssp_dev.c
170
struct ssp_data *data = timer_container_of(data, t, wdt_timer);
drivers/iio/dac/ad5360.c
224
struct spi_transfer t[] = {
drivers/iio/dac/ad5360.c
242
ret = spi_sync_transfer(st->spi, t, ARRAY_SIZE(t));
drivers/iio/dac/ad5421.c
161
struct spi_transfer t[] = {
drivers/iio/dac/ad5421.c
176
ret = spi_sync_transfer(st->spi, t, ARRAY_SIZE(t));
drivers/iio/dac/ad5449.c
103
struct spi_transfer t[] = {
drivers/iio/dac/ad5449.c
119
ret = spi_sync_transfer(st->spi, t, ARRAY_SIZE(t));
drivers/iio/dac/ad5504.c
79
struct spi_transfer t = {
drivers/iio/dac/ad5504.c
86
ret = spi_sync_transfer(st->spi, &t, 1);
drivers/iio/dac/ad5592r.c
22
struct spi_transfer t = {
drivers/iio/dac/ad5592r.c
30
return spi_sync_transfer(spi, &t, 1);
drivers/iio/dac/ad5686-spi.c
51
struct spi_transfer t[] = {
drivers/iio/dac/ad5686-spi.c
83
ret = spi_sync_transfer(spi, t, ARRAY_SIZE(t));
drivers/iio/dac/ad5755.c
299
struct spi_transfer t[] = {
drivers/iio/dac/ad5755.c
316
ret = spi_sync_transfer(st->spi, t, ARRAY_SIZE(t));
drivers/iio/dac/ad5758.c
188
struct spi_transfer t[] = {
drivers/iio/dac/ad5758.c
206
ret = spi_sync_transfer(st->spi, t, ARRAY_SIZE(t));
drivers/iio/dac/ad5764.c
144
struct spi_transfer t[] = {
drivers/iio/dac/ad5764.c
159
ret = spi_sync_transfer(st->spi, t, ARRAY_SIZE(t));
drivers/iio/dac/ad7293.c
173
struct spi_transfer t = {0};
drivers/iio/dac/ad7293.c
185
t.tx_buf = &st->data[0];
drivers/iio/dac/ad7293.c
186
t.rx_buf = &st->data[0];
drivers/iio/dac/ad7293.c
187
t.len = length + 1;
drivers/iio/dac/ad7293.c
189
ret = spi_sync_transfer(st->spi, &t, 1);
drivers/iio/frequency/ad9523.c
308
struct spi_transfer t[] = {
drivers/iio/frequency/ad9523.c
322
ret = spi_sync_transfer(st->spi, t, ARRAY_SIZE(t));
drivers/iio/frequency/ad9523.c
337
struct spi_transfer t[] = {
drivers/iio/frequency/ad9523.c
352
ret = spi_sync_transfer(st->spi, t, ARRAY_SIZE(t));
drivers/iio/frequency/admv1013.c
109
struct spi_transfer t = {0};
drivers/iio/frequency/admv1013.c
115
t.rx_buf = &st->data[0];
drivers/iio/frequency/admv1013.c
116
t.tx_buf = &st->data[0];
drivers/iio/frequency/admv1013.c
117
t.len = 3;
drivers/iio/frequency/admv1013.c
119
ret = spi_sync_transfer(st->spi, &t, 1);
drivers/iio/frequency/admv1014.c
139
struct spi_transfer t = {};
drivers/iio/frequency/admv1014.c
146
t.rx_buf = &st->data[0];
drivers/iio/frequency/admv1014.c
147
t.tx_buf = &st->data[0];
drivers/iio/frequency/admv1014.c
148
t.len = sizeof(st->data);
drivers/iio/frequency/admv1014.c
150
ret = spi_sync_transfer(st->spi, &t, 1);
drivers/iio/frequency/adrf6780.c
101
t.rx_buf = &st->data[0];
drivers/iio/frequency/adrf6780.c
102
t.tx_buf = &st->data[0];
drivers/iio/frequency/adrf6780.c
103
t.len = 3;
drivers/iio/frequency/adrf6780.c
105
ret = spi_sync_transfer(st->spi, &t, 1);
drivers/iio/frequency/adrf6780.c
95
struct spi_transfer t = {0};
drivers/iio/gyro/adis16080.c
56
struct spi_transfer t[] = {
drivers/iio/gyro/adis16080.c
69
ret = spi_sync_transfer(st->us, t, ARRAY_SIZE(t));
drivers/iio/gyro/adis16136.c
169
unsigned int t;
drivers/iio/gyro/adis16136.c
171
t = 32768 / freq;
drivers/iio/gyro/adis16136.c
172
if (t < 0xf)
drivers/iio/gyro/adis16136.c
173
t = 0xf;
drivers/iio/gyro/adis16136.c
174
else if (t > 0xffff)
drivers/iio/gyro/adis16136.c
175
t = 0xffff;
drivers/iio/gyro/adis16136.c
177
t--;
drivers/iio/gyro/adis16136.c
179
return adis_write_reg_16(&adis16136->adis, ADIS16136_REG_SMPL_PRD, t);
drivers/iio/gyro/adis16136.c
184
uint16_t t;
drivers/iio/gyro/adis16136.c
187
ret = __adis_read_reg_16(&adis16136->adis, ADIS16136_REG_SMPL_PRD, &t);
drivers/iio/gyro/adis16136.c
191
*freq = 32768 / (t + 1);
drivers/iio/gyro/adis16260.c
274
u8 t;
drivers/iio/gyro/adis16260.c
291
t = 256 / val;
drivers/iio/gyro/adis16260.c
293
t = 2048 / val;
drivers/iio/gyro/adis16260.c
295
if (t > ADIS16260_SMPL_PRD_DIV_MASK)
drivers/iio/gyro/adis16260.c
296
t = ADIS16260_SMPL_PRD_DIV_MASK;
drivers/iio/gyro/adis16260.c
297
else if (t > 0)
drivers/iio/gyro/adis16260.c
298
t--;
drivers/iio/gyro/adis16260.c
300
if (t >= 0x0A)
drivers/iio/gyro/adis16260.c
304
return __adis_write_reg_8(adis, ADIS16260_SMPL_PRD, t);
drivers/iio/gyro/adxrs290.c
288
unsigned int t;
drivers/iio/gyro/adxrs290.c
337
t = st->lpf_3db_freq_idx;
drivers/iio/gyro/adxrs290.c
338
*val = adxrs290_lpf_3db_freq_hz_table[t][0];
drivers/iio/gyro/adxrs290.c
339
*val2 = adxrs290_lpf_3db_freq_hz_table[t][1];
drivers/iio/gyro/adxrs290.c
347
t = st->hpf_3db_freq_idx;
drivers/iio/gyro/adxrs290.c
348
*val = adxrs290_hpf_3db_freq_hz_table[t][0];
drivers/iio/gyro/adxrs290.c
349
*val2 = adxrs290_hpf_3db_freq_hz_table[t][1];
drivers/iio/gyro/adxrs450.c
232
u32 t;
drivers/iio/gyro/adxrs450.c
238
ret = adxrs450_spi_initial(st, &t, 1);
drivers/iio/gyro/adxrs450.c
241
if (t != 0x01)
drivers/iio/gyro/adxrs450.c
245
ret = adxrs450_spi_initial(st, &t, 0);
drivers/iio/gyro/adxrs450.c
250
ret = adxrs450_spi_initial(st, &t, 0);
drivers/iio/gyro/adxrs450.c
253
if (((t & 0xff) | 0x01) != 0xff || ADXRS450_GET_ST(t) != 2) {
drivers/iio/gyro/adxrs450.c
258
ret = adxrs450_spi_initial(st, &t, 0);
drivers/iio/gyro/adxrs450.c
261
if (((t & 0xff) | 0x01) != 0xff || ADXRS450_GET_ST(t) != 2) {
drivers/iio/gyro/adxrs450.c
305
s16 t;
drivers/iio/gyro/adxrs450.c
311
ret = adxrs450_spi_sensor_data(indio_dev, &t);
drivers/iio/gyro/adxrs450.c
314
*val = t;
drivers/iio/gyro/adxrs450.c
319
ADXRS450_TEMP1, &t);
drivers/iio/gyro/adxrs450.c
322
*val = (t >> 6) + 225;
drivers/iio/gyro/adxrs450.c
344
ret = adxrs450_spi_read_reg_16(indio_dev, ADXRS450_QUAD1, &t);
drivers/iio/gyro/adxrs450.c
347
*val = t;
drivers/iio/gyro/adxrs450.c
351
ret = adxrs450_spi_read_reg_16(indio_dev, ADXRS450_DNC1, &t);
drivers/iio/gyro/adxrs450.c
354
*val = sign_extend32(t, 9);
drivers/iio/gyro/itg3200_core.c
139
u8 t;
drivers/iio/gyro/itg3200_core.c
148
ret = itg3200_read_reg_8(indio_dev, ITG3200_REG_DLPF, &t);
drivers/iio/gyro/itg3200_core.c
153
t = ((t & ITG3200_DLPF_CFG_MASK) ? 1000u : 8000u) / val - 1;
drivers/iio/gyro/itg3200_core.c
157
t);
drivers/iio/gyro/ssp_gyro_sensor.c
31
u32 t;
drivers/iio/gyro/ssp_gyro_sensor.c
36
t = ssp_get_sensor_delay(data, SSP_GYROSCOPE_SENSOR);
drivers/iio/gyro/ssp_gyro_sensor.c
37
ssp_convert_to_freq(t, val, val2);
drivers/iio/humidity/dht11.c
122
int i, t;
drivers/iio/humidity/dht11.c
127
t = dht11->edges[offset + 2 * i + 2].ts -
drivers/iio/humidity/dht11.c
135
bits[i] = t > DHT11_THRESHOLD;
drivers/iio/imu/adis16400.c
295
uint16_t t;
drivers/iio/imu/adis16400.c
297
ret = __adis_read_reg_16(&st->adis, ADIS16400_SMPL_PRD, &t);
drivers/iio/imu/adis16400.c
301
t >>= ADIS16334_RATE_DIV_SHIFT;
drivers/iio/imu/adis16400.c
303
return 819200 >> t;
drivers/iio/imu/adis16400.c
308
unsigned int t;
drivers/iio/imu/adis16400.c
311
t = ilog2(819200 / freq);
drivers/iio/imu/adis16400.c
313
t = 0;
drivers/iio/imu/adis16400.c
315
if (t > 0x31)
drivers/iio/imu/adis16400.c
316
t = 0x31;
drivers/iio/imu/adis16400.c
318
t <<= ADIS16334_RATE_DIV_SHIFT;
drivers/iio/imu/adis16400.c
319
t |= ADIS16334_RATE_INT_CLK;
drivers/iio/imu/adis16400.c
321
return __adis_write_reg_16(&st->adis, ADIS16400_SMPL_PRD, t);
drivers/iio/imu/adis16400.c
327
uint16_t t;
drivers/iio/imu/adis16400.c
329
ret = __adis_read_reg_16(&st->adis, ADIS16400_SMPL_PRD, &t);
drivers/iio/imu/adis16400.c
333
sps = (t & ADIS16400_SMPL_PRD_TIME_BASE) ? 52851 : 1638404;
drivers/iio/imu/adis16400.c
334
sps /= (t & ADIS16400_SMPL_PRD_DIV_MASK) + 1;
drivers/iio/imu/adis16400.c
341
unsigned int t;
drivers/iio/imu/adis16400.c
344
t = 1638404 / freq;
drivers/iio/imu/adis16400.c
345
if (t >= 128) {
drivers/iio/imu/adis16400.c
347
t = 52851 / freq;
drivers/iio/imu/adis16400.c
348
if (t >= 128)
drivers/iio/imu/adis16400.c
349
t = 127;
drivers/iio/imu/adis16400.c
350
} else if (t != 0) {
drivers/iio/imu/adis16400.c
351
t--;
drivers/iio/imu/adis16400.c
354
val |= t;
drivers/iio/imu/adis16400.c
356
if (t >= 0x0A || (val & ADIS16400_SMPL_PRD_TIME_BASE))
drivers/iio/imu/adis16460.c
145
int t;
drivers/iio/imu/adis16460.c
147
t = val * 1000 + val2 / 1000;
drivers/iio/imu/adis16460.c
148
if (t <= 0)
drivers/iio/imu/adis16460.c
151
t = 2048000 / t;
drivers/iio/imu/adis16460.c
152
if (t > 2048)
drivers/iio/imu/adis16460.c
153
t = 2048;
drivers/iio/imu/adis16460.c
155
if (t != 0)
drivers/iio/imu/adis16460.c
156
t--;
drivers/iio/imu/adis16460.c
158
return adis_write_reg_16(&st->adis, ADIS16460_REG_DEC_RATE, t);
drivers/iio/imu/adis16460.c
164
uint16_t t;
drivers/iio/imu/adis16460.c
168
ret = adis_read_reg_16(&st->adis, ADIS16460_REG_DEC_RATE, &t);
drivers/iio/imu/adis16460.c
172
freq = 2048000 / (t + 1);
drivers/iio/imu/adis16480.c
328
unsigned int t, sample_rate = st->clk_freq;
drivers/iio/imu/adis16480.c
334
t = val * 1000 + val2 / 1000;
drivers/iio/imu/adis16480.c
335
if (t == 0)
drivers/iio/imu/adis16480.c
352
unsigned long scaled_rate = lcm(st->clk_freq, t);
drivers/iio/imu/adis16480.c
386
t = DIV_ROUND_CLOSEST(sample_rate, t);
drivers/iio/imu/adis16480.c
387
if (t)
drivers/iio/imu/adis16480.c
388
t--;
drivers/iio/imu/adis16480.c
390
if (t > st->chip_info->max_dec_rate)
drivers/iio/imu/adis16480.c
391
t = st->chip_info->max_dec_rate;
drivers/iio/imu/adis16480.c
393
return __adis_write_reg_16(&st->adis, ADIS16480_REG_DEC_RATE, t);
drivers/iio/imu/adis16480.c
399
uint16_t t;
drivers/iio/imu/adis16480.c
415
ret = __adis_read_reg_16(&st->adis, ADIS16480_REG_DEC_RATE, &t);
drivers/iio/imu/adis16480.c
419
freq = DIV_ROUND_CLOSEST(sample_rate, (t + 1));
drivers/iio/imu/bmi160/bmi160_core.c
311
int bmi160_set_mode(struct bmi160_data *data, enum bmi160_sensor_type t,
drivers/iio/imu/bmi160/bmi160_core.c
318
cmd = bmi160_regs[t].pmu_cmd_normal;
drivers/iio/imu/bmi160/bmi160_core.c
320
cmd = bmi160_regs[t].pmu_cmd_suspend;
drivers/iio/imu/bmi160/bmi160_core.c
326
usleep_range(bmi160_pmu_time[t], bmi160_pmu_time[t] + 1000);
drivers/iio/imu/bmi160/bmi160_core.c
332
int bmi160_set_scale(struct bmi160_data *data, enum bmi160_sensor_type t,
drivers/iio/imu/bmi160/bmi160_core.c
337
for (i = 0; i < bmi160_scale_table[t].num; i++)
drivers/iio/imu/bmi160/bmi160_core.c
338
if (bmi160_scale_table[t].tbl[i].uscale == uscale)
drivers/iio/imu/bmi160/bmi160_core.c
341
if (i == bmi160_scale_table[t].num)
drivers/iio/imu/bmi160/bmi160_core.c
344
return regmap_write(data->regmap, bmi160_regs[t].range,
drivers/iio/imu/bmi160/bmi160_core.c
345
bmi160_scale_table[t].tbl[i].bits);
drivers/iio/imu/bmi160/bmi160_core.c
349
int bmi160_get_scale(struct bmi160_data *data, enum bmi160_sensor_type t,
drivers/iio/imu/bmi160/bmi160_core.c
354
ret = regmap_read(data->regmap, bmi160_regs[t].range, &val);
drivers/iio/imu/bmi160/bmi160_core.c
358
for (i = 0; i < bmi160_scale_table[t].num; i++)
drivers/iio/imu/bmi160/bmi160_core.c
359
if (bmi160_scale_table[t].tbl[i].bits == val) {
drivers/iio/imu/bmi160/bmi160_core.c
360
*uscale = bmi160_scale_table[t].tbl[i].uscale;
drivers/iio/imu/bmi160/bmi160_core.c
373
enum bmi160_sensor_type t = bmi160_to_sensor(chan_type);
drivers/iio/imu/bmi160/bmi160_core.c
375
reg = bmi160_regs[t].data + (axis - IIO_MOD_X) * sizeof(sample);
drivers/iio/imu/bmi160/bmi160_core.c
387
int bmi160_set_odr(struct bmi160_data *data, enum bmi160_sensor_type t,
drivers/iio/imu/bmi160/bmi160_core.c
392
for (i = 0; i < bmi160_odr_table[t].num; i++)
drivers/iio/imu/bmi160/bmi160_core.c
393
if (bmi160_odr_table[t].tbl[i].odr == odr &&
drivers/iio/imu/bmi160/bmi160_core.c
394
bmi160_odr_table[t].tbl[i].uodr == uodr)
drivers/iio/imu/bmi160/bmi160_core.c
397
if (i >= bmi160_odr_table[t].num)
drivers/iio/imu/bmi160/bmi160_core.c
401
bmi160_regs[t].config,
drivers/iio/imu/bmi160/bmi160_core.c
402
bmi160_regs[t].config_odr_mask,
drivers/iio/imu/bmi160/bmi160_core.c
403
bmi160_odr_table[t].tbl[i].bits);
drivers/iio/imu/bmi160/bmi160_core.c
406
static int bmi160_get_odr(struct bmi160_data *data, enum bmi160_sensor_type t,
drivers/iio/imu/bmi160/bmi160_core.c
411
ret = regmap_read(data->regmap, bmi160_regs[t].config, &val);
drivers/iio/imu/bmi160/bmi160_core.c
415
val &= bmi160_regs[t].config_odr_mask;
drivers/iio/imu/bmi160/bmi160_core.c
417
for (i = 0; i < bmi160_odr_table[t].num; i++)
drivers/iio/imu/bmi160/bmi160_core.c
418
if (val == bmi160_odr_table[t].tbl[i].bits)
drivers/iio/imu/bmi160/bmi160_core.c
421
if (i >= bmi160_odr_table[t].num)
drivers/iio/imu/bmi160/bmi160_core.c
424
*odr = bmi160_odr_table[t].tbl[i].odr;
drivers/iio/imu/bmi160/bmi160_core.c
425
*uodr = bmi160_odr_table[t].tbl[i].uodr;
drivers/iio/imu/fxos8700_core.c
330
enum fxos8700_sensor t, bool mode)
drivers/iio/imu/fxos8700_core.c
345
enum fxos8700_sensor t, int uscale)
drivers/iio/imu/fxos8700_core.c
352
if (t == FXOS8700_MAGN) {
drivers/iio/imu/fxos8700_core.c
392
enum fxos8700_sensor t, int *uscale)
drivers/iio/imu/fxos8700_core.c
397
if (t == FXOS8700_MAGN) {
drivers/iio/imu/fxos8700_core.c
479
static int fxos8700_set_odr(struct fxos8700_data *data, enum fxos8700_sensor t,
drivers/iio/imu/fxos8700_core.c
515
static int fxos8700_get_odr(struct fxos8700_data *data, enum fxos8700_sensor t,
drivers/iio/industrialio-buffer.c
2354
struct iio_demux_table *t;
drivers/iio/industrialio-buffer.c
2358
list_for_each_entry(t, &buffer->demux_list, l)
drivers/iio/industrialio-buffer.c
2359
memcpy(buffer->demux_bounce + t->to,
drivers/iio/industrialio-buffer.c
2360
datain + t->from, t->length);
drivers/iio/industrialio-core.c
1193
struct iio_dev_attr *iio_attr, *t;
drivers/iio/industrialio-core.c
1206
list_for_each_entry(t, attr_list, l)
drivers/iio/industrialio-core.c
1207
if (strcmp(t->dev_attr.attr.name,
drivers/iio/industrialio-core.c
1211
t->dev_attr.attr.name);
drivers/iio/industrialio-gts-helper.c
591
static int sanity_check_time(const struct iio_itime_sel_mul *t)
drivers/iio/industrialio-gts-helper.c
593
if (t->sel < 0 || t->time_us < 0 || t->mul <= 0)
drivers/iio/industrialio-gts-helper.c
609
int g, t, ret;
drivers/iio/industrialio-gts-helper.c
614
for (t = 0; t < gts->num_itime; t++) {
drivers/iio/industrialio-gts-helper.c
615
ret = sanity_check_time(&gts->itime_table[t]);
drivers/iio/industrialio-gts-helper.c
627
for (t = 0; t < gts->num_itime; t++) {
drivers/iio/industrialio-gts-helper.c
631
mul = gts->itime_table[t].mul;
drivers/iio/industrialio-sw-trigger.c
102
struct iio_sw_trigger *t;
drivers/iio/industrialio-sw-trigger.c
110
t = tt->ops->probe(name);
drivers/iio/industrialio-sw-trigger.c
111
if (IS_ERR(t))
drivers/iio/industrialio-sw-trigger.c
114
t->trigger_type = tt;
drivers/iio/industrialio-sw-trigger.c
116
return t;
drivers/iio/industrialio-sw-trigger.c
119
return t;
drivers/iio/industrialio-sw-trigger.c
123
void iio_sw_trigger_destroy(struct iio_sw_trigger *t)
drivers/iio/industrialio-sw-trigger.c
125
struct iio_sw_trigger_type *tt = t->trigger_type;
drivers/iio/industrialio-sw-trigger.c
127
tt->ops->remove(t);
drivers/iio/industrialio-sw-trigger.c
135
struct iio_sw_trigger *t;
drivers/iio/industrialio-sw-trigger.c
137
t = iio_sw_trigger_create(group->cg_item.ci_name, name);
drivers/iio/industrialio-sw-trigger.c
138
if (IS_ERR(t))
drivers/iio/industrialio-sw-trigger.c
139
return ERR_CAST(t);
drivers/iio/industrialio-sw-trigger.c
141
config_item_set_name(&t->group.cg_item, "%s", name);
drivers/iio/industrialio-sw-trigger.c
143
return &t->group;
drivers/iio/industrialio-sw-trigger.c
149
struct iio_sw_trigger *t = to_iio_sw_trigger(item);
drivers/iio/industrialio-sw-trigger.c
151
iio_sw_trigger_destroy(t);
drivers/iio/industrialio-sw-trigger.c
32
struct iio_sw_trigger_type *t = NULL, *iter;
drivers/iio/industrialio-sw-trigger.c
36
t = iter;
drivers/iio/industrialio-sw-trigger.c
40
return t;
drivers/iio/industrialio-sw-trigger.c
43
int iio_register_sw_trigger_type(struct iio_sw_trigger_type *t)
drivers/iio/industrialio-sw-trigger.c
49
iter = __iio_find_sw_trigger_type(t->name, strlen(t->name));
drivers/iio/industrialio-sw-trigger.c
53
list_add_tail(&t->list, &iio_trigger_types_list);
drivers/iio/industrialio-sw-trigger.c
59
t->group = configfs_register_default_group(iio_triggers_group, t->name,
drivers/iio/industrialio-sw-trigger.c
61
if (IS_ERR(t->group)) {
drivers/iio/industrialio-sw-trigger.c
63
list_del(&t->list);
drivers/iio/industrialio-sw-trigger.c
65
ret = PTR_ERR(t->group);
drivers/iio/industrialio-sw-trigger.c
72
void iio_unregister_sw_trigger_type(struct iio_sw_trigger_type *t)
drivers/iio/industrialio-sw-trigger.c
77
iter = __iio_find_sw_trigger_type(t->name, strlen(t->name));
drivers/iio/industrialio-sw-trigger.c
79
list_del(&t->list);
drivers/iio/industrialio-sw-trigger.c
82
configfs_unregister_default_group(t->group);
drivers/iio/industrialio-sw-trigger.c
89
struct iio_sw_trigger_type *t;
drivers/iio/industrialio-sw-trigger.c
92
t = __iio_find_sw_trigger_type(name, strlen(name));
drivers/iio/industrialio-sw-trigger.c
93
if (t && !try_module_get(t->owner))
drivers/iio/industrialio-sw-trigger.c
94
t = NULL;
drivers/iio/industrialio-sw-trigger.c
97
return t;
drivers/iio/light/zopt2201.c
195
unsigned long t = zopt2201_resolution[data->res].us;
drivers/iio/light/zopt2201.c
197
if (t <= 20000)
drivers/iio/light/zopt2201.c
198
usleep_range(t, t + 1000);
drivers/iio/light/zopt2201.c
200
msleep(t / 1000);
drivers/iio/magnetometer/als31300.c
106
u16 *t, s16 *x, s16 *y, s16 *z)
drivers/iio/magnetometer/als31300.c
137
*t = ALS31300_TEMPERATURE_GET(buf);
drivers/iio/magnetometer/als31300.c
154
u16 t;
drivers/iio/magnetometer/als31300.c
159
ret = als31300_get_measure(data, &t, &x, &y, &z);
drivers/iio/magnetometer/als31300.c
165
*val = t;
drivers/iio/magnetometer/als31300.c
236
u16 t;
drivers/iio/magnetometer/als31300.c
238
ret = als31300_get_measure(data, &t, &x, &y, &z);
drivers/iio/magnetometer/als31300.c
242
scan.temperature = t;
drivers/iio/magnetometer/tlv493d.c
194
s16 *z, s16 *t)
drivers/iio/magnetometer/tlv493d.c
230
*t = tlv493d_get_channel_data(buff, TLV493D_TEMPERATURE);
drivers/iio/magnetometer/tlv493d.c
274
s16 x, y, z, t;
drivers/iio/magnetometer/tlv493d.c
279
ret = tlv493d_get_measurements(data, &x, &y, &z, &t);
drivers/iio/magnetometer/tlv493d.c
294
*val = t;
drivers/iio/magnetometer/tlv493d.c
343
s16 x, y, z, t;
drivers/iio/magnetometer/tlv493d.c
354
ret = tlv493d_get_measurements(data, &x, &y, &z, &t);
drivers/iio/magnetometer/tlv493d.c
363
scan.temperature = t;
drivers/iio/magnetometer/tmag5273.c
156
static int tmag5273_get_measure(struct tmag5273_data *data, s16 *t, s16 *x,
drivers/iio/magnetometer/tmag5273.c
183
*t = be16_to_cpu(reg_data[0]);
drivers/iio/magnetometer/tmag5273.c
285
s16 t, x, y, z;
drivers/iio/magnetometer/tmag5273.c
295
ret = tmag5273_get_measure(data, &t, &x, &y, &z, &angle, &magnitude);
drivers/iio/magnetometer/tmag5273.c
304
*val = t;
drivers/iio/magnetometer/yamaha-yas530.c
1194
u16 t, x, y1, y2;
drivers/iio/magnetometer/yamaha-yas530.c
1237
ret = yas530_measure(yas5xx, &t, &x, &y1, &y2);
drivers/iio/magnetometer/yamaha-yas530.c
284
static int yas530_measure(struct yas5xx *yas5xx, u16 *t, u16 *x, u16 *y1, u16 *y2)
drivers/iio/magnetometer/yamaha-yas530.c
327
*t = val;
drivers/iio/magnetometer/yamaha-yas530.c
341
*t = val;
drivers/iio/magnetometer/yamaha-yas530.c
368
static int yas537_measure(struct yas5xx *yas5xx, u16 *t, u16 *x, u16 *y1, u16 *y2)
drivers/iio/magnetometer/yamaha-yas530.c
402
*t = get_unaligned_be16(&data[0]);
drivers/iio/magnetometer/yamaha-yas530.c
474
static s32 yas5xx_calc_temperature(struct yas5xx *yas5xx, u16 t)
drivers/iio/magnetometer/yamaha-yas530.c
486
to = (min_temp_x10 + ((ref_temp_x10 - min_temp_x10) * t / t_ref)) * 100;
drivers/iio/magnetometer/yamaha-yas530.c
505
u16 t_ref, t_comp, t, x, y1, y2;
drivers/iio/magnetometer/yamaha-yas530.c
511
ret = yas530_measure(yas5xx, &t, &x, &y1, &y2);
drivers/iio/magnetometer/yamaha-yas530.c
528
t_comp = t - t_ref;
drivers/iio/magnetometer/yamaha-yas530.c
530
t_comp = t;
drivers/iio/magnetometer/yamaha-yas530.c
552
*to = yas5xx_calc_temperature(yas5xx, t);
drivers/iio/magnetometer/yamaha-yas530.c
587
u16 t, x, y1, y2;
drivers/iio/magnetometer/yamaha-yas530.c
591
ret = yas537_measure(yas5xx, &t, &x, &y1, &y2);
drivers/iio/magnetometer/yamaha-yas530.c
596
*to = yas5xx_calc_temperature(yas5xx, t);
drivers/iio/magnetometer/yamaha-yas530.c
618
s32 t, x, y, z;
drivers/iio/magnetometer/yamaha-yas530.c
625
ret = ci->get_measure(yas5xx, &t, &x, &y, &z);
drivers/iio/magnetometer/yamaha-yas530.c
631
*val = t;
drivers/iio/magnetometer/yamaha-yas530.c
661
s32 t, x, y, z;
drivers/iio/magnetometer/yamaha-yas530.c
665
ret = ci->get_measure(yas5xx, &t, &x, &y, &z);
drivers/iio/magnetometer/yamaha-yas530.c
671
yas5xx->scan.channels[0] = t;
drivers/iio/potentiometer/mcp4131.c
150
struct spi_transfer t = {
drivers/iio/potentiometer/mcp4131.c
158
spi_message_add_tail(&t, &m);
drivers/iio/pressure/dps310.c
61
#define DPS310_POLL_SLEEP_US(t) min(30000, (t) / 8)
drivers/iio/pressure/dps310.c
641
s64 t;
drivers/iio/pressure/dps310.c
669
t = (s64)data->temp_raw;
drivers/iio/pressure/dps310.c
680
nums[4] = t * (s64)data->c01;
drivers/iio/pressure/dps310.c
682
nums[5] = t * p * (s64)data->c11;
drivers/iio/pressure/dps310.c
684
nums[6] = t * p * p * (s64)data->c21;
drivers/iio/pressure/dps310.c
752
s64 t;
drivers/iio/pressure/dps310.c
763
t = c0 + ((s64)data->temp_raw * (s64)data->c1);
drivers/iio/pressure/dps310.c
766
*val = (int)div_s64(t * 1000LL, kt);
drivers/iio/pressure/icp10100.c
283
int32_t t, t_square;
drivers/iio/pressure/icp10100.c
291
t = (int32_t)raw_temp - 32768;
drivers/iio/pressure/icp10100.c
292
t_square = t * t;
drivers/iio/pressure/ms5611_core.c
123
s32 t = *temp, p = *pressure;
drivers/iio/pressure/ms5611_core.c
126
dt = t - (st->prom[5] << 8);
drivers/iio/pressure/ms5611_core.c
130
t = 2000 + ((st->prom[6] * dt) >> 23);
drivers/iio/pressure/ms5611_core.c
131
if (t < 2000) {
drivers/iio/pressure/ms5611_core.c
135
off2 = (5 * (t - 2000) * (t - 2000)) >> 1;
drivers/iio/pressure/ms5611_core.c
138
if (t < -1500) {
drivers/iio/pressure/ms5611_core.c
139
s64 tmp = (t + 1500) * (t + 1500);
drivers/iio/pressure/ms5611_core.c
145
t -= t2;
drivers/iio/pressure/ms5611_core.c
150
*temp = t;
drivers/iio/pressure/ms5611_core.c
159
s32 t = *temp, p = *pressure;
drivers/iio/pressure/ms5611_core.c
162
dt = t - (st->prom[5] << 8);
drivers/iio/pressure/ms5611_core.c
166
t = 2000 + ((st->prom[6] * dt) >> 23);
drivers/iio/pressure/ms5611_core.c
167
if (t < 2000) {
drivers/iio/pressure/ms5611_core.c
171
tmp = (t - 2000) * (t - 2000);
drivers/iio/pressure/ms5611_core.c
175
if (t < -1500) {
drivers/iio/pressure/ms5611_core.c
176
tmp = (t + 1500) * (t + 1500);
drivers/iio/pressure/ms5611_core.c
181
t -= t2;
drivers/iio/pressure/ms5611_core.c
186
*temp = t;
drivers/iio/test/iio-test-rescale.c
576
static void case_to_desc(const struct rescale_tc_data *t, char *desc)
drivers/iio/test/iio-test-rescale.c
578
strcpy(desc, t->name);
drivers/iio/test/iio-test-rescale.c
648
struct rescale_tc_data *t = (struct rescale_tc_data *)test->param_value;
drivers/iio/test/iio-test-rescale.c
657
rescale.numerator = t->numerator;
drivers/iio/test/iio-test-rescale.c
658
rescale.denominator = t->denominator;
drivers/iio/test/iio-test-rescale.c
659
rescale.offset = t->offset;
drivers/iio/test/iio-test-rescale.c
660
values[0] = t->schan_val;
drivers/iio/test/iio-test-rescale.c
661
values[1] = t->schan_val2;
drivers/iio/test/iio-test-rescale.c
663
ret = rescale_process_scale(&rescale, t->schan_scale_type,
drivers/iio/test/iio-test-rescale.c
669
rel_ppm = iio_test_relative_error_ppm(buff, t->expected);
drivers/iio/test/iio-test-rescale.c
675
buff, t->expected);
drivers/iio/test/iio-test-rescale.c
680
struct rescale_tc_data *t = (struct rescale_tc_data *)test->param_value;
drivers/iio/test/iio-test-rescale.c
688
rescale.numerator = t->numerator;
drivers/iio/test/iio-test-rescale.c
689
rescale.denominator = t->denominator;
drivers/iio/test/iio-test-rescale.c
690
rescale.offset = t->offset;
drivers/iio/test/iio-test-rescale.c
691
values[0] = t->schan_val;
drivers/iio/test/iio-test-rescale.c
692
values[1] = t->schan_val2;
drivers/iio/test/iio-test-rescale.c
694
ret = rescale_process_offset(&rescale, t->schan_scale_type,
drivers/iio/test/iio-test-rescale.c
695
t->schan_val, t->schan_val2, t->schan_off,
drivers/iio/test/iio-test-rescale.c
701
KUNIT_EXPECT_STREQ(test, strim(buff_off), t->expected_off);
drivers/iio/trigger/iio-trig-sysfs.c
129
struct iio_sysfs_trig *t;
drivers/iio/trigger/iio-trig-sysfs.c
134
list_for_each_entry(t, &iio_sysfs_trig_list, l)
drivers/iio/trigger/iio-trig-sysfs.c
135
if (id == t->id) {
drivers/iio/trigger/iio-trig-sysfs.c
143
t = kmalloc_obj(*t);
drivers/iio/trigger/iio-trig-sysfs.c
144
if (t == NULL) {
drivers/iio/trigger/iio-trig-sysfs.c
148
t->id = id;
drivers/iio/trigger/iio-trig-sysfs.c
149
t->trig = iio_trigger_alloc(&iio_sysfs_trig_dev, "sysfstrig%d", id);
drivers/iio/trigger/iio-trig-sysfs.c
150
if (!t->trig) {
drivers/iio/trigger/iio-trig-sysfs.c
155
t->trig->dev.groups = iio_sysfs_trigger_attr_groups;
drivers/iio/trigger/iio-trig-sysfs.c
156
iio_trigger_set_drvdata(t->trig, t);
drivers/iio/trigger/iio-trig-sysfs.c
158
t->work = IRQ_WORK_INIT_HARD(iio_sysfs_trigger_work);
drivers/iio/trigger/iio-trig-sysfs.c
160
ret = iio_trigger_register(t->trig);
drivers/iio/trigger/iio-trig-sysfs.c
163
list_add(&t->l, &iio_sysfs_trig_list);
drivers/iio/trigger/iio-trig-sysfs.c
169
iio_trigger_free(t->trig);
drivers/iio/trigger/iio-trig-sysfs.c
171
kfree(t);
drivers/iio/trigger/iio-trig-sysfs.c
179
struct iio_sysfs_trig *t = NULL, *iter;
drivers/iio/trigger/iio-trig-sysfs.c
184
t = iter;
drivers/iio/trigger/iio-trig-sysfs.c
187
if (!t) {
drivers/iio/trigger/iio-trig-sysfs.c
192
iio_trigger_unregister(t->trig);
drivers/iio/trigger/iio-trig-sysfs.c
193
irq_work_sync(&t->work);
drivers/iio/trigger/iio-trig-sysfs.c
194
iio_trigger_free(t->trig);
drivers/iio/trigger/iio-trig-sysfs.c
196
list_del(&t->l);
drivers/iio/trigger/iio-trig-sysfs.c
197
kfree(t);
drivers/infiniband/core/mad.c
807
struct ib_rmpp_segment *s, *t;
drivers/infiniband/core/mad.c
809
list_for_each_entry_safe(s, t, &mad_send_wr->rmpp_list, list) {
drivers/infiniband/core/sysfs.c
1357
struct kobject *p, *t;
drivers/infiniband/core/sysfs.c
1359
list_for_each_entry_safe(p, t, &coredev->port_list, entry) {
drivers/infiniband/hw/bng_re/bng_fw.c
215
static void bng_re_service_creq(struct tasklet_struct *t)
drivers/infiniband/hw/bng_re/bng_fw.c
217
struct bng_re_rcfw *rcfw = from_tasklet(rcfw, t, creq.creq_tasklet);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
301
static void bnxt_qplib_service_nq(struct tasklet_struct *t)
drivers/infiniband/hw/bnxt_re/qplib_fp.c
303
struct bnxt_qplib_nq *nq = from_tasklet(nq, t, nq_tasklet);
drivers/infiniband/hw/bnxt_re/qplib_rcfw.c
54
static void bnxt_qplib_service_creq(struct tasklet_struct *t);
drivers/infiniband/hw/bnxt_re/qplib_rcfw.c
733
static void bnxt_qplib_service_creq(struct tasklet_struct *t)
drivers/infiniband/hw/bnxt_re/qplib_rcfw.c
735
struct bnxt_qplib_rcfw *rcfw = from_tasklet(rcfw, t, creq.creq_tasklet);
drivers/infiniband/hw/cxgb4/cm.c
1221
struct tid_info *t = dev->rdev.lldi.tids;
drivers/infiniband/hw/cxgb4/cm.c
1224
ep = lookup_atid(t, atid);
drivers/infiniband/hw/cxgb4/cm.c
1236
cxgb4_insert_tid(t, ep, tid, ep->com.local_addr.ss_family);
drivers/infiniband/hw/cxgb4/cm.c
1247
cxgb4_free_atid(t, atid);
drivers/infiniband/hw/cxgb4/cm.c
143
static void ep_timeout(struct timer_list *t);
drivers/infiniband/hw/cxgb4/cm.c
2274
struct tid_info *t = dev->rdev.lldi.tids;
drivers/infiniband/hw/cxgb4/cm.c
2282
ep = lookup_atid(t, atid);
drivers/infiniband/hw/cxgb4/cm.c
2339
cxgb4_free_atid(t, atid);
drivers/infiniband/hw/cxgb4/cm.c
2376
cxgb4_free_atid(t, atid);
drivers/infiniband/hw/cxgb4/cm.c
2519
struct tid_info *t = dev->rdev.lldi.tids;
drivers/infiniband/hw/cxgb4/cm.c
2652
cxgb4_insert_tid(t, child_ep, hwtid,
drivers/infiniband/hw/cxgb4/cm.c
3830
u64 t;
drivers/infiniband/hw/cxgb4/cm.c
3833
t = (thi << shift) | (tlo >> shift);
drivers/infiniband/hw/cxgb4/cm.c
3835
return t;
drivers/infiniband/hw/cxgb4/cm.c
3841
u64 t = be64_to_cpu(tcb[(31 - word) / 2]);
drivers/infiniband/hw/cxgb4/cm.c
3845
v = (t >> shift) & mask;
drivers/infiniband/hw/cxgb4/cm.c
4328
static void ep_timeout(struct timer_list *t)
drivers/infiniband/hw/cxgb4/cm.c
4330
struct c4iw_ep *ep = timer_container_of(ep, t, timer);
drivers/infiniband/hw/hfi1/aspm.c
170
static void aspm_ctx_timer_function(struct timer_list *t)
drivers/infiniband/hw/hfi1/aspm.c
172
struct hfi1_ctxtdata *rcd = timer_container_of(rcd, t, aspm_timer);
drivers/infiniband/hw/hfi1/chip.c
11602
int fm_get_table(struct hfi1_pportdata *ppd, int which, void *t)
drivers/infiniband/hw/hfi1/chip.c
11616
vl_arb_get_cache(vlc, t);
drivers/infiniband/hw/hfi1/chip.c
11626
vl_arb_get_cache(vlc, t);
drivers/infiniband/hw/hfi1/chip.c
11630
size = get_buffer_control(ppd->dd, t, NULL);
drivers/infiniband/hw/hfi1/chip.c
11633
size = get_sc2vlnt(ppd->dd, t);
drivers/infiniband/hw/hfi1/chip.c
11638
get_vlarb_preempt(ppd->dd, OPA_MAX_VLS, t);
drivers/infiniband/hw/hfi1/chip.c
11656
int fm_set_table(struct hfi1_pportdata *ppd, int which, void *t)
drivers/infiniband/hw/hfi1/chip.c
11664
if (vl_arb_match_cache(vlc, t)) {
drivers/infiniband/hw/hfi1/chip.c
11668
vl_arb_set_cache(vlc, t);
drivers/infiniband/hw/hfi1/chip.c
11671
VL_ARB_HIGH_PRIO_TABLE_SIZE, t);
drivers/infiniband/hw/hfi1/chip.c
11675
if (vl_arb_match_cache(vlc, t)) {
drivers/infiniband/hw/hfi1/chip.c
11679
vl_arb_set_cache(vlc, t);
drivers/infiniband/hw/hfi1/chip.c
11682
VL_ARB_LOW_PRIO_TABLE_SIZE, t);
drivers/infiniband/hw/hfi1/chip.c
11685
ret = set_buffer_control(ppd, t);
drivers/infiniband/hw/hfi1/chip.c
11688
set_sc2vlnt(ppd->dd, t);
drivers/infiniband/hw/hfi1/chip.c
12588
static void update_synth_timer(struct timer_list *t)
drivers/infiniband/hw/hfi1/chip.c
12590
struct hfi1_devdata *dd = timer_container_of(dd, t, synth_stats_timer);
drivers/infiniband/hw/hfi1/chip.c
5549
static void update_rcverr_timer(struct timer_list *t)
drivers/infiniband/hw/hfi1/chip.c
5551
struct hfi1_devdata *dd = timer_container_of(dd, t, rcverr_timer);
drivers/infiniband/hw/hfi1/driver.c
1316
static void run_led_override(struct timer_list *t)
drivers/infiniband/hw/hfi1/driver.c
1318
struct hfi1_pportdata *ppd = timer_container_of(ppd, t,
drivers/infiniband/hw/hfi1/hfi.h
1892
int fm_get_table(struct hfi1_pportdata *ppd, int which, void *t);
drivers/infiniband/hw/hfi1/hfi.h
1893
int fm_set_table(struct hfi1_pportdata *ppd, int which, void *t);
drivers/infiniband/hw/hfi1/init.c
535
static enum hrtimer_restart cca_timer_fn(struct hrtimer *t)
drivers/infiniband/hw/hfi1/init.c
545
cca_timer = container_of(t, struct cca_timer, hrtimer);
drivers/infiniband/hw/hfi1/init.c
577
hrtimer_forward_now(t, ns_to_ktime(nsec));
drivers/infiniband/hw/hfi1/mad.c
370
void hfi1_handle_trap_timer(struct timer_list *t)
drivers/infiniband/hw/hfi1/mad.c
372
struct hfi1_ibport *ibp = timer_container_of(ibp, t, rvp.trap_timer);
drivers/infiniband/hw/hfi1/mad.h
398
void hfi1_handle_trap_timer(struct timer_list *t);
drivers/infiniband/hw/hfi1/sdma.c
467
static void sdma_err_progress_check(struct timer_list *t)
drivers/infiniband/hw/hfi1/sdma.c
470
struct sdma_engine *sde = timer_container_of(sde, t,
drivers/infiniband/hw/hfi1/sdma.c
507
static void sdma_hw_clean_up_task(struct tasklet_struct *t)
drivers/infiniband/hw/hfi1/sdma.c
509
struct sdma_engine *sde = from_tasklet(sde, t,
drivers/infiniband/hw/hfi1/sdma.c
567
static void sdma_sw_clean_up_task(struct tasklet_struct *t)
drivers/infiniband/hw/hfi1/sdma.c
569
struct sdma_engine *sde = from_tasklet(sde, t, sdma_sw_clean_up_task);
drivers/infiniband/hw/hfi1/sysfs.c
509
#define temp_d(t) ((t) >> 2)
drivers/infiniband/hw/hfi1/sysfs.c
510
#define temp_f(t) (((t)&0x3) * 25u)
drivers/infiniband/hw/hfi1/tid_rdma.c
117
static void hfi1_tid_timeout(struct timer_list *t);
drivers/infiniband/hw/hfi1/tid_rdma.c
122
static void hfi1_tid_retry_timeout(struct timer_list *t);
drivers/infiniband/hw/hfi1/tid_rdma.c
3982
static void hfi1_tid_timeout(struct timer_list *t)
drivers/infiniband/hw/hfi1/tid_rdma.c
3984
struct hfi1_qp_priv *qpriv = timer_container_of(qpriv, t, s_tid_timer);
drivers/infiniband/hw/hfi1/tid_rdma.c
4798
static void hfi1_tid_retry_timeout(struct timer_list *t)
drivers/infiniband/hw/hfi1/tid_rdma.c
4800
struct hfi1_qp_priv *priv = timer_container_of(priv, t,
drivers/infiniband/hw/hfi1/user_sdma.c
1196
struct sdma_txreq *t, *p;
drivers/infiniband/hw/hfi1/user_sdma.c
1198
list_for_each_entry_safe(t, p, &req->txps, list) {
drivers/infiniband/hw/hfi1/user_sdma.c
1200
container_of(t, struct user_sdma_txreq, txreq);
drivers/infiniband/hw/hfi1/user_sdma.c
1201
list_del_init(&t->list);
drivers/infiniband/hw/hfi1/user_sdma.c
1202
sdma_txclean(req->pq->dd, t);
drivers/infiniband/hw/hfi1/verbs.c
555
static void mem_timer(struct timer_list *t)
drivers/infiniband/hw/hfi1/verbs.c
557
struct hfi1_ibdev *dev = timer_container_of(dev, t, mem_timer);
drivers/infiniband/hw/irdma/cm.c
1259
static void irdma_cm_timer_tick(struct timer_list *t)
drivers/infiniband/hw/irdma/cm.c
1266
struct irdma_cm_core *cm_core = timer_container_of(cm_core, t,
drivers/infiniband/hw/irdma/hw.c
536
static void irdma_dpc(struct tasklet_struct *t)
drivers/infiniband/hw/irdma/hw.c
538
struct irdma_pci_f *rf = from_tasklet(rf, t, dpc_tasklet);
drivers/infiniband/hw/irdma/hw.c
550
static void irdma_ceq_dpc(struct tasklet_struct *t)
drivers/infiniband/hw/irdma/hw.c
552
struct irdma_ceq *iwceq = from_tasklet(iwceq, t, dpc_tasklet);
drivers/infiniband/hw/irdma/utils.c
1637
static void irdma_hw_stats_timeout(struct timer_list *t)
drivers/infiniband/hw/irdma/utils.c
1640
timer_container_of(pf_devstat, t, stats_timer);
drivers/infiniband/hw/irdma/utils.c
985
static void irdma_terminate_timeout(struct timer_list *t)
drivers/infiniband/hw/irdma/utils.c
987
struct irdma_qp *iwqp = timer_container_of(iwqp, t, terminate_timer);
drivers/infiniband/hw/mlx4/qp.c
3444
u32 *t = dseg;
drivers/infiniband/hw/mlx4/qp.c
3447
t[1] = 0;
drivers/infiniband/hw/mlx4/sysfs.c
697
struct kobject *p, *t;
drivers/infiniband/hw/mlx4/sysfs.c
734
list_for_each_entry_safe(p, t,
drivers/infiniband/hw/mlx4/sysfs.c
772
struct kobject *p, *t;
drivers/infiniband/hw/mlx4/sysfs.c
779
list_for_each_entry_safe(p, t,
drivers/infiniband/hw/mlx5/mr.c
858
static void delay_time_func(struct timer_list *t)
drivers/infiniband/hw/mlx5/mr.c
860
struct mlx5_ib_dev *dev = timer_container_of(dev, t, delay_timer);
drivers/infiniband/hw/mthca/mthca_allocator.c
196
dma_addr_t t;
drivers/infiniband/hw/mthca/mthca_allocator.c
205
size, &t, GFP_KERNEL);
drivers/infiniband/hw/mthca/mthca_allocator.c
209
dma_unmap_addr_set(&buf->direct, mapping, t);
drivers/infiniband/hw/mthca/mthca_allocator.c
211
while (t & ((1 << shift) - 1)) {
drivers/infiniband/hw/mthca/mthca_allocator.c
222
dma_list[i] = t + i * (1 << shift);
drivers/infiniband/hw/mthca/mthca_allocator.c
243
&t, GFP_KERNEL);
drivers/infiniband/hw/mthca/mthca_allocator.c
247
dma_list[i] = t;
drivers/infiniband/hw/mthca/mthca_allocator.c
248
dma_unmap_addr_set(&buf->page_list[i], mapping, t);
drivers/infiniband/hw/mthca/mthca_catas.c
133
static void poll_catas(struct timer_list *t)
drivers/infiniband/hw/mthca/mthca_catas.c
135
struct mthca_dev *dev = timer_container_of(dev, t, catas_err.timer);
drivers/infiniband/hw/mthca/mthca_eq.c
472
dma_addr_t t;
drivers/infiniband/hw/mthca/mthca_eq.c
500
PAGE_SIZE, &t, GFP_KERNEL);
drivers/infiniband/hw/mthca/mthca_eq.c
504
dma_list[i] = t;
drivers/infiniband/hw/mthca/mthca_eq.c
505
dma_unmap_addr_set(&eq->page_list[i], mapping, t);
drivers/infiniband/sw/rdmavt/mr.c
447
static int rvt_check_refs(struct rvt_mregion *mr, const char *t)
drivers/infiniband/sw/rdmavt/mr.c
463
t, mr, mr->pd, mr->lkey,
drivers/infiniband/sw/rdmavt/qp.c
21
static void rvt_rc_timeout(struct timer_list *t);
drivers/infiniband/sw/rdmavt/qp.c
2585
static void rvt_rc_timeout(struct timer_list *t)
drivers/infiniband/sw/rdmavt/qp.c
2587
struct rvt_qp *qp = timer_container_of(qp, t, s_timer);
drivers/infiniband/sw/rdmavt/qp.c
2613
enum hrtimer_restart rvt_rc_rnr_retry(struct hrtimer *t)
drivers/infiniband/sw/rdmavt/qp.c
2615
struct rvt_qp *qp = container_of(t, struct rvt_qp, s_rnr_timer);
drivers/infiniband/sw/rxe/rxe_comp.c
115
void retransmit_timer(struct timer_list *t)
drivers/infiniband/sw/rxe/rxe_comp.c
117
struct rxe_qp *qp = timer_container_of(qp, t, retrans_timer);
drivers/infiniband/sw/rxe/rxe_loc.h
158
void retransmit_timer(struct timer_list *t);
drivers/infiniband/sw/rxe/rxe_loc.h
159
void rnr_nak_timer(struct timer_list *t);
drivers/infiniband/sw/rxe/rxe_req.c
100
struct rxe_qp *qp = timer_container_of(qp, t, rnr_nak_timer);
drivers/infiniband/sw/rxe/rxe_req.c
98
void rnr_nak_timer(struct timer_list *t)
drivers/infiniband/ulp/ipoib/ipoib_cm.c
354
} *t;
drivers/infiniband/ulp/ipoib/ipoib_cm.c
363
t = kmalloc_obj(*t);
drivers/infiniband/ulp/ipoib/ipoib_cm.c
364
if (!t) {
drivers/infiniband/ulp/ipoib/ipoib_cm.c
369
ipoib_cm_init_rx_wr(dev, &t->wr, t->sge);
drivers/infiniband/ulp/ipoib/ipoib_cm.c
391
ret = ipoib_cm_post_receive_nonsrq(dev, rx, &t->wr, t->sge, i);
drivers/infiniband/ulp/ipoib/ipoib_cm.c
402
kfree(t);
drivers/infiniband/ulp/ipoib/ipoib_cm.c
412
kfree(t);
drivers/infiniband/ulp/ipoib/ipoib_ib.c
457
int t;
drivers/infiniband/ulp/ipoib/ipoib_ib.c
466
t = min(IPOIB_NUM_WC, max);
drivers/infiniband/ulp/ipoib/ipoib_ib.c
467
n = ib_poll_cq(priv->recv_cq, t, priv->ibwc);
drivers/infiniband/ulp/ipoib/ipoib_ib.c
483
if (n != t)
drivers/infiniband/ulp/rtrs/rtrs-pri.h
309
struct rtrs_iu *rtrs_iu_alloc(u32 queue_num, size_t size, gfp_t t,
drivers/infiniband/ulp/srp/ib_srp.c
3188
struct srp_target_port *t;
drivers/infiniband/ulp/srp/ib_srp.c
3197
list_for_each_entry(t, &host->target_list, list) {
drivers/infiniband/ulp/srp/ib_srp.c
3198
if (t != target &&
drivers/infiniband/ulp/srp/ib_srp.c
3199
target->id_ext == t->id_ext &&
drivers/infiniband/ulp/srp/ib_srp.c
3200
target->ioc_guid == t->ioc_guid &&
drivers/infiniband/ulp/srp/ib_srp.c
3201
target->initiator_ext == t->initiator_ext) {
drivers/input/evdev.c
1037
unsigned int i, t, u, v;
drivers/input/evdev.c
1207
t = _IOC_NR(cmd) & ABS_MAX;
drivers/input/evdev.c
1208
abs = dev->absinfo[t];
drivers/input/evdev.c
1225
t = _IOC_NR(cmd) & ABS_MAX;
drivers/input/evdev.c
1235
if (t == ABS_MT_SLOT)
drivers/input/evdev.c
1244
dev->absinfo[t] = abs;
drivers/input/ff-memless.c
401
static void ml_effect_timer(struct timer_list *t)
drivers/input/ff-memless.c
403
struct ml_device *ml = timer_container_of(ml, t, timer);
drivers/input/gameport/gameport.c
100
tx = t;
drivers/input/gameport/gameport.c
104
t = 1000000 * 50;
drivers/input/gameport/gameport.c
106
t /= tx;
drivers/input/gameport/gameport.c
107
return t;
drivers/input/gameport/gameport.c
114
unsigned int i, t, t1, t2, t3, tx;
drivers/input/gameport/gameport.c
125
for (t = 0; t < 50; t++) gameport_read(gameport);
drivers/input/gameport/gameport.c
130
if ((t = DELTA(t2,t1) - DELTA(t3,t2)) < tx) tx = t;
drivers/input/gameport/gameport.c
138
unsigned int i, t;
drivers/input/gameport/gameport.c
149
for (t = 0; t < 50; t++) gameport_read(gameport);
drivers/input/gameport/gameport.c
162
unsigned int j, t = 0;
drivers/input/gameport/gameport.c
168
j = jiffies; while (j == jiffies) { t++; gameport_read(gameport); }
drivers/input/gameport/gameport.c
171
return t * HZ / 1000;
drivers/input/gameport/gameport.c
201
static void gameport_run_poll_handler(struct timer_list *t)
drivers/input/gameport/gameport.c
203
struct gameport *gameport = timer_container_of(gameport, t,
drivers/input/gameport/gameport.c
80
unsigned int i, t, tx;
drivers/input/gameport/gameport.c
92
for (t = 0; t < 50; t++)
drivers/input/gameport/gameport.c
98
t = (t2 - t1) - (t3 - t2);
drivers/input/gameport/gameport.c
99
if (t < tx)
drivers/input/gameport/lightning.c
179
int i, t;
drivers/input/gameport/lightning.c
187
t = (max[i] * cal[i]) / 200;
drivers/input/gameport/lightning.c
188
t = (t < 1) ? 1 : ((t > 255) ? 255 : t);
drivers/input/gameport/lightning.c
189
axes[i] = (axes[i] < 0) ? -1 : (axes[i] * cal[i]) / t;
drivers/input/gameport/lightning.c
191
cal[i] = t;
drivers/input/gameport/lightning.c
47
unsigned int t = L4_TIMEOUT;
drivers/input/gameport/lightning.c
49
while ((inb(L4_PORT) & L4_BUSY) && t > 0) t--;
drivers/input/gameport/lightning.c
50
return -(t <= 0);
drivers/input/input.c
2216
static void input_repeat_key(struct timer_list *t)
drivers/input/input.c
2218
struct input_dev *dev = timer_container_of(dev, t, timer);
drivers/input/joydev.c
914
int i, j, t, minor, dev_no;
drivers/input/joydev.c
976
t = (input_abs_get_max(dev, j) + input_abs_get_min(dev, j)) / 2;
drivers/input/joydev.c
977
joydev->corr[i].coef[0] = t - input_abs_get_flat(dev, j);
drivers/input/joydev.c
978
joydev->corr[i].coef[1] = t + input_abs_get_flat(dev, j);
drivers/input/joydev.c
980
t = (input_abs_get_max(dev, j) - input_abs_get_min(dev, j)) / 2
drivers/input/joydev.c
982
if (t) {
drivers/input/joydev.c
983
joydev->corr[i].coef[2] = (1 << 29) / t;
drivers/input/joydev.c
984
joydev->corr[i].coef[3] = (1 << 29) / t;
drivers/input/joystick/a3d.c
56
unsigned int t, s;
drivers/input/joystick/a3d.c
60
t = gameport_time(gameport, A3D_MAX_START);
drivers/input/joystick/a3d.c
67
while (t > 0 && i < length) {
drivers/input/joystick/a3d.c
68
t--;
drivers/input/joystick/a3d.c
72
t = s;
drivers/input/joystick/adi.c
124
int t[2], s[2], i;
drivers/input/joystick/adi.c
129
t[i] = gameport_time(gameport, ADI_MAX_START);
drivers/input/joystick/adi.c
142
t[i]--;
drivers/input/joystick/adi.c
144
if ((w & 0x30) < 0x30 && adi[i].ret < ADI_MAX_LENGTH && t[i] > 0) {
drivers/input/joystick/adi.c
146
t[i] = gameport_time(gameport, ADI_MAX_STROBE);
drivers/input/joystick/adi.c
147
} else t[i] = 0;
drivers/input/joystick/adi.c
150
} while (t[0] > 0 || t[1] > 0);
drivers/input/joystick/adi.c
202
int i, t;
drivers/input/joystick/adi.c
215
t = adi_get_bits(adi, 4);
drivers/input/joystick/adi.c
216
input_report_abs(dev, *abs++, ((t >> 2) & 1) - ( t & 1));
drivers/input/joystick/adi.c
217
input_report_abs(dev, *abs++, ((t >> 1) & 1) - ((t >> 3) & 1));
drivers/input/joystick/adi.c
223
if ((t = adi_get_bits(adi, 4)) > 8) t = 0;
drivers/input/joystick/adi.c
224
input_report_abs(dev, *abs++, adi_hat_to_axis[t].x);
drivers/input/joystick/adi.c
225
input_report_abs(dev, *abs++, adi_hat_to_axis[t].y);
drivers/input/joystick/adi.c
313
int i, t;
drivers/input/joystick/adi.c
318
if (adi->ret < (t = adi_get_bits(adi, 10))) {
drivers/input/joystick/adi.c
319
printk(KERN_WARNING "adi: Short ID packet: reported: %d != read: %d\n", t, adi->ret);
drivers/input/joystick/adi.c
325
if ((t = adi_get_bits(adi, 4)) & ADI_FLAG_HAT) adi->hats++;
drivers/input/joystick/adi.c
349
if (t & ADI_FLAG_10BIT) {
drivers/input/joystick/adi.c
354
t = adi_get_bits(adi, 4);
drivers/input/joystick/adi.c
356
for (i = 0; i < t; i++)
drivers/input/joystick/adi.c
360
t = 8 + adi->buttons + adi->axes10 * 10 + adi->axes8 * 8 + adi->hats * 4;
drivers/input/joystick/adi.c
361
if (adi->length != t && adi->length != t + (t & 1)) {
drivers/input/joystick/adi.c
362
printk(KERN_WARNING "adi: Expected length %d != data length %d\n", t, adi->length);
drivers/input/joystick/adi.c
386
int i, t;
drivers/input/joystick/adi.c
392
t = adi->id < ADI_ID_MAX ? adi->id : ADI_ID_MAX;
drivers/input/joystick/adi.c
394
snprintf(buf, ADI_MAX_PHYS_LENGTH, adi_names[t], adi->id);
drivers/input/joystick/adi.c
398
adi->abs = adi_abs[t];
drivers/input/joystick/adi.c
399
adi->key = adi_key[t];
drivers/input/joystick/adi.c
427
int i, t, x;
drivers/input/joystick/adi.c
434
t = adi->abs[i];
drivers/input/joystick/adi.c
435
x = input_abs_get_val(adi->dev, t);
drivers/input/joystick/adi.c
437
if (t == ABS_THROTTLE || t == ABS_RUDDER || adi->id == ADI_ID_WGPE)
drivers/input/joystick/adi.c
441
input_set_abs_params(adi->dev, t, 64, x * 2 - 64, 2, 16);
drivers/input/joystick/adi.c
443
input_set_abs_params(adi->dev, t, 48, x * 2 - 48, 1, 16);
drivers/input/joystick/adi.c
445
input_set_abs_params(adi->dev, t, -1, 1, 0, 0);
drivers/input/joystick/analog.c
220
int t = 1, i = 0;
drivers/input/joystick/analog.c
232
while ((~u & 0xf0) && (i < 16) && t) {
drivers/input/joystick/analog.c
236
t = strobe;
drivers/input/joystick/analog.c
238
while (((u = gameport_read(port->gameport)) & port->mask) && t) t--;
drivers/input/joystick/analog.c
242
return -(!t || (i == 16));
drivers/input/joystick/analog.c
310
unsigned int i, t, tx;
drivers/input/joystick/analog.c
319
for (t = 0; t < 50; t++) {
drivers/input/joystick/analog.c
326
t = ktime_sub(t2, t1) - ktime_sub(t3, t2);
drivers/input/joystick/analog.c
327
if (t < tx) tx = t;
drivers/input/joystick/analog.c
366
int i, j, t, v, w, x, y, z;
drivers/input/joystick/analog.c
396
t = analog_axes[j];
drivers/input/joystick/analog.c
413
input_set_abs_params(input_dev, t, v, (x << 1) - v, port->fuzz, w);
drivers/input/joystick/analog.c
420
t = analog_hats[j++];
drivers/input/joystick/analog.c
421
input_set_abs_params(input_dev, t, -1, 1, 0, 0);
drivers/input/joystick/analog.c
518
int i, t, u, v;
drivers/input/joystick/analog.c
529
t = gameport_read(gameport);
drivers/input/joystick/analog.c
531
port->mask = (gameport_read(gameport) ^ t) & t & 0xf;
drivers/input/joystick/analog.c
543
t = gameport_time(gameport, ANALOG_MAX_TIME * 1000);
drivers/input/joystick/analog.c
545
while ((gameport_read(port->gameport) & port->mask) && (u < t))
drivers/input/joystick/analog.c
548
t = gameport_time(gameport, ANALOG_SAITEK_TIME);
drivers/input/joystick/analog.c
550
while ((gameport_read(port->gameport) & port->mask) && (v < t))
drivers/input/joystick/cobra.c
42
int r[2], t[2];
drivers/input/joystick/cobra.c
49
t[i] = COBRA_MAX_STROBE;
drivers/input/joystick/cobra.c
57
t[0]--; t[1]--;
drivers/input/joystick/cobra.c
61
if ((w & 0x30) < 0x30 && r[i] < COBRA_LENGTH && t[i] > 0) {
drivers/input/joystick/cobra.c
63
t[i] = strobe;
drivers/input/joystick/cobra.c
65
} else t[i] = 0;
drivers/input/joystick/cobra.c
67
} while (t[0] > 0 || t[1] > 0);
drivers/input/joystick/db9.c
352
static void db9_timer(struct timer_list *t)
drivers/input/joystick/db9.c
354
struct db9 *db9 = timer_container_of(db9, t, timer);
drivers/input/joystick/gamecon.c
727
static void gc_timer(struct timer_list *t)
drivers/input/joystick/gamecon.c
729
struct gc *gc = timer_container_of(gc, t, timer);
drivers/input/joystick/gf2k.c
115
int i, t;
drivers/input/joystick/gf2k.c
122
t = gameport_time(gameport, GF2K_TIMEOUT * 1000);
drivers/input/joystick/gf2k.c
123
while ((gameport_read(gameport) & 1) && t) t--;
drivers/input/joystick/gf2k.c
158
int i, t;
drivers/input/joystick/gf2k.c
166
t = GB(40,4,0);
drivers/input/joystick/gf2k.c
169
input_report_abs(dev, ABS_HAT0X + i, gf2k_hat_to_axis[t][i]);
drivers/input/joystick/gf2k.c
171
t = GB(44,2,0) | GB(32,8,2) | GB(78,2,10);
drivers/input/joystick/gf2k.c
174
input_report_key(dev, gf2k_btn_joy[i], (t >> i) & 1);
drivers/input/joystick/gf2k.c
177
input_report_key(dev, gf2k_btn_pad[i], (t >> i) & 1);
drivers/input/joystick/gf2k.c
79
unsigned int t, p;
drivers/input/joystick/gf2k.c
82
t = gameport_time(gameport, GF2K_START);
drivers/input/joystick/gf2k.c
92
while (t > 0 && i < length) {
drivers/input/joystick/gf2k.c
93
t--; u = v;
drivers/input/joystick/gf2k.c
97
t = p;
drivers/input/joystick/grip.c
110
unsigned int t;
drivers/input/joystick/grip.c
117
t = strobe;
drivers/input/joystick/grip.c
124
t--;
drivers/input/joystick/grip.c
131
t = strobe;
drivers/input/joystick/grip.c
144
t = strobe;
drivers/input/joystick/grip.c
152
} while (status != 0xf && i < GRIP_MAX_BITS_XT && j < GRIP_MAX_CHUNKS_XT && t > 0);
drivers/input/joystick/grip.c
284
int i, j, t;
drivers/input/joystick/grip.c
354
for (j = 0; (t = grip_abs[grip->mode[i]][j]) >= 0; j++) {
drivers/input/joystick/grip.c
357
input_set_abs_params(input_dev, t, 14, 52, 1, 2);
drivers/input/joystick/grip.c
359
input_set_abs_params(input_dev, t, 3, 57, 1, 0);
drivers/input/joystick/grip.c
361
input_set_abs_params(input_dev, t, -1, 1, 0, 0);
drivers/input/joystick/grip.c
364
for (j = 0; (t = grip_btn[grip->mode[i]][j]) >= 0; j++)
drivers/input/joystick/grip.c
365
if (t > 0)
drivers/input/joystick/grip.c
366
set_bit(t, input_dev->keybit);
drivers/input/joystick/grip.c
69
unsigned int t;
drivers/input/joystick/grip.c
75
t = strobe;
drivers/input/joystick/grip.c
83
t--;
drivers/input/joystick/grip.c
87
t = strobe;
drivers/input/joystick/grip.c
89
} while (i < GRIP_LENGTH_GPP && t > 0);
drivers/input/joystick/grip_mp.c
588
int j, t;
drivers/input/joystick/grip_mp.c
609
for (j = 0; (t = grip_abs[port->mode][j]) >= 0; j++)
drivers/input/joystick/grip_mp.c
610
input_set_abs_params(input_dev, t, -1, 1, 0, 0);
drivers/input/joystick/grip_mp.c
612
for (j = 0; (t = grip_btn[port->mode][j]) >= 0; j++)
drivers/input/joystick/grip_mp.c
613
if (t > 0)
drivers/input/joystick/grip_mp.c
614
set_bit(t, input_dev->keybit);
drivers/input/joystick/guillemot.c
163
int i, t;
drivers/input/joystick/guillemot.c
221
for (i = 0; (t = guillemot->type->abs[i]) >= 0; i++)
drivers/input/joystick/guillemot.c
222
input_set_abs_params(input_dev, t, 0, 255, 0, 0);
drivers/input/joystick/guillemot.c
229
for (i = 0; (t = guillemot->type->btn[i]) >= 0; i++)
drivers/input/joystick/guillemot.c
230
set_bit(t, input_dev->keybit);
drivers/input/joystick/guillemot.c
69
unsigned int t, s;
drivers/input/joystick/guillemot.c
76
t = gameport_time(gameport, GUILLEMOT_MAX_START);
drivers/input/joystick/guillemot.c
83
while (t > 0 && i < GUILLEMOT_MAX_LENGTH * 8) {
drivers/input/joystick/guillemot.c
84
t--;
drivers/input/joystick/guillemot.c
89
t = s;
drivers/input/joystick/iforce/iforce-main.c
342
signed short t = iforce->type->abs[i];
drivers/input/joystick/iforce/iforce-main.c
344
switch (t) {
drivers/input/joystick/iforce/iforce-main.c
348
input_set_abs_params(input_dev, t, -1920, 1920, 16, 128);
drivers/input/joystick/iforce/iforce-main.c
349
set_bit(t, input_dev->ffbit);
drivers/input/joystick/iforce/iforce-main.c
355
input_set_abs_params(input_dev, t, 0, 255, 0, 0);
drivers/input/joystick/iforce/iforce-main.c
359
input_set_abs_params(input_dev, t, -128, 127, 0, 0);
drivers/input/joystick/iforce/iforce-main.c
366
input_set_abs_params(input_dev, t, -1, 1, 0, 0);
drivers/input/joystick/interact.c
192
int i, t;
drivers/input/joystick/interact.c
252
for (i = 0; (t = interact_type[interact->type].abs[i]) >= 0; i++) {
drivers/input/joystick/interact.c
254
input_set_abs_params(input_dev, t, 0, 255, 0, 0);
drivers/input/joystick/interact.c
256
input_set_abs_params(input_dev, t, -1, 1, 0, 0);
drivers/input/joystick/interact.c
259
for (i = 0; (t = interact_type[interact->type].btn[i]) >= 0; i++)
drivers/input/joystick/interact.c
260
__set_bit(t, input_dev->keybit);
drivers/input/joystick/interact.c
76
unsigned int t, s;
drivers/input/joystick/interact.c
81
t = gameport_time(gameport, INTERACT_MAX_START);
drivers/input/joystick/interact.c
88
while (t > 0 && i < length) {
drivers/input/joystick/interact.c
89
t--;
drivers/input/joystick/interact.c
96
t = s;
drivers/input/joystick/joydump.c
102
t = i;
drivers/input/joystick/joydump.c
113
for (i = 1; i < t; i++, dump++, prev++) {
drivers/input/joystick/joydump.c
35
int i, j, t, timeout;
drivers/input/joystick/joydump.c
70
t = 0;
drivers/input/joystick/joydump.c
78
dump->time = t;
drivers/input/joystick/joydump.c
83
while (i < BUF_SIZE && t < timeout) {
drivers/input/joystick/joydump.c
89
dump->time = t;
drivers/input/joystick/joydump.c
93
t++;
drivers/input/joystick/magellan.c
67
int i, t;
drivers/input/joystick/magellan.c
85
t = (data[1] << 1) | (data[2] << 5) | data[3];
drivers/input/joystick/magellan.c
86
for (i = 0; i < 9; i++) input_report_key(dev, magellan_buttons[i], (t >> i) & 1);
drivers/input/joystick/n64joy.c
143
static void n64joy_poll(struct timer_list *t)
drivers/input/joystick/n64joy.c
146
struct n64joy_priv *priv = container_of(t, struct n64joy_priv, timer);
drivers/input/joystick/sidewinder.c
226
int i, t;
drivers/input/joystick/sidewinder.c
233
t = gameport_time(gameport, SW_TIMEOUT * 1000);
drivers/input/joystick/sidewinder.c
234
while ((gameport_read(gameport) & 1) && t) t--; /* Wait for axis to fall back to 0 */
drivers/input/joystick/sidewinder.c
247
static int sw_parity(__u64 t)
drivers/input/joystick/sidewinder.c
249
int x = t ^ (t >> 32);
drivers/input/joystick/sidewinder.c
263
static int sw_check(__u64 t)
drivers/input/joystick/sidewinder.c
267
if ((t & 0x8080808080808080ULL) ^ 0x80) /* Sync */
drivers/input/joystick/sidewinder.c
270
while (t) { /* Sum */
drivers/input/joystick/sidewinder.c
271
sum += t & 0xf;
drivers/input/joystick/sidewinder.c
272
t >>= 4;
drivers/input/joystick/tmdc.c
131
int i[2], j[2], t[2], p, k;
drivers/input/joystick/tmdc.c
136
t[k] = gameport_time(gameport, TMDC_MAX_START);
drivers/input/joystick/tmdc.c
151
if (t[k] <= 0 || i[k] >= TMDC_MAX_LENGTH) continue;
drivers/input/joystick/tmdc.c
152
t[k] = p;
drivers/input/joystick/tmdc.c
154
if (~v & 1) t[k] = 0;
drivers/input/joystick/tmdc.c
158
if (v & 1) t[k] = 0;
drivers/input/joystick/tmdc.c
163
t[k]--;
drivers/input/joystick/tmdc.c
165
} while (t[0] > 0 || t[1] > 0);
drivers/input/joystick/turbografx.c
73
static void tgfx_timer(struct timer_list *t)
drivers/input/joystick/turbografx.c
75
struct tgfx *tgfx = timer_container_of(tgfx, t, timer);
drivers/input/keyboard/applespi.c
1016
struct touchpad_protocol *t)
drivers/input/keyboard/applespi.c
1030
for (i = 0; i < t->number_of_fingers; i++) {
drivers/input/keyboard/applespi.c
1031
f = &t->fingers[i];
drivers/input/keyboard/applespi.c
1047
&applespi->pos[i], &t->fingers[i]);
drivers/input/keyboard/applespi.c
1050
input_report_key(input, BTN_LEFT, t->clicked);
drivers/input/keyboard/gpio_keys.c
401
static enum hrtimer_restart gpio_keys_debounce_timer(struct hrtimer *t)
drivers/input/keyboard/gpio_keys.c
404
container_of(t, struct gpio_button_data, debounce_timer);
drivers/input/keyboard/gpio_keys.c
445
static enum hrtimer_restart gpio_keys_irq_timer(struct hrtimer *t)
drivers/input/keyboard/gpio_keys.c
447
struct gpio_button_data *bdata = container_of(t,
drivers/input/keyboard/imx_keypad.c
184
static void imx_keypad_check_for_events(struct timer_list *t)
drivers/input/keyboard/imx_keypad.c
186
struct imx_keypad *keypad = timer_container_of(keypad, t,
drivers/input/keyboard/jornada680_kbd.c
119
}, *t = matrix_switch;
drivers/input/keyboard/jornada680_kbd.c
148
__raw_writeb(*t++, PDDR);
drivers/input/keyboard/jornada680_kbd.c
149
__raw_writeb(*t++, PEDR);
drivers/input/keyboard/lm8323.c
103
#define PWM_RAMP(s, t, n, u) ((!!(s) << 14) | ((t) & 0x3f) << 8 | \
drivers/input/keyboard/locomokbd.c
195
static void locomokbd_timer_callback(struct timer_list *t)
drivers/input/keyboard/locomokbd.c
197
struct locomokbd *locomokbd = timer_container_of(locomokbd, t, timer);
drivers/input/keyboard/mpr121_touchkey.c
141
int i, t, vdd, ret;
drivers/input/keyboard/mpr121_touchkey.c
145
t = ELE0_TOUCH_THRESHOLD_ADDR + (i * 2);
drivers/input/keyboard/mpr121_touchkey.c
146
ret = i2c_smbus_write_byte_data(client, t, TOUCH_THRESHOLD);
drivers/input/keyboard/mpr121_touchkey.c
149
ret = i2c_smbus_write_byte_data(client, t + 1,
drivers/input/keyboard/snvs_pwrkey.c
47
static void imx_imx_snvs_check_for_events(struct timer_list *t)
drivers/input/keyboard/snvs_pwrkey.c
49
struct pwrkey_drv_data *pdata = timer_container_of(pdata, t,
drivers/input/keyboard/tegra-kbc.c
241
static void tegra_kbc_keypress_timer(struct timer_list *t)
drivers/input/keyboard/tegra-kbc.c
243
struct tegra_kbc *kbc = timer_container_of(kbc, t, timer);
drivers/input/keyboard/twl4030_keypad.c
95
#define KEYP_PERIOD_US(t, prescale) ((t) / (31 << ((prescale) + 1)) - 1)
drivers/input/misc/ariel-pwrbutton.c
37
struct spi_transfer t = {
drivers/input/misc/ariel-pwrbutton.c
46
return spi_sync_transfer(spi, &t, 1);
drivers/input/misc/hp_sdc_rtc.c
138
hp_sdc_transaction t;
drivers/input/misc/hp_sdc_rtc.c
153
t.endidx = numreg * 5;
drivers/input/misc/hp_sdc_rtc.c
156
tseq[t.endidx - 4] |= HP_SDC_ACT_SEMAPHORE; /* numreg assumed > 1 */
drivers/input/misc/hp_sdc_rtc.c
158
t.seq = tseq;
drivers/input/misc/hp_sdc_rtc.c
159
t.act.semaphore = &i8042tregs;
drivers/input/misc/hp_sdc_rtc.c
165
if (hp_sdc_enqueue_transaction(&t)) {
drivers/input/misc/hp_sdc_rtc.c
69
hp_sdc_transaction t;
drivers/input/misc/hp_sdc_rtc.c
85
t.endidx = 91;
drivers/input/misc/hp_sdc_rtc.c
86
t.seq = tseq;
drivers/input/misc/hp_sdc_rtc.c
87
t.act.semaphore = &tsem;
drivers/input/misc/hp_sdc_rtc.c
90
if (hp_sdc_enqueue_transaction(&t)) return -1;
drivers/input/misc/nxp-bbnsm-pwrkey.c
46
static void bbnsm_pwrkey_check_for_events(struct timer_list *t)
drivers/input/misc/nxp-bbnsm-pwrkey.c
48
struct bbnsm_pwrkey *bbnsm = timer_container_of(bbnsm, t, check_timer);
drivers/input/misc/yealink.c
61
#define _SEG(t, a, am, b, bm, c, cm, d, dm, e, em, f, fm, g, gm) \
drivers/input/misc/yealink.c
62
{ .type = (t), \
drivers/input/misc/yealink.c
66
#define _PIC(t, h, hm, n) \
drivers/input/misc/yealink.c
67
{ .type = (t), \
drivers/input/mouse/alps.c
1583
static void alps_flush_packet(struct timer_list *t)
drivers/input/mouse/alps.c
1585
struct alps_data *priv = timer_container_of(priv, t, timer);
drivers/input/mouse/byd.c
252
static void byd_clear_touch(struct timer_list *t)
drivers/input/mouse/byd.c
254
struct byd_data *priv = timer_container_of(priv, t, timer);
drivers/input/mouse/cyapa.h
241
#define PIP_DEV_SET_SLEEP_TIME(cyapa, t) ((cyapa)->dev_sleep_time = (t))
drivers/input/mouse/elantech.c
511
u32 t;
drivers/input/mouse/elantech.c
513
t = get_unaligned_le32(&packet[0]);
drivers/input/mouse/elantech.c
515
switch (t & ~7U) {
drivers/input/touchscreen/ad7877.c
376
static void ad7877_timer(struct timer_list *t)
drivers/input/touchscreen/ad7877.c
378
struct ad7877 *ts = timer_container_of(ts, t, timer);
drivers/input/touchscreen/ad7879.c
239
static void ad7879_timer(struct timer_list *t)
drivers/input/touchscreen/ad7879.c
241
struct ad7879 *ts = timer_container_of(ts, t, timer);
drivers/input/touchscreen/bu21029_ts.c
210
static void bu21029_touch_release(struct timer_list *t)
drivers/input/touchscreen/bu21029_ts.c
212
struct bu21029_ts_data *bu21029 = timer_container_of(bu21029, t,
drivers/input/touchscreen/cyttsp5.c
295
int i, t = 0, offset = 0;
drivers/input/touchscreen/cyttsp5.c
322
t = tch->abs[CY_TCH_T];
drivers/input/touchscreen/cyttsp5.c
323
input_mt_slot(ts->input, t);
drivers/input/touchscreen/cyttsp5.c
325
__set_bit(t, ids);
drivers/input/touchscreen/exc3000.c
106
static void exc3000_timer(struct timer_list *t)
drivers/input/touchscreen/exc3000.c
108
struct exc3000_data *data = timer_container_of(data, t, timer);
drivers/input/touchscreen/goodix_berlin_core.c
440
struct goodix_berlin_touch *t;
drivers/input/touchscreen/goodix_berlin_core.c
445
t = &touch_data[i];
drivers/input/touchscreen/goodix_berlin_core.c
447
type = FIELD_GET(GOODIX_BERLIN_POINT_TYPE_MASK, t->status);
drivers/input/touchscreen/goodix_berlin_core.c
454
id = FIELD_GET(GOODIX_BERLIN_TOUCH_ID_MASK, t->status);
drivers/input/touchscreen/goodix_berlin_core.c
464
__le16_to_cpu(t->x), __le16_to_cpu(t->y),
drivers/input/touchscreen/goodix_berlin_core.c
467
__le16_to_cpu(t->w));
drivers/input/touchscreen/sx8654.c
117
static void sx865x_penrelease_timer_handler(struct timer_list *t)
drivers/input/touchscreen/sx8654.c
119
struct sx8654 *ts = timer_container_of(ts, t, timer);
drivers/input/touchscreen/tsc200x-core.c
196
static void tsc200x_penup_timer(struct timer_list *t)
drivers/input/touchscreen/tsc200x-core.c
198
struct tsc200x *ts = timer_container_of(ts, t, penup_timer);
drivers/iommu/amd/iommu.c
50
#define CMD_SET_TYPE(cmd, t) ((cmd)->data[1] |= ((t) << 28))
drivers/iommu/dma-iommu.c
180
static void fq_flush_timeout(struct timer_list *t)
drivers/iommu/dma-iommu.c
182
struct iommu_dma_cookie *cookie = timer_container_of(cookie, t,
drivers/iommu/iommu.c
187
static const char *iommu_domain_type_str(unsigned int t)
drivers/iommu/iommu.c
189
switch (t) {
drivers/iommu/msm_iommu_hw-8xxx.h
37
int t = readl(addr); \
drivers/iommu/msm_iommu_hw-8xxx.h
38
writel((t & ~((mask) << (shift))) + (((v) & (mask)) << (shift)), addr);\
drivers/iommu/sprd-iommu.c
107
u32 t = sprd_iommu_read(sdev, reg);
drivers/iommu/sprd-iommu.c
109
t = (t & (~(mask << shift))) | ((val & mask) << shift);
drivers/iommu/sprd-iommu.c
110
sprd_iommu_write(sdev, reg, t);
drivers/irqchip/irq-ath79-misc.c
66
u32 t;
drivers/irqchip/irq-ath79-misc.c
68
t = __raw_readl(base + AR71XX_RESET_REG_MISC_INT_ENABLE);
drivers/irqchip/irq-ath79-misc.c
69
__raw_writel(t | BIT(irq), base + AR71XX_RESET_REG_MISC_INT_ENABLE);
drivers/irqchip/irq-ath79-misc.c
79
u32 t;
drivers/irqchip/irq-ath79-misc.c
81
t = __raw_readl(base + AR71XX_RESET_REG_MISC_INT_ENABLE);
drivers/irqchip/irq-ath79-misc.c
82
__raw_writel(t & ~BIT(irq), base + AR71XX_RESET_REG_MISC_INT_ENABLE);
drivers/irqchip/irq-ath79-misc.c
92
u32 t;
drivers/irqchip/irq-ath79-misc.c
94
t = __raw_readl(base + AR71XX_RESET_REG_MISC_INT_STATUS);
drivers/irqchip/irq-ath79-misc.c
95
__raw_writel(t & ~BIT(irq), base + AR71XX_RESET_REG_MISC_INT_STATUS);
drivers/irqchip/irq-imx-irqsteer.c
21
#define CHANMASK(n, t) (CTRL_STRIDE_OFF(t, 0) + 0x4 * (n) + 0x4)
drivers/irqchip/irq-imx-irqsteer.c
22
#define CHANSET(n, t) (CTRL_STRIDE_OFF(t, 1) + 0x4 * (n) + 0x4)
drivers/irqchip/irq-imx-irqsteer.c
23
#define CHANSTATUS(n, t) (CTRL_STRIDE_OFF(t, 2) + 0x4 * (n) + 0x4)
drivers/irqchip/irq-imx-irqsteer.c
24
#define CHAN_MINTDIS(t) (CTRL_STRIDE_OFF(t, 3) + 0x4)
drivers/irqchip/irq-imx-irqsteer.c
25
#define CHAN_MASTRSTAT(t) (CTRL_STRIDE_OFF(t, 3) + 0x8)
drivers/isdn/hardware/mISDN/hfcmulti.c
2813
hfcmulti_dbusy_timer(struct timer_list *t)
drivers/isdn/hardware/mISDN/hfcpci.c
1231
hfcpci_dbusy_timer(struct timer_list *t)
drivers/isdn/hardware/mISDN/hfcpci.c
293
hfcpci_Timer(struct timer_list *t)
drivers/isdn/hardware/mISDN/hfcpci.c
295
struct hfc_pci *hc = timer_container_of(hc, t, hw.timer);
drivers/isdn/hardware/mISDN/mISDNipac.c
714
dbusy_timer_handler(struct timer_list *t)
drivers/isdn/hardware/mISDN/mISDNipac.c
716
struct isac_hw *isac = timer_container_of(isac, t, dch.timer);
drivers/isdn/hardware/mISDN/mISDNisar.c
1136
ftimer_handler(struct timer_list *t)
drivers/isdn/hardware/mISDN/mISDNisar.c
1138
struct isar_ch *ch = timer_container_of(ch, t, ftimer);
drivers/isdn/hardware/mISDN/mISDNisar.c
131
int t = maxdelay;
drivers/isdn/hardware/mISDN/mISDNisar.c
135
while (t && !(irq & ISAR_IRQSTA)) {
drivers/isdn/hardware/mISDN/mISDNisar.c
137
t--;
drivers/isdn/hardware/mISDN/mISDNisar.c
139
if (t) {
drivers/isdn/hardware/mISDN/mISDNisar.c
144
isar->name, isar->clsb, maxdelay - t);
drivers/isdn/hardware/mISDN/mISDNisar.c
145
return t;
drivers/isdn/hardware/mISDN/mISDNisar.c
40
int t = timeout;
drivers/isdn/hardware/mISDN/mISDNisar.c
43
while ((val & 1) && t) {
drivers/isdn/hardware/mISDN/mISDNisar.c
45
t--;
drivers/isdn/hardware/mISDN/mISDNisar.c
48
pr_debug("%s: HIA after %dus\n", isar->name, timeout - t);
drivers/isdn/hardware/mISDN/w6692.c
803
dbusy_timer_handler(struct timer_list *t)
drivers/isdn/hardware/mISDN/w6692.c
805
struct dchannel *dch = timer_container_of(dch, t, timer);
drivers/isdn/mISDN/dsp.h
262
extern void dsp_tone_timeout(struct timer_list *t);
drivers/isdn/mISDN/dsp_cmx.c
1305
int r, rr, t, tt, o_r, o_rr;
drivers/isdn/mISDN/dsp_cmx.c
1360
t = dsp->tx_R; /* tx-pointers */
drivers/isdn/mISDN/dsp_cmx.c
1380
if (!dsp->tx_mix && t != tt) {
drivers/isdn/mISDN/dsp_cmx.c
1383
sprintf(debugbuf, "TX sending (%04x-%04x)%p: ", t, tt, p);
drivers/isdn/mISDN/dsp_cmx.c
1385
while (r != rr && t != tt) {
drivers/isdn/mISDN/dsp_cmx.c
1389
p[t]);
drivers/isdn/mISDN/dsp_cmx.c
1391
*d++ = p[t]; /* write tx_buff */
drivers/isdn/mISDN/dsp_cmx.c
1392
t = (t + 1) & CMX_BUFF_MASK;
drivers/isdn/mISDN/dsp_cmx.c
1396
dsp->tx_R = t;
drivers/isdn/mISDN/dsp_cmx.c
1412
while (r != rr && t != tt) {
drivers/isdn/mISDN/dsp_cmx.c
1413
*d++ = p[t]; /* write tx_buff */
drivers/isdn/mISDN/dsp_cmx.c
1414
t = (t + 1) & CMX_BUFF_MASK;
drivers/isdn/mISDN/dsp_cmx.c
1429
while (r != rr && t != tt) {
drivers/isdn/mISDN/dsp_cmx.c
1430
*d++ = dsp_audio_mix_law[(p[t] << 8) | q[r]];
drivers/isdn/mISDN/dsp_cmx.c
1431
t = (t + 1) & CMX_BUFF_MASK;
drivers/isdn/mISDN/dsp_cmx.c
1439
dsp->tx_R = t;
drivers/isdn/mISDN/dsp_cmx.c
1465
while (o_r != o_rr && t != tt) {
drivers/isdn/mISDN/dsp_cmx.c
1466
*d++ = dsp_audio_mix_law[(p[t] << 8) | o_q[o_r]];
drivers/isdn/mISDN/dsp_cmx.c
1467
t = (t + 1) & CMX_BUFF_MASK;
drivers/isdn/mISDN/dsp_cmx.c
1480
while (r != rr && t != tt) {
drivers/isdn/mISDN/dsp_cmx.c
1481
sample = dsp_audio_law_to_s32[p[t]] +
drivers/isdn/mISDN/dsp_cmx.c
1490
t = (t + 1) & CMX_BUFF_MASK;
drivers/isdn/mISDN/dsp_cmx.c
1500
dsp->tx_R = t;
drivers/isdn/mISDN/dsp_cmx.c
1510
while (r != rr && t != tt) {
drivers/isdn/mISDN/dsp_cmx.c
1511
sample = dsp_audio_law_to_s32[p[t]] + *c++ -
drivers/isdn/mISDN/dsp_cmx.c
1520
t = (t + 1) & CMX_BUFF_MASK;
drivers/isdn/mISDN/dsp_cmx.c
1538
while (r != rr && t != tt) {
drivers/isdn/mISDN/dsp_cmx.c
1539
sample = dsp_audio_law_to_s32[p[t]] + *c++;
drivers/isdn/mISDN/dsp_cmx.c
1546
t = (t + 1) & CMX_BUFF_MASK;
drivers/isdn/mISDN/dsp_cmx.c
1560
dsp->tx_R = t;
drivers/isdn/mISDN/dsp_tones.c
460
dsp_tone_timeout(struct timer_list *t)
drivers/isdn/mISDN/dsp_tones.c
462
struct dsp *dsp = timer_container_of(dsp, t, tone.tl);
drivers/isdn/mISDN/fsm.c
96
FsmExpireTimer(struct timer_list *t)
drivers/isdn/mISDN/fsm.c
98
struct FsmTimer *ft = timer_container_of(ft, t, tl);
drivers/isdn/mISDN/l1oip_core.c
823
l1oip_keepalive(struct timer_list *t)
drivers/isdn/mISDN/l1oip_core.c
825
struct l1oip *hc = timer_container_of(hc, t, keep_tl);
drivers/isdn/mISDN/l1oip_core.c
831
l1oip_timeout(struct timer_list *t)
drivers/isdn/mISDN/l1oip_core.c
833
struct l1oip *hc = timer_container_of(hc, t,
drivers/isdn/mISDN/timerdev.c
159
dev_expire_timer(struct timer_list *t)
drivers/isdn/mISDN/timerdev.c
161
struct mISDNtimer *timer = timer_container_of(timer, t, tl);
drivers/leds/flash/leds-as3645a.c
25
#define AS_TIMER_US_TO_CODE(t) (((t) / 1000 - 100) / 50)
drivers/leds/flash/leds-rt8515.c
166
static void rt8515_powerdown_timer(struct timer_list *t)
drivers/leds/flash/leds-rt8515.c
168
struct rt8515 *rt = timer_container_of(rt, t, powerdown_timer);
drivers/leds/flash/leds-sgm3140.c
136
static void sgm3140_powerdown_timer(struct timer_list *t)
drivers/leds/flash/leds-sgm3140.c
138
struct sgm3140 *priv = timer_container_of(priv, t, powerdown_timer);
drivers/leds/led-core.c
65
static void led_timer_function(struct timer_list *t)
drivers/leds/led-core.c
67
struct led_classdev *led_cdev = timer_container_of(led_cdev, t,
drivers/leds/leds-el15203000.c
159
int val, t;
drivers/leds/leds-el15203000.c
168
t = inv ? ~val & GENMASK(4, 0) : val;
drivers/leds/leds-el15203000.c
171
pattern[i].brightness != t)
drivers/leds/leds-lm3533.c
182
static u8 time_to_val(unsigned *t, unsigned t_min, unsigned t_step,
drivers/leds/leds-lm3533.c
187
val = (*t + t_step / 2 - t_min) / t_step + v_min;
drivers/leds/leds-lm3533.c
189
*t = t_step * (val - v_min) + t_min;
drivers/leds/leds-lm3533.c
209
unsigned t;
drivers/leds/leds-lm3533.c
212
t = *delay * 1000;
drivers/leds/leds-lm3533.c
214
if (t >= (LM3533_LED_DELAY2_TMAX + LM3533_LED_DELAY3_TMIN) / 2) {
drivers/leds/leds-lm3533.c
215
t = clamp(t, LM3533_LED_DELAY3_TMIN, LM3533_LED_DELAY3_TMAX);
drivers/leds/leds-lm3533.c
216
val = time_to_val(&t, LM3533_LED_DELAY3_TMIN,
drivers/leds/leds-lm3533.c
220
} else if (t >= (LM3533_LED_DELAY1_TMAX + LM3533_LED_DELAY2_TMIN) / 2) {
drivers/leds/leds-lm3533.c
221
t = clamp(t, LM3533_LED_DELAY2_TMIN, LM3533_LED_DELAY2_TMAX);
drivers/leds/leds-lm3533.c
222
val = time_to_val(&t, LM3533_LED_DELAY2_TMIN,
drivers/leds/leds-lm3533.c
227
t = clamp(t, LM3533_LED_DELAY1_TMIN, LM3533_LED_DELAY1_TMAX);
drivers/leds/leds-lm3533.c
228
val = time_to_val(&t, LM3533_LED_DELAY1_TMIN,
drivers/leds/leds-lm3533.c
234
*delay = (t + 500) / 1000;
drivers/leds/leds-lm3533.c
246
unsigned t;
drivers/leds/leds-lm3533.c
251
t = (unsigned)*delay;
drivers/leds/leds-lm3533.c
255
t = min(t, LM3533_LED_DELAY2_TMAX / 1000);
drivers/leds/leds-lm3533.c
257
val = lm3533_led_get_hw_delay(&t);
drivers/leds/leds-lm3533.c
260
*delay, t, val);
drivers/leds/leds-lm3533.c
266
*delay = t;
drivers/leds/leds-lm3533.c
271
static int lm3533_led_delay_on_set(struct lm3533_led *led, unsigned long *t)
drivers/leds/leds-lm3533.c
273
return lm3533_led_delay_set(led, LM3533_REG_PATTERN_HIGH_TIME_BASE, t);
drivers/leds/leds-lm3533.c
276
static int lm3533_led_delay_off_set(struct lm3533_led *led, unsigned long *t)
drivers/leds/leds-lm3533.c
278
return lm3533_led_delay_set(led, LM3533_REG_PATTERN_LOW_TIME_BASE, t);
drivers/leds/leds-sc27xx-bltc.c
136
u32 v, offset, t = *delta_t;
drivers/leds/leds-sc27xx-bltc.c
138
v = t + SC27XX_LEDS_STEP / 2;
drivers/leds/leds-tca6507.c
215
int t = time_codes[c1];
drivers/leds/leds-tca6507.c
216
if (t*2 < tmin)
drivers/leds/leds-tca6507.c
218
if (t > tmax)
drivers/leds/leds-tca6507.c
221
int tt = t + time_codes[c2];
drivers/leds/trigger/ledtrig-activity.c
33
static void led_activity_function(struct timer_list *t)
drivers/leds/trigger/ledtrig-activity.c
36
t, timer);
drivers/leds/trigger/ledtrig-heartbeat.c
33
static void led_heartbeat_function(struct timer_list *t)
drivers/leds/trigger/ledtrig-heartbeat.c
36
timer_container_of(heartbeat_data, t, timer);
drivers/leds/trigger/ledtrig-pattern.c
151
static void pattern_trig_timer_function(struct timer_list *t)
drivers/leds/trigger/ledtrig-pattern.c
153
struct pattern_trig_data *data = timer_container_of(data, t, timer);
drivers/leds/trigger/ledtrig-pattern.c
158
static enum hrtimer_restart pattern_trig_hrtimer_function(struct hrtimer *t)
drivers/leds/trigger/ledtrig-pattern.c
161
container_of(t, struct pattern_trig_data, hrtimer);
drivers/leds/trigger/ledtrig-transient.c
32
static void transient_timer_function(struct timer_list *t)
drivers/leds/trigger/ledtrig-transient.c
35
timer_container_of(transient_data, t, timer);
drivers/macintosh/smu.c
105
static void smu_i2c_retry(struct timer_list *t);
drivers/macintosh/windfarm_pm112.c
273
int i, t, target = 0;
drivers/macintosh/windfarm_pm112.c
309
t = wf_cpu_pid_run(sp, power, temp);
drivers/macintosh/windfarm_pm112.c
313
target = t;
drivers/mailbox/bcm-pdc-mailbox.c
76
#define NTXDACTIVE(h, t, max_mask) TXD((t) - (h), (max_mask))
drivers/mailbox/bcm-pdc-mailbox.c
77
#define NRXDACTIVE(h, t, max_mask) RXD((t) - (h), (max_mask))
drivers/mailbox/bcm-pdc-mailbox.c
961
static void pdc_work_cb(struct work_struct *t)
drivers/mailbox/bcm-pdc-mailbox.c
963
struct pdc_state *pdcs = from_work(pdcs, t, rx_work);
drivers/mailbox/imx-mailbox.c
506
static void imx_mu_txdb_work(struct work_struct *t)
drivers/mailbox/imx-mailbox.c
508
struct imx_mu_con_priv *cp = from_work(cp, t, txdb_work);
drivers/mailbox/mailbox-altera.c
131
static void altera_mbox_poll_rx(struct timer_list *t)
drivers/mailbox/mailbox-altera.c
133
struct altera_mbox *mbox = timer_container_of(mbox, t, rxpoll_timer);
drivers/mailbox/mailbox.c
247
int t;
drivers/mailbox/mailbox.c
252
t = add_to_rbuf(chan, mssg);
drivers/mailbox/mailbox.c
253
if (t < 0) {
drivers/mailbox/mailbox.c
255
return t;
drivers/mailbox/mailbox.c
271
t = -ETIME;
drivers/mailbox/mailbox.c
272
tx_tick(chan, t);
drivers/mailbox/mailbox.c
276
return t;
drivers/mailbox/mtk-cmdq-mailbox.c
25
#define CMDQ_NUM_CMD(t) (t->cmd_buf_size / CMDQ_INST_SIZE)
drivers/md/bcache/bset.c
1001
r = bset_bkey_last(t->data);
drivers/md/bcache/bset.c
1003
r = cacheline_to_bkey(t, inorder, f->m);
drivers/md/bcache/bset.c
1006
f = &t->tree[inorder_prev(j, t->size)];
drivers/md/bcache/bset.c
1007
l = cacheline_to_bkey(t, inorder, f->m);
drivers/md/bcache/bset.c
1009
l = t->data->start;
drivers/md/bcache/bset.c
1015
struct bkey *__bch_bset_search(struct btree_keys *b, struct bset_tree *t,
drivers/md/bcache/bset.c
1035
if (unlikely(!t->size)) {
drivers/md/bcache/bset.c
1036
i.l = t->data->start;
drivers/md/bcache/bset.c
1037
i.r = bset_bkey_last(t->data);
drivers/md/bcache/bset.c
1038
} else if (bset_written(b, t)) {
drivers/md/bcache/bset.c
1046
if (unlikely(bkey_cmp(search, &t->end) >= 0))
drivers/md/bcache/bset.c
1047
return bset_bkey_last(t->data);
drivers/md/bcache/bset.c
1049
if (unlikely(bkey_cmp(search, t->data->start) < 0))
drivers/md/bcache/bset.c
1050
return t->data->start;
drivers/md/bcache/bset.c
1052
i = bset_search_tree(t, search);
drivers/md/bcache/bset.c
1055
t->size < bkey_to_cacheline(t, bset_bkey_last(t->data)));
drivers/md/bcache/bset.c
1057
i = bset_search_write_set(t, search);
drivers/md/bcache/bset.c
1061
BUG_ON(bset_written(b, t) &&
drivers/md/bcache/bset.c
1062
i.l != t->data->start &&
drivers/md/bcache/bset.c
1063
bkey_cmp(tree_to_prev_bkey(t,
drivers/md/bcache/bset.c
1064
inorder_to_tree(bkey_to_cacheline(t, i.l), t)),
drivers/md/bcache/bset.c
1067
BUG_ON(i.r != bset_bkey_last(t->data) &&
drivers/md/bcache/bset.c
1372
struct bset_tree *t = &b->set[i];
drivers/md/bcache/bset.c
1373
size_t bytes = t->data->keys * sizeof(uint64_t);
drivers/md/bcache/bset.c
1376
if (bset_written(b, t)) {
drivers/md/bcache/bset.c
1380
stats->floats += t->size - 1;
drivers/md/bcache/bset.c
1382
for (j = 1; j < t->size; j++)
drivers/md/bcache/bset.c
1383
if (t->tree[j].exponent == 127)
drivers/md/bcache/bset.c
294
struct bset_tree *t = b->set;
drivers/md/bcache/bset.c
297
kfree(t->prev);
drivers/md/bcache/bset.c
299
free_pages((unsigned long) t->prev,
drivers/md/bcache/bset.c
303
kfree(t->tree);
drivers/md/bcache/bset.c
305
free_pages((unsigned long) t->tree,
drivers/md/bcache/bset.c
308
free_pages((unsigned long) t->data, b->page_order);
drivers/md/bcache/bset.c
310
t->prev = NULL;
drivers/md/bcache/bset.c
311
t->tree = NULL;
drivers/md/bcache/bset.c
312
t->data = NULL;
drivers/md/bcache/bset.c
319
struct bset_tree *t = b->set;
drivers/md/bcache/bset.c
321
BUG_ON(t->data);
drivers/md/bcache/bset.c
325
t->data = (void *) __get_free_pages(__GFP_COMP|gfp, b->page_order);
drivers/md/bcache/bset.c
326
if (!t->data)
drivers/md/bcache/bset.c
329
t->tree = bset_tree_bytes(b) < PAGE_SIZE
drivers/md/bcache/bset.c
332
if (!t->tree)
drivers/md/bcache/bset.c
335
t->prev = bset_prev_bytes(b) < PAGE_SIZE
drivers/md/bcache/bset.c
338
if (!t->prev)
drivers/md/bcache/bset.c
437
static unsigned int to_inorder(unsigned int j, struct bset_tree *t)
drivers/md/bcache/bset.c
439
return __to_inorder(j, t->size, t->extra);
drivers/md/bcache/bset.c
463
static unsigned int inorder_to_tree(unsigned int j, struct bset_tree *t)
drivers/md/bcache/bset.c
465
return __inorder_to_tree(j, t->size, t->extra);
drivers/md/bcache/bset.c
525
static struct bkey *cacheline_to_bkey(struct bset_tree *t,
drivers/md/bcache/bset.c
529
return ((void *) t->data) + cacheline * BSET_CACHELINE + offset * 8;
drivers/md/bcache/bset.c
532
static unsigned int bkey_to_cacheline(struct bset_tree *t, struct bkey *k)
drivers/md/bcache/bset.c
534
return ((void *) k - (void *) t->data) / BSET_CACHELINE;
drivers/md/bcache/bset.c
537
static unsigned int bkey_to_cacheline_offset(struct bset_tree *t,
drivers/md/bcache/bset.c
541
return (u64 *) k - (u64 *) cacheline_to_bkey(t, cacheline, 0);
drivers/md/bcache/bset.c
544
static struct bkey *tree_to_bkey(struct bset_tree *t, unsigned int j)
drivers/md/bcache/bset.c
546
return cacheline_to_bkey(t, to_inorder(j, t), t->tree[j].m);
drivers/md/bcache/bset.c
549
static struct bkey *tree_to_prev_bkey(struct bset_tree *t, unsigned int j)
drivers/md/bcache/bset.c
551
return (void *) (((uint64_t *) tree_to_bkey(t, j)) - t->prev[j]);
drivers/md/bcache/bset.c
558
static struct bkey *table_to_bkey(struct bset_tree *t, unsigned int cacheline)
drivers/md/bcache/bset.c
560
return cacheline_to_bkey(t, cacheline, t->prev[cacheline]);
drivers/md/bcache/bset.c
592
static void make_bfloat(struct bset_tree *t, unsigned int j)
drivers/md/bcache/bset.c
594
struct bkey_float *f = &t->tree[j];
drivers/md/bcache/bset.c
595
struct bkey *m = tree_to_bkey(t, j);
drivers/md/bcache/bset.c
596
struct bkey *p = tree_to_prev_bkey(t, j);
drivers/md/bcache/bset.c
599
? t->data->start
drivers/md/bcache/bset.c
600
: tree_to_prev_bkey(t, j >> ffs(j));
drivers/md/bcache/bset.c
603
? bset_bkey_idx(t->data, t->data->keys - bkey_u64s(&t->end))
drivers/md/bcache/bset.c
604
: tree_to_bkey(t, j >> (ffz(j) + 1));
drivers/md/bcache/bset.c
637
static void bset_alloc_tree(struct btree_keys *b, struct bset_tree *t)
drivers/md/bcache/bset.c
639
if (t != b->set) {
drivers/md/bcache/bset.c
640
unsigned int j = roundup(t[-1].size,
drivers/md/bcache/bset.c
643
t->tree = t[-1].tree + j;
drivers/md/bcache/bset.c
644
t->prev = t[-1].prev + j;
drivers/md/bcache/bset.c
647
while (t < b->set + MAX_BSETS)
drivers/md/bcache/bset.c
648
t++->size = 0;
drivers/md/bcache/bset.c
653
struct bset_tree *t = bset_tree_last(b);
drivers/md/bcache/bset.c
658
bset_alloc_tree(b, t);
drivers/md/bcache/bset.c
660
if (t->tree != b->set->tree + btree_keys_cachelines(b)) {
drivers/md/bcache/bset.c
661
t->prev[0] = bkey_to_cacheline_offset(t, 0, t->data->start);
drivers/md/bcache/bset.c
662
t->size = 1;
drivers/md/bcache/bset.c
692
struct bset_tree *t = bset_tree_last(b);
drivers/md/bcache/bset.c
693
struct bkey *prev = NULL, *k = t->data->start;
drivers/md/bcache/bset.c
698
bset_alloc_tree(b, t);
drivers/md/bcache/bset.c
700
t->size = min_t(unsigned int,
drivers/md/bcache/bset.c
701
bkey_to_cacheline(t, bset_bkey_last(t->data)),
drivers/md/bcache/bset.c
702
b->set->tree + btree_keys_cachelines(b) - t->tree);
drivers/md/bcache/bset.c
704
if (t->size < 2) {
drivers/md/bcache/bset.c
705
t->size = 0;
drivers/md/bcache/bset.c
709
t->extra = (t->size - rounddown_pow_of_two(t->size - 1)) << 1;
drivers/md/bcache/bset.c
712
for (j = inorder_next(0, t->size);
drivers/md/bcache/bset.c
714
j = inorder_next(j, t->size)) {
drivers/md/bcache/bset.c
715
while (bkey_to_cacheline(t, k) < cacheline) {
drivers/md/bcache/bset.c
720
t->prev[j] = bkey_u64s(prev);
drivers/md/bcache/bset.c
721
t->tree[j].m = bkey_to_cacheline_offset(t, cacheline++, k);
drivers/md/bcache/bset.c
724
while (bkey_next(k) != bset_bkey_last(t->data))
drivers/md/bcache/bset.c
727
t->end = *k;
drivers/md/bcache/bset.c
730
for (j = inorder_next(0, t->size);
drivers/md/bcache/bset.c
732
j = inorder_next(j, t->size))
drivers/md/bcache/bset.c
733
make_bfloat(t, j);
drivers/md/bcache/bset.c
740
struct bset_tree *t;
drivers/md/bcache/bset.c
743
for (t = b->set; t <= bset_tree_last(b); t++)
drivers/md/bcache/bset.c
744
if (k < bset_bkey_last(t->data))
drivers/md/bcache/bset.c
749
if (!t->size || !bset_written(b, t))
drivers/md/bcache/bset.c
752
inorder = bkey_to_cacheline(t, k);
drivers/md/bcache/bset.c
754
if (k == t->data->start)
drivers/md/bcache/bset.c
757
if (bkey_next(k) == bset_bkey_last(t->data)) {
drivers/md/bcache/bset.c
758
t->end = *k;
drivers/md/bcache/bset.c
762
j = inorder_to_tree(inorder, t);
drivers/md/bcache/bset.c
765
j < t->size &&
drivers/md/bcache/bset.c
766
k == tree_to_bkey(t, j))
drivers/md/bcache/bset.c
768
make_bfloat(t, j);
drivers/md/bcache/bset.c
770
} while (j < t->size);
drivers/md/bcache/bset.c
772
j = inorder_to_tree(inorder + 1, t);
drivers/md/bcache/bset.c
775
j < t->size &&
drivers/md/bcache/bset.c
776
k == tree_to_prev_bkey(t, j))
drivers/md/bcache/bset.c
778
make_bfloat(t, j);
drivers/md/bcache/bset.c
780
} while (j < t->size);
drivers/md/bcache/bset.c
784
struct bset_tree *t,
drivers/md/bcache/bset.c
788
unsigned int j = bkey_to_cacheline(t, k);
drivers/md/bcache/bset.c
791
if (!t->size)
drivers/md/bcache/bset.c
799
while (j < t->size &&
drivers/md/bcache/bset.c
800
table_to_bkey(t, j) <= k)
drivers/md/bcache/bset.c
807
for (; j < t->size; j++) {
drivers/md/bcache/bset.c
808
t->prev[j] += shift;
drivers/md/bcache/bset.c
810
if (t->prev[j] > 7) {
drivers/md/bcache/bset.c
811
k = table_to_bkey(t, j - 1);
drivers/md/bcache/bset.c
813
while (k < cacheline_to_bkey(t, j, 0))
drivers/md/bcache/bset.c
816
t->prev[j] = bkey_to_cacheline_offset(t, j, k);
drivers/md/bcache/bset.c
820
if (t->size == b->set->tree + btree_keys_cachelines(b) - t->tree)
drivers/md/bcache/bset.c
825
for (k = table_to_bkey(t, t->size - 1);
drivers/md/bcache/bset.c
826
k != bset_bkey_last(t->data);
drivers/md/bcache/bset.c
828
if (t->size == bkey_to_cacheline(t, k)) {
drivers/md/bcache/bset.c
829
t->prev[t->size] =
drivers/md/bcache/bset.c
830
bkey_to_cacheline_offset(t, t->size, k);
drivers/md/bcache/bset.c
831
t->size++;
drivers/md/bcache/bset.c
860
struct bset_tree *t = bset_tree_last(b);
drivers/md/bcache/bset.c
863
BUG_ON(bset_byte_offset(b, t->data) +
drivers/md/bcache/bset.c
864
__set_bytes(t->data, t->data->keys + bkey_u64s(insert)) >
drivers/md/bcache/bset.c
869
(void *) bset_bkey_last(t->data) - (void *) where);
drivers/md/bcache/bset.c
871
t->data->keys += bkey_u64s(insert);
drivers/md/bcache/bset.c
873
bch_bset_fix_lookup_table(b, t, where);
drivers/md/bcache/bset.c
939
static struct bset_search_iter bset_search_write_set(struct bset_tree *t,
drivers/md/bcache/bset.c
942
unsigned int li = 0, ri = t->size;
drivers/md/bcache/bset.c
947
if (bkey_cmp(table_to_bkey(t, m), search) > 0)
drivers/md/bcache/bset.c
954
table_to_bkey(t, li),
drivers/md/bcache/bset.c
955
ri < t->size ? table_to_bkey(t, ri) : bset_bkey_last(t->data)
drivers/md/bcache/bset.c
959
static struct bset_search_iter bset_search_tree(struct bset_tree *t,
drivers/md/bcache/bset.c
969
if (p < t->size)
drivers/md/bcache/bset.c
970
prefetch(&t->tree[p]);
drivers/md/bcache/bset.c
973
f = &t->tree[j];
drivers/md/bcache/bset.c
981
if (bkey_cmp(tree_to_bkey(t, j), search) > 0)
drivers/md/bcache/bset.c
986
} while (n < t->size);
drivers/md/bcache/bset.c
988
inorder = to_inorder(j, t);
drivers/md/bcache/bset.c
995
l = cacheline_to_bkey(t, inorder, f->m);
drivers/md/bcache/bset.c
997
if (++inorder != t->size) {
drivers/md/bcache/bset.c
998
f = &t->tree[inorder_next(j, t->size)];
drivers/md/bcache/bset.c
999
r = cacheline_to_bkey(t, inorder, f->m);
drivers/md/bcache/bset.h
239
static inline bool bset_written(struct btree_keys *b, struct bset_tree *t)
drivers/md/bcache/bset.h
241
return t <= b->set + b->nsets - b->last_set_unwritten;
drivers/md/bcache/bset.h
271
struct bset_tree *t = bset_tree_last(b);
drivers/md/bcache/bset.h
274
(bset_byte_offset(b, t->data) + set_bytes(t->data)));
drivers/md/bcache/bset.h
280
(bset_byte_offset(b, t->data) + set_bytes(t->data))) /
drivers/md/bcache/bset.h
351
struct bkey *__bch_bset_search(struct btree_keys *b, struct bset_tree *t,
drivers/md/bcache/bset.h
358
struct bset_tree *t,
drivers/md/bcache/bset.h
361
return search ? __bch_bset_search(b, t, search) : t->data->start;
drivers/md/bcache/btree.c
1311
struct bset_tree *t;
drivers/md/bcache/btree.c
1329
for (t = b->keys.set; t <= &b->keys.set[b->keys.nsets]; t++)
drivers/md/bcache/btree.c
1330
btree_bug_on(t->size &&
drivers/md/bcache/btree.c
1331
bset_written(&b->keys, t) &&
drivers/md/bcache/btree.c
1332
bkey_cmp(&b->key, &t->end) < 0,
drivers/md/bcache/btree.c
669
struct btree *b, *t;
drivers/md/bcache/btree.c
700
list_for_each_entry_safe_reverse(b, t, &c->btree_cache_freeable, list) {
drivers/md/bcache/btree.c
713
list_for_each_entry_safe_reverse(b, t, &c->btree_cache, list) {
drivers/md/bcache/io.c
148
unsigned int t = local_clock_us();
drivers/md/bcache/io.c
149
int us = t - b->submit_time_us;
drivers/md/bcache/io.c
155
c->congested_last_us = t;
drivers/md/bcache/journal.c
402
struct btree *b, *t, *btree_nodes[BTREE_FLUSH_NR];
drivers/md/bcache/journal.c
439
list_for_each_entry_safe_reverse(b, t, &c->btree_cache, list) {
drivers/md/bcache/request.c
350
static void add_sequential(struct task_struct *t)
drivers/md/bcache/request.c
352
ewma_add(t->sequential_io_avg,
drivers/md/bcache/request.c
353
t->sequential_io, 8, 0);
drivers/md/bcache/request.c
355
t->sequential_io = 0;
drivers/md/bcache/stats.c
150
static void scale_accounting(struct timer_list *t)
drivers/md/bcache/stats.c
152
struct cache_accounting *acc = timer_container_of(acc, t, timer);
drivers/md/bcache/stats.c
155
unsigned int t = atomic_xchg(&acc->collector.name, 0); \
drivers/md/bcache/stats.c
156
t <<= 16; \
drivers/md/bcache/stats.c
157
acc->five_minute.name += t; \
drivers/md/bcache/stats.c
158
acc->hour.name += t; \
drivers/md/bcache/stats.c
159
acc->day.name += t; \
drivers/md/bcache/stats.c
160
acc->total.name += t; \
drivers/md/bcache/super.c
1201
struct cached_dev *exist_dc, *t;
drivers/md/bcache/super.c
1226
list_for_each_entry_safe(exist_dc, t, &c->cached_devs, list) {
drivers/md/bcache/super.c
1981
struct cached_dev *dc, *t;
drivers/md/bcache/super.c
2132
list_for_each_entry_safe(dc, t, &uncached_devices, list)
drivers/md/bcache/super.c
2450
struct cached_dev *dc, *t;
drivers/md/bcache/super.c
2453
list_for_each_entry_safe(dc, t, &c->cached_devs, list)
drivers/md/bcache/super.c
2456
list_for_each_entry_safe(dc, t, &uncached_devices, list)
drivers/md/bcache/util.c
111
t = q & ~(~0 << 10);
drivers/md/bcache/util.c
119
return sprintf(buf, "-%llu.%i%c", q, t * 10 / 1024, units[u]);
drivers/md/bcache/util.c
121
return sprintf(buf, "%llu.%i%c", q, t * 10 / 1024, units[u]);
drivers/md/bcache/util.c
95
int u = 0, t;
drivers/md/bcache/util.h
284
#define ANYSINT_MAX(t) \
drivers/md/bcache/util.h
285
((((t) 1 << (sizeof(t) * 8 - 2)) - (t) 1) * (t) 2 + (t) 1)
drivers/md/dm-cache-policy.c
139
struct dm_cache_policy_type *t = p->private;
drivers/md/dm-cache-policy.c
142
put_policy(t);
drivers/md/dm-cache-policy.c
148
struct dm_cache_policy_type *t = p->private;
drivers/md/dm-cache-policy.c
151
if (t->real)
drivers/md/dm-cache-policy.c
152
return t->real->name;
drivers/md/dm-cache-policy.c
154
return t->name;
drivers/md/dm-cache-policy.c
160
struct dm_cache_policy_type *t = p->private;
drivers/md/dm-cache-policy.c
162
return t->version;
drivers/md/dm-cache-policy.c
168
struct dm_cache_policy_type *t = p->private;
drivers/md/dm-cache-policy.c
170
return t->hint_size;
drivers/md/dm-cache-policy.c
23
struct dm_cache_policy_type *t;
drivers/md/dm-cache-policy.c
25
list_for_each_entry(t, &register_list, list)
drivers/md/dm-cache-policy.c
26
if (!strcmp(t->name, name))
drivers/md/dm-cache-policy.c
27
return t;
drivers/md/dm-cache-policy.c
34
struct dm_cache_policy_type *t = __find_policy(name);
drivers/md/dm-cache-policy.c
36
if (t && !try_module_get(t->owner)) {
drivers/md/dm-cache-policy.c
38
t = ERR_PTR(-EINVAL);
drivers/md/dm-cache-policy.c
41
return t;
drivers/md/dm-cache-policy.c
46
struct dm_cache_policy_type *t;
drivers/md/dm-cache-policy.c
49
t = __get_policy_once(name);
drivers/md/dm-cache-policy.c
52
return t;
drivers/md/dm-cache-policy.c
57
struct dm_cache_policy_type *t;
drivers/md/dm-cache-policy.c
59
t = get_policy_once(name);
drivers/md/dm-cache-policy.c
60
if (IS_ERR(t))
drivers/md/dm-cache-policy.c
63
if (t)
drivers/md/dm-cache-policy.c
64
return t;
drivers/md/dm-cache-policy.c
68
t = get_policy_once(name);
drivers/md/dm-cache-policy.c
69
if (IS_ERR(t))
drivers/md/dm-cache-policy.c
72
return t;
drivers/md/dm-cache-policy.c
75
static void put_policy(struct dm_cache_policy_type *t)
drivers/md/dm-cache-policy.c
77
module_put(t->owner);
drivers/md/dm-core.h
230
static inline struct dm_target *dm_table_get_target(struct dm_table *t,
drivers/md/dm-core.h
233
BUG_ON(index >= t->num_targets);
drivers/md/dm-core.h
234
return t->targets + index;
drivers/md/dm-delay.c
57
static void handle_delayed_timer(struct timer_list *t)
drivers/md/dm-delay.c
59
struct delay_c *dc = timer_container_of(dc, t, delay_timer);
drivers/md/dm-integrity.c
1551
static void autocommit_fn(struct timer_list *t)
drivers/md/dm-integrity.c
1553
struct dm_integrity_c *ic = timer_container_of(ic, t,
drivers/md/dm-integrity.c
3127
__u8 *t;
drivers/md/dm-integrity.c
3228
t = recalc_tags;
drivers/md/dm-integrity.c
3233
integrity_sector_checksum(ic, &ahash_req, logical_sector + i, ptr_page, ptr_offset, t);
drivers/md/dm-integrity.c
3234
t += ic->tag_size;
drivers/md/dm-integrity.c
3239
r = dm_integrity_rw_tag(ic, recalc_tags, &metadata_block, &metadata_offset, t - recalc_tags, TAG_WRITE);
drivers/md/dm-integrity.c
3290
__u8 *t;
drivers/md/dm-integrity.c
3353
t = recalc_tags;
drivers/md/dm-integrity.c
3358
memset(t, 0, ic->tuple_size);
drivers/md/dm-integrity.c
3359
integrity_sector_checksum(ic, &ahash_req, range.logical_sector + i, ptr_page, ptr_offset, t);
drivers/md/dm-integrity.c
3360
t += ic->tuple_size;
drivers/md/dm-integrity.c
3374
ret = bio_integrity_add_page(bio, virt_to_page(recalc_tags), t - recalc_tags, offset_in_page(recalc_tags));
drivers/md/dm-integrity.c
3375
if (unlikely(ret != t - recalc_tags)) {
drivers/md/dm-ioctl.c
1011
t = __hash_remove(hc);
drivers/md/dm-ioctl.c
1014
if (t) {
drivers/md/dm-ioctl.c
1016
dm_table_destroy(t);
drivers/md/dm-ioctl.c
1518
struct dm_table *t, *old_map = NULL;
drivers/md/dm-ioctl.c
1526
r = dm_table_create(&t, get_mode(param), param->target_count, md);
drivers/md/dm-ioctl.c
1532
r = populate_table(t, param, param_size);
drivers/md/dm-ioctl.c
1536
dm_ima_measure_on_table_load(t, STATUSTYPE_IMA);
drivers/md/dm-ioctl.c
1540
(immutable_target_type != dm_table_get_immutable_target_type(t)) &&
drivers/md/dm-ioctl.c
1541
!dm_table_get_wildcard_target(t)) {
drivers/md/dm-ioctl.c
1550
r = dm_setup_md_queue(md, t);
drivers/md/dm-ioctl.c
1555
} else if (!is_valid_type(dm_get_md_type(md), dm_table_get_type(t))) {
drivers/md/dm-ioctl.c
1557
dm_get_md_type(md), dm_table_get_type(t));
drivers/md/dm-ioctl.c
1576
hc->new_map = t;
drivers/md/dm-ioctl.c
1594
dm_table_destroy(t);
drivers/md/dm-ioctl.c
2267
struct dm_table *t, *old_map;
drivers/md/dm-ioctl.c
2292
r = dm_table_create(&t, get_mode(dmi), dmi->target_count, md);
drivers/md/dm-ioctl.c
2298
r = dm_table_add_target(t, spec_array[i]->target_type,
drivers/md/dm-ioctl.c
2309
r = dm_table_complete(t);
drivers/md/dm-ioctl.c
2314
r = dm_setup_md_queue(md, t);
drivers/md/dm-ioctl.c
2322
old_map = dm_swap_table(md, t);
drivers/md/dm-ioctl.c
2339
dm_table_destroy(t);
drivers/md/dm-ioctl.c
342
struct dm_table *t;
drivers/md/dm-ioctl.c
361
t = __hash_remove(hc);
drivers/md/dm-ioctl.c
365
if (t) {
drivers/md/dm-ioctl.c
367
dm_table_destroy(t);
drivers/md/dm-ioctl.c
982
struct dm_table *t;
drivers/md/dm-kcopyd.c
123
static void io_job_start(struct dm_kcopyd_throttle *t)
drivers/md/dm-kcopyd.c
128
if (unlikely(!t))
drivers/md/dm-kcopyd.c
134
throttle = READ_ONCE(t->throttle);
drivers/md/dm-kcopyd.c
140
difference = now - t->last_jiffies;
drivers/md/dm-kcopyd.c
141
t->last_jiffies = now;
drivers/md/dm-kcopyd.c
142
if (t->num_io_jobs)
drivers/md/dm-kcopyd.c
143
t->io_period += difference;
drivers/md/dm-kcopyd.c
144
t->total_period += difference;
drivers/md/dm-kcopyd.c
149
if (unlikely(t->io_period > t->total_period))
drivers/md/dm-kcopyd.c
150
t->io_period = t->total_period;
drivers/md/dm-kcopyd.c
152
if (unlikely(t->total_period >= (1 << ACCOUNT_INTERVAL_SHIFT))) {
drivers/md/dm-kcopyd.c
153
int shift = fls(t->total_period >> ACCOUNT_INTERVAL_SHIFT);
drivers/md/dm-kcopyd.c
155
t->total_period >>= shift;
drivers/md/dm-kcopyd.c
156
t->io_period >>= shift;
drivers/md/dm-kcopyd.c
159
skew = t->io_period - throttle * t->total_period / 100;
drivers/md/dm-kcopyd.c
169
t->num_io_jobs++;
drivers/md/dm-kcopyd.c
174
static void io_job_finish(struct dm_kcopyd_throttle *t)
drivers/md/dm-kcopyd.c
178
if (unlikely(!t))
drivers/md/dm-kcopyd.c
183
t->num_io_jobs--;
drivers/md/dm-kcopyd.c
185
if (likely(READ_ONCE(t->throttle) >= 100))
drivers/md/dm-kcopyd.c
188
if (!t->num_io_jobs) {
drivers/md/dm-kcopyd.c
192
difference = now - t->last_jiffies;
drivers/md/dm-kcopyd.c
193
t->last_jiffies = now;
drivers/md/dm-kcopyd.c
195
t->io_period += difference;
drivers/md/dm-kcopyd.c
196
t->total_period += difference;
drivers/md/dm-kcopyd.c
201
if (unlikely(t->io_period > t->total_period))
drivers/md/dm-kcopyd.c
202
t->io_period = t->total_period;
drivers/md/dm-mpath.c
124
static void queue_if_no_path_timeout_work(struct timer_list *t);
drivers/md/dm-mpath.c
785
static void queue_if_no_path_timeout_work(struct timer_list *t)
drivers/md/dm-mpath.c
787
struct multipath *m = timer_container_of(m, t, nopath_timer);
drivers/md/dm-raid1.c
104
static void delayed_wake_fn(struct timer_list *t)
drivers/md/dm-raid1.c
106
struct mirror_set *ms = timer_container_of(ms, t, timer);
drivers/md/dm-rq.c
545
int dm_mq_init_request_queue(struct mapped_device *md, struct dm_table *t)
drivers/md/dm-rq.c
562
immutable_tgt = dm_table_get_immutable_target(t);
drivers/md/dm-rq.h
34
int dm_mq_init_request_queue(struct mapped_device *md, struct dm_table *t);
drivers/md/dm-table.c
1000
enum dm_queue_mode dm_table_get_type(struct dm_table *t)
drivers/md/dm-table.c
1002
return t->type;
drivers/md/dm-table.c
1005
struct target_type *dm_table_get_immutable_target_type(struct dm_table *t)
drivers/md/dm-table.c
1007
return t->immutable_target_type;
drivers/md/dm-table.c
1010
struct dm_target *dm_table_get_immutable_target(struct dm_table *t)
drivers/md/dm-table.c
1013
if (t->num_targets > 1 ||
drivers/md/dm-table.c
1014
!dm_target_is_immutable(t->targets[0].type))
drivers/md/dm-table.c
1017
return t->targets;
drivers/md/dm-table.c
1020
struct dm_target *dm_table_get_wildcard_target(struct dm_table *t)
drivers/md/dm-table.c
1022
for (unsigned int i = 0; i < t->num_targets; i++) {
drivers/md/dm-table.c
1023
struct dm_target *ti = dm_table_get_target(t, i);
drivers/md/dm-table.c
1032
bool dm_table_request_based(struct dm_table *t)
drivers/md/dm-table.c
1034
return __table_type_request_based(dm_table_get_type(t));
drivers/md/dm-table.c
1037
static int dm_table_alloc_md_mempools(struct dm_table *t, struct mapped_device *md)
drivers/md/dm-table.c
1039
enum dm_queue_mode type = dm_table_get_type(t);
drivers/md/dm-table.c
104
static int alloc_targets(struct dm_table *t, unsigned int num)
drivers/md/dm-table.c
1063
for (unsigned int i = 0; i < t->num_targets; i++) {
drivers/md/dm-table.c
1064
struct dm_target *ti = dm_table_get_target(t, i);
drivers/md/dm-table.c
1081
t->mempools = pools;
drivers/md/dm-table.c
1089
static int setup_indexes(struct dm_table *t)
drivers/md/dm-table.c
1096
for (i = t->depth - 2; i >= 0; i--) {
drivers/md/dm-table.c
1097
t->counts[i] = dm_div_up(t->counts[i + 1], CHILDREN_PER_NODE);
drivers/md/dm-table.c
1098
total += t->counts[i];
drivers/md/dm-table.c
1106
for (i = t->depth - 2; i >= 0; i--) {
drivers/md/dm-table.c
1107
t->index[i] = indexes;
drivers/md/dm-table.c
1108
indexes += (KEYS_PER_NODE * t->counts[i]);
drivers/md/dm-table.c
1109
setup_btree_index(i, t);
drivers/md/dm-table.c
1118
static int dm_table_build_index(struct dm_table *t)
drivers/md/dm-table.c
1124
leaf_nodes = dm_div_up(t->num_targets, KEYS_PER_NODE);
drivers/md/dm-table.c
1125
t->depth = 1 + int_log(leaf_nodes, CHILDREN_PER_NODE);
drivers/md/dm-table.c
1128
t->counts[t->depth - 1] = leaf_nodes;
drivers/md/dm-table.c
1129
t->index[t->depth - 1] = t->highs;
drivers/md/dm-table.c
1131
if (t->depth >= 2)
drivers/md/dm-table.c
1132
r = setup_indexes(t);
drivers/md/dm-table.c
1162
struct dm_table *t;
drivers/md/dm-table.c
1165
t = dm_get_live_table(md, &srcu_idx);
drivers/md/dm-table.c
1166
if (!t)
drivers/md/dm-table.c
1169
for (unsigned int i = 0; i < t->num_targets; i++) {
drivers/md/dm-table.c
1170
struct dm_target *ti = dm_table_get_target(t, i);
drivers/md/dm-table.c
121
t->num_allocated = num;
drivers/md/dm-table.c
122
t->highs = n_highs;
drivers/md/dm-table.c
123
t->targets = n_targets;
drivers/md/dm-table.c
1259
struct dm_table *t;
drivers/md/dm-table.c
1265
t = dm_get_live_table(md, &srcu_idx);
drivers/md/dm-table.c
1266
if (!t)
drivers/md/dm-table.c
1278
for (i = 0; i < t->num_targets; i++) {
drivers/md/dm-table.c
1279
ti = dm_table_get_target(t, i);
drivers/md/dm-table.c
131
struct dm_table *t;
drivers/md/dm-table.c
136
t = kzalloc_obj(*t);
drivers/md/dm-table.c
1372
static void dm_table_destroy_crypto_profile(struct dm_table *t)
drivers/md/dm-table.c
1374
dm_destroy_crypto_profile(t->crypto_profile);
drivers/md/dm-table.c
1375
t->crypto_profile = NULL;
drivers/md/dm-table.c
138
if (!t)
drivers/md/dm-table.c
1387
static int dm_table_construct_crypto_profile(struct dm_table *t)
drivers/md/dm-table.c
1397
dmcp->md = t->md;
drivers/md/dm-table.c
1407
for (i = 0; i < t->num_targets; i++) {
drivers/md/dm-table.c
1408
struct dm_target *ti = dm_table_get_target(t, i);
drivers/md/dm-table.c
141
INIT_LIST_HEAD(&t->devices);
drivers/md/dm-table.c
1428
if (t->md->queue &&
drivers/md/dm-table.c
1430
t->md->queue->crypto_profile)) {
drivers/md/dm-table.c
1457
t->crypto_profile = profile;
drivers/md/dm-table.c
1463
struct dm_table *t)
drivers/md/dm-table.c
1465
if (!t->crypto_profile)
drivers/md/dm-table.c
1470
blk_crypto_register(t->crypto_profile, q);
drivers/md/dm-table.c
1473
t->crypto_profile);
drivers/md/dm-table.c
1474
dm_destroy_crypto_profile(t->crypto_profile);
drivers/md/dm-table.c
1476
t->crypto_profile = NULL;
drivers/md/dm-table.c
1481
static int dm_table_construct_crypto_profile(struct dm_table *t)
drivers/md/dm-table.c
149
kfree(t);
drivers/md/dm-table.c
1490
static void dm_table_destroy_crypto_profile(struct dm_table *t)
drivers/md/dm-table.c
1495
struct dm_table *t)
drivers/md/dm-table.c
1505
int dm_table_complete(struct dm_table *t)
drivers/md/dm-table.c
1509
r = dm_table_determine_type(t);
drivers/md/dm-table.c
1515
r = dm_table_build_index(t);
drivers/md/dm-table.c
1521
r = dm_table_construct_crypto_profile(t);
drivers/md/dm-table.c
1527
r = dm_table_alloc_md_mempools(t, t->md);
drivers/md/dm-table.c
153
if (alloc_targets(t, num_targets)) {
drivers/md/dm-table.c
1535
void dm_table_event_callback(struct dm_table *t,
drivers/md/dm-table.c
1539
t->event_fn = fn;
drivers/md/dm-table.c
154
kfree(t);
drivers/md/dm-table.c
1540
t->event_context = context;
drivers/md/dm-table.c
1544
void dm_table_event(struct dm_table *t)
drivers/md/dm-table.c
1547
if (t->event_fn)
drivers/md/dm-table.c
1548
t->event_fn(t->event_context);
drivers/md/dm-table.c
1553
inline sector_t dm_table_get_size(struct dm_table *t)
drivers/md/dm-table.c
1555
return t->num_targets ? (t->highs[t->num_targets - 1] + 1) : 0;
drivers/md/dm-table.c
1565
struct dm_target *dm_table_find_target(struct dm_table *t, sector_t sector)
drivers/md/dm-table.c
1570
if (unlikely(sector >= dm_table_get_size(t)))
drivers/md/dm-table.c
1573
for (l = 0; l < t->depth; l++) {
drivers/md/dm-table.c
1575
node = get_node(t, l, n);
drivers/md/dm-table.c
158
t->type = DM_TYPE_NONE;
drivers/md/dm-table.c
1582
return &t->targets[(KEYS_PER_NODE * n) + k];
drivers/md/dm-table.c
159
t->mode = mode;
drivers/md/dm-table.c
160
t->md = md;
drivers/md/dm-table.c
1608
static bool dm_table_any_dev_attr(struct dm_table *t,
drivers/md/dm-table.c
161
t->flush_bypasses_map = true;
drivers/md/dm-table.c
1611
for (unsigned int i = 0; i < t->num_targets; i++) {
drivers/md/dm-table.c
1612
struct dm_target *ti = dm_table_get_target(t, i);
drivers/md/dm-table.c
162
*result = t;
drivers/md/dm-table.c
1638
bool dm_table_has_no_data_devices(struct dm_table *t)
drivers/md/dm-table.c
1640
for (unsigned int i = 0; i < t->num_targets; i++) {
drivers/md/dm-table.c
1641
struct dm_target *ti = dm_table_get_target(t, i);
drivers/md/dm-table.c
1655
bool dm_table_is_wildcard(struct dm_table *t)
drivers/md/dm-table.c
1657
for (unsigned int i = 0; i < t->num_targets; i++) {
drivers/md/dm-table.c
1658
struct dm_target *ti = dm_table_get_target(t, i);
drivers/md/dm-table.c
1688
static bool dm_table_supports_zoned(struct dm_table *t, bool zoned)
drivers/md/dm-table.c
1690
for (unsigned int i = 0; i < t->num_targets; i++) {
drivers/md/dm-table.c
1691
struct dm_target *ti = dm_table_get_target(t, i);
drivers/md/dm-table.c
1734
static int validate_hardware_zoned(struct dm_table *t, bool zoned,
drivers/md/dm-table.c
1740
if (!dm_table_supports_zoned(t, zoned)) {
drivers/md/dm-table.c
1742
dm_device_name(t->md));
drivers/md/dm-table.c
1750
if (dm_table_any_dev_attr(t, device_not_matches_zone_sectors, &zone_sectors)) {
drivers/md/dm-table.c
1752
dm_device_name(t->md));
drivers/md/dm-table.c
1762
int dm_calculate_queue_limits(struct dm_table *t,
drivers/md/dm-table.c
1771
t->integrity_supported = true;
drivers/md/dm-table.c
1772
for (unsigned int i = 0; i < t->num_targets; i++) {
drivers/md/dm-table.c
1773
struct dm_target *ti = dm_table_get_target(t, i);
drivers/md/dm-table.c
1776
t->integrity_supported = false;
drivers/md/dm-table.c
1779
for (unsigned int i = 0; i < t->num_targets; i++) {
drivers/md/dm-table.c
1780
struct dm_target *ti = dm_table_get_target(t, i);
drivers/md/dm-table.c
180
static void dm_table_destroy_crypto_profile(struct dm_table *t);
drivers/md/dm-table.c
182
void dm_table_destroy(struct dm_table *t)
drivers/md/dm-table.c
1826
dm_device_name(t->md),
drivers/md/dm-table.c
1830
if (t->integrity_supported ||
drivers/md/dm-table.c
1835
dm_device_name(t->md),
drivers/md/dm-table.c
1838
t->integrity_supported = false;
drivers/md/dm-table.c
184
if (!t)
drivers/md/dm-table.c
1858
if (validate_hardware_zoned(t, zoned, zone_sectors))
drivers/md/dm-table.c
1861
return validate_hardware_logical_block_alignment(t, limits);
drivers/md/dm-table.c
1868
static bool dm_table_supports_flush(struct dm_table *t)
drivers/md/dm-table.c
1870
for (unsigned int i = 0; i < t->num_targets; i++) {
drivers/md/dm-table.c
1871
struct dm_target *ti = dm_table_get_target(t, i);
drivers/md/dm-table.c
188
if (t->depth >= 2)
drivers/md/dm-table.c
189
kvfree(t->index[t->depth - 2]);
drivers/md/dm-table.c
1906
static bool dm_table_supports_write_zeroes(struct dm_table *t)
drivers/md/dm-table.c
1908
for (unsigned int i = 0; i < t->num_targets; i++) {
drivers/md/dm-table.c
1909
struct dm_target *ti = dm_table_get_target(t, i);
drivers/md/dm-table.c
192
for (unsigned int i = 0; i < t->num_targets; i++) {
drivers/md/dm-table.c
1922
static bool dm_table_supports_nowait(struct dm_table *t)
drivers/md/dm-table.c
1924
for (unsigned int i = 0; i < t->num_targets; i++) {
drivers/md/dm-table.c
1925
struct dm_target *ti = dm_table_get_target(t, i);
drivers/md/dm-table.c
193
struct dm_target *ti = dm_table_get_target(t, i);
drivers/md/dm-table.c
1940
static bool dm_table_supports_discards(struct dm_table *t)
drivers/md/dm-table.c
1942
for (unsigned int i = 0; i < t->num_targets; i++) {
drivers/md/dm-table.c
1943
struct dm_target *ti = dm_table_get_target(t, i);
drivers/md/dm-table.c
1969
static bool dm_table_supports_secure_erase(struct dm_table *t)
drivers/md/dm-table.c
1971
for (unsigned int i = 0; i < t->num_targets; i++) {
drivers/md/dm-table.c
1972
struct dm_target *ti = dm_table_get_target(t, i);
drivers/md/dm-table.c
1992
static bool dm_table_supports_atomic_writes(struct dm_table *t)
drivers/md/dm-table.c
1994
for (unsigned int i = 0; i < t->num_targets; i++) {
drivers/md/dm-table.c
1995
struct dm_target *ti = dm_table_get_target(t, i);
drivers/md/dm-table.c
201
kvfree(t->highs);
drivers/md/dm-table.c
2011
bool dm_table_supports_size_change(struct dm_table *t, sector_t old_size,
drivers/md/dm-table.c
2014
if (IS_ENABLED(CONFIG_BLK_DEV_ZONED) && dm_has_zone_plugs(t->md) &&
drivers/md/dm-table.c
2018
dm_device_name(t->md));
drivers/md/dm-table.c
2028
int dm_table_set_restrictions(struct dm_table *t, struct request_queue *q,
drivers/md/dm-table.c
2034
if (!dm_table_supports_nowait(t))
drivers/md/dm-table.c
204
free_devices(&t->devices, t->md);
drivers/md/dm-table.c
2041
if (!__table_type_bio_based(t->type))
drivers/md/dm-table.c
2044
if (!dm_table_supports_discards(t)) {
drivers/md/dm-table.c
2050
if (!dm_table_supports_write_zeroes(t)) {
drivers/md/dm-table.c
2055
if (!dm_table_supports_secure_erase(t))
drivers/md/dm-table.c
2058
if (dm_table_supports_flush(t))
drivers/md/dm-table.c
206
dm_free_md_mempools(t->mempools);
drivers/md/dm-table.c
2061
if (dm_table_supports_dax(t, device_not_dax_capable))
drivers/md/dm-table.c
2069
r = dm_set_zones_restrictions(t, q, limits);
drivers/md/dm-table.c
2072
} else if (dm_has_zone_plugs(t->md)) {
drivers/md/dm-table.c
2075
dm_device_name(t->md));
drivers/md/dm-table.c
208
dm_table_destroy_crypto_profile(t);
drivers/md/dm-table.c
2080
if (dm_table_supports_atomic_writes(t))
drivers/md/dm-table.c
2094
r = dm_revalidate_zones(t, q);
drivers/md/dm-table.c
210
kfree(t);
drivers/md/dm-table.c
2102
dm_finalize_zone_settings(t, limits);
drivers/md/dm-table.c
2104
if (dm_table_supports_dax(t, device_not_dax_synchronous_capable))
drivers/md/dm-table.c
2105
set_dax_synchronous(t->md->dax_dev);
drivers/md/dm-table.c
2107
if (dm_table_any_dev_attr(t, device_dax_write_cache_enabled, NULL))
drivers/md/dm-table.c
2108
dax_write_cache(t->md->dax_dev, true);
drivers/md/dm-table.c
2110
dm_update_crypto_profile(q, t);
drivers/md/dm-table.c
2114
struct list_head *dm_table_get_devices(struct dm_table *t)
drivers/md/dm-table.c
2116
return &t->devices;
drivers/md/dm-table.c
2119
blk_mode_t dm_table_get_mode(struct dm_table *t)
drivers/md/dm-table.c
2121
return t->mode;
drivers/md/dm-table.c
2131
static void suspend_targets(struct dm_table *t, enum suspend_mode mode)
drivers/md/dm-table.c
2133
lockdep_assert_held(&t->md->suspend_lock);
drivers/md/dm-table.c
2135
for (unsigned int i = 0; i < t->num_targets; i++) {
drivers/md/dm-table.c
2136
struct dm_target *ti = dm_table_get_target(t, i);
drivers/md/dm-table.c
2155
void dm_table_presuspend_targets(struct dm_table *t)
drivers/md/dm-table.c
2157
if (!t)
drivers/md/dm-table.c
2160
suspend_targets(t, PRESUSPEND);
drivers/md/dm-table.c
2163
void dm_table_presuspend_undo_targets(struct dm_table *t)
drivers/md/dm-table.c
2165
if (!t)
drivers/md/dm-table.c
2168
suspend_targets(t, PRESUSPEND_UNDO);
drivers/md/dm-table.c
2171
void dm_table_postsuspend_targets(struct dm_table *t)
drivers/md/dm-table.c
2173
if (!t)
drivers/md/dm-table.c
2176
suspend_targets(t, POSTSUSPEND);
drivers/md/dm-table.c
2179
int dm_table_resume_targets(struct dm_table *t)
drivers/md/dm-table.c
2184
lockdep_assert_held(&t->md->suspend_lock);
drivers/md/dm-table.c
2186
for (i = 0; i < t->num_targets; i++) {
drivers/md/dm-table.c
2187
struct dm_target *ti = dm_table_get_target(t, i);
drivers/md/dm-table.c
2195
dm_device_name(t->md), ti->type->name, r);
drivers/md/dm-table.c
2200
for (i = 0; i < t->num_targets; i++) {
drivers/md/dm-table.c
2201
struct dm_target *ti = dm_table_get_target(t, i);
drivers/md/dm-table.c
2210
struct mapped_device *dm_table_get_md(struct dm_table *t)
drivers/md/dm-table.c
2212
return t->md;
drivers/md/dm-table.c
2216
const char *dm_table_device_name(struct dm_table *t)
drivers/md/dm-table.c
2218
return dm_device_name(t->md);
drivers/md/dm-table.c
2222
void dm_table_run_md_queue_async(struct dm_table *t)
drivers/md/dm-table.c
2224
if (!dm_table_request_based(t))
drivers/md/dm-table.c
2227
if (t->md->queue)
drivers/md/dm-table.c
2228
blk_mq_run_hw_queues(t->md->queue, true);
drivers/md/dm-table.c
371
struct dm_table *t = ti->table;
drivers/md/dm-table.c
373
BUG_ON(!t);
drivers/md/dm-table.c
379
if (dev == disk_devt(t->md->disk))
drivers/md/dm-table.c
382
dd = find_device(&t->devices, dev);
drivers/md/dm-table.c
388
r = dm_get_table_device(t->md, dev, mode, &dd->dm_dev);
drivers/md/dm-table.c
395
list_add(&dd->list, &t->devices);
drivers/md/dm-table.c
399
r = upgrade_mode(dd, mode, t->md);
drivers/md/dm-table.c
482
static int adjoin(struct dm_table *t, struct dm_target *ti)
drivers/md/dm-table.c
486
if (!t->num_targets)
drivers/md/dm-table.c
489
prev = &t->targets[t->num_targets - 1];
drivers/md/dm-table.c
60
static inline sector_t *get_node(struct dm_table *t,
drivers/md/dm-table.c
603
static int validate_hardware_logical_block_alignment(struct dm_table *t,
drivers/md/dm-table.c
63
return t->index[l] + (n * KEYS_PER_NODE);
drivers/md/dm-table.c
631
for (i = 0; i < t->num_targets; i++) {
drivers/md/dm-table.c
632
ti = dm_table_get_target(t, i);
drivers/md/dm-table.c
660
dm_device_name(t->md), i,
drivers/md/dm-table.c
670
int dm_table_add_target(struct dm_table *t, const char *type,
drivers/md/dm-table.c
677
if (t->singleton) {
drivers/md/dm-table.c
679
dm_device_name(t->md), t->targets->type->name);
drivers/md/dm-table.c
683
BUG_ON(t->num_targets >= t->num_allocated);
drivers/md/dm-table.c
685
ti = t->targets + t->num_targets;
drivers/md/dm-table.c
689
DMERR("%s: zero-length target", dm_device_name(t->md));
drivers/md/dm-table.c
693
DMERR("%s: too large device", dm_device_name(t->md));
drivers/md/dm-table.c
699
DMERR("%s: %s: unknown target type", dm_device_name(t->md), type);
drivers/md/dm-table.c
70
static sector_t high(struct dm_table *t, unsigned int l, unsigned int n)
drivers/md/dm-table.c
704
if (t->num_targets) {
drivers/md/dm-table.c
708
t->singleton = true;
drivers/md/dm-table.c
712
!(t->mode & BLK_OPEN_WRITE)) {
drivers/md/dm-table.c
717
if (t->immutable_target_type) {
drivers/md/dm-table.c
718
if (t->immutable_target_type != ti->type) {
drivers/md/dm-table.c
72
for (; l < t->depth - 1; l++)
drivers/md/dm-table.c
723
if (t->num_targets) {
drivers/md/dm-table.c
727
t->immutable_target_type = ti->type;
drivers/md/dm-table.c
730
ti->table = t;
drivers/md/dm-table.c
738
if (!adjoin(t, ti)) {
drivers/md/dm-table.c
75
if (n >= t->counts[l])
drivers/md/dm-table.c
754
t->highs[t->num_targets++] = ti->begin + ti->len - 1;
drivers/md/dm-table.c
758
dm_device_name(t->md), type);
drivers/md/dm-table.c
764
t->flush_bypasses_map = false;
drivers/md/dm-table.c
769
DMERR("%s: %s: %s (%pe)", dm_device_name(t->md), type, ti->error, ERR_PTR(r));
drivers/md/dm-table.c
78
return get_node(t, l, n)[KEYS_PER_NODE - 1];
drivers/md/dm-table.c
843
void dm_table_set_type(struct dm_table *t, enum dm_queue_mode type)
drivers/md/dm-table.c
845
t->type = type;
drivers/md/dm-table.c
85
static int setup_btree_index(unsigned int l, struct dm_table *t)
drivers/md/dm-table.c
867
static bool dm_table_supports_dax(struct dm_table *t,
drivers/md/dm-table.c
871
for (unsigned int i = 0; i < t->num_targets; i++) {
drivers/md/dm-table.c
872
struct dm_target *ti = dm_table_get_target(t, i);
drivers/md/dm-table.c
899
static int dm_table_determine_type(struct dm_table *t)
drivers/md/dm-table.c
90
for (n = 0U; n < t->counts[l]; n++) {
drivers/md/dm-table.c
903
struct list_head *devices = dm_table_get_devices(t);
drivers/md/dm-table.c
904
enum dm_queue_mode live_md_type = dm_get_md_type(t->md);
drivers/md/dm-table.c
906
if (t->type != DM_TYPE_NONE) {
drivers/md/dm-table.c
908
if (t->type == DM_TYPE_BIO_BASED) {
drivers/md/dm-table.c
91
node = get_node(t, l, n);
drivers/md/dm-table.c
912
BUG_ON(t->type == DM_TYPE_DAX_BIO_BASED);
drivers/md/dm-table.c
916
for (unsigned int i = 0; i < t->num_targets; i++) {
drivers/md/dm-table.c
917
ti = dm_table_get_target(t, i);
drivers/md/dm-table.c
94
node[k] = high(t, l + 1, get_child(n, k));
drivers/md/dm-table.c
946
t->type = DM_TYPE_BIO_BASED;
drivers/md/dm-table.c
947
if (dm_table_supports_dax(t, device_not_dax_capable) ||
drivers/md/dm-table.c
949
t->type = DM_TYPE_DAX_BIO_BASED;
drivers/md/dm-table.c
956
t->type = DM_TYPE_REQUEST_BASED;
drivers/md/dm-table.c
965
if (t->num_targets > 1) {
drivers/md/dm-table.c
972
struct dm_table *live_table = dm_get_live_table(t->md, &srcu_idx);
drivers/md/dm-table.c
976
t->type = live_table->type;
drivers/md/dm-table.c
977
dm_put_live_table(t->md, srcu_idx);
drivers/md/dm-table.c
981
ti = dm_table_get_immutable_target(t);
drivers/md/dm-thin-metadata.c
308
static uint64_t pack_block_time(dm_block_t b, uint32_t t)
drivers/md/dm-thin-metadata.c
310
return (b << 24) | t;
drivers/md/dm-thin-metadata.c
313
static void unpack_block_time(uint64_t v, dm_block_t *b, uint32_t *t)
drivers/md/dm-thin-metadata.c
316
*t = v & ((1 << 24) - 1);
drivers/md/dm-thin-metadata.c
329
uint32_t t;
drivers/md/dm-thin-metadata.c
335
unpack_block_time(le64_to_cpu(*value_le), &b, &t);
drivers/md/dm-thin-metadata.c
372
uint32_t t;
drivers/md/dm-thin-metadata.c
376
unpack_block_time(le64_to_cpu(v1_le), &b1, &t);
drivers/md/dm-thin-metadata.c
377
unpack_block_time(le64_to_cpu(v2_le), &b2, &t);
drivers/md/dm-thin.c
154
static void throttle_init(struct throttle *t)
drivers/md/dm-thin.c
156
init_rwsem(&t->lock);
drivers/md/dm-thin.c
157
t->throttle_applied = false;
drivers/md/dm-thin.c
160
static void throttle_work_start(struct throttle *t)
drivers/md/dm-thin.c
162
t->threshold = jiffies + THROTTLE_THRESHOLD;
drivers/md/dm-thin.c
165
static void throttle_work_update(struct throttle *t)
drivers/md/dm-thin.c
167
if (!t->throttle_applied && time_is_before_jiffies(t->threshold)) {
drivers/md/dm-thin.c
168
down_write(&t->lock);
drivers/md/dm-thin.c
169
t->throttle_applied = true;
drivers/md/dm-thin.c
173
static void throttle_work_complete(struct throttle *t)
drivers/md/dm-thin.c
175
if (t->throttle_applied) {
drivers/md/dm-thin.c
176
t->throttle_applied = false;
drivers/md/dm-thin.c
177
up_write(&t->lock);
drivers/md/dm-thin.c
181
static void throttle_lock(struct throttle *t)
drivers/md/dm-thin.c
183
down_read(&t->lock);
drivers/md/dm-thin.c
186
static void throttle_unlock(struct throttle *t)
drivers/md/dm-thin.c
188
up_read(&t->lock);
drivers/md/dm-vdo/dedupe.c
2340
static void timeout_index_operations(struct timer_list *t)
drivers/md/dm-vdo/dedupe.c
2342
struct hash_zone *zone = timer_container_of(zone, t, timer);
drivers/md/dm-writecache.c
706
static void writecache_max_age_timer(struct timer_list *t)
drivers/md/dm-writecache.c
708
struct dm_writecache *wc = timer_container_of(wc, t, max_age_timer);
drivers/md/dm-writecache.c
866
static void writecache_autocommit_timer(struct timer_list *t)
drivers/md/dm-writecache.c
868
struct dm_writecache *wc = timer_container_of(wc, t, autocommit_timer);
drivers/md/dm-zone.c
172
int dm_revalidate_zones(struct dm_table *t, struct request_queue *q)
drivers/md/dm-zone.c
174
struct mapped_device *md = t->md;
drivers/md/dm-zone.c
19
static int dm_blk_do_report_zones(struct mapped_device *md, struct dm_table *t,
drivers/md/dm-zone.c
197
md->zone_revalidate_map = t;
drivers/md/dm-zone.c
219
static bool dm_table_supports_zone_append(struct dm_table *t)
drivers/md/dm-zone.c
221
for (unsigned int i = 0; i < t->num_targets; i++) {
drivers/md/dm-zone.c
222
struct dm_target *ti = dm_table_get_target(t, i);
drivers/md/dm-zone.c
27
tgt = dm_table_find_target(t, args->next_sector);
drivers/md/dm-zone.c
349
int dm_set_zones_restrictions(struct dm_table *t, struct request_queue *q,
drivers/md/dm-zone.c
352
struct mapped_device *md = t->md;
drivers/md/dm-zone.c
366
if (!dm_table_supports_zone_append(t))
drivers/md/dm-zone.c
376
for (unsigned int i = 0; i < t->num_targets; i++) {
drivers/md/dm-zone.c
377
struct dm_target *ti = dm_table_get_target(t, i);
drivers/md/dm-zone.c
411
if (get_capacity(disk) && dm_has_zone_plugs(t->md)) {
drivers/md/dm-zone.c
419
!dm_table_is_wildcard(t)) {
drivers/md/dm-zone.c
445
void dm_finalize_zone_settings(struct dm_table *t, struct queue_limits *lim)
drivers/md/dm-zone.c
447
struct mapped_device *md = t->md;
drivers/md/dm-zone.c
450
if (dm_table_supports_zone_append(t))
drivers/md/dm-zone.c
501
int dm_zone_get_reset_bitmap(struct mapped_device *md, struct dm_table *t,
drivers/md/dm-zone.c
513
ret = dm_blk_do_report_zones(md, t, nr_zones, &args);
drivers/md/dm.c
1530
struct dm_table *t = ci->map;
drivers/md/dm.c
1549
if (!t->flush_bypasses_map) {
drivers/md/dm.c
1550
for (unsigned int i = 0; i < t->num_targets; i++) {
drivers/md/dm.c
1552
struct dm_target *ti = dm_table_get_target(t, i);
drivers/md/dm.c
1568
struct list_head *devices = dm_table_get_devices(t);
drivers/md/dm.c
1907
struct dm_table *t = ci->map;
drivers/md/dm.c
1910
for (unsigned int i = 0; i < t->num_targets; i++) {
drivers/md/dm.c
1911
struct dm_target *ti = dm_table_get_target(t, i);
drivers/md/dm.c
2445
static struct dm_table *__bind(struct mapped_device *md, struct dm_table *t,
drivers/md/dm.c
2453
size = dm_table_get_size(t);
drivers/md/dm.c
2457
if (!dm_table_supports_size_change(t, old_size, size)) {
drivers/md/dm.c
2465
int ret = dm_table_set_restrictions(t, md->queue, limits);
drivers/md/dm.c
2479
dm_table_event_callback(t, event_callback, md);
drivers/md/dm.c
2481
if (dm_table_request_based(t)) {
drivers/md/dm.c
2486
md->immutable_target = dm_table_get_immutable_target(t);
drivers/md/dm.c
2497
md->mempools = t->mempools;
drivers/md/dm.c
2499
dm_free_md_mempools(t->mempools);
drivers/md/dm.c
2507
md->mempools = t->mempools;
drivers/md/dm.c
2509
t->mempools = NULL;
drivers/md/dm.c
2512
rcu_assign_pointer(md->map, (void *)t);
drivers/md/dm.c
2513
md->immutable_target_type = dm_table_get_immutable_target_type(t);
drivers/md/dm.c
2582
int dm_setup_md_queue(struct mapped_device *md, struct dm_table *t)
drivers/md/dm.c
2584
enum dm_queue_mode type = dm_table_get_type(t);
drivers/md/dm.c
2593
r = dm_mq_init_request_queue(md, t);
drivers/md/dm.c
2600
r = dm_calculate_queue_limits(t, &limits);
drivers/md/dm.c
2605
r = dm_table_set_restrictions(t, md->queue, &limits);
drivers/md/dm.h
100
#define dm_target_hybrid(t) (dm_target_bio_based(t) && dm_target_request_based(t))
drivers/md/dm.h
105
int dm_set_zones_restrictions(struct dm_table *t, struct request_queue *q,
drivers/md/dm.h
107
int dm_revalidate_zones(struct dm_table *t, struct request_queue *q);
drivers/md/dm.h
108
void dm_finalize_zone_settings(struct dm_table *t, struct queue_limits *lim);
drivers/md/dm.h
115
int dm_zone_get_reset_bitmap(struct mapped_device *md, struct dm_table *t,
drivers/md/dm.h
57
void dm_table_event_callback(struct dm_table *t,
drivers/md/dm.h
59
struct dm_target *dm_table_find_target(struct dm_table *t, sector_t sector);
drivers/md/dm.h
61
bool dm_table_is_wildcard(struct dm_table *t);
drivers/md/dm.h
64
int dm_table_set_restrictions(struct dm_table *t, struct request_queue *q,
drivers/md/dm.h
66
struct list_head *dm_table_get_devices(struct dm_table *t);
drivers/md/dm.h
67
void dm_table_presuspend_targets(struct dm_table *t);
drivers/md/dm.h
68
void dm_table_presuspend_undo_targets(struct dm_table *t);
drivers/md/dm.h
69
void dm_table_postsuspend_targets(struct dm_table *t);
drivers/md/dm.h
70
int dm_table_resume_targets(struct dm_table *t);
drivers/md/dm.h
71
enum dm_queue_mode dm_table_get_type(struct dm_table *t);
drivers/md/dm.h
72
struct target_type *dm_table_get_immutable_target_type(struct dm_table *t);
drivers/md/dm.h
73
struct dm_target *dm_table_get_immutable_target(struct dm_table *t);
drivers/md/dm.h
74
struct dm_target *dm_table_get_wildcard_target(struct dm_table *t);
drivers/md/dm.h
75
bool dm_table_request_based(struct dm_table *t);
drivers/md/dm.h
76
bool dm_table_supports_size_change(struct dm_table *t, sector_t old_size,
drivers/md/dm.h
84
int dm_setup_md_queue(struct mapped_device *md, struct dm_table *t);
drivers/md/dm.h
89
#define dm_target_bio_based(t) ((t)->type->map != NULL)
drivers/md/dm.h
94
#define dm_target_request_based(t) ((t)->type->clone_and_map_rq != NULL)
drivers/md/md.c
6437
static void md_safemode_timeout(struct timer_list *t)
drivers/md/md.c
6439
struct mddev *mddev = timer_container_of(mddev, t, safemode_timer);
drivers/md/md.c
669
static void md_safemode_timeout(struct timer_list *t);
drivers/md/md.c
8516
struct md_thread *t;
drivers/md/md.c
8519
t = rcu_dereference(*thread);
drivers/md/md.c
8520
if (t)
drivers/md/md.c
8521
wake_up_process(t->tsk);
drivers/md/md.c
8527
struct md_thread *t;
drivers/md/md.c
8529
t = rcu_dereference(thread);
drivers/md/md.c
8530
if (t) {
drivers/md/md.c
8531
pr_debug("md: waking up MD thread %s.\n", t->tsk->comm);
drivers/md/md.c
8532
set_bit(THREAD_WAKEUP, &t->flags);
drivers/md/md.c
8533
if (wq_has_sleeper(&t->wqueue))
drivers/md/md.c
8534
wake_up(&t->wqueue);
drivers/md/persistent-data/dm-block-manager.c
79
struct stack_store *t;
drivers/md/persistent-data/dm-block-manager.c
86
t = lock->traces + h;
drivers/md/persistent-data/dm-block-manager.c
87
t->nr_entries = stack_trace_save(t->entries, MAX_STACK, 2);
drivers/md/persistent-data/dm-space-map-metadata.c
33
static void threshold_init(struct threshold *t)
drivers/md/persistent-data/dm-space-map-metadata.c
35
t->threshold_set = false;
drivers/md/persistent-data/dm-space-map-metadata.c
36
t->value_set = false;
drivers/md/persistent-data/dm-space-map-metadata.c
39
static void set_threshold(struct threshold *t, dm_block_t value,
drivers/md/persistent-data/dm-space-map-metadata.c
42
t->threshold_set = true;
drivers/md/persistent-data/dm-space-map-metadata.c
43
t->threshold = value;
drivers/md/persistent-data/dm-space-map-metadata.c
44
t->fn = fn;
drivers/md/persistent-data/dm-space-map-metadata.c
45
t->context = context;
drivers/md/persistent-data/dm-space-map-metadata.c
48
static bool below_threshold(struct threshold *t, dm_block_t value)
drivers/md/persistent-data/dm-space-map-metadata.c
50
return t->threshold_set && value <= t->threshold;
drivers/md/persistent-data/dm-space-map-metadata.c
53
static bool threshold_already_triggered(struct threshold *t)
drivers/md/persistent-data/dm-space-map-metadata.c
55
return t->value_set && below_threshold(t, t->current_value);
drivers/md/persistent-data/dm-space-map-metadata.c
58
static void check_threshold(struct threshold *t, dm_block_t value)
drivers/md/persistent-data/dm-space-map-metadata.c
60
if (below_threshold(t, value) &&
drivers/md/persistent-data/dm-space-map-metadata.c
61
!threshold_already_triggered(t))
drivers/md/persistent-data/dm-space-map-metadata.c
62
t->fn(t->context);
drivers/md/persistent-data/dm-space-map-metadata.c
64
t->value_set = true;
drivers/md/persistent-data/dm-space-map-metadata.c
65
t->current_value = value;
drivers/md/raid5.c
372
struct stripe_head *sh, *t;
drivers/md/raid5.c
378
llist_for_each_entry_safe(sh, t, head, release_list) {
drivers/media/common/saa7146/saa7146_fops.c
154
void saa7146_buffer_timeout(struct timer_list *t)
drivers/media/common/saa7146/saa7146_fops.c
156
struct saa7146_dmaqueue *q = timer_container_of(q, t, timeout);
drivers/media/common/saa7146/saa7146_vbi.c
331
static void vbi_read_timeout(struct timer_list *t)
drivers/media/common/saa7146/saa7146_vbi.c
333
struct saa7146_vv *vv = timer_container_of(vv, t, vbi_read_timeout);
drivers/media/dvb-core/dmxdev.c
356
static void dvb_dmxdev_filter_timeout(struct timer_list *t)
drivers/media/dvb-core/dmxdev.c
359
t, timer);
drivers/media/dvb-frontends/atbm8830.c
109
u64 t;
drivers/media/dvb-frontends/atbm8830.c
115
t = (u64) 2 * 31416 * (freq - fs);
drivers/media/dvb-frontends/atbm8830.c
116
t <<= 22;
drivers/media/dvb-frontends/atbm8830.c
117
do_div(t, fs);
drivers/media/dvb-frontends/atbm8830.c
118
do_div(t, 1000);
drivers/media/dvb-frontends/atbm8830.c
119
val = t;
drivers/media/dvb-frontends/atbm8830.c
352
u8 t;
drivers/media/dvb-frontends/atbm8830.c
358
atbm8830_read_reg(priv, REG_FRAME_ERR_CNT + 1, &t);
drivers/media/dvb-frontends/atbm8830.c
359
frame_err = t & 0x7F;
drivers/media/dvb-frontends/atbm8830.c
361
atbm8830_read_reg(priv, REG_FRAME_ERR_CNT, &t);
drivers/media/dvb-frontends/atbm8830.c
362
frame_err |= t;
drivers/media/dvb-frontends/atbm8830.c
376
u8 t;
drivers/media/dvb-frontends/atbm8830.c
381
atbm8830_read_reg(priv, REG_AGC_PWM_VAL + 1, &t);
drivers/media/dvb-frontends/atbm8830.c
382
pwm = t & 0x03;
drivers/media/dvb-frontends/atbm8830.c
384
atbm8830_read_reg(priv, REG_AGC_PWM_VAL, &t);
drivers/media/dvb-frontends/atbm8830.c
385
pwm |= t;
drivers/media/dvb-frontends/atbm8830.c
91
u64 t;
drivers/media/dvb-frontends/atbm8830.c
94
t = (u64)0x100000 * freq;
drivers/media/dvb-frontends/atbm8830.c
95
do_div(t, 30400);
drivers/media/dvb-frontends/atbm8830.c
96
val = t;
drivers/media/dvb-frontends/bcm3510.c
158
unsigned long t;
drivers/media/dvb-frontends/bcm3510.c
190
t = jiffies + 1*HZ;
drivers/media/dvb-frontends/bcm3510.c
191
while (time_before(jiffies, t)) {
drivers/media/dvb-frontends/bcm3510.c
336
s32 t;
drivers/media/dvb-frontends/bcm3510.c
339
t = st->status2.SIGNAL;
drivers/media/dvb-frontends/bcm3510.c
341
if (t > 190)
drivers/media/dvb-frontends/bcm3510.c
342
t = 190;
drivers/media/dvb-frontends/bcm3510.c
343
if (t < 90)
drivers/media/dvb-frontends/bcm3510.c
344
t = 90;
drivers/media/dvb-frontends/bcm3510.c
346
t -= 90;
drivers/media/dvb-frontends/bcm3510.c
347
t = t * 0xff / 100;
drivers/media/dvb-frontends/bcm3510.c
349
*strength = (t << 8) | t;
drivers/media/dvb-frontends/bcm3510.c
683
unsigned long t;
drivers/media/dvb-frontends/bcm3510.c
690
t = jiffies + 3*HZ;
drivers/media/dvb-frontends/bcm3510.c
691
while (time_before(jiffies, t)) {
drivers/media/dvb-frontends/bcm3510.c
707
unsigned long t;
drivers/media/dvb-frontends/bcm3510.c
713
t = jiffies + 3*HZ;
drivers/media/dvb-frontends/bcm3510.c
714
while (time_before(jiffies, t)) {
drivers/media/dvb-frontends/dib8000.c
3889
u32 t = (s/n) << 16;
drivers/media/dvb-frontends/dib8000.c
3890
return t + ((s << 16) - n*t) / n;
drivers/media/dvb-frontends/dib9000.c
1588
u16 i, len, t, index_msg;
drivers/media/dvb-frontends/dib9000.c
1611
t = dib9000_read_word(state, 785);
drivers/media/dvb-frontends/dib9000.c
1612
msg[index_msg].buf[i] = (t >> 8) & 0xff;
drivers/media/dvb-frontends/dib9000.c
1613
msg[index_msg].buf[i + 1] = (t) & 0xff;
drivers/media/dvb-frontends/dib9000.c
2314
u32 t = (s / n) << 16;
drivers/media/dvb-frontends/dib9000.c
2315
return t + ((s << 16) - n * t) / n;
drivers/media/dvb-frontends/lgs8gxx.c
104
u8 t;
drivers/media/dvb-frontends/lgs8gxx.c
108
lgs8gxx_read_reg(priv, reg, &t);
drivers/media/dvb-frontends/lgs8gxx.c
110
if ((t & mask) == val)
drivers/media/dvb-frontends/lgs8gxx.c
182
u8 reg_addr, t;
drivers/media/dvb-frontends/lgs8gxx.c
191
lgs8gxx_read_reg(priv, reg_addr, &t);
drivers/media/dvb-frontends/lgs8gxx.c
193
v32 |= t;
drivers/media/dvb-frontends/lgs8gxx.c
206
u8 t;
drivers/media/dvb-frontends/lgs8gxx.c
213
lgs8gxx_read_reg(priv, 0x0C, &t);
drivers/media/dvb-frontends/lgs8gxx.c
214
t &= (~0x04);
drivers/media/dvb-frontends/lgs8gxx.c
215
lgs8gxx_write_reg(priv, 0x0C, t | 0x80);
drivers/media/dvb-frontends/lgs8gxx.c
223
lgs8gxx_read_reg(priv, 0x7E, &t);
drivers/media/dvb-frontends/lgs8gxx.c
224
lgs8gxx_write_reg(priv, 0x7E, t | 0x01);
drivers/media/dvb-frontends/lgs8gxx.c
227
lgs8gxx_read_reg(priv, 0xC5, &t);
drivers/media/dvb-frontends/lgs8gxx.c
228
lgs8gxx_write_reg(priv, 0xC5, t & 0xE0);
drivers/media/dvb-frontends/lgs8gxx.c
235
lgs8gxx_read_reg(priv, 0x7C, &t);
drivers/media/dvb-frontends/lgs8gxx.c
236
t = (t & 0x8C) | 0x03;
drivers/media/dvb-frontends/lgs8gxx.c
237
lgs8gxx_write_reg(priv, 0x7C, t);
drivers/media/dvb-frontends/lgs8gxx.c
240
lgs8gxx_read_reg(priv, 0xC3, &t);
drivers/media/dvb-frontends/lgs8gxx.c
241
t = (t & 0xEF) | 0x10;
drivers/media/dvb-frontends/lgs8gxx.c
242
lgs8gxx_write_reg(priv, 0xC3, t);
drivers/media/dvb-frontends/lgs8gxx.c
253
u8 t;
drivers/media/dvb-frontends/lgs8gxx.c
257
lgs8gxx_read_reg(priv, 0x0C, &t);
drivers/media/dvb-frontends/lgs8gxx.c
258
t &= (~0x80);
drivers/media/dvb-frontends/lgs8gxx.c
259
lgs8gxx_write_reg(priv, 0x0C, t);
drivers/media/dvb-frontends/lgs8gxx.c
261
lgs8gxx_read_reg(priv, 0x0C, &t);
drivers/media/dvb-frontends/lgs8gxx.c
264
if (((t&0x03) == 0x01) && (t2&0x01)) {
drivers/media/dvb-frontends/lgs8gxx.c
286
lgs8gxx_read_reg(priv, 0xC5, &t);
drivers/media/dvb-frontends/lgs8gxx.c
287
t = (t & 0xE0) | 0x06;
drivers/media/dvb-frontends/lgs8gxx.c
288
lgs8gxx_write_reg(priv, 0xC5, t);
drivers/media/dvb-frontends/lgs8gxx.c
298
u8 t;
drivers/media/dvb-frontends/lgs8gxx.c
301
ret = lgs8gxx_read_reg(priv, 0x13, &t);
drivers/media/dvb-frontends/lgs8gxx.c
303
ret = lgs8gxx_read_reg(priv, 0x4B, &t);
drivers/media/dvb-frontends/lgs8gxx.c
308
*locked = ((t & 0x80) == 0x80) ? 1 : 0;
drivers/media/dvb-frontends/lgs8gxx.c
310
*locked = ((t & 0xC0) == 0xC0) ? 1 : 0;
drivers/media/dvb-frontends/lgs8gxx.c
443
u8 t;
drivers/media/dvb-frontends/lgs8gxx.c
446
lgs8gxx_read_reg(priv, 0xA2, &t);
drivers/media/dvb-frontends/lgs8gxx.c
447
*detected_param = t;
drivers/media/dvb-frontends/lgs8gxx.c
449
lgs8gxx_read_reg(priv, 0x1F, &t);
drivers/media/dvb-frontends/lgs8gxx.c
450
*detected_param = t & 0x3F;
drivers/media/dvb-frontends/lgs8gxx.c
490
u8 t;
drivers/media/dvb-frontends/lgs8gxx.c
491
lgs8gxx_read_reg(priv, 0x19, &t);
drivers/media/dvb-frontends/lgs8gxx.c
492
t &= 0x81;
drivers/media/dvb-frontends/lgs8gxx.c
493
t |= detected_param << 1;
drivers/media/dvb-frontends/lgs8gxx.c
494
lgs8gxx_write_reg(priv, 0x19, t);
drivers/media/dvb-frontends/lgs8gxx.c
521
u8 t, reg_addr;
drivers/media/dvb-frontends/lgs8gxx.c
524
ret = lgs8gxx_read_reg(priv, reg_addr, &t);
drivers/media/dvb-frontends/lgs8gxx.c
528
t &= 0xF8;
drivers/media/dvb-frontends/lgs8gxx.c
529
t |= serial ? TS_SERIAL : TS_PARALLEL;
drivers/media/dvb-frontends/lgs8gxx.c
530
t |= clk_pol ? TS_CLK_INVERTED : TS_CLK_NORMAL;
drivers/media/dvb-frontends/lgs8gxx.c
531
t |= clk_gated ? TS_CLK_GATED : TS_CLK_FREERUN;
drivers/media/dvb-frontends/lgs8gxx.c
533
ret = lgs8gxx_write_reg(priv, reg_addr, t);
drivers/media/dvb-frontends/lgs8gxx.c
561
u8 t;
drivers/media/dvb-frontends/lgs8gxx.c
566
lgs8gxx_read_reg(priv, 0x7c, &t);
drivers/media/dvb-frontends/lgs8gxx.c
567
lgs8gxx_write_reg(priv, 0x7c, (t&0x8c) | 0x3);
drivers/media/dvb-frontends/lgs8gxx.c
570
lgs8gxx_read_reg(priv, 0xc3, &t);
drivers/media/dvb-frontends/lgs8gxx.c
571
lgs8gxx_write_reg(priv, 0xc3, t&0x10);
drivers/media/dvb-frontends/lgs8gxx.c
716
u8 t, locked = 0;
drivers/media/dvb-frontends/lgs8gxx.c
730
ret = lgs8gxx_read_reg(priv, 0x4B, &t);
drivers/media/dvb-frontends/lgs8gxx.c
734
dprintk("Reg 0x4B: 0x%02X\n", t);
drivers/media/dvb-frontends/lgs8gxx.c
738
if ((t & 0x40) == 0x40)
drivers/media/dvb-frontends/lgs8gxx.c
740
if ((t & 0x80) == 0x80)
drivers/media/dvb-frontends/lgs8gxx.c
744
if ((t & 0x80) == 0x80)
drivers/media/dvb-frontends/lgs8gxx.c
791
u8 t; s8 ret;
drivers/media/dvb-frontends/lgs8gxx.c
798
ret = lgs8gxx_read_reg(priv, 0x4B, &t);
drivers/media/dvb-frontends/lgs8gxx.c
803
if ((t & 0xC0) == 0xC0) {
drivers/media/dvb-frontends/lgs8gxx.c
826
lgs8gxx_read_reg(priv, 0x95, &t);
drivers/media/dvb-frontends/lgs8gxx.c
827
dprintk("%s: AVG Noise=0x%02X\n", __func__, t);
drivers/media/dvb-frontends/lgs8gxx.c
834
u8 t;
drivers/media/dvb-frontends/lgs8gxx.c
839
lgs8gxx_read_reg(priv, 0xB1, &t);
drivers/media/dvb-frontends/lgs8gxx.c
840
v |= t;
drivers/media/dvb-frontends/lgs8gxx.c
842
lgs8gxx_read_reg(priv, 0xB0, &t);
drivers/media/dvb-frontends/lgs8gxx.c
843
v |= t;
drivers/media/dvb-frontends/lgs8gxx.c
866
u8 t;
drivers/media/dvb-frontends/lgs8gxx.c
870
lgs8gxx_read_reg(priv, 0x34, &t);
drivers/media/dvb-frontends/lgs8gxx.c
872
lgs8gxx_read_reg(priv, 0x95, &t);
drivers/media/dvb-frontends/lgs8gxx.c
873
dprintk("AVG Noise=0x%02X\n", t);
drivers/media/dvb-frontends/lgs8gxx.c
874
*snr = 256 - t;
drivers/media/dvb-frontends/lgs8gxx.c
890
u8 orig, t;
drivers/media/dvb-frontends/lgs8gxx.c
895
t = orig | 0x10;
drivers/media/dvb-frontends/lgs8gxx.c
896
lgs8gxx_write_reg(priv, 0x30, t);
drivers/media/dvb-frontends/lgs8gxx.c
897
t = orig | 0x18;
drivers/media/dvb-frontends/lgs8gxx.c
898
lgs8gxx_write_reg(priv, 0x30, t);
drivers/media/dvb-frontends/lgs8gxx.c
899
t = orig | 0x10;
drivers/media/dvb-frontends/lgs8gxx.c
900
lgs8gxx_write_reg(priv, 0x30, t);
drivers/media/dvb-frontends/lgs8gxx.c
910
u8 t;
drivers/media/dvb-frontends/lgs8gxx.c
913
lgs8gxx_read_reg(priv, 0x30, &t);
drivers/media/dvb-frontends/lgs8gxx.c
914
t &= 0xE7;
drivers/media/dvb-frontends/lgs8gxx.c
915
lgs8gxx_write_reg(priv, 0x30, t);
drivers/media/dvb-frontends/lgs8gxx.c
924
u8 reg_err, reg_total, t;
drivers/media/dvb-frontends/lgs8gxx.c
942
lgs8gxx_read_reg(priv, reg_total+3-i, &t);
drivers/media/dvb-frontends/lgs8gxx.c
943
total_cnt |= t;
drivers/media/dvb-frontends/lgs8gxx.c
947
lgs8gxx_read_reg(priv, reg_err+3-i, &t);
drivers/media/dvb-frontends/lgs8gxx.c
948
err_cnt |= t;
drivers/media/dvb-frontends/mt312.c
393
const enum fe_sec_tone_mode t)
drivers/media/dvb-frontends/mt312.c
401
if (t > SEC_TONE_OFF)
drivers/media/dvb-frontends/mt312.c
409
(diseqc_mode & 0x40) | tone_tab[t]);
drivers/media/dvb-frontends/tda10048.c
295
u64 t;
drivers/media/dvb-frontends/tda10048.c
304
t = if_hz;
drivers/media/dvb-frontends/tda10048.c
305
t *= 10;
drivers/media/dvb-frontends/tda10048.c
306
t *= 32768;
drivers/media/dvb-frontends/tda10048.c
307
do_div(t, sample_freq_hz);
drivers/media/dvb-frontends/tda10048.c
308
t += 5;
drivers/media/dvb-frontends/tda10048.c
309
do_div(t, 10);
drivers/media/dvb-frontends/tda10048.c
312
t = sample_freq_hz - if_hz;
drivers/media/dvb-frontends/tda10048.c
313
t *= 10;
drivers/media/dvb-frontends/tda10048.c
314
t *= 32768;
drivers/media/dvb-frontends/tda10048.c
315
do_div(t, sample_freq_hz);
drivers/media/dvb-frontends/tda10048.c
316
t += 5;
drivers/media/dvb-frontends/tda10048.c
317
do_div(t, 10);
drivers/media/dvb-frontends/tda10048.c
318
t = ~t + 1;
drivers/media/dvb-frontends/tda10048.c
321
tda10048_writereg(state, TDA10048_FREQ_PHY2_LSB, (u8)t);
drivers/media/dvb-frontends/tda10048.c
322
tda10048_writereg(state, TDA10048_FREQ_PHY2_MSB, (u8)(t >> 8));
drivers/media/dvb-frontends/tda10048.c
331
u64 t;
drivers/media/dvb-frontends/tda10048.c
340
t = bw * 10;
drivers/media/dvb-frontends/tda10048.c
343
t *= (2048 * 1024);
drivers/media/dvb-frontends/tda10048.c
344
t *= 1024;
drivers/media/dvb-frontends/tda10048.c
351
do_div(t, z);
drivers/media/dvb-frontends/tda10048.c
352
t += 5;
drivers/media/dvb-frontends/tda10048.c
353
do_div(t, 10);
drivers/media/dvb-frontends/tda10048.c
355
tda10048_writereg(state, TDA10048_TIME_WREF_LSB, (u8)t);
drivers/media/dvb-frontends/tda10048.c
356
tda10048_writereg(state, TDA10048_TIME_WREF_MID1, (u8)(t >> 8));
drivers/media/dvb-frontends/tda10048.c
357
tda10048_writereg(state, TDA10048_TIME_WREF_MID2, (u8)(t >> 16));
drivers/media/dvb-frontends/tda10048.c
358
tda10048_writereg(state, TDA10048_TIME_WREF_MSB, (u8)(t >> 24));
drivers/media/dvb-frontends/tda10048.c
367
u64 t;
drivers/media/dvb-frontends/tda10048.c
375
t = sample_freq_hz;
drivers/media/dvb-frontends/tda10048.c
376
t *= 7;
drivers/media/dvb-frontends/tda10048.c
377
t *= 32;
drivers/media/dvb-frontends/tda10048.c
378
t *= 10;
drivers/media/dvb-frontends/tda10048.c
379
do_div(t, bw);
drivers/media/dvb-frontends/tda10048.c
380
t += 5;
drivers/media/dvb-frontends/tda10048.c
381
do_div(t, 10);
drivers/media/dvb-frontends/tda10048.c
383
tda10048_writereg(state, TDA10048_TIME_INVWREF_LSB, (u8)t);
drivers/media/dvb-frontends/tda10048.c
384
tda10048_writereg(state, TDA10048_TIME_INVWREF_MSB, (u8)(t >> 8));
drivers/media/i2c/adp1653.c
33
#define TIMEOUT_US_TO_CODE(t) ((TIMEOUT_MAX + (TIMEOUT_STEP / 2) - (t)) \
drivers/media/i2c/adv748x/adv748x.h
413
#define tx_read(t, r) adv748x_read(t->state, t->page, r)
drivers/media/i2c/adv748x/adv748x.h
414
#define tx_write(t, r, v) adv748x_write(t->state, t->page, r, v)
drivers/media/i2c/adv7604.c
230
static bool adv76xx_check_dv_timings(const struct v4l2_dv_timings *t, void *hdl)
drivers/media/i2c/adv7604.c
235
if (v4l2_match_dv_timings(t, adv76xx_timings_exceptions + i, 0, false))
drivers/media/i2c/adv7604.c
335
static inline unsigned htotal(const struct v4l2_bt_timings *t)
drivers/media/i2c/adv7604.c
337
return V4L2_DV_BT_FRAME_WIDTH(t);
drivers/media/i2c/adv7604.c
340
static inline unsigned vtotal(const struct v4l2_bt_timings *t)
drivers/media/i2c/adv7604.c
342
return V4L2_DV_BT_FRAME_HEIGHT(t);
drivers/media/i2c/adv7842.c
152
static bool adv7842_check_dv_timings(const struct v4l2_dv_timings *t, void *hdl)
drivers/media/i2c/adv7842.c
157
if (v4l2_match_dv_timings(t, adv7842_timings_exceptions + i, 0, false))
drivers/media/i2c/adv7842.c
262
static inline unsigned htotal(const struct v4l2_bt_timings *t)
drivers/media/i2c/adv7842.c
264
return V4L2_DV_BT_FRAME_WIDTH(t);
drivers/media/i2c/adv7842.c
267
static inline unsigned vtotal(const struct v4l2_bt_timings *t)
drivers/media/i2c/adv7842.c
269
return V4L2_DV_BT_FRAME_HEIGHT(t);
drivers/media/i2c/et8ek8/et8ek8_driver.c
453
#define TIMEPERFRAME_AVG_FPS(t) \
drivers/media/i2c/et8ek8/et8ek8_driver.c
454
(((t).denominator + ((t).numerator >> 1)) / (t).numerator)
drivers/media/i2c/ov7251.c
1182
static inline u32 avg_fps(const struct v4l2_fract *t)
drivers/media/i2c/ov7251.c
1184
return (t->denominator + (t->numerator >> 1)) / t->numerator;
drivers/media/i2c/sony-btf-mpx.c
157
static int mpx_setup(struct sony_btf_mpx *t)
drivers/media/i2c/sony-btf-mpx.c
159
struct i2c_client *client = v4l2_get_subdevdata(&t->sd);
drivers/media/i2c/sony-btf-mpx.c
163
int mode = t->mpxmode;
drivers/media/i2c/sony-btf-mpx.c
177
if (t->audmode != V4L2_TUNER_MODE_MONO)
drivers/media/i2c/sony-btf-mpx.c
181
switch (t->audmode) {
drivers/media/i2c/sony-btf-mpx.c
224
t->audmode == V4L2_TUNER_MODE_MONO ? 0x07f0 : 0x0190);
drivers/media/i2c/sony-btf-mpx.c
269
struct sony_btf_mpx *t = to_state(sd);
drivers/media/i2c/sony-btf-mpx.c
281
if (default_mpx_mode != t->mpxmode) {
drivers/media/i2c/sony-btf-mpx.c
282
t->mpxmode = default_mpx_mode;
drivers/media/i2c/sony-btf-mpx.c
283
mpx_setup(t);
drivers/media/i2c/sony-btf-mpx.c
290
struct sony_btf_mpx *t = to_state(sd);
drivers/media/i2c/sony-btf-mpx.c
298
vt->audmode = t->audmode;
drivers/media/i2c/sony-btf-mpx.c
304
struct sony_btf_mpx *t = to_state(sd);
drivers/media/i2c/sony-btf-mpx.c
309
if (vt->audmode != t->audmode) {
drivers/media/i2c/sony-btf-mpx.c
310
t->audmode = vt->audmode;
drivers/media/i2c/sony-btf-mpx.c
311
mpx_setup(t);
drivers/media/i2c/sony-btf-mpx.c
336
struct sony_btf_mpx *t;
drivers/media/i2c/sony-btf-mpx.c
345
t = devm_kzalloc(&client->dev, sizeof(*t), GFP_KERNEL);
drivers/media/i2c/sony-btf-mpx.c
346
if (t == NULL)
drivers/media/i2c/sony-btf-mpx.c
349
sd = &t->sd;
drivers/media/i2c/sony-btf-mpx.c
353
t->mpxmode = 0;
drivers/media/i2c/sony-btf-mpx.c
354
t->audmode = V4L2_TUNER_MODE_STEREO;
drivers/media/i2c/tc358743.c
1599
static void tc358743_irq_poll_timer(struct timer_list *t)
drivers/media/i2c/tc358743.c
1601
struct tc358743_state *state = timer_container_of(state, t, timer);
drivers/media/i2c/tc358743.c
330
static inline unsigned fps(const struct v4l2_bt_timings *t)
drivers/media/i2c/tc358743.c
332
if (!V4L2_DV_BT_FRAME_HEIGHT(t) || !V4L2_DV_BT_FRAME_WIDTH(t))
drivers/media/i2c/tc358743.c
335
return DIV_ROUND_CLOSEST((unsigned)t->pixelclock,
drivers/media/i2c/tc358743.c
336
V4L2_DV_BT_FRAME_HEIGHT(t) * V4L2_DV_BT_FRAME_WIDTH(t));
drivers/media/i2c/tda7432.c
274
struct tda7432 *t = to_state(sd);
drivers/media/i2c/tda7432.c
280
if (t->balance->val < 0) {
drivers/media/i2c/tda7432.c
282
rr = rf = -t->balance->val;
drivers/media/i2c/tda7432.c
284
} else if (t->balance->val > 0) {
drivers/media/i2c/tda7432.c
287
lr = lf = t->balance->val;
drivers/media/i2c/tda7432.c
293
if (t->mute->val) {
drivers/media/i2c/tda7432.c
313
bass = t->bass->val;
drivers/media/i2c/tda7432.c
314
treble = t->treble->val;
drivers/media/i2c/tda7432.c
348
struct tda7432 *t;
drivers/media/i2c/tda7432.c
354
t = devm_kzalloc(&client->dev, sizeof(*t), GFP_KERNEL);
drivers/media/i2c/tda7432.c
355
if (!t)
drivers/media/i2c/tda7432.c
357
sd = &t->sd;
drivers/media/i2c/tda7432.c
359
v4l2_ctrl_handler_init(&t->hdl, 5);
drivers/media/i2c/tda7432.c
360
v4l2_ctrl_new_std(&t->hdl, &tda7432_ctrl_ops,
drivers/media/i2c/tda7432.c
362
t->mute = v4l2_ctrl_new_std(&t->hdl, &tda7432_ctrl_ops,
drivers/media/i2c/tda7432.c
364
t->balance = v4l2_ctrl_new_std(&t->hdl, &tda7432_ctrl_ops,
drivers/media/i2c/tda7432.c
366
t->bass = v4l2_ctrl_new_std(&t->hdl, &tda7432_ctrl_ops,
drivers/media/i2c/tda7432.c
368
t->treble = v4l2_ctrl_new_std(&t->hdl, &tda7432_ctrl_ops,
drivers/media/i2c/tda7432.c
370
sd->ctrl_handler = &t->hdl;
drivers/media/i2c/tda7432.c
371
if (t->hdl.error) {
drivers/media/i2c/tda7432.c
372
int err = t->hdl.error;
drivers/media/i2c/tda7432.c
374
v4l2_ctrl_handler_free(&t->hdl);
drivers/media/i2c/tda7432.c
377
v4l2_ctrl_cluster(2, &t->bass);
drivers/media/i2c/tda7432.c
378
v4l2_ctrl_cluster(2, &t->mute);
drivers/media/i2c/tda7432.c
379
v4l2_ctrl_handler_setup(&t->hdl);
drivers/media/i2c/tda7432.c
395
struct tda7432 *t = to_state(sd);
drivers/media/i2c/tda7432.c
399
v4l2_ctrl_handler_free(&t->hdl);
drivers/media/i2c/tda9840.c
113
static int tda9840_g_tuner(struct v4l2_subdev *sd, struct v4l2_tuner *t)
drivers/media/i2c/tda9840.c
120
t->rxsubchans = V4L2_TUNER_SUB_MONO;
drivers/media/i2c/tda9840.c
124
t->rxsubchans = V4L2_TUNER_SUB_MONO;
drivers/media/i2c/tda9840.c
127
t->rxsubchans = V4L2_TUNER_SUB_LANG1 | V4L2_TUNER_SUB_LANG2;
drivers/media/i2c/tda9840.c
130
t->rxsubchans = V4L2_TUNER_SUB_STEREO | V4L2_TUNER_SUB_MONO;
drivers/media/i2c/tda9840.c
133
t->rxsubchans = V4L2_TUNER_MODE_MONO;
drivers/media/i2c/tda9840.c
81
static int tda9840_s_tuner(struct v4l2_subdev *sd, const struct v4l2_tuner *t)
drivers/media/i2c/tda9840.c
86
if (t->index)
drivers/media/i2c/tda9840.c
93
byte = (t->audmode == V4L2_TUNER_MODE_MONO) ?
drivers/media/i2c/tda9840.c
96
switch (t->audmode) {
drivers/media/i2c/ths8200.c
61
static inline unsigned htotal(const struct v4l2_bt_timings *t)
drivers/media/i2c/ths8200.c
63
return V4L2_DV_BT_FRAME_WIDTH(t);
drivers/media/i2c/ths8200.c
66
static inline unsigned vtotal(const struct v4l2_bt_timings *t)
drivers/media/i2c/ths8200.c
68
return V4L2_DV_BT_FRAME_HEIGHT(t);
drivers/media/i2c/tvaudio.c
1451
audiocmd *t = NULL;
drivers/media/i2c/tvaudio.c
1457
t = &ta8874z_mono;
drivers/media/i2c/tvaudio.c
1460
t = &ta8874z_stereo;
drivers/media/i2c/tvaudio.c
1463
t = &ta8874z_main;
drivers/media/i2c/tvaudio.c
1466
t = &ta8874z_sub;
drivers/media/i2c/tvaudio.c
1469
t = &ta8874z_both;
drivers/media/i2c/tvaudio.c
1476
chip_cmd(chip, "TA8874Z", t);
drivers/media/i2c/tvaudio.c
319
static void chip_thread_wake(struct timer_list *t)
drivers/media/i2c/tvaudio.c
321
struct CHIPSTATE *chip = timer_container_of(chip, t, wt);
drivers/media/i2c/tvaudio.c
441
int t = chip->shadow.bytes[TDA9840_SW + 1] & ~0x7e;
drivers/media/i2c/tvaudio.c
445
t |= TDA9840_MONO;
drivers/media/i2c/tvaudio.c
448
t |= TDA9840_STEREO;
drivers/media/i2c/tvaudio.c
451
t |= TDA9840_DUALA;
drivers/media/i2c/tvaudio.c
454
t |= TDA9840_DUALB;
drivers/media/i2c/tvaudio.c
457
t |= TDA9840_DUALAB;
drivers/media/i2c/tvaudio.c
464
chip_write(chip, TDA9840_SW, t);
drivers/media/i2c/tvp7002.c
562
const struct v4l2_bt_timings *t = &tvp7002_timings[i].timings.bt;
drivers/media/i2c/tvp7002.c
564
if (!memcmp(bt, t, &bt->standards - &bt->width)) {
drivers/media/pci/bt8xx/bttv-audio-hook.c
114
t->rxsubchans = V4L2_TUNER_SUB_LANG1 | V4L2_TUNER_SUB_LANG2;
drivers/media/pci/bt8xx/bttv-audio-hook.c
115
t->audmode = V4L2_TUNER_MODE_LANG1_LANG2;
drivers/media/pci/bt8xx/bttv-audio-hook.c
118
t->rxsubchans = V4L2_TUNER_SUB_LANG2;
drivers/media/pci/bt8xx/bttv-audio-hook.c
119
t->audmode = V4L2_TUNER_MODE_LANG1_LANG2;
drivers/media/pci/bt8xx/bttv-audio-hook.c
122
t->rxsubchans = V4L2_TUNER_SUB_LANG1;
drivers/media/pci/bt8xx/bttv-audio-hook.c
123
t->audmode = V4L2_TUNER_MODE_LANG1_LANG2;
drivers/media/pci/bt8xx/bttv-audio-hook.c
126
t->rxsubchans = V4L2_TUNER_SUB_STEREO;
drivers/media/pci/bt8xx/bttv-audio-hook.c
127
t->audmode = V4L2_TUNER_MODE_STEREO;
drivers/media/pci/bt8xx/bttv-audio-hook.c
130
t->rxsubchans = V4L2_TUNER_SUB_MONO;
drivers/media/pci/bt8xx/bttv-audio-hook.c
131
t->audmode = V4L2_TUNER_MODE_MONO;
drivers/media/pci/bt8xx/bttv-audio-hook.c
134
t->rxsubchans = V4L2_TUNER_SUB_MONO |
drivers/media/pci/bt8xx/bttv-audio-hook.c
138
t->audmode = V4L2_TUNER_MODE_LANG1;
drivers/media/pci/bt8xx/bttv-audio-hook.c
157
void avermedia_tvphone_audio(struct bttv *btv, struct v4l2_tuner *t, int set)
drivers/media/pci/bt8xx/bttv-audio-hook.c
163
t->audmode = V4L2_TUNER_MODE_LANG1;
drivers/media/pci/bt8xx/bttv-audio-hook.c
164
t->rxsubchans = V4L2_TUNER_SUB_MONO |
drivers/media/pci/bt8xx/bttv-audio-hook.c
172
switch (t->audmode) {
drivers/media/pci/bt8xx/bttv-audio-hook.c
188
void avermedia_tv_stereo_audio(struct bttv *btv, struct v4l2_tuner *t, int set)
drivers/media/pci/bt8xx/bttv-audio-hook.c
194
t->audmode = V4L2_TUNER_MODE_LANG1;
drivers/media/pci/bt8xx/bttv-audio-hook.c
195
t->rxsubchans = V4L2_TUNER_SUB_MONO |
drivers/media/pci/bt8xx/bttv-audio-hook.c
203
switch (t->audmode) {
drivers/media/pci/bt8xx/bttv-audio-hook.c
221
void lt9415_audio(struct bttv *btv, struct v4l2_tuner *t, int set)
drivers/media/pci/bt8xx/bttv-audio-hook.c
226
t->audmode = V4L2_TUNER_MODE_MONO;
drivers/media/pci/bt8xx/bttv-audio-hook.c
232
t->audmode = V4L2_TUNER_MODE_LANG1;
drivers/media/pci/bt8xx/bttv-audio-hook.c
233
t->rxsubchans = V4L2_TUNER_SUB_MONO |
drivers/media/pci/bt8xx/bttv-audio-hook.c
241
switch (t->audmode) {
drivers/media/pci/bt8xx/bttv-audio-hook.c
259
void terratv_audio(struct bttv *btv, struct v4l2_tuner *t, int set)
drivers/media/pci/bt8xx/bttv-audio-hook.c
265
t->audmode = V4L2_TUNER_MODE_LANG1;
drivers/media/pci/bt8xx/bttv-audio-hook.c
266
t->rxsubchans = V4L2_TUNER_SUB_MONO |
drivers/media/pci/bt8xx/bttv-audio-hook.c
275
switch (t->audmode) {
drivers/media/pci/bt8xx/bttv-audio-hook.c
292
void winfast2000_audio(struct bttv *btv, struct v4l2_tuner *t, int set)
drivers/media/pci/bt8xx/bttv-audio-hook.c
300
switch (t->audmode) {
drivers/media/pci/bt8xx/bttv-audio-hook.c
328
void pvbt878p9b_audio(struct bttv *btv, struct v4l2_tuner *t, int set)
drivers/media/pci/bt8xx/bttv-audio-hook.c
337
t->audmode = V4L2_TUNER_MODE_LANG1;
drivers/media/pci/bt8xx/bttv-audio-hook.c
338
t->rxsubchans = V4L2_TUNER_SUB_MONO |
drivers/media/pci/bt8xx/bttv-audio-hook.c
346
switch (t->audmode) {
drivers/media/pci/bt8xx/bttv-audio-hook.c
369
void fv2000s_audio(struct bttv *btv, struct v4l2_tuner *t, int set)
drivers/media/pci/bt8xx/bttv-audio-hook.c
378
t->audmode = V4L2_TUNER_MODE_LANG1;
drivers/media/pci/bt8xx/bttv-audio-hook.c
379
t->rxsubchans = V4L2_TUNER_SUB_MONO |
drivers/media/pci/bt8xx/bttv-audio-hook.c
387
switch (t->audmode) {
drivers/media/pci/bt8xx/bttv-audio-hook.c
408
void windvr_audio(struct bttv *btv, struct v4l2_tuner *t, int set)
drivers/media/pci/bt8xx/bttv-audio-hook.c
414
t->audmode = V4L2_TUNER_MODE_LANG1;
drivers/media/pci/bt8xx/bttv-audio-hook.c
415
t->rxsubchans = V4L2_TUNER_SUB_MONO |
drivers/media/pci/bt8xx/bttv-audio-hook.c
423
switch (t->audmode) {
drivers/media/pci/bt8xx/bttv-audio-hook.c
443
void adtvk503_audio(struct bttv *btv, struct v4l2_tuner *t, int set)
drivers/media/pci/bt8xx/bttv-audio-hook.c
451
t->audmode = V4L2_TUNER_MODE_LANG1;
drivers/media/pci/bt8xx/bttv-audio-hook.c
452
t->rxsubchans = V4L2_TUNER_SUB_MONO |
drivers/media/pci/bt8xx/bttv-audio-hook.c
461
switch (t->audmode) {
drivers/media/pci/bt8xx/bttv-audio-hook.c
55
void gvbctv3pci_audio(struct bttv *btv, struct v4l2_tuner *t, int set)
drivers/media/pci/bt8xx/bttv-audio-hook.c
61
t->audmode = V4L2_TUNER_MODE_LANG1;
drivers/media/pci/bt8xx/bttv-audio-hook.c
62
t->rxsubchans = V4L2_TUNER_SUB_MONO |
drivers/media/pci/bt8xx/bttv-audio-hook.c
71
switch (t->audmode) {
drivers/media/pci/bt8xx/bttv-audio-hook.c
86
void gvbctv5pci_audio(struct bttv *btv, struct v4l2_tuner *t, int set)
drivers/media/pci/bt8xx/bttv-audio-hook.c
95
switch (t->audmode) {
drivers/media/pci/bt8xx/bttv-driver.c
1688
const struct v4l2_tuner *t)
drivers/media/pci/bt8xx/bttv-driver.c
1692
if (t->index)
drivers/media/pci/bt8xx/bttv-driver.c
1695
bttv_call_all(btv, tuner, s_tuner, t);
drivers/media/pci/bt8xx/bttv-driver.c
1698
struct v4l2_tuner copy = *t;
drivers/media/pci/bt8xx/bttv-driver.c
2177
struct v4l2_tuner *t)
drivers/media/pci/bt8xx/bttv-driver.c
2181
if (0 != t->index)
drivers/media/pci/bt8xx/bttv-driver.c
2184
t->rxsubchans = V4L2_TUNER_SUB_MONO;
drivers/media/pci/bt8xx/bttv-driver.c
2185
t->capability = V4L2_TUNER_CAP_NORM;
drivers/media/pci/bt8xx/bttv-driver.c
2186
bttv_call_all(btv, tuner, g_tuner, t);
drivers/media/pci/bt8xx/bttv-driver.c
2187
strscpy(t->name, "Television", sizeof(t->name));
drivers/media/pci/bt8xx/bttv-driver.c
2188
t->type = V4L2_TUNER_ANALOG_TV;
drivers/media/pci/bt8xx/bttv-driver.c
2190
t->signal = 0xffff;
drivers/media/pci/bt8xx/bttv-driver.c
2193
btv->audio_mode_gpio(btv, t, 0);
drivers/media/pci/bt8xx/bttv-driver.c
2401
static int radio_g_tuner(struct file *file, void *priv, struct v4l2_tuner *t)
drivers/media/pci/bt8xx/bttv-driver.c
2405
if (0 != t->index)
drivers/media/pci/bt8xx/bttv-driver.c
2407
strscpy(t->name, "Radio", sizeof(t->name));
drivers/media/pci/bt8xx/bttv-driver.c
2408
t->type = V4L2_TUNER_RADIO;
drivers/media/pci/bt8xx/bttv-driver.c
2411
bttv_call_all(btv, tuner, g_tuner, t);
drivers/media/pci/bt8xx/bttv-driver.c
2414
btv->audio_mode_gpio(btv, t, 0);
drivers/media/pci/bt8xx/bttv-driver.c
2417
return snd_tea575x_g_tuner(&btv->tea, t);
drivers/media/pci/bt8xx/bttv-driver.c
2423
const struct v4l2_tuner *t)
drivers/media/pci/bt8xx/bttv-driver.c
2427
if (0 != t->index)
drivers/media/pci/bt8xx/bttv-driver.c
2431
bttv_call_all(btv, tuner, s_tuner, t);
drivers/media/pci/bt8xx/bttv-driver.c
2799
static void bttv_irq_timeout(struct timer_list *t)
drivers/media/pci/bt8xx/bttv-driver.c
2801
struct bttv *btv = timer_container_of(btv, t, timeout);
drivers/media/pci/bt8xx/bttv-input.c
127
static void bttv_input_timer(struct timer_list *t)
drivers/media/pci/bt8xx/bttv-input.c
129
struct bttv_ir *ir = timer_container_of(ir, t, timer);
drivers/media/pci/bt8xx/bttv-input.c
183
static void bttv_rc5_timer_end(struct timer_list *t)
drivers/media/pci/bt8xx/bttv-input.c
185
struct bttv_ir *ir = timer_container_of(ir, t, timer);
drivers/media/pci/bt8xx/dvb-bt8xx.c
42
static void dvb_bt8xx_work(struct work_struct *t)
drivers/media/pci/bt8xx/dvb-bt8xx.c
44
struct bt878 *bt = from_work(bt, t, bh_work);
drivers/media/pci/cx18/cx18-fileops.c
629
void cx18_vb_timeout(struct timer_list *t)
drivers/media/pci/cx18/cx18-fileops.c
631
struct cx18_stream *s = timer_container_of(s, t, vb_timeout);
drivers/media/pci/cx18/cx18-fileops.h
24
void cx18_vb_timeout(struct timer_list *t);
drivers/media/pci/cx23885/cx23885-417.c
1264
struct v4l2_tuner *t)
drivers/media/pci/cx23885/cx23885-417.c
1270
if (0 != t->index)
drivers/media/pci/cx23885/cx23885-417.c
1272
strscpy(t->name, "Television", sizeof(t->name));
drivers/media/pci/cx23885/cx23885-417.c
1273
call_all(dev, tuner, g_tuner, t);
drivers/media/pci/cx23885/cx23885-417.c
1275
dprintk(1, "VIDIOC_G_TUNER: tuner type %d\n", t->type);
drivers/media/pci/cx23885/cx23885-417.c
1281
const struct v4l2_tuner *t)
drivers/media/pci/cx23885/cx23885-417.c
1289
call_all(dev, tuner, s_tuner, t);
drivers/media/pci/cx23885/cx23885-video.c
895
struct v4l2_tuner *t)
drivers/media/pci/cx23885/cx23885-video.c
910
if (0 != t->index)
drivers/media/pci/cx23885/cx23885-video.c
913
strscpy(t->name, "Television", sizeof(t->name));
drivers/media/pci/cx23885/cx23885-video.c
915
call_all(dev, tuner, g_tuner, t);
drivers/media/pci/cx23885/cx23885-video.c
920
const struct v4l2_tuner *t)
drivers/media/pci/cx23885/cx23885-video.c
935
if (0 != t->index)
drivers/media/pci/cx23885/cx23885-video.c
938
call_all(dev, tuner, s_tuner, t);
drivers/media/pci/cx88/cx88-blackbird.c
1004
if (t->index != 0)
drivers/media/pci/cx88/cx88-blackbird.c
1007
cx88_set_stereo(core, t->audmode, 1);
drivers/media/pci/cx88/cx88-blackbird.c
974
struct v4l2_tuner *t)
drivers/media/pci/cx88/cx88-blackbird.c
982
if (t->index != 0)
drivers/media/pci/cx88/cx88-blackbird.c
985
strscpy(t->name, "Television", sizeof(t->name));
drivers/media/pci/cx88/cx88-blackbird.c
986
t->capability = V4L2_TUNER_CAP_NORM;
drivers/media/pci/cx88/cx88-blackbird.c
987
t->rangehigh = 0xffffffffUL;
drivers/media/pci/cx88/cx88-blackbird.c
988
call_all(core, tuner, g_tuner, t);
drivers/media/pci/cx88/cx88-blackbird.c
990
cx88_get_stereo(core, t);
drivers/media/pci/cx88/cx88-blackbird.c
992
t->signal = (reg & (1 << 5)) ? 0xffff : 0x0000;
drivers/media/pci/cx88/cx88-blackbird.c
997
const struct v4l2_tuner *t)
drivers/media/pci/cx88/cx88-tvaudio.c
1011
memset(&t, 0, sizeof(t));
drivers/media/pci/cx88/cx88-tvaudio.c
1012
cx88_get_stereo(core, &t);
drivers/media/pci/cx88/cx88-tvaudio.c
1019
if (t.rxsubchans & V4L2_TUNER_SUB_STEREO)
drivers/media/pci/cx88/cx88-tvaudio.c
808
void cx88_get_stereo(struct cx88_core *core, struct v4l2_tuner *t)
drivers/media/pci/cx88/cx88-tvaudio.c
826
t->capability = V4L2_TUNER_CAP_STEREO | V4L2_TUNER_CAP_SAP |
drivers/media/pci/cx88/cx88-tvaudio.c
828
t->rxsubchans = UNSET;
drivers/media/pci/cx88/cx88-tvaudio.c
829
t->audmode = V4L2_TUNER_MODE_MONO;
drivers/media/pci/cx88/cx88-tvaudio.c
833
t->audmode = V4L2_TUNER_MODE_STEREO;
drivers/media/pci/cx88/cx88-tvaudio.c
836
t->audmode = V4L2_TUNER_MODE_LANG2;
drivers/media/pci/cx88/cx88-tvaudio.c
839
t->audmode = V4L2_TUNER_MODE_MONO;
drivers/media/pci/cx88/cx88-tvaudio.c
842
t->audmode = V4L2_TUNER_MODE_SAP;
drivers/media/pci/cx88/cx88-tvaudio.c
853
t->rxsubchans = cx88_dsp_detect_stereo_sap(core);
drivers/media/pci/cx88/cx88-tvaudio.c
868
if (t->rxsubchans == UNSET) {
drivers/media/pci/cx88/cx88-tvaudio.c
869
t->rxsubchans = V4L2_TUNER_SUB_MONO;
drivers/media/pci/cx88/cx88-tvaudio.c
874
if (t->audmode == V4L2_TUNER_MODE_STEREO)
drivers/media/pci/cx88/cx88-tvaudio.c
875
t->rxsubchans |= V4L2_TUNER_SUB_STEREO;
drivers/media/pci/cx88/cx88-tvaudio.c
990
struct v4l2_tuner t;
drivers/media/pci/cx88/cx88-video.c
1022
struct v4l2_tuner *t)
drivers/media/pci/cx88/cx88-video.c
1027
if (unlikely(t->index > 0))
drivers/media/pci/cx88/cx88-video.c
1030
strscpy(t->name, "Radio", sizeof(t->name));
drivers/media/pci/cx88/cx88-video.c
1032
call_all(core, tuner, g_tuner, t);
drivers/media/pci/cx88/cx88-video.c
1037
const struct v4l2_tuner *t)
drivers/media/pci/cx88/cx88-video.c
1042
if (t->index != 0)
drivers/media/pci/cx88/cx88-video.c
1045
call_all(core, tuner, s_tuner, t);
drivers/media/pci/cx88/cx88-video.c
906
struct v4l2_tuner *t)
drivers/media/pci/cx88/cx88-video.c
914
if (t->index != 0)
drivers/media/pci/cx88/cx88-video.c
917
strscpy(t->name, "Television", sizeof(t->name));
drivers/media/pci/cx88/cx88-video.c
918
t->capability = V4L2_TUNER_CAP_NORM;
drivers/media/pci/cx88/cx88-video.c
919
t->rangehigh = 0xffffffffUL;
drivers/media/pci/cx88/cx88-video.c
920
call_all(core, tuner, g_tuner, t);
drivers/media/pci/cx88/cx88-video.c
922
cx88_get_stereo(core, t);
drivers/media/pci/cx88/cx88-video.c
924
t->signal = (reg & (1 << 5)) ? 0xffff : 0x0000;
drivers/media/pci/cx88/cx88-video.c
929
const struct v4l2_tuner *t)
drivers/media/pci/cx88/cx88-video.c
936
if (t->index != 0)
drivers/media/pci/cx88/cx88-video.c
939
cx88_set_stereo(core, t->audmode, 1);
drivers/media/pci/cx88/cx88.h
686
void cx88_get_stereo(struct cx88_core *core, struct v4l2_tuner *t);
drivers/media/pci/ivtv/ivtv-irq.c
1065
void ivtv_unfinished_dma(struct timer_list *t)
drivers/media/pci/ivtv/ivtv-irq.c
1067
struct ivtv *itv = timer_container_of(itv, t, dma_timer);
drivers/media/pci/ivtv/ivtv-irq.h
39
void ivtv_unfinished_dma(struct timer_list *t);
drivers/media/pci/mantis/mantis_dma.c
203
void mantis_dma_xfer(struct work_struct *t)
drivers/media/pci/mantis/mantis_dma.c
205
struct mantis_pci *mantis = from_work(mantis, t, bh_work);
drivers/media/pci/mantis/mantis_dma.h
16
extern void mantis_dma_xfer(struct work_struct *t);
drivers/media/pci/netup_unidvb/netup_unidvb_core.c
638
static void netup_unidvb_dma_timeout(struct timer_list *t)
drivers/media/pci/netup_unidvb/netup_unidvb_core.c
640
struct netup_dma *dma = timer_container_of(dma, t, timeout);
drivers/media/pci/netup_unidvb/netup_unidvb_spi.c
100
struct spi_transfer *t;
drivers/media/pci/netup_unidvb/netup_unidvb_spi.c
107
list_for_each_entry(t, &msg->transfers, transfer_list) {
drivers/media/pci/netup_unidvb/netup_unidvb_spi.c
108
tr_size = t->len;
drivers/media/pci/netup_unidvb/netup_unidvb_spi.c
110
u32 frag_offset = t->len - tr_size;
drivers/media/pci/netup_unidvb/netup_unidvb_spi.c
115
if (list_is_last(&t->transfer_list,
drivers/media/pci/netup_unidvb/netup_unidvb_spi.c
117
frag_offset + frag_size == t->len) {
drivers/media/pci/netup_unidvb/netup_unidvb_spi.c
120
if (t->tx_buf) {
drivers/media/pci/netup_unidvb/netup_unidvb_spi.c
122
t->tx_buf + frag_offset,
drivers/media/pci/netup_unidvb/netup_unidvb_spi.c
141
if (t->rx_buf) {
drivers/media/pci/netup_unidvb/netup_unidvb_spi.c
142
memcpy_fromio(t->rx_buf + frag_offset,
drivers/media/pci/ngene/ngene-core.c
53
static void event_bh_work(struct work_struct *t)
drivers/media/pci/ngene/ngene-core.c
55
struct ngene *dev = from_work(dev, t, event_bh_work);
drivers/media/pci/ngene/ngene-core.c
71
static void demux_bh_work(struct work_struct *t)
drivers/media/pci/ngene/ngene-core.c
73
struct ngene_channel *chan = from_work(chan, t, demux_bh_work);
drivers/media/pci/saa7134/saa7134-core.c
329
void saa7134_buffer_timeout(struct timer_list *t)
drivers/media/pci/saa7134/saa7134-core.c
331
struct saa7134_dmaqueue *q = timer_container_of(q, t, timeout);
drivers/media/pci/saa7134/saa7134-input.c
432
static void saa7134_input_timer(struct timer_list *t)
drivers/media/pci/saa7134/saa7134-input.c
434
struct saa7134_card_ir *ir = timer_container_of(ir, t, timer);
drivers/media/pci/saa7134/saa7134-reg.h
210
#define SAA7134_TASK_CONDITIONS(t) (0x000 +t)
drivers/media/pci/saa7134/saa7134-reg.h
211
#define SAA7134_FIELD_HANDLING(t) (0x001 +t)
drivers/media/pci/saa7134/saa7134-reg.h
212
#define SAA7134_DATA_PATH(t) (0x002 +t)
drivers/media/pci/saa7134/saa7134-reg.h
213
#define SAA7134_VBI_H_START1(t) (0x004 +t)
drivers/media/pci/saa7134/saa7134-reg.h
214
#define SAA7134_VBI_H_START2(t) (0x005 +t)
drivers/media/pci/saa7134/saa7134-reg.h
215
#define SAA7134_VBI_H_STOP1(t) (0x006 +t)
drivers/media/pci/saa7134/saa7134-reg.h
216
#define SAA7134_VBI_H_STOP2(t) (0x007 +t)
drivers/media/pci/saa7134/saa7134-reg.h
217
#define SAA7134_VBI_V_START1(t) (0x008 +t)
drivers/media/pci/saa7134/saa7134-reg.h
218
#define SAA7134_VBI_V_START2(t) (0x009 +t)
drivers/media/pci/saa7134/saa7134-reg.h
219
#define SAA7134_VBI_V_STOP1(t) (0x00a +t)
drivers/media/pci/saa7134/saa7134-reg.h
220
#define SAA7134_VBI_V_STOP2(t) (0x00b +t)
drivers/media/pci/saa7134/saa7134-reg.h
221
#define SAA7134_VBI_H_LEN1(t) (0x00c +t)
drivers/media/pci/saa7134/saa7134-reg.h
222
#define SAA7134_VBI_H_LEN2(t) (0x00d +t)
drivers/media/pci/saa7134/saa7134-reg.h
223
#define SAA7134_VBI_V_LEN1(t) (0x00e +t)
drivers/media/pci/saa7134/saa7134-reg.h
224
#define SAA7134_VBI_V_LEN2(t) (0x00f +t)
drivers/media/pci/saa7134/saa7134-reg.h
226
#define SAA7134_VIDEO_H_START1(t) (0x014 +t)
drivers/media/pci/saa7134/saa7134-reg.h
227
#define SAA7134_VIDEO_H_START2(t) (0x015 +t)
drivers/media/pci/saa7134/saa7134-reg.h
228
#define SAA7134_VIDEO_H_STOP1(t) (0x016 +t)
drivers/media/pci/saa7134/saa7134-reg.h
229
#define SAA7134_VIDEO_H_STOP2(t) (0x017 +t)
drivers/media/pci/saa7134/saa7134-reg.h
230
#define SAA7134_VIDEO_V_START1(t) (0x018 +t)
drivers/media/pci/saa7134/saa7134-reg.h
231
#define SAA7134_VIDEO_V_START2(t) (0x019 +t)
drivers/media/pci/saa7134/saa7134-reg.h
232
#define SAA7134_VIDEO_V_STOP1(t) (0x01a +t)
drivers/media/pci/saa7134/saa7134-reg.h
233
#define SAA7134_VIDEO_V_STOP2(t) (0x01b +t)
drivers/media/pci/saa7134/saa7134-reg.h
234
#define SAA7134_VIDEO_PIXELS1(t) (0x01c +t)
drivers/media/pci/saa7134/saa7134-reg.h
235
#define SAA7134_VIDEO_PIXELS2(t) (0x01d +t)
drivers/media/pci/saa7134/saa7134-reg.h
236
#define SAA7134_VIDEO_LINES1(t) (0x01e +t)
drivers/media/pci/saa7134/saa7134-reg.h
237
#define SAA7134_VIDEO_LINES2(t) (0x01f +t)
drivers/media/pci/saa7134/saa7134-reg.h
239
#define SAA7134_H_PRESCALE(t) (0x020 +t)
drivers/media/pci/saa7134/saa7134-reg.h
240
#define SAA7134_ACC_LENGTH(t) (0x021 +t)
drivers/media/pci/saa7134/saa7134-reg.h
241
#define SAA7134_LEVEL_CTRL(t) (0x022 +t)
drivers/media/pci/saa7134/saa7134-reg.h
242
#define SAA7134_FIR_PREFILTER_CTRL(t) (0x023 +t)
drivers/media/pci/saa7134/saa7134-reg.h
243
#define SAA7134_LUMA_BRIGHT(t) (0x024 +t)
drivers/media/pci/saa7134/saa7134-reg.h
244
#define SAA7134_LUMA_CONTRAST(t) (0x025 +t)
drivers/media/pci/saa7134/saa7134-reg.h
245
#define SAA7134_CHROMA_SATURATION(t) (0x026 +t)
drivers/media/pci/saa7134/saa7134-reg.h
246
#define SAA7134_VBI_H_SCALE_INC1(t) (0x028 +t)
drivers/media/pci/saa7134/saa7134-reg.h
247
#define SAA7134_VBI_H_SCALE_INC2(t) (0x029 +t)
drivers/media/pci/saa7134/saa7134-reg.h
248
#define SAA7134_VBI_PHASE_OFFSET_LUMA(t) (0x02a +t)
drivers/media/pci/saa7134/saa7134-reg.h
249
#define SAA7134_VBI_PHASE_OFFSET_CHROMA(t) (0x02b +t)
drivers/media/pci/saa7134/saa7134-reg.h
250
#define SAA7134_H_SCALE_INC1(t) (0x02c +t)
drivers/media/pci/saa7134/saa7134-reg.h
251
#define SAA7134_H_SCALE_INC2(t) (0x02d +t)
drivers/media/pci/saa7134/saa7134-reg.h
252
#define SAA7134_H_PHASE_OFF_LUMA(t) (0x02e +t)
drivers/media/pci/saa7134/saa7134-reg.h
253
#define SAA7134_H_PHASE_OFF_CHROMA(t) (0x02f +t)
drivers/media/pci/saa7134/saa7134-reg.h
254
#define SAA7134_V_SCALE_RATIO1(t) (0x030 +t)
drivers/media/pci/saa7134/saa7134-reg.h
255
#define SAA7134_V_SCALE_RATIO2(t) (0x031 +t)
drivers/media/pci/saa7134/saa7134-reg.h
256
#define SAA7134_V_FILTER(t) (0x032 +t)
drivers/media/pci/saa7134/saa7134-reg.h
257
#define SAA7134_V_PHASE_OFFSET0(t) (0x034 +t)
drivers/media/pci/saa7134/saa7134-reg.h
258
#define SAA7134_V_PHASE_OFFSET1(t) (0x035 +t)
drivers/media/pci/saa7134/saa7134-reg.h
259
#define SAA7134_V_PHASE_OFFSET2(t) (0x036 +t)
drivers/media/pci/saa7134/saa7134-reg.h
260
#define SAA7134_V_PHASE_OFFSET3(t) (0x037 +t)
drivers/media/pci/saa7134/saa7134-video.c
1362
struct v4l2_tuner *t)
drivers/media/pci/saa7134/saa7134-video.c
1367
if (0 != t->index)
drivers/media/pci/saa7134/saa7134-video.c
1369
memset(t, 0, sizeof(*t));
drivers/media/pci/saa7134/saa7134-video.c
1378
strscpy(t->name, "Television", sizeof(t->name));
drivers/media/pci/saa7134/saa7134-video.c
1379
t->type = V4L2_TUNER_ANALOG_TV;
drivers/media/pci/saa7134/saa7134-video.c
1380
saa_call_all(dev, tuner, g_tuner, t);
drivers/media/pci/saa7134/saa7134-video.c
1381
t->capability = V4L2_TUNER_CAP_NORM |
drivers/media/pci/saa7134/saa7134-video.c
1385
t->rxsubchans = saa7134_tvaudio_getstereo(dev);
drivers/media/pci/saa7134/saa7134-video.c
1386
t->audmode = saa7134_tvaudio_rx2mode(t->rxsubchans);
drivers/media/pci/saa7134/saa7134-video.c
1389
t->signal = 0xffff;
drivers/media/pci/saa7134/saa7134-video.c
1395
const struct v4l2_tuner *t)
drivers/media/pci/saa7134/saa7134-video.c
1400
if (0 != t->index)
drivers/media/pci/saa7134/saa7134-video.c
1408
if (mode != t->audmode)
drivers/media/pci/saa7134/saa7134-video.c
1409
dev->thread.mode = t->audmode;
drivers/media/pci/saa7134/saa7134-video.c
1477
struct v4l2_tuner *t)
drivers/media/pci/saa7134/saa7134-video.c
1481
if (0 != t->index)
drivers/media/pci/saa7134/saa7134-video.c
1484
strscpy(t->name, "Radio", sizeof(t->name));
drivers/media/pci/saa7134/saa7134-video.c
1486
saa_call_all(dev, tuner, g_tuner, t);
drivers/media/pci/saa7134/saa7134-video.c
1487
t->audmode &= V4L2_TUNER_MODE_MONO | V4L2_TUNER_MODE_STEREO;
drivers/media/pci/saa7134/saa7134-video.c
1489
t->signal = 0xf800 - ((saa_readb(0x581) & 0x1f) << 11);
drivers/media/pci/saa7134/saa7134-video.c
1490
t->rxsubchans = (saa_readb(0x529) & 0x08) ?
drivers/media/pci/saa7134/saa7134-video.c
1496
const struct v4l2_tuner *t)
drivers/media/pci/saa7134/saa7134-video.c
1500
if (0 != t->index)
drivers/media/pci/saa7134/saa7134-video.c
1503
saa_call_all(dev, tuner, s_tuner, t);
drivers/media/pci/saa7134/saa7134.h
758
void saa7134_buffer_timeout(struct timer_list *t);
drivers/media/pci/saa7134/saa7134.h
807
struct v4l2_tuner *t);
drivers/media/pci/saa7134/saa7134.h
809
const struct v4l2_tuner *t);
drivers/media/pci/saa7146/mxb.c
176
struct v4l2_tuner t = {
drivers/media/pci/saa7146/mxb.c
180
tda9840_call(mxb, tuner, s_tuner, &t);
drivers/media/pci/saa7146/mxb.c
533
static int vidioc_g_tuner(struct file *file, void *fh, struct v4l2_tuner *t)
drivers/media/pci/saa7146/mxb.c
538
if (t->index) {
drivers/media/pci/saa7146/mxb.c
540
t->index);
drivers/media/pci/saa7146/mxb.c
544
DEB_EE("VIDIOC_G_TUNER: %d\n", t->index);
drivers/media/pci/saa7146/mxb.c
546
memset(t, 0, sizeof(*t));
drivers/media/pci/saa7146/mxb.c
547
strscpy(t->name, "TV Tuner", sizeof(t->name));
drivers/media/pci/saa7146/mxb.c
548
t->type = V4L2_TUNER_ANALOG_TV;
drivers/media/pci/saa7146/mxb.c
549
t->capability = V4L2_TUNER_CAP_NORM | V4L2_TUNER_CAP_STEREO |
drivers/media/pci/saa7146/mxb.c
551
t->audmode = mxb->cur_mode;
drivers/media/pci/saa7146/mxb.c
552
return call_all(dev, tuner, g_tuner, t);
drivers/media/pci/saa7146/mxb.c
555
static int vidioc_s_tuner(struct file *file, void *fh, const struct v4l2_tuner *t)
drivers/media/pci/saa7146/mxb.c
560
if (t->index) {
drivers/media/pci/saa7146/mxb.c
562
t->index);
drivers/media/pci/saa7146/mxb.c
566
mxb->cur_mode = t->audmode;
drivers/media/pci/saa7146/mxb.c
567
return call_all(dev, tuner, s_tuner, t);
drivers/media/pci/saa7164/saa7164-api.c
1062
t = (struct tmComResDescrHeader *)
drivers/media/pci/saa7164/saa7164-api.c
1064
switch (t->subtype) {
drivers/media/pci/saa7164/saa7164-api.c
1067
(struct tmComResTSFormatDescrHeader *)t;
drivers/media/pci/saa7164/saa7164-api.c
1079
(struct tmComResPSFormatDescrHeader *)t;
drivers/media/pci/saa7164/saa7164-api.c
1091
(struct tmComResVBIFormatDescrHeader *)t;
drivers/media/pci/saa7164/saa7164-api.c
1118
t->subtype);
drivers/media/pci/saa7164/saa7164-api.c
1120
next_offset += t->len;
drivers/media/pci/saa7164/saa7164-api.c
1481
struct tmComResGPIO t;
drivers/media/pci/saa7164/saa7164-api.c
1489
t.pin = pin;
drivers/media/pci/saa7164/saa7164-api.c
1490
t.state = state;
drivers/media/pci/saa7164/saa7164-api.c
1493
EXU_GPIO_CONTROL, sizeof(t), &t);
drivers/media/pci/saa7164/saa7164-api.c
922
struct tmComResDescrHeader *hdr, *t;
drivers/media/pci/saa7164/saa7164-encoder.c
309
int saa7164_g_tuner(struct file *file, void *priv, struct v4l2_tuner *t)
drivers/media/pci/saa7164/saa7164-encoder.c
315
if (0 != t->index)
drivers/media/pci/saa7164/saa7164-encoder.c
318
strscpy(t->name, "tuner", sizeof(t->name));
drivers/media/pci/saa7164/saa7164-encoder.c
319
t->capability = V4L2_TUNER_CAP_NORM | V4L2_TUNER_CAP_STEREO;
drivers/media/pci/saa7164/saa7164-encoder.c
320
t->rangelow = SAA7164_TV_MIN_FREQ;
drivers/media/pci/saa7164/saa7164-encoder.c
321
t->rangehigh = SAA7164_TV_MAX_FREQ;
drivers/media/pci/saa7164/saa7164-encoder.c
323
dprintk(DBGLVL_ENC, "VIDIOC_G_TUNER: tuner type %d\n", t->type);
drivers/media/pci/saa7164/saa7164-encoder.c
329
const struct v4l2_tuner *t)
drivers/media/pci/saa7164/saa7164-encoder.c
331
if (0 != t->index)
drivers/media/pci/saa7164/saa7164.h
596
int saa7164_g_tuner(struct file *file, void *priv, struct v4l2_tuner *t);
drivers/media/pci/saa7164/saa7164.h
597
int saa7164_s_tuner(struct file *file, void *priv, const struct v4l2_tuner *t);
drivers/media/pci/smipcie/smipcie-main.c
283
static void smi_dma_xfer(struct work_struct *t)
drivers/media/pci/smipcie/smipcie-main.c
285
struct smi_port *port = from_work(port, t, bh_work);
drivers/media/pci/solo6x10/solo6x10-v4l2-enc.c
1041
struct v4l2_fract *t = &sp->parm.capture.timeperframe;
drivers/media/pci/solo6x10/solo6x10-v4l2-enc.c
1047
solo_enc->interval = calc_interval(fps, t->numerator, t->denominator);
drivers/media/pci/ttpci/budget-ci.c
104
static void msp430_ir_interrupt(struct work_struct *t)
drivers/media/pci/ttpci/budget-ci.c
106
struct budget_ci_ir *ir = from_work(ir, t, msp430_irq_bh_work);
drivers/media/pci/ttpci/budget-ci.c
353
static void ciintf_interrupt(struct work_struct *t)
drivers/media/pci/ttpci/budget-ci.c
355
struct budget_ci *budget_ci = from_work(budget_ci, t,
drivers/media/pci/ttpci/budget-core.c
175
static void vpeirq(struct work_struct *t)
drivers/media/pci/ttpci/budget-core.c
177
struct budget *budget = from_work(budget, t, vpe_bh_work);
drivers/media/pci/tw5864/tw5864-video.c
1317
static void tw5864_handle_frame_work(struct work_struct *t)
drivers/media/pci/tw5864/tw5864-video.c
1319
struct tw5864_dev *dev = from_work(dev, t, bh_work);
drivers/media/pci/tw5864/tw5864-video.c
179
static void tw5864_handle_frame_work(struct work_struct *t);
drivers/media/pci/tw5864/tw5864-video.c
803
struct v4l2_fract *t = &sp->parm.capture.timeperframe;
drivers/media/pci/tw5864/tw5864-video.c
811
if (!t->numerator || !t->denominator) {
drivers/media/pci/tw5864/tw5864-video.c
812
t->numerator = time_base.numerator * input->frame_interval;
drivers/media/pci/tw5864/tw5864-video.c
813
t->denominator = time_base.denominator;
drivers/media/pci/tw5864/tw5864-video.c
814
} else if (t->denominator != time_base.denominator) {
drivers/media/pci/tw5864/tw5864-video.c
815
t->numerator = t->numerator * time_base.denominator /
drivers/media/pci/tw5864/tw5864-video.c
816
t->denominator;
drivers/media/pci/tw5864/tw5864-video.c
817
t->denominator = time_base.denominator;
drivers/media/pci/tw5864/tw5864-video.c
820
input->frame_interval = t->numerator / time_base.numerator;
drivers/media/pci/tw686x/tw686x-core.c
126
static void tw686x_dma_delay(struct timer_list *t)
drivers/media/pci/tw686x/tw686x-core.c
128
struct tw686x_dev *dev = timer_container_of(dev, t, dma_delay_timer);
drivers/media/platform/arm/mali-c55/mali-c55-registers.h
413
#define MALI_C55_SCALER_TIMEOUT(t) ((t) << 16)
drivers/media/platform/aspeed/aspeed-video.c
574
u32 t = readl(video->base + reg);
drivers/media/platform/aspeed/aspeed-video.c
575
u32 before = t;
drivers/media/platform/aspeed/aspeed-video.c
577
t &= ~clear;
drivers/media/platform/aspeed/aspeed-video.c
578
t |= bits;
drivers/media/platform/aspeed/aspeed-video.c
579
writel(t, video->base + reg);
drivers/media/platform/aspeed/aspeed-video.c
586
u32 t = readl(video->base + reg);
drivers/media/platform/aspeed/aspeed-video.c
588
v4l2_dbg(3, debug, &video->v4l2_dev, "read %03x[%08x]\n", reg, t);
drivers/media/platform/aspeed/aspeed-video.c
589
return t;
drivers/media/platform/chips-media/coda/coda-jpeg.c
751
int t = (j == 0) ? AC_TABLE_INDEX0 :
drivers/media/platform/chips-media/coda/coda-jpeg.c
759
if (huff->size[t][i] == 0 && huff->code[t][i] == 0)
drivers/media/platform/chips-media/coda/coda-jpeg.c
763
((huff->size[t][i] - 1) << 16) |
drivers/media/platform/chips-media/coda/coda-jpeg.c
764
huff->code[t][i];
drivers/media/platform/intel/pxa_camera.c
1150
static void pxa_camera_eof_bh_work(struct work_struct *t)
drivers/media/platform/intel/pxa_camera.c
1152
struct pxa_camera_dev *pcdev = from_work(pcdev, t, eof_bh_work);
drivers/media/platform/marvell/mcam-core.c
442
static void mcam_frame_work(struct work_struct *t)
drivers/media/platform/marvell/mcam-core.c
444
struct mcam_camera *cam = from_work(cam, t, s_bh_work);
drivers/media/platform/mediatek/mdp3/mtk-mdp3-vpu.c
169
unsigned int t = MDP_VPU_MESSAGE_TIMEOUT;
drivers/media/platform/mediatek/mdp3/mtk-mdp3-vpu.c
183
msecs_to_jiffies(t));
drivers/media/platform/qcom/camss/camss-vfe.c
784
struct camss_buffer *t;
drivers/media/platform/qcom/camss/camss-vfe.c
786
list_for_each_entry_safe(buf, t, &output->pending_bufs, queue) {
drivers/media/platform/renesas/rcar_jpu.c
113
#define JCQTN_SHIFT(t) (((t) - 1) << 1)
drivers/media/platform/renesas/rcar_jpu.c
117
#define JCHTN_AC_SHIFT(t) (((t) << 1) - 1)
drivers/media/platform/renesas/rcar_jpu.c
118
#define JCHTN_DC_SHIFT(t) (((t) - 1) << 1)
drivers/media/platform/samsung/s5p-jpeg/jpeg-hw-exynos3250.c
182
void exynos3250_jpeg_qtbl(void __iomem *regs, unsigned int t, unsigned int n)
drivers/media/platform/samsung/s5p-jpeg/jpeg-hw-exynos3250.c
187
reg &= ~EXYNOS3250_QT_NUM_MASK(t);
drivers/media/platform/samsung/s5p-jpeg/jpeg-hw-exynos3250.c
188
reg |= (n << EXYNOS3250_QT_NUM_SHIFT(t)) &
drivers/media/platform/samsung/s5p-jpeg/jpeg-hw-exynos3250.c
189
EXYNOS3250_QT_NUM_MASK(t);
drivers/media/platform/samsung/s5p-jpeg/jpeg-hw-exynos3250.c
193
void exynos3250_jpeg_htbl_ac(void __iomem *regs, unsigned int t)
drivers/media/platform/samsung/s5p-jpeg/jpeg-hw-exynos3250.c
198
reg &= ~EXYNOS3250_HT_NUM_AC_MASK(t);
drivers/media/platform/samsung/s5p-jpeg/jpeg-hw-exynos3250.c
200
reg |= (0 << EXYNOS3250_HT_NUM_AC_SHIFT(t)) &
drivers/media/platform/samsung/s5p-jpeg/jpeg-hw-exynos3250.c
201
EXYNOS3250_HT_NUM_AC_MASK(t);
drivers/media/platform/samsung/s5p-jpeg/jpeg-hw-exynos3250.c
205
void exynos3250_jpeg_htbl_dc(void __iomem *regs, unsigned int t)
drivers/media/platform/samsung/s5p-jpeg/jpeg-hw-exynos3250.c
210
reg &= ~EXYNOS3250_HT_NUM_DC_MASK(t);
drivers/media/platform/samsung/s5p-jpeg/jpeg-hw-exynos3250.c
212
reg |= (0 << EXYNOS3250_HT_NUM_DC_SHIFT(t)) &
drivers/media/platform/samsung/s5p-jpeg/jpeg-hw-exynos3250.c
213
EXYNOS3250_HT_NUM_DC_MASK(t);
drivers/media/platform/samsung/s5p-jpeg/jpeg-hw-exynos3250.h
28
void exynos3250_jpeg_qtbl(void __iomem *regs, unsigned int t, unsigned int n);
drivers/media/platform/samsung/s5p-jpeg/jpeg-hw-exynos3250.h
29
void exynos3250_jpeg_htbl_ac(void __iomem *regs, unsigned int t);
drivers/media/platform/samsung/s5p-jpeg/jpeg-hw-exynos3250.h
30
void exynos3250_jpeg_htbl_dc(void __iomem *regs, unsigned int t);
drivers/media/platform/samsung/s5p-jpeg/jpeg-hw-s5p.c
100
void s5p_jpeg_qtbl(void __iomem *regs, unsigned int t, unsigned int n)
drivers/media/platform/samsung/s5p-jpeg/jpeg-hw-s5p.c
105
reg &= ~S5P_QT_NUMt_MASK(t);
drivers/media/platform/samsung/s5p-jpeg/jpeg-hw-s5p.c
106
reg |= (n << S5P_QT_NUMt_SHIFT(t)) & S5P_QT_NUMt_MASK(t);
drivers/media/platform/samsung/s5p-jpeg/jpeg-hw-s5p.c
110
void s5p_jpeg_htbl_ac(void __iomem *regs, unsigned int t)
drivers/media/platform/samsung/s5p-jpeg/jpeg-hw-s5p.c
115
reg &= ~S5P_HT_NUMt_AC_MASK(t);
drivers/media/platform/samsung/s5p-jpeg/jpeg-hw-s5p.c
117
reg |= (0 << S5P_HT_NUMt_AC_SHIFT(t)) & S5P_HT_NUMt_AC_MASK(t);
drivers/media/platform/samsung/s5p-jpeg/jpeg-hw-s5p.c
121
void s5p_jpeg_htbl_dc(void __iomem *regs, unsigned int t)
drivers/media/platform/samsung/s5p-jpeg/jpeg-hw-s5p.c
126
reg &= ~S5P_HT_NUMt_DC_MASK(t);
drivers/media/platform/samsung/s5p-jpeg/jpeg-hw-s5p.c
128
reg |= (0 << S5P_HT_NUMt_DC_SHIFT(t)) & S5P_HT_NUMt_DC_MASK(t);
drivers/media/platform/samsung/s5p-jpeg/jpeg-hw-s5p.h
33
void s5p_jpeg_qtbl(void __iomem *regs, unsigned int t, unsigned int n);
drivers/media/platform/samsung/s5p-jpeg/jpeg-hw-s5p.h
34
void s5p_jpeg_htbl_ac(void __iomem *regs, unsigned int t);
drivers/media/platform/samsung/s5p-jpeg/jpeg-hw-s5p.h
35
void s5p_jpeg_htbl_dc(void __iomem *regs, unsigned int t);
drivers/media/platform/samsung/s5p-jpeg/jpeg-regs.h
34
#define S5P_QT_NUMt_SHIFT(t) (((t) - 1) << 1)
drivers/media/platform/samsung/s5p-jpeg/jpeg-regs.h
35
#define S5P_QT_NUMt_MASK(t) (0x3 << S5P_QT_NUMt_SHIFT(t))
drivers/media/platform/samsung/s5p-jpeg/jpeg-regs.h
39
#define S5P_HT_NUMt_AC_SHIFT(t) (((t) << 1) - 1)
drivers/media/platform/samsung/s5p-jpeg/jpeg-regs.h
40
#define S5P_HT_NUMt_AC_MASK(t) (0x1 << S5P_HT_NUMt_AC_SHIFT(t))
drivers/media/platform/samsung/s5p-jpeg/jpeg-regs.h
42
#define S5P_HT_NUMt_DC_SHIFT(t) (((t) - 1) << 1)
drivers/media/platform/samsung/s5p-jpeg/jpeg-regs.h
420
#define EXYNOS3250_QT_NUM_SHIFT(t) ((((t) - 1) << 1) + 8)
drivers/media/platform/samsung/s5p-jpeg/jpeg-regs.h
421
#define EXYNOS3250_QT_NUM_MASK(t) (0x3 << EXYNOS3250_QT_NUM_SHIFT(t))
drivers/media/platform/samsung/s5p-jpeg/jpeg-regs.h
424
#define EXYNOS3250_HT_NUM_AC_SHIFT(t) (((t) << 1) - 1)
drivers/media/platform/samsung/s5p-jpeg/jpeg-regs.h
425
#define EXYNOS3250_HT_NUM_AC_MASK(t) (0x1 << EXYNOS3250_HT_NUM_AC_SHIFT(t))
drivers/media/platform/samsung/s5p-jpeg/jpeg-regs.h
427
#define EXYNOS3250_HT_NUM_DC_SHIFT(t) (((t) - 1) << 1)
drivers/media/platform/samsung/s5p-jpeg/jpeg-regs.h
428
#define EXYNOS3250_HT_NUM_DC_MASK(t) (0x1 << EXYNOS3250_HT_NUM_DC_SHIFT(t))
drivers/media/platform/samsung/s5p-jpeg/jpeg-regs.h
43
#define S5P_HT_NUMt_DC_MASK(t) (0x1 << S5P_HT_NUMt_DC_SHIFT(t))
drivers/media/platform/samsung/s5p-mfc/s5p_mfc.c
144
static void s5p_mfc_watchdog(struct timer_list *t)
drivers/media/platform/samsung/s5p-mfc/s5p_mfc.c
146
struct s5p_mfc_dev *dev = timer_container_of(dev, t, watchdog_timer);
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_dec.c
180
static const struct s5p_mfc_fmt *find_format(struct v4l2_format *f, unsigned int t)
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_dec.c
186
formats[i].type == t)
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_enc.c
114
static const struct s5p_mfc_fmt *find_format(struct v4l2_format *f, unsigned int t)
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_enc.c
120
formats[i].type == t)
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_enc.c
1725
static unsigned int t[V4L2_MPEG_VIDEO_H264_LEVEL_4_0 + 1] = {
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_enc.c
1739
return t[lvl];
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_enc.c
1744
static unsigned int t[V4L2_MPEG_VIDEO_MPEG4_LEVEL_5 + 1] = {
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_enc.c
1754
return t[lvl];
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_enc.c
1759
static unsigned int t[] = {
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_enc.c
1774
return t[lvl];
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_enc.c
1779
static unsigned int t[V4L2_MPEG_VIDEO_H264_VUI_SAR_IDC_EXTENDED + 1] = {
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_enc.c
1799
return t[sar];
drivers/media/platform/sunxi/sun6i-mipi-csi2/sun6i_mipi_csi2_reg.h
24
#define SUN6I_MIPI_CSI2_VCDT_RX_CH_DT(ch, t) (((t) & GENMASK(5, 0)) << \
drivers/media/platform/sunxi/sun8i-a83t-mipi-csi2/sun8i_a83t_mipi_csi2_reg.h
143
#define SUN8I_A83T_MIPI_CSI2_VCDT0_CH_DT(ch, t) (((t) & GENMASK(5, 0)) << \
drivers/media/platform/sunxi/sun8i-a83t-mipi-csi2/sun8i_a83t_mipi_csi2_reg.h
148
#define SUN8I_A83T_MIPI_CSI2_VCDT1_CH_DT(ch, t) (((t) & GENMASK(5, 0)) << \
drivers/media/platform/ti/omap/omap_vout_vrfb.c
67
struct vid_vrfb_dma *t = (struct vid_vrfb_dma *) data;
drivers/media/platform/ti/omap/omap_vout_vrfb.c
69
t->tx_status = 1;
drivers/media/platform/ti/omap/omap_vout_vrfb.c
70
wake_up_interruptible(&t->wait);
drivers/media/radio/radio-cadet.c
285
static void cadet_handler(struct timer_list *t)
drivers/media/radio/radio-cadet.c
287
struct cadet *dev = timer_container_of(dev, t, readtimer);
drivers/media/rc/ene_ir.c
660
static void ene_tx_irqsim(struct timer_list *t)
drivers/media/rc/ene_ir.c
662
struct ene_device *dev = timer_container_of(dev, t, tx_sim_timer);
drivers/media/rc/igorplugusb.c
132
static void igorplugusb_timer(struct timer_list *t)
drivers/media/rc/igorplugusb.c
134
struct igorplugusb *ir = timer_container_of(ir, t, timer);
drivers/media/rc/img-ir/img-ir-hw.c
866
static void img_ir_end_timer(struct timer_list *t)
drivers/media/rc/img-ir/img-ir-hw.c
868
struct img_ir_priv *priv = timer_container_of(priv, t, hw.end_timer);
drivers/media/rc/img-ir/img-ir-hw.c
880
static void img_ir_suspend_timer(struct timer_list *t)
drivers/media/rc/img-ir/img-ir-hw.c
882
struct img_ir_priv *priv = timer_container_of(priv, t,
drivers/media/rc/img-ir/img-ir-raw.c
66
static void img_ir_echo_timer(struct timer_list *t)
drivers/media/rc/img-ir/img-ir-raw.c
68
struct img_ir_priv *priv = timer_container_of(priv, t, raw.timer);
drivers/media/rc/imon.c
1101
static void imon_touch_display_timeout(struct timer_list *t)
drivers/media/rc/imon.c
1103
struct imon_context *ictx = timer_container_of(ictx, t, ttimer);
drivers/media/rc/imon.c
1590
ktime_t t;
drivers/media/rc/imon.c
1690
t = ktime_get();
drivers/media/rc/imon.c
1695
ktime_ms_delta(t, prev_time) < ictx->idev->rep[REP_DELAY]) {
drivers/media/rc/imon.c
1701
prev_time = t;
drivers/media/rc/ir-mce_kbd-decoder.c
110
static void mce_kbd_rx_timeout(struct timer_list *t)
drivers/media/rc/ir-mce_kbd-decoder.c
112
struct ir_raw_event_ctrl *raw = timer_container_of(raw, t,
drivers/media/rc/rc-ir-raw.c
553
static void ir_raw_edge_handle(struct timer_list *t)
drivers/media/rc/rc-ir-raw.c
555
struct ir_raw_event_ctrl *raw = timer_container_of(raw, t,
drivers/media/rc/rc-main.c
675
static void ir_timer_keyup(struct timer_list *t)
drivers/media/rc/rc-main.c
677
struct rc_dev *dev = timer_container_of(dev, t, timer_keyup);
drivers/media/rc/rc-main.c
704
static void ir_timer_repeat(struct timer_list *t)
drivers/media/rc/rc-main.c
706
struct rc_dev *dev = timer_container_of(dev, t, timer_repeat);
drivers/media/test-drivers/vidtv/vidtv_psi.c
1571
struct vidtv_psi_table_transport *t = nit->transport;
drivers/media/test-drivers/vidtv/vidtv_psi.c
1588
while (t) {
drivers/media/test-drivers/vidtv/vidtv_psi.c
1596
desc_loop_len = vidtv_psi_desc_comp_loop_len(t->descriptor);
drivers/media/test-drivers/vidtv/vidtv_psi.c
1597
vidtv_psi_set_desc_loop_len(&t->bitfield, desc_loop_len, 12);
drivers/media/test-drivers/vidtv/vidtv_psi.c
1601
t = t->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
1781
static void vidtv_psi_transport_destroy(struct vidtv_psi_table_transport *t)
drivers/media/test-drivers/vidtv/vidtv_psi.c
1784
struct vidtv_psi_table_transport *curr_t = t;
drivers/media/test-drivers/visl/visl-trace-av1.h
299
TP_PROTO(const struct v4l2_ctrl_av1_tile_group_entry *t),
drivers/media/test-drivers/visl/visl-trace-av1.h
300
TP_ARGS(t)
drivers/media/test-drivers/visl/visl-trace-av1.h
48
TP_PROTO(const struct v4l2_ctrl_av1_tile_group_entry *t),
drivers/media/test-drivers/visl/visl-trace-av1.h
49
TP_ARGS(t),
drivers/media/test-drivers/visl/visl-trace-av1.h
50
TP_STRUCT__entry(__field_struct(struct v4l2_ctrl_av1_tile_group_entry, t)),
drivers/media/test-drivers/visl/visl-trace-av1.h
51
TP_fast_assign(__entry->t = *t;),
drivers/media/test-drivers/visl/visl-trace-av1.h
53
__entry->t.tile_offset,
drivers/media/test-drivers/visl/visl-trace-av1.h
54
__entry->t.tile_size,
drivers/media/test-drivers/visl/visl-trace-av1.h
55
__entry->t.tile_row,
drivers/media/test-drivers/visl/visl-trace-av1.h
56
__entry->t.tile_col
drivers/media/tuners/max2165.c
181
u32 t;
drivers/media/tuners/max2165.c
204
t = priv->tf_balun_low_ref;
drivers/media/tuners/max2165.c
205
t += (priv->tf_balun_hi_ref - priv->tf_balun_low_ref)
drivers/media/tuners/max2165.c
208
tf = t;
drivers/media/tuners/xc2028.c
175
#define dump_firm_type(t) dump_firm_type_and_int_freq(t, 0)
drivers/media/tuners/xc4000.c
549
#define dump_firm_type(t) dump_firm_type_and_int_freq(t, 0)
drivers/media/usb/au0828/au0828-dvb.c
100
struct au0828_dev *dev = timer_container_of(dev, t, bulk_timeout);
drivers/media/usb/au0828/au0828-dvb.c
98
static void au0828_bulk_timeout(struct timer_list *t)
drivers/media/usb/au0828/au0828-video.c
1484
static int vidioc_g_tuner(struct file *file, void *priv, struct v4l2_tuner *t)
drivers/media/usb/au0828/au0828-video.c
1490
if (t->index != 0)
drivers/media/usb/au0828/au0828-video.c
1500
strscpy(t->name, "Auvitek tuner", sizeof(t->name));
drivers/media/usb/au0828/au0828-video.c
1504
v4l2_device_call_all(&dev->v4l2_dev, 0, tuner, g_tuner, t);
drivers/media/usb/au0828/au0828-video.c
1510
const struct v4l2_tuner *t)
drivers/media/usb/au0828/au0828-video.c
1514
if (t->index != 0)
drivers/media/usb/au0828/au0828-video.c
1522
v4l2_device_call_all(&dev->v4l2_dev, 0, tuner, s_tuner, t);
drivers/media/usb/au0828/au0828-video.c
1525
dprintk(1, "VIDIOC_S_TUNER: signal = %x, afc = %x\n", t->signal,
drivers/media/usb/au0828/au0828-video.c
1526
t->afc);
drivers/media/usb/au0828/au0828-video.c
949
static void au0828_vid_buffer_timeout(struct timer_list *t)
drivers/media/usb/au0828/au0828-video.c
951
struct au0828_dev *dev = timer_container_of(dev, t, vid_timeout);
drivers/media/usb/au0828/au0828-video.c
973
static void au0828_vbi_buffer_timeout(struct timer_list *t)
drivers/media/usb/au0828/au0828-video.c
975
struct au0828_dev *dev = timer_container_of(dev, t, vbi_timeout);
drivers/media/usb/cx231xx/cx231xx-video.c
1088
int cx231xx_g_tuner(struct file *file, void *priv, struct v4l2_tuner *t)
drivers/media/usb/cx231xx/cx231xx-video.c
1092
if (0 != t->index)
drivers/media/usb/cx231xx/cx231xx-video.c
1095
strscpy(t->name, "Tuner", sizeof(t->name));
drivers/media/usb/cx231xx/cx231xx-video.c
1097
t->type = V4L2_TUNER_ANALOG_TV;
drivers/media/usb/cx231xx/cx231xx-video.c
1098
t->capability = V4L2_TUNER_CAP_NORM;
drivers/media/usb/cx231xx/cx231xx-video.c
1099
t->rangehigh = 0xffffffffUL;
drivers/media/usb/cx231xx/cx231xx-video.c
1100
t->signal = 0xffff; /* LOCKED */
drivers/media/usb/cx231xx/cx231xx-video.c
1101
call_all(dev, tuner, g_tuner, t);
drivers/media/usb/cx231xx/cx231xx-video.c
1106
int cx231xx_s_tuner(struct file *file, void *priv, const struct v4l2_tuner *t)
drivers/media/usb/cx231xx/cx231xx-video.c
1108
if (0 != t->index)
drivers/media/usb/cx231xx/cx231xx-video.c
1465
static int radio_g_tuner(struct file *file, void *priv, struct v4l2_tuner *t)
drivers/media/usb/cx231xx/cx231xx-video.c
1469
if (t->index)
drivers/media/usb/cx231xx/cx231xx-video.c
1472
strscpy(t->name, "Radio", sizeof(t->name));
drivers/media/usb/cx231xx/cx231xx-video.c
1474
call_all(dev, tuner, g_tuner, t);
drivers/media/usb/cx231xx/cx231xx-video.c
1478
static int radio_s_tuner(struct file *file, void *priv, const struct v4l2_tuner *t)
drivers/media/usb/cx231xx/cx231xx-video.c
1482
if (t->index)
drivers/media/usb/cx231xx/cx231xx-video.c
1485
call_all(dev, tuner, s_tuner, t);
drivers/media/usb/cx231xx/cx231xx.h
929
int cx231xx_g_tuner(struct file *file, void *priv, struct v4l2_tuner *t);
drivers/media/usb/cx231xx/cx231xx.h
930
int cx231xx_s_tuner(struct file *file, void *priv, const struct v4l2_tuner *t);
drivers/media/usb/dvb-usb-v2/dvb_usb.h
54
#define dvb_usb_dbg_usb_control_msg(udev, r, t, v, i, b, l) { \
drivers/media/usb/dvb-usb-v2/dvb_usb.h
56
if (t == (USB_TYPE_VENDOR | USB_DIR_OUT)) \
drivers/media/usb/dvb-usb-v2/dvb_usb.h
61
"%s %*ph\n", __func__, t, r, v & 0xff, v >> 8, \
drivers/media/usb/em28xx/em28xx-video.c
1823
struct v4l2_tuner *t)
drivers/media/usb/em28xx/em28xx-video.c
1827
if (t->index != 0)
drivers/media/usb/em28xx/em28xx-video.c
1830
strscpy(t->name, "Tuner", sizeof(t->name));
drivers/media/usb/em28xx/em28xx-video.c
1832
v4l2_device_call_all(&dev->v4l2->v4l2_dev, 0, tuner, g_tuner, t);
drivers/media/usb/em28xx/em28xx-video.c
1837
const struct v4l2_tuner *t)
drivers/media/usb/em28xx/em28xx-video.c
1841
if (t->index != 0)
drivers/media/usb/em28xx/em28xx-video.c
1844
v4l2_device_call_all(&dev->v4l2->v4l2_dev, 0, tuner, s_tuner, t);
drivers/media/usb/em28xx/em28xx-video.c
2079
struct v4l2_tuner *t)
drivers/media/usb/em28xx/em28xx-video.c
2083
if (unlikely(t->index > 0))
drivers/media/usb/em28xx/em28xx-video.c
2086
strscpy(t->name, "Radio", sizeof(t->name));
drivers/media/usb/em28xx/em28xx-video.c
2088
v4l2_device_call_all(&dev->v4l2->v4l2_dev, 0, tuner, g_tuner, t);
drivers/media/usb/em28xx/em28xx-video.c
2094
const struct v4l2_tuner *t)
drivers/media/usb/em28xx/em28xx-video.c
2098
if (t->index != 0)
drivers/media/usb/em28xx/em28xx-video.c
2101
v4l2_device_call_all(&dev->v4l2->v4l2_dev, 0, tuner, s_tuner, t);
drivers/media/usb/go7007/go7007-v4l2.c
704
struct v4l2_tuner *t)
drivers/media/usb/go7007/go7007-v4l2.c
708
if (t->index != 0)
drivers/media/usb/go7007/go7007-v4l2.c
711
strscpy(t->name, "Tuner", sizeof(t->name));
drivers/media/usb/go7007/go7007-v4l2.c
712
return call_all(&go->v4l2_dev, tuner, g_tuner, t);
drivers/media/usb/go7007/go7007-v4l2.c
716
const struct v4l2_tuner *t)
drivers/media/usb/go7007/go7007-v4l2.c
720
if (t->index != 0)
drivers/media/usb/go7007/go7007-v4l2.c
723
return call_all(&go->v4l2_dev, tuner, s_tuner, t);
drivers/media/usb/msi2500/msi2500.c
1141
struct spi_transfer *t;
drivers/media/usb/msi2500/msi2500.c
1145
list_for_each_entry(t, &m->transfers, transfer_list) {
drivers/media/usb/msi2500/msi2500.c
1146
dev_dbg(dev->dev, "msg=%*ph\n", t->len, t->tx_buf);
drivers/media/usb/msi2500/msi2500.c
1148
data |= ((u8 *)t->tx_buf)[0] << 8;
drivers/media/usb/msi2500/msi2500.c
1149
data |= ((u8 *)t->tx_buf)[1] << 16;
drivers/media/usb/msi2500/msi2500.c
1150
data |= ((u8 *)t->tx_buf)[2] << 24;
drivers/media/usb/pvrusb2/pvrusb2-hdw.c
3561
static void pvr2_ctl_timeout(struct timer_list *t)
drivers/media/usb/pvrusb2/pvrusb2-hdw.c
3563
struct hdw_timer *timer = timer_container_of(timer, t, timer);
drivers/media/usb/pvrusb2/pvrusb2-hdw.c
4425
static void pvr2_hdw_quiescent_timeout(struct timer_list *t)
drivers/media/usb/pvrusb2/pvrusb2-hdw.c
4427
struct pvr2_hdw *hdw = timer_container_of(hdw, t, quiescent_timer);
drivers/media/usb/pvrusb2/pvrusb2-hdw.c
4436
static void pvr2_hdw_decoder_stabilization_timeout(struct timer_list *t)
drivers/media/usb/pvrusb2/pvrusb2-hdw.c
4438
struct pvr2_hdw *hdw = timer_container_of(hdw, t,
drivers/media/usb/pvrusb2/pvrusb2-hdw.c
4448
static void pvr2_hdw_encoder_wait_timeout(struct timer_list *t)
drivers/media/usb/pvrusb2/pvrusb2-hdw.c
4450
struct pvr2_hdw *hdw = timer_container_of(hdw, t, encoder_wait_timer);
drivers/media/usb/pvrusb2/pvrusb2-hdw.c
4459
static void pvr2_hdw_encoder_run_timeout(struct timer_list *t)
drivers/media/usb/pvrusb2/pvrusb2-hdw.c
4461
struct pvr2_hdw *hdw = timer_container_of(hdw, t, encoder_run_timer);
drivers/media/usb/pvrusb2/pvrusb2-std.c
125
v4l2_std_id t;
drivers/media/usb/pvrusb2/pvrusb2-std.c
159
t = sp->id & cmsk;
drivers/media/usb/pvrusb2/pvrusb2-std.c
160
if (!t) return 0; // Specific color + modulation system illegal
drivers/media/usb/pvrusb2/pvrusb2-std.c
161
id |= t;
drivers/media/usb/pvrusb2/pvrusb2-util.h
11
(t)[i] = (d) & 0xff;\
drivers/media/usb/pvrusb2/pvrusb2-util.h
12
(t)[i+1] = ((d) >> 8) & 0xff;\
drivers/media/usb/pvrusb2/pvrusb2-util.h
13
(t)[i+2] = ((d) >> 16) & 0xff;\
drivers/media/usb/pvrusb2/pvrusb2-util.h
14
(t)[i+3] = ((d) >> 24) & 0xff;\
drivers/media/usb/pvrusb2/pvrusb2-util.h
17
#define PVR2_DECOMPOSE_BE(t,i,d) \
drivers/media/usb/pvrusb2/pvrusb2-util.h
19
(t)[i+3] = (d) & 0xff;\
drivers/media/usb/pvrusb2/pvrusb2-util.h
20
(t)[i+2] = ((d) >> 8) & 0xff;\
drivers/media/usb/pvrusb2/pvrusb2-util.h
21
(t)[i+1] = ((d) >> 16) & 0xff;\
drivers/media/usb/pvrusb2/pvrusb2-util.h
22
(t)[i] = ((d) >> 24) & 0xff;\
drivers/media/usb/pvrusb2/pvrusb2-util.h
25
#define PVR2_COMPOSE_LE(t,i) \
drivers/media/usb/pvrusb2/pvrusb2-util.h
26
((((u32)((t)[i+3])) << 24) | \
drivers/media/usb/pvrusb2/pvrusb2-util.h
27
(((u32)((t)[i+2])) << 16) | \
drivers/media/usb/pvrusb2/pvrusb2-util.h
28
(((u32)((t)[i+1])) << 8) | \
drivers/media/usb/pvrusb2/pvrusb2-util.h
29
((u32)((t)[i])))
drivers/media/usb/pvrusb2/pvrusb2-util.h
31
#define PVR2_COMPOSE_BE(t,i) \
drivers/media/usb/pvrusb2/pvrusb2-util.h
32
((((u32)((t)[i])) << 24) | \
drivers/media/usb/pvrusb2/pvrusb2-util.h
33
(((u32)((t)[i+1])) << 16) | \
drivers/media/usb/pvrusb2/pvrusb2-util.h
34
(((u32)((t)[i+2])) << 8) | \
drivers/media/usb/pvrusb2/pvrusb2-util.h
35
((u32)((t)[i+3])))
drivers/media/usb/pvrusb2/pvrusb2-util.h
9
#define PVR2_DECOMPOSE_LE(t,i,d) \
drivers/media/usb/s2255/s2255drv.c
472
static void s2255_timer(struct timer_list *t)
drivers/media/usb/s2255/s2255drv.c
474
struct s2255_dev *dev = timer_container_of(dev, t, timer);
drivers/media/usb/ttusb-dec/ttusb_dec.c
770
static void ttusb_dec_process_urb_frame_list(struct work_struct *t)
drivers/media/usb/ttusb-dec/ttusb_dec.c
772
struct ttusb_dec *dec = from_work(dec, t, urb_bh_work);
drivers/media/v4l2-core/tuner-core.c
1031
struct tuner *t = to_tuner(i2c_get_clientdata(c));
drivers/media/v4l2-core/tuner-core.c
1032
struct analog_demod_ops *analog_ops = &t->fe.ops.analog_ops;
drivers/media/v4l2-core/tuner-core.c
1035
.mode = t->mode,
drivers/media/v4l2-core/tuner-core.c
1036
.audmode = t->audmode,
drivers/media/v4l2-core/tuner-core.c
1037
.std = t->std
drivers/media/v4l2-core/tuner-core.c
1040
if (t->type == UNSET) {
drivers/media/v4l2-core/tuner-core.c
1062
t->radio_freq = freq;
drivers/media/v4l2-core/tuner-core.c
1063
t->standby = false;
drivers/media/v4l2-core/tuner-core.c
1065
analog_ops->set_params(&t->fe, &params);
drivers/media/v4l2-core/tuner-core.c
1070
t->audmode = params.audmode;
drivers/media/v4l2-core/tuner-core.c
1086
struct tuner *t = fe->analog_demod_priv;
drivers/media/v4l2-core/tuner-core.c
1092
switch (t->mode) {
drivers/media/v4l2-core/tuner-core.c
1104
if (t->mode == V4L2_TUNER_RADIO) {
drivers/media/v4l2-core/tuner-core.c
1105
freq = t->radio_freq / 16000;
drivers/media/v4l2-core/tuner-core.c
1106
freq_fraction = (t->radio_freq % 16000) * 100 / 16000;
drivers/media/v4l2-core/tuner-core.c
1108
freq = t->tv_freq / 16;
drivers/media/v4l2-core/tuner-core.c
1109
freq_fraction = (t->tv_freq % 16) * 100 / 16;
drivers/media/v4l2-core/tuner-core.c
1112
t->standby ? " on standby mode" : "");
drivers/media/v4l2-core/tuner-core.c
1114
pr_info("Standard: 0x%08lx\n", (unsigned long)t->std);
drivers/media/v4l2-core/tuner-core.c
1115
if (t->mode != V4L2_TUNER_RADIO)
drivers/media/v4l2-core/tuner-core.c
1120
fe_tuner_ops->get_status(&t->fe, &tuner_status);
drivers/media/v4l2-core/tuner-core.c
1140
struct tuner *t = to_tuner(sd);
drivers/media/v4l2-core/tuner-core.c
1142
if (set_mode(t, V4L2_TUNER_RADIO) == 0)
drivers/media/v4l2-core/tuner-core.c
1143
set_freq(t, 0);
drivers/media/v4l2-core/tuner-core.c
1157
struct tuner *t = to_tuner(sd);
drivers/media/v4l2-core/tuner-core.c
1158
struct analog_demod_ops *analog_ops = &t->fe.ops.analog_ops;
drivers/media/v4l2-core/tuner-core.c
1161
t->standby = true;
drivers/media/v4l2-core/tuner-core.c
1163
analog_ops->standby(&t->fe);
drivers/media/v4l2-core/tuner-core.c
1169
struct tuner *t = to_tuner(sd);
drivers/media/v4l2-core/tuner-core.c
1171
if (set_mode(t, V4L2_TUNER_ANALOG_TV))
drivers/media/v4l2-core/tuner-core.c
1174
t->std = tuner_fixup_std(t, std);
drivers/media/v4l2-core/tuner-core.c
1175
if (t->std != std)
drivers/media/v4l2-core/tuner-core.c
1176
dprintk("Fixup standard %llx to %llx\n", std, t->std);
drivers/media/v4l2-core/tuner-core.c
1177
set_freq(t, 0);
drivers/media/v4l2-core/tuner-core.c
1183
struct tuner *t = to_tuner(sd);
drivers/media/v4l2-core/tuner-core.c
1185
if (set_mode(t, f->type) == 0)
drivers/media/v4l2-core/tuner-core.c
1186
set_freq(t, f->frequency);
drivers/media/v4l2-core/tuner-core.c
1202
struct tuner *t = to_tuner(sd);
drivers/media/v4l2-core/tuner-core.c
1203
struct dvb_tuner_ops *fe_tuner_ops = &t->fe.ops.tuner_ops;
drivers/media/v4l2-core/tuner-core.c
1205
if (check_mode(t, f->type) == -EINVAL)
drivers/media/v4l2-core/tuner-core.c
1207
if (f->type == t->mode && fe_tuner_ops->get_frequency && !t->standby) {
drivers/media/v4l2-core/tuner-core.c
1210
fe_tuner_ops->get_frequency(&t->fe, &abs_freq);
drivers/media/v4l2-core/tuner-core.c
1211
f->frequency = (V4L2_TUNER_RADIO == t->mode) ?
drivers/media/v4l2-core/tuner-core.c
1216
t->radio_freq : t->tv_freq;
drivers/media/v4l2-core/tuner-core.c
1233
struct tuner *t = to_tuner(sd);
drivers/media/v4l2-core/tuner-core.c
1234
struct analog_demod_ops *analog_ops = &t->fe.ops.analog_ops;
drivers/media/v4l2-core/tuner-core.c
1235
struct dvb_tuner_ops *fe_tuner_ops = &t->fe.ops.tuner_ops;
drivers/media/v4l2-core/tuner-core.c
1237
if (check_mode(t, vt->type) == -EINVAL)
drivers/media/v4l2-core/tuner-core.c
1239
if (vt->type == t->mode && analog_ops->get_afc)
drivers/media/v4l2-core/tuner-core.c
1240
analog_ops->get_afc(&t->fe, &vt->afc);
drivers/media/v4l2-core/tuner-core.c
1241
if (vt->type == t->mode && analog_ops->has_signal) {
drivers/media/v4l2-core/tuner-core.c
1244
if (!analog_ops->has_signal(&t->fe, &signal))
drivers/media/v4l2-core/tuner-core.c
1255
if (vt->type == t->mode) {
drivers/media/v4l2-core/tuner-core.c
1260
fe_tuner_ops->get_status(&t->fe, &tuner_status);
drivers/media/v4l2-core/tuner-core.c
1266
vt->audmode = t->audmode;
drivers/media/v4l2-core/tuner-core.c
1286
struct tuner *t = to_tuner(sd);
drivers/media/v4l2-core/tuner-core.c
1288
if (set_mode(t, vt->type))
drivers/media/v4l2-core/tuner-core.c
1291
if (t->mode == V4L2_TUNER_RADIO) {
drivers/media/v4l2-core/tuner-core.c
1292
t->audmode = vt->audmode;
drivers/media/v4l2-core/tuner-core.c
1299
if (t->audmode != V4L2_TUNER_MODE_MONO &&
drivers/media/v4l2-core/tuner-core.c
1300
t->audmode != V4L2_TUNER_MODE_STEREO)
drivers/media/v4l2-core/tuner-core.c
1301
t->audmode = V4L2_TUNER_MODE_STEREO;
drivers/media/v4l2-core/tuner-core.c
1303
set_freq(t, 0);
drivers/media/v4l2-core/tuner-core.c
1310
struct tuner *t = to_tuner(sd);
drivers/media/v4l2-core/tuner-core.c
1311
struct analog_demod_ops *analog_ops = &t->fe.ops.analog_ops;
drivers/media/v4l2-core/tuner-core.c
1314
analog_ops->tuner_status(&t->fe);
drivers/media/v4l2-core/tuner-core.c
1322
struct tuner *t = to_tuner(i2c_get_clientdata(c));
drivers/media/v4l2-core/tuner-core.c
1323
struct analog_demod_ops *analog_ops = &t->fe.ops.analog_ops;
drivers/media/v4l2-core/tuner-core.c
1327
if (t->fe.ops.tuner_ops.suspend)
drivers/media/v4l2-core/tuner-core.c
1328
t->fe.ops.tuner_ops.suspend(&t->fe);
drivers/media/v4l2-core/tuner-core.c
1329
else if (!t->standby && analog_ops->standby)
drivers/media/v4l2-core/tuner-core.c
1330
analog_ops->standby(&t->fe);
drivers/media/v4l2-core/tuner-core.c
1338
struct tuner *t = to_tuner(i2c_get_clientdata(c));
drivers/media/v4l2-core/tuner-core.c
1342
if (t->fe.ops.tuner_ops.resume)
drivers/media/v4l2-core/tuner-core.c
1343
t->fe.ops.tuner_ops.resume(&t->fe);
drivers/media/v4l2-core/tuner-core.c
1344
else if (!t->standby)
drivers/media/v4l2-core/tuner-core.c
1345
if (set_mode(t, t->mode) == 0)
drivers/media/v4l2-core/tuner-core.c
1346
set_freq(t, 0);
drivers/media/v4l2-core/tuner-core.c
240
struct tuner *t = fe->analog_demod_priv;
drivers/media/v4l2-core/tuner-core.c
260
struct tuner *t = fe->analog_demod_priv;
drivers/media/v4l2-core/tuner-core.c
301
struct tuner *t = to_tuner(i2c_get_clientdata(c));
drivers/media/v4l2-core/tuner-core.c
302
struct dvb_tuner_ops *fe_tuner_ops = &t->fe.ops.tuner_ops;
drivers/media/v4l2-core/tuner-core.c
303
struct analog_demod_ops *analog_ops = &t->fe.ops.analog_ops;
drivers/media/v4l2-core/tuner-core.c
312
t->type = type;
drivers/media/v4l2-core/tuner-core.c
313
t->config = new_config;
drivers/media/v4l2-core/tuner-core.c
316
t->fe.callback = tuner_callback;
drivers/media/v4l2-core/tuner-core.c
320
tuner_detach(&t->fe);
drivers/media/v4l2-core/tuner-core.c
321
t->fe.analog_demod_priv = NULL;
drivers/media/v4l2-core/tuner-core.c
323
switch (t->type) {
drivers/media/v4l2-core/tuner-core.c
326
&t->fe, t->i2c->adapter, t->i2c->addr))
drivers/media/v4l2-core/tuner-core.c
331
if (!dvb_attach(tda829x_attach, &t->fe, t->i2c->adapter,
drivers/media/v4l2-core/tuner-core.c
332
t->i2c->addr, t->config))
drivers/media/v4l2-core/tuner-core.c
337
if (!dvb_attach(tea5767_attach, &t->fe,
drivers/media/v4l2-core/tuner-core.c
338
t->i2c->adapter, t->i2c->addr))
drivers/media/v4l2-core/tuner-core.c
340
t->mode_mask = T_RADIO;
drivers/media/v4l2-core/tuner-core.c
343
if (!dvb_attach(tea5761_attach, &t->fe,
drivers/media/v4l2-core/tuner-core.c
344
t->i2c->adapter, t->i2c->addr))
drivers/media/v4l2-core/tuner-core.c
346
t->mode_mask = T_RADIO;
drivers/media/v4l2-core/tuner-core.c
359
if (!dvb_attach(simple_tuner_attach, &t->fe,
drivers/media/v4l2-core/tuner-core.c
360
t->i2c->adapter, t->i2c->addr, t->type))
drivers/media/v4l2-core/tuner-core.c
369
if (!dvb_attach(simple_tuner_attach, &t->fe,
drivers/media/v4l2-core/tuner-core.c
370
t->i2c->adapter, t->i2c->addr, t->type))
drivers/media/v4l2-core/tuner-core.c
376
.i2c_adap = t->i2c->adapter,
drivers/media/v4l2-core/tuner-core.c
377
.i2c_addr = t->i2c->addr,
drivers/media/v4l2-core/tuner-core.c
379
if (!dvb_attach(xc2028_attach, &t->fe, &cfg))
drivers/media/v4l2-core/tuner-core.c
386
&t->fe, t->i2c->adapter, t->i2c->addr))
drivers/media/v4l2-core/tuner-core.c
392
.i2c_address = t->i2c->addr,
drivers/media/v4l2-core/tuner-core.c
398
&t->fe, t->i2c->adapter, &xc5000_cfg))
drivers/media/v4l2-core/tuner-core.c
406
.i2c_address = t->i2c->addr,
drivers/media/v4l2-core/tuner-core.c
413
&t->fe, t->i2c->adapter, &xc5000c_cfg))
drivers/media/v4l2-core/tuner-core.c
424
if (!dvb_attach(tda18271_attach, &t->fe, t->i2c->addr,
drivers/media/v4l2-core/tuner-core.c
425
t->i2c->adapter, &cfg))
drivers/media/v4l2-core/tuner-core.c
433
.i2c_address = t->i2c->addr,
drivers/media/v4l2-core/tuner-core.c
442
&t->fe, t->i2c->adapter, &xc4000_cfg))
drivers/media/v4l2-core/tuner-core.c
448
if (!dvb_attach(simple_tuner_attach, &t->fe,
drivers/media/v4l2-core/tuner-core.c
449
t->i2c->adapter, t->i2c->addr, t->type))
drivers/media/v4l2-core/tuner-core.c
458
t->name = fe_tuner_ops->info.name;
drivers/media/v4l2-core/tuner-core.c
460
t->fe.analog_demod_priv = t;
drivers/media/v4l2-core/tuner-core.c
470
t->name = analog_ops->info.name;
drivers/media/v4l2-core/tuner-core.c
474
t->sd.entity.name = t->name;
drivers/media/v4l2-core/tuner-core.c
477
dprintk("type set to %s\n", t->name);
drivers/media/v4l2-core/tuner-core.c
479
t->mode_mask = new_mode_mask;
drivers/media/v4l2-core/tuner-core.c
488
if (V4L2_TUNER_RADIO == t->mode)
drivers/media/v4l2-core/tuner-core.c
489
set_radio_freq(c, t->radio_freq);
drivers/media/v4l2-core/tuner-core.c
491
set_tv_freq(c, t->tv_freq);
drivers/media/v4l2-core/tuner-core.c
496
t->mode_mask);
drivers/media/v4l2-core/tuner-core.c
500
dprintk("Tuner attach for type = %d failed.\n", t->type);
drivers/media/v4l2-core/tuner-core.c
501
t->type = TUNER_ABSENT;
drivers/media/v4l2-core/tuner-core.c
523
struct tuner *t = to_tuner(sd);
drivers/media/v4l2-core/tuner-core.c
532
if ((t->type == UNSET && ((tun_setup->addr == ADDR_UNSET) &&
drivers/media/v4l2-core/tuner-core.c
533
(t->mode_mask & tun_setup->mode_mask))) ||
drivers/media/v4l2-core/tuner-core.c
539
t->type, t->mode_mask,
drivers/media/v4l2-core/tuner-core.c
557
struct tuner *t = to_tuner(sd);
drivers/media/v4l2-core/tuner-core.c
558
struct analog_demod_ops *analog_ops = &t->fe.ops.analog_ops;
drivers/media/v4l2-core/tuner-core.c
560
if (t->type != cfg->tuner)
drivers/media/v4l2-core/tuner-core.c
564
analog_ops->set_config(&t->fe, cfg->priv);
drivers/media/v4l2-core/tuner-core.c
629
struct tuner *t;
drivers/media/v4l2-core/tuner-core.c
636
t = kzalloc_obj(struct tuner);
drivers/media/v4l2-core/tuner-core.c
637
if (NULL == t)
drivers/media/v4l2-core/tuner-core.c
639
v4l2_i2c_subdev_init(&t->sd, client, &tuner_ops);
drivers/media/v4l2-core/tuner-core.c
640
t->i2c = client;
drivers/media/v4l2-core/tuner-core.c
641
t->name = "(tuner unset)";
drivers/media/v4l2-core/tuner-core.c
642
t->type = UNSET;
drivers/media/v4l2-core/tuner-core.c
643
t->audmode = V4L2_TUNER_MODE_STEREO;
drivers/media/v4l2-core/tuner-core.c
644
t->standby = true;
drivers/media/v4l2-core/tuner-core.c
645
t->radio_freq = 87.5 * 16000; /* Initial freq range */
drivers/media/v4l2-core/tuner-core.c
646
t->tv_freq = 400 * 16; /* Sets freq to VHF High - needed for some PLL's to properly start */
drivers/media/v4l2-core/tuner-core.c
663
t->i2c->adapter,
drivers/media/v4l2-core/tuner-core.c
664
t->i2c->addr) >= 0) {
drivers/media/v4l2-core/tuner-core.c
665
t->type = TUNER_TEA5761;
drivers/media/v4l2-core/tuner-core.c
666
t->mode_mask = T_RADIO;
drivers/media/v4l2-core/tuner-core.c
667
tuner_lookup(t->i2c->adapter, &radio, &tv);
drivers/media/v4l2-core/tuner-core.c
673
kfree(t);
drivers/media/v4l2-core/tuner-core.c
681
if (tuner_symbol_probe(tda829x_probe, t->i2c->adapter,
drivers/media/v4l2-core/tuner-core.c
682
t->i2c->addr) >= 0) {
drivers/media/v4l2-core/tuner-core.c
686
t->type = TUNER_TDA9887;
drivers/media/v4l2-core/tuner-core.c
687
t->mode_mask = T_RADIO | T_ANALOG_TV;
drivers/media/v4l2-core/tuner-core.c
693
t->i2c->adapter, t->i2c->addr)
drivers/media/v4l2-core/tuner-core.c
695
t->type = TUNER_TEA5767;
drivers/media/v4l2-core/tuner-core.c
696
t->mode_mask = T_RADIO;
drivers/media/v4l2-core/tuner-core.c
698
tuner_lookup(t->i2c->adapter, &radio, &tv);
drivers/media/v4l2-core/tuner-core.c
715
tuner_lookup(t->i2c->adapter, &radio, &tv);
drivers/media/v4l2-core/tuner-core.c
717
t->mode_mask = T_ANALOG_TV;
drivers/media/v4l2-core/tuner-core.c
719
t->mode_mask |= T_RADIO;
drivers/media/v4l2-core/tuner-core.c
720
dprintk("Setting mode_mask to 0x%02x\n", t->mode_mask);
drivers/media/v4l2-core/tuner-core.c
726
t->sd.entity.name = t->name;
drivers/media/v4l2-core/tuner-core.c
732
if (t->type == TUNER_TDA9887) {
drivers/media/v4l2-core/tuner-core.c
733
t->pad[IF_VID_DEC_PAD_IF_INPUT].flags = MEDIA_PAD_FL_SINK;
drivers/media/v4l2-core/tuner-core.c
734
t->pad[IF_VID_DEC_PAD_IF_INPUT].sig_type = PAD_SIGNAL_ANALOG;
drivers/media/v4l2-core/tuner-core.c
735
t->pad[IF_VID_DEC_PAD_OUT].flags = MEDIA_PAD_FL_SOURCE;
drivers/media/v4l2-core/tuner-core.c
736
t->pad[IF_VID_DEC_PAD_OUT].sig_type = PAD_SIGNAL_ANALOG;
drivers/media/v4l2-core/tuner-core.c
737
ret = media_entity_pads_init(&t->sd.entity,
drivers/media/v4l2-core/tuner-core.c
739
&t->pad[0]);
drivers/media/v4l2-core/tuner-core.c
740
t->sd.entity.function = MEDIA_ENT_F_IF_VID_DECODER;
drivers/media/v4l2-core/tuner-core.c
742
t->pad[TUNER_PAD_RF_INPUT].flags = MEDIA_PAD_FL_SINK;
drivers/media/v4l2-core/tuner-core.c
743
t->pad[TUNER_PAD_RF_INPUT].sig_type = PAD_SIGNAL_ANALOG;
drivers/media/v4l2-core/tuner-core.c
744
t->pad[TUNER_PAD_OUTPUT].flags = MEDIA_PAD_FL_SOURCE;
drivers/media/v4l2-core/tuner-core.c
745
t->pad[TUNER_PAD_OUTPUT].sig_type = PAD_SIGNAL_ANALOG;
drivers/media/v4l2-core/tuner-core.c
746
t->pad[TUNER_PAD_AUD_OUT].flags = MEDIA_PAD_FL_SOURCE;
drivers/media/v4l2-core/tuner-core.c
747
t->pad[TUNER_PAD_AUD_OUT].sig_type = PAD_SIGNAL_AUDIO;
drivers/media/v4l2-core/tuner-core.c
748
ret = media_entity_pads_init(&t->sd.entity, TUNER_NUM_PADS,
drivers/media/v4l2-core/tuner-core.c
749
&t->pad[0]);
drivers/media/v4l2-core/tuner-core.c
750
t->sd.entity.function = MEDIA_ENT_F_TUNER;
drivers/media/v4l2-core/tuner-core.c
755
kfree(t);
drivers/media/v4l2-core/tuner-core.c
760
if (t->mode_mask & T_ANALOG_TV)
drivers/media/v4l2-core/tuner-core.c
761
t->mode = V4L2_TUNER_ANALOG_TV;
drivers/media/v4l2-core/tuner-core.c
763
t->mode = V4L2_TUNER_RADIO;
drivers/media/v4l2-core/tuner-core.c
764
set_type(client, t->type, t->mode_mask, t->config, t->fe.callback);
drivers/media/v4l2-core/tuner-core.c
765
list_add_tail(&t->list, &tuner_list);
drivers/media/v4l2-core/tuner-core.c
768
t->type,
drivers/media/v4l2-core/tuner-core.c
769
t->mode_mask & T_RADIO ? " Radio" : "",
drivers/media/v4l2-core/tuner-core.c
770
t->mode_mask & T_ANALOG_TV ? " TV" : "");
drivers/media/v4l2-core/tuner-core.c
782
struct tuner *t = to_tuner(i2c_get_clientdata(client));
drivers/media/v4l2-core/tuner-core.c
784
v4l2_device_unregister_subdev(&t->sd);
drivers/media/v4l2-core/tuner-core.c
785
tuner_detach(&t->fe);
drivers/media/v4l2-core/tuner-core.c
786
t->fe.analog_demod_priv = NULL;
drivers/media/v4l2-core/tuner-core.c
788
list_del(&t->list);
drivers/media/v4l2-core/tuner-core.c
789
kfree(t);
drivers/media/v4l2-core/tuner-core.c
816
static inline int check_mode(struct tuner *t, enum v4l2_tuner_type mode)
drivers/media/v4l2-core/tuner-core.c
824
if ((t_mode & t->mode_mask) == 0)
drivers/media/v4l2-core/tuner-core.c
839
static int set_mode(struct tuner *t, enum v4l2_tuner_type mode)
drivers/media/v4l2-core/tuner-core.c
841
struct analog_demod_ops *analog_ops = &t->fe.ops.analog_ops;
drivers/media/v4l2-core/tuner-core.c
843
if (mode != t->mode) {
drivers/media/v4l2-core/tuner-core.c
844
if (check_mode(t, mode) == -EINVAL) {
drivers/media/v4l2-core/tuner-core.c
847
t->standby = true;
drivers/media/v4l2-core/tuner-core.c
849
analog_ops->standby(&t->fe);
drivers/media/v4l2-core/tuner-core.c
852
t->mode = mode;
drivers/media/v4l2-core/tuner-core.c
863
static void set_freq(struct tuner *t, unsigned int freq)
drivers/media/v4l2-core/tuner-core.c
865
struct i2c_client *client = v4l2_get_subdevdata(&t->sd);
drivers/media/v4l2-core/tuner-core.c
867
if (t->mode == V4L2_TUNER_RADIO) {
drivers/media/v4l2-core/tuner-core.c
869
freq = t->radio_freq;
drivers/media/v4l2-core/tuner-core.c
873
freq = t->tv_freq;
drivers/media/v4l2-core/tuner-core.c
89
i2c_adapter_id(t->i2c->adapter), t->i2c->addr
drivers/media/v4l2-core/tuner-core.c
890
struct tuner *t = to_tuner(i2c_get_clientdata(c));
drivers/media/v4l2-core/tuner-core.c
891
struct analog_demod_ops *analog_ops = &t->fe.ops.analog_ops;
drivers/media/v4l2-core/tuner-core.c
894
.mode = t->mode,
drivers/media/v4l2-core/tuner-core.c
895
.audmode = t->audmode,
drivers/media/v4l2-core/tuner-core.c
896
.std = t->std
drivers/media/v4l2-core/tuner-core.c
899
if (t->type == UNSET) {
drivers/media/v4l2-core/tuner-core.c
921
t->tv_freq = freq;
drivers/media/v4l2-core/tuner-core.c
922
t->standby = false;
drivers/media/v4l2-core/tuner-core.c
924
analog_ops->set_params(&t->fe, &params);
drivers/media/v4l2-core/tuner-core.c
941
static v4l2_std_id tuner_fixup_std(struct tuner *t, v4l2_std_id std)
drivers/media/v4l2-core/v4l2-dv-timings.c
140
bool v4l2_valid_dv_timings(const struct v4l2_dv_timings *t,
drivers/media/v4l2-core/v4l2-dv-timings.c
145
const struct v4l2_bt_timings *bt = &t->bt;
drivers/media/v4l2-core/v4l2-dv-timings.c
151
if (t->type != V4L2_DV_BT_656_1120)
drivers/media/v4l2-core/v4l2-dv-timings.c
153
if (t->type != dvcap->type ||
drivers/media/v4l2-core/v4l2-dv-timings.c
186
return fnc == NULL || fnc(t, fnc_handle);
drivers/media/v4l2-core/v4l2-dv-timings.c
190
int v4l2_enum_dv_timings_cap(struct v4l2_enum_dv_timings *t,
drivers/media/v4l2-core/v4l2-dv-timings.c
197
memset(t->reserved, 0, sizeof(t->reserved));
drivers/media/v4l2-core/v4l2-dv-timings.c
201
idx++ == t->index) {
drivers/media/v4l2-core/v4l2-dv-timings.c
202
t->timings = v4l2_dv_timings_presets[i];
drivers/media/v4l2-core/v4l2-dv-timings.c
210
bool v4l2_find_dv_timings_cap(struct v4l2_dv_timings *t,
drivers/media/v4l2-core/v4l2-dv-timings.c
218
if (!v4l2_valid_dv_timings(t, cap, fnc, fnc_handle))
drivers/media/v4l2-core/v4l2-dv-timings.c
224
v4l2_match_dv_timings(t, v4l2_dv_timings_presets + i,
drivers/media/v4l2-core/v4l2-dv-timings.c
226
u32 flags = t->bt.flags & V4L2_DV_FL_REDUCED_FPS;
drivers/media/v4l2-core/v4l2-dv-timings.c
228
*t = v4l2_dv_timings_presets[i];
drivers/media/v4l2-core/v4l2-dv-timings.c
229
if (can_reduce_fps(&t->bt))
drivers/media/v4l2-core/v4l2-dv-timings.c
230
t->bt.flags |= flags;
drivers/media/v4l2-core/v4l2-dv-timings.c
239
bool v4l2_find_dv_timings_cea861_vic(struct v4l2_dv_timings *t, u8 vic)
drivers/media/v4l2-core/v4l2-dv-timings.c
249
*t = v4l2_dv_timings_presets[i];
drivers/media/v4l2-core/v4l2-dv-timings.c
298
const struct v4l2_dv_timings *t, bool detailed)
drivers/media/v4l2-core/v4l2-dv-timings.c
300
const struct v4l2_bt_timings *bt = &t->bt;
drivers/media/v4l2-core/v4l2-dv-timings.c
304
if (t->type != V4L2_DV_BT_656_1120)
drivers/media/v4l2-core/v4l2-dv-timings.c
378
struct v4l2_fract v4l2_dv_timings_aspect_ratio(const struct v4l2_dv_timings *t)
drivers/media/v4l2-core/v4l2-dv-timings.c
383
if (t->type != V4L2_DV_BT_656_1120)
drivers/media/v4l2-core/v4l2-dv-timings.c
385
if (!(t->bt.flags & V4L2_DV_FL_HAS_PICTURE_ASPECT))
drivers/media/v4l2-core/v4l2-dv-timings.c
388
ratio.numerator = t->bt.width * t->bt.picture_aspect.denominator;
drivers/media/v4l2-core/v4l2-dv-timings.c
389
ratio.denominator = t->bt.height * t->bt.picture_aspect.numerator;
drivers/media/v4l2-core/v4l2-dv-timings.c
407
struct v4l2_fract v4l2_calc_timeperframe(const struct v4l2_dv_timings *t)
drivers/media/v4l2-core/v4l2-dv-timings.c
409
const struct v4l2_bt_timings *bt = &t->bt;
drivers/media/v4l2-core/v4l2-dv-timings.c
415
if (t->type != V4L2_DV_BT_656_1120)
drivers/media/v4l2-core/v4l2-dv-timings.c
500
struct v4l2_dv_timings t = {};
drivers/media/v4l2-core/v4l2-dv-timings.c
631
t.type = V4L2_DV_BT_656_1120;
drivers/media/v4l2-core/v4l2-dv-timings.c
632
t.bt.polarities = polarities;
drivers/media/v4l2-core/v4l2-dv-timings.c
633
t.bt.width = image_width;
drivers/media/v4l2-core/v4l2-dv-timings.c
634
t.bt.height = image_height;
drivers/media/v4l2-core/v4l2-dv-timings.c
635
t.bt.hfrontporch = h_fp;
drivers/media/v4l2-core/v4l2-dv-timings.c
636
t.bt.vfrontporch = v_fp;
drivers/media/v4l2-core/v4l2-dv-timings.c
637
t.bt.hsync = hsync;
drivers/media/v4l2-core/v4l2-dv-timings.c
638
t.bt.vsync = vsync;
drivers/media/v4l2-core/v4l2-dv-timings.c
639
t.bt.hbackporch = frame_width - image_width - h_fp - hsync;
drivers/media/v4l2-core/v4l2-dv-timings.c
642
t.bt.vbackporch = frame_height - image_height - v_fp - vsync;
drivers/media/v4l2-core/v4l2-dv-timings.c
643
t.bt.interlaced = V4L2_DV_PROGRESSIVE;
drivers/media/v4l2-core/v4l2-dv-timings.c
645
t.bt.vbackporch = (frame_height - image_height - 2 * v_fp -
drivers/media/v4l2-core/v4l2-dv-timings.c
647
t.bt.il_vbackporch = frame_height - image_height - 2 * v_fp -
drivers/media/v4l2-core/v4l2-dv-timings.c
648
2 * vsync - t.bt.vbackporch;
drivers/media/v4l2-core/v4l2-dv-timings.c
649
t.bt.il_vfrontporch = v_fp;
drivers/media/v4l2-core/v4l2-dv-timings.c
650
t.bt.il_vsync = vsync;
drivers/media/v4l2-core/v4l2-dv-timings.c
651
t.bt.flags |= V4L2_DV_FL_HALF_LINE;
drivers/media/v4l2-core/v4l2-dv-timings.c
652
t.bt.interlaced = V4L2_DV_INTERLACED;
drivers/media/v4l2-core/v4l2-dv-timings.c
655
t.bt.pixelclock = pix_clk;
drivers/media/v4l2-core/v4l2-dv-timings.c
656
t.bt.standards = V4L2_DV_BT_STD_CVT;
drivers/media/v4l2-core/v4l2-dv-timings.c
659
t.bt.flags |= V4L2_DV_FL_REDUCED_BLANKING;
drivers/media/v4l2-core/v4l2-dv-timings.c
661
if (!v4l2_valid_dv_timings(&t, cap, NULL, NULL))
drivers/media/v4l2-core/v4l2-dv-timings.c
663
*timings = t;
drivers/media/v4l2-core/v4l2-dv-timings.c
724
struct v4l2_dv_timings t = {};
drivers/media/v4l2-core/v4l2-dv-timings.c
795
t.type = V4L2_DV_BT_656_1120;
drivers/media/v4l2-core/v4l2-dv-timings.c
796
t.bt.polarities = polarities;
drivers/media/v4l2-core/v4l2-dv-timings.c
797
t.bt.width = image_width;
drivers/media/v4l2-core/v4l2-dv-timings.c
798
t.bt.height = image_height;
drivers/media/v4l2-core/v4l2-dv-timings.c
799
t.bt.hfrontporch = h_fp;
drivers/media/v4l2-core/v4l2-dv-timings.c
800
t.bt.vfrontporch = v_fp;
drivers/media/v4l2-core/v4l2-dv-timings.c
801
t.bt.hsync = hsync;
drivers/media/v4l2-core/v4l2-dv-timings.c
802
t.bt.vsync = vsync;
drivers/media/v4l2-core/v4l2-dv-timings.c
803
t.bt.hbackporch = frame_width - image_width - h_fp - hsync;
drivers/media/v4l2-core/v4l2-dv-timings.c
806
t.bt.vbackporch = frame_height - image_height - v_fp - vsync;
drivers/media/v4l2-core/v4l2-dv-timings.c
807
t.bt.interlaced = V4L2_DV_PROGRESSIVE;
drivers/media/v4l2-core/v4l2-dv-timings.c
809
t.bt.vbackporch = (frame_height - image_height - 2 * v_fp -
drivers/media/v4l2-core/v4l2-dv-timings.c
811
t.bt.il_vbackporch = frame_height - image_height - 2 * v_fp -
drivers/media/v4l2-core/v4l2-dv-timings.c
812
2 * vsync - t.bt.vbackporch;
drivers/media/v4l2-core/v4l2-dv-timings.c
813
t.bt.il_vfrontporch = v_fp;
drivers/media/v4l2-core/v4l2-dv-timings.c
814
t.bt.il_vsync = vsync;
drivers/media/v4l2-core/v4l2-dv-timings.c
815
t.bt.flags |= V4L2_DV_FL_HALF_LINE;
drivers/media/v4l2-core/v4l2-dv-timings.c
816
t.bt.interlaced = V4L2_DV_INTERLACED;
drivers/media/v4l2-core/v4l2-dv-timings.c
819
t.bt.pixelclock = pix_clk;
drivers/media/v4l2-core/v4l2-dv-timings.c
820
t.bt.standards = V4L2_DV_BT_STD_GTF;
drivers/media/v4l2-core/v4l2-dv-timings.c
823
t.bt.flags |= V4L2_DV_FL_REDUCED_BLANKING;
drivers/media/v4l2-core/v4l2-dv-timings.c
825
if (!v4l2_valid_dv_timings(&t, cap, NULL, NULL))
drivers/media/v4l2-core/v4l2-dv-timings.c
827
*timings = t;
drivers/media/v4l2-core/v4l2-ioctl.c
2812
struct v4l2_tuner t = {
drivers/media/v4l2-core/v4l2-ioctl.c
2819
err = ops->vidioc_g_tuner(file, NULL, &t);
drivers/media/v4l2-core/v4l2-ioctl.c
2822
p->capability = t.capability | V4L2_TUNER_CAP_FREQ_BANDS;
drivers/media/v4l2-core/v4l2-ioctl.c
2823
p->rangelow = t.rangelow;
drivers/media/v4l2-core/v4l2-ioctl.c
2824
p->rangehigh = t.rangehigh;
drivers/memory/omap-gpmc.c
1188
static void gpmc_omap_onenand_calc_sync_timings(struct gpmc_timings *t,
drivers/memory/omap-gpmc.c
1264
gpmc_calc_timings(t, s, &dev_t);
drivers/memory/omap-gpmc.c
1795
static void gpmc_convert_ps_to_ns(struct gpmc_timings *t)
drivers/memory/omap-gpmc.c
1797
t->cs_on /= 1000;
drivers/memory/omap-gpmc.c
1798
t->cs_rd_off /= 1000;
drivers/memory/omap-gpmc.c
1799
t->cs_wr_off /= 1000;
drivers/memory/omap-gpmc.c
1800
t->adv_on /= 1000;
drivers/memory/omap-gpmc.c
1801
t->adv_rd_off /= 1000;
drivers/memory/omap-gpmc.c
1802
t->adv_wr_off /= 1000;
drivers/memory/omap-gpmc.c
1803
t->we_on /= 1000;
drivers/memory/omap-gpmc.c
1804
t->we_off /= 1000;
drivers/memory/omap-gpmc.c
1805
t->oe_on /= 1000;
drivers/memory/omap-gpmc.c
1806
t->oe_off /= 1000;
drivers/memory/omap-gpmc.c
1807
t->page_burst_access /= 1000;
drivers/memory/omap-gpmc.c
1808
t->access /= 1000;
drivers/memory/omap-gpmc.c
1809
t->rd_cycle /= 1000;
drivers/memory/omap-gpmc.c
1810
t->wr_cycle /= 1000;
drivers/memory/omap-gpmc.c
1811
t->bus_turnaround /= 1000;
drivers/memory/omap-gpmc.c
1812
t->cycle2cycle_delay /= 1000;
drivers/memory/omap-gpmc.c
1813
t->wait_monitoring /= 1000;
drivers/memory/omap-gpmc.c
1814
t->clk_activation /= 1000;
drivers/memory/omap-gpmc.c
1815
t->wr_access /= 1000;
drivers/memory/omap-gpmc.c
1816
t->wr_data_mux_bus /= 1000;
drivers/memory/omap-gpmc.c
700
int gpmc_cs_set_timings(int cs, const struct gpmc_timings *t,
drivers/memory/omap-gpmc.c
706
div = gpmc_calc_divider(t->sync_clk);
drivers/memory/omap-gpmc.c
726
div = gpmc_calc_waitmonitoring_divider(t->wait_monitoring);
drivers/memory/omap-gpmc.c
730
t->wait_monitoring
drivers/memory/omap-gpmc.c
737
ret |= set_gpmc_timing_reg(cs, GPMC_CS_CONFIG2, 0, 3, 0, t->cs_on,
drivers/memory/omap-gpmc.c
739
ret |= set_gpmc_timing_reg(cs, GPMC_CS_CONFIG2, 8, 12, 0, t->cs_rd_off,
drivers/memory/omap-gpmc.c
741
ret |= set_gpmc_timing_reg(cs, GPMC_CS_CONFIG2, 16, 20, 0, t->cs_wr_off,
drivers/memory/omap-gpmc.c
746
ret |= set_gpmc_timing_reg(cs, GPMC_CS_CONFIG3, 0, 3, 0, t->adv_on,
drivers/memory/omap-gpmc.c
748
ret |= set_gpmc_timing_reg(cs, GPMC_CS_CONFIG3, 8, 12, 0, t->adv_rd_off,
drivers/memory/omap-gpmc.c
750
ret |= set_gpmc_timing_reg(cs, GPMC_CS_CONFIG3, 16, 20, 0, t->adv_wr_off,
drivers/memory/omap-gpmc.c
757
t->adv_aad_mux_on, GPMC_CD_FCLK,
drivers/memory/omap-gpmc.c
760
t->adv_aad_mux_rd_off, GPMC_CD_FCLK,
drivers/memory/omap-gpmc.c
763
t->adv_aad_mux_wr_off, GPMC_CD_FCLK,
drivers/memory/omap-gpmc.c
769
ret |= set_gpmc_timing_reg(cs, GPMC_CS_CONFIG4, 0, 3, 0, t->oe_on,
drivers/memory/omap-gpmc.c
771
ret |= set_gpmc_timing_reg(cs, GPMC_CS_CONFIG4, 8, 12, 0, t->oe_off,
drivers/memory/omap-gpmc.c
775
t->oe_aad_mux_on, GPMC_CD_FCLK,
drivers/memory/omap-gpmc.c
778
t->oe_aad_mux_off, GPMC_CD_FCLK,
drivers/memory/omap-gpmc.c
781
ret |= set_gpmc_timing_reg(cs, GPMC_CS_CONFIG4, 16, 19, 0, t->we_on,
drivers/memory/omap-gpmc.c
783
ret |= set_gpmc_timing_reg(cs, GPMC_CS_CONFIG4, 24, 28, 0, t->we_off,
drivers/memory/omap-gpmc.c
788
ret |= set_gpmc_timing_reg(cs, GPMC_CS_CONFIG5, 0, 4, 0, t->rd_cycle,
drivers/memory/omap-gpmc.c
790
ret |= set_gpmc_timing_reg(cs, GPMC_CS_CONFIG5, 8, 12, 0, t->wr_cycle,
drivers/memory/omap-gpmc.c
792
ret |= set_gpmc_timing_reg(cs, GPMC_CS_CONFIG5, 16, 20, 0, t->access,
drivers/memory/omap-gpmc.c
795
t->page_burst_access, GPMC_CD_FCLK,
drivers/memory/omap-gpmc.c
801
t->bus_turnaround, GPMC_CD_FCLK,
drivers/memory/omap-gpmc.c
804
t->cycle2cycle_delay, GPMC_CD_FCLK,
drivers/memory/omap-gpmc.c
811
t->wr_data_mux_bus, GPMC_CD_FCLK,
drivers/memory/omap-gpmc.c
818
t->wr_access, GPMC_CD_FCLK,
drivers/memory/omap-gpmc.c
832
t->wait_monitoring, GPMC_CD_CLK,
drivers/memory/omap-gpmc.c
836
t->clk_activation, GPMC_CD_FCLK,
drivers/memory/omap-gpmc.c
846
gpmc_cs_bool_timings(cs, &t->bool_timings);
drivers/memory/tegra/tegra210-emc-cc-r21021.c
1153
s32 t = (s32)(ramp_up_wait + ramp_down_wait) /
drivers/memory/tegra/tegra210-emc-cc-r21021.c
1155
zq_latch_dvfs_wait_time = (s32)tZQCAL_lpddr4_fc_adj - t;
drivers/memstick/core/ms_block.c
1498
static void msb_cache_flush_timer(struct timer_list *t)
drivers/memstick/core/ms_block.c
1500
struct msb_data *msb = timer_container_of(msb, t, cache_flush_timer);
drivers/memstick/host/jmb38x_ms.c
590
static void jmb38x_ms_abort(struct timer_list *t)
drivers/memstick/host/jmb38x_ms.c
592
struct jmb38x_ms_host *host = timer_container_of(host, t, timer);
drivers/memstick/host/r592.c
615
static void r592_detect_timer(struct timer_list *t)
drivers/memstick/host/r592.c
617
struct r592_device *dev = timer_container_of(dev, t, detect_timer);
drivers/memstick/host/tifm_ms.c
535
static void tifm_ms_abort(struct timer_list *t)
drivers/memstick/host/tifm_ms.c
537
struct tifm_ms *host = timer_container_of(host, t, timer);
drivers/message/fusion/lsi/mpi_targ.h
587
#define SET_IO_INDEX(t, i) \
drivers/message/fusion/lsi/mpi_targ.h
588
((t) = ((t) & ~TARGET_MODE_REPLY_IO_INDEX_MASK) | \
drivers/message/fusion/lsi/mpi_targ.h
595
#define SET_INITIATOR_INDEX(t, ii) \
drivers/message/fusion/lsi/mpi_targ.h
596
((t) = ((t) & ~TARGET_MODE_REPLY_INITIATOR_INDEX_MASK) | \
drivers/message/fusion/lsi/mpi_targ.h
603
#define SET_ALIAS(t, a) ((t) = ((t) & ~TARGET_MODE_REPLY_ALIAS_MASK) | \
drivers/message/fusion/lsi/mpi_targ.h
610
#define SET_PORT(t, p) ((t) = ((t) & ~TARGET_MODE_REPLY_PORT_MASK) | \
drivers/message/fusion/lsi/mpi_targ.h
627
#define SET_HOST_INDEX_0100(t, hi) \
drivers/message/fusion/lsi/mpi_targ.h
628
((t) = ((t) & ~TARGET_MODE_REPLY_0100_MASK_HOST_INDEX) | \
drivers/message/fusion/lsi/mpi_targ.h
635
#define SET_IOC_INDEX_0100(t, ii) \
drivers/message/fusion/lsi/mpi_targ.h
636
((t) = ((t) & ~TARGET_MODE_REPLY_0100_MASK_IOC_INDEX) | \
drivers/message/fusion/lsi/mpi_targ.h
644
#define SET_INITIATOR_INDEX_0100(t, ii) \
drivers/message/fusion/lsi/mpi_targ.h
645
((t) = ((t) & ~TARGET_MODE_REPLY_0100_MASK_INITIATOR_INDEX) | \
drivers/message/fusion/mptbase.c
4654
int t;
drivers/message/fusion/mptbase.c
4676
if ((t = WaitForDoorbellInt(ioc, 5, sleepFlag)) < 0)
drivers/message/fusion/mptbase.c
4680
ioc->name, reqBytes, t, failcnt ? " - MISSING DOORBELL HANDSHAKE!" : ""));
drivers/message/fusion/mptbase.c
4692
if (!failcnt && (t = WaitForDoorbellAck(ioc, 5, sleepFlag)) < 0)
drivers/message/fusion/mptbase.c
4710
if ((t = WaitForDoorbellAck(ioc, 5, sleepFlag)) < 0)
drivers/message/fusion/mptbase.c
4718
ioc->name, t, failcnt ? " - MISSING DOORBELL ACK!" : ""));
drivers/message/fusion/mptbase.c
4723
if (!failcnt && (t = WaitForDoorbellReply(ioc, maxwait, sleepFlag)) < 0)
drivers/message/fusion/mptbase.c
4727
ioc->name, t, failcnt ? " - MISSING DOORBELL REPLY!" : ""));
drivers/message/fusion/mptbase.c
4859
int t;
drivers/message/fusion/mptbase.c
4870
if ((t = WaitForDoorbellInt(ioc, howlong, sleepFlag)) < 0) {
drivers/message/fusion/mptbase.c
4875
if ((t = WaitForDoorbellInt(ioc, 5, sleepFlag)) < 0)
drivers/message/fusion/mptbase.c
4884
ioc->name, t, le32_to_cpu(*(u32 *)hs_reply),
drivers/message/fusion/mptbase.c
4892
if ((t = WaitForDoorbellInt(ioc, 5, sleepFlag)) < 0)
drivers/message/fusion/mptbase.c
4901
if (!failcnt && (t = WaitForDoorbellInt(ioc, 5, sleepFlag)) < 0)
drivers/message/fusion/mptbase.c
4923
ioc->name, t, u16cnt/2));
drivers/mfd/ezx-pcap.c
52
struct spi_transfer t;
drivers/mfd/ezx-pcap.c
56
memset(&t, 0, sizeof(t));
drivers/mfd/ezx-pcap.c
58
t.len = sizeof(u32);
drivers/mfd/ezx-pcap.c
59
spi_message_add_tail(&t, &m);
drivers/mfd/ezx-pcap.c
62
t.tx_buf = (u8 *) &pcap->buf;
drivers/mfd/ezx-pcap.c
63
t.rx_buf = (u8 *) &pcap->buf;
drivers/mfd/mc13xxx-spi.c
67
struct spi_transfer t = {
drivers/mfd/mc13xxx-spi.c
80
spi_message_add_tail(&t, &m);
drivers/mfd/menelaus.c
732
int t, ret;
drivers/mfd/menelaus.c
745
t = (GPIO_CTRL_SLPCTLEN | GPIO3_DIR_INPUT);
drivers/mfd/menelaus.c
747
ret |= t;
drivers/mfd/menelaus.c
749
ret &= ~t;
drivers/mfd/menelaus.c
827
static void menelaus_to_time(char *regs, struct rtc_time *t)
drivers/mfd/menelaus.c
829
t->tm_sec = bcd2bin(regs[0]);
drivers/mfd/menelaus.c
830
t->tm_min = bcd2bin(regs[1]);
drivers/mfd/menelaus.c
832
t->tm_hour = bcd2bin(regs[2] & 0x1f) - 1;
drivers/mfd/menelaus.c
834
t->tm_hour += 12;
drivers/mfd/menelaus.c
836
t->tm_hour = bcd2bin(regs[2] & 0x3f);
drivers/mfd/menelaus.c
837
t->tm_mday = bcd2bin(regs[3]);
drivers/mfd/menelaus.c
838
t->tm_mon = bcd2bin(regs[4]) - 1;
drivers/mfd/menelaus.c
839
t->tm_year = bcd2bin(regs[5]) + 100;
drivers/mfd/menelaus.c
842
static int time_to_menelaus(struct rtc_time *t, int regnum)
drivers/mfd/menelaus.c
846
status = menelaus_write_reg(regnum++, bin2bcd(t->tm_sec));
drivers/mfd/menelaus.c
850
status = menelaus_write_reg(regnum++, bin2bcd(t->tm_min));
drivers/mfd/menelaus.c
855
hour = t->tm_hour + 1;
drivers/mfd/menelaus.c
861
hour = bin2bcd(t->tm_hour);
drivers/mfd/menelaus.c
866
status = menelaus_write_reg(regnum++, bin2bcd(t->tm_mday));
drivers/mfd/menelaus.c
870
status = menelaus_write_reg(regnum++, bin2bcd(t->tm_mon + 1));
drivers/mfd/menelaus.c
874
status = menelaus_write_reg(regnum++, bin2bcd(t->tm_year - 100));
drivers/mfd/menelaus.c
885
static int menelaus_read_time(struct device *dev, struct rtc_time *t)
drivers/mfd/menelaus.c
910
menelaus_to_time(regs, t);
drivers/mfd/menelaus.c
911
t->tm_wday = bcd2bin(regs[6]);
drivers/mfd/menelaus.c
916
static int menelaus_set_time(struct device *dev, struct rtc_time *t)
drivers/mfd/menelaus.c
921
status = time_to_menelaus(t, MENELAUS_RTC_SEC);
drivers/mfd/menelaus.c
924
status = menelaus_write_reg(MENELAUS_RTC_WKDAY, bin2bcd(t->tm_wday));
drivers/mfd/rohm-bd71828.c
782
const struct regmap_irq_type *t = &irq_data->type;
drivers/mfd/rohm-bd71828.c
791
buf[0][idx] &= ~t->type_reg_mask;
drivers/misc/bcm-vk/bcm_vk_tty.c
44
static void bcm_vk_tty_poll(struct timer_list *t)
drivers/misc/bcm-vk/bcm_vk_tty.c
46
struct bcm_vk *vk = timer_container_of(vk, t, serial_timer);
drivers/misc/cardreader/rtsx_usb.c
32
static void rtsx_usb_sg_timed_out(struct timer_list *t)
drivers/misc/cardreader/rtsx_usb.c
34
struct rtsx_ucr *ucr = timer_container_of(ucr, t, sg_timer);
drivers/misc/cb710/debug.c
29
#define CB710_READ_REGS_TEMPLATE(t) \
drivers/misc/cb710/debug.c
30
static void cb710_read_regs_##t(void __iomem *iobase, \
drivers/misc/cb710/debug.c
31
u##t *reg, unsigned select) \
drivers/misc/cb710/debug.c
35
for (i = 0; i < ARRAY_SIZE(allow); ++i, reg += 16/(t/8)) { \
drivers/misc/cb710/debug.c
39
for (j = 0; j < 0x10/(t/8); ++j) { \
drivers/misc/cb710/debug.c
40
if (!allow_reg_read(i, j, t)) \
drivers/misc/cb710/debug.c
42
reg[j] = ioread##t(iobase \
drivers/misc/cb710/debug.c
43
+ (i << 4) + (j * (t/8))); \
drivers/misc/cb710/debug.c
53
#define CB710_DUMP_REGS_TEMPLATE(t) \
drivers/misc/cb710/debug.c
54
static void cb710_dump_regs_##t(struct device *dev, \
drivers/misc/cb710/debug.c
55
const u##t *reg, unsigned select) \
drivers/misc/cb710/debug.c
57
const char *const xp = &cb710_xes[8 - t/4]; \
drivers/misc/cb710/debug.c
58
const char *const format = cb710_regf_##t; \
drivers/misc/cb710/debug.c
63
for (i = 0; i < ARRAY_SIZE(allow); ++i, reg += 16/(t/8)) { \
drivers/misc/cb710/debug.c
67
for (j = 0; j < 0x10/(t/8); ++j) { \
drivers/misc/cb710/debug.c
69
if (j == 8/(t/8)) \
drivers/misc/cb710/debug.c
71
if (allow_reg_read(i, j, t)) \
drivers/misc/cb710/debug.c
80
#define CB710_READ_AND_DUMP_REGS_TEMPLATE(t) \
drivers/misc/cb710/debug.c
81
static void cb710_read_and_dump_regs_##t(struct cb710_chip *chip, \
drivers/misc/cb710/debug.c
84
u##t regs[CB710_REG_COUNT/sizeof(u##t)]; \
drivers/misc/cb710/debug.c
87
cb710_read_regs_##t(chip->iobase, regs, select); \
drivers/misc/cb710/debug.c
88
cb710_dump_regs_##t(cb710_chip_dev(chip), regs, select); \
drivers/misc/cb710/debug.c
91
#define CB710_REG_ACCESS_TEMPLATES(t) \
drivers/misc/cb710/debug.c
92
CB710_READ_REGS_TEMPLATE(t) \
drivers/misc/cb710/debug.c
93
CB710_DUMP_REGS_TEMPLATE(t) \
drivers/misc/cb710/debug.c
94
CB710_READ_AND_DUMP_REGS_TEMPLATE(t)
drivers/misc/cs5535-mfgpt.c
170
unsigned long t;
drivers/misc/cs5535-mfgpt.c
173
t = find_first_bit(mfgpt->avail, max);
drivers/misc/cs5535-mfgpt.c
175
timer_nr = t < max ? (int) t : -1;
drivers/misc/cs5535-mfgpt.c
267
struct cs5535_mfgpt_timer t;
drivers/misc/cs5535-mfgpt.c
270
t.nr = i;
drivers/misc/cs5535-mfgpt.c
272
cs5535_mfgpt_toggle_event(&t, MFGPT_CMP1, MFGPT_EVENT_RESET, 0);
drivers/misc/cs5535-mfgpt.c
273
cs5535_mfgpt_toggle_event(&t, MFGPT_CMP2, MFGPT_EVENT_RESET, 0);
drivers/misc/cs5535-mfgpt.c
274
cs5535_mfgpt_toggle_event(&t, MFGPT_CMP1, MFGPT_EVENT_NMI, 0);
drivers/misc/cs5535-mfgpt.c
275
cs5535_mfgpt_toggle_event(&t, MFGPT_CMP2, MFGPT_EVENT_NMI, 0);
drivers/misc/cs5535-mfgpt.c
276
cs5535_mfgpt_toggle_event(&t, MFGPT_CMP1, MFGPT_EVENT_IRQ, 0);
drivers/misc/cs5535-mfgpt.c
277
cs5535_mfgpt_toggle_event(&t, MFGPT_CMP2, MFGPT_EVENT_IRQ, 0);
drivers/misc/cs5535-mfgpt.c
320
int err = -EIO, t;
drivers/misc/cs5535-mfgpt.c
354
t = scan_timers(&cs5535_mfgpt_chip);
drivers/misc/cs5535-mfgpt.c
355
dev_info(&pdev->dev, "%d MFGPT timers available\n", t);
drivers/misc/eeprom/eeprom_93xx46.c
130
struct spi_transfer t[2] = {};
drivers/misc/eeprom/eeprom_93xx46.c
152
t[0].tx_buf = (char *)&cmd_addr;
drivers/misc/eeprom/eeprom_93xx46.c
153
t[0].len = 2;
drivers/misc/eeprom/eeprom_93xx46.c
154
t[0].bits_per_word = bits;
drivers/misc/eeprom/eeprom_93xx46.c
156
t[1].rx_buf = buf;
drivers/misc/eeprom/eeprom_93xx46.c
157
t[1].len = count;
drivers/misc/eeprom/eeprom_93xx46.c
158
t[1].bits_per_word = 8;
drivers/misc/eeprom/eeprom_93xx46.c
160
spi_message_init_with_transfers(&m, t, ARRAY_SIZE(t));
drivers/misc/eeprom/eeprom_93xx46.c
187
struct spi_transfer t = {};
drivers/misc/eeprom/eeprom_93xx46.c
205
t.tx_buf = &cmd_addr;
drivers/misc/eeprom/eeprom_93xx46.c
206
t.len = 2;
drivers/misc/eeprom/eeprom_93xx46.c
207
t.bits_per_word = bits;
drivers/misc/eeprom/eeprom_93xx46.c
209
spi_message_init_with_transfers(&m, &t, 1);
drivers/misc/eeprom/eeprom_93xx46.c
233
struct spi_transfer t[2] = {};
drivers/misc/eeprom/eeprom_93xx46.c
255
t[0].tx_buf = (char *)&cmd_addr;
drivers/misc/eeprom/eeprom_93xx46.c
256
t[0].len = 2;
drivers/misc/eeprom/eeprom_93xx46.c
257
t[0].bits_per_word = bits;
drivers/misc/eeprom/eeprom_93xx46.c
259
t[1].tx_buf = buf;
drivers/misc/eeprom/eeprom_93xx46.c
260
t[1].len = data_len;
drivers/misc/eeprom/eeprom_93xx46.c
261
t[1].bits_per_word = 8;
drivers/misc/eeprom/eeprom_93xx46.c
263
spi_message_init_with_transfers(&m, t, ARRAY_SIZE(t));
drivers/misc/eeprom/eeprom_93xx46.c
321
struct spi_transfer t = {};
drivers/misc/eeprom/eeprom_93xx46.c
338
t.tx_buf = &cmd_addr;
drivers/misc/eeprom/eeprom_93xx46.c
339
t.len = 2;
drivers/misc/eeprom/eeprom_93xx46.c
340
t.bits_per_word = bits;
drivers/misc/eeprom/eeprom_93xx46.c
342
spi_message_init_with_transfers(&m, &t, 1);
drivers/misc/genwqe/card_ddcb.c
623
unsigned int t;
drivers/misc/genwqe/card_ddcb.c
640
for (t = 0; t < GENWQE_DDCB_SOFTWARE_TIMEOUT * 10; t++) {
drivers/misc/genwqe/card_debugfs.c
197
u64 t;
drivers/misc/genwqe/card_debugfs.c
199
t = genwqe_read_vreg(cd, IO_SLC_VF_QUEUE_WTIME, 0);
drivers/misc/genwqe/card_debugfs.c
200
seq_printf(s, " PF 0x%016llx\n", t);
drivers/misc/genwqe/card_debugfs.c
203
t = genwqe_read_vreg(cd, IO_SLC_VF_QUEUE_WTIME, vf_num + 1);
drivers/misc/genwqe/card_debugfs.c
204
seq_printf(s, " VF%-2d 0x%016llx\n", vf_num, t);
drivers/misc/genwqe/card_sysfs.c
100
u64 t;
drivers/misc/genwqe/card_sysfs.c
103
t = __genwqe_readq(cd, IO_SLC_FREE_RUNNING_TIMER);
drivers/misc/genwqe/card_sysfs.c
104
return sprintf(buf, "%016llx\n", t);
drivers/misc/genwqe/card_sysfs.c
112
u64 t;
drivers/misc/genwqe/card_sysfs.c
115
t = __genwqe_readq(cd, IO_SLC_QUEUE_WTIME);
drivers/misc/genwqe/card_sysfs.c
116
return sprintf(buf, "%016llx\n", t);
drivers/misc/sgi-xp/xpc_main.c
165
xpc_timeout_partition_disengage(struct timer_list *t)
drivers/misc/sgi-xp/xpc_main.c
167
struct xpc_partition *part = timer_container_of(part, t,
drivers/mmc/core/host.c
213
static void mmc_retune_timer(struct timer_list *t)
drivers/mmc/core/host.c
215
struct mmc_host *host = timer_container_of(host, t, retune_timer);
drivers/mmc/core/mmc_test.c
1377
struct mmc_test_area *t = &test->area;
drivers/mmc/core/mmc_test.c
1381
t->blocks = sz >> 9;
drivers/mmc/core/mmc_test.c
1384
err = mmc_test_map_sg_max_scatter(t->mem, sz, t->sg,
drivers/mmc/core/mmc_test.c
1385
t->max_segs, t->max_seg_sz,
drivers/mmc/core/mmc_test.c
1386
&t->sg_len);
drivers/mmc/core/mmc_test.c
1388
err = mmc_test_map_sg(t->mem, sz, t->sg, 1, t->max_segs,
drivers/mmc/core/mmc_test.c
1389
t->max_seg_sz, &t->sg_len, min_sg_len);
drivers/mmc/core/mmc_test.c
1396
err = mmc_test_map_sg_max_scatter(t->mem, sz, t->sg_areq,
drivers/mmc/core/mmc_test.c
1397
t->max_segs, t->max_seg_sz,
drivers/mmc/core/mmc_test.c
1400
err = mmc_test_map_sg(t->mem, sz, t->sg_areq, 1, t->max_segs,
drivers/mmc/core/mmc_test.c
1401
t->max_seg_sz, &sg_len, min_sg_len);
drivers/mmc/core/mmc_test.c
1403
if (!err && sg_len != t->sg_len)
drivers/mmc/core/mmc_test.c
1419
struct mmc_test_area *t = &test->area;
drivers/mmc/core/mmc_test.c
1421
return mmc_test_simple_transfer(test, t->sg, t->sg_len, dev_addr,
drivers/mmc/core/mmc_test.c
1422
t->blocks, 512, write);
drivers/mmc/core/mmc_test.c
1442
struct mmc_test_area *t = &test->area;
drivers/mmc/core/mmc_test.c
1445
if (t->max_seg_sz >= PAGE_SIZE)
drivers/mmc/core/mmc_test.c
1446
max_tfr = t->max_segs * PAGE_SIZE;
drivers/mmc/core/mmc_test.c
1448
max_tfr = t->max_segs * t->max_seg_sz;
drivers/mmc/core/mmc_test.c
1492
struct mmc_test_area *t = &test->area;
drivers/mmc/core/mmc_test.c
1494
return mmc_test_area_io(test, t->max_tfr, t->dev_addr, 1, 0, 0);
drivers/mmc/core/mmc_test.c
1502
struct mmc_test_area *t = &test->area;
drivers/mmc/core/mmc_test.c
1507
return mmc_erase(test->card, t->dev_addr, t->max_sz >> 9,
drivers/mmc/core/mmc_test.c
1516
struct mmc_test_area *t = &test->area;
drivers/mmc/core/mmc_test.c
1518
kfree(t->sg);
drivers/mmc/core/mmc_test.c
1519
kfree(t->sg_areq);
drivers/mmc/core/mmc_test.c
1520
mmc_test_free_mem(t->mem);
drivers/mmc/core/mmc_test.c
1534
struct mmc_test_area *t = &test->area;
drivers/mmc/core/mmc_test.c
1544
t->max_sz = sz;
drivers/mmc/core/mmc_test.c
1545
while (t->max_sz < 4 * 1024 * 1024)
drivers/mmc/core/mmc_test.c
1546
t->max_sz += sz;
drivers/mmc/core/mmc_test.c
1547
while (t->max_sz > TEST_AREA_MAX_SIZE && t->max_sz > sz)
drivers/mmc/core/mmc_test.c
1548
t->max_sz -= sz;
drivers/mmc/core/mmc_test.c
1550
t->max_segs = test->card->host->max_segs;
drivers/mmc/core/mmc_test.c
1551
t->max_seg_sz = test->card->host->max_seg_size;
drivers/mmc/core/mmc_test.c
1552
t->max_seg_sz -= t->max_seg_sz % 512;
drivers/mmc/core/mmc_test.c
1554
t->max_tfr = t->max_sz;
drivers/mmc/core/mmc_test.c
1555
if (t->max_tfr >> 9 > test->card->host->max_blk_count)
drivers/mmc/core/mmc_test.c
1556
t->max_tfr = test->card->host->max_blk_count << 9;
drivers/mmc/core/mmc_test.c
1557
if (t->max_tfr > test->card->host->max_req_size)
drivers/mmc/core/mmc_test.c
1558
t->max_tfr = test->card->host->max_req_size;
drivers/mmc/core/mmc_test.c
1559
if (t->max_tfr / t->max_seg_sz > t->max_segs)
drivers/mmc/core/mmc_test.c
1560
t->max_tfr = t->max_segs * t->max_seg_sz;
drivers/mmc/core/mmc_test.c
1568
t->mem = mmc_test_alloc_mem(min_sz, t->max_tfr, t->max_segs,
drivers/mmc/core/mmc_test.c
1569
t->max_seg_sz);
drivers/mmc/core/mmc_test.c
1570
if (!t->mem)
drivers/mmc/core/mmc_test.c
1573
t->sg = kmalloc_objs(*t->sg, t->max_segs);
drivers/mmc/core/mmc_test.c
1574
if (!t->sg) {
drivers/mmc/core/mmc_test.c
1579
t->sg_areq = kmalloc_objs(*t->sg_areq, t->max_segs);
drivers/mmc/core/mmc_test.c
1580
if (!t->sg_areq) {
drivers/mmc/core/mmc_test.c
1585
t->dev_addr = mmc_test_capacity(test->card) / 2;
drivers/mmc/core/mmc_test.c
1586
t->dev_addr -= t->dev_addr % (t->max_sz >> 9);
drivers/mmc/core/mmc_test.c
1642
struct mmc_test_area *t = &test->area;
drivers/mmc/core/mmc_test.c
1644
return mmc_test_area_io(test, t->max_tfr, t->dev_addr, write,
drivers/mmc/core/mmc_test.c
1685
struct mmc_test_area *t = &test->area;
drivers/mmc/core/mmc_test.c
1690
for (sz = 512; sz < t->max_tfr; sz <<= 1) {
drivers/mmc/core/mmc_test.c
1691
dev_addr = t->dev_addr + (sz >> 9);
drivers/mmc/core/mmc_test.c
1696
sz = t->max_tfr;
drivers/mmc/core/mmc_test.c
1697
dev_addr = t->dev_addr;
drivers/mmc/core/mmc_test.c
1706
struct mmc_test_area *t = &test->area;
drivers/mmc/core/mmc_test.c
1714
for (sz = 512; sz < t->max_tfr; sz <<= 1) {
drivers/mmc/core/mmc_test.c
1715
dev_addr = t->dev_addr + (sz >> 9);
drivers/mmc/core/mmc_test.c
1723
sz = t->max_tfr;
drivers/mmc/core/mmc_test.c
1724
dev_addr = t->dev_addr;
drivers/mmc/core/mmc_test.c
1733
struct mmc_test_area *t = &test->area;
drivers/mmc/core/mmc_test.c
1745
for (sz = 512; sz < t->max_sz; sz <<= 1) {
drivers/mmc/core/mmc_test.c
1746
dev_addr = t->dev_addr + (sz >> 9);
drivers/mmc/core/mmc_test.c
1754
dev_addr = t->dev_addr;
drivers/mmc/core/mmc_test.c
1766
struct mmc_test_area *t = &test->area;
drivers/mmc/core/mmc_test.c
1771
cnt = t->max_sz / sz;
drivers/mmc/core/mmc_test.c
1772
dev_addr = t->dev_addr;
drivers/mmc/core/mmc_test.c
1790
struct mmc_test_area *t = &test->area;
drivers/mmc/core/mmc_test.c
1794
for (sz = 512; sz < t->max_tfr; sz <<= 1) {
drivers/mmc/core/mmc_test.c
1799
sz = t->max_tfr;
drivers/mmc/core/mmc_test.c
1805
struct mmc_test_area *t = &test->area;
drivers/mmc/core/mmc_test.c
1813
cnt = t->max_sz / sz;
drivers/mmc/core/mmc_test.c
1814
dev_addr = t->dev_addr;
drivers/mmc/core/mmc_test.c
1832
struct mmc_test_area *t = &test->area;
drivers/mmc/core/mmc_test.c
1836
for (sz = 512; sz < t->max_tfr; sz <<= 1) {
drivers/mmc/core/mmc_test.c
1841
sz = t->max_tfr;
drivers/mmc/core/mmc_test.c
1850
struct mmc_test_area *t = &test->area;
drivers/mmc/core/mmc_test.c
1862
for (sz = 512; sz <= t->max_sz; sz <<= 1) {
drivers/mmc/core/mmc_test.c
1869
cnt = t->max_sz / sz;
drivers/mmc/core/mmc_test.c
1870
dev_addr = t->dev_addr;
drivers/mmc/core/mmc_test.c
1935
struct mmc_test_area *t = &test->area;
drivers/mmc/core/mmc_test.c
1940
for (sz = 512; sz < t->max_tfr; sz <<= 1) {
drivers/mmc/core/mmc_test.c
1957
sz = t->max_tfr;
drivers/mmc/core/mmc_test.c
1998
struct mmc_test_area *t = &test->area;
drivers/mmc/core/mmc_test.c
2003
sz = t->max_tfr;
drivers/mmc/core/mmc_test.c
2012
if (t->max_seg_sz >= PAGE_SIZE)
drivers/mmc/core/mmc_test.c
2013
max_tfr = t->max_segs * PAGE_SIZE;
drivers/mmc/core/mmc_test.c
2015
max_tfr = t->max_segs * t->max_seg_sz;
drivers/mmc/core/mmc_test.c
2087
struct mmc_test_area *t = &test->area;
drivers/mmc/core/mmc_test.c
2093
if (reqsize > t->max_tfr)
drivers/mmc/core/mmc_test.c
2094
reqsize = t->max_tfr;
drivers/mmc/core/mmc_test.c
2363
struct mmc_test_area *t = &test->area;
drivers/mmc/core/mmc_test.c
2379
mmc_test_prepare_mrq(test, mrq, t->sg, t->sg_len, dev_addr, t->blocks,
drivers/mmc/core/mmc_test.c
2382
if (use_sbc && t->blocks > 1 && !mrq->sbc) {
drivers/mmc/core/mmc_test.c
2459
if (repeat_cmd && (t->blocks + 1) << 9 > t->max_tfr)
drivers/mmc/core/mmc_test.c
2461
mmc_hostname(test->card->host), count, t->blocks);
drivers/mmc/core/mmc_test.c
2475
struct mmc_test_area *t = &test->area;
drivers/mmc/core/mmc_test.c
2485
ret = mmc_test_ongoing_transfer(test, t->dev_addr, use_sbc, 0, write,
drivers/mmc/core/mmc_test.c
2490
return mmc_test_ongoing_transfer(test, t->dev_addr, use_sbc, 1, write,
drivers/mmc/core/mmc_test.c
2497
struct mmc_test_area *t = &test->area;
drivers/mmc/core/mmc_test.c
2501
for (sz = 512; sz <= t->max_tfr; sz += 512) {
drivers/mmc/core/mmc_test.c
824
struct mmc_test_area *t = &test->area;
drivers/mmc/core/mmc_test.c
825
struct scatterlist *sg = t->sg;
drivers/mmc/core/mmc_test.c
826
struct scatterlist *sg_areq = t->sg_areq;
drivers/mmc/core/mmc_test.c
840
mmc_test_prepare_mrq(test, mrq, sg, t->sg_len, dev_addr,
drivers/mmc/core/mmc_test.c
841
t->blocks, 512, write);
drivers/mmc/core/mmc_test.c
851
dev_addr += t->blocks;
drivers/mmc/host/atmel-mci.c
1653
static void atmci_detect_change(struct timer_list *t)
drivers/mmc/host/atmel-mci.c
1655
struct atmel_mci_slot *slot = timer_container_of(slot, t,
drivers/mmc/host/atmel-mci.c
1750
static void atmci_work_func(struct work_struct *t)
drivers/mmc/host/atmel-mci.c
1752
struct atmel_mci *host = from_work(host, t, bh_work);
drivers/mmc/host/atmel-mci.c
715
static void atmci_timeout_timer(struct timer_list *t)
drivers/mmc/host/atmel-mci.c
717
struct atmel_mci *host = timer_container_of(host, t, timer);
drivers/mmc/host/au1xmmc.c
257
static void au1xmmc_finish_bh_work(struct work_struct *t)
drivers/mmc/host/au1xmmc.c
259
struct au1xmmc_host *host = from_work(host, t, finish_bh_work);
drivers/mmc/host/au1xmmc.c
367
static void au1xmmc_data_bh_work(struct work_struct *t)
drivers/mmc/host/au1xmmc.c
369
struct au1xmmc_host *host = from_work(host, t, data_bh_work);
drivers/mmc/host/cb710-mmc.c
650
static void cb710_mmc_finish_request_bh_work(struct work_struct *t)
drivers/mmc/host/cb710-mmc.c
652
struct cb710_mmc_reader *reader = from_work(reader, t,
drivers/mmc/host/dw_mmc.c
1829
static enum hrtimer_restart dw_mci_fault_timer(struct hrtimer *t)
drivers/mmc/host/dw_mmc.c
1831
struct dw_mci *host = container_of(t, struct dw_mci, fault_timer);
drivers/mmc/host/dw_mmc.c
2064
static void dw_mci_work_func(struct work_struct *t)
drivers/mmc/host/dw_mmc.c
2066
struct dw_mci *host = from_work(host, t, bh_work);
drivers/mmc/host/dw_mmc.c
3174
static void dw_mci_cmd11_timer(struct timer_list *t)
drivers/mmc/host/dw_mmc.c
3176
struct dw_mci *host = timer_container_of(host, t, cmd11_timer);
drivers/mmc/host/dw_mmc.c
3188
static void dw_mci_cto_timer(struct timer_list *t)
drivers/mmc/host/dw_mmc.c
3190
struct dw_mci *host = timer_container_of(host, t, cto_timer);
drivers/mmc/host/dw_mmc.c
3243
static void dw_mci_dto_timer(struct timer_list *t)
drivers/mmc/host/dw_mmc.c
3245
struct dw_mci *host = timer_container_of(host, t, dto_timer);
drivers/mmc/host/dw_mmc.h
423
#define SDMMC_SET_FIFOTH(m, r, t) (((m) & 0x7) << 28 | \
drivers/mmc/host/dw_mmc.h
425
((t) & 0xFFF))
drivers/mmc/host/jz4740_mmc.c
642
static void jz4740_mmc_timeout(struct timer_list *t)
drivers/mmc/host/jz4740_mmc.c
644
struct jz4740_mmc_host *host = timer_container_of(host, t,
drivers/mmc/host/meson-mx-sdio.c
453
static void meson_mx_mmc_timeout(struct timer_list *t)
drivers/mmc/host/meson-mx-sdio.c
455
struct meson_mx_mmc_host *host = timer_container_of(host, t,
drivers/mmc/host/mmc_spi.c
114
struct spi_transfer token, t, crc, early_status;
drivers/mmc/host/mmc_spi.c
219
u8 *end = cp + host->t.len;
drivers/mmc/host/mmc_spi.c
405
struct spi_transfer *t;
drivers/mmc/host/mmc_spi.c
482
t = &host->t;
drivers/mmc/host/mmc_spi.c
483
memset(t, 0, sizeof(*t));
drivers/mmc/host/mmc_spi.c
484
t->tx_buf = t->rx_buf = data->status;
drivers/mmc/host/mmc_spi.c
485
t->len = cp - data->status;
drivers/mmc/host/mmc_spi.c
486
t->cs_change = 1;
drivers/mmc/host/mmc_spi.c
487
spi_message_add_tail(t, &host->m);
drivers/mmc/host/mmc_spi.c
511
struct spi_transfer *t;
drivers/mmc/host/mmc_spi.c
520
t = &host->token;
drivers/mmc/host/mmc_spi.c
521
memset(t, 0, sizeof(*t));
drivers/mmc/host/mmc_spi.c
522
t->len = 1;
drivers/mmc/host/mmc_spi.c
527
t->tx_buf = &scratch->data_token;
drivers/mmc/host/mmc_spi.c
528
spi_message_add_tail(t, &host->m);
drivers/mmc/host/mmc_spi.c
534
t = &host->t;
drivers/mmc/host/mmc_spi.c
535
memset(t, 0, sizeof(*t));
drivers/mmc/host/mmc_spi.c
536
t->tx_buf = host->ones;
drivers/mmc/host/mmc_spi.c
538
spi_message_add_tail(t, &host->m);
drivers/mmc/host/mmc_spi.c
540
t = &host->crc;
drivers/mmc/host/mmc_spi.c
541
memset(t, 0, sizeof(*t));
drivers/mmc/host/mmc_spi.c
542
t->len = 2;
drivers/mmc/host/mmc_spi.c
545
t->tx_buf = &scratch->crc_val;
drivers/mmc/host/mmc_spi.c
547
t->tx_buf = host->ones;
drivers/mmc/host/mmc_spi.c
548
t->rx_buf = &scratch->crc_val;
drivers/mmc/host/mmc_spi.c
550
spi_message_add_tail(t, &host->m);
drivers/mmc/host/mmc_spi.c
567
t = &host->early_status;
drivers/mmc/host/mmc_spi.c
568
memset(t, 0, sizeof(*t));
drivers/mmc/host/mmc_spi.c
569
t->len = sizeof(scratch->status);
drivers/mmc/host/mmc_spi.c
570
t->tx_buf = host->ones;
drivers/mmc/host/mmc_spi.c
571
t->rx_buf = scratch->status;
drivers/mmc/host/mmc_spi.c
572
t->cs_change = 1;
drivers/mmc/host/mmc_spi.c
573
spi_message_add_tail(t, &host->m);
drivers/mmc/host/mmc_spi.c
590
mmc_spi_writeblock(struct mmc_spi_host *host, struct spi_transfer *t,
drivers/mmc/host/mmc_spi.c
599
scratch->crc_val = cpu_to_be16(crc_itu_t(0, t->tx_buf, t->len));
drivers/mmc/host/mmc_spi.c
654
t->tx_buf += t->len;
drivers/mmc/host/mmc_spi.c
684
mmc_spi_readblock(struct mmc_spi_host *host, struct spi_transfer *t,
drivers/mmc/host/mmc_spi.c
728
u8 *cp = t->rx_buf;
drivers/mmc/host/mmc_spi.c
732
for (len = t->len; len; len--) {
drivers/mmc/host/mmc_spi.c
746
u16 crc = crc_itu_t(0, t->rx_buf, t->len);
drivers/mmc/host/mmc_spi.c
752
scratch->crc_val, crc, t->len);
drivers/mmc/host/mmc_spi.c
757
t->rx_buf += t->len;
drivers/mmc/host/mmc_spi.c
772
struct spi_transfer *t;
drivers/mmc/host/mmc_spi.c
782
t = &host->t;
drivers/mmc/host/mmc_spi.c
784
if (t->speed_hz)
drivers/mmc/host/mmc_spi.c
785
clock_rate = t->speed_hz;
drivers/mmc/host/mmc_spi.c
804
t->tx_buf = kmap_addr + sg->offset;
drivers/mmc/host/mmc_spi.c
806
t->rx_buf = kmap_addr + sg->offset;
drivers/mmc/host/mmc_spi.c
810
t->len = min(length, blk_size);
drivers/mmc/host/mmc_spi.c
812
dev_dbg(&spi->dev, " %s block, %d bytes\n", write_or_read, t->len);
drivers/mmc/host/mmc_spi.c
815
status = mmc_spi_writeblock(host, t, timeout);
drivers/mmc/host/mmc_spi.c
817
status = mmc_spi_readblock(host, t, timeout);
drivers/mmc/host/mmc_spi.c
821
data->bytes_xfered += t->len;
drivers/mmc/host/mmc_spi.c
822
length -= t->len;
drivers/mmc/host/mvsdio.c
510
static void mvsd_timeout_timer(struct timer_list *t)
drivers/mmc/host/mvsdio.c
512
struct mvsd_host *host = timer_container_of(host, t, timer);
drivers/mmc/host/mvsdio.c
72
unsigned long t = jiffies + HZ;
drivers/mmc/host/mvsdio.c
76
if (time_after(jiffies, t)) {
drivers/mmc/host/mvsdio.c
84
hw_state, count, jiffies - (t - HZ));
drivers/mmc/host/mxcmmc.c
956
static void mxcmci_watchdog(struct timer_list *t)
drivers/mmc/host/mxcmmc.c
958
struct mxcmci_host *host = timer_container_of(host, t, watchdog);
drivers/mmc/host/omap.c
640
mmc_omap_cmd_timer(struct timer_list *t)
drivers/mmc/host/omap.c
642
struct mmc_omap_host *host = timer_container_of(host, t,
drivers/mmc/host/omap.c
657
mmc_omap_clk_timer(struct timer_list *t)
drivers/mmc/host/omap.c
659
struct mmc_omap_host *host = timer_container_of(host, t, clk_timer);
drivers/mmc/host/omap.c
881
static void mmc_omap_cover_timer(struct timer_list *t)
drivers/mmc/host/omap.c
883
struct mmc_omap_slot *slot = timer_container_of(slot, t, cover_timer);
drivers/mmc/host/omap.c
887
static void mmc_omap_cover_bh_handler(struct work_struct *t)
drivers/mmc/host/omap.c
889
struct mmc_omap_slot *slot = from_work(slot, t, cover_bh_work);
drivers/mmc/host/sdhci.c
3238
static void sdhci_timeout_timer(struct timer_list *t)
drivers/mmc/host/sdhci.c
3243
host = timer_container_of(host, t, timer);
drivers/mmc/host/sdhci.c
3260
static void sdhci_timeout_data_timer(struct timer_list *t)
drivers/mmc/host/sdhci.c
3265
host = timer_container_of(host, t, data_timer);
drivers/mmc/host/tifm_sd.c
727
static void tifm_sd_end_cmd(struct work_struct *t)
drivers/mmc/host/tifm_sd.c
729
struct tifm_sd *host = from_work(host, t, finish_bh_work);
drivers/mmc/host/tifm_sd.c
778
static void tifm_sd_abort(struct timer_list *t)
drivers/mmc/host/tifm_sd.c
780
struct tifm_sd *host = timer_container_of(host, t, timer);
drivers/mmc/host/uniphier-sd.c
239
static void uniphier_sd_internal_dma_issue(struct work_struct *t)
drivers/mmc/host/uniphier-sd.c
241
struct tmio_mmc_host *host = from_work(host, t, dma_issue);
drivers/mmc/host/uniphier-sd.c
93
static void uniphier_sd_external_dma_issue(struct work_struct *t)
drivers/mmc/host/uniphier-sd.c
95
struct tmio_mmc_host *host = from_work(host, t, dma_issue);
drivers/mmc/host/via-sdmmc.c
935
static void via_sdc_timeout(struct timer_list *t)
drivers/mmc/host/via-sdmmc.c
940
sdhost = timer_container_of(sdhost, t, timer);
drivers/mmc/host/via-sdmmc.c
966
static void via_sdc_finish_bh_work(struct work_struct *t)
drivers/mmc/host/via-sdmmc.c
968
struct via_crdr_mmc_host *host = from_work(host, t, finish_bh_work);
drivers/mmc/host/vub300.c
1182
static void vub300_sg_timed_out(struct timer_list *t)
drivers/mmc/host/vub300.c
1184
struct vub300_mmc_host *vub300 = timer_container_of(vub300, t,
drivers/mmc/host/vub300.c
742
static void vub300_inactivity_timer_expired(struct timer_list *t)
drivers/mmc/host/vub300.c
744
struct vub300_mmc_host *vub300 = timer_container_of(vub300, t,
drivers/mmc/host/wbsd.c
1039
static void wbsd_fifo_bh_work(struct work_struct *t)
drivers/mmc/host/wbsd.c
1041
struct wbsd_host *host = from_work(host, t, fifo_bh_work);
drivers/mmc/host/wbsd.c
1070
static void wbsd_crc_bh_work(struct work_struct *t)
drivers/mmc/host/wbsd.c
1072
struct wbsd_host *host = from_work(host, t, crc_bh_work);
drivers/mmc/host/wbsd.c
1094
static void wbsd_timeout_bh_work(struct work_struct *t)
drivers/mmc/host/wbsd.c
1096
struct wbsd_host *host = from_work(host, t, timeout_bh_work);
drivers/mmc/host/wbsd.c
1118
static void wbsd_finish_bh_work(struct work_struct *t)
drivers/mmc/host/wbsd.c
1120
struct wbsd_host *host = from_work(host, t, finish_bh_work);
drivers/mmc/host/wbsd.c
948
static void wbsd_reset_ignore(struct timer_list *t)
drivers/mmc/host/wbsd.c
950
struct wbsd_host *host = timer_container_of(host, t, ignore_timer);
drivers/mmc/host/wbsd.c
990
static void wbsd_card_bh_work(struct work_struct *t)
drivers/mmc/host/wbsd.c
992
struct wbsd_host *host = from_work(host, t, card_bh_work);
drivers/most/most_usb.c
668
static void link_stat_timer_handler(struct timer_list *t)
drivers/most/most_usb.c
670
struct most_dev *mdev = timer_container_of(mdev, t, link_stat_timer);
drivers/mtd/devices/mtd_dataflash.c
462
struct spi_transfer t;
drivers/mtd/devices/mtd_dataflash.c
483
memset(&t, 0, sizeof t);
drivers/mtd/devices/mtd_dataflash.c
484
t.tx_buf = scratch;
drivers/mtd/devices/mtd_dataflash.c
485
t.rx_buf = scratch;
drivers/mtd/devices/mtd_dataflash.c
486
t.len = l;
drivers/mtd/devices/mtd_dataflash.c
487
spi_message_add_tail(&t, &m);
drivers/mtd/devices/mtd_dataflash.c
541
struct spi_transfer t;
drivers/mtd/devices/mtd_dataflash.c
569
memset(&t, 0, sizeof t);
drivers/mtd/devices/mtd_dataflash.c
570
t.tx_buf = scratch;
drivers/mtd/devices/mtd_dataflash.c
571
t.len = l;
drivers/mtd/devices/mtd_dataflash.c
572
spi_message_add_tail(&t, &m);
drivers/mtd/devices/sst25l.c
312
struct spi_transfer t;
drivers/mtd/devices/sst25l.c
318
memset(&t, 0, sizeof(struct spi_transfer));
drivers/mtd/devices/sst25l.c
326
t.tx_buf = cmd_resp;
drivers/mtd/devices/sst25l.c
327
t.rx_buf = cmd_resp;
drivers/mtd/devices/sst25l.c
328
t.len = sizeof(cmd_resp);
drivers/mtd/devices/sst25l.c
329
spi_message_add_tail(&t, &m);
drivers/mtd/devices/sst25l.c
70
struct spi_transfer t;
drivers/mtd/devices/sst25l.c
75
memset(&t, 0, sizeof(struct spi_transfer));
drivers/mtd/devices/sst25l.c
79
t.tx_buf = cmd_resp;
drivers/mtd/devices/sst25l.c
80
t.rx_buf = cmd_resp;
drivers/mtd/devices/sst25l.c
81
t.len = sizeof(cmd_resp);
drivers/mtd/devices/sst25l.c
82
spi_message_add_tail(&t, &m);
drivers/mtd/devices/st_spi_fsm.c
1586
uint8_t *t = (uint8_t *)&tmp;
drivers/mtd/devices/st_spi_fsm.c
1640
memset(t, 0xff, write_mask + 1); /* fill with 0xff's */
drivers/mtd/devices/st_spi_fsm.c
1642
t[i] = *p++;
drivers/mtd/devices/st_spi_fsm.c
918
uint8_t *t = (uint8_t *)&tmp;
drivers/mtd/devices/st_spi_fsm.c
934
data[i] = t[i];
drivers/mtd/maps/lantiq-flash.c
104
*t++ = *f++;
drivers/mtd/maps/lantiq-flash.c
85
unsigned char *t = (unsigned char *)to;
drivers/mtd/maps/lantiq-flash.c
90
*t++ = *f++;
drivers/mtd/maps/lantiq-flash.c
99
unsigned char *t = (unsigned char *)map->virt + to;
drivers/mtd/maps/pcmciamtd.c
340
cistpl_format_t *t = &parse.format;
drivers/mtd/maps/pcmciamtd.c
341
(void)t; /* Shut up, gcc */
drivers/mtd/maps/pcmciamtd.c
343
t->type, t->edc, t->offset, t->length);
drivers/mtd/maps/pcmciamtd.c
356
cistpl_jedec_t *t = &parse.jedec;
drivers/mtd/maps/pcmciamtd.c
357
for (i = 0; i < t->nid; i++)
drivers/mtd/maps/pcmciamtd.c
359
t->id[i].mfr, t->id[i].info);
drivers/mtd/maps/pcmciamtd.c
370
cistpl_device_t *t = &parse.device;
drivers/mtd/maps/pcmciamtd.c
377
dev->pcmcia_map.size = t->dev[0].size;
drivers/mtd/maps/pcmciamtd.c
379
for (i = 0; i < t->ndev; i++) {
drivers/mtd/maps/pcmciamtd.c
380
pr_debug("Region %d, type = %u\n", i, t->dev[i].type);
drivers/mtd/maps/pcmciamtd.c
381
pr_debug("Region %d, wp = %u\n", i, t->dev[i].wp);
drivers/mtd/maps/pcmciamtd.c
382
pr_debug("Region %d, speed = %u ns\n", i, t->dev[i].speed);
drivers/mtd/maps/pcmciamtd.c
383
pr_debug("Region %d, size = %u bytes\n", i, t->dev[i].size);
drivers/mtd/maps/pcmciamtd.c
394
cistpl_device_geo_t *t = &parse.device_geo;
drivers/mtd/maps/pcmciamtd.c
400
dev->pcmcia_map.bankwidth = t->geo[0].buswidth;
drivers/mtd/maps/pcmciamtd.c
402
for (i = 0; i < t->ngeo; i++) {
drivers/mtd/maps/pcmciamtd.c
403
pr_debug("region: %d bankwidth = %u\n", i, t->geo[i].buswidth);
drivers/mtd/maps/pcmciamtd.c
404
pr_debug("region: %d erase_block = %u\n", i, t->geo[i].erase_block);
drivers/mtd/maps/pcmciamtd.c
405
pr_debug("region: %d read_block = %u\n", i, t->geo[i].read_block);
drivers/mtd/maps/pcmciamtd.c
406
pr_debug("region: %d write_block = %u\n", i, t->geo[i].write_block);
drivers/mtd/maps/pcmciamtd.c
407
pr_debug("region: %d partition = %u\n", i, t->geo[i].partition);
drivers/mtd/maps/pcmciamtd.c
408
pr_debug("region: %d interleave = %u\n", i, t->geo[i].interleave);
drivers/mtd/nand/ecc-sw-bch.c
113
unsigned int m, t, i;
drivers/mtd/nand/ecc-sw-bch.c
118
t = (eccbytes * 8) / m;
drivers/mtd/nand/ecc-sw-bch.c
120
engine_conf->bch = bch_init(m, t, 0, false);
drivers/mtd/nand/ecc-sw-bch.c
125
engine_conf->errloc = kmalloc_array(t, sizeof(*engine_conf->errloc),
drivers/mtd/nand/raw/cadence-nand-controller.c
1370
struct cadence_nand_timings *t)
drivers/mtd/nand/raw/cadence-nand-controller.c
1372
writel_relaxed(t->async_toggle_timings,
drivers/mtd/nand/raw/cadence-nand-controller.c
1374
writel_relaxed(t->timings0, cdns_ctrl->reg + TIMINGS0);
drivers/mtd/nand/raw/cadence-nand-controller.c
1375
writel_relaxed(t->timings1, cdns_ctrl->reg + TIMINGS1);
drivers/mtd/nand/raw/cadence-nand-controller.c
1376
writel_relaxed(t->timings2, cdns_ctrl->reg + TIMINGS2);
drivers/mtd/nand/raw/cadence-nand-controller.c
1379
writel_relaxed(t->dll_phy_ctrl, cdns_ctrl->reg + DLL_PHY_CTRL);
drivers/mtd/nand/raw/cadence-nand-controller.c
1381
writel_relaxed(t->phy_ctrl, cdns_ctrl->reg + PHY_CTRL);
drivers/mtd/nand/raw/cadence-nand-controller.c
1386
writel_relaxed(t->phy_dqs_timing,
drivers/mtd/nand/raw/cadence-nand-controller.c
1388
writel_relaxed(t->phy_gate_lpbk_ctrl,
drivers/mtd/nand/raw/cadence-nand-controller.c
2404
struct cadence_nand_timings *t = &cdns_chip->timings;
drivers/mtd/nand/raw/cadence-nand-controller.c
2422
memset(t, 0, sizeof(*t));
drivers/mtd/nand/raw/cadence-nand-controller.c
2554
t->async_toggle_timings = reg;
drivers/mtd/nand/raw/cadence-nand-controller.c
2574
t->timings0 = reg;
drivers/mtd/nand/raw/cadence-nand-controller.c
2595
t->timings1 = reg;
drivers/mtd/nand/raw/cadence-nand-controller.c
2608
t->timings2 = reg;
drivers/mtd/nand/raw/cadence-nand-controller.c
2620
t->dll_phy_ctrl = reg;
drivers/mtd/nand/raw/cadence-nand-controller.c
2661
t->phy_ctrl = reg;
drivers/mtd/nand/raw/cadence-nand-controller.c
2669
t->phy_dqs_timing = dll_phy_dqs_timing;
drivers/mtd/nand/raw/cadence-nand-controller.c
2674
t->phy_gate_lpbk_ctrl = reg;
drivers/mtd/nand/raw/cadence-nand-controller.c
2689
struct cadence_nand_timings *t = &cdns_chip->timings;
drivers/mtd/nand/raw/cadence-nand-controller.c
2706
memset(t, 0, sizeof(*t));
drivers/mtd/nand/raw/cadence-nand-controller.c
2712
t->sync_timings = reg;
drivers/mtd/nand/raw/cadence-nand-controller.c
2723
t->timings0 = reg;
drivers/mtd/nand/raw/cadence-nand-controller.c
2741
t->timings1 = reg;
drivers/mtd/nand/raw/cadence-nand-controller.c
2753
t->timings2 = reg;
drivers/mtd/nand/raw/cadence-nand-controller.c
2758
t->dll_phy_ctrl = reg;
drivers/mtd/nand/raw/cadence-nand-controller.c
2762
t->phy_ctrl = reg;
drivers/mtd/nand/raw/cadence-nand-controller.c
2784
t->sync_timings = reg;
drivers/mtd/nand/raw/cadence-nand-controller.c
2803
t->phy_gate_lpbk_ctrl = reg;
drivers/mtd/nand/raw/cadence-nand-controller.c
2820
t->phy_dq_timing = reg;
drivers/mtd/nand/raw/cadence-nand-controller.c
2826
t->phy_dqs_timing = reg;
drivers/mtd/nand/raw/cadence-nand-controller.c
2832
t->phy_ie_timing = reg;
drivers/mtd/nand/raw/cadence-nand-controller.c
2847
writel_relaxed(t->sync_timings, cdns_ctrl->reg + SYNC_TIMINGS);
drivers/mtd/nand/raw/cadence-nand-controller.c
2848
writel_relaxed(t->timings0, cdns_ctrl->reg + TIMINGS0);
drivers/mtd/nand/raw/cadence-nand-controller.c
2849
writel_relaxed(t->timings1, cdns_ctrl->reg + TIMINGS1);
drivers/mtd/nand/raw/cadence-nand-controller.c
2850
writel_relaxed(t->timings2, cdns_ctrl->reg + TIMINGS2);
drivers/mtd/nand/raw/cadence-nand-controller.c
2851
writel_relaxed(t->dll_phy_ctrl, cdns_ctrl->reg + DLL_PHY_CTRL);
drivers/mtd/nand/raw/cadence-nand-controller.c
2852
writel_relaxed(t->phy_ctrl, cdns_ctrl->reg + PHY_CTRL);
drivers/mtd/nand/raw/cadence-nand-controller.c
2854
writel_relaxed(t->phy_dq_timing, cdns_ctrl->reg + PHY_DQ_TIMING);
drivers/mtd/nand/raw/cadence-nand-controller.c
2855
writel_relaxed(t->phy_dqs_timing, cdns_ctrl->reg + PHY_DQS_TIMING);
drivers/mtd/nand/raw/cadence-nand-controller.c
2856
writel_relaxed(t->phy_gate_lpbk_ctrl,
drivers/mtd/nand/raw/cadence-nand-controller.c
2862
writel_relaxed(t->phy_ie_timing, cdns_ctrl->reg + PHY_IE_TIMING);
drivers/mtd/nand/raw/mxc_nand.c
1477
__iomem u16 *t = buf;
drivers/mtd/nand/raw/mxc_nand.c
1481
__raw_writew(0xffff, t++);
drivers/mtd/nand/raw/mxc_nand.c
195
u32 *t = trg;
drivers/mtd/nand/raw/mxc_nand.c
199
*t++ = __raw_readl(s++);
drivers/mtd/nand/raw/mxc_nand.c
205
u16 *t = trg;
drivers/mtd/nand/raw/mxc_nand.c
215
*t++ = __raw_readw(s++);
drivers/mtd/nand/raw/mxc_nand.c
227
__iomem u16 *t = trg;
drivers/mtd/nand/raw/mxc_nand.c
237
__raw_writew(*s++, t++);
drivers/mtd/nand/raw/mxic_nand.c
26
#define HC_CFG_TYPE(s, t) ((t) << (23 + ((s) * 2)))
drivers/mtd/sm_ftl.c
992
static void sm_cache_flush_timer(struct timer_list *t)
drivers/mtd/sm_ftl.c
994
struct sm_ftl *ftl = timer_container_of(ftl, t, timer);
drivers/net/amt.c
102
hlist_for_each_entry_safe(snode, t, &gc_list, node) {
drivers/net/amt.c
1328
struct hlist_node *t;
drivers/net/amt.c
1334
hlist_for_each_entry_safe(gnode, t, &tunnel->groups[i], node)
drivers/net/amt.c
1362
struct hlist_node *t;
drivers/net/amt.c
1367
hlist_for_each_entry_safe(snode, t, &gnode->sources[i], node) {
drivers/net/amt.c
1488
struct hlist_node *t;
drivers/net/amt.c
1525
hlist_for_each_entry_safe(snode, t, &gnode->sources[i],
drivers/net/amt.c
1548
hlist_for_each_entry_safe(snode, t, &gnode->sources[i],
drivers/net/amt.c
258
struct hlist_node *t;
drivers/net/amt.c
275
hlist_for_each_entry_safe(snode, t, &gnode->sources[i], node)
drivers/net/amt.c
419
struct hlist_node *t;
drivers/net/amt.c
433
hlist_for_each_entry_safe(snode, t,
drivers/net/amt.c
96
struct hlist_node *t;
drivers/net/arcnet/arcnet.c
383
static void arcnet_timer(struct timer_list *t)
drivers/net/arcnet/arcnet.c
385
struct arcnet_local *lp = timer_container_of(lp, t, timer);
drivers/net/arcnet/arcnet.c
428
static void arcnet_reply_work(struct work_struct *t)
drivers/net/arcnet/arcnet.c
430
struct arcnet_local *lp = from_work(lp, t, reply_work);
drivers/net/bareudp.c
636
struct bareudp_dev *bareudp, *t = NULL;
drivers/net/bareudp.c
640
t = bareudp;
drivers/net/bareudp.c
642
return t;
drivers/net/bareudp.c
650
struct bareudp_dev *t, *bareudp = netdev_priv(dev);
drivers/net/bareudp.c
655
t = bareudp_find_dev(bn, conf);
drivers/net/bareudp.c
656
if (t) {
drivers/net/caif/caif_virtio.c
601
static void cfv_tx_release_tasklet(struct tasklet_struct *t)
drivers/net/caif/caif_virtio.c
603
struct cfv_info *cfv = from_tasklet(cfv, t, tx_release_tasklet);
drivers/net/can/grcan.c
807
static void grcan_running_reset(struct timer_list *t)
drivers/net/can/grcan.c
809
struct grcan_priv *priv = timer_container_of(priv, t, rr_timer);
drivers/net/can/grcan.c
898
static void grcan_initiate_running_reset(struct timer_list *t)
drivers/net/can/grcan.c
900
struct grcan_priv *priv = timer_container_of(priv, t, hang_timer);
drivers/net/can/sja1000/peak_pcmcia.c
375
static void pcan_led_timer(struct timer_list *t)
drivers/net/can/sja1000/peak_pcmcia.c
377
struct pcan_pccard *card = timer_container_of(card, t, led_timer);
drivers/net/can/spi/hi311x.c
203
struct spi_transfer t = {
drivers/net/can/spi/hi311x.c
213
spi_message_add_tail(&t, &m);
drivers/net/can/spi/mcp251x.c
299
struct spi_transfer t = {
drivers/net/can/spi/mcp251x.c
309
spi_message_add_tail(&t, &m);
drivers/net/can/spi/mcp251xfd/mcp251xfd-ring.c
413
static enum hrtimer_restart mcp251xfd_rx_irq_timer(struct hrtimer *t)
drivers/net/can/spi/mcp251xfd/mcp251xfd-ring.c
415
struct mcp251xfd_priv *priv = container_of(t, struct mcp251xfd_priv,
drivers/net/can/spi/mcp251xfd/mcp251xfd-ring.c
427
static enum hrtimer_restart mcp251xfd_tx_irq_timer(struct hrtimer *t)
drivers/net/can/spi/mcp251xfd/mcp251xfd-ring.c
429
struct mcp251xfd_priv *priv = container_of(t, struct mcp251xfd_priv,
drivers/net/can/usb/peak_usb/pcan_usb.c
320
static void pcan_usb_restart(struct timer_list *t)
drivers/net/can/usb/peak_usb/pcan_usb.c
322
struct pcan_usb *pdev = timer_container_of(pdev, t, restart_timer);
drivers/net/dsa/ks8995.c
252
struct spi_transfer t[2];
drivers/net/dsa/ks8995.c
259
memset(&t, 0, sizeof(t));
drivers/net/dsa/ks8995.c
261
t[0].tx_buf = &cmd;
drivers/net/dsa/ks8995.c
262
t[0].len = sizeof(cmd);
drivers/net/dsa/ks8995.c
263
spi_message_add_tail(&t[0], &m);
drivers/net/dsa/ks8995.c
265
t[1].rx_buf = buf;
drivers/net/dsa/ks8995.c
266
t[1].len = count;
drivers/net/dsa/ks8995.c
267
spi_message_add_tail(&t[1], &m);
drivers/net/dsa/ks8995.c
280
struct spi_transfer t[2];
drivers/net/dsa/ks8995.c
287
memset(&t, 0, sizeof(t));
drivers/net/dsa/ks8995.c
289
t[0].tx_buf = &cmd;
drivers/net/dsa/ks8995.c
290
t[0].len = sizeof(cmd);
drivers/net/dsa/ks8995.c
291
spi_message_add_tail(&t[0], &m);
drivers/net/dsa/ks8995.c
293
t[1].tx_buf = buf;
drivers/net/dsa/ks8995.c
294
t[1].len = count;
drivers/net/dsa/ks8995.c
295
spi_message_add_tail(&t[1], &m);
drivers/net/dsa/lantiq/lantiq_gswip.h
204
#define GSWIP_MAC_CTRL_4_GWAIT(t) u16_encode_bits((t), GSWIP_MAC_CTRL_4_GWAIT_MASK)
drivers/net/dsa/lantiq/lantiq_gswip.h
206
#define GSWIP_MAC_CTRL_4_WAIT(t) u16_encode_bits((t), GSWIP_MAC_CTRL_4_WAIT_MASK)
drivers/net/dsa/mv88e6xxx/phy.c
183
static void mv88e6xxx_phy_ppu_reenable_timer(struct timer_list *t)
drivers/net/dsa/mv88e6xxx/phy.c
185
struct mv88e6xxx_chip *chip = timer_container_of(chip, t, ppu_timer);
drivers/net/dsa/realtek/realtek-smi.c
174
u32 t;
drivers/net/dsa/realtek/realtek-smi.c
177
realtek_smi_read_bits(priv, 8, &t);
drivers/net/dsa/realtek/realtek-smi.c
178
*data = (t & 0xff);
drivers/net/dsa/realtek/realtek-smi.c
188
u32 t;
drivers/net/dsa/realtek/realtek-smi.c
191
realtek_smi_read_bits(priv, 8, &t);
drivers/net/dsa/realtek/realtek-smi.c
192
*data = (t & 0xff);
drivers/net/dsa/sja1105/sja1105_ptp.c
54
#define extts_to_data(t) \
drivers/net/dsa/sja1105/sja1105_ptp.c
55
container_of((t), struct sja1105_ptp_data, extts_timer)
drivers/net/dsa/sja1105/sja1105_ptp.c
681
static void sja1105_ptp_extts_timer(struct timer_list *t)
drivers/net/dsa/sja1105/sja1105_ptp.c
683
struct sja1105_ptp_data *ptp_data = extts_to_data(t);
drivers/net/dsa/vitesse-vsc73xx-spi.c
109
memset(&t, 0, sizeof(t));
drivers/net/dsa/vitesse-vsc73xx-spi.c
111
t[0].tx_buf = cmd;
drivers/net/dsa/vitesse-vsc73xx-spi.c
112
t[0].len = sizeof(cmd);
drivers/net/dsa/vitesse-vsc73xx-spi.c
113
spi_message_add_tail(&t[0], &m);
drivers/net/dsa/vitesse-vsc73xx-spi.c
115
t[1].tx_buf = buf;
drivers/net/dsa/vitesse-vsc73xx-spi.c
116
t[1].len = sizeof(buf);
drivers/net/dsa/vitesse-vsc73xx-spi.c
117
spi_message_add_tail(&t[1], &m);
drivers/net/dsa/vitesse-vsc73xx-spi.c
56
struct spi_transfer t[2];
drivers/net/dsa/vitesse-vsc73xx-spi.c
67
memset(&t, 0, sizeof(t));
drivers/net/dsa/vitesse-vsc73xx-spi.c
69
t[0].tx_buf = cmd;
drivers/net/dsa/vitesse-vsc73xx-spi.c
70
t[0].len = sizeof(cmd);
drivers/net/dsa/vitesse-vsc73xx-spi.c
71
spi_message_add_tail(&t[0], &m);
drivers/net/dsa/vitesse-vsc73xx-spi.c
73
t[1].rx_buf = buf;
drivers/net/dsa/vitesse-vsc73xx-spi.c
74
t[1].len = sizeof(buf);
drivers/net/dsa/vitesse-vsc73xx-spi.c
75
spi_message_add_tail(&t[1], &m);
drivers/net/dsa/vitesse-vsc73xx-spi.c
98
struct spi_transfer t[2];
drivers/net/dsa/xrs700x/xrs700x_reg.h
190
#define XRS_TX_TS_NS_LO(t) (XRS_SWITCH_TS_BASE + 0x80 * (t) + 0x0)
drivers/net/dsa/xrs700x/xrs700x_reg.h
191
#define XRS_TX_TS_NS_HI(t) (XRS_SWITCH_TS_BASE + 0x80 * (t) + 0x2)
drivers/net/dsa/xrs700x/xrs700x_reg.h
192
#define XRS_TX_TS_S_LO(t) (XRS_SWITCH_TS_BASE + 0x80 * (t) + 0x4)
drivers/net/dsa/xrs700x/xrs700x_reg.h
193
#define XRS_TX_TS_S_HI(t) (XRS_SWITCH_TS_BASE + 0x80 * (t) + 0x6)
drivers/net/dsa/xrs700x/xrs700x_reg.h
194
#define XRS_TX_TS_HDR(t, h) (XRS_SWITCH_TS_BASE + 0x80 * (t) + \
drivers/net/dsa/xrs700x/xrs700x_reg.h
196
#define XRS_RX_TS_NS_LO(t) (XRS_SWITCH_TS_BASE + 0x80 * (t) + \
drivers/net/dsa/xrs700x/xrs700x_reg.h
198
#define XRS_RX_TS_NS_HI(t) (XRS_SWITCH_TS_BASE + 0x80 * (t) + \
drivers/net/dsa/xrs700x/xrs700x_reg.h
200
#define XRS_RX_TS_S_LO(t) (XRS_SWITCH_TS_BASE + 0x80 * (t) + \
drivers/net/dsa/xrs700x/xrs700x_reg.h
202
#define XRS_RX_TS_S_HI(t) (XRS_SWITCH_TS_BASE + 0x80 * (t) + \
drivers/net/dsa/xrs700x/xrs700x_reg.h
204
#define XRS_RX_TS_HDR(t, h) (XRS_SWITCH_TS_BASE + 0x80 * (t) + \
drivers/net/eql.c
144
static void eql_timer(struct timer_list *t)
drivers/net/eql.c
146
equalizer_t *eql = timer_container_of(eql, t, timer);
drivers/net/ethernet/3com/3c515.c
363
static void corkscrew_timer(struct timer_list *t);
drivers/net/ethernet/3com/3c515.c
859
static void corkscrew_timer(struct timer_list *t)
drivers/net/ethernet/3com/3c515.c
862
struct corkscrew_private *vp = timer_container_of(vp, t, timer);
drivers/net/ethernet/3com/3c574_cs.c
228
static void media_check(struct timer_list *t);
drivers/net/ethernet/3com/3c574_cs.c
859
static void media_check(struct timer_list *t)
drivers/net/ethernet/3com/3c574_cs.c
861
struct el3_private *lp = timer_container_of(lp, t, media);
drivers/net/ethernet/3com/3c589_cs.c
165
static void media_check(struct timer_list *t);
drivers/net/ethernet/3com/3c589_cs.c
686
static void media_check(struct timer_list *t)
drivers/net/ethernet/3com/3c589_cs.c
688
struct el3_private *lp = timer_container_of(lp, t, media);
drivers/net/ethernet/3com/3c59x.c
1784
vortex_timer(struct timer_list *t)
drivers/net/ethernet/3com/3c59x.c
1786
struct vortex_private *vp = timer_container_of(vp, t, timer);
drivers/net/ethernet/3com/3c59x.c
761
static void vortex_timer(struct timer_list *t);
drivers/net/ethernet/8390/axnet_cs.c
551
static void ei_watchdog(struct timer_list *t)
drivers/net/ethernet/8390/axnet_cs.c
553
struct axnet_dev *info = timer_container_of(info, t, watchdog);
drivers/net/ethernet/8390/axnet_cs.c
88
static void ei_watchdog(struct timer_list *t);
drivers/net/ethernet/8390/pcnet_cs.c
100
static void ei_watchdog(struct timer_list *t);
drivers/net/ethernet/8390/pcnet_cs.c
1019
static void ei_watchdog(struct timer_list *t)
drivers/net/ethernet/8390/pcnet_cs.c
1021
struct pcnet_dev *info = timer_container_of(info, t, watchdog);
drivers/net/ethernet/adi/adin1110.c
199
struct spi_transfer t = {0};
drivers/net/ethernet/adi/adin1110.c
216
t.tx_buf = &priv->data[0];
drivers/net/ethernet/adi/adin1110.c
217
t.rx_buf = &priv->data[0];
drivers/net/ethernet/adi/adin1110.c
218
t.len = read_len + header_len;
drivers/net/ethernet/adi/adin1110.c
220
ret = spi_sync_transfer(priv->spidev, &t, 1);
drivers/net/ethernet/adi/adin1110.c
297
struct spi_transfer t = {0};
drivers/net/ethernet/adi/adin1110.c
344
t.tx_buf = &priv->data[0];
drivers/net/ethernet/adi/adin1110.c
345
t.rx_buf = &rxb->data[0];
drivers/net/ethernet/adi/adin1110.c
346
t.len = header_len + round_len;
drivers/net/ethernet/adi/adin1110.c
348
ret = spi_sync_transfer(priv->spidev, &t, 1);
drivers/net/ethernet/agere/et131x.c
3113
static void et131x_error_timer_handler(struct timer_list *t)
drivers/net/ethernet/agere/et131x.c
3115
struct et131x_adapter *adapter = timer_container_of(adapter, t,
drivers/net/ethernet/airoha/airoha_ppe.c
402
struct airoha_foe_ipv4_tuple *t;
drivers/net/ethernet/airoha/airoha_ppe.c
407
t = &hwe->ipv4.new_tuple;
drivers/net/ethernet/airoha/airoha_ppe.c
413
t = &hwe->ipv4.orig_tuple;
drivers/net/ethernet/airoha/airoha_ppe.c
420
t->src_ip = be32_to_cpu(data->v4.src_addr);
drivers/net/ethernet/airoha/airoha_ppe.c
421
t->dest_ip = be32_to_cpu(data->v4.dst_addr);
drivers/net/ethernet/airoha/airoha_ppe.c
424
t->src_port = be16_to_cpu(data->src_port);
drivers/net/ethernet/airoha/airoha_ppe.c
425
t->dest_port = be16_to_cpu(data->dst_port);
drivers/net/ethernet/alteon/acenic.c
965
int t;
drivers/net/ethernet/alteon/acenic.c
968
t = read_eeprom_byte(dev, 0x8c+i);
drivers/net/ethernet/alteon/acenic.c
969
if (t < 0) {
drivers/net/ethernet/alteon/acenic.c
973
mac1 |= (t & 0xff);
drivers/net/ethernet/alteon/acenic.c
977
int t;
drivers/net/ethernet/alteon/acenic.c
980
t = read_eeprom_byte(dev, 0x8c+i);
drivers/net/ethernet/alteon/acenic.c
981
if (t < 0) {
drivers/net/ethernet/alteon/acenic.c
985
mac2 |= (t & 0xff);
drivers/net/ethernet/amazon/ena/ena_netdev.c
3682
static void ena_timer_service(struct timer_list *t)
drivers/net/ethernet/amazon/ena/ena_netdev.c
3684
struct ena_adapter *adapter = timer_container_of(adapter, t,
drivers/net/ethernet/amd/7990.c
100
t, ib->brx_ring[t].rmd1_hadr, ib->brx_ring[t].rmd0, \
drivers/net/ethernet/amd/7990.c
101
ib->brx_ring[t].length, \
drivers/net/ethernet/amd/7990.c
102
ib->brx_ring[t].mblength, ib->brx_ring[t].rmd1_bits); \
drivers/net/ethernet/amd/7990.c
104
for (t = 0; t < TX_RING_SIZE; t++) { \
drivers/net/ethernet/amd/7990.c
106
t, ib->btx_ring[t].tmd1_hadr, ib->btx_ring[t].tmd0, \
drivers/net/ethernet/amd/7990.c
107
ib->btx_ring[t].length, \
drivers/net/ethernet/amd/7990.c
108
ib->btx_ring[t].misc, ib->btx_ring[t].tmd1_bits); \
drivers/net/ethernet/amd/7990.c
97
int t; \
drivers/net/ethernet/amd/7990.c
98
for (t = 0; t < RX_RING_SIZE; t++) { \
drivers/net/ethernet/amd/a2065.c
637
static void lance_set_multicast_retry(struct timer_list *t)
drivers/net/ethernet/amd/a2065.c
639
struct lance_private *lp = timer_container_of(lp, t, multicast_timer);
drivers/net/ethernet/amd/amd8111e.c
1642
static void amd8111e_config_ipg(struct timer_list *t)
drivers/net/ethernet/amd/amd8111e.c
1644
struct amd8111e_priv *lp = timer_container_of(lp, t,
drivers/net/ethernet/amd/ariadne.c
130
volatile struct TDRE *t = &lancedata->tx_ring[i];
drivers/net/ethernet/amd/ariadne.c
131
t->TMD0 = swloww(ARIADNE_RAM +
drivers/net/ethernet/amd/ariadne.c
133
t->TMD1 = swhighw(ARIADNE_RAM +
drivers/net/ethernet/amd/ariadne.c
136
t->TMD2 = swapw((u_short)-PKT_BUF_SIZE);
drivers/net/ethernet/amd/ariadne.c
137
t->TMD3 = 0;
drivers/net/ethernet/amd/declance.c
1005
static void lance_set_multicast_retry(struct timer_list *t)
drivers/net/ethernet/amd/declance.c
1007
struct lance_private *lp = timer_container_of(lp, t, multicast_timer);
drivers/net/ethernet/amd/pcnet32.c
2902
static void pcnet32_watchdog(struct timer_list *t)
drivers/net/ethernet/amd/pcnet32.c
2904
struct pcnet32_private *lp = timer_container_of(lp, t, watchdog_timer);
drivers/net/ethernet/amd/pds_core/main.c
24
static void pdsc_wdtimer_cb(struct timer_list *t)
drivers/net/ethernet/amd/pds_core/main.c
26
struct pdsc *pdsc = timer_container_of(pdsc, t, wdtimer);
drivers/net/ethernet/amd/sunlance.c
1247
static void lance_set_multicast_retry(struct timer_list *t)
drivers/net/ethernet/amd/sunlance.c
1249
struct lance_private *lp = timer_container_of(lp, t, multicast_timer);
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
566
static void xgbe_tx_timer(struct timer_list *t)
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
568
struct xgbe_channel *channel = timer_container_of(channel, t,
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
605
static void xgbe_service_timer(struct timer_list *t)
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
607
struct xgbe_prv_data *pdata = timer_container_of(pdata, t,
drivers/net/ethernet/apple/bmac.c
1411
static void bmac_tx_timeout(struct timer_list *t)
drivers/net/ethernet/apple/bmac.c
1413
struct bmac_data *bp = timer_container_of(bp, t, tx_timeout);
drivers/net/ethernet/apple/bmac.c
155
static void bmac_tx_timeout(struct timer_list *t);
drivers/net/ethernet/apple/mace.c
806
static void mace_tx_timeout(struct timer_list *t)
drivers/net/ethernet/apple/mace.c
808
struct mace_data *mp = timer_container_of(mp, t, tx_timeout);
drivers/net/ethernet/apple/mace.c
89
static void mace_tx_timeout(struct timer_list *t);
drivers/net/ethernet/aquantia/atlantic/aq_nic.c
255
static void aq_nic_service_timer_cb(struct timer_list *t)
drivers/net/ethernet/aquantia/atlantic/aq_nic.c
257
struct aq_nic_s *self = timer_container_of(self, t, service_timer);
drivers/net/ethernet/aquantia/atlantic/aq_nic.c
265
static void aq_nic_polling_timer_cb(struct timer_list *t)
drivers/net/ethernet/aquantia/atlantic/aq_nic.c
267
struct aq_nic_s *self = timer_container_of(self, t, polling_timer);
drivers/net/ethernet/aquantia/atlantic/aq_ptp.c
393
struct ptp_clock_time *t = &rq->perout.period;
drivers/net/ethernet/aquantia/atlantic/aq_ptp.c
406
if (t->sec > 4 || t->sec < 0)
drivers/net/ethernet/aquantia/atlantic/aq_ptp.c
412
period = on ? t->sec * NSEC_PER_SEC + t->nsec : 0;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
246
unsigned int t)
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
248
return (h < t) ? ((h < i) && (i < t)) : ((h < i) || (i < t));
drivers/net/ethernet/asix/ax88796c_main.c
277
size_t t = max((padlen + TX_EOP_SIZE) - tailroom, 0);
drivers/net/ethernet/asix/ax88796c_main.c
279
if (pskb_expand_head(skb, h, t, GFP_KERNEL))
drivers/net/ethernet/asix/ax88796c_main.c
801
u16 t;
drivers/net/ethernet/asix/ax88796c_main.c
825
t = AX_READ(&ax_local->ax_spi, P1_RXBSPCR);
drivers/net/ethernet/asix/ax88796c_main.c
826
t &= ~RXBSPCR_STUF_ENABLE;
drivers/net/ethernet/asix/ax88796c_main.c
827
AX_WRITE(&ax_local->ax_spi, t, P1_RXBSPCR);
drivers/net/ethernet/asix/ax88796c_main.c
832
t = AX_READ(&ax_local->ax_spi, P0_FER);
drivers/net/ethernet/asix/ax88796c_main.c
833
t |= FER_RXEN | FER_TXEN | FER_BSWAP | FER_IRQ_PULL;
drivers/net/ethernet/asix/ax88796c_main.c
834
AX_WRITE(&ax_local->ax_spi, t, P0_FER);
drivers/net/ethernet/atheros/ag71xx.c
1573
static void ag71xx_oom_timer_handler(struct timer_list *t)
drivers/net/ethernet/atheros/ag71xx.c
1575
struct ag71xx *ag = timer_container_of(ag, t, oom_timer);
drivers/net/ethernet/atheros/ag71xx.c
656
unsigned long t;
drivers/net/ethernet/atheros/ag71xx.c
658
t = ref_clock / table[i];
drivers/net/ethernet/atheros/ag71xx.c
659
if (t <= AG71XX_MDIO_MAX_CLK) {
drivers/net/ethernet/atheros/ag71xx.c
672
u32 t;
drivers/net/ethernet/atheros/ag71xx.c
674
err = ag71xx_mdio_get_divider(ag, &t);
drivers/net/ethernet/atheros/ag71xx.c
678
ag71xx_wr(ag, AG71XX_REG_MII_CFG, t | MII_CFG_RESET);
drivers/net/ethernet/atheros/ag71xx.c
681
ag71xx_wr(ag, AG71XX_REG_MII_CFG, t);
drivers/net/ethernet/atheros/ag71xx.c
931
u32 t;
drivers/net/ethernet/atheros/ag71xx.c
933
t = (((u32)mac[5]) << 24) | (((u32)mac[4]) << 16)
drivers/net/ethernet/atheros/ag71xx.c
936
ag71xx_wr(ag, AG71XX_REG_MAC_ADDR1, t);
drivers/net/ethernet/atheros/ag71xx.c
938
t = (((u32)mac[1]) << 24) | (((u32)mac[0]) << 16);
drivers/net/ethernet/atheros/ag71xx.c
939
ag71xx_wr(ag, AG71XX_REG_MAC_ADDR2, t);
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
232
static void atl1c_phy_config(struct timer_list *t)
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
234
struct atl1c_adapter *adapter = timer_container_of(adapter, t,
drivers/net/ethernet/atheros/atl1e/atl1e_main.c
116
static void atl1e_phy_config(struct timer_list *t)
drivers/net/ethernet/atheros/atl1e/atl1e_main.c
118
struct atl1e_adapter *adapter = timer_container_of(adapter, t,
drivers/net/ethernet/atheros/atlx/atl1.c
2592
static void atl1_phy_config(struct timer_list *t)
drivers/net/ethernet/atheros/atlx/atl1.c
2594
struct atl1_adapter *adapter = timer_container_of(adapter, t,
drivers/net/ethernet/atheros/atlx/atl2.c
1011
static void atl2_watchdog(struct timer_list *t)
drivers/net/ethernet/atheros/atlx/atl2.c
1013
struct atl2_adapter *adapter = timer_container_of(adapter, t,
drivers/net/ethernet/atheros/atlx/atl2.c
1037
static void atl2_phy_config(struct timer_list *t)
drivers/net/ethernet/atheros/atlx/atl2.c
1039
struct atl2_adapter *adapter = timer_container_of(adapter, t,
drivers/net/ethernet/broadcom/b44.c
577
static void b44_timer(struct timer_list *t)
drivers/net/ethernet/broadcom/b44.c
579
struct b44 *bp = timer_container_of(bp, t, timer);
drivers/net/ethernet/broadcom/bcm63xx_enet.c
1382
static void bcm_enet_update_mib_counters_defer(struct work_struct *t)
drivers/net/ethernet/broadcom/bcm63xx_enet.c
1386
priv = container_of(t, struct bcm_enet_priv, mib_update_task);
drivers/net/ethernet/broadcom/bcm63xx_enet.c
2001
static void swphy_poll_timer(struct timer_list *t)
drivers/net/ethernet/broadcom/bcm63xx_enet.c
2003
struct bcm_enet_priv *priv = timer_container_of(priv, t, swphy_poll);
drivers/net/ethernet/broadcom/bcm63xx_enet.c
287
static void bcm_enet_refill_rx_timer(struct timer_list *t)
drivers/net/ethernet/broadcom/bcm63xx_enet.c
289
struct bcm_enet_priv *priv = timer_container_of(priv, t, rx_timeout);
drivers/net/ethernet/broadcom/bnx2.c
6164
bnx2_timer(struct timer_list *t)
drivers/net/ethernet/broadcom/bnx2.c
6166
struct bnx2 *bp = timer_container_of(bp, t, timer);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c
5787
static void bnx2x_timer(struct timer_list *t)
drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c
5789
struct bnx2x *bp = timer_container_of(bp, t, timer);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
394
#define UPDATE_STAT64(s, t) \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
396
DIFF_64(diff.hi, new->s##_hi, pstats->mac_stx[0].t##_hi, \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
397
diff.lo, new->s##_lo, pstats->mac_stx[0].t##_lo); \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
398
pstats->mac_stx[0].t##_hi = new->s##_hi; \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
399
pstats->mac_stx[0].t##_lo = new->s##_lo; \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
400
ADD_64(pstats->mac_stx[1].t##_hi, diff.hi, \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
401
pstats->mac_stx[1].t##_lo, diff.lo); \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
404
#define UPDATE_STAT64_NIG(s, t) \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
408
ADD_64(estats->t##_hi, diff.hi, \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
409
estats->t##_lo, diff.lo); \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
419
#define ADD_STAT64(diff, t) \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
421
ADD_64(pstats->mac_stx[1].t##_hi, new->diff##_hi, \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
422
pstats->mac_stx[1].t##_lo, new->diff##_lo); \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
432
#define UPDATE_EXTEND_TSTAT_X(s, t, size) \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
437
ADD_EXTEND_64(qstats->t##_hi, qstats->t##_lo, diff); \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
440
#define UPDATE_EXTEND_TSTAT(s, t) UPDATE_EXTEND_TSTAT_X(s, t, 32)
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
442
#define UPDATE_EXTEND_E_TSTAT(s, t, size) \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
444
UPDATE_EXTEND_TSTAT_X(s, t, size); \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
445
ADD_EXTEND_64(estats->t##_hi, estats->t##_lo, diff); \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
448
#define UPDATE_EXTEND_USTAT(s, t) \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
452
ADD_EXTEND_64(qstats->t##_hi, qstats->t##_lo, diff); \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
455
#define UPDATE_EXTEND_E_USTAT(s, t) \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
457
UPDATE_EXTEND_USTAT(s, t); \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
458
ADD_EXTEND_64(estats->t##_hi, estats->t##_lo, diff); \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
461
#define UPDATE_EXTEND_XSTAT(s, t) \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
465
ADD_EXTEND_64(qstats->t##_hi, qstats->t##_lo, diff); \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
468
#define UPDATE_QSTAT(s, t) \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
470
qstats->t##_lo = qstats_old->t##_lo + le32_to_cpu(s.lo); \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
471
qstats->t##_hi = qstats_old->t##_hi + le32_to_cpu(s.hi) \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
472
+ ((qstats->t##_lo < qstats_old->t##_lo) ? 1 : 0); \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
519
#define UPDATE_ESTAT(s, t) \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
521
SUB_64(estats->s##_hi, estats_old->t##_hi, \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
522
estats->s##_lo, estats_old->t##_lo); \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
523
ADD_64(estats->s##_hi, estats->t##_hi, \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
524
estats->s##_lo, estats->t##_lo); \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
525
estats_old->t##_hi = estats->t##_hi; \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
526
estats_old->t##_lo = estats->t##_lo; \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
541
#define SUB_EXTEND_USTAT(s, t) \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_stats.h
544
SUB_EXTEND_64(qstats->t##_hi, qstats->t##_lo, diff); \
drivers/net/ethernet/broadcom/bnxt/bnxt.c
14229
static void bnxt_timer(struct timer_list *t)
drivers/net/ethernet/broadcom/bnxt/bnxt.c
14231
struct bnxt *bp = timer_container_of(bp, t, timer);
drivers/net/ethernet/broadcom/tg3.c
11059
static void tg3_timer(struct timer_list *t)
drivers/net/ethernet/broadcom/tg3.c
11061
struct tg3 *tp = timer_container_of(tp, t, timer);
drivers/net/ethernet/brocade/bna/bfa_cs.h
22
#define BFA_SM_TABLE(n, s, e, t) \
drivers/net/ethernet/brocade/bna/bfa_cs.h
25
typedef void (*t)(struct s *, enum e); \
drivers/net/ethernet/brocade/bna/bfa_cs.h
28
t sm; /* state machine function */ \
drivers/net/ethernet/brocade/bna/bfa_cs.h
34
n ## _sm_to_state(struct n ## _sm_table_s *smt, t sm) \
drivers/net/ethernet/brocade/bna/bnad.c
1686
bnad_ioc_timeout(struct timer_list *t)
drivers/net/ethernet/brocade/bna/bnad.c
1688
struct bnad *bnad = timer_container_of(bnad, t,
drivers/net/ethernet/brocade/bna/bnad.c
1698
bnad_ioc_hb_check(struct timer_list *t)
drivers/net/ethernet/brocade/bna/bnad.c
1700
struct bnad *bnad = timer_container_of(bnad, t,
drivers/net/ethernet/brocade/bna/bnad.c
1710
bnad_iocpf_timeout(struct timer_list *t)
drivers/net/ethernet/brocade/bna/bnad.c
1712
struct bnad *bnad = timer_container_of(bnad, t,
drivers/net/ethernet/brocade/bna/bnad.c
1722
bnad_iocpf_sem_timeout(struct timer_list *t)
drivers/net/ethernet/brocade/bna/bnad.c
1724
struct bnad *bnad = timer_container_of(bnad, t,
drivers/net/ethernet/brocade/bna/bnad.c
1745
bnad_dim_timeout(struct timer_list *t)
drivers/net/ethernet/brocade/bna/bnad.c
1747
struct bnad *bnad = timer_container_of(bnad, t, dim_timer);
drivers/net/ethernet/brocade/bna/bnad.c
1778
bnad_stats_timeout(struct timer_list *t)
drivers/net/ethernet/brocade/bna/bnad.c
1780
struct bnad *bnad = timer_container_of(bnad, t, stats_timer);
drivers/net/ethernet/calxeda/xgmac.c
397
#define dma_ring_space(h, t, s) CIRC_SPACE(h, t, s)
drivers/net/ethernet/calxeda/xgmac.c
398
#define dma_ring_cnt(h, t, s) CIRC_CNT(h, t, s)
drivers/net/ethernet/cavium/liquidio/lio_main.c
153
static void octeon_droq_bh(struct tasklet_struct *t)
drivers/net/ethernet/cavium/liquidio/lio_main.c
157
struct octeon_device_priv *oct_priv = from_tasklet(oct_priv, t,
drivers/net/ethernet/cavium/liquidio/octeon_mem_ops.c
202
__be32 t = cpu_to_be32(val);
drivers/net/ethernet/cavium/liquidio/octeon_mem_ops.c
204
__octeon_pci_rw_core_mem(oct, coreaddr, (u8 *)&t, 4, 0);
drivers/net/ethernet/cavium/octeon/octeon_mgmt.c
318
static void octeon_mgmt_clean_tx_tasklet(struct tasklet_struct *t)
drivers/net/ethernet/cavium/octeon/octeon_mgmt.c
320
struct octeon_mgmt *p = from_tasklet(p, t, tx_clean_tasklet);
drivers/net/ethernet/cavium/thunder/nicvf_main.c
985
static void nicvf_handle_qs_err(struct tasklet_struct *t)
drivers/net/ethernet/cavium/thunder/nicvf_main.c
987
struct nicvf *nic = from_tasklet(nic, t, qs_err_task);
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
463
void nicvf_rbdr_task(struct tasklet_struct *t)
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
465
struct nicvf *nic = from_tasklet(nic, t, rbdr_task);
drivers/net/ethernet/cavium/thunder/nicvf_queues.h
351
void nicvf_rbdr_task(struct tasklet_struct *t);
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
459
const struct sge_intr_counts *t;
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
463
t = t1_sge_get_intr_counts(adapter->sge);
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
511
*data++ = t->rx_drops;
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
512
*data++ = t->pure_rsps;
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
513
*data++ = t->unhandled_irqs;
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
514
*data++ = t->respQ_empty;
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
515
*data++ = t->respQ_overflow;
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
516
*data++ = t->freelistQ_empty;
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
517
*data++ = t->pkt_too_big;
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
518
*data++ = t->pkt_mismatch;
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
519
*data++ = t->cmdQ_full[0];
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
520
*data++ = t->cmdQ_full[1];
drivers/net/ethernet/chelsio/cxgb/sge.c
1306
static void restart_sched(struct tasklet_struct *t)
drivers/net/ethernet/chelsio/cxgb/sge.c
1308
struct sched *s = from_tasklet(s, t, sched_tsk);
drivers/net/ethernet/chelsio/cxgb/sge.c
1922
static void sge_tx_reclaim_cb(struct timer_list *t)
drivers/net/ethernet/chelsio/cxgb/sge.c
1925
struct sge *sge = timer_container_of(sge, t, tx_reclaim_timer);
drivers/net/ethernet/chelsio/cxgb/sge.c
2018
static void espibug_workaround_t204(struct timer_list *t)
drivers/net/ethernet/chelsio/cxgb/sge.c
2020
struct sge *sge = timer_container_of(sge, t, espibug_timer);
drivers/net/ethernet/chelsio/cxgb/sge.c
2061
static void espibug_workaround(struct timer_list *t)
drivers/net/ethernet/chelsio/cxgb/sge.c
2063
struct sge *sge = timer_container_of(sge, t, espibug_timer);
drivers/net/ethernet/chelsio/cxgb/sge.c
236
static void restart_sched(struct tasklet_struct *t);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h
101
e = atid2entry(t, tid);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h
102
if ((void *)e->next >= (void *)t->tid_tab &&
drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h
103
(void *)e->next < (void *)&t->atid_tab[t->natids])
drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h
47
static inline union active_open_entry *atid2entry(const struct tid_info *t,
drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h
50
return &t->atid_tab[atid - t->atid_base];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h
53
static inline union listen_entry *stid2entry(const struct tid_info *t,
drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h
56
return &t->stid_tab[stid - t->stid_base];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h
62
static inline struct t3c_tid_entry *lookup_tid(const struct tid_info *t,
drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h
65
struct t3c_tid_entry *t3c_tid = tid < t->ntids ?
drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h
66
&(t->tid_tab[tid]) : NULL;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h
74
static inline struct t3c_tid_entry *lookup_stid(const struct tid_info *t,
drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h
79
if (tid < t->stid_base || tid >= t->stid_base + t->nstids)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h
82
e = stid2entry(t, tid);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h
83
if ((void *)e->next >= (void *)t->tid_tab &&
drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h
84
(void *)e->next < (void *)&t->atid_tab[t->natids])
drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h
93
static inline struct t3c_tid_entry *lookup_atid(const struct tid_info *t,
drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h
98
if (tid < t->atid_base || tid >= t->atid_base + t->natids)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2152
struct ch_qset_params t;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2158
if (copy_from_user(&t, useraddr, sizeof(t)))
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2160
if (t.cmd != CHELSIO_SET_QSET_PARAMS)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2162
if (t.qset_idx >= SGE_QSETS)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2164
if (!cxgb_in_range(t.intr_lat, 0, M_NEWTIMER) ||
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2165
!cxgb_in_range(t.cong_thres, 0, 255) ||
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2166
!cxgb_in_range(t.txq_size[0], MIN_TXQ_ENTRIES,
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2168
!cxgb_in_range(t.txq_size[1], MIN_TXQ_ENTRIES,
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2170
!cxgb_in_range(t.txq_size[2], MIN_CTRL_TXQ_ENTRIES,
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2172
!cxgb_in_range(t.fl_size[0], MIN_FL_ENTRIES,
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2174
!cxgb_in_range(t.fl_size[1], MIN_FL_ENTRIES,
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2176
!cxgb_in_range(t.rspq_size, MIN_RSPQ_ENTRIES,
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2181
(t.rspq_size >= 0 || t.fl_size[0] >= 0 ||
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2182
t.fl_size[1] >= 0 || t.txq_size[0] >= 0 ||
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2183
t.txq_size[1] >= 0 || t.txq_size[2] >= 0 ||
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2184
t.polling >= 0 || t.cong_thres >= 0))
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2196
if (t.qset_idx < q1)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2198
if (t.qset_idx > q1 + nqsets - 1)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2201
q = &adapter->params.sge.qset[t.qset_idx];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2203
if (t.rspq_size >= 0)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2204
q->rspq_size = t.rspq_size;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2205
if (t.fl_size[0] >= 0)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2206
q->fl_size = t.fl_size[0];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2207
if (t.fl_size[1] >= 0)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2208
q->jumbo_size = t.fl_size[1];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2209
if (t.txq_size[0] >= 0)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2210
q->txq_size[0] = t.txq_size[0];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2211
if (t.txq_size[1] >= 0)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2212
q->txq_size[1] = t.txq_size[1];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2213
if (t.txq_size[2] >= 0)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2214
q->txq_size[2] = t.txq_size[2];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2215
if (t.cong_thres >= 0)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2216
q->cong_thres = t.cong_thres;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2217
if (t.intr_lat >= 0) {
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2219
&adapter->sge.qs[t.qset_idx];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2221
q->coalesce_usecs = t.intr_lat;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2224
if (t.polling >= 0) {
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2226
q->polling = t.polling;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2231
t.polling = 0;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2236
q->polling = t.polling;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2241
if (t.lro >= 0) {
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2242
if (t.lro)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2253
struct ch_qset_params t;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2258
if (copy_from_user(&t, useraddr, sizeof(t)))
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2261
if (t.cmd != CHELSIO_GET_QSET_PARAMS)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2273
if (t.qset_idx >= nqsets)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2275
t.qset_idx = array_index_nospec(t.qset_idx, nqsets);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2277
q = &adapter->params.sge.qset[q1 + t.qset_idx];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2278
t.rspq_size = q->rspq_size;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2279
t.txq_size[0] = q->txq_size[0];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2280
t.txq_size[1] = q->txq_size[1];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2281
t.txq_size[2] = q->txq_size[2];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2282
t.fl_size[0] = q->fl_size;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2283
t.fl_size[1] = q->jumbo_size;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2284
t.polling = q->polling;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2285
t.lro = !!(dev->features & NETIF_F_GRO);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2286
t.intr_lat = q->coalesce_usecs;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2287
t.cong_thres = q->cong_thres;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2288
t.qnum = q1;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2291
t.vector = adapter->msix_info[q1 + t.qset_idx + 1].vec;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2293
t.vector = adapter->pdev->irq;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2295
if (copy_to_user(useraddr, &t, sizeof(t)))
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2345
struct ch_mem_range t;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2349
if (copy_from_user(&t, useraddr, sizeof(t)))
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2351
if (t.cmd != CHELSIO_LOAD_FW)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2354
fw_data = memdup_user(useraddr + sizeof(t), t.len);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2358
ret = t3_load_fw(adapter, fw_data, t.len);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2444
struct ch_mem_range t;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2454
if (copy_from_user(&t, useraddr, sizeof(t)))
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2456
if (t.cmd != CHELSIO_GET_MEM)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2458
if ((t.addr & 7) || (t.len & 7))
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2460
if (t.mem_id == MEM_CM)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2462
else if (t.mem_id == MEM_PMRX)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2464
else if (t.mem_id == MEM_PMTX)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2474
t.version = 3 | (adapter->params.rev << 10);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2475
if (copy_to_user(useraddr, &t, sizeof(t)))
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2482
useraddr += sizeof(t); /* advance to start of buffer */
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2483
while (t.len) {
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2485
min_t(unsigned int, t.len, sizeof(buf));
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2488
t3_mc7_bd_read(mem, t.addr / 8, chunk / 8,
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2495
t.addr += chunk;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2496
t.len -= chunk;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2501
struct ch_trace t;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2508
if (copy_from_user(&t, useraddr, sizeof(t)))
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2510
if (t.cmd != CHELSIO_SET_TRACE_FILTER)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2513
tp = (const struct trace_params *)&t.sip;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2514
if (t.config_tx)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2516
t.invert_match,
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2517
t.trace_tx);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2518
if (t.config_rx)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2520
t.invert_match,
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2521
t.trace_rx);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1118
static int init_tid_tabs(struct tid_info *t, unsigned int ntids,
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1122
unsigned long size = ntids * sizeof(*t->tid_tab) +
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1123
natids * sizeof(*t->atid_tab) + nstids * sizeof(*t->stid_tab);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1125
t->tid_tab = kvzalloc(size, GFP_KERNEL);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1126
if (!t->tid_tab)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1129
t->stid_tab = (union listen_entry *)&t->tid_tab[ntids];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1130
t->atid_tab = (union active_open_entry *)&t->stid_tab[nstids];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1131
t->ntids = ntids;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1132
t->nstids = nstids;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1133
t->stid_base = stid_base;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1134
t->sfree = NULL;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1135
t->natids = natids;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1136
t->atid_base = atid_base;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1137
t->afree = NULL;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1138
t->stids_in_use = t->atids_in_use = 0;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1139
atomic_set(&t->tids_in_use, 0);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1140
spin_lock_init(&t->stid_lock);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1141
spin_lock_init(&t->atid_lock);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1148
t->stid_tab[nstids - 1].next = &t->stid_tab[nstids];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1149
t->sfree = t->stid_tab;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1153
t->atid_tab[natids - 1].next = &t->atid_tab[natids];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1154
t->afree = t->atid_tab;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1159
static void free_tid_maps(struct tid_info *t)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1161
kvfree(t->tid_tab);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1182
struct t3c_data *t;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1188
t = kzalloc_obj(*t);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1189
if (!t)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1193
if (dev->ctl(dev, GET_TX_MAX_CHUNK, &t->tx_max_chunk) < 0 ||
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1194
dev->ctl(dev, GET_MAX_OUTSTANDING_WR, &t->max_wrs) < 0 ||
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1207
err = init_tid_tabs(&t->tid_maps, tid_range.num, natids,
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1212
t->mtus = mtutab.mtus;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1213
t->nmtus = mtutab.size;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1215
INIT_WORK(&t->tid_release_task, t3_process_tid_release_list);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1216
spin_lock_init(&t->tid_release_lock);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1217
INIT_LIST_HEAD(&t->list_node);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1218
t->dev = dev;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1221
T3C_DATA(dev) = t;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1229
t->nofail_skb = alloc_skb(sizeof(struct cpl_tid_release), GFP_KERNEL);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1230
t->release_list_incomplete = 0;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1238
kfree(t);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1252
struct t3c_data *t = T3C_DATA(tdev);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1259
free_tid_maps(&t->tid_maps);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1266
kfree_skb(t->nofail_skb);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1267
kfree(t);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
310
struct ch_mem_range *t = data;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
313
if ((t->addr & 7) || (t->len & 7))
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
315
if (t->mem_id == MEM_CM)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
317
else if (t->mem_id == MEM_PMRX)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
319
else if (t->mem_id == MEM_PMTX)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
325
t3_mc7_bd_read(mem, t->addr / 8, t->len / 8,
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
326
(u64 *) t->buf);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
503
struct tid_info *t = &(T3C_DATA(tdev))->tid_maps;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
504
union active_open_entry *p = atid2entry(t, atid);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
507
spin_lock_bh(&t->atid_lock);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
508
p->next = t->afree;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
509
t->afree = p;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
510
t->atids_in_use--;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
511
spin_unlock_bh(&t->atid_lock);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
521
struct tid_info *t = &(T3C_DATA(tdev))->tid_maps;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
523
t->tid_tab[tid].client = client;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
524
t->tid_tab[tid].ctx = ctx;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
525
atomic_inc(&t->tids_in_use);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
612
struct tid_info *t = &(T3C_DATA(tdev))->tid_maps;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
614
BUG_ON(tid >= t->ntids);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
616
(void)cmpxchg(&t->tid_tab[tid].ctx, ctx, NULL);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
624
t->tid_tab[tid].ctx = NULL;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
628
atomic_dec(&t->tids_in_use);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
637
struct tid_info *t = &(T3C_DATA(tdev))->tid_maps;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
639
spin_lock_bh(&t->atid_lock);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
640
if (t->afree &&
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
641
t->atids_in_use + atomic_read(&t->tids_in_use) + MC5_MIN_TIDS <=
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
642
t->ntids) {
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
643
union active_open_entry *p = t->afree;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
645
atid = (p - t->atid_tab) + t->atid_base;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
646
t->afree = p->next;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
649
t->atids_in_use++;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
651
spin_unlock_bh(&t->atid_lock);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
760
struct tid_info *t = &(T3C_DATA(dev))->tid_maps;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
764
if (unlikely(tid >= t->ntids)) {
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
771
t3c_tid = lookup_stid(t, stid);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
854
struct tid_info *t = &(T3C_DATA(dev))->tid_maps;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
858
if (unlikely(tid >= t->ntids)) {
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
865
t3c_tid = lookup_atid(t, atid);
drivers/net/ethernet/chelsio/cxgb3/l2t.h
128
static inline void l2t_release(struct t3cdev *t, struct l2t_entry *e)
drivers/net/ethernet/chelsio/cxgb3/l2t.h
133
d = L2DATA(t);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2907
static void sge_timer_tx(struct timer_list *t)
drivers/net/ethernet/chelsio/cxgb3/sge.c
2909
struct sge_qset *qs = timer_container_of(qs, t, tx_reclaim_timer);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2947
static void sge_timer_rx(struct timer_list *t)
drivers/net/ethernet/chelsio/cxgb3/sge.c
2950
struct sge_qset *qs = timer_container_of(qs, t, rx_reclaim_timer);
drivers/net/ethernet/chelsio/cxgb4/cxgb4.h
2105
void cxgb4_ethofld_restart(struct tasklet_struct *t);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3315
const struct tid_info *t;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3318
t = &adap->tids;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3323
atomic_read(&t->conns_in_use));
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3331
seq_printf(seq, "TID range: %u..%u/%u..%u", t->tid_base,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3333
t->tid_base + t->ntids - 1);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3335
atomic_read(&t->tids_in_use),
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3336
atomic_read(&t->hash_tids_in_use));
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3339
t->aftid_base,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3340
t->aftid_end,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3342
t->tid_base + t->ntids - 1);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3344
atomic_read(&t->tids_in_use),
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3345
atomic_read(&t->hash_tids_in_use));
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3349
t->tid_base + t->ntids - 1);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3351
atomic_read(&t->hash_tids_in_use));
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3353
} else if (t->ntids) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3355
atomic_read(&t->conns_in_use));
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3357
seq_printf(seq, "TID range: %u..%u", t->tid_base,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3358
t->tid_base + t->ntids - 1);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3360
atomic_read(&t->tids_in_use));
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3363
if (t->nstids)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3365
(!t->stid_base &&
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3367
t->stid_base + 1 : t->stid_base,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3368
t->stid_base + t->nstids - 1,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3369
t->stids_in_use - t->v6_stids_in_use,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3370
t->v6_stids_in_use);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3372
if (t->natids)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3374
t->natids - 1, t->atids_in_use);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3375
seq_printf(seq, "FTID range: %u..%u\n", t->ftid_base,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3376
t->ftid_base + t->nftids - 1);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3377
if (t->nsftids)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3379
t->sftid_base, t->sftid_base + t->nsftids - 2,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3380
t->sftids_in_use);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3381
if (t->nhpftids)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3382
seq_printf(seq, "HPFTID range: %u..%u\n", t->hpftid_base,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3383
t->hpftid_base + t->nhpftids - 1);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3384
if (t->neotids)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3386
t->eotid_base, t->eotid_base + t->neotids - 1,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3387
atomic_read(&t->eotids_in_use));
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3388
if (t->ntids)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3429
unsigned long *t;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3432
t = bitmap_zalloc(adap->sge.egr_sz, GFP_KERNEL);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3433
if (!t)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3436
err = bitmap_parse_user(ubuf, count, t, adap->sge.egr_sz);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3438
bitmap_free(t);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3442
bitmap_copy(adap->sge.blocked_fl, t, adap->sge.egr_sz);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3443
bitmap_free(t);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
1631
struct tid_info *t = &adap->tids;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
1633
if (ftid >= t->hpftid_base && ftid < t->hpftid_base + t->nhpftids)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
1634
return &t->hpftid_tab[ftid - t->hpftid_base];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
1636
if (ftid >= t->ftid_base && ftid < t->ftid_base + t->nftids)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
1637
return &t->ftid_tab[ftid - t->ftid_base];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
1639
return lookup_tid(t, ftid);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
1376
struct tid_info *t = &adapter->tids;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
1430
atid = cxgb4_alloc_atid(t, f);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
1504
cxgb4_free_atid(t, atid);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
1739
struct tid_info *t = &adapter->tids;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
1753
if (tid_out_of_range(t, filter_id))
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
1756
f = lookup_tid(t, filter_id);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
1972
struct tid_info *t = &adap->tids;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
1980
f = lookup_tid(t, tid);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
1989
cxgb4_remove_tid(t, 0, tid, 0);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
2001
struct tid_info *t = &adap->tids;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
2009
f = lookup_atid(t, ftid);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
2023
cxgb4_insert_tid(t, f, f->tid, 0);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
2024
cxgb4_free_atid(t, ftid);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
2031
cxgb4_remove_tid(t, 0, tid, 0);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
2067
cxgb4_free_atid(t, ftid);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
445
static bool cxgb4_filter_prio_in_range(struct tid_info *t, u32 idx, u8 nslots,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
461
if (idx < t->nhpftids) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
465
if (test_bit(idx, t->hpftid_bmap))
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
468
next_tab = t->hpftid_tab;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
469
next_ftid = find_next_bit(t->hpftid_bmap, t->nhpftids, idx);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
470
if (next_ftid >= t->nhpftids) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
475
next_ftid = find_first_bit(t->ftid_bmap, t->nftids);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
476
if (next_ftid >= t->nftids)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
479
next_tab = t->ftid_tab;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
487
prev_ftid = find_last_bit(t->hpftid_bmap, idx);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
491
prev_tab = t->hpftid_tab;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
493
idx -= t->nhpftids;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
498
if (test_bit(idx, t->ftid_bmap))
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
501
prev_tab = t->ftid_tab;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
502
prev_ftid = find_last_bit(t->ftid_bmap, idx);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
508
prev_ftid = find_last_bit(t->hpftid_bmap, t->nhpftids);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
509
if (prev_ftid >= t->nhpftids)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
512
prev_tab = t->hpftid_tab;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
521
next_ftid = find_next_bit(t->ftid_bmap, t->nftids, idx);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
522
if (next_ftid >= t->nftids)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
525
next_tab = t->ftid_tab;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
550
struct tid_info *t = &adap->tids;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
598
spin_lock_bh(&t->ftid_lock);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
600
ftid = (tc_prio <= t->nhpftids) ? 0 : t->nhpftids;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
601
max_ftid = t->nftids + t->nhpftids;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
603
if (ftid < t->nhpftids) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
612
if ((t->tc_hash_tids_max_prio &&
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
613
tc_prio > t->tc_hash_tids_max_prio) ||
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
614
(ftid + n) > t->nhpftids) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
615
ftid = t->nhpftids;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
619
bmap = t->hpftid_bmap;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
625
ftid = find_last_bit(t->hpftid_bmap, t->nhpftids);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
626
if (ftid < t->nhpftids) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
627
f = &t->hpftid_tab[ftid];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
635
ftid = find_first_bit(t->ftid_bmap, t->nftids);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
636
if (ftid < t->nftids) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
637
f = &t->ftid_tab[ftid];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
643
ftid = t->nhpftids;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
651
if (t->tc_hash_tids_max_prio &&
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
652
tc_prio < t->tc_hash_tids_max_prio)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
658
bmap = t->ftid_bmap;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
659
bmap_ftid = ftid - t->nhpftids;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
673
if (cxgb4_filter_prio_in_range(t, ftid, n,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
685
spin_unlock_bh(&t->ftid_lock);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
689
static int cxgb4_set_ftid(struct tid_info *t, int fidx, int family,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
692
spin_lock_bh(&t->ftid_lock);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
694
if (test_bit(fidx, t->ftid_bmap)) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
695
spin_unlock_bh(&t->ftid_lock);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
700
__set_bit(fidx, t->ftid_bmap);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
703
bitmap_allocate_region(t->ftid_bmap, fidx, 2);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
705
bitmap_allocate_region(t->ftid_bmap, fidx, 1);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
708
spin_unlock_bh(&t->ftid_lock);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
712
static int cxgb4_set_hpftid(struct tid_info *t, int fidx, int family)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
714
spin_lock_bh(&t->ftid_lock);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
716
if (test_bit(fidx, t->hpftid_bmap)) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
717
spin_unlock_bh(&t->ftid_lock);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
722
__set_bit(fidx, t->hpftid_bmap);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
724
bitmap_allocate_region(t->hpftid_bmap, fidx, 1);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
726
spin_unlock_bh(&t->ftid_lock);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
730
static void cxgb4_clear_ftid(struct tid_info *t, int fidx, int family,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
733
spin_lock_bh(&t->ftid_lock);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
735
__clear_bit(fidx, t->ftid_bmap);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
738
bitmap_release_region(t->ftid_bmap, fidx, 2);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
740
bitmap_release_region(t->ftid_bmap, fidx, 1);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
742
spin_unlock_bh(&t->ftid_lock);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
745
static void cxgb4_clear_hpftid(struct tid_info *t, int fidx, int family)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
747
spin_lock_bh(&t->ftid_lock);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
750
__clear_bit(fidx, t->hpftid_bmap);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
752
bitmap_release_region(t->hpftid_bmap, fidx, 1);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
754
spin_unlock_bh(&t->ftid_lock);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1062
struct sge_eth_txq *t = &s->ethtxq[pi->first_qset];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1091
for (j = 0; j < pi->nqsets; j++, t++, q++) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1092
err = t4_sge_alloc_eth_txq(adap, t, dev,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1593
int cxgb4_alloc_atid(struct tid_info *t, void *data)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1597
spin_lock_bh(&t->atid_lock);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1598
if (t->afree) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1599
union aopen_entry *p = t->afree;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1601
atid = (p - t->atid_tab) + t->atid_base;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1602
t->afree = p->next;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1604
t->atids_in_use++;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1606
spin_unlock_bh(&t->atid_lock);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1614
void cxgb4_free_atid(struct tid_info *t, unsigned int atid)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1616
union aopen_entry *p = &t->atid_tab[atid - t->atid_base];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1618
spin_lock_bh(&t->atid_lock);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1619
p->next = t->afree;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1620
t->afree = p;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1621
t->atids_in_use--;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1622
spin_unlock_bh(&t->atid_lock);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1629
int cxgb4_alloc_stid(struct tid_info *t, int family, void *data)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1633
spin_lock_bh(&t->stid_lock);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1635
stid = find_first_zero_bit(t->stid_bmap, t->nstids);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1636
if (stid < t->nstids)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1637
__set_bit(stid, t->stid_bmap);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1641
stid = bitmap_find_free_region(t->stid_bmap, t->nstids, 1);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1646
t->stid_tab[stid].data = data;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1647
stid += t->stid_base;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1653
t->stids_in_use += 2;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1654
t->v6_stids_in_use += 2;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1656
t->stids_in_use++;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1659
spin_unlock_bh(&t->stid_lock);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1666
int cxgb4_alloc_sftid(struct tid_info *t, int family, void *data)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1670
spin_lock_bh(&t->stid_lock);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1672
stid = find_next_zero_bit(t->stid_bmap,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1673
t->nstids + t->nsftids, t->nstids);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1674
if (stid < (t->nstids + t->nsftids))
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1675
__set_bit(stid, t->stid_bmap);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1682
t->stid_tab[stid].data = data;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1683
stid -= t->nstids;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1684
stid += t->sftid_base;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1685
t->sftids_in_use++;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1687
spin_unlock_bh(&t->stid_lock);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1694
void cxgb4_free_stid(struct tid_info *t, unsigned int stid, int family)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1697
if (t->nsftids && (stid >= t->sftid_base)) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1698
stid -= t->sftid_base;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1699
stid += t->nstids;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1701
stid -= t->stid_base;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1704
spin_lock_bh(&t->stid_lock);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1706
__clear_bit(stid, t->stid_bmap);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1708
bitmap_release_region(t->stid_bmap, stid, 1);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1709
t->stid_tab[stid].data = NULL;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1710
if (stid < t->nstids) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1712
t->stids_in_use -= 2;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1713
t->v6_stids_in_use -= 2;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1715
t->stids_in_use--;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1718
t->sftids_in_use--;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1721
spin_unlock_bh(&t->stid_lock);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1743
static void cxgb4_queue_tid_release(struct tid_info *t, unsigned int chan,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1746
struct adapter *adap = container_of(t, struct adapter, tids);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1747
void **p = &t->tid_tab[tid - t->tid_base];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1796
void cxgb4_remove_tid(struct tid_info *t, unsigned int chan, unsigned int tid,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1799
struct adapter *adap = container_of(t, struct adapter, tids);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1807
if (t->tid_tab[tid - adap->tids.tid_base]) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1808
t->tid_tab[tid - adap->tids.tid_base] = NULL;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1809
atomic_dec(&t->conns_in_use);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1810
if (t->hash_base && (tid >= t->hash_base)) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1812
atomic_sub(2, &t->hash_tids_in_use);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1814
atomic_dec(&t->hash_tids_in_use);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1817
atomic_sub(2, &t->tids_in_use);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1819
atomic_dec(&t->tids_in_use);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1828
cxgb4_queue_tid_release(t, chan, tid);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1835
static int tid_init(struct tid_info *t)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1837
struct adapter *adap = container_of(t, struct adapter, tids);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1838
unsigned int max_ftids = t->nftids + t->nsftids;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1839
unsigned int natids = t->natids;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1846
stid_bmap_size = BITS_TO_LONGS(t->nstids + t->nsftids);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1847
ftid_bmap_size = BITS_TO_LONGS(t->nftids);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1848
hpftid_bmap_size = BITS_TO_LONGS(t->nhpftids);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1849
eotid_bmap_size = BITS_TO_LONGS(t->neotids);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1850
size = t->ntids * sizeof(*t->tid_tab) +
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1851
natids * sizeof(*t->atid_tab) +
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1852
t->nstids * sizeof(*t->stid_tab) +
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1853
t->nsftids * sizeof(*t->stid_tab) +
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1855
t->nhpftids * sizeof(*t->hpftid_tab) +
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1857
max_ftids * sizeof(*t->ftid_tab) +
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1859
t->neotids * sizeof(*t->eotid_tab) +
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1862
t->tid_tab = kvzalloc(size, GFP_KERNEL);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1863
if (!t->tid_tab)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1866
t->atid_tab = (union aopen_entry *)&t->tid_tab[t->ntids];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1867
t->stid_tab = (struct serv_entry *)&t->atid_tab[natids];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1868
t->stid_bmap = (unsigned long *)&t->stid_tab[t->nstids + t->nsftids];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1869
t->hpftid_tab = (struct filter_entry *)&t->stid_bmap[stid_bmap_size];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1870
t->hpftid_bmap = (unsigned long *)&t->hpftid_tab[t->nhpftids];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1871
t->ftid_tab = (struct filter_entry *)&t->hpftid_bmap[hpftid_bmap_size];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1872
t->ftid_bmap = (unsigned long *)&t->ftid_tab[max_ftids];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1873
t->eotid_tab = (struct eotid_entry *)&t->ftid_bmap[ftid_bmap_size];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1874
t->eotid_bmap = (unsigned long *)&t->eotid_tab[t->neotids];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1875
spin_lock_init(&t->stid_lock);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1876
spin_lock_init(&t->atid_lock);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1877
spin_lock_init(&t->ftid_lock);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1879
t->stids_in_use = 0;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1880
t->v6_stids_in_use = 0;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1881
t->sftids_in_use = 0;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1882
t->afree = NULL;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1883
t->atids_in_use = 0;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1884
atomic_set(&t->tids_in_use, 0);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1885
atomic_set(&t->conns_in_use, 0);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1886
atomic_set(&t->hash_tids_in_use, 0);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1887
atomic_set(&t->eotids_in_use, 0);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1892
t->atid_tab[natids - 1].next = &t->atid_tab[natids];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1893
t->afree = t->atid_tab;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1897
bitmap_zero(t->stid_bmap, t->nstids + t->nsftids);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1899
if (!t->stid_base &&
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1901
__set_bit(0, t->stid_bmap);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1903
if (t->neotids)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1904
bitmap_zero(t->eotid_bmap, t->neotids);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1907
if (t->nhpftids)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1908
bitmap_zero(t->hpftid_bmap, t->nhpftids);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1909
bitmap_zero(t->ftid_bmap, t->nftids);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.c
1060
static void ch_flower_stats_cb(struct timer_list *t)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.c
1062
struct adapter *adap = timer_container_of(adap, t, flower_stats_timer);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.c
793
struct tid_info *t = &adap->tids;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.c
798
spin_lock_bh(&t->ftid_lock);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.c
802
if (t->tc_hash_tids_max_prio != tc_prio)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.c
815
fe->fs.tc_prio <= t->tc_hash_tids_max_prio) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.c
816
t->tc_hash_tids_max_prio = fe->fs.tc_prio;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.c
835
t->tc_hash_tids_max_prio = 0;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.c
838
spin_unlock_bh(&t->ftid_lock);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
156
struct cxgb4_tc_u32_table *t;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
184
t = adapter->tc_u32;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
191
if (uhtid != 0x800 && uhtid >= t->size)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
195
if (link_uhtid >= t->size)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
215
if (!t->table[uhtid - 1].link_handle)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
219
link_start = t->table[uhtid - 1].match_field;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
234
if (t->table[link_uhtid - 1].link_handle) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
281
link = &t->table[link_uhtid - 1];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
299
if (uhtid != 0x800 && t->table[uhtid - 1].link_handle) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
301
memcpy(&fs, &t->table[uhtid - 1].fs, sizeof(fs));
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
345
if (uhtid != 0x800 && t->table[uhtid - 1].link_handle)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
346
set_bit(filter_id, t->table[uhtid - 1].tid_map);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
357
struct cxgb4_tc_u32_table *t;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
420
t = adapter->tc_u32;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
427
if (uhtid != 0x800 && uhtid >= t->size)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
432
link = &t->table[uhtid - 1];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
450
for (i = 0; i < t->size; i++) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
451
link = &t->table[i];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
479
struct cxgb4_tc_u32_table *t;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
486
t = adap->tc_u32;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
487
for (i = 0; i < t->size; i++) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
488
struct cxgb4_link *link = &t->table[i];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
498
struct cxgb4_tc_u32_table *t;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
504
t = kvzalloc_flex(*t, table, max_tids);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
505
if (!t)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
508
t->size = max_tids;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
510
for (i = 0; i < t->size; i++) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
511
struct cxgb4_link *link = &t->table[i];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
522
return t;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
525
for (i = 0; i < t->size; i++) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
526
struct cxgb4_link *link = &t->table[i];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c
529
kvfree(t);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c
666
const struct tid_info *t = &adap->tids;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c
668
return (atomic_read(&t->conns_in_use) || t->stids_in_use);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
166
static inline void *lookup_tid(const struct tid_info *t, unsigned int tid)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
168
tid -= t->tid_base;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
169
return tid < t->ntids ? t->tid_tab[tid] : NULL;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
172
static inline bool tid_out_of_range(const struct tid_info *t, unsigned int tid)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
174
return ((tid - t->tid_base) >= t->ntids);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
177
static inline void *lookup_atid(const struct tid_info *t, unsigned int atid)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
179
return atid < t->natids ? t->atid_tab[atid].data : NULL;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
182
static inline void *lookup_stid(const struct tid_info *t, unsigned int stid)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
185
if (t->nsftids && (stid >= t->sftid_base)) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
186
stid -= t->sftid_base;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
187
stid += t->nstids;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
189
stid -= t->stid_base;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
192
return stid < (t->nstids + t->nsftids) ? t->stid_tab[stid].data : NULL;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
195
static inline void cxgb4_insert_tid(struct tid_info *t, void *data,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
198
t->tid_tab[tid - t->tid_base] = data;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
199
if (t->hash_base && (tid >= t->hash_base)) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
201
atomic_add(2, &t->hash_tids_in_use);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
203
atomic_inc(&t->hash_tids_in_use);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
206
atomic_add(2, &t->tids_in_use);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
208
atomic_inc(&t->tids_in_use);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
210
atomic_inc(&t->conns_in_use);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
213
static inline struct eotid_entry *cxgb4_lookup_eotid(struct tid_info *t,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
216
return eotid < t->neotids ? &t->eotid_tab[eotid] : NULL;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
219
static inline int cxgb4_get_free_eotid(struct tid_info *t)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
223
eotid = find_first_zero_bit(t->eotid_bmap, t->neotids);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
224
if (eotid >= t->neotids)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
230
static inline void cxgb4_alloc_eotid(struct tid_info *t, u32 eotid, void *data)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
232
set_bit(eotid, t->eotid_bmap);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
233
t->eotid_tab[eotid].data = data;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
234
atomic_inc(&t->eotids_in_use);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
237
static inline void cxgb4_free_eotid(struct tid_info *t, u32 eotid)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
239
clear_bit(eotid, t->eotid_bmap);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
240
t->eotid_tab[eotid].data = NULL;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
241
atomic_dec(&t->eotids_in_use);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
244
int cxgb4_alloc_atid(struct tid_info *t, void *data);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
245
int cxgb4_alloc_stid(struct tid_info *t, int family, void *data);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
246
int cxgb4_alloc_sftid(struct tid_info *t, int family, void *data);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
247
void cxgb4_free_atid(struct tid_info *t, unsigned int atid);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
248
void cxgb4_free_stid(struct tid_info *t, unsigned int stid, int family);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
249
void cxgb4_remove_tid(struct tid_info *t, unsigned int qid, unsigned int tid,
drivers/net/ethernet/chelsio/cxgb4/sge.c
2777
static void restart_ctrlq(struct tasklet_struct *t)
drivers/net/ethernet/chelsio/cxgb4/sge.c
2781
struct sge_ctrl_txq *q = from_tasklet(q, t, qresume_tsk);
drivers/net/ethernet/chelsio/cxgb4/sge.c
3083
static void restart_ofldq(struct tasklet_struct *t)
drivers/net/ethernet/chelsio/cxgb4/sge.c
3085
struct sge_uld_txq *q = from_tasklet(q, t, qresume_tsk);
drivers/net/ethernet/chelsio/cxgb4/sge.c
4024
void cxgb4_ethofld_restart(struct tasklet_struct *t)
drivers/net/ethernet/chelsio/cxgb4/sge.c
4026
struct sge_eosw_txq *eosw_txq = from_tasklet(eosw_txq, t,
drivers/net/ethernet/chelsio/cxgb4/sge.c
4233
static void sge_rx_timer_cb(struct timer_list *t)
drivers/net/ethernet/chelsio/cxgb4/sge.c
4237
struct adapter *adap = timer_container_of(adap, t, sge.rx_timer);
drivers/net/ethernet/chelsio/cxgb4/sge.c
4270
static void sge_tx_timer_cb(struct timer_list *t)
drivers/net/ethernet/chelsio/cxgb4/sge.c
4272
struct adapter *adap = timer_container_of(adap, t, sge.tx_timer);
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
2063
static void sge_rx_timer_cb(struct timer_list *t)
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
2065
struct adapter *adapter = timer_container_of(adapter, t, sge.rx_timer);
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
2122
static void sge_tx_timer_cb(struct timer_list *t)
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
2124
struct adapter *adapter = timer_container_of(adapter, t, sge.tx_timer);
drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/chcr_ktls.c
279
struct tid_info *t = &tx_info->adap->tids;
drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/chcr_ktls.c
282
atid = cxgb4_alloc_atid(t, tx_info);
drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/chcr_ktls.c
316
cxgb4_free_atid(t, atid);
drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/chcr_ktls.c
650
struct tid_info *t;
drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/chcr_ktls.c
657
t = &adap->tids;
drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/chcr_ktls.c
658
tx_info = lookup_atid(t, atid);
drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/chcr_ktls.c
665
cxgb4_free_atid(t, atid);
drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/chcr_ktls.c
682
cxgb4_insert_tid(t, tx_info, tx_info->tid, tx_info->ip_family);
drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/chcr_ktls.c
715
struct tid_info *t;
drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/chcr_ktls.c
720
t = &adap->tids;
drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/chcr_ktls.c
721
tx_info = lookup_tid(t, tid);
drivers/net/ethernet/cisco/enic/enic_clsf.c
126
void enic_flow_may_expire(struct timer_list *t)
drivers/net/ethernet/cisco/enic/enic_clsf.c
128
struct enic *enic = timer_container_of(enic, t, rfs_h.rfs_may_expire);
drivers/net/ethernet/cisco/enic/enic_clsf.h
19
void enic_flow_may_expire(struct timer_list *t);
drivers/net/ethernet/cisco/enic/enic_main.c
1511
static void enic_notify_timer(struct timer_list *t)
drivers/net/ethernet/cisco/enic/enic_main.c
1513
struct enic *enic = timer_container_of(enic, t, notify_timer);
drivers/net/ethernet/dec/tulip/de2104x.c
1045
static void de21041_media_timer (struct timer_list *t)
drivers/net/ethernet/dec/tulip/de2104x.c
1047
struct de_private *de = timer_container_of(de, t, media_timer);
drivers/net/ethernet/dec/tulip/de2104x.c
330
static void de21040_media_timer (struct timer_list *t);
drivers/net/ethernet/dec/tulip/de2104x.c
331
static void de21041_media_timer (struct timer_list *t);
drivers/net/ethernet/dec/tulip/de2104x.c
964
static void de21040_media_timer (struct timer_list *t)
drivers/net/ethernet/dec/tulip/de2104x.c
966
struct de_private *de = timer_container_of(de, t, media_timer);
drivers/net/ethernet/dec/tulip/dmfe.c
1116
static void dmfe_timer(struct timer_list *t)
drivers/net/ethernet/dec/tulip/dmfe.c
1118
struct dmfe_board_info *db = timer_container_of(db, t, timer);
drivers/net/ethernet/dec/tulip/interrupt.c
105
void oom_timer(struct timer_list *t)
drivers/net/ethernet/dec/tulip/interrupt.c
107
struct tulip_private *tp = timer_container_of(tp, t, oom_timer);
drivers/net/ethernet/dec/tulip/pnic.c
87
void pnic_timer(struct timer_list *t)
drivers/net/ethernet/dec/tulip/pnic.c
89
struct tulip_private *tp = timer_container_of(tp, t, timer);
drivers/net/ethernet/dec/tulip/pnic2.c
79
void pnic2_timer(struct timer_list *t)
drivers/net/ethernet/dec/tulip/pnic2.c
81
struct tulip_private *tp = timer_container_of(tp, t, timer);
drivers/net/ethernet/dec/tulip/timer.c
140
void mxic_timer(struct timer_list *t)
drivers/net/ethernet/dec/tulip/timer.c
142
struct tulip_private *tp = timer_container_of(tp, t, timer);
drivers/net/ethernet/dec/tulip/timer.c
157
void comet_timer(struct timer_list *t)
drivers/net/ethernet/dec/tulip/timer.c
159
struct tulip_private *tp = timer_container_of(tp, t, timer);
drivers/net/ethernet/dec/tulip/tulip.h
479
void pnic2_timer(struct timer_list *t);
drivers/net/ethernet/dec/tulip/tulip.h
506
void pnic_timer(struct timer_list *t);
drivers/net/ethernet/dec/tulip/tulip.h
510
void mxic_timer(struct timer_list *t);
drivers/net/ethernet/dec/tulip/tulip.h
511
void comet_timer(struct timer_list *t);
drivers/net/ethernet/dec/tulip/tulip.h
518
void oom_timer(struct timer_list *t);
drivers/net/ethernet/dec/tulip/tulip_core.c
115
static void tulip_timer(struct timer_list *t)
drivers/net/ethernet/dec/tulip/tulip_core.c
117
struct tulip_private *tp = timer_container_of(tp, t, timer);
drivers/net/ethernet/dec/tulip/uli526x.c
1015
static void uli526x_timer(struct timer_list *t)
drivers/net/ethernet/dec/tulip/uli526x.c
1017
struct uli526x_board_info *db = timer_container_of(db, t, timer);
drivers/net/ethernet/dec/tulip/uli526x.c
230
static void uli526x_timer(struct timer_list *t);
drivers/net/ethernet/dec/tulip/winbond-840.c
320
static void netdev_timer(struct timer_list *t);
drivers/net/ethernet/dec/tulip/winbond-840.c
739
int t;
drivers/net/ethernet/dec/tulip/winbond-840.c
741
t = (csr5 >> 17) & 0x07;
drivers/net/ethernet/dec/tulip/winbond-840.c
742
if (t==0||t==1) {
drivers/net/ethernet/dec/tulip/winbond-840.c
744
t = (csr5 >> 20) & 0x07;
drivers/net/ethernet/dec/tulip/winbond-840.c
745
if (t==0||t==1)
drivers/net/ethernet/dec/tulip/winbond-840.c
764
static void netdev_timer(struct timer_list *t)
drivers/net/ethernet/dec/tulip/winbond-840.c
766
struct netdev_private *np = timer_container_of(np, t, timer);
drivers/net/ethernet/dlink/dl2k.c
63
static void rio_timer (struct timer_list *t);
drivers/net/ethernet/dlink/dl2k.c
668
rio_timer (struct timer_list *t)
drivers/net/ethernet/dlink/dl2k.c
670
struct netdev_private *np = timer_container_of(np, t, timer);
drivers/net/ethernet/dlink/sundance.c
1062
static void tx_poll(struct tasklet_struct *t)
drivers/net/ethernet/dlink/sundance.c
1064
struct netdev_private *np = from_tasklet(np, t, tx_tasklet);
drivers/net/ethernet/dlink/sundance.c
1321
static void rx_poll(struct tasklet_struct *t)
drivers/net/ethernet/dlink/sundance.c
1323
struct netdev_private *np = from_tasklet(np, t, rx_tasklet);
drivers/net/ethernet/dlink/sundance.c
427
static void netdev_timer(struct timer_list *t);
drivers/net/ethernet/dlink/sundance.c
433
static void rx_poll(struct tasklet_struct *t);
drivers/net/ethernet/dlink/sundance.c
434
static void tx_poll(struct tasklet_struct *t);
drivers/net/ethernet/dlink/sundance.c
943
static void netdev_timer(struct timer_list *t)
drivers/net/ethernet/dlink/sundance.c
945
struct netdev_private *np = timer_container_of(np, t, timer);
drivers/net/ethernet/fealnx.c
1075
static void netdev_timer(struct timer_list *t)
drivers/net/ethernet/fealnx.c
1077
struct netdev_private *np = timer_container_of(np, t, timer);
drivers/net/ethernet/fealnx.c
1164
static void reset_timer(struct timer_list *t)
drivers/net/ethernet/fealnx.c
1166
struct netdev_private *np = timer_container_of(np, t, reset_timer);
drivers/net/ethernet/fealnx.c
422
static void netdev_timer(struct timer_list *t);
drivers/net/ethernet/fealnx.c
423
static void reset_timer(struct timer_list *t);
drivers/net/ethernet/freescale/enetc/enetc_hw.h
116
#define ENETC_BDR(t, i, r) (0x8000 + (t) * 0x100 + ENETC_BDR_OFF(i) + (r))
drivers/net/ethernet/freescale/enetc/enetc_hw.h
548
#define enetc_bdr_rd(hw, t, n, off) \
drivers/net/ethernet/freescale/enetc/enetc_hw.h
549
enetc_rd(hw, ENETC_BDR(t, n, off))
drivers/net/ethernet/freescale/enetc/enetc_hw.h
550
#define enetc_bdr_wr(hw, t, n, off, val) \
drivers/net/ethernet/freescale/enetc/enetc_hw.h
551
enetc_wr(hw, ENETC_BDR(t, n, off), val)
drivers/net/ethernet/google/gve/gve_main.c
270
static void gve_stats_report_timer(struct timer_list *t)
drivers/net/ethernet/google/gve/gve_main.c
272
struct gve_priv *priv = timer_container_of(priv, t,
drivers/net/ethernet/hisilicon/hns/hns_enet.c
2076
static void hns_nic_service_timer(struct timer_list *t)
drivers/net/ethernet/hisilicon/hns/hns_enet.c
2078
struct hns_nic_priv *priv = timer_container_of(priv, t, service_timer);
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
111
#define HNS3_RX_PTYPE_ENTRY(ptype, l, s, t, h) \
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
115
HNS3_L3_TYPE_##t, \
drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c
4504
static void hclge_reset_timer(struct timer_list *t)
drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c
4506
struct hclge_dev *hdev = timer_container_of(hdev, t, reset_timer);
drivers/net/ethernet/hisilicon/hns3/hns3vf/hclgevf_main.c
2058
static void hclgevf_reset_timer(struct timer_list *t)
drivers/net/ethernet/hisilicon/hns3/hns3vf/hclgevf_main.c
2060
struct hclgevf_dev *hdev = timer_container_of(hdev, t, reset_timer);
drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c
374
static void ceq_tasklet(struct tasklet_struct *t)
drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c
376
struct hinic_eq *ceq = from_tasklet(ceq, t, ceq_tasklet);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1219
static void ehea_neq_tasklet(struct tasklet_struct *t)
drivers/net/ethernet/ibm/ehea/ehea_main.c
1221
struct ehea_adapter *adapter = from_tasklet(adapter, t, neq_tasklet);
drivers/net/ethernet/ibm/ibmvnic.c
6165
static void ibmvnic_tasklet(struct tasklet_struct *t)
drivers/net/ethernet/ibm/ibmvnic.c
6167
struct ibmvnic_adapter *adapter = from_tasklet(adapter, t, tasklet);
drivers/net/ethernet/intel/e100.c
1683
static void e100_watchdog(struct timer_list *t)
drivers/net/ethernet/intel/e100.c
1685
struct nic *nic = timer_container_of(nic, t, watchdog);
drivers/net/ethernet/intel/e1000e/netdev.c
4860
static void e1000_update_phy_info(struct timer_list *t)
drivers/net/ethernet/intel/e1000e/netdev.c
4862
struct e1000_adapter *adapter = timer_container_of(adapter, t,
drivers/net/ethernet/intel/e1000e/netdev.c
5197
static void e1000_watchdog(struct timer_list *t)
drivers/net/ethernet/intel/e1000e/netdev.c
5199
struct e1000_adapter *adapter = timer_container_of(adapter, t,
drivers/net/ethernet/intel/fm10k/fm10k_pci.c
200
static void fm10k_service_timer(struct timer_list *t)
drivers/net/ethernet/intel/fm10k/fm10k_pci.c
202
struct fm10k_intfc *interface = timer_container_of(interface, t,
drivers/net/ethernet/intel/i40e/i40e_main.c
11390
static void i40e_service_timer(struct timer_list *t)
drivers/net/ethernet/intel/i40e/i40e_main.c
11392
struct i40e_pf *pf = timer_container_of(pf, t, service_timer);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
1142
hw->blk[blk].xlt2.t[vsi] = vsig;
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
1240
if (memcmp(&es->t[off], fv, es->fvw * sizeof(*fv)))
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
1735
memset(&hw->blk[blk].es.t[off], 0,
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
1739
memcpy(&hw->blk[blk].es.t[off], fv,
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
1825
ptg = hw->blk[blk].xlt1.t[pt];
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
1845
vsig = hw->blk[blk].xlt2.t[vsi];
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
1918
sizeof(*hw->blk[block_id].xlt1.t);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
1919
dst = hw->blk[block_id].xlt1.t;
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
1921
sizeof(*hw->blk[block_id].xlt1.t);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
1931
sizeof(*hw->blk[block_id].xlt2.t);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
1932
dst = (u8 *)hw->blk[block_id].xlt2.t;
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
1934
sizeof(*hw->blk[block_id].xlt2.t);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
1944
sizeof(*hw->blk[block_id].prof.t);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
1945
dst = (u8 *)hw->blk[block_id].prof.t;
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
1947
sizeof(*hw->blk[block_id].prof.t);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
1957
sizeof(*hw->blk[block_id].prof_redir.t);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
1958
dst = hw->blk[block_id].prof_redir.t;
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
1960
sizeof(*hw->blk[block_id].prof_redir.t);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
1971
sizeof(*hw->blk[block_id].es.t);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
1972
dst = (u8 *)hw->blk[block_id].es.t;
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
1975
sizeof(*hw->blk[block_id].es.t);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
2056
struct ice_flow_entry *e, *t;
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
2058
list_for_each_entry_safe(e, t, &p->entries, l_entry)
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
2116
devm_kfree(ice_hw_to_dev(hw), hw->blk[i].xlt1.t);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
2117
devm_kfree(ice_hw_to_dev(hw), hw->blk[i].xlt2.t);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
2120
devm_kfree(ice_hw_to_dev(hw), hw->blk[i].prof.t);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
2121
devm_kfree(ice_hw_to_dev(hw), hw->blk[i].prof_redir.t);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
2122
devm_kfree(ice_hw_to_dev(hw), hw->blk[i].es.t);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
2176
memset(xlt1->t, 0, xlt1->count * sizeof(*xlt1->t));
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
2181
memset(xlt2->t, 0, xlt2->count * sizeof(*xlt2->t));
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
2183
memset(prof->t, 0, prof->count * sizeof(*prof->t));
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
2184
memset(prof_redir->t, 0,
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
2185
prof_redir->count * sizeof(*prof_redir->t));
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
2187
memset(es->t, 0, es->count * sizeof(*es->t) * es->fvw);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
2244
xlt1->t = devm_kcalloc(ice_hw_to_dev(hw), xlt1->count,
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
2245
sizeof(*xlt1->t), GFP_KERNEL);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
2246
if (!xlt1->t)
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
2267
xlt2->t = devm_kcalloc(ice_hw_to_dev(hw), xlt2->count,
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
2268
sizeof(*xlt2->t), GFP_KERNEL);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
2269
if (!xlt2->t)
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
2276
prof->t = devm_kcalloc(ice_hw_to_dev(hw), prof->count,
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
2277
sizeof(*prof->t), GFP_KERNEL);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
2279
if (!prof->t)
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
2284
prof_redir->t = devm_kcalloc(ice_hw_to_dev(hw),
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
2286
sizeof(*prof_redir->t),
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
2289
if (!prof_redir->t)
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
2295
es->t = devm_kcalloc(ice_hw_to_dev(hw),
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
2297
sizeof(*es->t), GFP_KERNEL);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
2298
if (!es->t)
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
2415
dc_msk, nm_msk, hw->blk[blk].prof.t[idx].key);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
2417
hw->blk[blk].prof.t[idx].addr = cpu_to_le16(idx);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
2418
hw->blk[blk].prof.t[idx].prof_id = prof_id;
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
2506
memcpy(p->es, &hw->blk[blk].es.t[off], vec_size);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
2542
&hw->blk[blk].prof.t[tmp->tcam_idx].key,
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
2543
sizeof(hw->blk[blk].prof.t->key));
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3247
struct ice_vsig_prof *d, *t;
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3250
list_for_each_entry_safe(d, t,
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3304
struct ice_vsig_prof *p, *t;
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3306
list_for_each_entry_safe(p, t,
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3732
struct ice_vsig_prof *t;
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3753
list_for_each_entry(t, &hw->blk[blk].xlt2.vsig_tbl[idx].prop_lst,
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3757
for (i = 0; i < t->tcam_count; i++) {
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3763
used = ice_ptg_attr_in_use(&t->tcam[i], ptgs_used,
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3766
if (used && t->tcam[i].in_use) {
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3773
&t->tcam[i],
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3777
} else if (!used && !t->tcam[i].in_use) {
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3783
&t->tcam[i],
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3790
set_bit(t->tcam[i].ptg, ptgs_used);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3792
attr_used[attr_used_cnt++] = &t->tcam[i];
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3821
struct ice_vsig_prof *t;
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3831
t = devm_kzalloc(ice_hw_to_dev(hw), sizeof(*t), GFP_KERNEL);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3832
if (!t)
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3843
t->profile_cookie = map->profile_cookie;
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3844
t->prof_id = map->prof_id;
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3845
t->tcam_count = map->ptg_cnt;
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3870
t->tcam[i].ptg = map->ptg[i];
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3871
t->tcam[i].prof_id = map->prof_id;
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3872
t->tcam[i].tcam_idx = tcam_idx;
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3873
t->tcam[i].attr = map->attr[i];
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3874
t->tcam[i].in_use = true;
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3878
p->prof_id = t->tcam[i].prof_id;
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3879
p->ptg = t->tcam[i].ptg;
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3881
p->tcam_idx = t->tcam[i].tcam_idx;
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3884
ice_set_tcam_flags(t->tcam[i].attr.mask, dc_msk);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3887
status = ice_tcam_write_entry(hw, blk, t->tcam[i].tcam_idx,
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3888
t->tcam[i].prof_id,
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3889
t->tcam[i].ptg, vsig, 0,
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3890
t->tcam[i].attr.flags, vl_msk,
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3904
list_add_tail(&t->list,
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3907
list_add(&t->list,
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3916
devm_kfree(ice_hw_to_dev(hw), t);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3983
struct ice_vsig_prof *t;
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3995
list_for_each_entry(t, lst, list) {
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
3997
status = ice_add_prof_id_vsig(hw, blk, vsig, t->profile_cookie,
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
4018
struct ice_vsig_prof *t;
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
4024
t = kzalloc_obj(*t);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
4025
if (!t)
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
4028
t->profile_cookie = hdl;
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
4029
list_add(&t->list, &lst);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
4033
list_del(&t->list);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
4034
kfree(t);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
637
fv_ext = hw->blk[blk].es.t + (prof * hw->blk[blk].es.fvw);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
774
hw->blk[blk].xlt1.t[ptype] = ptg;
drivers/net/ethernet/intel/ice/ice_flex_type.h
149
struct ice_fv_word *t;
drivers/net/ethernet/intel/ice/ice_flex_type.h
228
u8 *t;
drivers/net/ethernet/intel/ice/ice_flex_type.h
252
u16 *t;
drivers/net/ethernet/intel/ice/ice_flex_type.h
285
struct ice_prof_tcam_entry *t;
drivers/net/ethernet/intel/ice/ice_flex_type.h
290
u8 *t;
drivers/net/ethernet/intel/ice/ice_flow.c
1545
struct ice_flow_entry *e, *t;
drivers/net/ethernet/intel/ice/ice_flow.c
1549
list_for_each_entry_safe(e, t, &prof->entries, l_entry) {
drivers/net/ethernet/intel/ice/ice_flow.c
1955
enum ice_flow_fld_match_type t = range ?
drivers/net/ethernet/intel/ice/ice_flow.c
1958
ice_flow_set_fld_ext(seg, fld, t, val_loc, mask_loc, last_loc);
drivers/net/ethernet/intel/ice/ice_flow.c
2024
struct ice_flow_entry *e, *t;
drivers/net/ethernet/intel/ice/ice_flow.c
2027
list_for_each_entry_safe(e, t, &prof->entries, l_entry) {
drivers/net/ethernet/intel/ice/ice_flow.c
2150
struct ice_flow_prof *p, *t;
drivers/net/ethernet/intel/ice/ice_flow.c
2160
list_for_each_entry_safe(p, t, &hw->fl_profs[blk], l_entry)
drivers/net/ethernet/intel/ice/ice_main.c
1726
static void ice_service_timer(struct timer_list *t)
drivers/net/ethernet/intel/ice/ice_main.c
1728
struct ice_pf *pf = timer_container_of(pf, t, serv_tmr);
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5615
static const struct ice_cgu_pin_desc *t;
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5620
t = ice_e823_zl_cgu_inputs;
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5623
t = ice_e823_zl_cgu_outputs;
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5629
t = ice_e823_si_cgu_inputs;
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5632
t = ice_e823_si_cgu_outputs;
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5636
t = NULL;
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5640
return t;
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5654
const struct ice_cgu_pin_desc *t = NULL;
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5659
t = ice_e810t_sfp_cgu_inputs;
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5662
t = ice_e810t_sfp_cgu_outputs;
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5668
t = ice_e810t_qsfp_cgu_inputs;
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5671
t = ice_e810t_qsfp_cgu_outputs;
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5685
t = ice_cgu_get_pin_desc_e823(hw, input, size);
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5691
return t;
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5703
const struct ice_cgu_pin_desc *t;
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5706
t = ice_cgu_get_pin_desc(hw, input, &size);
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5707
if (t)
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5723
const struct ice_cgu_pin_desc *t;
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5726
t = ice_cgu_get_pin_desc(hw, input, &t_size);
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5728
if (!t)
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5734
return t[pin].type;
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5751
const struct ice_cgu_pin_desc *t;
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5755
t = ice_cgu_get_pin_desc(hw, input, &t_size);
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5756
if (!t)
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5760
*num = t[pin].freq_supp_num;
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5762
return t[pin].freq_supp;
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5777
const struct ice_cgu_pin_desc *t;
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5780
t = ice_cgu_get_pin_desc(hw, input, &t_size);
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5782
if (!t)
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5788
return t[pin].name;
drivers/net/ethernet/intel/ice/virt/fdir.c
1451
static void ice_vf_fdir_timer(struct timer_list *t)
drivers/net/ethernet/intel/ice/virt/fdir.c
1453
struct ice_vf_fdir_ctx *ctx_irq = timer_container_of(ctx_irq, t,
drivers/net/ethernet/intel/ice/virt/fdir.c
1993
ptg = hw->blk[ICE_BLK_FD].xlt1.t[id];
drivers/net/ethernet/intel/ice/virt/fdir.c
2265
ptg = hw->blk[ICE_BLK_FD].xlt1.t[id];
drivers/net/ethernet/intel/ice/virt/rss.c
1451
ptg = hw->blk[ICE_BLK_RSS].xlt1.t[id];
drivers/net/ethernet/intel/ice/virt/rss.c
1526
ptg = hw->blk[ICE_BLK_RSS].xlt1.t[id];
drivers/net/ethernet/intel/igb/igb_main.c
5463
static void igb_update_phy_info(struct timer_list *t)
drivers/net/ethernet/intel/igb/igb_main.c
5465
struct igb_adapter *adapter = timer_container_of(adapter, t,
drivers/net/ethernet/intel/igb/igb_main.c
5554
static void igb_watchdog(struct timer_list *t)
drivers/net/ethernet/intel/igb/igb_main.c
5556
struct igb_adapter *adapter = timer_container_of(adapter, t,
drivers/net/ethernet/intel/igbvf/netdev.c
1892
static void igbvf_watchdog(struct timer_list *t)
drivers/net/ethernet/intel/igbvf/netdev.c
1894
struct igbvf_adapter *adapter = timer_container_of(adapter, t,
drivers/net/ethernet/intel/igc/igc_main.c
5781
static void igc_update_phy_info(struct timer_list *t)
drivers/net/ethernet/intel/igc/igc_main.c
5783
struct igc_adapter *adapter = timer_container_of(adapter, t,
drivers/net/ethernet/intel/igc/igc_main.c
5824
static void igc_watchdog(struct timer_list *t)
drivers/net/ethernet/intel/igc/igc_main.c
5826
struct igc_adapter *adapter = timer_container_of(adapter, t,
drivers/net/ethernet/intel/ixgbe/ixgbe_ipsec.c
320
struct tx_sa *t = &ipsec->tx_tbl[i];
drivers/net/ethernet/intel/ixgbe/ixgbe_ipsec.c
331
if (t->used) {
drivers/net/ethernet/intel/ixgbe/ixgbe_ipsec.c
332
if (t->mode & IXGBE_RXTXMOD_VF)
drivers/net/ethernet/intel/ixgbe/ixgbe_ipsec.c
333
ixgbe_ipsec_del_sa(adapter->netdev, t->xs);
drivers/net/ethernet/intel/ixgbe/ixgbe_ipsec.c
335
ixgbe_ipsec_set_tx_sa(hw, i, t->key, t->salt);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
8233
u32 h, t;
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
8236
t = IXGBE_READ_REG(hw, IXGBE_PVFTDTN(q_per_pool, i, j));
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
8238
if (h != t)
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
8498
static void ixgbe_service_timer(struct timer_list *t)
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
8500
struct ixgbe_adapter *adapter = timer_container_of(adapter, t,
drivers/net/ethernet/intel/ixgbevf/ipsec.c
113
struct tx_sa *t = &ipsec->tx_tbl[i];
drivers/net/ethernet/intel/ixgbevf/ipsec.c
123
if (t->used) {
drivers/net/ethernet/intel/ixgbevf/ipsec.c
124
ret = ixgbevf_ipsec_set_pf_sa(adapter, t->xs);
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
3206
static void ixgbevf_service_timer(struct timer_list *t)
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
3208
struct ixgbevf_adapter *adapter = timer_container_of(adapter, t,
drivers/net/ethernet/jme.c
1183
jme_pcc_tasklet(struct tasklet_struct *t)
drivers/net/ethernet/jme.c
1185
struct jme_adapter *jme = from_tasklet(jme, t, pcc_task);
drivers/net/ethernet/jme.c
1340
jme_rx_clean_tasklet(struct tasklet_struct *t)
drivers/net/ethernet/jme.c
1342
struct jme_adapter *jme = from_tasklet(jme, t, rxclean_task);
drivers/net/ethernet/jme.c
1375
jme_rx_empty_tasklet(struct tasklet_struct *t)
drivers/net/ethernet/jme.c
1377
struct jme_adapter *jme = from_tasklet(jme, t, rxempty_task);
drivers/net/ethernet/jme.c
1411
static void jme_tx_clean_tasklet(struct tasklet_struct *t)
drivers/net/ethernet/jme.c
1413
struct jme_adapter *jme = from_tasklet(jme, t, txclean_task);
drivers/net/ethernet/korina.c
901
static void korina_poll_media(struct timer_list *t)
drivers/net/ethernet/korina.c
903
struct korina_private *lp = timer_container_of(lp, t,
drivers/net/ethernet/marvell/mv643xx_eth.c
1334
static void mib_counters_timer_wrapper(struct timer_list *t)
drivers/net/ethernet/marvell/mv643xx_eth.c
1336
struct mv643xx_eth_private *mp = timer_container_of(mp, t,
drivers/net/ethernet/marvell/mv643xx_eth.c
2307
static inline void oom_timer_wrapper(struct timer_list *t)
drivers/net/ethernet/marvell/mv643xx_eth.c
2309
struct mv643xx_eth_private *mp = timer_container_of(mp, t, rx_oom);
drivers/net/ethernet/marvell/pxa168_eth.c
354
static inline void rxq_refill_timer_wrapper(struct timer_list *t)
drivers/net/ethernet/marvell/pxa168_eth.c
356
struct pxa168_eth_private *pep = timer_container_of(pep, t, timeout);
drivers/net/ethernet/marvell/skge.c
1494
static void xm_link_timer(struct timer_list *t)
drivers/net/ethernet/marvell/skge.c
1496
struct skge_port *skge = timer_container_of(skge, t, link_timer);
drivers/net/ethernet/marvell/skge.c
3347
static void skge_extirq(struct tasklet_struct *t)
drivers/net/ethernet/marvell/skge.c
3349
struct skge_hw *hw = from_tasklet(hw, t, phy_task);
drivers/net/ethernet/marvell/skge.c
3706
const struct skge_tx_desc *t = e->desc;
drivers/net/ethernet/marvell/skge.c
3708
t->control, t->dma_hi, t->dma_lo, t->status,
drivers/net/ethernet/marvell/skge.c
3709
t->csum_offs, t->csum_write, t->csum_start);
drivers/net/ethernet/marvell/sky2.c
2960
static void sky2_watchdog(struct timer_list *t)
drivers/net/ethernet/marvell/sky2.c
2962
struct sky2_hw *hw = timer_container_of(hw, t, watchdog_timer);
drivers/net/ethernet/mediatek/mtk_ppe.c
303
struct mtk_ipv4_tuple *t;
drivers/net/ethernet/mediatek/mtk_ppe.c
308
t = &entry->ipv4.new;
drivers/net/ethernet/mediatek/mtk_ppe.c
314
t = &entry->ipv4.orig;
drivers/net/ethernet/mediatek/mtk_ppe.c
325
t->src_ip = be32_to_cpu(src_addr);
drivers/net/ethernet/mediatek/mtk_ppe.c
326
t->dest_ip = be32_to_cpu(dest_addr);
drivers/net/ethernet/mediatek/mtk_ppe.c
331
t->src_port = be16_to_cpu(src_port);
drivers/net/ethernet/mediatek/mtk_ppe.c
332
t->dest_port = be16_to_cpu(dest_port);
drivers/net/ethernet/mediatek/mtk_wed_wo.c
235
static void mtk_wed_wo_irq_tasklet(struct tasklet_struct *t)
drivers/net/ethernet/mediatek/mtk_wed_wo.c
237
struct mtk_wed_wo *wo = from_tasklet(wo, t, mmio.irq_tasklet);
drivers/net/ethernet/mellanox/mlx4/alloc.c
558
dma_addr_t t;
drivers/net/ethernet/mellanox/mlx4/alloc.c
564
dma_alloc_coherent(&dev->persist->pdev->dev, size, &t,
drivers/net/ethernet/mellanox/mlx4/alloc.c
569
buf->direct.map = t;
drivers/net/ethernet/mellanox/mlx4/alloc.c
571
while (t & ((1 << buf->page_shift) - 1)) {
drivers/net/ethernet/mellanox/mlx4/alloc.c
590
dma_addr_t t;
drivers/net/ethernet/mellanox/mlx4/alloc.c
604
PAGE_SIZE, &t, GFP_KERNEL);
drivers/net/ethernet/mellanox/mlx4/alloc.c
608
buf->page_list[i].map = t;
drivers/net/ethernet/mellanox/mlx4/catas.c
237
static void poll_catas(struct timer_list *t)
drivers/net/ethernet/mellanox/mlx4/catas.c
239
struct mlx4_priv *priv = timer_container_of(priv, t, catas_err.timer);
drivers/net/ethernet/mellanox/mlx4/cq.c
58
void mlx4_cq_tasklet_cb(struct tasklet_struct *t)
drivers/net/ethernet/mellanox/mlx4/cq.c
62
struct mlx4_eq_tasklet *ctx = from_tasklet(ctx, t, task);
drivers/net/ethernet/mellanox/mlx4/en_ethtool.c
1935
int t;
drivers/net/ethernet/mellanox/mlx4/en_ethtool.c
1940
for (t = 0; t < MLX4_EN_NUM_TX_TYPES; t++)
drivers/net/ethernet/mellanox/mlx4/en_ethtool.c
1941
for (i = 0; i < priv->tx_ring_num[t]; i++)
drivers/net/ethernet/mellanox/mlx4/en_ethtool.c
1943
priv->tx_ring[t][i]->bf_alloced;
drivers/net/ethernet/mellanox/mlx4/en_ethtool.c
1955
for (t = 0; t < MLX4_EN_NUM_TX_TYPES; t++)
drivers/net/ethernet/mellanox/mlx4/en_ethtool.c
1956
for (i = 0; i < priv->tx_ring_num[t]; i++)
drivers/net/ethernet/mellanox/mlx4/en_ethtool.c
1957
priv->tx_ring[t][i]->bf_enabled =
drivers/net/ethernet/mellanox/mlx4/en_ethtool.c
53
int i, t;
drivers/net/ethernet/mellanox/mlx4/en_ethtool.c
56
for (t = 0 ; t < MLX4_EN_NUM_TX_TYPES; t++) {
drivers/net/ethernet/mellanox/mlx4/en_ethtool.c
57
for (i = 0; i < priv->tx_ring_num[t]; i++) {
drivers/net/ethernet/mellanox/mlx4/en_ethtool.c
58
priv->tx_cq[t][i]->moder_cnt = priv->tx_frames;
drivers/net/ethernet/mellanox/mlx4/en_ethtool.c
59
priv->tx_cq[t][i]->moder_time = priv->tx_usecs;
drivers/net/ethernet/mellanox/mlx4/en_ethtool.c
62
priv->tx_cq[t][i]);
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
1408
int i, t;
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
1433
for (t = 0 ; t < MLX4_EN_NUM_TX_TYPES; t++) {
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
1434
for (i = 0; i < priv->tx_ring_num[t]; i++) {
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
1435
cq = priv->tx_cq[t][i];
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
1637
int i, t;
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
1722
for (t = 0 ; t < MLX4_EN_NUM_TX_TYPES; t++) {
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
1723
u8 num_tx_rings_p_up = t == TX ?
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
1724
priv->num_tx_rings_p_up : priv->tx_ring_num[t];
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
1726
for (i = 0; i < priv->tx_ring_num[t]; i++) {
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
1728
cq = priv->tx_cq[t][i];
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
1745
tx_ring = priv->tx_ring[t][i];
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
1755
if (t != TX_XDP) {
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
1858
if (t == MLX4_EN_NUM_TX_TYPES) {
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
1859
t--;
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
1860
i = priv->tx_ring_num[t];
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
1862
while (t >= 0) {
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
1864
mlx4_en_deactivate_tx_ring(priv, priv->tx_ring[t][i]);
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
1865
mlx4_en_deactivate_cq(priv, priv->tx_cq[t][i]);
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
1867
if (!t--)
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
1869
i = priv->tx_ring_num[t];
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
1894
int i, t;
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
1984
for (t = 0; t < MLX4_EN_NUM_TX_TYPES; t++) {
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
1985
for (i = 0; i < priv->tx_ring_num[t]; i++) {
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
1986
mlx4_en_deactivate_tx_ring(priv, priv->tx_ring[t][i]);
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
1987
mlx4_en_deactivate_cq(priv, priv->tx_cq[t][i]);
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
1992
for (t = 0; t < MLX4_EN_NUM_TX_TYPES; t++)
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
1993
for (i = 0; i < priv->tx_ring_num[t]; i++)
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
1994
mlx4_en_free_tx_buf(dev, priv->tx_ring[t][i]);
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2128
int i, t;
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2134
for (t = 0; t < MLX4_EN_NUM_TX_TYPES; t++) {
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2135
for (i = 0; i < priv->tx_ring_num[t]; i++) {
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2136
if (priv->tx_ring[t] && priv->tx_ring[t][i])
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2138
&priv->tx_ring[t][i]);
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2139
if (priv->tx_cq[t] && priv->tx_cq[t][i])
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2140
mlx4_en_destroy_cq(priv, &priv->tx_cq[t][i]);
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2142
kfree(priv->tx_ring[t]);
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2143
kfree(priv->tx_cq[t]);
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2159
int i, t;
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2163
for (t = 0; t < MLX4_EN_NUM_TX_TYPES; t++) {
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2164
for (i = 0; i < priv->tx_ring_num[t]; i++) {
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2166
if (mlx4_en_create_cq(priv, &priv->tx_cq[t][i],
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2167
prof->tx_ring_size, i, t, node))
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2170
if (mlx4_en_create_tx_ring(priv, &priv->tx_ring[t][i],
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2207
for (t = 0; t < MLX4_EN_NUM_TX_TYPES; t++) {
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2208
for (i = 0; i < priv->tx_ring_num[t]; i++) {
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2209
if (priv->tx_ring[t][i])
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2211
&priv->tx_ring[t][i]);
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2212
if (priv->tx_cq[t][i])
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2213
mlx4_en_destroy_cq(priv, &priv->tx_cq[t][i]);
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2224
int t;
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2238
for (t = 0; t < MLX4_EN_NUM_TX_TYPES; t++) {
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2239
dst->tx_ring_num[t] = prof->tx_ring_num[t];
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2240
if (!dst->tx_ring_num[t])
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2243
dst->tx_ring[t] = kzalloc_objs(struct mlx4_en_tx_ring *,
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2245
if (!dst->tx_ring[t])
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2248
dst->tx_cq[t] = kzalloc_objs(struct mlx4_en_cq *, MAX_TX_RINGS);
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2249
if (!dst->tx_cq[t]) {
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2250
kfree(dst->tx_ring[t]);
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2258
while (t--) {
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2259
kfree(dst->tx_ring[t]);
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2260
kfree(dst->tx_cq[t]);
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2268
int t;
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2275
for (t = 0; t < MLX4_EN_NUM_TX_TYPES; t++) {
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2276
dst->tx_ring_num[t] = src->tx_ring_num[t];
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2277
dst->tx_ring[t] = src->tx_ring[t];
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2278
dst->tx_cq[t] = src->tx_cq[t];
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2291
int i, t, ret;
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2304
for (t = 0; t < MLX4_EN_NUM_TX_TYPES; t++) {
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2305
kfree(tmp->tx_ring[t]);
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2306
kfree(tmp->tx_cq[t]);
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
3166
int i, t;
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
3212
for (t = 0; t < MLX4_EN_NUM_TX_TYPES; t++) {
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
3213
priv->tx_ring_num[t] = prof->tx_ring_num[t];
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
3214
if (!priv->tx_ring_num[t])
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
3217
priv->tx_ring[t] = kzalloc_objs(struct mlx4_en_tx_ring *,
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
3219
if (!priv->tx_ring[t]) {
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
3223
priv->tx_cq[t] = kzalloc_objs(struct mlx4_en_cq *, MAX_TX_RINGS);
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
3224
if (!priv->tx_cq[t]) {
drivers/net/ethernet/mellanox/mlx4/eq.c
1009
PAGE_SIZE, &t,
drivers/net/ethernet/mellanox/mlx4/eq.c
1014
dma_list[i] = t;
drivers/net/ethernet/mellanox/mlx4/eq.c
1015
eq->page_list[i].map = t;
drivers/net/ethernet/mellanox/mlx4/eq.c
978
dma_addr_t t;
drivers/net/ethernet/mellanox/mlx4/mlx4.h
1231
void mlx4_cq_tasklet_cb(struct tasklet_struct *t);
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
505
int t;
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
513
for (t = 0; t < MLX4_NUM_OF_RESOURCE_TYPE; ++t)
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
515
slave_list[i].res_list[t]);
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
546
for (t = 0; t < dev->persist->num_vfs + 1; t++) {
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
548
mlx4_get_active_ports(dev, t);
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
552
t, dev->caps.num_qps -
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
558
t, dev->caps.num_cqs -
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
563
t, dev->caps.num_srqs -
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
568
t, dev->caps.num_mpts -
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
573
t, dev->caps.num_mtts -
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
577
if (t == mlx4_master_func_num(dev)) {
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
592
res_alloc->quota[t] =
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
595
res_alloc->guaranteed[t] = 2;
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
600
res_alloc->quota[t] = MLX4_MAX_MAC_NUM;
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
601
res_alloc->guaranteed[t] = 2;
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
605
if (t == mlx4_master_func_num(dev)) {
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
606
res_alloc->quota[t] = MLX4_MAX_VLAN_NUM;
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
607
res_alloc->guaranteed[t] = MLX4_MAX_VLAN_NUM / 2;
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
610
res_alloc->quota[t];
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
612
res_alloc->quota[t] = MLX4_MAX_VLAN_NUM / 2;
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
613
res_alloc->guaranteed[t] = 0;
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
617
res_alloc->quota[t] = dev->caps.max_counters;
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
618
res_alloc->guaranteed[t] =
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
619
mlx4_calc_res_counter_guaranteed(dev, res_alloc, t);
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
628
res_alloc->guaranteed[t];
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
630
res_alloc->res_reserved += res_alloc->guaranteed[t];
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
846
static const char *mlx4_resource_type_to_str(enum mlx4_resource t)
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
848
switch (t) {
drivers/net/ethernet/mellanox/mlx5/core/cq.c
44
void mlx5_cq_tasklet_cb(struct tasklet_struct *t)
drivers/net/ethernet/mellanox/mlx5/core/cq.c
48
struct mlx5_eq_tasklet *ctx = from_tasklet(ctx, t, task);
drivers/net/ethernet/mellanox/mlx5/core/en/fs.h
27
struct mlx5_flow_table *t;
drivers/net/ethernet/mellanox/mlx5/core/en/fs_tt_redirect.c
127
rule = mlx5_add_flow_rules(fs_udp_t->t, NULL, &flow_act, &dest, 1);
drivers/net/ethernet/mellanox/mlx5/core/en/fs_tt_redirect.c
181
ft->g[ft->num_groups] = mlx5_create_flow_group(ft->t, in);
drivers/net/ethernet/mellanox/mlx5/core/en/fs_tt_redirect.c
191
ft->g[ft->num_groups] = mlx5_create_flow_group(ft->t, in);
drivers/net/ethernet/mellanox/mlx5/core/en/fs_tt_redirect.c
223
ft->t = mlx5_create_flow_table(ns, &ft_attr);
drivers/net/ethernet/mellanox/mlx5/core/en/fs_tt_redirect.c
224
if (IS_ERR(ft->t)) {
drivers/net/ethernet/mellanox/mlx5/core/en/fs_tt_redirect.c
225
err = PTR_ERR(ft->t);
drivers/net/ethernet/mellanox/mlx5/core/en/fs_tt_redirect.c
226
ft->t = NULL;
drivers/net/ethernet/mellanox/mlx5/core/en/fs_tt_redirect.c
231
fs_udp_type2str(type), ft->t->id, ft->t->level);
drivers/net/ethernet/mellanox/mlx5/core/en/fs_tt_redirect.c
250
if (IS_ERR_OR_NULL(fs_udp->tables[i].t))
drivers/net/ethernet/mellanox/mlx5/core/en/fs_tt_redirect.c
255
fs_udp->tables[i].t = NULL;
drivers/net/ethernet/mellanox/mlx5/core/en/fs_tt_redirect.c
285
dest.ft = udp->tables[i].t;
drivers/net/ethernet/mellanox/mlx5/core/en/fs_tt_redirect.c
379
ft = fs_any->table.t;
drivers/net/ethernet/mellanox/mlx5/core/en/fs_tt_redirect.c
408
rule = mlx5_add_flow_rules(fs_any_t->t, NULL, &flow_act, &dest, 1);
drivers/net/ethernet/mellanox/mlx5/core/en/fs_tt_redirect.c
452
ft->g[ft->num_groups] = mlx5_create_flow_group(ft->t, in);
drivers/net/ethernet/mellanox/mlx5/core/en/fs_tt_redirect.c
462
ft->g[ft->num_groups] = mlx5_create_flow_group(ft->t, in);
drivers/net/ethernet/mellanox/mlx5/core/en/fs_tt_redirect.c
492
ft->t = mlx5_create_flow_table(ns, &ft_attr);
drivers/net/ethernet/mellanox/mlx5/core/en/fs_tt_redirect.c
493
if (IS_ERR(ft->t)) {
drivers/net/ethernet/mellanox/mlx5/core/en/fs_tt_redirect.c
494
err = PTR_ERR(ft->t);
drivers/net/ethernet/mellanox/mlx5/core/en/fs_tt_redirect.c
495
ft->t = NULL;
drivers/net/ethernet/mellanox/mlx5/core/en/fs_tt_redirect.c
500
ft->t->id, ft->t->level);
drivers/net/ethernet/mellanox/mlx5/core/en/fs_tt_redirect.c
541
dest.ft = any->table.t;
drivers/net/ethernet/mellanox/mlx5/core/en/fs_tt_redirect.c
556
if (IS_ERR_OR_NULL(fs_any->table.t))
drivers/net/ethernet/mellanox/mlx5/core/en/fs_tt_redirect.c
561
fs_any->table.t = NULL;
drivers/net/ethernet/mellanox/mlx5/core/en/fs_tt_redirect.c
97
ft = fs_udp->tables[type].t;
drivers/net/ethernet/mellanox/mlx5/core/en/trap.c
122
int max_mtu, struct mlx5e_trap *t)
drivers/net/ethernet/mellanox/mlx5/core/en/trap.c
124
struct mlx5e_params *params = &t->params;
drivers/net/ethernet/mellanox/mlx5/core/en/trap.c
129
mlx5e_build_rq_param(mdev, params, NULL, &t->rq_param);
drivers/net/ethernet/mellanox/mlx5/core/en/trap.c
136
struct mlx5e_trap *t;
drivers/net/ethernet/mellanox/mlx5/core/en/trap.c
139
t = kvzalloc_node(sizeof(*t), GFP_KERNEL, cpu_to_node(cpu));
drivers/net/ethernet/mellanox/mlx5/core/en/trap.c
140
if (!t)
drivers/net/ethernet/mellanox/mlx5/core/en/trap.c
143
mlx5e_build_trap_params(priv->mdev, netdev->max_mtu, t);
drivers/net/ethernet/mellanox/mlx5/core/en/trap.c
145
t->priv = priv;
drivers/net/ethernet/mellanox/mlx5/core/en/trap.c
146
t->mdev = priv->mdev;
drivers/net/ethernet/mellanox/mlx5/core/en/trap.c
147
t->pdev = mlx5_core_dma_dev(priv->mdev);
drivers/net/ethernet/mellanox/mlx5/core/en/trap.c
148
t->netdev = priv->netdev;
drivers/net/ethernet/mellanox/mlx5/core/en/trap.c
149
t->mkey_be = cpu_to_be32(priv->mdev->mlx5e_res.hw_objs.mkey);
drivers/net/ethernet/mellanox/mlx5/core/en/trap.c
150
t->stats = &priv->trap_stats.ch;
drivers/net/ethernet/mellanox/mlx5/core/en/trap.c
152
netif_napi_add_locked(netdev, &t->napi, mlx5e_trap_napi_poll);
drivers/net/ethernet/mellanox/mlx5/core/en/trap.c
154
err = mlx5e_open_trap_rq(priv, t);
drivers/net/ethernet/mellanox/mlx5/core/en/trap.c
158
err = mlx5e_create_trap_direct_rq_tir(t->mdev, &t->tir, t->rq.rqn);
drivers/net/ethernet/mellanox/mlx5/core/en/trap.c
162
return t;
drivers/net/ethernet/mellanox/mlx5/core/en/trap.c
165
mlx5e_close_trap_rq(&t->rq);
drivers/net/ethernet/mellanox/mlx5/core/en/trap.c
167
netif_napi_del_locked(&t->napi);
drivers/net/ethernet/mellanox/mlx5/core/en/trap.c
168
kvfree(t);
drivers/net/ethernet/mellanox/mlx5/core/en/trap.c
39
static void mlx5e_init_trap_rq(struct mlx5e_trap *t, struct mlx5e_params *params,
drivers/net/ethernet/mellanox/mlx5/core/en/trap.c
42
struct mlx5_core_dev *mdev = t->mdev;
drivers/net/ethernet/mellanox/mlx5/core/en/trap.c
43
struct mlx5e_priv *priv = t->priv;
drivers/net/ethernet/mellanox/mlx5/core/en/trap.c
46
rq->pdev = t->pdev;
drivers/net/ethernet/mellanox/mlx5/core/en/trap.c
59
static int mlx5e_open_trap_rq(struct mlx5e_priv *priv, struct mlx5e_trap *t)
drivers/net/ethernet/mellanox/mlx5/core/en/trap.c
61
struct mlx5e_rq_param *rq_param = &t->rq_param;
drivers/net/ethernet/mellanox/mlx5/core/en/trap.c
65
struct mlx5e_rq *rq = &t->rq;
drivers/net/ethernet/mellanox/mlx5/core/en/trap.c
76
ccp.ch_stats = t->stats;
drivers/net/ethernet/mellanox/mlx5/core/en/trap.c
77
ccp.napi = &t->napi;
drivers/net/ethernet/mellanox/mlx5/core/en/trap.c
84
mlx5e_init_trap_rq(t, &t->params, rq);
drivers/net/ethernet/mellanox/mlx5/core/en/trap.c
85
err = mlx5e_open_rq(&t->params, rq_param, NULL, node, q_counter, rq);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/fs_tcp.c
138
flow = mlx5_add_flow_rules(ft->t, spec, &flow_act, &dest, 1);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/fs_tcp.c
162
rule = mlx5_add_flow_rules(accel_fs_t->t, NULL, &flow_act, &dest, 1);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/fs_tcp.c
238
ft->g[ft->num_groups] = mlx5_create_flow_group(ft->t, in);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/fs_tcp.c
248
ft->g[ft->num_groups] = mlx5_create_flow_group(ft->t, in);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/fs_tcp.c
279
ft->t = mlx5_create_flow_table(ns, &ft_attr);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/fs_tcp.c
280
if (IS_ERR(ft->t)) {
drivers/net/ethernet/mellanox/mlx5/core/en_accel/fs_tcp.c
281
err = PTR_ERR(ft->t);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/fs_tcp.c
282
ft->t = NULL;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/fs_tcp.c
287
ft->t->id, ft->t->level);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/fs_tcp.c
331
dest.ft = accel_tcp->tables[i].t;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/fs_tcp.c
348
if (IS_ERR_OR_NULL(fs_tcp->tables[i].t))
drivers/net/ethernet/mellanox/mlx5/core/en_accel/fs_tcp.c
353
fs_tcp->tables[i].t = NULL;
drivers/net/ethernet/mellanox/mlx5/core/en_arfs.c
168
dest.ft = arfs->arfs_tables[i].ft.t;
drivers/net/ethernet/mellanox/mlx5/core/en_arfs.c
197
if (!IS_ERR_OR_NULL(arfs->arfs_tables[i].ft.t))
drivers/net/ethernet/mellanox/mlx5/core/en_arfs.c
236
arfs_t->default_rule = mlx5_add_flow_rules(arfs_t->ft.t, NULL,
drivers/net/ethernet/mellanox/mlx5/core/en_arfs.c
319
ft->g[ft->num_groups] = mlx5_create_flow_group(ft->t, in);
drivers/net/ethernet/mellanox/mlx5/core/en_arfs.c
328
ft->g[ft->num_groups] = mlx5_create_flow_group(ft->t, in);
drivers/net/ethernet/mellanox/mlx5/core/en_arfs.c
363
ft->t = mlx5_create_flow_table(ns, &ft_attr);
drivers/net/ethernet/mellanox/mlx5/core/en_arfs.c
364
if (IS_ERR(ft->t)) {
drivers/net/ethernet/mellanox/mlx5/core/en_arfs.c
365
err = PTR_ERR(ft->t);
drivers/net/ethernet/mellanox/mlx5/core/en_arfs.c
366
ft->t = NULL;
drivers/net/ethernet/mellanox/mlx5/core/en_arfs.c
544
ft = arfs_table->ft.t;
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
1060
ft->g[ft->num_groups] = mlx5_create_flow_group(ft->t, in);
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
1071
ft->g[ft->num_groups] = mlx5_create_flow_group(ft->t, in);
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
1081
ft->g[ft->num_groups] = mlx5_create_flow_group(ft->t, in);
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
1117
ft->t = mlx5_create_flow_table(fs->ns, &ft_attr);
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
1118
if (IS_ERR(ft->t)) {
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
1119
err = PTR_ERR(ft->t);
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
1120
ft->t = NULL;
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
1131
mlx5_destroy_flow_table(ft->t);
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
1132
ft->t = NULL;
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
1163
ft->g[ft->num_groups] = mlx5_create_flow_group(ft->t, in);
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
1175
ft->g[ft->num_groups] = mlx5_create_flow_group(ft->t, in);
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
1186
ft->g[ft->num_groups] = mlx5_create_flow_group(ft->t, in);
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
1197
ft->g[ft->num_groups] = mlx5_create_flow_group(ft->t, in);
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
1206
ft->g[ft->num_groups] = mlx5_create_flow_group(ft->t, in);
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
1250
ft->t = mlx5_create_flow_table(fs->ns, &ft_attr);
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
1251
if (IS_ERR(ft->t))
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
1252
return PTR_ERR(ft->t);
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
1271
mlx5_destroy_flow_table(ft->t);
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
167
return vlan->ft.t;
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
223
struct mlx5_flow_table *ft = fs->vlan->ft.t;
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
230
dest.ft = fs->l2.ft.t;
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
390
struct mlx5_flow_table *ft = fs->vlan->ft.t;
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
415
struct mlx5_flow_table *ft = fs->l2.ft.t;
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
750
struct mlx5_flow_table *ft = fs->promisc.ft.t;
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
785
ft->t = mlx5_create_auto_grouped_flow_table(fs->ns, &ft_attr);
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
786
if (IS_ERR(ft->t)) {
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
787
err = PTR_ERR(ft->t);
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
788
ft->t = NULL;
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
800
mlx5_destroy_flow_table(ft->t);
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
801
ft->t = NULL;
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
816
if (!fs->promisc.ft.t)
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
819
mlx5_destroy_flow_table(fs->promisc.ft.t);
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
820
fs->promisc.ft.t = NULL;
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
891
mlx5_destroy_flow_table(ft->t);
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
892
ft->t = NULL;
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c
979
struct mlx5_flow_table *ft = fs->l2.ft.t;
drivers/net/ethernet/mellanox/mlx5/core/en_tc.c
1297
if (IS_ERR_OR_NULL(tc->t)) {
drivers/net/ethernet/mellanox/mlx5/core/en_tc.c
1299
tc->t =
drivers/net/ethernet/mellanox/mlx5/core/en_tc.c
1302
if (IS_ERR(tc->t)) {
drivers/net/ethernet/mellanox/mlx5/core/en_tc.c
1306
rule = ERR_CAST(tc->t);
drivers/net/ethernet/mellanox/mlx5/core/en_tc.c
1432
!IS_ERR_OR_NULL(tc->t)) {
drivers/net/ethernet/mellanox/mlx5/core/en_tc.c
1434
tc->t = NULL;
drivers/net/ethernet/mellanox/mlx5/core/en_tc.c
5360
if (!IS_ERR_OR_NULL(tc->t)) {
drivers/net/ethernet/mellanox/mlx5/core/en_tc.c
5362
tc->t = NULL;
drivers/net/ethernet/mellanox/mlx5/core/en_tc.c
84
struct mlx5_flow_table *t;
drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c
391
static void mlx5_fpga_conn_cq_tasklet(struct tasklet_struct *t)
drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c
393
struct mlx5_fpga_conn *conn = from_tasklet(conn, t, cq.tasklet);
drivers/net/ethernet/mellanox/mlx5/core/fw_reset.c
297
static void poll_sync_reset(struct timer_list *t)
drivers/net/ethernet/mellanox/mlx5/core/fw_reset.c
299
struct mlx5_fw_reset *fw_reset = timer_container_of(fw_reset, t,
drivers/net/ethernet/mellanox/mlx5/core/health.c
787
static void poll_health(struct timer_list *t)
drivers/net/ethernet/mellanox/mlx5/core/health.c
789
struct mlx5_core_dev *dev = timer_container_of(dev, t,
drivers/net/ethernet/mellanox/mlx5/core/lib/eq.h
87
void mlx5_cq_tasklet_cb(struct tasklet_struct *t);
drivers/net/ethernet/mellanox/mlx5/core/lib/fs_ttc.c
1045
rule = mlx5_add_flow_rules(ttc->t, spec, &flow_act, &dest, 1);
drivers/net/ethernet/mellanox/mlx5/core/lib/fs_ttc.c
1090
rule = mlx5_add_flow_rules(ttc->t, spec, &flow_act, &dest, 1);
drivers/net/ethernet/mellanox/mlx5/core/lib/fs_ttc.c
36
struct mlx5_flow_table *t;
drivers/net/ethernet/mellanox/mlx5/core/lib/fs_ttc.c
409
ft = ttc->t;
drivers/net/ethernet/mellanox/mlx5/core/lib/fs_ttc.c
46
return ttc->t;
drivers/net/ethernet/mellanox/mlx5/core/lib/fs_ttc.c
477
ttc->g[ttc->num_groups] = mlx5_create_flow_group(ttc->t, in);
drivers/net/ethernet/mellanox/mlx5/core/lib/fs_ttc.c
495
ttc->g[ttc->num_groups] = mlx5_create_flow_group(ttc->t, in);
drivers/net/ethernet/mellanox/mlx5/core/lib/fs_ttc.c
516
ttc->g[ttc->num_groups] = mlx5_create_flow_group(ttc->t, in);
drivers/net/ethernet/mellanox/mlx5/core/lib/fs_ttc.c
562
ttc->g[ttc->num_groups] = mlx5_create_flow_group(ttc->t, in);
drivers/net/ethernet/mellanox/mlx5/core/lib/fs_ttc.c
575
ttc->g[ttc->num_groups] = mlx5_create_flow_group(ttc->t, in);
drivers/net/ethernet/mellanox/mlx5/core/lib/fs_ttc.c
595
ttc->g[ttc->num_groups] = mlx5_create_flow_group(ttc->t, in);
drivers/net/ethernet/mellanox/mlx5/core/lib/fs_ttc.c
605
ttc->g[ttc->num_groups] = mlx5_create_flow_group(ttc->t, in);
drivers/net/ethernet/mellanox/mlx5/core/lib/fs_ttc.c
675
ft = ttc->t;
drivers/net/ethernet/mellanox/mlx5/core/lib/fs_ttc.c
736
ttc->g[ttc->num_groups] = mlx5_create_flow_group(ttc->t, in);
drivers/net/ethernet/mellanox/mlx5/core/lib/fs_ttc.c
749
ttc->g[ttc->num_groups] = mlx5_create_flow_group(ttc->t, in);
drivers/net/ethernet/mellanox/mlx5/core/lib/fs_ttc.c
759
ttc->g[ttc->num_groups] = mlx5_create_flow_group(ttc->t, in);
drivers/net/ethernet/mellanox/mlx5/core/lib/fs_ttc.c
769
ttc->g[ttc->num_groups] = mlx5_create_flow_group(ttc->t, in);
drivers/net/ethernet/mellanox/mlx5/core/lib/fs_ttc.c
822
ttc->t = mlx5_create_flow_table(ns, &params->ft_attr);
drivers/net/ethernet/mellanox/mlx5/core/lib/fs_ttc.c
823
if (IS_ERR(ttc->t)) {
drivers/net/ethernet/mellanox/mlx5/core/lib/fs_ttc.c
824
err = PTR_ERR(ttc->t);
drivers/net/ethernet/mellanox/mlx5/core/lib/fs_ttc.c
856
mlx5_destroy_flow_table(ttc->t);
drivers/net/ethernet/mellanox/mlx5/core/lib/fs_ttc.c
899
ttc->t = mlx5_create_flow_table(ns, &params->ft_attr);
drivers/net/ethernet/mellanox/mlx5/core/lib/fs_ttc.c
900
if (IS_ERR(ttc->t)) {
drivers/net/ethernet/mellanox/mlx5/core/lib/fs_ttc.c
901
err = PTR_ERR(ttc->t);
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
1084
ft->g[ft->num_groups] = mlx5_create_flow_group(ft->t, in);
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
1104
ft->g[ft->num_groups] = mlx5_create_flow_group(ft->t, in);
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
1115
ft->g[ft->num_groups] = mlx5_create_flow_group(ft->t, in);
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
1540
ft_crypto->t = flow_table;
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
1552
rule = mlx5_add_flow_rules(ft_crypto->t, NULL, &flow_act, NULL, 0);
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
1788
rule = mlx5_add_flow_rules(ft_crypto->t, spec, &flow_act, &dest, 1);
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
1813
rule = mlx5_add_flow_rules(ft_crypto->t, spec, &flow_act, &dest, 1);
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
212
mlx5_destroy_flow_table(ft->t);
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
213
ft->t = NULL;
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
2239
dest.ft = tx_fs->tables.ft_crypto.t;
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
290
ft->g[ft->num_groups] = mlx5_create_flow_group(ft->t, in);
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
305
ft->g[ft->num_groups] = mlx5_create_flow_group(ft->t, in);
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
316
ft->g[ft->num_groups] = mlx5_create_flow_group(ft->t, in);
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
431
ft_crypto->t = flow_table;
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
449
rule = mlx5_add_flow_rules(ft_crypto->t, spec, &flow_act, NULL, 0);
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
460
rule = mlx5_add_flow_rules(ft_crypto->t, NULL, &flow_act, NULL, 0);
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
75
struct mlx5_flow_table *t;
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
867
rule = mlx5_add_flow_rules(tx_tables->ft_crypto.t, spec, &flow_act, &dest, 1);
drivers/net/ethernet/mellanox/mlxsw/pci.c
1115
static void mlxsw_pci_eq_tasklet(struct tasklet_struct *t)
drivers/net/ethernet/mellanox/mlxsw/pci.c
1118
struct mlxsw_pci_queue *q = from_tasklet(q, t, u.eq.tasklet);
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
525
struct ip6_tnl *t = netdev_priv(to_dev);
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
526
struct flowi6 fl6 = t->fl.u.ip6;
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
534
fl6.flowi6_mark = t->parms.fwmark;
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
535
if (!ip6_tnl_xmit_ctl(t, &fl6.saddr, &fl6.daddr))
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
538
dst = ip6_route_output(t->net, NULL, &fl6);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2231
int i, t;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2234
for (t = 0; t < nv->txt_count; t++) {
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2235
struct fbnic_q_triad *qt = &nv->qt[t];
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2243
for (i = 0; i < nv->rxt_count; i++, t++) {
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2244
struct fbnic_q_triad *qt = &nv->qt[t];
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2409
int j, t;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2412
for (t = 0; t < nv->txt_count; t++) {
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2413
struct fbnic_q_triad *qt = &nv->qt[t];
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2435
for (j = 0; j < nv->rxt_count; j++, t++) {
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2436
struct fbnic_q_triad *qt = &nv->qt[t];
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2460
int j, t;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2465
for (j = 0, t = nv->txt_count; j < nv->rxt_count; j++, t++) {
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2466
struct fbnic_q_triad *qt = &nv->qt[t];
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2617
int i, t;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2624
for (t = 0; t < nv->rxt_count; t++) {
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2625
struct fbnic_q_triad *qt = &nv->qt[nv->txt_count + t];
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2721
int j, t;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2724
for (t = 0; t < nv->txt_count; t++) {
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2725
struct fbnic_q_triad *qt = &nv->qt[t];
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2733
for (j = 0; j < nv->rxt_count; j++, t++) {
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2734
struct fbnic_q_triad *qt = &nv->qt[t];
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2805
int i, j, t;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2811
for (t = nv->txt_count, j = 0; j < nv->rxt_count; j++, t++) {
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2815
if (fbnic_desc_used(&nv->qt[t].sub0) < 4 ||
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2816
fbnic_desc_used(&nv->qt[t].sub1) < 4)
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2913
int i, t;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2927
for (t = 0; t < nv->txt_count + nv->rxt_count; t++) {
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2928
err = fbnic_wait_queue_idle(fbn, t >= nv->txt_count,
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2929
nv->qt[t].sub0.q_idx);
drivers/net/ethernet/micrel/ks8842.c
592
static void ks8842_rx_frame_dma_tasklet(struct tasklet_struct *t)
drivers/net/ethernet/micrel/ks8842.c
594
struct ks8842_adapter *adapter = from_tasklet(adapter, t, dma_rx.tasklet);
drivers/net/ethernet/micrel/ks8842.c
725
static void ks8842_tasklet(struct tasklet_struct *t)
drivers/net/ethernet/micrel/ks8842.c
727
struct ks8842_adapter *adapter = from_tasklet(adapter, t, tasklet);
drivers/net/ethernet/micrel/ksz884x.c
4795
static void rx_proc_task(struct tasklet_struct *t)
drivers/net/ethernet/micrel/ksz884x.c
4797
struct dev_info *hw_priv = from_tasklet(hw_priv, t, rx_tasklet);
drivers/net/ethernet/micrel/ksz884x.c
4817
static void tx_proc_task(struct tasklet_struct *t)
drivers/net/ethernet/micrel/ksz884x.c
4819
struct dev_info *hw_priv = from_tasklet(hw_priv, t, tx_tasklet);
drivers/net/ethernet/micrel/ksz884x.c
6305
static void mib_monitor(struct timer_list *t)
drivers/net/ethernet/micrel/ksz884x.c
6307
struct dev_info *hw_priv = timer_container_of(hw_priv, t,
drivers/net/ethernet/micrel/ksz884x.c
6333
static void dev_monitor(struct timer_list *t)
drivers/net/ethernet/micrel/ksz884x.c
6335
struct dev_priv *priv = timer_container_of(priv, t,
drivers/net/ethernet/microchip/encx24j600-regmap.c
119
struct spi_transfer t[3] = { { .tx_buf = &cmd, .len = sizeof(cmd), },
drivers/net/ethernet/microchip/encx24j600-regmap.c
157
spi_message_add_tail(&t[0], &m);
drivers/net/ethernet/microchip/encx24j600-regmap.c
160
t[1].tx_buf = &reg;
drivers/net/ethernet/microchip/encx24j600-regmap.c
161
spi_message_add_tail(&t[1], &m);
drivers/net/ethernet/microchip/encx24j600-regmap.c
164
spi_message_add_tail(&t[2], &m);
drivers/net/ethernet/microchip/encx24j600-regmap.c
37
struct spi_transfer t[2] = { { .tx_buf = &opcode, .len = 1, },
drivers/net/ethernet/microchip/encx24j600-regmap.c
40
spi_message_add_tail(&t[0], &m);
drivers/net/ethernet/microchip/encx24j600-regmap.c
41
spi_message_add_tail(&t[1], &m);
drivers/net/ethernet/microchip/lan966x/lan966x_regs.h
1757
#define VCAP_UPDATE_CTRL(t) __REG(TARGET_VCAP, t, 3, 0, 0, 1, 8, 0, 0, 1, 4)
drivers/net/ethernet/microchip/lan966x/lan966x_regs.h
1808
#define VCAP_MV_CFG(t) __REG(TARGET_VCAP, t, 3, 0, 0, 1, 8, 4, 0, 1, 4)
drivers/net/ethernet/microchip/lan966x/lan966x_regs.h
1823
#define VCAP_ENTRY_DAT(t, r) __REG(TARGET_VCAP, t, 3, 8, 0, 1, 904, 0, r, 64, 4)
drivers/net/ethernet/microchip/lan966x/lan966x_regs.h
1826
#define VCAP_MASK_DAT(t, r) __REG(TARGET_VCAP, t, 3, 8, 0, 1, 904, 256, r, 64, 4)
drivers/net/ethernet/microchip/lan966x/lan966x_regs.h
1829
#define VCAP_ACTION_DAT(t, r) __REG(TARGET_VCAP, t, 3, 8, 0, 1, 904, 512, r, 64, 4)
drivers/net/ethernet/microchip/lan966x/lan966x_regs.h
1832
#define VCAP_CNT_DAT(t, r) __REG(TARGET_VCAP, t, 3, 8, 0, 1, 904, 768, r, 32, 4)
drivers/net/ethernet/microchip/lan966x/lan966x_regs.h
1835
#define VCAP_CNT_FW_DAT(t) __REG(TARGET_VCAP, t, 3, 8, 0, 1, 904, 896, 0, 1, 4)
drivers/net/ethernet/microchip/lan966x/lan966x_regs.h
1838
#define VCAP_TG_DAT(t) __REG(TARGET_VCAP, t, 3, 8, 0, 1, 904, 900, 0, 1, 4)
drivers/net/ethernet/microchip/lan966x/lan966x_regs.h
1841
#define VCAP_CORE_IDX(t) __REG(TARGET_VCAP, t, 3, 912, 0, 1, 8, 0, 0, 1, 4)
drivers/net/ethernet/microchip/lan966x/lan966x_regs.h
1850
#define VCAP_CORE_MAP(t) __REG(TARGET_VCAP, t, 3, 912, 0, 1, 8, 4, 0, 1, 4)
drivers/net/ethernet/microchip/lan966x/lan966x_regs.h
1859
#define VCAP_VER(t) __REG(TARGET_VCAP, t, 3, 924, 0, 1, 40, 0, 0, 1, 4)
drivers/net/ethernet/microchip/lan966x/lan966x_regs.h
1862
#define VCAP_ENTRY_WIDTH(t) __REG(TARGET_VCAP, t, 3, 924, 0, 1, 40, 4, 0, 1, 4)
drivers/net/ethernet/microchip/lan966x/lan966x_regs.h
1865
#define VCAP_ENTRY_CNT(t) __REG(TARGET_VCAP, t, 3, 924, 0, 1, 40, 8, 0, 1, 4)
drivers/net/ethernet/microchip/lan966x/lan966x_regs.h
1868
#define VCAP_ENTRY_SWCNT(t) __REG(TARGET_VCAP, t, 3, 924, 0, 1, 40, 12, 0, 1, 4)
drivers/net/ethernet/microchip/lan966x/lan966x_regs.h
1871
#define VCAP_ENTRY_TG_WIDTH(t) __REG(TARGET_VCAP, t, 3, 924, 0, 1, 40, 16, 0, 1, 4)
drivers/net/ethernet/microchip/lan966x/lan966x_regs.h
1874
#define VCAP_ACTION_DEF_CNT(t) __REG(TARGET_VCAP, t, 3, 924, 0, 1, 40, 20, 0, 1, 4)
drivers/net/ethernet/microchip/lan966x/lan966x_regs.h
1877
#define VCAP_ACTION_WIDTH(t) __REG(TARGET_VCAP, t, 3, 924, 0, 1, 40, 24, 0, 1, 4)
drivers/net/ethernet/microchip/lan966x/lan966x_regs.h
1880
#define VCAP_CNT_WIDTH(t) __REG(TARGET_VCAP, t, 3, 924, 0, 1, 40, 28, 0, 1, 4)
drivers/net/ethernet/microchip/lan966x/lan966x_regs.h
1883
#define VCAP_CORE_CNT(t) __REG(TARGET_VCAP, t, 3, 924, 0, 1, 40, 32, 0, 1, 4)
drivers/net/ethernet/microchip/lan966x/lan966x_regs.h
1886
#define VCAP_IF_CNT(t) __REG(TARGET_VCAP, t, 3, 924, 0, 1, 40, 36, 0, 1, 4)
drivers/net/ethernet/microchip/lan966x/lan966x_regs.h
713
#define DEV_CLOCK_CFG(t) __REG(TARGET_DEV, t, 8, 0, 0, 1, 28, 0, 0, 1, 4)
drivers/net/ethernet/microchip/lan966x/lan966x_regs.h
752
#define DEV_MAC_ENA_CFG(t) __REG(TARGET_DEV, t, 8, 28, 0, 1, 44, 0, 0, 1, 4)
drivers/net/ethernet/microchip/lan966x/lan966x_regs.h
767
#define DEV_MAC_MODE_CFG(t) __REG(TARGET_DEV, t, 8, 28, 0, 1, 44, 4, 0, 1, 4)
drivers/net/ethernet/microchip/lan966x/lan966x_regs.h
776
#define DEV_MAC_MAXLEN_CFG(t) __REG(TARGET_DEV, t, 8, 28, 0, 1, 44, 8, 0, 1, 4)
drivers/net/ethernet/microchip/lan966x/lan966x_regs.h
785
#define DEV_MAC_TAGS_CFG(t) __REG(TARGET_DEV, t, 8, 28, 0, 1, 44, 12, 0, 1, 4)
drivers/net/ethernet/microchip/lan966x/lan966x_regs.h
800
#define DEV_MAC_IFG_CFG(t) __REG(TARGET_DEV, t, 8, 28, 0, 1, 44, 20, 0, 1, 4)
drivers/net/ethernet/microchip/lan966x/lan966x_regs.h
821
#define DEV_MAC_HDX_CFG(t) __REG(TARGET_DEV, t, 8, 28, 0, 1, 44, 24, 0, 1, 4)
drivers/net/ethernet/microchip/lan966x/lan966x_regs.h
836
#define DEV_FC_MAC_LOW_CFG(t) __REG(TARGET_DEV, t, 8, 28, 0, 1, 44, 32, 0, 1, 4)
drivers/net/ethernet/microchip/lan966x/lan966x_regs.h
839
#define DEV_FC_MAC_HIGH_CFG(t) __REG(TARGET_DEV, t, 8, 28, 0, 1, 44, 36, 0, 1, 4)
drivers/net/ethernet/microchip/lan966x/lan966x_regs.h
842
#define DEV_PCS1G_CFG(t) __REG(TARGET_DEV, t, 8, 72, 0, 1, 68, 0, 0, 1, 4)
drivers/net/ethernet/microchip/lan966x/lan966x_regs.h
851
#define DEV_PCS1G_MODE_CFG(t) __REG(TARGET_DEV, t, 8, 72, 0, 1, 68, 4, 0, 1, 4)
drivers/net/ethernet/microchip/lan966x/lan966x_regs.h
866
#define DEV_PCS1G_SD_CFG(t) __REG(TARGET_DEV, t, 8, 72, 0, 1, 68, 8, 0, 1, 4)
drivers/net/ethernet/microchip/lan966x/lan966x_regs.h
875
#define DEV_PCS1G_ANEG_CFG(t) __REG(TARGET_DEV, t, 8, 72, 0, 1, 68, 12, 0, 1, 4)
drivers/net/ethernet/microchip/lan966x/lan966x_regs.h
902
#define DEV_PCS1G_ANEG_STATUS(t) __REG(TARGET_DEV, t, 8, 72, 0, 1, 68, 32, 0, 1, 4)
drivers/net/ethernet/microchip/lan966x/lan966x_regs.h
917
#define DEV_PCS1G_LINK_STATUS(t) __REG(TARGET_DEV, t, 8, 72, 0, 1, 68, 40, 0, 1, 4)
drivers/net/ethernet/microchip/lan966x/lan966x_regs.h
932
#define DEV_PCS1G_STICKY(t) __REG(TARGET_DEV, t, 8, 72, 0, 1, 68, 48, 0, 1, 4)
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
2672
#define DEV2G5_PHAD_CTRL(t, g) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
2673
__REG(TARGET_DEV2G5, t, regs->tsize[TC_DEV2G5], 200, g, 2, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
2691
#define DEV2G5_PHAD_CTRL(t, g) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
2692
__REG(TARGET_DEV2G5, t, regs->tsize[TC_DEV2G5], 200, g, 2, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
2710
#define DEV10G_MAC_ENA_CFG(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
2711
__REG(TARGET_DEV10G, t, regs->tsize[TC_DEV10G], 0, 0, 1, 60, 0, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
2727
#define DEV10G_MAC_MAXLEN_CFG(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
2728
__REG(TARGET_DEV10G, t, regs->tsize[TC_DEV10G], 0, 0, 1, 60, 8, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
2744
#define DEV10G_MAC_NUM_TAGS_CFG(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
2745
__REG(TARGET_DEV10G, t, regs->tsize[TC_DEV10G], 0, 0, 1, 60, 12, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
2755
#define DEV10G_MAC_TAGS_CFG(t, r) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
2756
__REG(TARGET_DEV10G, t, regs->tsize[TC_DEV10G], 0, 0, 1, 60, 16, r, 3, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
2772
#define DEV10G_MAC_ADV_CHK_CFG(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
2773
__REG(TARGET_DEV10G, t, regs->tsize[TC_DEV10G], 0, 0, 1, 60, 28, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
2819
#define DEV10G_MAC_TX_MONITOR_STICKY(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
2820
__REG(TARGET_DEV10G, t, regs->tsize[TC_DEV10G], 0, 0, 1, 60, 48, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
2854
#define DEV10G_DEV_RST_CTRL(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
2855
__REG(TARGET_DEV10G, t, regs->tsize[TC_DEV10G], 436, 0, 1, 52, 0, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
2913
#define DEV10G_PTP_STAMPER_CFG(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
2914
__REG(TARGET_DEV10G, t, regs->tsize[TC_DEV10G], 436, 0, 1, 52, 20, 0, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
2918
#define DEV10G_PCS25G_CFG(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
2919
__REG(TARGET_DEV10G, t, regs->tsize[TC_DEV10G], 488, 0, 1, 32, 0, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
2930
#define DEV25G_MAC_ENA_CFG(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
2931
__REG(TARGET_DEV25G, t, 8, 0, 0, 1, 60, 0, 0, 1, 4)
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
2947
#define DEV25G_MAC_MAXLEN_CFG(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
2948
__REG(TARGET_DEV25G, t, 8, 0, 0, 1, 60, 8, 0, 1, 4)
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
2964
#define DEV25G_MAC_ADV_CHK_CFG(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
2965
__REG(TARGET_DEV25G, t, 8, 0, 0, 1, 60, 28, 0, 1, 4)
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3011
#define DEV25G_DEV_RST_CTRL(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3012
__REG(TARGET_DEV25G, t, 8, 436, 0, 1, 52, 0, 0, 1, 4)
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3070
#define DEV25G_PCS25G_CFG(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3071
__REG(TARGET_DEV25G, t, 8, 488, 0, 1, 32, 0, 0, 1, 4)
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3081
#define DEV25G_PCS25G_SD_CFG(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3082
__REG(TARGET_DEV25G, t, 8, 488, 0, 1, 32, 4, 0, 1, 4)
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3103
#define DEV2G5_DEV_RST_CTRL(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3104
__REG(TARGET_DEV2G5, t, regs->tsize[TC_DEV2G5], 0, 0, 1, 36, 0, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3156
#define DEV2G5_MAC_ENA_CFG(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3157
__REG(TARGET_DEV2G5, t, regs->tsize[TC_DEV2G5], 52, 0, 1, 36, 0, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3173
#define DEV2G5_MAC_MODE_CFG(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3174
__REG(TARGET_DEV2G5, t, regs->tsize[TC_DEV2G5], 52, 0, 1, 36, 4, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3196
#define DEV2G5_MAC_MAXLEN_CFG(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3197
__REG(TARGET_DEV2G5, t, regs->tsize[TC_DEV2G5], 52, 0, 1, 36, 8, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3207
#define DEV2G5_MAC_TAGS_CFG(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3208
__REG(TARGET_DEV2G5, t, regs->tsize[TC_DEV2G5], 52, 0, 1, 36, 12, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3236
#define DEV2G5_MAC_TAGS_CFG2(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3237
__REG(TARGET_DEV2G5, t, regs->tsize[TC_DEV2G5], 52, 0, 1, 36, 16, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3253
#define DEV2G5_MAC_ADV_CHK_CFG(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3254
__REG(TARGET_DEV2G5, t, regs->tsize[TC_DEV2G5], 52, 0, 1, 36, 20, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3264
#define DEV2G5_MAC_IFG_CFG(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3265
__REG(TARGET_DEV2G5, t, regs->tsize[TC_DEV2G5], 52, 0, 1, 36, 24, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3293
#define DEV2G5_MAC_HDX_CFG(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3294
__REG(TARGET_DEV2G5, t, regs->tsize[TC_DEV2G5], 52, 0, 1, 36, 28, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3328
#define DEV2G5_PCS1G_CFG(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3329
__REG(TARGET_DEV2G5, t, regs->tsize[TC_DEV2G5], 88, 0, 1, 68, 0, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3351
#define DEV2G5_PCS1G_MODE_CFG(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3352
__REG(TARGET_DEV2G5, t, regs->tsize[TC_DEV2G5], 88, 0, 1, 68, 4, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3374
#define DEV2G5_PCS1G_SD_CFG(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3375
__REG(TARGET_DEV2G5, t, regs->tsize[TC_DEV2G5], 88, 0, 1, 68, 8, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3397
#define DEV2G5_PCS1G_ANEG_CFG(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3398
__REG(TARGET_DEV2G5, t, regs->tsize[TC_DEV2G5], 88, 0, 1, 68, 12, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3426
#define DEV2G5_PCS1G_LB_CFG(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3427
__REG(TARGET_DEV2G5, t, regs->tsize[TC_DEV2G5], 88, 0, 1, 68, 20, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3449
#define DEV2G5_PCS1G_ANEG_STATUS(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3450
__REG(TARGET_DEV2G5, t, regs->tsize[TC_DEV2G5], 88, 0, 1, 68, 32, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3478
#define DEV2G5_PCS1G_LINK_STATUS(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3479
__REG(TARGET_DEV2G5, t, regs->tsize[TC_DEV2G5], 88, 0, 1, 68, 40, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3507
#define DEV2G5_PCS1G_STICKY(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3508
__REG(TARGET_DEV2G5, t, regs->tsize[TC_DEV2G5], 88, 0, 1, 68, 48, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3524
#define DEV2G5_PCS_FX100_CFG(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3525
__REG(TARGET_DEV2G5, t, regs->tsize[TC_DEV2G5], 164, 0, 1, 4, 0, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3607
#define DEV2G5_PCS_FX100_STATUS(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3608
__REG(TARGET_DEV2G5, t, regs->tsize[TC_DEV2G5], 168, 0, 1, 4, 0, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3660
#define DEV5G_MAC_ENA_CFG(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3661
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 0, 0, 1, 60, 0, 0, 1, 4)
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3676
#define DEV5G_MAC_MAXLEN_CFG(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3677
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 0, 0, 1, 60, 8, 0, 1, 4)
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3692
#define DEV5G_MAC_ADV_CHK_CFG(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3693
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 0, 0, 1, 60, 28, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3739
#define DEV5G_RX_SYMBOL_ERR_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3740
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 0, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3744
#define DEV5G_RX_PAUSE_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3745
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 4, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3749
#define DEV5G_RX_UNSUP_OPCODE_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3750
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 8, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3754
#define DEV5G_RX_UC_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3755
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 12, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3759
#define DEV5G_RX_MC_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3760
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 16, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3764
#define DEV5G_RX_BC_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3765
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 20, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3769
#define DEV5G_RX_CRC_ERR_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3770
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 24, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3774
#define DEV5G_RX_UNDERSIZE_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3775
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 28, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3779
#define DEV5G_RX_FRAGMENTS_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3780
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 32, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3784
#define DEV5G_RX_IN_RANGE_LEN_ERR_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3785
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 36, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3789
#define DEV5G_RX_OUT_OF_RANGE_LEN_ERR_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3790
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 40, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3794
#define DEV5G_RX_OVERSIZE_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3795
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 44, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3799
#define DEV5G_RX_JABBERS_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3800
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 48, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3804
#define DEV5G_RX_SIZE64_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3805
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 52, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3809
#define DEV5G_RX_SIZE65TO127_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3810
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 56, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3814
#define DEV5G_RX_SIZE128TO255_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3815
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 60, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3819
#define DEV5G_RX_SIZE256TO511_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3820
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 64, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3824
#define DEV5G_RX_SIZE512TO1023_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3825
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 68, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3829
#define DEV5G_RX_SIZE1024TO1518_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3830
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 72, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3834
#define DEV5G_RX_SIZE1519TOMAX_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3835
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 76, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3839
#define DEV5G_RX_IPG_SHRINK_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3840
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 80, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3844
#define DEV5G_TX_PAUSE_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3845
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 84, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3849
#define DEV5G_TX_UC_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3850
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 88, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3854
#define DEV5G_TX_MC_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3855
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 92, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3859
#define DEV5G_TX_BC_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3860
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 96, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3864
#define DEV5G_TX_SIZE64_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3865
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 100, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3869
#define DEV5G_TX_SIZE65TO127_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3870
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 104, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3874
#define DEV5G_TX_SIZE128TO255_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3875
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 108, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3879
#define DEV5G_TX_SIZE256TO511_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3880
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 112, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3884
#define DEV5G_TX_SIZE512TO1023_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3885
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 116, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3889
#define DEV5G_TX_SIZE1024TO1518_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3890
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 120, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3894
#define DEV5G_TX_SIZE1519TOMAX_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3895
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 124, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3899
#define DEV5G_RX_ALIGNMENT_LOST_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3900
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 128, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3904
#define DEV5G_RX_TAGGED_FRMS_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3905
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 132, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3909
#define DEV5G_RX_UNTAGGED_FRMS_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3910
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 136, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3914
#define DEV5G_TX_TAGGED_FRMS_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3915
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 140, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3919
#define DEV5G_TX_UNTAGGED_FRMS_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3920
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 144, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3924
#define DEV5G_PMAC_RX_SYMBOL_ERR_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3925
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 148, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3929
#define DEV5G_PMAC_RX_PAUSE_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3930
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 152, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3934
#define DEV5G_PMAC_RX_UNSUP_OPCODE_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3935
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 156, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3939
#define DEV5G_PMAC_RX_UC_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3940
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 160, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3944
#define DEV5G_PMAC_RX_MC_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3945
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 164, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3949
#define DEV5G_PMAC_RX_BC_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3950
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 168, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3954
#define DEV5G_PMAC_RX_CRC_ERR_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3955
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 172, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3959
#define DEV5G_PMAC_RX_UNDERSIZE_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3960
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 176, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3964
#define DEV5G_PMAC_RX_FRAGMENTS_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3965
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 180, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3969
#define DEV5G_PMAC_RX_IN_RANGE_LEN_ERR_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3970
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 184, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3974
#define DEV5G_PMAC_RX_OUT_OF_RANGE_LEN_ERR_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3975
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 188, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3979
#define DEV5G_PMAC_RX_OVERSIZE_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3980
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 192, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3984
#define DEV5G_PMAC_RX_JABBERS_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3985
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 196, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3989
#define DEV5G_PMAC_RX_SIZE64_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3990
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 200, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3994
#define DEV5G_PMAC_RX_SIZE65TO127_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3995
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 204, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
3999
#define DEV5G_PMAC_RX_SIZE128TO255_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4000
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 208, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4004
#define DEV5G_PMAC_RX_SIZE256TO511_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4005
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 212, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4009
#define DEV5G_PMAC_RX_SIZE512TO1023_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4010
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 216, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4014
#define DEV5G_PMAC_RX_SIZE1024TO1518_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4015
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 220, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4019
#define DEV5G_PMAC_RX_SIZE1519TOMAX_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4020
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 224, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4024
#define DEV5G_PMAC_TX_PAUSE_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4025
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 228, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4029
#define DEV5G_PMAC_TX_UC_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4030
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 232, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4034
#define DEV5G_PMAC_TX_MC_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4035
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 236, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4039
#define DEV5G_PMAC_TX_BC_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4040
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 240, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4044
#define DEV5G_PMAC_TX_SIZE64_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4045
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 244, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4049
#define DEV5G_PMAC_TX_SIZE65TO127_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4050
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 248, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4054
#define DEV5G_PMAC_TX_SIZE128TO255_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4055
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 252, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4059
#define DEV5G_PMAC_TX_SIZE256TO511_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4060
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 256, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4064
#define DEV5G_PMAC_TX_SIZE512TO1023_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4065
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 260, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4069
#define DEV5G_PMAC_TX_SIZE1024TO1518_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4070
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 264, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4074
#define DEV5G_PMAC_TX_SIZE1519TOMAX_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4075
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 268, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4079
#define DEV5G_PMAC_RX_ALIGNMENT_LOST_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4080
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 272, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4084
#define DEV5G_MM_RX_ASSEMBLY_ERR_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4085
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 276, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4089
#define DEV5G_MM_RX_SMD_ERR_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4090
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 280, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4094
#define DEV5G_MM_RX_ASSEMBLY_OK_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4095
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 284, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4099
#define DEV5G_MM_RX_MERGE_FRAG_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4100
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 288, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4104
#define DEV5G_MM_TX_PFRAGMENT_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4105
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 292, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4109
#define DEV5G_RX_HIH_CKSM_ERR_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4110
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 296, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4114
#define DEV5G_RX_XGMII_PROT_ERR_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4115
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 300, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4119
#define DEV5G_PMAC_RX_HIH_CKSM_ERR_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4120
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 304, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4124
#define DEV5G_PMAC_RX_XGMII_PROT_ERR_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4125
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 60, 0, 1, 312, 308, 0, 1,\
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4129
#define DEV5G_RX_IN_BYTES_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4130
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 372, 0, 1, 64, 0, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4134
#define DEV5G_RX_IN_BYTES_MSB_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4135
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 372, 0, 1, 64, 4, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4145
#define DEV5G_RX_OK_BYTES_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4146
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 372, 0, 1, 64, 8, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4150
#define DEV5G_RX_OK_BYTES_MSB_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4151
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 372, 0, 1, 64, 12, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4161
#define DEV5G_RX_BAD_BYTES_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4162
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 372, 0, 1, 64, 16, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4166
#define DEV5G_RX_BAD_BYTES_MSB_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4167
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 372, 0, 1, 64, 20, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4177
#define DEV5G_TX_OUT_BYTES_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4178
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 372, 0, 1, 64, 24, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4182
#define DEV5G_TX_OUT_BYTES_MSB_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4183
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 372, 0, 1, 64, 28, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4193
#define DEV5G_TX_OK_BYTES_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4194
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 372, 0, 1, 64, 32, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4198
#define DEV5G_TX_OK_BYTES_MSB_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4199
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 372, 0, 1, 64, 36, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4209
#define DEV5G_PMAC_RX_OK_BYTES_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4210
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 372, 0, 1, 64, 40, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4214
#define DEV5G_PMAC_RX_OK_BYTES_MSB_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4215
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 372, 0, 1, 64, 44, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4225
#define DEV5G_PMAC_RX_BAD_BYTES_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4226
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 372, 0, 1, 64, 48, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4230
#define DEV5G_PMAC_RX_BAD_BYTES_MSB_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4231
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 372, 0, 1, 64, 52, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4241
#define DEV5G_PMAC_TX_OK_BYTES_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4242
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 372, 0, 1, 64, 56, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4246
#define DEV5G_PMAC_TX_OK_BYTES_MSB_CNT(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4247
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 372, 0, 1, 64, 60, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4257
#define DEV5G_DEV_RST_CTRL(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4258
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 436, 0, 1, 52, 0, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4316
#define DEV5G_PTP_STAMPER_CFG(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
4317
__REG(TARGET_DEV5G, t, regs->tsize[TC_DEV5G], 436, 0, 1, 52, 20, 0, 1, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
5898
#define PCS10G_BR_PCS_CFG(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
5899
__REG(TARGET_PCS10G_BR, t, regs->tsize[TC_PCS10G_BR], 0, 0, 1, 56, 0, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
5975
#define PCS10G_BR_PCS_SD_CFG(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
5976
__REG(TARGET_PCS10G_BR, t, regs->tsize[TC_PCS10G_BR], 0, 0, 1, 56, 4, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
5999
#define PCS25G_BR_PCS_CFG(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
6000
__REG(TARGET_PCS25G_BR, t, 8, 0, 0, 1, 56, 0, 0, 1, 4)
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
6076
#define PCS25G_BR_PCS_SD_CFG(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
6077
__REG(TARGET_PCS25G_BR, t, 8, 0, 0, 1, 56, 4, 0, 1, 4)
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
6098
#define PCS5G_BR_PCS_CFG(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
6099
__REG(TARGET_PCS5G_BR, t, regs->tsize[TC_PCS5G_BR], 0, 0, 1, 56, 0, 0, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
6175
#define PCS5G_BR_PCS_SD_CFG(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
6176
__REG(TARGET_PCS5G_BR, t, regs->tsize[TC_PCS5G_BR], 0, 0, 1, 56, 4, 0, \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
8180
#define DEVRGMII_DEV_RST_CTRL(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
8181
__REG(TARGET_DEVRGMII, t, 2, 0, 0, 1, 36, 0, 0, 1, 4)
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
8191
#define DEVRGMII_MAC_ENA_CFG(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
8192
__REG(TARGET_DEVRGMII, t, 2, 36, 0, 1, 36, 0, 0, 1, 4)
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
8208
#define DEVRGMII_MAC_TAGS_CFG(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
8209
__REG(TARGET_DEVRGMII, t, 2, 36, 0, 1, 36, 12, 0, 1, 4)
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
8237
#define DEVRGMII_MAC_IFG_CFG(t) \
drivers/net/ethernet/microchip/sparx5/sparx5_main_regs.h
8238
__REG(TARGET_DEVRGMII, t, 2, 36, 0, 1, 36, 24, 0, 1, 4)
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
3495
static void myri10ge_watchdog_timer(struct timer_list *t)
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
3503
mgp = timer_container_of(mgp, t, watchdog_timer);
drivers/net/ethernet/natsemi/natsemi.c
1787
static void netdev_timer(struct timer_list *t)
drivers/net/ethernet/natsemi/natsemi.c
1789
struct netdev_private *np = timer_container_of(np, t, timer);
drivers/net/ethernet/natsemi/natsemi.c
613
static void netdev_timer(struct timer_list *t);
drivers/net/ethernet/natsemi/ns83820.c
1587
static void ns83820_tx_watch(struct timer_list *t)
drivers/net/ethernet/natsemi/ns83820.c
1589
struct ns83820 *dev = timer_container_of(dev, t, tx_watchdog);
drivers/net/ethernet/natsemi/ns83820.c
927
static void rx_action(struct tasklet_struct *t)
drivers/net/ethernet/natsemi/ns83820.c
929
struct ns83820 *dev = from_tasklet(dev, t, rx_tasklet);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
1397
void nfp_nfd3_ctrl_poll(struct tasklet_struct *t)
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
1399
struct nfp_net_r_vector *r_vec = from_tasklet(r_vec, t, tasklet);
drivers/net/ethernet/netronome/nfp/nfd3/nfd3.h
100
void nfp_nfd3_ctrl_poll(struct tasklet_struct *t);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1559
void nfp_nfdk_ctrl_poll(struct tasklet_struct *t)
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1561
struct nfp_net_r_vector *r_vec = from_tasklet(r_vec, t, tasklet);
drivers/net/ethernet/netronome/nfp/nfdk/nfdk.h
125
void nfp_nfdk_ctrl_poll(struct tasklet_struct *t);
drivers/net/ethernet/netronome/nfp/nfp_net_common.c
160
static void nfp_net_reconfig_timer(struct timer_list *t)
drivers/net/ethernet/netronome/nfp/nfp_net_common.c
162
struct nfp_net *nn = timer_container_of(nn, t, reconfig_timer);
drivers/net/ethernet/netronome/nfp/nfp_net_dp.h
144
void (*ctrl_poll)(struct tasklet_struct *t);
drivers/net/ethernet/ni/nixge.c
790
static void nixge_dma_err_handler(struct tasklet_struct *t)
drivers/net/ethernet/ni/nixge.c
792
struct nixge_priv *lp = from_tasklet(lp, t, dma_err_tasklet);
drivers/net/ethernet/nvidia/forcedeth.c
1892
static void nv_do_rx_refill(struct timer_list *t)
drivers/net/ethernet/nvidia/forcedeth.c
1894
struct fe_priv *np = timer_container_of(np, t, oom_kick);
drivers/net/ethernet/nvidia/forcedeth.c
4141
static void nv_do_nic_poll(struct timer_list *t)
drivers/net/ethernet/nvidia/forcedeth.c
4143
struct fe_priv *np = timer_container_of(np, t, nic_poll);
drivers/net/ethernet/nvidia/forcedeth.c
4258
static void nv_do_stats_poll(struct timer_list *t)
drivers/net/ethernet/nvidia/forcedeth.c
4262
struct fe_priv *np = timer_container_of(np, t, stats_poll);
drivers/net/ethernet/oki-semi/pch_gbe/pch_gbe_main.c
1024
static void pch_gbe_watchdog(struct timer_list *t)
drivers/net/ethernet/oki-semi/pch_gbe/pch_gbe_main.c
1026
struct pch_gbe_adapter *adapter = timer_container_of(adapter, t,
drivers/net/ethernet/packetengines/hamachi.c
1026
static void hamachi_timer(struct timer_list *t)
drivers/net/ethernet/packetengines/hamachi.c
1028
struct hamachi_private *hmp = timer_container_of(hmp, t, timer);
drivers/net/ethernet/packetengines/hamachi.c
416
static void hamachi_timer(struct timer_list *t);
drivers/net/ethernet/packetengines/hamachi.c
422
void (*media_timer)(struct timer_list *t);
drivers/net/ethernet/packetengines/hamachi.c
552
static void hamachi_timer(struct timer_list *t);
drivers/net/ethernet/packetengines/yellowfin.c
346
static void yellowfin_timer(struct timer_list *t);
drivers/net/ethernet/packetengines/yellowfin.c
653
static void yellowfin_timer(struct timer_list *t)
drivers/net/ethernet/packetengines/yellowfin.c
655
struct yellowfin_private *yp = timer_container_of(yp, t, timer);
drivers/net/ethernet/pasemi/pasemi_mac.c
932
static void pasemi_mac_tx_timer(struct timer_list *t)
drivers/net/ethernet/pasemi/pasemi_mac.c
934
struct pasemi_mac_txring *txring = timer_container_of(txring, t,
drivers/net/ethernet/pensando/ionic/ionic_dev.c
14
static void ionic_watchdog_cb(struct timer_list *t)
drivers/net/ethernet/pensando/ionic/ionic_dev.c
16
struct ionic *ionic = timer_container_of(ionic, t, watchdog_timer);
drivers/net/ethernet/qlogic/qed/qed_int.c
1239
void qed_int_sp_dpc(struct tasklet_struct *t)
drivers/net/ethernet/qlogic/qed/qed_int.c
1241
struct qed_hwfn *p_hwfn = from_tasklet(p_hwfn, t, sp_dpc);
drivers/net/ethernet/qlogic/qed/qed_int.h
150
void qed_int_sp_dpc(struct tasklet_struct *t);
drivers/net/ethernet/qlogic/qede/qede_filter.c
1396
static int qede_flow_get_min_header_size(struct qede_arfs_tuple *t)
drivers/net/ethernet/qlogic/qede/qede_filter.c
1400
if (t->eth_proto == htons(ETH_P_IP))
drivers/net/ethernet/qlogic/qede/qede_filter.c
1405
if (t->ip_proto == IPPROTO_TCP)
drivers/net/ethernet/qlogic/qede/qede_filter.c
1424
static void qede_flow_build_ipv4_hdr(struct qede_arfs_tuple *t,
drivers/net/ethernet/qlogic/qede/qede_filter.c
1431
eth->h_proto = t->eth_proto;
drivers/net/ethernet/qlogic/qede/qede_filter.c
1432
ip->saddr = t->src_ipv4;
drivers/net/ethernet/qlogic/qede/qede_filter.c
1433
ip->daddr = t->dst_ipv4;
drivers/net/ethernet/qlogic/qede/qede_filter.c
1436
ip->protocol = t->ip_proto;
drivers/net/ethernet/qlogic/qede/qede_filter.c
1437
ip->tot_len = cpu_to_be16(qede_flow_get_min_header_size(t) - ETH_HLEN);
drivers/net/ethernet/qlogic/qede/qede_filter.c
1440
ports[0] = t->src_port;
drivers/net/ethernet/qlogic/qede/qede_filter.c
1441
ports[1] = t->dst_port;
drivers/net/ethernet/qlogic/qede/qede_filter.c
1444
static void qede_flow_stringify_ipv4_hdr(struct qede_arfs_tuple *t,
drivers/net/ethernet/qlogic/qede/qede_filter.c
1447
const char *prefix = t->ip_proto == IPPROTO_TCP ? "TCP" : "UDP";
drivers/net/ethernet/qlogic/qede/qede_filter.c
1451
prefix, &t->src_ipv4, t->src_port,
drivers/net/ethernet/qlogic/qede/qede_filter.c
1452
&t->dst_ipv4, t->dst_port);
drivers/net/ethernet/qlogic/qede/qede_filter.c
1471
static void qede_flow_build_ipv6_hdr(struct qede_arfs_tuple *t,
drivers/net/ethernet/qlogic/qede/qede_filter.c
1478
eth->h_proto = t->eth_proto;
drivers/net/ethernet/qlogic/qede/qede_filter.c
1479
memcpy(&ip6->saddr, &t->src_ipv6, sizeof(struct in6_addr));
drivers/net/ethernet/qlogic/qede/qede_filter.c
1480
memcpy(&ip6->daddr, &t->dst_ipv6, sizeof(struct in6_addr));
drivers/net/ethernet/qlogic/qede/qede_filter.c
1483
if (t->ip_proto == IPPROTO_TCP) {
drivers/net/ethernet/qlogic/qede/qede_filter.c
1492
ports[0] = t->src_port;
drivers/net/ethernet/qlogic/qede/qede_filter.c
1493
ports[1] = t->dst_port;
drivers/net/ethernet/qlogic/qede/qede_filter.c
1520
static int qede_set_v4_tuple_to_profile(struct qede_arfs_tuple *t,
drivers/net/ethernet/qlogic/qede/qede_filter.c
1526
if (t->src_port && t->dst_port && t->src_ipv4 && t->dst_ipv4) {
drivers/net/ethernet/qlogic/qede/qede_filter.c
1527
t->mode = QED_FILTER_CONFIG_MODE_5_TUPLE;
drivers/net/ethernet/qlogic/qede/qede_filter.c
1528
} else if (!t->src_port && t->dst_port &&
drivers/net/ethernet/qlogic/qede/qede_filter.c
1529
!t->src_ipv4 && !t->dst_ipv4) {
drivers/net/ethernet/qlogic/qede/qede_filter.c
1530
t->mode = QED_FILTER_CONFIG_MODE_L4_PORT;
drivers/net/ethernet/qlogic/qede/qede_filter.c
1531
} else if (!t->src_port && !t->dst_port &&
drivers/net/ethernet/qlogic/qede/qede_filter.c
1532
!t->dst_ipv4 && t->src_ipv4) {
drivers/net/ethernet/qlogic/qede/qede_filter.c
1533
t->mode = QED_FILTER_CONFIG_MODE_IP_SRC;
drivers/net/ethernet/qlogic/qede/qede_filter.c
1534
} else if (!t->src_port && !t->dst_port &&
drivers/net/ethernet/qlogic/qede/qede_filter.c
1535
t->dst_ipv4 && !t->src_ipv4) {
drivers/net/ethernet/qlogic/qede/qede_filter.c
1536
t->mode = QED_FILTER_CONFIG_MODE_IP_DEST;
drivers/net/ethernet/qlogic/qede/qede_filter.c
1542
t->ip_comp = qede_flow_spec_ipv4_cmp;
drivers/net/ethernet/qlogic/qede/qede_filter.c
1543
t->build_hdr = qede_flow_build_ipv4_hdr;
drivers/net/ethernet/qlogic/qede/qede_filter.c
1544
t->stringify = qede_flow_stringify_ipv4_hdr;
drivers/net/ethernet/qlogic/qede/qede_filter.c
1549
static int qede_set_v6_tuple_to_profile(struct qede_arfs_tuple *t,
drivers/net/ethernet/qlogic/qede/qede_filter.c
1556
if (t->src_port && t->dst_port &&
drivers/net/ethernet/qlogic/qede/qede_filter.c
1557
memcmp(&t->src_ipv6, zaddr, sizeof(struct in6_addr)) &&
drivers/net/ethernet/qlogic/qede/qede_filter.c
1558
memcmp(&t->dst_ipv6, zaddr, sizeof(struct in6_addr))) {
drivers/net/ethernet/qlogic/qede/qede_filter.c
1559
t->mode = QED_FILTER_CONFIG_MODE_5_TUPLE;
drivers/net/ethernet/qlogic/qede/qede_filter.c
1560
} else if (!t->src_port && t->dst_port &&
drivers/net/ethernet/qlogic/qede/qede_filter.c
1561
!memcmp(&t->src_ipv6, zaddr, sizeof(struct in6_addr)) &&
drivers/net/ethernet/qlogic/qede/qede_filter.c
1562
!memcmp(&t->dst_ipv6, zaddr, sizeof(struct in6_addr))) {
drivers/net/ethernet/qlogic/qede/qede_filter.c
1563
t->mode = QED_FILTER_CONFIG_MODE_L4_PORT;
drivers/net/ethernet/qlogic/qede/qede_filter.c
1564
} else if (!t->src_port && !t->dst_port &&
drivers/net/ethernet/qlogic/qede/qede_filter.c
1565
!memcmp(&t->dst_ipv6, zaddr, sizeof(struct in6_addr)) &&
drivers/net/ethernet/qlogic/qede/qede_filter.c
1566
memcmp(&t->src_ipv6, zaddr, sizeof(struct in6_addr))) {
drivers/net/ethernet/qlogic/qede/qede_filter.c
1567
t->mode = QED_FILTER_CONFIG_MODE_IP_SRC;
drivers/net/ethernet/qlogic/qede/qede_filter.c
1568
} else if (!t->src_port && !t->dst_port &&
drivers/net/ethernet/qlogic/qede/qede_filter.c
1569
memcmp(&t->dst_ipv6, zaddr, sizeof(struct in6_addr)) &&
drivers/net/ethernet/qlogic/qede/qede_filter.c
1570
!memcmp(&t->src_ipv6, zaddr, sizeof(struct in6_addr))) {
drivers/net/ethernet/qlogic/qede/qede_filter.c
1571
t->mode = QED_FILTER_CONFIG_MODE_IP_DEST;
drivers/net/ethernet/qlogic/qede/qede_filter.c
1577
t->ip_comp = qede_flow_spec_ipv6_cmp;
drivers/net/ethernet/qlogic/qede/qede_filter.c
1578
t->build_hdr = qede_flow_build_ipv6_hdr;
drivers/net/ethernet/qlogic/qede/qede_filter.c
1585
qede_flow_find_fltr(struct qede_dev *edev, struct qede_arfs_tuple *t)
drivers/net/ethernet/qlogic/qede/qede_filter.c
1594
if (fltr->tuple.ip_proto == t->ip_proto &&
drivers/net/ethernet/qlogic/qede/qede_filter.c
1595
fltr->tuple.src_port == t->src_port &&
drivers/net/ethernet/qlogic/qede/qede_filter.c
1596
fltr->tuple.dst_port == t->dst_port &&
drivers/net/ethernet/qlogic/qede/qede_filter.c
1597
t->ip_comp(&fltr->tuple, t))
drivers/net/ethernet/qlogic/qede/qede_filter.c
1701
qede_flow_parse_ports(struct flow_rule *rule, struct qede_arfs_tuple *t,
drivers/net/ethernet/qlogic/qede/qede_filter.c
1715
t->src_port = match.key->src;
drivers/net/ethernet/qlogic/qede/qede_filter.c
1716
t->dst_port = match.key->dst;
drivers/net/ethernet/qlogic/qede/qede_filter.c
1724
struct qede_arfs_tuple *t,
drivers/net/ethernet/qlogic/qede/qede_filter.c
1746
memcpy(&t->src_ipv6, &match.key->src, sizeof(addr));
drivers/net/ethernet/qlogic/qede/qede_filter.c
1747
memcpy(&t->dst_ipv6, &match.key->dst, sizeof(addr));
drivers/net/ethernet/qlogic/qede/qede_filter.c
1750
err = qede_flow_parse_ports(rule, t, extack);
drivers/net/ethernet/qlogic/qede/qede_filter.c
1754
return qede_set_v6_tuple_to_profile(t, &zero_addr, extack);
drivers/net/ethernet/qlogic/qede/qede_filter.c
1759
struct qede_arfs_tuple *t,
drivers/net/ethernet/qlogic/qede/qede_filter.c
1775
t->src_ipv4 = match.key->src;
drivers/net/ethernet/qlogic/qede/qede_filter.c
1776
t->dst_ipv4 = match.key->dst;
drivers/net/ethernet/qlogic/qede/qede_filter.c
1779
err = qede_flow_parse_ports(rule, t, extack);
drivers/net/ethernet/qlogic/qede/qede_filter.c
1783
return qede_set_v4_tuple_to_profile(t, extack);
drivers/net/ethernet/qlogic/qede/qede_filter.c
1884
struct qede_arfs_tuple t;
drivers/net/ethernet/qlogic/qede/qede_filter.c
1895
rc = qede_parse_flow_attr(proto, f->rule, &t, extack);
drivers/net/ethernet/qlogic/qede/qede_filter.c
1900
if ((edev->arfs->filter_count && edev->arfs->mode != t.mode) ||
drivers/net/ethernet/qlogic/qede/qede_filter.c
1904
t.mode, edev->arfs->mode, edev->arfs->filter_count);
drivers/net/ethernet/qlogic/qede/qede_filter.c
1914
if (qede_flow_find_fltr(edev, &t)) {
drivers/net/ethernet/qlogic/qede/qede_filter.c
1925
min_hlen = qede_flow_get_min_header_size(&t);
drivers/net/ethernet/qlogic/qede/qede_filter.c
1934
memcpy(&n->tuple, &t, sizeof(n->tuple));
drivers/net/ethernet/qlogic/qede/qede_filter.c
1956
struct qede_arfs_tuple *t,
drivers/net/ethernet/qlogic/qede/qede_filter.c
1975
edev->arfs->mode != t->mode) {
drivers/net/ethernet/qlogic/qede/qede_filter.c
1978
t->mode, edev->arfs->filter_count);
drivers/net/ethernet/qlogic/qede/qede_filter.c
1990
struct qede_arfs_tuple *t,
drivers/net/ethernet/qlogic/qede/qede_filter.c
2023
err = qede_parse_flow_attr(proto, flow->rule, t, &extack);
drivers/net/ethernet/qlogic/qede/qede_filter.c
2028
err = qede_flow_spec_validate(edev, &flow->rule->action, t,
drivers/net/ethernet/qlogic/qede/qede_filter.c
2042
struct qede_arfs_tuple t;
drivers/net/ethernet/qlogic/qede/qede_filter.c
2053
rc = qede_flow_spec_to_rule(edev, &t, fsp);
drivers/net/ethernet/qlogic/qede/qede_filter.c
2057
if (qede_flow_find_fltr(edev, &t)) {
drivers/net/ethernet/qlogic/qede/qede_filter.c
2068
min_hlen = qede_flow_get_min_header_size(&t);
drivers/net/ethernet/qlogic/qede/qede_filter.c
2080
memcpy(&n->tuple, &t, sizeof(n->tuple));
drivers/net/ethernet/qlogic/qede/qede_filter.c
38
void (*build_hdr)(struct qede_arfs_tuple *t, void *header);
drivers/net/ethernet/qlogic/qede/qede_filter.c
41
void (*stringify)(struct qede_arfs_tuple *t, void *buffer);
drivers/net/ethernet/qlogic/qla3xxx.c
3735
static void ql3xxx_timer(struct timer_list *t)
drivers/net/ethernet/qlogic/qla3xxx.c
3737
struct ql3_adapter *qdev = timer_container_of(qdev, t, adapter_timer);
drivers/net/ethernet/qualcomm/rmnet/rmnet_map_data.c
571
static enum hrtimer_restart rmnet_map_flush_tx_packet_queue(struct hrtimer *t)
drivers/net/ethernet/qualcomm/rmnet/rmnet_map_data.c
575
port = container_of(t, struct rmnet_port, hrtimer);
drivers/net/ethernet/rocker/rocker_ofdpa.c
1983
static void ofdpa_fdb_cleanup(struct timer_list *t)
drivers/net/ethernet/rocker/rocker_ofdpa.c
1985
struct ofdpa *ofdpa = timer_container_of(ofdpa, t, fdb_cleanup_timer);
drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c
1012
static void sxgbe_tx_timer(struct timer_list *t)
drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c
1014
struct sxgbe_tx_queue *p = timer_container_of(p, t, txtimer);
drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c
105
static void sxgbe_eee_ctrl_timer(struct timer_list *t)
drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c
107
struct sxgbe_priv_data *priv = timer_container_of(priv, t,
drivers/net/ethernet/seeq/ether3.c
171
static void ether3_ledoff(struct timer_list *t)
drivers/net/ethernet/seeq/ether3.c
173
struct dev_priv *private = timer_container_of(private, t, timer);
drivers/net/ethernet/seeq/sgiseeq.c
253
struct sgiseeq_tx_desc *t = gpriv->tx_desc;
drivers/net/ethernet/seeq/sgiseeq.c
272
i, (&t[i]), t[i].tdma.pbuf, t[i].tdma.cntinfo,
drivers/net/ethernet/seeq/sgiseeq.c
273
t[i].tdma.pnext);
drivers/net/ethernet/seeq/sgiseeq.c
276
i, (&t[i]), t[i].tdma.pbuf, t[i].tdma.cntinfo,
drivers/net/ethernet/seeq/sgiseeq.c
277
t[i].tdma.pnext);
drivers/net/ethernet/sfc/falcon/efx.h
45
void ef4_rx_slow_fill(struct timer_list *t);
drivers/net/ethernet/sfc/falcon/falcon.c
1454
static void falcon_stats_timer_func(struct timer_list *t)
drivers/net/ethernet/sfc/falcon/falcon.c
1456
struct falcon_nic_data *nic_data = timer_container_of(nic_data, t,
drivers/net/ethernet/sfc/falcon/rx.c
383
void ef4_rx_slow_fill(struct timer_list *t)
drivers/net/ethernet/sfc/falcon/rx.c
385
struct ef4_rx_queue *rx_queue = timer_container_of(rx_queue, t,
drivers/net/ethernet/sfc/mcdi.c
47
static void efx_mcdi_timeout_async(struct timer_list *t);
drivers/net/ethernet/sfc/mcdi.c
606
static void efx_mcdi_timeout_async(struct timer_list *t)
drivers/net/ethernet/sfc/mcdi.c
608
struct efx_mcdi_iface *mcdi = timer_container_of(mcdi, t, async_timer);
drivers/net/ethernet/sfc/rx_common.c
347
void efx_rx_slow_fill(struct timer_list *t)
drivers/net/ethernet/sfc/rx_common.c
349
struct efx_rx_queue *rx_queue = timer_container_of(rx_queue, t,
drivers/net/ethernet/sfc/rx_common.h
46
void efx_rx_slow_fill(struct timer_list *t);
drivers/net/ethernet/sfc/siena/mcdi.c
48
static void efx_mcdi_timeout_async(struct timer_list *t);
drivers/net/ethernet/sfc/siena/mcdi.c
610
static void efx_mcdi_timeout_async(struct timer_list *t)
drivers/net/ethernet/sfc/siena/mcdi.c
612
struct efx_mcdi_iface *mcdi = timer_container_of(mcdi, t, async_timer);
drivers/net/ethernet/sfc/siena/rx_common.c
344
void efx_siena_rx_slow_fill(struct timer_list *t)
drivers/net/ethernet/sfc/siena/rx_common.c
346
struct efx_rx_queue *rx_queue = timer_container_of(rx_queue, t,
drivers/net/ethernet/sfc/siena/rx_common.h
46
void efx_siena_rx_slow_fill(struct timer_list *t);
drivers/net/ethernet/sgi/ioc3-eth.c
571
static void ioc3_timer(struct timer_list *t)
drivers/net/ethernet/sgi/ioc3-eth.c
573
struct ioc3_private *ip = timer_container_of(ip, t, ioc3_timer);
drivers/net/ethernet/silan/sc92031.c
833
static void sc92031_tasklet(struct tasklet_struct *t)
drivers/net/ethernet/silan/sc92031.c
835
struct sc92031_priv *priv = from_tasklet(priv, t, tasklet);
drivers/net/ethernet/sis/sis190.c
1024
static void sis190_phy_timer(struct timer_list *t)
drivers/net/ethernet/sis/sis190.c
1026
struct sis190_private *tp = timer_container_of(tp, t, timer);
drivers/net/ethernet/sis/sis900.c
1304
static void sis900_timer(struct timer_list *t)
drivers/net/ethernet/sis/sis900.c
1306
struct sis900_private *sis_priv = timer_container_of(sis_priv, t,
drivers/net/ethernet/sis/sis900.c
219
static void sis900_timer(struct timer_list *t);
drivers/net/ethernet/smsc/epic100.c
286
static void epic_timer(struct timer_list *t);
drivers/net/ethernet/smsc/epic100.c
838
static void epic_timer(struct timer_list *t)
drivers/net/ethernet/smsc/epic100.c
840
struct epic_private *ep = timer_container_of(ep, t, timer);
drivers/net/ethernet/smsc/smc91c92_cs.c
1714
static void media_check(struct timer_list *t)
drivers/net/ethernet/smsc/smc91c92_cs.c
1716
struct smc_private *smc = timer_container_of(smc, t, media);
drivers/net/ethernet/smsc/smc91c92_cs.c
283
static void media_check(struct timer_list *t);
drivers/net/ethernet/smsc/smc91x.c
531
static void smc_hardware_send_pkt(struct tasklet_struct *t)
drivers/net/ethernet/smsc/smc91x.c
533
struct smc_local *lp = from_tasklet(lp, t, tx_task);
drivers/net/ethernet/spacemit/k1_emac.c
1148
static void emac_stats_timer(struct timer_list *t)
drivers/net/ethernet/spacemit/k1_emac.c
1150
struct emac_priv *priv = timer_container_of(priv, t, stats_timer);
drivers/net/ethernet/spacemit/k1_emac.c
299
static void emac_tx_coal_timer(struct timer_list *t)
drivers/net/ethernet/spacemit/k1_emac.c
301
struct emac_priv *priv = timer_container_of(priv, t, txtimer);
drivers/net/ethernet/stmicro/stmmac/dwmac-sun8i.c
615
u32 t, r;
drivers/net/ethernet/stmicro/stmmac/dwmac-sun8i.c
617
t = readl(ioaddr + EMAC_TX_CTL0);
drivers/net/ethernet/stmicro/stmmac/dwmac-sun8i.c
620
t |= EMAC_TX_TRANSMITTER_EN;
drivers/net/ethernet/stmicro/stmmac/dwmac-sun8i.c
623
t &= ~EMAC_TX_TRANSMITTER_EN;
drivers/net/ethernet/stmicro/stmmac/dwmac-sun8i.c
626
writel(t, ioaddr + EMAC_TX_CTL0);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3340
static enum hrtimer_restart stmmac_tx_timer(struct hrtimer *t)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3342
struct stmmac_tx_queue *tx_q = container_of(t, struct stmmac_tx_queue, txtimer);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
446
static void stmmac_eee_ctrl_timer(struct timer_list *t)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
448
struct stmmac_priv *priv = timer_container_of(priv, t, eee_ctrl_timer);
drivers/net/ethernet/sun/cassini.c
4022
static void cas_link_timer(struct timer_list *t)
drivers/net/ethernet/sun/cassini.c
4024
struct cas *cp = timer_container_of(cp, t, link_timer);
drivers/net/ethernet/sun/niu.c
2226
static void niu_timer(struct timer_list *t)
drivers/net/ethernet/sun/niu.c
2228
struct niu *np = timer_container_of(np, t, timer);
drivers/net/ethernet/sun/sunbmac.c
527
static void bigmac_timer(struct timer_list *t)
drivers/net/ethernet/sun/sunbmac.c
529
struct bigmac *bp = timer_container_of(bp, t, bigmac_timer);
drivers/net/ethernet/sun/sungem.c
1482
static void gem_link_timer(struct timer_list *t)
drivers/net/ethernet/sun/sungem.c
1484
struct gem *gp = timer_container_of(gp, t, link_timer);
drivers/net/ethernet/sun/sunhme.c
722
static void happy_meal_timer(struct timer_list *t)
drivers/net/ethernet/sun/sunhme.c
724
struct happy_meal *hp = timer_container_of(hp, t, happy_timer);
drivers/net/ethernet/sun/sunvnet_common.c
1045
void sunvnet_clean_timer_expire_common(struct timer_list *t)
drivers/net/ethernet/sun/sunvnet_common.c
1047
struct vnet_port *port = timer_container_of(port, t, clean_timer);
drivers/net/ethernet/sun/sunvnet_common.h
133
void sunvnet_clean_timer_expire_common(struct timer_list *t);
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
361
static void xlgmac_tx_timer(struct timer_list *t)
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
363
struct xlgmac_channel *channel = timer_container_of(channel, t,
drivers/net/ethernet/ti/cpsw_ale.c
1226
static void cpsw_ale_timer(struct timer_list *t)
drivers/net/ethernet/ti/cpsw_ale.c
1228
struct cpsw_ale *ale = timer_container_of(ale, t, timer);
drivers/net/ethernet/ti/icssm/icssm_prueth_switch.c
213
struct fdb_tbl *t = prueth->fdb_tbl;
drivers/net/ethernet/ti/icssm/icssm_prueth_switch.c
219
t->index_a = sram_base + V2_1_FDB_TBL_OFFSET;
drivers/net/ethernet/ti/icssm/icssm_prueth_switch.c
220
t->mac_tbl_a = sram_base + FDB_MAC_TBL_OFFSET;
drivers/net/ethernet/ti/icssm/icssm_prueth_switch.c
221
t->port1_stp_cfg = sram_base + FDB_PORT1_STP_CFG_OFFSET;
drivers/net/ethernet/ti/icssm/icssm_prueth_switch.c
222
t->port2_stp_cfg = sram_base + FDB_PORT2_STP_CFG_OFFSET;
drivers/net/ethernet/ti/icssm/icssm_prueth_switch.c
223
t->flood_enable_flags = sram_base + FDB_FLOOD_ENABLE_FLAGS_OFFSET;
drivers/net/ethernet/ti/icssm/icssm_prueth_switch.c
224
t->locks = sram_base + FDB_LOCKS_OFFSET;
drivers/net/ethernet/ti/icssm/icssm_prueth_switch.c
226
val = readb(t->flood_enable_flags);
drivers/net/ethernet/ti/icssm/icssm_prueth_switch.c
233
writeb(val, t->flood_enable_flags);
drivers/net/ethernet/ti/icssm/icssm_prueth_switch.c
235
writeb(0, &t->locks->host_lock);
drivers/net/ethernet/ti/icssm/icssm_prueth_switch.c
236
t->total_entries = 0;
drivers/net/ethernet/ti/icssm/icssm_switchdev.c
30
struct fdb_tbl *t = prueth->fdb_tbl;
drivers/net/ethernet/ti/icssm/icssm_switchdev.c
32
writeb(state, port - 1 ? (void __iomem *)&t->port2_stp_cfg->state :
drivers/net/ethernet/ti/icssm/icssm_switchdev.c
33
(void __iomem *)&t->port1_stp_cfg->state);
drivers/net/ethernet/ti/icssm/icssm_switchdev.c
38
struct fdb_tbl *t = prueth->fdb_tbl;
drivers/net/ethernet/ti/icssm/icssm_switchdev.c
41
state = readb(port - 1 ? (void __iomem *)&t->port2_stp_cfg->state :
drivers/net/ethernet/ti/icssm/icssm_switchdev.c
42
(void __iomem *)&t->port1_stp_cfg->state);
drivers/net/ethernet/ti/netcp_ethss.c
2838
static void netcp_ethss_timer(struct timer_list *t)
drivers/net/ethernet/ti/netcp_ethss.c
2840
struct gbe_priv *gbe_dev = timer_container_of(gbe_dev, t, timer);
drivers/net/ethernet/ti/tlan.c
177
static void tlan_timer(struct timer_list *t);
drivers/net/ethernet/ti/tlan.c
178
static void tlan_phy_monitor(struct timer_list *t);
drivers/net/ethernet/ti/tlan.c
1816
static void tlan_timer(struct timer_list *t)
drivers/net/ethernet/ti/tlan.c
1818
struct tlan_priv *priv = timer_container_of(priv, t, timer);
drivers/net/ethernet/ti/tlan.c
2747
static void tlan_phy_monitor(struct timer_list *t)
drivers/net/ethernet/ti/tlan.c
2749
struct tlan_priv *priv = timer_container_of(priv, t, media_timer);
drivers/net/ethernet/toshiba/ps3_gelic_net.c
1101
static void gelic_rx_oom_timer(struct timer_list *t)
drivers/net/ethernet/toshiba/ps3_gelic_net.c
1103
struct gelic_card *card = timer_container_of(card, t, rx_oom_timer);
drivers/net/ethernet/tundra/tsi108_eth.c
144
static void tsi108_timed_checker(struct timer_list *t);
drivers/net/ethernet/tundra/tsi108_eth.c
1653
static void tsi108_timed_checker(struct timer_list *t)
drivers/net/ethernet/tundra/tsi108_eth.c
1655
struct tsi108_prv_data *data = timer_container_of(data, t, timer);
drivers/net/ethernet/wangxun/libwx/wx_lib.c
3333
void wx_service_timer(struct timer_list *t)
drivers/net/ethernet/wangxun/libwx/wx_lib.c
3335
struct wx *wx = timer_container_of(wx, t, service_timer);
drivers/net/ethernet/wangxun/libwx/wx_lib.h
43
void wx_service_timer(struct timer_list *t);
drivers/net/fddi/defza.c
1045
static void fza_reset_timer(struct timer_list *t)
drivers/net/fddi/defza.c
1047
struct fza_private *fp = timer_container_of(fp, t, reset_timer);
drivers/net/fddi/defza.c
1156
long t;
drivers/net/fddi/defza.c
1200
t = wait_event_timeout(fp->cmd_done_wait, fp->cmd_done_flag, 3 * HZ);
drivers/net/fddi/defza.c
1214
(3 * HZ - t) * 1000 / HZ);
drivers/net/fddi/defza.c
1224
long t;
drivers/net/fddi/defza.c
1240
t = wait_event_timeout(fp->state_chg_wait, fp->state_chg_flag,
drivers/net/fddi/defza.c
1252
(15 * HZ - t) * 1000 / HZ);
drivers/net/fddi/defza.c
252
long t;
drivers/net/fddi/defza.c
265
t = wait_event_timeout(fp->state_chg_wait, fp->state_chg_flag,
drivers/net/fddi/defza.c
280
(45 * HZ - t) * 1000 / HZ);
drivers/net/fddi/defza.c
384
long t;
drivers/net/fddi/defza.c
397
t = wait_event_timeout(fp->cmd_done_wait, fp->cmd_done_flag, 3 * HZ);
drivers/net/fddi/defza.c
411
(3 * HZ - t) * 1000 / HZ);
drivers/net/fddi/skfp/h/cmtdef.h
382
#define MIB2US(t) ((t)/12)
drivers/net/fddi/skfp/h/fplustm.h
83
struct s_smt_fp_txd t ; /* pointer to the TxD */
drivers/net/fddi/skfp/h/mbuf.h
43
#define smtod(x,t) ((t)((x)->sm_data + (x)->sm_off))
drivers/net/fddi/skfp/h/mbuf.h
44
#define smtodoff(x,t,o) ((t)((x)->sm_data + (o)))
drivers/net/fddi/skfp/h/skfbi.h
892
#define COUNT(t) ((t)<<6) /* counter */
drivers/net/fddi/skfp/h/smc.h
450
struct s_timer t ; /* timer */
drivers/net/fddi/skfp/h/smt.h
110
#define SMTSETPARA(p,t) (p)->para.p_type = (t),\
drivers/net/fddi/skfp/h/supern_2.h
179
} t ;
drivers/net/fddi/skfp/h/supern_2.h
209
} t ;
drivers/net/fddi/skfp/hwmtm.c
1623
struct s_smt_fp_txd volatile *t ;
drivers/net/fddi/skfp/hwmtm.c
1635
t = queue->tx_curr_put ;
drivers/net/fddi/skfp/hwmtm.c
1640
DB_TX(3, "LAN_TX: TxD = %p, virt = %p", t, virt);
drivers/net/fddi/skfp/hwmtm.c
1641
t->txd_virt = virt ;
drivers/net/fddi/skfp/hwmtm.c
1642
t->txd_txdscr = cpu_to_le32(smc->os.hwm.tx_descr) ;
drivers/net/fddi/skfp/hwmtm.c
1643
t->txd_tbadr = cpu_to_le32(phys) ;
drivers/net/fddi/skfp/hwmtm.c
1647
t->txd_tbctrl = tbctrl ;
drivers/net/fddi/skfp/hwmtm.c
1650
DRV_BUF_FLUSH(t,DDI_DMA_SYNC_FORDEV) ;
drivers/net/fddi/skfp/hwmtm.c
1653
DRV_BUF_FLUSH(t,DDI_DMA_SYNC_FORDEV) ;
drivers/net/fddi/skfp/hwmtm.c
1663
queue->tx_curr_put = t->txd_next ;
drivers/net/fddi/skfp/hwmtm.c
1680
hwm_cpy_txd2mb(t,smc->os.hwm.tx_data,
drivers/net/fddi/skfp/hwmtm.c
1702
hwm_cpy_txd2mb(t,smc->os.hwm.tx_data,
drivers/net/fddi/skfp/hwmtm.c
1719
NDD_TRACE("THfE",t,queue->tx_free,0) ;
drivers/net/fddi/skfp/hwmtm.c
1809
struct s_smt_fp_txd volatile *t ;
drivers/net/fddi/skfp/hwmtm.c
1873
t = queue->tx_curr_put ;
drivers/net/fddi/skfp/hwmtm.c
1876
DB_TX(5, "init TxD = 0x%p", t);
drivers/net/fddi/skfp/hwmtm.c
1879
t->txd_txdscr = cpu_to_le32(TX_DESCRIPTOR |
drivers/net/fddi/skfp/hwmtm.c
1882
t->txd_virt = virt[i] ;
drivers/net/fddi/skfp/hwmtm.c
1885
t->txd_tbadr = cpu_to_le32(phys) ;
drivers/net/fddi/skfp/hwmtm.c
1889
t->txd_tbctrl = tbctrl ;
drivers/net/fddi/skfp/hwmtm.c
1891
DRV_BUF_FLUSH(t,DDI_DMA_SYNC_FORDEV) ;
drivers/net/fddi/skfp/hwmtm.c
1894
DRV_BUF_FLUSH(t,DDI_DMA_SYNC_FORDEV) ;
drivers/net/fddi/skfp/hwmtm.c
1898
queue->tx_curr_put = t = t->txd_next ;
drivers/net/fddi/skfp/hwmtm.c
1916
NDD_TRACE("THSE",t,queue->tx_free,frag_count) ;
drivers/net/fddi/skfp/hwmtm.c
2025
struct s_smt_fp_txd volatile *t ;
drivers/net/fddi/skfp/hwmtm.c
2043
t = queue->tx_curr_get ;
drivers/net/fddi/skfp/hwmtm.c
2046
DRV_BUF_FLUSH(t,DDI_DMA_SYNC_FORCPU) ;
drivers/net/fddi/skfp/hwmtm.c
2047
DB_TX(5, "switch OWN bit of TxD 0x%p", t);
drivers/net/fddi/skfp/hwmtm.c
2048
t->txd_tbctrl &= ~cpu_to_le32(BMU_OWN) ;
drivers/net/fddi/skfp/hwmtm.c
2049
DRV_BUF_FLUSH(t,DDI_DMA_SYNC_FORDEV) ;
drivers/net/fddi/skfp/hwmtm.c
2050
t = t->txd_next ;
drivers/net/fddi/skfp/hwmtm.c
2062
t = queue->tx_curr_get ;
drivers/net/fddi/skfp/hwmtm.c
2070
outpd(ADDR(B5_XS_DA),le32_to_cpu(t->txd_ntdadr)) ;
drivers/net/fddi/skfp/hwmtm.c
2073
outpd(ADDR(B5_XA_DA),le32_to_cpu(t->txd_ntdadr)) ;
drivers/net/fddi/skfp/hwmtm.c
584
struct s_smt_fp_txd volatile *t ;
drivers/net/fddi/skfp/hwmtm.c
588
t = queue->tx_curr_get ;
drivers/net/fddi/skfp/hwmtm.c
591
t = t->txd_next ;
drivers/net/fddi/skfp/hwmtm.c
593
phys = le32_to_cpu(t->txd_ntdadr) ;
drivers/net/fddi/skfp/hwmtm.c
595
t = queue->tx_curr_get ;
drivers/net/fddi/skfp/hwmtm.c
597
DRV_BUF_FLUSH(t,DDI_DMA_SYNC_FORCPU) ;
drivers/net/fddi/skfp/hwmtm.c
598
tbctrl = le32_to_cpu(t->txd_tbctrl) ;
drivers/net/fddi/skfp/hwmtm.c
608
t->txd_tbctrl &= ~cpu_to_le32(BMU_OWN) ;
drivers/net/fddi/skfp/hwmtm.c
611
phys = le32_to_cpu(t->txd_ntdadr) ;
drivers/net/fddi/skfp/hwmtm.c
612
DRV_BUF_FLUSH(t,DDI_DMA_SYNC_FORDEV) ;
drivers/net/fddi/skfp/hwmtm.c
613
t = t->txd_next ;
drivers/net/fddi/skfp/smttimer.c
104
smt_timer_start(smc,&smc->t.st_fast,32L, EV_TOKEN(EVENT_SMT,SM_FAST));
drivers/net/fddi/skfp/smttimer.c
121
last = &smc->t.st_queue ;
drivers/net/fddi/skfp/smttimer.c
122
tm = smc->t.st_queue ;
drivers/net/fddi/skfp/smttimer.c
137
next = smc->t.st_queue ;
drivers/net/fddi/skfp/smttimer.c
138
smc->t.st_queue = tm ;
drivers/net/fddi/skfp/smttimer.c
145
if (restart && smc->t.st_queue)
drivers/net/fddi/skfp/smttimer.c
146
hwt_start(smc,smc->t.st_queue->tm_delta) ;
drivers/net/fddi/skfp/smttimer.c
25
smc->t.st_queue = NULL;
drivers/net/fddi/skfp/smttimer.c
26
smc->t.st_fast.tm_active = FALSE ;
drivers/net/fddi/skfp/smttimer.c
27
smc->t.st_fast.tm_next = NULL;
drivers/net/fddi/skfp/smttimer.c
40
if (smc->t.st_queue == timer && !timer->tm_next) {
drivers/net/fddi/skfp/smttimer.c
43
for (prev = &smc->t.st_queue ; (tm = *prev) ; prev = &tm->tm_next ) {
drivers/net/fddi/skfp/smttimer.c
68
if (!smc->t.st_queue) {
drivers/net/fddi/skfp/smttimer.c
69
smc->t.st_queue = timer ;
drivers/net/fddi/skfp/smttimer.c
84
for (prev = &smc->t.st_queue ; (tm = *prev) ; prev = &tm->tm_next ) {
drivers/net/fddi/skfp/smttimer.c
99
hwt_start(smc,smc->t.st_queue->tm_delta) ;
drivers/net/geneve.c
1791
struct geneve_dev *geneve, *t = NULL;
drivers/net/geneve.c
1803
t = geneve;
drivers/net/geneve.c
1805
return t;
drivers/net/geneve.c
1830
struct geneve_dev *t, *geneve = netdev_priv(dev);
drivers/net/geneve.c
1844
t = geneve_find_dev(gn, info, &tun_on_same_port, &tun_collect_md);
drivers/net/geneve.c
1845
if (t)
drivers/net/hamradio/6pack.c
132
static void sp_xmit_on_air(struct timer_list *t)
drivers/net/hamradio/6pack.c
134
struct sixpack *sp = timer_container_of(sp, t, tx_t);
drivers/net/hamradio/6pack.c
457
static void resync_tnc(struct timer_list *t)
drivers/net/hamradio/6pack.c
459
struct sixpack *sp = timer_container_of(sp, t, resync_t);
drivers/net/hamradio/scc.c
1017
void (*handler)(struct timer_list *t),
drivers/net/hamradio/scc.c
1128
static void t_dwait(struct timer_list *t)
drivers/net/hamradio/scc.c
1130
struct scc_channel *scc = timer_container_of(scc, t, tx_t);
drivers/net/hamradio/scc.c
1170
static void t_txdelay(struct timer_list *t)
drivers/net/hamradio/scc.c
1172
struct scc_channel *scc = timer_container_of(scc, t, tx_t);
drivers/net/hamradio/scc.c
1191
static void t_tail(struct timer_list *t)
drivers/net/hamradio/scc.c
1193
struct scc_channel *scc = timer_container_of(scc, t, tx_t);
drivers/net/hamradio/scc.c
1218
static void t_busy(struct timer_list *t)
drivers/net/hamradio/scc.c
1220
struct scc_channel *scc = timer_container_of(scc, t, tx_wdog);
drivers/net/hamradio/scc.c
1237
static void t_maxkeyup(struct timer_list *t)
drivers/net/hamradio/scc.c
1239
struct scc_channel *scc = timer_container_of(scc, t, tx_wdog);
drivers/net/hamradio/scc.c
1271
static void t_idle(struct timer_list *t)
drivers/net/hamradio/scc.c
1273
struct scc_channel *scc = timer_container_of(scc, t, tx_t);
drivers/net/hamradio/scc.c
1404
static void scc_stop_calibrate(struct timer_list *t)
drivers/net/hamradio/scc.c
1406
struct scc_channel *scc = timer_container_of(scc, t, tx_wdog);
drivers/net/hamradio/scc.c
189
static void t_dwait(struct timer_list *t);
drivers/net/hamradio/scc.c
190
static void t_txdelay(struct timer_list *t);
drivers/net/hamradio/scc.c
191
static void t_tail(struct timer_list *t);
drivers/net/hamradio/scc.c
194
static void t_idle(struct timer_list *t);
drivers/net/hamradio/scc.c
999
void (*handler)(struct timer_list *t),
drivers/net/hyperv/netvsc_drv.c
1314
struct rtnl_link_stats64 *t)
drivers/net/hyperv/netvsc_drv.c
1327
netdev_stats_to_stats64(t, &net->stats);
drivers/net/hyperv/netvsc_drv.c
1330
t->rx_packets += vf_tot.rx_packets;
drivers/net/hyperv/netvsc_drv.c
1331
t->tx_packets += vf_tot.tx_packets;
drivers/net/hyperv/netvsc_drv.c
1332
t->rx_bytes += vf_tot.rx_bytes;
drivers/net/hyperv/netvsc_drv.c
1333
t->tx_bytes += vf_tot.tx_bytes;
drivers/net/hyperv/netvsc_drv.c
1334
t->tx_dropped += vf_tot.tx_dropped;
drivers/net/hyperv/netvsc_drv.c
1350
t->tx_bytes += bytes;
drivers/net/hyperv/netvsc_drv.c
1351
t->tx_packets += packets;
drivers/net/hyperv/netvsc_drv.c
1361
t->rx_bytes += bytes;
drivers/net/hyperv/netvsc_drv.c
1362
t->rx_packets += packets;
drivers/net/hyperv/netvsc_drv.c
1363
t->multicast += multicast;
drivers/net/ifb.c
89
static void ifb_ri_tasklet(struct tasklet_struct *t)
drivers/net/ifb.c
91
struct ifb_q_private *txp = from_tasklet(txp, t, ifb_tasklet);
drivers/net/ntb_netdev.c
230
static void ntb_netdev_tx_timer(struct timer_list *t)
drivers/net/ntb_netdev.c
232
struct ntb_netdev *dev = timer_container_of(dev, t, tx_timer);
drivers/net/phy/phylink.c
1750
static void phylink_fixed_poll(struct timer_list *t)
drivers/net/phy/phylink.c
1752
struct phylink *pl = container_of(t, struct phylink, link_poll);
drivers/net/phy/phylink.c
1754
mod_timer(t, jiffies + HZ);
drivers/net/ppp/ppp_async.c
105
static void ppp_async_process(struct tasklet_struct *t);
drivers/net/ppp/ppp_async.c
486
static void ppp_async_process(struct tasklet_struct *t)
drivers/net/ppp/ppp_async.c
488
struct asyncppp *ap = from_tasklet(ap, t, tsk);
drivers/net/ppp/ppp_synctty.c
477
static void ppp_sync_process(struct tasklet_struct *t)
drivers/net/ppp/ppp_synctty.c
479
struct syncppp *ap = from_tasklet(ap, t, tsk);
drivers/net/ppp/ppp_synctty.c
93
static void ppp_sync_process(struct tasklet_struct *t);
drivers/net/slip/slip.c
110
static void sl_keepalive(struct timer_list *t);
drivers/net/slip/slip.c
111
static void sl_outfill(struct timer_list *t);
drivers/net/slip/slip.c
1378
static void sl_outfill(struct timer_list *t)
drivers/net/slip/slip.c
1380
struct slip *sl = timer_container_of(sl, t, outfill_timer);
drivers/net/slip/slip.c
1409
static void sl_keepalive(struct timer_list *t)
drivers/net/slip/slip.c
1411
struct slip *sl = timer_container_of(sl, t, keepalive_timer);
drivers/net/tun.c
379
static void tun_flow_cleanup(struct timer_list *t)
drivers/net/tun.c
381
struct tun_struct *tun = timer_container_of(tun, t, flow_gc_timer);
drivers/net/usb/catc.c
607
static void catc_stats_timer(struct timer_list *t)
drivers/net/usb/catc.c
609
struct catc *catc = timer_container_of(catc, t, timer);
drivers/net/usb/cdc_ncm.c
1501
static void cdc_ncm_txpath_bh(struct tasklet_struct *t)
drivers/net/usb/cdc_ncm.c
1503
struct cdc_ncm_ctx *ctx = from_tasklet(ctx, t, bh);
drivers/net/usb/cdc_ncm.c
65
static void cdc_ncm_txpath_bh(struct tasklet_struct *t);
drivers/net/usb/hso.c
1212
static void hso_unthrottle_tasklet(struct tasklet_struct *t)
drivers/net/usb/hso.c
1214
struct hso_serial *serial = from_tasklet(serial, t,
drivers/net/usb/lan78xx.c
4616
static void lan78xx_stat_monitor(struct timer_list *t)
drivers/net/usb/lan78xx.c
4618
struct lan78xx_net *dev = timer_container_of(dev, t, stat_monitor);
drivers/net/usb/pegasus.c
192
__le16 *t = (__le16 *) & data[1];
drivers/net/usb/pegasus.c
193
*t = cpu_to_le16(*regd);
drivers/net/usb/pegasus.c
575
static void rx_fixup(struct tasklet_struct *t)
drivers/net/usb/pegasus.c
577
pegasus_t *pegasus = from_tasklet(pegasus, t, rx_tl);
drivers/net/usb/r8152.c
2750
static void bottom_half(struct tasklet_struct *t)
drivers/net/usb/r8152.c
2752
struct r8152 *tp = from_tasklet(tp, t, tx_tl);
drivers/net/usb/rtl8150.c
588
static void rx_fixup(struct tasklet_struct *t)
drivers/net/usb/rtl8150.c
590
struct rtl8150 *dev = from_tasklet(dev, t, tl);
drivers/net/usb/sierra_net.c
572
static void sierra_sync_timer(struct timer_list *t)
drivers/net/usb/sierra_net.c
574
struct sierra_net_data *priv = timer_container_of(priv, t, sync_timer);
drivers/net/usb/usbnet.c
1582
static void usbnet_bh(struct timer_list *t)
drivers/net/usb/usbnet.c
1584
struct usbnet *dev = timer_container_of(dev, t, delay);
drivers/net/vxlan/vxlan_core.c
2845
static void vxlan_cleanup(struct timer_list *t)
drivers/net/vxlan/vxlan_core.c
2847
struct vxlan_dev *vxlan = timer_container_of(vxlan, t, age_timer);
drivers/net/wan/hdlc_cisco.c
245
static void cisco_timer(struct timer_list *t)
drivers/net/wan/hdlc_cisco.c
247
struct cisco_state *st = timer_container_of(st, t, timer);
drivers/net/wan/hdlc_fr.c
582
static void fr_timer(struct timer_list *t)
drivers/net/wan/hdlc_fr.c
584
struct frad_state *st = timer_container_of(st, t, timer);
drivers/net/wan/hdlc_ppp.c
560
static void ppp_timer(struct timer_list *t)
drivers/net/wan/hdlc_ppp.c
562
struct proto *proto = timer_container_of(proto, t, timer);
drivers/net/wan/hdlc_x25.c
39
static void x25_rx_queue_kick(struct tasklet_struct *t)
drivers/net/wan/hdlc_x25.c
41
struct x25_state *x25st = from_tasklet(x25st, t, rx_tasklet);
drivers/net/wireguard/noise.c
595
u8 t[NOISE_TIMESTAMP_LEN];
drivers/net/wireguard/noise.c
628
if (!message_decrypt(t, src->encrypted_timestamp,
drivers/net/wireguard/noise.c
633
replay_attack = memcmp(t, handshake->latest_timestamp,
drivers/net/wireguard/noise.c
645
if (memcmp(t, handshake->latest_timestamp, NOISE_TIMESTAMP_LEN) > 0)
drivers/net/wireguard/noise.c
646
memcpy(handshake->latest_timestamp, t, NOISE_TIMESTAMP_LEN);
drivers/net/wireguard/selftest/allowedips.c
261
struct allowedips t;
drivers/net/wireguard/selftest/allowedips.c
266
wg_allowedips_init(&t);
drivers/net/wireguard/selftest/allowedips.c
290
if (wg_allowedips_insert_v4(&t, (struct in_addr *)ip, cidr,
drivers/net/wireguard/selftest/allowedips.c
316
if (wg_allowedips_insert_v4(&t,
drivers/net/wireguard/selftest/allowedips.c
334
if (wg_allowedips_insert_v6(&t, (struct in6_addr *)ip, cidr,
drivers/net/wireguard/selftest/allowedips.c
360
if (wg_allowedips_insert_v6(&t,
drivers/net/wireguard/selftest/allowedips.c
378
print_tree(t.root4, 32);
drivers/net/wireguard/selftest/allowedips.c
379
print_tree(t.root6, 128);
drivers/net/wireguard/selftest/allowedips.c
385
if (lookup(t.root4, 32, ip) != horrible_allowedips_lookup_v4(&h, (struct in_addr *)ip)) {
drivers/net/wireguard/selftest/allowedips.c
390
if (lookup(t.root6, 128, ip) != horrible_allowedips_lookup_v6(&h, (struct in6_addr *)ip)) {
drivers/net/wireguard/selftest/allowedips.c
398
wg_allowedips_remove_by_peer(&t, peers[j], &mutex);
drivers/net/wireguard/selftest/allowedips.c
403
if (t.root4 || t.root6) {
drivers/net/wireguard/selftest/allowedips.c
413
wg_allowedips_free(&t, &mutex);
drivers/net/wireguard/selftest/allowedips.c
460
wg_allowedips_insert_v##version(&t, ip##version(ipa, ipb, ipc, ipd), \
drivers/net/wireguard/selftest/allowedips.c
464
wg_allowedips_remove_v##version(&t, ip##version(ipa, ipb, ipc, ipd), \
drivers/net/wireguard/selftest/allowedips.c
476
bool _s = lookup(t.root##version, (version) == 4 ? 32 : 128, \
drivers/net/wireguard/selftest/allowedips.c
482
bool _s = lookup(t.root##version, (version) == 4 ? 32 : 128, \
drivers/net/wireguard/selftest/allowedips.c
501
struct allowedips t;
drivers/net/wireguard/selftest/allowedips.c
509
wg_allowedips_init(&t);
drivers/net/wireguard/selftest/allowedips.c
545
print_tree(t.root4, 32);
drivers/net/wireguard/selftest/allowedips.c
546
print_tree(t.root6, 128);
drivers/net/wireguard/selftest/allowedips.c
585
wg_allowedips_remove_by_peer(&t, a, &mutex);
drivers/net/wireguard/selftest/allowedips.c
636
wg_allowedips_free(&t, &mutex);
drivers/net/wireguard/selftest/allowedips.c
637
wg_allowedips_init(&t);
drivers/net/wireguard/selftest/allowedips.c
640
wg_allowedips_remove_by_peer(&t, a, &mutex);
drivers/net/wireguard/selftest/allowedips.c
650
wg_allowedips_insert_v6(&t, &ip, 128, a, &mutex);
drivers/net/wireguard/selftest/allowedips.c
653
wg_allowedips_insert_v6(&t, &ip, 128, a, &mutex);
drivers/net/wireguard/selftest/allowedips.c
656
wg_allowedips_insert_v6(&t, &ip, 128, a, &mutex);
drivers/net/wireguard/selftest/allowedips.c
657
wg_allowedips_free(&t, &mutex);
drivers/net/wireguard/selftest/allowedips.c
659
wg_allowedips_init(&t);
drivers/net/wireguard/selftest/allowedips.c
707
wg_allowedips_free(&t, &mutex);
drivers/net/wireless/ath/ar5523/ar5523.c
903
static void ar5523_tx_wd_timer(struct timer_list *t)
drivers/net/wireless/ath/ar5523/ar5523.c
905
struct ar5523 *ar = timer_container_of(ar, t, tx_wd_timer);
drivers/net/wireless/ath/ath10k/htt_rx.c
261
static void ath10k_htt_rx_ring_refill_retry(struct timer_list *t)
drivers/net/wireless/ath/ath10k/htt_rx.c
263
struct ath10k_htt *htt = timer_container_of(htt, t,
drivers/net/wireless/ath/ath10k/pci.c
593
static void ath10k_pci_ps_timer(struct timer_list *t)
drivers/net/wireless/ath/ath10k/pci.c
595
struct ath10k_pci *ar_pci = timer_container_of(ar_pci, t, ps_timer);
drivers/net/wireless/ath/ath10k/pci.c
846
void ath10k_pci_rx_replenish_retry(struct timer_list *t)
drivers/net/wireless/ath/ath10k/pci.c
848
struct ath10k_pci *ar_pci = timer_container_of(ar_pci, t,
drivers/net/wireless/ath/ath10k/pci.h
239
void ath10k_pci_rx_replenish_retry(struct timer_list *t);
drivers/net/wireless/ath/ath10k/sdio.c
1448
static void ath10k_sdio_sleep_timer_handler(struct timer_list *t)
drivers/net/wireless/ath/ath10k/sdio.c
1450
struct ath10k_sdio *ar_sdio = timer_container_of(ar_sdio, t,
drivers/net/wireless/ath/ath10k/snoc.c
647
static void ath10k_snoc_rx_replenish_retry(struct timer_list *t)
drivers/net/wireless/ath/ath10k/snoc.c
649
struct ath10k_snoc *ar_snoc = timer_container_of(ar_snoc, t,
drivers/net/wireless/ath/ath10k/wmi.c
3584
__le32 t;
drivers/net/wireless/ath/ath10k/wmi.c
3605
t = tim_info->tim_bitmap[i / 4];
drivers/net/wireless/ath/ath10k/wmi.c
3606
v = __le32_to_cpu(t);
drivers/net/wireless/ath/ath11k/ahb.c
467
static void ath11k_ahb_ce_tasklet(struct tasklet_struct *t)
drivers/net/wireless/ath/ath11k/ahb.c
469
struct ath11k_ce_pipe *ce_pipe = from_tasklet(ce_pipe, t, intr_tq);
drivers/net/wireless/ath/ath11k/ce.c
908
void ath11k_ce_rx_replenish_retry(struct timer_list *t)
drivers/net/wireless/ath/ath11k/ce.c
910
struct ath11k_base *ab = timer_container_of(ab, t, rx_replenish_retry);
drivers/net/wireless/ath/ath11k/ce.h
197
void ath11k_ce_rx_replenish_retry(struct timer_list *t);
drivers/net/wireless/ath/ath11k/dp.c
1120
static void ath11k_dp_shadow_timer_handler(struct timer_list *t)
drivers/net/wireless/ath/ath11k/dp.c
1123
t,
drivers/net/wireless/ath/ath11k/dp_rx.c
309
static void ath11k_dp_service_mon_ring(struct timer_list *t)
drivers/net/wireless/ath/ath11k/dp_rx.c
311
struct ath11k_base *ab = timer_container_of(ab, t, mon_reap_timer);
drivers/net/wireless/ath/ath11k/pcic.c
406
static void ath11k_pcic_ce_tasklet(struct tasklet_struct *t)
drivers/net/wireless/ath/ath11k/pcic.c
408
struct ath11k_ce_pipe *ce_pipe = from_tasklet(ce_pipe, t, intr_tq);
drivers/net/wireless/ath/ath12k/ce.c
567
void ath12k_ce_rx_replenish_retry(struct timer_list *t)
drivers/net/wireless/ath/ath12k/ce.c
569
struct ath12k_base *ab = timer_container_of(ab, t, rx_replenish_retry);
drivers/net/wireless/ath/ath12k/ce.h
178
void ath12k_ce_rx_replenish_retry(struct timer_list *t);
drivers/net/wireless/ath/ath5k/base.c
1584
ath5k_tasklet_rx(struct tasklet_struct *t)
drivers/net/wireless/ath/ath5k/base.c
1589
struct ath5k_hw *ah = from_tasklet(ah, t, rxtq);
drivers/net/wireless/ath/ath5k/base.c
1832
ath5k_tasklet_tx(struct tasklet_struct *t)
drivers/net/wireless/ath/ath5k/base.c
1835
struct ath5k_hw *ah = from_tasklet(ah, t, txtq);
drivers/net/wireless/ath/ath5k/base.c
2224
static void ath5k_tasklet_beacon(struct tasklet_struct *t)
drivers/net/wireless/ath/ath5k/base.c
2226
struct ath5k_hw *ah = from_tasklet(ah, t, beacontq);
drivers/net/wireless/ath/ath5k/base.c
2495
ath5k_tasklet_ani(struct tasklet_struct *t)
drivers/net/wireless/ath/ath5k/base.c
2497
struct ath5k_hw *ah = from_tasklet(ah, t, ani_tasklet);
drivers/net/wireless/ath/ath5k/rfkill.c
76
ath5k_tasklet_rfkill_toggle(struct tasklet_struct *t)
drivers/net/wireless/ath/ath5k/rfkill.c
78
struct ath5k_hw *ah = from_tasklet(ah, t, rf_kill.toggleq);
drivers/net/wireless/ath/ath6kl/core.h
896
void disconnect_timer_handler(struct timer_list *t);
drivers/net/wireless/ath/ath6kl/main.c
504
void disconnect_timer_handler(struct timer_list *t)
drivers/net/wireless/ath/ath6kl/main.c
506
struct ath6kl_vif *vif = timer_container_of(vif, t, disconnect_timer);
drivers/net/wireless/ath/ath6kl/recovery.c
63
static void ath6kl_recovery_hb_timer(struct timer_list *t)
drivers/net/wireless/ath/ath6kl/recovery.c
65
struct ath6kl *ar = timer_container_of(ar, t, fw_recovery.hb_timer);
drivers/net/wireless/ath/ath6kl/txrx.c
1623
static void aggr_timeout(struct timer_list *t)
drivers/net/wireless/ath/ath6kl/txrx.c
1626
struct aggr_info_conn *aggr_conn = timer_container_of(aggr_conn, t,
drivers/net/wireless/ath/ath6kl/wmi.c
1079
void ath6kl_wmi_sscan_timer(struct timer_list *t)
drivers/net/wireless/ath/ath6kl/wmi.c
1081
struct ath6kl_vif *vif = timer_container_of(vif, t, sched_scan_timer);
drivers/net/wireless/ath/ath6kl/wmi.h
2722
void ath6kl_wmi_sscan_timer(struct timer_list *t);
drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
3841
int f[3], t[3];
drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
3849
t[0] = eep->base_ext2.xatten1DBLow[chain];
drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
3851
t[1] = eep->modalHeader5G.xatten1DB[chain];
drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
3853
t[2] = eep->base_ext2.xatten1DBHigh[chain];
drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
3856
f, t, 3);
drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
3869
int f[3], t[3];
drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
3877
t[0] = eep->base_ext2.xatten1MarginLow[chain];
drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
3879
t[1] = eep->modalHeader5G.xatten1Margin[chain];
drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
3881
t[2] = eep->base_ext2.xatten1MarginHigh[chain];
drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
3884
f, t, 3);
drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
4070
s32 t[3], f[3] = {5180, 5500, 5785};
drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
4079
t[0] = eep->base_ext1.quick_drop_low;
drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
4080
t[1] = eep->modalHeader5G.quick_drop;
drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
4081
t[2] = eep->base_ext1.quick_drop_high;
drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
4082
quick_drop = ar9003_hw_power_interpolate(freq, f, t, 3);
drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
4810
int f[8], t[8], t1[3], t2[3], i;
drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
4845
t[0] = eep->base_ext1.tempslopextension[2];
drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
4850
t[1] = eep->modalHeader5G.tempSlope;
drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
4855
t[2] = eep->base_ext1.tempslopextension[5];
drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
4861
f, t, 3);
drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
4872
t[i] = eep->base_ext1.tempslopextension[i];
drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
4876
f, t, 8);
drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
4878
t[0] = eep->base_ext2.tempSlopeLow;
drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
4880
t[1] = eep->modalHeader5G.tempSlope;
drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
4882
t[2] = eep->base_ext2.tempSlopeHigh;
drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
4885
f, t, 3);
drivers/net/wireless/ath/ath9k/ath9k.h
1124
void ath9k_tasklet(struct tasklet_struct *t);
drivers/net/wireless/ath/ath9k/ath9k.h
717
void ath9k_beacon_tasklet(struct tasklet_struct *t);
drivers/net/wireless/ath/ath9k/ath9k.h
747
void ath_ani_calibrate(struct timer_list *t);
drivers/net/wireless/ath/ath9k/ath9k.h
754
void ath_ps_full_sleep(struct timer_list *t);
drivers/net/wireless/ath/ath9k/beacon.c
388
void ath9k_beacon_tasklet(struct tasklet_struct *t)
drivers/net/wireless/ath/ath9k/beacon.c
390
struct ath_softc *sc = from_tasklet(sc, t, bcon_tasklet);
drivers/net/wireless/ath/ath9k/channel.c
1041
static void ath_chanctx_timer(struct timer_list *t)
drivers/net/wireless/ath/ath9k/channel.c
1043
struct ath_softc *sc = timer_container_of(sc, t, sched.timer);
drivers/net/wireless/ath/ath9k/channel.c
1052
static void ath_offchannel_timer(struct timer_list *t)
drivers/net/wireless/ath/ath9k/channel.c
1054
struct ath_softc *sc = timer_container_of(sc, t, offchannel.timer);
drivers/net/wireless/ath/ath9k/gpio.c
194
static void ath_btcoex_period_timer(struct timer_list *t)
drivers/net/wireless/ath/ath9k/gpio.c
196
struct ath_softc *sc = timer_container_of(sc, t, btcoex.period_timer);
drivers/net/wireless/ath/ath9k/gpio.c
255
static void ath_btcoex_no_stomp_timer(struct timer_list *t)
drivers/net/wireless/ath/ath9k/gpio.c
257
struct ath_softc *sc = timer_container_of(sc, t,
drivers/net/wireless/ath/ath9k/htc.h
592
void ath9k_tx_failed_tasklet(struct tasklet_struct *t);
drivers/net/wireless/ath/ath9k/htc.h
593
void ath9k_htc_tx_cleanup_timer(struct timer_list *t);
drivers/net/wireless/ath/ath9k/htc.h
599
void ath9k_rx_tasklet(struct tasklet_struct *t);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
1081
void ath9k_rx_tasklet(struct tasklet_struct *t)
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
1083
struct ath9k_htc_priv *priv = from_tasklet(priv, t, rx_tasklet);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
578
void ath9k_tx_failed_tasklet(struct tasklet_struct *t)
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
580
struct ath9k_htc_priv *priv = from_tasklet(priv, t, tx_failed_tasklet);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
761
void ath9k_htc_tx_cleanup_timer(struct timer_list *t)
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
763
struct ath9k_htc_priv *priv = timer_container_of(priv, t,
drivers/net/wireless/ath/ath9k/link.c
333
void ath_ani_calibrate(struct timer_list *t)
drivers/net/wireless/ath/ath9k/link.c
335
struct ath_common *common = timer_container_of(common, t, ani.timer);
drivers/net/wireless/ath/ath9k/main.c
101
struct ath_softc *sc = timer_container_of(sc, t, sleep_timer);
drivers/net/wireless/ath/ath9k/main.c
377
void ath9k_tasklet(struct tasklet_struct *t)
drivers/net/wireless/ath/ath9k/main.c
379
struct ath_softc *sc = from_tasklet(sc, t, intr_tq);
drivers/net/wireless/ath/ath9k/main.c
99
void ath_ps_full_sleep(struct timer_list *t)
drivers/net/wireless/ath/ath9k/wmi.c
138
void ath9k_wmi_event_tasklet(struct tasklet_struct *t)
drivers/net/wireless/ath/ath9k/wmi.c
140
struct wmi *wmi = from_tasklet(wmi, t, wmi_event_tasklet);
drivers/net/wireless/ath/ath9k/wmi.h
188
void ath9k_wmi_event_tasklet(struct tasklet_struct *t);
drivers/net/wireless/ath/ath9k/xmit.c
32
#define TIME_SYMBOLS(t) ((t) >> 2)
drivers/net/wireless/ath/ath9k/xmit.c
33
#define TIME_SYMBOLS_HALFGI(t) (((t) * 5 - 4) / 18)
drivers/net/wireless/ath/carl9170/tx.c
663
unsigned int r, t, q;
drivers/net/wireless/ath/carl9170/tx.c
683
t = (info & CARL9170_TX_STATUS_TRIES) >> CARL9170_TX_STATUS_TRIES_S;
drivers/net/wireless/ath/carl9170/tx.c
685
carl9170_tx_fill_rateinfo(ar, r, t, txinfo);
drivers/net/wireless/ath/carl9170/usb.c
380
static void carl9170_usb_tasklet(struct tasklet_struct *t)
drivers/net/wireless/ath/carl9170/usb.c
382
struct ar9170 *ar = from_tasklet(ar, t, usb_tasklet);
drivers/net/wireless/ath/carl9170/wlan.h
401
static inline u8 ar9170_get_decrypt_type(struct ar9170_rx_macstatus *t)
drivers/net/wireless/ath/carl9170/wlan.h
403
return (t->SAidx & 0xc0) >> 4 |
drivers/net/wireless/ath/carl9170/wlan.h
404
(t->DAidx & 0xc0) >> 6;
drivers/net/wireless/ath/wcn36xx/dxe.c
374
static void wcn36xx_dxe_tx_timer(struct timer_list *t)
drivers/net/wireless/ath/wcn36xx/dxe.c
376
struct wcn36xx *wcn = timer_container_of(wcn, t, tx_ack_timer);
drivers/net/wireless/ath/wil6210/cfg80211.c
2367
struct wil_probe_client_req *req, *t;
drivers/net/wireless/ath/wil6210/cfg80211.c
2374
list_for_each_entry_safe(req, t, &vif->probe_client_pending, list) {
drivers/net/wireless/ath/wil6210/debugfs.c
1332
static void print_temp(struct seq_file *s, const char *prefix, s32 t)
drivers/net/wireless/ath/wil6210/debugfs.c
1334
switch (t) {
drivers/net/wireless/ath/wil6210/debugfs.c
1340
seq_printf(s, "%s %s%d.%03d\n", prefix, (t < 0 ? "-" : ""),
drivers/net/wireless/ath/wil6210/debugfs.c
1341
abs(t / 1000), abs(t % 1000));
drivers/net/wireless/ath/wil6210/netdev.c
228
static void wil_connect_timer_fn(struct timer_list *t)
drivers/net/wireless/ath/wil6210/netdev.c
230
struct wil6210_vif *vif = timer_container_of(vif, t, connect_timer);
drivers/net/wireless/ath/wil6210/netdev.c
244
static void wil_scan_timer_fn(struct timer_list *t)
drivers/net/wireless/ath/wil6210/netdev.c
246
struct wil6210_vif *vif = timer_container_of(vif, t, scan_timer);
drivers/net/wireless/ath/wil6210/netdev.c
254
static void wil_p2p_discovery_timer_fn(struct timer_list *t)
drivers/net/wireless/ath/wil6210/netdev.c
256
struct wil6210_vif *vif = timer_container_of(vif, t,
drivers/net/wireless/ath/wil6210/wmi.c
3070
struct pending_wmi_event *evt, *t;
drivers/net/wireless/ath/wil6210/wmi.c
3076
list_for_each_entry_safe(evt, t, &wil->pending_wmi_ev, list) {
drivers/net/wireless/atmel/at76c50x-usb.c
1541
static void at76_rx_tasklet(struct tasklet_struct *t)
drivers/net/wireless/atmel/at76c50x-usb.c
1543
struct at76_priv *priv = from_tasklet(priv, t, rx_tasklet);
drivers/net/wireless/broadcom/b43legacy/main.c
1278
static void b43legacy_interrupt_tasklet(struct tasklet_struct *t)
drivers/net/wireless/broadcom/b43legacy/main.c
1280
struct b43legacy_wldev *dev = from_tasklet(dev, t, isr_tasklet);
drivers/net/wireless/broadcom/b43legacy/pio.c
267
static void tx_tasklet(struct tasklet_struct *t)
drivers/net/wireless/broadcom/b43legacy/pio.c
269
struct b43legacy_pioqueue *queue = from_tasklet(queue, t, txtask);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/btcoex.c
273
static void brcmf_btcoex_timerfunc(struct timer_list *t)
drivers/net/wireless/broadcom/brcm80211/brcmfmac/btcoex.c
275
struct brcmf_btcoex_info *bt_local = timer_container_of(bt_local, t,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
3574
static void brcmf_escan_timeout(struct timer_list *t)
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
3577
timer_container_of(cfg, t, escan_timeout);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/pcie.c
2332
brcmf_pcie_fwcon(struct timer_list *t)
drivers/net/wireless/broadcom/brcm80211/brcmfmac/pcie.c
2334
struct brcmf_pciedev_info *devinfo = timer_container_of(devinfo, t,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
4124
brcmf_sdio_watchdog(struct timer_list *t)
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
4126
struct brcmf_sdio *bus = timer_container_of(bus, t, timer);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/brcms_trace_brcmsmac.h
31
TP_PROTO(struct brcms_timer *t),
drivers/net/wireless/broadcom/brcm80211/brcmsmac/brcms_trace_brcmsmac.h
36
TP_ARGS(t),
drivers/net/wireless/broadcom/brcm80211/brcmsmac/brcms_trace_brcmsmac.h
48
__entry->ms = t->ms;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/brcms_trace_brcmsmac.h
49
__entry->set = t->set;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/brcms_trace_brcmsmac.h
50
__entry->periodic = t->periodic;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c
338
static uint ntxdactive(struct dma_info *di, uint h, uint t)
drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c
340
return txd(di, t-h);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c
343
static uint nrxdactive(struct dma_info *di, uint h, uint t)
drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c
345
return rxd(di, t-h);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1467
struct brcms_timer *t = container_of(work, struct brcms_timer,
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1470
spin_lock_bh(&t->wl->lock);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1472
if (t->set) {
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1473
if (t->periodic) {
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1474
atomic_inc(&t->wl->callbacks);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1475
ieee80211_queue_delayed_work(t->wl->pub->ieee_hw,
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1476
&t->dly_wrk,
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1477
msecs_to_jiffies(t->ms));
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1479
t->set = false;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1482
t->fn(t->arg);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1485
atomic_dec(&t->wl->callbacks);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1487
spin_unlock_bh(&t->wl->lock);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1500
struct brcms_timer *t;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1502
t = kzalloc_obj(*t, GFP_ATOMIC);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1503
if (!t)
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1506
INIT_DELAYED_WORK(&t->dly_wrk, _brcms_timer);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1507
t->wl = wl;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1508
t->fn = fn;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1509
t->arg = arg;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1510
t->next = wl->timers;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1511
wl->timers = t;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1514
t->name = kstrdup(name, GFP_ATOMIC);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1517
return t;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1526
void brcms_add_timer(struct brcms_timer *t, uint ms, int periodic)
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1528
struct ieee80211_hw *hw = t->wl->pub->ieee_hw;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1531
if (t->set)
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1532
brcms_dbg_info(t->wl->wlc->hw->d11core,
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1534
__func__, t->name, periodic);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1536
t->ms = ms;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1537
t->periodic = (bool) periodic;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1538
if (!t->set) {
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1539
t->set = true;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1540
atomic_inc(&t->wl->callbacks);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1543
ieee80211_queue_delayed_work(hw, &t->dly_wrk, msecs_to_jiffies(ms));
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1551
bool brcms_del_timer(struct brcms_timer *t)
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1553
if (t->set) {
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1554
t->set = false;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1555
if (!cancel_delayed_work(&t->dly_wrk))
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1558
atomic_dec(&t->wl->callbacks);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1567
void brcms_free_timer(struct brcms_timer *t)
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1569
struct brcms_info *wl = t->wl;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1573
brcms_del_timer(t);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1575
if (wl->timers == t) {
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1578
kfree(t->name);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1580
kfree(t);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1587
if (tmp->next == t) {
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1588
tmp->next = t->next;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1590
kfree(t->name);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
1592
kfree(t);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
288
struct brcms_timer *t, *next;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
318
for (t = wl->timers; t; t = next) {
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
319
next = t->next;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
321
kfree(t->name);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
323
kfree(t);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
989
void brcms_dpc(struct tasklet_struct *t)
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
993
wl = from_tasklet(wl, t, tasklet);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.h
109
void brcms_dpc(struct tasklet_struct *t);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.h
110
void brcms_timer(struct brcms_timer *t);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c
3337
u16 num_samps, t, k;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c
3371
for (t = 0; t < num_samps; t++) {
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c
3379
data_buf[t] = (i_samp << 10) | q_samp;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c
23007
u16 t;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c
23017
for (t = 0; t < num_samps; t++)
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c
23018
data_buf[t] = ((((unsigned int)tone_buf[t].i) & 0x3ff) << 10) |
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c
23019
(((unsigned int)tone_buf[t].q) & 0x3ff);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c
23034
u16 num_samps, t, spur;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c
23060
for (t = 0; t < num_samps; t++) {
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c
23062
tone_buf[t] = cordic_calc_iq(theta);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c
23066
tone_buf[t].q = (s32)CORDIC_FLOAT(tone_buf[t].q * max_val);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c
23067
tone_buf[t].i = (s32)CORDIC_FLOAT(tone_buf[t].i * max_val);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy_shim.c
67
void wlapi_free_timer(struct wlapi_timer *t)
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy_shim.c
69
brcms_free_timer((struct brcms_timer *)t);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy_shim.c
73
wlapi_add_timer(struct wlapi_timer *t, uint ms, int periodic)
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy_shim.c
75
brcms_add_timer((struct brcms_timer *)t, ms, periodic);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy_shim.c
78
bool wlapi_del_timer(struct wlapi_timer *t)
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy_shim.c
80
return brcms_del_timer((struct brcms_timer *)t);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy_shim.h
136
void wlapi_free_timer(struct wlapi_timer *t);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy_shim.h
137
void wlapi_add_timer(struct wlapi_timer *t, uint ms, int periodic);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy_shim.h
138
bool wlapi_del_timer(struct wlapi_timer *t);
drivers/net/wireless/intel/ipw2x00/ipw2100.c
3189
static void ipw2100_irq_tasklet(struct tasklet_struct *t)
drivers/net/wireless/intel/ipw2x00/ipw2100.c
3191
struct ipw2100_priv *priv = from_tasklet(priv, t, irq_tasklet);
drivers/net/wireless/intel/ipw2x00/ipw2100.c
5983
static void ipw2100_irq_tasklet(struct tasklet_struct *t);
drivers/net/wireless/intel/ipw2x00/ipw2200.c
1916
static void ipw_irq_tasklet(struct tasklet_struct *t)
drivers/net/wireless/intel/ipw2x00/ipw2200.c
1918
struct ipw_priv *priv = from_tasklet(priv, t, irq_tasklet);
drivers/net/wireless/intel/ipw2x00/libipw_crypto.c
111
static void libipw_crypt_deinit_handler(struct timer_list *t)
drivers/net/wireless/intel/ipw2x00/libipw_crypto.c
113
struct libipw_crypt_info *info = timer_container_of(info, t,
drivers/net/wireless/intel/ipw2x00/libipw_crypto.c
39
static void libipw_crypt_deinit_handler(struct timer_list *t);
drivers/net/wireless/intel/ipw2x00/libipw_crypto_tkip.c
202
u16 t = Sbox[Hi8(v)];
drivers/net/wireless/intel/ipw2x00/libipw_crypto_tkip.c
203
return Sbox[Lo8(v)] ^ ((t << 8) | (t >> 8));
drivers/net/wireless/intel/iwlegacy/3945-mac.c
1331
il3945_irq_tasklet(struct tasklet_struct *t)
drivers/net/wireless/intel/iwlegacy/3945-mac.c
1333
struct il_priv *il = from_tasklet(il, t, irq_tasklet);
drivers/net/wireless/intel/iwlegacy/3945-rs.c
169
il3945_bg_rate_scale_flush(struct timer_list *t)
drivers/net/wireless/intel/iwlegacy/3945-rs.c
171
struct il3945_rs_sta *rs_sta = timer_container_of(rs_sta, t,
drivers/net/wireless/intel/iwlegacy/4965-mac.c
4055
il4965_bg_stats_periodic(struct timer_list *t)
drivers/net/wireless/intel/iwlegacy/4965-mac.c
4057
struct il_priv *il = timer_container_of(il, t, stats_periodic);
drivers/net/wireless/intel/iwlegacy/4965-mac.c
4338
il4965_irq_tasklet(struct tasklet_struct *t)
drivers/net/wireless/intel/iwlegacy/4965-mac.c
4340
struct il_priv *il = from_tasklet(il, t, irq_tasklet);
drivers/net/wireless/intel/iwlegacy/4965.h
252
#define IL_TX_POWER_TEMPERATURE_OUT_OF_RANGE(t) \
drivers/net/wireless/intel/iwlegacy/4965.h
253
((t) < IL_TX_POWER_TEMPERATURE_MIN || \
drivers/net/wireless/intel/iwlegacy/4965.h
254
(t) > IL_TX_POWER_TEMPERATURE_MAX)
drivers/net/wireless/intel/iwlegacy/common.c
111
int t = 0;
drivers/net/wireless/intel/iwlegacy/common.c
115
return t;
drivers/net/wireless/intel/iwlegacy/common.c
117
t += interval;
drivers/net/wireless/intel/iwlegacy/common.c
118
} while (t < timeout);
drivers/net/wireless/intel/iwlegacy/common.c
30
int t = 0;
drivers/net/wireless/intel/iwlegacy/common.c
34
return t;
drivers/net/wireless/intel/iwlegacy/common.c
36
t += interval;
drivers/net/wireless/intel/iwlegacy/common.c
37
} while (t < timeout);
drivers/net/wireless/intel/iwlegacy/common.c
4799
il_bg_watchdog(struct timer_list *t)
drivers/net/wireless/intel/iwlegacy/common.c
4801
struct il_priv *il = timer_container_of(il, t, watchdog);
drivers/net/wireless/intel/iwlegacy/common.h
1829
void il_bg_watchdog(struct timer_list *t);
drivers/net/wireless/intel/iwlwifi/dvm/main.c
382
static void iwl_bg_statistics_periodic(struct timer_list *t)
drivers/net/wireless/intel/iwlwifi/dvm/main.c
384
struct iwl_priv *priv = timer_container_of(priv, t,
drivers/net/wireless/intel/iwlwifi/dvm/main.c
539
static void iwl_bg_ucode_trace(struct timer_list *t)
drivers/net/wireless/intel/iwlwifi/dvm/main.c
541
struct iwl_priv *priv = timer_container_of(priv, t, ucode_trace);
drivers/net/wireless/intel/iwlwifi/dvm/tt.c
138
static void iwl_tt_check_exit_ct_kill(struct timer_list *t)
drivers/net/wireless/intel/iwlwifi/dvm/tt.c
140
struct iwl_priv *priv = timer_container_of(priv, t,
drivers/net/wireless/intel/iwlwifi/dvm/tt.c
188
static void iwl_tt_ready_for_ct_kill(struct timer_list *t)
drivers/net/wireless/intel/iwlwifi/dvm/tt.c
190
struct iwl_priv *priv = timer_container_of(priv, t,
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
949
static void iwl_dbg_tlv_periodic_trig_handler(struct timer_list *t)
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
952
timer_container_of(timer_node, t, timer);
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
966
mod_timer(t, jiffies + msecs_to_jiffies(collect_interval));
drivers/net/wireless/intel/iwlwifi/iwl-io.c
103
return t;
drivers/net/wireless/intel/iwlwifi/iwl-io.c
105
t += IWL_POLL_INTERVAL;
drivers/net/wireless/intel/iwlwifi/iwl-io.c
106
} while (t < timeout);
drivers/net/wireless/intel/iwlwifi/iwl-io.c
159
int t = 0;
drivers/net/wireless/intel/iwlwifi/iwl-io.c
165
t += IWL_POLL_INTERVAL;
drivers/net/wireless/intel/iwlwifi/iwl-io.c
166
} while (t < timeout);
drivers/net/wireless/intel/iwlwifi/iwl-io.c
53
int t = 0;
drivers/net/wireless/intel/iwlwifi/iwl-io.c
59
t += IWL_POLL_INTERVAL;
drivers/net/wireless/intel/iwlwifi/iwl-io.c
60
} while (t < timeout);
drivers/net/wireless/intel/iwlwifi/iwl-io.c
99
int t = 0;
drivers/net/wireless/intel/iwlwifi/mld/agg.c
322
static void iwl_mld_rx_agg_session_expired(struct timer_list *t)
drivers/net/wireless/intel/iwlwifi/mld/agg.c
325
timer_container_of(data, t, session_timer);
drivers/net/wireless/intel/iwlwifi/mld/tests/hcmd.c
32
static struct iwl_trans t = {
drivers/net/wireless/intel/iwlwifi/mld/tests/hcmd.c
36
t.conf.command_groups_size = global_iwl_mld_goups_size;
drivers/net/wireless/intel/iwlwifi/mld/tests/hcmd.c
44
name = iwl_get_cmd_string(&t, rxh->cmd_id);
drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c
51
struct iwl_mvm_loc_entry *e, *t;
drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c
58
list_for_each_entry_safe(e, t, &mvm->ftm_initiator.loc_list, list) {
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
253
static void iwl_mvm_rx_agg_session_expired(struct timer_list *t)
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
256
timer_container_of(data, t, session_timer);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
564
int t = 0;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
590
t += 200;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
591
} while (t < 150000);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
703
static void iwl_txq_stuck_timer(struct timer_list *t)
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
705
struct iwl_txq *txq = timer_container_of(txq, t, stuck_timer);
drivers/net/wireless/intel/iwlwifi/tests/devinfo.c
206
const struct pci_device_id *s, *t;
drivers/net/wireless/intel/iwlwifi/tests/devinfo.c
215
t = pci_match_id(iwl_hw_card_ids, dev);
drivers/net/wireless/intel/iwlwifi/tests/devinfo.c
216
KUNIT_EXPECT_PTR_EQ(test, t, s);
drivers/net/wireless/intersil/p54/p54pci.c
281
static void p54p_tasklet(struct tasklet_struct *t)
drivers/net/wireless/intersil/p54/p54pci.c
283
struct p54p_priv *priv = from_tasklet(priv, t, tasklet);
drivers/net/wireless/intersil/p54/p54spi.c
49
struct spi_transfer t[2];
drivers/net/wireless/intersil/p54/p54spi.c
57
memset(t, 0, sizeof(t));
drivers/net/wireless/intersil/p54/p54spi.c
59
t[0].tx_buf = &addr;
drivers/net/wireless/intersil/p54/p54spi.c
60
t[0].len = sizeof(addr);
drivers/net/wireless/intersil/p54/p54spi.c
61
spi_message_add_tail(&t[0], &m);
drivers/net/wireless/intersil/p54/p54spi.c
63
t[1].rx_buf = buf;
drivers/net/wireless/intersil/p54/p54spi.c
64
t[1].len = len;
drivers/net/wireless/intersil/p54/p54spi.c
65
spi_message_add_tail(&t[1], &m);
drivers/net/wireless/intersil/p54/p54spi.c
74
struct spi_transfer t[3];
drivers/net/wireless/intersil/p54/p54spi.c
82
memset(t, 0, sizeof(t));
drivers/net/wireless/intersil/p54/p54spi.c
84
t[0].tx_buf = &addr;
drivers/net/wireless/intersil/p54/p54spi.c
85
t[0].len = sizeof(addr);
drivers/net/wireless/intersil/p54/p54spi.c
86
spi_message_add_tail(&t[0], &m);
drivers/net/wireless/intersil/p54/p54spi.c
88
t[1].tx_buf = buf;
drivers/net/wireless/intersil/p54/p54spi.c
89
t[1].len = len & ~1;
drivers/net/wireless/intersil/p54/p54spi.c
90
spi_message_add_tail(&t[1], &m);
drivers/net/wireless/intersil/p54/p54spi.c
96
t[2].tx_buf = &last_word;
drivers/net/wireless/intersil/p54/p54spi.c
97
t[2].len = sizeof(last_word);
drivers/net/wireless/intersil/p54/p54spi.c
98
spi_message_add_tail(&t[2], &m);
drivers/net/wireless/marvell/libertas/cmd.c
637
struct ieee80211_country_ie_triplet *t;
drivers/net/wireless/marvell/libertas/cmd.c
702
t = &domain->triplet[num_triplet];
drivers/net/wireless/marvell/libertas/cmd.c
703
t->chans.first_channel = first_channel;
drivers/net/wireless/marvell/libertas/cmd.c
704
t->chans.num_channels = num_parsed_chan;
drivers/net/wireless/marvell/libertas/cmd.c
705
t->chans.max_power = max_pwr;
drivers/net/wireless/marvell/libertas/cmd.c
715
t = &domain->triplet[num_triplet];
drivers/net/wireless/marvell/libertas/cmd.c
716
t->chans.first_channel = first_channel;
drivers/net/wireless/marvell/libertas/cmd.c
717
t->chans.num_channels = num_parsed_chan;
drivers/net/wireless/marvell/libertas/cmd.c
718
t->chans.max_power = max_pwr;
drivers/net/wireless/marvell/libertas/if_usb.c
166
static void if_usb_fw_timeo(struct timer_list *t)
drivers/net/wireless/marvell/libertas/if_usb.c
168
struct if_usb_card *cardp = timer_container_of(cardp, t, fw_timeout);
drivers/net/wireless/marvell/libertas/main.c
694
static void lbs_cmd_timeout_handler(struct timer_list *t)
drivers/net/wireless/marvell/libertas/main.c
696
struct lbs_private *priv = timer_container_of(priv, t, command_timer);
drivers/net/wireless/marvell/libertas/main.c
728
static void lbs_tx_lockup_handler(struct timer_list *t)
drivers/net/wireless/marvell/libertas/main.c
730
struct lbs_private *priv = timer_container_of(priv, t,
drivers/net/wireless/marvell/libertas_tf/if_usb.c
114
static void if_usb_fw_timeo(struct timer_list *t)
drivers/net/wireless/marvell/libertas_tf/if_usb.c
116
struct if_usb_card *cardp = timer_container_of(cardp, t, fw_timeout);
drivers/net/wireless/marvell/libertas_tf/main.c
127
static void command_timer_fn(struct timer_list *t)
drivers/net/wireless/marvell/libertas_tf/main.c
129
struct lbtf_private *priv = timer_container_of(priv, t, command_timer);
drivers/net/wireless/marvell/mwifiex/11n_rxreorder.c
300
mwifiex_flush_data(struct timer_list *t)
drivers/net/wireless/marvell/mwifiex/11n_rxreorder.c
303
timer_container_of(ctx, t, timer);
drivers/net/wireless/marvell/mwifiex/cfg80211.c
559
struct ieee80211_country_ie_triplet *t;
drivers/net/wireless/marvell/mwifiex/cfg80211.c
603
t = &domain_info->triplet[no_of_triplet];
drivers/net/wireless/marvell/mwifiex/cfg80211.c
604
t->chans.first_channel = first_chan;
drivers/net/wireless/marvell/mwifiex/cfg80211.c
605
t->chans.num_channels = no_of_parsed_chan;
drivers/net/wireless/marvell/mwifiex/cfg80211.c
606
t->chans.max_power = max_pwr;
drivers/net/wireless/marvell/mwifiex/cfg80211.c
616
t = &domain_info->triplet[no_of_triplet];
drivers/net/wireless/marvell/mwifiex/cfg80211.c
617
t->chans.first_channel = first_chan;
drivers/net/wireless/marvell/mwifiex/cfg80211.c
618
t->chans.num_channels = no_of_parsed_chan;
drivers/net/wireless/marvell/mwifiex/cfg80211.c
619
t->chans.max_power = max_pwr;
drivers/net/wireless/marvell/mwifiex/cmdevt.c
935
mwifiex_cmd_timeout_func(struct timer_list *t)
drivers/net/wireless/marvell/mwifiex/cmdevt.c
937
struct mwifiex_adapter *adapter = timer_container_of(adapter, t,
drivers/net/wireless/marvell/mwifiex/init.c
42
static void wakeup_timer_fn(struct timer_list *t)
drivers/net/wireless/marvell/mwifiex/init.c
44
struct mwifiex_adapter *adapter = timer_container_of(adapter, t,
drivers/net/wireless/marvell/mwifiex/main.h
1082
void mwifiex_cmd_timeout_func(struct timer_list *t);
drivers/net/wireless/marvell/mwifiex/main.h
1625
void mwifiex_check_auto_tdls(struct timer_list *t);
drivers/net/wireless/marvell/mwifiex/tdls.c
1416
void mwifiex_check_auto_tdls(struct timer_list *t)
drivers/net/wireless/marvell/mwifiex/tdls.c
1418
struct mwifiex_private *priv = timer_container_of(priv, t,
drivers/net/wireless/marvell/mwifiex/usb.c
1123
static void mwifiex_usb_tx_aggr_tmo(struct timer_list *t)
drivers/net/wireless/marvell/mwifiex/usb.c
1128
timer_container_of(timer_context, t, hold_timer);
drivers/net/wireless/marvell/mwl8k.c
4711
static void mwl8k_tx_poll(struct tasklet_struct *t)
drivers/net/wireless/marvell/mwl8k.c
4713
struct mwl8k_priv *priv = from_tasklet(priv, t, poll_tx_task);
drivers/net/wireless/marvell/mwl8k.c
4740
static void mwl8k_rx_poll(struct tasklet_struct *t)
drivers/net/wireless/marvell/mwl8k.c
4742
struct mwl8k_priv *priv = from_tasklet(priv, t, poll_rx_task);
drivers/net/wireless/mediatek/mt76/dma.c
105
mt76_put_txwi(struct mt76_dev *dev, struct mt76_txwi_cache *t)
drivers/net/wireless/mediatek/mt76/dma.c
107
if (!t)
drivers/net/wireless/mediatek/mt76/dma.c
111
list_add(&t->list, &dev->txwi_cache);
drivers/net/wireless/mediatek/mt76/dma.c
117
mt76_put_rxwi(struct mt76_dev *dev, struct mt76_txwi_cache *t)
drivers/net/wireless/mediatek/mt76/dma.c
119
if (!t)
drivers/net/wireless/mediatek/mt76/dma.c
123
list_add(&t->list, &dev->rxwi_cache);
drivers/net/wireless/mediatek/mt76/dma.c
13
struct mt76_txwi_cache *t;
drivers/net/wireless/mediatek/mt76/dma.c
131
struct mt76_txwi_cache *t;
drivers/net/wireless/mediatek/mt76/dma.c
134
while ((t = __mt76_get_txwi(dev)) != NULL) {
drivers/net/wireless/mediatek/mt76/dma.c
135
dma_unmap_single(dev->dma_dev, t->dma_addr, dev->drv->txwi_size,
drivers/net/wireless/mediatek/mt76/dma.c
137
kfree(mt76_get_txwi_ptr(dev, t));
drivers/net/wireless/mediatek/mt76/dma.c
145
struct mt76_txwi_cache *t;
drivers/net/wireless/mediatek/mt76/dma.c
148
while ((t = __mt76_get_rxwi(dev)) != NULL) {
drivers/net/wireless/mediatek/mt76/dma.c
149
if (t->ptr)
drivers/net/wireless/mediatek/mt76/dma.c
150
mt76_put_page_pool_buf(t->ptr, false);
drivers/net/wireless/mediatek/mt76/dma.c
151
kfree(t);
drivers/net/wireless/mediatek/mt76/dma.c
18
size = L1_CACHE_ALIGN(dev->drv->txwi_size + sizeof(*t));
drivers/net/wireless/mediatek/mt76/dma.c
30
t = (struct mt76_txwi_cache *)(txwi + dev->drv->txwi_size);
drivers/net/wireless/mediatek/mt76/dma.c
31
t->dma_addr = addr;
drivers/net/wireless/mediatek/mt76/dma.c
33
return t;
drivers/net/wireless/mediatek/mt76/dma.c
39
struct mt76_txwi_cache *t;
drivers/net/wireless/mediatek/mt76/dma.c
41
t = kzalloc(L1_CACHE_ALIGN(sizeof(*t)), GFP_ATOMIC);
drivers/net/wireless/mediatek/mt76/dma.c
42
if (!t)
drivers/net/wireless/mediatek/mt76/dma.c
45
t->ptr = NULL;
drivers/net/wireless/mediatek/mt76/dma.c
455
struct mt76_txwi_cache *t;
drivers/net/wireless/mediatek/mt76/dma.c
46
return t;
drivers/net/wireless/mediatek/mt76/dma.c
461
t = mt76_rx_token_release(dev, rx_token_id);
drivers/net/wireless/mediatek/mt76/dma.c
462
if (!t)
drivers/net/wireless/mediatek/mt76/dma.c
465
q = &dev->q_rx[t->qid];
drivers/net/wireless/mediatek/mt76/dma.c
466
dma_sync_single_for_cpu(dev->dma_dev, t->dma_addr,
drivers/net/wireless/mediatek/mt76/dma.c
475
buf = t->ptr;
drivers/net/wireless/mediatek/mt76/dma.c
482
t->ptr = NULL;
drivers/net/wireless/mediatek/mt76/dma.c
483
t->dma_addr = 0;
drivers/net/wireless/mediatek/mt76/dma.c
485
mt76_put_rxwi(dev, t);
drivers/net/wireless/mediatek/mt76/dma.c
52
struct mt76_txwi_cache *t = NULL;
drivers/net/wireless/mediatek/mt76/dma.c
520
struct mt76_txwi_cache *t = mt76_rx_token_release(dev, token);
drivers/net/wireless/mediatek/mt76/dma.c
522
if (!t)
drivers/net/wireless/mediatek/mt76/dma.c
525
dma_sync_single_for_cpu(dev->dma_dev, t->dma_addr,
drivers/net/wireless/mediatek/mt76/dma.c
529
buf = t->ptr;
drivers/net/wireless/mediatek/mt76/dma.c
530
t->dma_addr = 0;
drivers/net/wireless/mediatek/mt76/dma.c
531
t->ptr = NULL;
drivers/net/wireless/mediatek/mt76/dma.c
533
mt76_put_rxwi(dev, t);
drivers/net/wireless/mediatek/mt76/dma.c
56
t = list_first_entry(&dev->txwi_cache, struct mt76_txwi_cache,
drivers/net/wireless/mediatek/mt76/dma.c
58
list_del(&t->list);
drivers/net/wireless/mediatek/mt76/dma.c
62
return t;
drivers/net/wireless/mediatek/mt76/dma.c
650
struct mt76_txwi_cache *t;
drivers/net/wireless/mediatek/mt76/dma.c
662
t = mt76_get_txwi(dev);
drivers/net/wireless/mediatek/mt76/dma.c
663
if (!t)
drivers/net/wireless/mediatek/mt76/dma.c
666
txwi = mt76_get_txwi_ptr(dev, t);
drivers/net/wireless/mediatek/mt76/dma.c
677
tx_info.buf[n].addr = t->dma_addr;
drivers/net/wireless/mediatek/mt76/dma.c
68
struct mt76_txwi_cache *t = NULL;
drivers/net/wireless/mediatek/mt76/dma.c
701
dma_sync_single_for_cpu(dev->dma_dev, t->dma_addr, dev->drv->txwi_size,
drivers/net/wireless/mediatek/mt76/dma.c
704
dma_sync_single_for_device(dev->dma_dev, t->dma_addr, dev->drv->txwi_size,
drivers/net/wireless/mediatek/mt76/dma.c
713
tx_info.info, tx_info.skb, t);
drivers/net/wireless/mediatek/mt76/dma.c
72
t = list_first_entry(&dev->rxwi_cache, struct mt76_txwi_cache,
drivers/net/wireless/mediatek/mt76/dma.c
731
mt76_put_txwi(dev, t);
drivers/net/wireless/mediatek/mt76/dma.c
74
list_del(&t->list);
drivers/net/wireless/mediatek/mt76/dma.c
78
return t;
drivers/net/wireless/mediatek/mt76/dma.c
84
struct mt76_txwi_cache *t = __mt76_get_txwi(dev);
drivers/net/wireless/mediatek/mt76/dma.c
86
if (t)
drivers/net/wireless/mediatek/mt76/dma.c
87
return t;
drivers/net/wireless/mediatek/mt76/dma.c
95
struct mt76_txwi_cache *t = __mt76_get_rxwi(dev);
drivers/net/wireless/mediatek/mt76/dma.c
97
if (t)
drivers/net/wireless/mediatek/mt76/dma.c
98
return t;
drivers/net/wireless/mediatek/mt76/mt76.h
1385
mt76_get_txwi_ptr(struct mt76_dev *dev, struct mt76_txwi_cache *t)
drivers/net/wireless/mediatek/mt76/mt76.h
1387
return (u8 *)t - dev->drv->txwi_size;
drivers/net/wireless/mediatek/mt76/mt76.h
1764
void mt76_put_txwi(struct mt76_dev *dev, struct mt76_txwi_cache *t);
drivers/net/wireless/mediatek/mt76/mt76.h
1765
void mt76_put_rxwi(struct mt76_dev *dev, struct mt76_txwi_cache *t);
drivers/net/wireless/mediatek/mt76/mt7603/beacon.c
101
void mt7603_pre_tbtt_tasklet(struct tasklet_struct *t)
drivers/net/wireless/mediatek/mt76/mt7603/beacon.c
103
struct mt7603_dev *dev = from_tasklet(dev, t, mt76.pre_tbtt_tasklet);
drivers/net/wireless/mediatek/mt76/mt7603/mt7603.h
254
void mt7603_pre_tbtt_tasklet(struct tasklet_struct *t);
drivers/net/wireless/mediatek/mt76/mt7615/mmio.c
86
static void mt7615_irq_tasklet(struct tasklet_struct *t)
drivers/net/wireless/mediatek/mt76/mt7615/mmio.c
88
struct mt7615_dev *dev = from_tasklet(dev, t, mt76.irq_tasklet);
drivers/net/wireless/mediatek/mt76/mt7615/pci_mac.c
70
struct mt76_txwi_cache *t;
drivers/net/wireless/mediatek/mt76/mt7615/pci_mac.c
91
t = (struct mt76_txwi_cache *)(txwi + mdev->drv->txwi_size);
drivers/net/wireless/mediatek/mt76/mt7615/pci_mac.c
92
t->skb = tx_info->skb;
drivers/net/wireless/mediatek/mt76/mt7615/pci_mac.c
94
id = mt76_token_get(mdev, &t);
drivers/net/wireless/mediatek/mt76/mt76_connac.h
310
mt76_connac_txwi_to_txp(struct mt76_dev *dev, struct mt76_txwi_cache *t)
drivers/net/wireless/mediatek/mt76/mt76_connac.h
314
if (!t)
drivers/net/wireless/mediatek/mt76/mt76_connac.h
317
txwi = mt76_get_txwi_ptr(dev, t);
drivers/net/wireless/mediatek/mt76/mt76_connac.h
451
void mt76_connac2_txwi_free(struct mt76_dev *dev, struct mt76_txwi_cache *t,
drivers/net/wireless/mediatek/mt76/mt76_connac_mac.c
1161
void mt76_connac2_txwi_free(struct mt76_dev *dev, struct mt76_txwi_cache *t,
drivers/net/wireless/mediatek/mt76/mt76_connac_mac.c
1169
mt76_connac_txp_skb_unmap(dev, t);
drivers/net/wireless/mediatek/mt76/mt76_connac_mac.c
1170
if (!t->skb)
drivers/net/wireless/mediatek/mt76/mt76_connac_mac.c
1173
txwi = (__le32 *)mt76_get_txwi_ptr(dev, t);
drivers/net/wireless/mediatek/mt76/mt76_connac_mac.c
1188
if (sta && likely(t->skb->protocol != cpu_to_be16(ETH_P_PAE)))
drivers/net/wireless/mediatek/mt76/mt76_connac_mac.c
1191
__mt76_tx_complete_skb(dev, wcid_idx, t->skb, free_list);
drivers/net/wireless/mediatek/mt76/mt76_connac_mac.c
1193
t->skb = NULL;
drivers/net/wireless/mediatek/mt76/mt76_connac_mac.c
1194
mt76_put_txwi(dev, t);
drivers/net/wireless/mediatek/mt76/mt76_connac_mac.c
249
struct mt76_txwi_cache *t)
drivers/net/wireless/mediatek/mt76/mt76_connac_mac.c
253
txp = mt76_connac_txwi_to_txp(dev, t);
drivers/net/wireless/mediatek/mt76/mt76x0/eeprom.c
154
struct mt76x02_rate_power *t)
drivers/net/wireless/mediatek/mt76/mt76x0/eeprom.c
160
memset(t, 0, sizeof(*t));
drivers/net/wireless/mediatek/mt76/mt76x0/eeprom.c
164
t->cck[0] = t->cck[1] = s6_to_s8(val);
drivers/net/wireless/mediatek/mt76/mt76x0/eeprom.c
165
t->cck[2] = t->cck[3] = s6_to_s8(val >> 8);
drivers/net/wireless/mediatek/mt76/mt76x0/eeprom.c
170
t->ofdm[0] = t->ofdm[1] = s6_to_s8(val);
drivers/net/wireless/mediatek/mt76/mt76x0/eeprom.c
171
t->ofdm[2] = t->ofdm[3] = s6_to_s8(val >> 8);
drivers/net/wireless/mediatek/mt76/mt76x0/eeprom.c
176
t->ofdm[4] = t->ofdm[5] = s6_to_s8(val);
drivers/net/wireless/mediatek/mt76/mt76x0/eeprom.c
177
t->ofdm[6] = t->ofdm[7] = s6_to_s8(val >> 8);
drivers/net/wireless/mediatek/mt76/mt76x0/eeprom.c
182
t->ht[0] = t->ht[1] = s6_to_s8(val);
drivers/net/wireless/mediatek/mt76/mt76x0/eeprom.c
183
t->ht[2] = t->ht[3] = s6_to_s8(val >> 8);
drivers/net/wireless/mediatek/mt76/mt76x0/eeprom.c
188
t->ht[4] = t->ht[5] = s6_to_s8(val);
drivers/net/wireless/mediatek/mt76/mt76x0/eeprom.c
189
t->ht[6] = t->ht[7] = s6_to_s8(val >> 8);
drivers/net/wireless/mediatek/mt76/mt76x0/eeprom.c
193
t->vht[0] = s6_to_s8(val);
drivers/net/wireless/mediatek/mt76/mt76x0/eeprom.c
194
t->vht[1] = s6_to_s8(val >> 8);
drivers/net/wireless/mediatek/mt76/mt76x0/eeprom.c
197
mt76x02_add_rate_power_offset(t, delta);
drivers/net/wireless/mediatek/mt76/mt76x0/eeprom.h
22
struct mt76x02_rate_power *t);
drivers/net/wireless/mediatek/mt76/mt76x0/init.c
220
struct mt76x02_rate_power t;
drivers/net/wireless/mediatek/mt76/mt76x0/init.c
227
mt76x0_get_tx_power_per_rate(dev, chan, &t);
drivers/net/wireless/mediatek/mt76/mt76x0/init.c
230
chan->orig_mpwr = (mt76x02_get_max_rate_power(&t) + tp) / 2;
drivers/net/wireless/mediatek/mt76/mt76x0/phy.c
846
struct mt76x02_rate_power *t = &dev->rate_power;
drivers/net/wireless/mediatek/mt76/mt76x0/phy.c
849
mt76x0_get_tx_power_per_rate(dev, dev->mphy.chandef.chan, t);
drivers/net/wireless/mediatek/mt76/mt76x0/phy.c
852
mt76x02_add_rate_power_offset(t, info);
drivers/net/wireless/mediatek/mt76/mt76x0/phy.c
853
mt76x02_limit_rate_power(t, dev->txpower_conf);
drivers/net/wireless/mediatek/mt76/mt76x0/phy.c
854
dev->mphy.txpower_cur = mt76x02_get_max_rate_power(t);
drivers/net/wireless/mediatek/mt76/mt76x0/phy.c
855
mt76x02_add_rate_power_offset(t, -info);
drivers/net/wireless/mediatek/mt76/mt76x02_dfs.c
612
static void mt76x02_dfs_tasklet(struct tasklet_struct *t)
drivers/net/wireless/mediatek/mt76/mt76x02_dfs.c
614
struct mt76x02_dfs_pattern_detector *dfs_pd = from_tasklet(dfs_pd, t,
drivers/net/wireless/mediatek/mt76/mt76x02_mmio.c
14
static void mt76x02_pre_tbtt_tasklet(struct tasklet_struct *t)
drivers/net/wireless/mediatek/mt76/mt76x02_mmio.c
16
struct mt76x02_dev *dev = from_tasklet(dev, t, mt76.pre_tbtt_tasklet);
drivers/net/wireless/mediatek/mt76/mt76x02_phy.c
101
mt76x02_tx_power_mask(t->cck[0], t->cck[2], t->ofdm[0],
drivers/net/wireless/mediatek/mt76/mt76x02_phy.c
102
t->ofdm[2]));
drivers/net/wireless/mediatek/mt76/mt76x02_phy.c
104
mt76x02_tx_power_mask(t->ofdm[4], t->ofdm[6], t->ht[0],
drivers/net/wireless/mediatek/mt76/mt76x02_phy.c
105
t->ht[2]));
drivers/net/wireless/mediatek/mt76/mt76x02_phy.c
107
mt76x02_tx_power_mask(t->ht[4], t->ht[6], t->ht[8],
drivers/net/wireless/mediatek/mt76/mt76x02_phy.c
108
t->ht[10]));
drivers/net/wireless/mediatek/mt76/mt76x02_phy.c
110
mt76x02_tx_power_mask(t->ht[12], t->ht[14], t->ht[0],
drivers/net/wireless/mediatek/mt76/mt76x02_phy.c
111
t->ht[2]));
drivers/net/wireless/mediatek/mt76/mt76x02_phy.c
113
mt76x02_tx_power_mask(t->ht[4], t->ht[6], 0, 0));
drivers/net/wireless/mediatek/mt76/mt76x02_phy.c
115
mt76x02_tx_power_mask(t->ofdm[7], t->vht[0], t->ht[7],
drivers/net/wireless/mediatek/mt76/mt76x02_phy.c
116
t->vht[1]));
drivers/net/wireless/mediatek/mt76/mt76x02_phy.c
118
mt76x02_tx_power_mask(t->ht[14], 0, t->vht[0], t->vht[1]));
drivers/net/wireless/mediatek/mt76/mt76x02_phy.c
120
mt76x02_tx_power_mask(t->ht[7], 0, t->vht[0], t->vht[1]));
drivers/net/wireless/mediatek/mt76/mt76x02_phy.c
95
struct mt76x02_rate_power *t = &dev->rate_power;
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
292
void mt76x2_get_rate_power(struct mt76x02_dev *dev, struct mt76x02_rate_power *t,
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
300
memset(t, 0, sizeof(*t));
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
303
t->cck[0] = t->cck[1] = mt76x02_rate_power_val(val);
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
304
t->cck[2] = t->cck[3] = mt76x02_rate_power_val(val >> 8);
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
310
t->ofdm[0] = t->ofdm[1] = mt76x02_rate_power_val(val);
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
311
t->ofdm[2] = t->ofdm[3] = mt76x02_rate_power_val(val >> 8);
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
317
t->ofdm[4] = t->ofdm[5] = mt76x02_rate_power_val(val);
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
318
t->ofdm[6] = t->ofdm[7] = mt76x02_rate_power_val(val >> 8);
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
321
t->ht[0] = t->ht[1] = mt76x02_rate_power_val(val);
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
322
t->ht[2] = t->ht[3] = mt76x02_rate_power_val(val >> 8);
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
325
t->ht[4] = t->ht[5] = mt76x02_rate_power_val(val);
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
326
t->ht[6] = t->ht[7] = mt76x02_rate_power_val(val >> 8);
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
329
t->ht[8] = t->ht[9] = mt76x02_rate_power_val(val);
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
330
t->ht[10] = t->ht[11] = mt76x02_rate_power_val(val >> 8);
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
333
t->ht[12] = t->ht[13] = mt76x02_rate_power_val(val);
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
334
t->ht[14] = t->ht[15] = mt76x02_rate_power_val(val >> 8);
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
339
t->vht[0] = t->vht[1] = mt76x02_rate_power_val(val >> 8);
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
345
struct mt76x2_tx_power_info *t,
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
363
t->chain[chain].tssi_slope = data[0];
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
364
t->chain[chain].tssi_offset = data[1];
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
365
t->chain[chain].target_power = data[2];
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
366
t->chain[chain].delta =
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
370
t->target_power = val >> 8;
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
375
struct mt76x2_tx_power_info *t,
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
415
t->chain[chain].tssi_slope = data[0];
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
416
t->chain[chain].tssi_offset = data[1];
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
417
t->chain[chain].target_power = data[2];
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
418
t->chain[chain].delta =
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
422
t->target_power = val & 0xff;
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
426
struct mt76x2_tx_power_info *t,
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
431
memset(t, 0, sizeof(*t));
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
438
mt76x2_get_power_info_5g(dev, t, chan, 0,
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
440
mt76x2_get_power_info_5g(dev, t, chan, 1,
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
443
mt76x2_get_power_info_2g(dev, t, chan, 0,
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
445
mt76x2_get_power_info_2g(dev, t, chan, 1,
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
450
!mt76x02_field_valid(t->target_power))
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
451
t->target_power = t->chain[0].target_power;
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
453
t->delta_bw40 = mt76x02_rate_power_val(bw40);
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
454
t->delta_bw80 = mt76x02_rate_power_val(bw80);
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
458
int mt76x2_get_temp_comp(struct mt76x02_dev *dev, struct mt76x2_temp_comp *t)
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
464
memset(t, 0, sizeof(*t));
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
473
t->temp_25_ref = val & 0x7f;
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
483
t->high_slope = slope & 0xff;
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
484
t->low_slope = slope >> 8;
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
485
t->lower_bound = 0 - (bounds & 0xf);
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
486
t->upper_bound = (bounds >> 4) & 0xf;
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.h
43
void mt76x2_get_rate_power(struct mt76x02_dev *dev, struct mt76x02_rate_power *t,
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.h
46
struct mt76x2_tx_power_info *t,
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.h
48
int mt76x2_get_temp_comp(struct mt76x02_dev *dev, struct mt76x2_temp_comp *t);
drivers/net/wireless/mediatek/mt76/mt76x2/init.c
185
struct mt76x02_rate_power t = {};
drivers/net/wireless/mediatek/mt76/mt76x2/init.c
192
mt76x2_get_rate_power(dev, &t, chan);
drivers/net/wireless/mediatek/mt76/mt76x2/init.c
194
chan->orig_mpwr = mt76x02_get_max_rate_power(&t) +
drivers/net/wireless/mediatek/mt76/mt76x2/pci_phy.c
256
struct mt76x2_temp_comp t;
drivers/net/wireless/mediatek/mt76/mt76x2/pci_phy.c
259
if (mt76x2_get_temp_comp(dev, &t))
drivers/net/wireless/mediatek/mt76/mt76x2/pci_phy.c
263
temp -= t.temp_25_ref;
drivers/net/wireless/mediatek/mt76/mt76x2/pci_phy.c
268
db_diff = (temp - 25) / t.high_slope;
drivers/net/wireless/mediatek/mt76/mt76x2/pci_phy.c
270
db_diff = (25 - temp) / t.low_slope;
drivers/net/wireless/mediatek/mt76/mt76x2/pci_phy.c
272
db_diff = min(db_diff, t.upper_bound);
drivers/net/wireless/mediatek/mt76/mt76x2/pci_phy.c
273
db_diff = max(db_diff, t.lower_bound);
drivers/net/wireless/mediatek/mt76/mt76x2/phy.c
143
struct mt76x02_rate_power t = {};
drivers/net/wireless/mediatek/mt76/mt76x2/phy.c
153
mt76x2_get_rate_power(dev, &t, chan);
drivers/net/wireless/mediatek/mt76/mt76x2/phy.c
154
mt76x02_add_rate_power_offset(&t, txp.target_power + delta);
drivers/net/wireless/mediatek/mt76/mt76x2/phy.c
155
mt76x02_limit_rate_power(&t, dev->txpower_conf);
drivers/net/wireless/mediatek/mt76/mt76x2/phy.c
156
dev->mphy.txpower_cur = mt76x02_get_max_rate_power(&t);
drivers/net/wireless/mediatek/mt76/mt76x2/phy.c
158
base_power = mt76x2_get_min_rate_power(&t);
drivers/net/wireless/mediatek/mt76/mt76x2/phy.c
174
mt76x02_add_rate_power_offset(&t, -base_power);
drivers/net/wireless/mediatek/mt76/mt76x2/phy.c
178
dev->rate_power = t;
drivers/net/wireless/mediatek/mt76/mt76x2/phy.c
207
struct mt76x2_tssi_comp t = {};
drivers/net/wireless/mediatek/mt76/mt76x2/phy.c
214
t.cal_mode = BIT(0);
drivers/net/wireless/mediatek/mt76/mt76x2/phy.c
215
mt76x2_mcu_tssi_comp(dev, &t);
drivers/net/wireless/mediatek/mt76/mt76x2/phy.c
225
t.pa_mode = 1;
drivers/net/wireless/mediatek/mt76/mt76x2/phy.c
227
t.cal_mode = BIT(1);
drivers/net/wireless/mediatek/mt76/mt76x2/phy.c
228
t.slope0 = txp.chain[0].tssi_slope;
drivers/net/wireless/mediatek/mt76/mt76x2/phy.c
229
t.offset0 = txp.chain[0].tssi_offset;
drivers/net/wireless/mediatek/mt76/mt76x2/phy.c
230
t.slope1 = txp.chain[1].tssi_slope;
drivers/net/wireless/mediatek/mt76/mt76x2/phy.c
231
t.offset1 = txp.chain[1].tssi_offset;
drivers/net/wireless/mediatek/mt76/mt76x2/phy.c
232
mt76x2_mcu_tssi_comp(dev, &t);
drivers/net/wireless/mediatek/mt76/mt76x2/phy.c
234
if (t.pa_mode || dev->cal.dpd_cal_done || dev->ed_tx_blocked)
drivers/net/wireless/mediatek/mt76/mt7915/coredump.h
43
struct trace t;
drivers/net/wireless/mediatek/mt76/mt7915/mac.c
752
struct mt76_txwi_cache *t;
drivers/net/wireless/mediatek/mt76/mt7915/mac.c
774
t = (struct mt76_txwi_cache *)(txwi + mdev->drv->txwi_size);
drivers/net/wireless/mediatek/mt76/mt7915/mac.c
775
t->skb = tx_info->skb;
drivers/net/wireless/mediatek/mt76/mt7915/mac.c
777
id = mt76_token_consume(mdev, &t);
drivers/net/wireless/mediatek/mt76/mt7915/mcu.c
288
struct mt7915_mcu_thermal_notify *t;
drivers/net/wireless/mediatek/mt76/mt7915/mcu.c
291
t = (struct mt7915_mcu_thermal_notify *)skb->data;
drivers/net/wireless/mediatek/mt76/mt7915/mcu.c
292
if (t->ctrl.ctrl_id != THERMAL_PROTECT_ENABLE)
drivers/net/wireless/mediatek/mt76/mt7915/mcu.c
295
if (t->ctrl.band_idx > MT_BAND1)
drivers/net/wireless/mediatek/mt76/mt7915/mcu.c
298
if ((t->ctrl.band_idx && !dev->phy.mt76->band_idx) &&
drivers/net/wireless/mediatek/mt76/mt7915/mcu.c
303
phy->throttle_state = t->ctrl.duty.duty_cycle;
drivers/net/wireless/mediatek/mt76/mt7915/mmio.c
826
static void mt7915_irq_tasklet(struct tasklet_struct *t)
drivers/net/wireless/mediatek/mt76/mt7915/mmio.c
828
struct mt7915_dev *dev = from_tasklet(dev, t, mt76.irq_tasklet);
drivers/net/wireless/mediatek/mt76/mt7921/pci_mac.c
17
struct mt76_txwi_cache *t;
drivers/net/wireless/mediatek/mt76/mt7921/pci_mac.c
27
t = (struct mt76_txwi_cache *)(txwi + mdev->drv->txwi_size);
drivers/net/wireless/mediatek/mt76/mt7921/pci_mac.c
28
t->skb = tx_info->skb;
drivers/net/wireless/mediatek/mt76/mt7921/pci_mac.c
30
id = mt76_token_consume(mdev, &t);
drivers/net/wireless/mediatek/mt76/mt7925/mac.c
1061
void mt7925_txwi_free(struct mt792x_dev *dev, struct mt76_txwi_cache *t,
drivers/net/wireless/mediatek/mt76/mt7925/mac.c
1069
mt76_connac_txp_skb_unmap(mdev, t);
drivers/net/wireless/mediatek/mt76/mt7925/mac.c
1070
if (!t->skb)
drivers/net/wireless/mediatek/mt76/mt7925/mac.c
1073
txwi = (__le32 *)mt76_get_txwi_ptr(mdev, t);
drivers/net/wireless/mediatek/mt76/mt7925/mac.c
1075
if (likely(t->skb->protocol != cpu_to_be16(ETH_P_PAE)))
drivers/net/wireless/mediatek/mt76/mt7925/mac.c
1076
mt7925_tx_check_aggr(sta, t->skb, wcid);
drivers/net/wireless/mediatek/mt76/mt7925/mac.c
1083
__mt76_tx_complete_skb(mdev, wcid_idx, t->skb, free_list);
drivers/net/wireless/mediatek/mt76/mt7925/mac.c
1085
t->skb = NULL;
drivers/net/wireless/mediatek/mt76/mt7925/mac.c
1086
mt76_put_txwi(mdev, t);
drivers/net/wireless/mediatek/mt76/mt7925/mt7925.h
312
void mt7925_txwi_free(struct mt792x_dev *dev, struct mt76_txwi_cache *t,
drivers/net/wireless/mediatek/mt76/mt7925/pci_mac.c
17
struct mt76_txwi_cache *t;
drivers/net/wireless/mediatek/mt76/mt7925/pci_mac.c
27
t = (struct mt76_txwi_cache *)(txwi + mdev->drv->txwi_size);
drivers/net/wireless/mediatek/mt76/mt7925/pci_mac.c
28
t->skb = tx_info->skb;
drivers/net/wireless/mediatek/mt76/mt7925/pci_mac.c
30
id = mt76_token_consume(mdev, &t);
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1052
struct mt76_txwi_cache *t;
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1089
t = (struct mt76_txwi_cache *)(txwi + mdev->drv->txwi_size);
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1090
t->skb = tx_info->skb;
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1092
id = mt76_token_consume(mdev, &t);
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1278
mt7996_txwi_free(struct mt7996_dev *dev, struct mt76_txwi_cache *t,
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1286
mt76_connac_txp_skb_unmap(mdev, t);
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1287
if (!t->skb)
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1290
txwi = (__le32 *)mt76_get_txwi_ptr(mdev, t);
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1293
if (likely(t->skb->protocol != cpu_to_be16(ETH_P_PAE))) {
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1299
t->skb);
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1305
__mt76_tx_complete_skb(mdev, wcid_idx, t->skb, free_list);
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1308
t->skb = NULL;
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1309
mt76_put_txwi(mdev, t);
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1809
struct mt76_txwi_cache *t;
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1811
t = mt76_rx_token_release(&dev->mt76, i);
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1812
if (!t || !t->ptr)
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1815
mt76_put_page_pool_buf(t->ptr, false);
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1816
t->dma_addr = 0;
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1817
t->ptr = NULL;
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1819
mt76_put_rxwi(&dev->mt76, t);
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1913
struct mt76_txwi_cache *t;
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1960
t = mt76_rx_token_release(mdev, rx_token_id);
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1961
if (!t)
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1964
qid = t->qid;
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1965
buf = t->ptr;
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1967
dma_sync_single_for_cpu(mdev->dma_dev, t->dma_addr,
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1971
t->dma_addr = 0;
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1972
t->ptr = NULL;
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1973
mt76_put_rxwi(mdev, t);
drivers/net/wireless/mediatek/mt76/mt7996/mmio.c
725
static void mt7996_irq_tasklet(struct tasklet_struct *t)
drivers/net/wireless/mediatek/mt76/mt7996/mmio.c
727
struct mt7996_dev *dev = from_tasklet(dev, t, mt76.irq_tasklet);
drivers/net/wireless/mediatek/mt76/tx.c
872
struct mt76_txwi_cache *t, dma_addr_t phys)
drivers/net/wireless/mediatek/mt76/tx.c
877
token = idr_alloc(&dev->rx_token, t, 0, dev->rx_token_size,
drivers/net/wireless/mediatek/mt76/tx.c
880
t->ptr = ptr;
drivers/net/wireless/mediatek/mt76/tx.c
881
t->dma_addr = phys;
drivers/net/wireless/mediatek/mt76/tx.c
921
struct mt76_txwi_cache *t;
drivers/net/wireless/mediatek/mt76/tx.c
924
t = idr_remove(&dev->rx_token, token);
drivers/net/wireless/mediatek/mt76/tx.c
927
return t;
drivers/net/wireless/mediatek/mt76/wed.c
15
struct mt76_txwi_cache *t;
drivers/net/wireless/mediatek/mt76/wed.c
17
t = mt76_rx_token_release(dev, i);
drivers/net/wireless/mediatek/mt76/wed.c
18
if (!t || !t->ptr)
drivers/net/wireless/mediatek/mt76/wed.c
21
mt76_put_page_pool_buf(t->ptr, false);
drivers/net/wireless/mediatek/mt76/wed.c
22
t->ptr = NULL;
drivers/net/wireless/mediatek/mt76/wed.c
24
mt76_put_rxwi(dev, t);
drivers/net/wireless/mediatek/mt76/wed.c
37
struct mt76_txwi_cache *t = NULL;
drivers/net/wireless/mediatek/mt76/wed.c
46
t = mt76_get_rxwi(dev);
drivers/net/wireless/mediatek/mt76/wed.c
47
if (!t)
drivers/net/wireless/mediatek/mt76/wed.c
56
token = mt76_rx_token_consume(dev, buf, t, addr);
drivers/net/wireless/mediatek/mt76/wed.c
73
if (t)
drivers/net/wireless/mediatek/mt76/wed.c
74
mt76_put_rxwi(dev, t);
drivers/net/wireless/mediatek/mt7601u/dma.c
225
static void mt7601u_rx_tasklet(struct tasklet_struct *t)
drivers/net/wireless/mediatek/mt7601u/dma.c
227
struct mt7601u_dev *dev = from_tasklet(dev, t, rx_tasklet);
drivers/net/wireless/mediatek/mt7601u/dma.c
280
static void mt7601u_tx_tasklet(struct tasklet_struct *t)
drivers/net/wireless/mediatek/mt7601u/dma.c
282
struct mt7601u_dev *dev = from_tasklet(dev, t, tx_tasklet);
drivers/net/wireless/mediatek/mt7601u/eeprom.c
265
struct mt7601u_rate_power *t = &dev->ee->power_rate_table;
drivers/net/wireless/mediatek/mt7601u/eeprom.c
269
mt7601u_set_power_rate(&t->cck[0], delta, (val >> 0) & 0xff);
drivers/net/wireless/mediatek/mt7601u/eeprom.c
270
mt7601u_set_power_rate(&t->cck[1], delta, (val >> 8) & 0xff);
drivers/net/wireless/mediatek/mt7601u/eeprom.c
272
dev->ee->real_cck_bw20[0] = t->cck[0].bw20;
drivers/net/wireless/mediatek/mt7601u/eeprom.c
273
dev->ee->real_cck_bw20[1] = t->cck[1].bw20;
drivers/net/wireless/mediatek/mt7601u/eeprom.c
275
mt7601u_set_power_rate(&t->ofdm[0], delta, (val >> 16) & 0xff);
drivers/net/wireless/mediatek/mt7601u/eeprom.c
276
mt7601u_set_power_rate(&t->ofdm[1], delta, (val >> 24) & 0xff);
drivers/net/wireless/mediatek/mt7601u/eeprom.c
279
mt7601u_set_power_rate(&t->ofdm[2], delta, (val >> 0) & 0xff);
drivers/net/wireless/mediatek/mt7601u/eeprom.c
280
mt7601u_set_power_rate(&t->ofdm[3], delta, (val >> 8) & 0xff);
drivers/net/wireless/mediatek/mt7601u/eeprom.c
281
mt7601u_set_power_rate(&t->ht[0], delta, (val >> 16) & 0xff);
drivers/net/wireless/mediatek/mt7601u/eeprom.c
282
mt7601u_set_power_rate(&t->ht[1], delta, (val >> 24) & 0xff);
drivers/net/wireless/mediatek/mt7601u/eeprom.c
285
mt7601u_set_power_rate(&t->ht[2], delta, (val >> 0) & 0xff);
drivers/net/wireless/mediatek/mt7601u/eeprom.c
286
mt7601u_set_power_rate(&t->ht[3], delta, (val >> 8) & 0xff);
drivers/net/wireless/mediatek/mt7601u/phy.c
291
const struct reg_table *t;
drivers/net/wireless/mediatek/mt7601u/phy.c
296
t = &bbp_mode_table[dev->temp_mode][dev->bw];
drivers/net/wireless/mediatek/mt7601u/phy.c
298
return mt7601u_write_reg_pairs(dev, MT_MCU_MEMMAP_BBP, t->regs, t->n);
drivers/net/wireless/mediatek/mt7601u/phy.c
303
const struct reg_table *t;
drivers/net/wireless/mediatek/mt7601u/phy.c
312
t = bbp_mode_table[dev->temp_mode];
drivers/net/wireless/mediatek/mt7601u/phy.c
314
t[2].regs, t[2].n);
drivers/net/wireless/mediatek/mt7601u/phy.c
319
t[dev->bw].regs, t[dev->bw].n);
drivers/net/wireless/mediatek/mt7601u/phy.c
324
struct mt7601u_rate_power *t = &dev->ee->power_rate_table;
drivers/net/wireless/mediatek/mt7601u/phy.c
330
t->cck[0].bw20 = dev->ee->real_cck_bw20[0];
drivers/net/wireless/mediatek/mt7601u/phy.c
331
t->cck[1].bw20 = dev->ee->real_cck_bw20[1];
drivers/net/wireless/mediatek/mt7601u/phy.c
337
t->cck[0].bw20 = dev->ee->real_cck_bw20[0] - 2;
drivers/net/wireless/mediatek/mt7601u/phy.c
338
t->cck[1].bw20 = dev->ee->real_cck_bw20[1] - 2;
drivers/net/wireless/mediatek/mt7601u/phy.c
374
struct mt7601u_rate_power *t = &dev->ee->power_rate_table;
drivers/net/wireless/mediatek/mt7601u/phy.c
429
mt7601u_wr(dev, MT_TX_PWR_CFG_0, int_to_s6(t->ofdm[1].bw20) << 24 |
drivers/net/wireless/mediatek/mt7601u/phy.c
430
int_to_s6(t->ofdm[0].bw20) << 16 |
drivers/net/wireless/mediatek/mt7601u/phy.c
431
int_to_s6(t->cck[1].bw20) << 8 |
drivers/net/wireless/mediatek/mt7601u/phy.c
432
int_to_s6(t->cck[0].bw20));
drivers/net/wireless/microchip/wilc1000/hif.c
1076
static void timer_scan_cb(struct timer_list *t)
drivers/net/wireless/microchip/wilc1000/hif.c
1078
struct host_if_drv *hif_drv = timer_container_of(hif_drv, t,
drivers/net/wireless/microchip/wilc1000/hif.c
1093
static void timer_connect_cb(struct timer_list *t)
drivers/net/wireless/microchip/wilc1000/hif.c
1095
struct host_if_drv *hif_drv = timer_container_of(hif_drv, t,
drivers/net/wireless/microchip/wilc1000/hif.c
1499
static void get_periodic_rssi(struct timer_list *t)
drivers/net/wireless/microchip/wilc1000/hif.c
1501
struct wilc_vif *vif = timer_container_of(vif, t, periodic_rssi);
drivers/net/wireless/microchip/wilc1000/hif.c
954
static void listen_timer_cb(struct timer_list *t)
drivers/net/wireless/microchip/wilc1000/hif.c
956
struct host_if_drv *hif_drv = timer_container_of(hif_drv, t,
drivers/net/wireless/purelifi/plfxlc/usb.c
549
static void slif_data_plane_sap_timer_callb(struct timer_list *t)
drivers/net/wireless/purelifi/plfxlc/usb.c
551
struct plfxlc_usb *usb = timer_container_of(usb, t, tx.tx_retry_timer);
drivers/net/wireless/purelifi/plfxlc/usb.c
559
static void sta_queue_cleanup_timer_callb(struct timer_list *t)
drivers/net/wireless/purelifi/plfxlc/usb.c
561
struct plfxlc_usb *usb = timer_container_of(usb, t, sta_queue_cleanup);
drivers/net/wireless/quantenna/qtnfmac/pcie/pearl_pcie.c
1094
static void qtnf_pearl_reclaim_tasklet_fn(struct tasklet_struct *t)
drivers/net/wireless/quantenna/qtnfmac/pcie/pearl_pcie.c
1096
struct qtnf_pcie_pearl_state *ps = from_tasklet(ps, t, base.reclaim_tq);
drivers/net/wireless/quantenna/qtnfmac/pcie/topaz_pcie.c
1108
static void qtnf_reclaim_tasklet_fn(struct tasklet_struct *t)
drivers/net/wireless/quantenna/qtnfmac/pcie/topaz_pcie.c
1110
struct qtnf_pcie_topaz_state *ts = from_tasklet(ts, t, base.reclaim_tq);
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1322
static void rt2400pci_txstatus_tasklet(struct tasklet_struct *t)
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1324
struct rt2x00_dev *rt2x00dev = from_tasklet(rt2x00dev, t,
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1351
static void rt2400pci_tbtt_tasklet(struct tasklet_struct *t)
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1353
struct rt2x00_dev *rt2x00dev = from_tasklet(rt2x00dev, t, tbtt_tasklet);
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1359
static void rt2400pci_rxdone_tasklet(struct tasklet_struct *t)
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1361
struct rt2x00_dev *rt2x00dev = from_tasklet(rt2x00dev, t,
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
1450
static void rt2500pci_txstatus_tasklet(struct tasklet_struct *t)
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
1452
struct rt2x00_dev *rt2x00dev = from_tasklet(rt2x00dev, t,
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
1479
static void rt2500pci_tbtt_tasklet(struct tasklet_struct *t)
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
1481
struct rt2x00_dev *rt2x00dev = from_tasklet(rt2x00dev, t, tbtt_tasklet);
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
1487
static void rt2500pci_rxdone_tasklet(struct tasklet_struct *t)
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
1489
struct rt2x00_dev *rt2x00dev = from_tasklet(rt2x00dev, t,
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
5212
u8 t, i;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
5244
t = eeprom & 0x3f;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
5245
if (t == 32)
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
5246
t++;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
5248
gdata = t;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
5250
t = (eeprom & 0x3f00) >> 8;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
5251
if (t == 32)
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
5252
t++;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
5254
gdata |= (t << 8);
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
5260
t = eeprom & 0x3f;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
5261
if (t == 32)
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
5262
t++;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
5264
gdata |= (t << 16);
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
5266
t = (eeprom & 0x3f00) >> 8;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
5267
if (t == 32)
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
5268
t++;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
5270
gdata |= (t << 24);
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
5301
t = rt2x00_get_field32(reg, TX_PWR_CFG_1B_48MBS);
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
5302
rt2x00_set_field32(&pwreg, TX_PWR_CFG_7B_54MBS, t);
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
5306
t = rt2x00_get_field32(reg, TX_PWR_CFG_2B_MCS6_MCS7);
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
5307
rt2x00_set_field32(&pwreg, TX_PWR_CFG_7B_MCS7, t);
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
5313
t = rt2x00_get_field32(reg, TX_PWR_CFG_3B_MCS14);
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
5314
rt2x00_set_field32(&pwreg, TX_PWR_CFG_8B_MCS15, t);
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
5320
t = rt2x00_get_field32(reg, TX_PWR_CFG_4B_STBC_MCS6);
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
5321
rt2x00_set_field32(&pwreg, TX_PWR_CFG_9B_STBC_MCS7, t);
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
213
void rt2800mmio_pretbtt_tasklet(struct tasklet_struct *t)
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
215
struct rt2x00_dev *rt2x00dev = from_tasklet(rt2x00dev, t,
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
223
void rt2800mmio_tbtt_tasklet(struct tasklet_struct *t)
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
225
struct rt2x00_dev *rt2x00dev = from_tasklet(rt2x00dev, t, tbtt_tasklet);
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
258
void rt2800mmio_rxdone_tasklet(struct tasklet_struct *t)
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
260
struct rt2x00_dev *rt2x00dev = from_tasklet(rt2x00dev, t,
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
269
void rt2800mmio_autowake_tasklet(struct tasklet_struct *t)
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
271
struct rt2x00_dev *rt2x00dev = from_tasklet(rt2x00dev, t,
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
313
void rt2800mmio_txstatus_tasklet(struct tasklet_struct *t)
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
315
struct rt2x00_dev *rt2x00dev = from_tasklet(rt2x00dev, t,
drivers/net/wireless/ralink/rt2x00/rt2800mmio.h
129
void rt2800mmio_txstatus_tasklet(struct tasklet_struct *t);
drivers/net/wireless/ralink/rt2x00/rt2800mmio.h
130
void rt2800mmio_pretbtt_tasklet(struct tasklet_struct *t);
drivers/net/wireless/ralink/rt2x00/rt2800mmio.h
131
void rt2800mmio_tbtt_tasklet(struct tasklet_struct *t);
drivers/net/wireless/ralink/rt2x00/rt2800mmio.h
132
void rt2800mmio_rxdone_tasklet(struct tasklet_struct *t);
drivers/net/wireless/ralink/rt2x00/rt2800mmio.h
133
void rt2800mmio_autowake_tasklet(struct tasklet_struct *t);
drivers/net/wireless/ralink/rt2x00/rt2x00.h
520
void (*txstatus_tasklet) (struct tasklet_struct *t);
drivers/net/wireless/ralink/rt2x00/rt2x00.h
521
void (*pretbtt_tasklet) (struct tasklet_struct *t);
drivers/net/wireless/ralink/rt2x00/rt2x00.h
522
void (*tbtt_tasklet) (struct tasklet_struct *t);
drivers/net/wireless/ralink/rt2x00/rt2x00.h
523
void (*rxdone_tasklet) (struct tasklet_struct *t);
drivers/net/wireless/ralink/rt2x00/rt2x00.h
524
void (*autowake_tasklet) (struct tasklet_struct *t);
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2193
static void rt61pci_txstatus_tasklet(struct tasklet_struct *t)
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2195
struct rt2x00_dev *rt2x00dev = from_tasklet(rt2x00dev, t,
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2203
static void rt61pci_tbtt_tasklet(struct tasklet_struct *t)
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2205
struct rt2x00_dev *rt2x00dev = from_tasklet(rt2x00dev, t, tbtt_tasklet);
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2211
static void rt61pci_rxdone_tasklet(struct tasklet_struct *t)
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2213
struct rt2x00_dev *rt2x00dev = from_tasklet(rt2x00dev, t,
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2221
static void rt61pci_autowake_tasklet(struct tasklet_struct *t)
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2223
struct rt2x00_dev *rt2x00dev = from_tasklet(rt2x00dev, t,
drivers/net/wireless/realtek/rtl8xxxu/8188e.c
751
int result[][8], int t)
drivers/net/wireless/realtek/rtl8xxxu/8188e.c
783
if (t == 0) {
drivers/net/wireless/realtek/rtl8xxxu/8188e.c
794
if (t == 0) {
drivers/net/wireless/realtek/rtl8xxxu/8188e.c
845
result[t][0] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/8188e.c
848
result[t][1] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/8188e.c
861
result[t][2] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/8188e.c
864
result[t][3] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/8188e.c
878
if (t == 0)
drivers/net/wireless/realtek/rtl8xxxu/8188f.c
1062
int result[][8], int t)
drivers/net/wireless/realtek/rtl8xxxu/8188f.c
1097
if (t == 0) {
drivers/net/wireless/realtek/rtl8xxxu/8188f.c
1108
if (t == 0) {
drivers/net/wireless/realtek/rtl8xxxu/8188f.c
1144
result[t][0] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/8188f.c
1148
result[t][1] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/8188f.c
1158
result[t][2] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/8188f.c
1162
result[t][3] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/8188f.c
1175
if (t == 0)
drivers/net/wireless/realtek/rtl8xxxu/8192e.c
1051
int result[][8], int t)
drivers/net/wireless/realtek/rtl8xxxu/8192e.c
1085
if (t == 0) {
drivers/net/wireless/realtek/rtl8xxxu/8192e.c
1127
result[t][0] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/8192e.c
1130
result[t][1] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/8192e.c
1144
result[t][2] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/8192e.c
1147
result[t][3] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/8192e.c
1173
result[t][4] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/8192e.c
1175
result[t][5] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/8192e.c
1188
result[t][6] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/8192e.c
1191
result[t][7] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/8192e.c
1203
if (t) {
drivers/net/wireless/realtek/rtl8xxxu/8192f.c
1256
int result[][8], int t)
drivers/net/wireless/realtek/rtl8xxxu/8192f.c
1288
if (t == 0) {
drivers/net/wireless/realtek/rtl8xxxu/8192f.c
1339
result[t][0] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/8192f.c
1342
result[t][1] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/8192f.c
1345
result[t][0] = 0x100;
drivers/net/wireless/realtek/rtl8xxxu/8192f.c
1346
result[t][1] = 0x0;
drivers/net/wireless/realtek/rtl8xxxu/8192f.c
1355
result[t][2] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/8192f.c
1358
result[t][3] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/8192f.c
1361
result[t][2] = 0x100;
drivers/net/wireless/realtek/rtl8xxxu/8192f.c
1362
result[t][3] = 0x0;
drivers/net/wireless/realtek/rtl8xxxu/8192f.c
1375
result[t][4] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/8192f.c
1378
result[t][5] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/8192f.c
1381
result[t][4] = 0x100;
drivers/net/wireless/realtek/rtl8xxxu/8192f.c
1382
result[t][5] = 0x0;
drivers/net/wireless/realtek/rtl8xxxu/8192f.c
1391
result[t][6] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/8192f.c
1394
result[t][7] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/8192f.c
1397
result[t][6] = 0x100;
drivers/net/wireless/realtek/rtl8xxxu/8192f.c
1398
result[t][7] = 0x0;
drivers/net/wireless/realtek/rtl8xxxu/8710b.c
1255
int result[][8], int t)
drivers/net/wireless/realtek/rtl8xxxu/8710b.c
1290
if (t == 0) {
drivers/net/wireless/realtek/rtl8xxxu/8710b.c
1301
if (t == 0) {
drivers/net/wireless/realtek/rtl8xxxu/8710b.c
1342
result[t][0] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/8710b.c
1345
result[t][1] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/8710b.c
1348
result[t][0] = 0x100;
drivers/net/wireless/realtek/rtl8xxxu/8710b.c
1349
result[t][1] = 0x0;
drivers/net/wireless/realtek/rtl8xxxu/8710b.c
1358
result[t][2] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/8710b.c
1361
result[t][3] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/8710b.c
1364
result[t][2] = 0x100;
drivers/net/wireless/realtek/rtl8xxxu/8710b.c
1365
result[t][3] = 0x0;
drivers/net/wireless/realtek/rtl8xxxu/8710b.c
1377
if (t == 0)
drivers/net/wireless/realtek/rtl8xxxu/8723b.c
1002
result[t][3] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/8723b.c
1036
result[t][4] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/8723b.c
1038
result[t][5] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/8723b.c
1051
result[t][6] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/8723b.c
1054
result[t][7] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/8723b.c
1069
if (t) {
drivers/net/wireless/realtek/rtl8xxxu/8723b.c
895
int result[][8], int t)
drivers/net/wireless/realtek/rtl8xxxu/8723b.c
929
if (t == 0) {
drivers/net/wireless/realtek/rtl8xxxu/8723b.c
982
result[t][0] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/8723b.c
985
result[t][1] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/8723b.c
999
result[t][2] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/core.c
3183
int result[][8], int t)
drivers/net/wireless/realtek/rtl8xxxu/core.c
3215
if (t == 0) {
drivers/net/wireless/realtek/rtl8xxxu/core.c
3226
if (t == 0) {
drivers/net/wireless/realtek/rtl8xxxu/core.c
3284
result[t][0] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/core.c
3287
result[t][1] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/core.c
3290
result[t][2] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/core.c
3293
result[t][3] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/core.c
3302
result[t][0] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/core.c
3305
result[t][1] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/core.c
3327
result[t][4] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/core.c
3329
result[t][5] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/core.c
3331
result[t][6] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/core.c
3333
result[t][7] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/core.c
3338
result[t][4] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/core.c
3340
result[t][5] = (val32 >> 16) & 0x3ff;
drivers/net/wireless/realtek/rtl8xxxu/core.c
3351
if (t) {
drivers/net/wireless/realtek/rtlwifi/base.c
2221
void rtl_watch_dog_timer_callback(struct timer_list *t)
drivers/net/wireless/realtek/rtlwifi/base.c
2223
struct rtl_priv *rtlpriv = timer_container_of(rtlpriv, t,
drivers/net/wireless/realtek/rtlwifi/base.h
74
void rtl_watch_dog_timer_callback(struct timer_list *t);
drivers/net/wireless/realtek/rtlwifi/pci.c
1003
static void _rtl_pci_prepare_bcn_tasklet(struct tasklet_struct *t)
drivers/net/wireless/realtek/rtlwifi/pci.c
1005
struct rtl_priv *rtlpriv = from_tasklet(rtlpriv, t,
drivers/net/wireless/realtek/rtlwifi/pci.c
996
static void _rtl_pci_irq_tasklet(struct tasklet_struct *t)
drivers/net/wireless/realtek/rtlwifi/pci.c
998
struct rtl_priv *rtlpriv = from_tasklet(rtlpriv, t, works.irq_tasklet);
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/dm.c
1684
void rtl88e_dm_fast_antenna_training_callback(struct timer_list *t)
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/dm.c
1687
timer_container_of(rtlpriv, t,
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/dm.h
251
void rtl88e_dm_fast_antenna_training_callback(struct timer_list *t);
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c
238
void rtl88ee_fw_clk_off_timer_callback(struct timer_list *t)
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c
240
struct rtl_priv *rtlpriv = timer_container_of(rtlpriv, t,
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.h
39
void rtl88ee_fw_clk_off_timer_callback(struct timer_list *t);
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/phy.c
1674
long result[][8], u8 t, bool is2t)
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/phy.c
1696
if (t == 0) {
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/phy.c
1706
if (t == 0) {
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/phy.c
1742
result[t][0] = (rtl_get_bbreg(hw, 0xe94, MASKDWORD) &
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/phy.c
1744
result[t][1] = (rtl_get_bbreg(hw, 0xe9c, MASKDWORD) &
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/phy.c
1755
result[t][2] = (rtl_get_bbreg(hw, 0xea4, MASKDWORD) &
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/phy.c
1757
result[t][3] = (rtl_get_bbreg(hw, 0xeac, MASKDWORD) &
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/phy.c
1775
result[t][4] = (rtl_get_bbreg(hw,
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/phy.c
1779
result[t][5] =
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/phy.c
1782
result[t][6] =
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/phy.c
1785
result[t][7] =
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/phy.c
1790
result[t][4] = (rtl_get_bbreg(hw,
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/phy.c
1795
result[t][5] = (rtl_get_bbreg(hw, 0xebc, MASKDWORD) &
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/phy.c
1802
if (t != 0) {
drivers/net/wireless/realtek/rtlwifi/rtl8192c/phy_common.c
1195
long result[][8], u8 t, bool is2t)
drivers/net/wireless/realtek/rtlwifi/rtl8192c/phy_common.c
1212
if (t == 0) {
drivers/net/wireless/realtek/rtlwifi/rtl8192c/phy_common.c
1221
if (t == 0) {
drivers/net/wireless/realtek/rtlwifi/rtl8192c/phy_common.c
1229
if (t == 0) {
drivers/net/wireless/realtek/rtlwifi/rtl8192c/phy_common.c
1252
result[t][0] = (rtl_get_bbreg(hw, 0xe94, MASKDWORD) &
drivers/net/wireless/realtek/rtlwifi/rtl8192c/phy_common.c
1254
result[t][1] = (rtl_get_bbreg(hw, 0xe9c, MASKDWORD) &
drivers/net/wireless/realtek/rtlwifi/rtl8192c/phy_common.c
1256
result[t][2] = (rtl_get_bbreg(hw, 0xea4, MASKDWORD) &
drivers/net/wireless/realtek/rtlwifi/rtl8192c/phy_common.c
1258
result[t][3] = (rtl_get_bbreg(hw, 0xeac, MASKDWORD) &
drivers/net/wireless/realtek/rtlwifi/rtl8192c/phy_common.c
1263
result[t][0] = (rtl_get_bbreg(hw, 0xe94,
drivers/net/wireless/realtek/rtlwifi/rtl8192c/phy_common.c
1266
result[t][1] =
drivers/net/wireless/realtek/rtlwifi/rtl8192c/phy_common.c
1277
result[t][4] = (rtl_get_bbreg(hw,
drivers/net/wireless/realtek/rtlwifi/rtl8192c/phy_common.c
1281
result[t][5] =
drivers/net/wireless/realtek/rtlwifi/rtl8192c/phy_common.c
1284
result[t][6] =
drivers/net/wireless/realtek/rtlwifi/rtl8192c/phy_common.c
1287
result[t][7] =
drivers/net/wireless/realtek/rtlwifi/rtl8192c/phy_common.c
1292
result[t][4] = (rtl_get_bbreg(hw,
drivers/net/wireless/realtek/rtlwifi/rtl8192c/phy_common.c
1297
result[t][5] = (rtl_get_bbreg(hw, 0xebc, MASKDWORD) &
drivers/net/wireless/realtek/rtlwifi/rtl8192c/phy_common.c
1308
if (t != 0) {
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
1200
u8 t, bool is2t)
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
1226
if (t == 0) {
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
1243
if (t == 0)
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
1279
result[t][0] = (rtl_get_bbreg(hw, 0xe94, MASKDWORD) &
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
1281
result[t][1] = (rtl_get_bbreg(hw, 0xe9c, MASKDWORD) &
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
1283
result[t][2] = (rtl_get_bbreg(hw, 0xea4, MASKDWORD) &
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
1285
result[t][3] = (rtl_get_bbreg(hw, 0xeac, MASKDWORD) &
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
1293
result[t][0] = (rtl_get_bbreg(hw, 0xe94, MASKDWORD) &
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
1295
result[t][1] = (rtl_get_bbreg(hw, 0xe9c, MASKDWORD) &
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
1310
result[t][4] = (rtl_get_bbreg(hw, 0xeb4,
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
1312
result[t][5] = (rtl_get_bbreg(hw, 0xebc,
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
1314
result[t][6] = (rtl_get_bbreg(hw, 0xec4,
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
1316
result[t][7] = (rtl_get_bbreg(hw, 0xecc,
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
1323
result[t][4] = (rtl_get_bbreg(hw, 0xeb4,
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
1325
result[t][5] = (rtl_get_bbreg(hw, 0xebc,
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
1339
if (t != 0) {
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
1365
long result[][8], u8 t)
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
1395
if (t == 0) {
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
1419
if (t == 0)
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
1443
result[t][0] = (rtl_get_bbreg(hw, 0xe94, MASKDWORD) &
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
1445
result[t][1] = (rtl_get_bbreg(hw, 0xe9c, MASKDWORD) &
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
1447
result[t][2] = (rtl_get_bbreg(hw, 0xea4, MASKDWORD) &
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
1449
result[t][3] = (rtl_get_bbreg(hw, 0xeac, MASKDWORD) &
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
1455
result[t][0] = (rtl_get_bbreg(hw, 0xe94, MASKDWORD) &
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
1457
result[t][1] = (rtl_get_bbreg(hw, 0xe9c, MASKDWORD) &
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
1470
result[t][4] = (rtl_get_bbreg(hw, 0xeb4, MASKDWORD) &
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
1472
result[t][5] = (rtl_get_bbreg(hw, 0xebc, MASKDWORD) &
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
1474
result[t][6] = (rtl_get_bbreg(hw, 0xec4, MASKDWORD) &
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
1476
result[t][7] = (rtl_get_bbreg(hw, 0xecc, MASKDWORD) &
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
1481
result[t][4] = (rtl_get_bbreg(hw, 0xeb4, MASKDWORD) &
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
1483
result[t][5] = (rtl_get_bbreg(hw, 0xebc, MASKDWORD) &
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
1495
if (t != 0) {
drivers/net/wireless/realtek/rtlwifi/rtl8192du/phy.c
1345
u8 t, bool is2t)
drivers/net/wireless/realtek/rtlwifi/rtl8192du/phy.c
1371
if (t == 0) {
drivers/net/wireless/realtek/rtlwifi/rtl8192du/phy.c
1391
if (t == 0)
drivers/net/wireless/realtek/rtlwifi/rtl8192du/phy.c
1431
result[t][0] = rtl_get_bbreg(hw, RTX_POWER_BEFORE_IQK_A,
drivers/net/wireless/realtek/rtlwifi/rtl8192du/phy.c
1433
result[t][1] = rtl_get_bbreg(hw, RTX_POWER_AFTER_IQK_A,
drivers/net/wireless/realtek/rtlwifi/rtl8192du/phy.c
1435
result[t][2] = rtl_get_bbreg(hw, RRX_POWER_BEFORE_IQK_A_2,
drivers/net/wireless/realtek/rtlwifi/rtl8192du/phy.c
1437
result[t][3] = rtl_get_bbreg(hw, RRX_POWER_AFTER_IQK_A_2,
drivers/net/wireless/realtek/rtlwifi/rtl8192du/phy.c
1445
result[t][0] = rtl_get_bbreg(hw, RTX_POWER_BEFORE_IQK_A,
drivers/net/wireless/realtek/rtlwifi/rtl8192du/phy.c
1447
result[t][1] = rtl_get_bbreg(hw, RTX_POWER_AFTER_IQK_A,
drivers/net/wireless/realtek/rtlwifi/rtl8192du/phy.c
1464
result[t][4] = rtl_get_bbreg(hw, RTX_POWER_BEFORE_IQK_B,
drivers/net/wireless/realtek/rtlwifi/rtl8192du/phy.c
1466
result[t][5] = rtl_get_bbreg(hw, RTX_POWER_AFTER_IQK_B,
drivers/net/wireless/realtek/rtlwifi/rtl8192du/phy.c
1468
result[t][6] = rtl_get_bbreg(hw, RRX_POWER_BEFORE_IQK_B_2,
drivers/net/wireless/realtek/rtlwifi/rtl8192du/phy.c
1470
result[t][7] = rtl_get_bbreg(hw, RRX_POWER_AFTER_IQK_B_2,
drivers/net/wireless/realtek/rtlwifi/rtl8192du/phy.c
1477
result[t][4] = rtl_get_bbreg(hw, RTX_POWER_BEFORE_IQK_B,
drivers/net/wireless/realtek/rtlwifi/rtl8192du/phy.c
1479
result[t][5] = rtl_get_bbreg(hw, RTX_POWER_AFTER_IQK_B,
drivers/net/wireless/realtek/rtlwifi/rtl8192du/phy.c
1494
if (t != 0) {
drivers/net/wireless/realtek/rtlwifi/rtl8192du/phy.c
1525
long result[][8], u8 t)
drivers/net/wireless/realtek/rtlwifi/rtl8192du/phy.c
1559
if (t == 0) {
drivers/net/wireless/realtek/rtlwifi/rtl8192du/phy.c
1584
if (t == 0)
drivers/net/wireless/realtek/rtlwifi/rtl8192du/phy.c
1621
result[t][0] = rtl_get_bbreg(hw, RTX_POWER_BEFORE_IQK_A,
drivers/net/wireless/realtek/rtlwifi/rtl8192du/phy.c
1623
result[t][1] = rtl_get_bbreg(hw, RTX_POWER_AFTER_IQK_A,
drivers/net/wireless/realtek/rtlwifi/rtl8192du/phy.c
1625
result[t][2] = rtl_get_bbreg(hw, RRX_POWER_BEFORE_IQK_A_2,
drivers/net/wireless/realtek/rtlwifi/rtl8192du/phy.c
1627
result[t][3] = rtl_get_bbreg(hw, RRX_POWER_AFTER_IQK_A_2,
drivers/net/wireless/realtek/rtlwifi/rtl8192du/phy.c
1633
result[t][0] = rtl_get_bbreg(hw, RTX_POWER_BEFORE_IQK_A,
drivers/net/wireless/realtek/rtlwifi/rtl8192du/phy.c
1635
result[t][1] = rtl_get_bbreg(hw, RTX_POWER_AFTER_IQK_A,
drivers/net/wireless/realtek/rtlwifi/rtl8192du/phy.c
1656
result[t][4] = rtl_get_bbreg(hw, RTX_POWER_BEFORE_IQK_B,
drivers/net/wireless/realtek/rtlwifi/rtl8192du/phy.c
1658
result[t][5] = rtl_get_bbreg(hw, RTX_POWER_AFTER_IQK_B,
drivers/net/wireless/realtek/rtlwifi/rtl8192du/phy.c
1660
result[t][6] = rtl_get_bbreg(hw, RRX_POWER_BEFORE_IQK_B_2,
drivers/net/wireless/realtek/rtlwifi/rtl8192du/phy.c
1662
result[t][7] = rtl_get_bbreg(hw, RRX_POWER_AFTER_IQK_B_2,
drivers/net/wireless/realtek/rtlwifi/rtl8192du/phy.c
1667
result[t][4] = rtl_get_bbreg(hw, RTX_POWER_BEFORE_IQK_B,
drivers/net/wireless/realtek/rtlwifi/rtl8192du/phy.c
1669
result[t][5] = rtl_get_bbreg(hw, RTX_POWER_AFTER_IQK_B,
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/phy.c
2475
long result[][8], u8 t, bool is2t)
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/phy.c
2500
if (t == 0) {
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/phy.c
2538
result[t][0] = (rtl_get_bbreg(hw,
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/phy.c
2542
result[t][1] = (rtl_get_bbreg(hw, RTX_POWER_AFTER_IQK_A,
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/phy.c
2558
result[t][2] = (rtl_get_bbreg(hw,
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/phy.c
2562
result[t][3] = (rtl_get_bbreg(hw,
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/phy.c
2591
result[t][4] = (rtl_get_bbreg(hw,
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/phy.c
2595
result[t][5] = (rtl_get_bbreg(hw,
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/phy.c
2611
result[t][6] = (rtl_get_bbreg(hw,
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/phy.c
2615
result[t][7] = (rtl_get_bbreg(hw,
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/phy.c
2635
if (t != 0) {
drivers/net/wireless/realtek/rtlwifi/rtl8723ae/phy.c
1131
long result[][8], u8 t, bool is2t)
drivers/net/wireless/realtek/rtlwifi/rtl8723ae/phy.c
1150
if (t == 0) {
drivers/net/wireless/realtek/rtlwifi/rtl8723ae/phy.c
1159
if (t == 0) {
drivers/net/wireless/realtek/rtlwifi/rtl8723ae/phy.c
1167
if (t == 0) {
drivers/net/wireless/realtek/rtlwifi/rtl8723ae/phy.c
1190
result[t][0] = (rtl_get_bbreg(hw, 0xe94, MASKDWORD) &
drivers/net/wireless/realtek/rtlwifi/rtl8723ae/phy.c
1192
result[t][1] = (rtl_get_bbreg(hw, 0xe9c, MASKDWORD) &
drivers/net/wireless/realtek/rtlwifi/rtl8723ae/phy.c
1194
result[t][2] = (rtl_get_bbreg(hw, 0xea4, MASKDWORD) &
drivers/net/wireless/realtek/rtlwifi/rtl8723ae/phy.c
1196
result[t][3] = (rtl_get_bbreg(hw, 0xeac, MASKDWORD) &
drivers/net/wireless/realtek/rtlwifi/rtl8723ae/phy.c
1201
result[t][0] = (rtl_get_bbreg(hw, 0xe94,
drivers/net/wireless/realtek/rtlwifi/rtl8723ae/phy.c
1204
result[t][1] =
drivers/net/wireless/realtek/rtlwifi/rtl8723ae/phy.c
1215
result[t][4] = (rtl_get_bbreg(hw,
drivers/net/wireless/realtek/rtlwifi/rtl8723ae/phy.c
1219
result[t][5] =
drivers/net/wireless/realtek/rtlwifi/rtl8723ae/phy.c
1222
result[t][6] =
drivers/net/wireless/realtek/rtlwifi/rtl8723ae/phy.c
1225
result[t][7] =
drivers/net/wireless/realtek/rtlwifi/rtl8723ae/phy.c
1230
result[t][4] = (rtl_get_bbreg(hw,
drivers/net/wireless/realtek/rtlwifi/rtl8723ae/phy.c
1235
result[t][5] = (rtl_get_bbreg(hw, 0xebc, MASKDWORD) &
drivers/net/wireless/realtek/rtlwifi/rtl8723ae/phy.c
1246
if (t != 0) {
drivers/net/wireless/realtek/rtlwifi/rtl8723be/phy.c
2006
long result[][8], u8 t, bool is2t)
drivers/net/wireless/realtek/rtlwifi/rtl8723be/phy.c
2037
if (t == 0) {
drivers/net/wireless/realtek/rtlwifi/rtl8723be/phy.c
2047
if (t == 0) {
drivers/net/wireless/realtek/rtlwifi/rtl8723be/phy.c
2069
result[t][0] = (rtl_get_bbreg(hw, 0xe94, MASKDWORD) &
drivers/net/wireless/realtek/rtlwifi/rtl8723be/phy.c
2071
result[t][1] = (rtl_get_bbreg(hw, 0xe9c, MASKDWORD) &
drivers/net/wireless/realtek/rtlwifi/rtl8723be/phy.c
2085
result[t][2] = (rtl_get_bbreg(hw, 0xea4, MASKDWORD) &
drivers/net/wireless/realtek/rtlwifi/rtl8723be/phy.c
2087
result[t][3] = (rtl_get_bbreg(hw, 0xeac, MASKDWORD) &
drivers/net/wireless/realtek/rtlwifi/rtl8723be/phy.c
2105
result[t][4] = (rtl_get_bbreg(hw, 0xe94,
drivers/net/wireless/realtek/rtlwifi/rtl8723be/phy.c
2108
result[t][5] = (rtl_get_bbreg(hw, 0xe9c,
drivers/net/wireless/realtek/rtlwifi/rtl8723be/phy.c
2122
result[t][6] = (rtl_get_bbreg(hw, 0xea4,
drivers/net/wireless/realtek/rtlwifi/rtl8723be/phy.c
2125
result[t][7] = (rtl_get_bbreg(hw, 0xeac,
drivers/net/wireless/realtek/rtlwifi/rtl8723be/phy.c
2138
if (t != 0) {
drivers/net/wireless/realtek/rtlwifi/usb.c
211
static void _rtl_rx_work(struct tasklet_struct *t);
drivers/net/wireless/realtek/rtlwifi/usb.c
480
static void _rtl_rx_work(struct tasklet_struct *t)
drivers/net/wireless/realtek/rtlwifi/usb.c
482
struct rtl_usb *rtlusb = from_tasklet(rtlusb, t, rx_work_tasklet);
drivers/net/wireless/realtek/rtlwifi/wifi.h
2826
#define MSECS(t) msecs_to_jiffies(t)
drivers/net/wireless/realtek/rtw88/main.h
2242
void rtw_tx_report_purge_timer(struct timer_list *t);
drivers/net/wireless/realtek/rtw88/rtw8703b.c
1002
dur = ktime_us_delta(ktime_get(), t);
drivers/net/wireless/realtek/rtw88/rtw8703b.c
1213
void rtw8703b_iqk_one_round(struct rtw_dev *rtwdev, s32 result[][IQK_NR], u8 t,
drivers/net/wireless/realtek/rtw88/rtw8703b.c
1220
"[IQK] IQ Calibration for 1T1R_S0/S1 for %d times\n", t);
drivers/net/wireless/realtek/rtw88/rtw8703b.c
1234
result[t][IQK_S1_TX_X] =
drivers/net/wireless/realtek/rtw88/rtw8703b.c
1237
result[t][IQK_S1_TX_Y] =
drivers/net/wireless/realtek/rtw88/rtw8703b.c
1244
result[t][IQK_S1_TX_X] = 0x100;
drivers/net/wireless/realtek/rtw88/rtw8703b.c
1245
result[t][IQK_S1_TX_Y] = 0x0;
drivers/net/wireless/realtek/rtw88/rtw8703b.c
1253
result[t][IQK_S1_RX_X] =
drivers/net/wireless/realtek/rtw88/rtw8703b.c
1256
result[t][IQK_S1_RX_Y] =
drivers/net/wireless/realtek/rtw88/rtw8703b.c
1263
result[t][IQK_S1_RX_X] = 0x100;
drivers/net/wireless/realtek/rtw88/rtw8703b.c
1264
result[t][IQK_S1_RX_Y] = 0x0;
drivers/net/wireless/realtek/rtw88/rtw8703b.c
985
ktime_t t;
drivers/net/wireless/realtek/rtw88/rtw8703b.c
997
t = ktime_get();
drivers/net/wireless/realtek/rtw88/rtw8723d.c
824
void rtw8723d_iqk_one_round(struct rtw_dev *rtwdev, s32 result[][IQK_NR], u8 t,
drivers/net/wireless/realtek/rtw88/rtw8723d.c
831
"[IQK] IQ Calibration for 1T1R_S0/S1 for %d times\n", t);
drivers/net/wireless/realtek/rtw88/rtw8723d.c
846
result[t][IQK_S1_TX_X] =
drivers/net/wireless/realtek/rtw88/rtw8723d.c
848
result[t][IQK_S1_TX_Y] =
drivers/net/wireless/realtek/rtw88/rtw8723d.c
854
result[t][IQK_S1_TX_X] = 0x100;
drivers/net/wireless/realtek/rtw88/rtw8723d.c
855
result[t][IQK_S1_TX_Y] = 0x0;
drivers/net/wireless/realtek/rtw88/rtw8723d.c
863
result[t][IQK_S1_RX_X] =
drivers/net/wireless/realtek/rtw88/rtw8723d.c
865
result[t][IQK_S1_RX_Y] =
drivers/net/wireless/realtek/rtw88/rtw8723d.c
871
result[t][IQK_S1_RX_X] = 0x100;
drivers/net/wireless/realtek/rtw88/rtw8723d.c
872
result[t][IQK_S1_RX_Y] = 0x0;
drivers/net/wireless/realtek/rtw88/rtw8723d.c
885
result[t][IQK_S0_TX_X] =
drivers/net/wireless/realtek/rtw88/rtw8723d.c
887
result[t][IQK_S0_TX_Y] =
drivers/net/wireless/realtek/rtw88/rtw8723d.c
893
result[t][IQK_S0_TX_X] = 0x100;
drivers/net/wireless/realtek/rtw88/rtw8723d.c
894
result[t][IQK_S0_TX_Y] = 0x0;
drivers/net/wireless/realtek/rtw88/rtw8723d.c
903
result[t][IQK_S0_RX_X] =
drivers/net/wireless/realtek/rtw88/rtw8723d.c
905
result[t][IQK_S0_RX_Y] =
drivers/net/wireless/realtek/rtw88/rtw8723d.c
911
result[t][IQK_S0_RX_X] = 0x100;
drivers/net/wireless/realtek/rtw88/rtw8723d.c
912
result[t][IQK_S0_RX_Y] = 0x0;
drivers/net/wireless/realtek/rtw88/rtw8723x.h
355
s32 t;
drivers/net/wireless/realtek/rtw88/rtw8723x.h
357
t = x * y;
drivers/net/wireless/realtek/rtw88/rtw8723x.h
359
*ext = (t >> 7) & 0x1; /* Q.16 --> Q.9; get LSB of Q.9 */
drivers/net/wireless/realtek/rtw88/rtw8723x.h
361
return (t >> 8); /* Q.16 --> Q.8 */
drivers/net/wireless/realtek/rtw88/rtw8822c.c
212
u32 p, m, t, i;
drivers/net/wireless/realtek/rtw88/rtw8822c.c
224
t = p - m;
drivers/net/wireless/realtek/rtw88/rtw8822c.c
225
t = t / (DACK_SN_8822C - 20);
drivers/net/wireless/realtek/rtw88/rtw8822c.c
227
t = m - p;
drivers/net/wireless/realtek/rtw88/rtw8822c.c
228
t = t / (DACK_SN_8822C - 20);
drivers/net/wireless/realtek/rtw88/rtw8822c.c
229
if (t != 0x0)
drivers/net/wireless/realtek/rtw88/rtw8822c.c
230
t = 0x400 - t;
drivers/net/wireless/realtek/rtw88/rtw8822c.c
233
*val = t;
drivers/net/wireless/realtek/rtw88/tx.c
179
void rtw_tx_report_purge_timer(struct timer_list *t)
drivers/net/wireless/realtek/rtw88/tx.c
181
struct rtw_dev *rtwdev = timer_container_of(rtwdev, t,
drivers/net/wireless/realtek/rtw89/coex.c
3686
struct rtw89_btc_fbtc_tdma *t = &dm->tdma;
drivers/net/wireless/realtek/rtw89/coex.c
3709
*t = t_def[CXTD_OFF];
drivers/net/wireless/realtek/rtw89/coex.c
3716
*t = t_def[CXTD_OFF];
drivers/net/wireless/realtek/rtw89/coex.c
3751
*t = t_def[CXTD_OFF_B2];
drivers/net/wireless/realtek/rtw89/coex.c
3762
*t = t_def[CXTD_OFF_EXT];
drivers/net/wireless/realtek/rtw89/coex.c
3780
*t = t_def[CXTD_FIX];
drivers/net/wireless/realtek/rtw89/coex.c
3836
*t = t_def[CXTD_PFIX];
drivers/net/wireless/realtek/rtw89/coex.c
3869
*t = t_def[CXTD_AUTO];
drivers/net/wireless/realtek/rtw89/coex.c
3893
*t = t_def[CXTD_PAUTO];
drivers/net/wireless/realtek/rtw89/coex.c
3917
*t = t_def[CXTD_AUTO2];
drivers/net/wireless/realtek/rtw89/coex.c
3954
*t = t_def[CXTD_PAUTO2];
drivers/net/wireless/realtek/rtw89/coex.c
3997
struct rtw89_btc_fbtc_tdma *t = &dm->tdma;
drivers/net/wireless/realtek/rtw89/coex.c
4053
*t = t_def[CXTD_OFF];
drivers/net/wireless/realtek/rtw89/coex.c
4060
*t = t_def[CXTD_OFF];
drivers/net/wireless/realtek/rtw89/coex.c
4112
*t = t_def[CXTD_OFF_B2];
drivers/net/wireless/realtek/rtw89/coex.c
4126
*t = t_def[CXTD_OFF_EXT];
drivers/net/wireless/realtek/rtw89/coex.c
4191
*t = t_def[CXTD_FIX];
drivers/net/wireless/realtek/rtw89/coex.c
4250
*t = t_def[CXTD_PFIX];
drivers/net/wireless/realtek/rtw89/coex.c
4289
*t = t_def[CXTD_AUTO];
drivers/net/wireless/realtek/rtw89/coex.c
4316
*t = t_def[CXTD_PAUTO];
drivers/net/wireless/realtek/rtw89/coex.c
4343
*t = t_def[CXTD_AUTO2];
drivers/net/wireless/realtek/rtw89/coex.c
4385
*t = t_def[CXTD_PAUTO2];
drivers/net/wireless/realtek/rtw89/coex.c
8898
struct rtw89_traffic_stats *t;
drivers/net/wireless/realtek/rtw89/coex.c
8932
t = &plink->stat.traffic;
drivers/net/wireless/realtek/rtw89/coex.c
8936
t->tx_rate, t->tx_tfc_lv);
drivers/net/wireless/realtek/rtw89/coex.c
8940
t->rx_rate,
drivers/net/wireless/realtek/rtw89/coex.c
8941
t->rx_tfc_lv, plink->rx_rate_drop_cnt);
drivers/net/wireless/realtek/rtw89/coex.c
9735
struct rtw89_btc_fbtc_tdma *t = NULL;
drivers/net/wireless/realtek/rtw89/coex.c
9743
t = &pfwinfo->rpt_fbtc_tdma.finfo.v1;
drivers/net/wireless/realtek/rtw89/coex.c
9745
t = &pfwinfo->rpt_fbtc_tdma.finfo.v3.tdma;
drivers/net/wireless/realtek/rtw89/coex.c
9751
(u32)t->type,
drivers/net/wireless/realtek/rtw89/coex.c
9752
t->rxflctrl, t->txpause);
drivers/net/wireless/realtek/rtw89/coex.c
9756
t->wtgle_n, t->leak_n, t->ext_ctrl);
drivers/net/wireless/realtek/rtw89/core.c
2077
u32 t;
drivers/net/wireless/realtek/rtw89/core.c
2104
t = le32_get_bits(ie->w1, RTW89_PHY_STS_IE01_W1_FD_CFO);
drivers/net/wireless/realtek/rtw89/core.c
2105
cfo = sign_extend32(t, 11);
drivers/net/wireless/realtek/rtw89/core.c
2107
t = le32_get_bits(ie->w1, RTW89_PHY_STS_IE01_W1_PREMB_CFO);
drivers/net/wireless/realtek/rtw89/core.c
2108
cfo = sign_extend32(t, 11);
drivers/net/wireless/realtek/rtw89/mac.c
6643
u8 sound_dim = 3, t;
drivers/net/wireless/realtek/rtw89/mac.c
6662
t = FIELD_GET(IEEE80211_HE_PHY_CAP5_BEAMFORMEE_NUM_SND_DIM_UNDER_80MHZ_MASK,
drivers/net/wireless/realtek/rtw89/mac.c
6664
sound_dim = min(sound_dim, t);
drivers/net/wireless/realtek/rtw89/mac.c
6670
t = FIELD_GET(IEEE80211_VHT_CAP_SOUNDING_DIMENSIONS_MASK,
drivers/net/wireless/realtek/rtw89/mac.c
6672
sound_dim = min(sound_dim, t);
drivers/net/wireless/realtek/rtw89/mac_be.c
2720
u8 sound_dim = 3, t;
drivers/net/wireless/realtek/rtw89/mac_be.c
2739
t = u8_get_bits(phy_cap[5],
drivers/net/wireless/realtek/rtw89/mac_be.c
2741
sound_dim = min(sound_dim, t);
drivers/net/wireless/realtek/rtw89/mac_be.c
2748
t = u32_get_bits(link_sta->vht_cap.cap,
drivers/net/wireless/realtek/rtw89/mac_be.c
2750
sound_dim = min(sound_dim, t);
drivers/net/wireless/realtek/rtw89/phy.c
3201
u8 t;
drivers/net/wireless/realtek/rtw89/phy.c
3213
t = le32_get_bits(c2h->w2, RTW89_C2H_RA_RPT_W2_MCSNSS_B7);
drivers/net/wireless/realtek/rtw89/phy.c
3214
rate |= u8_encode_bits(t, BIT(7));
drivers/net/wireless/realtek/rtw89/phy.c
3215
t = le32_get_bits(c2h->w3, RTW89_C2H_RA_RPT_W3_BW_B2);
drivers/net/wireless/realtek/rtw89/phy.c
3216
bw |= u8_encode_bits(t, BIT(2));
drivers/net/wireless/realtek/rtw89/phy.c
3217
t = le32_get_bits(c2h->w3, RTW89_C2H_RA_RPT_W3_MD_SEL_B2);
drivers/net/wireless/realtek/rtw89/phy.c
3218
mode |= u8_encode_bits(t, BIT(2));
drivers/net/wireless/realtek/rtw89/phy_be.c
485
u32 addr, t;
drivers/net/wireless/realtek/rtw89/phy_be.c
491
t = rtw89_mac_reg_by_idx(rtwdev, addr, mac_idx);
drivers/net/wireless/realtek/rtw89/phy_be.c
492
rtw89_write32(rtwdev, t, 0);
drivers/net/wireless/realtek/rtw89/phy_be.c
495
t = rtw89_mac_reg_by_idx(rtwdev, addr, mac_idx);
drivers/net/wireless/realtek/rtw89/phy_be.c
496
rtw89_write32(rtwdev, t, 0);
drivers/net/wireless/realtek/rtw89/phy_be.c
499
t = rtw89_mac_reg_by_idx(rtwdev, addr, mac_idx);
drivers/net/wireless/realtek/rtw89/phy_be.c
500
rtw89_write32(rtwdev, t, 0);
drivers/net/wireless/realtek/rtw89/rtw8851b_rfk.c
519
u8 t;
drivers/net/wireless/realtek/rtw89/rtw8851b_rfk.c
530
t = dack->msbk_d[0][0][i];
drivers/net/wireless/realtek/rtw89/rtw8851b_rfk.c
531
rtw89_debug(rtwdev, RTW89_DBG_RFK, "[DACK]0x%x\n", t);
drivers/net/wireless/realtek/rtw89/rtw8851b_rfk.c
536
t = dack->msbk_d[0][1][i];
drivers/net/wireless/realtek/rtw89/rtw8851b_rfk.c
537
rtw89_debug(rtwdev, RTW89_DBG_RFK, "[DACK]0x%x\n", t);
drivers/net/wireless/realtek/rtw89/rtw8852a_rfk.c
115
u8 t;
drivers/net/wireless/realtek/rtw89/rtw8852a_rfk.c
139
t = dack->msbk_d[0][0][i];
drivers/net/wireless/realtek/rtw89/rtw8852a_rfk.c
140
rtw89_debug(rtwdev, RTW89_DBG_RFK, "[DACK]0x%x\n", t);
drivers/net/wireless/realtek/rtw89/rtw8852a_rfk.c
144
t = dack->msbk_d[0][1][i];
drivers/net/wireless/realtek/rtw89/rtw8852a_rfk.c
145
rtw89_debug(rtwdev, RTW89_DBG_RFK, "[DACK]0x%x\n", t);
drivers/net/wireless/realtek/rtw89/rtw8852a_rfk.c
149
t = dack->msbk_d[1][0][i];
drivers/net/wireless/realtek/rtw89/rtw8852a_rfk.c
150
rtw89_debug(rtwdev, RTW89_DBG_RFK, "[DACK]0x%x\n", t);
drivers/net/wireless/realtek/rtw89/rtw8852a_rfk.c
154
t = dack->msbk_d[1][1][i];
drivers/net/wireless/realtek/rtw89/rtw8852a_rfk.c
155
rtw89_debug(rtwdev, RTW89_DBG_RFK, "[DACK]0x%x\n", t);
drivers/net/wireless/realtek/rtw89/rtw8852b_rfk.c
710
u8 t;
drivers/net/wireless/realtek/rtw89/rtw8852b_rfk.c
733
t = dack->msbk_d[0][0][i];
drivers/net/wireless/realtek/rtw89/rtw8852b_rfk.c
734
rtw89_debug(rtwdev, RTW89_DBG_RFK, "[DACK]0x%x\n", t);
drivers/net/wireless/realtek/rtw89/rtw8852b_rfk.c
739
t = dack->msbk_d[0][1][i];
drivers/net/wireless/realtek/rtw89/rtw8852b_rfk.c
740
rtw89_debug(rtwdev, RTW89_DBG_RFK, "[DACK]0x%x\n", t);
drivers/net/wireless/realtek/rtw89/rtw8852b_rfk.c
745
t = dack->msbk_d[1][0][i];
drivers/net/wireless/realtek/rtw89/rtw8852b_rfk.c
746
rtw89_debug(rtwdev, RTW89_DBG_RFK, "[DACK]0x%x\n", t);
drivers/net/wireless/realtek/rtw89/rtw8852b_rfk.c
751
t = dack->msbk_d[1][1][i];
drivers/net/wireless/realtek/rtw89/rtw8852b_rfk.c
752
rtw89_debug(rtwdev, RTW89_DBG_RFK, "[DACK]0x%x\n", t);
drivers/net/wireless/realtek/rtw89/rtw8852bt_rfk.c
732
u8 t;
drivers/net/wireless/realtek/rtw89/rtw8852bt_rfk.c
755
t = dack->msbk_d[0][0][i];
drivers/net/wireless/realtek/rtw89/rtw8852bt_rfk.c
756
rtw89_debug(rtwdev, RTW89_DBG_RFK, "[DACK]0x%x\n", t);
drivers/net/wireless/realtek/rtw89/rtw8852bt_rfk.c
761
t = dack->msbk_d[0][1][i];
drivers/net/wireless/realtek/rtw89/rtw8852bt_rfk.c
762
rtw89_debug(rtwdev, RTW89_DBG_RFK, "[DACK]0x%x\n", t);
drivers/net/wireless/realtek/rtw89/rtw8852bt_rfk.c
767
t = dack->msbk_d[1][0][i];
drivers/net/wireless/realtek/rtw89/rtw8852bt_rfk.c
768
rtw89_debug(rtwdev, RTW89_DBG_RFK, "[DACK]0x%x\n", t);
drivers/net/wireless/realtek/rtw89/rtw8852bt_rfk.c
773
t = dack->msbk_d[1][1][i];
drivers/net/wireless/realtek/rtw89/rtw8852bt_rfk.c
774
rtw89_debug(rtwdev, RTW89_DBG_RFK, "[DACK]0x%x\n", t);
drivers/net/wireless/realtek/rtw89/rtw8852c_rfk.c
174
u8 t;
drivers/net/wireless/realtek/rtw89/rtw8852c_rfk.c
198
t = dack->msbk_d[0][0][i];
drivers/net/wireless/realtek/rtw89/rtw8852c_rfk.c
199
rtw89_debug(rtwdev, RTW89_DBG_RFK, "[DACK]0x%x\n", t);
drivers/net/wireless/realtek/rtw89/rtw8852c_rfk.c
203
t = dack->msbk_d[0][1][i];
drivers/net/wireless/realtek/rtw89/rtw8852c_rfk.c
204
rtw89_debug(rtwdev, RTW89_DBG_RFK, "[DACK]0x%x\n", t);
drivers/net/wireless/realtek/rtw89/rtw8852c_rfk.c
208
t = dack->msbk_d[1][0][i];
drivers/net/wireless/realtek/rtw89/rtw8852c_rfk.c
209
rtw89_debug(rtwdev, RTW89_DBG_RFK, "[DACK]0x%x\n", t);
drivers/net/wireless/realtek/rtw89/rtw8852c_rfk.c
213
t = dack->msbk_d[1][1][i];
drivers/net/wireless/realtek/rtw89/rtw8852c_rfk.c
214
rtw89_debug(rtwdev, RTW89_DBG_RFK, "[DACK]0x%x\n", t);
drivers/net/wireless/realtek/rtw89/rtw8922a.c
542
u8 t;
drivers/net/wireless/realtek/rtw89/rtw8922a.c
573
t = gain->offset[i][j];
drivers/net/wireless/realtek/rtw89/rtw8922a.c
574
if (t != 0xff)
drivers/net/wireless/realtek/rtw89/rtw8922a.c
576
if (t != 0x0)
drivers/net/wireless/realtek/rtw89/rtw8922a.c
580
if (t & 0x80)
drivers/net/wireless/realtek/rtw89/rtw8922a.c
581
gain->offset[i][j] = (t ^ 0x7f) + 1;
drivers/net/wireless/rsi/rsi_91x_hal.c
491
static void bl_cmd_timeout(struct timer_list *t)
drivers/net/wireless/rsi/rsi_91x_hal.c
493
struct rsi_hw *adapter = timer_container_of(adapter, t, bl_cmd_timer);
drivers/net/wireless/rsi/rsi_91x_mac80211.c
1754
void rsi_roc_timeout(struct timer_list *t)
drivers/net/wireless/rsi/rsi_91x_mac80211.c
1756
struct rsi_common *common = timer_container_of(common, t, roc_timer);
drivers/net/wireless/rsi/rsi_common.h
91
void rsi_roc_timeout(struct timer_list *t);
drivers/net/wireless/st/cw1200/queue.c
132
static void cw1200_queue_gc(struct timer_list *t)
drivers/net/wireless/st/cw1200/queue.c
136
timer_container_of(queue, t, gc);
drivers/net/wireless/st/cw1200/sta.c
2114
void cw1200_mcast_timeout(struct timer_list *t)
drivers/net/wireless/st/cw1200/sta.c
2116
struct cw1200_common *priv = timer_container_of(priv, t,
drivers/net/wireless/st/cw1200/sta.h
119
void cw1200_mcast_timeout(struct timer_list *t);
drivers/net/wireless/st/cw1200/txrx.c
438
struct cw1200_txinfo *t)
drivers/net/wireless/st/cw1200/txrx.c
440
if (t->sta && t->sta_priv->link_id)
drivers/net/wireless/st/cw1200/txrx.c
441
t->txpriv.raw_link_id =
drivers/net/wireless/st/cw1200/txrx.c
442
t->txpriv.link_id =
drivers/net/wireless/st/cw1200/txrx.c
443
t->sta_priv->link_id;
drivers/net/wireless/st/cw1200/txrx.c
445
t->txpriv.raw_link_id =
drivers/net/wireless/st/cw1200/txrx.c
446
t->txpriv.link_id = 0;
drivers/net/wireless/st/cw1200/txrx.c
447
else if (is_multicast_ether_addr(t->da)) {
drivers/net/wireless/st/cw1200/txrx.c
449
t->txpriv.raw_link_id = 0;
drivers/net/wireless/st/cw1200/txrx.c
450
t->txpriv.link_id = CW1200_LINK_ID_AFTER_DTIM;
drivers/net/wireless/st/cw1200/txrx.c
452
t->txpriv.raw_link_id = 0;
drivers/net/wireless/st/cw1200/txrx.c
453
t->txpriv.link_id = 0;
drivers/net/wireless/st/cw1200/txrx.c
456
t->txpriv.link_id = cw1200_find_link_id(priv, t->da);
drivers/net/wireless/st/cw1200/txrx.c
457
if (!t->txpriv.link_id)
drivers/net/wireless/st/cw1200/txrx.c
458
t->txpriv.link_id = cw1200_alloc_link_id(priv, t->da);
drivers/net/wireless/st/cw1200/txrx.c
459
if (!t->txpriv.link_id) {
drivers/net/wireless/st/cw1200/txrx.c
464
t->txpriv.raw_link_id = t->txpriv.link_id;
drivers/net/wireless/st/cw1200/txrx.c
466
if (t->txpriv.raw_link_id)
drivers/net/wireless/st/cw1200/txrx.c
467
priv->link_id_db[t->txpriv.raw_link_id - 1].timestamp =
drivers/net/wireless/st/cw1200/txrx.c
469
if (t->sta && (t->sta->uapsd_queues & BIT(t->queue)))
drivers/net/wireless/st/cw1200/txrx.c
470
t->txpriv.link_id = CW1200_LINK_ID_UAPSD;
drivers/net/wireless/st/cw1200/txrx.c
476
struct cw1200_txinfo *t)
drivers/net/wireless/st/cw1200/txrx.c
478
if (ieee80211_is_auth(t->hdr->frame_control)) {
drivers/net/wireless/st/cw1200/txrx.c
479
u32 mask = ~BIT(t->txpriv.raw_link_id);
drivers/net/wireless/st/cw1200/txrx.c
489
struct cw1200_txinfo *t)
drivers/net/wireless/st/cw1200/txrx.c
491
if (ieee80211_is_data_qos(t->hdr->frame_control)) {
drivers/net/wireless/st/cw1200/txrx.c
492
u8 *qos = ieee80211_get_qos_ctl(t->hdr);
drivers/net/wireless/st/cw1200/txrx.c
493
t->txpriv.tid = qos[0] & IEEE80211_QOS_CTL_TID_MASK;
drivers/net/wireless/st/cw1200/txrx.c
494
} else if (ieee80211_is_data(t->hdr->frame_control)) {
drivers/net/wireless/st/cw1200/txrx.c
495
t->txpriv.tid = 0;
drivers/net/wireless/st/cw1200/txrx.c
501
struct cw1200_txinfo *t)
drivers/net/wireless/st/cw1200/txrx.c
503
if (!t->tx_info->control.hw_key ||
drivers/net/wireless/st/cw1200/txrx.c
504
!ieee80211_has_protected(t->hdr->frame_control))
drivers/net/wireless/st/cw1200/txrx.c
507
t->hdrlen += t->tx_info->control.hw_key->iv_len;
drivers/net/wireless/st/cw1200/txrx.c
508
skb_put(t->skb, t->tx_info->control.hw_key->icv_len);
drivers/net/wireless/st/cw1200/txrx.c
510
if (t->tx_info->control.hw_key->cipher == WLAN_CIPHER_SUITE_TKIP)
drivers/net/wireless/st/cw1200/txrx.c
511
skb_put(t->skb, 8); /* MIC space */
drivers/net/wireless/st/cw1200/txrx.c
518
struct cw1200_txinfo *t,
drivers/net/wireless/st/cw1200/txrx.c
521
size_t offset = (size_t)t->skb->data & 3;
drivers/net/wireless/st/cw1200/txrx.c
533
if (skb_headroom(t->skb) < offset) {
drivers/net/wireless/st/cw1200/txrx.c
536
skb_headroom(t->skb));
drivers/net/wireless/st/cw1200/txrx.c
539
skb_push(t->skb, offset);
drivers/net/wireless/st/cw1200/txrx.c
540
t->hdrlen += offset;
drivers/net/wireless/st/cw1200/txrx.c
541
t->txpriv.offset += offset;
drivers/net/wireless/st/cw1200/txrx.c
549
struct cw1200_txinfo *t)
drivers/net/wireless/st/cw1200/txrx.c
552
(struct ieee80211_mgmt *)t->hdr;
drivers/net/wireless/st/cw1200/txrx.c
553
if (ieee80211_is_action(t->hdr->frame_control) &&
drivers/net/wireless/st/cw1200/txrx.c
563
struct cw1200_txinfo *t)
drivers/net/wireless/st/cw1200/txrx.c
567
if (skb_headroom(t->skb) < sizeof(struct wsm_tx)) {
drivers/net/wireless/st/cw1200/txrx.c
570
skb_headroom(t->skb));
drivers/net/wireless/st/cw1200/txrx.c
574
wsm = skb_push(t->skb, sizeof(struct wsm_tx));
drivers/net/wireless/st/cw1200/txrx.c
575
t->txpriv.offset += sizeof(struct wsm_tx);
drivers/net/wireless/st/cw1200/txrx.c
577
wsm->hdr.len = __cpu_to_le16(t->skb->len);
drivers/net/wireless/st/cw1200/txrx.c
579
wsm->queue_id = wsm_queue_id_to_wsm(t->queue);
drivers/net/wireless/st/cw1200/txrx.c
586
struct cw1200_txinfo *t,
drivers/net/wireless/st/cw1200/txrx.c
594
if (ieee80211_is_nullfunc(t->hdr->frame_control)) {
drivers/net/wireless/st/cw1200/txrx.c
596
} else if (ieee80211_is_data(t->hdr->frame_control)) {
drivers/net/wireless/st/cw1200/txrx.c
598
u8 *payload = &t->skb->data[t->hdrlen];
drivers/net/wireless/st/cw1200/txrx.c
602
} else if (ieee80211_is_assoc_req(t->hdr->frame_control) ||
drivers/net/wireless/st/cw1200/txrx.c
603
ieee80211_is_reassoc_req(t->hdr->frame_control)) {
drivers/net/wireless/st/cw1200/txrx.c
605
(struct ieee80211_mgmt *)t->hdr;
drivers/net/wireless/st/cw1200/txrx.c
620
if (ieee80211_is_action(t->hdr->frame_control))
drivers/net/wireless/st/cw1200/txrx.c
622
else if (ieee80211_is_mgmt(t->hdr->frame_control))
drivers/net/wireless/st/cw1200/txrx.c
639
struct cw1200_txinfo *t,
drivers/net/wireless/st/cw1200/txrx.c
644
t->txpriv.rate_id = tx_policy_get(priv,
drivers/net/wireless/st/cw1200/txrx.c
645
t->tx_info->control.rates, IEEE80211_TX_MAX_RATES,
drivers/net/wireless/st/cw1200/txrx.c
647
if (t->txpriv.rate_id == CW1200_INVALID_RATE_ID)
drivers/net/wireless/st/cw1200/txrx.c
650
wsm->flags |= t->txpriv.rate_id << 4;
drivers/net/wireless/st/cw1200/txrx.c
652
t->rate = cw1200_get_tx_rate(priv,
drivers/net/wireless/st/cw1200/txrx.c
653
&t->tx_info->control.rates[0]);
drivers/net/wireless/st/cw1200/txrx.c
654
wsm->max_tx_rate = t->rate->hw_value;
drivers/net/wireless/st/cw1200/txrx.c
655
if (t->rate->flags & IEEE80211_TX_RC_MCS) {
drivers/net/wireless/st/cw1200/txrx.c
683
struct cw1200_txinfo *t)
drivers/net/wireless/st/cw1200/txrx.c
687
if (t->txpriv.link_id == CW1200_LINK_ID_AFTER_DTIM &&
drivers/net/wireless/st/cw1200/txrx.c
695
if (t->txpriv.raw_link_id && t->txpriv.tid < CW1200_MAX_TID)
drivers/net/wireless/st/cw1200/txrx.c
696
was_buffered = priv->link_id_db[t->txpriv.raw_link_id - 1].buffered[t->txpriv.tid]++;
drivers/net/wireless/st/cw1200/txrx.c
708
struct cw1200_txinfo t = {
drivers/net/wireless/st/cw1200/txrx.c
725
t.hdrlen = ieee80211_hdrlen(t.hdr->frame_control);
drivers/net/wireless/st/cw1200/txrx.c
726
t.da = ieee80211_get_DA(t.hdr);
drivers/net/wireless/st/cw1200/txrx.c
728
t.sta = control->sta;
drivers/net/wireless/st/cw1200/txrx.c
729
t.sta_priv = (struct cw1200_sta_priv *)&t.sta->drv_priv;
drivers/net/wireless/st/cw1200/txrx.c
732
if (WARN_ON(t.queue >= 4))
drivers/net/wireless/st/cw1200/txrx.c
735
ret = cw1200_tx_h_calc_link_ids(priv, &t);
drivers/net/wireless/st/cw1200/txrx.c
740
skb->len, t.queue, t.txpriv.link_id,
drivers/net/wireless/st/cw1200/txrx.c
741
t.txpriv.raw_link_id);
drivers/net/wireless/st/cw1200/txrx.c
743
cw1200_tx_h_pm(priv, &t);
drivers/net/wireless/st/cw1200/txrx.c
744
cw1200_tx_h_calc_tid(priv, &t);
drivers/net/wireless/st/cw1200/txrx.c
745
ret = cw1200_tx_h_crypt(priv, &t);
drivers/net/wireless/st/cw1200/txrx.c
748
ret = cw1200_tx_h_align(priv, &t, &flags);
drivers/net/wireless/st/cw1200/txrx.c
751
ret = cw1200_tx_h_action(priv, &t);
drivers/net/wireless/st/cw1200/txrx.c
754
wsm = cw1200_tx_h_wsm(priv, &t);
drivers/net/wireless/st/cw1200/txrx.c
760
cw1200_tx_h_bt(priv, &t, wsm);
drivers/net/wireless/st/cw1200/txrx.c
761
ret = cw1200_tx_h_rate_policy(priv, &t, wsm);
drivers/net/wireless/st/cw1200/txrx.c
765
sta = t.sta;
drivers/net/wireless/st/cw1200/txrx.c
769
tid_update = cw1200_tx_h_pm_state(priv, &t);
drivers/net/wireless/st/cw1200/txrx.c
770
BUG_ON(cw1200_queue_put(&priv->tx_queue[t.queue],
drivers/net/wireless/st/cw1200/txrx.c
771
t.skb, &t.txpriv));
drivers/net/wireless/st/cw1200/txrx.c
776
ieee80211_sta_set_buffered(sta, t.txpriv.tid, true);
drivers/net/wireless/st/cw1200/txrx.c
783
cw1200_skb_dtor(priv, skb, &t.txpriv);
drivers/net/wireless/ti/wl1251/spi.c
113
t.tx_buf = cmd;
drivers/net/wireless/ti/wl1251/spi.c
114
t.len = WSPI_INIT_CMD_LEN;
drivers/net/wireless/ti/wl1251/spi.c
115
spi_message_add_tail(&t, &m);
drivers/net/wireless/ti/wl1251/spi.c
134
struct spi_transfer t[3];
drivers/net/wireless/ti/wl1251/spi.c
148
memset(t, 0, sizeof(t));
drivers/net/wireless/ti/wl1251/spi.c
150
t[0].tx_buf = cmd;
drivers/net/wireless/ti/wl1251/spi.c
151
t[0].len = 4;
drivers/net/wireless/ti/wl1251/spi.c
152
spi_message_add_tail(&t[0], &m);
drivers/net/wireless/ti/wl1251/spi.c
155
t[1].rx_buf = busy_buf;
drivers/net/wireless/ti/wl1251/spi.c
156
t[1].len = WL1251_BUSY_WORD_LEN;
drivers/net/wireless/ti/wl1251/spi.c
157
spi_message_add_tail(&t[1], &m);
drivers/net/wireless/ti/wl1251/spi.c
159
t[2].rx_buf = buf;
drivers/net/wireless/ti/wl1251/spi.c
160
t[2].len = len;
drivers/net/wireless/ti/wl1251/spi.c
161
spi_message_add_tail(&t[2], &m);
drivers/net/wireless/ti/wl1251/spi.c
175
struct spi_transfer t[2];
drivers/net/wireless/ti/wl1251/spi.c
187
memset(t, 0, sizeof(t));
drivers/net/wireless/ti/wl1251/spi.c
189
t[0].tx_buf = cmd;
drivers/net/wireless/ti/wl1251/spi.c
190
t[0].len = sizeof(*cmd);
drivers/net/wireless/ti/wl1251/spi.c
191
spi_message_add_tail(&t[0], &m);
drivers/net/wireless/ti/wl1251/spi.c
193
t[1].tx_buf = buf;
drivers/net/wireless/ti/wl1251/spi.c
194
t[1].len = len;
drivers/net/wireless/ti/wl1251/spi.c
195
spi_message_add_tail(&t[1], &m);
drivers/net/wireless/ti/wl1251/spi.c
46
struct spi_transfer t;
drivers/net/wireless/ti/wl1251/spi.c
55
memset(&t, 0, sizeof(t));
drivers/net/wireless/ti/wl1251/spi.c
60
t.tx_buf = cmd;
drivers/net/wireless/ti/wl1251/spi.c
61
t.len = WSPI_INIT_CMD_LEN;
drivers/net/wireless/ti/wl1251/spi.c
62
spi_message_add_tail(&t, &m);
drivers/net/wireless/ti/wl1251/spi.c
74
struct spi_transfer t;
drivers/net/wireless/ti/wl1251/spi.c
83
memset(&t, 0, sizeof(t));
drivers/net/wireless/ti/wlcore/main.c
188
static void wl1271_rx_streaming_timer(struct timer_list *t)
drivers/net/wireless/ti/wlcore/main.c
190
struct wl12xx_vif *wlvif = timer_container_of(wlvif, t,
drivers/net/wireless/ti/wlcore/spi.c
103
memset(&t, 0, sizeof(t));
drivers/net/wireless/ti/wlcore/spi.c
108
t.tx_buf = cmd;
drivers/net/wireless/ti/wlcore/spi.c
109
t.len = WSPI_INIT_CMD_LEN;
drivers/net/wireless/ti/wlcore/spi.c
110
spi_message_add_tail(&t, &m);
drivers/net/wireless/ti/wlcore/spi.c
120
struct spi_transfer t;
drivers/net/wireless/ti/wlcore/spi.c
131
memset(&t, 0, sizeof(t));
drivers/net/wireless/ti/wlcore/spi.c
163
t.tx_buf = cmd;
drivers/net/wireless/ti/wlcore/spi.c
164
t.len = WSPI_INIT_CMD_LEN;
drivers/net/wireless/ti/wlcore/spi.c
165
spi_message_add_tail(&t, &m);
drivers/net/wireless/ti/wlcore/spi.c
182
t.tx_buf = cmd;
drivers/net/wireless/ti/wlcore/spi.c
183
t.len = 4;
drivers/net/wireless/ti/wlcore/spi.c
184
spi_message_add_tail(&t, &m);
drivers/net/wireless/ti/wlcore/spi.c
199
struct spi_transfer t[1];
drivers/net/wireless/ti/wlcore/spi.c
214
memset(t, 0, sizeof(t));
drivers/net/wireless/ti/wlcore/spi.c
215
t[0].rx_buf = busy_buf;
drivers/net/wireless/ti/wlcore/spi.c
216
t[0].len = sizeof(u32);
drivers/net/wireless/ti/wlcore/spi.c
217
t[0].cs_change = true;
drivers/net/wireless/ti/wlcore/spi.c
218
spi_message_add_tail(&t[0], &m);
drivers/net/wireless/ti/wlcore/spi.c
235
struct spi_transfer t[2];
drivers/net/wireless/ti/wlcore/spi.c
257
memset(t, 0, sizeof(t));
drivers/net/wireless/ti/wlcore/spi.c
259
t[0].tx_buf = cmd;
drivers/net/wireless/ti/wlcore/spi.c
260
t[0].len = 4;
drivers/net/wireless/ti/wlcore/spi.c
261
t[0].cs_change = true;
drivers/net/wireless/ti/wlcore/spi.c
262
spi_message_add_tail(&t[0], &m);
drivers/net/wireless/ti/wlcore/spi.c
265
t[1].rx_buf = busy_buf;
drivers/net/wireless/ti/wlcore/spi.c
266
t[1].len = WL1271_BUSY_WORD_LEN;
drivers/net/wireless/ti/wlcore/spi.c
267
t[1].cs_change = true;
drivers/net/wireless/ti/wlcore/spi.c
268
spi_message_add_tail(&t[1], &m);
drivers/net/wireless/ti/wlcore/spi.c
279
memset(t, 0, sizeof(t));
drivers/net/wireless/ti/wlcore/spi.c
281
t[0].rx_buf = buf;
drivers/net/wireless/ti/wlcore/spi.c
282
t[0].len = chunk_len;
drivers/net/wireless/ti/wlcore/spi.c
283
t[0].cs_change = true;
drivers/net/wireless/ti/wlcore/spi.c
284
spi_message_add_tail(&t[0], &m);
drivers/net/wireless/ti/wlcore/spi.c
301
struct spi_transfer *t;
drivers/net/wireless/ti/wlcore/spi.c
309
t = kzalloc(sizeof(*t) * 2 * WSPI_MAX_NUM_OF_CHUNKS, GFP_KERNEL);
drivers/net/wireless/ti/wlcore/spi.c
310
if (!t)
drivers/net/wireless/ti/wlcore/spi.c
331
t[i].tx_buf = cmd;
drivers/net/wireless/ti/wlcore/spi.c
332
t[i].len = sizeof(*cmd);
drivers/net/wireless/ti/wlcore/spi.c
333
spi_message_add_tail(&t[i++], &m);
drivers/net/wireless/ti/wlcore/spi.c
335
t[i].tx_buf = buf;
drivers/net/wireless/ti/wlcore/spi.c
336
t[i].len = chunk_len;
drivers/net/wireless/ti/wlcore/spi.c
337
spi_message_add_tail(&t[i++], &m);
drivers/net/wireless/ti/wlcore/spi.c
348
kfree(t);
drivers/net/wireless/ti/wlcore/spi.c
93
struct spi_transfer t;
drivers/net/wireless/zydas/zd1211rw/zd_chip.c
1263
int r, t;
drivers/net/wireless/zydas/zd1211rw/zd_chip.c
1283
t = zd_chip_unlock_phy_regs(chip);
drivers/net/wireless/zydas/zd1211rw/zd_chip.c
1284
if (t && !r)
drivers/net/wireless/zydas/zd1211rw/zd_chip.c
1285
r = t;
drivers/net/wireless/zydas/zd1211rw/zd_chip.c
188
unsigned int i, j, t, max;
drivers/net/wireless/zydas/zd1211rw/zd_chip.c
193
for (i = 0; i < count; i += j + t) {
drivers/net/wireless/zydas/zd1211rw/zd_chip.c
194
t = 0;
drivers/net/wireless/zydas/zd1211rw/zd_chip.c
200
t = 1;
drivers/net/wireless/zydas/zd1211rw/zd_chip.c
226
unsigned int i, j, t, max;
drivers/net/wireless/zydas/zd1211rw/zd_chip.c
230
for (i = 0; i < count; i += j + t) {
drivers/net/wireless/zydas/zd1211rw/zd_chip.c
231
t = 0;
drivers/net/wireless/zydas/zd1211rw/zd_chip.c
237
t = 1;
drivers/net/wireless/zydas/zd1211rw/zd_chip.c
662
int r, t;
drivers/net/wireless/zydas/zd1211rw/zd_chip.c
676
t = zd_chip_unlock_phy_regs(chip);
drivers/net/wireless/zydas/zd1211rw/zd_chip.c
677
if (t && !r)
drivers/net/wireless/zydas/zd1211rw/zd_chip.c
678
r = t;
drivers/net/wireless/zydas/zd1211rw/zd_chip.c
744
int r, t;
drivers/net/wireless/zydas/zd1211rw/zd_chip.c
753
t = zd_chip_unlock_phy_regs(chip);
drivers/net/wireless/zydas/zd1211rw/zd_chip.c
754
if (t && !r)
drivers/net/wireless/zydas/zd1211rw/zd_chip.c
755
r = t;
drivers/net/wireless/zydas/zd1211rw/zd_mac.c
632
u32 t = bits % 11;
drivers/net/wireless/zydas/zd1211rw/zd_mac.c
634
if (0 < t && t <= 3) {
drivers/net/wireless/zydas/zd1211rw/zd_rf.c
126
int r, t;
drivers/net/wireless/zydas/zd1211rw/zd_rf.c
133
t = rf->switch_radio_on(rf);
drivers/net/wireless/zydas/zd1211rw/zd_rf.c
135
if (t)
drivers/net/wireless/zydas/zd1211rw/zd_rf.c
136
r = t;
drivers/net/wireless/zydas/zd1211rw/zd_rf.c
142
int r, t;
drivers/net/wireless/zydas/zd1211rw/zd_rf.c
150
t = rf->switch_radio_off(rf);
drivers/net/wireless/zydas/zd1211rw/zd_rf.c
152
if (t)
drivers/net/wireless/zydas/zd1211rw/zd_rf.c
153
r = t;
drivers/net/wireless/zydas/zd1211rw/zd_rf.c
61
int t;
drivers/net/wireless/zydas/zd1211rw/zd_rf.c
95
t = rf->init_hw(rf);
drivers/net/wireless/zydas/zd1211rw/zd_rf.c
97
if (t)
drivers/net/wireless/zydas/zd1211rw/zd_rf.c
98
r = t;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1143
static void zd_usb_reset_rx_idle_timer_tasklet(struct tasklet_struct *t)
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1145
struct zd_usb *usb = from_tasklet(usb, t, rx.reset_timer_tasklet);
drivers/net/wwan/wwan_hwsim.c
234
static void wwan_hwsim_nmea_emul_timer(struct timer_list *t)
drivers/net/wwan/wwan_hwsim.c
239
struct wwan_hwsim_port *port = timer_container_of(port, t, nmea_emul.timer);
drivers/net/xen-netback/common.h
344
void xenvif_tx_credit_callback(struct timer_list *t);
drivers/net/xen-netback/netback.c
199
void xenvif_tx_credit_callback(struct timer_list *t)
drivers/net/xen-netback/netback.c
201
struct xenvif_queue *queue = timer_container_of(queue, t,
drivers/net/xen-netfront.c
246
static void rx_refill_timeout(struct timer_list *t)
drivers/net/xen-netfront.c
248
struct netfront_queue *queue = timer_container_of(queue, t,
drivers/nfc/nfcmrvl/fw_dnld.c
120
static void fw_dnld_timeout(struct timer_list *t)
drivers/nfc/nfcmrvl/fw_dnld.c
122
struct nfcmrvl_private *priv = timer_container_of(priv, t,
drivers/nfc/pn533/pn533.c
1234
static void pn533_listen_mode_timer(struct timer_list *t)
drivers/nfc/pn533/pn533.c
1236
struct pn533 *dev = timer_container_of(dev, t, listen_timer);
drivers/nfc/pn533/pn533.c
2067
struct sk_buff *skb, *tmp, *t;
drivers/nfc/pn533/pn533.c
2078
skb_queue_walk_safe(&dev->resp_q, tmp, t)
drivers/nfc/pn533/pn533.c
2090
skb_queue_walk_safe(&dev->resp_q, tmp, t) {
drivers/nfc/pn533/uart.c
134
static void pn532_cmd_timeout(struct timer_list *t)
drivers/nfc/pn533/uart.c
136
struct pn532_uart_phy *dev = timer_container_of(dev, t, cmd_timeout);
drivers/nfc/st-nci/ndlc.c
238
static void ndlc_t1_timeout(struct timer_list *t)
drivers/nfc/st-nci/ndlc.c
240
struct llt_ndlc *ndlc = timer_container_of(ndlc, t, t1_timer);
drivers/nfc/st-nci/ndlc.c
245
static void ndlc_t2_timeout(struct timer_list *t)
drivers/nfc/st-nci/ndlc.c
247
struct llt_ndlc *ndlc = timer_container_of(ndlc, t, t2_timer);
drivers/nfc/st-nci/se.c
686
static void st_nci_se_wt_timeout(struct timer_list *t)
drivers/nfc/st-nci/se.c
699
struct st_nci_info *info = timer_container_of(info, t,
drivers/nfc/st-nci/se.c
716
static void st_nci_se_activation_timeout(struct timer_list *t)
drivers/nfc/st-nci/se.c
718
struct st_nci_info *info = timer_container_of(info, t,
drivers/nfc/st21nfca/se.c
281
static void st21nfca_se_wt_timeout(struct timer_list *t)
drivers/nfc/st21nfca/se.c
283
struct st21nfca_hci_info *info = timer_container_of(info, t,
drivers/nfc/st21nfca/se.c
289
static void st21nfca_se_activation_timeout(struct timer_list *t)
drivers/nfc/st21nfca/se.c
291
struct st21nfca_hci_info *info = timer_container_of(info, t,
drivers/nfc/st95hf/spi.c
133
struct spi_transfer t[2] = {
drivers/nfc/st95hf/spi.c
144
spi_message_add_tail(&t[0], &m);
drivers/nfc/st95hf/spi.c
145
spi_message_add_tail(&t[1], &m);
drivers/nfc/st95hf/spi.c
76
struct spi_transfer t[2] = {
drivers/nfc/st95hf/spi.c
89
spi_message_add_tail(&t[0], &m);
drivers/nfc/st95hf/spi.c
90
spi_message_add_tail(&t[1], &m);
drivers/nfc/trf7970a.c
498
struct spi_transfer t[2];
drivers/nfc/trf7970a.c
506
memset(&t, 0, sizeof(t));
drivers/nfc/trf7970a.c
508
t[0].tx_buf = &addr;
drivers/nfc/trf7970a.c
509
t[0].len = sizeof(addr);
drivers/nfc/trf7970a.c
510
spi_message_add_tail(&t[0], &m);
drivers/nfc/trf7970a.c
512
t[1].rx_buf = buf;
drivers/nfc/trf7970a.c
513
t[1].len = len;
drivers/nfc/trf7970a.c
514
spi_message_add_tail(&t[1], &m);
drivers/nfc/trf7970a.c
693
struct spi_transfer t[2];
drivers/nfc/trf7970a.c
703
memset(&t, 0, sizeof(t));
drivers/nfc/trf7970a.c
705
t[0].tx_buf = prefix;
drivers/nfc/trf7970a.c
706
t[0].len = prefix_len;
drivers/nfc/trf7970a.c
707
spi_message_add_tail(&t[0], &m);
drivers/nfc/trf7970a.c
709
t[1].tx_buf = skb->data;
drivers/nfc/trf7970a.c
710
t[1].len = len;
drivers/nfc/trf7970a.c
711
spi_message_add_tail(&t[1], &m);
drivers/ntb/test/ntb_pingpong.c
214
static enum hrtimer_restart pp_timer_func(struct hrtimer *t)
drivers/ntb/test/ntb_pingpong.c
216
struct pp_ctx *pp = to_pp_timer(t);
drivers/nubus/nubus.c
177
unsigned char *t = dest;
drivers/nubus/nubus.c
181
*t++ = nubus_get_rom(&p, 1, dirent->mask);
drivers/nubus/nubus.c
190
char *t = dest;
drivers/nubus/nubus.c
198
*t++ = c;
drivers/nubus/nubus.c
202
*t = '\0';
drivers/nubus/nubus.c
203
return t - dest;
drivers/nvme/host/multipath.c
1001
static void nvme_anatt_timeout(struct timer_list *t)
drivers/nvme/host/multipath.c
1003
struct nvme_ctrl *ctrl = timer_container_of(ctrl, t, anatt_timer);
drivers/nvmem/lan9662-otpc.c
11
#define OTP_OTP_ADDR_HI(t) (t + 0x04)
drivers/nvmem/lan9662-otpc.c
12
#define OTP_OTP_ADDR_LO(t) (t + 0x08)
drivers/nvmem/lan9662-otpc.c
13
#define OTP_OTP_PRGM_DATA(t) (t + 0x10)
drivers/nvmem/lan9662-otpc.c
14
#define OTP_OTP_PRGM_MODE(t) (t + 0x14)
drivers/nvmem/lan9662-otpc.c
16
#define OTP_OTP_RD_DATA(t) (t + 0x18)
drivers/nvmem/lan9662-otpc.c
17
#define OTP_OTP_FUNC_CMD(t) (t + 0x20)
drivers/nvmem/lan9662-otpc.c
20
#define OTP_OTP_CMD_GO(t) (t + 0x28)
drivers/nvmem/lan9662-otpc.c
22
#define OTP_OTP_PASS_FAIL(t) (t + 0x2c)
drivers/nvmem/lan9662-otpc.c
26
#define OTP_OTP_STATUS(t) (t + 0x30)
drivers/nvmem/lan9662-otpc.c
9
#define OTP_OTP_PWR_DN(t) (t + 0x00)
drivers/parisc/iosapic.c
620
u32 *t = (u32 *) ((ulong) vi->eoi_addr & ~0xffUL);
drivers/parisc/iosapic.c
622
for ( ; t < vi->eoi_addr; t++)
drivers/parisc/iosapic.c
623
printk(" %x", readl(t));
drivers/parisc/lba_pci.c
890
u##size t; \
drivers/parisc/lba_pci.c
891
t = READ_REG##size(astro_iop_base + addr); \
drivers/parisc/lba_pci.c
892
DBG_PORT(" 0x%x\n", t); \
drivers/parisc/lba_pci.c
893
return (t); \
drivers/parisc/lba_pci.c
972
u##size t; \
drivers/parisc/lba_pci.c
974
t = READ_REG##size(PIOP_TO_GMMIO(LBA_DEV(l), addr)); \
drivers/parisc/lba_pci.c
975
DBG_PORT(" 0x%x\n", t); \
drivers/parisc/lba_pci.c
976
return (t); \
drivers/parisc/sba_iommu.c
2007
char t = sba_dev->id.hw_type;
drivers/parisc/sba_iommu.c
2010
WARN_ON((t != HPHW_IOA) && (t != HPHW_BCPORT));
drivers/parisc/sba_iommu.c
2028
char t = sba_dev->id.hw_type;
drivers/parisc/sba_iommu.c
2032
BUG_ON((t!=HPHW_IOA) && (t!=HPHW_BCPORT));
drivers/parisc/sba_iommu.c
2071
char t = sba_dev->id.hw_type;
drivers/parisc/sba_iommu.c
2075
BUG_ON((t!=HPHW_IOA) && (t!=HPHW_BCPORT));
drivers/parport/ieee1284.c
41
static void timeout_waiting_on_port (struct timer_list *t)
drivers/parport/ieee1284.c
43
struct parport *port = timer_container_of(port, t, timer);
drivers/parport/procfs.c
417
struct parport_sysctl_table *t;
drivers/parport/procfs.c
421
t = kmemdup(&parport_sysctl_template, sizeof(*t), GFP_KERNEL);
drivers/parport/procfs.c
422
if (t == NULL)
drivers/parport/procfs.c
425
t->device_dir[0].extra1 = port;
drivers/parport/procfs.c
427
t->vars[0].data = &port->spintime;
drivers/parport/procfs.c
429
t->vars[i].extra1 = port;
drivers/parport/procfs.c
431
t->vars[5 + i].extra2 = &port->probe_info[i];
drivers/parport/procfs.c
441
t->devices_header = register_sysctl(tmp_dir_path, t->device_dir);
drivers/parport/procfs.c
442
if (t->devices_header == NULL) {
drivers/parport/procfs.c
455
t->port_header = register_sysctl(tmp_dir_path, t->vars);
drivers/parport/procfs.c
456
if (t->port_header == NULL) {
drivers/parport/procfs.c
461
port->sysctl_table = t;
drivers/parport/procfs.c
467
unregister_sysctl_table(t->devices_header);
drivers/parport/procfs.c
473
kfree(t);
drivers/parport/procfs.c
480
struct parport_sysctl_table *t = port->sysctl_table;
drivers/parport/procfs.c
482
unregister_sysctl_table(t->devices_header);
drivers/parport/procfs.c
483
unregister_sysctl_table(t->port_header);
drivers/parport/procfs.c
484
kfree(t);
drivers/parport/procfs.c
491
struct parport_device_sysctl_table *t;
drivers/parport/procfs.c
496
t = kmemdup(&parport_device_sysctl_template, sizeof(*t), GFP_KERNEL);
drivers/parport/procfs.c
497
if (t == NULL)
drivers/parport/procfs.c
507
t->vars[0].data = &device->timeslice;
drivers/parport/procfs.c
509
t->sysctl_header = register_sysctl(tmp_dir_path, t->vars);
drivers/parport/procfs.c
510
if (t->sysctl_header == NULL) {
drivers/parport/procfs.c
511
kfree(t);
drivers/parport/procfs.c
512
t = NULL;
drivers/parport/procfs.c
514
device->sysctl_table = t;
drivers/parport/procfs.c
520
kfree(t);
drivers/parport/procfs.c
528
struct parport_device_sysctl_table *t = device->sysctl_table;
drivers/parport/procfs.c
530
unregister_sysctl_table(t->sysctl_header);
drivers/parport/procfs.c
531
kfree(t);
drivers/pci/hotplug/cpqphp.h
399
void cpqhp_pushbutton_thread(struct timer_list *t);
drivers/pci/hotplug/cpqphp_ctrl.c
1716
static void pushbutton_helper_thread(struct timer_list *t)
drivers/pci/hotplug/cpqphp_ctrl.c
1718
pushbutton_pending = t;
drivers/pci/hotplug/cpqphp_ctrl.c
1882
void cpqhp_pushbutton_thread(struct timer_list *t)
drivers/pci/hotplug/cpqphp_ctrl.c
1886
struct slot *p_slot = timer_container_of(p_slot, t, task_event);
drivers/pci/hotplug/shpchp_hpc.c
212
static void int_poll_timeout(struct timer_list *t)
drivers/pci/hotplug/shpchp_hpc.c
214
struct controller *ctrl = timer_container_of(ctrl, t, poll_timer);
drivers/pci/pci.c
1072
pci_power_t t)
drivers/pci/pci.c
1075
return mid_pci_set_power_state(dev, t);
drivers/pci/pci.c
1077
return acpi_pci_set_power_state(dev, t);
drivers/pci/pcie/aer.c
432
#define AER_AGENT_REQUESTER_MASK(t) ((t == AER_CORRECTABLE) ? \
drivers/pci/pcie/aer.c
434
#define AER_AGENT_COMPLETER_MASK(t) ((t == AER_CORRECTABLE) ? \
drivers/pci/pcie/aer.c
436
#define AER_AGENT_TRANSMITTER_MASK(t) ((t == AER_CORRECTABLE) ? \
drivers/pci/pcie/aer.c
439
#define AER_GET_AGENT(t, e) \
drivers/pci/pcie/aer.c
440
((e & AER_AGENT_COMPLETER_MASK(t)) ? AER_AGENT_COMPLETER : \
drivers/pci/pcie/aer.c
441
(e & AER_AGENT_REQUESTER_MASK(t)) ? AER_AGENT_REQUESTER : \
drivers/pci/pcie/aer.c
442
(e & AER_AGENT_TRANSMITTER_MASK(t)) ? AER_AGENT_TRANSMITTER : \
drivers/pci/pcie/aer.c
449
#define AER_PHYSICAL_LAYER_ERROR_MASK(t) ((t == AER_CORRECTABLE) ? \
drivers/pci/pcie/aer.c
451
#define AER_DATA_LINK_LAYER_ERROR_MASK(t) ((t == AER_CORRECTABLE) ? \
drivers/pci/pcie/aer.c
457
#define AER_GET_LAYER_ERROR(t, e) \
drivers/pci/pcie/aer.c
458
((e & AER_PHYSICAL_LAYER_ERROR_MASK(t)) ? AER_PHYSICAL_LAYER_ERROR : \
drivers/pci/pcie/aer.c
459
(e & AER_DATA_LINK_LAYER_ERROR_MASK(t)) ? AER_DATA_LINK_LAYER_ERROR : \
drivers/pci/xen-pcifront.c
554
struct pci_bus_entry *bus_entry, *t;
drivers/pci/xen-pcifront.c
559
list_for_each_entry_safe(bus_entry, t, &pdev->root_buses, list) {
drivers/pcmcia/bcm63xx_pcmcia.c
266
static void bcm63xx_pcmcia_poll(struct timer_list *t)
drivers/pcmcia/bcm63xx_pcmcia.c
271
skt = timer_container_of(skt, t, timer);
drivers/pcmcia/electra_cf.c
69
static void electra_cf_timer(struct timer_list *t)
drivers/pcmcia/electra_cf.c
71
struct electra_cf_socket *cf = timer_container_of(cf, t, timer);
drivers/pcmcia/i82365.c
1008
if (t->flags & IS_CIRRUS) {
drivers/pcmcia/i82365.c
1015
reg = t->cs_irq << 4;
drivers/pcmcia/i82365.c
323
struct i82365_socket *t = &socket[s];
drivers/pcmcia/i82365.c
349
if (!(t->flags & IS_VIA)) {
drivers/pcmcia/i82365.c
427
struct i82365_socket *t = &socket[s];
drivers/pcmcia/i82365.c
428
if (t->flags & IS_CIRRUS)
drivers/pcmcia/i82365.c
430
else if (t->flags & IS_VADEM)
drivers/pcmcia/i82365.c
436
struct i82365_socket *t = &socket[s];
drivers/pcmcia/i82365.c
437
if (t->flags & IS_CIRRUS)
drivers/pcmcia/i82365.c
443
i365_bflip(s, I365_INTCTL, I365_INTR_ENA, t->intr);
drivers/pcmcia/i82365.c
444
if (t->flags & IS_VADEM)
drivers/pcmcia/i82365.c
677
struct i82365_socket *t = &socket[sockets-ns];
drivers/pcmcia/i82365.c
683
t->ioaddr, t->psock*0x40);
drivers/pcmcia/i82365.c
728
t[i].socket.features |= SS_CAP_PCCARD;
drivers/pcmcia/i82365.c
729
t[i].socket.map_size = 0x1000;
drivers/pcmcia/i82365.c
730
t[i].socket.irq_mask = mask;
drivers/pcmcia/i82365.c
731
t[i].cs_irq = isa_irq;
drivers/pcmcia/i82365.c
924
struct i82365_socket *t = &socket[sock];
drivers/pcmcia/i82365.c
935
reg = t->intr;
drivers/pcmcia/i82365.c
945
if (t->flags & IS_CIRRUS) {
drivers/pcmcia/i82365.c
961
} else if (t->flags & IS_VG_PWR) {
drivers/pcmcia/i82365.c
977
} else if (t->flags & IS_DF_PWR) {
drivers/pcmcia/omap_cf.c
78
static void omap_cf_timer(struct timer_list *t)
drivers/pcmcia/omap_cf.c
80
struct omap_cf_socket *cf = timer_container_of(cf, t, timer);
drivers/pcmcia/pd6729.c
237
static void pd6729_interrupt_wrapper(struct timer_list *t)
drivers/pcmcia/pd6729.c
239
struct pd6729_socket *socket = timer_container_of(socket, t,
drivers/pcmcia/pd6729.c
252
struct pd6729_socket *t;
drivers/pcmcia/pd6729.c
286
t = (socket->number) ? socket : socket + 1;
drivers/pcmcia/pd6729.c
287
indirect_write(t, PD67_EXT_INDEX, PD67_EXTERN_DATA);
drivers/pcmcia/pd6729.c
288
data = indirect_read16(t, PD67_EXT_DATA);
drivers/pcmcia/sa11xx_base.h
108
unsigned int t = ((pcmcia_cycle_ns * cpu_clock_khz) / 6) - 1000000;
drivers/pcmcia/sa11xx_base.h
109
return (t / 1000000) + (((t % 1000000) == 0) ? 0 : 1);
drivers/pcmcia/soc_common.c
461
static void soc_common_pcmcia_poll_event(struct timer_list *t)
drivers/pcmcia/soc_common.c
463
struct soc_pcmcia_socket *skt = timer_container_of(skt, t, poll_timer);
drivers/pcmcia/yenta_socket.c
540
static void yenta_interrupt_wrapper(struct timer_list *t)
drivers/pcmcia/yenta_socket.c
542
struct yenta_socket *socket = timer_container_of(socket, t,
drivers/phy/microchip/sparx5_serdes_regs.h
1018
#define SD10G_LANE_LANE_A2(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1019
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 640, 0, 1, 128, 8, 0,\
drivers/phy/microchip/sparx5_serdes_regs.h
1029
#define SD10G_LANE_LANE_DF(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
103
#define SD10G_LANE_LANE_03(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1030
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 832, 0, 1, 84, 60, 0,\
drivers/phy/microchip/sparx5_serdes_regs.h
104
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 0, 0, 1, 288, 12, 0, \
drivers/phy/microchip/sparx5_serdes_regs.h
1059
#define SD25G_LANE_CMU_09(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1060
__REG(TARGET_SD25G_LANE, t, 8, 0, 0, 1, 132, 36, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
1094
#define SD25G_LANE_CMU_0B(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1095
__REG(TARGET_SD25G_LANE, t, 8, 0, 0, 1, 132, 44, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
114
#define SD10G_LANE_LANE_04(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1147
#define SD25G_LANE_CMU_0C(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1148
__REG(TARGET_SD25G_LANE, t, 8, 0, 0, 1, 132, 48, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
115
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 0, 0, 1, 288, 16, 0, \
drivers/phy/microchip/sparx5_serdes_regs.h
1182
#define SD25G_LANE_CMU_0D(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1183
__REG(TARGET_SD25G_LANE, t, 8, 0, 0, 1, 132, 52, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
1217
#define SD25G_LANE_CMU_0E(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1218
__REG(TARGET_SD25G_LANE, t, 8, 0, 0, 1, 132, 56, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
1234
#define SD25G_LANE_CMU_13(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1235
__REG(TARGET_SD25G_LANE, t, 8, 0, 0, 1, 132, 76, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
125
#define SD10G_LANE_LANE_06(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1251
#define SD25G_LANE_CMU_18(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1252
__REG(TARGET_SD25G_LANE, t, 8, 0, 0, 1, 132, 96, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
126
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 0, 0, 1, 288, 24, 0, \
drivers/phy/microchip/sparx5_serdes_regs.h
1280
#define SD25G_LANE_CMU_19(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1281
__REG(TARGET_SD25G_LANE, t, 8, 0, 0, 1, 132, 100, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
1297
#define SD25G_LANE_CMU_1A(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1298
__REG(TARGET_SD25G_LANE, t, 8, 0, 0, 1, 132, 104, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
1326
#define SD25G_LANE_CMU_2A(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1327
__REG(TARGET_SD25G_LANE, t, 8, 132, 0, 1, 124, 36, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
1349
#define SD25G_LANE_CMU_30(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1350
__REG(TARGET_SD25G_LANE, t, 8, 132, 0, 1, 124, 60, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
1366
#define SD25G_LANE_CMU_31(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1367
__REG(TARGET_SD25G_LANE, t, 8, 132, 0, 1, 124, 64, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
1377
#define SD25G_LANE_CMU_40(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1378
__REG(TARGET_SD25G_LANE, t, 8, 256, 0, 1, 512, 0, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
1418
#define SD25G_LANE_CMU_45(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1419
__REG(TARGET_SD25G_LANE, t, 8, 256, 0, 1, 512, 20, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
1429
#define SD25G_LANE_CMU_46(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1430
__REG(TARGET_SD25G_LANE, t, 8, 256, 0, 1, 512, 24, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
1440
#define SD25G_LANE_CMU_C0(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1441
__REG(TARGET_SD25G_LANE, t, 8, 768, 0, 1, 252, 0, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
1457
#define SD25G_LANE_CMU_FF(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1458
__REG(TARGET_SD25G_LANE, t, 8, 1020, 0, 1, 4, 0, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
1468
#define SD25G_LANE_LANE_00(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1469
__REG(TARGET_SD25G_LANE, t, 8, 1024, 0, 1, 768, 0, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
1485
#define SD25G_LANE_LANE_01(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1486
__REG(TARGET_SD25G_LANE, t, 8, 1024, 0, 1, 768, 4, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
1502
#define SD25G_LANE_LANE_03(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1503
__REG(TARGET_SD25G_LANE, t, 8, 1024, 0, 1, 768, 12, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
1513
#define SD25G_LANE_LANE_04(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1514
__REG(TARGET_SD25G_LANE, t, 8, 1024, 0, 1, 768, 16, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
1554
#define SD25G_LANE_LANE_05(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1555
__REG(TARGET_SD25G_LANE, t, 8, 1024, 0, 1, 768, 20, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
1571
#define SD25G_LANE_LANE_06(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1572
__REG(TARGET_SD25G_LANE, t, 8, 1024, 0, 1, 768, 24, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
1588
#define SD25G_LANE_LANE_07(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1589
__REG(TARGET_SD25G_LANE, t, 8, 1024, 0, 1, 768, 28, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
1611
#define SD25G_LANE_LANE_09(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1612
__REG(TARGET_SD25G_LANE, t, 8, 1024, 0, 1, 768, 36, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
1622
#define SD25G_LANE_LANE_0A(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1623
__REG(TARGET_SD25G_LANE, t, 8, 1024, 0, 1, 768, 40, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
1633
#define SD25G_LANE_LANE_0B(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1634
__REG(TARGET_SD25G_LANE, t, 8, 1024, 0, 1, 768, 44, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
1656
#define SD25G_LANE_LANE_0C(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1657
__REG(TARGET_SD25G_LANE, t, 8, 1024, 0, 1, 768, 48, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
166
#define SD10G_LANE_LANE_0B(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
167
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 0, 0, 1, 288, 44, 0, \
drivers/phy/microchip/sparx5_serdes_regs.h
1679
#define SD25G_LANE_LANE_0D(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1680
__REG(TARGET_SD25G_LANE, t, 8, 1024, 0, 1, 768, 52, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
1714
#define SD25G_LANE_LANE_0E(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1715
__REG(TARGET_SD25G_LANE, t, 8, 1024, 0, 1, 768, 56, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
1743
#define SD25G_LANE_LANE_0F(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1744
__REG(TARGET_SD25G_LANE, t, 8, 1024, 0, 1, 768, 60, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
1754
#define SD25G_LANE_LANE_18(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1755
__REG(TARGET_SD25G_LANE, t, 8, 1024, 0, 1, 768, 96, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
1789
#define SD25G_LANE_LANE_19(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1790
__REG(TARGET_SD25G_LANE, t, 8, 1024, 0, 1, 768, 100, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
1842
#define SD25G_LANE_LANE_1A(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1843
__REG(TARGET_SD25G_LANE, t, 8, 1024, 0, 1, 768, 104, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
1859
#define SD25G_LANE_LANE_1B(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1860
__REG(TARGET_SD25G_LANE, t, 8, 1024, 0, 1, 768, 108, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
1870
#define SD25G_LANE_LANE_1C(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1871
__REG(TARGET_SD25G_LANE, t, 8, 1024, 0, 1, 768, 112, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
1899
#define SD25G_LANE_LANE_1D(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1900
__REG(TARGET_SD25G_LANE, t, 8, 1024, 0, 1, 768, 116, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
1952
#define SD25G_LANE_LANE_1E(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1953
__REG(TARGET_SD25G_LANE, t, 8, 1024, 0, 1, 768, 120, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
1987
#define SD25G_LANE_LANE_21(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1988
__REG(TARGET_SD25G_LANE, t, 8, 1024, 0, 1, 768, 132, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
1998
#define SD25G_LANE_LANE_22(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
1999
__REG(TARGET_SD25G_LANE, t, 8, 1024, 0, 1, 768, 136, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
2009
#define SD25G_LANE_LANE_25(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
201
#define SD10G_LANE_LANE_0C(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2010
__REG(TARGET_SD25G_LANE, t, 8, 1024, 0, 1, 768, 148, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
202
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 0, 0, 1, 288, 48, 0, \
drivers/phy/microchip/sparx5_serdes_regs.h
2020
#define SD25G_LANE_LANE_26(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2021
__REG(TARGET_SD25G_LANE, t, 8, 1024, 0, 1, 768, 152, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
2031
#define SD25G_LANE_LANE_28(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2032
__REG(TARGET_SD25G_LANE, t, 8, 1024, 0, 1, 768, 160, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
2060
#define SD25G_LANE_LANE_2B(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2061
__REG(TARGET_SD25G_LANE, t, 8, 1024, 0, 1, 768, 172, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
2083
#define SD25G_LANE_LANE_2C(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2084
__REG(TARGET_SD25G_LANE, t, 8, 1024, 0, 1, 768, 176, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
2100
#define SD25G_LANE_LANE_2D(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2101
__REG(TARGET_SD25G_LANE, t, 8, 1024, 0, 1, 768, 180, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
2117
#define SD25G_LANE_LANE_2E(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2118
__REG(TARGET_SD25G_LANE, t, 8, 1024, 0, 1, 768, 184, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
2170
#define SD25G_LANE_LANE_40(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2171
__REG(TARGET_SD25G_LANE, t, 8, 1024, 0, 1, 768, 256, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
2217
#define SD25G_LANE_LANE_42(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2218
__REG(TARGET_SD25G_LANE, t, 8, 1024, 0, 1, 768, 264, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
2228
#define SD25G_LANE_LANE_43(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2229
__REG(TARGET_SD25G_LANE, t, 8, 1024, 0, 1, 768, 268, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
2239
#define SD25G_LANE_LANE_44(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2240
__REG(TARGET_SD25G_LANE, t, 8, 1024, 0, 1, 768, 272, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
2250
#define SD25G_LANE_LANE_45(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2251
__REG(TARGET_SD25G_LANE, t, 8, 1024, 0, 1, 768, 276, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
2261
#define SD25G_LANE_LANE_DE(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2262
__REG(TARGET_SD25G_LANE, t, 8, 1792, 0, 1, 128, 120, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
2290
#define SD6G_LANE_LANE_DF(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2291
__REG(TARGET_SD6G_LANE, t, 13, 832, 0, 1, 84, 60, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
2318
#define SD_CMU_CMU_00(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2319
__REG(TARGET_SD_CMU, t, TSIZE(TC_SD_CMU), 0, 0, 1, 20, 0, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
2346
#define SD_CMU_CMU_05(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2347
__REG(TARGET_SD_CMU, t, TSIZE(TC_SD_CMU), 20, 0, 1, 72, 0, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
2362
#define SD_CMU_CMU_06(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2363
__REG(TARGET_SD_CMU, t, TSIZE(TC_SD_CMU), 20, 0, 1, 72, 4, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
2414
#define SD_CMU_CMU_08(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2415
__REG(TARGET_SD_CMU, t, TSIZE(TC_SD_CMU), 20, 0, 1, 72, 12, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
2448
#define SD_CMU_CMU_09(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2449
__REG(TARGET_SD_CMU, t, TSIZE(TC_SD_CMU), 20, 0, 1, 72, 16, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
2476
#define SD_CMU_CMU_0D(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2477
__REG(TARGET_SD_CMU, t, TSIZE(TC_SD_CMU), 20, 0, 1, 72, 32, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
2510
#define SD_CMU_CMU_1B(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2511
__REG(TARGET_SD_CMU, t, TSIZE(TC_SD_CMU), 104, 0, 1, 20, 4, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
2520
#define SD_CMU_CMU_1F(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2521
__REG(TARGET_SD_CMU, t, TSIZE(TC_SD_CMU), 124, 0, 1, 68, 0, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
254
#define SD10G_LANE_LANE_0D(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2548
#define SD_CMU_CMU_30(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2549
__REG(TARGET_SD_CMU, t, TSIZE(TC_SD_CMU), 192, 0, 1, 72, 0, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
255
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 0, 0, 1, 288, 52, 0, \
drivers/phy/microchip/sparx5_serdes_regs.h
2558
#define SD_CMU_CMU_44(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2559
__REG(TARGET_SD_CMU, t, TSIZE(TC_SD_CMU), 264, 0, 1, 632, 8, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
2574
#define SD_CMU_CMU_45(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2575
__REG(TARGET_SD_CMU, t, TSIZE(TC_SD_CMU), 264, 0, 1, 632, 12, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
2626
#define SD_CMU_CMU_47(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2627
__REG(TARGET_SD_CMU, t, TSIZE(TC_SD_CMU), 264, 0, 1, 632, 20, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
2636
#define SD_CMU_CMU_E0(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2637
__REG(TARGET_SD_CMU, t, TSIZE(TC_SD_CMU), 896, 0, 1, 8, 0, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
2652
#define SD_CMU_CFG_SD_CMU_CFG(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2653
__REG(TARGET_SD_CMU_CFG, t, TSIZE(TC_SD_CMU_CFG), 0, 0, 1, 8, 0, 0, 1, \
drivers/phy/microchip/sparx5_serdes_regs.h
2669
#define SD_LANE_SD_SER_RST(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2670
__REG(TARGET_SD_LANE, t, TSIZE(TC_SD_LANE), 0, 0, 1, 8, 0, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
2679
#define SD_LANE_SD_DES_RST(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2680
__REG(TARGET_SD_LANE, t, TSIZE(TC_SD_LANE), 0, 0, 1, 8, 4, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
2689
#define SD_LANE_SD_LANE_CFG(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2690
__REG(TARGET_SD_LANE, t, TSIZE(TC_SD_LANE), 8, 0, 1, 8, 0, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
271
#define SD10G_LANE_LANE_0E(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
272
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 0, 0, 1, 288, 56, 0, \
drivers/phy/microchip/sparx5_serdes_regs.h
2735
#define SD_LANE_SD_LANE_STAT(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2736
__REG(TARGET_SD_LANE, t, TSIZE(TC_SD_LANE), 8, 0, 1, 8, 4, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
2757
#define SD_LANE_QUIET_MODE_6G(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2758
__REG(TARGET_SD_LANE, t, TSIZE(TC_SD_LANE), 24, 0, 1, 8, 4, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
2767
#define SD_LANE_MISC(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2768
__REG(TARGET_SD_LANE, t, TSIZE(TC_SD_LANE), 56, 0, 1, 56, 0, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
2796
#define SD_LANE_M_STAT_MISC(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2797
__REG(TARGET_SD_LANE, t, TSIZE(TC_SD_LANE), 56, 0, 1, 56, 36, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
2813
#define SD_LANE_25G_SD_SER_RST(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2814
__REG(TARGET_SD_LANE_25G, t, 8, 0, 0, 1, 8, 0, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
2824
#define SD_LANE_25G_SD_DES_RST(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2825
__REG(TARGET_SD_LANE_25G, t, 8, 0, 0, 1, 8, 4, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
2835
#define SD_LANE_25G_SD_LANE_CFG(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2836
__REG(TARGET_SD_LANE_25G, t, 8, 8, 0, 1, 12, 0, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
2936
#define SD_LANE_25G_SD_LANE_CFG2(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
2937
__REG(TARGET_SD_LANE_25G, t, 8, 8, 0, 1, 12, 4, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
300
#define SD10G_LANE_LANE_0F(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
3007
#define SD_LANE_25G_SD_LANE_STAT(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
3008
__REG(TARGET_SD_LANE_25G, t, 8, 8, 0, 1, 12, 8, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
301
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 0, 0, 1, 288, 60, 0, \
drivers/phy/microchip/sparx5_serdes_regs.h
3030
#define SD_LANE_25G_QUIET_MODE_6G(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
3031
__REG(TARGET_SD_LANE_25G, t, 8, 28, 0, 1, 8, 4, 0, 1, 4)
drivers/phy/microchip/sparx5_serdes_regs.h
311
#define SD10G_LANE_LANE_13(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
312
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 0, 0, 1, 288, 76, 0, \
drivers/phy/microchip/sparx5_serdes_regs.h
334
#define SD10G_LANE_LANE_14(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
335
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 0, 0, 1, 288, 80, 0, \
drivers/phy/microchip/sparx5_serdes_regs.h
345
#define SD10G_LANE_LANE_15(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
346
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 0, 0, 1, 288, 84, 0, \
drivers/phy/microchip/sparx5_serdes_regs.h
356
#define SD10G_LANE_LANE_16(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
357
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 0, 0, 1, 288, 88, 0, \
drivers/phy/microchip/sparx5_serdes_regs.h
367
#define SD10G_LANE_LANE_1A(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
368
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 0, 0, 1, 288, 104, 0,\
drivers/phy/microchip/sparx5_serdes_regs.h
402
#define SD10G_LANE_LANE_22(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
403
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 0, 0, 1, 288, 136, 0,\
drivers/phy/microchip/sparx5_serdes_regs.h
413
#define SD10G_LANE_LANE_23(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
414
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 0, 0, 1, 288, 140, 0,\
drivers/phy/microchip/sparx5_serdes_regs.h
448
#define SD10G_LANE_LANE_24(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
449
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 0, 0, 1, 288, 144, 0,\
drivers/phy/microchip/sparx5_serdes_regs.h
45
#define SD10G_LANE_LANE_01(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
46
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 0, 0, 1, 288, 4, 0, \
drivers/phy/microchip/sparx5_serdes_regs.h
465
#define SD10G_LANE_LANE_26(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
466
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 0, 0, 1, 288, 152, 0,\
drivers/phy/microchip/sparx5_serdes_regs.h
476
#define SD10G_LANE_LANE_2F(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
477
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 0, 0, 1, 288, 188, 0,\
drivers/phy/microchip/sparx5_serdes_regs.h
493
#define SD10G_LANE_LANE_30(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
494
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 0, 0, 1, 288, 192, 0,\
drivers/phy/microchip/sparx5_serdes_regs.h
510
#define SD10G_LANE_LANE_31(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
511
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 0, 0, 1, 288, 196, 0,\
drivers/phy/microchip/sparx5_serdes_regs.h
551
#define SD10G_LANE_LANE_32(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
552
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 0, 0, 1, 288, 200, 0,\
drivers/phy/microchip/sparx5_serdes_regs.h
568
#define SD10G_LANE_LANE_33(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
569
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 0, 0, 1, 288, 204, 0,\
drivers/phy/microchip/sparx5_serdes_regs.h
585
#define SD10G_LANE_LANE_35(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
586
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 0, 0, 1, 288, 212, 0,\
drivers/phy/microchip/sparx5_serdes_regs.h
602
#define SD10G_LANE_LANE_36(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
603
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 0, 0, 1, 288, 216, 0,\
drivers/phy/microchip/sparx5_serdes_regs.h
637
#define SD10G_LANE_LANE_37(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
638
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 0, 0, 1, 288, 220, 0,\
drivers/phy/microchip/sparx5_serdes_regs.h
666
#define SD10G_LANE_LANE_39(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
667
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 0, 0, 1, 288, 228, 0,\
drivers/phy/microchip/sparx5_serdes_regs.h
68
#define SD10G_LANE_LANE_02(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
683
#define SD10G_LANE_LANE_3A(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
684
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 0, 0, 1, 288, 232, 0,\
drivers/phy/microchip/sparx5_serdes_regs.h
69
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 0, 0, 1, 288, 8, 0, \
drivers/phy/microchip/sparx5_serdes_regs.h
700
#define SD10G_LANE_LANE_3C(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
701
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 0, 0, 1, 288, 240, 0,\
drivers/phy/microchip/sparx5_serdes_regs.h
717
#define SD10G_LANE_LANE_40(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
718
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 0, 0, 1, 288, 256, 0,\
drivers/phy/microchip/sparx5_serdes_regs.h
728
#define SD10G_LANE_LANE_41(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
729
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 0, 0, 1, 288, 260, 0,\
drivers/phy/microchip/sparx5_serdes_regs.h
739
#define SD10G_LANE_LANE_42(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
740
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 0, 0, 1, 288, 264, 0,\
drivers/phy/microchip/sparx5_serdes_regs.h
756
#define SD10G_LANE_LANE_48(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
757
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 288, 0, 1, 40, 0, 0, \
drivers/phy/microchip/sparx5_serdes_regs.h
779
#define SD10G_LANE_LANE_50(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
780
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 288, 0, 1, 40, 32, 0,\
drivers/phy/microchip/sparx5_serdes_regs.h
814
#define SD10G_LANE_LANE_52(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
815
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 328, 0, 1, 24, 0, 0, \
drivers/phy/microchip/sparx5_serdes_regs.h
825
#define SD10G_LANE_LANE_83(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
826
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 464, 0, 1, 112, 60, \
drivers/phy/microchip/sparx5_serdes_regs.h
872
#define SD10G_LANE_LANE_93(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
873
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 576, 0, 1, 64, 12, 0,\
drivers/phy/microchip/sparx5_serdes_regs.h
925
#define SD10G_LANE_LANE_94(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
926
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 576, 0, 1, 64, 16, 0,\
drivers/phy/microchip/sparx5_serdes_regs.h
960
#define SD10G_LANE_LANE_9E(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
961
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 576, 0, 1, 64, 56, 0,\
drivers/phy/microchip/sparx5_serdes_regs.h
983
#define SD10G_LANE_LANE_A1(t) \
drivers/phy/microchip/sparx5_serdes_regs.h
984
__REG(TARGET_SD10G_LANE, t, TSIZE(TC_SD10G_LANE), 640, 0, 1, 128, 4, 0,\
drivers/phy/qualcomm/phy-qcom-qmp-common.h
41
const struct qmp_phy_init_tbl *t = tbl;
drivers/phy/qualcomm/phy-qcom-qmp-common.h
43
if (!t)
drivers/phy/qualcomm/phy-qcom-qmp-common.h
46
for (i = 0; i < num; i++, t++) {
drivers/phy/qualcomm/phy-qcom-qmp-common.h
47
if (!(t->lane_mask & lane_mask))
drivers/phy/qualcomm/phy-qcom-qmp-common.h
51
t->name, t->offset, t->val);
drivers/phy/qualcomm/phy-qcom-qmp-common.h
52
writel(t->val, base + t->offset);
drivers/phy/qualcomm/phy-qcom-qmp-usb-legacy.c
707
const struct qmp_phy_init_tbl *t = tbl;
drivers/phy/qualcomm/phy-qcom-qmp-usb-legacy.c
709
if (!t)
drivers/phy/qualcomm/phy-qcom-qmp-usb-legacy.c
712
for (i = 0; i < num; i++, t++) {
drivers/phy/qualcomm/phy-qcom-qmp-usb-legacy.c
713
if (!(t->lane_mask & lane_mask))
drivers/phy/qualcomm/phy-qcom-qmp-usb-legacy.c
716
writel(t->val, base + t->offset);
drivers/phy/ralink/phy-ralink-usb.c
119
regmap_read(phy->sysctl, RT_SYSC_REG_USB_PHY_CFG, &t);
drivers/phy/ralink/phy-ralink-usb.c
121
(t & UDEV_WAKEUP) ? ("enabled") : ("disabled"));
drivers/phy/ralink/phy-ralink-usb.c
122
if (t & USB_PHY_UTMI_8B60M)
drivers/phy/ralink/phy-ralink-usb.c
94
u32 t;
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
1187
u32 t, t2;
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
1190
err = mtk_hw_get_value(hw, desc, PINCTRL_PIN_REG_PUPD, &t);
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
1208
if (pullup ^ !t)
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
1212
err = mtk_hw_get_value(hw, desc, PINCTRL_PIN_REG_R0, &t);
drivers/pinctrl/mediatek/pinctrl-mtk-common-v2.c
1220
*val = (t | t2 << 1) & 0x7;
drivers/pinctrl/mvebu/pinctrl-armada-37xx.c
691
u32 t = irq_get_trigger_type(virq);
drivers/pinctrl/mvebu/pinctrl-armada-37xx.c
693
if ((t & IRQ_TYPE_SENSE_MASK) == IRQ_TYPE_EDGE_BOTH) {
drivers/pinctrl/pinctrl-lpc18xx.c
213
#define LPC_P(port, pin, f0, f1, f2, f3, f4, f5, f6, f7, a, t) \
drivers/pinctrl/pinctrl-lpc18xx.c
222
.type = TYPE_##t, \
drivers/pinctrl/pinctrl-lpc18xx.c
225
#define LPC_N(pname, off, f0, f1, f2, f3, f4, f5, f6, f7, a, t) \
drivers/pinctrl/pinctrl-lpc18xx.c
234
.type = TYPE_##t, \
drivers/pinctrl/pinctrl-mcp23s08_spi.c
31
struct spi_transfer t[2] = { { .tx_buf = &mcp->addr, .len = 1, },
drivers/pinctrl/pinctrl-mcp23s08_spi.c
35
spi_message_add_tail(&t[0], &m);
drivers/pinctrl/pinctrl-mcp23s08_spi.c
36
spi_message_add_tail(&t[1], &m);
drivers/pinctrl/pinctrl-mcp23s08_spi.c
48
struct spi_transfer t[3] = { { .tx_buf = &mcp->addr, .len = 1, },
drivers/pinctrl/pinctrl-mcp23s08_spi.c
53
spi_message_add_tail(&t[0], &m);
drivers/pinctrl/pinctrl-mcp23s08_spi.c
54
spi_message_add_tail(&t[1], &m);
drivers/pinctrl/pinctrl-mcp23s08_spi.c
55
spi_message_add_tail(&t[2], &m);
drivers/pinctrl/renesas/pinctrl-rzg2l.c
175
#define RZG2L_PACK_HWIRQ(t, i) (((t) << 16) | (i))
drivers/platform/mellanox/mlxbf-tmfifo.c
439
static void mlxbf_tmfifo_timer(struct timer_list *t)
drivers/platform/mellanox/mlxbf-tmfifo.c
441
struct mlxbf_tmfifo *fifo = container_of(t, struct mlxbf_tmfifo, timer);
drivers/platform/x86/acerhdf.c
366
static int acerhdf_get_ec_temp(struct thermal_zone_device *thermal, int *t)
drivers/platform/x86/acerhdf.c
377
*t = temp;
drivers/platform/x86/dell/dell-smbios-base.c
188
u16 t = 0;
drivers/platform/x86/dell/dell-smbios-base.c
231
t = da_tokens[i].tokenID;
drivers/platform/x86/dell/dell-smbios-base.c
236
if (!t) {
drivers/platform/x86/dell/dell-smbios-base.c
246
if (t >= token_blacklist[i].min &&
drivers/platform/x86/dell/dell-smbios-base.c
247
t <= token_blacklist[i].max)
drivers/platform/x86/dell/dell-smbios-base.c
255
if (t < token_whitelist[i].min ||
drivers/platform/x86/dell/dell-smbios-base.c
256
t > token_whitelist[i].max)
drivers/platform/x86/dell/dell-smbios-base.c
260
dev_dbg(d, "whitelisted token: %x\n", t);
drivers/platform/x86/intel/ifs/runtest.c
162
static void wait_for_sibling_cpu(atomic_t *t, long long timeout)
drivers/platform/x86/intel/ifs/runtest.c
168
atomic_inc(t);
drivers/platform/x86/intel/ifs/runtest.c
169
while (atomic_read(t) < all_cpus) {
drivers/platform/x86/intel_ips.c
935
static void monitor_timeout(struct timer_list *t)
drivers/platform/x86/intel_ips.c
937
struct ips_driver *ips = timer_container_of(ips, t, timer);
drivers/platform/x86/lenovo/thinkpad_acpi.c
11720
char t;
drivers/platform/x86/lenovo/thinkpad_acpi.c
11742
t = tpacpi_parse_fw_id(tp->bios_version_str,
drivers/platform/x86/lenovo/thinkpad_acpi.c
11744
if (t != 'E' && t != 'C')
drivers/platform/x86/lenovo/thinkpad_acpi.c
11773
t = tpacpi_parse_fw_id(ec_fw_string,
drivers/platform/x86/lenovo/thinkpad_acpi.c
11775
if (t != 'H') {
drivers/platform/x86/lenovo/thinkpad_acpi.c
1293
unsigned long t;
drivers/platform/x86/lenovo/thinkpad_acpi.c
1298
if (parse_strtoul(buf, 1, &t))
drivers/platform/x86/lenovo/thinkpad_acpi.c
1301
tpacpi_disclose_usertask(attr->attr.name, "set to %ld\n", t);
drivers/platform/x86/lenovo/thinkpad_acpi.c
1304
if (tpacpi_rfk_check_hwblock_state() && !!t)
drivers/platform/x86/lenovo/thinkpad_acpi.c
1307
res = tpacpi_rfkill_switches[id]->ops->set_status((!!t) ?
drivers/platform/x86/lenovo/thinkpad_acpi.c
1388
unsigned long t;
drivers/platform/x86/lenovo/thinkpad_acpi.c
1390
if (parse_strtoul(buf, 0xffff, &t))
drivers/platform/x86/lenovo/thinkpad_acpi.c
1393
dbg_level = t;
drivers/platform/x86/lenovo/thinkpad_acpi.c
1420
unsigned long t;
drivers/platform/x86/lenovo/thinkpad_acpi.c
1422
if (parse_strtoul(buf, 1, &t))
drivers/platform/x86/lenovo/thinkpad_acpi.c
1425
if (tpacpi_wlsw_emulstate != !!t) {
drivers/platform/x86/lenovo/thinkpad_acpi.c
1426
tpacpi_wlsw_emulstate = !!t;
drivers/platform/x86/lenovo/thinkpad_acpi.c
1427
tpacpi_rfk_update_hwblock_state(!t); /* negative logic */
drivers/platform/x86/lenovo/thinkpad_acpi.c
1443
unsigned long t;
drivers/platform/x86/lenovo/thinkpad_acpi.c
1445
if (parse_strtoul(buf, 1, &t))
drivers/platform/x86/lenovo/thinkpad_acpi.c
1448
tpacpi_bluetooth_emulstate = !!t;
drivers/platform/x86/lenovo/thinkpad_acpi.c
1463
unsigned long t;
drivers/platform/x86/lenovo/thinkpad_acpi.c
1465
if (parse_strtoul(buf, 1, &t))
drivers/platform/x86/lenovo/thinkpad_acpi.c
1468
tpacpi_wwan_emulstate = !!t;
drivers/platform/x86/lenovo/thinkpad_acpi.c
1483
unsigned long t;
drivers/platform/x86/lenovo/thinkpad_acpi.c
1485
if (parse_strtoul(buf, 1, &t))
drivers/platform/x86/lenovo/thinkpad_acpi.c
1488
tpacpi_uwb_emulstate = !!t;
drivers/platform/x86/lenovo/thinkpad_acpi.c
2506
unsigned long t;
drivers/platform/x86/lenovo/thinkpad_acpi.c
2518
t = 0;
drivers/platform/x86/lenovo/thinkpad_acpi.c
2531
if (t == 0) {
drivers/platform/x86/lenovo/thinkpad_acpi.c
2533
t = 1000/poll_freq;
drivers/platform/x86/lenovo/thinkpad_acpi.c
2535
t = 100; /* should never happen... */
drivers/platform/x86/lenovo/thinkpad_acpi.c
2537
t = msleep_interruptible(t);
drivers/platform/x86/lenovo/thinkpad_acpi.c
2541
if (t > 0 && !was_frozen)
drivers/platform/x86/lenovo/thinkpad_acpi.c
2548
t = 0;
drivers/platform/x86/lenovo/thinkpad_acpi.c
2688
unsigned long t;
drivers/platform/x86/lenovo/thinkpad_acpi.c
2693
if (parse_strtoul(buf, 1, &t))
drivers/platform/x86/lenovo/thinkpad_acpi.c
2696
if (t == 0)
drivers/platform/x86/lenovo/thinkpad_acpi.c
2716
unsigned long t;
drivers/platform/x86/lenovo/thinkpad_acpi.c
2719
if (parse_strtoul(buf, 0xffffffffUL, &t))
drivers/platform/x86/lenovo/thinkpad_acpi.c
2725
res = hotkey_user_mask_set(t);
drivers/platform/x86/lenovo/thinkpad_acpi.c
2733
tpacpi_disclose_usertask("hotkey_mask", "set to 0x%08lx\n", t);
drivers/platform/x86/lenovo/thinkpad_acpi.c
2810
unsigned long t;
drivers/platform/x86/lenovo/thinkpad_acpi.c
2814
if (parse_strtoul(buf, 0xffffffffUL, &t) ||
drivers/platform/x86/lenovo/thinkpad_acpi.c
2815
((t & ~TPACPI_HKEY_NVRAM_KNOWN_MASK) != 0))
drivers/platform/x86/lenovo/thinkpad_acpi.c
2822
hotkey_source_mask = t;
drivers/platform/x86/lenovo/thinkpad_acpi.c
2842
tpacpi_disclose_usertask("hotkey_source_mask", "set to 0x%08lx\n", t);
drivers/platform/x86/lenovo/thinkpad_acpi.c
2861
unsigned long t;
drivers/platform/x86/lenovo/thinkpad_acpi.c
2863
if (parse_strtoul(buf, 25, &t))
drivers/platform/x86/lenovo/thinkpad_acpi.c
2869
hotkey_poll_set_freq(t);
drivers/platform/x86/lenovo/thinkpad_acpi.c
2874
tpacpi_disclose_usertask("hotkey_poll_freq", "set to %lu\n", t);
drivers/platform/x86/lenovo/thinkpad_acpi.c
2994
unsigned long t;
drivers/platform/x86/lenovo/thinkpad_acpi.c
2997
if (parse_strtoul(buf, LAYFLAT_MODE, &t))
drivers/platform/x86/lenovo/thinkpad_acpi.c
3000
res = adaptive_keyboard_set_mode(t);
drivers/platform/x86/lenovo/thinkpad_acpi.c
6042
u8 t, ta1, ta2, ver = 0;
drivers/platform/x86/lenovo/thinkpad_acpi.c
6074
if (acpi_ec_read(TP_EC_THERMAL_TMP0 + i, &t)) {
drivers/platform/x86/lenovo/thinkpad_acpi.c
6075
ta1 |= t;
drivers/platform/x86/lenovo/thinkpad_acpi.c
6081
if (acpi_ec_read(TP_EC_THERMAL_TMP8 + i, &t)) {
drivers/platform/x86/lenovo/thinkpad_acpi.c
6082
ta2 |= t;
drivers/platform/x86/lenovo/thinkpad_acpi.c
6124
int t;
drivers/platform/x86/lenovo/thinkpad_acpi.c
6128
t = TP_EC_THERMAL_TMP0;
drivers/platform/x86/lenovo/thinkpad_acpi.c
6134
t = TP_EC_THERMAL_TMP8;
drivers/platform/x86/lenovo/thinkpad_acpi.c
6141
if (!acpi_ec_read(t + idx, &tmp))
drivers/platform/x86/lenovo/thinkpad_acpi.c
6153
t = idx < 8 ? TP_EC_THERMAL_TMP0_NS + idx :
drivers/platform/x86/lenovo/thinkpad_acpi.c
6156
if (!acpi_ec_read(t, &tmp))
drivers/platform/x86/lenovo/thinkpad_acpi.c
6167
if (!acpi_evalf(ec_handle, &t, tmpi, "d"))
drivers/platform/x86/lenovo/thinkpad_acpi.c
6169
*value = (t - 2732) * 100;
drivers/platform/x86/lenovo/thinkpad_acpi.c
6177
if (!acpi_evalf(ec_handle, &t, tmpi, "d"))
drivers/platform/x86/lenovo/thinkpad_acpi.c
6179
if (t > 127 || t < -127)
drivers/platform/x86/lenovo/thinkpad_acpi.c
6180
t = TP_EC_THERMAL_TMP_NA;
drivers/platform/x86/lenovo/thinkpad_acpi.c
6181
*value = t * 1000;
drivers/platform/x86/lenovo/thinkpad_acpi.c
6220
struct ibm_thermal_sensors_struct t;
drivers/platform/x86/lenovo/thinkpad_acpi.c
6222
n = thermal_get_sensors(&t);
drivers/platform/x86/lenovo/thinkpad_acpi.c
6229
if (t.temp[i] != TPACPI_THERMAL_SENSOR_NA)
drivers/platform/x86/lenovo/thinkpad_acpi.c
6230
pr_cont(" %d", (int)(t.temp[i] / 1000));
drivers/platform/x86/lenovo/thinkpad_acpi.c
6395
struct ibm_thermal_sensors_struct t;
drivers/platform/x86/lenovo/thinkpad_acpi.c
6397
n = thermal_get_sensors(&t);
drivers/platform/x86/lenovo/thinkpad_acpi.c
6405
seq_printf(m, "%d ", t.temp[i] / 1000);
drivers/platform/x86/lenovo/thinkpad_acpi.c
6406
seq_printf(m, "%d\n", t.temp[i] / 1000);
drivers/platform/x86/lenovo/thinkpad_acpi.c
8614
unsigned long t;
drivers/platform/x86/lenovo/thinkpad_acpi.c
8617
if (parse_strtoul(buf, 2, &t))
drivers/platform/x86/lenovo/thinkpad_acpi.c
8621
"set fan mode to %lu\n", t);
drivers/platform/x86/lenovo/thinkpad_acpi.c
8623
switch (t) {
drivers/platform/x86/lenovo/thinkpad_acpi.c
8765
unsigned long t;
drivers/platform/x86/lenovo/thinkpad_acpi.c
8767
if (parse_strtoul(buf, 120, &t))
drivers/platform/x86/lenovo/thinkpad_acpi.c
8773
fan_watchdog_maxinterval = t;
drivers/platform/x86/lenovo/thinkpad_acpi.c
8776
tpacpi_disclose_usertask("fan_watchdog", "set to %lu\n", t);
drivers/platform/x86/lenovo/thinkpad_acpi.c
9346
static int mute_led_on_off(struct tp_led_table *t, bool state)
drivers/platform/x86/lenovo/thinkpad_acpi.c
9351
if (ACPI_FAILURE(acpi_get_handle(hkey_handle, t->name, &temp))) {
drivers/platform/x86/lenovo/thinkpad_acpi.c
9352
pr_warn("Thinkpad ACPI has no %s interface.\n", t->name);
drivers/platform/x86/lenovo/thinkpad_acpi.c
9356
if (!acpi_evalf(hkey_handle, &output, t->name, "dd",
drivers/platform/x86/lenovo/thinkpad_acpi.c
9357
state ? t->on_value : t->off_value))
drivers/platform/x86/lenovo/thinkpad_acpi.c
9360
t->state = state;
drivers/platform/x86/lenovo/thinkpad_acpi.c
9366
struct tp_led_table *t;
drivers/platform/x86/lenovo/thinkpad_acpi.c
9368
t = &led_tables[whichled];
drivers/platform/x86/lenovo/thinkpad_acpi.c
9369
if (t->state < 0 || t->state == on)
drivers/platform/x86/lenovo/thinkpad_acpi.c
9370
return t->state;
drivers/platform/x86/lenovo/thinkpad_acpi.c
9371
return mute_led_on_off(t, on);
drivers/platform/x86/lenovo/thinkpad_acpi.c
9407
struct tp_led_table *t = &led_tables[i];
drivers/platform/x86/lenovo/thinkpad_acpi.c
9408
if (ACPI_FAILURE(acpi_get_handle(hkey_handle, t->name, &temp))) {
drivers/platform/x86/lenovo/thinkpad_acpi.c
9409
t->state = -ENODEV;
drivers/platform/x86/lenovo/thinkpad_acpi.c
9438
struct tp_led_table *t = &led_tables[i];
drivers/platform/x86/lenovo/thinkpad_acpi.c
9439
if (t->state >= 0)
drivers/platform/x86/lenovo/thinkpad_acpi.c
9440
mute_led_on_off(t, t->state);
drivers/power/supply/ab8500_charger.c
1959
int t = 10;
drivers/power/supply/ab8500_charger.c
1990
t = 1;
drivers/power/supply/ab8500_charger.c
1992
queue_delayed_work(di->charger_wq, &di->check_vbat_work, t * HZ);
drivers/power/supply/bd71828-power.c
215
int t;
drivers/power/supply/bd71828-power.c
217
ret = regmap_read(regmap, pwr->regs->btemp_vth, &t);
drivers/power/supply/bd71828-power.c
221
t = 200 - t;
drivers/power/supply/bd71828-power.c
223
if (t > 200) {
drivers/power/supply/bd71828-power.c
234
u16 t;
drivers/power/supply/bd71828-power.c
239
BD71828_MASK_VM_BTMP_U, &t);
drivers/power/supply/bd71828-power.c
243
if (t > 3200) {
drivers/power/supply/bd71828-power.c
249
tmp -= 625ULL * (unsigned int)t;
drivers/power/supply/mm8013.c
40
#define DECIKELVIN_TO_DECIDEGC(t) (t - 2731)
drivers/powercap/intel_rapl_common.c
596
int t;
drivers/powercap/intel_rapl_common.c
617
for (t = POWER_LIMIT1; t < NR_POWER_LIMITS; t++) {
drivers/powercap/intel_rapl_common.c
618
if (rp->priv->limits[i] & BIT(t))
drivers/powercap/intel_rapl_common.c
619
rd->rpl[t].name = pl_names[t];
drivers/pps/clients/pps-gpio.c
101
info = timer_container_of(info, t, echo_timer);
drivers/pps/clients/pps-gpio.c
97
static void pps_gpio_echo_timer_callback(struct timer_list *t)
drivers/ptp/ptp_clock.c
72
dst->t.sec = seconds;
drivers/ptp/ptp_clock.c
73
dst->t.nsec = remainder;
drivers/ptp/ptp_ocp.c
1530
ptp_ocp_watchdog(struct timer_list *t)
drivers/ptp/ptp_ocp.c
1532
struct ptp_ocp *bp = timer_container_of(bp, t, watchdog);
drivers/ptp/ptp_sysfs.c
105
event.index, event.t.sec, event.t.nsec);
drivers/pwm/pwm-mc33xs2410.c
101
t.len = len * 2;
drivers/pwm/pwm-mc33xs2410.c
106
ret = spi_sync_transfer(spi, &t, 1);
drivers/pwm/pwm-mc33xs2410.c
91
struct spi_transfer t = {
drivers/pwm/pwm-vt8500.c
57
#define msecs_to_loops(t) (loops_per_jiffy / 1000 * HZ * t)
drivers/regulator/s2dos05-regulator.c
35
#define BUCK_DESC(_name, _id, _ops, m, s, v, e, em, t, a) { \
drivers/regulator/s2dos05-regulator.c
51
.enable_time = t, \
drivers/regulator/s2dos05-regulator.c
58
#define LDO_DESC(_name, _id, _ops, m, s, v, e, em, t, a) { \
drivers/regulator/s2dos05-regulator.c
74
.enable_time = t, \
drivers/regulator/tps6524x-regulator.c
148
struct spi_transfer t[3];
drivers/regulator/tps6524x-regulator.c
151
memset(t, 0, sizeof(t));
drivers/regulator/tps6524x-regulator.c
153
t[0].tx_buf = &cmd;
drivers/regulator/tps6524x-regulator.c
154
t[0].len = 2;
drivers/regulator/tps6524x-regulator.c
155
t[0].bits_per_word = 12;
drivers/regulator/tps6524x-regulator.c
156
spi_message_add_tail(&t[0], &m);
drivers/regulator/tps6524x-regulator.c
158
t[1].rx_buf = &in;
drivers/regulator/tps6524x-regulator.c
159
t[1].len = 2;
drivers/regulator/tps6524x-regulator.c
160
t[1].bits_per_word = 16;
drivers/regulator/tps6524x-regulator.c
161
spi_message_add_tail(&t[1], &m);
drivers/regulator/tps6524x-regulator.c
163
t[2].rx_buf = &status;
drivers/regulator/tps6524x-regulator.c
164
t[2].len = 1;
drivers/regulator/tps6524x-regulator.c
165
t[2].bits_per_word = 4;
drivers/regulator/tps6524x-regulator.c
166
spi_message_add_tail(&t[2], &m);
drivers/regulator/tps6524x-regulator.c
201
struct spi_transfer t[3];
drivers/regulator/tps6524x-regulator.c
204
memset(t, 0, sizeof(t));
drivers/regulator/tps6524x-regulator.c
206
t[0].tx_buf = &cmd;
drivers/regulator/tps6524x-regulator.c
207
t[0].len = 2;
drivers/regulator/tps6524x-regulator.c
208
t[0].bits_per_word = 12;
drivers/regulator/tps6524x-regulator.c
209
spi_message_add_tail(&t[0], &m);
drivers/regulator/tps6524x-regulator.c
211
t[1].tx_buf = &out;
drivers/regulator/tps6524x-regulator.c
212
t[1].len = 2;
drivers/regulator/tps6524x-regulator.c
213
t[1].bits_per_word = 16;
drivers/regulator/tps6524x-regulator.c
214
spi_message_add_tail(&t[1], &m);
drivers/regulator/tps6524x-regulator.c
216
t[2].rx_buf = &status;
drivers/regulator/tps6524x-regulator.c
217
t[2].len = 1;
drivers/regulator/tps6524x-regulator.c
218
t[2].bits_per_word = 4;
drivers/regulator/tps6524x-regulator.c
219
spi_message_add_tail(&t[2], &m);
drivers/remoteproc/remoteproc_debugfs.c
282
struct fw_rsc_trace *t;
drivers/remoteproc/remoteproc_debugfs.c
318
t = rsc;
drivers/remoteproc/remoteproc_debugfs.c
320
seq_printf(seq, " Device Address 0x%x\n", t->da);
drivers/remoteproc/remoteproc_debugfs.c
321
seq_printf(seq, " Length 0x%x Bytes\n", t->len);
drivers/remoteproc/remoteproc_debugfs.c
322
seq_printf(seq, " Reserved (should be zero) [%d]\n", t->reserved);
drivers/remoteproc/remoteproc_debugfs.c
323
seq_printf(seq, " Name %s\n\n", t->name);
drivers/rtc/dev.c
73
static void rtc_uie_timer(struct timer_list *t)
drivers/rtc/dev.c
75
struct rtc_device *rtc = timer_container_of(rtc, t, uie_timer);
drivers/rtc/rtc-abx80x.c
285
static int abx80x_read_alarm(struct device *dev, struct rtc_wkalrm *t)
drivers/rtc/rtc-abx80x.c
304
t->time.tm_sec = bcd2bin(buf[0] & 0x7F);
drivers/rtc/rtc-abx80x.c
305
t->time.tm_min = bcd2bin(buf[1] & 0x7F);
drivers/rtc/rtc-abx80x.c
306
t->time.tm_hour = bcd2bin(buf[2] & 0x3F);
drivers/rtc/rtc-abx80x.c
307
t->time.tm_mday = bcd2bin(buf[3] & 0x3F);
drivers/rtc/rtc-abx80x.c
308
t->time.tm_mon = bcd2bin(buf[4] & 0x1F) - 1;
drivers/rtc/rtc-abx80x.c
309
t->time.tm_wday = buf[5] & 0x7;
drivers/rtc/rtc-abx80x.c
311
t->enabled = !!(irq_mask & ABX8XX_IRQ_AIE);
drivers/rtc/rtc-abx80x.c
312
t->pending = (buf[6] & ABX8XX_STATUS_AF) && t->enabled;
drivers/rtc/rtc-abx80x.c
317
static int abx80x_set_alarm(struct device *dev, struct rtc_wkalrm *t)
drivers/rtc/rtc-abx80x.c
327
alarm[1] = bin2bcd(t->time.tm_sec);
drivers/rtc/rtc-abx80x.c
328
alarm[2] = bin2bcd(t->time.tm_min);
drivers/rtc/rtc-abx80x.c
329
alarm[3] = bin2bcd(t->time.tm_hour);
drivers/rtc/rtc-abx80x.c
330
alarm[4] = bin2bcd(t->time.tm_mday);
drivers/rtc/rtc-abx80x.c
331
alarm[5] = bin2bcd(t->time.tm_mon + 1);
drivers/rtc/rtc-abx80x.c
340
if (t->enabled) {
drivers/rtc/rtc-au1xxx.c
30
unsigned long t;
drivers/rtc/rtc-au1xxx.c
32
t = alchemy_rdsys(AU1000_SYS_TOYREAD);
drivers/rtc/rtc-au1xxx.c
34
rtc_time64_to_tm(t, tm);
drivers/rtc/rtc-au1xxx.c
41
unsigned long t;
drivers/rtc/rtc-au1xxx.c
43
t = rtc_tm_to_time64(tm);
drivers/rtc/rtc-au1xxx.c
45
alchemy_wrsys(t, AU1000_SYS_TOYWRITE);
drivers/rtc/rtc-au1xxx.c
64
unsigned long t;
drivers/rtc/rtc-au1xxx.c
66
t = alchemy_rdsys(AU1000_SYS_CNTRCTRL);
drivers/rtc/rtc-au1xxx.c
67
if (!(t & CNTR_OK)) {
drivers/rtc/rtc-au1xxx.c
75
t = 0x00100000;
drivers/rtc/rtc-au1xxx.c
76
while ((alchemy_rdsys(AU1000_SYS_CNTRCTRL) & SYS_CNTRL_T0S) && --t)
drivers/rtc/rtc-au1xxx.c
79
if (!t) {
drivers/rtc/rtc-bd70528.c
101
t->tm_hour += 12;
drivers/rtc/rtc-bd70528.c
103
t->tm_mday = bcd2bin(r->day & BD70528_MASK_RTC_DAY);
drivers/rtc/rtc-bd70528.c
104
t->tm_mon = bcd2bin(r->month & BD70528_MASK_RTC_MONTH) - 1;
drivers/rtc/rtc-bd70528.c
105
t->tm_year = 100 + bcd2bin(r->year & BD70528_MASK_RTC_YEAR);
drivers/rtc/rtc-bd70528.c
106
t->tm_wday = bcd2bin(r->week & BD70528_MASK_RTC_WEEK);
drivers/rtc/rtc-bd70528.c
161
static int bd71828_set_time(struct device *dev, struct rtc_time *t)
drivers/rtc/rtc-bd70528.c
173
tm2rtc(t, &rtc_data);
drivers/rtc/rtc-bd70528.c
183
static int bd70528_get_time(struct device *dev, struct rtc_time *t)
drivers/rtc/rtc-bd70528.c
197
rtc2tm(&rtc_data, t);
drivers/rtc/rtc-bd70528.c
307
struct rtc_time t;
drivers/rtc/rtc-bd70528.c
309
ret = rtc_ops->read_time(&pdev->dev, &t);
drivers/rtc/rtc-bd70528.c
312
ret = rtc_ops->set_time(&pdev->dev, &t);
drivers/rtc/rtc-bd70528.c
57
static inline void tmday2rtc(struct rtc_time *t, struct bd70528_rtc_day *d)
drivers/rtc/rtc-bd70528.c
62
d->sec |= bin2bcd(t->tm_sec);
drivers/rtc/rtc-bd70528.c
63
d->min |= bin2bcd(t->tm_min);
drivers/rtc/rtc-bd70528.c
64
d->hour |= bin2bcd(t->tm_hour);
drivers/rtc/rtc-bd70528.c
67
static inline void tm2rtc(struct rtc_time *t, struct bd70528_rtc_data *r)
drivers/rtc/rtc-bd70528.c
78
tmday2rtc(t, &r->time);
drivers/rtc/rtc-bd70528.c
83
r->day |= bin2bcd(t->tm_mday);
drivers/rtc/rtc-bd70528.c
84
r->week |= bin2bcd(t->tm_wday);
drivers/rtc/rtc-bd70528.c
85
r->month |= bin2bcd(t->tm_mon + 1);
drivers/rtc/rtc-bd70528.c
86
r->year = bin2bcd(t->tm_year - 100);
drivers/rtc/rtc-bd70528.c
89
static inline void rtc2tm(struct bd70528_rtc_data *r, struct rtc_time *t)
drivers/rtc/rtc-bd70528.c
91
t->tm_sec = bcd2bin(r->time.sec & BD70528_MASK_RTC_SEC);
drivers/rtc/rtc-bd70528.c
92
t->tm_min = bcd2bin(r->time.min & BD70528_MASK_RTC_MINUTE);
drivers/rtc/rtc-bd70528.c
93
t->tm_hour = bcd2bin(r->time.hour & BD70528_MASK_RTC_HOUR);
drivers/rtc/rtc-bd70528.c
99
t->tm_hour %= 12;
drivers/rtc/rtc-brcmstb-waketimer.c
135
struct wktmr_time *t)
drivers/rtc/rtc-brcmstb-waketimer.c
140
t->sec = readl_relaxed(timer->base + BRCMSTB_WKTMR_COUNTER);
drivers/rtc/rtc-brcmstb-waketimer.c
144
t->pre = timer->rate - tmp;
drivers/rtc/rtc-cmos.c
218
static int cmos_read_time(struct device *dev, struct rtc_time *t)
drivers/rtc/rtc-cmos.c
229
ret = mc146818_get_time(t, 1000);
drivers/rtc/rtc-cmos.c
238
static int cmos_set_time(struct device *dev, struct rtc_time *t)
drivers/rtc/rtc-cmos.c
244
return mc146818_set_time(t);
drivers/rtc/rtc-cmos.c
280
static int cmos_read_alarm(struct device *dev, struct rtc_wkalrm *t)
drivers/rtc/rtc-cmos.c
285
.time = &t->time,
drivers/rtc/rtc-cmos.c
309
if (((unsigned)t->time.tm_sec) < 0x60)
drivers/rtc/rtc-cmos.c
310
t->time.tm_sec = bcd2bin(t->time.tm_sec);
drivers/rtc/rtc-cmos.c
312
t->time.tm_sec = -1;
drivers/rtc/rtc-cmos.c
313
if (((unsigned)t->time.tm_min) < 0x60)
drivers/rtc/rtc-cmos.c
314
t->time.tm_min = bcd2bin(t->time.tm_min);
drivers/rtc/rtc-cmos.c
316
t->time.tm_min = -1;
drivers/rtc/rtc-cmos.c
317
if (((unsigned)t->time.tm_hour) < 0x24)
drivers/rtc/rtc-cmos.c
318
t->time.tm_hour = bcd2bin(t->time.tm_hour);
drivers/rtc/rtc-cmos.c
320
t->time.tm_hour = -1;
drivers/rtc/rtc-cmos.c
323
if (((unsigned)t->time.tm_mday) <= 0x31)
drivers/rtc/rtc-cmos.c
324
t->time.tm_mday = bcd2bin(t->time.tm_mday);
drivers/rtc/rtc-cmos.c
326
t->time.tm_mday = -1;
drivers/rtc/rtc-cmos.c
329
if (((unsigned)t->time.tm_mon) <= 0x12)
drivers/rtc/rtc-cmos.c
330
t->time.tm_mon = bcd2bin(t->time.tm_mon)-1;
drivers/rtc/rtc-cmos.c
332
t->time.tm_mon = -1;
drivers/rtc/rtc-cmos.c
337
t->enabled = !!(p.rtc_control & RTC_AIE);
drivers/rtc/rtc-cmos.c
338
t->pending = 0;
drivers/rtc/rtc-cmos.c
401
static int cmos_validate_alarm(struct device *dev, struct rtc_wkalrm *t)
drivers/rtc/rtc-cmos.c
414
t_alrm = rtc_tm_to_time64(&t->time);
drivers/rtc/rtc-cmos.c
438
t_alrm = rtc_tm_to_time64(&t->time);
drivers/rtc/rtc-cmos.c
457
t_alrm = rtc_tm_to_time64(&t->time);
drivers/rtc/rtc-cmos.c
471
struct rtc_wkalrm *t;
drivers/rtc/rtc-cmos.c
503
hpet_set_alarm_time(p->t->time.tm_hour, p->t->time.tm_min,
drivers/rtc/rtc-cmos.c
504
p->t->time.tm_sec);
drivers/rtc/rtc-cmos.c
507
if (p->t->enabled)
drivers/rtc/rtc-cmos.c
511
static int cmos_set_alarm(struct device *dev, struct rtc_wkalrm *t)
drivers/rtc/rtc-cmos.c
516
.t = t
drivers/rtc/rtc-cmos.c
525
ret = cmos_validate_alarm(dev, t);
drivers/rtc/rtc-cmos.c
529
p.mon = t->time.tm_mon + 1;
drivers/rtc/rtc-cmos.c
530
p.mday = t->time.tm_mday;
drivers/rtc/rtc-cmos.c
531
p.hrs = t->time.tm_hour;
drivers/rtc/rtc-cmos.c
532
p.min = t->time.tm_min;
drivers/rtc/rtc-cmos.c
533
p.sec = t->time.tm_sec;
drivers/rtc/rtc-cmos.c
557
cmos->alarm_expires = rtc_tm_to_time64(&t->time);
drivers/rtc/rtc-ds1307.c
210
static int ds1307_get_time(struct device *dev, struct rtc_time *t)
drivers/rtc/rtc-ds1307.c
305
t->tm_sec = bcd2bin(regs[DS1307_REG_SECS] & 0x7f);
drivers/rtc/rtc-ds1307.c
306
t->tm_min = bcd2bin(regs[DS1307_REG_MIN] & 0x7f);
drivers/rtc/rtc-ds1307.c
308
t->tm_hour = bcd2bin(tmp);
drivers/rtc/rtc-ds1307.c
311
t->tm_wday = fls(regs[DS1307_REG_WDAY] & 0x7f);
drivers/rtc/rtc-ds1307.c
313
t->tm_wday = bcd2bin(regs[DS1307_REG_WDAY] & 0x07) - 1;
drivers/rtc/rtc-ds1307.c
314
t->tm_mday = bcd2bin(regs[DS1307_REG_MDAY] & 0x3f);
drivers/rtc/rtc-ds1307.c
316
t->tm_mon = bcd2bin(tmp) - 1;
drivers/rtc/rtc-ds1307.c
317
t->tm_year = bcd2bin(regs[DS1307_REG_YEAR]) + 100;
drivers/rtc/rtc-ds1307.c
321
t->tm_year += 100;
drivers/rtc/rtc-ds1307.c
325
"read", t->tm_sec, t->tm_min,
drivers/rtc/rtc-ds1307.c
326
t->tm_hour, t->tm_mday,
drivers/rtc/rtc-ds1307.c
327
t->tm_mon, t->tm_year, t->tm_wday);
drivers/rtc/rtc-ds1307.c
332
static int ds1307_set_time(struct device *dev, struct rtc_time *t)
drivers/rtc/rtc-ds1307.c
342
"write", t->tm_sec, t->tm_min,
drivers/rtc/rtc-ds1307.c
343
t->tm_hour, t->tm_mday,
drivers/rtc/rtc-ds1307.c
344
t->tm_mon, t->tm_year, t->tm_wday);
drivers/rtc/rtc-ds1307.c
346
if (t->tm_year < 100)
drivers/rtc/rtc-ds1307.c
350
if (t->tm_year > (chip->century_bit ? 299 : 199))
drivers/rtc/rtc-ds1307.c
353
if (t->tm_year > 199)
drivers/rtc/rtc-ds1307.c
357
regs[DS1307_REG_SECS] = bin2bcd(t->tm_sec);
drivers/rtc/rtc-ds1307.c
358
regs[DS1307_REG_MIN] = bin2bcd(t->tm_min);
drivers/rtc/rtc-ds1307.c
359
regs[DS1307_REG_HOUR] = bin2bcd(t->tm_hour);
drivers/rtc/rtc-ds1307.c
362
regs[DS1307_REG_WDAY] = 1 << t->tm_wday;
drivers/rtc/rtc-ds1307.c
364
regs[DS1307_REG_WDAY] = bin2bcd(t->tm_wday + 1);
drivers/rtc/rtc-ds1307.c
365
regs[DS1307_REG_MDAY] = bin2bcd(t->tm_mday);
drivers/rtc/rtc-ds1307.c
366
regs[DS1307_REG_MONTH] = bin2bcd(t->tm_mon + 1);
drivers/rtc/rtc-ds1307.c
369
tmp = t->tm_year % 100;
drivers/rtc/rtc-ds1307.c
374
if (t->tm_year > 199 && chip->century_bit)
drivers/rtc/rtc-ds1307.c
430
static int ds1337_read_alarm(struct device *dev, struct rtc_wkalrm *t)
drivers/rtc/rtc-ds1307.c
451
t->time.tm_sec = bcd2bin(regs[0] & 0x7f);
drivers/rtc/rtc-ds1307.c
452
t->time.tm_min = bcd2bin(regs[1] & 0x7f);
drivers/rtc/rtc-ds1307.c
453
t->time.tm_hour = bcd2bin(regs[2] & 0x3f);
drivers/rtc/rtc-ds1307.c
454
t->time.tm_mday = bcd2bin(regs[3] & 0x3f);
drivers/rtc/rtc-ds1307.c
457
t->enabled = !!(regs[7] & DS1337_BIT_A1IE);
drivers/rtc/rtc-ds1307.c
458
t->pending = !!(regs[8] & DS1337_BIT_A1I);
drivers/rtc/rtc-ds1307.c
462
"alarm read", t->time.tm_sec, t->time.tm_min,
drivers/rtc/rtc-ds1307.c
463
t->time.tm_hour, t->time.tm_mday,
drivers/rtc/rtc-ds1307.c
464
t->enabled, t->pending);
drivers/rtc/rtc-ds1307.c
469
static int ds1337_set_alarm(struct device *dev, struct rtc_wkalrm *t)
drivers/rtc/rtc-ds1307.c
478
"alarm set", t->time.tm_sec, t->time.tm_min,
drivers/rtc/rtc-ds1307.c
479
t->time.tm_hour, t->time.tm_mday,
drivers/rtc/rtc-ds1307.c
480
t->enabled, t->pending);
drivers/rtc/rtc-ds1307.c
496
regs[0] = bin2bcd(t->time.tm_sec);
drivers/rtc/rtc-ds1307.c
497
regs[1] = bin2bcd(t->time.tm_min);
drivers/rtc/rtc-ds1307.c
498
regs[2] = bin2bcd(t->time.tm_hour);
drivers/rtc/rtc-ds1307.c
499
regs[3] = bin2bcd(t->time.tm_mday);
drivers/rtc/rtc-ds1307.c
518
if (t->enabled) {
drivers/rtc/rtc-ds1307.c
602
static int rx8130_read_alarm(struct device *dev, struct rtc_wkalrm *t)
drivers/rtc/rtc-ds1307.c
620
t->enabled = !!(ctl[2] & RX8130_REG_CONTROL0_AIE);
drivers/rtc/rtc-ds1307.c
621
t->pending = !!(ctl[1] & RX8130_REG_FLAG_AF);
drivers/rtc/rtc-ds1307.c
624
t->time.tm_sec = -1;
drivers/rtc/rtc-ds1307.c
625
t->time.tm_min = bcd2bin(ald[0] & 0x7f);
drivers/rtc/rtc-ds1307.c
626
t->time.tm_hour = bcd2bin(ald[1] & 0x7f);
drivers/rtc/rtc-ds1307.c
627
t->time.tm_wday = -1;
drivers/rtc/rtc-ds1307.c
628
t->time.tm_mday = bcd2bin(ald[2] & 0x7f);
drivers/rtc/rtc-ds1307.c
629
t->time.tm_mon = -1;
drivers/rtc/rtc-ds1307.c
630
t->time.tm_year = -1;
drivers/rtc/rtc-ds1307.c
631
t->time.tm_yday = -1;
drivers/rtc/rtc-ds1307.c
632
t->time.tm_isdst = -1;
drivers/rtc/rtc-ds1307.c
635
__func__, t->time.tm_sec, t->time.tm_min, t->time.tm_hour,
drivers/rtc/rtc-ds1307.c
636
t->time.tm_wday, t->time.tm_mday, t->time.tm_mon, t->enabled);
drivers/rtc/rtc-ds1307.c
641
static int rx8130_set_alarm(struct device *dev, struct rtc_wkalrm *t)
drivers/rtc/rtc-ds1307.c
649
t->time.tm_sec, t->time.tm_min, t->time.tm_hour,
drivers/rtc/rtc-ds1307.c
650
t->time.tm_wday, t->time.tm_mday, t->time.tm_mon,
drivers/rtc/rtc-ds1307.c
651
t->enabled, t->pending);
drivers/rtc/rtc-ds1307.c
669
ald[0] = bin2bcd(t->time.tm_min);
drivers/rtc/rtc-ds1307.c
670
ald[1] = bin2bcd(t->time.tm_hour);
drivers/rtc/rtc-ds1307.c
671
ald[2] = bin2bcd(t->time.tm_mday);
drivers/rtc/rtc-ds1307.c
678
if (!t->enabled)
drivers/rtc/rtc-ds1307.c
736
static int mcp794xx_read_alarm(struct device *dev, struct rtc_wkalrm *t)
drivers/rtc/rtc-ds1307.c
748
t->enabled = !!(regs[0] & MCP794XX_BIT_ALM0_EN);
drivers/rtc/rtc-ds1307.c
751
t->time.tm_sec = bcd2bin(regs[3] & 0x7f);
drivers/rtc/rtc-ds1307.c
752
t->time.tm_min = bcd2bin(regs[4] & 0x7f);
drivers/rtc/rtc-ds1307.c
753
t->time.tm_hour = bcd2bin(regs[5] & 0x3f);
drivers/rtc/rtc-ds1307.c
754
t->time.tm_wday = bcd2bin(regs[6] & 0x7) - 1;
drivers/rtc/rtc-ds1307.c
755
t->time.tm_mday = bcd2bin(regs[7] & 0x3f);
drivers/rtc/rtc-ds1307.c
756
t->time.tm_mon = bcd2bin(regs[8] & 0x1f) - 1;
drivers/rtc/rtc-ds1307.c
757
t->time.tm_year = -1;
drivers/rtc/rtc-ds1307.c
758
t->time.tm_yday = -1;
drivers/rtc/rtc-ds1307.c
759
t->time.tm_isdst = -1;
drivers/rtc/rtc-ds1307.c
763
t->time.tm_sec, t->time.tm_min, t->time.tm_hour,
drivers/rtc/rtc-ds1307.c
764
t->time.tm_wday, t->time.tm_mday, t->time.tm_mon, t->enabled,
drivers/rtc/rtc-ds1307.c
791
static int mcp794xx_set_alarm(struct device *dev, struct rtc_wkalrm *t)
drivers/rtc/rtc-ds1307.c
797
wday = mcp794xx_alm_weekday(dev, &t->time);
drivers/rtc/rtc-ds1307.c
803
t->time.tm_sec, t->time.tm_min, t->time.tm_hour,
drivers/rtc/rtc-ds1307.c
804
t->time.tm_wday, t->time.tm_mday, t->time.tm_mon,
drivers/rtc/rtc-ds1307.c
805
t->enabled, t->pending);
drivers/rtc/rtc-ds1307.c
814
regs[3] = bin2bcd(t->time.tm_sec);
drivers/rtc/rtc-ds1307.c
815
regs[4] = bin2bcd(t->time.tm_min);
drivers/rtc/rtc-ds1307.c
816
regs[5] = bin2bcd(t->time.tm_hour);
drivers/rtc/rtc-ds1307.c
818
regs[7] = bin2bcd(t->time.tm_mday);
drivers/rtc/rtc-ds1307.c
819
regs[8] = bin2bcd(t->time.tm_mon + 1);
drivers/rtc/rtc-ds1307.c
833
if (!t->enabled)
drivers/rtc/rtc-fm3130.c
117
t->tm_sec = bcd2bin(fm3130->regs[FM3130_RTC_SECONDS] & 0x7f);
drivers/rtc/rtc-fm3130.c
118
t->tm_min = bcd2bin(fm3130->regs[FM3130_RTC_MINUTES] & 0x7f);
drivers/rtc/rtc-fm3130.c
120
t->tm_hour = bcd2bin(tmp);
drivers/rtc/rtc-fm3130.c
121
t->tm_wday = bcd2bin(fm3130->regs[FM3130_RTC_DAY] & 0x07) - 1;
drivers/rtc/rtc-fm3130.c
122
t->tm_mday = bcd2bin(fm3130->regs[FM3130_RTC_DATE] & 0x3f);
drivers/rtc/rtc-fm3130.c
124
t->tm_mon = bcd2bin(tmp) - 1;
drivers/rtc/rtc-fm3130.c
127
t->tm_year = bcd2bin(fm3130->regs[FM3130_RTC_YEARS]) + 100;
drivers/rtc/rtc-fm3130.c
131
"read", t->tm_sec, t->tm_min,
drivers/rtc/rtc-fm3130.c
132
t->tm_hour, t->tm_mday,
drivers/rtc/rtc-fm3130.c
133
t->tm_mon, t->tm_year, t->tm_wday);
drivers/rtc/rtc-fm3130.c
139
static int fm3130_set_time(struct device *dev, struct rtc_time *t)
drivers/rtc/rtc-fm3130.c
147
"write", t->tm_sec, t->tm_min,
drivers/rtc/rtc-fm3130.c
148
t->tm_hour, t->tm_mday,
drivers/rtc/rtc-fm3130.c
149
t->tm_mon, t->tm_year, t->tm_wday);
drivers/rtc/rtc-fm3130.c
152
buf[FM3130_RTC_SECONDS] = bin2bcd(t->tm_sec);
drivers/rtc/rtc-fm3130.c
153
buf[FM3130_RTC_MINUTES] = bin2bcd(t->tm_min);
drivers/rtc/rtc-fm3130.c
154
buf[FM3130_RTC_HOURS] = bin2bcd(t->tm_hour);
drivers/rtc/rtc-fm3130.c
155
buf[FM3130_RTC_DAY] = bin2bcd(t->tm_wday + 1);
drivers/rtc/rtc-fm3130.c
156
buf[FM3130_RTC_DATE] = bin2bcd(t->tm_mday);
drivers/rtc/rtc-fm3130.c
157
buf[FM3130_RTC_MONTHS] = bin2bcd(t->tm_mon + 1);
drivers/rtc/rtc-fm3130.c
160
tmp = t->tm_year - 100;
drivers/rtc/rtc-fm3130.c
92
static int fm3130_get_time(struct device *dev, struct rtc_time *t)
drivers/rtc/rtc-gamecube.c
161
static int gamecube_rtc_read_time(struct device *dev, struct rtc_time *t)
drivers/rtc/rtc-gamecube.c
174
rtc_time64_to_tm(timestamp, t);
drivers/rtc/rtc-gamecube.c
179
static int gamecube_rtc_set_time(struct device *dev, struct rtc_time *t)
drivers/rtc/rtc-gamecube.c
185
timestamp = rtc_tm_to_time64(t);
drivers/rtc/rtc-rs5c372.c
340
int t = tmp & 0x3f;
drivers/rtc/rtc-rs5c372.c
343
t = (~t | (s8)0xc0) + 1;
drivers/rtc/rtc-rs5c372.c
345
t = t - 1;
drivers/rtc/rtc-rs5c372.c
347
tmp = t * 2;
drivers/rtc/rtc-rs5c372.c
398
static int rs5c_read_alarm(struct device *dev, struct rtc_wkalrm *t)
drivers/rtc/rtc-rs5c372.c
409
t->time.tm_sec = 0;
drivers/rtc/rtc-rs5c372.c
410
t->time.tm_min = bcd2bin(rs5c->regs[RS5C_REG_ALARM_A_MIN] & 0x7f);
drivers/rtc/rtc-rs5c372.c
411
t->time.tm_hour = rs5c_reg2hr(rs5c, rs5c->regs[RS5C_REG_ALARM_A_HOURS]);
drivers/rtc/rtc-rs5c372.c
414
t->enabled = !!(rs5c->regs[RS5C_REG_CTRL1] & RS5C_CTRL1_AALE);
drivers/rtc/rtc-rs5c372.c
415
t->pending = !!(rs5c->regs[RS5C_REG_CTRL2] & RS5C_CTRL2_AAFG);
drivers/rtc/rtc-rs5c372.c
420
static int rs5c_set_alarm(struct device *dev, struct rtc_wkalrm *t)
drivers/rtc/rtc-rs5c372.c
428
if (t->time.tm_mday != -1
drivers/rtc/rtc-rs5c372.c
429
|| t->time.tm_mon != -1
drivers/rtc/rtc-rs5c372.c
430
|| t->time.tm_year != -1)
drivers/rtc/rtc-rs5c372.c
450
buf[0] = bin2bcd(t->time.tm_min);
drivers/rtc/rtc-rs5c372.c
451
buf[1] = rs5c_hr2reg(rs5c, t->time.tm_hour);
drivers/rtc/rtc-rs5c372.c
463
if (t->enabled) {
drivers/rtc/rtc-rx8010.c
225
static int rx8010_read_alarm(struct device *dev, struct rtc_wkalrm *t)
drivers/rtc/rtc-rx8010.c
239
t->time.tm_sec = 0;
drivers/rtc/rtc-rx8010.c
240
t->time.tm_min = bcd2bin(alarmvals[0] & 0x7f);
drivers/rtc/rtc-rx8010.c
241
t->time.tm_hour = bcd2bin(alarmvals[1] & 0x3f);
drivers/rtc/rtc-rx8010.c
244
t->time.tm_mday = bcd2bin(alarmvals[2] & 0x7f);
drivers/rtc/rtc-rx8010.c
246
t->enabled = !!(rx8010->ctrlreg & RX8010_CTRL_AIE);
drivers/rtc/rtc-rx8010.c
247
t->pending = (flagreg & RX8010_FLAG_AF) && t->enabled;
drivers/rtc/rtc-rx8010.c
252
static int rx8010_set_alarm(struct device *dev, struct rtc_wkalrm *t)
drivers/rtc/rtc-rx8010.c
269
alarmvals[0] = bin2bcd(t->time.tm_min);
drivers/rtc/rtc-rx8010.c
270
alarmvals[1] = bin2bcd(t->time.tm_hour);
drivers/rtc/rtc-rx8010.c
271
alarmvals[2] = bin2bcd(t->time.tm_mday);
drivers/rtc/rtc-rx8010.c
288
if (t->enabled) {
drivers/rtc/rtc-rx8025.c
326
static int rx8025_read_alarm(struct device *dev, struct rtc_wkalrm *t)
drivers/rtc/rtc-rx8025.c
345
t->time.tm_sec = 0;
drivers/rtc/rtc-rx8025.c
346
t->time.tm_min = bcd2bin(ald[0] & 0x7f);
drivers/rtc/rtc-rx8025.c
348
t->time.tm_hour = bcd2bin(ald[1] & 0x3f);
drivers/rtc/rtc-rx8025.c
350
t->time.tm_hour = bcd2bin(ald[1] & 0x1f) % 12
drivers/rtc/rtc-rx8025.c
353
dev_dbg(dev, "%s: date: %ptRr\n", __func__, &t->time);
drivers/rtc/rtc-rx8025.c
354
t->enabled = !!(rx8025->ctrl1 & RX8025_BIT_CTRL1_DALE);
drivers/rtc/rtc-rx8025.c
355
t->pending = (ctrl2 & RX8025_BIT_CTRL2_DAFG) && t->enabled;
drivers/rtc/rtc-rx8025.c
360
static int rx8025_set_alarm(struct device *dev, struct rtc_wkalrm *t)
drivers/rtc/rtc-rx8025.c
367
ald[0] = bin2bcd(t->time.tm_min);
drivers/rtc/rtc-rx8025.c
369
ald[1] = bin2bcd(t->time.tm_hour);
drivers/rtc/rtc-rx8025.c
371
ald[1] = (t->time.tm_hour >= 12 ? 0x20 : 0)
drivers/rtc/rtc-rx8025.c
372
| bin2bcd((t->time.tm_hour + 11) % 12 + 1);
drivers/rtc/rtc-rx8025.c
387
if (t->enabled) {
drivers/rtc/rtc-spacemit-p1.c
100
time[0] = t->tm_sec;
drivers/rtc/rtc-spacemit-p1.c
101
time[1] = t->tm_min;
drivers/rtc/rtc-spacemit-p1.c
102
time[2] = t->tm_hour;
drivers/rtc/rtc-spacemit-p1.c
103
time[3] = t->tm_mday - 1;
drivers/rtc/rtc-spacemit-p1.c
104
time[4] = t->tm_mon;
drivers/rtc/rtc-spacemit-p1.c
105
time[5] = t->tm_year - 100;
drivers/rtc/rtc-spacemit-p1.c
51
static int p1_rtc_read_time(struct device *dev, struct rtc_time *t)
drivers/rtc/rtc-spacemit-p1.c
77
t->tm_sec = time[0] & GENMASK(5, 0);
drivers/rtc/rtc-spacemit-p1.c
78
t->tm_min = time[1] & GENMASK(5, 0);
drivers/rtc/rtc-spacemit-p1.c
79
t->tm_hour = time[2] & GENMASK(4, 0);
drivers/rtc/rtc-spacemit-p1.c
80
t->tm_mday = (time[3] & GENMASK(4, 0)) + 1;
drivers/rtc/rtc-spacemit-p1.c
81
t->tm_mon = time[4] & GENMASK(3, 0);
drivers/rtc/rtc-spacemit-p1.c
82
t->tm_year = (time[5] & GENMASK(5, 0)) + 100;
drivers/rtc/rtc-spacemit-p1.c
93
static int p1_rtc_set_time(struct device *dev, struct rtc_time *t)
drivers/rtc/rtc-st-lpc.c
152
static int st_rtc_set_alarm(struct device *dev, struct rtc_wkalrm *t)
drivers/rtc/rtc-st-lpc.c
162
alarm_secs = rtc_tm_to_time64(&t->time);
drivers/rtc/rtc-st-lpc.c
164
memcpy(&rtc->alarm, t, sizeof(struct rtc_wkalrm));
drivers/rtc/rtc-st-lpc.c
171
st_rtc_alarm_irq_enable(dev, t->enabled);
drivers/rtc/rtc-test.c
108
static void test_rtc_alarm_handler(struct timer_list *t)
drivers/rtc/rtc-test.c
110
struct rtc_test_data *rtd = timer_container_of(rtd, t, alarm);
drivers/s390/block/dasd.c
1448
static void dasd_device_timeout(struct timer_list *t)
drivers/s390/block/dasd.c
1453
device = timer_container_of(device, t, timer);
drivers/s390/block/dasd.c
2632
static void dasd_block_timeout(struct timer_list *t)
drivers/s390/block/dasd.c
2637
block = timer_container_of(block, t, timer);
drivers/s390/char/con3215.c
285
static void raw3215_timeout(struct timer_list *t)
drivers/s390/char/con3215.c
287
struct raw3215_info *raw = timer_container_of(raw, t, timer);
drivers/s390/char/con3270.c
525
static void tty3270_update(struct timer_list *t)
drivers/s390/char/con3270.c
527
struct tty3270 *tp = timer_container_of(tp, t, timer);
drivers/s390/char/sclp.c
852
struct sclp_register *t;
drivers/s390/char/sclp.c
857
t = list_entry(l, struct sclp_register, list);
drivers/s390/char/sclp.c
858
*receive_mask |= t->receive_mask;
drivers/s390/char/sclp.c
859
*send_mask |= t->send_mask;
drivers/s390/char/sclp_tty.c
287
struct sclp_buffer *t;
drivers/s390/char/sclp_tty.c
293
list_for_each_entry(t, &sclp_tty_outqueue, list) {
drivers/s390/char/sclp_tty.c
294
count += sclp_chars_in_buffer(t);
drivers/s390/char/tape_core.c
37
static void tape_long_busy_timeout(struct timer_list *t);
drivers/s390/char/tape_core.c
851
static void tape_long_busy_timeout(struct timer_list *t)
drivers/s390/char/tape_core.c
853
struct tape_device *device = timer_container_of(device, t, lb_timeout);
drivers/s390/char/tape_std.c
35
tape_std_assign_timeout(struct timer_list *t)
drivers/s390/char/tape_std.c
37
struct tape_request * request = timer_container_of(request, t,
drivers/s390/cio/chp.c
753
enum cfg_task_t t = cfg_none;
drivers/s390/cio/chp.c
756
t = cfg_get_task(*chpid);
drivers/s390/cio/chp.c
757
if (t != cfg_none)
drivers/s390/cio/chp.c
761
return t;
drivers/s390/cio/chp.c
769
enum cfg_task_t t;
drivers/s390/cio/chp.c
773
t = chp_cfg_fetch_task(&chpid);
drivers/s390/cio/chp.c
776
switch (t) {
drivers/s390/cio/chp.c
804
if (t == cfg_get_task(chpid))
drivers/s390/cio/chp.c
846
enum cfg_task_t t;
drivers/s390/cio/chp.c
849
t = chp_cfg_fetch_task(&chpid);
drivers/s390/cio/chp.c
852
return t == cfg_none;
drivers/s390/cio/device.h
137
void ccw_device_timeout(struct timer_list *t);
drivers/s390/cio/device_fsm.c
100
ccw_device_timeout(struct timer_list *t)
drivers/s390/cio/device_fsm.c
102
struct ccw_device_private *priv = timer_container_of(priv, t, timer);
drivers/s390/cio/eadm_sch.c
100
static void eadm_subchannel_timeout(struct timer_list *t)
drivers/s390/cio/eadm_sch.c
102
struct eadm_private *private = timer_container_of(private, t, timer);
drivers/s390/crypto/ap_bus.c
439
void ap_request_timeout(struct timer_list *t)
drivers/s390/crypto/ap_bus.c
441
struct ap_queue *aq = timer_container_of(aq, t, timeout);
drivers/s390/crypto/pkey_api.c
171
struct clearkeytoken *t;
drivers/s390/crypto/pkey_api.c
187
tmpbuf = kzalloc(sizeof(*t) + keylen, GFP_KERNEL);
drivers/s390/crypto/pkey_api.c
192
t = (struct clearkeytoken *)tmpbuf;
drivers/s390/crypto/pkey_api.c
193
t->type = TOKTYPE_NON_CCA;
drivers/s390/crypto/pkey_api.c
194
t->version = TOKVER_CLEAR_KEY;
drivers/s390/crypto/pkey_api.c
195
t->keytype = (keylen - 8) >> 3;
drivers/s390/crypto/pkey_api.c
196
t->len = keylen;
drivers/s390/crypto/pkey_api.c
197
memcpy(t->clearkey, kcp.clrkey.clrkey, keylen);
drivers/s390/crypto/pkey_api.c
201
tmpbuf, sizeof(*t) + keylen,
drivers/s390/crypto/pkey_api.c
329
struct protaeskeytoken *t;
drivers/s390/crypto/pkey_api.c
346
tmpbuf = kzalloc(sizeof(*t), GFP_KERNEL);
drivers/s390/crypto/pkey_api.c
351
t = (struct protaeskeytoken *)tmpbuf;
drivers/s390/crypto/pkey_api.c
352
t->type = TOKTYPE_NON_CCA;
drivers/s390/crypto/pkey_api.c
353
t->version = TOKVER_PROTECTED_KEY;
drivers/s390/crypto/pkey_api.c
354
t->keytype = keytype;
drivers/s390/crypto/pkey_api.c
355
t->len = kvp.protkey.len;
drivers/s390/crypto/pkey_api.c
356
memcpy(t->protkey, kvp.protkey.protkey, kvp.protkey.len);
drivers/s390/crypto/pkey_api.c
358
rc = pkey_handler_verify_key(tmpbuf, sizeof(*t),
drivers/s390/crypto/pkey_cca.c
101
struct cipherkeytoken *t = (struct cipherkeytoken *)key;
drivers/s390/crypto/pkey_cca.c
105
cur_mkvp = t->mkvp0;
drivers/s390/crypto/pkey_cca.c
107
old_mkvp = t->mkvp0;
drivers/s390/crypto/pkey_cca.c
119
struct eccprivkeytoken *t = (struct eccprivkeytoken *)key;
drivers/s390/crypto/pkey_cca.c
122
if (t->secid == 0x20) {
drivers/s390/crypto/pkey_cca.c
124
cur_mkvp = t->mkvp;
drivers/s390/crypto/pkey_cca.c
126
old_mkvp = t->mkvp;
drivers/s390/crypto/pkey_cca.c
481
struct secaeskeytoken *t = (struct secaeskeytoken *)key;
drivers/s390/crypto/pkey_cca.c
487
*keybitsize = t->bitsize;
drivers/s390/crypto/pkey_cca.c
490
t->mkvp, 0, xflags);
drivers/s390/crypto/pkey_cca.c
497
0, t->mkvp, xflags);
drivers/s390/crypto/pkey_cca.c
509
struct cipherkeytoken *t = (struct cipherkeytoken *)key;
drivers/s390/crypto/pkey_cca.c
516
if (!t->plfver && t->wpllen == 512)
drivers/s390/crypto/pkey_cca.c
518
else if (!t->plfver && t->wpllen == 576)
drivers/s390/crypto/pkey_cca.c
520
else if (!t->plfver && t->wpllen == 640)
drivers/s390/crypto/pkey_cca.c
524
t->mkvp0, 0, xflags);
drivers/s390/crypto/pkey_cca.c
531
0, t->mkvp0, xflags);
drivers/s390/crypto/pkey_cca.c
565
const struct clearkeytoken *t = (const struct clearkeytoken *)key;
drivers/s390/crypto/pkey_cca.c
575
keysize = pkey_keytype_aes_to_size(t->keytype);
drivers/s390/crypto/pkey_cca.c
576
if (!keysize || t->len != keysize)
drivers/s390/crypto/pkey_cca.c
582
rc = cca_clr2key(NULL, 0, t->keytype, PKEY_TYPE_CCA_DATA,
drivers/s390/crypto/pkey_cca.c
583
8 * keysize, 0, t->clearkey, t->len,
drivers/s390/crypto/pkey_cca.c
94
struct secaeskeytoken *t = (struct secaeskeytoken *)key;
drivers/s390/crypto/pkey_cca.c
97
cur_mkvp = t->mkvp;
drivers/s390/crypto/pkey_cca.c
99
old_mkvp = t->mkvp;
drivers/s390/crypto/pkey_ep11.c
511
const struct clearkeytoken *t = (const struct clearkeytoken *)key;
drivers/s390/crypto/pkey_ep11.c
521
keysize = pkey_keytype_aes_to_size(t->keytype);
drivers/s390/crypto/pkey_ep11.c
522
if (!keysize || t->len != keysize)
drivers/s390/crypto/pkey_ep11.c
528
rc = ep11_clr2key(NULL, 0, t->keytype, PKEY_TYPE_EP11,
drivers/s390/crypto/pkey_ep11.c
529
8 * keysize, 0, t->clearkey, t->len,
drivers/s390/crypto/pkey_pckmo.c
231
struct protkeytoken *t = (struct protkeytoken *)key;
drivers/s390/crypto/pkey_pckmo.c
234
if (keylen < sizeof(*t))
drivers/s390/crypto/pkey_pckmo.c
236
keysize = pkey_keytype_to_size(t->keytype);
drivers/s390/crypto/pkey_pckmo.c
239
__func__, t->keytype);
drivers/s390/crypto/pkey_pckmo.c
242
switch (t->keytype) {
drivers/s390/crypto/pkey_pckmo.c
246
if (t->len != keysize + AES_WK_VP_SIZE ||
drivers/s390/crypto/pkey_pckmo.c
249
rc = pckmo_verify_protkey(t->protkey, t->len,
drivers/s390/crypto/pkey_pckmo.c
250
t->keytype);
drivers/s390/crypto/pkey_pckmo.c
255
if (t->len != keysize + AES_WK_VP_SIZE ||
drivers/s390/crypto/pkey_pckmo.c
256
keylen < sizeof(*t) + keysize + AES_WK_VP_SIZE)
drivers/s390/crypto/pkey_pckmo.c
260
memcpy(protkey, t->protkey, t->len);
drivers/s390/crypto/pkey_pckmo.c
261
*protkeylen = t->len;
drivers/s390/crypto/pkey_pckmo.c
262
*protkeytype = t->keytype;
drivers/s390/crypto/pkey_pckmo.c
267
struct clearkeytoken *t = (struct clearkeytoken *)key;
drivers/s390/crypto/pkey_pckmo.c
275
if (keylen < sizeof(*t) ||
drivers/s390/crypto/pkey_pckmo.c
276
keylen < sizeof(*t) + t->len)
drivers/s390/crypto/pkey_pckmo.c
278
keysize = pkey_keytype_to_size(t->keytype);
drivers/s390/crypto/pkey_pckmo.c
281
__func__, t->keytype);
drivers/s390/crypto/pkey_pckmo.c
284
if (t->len != keysize) {
drivers/s390/crypto/pkey_pckmo.c
286
__func__, t->len);
drivers/s390/crypto/pkey_pckmo.c
289
rc = pckmo_clr2protkey(t->keytype, t->clearkey, t->len,
drivers/s390/crypto/pkey_pckmo.c
30
struct clearkeytoken *t = (struct clearkeytoken *)key;
drivers/s390/crypto/pkey_pckmo.c
373
struct protkeytoken *t = (struct protkeytoken *)key;
drivers/s390/crypto/pkey_pckmo.c
376
if (keylen < sizeof(*t))
drivers/s390/crypto/pkey_pckmo.c
378
keysize = pkey_keytype_to_size(t->keytype);
drivers/s390/crypto/pkey_pckmo.c
379
if (!keysize || t->len != keysize + AES_WK_VP_SIZE)
drivers/s390/crypto/pkey_pckmo.c
381
switch (t->keytype) {
drivers/s390/crypto/pkey_pckmo.c
389
if (keylen < sizeof(*t) + keysize + AES_WK_VP_SIZE)
drivers/s390/crypto/pkey_pckmo.c
39
if (pkey_keytype_to_size(t->keytype))
drivers/s390/crypto/pkey_pckmo.c
393
rc = pckmo_verify_protkey(t->protkey, t->len, t->keytype);
drivers/s390/crypto/pkey_sysfs.c
109
struct protkeytoken *t = (struct protkeytoken *)buf;
drivers/s390/crypto/pkey_sysfs.c
124
if (off != 0 || count < sizeof(*t) + protlen)
drivers/s390/crypto/pkey_sysfs.c
127
memset(t, 0, sizeof(*t) + protlen);
drivers/s390/crypto/pkey_sysfs.c
128
t->type = TOKTYPE_NON_CCA;
drivers/s390/crypto/pkey_sysfs.c
129
t->version = TOKVER_PROTECTED_KEY;
drivers/s390/crypto/pkey_sysfs.c
130
t->keytype = keytype;
drivers/s390/crypto/pkey_sysfs.c
133
t->protkey, &protlen, &prottype);
drivers/s390/crypto/pkey_sysfs.c
137
t->len = protlen;
drivers/s390/crypto/pkey_sysfs.c
139
return sizeof(*t) + protlen;
drivers/s390/crypto/pkey_sysfs.c
151
struct protkeytoken *t = (struct protkeytoken *)buf;
drivers/s390/crypto/pkey_sysfs.c
166
if (off != 0 || count < sizeof(*t) + protlen)
drivers/s390/crypto/pkey_sysfs.c
169
memset(t, 0, sizeof(*t) + protlen);
drivers/s390/crypto/pkey_sysfs.c
170
t->type = TOKTYPE_NON_CCA;
drivers/s390/crypto/pkey_sysfs.c
171
t->version = TOKVER_PROTECTED_KEY;
drivers/s390/crypto/pkey_sysfs.c
172
t->keytype = keytype;
drivers/s390/crypto/pkey_sysfs.c
175
t->protkey, &protlen, &prottype);
drivers/s390/crypto/pkey_sysfs.c
179
t->len = protlen;
drivers/s390/crypto/pkey_sysfs.c
181
return sizeof(*t) + protlen;
drivers/s390/crypto/pkey_uv.c
196
struct uvsecrettoken *t = (struct uvsecrettoken *)key;
drivers/s390/crypto/pkey_uv.c
201
rc = uv_get_size_and_type(t->secret_type, &pkeysize, &pkeytype);
drivers/s390/crypto/pkey_uv.c
212
rc = retrieve_secret(t->secret_id, &secret_type, protkey, protkeylen);
drivers/s390/crypto/pkey_uv.c
218
if (secret_type != t->secret_type) {
drivers/s390/crypto/pkey_uv.c
220
__func__, secret_type, t->secret_type);
drivers/s390/crypto/pkey_uv.c
239
struct uvsecrettoken *t = (struct uvsecrettoken *)key;
drivers/s390/crypto/pkey_uv.c
244
rc = uv_get_size_and_type(t->secret_type, &pkeysize, &pkeytype);
drivers/s390/crypto/pkey_uv.c
248
rc = get_secret_metadata(t->secret_id, &secret_meta_data);
drivers/s390/crypto/pkey_uv.c
252
if (secret_meta_data.type != t->secret_type) {
drivers/s390/crypto/pkey_uv.c
49
struct uvsecrettoken *t = (struct uvsecrettoken *)key;
drivers/s390/crypto/pkey_uv.c
51
if (keylen < sizeof(*t))
drivers/s390/crypto/pkey_uv.c
54
switch (t->type) {
drivers/s390/crypto/pkey_uv.c
56
switch (t->version) {
drivers/s390/crypto/pkey_uv.c
58
switch (t->secret_type) {
drivers/s390/crypto/zcrypt_ccamisc.c
1021
struct cipherkeytoken *t;
drivers/s390/crypto/zcrypt_ccamisc.c
107
struct cipherkeytoken *t = (struct cipherkeytoken *)token;
drivers/s390/crypto/zcrypt_ccamisc.c
1109
t = (struct cipherkeytoken *)prepparm->kb.tlv1.key_token;
drivers/s390/crypto/zcrypt_ccamisc.c
1110
memcpy(key_token, t, t->len);
drivers/s390/crypto/zcrypt_ccamisc.c
1111
*key_token_size = t->len;
drivers/s390/crypto/zcrypt_ccamisc.c
112
if (t->type != TOKTYPE_CCA_INTERNAL) {
drivers/s390/crypto/zcrypt_ccamisc.c
1128
struct cipherkeytoken *t;
drivers/s390/crypto/zcrypt_ccamisc.c
115
__func__, (int)t->type, TOKTYPE_CCA_INTERNAL);
drivers/s390/crypto/zcrypt_ccamisc.c
1152
t = (struct cipherkeytoken *)token;
drivers/s390/crypto/zcrypt_ccamisc.c
1153
t->kmf1 |= (u16)(keygenflags & 0x0000FF00);
drivers/s390/crypto/zcrypt_ccamisc.c
1154
t->kmf1 &= (u16)~(keygenflags & 0x000000FF);
drivers/s390/crypto/zcrypt_ccamisc.c
118
if (t->version != TOKVER_CCA_VLSC) {
drivers/s390/crypto/zcrypt_ccamisc.c
121
__func__, (int)t->version, TOKVER_CCA_VLSC);
drivers/s390/crypto/zcrypt_ccamisc.c
124
if (t->algtype != 0x02) {
drivers/s390/crypto/zcrypt_ccamisc.c
127
__func__, (int)t->algtype);
drivers/s390/crypto/zcrypt_ccamisc.c
130
if (t->keytype != 0x0001) {
drivers/s390/crypto/zcrypt_ccamisc.c
133
__func__, (int)t->keytype);
drivers/s390/crypto/zcrypt_ccamisc.c
136
if (t->plfver != 0x00 && t->plfver != 0x01) {
drivers/s390/crypto/zcrypt_ccamisc.c
139
__func__, (int)t->plfver);
drivers/s390/crypto/zcrypt_ccamisc.c
142
if (t->wpllen != 512 && t->wpllen != 576 && t->wpllen != 640) {
drivers/s390/crypto/zcrypt_ccamisc.c
145
__func__, (int)t->wpllen);
drivers/s390/crypto/zcrypt_ccamisc.c
151
if (t->wpllen != (t->plfver ? 640 : 512))
drivers/s390/crypto/zcrypt_ccamisc.c
155
if (t->wpllen != (t->plfver ? 640 : 576))
drivers/s390/crypto/zcrypt_ccamisc.c
159
if (t->wpllen != 640)
drivers/s390/crypto/zcrypt_ccamisc.c
173
if (checkcpacfexport && !(t->kmf1 & KMF1_XPRT_CPAC)) {
drivers/s390/crypto/zcrypt_ccamisc.c
194
struct eccprivkeytoken *t = (struct eccprivkeytoken *)token;
drivers/s390/crypto/zcrypt_ccamisc.c
198
if (t->type != TOKTYPE_CCA_INTERNAL_PKA) {
drivers/s390/crypto/zcrypt_ccamisc.c
201
__func__, (int)t->type, TOKTYPE_CCA_INTERNAL_PKA);
drivers/s390/crypto/zcrypt_ccamisc.c
204
if (t->len > keysize) {
drivers/s390/crypto/zcrypt_ccamisc.c
207
__func__, (int)t->len, keysize);
drivers/s390/crypto/zcrypt_ccamisc.c
210
if (t->secid != 0x20) {
drivers/s390/crypto/zcrypt_ccamisc.c
213
__func__, (int)t->secid);
drivers/s390/crypto/zcrypt_ccamisc.c
216
if (checkcpacfexport && !(t->kutc & 0x01)) {
drivers/s390/crypto/zcrypt_ccamisc.c
67
struct secaeskeytoken *t = (struct secaeskeytoken *)token;
drivers/s390/crypto/zcrypt_ccamisc.c
71
if (t->type != TOKTYPE_CCA_INTERNAL) {
drivers/s390/crypto/zcrypt_ccamisc.c
74
__func__, (int)t->type, TOKTYPE_CCA_INTERNAL);
drivers/s390/crypto/zcrypt_ccamisc.c
77
if (t->version != TOKVER_CCA_AES) {
drivers/s390/crypto/zcrypt_ccamisc.c
80
__func__, (int)t->version, TOKVER_CCA_AES);
drivers/s390/crypto/zcrypt_ccamisc.c
83
if (keybitsize > 0 && t->bitsize != keybitsize) {
drivers/s390/crypto/zcrypt_ccamisc.c
844
struct cipherkeytoken *t;
drivers/s390/crypto/zcrypt_ccamisc.c
86
__func__, (int)t->bitsize, keybitsize);
drivers/s390/crypto/zcrypt_ccamisc.c
898
t = (struct cipherkeytoken *)preqparm->kb.tlv3.gen_key_id_1;
drivers/s390/crypto/zcrypt_ccamisc.c
899
t->kmf1 |= (u16)(keygenflags & 0x0000FF00);
drivers/s390/crypto/zcrypt_ccamisc.c
900
t->kmf1 &= (u16)~(keygenflags & 0x000000FF);
drivers/s390/crypto/zcrypt_ccamisc.c
948
t = (struct cipherkeytoken *)prepparm->kb.tlv1.gen_key;
drivers/s390/crypto/zcrypt_ccamisc.c
950
if (*keybufsize >= t->len)
drivers/s390/crypto/zcrypt_ccamisc.c
951
memcpy(keybuf, t, t->len);
drivers/s390/crypto/zcrypt_ccamisc.c
955
*keybufsize = t->len;
drivers/s390/crypto/zcrypt_ep11misc.c
453
struct ep11_target_dev *t, int nt,
drivers/s390/crypto/zcrypt_ep11misc.c
458
u->targets = (u8 __user *)t;
drivers/s390/net/fsm.c
134
fsm_expire_timer(struct timer_list *t)
drivers/s390/net/fsm.c
136
fsm_timer *this = timer_container_of(this, t, tl);
drivers/s390/scsi/zfcp_erp.c
615
void zfcp_erp_timeout_handler(struct timer_list *t)
drivers/s390/scsi/zfcp_erp.c
617
struct zfcp_fsf_req *fsf_req = timer_container_of(fsf_req, t, timer);
drivers/s390/scsi/zfcp_erp.c
629
static void zfcp_erp_memwait_handler(struct timer_list *t)
drivers/s390/scsi/zfcp_erp.c
631
struct zfcp_erp_action *act = timer_container_of(act, t, timer);
drivers/s390/scsi/zfcp_ext.h
88
extern void zfcp_erp_timeout_handler(struct timer_list *t);
drivers/s390/scsi/zfcp_fsf.c
36
static void zfcp_fsf_request_timeout_handler(struct timer_list *t)
drivers/s390/scsi/zfcp_fsf.c
38
struct zfcp_fsf_req *fsf_req = timer_container_of(fsf_req, t, timer);
drivers/scsi/aic7xxx/aic79xx_core.c
202
static void ahd_stat_timer(struct timer_list *t);
drivers/scsi/aic7xxx/aic79xx_core.c
8785
ahd_stat_timer(struct timer_list *t)
drivers/scsi/aic7xxx/aic79xx_core.c
8787
struct ahd_softc *ahd = timer_container_of(ahd, t, stat_timer);
drivers/scsi/aic94xx/aic94xx_hwi.c
1132
struct asd_dma_tok t = ascb->dma_scb;
drivers/scsi/aic94xx/aic94xx_hwi.c
1137
seq->next_scb = t;
drivers/scsi/aic94xx/aic94xx_hwi.h
376
void asd_ascb_timedout(struct timer_list *t);
drivers/scsi/aic94xx/aic94xx_scb.c
863
void asd_ascb_timedout(struct timer_list *t)
drivers/scsi/aic94xx/aic94xx_scb.c
865
struct asd_ascb *ascb = timer_container_of(ascb, t, timer);
drivers/scsi/aic94xx/aic94xx_task.c
525
struct sas_task *t = task;
drivers/scsi/aic94xx/aic94xx_task.c
542
a->uldd_task = t;
drivers/scsi/aic94xx/aic94xx_task.c
543
t->lldd_task = a;
drivers/scsi/aic94xx/aic94xx_task.c
547
t = a->uldd_task;
drivers/scsi/aic94xx/aic94xx_task.c
549
if (t->task_proto & SAS_PROTOCOL_STP)
drivers/scsi/aic94xx/aic94xx_task.c
550
t->task_proto = SAS_PROTOCOL_STP;
drivers/scsi/aic94xx/aic94xx_task.c
551
switch (t->task_proto) {
drivers/scsi/aic94xx/aic94xx_task.c
554
res = asd_build_ata_ascb(a, t, gfp_flags);
drivers/scsi/aic94xx/aic94xx_task.c
557
res = asd_build_smp_ascb(a, t, gfp_flags);
drivers/scsi/aic94xx/aic94xx_task.c
560
res = asd_build_ssp_ascb(a, t, gfp_flags);
drivers/scsi/aic94xx/aic94xx_task.c
564
t->task_proto);
drivers/scsi/aic94xx/aic94xx_task.c
587
t = a->uldd_task;
drivers/scsi/aic94xx/aic94xx_task.c
588
switch (t->task_proto) {
drivers/scsi/aic94xx/aic94xx_task.c
602
t->lldd_task = NULL;
drivers/scsi/aic94xx/aic94xx_tmf.c
20
void (*timed_out)(struct timer_list *t))
drivers/scsi/aic94xx/aic94xx_tmf.c
245
static void asd_tmf_timedout(struct timer_list *t)
drivers/scsi/aic94xx/aic94xx_tmf.c
247
struct asd_ascb *ascb = timer_container_of(ascb, t, timer);
drivers/scsi/aic94xx/aic94xx_tmf.c
71
static void asd_clear_nexus_timedout(struct timer_list *t)
drivers/scsi/aic94xx/aic94xx_tmf.c
73
struct asd_ascb *ascb = timer_container_of(ascb, t, timer);
drivers/scsi/arcmsr/arcmsr_hba.c
132
static void arcmsr_request_device_map(struct timer_list *t);
drivers/scsi/arcmsr/arcmsr_hba.c
3937
static void arcmsr_set_iop_datetime(struct timer_list *t)
drivers/scsi/arcmsr/arcmsr_hba.c
3939
struct AdapterControlBlock *pacb = timer_container_of(pacb, t,
drivers/scsi/arcmsr/arcmsr_hba.c
4266
static void arcmsr_request_device_map(struct timer_list *t)
drivers/scsi/arcmsr/arcmsr_hba.c
4268
struct AdapterControlBlock *acb = timer_container_of(acb, t,
drivers/scsi/arm/fas216.c
2329
static void fas216_eh_timer(struct timer_list *t)
drivers/scsi/arm/fas216.c
2331
FAS216_Info *info = timer_container_of(info, t, eh_timer);
drivers/scsi/be2iscsi/be_main.c
5226
static void beiscsi_hw_tpe_check(struct timer_list *t)
drivers/scsi/be2iscsi/be_main.c
5228
struct beiscsi_hba *phba = timer_container_of(phba, t, hw_check);
drivers/scsi/be2iscsi/be_main.c
5243
static void beiscsi_hw_health_check(struct timer_list *t)
drivers/scsi/be2iscsi/be_main.c
5245
struct beiscsi_hba *phba = timer_container_of(phba, t, hw_check);
drivers/scsi/bfa/bfa_defs.h
74
u32 t = ((u32)(m)[0] << 16) | ((u32)(m)[1] << 8) | \
drivers/scsi/bfa/bfa_defs.h
76
t += (i); \
drivers/scsi/bfa/bfa_defs.h
77
(m)[0] = (t >> 16) & 0xFF; \
drivers/scsi/bfa/bfa_defs.h
78
(m)[1] = (t >> 8) & 0xFF; \
drivers/scsi/bfa/bfa_defs.h
79
(m)[2] = t & 0xFF; \
drivers/scsi/bfa/bfad.c
686
bfad_bfa_tmo(struct timer_list *t)
drivers/scsi/bfa/bfad.c
688
struct bfad_s *bfad = timer_container_of(bfad, t,
drivers/scsi/bfa/bfad_drv.h
311
void bfad_bfa_tmo(struct timer_list *t);
drivers/scsi/bnx2fc/bnx2fc_fcoe.c
838
static void bnx2fc_destroy_timer(struct timer_list *t)
drivers/scsi/bnx2fc/bnx2fc_fcoe.c
840
struct bnx2fc_hba *hba = timer_container_of(hba, t, destroy_timer);
drivers/scsi/bnx2fc/bnx2fc_tgt.c
17
static void bnx2fc_upld_timer(struct timer_list *t);
drivers/scsi/bnx2fc/bnx2fc_tgt.c
18
static void bnx2fc_ofld_timer(struct timer_list *t);
drivers/scsi/bnx2fc/bnx2fc_tgt.c
30
static void bnx2fc_upld_timer(struct timer_list *t)
drivers/scsi/bnx2fc/bnx2fc_tgt.c
33
struct bnx2fc_rport *tgt = timer_container_of(tgt, t, upld_timer);
drivers/scsi/bnx2fc/bnx2fc_tgt.c
43
static void bnx2fc_ofld_timer(struct timer_list *t)
drivers/scsi/bnx2fc/bnx2fc_tgt.c
46
struct bnx2fc_rport *tgt = timer_container_of(tgt, t, ofld_timer);
drivers/scsi/bnx2i/bnx2i.h
856
extern void bnx2i_ep_ofld_timer(struct timer_list *t);
drivers/scsi/bnx2i/bnx2i_hwi.c
686
void bnx2i_ep_ofld_timer(struct timer_list *t)
drivers/scsi/bnx2i/bnx2i_hwi.c
688
struct bnx2i_endpoint *ep = timer_container_of(ep, t, ofld_timer);
drivers/scsi/csiostor/csio_hw.c
3739
csio_hw_mb_timer(struct timer_list *t)
drivers/scsi/csiostor/csio_hw.c
3741
struct csio_mbm *mbm = timer_container_of(mbm, t, timer);
drivers/scsi/csiostor/csio_hw.c
4108
csio_mgmt_tmo_handler(struct timer_list *t)
drivers/scsi/csiostor/csio_hw.c
4110
struct csio_mgmtm *mgmtm = timer_container_of(mgmtm, t, mgmt_timer);
drivers/scsi/cxgbi/cxgb3i/cxgb3i.c
548
static void act_open_retry_timer(struct timer_list *t)
drivers/scsi/cxgbi/cxgb3i/cxgb3i.c
550
struct cxgbi_sock *csk = timer_container_of(csk, t, retry_timer);
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
1052
struct tid_info *t = lldi->tids;
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
1054
csk = lookup_atid(t, atid);
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
1098
struct tid_info *t = lldi->tids;
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
1100
csk = lookup_tid(t, tid);
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
1119
struct tid_info *t = lldi->tids;
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
1121
csk = lookup_tid(t, tid);
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
1158
struct tid_info *t = lldi->tids;
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
1161
csk = lookup_tid(t, tid);
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
1206
struct tid_info *t = lldi->tids;
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
1208
csk = lookup_tid(t, tid);
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
1230
struct tid_info *t = lldi->tids;
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
1232
csk = lookup_tid(t, tid);
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
1252
struct tid_info *t = lldi->tids;
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
1254
csk = lookup_tid(t, tid);
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
1355
struct tid_info *t = lldi->tids;
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
1360
csk = lookup_tid(t, tid);
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
1453
struct tid_info *t = lldi->tids;
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
1456
csk = lookup_tid(t, tid);
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
1517
struct tid_info *t = lldi->tids;
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
1524
csk = lookup_tid(t, tid);
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
1609
struct tid_info *t = lldi->tids;
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
1611
csk = lookup_tid(t, tid);
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
1629
struct tid_info *t = lldi->tids;
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
1632
csk = lookup_tid(t, tid);
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
2217
struct tid_info *t;
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
2263
t = &adap->tids;
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
2264
if (t->ntids <= CXGBI_MAX_CONN)
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
2265
max_conn = t->ntids;
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
2272
ndev->name, cdev, t->ntids);
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
899
struct tid_info *t = lldi->tids;
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
902
csk = lookup_atid(t, atid);
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
988
static void csk_act_open_retry_timer(struct timer_list *t)
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
991
struct cxgbi_sock *csk = timer_container_of(csk, t, retry_timer);
drivers/scsi/dc395x.c
323
static void waiting_timeout(struct timer_list *t);
drivers/scsi/dc395x.c
751
static void waiting_timeout(struct timer_list *t)
drivers/scsi/dc395x.c
754
struct AdapterCtlBlk *acb = timer_container_of(acb, t, waiting_timer);
drivers/scsi/elx/efct/efct_xport.c
181
efct_xport_stats_timer_cb(struct timer_list *t)
drivers/scsi/elx/efct/efct_xport.c
183
struct efct_xport *xport = timer_container_of(xport, t, stats_timer);
drivers/scsi/elx/libefc/efc_els.c
148
efc_els_delay_timer_cb(struct timer_list *t)
drivers/scsi/elx/libefc/efc_els.c
150
struct efc_els_io_req *els = timer_container_of(els, t, delay_timer);
drivers/scsi/elx/libefc/efc_fabric.c
887
gidpt_delay_timer_cb(struct timer_list *t)
drivers/scsi/elx/libefc/efc_fabric.c
889
struct efc_node *node = timer_container_of(node, t, gidpt_delay_timer);
drivers/scsi/esas2r/esas2r.h
1114
void esas2r_targ_db_remove(struct esas2r_adapter *a, struct esas2r_target *t);
drivers/scsi/esas2r/esas2r.h
1382
static inline u16 esas2r_targ_get_id(struct esas2r_target *t,
drivers/scsi/esas2r/esas2r.h
1385
return (u16)(uintptr_t)(t - a->targetdb);
drivers/scsi/esas2r/esas2r_disc.c
1049
struct esas2r_target *t;
drivers/scsi/esas2r/esas2r_disc.c
1056
for (t = a->targetdb; t < a->targetdb_end; t++) {
drivers/scsi/esas2r/esas2r_disc.c
1057
if (t->new_target_state != TS_NOT_PRESENT)
drivers/scsi/esas2r/esas2r_disc.c
1060
t->new_target_state = TS_INVALID;
drivers/scsi/esas2r/esas2r_disc.c
1066
esas2r_targ_get_id(t,
drivers/scsi/esas2r/esas2r_disc.c
1088
struct esas2r_target *t = dc->curr_targ;
drivers/scsi/esas2r/esas2r_disc.c
1090
if (t >= a->targetdb_end) {
drivers/scsi/esas2r/esas2r_disc.c
1094
} else if (t->new_target_state == TS_PRESENT) {
drivers/scsi/esas2r/esas2r_disc.c
1095
struct atto_vda_ae_lu *luevt = &t->lu_event;
drivers/scsi/esas2r/esas2r_disc.c
1101
t->new_target_state = TS_INVALID;
drivers/scsi/esas2r/esas2r_disc.c
1105
dc->curr_virt_id = esas2r_targ_get_id(t, a);
drivers/scsi/esas2r/esas2r_disc.c
1162
struct esas2r_target *t;
drivers/scsi/esas2r/esas2r_disc.c
1173
t = a->targetdb + rq->target_id;
drivers/scsi/esas2r/esas2r_disc.c
1175
if (t->target_state == TS_PRESENT)
drivers/scsi/esas2r/esas2r_disc.c
1177
t->virt_targ_id);
drivers/scsi/esas2r/esas2r_disc.c
944
struct esas2r_target *t = NULL;
drivers/scsi/esas2r/esas2r_disc.c
974
t = esas2r_targ_db_add_pthru(a,
drivers/scsi/esas2r/esas2r_disc.c
983
if (t)
drivers/scsi/esas2r/esas2r_disc.c
984
memcpy(&t->sas_addr, &dc->sas_addr,
drivers/scsi/esas2r/esas2r_disc.c
985
sizeof(t->sas_addr));
drivers/scsi/esas2r/esas2r_int.c
706
struct esas2r_target *t = a->targetdb + target;
drivers/scsi/esas2r/esas2r_int.c
710
if (cplen > sizeof(t->lu_event))
drivers/scsi/esas2r/esas2r_int.c
711
cplen = sizeof(t->lu_event);
drivers/scsi/esas2r/esas2r_int.c
718
t->new_target_state = TS_INVALID;
drivers/scsi/esas2r/esas2r_int.c
721
t->new_target_state = TS_NOT_PRESENT;
drivers/scsi/esas2r/esas2r_int.c
728
t->new_target_state = TS_NOT_PRESENT;
drivers/scsi/esas2r/esas2r_int.c
733
t->new_target_state = TS_PRESENT;
drivers/scsi/esas2r/esas2r_int.c
738
if (t->new_target_state != TS_INVALID) {
drivers/scsi/esas2r/esas2r_int.c
739
memcpy(&t->lu_event, &ae->lu, cplen);
drivers/scsi/esas2r/esas2r_io.c
48
struct esas2r_target *t = NULL;
drivers/scsi/esas2r/esas2r_io.c
529
struct esas2r_target *t = a->targetdb + rq->target_id;
drivers/scsi/esas2r/esas2r_io.c
541
&& t->target_state == TS_PRESENT
drivers/scsi/esas2r/esas2r_io.c
542
&& !(t->flags & TF_PASS_THRU)) {
drivers/scsi/esas2r/esas2r_io.c
59
t = a->targetdb + rq->target_id;
drivers/scsi/esas2r/esas2r_io.c
594
startlba = t->inter_block - (lbalo & (t->inter_block -
drivers/scsi/esas2r/esas2r_io.c
596
sgc->length = startlba * t->block_size;
drivers/scsi/esas2r/esas2r_io.c
599
if ((lbalo & (t->inter_block - 1)) == 0)
drivers/scsi/esas2r/esas2r_io.c
61
if (unlikely(t >= a->targetdb_end
drivers/scsi/esas2r/esas2r_io.c
62
|| !(t->flags & TF_USED))) {
drivers/scsi/esas2r/esas2r_io.c
641
if (len > t->inter_byte)
drivers/scsi/esas2r/esas2r_io.c
642
sgc->length = t->inter_byte;
drivers/scsi/esas2r/esas2r_io.c
66
rq->vrq->scsi.target_id = cpu_to_le16(t->virt_targ_id);
drivers/scsi/esas2r/esas2r_io.c
73
if (unlikely(t->target_state != TS_PRESENT &&
drivers/scsi/esas2r/esas2r_ioctl.c
1056
struct esas2r_target *t;
drivers/scsi/esas2r/esas2r_ioctl.c
1076
t = a->targetdb + (u16)gda->target_id;
drivers/scsi/esas2r/esas2r_ioctl.c
1078
if (t->target_state != TS_PRESENT) {
drivers/scsi/esas2r/esas2r_ioctl.c
1081
if (t->sas_addr == 0) {
drivers/scsi/esas2r/esas2r_ioctl.c
1084
*(u64 *)gda->address = t->sas_addr;
drivers/scsi/esas2r/esas2r_ioctl.c
1152
struct esas2r_target *t;
drivers/scsi/esas2r/esas2r_ioctl.c
1172
t = a->targetdb + (u16)gdi->target_id;
drivers/scsi/esas2r/esas2r_ioctl.c
1180
if (t->target_state != TS_PRESENT) {
drivers/scsi/esas2r/esas2r_ioctl.c
403
struct esas2r_target *t;
drivers/scsi/esas2r/esas2r_ioctl.c
540
t = esas2r_targ_db_find_by_sas_addr(a, (u64 *)gsa->sas_addr);
drivers/scsi/esas2r/esas2r_ioctl.c
543
if (t == NULL) {
drivers/scsi/esas2r/esas2r_ioctl.c
550
rq->target_id = esas2r_targ_get_id(t, a);
drivers/scsi/esas2r/esas2r_ioctl.c
559
t = a->targetdb + rq->target_id;
drivers/scsi/esas2r/esas2r_ioctl.c
561
if (t >= a->targetdb_end
drivers/scsi/esas2r/esas2r_ioctl.c
562
|| t->target_state != TS_PRESENT
drivers/scsi/esas2r/esas2r_ioctl.c
563
|| t->sas_addr == 0) {
drivers/scsi/esas2r/esas2r_ioctl.c
569
*(u64 *)gda->sas_addr = t->sas_addr;
drivers/scsi/esas2r/esas2r_ioctl.c
578
t = a->targetdb + rq->target_id;
drivers/scsi/esas2r/esas2r_ioctl.c
580
if (t >= a->targetdb_end
drivers/scsi/esas2r/esas2r_ioctl.c
581
|| t->target_state != TS_PRESENT
drivers/scsi/esas2r/esas2r_ioctl.c
582
|| !(t->flags & TF_PASS_THRU)) {
drivers/scsi/esas2r/esas2r_ioctl.c
589
t->phys_targ_id)) {
drivers/scsi/esas2r/esas2r_main.c
1574
static void esas2r_timer_callback(struct timer_list *t);
drivers/scsi/esas2r/esas2r_main.c
1586
static void esas2r_timer_callback(struct timer_list *t)
drivers/scsi/esas2r/esas2r_main.c
1588
struct esas2r_adapter *a = timer_container_of(a, t, timer);
drivers/scsi/esas2r/esas2r_main.c
658
struct esas2r_target *t;
drivers/scsi/esas2r/esas2r_main.c
694
for (t = a->targetdb; t < a->targetdb_end; t++)
drivers/scsi/esas2r/esas2r_main.c
695
if (t->buffered_target_state == TS_PRESENT) {
drivers/scsi/esas2r/esas2r_main.c
698
(u16)(uintptr_t)(t - a->targetdb));
drivers/scsi/esas2r/esas2r_targdb.c
105
t,
drivers/scsi/esas2r/esas2r_targdb.c
110
esas2r_targ_get_id(t,
drivers/scsi/esas2r/esas2r_targdb.c
123
struct esas2r_target *t;
drivers/scsi/esas2r/esas2r_targdb.c
133
t = a->targetdb + dc->curr_virt_id;
drivers/scsi/esas2r/esas2r_targdb.c
135
if (t->target_state == TS_PRESENT) {
drivers/scsi/esas2r/esas2r_targdb.c
142
t,
drivers/scsi/esas2r/esas2r_targdb.c
156
t->block_size = dc->block_size;
drivers/scsi/esas2r/esas2r_targdb.c
157
t->inter_byte = dc->interleave;
drivers/scsi/esas2r/esas2r_targdb.c
158
t->inter_block = dc->interleave / dc->block_size;
drivers/scsi/esas2r/esas2r_targdb.c
159
t->virt_targ_id = dc->curr_virt_id;
drivers/scsi/esas2r/esas2r_targdb.c
160
t->phys_targ_id = ESAS2R_TARG_ID_INV;
drivers/scsi/esas2r/esas2r_targdb.c
162
t->flags &= ~TF_PASS_THRU;
drivers/scsi/esas2r/esas2r_targdb.c
163
t->flags |= TF_USED;
drivers/scsi/esas2r/esas2r_targdb.c
165
t->identifier_len = 0;
drivers/scsi/esas2r/esas2r_targdb.c
167
t->target_state = TS_PRESENT;
drivers/scsi/esas2r/esas2r_targdb.c
169
return t;
drivers/scsi/esas2r/esas2r_targdb.c
177
struct esas2r_target *t;
drivers/scsi/esas2r/esas2r_targdb.c
189
t = esas2r_targ_db_find_by_ident(a, ident, ident_len);
drivers/scsi/esas2r/esas2r_targdb.c
191
if (t == NULL) {
drivers/scsi/esas2r/esas2r_targdb.c
192
t = a->targetdb + dc->curr_virt_id;
drivers/scsi/esas2r/esas2r_targdb.c
194
if (ident_len > sizeof(t->identifier)
drivers/scsi/esas2r/esas2r_targdb.c
195
|| t->target_state == TS_PRESENT) {
drivers/scsi/esas2r/esas2r_targdb.c
201
esas2r_hdebug("add PT; T:%d, V:%d, P:%d", esas2r_targ_get_id(t, a),
drivers/scsi/esas2r/esas2r_targdb.c
205
t->block_size = 0;
drivers/scsi/esas2r/esas2r_targdb.c
206
t->inter_byte = 0;
drivers/scsi/esas2r/esas2r_targdb.c
207
t->inter_block = 0;
drivers/scsi/esas2r/esas2r_targdb.c
208
t->virt_targ_id = dc->curr_virt_id;
drivers/scsi/esas2r/esas2r_targdb.c
209
t->phys_targ_id = dc->curr_phys_id;
drivers/scsi/esas2r/esas2r_targdb.c
210
t->identifier_len = ident_len;
drivers/scsi/esas2r/esas2r_targdb.c
212
memcpy(t->identifier, ident, ident_len);
drivers/scsi/esas2r/esas2r_targdb.c
214
t->flags |= TF_PASS_THRU | TF_USED;
drivers/scsi/esas2r/esas2r_targdb.c
216
t->target_state = TS_PRESENT;
drivers/scsi/esas2r/esas2r_targdb.c
218
return t;
drivers/scsi/esas2r/esas2r_targdb.c
221
void esas2r_targ_db_remove(struct esas2r_adapter *a, struct esas2r_target *t)
drivers/scsi/esas2r/esas2r_targdb.c
225
t->target_state = TS_NOT_PRESENT;
drivers/scsi/esas2r/esas2r_targdb.c
227
esas2r_trace("remove id:%d", esas2r_targ_get_id(t, a));
drivers/scsi/esas2r/esas2r_targdb.c
235
struct esas2r_target *t;
drivers/scsi/esas2r/esas2r_targdb.c
237
for (t = a->targetdb; t < a->targetdb_end; t++)
drivers/scsi/esas2r/esas2r_targdb.c
238
if (t->sas_addr == *sas_addr)
drivers/scsi/esas2r/esas2r_targdb.c
239
return t;
drivers/scsi/esas2r/esas2r_targdb.c
248
struct esas2r_target *t;
drivers/scsi/esas2r/esas2r_targdb.c
250
for (t = a->targetdb; t < a->targetdb_end; t++) {
drivers/scsi/esas2r/esas2r_targdb.c
251
if (ident_len == t->identifier_len
drivers/scsi/esas2r/esas2r_targdb.c
252
&& memcmp(&t->identifier[0], identifier,
drivers/scsi/esas2r/esas2r_targdb.c
254
return t;
drivers/scsi/esas2r/esas2r_targdb.c
265
struct esas2r_target *t = a->targetdb + id;
drivers/scsi/esas2r/esas2r_targdb.c
267
if (t->target_state == TS_PRESENT)
drivers/scsi/esas2r/esas2r_targdb.c
279
struct esas2r_target *t;
drivers/scsi/esas2r/esas2r_targdb.c
281
for (t = a->targetdb; t < a->targetdb_end; t++) {
drivers/scsi/esas2r/esas2r_targdb.c
282
if (t->target_state != TS_PRESENT)
drivers/scsi/esas2r/esas2r_targdb.c
285
if (t->virt_targ_id == virt_id)
drivers/scsi/esas2r/esas2r_targdb.c
286
return t;
drivers/scsi/esas2r/esas2r_targdb.c
295
struct esas2r_target *t;
drivers/scsi/esas2r/esas2r_targdb.c
299
for (t = a->targetdb; t < a->targetdb_end; t++)
drivers/scsi/esas2r/esas2r_targdb.c
300
if (t->target_state == TS_PRESENT)
drivers/scsi/esas2r/esas2r_targdb.c
48
struct esas2r_target *t;
drivers/scsi/esas2r/esas2r_targdb.c
50
for (t = a->targetdb; t < a->targetdb_end; t++) {
drivers/scsi/esas2r/esas2r_targdb.c
51
memset(t, 0, sizeof(struct esas2r_target));
drivers/scsi/esas2r/esas2r_targdb.c
53
t->target_state = TS_NOT_PRESENT;
drivers/scsi/esas2r/esas2r_targdb.c
54
t->buffered_target_state = TS_NOT_PRESENT;
drivers/scsi/esas2r/esas2r_targdb.c
55
t->new_target_state = TS_INVALID;
drivers/scsi/esas2r/esas2r_targdb.c
61
struct esas2r_target *t;
drivers/scsi/esas2r/esas2r_targdb.c
64
for (t = a->targetdb; t < a->targetdb_end; t++) {
drivers/scsi/esas2r/esas2r_targdb.c
65
if (t->target_state != TS_PRESENT)
drivers/scsi/esas2r/esas2r_targdb.c
69
esas2r_targ_db_remove(a, t);
drivers/scsi/esas2r/esas2r_targdb.c
73
esas2r_trace("remove id:%d", esas2r_targ_get_id(t,
drivers/scsi/esas2r/esas2r_targdb.c
75
esas2r_target_state_changed(a, esas2r_targ_get_id(t,
drivers/scsi/esas2r/esas2r_targdb.c
84
struct esas2r_target *t;
drivers/scsi/esas2r/esas2r_targdb.c
94
for (t = a->targetdb; t < a->targetdb_end; t++) {
drivers/scsi/esas2r/esas2r_targdb.c
98
if (t->buffered_target_state != t->target_state)
drivers/scsi/esas2r/esas2r_targdb.c
99
state = t->buffered_target_state = t->target_state;
drivers/scsi/fcoe/fcoe_ctlr.c
1773
static void fcoe_ctlr_timeout(struct timer_list *t)
drivers/scsi/fcoe/fcoe_ctlr.c
1775
struct fcoe_ctlr *fip = timer_container_of(fip, t, timer);
drivers/scsi/fcoe/fcoe_ctlr.c
897
unsigned long t;
drivers/scsi/fcoe/fcoe_ctlr.c
974
t = ntohl(fka->fd_fka_period);
drivers/scsi/fcoe/fcoe_ctlr.c
975
if (t >= FCOE_CTLR_MIN_FKA)
drivers/scsi/fcoe/fcoe_ctlr.c
976
fcf->fka_period = msecs_to_jiffies(t);
drivers/scsi/fcoe/fcoe_transport.c
448
void fcoe_queue_timer(struct timer_list *t)
drivers/scsi/fcoe/fcoe_transport.c
450
struct fcoe_port *port = timer_container_of(port, t, timer);
drivers/scsi/fnic/fdls_disc.c
2111
void fdls_fabric_timer_callback(struct timer_list *t)
drivers/scsi/fnic/fdls_disc.c
2113
struct fnic_fdls_fabric_s *fabric = timer_container_of(fabric, t,
drivers/scsi/fnic/fdls_disc.c
2299
void fdls_fdmi_timer_callback(struct timer_list *t)
drivers/scsi/fnic/fdls_disc.c
2301
struct fnic_fdls_fabric_s *fabric = timer_container_of(fabric, t,
drivers/scsi/fnic/fdls_disc.c
2381
static void fdls_tport_timer_callback(struct timer_list *t)
drivers/scsi/fnic/fdls_disc.c
2383
struct fnic_tport_s *tport = timer_container_of(tport, t, retry_timer);
drivers/scsi/fnic/fdls_disc.c
89
static void fdls_tport_timer_callback(struct timer_list *t);
drivers/scsi/fnic/fip.c
1001
struct fnic *fnic = timer_container_of(fnic, t, fcs_ka_timer);
drivers/scsi/fnic/fip.c
778
void fnic_handle_fip_timer(struct timer_list *t)
drivers/scsi/fnic/fip.c
780
struct fnic *fnic = timer_container_of(fnic, t, retry_fip_timer);
drivers/scsi/fnic/fip.c
790
void fnic_handle_enode_ka_timer(struct timer_list *t)
drivers/scsi/fnic/fip.c
793
struct fnic *fnic = timer_container_of(fnic, t, enode_ka_timer);
drivers/scsi/fnic/fip.c
843
void fnic_handle_vn_ka_timer(struct timer_list *t)
drivers/scsi/fnic/fip.c
846
struct fnic *fnic = timer_container_of(fnic, t, vn_ka_timer);
drivers/scsi/fnic/fip.c
999
void fnic_handle_fcs_ka_timer(struct timer_list *t)
drivers/scsi/fnic/fnic.h
535
extern void fnic_handle_fip_timer(struct timer_list *t);
drivers/scsi/fnic/fnic_fdls.h
396
void fdls_fdmi_timer_callback(struct timer_list *t);
drivers/scsi/fnic/fnic_fdls.h
418
void fnic_handle_fcs_ka_timer(struct timer_list *t);
drivers/scsi/fnic/fnic_fdls.h
419
void fnic_handle_enode_ka_timer(struct timer_list *t);
drivers/scsi/fnic/fnic_fdls.h
420
void fnic_handle_vn_ka_timer(struct timer_list *t);
drivers/scsi/fnic/fnic_fdls.h
421
void fnic_handle_fip_timer(struct timer_list *t);
drivers/scsi/fnic/fnic_fdls.h
422
extern void fdls_fabric_timer_callback(struct timer_list *t);
drivers/scsi/fnic/fnic_fdls.h
426
extern void fdls_fabric_timer_callback(struct timer_list *t);
drivers/scsi/fnic/fnic_main.c
447
static void fnic_notify_timer(struct timer_list *t)
drivers/scsi/fnic/fnic_main.c
449
struct fnic *fnic = timer_container_of(fnic, t, notify_timer);
drivers/scsi/fnic/fnic_pci_subsys_devid.c
107
struct fnic_pcie_device *t = fnic_pcie_device_table;
drivers/scsi/fnic/fnic_pci_subsys_devid.c
113
while (t->device != 0) {
drivers/scsi/fnic/fnic_pci_subsys_devid.c
116
(char *) &t->subsystem_device, sizeof(short)) == 0)
drivers/scsi/fnic/fnic_pci_subsys_devid.c
118
t++;
drivers/scsi/hisi_sas/hisi_sas_main.c
1017
static void hisi_sas_wait_phyup_timedout(struct timer_list *t)
drivers/scsi/hisi_sas/hisi_sas_main.c
1019
struct hisi_sas_phy *phy = timer_container_of(phy, t, timer);
drivers/scsi/hisi_sas/hisi_sas_v1_hw.c
796
static void start_phys_v1_hw(struct timer_list *t)
drivers/scsi/hisi_sas/hisi_sas_v1_hw.c
798
struct hisi_hba *hisi_hba = timer_container_of(hisi_hba, t, timer);
drivers/scsi/hisi_sas/hisi_sas_v2_hw.c
1328
static void link_timeout_enable_link(struct timer_list *t)
drivers/scsi/hisi_sas/hisi_sas_v2_hw.c
1330
struct hisi_hba *hisi_hba = timer_container_of(hisi_hba, t, timer);
drivers/scsi/hisi_sas/hisi_sas_v2_hw.c
1349
static void link_timeout_disable_link(struct timer_list *t)
drivers/scsi/hisi_sas/hisi_sas_v2_hw.c
1351
struct hisi_hba *hisi_hba = timer_container_of(hisi_hba, t, timer);
drivers/scsi/hisi_sas/hisi_sas_v2_hw.c
2580
static void hisi_sas_internal_abort_quirk_timeout(struct timer_list *t)
drivers/scsi/hisi_sas/hisi_sas_v2_hw.c
2582
struct hisi_sas_slot *slot = timer_container_of(slot, t,
drivers/scsi/hisi_sas/hisi_sas_v2_hw.c
727
static void link_timeout_disable_link(struct timer_list *t);
drivers/scsi/hpsa_cmd.h
363
#define TYPE_ATTR_DIR(t, a, d) ((((d) & 0x03) << 6) |\
drivers/scsi/hpsa_cmd.h
365
((t) & 0x07))
drivers/scsi/ibmvscsi/ibmvfc.c
1698
static void ibmvfc_timeout(struct timer_list *t)
drivers/scsi/ibmvscsi/ibmvfc.c
1700
struct ibmvfc_event *evt = timer_container_of(evt, t, timer);
drivers/scsi/ibmvscsi/ibmvfc.c
4632
static void ibmvfc_adisc_timeout(struct timer_list *t)
drivers/scsi/ibmvscsi/ibmvfc.c
4634
struct ibmvfc_target *tgt = timer_container_of(tgt, t, timer);
drivers/scsi/ibmvscsi/ibmvfc.h
921
#define tgt_dbg(t, fmt, ...) \
drivers/scsi/ibmvscsi/ibmvfc.h
922
DBG_CMD(dev_info((t)->vhost->dev, "%llX: " fmt, (t)->scsi_id, ##__VA_ARGS__))
drivers/scsi/ibmvscsi/ibmvfc.h
924
#define tgt_info(t, fmt, ...) \
drivers/scsi/ibmvscsi/ibmvfc.h
925
dev_info((t)->vhost->dev, "%llX: " fmt, (t)->scsi_id, ##__VA_ARGS__)
drivers/scsi/ibmvscsi/ibmvfc.h
927
#define tgt_err(t, fmt, ...) \
drivers/scsi/ibmvscsi/ibmvfc.h
928
dev_err((t)->vhost->dev, "%llX: " fmt, (t)->scsi_id, ##__VA_ARGS__)
drivers/scsi/ibmvscsi/ibmvfc.h
930
#define tgt_log(t, level, fmt, ...) \
drivers/scsi/ibmvscsi/ibmvfc.h
932
if ((t)->vhost->log_level >= level) \
drivers/scsi/ibmvscsi/ibmvfc.h
933
tgt_err(t, fmt, ##__VA_ARGS__); \
drivers/scsi/ibmvscsi/ibmvscsi.c
846
static void ibmvscsi_timeout(struct timer_list *t)
drivers/scsi/ibmvscsi/ibmvscsi.c
849
t, timer);
drivers/scsi/ipr.c
2590
static void ipr_timeout(struct timer_list *t)
drivers/scsi/ipr.c
2592
struct ipr_cmnd *ipr_cmd = timer_container_of(ipr_cmd, t, timer);
drivers/scsi/ipr.c
2623
static void ipr_oper_timeout(struct timer_list *t)
drivers/scsi/ipr.c
2625
struct ipr_cmnd *ipr_cmd = timer_container_of(ipr_cmd, t, timer);
drivers/scsi/ipr.c
5152
static void ipr_abort_timeout(struct timer_list *t)
drivers/scsi/ipr.c
5154
struct ipr_cmnd *ipr_cmd = timer_container_of(ipr_cmd, t, timer);
drivers/scsi/ipr.c
7477
static void ipr_reset_timer_done(struct timer_list *t)
drivers/scsi/ipr.c
7479
struct ipr_cmnd *ipr_cmd = timer_container_of(ipr_cmd, t, timer);
drivers/scsi/ipr.c
9097
struct ipr_interrupts *t;
drivers/scsi/ipr.c
9101
t = &ioa_cfg->regs;
drivers/scsi/ipr.c
9104
t->set_interrupt_mask_reg = base + p->set_interrupt_mask_reg;
drivers/scsi/ipr.c
9105
t->clr_interrupt_mask_reg = base + p->clr_interrupt_mask_reg;
drivers/scsi/ipr.c
9106
t->clr_interrupt_mask_reg32 = base + p->clr_interrupt_mask_reg32;
drivers/scsi/ipr.c
9107
t->sense_interrupt_mask_reg = base + p->sense_interrupt_mask_reg;
drivers/scsi/ipr.c
9108
t->sense_interrupt_mask_reg32 = base + p->sense_interrupt_mask_reg32;
drivers/scsi/ipr.c
9109
t->clr_interrupt_reg = base + p->clr_interrupt_reg;
drivers/scsi/ipr.c
9110
t->clr_interrupt_reg32 = base + p->clr_interrupt_reg32;
drivers/scsi/ipr.c
9111
t->sense_interrupt_reg = base + p->sense_interrupt_reg;
drivers/scsi/ipr.c
9112
t->sense_interrupt_reg32 = base + p->sense_interrupt_reg32;
drivers/scsi/ipr.c
9113
t->ioarrin_reg = base + p->ioarrin_reg;
drivers/scsi/ipr.c
9114
t->sense_uproc_interrupt_reg = base + p->sense_uproc_interrupt_reg;
drivers/scsi/ipr.c
9115
t->sense_uproc_interrupt_reg32 = base + p->sense_uproc_interrupt_reg32;
drivers/scsi/ipr.c
9116
t->set_uproc_interrupt_reg = base + p->set_uproc_interrupt_reg;
drivers/scsi/ipr.c
9117
t->set_uproc_interrupt_reg32 = base + p->set_uproc_interrupt_reg32;
drivers/scsi/ipr.c
9118
t->clr_uproc_interrupt_reg = base + p->clr_uproc_interrupt_reg;
drivers/scsi/ipr.c
9119
t->clr_uproc_interrupt_reg32 = base + p->clr_uproc_interrupt_reg32;
drivers/scsi/ipr.c
9122
t->init_feedback_reg = base + p->init_feedback_reg;
drivers/scsi/ipr.c
9123
t->dump_addr_reg = base + p->dump_addr_reg;
drivers/scsi/ipr.c
9124
t->dump_data_reg = base + p->dump_data_reg;
drivers/scsi/ipr.c
9125
t->endian_swap_reg = base + p->endian_swap_reg;
drivers/scsi/isci/host.c
1593
static void controller_timeout(struct timer_list *t)
drivers/scsi/isci/host.c
1595
struct sci_timer *tmr = timer_container_of(tmr, t, timer);
drivers/scsi/isci/host.c
1738
static void power_control_timeout(struct timer_list *t)
drivers/scsi/isci/host.c
1740
struct sci_timer *tmr = timer_container_of(tmr, t, timer);
drivers/scsi/isci/host.c
959
static void phy_startup_timeout(struct timer_list *t)
drivers/scsi/isci/host.c
961
struct sci_timer *tmr = timer_container_of(tmr, t, timer);
drivers/scsi/isci/isci.h
494
void sci_init_timer(struct sci_timer *tmr, void (*fn)(struct timer_list *t))
drivers/scsi/isci/phy.c
318
static void phy_sata_timeout(struct timer_list *t)
drivers/scsi/isci/phy.c
320
struct sci_timer *tmr = timer_container_of(tmr, t, timer);
drivers/scsi/isci/port.c
776
static void port_timeout(struct timer_list *t)
drivers/scsi/isci/port.c
778
struct sci_timer *tmr = timer_container_of(tmr, t, timer);
drivers/scsi/isci/port_config.c
321
static void mpc_agent_timeout(struct timer_list *t)
drivers/scsi/isci/port_config.c
324
struct sci_timer *tmr = timer_container_of(tmr, t, timer);
drivers/scsi/isci/port_config.c
659
static void apc_agent_timeout(struct timer_list *t)
drivers/scsi/isci/port_config.c
662
struct sci_timer *tmr = timer_container_of(tmr, t, timer);
drivers/scsi/libfc/fc_fcp.c
1284
static void fc_lun_reset_send(struct timer_list *t)
drivers/scsi/libfc/fc_fcp.c
1286
struct fc_fcp_pkt *fsp = timer_container_of(fsp, t, timer);
drivers/scsi/libfc/fc_fcp.c
1417
static void fc_fcp_timeout(struct timer_list *t)
drivers/scsi/libfc/fc_fcp.c
1419
struct fc_fcp_pkt *fsp = timer_container_of(fsp, t, timer);
drivers/scsi/libiscsi.c
1900
static void iscsi_tmf_timedout(struct timer_list *t)
drivers/scsi/libiscsi.c
1902
struct iscsi_session *session = timer_container_of(session, t,
drivers/scsi/libiscsi.c
2243
static void iscsi_check_transport_timeouts(struct timer_list *t)
drivers/scsi/libiscsi.c
2245
struct iscsi_conn *conn = timer_container_of(conn, t, transport_timer);
drivers/scsi/libsas/sas_internal.h
109
void sas_task_internal_timedout(struct timer_list *t);
drivers/scsi/libsas/sas_scsi_host.c
868
void sas_task_internal_timedout(struct timer_list *t)
drivers/scsi/libsas/sas_scsi_host.c
870
struct sas_task_slow *slow = timer_container_of(slow, t, timer);
drivers/scsi/lpfc/lpfc_crtn.h
256
void lpfc_sli4_poll_hbtimer(struct timer_list *t);
drivers/scsi/lpfc/lpfc_crtn.h
322
void lpfc_poll_timeout(struct timer_list *t);
drivers/scsi/lpfc/lpfc_crtn.h
414
void lpfc_mbox_timeout(struct timer_list *t);
drivers/scsi/lpfc/lpfc_ct.c
3458
lpfc_delayed_disc_tmo(struct timer_list *t)
drivers/scsi/lpfc/lpfc_ct.c
3460
struct lpfc_vport *vport = timer_container_of(vport, t,
drivers/scsi/lpfc/lpfc_els.c
11767
lpfc_fabric_block_timeout(struct timer_list *t)
drivers/scsi/lpfc/lpfc_els.c
11769
struct lpfc_hba *phba = timer_container_of(phba, t,
drivers/scsi/lpfc/lpfc_els.c
4507
lpfc_els_retry_delay(struct timer_list *t)
drivers/scsi/lpfc/lpfc_els.c
4509
struct lpfc_nodelist *ndlp = timer_container_of(ndlp, t,
drivers/scsi/lpfc/lpfc_els.c
9550
lpfc_els_timeout(struct timer_list *t)
drivers/scsi/lpfc/lpfc_els.c
9552
struct lpfc_vport *vport = timer_container_of(vport, t, els_tmofunc);
drivers/scsi/lpfc/lpfc_hbadisc.c
6064
lpfc_disc_timeout(struct timer_list *t)
drivers/scsi/lpfc/lpfc_hbadisc.c
6066
struct lpfc_vport *vport = timer_container_of(vport, t, fc_disctmo);
drivers/scsi/lpfc/lpfc_init.c
1193
lpfc_hb_timeout(struct timer_list *t)
drivers/scsi/lpfc/lpfc_init.c
1199
phba = timer_container_of(phba, t, hb_tmofunc);
drivers/scsi/lpfc/lpfc_init.c
1227
lpfc_rrq_timeout(struct timer_list *t)
drivers/scsi/lpfc/lpfc_init.c
1231
phba = timer_container_of(phba, t, rrq_tmr);
drivers/scsi/lpfc/lpfc_init.c
2947
int t;
drivers/scsi/lpfc/lpfc_init.c
2950
t = 16;
drivers/scsi/lpfc/lpfc_init.c
2952
HashWorkingPointer[t] =
drivers/scsi/lpfc/lpfc_init.c
2954
HashWorkingPointer[t - 3] ^ HashWorkingPointer[t -
drivers/scsi/lpfc/lpfc_init.c
2956
HashWorkingPointer[t - 14] ^ HashWorkingPointer[t - 16]);
drivers/scsi/lpfc/lpfc_init.c
2957
} while (++t <= 79);
drivers/scsi/lpfc/lpfc_init.c
2958
t = 0;
drivers/scsi/lpfc/lpfc_init.c
2966
if (t < 20) {
drivers/scsi/lpfc/lpfc_init.c
2968
} else if (t < 40) {
drivers/scsi/lpfc/lpfc_init.c
2970
} else if (t < 60) {
drivers/scsi/lpfc/lpfc_init.c
2975
TEMP += S(5, A) + E + HashWorkingPointer[t];
drivers/scsi/lpfc/lpfc_init.c
2981
} while (++t <= 79);
drivers/scsi/lpfc/lpfc_init.c
3017
int t;
drivers/scsi/lpfc/lpfc_init.c
3028
for (t = 0; t < 7; t++)
drivers/scsi/lpfc/lpfc_init.c
3029
lpfc_challenge_key(phba->RandomData + t, HashWorking + t);
drivers/scsi/lpfc/lpfc_init.c
5133
lpfc_sli4_fcf_redisc_wait_tmo(struct timer_list *t)
drivers/scsi/lpfc/lpfc_init.c
5135
struct lpfc_hba *phba = timer_container_of(phba, t, fcf.redisc_wait);
drivers/scsi/lpfc/lpfc_init.c
5164
lpfc_vmid_poll(struct timer_list *t)
drivers/scsi/lpfc/lpfc_init.c
5166
struct lpfc_hba *phba = timer_container_of(phba, t,
drivers/scsi/lpfc/lpfc_scsi.c
5196
void lpfc_poll_timeout(struct timer_list *t)
drivers/scsi/lpfc/lpfc_scsi.c
5198
struct lpfc_hba *phba = timer_container_of(phba, t, fcp_poll_timer);
drivers/scsi/lpfc/lpfc_sli.c
15719
void lpfc_sli4_poll_hbtimer(struct timer_list *t)
drivers/scsi/lpfc/lpfc_sli.c
15721
struct lpfc_hba *phba = timer_container_of(phba, t, cpuhp_poll_timer);
drivers/scsi/lpfc/lpfc_sli.c
3923
void lpfc_poll_eratt(struct timer_list *t)
drivers/scsi/lpfc/lpfc_sli.c
3929
phba = timer_container_of(phba, t, eratt_poll);
drivers/scsi/lpfc/lpfc_sli.c
9191
lpfc_mbox_timeout(struct timer_list *t)
drivers/scsi/lpfc/lpfc_sli.c
9193
struct lpfc_hba *phba = timer_container_of(phba, t, sli.mbox_tmo);
drivers/scsi/megaraid/megaraid_mbox.c
3356
uint8_t t;
drivers/scsi/megaraid/megaraid_mbox.c
3361
for (t = 0; t < LSI_MAX_LOGICAL_DRIVES_64LD; t++)
drivers/scsi/megaraid/megaraid_mbox.c
3362
adapter->device_ids[adapter->max_channel][t] =
drivers/scsi/megaraid/megaraid_mbox.c
3363
(t < adapter->init_id) ? t : t - 1;
drivers/scsi/megaraid/megaraid_mbox.c
3371
for (t = 0; t < LSI_MAX_LOGICAL_DRIVES_64LD; t++)
drivers/scsi/megaraid/megaraid_mbox.c
3372
adapter->device_ids[c][t] = (c << 8) | t;
drivers/scsi/megaraid/megaraid_mbox.c
3840
megaraid_sysfs_get_ldmap_timeout(struct timer_list *t)
drivers/scsi/megaraid/megaraid_mbox.c
3842
struct uioc_timeout *timeout = timer_container_of(timeout, t, timer);
drivers/scsi/megaraid/megaraid_mm.c
784
lld_timedout(struct timer_list *t)
drivers/scsi/megaraid/megaraid_mm.c
786
struct uioc_timeout *timeout = timer_container_of(timeout, t, timer);
drivers/scsi/megaraid/megaraid_sas_base.c
2327
static void megasas_sriov_heartbeat_handler(struct timer_list *t);
drivers/scsi/megaraid/megaraid_sas_base.c
2723
static void megasas_sriov_heartbeat_handler(struct timer_list *t)
drivers/scsi/megaraid/megaraid_sas_base.c
2726
timer_container_of(instance, t, sriov_heartbeat_timer);
drivers/scsi/mesh.c
1320
int t, nb;
drivers/scsi/mesh.c
1324
t = 50;
drivers/scsi/mesh.c
1325
while (t > 0 && in_8(&mr->fifo_count) != 0
drivers/scsi/mesh.c
1327
--t;
drivers/scsi/mesh.c
1445
int seq, n, t;
drivers/scsi/mesh.c
1492
t = 30; /* wait up to 30us */
drivers/scsi/mesh.c
1493
while ((in_8(&mr->bus_status0) & BS0_REQ) == 0 && --t >= 0)
drivers/scsi/mesh.c
1565
t = 230; /* wait up to 230us */
drivers/scsi/mesh.c
1567
if (--t < 0) {
drivers/scsi/mesh.c
238
static void dumplog(struct mesh_state *ms, int t)
drivers/scsi/mesh.c
240
struct mesh_target *tp = &ms->tgts[t];
drivers/scsi/mesh.c
253
t, lp->bs1, lp->bs0, lp->phase);
drivers/scsi/mesh.c
307
int t;
drivers/scsi/mesh.c
326
for (t = 0; t < 8; ++t) {
drivers/scsi/mesh.c
327
tp = &ms->tgts[t];
drivers/scsi/mesh.c
331
t, tp->current_req, tp->data_goes_out, tp->saved_ptr);
drivers/scsi/mesh.c
399
int t, id;
drivers/scsi/mesh.c
451
for (t = 100; t > 0; --t) {
drivers/scsi/mesh.c
506
for (t = 230; t > 0; --t) {
drivers/scsi/mesh.c
525
for (t = 10; t > 0 && in_8(&mr->interrupt) == 0; --t)
drivers/scsi/mesh.c
855
int b, t, prev;
drivers/scsi/mesh.c
931
for (t = 0; t < 8; ++t)
drivers/scsi/mesh.c
932
if ((b & (1 << t)) != 0 && t != ms->host->this_id)
drivers/scsi/mesh.c
934
if (b != (1 << t) + (1 << ms->host->this_id)) {
drivers/scsi/mesh.c
944
ms->conn_tgt = t;
drivers/scsi/mesh.c
945
tp = &ms->tgts[t];
drivers/scsi/mesh.c
947
if (ALLOW_DEBUG(t)) {
drivers/scsi/mesh.c
948
printk(KERN_DEBUG "mesh: reselected by target %d\n", t);
drivers/scsi/mesh.c
954
printk(KERN_ERR "mesh: reselected by tgt %d but no cmd!\n", t);
drivers/scsi/mvsas/mv_sas.c
1756
static void mvs_sig_time_out(struct timer_list *t)
drivers/scsi/mvsas/mv_sas.c
1758
struct mvs_phy *phy = timer_container_of(phy, t, timer);
drivers/scsi/ncr53c8xx.c
7924
static void ncr53c8xx_timeout(struct timer_list *t)
drivers/scsi/ncr53c8xx.c
7926
struct ncb *np = timer_container_of(np, t, timer);
drivers/scsi/ncr53c8xx.c
793
int c, h, t, u, v;
drivers/scsi/ncr53c8xx.c
798
t = NO_TARGET;
drivers/scsi/ncr53c8xx.c
805
t = ALL_TARGETS;
drivers/scsi/ncr53c8xx.c
809
if (t != target)
drivers/scsi/ncr53c8xx.c
810
t = (target == v) ? v : NO_TARGET;
drivers/scsi/ncr53c8xx.c
819
(t == ALL_TARGETS || t == target) &&
drivers/scsi/ncr53c8xx.c
824
t = ALL_TARGETS;
drivers/scsi/pm8001/pm8001_hwi.c
1471
struct sas_task *t = (struct sas_task *)pm8001_dev;
drivers/scsi/pm8001/pm8001_hwi.c
1478
if (pm8001_query_task(t) == TMF_RESP_FUNC_SUCC)
drivers/scsi/pm8001/pm8001_hwi.c
1482
spin_lock_irqsave(&t->task_state_lock, flags1);
drivers/scsi/pm8001/pm8001_hwi.c
1483
if (unlikely((t->task_state_flags & SAS_TASK_STATE_DONE))) {
drivers/scsi/pm8001/pm8001_hwi.c
1484
spin_unlock_irqrestore(&t->task_state_lock, flags1);
drivers/scsi/pm8001/pm8001_hwi.c
1488
spin_unlock_irqrestore(&t->task_state_lock, flags1);
drivers/scsi/pm8001/pm8001_hwi.c
1494
(ccb->task == t))
drivers/scsi/pm8001/pm8001_hwi.c
1501
ts = &t->task_status;
drivers/scsi/pm8001/pm8001_hwi.c
1508
spin_lock_irqsave(&t->task_state_lock, flags1);
drivers/scsi/pm8001/pm8001_hwi.c
1509
t->task_state_flags &= ~SAS_TASK_STATE_PENDING;
drivers/scsi/pm8001/pm8001_hwi.c
1510
t->task_state_flags |= SAS_TASK_STATE_DONE;
drivers/scsi/pm8001/pm8001_hwi.c
1511
if (unlikely((t->task_state_flags & SAS_TASK_STATE_ABORTED))) {
drivers/scsi/pm8001/pm8001_hwi.c
1512
spin_unlock_irqrestore(&t->task_state_lock, flags1);
drivers/scsi/pm8001/pm8001_hwi.c
1514
t, pw->handler, ts->resp, ts->stat);
drivers/scsi/pm8001/pm8001_hwi.c
1518
spin_unlock_irqrestore(&t->task_state_lock, flags1);
drivers/scsi/pm8001/pm8001_hwi.c
1522
t->task_done(t);
drivers/scsi/pm8001/pm8001_hwi.c
1527
struct sas_task *t = (struct sas_task *)pm8001_dev;
drivers/scsi/pm8001/pm8001_hwi.c
1535
ret = pm8001_query_task(t);
drivers/scsi/pm8001/pm8001_hwi.c
1546
spin_lock_irqsave(&t->task_state_lock, flags1);
drivers/scsi/pm8001/pm8001_hwi.c
1548
if (unlikely((t->task_state_flags & SAS_TASK_STATE_DONE))) {
drivers/scsi/pm8001/pm8001_hwi.c
1549
spin_unlock_irqrestore(&t->task_state_lock, flags1);
drivers/scsi/pm8001/pm8001_hwi.c
1552
(void)pm8001_abort_task(t);
drivers/scsi/pm8001/pm8001_hwi.c
1556
spin_unlock_irqrestore(&t->task_state_lock, flags1);
drivers/scsi/pm8001/pm8001_hwi.c
1562
(ccb->task == t))
drivers/scsi/pm8001/pm8001_hwi.c
1568
(void)pm8001_abort_task(t);
drivers/scsi/pm8001/pm8001_hwi.c
1579
ret = pm8001_abort_task(t);
drivers/scsi/pm8001/pm8001_hwi.c
1606
t = NULL;
drivers/scsi/pm8001/pm8001_hwi.c
1607
pm8001_open_reject_retry(pm8001_ha, t, pm8001_dev);
drivers/scsi/pm8001/pm8001_hwi.c
1716
struct sas_task *t;
drivers/scsi/pm8001/pm8001_hwi.c
1738
t = ccb->task;
drivers/scsi/pm8001/pm8001_hwi.c
1742
if (unlikely(!t || !t->lldd_task || !t->dev))
drivers/scsi/pm8001/pm8001_hwi.c
1744
ts = &t->task_status;
drivers/scsi/pm8001/pm8001_hwi.c
1749
SAS_ADDR(t->dev->sas_addr));
drivers/scsi/pm8001/pm8001_hwi.c
1754
status, tag, t);
drivers/scsi/pm8001/pm8001_hwi.c
1768
sas_ssp_task_response(pm8001_ha->dev, t, iu);
drivers/scsi/pm8001/pm8001_hwi.c
1831
if (!t->uldd_task)
drivers/scsi/pm8001/pm8001_hwi.c
1892
if (!t->uldd_task)
drivers/scsi/pm8001/pm8001_hwi.c
1928
spin_lock_irqsave(&t->task_state_lock, flags);
drivers/scsi/pm8001/pm8001_hwi.c
1929
t->task_state_flags &= ~SAS_TASK_STATE_PENDING;
drivers/scsi/pm8001/pm8001_hwi.c
1930
t->task_state_flags |= SAS_TASK_STATE_DONE;
drivers/scsi/pm8001/pm8001_hwi.c
1931
if (unlikely((t->task_state_flags & SAS_TASK_STATE_ABORTED))) {
drivers/scsi/pm8001/pm8001_hwi.c
1932
spin_unlock_irqrestore(&t->task_state_lock, flags);
drivers/scsi/pm8001/pm8001_hwi.c
1934
t, status, ts->resp, ts->stat);
drivers/scsi/pm8001/pm8001_hwi.c
1937
spin_unlock_irqrestore(&t->task_state_lock, flags);
drivers/scsi/pm8001/pm8001_hwi.c
1940
t->task_done(t);
drivers/scsi/pm8001/pm8001_hwi.c
1947
struct sas_task *t;
drivers/scsi/pm8001/pm8001_hwi.c
1960
t = ccb->task;
drivers/scsi/pm8001/pm8001_hwi.c
1964
if (unlikely(!t || !t->lldd_task || !t->dev))
drivers/scsi/pm8001/pm8001_hwi.c
1966
ts = &t->task_status;
drivers/scsi/pm8001/pm8001_hwi.c
1980
pm8001_handle_event(pm8001_ha, t, IO_XFER_ERROR_BREAK);
drivers/scsi/pm8001/pm8001_hwi.c
2012
if (!t->uldd_task)
drivers/scsi/pm8001/pm8001_hwi.c
2050
pm8001_handle_event(pm8001_ha, t, IO_XFER_OPEN_RETRY_TIMEOUT);
drivers/scsi/pm8001/pm8001_hwi.c
2095
spin_lock_irqsave(&t->task_state_lock, flags);
drivers/scsi/pm8001/pm8001_hwi.c
2096
t->task_state_flags &= ~SAS_TASK_STATE_PENDING;
drivers/scsi/pm8001/pm8001_hwi.c
2097
t->task_state_flags |= SAS_TASK_STATE_DONE;
drivers/scsi/pm8001/pm8001_hwi.c
2098
if (unlikely((t->task_state_flags & SAS_TASK_STATE_ABORTED))) {
drivers/scsi/pm8001/pm8001_hwi.c
2099
spin_unlock_irqrestore(&t->task_state_lock, flags);
drivers/scsi/pm8001/pm8001_hwi.c
2101
t, event, ts->resp, ts->stat);
drivers/scsi/pm8001/pm8001_hwi.c
2104
spin_unlock_irqrestore(&t->task_state_lock, flags);
drivers/scsi/pm8001/pm8001_hwi.c
2107
t->task_done(t);
drivers/scsi/pm8001/pm8001_hwi.c
2115
struct sas_task *t;
drivers/scsi/pm8001/pm8001_hwi.c
2138
t = ccb->task;
drivers/scsi/pm8001/pm8001_hwi.c
2141
if (t) {
drivers/scsi/pm8001/pm8001_hwi.c
2142
if (t->dev && (t->dev->lldd_dev))
drivers/scsi/pm8001/pm8001_hwi.c
2143
pm8001_dev = t->dev->lldd_dev;
drivers/scsi/pm8001/pm8001_hwi.c
2151
if (pm8001_dev && unlikely(!t || !t->lldd_task || !t->dev)) {
drivers/scsi/pm8001/pm8001_hwi.c
2156
ts = &t->task_status;
drivers/scsi/pm8001/pm8001_hwi.c
2161
status, tag, t);
drivers/scsi/pm8001/pm8001_hwi.c
2166
if (!dev_parent_is_expander(t->dev)) {
drivers/scsi/pm8001/pm8001_hwi.c
2199
SAS_ADDR(t->dev->sas_addr));
drivers/scsi/pm8001/pm8001_hwi.c
2218
if (t->ata_task.dma_xfer == 0 &&
drivers/scsi/pm8001/pm8001_hwi.c
2219
t->data_dir == DMA_FROM_DEVICE) {
drivers/scsi/pm8001/pm8001_hwi.c
2223
} else if (t->ata_task.use_ncq &&
drivers/scsi/pm8001/pm8001_hwi.c
2224
t->data_dir != DMA_NONE) {
drivers/scsi/pm8001/pm8001_hwi.c
2312
if (!t->uldd_task) {
drivers/scsi/pm8001/pm8001_hwi.c
2328
if (!t->uldd_task) {
drivers/scsi/pm8001/pm8001_hwi.c
2350
if (!t->uldd_task) {
drivers/scsi/pm8001/pm8001_hwi.c
2422
if (!t->uldd_task) {
drivers/scsi/pm8001/pm8001_hwi.c
2442
if (!t->uldd_task) {
drivers/scsi/pm8001/pm8001_hwi.c
2469
spin_lock_irqsave(&t->task_state_lock, flags);
drivers/scsi/pm8001/pm8001_hwi.c
2470
t->task_state_flags &= ~SAS_TASK_STATE_PENDING;
drivers/scsi/pm8001/pm8001_hwi.c
2471
t->task_state_flags |= SAS_TASK_STATE_DONE;
drivers/scsi/pm8001/pm8001_hwi.c
2472
if (unlikely((t->task_state_flags & SAS_TASK_STATE_ABORTED))) {
drivers/scsi/pm8001/pm8001_hwi.c
2473
spin_unlock_irqrestore(&t->task_state_lock, flags);
drivers/scsi/pm8001/pm8001_hwi.c
2476
t, status, ts->resp, ts->stat);
drivers/scsi/pm8001/pm8001_hwi.c
2479
spin_unlock_irqrestore(&t->task_state_lock, flags);
drivers/scsi/pm8001/pm8001_hwi.c
2487
struct sas_task *t;
drivers/scsi/pm8001/pm8001_hwi.c
2513
t = ccb->task;
drivers/scsi/pm8001/pm8001_hwi.c
2518
if (unlikely(!t)) {
drivers/scsi/pm8001/pm8001_hwi.c
2525
if (unlikely(!t->lldd_task || !t->dev))
drivers/scsi/pm8001/pm8001_hwi.c
2528
ts = &t->task_status;
drivers/scsi/pm8001/pm8001_hwi.c
2573
if (!t->uldd_task) {
drivers/scsi/pm8001/pm8001_hwi.c
2670
struct sas_task *t;
drivers/scsi/pm8001/pm8001_hwi.c
2684
t = ccb->task;
drivers/scsi/pm8001/pm8001_hwi.c
2685
ts = &t->task_status;
drivers/scsi/pm8001/pm8001_hwi.c
2691
status, tag, t);
drivers/scsi/pm8001/pm8001_hwi.c
2693
if (unlikely(!t || !t->lldd_task || !t->dev))
drivers/scsi/pm8001/pm8001_hwi.c
2835
spin_lock_irqsave(&t->task_state_lock, flags);
drivers/scsi/pm8001/pm8001_hwi.c
2836
t->task_state_flags &= ~SAS_TASK_STATE_PENDING;
drivers/scsi/pm8001/pm8001_hwi.c
2837
t->task_state_flags |= SAS_TASK_STATE_DONE;
drivers/scsi/pm8001/pm8001_hwi.c
2838
if (unlikely((t->task_state_flags & SAS_TASK_STATE_ABORTED))) {
drivers/scsi/pm8001/pm8001_hwi.c
2839
spin_unlock_irqrestore(&t->task_state_lock, flags);
drivers/scsi/pm8001/pm8001_hwi.c
2841
t, status, ts->resp, ts->stat);
drivers/scsi/pm8001/pm8001_hwi.c
2844
spin_unlock_irqrestore(&t->task_state_lock, flags);
drivers/scsi/pm8001/pm8001_hwi.c
3441
struct sas_task *t;
drivers/scsi/pm8001/pm8001_hwi.c
3457
t = ccb->task;
drivers/scsi/pm8001/pm8001_hwi.c
3460
if (!t) {
drivers/scsi/pm8001/pm8001_hwi.c
3465
if (t->task_proto == SAS_PROTOCOL_INTERNAL_ABORT)
drivers/scsi/pm8001/pm8001_hwi.c
3468
ts = &t->task_status;
drivers/scsi/pm8001/pm8001_hwi.c
3484
spin_lock_irqsave(&t->task_state_lock, flags);
drivers/scsi/pm8001/pm8001_hwi.c
3485
t->task_state_flags &= ~SAS_TASK_STATE_PENDING;
drivers/scsi/pm8001/pm8001_hwi.c
3486
t->task_state_flags |= SAS_TASK_STATE_DONE;
drivers/scsi/pm8001/pm8001_hwi.c
3487
spin_unlock_irqrestore(&t->task_state_lock, flags);
drivers/scsi/pm8001/pm8001_hwi.c
3491
t->task_done(t);
drivers/scsi/pm8001/pm80xx_hwi.c
1823
struct sas_task *t;
drivers/scsi/pm8001/pm80xx_hwi.c
1844
t = ccb->task;
drivers/scsi/pm8001/pm80xx_hwi.c
1848
if (unlikely(!t || !t->lldd_task || !t->dev))
drivers/scsi/pm8001/pm80xx_hwi.c
1850
ts = &t->task_status;
drivers/scsi/pm8001/pm80xx_hwi.c
1853
"tag::0x%x, status::0x%x task::0x%p\n", tag, status, t);
drivers/scsi/pm8001/pm80xx_hwi.c
1859
SAS_ADDR(t->dev->sas_addr));
drivers/scsi/pm8001/pm80xx_hwi.c
1873
sas_ssp_task_response(pm8001_ha->dev, t, iu);
drivers/scsi/pm8001/pm80xx_hwi.c
1964
if (!t->uldd_task)
drivers/scsi/pm8001/pm80xx_hwi.c
2044
if (!t->uldd_task)
drivers/scsi/pm8001/pm80xx_hwi.c
2090
spin_lock_irqsave(&t->task_state_lock, flags);
drivers/scsi/pm8001/pm80xx_hwi.c
2091
t->task_state_flags &= ~SAS_TASK_STATE_PENDING;
drivers/scsi/pm8001/pm80xx_hwi.c
2092
t->task_state_flags |= SAS_TASK_STATE_DONE;
drivers/scsi/pm8001/pm80xx_hwi.c
2093
if (unlikely((t->task_state_flags & SAS_TASK_STATE_ABORTED))) {
drivers/scsi/pm8001/pm80xx_hwi.c
2094
spin_unlock_irqrestore(&t->task_state_lock, flags);
drivers/scsi/pm8001/pm80xx_hwi.c
2097
t, status, ts->resp, ts->stat);
drivers/scsi/pm8001/pm80xx_hwi.c
2099
if (t->slow_task)
drivers/scsi/pm8001/pm80xx_hwi.c
2100
complete(&t->slow_task->completion);
drivers/scsi/pm8001/pm80xx_hwi.c
2102
spin_unlock_irqrestore(&t->task_state_lock, flags);
drivers/scsi/pm8001/pm80xx_hwi.c
2110
struct sas_task *t;
drivers/scsi/pm8001/pm80xx_hwi.c
2122
t = ccb->task;
drivers/scsi/pm8001/pm80xx_hwi.c
2126
if (unlikely(!t || !t->lldd_task || !t->dev))
drivers/scsi/pm8001/pm80xx_hwi.c
2128
ts = &t->task_status;
drivers/scsi/pm8001/pm80xx_hwi.c
2142
pm8001_handle_event(pm8001_ha, t, IO_XFER_ERROR_BREAK);
drivers/scsi/pm8001/pm80xx_hwi.c
2180
if (!t->uldd_task)
drivers/scsi/pm8001/pm80xx_hwi.c
2219
pm8001_handle_event(pm8001_ha, t, IO_XFER_OPEN_RETRY_TIMEOUT);
drivers/scsi/pm8001/pm80xx_hwi.c
2271
spin_lock_irqsave(&t->task_state_lock, flags);
drivers/scsi/pm8001/pm80xx_hwi.c
2272
t->task_state_flags &= ~SAS_TASK_STATE_PENDING;
drivers/scsi/pm8001/pm80xx_hwi.c
2273
t->task_state_flags |= SAS_TASK_STATE_DONE;
drivers/scsi/pm8001/pm80xx_hwi.c
2274
if (unlikely((t->task_state_flags & SAS_TASK_STATE_ABORTED))) {
drivers/scsi/pm8001/pm80xx_hwi.c
2275
spin_unlock_irqrestore(&t->task_state_lock, flags);
drivers/scsi/pm8001/pm80xx_hwi.c
2278
t, event, ts->resp, ts->stat);
drivers/scsi/pm8001/pm80xx_hwi.c
2281
spin_unlock_irqrestore(&t->task_state_lock, flags);
drivers/scsi/pm8001/pm80xx_hwi.c
2291
struct sas_task *t;
drivers/scsi/pm8001/pm80xx_hwi.c
2314
t = ccb->task;
drivers/scsi/pm8001/pm80xx_hwi.c
2317
if (t) {
drivers/scsi/pm8001/pm80xx_hwi.c
2318
if (t->dev && (t->dev->lldd_dev)) {
drivers/scsi/pm8001/pm80xx_hwi.c
2319
pm8001_dev = t->dev->lldd_dev;
drivers/scsi/pm8001/pm80xx_hwi.c
2320
qc = t->uldd_task;
drivers/scsi/pm8001/pm80xx_hwi.c
2330
if (pm8001_dev && unlikely(!t->lldd_task || !t->dev))
drivers/scsi/pm8001/pm80xx_hwi.c
2333
ts = &t->task_status;
drivers/scsi/pm8001/pm80xx_hwi.c
2343
if (!dev_parent_is_expander(t->dev)) {
drivers/scsi/pm8001/pm80xx_hwi.c
2377
SAS_ADDR(t->dev->sas_addr));
drivers/scsi/pm8001/pm80xx_hwi.c
2396
if (t->ata_task.dma_xfer == 0 &&
drivers/scsi/pm8001/pm80xx_hwi.c
2397
t->data_dir == DMA_FROM_DEVICE) {
drivers/scsi/pm8001/pm80xx_hwi.c
2401
} else if (t->ata_task.use_ncq &&
drivers/scsi/pm8001/pm80xx_hwi.c
2402
t->data_dir != DMA_NONE) {
drivers/scsi/pm8001/pm80xx_hwi.c
2496
if (!t->uldd_task) {
drivers/scsi/pm8001/pm80xx_hwi.c
2516
if (!t->uldd_task) {
drivers/scsi/pm8001/pm80xx_hwi.c
2544
if (!t->uldd_task) {
drivers/scsi/pm8001/pm80xx_hwi.c
2620
if (!t->uldd_task) {
drivers/scsi/pm8001/pm80xx_hwi.c
2644
if (!t->uldd_task) {
drivers/scsi/pm8001/pm80xx_hwi.c
2677
spin_lock_irqsave(&t->task_state_lock, flags);
drivers/scsi/pm8001/pm80xx_hwi.c
2678
t->task_state_flags &= ~SAS_TASK_STATE_PENDING;
drivers/scsi/pm8001/pm80xx_hwi.c
2679
t->task_state_flags |= SAS_TASK_STATE_DONE;
drivers/scsi/pm8001/pm80xx_hwi.c
2680
if (unlikely((t->task_state_flags & SAS_TASK_STATE_ABORTED))) {
drivers/scsi/pm8001/pm80xx_hwi.c
2681
spin_unlock_irqrestore(&t->task_state_lock, flags);
drivers/scsi/pm8001/pm80xx_hwi.c
2684
t, status, ts->resp, ts->stat);
drivers/scsi/pm8001/pm80xx_hwi.c
2686
if (t->slow_task)
drivers/scsi/pm8001/pm80xx_hwi.c
2687
complete(&t->slow_task->completion);
drivers/scsi/pm8001/pm80xx_hwi.c
2689
spin_unlock_irqrestore(&t->task_state_lock, flags);
drivers/scsi/pm8001/pm80xx_hwi.c
2702
struct sas_task *t;
drivers/scsi/pm8001/pm80xx_hwi.c
2735
t = ccb->task;
drivers/scsi/pm8001/pm80xx_hwi.c
2737
if (unlikely(!t)) {
drivers/scsi/pm8001/pm80xx_hwi.c
2744
if (unlikely(!t->lldd_task || !t->dev))
drivers/scsi/pm8001/pm80xx_hwi.c
2747
ts = &t->task_status;
drivers/scsi/pm8001/pm80xx_hwi.c
2798
if (!t->uldd_task) {
drivers/scsi/pm8001/pm80xx_hwi.c
2910
struct sas_task *t;
drivers/scsi/pm8001/pm80xx_hwi.c
2925
t = ccb->task;
drivers/scsi/pm8001/pm80xx_hwi.c
2926
ts = &t->task_status;
drivers/scsi/pm8001/pm80xx_hwi.c
2930
if (unlikely(!t || !t->lldd_task || !t->dev))
drivers/scsi/pm8001/pm80xx_hwi.c
2944
struct scatterlist *sg_resp = &t->smp_task.smp_resp;
drivers/scsi/pm8001/pm80xx_hwi.c
3100
spin_lock_irqsave(&t->task_state_lock, flags);
drivers/scsi/pm8001/pm80xx_hwi.c
3101
t->task_state_flags &= ~SAS_TASK_STATE_PENDING;
drivers/scsi/pm8001/pm80xx_hwi.c
3102
t->task_state_flags |= SAS_TASK_STATE_DONE;
drivers/scsi/pm8001/pm80xx_hwi.c
3103
if (unlikely((t->task_state_flags & SAS_TASK_STATE_ABORTED))) {
drivers/scsi/pm8001/pm80xx_hwi.c
3104
spin_unlock_irqrestore(&t->task_state_lock, flags);
drivers/scsi/pm8001/pm80xx_hwi.c
3107
t, status, ts->resp, ts->stat);
drivers/scsi/pm8001/pm80xx_hwi.c
3110
spin_unlock_irqrestore(&t->task_state_lock, flags);
drivers/scsi/pm8001/pm80xx_hwi.c
3113
t->task_done(t);
drivers/scsi/pmcraid.c
545
static void pmcraid_bist_done(struct timer_list *t)
drivers/scsi/pmcraid.c
547
struct pmcraid_cmd *cmd = timer_container_of(cmd, t, timer);
drivers/scsi/pmcraid.c
602
static void pmcraid_reset_alert_done(struct timer_list *t)
drivers/scsi/pmcraid.c
604
struct pmcraid_cmd *cmd = timer_container_of(cmd, t, timer);
drivers/scsi/pmcraid.c
686
static void pmcraid_timeout_handler(struct timer_list *t)
drivers/scsi/pmcraid.c
688
struct pmcraid_cmd *cmd = timer_container_of(cmd, t, timer);
drivers/scsi/qla1280.c
724
static void qla1280_mailbox_timeout(struct timer_list *t)
drivers/scsi/qla1280.c
726
struct scsi_qla_host *ha = timer_container_of(ha, t, mailbox_timer);
drivers/scsi/qla2xxx/qla_edif.c
95
static void qla2x00_sa_replace_iocb_timeout(struct timer_list *t)
drivers/scsi/qla2xxx/qla_edif.c
97
struct edif_list_entry *edif_entry = timer_container_of(edif_entry, t,
drivers/scsi/qla2xxx/qla_init.c
2263
struct tmf_arg *t;
drivers/scsi/qla2xxx/qla_init.c
2266
list_for_each_entry(t, &ha->tmf_active, tmf_elem) {
drivers/scsi/qla2xxx/qla_init.c
2267
if (t->fcport == arg->fcport && t->lun == arg->lun) {
drivers/scsi/qla2xxx/qla_init.c
46
qla2x00_sp_timeout(struct timer_list *t)
drivers/scsi/qla2xxx/qla_init.c
48
srb_t *sp = timer_container_of(sp, t, u.iocb_cmd.timer);
drivers/scsi/qla2xxx/qla_os.c
1329
__qla2x00_eh_wait_for_pending_commands(struct qla_qpair *qpair, unsigned int t,
drivers/scsi/qla2xxx/qla_os.c
1364
match = sp->fcport->d_id.b24 == t;
drivers/scsi/qla2xxx/qla_os.c
1370
match = (sp->fcport->d_id.b24 == t &&
drivers/scsi/qla2xxx/qla_os.c
1410
qla2x00_eh_wait_for_pending_commands(scsi_qla_host_t *vha, unsigned int t,
drivers/scsi/qla2xxx/qla_os.c
1417
status = __qla2x00_eh_wait_for_pending_commands(ha->base_qpair, t, l,
drivers/scsi/qla2xxx/qla_os.c
1423
status = __qla2x00_eh_wait_for_pending_commands(qpair, t, l,
drivers/scsi/qla2xxx/qla_os.c
3633
struct task_struct *t = ha->dpc_thread;
drivers/scsi/qla2xxx/qla_os.c
3636
kthread_stop(t);
drivers/scsi/qla2xxx/qla_os.c
3816
struct task_struct *t = ha->dpc_thread;
drivers/scsi/qla2xxx/qla_os.c
3823
kthread_stop(t);
drivers/scsi/qla2xxx/qla_os.c
7245
struct task_struct *t = ha->dpc_thread;
drivers/scsi/qla2xxx/qla_os.c
7247
if (!test_bit(UNLOADING, &vha->dpc_flags) && t)
drivers/scsi/qla2xxx/qla_os.c
7248
wake_up_process(t);
drivers/scsi/qla2xxx/qla_os.c
7385
qla2x00_timer(struct timer_list *t)
drivers/scsi/qla2xxx/qla_os.c
7387
scsi_qla_host_t *vha = timer_container_of(vha, t, timer);
drivers/scsi/qla2xxx/qla_target.c
255
struct qla_tgt_sess_op *u, *t;
drivers/scsi/qla2xxx/qla_target.c
261
list_for_each_entry_safe(u, t, &vha->unknown_atio_list, cmd_list) {
drivers/scsi/qla2xxx/qla_target.c
653
fc_port_t *t;
drivers/scsi/qla2xxx/qla_target.c
657
t = e->u.nack.fcport;
drivers/scsi/qla2xxx/qla_target.c
658
flush_work(&t->del_work);
drivers/scsi/qla2xxx/qla_target.c
659
flush_work(&t->free_work);
drivers/scsi/qla2xxx/qla_target.c
661
t = qlt_create_sess(vha, e->u.nack.fcport, 0);
drivers/scsi/qla2xxx/qla_target.c
663
if (t) {
drivers/scsi/qla2xxx/qla_target.c
665
"%s create sess success %p", __func__, t);
drivers/scsi/qla4xxx/ql4_os.c
4012
static void qla4xxx_timer(struct timer_list *t);
drivers/scsi/qla4xxx/ql4_os.c
4556
static void qla4xxx_timer(struct timer_list *t)
drivers/scsi/qla4xxx/ql4_os.c
4558
struct scsi_qla_host *ha = timer_container_of(ha, t, timer);
drivers/scsi/scsi_priv.h
181
static inline void scsi_autopm_get_target(struct scsi_target *t) {}
drivers/scsi/scsi_priv.h
182
static inline void scsi_autopm_put_target(struct scsi_target *t) {}
drivers/scsi/scsi_sas_internal.h
13
struct scsi_transport_template t;
drivers/scsi/scsi_sas_internal.h
41
#define to_sas_internal(tmpl) container_of(tmpl, struct sas_internal, t)
drivers/scsi/scsi_transport_fc.c
2456
return &i->t.host_attrs.ac == cont;
drivers/scsi/scsi_transport_fc.c
2475
return &i->t.target_attrs.ac == cont;
drivers/scsi/scsi_transport_fc.c
2659
i->t.target_attrs.ac.attrs = &i->starget_attrs[0];
drivers/scsi/scsi_transport_fc.c
2660
i->t.target_attrs.ac.class = &fc_transport_class.class;
drivers/scsi/scsi_transport_fc.c
2661
i->t.target_attrs.ac.match = fc_target_match;
drivers/scsi/scsi_transport_fc.c
2662
i->t.target_size = sizeof(struct fc_starget_attrs);
drivers/scsi/scsi_transport_fc.c
2663
transport_container_register(&i->t.target_attrs);
drivers/scsi/scsi_transport_fc.c
2665
i->t.host_attrs.ac.attrs = &i->host_attrs[0];
drivers/scsi/scsi_transport_fc.c
2666
i->t.host_attrs.ac.class = &fc_host_class.class;
drivers/scsi/scsi_transport_fc.c
2667
i->t.host_attrs.ac.match = fc_host_match;
drivers/scsi/scsi_transport_fc.c
2668
i->t.host_size = sizeof(struct fc_host_attrs);
drivers/scsi/scsi_transport_fc.c
2670
i->t.host_attrs.statistics = &fc_statistics_group;
drivers/scsi/scsi_transport_fc.c
2671
transport_container_register(&i->t.host_attrs);
drivers/scsi/scsi_transport_fc.c
2689
i->t.create_work_queue = 1;
drivers/scsi/scsi_transport_fc.c
2691
i->t.user_scan = fc_user_scan;
drivers/scsi/scsi_transport_fc.c
2790
return &i->t;
drivers/scsi/scsi_transport_fc.c
2794
void fc_release_transport(struct scsi_transport_template *t)
drivers/scsi/scsi_transport_fc.c
2796
struct fc_internal *i = to_fc_internal(t);
drivers/scsi/scsi_transport_fc.c
2798
transport_container_unregister(&i->t.target_attrs);
drivers/scsi/scsi_transport_fc.c
2799
transport_container_unregister(&i->t.host_attrs);
drivers/scsi/scsi_transport_fc.c
326
struct scsi_transport_template t;
drivers/scsi/scsi_transport_fc.c
355
#define to_fc_internal(tmpl) container_of(tmpl, struct fc_internal, t)
drivers/scsi/scsi_transport_iscsi.c
1027
struct iscsi_transport *t = fnode_conn->transport; \
drivers/scsi/scsi_transport_iscsi.c
1028
return t->get_flashnode_param(fnode_sess, param, buf); \
drivers/scsi/scsi_transport_iscsi.c
106
container_of(tmpl, struct iscsi_internal, t)
drivers/scsi/scsi_transport_iscsi.c
1115
struct iscsi_transport *t = fnode_conn->transport;
drivers/scsi/scsi_transport_iscsi.c
1175
return t->attr_is_visible(ISCSI_FLASHNODE_PARAM, param);
drivers/scsi/scsi_transport_iscsi.c
325
struct iscsi_transport *t = iface->transport; \
drivers/scsi/scsi_transport_iscsi.c
326
return t->get_iface_param(iface, param_type, param, buf); \
drivers/scsi/scsi_transport_iscsi.c
4120
struct iscsi_transport *t = conn->transport; \
drivers/scsi/scsi_transport_iscsi.c
4121
return t->get_conn_param(conn, param, buf); \
drivers/scsi/scsi_transport_iscsi.c
4187
struct iscsi_transport *t = conn->transport; \
drivers/scsi/scsi_transport_iscsi.c
4197
if (!ep && t->ep_connect) { \
drivers/scsi/scsi_transport_iscsi.c
4203
rc = t->get_ep_param(ep, param, buf); \
drivers/scsi/scsi_transport_iscsi.c
4205
rc = t->get_conn_param(conn, param, buf); \
drivers/scsi/scsi_transport_iscsi.c
4258
struct iscsi_transport *t = conn->transport;
drivers/scsi/scsi_transport_iscsi.c
4328
return t->attr_is_visible(ISCSI_PARAM, param);
drivers/scsi/scsi_transport_iscsi.c
4346
struct iscsi_transport *t = session->transport; \
drivers/scsi/scsi_transport_iscsi.c
4350
return t->get_session_param(session, param, buf); \
drivers/scsi/scsi_transport_iscsi.c
446
struct iscsi_transport *t = iface->transport;
drivers/scsi/scsi_transport_iscsi.c
4541
struct iscsi_transport *t = session->transport;
drivers/scsi/scsi_transport_iscsi.c
4639
return t->attr_is_visible(ISCSI_PARAM, param);
drivers/scsi/scsi_transport_iscsi.c
4834
return &priv->t.host_attrs.ac == cont;
drivers/scsi/scsi_transport_iscsi.c
4856
priv->t.user_scan = iscsi_user_scan;
drivers/scsi/scsi_transport_iscsi.c
4869
priv->t.host_attrs.ac.class = &iscsi_host_class.class;
drivers/scsi/scsi_transport_iscsi.c
487
return t->attr_is_visible(ISCSI_IFACE_PARAM, param);
drivers/scsi/scsi_transport_iscsi.c
4870
priv->t.host_attrs.ac.match = iscsi_host_match;
drivers/scsi/scsi_transport_iscsi.c
4871
priv->t.host_attrs.ac.grp = &iscsi_host_group;
drivers/scsi/scsi_transport_iscsi.c
4872
priv->t.host_size = sizeof(struct iscsi_cls_host);
drivers/scsi/scsi_transport_iscsi.c
4873
transport_container_register(&priv->t.host_attrs);
drivers/scsi/scsi_transport_iscsi.c
4892
return &priv->t;
drivers/scsi/scsi_transport_iscsi.c
4921
transport_container_unregister(&priv->t.host_attrs);
drivers/scsi/scsi_transport_iscsi.c
609
return t->attr_is_visible(ISCSI_NET_PARAM, param);
drivers/scsi/scsi_transport_iscsi.c
80
struct scsi_transport_template t;
drivers/scsi/scsi_transport_iscsi.c
814
struct iscsi_transport *t = fnode_sess->transport; \
drivers/scsi/scsi_transport_iscsi.c
815
return t->get_flashnode_param(fnode_sess, param, buf); \
drivers/scsi/scsi_transport_iscsi.c
914
struct iscsi_transport *t = fnode_sess->transport;
drivers/scsi/scsi_transport_iscsi.c
988
return t->attr_is_visible(ISCSI_FLASHNODE_PARAM, param);
drivers/scsi/scsi_transport_sas.c
1822
i->t.user_scan = sas_user_scan;
drivers/scsi/scsi_transport_sas.c
1824
i->t.host_attrs.ac.attrs = &i->host_attrs[0];
drivers/scsi/scsi_transport_sas.c
1825
i->t.host_attrs.ac.class = &sas_host_class.class;
drivers/scsi/scsi_transport_sas.c
1826
i->t.host_attrs.ac.match = sas_host_match;
drivers/scsi/scsi_transport_sas.c
1827
transport_container_register(&i->t.host_attrs);
drivers/scsi/scsi_transport_sas.c
1828
i->t.host_size = sizeof(struct sas_host_attrs);
drivers/scsi/scsi_transport_sas.c
1913
return &i->t;
drivers/scsi/scsi_transport_sas.c
1921
void sas_release_transport(struct scsi_transport_template *t)
drivers/scsi/scsi_transport_sas.c
1923
struct sas_internal *i = to_sas_internal(t);
drivers/scsi/scsi_transport_sas.c
1925
transport_container_unregister(&i->t.host_attrs);
drivers/scsi/scsi_transport_sas.c
280
return &i->t.host_attrs.ac == cont;
drivers/scsi/scsi_transport_spi.c
1423
return &i->t.target_attrs.ac == cont;
drivers/scsi/scsi_transport_spi.c
1578
i->t.target_attrs.ac.class = &spi_transport_class.class;
drivers/scsi/scsi_transport_spi.c
1579
i->t.target_attrs.ac.grp = &target_attribute_group;
drivers/scsi/scsi_transport_spi.c
1580
i->t.target_attrs.ac.match = spi_target_match;
drivers/scsi/scsi_transport_spi.c
1581
transport_container_register(&i->t.target_attrs);
drivers/scsi/scsi_transport_spi.c
1582
i->t.target_size = sizeof(struct spi_transport_attrs);
drivers/scsi/scsi_transport_spi.c
1583
i->t.host_attrs.ac.class = &spi_host_class.class;
drivers/scsi/scsi_transport_spi.c
1584
i->t.host_attrs.ac.grp = &host_attribute_group;
drivers/scsi/scsi_transport_spi.c
1585
i->t.host_attrs.ac.match = spi_host_match;
drivers/scsi/scsi_transport_spi.c
1586
transport_container_register(&i->t.host_attrs);
drivers/scsi/scsi_transport_spi.c
1587
i->t.host_size = sizeof(struct spi_host_attrs);
drivers/scsi/scsi_transport_spi.c
1590
return &i->t;
drivers/scsi/scsi_transport_spi.c
1594
void spi_release_transport(struct scsi_transport_template *t)
drivers/scsi/scsi_transport_spi.c
1596
struct spi_internal *i = to_spi_internal(t);
drivers/scsi/scsi_transport_spi.c
1598
transport_container_unregister(&i->t.target_attrs);
drivers/scsi/scsi_transport_spi.c
1599
transport_container_unregister(&i->t.host_attrs);
drivers/scsi/scsi_transport_spi.c
60
struct scsi_transport_template t;
drivers/scsi/scsi_transport_spi.c
64
#define to_spi_internal(tmpl) container_of(tmpl, struct spi_internal, t)
drivers/scsi/scsi_transport_srp.c
31
struct scsi_transport_template t;
drivers/scsi/scsi_transport_srp.c
42
#define to_srp_internal(tmpl) container_of(tmpl, struct srp_internal, t)
drivers/scsi/scsi_transport_srp.c
665
return &i->t.host_attrs.ac == cont;
drivers/scsi/scsi_transport_srp.c
821
i->t.host_size = sizeof(struct srp_host_attrs);
drivers/scsi/scsi_transport_srp.c
822
i->t.host_attrs.ac.attrs = &i->host_attrs[0];
drivers/scsi/scsi_transport_srp.c
823
i->t.host_attrs.ac.class = &srp_host_class.class;
drivers/scsi/scsi_transport_srp.c
824
i->t.host_attrs.ac.match = srp_host_match;
drivers/scsi/scsi_transport_srp.c
826
transport_container_register(&i->t.host_attrs);
drivers/scsi/scsi_transport_srp.c
853
return &i->t;
drivers/scsi/scsi_transport_srp.c
861
void srp_release_transport(struct scsi_transport_template *t)
drivers/scsi/scsi_transport_srp.c
863
struct srp_internal *i = to_srp_internal(t);
drivers/scsi/scsi_transport_srp.c
865
transport_container_unregister(&i->t.host_attrs);
drivers/scsi/smartpqi/smartpqi_init.c
3830
static void pqi_heartbeat_timer_handler(struct timer_list *t)
drivers/scsi/smartpqi/smartpqi_init.c
3834
struct pqi_ctrl_info *ctrl_info = timer_container_of(ctrl_info, t,
drivers/scsi/snic/snic_disc.h
96
#define snic_tgt_to_shost(t) \
drivers/scsi/snic/snic_disc.h
97
dev_to_shost(t->dev.parent)
drivers/scsi/st.c
314
#define st_printk(prefix, t, fmt, a...) \
drivers/scsi/st.c
315
sdev_prefix_printk(prefix, (t)->device, (t)->name, fmt, ##a)
drivers/scsi/st.c
317
#define DEBC_printk(t, fmt, a...) \
drivers/scsi/st.c
318
if (debugging) { st_printk(ST_DEB_MSG, t, fmt, ##a ); }
drivers/scsi/st.c
320
#define DEBC_printk(t, fmt, a...)
drivers/scsi/storvsc_drv.c
1644
int ret, t;
drivers/scsi/storvsc_drv.c
1668
t = wait_for_completion_timeout(&request->wait_event, storvsc_timeout * HZ);
drivers/scsi/storvsc_drv.c
1669
if (t == 0)
drivers/scsi/storvsc_drv.c
725
int ret, t;
drivers/scsi/storvsc_drv.c
771
t = wait_for_completion_timeout(&request->wait_event, storvsc_timeout * HZ);
drivers/scsi/storvsc_drv.c
772
if (t == 0) {
drivers/scsi/storvsc_drv.c
817
int ret, t;
drivers/scsi/storvsc_drv.c
836
t = wait_for_completion_timeout(&request->wait_event, storvsc_timeout * HZ);
drivers/scsi/storvsc_drv.c
837
if (t == 0)
drivers/scsi/sym53c8xx_2/sym_glue.c
546
static void sym53c8xx_timer(struct timer_list *t)
drivers/scsi/sym53c8xx_2/sym_glue.c
548
struct sym_hcb *np = timer_container_of(np, t, s.timer);
drivers/scsi/sym53c8xx_2/sym_glue.c
942
int t, l;
drivers/scsi/sym53c8xx_2/sym_glue.c
961
for (t = 0; t < SYM_CONF_MAX_TARGET; t++) {
drivers/scsi/sym53c8xx_2/sym_glue.c
962
if (!((uc->target >> t) & 1))
drivers/scsi/sym53c8xx_2/sym_glue.c
964
tp = &np->target[t];
drivers/soc/apple/mailbox.c
108
long t;
drivers/soc/apple/mailbox.c
144
t = wait_for_completion_interruptible_timeout(
drivers/soc/apple/mailbox.c
147
if (t < 0)
drivers/soc/apple/mailbox.c
148
return t;
drivers/soc/apple/mailbox.c
149
else if (t == 0)
drivers/soc/apple/rtkit.c
721
long t;
drivers/soc/apple/rtkit.c
723
t = wait_for_completion_interruptible_timeout(c,
drivers/soc/apple/rtkit.c
725
if (t < 0)
drivers/soc/apple/rtkit.c
726
return t;
drivers/soc/apple/rtkit.c
727
else if (t == 0)
drivers/soc/fsl/qbman/qman_test_api.c
74
u64 t = qm_fd_addr_get64(fd);
drivers/soc/fsl/qbman/qman_test_api.c
75
int z = t >> 40;
drivers/soc/fsl/qbman/qman_test_api.c
79
t <<= 1;
drivers/soc/fsl/qbman/qman_test_api.c
81
t |= 1;
drivers/soc/fsl/qbman/qman_test_api.c
82
qm_fd_addr_set64(fd, t);
drivers/spi/spi-altera-core.c
143
struct spi_device *spi, struct spi_transfer *t)
drivers/spi/spi-altera-core.c
148
hw->tx = t->tx_buf;
drivers/spi/spi-altera-core.c
149
hw->rx = t->rx_buf;
drivers/spi/spi-altera-core.c
151
hw->bytes_per_word = DIV_ROUND_UP(t->bits_per_word, 8);
drivers/spi/spi-altera-core.c
152
hw->len = t->len / hw->bytes_per_word;
drivers/spi/spi-apple.c
196
static bool apple_spi_prep_transfer(struct apple_spi *spi, struct spi_transfer *t)
drivers/spi/spi-apple.c
201
cr = DIV_ROUND_UP(clk_get_rate(spi->clk), t->speed_hz);
drivers/spi/spi-apple.c
206
FIELD_PREP(APPLE_SPI_SHIFTCFG_BITS, t->bits_per_word));
drivers/spi/spi-apple.c
215
return (200000 * t->bits_per_word * fifo_threshold) <= t->speed_hz;
drivers/spi/spi-apple.c
358
struct spi_transfer *t)
drivers/spi/spi-apple.c
361
bool poll = apple_spi_prep_transfer(spi, t);
drivers/spi/spi-apple.c
362
const void *tx_ptr = t->tx_buf;
drivers/spi/spi-apple.c
363
void *rx_ptr = t->rx_buf;
drivers/spi/spi-apple.c
371
if (t->bits_per_word > 16)
drivers/spi/spi-apple.c
373
else if (t->bits_per_word > 8)
drivers/spi/spi-apple.c
378
words = t->len / bytes_per_word;
drivers/spi/spi-ar934x.c
100
div = ar934x_spi_clk_div(sp, t->speed_hz);
drivers/spi/spi-ar934x.c
114
for (trx_done = 0; trx_done < t->len; trx_done += bpw) {
drivers/spi/spi-ar934x.c
115
trx_cur = t->len - trx_done;
drivers/spi/spi-ar934x.c
118
else if (list_is_last(&t->transfer_list, &m->transfers))
drivers/spi/spi-ar934x.c
121
if (t->tx_buf) {
drivers/spi/spi-ar934x.c
122
tx_buf = t->tx_buf + trx_done;
drivers/spi/spi-ar934x.c
138
if (t->rx_buf) {
drivers/spi/spi-ar934x.c
140
buf = t->rx_buf + trx_done;
drivers/spi/spi-ar934x.c
146
spi_delay_exec(&t->word_delay, t);
drivers/spi/spi-ar934x.c
148
m->actual_length += t->len;
drivers/spi/spi-ar934x.c
149
spi_transfer_delay_exec(t);
drivers/spi/spi-ar934x.c
82
struct spi_transfer *t = NULL;
drivers/spi/spi-ar934x.c
93
list_for_each_entry(t, &m->transfers, transfer_list) {
drivers/spi/spi-ar934x.c
94
if (t->bits_per_word >= 8 && t->bits_per_word < 32)
drivers/spi/spi-ar934x.c
95
bpw = t->bits_per_word >> 3;
drivers/spi/spi-ar934x.c
99
if (t->speed_hz)
drivers/spi/spi-au1550.c
220
static int au1550_spi_setupxfer(struct spi_device *spi, struct spi_transfer *t)
drivers/spi/spi-au1550.c
226
if (t) {
drivers/spi/spi-au1550.c
227
bpw = t->bits_per_word;
drivers/spi/spi-au1550.c
228
hz = t->speed_hz;
drivers/spi/spi-au1550.c
304
static int au1550_spi_dma_txrxb(struct spi_device *spi, struct spi_transfer *t)
drivers/spi/spi-au1550.c
311
hw->len = t->len;
drivers/spi/spi-au1550.c
315
hw->tx = t->tx_buf;
drivers/spi/spi-au1550.c
316
hw->rx = t->rx_buf;
drivers/spi/spi-au1550.c
325
if (t->tx_buf) {
drivers/spi/spi-au1550.c
326
dma_tx_addr = dma_map_single(hw->dev, (void *)t->tx_buf,
drivers/spi/spi-au1550.c
327
t->len, DMA_TO_DEVICE);
drivers/spi/spi-au1550.c
332
if (t->rx_buf) {
drivers/spi/spi-au1550.c
333
dma_rx_addr = dma_map_single(hw->dev, (void *)t->rx_buf,
drivers/spi/spi-au1550.c
334
t->len, DMA_FROM_DEVICE);
drivers/spi/spi-au1550.c
338
if (t->len > hw->dma_rx_tmpbuf_size) {
drivers/spi/spi-au1550.c
342
ret = au1550_spi_dma_rxtmp_alloc(hw, max(t->len,
drivers/spi/spi-au1550.c
350
t->len, DMA_FROM_DEVICE);
drivers/spi/spi-au1550.c
353
if (!t->tx_buf) {
drivers/spi/spi-au1550.c
355
t->len, DMA_BIDIRECTIONAL);
drivers/spi/spi-au1550.c
361
t->len, DDMA_FLAGS_IE);
drivers/spi/spi-au1550.c
366
t->len, DDMA_FLAGS_IE);
drivers/spi/spi-au1550.c
386
if (!t->rx_buf) {
drivers/spi/spi-au1550.c
388
dma_sync_single_for_cpu(hw->dev, dma_rx_addr, t->len,
drivers/spi/spi-au1550.c
392
if (t->rx_buf)
drivers/spi/spi-au1550.c
393
dma_unmap_single(hw->dev, dma_rx_addr, t->len,
drivers/spi/spi-au1550.c
395
if (t->tx_buf)
drivers/spi/spi-au1550.c
396
dma_unmap_single(hw->dev, dma_tx_addr, t->len,
drivers/spi/spi-au1550.c
493
static int au1550_spi_pio_txrxb(struct spi_device *spi, struct spi_transfer *t)
drivers/spi/spi-au1550.c
498
hw->tx = t->tx_buf;
drivers/spi/spi-au1550.c
499
hw->rx = t->rx_buf;
drivers/spi/spi-au1550.c
500
hw->len = t->len;
drivers/spi/spi-au1550.c
54
int (*txrx_bufs)(struct spi_device *spi, struct spi_transfer *t);
drivers/spi/spi-au1550.c
626
static int au1550_spi_txrx_bufs(struct spi_device *spi, struct spi_transfer *t)
drivers/spi/spi-au1550.c
630
return hw->txrx_bufs(spi, t);
drivers/spi/spi-axi-spi-engine.c
261
unsigned int t;
drivers/spi/spi-axi-spi-engine.c
272
t = DIV_ROUND_UP_ULL((u64)(delay_ns - inst_ns) * sclk_hz, NSEC_PER_SEC);
drivers/spi/spi-axi-spi-engine.c
273
while (t) {
drivers/spi/spi-axi-spi-engine.c
274
unsigned int n = min(t, 256U);
drivers/spi/spi-axi-spi-engine.c
277
t -= n;
drivers/spi/spi-bcm-qspi.c
1155
struct spi_transfer t[2];
drivers/spi/spi-bcm-qspi.c
1160
memset(t, 0, sizeof(t));
drivers/spi/spi-bcm-qspi.c
1168
t[0].tx_buf = cmd;
drivers/spi/spi-bcm-qspi.c
1169
t[0].len = op->addr.nbytes + op->dummy.nbytes + 1;
drivers/spi/spi-bcm-qspi.c
1170
t[0].bits_per_word = spi->bits_per_word;
drivers/spi/spi-bcm-qspi.c
1171
t[0].tx_nbits = op->cmd.buswidth;
drivers/spi/spi-bcm-qspi.c
1174
ret = bcm_qspi_transfer_one(host, spi, &t[0]);
drivers/spi/spi-bcm-qspi.c
1180
t[1].rx_buf = op->data.buf.in;
drivers/spi/spi-bcm-qspi.c
1181
t[1].len = op->data.nbytes;
drivers/spi/spi-bcm-qspi.c
1182
t[1].rx_nbits = op->data.buswidth;
drivers/spi/spi-bcm-qspi.c
1183
t[1].bits_per_word = spi->bits_per_word;
drivers/spi/spi-bcm-qspi.c
1184
ret = bcm_qspi_transfer_one(host, spi, &t[1]);
drivers/spi/spi-bcm63xx-hsspi.c
273
struct spi_transfer *t;
drivers/spi/spi-bcm63xx-hsspi.c
286
list_for_each_entry(t, &msg->transfers, transfer_list) {
drivers/spi/spi-bcm63xx-hsspi.c
287
if ((spi_delay_to_ns(&t->delay, t) > 0) || t->cs_change) {
drivers/spi/spi-bcm63xx-hsspi.c
294
if (t->tx_buf && !t->rx_buf) {
drivers/spi/spi-bcm63xx-hsspi.c
296
if (bs->prepend_cnt + t->len >
drivers/spi/spi-bcm63xx-hsspi.c
303
if (t->tx_nbits == SPI_NBITS_SINGLE &&
drivers/spi/spi-bcm63xx-hsspi.c
304
!list_is_last(&t->transfer_list, &msg->transfers) &&
drivers/spi/spi-bcm63xx-hsspi.c
311
if (t->tx_nbits > SPI_NBITS_SINGLE)
drivers/spi/spi-bcm63xx-hsspi.c
314
memcpy(bs->prepend_buf + bs->prepend_cnt, t->tx_buf, t->len);
drivers/spi/spi-bcm63xx-hsspi.c
315
bs->prepend_cnt += t->len;
drivers/spi/spi-bcm63xx-hsspi.c
317
if (t->tx_nbits == SPI_NBITS_SINGLE)
drivers/spi/spi-bcm63xx-hsspi.c
318
bs->md_start += t->len;
drivers/spi/spi-bcm63xx-hsspi.c
321
if (!list_is_last(&t->transfer_list, &msg->transfers)) {
drivers/spi/spi-bcm63xx-hsspi.c
327
if (t->rx_buf && t->rx_nbits == SPI_NBITS_SINGLE &&
drivers/spi/spi-bcm63xx-hsspi.c
335
if (list_is_last(&t->transfer_list, &msg->transfers)) {
drivers/spi/spi-bcm63xx-hsspi.c
336
memcpy(t_prepend, t, sizeof(struct spi_transfer));
drivers/spi/spi-bcm63xx-hsspi.c
374
struct spi_transfer *t)
drivers/spi/spi-bcm63xx-hsspi.c
379
const u8 *tx = t->tx_buf;
drivers/spi/spi-bcm63xx-hsspi.c
380
u8 *rx = t->rx_buf;
drivers/spi/spi-bcm63xx-hsspi.c
387
if (t->len + bs->prepend_cnt > (HSSPI_BUFFER_LEN - HSSPI_OPCODE_LEN)) {
drivers/spi/spi-bcm63xx-hsspi.c
390
t->len, bs->prepend_cnt);
drivers/spi/spi-bcm63xx-hsspi.c
394
bcm63xx_hsspi_set_clk(bs, spi, t->speed_hz);
drivers/spi/spi-bcm63xx-hsspi.c
403
if ((opcode == HSSPI_OP_READ && t->rx_nbits == SPI_NBITS_DUAL) ||
drivers/spi/spi-bcm63xx-hsspi.c
404
(opcode == HSSPI_OP_WRITE && t->tx_nbits == SPI_NBITS_DUAL)) {
drivers/spi/spi-bcm63xx-hsspi.c
407
if (t->rx_nbits == SPI_NBITS_DUAL) {
drivers/spi/spi-bcm63xx-hsspi.c
411
if (t->tx_nbits == SPI_NBITS_DUAL) {
drivers/spi/spi-bcm63xx-hsspi.c
427
t->len);
drivers/spi/spi-bcm63xx-hsspi.c
429
*(__be16 *)(&val) = cpu_to_be16(opcode | t->len);
drivers/spi/spi-bcm63xx-hsspi.c
445
memcpy_fromio(rx, bs->fifo, t->len);
drivers/spi/spi-bcm63xx-hsspi.c
493
static int bcm63xx_hsspi_do_txrx(struct spi_device *spi, struct spi_transfer *t)
drivers/spi/spi-bcm63xx-hsspi.c
498
int pending = t->len;
drivers/spi/spi-bcm63xx-hsspi.c
500
const u8 *tx = t->tx_buf;
drivers/spi/spi-bcm63xx-hsspi.c
501
u8 *rx = t->rx_buf;
drivers/spi/spi-bcm63xx-hsspi.c
504
bcm63xx_hsspi_set_clk(bs, spi, t->speed_hz);
drivers/spi/spi-bcm63xx-hsspi.c
505
if (!t->cs_off)
drivers/spi/spi-bcm63xx-hsspi.c
518
if ((opcode == HSSPI_OP_READ && t->rx_nbits == SPI_NBITS_DUAL) ||
drivers/spi/spi-bcm63xx-hsspi.c
519
(opcode == HSSPI_OP_WRITE && t->tx_nbits == SPI_NBITS_DUAL)) {
drivers/spi/spi-bcm63xx-hsspi.c
522
if (t->rx_nbits == SPI_NBITS_DUAL)
drivers/spi/spi-bcm63xx-hsspi.c
524
if (t->tx_nbits == SPI_NBITS_DUAL)
drivers/spi/spi-bcm63xx-hsspi.c
611
struct spi_transfer *t;
drivers/spi/spi-bcm63xx-hsspi.c
631
list_for_each_entry(t, &msg->transfers, transfer_list) {
drivers/spi/spi-bcm63xx-hsspi.c
639
if (t->speed_hz > HSSPI_MAX_SYNC_CLOCK) {
drivers/spi/spi-bcm63xx-hsspi.c
640
t->speed_hz = HSSPI_MAX_SYNC_CLOCK;
drivers/spi/spi-bcm63xx-hsspi.c
643
t->speed_hz);
drivers/spi/spi-bcm63xx-hsspi.c
647
status = bcm63xx_hsspi_do_txrx(spi, t);
drivers/spi/spi-bcm63xx-hsspi.c
651
msg->actual_length += t->len;
drivers/spi/spi-bcm63xx-hsspi.c
653
spi_transfer_delay_exec(t);
drivers/spi/spi-bcm63xx-hsspi.c
656
if (t->cs_change) {
drivers/spi/spi-bcm63xx-hsspi.c
657
if (list_is_last(&t->transfer_list, &msg->transfers)) {
drivers/spi/spi-bcm63xx-hsspi.c
660
if (!t->cs_off)
drivers/spi/spi-bcm63xx-hsspi.c
663
spi_transfer_cs_change_delay_exec(msg, t);
drivers/spi/spi-bcm63xx-hsspi.c
665
if (!list_next_entry(t, transfer_list)->cs_off)
drivers/spi/spi-bcm63xx-hsspi.c
668
} else if (!list_is_last(&t->transfer_list, &msg->transfers) &&
drivers/spi/spi-bcm63xx-hsspi.c
669
t->cs_off != list_next_entry(t, transfer_list)->cs_off) {
drivers/spi/spi-bcm63xx-hsspi.c
670
bcm63xx_hsspi_set_cs(bs, spi_get_chipselect(spi, 0), t->cs_off);
drivers/spi/spi-bcm63xx.c
187
struct spi_transfer *t)
drivers/spi/spi-bcm63xx.c
198
if (t->speed_hz >= bcm63xx_spi_freq_table[i][0]) {
drivers/spi/spi-bcm63xx.c
211
clk_cfg, t->speed_hz);
drivers/spi/spi-bcm63xx.c
224
struct spi_transfer *t = first;
drivers/spi/spi-bcm63xx.c
232
t->tx_buf, t->rx_buf, t->len);
drivers/spi/spi-bcm63xx.c
234
if (num_transfers > 1 && t->tx_buf && t->len <= BCM63XX_SPI_MAX_PREPEND)
drivers/spi/spi-bcm63xx.c
235
prepend_len = t->len;
drivers/spi/spi-bcm63xx.c
239
if (t->tx_buf) {
drivers/spi/spi-bcm63xx.c
241
memcpy_toio(bs->tx_io + len, t->tx_buf, t->len);
drivers/spi/spi-bcm63xx.c
244
if (t != first)
drivers/spi/spi-bcm63xx.c
248
if (t->rx_buf) {
drivers/spi/spi-bcm63xx.c
261
if (!t->tx_buf)
drivers/spi/spi-bcm63xx.c
262
memset_io(bs->tx_io + len, 0xFF, t->len);
drivers/spi/spi-bcm63xx.c
265
if (t == first)
drivers/spi/spi-bcm63xx.c
269
len += t->len;
drivers/spi/spi-bcm63xx.c
271
t = list_entry(t->transfer_list.next, struct spi_transfer,
drivers/spi/spi-bcm63xx.c
313
t = first;
drivers/spi/spi-bcm63xx.c
316
if (t->rx_buf)
drivers/spi/spi-bcm63xx.c
317
memcpy_fromio(t->rx_buf, bs->rx_io + len, t->len);
drivers/spi/spi-bcm63xx.c
319
if (t != first || prepend_len == 0)
drivers/spi/spi-bcm63xx.c
320
len += t->len;
drivers/spi/spi-bcm63xx.c
322
t = list_entry(t->transfer_list.next, struct spi_transfer,
drivers/spi/spi-bcm63xx.c
333
struct spi_transfer *t, *first = NULL;
drivers/spi/spi-bcm63xx.c
345
list_for_each_entry(t, &m->transfers, transfer_list) {
drivers/spi/spi-bcm63xx.c
347
first = t;
drivers/spi/spi-bcm63xx.c
350
total_len += t->len;
drivers/spi/spi-bcm63xx.c
352
if (n_transfers == 2 && !first->rx_buf && !t->tx_buf &&
drivers/spi/spi-bcm63xx.c
355
else if (can_use_prepend && t->tx_buf)
drivers/spi/spi-bcm63xx.c
369
if (t->speed_hz != first->speed_hz) {
drivers/spi/spi-bcm63xx.c
376
if (t->delay.value) {
drivers/spi/spi-bcm63xx.c
382
if (t->cs_change ||
drivers/spi/spi-bcm63xx.c
383
list_is_last(&t->transfer_list, &m->transfers)) {
drivers/spi/spi-bcmbca-hsspi.c
250
static int bcmbca_hsspi_do_txrx(struct spi_device *spi, struct spi_transfer *t,
drivers/spi/spi-bcmbca-hsspi.c
256
int pending = t->len;
drivers/spi/spi-bcmbca-hsspi.c
258
const u8 *tx = t->tx_buf;
drivers/spi/spi-bcmbca-hsspi.c
259
u8 *rx = t->rx_buf;
drivers/spi/spi-bcmbca-hsspi.c
262
bcmbca_hsspi_set_clk(bs, spi, t->speed_hz);
drivers/spi/spi-bcmbca-hsspi.c
274
if ((opcode == HSSPI_OP_READ && t->rx_nbits == SPI_NBITS_DUAL) ||
drivers/spi/spi-bcmbca-hsspi.c
275
(opcode == HSSPI_OP_WRITE && t->tx_nbits == SPI_NBITS_DUAL)) {
drivers/spi/spi-bcmbca-hsspi.c
278
if (t->rx_nbits == SPI_NBITS_DUAL)
drivers/spi/spi-bcmbca-hsspi.c
280
if (t->tx_nbits == SPI_NBITS_DUAL)
drivers/spi/spi-bcmbca-hsspi.c
373
struct spi_transfer *t;
drivers/spi/spi-bcmbca-hsspi.c
379
list_for_each_entry(t, &msg->transfers, transfer_list) {
drivers/spi/spi-bcmbca-hsspi.c
380
status = bcmbca_hsspi_do_txrx(spi, t, msg);
drivers/spi/spi-bcmbca-hsspi.c
384
spi_transfer_delay_exec(t);
drivers/spi/spi-bcmbca-hsspi.c
386
if (t->cs_change) {
drivers/spi/spi-bcmbca-hsspi.c
387
if (list_is_last(&t->transfer_list, &msg->transfers)) {
drivers/spi/spi-bcmbca-hsspi.c
390
if (!t->cs_off)
drivers/spi/spi-bcmbca-hsspi.c
393
spi_transfer_cs_change_delay_exec(msg, t);
drivers/spi/spi-bcmbca-hsspi.c
395
if (!list_next_entry(t, transfer_list)->cs_off)
drivers/spi/spi-bcmbca-hsspi.c
398
} else if (!list_is_last(&t->transfer_list, &msg->transfers) &&
drivers/spi/spi-bcmbca-hsspi.c
399
t->cs_off != list_next_entry(t, transfer_list)->cs_off) {
drivers/spi/spi-bcmbca-hsspi.c
400
bcmbca_hsspi_set_cs(bs, spi_get_chipselect(spi, 0), t->cs_off);
drivers/spi/spi-bcmbca-hsspi.c
403
msg->actual_length += t->len;
drivers/spi/spi-bitbang.c
110
return t->len - count;
drivers/spi/spi-bitbang.c
116
struct spi_transfer *t,
drivers/spi/spi-bitbang.c
120
unsigned int bits = t->bits_per_word;
drivers/spi/spi-bitbang.c
121
unsigned int count = t->len;
drivers/spi/spi-bitbang.c
122
const u32 *tx = t->tx_buf;
drivers/spi/spi-bitbang.c
123
u32 *rx = t->rx_buf;
drivers/spi/spi-bitbang.c
141
return t->len - count;
drivers/spi/spi-bitbang.c
144
int spi_bitbang_setup_transfer(struct spi_device *spi, struct spi_transfer *t)
drivers/spi/spi-bitbang.c
150
if (t) {
drivers/spi/spi-bitbang.c
151
bits_per_word = t->bits_per_word;
drivers/spi/spi-bitbang.c
152
hz = t->speed_hz;
drivers/spi/spi-bitbang.c
239
static int spi_bitbang_bufs(struct spi_device *spi, struct spi_transfer *t)
drivers/spi/spi-bitbang.c
249
err = bitbang->set_line_direction(spi, !!(t->tx_buf));
drivers/spi/spi-bitbang.c
257
flags = t->tx_buf ? SPI_CONTROLLER_NO_RX : SPI_CONTROLLER_NO_TX;
drivers/spi/spi-bitbang.c
258
return cs->txrx_bufs(spi, cs->txrx_word, nsecs, t, flags);
drivers/spi/spi-bitbang.c
260
return cs->txrx_bufs(spi, cs->txrx_word, nsecs, t, 0);
drivers/spi/spi-bitbang.c
54
struct spi_transfer *t,
drivers/spi/spi-bitbang.c
58
unsigned int bits = t->bits_per_word;
drivers/spi/spi-bitbang.c
59
unsigned int count = t->len;
drivers/spi/spi-bitbang.c
60
const u8 *tx = t->tx_buf;
drivers/spi/spi-bitbang.c
61
u8 *rx = t->rx_buf;
drivers/spi/spi-bitbang.c
79
return t->len - count;
drivers/spi/spi-bitbang.c
85
struct spi_transfer *t,
drivers/spi/spi-bitbang.c
89
unsigned int bits = t->bits_per_word;
drivers/spi/spi-bitbang.c
90
unsigned int count = t->len;
drivers/spi/spi-bitbang.c
91
const u16 *tx = t->tx_buf;
drivers/spi/spi-bitbang.c
92
u16 *rx = t->rx_buf;
drivers/spi/spi-cadence-xspi.c
1035
struct spi_transfer *t = NULL;
drivers/spi/spi-cadence-xspi.c
1057
list_for_each_entry(t, &m->transfers, transfer_list) {
drivers/spi/spi-cadence-xspi.c
1058
u8 *txd = (u8 *) t->tx_buf;
drivers/spi/spi-cadence-xspi.c
1059
u8 *rxd = (u8 *) t->rx_buf;
drivers/spi/spi-cadence-xspi.c
1069
while (t->len) {
drivers/spi/spi-cadence-xspi.c
1071
current_transfer_len = min(max_len, t->len);
drivers/spi/spi-cadence-xspi.c
1109
cs_change = t->cs_change;
drivers/spi/spi-cadence-xspi.c
1110
t->len -= current_transfer_len;
drivers/spi/spi-cadence-xspi.c
1112
spi_transfer_delay_exec(t);
drivers/spi/spi-coldfire-qspi.c
300
struct spi_transfer *t)
drivers/spi/spi-coldfire-qspi.c
305
qmr |= t->bits_per_word << 10;
drivers/spi/spi-coldfire-qspi.c
310
qmr |= mcfqspi_qmr_baud(t->speed_hz);
drivers/spi/spi-coldfire-qspi.c
314
if (t->bits_per_word == 8)
drivers/spi/spi-coldfire-qspi.c
315
mcfqspi_transfer_msg8(mcfqspi, t->len, t->tx_buf, t->rx_buf);
drivers/spi/spi-coldfire-qspi.c
317
mcfqspi_transfer_msg16(mcfqspi, t->len / 2, t->tx_buf,
drivers/spi/spi-coldfire-qspi.c
318
t->rx_buf);
drivers/spi/spi-davinci.c
325
struct spi_transfer *t)
drivers/spi/spi-davinci.c
339
if (t) {
drivers/spi/spi-davinci.c
340
bits_per_word = t->bits_per_word;
drivers/spi/spi-davinci.c
341
hz = t->speed_hz;
drivers/spi/spi-davinci.c
625
static int davinci_spi_bufs(struct spi_device *spi, struct spi_transfer *t)
drivers/spi/spi-davinci.c
644
dspi->tx = t->tx_buf;
drivers/spi/spi-davinci.c
645
dspi->rx = t->rx_buf;
drivers/spi/spi-davinci.c
646
dspi->wcount = t->len / data_type;
drivers/spi/spi-davinci.c
656
if (!davinci_spi_can_dma(spi->controller, spi, t)) {
drivers/spi/spi-davinci.c
685
t->rx_sg.sgl, t->rx_sg.nents, DMA_DEV_TO_MEM,
drivers/spi/spi-davinci.c
690
if (!t->tx_buf) {
drivers/spi/spi-davinci.c
696
t->tx_sg.sgl = t->rx_sg.sgl;
drivers/spi/spi-davinci.c
697
t->tx_sg.nents = t->rx_sg.nents;
drivers/spi/spi-davinci.c
701
t->tx_sg.sgl, t->tx_sg.nents, DMA_MEM_TO_DEV,
drivers/spi/spi-davinci.c
725
timeout = DIV_ROUND_UP(t->speed_hz, MSEC_PER_SEC);
drivers/spi/spi-davinci.c
726
timeout = DIV_ROUND_UP(t->len * 8, timeout);
drivers/spi/spi-davinci.c
742
if (davinci_spi_can_dma(spi->controller, spi, t))
drivers/spi/spi-davinci.c
764
return t->len;
drivers/spi/spi-falcon.c
103
const u8 *txp = t->tx_buf;
drivers/spi/spi-falcon.c
104
u8 *rxp = t->rx_buf;
drivers/spi/spi-falcon.c
105
unsigned int bytelen = ((8 * t->len + 7) / 8);
drivers/spi/spi-falcon.c
359
struct spi_transfer *t;
drivers/spi/spi-falcon.c
368
list_for_each_entry(t, &m->transfers, transfer_list) {
drivers/spi/spi-falcon.c
369
if (list_is_last(&t->transfer_list, &m->transfers))
drivers/spi/spi-falcon.c
373
ret = falcon_sflash_xfer(m->spi, t, spi_flags);
drivers/spi/spi-falcon.c
379
m->actual_length += t->len;
drivers/spi/spi-falcon.c
381
WARN_ON(t->delay.value || t->cs_change);
drivers/spi/spi-falcon.c
98
falcon_sflash_xfer(struct spi_device *spi, struct spi_transfer *t,
drivers/spi/spi-fsl-cpm.c
101
int fsl_spi_cpm_bufs(struct mpc8xxx_spi *mspi, struct spi_transfer *t)
drivers/spi/spi-fsl-cpm.c
109
if (!t->tx_buf) {
drivers/spi/spi-fsl-cpm.c
114
if (!t->rx_buf) {
drivers/spi/spi-fsl-cpm.c
118
if (t->bits_per_word == 16 && t->tx_buf) {
drivers/spi/spi-fsl-cpm.c
119
const u16 *src = t->tx_buf;
drivers/spi/spi-fsl-cpm.c
123
dst = kmalloc(t->len, GFP_KERNEL);
drivers/spi/spi-fsl-cpm.c
127
for (i = 0; i < t->len >> 1; i++)
drivers/spi/spi-fsl-cpm.c
137
mspi->tx_dma = dma_map_single(dev, nonconst_tx, t->len,
drivers/spi/spi-fsl-cpm.c
143
} else if (t->tx_buf) {
drivers/spi/spi-fsl-cpm.c
148
mspi->rx_dma = dma_map_single(dev, mspi->rx, t->len,
drivers/spi/spi-fsl-cpm.c
154
} else if (t->rx_buf) {
drivers/spi/spi-fsl-cpm.c
155
mspi->rx_dma = t->rx_dma;
drivers/spi/spi-fsl-cpm.c
161
mspi->xfer_in_progress = t;
drivers/spi/spi-fsl-cpm.c
162
mspi->count = t->len;
drivers/spi/spi-fsl-cpm.c
171
dma_unmap_single(dev, mspi->tx_dma, t->len, DMA_TO_DEVICE);
drivers/spi/spi-fsl-cpm.c
179
struct spi_transfer *t = mspi->xfer_in_progress;
drivers/spi/spi-fsl-cpm.c
182
dma_unmap_single(dev, mspi->tx_dma, t->len, DMA_TO_DEVICE);
drivers/spi/spi-fsl-cpm.c
184
dma_unmap_single(dev, mspi->rx_dma, t->len, DMA_FROM_DEVICE);
drivers/spi/spi-fsl-cpm.c
187
if (t->bits_per_word == 16 && t->rx_buf) {
drivers/spi/spi-fsl-cpm.c
190
for (i = 0; i < t->len; i += 2)
drivers/spi/spi-fsl-cpm.c
191
le16_to_cpus(t->rx_buf + i);
drivers/spi/spi-fsl-cpm.h
23
struct spi_transfer *t);
drivers/spi/spi-fsl-cpm.h
31
struct spi_transfer *t) { return 0; }
drivers/spi/spi-fsl-espi.c
152
struct spi_transfer *t, *first;
drivers/spi/spi-fsl-espi.c
163
list_for_each_entry(t, &m->transfers, transfer_list) {
drivers/spi/spi-fsl-espi.c
164
if (first->bits_per_word != t->bits_per_word ||
drivers/spi/spi-fsl-espi.c
165
first->speed_hz != t->speed_hz) {
drivers/spi/spi-fsl-espi.c
185
struct spi_transfer *t;
drivers/spi/spi-fsl-espi.c
198
list_for_each_entry(t, &m->transfers, transfer_list) {
drivers/spi/spi-fsl-espi.c
200
if (!t->tx_buf || t->rx_buf ||
drivers/spi/spi-fsl-espi.c
201
t->len > FSL_ESPI_FIFO_SIZE)
drivers/spi/spi-fsl-espi.c
203
rxskip = t->len;
drivers/spi/spi-fsl-espi.c
205
if (t->tx_buf || !t->rx_buf)
drivers/spi/spi-fsl-espi.c
324
struct spi_transfer *t)
drivers/spi/spi-fsl-espi.c
327
int bits_per_word = t ? t->bits_per_word : spi->bits_per_word;
drivers/spi/spi-fsl-espi.c
328
u32 pm, hz = t ? t->speed_hz : spi->max_speed_hz;
drivers/spi/spi-fsl-espi.c
352
static int fsl_espi_bufs(struct spi_device *spi, struct spi_transfer *t)
drivers/spi/spi-fsl-espi.c
355
unsigned int rx_len = t->len;
drivers/spi/spi-fsl-espi.c
363
spcom |= SPCOM_TRANLEN(t->len - 1);
drivers/spi/spi-fsl-espi.c
368
rx_len = t->len - espi->rxskip;
drivers/spi/spi-fsl-espi.c
369
if (t->rx_nbits == SPI_NBITS_DUAL)
drivers/spi/spi-fsl-espi.c
439
struct spi_transfer *t, trans = {};
drivers/spi/spi-fsl-espi.c
446
list_for_each_entry(t, &m->transfers, transfer_list) {
drivers/spi/spi-fsl-espi.c
447
unsigned int delay = spi_delay_to_ns(&t->delay, t);
drivers/spi/spi-fsl-espi.c
451
if (t->rx_nbits > rx_nbits)
drivers/spi/spi-fsl-espi.c
452
rx_nbits = t->rx_nbits;
drivers/spi/spi-fsl-espi.c
455
t = list_first_entry(&m->transfers, struct spi_transfer,
drivers/spi/spi-fsl-espi.c
459
trans.speed_hz = t->speed_hz;
drivers/spi/spi-fsl-espi.c
460
trans.bits_per_word = t->bits_per_word;
drivers/spi/spi-fsl-lpspi.c
464
struct spi_transfer *t)
drivers/spi/spi-fsl-lpspi.c
469
if (t == NULL)
drivers/spi/spi-fsl-lpspi.c
473
fsl_lpspi->config.bpw = t->bits_per_word;
drivers/spi/spi-fsl-lpspi.c
474
fsl_lpspi->config.speed_hz = t->speed_hz;
drivers/spi/spi-fsl-lpspi.c
497
fsl_lpspi->watermark = min(fsl_lpspi->txfifosize, t->len);
drivers/spi/spi-fsl-lpspi.c
508
struct spi_transfer *t;
drivers/spi/spi-fsl-lpspi.c
511
t = list_first_entry_or_null(&msg->transfers, struct spi_transfer,
drivers/spi/spi-fsl-lpspi.c
513
if (!t)
drivers/spi/spi-fsl-lpspi.c
518
ret = fsl_lpspi_setup_transfer(controller, spi, t);
drivers/spi/spi-fsl-lpspi.c
520
if (fsl_lpspi_can_dma(controller, spi, t))
drivers/spi/spi-fsl-lpspi.c
767
struct spi_transfer *t)
drivers/spi/spi-fsl-lpspi.c
773
fsl_lpspi->tx_buf = t->tx_buf;
drivers/spi/spi-fsl-lpspi.c
774
fsl_lpspi->rx_buf = t->rx_buf;
drivers/spi/spi-fsl-lpspi.c
775
fsl_lpspi->remain = t->len;
drivers/spi/spi-fsl-lpspi.c
791
struct spi_transfer *t)
drivers/spi/spi-fsl-lpspi.c
797
if (fsl_lpspi_can_dma(controller, spi, t))
drivers/spi/spi-fsl-lpspi.c
802
ret = fsl_lpspi_setup_transfer(controller, spi, t);
drivers/spi/spi-fsl-lpspi.c
806
t->effective_speed_hz = fsl_lpspi->config.effective_speed_hz;
drivers/spi/spi-fsl-lpspi.c
812
ret = fsl_lpspi_dma_transfer(controller, fsl_lpspi, t);
drivers/spi/spi-fsl-lpspi.c
814
ret = fsl_lpspi_pio_transfer(controller, t);
drivers/spi/spi-fsl-spi.c
178
struct spi_transfer *t)
drivers/spi/spi-fsl-spi.c
188
if (t) {
drivers/spi/spi-fsl-spi.c
189
bits_per_word = t->bits_per_word;
drivers/spi/spi-fsl-spi.c
190
hz = t->speed_hz;
drivers/spi/spi-fsl-spi.c
235
struct spi_transfer *t, unsigned int len)
drivers/spi/spi-fsl-spi.c
252
static int fsl_spi_bufs(struct spi_device *spi, struct spi_transfer *t)
drivers/spi/spi-fsl-spi.c
256
unsigned int len = t->len;
drivers/spi/spi-fsl-spi.c
262
if (t->bits_per_word)
drivers/spi/spi-fsl-spi.c
263
bits_per_word = t->bits_per_word;
drivers/spi/spi-fsl-spi.c
270
mpc8xxx_spi->tx = t->tx_buf;
drivers/spi/spi-fsl-spi.c
271
mpc8xxx_spi->rx = t->rx_buf;
drivers/spi/spi-fsl-spi.c
276
ret = fsl_spi_cpm_bufs(mpc8xxx_spi, t);
drivers/spi/spi-fsl-spi.c
278
ret = fsl_spi_cpu_bufs(mpc8xxx_spi, t, len);
drivers/spi/spi-fsl-spi.c
297
struct spi_transfer *t;
drivers/spi/spi-fsl-spi.c
311
list_for_each_entry(t, &m->transfers, transfer_list) {
drivers/spi/spi-fsl-spi.c
312
if (t->speed_hz != first->speed_hz) {
drivers/spi/spi-fsl-spi.c
318
if (t->len < 256 || t->bits_per_word != 8)
drivers/spi/spi-fsl-spi.c
320
if ((t->len & 3) == 0)
drivers/spi/spi-fsl-spi.c
321
t->bits_per_word = 32;
drivers/spi/spi-fsl-spi.c
322
else if ((t->len & 1) == 0)
drivers/spi/spi-fsl-spi.c
323
t->bits_per_word = 16;
drivers/spi/spi-fsl-spi.c
333
if (m->spi->mode & SPI_LSB_FIRST && t->bits_per_word > 8)
drivers/spi/spi-fsl-spi.c
335
if (t->bits_per_word == 16 || t->bits_per_word == 32)
drivers/spi/spi-fsl-spi.c
336
t->bits_per_word = 8; /* pretend its 8 bits */
drivers/spi/spi-fsl-spi.c
337
if (t->bits_per_word == 8 && t->len >= 256 &&
drivers/spi/spi-fsl-spi.c
338
!(t->len & 1) && (mpc8xxx_spi->flags & SPI_CPM1))
drivers/spi/spi-fsl-spi.c
339
t->bits_per_word = 16;
drivers/spi/spi-fsl-spi.c
347
struct spi_transfer *t)
drivers/spi/spi-fsl-spi.c
351
status = fsl_spi_setup_transfer(spi, t);
drivers/spi/spi-fsl-spi.c
354
if (t->len)
drivers/spi/spi-fsl-spi.c
355
status = fsl_spi_bufs(spi, t);
drivers/spi/spi-imx.c
1073
struct spi_device *spi, struct spi_transfer *t)
drivers/spi/spi-imx.c
1317
struct spi_transfer *t)
drivers/spi/spi-imx.c
1321
if (!t)
drivers/spi/spi-imx.c
1325
if (!t->speed_hz) {
drivers/spi/spi-imx.c
1333
spi_imx->spi_bus_clk = t->speed_hz;
drivers/spi/spi-imx.c
1337
spi_imx->bits_per_word = t->bits_per_word;
drivers/spi/spi-imx.c
1338
spi_imx->count = t->len;
drivers/spi/spi-imx.c
1348
!(t->word_delay.value) &&
drivers/spi/spi-imx.c
1371
if (spi_imx_can_dma(spi_imx->controller, spi, t))
drivers/spi/spi-imx.c
1376
spi_imx->rx_only = ((t->tx_buf == NULL)
drivers/spi/spi-imx.c
1377
|| (t->tx_buf == spi->controller->dummy_tx));
drivers/spi/spi-imx.c
1382
spi_imx->target_burst = t->len;
drivers/spi/spi-imx.c
1385
spi_imx->devtype_data->prepare_transfer(spi_imx, spi, t);
drivers/spi/spi-imx.c
723
struct spi_device *spi, struct spi_transfer *t)
drivers/spi/spi-imx.c
762
if (t->word_delay.value == 0) {
drivers/spi/spi-imx.c
764
} else if (t->word_delay.unit == SPI_DELAY_UNIT_SCK) {
drivers/spi/spi-imx.c
765
word_delay_sck = t->word_delay.value;
drivers/spi/spi-imx.c
776
word_delay_ns = spi_delay_to_ns(&t->word_delay, t);
drivers/spi/spi-imx.c
81
struct spi_transfer *t);
drivers/spi/spi-imx.c
894
struct spi_device *spi, struct spi_transfer *t)
drivers/spi/spi-imx.c
998
struct spi_device *spi, struct spi_transfer *t)
drivers/spi/spi-jcore.c
112
jcore_spi_baudrate(hw, t->speed_hz);
drivers/spi/spi-jcore.c
115
tx = t->tx_buf;
drivers/spi/spi-jcore.c
116
rx = t->rx_buf;
drivers/spi/spi-jcore.c
117
len = t->len;
drivers/spi/spi-jcore.c
98
struct spi_transfer *t)
drivers/spi/spi-kspi2.c
164
static int kspi2_process_transfer(struct kspi2 *kspi, struct spi_transfer *t)
drivers/spi/spi-kspi2.c
171
for (i = 0; i < t->len; i++) {
drivers/spi/spi-kspi2.c
172
if (t->tx_buf)
drivers/spi/spi-kspi2.c
173
tx = ((const u8 *)t->tx_buf)[i];
drivers/spi/spi-kspi2.c
179
if (t->rx_buf)
drivers/spi/spi-kspi2.c
180
((u8 *)t->rx_buf)[i] = rx;
drivers/spi/spi-kspi2.c
188
struct spi_transfer *t)
drivers/spi/spi-kspi2.c
197
if (t->bits_per_word && ((t->bits_per_word % 8) != 0)) {
drivers/spi/spi-kspi2.c
199
t->bits_per_word);
drivers/spi/spi-kspi2.c
204
if (t->speed_hz && (t->speed_hz < max_speed_hz))
drivers/spi/spi-kspi2.c
205
max_speed_hz = t->speed_hz;
drivers/spi/spi-kspi2.c
215
struct spi_transfer *t)
drivers/spi/spi-kspi2.c
220
ret = kspi2_setup_transfer(kspi, spi, t);
drivers/spi/spi-kspi2.c
224
if (t->len) {
drivers/spi/spi-kspi2.c
225
ret = kspi2_process_transfer(kspi, t);
drivers/spi/spi-lantiq-ssc.c
432
struct spi_device *spidev, struct spi_transfer *t)
drivers/spi/spi-lantiq-ssc.c
434
unsigned int speed_hz = t->speed_hz;
drivers/spi/spi-lantiq-ssc.c
435
unsigned int bits_per_word = t->bits_per_word;
drivers/spi/spi-lantiq-ssc.c
451
if (t->tx_buf)
drivers/spi/spi-lantiq-ssc.c
456
if (t->rx_buf)
drivers/spi/spi-lantiq-ssc.c
723
struct spi_transfer *t)
drivers/spi/spi-lantiq-ssc.c
729
spi->tx = t->tx_buf;
drivers/spi/spi-lantiq-ssc.c
730
spi->rx = t->rx_buf;
drivers/spi/spi-lantiq-ssc.c
732
if (t->tx_buf) {
drivers/spi/spi-lantiq-ssc.c
733
spi->tx_todo = t->len;
drivers/spi/spi-lantiq-ssc.c
740
spi->rx_todo = t->len;
drivers/spi/spi-lantiq-ssc.c
749
return t->len;
drivers/spi/spi-lantiq-ssc.c
813
struct spi_transfer *t)
drivers/spi/spi-lantiq-ssc.c
817
hw_setup_transfer(spi, spidev, t);
drivers/spi/spi-lantiq-ssc.c
819
return transfer_start(spi, spidev, t);
drivers/spi/spi-loongson-core.c
79
struct spi_device *spi, struct spi_transfer *t)
drivers/spi/spi-loongson-core.c
81
if (t && loongson_spi->hz != t->speed_hz)
drivers/spi/spi-loongson-core.c
82
loongson_spi_set_clk(loongson_spi, t->speed_hz);
drivers/spi/spi-lp8841-rtc.c
100
struct spi_transfer *t)
drivers/spi/spi-lp8841-rtc.c
103
unsigned count = t->len;
drivers/spi/spi-lp8841-rtc.c
104
const u8 *tx = t->tx_buf;
drivers/spi/spi-lp8841-rtc.c
105
u8 *rx = t->rx_buf;
drivers/spi/spi-meson-spicc.c
234
struct spi_transfer *t)
drivers/spi/spi-meson-spicc.c
238
if (!(t->tx_buf && t->rx_buf))
drivers/spi/spi-meson-spicc.c
241
t->tx_dma = dma_map_single(dev, (void *)t->tx_buf, t->len, DMA_TO_DEVICE);
drivers/spi/spi-meson-spicc.c
242
if (dma_mapping_error(dev, t->tx_dma))
drivers/spi/spi-meson-spicc.c
245
t->rx_dma = dma_map_single(dev, t->rx_buf, t->len, DMA_FROM_DEVICE);
drivers/spi/spi-meson-spicc.c
246
if (dma_mapping_error(dev, t->rx_dma))
drivers/spi/spi-meson-spicc.c
249
spicc->tx_dma = t->tx_dma;
drivers/spi/spi-meson-spicc.c
250
spicc->rx_dma = t->rx_dma;
drivers/spi/spi-meson-spicc.c
256
struct spi_transfer *t)
drivers/spi/spi-meson-spicc.c
260
if (t->tx_dma)
drivers/spi/spi-meson-spicc.c
261
dma_unmap_single(dev, t->tx_dma, t->len, DMA_TO_DEVICE);
drivers/spi/spi-meson-spicc.c
262
if (t->rx_dma)
drivers/spi/spi-meson-spicc.c
263
dma_unmap_single(dev, t->rx_dma, t->len, DMA_FROM_DEVICE);
drivers/spi/spi-microchip-core-qspi.c
597
struct spi_transfer *t = NULL;
drivers/spi/spi-microchip-core-qspi.c
623
list_for_each_entry(t, &m->transfers, transfer_list) {
drivers/spi/spi-microchip-core-qspi.c
624
total_bytes += t->len;
drivers/spi/spi-microchip-core-qspi.c
625
if (!cmd_bytes && !(t->tx_buf && t->rx_buf))
drivers/spi/spi-microchip-core-qspi.c
626
cmd_bytes = t->len;
drivers/spi/spi-microchip-core-qspi.c
627
if (!t->rx_buf)
drivers/spi/spi-microchip-core-qspi.c
629
if (t->tx_nbits == SPI_NBITS_QUAD || t->rx_nbits == SPI_NBITS_QUAD)
drivers/spi/spi-microchip-core-qspi.c
631
else if (t->tx_nbits == SPI_NBITS_DUAL || t->rx_nbits == SPI_NBITS_DUAL)
drivers/spi/spi-microchip-core-qspi.c
662
struct spi_transfer *t)
drivers/spi/spi-microchip-core-qspi.c
666
qspi->tx_len = t->len;
drivers/spi/spi-microchip-core-qspi.c
668
if (t->tx_buf)
drivers/spi/spi-microchip-core-qspi.c
669
qspi->txbuf = (u8 *)t->tx_buf;
drivers/spi/spi-microchip-core-qspi.c
671
if (!t->rx_buf) {
drivers/spi/spi-microchip-core-qspi.c
674
qspi->rxbuf = (u8 *)t->rx_buf;
drivers/spi/spi-microchip-core-qspi.c
675
qspi->rx_len = t->len;
drivers/spi/spi-mpc512x-psc.c
144
struct spi_transfer *t)
drivers/spi/spi-mpc512x-psc.c
148
size_t tx_len = t->len;
drivers/spi/spi-mpc512x-psc.c
149
size_t rx_len = t->len;
drivers/spi/spi-mpc512x-psc.c
150
u8 *tx_buf = (u8 *)t->tx_buf;
drivers/spi/spi-mpc512x-psc.c
151
u8 *rx_buf = (u8 *)t->rx_buf;
drivers/spi/spi-mpc512x-psc.c
153
if (!tx_buf && !rx_buf && t->len)
drivers/spi/spi-mpc512x-psc.c
177
if (tx_len == EOFBYTE && t->cs_change)
drivers/spi/spi-mpc512x-psc.c
289
struct spi_transfer *t;
drivers/spi/spi-mpc512x-psc.c
294
list_for_each_entry(t, &m->transfers, transfer_list) {
drivers/spi/spi-mpc512x-psc.c
295
status = mpc512x_psc_spi_transfer_setup(spi, t);
drivers/spi/spi-mpc512x-psc.c
301
cs_change = t->cs_change;
drivers/spi/spi-mpc512x-psc.c
303
status = mpc512x_psc_spi_transfer_rxtx(spi, t);
drivers/spi/spi-mpc512x-psc.c
306
m->actual_length += t->len;
drivers/spi/spi-mpc512x-psc.c
308
spi_transfer_delay_exec(t);
drivers/spi/spi-mpc512x-psc.c
73
struct spi_transfer *t)
drivers/spi/spi-mpc512x-psc.c
77
cs->speed_hz = (t && t->speed_hz)
drivers/spi/spi-mpc512x-psc.c
78
? t->speed_hz : spi->max_speed_hz;
drivers/spi/spi-mpc512x-psc.c
79
cs->bits_per_word = (t && t->bits_per_word)
drivers/spi/spi-mpc512x-psc.c
80
? t->bits_per_word : spi->bits_per_word;
drivers/spi/spi-mpc52xx-psc.c
105
struct spi_transfer *t)
drivers/spi/spi-mpc52xx-psc.c
112
unsigned char *rx_buf = (unsigned char *)t->rx_buf;
drivers/spi/spi-mpc52xx-psc.c
113
unsigned char *tx_buf = (unsigned char *)t->tx_buf;
drivers/spi/spi-mpc52xx-psc.c
119
if (!t->tx_buf && !t->rx_buf && t->len)
drivers/spi/spi-mpc52xx-psc.c
124
while (rb < t->len) {
drivers/spi/spi-mpc52xx-psc.c
125
if (t->len - rb > MPC52xx_PSC_BUFSIZE) {
drivers/spi/spi-mpc52xx-psc.c
129
send_at_once = t->len - sb;
drivers/spi/spi-mpc52xx-psc.c
130
rfalarm = MPC52xx_PSC_BUFSIZE - (t->len - rb);
drivers/spi/spi-mpc52xx-psc.c
152
if (t->len - rb == 1) {
drivers/spi/spi-mpc52xx-psc.c
182
struct spi_transfer *t = NULL;
drivers/spi/spi-mpc52xx-psc.c
189
list_for_each_entry (t, &m->transfers, transfer_list) {
drivers/spi/spi-mpc52xx-psc.c
190
if (t->bits_per_word || t->speed_hz) {
drivers/spi/spi-mpc52xx-psc.c
191
status = mpc52xx_psc_spi_transfer_setup(spi, t);
drivers/spi/spi-mpc52xx-psc.c
198
cs_change = t->cs_change;
drivers/spi/spi-mpc52xx-psc.c
200
status = mpc52xx_psc_spi_transfer_rxtx(spi, t);
drivers/spi/spi-mpc52xx-psc.c
203
m->actual_length += t->len;
drivers/spi/spi-mpc52xx-psc.c
205
spi_transfer_delay_exec(t);
drivers/spi/spi-mpc52xx-psc.c
48
struct spi_transfer *t)
drivers/spi/spi-mpc52xx-psc.c
52
cs->speed_hz = (t && t->speed_hz)
drivers/spi/spi-mpc52xx-psc.c
53
? t->speed_hz : spi->max_speed_hz;
drivers/spi/spi-mpc52xx-psc.c
54
cs->bits_per_word = (t && t->bits_per_word)
drivers/spi/spi-mpc52xx-psc.c
55
? t->bits_per_word : spi->bits_per_word;
drivers/spi/spi-mt7621.c
166
struct spi_transfer *t = NULL;
drivers/spi/spi-mt7621.c
170
list_for_each_entry(t, &m->transfers, transfer_list)
drivers/spi/spi-mt7621.c
171
if (t->speed_hz < speed)
drivers/spi/spi-mt7621.c
172
speed = t->speed_hz;
drivers/spi/spi-mt7621.c
271
struct spi_transfer *t)
drivers/spi/spi-mt7621.c
275
if ((t->rx_buf) && (t->tx_buf)) {
drivers/spi/spi-mt7621.c
284
} else if (t->rx_buf) {
drivers/spi/spi-mt7621.c
285
mt7621_spi_read_half_duplex(rs, t->len, t->rx_buf);
drivers/spi/spi-mt7621.c
286
} else if (t->tx_buf) {
drivers/spi/spi-mt7621.c
287
mt7621_spi_write_half_duplex(rs, t->len, t->tx_buf);
drivers/spi/spi-mtk-nor.c
658
struct spi_transfer *t = NULL;
drivers/spi/spi-mtk-nor.c
667
list_for_each_entry(t, &m->transfers, transfer_list) {
drivers/spi/spi-mtk-nor.c
668
txbuf = t->tx_buf;
drivers/spi/spi-mtk-nor.c
669
for (i = 0; i < t->len; i++, reg_offset--) {
drivers/spi/spi-mtk-nor.c
676
trx_len += t->len;
drivers/spi/spi-mtk-nor.c
687
list_for_each_entry(t, &m->transfers, transfer_list) {
drivers/spi/spi-mtk-nor.c
688
rxbuf = t->rx_buf;
drivers/spi/spi-mtk-nor.c
689
for (i = 0; i < t->len; i++, reg_offset--) {
drivers/spi/spi-mxic.c
27
#define HC_CFG_TYPE(s, t) ((t) << (23 + ((s) * 2)))
drivers/spi/spi-mxic.c
607
struct spi_transfer *t)
drivers/spi/spi-mxic.c
613
if (t->rx_buf && t->tx_buf) {
drivers/spi/spi-mxic.c
621
ret = mxic_spi_set_freq(mxic, t->speed_hz);
drivers/spi/spi-mxic.c
625
if (t->tx_buf) {
drivers/spi/spi-mxic.c
630
} else if (t->rx_buf) {
drivers/spi/spi-mxic.c
638
OP_DATA_BUSW(busw) | (t->rx_buf ? OP_READ : 0),
drivers/spi/spi-mxic.c
641
ret = mxic_spi_data_xfer(mxic, t->tx_buf, t->rx_buf, t->len);
drivers/spi/spi-mxs.c
366
struct spi_transfer *t;
drivers/spi/spi-mxs.c
376
list_for_each_entry(t, &m->transfers, transfer_list) {
drivers/spi/spi-mxs.c
378
trace_spi_transfer_start(m, t);
drivers/spi/spi-mxs.c
380
status = mxs_spi_setup_transfer(m->spi, t);
drivers/spi/spi-mxs.c
384
t->effective_speed_hz = ssp->clk_rate;
drivers/spi/spi-mxs.c
387
flag = (&t->transfer_list == m->transfers.prev) ^ t->cs_change ?
drivers/spi/spi-mxs.c
399
if (t->len < 32) {
drivers/spi/spi-mxs.c
404
if (t->tx_buf)
drivers/spi/spi-mxs.c
406
(void *)t->tx_buf,
drivers/spi/spi-mxs.c
407
t->len, flag | TXRX_WRITE);
drivers/spi/spi-mxs.c
408
if (t->rx_buf)
drivers/spi/spi-mxs.c
410
t->rx_buf, t->len,
drivers/spi/spi-mxs.c
417
if (t->tx_buf)
drivers/spi/spi-mxs.c
419
(void *)t->tx_buf, t->len,
drivers/spi/spi-mxs.c
421
if (t->rx_buf)
drivers/spi/spi-mxs.c
423
t->rx_buf, t->len,
drivers/spi/spi-mxs.c
427
trace_spi_transfer_stop(m, t);
drivers/spi/spi-mxs.c
434
m->actual_length += t->len;
drivers/spi/spi-mxs.c
65
const struct spi_transfer *t)
drivers/spi/spi-mxs.c
69
const unsigned int hz = min(dev->max_speed_hz, t->speed_hz);
drivers/spi/spi-npcm-pspi.c
160
struct spi_transfer *t)
drivers/spi/spi-npcm-pspi.c
164
priv->tx_buf = t->tx_buf;
drivers/spi/spi-npcm-pspi.c
165
priv->rx_buf = t->rx_buf;
drivers/spi/spi-npcm-pspi.c
166
priv->tx_bytes = t->len;
drivers/spi/spi-npcm-pspi.c
167
priv->rx_bytes = t->len;
drivers/spi/spi-npcm-pspi.c
178
if (priv->bits_per_word == 8 && !(t->len & 0x1))
drivers/spi/spi-npcm-pspi.c
179
t->bits_per_word = 16;
drivers/spi/spi-npcm-pspi.c
181
if (!priv->is_save_param || priv->bits_per_word != t->bits_per_word) {
drivers/spi/spi-npcm-pspi.c
182
npcm_pspi_set_transfer_size(priv, t->bits_per_word);
drivers/spi/spi-npcm-pspi.c
183
priv->bits_per_word = t->bits_per_word;
drivers/spi/spi-npcm-pspi.c
186
if (!priv->is_save_param || priv->speed_hz != t->speed_hz) {
drivers/spi/spi-npcm-pspi.c
187
npcm_pspi_set_baudrate(priv, t->speed_hz);
drivers/spi/spi-npcm-pspi.c
188
priv->speed_hz = t->speed_hz;
drivers/spi/spi-npcm-pspi.c
250
struct spi_transfer *t)
drivers/spi/spi-npcm-pspi.c
255
npcm_pspi_setup_transfer(spi, t);
drivers/spi/spi-oc-tiny.c
107
static int tiny_spi_txrx_bufs(struct spi_device *spi, struct spi_transfer *t)
drivers/spi/spi-oc-tiny.c
110
const u8 *txp = t->tx_buf;
drivers/spi/spi-oc-tiny.c
111
u8 *rxp = t->rx_buf;
drivers/spi/spi-oc-tiny.c
116
hw->len = t->len;
drivers/spi/spi-oc-tiny.c
117
hw->txp = t->tx_buf;
drivers/spi/spi-oc-tiny.c
118
hw->rxp = t->rx_buf;
drivers/spi/spi-oc-tiny.c
123
if (t->len > 1) {
drivers/spi/spi-oc-tiny.c
142
for (i = 1; i < t->len; i++) {
drivers/spi/spi-oc-tiny.c
145
if (rxp || (i != t->len - 1))
drivers/spi/spi-oc-tiny.c
155
return t->len;
drivers/spi/spi-oc-tiny.c
67
struct spi_transfer *t)
drivers/spi/spi-oc-tiny.c
72
if (t) {
drivers/spi/spi-oc-tiny.c
73
if (t->speed_hz && t->speed_hz != hw->speed_hz)
drivers/spi/spi-oc-tiny.c
74
baud = tiny_spi_baud(spi, t->speed_hz);
drivers/spi/spi-omap-uwire.c
202
static int uwire_txrx(struct spi_device *spi, struct spi_transfer *t)
drivers/spi/spi-omap-uwire.c
204
unsigned len = t->len;
drivers/spi/spi-omap-uwire.c
205
unsigned bits = t->bits_per_word;
drivers/spi/spi-omap-uwire.c
210
if (!t->tx_buf && !t->rx_buf)
drivers/spi/spi-omap-uwire.c
216
if (t->tx_buf) {
drivers/spi/spi-omap-uwire.c
217
const u8 *buf = t->tx_buf;
drivers/spi/spi-omap-uwire.c
263
} else if (t->rx_buf) {
drivers/spi/spi-omap-uwire.c
264
u8 *buf = t->rx_buf;
drivers/spi/spi-omap-uwire.c
306
static int uwire_setup_transfer(struct spi_device *spi, struct spi_transfer *t)
drivers/spi/spi-omap-uwire.c
344
if (t != NULL)
drivers/spi/spi-omap-uwire.c
345
hz = t->speed_hz;
drivers/spi/spi-omap2-mcspi.c
1142
struct spi_transfer *t)
drivers/spi/spi-omap2-mcspi.c
1181
(t->speed_hz != spi->max_speed_hz) ||
drivers/spi/spi-omap2-mcspi.c
1182
(t->bits_per_word != spi->bits_per_word)) {
drivers/spi/spi-omap2-mcspi.c
1184
status = omap2_mcspi_setup_transfer(spi, t);
drivers/spi/spi-omap2-mcspi.c
1187
if (t->speed_hz == spi->max_speed_hz &&
drivers/spi/spi-omap2-mcspi.c
1188
t->bits_per_word == spi->bits_per_word)
drivers/spi/spi-omap2-mcspi.c
1196
if (t->tx_buf == NULL)
drivers/spi/spi-omap2-mcspi.c
1198
else if (t->rx_buf == NULL)
drivers/spi/spi-omap2-mcspi.c
1201
if (cd && cd->turbo_mode && t->tx_buf == NULL) {
drivers/spi/spi-omap2-mcspi.c
1203
if (t->len > ((cs->word_len + 7) >> 3))
drivers/spi/spi-omap2-mcspi.c
1209
if (t->len) {
drivers/spi/spi-omap2-mcspi.c
1213
spi_xfer_is_dma_mapped(ctlr, spi, t))
drivers/spi/spi-omap2-mcspi.c
1214
omap2_mcspi_set_fifo(spi, t, 1);
drivers/spi/spi-omap2-mcspi.c
1219
if (t->tx_buf == NULL)
drivers/spi/spi-omap2-mcspi.c
1224
spi_xfer_is_dma_mapped(ctlr, spi, t))
drivers/spi/spi-omap2-mcspi.c
1225
count = omap2_mcspi_txrx_dma(spi, t);
drivers/spi/spi-omap2-mcspi.c
1227
count = omap2_mcspi_txrx_pio(spi, t);
drivers/spi/spi-omap2-mcspi.c
1229
if (count != t->len) {
drivers/spi/spi-omap2-mcspi.c
1238
omap2_mcspi_set_fifo(spi, t, 0);
drivers/spi/spi-omap2-mcspi.c
1252
if (mcspi->fifo_depth > 0 && t)
drivers/spi/spi-omap2-mcspi.c
1253
omap2_mcspi_set_fifo(spi, t, 0);
drivers/spi/spi-omap2-mcspi.c
307
struct spi_transfer *t, int enable)
drivers/spi/spi-omap2-mcspi.c
321
if (t->len % bytes_per_word != 0)
drivers/spi/spi-omap2-mcspi.c
324
if (t->rx_buf != NULL && t->tx_buf != NULL)
drivers/spi/spi-omap2-mcspi.c
329
wcnt = t->len / bytes_per_word;
drivers/spi/spi-omap2-mcspi.c
334
if (t->rx_buf != NULL) {
drivers/spi/spi-omap2-mcspi.c
339
if (t->tx_buf != NULL) {
drivers/spi/spi-omap2-mcspi.c
352
if (t->rx_buf != NULL)
drivers/spi/spi-omap2-mcspi.c
355
if (t->tx_buf != NULL)
drivers/spi/spi-omap2-mcspi.c
911
struct spi_transfer *t)
drivers/spi/spi-omap2-mcspi.c
921
if (t != NULL && t->bits_per_word)
drivers/spi/spi-omap2-mcspi.c
922
word_len = t->bits_per_word;
drivers/spi/spi-omap2-mcspi.c
926
if (t && t->speed_hz)
drivers/spi/spi-omap2-mcspi.c
927
speed_hz = t->speed_hz;
drivers/spi/spi-orion.c
293
orion_spi_setup_transfer(struct spi_device *spi, struct spi_transfer *t)
drivers/spi/spi-orion.c
302
if ((t != NULL) && t->speed_hz)
drivers/spi/spi-orion.c
303
speed = t->speed_hz;
drivers/spi/spi-orion.c
305
if ((t != NULL) && t->bits_per_word)
drivers/spi/spi-orion.c
306
bits_per_word = t->bits_per_word;
drivers/spi/spi-orion.c
533
struct spi_transfer *t)
drivers/spi/spi-orion.c
537
status = orion_spi_setup_transfer(spi, t);
drivers/spi/spi-orion.c
541
if (t->len)
drivers/spi/spi-orion.c
542
orion_spi_write_read(spi, t);
drivers/spi/spi-ppc4xx.c
138
static int spi_ppc4xx_txrx(struct spi_device *spi, struct spi_transfer *t)
drivers/spi/spi-ppc4xx.c
144
t->tx_buf, t->rx_buf, t->len);
drivers/spi/spi-ppc4xx.c
148
hw->tx = t->tx_buf;
drivers/spi/spi-ppc4xx.c
149
hw->rx = t->rx_buf;
drivers/spi/spi-ppc4xx.c
150
hw->len = t->len;
drivers/spi/spi-ppc4xx.c
162
static int spi_ppc4xx_setupxfer(struct spi_device *spi, struct spi_transfer *t)
drivers/spi/spi-ppc4xx.c
177
if (t) {
drivers/spi/spi-ppc4xx.c
178
if (t->speed_hz)
drivers/spi/spi-ppc4xx.c
179
speed = min(t->speed_hz, spi->max_speed_hz);
drivers/spi/spi-rb4xx.c
105
struct spi_device *spi, struct spi_transfer *t)
drivers/spi/spi-rb4xx.c
126
tx_buf = t->tx_buf;
drivers/spi/spi-rb4xx.c
127
rx_buf = t->rx_buf;
drivers/spi/spi-rb4xx.c
128
for (i = 0; i < t->len; ++i) {
drivers/spi/spi-rb4xx.c
129
if (t->tx_nbits == SPI_NBITS_DUAL)
drivers/spi/spi-rzv2m-csi.c
383
struct spi_transfer *t)
drivers/spi/spi-rzv2m-csi.c
385
if (t->rx_buf && !t->tx_buf)
drivers/spi/spi-rzv2m-csi.c
392
csi->bytes_per_word = t->bits_per_word / 8;
drivers/spi/spi-s3c64xx.c
126
#define msecs_to_loops(t) (loops_per_jiffy / 1000 * HZ * t)
drivers/spi/spi-sc18is602.c
113
if (t->rx_buf) {
drivers/spi/spi-sc18is602.c
124
memcpy(t->rx_buf, &hw->buffer[hw->rindex], len);
drivers/spi/spi-sc18is602.c
176
struct spi_transfer *t, int tlen)
drivers/spi/spi-sc18is602.c
178
if (t && t->len + tlen > SC18IS602_BUFSIZ + 1)
drivers/spi/spi-sc18is602.c
189
struct spi_transfer *t;
drivers/spi/spi-sc18is602.c
193
list_for_each_entry(t, &m->transfers, transfer_list) {
drivers/spi/spi-sc18is602.c
196
status = sc18is602_check_transfer(spi, t, hw->tlen);
drivers/spi/spi-sc18is602.c
200
status = sc18is602_setup_transfer(hw, t->speed_hz, spi->mode);
drivers/spi/spi-sc18is602.c
204
do_transfer = t->cs_change || list_is_last(&t->transfer_list,
drivers/spi/spi-sc18is602.c
207
if (t->len) {
drivers/spi/spi-sc18is602.c
208
status = sc18is602_txrx(hw, m, t, do_transfer);
drivers/spi/spi-sc18is602.c
215
spi_transfer_delay_exec(t);
drivers/spi/spi-sc18is602.c
67
struct spi_transfer *t, bool do_transfer)
drivers/spi/spi-sc18is602.c
69
unsigned int len = t->len;
drivers/spi/spi-sc18is602.c
83
if (t->tx_buf) {
drivers/spi/spi-sc18is602.c
84
memcpy(&hw->buffer[hw->tlen], t->tx_buf, len);
drivers/spi/spi-sc18is602.c
86
if (t->rx_buf)
drivers/spi/spi-sc18is602.c
90
} else if (t->rx_buf) {
drivers/spi/spi-sh-hspi.c
123
tmp = abs(t->speed_hz - rate);
drivers/spi/spi-sh-hspi.c
136
dev_dbg(dev, "speed %d/%d\n", t->speed_hz, best_rate);
drivers/spi/spi-sh-hspi.c
147
struct spi_transfer *t;
drivers/spi/spi-sh-hspi.c
158
list_for_each_entry(t, &msg->transfers, transfer_list) {
drivers/spi/spi-sh-hspi.c
161
hspi_hw_setup(hspi, msg, t);
drivers/spi/spi-sh-hspi.c
165
cs_change = t->cs_change;
drivers/spi/spi-sh-hspi.c
167
for (i = 0; i < t->len; i++) {
drivers/spi/spi-sh-hspi.c
175
if (t->tx_buf)
drivers/spi/spi-sh-hspi.c
176
tx = (u32)((u8 *)t->tx_buf)[i];
drivers/spi/spi-sh-hspi.c
186
if (t->rx_buf)
drivers/spi/spi-sh-hspi.c
187
((u8 *)t->rx_buf)[i] = (u8)rx;
drivers/spi/spi-sh-hspi.c
191
msg->actual_length += t->len;
drivers/spi/spi-sh-hspi.c
193
spi_transfer_delay_exec(t);
drivers/spi/spi-sh-hspi.c
71
int t = 256;
drivers/spi/spi-sh-hspi.c
73
while (t--) {
drivers/spi/spi-sh-hspi.c
97
struct spi_transfer *t)
drivers/spi/spi-sh-msiof.c
131
struct spi_transfer *t)
drivers/spi/spi-sh-msiof.c
135
u32 spi_hz = t->speed_hz;
drivers/spi/spi-sh-msiof.c
166
t->effective_speed_hz = parent_rate / (brps << div_pow);
drivers/spi/spi-sh-msiof.c
817
struct spi_transfer *t)
drivers/spi/spi-sh-msiof.c
826
const void *tx_buf = t->tx_buf;
drivers/spi/spi-sh-msiof.c
827
void *rx_buf = t->rx_buf;
drivers/spi/spi-sh-msiof.c
828
unsigned int len = t->len;
drivers/spi/spi-sh-msiof.c
829
unsigned int bits = t->bits_per_word;
drivers/spi/spi-sh-msiof.c
841
sh_msiof_spi_set_clk_regs(p, t);
drivers/spi/spi-sh-msiof.c
951
bits = t->bits_per_word;
drivers/spi/spi-sh.c
151
struct spi_transfer *t)
drivers/spi/spi-sh.c
154
int remain = t->len;
drivers/spi/spi-sh.c
159
if (t->len)
drivers/spi/spi-sh.c
162
data = (unsigned char *)t->tx_buf;
drivers/spi/spi-sh.c
197
if (list_is_last(&t->transfer_list, &mesg->transfers)) {
drivers/spi/spi-sh.c
216
struct spi_transfer *t)
drivers/spi/spi-sh.c
219
int remain = t->len;
drivers/spi/spi-sh.c
224
if (t->len > SPI_SH_MAX_BYTE)
drivers/spi/spi-sh.c
227
spi_sh_write(ss, t->len, SPI_SH_CR3);
drivers/spi/spi-sh.c
234
data = (unsigned char *)t->rx_buf;
drivers/spi/spi-sh.c
261
if (t->len > SPI_SH_MAX_BYTE) {
drivers/spi/spi-sh.c
275
struct spi_transfer *t;
drivers/spi/spi-sh.c
282
list_for_each_entry(t, &mesg->transfers, transfer_list) {
drivers/spi/spi-sh.c
284
t->tx_buf, t->rx_buf);
drivers/spi/spi-sh.c
286
t->len, t->delay.value);
drivers/spi/spi-sh.c
288
if (t->tx_buf) {
drivers/spi/spi-sh.c
289
ret = spi_sh_send(ss, mesg, t);
drivers/spi/spi-sh.c
293
if (t->rx_buf) {
drivers/spi/spi-sh.c
294
ret = spi_sh_receive(ss, mesg, t);
drivers/spi/spi-sh.c
298
mesg->actual_length += t->len;
drivers/spi/spi-sifive.c
168
struct spi_transfer *t)
drivers/spi/spi-sifive.c
174
cr = DIV_ROUND_UP(clk_get_rate(spi->clk) >> 1, t->speed_hz) - 1;
drivers/spi/spi-sifive.c
178
mode = max_t(unsigned int, t->rx_nbits, t->tx_nbits);
drivers/spi/spi-sifive.c
181
cr = SIFIVE_SPI_FMT_LEN(t->bits_per_word);
drivers/spi/spi-sifive.c
195
if (!t->rx_buf)
drivers/spi/spi-sifive.c
205
return 1600000 * spi->fifo_depth <= t->speed_hz * mode;
drivers/spi/spi-sifive.c
256
struct spi_transfer *t)
drivers/spi/spi-sifive.c
259
int poll = sifive_spi_prep_transfer(spi, device, t);
drivers/spi/spi-sifive.c
260
const u8 *tx_ptr = t->tx_buf;
drivers/spi/spi-sifive.c
261
u8 *rx_ptr = t->rx_buf;
drivers/spi/spi-sifive.c
262
unsigned int remaining_words = t->len;
drivers/spi/spi-sprd-adi.c
335
struct spi_transfer *t)
drivers/spi/spi-sprd-adi.c
341
if (t->rx_buf) {
drivers/spi/spi-sprd-adi.c
342
reg = *(u32 *)t->rx_buf;
drivers/spi/spi-sprd-adi.c
344
*(u32 *)t->rx_buf = val;
drivers/spi/spi-sprd-adi.c
345
} else if (t->tx_buf) {
drivers/spi/spi-sprd-adi.c
346
u32 *p = (u32 *)t->tx_buf;
drivers/spi/spi-sprd.c
173
struct spi_transfer *t)
drivers/spi/spi-sprd.c
179
u32 size = t->bits_per_word * SPRD_SPI_FIFO_SIZE;
drivers/spi/spi-sprd.c
193
static int sprd_spi_wait_for_tx_end(struct sprd_spi *ss, struct spi_transfer *t)
drivers/spi/spi-sprd.c
198
us = sprd_spi_transfer_max_timeout(ss, t);
drivers/spi/spi-sprd.c
218
static int sprd_spi_wait_for_rx_end(struct sprd_spi *ss, struct spi_transfer *t)
drivers/spi/spi-sprd.c
223
us = sprd_spi_transfer_max_timeout(ss, t);
drivers/spi/spi-sprd.c
401
static int sprd_spi_txrx_bufs(struct spi_device *sdev, struct spi_transfer *t)
drivers/spi/spi-sprd.c
421
ret = sprd_spi_wait_for_tx_end(ss, t);
drivers/spi/spi-sprd.c
434
ret = sprd_spi_wait_for_rx_end(ss, t);
drivers/spi/spi-sprd.c
516
static int sprd_spi_dma_rx_config(struct sprd_spi *ss, struct spi_transfer *t)
drivers/spi/spi-sprd.c
527
ret = sprd_spi_dma_submit(dma_chan, &config, &t->rx_sg, DMA_DEV_TO_MEM);
drivers/spi/spi-sprd.c
534
static int sprd_spi_dma_tx_config(struct sprd_spi *ss, struct spi_transfer *t)
drivers/spi/spi-sprd.c
545
ret = sprd_spi_dma_submit(dma_chan, &config, &t->tx_sg, DMA_MEM_TO_DEV);
drivers/spi/spi-sprd.c
549
return t->len;
drivers/spi/spi-sprd.c
579
struct spi_transfer *t)
drivers/spi/spi-sprd.c
588
write_size = sprd_spi_dma_tx_config(ss, t);
drivers/spi/spi-sprd.c
624
ss->dma.rx_len = t->len > ss->dma.fragmens_len ?
drivers/spi/spi-sprd.c
625
(t->len - t->len % ss->dma.fragmens_len) :
drivers/spi/spi-sprd.c
626
t->len;
drivers/spi/spi-sprd.c
627
ret = sprd_spi_dma_rx_config(ss, t);
drivers/spi/spi-sprd.c
664
static int sprd_spi_init_hw(struct sprd_spi *ss, struct spi_transfer *t)
drivers/spi/spi-sprd.c
666
struct spi_delay *d = &t->word_delay;
drivers/spi/spi-sprd.c
715
struct spi_transfer *t)
drivers/spi/spi-sprd.c
718
u8 bits_per_word = t->bits_per_word;
drivers/spi/spi-sprd.c
722
ss->len = t->len;
drivers/spi/spi-sprd.c
723
ss->tx_buf = t->tx_buf;
drivers/spi/spi-sprd.c
724
ss->rx_buf = t->rx_buf;
drivers/spi/spi-sprd.c
727
ret = sprd_spi_init_hw(ss, t);
drivers/spi/spi-sprd.c
732
sprd_spi_set_speed(ss, t->speed_hz);
drivers/spi/spi-sprd.c
742
ss->trans_len = t->len;
drivers/spi/spi-sprd.c
749
ss->trans_len = t->len >> 1;
drivers/spi/spi-sprd.c
756
ss->trans_len = t->len >> 2;
drivers/spi/spi-sprd.c
769
if (t->tx_buf)
drivers/spi/spi-sprd.c
771
if (t->rx_buf)
drivers/spi/spi-sprd.c
790
struct spi_transfer *t)
drivers/spi/spi-sprd.c
794
ret = sprd_spi_setup_transfer(sdev, t);
drivers/spi/spi-sprd.c
798
if (sctlr->can_dma(sctlr, sdev, t))
drivers/spi/spi-sprd.c
799
ret = sprd_spi_dma_txrx_bufs(sdev, t);
drivers/spi/spi-sprd.c
801
ret = sprd_spi_txrx_bufs(sdev, t);
drivers/spi/spi-sprd.c
803
if (ret == t->len)
drivers/spi/spi-sprd.c
890
struct spi_device *spi, struct spi_transfer *t)
drivers/spi/spi-sprd.c
894
return ss->dma.enable && (t->len > SPRD_SPI_FIFO_SIZE);
drivers/spi/spi-st-ssc4.c
119
struct spi_device *spi, struct spi_transfer *t)
drivers/spi/spi-st-ssc4.c
125
spi_st->tx_ptr = t->tx_buf;
drivers/spi/spi-st-ssc4.c
126
spi_st->rx_ptr = t->rx_buf;
drivers/spi/spi-st-ssc4.c
134
spi_st->words_remaining = t->len / 2;
drivers/spi/spi-st-ssc4.c
136
} else if (spi->bits_per_word == 8 && !(t->len & 0x1)) {
drivers/spi/spi-st-ssc4.c
142
spi_st->words_remaining = t->len / 2;
drivers/spi/spi-st-ssc4.c
152
spi_st->words_remaining = t->len;
drivers/spi/spi-st-ssc4.c
170
return t->len;
drivers/spi/spi-tegra114.c
1124
struct spi_transfer *t = tspi->curr_xfer;
drivers/spi/spi-tegra114.c
1144
tegra_spi_read_rx_fifo_to_client_rxbuf(tspi, t);
drivers/spi/spi-tegra114.c
1151
if (tspi->cur_pos == t->len) {
drivers/spi/spi-tegra114.c
1156
tegra_spi_calculate_curr_xfer_param(tspi->cur_spi, tspi, t);
drivers/spi/spi-tegra114.c
1157
tegra_spi_start_cpu_based_transfer(tspi, t);
drivers/spi/spi-tegra114.c
1165
struct spi_transfer *t = tspi->curr_xfer;
drivers/spi/spi-tegra114.c
1219
tegra_spi_copy_spi_rxbuf_to_client_rxbuf(tspi, t);
drivers/spi/spi-tegra114.c
1226
if (tspi->cur_pos == t->len) {
drivers/spi/spi-tegra114.c
1233
tspi, t);
drivers/spi/spi-tegra114.c
1235
err = tegra_spi_start_dma_based_transfer(tspi, t);
drivers/spi/spi-tegra114.c
1237
err = tegra_spi_start_cpu_based_transfer(tspi, t);
drivers/spi/spi-tegra114.c
259
struct spi_transfer *t)
drivers/spi/spi-tegra114.c
261
unsigned remain_len = t->len - tspi->cur_pos;
drivers/spi/spi-tegra114.c
263
unsigned bits_per_word = t->bits_per_word;
drivers/spi/spi-tegra114.c
270
bits_per_word == 32) && t->len > 3) {
drivers/spi/spi-tegra114.c
292
struct tegra_spi_data *tspi, struct spi_transfer *t)
drivers/spi/spi-tegra114.c
301
u8 *tx_buf = (u8 *)t->tx_buf + tspi->cur_tx_pos;
drivers/spi/spi-tegra114.c
325
if (nbytes > t->len - tspi->cur_pos)
drivers/spi/spi-tegra114.c
326
nbytes = t->len - tspi->cur_pos;
drivers/spi/spi-tegra114.c
344
struct tegra_spi_data *tspi, struct spi_transfer *t)
drivers/spi/spi-tegra114.c
351
u8 *rx_buf = (u8 *)t->rx_buf + tspi->cur_rx_pos;
drivers/spi/spi-tegra114.c
366
u32 rx_mask = ((u32)1 << t->bits_per_word) - 1;
drivers/spi/spi-tegra114.c
371
if (len > t->len - tspi->cur_pos)
drivers/spi/spi-tegra114.c
372
len = t->len - tspi->cur_pos;
drivers/spi/spi-tegra114.c
388
struct tegra_spi_data *tspi, struct spi_transfer *t)
drivers/spi/spi-tegra114.c
397
memcpy(tspi->tx_dma_buf, t->tx_buf + tspi->cur_pos, len);
drivers/spi/spi-tegra114.c
402
u8 *tx_buf = (u8 *)t->tx_buf + tspi->cur_tx_pos;
drivers/spi/spi-tegra114.c
406
if (consume > t->len - tspi->cur_pos)
drivers/spi/spi-tegra114.c
407
consume = t->len - tspi->cur_pos;
drivers/spi/spi-tegra114.c
427
struct tegra_spi_data *tspi, struct spi_transfer *t)
drivers/spi/spi-tegra114.c
436
memcpy(t->rx_buf + tspi->cur_rx_pos, tspi->rx_dma_buf, len);
drivers/spi/spi-tegra114.c
441
unsigned char *rx_buf = t->rx_buf + tspi->cur_rx_pos;
drivers/spi/spi-tegra114.c
442
u32 rx_mask = ((u32)1 << t->bits_per_word) - 1;
drivers/spi/spi-tegra114.c
446
if (consume > t->len - tspi->cur_pos)
drivers/spi/spi-tegra114.c
447
consume = t->len - tspi->cur_pos;
drivers/spi/spi-tegra114.c
535
struct tegra_spi_data *tspi, struct spi_transfer *t)
drivers/spi/spi-tegra114.c
587
tegra_spi_copy_client_txbuf_to_spi_txbuf(tspi, t);
drivers/spi/spi-tegra114.c
629
struct tegra_spi_data *tspi, struct spi_transfer *t)
drivers/spi/spi-tegra114.c
635
cur_words = tegra_spi_fill_tx_fifo_from_client_txbuf(tspi, t);
drivers/spi/spi-tegra114.c
771
struct spi_transfer *t,
drivers/spi/spi-tegra114.c
777
u32 speed = t->speed_hz;
drivers/spi/spi-tegra114.c
778
u8 bits_per_word = t->bits_per_word;
drivers/spi/spi-tegra114.c
792
tspi->curr_xfer = t;
drivers/spi/spi-tegra114.c
832
if (is_single_xfer && !(t->cs_change)) {
drivers/spi/spi-tegra114.c
866
struct spi_transfer *t, u32 command1)
drivers/spi/spi-tegra114.c
872
total_fifo_words = tegra_spi_calculate_curr_xfer_param(spi, tspi, t);
drivers/spi/spi-tegra114.c
874
if (t->rx_nbits == SPI_NBITS_DUAL || t->tx_nbits == SPI_NBITS_DUAL)
drivers/spi/spi-tegra114.c
886
if (t->rx_buf) {
drivers/spi/spi-tegra114.c
890
if (t->tx_buf) {
drivers/spi/spi-tegra114.c
905
ret = tegra_spi_start_dma_based_transfer(tspi, t);
drivers/spi/spi-tegra114.c
907
ret = tegra_spi_start_cpu_based_transfer(tspi, t);
drivers/spi/spi-tegra20-sflash.c
159
struct spi_transfer *t)
drivers/spi/spi-tegra20-sflash.c
161
unsigned remain_len = t->len - tsd->cur_pos;
drivers/spi/spi-tegra20-sflash.c
164
tsd->bytes_per_word = DIV_ROUND_UP(t->bits_per_word, 8);
drivers/spi/spi-tegra20-sflash.c
173
struct tegra_sflash_data *tsd, struct spi_transfer *t)
drivers/spi/spi-tegra20-sflash.c
178
u8 *tx_buf = (u8 *)t->tx_buf + tsd->cur_tx_pos;
drivers/spi/spi-tegra20-sflash.c
203
struct tegra_sflash_data *tsd, struct spi_transfer *t)
drivers/spi/spi-tegra20-sflash.c
207
u8 *rx_buf = (u8 *)t->rx_buf + tsd->cur_rx_pos;
drivers/spi/spi-tegra20-sflash.c
224
struct tegra_sflash_data *tsd, struct spi_transfer *t)
drivers/spi/spi-tegra20-sflash.c
239
cur_words = tegra_sflash_fill_tx_fifo_from_client_txbuf(tsd, t);
drivers/spi/spi-tegra20-sflash.c
251
struct spi_transfer *t, bool is_first_of_msg,
drivers/spi/spi-tegra20-sflash.c
258
speed = t->speed_hz;
drivers/spi/spi-tegra20-sflash.c
268
tsd->curr_xfer = t;
drivers/spi/spi-tegra20-sflash.c
269
tegra_sflash_calculate_curr_xfer_param(spi, tsd, t);
drivers/spi/spi-tegra20-sflash.c
272
command |= SPI_BIT_LENGTH(t->bits_per_word - 1);
drivers/spi/spi-tegra20-sflash.c
287
command |= SPI_BIT_LENGTH(t->bits_per_word - 1);
drivers/spi/spi-tegra20-sflash.c
292
if (t->rx_buf) {
drivers/spi/spi-tegra20-sflash.c
296
if (t->tx_buf) {
drivers/spi/spi-tegra20-sflash.c
303
return tegra_sflash_start_cpu_based_transfer(tsd, t);
drivers/spi/spi-tegra20-sflash.c
360
struct spi_transfer *t = tsd->curr_xfer;
drivers/spi/spi-tegra20-sflash.c
377
tegra_sflash_read_rx_fifo_to_client_rxbuf(tsd, t);
drivers/spi/spi-tegra20-sflash.c
384
if (tsd->cur_pos == t->len) {
drivers/spi/spi-tegra20-sflash.c
389
tegra_sflash_calculate_curr_xfer_param(tsd->cur_spi, tsd, t);
drivers/spi/spi-tegra20-sflash.c
390
tegra_sflash_start_cpu_based_transfer(tsd, t);
drivers/spi/spi-tegra20-slink.c
237
struct spi_transfer *t)
drivers/spi/spi-tegra20-slink.c
255
struct spi_transfer *t)
drivers/spi/spi-tegra20-slink.c
257
unsigned remain_len = t->len - tspi->cur_pos;
drivers/spi/spi-tegra20-slink.c
263
bits_per_word = t->bits_per_word;
drivers/spi/spi-tegra20-slink.c
273
tspi->packed_size = tegra_slink_get_packed_size(tspi, t);
drivers/spi/spi-tegra20-slink.c
289
struct tegra_slink_data *tspi, struct spi_transfer *t)
drivers/spi/spi-tegra20-slink.c
298
u8 *tx_buf = (u8 *)t->tx_buf + tspi->cur_tx_pos;
drivers/spi/spi-tegra20-slink.c
331
struct tegra_slink_data *tspi, struct spi_transfer *t)
drivers/spi/spi-tegra20-slink.c
338
u8 *rx_buf = (u8 *)t->rx_buf + tspi->cur_rx_pos;
drivers/spi/spi-tegra20-slink.c
364
struct tegra_slink_data *tspi, struct spi_transfer *t)
drivers/spi/spi-tegra20-slink.c
372
memcpy(tspi->tx_dma_buf, t->tx_buf + tspi->cur_pos, len);
drivers/spi/spi-tegra20-slink.c
376
u8 *tx_buf = (u8 *)t->tx_buf + tspi->cur_tx_pos;
drivers/spi/spi-tegra20-slink.c
395
struct tegra_slink_data *tspi, struct spi_transfer *t)
drivers/spi/spi-tegra20-slink.c
405
memcpy(t->rx_buf + tspi->cur_rx_pos, tspi->rx_dma_buf, len);
drivers/spi/spi-tegra20-slink.c
409
unsigned char *rx_buf = t->rx_buf + tspi->cur_rx_pos;
drivers/spi/spi-tegra20-slink.c
410
u32 rx_mask = ((u32)1 << t->bits_per_word) - 1;
drivers/spi/spi-tegra20-slink.c
471
struct tegra_slink_data *tspi, struct spi_transfer *t)
drivers/spi/spi-tegra20-slink.c
512
tegra_slink_copy_client_txbuf_to_spi_txbuf(tspi, t);
drivers/spi/spi-tegra20-slink.c
556
struct tegra_slink_data *tspi, struct spi_transfer *t)
drivers/spi/spi-tegra20-slink.c
572
cur_words = tegra_slink_fill_tx_fifo_from_client_txbuf(tspi, t);
drivers/spi/spi-tegra20-slink.c
672
struct spi_transfer *t)
drivers/spi/spi-tegra20-slink.c
682
bits_per_word = t->bits_per_word;
drivers/spi/spi-tegra20-slink.c
683
speed = t->speed_hz;
drivers/spi/spi-tegra20-slink.c
693
tspi->curr_xfer = t;
drivers/spi/spi-tegra20-slink.c
694
total_fifo_words = tegra_slink_calculate_curr_xfer_param(spi, tspi, t);
drivers/spi/spi-tegra20-slink.c
704
if (t->rx_buf) {
drivers/spi/spi-tegra20-slink.c
708
if (t->tx_buf) {
drivers/spi/spi-tegra20-slink.c
725
ret = tegra_slink_start_dma_based_transfer(tspi, t);
drivers/spi/spi-tegra20-slink.c
727
ret = tegra_slink_start_cpu_based_transfer(tspi, t);
drivers/spi/spi-tegra20-slink.c
841
struct spi_transfer *t = tspi->curr_xfer;
drivers/spi/spi-tegra20-slink.c
860
tegra_slink_read_rx_fifo_to_client_rxbuf(tspi, t);
drivers/spi/spi-tegra20-slink.c
867
if (tspi->cur_pos == t->len) {
drivers/spi/spi-tegra20-slink.c
872
tegra_slink_calculate_curr_xfer_param(tspi->cur_spi, tspi, t);
drivers/spi/spi-tegra20-slink.c
873
tegra_slink_start_cpu_based_transfer(tspi, t);
drivers/spi/spi-tegra20-slink.c
881
struct spi_transfer *t = tspi->curr_xfer;
drivers/spi/spi-tegra20-slink.c
934
tegra_slink_copy_spi_rxbuf_to_client_rxbuf(tspi, t);
drivers/spi/spi-tegra20-slink.c
941
if (tspi->cur_pos == t->len) {
drivers/spi/spi-tegra20-slink.c
948
tspi, t);
drivers/spi/spi-tegra20-slink.c
950
err = tegra_slink_start_dma_based_transfer(tspi, t);
drivers/spi/spi-tegra20-slink.c
952
err = tegra_slink_start_cpu_based_transfer(tspi, t);
drivers/spi/spi-tegra210-quad.c
1454
struct spi_transfer *t;
drivers/spi/spi-tegra210-quad.c
1458
t = tqspi->curr_xfer;
drivers/spi/spi-tegra210-quad.c
1460
if (!t) {
drivers/spi/spi-tegra210-quad.c
1472
tegra_qspi_read_rx_fifo_to_client_rxbuf(tqspi, t);
drivers/spi/spi-tegra210-quad.c
1479
if (tqspi->cur_pos == t->len) {
drivers/spi/spi-tegra210-quad.c
1484
tegra_qspi_calculate_curr_xfer_param(tqspi, t);
drivers/spi/spi-tegra210-quad.c
1485
tegra_qspi_start_cpu_based_transfer(tqspi, t);
drivers/spi/spi-tegra210-quad.c
1494
struct spi_transfer *t;
drivers/spi/spi-tegra210-quad.c
1533
t = tqspi->curr_xfer;
drivers/spi/spi-tegra210-quad.c
1535
if (!t) {
drivers/spi/spi-tegra210-quad.c
1541
tegra_qspi_dma_unmap_xfer(tqspi, t);
drivers/spi/spi-tegra210-quad.c
1548
tegra_qspi_copy_qspi_rxbuf_to_client_rxbuf(tqspi, t);
drivers/spi/spi-tegra210-quad.c
1555
if (tqspi->cur_pos == t->len) {
drivers/spi/spi-tegra210-quad.c
1556
tegra_qspi_dma_unmap_xfer(tqspi, t);
drivers/spi/spi-tegra210-quad.c
1561
tegra_qspi_dma_unmap_xfer(tqspi, t);
drivers/spi/spi-tegra210-quad.c
1564
total_fifo_words = tegra_qspi_calculate_curr_xfer_param(tqspi, t);
drivers/spi/spi-tegra210-quad.c
1566
num_errors = tegra_qspi_start_dma_based_transfer(tqspi, t);
drivers/spi/spi-tegra210-quad.c
1568
num_errors = tegra_qspi_start_cpu_based_transfer(tqspi, t);
drivers/spi/spi-tegra210-quad.c
276
tegra_qspi_calculate_curr_xfer_param(struct tegra_qspi *tqspi, struct spi_transfer *t)
drivers/spi/spi-tegra210-quad.c
279
unsigned int remain_len = t->len - tqspi->cur_pos;
drivers/spi/spi-tegra210-quad.c
280
unsigned int bits_per_word = t->bits_per_word;
drivers/spi/spi-tegra210-quad.c
292
bits_per_word == 32) && t->len > 3) {
drivers/spi/spi-tegra210-quad.c
315
tegra_qspi_fill_tx_fifo_from_client_txbuf(struct tegra_qspi *tqspi, struct spi_transfer *t)
drivers/spi/spi-tegra210-quad.c
319
u8 *tx_buf = (u8 *)t->tx_buf + tqspi->cur_tx_pos;
drivers/spi/spi-tegra210-quad.c
346
if (len > t->len - tqspi->cur_pos)
drivers/spi/spi-tegra210-quad.c
347
len = t->len - tqspi->cur_pos;
drivers/spi/spi-tegra210-quad.c
364
tegra_qspi_read_rx_fifo_to_client_rxbuf(struct tegra_qspi *tqspi, struct spi_transfer *t)
drivers/spi/spi-tegra210-quad.c
366
u8 *rx_buf = (u8 *)t->rx_buf + tqspi->cur_rx_pos;
drivers/spi/spi-tegra210-quad.c
385
u32 rx_mask = ((u32)1 << t->bits_per_word) - 1;
drivers/spi/spi-tegra210-quad.c
390
if (len > t->len - tqspi->cur_pos)
drivers/spi/spi-tegra210-quad.c
391
len = t->len - tqspi->cur_pos;
drivers/spi/spi-tegra210-quad.c
408
tegra_qspi_copy_client_txbuf_to_qspi_txbuf(struct tegra_qspi *tqspi, struct spi_transfer *t)
drivers/spi/spi-tegra210-quad.c
421
u8 *tx_buf = (u8 *)t->tx_buf + tqspi->cur_tx_pos;
drivers/spi/spi-tegra210-quad.c
429
if (consume > t->len - tqspi->cur_pos)
drivers/spi/spi-tegra210-quad.c
430
consume = t->len - tqspi->cur_pos;
drivers/spi/spi-tegra210-quad.c
445
tegra_qspi_copy_qspi_rxbuf_to_client_rxbuf(struct tegra_qspi *tqspi, struct spi_transfer *t)
drivers/spi/spi-tegra210-quad.c
450
unsigned char *rx_buf = t->rx_buf + tqspi->cur_rx_pos;
drivers/spi/spi-tegra210-quad.c
451
u32 rx_mask = ((u32)1 << t->bits_per_word) - 1;
drivers/spi/spi-tegra210-quad.c
460
if (consume > t->len - tqspi->cur_pos)
drivers/spi/spi-tegra210-quad.c
461
consume = t->len - tqspi->cur_pos;
drivers/spi/spi-tegra210-quad.c
481
static int tegra_qspi_start_tx_dma(struct tegra_qspi *tqspi, struct spi_transfer *t, int len)
drivers/spi/spi-tegra210-quad.c
488
tx_dma_phys = t->tx_dma;
drivers/spi/spi-tegra210-quad.c
509
static int tegra_qspi_start_rx_dma(struct tegra_qspi *tqspi, struct spi_transfer *t, int len)
drivers/spi/spi-tegra210-quad.c
516
rx_dma_phys = t->rx_dma;
drivers/spi/spi-tegra210-quad.c
568
static int tegra_qspi_dma_map_xfer(struct tegra_qspi *tqspi, struct spi_transfer *t)
drivers/spi/spi-tegra210-quad.c
570
u8 *tx_buf = (u8 *)t->tx_buf + tqspi->cur_tx_pos;
drivers/spi/spi-tegra210-quad.c
571
u8 *rx_buf = (u8 *)t->rx_buf + tqspi->cur_rx_pos;
drivers/spi/spi-tegra210-quad.c
576
if (t->tx_buf) {
drivers/spi/spi-tegra210-quad.c
577
t->tx_dma = dma_map_single(tqspi->dev, (void *)tx_buf, len, DMA_TO_DEVICE);
drivers/spi/spi-tegra210-quad.c
578
if (dma_mapping_error(tqspi->dev, t->tx_dma))
drivers/spi/spi-tegra210-quad.c
582
if (t->rx_buf) {
drivers/spi/spi-tegra210-quad.c
583
t->rx_dma = dma_map_single(tqspi->dev, (void *)rx_buf, len, DMA_FROM_DEVICE);
drivers/spi/spi-tegra210-quad.c
584
if (dma_mapping_error(tqspi->dev, t->rx_dma)) {
drivers/spi/spi-tegra210-quad.c
585
dma_unmap_single(tqspi->dev, t->tx_dma, len, DMA_TO_DEVICE);
drivers/spi/spi-tegra210-quad.c
593
static void tegra_qspi_dma_unmap_xfer(struct tegra_qspi *tqspi, struct spi_transfer *t)
drivers/spi/spi-tegra210-quad.c
599
if (t->tx_buf)
drivers/spi/spi-tegra210-quad.c
600
dma_unmap_single(tqspi->dev, t->tx_dma, len, DMA_TO_DEVICE);
drivers/spi/spi-tegra210-quad.c
601
if (t->rx_buf)
drivers/spi/spi-tegra210-quad.c
602
dma_unmap_single(tqspi->dev, t->rx_dma, len, DMA_FROM_DEVICE);
drivers/spi/spi-tegra210-quad.c
605
static int tegra_qspi_start_dma_based_transfer(struct tegra_qspi *tqspi, struct spi_transfer *t)
drivers/spi/spi-tegra210-quad.c
615
ret = tegra_qspi_dma_map_xfer(tqspi, t);
drivers/spi/spi-tegra210-quad.c
662
tegra_qspi_copy_client_txbuf_to_qspi_txbuf(tqspi, t);
drivers/spi/spi-tegra210-quad.c
663
ret = tegra_qspi_start_tx_dma(tqspi, t, len);
drivers/spi/spi-tegra210-quad.c
670
tx_dma_phys = t->tx_dma;
drivers/spi/spi-tegra210-quad.c
673
tegra_qspi_copy_client_txbuf_to_qspi_txbuf(tqspi, t);
drivers/spi/spi-tegra210-quad.c
692
ret = tegra_qspi_start_rx_dma(tqspi, t, len);
drivers/spi/spi-tegra210-quad.c
701
rx_dma_phys = t->rx_dma;
drivers/spi/spi-tegra210-quad.c
722
static int tegra_qspi_start_cpu_based_transfer(struct tegra_qspi *qspi, struct spi_transfer *t)
drivers/spi/spi-tegra210-quad.c
728
cur_words = tegra_qspi_fill_tx_fifo_from_client_txbuf(qspi, t);
drivers/spi/spi-tegra210-quad.c
834
static u32 tegra_qspi_setup_transfer_one(struct spi_device *spi, struct spi_transfer *t,
drivers/spi/spi-tegra210-quad.c
839
u32 command1, command2, speed = t->speed_hz;
drivers/spi/spi-tegra210-quad.c
840
u8 bits_per_word = t->bits_per_word;
drivers/spi/spi-tegra210-quad.c
854
tqspi->curr_xfer = t;
drivers/spi/spi-tegra210-quad.c
899
struct spi_transfer *t, u32 command1)
drivers/spi/spi-tegra210-quad.c
906
total_fifo_words = tegra_qspi_calculate_curr_xfer_param(tqspi, t);
drivers/spi/spi-tegra210-quad.c
916
if (t->rx_buf) {
drivers/spi/spi-tegra210-quad.c
919
bus_width = t->rx_nbits;
drivers/spi/spi-tegra210-quad.c
922
if (t->tx_buf) {
drivers/spi/spi-tegra210-quad.c
925
bus_width = t->tx_nbits;
drivers/spi/spi-tegra210-quad.c
946
ret = tegra_qspi_start_dma_based_transfer(tqspi, t);
drivers/spi/spi-tegra210-quad.c
948
ret = tegra_qspi_start_cpu_based_transfer(tqspi, t);
drivers/spi/spi-ti-qspi.c
240
static int qspi_write_msg(struct ti_qspi *qspi, struct spi_transfer *t,
drivers/spi/spi-ti-qspi.c
248
txbuf = t->tx_buf;
drivers/spi/spi-ti-qspi.c
250
wlen = t->bits_per_word >> 3; /* in bytes */
drivers/spi/spi-ti-qspi.c
309
static int qspi_read_msg(struct ti_qspi *qspi, struct spi_transfer *t,
drivers/spi/spi-ti-qspi.c
318
rxbuf = t->rx_buf;
drivers/spi/spi-ti-qspi.c
320
switch (t->rx_nbits) {
drivers/spi/spi-ti-qspi.c
331
wlen = t->bits_per_word >> 3; /* in bytes */
drivers/spi/spi-ti-qspi.c
408
static int qspi_transfer_msg(struct ti_qspi *qspi, struct spi_transfer *t,
drivers/spi/spi-ti-qspi.c
413
if (t->tx_buf) {
drivers/spi/spi-ti-qspi.c
414
ret = qspi_write_msg(qspi, t, count);
drivers/spi/spi-ti-qspi.c
421
if (t->rx_buf) {
drivers/spi/spi-ti-qspi.c
422
ret = qspi_read_msg(qspi, t, count);
drivers/spi/spi-ti-qspi.c
668
struct spi_transfer *t;
drivers/spi/spi-ti-qspi.c
684
list_for_each_entry(t, &m->transfers, transfer_list)
drivers/spi/spi-ti-qspi.c
685
frame_len_words += t->len / (t->bits_per_word >> 3);
drivers/spi/spi-ti-qspi.c
700
list_for_each_entry(t, &m->transfers, transfer_list) {
drivers/spi/spi-ti-qspi.c
702
QSPI_WLEN(t->bits_per_word));
drivers/spi/spi-ti-qspi.c
704
wlen = t->bits_per_word >> 3;
drivers/spi/spi-ti-qspi.c
705
transfer_len_words = min(t->len / wlen, frame_len_words);
drivers/spi/spi-ti-qspi.c
707
ti_qspi_setup_clk(qspi, t->speed_hz);
drivers/spi/spi-ti-qspi.c
708
ret = qspi_transfer_msg(qspi, t, transfer_len_words * wlen);
drivers/spi/spi-uniphier.c
218
struct spi_transfer *t)
drivers/spi/spi-uniphier.c
224
priv->tx_buf = t->tx_buf;
drivers/spi/spi-uniphier.c
225
priv->rx_buf = t->rx_buf;
drivers/spi/spi-uniphier.c
226
priv->tx_bytes = priv->rx_bytes = t->len;
drivers/spi/spi-uniphier.c
234
if (!priv->is_save_param || priv->bits_per_word != t->bits_per_word) {
drivers/spi/spi-uniphier.c
235
uniphier_spi_set_transfer_size(spi, t->bits_per_word);
drivers/spi/spi-uniphier.c
236
priv->bits_per_word = t->bits_per_word;
drivers/spi/spi-uniphier.c
239
if (!priv->is_save_param || priv->speed_hz != t->speed_hz) {
drivers/spi/spi-uniphier.c
240
uniphier_spi_set_baudrate(spi, t->speed_hz);
drivers/spi/spi-uniphier.c
241
priv->speed_hz = t->speed_hz;
drivers/spi/spi-uniphier.c
351
struct spi_transfer *t)
drivers/spi/spi-uniphier.c
357
|| (!host->dma_tx && t->tx_buf)
drivers/spi/spi-uniphier.c
358
|| (!host->dma_rx && t->rx_buf))
drivers/spi/spi-uniphier.c
361
return DIV_ROUND_UP(t->len, bpw) > SSI_FIFO_DEPTH;
drivers/spi/spi-uniphier.c
390
struct spi_transfer *t)
drivers/spi/spi-uniphier.c
419
t->rx_sg.sgl, t->rx_sg.nents,
drivers/spi/spi-uniphier.c
446
t->tx_sg.sgl, t->tx_sg.nents,
drivers/spi/spi-uniphier.c
473
struct spi_transfer *t)
drivers/spi/spi-uniphier.c
500
struct spi_transfer *t)
drivers/spi/spi-uniphier.c
523
return uniphier_spi_transfer_one_irq(host, spi, t);
drivers/spi/spi-uniphier.c
528
struct spi_transfer *t)
drivers/spi/spi-uniphier.c
535
if (!t->len)
drivers/spi/spi-uniphier.c
538
uniphier_spi_setup_transfer(spi, t);
drivers/spi/spi-uniphier.c
540
use_dma = host->can_dma ? host->can_dma(host, spi, t) : false;
drivers/spi/spi-uniphier.c
542
return uniphier_spi_transfer_one_dma(host, spi, t);
drivers/spi/spi-uniphier.c
550
if (t->len > threshold)
drivers/spi/spi-uniphier.c
551
return uniphier_spi_transfer_one_irq(host, spi, t);
drivers/spi/spi-uniphier.c
553
return uniphier_spi_transfer_one_poll(host, spi, t);
drivers/spi/spi-xcomm.c
113
struct spi_device *spi, struct spi_transfer *t,
drivers/spi/spi-xcomm.c
116
if (t->len > 62)
drivers/spi/spi-xcomm.c
119
if (t->speed_hz != spi_xcomm->current_speed) {
drivers/spi/spi-xcomm.c
122
divider = DIV_ROUND_UP(SPI_XCOMM_CLOCK, t->speed_hz);
drivers/spi/spi-xcomm.c
130
spi_xcomm->current_speed = t->speed_hz;
drivers/spi/spi-xcomm.c
152
struct spi_device *spi, struct spi_transfer *t)
drivers/spi/spi-xcomm.c
156
if (t->tx_buf) {
drivers/spi/spi-xcomm.c
158
memcpy(spi_xcomm->buf + 1, t->tx_buf, t->len);
drivers/spi/spi-xcomm.c
160
ret = i2c_master_send(spi_xcomm->i2c, spi_xcomm->buf, t->len + 1);
drivers/spi/spi-xcomm.c
163
if (ret != t->len + 1)
drivers/spi/spi-xcomm.c
165
} else if (t->rx_buf) {
drivers/spi/spi-xcomm.c
166
ret = i2c_master_recv(spi_xcomm->i2c, t->rx_buf, t->len);
drivers/spi/spi-xcomm.c
169
if (ret != t->len)
drivers/spi/spi-xcomm.c
173
return t->len;
drivers/spi/spi-xcomm.c
183
struct spi_transfer *t;
drivers/spi/spi-xcomm.c
190
list_for_each_entry(t, &msg->transfers, transfer_list) {
drivers/spi/spi-xcomm.c
191
if (!t->tx_buf && !t->rx_buf && t->len) {
drivers/spi/spi-xcomm.c
196
status = spi_xcomm_setup_transfer(spi_xcomm, spi, t, &settings);
drivers/spi/spi-xcomm.c
200
is_last = list_is_last(&t->transfer_list, &msg->transfers);
drivers/spi/spi-xcomm.c
201
cs_change = t->cs_change;
drivers/spi/spi-xcomm.c
208
if (t->rx_buf) {
drivers/spi/spi-xcomm.c
210
status = spi_xcomm_sync_config(spi_xcomm, t->len);
drivers/spi/spi-xcomm.c
220
if (t->len) {
drivers/spi/spi-xcomm.c
221
status = spi_xcomm_txrx_bufs(spi_xcomm, spi, t);
drivers/spi/spi-xcomm.c
231
spi_transfer_delay_exec(t);
drivers/spi/spi-xilinx.c
226
struct spi_transfer *t)
drivers/spi/spi-xilinx.c
238
static int xilinx_spi_txrx_bufs(struct spi_device *spi, struct spi_transfer *t)
drivers/spi/spi-xilinx.c
247
xspi->tx_ptr = t->tx_buf;
drivers/spi/spi-xilinx.c
248
xspi->rx_ptr = t->rx_buf;
drivers/spi/spi-xilinx.c
249
remaining_words = t->len / xspi->bytes_per_word;
drivers/spi/spi-xilinx.c
335
return t->len;
drivers/spi/spi-xlp.c
319
static int xlp_spi_txrx_bufs(struct xlp_spi_priv *xs, struct spi_transfer *t)
drivers/spi/spi-xlp.c
325
tx_buf = t->tx_buf;
drivers/spi/spi-xlp.c
326
rx_buf = t->rx_buf;
drivers/spi/spi-xlp.c
327
bytesleft = t->len;
drivers/spi/spi-xlp.c
348
struct spi_transfer *t)
drivers/spi/spi-xlp.c
356
if (spi_transfer_is_last(host, t))
drivers/spi/spi-xlp.c
361
if (xlp_spi_txrx_bufs(xspi, t))
drivers/spi/spidev.c
107
struct spi_transfer t = {
drivers/spi/spidev.c
115
spi_message_add_tail(&t, &m);
drivers/spi/spidev.c
123
struct spi_transfer t = {
drivers/spi/spidev.c
131
spi_message_add_tail(&t, &m);
drivers/ssb/pci.c
195
static const u8 t[] = {
drivers/ssb/pci.c
229
return t[crc ^ data];
drivers/staging/fbtft/fb_ra8875.c
19
struct spi_transfer t = {
drivers/staging/fbtft/fb_ra8875.c
36
spi_message_add_tail(&t, &m);
drivers/staging/fbtft/fbtft-io.c
10
struct spi_transfer t = {
drivers/staging/fbtft/fbtft-io.c
110
t.tx_buf = txbuf;
drivers/staging/fbtft/fbtft-io.c
117
spi_message_add_tail(&t, &m);
drivers/staging/fbtft/fbtft-io.c
26
spi_message_add_tail(&t, &m);
drivers/staging/fbtft/fbtft-io.c
89
struct spi_transfer t = {
drivers/staging/media/atomisp/pci/ia_css_acc_types.h
437
#define IA_CSS_ACC_OFFSET(t, f, n) ((t)((uint8_t *)(f) + (f->header.n)))
drivers/staging/media/atomisp/pci/ia_css_isp_params.c
72
struct sh_css_isp_aa_params *t = (struct sh_css_isp_aa_params *)
drivers/staging/media/atomisp/pci/ia_css_isp_params.c
74
t->strength = params->aa_config.strength;
drivers/staging/media/atomisp/pci/runtime/debug/src/ia_css_debug.c
748
int t, n;
drivers/staging/media/atomisp/pci/runtime/debug/src/ia_css_debug.c
752
for (t = 0; t < SH_CSS_SP_DBG_NR_OF_TRACES; t++) {
drivers/staging/media/atomisp/pci/runtime/debug/src/ia_css_debug.c
753
int sp_index_last = state->index_last[t];
drivers/staging/media/atomisp/pci/runtime/debug/src/ia_css_debug.c
755
if (sp_index_last < host_index_last[t]) {
drivers/staging/media/atomisp/pci/runtime/debug/src/ia_css_debug.c
757
host_index_last[t] = 0;
drivers/staging/media/atomisp/pci/runtime/debug/src/ia_css_debug.c
760
if ((host_index_last[t] + SH_CSS_SP_DBG_TRACE_DEPTH) <
drivers/staging/media/atomisp/pci/runtime/debug/src/ia_css_debug.c
766
trace_name[t],
drivers/staging/media/atomisp/pci/runtime/debug/src/ia_css_debug.c
768
(host_index_last[t] +
drivers/staging/media/atomisp/pci/runtime/debug/src/ia_css_debug.c
771
host_index_last[t] =
drivers/staging/media/atomisp/pci/runtime/debug/src/ia_css_debug.c
775
for (n = host_index_last[t]; n < sp_index_last; n++) {
drivers/staging/media/atomisp/pci/runtime/debug/src/ia_css_debug.c
777
int l = state->trace[t][i].location &
drivers/staging/media/atomisp/pci/runtime/debug/src/ia_css_debug.c
779
int fid = state->trace[t][i].location >>
drivers/staging/media/atomisp/pci/runtime/debug/src/ia_css_debug.c
781
int ts = state->trace[t][i].time_stamp;
drivers/staging/media/atomisp/pci/runtime/debug/src/ia_css_debug.c
787
trace_name[t],
drivers/staging/media/atomisp/pci/runtime/debug/src/ia_css_debug.c
789
state->trace[t][i].data);
drivers/staging/media/atomisp/pci/runtime/debug/src/ia_css_debug.c
792
host_index_last[t] = sp_index_last;
drivers/staging/media/av7110/av7110.c
1221
static void vpeirq(struct tasklet_struct *t)
drivers/staging/media/av7110/av7110.c
1223
struct av7110 *budget = from_tasklet(budget, t, vpe_tasklet);
drivers/staging/media/av7110/av7110.c
349
static void debiirq(struct tasklet_struct *t)
drivers/staging/media/av7110/av7110.c
351
struct av7110 *av7110 = from_tasklet(av7110, t, debi_tasklet);
drivers/staging/media/av7110/av7110.c
450
static void gpioirq(struct tasklet_struct *t)
drivers/staging/media/av7110/av7110.c
452
struct av7110 *av7110 = from_tasklet(av7110, t, gpio_tasklet);
drivers/staging/media/av7110/av7110_v4l.c
293
static int vidioc_g_tuner(struct file *file, void *fh, struct v4l2_tuner *t)
drivers/staging/media/av7110/av7110_v4l.c
300
dprintk(2, "VIDIOC_G_TUNER: %d\n", t->index);
drivers/staging/media/av7110/av7110_v4l.c
302
if (!av7110->analog_tuner_flags || t->index != 0)
drivers/staging/media/av7110/av7110_v4l.c
305
memset(t, 0, sizeof(*t));
drivers/staging/media/av7110/av7110_v4l.c
306
strscpy((char *)t->name, "Television", sizeof(t->name));
drivers/staging/media/av7110/av7110_v4l.c
308
t->type = V4L2_TUNER_ANALOG_TV;
drivers/staging/media/av7110/av7110_v4l.c
309
t->capability = V4L2_TUNER_CAP_NORM | V4L2_TUNER_CAP_STEREO |
drivers/staging/media/av7110/av7110_v4l.c
311
t->rangelow = 772; /* 48.25 MHZ / 62.5 kHz = 772, see fi1216mk2-specs, page 2 */
drivers/staging/media/av7110/av7110_v4l.c
312
t->rangehigh = 13684; /* 855.25 MHz / 62.5 kHz = 13684 */
drivers/staging/media/av7110/av7110_v4l.c
314
t->signal = 0xffff;
drivers/staging/media/av7110/av7110_v4l.c
315
t->afc = 0;
drivers/staging/media/av7110/av7110_v4l.c
325
t->rxsubchans = V4L2_TUNER_SUB_STEREO | V4L2_TUNER_SUB_MONO;
drivers/staging/media/av7110/av7110_v4l.c
326
t->audmode = V4L2_TUNER_MODE_STEREO;
drivers/staging/media/av7110/av7110_v4l.c
329
t->rxsubchans = V4L2_TUNER_SUB_LANG1 | V4L2_TUNER_SUB_LANG2;
drivers/staging/media/av7110/av7110_v4l.c
330
t->audmode = V4L2_TUNER_MODE_LANG1;
drivers/staging/media/av7110/av7110_v4l.c
333
t->rxsubchans = V4L2_TUNER_SUB_MONO;
drivers/staging/media/av7110/av7110_v4l.c
339
static int vidioc_s_tuner(struct file *file, void *fh, const struct v4l2_tuner *t)
drivers/staging/media/av7110/av7110_v4l.c
345
dprintk(2, "VIDIOC_S_TUNER: %d\n", t->index);
drivers/staging/media/av7110/av7110_v4l.c
350
switch (t->audmode) {
drivers/staging/media/imx/imx-ic-prpencvf.c
296
static void prp_eof_timeout(struct timer_list *t)
drivers/staging/media/imx/imx-ic-prpencvf.c
298
struct prp_priv *priv = timer_container_of(priv, t, eof_timeout_timer);
drivers/staging/media/imx/imx-media-csi.c
357
static void csi_idmac_eof_timeout(struct timer_list *t)
drivers/staging/media/imx/imx-media-csi.c
359
struct csi_priv *priv = timer_container_of(priv, t, eof_timeout_timer);
drivers/staging/media/starfive/camss/stf-capture.c
361
struct stfcamss_buffer *t;
drivers/staging/media/starfive/camss/stf-capture.c
363
list_for_each_entry_safe(buf, t, &output->pending_bufs, queue) {
drivers/staging/media/starfive/camss/stf-capture.c
367
list_for_each_entry_safe(buf, t, &output->ready_bufs, queue) {
drivers/staging/media/sunxi/cedrus/cedrus_regs.h
95
#define VE_DEC_MPEG_MP12HDR_SLICE_TYPE(t) SHIFT_AND_MASK_BITS(t, 30, 28)
drivers/staging/octeon/ethernet-tx.c
105
struct sk_buff *t = to_free_list;
drivers/staging/octeon/ethernet-tx.c
108
dev_kfree_skb_any(t);
drivers/staging/octeon/ethernet-tx.c
457
struct sk_buff *t = __skb_dequeue(&priv->tx_free_list[qos]);
drivers/staging/octeon/ethernet-tx.c
459
t->next = to_free_list;
drivers/staging/octeon/ethernet-tx.c
460
to_free_list = t;
drivers/staging/octeon/ethernet-tx.c
468
struct sk_buff *t = to_free_list;
drivers/staging/octeon/ethernet-tx.c
471
dev_kfree_skb_any(t);
drivers/staging/octeon/ethernet-tx.c
94
struct sk_buff *t;
drivers/staging/octeon/ethernet-tx.c
96
t = __skb_dequeue(&priv->tx_free_list[qos]);
drivers/staging/octeon/ethernet-tx.c
97
t->next = to_free_list;
drivers/staging/octeon/ethernet-tx.c
98
to_free_list = t;
drivers/staging/rtl8723bs/core/rtw_mlme.c
12
static void _dynamic_check_timer_handler(struct timer_list *t)
drivers/staging/rtl8723bs/core/rtw_mlme.c
15
timer_container_of(adapter, t, mlmepriv.dynamic_chk_timer);
drivers/staging/rtl8723bs/core/rtw_mlme.c
1531
void _rtw_join_timeout_handler(struct timer_list *t)
drivers/staging/rtl8723bs/core/rtw_mlme.c
1533
struct adapter *adapter = timer_container_of(adapter, t,
drivers/staging/rtl8723bs/core/rtw_mlme.c
1574
void rtw_scan_timeout_handler(struct timer_list *t)
drivers/staging/rtl8723bs/core/rtw_mlme.c
1576
struct adapter *adapter = timer_container_of(adapter, t,
drivers/staging/rtl8723bs/core/rtw_mlme.c
22
static void _rtw_set_scan_deny_timer_hdl(struct timer_list *t)
drivers/staging/rtl8723bs/core/rtw_mlme.c
25
timer_container_of(adapter, t, mlmepriv.set_scan_deny_timer);
drivers/staging/rtl8723bs/core/rtw_mlme_ext.c
5057
void survey_timer_hdl(struct timer_list *t)
drivers/staging/rtl8723bs/core/rtw_mlme_ext.c
5060
timer_container_of(padapter, t, mlmeextpriv.survey_timer);
drivers/staging/rtl8723bs/core/rtw_mlme_ext.c
5092
void link_timer_hdl(struct timer_list *t)
drivers/staging/rtl8723bs/core/rtw_mlme_ext.c
5095
timer_container_of(padapter, t, mlmeextpriv.link_timer);
drivers/staging/rtl8723bs/core/rtw_mlme_ext.c
5127
void addba_timer_hdl(struct timer_list *t)
drivers/staging/rtl8723bs/core/rtw_mlme_ext.c
5129
struct sta_info *psta = timer_container_of(psta, t, addba_retry_timer);
drivers/staging/rtl8723bs/core/rtw_mlme_ext.c
5144
void sa_query_timer_hdl(struct timer_list *t)
drivers/staging/rtl8723bs/core/rtw_mlme_ext.c
5147
timer_container_of(padapter, t, mlmeextpriv.sa_query_timer);
drivers/staging/rtl8723bs/core/rtw_pwrctrl.c
171
static void pwr_state_check_handler(struct timer_list *t)
drivers/staging/rtl8723bs/core/rtw_pwrctrl.c
174
timer_container_of(pwrctrlpriv, t, pwr_state_check_timer);
drivers/staging/rtl8723bs/core/rtw_pwrctrl.c
665
static void pwr_rpwm_timeout_handler(struct timer_list *t)
drivers/staging/rtl8723bs/core/rtw_pwrctrl.c
667
struct pwrctrl_priv *pwrpriv = timer_container_of(pwrpriv, t,
drivers/staging/rtl8723bs/core/rtw_recv.c
16
static void rtw_signal_stat_timer_hdl(struct timer_list *t);
drivers/staging/rtl8723bs/core/rtw_recv.c
2065
void rtw_reordering_ctrl_timeout_handler(struct timer_list *t)
drivers/staging/rtl8723bs/core/rtw_recv.c
2068
timer_container_of(preorder_ctrl, t, reordering_ctrl_timer);
drivers/staging/rtl8723bs/core/rtw_recv.c
2244
static void rtw_signal_stat_timer_hdl(struct timer_list *t)
drivers/staging/rtl8723bs/core/rtw_recv.c
2247
timer_container_of(adapter, t, recvpriv.signal_stat_timer);
drivers/staging/rtl8723bs/hal/HalPhyRf_8723B.c
1302
u8 t,
drivers/staging/rtl8723bs/hal/HalPhyRf_8723B.c
1358
if (t == 0) {
drivers/staging/rtl8723bs/hal/HalPhyRf_8723B.c
1410
result[t][0] = (PHY_QueryBBReg(pDM_Odm->Adapter, rTx_Power_Before_IQK_A, bMaskDWord)&0x3FF0000)>>16;
drivers/staging/rtl8723bs/hal/HalPhyRf_8723B.c
1411
result[t][1] = (PHY_QueryBBReg(pDM_Odm->Adapter, rTx_Power_After_IQK_A, bMaskDWord)&0x3FF0000)>>16;
drivers/staging/rtl8723bs/hal/HalPhyRf_8723B.c
1422
result[t][2] = (PHY_QueryBBReg(pDM_Odm->Adapter, rRx_Power_Before_IQK_A_2, bMaskDWord)&0x3FF0000)>>16;
drivers/staging/rtl8723bs/hal/HalPhyRf_8723B.c
1423
result[t][3] = (PHY_QueryBBReg(pDM_Odm->Adapter, rRx_Power_After_IQK_A_2, bMaskDWord)&0x3FF0000)>>16;
drivers/staging/rtl8723bs/hal/HalPhyRf_8723B.c
1442
result[t][4] = (PHY_QueryBBReg(pDM_Odm->Adapter, rTx_Power_Before_IQK_A, bMaskDWord)&0x3FF0000)>>16;
drivers/staging/rtl8723bs/hal/HalPhyRf_8723B.c
1443
result[t][5] = (PHY_QueryBBReg(pDM_Odm->Adapter, rTx_Power_After_IQK_A, bMaskDWord)&0x3FF0000)>>16;
drivers/staging/rtl8723bs/hal/HalPhyRf_8723B.c
1454
result[t][6] = (PHY_QueryBBReg(pDM_Odm->Adapter, rRx_Power_Before_IQK_A_2, bMaskDWord)&0x3FF0000)>>16;
drivers/staging/rtl8723bs/hal/HalPhyRf_8723B.c
1455
result[t][7] = (PHY_QueryBBReg(pDM_Odm->Adapter, rRx_Power_After_IQK_A_2, bMaskDWord)&0x3FF0000)>>16;
drivers/staging/rtl8723bs/hal/HalPhyRf_8723B.c
1466
if (t != 0) {
drivers/staging/rtl8723bs/hal/rtl8723bs_recv.c
216
static void rtl8723bs_recv_tasklet(struct tasklet_struct *t)
drivers/staging/rtl8723bs/hal/rtl8723bs_recv.c
218
struct adapter *padapter = from_tasklet(padapter, t,
drivers/staging/rtl8723bs/include/rtw_mlme.h
263
extern void rtw_join_timeout_handler(struct timer_list *t);
drivers/staging/rtl8723bs/include/rtw_mlme.h
264
extern void _rtw_scan_timeout_handler(struct timer_list *t);
drivers/staging/rtl8723bs/include/rtw_mlme.h
340
extern void _rtw_join_timeout_handler(struct timer_list *t);
drivers/staging/rtl8723bs/include/rtw_mlme.h
341
extern void rtw_scan_timeout_handler(struct timer_list *t);
drivers/staging/rtl8723bs/include/rtw_mlme_ext.h
592
void survey_timer_hdl(struct timer_list *t);
drivers/staging/rtl8723bs/include/rtw_mlme_ext.h
593
void link_timer_hdl(struct timer_list *t);
drivers/staging/rtl8723bs/include/rtw_mlme_ext.h
594
void addba_timer_hdl(struct timer_list *t);
drivers/staging/rtl8723bs/include/rtw_mlme_ext.h
595
void sa_query_timer_hdl(struct timer_list *t);
drivers/staging/rtl8723bs/include/rtw_recv.h
343
void rtw_reordering_ctrl_timeout_handler(struct timer_list *t);
drivers/target/iscsi/cxgbit/cxgbit_cm.c
1223
struct tid_info *t = cdev->lldi.tids;
drivers/target/iscsi/cxgbit/cxgbit_cm.c
1237
cnp = lookup_stid(t, stid);
drivers/target/iscsi/cxgbit/cxgbit_cm.c
1250
csk = lookup_tid(t, tid);
drivers/target/iscsi/cxgbit/cxgbit_cm.c
1378
cxgb4_insert_tid(t, csk, tid, csk->com.local_addr.ss_family);
drivers/target/iscsi/cxgbit/cxgbit_cm.c
1582
struct tid_info *t = cdev->lldi.tids;
drivers/target/iscsi/cxgbit/cxgbit_cm.c
1584
struct cxgbit_np *cnp = lookup_stid(t, stid);
drivers/target/iscsi/cxgbit/cxgbit_cm.c
1604
struct tid_info *t = cdev->lldi.tids;
drivers/target/iscsi/cxgbit/cxgbit_cm.c
1606
struct cxgbit_np *cnp = lookup_stid(t, stid);
drivers/target/iscsi/cxgbit/cxgbit_cm.c
1626
struct tid_info *t = cdev->lldi.tids;
drivers/target/iscsi/cxgbit/cxgbit_cm.c
1634
csk = lookup_tid(t, tid);
drivers/target/iscsi/cxgbit/cxgbit_cm.c
1903
struct tid_info *t = lldi->tids;
drivers/target/iscsi/cxgbit/cxgbit_cm.c
1905
csk = lookup_tid(t, tid);
drivers/target/iscsi/cxgbit/cxgbit_cm.c
1924
struct tid_info *t = lldi->tids;
drivers/target/iscsi/cxgbit/cxgbit_cm.c
1926
csk = lookup_tid(t, tid);
drivers/target/iscsi/cxgbit/cxgbit_cm.c
1964
struct tid_info *t = lldi->tids;
drivers/target/iscsi/cxgbit/cxgbit_cm.c
1990
csk = lookup_tid(t, tid);
drivers/target/iscsi/iscsi_target.c
3714
struct iscsit_transport *t = conn->conn_transport;
drivers/target/iscsi/iscsi_target.c
3726
ret = t->iscsit_immediate_queue(conn, cmd, state);
drivers/target/iscsi/iscsi_target.c
3843
struct iscsit_transport *t = conn->conn_transport;
drivers/target/iscsi/iscsi_target.c
3854
ret = t->iscsit_response_queue(conn, cmd, state);
drivers/target/iscsi/iscsi_target_erl0.c
744
void iscsit_handle_time2retain_timeout(struct timer_list *t)
drivers/target/iscsi/iscsi_target_erl0.c
746
struct iscsit_session *sess = timer_container_of(sess, t,
drivers/target/iscsi/iscsi_target_erl0.h
15
extern void iscsit_handle_time2retain_timeout(struct timer_list *t);
drivers/target/iscsi/iscsi_target_erl1.c
1093
void iscsit_handle_dataout_timeout(struct timer_list *t)
drivers/target/iscsi/iscsi_target_erl1.c
1097
struct iscsit_cmd *cmd = timer_container_of(cmd, t, dataout_timer);
drivers/target/iscsi/iscsi_target_erl1.h
33
extern void iscsit_handle_dataout_timeout(struct timer_list *t);
drivers/target/iscsi/iscsi_target_login.c
848
struct iscsit_transport *t;
drivers/target/iscsi/iscsi_target_login.c
851
t = iscsit_get_transport(np->np_network_transport);
drivers/target/iscsi/iscsi_target_login.c
852
if (!t)
drivers/target/iscsi/iscsi_target_login.c
855
rc = t->iscsit_setup_np(np, sockaddr);
drivers/target/iscsi/iscsi_target_login.c
857
iscsit_put_transport(t);
drivers/target/iscsi/iscsi_target_login.c
861
np->np_transport = t;
drivers/target/iscsi/iscsi_target_login.c
982
iscsit_conn_set_transport(struct iscsit_conn *conn, struct iscsit_transport *t)
drivers/target/iscsi/iscsi_target_login.c
986
if (!t->owner) {
drivers/target/iscsi/iscsi_target_login.c
987
conn->conn_transport = t;
drivers/target/iscsi/iscsi_target_login.c
991
rc = try_module_get(t->owner);
drivers/target/iscsi/iscsi_target_login.c
993
pr_err("try_module_get() failed for %s\n", t->name);
drivers/target/iscsi/iscsi_target_login.c
997
conn->conn_transport = t;
drivers/target/iscsi/iscsi_target_transport.c
12
struct iscsit_transport *t;
drivers/target/iscsi/iscsi_target_transport.c
15
list_for_each_entry(t, &g_transport_list, t_node) {
drivers/target/iscsi/iscsi_target_transport.c
16
if (t->transport_type == type) {
drivers/target/iscsi/iscsi_target_transport.c
17
if (t->owner && !try_module_get(t->owner)) {
drivers/target/iscsi/iscsi_target_transport.c
18
t = NULL;
drivers/target/iscsi/iscsi_target_transport.c
21
return t;
drivers/target/iscsi/iscsi_target_transport.c
29
void iscsit_put_transport(struct iscsit_transport *t)
drivers/target/iscsi/iscsi_target_transport.c
31
module_put(t->owner);
drivers/target/iscsi/iscsi_target_transport.c
34
void iscsit_register_transport(struct iscsit_transport *t)
drivers/target/iscsi/iscsi_target_transport.c
36
INIT_LIST_HEAD(&t->t_node);
drivers/target/iscsi/iscsi_target_transport.c
39
list_add_tail(&t->t_node, &g_transport_list);
drivers/target/iscsi/iscsi_target_transport.c
42
pr_debug("Registered iSCSI transport: %s\n", t->name);
drivers/target/iscsi/iscsi_target_transport.c
46
void iscsit_unregister_transport(struct iscsit_transport *t)
drivers/target/iscsi/iscsi_target_transport.c
49
list_del(&t->t_node);
drivers/target/iscsi/iscsi_target_transport.c
52
pr_debug("Unregistered iSCSI transport: %s\n", t->name);
drivers/target/iscsi/iscsi_target_util.c
1006
void iscsit_login_timeout(struct timer_list *t)
drivers/target/iscsi/iscsi_target_util.c
1008
struct iscsit_conn *conn = timer_container_of(conn, t, login_timer);
drivers/target/iscsi/iscsi_target_util.c
858
void iscsit_handle_nopin_response_timeout(struct timer_list *t)
drivers/target/iscsi/iscsi_target_util.c
860
struct iscsit_conn *conn = timer_container_of(conn, t,
drivers/target/iscsi/iscsi_target_util.c
939
void iscsit_handle_nopin_timeout(struct timer_list *t)
drivers/target/iscsi/iscsi_target_util.c
941
struct iscsit_conn *conn = timer_container_of(conn, t, nopin_timer);
drivers/target/iscsi/iscsi_target_util.h
48
extern void iscsit_handle_nopin_response_timeout(struct timer_list *t);
drivers/target/iscsi/iscsi_target_util.h
52
extern void iscsit_handle_nopin_timeout(struct timer_list *t);
drivers/target/iscsi/iscsi_target_util.h
56
extern void iscsit_login_timeout(struct timer_list *t);
drivers/target/target_core_configfs.c
3045
struct t10_alua_tg_pt_gp *t = to_tg_pt_gp(item); \
drivers/target/target_core_configfs.c
3047
!!(t->tg_pt_gp_alua_supported_states & _bit)); \
drivers/target/target_core_configfs.c
3053
struct t10_alua_tg_pt_gp *t = to_tg_pt_gp(item); \
drivers/target/target_core_configfs.c
3057
if (!t->tg_pt_gp_valid_id) { \
drivers/target/target_core_configfs.c
3072
t->tg_pt_gp_alua_supported_states |= _bit; \
drivers/target/target_core_configfs.c
3074
t->tg_pt_gp_alua_supported_states &= ~_bit; \
drivers/target/target_core_configfs.c
504
struct target_fabric_configfs *t;
drivers/target/target_core_configfs.c
507
list_for_each_entry(t, &g_tf_list, tf_list) {
drivers/target/target_core_configfs.c
508
if (!strcmp(t->tf_ops->fabric_name, fo->fabric_name)) {
drivers/target/target_core_configfs.c
509
BUG_ON(atomic_read(&t->tf_access_cnt));
drivers/target/target_core_configfs.c
510
list_del(&t->tf_list);
drivers/target/target_core_configfs.c
518
kfree(t->tf_tpg_base_cit.ct_attrs);
drivers/target/target_core_configfs.c
519
kfree(t->tf_ops);
drivers/target/target_core_configfs.c
520
kfree(t);
drivers/target/target_core_transport.c
3088
static const char *cmd_state_name(enum transport_state_table t)
drivers/target/target_core_transport.c
3090
switch (t) {
drivers/target/target_core_user.c
1565
static void tcmu_cmd_timedout(struct timer_list *t)
drivers/target/target_core_user.c
1567
struct tcmu_dev *udev = timer_container_of(udev, t, cmd_timer);
drivers/target/target_core_user.c
1573
static void tcmu_qfull_timedout(struct timer_list *t)
drivers/target/target_core_user.c
1575
struct tcmu_dev *udev = timer_container_of(udev, t, qfull_timer);
drivers/tee/amdtee/amdtee_if.h
62
#define TEE_PARAM_TYPE_GET(t, i) (((t) >> ((i) * 4)) & 0xF)
drivers/tee/qcomtee/qcomtee_msg.h
165
#define qcomtee_msg_buffer_args(t, n) \
drivers/tee/qcomtee/qcomtee_msg.h
166
qcomtee_msg_offset_align(struct_size_t(t, args, n))
drivers/thermal/broadcom/brcmstb_thermal.c
157
long t;
drivers/thermal/broadcom/brcmstb_thermal.c
166
t = avs_tmon_code_to_temp(priv, val);
drivers/thermal/broadcom/brcmstb_thermal.c
167
if (t < 0)
drivers/thermal/broadcom/brcmstb_thermal.c
170
*temp = t;
drivers/thermal/da9062-thermal.c
42
#define DA9062_MILLI_CELSIUS(t) ((t) * 1000)
drivers/thermal/tegra/soctherm.c
420
int t;
drivers/thermal/tegra/soctherm.c
422
t = ((val & READBACK_VALUE_MASK) >> READBACK_VALUE_SHIFT) * 1000;
drivers/thermal/tegra/soctherm.c
424
t += 500;
drivers/thermal/tegra/soctherm.c
426
t *= -1;
drivers/thermal/tegra/soctherm.c
428
return t;
drivers/thermal/thermal_mmio.c
25
int t;
drivers/thermal/thermal_mmio.c
28
t = sensor->read_mmio(sensor->mmio_base) & sensor->mask;
drivers/thermal/thermal_mmio.c
29
t *= sensor->factor;
drivers/thermal/thermal_mmio.c
31
*temp = t;
drivers/thermal/thermal_of.c
47
const char *t;
drivers/thermal/thermal_of.c
50
err = of_property_read_string(np, "type", &t);
drivers/thermal/thermal_of.c
55
if (!strcasecmp(t, trip_types[i])) {
drivers/thermal/thermal_thresholds.c
100
if (temperature <= t->temperature &&
drivers/thermal/thermal_thresholds.c
101
last_temperature > t->temperature)
drivers/thermal/thermal_thresholds.c
111
struct user_threshold *t;
drivers/thermal/thermal_thresholds.c
113
list_for_each_entry(t, thresholds, list_node) {
drivers/thermal/thermal_thresholds.c
114
if (temperature < t->temperature &&
drivers/thermal/thermal_thresholds.c
115
(t->direction & THERMAL_THRESHOLD_WAY_UP) &&
drivers/thermal/thermal_thresholds.c
116
*high > t->temperature)
drivers/thermal/thermal_thresholds.c
117
*high = t->temperature;
drivers/thermal/thermal_thresholds.c
120
list_for_each_entry_reverse(t, thresholds, list_node) {
drivers/thermal/thermal_thresholds.c
121
if (temperature > t->temperature &&
drivers/thermal/thermal_thresholds.c
122
(t->direction & THERMAL_THRESHOLD_WAY_DOWN) &&
drivers/thermal/thermal_thresholds.c
123
*low < t->temperature)
drivers/thermal/thermal_thresholds.c
124
*low = t->temperature;
drivers/thermal/thermal_thresholds.c
172
struct user_threshold *t;
drivers/thermal/thermal_thresholds.c
176
t = __thermal_thresholds_find(thresholds, temperature);
drivers/thermal/thermal_thresholds.c
177
if (t) {
drivers/thermal/thermal_thresholds.c
178
if (t->direction == direction)
drivers/thermal/thermal_thresholds.c
181
t->direction |= direction;
drivers/thermal/thermal_thresholds.c
184
t = kmalloc_obj(*t);
drivers/thermal/thermal_thresholds.c
185
if (!t)
drivers/thermal/thermal_thresholds.c
188
INIT_LIST_HEAD(&t->list_node);
drivers/thermal/thermal_thresholds.c
189
t->temperature = temperature;
drivers/thermal/thermal_thresholds.c
190
t->direction = direction;
drivers/thermal/thermal_thresholds.c
191
list_add(&t->list_node, thresholds);
drivers/thermal/thermal_thresholds.c
206
struct user_threshold *t;
drivers/thermal/thermal_thresholds.c
210
t = __thermal_thresholds_find(thresholds, temperature);
drivers/thermal/thermal_thresholds.c
211
if (!t)
drivers/thermal/thermal_thresholds.c
214
if (t->direction == direction) {
drivers/thermal/thermal_thresholds.c
215
list_del(&t->list_node);
drivers/thermal/thermal_thresholds.c
216
kfree(t);
drivers/thermal/thermal_thresholds.c
218
t->direction &= ~direction;
drivers/thermal/thermal_thresholds.c
63
struct user_threshold *t;
drivers/thermal/thermal_thresholds.c
65
list_for_each_entry(t, thresholds, list_node)
drivers/thermal/thermal_thresholds.c
66
if (t->temperature == temperature)
drivers/thermal/thermal_thresholds.c
67
return t;
drivers/thermal/thermal_thresholds.c
75
struct user_threshold *t;
drivers/thermal/thermal_thresholds.c
77
list_for_each_entry(t, thresholds, list_node) {
drivers/thermal/thermal_thresholds.c
79
if (!(t->direction & THERMAL_THRESHOLD_WAY_UP))
drivers/thermal/thermal_thresholds.c
82
if (temperature >= t->temperature &&
drivers/thermal/thermal_thresholds.c
83
last_temperature < t->temperature)
drivers/thermal/thermal_thresholds.c
93
struct user_threshold *t;
drivers/thermal/thermal_thresholds.c
95
list_for_each_entry_reverse(t, thresholds, list_node) {
drivers/thermal/thermal_thresholds.c
97
if (!(t->direction & THERMAL_THRESHOLD_WAY_DOWN))
drivers/thermal/ti-soc-thermal/ti-bandgap.c
292
int ti_bandgap_adc_to_mcelsius(struct ti_bandgap *bgp, int adc_val, int *t)
drivers/thermal/ti-soc-thermal/ti-bandgap.c
300
*t = bgp->conf->conv_table[adc_val - conf->adc_start_val];
drivers/thermal/ti-soc-thermal/ti-bandgap.c
80
struct temp_sensor_registers *t; \
drivers/thermal/ti-soc-thermal/ti-bandgap.c
83
t = bgp->conf->sensors[(id)].registers; \
drivers/thermal/ti-soc-thermal/ti-bandgap.c
84
r = ti_bandgap_readl(bgp, t->reg); \
drivers/thermal/ti-soc-thermal/ti-bandgap.c
85
r &= ~t->mask; \
drivers/thermal/ti-soc-thermal/ti-bandgap.c
86
r |= (val) << __ffs(t->mask); \
drivers/thermal/ti-soc-thermal/ti-bandgap.c
87
ti_bandgap_writel(bgp, r, t->reg); \
drivers/thermal/ti-soc-thermal/ti-thermal-common.c
58
static inline int ti_thermal_hotspot_temperature(int t, int s, int c)
drivers/thermal/ti-soc-thermal/ti-thermal-common.c
60
int delta = t * s / 1000 + c;
drivers/thermal/ti-soc-thermal/ti-thermal-common.c
65
return t + delta;
drivers/tty/ipwireless/hardware.c
1009
static void ipwireless_do_tasklet(struct tasklet_struct *t)
drivers/tty/ipwireless/hardware.c
1011
struct ipw_hardware *hw = from_tasklet(hw, t, tasklet);
drivers/tty/ipwireless/hardware.c
1677
static void ipwireless_setup_timer(struct timer_list *t)
drivers/tty/ipwireless/hardware.c
1679
struct ipw_hardware *hw = timer_container_of(hw, t, setup_timer);
drivers/tty/ipwireless/hardware.c
36
static void ipwireless_setup_timer(struct timer_list *t);
drivers/tty/mips_ejtag_fdc.c
684
static void mips_ejtag_fdc_tty_timer(struct timer_list *t)
drivers/tty/mips_ejtag_fdc.c
686
struct mips_ejtag_fdc_tty *priv = timer_container_of(priv, t,
drivers/tty/n_gsm.c
1973
static void gsm_control_keep_alive(struct timer_list *t)
drivers/tty/n_gsm.c
1975
struct gsm_mux *gsm = timer_container_of(gsm, t, ka_timer);
drivers/tty/n_gsm.c
2030
static void gsm_control_retransmit(struct timer_list *t)
drivers/tty/n_gsm.c
2032
struct gsm_mux *gsm = timer_container_of(gsm, t, t2_timer);
drivers/tty/n_gsm.c
2230
static void gsm_dlci_t1(struct timer_list *t)
drivers/tty/n_gsm.c
2232
struct gsm_dlci *dlci = timer_container_of(dlci, t, t1);
drivers/tty/n_gsm.c
2490
static void gsm_kick_timer(struct timer_list *t)
drivers/tty/n_gsm.c
2492
struct gsm_mux *gsm = timer_container_of(gsm, t, kick_timer);
drivers/tty/serial/8250/8250_bcm7271.c
839
static enum hrtimer_restart brcmuart_hrtimer_func(struct hrtimer *t)
drivers/tty/serial/8250/8250_bcm7271.c
841
struct brcmuart_priv *priv = container_of(t, struct brcmuart_priv, hrt);
drivers/tty/serial/8250/8250_core.c
209
static void serial8250_timeout(struct timer_list *t)
drivers/tty/serial/8250/8250_core.c
211
struct uart_8250_port *up = timer_container_of(up, t, timer);
drivers/tty/serial/8250/8250_core.c
217
static void serial8250_backup_timeout(struct timer_list *t)
drivers/tty/serial/8250/8250_core.c
219
struct uart_8250_port *up = timer_container_of(up, t, timer);
drivers/tty/serial/8250/8250_port.c
1316
static enum hrtimer_restart serial8250_em485_handle_stop_tx(struct hrtimer *t)
drivers/tty/serial/8250/8250_port.c
1318
struct uart_8250_em485 *em485 = container_of(t, struct uart_8250_em485,
drivers/tty/serial/8250/8250_port.c
1503
static enum hrtimer_restart serial8250_em485_handle_start_tx(struct hrtimer *t)
drivers/tty/serial/8250/8250_port.c
1505
struct uart_8250_em485 *em485 = container_of(t, struct uart_8250_em485,
drivers/tty/serial/8250/8250_port.c
503
static enum hrtimer_restart serial8250_em485_handle_start_tx(struct hrtimer *t);
drivers/tty/serial/8250/8250_port.c
504
static enum hrtimer_restart serial8250_em485_handle_stop_tx(struct hrtimer *t);
drivers/tty/serial/altera_uart.c
276
static void altera_uart_timer(struct timer_list *t)
drivers/tty/serial/altera_uart.c
278
struct altera_uart *pp = timer_container_of(pp, t, tmr);
drivers/tty/serial/amba-pl011.c
1047
static void pl011_dma_rx_poll(struct timer_list *t)
drivers/tty/serial/amba-pl011.c
1049
struct uart_amba_port *uap = timer_container_of(uap, t, dmarx.timer);
drivers/tty/serial/amba-pl011.c
1399
static enum hrtimer_restart pl011_trigger_start_tx(struct hrtimer *t)
drivers/tty/serial/amba-pl011.c
1402
container_of(t, struct uart_amba_port, trigger_start_tx);
drivers/tty/serial/amba-pl011.c
1413
static enum hrtimer_restart pl011_trigger_stop_tx(struct hrtimer *t)
drivers/tty/serial/amba-pl011.c
1416
container_of(t, struct uart_amba_port, trigger_stop_tx);
drivers/tty/serial/ar933x_uart.c
240
u64 t;
drivers/tty/serial/ar933x_uart.c
244
t = clk;
drivers/tty/serial/ar933x_uart.c
245
t *= step;
drivers/tty/serial/ar933x_uart.c
246
t += (div / 2);
drivers/tty/serial/ar933x_uart.c
247
do_div(t, div);
drivers/tty/serial/ar933x_uart.c
249
return t;
drivers/tty/serial/ar933x_uart.c
74
unsigned int t;
drivers/tty/serial/ar933x_uart.c
76
t = ar933x_uart_read(up, offset);
drivers/tty/serial/ar933x_uart.c
77
t &= ~mask;
drivers/tty/serial/ar933x_uart.c
78
t |= val;
drivers/tty/serial/ar933x_uart.c
79
ar933x_uart_write(up, offset, t);
drivers/tty/serial/atmel_serial.c
1267
static void atmel_uart_timer_callback(struct timer_list *t)
drivers/tty/serial/atmel_serial.c
1269
struct atmel_uart_port *atmel_port = timer_container_of(atmel_port, t,
drivers/tty/serial/atmel_serial.c
1699
static void atmel_tasklet_rx_func(struct tasklet_struct *t)
drivers/tty/serial/atmel_serial.c
1701
struct atmel_uart_port *atmel_port = from_tasklet(atmel_port, t,
drivers/tty/serial/atmel_serial.c
1711
static void atmel_tasklet_tx_func(struct tasklet_struct *t)
drivers/tty/serial/atmel_serial.c
1713
struct atmel_uart_port *atmel_port = from_tasklet(atmel_port, t,
drivers/tty/serial/atmel_serial.c
288
struct tasklet_struct *t)
drivers/tty/serial/atmel_serial.c
291
tasklet_schedule(t);
drivers/tty/serial/fsl_lpuart.c
1308
static void lpuart_timer_func(struct timer_list *t)
drivers/tty/serial/fsl_lpuart.c
1310
struct lpuart_port *sport = timer_container_of(sport, t, lpuart_timer);
drivers/tty/serial/imx.c
1167
static void imx_uart_timeout(struct timer_list *t)
drivers/tty/serial/imx.c
1169
struct imx_port *sport = timer_container_of(sport, t, timer);
drivers/tty/serial/imx.c
2371
static enum hrtimer_restart imx_trigger_start_tx(struct hrtimer *t)
drivers/tty/serial/imx.c
2373
struct imx_port *sport = container_of(t, struct imx_port, trigger_start_tx);
drivers/tty/serial/imx.c
2384
static enum hrtimer_restart imx_trigger_stop_tx(struct hrtimer *t)
drivers/tty/serial/imx.c
2386
struct imx_port *sport = container_of(t, struct imx_port, trigger_stop_tx);
drivers/tty/serial/liteuart.c
153
static void liteuart_timer(struct timer_list *t)
drivers/tty/serial/liteuart.c
155
struct liteuart_port *uart = timer_container_of(uart, t, timer);
drivers/tty/serial/max3100.c
310
static void max3100_timeout(struct timer_list *t)
drivers/tty/serial/max3100.c
312
struct max3100_port *s = timer_container_of(s, t, timer);
drivers/tty/serial/pmac_zilog.c
1061
int t, version;
drivers/tty/serial/pmac_zilog.c
1102
t = 10000;
drivers/tty/serial/pmac_zilog.c
1105
if (--t <= 0) {
drivers/tty/serial/pmac_zilog.c
1113
t = 100;
drivers/tty/serial/pmac_zilog.c
1121
if (--t <= 0) {
drivers/tty/serial/pmac_zilog.c
1140
t = 5000;
drivers/tty/serial/pmac_zilog.c
1142
if (--t <= 0) {
drivers/tty/serial/pmac_zilog.c
1157
t = 5000;
drivers/tty/serial/pmac_zilog.c
1159
if (--t <= 0) {
drivers/tty/serial/pmac_zilog.c
1165
t = read_zsdata(uap);
drivers/tty/serial/pmac_zilog.c
1166
if (t != cmdbyte)
drivers/tty/serial/pmac_zilog.c
1167
pmz_error("irda_setup speed mode byte = %x (%x)\n", t, cmdbyte);
drivers/tty/serial/sa1100.c
112
static void sa1100_timeout(struct timer_list *t)
drivers/tty/serial/sa1100.c
114
struct sa1100_port *sport = timer_container_of(sport, t, timer);
drivers/tty/serial/samsung_tty.c
540
struct tty_port *t = &port->state->port;
drivers/tty/serial/samsung_tty.c
571
s3c24xx_uart_copy_rx_to_tty(ourport, t, received);
drivers/tty/serial/samsung_tty.c
612
struct tty_port *t = &port->state->port;
drivers/tty/serial/samsung_tty.c
626
s3c24xx_uart_copy_rx_to_tty(ourport, t, received);
drivers/tty/serial/samsung_tty.c
629
tty_flip_buffer_push(t);
drivers/tty/serial/samsung_tty.c
718
struct tty_port *t = &port->state->port;
drivers/tty/serial/samsung_tty.c
740
s3c24xx_uart_copy_rx_to_tty(ourport, t, received);
drivers/tty/serial/samsung_tty.c
748
tty_flip_buffer_push(t);
drivers/tty/serial/sccnxp.c
501
static void sccnxp_timer(struct timer_list *t)
drivers/tty/serial/sccnxp.c
503
struct sccnxp_port *s = timer_container_of(s, t, timer);
drivers/tty/serial/sh-sci.c
1321
static void rx_fifo_timer_fn(struct timer_list *t)
drivers/tty/serial/sh-sci.c
1323
struct sci_port *s = timer_container_of(s, t, rx_fifo_timer);
drivers/tty/serial/sh-sci.c
1506
ktime_t t = ktime_set(sec, nsec);
drivers/tty/serial/sh-sci.c
1508
hrtimer_start(hrt, t, HRTIMER_MODE_REL);
drivers/tty/serial/sh-sci.c
1705
static enum hrtimer_restart sci_dma_rx_timer_fn(struct hrtimer *t)
drivers/tty/serial/sh-sci.c
1707
struct sci_port *s = container_of(t, struct sci_port, rx_timer);
drivers/tty/serial/timbuart.c
169
static void timbuart_tasklet(struct tasklet_struct *t)
drivers/tty/serial/timbuart.c
171
struct timbuart_port *uart = from_tasklet(uart, t, tasklet);
drivers/tty/serial/xilinx_uartps.c
392
static enum hrtimer_restart cdns_rs485_rx_callback(struct hrtimer *t)
drivers/tty/serial/xilinx_uartps.c
394
struct cdns_uart *cdns_uart = container_of(t, struct cdns_uart, tx_timer);
drivers/tty/serial/xilinx_uartps.c
694
static enum hrtimer_restart cdns_rs485_tx_callback(struct hrtimer *t)
drivers/tty/serial/xilinx_uartps.c
696
struct cdns_uart *cdns_uart = container_of(t, struct cdns_uart, tx_timer);
drivers/tty/synclink_gt.c
444
static void tx_timeout(struct timer_list *t);
drivers/tty/synclink_gt.c
445
static void rx_timeout(struct timer_list *t);
drivers/tty/synclink_gt.c
5003
static void tx_timeout(struct timer_list *t)
drivers/tty/synclink_gt.c
5005
struct slgt_info *info = timer_container_of(info, t, tx_timer);
drivers/tty/synclink_gt.c
5027
static void rx_timeout(struct timer_list *t)
drivers/tty/synclink_gt.c
5029
struct slgt_info *info = timer_container_of(info, t, rx_timer);
drivers/tty/sysrq.c
1127
struct ctl_table t = *table;
drivers/tty/sysrq.c
1130
t.data = &tmp;
drivers/tty/sysrq.c
1135
ret = proc_dointvec_minmax(&t, write, buffer, lenp, ppos);
drivers/tty/sysrq.c
713
static void sysrq_do_reset(struct timer_list *t)
drivers/tty/sysrq.c
715
struct sysrq_state *state = timer_container_of(state, t,
drivers/tty/tty_io.c
2990
static int this_tty(const void *t, struct file *file, unsigned fd)
drivers/tty/tty_io.c
2994
return file_tty(file) != t ? 0 : fd + 1;
drivers/tty/vcc.c
357
static void vcc_rx_timer(struct timer_list *t)
drivers/tty/vcc.c
359
struct vcc_port *port = timer_container_of(port, t, rx_timer);
drivers/tty/vcc.c
383
static void vcc_tx_timer(struct timer_list *t)
drivers/tty/vcc.c
385
struct vcc_port *port = timer_container_of(port, t, tx_timer);
drivers/tty/vt/consolemap.c
229
unsigned short *t = translations[m];
drivers/tty/vt/consolemap.c
245
int glyph = conv_uni_to_pc(conp, t[ch]);
drivers/usb/atm/cxacru.c
583
static void cxacru_timeout_kill(struct timer_list *t)
drivers/usb/atm/cxacru.c
585
struct cxacru_timer *timer = timer_container_of(timer, t, timer);
drivers/usb/atm/speedtch.c
561
static void speedtch_status_poll(struct timer_list *t)
drivers/usb/atm/speedtch.c
564
t,
drivers/usb/atm/speedtch.c
576
static void speedtch_resubmit_int(struct timer_list *t)
drivers/usb/atm/speedtch.c
579
t,
drivers/usb/atm/ueagle-atm.c
415
#define E1_MAKEFUNCTION(t, s) (((t) & 0xf) << 4 | ((s) & 0xf))
drivers/usb/atm/ueagle-atm.c
416
#define E4_MAKEFUNCTION(t, st, s) (((t) & 0xf) << 8 | \
drivers/usb/atm/usbatm.c
514
static void usbatm_rx_process(struct tasklet_struct *t)
drivers/usb/atm/usbatm.c
516
struct usbatm_data *instance = from_tasklet(instance, t,
drivers/usb/atm/usbatm.c
568
static void usbatm_tx_process(struct tasklet_struct *t)
drivers/usb/atm/usbatm.c
570
struct usbatm_data *instance = from_tasklet(instance, t,
drivers/usb/atm/usbatm.c
977
struct task_struct *t;
drivers/usb/atm/usbatm.c
979
t = kthread_create(usbatm_do_heavy_init, instance, "%s",
drivers/usb/atm/usbatm.c
981
if (IS_ERR(t)) {
drivers/usb/atm/usbatm.c
983
__func__, PTR_ERR(t));
drivers/usb/atm/usbatm.c
984
return PTR_ERR(t);
drivers/usb/atm/usbatm.c
987
instance->thread = t;
drivers/usb/atm/usbatm.c
988
wake_up_process(t);
drivers/usb/atm/usbatm.c
994
static void usbatm_tasklet_schedule(struct timer_list *t)
drivers/usb/atm/usbatm.c
996
struct usbatm_channel *channel = timer_container_of(channel, t, delay);
drivers/usb/chipidea/otg_fsm.c
124
unsigned size, t;
drivers/usb/chipidea/otg_fsm.c
129
t = scnprintf(next, size, "%d\n", ci->fsm.b_bus_req);
drivers/usb/chipidea/otg_fsm.c
130
size -= t;
drivers/usb/chipidea/otg_fsm.c
131
next += t;
drivers/usb/chipidea/otg_fsm.c
219
static void ci_otg_add_timer(struct ci_hdrc *ci, enum otg_fsm_timer t)
drivers/usb/chipidea/otg_fsm.c
223
if (t >= NUM_OTG_FSM_TIMERS)
drivers/usb/chipidea/otg_fsm.c
227
timer_sec = otg_timer_ms[t] / MSEC_PER_SEC;
drivers/usb/chipidea/otg_fsm.c
228
timer_nsec = (otg_timer_ms[t] % MSEC_PER_SEC) * NSEC_PER_MSEC;
drivers/usb/chipidea/otg_fsm.c
229
ci->hr_timeouts[t] = ktime_add(ktime_get(),
drivers/usb/chipidea/otg_fsm.c
231
ci->enabled_otg_timer_bits |= (1 << t);
drivers/usb/chipidea/otg_fsm.c
234
ci->hr_timeouts[t])) {
drivers/usb/chipidea/otg_fsm.c
235
ci->next_otg_timer = t;
drivers/usb/chipidea/otg_fsm.c
237
ci->hr_timeouts[t], NSEC_PER_MSEC,
drivers/usb/chipidea/otg_fsm.c
246
static void ci_otg_del_timer(struct ci_hdrc *ci, enum otg_fsm_timer t)
drivers/usb/chipidea/otg_fsm.c
251
if ((t >= NUM_OTG_FSM_TIMERS) ||
drivers/usb/chipidea/otg_fsm.c
252
!(ci->enabled_otg_timer_bits & (1 << t)))
drivers/usb/chipidea/otg_fsm.c
256
ci->enabled_otg_timer_bits &= ~(1 << t);
drivers/usb/chipidea/otg_fsm.c
257
if (ci->next_otg_timer == t) {
drivers/usb/chipidea/otg_fsm.c
35
unsigned size, t;
drivers/usb/chipidea/otg_fsm.c
383
static enum hrtimer_restart ci_otg_hrtimer_func(struct hrtimer *t)
drivers/usb/chipidea/otg_fsm.c
385
struct ci_hdrc *ci = container_of(t, struct ci_hdrc, otg_fsm_hrtimer);
drivers/usb/chipidea/otg_fsm.c
40
t = scnprintf(next, size, "%d\n", ci->fsm.a_bus_req);
drivers/usb/chipidea/otg_fsm.c
41
size -= t;
drivers/usb/chipidea/otg_fsm.c
42
next += t;
drivers/usb/chipidea/otg_fsm.c
435
static void ci_otg_fsm_add_timer(struct otg_fsm *fsm, enum otg_fsm_timer t)
drivers/usb/chipidea/otg_fsm.c
439
if (t < NUM_OTG_FSM_TIMERS)
drivers/usb/chipidea/otg_fsm.c
440
ci_otg_add_timer(ci, t);
drivers/usb/chipidea/otg_fsm.c
444
static void ci_otg_fsm_del_timer(struct otg_fsm *fsm, enum otg_fsm_timer t)
drivers/usb/chipidea/otg_fsm.c
448
if (t < NUM_OTG_FSM_TIMERS)
drivers/usb/chipidea/otg_fsm.c
449
ci_otg_del_timer(ci, t);
drivers/usb/chipidea/otg_fsm.c
84
unsigned size, t;
drivers/usb/chipidea/otg_fsm.c
89
t = scnprintf(next, size, "%d\n", ci->fsm.a_bus_drop);
drivers/usb/chipidea/otg_fsm.c
90
size -= t;
drivers/usb/chipidea/otg_fsm.c
91
next += t;
drivers/usb/core/devio.c
497
const char *t, *d;
drivers/usb/core/devio.c
503
t = types[usb_pipetype(pipe)];
drivers/usb/core/devio.c
510
userurb, ep, t, d, length);
drivers/usb/core/devio.c
514
userurb, ep, t, d, length,
drivers/usb/core/devio.c
520
ep, t, d, length, timeout_or_status);
drivers/usb/core/devio.c
524
ep, t, d, length, timeout_or_status);
drivers/usb/core/hcd.c
378
unsigned n, t = 2 + 2*strlen(s);
drivers/usb/core/hcd.c
380
if (t > 254)
drivers/usb/core/hcd.c
381
t = 254; /* Longest possible UTF string descriptor */
drivers/usb/core/hcd.c
382
if (len > t)
drivers/usb/core/hcd.c
383
len = t;
drivers/usb/core/hcd.c
385
t += USB_DT_STRING << 8; /* Now t is first 16 bits to store */
drivers/usb/core/hcd.c
389
*buf++ = t;
drivers/usb/core/hcd.c
392
*buf++ = t >> 8;
drivers/usb/core/hcd.c
393
t = (unsigned char)*s++;
drivers/usb/core/hcd.c
772
static void rh_timer_func (struct timer_list *t)
drivers/usb/core/hcd.c
774
struct usb_hcd *_hcd = timer_container_of(_hcd, t, rh_timer);
drivers/usb/core/hub.c
700
static void hub_retry_irq_urb(struct timer_list *t)
drivers/usb/core/hub.c
702
struct usb_hub *hub = timer_container_of(hub, t, irq_urb_retry);
drivers/usb/dwc2/hcd.c
3228
static void dwc2_wakeup_detected(struct timer_list *t)
drivers/usb/dwc2/hcd.c
3230
struct dwc2_hsotg *hsotg = timer_container_of(hsotg, t, wkp_timer);
drivers/usb/dwc2/hcd_queue.c
1216
static void dwc2_unreserve_timer_fn(struct timer_list *t)
drivers/usb/dwc2/hcd_queue.c
1218
struct dwc2_qh *qh = timer_container_of(qh, t, unreserve_timer);
drivers/usb/dwc2/hcd_queue.c
1404
static enum hrtimer_restart dwc2_wait_timer_fn(struct hrtimer *t)
drivers/usb/dwc2/hcd_queue.c
1406
struct dwc2_qh *qh = container_of(t, struct dwc2_qh, wait_timer);
drivers/usb/dwc3/gadget.c
2165
struct dwc3_request *t;
drivers/usb/dwc3/gadget.c
2174
list_for_each_entry_safe(r, t, &dep->started_list, list)
drivers/usb/fotg210/fotg210-hcd.c
1330
static enum hrtimer_restart fotg210_hrtimer_func(struct hrtimer *t)
drivers/usb/fotg210/fotg210-hcd.c
1333
container_of(t, struct fotg210_hcd, hrtimer);
drivers/usb/fotg210/fotg210-hcd.c
4424
u32 t;
drivers/usb/fotg210/fotg210-hcd.c
4438
t = hc32_to_cpup(fotg210, &itd->hw_transaction[uframe]);
drivers/usb/fotg210/fotg210-hcd.c
4442
if (unlikely(t & ISO_ERRS)) {
drivers/usb/fotg210/fotg210-hcd.c
4444
if (t & FOTG210_ISOC_BUF_ERR)
drivers/usb/fotg210/fotg210-hcd.c
4448
else if (t & FOTG210_ISOC_BABBLE)
drivers/usb/fotg210/fotg210-hcd.c
4454
if (!(t & FOTG210_ISOC_BABBLE)) {
drivers/usb/fotg210/fotg210-hcd.c
4455
desc->actual_length = FOTG210_ITD_LENGTH(t);
drivers/usb/fotg210/fotg210-hcd.c
4458
} else if (likely((t & FOTG210_ISOC_ACTIVE) == 0)) {
drivers/usb/fotg210/fotg210-hcd.c
4460
desc->actual_length = FOTG210_ITD_LENGTH(t);
drivers/usb/gadget/composite.c
112
next_desc(struct usb_descriptor_header **t, u8 desc_type)
drivers/usb/gadget/composite.c
114
for (; *t; t++) {
drivers/usb/gadget/composite.c
115
if ((*t)->bDescriptorType == desc_type)
drivers/usb/gadget/composite.c
116
return t;
drivers/usb/gadget/function/f_fs.c
3039
struct usb_gadget_strings **stringtabs, *t;
drivers/usb/gadget/function/f_fs.c
3087
t = vla_ptr(vlabuf, d, stringtab);
drivers/usb/gadget/function/f_fs.c
3090
*stringtabs++ = t++;
drivers/usb/gadget/function/f_fs.c
3096
t = vla_ptr(vlabuf, d, stringtab);
drivers/usb/gadget/function/f_fs.c
3110
t->language = get_unaligned_le16(data);
drivers/usb/gadget/function/f_fs.c
3111
t->strings = s;
drivers/usb/gadget/function/f_fs.c
3112
++t;
drivers/usb/gadget/function/f_fs.c
3423
struct usb_os_desc_table *t;
drivers/usb/gadget/function/f_fs.c
3425
t = &func->function.os_desc_table[desc->bFirstInterfaceNumber];
drivers/usb/gadget/function/f_fs.c
3426
t->if_id = func->interfaces_nums[desc->bFirstInterfaceNumber];
drivers/usb/gadget/function/f_fs.c
3427
memcpy(t->os_desc->ext_compat_id, &desc->IDs,
drivers/usb/gadget/function/f_fs.c
3434
struct usb_os_desc_table *t;
drivers/usb/gadget/function/f_fs.c
3439
t = &func->function.os_desc_table[h->interface];
drivers/usb/gadget/function/f_fs.c
3440
t->if_id = func->interfaces_nums[h->interface];
drivers/usb/gadget/function/f_fs.c
3478
t->os_desc->ext_prop_len +=
drivers/usb/gadget/function/f_fs.c
3480
++t->os_desc->ext_prop_count;
drivers/usb/gadget/function/f_fs.c
3481
list_add_tail(&ext_prop->entry, &t->os_desc->ext_prop);
drivers/usb/gadget/function/u_uac1_legacy.c
31
struct snd_interval t;
drivers/usb/gadget/function/u_uac1_legacy.c
32
t.empty = 0;
drivers/usb/gadget/function/u_uac1_legacy.c
33
t.min = t.max = val;
drivers/usb/gadget/function/u_uac1_legacy.c
34
t.openmin = t.openmax = 0;
drivers/usb/gadget/function/u_uac1_legacy.c
35
t.integer = 1;
drivers/usb/gadget/function/u_uac1_legacy.c
36
return snd_interval_refine(i, &t);
drivers/usb/gadget/function/u_uac1_legacy.c
65
struct snd_interval t;
drivers/usb/gadget/function/u_uac1_legacy.c
66
t.openmin = 1;
drivers/usb/gadget/function/u_uac1_legacy.c
67
t.openmax = 1;
drivers/usb/gadget/function/u_uac1_legacy.c
68
t.empty = 0;
drivers/usb/gadget/function/u_uac1_legacy.c
69
t.integer = 0;
drivers/usb/gadget/function/u_uac1_legacy.c
71
t.min = val - 1;
drivers/usb/gadget/function/u_uac1_legacy.c
72
t.max = val;
drivers/usb/gadget/function/u_uac1_legacy.c
74
t.min = val;
drivers/usb/gadget/function/u_uac1_legacy.c
75
t.max = val+1;
drivers/usb/gadget/function/u_uac1_legacy.c
77
changed = snd_interval_refine(i, &t);
drivers/usb/gadget/udc/at91_udc.c
1542
static void at91_vbus_timer(struct timer_list *t)
drivers/usb/gadget/udc/at91_udc.c
1544
struct at91_udc *udc = timer_container_of(udc, t, vbus_timer);
drivers/usb/gadget/udc/dummy_hcd.c
1798
static enum hrtimer_restart dummy_timer(struct hrtimer *t)
drivers/usb/gadget/udc/dummy_hcd.c
1800
struct dummy_hcd *dum_hcd = timer_container_of(dum_hcd, t,
drivers/usb/gadget/udc/fsl_qe_udc.c
933
static void ep_rx_tasklet(struct tasklet_struct *t)
drivers/usb/gadget/udc/fsl_qe_udc.c
935
struct qe_udc *udc = from_tasklet(udc, t, rx_tasklet);
drivers/usb/gadget/udc/m66592-udc.c
1262
static void m66592_timer(struct timer_list *t)
drivers/usb/gadget/udc/m66592-udc.c
1264
struct m66592 *m66592 = timer_container_of(m66592, t, timer);
drivers/usb/gadget/udc/max3420_udc.c
1045
struct max3420_req *t = NULL;
drivers/usb/gadget/udc/max3420_udc.c
1058
t = iter;
drivers/usb/gadget/udc/max3420_udc.c
1064
if (t)
drivers/usb/gadget/udc/net2280.c
1655
unsigned size, t;
drivers/usb/gadget/udc/net2280.c
1672
t = scnprintf(next, size, "%s version " DRIVER_VERSION
drivers/usb/gadget/udc/net2280.c
1685
size -= t;
drivers/usb/gadget/udc/net2280.c
1686
next += t;
drivers/usb/gadget/udc/net2280.c
1701
t = scnprintf(next, size,
drivers/usb/gadget/udc/net2280.c
1706
size -= t;
drivers/usb/gadget/udc/net2280.c
1707
next += t;
drivers/usb/gadget/udc/net2280.c
1723
t = scnprintf(next, size,
drivers/usb/gadget/udc/net2280.c
1744
size -= t;
drivers/usb/gadget/udc/net2280.c
1745
next += t;
drivers/usb/gadget/udc/net2280.c
1747
t = scnprintf(next, size,
drivers/usb/gadget/udc/net2280.c
1755
size -= t;
drivers/usb/gadget/udc/net2280.c
1756
next += t;
drivers/usb/gadget/udc/net2280.c
1761
t = scnprintf(next, size,
drivers/usb/gadget/udc/net2280.c
1769
size -= t;
drivers/usb/gadget/udc/net2280.c
1770
next += t;
drivers/usb/gadget/udc/net2280.c
1777
t = scnprintf(next, size, "\nirqs: ");
drivers/usb/gadget/udc/net2280.c
1778
size -= t;
drivers/usb/gadget/udc/net2280.c
1779
next += t;
drivers/usb/gadget/udc/net2280.c
1786
t = scnprintf(next, size, " %s/%lu", ep->ep.name, ep->irqs);
drivers/usb/gadget/udc/net2280.c
1787
size -= t;
drivers/usb/gadget/udc/net2280.c
1788
next += t;
drivers/usb/gadget/udc/net2280.c
1791
t = scnprintf(next, size, "\n");
drivers/usb/gadget/udc/net2280.c
1792
size -= t;
drivers/usb/gadget/udc/net2280.c
1793
next += t;
drivers/usb/gadget/udc/net2280.c
1818
int t;
drivers/usb/gadget/udc/net2280.c
1826
t = d->bEndpointAddress;
drivers/usb/gadget/udc/net2280.c
1827
t = scnprintf(next, size,
drivers/usb/gadget/udc/net2280.c
1829
ep->ep.name, t & USB_ENDPOINT_NUMBER_MASK,
drivers/usb/gadget/udc/net2280.c
1830
(t & USB_DIR_IN) ? "in" : "out",
drivers/usb/gadget/udc/net2280.c
1836
t = scnprintf(next, size, "ep0 max 64 pio %s\n",
drivers/usb/gadget/udc/net2280.c
1838
if (t <= 0 || t > size)
drivers/usb/gadget/udc/net2280.c
1840
size -= t;
drivers/usb/gadget/udc/net2280.c
1841
next += t;
drivers/usb/gadget/udc/net2280.c
1844
t = scnprintf(next, size, "\t(nothing queued)\n");
drivers/usb/gadget/udc/net2280.c
1845
if (t <= 0 || t > size)
drivers/usb/gadget/udc/net2280.c
1847
size -= t;
drivers/usb/gadget/udc/net2280.c
1848
next += t;
drivers/usb/gadget/udc/net2280.c
1853
t = scnprintf(next, size,
drivers/usb/gadget/udc/net2280.c
1860
t = scnprintf(next, size,
drivers/usb/gadget/udc/net2280.c
1864
if (t <= 0 || t > size)
drivers/usb/gadget/udc/net2280.c
1866
size -= t;
drivers/usb/gadget/udc/net2280.c
1867
next += t;
drivers/usb/gadget/udc/net2280.c
1873
t = scnprintf(next, size, "\t td %08x "
drivers/usb/gadget/udc/net2280.c
1879
if (t <= 0 || t > size)
drivers/usb/gadget/udc/net2280.c
1881
size -= t;
drivers/usb/gadget/udc/net2280.c
1882
next += t;
drivers/usb/gadget/udc/net2280.c
2519
u32 t;
drivers/usb/gadget/udc/net2280.c
2530
t = readl(&ep->regs->ep_stat);
drivers/usb/gadget/udc/net2280.c
2534
ep->ep.name, t, req ? &req->req : NULL);
drivers/usb/gadget/udc/net2280.c
2537
writel(t & ~BIT(NAK_OUT_PACKETS), &ep->regs->ep_stat);
drivers/usb/gadget/udc/net2280.c
2540
writel(t, &ep->regs->ep_stat);
drivers/usb/gadget/udc/net2280.c
2555
if (t & BIT(DATA_OUT_PING_TOKEN_INTERRUPT)) {
drivers/usb/gadget/udc/net2280.c
2564
} else if (t & BIT(DATA_IN_TOKEN_INTERRUPT)) {
drivers/usb/gadget/udc/net2280.c
2575
if (t & BIT(DATA_IN_TOKEN_INTERRUPT)) {
drivers/usb/gadget/udc/net2280.c
2582
} else if (((t & BIT(DATA_OUT_PING_TOKEN_INTERRUPT)) &&
drivers/usb/gadget/udc/net2280.c
2601
if (t & BIT(SHORT_PACKET_TRANSFERRED_INTERRUPT)) {
drivers/usb/gadget/udc/net2280.c
2613
for (count = 0; ; t = readl(&ep->regs->ep_stat)) {
drivers/usb/gadget/udc/net2280.c
2630
if (likely(t & BIT(FIFO_EMPTY))) {
drivers/usb/gadget/udc/net2280.c
2672
t = readl(&ep->regs->ep_avail);
drivers/usb/gadget/udc/net2280.c
2674
(ep->out_overflow || t)
drivers/usb/gadget/udc/net2280.c
2693
ep->ep.name, t);
drivers/usb/gadget/udc/net2280.c
2697
} else if (t & BIT(DATA_PACKET_RECEIVED_INTERRUPT)) {
drivers/usb/gadget/udc/net2280.c
2702
} else if (t & BIT(DATA_PACKET_TRANSMITTED_INTERRUPT)) {
drivers/usb/gadget/udc/net2280.c
2751
if (t & BIT(DATA_PACKET_TRANSMITTED_INTERRUPT))
drivers/usb/gadget/udc/net2280.c
3332
u32 t;
drivers/usb/gadget/udc/net2280.c
3335
t = BIT(num);
drivers/usb/gadget/udc/net2280.c
3336
if ((scratch & t) == 0)
drivers/usb/gadget/udc/net2280.c
3338
scratch ^= t;
drivers/usb/gadget/udc/omap_udc.c
1861
static void pio_out_timer(struct timer_list *t)
drivers/usb/gadget/udc/omap_udc.c
1863
struct omap_ep *ep = timer_container_of(ep, t, timer);
drivers/usb/gadget/udc/pxa25x_udc.c
1575
static void udc_watchdog(struct timer_list *t)
drivers/usb/gadget/udc/pxa25x_udc.c
1577
struct pxa25x_udc *dev = timer_container_of(dev, t, timer);
drivers/usb/gadget/udc/r8a66597-udc.c
1517
static void r8a66597_timer(struct timer_list *t)
drivers/usb/gadget/udc/r8a66597-udc.c
1519
struct r8a66597 *r8a66597 = timer_container_of(r8a66597, t, timer);
drivers/usb/host/ehci-platform.c
199
static void quirk_poll_timer(struct timer_list *t)
drivers/usb/host/ehci-platform.c
201
struct ehci_platform_priv *priv = timer_container_of(priv, t,
drivers/usb/host/ehci-sched.c
1824
u32 t;
drivers/usb/host/ehci-sched.c
1837
t = hc32_to_cpup(ehci, &itd->hw_transaction[uframe]);
drivers/usb/host/ehci-sched.c
1841
if (unlikely(t & ISO_ERRS)) {
drivers/usb/host/ehci-sched.c
1843
if (t & EHCI_ISOC_BUF_ERR)
drivers/usb/host/ehci-sched.c
1847
else if (t & EHCI_ISOC_BABBLE)
drivers/usb/host/ehci-sched.c
1853
if (!(t & EHCI_ISOC_BABBLE)) {
drivers/usb/host/ehci-sched.c
1854
desc->actual_length = EHCI_ITD_LENGTH(t);
drivers/usb/host/ehci-sched.c
1857
} else if (likely((t & EHCI_ISOC_ACTIVE) == 0)) {
drivers/usb/host/ehci-sched.c
1859
desc->actual_length = EHCI_ITD_LENGTH(t);
drivers/usb/host/ehci-sched.c
2218
u32 t;
drivers/usb/host/ehci-sched.c
2225
t = hc32_to_cpup(ehci, &sitd->hw_results);
drivers/usb/host/ehci-sched.c
2228
if (unlikely(t & SITD_ERRS)) {
drivers/usb/host/ehci-sched.c
2230
if (t & SITD_STS_DBE)
drivers/usb/host/ehci-sched.c
2234
else if (t & SITD_STS_BABBLE)
drivers/usb/host/ehci-sched.c
2238
} else if (unlikely(t & SITD_STS_ACTIVE)) {
drivers/usb/host/ehci-sched.c
2243
desc->actual_length = desc->length - SITD_LENGTH(t);
drivers/usb/host/ehci-timer.c
398
static enum hrtimer_restart ehci_hrtimer_func(struct hrtimer *t)
drivers/usb/host/ehci-timer.c
400
struct ehci_hcd *ehci = container_of(t, struct ehci_hcd, hrtimer);
drivers/usb/host/isp116x-hcd.c
1249
unsigned long t;
drivers/usb/host/isp116x-hcd.c
1257
t = jiffies + msecs_to_jiffies(timeout);
drivers/usb/host/isp116x-hcd.c
1258
while (time_before_eq(jiffies, t)) {
drivers/usb/host/isp116x-hcd.c
972
unsigned long flags, t;
drivers/usb/host/isp116x-hcd.c
976
t = jiffies + msecs_to_jiffies(100);
drivers/usb/host/isp116x-hcd.c
978
while (time_before(jiffies, t)) {
drivers/usb/host/octeon-hcd.c
1848
struct cvmx_usb_transaction *t =
drivers/usb/host/octeon-hcd.c
1849
list_first_entry(&pipe->transactions, typeof(*t),
drivers/usb/host/octeon-hcd.c
1851
if (!(pipe->flags & CVMX_USB_PIPE_FLAGS_SCHEDULED) && t &&
drivers/usb/host/octeon-hcd.c
1856
(!usb->active_split || (usb->active_split == t))) {
drivers/usb/host/octeon-hcd.c
1857
prefetch(t);
drivers/usb/host/ohci-hcd.c
747
static void io_watchdog_func(struct timer_list *t)
drivers/usb/host/ohci-hcd.c
749
struct ohci_hcd *ohci = timer_container_of(ohci, t,
drivers/usb/host/ohci-hcd.c
85
static void io_watchdog_func(struct timer_list *t);
drivers/usb/host/oxu210hp-hcd.c
2956
static void oxu_watchdog(struct timer_list *t)
drivers/usb/host/oxu210hp-hcd.c
2958
struct oxu_hcd *oxu = timer_container_of(oxu, t, watchdog);
drivers/usb/host/oxu210hp-hcd.c
676
unsigned long t;
drivers/usb/host/oxu210hp-hcd.c
680
t = EHCI_IAA_JIFFIES;
drivers/usb/host/oxu210hp-hcd.c
683
t = EHCI_IO_JIFFIES;
drivers/usb/host/oxu210hp-hcd.c
686
t = EHCI_ASYNC_JIFFIES;
drivers/usb/host/oxu210hp-hcd.c
690
t = EHCI_SHRINK_JIFFIES;
drivers/usb/host/oxu210hp-hcd.c
693
t += jiffies;
drivers/usb/host/oxu210hp-hcd.c
700
&& t > oxu->watchdog.expires
drivers/usb/host/oxu210hp-hcd.c
703
mod_timer(&oxu->watchdog, t);
drivers/usb/host/r8a66597-hcd.c
1721
static void r8a66597_interval_timer(struct timer_list *t)
drivers/usb/host/r8a66597-hcd.c
1723
struct r8a66597_timers *timers = timer_container_of(timers, t,
drivers/usb/host/r8a66597-hcd.c
1746
static void r8a66597_td_timer(struct timer_list *t)
drivers/usb/host/r8a66597-hcd.c
1748
struct r8a66597_timers *timers = timer_container_of(timers, t, td);
drivers/usb/host/r8a66597-hcd.c
1800
static void r8a66597_timer(struct timer_list *t)
drivers/usb/host/r8a66597-hcd.c
1802
struct r8a66597 *r8a66597 = timer_container_of(r8a66597, t, rh_timer);
drivers/usb/host/sl811-hcd.c
1125
sl811h_timer(struct timer_list *t)
drivers/usb/host/sl811-hcd.c
1127
struct sl811 *sl811 = timer_container_of(sl811, t, timer);
drivers/usb/host/sl811-hcd.c
1409
u8 t = sl811_read(sl811, SL11H_CTLREG1);
drivers/usb/host/sl811-hcd.c
1411
seq_printf(s, "ctrl1 %02x%s%s%s%s\n", t,
drivers/usb/host/sl811-hcd.c
1412
(t & SL11H_CTL1MASK_SOF_ENA) ? " sofgen" : "",
drivers/usb/host/sl811-hcd.c
1413
({char *s; switch (t & SL11H_CTL1MASK_FORCE) {
drivers/usb/host/sl811-hcd.c
1419
(t & SL11H_CTL1MASK_LSPD) ? " lowspeed" : "",
drivers/usb/host/sl811-hcd.c
1420
(t & SL11H_CTL1MASK_SUSPEND) ? " suspend" : "");
drivers/usb/host/uhci-q.c
92
static void uhci_fsbr_timeout(struct timer_list *t)
drivers/usb/host/uhci-q.c
94
struct uhci_hcd *uhci = timer_container_of(uhci, t, fsbr_timer);
drivers/usb/host/xen-hcd.c
131
unsigned long t;
drivers/usb/host/xen-hcd.c
135
t = XENHCD_RING_JIFFIES;
drivers/usb/host/xen-hcd.c
138
t = XENHCD_SCAN_JIFFIES;
drivers/usb/host/xen-hcd.c
141
mod_timer(&info->watchdog, t + jiffies);
drivers/usb/host/xhci-dbgcap.h
197
#define dbc_epctx_info2(t, p, b) \
drivers/usb/host/xhci-dbgcap.h
198
cpu_to_le32(EP_TYPE(t) | MAX_PACKET(p) | MAX_BURST(b))
drivers/usb/host/xhci-dbgtty.c
403
static void dbc_rx_push(struct tasklet_struct *t)
drivers/usb/host/xhci-dbgtty.c
409
struct dbc_port *port = from_tasklet(port, t, push);
drivers/usb/host/xhci.c
369
static void compliance_mode_recovery(struct timer_list *t)
drivers/usb/host/xhci.c
377
xhci = timer_container_of(xhci, t, comp_mode_recovery_timer);
drivers/usb/isp1760/isp1760-udc.c
1424
static void isp1760_udc_vbus_poll(struct timer_list *t)
drivers/usb/isp1760/isp1760-udc.c
1426
struct isp1760_udc *udc = timer_container_of(udc, t, vbus_timer);
drivers/usb/misc/usbtest.c
593
static void sg_timeout(struct timer_list *t)
drivers/usb/misc/usbtest.c
595
struct sg_timeout *timeout = timer_container_of(timeout, t, timer);
drivers/usb/musb/da8xx.c
124
static void otg_timer(struct timer_list *t)
drivers/usb/musb/da8xx.c
126
struct musb *musb = timer_container_of(musb, t,
drivers/usb/musb/mpfs.c
92
static void otg_timer(struct timer_list *t)
drivers/usb/musb/mpfs.c
94
struct musb *musb = timer_container_of(musb, t,
drivers/usb/musb/musb_core.c
597
static void musb_otg_timer_func(struct timer_list *t)
drivers/usb/musb/musb_core.c
599
struct musb *musb = timer_container_of(musb, t, otg_timer);
drivers/usb/musb/musb_dsps.c
279
static void otg_timer(struct timer_list *t)
drivers/usb/musb/musb_dsps.c
281
struct musb *musb = timer_container_of(musb, t, dev_timer);
drivers/usb/musb/tusb6010.c
458
static void musb_do_idle(struct timer_list *t)
drivers/usb/musb/tusb6010.c
460
struct musb *musb = timer_container_of(musb, t, dev_timer);
drivers/usb/phy/phy-fsl-usb.c
359
static struct fsl_otg_timer *fsl_otg_get_timer(enum otg_fsm_timer t)
drivers/usb/phy/phy-fsl-usb.c
364
switch (t) {
drivers/usb/phy/phy-fsl-usb.c
412
static void fsl_otg_fsm_add_timer(struct otg_fsm *fsm, enum otg_fsm_timer t)
drivers/usb/phy/phy-fsl-usb.c
416
timer = fsl_otg_get_timer(t);
drivers/usb/phy/phy-fsl-usb.c
434
static void fsl_otg_fsm_del_timer(struct otg_fsm *fsm, enum otg_fsm_timer t)
drivers/usb/phy/phy-fsl-usb.c
438
timer = fsl_otg_get_timer(t);
drivers/usb/renesas_usbhs/mod_host.c
205
int t = 0;
drivers/usb/renesas_usbhs/mod_host.c
228
t = len / maxp;
drivers/usb/renesas_usbhs/mod_host.c
230
t++;
drivers/usb/renesas_usbhs/mod_host.c
232
t++;
drivers/usb/renesas_usbhs/mod_host.c
233
t %= 2;
drivers/usb/renesas_usbhs/mod_host.c
235
if (t)
drivers/usb/renesas_usbhs/pipe.h
108
#define usbhs_pipe_type_is(p, t) ((p)->pipe_type == t)
drivers/usb/serial/garmin_gps.c
1358
static void timeout_handler(struct timer_list *t)
drivers/usb/serial/garmin_gps.c
1361
t, timer);
drivers/usb/serial/mos7840.c
384
static void mos7840_led_off(struct timer_list *t)
drivers/usb/serial/mos7840.c
386
struct moschip_port *mcs = timer_container_of(mcs, t, led_timer1);
drivers/usb/serial/mos7840.c
394
static void mos7840_led_flag_off(struct timer_list *t)
drivers/usb/serial/mos7840.c
396
struct moschip_port *mcs = timer_container_of(mcs, t, led_timer2);
drivers/usb/serial/whiteheat.c
537
int t;
drivers/usb/serial/whiteheat.c
561
t = wait_event_timeout(command_info->wait_command,
drivers/usb/serial/whiteheat.c
563
if (!t)
drivers/usb/storage/realtek_cr.c
755
static void rts51x_suspend_timer_fn(struct timer_list *t)
drivers/usb/storage/realtek_cr.c
757
struct rts51x_chip *chip = timer_container_of(chip, t,
drivers/usb/usbip/vudc_transfer.c
302
static void v_timer(struct timer_list *t)
drivers/usb/usbip/vudc_transfer.c
304
struct vudc *udc = timer_container_of(udc, t, tr_timer.timer);
drivers/usb/usbip/vudc_transfer.c
448
struct transfer_timer *t = &udc->tr_timer;
drivers/usb/usbip/vudc_transfer.c
450
timer_setup(&t->timer, v_timer, 0);
drivers/usb/usbip/vudc_transfer.c
451
t->state = VUDC_TR_STOPPED;
drivers/usb/usbip/vudc_transfer.c
456
struct transfer_timer *t = &udc->tr_timer;
drivers/usb/usbip/vudc_transfer.c
459
switch (t->state) {
drivers/usb/usbip/vudc_transfer.c
465
t->state = VUDC_TR_IDLE;
drivers/usb/usbip/vudc_transfer.c
466
t->frame_start = jiffies;
drivers/usb/usbip/vudc_transfer.c
467
t->frame_limit = get_frame_limit(udc->gadget.speed);
drivers/usb/usbip/vudc_transfer.c
474
struct transfer_timer *t = &udc->tr_timer;
drivers/usb/usbip/vudc_transfer.c
477
switch (t->state) {
drivers/usb/usbip/vudc_transfer.c
481
t->state = VUDC_TR_RUNNING;
drivers/usb/usbip/vudc_transfer.c
485
mod_timer(&t->timer, time);
drivers/usb/usbip/vudc_transfer.c
491
struct transfer_timer *t = &udc->tr_timer;
drivers/usb/usbip/vudc_transfer.c
495
t->state = VUDC_TR_STOPPED;
drivers/vhost/scsi.c
1982
struct vhost_scsi_target *t)
drivers/vhost/scsi.c
2030
if (!strcmp(tv_tport->tport_name, t->vhost_wwpn)) {
drivers/vhost/scsi.c
2055
memcpy(vs->vs_vhost_wwpn, t->vhost_wwpn,
drivers/vhost/scsi.c
2114
struct vhost_scsi_target *t)
drivers/vhost/scsi.c
2150
if (strcmp(tv_tport->tport_name, t->vhost_wwpn)) {
drivers/vhost/scsi.c
2154
t->vhost_wwpn, t->vhost_tpgt);
drivers/vhost/scsi.c
2346
struct vhost_scsi_target t;
drivers/vhost/scsi.c
2349
memcpy(t.vhost_wwpn, vs->vs_vhost_wwpn, sizeof(t.vhost_wwpn));
drivers/vhost/scsi.c
2351
vhost_scsi_clear_endpoint(vs, &t);
drivers/vhost/scsi.c
472
struct vhost_scsi_cmd *cmd, *t;
drivers/vhost/scsi.c
476
llist_for_each_entry_safe(cmd, t, llnode, tvc_completion_list)
drivers/vhost/scsi.c
636
struct vhost_scsi_evt *evt, *t;
drivers/vhost/scsi.c
641
llist_for_each_entry_safe(evt, t, llnode, list) {
drivers/vhost/scsi.c
691
struct vhost_scsi_cmd *cmd, *t;
drivers/vhost/scsi.c
702
llist_for_each_entry_safe(cmd, t, llnode, tvc_completion_list) {
drivers/vhost/vhost.c
1334
struct iov_iter t;
drivers/vhost/vhost.c
1347
iov_iter_init(&t, ITER_DEST, vq->iotlb_iov, ret, size);
drivers/vhost/vhost.c
1348
ret = copy_to_iter(from, size, &t);
drivers/video/console/mdacon.c
494
static bool mdacon_scroll(struct vc_data *c, unsigned int t, unsigned int b,
drivers/video/console/mdacon.c
508
scr_memmovew(mda_addr(0, t), mda_addr(0, t + lines),
drivers/video/console/mdacon.c
509
(b-t-lines)*mda_num_columns*2);
drivers/video/console/mdacon.c
515
scr_memmovew(mda_addr(0, t + lines), mda_addr(0, t),
drivers/video/console/mdacon.c
516
(b-t-lines)*mda_num_columns*2);
drivers/video/console/mdacon.c
517
scr_memsetw(mda_addr(0, t), eattr, lines*mda_num_columns*2);
drivers/video/console/newport_con.c
579
static bool newport_scroll(struct vc_data *vc, unsigned int t, unsigned int b,
drivers/video/console/newport_con.c
588
if (t == 0 && b == vc->vc_rows) {
drivers/video/console/newport_con.c
603
count = (b - t - lines) * vc->vc_cols;
drivers/video/console/newport_con.c
606
y = t;
drivers/video/console/newport_con.c
608
vc->vc_size_row * (t + lines));
drivers/video/console/newport_con.c
610
vc->vc_size_row * t);
drivers/video/console/newport_con.c
659
vc->vc_size_row * t);
drivers/video/console/newport_con.c
661
y = t;
drivers/video/console/sticon.c
116
static bool sticon_scroll(struct vc_data *conp, unsigned int t,
drivers/video/console/sticon.c
128
sti_bmove(sti, t + count, 0, t, 0, b - t - count, conp->vc_cols,
drivers/video/console/sticon.c
135
sti_bmove(sti, t, 0, t + count, 0, b - t - count, conp->vc_cols,
drivers/video/console/sticon.c
137
sti_clear(sti, t, 0, count, conp->vc_cols,
drivers/video/console/vgacon.c
1136
static bool vgacon_scroll(struct vc_data *c, unsigned int t, unsigned int b,
drivers/video/console/vgacon.c
1142
if (t || b != c->vc_rows || vga_is_gfx || c->vc_mode != KD_TEXT)
drivers/video/fbdev/atafb.c
1034
if (var->pixclock > f32.t * plen)
drivers/video/fbdev/atafb.c
1096
if (f25.t * i >= var->pixclock &&
drivers/video/fbdev/atafb.c
1097
f25.t * i < pcl) {
drivers/video/fbdev/atafb.c
1098
pcl = f25.t * i;
drivers/video/fbdev/atafb.c
1101
if (f32.t * i >= var->pixclock &&
drivers/video/fbdev/atafb.c
1102
f32.t * i < pcl) {
drivers/video/fbdev/atafb.c
1103
pcl = f32.t * i;
drivers/video/fbdev/atafb.c
1106
if (fext.t && fext.t * i >= var->pixclock &&
drivers/video/fbdev/atafb.c
1107
fext.t * i < pcl) {
drivers/video/fbdev/atafb.c
1108
pcl = fext.t * i;
drivers/video/fbdev/atafb.c
1114
plen = pcl / pclock->t;
drivers/video/fbdev/atafb.c
1360
var->pixclock = hw->sync & 0x1 ? fext.t :
drivers/video/fbdev/atafb.c
1361
hw->vid_control & VCO_CLOCK25 ? f25.t : f32.t;
drivers/video/fbdev/atafb.c
1750
f25.hsync = h_syncs[mon_type] / f25.t;
drivers/video/fbdev/atafb.c
1751
f32.hsync = h_syncs[mon_type] / f32.t;
drivers/video/fbdev/atafb.c
1752
if (fext.t)
drivers/video/fbdev/atafb.c
1753
fext.hsync = h_syncs[mon_type] / fext.t;
drivers/video/fbdev/atafb.c
2975
fext.t = 1000000000 / fext.f;
drivers/video/fbdev/atafb.c
824
unsigned long t; /* t/[ps] (=1/f) */
drivers/video/fbdev/aty/radeon_base.c
1444
static void radeon_lvds_timer_func(struct timer_list *t)
drivers/video/fbdev/aty/radeon_base.c
1446
struct radeonfb_info *rinfo = timer_container_of(rinfo, t, lvds_timer);
drivers/video/fbdev/c2p_core.h
25
u32 t = (d[i1] ^ (d[i2] >> shift)) & mask;
drivers/video/fbdev/c2p_core.h
27
d[i1] ^= t;
drivers/video/fbdev/c2p_core.h
28
d[i2] ^= t << shift;
drivers/video/fbdev/core/fbcon.c
1048
struct fbcon_display *t, *p = &fb_display[vc->vc_num];
drivers/video/fbdev/core/fbcon.c
1076
t = &fb_display[fg_console];
drivers/video/fbdev/core/fbcon.c
1078
if (t->fontdata) {
drivers/video/fbdev/core/fbcon.c
1086
p->userfont = t->userfont;
drivers/video/fbdev/core/fbcon.c
1391
struct fbcon_display *p, *t;
drivers/video/fbdev/core/fbcon.c
1410
t = &fb_display[svc->vc_num];
drivers/video/fbdev/core/fbcon.c
1413
vc->vc_font.data = (void *)(p->fontdata = t->fontdata);
drivers/video/fbdev/core/fbcon.c
1417
p->userfont = t->userfont;
drivers/video/fbdev/core/fbcon.c
1515
static __inline__ void ypan_up_redraw(struct vc_data *vc, int t, int count)
drivers/video/fbdev/core/fbcon.c
1525
fbcon_redraw_move(vc, p, t + count, vc->vc_rows - count, t);
drivers/video/fbdev/core/fbcon.c
1563
static __inline__ void ypan_down_redraw(struct vc_data *vc, int t, int count)
drivers/video/fbdev/core/fbcon.c
1573
fbcon_redraw_move(vc, p, t, vc->vc_rows - count, t + count);
drivers/video/fbdev/core/fbcon.c
1784
static bool fbcon_scroll(struct vc_data *vc, unsigned int t, unsigned int b,
drivers/video/fbdev/core/fbcon.c
1808
fbcon_redraw_blit(vc, info, p, t, b - t - count,
drivers/video/fbdev/core/fbcon.c
1819
if (b - t - count > 3 * vc->vc_rows >> 2) {
drivers/video/fbdev/core/fbcon.c
1820
if (t > 0)
drivers/video/fbdev/core/fbcon.c
1821
fbcon_bmove(vc, 0, 0, count, 0, t,
drivers/video/fbdev/core/fbcon.c
1829
fbcon_bmove(vc, t + count, 0, t, 0,
drivers/video/fbdev/core/fbcon.c
1830
b - t - count, vc->vc_cols);
drivers/video/fbdev/core/fbcon.c
1839
&& ((!scroll_partial && (b - t == vc->vc_rows))
drivers/video/fbdev/core/fbcon.c
1841
&& (b - t - count >
drivers/video/fbdev/core/fbcon.c
1843
if (t > 0)
drivers/video/fbdev/core/fbcon.c
1844
fbcon_redraw_move(vc, p, 0, t, count);
drivers/video/fbdev/core/fbcon.c
1845
ypan_up_redraw(vc, t, count);
drivers/video/fbdev/core/fbcon.c
1850
fbcon_redraw_move(vc, p, t + count, b - t - count, t);
drivers/video/fbdev/core/fbcon.c
1857
&& ((!scroll_partial && (b - t == vc->vc_rows))
drivers/video/fbdev/core/fbcon.c
1859
&& (b - t - count >
drivers/video/fbdev/core/fbcon.c
1861
if (t > 0)
drivers/video/fbdev/core/fbcon.c
1862
fbcon_bmove(vc, 0, 0, count, 0, t,
drivers/video/fbdev/core/fbcon.c
1870
fbcon_bmove(vc, t + count, 0, t, 0,
drivers/video/fbdev/core/fbcon.c
1871
b - t - count, vc->vc_cols);
drivers/video/fbdev/core/fbcon.c
1879
fbcon_redraw(vc, t, b - t - count,
drivers/video/fbdev/core/fbcon.c
1896
fbcon_redraw_blit(vc, info, p, b - 1, b - t - count,
drivers/video/fbdev/core/fbcon.c
1898
__fbcon_clear(vc, t, 0, count, vc->vc_cols);
drivers/video/fbdev/core/fbcon.c
1901
t),
drivers/video/fbdev/core/fbcon.c
1907
if (b - t - count > 3 * vc->vc_rows >> 2) {
drivers/video/fbdev/core/fbcon.c
1913
if (t > 0)
drivers/video/fbdev/core/fbcon.c
1914
fbcon_bmove(vc, count, 0, 0, 0, t,
drivers/video/fbdev/core/fbcon.c
1917
fbcon_bmove(vc, t, 0, t + count, 0,
drivers/video/fbdev/core/fbcon.c
1918
b - t - count, vc->vc_cols);
drivers/video/fbdev/core/fbcon.c
1921
__fbcon_clear(vc, t, 0, count, vc->vc_cols);
drivers/video/fbdev/core/fbcon.c
1926
&& ((!scroll_partial && (b - t == vc->vc_rows))
drivers/video/fbdev/core/fbcon.c
1928
&& (b - t - count >
drivers/video/fbdev/core/fbcon.c
1935
if (t > 0)
drivers/video/fbdev/core/fbcon.c
1936
fbcon_bmove(vc, count, 0, 0, 0, t,
drivers/video/fbdev/core/fbcon.c
1939
fbcon_bmove(vc, t, 0, t + count, 0,
drivers/video/fbdev/core/fbcon.c
1940
b - t - count, vc->vc_cols);
drivers/video/fbdev/core/fbcon.c
1943
__fbcon_clear(vc, t, 0, count, vc->vc_cols);
drivers/video/fbdev/core/fbcon.c
1948
&& ((!scroll_partial && (b - t == vc->vc_rows))
drivers/video/fbdev/core/fbcon.c
1950
&& (b - t - count >
drivers/video/fbdev/core/fbcon.c
1955
ypan_down_redraw(vc, t, count);
drivers/video/fbdev/core/fbcon.c
1956
if (t > 0)
drivers/video/fbdev/core/fbcon.c
1957
fbcon_redraw_move(vc, p, count, t, 0);
drivers/video/fbdev/core/fbcon.c
1959
fbcon_redraw_move(vc, p, t, b - t - count, t + count);
drivers/video/fbdev/core/fbcon.c
1960
__fbcon_clear(vc, t, 0, count, vc->vc_cols);
drivers/video/fbdev/core/fbcon.c
1965
fbcon_redraw(vc, b - 1, b - t - count,
drivers/video/fbdev/core/fbcon.c
1967
__fbcon_clear(vc, t, 0, count, vc->vc_cols);
drivers/video/fbdev/core/fbcon.c
1970
t),
drivers/video/fbdev/core/fbcon.c
1987
u16 t = 0;
drivers/video/fbdev/core/fbcon.c
1989
int ywrap = FBCON_SWAP(par->rotate, info->fix.ywrapstep, t);
drivers/video/fbdev/fm2fb.c
180
unsigned char t = FRAMEMASTER_ROM;
drivers/video/fbdev/fm2fb.c
183
t |= FRAMEMASTER_ENABLE | FRAMEMASTER_NOLACE;
drivers/video/fbdev/fm2fb.c
184
fm2fb_reg[0] = t;
drivers/video/fbdev/gxt4500.c
240
int pll_period, best_error, t, intf;
drivers/video/fbdev/gxt4500.c
261
t = par->refclk_ps * m * postdiv / n;
drivers/video/fbdev/gxt4500.c
262
t -= period_ps;
drivers/video/fbdev/gxt4500.c
263
if (t >= 0 && t < best_error) {
drivers/video/fbdev/gxt4500.c
268
best_error = t;
drivers/video/fbdev/metronomefb.c
170
static int load_waveform(u8 *mem, size_t size, int m, int t,
drivers/video/fbdev/metronomefb.c
228
if (mem[i] > t) {
drivers/video/fbdev/mmp/hw/mmp_spi.c
109
struct spi_transfer *t;
drivers/video/fbdev/mmp/hw/mmp_spi.c
112
list_for_each_entry(t, &m->transfers, transfer_list) {
drivers/video/fbdev/mmp/hw/mmp_spi.c
115
for (i = 0; i < t->len; i++)
drivers/video/fbdev/mmp/hw/mmp_spi.c
116
lcd_spi_write(spi, ((u8 *)t->tx_buf)[i]);
drivers/video/fbdev/mmp/hw/mmp_spi.c
119
for (i = 0; i < t->len/2; i++)
drivers/video/fbdev/mmp/hw/mmp_spi.c
120
lcd_spi_write(spi, ((u16 *)t->tx_buf)[i]);
drivers/video/fbdev/mmp/hw/mmp_spi.c
123
for (i = 0; i < t->len/4; i++)
drivers/video/fbdev/mmp/hw/mmp_spi.c
124
lcd_spi_write(spi, ((u32 *)t->tx_buf)[i]);
drivers/video/fbdev/nvidia/nv_type.h
17
#define BITMASK(t,b) (((unsigned)(1U << (((t)-(b)+1)))-1) << (b))
drivers/video/fbdev/omap/hwa742.c
637
struct extif_timings *t;
drivers/video/fbdev/omap/hwa742.c
652
t = &hwa742.reg_timings;
drivers/video/fbdev/omap/hwa742.c
653
memset(t, 0, sizeof(*t));
drivers/video/fbdev/omap/hwa742.c
654
t->clk_div = div;
drivers/video/fbdev/omap/hwa742.c
655
t->cs_on_time = 0;
drivers/video/fbdev/omap/hwa742.c
656
t->we_on_time = round_to_extif_ticks(t->cs_on_time + 2000, div);
drivers/video/fbdev/omap/hwa742.c
657
t->re_on_time = round_to_extif_ticks(t->cs_on_time + 2000, div);
drivers/video/fbdev/omap/hwa742.c
658
t->access_time = round_to_extif_ticks(t->re_on_time + 12200, div);
drivers/video/fbdev/omap/hwa742.c
659
t->we_off_time = round_to_extif_ticks(t->we_on_time + 1000, div);
drivers/video/fbdev/omap/hwa742.c
660
t->re_off_time = round_to_extif_ticks(t->re_on_time + 16000, div);
drivers/video/fbdev/omap/hwa742.c
661
t->cs_off_time = round_to_extif_ticks(t->re_off_time + 1000, div);
drivers/video/fbdev/omap/hwa742.c
662
t->we_cycle_time = round_to_extif_ticks(2 * systim + 2000, div);
drivers/video/fbdev/omap/hwa742.c
663
if (t->we_cycle_time < t->we_off_time)
drivers/video/fbdev/omap/hwa742.c
664
t->we_cycle_time = t->we_off_time;
drivers/video/fbdev/omap/hwa742.c
665
t->re_cycle_time = round_to_extif_ticks(2 * systim + 2000, div);
drivers/video/fbdev/omap/hwa742.c
666
if (t->re_cycle_time < t->re_off_time)
drivers/video/fbdev/omap/hwa742.c
667
t->re_cycle_time = t->re_off_time;
drivers/video/fbdev/omap/hwa742.c
668
t->cs_pulse_width = 0;
drivers/video/fbdev/omap/hwa742.c
671
t->cs_on_time, t->cs_off_time, t->re_on_time, t->re_off_time);
drivers/video/fbdev/omap/hwa742.c
673
t->we_on_time, t->we_off_time, t->re_cycle_time,
drivers/video/fbdev/omap/hwa742.c
674
t->we_cycle_time);
drivers/video/fbdev/omap/hwa742.c
676
t->access_time, t->cs_pulse_width);
drivers/video/fbdev/omap/hwa742.c
678
return hwa742.extif->convert_timings(t);
drivers/video/fbdev/omap/hwa742.c
683
struct extif_timings *t;
drivers/video/fbdev/omap/hwa742.c
699
t = &hwa742.lut_timings;
drivers/video/fbdev/omap/hwa742.c
700
memset(t, 0, sizeof(*t));
drivers/video/fbdev/omap/hwa742.c
702
t->clk_div = div;
drivers/video/fbdev/omap/hwa742.c
704
t->cs_on_time = 0;
drivers/video/fbdev/omap/hwa742.c
705
t->we_on_time = round_to_extif_ticks(t->cs_on_time + 2000, div);
drivers/video/fbdev/omap/hwa742.c
706
t->re_on_time = round_to_extif_ticks(t->cs_on_time + 2000, div);
drivers/video/fbdev/omap/hwa742.c
707
t->access_time = round_to_extif_ticks(t->re_on_time + 4 * systim +
drivers/video/fbdev/omap/hwa742.c
709
t->we_off_time = round_to_extif_ticks(t->we_on_time + 1000, div);
drivers/video/fbdev/omap/hwa742.c
710
t->re_off_time = round_to_extif_ticks(t->re_on_time + 4 * systim +
drivers/video/fbdev/omap/hwa742.c
712
t->cs_off_time = round_to_extif_ticks(t->re_off_time + 1000, div);
drivers/video/fbdev/omap/hwa742.c
713
t->we_cycle_time = round_to_extif_ticks(2 * systim + 2000, div);
drivers/video/fbdev/omap/hwa742.c
714
if (t->we_cycle_time < t->we_off_time)
drivers/video/fbdev/omap/hwa742.c
715
t->we_cycle_time = t->we_off_time;
drivers/video/fbdev/omap/hwa742.c
716
t->re_cycle_time = round_to_extif_ticks(2000 + 4 * systim + 26000, div);
drivers/video/fbdev/omap/hwa742.c
717
if (t->re_cycle_time < t->re_off_time)
drivers/video/fbdev/omap/hwa742.c
718
t->re_cycle_time = t->re_off_time;
drivers/video/fbdev/omap/hwa742.c
719
t->cs_pulse_width = 0;
drivers/video/fbdev/omap/hwa742.c
722
t->cs_on_time, t->cs_off_time, t->re_on_time, t->re_off_time);
drivers/video/fbdev/omap/hwa742.c
724
t->we_on_time, t->we_off_time, t->re_cycle_time,
drivers/video/fbdev/omap/hwa742.c
725
t->we_cycle_time);
drivers/video/fbdev/omap/hwa742.c
727
t->access_time, t->cs_pulse_width);
drivers/video/fbdev/omap/hwa742.c
729
return hwa742.extif->convert_timings(t);
drivers/video/fbdev/omap/sossi.c
120
static int calc_rd_timings(struct extif_timings *t)
drivers/video/fbdev/omap/sossi.c
124
int div = t->clk_div;
drivers/video/fbdev/omap/sossi.c
130
reon = ps_to_sossi_ticks(t->re_on_time, div);
drivers/video/fbdev/omap/sossi.c
135
reoff = ps_to_sossi_ticks(t->re_off_time, div);
drivers/video/fbdev/omap/sossi.c
144
recyc = ps_to_sossi_ticks(t->re_cycle_time, div);
drivers/video/fbdev/omap/sossi.c
155
actim = ps_to_sossi_ticks(t->access_time, div);
drivers/video/fbdev/omap/sossi.c
165
t->tim[0] = tw0 - 1;
drivers/video/fbdev/omap/sossi.c
166
t->tim[1] = tw1 - 1;
drivers/video/fbdev/omap/sossi.c
171
static int calc_wr_timings(struct extif_timings *t)
drivers/video/fbdev/omap/sossi.c
175
int div = t->clk_div;
drivers/video/fbdev/omap/sossi.c
181
weon = ps_to_sossi_ticks(t->we_on_time, div);
drivers/video/fbdev/omap/sossi.c
186
weoff = ps_to_sossi_ticks(t->we_off_time, div);
drivers/video/fbdev/omap/sossi.c
193
wecyc = ps_to_sossi_ticks(t->we_cycle_time, div);
drivers/video/fbdev/omap/sossi.c
204
t->tim[2] = tw0 - 1;
drivers/video/fbdev/omap/sossi.c
205
t->tim[3] = tw1 - 1;
drivers/video/fbdev/omap/sossi.c
314
static int sossi_convert_timings(struct extif_timings *t)
drivers/video/fbdev/omap/sossi.c
317
int div = t->clk_div;
drivers/video/fbdev/omap/sossi.c
319
t->converted = 0;
drivers/video/fbdev/omap/sossi.c
325
if ((r = calc_rd_timings(t)) < 0)
drivers/video/fbdev/omap/sossi.c
328
if ((r = calc_wr_timings(t)) < 0)
drivers/video/fbdev/omap/sossi.c
331
t->tim[4] = div;
drivers/video/fbdev/omap/sossi.c
333
t->converted = 1;
drivers/video/fbdev/omap/sossi.c
338
static void sossi_set_timings(const struct extif_timings *t)
drivers/video/fbdev/omap/sossi.c
340
BUG_ON(!t->converted);
drivers/video/fbdev/omap/sossi.c
342
sossi.clk_tw0[RD_ACCESS] = t->tim[0];
drivers/video/fbdev/omap/sossi.c
343
sossi.clk_tw1[RD_ACCESS] = t->tim[1];
drivers/video/fbdev/omap/sossi.c
345
sossi.clk_tw0[WR_ACCESS] = t->tim[2];
drivers/video/fbdev/omap/sossi.c
346
sossi.clk_tw1[WR_ACCESS] = t->tim[3];
drivers/video/fbdev/omap/sossi.c
348
sossi.clk_div = t->tim[4];
drivers/video/fbdev/omap2/omapfb/displays/panel-dsi-cm.c
441
unsigned long t;
drivers/video/fbdev/omap2/omapfb/displays/panel-dsi-cm.c
444
r = kstrtoul(buf, 0, &t);
drivers/video/fbdev/omap2/omapfb/displays/panel-dsi-cm.c
453
if (t)
drivers/video/fbdev/omap2/omapfb/displays/panel-dsi-cm.c
474
unsigned t;
drivers/video/fbdev/omap2/omapfb/displays/panel-dsi-cm.c
477
t = ddata->ulps_enabled;
drivers/video/fbdev/omap2/omapfb/displays/panel-dsi-cm.c
480
return sysfs_emit(buf, "%u\n", t);
drivers/video/fbdev/omap2/omapfb/displays/panel-dsi-cm.c
489
unsigned long t;
drivers/video/fbdev/omap2/omapfb/displays/panel-dsi-cm.c
492
r = kstrtoul(buf, 0, &t);
drivers/video/fbdev/omap2/omapfb/displays/panel-dsi-cm.c
497
ddata->ulps_timeout = t;
drivers/video/fbdev/omap2/omapfb/displays/panel-dsi-cm.c
519
unsigned t;
drivers/video/fbdev/omap2/omapfb/displays/panel-dsi-cm.c
522
t = ddata->ulps_timeout;
drivers/video/fbdev/omap2/omapfb/displays/panel-dsi-cm.c
525
return sysfs_emit(buf, "%u\n", t);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
392
unsigned long t;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
408
t = msecs_to_jiffies(500);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
409
r = wait_for_completion_timeout(&extra_updated_completion, t);
drivers/video/fbdev/omap2/omapfb/dss/dispc.c
2094
const struct omap_video_timings *t, u16 pos_x,
drivers/video/fbdev/omap2/omapfb/dss/dispc.c
2104
nonactive = t->x_res + t->hfp + t->hsw + t->hbp - out_width;
drivers/video/fbdev/omap2/omapfb/dss/dispc.c
2111
blank = div_u64((u64)(t->hbp + t->hsw + t->hfp) * lclk, pclk);
drivers/video/fbdev/omap2/omapfb/dss/dispc.c
3125
struct omap_video_timings t = *timings;
drivers/video/fbdev/omap2/omapfb/dss/dispc.c
3127
DSSDBG("channel %d xres %u yres %u\n", channel, t.x_res, t.y_res);
drivers/video/fbdev/omap2/omapfb/dss/dispc.c
3129
if (!dispc_mgr_timings_ok(channel, &t)) {
drivers/video/fbdev/omap2/omapfb/dss/dispc.c
3135
_dispc_mgr_set_lcd_timings(channel, t.hsw, t.hfp, t.hbp, t.vsw,
drivers/video/fbdev/omap2/omapfb/dss/dispc.c
3136
t.vfp, t.vbp, t.vsync_level, t.hsync_level,
drivers/video/fbdev/omap2/omapfb/dss/dispc.c
3137
t.data_pclk_edge, t.de_level, t.sync_pclk_edge);
drivers/video/fbdev/omap2/omapfb/dss/dispc.c
3139
xtot = t.x_res + t.hfp + t.hsw + t.hbp;
drivers/video/fbdev/omap2/omapfb/dss/dispc.c
3140
ytot = t.y_res + t.vfp + t.vsw + t.vbp;
drivers/video/fbdev/omap2/omapfb/dss/dispc.c
3147
t.hsw, t.hfp, t.hbp, t.vsw, t.vfp, t.vbp);
drivers/video/fbdev/omap2/omapfb/dss/dispc.c
3149
t.vsync_level, t.hsync_level, t.data_pclk_edge,
drivers/video/fbdev/omap2/omapfb/dss/dispc.c
3150
t.de_level, t.sync_pclk_edge);
drivers/video/fbdev/omap2/omapfb/dss/dispc.c
3154
if (t.interlace)
drivers/video/fbdev/omap2/omapfb/dss/dispc.c
3155
t.y_res /= 2;
drivers/video/fbdev/omap2/omapfb/dss/dispc.c
3158
dispc_mgr_set_size(channel, t.x_res, t.y_res);
drivers/video/fbdev/omap2/omapfb/dss/display-sysfs.c
100
t.y_res, t.vfp, t.vbp, t.vsw);
drivers/video/fbdev/omap2/omapfb/dss/display-sysfs.c
106
struct omap_video_timings t = dssdev->panel.timings;
drivers/video/fbdev/omap2/omapfb/dss/display-sysfs.c
115
t = omap_dss_pal_timings;
drivers/video/fbdev/omap2/omapfb/dss/display-sysfs.c
118
t = omap_dss_ntsc_timings;
drivers/video/fbdev/omap2/omapfb/dss/display-sysfs.c
123
&t.pixelclock,
drivers/video/fbdev/omap2/omapfb/dss/display-sysfs.c
124
&t.x_res, &t.hfp, &t.hbp, &t.hsw,
drivers/video/fbdev/omap2/omapfb/dss/display-sysfs.c
125
&t.y_res, &t.vfp, &t.vbp, &t.vsw) != 9)
drivers/video/fbdev/omap2/omapfb/dss/display-sysfs.c
128
r = dssdev->driver->check_timings(dssdev, &t);
drivers/video/fbdev/omap2/omapfb/dss/display-sysfs.c
133
dssdev->driver->set_timings(dssdev, &t);
drivers/video/fbdev/omap2/omapfb/dss/display-sysfs.c
90
struct omap_video_timings t;
drivers/video/fbdev/omap2/omapfb/dss/display-sysfs.c
95
dssdev->driver->get_timings(dssdev, &t);
drivers/video/fbdev/omap2/omapfb/dss/display-sysfs.c
98
t.pixelclock,
drivers/video/fbdev/omap2/omapfb/dss/display-sysfs.c
99
t.x_res, t.hfp, t.hbp, t.hsw,
drivers/video/fbdev/omap2/omapfb/dss/dpi.c
327
struct omap_video_timings *t = &dpi->timings;
drivers/video/fbdev/omap2/omapfb/dss/dpi.c
334
r = dpi_set_dsi_clk(dpi, mgr->id, t->pixelclock, &fck,
drivers/video/fbdev/omap2/omapfb/dss/dpi.c
337
r = dpi_set_dispc_clk(dpi, t->pixelclock, &fck,
drivers/video/fbdev/omap2/omapfb/dss/dpi.c
344
if (pck != t->pixelclock) {
drivers/video/fbdev/omap2/omapfb/dss/dpi.c
346
t->pixelclock, pck);
drivers/video/fbdev/omap2/omapfb/dss/dpi.c
348
t->pixelclock = pck;
drivers/video/fbdev/omap2/omapfb/dss/dpi.c
351
dss_mgr_set_timings(mgr, t);
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
1342
int t = 0;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
1354
if (++t > 1000) {
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
1754
int t = 0;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
1762
if (++t > 1000) {
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
1826
unsigned t;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
1828
for (t = 0; t < dsi->num_lanes_supported; ++t)
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
1829
if (dsi->lanes[t].function == functions[i])
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
1832
if (t == dsi->num_lanes_supported)
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
1835
lane_number = t;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
1836
polarity = dsi->lanes[t].polarity;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
2002
int t, i;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
2016
t = 100000;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
2032
if (--t == 0) {
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4319
const struct omap_dss_dsi_videomode_timings *t)
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4321
unsigned long byteclk = t->hsclk / 4;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4324
wc = DIV_ROUND_UP(t->hact * t->bitspp, 8);
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4325
pps = DIV_ROUND_UP(wc + 6, t->ndl); /* pixel packet size */
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4326
bl = t->hss + t->hsa + t->hse + t->hbp + t->hfp;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4335
t->hss, t->hsa, t->hse, t->hbp, pps, t->hfp,
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4337
TO_DSI_T(t->hss),
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4338
TO_DSI_T(t->hsa),
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4339
TO_DSI_T(t->hse),
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4340
TO_DSI_T(t->hbp),
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4342
TO_DSI_T(t->hfp),
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4351
static void print_dispc_vm(const char *str, const struct omap_video_timings *t)
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4353
unsigned long pck = t->pixelclock;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4356
hact = t->x_res;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4357
bl = t->hsw + t->hbp + t->hfp;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4366
t->hsw, t->hbp, hact, t->hfp,
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4368
TO_DISPC_T(t->hsw),
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4369
TO_DISPC_T(t->hbp),
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4371
TO_DISPC_T(t->hfp),
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4380
const struct omap_dss_dsi_videomode_timings *t)
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4383
unsigned long byteclk = t->hsclk / 4;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4388
dsi_tput = (u64)byteclk * t->ndl * 8;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4389
pck = (u32)div64_u64(dsi_tput, t->bitspp);
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4390
dsi_hact = DIV_ROUND_UP(DIV_ROUND_UP(t->hact * t->bitspp, 8) + 6, t->ndl);
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4391
dsi_htot = t->hss + t->hsa + t->hse + t->hbp + dsi_hact + t->hfp;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4394
vm.hsw = div64_u64((u64)(t->hsa + t->hse) * pck, byteclk);
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4395
vm.hbp = div64_u64((u64)t->hbp * pck, byteclk);
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4396
vm.hfp = div64_u64((u64)t->hfp * pck, byteclk);
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4397
vm.x_res = t->hact;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4407
struct omap_video_timings *t = &ctx->dispc_vm;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4414
*t = *ctx->config->timings;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4415
t->pixelclock = pck;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4416
t->x_res = ctx->config->timings->x_res;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4417
t->y_res = ctx->config->timings->y_res;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4418
t->hsw = t->hfp = t->hbp = t->vsw = 1;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4419
t->vfp = t->vbp = 0;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4603
int t;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4606
t = 1 - hfp;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4607
hbp = max(hbp - t, 1);
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4612
t = 1 - hfp;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4613
hsa = max(hsa - t, 1);
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4662
int t;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4665
t = 1 - hfp;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4666
hbp = max(hbp - t, 1);
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4671
t = 1 - hfp;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4672
hsa = max(hsa - t, 1);
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4753
const struct omap_video_timings *t = cfg->timings;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4769
ctx->req_pck_min = t->pixelclock - 1000;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4770
ctx->req_pck_nom = t->pixelclock;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
4771
ctx->req_pck_max = t->pixelclock + 1000;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
497
int t;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
500
t = 100;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
501
while (t-- > 0) {
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
552
ktime_t t, setup_time, trans_time;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
559
t = ktime_get();
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
566
trans_time = ktime_sub(t, dsi->perf_start_time);
drivers/video/fbdev/omap2/omapfb/dss/hdmi.h
267
u32 t = 0, v;
drivers/video/fbdev/omap2/omapfb/dss/hdmi.h
269
if (t++ > 10000)
drivers/video/fbdev/omap2/omapfb/dss/hdmi4_core.c
127
int t;
drivers/video/fbdev/omap2/omapfb/dss/hdmi4_core.c
135
t = 0;
drivers/video/fbdev/omap2/omapfb/dss/hdmi4_core.c
138
if (t++ > 10000) {
drivers/video/fbdev/omap2/omapfb/dss/pll.c
191
int t;
drivers/video/fbdev/omap2/omapfb/dss/pll.c
194
t = 100;
drivers/video/fbdev/omap2/omapfb/dss/pll.c
195
while (t-- > 0) {
drivers/video/fbdev/omap2/omapfb/dss/pll.c
226
int t = 100;
drivers/video/fbdev/omap2/omapfb/dss/pll.c
228
while (t-- > 0) {
drivers/video/fbdev/omap2/omapfb/dss/sdi.c
122
struct omap_video_timings *t = &sdi.timings;
drivers/video/fbdev/omap2/omapfb/dss/sdi.c
142
t->data_pclk_edge = OMAPDSS_DRIVE_SIG_RISING_EDGE;
drivers/video/fbdev/omap2/omapfb/dss/sdi.c
143
t->sync_pclk_edge = OMAPDSS_DRIVE_SIG_RISING_EDGE;
drivers/video/fbdev/omap2/omapfb/dss/sdi.c
145
r = sdi_calc_clock_div(t->pixelclock, &fck, &dispc_cinfo);
drivers/video/fbdev/omap2/omapfb/dss/sdi.c
153
if (pck != t->pixelclock) {
drivers/video/fbdev/omap2/omapfb/dss/sdi.c
155
t->pixelclock, pck);
drivers/video/fbdev/omap2/omapfb/dss/sdi.c
157
t->pixelclock = pck;
drivers/video/fbdev/omap2/omapfb/dss/sdi.c
161
dss_mgr_set_timings(out->manager, t);
drivers/video/fbdev/omap2/omapfb/dss/venc.c
329
int t = 1000;
drivers/video/fbdev/omap2/omapfb/dss/venc.c
333
if (--t == 0) {
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
122
int t = x * 3 / w;
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
126
if (t == 0)
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
128
else if (t == 1)
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
130
else if (t == 2)
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
133
if (t == 0)
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
135
else if (t == 1)
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
137
else if (t == 2)
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
2200
struct omap_video_timings *t)
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
2203
display->driver->get_timings(display, t);
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
2205
t->data_pclk_edge = OMAPDSS_DRIVE_SIG_RISING_EDGE;
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
2206
t->de_level = OMAPDSS_SIG_ACTIVE_HIGH;
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
2207
t->sync_pclk_edge = OMAPDSS_DRIVE_SIG_FALLING_EDGE;
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
2210
t->x_res = m->xres;
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
2211
t->y_res = m->yres;
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
2212
t->pixelclock = PICOS2KHZ(m->pixclock) * 1000;
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
2213
t->hsw = m->hsync_len;
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
2214
t->hfp = m->right_margin;
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
2215
t->hbp = m->left_margin;
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
2216
t->vsw = m->vsync_len;
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
2217
t->vfp = m->lower_margin;
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
2218
t->vbp = m->upper_margin;
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
2219
t->hsync_level = m->sync & FB_SYNC_HOR_HIGH_ACT ?
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
2222
t->vsync_level = m->sync & FB_SYNC_VERT_HIGH_ACT ?
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
2225
t->interlace = m->vmode & FB_VMODE_INTERLACED;
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
2259
struct omap_video_timings t;
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
2274
fb_videomode_to_omap_timings(m, display, &t);
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
2276
r = display->driver->check_timings(display, &t);
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
2544
struct omap_video_timings t;
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
2546
r = omapfb_find_best_mode(def_display, &t);
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
2549
def_display->driver->set_timings(def_display, &t);
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
138
int t;
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
143
for (t = 0; t < ofbi->num_overlays; t++) {
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
144
struct omap_overlay *ovl = ofbi->overlays[t];
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
152
t == 0 ? "" : ",", ovlnum);
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
166
int i, t;
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
171
for (t = 0; t < ofbi->num_overlays; t++) {
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
172
if (ofbi->overlays[t] == ovl)
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
244
int t, found;
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
250
for (t = 0; t < num_ovls; ++t) {
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
251
if (ovl == ovls[t]) {
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
271
for (t = i + 1; t < ofbi->num_overlays; t++) {
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
272
ofbi->rotation[t-1] = ofbi->rotation[t];
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
273
ofbi->overlays[t-1] = ofbi->overlays[t];
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
281
int t, found;
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
287
for (t = 0; t < ofbi->num_overlays; ++t) {
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
288
if (ovl == ofbi->overlays[t]) {
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
327
int t;
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
331
for (t = 0; t < ofbi->num_overlays; t++) {
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
333
t == 0 ? "" : ",", ofbi->rotation[t]);
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
562
int t;
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
568
for (t = 0; t < ARRAY_SIZE(omapfb_attrs); t++) {
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
569
r = device_create_file(dev, &omapfb_attrs[t]);
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
584
int i, t;
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
594
for (t = 0; t < ARRAY_SIZE(omapfb_attrs); t++)
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
595
device_remove_file(dev, &omapfb_attrs[t]);
drivers/video/fbdev/pxafb.c
1214
unsigned int t = (time_ns * (lcd_clk / 1000000) / 1000);
drivers/video/fbdev/pxafb.c
1215
return (t == 0) ? 1 : t;
drivers/video/fbdev/pxafb.c
310
#define SET_PIXFMT(v, r, g, b, t) \
drivers/video/fbdev/pxafb.c
312
(v)->transp.offset = (t) ? (r) + (g) + (b) : 0; \
drivers/video/fbdev/pxafb.c
313
(v)->transp.length = (t) ? (t) : 0; \
drivers/video/fbdev/riva/nvreg.h
30
#define BITMASK(t,b) (((unsigned)(1U << (((t)-(b)+1)))-1) << (b))
drivers/video/fbdev/savage/savagefb.h
116
#define BCI_CLIP_TL(t, l) ((((t) << 16) | (l)) & 0x0FFF0FFF)
drivers/video/fbdev/sstfb.c
1000
dac_i_write(DACREG_AC1_I, t->p << 6 | t->n);
drivers/video/fbdev/sstfb.c
1005
dac_i_write(DACREG_BD0_I, t->m);
drivers/video/fbdev/sstfb.c
1006
dac_i_write(DACREG_BD1_I, t->p << 6 | t->n);
drivers/video/fbdev/sstfb.c
1024
const struct pll_timing *t, const int clock)
drivers/video/fbdev/sstfb.c
1034
sst_dac_write(DACREG_ICS_PLLDATA, t->m);
drivers/video/fbdev/sstfb.c
1035
sst_dac_write(DACREG_ICS_PLLDATA, t->p << 5 | t->n);
drivers/video/fbdev/sstfb.c
1045
sst_dac_write(DACREG_ICS_PLLDATA, t->m);
drivers/video/fbdev/sstfb.c
1046
sst_dac_write(DACREG_ICS_PLLDATA, t->p << 5 | t->n);
drivers/video/fbdev/sstfb.c
292
static int sst_calc_pll(const int freq, int *freq_out, struct pll_timing *t)
drivers/video/fbdev/sstfb.c
323
t->p = p;
drivers/video/fbdev/sstfb.c
324
t->n = best_n;
drivers/video/fbdev/sstfb.c
325
t->m = best_m;
drivers/video/fbdev/sstfb.c
326
*freq_out = (DAC_FREF * (t->m + 2)) / ((1 << t->p) * (t->n + 2));
drivers/video/fbdev/sstfb.c
328
t->m, t->n, t->p, *freq_out);
drivers/video/fbdev/sstfb.c
970
const struct pll_timing *t, const int clock)
drivers/video/fbdev/sstfb.c
999
dac_i_write(DACREG_AC0_I, t->m);
drivers/video/fbdev/uvesafb.c
103
memcpy(&task->t, utask, sizeof(*utask));
drivers/video/fbdev/uvesafb.c
105
if (task->t.buf_len && task->buf)
drivers/video/fbdev/uvesafb.c
106
memcpy(task->buf, utask + 1, task->t.buf_len);
drivers/video/fbdev/uvesafb.c
1139
task->t.regs.eax = 0x4f10;
drivers/video/fbdev/uvesafb.c
1142
task->t.regs.ebx = 0x0001;
drivers/video/fbdev/uvesafb.c
1145
task->t.regs.ebx = 0x0101; /* standby */
drivers/video/fbdev/uvesafb.c
1148
task->t.regs.ebx = 0x0401; /* powerdown */
drivers/video/fbdev/uvesafb.c
1155
if (err || (task->t.regs.eax & 0xffff) != 0x004f)
drivers/video/fbdev/uvesafb.c
1199
task->t.regs.eax = 0x0003;
drivers/video/fbdev/uvesafb.c
1236
task->t.regs.eax = 0x4f02;
drivers/video/fbdev/uvesafb.c
1237
task->t.regs.ebx = mode->mode_id | 0x4000; /* use LFB */
drivers/video/fbdev/uvesafb.c
1241
task->t.regs.ebx |= 0x0800; /* use CRTC data */
drivers/video/fbdev/uvesafb.c
1242
task->t.flags = TF_BUF_ESDI;
drivers/video/fbdev/uvesafb.c
1273
task->t.buf_len = sizeof(struct vbe_crtc_ib);
drivers/video/fbdev/uvesafb.c
1277
if (err || (task->t.regs.eax & 0xffff) != 0x004f) {
drivers/video/fbdev/uvesafb.c
1284
task->t.regs.eax, err);
drivers/video/fbdev/uvesafb.c
1292
task->t.regs.eax, err);
drivers/video/fbdev/uvesafb.c
1303
task->t.regs.eax = 0x4f08;
drivers/video/fbdev/uvesafb.c
1304
task->t.regs.ebx = 0x0800;
drivers/video/fbdev/uvesafb.c
1307
if (err || (task->t.regs.eax & 0xffff) != 0x004f ||
drivers/video/fbdev/uvesafb.c
1308
((task->t.regs.ebx & 0xff00) >> 8) != 8) {
drivers/video/fbdev/uvesafb.c
149
int len = sizeof(task->t) + task->t.buf_len;
drivers/video/fbdev/uvesafb.c
173
memcpy(m + 1, &task->t, sizeof(task->t));
drivers/video/fbdev/uvesafb.c
176
memcpy((u8 *)(m + 1) + sizeof(task->t), task->buf, task->t.buf_len);
drivers/video/fbdev/uvesafb.c
1922
task->t.flags = TF_EXIT;
drivers/video/fbdev/uvesafb.c
216
if (!err && !(task->t.flags & TF_EXIT))
drivers/video/fbdev/uvesafb.c
371
task->t.regs.eax = 0x4f04;
drivers/video/fbdev/uvesafb.c
372
task->t.regs.ecx = 0x000f;
drivers/video/fbdev/uvesafb.c
373
task->t.regs.edx = 0x0001;
drivers/video/fbdev/uvesafb.c
374
task->t.flags = TF_BUF_RET | TF_BUF_ESBX;
drivers/video/fbdev/uvesafb.c
375
task->t.buf_len = par->vbe_state_size;
drivers/video/fbdev/uvesafb.c
379
if (err || (task->t.regs.eax & 0xffff) != 0x004f) {
drivers/video/fbdev/uvesafb.c
381
task->t.regs.eax, err);
drivers/video/fbdev/uvesafb.c
402
task->t.regs.eax = 0x4f04;
drivers/video/fbdev/uvesafb.c
403
task->t.regs.ecx = 0x000f;
drivers/video/fbdev/uvesafb.c
404
task->t.regs.edx = 0x0002;
drivers/video/fbdev/uvesafb.c
405
task->t.buf_len = par->vbe_state_size;
drivers/video/fbdev/uvesafb.c
406
task->t.flags = TF_BUF_ESBX;
drivers/video/fbdev/uvesafb.c
410
if (err || (task->t.regs.eax & 0xffff) != 0x004f)
drivers/video/fbdev/uvesafb.c
412
task->t.regs.eax, err);
drivers/video/fbdev/uvesafb.c
422
task->t.regs.eax = 0x4f00;
drivers/video/fbdev/uvesafb.c
423
task->t.flags = TF_VBEIB;
drivers/video/fbdev/uvesafb.c
424
task->t.buf_len = sizeof(struct vbe_ib);
drivers/video/fbdev/uvesafb.c
429
if (err || (task->t.regs.eax & 0xffff) != 0x004f) {
drivers/video/fbdev/uvesafb.c
431
(u32)task->t.regs.eax, err);
drivers/video/fbdev/uvesafb.c
500
task->t.regs.eax = 0x4f01;
drivers/video/fbdev/uvesafb.c
501
task->t.regs.ecx = (u32) *mode;
drivers/video/fbdev/uvesafb.c
502
task->t.flags = TF_BUF_RET | TF_BUF_ESDI;
drivers/video/fbdev/uvesafb.c
503
task->t.buf_len = sizeof(struct vbe_mode_ib);
drivers/video/fbdev/uvesafb.c
507
if (err || (task->t.regs.eax & 0xffff) != 0x004f) {
drivers/video/fbdev/uvesafb.c
509
*mode, (u32)task->t.regs.eax, err);
drivers/video/fbdev/uvesafb.c
558
task->t.regs.eax = 0x4f0a;
drivers/video/fbdev/uvesafb.c
559
task->t.regs.ebx = 0x0;
drivers/video/fbdev/uvesafb.c
564
if ((task->t.regs.eax & 0xffff) != 0x4f || task->t.regs.es < 0xc000) {
drivers/video/fbdev/uvesafb.c
567
par->pmi_base = (u16 *)phys_to_virt(((u32)task->t.regs.es << 4)
drivers/video/fbdev/uvesafb.c
568
+ task->t.regs.edi);
drivers/video/fbdev/uvesafb.c
572
(u16)task->t.regs.es, (u16)task->t.regs.edi);
drivers/video/fbdev/uvesafb.c
621
task->t.regs.eax = 0x4f15;
drivers/video/fbdev/uvesafb.c
622
task->t.regs.ebx = 0;
drivers/video/fbdev/uvesafb.c
623
task->t.regs.ecx = 0;
drivers/video/fbdev/uvesafb.c
624
task->t.buf_len = 0;
drivers/video/fbdev/uvesafb.c
625
task->t.flags = 0;
drivers/video/fbdev/uvesafb.c
629
if ((task->t.regs.eax & 0xffff) != 0x004f || err)
drivers/video/fbdev/uvesafb.c
632
if ((task->t.regs.ebx & 0x3) == 3) {
drivers/video/fbdev/uvesafb.c
634
} else if ((task->t.regs.ebx & 0x3) == 2) {
drivers/video/fbdev/uvesafb.c
636
} else if ((task->t.regs.ebx & 0x3) == 1) {
drivers/video/fbdev/uvesafb.c
643
task->t.regs.eax = 0x4f15;
drivers/video/fbdev/uvesafb.c
644
task->t.regs.ebx = 1;
drivers/video/fbdev/uvesafb.c
645
task->t.regs.ecx = task->t.regs.edx = 0;
drivers/video/fbdev/uvesafb.c
646
task->t.flags = TF_BUF_RET | TF_BUF_ESDI;
drivers/video/fbdev/uvesafb.c
647
task->t.buf_len = EDID_LENGTH;
drivers/video/fbdev/uvesafb.c
654
if ((task->t.regs.eax & 0xffff) == 0x004f && !err) {
drivers/video/fbdev/uvesafb.c
764
task->t.regs.eax = 0x4f04;
drivers/video/fbdev/uvesafb.c
765
task->t.regs.ecx = 0x000f;
drivers/video/fbdev/uvesafb.c
766
task->t.regs.edx = 0x0000;
drivers/video/fbdev/uvesafb.c
767
task->t.flags = 0;
drivers/video/fbdev/uvesafb.c
771
if (err || (task->t.regs.eax & 0xffff) != 0x004f) {
drivers/video/fbdev/uvesafb.c
773
task->t.regs.eax, err);
drivers/video/fbdev/uvesafb.c
778
par->vbe_state_size = 64 * (task->t.regs.ebx & 0xffff);
drivers/video/fbdev/uvesafb.c
94
if (task->t.buf_len < utask->buf_len ||
drivers/video/fbdev/uvesafb.c
970
task->t.regs.eax = 0x4f09;
drivers/video/fbdev/uvesafb.c
971
task->t.regs.ebx = 0x0;
drivers/video/fbdev/uvesafb.c
972
task->t.regs.ecx = count;
drivers/video/fbdev/uvesafb.c
973
task->t.regs.edx = start;
drivers/video/fbdev/uvesafb.c
974
task->t.flags = TF_BUF_ESDI;
drivers/video/fbdev/uvesafb.c
975
task->t.buf_len = sizeof(struct uvesafb_pal_entry) * count;
drivers/video/fbdev/uvesafb.c
979
if ((task->t.regs.eax & 0xffff) != 0x004f)
drivers/virt/vboxguest/vboxguest_core.c
419
static void vbg_heartbeat_timer(struct timer_list *t)
drivers/virt/vboxguest/vboxguest_core.c
421
struct vbg_dev *gdev = timer_container_of(gdev, t, heartbeat_timer);
drivers/w1/slaves/w1_therm.c
698
s16 t;
drivers/w1/slaves/w1_therm.c
708
t = (s16) bv; /* Degrees, lowest bit is 2^-6 */
drivers/w1/slaves/w1_therm.c
709
return (int)t * 1000 / 64; /* Sign-extend to int; millidegrees */
drivers/w1/slaves/w1_therm.c
711
t = (s16)bv; /* Degrees, lowest bit is 2^-4 */
drivers/w1/slaves/w1_therm.c
712
return (int)t * 1000 / 16; /* Sign-extend to int; millidegrees */
drivers/w1/slaves/w1_therm.c
725
int t, h;
drivers/w1/slaves/w1_therm.c
733
t = ((s32)rom[0] >> 1)*1000;
drivers/w1/slaves/w1_therm.c
735
t = 1000*(-1*(s32)(0x100-rom[0]) >> 1);
drivers/w1/slaves/w1_therm.c
737
t -= 250;
drivers/w1/slaves/w1_therm.c
740
t += h;
drivers/w1/slaves/w1_therm.c
742
return t;
drivers/w1/slaves/w1_therm.c
758
s16 t;
drivers/w1/slaves/w1_therm.c
769
t = (s16)bv; /* Degrees, lowest bit is 2^-4 */
drivers/w1/slaves/w1_therm.c
770
return (int)t * 1000 / 16; /* Sign-extend to int; millidegrees */
drivers/watchdog/advantech_ec_wdt.c
111
wdd->timeout = t;
drivers/watchdog/advantech_ec_wdt.c
86
static int adv_ec_wdt_set_timeout(struct watchdog_device *wdd, unsigned int t)
drivers/watchdog/advantech_ec_wdt.c
91
val = t * 10;
drivers/watchdog/advantechwdt.c
101
timeout = t;
drivers/watchdog/advantechwdt.c
97
static int advwdt_set_heartbeat(int t)
drivers/watchdog/advantechwdt.c
99
if (t < 1 || t > 63)
drivers/watchdog/alim1535_wdt.c
104
static int ali_settimer(int t)
drivers/watchdog/alim1535_wdt.c
106
if (t < 0)
drivers/watchdog/alim1535_wdt.c
108
else if (t < 60)
drivers/watchdog/alim1535_wdt.c
109
ali_timeout_bits = t|(1 << 6);
drivers/watchdog/alim1535_wdt.c
110
else if (t < 3600)
drivers/watchdog/alim1535_wdt.c
111
ali_timeout_bits = (t / 60)|(1 << 7);
drivers/watchdog/alim1535_wdt.c
112
else if (t < 18000)
drivers/watchdog/alim1535_wdt.c
113
ali_timeout_bits = (t / 300)|(1 << 6)|(1 << 7);
drivers/watchdog/alim1535_wdt.c
117
timeout = t;
drivers/watchdog/at91sam9_wdt.c
120
static void at91_ping(struct timer_list *t)
drivers/watchdog/at91sam9_wdt.c
122
struct at91wdt *wdt = timer_container_of(wdt, t, timer);
drivers/watchdog/at91sam9_wdt.c
49
#define ticks_to_hz_rounddown(t) ((((t) + 1) * HZ) >> 8)
drivers/watchdog/at91sam9_wdt.c
50
#define ticks_to_hz_roundup(t) (((((t) + 1) * HZ) + 255) >> 8)
drivers/watchdog/at91sam9_wdt.c
51
#define ticks_to_secs(t) (((t) + 1) >> 8)
drivers/watchdog/ath79_wdt.c
189
int t;
drivers/watchdog/ath79_wdt.c
211
err = get_user(t, p);
drivers/watchdog/ath79_wdt.c
215
err = ath79_wdt_set_timeout(t);
drivers/watchdog/bcm47xx_wdt.c
106
static void bcm47xx_wdt_soft_timer_tick(struct timer_list *t)
drivers/watchdog/bcm47xx_wdt.c
108
struct bcm47xx_wdt *wdt = timer_container_of(wdt, t, soft_timer);
drivers/watchdog/bcm7038_wdt.c
100
wdog->timeout = t;
drivers/watchdog/bcm7038_wdt.c
96
unsigned int t)
drivers/watchdog/bcm_kona_wdt.c
210
unsigned int t)
drivers/watchdog/bcm_kona_wdt.c
212
wdog->timeout = t;
drivers/watchdog/cros_ec_wdt.c
92
static int cros_ec_wdt_set_timeout(struct watchdog_device *wdd, unsigned int t)
drivers/watchdog/cros_ec_wdt.c
97
wdd->timeout = t;
drivers/watchdog/digicolor_wdt.c
78
static int dc_wdt_set_timeout(struct watchdog_device *wdog, unsigned int t)
drivers/watchdog/digicolor_wdt.c
82
dc_wdt_set(wdt, t * clk_get_rate(wdt->clk));
drivers/watchdog/digicolor_wdt.c
83
wdog->timeout = t;
drivers/watchdog/ebc-c384_wdt.c
42
unsigned t = wdev->timeout;
drivers/watchdog/ebc-c384_wdt.c
45
if (t > 255)
drivers/watchdog/ebc-c384_wdt.c
46
t = DIV_ROUND_UP(t, 60);
drivers/watchdog/ebc-c384_wdt.c
48
outb(t, PET_ADDR);
drivers/watchdog/ebc-c384_wdt.c
60
static int ebc_c384_wdt_set_timeout(struct watchdog_device *wdev, unsigned t)
drivers/watchdog/ebc-c384_wdt.c
63
if (t > 255) {
drivers/watchdog/ebc-c384_wdt.c
65
wdev->timeout = roundup(t, 60);
drivers/watchdog/ebc-c384_wdt.c
70
wdev->timeout = t;
drivers/watchdog/exar_wdt.c
194
static int exar_wdt_set_timeout(struct watchdog_device *wdog, unsigned int t)
drivers/watchdog/exar_wdt.c
203
if (t > 255) {
drivers/watchdog/exar_wdt.c
205
t = DIV_ROUND_UP(t, 60);
drivers/watchdog/exar_wdt.c
210
priv->timeout = t;
drivers/watchdog/exar_wdt.c
212
wdog->timeout = unit_min ? t * 60 : t;
drivers/watchdog/exar_wdt.c
371
struct wdt_pdev_node *n, *t;
drivers/watchdog/exar_wdt.c
373
list_for_each_entry_safe(n, t, &pdev_list, list) {
drivers/watchdog/iTCO_wdt.c
345
static int iTCO_wdt_set_timeout(struct watchdog_device *wd_dev, unsigned int t)
drivers/watchdog/iTCO_wdt.c
352
tmrval = seconds_to_ticks(p, t);
drivers/watchdog/iTCO_wdt.c
387
wd_dev->timeout = t;
drivers/watchdog/ib700wdt.c
131
static int ibwdt_set_heartbeat(int t)
drivers/watchdog/ib700wdt.c
133
if (t < 0 || t > 30)
drivers/watchdog/ib700wdt.c
136
timeout = t;
drivers/watchdog/ie6xx_wdt.c
104
preload = (t * clock) >> 15;
drivers/watchdog/ie6xx_wdt.c
128
wdd->timeout = t;
drivers/watchdog/ie6xx_wdt.c
95
static int ie6xx_wdt_set_timeout(struct watchdog_device *wdd, unsigned int t)
drivers/watchdog/intel_oc_wdt.c
92
unsigned int t)
drivers/watchdog/intel_oc_wdt.c
96
outl((inl(INTEL_OC_WDT_CTRL_REG(oc_wdt)) & ~INTEL_OC_WDT_TOV) | (t - 1),
drivers/watchdog/intel_oc_wdt.c
99
wdd->timeout = t;
drivers/watchdog/it87_wdt.c
170
static void _wdt_update_timeout(unsigned int t)
drivers/watchdog/it87_wdt.c
177
if (t <= max_units)
drivers/watchdog/it87_wdt.c
180
t /= 60;
drivers/watchdog/it87_wdt.c
186
superio_outb(t, WDTVALLSB);
drivers/watchdog/it87_wdt.c
188
superio_outb(t >> 8, WDTVALMSB);
drivers/watchdog/it87_wdt.c
197
static int wdt_update_timeout(unsigned int t)
drivers/watchdog/it87_wdt.c
206
_wdt_update_timeout(t);
drivers/watchdog/it87_wdt.c
212
static int wdt_round_time(int t)
drivers/watchdog/it87_wdt.c
214
t += 59;
drivers/watchdog/it87_wdt.c
215
t -= t % 60;
drivers/watchdog/it87_wdt.c
216
return t;
drivers/watchdog/it87_wdt.c
245
static int wdt_set_timeout(struct watchdog_device *wdd, unsigned int t)
drivers/watchdog/it87_wdt.c
249
if (t > max_units)
drivers/watchdog/it87_wdt.c
250
t = wdt_round_time(t);
drivers/watchdog/it87_wdt.c
252
wdd->timeout = t;
drivers/watchdog/it87_wdt.c
255
ret = wdt_update_timeout(t);
drivers/watchdog/keembay_wdt.c
113
static int keembay_wdt_set_timeout(struct watchdog_device *wdog, u32 t)
drivers/watchdog/keembay_wdt.c
115
wdog->timeout = t;
drivers/watchdog/keembay_wdt.c
122
static int keembay_wdt_set_pretimeout(struct watchdog_device *wdog, u32 t)
drivers/watchdog/keembay_wdt.c
124
if (t > wdog->timeout)
drivers/watchdog/keembay_wdt.c
127
wdog->pretimeout = t;
drivers/watchdog/lpc18xx_wdt.c
78
static void lpc18xx_wdt_timer_feed(struct timer_list *t)
drivers/watchdog/lpc18xx_wdt.c
81
t, timer);
drivers/watchdog/mt7621_wdt.c
101
t = rt_wdt_r32(drvdata->base, TIMER_REG_TMR1CTL);
drivers/watchdog/mt7621_wdt.c
102
t &= ~TMR1CTL_ENABLE;
drivers/watchdog/mt7621_wdt.c
103
rt_wdt_w32(drvdata->base, TIMER_REG_TMR1CTL, t);
drivers/watchdog/mt7621_wdt.c
66
static int mt7621_wdt_set_timeout(struct watchdog_device *w, unsigned int t)
drivers/watchdog/mt7621_wdt.c
70
w->timeout = t;
drivers/watchdog/mt7621_wdt.c
71
rt_wdt_w32(drvdata->base, TIMER_REG_TMR1LOAD, t * 1000);
drivers/watchdog/mt7621_wdt.c
80
u32 t;
drivers/watchdog/mt7621_wdt.c
87
t = rt_wdt_r32(drvdata->base, TIMER_REG_TMR1CTL);
drivers/watchdog/mt7621_wdt.c
88
t |= TMR1CTL_ENABLE;
drivers/watchdog/mt7621_wdt.c
89
rt_wdt_w32(drvdata->base, TIMER_REG_TMR1CTL, t);
drivers/watchdog/mt7621_wdt.c
97
u32 t;
drivers/watchdog/nv_tco.c
108
static int tco_timer_set_heartbeat(int t)
drivers/watchdog/nv_tco.c
120
if (t < 0 || t > 0x3f)
drivers/watchdog/nv_tco.c
122
tmrval = seconds_to_ticks(t);
drivers/watchdog/nv_tco.c
143
heartbeat = t;
drivers/watchdog/octeon-wdt-main.c
425
static void octeon_wdt_calc_parameters(int t)
drivers/watchdog/octeon-wdt-main.c
436
while ((t % timeout_sec) != 0)
drivers/watchdog/octeon-wdt-main.c
439
periods = t / timeout_sec;
drivers/watchdog/octeon-wdt-main.c
447
heartbeat = t;
drivers/watchdog/octeon-wdt-main.c
452
unsigned int t)
drivers/watchdog/octeon-wdt-main.c
459
if (t <= 0)
drivers/watchdog/octeon-wdt-main.c
462
octeon_wdt_calc_parameters(t);
drivers/watchdog/pcwd.c
469
static int pcwd_set_heartbeat(int t)
drivers/watchdog/pcwd.c
471
if (t < 2 || t > 7200) /* arbitrary upper limit */
drivers/watchdog/pcwd.c
474
heartbeat = t;
drivers/watchdog/pcwd_pci.c
319
static int pcipcwd_set_heartbeat(int t)
drivers/watchdog/pcwd_pci.c
321
int t_msb = t / 256;
drivers/watchdog/pcwd_pci.c
322
int t_lsb = t % 256;
drivers/watchdog/pcwd_pci.c
324
if ((t < 0x0001) || (t > 0xFFFF))
drivers/watchdog/pcwd_pci.c
330
heartbeat = t;
drivers/watchdog/pcwd_usb.c
310
static int usb_pcwd_set_heartbeat(struct usb_pcwd_private *usb_pcwd, int t)
drivers/watchdog/pcwd_usb.c
312
unsigned char msb = t / 256;
drivers/watchdog/pcwd_usb.c
313
unsigned char lsb = t % 256;
drivers/watchdog/pcwd_usb.c
315
if ((t < 0x0001) || (t > 0xFFFF))
drivers/watchdog/pcwd_usb.c
321
heartbeat = t;
drivers/watchdog/rn5t618_wdt.c
48
unsigned int t)
drivers/watchdog/rn5t618_wdt.c
54
if (rn5t618_wdt_map[i].time + 1 >= t)
drivers/watchdog/rt2880_wdt.c
100
u32 t;
drivers/watchdog/rt2880_wdt.c
104
t = rt_wdt_r32(drvdata->base, TIMER_REG_TMR1CTL);
drivers/watchdog/rt2880_wdt.c
105
t &= ~TMR1CTL_ENABLE;
drivers/watchdog/rt2880_wdt.c
106
rt_wdt_w32(drvdata->base, TIMER_REG_TMR1CTL, t);
drivers/watchdog/rt2880_wdt.c
111
static int rt288x_wdt_set_timeout(struct watchdog_device *w, unsigned int t)
drivers/watchdog/rt2880_wdt.c
113
w->timeout = t;
drivers/watchdog/rt2880_wdt.c
79
u32 t;
drivers/watchdog/rt2880_wdt.c
81
t = rt_wdt_r32(drvdata->base, TIMER_REG_TMR1CTL);
drivers/watchdog/rt2880_wdt.c
82
t &= ~(TMR1CTL_MODE_MASK << TMR1CTL_MODE_SHIFT |
drivers/watchdog/rt2880_wdt.c
84
t |= (TMR1CTL_MODE_WDT << TMR1CTL_MODE_SHIFT |
drivers/watchdog/rt2880_wdt.c
86
rt_wdt_w32(drvdata->base, TIMER_REG_TMR1CTL, t);
drivers/watchdog/rt2880_wdt.c
90
t = rt_wdt_r32(drvdata->base, TIMER_REG_TMR1CTL);
drivers/watchdog/rt2880_wdt.c
91
t |= TMR1CTL_ENABLE;
drivers/watchdog/rt2880_wdt.c
92
rt_wdt_w32(drvdata->base, TIMER_REG_TMR1CTL, t);
drivers/watchdog/sb_wdog.c
70
static void sbwdog_set(char __iomem *wdog, unsigned long t)
drivers/watchdog/sb_wdog.c
74
__raw_writeq(t & 0x7fffffUL, wdog - 0x10);
drivers/watchdog/sbc7240_wdt.c
71
static int wdt_set_timeout(int t)
drivers/watchdog/sbc7240_wdt.c
73
if (t < 1 || t > SBC7240_MAX_TIMEOUT) {
drivers/watchdog/sbc7240_wdt.c
78
outb_p((unsigned)t, SBC7240_SET_TIMEOUT_PORT);
drivers/watchdog/sbc7240_wdt.c
79
timeout = t;
drivers/watchdog/sbc7240_wdt.c
80
pr_info("timeout set to %d seconds\n", t);
drivers/watchdog/sc520_wdt.c
205
static int wdt_set_heartbeat(int t)
drivers/watchdog/sc520_wdt.c
207
if ((t < 1) || (t > 3600)) /* arbitrary upper limit */
drivers/watchdog/sc520_wdt.c
210
timeout = t;
drivers/watchdog/sch311x_wdt.c
109
static void sch311x_wdt_set_timeout(int t)
drivers/watchdog/sch311x_wdt.c
114
if (t > 255) {
drivers/watchdog/sch311x_wdt.c
116
t /= 60;
drivers/watchdog/sch311x_wdt.c
129
outb(t, sch311x_wdt_data.runtime_reg + WDT_VAL);
drivers/watchdog/sch311x_wdt.c
134
unsigned char t;
drivers/watchdog/sch311x_wdt.c
149
t = inb(sch311x_wdt_data.runtime_reg + GP60);
drivers/watchdog/sch311x_wdt.c
150
outb((t & ~0x0d) | 0x0c, sch311x_wdt_data.runtime_reg + GP60);
drivers/watchdog/sch311x_wdt.c
158
unsigned char t;
drivers/watchdog/sch311x_wdt.c
163
t = inb(sch311x_wdt_data.runtime_reg + GP60);
drivers/watchdog/sch311x_wdt.c
164
outb((t & ~0x0d) | 0x01, sch311x_wdt_data.runtime_reg + GP60);
drivers/watchdog/sch311x_wdt.c
178
static int sch311x_wdt_set_heartbeat(int t)
drivers/watchdog/sch311x_wdt.c
180
if (t < 1 || t > (255*60))
drivers/watchdog/sch311x_wdt.c
185
if (t > 255)
drivers/watchdog/sch311x_wdt.c
186
t = (((t - 1) / 60) + 1) * 60;
drivers/watchdog/sch311x_wdt.c
188
timeout = t;
drivers/watchdog/shwdt.c
158
static int sh_wdt_set_heartbeat(struct watchdog_device *wdt_dev, unsigned t)
drivers/watchdog/shwdt.c
163
if (unlikely(t < 1 || t > 3600)) /* arbitrary upper limit */
drivers/watchdog/shwdt.c
167
heartbeat = t;
drivers/watchdog/shwdt.c
168
wdt_dev->timeout = t;
drivers/watchdog/shwdt.c
174
static void sh_wdt_ping(struct timer_list *t)
drivers/watchdog/shwdt.c
176
struct sh_wdt *wdt = timer_container_of(wdt, t, timer);
drivers/watchdog/simatic-ipc-wdt.c
83
static int wd_set_timeout(struct watchdog_device *wdd, unsigned int t)
drivers/watchdog/simatic-ipc-wdt.c
85
int timeout_idx = find_closest(t, wd_timeout_table,
drivers/watchdog/sp5100_tco.c
150
unsigned int t)
drivers/watchdog/sp5100_tco.c
155
writel(t, SP5100_WDT_COUNT(tco->tcobase));
drivers/watchdog/sp5100_tco.c
157
wdd->timeout = t;
drivers/watchdog/tqmx86_wdt.c
44
static int tqmx86_wdt_set_timeout(struct watchdog_device *wdd, unsigned int t)
drivers/watchdog/tqmx86_wdt.c
49
t = roundup_pow_of_two(t);
drivers/watchdog/tqmx86_wdt.c
50
val = ilog2(t) | 0x90;
drivers/watchdog/tqmx86_wdt.c
54
wdd->timeout = t;
drivers/watchdog/uniphier_wdt.c
142
unsigned int t)
drivers/watchdog/uniphier_wdt.c
148
tmp_timeout = roundup_pow_of_two(t);
drivers/watchdog/w83627hf_wdt.c
157
unsigned char t;
drivers/watchdog/w83627hf_wdt.c
166
t = superio_inb(0x30);
drivers/watchdog/w83627hf_wdt.c
167
if (!(t & 0x01))
drivers/watchdog/w83627hf_wdt.c
168
superio_outb(0x30, t | 0x01);
drivers/watchdog/w83627hf_wdt.c
173
t = superio_inb(0x2B) & ~0x10;
drivers/watchdog/w83627hf_wdt.c
174
superio_outb(0x2B, t); /* set GPIO24 to WDT0 */
drivers/watchdog/w83627hf_wdt.c
178
t = superio_inb(0x29) & ~0x60;
drivers/watchdog/w83627hf_wdt.c
179
t |= 0x20;
drivers/watchdog/w83627hf_wdt.c
180
superio_outb(0x29, t);
drivers/watchdog/w83627hf_wdt.c
184
t = superio_inb(0x2b) & ~0x04;
drivers/watchdog/w83627hf_wdt.c
185
superio_outb(0x2b, t);
drivers/watchdog/w83627hf_wdt.c
188
t = (superio_inb(0x2B) & ~0x08) | 0x04;
drivers/watchdog/w83627hf_wdt.c
189
superio_outb(0x2B, t); /* set GPIO3 to WDT0 */
drivers/watchdog/w83627hf_wdt.c
193
t = superio_inb(0x2D) & ~0x01; /* PIN77 -> WDT0# */
drivers/watchdog/w83627hf_wdt.c
194
superio_outb(0x2D, t); /* set GPIO5 to WDT0 */
drivers/watchdog/w83627hf_wdt.c
195
t = superio_inb(cr_wdt_control);
drivers/watchdog/w83627hf_wdt.c
196
t |= 0x02; /* enable the WDTO# output low pulse
drivers/watchdog/w83627hf_wdt.c
198
superio_outb(cr_wdt_control, t);
drivers/watchdog/w83627hf_wdt.c
203
t = superio_inb(0x2C) & ~0x80; /* PIN47 -> WDT0# */
drivers/watchdog/w83627hf_wdt.c
204
superio_outb(0x2C, t);
drivers/watchdog/w83627hf_wdt.c
226
t = superio_inb(cr_wdt_control);
drivers/watchdog/w83627hf_wdt.c
227
t |= 0x02; /* enable the WDTO# output low pulse
drivers/watchdog/w83627hf_wdt.c
229
superio_outb(cr_wdt_control, t);
drivers/watchdog/w83627hf_wdt.c
235
t = superio_inb(cr_wdt_timeout);
drivers/watchdog/w83627hf_wdt.c
236
if (t != 0) {
drivers/watchdog/w83627hf_wdt.c
248
t = superio_inb(cr_wdt_control) & ~0x0C;
drivers/watchdog/w83627hf_wdt.c
249
superio_outb(cr_wdt_control, t);
drivers/watchdog/w83627hf_wdt.c
251
t = superio_inb(cr_wdt_csr);
drivers/watchdog/w83627hf_wdt.c
252
if (t & WDT_CSR_STATUS)
drivers/watchdog/w83627hf_wdt.c
256
t &= ~(WDT_CSR_STATUS | WDT_CSR_KBD | WDT_CSR_MOUSE);
drivers/watchdog/w83627hf_wdt.c
257
superio_outb(cr_wdt_csr, t);
drivers/watchdog/w83977f_wdt.c
220
static int wdt_set_timeout(int t)
drivers/watchdog/w83977f_wdt.c
231
if (t < 15)
drivers/watchdog/w83977f_wdt.c
234
tmrval = ((t + 15) + 29) / 30;
drivers/watchdog/watchdog_core.c
125
unsigned int t = 0;
drivers/watchdog/watchdog_core.c
142
if (dev && device_property_read_u32(dev, "timeout-sec", &t) == 0) {
drivers/watchdog/watchdog_core.c
143
if (t && !watchdog_timeout_invalid(wdd, t)) {
drivers/watchdog/watchdog_core.c
144
wdd->timeout = t;
drivers/watchdog/watchdog_core.c
147
pr_err("%s: DT supplied timeout (%u) out of range\n", dev_str, t);
drivers/watchdog/watchdog_dev.c
134
ktime_t t = watchdog_next_keepalive(wdd);
drivers/watchdog/watchdog_dev.c
136
if (t > 0)
drivers/watchdog/watchdog_dev.c
137
hrtimer_start(&wd_data->timer, t,
drivers/watchdog/watchdog_dev.c
80
unsigned int t = wdd->timeout * 1000;
drivers/watchdog/watchdog_dev.c
95
return (hm && watchdog_active(wdd) && t > hm) ||
drivers/watchdog/watchdog_dev.c
96
(t && !watchdog_active(wdd) && watchdog_hw_running(wdd));
drivers/watchdog/watchdog_pretimeout.c
157
struct governor_priv *priv, *t;
drivers/watchdog/watchdog_pretimeout.c
161
list_for_each_entry_safe(priv, t, &governor_list, entry) {
drivers/watchdog/watchdog_pretimeout.c
201
struct watchdog_pretimeout *p, *t;
drivers/watchdog/watchdog_pretimeout.c
209
list_for_each_entry_safe(p, t, &pretimeout_list, entry) {
drivers/watchdog/wdt.c
182
static int wdt_set_heartbeat(int t)
drivers/watchdog/wdt.c
184
if (t < 1 || t > 65535)
drivers/watchdog/wdt.c
187
heartbeat = t;
drivers/watchdog/wdt.c
188
wd_heartbeat = t * 100;
drivers/watchdog/wdt977.c
199
static int wdt977_set_timeout(int t)
drivers/watchdog/wdt977.c
204
tmrval = (t + 59) / 60;
drivers/watchdog/wdt977.c
219
timeout = t;
drivers/watchdog/wdt_pci.c
216
static int wdtpci_set_heartbeat(int t)
drivers/watchdog/wdt_pci.c
219
if (t < 1 || t > 65535)
drivers/watchdog/wdt_pci.c
222
heartbeat = t;
drivers/watchdog/wdt_pci.c
223
wd_heartbeat = t * 100;
drivers/xen/xen-pciback/conf_space.c
332
struct config_field_entry *cfg_entry, *t;
drivers/xen/xen-pciback/conf_space.c
340
list_for_each_entry_safe(cfg_entry, t, &dev_data->config_fields, list) {
drivers/xen/xen-pciback/conf_space.c
376
struct config_field_entry *cfg_entry, *t;
drivers/xen/xen-pciback/conf_space.c
383
list_for_each_entry_safe(cfg_entry, t, &dev_data->config_fields, list) {
drivers/xen/xen-pciback/passthrough.c
162
struct pci_dev_entry *dev_entry, *t;
drivers/xen/xen-pciback/passthrough.c
164
list_for_each_entry_safe(dev_entry, t, &dev_data->dev_list, list) {
drivers/xen/xen-pciback/passthrough.c
76
struct pci_dev_entry *dev_entry, *t;
drivers/xen/xen-pciback/passthrough.c
81
list_for_each_entry_safe(dev_entry, t, &dev_data->dev_list, list) {
drivers/xen/xen-pciback/pci_stub.c
1146
struct pcistub_device_id *pci_dev_id, *t;
drivers/xen/xen-pciback/pci_stub.c
1151
list_for_each_entry_safe(pci_dev_id, t, &pcistub_device_ids,
drivers/xen/xen-pciback/vpci.c
107
t = list_entry(list_first(&vpci_dev->dev_list[slot]),
drivers/xen/xen-pciback/vpci.c
109
if (t->dev->is_virtfn && !PCI_FUNC(t->dev->devfn))
drivers/xen/xen-pciback/vpci.c
112
if (match_slot(dev, t->dev)) {
drivers/xen/xen-pciback/vpci.c
74
struct pci_dev_entry *t, *dev_entry;
drivers/xen/xenbus/xenbus_xs.c
315
static void *xs_talkv(struct xenbus_transaction t,
drivers/xen/xenbus/xenbus_xs.c
338
msg.tx_id = t.id;
drivers/xen/xenbus/xenbus_xs.c
369
static void *xs_single(struct xenbus_transaction t,
drivers/xen/xenbus/xenbus_xs.c
378
return xs_talkv(t, type, &iovec, 1, len);
drivers/xen/xenbus/xenbus_xs.c
436
char **xenbus_directory(struct xenbus_transaction t,
drivers/xen/xenbus/xenbus_xs.c
446
strings = xs_single(t, XS_DIRECTORY, path, &len);
drivers/xen/xenbus/xenbus_xs.c
456
int xenbus_exists(struct xenbus_transaction t,
drivers/xen/xenbus/xenbus_xs.c
462
d = xenbus_directory(t, dir, node, &dir_n);
drivers/xen/xenbus/xenbus_xs.c
474
void *xenbus_read(struct xenbus_transaction t,
drivers/xen/xenbus/xenbus_xs.c
484
ret = xs_single(t, XS_READ, path, len);
drivers/xen/xenbus/xenbus_xs.c
493
int xenbus_write(struct xenbus_transaction t,
drivers/xen/xenbus/xenbus_xs.c
509
ret = xs_error(xs_talkv(t, XS_WRITE, iovec, ARRAY_SIZE(iovec), NULL));
drivers/xen/xenbus/xenbus_xs.c
516
int xenbus_rm(struct xenbus_transaction t, const char *dir, const char *node)
drivers/xen/xenbus/xenbus_xs.c
525
ret = xs_error(xs_single(t, XS_RM, path, NULL));
drivers/xen/xenbus/xenbus_xs.c
534
int xenbus_transaction_start(struct xenbus_transaction *t)
drivers/xen/xenbus/xenbus_xs.c
542
t->id = simple_strtoul(id_str, NULL, 0);
drivers/xen/xenbus/xenbus_xs.c
551
int xenbus_transaction_end(struct xenbus_transaction t, bool abort)
drivers/xen/xenbus/xenbus_xs.c
553
return xs_error(xs_single(t, XS_TRANSACTION_END, abort ? "F" : "T",
drivers/xen/xenbus/xenbus_xs.c
559
int xenbus_scanf(struct xenbus_transaction t,
drivers/xen/xenbus/xenbus_xs.c
566
val = xenbus_read(t, dir, node, NULL);
drivers/xen/xenbus/xenbus_xs.c
597
int xenbus_printf(struct xenbus_transaction t,
drivers/xen/xenbus/xenbus_xs.c
611
ret = xenbus_write(t, dir, node, buf);
drivers/xen/xenbus/xenbus_xs.c
620
int xenbus_gather(struct xenbus_transaction t, const char *dir, ...)
drivers/xen/xenbus/xenbus_xs.c
632
p = xenbus_read(t, dir, name, NULL);
fs/adfs/dir_fplus.c
107
struct adfs_bigdirtail *t;
fs/adfs/dir_fplus.c
135
dir->bigtail = t = (struct adfs_bigdirtail *)
fs/adfs/dir_fplus.c
138
ret = adfs_fplus_validate_tail(h, t);
fs/adfs/dir_fplus.c
144
if (adfs_fplus_checkbyte(dir) != t->bigdircheckbyte) {
fs/adfs/dir_fplus.c
60
const struct adfs_bigdirtail *t)
fs/adfs/dir_fplus.c
62
if (t->bigdirendname != cpu_to_le32(BIGDIRENDNAME) ||
fs/adfs/dir_fplus.c
63
t->bigdirendmasseq != h->startmasseq ||
fs/adfs/dir_fplus.c
64
t->reserved[0] != 0 || t->reserved[1] != 0)
fs/adfs/dir_fplus.c
73
struct adfs_bigdirtail *t = dir->bigtail;
fs/adfs/dir_fplus.c
95
dircheck = ror32(dircheck, 13) ^ le32_to_cpu(t->bigdirendname);
fs/adfs/dir_fplus.c
96
dircheck = ror32(dircheck, 13) ^ t->bigdirendmasseq;
fs/adfs/dir_fplus.c
97
dircheck = ror32(dircheck, 13) ^ t->reserved[0];
fs/adfs/dir_fplus.c
98
dircheck = ror32(dircheck, 13) ^ t->reserved[1];
fs/afs/inode.c
159
struct timespec64 t;
fs/afs/inode.c
177
t = status->mtime_client;
fs/afs/inode.c
178
inode_set_ctime_to_ts(inode, t);
fs/afs/inode.c
179
inode_set_mtime_to_ts(inode, t);
fs/afs/inode.c
180
inode_set_atime_to_ts(inode, t);
fs/afs/inode.c
257
struct timespec64 t;
fs/afs/inode.c
295
t = status->mtime_client;
fs/afs/inode.c
296
inode_set_mtime_to_ts(inode, t);
fs/afs/inode.c
351
inode_set_ctime_to_ts(inode, t);
fs/afs/inode.c
352
inode_set_atime_to_ts(inode, t);
fs/afs/yfsclient.c
107
static struct timespec64 yfs_time_to_linux(s64 t)
fs/afs/yfsclient.c
117
if (t < 0) {
fs/afs/yfsclient.c
118
abs_t = -t;
fs/afs/yfsclient.c
123
abs_t = t;
fs/afs/yfsclient.c
133
s64 t = xdr_to_u64(xdr);
fs/afs/yfsclient.c
135
return yfs_time_to_linux(t);
fs/afs/yfsclient.c
80
static s64 linux_to_yfs_time(const struct timespec64 *t)
fs/afs/yfsclient.c
83
return (u64)t->tv_sec * 10000000 + t->tv_nsec/100;
fs/afs/yfsclient.c
87
const struct timespec64 *t)
fs/afs/yfsclient.c
91
s64 mtime = linux_to_yfs_time(t);
fs/aio.c
1315
struct hrtimer_sleeper t;
fs/aio.c
1337
hrtimer_setup_sleeper_on_stack(&t, CLOCK_MONOTONIC, HRTIMER_MODE_REL);
fs/aio.c
1339
hrtimer_set_expires_range_ns(&t.timer, until, current->timer_slack_ns);
fs/aio.c
1340
hrtimer_sleeper_start_expires(&t, HRTIMER_MODE_REL);
fs/aio.c
1351
if (!ret2 && !t.task)
fs/aio.c
1362
hrtimer_cancel(&t.timer);
fs/aio.c
1363
destroy_hrtimer_on_stack(&t.timer);
fs/aio.c
2354
struct timespec64 t;
fs/aio.c
2357
if (timeout && get_old_timespec32(&t, timeout))
fs/aio.c
2360
ret = do_io_getevents(ctx_id, min_nr, nr, events, timeout ? &t : NULL);
fs/aio.c
2386
struct timespec64 t;
fs/aio.c
2390
if (timeout && get_old_timespec32(&t, timeout))
fs/aio.c
2400
ret = do_io_getevents(ctx_id, min_nr, nr, events, timeout ? &t : NULL);
fs/aio.c
2421
struct timespec64 t;
fs/aio.c
2425
if (timeout && get_timespec64(&t, timeout))
fs/aio.c
2435
ret = do_io_getevents(ctx_id, min_nr, nr, events, timeout ? &t : NULL);
fs/binfmt_elf.c
1721
static int fill_thread_core_info(struct elf_thread_core_info *t,
fs/binfmt_elf.c
1733
fill_prstatus(&t->prstatus.common, t->task, signr);
fs/binfmt_elf.c
1734
regset_get(t->task, &view->regsets[0],
fs/binfmt_elf.c
1735
sizeof(t->prstatus.pr_reg), &t->prstatus.pr_reg);
fs/binfmt_elf.c
1737
fill_note(&t->notes[0], PRSTATUS, PRSTATUS_SIZE, &t->prstatus);
fs/binfmt_elf.c
1738
info->size += notesize(&t->notes[0]);
fs/binfmt_elf.c
1740
do_thread_regset_writeback(t->task, &view->regsets[0]);
fs/binfmt_elf.c
1755
do_thread_regset_writeback(t->task, regset);
fs/binfmt_elf.c
1758
if (regset->active && regset->active(t->task, regset) <= 0)
fs/binfmt_elf.c
1761
ret = regset_get_alloc(t->task, regset, ~0U, &data);
fs/binfmt_elf.c
1769
SET_PR_FPVALID(&t->prstatus);
fs/binfmt_elf.c
1779
__fill_note(&t->notes[note_iter], note_name, note_type,
fs/binfmt_elf.c
1782
info->size += notesize(&t->notes[note_iter]);
fs/binfmt_elf.c
1789
static int fill_thread_core_info(struct elf_thread_core_info *t,
fs/binfmt_elf.c
1793
struct task_struct *p = t->task;
fs/binfmt_elf.c
1796
fill_prstatus(&t->prstatus.common, p, signr);
fs/binfmt_elf.c
1797
elf_core_copy_task_regs(p, &t->prstatus.pr_reg);
fs/binfmt_elf.c
1799
fill_note(&t->notes[0], PRSTATUS, sizeof(t->prstatus), &t->prstatus);
fs/binfmt_elf.c
1800
info->size += notesize(&t->notes[0]);
fs/binfmt_elf.c
1808
t->prstatus.pr_fpvalid = 1;
fs/binfmt_elf.c
1809
fill_note(&t->notes[1], PRFPREG, sizeof(*fpu), fpu);
fs/binfmt_elf.c
1810
info->size += notesize(&t->notes[1]);
fs/binfmt_elf.c
1822
struct elf_thread_core_info *t;
fs/binfmt_elf.c
1883
t = kzalloc_flex(*t, notes, info->thread_notes);
fs/binfmt_elf.c
1884
if (unlikely(!t))
fs/binfmt_elf.c
1887
t->task = ct->task;
fs/binfmt_elf.c
1888
t->next = info->thread->next;
fs/binfmt_elf.c
1889
info->thread->next = t;
fs/binfmt_elf.c
1895
for (t = info->thread; t != NULL; t = t->next)
fs/binfmt_elf.c
1896
if (!fill_thread_core_info(t, view, cprm->siginfo->si_signo, info))
fs/binfmt_elf.c
1925
struct elf_thread_core_info *t = info->thread;
fs/binfmt_elf.c
1930
if (!writenote(&t->notes[0], cprm))
fs/binfmt_elf.c
1944
if (t->notes[i].data &&
fs/binfmt_elf.c
1945
!writenote(&t->notes[i], cprm))
fs/binfmt_elf.c
1949
t = t->next;
fs/binfmt_elf.c
1950
} while (t);
fs/binfmt_elf.c
1960
struct elf_thread_core_info *t = threads;
fs/binfmt_elf.c
1961
threads = t->next;
fs/binfmt_elf.c
1962
WARN_ON(t->notes[0].data && t->notes[0].data != &t->prstatus);
fs/binfmt_elf.c
1964
kvfree(t->notes[i].data);
fs/binfmt_elf.c
1965
kfree(t);
fs/binfmt_elf_fdpic.c
1397
struct elf_thread_status *t;
fs/binfmt_elf_fdpic.c
1400
t = kzalloc_obj(struct elf_thread_status);
fs/binfmt_elf_fdpic.c
1401
if (!t)
fs/binfmt_elf_fdpic.c
1402
return t;
fs/binfmt_elf_fdpic.c
1404
fill_prstatus(&t->prstatus.common, p, signr);
fs/binfmt_elf_fdpic.c
1405
t->prstatus.pr_exec_fdpic_loadmap = p->mm->context.exec_fdpic_loadmap;
fs/binfmt_elf_fdpic.c
1406
t->prstatus.pr_interp_fdpic_loadmap = p->mm->context.interp_fdpic_loadmap;
fs/binfmt_elf_fdpic.c
1408
sizeof(t->prstatus.pr_reg), &t->prstatus.pr_reg);
fs/binfmt_elf_fdpic.c
1410
fill_note(&t->notes[0], PRSTATUS, sizeof(t->prstatus), &t->prstatus);
fs/binfmt_elf_fdpic.c
1411
t->num_notes++;
fs/binfmt_elf_fdpic.c
1412
*sz += notesize(&t->notes[0]);
fs/binfmt_elf_fdpic.c
1420
ret = regset_get(p, regset, sizeof(t->fpu), &t->fpu);
fs/binfmt_elf_fdpic.c
1422
t->prstatus.pr_fpvalid = 1;
fs/binfmt_elf_fdpic.c
1426
if (t->prstatus.pr_fpvalid) {
fs/binfmt_elf_fdpic.c
1427
fill_note(&t->notes[1], PRFPREG, sizeof(t->fpu), &t->fpu);
fs/binfmt_elf_fdpic.c
1428
t->num_notes++;
fs/binfmt_elf_fdpic.c
1429
*sz += notesize(&t->notes[1]);
fs/binfmt_elf_fdpic.c
1431
return t;
fs/btrfs/ctree.c
2571
struct extent_buffer *t;
fs/btrfs/ctree.c
2579
t = path->nodes[i];
fs/btrfs/ctree.c
2580
ret = btrfs_tree_mod_log_insert_key(t, tslot,
fs/btrfs/ctree.c
2583
btrfs_set_node_key(t, key, tslot);
fs/btrfs/disk-io.c
4869
struct btrfs_transaction *t;
fs/btrfs/disk-io.c
4875
t = list_first_entry(&fs_info->trans_list,
fs/btrfs/disk-io.c
4877
if (t->state >= TRANS_STATE_COMMIT_PREP) {
fs/btrfs/disk-io.c
4878
refcount_inc(&t->use_count);
fs/btrfs/disk-io.c
4880
btrfs_wait_for_commit(fs_info, t->transid);
fs/btrfs/disk-io.c
4881
btrfs_put_transaction(t);
fs/btrfs/disk-io.c
4885
if (t == fs_info->running_transaction) {
fs/btrfs/disk-io.c
4886
t->state = TRANS_STATE_COMMIT_DOING;
fs/btrfs/disk-io.c
4892
wait_event(t->writer_wait,
fs/btrfs/disk-io.c
4893
atomic_read(&t->num_writers) == 0);
fs/btrfs/disk-io.c
4897
btrfs_cleanup_one_transaction(t);
fs/btrfs/disk-io.c
4900
if (t == fs_info->running_transaction)
fs/btrfs/disk-io.c
4902
list_del_init(&t->list);
fs/btrfs/disk-io.c
4905
btrfs_put_transaction(t);
fs/btrfs/fs.c
27
u16 t = btrfs_super_csum_type(s);
fs/btrfs/fs.c
30
return btrfs_csum_type_size(t);
fs/btrfs/transaction.c
948
struct btrfs_transaction *cur_trans = NULL, *t;
fs/btrfs/transaction.c
957
list_for_each_entry(t, &fs_info->trans_list, list) {
fs/btrfs/transaction.c
958
if (t->transid == transid) {
fs/btrfs/transaction.c
959
cur_trans = t;
fs/btrfs/transaction.c
964
if (t->transid > transid) {
fs/btrfs/transaction.c
983
list_for_each_entry_reverse(t, &fs_info->trans_list,
fs/btrfs/transaction.c
985
if (t->state >= TRANS_STATE_COMMIT_START) {
fs/btrfs/transaction.c
986
if (t->state == TRANS_STATE_COMPLETED)
fs/btrfs/transaction.c
988
cur_trans = t;
fs/ceph/file.c
2630
u64 nearly, t;
fs/ceph/file.c
2634
t = nearly;
fs/ceph/file.c
2635
nearly -= do_div(t, object_set_size);
fs/ceph/inode.c
370
u32 t = ceph_frag_make(0, 0);
fs/ceph/inode.c
379
WARN_ON(!ceph_frag_contains_value(t, v));
fs/ceph/inode.c
380
frag = __ceph_find_frag(ci, t);
fs/ceph/inode.c
393
doutc(cl, "frag(%x) %x splits by %d (%d ways)\n", v, t,
fs/ceph/inode.c
396
n = ceph_frag_make_child(t, frag->split_by, i);
fs/ceph/inode.c
398
t = n;
fs/ceph/inode.c
404
doutc(cl, "frag(%x) = %x\n", v, t);
fs/ceph/inode.c
406
return t;
fs/ceph/metric.c
13
struct timespec64 t = ktime_to_timespec64(val);
fs/ceph/metric.c
14
ceph_encode_timespec64(ts, &t);
fs/ceph/super.h
791
extern int __ceph_caps_issued_mask(struct ceph_inode_info *ci, int mask, int t);
fs/ceph/super.h
793
int t);
fs/configfs/dir.c
606
const struct config_item_type *t = item->ci_type;
fs/configfs/dir.c
613
if (!t)
fs/configfs/dir.c
616
ops = t->ct_group_ops;
fs/configfs/dir.c
618
if (t->ct_attrs) {
fs/configfs/dir.c
619
for (i = 0; (attr = t->ct_attrs[i]) != NULL; i++) {
fs/configfs/dir.c
627
if (!error && t->ct_bin_attrs) {
fs/configfs/dir.c
628
for (i = 0; (bin_attr = t->ct_bin_attrs[i]) != NULL; i++) {
fs/configfs/item.c
119
const struct config_item_type *t = item->ci_type;
fs/configfs/item.c
127
if (t && t->ct_item_ops && t->ct_item_ops->release)
fs/configfs/item.c
128
t->ct_item_ops->release(item);
fs/coredump.c
483
struct task_struct *t;
fs/coredump.c
490
__for_each_thread(signal, t) {
fs/coredump.c
491
task_clear_jobctl_pending(t, JOBCTL_PENDING_MASK);
fs/coredump.c
492
if (t != current && !(t->flags & PF_POSTCOREDUMP)) {
fs/coredump.c
493
sigaddset(&t->pending.signal, SIGKILL);
fs/coredump.c
494
signal_wake_up(t, 1);
fs/erofs/sysfs.c
159
unsigned long t;
fs/erofs/sysfs.c
166
ret = kstrtoul(skip_spaces(buf), 0, &t);
fs/erofs/sysfs.c
169
if (t != (unsigned int)t)
fs/erofs/sysfs.c
173
(t > EROFS_SYNC_DECOMPRESS_FORCE_OFF))
fs/erofs/sysfs.c
175
*(unsigned int *)ptr = t;
fs/erofs/sysfs.c
180
ret = kstrtoul(skip_spaces(buf), 0, &t);
fs/erofs/sysfs.c
183
if (t != 0 && t != 1)
fs/erofs/sysfs.c
185
*(bool *)ptr = !!t;
fs/erofs/sysfs.c
189
ret = kstrtoul(skip_spaces(buf), 0, &t);
fs/erofs/sysfs.c
192
if (t < 1 || t > 3)
fs/erofs/sysfs.c
195
if (t & 2)
fs/erofs/sysfs.c
197
if (t & 1)
fs/erofs/sysfs.c
206
t = strcspn(buf, "\n");
fs/erofs/sysfs.c
207
ret = z_erofs_crypto_enable_engine(buf, t);
fs/erofs/sysfs.c
210
buf += buf[t] != '\0' ? t + 1 : t;
fs/exec.c
1488
struct task_struct *p = current, *t;
fs/exec.c
1514
for_other_threads(p, t) {
fs/exec.c
1515
if (t->fs == p->fs)
fs/exfat/misc.c
100
t = (tm.tm_hour << 11) | (tm.tm_min << 5) | (tm.tm_sec >> 1);
fs/exfat/misc.c
103
*time = cpu_to_le16(t);
fs/exfat/misc.c
71
u16 t = le16_to_cpu(time);
fs/exfat/misc.c
75
t >> 11, (t >> 5) & 0x003F, (t & 0x001F) << 1);
fs/exfat/misc.c
97
u16 t, d;
fs/exfat/namei.c
105
struct nls_table *t = EXFAT_SB(sb)->nls_io;
fs/exfat/namei.c
117
charlen = t->char2uni(&name->name[i], alen - i, &c1);
fs/exfat/namei.c
120
if (charlen != t->char2uni(&str[i], blen - i, &c2))
fs/exfat/namei.c
82
struct nls_table *t = EXFAT_SB(sb)->nls_io;
fs/exfat/namei.c
91
charlen = t->char2uni(&name[i], len - i, &c);
fs/ext4/ext4.h
3207
extern void print_daily_error_info(struct timer_list *t);
fs/ext4/inode-test.c
214
static void timestamp_expectation_to_desc(const struct timestamp_expectation *t,
fs/ext4/inode-test.c
217
strscpy(desc, t->test_case_name, KUNIT_PARAM_DESC_SIZE);
fs/ext4/namei.c
304
struct ext4_dir_entry_tail *t = EXT4_DIRENT_TAIL(bh->b_data, blocksize);
fs/ext4/namei.c
306
memset(t, 0, sizeof(struct ext4_dir_entry_tail));
fs/ext4/namei.c
307
t->det_rec_len = ext4_rec_len_to_disk(
fs/ext4/namei.c
309
t->det_reserved_ft = EXT4_FT_DIR_CSUM;
fs/ext4/namei.c
316
struct ext4_dir_entry_tail *t;
fs/ext4/namei.c
332
t = (struct ext4_dir_entry_tail *)d;
fs/ext4/namei.c
334
t = EXT4_DIRENT_TAIL(bh->b_data, EXT4_BLOCK_SIZE(inode->i_sb));
fs/ext4/namei.c
337
if (t->det_reserved_zero1 ||
fs/ext4/namei.c
338
(ext4_rec_len_from_disk(t->det_rec_len, blocksize) !=
fs/ext4/namei.c
340
t->det_reserved_zero2 ||
fs/ext4/namei.c
341
t->det_reserved_ft != EXT4_FT_DIR_CSUM)
fs/ext4/namei.c
344
return t;
fs/ext4/namei.c
368
struct ext4_dir_entry_tail *t;
fs/ext4/namei.c
373
t = get_dirent_tail(inode, bh);
fs/ext4/namei.c
374
if (!t) {
fs/ext4/namei.c
379
if (t->det_checksum != ext4_dirblock_csum(inode, bh->b_data,
fs/ext4/namei.c
380
(char *)t - bh->b_data))
fs/ext4/namei.c
389
struct ext4_dir_entry_tail *t;
fs/ext4/namei.c
394
t = get_dirent_tail(inode, bh);
fs/ext4/namei.c
395
if (!t) {
fs/ext4/namei.c
400
t->det_checksum = ext4_dirblock_csum(inode, bh->b_data,
fs/ext4/namei.c
401
(char *)t - bh->b_data);
fs/ext4/namei.c
442
int count_offset, int count, struct dx_tail *t)
fs/ext4/namei.c
452
csum = ext4_chksum(csum, (__u8 *)t, offset);
fs/ext4/namei.c
462
struct dx_tail *t;
fs/ext4/namei.c
480
t = (struct dx_tail *)(((struct dx_entry *)c) + limit);
fs/ext4/namei.c
482
if (t->dt_checksum != ext4_dx_csum(inode, dirent, count_offset,
fs/ext4/namei.c
483
count, t))
fs/ext4/namei.c
491
struct dx_tail *t;
fs/ext4/namei.c
509
t = (struct dx_tail *)(((struct dx_entry *)c) + limit);
fs/ext4/namei.c
511
t->dt_checksum = ext4_dx_csum(inode, dirent, count_offset, count, t);
fs/ext4/super.c
1604
#define QTYPE2NAME(t) (quotatypes[t])
fs/ext4/super.c
3678
void print_daily_error_info(struct timer_list *t)
fs/ext4/super.c
3680
struct ext4_sb_info *sbi = timer_container_of(sbi, t, s_err_report);
fs/ext4/sysfs.c
137
unsigned long t;
fs/ext4/sysfs.c
140
ret = kstrtoul(skip_spaces(buf), 0, &t);
fs/ext4/sysfs.c
145
if (t > (365*24*60*60))
fs/ext4/sysfs.c
148
if (sbi->s_err_report_sec == t) /*nothing to do*/
fs/ext4/sysfs.c
150
else if (!sbi->s_err_report_sec && t) {
fs/ext4/sysfs.c
152
} else if (sbi->s_err_report_sec && !t) {
fs/ext4/sysfs.c
157
sbi->s_err_report_sec = t;
fs/ext4/sysfs.c
496
unsigned int t;
fs/ext4/sysfs.c
505
ret = kstrtouint(skip_spaces(buf), 0, &t);
fs/ext4/sysfs.c
508
if ((int)t < 0)
fs/ext4/sysfs.c
510
*((unsigned int *) ptr) = t;
fs/ext4/sysfs.c
513
ret = kstrtouint(skip_spaces(buf), 0, &t);
fs/ext4/sysfs.c
517
*((__le32 *) ptr) = cpu_to_le32(t);
fs/ext4/sysfs.c
519
*((unsigned int *) ptr) = t;
fs/ext4/sysfs.c
522
ret = kstrtouint(skip_spaces(buf), 0, &t);
fs/ext4/sysfs.c
525
if (t > 64)
fs/ext4/sysfs.c
527
*((unsigned int *) ptr) = t;
fs/ext4/sysfs.c
530
ret = kstrtouint(skip_spaces(buf), 0, &t);
fs/ext4/sysfs.c
533
if (t > sbi->s_clusters_per_group)
fs/ext4/sysfs.c
535
*((unsigned int *) ptr) = t;
fs/ext4/sysfs.c
88
unsigned long t;
fs/ext4/sysfs.c
91
ret = kstrtoul(skip_spaces(buf), 0, &t);
fs/ext4/sysfs.c
95
if (t && (!is_power_of_2(t) || t > 0x40000000))
fs/ext4/sysfs.c
98
sbi->s_inode_readahead_blks = t;
fs/f2fs/f2fs.h
659
struct f2fs_dentry_ptr *d, struct f2fs_dentry_block *t)
fs/f2fs/f2fs.h
664
d->bitmap = t->dentry_bitmap;
fs/f2fs/f2fs.h
665
d->dentry = t->dentry;
fs/f2fs/f2fs.h
666
d->filename = t->filename;
fs/f2fs/f2fs.h
670
struct f2fs_dentry_ptr *d, void *t)
fs/f2fs/f2fs.h
679
d->bitmap = t;
fs/f2fs/f2fs.h
680
d->dentry = t + bitmap_size + reserved_size;
fs/f2fs/f2fs.h
681
d->filename = t + bitmap_size + reserved_size +
fs/f2fs/segment.c
785
enum dirty_type t = sentry->type;
fs/f2fs/segment.c
787
if (unlikely(t >= DIRTY)) {
fs/f2fs/segment.c
791
if (!test_and_set_bit(segno, dirty_i->dirty_segmap[t]))
fs/f2fs/segment.c
792
dirty_i->nr_dirty[t]++;
fs/f2fs/segment.c
821
enum dirty_type t = sentry->type;
fs/f2fs/segment.c
823
if (test_and_clear_bit(segno, dirty_i->dirty_segmap[t]))
fs/f2fs/segment.c
824
dirty_i->nr_dirty[t]--;
fs/f2fs/segment.h
27
#define IS_DATASEG(t) ((t) <= CURSEG_COLD_DATA)
fs/f2fs/segment.h
28
#define IS_NODESEG(t) ((t) >= CURSEG_HOT_NODE && (t) <= CURSEG_COLD_NODE)
fs/f2fs/super.c
562
#define QTYPE2NAME(t) (quotatypes[t])
fs/f2fs/sysfs.c
1099
unsigned long t;
fs/f2fs/sysfs.c
1102
ret = kstrtoul(skip_spaces(buf), 0, &t);
fs/f2fs/sysfs.c
1107
f2fs_reclaim_caches(t);
fs/f2fs/sysfs.c
484
unsigned long t;
fs/f2fs/sysfs.c
561
ret = kstrtoul(skip_spaces(buf), 0, &t);
fs/f2fs/sysfs.c
566
if (f2fs_build_fault_attr(sbi, 0, t, FAULT_TYPE))
fs/f2fs/sysfs.c
571
if (f2fs_build_fault_attr(sbi, t, 0, FAULT_RATE))
fs/f2fs/sysfs.c
576
if (f2fs_build_fault_attr(sbi, 0, t, FAULT_TIMEOUT))
fs/f2fs/sysfs.c
584
if (t > (unsigned long)(sbi->user_block_count -
fs/f2fs/sysfs.c
589
*ui = t;
fs/f2fs/sysfs.c
597
if (t > MAX_PLIST_NUM)
fs/f2fs/sysfs.c
601
if (t == *ui)
fs/f2fs/sysfs.c
603
*ui = t;
fs/f2fs/sysfs.c
608
if (t == 0 || t > MAX_PLIST_NUM)
fs/f2fs/sysfs.c
612
if (t == *ui)
fs/f2fs/sysfs.c
614
*ui = t;
fs/f2fs/sysfs.c
619
if (t == 0 || t > MAX_PLIST_NUM)
fs/f2fs/sysfs.c
623
*ui = t;
fs/f2fs/sysfs.c
628
if (t > 100)
fs/f2fs/sysfs.c
630
*ui = t;
fs/f2fs/sysfs.c
635
if (t >= DPOLICY_IO_AWARE_MAX)
fs/f2fs/sysfs.c
637
*ui = t;
fs/f2fs/sysfs.c
642
if (t == 0 || t > SEGS_PER_SEC(sbi))
fs/f2fs/sysfs.c
647
if (t == 0 || t > SEGS_PER_SEC(sbi))
fs/f2fs/sysfs.c
652
if (t == 0) {
fs/f2fs/sysfs.c
654
} else if (t == 1) {
fs/f2fs/sysfs.c
662
} else if (t == 2) {
fs/f2fs/sysfs.c
664
} else if (t == 3) {
fs/f2fs/sysfs.c
677
if (t == GC_IDLE_CB) {
fs/f2fs/sysfs.c
679
} else if (t == GC_IDLE_GREEDY) {
fs/f2fs/sysfs.c
681
} else if (t == GC_IDLE_AT) {
fs/f2fs/sysfs.c
693
sbi->gc_remaining_trials = t;
fs/f2fs/sysfs.c
700
if (t > 100)
fs/f2fs/sysfs.c
702
*ui = (unsigned int)t;
fs/f2fs/sysfs.c
707
if (t > 100)
fs/f2fs/sysfs.c
709
*ui = (unsigned int)t;
fs/f2fs/sysfs.c
714
if (t > 100)
fs/f2fs/sysfs.c
716
*ui = (unsigned int)t;
fs/f2fs/sysfs.c
722
sbi->iostat_enable = !!t;
fs/f2fs/sysfs.c
729
if (t < MIN_IOSTAT_PERIOD_MS || t > MAX_IOSTAT_PERIOD_MS)
fs/f2fs/sysfs.c
732
sbi->iostat_period_ms = (unsigned int)t;
fs/f2fs/sysfs.c
740
if (t < BLKZONE_ALLOC_PRIOR_SEQ || t > BLKZONE_ALLOC_PRIOR_CONV)
fs/f2fs/sysfs.c
742
sbi->blkzone_alloc_policy = t;
fs/f2fs/sysfs.c
750
if (t != 0)
fs/f2fs/sysfs.c
758
if (t != 0)
fs/f2fs/sysfs.c
765
if (t == 0 || t > 100)
fs/f2fs/sysfs.c
767
*ui = t;
fs/f2fs/sysfs.c
772
if (t == 0 || t > 100)
fs/f2fs/sysfs.c
774
*ui = t;
fs/f2fs/sysfs.c
780
if (t > 100)
fs/f2fs/sysfs.c
782
sbi->am.candidate_ratio = t;
fs/f2fs/sysfs.c
787
if (t > 100)
fs/f2fs/sysfs.c
789
sbi->am.age_weight = t;
fs/f2fs/sysfs.c
794
if (t < MAX_GC_MODE)
fs/f2fs/sysfs.c
795
sbi->gc_segment_mode = t;
fs/f2fs/sysfs.c
802
if (t > MAX_GC_FAILED_PINNED_FILES)
fs/f2fs/sysfs.c
804
sbi->gc_pin_file_threshold = t;
fs/f2fs/sysfs.c
809
if (t != 0)
fs/f2fs/sysfs.c
816
if (t >= MIN_RA_MUL && t <= MAX_RA_MUL)
fs/f2fs/sysfs.c
817
sbi->seq_file_ra_mul = t;
fs/f2fs/sysfs.c
824
if (t >= MIN_FRAGMENT_SIZE && t <= MAX_FRAGMENT_SIZE)
fs/f2fs/sysfs.c
825
sbi->max_fragment_chunk = t;
fs/f2fs/sysfs.c
832
if (t >= MIN_FRAGMENT_SIZE && t <= MAX_FRAGMENT_SIZE)
fs/f2fs/sysfs.c
833
sbi->max_fragment_hole = t;
fs/f2fs/sysfs.c
840
if (t != 0)
fs/f2fs/sysfs.c
847
if (t != 0)
fs/f2fs/sysfs.c
854
if (t != 0)
fs/f2fs/sysfs.c
861
sbi->readdir_ra = !!t;
fs/f2fs/sysfs.c
866
if (t == 0 || t >= sbi->warm_data_age_threshold)
fs/f2fs/sysfs.c
868
if (t == *ui)
fs/f2fs/sysfs.c
870
*ui = (unsigned int)t;
fs/f2fs/sysfs.c
875
if (t <= sbi->hot_data_age_threshold)
fs/f2fs/sysfs.c
877
if (t == *ui)
fs/f2fs/sysfs.c
879
*ui = (unsigned int)t;
fs/f2fs/sysfs.c
884
if (t > 100)
fs/f2fs/sysfs.c
886
if (t == *ui)
fs/f2fs/sysfs.c
888
*ui = (unsigned int)t;
fs/f2fs/sysfs.c
893
if (t > UINT_MAX)
fs/f2fs/sysfs.c
895
*ui = (unsigned int)t;
fs/f2fs/sysfs.c
900
if (t >= BIT(F2FS_IPU_MAX))
fs/f2fs/sysfs.c
903
if (f2fs_lfs_mode(sbi) && (t & ~BIT(F2FS_IPU_NOCACHE)))
fs/f2fs/sysfs.c
905
SM_I(sbi)->ipu_policy = (unsigned int)t;
fs/f2fs/sysfs.c
910
if (t > MAX_DIR_HASH_DEPTH)
fs/f2fs/sysfs.c
912
sbi->dir_level = t;
fs/f2fs/sysfs.c
917
if (t > GET_SEC_FROM_SEG(sbi, overprovision_segments(sbi)))
fs/f2fs/sysfs.c
919
*ui = (unsigned int)t;
fs/f2fs/sysfs.c
924
if (t < 1 || t > SEGS_PER_SEC(sbi))
fs/f2fs/sysfs.c
926
sbi->gc_thread->boost_gc_multiple = (unsigned int)t;
fs/f2fs/sysfs.c
931
if (t > GC_GREEDY)
fs/f2fs/sysfs.c
933
sbi->gc_thread->boost_gc_greedy = (unsigned int)t;
fs/f2fs/sysfs.c
938
if (t < AWARE_ALL_IO || t > AWARE_NONE)
fs/f2fs/sysfs.c
940
sbi->bggc_io_aware = t;
fs/f2fs/sysfs.c
945
if (t < 0 || t > MAIN_SECS(sbi))
fs/f2fs/sysfs.c
947
sbi->allocate_section_hint = t;
fs/f2fs/sysfs.c
952
if (t < ALLOCATE_FORWARD_NOHINT || t > ALLOCATE_FORWARD_FROM_HINT)
fs/f2fs/sysfs.c
954
sbi->allocate_section_policy = t;
fs/f2fs/sysfs.c
959
if (t >= BIT(LOCK_NAME_MAX - 1))
fs/f2fs/sysfs.c
961
sbi->adjust_lock_priority = t;
fs/f2fs/sysfs.c
966
if (t < NICE_TO_PRIO(MIN_NICE) || t > NICE_TO_PRIO(MAX_NICE))
fs/f2fs/sysfs.c
968
sbi->lock_duration_priority = t;
fs/f2fs/sysfs.c
973
if (t < NICE_TO_PRIO(MIN_NICE) || t > NICE_TO_PRIO(MAX_NICE))
fs/f2fs/sysfs.c
977
sbi->critical_task_priority = t;
fs/f2fs/sysfs.c
987
__sbi_store_value(a, sbi, ptr + a->offset, t);
fs/fat/dir.c
195
fat_short2uni(struct nls_table *t, unsigned char *c, int clen, wchar_t *uni)
fs/fat/dir.c
199
charlen = t->char2uni(c, clen, uni);
fs/fat/dir.c
208
fat_short2lower_uni(struct nls_table *t, unsigned char *c,
fs/fat/dir.c
214
charlen = t->char2uni(c, clen, &wc);
fs/fat/dir.c
219
unsigned char nc = t->charset2lower[*c];
fs/fat/dir.c
224
charlen = t->char2uni(&nc, 1, uni);
fs/fat/fat_test.c
123
static void time_testcase_desc(struct fat_timestamp_testcase *t,
fs/fat/fat_test.c
126
strscpy(desc, t->name, KUNIT_PARAM_DESC_SIZE);
fs/fat/namei_vfat.c
134
struct nls_table *t = MSDOS_SB(dentry->d_sb)->nls_io;
fs/fat/namei_vfat.c
144
hash = partial_name_hash(nls_tolower(t, *name++), hash);
fs/fat/namei_vfat.c
156
struct nls_table *t = MSDOS_SB(dentry->d_sb)->nls_io;
fs/fat/namei_vfat.c
163
if (nls_strnicmp(t, name->name, str, alen) == 0)
fs/file_table.c
489
struct file *f, *t;
fs/file_table.c
491
llist_for_each_entry_safe(f, t, node, f_llist)
fs/fs-writeback.c
1477
static bool inode_dirtied_after(struct inode *inode, unsigned long t)
fs/fs-writeback.c
1479
bool ret = time_after(inode->dirtied_when, t);
fs/fuse/control.c
81
unsigned long t;
fs/fuse/control.c
88
err = kstrtoul_from_user(buf, count, 0, &t);
fs/fuse/control.c
95
if (t > limit)
fs/fuse/control.c
98
*val = t;
fs/gfs2/bmap.c
296
const __be64 *t;
fs/gfs2/bmap.c
298
for (t = start; t < end; t++) {
fs/gfs2/bmap.c
301
if (!*t)
fs/gfs2/bmap.c
304
rabh = gfs2_getbuf(gl, be64_to_cpu(*t), CREATE);
fs/gfs2/log.c
1292
unsigned long t = 1;
fs/gfs2/log.c
1299
if (gfs2_jrnl_flush_reqd(sdp) || t == 0) {
fs/gfs2/log.c
1315
t = gfs2_tune_get(sdp, gt_logd_secs) * HZ;
fs/gfs2/log.c
1317
t = wait_event_freezable_timeout(sdp->sd_logd_waitq,
fs/gfs2/log.c
1323
t);
fs/gfs2/quota.c
1585
unsigned long t;
fs/gfs2/quota.c
1615
t = min(statfs_deadline - now, quotad_deadline - now);
fs/gfs2/quota.c
1620
t);
fs/gfs2/quota.c
1623
t = 0;
fs/gfs2/util.c
394
u16 type, u16 t, const char *function,
fs/gfs2/util.c
401
(unsigned long long)bh->b_blocknr, type, t,
fs/gfs2/util.h
102
u16 type, u16 t,
fs/gfs2/util.h
114
u16 t = be32_to_cpu(mh->mh_type);
fs/gfs2/util.h
120
if (unlikely(t != type)) {
fs/gfs2/util.h
121
gfs2_metatype_check_ii(sdp, bh, type, t, function,
fs/hfsplus/hfsplus_fs.h
598
#define hfsp_mt2ut(t) (struct timespec64){ .tv_sec = __hfsp_mt2ut(t) }
fs/hfsplus/hfsplus_fs.h
599
#define hfsp_ut2mt(t) __hfsp_ut2mt((t).tv_sec)
fs/hfsplus/unicode.c
364
int l, v, t;
fs/hfsplus/unicode.c
372
t = Hangul_TBase + index % Hangul_TCount;
fs/hfsplus/unicode.c
376
if (t != Hangul_TBase) {
fs/hfsplus/unicode.c
377
result[2] = t;
fs/hpfs/dnode.c
277
loff_t t;
fs/hpfs/dnode.c
279
t = get_pos(d, de);
fs/hpfs/dnode.c
280
for_all_poss(i, hpfs_pos_ins, t, 1);
fs/hpfs/dnode.c
281
for_all_poss(i, hpfs_pos_subst, 4, t);
fs/hpfs/dnode.c
282
for_all_poss(i, hpfs_pos_subst, 5, t + 1);
fs/hpfs/dnode.c
441
loff_t t;
fs/hpfs/dnode.c
493
t = get_pos(dnode, de);
fs/hpfs/dnode.c
494
for_all_poss(i, hpfs_pos_subst, t, 4);
fs/hpfs/dnode.c
495
for_all_poss(i, hpfs_pos_subst, t + 1, 5);
fs/hpfs/dnode.c
716
loff_t t;
fs/hpfs/dnode.c
729
for_all_poss(i, hpfs_pos_del, (t = get_pos(dnode, de)) + 1, 1);
fs/hpfs/dnode.c
735
for_all_poss(i, hpfs_pos_subst, 5, t);
fs/hpfs/dnode.c
86
static void hpfs_pos_subst(loff_t *p, loff_t f, loff_t t)
fs/hpfs/dnode.c
88
if (*p == f) *p = t;
fs/hpfs/dnode.c
926
int t = hpfs_compare_names(inode->i_sb, name, len, de->name, de->namelen, de->last);
fs/hpfs/dnode.c
927
if (!t) {
fs/hpfs/dnode.c
931
if (t < 0) {
fs/hpfs/hpfs_fn.h
337
static inline time64_t local_to_gmt(struct super_block *s, time64_t t)
fs/hpfs/hpfs_fn.h
340
return t + sys_tz.tz_minuteswest * 60 + hpfs_sb(s)->sb_timeshift;
fs/hpfs/hpfs_fn.h
343
static inline time32_t gmt_to_local(struct super_block *s, time64_t t)
fs/hpfs/hpfs_fn.h
346
return t - sys_tz.tz_minuteswest * 60 - hpfs_sb(s)->sb_timeshift;
fs/hpfs/namei.c
14
time64_t t = local_to_gmt(dir->i_sb, local_get_seconds(dir->i_sb));
fs/hpfs/namei.c
15
if (t == inode_get_mtime_sec(dir) &&
fs/hpfs/namei.c
16
t == inode_get_ctime_sec(dir))
fs/hpfs/namei.c
18
inode_set_mtime_to_ts(dir, inode_set_ctime(dir, t, 0));
fs/inode.c
2793
struct timespec64 timestamp_truncate(struct timespec64 t, struct inode *inode)
fs/inode.c
2798
t.tv_sec = clamp(t.tv_sec, sb->s_time_min, sb->s_time_max);
fs/inode.c
2799
if (unlikely(t.tv_sec == sb->s_time_max || t.tv_sec == sb->s_time_min))
fs/inode.c
2800
t.tv_nsec = 0;
fs/inode.c
2806
t.tv_nsec = 0;
fs/inode.c
2808
t.tv_nsec -= t.tv_nsec % gran;
fs/inode.c
2811
return t;
fs/jbd2/checkpoint.c
202
transaction_t *t = jh->b_transaction;
fs/jbd2/checkpoint.c
203
tid_t tid = t->t_tid;
fs/jbd2/journal.c
135
static void commit_timeout(struct timer_list *t)
fs/jbd2/journal.c
137
journal_t *journal = timer_container_of(journal, t, j_commit_timer);
fs/jbd2/journal.c
259
struct task_struct *t;
fs/jbd2/journal.c
261
t = kthread_run(kjournald2, journal, "jbd2/%s",
fs/jbd2/journal.c
263
if (IS_ERR(t))
fs/jbd2/journal.c
264
return PTR_ERR(t);
fs/jbd2/transaction.c
212
transaction_t *t = journal->j_running_transaction;
fs/jbd2/transaction.c
220
if (t->t_state != T_RUNNING) {
fs/jbd2/transaction.c
221
WARN_ON_ONCE(t->t_state >= T_FLUSH);
fs/jbd2/transaction.c
232
needed = atomic_add_return(total, &t->t_outstanding_credits);
fs/jbd2/transaction.c
239
atomic_sub(total, &t->t_outstanding_credits);
fs/jbd2/transaction.c
273
atomic_sub(total, &t->t_outstanding_credits);
fs/jbd2/transaction.c
293
atomic_sub(total, &t->t_outstanding_credits);
fs/jbd2/transaction.c
544
static void __jbd2_journal_unreserve_handle(handle_t *handle, transaction_t *t)
fs/jbd2/transaction.c
550
if (t)
fs/jbd2/transaction.c
551
atomic_sub(handle->h_total_credits, &t->t_outstanding_credits);
fs/jffs2/os-linux.h
34
#define JFFS2_CLAMP_TIME(t) ((uint32_t)clamp_t(time64_t, (t), 0, U32_MAX))
fs/jfs/jfs_dtree.c
2733
struct dtslot *t;
fs/jfs/jfs_dtree.c
2974
t = (struct dtslot *) & p->slot[next];
fs/jfs/jfs_dtree.c
2987
outlen = jfs_strfromUCS_le(name_ptr, t->name,
fs/jfs/jfs_dtree.c
2991
next = t->next;
fs/jfs/jfs_dtree.c
3323
struct dtslot *t;
fs/jfs/jfs_dtree.c
3362
t = (struct dtslot *) & p->slot[si];
fs/jfs/jfs_dtree.c
3365
name = t->name;
fs/jfs/jfs_dtree.c
3372
si = t->next;
fs/jfs/jfs_dtree.c
3401
struct dtslot *t;
fs/jfs/jfs_dtree.c
3465
t = (struct dtslot *) & p->slot[si];
fs/jfs/jfs_dtree.c
3468
name = t->name;
fs/jfs/jfs_dtree.c
3482
si = t->next;
fs/jfs/jfs_dtree.c
3573
struct dtslot *t;
fs/jfs/jfs_dtree.c
3611
t = &p->slot[si];
fs/jfs/jfs_dtree.c
3615
UniStrncpy_from_le(kname, t->name, len);
fs/jfs/jfs_dtree.c
3617
si = t->next;
fs/jfs/jfs_dtree.c
3633
struct dtslot *h, *t;
fs/jfs/jfs_dtree.c
3695
t = h;
fs/jfs/jfs_dtree.c
3700
t = &p->slot[fsi];
fs/jfs/jfs_dtree.c
3701
p->header.freelist = t->next;
fs/jfs/jfs_dtree.c
3724
UniStrncpy_to_le(t->name, kname, len);
fs/jfs/jfs_dtree.c
3738
if (h == t) {
fs/jfs/jfs_dtree.c
3746
t->next = -1;
fs/jfs/jfs_dtree.c
3993
struct dtslot *t;
fs/jfs/jfs_dtree.c
4011
t = &p->slot[fsi];
fs/jfs/jfs_dtree.c
4013
si = ((struct ldtentry *) t)->next;
fs/jfs/jfs_dtree.c
4015
si = ((struct idtentry *) t)->next;
fs/jfs/jfs_dtree.c
4016
t->next = si;
fs/jfs/jfs_dtree.c
4017
t->cnt = 1;
fs/jfs/jfs_dtree.c
4046
t = &p->slot[si];
fs/jfs/jfs_dtree.c
4047
t->cnt = 1;
fs/jfs/jfs_dtree.c
4048
si = t->next;
fs/jfs/jfs_dtree.c
4058
t->next = p->header.freelist;
fs/jfs/jfs_dtree.c
4088
struct dtslot *t;
fs/jfs/jfs_dtree.c
4106
t = &p->slot[tsi];
fs/jfs/jfs_dtree.c
4108
((struct idtentry *) t)->namlen = 0;
fs/jfs/jfs_dtree.c
4109
si = ((struct idtentry *) t)->next;
fs/jfs/jfs_dtree.c
4110
((struct idtentry *) t)->next = -1;
fs/jfs/jfs_dtree.c
4141
t = &p->slot[si];
fs/jfs/jfs_dtree.c
4142
t->cnt = 1;
fs/jfs/jfs_dtree.c
4143
si = t->next;
fs/jfs/jfs_dtree.c
4155
t->next = p->header.freelist;
fs/jfs/jfs_dtree.c
4169
struct dtslot *t;
fs/jfs/jfs_dtree.c
4188
t = &p->slot[fsi];
fs/jfs/jfs_dtree.c
4189
si = t->next;
fs/jfs/jfs_dtree.c
4214
t = &p->slot[si];
fs/jfs/jfs_dtree.c
4215
si = t->next;
fs/jfs/jfs_logmgr.c
1370
lp->h.eor = lp->t.eor = cpu_to_le16(bp->l_eor);
fs/jfs/jfs_logmgr.c
1657
lp->h.eor = lp->t.eor = cpu_to_le16(bp->l_eor);
fs/jfs/jfs_logmgr.c
2432
lp->h.page = lp->t.page = cpu_to_le32(npages - 3);
fs/jfs/jfs_logmgr.c
2433
lp->h.eor = lp->t.eor = cpu_to_le16(LOGPHDRSIZE + LOGRDSIZE);
fs/jfs/jfs_logmgr.c
2452
lp->h.page = lp->t.page = cpu_to_le32(lspn);
fs/jfs/jfs_logmgr.c
2453
lp->h.eor = lp->t.eor = cpu_to_le16(LOGPHDRSIZE);
fs/jfs/jfs_logmgr.c
615
lp->h.eor = lp->t.eor = cpu_to_le16(bp->l_ceor);
fs/jfs/jfs_logmgr.c
636
lp->h.page = lp->t.page = cpu_to_le32(lspn + 1);
fs/jfs/jfs_logmgr.c
637
lp->h.eor = lp->t.eor = cpu_to_le16(LOGPHDRSIZE);
fs/jfs/jfs_logmgr.c
776
lp->h.eor = lp->t.eor = cpu_to_le16(bp->l_ceor);
fs/jfs/jfs_logmgr.c
784
lp->h.eor = lp->t.eor = cpu_to_le16(bp->l_ceor);
fs/jfs/jfs_logmgr.c
868
lp->h.eor = lp->t.eor = cpu_to_le16(bp->l_eor);
fs/jfs/jfs_logmgr.h
122
} t;
fs/jfs/jfs_txnmgr.c
351
tid_t t;
fs/jfs/jfs_txnmgr.c
395
if ((t = TxAnchor.freetid) == 0) {
fs/jfs/jfs_txnmgr.c
402
tblk = tid_to_tblock(t);
fs/jfs/jfs_txnmgr.c
432
HIGHWATERMARK(stattx.maxtid, t); /* statistics */
fs/jfs/jfs_txnmgr.c
437
jfs_info("txBegin: returning tid = %d", t);
fs/jfs/jfs_txnmgr.c
439
return t;
fs/namespace.c
1326
struct mount *m, *t;
fs/namespace.c
1328
llist_for_each_entry_safe(m, t, node, mnt_llist)
fs/nfs/flexfilelayout/flexfilelayout.c
2793
ktime_t t)
fs/nfs/flexfilelayout/flexfilelayout.c
2799
ts = ktime_to_timespec64(t);
fs/nfs/internal.h
221
int nfs_init_server_rpcclient(struct nfs_server *, const struct rpc_timeout *t,
fs/nfs/nfs4xdr.c
1077
xdr_encode_nfstime4(__be32 *p, const struct timespec64 *t)
fs/nfs/nfs4xdr.c
1079
p = xdr_encode_hyper(p, t->tv_sec);
fs/nfs/nfs4xdr.c
1080
*p++ = cpu_to_be32(t->tv_nsec);
fs/nfs/nfs4xdr.c
4182
xdr_decode_nfstime4(__be32 *p, struct timespec64 *t)
fs/nfs/nfs4xdr.c
4187
t-> tv_sec = sec;
fs/nfs/nfs4xdr.c
4188
t->tv_nsec = be32_to_cpup(p++);
fs/nfs/pnfs.c
2007
struct nfs4_threshold *t = ctx->mdsthreshold;
fs/nfs/pnfs.c
2012
if (t == NULL)
fs/nfs/pnfs.c
2016
__func__, t->bm, t->rd_sz, t->wr_sz, t->rd_io_sz, t->wr_io_sz);
fs/nfs/pnfs.c
2020
if (t->bm & THRESHOLD_RD) {
fs/nfs/pnfs.c
2023
if (fsize < t->rd_sz)
fs/nfs/pnfs.c
2026
if (t->bm & THRESHOLD_RD_IO) {
fs/nfs/pnfs.c
2030
if (nfsi->read_io < t->rd_io_sz)
fs/nfs/pnfs.c
2035
if (t->bm & THRESHOLD_WR) {
fs/nfs/pnfs.c
2038
if (fsize < t->wr_sz)
fs/nfs/pnfs.c
2041
if (t->bm & THRESHOLD_WR_IO) {
fs/nfs/pnfs.c
2045
if (nfsi->write_io < t->wr_io_sz)
fs/nfsd/nfs4layouts.c
553
struct nfs4_layout *lp, *t;
fs/nfsd/nfs4layouts.c
569
list_for_each_entry_safe(lp, t, &ls->ls_layouts, lo_perstate) {
fs/nfsd/nfs4state.c
2731
find_stateid_locked(struct nfs4_client *cl, stateid_t *t)
fs/nfsd/nfs4state.c
2735
ret = idr_find(&cl->cl_stateids, t->si_opaque.so_id);
fs/nfsd/nfs4state.c
2742
find_stateid_by_type(struct nfs4_client *cl, stateid_t *t,
fs/nfsd/nfs4state.c
2748
s = find_stateid_locked(cl, t);
fs/nfsd/nfs4state.c
7033
time64_t t;
fs/nfsd/nfs4state.c
7038
t = nfs4_laundromat(nn);
fs/nfsd/nfs4state.c
7039
queue_delayed_work(laundry_wq, &nn->laundromat_work, t*HZ);
fs/nilfs2/dir.c
268
unsigned char t;
fs/nilfs2/dir.c
270
t = fs_ftype_to_dtype(de->file_type);
fs/nilfs2/dir.c
273
le64_to_cpu(de->inode), t)) {
fs/nilfs2/segbuf.h
93
#define nilfs_for_each_segbuf_before(s, t, h) \
fs/nilfs2/segbuf.h
94
for ((s) = NILFS_FIRST_SEGBUF(h); (s) != (t); \
fs/nilfs2/segment.c
2486
static void nilfs_construction_timeout(struct timer_list *t)
fs/nilfs2/segment.c
2488
struct nilfs_sc_info *sci = timer_container_of(sci, t, sc_timer);
fs/nilfs2/segment.c
2836
struct task_struct *t;
fs/nilfs2/segment.c
2854
t = kthread_create(nilfs_segctor_thread, sci, "segctord");
fs/nilfs2/segment.c
2855
if (IS_ERR(t)) {
fs/nilfs2/segment.c
2856
err = PTR_ERR(t);
fs/nilfs2/segment.c
2861
sci->sc_task = t;
fs/nilfs2/sufile.c
49
__u64 t = segnum + NILFS_MDT(sufile)->mi_first_entry_offset;
fs/nilfs2/sufile.c
51
t = div64_ul(t, nilfs_sufile_segment_usages_per_block(sufile));
fs/nilfs2/sufile.c
52
return (unsigned long)t;
fs/nilfs2/sufile.c
58
__u64 t = segnum + NILFS_MDT(sufile)->mi_first_entry_offset;
fs/nilfs2/sufile.c
60
return do_div(t, nilfs_sufile_segment_usages_per_block(sufile));
fs/nilfs2/super.c
281
time64_t t;
fs/nilfs2/super.c
284
t = ktime_get_real_seconds();
fs/nilfs2/super.c
285
nilfs->ns_sbwtime = t;
fs/nilfs2/super.c
286
sbp[0]->s_wtime = cpu_to_le64(t);
fs/nilfs2/the_nilfs.h
253
u64 t = ktime_get_real_seconds();
fs/nilfs2/the_nilfs.h
255
return t < nilfs->ns_sbwtime ||
fs/nilfs2/the_nilfs.h
256
t > nilfs->ns_sbwtime + nilfs->ns_sb_update_freq;
fs/nls/nls_base.c
103
for (t = utf8_table; t->cmask && maxout; t++, maxout--) {
fs/nls/nls_base.c
105
if (l <= t->lmask) {
fs/nls/nls_base.c
106
c = t->shift;
fs/nls/nls_base.c
107
*s = (u8) (t->cval | (l >> c));
fs/nls/nls_base.c
59
const struct utf8_table *t;
fs/nls/nls_base.c
64
for (t = utf8_table; t->cmask; t++) {
fs/nls/nls_base.c
66
if ((c0 & t->cmask) == t->cval) {
fs/nls/nls_base.c
67
l &= t->lmask;
fs/nls/nls_base.c
68
if (l < t->lval || l > UNICODE_MAX ||
fs/nls/nls_base.c
93
const struct utf8_table *t;
fs/ntfs3/bitmap.c
1146
size_t t = max_alloc + wbits - 1;
fs/ntfs3/bitmap.c
1148
nwnd = likely(t > max_alloc) ? (t >> log2_bits) : wnd->nwnd;
fs/ntfs3/bitmap.c
1171
size_t t = max_alloc & (wbits - 1);
fs/ntfs3/bitmap.c
1173
if (t) {
fs/ntfs3/bitmap.c
1174
wbits = t;
fs/ntfs3/fslog.c
621
static inline void *enum_rstbl(struct RESTART_TABLE *t, void *c)
fs/ntfs3/fslog.c
627
if (!t)
fs/ntfs3/fslog.c
630
rsize = le16_to_cpu(t->size);
fs/ntfs3/fslog.c
634
if (!t->total)
fs/ntfs3/fslog.c
636
e = Add2Ptr(t, sizeof(struct RESTART_TABLE));
fs/ntfs3/fslog.c
642
for (bprt = bytes_per_rt(t); PtrOffset(t, e) < bprt;
fs/ntfs3/fslog.c
816
struct RESTART_TABLE *t = kzalloc(bytes, GFP_NOFS);
fs/ntfs3/fslog.c
818
if (!t)
fs/ntfs3/fslog.c
821
t->size = cpu_to_le16(esize);
fs/ntfs3/fslog.c
822
t->used = cpu_to_le16(used);
fs/ntfs3/fslog.c
823
t->free_goal = cpu_to_le32(~0u);
fs/ntfs3/fslog.c
824
t->first_free = cpu_to_le32(sizeof(struct RESTART_TABLE));
fs/ntfs3/fslog.c
825
t->last_free = cpu_to_le32(lf);
fs/ntfs3/fslog.c
827
e = (__le32 *)(t + 1);
fs/ntfs3/fslog.c
828
last_free = Add2Ptr(t, lf);
fs/ntfs3/fslog.c
834
return t;
fs/ntfs3/fslog.c
878
struct RESTART_TABLE *t = *tbl;
fs/ntfs3/fslog.c
880
if (!t->first_free) {
fs/ntfs3/fslog.c
881
*tbl = t = extend_rsttbl(t, 16, ~0u);
fs/ntfs3/fslog.c
882
if (!t)
fs/ntfs3/fslog.c
886
off = le32_to_cpu(t->first_free);
fs/ntfs3/fslog.c
889
e = Add2Ptr(t, off);
fs/ntfs3/fslog.c
891
t->first_free = *e;
fs/ntfs3/fslog.c
893
memset(e, 0, le16_to_cpu(t->size));
fs/ntfs3/fslog.c
898
if (!t->first_free)
fs/ntfs3/fslog.c
899
t->last_free = 0;
fs/ntfs3/fslog.c
901
le16_add_cpu(&t->total, 1);
fs/ntfs3/fslog.c
903
return Add2Ptr(t, off);
fs/ntfs3/ntfs_fs.h
1050
s64 t = le64_to_cpu(tm) - _100ns2seconds * SecondsToStartOf1970;
fs/ntfs3/ntfs_fs.h
1052
ts->tv_sec = div_s64_rem(t, _100ns2seconds, &t32);
fs/ntfs3/run.c
342
struct ntfs_run *t = run->runs + index - 1;
fs/ntfs3/run.c
344
if (t->vcn + t->len == vcn &&
fs/ntfs3/run.c
345
(t->lcn == SPARSE_LCN) == (lcn == SPARSE_LCN) &&
fs/ntfs3/run.c
346
(lcn == SPARSE_LCN || lcn == t->lcn + t->len)) {
fs/ntfs3/super.c
1262
struct ATTR_DEF_ENTRY *t;
fs/ntfs3/super.c
1542
sbi->def_table = t = kvmalloc(bytes, GFP_KERNEL);
fs/ntfs3/super.c
1543
if (!t) {
fs/ntfs3/super.c
1555
if (ATTR_STD != t->type) {
fs/ntfs3/super.c
1561
t += 1;
fs/ntfs3/super.c
1566
u32 t32 = le32_to_cpu(t->type);
fs/ntfs3/super.c
1567
u64 sz = le64_to_cpu(t->max_sz);
fs/ntfs3/super.c
1569
if ((t32 & 0xF) || le32_to_cpu(t[-1].type) >= t32)
fs/ntfs3/super.c
1572
if (t->type == ATTR_REPARSE)
fs/ntfs3/super.c
1574
else if (t->type == ATTR_EA)
fs/ntfs3/super.c
1578
t += 1;
fs/ocfs2/cluster/tcp.c
128
static void o2net_idle_timer(struct timer_list *t);
fs/ocfs2/cluster/tcp.c
1489
static void o2net_idle_timer(struct timer_list *t)
fs/ocfs2/cluster/tcp.c
1491
struct o2net_sock_container *sc = timer_container_of(sc, t,
fs/orangefs/orangefs-bufmap.c
102
t = schedule_timeout(n);
fs/orangefs/orangefs-bufmap.c
104
if (unlikely(!t) && n != left && m->c < 0)
fs/orangefs/orangefs-bufmap.c
105
left = t;
fs/orangefs/orangefs-bufmap.c
107
left = t + (left - n);
fs/orangefs/orangefs-bufmap.c
87
long n = left, t;
fs/pipe.c
1322
unsigned int t = tail & mask;
fs/pipe.c
1323
if (h > t) {
fs/pipe.c
1324
memcpy(bufs, pipe->bufs + t,
fs/pipe.c
1327
unsigned int tsize = pipe->ring_size - t;
fs/pipe.c
1331
memcpy(bufs, pipe->bufs + t,
fs/proc/array.c
550
struct task_struct *t;
fs/proc/array.c
556
__for_each_thread(sig, t) {
fs/proc/array.c
557
min_flt += t->min_flt;
fs/proc/array.c
558
maj_flt += t->maj_flt;
fs/proc/array.c
559
gtime += task_gtime(t);
fs/proc/base.c
3048
struct task_struct *t;
fs/proc/base.c
3053
__for_each_thread(sig, t)
fs/proc/base.c
3054
task_io_accounting_add(&acct, &t->ioac);
fs/resctrl/ctrlmondata.c
224
enum resctrl_conf_type t = s->conf_type;
fs/resctrl/ctrlmondata.c
272
cfg = &d->staged_config[t];
fs/resctrl/rdtgroup.c
2993
struct task_struct *p, *t;
fs/resctrl/rdtgroup.c
2996
for_each_process_thread(p, t) {
fs/resctrl/rdtgroup.c
2997
if (!from || is_closid_match(t, from) ||
fs/resctrl/rdtgroup.c
2998
is_rmid_match(t, from)) {
fs/resctrl/rdtgroup.c
2999
resctrl_arch_set_closid_rmid(t, to->closid,
fs/resctrl/rdtgroup.c
3017
if (IS_ENABLED(CONFIG_SMP) && mask && task_curr(t))
fs/resctrl/rdtgroup.c
3018
cpumask_set_cpu(task_cpu(t), mask);
fs/resctrl/rdtgroup.c
3528
enum resctrl_conf_type t = s->conf_type;
fs/resctrl/rdtgroup.c
3537
cfg = &d->staged_config[t];
fs/resctrl/rdtgroup.c
610
static void update_task_closid_rmid(struct task_struct *t)
fs/resctrl/rdtgroup.c
612
if (IS_ENABLED(CONFIG_SMP) && task_curr(t))
fs/resctrl/rdtgroup.c
613
smp_call_function_single(task_cpu(t), _update_task_closid_rmid, t, 1);
fs/resctrl/rdtgroup.c
615
_update_task_closid_rmid(t);
fs/resctrl/rdtgroup.c
680
static bool is_closid_match(struct task_struct *t, struct rdtgroup *r)
fs/resctrl/rdtgroup.c
683
resctrl_arch_match_closid(t, r->closid));
fs/resctrl/rdtgroup.c
686
static bool is_rmid_match(struct task_struct *t, struct rdtgroup *r)
fs/resctrl/rdtgroup.c
689
resctrl_arch_match_rmid(t, r->mon.parent->closid,
fs/resctrl/rdtgroup.c
701
struct task_struct *p, *t;
fs/resctrl/rdtgroup.c
707
for_each_process_thread(p, t) {
fs/resctrl/rdtgroup.c
708
if (is_closid_match(t, r) || is_rmid_match(t, r)) {
fs/resctrl/rdtgroup.c
821
struct task_struct *p, *t;
fs/resctrl/rdtgroup.c
825
for_each_process_thread(p, t) {
fs/resctrl/rdtgroup.c
826
if (is_closid_match(t, r) || is_rmid_match(t, r)) {
fs/resctrl/rdtgroup.c
827
pid = task_pid_vnr(t);
fs/smb/client/cifsproto.h
158
u64 cifs_UnixTimeToNT(struct timespec64 t);
fs/smb/client/dfs_cache.c
118
struct cache_dfs_tgt *t, *n;
fs/smb/client/dfs_cache.c
120
list_for_each_entry_safe(t, n, &ce->tlist, list) {
fs/smb/client/dfs_cache.c
121
list_del(&t->list);
fs/smb/client/dfs_cache.c
122
kfree(t->name);
fs/smb/client/dfs_cache.c
123
kfree(t);
fs/smb/client/dfs_cache.c
160
struct cache_dfs_tgt *t;
fs/smb/client/dfs_cache.c
179
list_for_each_entry(t, &ce->tlist, list) {
fs/smb/client/dfs_cache.c
181
t->name,
fs/smb/client/dfs_cache.c
182
READ_ONCE(ce->tgthint) == t ? " (target hint)" : "");
fs/smb/client/dfs_cache.c
229
struct cache_dfs_tgt *t;
fs/smb/client/dfs_cache.c
232
list_for_each_entry(t, &ce->tlist, list) {
fs/smb/client/dfs_cache.c
233
cifs_dbg(FYI, " %s%s\n", t->name,
fs/smb/client/dfs_cache.c
234
READ_ONCE(ce->tgthint) == t ? " (target hint)" : "");
fs/smb/client/dfs_cache.c
343
struct cache_dfs_tgt *t = READ_ONCE(ce->tgthint);
fs/smb/client/dfs_cache.c
345
return t ? t->name : ERR_PTR(-ENOENT);
fs/smb/client/dfs_cache.c
364
struct cache_dfs_tgt *t;
fs/smb/client/dfs_cache.c
366
t = kmalloc_obj(*t, GFP_ATOMIC);
fs/smb/client/dfs_cache.c
367
if (!t)
fs/smb/client/dfs_cache.c
369
t->name = kstrdup(name, GFP_ATOMIC);
fs/smb/client/dfs_cache.c
370
if (!t->name) {
fs/smb/client/dfs_cache.c
371
kfree(t);
fs/smb/client/dfs_cache.c
374
t->path_consumed = path_consumed;
fs/smb/client/dfs_cache.c
375
INIT_LIST_HEAD(&t->list);
fs/smb/client/dfs_cache.c
376
return t;
fs/smb/client/dfs_cache.c
397
struct cache_dfs_tgt *t;
fs/smb/client/dfs_cache.c
399
t = alloc_target(refs[i].node_name, refs[i].path_consumed);
fs/smb/client/dfs_cache.c
400
if (IS_ERR(t)) {
fs/smb/client/dfs_cache.c
402
return PTR_ERR(t);
fs/smb/client/dfs_cache.c
404
if (tgthint && !strcasecmp(t->name, tgthint)) {
fs/smb/client/dfs_cache.c
405
list_add(&t->list, &ce->tlist);
fs/smb/client/dfs_cache.c
408
list_add_tail(&t->list, &ce->tlist);
fs/smb/client/dfs_cache.c
792
struct cache_dfs_tgt *t;
fs/smb/client/dfs_cache.c
798
list_for_each_entry(t, &ce->tlist, list) {
fs/smb/client/dfs_cache.c
805
it->it_name = kstrdup(t->name, GFP_ATOMIC);
fs/smb/client/dfs_cache.c
811
it->it_path_consumed = t->path_consumed;
fs/smb/client/dfs_cache.c
813
if (READ_ONCE(ce->tgthint) == t)
fs/smb/client/dfs_cache.c
946
struct cache_dfs_tgt *t;
fs/smb/client/dfs_cache.c
960
t = READ_ONCE(ce->tgthint);
fs/smb/client/dfs_cache.c
962
if (unlikely(!strcasecmp(it->it_name, t->name)))
fs/smb/client/dfs_cache.c
965
list_for_each_entry(t, &ce->tlist, list) {
fs/smb/client/dfs_cache.c
966
if (!strcasecmp(t->name, it->it_name)) {
fs/smb/client/dfs_cache.c
967
WRITE_ONCE(ce->tgthint, t);
fs/smb/client/netmisc.c
125
s64 t = le64_to_cpu(ntutc) - NTFS_TIME_OFFSET;
fs/smb/client/netmisc.c
133
if (t < 0) {
fs/smb/client/netmisc.c
134
abs_t = -t;
fs/smb/client/netmisc.c
139
abs_t = t;
fs/smb/client/netmisc.c
149
cifs_UnixTimeToNT(struct timespec64 t)
fs/smb/client/netmisc.c
152
return (u64) t.tv_sec * 10000000 + t.tv_nsec/100 + NTFS_TIME_OFFSET;
fs/smb/client/transport.c
431
long int t;
fs/smb/client/transport.c
435
t = MAX_JIFFY_OFFSET;
fs/smb/client/transport.c
437
t = msecs_to_jiffies(timeout);
fs/smb/client/transport.c
485
has_credits(server, credits, num_credits), t);
fs/smb/client/transport.c
527
t);
fs/smb/server/connection.c
154
struct ksmbd_conn *t;
fs/smb/server/connection.c
159
hash_for_each(conn_list, bkt, t, hlist) {
fs/smb/server/connection.c
160
if (memcmp(t->ClientGUID, c->ClientGUID, SMB2_CLIENT_GUID_SIZE))
fs/smb/server/connection.c
32
struct timespec64 now, t;
fs/smb/server/connection.c
370
struct ksmbd_transport *t = conn->transport;
fs/smb/server/connection.c
378
if (t->ops->prepare && t->ops->prepare(t))
fs/smb/server/connection.c
398
size = t->ops->read(t, hdr_buf, sizeof(hdr_buf), -1);
fs/smb/server/connection.c
41
jiffies_to_timespec64(jiffies - conn->last_active, &t);
fs/smb/server/connection.c
43
t = timespec64_sub(now, t);
fs/smb/server/connection.c
440
size = t->ops->read(t, conn->request_buf + 4, pdu_size, 2);
fs/smb/server/connection.c
483
t->ops->disconnect(t);
fs/smb/server/connection.c
539
struct ksmbd_transport *t;
fs/smb/server/connection.c
545
t = conn->transport;
fs/smb/server/connection.c
547
if (t->ops->shutdown) {
fs/smb/server/connection.c
549
t->ops->shutdown(t);
fs/smb/server/connection.c
55
&t);
fs/smb/server/connection.h
130
int (*prepare)(struct ksmbd_transport *t);
fs/smb/server/connection.h
131
void (*disconnect)(struct ksmbd_transport *t);
fs/smb/server/connection.h
132
void (*shutdown)(struct ksmbd_transport *t);
fs/smb/server/connection.h
133
int (*read)(struct ksmbd_transport *t, char *buf,
fs/smb/server/connection.h
135
int (*writev)(struct ksmbd_transport *t, struct kvec *iovs, int niov,
fs/smb/server/connection.h
138
int (*rdma_read)(struct ksmbd_transport *t,
fs/smb/server/connection.h
142
int (*rdma_write)(struct ksmbd_transport *t,
fs/smb/server/misc.c
353
s64 t = le64_to_cpu(ntutc) - NTFS_TIME_OFFSET;
fs/smb/server/misc.c
361
if (t < 0) {
fs/smb/server/misc.c
362
abs_t = -t;
fs/smb/server/misc.c
367
abs_t = t;
fs/smb/server/misc.c
376
inline u64 ksmbd_UnixTimeToNT(struct timespec64 t)
fs/smb/server/misc.c
379
return (u64)t.tv_sec * 10000000 + t.tv_nsec / 100 + NTFS_TIME_OFFSET;
fs/smb/server/misc.h
38
u64 ksmbd_UnixTimeToNT(struct timespec64 t);
fs/smb/server/transport_rdma.c
117
#define KSMBD_TRANS(t) (&(t)->transport)
fs/smb/server/transport_rdma.c
118
#define SMBD_TRANS(t) (container_of(t, \
fs/smb/server/transport_rdma.c
131
struct smb_direct_transport *t;
fs/smb/server/transport_rdma.c
138
t = SMBD_TRANS(kt);
fs/smb/server/transport_rdma.c
139
sc = &t->socket;
fs/smb/server/transport_rdma.c
1611
static int smb_direct_writev(struct ksmbd_transport *t,
fs/smb/server/transport_rdma.c
1615
struct smb_direct_transport *st = SMBD_TRANS(t);
fs/smb/server/transport_rdma.c
1759
static void smb_direct_free_rdma_rw_msg(struct smb_direct_transport *t,
fs/smb/server/transport_rdma.c
1763
struct smbdirect_socket *sc = &t->socket;
fs/smb/server/transport_rdma.c
1799
static int smb_direct_rdma_xmit(struct smb_direct_transport *t,
fs/smb/server/transport_rdma.c
1805
struct smbdirect_socket *sc = &t->socket;
fs/smb/server/transport_rdma.c
1917
smb_direct_free_rdma_rw_msg(t, msg,
fs/smb/server/transport_rdma.c
1931
static int smb_direct_rdma_write(struct ksmbd_transport *t,
fs/smb/server/transport_rdma.c
1936
return smb_direct_rdma_xmit(SMBD_TRANS(t), buf, buflen,
fs/smb/server/transport_rdma.c
1940
static int smb_direct_rdma_read(struct ksmbd_transport *t,
fs/smb/server/transport_rdma.c
1945
return smb_direct_rdma_xmit(SMBD_TRANS(t), buf, buflen,
fs/smb/server/transport_rdma.c
1949
static void smb_direct_disconnect(struct ksmbd_transport *t)
fs/smb/server/transport_rdma.c
1951
struct smb_direct_transport *st = SMBD_TRANS(t);
fs/smb/server/transport_rdma.c
1959
static void smb_direct_shutdown(struct ksmbd_transport *t)
fs/smb/server/transport_rdma.c
1961
struct smb_direct_transport *st = SMBD_TRANS(t);
fs/smb/server/transport_rdma.c
2503
static int smb_direct_prepare(struct ksmbd_transport *t)
fs/smb/server/transport_rdma.c
2505
struct smb_direct_transport *st = SMBD_TRANS(t);
fs/smb/server/transport_rdma.c
2660
struct smb_direct_transport *t;
fs/smb/server/transport_rdma.c
2675
t = alloc_transport(new_cm_id);
fs/smb/server/transport_rdma.c
2676
if (!t)
fs/smb/server/transport_rdma.c
2678
sc = &t->socket;
fs/smb/server/transport_rdma.c
2747
KSMBD_TRANS(t)->conn, "ksmbd:r%u",
fs/smb/server/transport_rdma.c
2757
free_transport(t);
fs/smb/server/transport_rdma.c
409
struct smb_direct_transport *t;
fs/smb/server/transport_rdma.c
414
t = kzalloc_obj(*t, KSMBD_DEFAULT_GFP);
fs/smb/server/transport_rdma.c
415
if (!t)
fs/smb/server/transport_rdma.c
417
sc = &t->socket;
fs/smb/server/transport_rdma.c
452
conn->transport = KSMBD_TRANS(t);
fs/smb/server/transport_rdma.c
453
KSMBD_TRANS(t)->conn = conn;
fs/smb/server/transport_rdma.c
454
KSMBD_TRANS(t)->ops = &ksmbd_smb_direct_transport_ops;
fs/smb/server/transport_rdma.c
455
return t;
fs/smb/server/transport_rdma.c
457
kfree(t);
fs/smb/server/transport_rdma.c
466
static void free_transport(struct smb_direct_transport *t)
fs/smb/server/transport_rdma.c
468
struct smbdirect_socket *sc = &t->socket;
fs/smb/server/transport_rdma.c
528
ksmbd_conn_free(KSMBD_TRANS(t)->conn);
fs/smb/server/transport_rdma.c
899
static int smb_direct_read(struct ksmbd_transport *t, char *buf,
fs/smb/server/transport_rdma.c
907
struct smb_direct_transport *st = SMBD_TRANS(t);
fs/smb/server/transport_tcp.c
101
sock_release(t->sock);
fs/smb/server/transport_tcp.c
102
kfree(t->iov);
fs/smb/server/transport_tcp.c
103
kfree(t);
fs/smb/server/transport_tcp.c
106
static void free_transport(struct tcp_transport *t)
fs/smb/server/transport_tcp.c
108
kernel_sock_shutdown(t->sock, SHUT_RDWR);
fs/smb/server/transport_tcp.c
109
ksmbd_conn_free(KSMBD_TRANS(t)->conn);
fs/smb/server/transport_tcp.c
151
static struct kvec *get_conn_iovec(struct tcp_transport *t, unsigned int nr_segs)
fs/smb/server/transport_tcp.c
155
if (t->iov && nr_segs <= t->nr_iov)
fs/smb/server/transport_tcp.c
156
return t->iov;
fs/smb/server/transport_tcp.c
161
kfree(t->iov);
fs/smb/server/transport_tcp.c
162
t->iov = new_iov;
fs/smb/server/transport_tcp.c
163
t->nr_iov = nr_segs;
fs/smb/server/transport_tcp.c
180
struct tcp_transport *t;
fs/smb/server/transport_tcp.c
183
t = alloc_transport(client_sk);
fs/smb/server/transport_tcp.c
184
if (!t) {
fs/smb/server/transport_tcp.c
192
KSMBD_TRANS(t)->conn, "ksmbd:%pI6c",
fs/smb/server/transport_tcp.c
193
&KSMBD_TRANS(t)->conn->inet6_addr);
fs/smb/server/transport_tcp.c
196
KSMBD_TRANS(t)->conn, "ksmbd:%pI4",
fs/smb/server/transport_tcp.c
197
&KSMBD_TRANS(t)->conn->inet_addr);
fs/smb/server/transport_tcp.c
200
KSMBD_TRANS(t)->conn, "ksmbd:%pI4",
fs/smb/server/transport_tcp.c
201
&KSMBD_TRANS(t)->conn->inet_addr);
fs/smb/server/transport_tcp.c
206
ksmbd_tcp_disconnect(KSMBD_TRANS(t));
fs/smb/server/transport_tcp.c
338
static int ksmbd_tcp_readv(struct tcp_transport *t, struct kvec *iov_orig,
fs/smb/server/transport_tcp.c
347
struct ksmbd_conn *conn = KSMBD_TRANS(t)->conn;
fs/smb/server/transport_tcp.c
349
iov = get_conn_iovec(t, nr_segs);
fs/smb/server/transport_tcp.c
365
length = kernel_recvmsg(t->sock, &ksmbd_msg,
fs/smb/server/transport_tcp.c
407
static int ksmbd_tcp_read(struct ksmbd_transport *t, char *buf,
fs/smb/server/transport_tcp.c
415
return ksmbd_tcp_readv(TCP_TRANS(t), &iov, 1, to_read, max_retries);
fs/smb/server/transport_tcp.c
418
static int ksmbd_tcp_writev(struct ksmbd_transport *t, struct kvec *iov,
fs/smb/server/transport_tcp.c
425
return kernel_sendmsg(TCP_TRANS(t)->sock, &smb_msg, iov, nvecs, size);
fs/smb/server/transport_tcp.c
428
static void ksmbd_tcp_disconnect(struct ksmbd_transport *t)
fs/smb/server/transport_tcp.c
43
static void ksmbd_tcp_disconnect(struct ksmbd_transport *t);
fs/smb/server/transport_tcp.c
430
free_transport(TCP_TRANS(t));
fs/smb/server/transport_tcp.c
45
#define KSMBD_TRANS(t) (&(t)->transport)
fs/smb/server/transport_tcp.c
46
#define TCP_TRANS(t) ((struct tcp_transport *)container_of(t, \
fs/smb/server/transport_tcp.c
61
struct tcp_transport *t;
fs/smb/server/transport_tcp.c
64
t = kzalloc_obj(*t, KSMBD_DEFAULT_GFP);
fs/smb/server/transport_tcp.c
65
if (!t)
fs/smb/server/transport_tcp.c
67
t->sock = client_sk;
fs/smb/server/transport_tcp.c
71
kfree(t);
fs/smb/server/transport_tcp.c
91
conn->transport = KSMBD_TRANS(t);
fs/smb/server/transport_tcp.c
92
KSMBD_TRANS(t)->conn = conn;
fs/smb/server/transport_tcp.c
93
KSMBD_TRANS(t)->ops = &ksmbd_tcp_transport_ops;
fs/smb/server/transport_tcp.c
94
return t;
fs/smb/server/transport_tcp.c
99
struct tcp_transport *t = TCP_TRANS(kt);
fs/timerfd.c
180
remaining = alarm_expires_remaining(&ctx->t.alarm);
fs/timerfd.c
182
remaining = hrtimer_expires_remaining_adjusted(&ctx->t.tmr);
fs/timerfd.c
203
alarm_init(&ctx->t.alarm,
fs/timerfd.c
208
hrtimer_setup(&ctx->t.tmr, timerfd_tmrproc, clockid, htmode);
fs/timerfd.c
209
hrtimer_set_expires(&ctx->t.tmr, texp);
fs/timerfd.c
217
alarm_start(&ctx->t.alarm, texp);
fs/timerfd.c
219
alarm_start_relative(&ctx->t.alarm, texp);
fs/timerfd.c
221
hrtimer_start(&ctx->t.tmr, texp, htmode);
fs/timerfd.c
239
alarm_cancel(&ctx->t.alarm);
fs/timerfd.c
241
hrtimer_cancel(&ctx->t.tmr);
fs/timerfd.c
301
&ctx->t.alarm, ctx->tintv) - 1;
fs/timerfd.c
302
alarm_restart(&ctx->t.alarm);
fs/timerfd.c
304
ticks += hrtimer_forward_now(&ctx->t.tmr,
fs/timerfd.c
306
hrtimer_restart(&ctx->t.tmr);
fs/timerfd.c
35
} t;
fs/timerfd.c
425
alarm_init(&ctx->t.alarm,
fs/timerfd.c
430
hrtimer_setup(&ctx->t.tmr, timerfd_tmrproc, clockid, HRTIMER_MODE_ABS);
fs/timerfd.c
476
if (alarm_try_to_cancel(&ctx->t.alarm) >= 0)
fs/timerfd.c
479
if (hrtimer_try_to_cancel(&ctx->t.tmr) >= 0)
fs/timerfd.c
485
hrtimer_cancel_wait_running(&ctx->t.alarm.timer);
fs/timerfd.c
487
hrtimer_cancel_wait_running(&ctx->t.tmr);
fs/timerfd.c
498
alarm_forward_now(&ctx->t.alarm, ctx->tintv);
fs/timerfd.c
500
hrtimer_forward_now(&ctx->t.tmr, ctx->tintv);
fs/timerfd.c
515
static int do_timerfd_gettime(int ufd, struct itimerspec64 *t)
fs/timerfd.c
533
&ctx->t.alarm, ctx->tintv) - 1;
fs/timerfd.c
534
alarm_restart(&ctx->t.alarm);
fs/timerfd.c
537
hrtimer_forward_now(&ctx->t.tmr, ctx->tintv)
fs/timerfd.c
539
hrtimer_restart(&ctx->t.tmr);
fs/timerfd.c
542
t->it_value = ktime_to_timespec64(timerfd_get_remaining(ctx));
fs/timerfd.c
543
t->it_interval = ktime_to_timespec64(ctx->tintv);
fs/timerfd.c
77
t.tmr);
fs/timerfd.c
85
t.alarm);
fs/ubifs/key.h
435
union ubifs_key *t = to;
fs/ubifs/key.h
437
t->j32[0] = cpu_to_le32(from->u32[0]);
fs/ubifs/key.h
438
t->j32[1] = cpu_to_le32(from->u32[1]);
fs/ubifs/key.h
451
union ubifs_key *t = to;
fs/ubifs/key.h
453
t->j32[0] = cpu_to_le32(from->u32[0]);
fs/ubifs/key.h
454
t->j32[1] = cpu_to_le32(from->u32[1]);
fs/ubifs/log.c
100
return t - h;
fs/ubifs/log.c
92
long long h, t;
fs/ubifs/log.c
95
t = (long long)c->ltail_lnum * c->leb_size;
fs/ubifs/log.c
97
if (h > t)
fs/ubifs/log.c
98
return c->log_bytes - h + t;
fs/ubifs/log.c
99
else if (h != t)
fs/udf/misc.c
276
u8 udf_tag_checksum(const struct tag *t)
fs/udf/misc.c
278
u8 *data = (u8 *)t;
fs/udf/udfdecl.h
66
u8 udf_tag_checksum(const struct tag *t);
fs/unicode/mkutf8data.c
3177
char *t;
fs/unicode/mkutf8data.c
3183
t = buf3;
fs/unicode/mkutf8data.c
3187
if (c != (unsigned char)*t++)
fs/unicode/mkutf8data.c
3191
if (*t != 0)
fs/unicode/mkutf8data.c
3198
t = buf3;
fs/unicode/mkutf8data.c
3202
if (c != (unsigned char)*t++)
fs/unicode/mkutf8data.c
3206
if (*t != 0)
fs/unicode/mkutf8data.c
3218
char *t;
fs/unicode/mkutf8data.c
3237
t = buf2;
fs/unicode/mkutf8data.c
3240
t += utf8encode(t, unichar);
fs/unicode/mkutf8data.c
3242
*t = '\0';
fs/unicode/mkutf8data.c
3246
t = buf3;
fs/unicode/mkutf8data.c
3253
t += utf8encode(t, unichar);
fs/unicode/mkutf8data.c
3255
*t = '\0';
fs/unicode/mkutf8data.c
3280
int t;
fs/unicode/mkutf8data.c
3302
t = 0;
fs/unicode/mkutf8data.c
3305
ages[gen], trees[t].index,
fs/unicode/mkutf8data.c
3307
if (trees[t].maxage == ages[gen])
fs/unicode/mkutf8data.c
3308
t += 2;
fs/unicode/mkutf8data.c
3313
t = 1;
fs/unicode/mkutf8data.c
3316
ages[gen], trees[t].index,
fs/unicode/mkutf8data.c
3318
if (trees[t].maxage == ages[gen])
fs/unicode/mkutf8data.c
3319
t += 2;
fs/unicode/mkutf8data.c
3325
t = 0;
fs/unicode/mkutf8data.c
3327
if (i == trees[t].index) {
fs/unicode/mkutf8data.c
3329
trees[t].type, trees[t].maxage);
fs/unicode/mkutf8data.c
3330
if (t < trees_count-1)
fs/unicode/mkutf8data.c
3331
t++;
fs/utimes.c
231
struct old_utimbuf32 __user *, t)
fs/utimes.c
235
if (t) {
fs/utimes.c
236
if (get_user(tv[0].tv_sec, &t->actime) ||
fs/utimes.c
237
get_user(tv[1].tv_sec, &t->modtime))
fs/utimes.c
242
return do_utimes(AT_FDCWD, filename, t ? tv : NULL, 0);
fs/utimes.c
246
SYSCALL_DEFINE4(utimensat_time32, unsigned int, dfd, const char __user *, filename, struct old_timespec32 __user *, t, int, flags)
fs/utimes.c
250
if (t) {
fs/utimes.c
251
if (get_old_timespec32(&tv[0], &t[0]) ||
fs/utimes.c
252
get_old_timespec32(&tv[1], &t[1]))
fs/utimes.c
258
return do_utimes(dfd, filename, t ? tv : NULL, flags);
fs/utimes.c
263
struct old_timeval32 __user *t)
fs/utimes.c
267
if (t) {
fs/utimes.c
268
if (get_user(tv[0].tv_sec, &t[0].tv_sec) ||
fs/utimes.c
269
get_user(tv[0].tv_nsec, &t[0].tv_usec) ||
fs/utimes.c
270
get_user(tv[1].tv_sec, &t[1].tv_sec) ||
fs/utimes.c
271
get_user(tv[1].tv_nsec, &t[1].tv_usec))
fs/utimes.c
279
return do_utimes(dfd, filename, t ? tv : NULL, 0);
fs/utimes.c
284
struct old_timeval32 __user *, t)
fs/utimes.c
286
return do_compat_futimesat(dfd, filename, t);
fs/utimes.c
289
SYSCALL_DEFINE2(utimes_time32, const char __user *, filename, struct old_timeval32 __user *, t)
fs/utimes.c
291
return do_compat_futimesat(AT_FDCWD, filename, t);
fs/xfs/libxfs/xfs_dquot_buf.c
308
uint32_t t = be32_to_cpu(dtimer);
fs/xfs/libxfs/xfs_dquot_buf.c
310
if (t != 0 && (ddq->d_type & XFS_DQTYPE_BIGTIME))
fs/xfs/libxfs/xfs_dquot_buf.c
311
return xfs_dq_bigtime_to_unix(t);
fs/xfs/libxfs/xfs_dquot_buf.c
313
return t;
fs/xfs/libxfs/xfs_dquot_buf.c
322
uint32_t t = timer;
fs/xfs/libxfs/xfs_dquot_buf.c
325
t = xfs_dq_unix_to_bigtime(timer);
fs/xfs/libxfs/xfs_dquot_buf.c
327
return cpu_to_be32(t);
fs/xfs/libxfs/xfs_rmap.c
955
# define xfs_rmap_update_hook(t, p, o, s, b, u, oi) do { } while (0)
fs/xfs/libxfs/xfs_rtbitmap.h
431
# define xfs_rtfree_extent(t,b,l) (-ENOSYS)
fs/xfs/libxfs/xfs_rtbitmap.h
439
# define xfs_rtalloc_query_range(m,t,l,h,f,p) (-ENOSYS)
fs/xfs/libxfs/xfs_rtbitmap.h
440
# define xfs_rtalloc_query_all(m,t,f,p) (-ENOSYS)
fs/xfs/libxfs/xfs_rtbitmap.h
444
# define xfs_rtalloc_extent_is_free(m,t,s,l,i) (-ENOSYS)
fs/xfs/scrub/quota_repair.c
276
uint32_t t;
fs/xfs/scrub/quota_repair.c
284
t = xfs_dq_unix_to_bigtime(new_timer);
fs/xfs/scrub/quota_repair.c
286
t = new_timer;
fs/xfs/scrub/quota_repair.c
288
*timer = cpu_to_be32(t);
fs/zonefs/super.c
445
enum zonefs_ztype t;
fs/zonefs/super.c
460
for (t = 0; t < ZONEFS_ZTYPE_MAX; t++) {
fs/zonefs/super.c
461
if (sbi->s_zgroup[t].g_nr_zones)
fs/zonefs/super.c
462
buf->f_files += sbi->s_zgroup[t].g_nr_zones + 1;
include/acpi/actypes.h
501
#define ACPI_CAST_PTR(t, p) ((t *) (acpi_uintptr_t) (p))
include/acpi/actypes.h
502
#define ACPI_CAST_INDIRECT_PTR(t, p) ((t **) (acpi_uintptr_t) (p))
include/acpi/actypes.h
503
#define ACPI_ADD_PTR(t, a, b) ACPI_CAST_PTR (t, (ACPI_CAST_PTR (u8, (a)) + (acpi_size)(b)))
include/acpi/actypes.h
504
#define ACPI_SUB_PTR(t, a, b) ACPI_CAST_PTR (t, (ACPI_CAST_PTR (u8, (a)) - (acpi_size)(b)))
include/acpi/platform/aclinux.h
78
#define ACPI_GLOBAL(t,a)
include/acpi/platform/aclinux.h
79
#define ACPI_INIT_GLOBAL(t,a,b)
include/asm-generic/ioctl.h
12
#define _IOC_TYPECHECK(t) \
include/asm-generic/ioctl.h
13
((sizeof(t) == sizeof(t[1]) && \
include/asm-generic/ioctl.h
14
sizeof(t) < (1 << _IOC_SIZEBITS)) ? \
include/asm-generic/ioctl.h
15
sizeof(t) : __invalid_size_argument_for_IOC)
include/asm-generic/ioctl.h
8
#define _IOC_TYPECHECK(t) (sizeof(t))
include/asm-generic/rwonce.h
35
#define compiletime_assert_rwonce_type(t) \
include/asm-generic/rwonce.h
36
compiletime_assert(__native_word(t) || sizeof(t) == sizeof(long long), \
include/asm-generic/vmlinux.lds.h
972
#define RUNTIME_CONST(t,x) NAMED_SECTION(runtime_##t##_##x)
include/crypto/blake2b.h
33
u64 t[2];
include/crypto/blake2b.h
62
ctx->t[0] = 0;
include/crypto/blake2b.h
63
ctx->t[1] = 0;
include/crypto/blake2s.h
37
u32 t[2];
include/crypto/blake2s.h
66
ctx->t[0] = 0;
include/crypto/blake2s.h
67
ctx->t[1] = 0;
include/crypto/gf128mul.h
221
be128 t[256];
include/crypto/gf128mul.h
225
void gf128mul_4k_lle(be128 *a, const struct gf128mul_4k *t);
include/crypto/gf128mul.h
227
static inline void gf128mul_free_4k(struct gf128mul_4k *t)
include/crypto/gf128mul.h
229
kfree_sensitive(t);
include/crypto/gf128mul.h
236
struct gf128mul_4k *t[16];
include/crypto/gf128mul.h
245
void gf128mul_free_64k(struct gf128mul_64k *t);
include/crypto/gf128mul.h
246
void gf128mul_64k_bbe(be128 *a, const struct gf128mul_64k *t);
include/drm/drm_modes.h
134
#define DRM_MODE(nm, t, c, hd, hss, hse, ht, hsk, vd, vss, vse, vt, vs, f) \
include/drm/drm_modes.h
135
.name = nm, .status = 0, .type = (t), .clock = (c), \
include/dt-bindings/usb/pd.h
14
#define PDO_TYPE(t) ((t) << PDO_TYPE_SHIFT)
include/dt-bindings/usb/pd.h
67
#define PDO_APDO_TYPE(t) ((t) << PDO_APDO_TYPE_SHIFT)
include/kvm/arm_arch_timer.h
122
#define vcpu_get_timer(v,t) (&vcpu_timer(v)->timers[(t)])
include/linux/acpi.h
1418
#define acpi_probe_device_table(t) \
include/linux/acpi.h
1420
extern struct acpi_probe_entry ACPI_PROBE_TABLE(t), \
include/linux/acpi.h
1421
ACPI_PROBE_TABLE_END(t); \
include/linux/acpi.h
1422
__acpi_probe_device_table(&ACPI_PROBE_TABLE(t), \
include/linux/acpi.h
1423
(&ACPI_PROBE_TABLE_END(t) - \
include/linux/acpi.h
1424
&ACPI_PROBE_TABLE(t))); \
include/linux/acpi.h
1482
#define acpi_probe_device_table(t) ({ int __r = 0; __r;})
include/linux/assoc_array_priv.h
155
struct assoc_array_ptr *__assoc_array_x_to_ptr(const void *p, unsigned long t)
include/linux/assoc_array_priv.h
157
return (struct assoc_array_ptr *)((unsigned long)p | t);
include/linux/audit.h
214
extern int audit_signal_info(int sig, struct task_struct *t);
include/linux/audit.h
288
static inline int audit_signal_info(int sig, struct task_struct *t)
include/linux/audit.h
330
extern void __audit_ptrace(struct task_struct *t);
include/linux/audit.h
414
static inline void audit_ptrace(struct task_struct *t)
include/linux/audit.h
417
__audit_ptrace(t);
include/linux/audit.h
716
static inline void audit_ptrace(struct task_struct *t)
include/linux/bch.h
41
unsigned int t;
include/linux/bch.h
58
struct bch_control *bch_init(int m, int t, unsigned int prim_poly,
include/linux/bits.h
46
#define GENMASK_TYPE(t, h, l) \
include/linux/bits.h
47
((t)(GENMASK_INPUT_CHECK(h, l) + \
include/linux/bits.h
48
(type_max(t) << (l) & \
include/linux/bits.h
49
type_max(t) >> (BITS_PER_TYPE(t) - 1 - (h)))))
include/linux/blk-integrity.h
25
bool queue_limits_stack_integrity(struct queue_limits *t,
include/linux/blk-integrity.h
27
static inline bool queue_limits_stack_integrity_bdev(struct queue_limits *t,
include/linux/blk-integrity.h
30
return queue_limits_stack_integrity(t, &bdev->bd_disk->queue->limits);
include/linux/blkdev.h
1139
extern int blk_stack_limits(struct queue_limits *t, struct queue_limits *b,
include/linux/blkdev.h
1141
void queue_limits_stack_bdev(struct queue_limits *t, struct block_device *bdev,
include/linux/bpf.h
1990
int (*check_member)(const struct btf_type *t,
include/linux/bpf.h
1993
int (*init_member)(const struct btf_type *t,
include/linux/bpf.h
3026
struct btf *btf, const struct btf_type *t);
include/linux/bpf_mprog.h
115
#define bpf_mprog_foreach_tuple(entry, fp, cp, t) \
include/linux/bpf_mprog.h
118
t.prog = READ_ONCE(fp->prog); \
include/linux/bpf_mprog.h
119
t.link = cp->link; \
include/linux/bpf_mprog.h
120
t.prog; \
include/linux/btf.h
225
bool btf_type_is_i32(const struct btf_type *t);
include/linux/btf.h
226
bool btf_type_is_i64(const struct btf_type *t);
include/linux/btf.h
227
bool btf_type_is_primitive(const struct btf_type *t);
include/linux/btf.h
231
struct btf_record *btf_parse_fields(const struct btf *btf, const struct btf_type *t,
include/linux/btf.h
234
bool btf_type_is_void(const struct btf_type *t);
include/linux/btf.h
246
const char *btf_type_str(const struct btf_type *t);
include/linux/btf.h
258
static inline bool btf_type_is_ptr(const struct btf_type *t)
include/linux/btf.h
260
return BTF_INFO_KIND(t->info) == BTF_KIND_PTR;
include/linux/btf.h
263
static inline bool btf_type_is_int(const struct btf_type *t)
include/linux/btf.h
265
return BTF_INFO_KIND(t->info) == BTF_KIND_INT;
include/linux/btf.h
268
static inline bool btf_type_is_small_int(const struct btf_type *t)
include/linux/btf.h
270
return btf_type_is_int(t) && t->size <= sizeof(u64);
include/linux/btf.h
273
static inline u8 btf_int_encoding(const struct btf_type *t)
include/linux/btf.h
275
return BTF_INT_ENCODING(*(u32 *)(t + 1));
include/linux/btf.h
278
static inline bool btf_type_is_signed_int(const struct btf_type *t)
include/linux/btf.h
280
return btf_type_is_int(t) && (btf_int_encoding(t) & BTF_INT_SIGNED);
include/linux/btf.h
283
static inline bool btf_type_is_enum(const struct btf_type *t)
include/linux/btf.h
285
return BTF_INFO_KIND(t->info) == BTF_KIND_ENUM;
include/linux/btf.h
288
static inline bool btf_is_any_enum(const struct btf_type *t)
include/linux/btf.h
290
return BTF_INFO_KIND(t->info) == BTF_KIND_ENUM ||
include/linux/btf.h
291
BTF_INFO_KIND(t->info) == BTF_KIND_ENUM64;
include/linux/btf.h
306
static inline u16 btf_kind(const struct btf_type *t)
include/linux/btf.h
308
return BTF_INFO_KIND(t->info);
include/linux/btf.h
311
static inline bool btf_is_enum(const struct btf_type *t)
include/linux/btf.h
313
return btf_kind(t) == BTF_KIND_ENUM;
include/linux/btf.h
316
static inline bool btf_is_enum64(const struct btf_type *t)
include/linux/btf.h
318
return btf_kind(t) == BTF_KIND_ENUM64;
include/linux/btf.h
326
static inline bool btf_is_composite(const struct btf_type *t)
include/linux/btf.h
328
u16 kind = btf_kind(t);
include/linux/btf.h
333
static inline bool btf_is_array(const struct btf_type *t)
include/linux/btf.h
335
return btf_kind(t) == BTF_KIND_ARRAY;
include/linux/btf.h
338
static inline bool btf_is_int(const struct btf_type *t)
include/linux/btf.h
340
return btf_kind(t) == BTF_KIND_INT;
include/linux/btf.h
343
static inline bool btf_is_ptr(const struct btf_type *t)
include/linux/btf.h
345
return btf_kind(t) == BTF_KIND_PTR;
include/linux/btf.h
348
static inline u8 btf_int_offset(const struct btf_type *t)
include/linux/btf.h
350
return BTF_INT_OFFSET(*(u32 *)(t + 1));
include/linux/btf.h
353
static inline __u8 btf_int_bits(const struct btf_type *t)
include/linux/btf.h
355
return BTF_INT_BITS(*(__u32 *)(t + 1));
include/linux/btf.h
358
static inline bool btf_type_is_scalar(const struct btf_type *t)
include/linux/btf.h
360
return btf_type_is_int(t) || btf_type_is_enum(t);
include/linux/btf.h
363
static inline bool btf_type_is_fwd(const struct btf_type *t)
include/linux/btf.h
365
return BTF_INFO_KIND(t->info) == BTF_KIND_FWD;
include/linux/btf.h
368
static inline bool btf_type_is_typedef(const struct btf_type *t)
include/linux/btf.h
370
return BTF_INFO_KIND(t->info) == BTF_KIND_TYPEDEF;
include/linux/btf.h
373
static inline bool btf_type_is_volatile(const struct btf_type *t)
include/linux/btf.h
375
return BTF_INFO_KIND(t->info) == BTF_KIND_VOLATILE;
include/linux/btf.h
378
static inline bool btf_type_is_func(const struct btf_type *t)
include/linux/btf.h
380
return BTF_INFO_KIND(t->info) == BTF_KIND_FUNC;
include/linux/btf.h
383
static inline bool btf_type_is_func_proto(const struct btf_type *t)
include/linux/btf.h
385
return BTF_INFO_KIND(t->info) == BTF_KIND_FUNC_PROTO;
include/linux/btf.h
388
static inline bool btf_type_is_var(const struct btf_type *t)
include/linux/btf.h
390
return BTF_INFO_KIND(t->info) == BTF_KIND_VAR;
include/linux/btf.h
393
static inline bool btf_type_is_type_tag(const struct btf_type *t)
include/linux/btf.h
395
return BTF_INFO_KIND(t->info) == BTF_KIND_TYPE_TAG;
include/linux/btf.h
401
static inline bool btf_type_is_struct(const struct btf_type *t)
include/linux/btf.h
403
u8 kind = BTF_INFO_KIND(t->info);
include/linux/btf.h
408
static inline bool __btf_type_is_struct(const struct btf_type *t)
include/linux/btf.h
410
return BTF_INFO_KIND(t->info) == BTF_KIND_STRUCT;
include/linux/btf.h
413
static inline bool btf_type_is_array(const struct btf_type *t)
include/linux/btf.h
415
return BTF_INFO_KIND(t->info) == BTF_KIND_ARRAY;
include/linux/btf.h
418
static inline u16 btf_type_vlen(const struct btf_type *t)
include/linux/btf.h
420
return BTF_INFO_VLEN(t->info);
include/linux/btf.h
423
static inline u16 btf_vlen(const struct btf_type *t)
include/linux/btf.h
425
return btf_type_vlen(t);
include/linux/btf.h
428
static inline u16 btf_func_linkage(const struct btf_type *t)
include/linux/btf.h
430
return BTF_INFO_VLEN(t->info);
include/linux/btf.h
433
static inline bool btf_type_kflag(const struct btf_type *t)
include/linux/btf.h
435
return BTF_INFO_KFLAG(t->info);
include/linux/btf.h
452
static inline struct btf_member *btf_members(const struct btf_type *t)
include/linux/btf.h
454
return (struct btf_member *)(t + 1);
include/linux/btf.h
457
static inline u32 btf_member_bit_offset(const struct btf_type *t, u32 member_idx)
include/linux/btf.h
459
const struct btf_member *m = btf_members(t) + member_idx;
include/linux/btf.h
461
return __btf_member_bit_offset(t, m);
include/linux/btf.h
464
static inline u32 btf_member_bitfield_size(const struct btf_type *t, u32 member_idx)
include/linux/btf.h
466
const struct btf_member *m = btf_members(t) + member_idx;
include/linux/btf.h
468
return __btf_member_bitfield_size(t, m);
include/linux/btf.h
471
static inline const struct btf_member *btf_type_member(const struct btf_type *t)
include/linux/btf.h
473
return (const struct btf_member *)(t + 1);
include/linux/btf.h
476
static inline struct btf_array *btf_array(const struct btf_type *t)
include/linux/btf.h
478
return (struct btf_array *)(t + 1);
include/linux/btf.h
481
static inline struct btf_enum *btf_enum(const struct btf_type *t)
include/linux/btf.h
483
return (struct btf_enum *)(t + 1);
include/linux/btf.h
486
static inline struct btf_enum64 *btf_enum64(const struct btf_type *t)
include/linux/btf.h
488
return (struct btf_enum64 *)(t + 1);
include/linux/btf.h
492
const struct btf_type *t)
include/linux/btf.h
494
return (const struct btf_var_secinfo *)(t + 1);
include/linux/btf.h
497
static inline struct btf_param *btf_params(const struct btf_type *t)
include/linux/btf.h
499
return (struct btf_param *)(t + 1);
include/linux/btf.h
502
static inline struct btf_decl_tag *btf_decl_tag(const struct btf_type *t)
include/linux/btf.h
504
return (struct btf_decl_tag *)(t + 1);
include/linux/btf.h
571
int btf_field_iter_init(struct btf_field_iter *it, struct btf_type *t,
include/linux/btf.h
592
const struct btf_type *t, enum bpf_prog_type prog_type,
include/linux/btf.h
599
static inline bool btf_type_is_struct_ptr(struct btf *btf, const struct btf_type *t)
include/linux/btf.h
601
if (!btf_type_is_ptr(t))
include/linux/btf.h
604
t = btf_type_skip_modifiers(btf, t->type, NULL);
include/linux/btf.h
606
return btf_type_is_struct(t);
include/linux/btf.h
625
static inline int btf_field_iter_init(struct btf_field_iter *it, struct btf_type *t,
include/linux/btf.h
668
const struct btf_type *t, enum bpf_prog_type prog_type,
include/linux/can/dev/peak_canfd.h
62
#define PUCAN_TSLOW_SJW_T(s, t) (((s) & PUCAN_TSLOW_SJW_MASK) | \
include/linux/can/dev/peak_canfd.h
63
((!!(t)) << 7))
include/linux/can/dev/peak_canfd.h
64
#define PUCAN_TSLOW_TSEG2(t) ((t) & PUCAN_TSLOW_TSEG2_MASK)
include/linux/can/dev/peak_canfd.h
65
#define PUCAN_TSLOW_TSEG1(t) ((t) & PUCAN_TSLOW_TSEG1_MASK)
include/linux/can/dev/peak_canfd.h
91
#define PUCAN_TFAST_TSEG2(t) ((t) & PUCAN_TFAST_TSEG2_MASK)
include/linux/can/dev/peak_canfd.h
92
#define PUCAN_TFAST_TSEG1(t) ((t) & PUCAN_TFAST_TSEG1_MASK)
include/linux/capability.h
142
extern bool has_ns_capability(struct task_struct *t,
include/linux/capability.h
144
extern bool has_capability_noaudit(struct task_struct *t, int cap);
include/linux/capability.h
145
extern bool has_ns_capability_noaudit(struct task_struct *t,
include/linux/capability.h
152
static inline bool has_ns_capability(struct task_struct *t,
include/linux/capability.h
157
static inline bool has_capability_noaudit(struct task_struct *t, int cap)
include/linux/capability.h
161
static inline bool has_ns_capability_noaudit(struct task_struct *t,
include/linux/cb710.h
51
#define CB710_PORT_ACCESSORS(t) \
include/linux/cb710.h
52
static inline void cb710_write_port_##t(struct cb710_slot *slot, \
include/linux/cb710.h
53
unsigned port, u##t value) \
include/linux/cb710.h
55
iowrite##t(value, slot->iobase + port); \
include/linux/cb710.h
58
static inline u##t cb710_read_port_##t(struct cb710_slot *slot, \
include/linux/cb710.h
61
return ioread##t(slot->iobase + port); \
include/linux/cb710.h
64
static inline void cb710_modify_port_##t(struct cb710_slot *slot, \
include/linux/cb710.h
65
unsigned port, u##t set, u##t clear) \
include/linux/cb710.h
67
iowrite##t( \
include/linux/cb710.h
68
(ioread##t(slot->iobase + port) & ~clear)|set, \
include/linux/ceph/libceph.h
191
static bool __insert_##name(struct rb_root *root, type *t) \
include/linux/ceph/libceph.h
196
BUG_ON(!RB_EMPTY_NODE(&t->nodefld)); \
include/linux/ceph/libceph.h
203
cmp = cmpexp(keyexp(t->keyfld), keyexp(cur->keyfld)); \
include/linux/ceph/libceph.h
212
rb_link_node(&t->nodefld, parent, n); \
include/linux/ceph/libceph.h
213
rb_insert_color(&t->nodefld, root); \
include/linux/ceph/libceph.h
216
static void __maybe_unused insert_##name(struct rb_root *root, type *t) \
include/linux/ceph/libceph.h
218
if (!__insert_##name(root, t)) \
include/linux/ceph/libceph.h
221
static void erase_##name(struct rb_root *root, type *t) \
include/linux/ceph/libceph.h
223
BUG_ON(RB_EMPTY_NODE(&t->nodefld)); \
include/linux/ceph/libceph.h
224
rb_erase(&t->nodefld, root); \
include/linux/ceph/libceph.h
225
RB_CLEAR_NODE(&t->nodefld); \
include/linux/ceph/osd_client.h
323
struct ceph_osd_request_target t;
include/linux/cgroup.h
673
struct task_struct *t) { return 0; }
include/linux/cleanup.h
287
{ _type t = _init; return t; }
include/linux/cleanup.h
296
{ class_##_name##_t t = _init; return t; }
include/linux/compat.h
42
#define __SC_DELOUSE(t,v) ((__force t)(unsigned long)(v))
include/linux/compat.h
548
struct task_struct *t = current; \
include/linux/compat.h
549
unsafe_put_user(ptr_to_compat((void __user *)t->sas_ss_sp), \
include/linux/compat.h
551
unsafe_put_user(t->sas_ss_flags, &__uss->ss_flags, label); \
include/linux/compat.h
552
unsafe_put_user(t->sas_ss_size, &__uss->ss_size, label); \
include/linux/compiler_types.h
666
#define __native_word(t) \
include/linux/compiler_types.h
667
(sizeof(t) == sizeof(char) || sizeof(t) == sizeof(short) || \
include/linux/compiler_types.h
668
sizeof(t) == sizeof(int) || sizeof(t) == sizeof(long))
include/linux/compiler_types.h
708
#define compiletime_assert_atomic_type(t) \
include/linux/compiler_types.h
709
compiletime_assert(__native_word(t), \
include/linux/cpu.h
232
int arch_get_indir_br_lp_status(struct task_struct *t, unsigned long __user *status);
include/linux/cpu.h
233
int arch_set_indir_br_lp_status(struct task_struct *t, unsigned long status);
include/linux/cpu.h
234
int arch_lock_indir_br_lp_status(struct task_struct *t, unsigned long status);
include/linux/damon.h
449
struct damon_target *t, struct damon_region *r,
include/linux/damon.h
650
struct damon_target *t, struct damon_region *r,
include/linux/damon.h
653
struct damon_target *t, struct damon_region *r,
include/linux/damon.h
655
bool (*target_valid)(struct damon_target *t);
include/linux/damon.h
656
void (*cleanup_target)(struct damon_target *t);
include/linux/damon.h
843
static inline struct damon_region *damon_last_region(struct damon_target *t)
include/linux/damon.h
845
return list_last_entry(&t->regions_list, struct damon_region, list);
include/linux/damon.h
848
static inline struct damon_region *damon_first_region(struct damon_target *t)
include/linux/damon.h
850
return list_first_entry(&t->regions_list, struct damon_region, list);
include/linux/damon.h
859
#define damon_for_each_region(r, t) \
include/linux/damon.h
860
list_for_each_entry(r, &t->regions_list, list)
include/linux/damon.h
862
#define damon_for_each_region_from(r, t) \
include/linux/damon.h
863
list_for_each_entry_from(r, &t->regions_list, list)
include/linux/damon.h
865
#define damon_for_each_region_safe(r, next, t) \
include/linux/damon.h
866
list_for_each_entry_safe(r, next, &t->regions_list, list)
include/linux/damon.h
868
#define damon_for_each_target(t, ctx) \
include/linux/damon.h
869
list_for_each_entry(t, &(ctx)->adaptive_targets, list)
include/linux/damon.h
871
#define damon_for_each_target_safe(t, next, ctx) \
include/linux/damon.h
872
list_for_each_entry_safe(t, next, &(ctx)->adaptive_targets, list)
include/linux/damon.h
907
struct damon_target *t)
include/linux/damon.h
910
t->nr_regions++;
include/linux/damon.h
913
void damon_add_region(struct damon_region *r, struct damon_target *t);
include/linux/damon.h
914
void damon_destroy_region(struct damon_region *r, struct damon_target *t);
include/linux/damon.h
915
int damon_set_regions(struct damon_target *t, struct damon_addr_range *ranges,
include/linux/damon.h
943
void damon_add_target(struct damon_ctx *ctx, struct damon_target *t);
include/linux/damon.h
945
void damon_free_target(struct damon_target *t);
include/linux/damon.h
946
void damon_destroy_target(struct damon_target *t, struct damon_ctx *ctx);
include/linux/damon.h
947
unsigned int damon_nr_regions(struct damon_target *t);
include/linux/damon.h
982
int damon_set_region_biggest_system_ram_default(struct damon_target *t,
include/linux/device-mapper.h
439
int dm_register_target(struct target_type *t);
include/linux/device-mapper.h
440
void dm_unregister_target(struct target_type *t);
include/linux/device-mapper.h
589
int dm_table_add_target(struct dm_table *t, const char *type,
include/linux/device-mapper.h
598
void dm_table_set_type(struct dm_table *t, enum dm_queue_mode type);
include/linux/device-mapper.h
603
int dm_table_complete(struct dm_table *t);
include/linux/device-mapper.h
608
void dm_table_destroy(struct dm_table *t);
include/linux/device-mapper.h
625
sector_t dm_table_get_size(struct dm_table *t);
include/linux/device-mapper.h
626
blk_mode_t dm_table_get_mode(struct dm_table *t);
include/linux/device-mapper.h
627
struct mapped_device *dm_table_get_md(struct dm_table *t);
include/linux/device-mapper.h
628
const char *dm_table_device_name(struct dm_table *t);
include/linux/device-mapper.h
633
void dm_table_event(struct dm_table *t);
include/linux/device-mapper.h
638
void dm_table_run_md_queue_async(struct dm_table *t);
include/linux/device-mapper.h
645
struct dm_table *t);
include/linux/dma-mapping.h
599
#define dma_get_sgtable(d, t, v, h, s) dma_get_sgtable_attrs(d, t, v, h, s, 0)
include/linux/elfcore.h
87
static inline int elf_core_copy_task_regs(struct task_struct *t, elf_gregset_t* elfregs)
include/linux/elfcore.h
90
return ELF_CORE_COPY_TASK_REGS(t, elfregs);
include/linux/elfcore.h
92
elf_core_copy_regs(elfregs, task_pt_regs(t));
include/linux/elfcore.h
97
int elf_core_copy_task_fpregs(struct task_struct *t, elf_fpregset_t *fpu);
include/linux/filter.h
579
#define __BPF_MAP_1(m, v, t, a, ...) m(t, a)
include/linux/filter.h
580
#define __BPF_MAP_2(m, v, t, a, ...) m(t, a), __BPF_MAP_1(m, v, __VA_ARGS__)
include/linux/filter.h
581
#define __BPF_MAP_3(m, v, t, a, ...) m(t, a), __BPF_MAP_2(m, v, __VA_ARGS__)
include/linux/filter.h
582
#define __BPF_MAP_4(m, v, t, a, ...) m(t, a), __BPF_MAP_3(m, v, __VA_ARGS__)
include/linux/filter.h
583
#define __BPF_MAP_5(m, v, t, a, ...) m(t, a), __BPF_MAP_4(m, v, __VA_ARGS__)
include/linux/filter.h
595
#define __BPF_CAST(t, a) \
include/linux/filter.h
596
(__force t) \
include/linux/filter.h
598
typeof(__builtin_choose_expr(sizeof(t) == sizeof(unsigned long), \
include/linux/filter.h
599
(unsigned long)0, (t)0))) a
include/linux/filter.h
603
#define __BPF_DECL_ARGS(t, a) t a
include/linux/filter.h
604
#define __BPF_DECL_REGS(t, a) u64 a
include/linux/firewire.h
392
void __fw_send_request(struct fw_card *card, struct fw_transaction *t, int tcode,
include/linux/firewire.h
419
static inline void fw_send_request(struct fw_card *card, struct fw_transaction *t, int tcode,
include/linux/firewire.h
427
__fw_send_request(card, t, tcode, destination_id, generation, speed, offset, payload,
include/linux/firewire.h
452
static inline void fw_send_request_with_tstamp(struct fw_card *card, struct fw_transaction *t,
include/linux/firewire.h
460
__fw_send_request(card, t, tcode, destination_id, generation, speed, offset, payload,
include/linux/freezer.h
47
extern void __thaw_task(struct task_struct *t);
include/linux/freezer.h
81
static inline void __thaw_task(struct task_struct *t) {}
include/linux/fs.h
975
struct timespec64 timestamp_truncate(struct timespec64 t, struct inode *inode);
include/linux/fsi.h
45
#define FSI_DEVICE(t) \
include/linux/fsi.h
46
.engine_type = (t), .version = FSI_VERSION_ANY,
include/linux/fsi.h
48
#define FSI_DEVICE_VERSIONED(t, v) \
include/linux/fsi.h
49
.engine_type = (t), .version = (v),
include/linux/ftrace.h
1335
extern void ftrace_graph_init_task(struct task_struct *t);
include/linux/ftrace.h
1336
extern void ftrace_graph_exit_task(struct task_struct *t);
include/linux/ftrace.h
1337
extern void ftrace_graph_init_idle_task(struct task_struct *t, int cpu);
include/linux/ftrace.h
1355
static inline void ftrace_graph_init_task(struct task_struct *t) { }
include/linux/ftrace.h
1356
static inline void ftrace_graph_exit_task(struct task_struct *t) { }
include/linux/ftrace.h
1357
static inline void ftrace_graph_init_idle_task(struct task_struct *t, int cpu) { }
include/linux/i2c.h
914
void i2c_parse_fw_timings(struct device *dev, struct i2c_timings *t, bool use_defaults);
include/linux/iio/sw_trigger.h
55
void iio_swt_group_init_type_name(struct iio_sw_trigger *t,
include/linux/iio/sw_trigger.h
60
config_group_init_type_name(&t->group, name, type);
include/linux/interrupt.h
699
void (*callback)(struct tasklet_struct *t);
include/linux/interrupt.h
740
static inline int tasklet_trylock(struct tasklet_struct *t)
include/linux/interrupt.h
742
return !test_and_set_bit(TASKLET_STATE_RUN, &(t)->state);
include/linux/interrupt.h
745
void tasklet_unlock(struct tasklet_struct *t);
include/linux/interrupt.h
746
void tasklet_unlock_wait(struct tasklet_struct *t);
include/linux/interrupt.h
747
void tasklet_unlock_spin_wait(struct tasklet_struct *t);
include/linux/interrupt.h
750
static inline int tasklet_trylock(struct tasklet_struct *t) { return 1; }
include/linux/interrupt.h
751
static inline void tasklet_unlock(struct tasklet_struct *t) { }
include/linux/interrupt.h
752
static inline void tasklet_unlock_wait(struct tasklet_struct *t) { }
include/linux/interrupt.h
753
static inline void tasklet_unlock_spin_wait(struct tasklet_struct *t) { }
include/linux/interrupt.h
756
extern void __tasklet_schedule(struct tasklet_struct *t);
include/linux/interrupt.h
758
static inline void tasklet_schedule(struct tasklet_struct *t)
include/linux/interrupt.h
760
if (!test_and_set_bit(TASKLET_STATE_SCHED, &t->state))
include/linux/interrupt.h
761
__tasklet_schedule(t);
include/linux/interrupt.h
764
extern void __tasklet_hi_schedule(struct tasklet_struct *t);
include/linux/interrupt.h
766
static inline void tasklet_hi_schedule(struct tasklet_struct *t)
include/linux/interrupt.h
768
if (!test_and_set_bit(TASKLET_STATE_SCHED, &t->state))
include/linux/interrupt.h
769
__tasklet_hi_schedule(t);
include/linux/interrupt.h
772
static inline void tasklet_disable_nosync(struct tasklet_struct *t)
include/linux/interrupt.h
774
atomic_inc(&t->count);
include/linux/interrupt.h
782
static inline void tasklet_disable_in_atomic(struct tasklet_struct *t)
include/linux/interrupt.h
784
tasklet_disable_nosync(t);
include/linux/interrupt.h
785
tasklet_unlock_spin_wait(t);
include/linux/interrupt.h
789
static inline void tasklet_disable(struct tasklet_struct *t)
include/linux/interrupt.h
791
tasklet_disable_nosync(t);
include/linux/interrupt.h
792
tasklet_unlock_wait(t);
include/linux/interrupt.h
796
static inline void tasklet_enable(struct tasklet_struct *t)
include/linux/interrupt.h
799
atomic_dec(&t->count);
include/linux/interrupt.h
802
extern void tasklet_kill(struct tasklet_struct *t);
include/linux/interrupt.h
803
extern void tasklet_init(struct tasklet_struct *t,
include/linux/interrupt.h
805
extern void tasklet_setup(struct tasklet_struct *t,
include/linux/iversion.h
259
static inline u64 time_to_chattr(const struct timespec64 *t)
include/linux/iversion.h
261
u64 chattr = t->tv_sec;
include/linux/iversion.h
264
chattr += t->tv_nsec;
include/linux/kcov.h
30
void kcov_task_init(struct task_struct *t);
include/linux/kcov.h
31
void kcov_task_exit(struct task_struct *t);
include/linux/kcov.h
33
#define kcov_prepare_switch(t) \
include/linux/kcov.h
35
(t)->kcov_mode |= KCOV_IN_CTXSW; \
include/linux/kcov.h
38
#define kcov_finish_switch(t) \
include/linux/kcov.h
40
(t)->kcov_mode &= ~KCOV_IN_CTXSW; \
include/linux/kcov.h
96
static inline void kcov_task_init(struct task_struct *t) {}
include/linux/kcov.h
97
static inline void kcov_task_exit(struct task_struct *t) {}
include/linux/kcov.h
98
static inline void kcov_prepare_switch(struct task_struct *t) {}
include/linux/kcov.h
99
static inline void kcov_finish_switch(struct task_struct *t) {}
include/linux/kstack_erase.h
72
static inline void stackleak_task_init(struct task_struct *t)
include/linux/kstack_erase.h
74
t->lowest_stack = stackleak_task_low_bound(t);
include/linux/kstack_erase.h
76
t->prev_lowest_stack = t->lowest_stack;
include/linux/kstack_erase.h
86
static inline void stackleak_task_init(struct task_struct *t) { }
include/linux/kthread.h
138
void kthread_delayed_work_timer_fn(struct timer_list *t);
include/linux/lockdep.h
336
# define lock_acquire(l, s, t, r, c, n, i) do { } while (0)
include/linux/lockdep.h
509
#define lock_acquire_exclusive(l, s, t, n, i) lock_acquire(l, s, t, 0, 1, n, i)
include/linux/lockdep.h
510
#define lock_acquire_shared(l, s, t, n, i) lock_acquire(l, s, t, 1, 1, n, i)
include/linux/lockdep.h
511
#define lock_acquire_shared_recursive(l, s, t, n, i) lock_acquire(l, s, t, 2, 1, n, i)
include/linux/lockdep.h
513
#define spin_acquire(l, s, t, i) lock_acquire_exclusive(l, s, t, NULL, i)
include/linux/lockdep.h
514
#define spin_acquire_nest(l, s, t, n, i) lock_acquire_exclusive(l, s, t, n, i)
include/linux/lockdep.h
517
#define rwlock_acquire(l, s, t, i) lock_acquire_exclusive(l, s, t, NULL, i)
include/linux/lockdep.h
518
#define rwlock_acquire_read(l, s, t, i) \
include/linux/lockdep.h
521
lock_acquire_shared_recursive(l, s, t, NULL, i); \
include/linux/lockdep.h
523
lock_acquire_shared(l, s, t, NULL, i); \
include/linux/lockdep.h
528
#define seqcount_acquire(l, s, t, i) lock_acquire_exclusive(l, s, t, NULL, i)
include/linux/lockdep.h
529
#define seqcount_acquire_read(l, s, t, i) lock_acquire_shared_recursive(l, s, t, NULL, i)
include/linux/lockdep.h
532
#define mutex_acquire(l, s, t, i) lock_acquire_exclusive(l, s, t, NULL, i)
include/linux/lockdep.h
533
#define mutex_acquire_nest(l, s, t, n, i) lock_acquire_exclusive(l, s, t, n, i)
include/linux/lockdep.h
536
#define rwsem_acquire(l, s, t, i) lock_acquire_exclusive(l, s, t, NULL, i)
include/linux/lockdep.h
537
#define rwsem_acquire_nest(l, s, t, n, i) lock_acquire_exclusive(l, s, t, n, i)
include/linux/lockdep.h
538
#define rwsem_acquire_read(l, s, t, i) lock_acquire_shared(l, s, t, NULL, i)
include/linux/min_heap.h
122
u64 t = *(u64 *)(a + (n -= 8));
include/linux/min_heap.h
124
*(u64 *)(b + n) = t;
include/linux/min_heap.h
127
u32 t = *(u32 *)(a + (n -= 4));
include/linux/min_heap.h
129
*(u32 *)(b + n) = t;
include/linux/min_heap.h
131
t = *(u32 *)(a + (n -= 4));
include/linux/min_heap.h
133
*(u32 *)(b + n) = t;
include/linux/min_heap.h
150
char t = ((char *)a)[--n];
include/linux/min_heap.h
152
((char *)b)[n] = t;
include/linux/min_heap.h
95
u32 t = *(u32 *)(a + (n -= 4));
include/linux/min_heap.h
97
*(u32 *)(b + n) = t;
include/linux/mm.h
4812
int arch_get_shadow_stack_status(struct task_struct *t, unsigned long __user *status);
include/linux/mm.h
4813
int arch_set_shadow_stack_status(struct task_struct *t, unsigned long status);
include/linux/mm.h
4814
int arch_lock_shadow_stack_status(struct task_struct *t, unsigned long status);
include/linux/mroute_base.h
281
void (*expire_func)(struct timer_list *t),
include/linux/netfilter/ipset/ip_set.h
480
ip_set_timeout_expired(const unsigned long *t)
include/linux/netfilter/ipset/ip_set.h
482
return *t != IPSET_ELEM_PERMANENT && time_is_before_jiffies(*t);
include/linux/netfilter/ipset/ip_set.h
488
unsigned long t;
include/linux/netfilter/ipset/ip_set.h
495
t = msecs_to_jiffies(value * MSEC_PER_SEC) + jiffies;
include/linux/netfilter/ipset/ip_set.h
496
if (t == IPSET_ELEM_PERMANENT)
include/linux/netfilter/ipset/ip_set.h
498
t--;
include/linux/netfilter/ipset/ip_set.h
499
*timeout = t;
include/linux/netfilter/nf_conntrack_proto_gre.h
23
struct nf_conntrack_tuple *t);
include/linux/netfilter/x_tables.h
289
int xt_target_to_user(const struct xt_entry_target *t,
include/linux/netfilter/x_tables.h
320
void xt_table_unlock(struct xt_table *t);
include/linux/netfilter/x_tables.h
442
int xt_register_template(const struct xt_table *t, int(*table_init)(struct net *net));
include/linux/netfilter/x_tables.h
443
void xt_unregister_template(const struct xt_table *t);
include/linux/netfilter/x_tables.h
518
void xt_compat_target_from_user(struct xt_entry_target *t, void **dstptr,
include/linux/netfilter/x_tables.h
520
int xt_compat_target_to_user(const struct xt_entry_target *t,
include/linux/netfilter_bridge/ebtables.h
125
int ebt_register_template(const struct ebt_table *t, int(*table_init)(struct net *net));
include/linux/netfilter_bridge/ebtables.h
126
void ebt_unregister_template(const struct ebt_table *t);
include/linux/nls.h
62
static inline unsigned char nls_tolower(struct nls_table *t, unsigned char c)
include/linux/nls.h
64
unsigned char nc = t->charset2lower[c];
include/linux/nls.h
69
static inline unsigned char nls_toupper(struct nls_table *t, unsigned char c)
include/linux/nls.h
71
unsigned char nc = t->charset2upper[c];
include/linux/nls.h
76
static inline int nls_strnicmp(struct nls_table *t, const unsigned char *s1,
include/linux/nls.h
80
if (nls_tolower(t, *s1++) != nls_tolower(t, *s2++))
include/linux/omap-gpmc.h
71
extern int gpmc_cs_set_timings(int cs, const struct gpmc_timings *t,
include/linux/overflow.h
35
#define type_max(t) __type_max(typeof(t))
include/linux/overflow.h
37
#define type_min(t) __type_min(typeof(t))
include/linux/page_ref.h
27
#define page_ref_tracepoint_active(t) tracepoint_enabled(t)
include/linux/page_ref.h
39
#define page_ref_tracepoint_active(t) false
include/linux/pci.h
2094
#define _PCI_NOP(o, s, t) \
include/linux/pci.h
2096
int where, t val) \
include/linux/platform_data/cros_ec_commands.h
3255
int32_t t; /* In 1/100 K */
include/linux/psi.h
29
void psi_trigger_destroy(struct psi_trigger *t);
include/linux/rcupdate.h
174
# define rcu_tasks_classic_qs(t, preempt) \
include/linux/rcupdate.h
176
if (!(preempt) && READ_ONCE((t)->rcu_tasks_holdout)) \
include/linux/rcupdate.h
177
WRITE_ONCE((t)->rcu_tasks_holdout, false); \
include/linux/rcupdate.h
183
# define rcu_tasks_classic_qs(t, preempt) do { } while (0)
include/linux/rcupdate.h
188
#define rcu_tasks_qs(t, preempt) rcu_tasks_classic_qs((t), (preempt))
include/linux/rcupdate.h
195
#define rcu_note_voluntary_context_switch(t) rcu_tasks_qs(t, false)
include/linux/rcupdate.h
199
#define rcu_tasks_classic_qs(t, preempt) do { } while (0)
include/linux/rcupdate.h
200
#define rcu_tasks_qs(t, preempt) do { } while (0)
include/linux/rcupdate.h
201
#define rcu_note_voluntary_context_switch(t) do { } while (0)
include/linux/rcupdate_trace.h
101
if (t->trc_reader_nesting++) {
include/linux/rcupdate_trace.h
106
t->trc_reader_scp = __srcu_read_lock_fast(&rcu_tasks_trace_srcu_struct);
include/linux/rcupdate_trace.h
123
struct task_struct *t = current;
include/linux/rcupdate_trace.h
125
scp = t->trc_reader_scp;
include/linux/rcupdate_trace.h
127
if (!--t->trc_reader_nesting) {
include/linux/rcupdate_trace.h
98
struct task_struct *t = current;
include/linux/rcutiny.h
111
static inline void rcu_request_urgent_qs_task(struct task_struct *t) { }
include/linux/rcutiny.h
122
static inline bool rcu_preempt_need_deferred_qs(struct task_struct *t)
include/linux/rcutiny.h
126
static inline void rcu_preempt_deferred_qs(struct task_struct *t) { }
include/linux/rcutree.h
24
void rcu_request_urgent_qs_task(struct task_struct *t);
include/linux/rcutree.h
94
void rcu_preempt_deferred_qs(struct task_struct *t);
include/linux/reciprocal_div.h
35
u32 t = (u32)(((u64)a * R.m) >> 32);
include/linux/reciprocal_div.h
36
return (t + ((a - t) >> R.sh1)) >> R.sh2;
include/linux/refcount.h
125
void refcount_warn_saturate(refcount_t *r, enum refcount_saturation_type t);
include/linux/resctrl.h
508
u32 closid, enum resctrl_conf_type t, u32 cfg_val);
include/linux/rseq.h
120
static inline void rseq_reset(struct task_struct *t)
include/linux/rseq.h
122
memset(&t->rseq, 0, sizeof(t->rseq));
include/linux/rseq.h
123
t->rseq.ids.cpu_id = RSEQ_CPU_ID_UNINITIALIZED;
include/linux/rseq.h
126
static inline void rseq_execve(struct task_struct *t)
include/linux/rseq.h
128
rseq_reset(t);
include/linux/rseq.h
141
static inline void rseq_fork(struct task_struct *t, u64 clone_flags)
include/linux/rseq.h
144
rseq_reset(t);
include/linux/rseq.h
146
t->rseq = current->rseq;
include/linux/rseq.h
164
static inline void rseq_sched_switch_event(struct task_struct *t) { }
include/linux/rseq.h
165
static inline void rseq_sched_set_ids_changed(struct task_struct *t) { }
include/linux/rseq.h
168
static inline void rseq_fork(struct task_struct *t, u64 clone_flags) { }
include/linux/rseq.h
169
static inline void rseq_execve(struct task_struct *t) { }
include/linux/rseq.h
43
static inline void rseq_raise_notify_resume(struct task_struct *t)
include/linux/rseq.h
45
set_tsk_thread_flag(t, TIF_RSEQ);
include/linux/rseq.h
49
static __always_inline void rseq_sched_switch_event(struct task_struct *t)
include/linux/rseq.h
51
struct rseq_event *ev = &t->rseq.event;
include/linux/rseq.h
65
rseq_raise_notify_resume(t);
include/linux/rseq.h
69
t->rseq.event.sched_switch = true;
include/linux/rseq.h
70
rseq_raise_notify_resume(t);
include/linux/rseq.h
82
static __always_inline void rseq_sched_set_ids_changed(struct task_struct *t)
include/linux/rseq.h
84
t->rseq.event.ids_changed = true;
include/linux/rseq_entry.h
106
static __always_inline void rseq_slice_clear_grant(struct task_struct *t)
include/linux/rseq_entry.h
108
if (IS_ENABLED(CONFIG_RSEQ_STATS) && t->rseq.slice.state.granted)
include/linux/rseq_entry.h
110
t->rseq.slice.state.granted = false;
include/linux/rseq_entry.h
221
static __always_inline void rseq_slice_clear_grant(struct task_struct *t) { }
include/linux/rseq_entry.h
225
bool rseq_debug_update_user_cs(struct task_struct *t, struct pt_regs *regs, unsigned long csaddr);
include/linux/rseq_entry.h
226
bool rseq_debug_validate_ids(struct task_struct *t);
include/linux/rseq_entry.h
267
bool rseq_debug_update_user_cs(struct task_struct *t, struct pt_regs *regs,
include/linux/rseq_entry.h
318
if (unlikely(usig != t->rseq.sig))
include/linux/rseq_entry.h
324
if (unlikely(!t->rseq.event.user_irq))
include/linux/rseq_entry.h
327
unsafe_put_user(0ULL, &t->rseq.usrptr->rseq_cs, efault);
include/linux/rseq_entry.h
332
unsafe_put_user(0ULL, &t->rseq.usrptr->rseq_cs, efault);
include/linux/rseq_entry.h
341
t->rseq.event.fatal = true;
include/linux/rseq_entry.h
350
bool rseq_debug_validate_ids(struct task_struct *t)
include/linux/rseq_entry.h
352
struct rseq __user *rseq = t->rseq.usrptr;
include/linux/rseq_entry.h
359
node_id = t->rseq.ids.cpu_id != RSEQ_CPU_ID_UNINITIALIZED ?
include/linux/rseq_entry.h
360
cpu_to_node(t->rseq.ids.cpu_id) : 0;
include/linux/rseq_entry.h
364
if (cpu_id != t->rseq.ids.cpu_id)
include/linux/rseq_entry.h
373
if (uval != t->rseq.ids.mm_cid)
include/linux/rseq_entry.h
378
t->rseq.event.fatal = true;
include/linux/rseq_entry.h
392
rseq_update_user_cs(struct task_struct *t, struct pt_regs *regs, unsigned long csaddr)
include/linux/rseq_entry.h
403
t->rseq.event.fatal = true;
include/linux/rseq_entry.h
408
return rseq_debug_update_user_cs(t, regs, csaddr);
include/linux/rseq_entry.h
445
if (unlikely(usig != t->rseq.sig))
include/linux/rseq_entry.h
449
unsafe_put_user(0ULL, &t->rseq.usrptr->rseq_cs, efault);
include/linux/rseq_entry.h
455
unsafe_put_user(0ULL, &t->rseq.usrptr->rseq_cs, efault);
include/linux/rseq_entry.h
464
t->rseq.event.fatal = true;
include/linux/rseq_entry.h
492
bool rseq_set_ids_get_csaddr(struct task_struct *t, struct rseq_ids *ids,
include/linux/rseq_entry.h
495
struct rseq __user *rseq = t->rseq.usrptr;
include/linux/rseq_entry.h
498
if (!rseq_debug_validate_ids(t))
include/linux/rseq_entry.h
517
rseq_slice_clear_grant(t);
include/linux/rseq_entry.h
519
t->rseq.ids.cpu_cid = ids->cpu_cid;
include/linux/rseq_entry.h
521
rseq_trace_update(t, ids);
include/linux/rseq_entry.h
53
void __rseq_trace_update(struct task_struct *t);
include/linux/rseq_entry.h
531
static rseq_inline bool rseq_update_usr(struct task_struct *t, struct pt_regs *regs,
include/linux/rseq_entry.h
536
if (!rseq_set_ids_get_csaddr(t, ids, node_id, &csaddr))
include/linux/rseq_entry.h
546
if (likely(!t->rseq.event.user_irq))
include/linux/rseq_entry.h
553
return rseq_update_user_cs(t, regs, csaddr);
include/linux/rseq_entry.h
57
static inline void rseq_trace_update(struct task_struct *t, struct rseq_ids *ids)
include/linux/rseq_entry.h
598
static __always_inline bool rseq_exit_user_update(struct pt_regs *regs, struct task_struct *t)
include/linux/rseq_entry.h
60
__rseq_trace_update(t);
include/linux/rseq_entry.h
605
if (likely(!t->rseq.event.ids_changed)) {
include/linux/rseq_entry.h
606
struct rseq __user *rseq = t->rseq.usrptr;
include/linux/rseq_entry.h
623
rseq_slice_clear_grant(t);
include/linux/rseq_entry.h
626
if (unlikely(!rseq_update_user_cs(t, regs, csaddr)))
include/linux/rseq_entry.h
633
.cpu_id = task_cpu(t),
include/linux/rseq_entry.h
634
.mm_cid = task_mm_cid(t),
include/linux/rseq_entry.h
638
return rseq_update_usr(t, regs, &ids, node_id);
include/linux/rseq_entry.h
645
struct task_struct *t = current;
include/linux/rseq_entry.h
664
if (unlikely((t->rseq.event.sched_switch))) {
include/linux/rseq_entry.h
667
if (unlikely(!rseq_exit_user_update(regs, t)))
include/linux/rseq_entry.h
671
t->rseq.event.events = 0;
include/linux/rseq_entry.h
71
static inline void rseq_trace_update(struct task_struct *t, struct rseq_ids *ids) { }
include/linux/sched.h
2355
void sched_mm_cid_before_execve(struct task_struct *t);
include/linux/sched.h
2356
void sched_mm_cid_after_execve(struct task_struct *t);
include/linux/sched.h
2357
void sched_mm_cid_exit(struct task_struct *t);
include/linux/sched.h
2358
static __always_inline int task_mm_cid(struct task_struct *t)
include/linux/sched.h
2360
return t->mm_cid.cid & ~(MM_CID_ONCPU | MM_CID_TRANSIT);
include/linux/sched.h
2363
static inline void sched_mm_cid_before_execve(struct task_struct *t) { }
include/linux/sched.h
2364
static inline void sched_mm_cid_after_execve(struct task_struct *t) { }
include/linux/sched.h
2365
static inline void sched_mm_cid_exit(struct task_struct *t) { }
include/linux/sched.h
2366
static __always_inline int task_mm_cid(struct task_struct *t)
include/linux/sched.h
2373
return task_cpu(t);
include/linux/sched/cputime.h
13
extern bool task_cputime(struct task_struct *t,
include/linux/sched/cputime.h
15
extern u64 task_gtime(struct task_struct *t);
include/linux/sched/cputime.h
17
static inline bool task_cputime(struct task_struct *t,
include/linux/sched/cputime.h
20
*utime = t->utime;
include/linux/sched/cputime.h
21
*stime = t->stime;
include/linux/sched/cputime.h
25
static inline u64 task_gtime(struct task_struct *t)
include/linux/sched/cputime.h
27
return t->gtime;
include/linux/sched/cputime.h
32
static inline void task_cputime_scaled(struct task_struct *t,
include/linux/sched/cputime.h
36
*utimescaled = t->utimescaled;
include/linux/sched/cputime.h
37
*stimescaled = t->stimescaled;
include/linux/sched/cputime.h
40
static inline void task_cputime_scaled(struct task_struct *t,
include/linux/sched/cputime.h
44
task_cputime(t, utimescaled, stimescaled);
include/linux/sched/isolation.h
40
extern void housekeeping_affine(struct task_struct *t, enum hk_type type);
include/linux/sched/isolation.h
62
static inline void housekeeping_affine(struct task_struct *t,
include/linux/sched/signal.h
314
struct task_struct *t);
include/linux/sched/signal.h
316
int send_sig_fault(int sig, int code, void __user *addr, struct task_struct *t);
include/linux/sched/signal.h
328
struct task_struct *t);
include/linux/sched/signal.h
441
extern void signal_wake_up_state(struct task_struct *t, unsigned int state);
include/linux/sched/signal.h
443
static inline void signal_wake_up(struct task_struct *t, bool fatal)
include/linux/sched/signal.h
446
if (fatal && !(t->jobctl & JOBCTL_PTRACE_FROZEN)) {
include/linux/sched/signal.h
447
t->jobctl &= ~(JOBCTL_STOPPED | JOBCTL_TRACED);
include/linux/sched/signal.h
450
signal_wake_up_state(t, state);
include/linux/sched/signal.h
452
static inline void ptrace_signal_wake_up(struct task_struct *t, bool resume)
include/linux/sched/signal.h
456
t->jobctl &= ~JOBCTL_TRACED;
include/linux/sched/signal.h
459
signal_wake_up_state(t, state);
include/linux/sched/signal.h
648
#define while_each_thread(g, t) \
include/linux/sched/signal.h
649
while ((t = next_thread(t)) != g)
include/linux/sched/signal.h
651
#define for_other_threads(p, t) \
include/linux/sched/signal.h
652
for (t = p; (t = next_thread(t)) != p; )
include/linux/sched/signal.h
654
#define __for_each_thread(signal, t) \
include/linux/sched/signal.h
655
list_for_each_entry_rcu(t, &(signal)->thread_head, thread_node, \
include/linux/sched/signal.h
658
#define for_each_thread(p, t) \
include/linux/sched/signal.h
659
__for_each_thread((p)->signal, t)
include/linux/sched/signal.h
662
#define for_each_process_thread(p, t) \
include/linux/sched/signal.h
663
for_each_process(p) for_each_thread(p, t)
include/linux/sched/task.h
114
static inline struct task_struct *get_task_struct(struct task_struct *t)
include/linux/sched/task.h
116
refcount_inc(&t->usage);
include/linux/sched/task.h
117
return t;
include/linux/sched/task.h
120
static inline struct task_struct *tryget_task_struct(struct task_struct *t)
include/linux/sched/task.h
122
return refcount_inc_not_zero(&t->usage) ? t : NULL;
include/linux/sched/task.h
125
extern void __put_task_struct(struct task_struct *t);
include/linux/sched/task.h
128
static inline void put_task_struct(struct task_struct *t)
include/linux/sched/task.h
130
if (!refcount_dec_and_test(&t->usage))
include/linux/sched/task.h
159
call_rcu(&t->rcu, __put_task_struct_rcu_cb);
include/linux/sched/task.h
164
static inline void put_task_struct_many(struct task_struct *t, int nr)
include/linux/sched/task.h
166
if (refcount_sub_and_test(nr, &t->usage))
include/linux/sched/task.h
167
__put_task_struct(t);
include/linux/sched/task.h
196
static inline struct vm_struct *task_stack_vm_area(const struct task_struct *t)
include/linux/sched/task.h
198
return t->stack_vm_area;
include/linux/sched/task.h
201
static inline struct vm_struct *task_stack_vm_area(const struct task_struct *t)
include/linux/signal.h
451
#define sig_fatal(t, signr) \
include/linux/signal.h
453
(t)->sighand->action[(signr)-1].sa.sa_handler == SIG_DFL)
include/linux/signal.h
462
struct task_struct *t = current; \
include/linux/signal.h
463
unsafe_put_user((void __user *)t->sas_ss_sp, &__uss->ss_sp, label); \
include/linux/signal.h
464
unsafe_put_user(t->sas_ss_flags, &__uss->ss_flags, label); \
include/linux/signal.h
465
unsafe_put_user(t->sas_ss_size, &__uss->ss_size, label); \
include/linux/skbuff.h
4427
static inline ktime_t net_timedelta(ktime_t t)
include/linux/skbuff.h
4429
return ktime_sub(ktime_get_real(), t);
include/linux/spi/spi.h
1258
spi_message_add_tail(struct spi_transfer *t, struct spi_message *m)
include/linux/spi/spi.h
1260
list_add_tail(&t->transfer_list, &m->transfers);
include/linux/spi/spi.h
1264
spi_transfer_del(struct spi_transfer *t)
include/linux/spi/spi.h
1266
list_del(&t->transfer_list);
include/linux/spi/spi.h
1270
spi_transfer_delay_exec(struct spi_transfer *t)
include/linux/spi/spi.h
1272
return spi_delay_exec(&t->delay, t);
include/linux/spi/spi.h
1303
struct spi_transfer t[];
include/linux/spi/spi.h
1307
mwt = kzalloc_flex(*mwt, t, ntrans, flags);
include/linux/spi/spi.h
1313
spi_message_add_tail(&mwt->t[i], &mwt->m);
include/linux/spi/spi.h
1518
struct spi_transfer t = {
include/linux/spi/spi.h
1523
return spi_sync_transfer(spi, &t, 1);
include/linux/spi/spi.h
1541
struct spi_transfer t = {
include/linux/spi/spi.h
1546
return spi_sync_transfer(spi, &t, 1);
include/linux/spi/spi_bitbang.h
21
struct spi_transfer *t);
include/linux/spi/spi_bitbang.h
31
int (*txrx_bufs)(struct spi_device *spi, struct spi_transfer *t);
include/linux/spi/spi_bitbang.h
45
struct spi_transfer *t);
include/linux/sunrpc/sched.h
148
#define RPC_IS_ASYNC(t) ((t)->tk_flags & RPC_TASK_ASYNC)
include/linux/sunrpc/sched.h
149
#define RPC_IS_SWAPPER(t) ((t)->tk_flags & RPC_TASK_SWAPPER)
include/linux/sunrpc/sched.h
150
#define RPC_IS_SOFT(t) ((t)->tk_flags & (RPC_TASK_SOFT|RPC_TASK_TIMEOUT))
include/linux/sunrpc/sched.h
151
#define RPC_IS_SOFTCONN(t) ((t)->tk_flags & RPC_TASK_SOFTCONN)
include/linux/sunrpc/sched.h
152
#define RPC_WAS_SENT(t) ((t)->tk_flags & RPC_TASK_SENT)
include/linux/sunrpc/sched.h
153
#define RPC_IS_MOVEABLE(t) ((t)->tk_flags & RPC_TASK_MOVEABLE)
include/linux/sunrpc/sched.h
164
#define rpc_test_and_set_running(t) \
include/linux/sunrpc/sched.h
165
test_and_set_bit(RPC_TASK_RUNNING, &(t)->tk_runstate)
include/linux/sunrpc/sched.h
166
#define rpc_clear_running(t) clear_bit(RPC_TASK_RUNNING, &(t)->tk_runstate)
include/linux/sunrpc/sched.h
168
#define RPC_IS_QUEUED(t) test_bit(RPC_TASK_QUEUED, &(t)->tk_runstate)
include/linux/sunrpc/sched.h
169
#define rpc_set_queued(t) set_bit(RPC_TASK_QUEUED, &(t)->tk_runstate)
include/linux/sunrpc/sched.h
170
#define rpc_clear_queued(t) clear_bit(RPC_TASK_QUEUED, &(t)->tk_runstate)
include/linux/sunrpc/sched.h
172
#define RPC_IS_ACTIVATED(t) test_bit(RPC_TASK_ACTIVE, &(t)->tk_runstate)
include/linux/sunrpc/sched.h
174
#define RPC_SIGNALLED(t) (READ_ONCE(task->tk_rpc_status) == -ERESTARTSYS)
include/linux/sunrpc/timer.h
29
int *t;
include/linux/sunrpc/timer.h
32
t = &rt->ntimeouts[timer-1];
include/linux/sunrpc/timer.h
33
if (ntimeo < *t) {
include/linux/sunrpc/timer.h
34
if (*t > 0)
include/linux/sunrpc/timer.h
35
(*t)--;
include/linux/sunrpc/timer.h
39
*t = ntimeo;
include/linux/suspend.h
453
extern void pm_report_hw_sleep_time(u64 t);
include/linux/suspend.h
454
extern void pm_report_max_hw_sleep(u64 t);
include/linux/suspend.h
501
static inline void pm_report_hw_sleep_time(u64 t) {};
include/linux/suspend.h
502
static inline void pm_report_max_hw_sleep(u64 t) {};
include/linux/syscalls.h
1094
struct old_timeval32 __user *t);
include/linux/syscalls.h
1096
struct old_utimbuf32 __user *t);
include/linux/syscalls.h
1098
struct old_timeval32 __user *t);
include/linux/syscalls.h
118
#define __MAP1(m,t,a,...) m(t,a)
include/linux/syscalls.h
119
#define __MAP2(m,t,a,...) m(t,a), __MAP1(m,__VA_ARGS__)
include/linux/syscalls.h
120
#define __MAP3(m,t,a,...) m(t,a), __MAP2(m,__VA_ARGS__)
include/linux/syscalls.h
121
#define __MAP4(m,t,a,...) m(t,a), __MAP3(m,__VA_ARGS__)
include/linux/syscalls.h
122
#define __MAP5(m,t,a,...) m(t,a), __MAP4(m,__VA_ARGS__)
include/linux/syscalls.h
123
#define __MAP6(m,t,a,...) m(t,a), __MAP5(m,__VA_ARGS__)
include/linux/syscalls.h
126
#define __SC_DECL(t, a) t a
include/linux/syscalls.h
127
#define __TYPE_AS(t, v) __same_type((__force t)0, v)
include/linux/syscalls.h
128
#define __TYPE_IS_L(t) (__TYPE_AS(t, 0L))
include/linux/syscalls.h
129
#define __TYPE_IS_UL(t) (__TYPE_AS(t, 0UL))
include/linux/syscalls.h
130
#define __TYPE_IS_LL(t) (__TYPE_AS(t, 0LL) || __TYPE_AS(t, 0ULL))
include/linux/syscalls.h
131
#define __SC_LONG(t, a) __typeof(__builtin_choose_expr(__TYPE_IS_LL(t), 0LL, 0L)) a
include/linux/syscalls.h
132
#define __SC_CAST(t, a) (__force t) a
include/linux/syscalls.h
133
#define __SC_TYPE(t, a) t
include/linux/syscalls.h
134
#define __SC_ARGS(t, a) a
include/linux/syscalls.h
135
#define __SC_TEST(t, a) (void)BUILD_BUG_ON_ZERO(!__TYPE_IS_LL(t) && sizeof(t) > sizeof(long))
include/linux/syscalls.h
138
#define __SC_STR_ADECL(t, a) #a
include/linux/syscalls.h
139
#define __SC_STR_TDECL(t, a) #t
include/linux/syscalls.h
555
struct old_timespec32 __user *t, int flags);
include/linux/sysfs.h
613
static inline int sysfs_rename_link_ns(struct kobject *k, struct kobject *t,
include/linux/sysfs.h
620
static inline void sysfs_delete_link(struct kobject *k, struct kobject *t,
include/linux/thread_info.h
160
#define read_task_thread_flags(t) \
include/linux/thread_info.h
161
read_ti_thread_flags(task_thread_info(t))
include/linux/thread_info.h
171
#define set_task_syscall_work(t, fl) \
include/linux/thread_info.h
172
set_bit(SYSCALL_WORK_BIT_##fl, &task_thread_info(t)->syscall_work)
include/linux/thread_info.h
173
#define test_task_syscall_work(t, fl) \
include/linux/thread_info.h
174
test_bit(SYSCALL_WORK_BIT_##fl, &task_thread_info(t)->syscall_work)
include/linux/thread_info.h
175
#define clear_task_syscall_work(t, fl) \
include/linux/thread_info.h
176
clear_bit(SYSCALL_WORK_BIT_##fl, &task_thread_info(t)->syscall_work)
include/linux/thread_info.h
187
#define set_task_syscall_work(t, fl) \
include/linux/thread_info.h
188
set_ti_thread_flag(task_thread_info(t), TIF_##fl)
include/linux/thread_info.h
189
#define test_task_syscall_work(t, fl) \
include/linux/thread_info.h
190
test_ti_thread_flag(task_thread_info(t), TIF_##fl)
include/linux/thread_info.h
191
#define clear_task_syscall_work(t, fl) \
include/linux/thread_info.h
192
clear_ti_thread_flag(task_thread_info(t), TIF_##fl)
include/linux/time.h
98
#define time_between32(t, l, h) ((u32)(h) - (u32)(l) >= (u32)(t) - (u32)(l))
include/linux/tnum.h
135
u64 tnum_step(struct tnum t, u64 z);
include/linux/units.h
100
t = millicelsius_to_milli_kelvin(t * MILLIDEGREE_PER_DEGREE);
include/linux/units.h
102
return DIV_ROUND_CLOSEST(t, MILLIDEGREE_PER_DECIDEGREE);
include/linux/units.h
112
static inline long deci_kelvin_to_millicelsius_with_offset(long t, long offset)
include/linux/units.h
114
return t * MILLIDEGREE_PER_DECIDEGREE - offset;
include/linux/units.h
117
static inline long deci_kelvin_to_millicelsius(long t)
include/linux/units.h
119
return milli_kelvin_to_millicelsius(t * MILLIDEGREE_PER_DECIDEGREE);
include/linux/units.h
122
static inline long millicelsius_to_deci_kelvin(long t)
include/linux/units.h
124
t = millicelsius_to_milli_kelvin(t);
include/linux/units.h
126
return DIV_ROUND_CLOSEST(t, MILLIDEGREE_PER_DECIDEGREE);
include/linux/units.h
129
static inline long kelvin_to_celsius(long t)
include/linux/units.h
131
return t + DIV_ROUND_CLOSEST(ABSOLUTE_ZERO_MILLICELSIUS,
include/linux/units.h
135
static inline long celsius_to_kelvin(long t)
include/linux/units.h
137
return t - DIV_ROUND_CLOSEST(ABSOLUTE_ZERO_MILLICELSIUS,
include/linux/units.h
66
static inline long milli_kelvin_to_millicelsius(long t)
include/linux/units.h
68
return t + ABSOLUTE_ZERO_MILLICELSIUS;
include/linux/units.h
71
static inline long millicelsius_to_milli_kelvin(long t)
include/linux/units.h
73
return t - ABSOLUTE_ZERO_MILLICELSIUS;
include/linux/units.h
79
static inline long kelvin_to_millicelsius(long t)
include/linux/units.h
81
return milli_kelvin_to_millicelsius(t * MILLIDEGREE_PER_DEGREE);
include/linux/units.h
84
static inline long millicelsius_to_kelvin(long t)
include/linux/units.h
86
t = millicelsius_to_milli_kelvin(t);
include/linux/units.h
88
return DIV_ROUND_CLOSEST(t, MILLIDEGREE_PER_DEGREE);
include/linux/units.h
91
static inline long deci_kelvin_to_celsius(long t)
include/linux/units.h
93
t = milli_kelvin_to_millicelsius(t * MILLIDEGREE_PER_DECIDEGREE);
include/linux/units.h
95
return DIV_ROUND_CLOSEST(t, MILLIDEGREE_PER_DEGREE);
include/linux/units.h
98
static inline long celsius_to_deci_kelvin(long t)
include/linux/uprobes.h
218
extern void uprobe_free_utask(struct task_struct *t);
include/linux/uprobes.h
219
extern void uprobe_copy_process(struct task_struct *t, u64 flags);
include/linux/uprobes.h
298
static inline void uprobe_free_utask(struct task_struct *t)
include/linux/uprobes.h
301
static inline void uprobe_copy_process(struct task_struct *t, u64 flags)
include/linux/usb/pd.h
221
#define PDO_TYPE(t) ((t) << PDO_TYPE_SHIFT)
include/linux/usb/pd.h
282
#define PDO_APDO_TYPE(t) ((t) << PDO_APDO_TYPE_SHIFT)
include/linux/user_events.h
30
extern void user_event_mm_dup(struct task_struct *t,
include/linux/user_events.h
33
extern void user_event_mm_remove(struct task_struct *t);
include/linux/user_events.h
35
static inline void user_events_fork(struct task_struct *t,
include/linux/user_events.h
40
if (!t || !current->user_event_mm)
include/linux/user_events.h
46
t->user_event_mm = old_mm;
include/linux/user_events.h
51
user_event_mm_dup(t, old_mm);
include/linux/user_events.h
54
static inline void user_events_execve(struct task_struct *t)
include/linux/user_events.h
56
if (!t || !t->user_event_mm)
include/linux/user_events.h
59
user_event_mm_remove(t);
include/linux/user_events.h
62
static inline void user_events_exit(struct task_struct *t)
include/linux/user_events.h
64
if (!t || !t->user_event_mm)
include/linux/user_events.h
67
user_event_mm_remove(t);
include/linux/user_events.h
70
static inline void user_events_fork(struct task_struct *t,
include/linux/user_events.h
75
static inline void user_events_execve(struct task_struct *t)
include/linux/user_events.h
79
static inline void user_events_exit(struct task_struct *t)
include/linux/virtio_vsock.h
283
void virtio_transport_recv_pkt(struct virtio_transport *t,
include/linux/watchdog.h
173
unsigned int t)
include/linux/watchdog.h
186
return t > UINT_MAX / 1000 || t < wdd->min_timeout ||
include/linux/watchdog.h
188
t > wdd->max_timeout);
include/linux/watchdog.h
193
unsigned int t)
include/linux/watchdog.h
195
return t && wdd->timeout && t >= wdd->timeout;
include/linux/win_minmax.h
13
u32 t; /* time measurement was taken */
include/linux/win_minmax.h
27
static inline u32 minmax_reset(struct minmax *m, u32 t, u32 meas)
include/linux/win_minmax.h
29
struct minmax_sample val = { .t = t, .v = meas };
include/linux/win_minmax.h
35
u32 minmax_running_max(struct minmax *m, u32 win, u32 t, u32 meas);
include/linux/win_minmax.h
36
u32 minmax_running_min(struct minmax *m, u32 win, u32 t, u32 meas);
include/linux/workqueue_types.h
14
void delayed_work_timer_fn(struct timer_list *t);
include/media/drv-intf/msp3400.h
128
#define MSP_INPUT(sc, t, main_aux_src, sc_i2s_src) \
include/media/drv-intf/msp3400.h
130
MSP_TUNER_TO_DSP(t) | \
include/media/drv-intf/saa7146_vv.h
148
void saa7146_buffer_timeout(struct timer_list *t);
include/media/i2c/ov772x.h
41
#define OV772X_MANUAL_EDGECTRL(s, t) \
include/media/i2c/ov772x.h
45
.threshold = (t & OV772X_EDGE_THRESHOLD_MASK), \
include/media/v4l2-dv-timings.h
110
bool v4l2_find_dv_timings_cea861_vic(struct v4l2_dv_timings *t, u8 vic);
include/media/v4l2-dv-timings.h
136
const struct v4l2_dv_timings *t, bool detailed);
include/media/v4l2-dv-timings.h
209
struct v4l2_fract v4l2_dv_timings_aspect_ratio(const struct v4l2_dv_timings *t);
include/media/v4l2-dv-timings.h
23
struct v4l2_fract v4l2_calc_timeperframe(const struct v4l2_dv_timings *t);
include/media/v4l2-dv-timings.h
38
typedef bool v4l2_check_dv_timings_fnc(const struct v4l2_dv_timings *t, void *handle);
include/media/v4l2-dv-timings.h
52
bool v4l2_valid_dv_timings(const struct v4l2_dv_timings *t,
include/media/v4l2-dv-timings.h
73
int v4l2_enum_dv_timings_cap(struct v4l2_enum_dv_timings *t,
include/media/v4l2-dv-timings.h
96
bool v4l2_find_dv_timings_cap(struct v4l2_dv_timings *t,
include/net/af_vsock.h
191
int vsock_core_register(const struct vsock_transport *t, int features);
include/net/af_vsock.h
192
void vsock_core_unregister(const struct vsock_transport *t);
include/net/af_vsock.h
261
static inline bool vsock_msgzerocopy_allow(const struct vsock_transport *t)
include/net/af_vsock.h
263
return t->msgzerocopy_allow && t->msgzerocopy_allow();
include/net/bluetooth/l2cap.h
865
#define __set_chan_timer(c, t) l2cap_set_timer(c, &c->chan_timer, (t))
include/net/cfg80211.h
6500
void wiphy_delayed_work_timer(struct timer_list *t);
include/net/cfg80211.h
6603
enum hrtimer_restart wiphy_hrtimer_work_timer(struct hrtimer *t);
include/net/codel_impl.h
101
return t + reciprocal_scale(interval, rec_inv_sqrt << REC_INV_SQRT_SHIFT);
include/net/codel_impl.h
97
static codel_time_t codel_control_law(codel_time_t t,
include/net/erspan.h
104
t:1,
include/net/erspan.h
114
t:1,
include/net/erspan.h
211
ershdr->t = truncate;
include/net/erspan.h
305
ershdr->t = truncate;
include/net/ieee802154_netdev.h
480
struct ieee802154_llsec_table **t);
include/net/inet_frag.h
113
void (*frag_expire)(struct timer_list *t);
include/net/ip6_tunnel.h
109
static inline int ip6_tnl_encap(struct sk_buff *skb, struct ip6_tnl *t,
include/net/ip6_tunnel.h
115
if (t->encap.type == TUNNEL_ENCAP_NONE)
include/net/ip6_tunnel.h
118
if (t->encap.type >= MAX_IPTUN_ENCAP_OPS)
include/net/ip6_tunnel.h
122
ops = rcu_dereference(ip6tun_encaps[t->encap.type]);
include/net/ip6_tunnel.h
124
ret = ops->build_header(skb, &t->encap, protocol, fl6);
include/net/ip6_tunnel.h
138
int ip6_tnl_rcv_ctl(struct ip6_tnl *t, const struct in6_addr *laddr,
include/net/ip6_tunnel.h
143
int ip6_tnl_xmit_ctl(struct ip6_tnl *t, const struct in6_addr *laddr,
include/net/ip6_tunnel.h
148
__u32 ip6_tnl_get_cap(struct ip6_tnl *t, const struct in6_addr *laddr,
include/net/ip6_tunnel.h
86
int ip6_tnl_encap_setup(struct ip6_tnl *t,
include/net/ip_tunnels.h
460
int ip_tunnel_encap_setup(struct ip_tunnel *t,
include/net/ip_vs.h
1304
#define IP_VS_INIT_HASH_TABLE(t) ip_vs_init_hash_table((t), ARRAY_SIZE((t)))
include/net/llc_c_ac.h
180
void llc_conn_busy_tmr_cb(struct timer_list *t);
include/net/llc_c_ac.h
181
void llc_conn_pf_cycle_tmr_cb(struct timer_list *t);
include/net/llc_c_ac.h
182
void llc_conn_ack_tmr_cb(struct timer_list *t);
include/net/llc_c_ac.h
183
void llc_conn_rej_tmr_cb(struct timer_list *t);
include/net/netfilter/nf_conntrack_l4proto.h
150
struct nf_conntrack_tuple *t,
include/net/netfilter/nf_conntrack_l4proto.h
41
const struct nf_conntrack_tuple *t);
include/net/netfilter/nf_conntrack_l4proto.h
45
struct nf_conntrack_tuple *t,
include/net/netfilter/nf_conntrack_timeout.h
26
nf_ct_timeout_data(const struct nf_conn_timeout *t)
include/net/netfilter/nf_conntrack_timeout.h
31
timeout = rcu_dereference(t->timeout);
include/net/netfilter/nf_conntrack_tuple.h
100
t->src.u3.all, ntohs(t->src.u.all),
include/net/netfilter/nf_conntrack_tuple.h
101
t->dst.u3.all, ntohs(t->dst.u.all));
include/net/netfilter/nf_conntrack_tuple.h
105
static inline void nf_ct_dump_tuple(const struct nf_conntrack_tuple *t)
include/net/netfilter/nf_conntrack_tuple.h
107
switch (t->src.l3num) {
include/net/netfilter/nf_conntrack_tuple.h
109
nf_ct_dump_tuple_ip(t);
include/net/netfilter/nf_conntrack_tuple.h
112
nf_ct_dump_tuple_ipv6(t);
include/net/netfilter/nf_conntrack_tuple.h
182
nf_ct_tuple_mask_cmp(const struct nf_conntrack_tuple *t,
include/net/netfilter/nf_conntrack_tuple.h
186
return nf_ct_tuple_src_mask_cmp(t, tuple, mask) &&
include/net/netfilter/nf_conntrack_tuple.h
187
__nf_ct_tuple_dst_equal(t, tuple);
include/net/netfilter/nf_conntrack_tuple.h
85
static inline void nf_ct_dump_tuple_ip(const struct nf_conntrack_tuple *t)
include/net/netfilter/nf_conntrack_tuple.h
89
t, t->dst.protonum,
include/net/netfilter/nf_conntrack_tuple.h
90
&t->src.u3.ip, ntohs(t->src.u.all),
include/net/netfilter/nf_conntrack_tuple.h
91
&t->dst.u3.ip, ntohs(t->dst.u.all));
include/net/netfilter/nf_conntrack_tuple.h
95
static inline void nf_ct_dump_tuple_ipv6(const struct nf_conntrack_tuple *t)
include/net/netfilter/nf_conntrack_tuple.h
99
t, t->dst.protonum,
include/net/netfilter/nf_tables.h
1725
#define nft_trans_container_set(t) \
include/net/netfilter/nf_tables.h
1726
container_of(t, struct nft_trans_set, nft_trans_binding.nft_trans)
include/net/netfilter/nf_tables.h
1755
#define nft_trans_container_chain(t) \
include/net/netfilter/nf_tables.h
1756
container_of(t, struct nft_trans_chain, nft_trans_binding.nft_trans)
include/net/netfilter/nf_tables.h
1810
#define nft_trans_container_elem(t) \
include/net/netfilter/nf_tables.h
1811
container_of(t, struct nft_trans_elem, nft_trans)
include/net/netfilter/nf_tables.h
1824
#define nft_trans_container_obj(t) \
include/net/netfilter/nf_tables.h
1825
container_of(t, struct nft_trans_obj, nft_trans)
include/net/netfilter/nf_tables.h
1841
#define nft_trans_container_flowtable(t) \
include/net/netfilter/nf_tables.h
1842
container_of(t, struct nft_trans_flowtable, nft_trans)
include/net/pkt_cls.h
524
#define tcf_em_tree_validate(tp, tb, t) ((void)(t), 0)
include/net/pkt_cls.h
525
#define tcf_em_tree_destroy(t) do { (void)(t); } while(0)
include/net/pkt_cls.h
526
#define tcf_em_tree_dump(skb, t, tlv) (0)
include/net/pkt_cls.h
527
#define tcf_em_tree_match(skb, t, info) ((void)(info), 1)
include/net/sch_generic.h
406
struct tcmsg *t, bool rtnl_held);
include/net/sctp/sctp.h
144
struct sctp_transport *t, __u32 pmtu);
include/net/sctp/sctp.h
149
struct sctp_transport *t);
include/net/sctp/sctp.h
152
int sctp_hash_transport(struct sctp_transport *t);
include/net/sctp/sctp.h
153
void sctp_unhash_transport(struct sctp_transport *t);
include/net/sctp/sctp.h
558
static inline struct dst_entry *sctp_transport_dst_check(struct sctp_transport *t)
include/net/sctp/sctp.h
560
if (t->dst && !dst_check(t->dst, t->dst_cookie))
include/net/sctp/sctp.h
561
sctp_transport_dst_release(t);
include/net/sctp/sctp.h
563
return t->dst;
include/net/sctp/sctp.h
570
const struct sctp_transport *t,
include/net/sctp/sctp.h
577
if (sp->udp_port && (!t || t->encap_port))
include/net/sctp/sctp.h
601
static inline bool sctp_transport_pmtu_check(struct sctp_transport *t)
include/net/sctp/sctp.h
603
__u32 pmtu = sctp_dst_mtu(t->dst);
include/net/sctp/sctp.h
605
if (t->pathmtu == pmtu)
include/net/sctp/sctp.h
608
t->pathmtu = pmtu;
include/net/sctp/sctp.h
618
static inline int sctp_transport_pl_hlen(struct sctp_transport *t)
include/net/sctp/sctp.h
620
return __sctp_mtu_payload(sctp_sk(t->asoc->base.sk), t, 0, 0) -
include/net/sctp/sctp.h
624
static inline void sctp_transport_pl_reset(struct sctp_transport *t)
include/net/sctp/sctp.h
626
if (t->probe_interval && (t->param_flags & SPP_PMTUD_ENABLE) &&
include/net/sctp/sctp.h
627
(t->state == SCTP_ACTIVE || t->state == SCTP_UNKNOWN)) {
include/net/sctp/sctp.h
628
if (t->pl.state == SCTP_PL_DISABLED) {
include/net/sctp/sctp.h
629
t->pl.state = SCTP_PL_BASE;
include/net/sctp/sctp.h
630
t->pl.pmtu = SCTP_BASE_PLPMTU;
include/net/sctp/sctp.h
631
t->pl.probe_size = SCTP_BASE_PLPMTU;
include/net/sctp/sctp.h
632
sctp_transport_reset_probe_timer(t);
include/net/sctp/sctp.h
635
if (t->pl.state != SCTP_PL_DISABLED) {
include/net/sctp/sctp.h
636
if (timer_delete(&t->probe_timer))
include/net/sctp/sctp.h
637
sctp_transport_put(t);
include/net/sctp/sctp.h
638
t->pl.state = SCTP_PL_DISABLED;
include/net/sctp/sctp.h
643
static inline void sctp_transport_pl_update(struct sctp_transport *t)
include/net/sctp/sctp.h
645
if (t->pl.state == SCTP_PL_DISABLED)
include/net/sctp/sctp.h
648
t->pl.state = SCTP_PL_BASE;
include/net/sctp/sctp.h
649
t->pl.pmtu = SCTP_BASE_PLPMTU;
include/net/sctp/sctp.h
650
t->pl.probe_size = SCTP_BASE_PLPMTU;
include/net/sctp/sctp.h
651
sctp_transport_reset_probe_timer(t);
include/net/sctp/sctp.h
654
static inline bool sctp_transport_pl_enabled(struct sctp_transport *t)
include/net/sctp/sctp.h
656
return t->pl.state != SCTP_PL_DISABLED;
include/net/sctp/structs.h
1006
void sctp_transport_lower_cwnd(struct sctp_transport *t,
include/net/sctp/structs.h
1011
void sctp_transport_reset(struct sctp_transport *t);
include/net/sctp/structs.h
1012
bool sctp_transport_update_pmtu(struct sctp_transport *t, u32 pmtu);
include/net/sctp/structs.h
1014
void sctp_transport_dst_release(struct sctp_transport *t);
include/net/sctp/structs.h
1015
void sctp_transport_dst_confirm(struct sctp_transport *t);
include/net/sctp/structs.h
1016
void sctp_transport_pl_send(struct sctp_transport *t);
include/net/sctp/structs.h
1017
bool sctp_transport_pl_recv(struct sctp_transport *t);
include/net/sctp/structs.h
438
void (*get_dst) (struct sctp_transport *t,
include/net/sctp/structs.h
443
struct sctp_transport *t,
include/net/tc_act/tc_tunnel_key.h
31
struct tcf_tunnel_key *t = to_tunnel_key(a);
include/net/tc_act/tc_tunnel_key.h
34
params = rcu_dereference_protected(t->params,
include/net/tc_act/tc_tunnel_key.h
45
struct tcf_tunnel_key *t = to_tunnel_key(a);
include/net/tc_act/tc_tunnel_key.h
48
params = rcu_dereference_protected(t->params,
include/net/tc_act/tc_tunnel_key.h
59
struct tcf_tunnel_key *t = to_tunnel_key(a);
include/net/tc_act/tc_tunnel_key.h
62
params = rcu_dereference_protected(t->params,
include/net/xfrm.h
1823
struct ip6_tnl *t);
include/net/xfrm.h
1827
int xfrm6_rcv_tnl(struct sk_buff *skb, struct ip6_tnl *t);
include/net/xfrm.h
403
int km_query(struct xfrm_state *x, struct xfrm_tmpl *t,
include/net/xsk_buff_pool.h
34
#define XSK_CHECK_PRIV_TYPE(t) BUILD_BUG_ON(sizeof(t) > offsetofend(struct xdp_buff_xsk, cb))
include/net/xsk_buff_pool.h
35
#define XSK_TX_COMPL_FITS(t) BUILD_BUG_ON(sizeof(struct xsk_tx_metadata_compl) > sizeof(t))
include/rdma/rdmavt_qp.h
877
enum hrtimer_restart rvt_rc_rnr_retry(struct hrtimer *t);
include/scsi/libfcoe.h
386
void fcoe_queue_timer(struct timer_list *t);
include/scsi/scsi_host.h
814
extern void scsi_host_put(struct Scsi_Host *t);
include/scsi/scsi_transport.h
57
scsi_transport_reserve_target(struct scsi_transport_template * t, int space)
include/scsi/scsi_transport.h
59
BUG_ON(t->target_private_offset != 0);
include/scsi/scsi_transport.h
60
t->target_private_offset = ALIGN(t->target_size, sizeof(void *));
include/scsi/scsi_transport.h
61
t->target_size = t->target_private_offset + space;
include/scsi/scsi_transport.h
64
scsi_transport_reserve_device(struct scsi_transport_template * t, int space)
include/scsi/scsi_transport.h
66
BUG_ON(t->device_private_offset != 0);
include/scsi/scsi_transport.h
67
t->device_private_offset = ALIGN(t->device_size, sizeof(void *));
include/scsi/scsi_transport.h
68
t->device_size = t->device_private_offset + space;
include/scsi/scsi_transport_iscsi.h
467
struct iscsi_transport *t,
include/trace/bpf_probe.h
13
#define __perf_task(t) (t)
include/trace/events/f2fs.h
74
#define F2FS_BIO_FLAG_MASK(t) (__force u32)((t) & F2FS_OP_FLAGS)
include/trace/events/irq.h
165
TP_PROTO(struct tasklet_struct *t, void *func),
include/trace/events/irq.h
167
TP_ARGS(t, func),
include/trace/events/irq.h
175
__entry->tasklet = t;
include/trace/events/irq.h
191
TP_PROTO(struct tasklet_struct *t, void *func),
include/trace/events/irq.h
193
TP_ARGS(t, func)
include/trace/events/irq.h
205
TP_PROTO(struct tasklet_struct *t, void *func),
include/trace/events/irq.h
207
TP_ARGS(t, func)
include/trace/events/osnoise.h
107
TP_PROTO(struct task_struct *t, u64 start, u64 duration),
include/trace/events/osnoise.h
109
TP_ARGS(t, start, duration),
include/trace/events/osnoise.h
119
memcpy(__entry->comm, t->comm, TASK_COMM_LEN);
include/trace/events/osnoise.h
120
__entry->pid = t->pid;
include/trace/events/rseq.h
13
TP_PROTO(struct task_struct *t),
include/trace/events/rseq.h
15
TP_ARGS(t),
include/trace/events/rseq.h
24
__entry->cpu_id = t->rseq.ids.cpu_id;
include/trace/events/rseq.h
26
__entry->mm_cid = t->rseq.ids.mm_cid;
include/trace/events/sched.h
18
TP_PROTO(struct task_struct *t),
include/trace/events/sched.h
20
TP_ARGS(t),
include/trace/events/sched.h
23
__string( comm, t->comm )
include/trace/events/sched.h
29
__entry->pid = t->pid;
include/trace/perf.h
13
#define __perf_task(t) (__task = (t))
include/trace/stages/stage6_event_callback.h
139
#define __perf_task(t) (t)
include/uapi/asm-generic/ioctl.h
76
#define _IOC_TYPECHECK(t) (sizeof(t))
include/uapi/linux/atm.h
58
#define __SO_ENCODE(l,n,t) ((((l) & 0x1FF) << 22) | ((n) << 16) | \
include/uapi/linux/atm.h
59
sizeof(t))
include/uapi/linux/keyboard.h
45
#define K(t,v) (((t)<<8)|(v))
include/uapi/linux/mroute6.h
66
#define IF_COPY(f, t) bcopy(f, t, sizeof(*(f)))
include/uapi/linux/ptp_clock.h
257
struct ptp_clock_time t; /* Time event occurred. */
include/uapi/linux/serial.h
181
#define SER_ISO7816_T(t) (((t) & 0x0f) << 4)
include/uapi/linux/soundcard.h
116
#define _SIOR(x,y,t) ((int)(SIOC_OUT|((sizeof(t)&SIOCPARM_MASK)<<16)|(x<<8)|y))
include/uapi/linux/soundcard.h
117
#define _SIOW(x,y,t) ((int)(SIOC_IN|((sizeof(t)&SIOCPARM_MASK)<<16)|(x<<8)|y))
include/uapi/linux/soundcard.h
119
#define _SIOWR(x,y,t) ((int)(SIOC_INOUT|((sizeof(t)&SIOCPARM_MASK)<<16)|(x<<8)|y))
include/uapi/scsi/scsi_netlink.h
102
(hdr)->transport = t; \
include/uapi/scsi/scsi_netlink.h
99
#define INIT_SCSI_NL_HDR(hdr, t, mtype, mlen) \
include/ufs/ufshci.h
314
#define UIC_ARG_ATTR_TYPE(t) (((t) & 0xFF) << 16)
include/ufs/ufshci.h
359
#define INT_AGGR_TIMEOUT_VAL(t) (((t) & 0xFF) << 0)
include/vdso/gettime.h
20
__kernel_old_time_t __vdso_time(__kernel_old_time_t *t);
include/video/newport.h
455
int t = BUSY_TIMEOUT;
include/video/newport.h
457
while (--t)
include/video/newport.h
460
return !t;
include/video/newport.h
465
int t = BUSY_TIMEOUT;
include/video/newport.h
467
while (--t)
include/video/newport.h
470
return !t;
include/video/sstfb.h
327
int (*set_pll) (struct fb_info *info, const struct pll_timing *t, const int clock);
include/video/uvesafb.h
106
struct uvesafb_task t;
include/xen/xenbus.h
152
char **xenbus_directory(struct xenbus_transaction t,
include/xen/xenbus.h
154
void *xenbus_read(struct xenbus_transaction t,
include/xen/xenbus.h
156
int xenbus_write(struct xenbus_transaction t,
include/xen/xenbus.h
158
int xenbus_exists(struct xenbus_transaction t,
include/xen/xenbus.h
160
int xenbus_rm(struct xenbus_transaction t, const char *dir, const char *node);
include/xen/xenbus.h
161
int xenbus_transaction_start(struct xenbus_transaction *t);
include/xen/xenbus.h
162
int xenbus_transaction_end(struct xenbus_transaction t, bool abort);
include/xen/xenbus.h
166
int xenbus_scanf(struct xenbus_transaction t,
include/xen/xenbus.h
175
int xenbus_printf(struct xenbus_transaction t,
include/xen/xenbus.h
180
int xenbus_gather(struct xenbus_transaction t, const char *dir, ...);
init/initramfs.c
135
struct timespec64 t[2] = { { .tv_sec = mtime }, { .tv_sec = mtime } };
init/initramfs.c
136
init_utimes(filename, t);
init/initramfs.c
141
struct timespec64 t[2] = { { .tv_sec = mtime }, { .tv_sec = mtime } };
init/initramfs.c
142
vfs_utimes(path, t);
io_uring/io_uring.h
137
struct hrtimer t;
io_uring/napi.c
35
static inline ktime_t net_to_ktime(unsigned long t)
io_uring/napi.c
38
return ns_to_ktime(t << 10);
io_uring/wait.c
107
hrtimer_update_function(&iowq->t, io_cqring_timer_wakeup);
io_uring/wait.c
121
hrtimer_setup_on_stack(&iowq->t, io_cqring_min_timer_wakeup, clock_id,
io_uring/wait.c
125
hrtimer_setup_on_stack(&iowq->t, io_cqring_timer_wakeup, clock_id,
io_uring/wait.c
129
hrtimer_set_expires_range_ns(&iowq->t, timeout, 0);
io_uring/wait.c
130
hrtimer_start_expires(&iowq->t, HRTIMER_MODE_ABS);
io_uring/wait.c
135
hrtimer_cancel(&iowq->t);
io_uring/wait.c
136
destroy_hrtimer_on_stack(&iowq->t);
io_uring/wait.c
56
struct io_wait_queue *iowq = container_of(timer, struct io_wait_queue, t);
io_uring/wait.c
71
struct io_wait_queue *iowq = container_of(timer, struct io_wait_queue, t);
ipc/msg.c
215
struct msg_sender *mss, *t;
ipc/msg.c
219
list_for_each_entry_safe(mss, t, h, list) {
ipc/msg.c
252
struct msg_receiver *msr, *t;
ipc/msg.c
254
list_for_each_entry_safe(msr, t, &msq->q_receivers, r_list) {
ipc/msg.c
277
struct msg_msg *msg, *t;
ipc/msg.c
288
list_for_each_entry_safe(msg, t, &msq->q_messages, m_list) {
ipc/msg.c
819
struct msg_receiver *msr, *t;
ipc/msg.c
821
list_for_each_entry_safe(msr, t, &msq->q_receivers, r_list) {
ipc/namespace.c
174
struct ipc_namespace *n, *t;
ipc/namespace.c
176
llist_for_each_entry_safe(n, t, node, mnt_llist)
ipc/namespace.c
182
llist_for_each_entry_safe(n, t, node, mnt_llist)
kernel/audit.c
1539
unsigned int t;
kernel/audit.c
1541
t = READ_ONCE(current->signal->audit_tty);
kernel/audit.c
1542
s.enabled = t & AUDIT_TTY_ENABLE;
kernel/audit.c
1543
s.log_passwd = !!(t & AUDIT_TTY_LOG_PASSWD);
kernel/audit.c
1551
unsigned int t;
kernel/audit.c
1562
t = READ_ONCE(current->signal->audit_tty);
kernel/audit.c
1564
t = s.enabled | (-s.log_passwd & AUDIT_TTY_LOG_PASSWD);
kernel/audit.c
1565
t = xchg(&current->signal->audit_tty, t);
kernel/audit.c
1567
old.enabled = t & AUDIT_TTY_ENABLE;
kernel/audit.c
1568
old.log_passwd = !!(t & AUDIT_TTY_LOG_PASSWD);
kernel/audit.c
2743
int audit_signal_info(int sig, struct task_struct *t)
kernel/audit.c
2747
if (auditd_test_task(t) &&
kernel/audit.c
2759
return audit_signal_info_syscall(t);
kernel/audit.c
985
int t = done ? NLMSG_DONE : type;
kernel/audit.c
991
nlh = nlmsg_put(skb, 0, seq, t, size, flags);
kernel/audit.h
307
extern int audit_signal_info_syscall(struct task_struct *t);
kernel/audit.h
326
#define audit_exe_compare(t, m) (-EINVAL)
kernel/audit.h
338
static inline int audit_signal_info_syscall(struct task_struct *t)
kernel/audit.h
343
#define audit_filter_inodes(t, c) do { } while (0)
kernel/auditsc.c
2665
void __audit_ptrace(struct task_struct *t)
kernel/auditsc.c
2669
context->target_pid = task_tgid_nr(t);
kernel/auditsc.c
2670
context->target_auid = audit_get_loginuid(t);
kernel/auditsc.c
2671
context->target_uid = task_uid(t);
kernel/auditsc.c
2672
context->target_sessionid = audit_get_sessionid(t);
kernel/auditsc.c
2673
strscpy(context->target_comm, t->comm);
kernel/auditsc.c
2674
security_task_getlsmprop_obj(t, &context->target_ref);
kernel/auditsc.c
2684
int audit_signal_info_syscall(struct task_struct *t)
kernel/auditsc.c
2688
kuid_t t_uid = task_uid(t);
kernel/auditsc.c
2696
ctx->target_pid = task_tgid_nr(t);
kernel/auditsc.c
2697
ctx->target_auid = audit_get_loginuid(t);
kernel/auditsc.c
2699
ctx->target_sessionid = audit_get_sessionid(t);
kernel/auditsc.c
2700
strscpy(ctx->target_comm, t->comm);
kernel/auditsc.c
2701
security_task_getlsmprop_obj(t, &ctx->target_ref);
kernel/auditsc.c
2717
axp->target_pid[axp->pid_count] = task_tgid_nr(t);
kernel/auditsc.c
2718
axp->target_auid[axp->pid_count] = audit_get_loginuid(t);
kernel/auditsc.c
2720
axp->target_sessionid[axp->pid_count] = audit_get_sessionid(t);
kernel/auditsc.c
2721
security_task_getlsmprop_obj(t, &axp->target_ref[axp->pid_count]);
kernel/auditsc.c
2722
strscpy(axp->target_comm[axp->pid_count], t->comm);
kernel/bpf/arena.c
670
struct llist_node *pos, *t;
kernel/bpf/arena.c
714
llist_for_each_safe(pos, t, __llist_del_all(&free_pages)) {
kernel/bpf/arena.c
788
struct llist_node *list, *pos, *t;
kernel/bpf/arena.c
824
llist_for_each_safe(pos, t, list) {
kernel/bpf/arena.c
840
llist_for_each_safe(pos, t, __llist_del_all(&free_pages)) {
kernel/bpf/bpf_lru_list.c
17
#define LOCAL_LIST_IDX(t) ((t) - BPF_LOCAL_LIST_T_OFFSET)
kernel/bpf/bpf_lru_list.c
20
#define IS_LOCAL_LIST_TYPE(t) ((t) >= BPF_LOCAL_LIST_T_OFFSET)
kernel/bpf/bpf_struct_ops.c
1029
static u32 count_func_ptrs(const struct btf *btf, const struct btf_type *t)
kernel/bpf/bpf_struct_ops.c
1036
for_each_member(i, t, member)
kernel/bpf/bpf_struct_ops.c
1047
const struct btf_type *t, *vt;
kernel/bpf/bpf_struct_ops.c
1090
t = st_ops_desc->type;
kernel/bpf/bpf_struct_ops.c
1108
st_map->funcs_cnt = count_func_ptrs(btf, t);
kernel/bpf/bpf_struct_ops.c
312
const struct btf_type *t;
kernel/bpf/bpf_struct_ops.c
314
t = btf_type_resolve_ptr(btf, id, NULL);
kernel/bpf/bpf_struct_ops.c
315
if (!t)
kernel/bpf/bpf_struct_ops.c
318
if (!__btf_type_is_struct(t) && !btf_type_is_fwd(t))
kernel/bpf/bpf_struct_ops.c
321
return !strcmp(btf_name_by_offset(btf, t->name_off), "module");
kernel/bpf/bpf_struct_ops.c
338
const struct btf_type *t;
kernel/bpf/bpf_struct_ops.c
364
t = btf_type_by_id(btf, type_id);
kernel/bpf/bpf_struct_ops.c
365
if (btf_type_vlen(t) > BPF_STRUCT_OPS_MAX_NR_MEMBERS) {
kernel/bpf/bpf_struct_ops.c
367
btf_type_vlen(t), st_ops->name);
kernel/bpf/bpf_struct_ops.c
378
if (!is_valid_value_type(btf, value_id, t, value_name))
kernel/bpf/bpf_struct_ops.c
381
arg_info = kzalloc_objs(*arg_info, btf_type_vlen(t));
kernel/bpf/bpf_struct_ops.c
386
st_ops_desc->type = t;
kernel/bpf/bpf_struct_ops.c
391
for_each_member(i, t, member) {
kernel/bpf/bpf_struct_ops.c
396
moff = __btf_member_bit_offset(t, member) / 8;
kernel/bpf/bpf_struct_ops.c
405
if (__btf_member_bitfield_size(t, member)) {
kernel/bpf/bpf_struct_ops.c
555
static int check_zero_holes(const struct btf *btf, const struct btf_type *t, void *data)
kernel/bpf/bpf_struct_ops.c
561
for_each_member(i, t, member) {
kernel/bpf/bpf_struct_ops.c
562
moff = __btf_member_bit_offset(t, member) / 8;
kernel/bpf/bpf_struct_ops.c
574
if (t->size > prev_mend &&
kernel/bpf/bpf_struct_ops.c
575
memchr_inv(data + prev_mend, 0, t->size - prev_mend))
kernel/bpf/bpf_struct_ops.c
695
const struct btf_type *t = st_ops_desc->type;
kernel/bpf/bpf_struct_ops.c
716
err = check_zero_holes(st_map->btf, t, uvalue->data);
kernel/bpf/bpf_struct_ops.c
744
tname = btf_name_by_offset(st_map->btf, t->name_off);
kernel/bpf/bpf_struct_ops.c
746
for_each_member(i, t, member) {
kernel/bpf/bpf_struct_ops.c
753
moff = __btf_member_bit_offset(t, member) / 8;
kernel/bpf/bpf_struct_ops.c
763
err = st_ops->init_member(t, member, kdata, udata);
kernel/bpf/btf.c
1000
bool btf_type_is_i64(const struct btf_type *t)
kernel/bpf/btf.c
1002
return btf_type_is_int(t) && __btf_type_int_is_regular(t, 8);
kernel/bpf/btf.c
1005
bool btf_type_is_primitive(const struct btf_type *t)
kernel/bpf/btf.c
1007
return (btf_type_is_int(t) && btf_type_int_is_regular(t)) ||
kernel/bpf/btf.c
1008
btf_is_any_enum(t);
kernel/bpf/btf.c
1019
const struct btf_type *t;
kernel/bpf/btf.c
1024
t = btf_type_id_size(btf, &id, NULL);
kernel/bpf/btf.c
1025
if (!t || !btf_type_is_int(t))
kernel/bpf/btf.c
1028
int_data = btf_type_int(t);
kernel/bpf/btf.c
1056
const struct btf_type *t = btf_type_by_id(btf, id);
kernel/bpf/btf.c
1058
while (btf_type_is_modifier(t) &&
kernel/bpf/btf.c
1059
BTF_INFO_KIND(t->info) != BTF_KIND_TYPEDEF) {
kernel/bpf/btf.c
1060
t = btf_type_by_id(btf, t->type);
kernel/bpf/btf.c
1063
return t;
kernel/bpf/btf.c
1086
const struct btf_type *t;
kernel/bpf/btf.c
1118
t = btf_type_by_id(show->btf, id);
kernel/bpf/btf.c
1119
if (!t)
kernel/bpf/btf.c
1146
switch (BTF_INFO_KIND(t->info)) {
kernel/bpf/btf.c
1150
t->name_off);
kernel/bpf/btf.c
1152
id = t->type;
kernel/bpf/btf.c
1157
if (!t)
kernel/bpf/btf.c
1159
array = btf_type_array(t);
kernel/bpf/btf.c
1168
id = t->type;
kernel/bpf/btf.c
1176
t = btf_type_skip_qualifiers(show->btf, id);
kernel/bpf/btf.c
1183
name = btf_name_by_offset(show->btf, t->name_off);
kernel/bpf/btf.c
1185
switch (BTF_INFO_KIND(t->info)) {
kernel/bpf/btf.c
1188
prefix = BTF_INFO_KIND(t->info) == BTF_KIND_STRUCT ?
kernel/bpf/btf.c
1363
const struct btf_type *t,
kernel/bpf/btf.c
1373
rt = btf_resolve_size(show->btf, t, &size);
kernel/bpf/btf.c
1439
const struct btf_type *t,
kernel/bpf/btf.c
1442
show->state.type = t;
kernel/bpf/btf.c
1446
return btf_show_obj_safe(show, t, data);
kernel/bpf/btf.c
1457
const struct btf_type *t,
kernel/bpf/btf.c
1460
void *safe_data = btf_show_start_type(show, t, type_id, data);
kernel/bpf/btf.c
1505
const struct btf_type *t,
kernel/bpf/btf.c
1512
return btf_show_start_aggr_type(show, t, type_id, data);
kernel/bpf/btf.c
1523
const struct btf_type *t,
kernel/bpf/btf.c
1527
return btf_show_start_aggr_type(show, t, type_id, data);
kernel/bpf/btf.c
1560
const struct btf_type *t,
kernel/bpf/btf.c
1586
btf_type_str(t),
kernel/bpf/btf.c
1587
__btf_name_by_offset(btf, t->name_off),
kernel/bpf/btf.c
1591
btf_type_ops(t)->log_details(env, t);
kernel/bpf/btf.c
1603
#define btf_verifier_log_type(env, t, ...) \
kernel/bpf/btf.c
1604
__btf_verifier_log_type((env), (t), true, __VA_ARGS__)
kernel/bpf/btf.c
1605
#define btf_verifier_log_basic(env, t, ...) \
kernel/bpf/btf.c
1606
__btf_verifier_log_type((env), (t), false, __VA_ARGS__)
kernel/bpf/btf.c
1713
static int btf_add_type(struct btf_verifier_env *env, struct btf_type *t)
kernel/bpf/btf.c
1753
btf->types[btf->nr_types++] = t;
kernel/bpf/btf.c
1998
const struct btf_type *t, u32 type_id)
kernel/bpf/btf.c
2013
v->t = t;
kernel/bpf/btf.c
2018
if (btf_type_is_ptr(t))
kernel/bpf/btf.c
2020
else if (btf_type_is_struct(t) || btf_type_is_array(t))
kernel/bpf/btf.c
2260
btf_verifier_log_basic(env, v->t, "Unsupported resolve");
kernel/bpf/btf.c
2264
static void btf_df_show(const struct btf *btf, const struct btf_type *t,
kernel/bpf/btf.c
2268
btf_show(show, "<unsupported kind:%u>", BTF_INFO_KIND(t->info));
kernel/bpf/btf.c
2366
const struct btf_type *t,
kernel/bpf/btf.c
2373
btf_verifier_log_basic(env, t,
kernel/bpf/btf.c
2379
if (btf_type_vlen(t)) {
kernel/bpf/btf.c
2380
btf_verifier_log_type(env, t, "vlen != 0");
kernel/bpf/btf.c
2384
if (btf_type_kflag(t)) {
kernel/bpf/btf.c
2385
btf_verifier_log_type(env, t, "Invalid btf_info kind_flag");
kernel/bpf/btf.c
2389
int_data = btf_type_int(t);
kernel/bpf/btf.c
2391
btf_verifier_log_basic(env, t, "Invalid int_data:%x",
kernel/bpf/btf.c
2399
btf_verifier_log_type(env, t, "nr_bits exceeds %zu",
kernel/bpf/btf.c
2404
if (BITS_ROUNDUP_BYTES(nr_bits) > t->size) {
kernel/bpf/btf.c
2405
btf_verifier_log_type(env, t, "nr_bits exceeds type_size");
kernel/bpf/btf.c
2420
btf_verifier_log_type(env, t, "Unsupported encoding");
kernel/bpf/btf.c
2424
btf_verifier_log_type(env, t, NULL);
kernel/bpf/btf.c
2430
const struct btf_type *t)
kernel/bpf/btf.c
2432
int int_data = btf_type_int(t);
kernel/bpf/btf.c
2436
t->size, BTF_INT_OFFSET(int_data),
kernel/bpf/btf.c
2533
const struct btf_type *t,
kernel/bpf/btf.c
2537
u32 int_data = btf_type_int(t);
kernel/bpf/btf.c
2551
static void btf_int_show(const struct btf *btf, const struct btf_type *t,
kernel/bpf/btf.c
2555
u32 int_data = btf_type_int(t);
kernel/bpf/btf.c
2561
safe_data = btf_show_start_type(show, t, type_id, data);
kernel/bpf/btf.c
2567
btf_int_bits_show(btf, t, safe_data, bits_offset, show);
kernel/bpf/btf.c
2614
btf_int_bits_show(btf, t, safe_data, bits_offset, show);
kernel/bpf/btf.c
2707
const struct btf_type *t,
kernel/bpf/btf.c
2712
if (btf_type_vlen(t)) {
kernel/bpf/btf.c
2713
btf_verifier_log_type(env, t, "vlen != 0");
kernel/bpf/btf.c
2717
if (btf_type_kflag(t) && !btf_type_is_type_tag(t)) {
kernel/bpf/btf.c
2718
btf_verifier_log_type(env, t, "Invalid btf_info kind_flag");
kernel/bpf/btf.c
2722
if (!BTF_TYPE_ID_VALID(t->type)) {
kernel/bpf/btf.c
2723
btf_verifier_log_type(env, t, "Invalid type_id");
kernel/bpf/btf.c
2730
if (BTF_INFO_KIND(t->info) == BTF_KIND_TYPEDEF) {
kernel/bpf/btf.c
2731
if (!t->name_off ||
kernel/bpf/btf.c
2732
!btf_name_valid_identifier(env->btf, t->name_off)) {
kernel/bpf/btf.c
2733
btf_verifier_log_type(env, t, "Invalid name");
kernel/bpf/btf.c
2736
} else if (BTF_INFO_KIND(t->info) == BTF_KIND_TYPE_TAG) {
kernel/bpf/btf.c
2737
value = btf_name_by_offset(env->btf, t->name_off);
kernel/bpf/btf.c
2739
btf_verifier_log_type(env, t, "Invalid name");
kernel/bpf/btf.c
2743
if (t->name_off) {
kernel/bpf/btf.c
2744
btf_verifier_log_type(env, t, "Invalid name");
kernel/bpf/btf.c
2749
btf_verifier_log_type(env, t, NULL);
kernel/bpf/btf.c
2757
const struct btf_type *t = v->t;
kernel/bpf/btf.c
2759
u32 next_type_id = t->type;
kernel/bpf/btf.c
2764
btf_verifier_log_type(env, v->t, "Invalid type_id");
kernel/bpf/btf.c
2786
btf_verifier_log_type(env, v->t, "Invalid type_id");
kernel/bpf/btf.c
2800
const struct btf_type *t = v->t;
kernel/bpf/btf.c
2801
u32 next_type_id = t->type;
kernel/bpf/btf.c
2806
btf_verifier_log_type(env, v->t, "Invalid type_id");
kernel/bpf/btf.c
2833
btf_verifier_log_type(env, v->t, "Invalid type_id");
kernel/bpf/btf.c
2846
const struct btf_type *t = v->t;
kernel/bpf/btf.c
2847
u32 next_type_id = t->type;
kernel/bpf/btf.c
2852
btf_verifier_log_type(env, v->t, "Invalid type_id");
kernel/bpf/btf.c
2889
btf_verifier_log_type(env, v->t, "Invalid type_id");
kernel/bpf/btf.c
289
const struct btf_type *t;
kernel/bpf/btf.c
2900
const struct btf_type *t,
kernel/bpf/btf.c
2905
t = btf_type_id_resolve(btf, &type_id);
kernel/bpf/btf.c
2907
t = btf_type_skip_modifiers(btf, type_id, NULL);
kernel/bpf/btf.c
2909
btf_type_ops(t)->show(btf, t, type_id, data, bits_offset, show);
kernel/bpf/btf.c
2912
static void btf_var_show(const struct btf *btf, const struct btf_type *t,
kernel/bpf/btf.c
2916
t = btf_type_id_resolve(btf, &type_id);
kernel/bpf/btf.c
2918
btf_type_ops(t)->show(btf, t, type_id, data, bits_offset, show);
kernel/bpf/btf.c
2921
static void btf_ptr_show(const struct btf *btf, const struct btf_type *t,
kernel/bpf/btf.c
2927
safe_data = btf_show_start_type(show, t, type_id, data);
kernel/bpf/btf.c
2940
const struct btf_type *t)
kernel/bpf/btf.c
2942
btf_verifier_log(env, "type_id=%u", t->type);
kernel/bpf/btf.c
2964
const struct btf_type *t,
kernel/bpf/btf.c
2967
if (btf_type_vlen(t)) {
kernel/bpf/btf.c
2968
btf_verifier_log_type(env, t, "vlen != 0");
kernel/bpf/btf.c
2972
if (t->type) {
kernel/bpf/btf.c
2973
btf_verifier_log_type(env, t, "type != 0");
kernel/bpf/btf.c
2978
if (!t->name_off ||
kernel/bpf/btf.c
2979
!btf_name_valid_identifier(env->btf, t->name_off)) {
kernel/bpf/btf.c
2980
btf_verifier_log_type(env, t, "Invalid name");
kernel/bpf/btf.c
2984
btf_verifier_log_type(env, t, NULL);
kernel/bpf/btf.c
2990
const struct btf_type *t)
kernel/bpf/btf.c
2992
btf_verifier_log(env, "%s", btf_type_kflag(t) ? "union" : "struct");
kernel/bpf/btf.c
3034
const struct btf_type *t,
kernel/bpf/btf.c
3037
const struct btf_array *array = btf_type_array(t);
kernel/bpf/btf.c
3041
btf_verifier_log_basic(env, t,
kernel/bpf/btf.c
3048
if (t->name_off) {
kernel/bpf/btf.c
3049
btf_verifier_log_type(env, t, "Invalid name");
kernel/bpf/btf.c
3053
if (btf_type_vlen(t)) {
kernel/bpf/btf.c
3054
btf_verifier_log_type(env, t, "vlen != 0");
kernel/bpf/btf.c
3058
if (btf_type_kflag(t)) {
kernel/bpf/btf.c
3059
btf_verifier_log_type(env, t, "Invalid btf_info kind_flag");
kernel/bpf/btf.c
3063
if (t->size) {
kernel/bpf/btf.c
3064
btf_verifier_log_type(env, t, "size != 0");
kernel/bpf/btf.c
3072
btf_verifier_log_type(env, t, "Invalid elem");
kernel/bpf/btf.c
3077
btf_verifier_log_type(env, t, "Invalid index");
kernel/bpf/btf.c
3081
btf_verifier_log_type(env, t, NULL);
kernel/bpf/btf.c
3089
const struct btf_array *array = btf_type_array(v->t);
kernel/bpf/btf.c
3100
btf_verifier_log_type(env, v->t, "Invalid index");
kernel/bpf/btf.c
3111
btf_verifier_log_type(env, v->t, "Invalid index");
kernel/bpf/btf.c
3120
btf_verifier_log_type(env, v->t,
kernel/bpf/btf.c
3131
btf_verifier_log_type(env, v->t, "Invalid elem");
kernel/bpf/btf.c
3136
btf_verifier_log_type(env, v->t, "Invalid array of int");
kernel/bpf/btf.c
3141
btf_verifier_log_type(env, v->t,
kernel/bpf/btf.c
3152
const struct btf_type *t)
kernel/bpf/btf.c
3154
const struct btf_array *array = btf_type_array(t);
kernel/bpf/btf.c
3160
static void __btf_array_show(const struct btf *btf, const struct btf_type *t,
kernel/bpf/btf.c
3164
const struct btf_array *array = btf_type_array(t);
kernel/bpf/btf.c
3189
if (!btf_show_start_array_type(show, t, type_id, encoding, data))
kernel/bpf/btf.c
3213
static void btf_array_show(const struct btf *btf, const struct btf_type *t,
kernel/bpf/btf.c
3229
__btf_array_show(btf, t, type_id, data, bits_offset, show);
kernel/bpf/btf.c
3243
__btf_array_show(btf, t, type_id, data, bits_offset, show);
kernel/bpf/btf.c
3281
const struct btf_type *t,
kernel/bpf/btf.c
3284
bool is_union = BTF_INFO_KIND(t->info) == BTF_KIND_UNION;
kernel/bpf/btf.c
3288
u32 struct_size = t->size;
kernel/bpf/btf.c
3292
meta_needed = btf_type_vlen(t) * sizeof(*member);
kernel/bpf/btf.c
3294
btf_verifier_log_basic(env, t,
kernel/bpf/btf.c
3301
if (t->name_off &&
kernel/bpf/btf.c
3302
!btf_name_valid_identifier(env->btf, t->name_off)) {
kernel/bpf/btf.c
3303
btf_verifier_log_type(env, t, "Invalid name");
kernel/bpf/btf.c
3307
btf_verifier_log_type(env, t, NULL);
kernel/bpf/btf.c
3310
for_each_member(i, t, member) {
kernel/bpf/btf.c
3312
btf_verifier_log_member(env, t, member,
kernel/bpf/btf.c
3321
btf_verifier_log_member(env, t, member, "Invalid name");
kernel/bpf/btf.c
3326
btf_verifier_log_member(env, t, member,
kernel/bpf/btf.c
3331
offset = __btf_member_bit_offset(t, member);
kernel/bpf/btf.c
3333
btf_verifier_log_member(env, t, member,
kernel/bpf/btf.c
3343
btf_verifier_log_member(env, t, member,
kernel/bpf/btf.c
3349
btf_verifier_log_member(env, t, member,
kernel/bpf/btf.c
3354
btf_verifier_log_member(env, t, member, NULL);
kernel/bpf/btf.c
3377
last_member = btf_type_member(v->t) + v->next_member - 1;
kernel/bpf/btf.c
3385
if (btf_type_kflag(v->t))
kernel/bpf/btf.c
3386
err = btf_type_ops(last_member_type)->check_kflag_member(env, v->t,
kernel/bpf/btf.c
3390
err = btf_type_ops(last_member_type)->check_member(env, v->t,
kernel/bpf/btf.c
3397
for_each_member_from(i, v->next_member, v->t, member) {
kernel/bpf/btf.c
3404
btf_verifier_log_member(env, v->t, member,
kernel/bpf/btf.c
3415
if (btf_type_kflag(v->t))
kernel/bpf/btf.c
3416
err = btf_type_ops(member_type)->check_kflag_member(env, v->t,
kernel/bpf/btf.c
3420
err = btf_type_ops(member_type)->check_member(env, v->t,
kernel/bpf/btf.c
3433
const struct btf_type *t)
kernel/bpf/btf.c
3435
btf_verifier_log(env, "size=%u vlen=%u", t->size, btf_type_vlen(t));
kernel/bpf/btf.c
3457
static int btf_find_struct(const struct btf *btf, const struct btf_type *t,
kernel/bpf/btf.c
3461
if (!__btf_type_is_struct(t))
kernel/bpf/btf.c
3463
if (t->size != sz)
kernel/bpf/btf.c
3470
static int btf_find_kptr(const struct btf *btf, const struct btf_type *t,
kernel/bpf/btf.c
3479
if (btf_type_is_volatile(t))
kernel/bpf/btf.c
3480
t = btf_type_by_id(btf, t->type);
kernel/bpf/btf.c
3482
if (!btf_type_is_ptr(t))
kernel/bpf/btf.c
3484
t = btf_type_by_id(btf, t->type);
kernel/bpf/btf.c
3485
is_type_tag = btf_type_is_type_tag(t) && !btf_type_kflag(t);
kernel/bpf/btf.c
3489
if (btf_type_is_type_tag(btf_type_by_id(btf, t->type)))
kernel/bpf/btf.c
349
const char *btf_type_str(const struct btf_type *t)
kernel/bpf/btf.c
3491
tag_value = __btf_name_by_offset(btf, t->name_off);
kernel/bpf/btf.c
3507
t = btf_type_skip_modifiers(btf, t->type, &res_id);
kernel/bpf/btf.c
3509
if (!__btf_type_is_struct(t))
kernel/bpf/btf.c
351
return btf_kind_str[BTF_INFO_KIND(t->info)];
kernel/bpf/btf.c
3525
const struct btf_type *t = btf_type_by_id(btf, i);
kernel/bpf/btf.c
3527
if (!btf_type_is_decl_tag(t))
kernel/bpf/btf.c
3529
if (pt != btf_type_by_id(btf, t->type))
kernel/bpf/btf.c
3531
if (btf_type_decl_tag(t)->component_idx != comp_idx)
kernel/bpf/btf.c
3533
if (strncmp(__btf_name_by_offset(btf, t->name_off), tag_key, len))
kernel/bpf/btf.c
3544
const struct btf_type *t;
kernel/bpf/btf.c
3552
t = btf_type_by_id(btf, id);
kernel/bpf/btf.c
3554
value = __btf_name_by_offset(btf, t->name_off) + len;
kernel/bpf/btf.c
3566
const struct btf_type *t, int comp_idx, u32 off,
kernel/bpf/btf.c
3574
if (!__btf_type_is_struct(t))
kernel/bpf/btf.c
3576
if (t->size != sz)
kernel/bpf/btf.c
3700
const struct btf_type *t, u32 field_mask,
kernel/bpf/btf.c
3710
static int btf_find_nested_struct(const struct btf *btf, const struct btf_type *t,
kernel/bpf/btf.c
3721
ret = btf_find_struct_field(btf, t, field_mask, info, info_cnt, level);
kernel/bpf/btf.c
3733
err = btf_repeat_fields(info, info_cnt, ret, nelems - 1, t->size);
kernel/bpf/btf.c
3841
const struct btf_type *t, u32 field_mask,
kernel/bpf/btf.c
3849
for_each_member(i, t, member) {
kernel/bpf/btf.c
3853
off = __btf_member_bit_offset(t, member);
kernel/bpf/btf.c
3859
ret = btf_find_field_one(btf, t, member_type, i,
kernel/bpf/btf.c
3870
static int btf_find_datasec_var(const struct btf *btf, const struct btf_type *t,
kernel/bpf/btf.c
3878
for_each_vsi(i, t, vsi) {
kernel/bpf/btf.c
3894
static int btf_find_field(const struct btf *btf, const struct btf_type *t,
kernel/bpf/btf.c
3898
if (__btf_type_is_struct(t))
kernel/bpf/btf.c
3899
return btf_find_struct_field(btf, t, field_mask, info, info_cnt, 0);
kernel/bpf/btf.c
3900
else if (btf_type_is_datasec(t))
kernel/bpf/btf.c
3901
return btf_find_datasec_var(btf, t, field_mask, info, info_cnt, 0);
kernel/bpf/btf.c
3910
const struct btf_type *t;
kernel/bpf/btf.c
3921
t = btf_type_by_id(btf, info->kptr.type_id);
kernel/bpf/btf.c
3922
id = bpf_find_btf_id(__btf_name_by_offset(btf, t->name_off), BTF_INFO_KIND(t->info),
kernel/bpf/btf.c
4002
const struct btf_type *t, *n = NULL;
kernel/bpf/btf.c
4007
t = btf_type_by_id(btf, info->graph_root.value_btf_id);
kernel/bpf/btf.c
4012
for_each_member(i, t, member) {
kernel/bpf/btf.c
4066
struct btf_record *btf_parse_fields(const struct btf *btf, const struct btf_type *t,
kernel/bpf/btf.c
4074
ret = btf_find_field(btf, t, field_mask, info_arr, ARRAY_SIZE(info_arr));
kernel/bpf/btf.c
4215
const struct btf_type *t;
kernel/bpf/btf.c
4224
t = btf_type_by_id(rec->fields[i].kptr.btf,
kernel/bpf/btf.c
4226
if (!t->size)
kernel/bpf/btf.c
4228
if (t->size > PAGE_SIZE)
kernel/bpf/btf.c
4287
static void __btf_struct_show(const struct btf *btf, const struct btf_type *t,
kernel/bpf/btf.c
4295
safe_data = btf_show_start_struct_type(show, t, type_id, data);
kernel/bpf/btf.c
4299
for_each_member(i, t, member) {
kernel/bpf/btf.c
4309
member_offset = __btf_member_bit_offset(t, member);
kernel/bpf/btf.c
4310
bitfield_size = __btf_member_bitfield_size(t, member);
kernel/bpf/btf.c
4334
static void btf_struct_show(const struct btf *btf, const struct btf_type *t,
kernel/bpf/btf.c
4350
__btf_struct_show(btf, t, type_id, data, bits_offset, show);
kernel/bpf/btf.c
4365
__btf_struct_show(btf, t, type_id, data, bits_offset, show);
kernel/bpf/btf.c
4438
const struct btf_type *t,
kernel/bpf/btf.c
4441
const struct btf_enum *enums = btf_type_enum(t);
kernel/bpf/btf.c
4447
nr_enums = btf_type_vlen(t);
kernel/bpf/btf.c
4451
btf_verifier_log_basic(env, t,
kernel/bpf/btf.c
4457
if (t->size > 8 || !is_power_of_2(t->size)) {
kernel/bpf/btf.c
4458
btf_verifier_log_type(env, t, "Unexpected size");
kernel/bpf/btf.c
4463
if (t->name_off &&
kernel/bpf/btf.c
4464
!btf_name_valid_identifier(env->btf, t->name_off)) {
kernel/bpf/btf.c
4465
btf_verifier_log_type(env, t, "Invalid name");
kernel/bpf/btf.c
4469
btf_verifier_log_type(env, t, NULL);
kernel/bpf/btf.c
447
const struct btf_type *t,
kernel/bpf/btf.c
4481
btf_verifier_log_type(env, t, "Invalid name");
kernel/bpf/btf.c
4487
fmt_str = btf_type_kflag(t) ? "\t%s val=%d\n" : "\t%s val=%u\n";
kernel/bpf/btf.c
4497
const struct btf_type *t)
kernel/bpf/btf.c
4499
btf_verifier_log(env, "size=%u vlen=%u", t->size, btf_type_vlen(t));
kernel/bpf/btf.c
4502
static void btf_enum_show(const struct btf *btf, const struct btf_type *t,
kernel/bpf/btf.c
4506
const struct btf_enum *enums = btf_type_enum(t);
kernel/bpf/btf.c
4507
u32 i, nr_enums = btf_type_vlen(t);
kernel/bpf/btf.c
4511
safe_data = btf_show_start_type(show, t, type_id, data);
kernel/bpf/btf.c
4529
if (btf_type_kflag(t))
kernel/bpf/btf.c
4546
const struct btf_type *t,
kernel/bpf/btf.c
4549
const struct btf_enum64 *enums = btf_type_enum64(t);
kernel/bpf/btf.c
4555
nr_enums = btf_type_vlen(t);
kernel/bpf/btf.c
4559
btf_verifier_log_basic(env, t,
kernel/bpf/btf.c
4565
if (t->size > 8 || !is_power_of_2(t->size)) {
kernel/bpf/btf.c
4566
btf_verifier_log_type(env, t, "Unexpected size");
kernel/bpf/btf.c
4571
if (t->name_off &&
kernel/bpf/btf.c
4572
!btf_name_valid_identifier(env->btf, t->name_off)) {
kernel/bpf/btf.c
4573
btf_verifier_log_type(env, t, "Invalid name");
kernel/bpf/btf.c
4577
btf_verifier_log_type(env, t, NULL);
kernel/bpf/btf.c
4589
btf_verifier_log_type(env, t, "Invalid name");
kernel/bpf/btf.c
4596
fmt_str = btf_type_kflag(t) ? "\t%s val=%lld\n" : "\t%s val=%llu\n";
kernel/bpf/btf.c
460
const struct btf_type *t);
kernel/bpf/btf.c
4605
static void btf_enum64_show(const struct btf *btf, const struct btf_type *t,
kernel/bpf/btf.c
4609
const struct btf_enum64 *enums = btf_type_enum64(t);
kernel/bpf/btf.c
461
void (*show)(const struct btf *btf, const struct btf_type *t,
kernel/bpf/btf.c
4610
u32 i, nr_enums = btf_type_vlen(t);
kernel/bpf/btf.c
4614
safe_data = btf_show_start_type(show, t, type_id, data);
kernel/bpf/btf.c
4632
if (btf_type_kflag(t))
kernel/bpf/btf.c
4649
const struct btf_type *t,
kernel/bpf/btf.c
4652
u32 meta_needed = btf_type_vlen(t) * sizeof(struct btf_param);
kernel/bpf/btf.c
4655
btf_verifier_log_basic(env, t,
kernel/bpf/btf.c
4661
if (t->name_off) {
kernel/bpf/btf.c
4662
btf_verifier_log_type(env, t, "Invalid name");
kernel/bpf/btf.c
4666
if (btf_type_kflag(t)) {
kernel/bpf/btf.c
4667
btf_verifier_log_type(env, t, "Invalid btf_info kind_flag");
kernel/bpf/btf.c
4671
btf_verifier_log_type(env, t, NULL);
kernel/bpf/btf.c
4677
const struct btf_type *t)
kernel/bpf/btf.c
4679
const struct btf_param *args = (const struct btf_param *)(t + 1);
kernel/bpf/btf.c
4680
u16 nr_args = btf_type_vlen(t), i;
kernel/bpf/btf.c
4682
btf_verifier_log(env, "return=%u args=(", t->type);
kernel/bpf/btf.c
470
const struct btf_type *t, u32 type_id);
kernel/bpf/btf.c
473
const struct btf_type *t);
kernel/bpf/btf.c
4736
const struct btf_type *t,
kernel/bpf/btf.c
4739
if (!t->name_off ||
kernel/bpf/btf.c
4740
!btf_name_valid_identifier(env->btf, t->name_off)) {
kernel/bpf/btf.c
4741
btf_verifier_log_type(env, t, "Invalid name");
kernel/bpf/btf.c
4745
if (btf_type_vlen(t) > BTF_FUNC_GLOBAL) {
kernel/bpf/btf.c
4746
btf_verifier_log_type(env, t, "Invalid func linkage");
kernel/bpf/btf.c
475
static bool btf_type_is_modifier(const struct btf_type *t)
kernel/bpf/btf.c
4750
if (btf_type_kflag(t)) {
kernel/bpf/btf.c
4751
btf_verifier_log_type(env, t, "Invalid btf_info kind_flag");
kernel/bpf/btf.c
4755
btf_verifier_log_type(env, t, NULL);
kernel/bpf/btf.c
4763
const struct btf_type *t = v->t;
kernel/bpf/btf.c
4764
u32 next_type_id = t->type;
kernel/bpf/btf.c
4767
err = btf_func_check(env, t);
kernel/bpf/btf.c
4785
const struct btf_type *t,
kernel/bpf/btf.c
4792
btf_verifier_log_basic(env, t,
kernel/bpf/btf.c
4798
if (btf_type_vlen(t)) {
kernel/bpf/btf.c
4799
btf_verifier_log_type(env, t, "vlen != 0");
kernel/bpf/btf.c
4803
if (btf_type_kflag(t)) {
kernel/bpf/btf.c
4804
btf_verifier_log_type(env, t, "Invalid btf_info kind_flag");
kernel/bpf/btf.c
4808
if (!t->name_off ||
kernel/bpf/btf.c
4809
!btf_name_valid_identifier(env->btf, t->name_off)) {
kernel/bpf/btf.c
4810
btf_verifier_log_type(env, t, "Invalid name");
kernel/bpf/btf.c
4815
if (!t->type || !BTF_TYPE_ID_VALID(t->type)) {
kernel/bpf/btf.c
4816
btf_verifier_log_type(env, t, "Invalid type_id");
kernel/bpf/btf.c
4820
var = btf_type_var(t);
kernel/bpf/btf.c
4823
btf_verifier_log_type(env, t, "Linkage not supported");
kernel/bpf/btf.c
4827
btf_verifier_log_type(env, t, NULL);
kernel/bpf/btf.c
4832
static void btf_var_log(struct btf_verifier_env *env, const struct btf_type *t)
kernel/bpf/btf.c
4834
const struct btf_var *var = btf_type_var(t);
kernel/bpf/btf.c
4836
btf_verifier_log(env, "type_id=%u linkage=%u", t->type, var->linkage);
kernel/bpf/btf.c
4849
const struct btf_type *t,
kernel/bpf/btf.c
4856
meta_needed = btf_type_vlen(t) * sizeof(*vsi);
kernel/bpf/btf.c
4858
btf_verifier_log_basic(env, t,
kernel/bpf/btf.c
4864
if (!t->size) {
kernel/bpf/btf.c
4865
btf_verifier_log_type(env, t, "size == 0");
kernel/bpf/btf.c
4869
if (btf_type_kflag(t)) {
kernel/bpf/btf.c
487
switch (BTF_INFO_KIND(t->info)) {
kernel/bpf/btf.c
4870
btf_verifier_log_type(env, t, "Invalid btf_info kind_flag");
kernel/bpf/btf.c
4874
if (!t->name_off ||
kernel/bpf/btf.c
4875
!btf_name_valid_section(env->btf, t->name_off)) {
kernel/bpf/btf.c
4876
btf_verifier_log_type(env, t, "Invalid name");
kernel/bpf/btf.c
4880
btf_verifier_log_type(env, t, NULL);
kernel/bpf/btf.c
4882
for_each_vsi(i, t, vsi) {
kernel/bpf/btf.c
4885
btf_verifier_log_vsi(env, t, vsi,
kernel/bpf/btf.c
4890
if (vsi->offset < last_vsi_end_off || vsi->offset >= t->size) {
kernel/bpf/btf.c
4891
btf_verifier_log_vsi(env, t, vsi,
kernel/bpf/btf.c
4896
if (!vsi->size || vsi->size > t->size) {
kernel/bpf/btf.c
4897
btf_verifier_log_vsi(env, t, vsi,
kernel/bpf/btf.c
4903
if (last_vsi_end_off > t->size) {
kernel/bpf/btf.c
4904
btf_verifier_log_vsi(env, t, vsi,
kernel/bpf/btf.c
4909
btf_verifier_log_vsi(env, t, vsi, NULL);
kernel/bpf/btf.c
4913
if (t->size < sum) {
kernel/bpf/btf.c
4914
btf_verifier_log_type(env, t, "Invalid btf_info size");
kernel/bpf/btf.c
4929
for_each_vsi_from(i, v->next_member, v->t, vsi) {
kernel/bpf/btf.c
4934
btf_verifier_log_vsi(env, v->t, vsi,
kernel/bpf/btf.c
4947
btf_verifier_log_vsi(env, v->t, vsi, "Invalid type");
kernel/bpf/btf.c
4952
btf_verifier_log_vsi(env, v->t, vsi, "Invalid size");
kernel/bpf/btf.c
4962
const struct btf_type *t)
kernel/bpf/btf.c
4964
btf_verifier_log(env, "size=%u vlen=%u", t->size, btf_type_vlen(t));
kernel/bpf/btf.c
4968
const struct btf_type *t, u32 type_id,
kernel/bpf/btf.c
4976
if (!btf_show_start_type(show, t, type_id, data))
kernel/bpf/btf.c
4980
__btf_name_by_offset(btf, t->name_off));
kernel/bpf/btf.c
4981
for_each_vsi(i, t, vsi) {
kernel/bpf/btf.c
5001
const struct btf_type *t,
kernel/bpf/btf.c
5004
if (btf_type_vlen(t)) {
kernel/bpf/btf.c
5005
btf_verifier_log_type(env, t, "vlen != 0");
kernel/bpf/btf.c
5009
if (btf_type_kflag(t)) {
kernel/bpf/btf.c
5010
btf_verifier_log_type(env, t, "Invalid btf_info kind_flag");
kernel/bpf/btf.c
5014
if (t->size != 2 && t->size != 4 && t->size != 8 && t->size != 12 &&
kernel/bpf/btf.c
5015
t->size != 16) {
kernel/bpf/btf.c
5016
btf_verifier_log_type(env, t, "Invalid type_size");
kernel/bpf/btf.c
5020
btf_verifier_log_type(env, t, NULL);
kernel/bpf/btf.c
504
bool btf_type_is_void(const struct btf_type *t)
kernel/bpf/btf.c
506
return t == &btf_void;
kernel/bpf/btf.c
5061
const struct btf_type *t)
kernel/bpf/btf.c
5063
btf_verifier_log(env, "size=%u", t->size);
kernel/bpf/btf.c
5076
const struct btf_type *t,
kernel/bpf/btf.c
5085
btf_verifier_log_basic(env, t,
kernel/bpf/btf.c
509
static bool btf_type_is_datasec(const struct btf_type *t)
kernel/bpf/btf.c
5091
value = btf_name_by_offset(env->btf, t->name_off);
kernel/bpf/btf.c
5093
btf_verifier_log_type(env, t, "Invalid value");
kernel/bpf/btf.c
5097
if (btf_type_vlen(t)) {
kernel/bpf/btf.c
5098
btf_verifier_log_type(env, t, "vlen != 0");
kernel/bpf/btf.c
5102
component_idx = btf_type_decl_tag(t)->component_idx;
kernel/bpf/btf.c
5104
btf_verifier_log_type(env, t, "Invalid component_idx");
kernel/bpf/btf.c
5108
btf_verifier_log_type(env, t, NULL);
kernel/bpf/btf.c
511
return BTF_INFO_KIND(t->info) == BTF_KIND_DATASEC;
kernel/bpf/btf.c
5117
const struct btf_type *t = v->t;
kernel/bpf/btf.c
5118
u32 next_type_id = t->type;
kernel/bpf/btf.c
5125
btf_verifier_log_type(env, v->t, "Invalid type_id");
kernel/bpf/btf.c
5133
component_idx = btf_type_decl_tag(t)->component_idx;
kernel/bpf/btf.c
5136
btf_verifier_log_type(env, v->t, "Invalid component_idx");
kernel/bpf/btf.c
514
static bool btf_type_is_decl_tag(const struct btf_type *t)
kernel/bpf/btf.c
5149
btf_verifier_log_type(env, v->t, "Invalid component_idx");
kernel/bpf/btf.c
5159
static void btf_decl_tag_log(struct btf_verifier_env *env, const struct btf_type *t)
kernel/bpf/btf.c
516
return BTF_INFO_KIND(t->info) == BTF_KIND_DECL_TAG;
kernel/bpf/btf.c
5161
btf_verifier_log(env, "type=%u component_idx=%d", t->type,
kernel/bpf/btf.c
5162
btf_type_decl_tag(t)->component_idx);
kernel/bpf/btf.c
5175
const struct btf_type *t)
kernel/bpf/btf.c
5184
args = (const struct btf_param *)(t + 1);
kernel/bpf/btf.c
5185
nr_args = btf_type_vlen(t);
kernel/bpf/btf.c
5188
if (t->type) {
kernel/bpf/btf.c
5189
u32 ret_type_id = t->type;
kernel/bpf/btf.c
519
static bool btf_type_nosize(const struct btf_type *t)
kernel/bpf/btf.c
5193
btf_verifier_log_type(env, t, "Invalid return type");
kernel/bpf/btf.c
5198
btf_verifier_log_type(env, t, "Invalid return type");
kernel/bpf/btf.c
521
return btf_type_is_void(t) || btf_type_is_fwd(t) ||
kernel/bpf/btf.c
5211
btf_verifier_log_type(env, t, "Invalid return type");
kernel/bpf/btf.c
522
btf_type_is_func(t) || btf_type_is_func_proto(t) ||
kernel/bpf/btf.c
5222
btf_verifier_log_type(env, t, "Invalid arg#%u",
kernel/bpf/btf.c
523
btf_type_is_decl_tag(t);
kernel/bpf/btf.c
5236
btf_verifier_log_type(env, t, "Invalid arg#%u", i + 1);
kernel/bpf/btf.c
5241
btf_verifier_log_type(env, t, "Invalid arg#%u", i + 1);
kernel/bpf/btf.c
5248
btf_verifier_log_type(env, t,
kernel/bpf/btf.c
526
static bool btf_type_nosize_or_null(const struct btf_type *t)
kernel/bpf/btf.c
5261
btf_verifier_log_type(env, t, "Invalid arg#%u", i + 1);
kernel/bpf/btf.c
5270
const struct btf_type *t)
kernel/bpf/btf.c
5278
proto_type = btf_type_by_id(btf, t->type);
kernel/bpf/btf.c
528
return !t || btf_type_nosize(t);
kernel/bpf/btf.c
5281
btf_verifier_log_type(env, t, "Invalid type_id");
kernel/bpf/btf.c
5289
btf_verifier_log_type(env, t, "Invalid arg#%u", i + 1);
kernel/bpf/btf.c
531
static bool btf_type_is_decl_tag_target(const struct btf_type *t)
kernel/bpf/btf.c
5320
const struct btf_type *t,
kernel/bpf/btf.c
5326
if (meta_left < sizeof(*t)) {
kernel/bpf/btf.c
5328
env->log_type_id, meta_left, sizeof(*t));
kernel/bpf/btf.c
533
return btf_type_is_func(t) || btf_type_is_struct(t) ||
kernel/bpf/btf.c
5331
meta_left -= sizeof(*t);
kernel/bpf/btf.c
5333
if (t->info & ~BTF_INFO_MASK) {
kernel/bpf/btf.c
5335
env->log_type_id, t->info);
kernel/bpf/btf.c
5339
if (BTF_INFO_KIND(t->info) > BTF_KIND_MAX ||
kernel/bpf/btf.c
534
btf_type_is_var(t) || btf_type_is_typedef(t);
kernel/bpf/btf.c
5340
BTF_INFO_KIND(t->info) == BTF_KIND_UNKN) {
kernel/bpf/btf.c
5342
env->log_type_id, BTF_INFO_KIND(t->info));
kernel/bpf/btf.c
5346
if (!btf_name_offset_valid(env->btf, t->name_off)) {
kernel/bpf/btf.c
5348
env->log_type_id, t->name_off);
kernel/bpf/btf.c
5352
var_meta_size = btf_type_ops(t)->check_meta(env, t, meta_left);
kernel/bpf/btf.c
5373
struct btf_type *t = cur;
kernel/bpf/btf.c
5376
meta_size = btf_check_meta(env, t, end - cur);
kernel/bpf/btf.c
5380
btf_add_type(env, t);
kernel/bpf/btf.c
5389
const struct btf_type *t,
kernel/bpf/btf.c
5397
if (btf_type_is_struct(t) || btf_type_is_datasec(t))
kernel/bpf/btf.c
5401
if (btf_type_is_decl_tag(t) || btf_type_is_func(t))
kernel/bpf/btf.c
5405
if (btf_type_is_modifier(t) || btf_type_is_ptr(t) ||
kernel/bpf/btf.c
5406
btf_type_is_var(t)) {
kernel/bpf/btf.c
5407
t = btf_type_id_resolve(btf, &type_id);
kernel/bpf/btf.c
5408
return t &&
kernel/bpf/btf.c
5409
!btf_type_is_modifier(t) &&
kernel/bpf/btf.c
5410
!btf_type_is_var(t) &&
kernel/bpf/btf.c
5411
!btf_type_is_datasec(t);
kernel/bpf/btf.c
5414
if (btf_type_is_array(t)) {
kernel/bpf/btf.c
5415
const struct btf_array *array = btf_type_array(t);
kernel/bpf/btf.c
5430
const struct btf_type *t, u32 type_id)
kernel/bpf/btf.c
5437
env_stack_push(env, t, type_id);
kernel/bpf/btf.c
5440
err = btf_type_ops(v->t)->resolve(env, v);
kernel/bpf/btf.c
5445
btf_verifier_log_type(env, t,
kernel/bpf/btf.c
5449
btf_verifier_log_type(env, t, "Loop detected");
kernel/bpf/btf.c
5453
if (!err && !btf_resolve_valid(env, t, type_id)) {
kernel/bpf/btf.c
5454
btf_verifier_log_type(env, t, "Invalid resolve state");
kernel/bpf/btf.c
5465
const struct btf_type *t;
kernel/bpf/btf.c
5476
t = btf_type_by_id(btf, type_id);
kernel/bpf/btf.c
5479
if (btf_type_needs_resolve(t) &&
kernel/bpf/btf.c
5481
err = btf_resolve(env, t, type_id);
kernel/bpf/btf.c
5486
if (btf_type_is_func_proto(t)) {
kernel/bpf/btf.c
5487
err = btf_func_proto_check(env, t);
kernel/bpf/btf.c
565
const struct btf_type *t = btf_type_by_id(btf, i);
kernel/bpf/btf.c
566
const char *n = btf_name_by_offset(btf, t->name_off);
kernel/bpf/btf.c
5730
const struct btf_type *t;
kernel/bpf/btf.c
5732
t = btf_type_by_id(btf, i);
kernel/bpf/btf.c
5733
if (!t) {
kernel/bpf/btf.c
5738
ret = btf_find_kptr(btf, t, 0, 0, &tmp, BPF_KPTR);
kernel/bpf/btf.c
5763
const struct btf_type *t;
kernel/bpf/btf.c
5766
t = btf_type_by_id(btf, i);
kernel/bpf/btf.c
5767
if (!__btf_type_is_struct(t))
kernel/bpf/btf.c
5772
for_each_member(j, t, member) {
kernel/bpf/btf.c
5791
record = btf_parse_fields(btf, t, BPF_SPIN_LOCK | BPF_RES_SPIN_LOCK | BPF_LIST_HEAD | BPF_LIST_NODE |
kernel/bpf/btf.c
5793
BPF_KPTR, t->size);
kernel/bpf/btf.c
5830
const struct btf_type *t;
kernel/bpf/btf.c
5834
t = btf_type_by_id(btf, i);
kernel/bpf/btf.c
5835
if (!t)
kernel/bpf/btf.c
5837
if (!btf_type_is_modifier(t))
kernel/bpf/btf.c
5842
in_tags = btf_type_is_type_tag(t);
kernel/bpf/btf.c
5843
while (btf_type_is_modifier(t)) {
kernel/bpf/btf.c
5848
if (btf_type_is_type_tag(t)) {
kernel/bpf/btf.c
5859
cur_id = t->type;
kernel/bpf/btf.c
5860
t = btf_type_by_id(btf, cur_id);
kernel/bpf/btf.c
5861
if (!t)
kernel/bpf/btf.c
6003
const struct btf_type *t;
kernel/bpf/btf.c
6027
conv_struct = bpf_ctx_convert.t;
kernel/bpf/btf.c
6043
conv_struct = bpf_ctx_convert.t;
kernel/bpf/btf.c
6064
const struct btf_type *t, enum bpf_prog_type prog_type,
kernel/bpf/btf.c
6070
t = btf_type_by_id(btf, t->type);
kernel/bpf/btf.c
6076
while (btf_type_is_modifier(t) && !btf_type_is_typedef(t))
kernel/bpf/btf.c
6077
t = btf_type_by_id(btf, t->type);
kernel/bpf/btf.c
6079
if (btf_type_is_typedef(t)) {
kernel/bpf/btf.c
6080
tname = btf_name_by_offset(btf, t->name_off);
kernel/bpf/btf.c
6086
while (btf_type_is_modifier(t))
kernel/bpf/btf.c
6087
t = btf_type_by_id(btf, t->type);
kernel/bpf/btf.c
6088
if (!btf_type_is_struct(t)) {
kernel/bpf/btf.c
6096
tname = btf_name_by_offset(btf, t->name_off);
kernel/bpf/btf.c
6150
const struct btf_type *t, int arg,
kernel/bpf/btf.c
6157
if (!btf_is_ptr(t)) {
kernel/bpf/btf.c
6161
t = btf_type_by_id(btf, t->type);
kernel/bpf/btf.c
6165
while (btf_type_is_modifier(t) && !btf_type_is_typedef(t))
kernel/bpf/btf.c
6166
t = btf_type_by_id(btf, t->type);
kernel/bpf/btf.c
6168
if (btf_type_is_typedef(t)) {
kernel/bpf/btf.c
6169
tname = btf_name_by_offset(btf, t->name_off);
kernel/bpf/btf.c
6176
while (btf_type_is_modifier(t))
kernel/bpf/btf.c
6177
t = btf_type_by_id(btf, t->type);
kernel/bpf/btf.c
618
const struct btf_type *t;
kernel/bpf/btf.c
6180
if (btf_type_is_void(t))
kernel/bpf/btf.c
6183
tname = btf_name_by_offset(btf, t->name_off);
kernel/bpf/btf.c
6192
if (__btf_type_is_struct(t) && strcmp(tname, "pt_regs") == 0)
kernel/bpf/btf.c
6197
__btf_type_is_struct(t) && strcmp(tname, "pt_regs") == 0)
kernel/bpf/btf.c
6200
__btf_type_is_struct(t) && strcmp(tname, "user_pt_regs") == 0)
kernel/bpf/btf.c
6203
__btf_type_is_struct(t) && strcmp(tname, "user_regs_struct") == 0)
kernel/bpf/btf.c
6209
if (btf_is_int(t) && t->size == 8)
kernel/bpf/btf.c
6216
if (__btf_type_is_struct(t) &&
kernel/bpf/btf.c
6220
if (btf_is_int(t) && t->size == 8)
kernel/bpf/btf.c
6225
if (__btf_type_is_struct(t) &&
kernel/bpf/btf.c
6234
if (btf_is_int(t) && t->size == 8)
kernel/bpf/btf.c
6244
if (btf_is_int(t) && t->size == 8)
kernel/bpf/btf.c
626
t = btf_type_by_id(btf, m);
kernel/bpf/btf.c
627
tname = btf_name_by_offset(btf, t->name_off);
kernel/bpf/btf.c
6276
if (!__btf_type_is_struct(t) || strcmp(ctx_tname, tname) != 0) {
kernel/bpf/btf.c
6286
const struct btf_type *t,
kernel/bpf/btf.c
6290
if (!btf_is_prog_ctx_type(log, btf, t, prog_type, arg))
kernel/bpf/btf.c
6302
conv_struct = bpf_ctx_convert.t;
kernel/bpf/btf.c
6388
bpf_ctx_convert.t = btf_type_by_id(btf, bpf_ctx_convert_btf_id[0]);
kernel/bpf/btf.c
643
const struct btf_type *t;
kernel/bpf/btf.c
6520
static bool is_void_or_int_ptr(struct btf *btf, const struct btf_type *t)
kernel/bpf/btf.c
6523
t = btf_type_skip_modifiers(btf, t->type, NULL);
kernel/bpf/btf.c
6524
return btf_type_is_void(t) || btf_type_is_int(t);
kernel/bpf/btf.c
6531
const struct btf_type *t;
kernel/bpf/btf.c
6541
t = btf_type_skip_modifiers(btf, args[i].type, NULL);
kernel/bpf/btf.c
6542
offset += btf_type_is_ptr(t) ? 8 : roundup(t->size, 8);
kernel/bpf/btf.c
6547
t = btf_type_skip_modifiers(btf, func_proto->type, NULL);
kernel/bpf/btf.c
6548
offset += btf_type_is_ptr(t) ? 8 : roundup(t->size, 8);
kernel/bpf/btf.c
657
t = btf_type_by_id(btf, id);
kernel/bpf/btf.c
6575
const struct btf_type *t;
kernel/bpf/btf.c
658
tname = btf_name_by_offset(btf, t->name_off);
kernel/bpf/btf.c
6581
t = btf_type_by_id(btf, args[i].type);
kernel/bpf/btf.c
6582
t = btf_resolve_size(btf, t, &sz);
kernel/bpf/btf.c
6583
if (IS_ERR(t))
kernel/bpf/btf.c
6584
return PTR_ERR(t);
kernel/bpf/btf.c
661
if (BTF_INFO_KIND(t->info) == kind)
kernel/bpf/btf.c
666
t = btf_type_by_id(btf, id);
kernel/bpf/btf.c
667
if (BTF_INFO_KIND(t->info) != kind)
kernel/bpf/btf.c
669
tname = btf_name_by_offset(btf, t->name_off);
kernel/bpf/btf.c
6792
const struct btf_type *t = prog->aux->attach_func_proto;
kernel/bpf/btf.c
6808
arg = btf_ctx_arg_idx(btf, t, off);
kernel/bpf/btf.c
6809
args = (const struct btf_param *)(t + 1);
kernel/bpf/btf.c
6813
nr_args = t ? btf_type_vlen(t) : MAX_BPF_FUNC_REG_ARGS;
kernel/bpf/btf.c
6848
if (!t)
kernel/bpf/btf.c
6850
t = btf_type_by_id(btf, t->type);
kernel/bpf/btf.c
6856
if (!t)
kernel/bpf/btf.c
6859
t = btf_type_skip_modifiers(btf, t->type, NULL);
kernel/bpf/btf.c
6860
if (!btf_type_is_small_int(t)) {
kernel/bpf/btf.c
6863
btf_type_str(t));
kernel/bpf/btf.c
6873
if (!t)
kernel/bpf/btf.c
6876
t = btf_type_by_id(btf, args[arg].type);
kernel/bpf/btf.c
6880
while (btf_type_is_modifier(t))
kernel/bpf/btf.c
6881
t = btf_type_by_id(btf, t->type);
kernel/bpf/btf.c
6882
if (btf_type_is_small_int(t) || btf_is_any_enum(t) || btf_type_is_struct(t))
kernel/bpf/btf.c
6885
if (!btf_type_is_ptr(t)) {
kernel/bpf/btf.c
6889
__btf_name_by_offset(btf, t->name_off),
kernel/bpf/btf.c
6890
btf_type_str(t));
kernel/bpf/btf.c
6918
if (is_void_or_int_ptr(btf, t))
kernel/bpf/btf.c
6948
const struct btf_type *t;
kernel/bpf/btf.c
6952
t = btf_type_by_id(btf, prog->aux->attach_btf_id);
kernel/bpf/btf.c
6953
if (!t)
kernel/bpf/btf.c
6955
tname = btf_name_by_offset(btf, t->name_off);
kernel/bpf/btf.c
6987
ret = btf_translate_to_vmlinux(log, btf, t, tgt_type, arg);
kernel/bpf/btf.c
6998
info->btf_id = t->type;
kernel/bpf/btf.c
6999
t = btf_type_by_id(btf, t->type);
kernel/bpf/btf.c
7001
if (btf_type_is_type_tag(t) && !btf_type_kflag(t)) {
kernel/bpf/btf.c
7002
tag_value = __btf_name_by_offset(btf, t->name_off);
kernel/bpf/btf.c
7010
while (btf_type_is_modifier(t)) {
kernel/bpf/btf.c
7011
info->btf_id = t->type;
kernel/bpf/btf.c
7012
t = btf_type_by_id(btf, t->type);
kernel/bpf/btf.c
7014
if (!btf_type_is_struct(t)) {
kernel/bpf/btf.c
7017
tname, arg, btf_type_str(t));
kernel/bpf/btf.c
7021
tname, arg, info->btf_id, btf_type_str(t),
kernel/bpf/btf.c
7022
__btf_name_by_offset(btf, t->name_off));
kernel/bpf/btf.c
7045
const struct btf_type *t, int off, int size,
kernel/bpf/btf.c
7056
if (btf_type_is_modifier(t))
kernel/bpf/btf.c
7057
t = btf_type_skip_modifiers(btf, t->type, NULL);
kernel/bpf/btf.c
7058
tname = __btf_name_by_offset(btf, t->name_off);
kernel/bpf/btf.c
7059
if (!btf_type_is_struct(t)) {
kernel/bpf/btf.c
7064
vlen = btf_type_vlen(t);
kernel/bpf/btf.c
7065
if (BTF_INFO_KIND(t->info) == BTF_KIND_UNION && vlen != 1 && !(*flag & PTR_UNTRUSTED))
kernel/bpf/btf.c
7073
if (off + size > t->size) {
kernel/bpf/btf.c
7082
member = btf_type_member(t) + vlen - 1;
kernel/bpf/btf.c
7092
moff = __btf_member_bit_offset(t, member) / 8;
kernel/bpf/btf.c
7097
t = btf_type_skip_modifiers(btf, array_elem->type,
kernel/bpf/btf.c
7100
if (btf_type_is_int(t))
kernel/bpf/btf.c
7103
if (!btf_type_is_struct(t))
kernel/bpf/btf.c
7106
off = (off - moff) % t->size;
kernel/bpf/btf.c
7115
for_each_member(i, t, member) {
kernel/bpf/btf.c
7117
moff = __btf_member_bit_offset(t, member) / 8;
kernel/bpf/btf.c
7122
if (__btf_member_bitfield_size(t, member)) {
kernel/bpf/btf.c
7123
u32 end_bit = __btf_member_bit_offset(t, member) +
kernel/bpf/btf.c
7124
__btf_member_bitfield_size(t, member);
kernel/bpf/btf.c
7234
t = mtype;
kernel/bpf/btf.c
7248
const struct btf_type *stype, *t;
kernel/bpf/btf.c
726
const struct btf_type *t = btf_type_by_id(btf, id);
kernel/bpf/btf.c
7260
t = btf_type_by_id(btf, mtype->type);
kernel/bpf/btf.c
7261
if (btf_type_is_type_tag(t) && !btf_type_kflag(t)) {
kernel/bpf/btf.c
7262
tag_value = __btf_name_by_offset(btf, t->name_off);
kernel/bpf/btf.c
728
while (btf_type_is_modifier(t)) {
kernel/bpf/btf.c
729
id = t->type;
kernel/bpf/btf.c
730
t = btf_type_by_id(btf, t->type);
kernel/bpf/btf.c
7313
const struct btf_type *t;
kernel/bpf/btf.c
7339
t = btf_type_by_id(btf, id);
kernel/bpf/btf.c
7341
err = btf_struct_walk(log, btf, t, off, size, &id, &tmp_flag, field_name);
kernel/bpf/btf.c
736
return t;
kernel/bpf/btf.c
7366
t = btf_type_by_id(btf, id);
kernel/bpf/btf.c
7377
} while (t);
kernel/bpf/btf.c
742
const struct btf_type *t;
kernel/bpf/btf.c
7439
const struct btf_type *t;
kernel/bpf/btf.c
744
t = btf_type_skip_modifiers(btf, id, NULL);
kernel/bpf/btf.c
7445
t = btf_type_by_id(btf, btf_id);
kernel/bpf/btf.c
7446
while (t && btf_type_is_modifier(t))
kernel/bpf/btf.c
7447
t = btf_type_by_id(btf, t->type);
kernel/bpf/btf.c
7448
if (!t)
kernel/bpf/btf.c
745
if (!btf_type_is_ptr(t))
kernel/bpf/btf.c
7450
*ret_type = t;
kernel/bpf/btf.c
7451
if (btf_type_is_ptr(t))
kernel/bpf/btf.c
7454
if (btf_type_is_int(t) || btf_is_any_enum(t) || btf_type_is_struct(t))
kernel/bpf/btf.c
7455
return t->size;
kernel/bpf/btf.c
7459
static u8 __get_type_fmodel_flags(const struct btf_type *t)
kernel/bpf/btf.c
7463
if (btf_type_is_struct(t))
kernel/bpf/btf.c
7465
if (btf_type_is_signed_int(t))
kernel/bpf/btf.c
7478
const struct btf_type *t;
kernel/bpf/btf.c
748
return btf_type_skip_modifiers(btf, t->type, res_id);
kernel/bpf/btf.c
7503
ret = __get_type_size(btf, func->type, &t);
kernel/bpf/btf.c
7504
if (ret < 0 || btf_type_is_struct(t)) {
kernel/bpf/btf.c
7507
tname, btf_type_str(t));
kernel/bpf/btf.c
7511
m->ret_flags = __get_type_fmodel_flags(t);
kernel/bpf/btf.c
7520
ret = __get_type_size(btf, args[i].type, &t);
kernel/bpf/btf.c
7526
tname, i, btf_type_str(t));
kernel/bpf/btf.c
7536
m->arg_flags[i] = __get_type_fmodel_flags(t);
kernel/bpf/btf.c
766
static bool btf_type_is_resolve_source_only(const struct btf_type *t)
kernel/bpf/btf.c
768
return btf_type_is_var(t) ||
kernel/bpf/btf.c
7684
static bool btf_is_dynptr_ptr(const struct btf *btf, const struct btf_type *t)
kernel/bpf/btf.c
7688
t = btf_type_by_id(btf, t->type); /* skip PTR */
kernel/bpf/btf.c
769
btf_type_is_decl_tag(t) ||
kernel/bpf/btf.c
7690
while (btf_type_is_modifier(t))
kernel/bpf/btf.c
7691
t = btf_type_by_id(btf, t->type);
kernel/bpf/btf.c
7694
if (btf_type_is_struct(t) ||
kernel/bpf/btf.c
7695
(btf_type_is_fwd(t) && btf_type_kflag(t) == 0)) {
kernel/bpf/btf.c
7696
name = btf_str_by_offset(btf, t->name_off);
kernel/bpf/btf.c
770
btf_type_is_datasec(t);
kernel/bpf/btf.c
7720
const struct btf *btf, const struct btf_type *t)
kernel/bpf/btf.c
7731
type_id = t->type;
kernel/bpf/btf.c
7732
t = btf_type_by_id(btf, t->type);
kernel/bpf/btf.c
7733
while (btf_type_is_modifier(t)) {
kernel/bpf/btf.c
7734
type_id = t->type;
kernel/bpf/btf.c
7735
t = btf_type_by_id(btf, t->type);
kernel/bpf/btf.c
7743
arg_idx, btf_type_str(t), __btf_name_by_offset(btf, t->name_off),
kernel/bpf/btf.c
7749
arg_idx, btf_type_str(t), __btf_name_by_offset(btf, t->name_off),
kernel/bpf/btf.c
7756
arg_idx, btf_type_str(t), __btf_name_by_offset(btf, t->name_off));
kernel/bpf/btf.c
7796
const struct btf_type *t, *ref_t, *fn_t;
kernel/bpf/btf.c
7834
t = btf_type_by_id(btf, fn_t->type);
kernel/bpf/btf.c
7835
if (!t || !btf_type_is_func_proto(t)) {
kernel/bpf/btf.c
7839
args = (const struct btf_param *)(t + 1);
kernel/bpf/btf.c
7840
nargs = btf_type_vlen(t);
kernel/bpf/btf.c
7849
t = btf_type_by_id(btf, t->type);
kernel/bpf/btf.c
7850
while (btf_type_is_modifier(t))
kernel/bpf/btf.c
7851
t = btf_type_by_id(btf, t->type);
kernel/bpf/btf.c
7852
if (!btf_type_is_int(t) && !btf_is_any_enum(t)) {
kernel/bpf/btf.c
789
static bool btf_type_needs_resolve(const struct btf_type *t)
kernel/bpf/btf.c
7903
t = btf_type_by_id(btf, args[i].type);
kernel/bpf/btf.c
7904
while (btf_type_is_modifier(t))
kernel/bpf/btf.c
7905
t = btf_type_by_id(btf, t->type);
kernel/bpf/btf.c
7906
if (!btf_type_is_ptr(t))
kernel/bpf/btf.c
7909
if ((tags & ARG_TAG_CTX) || btf_is_prog_ctx_type(log, btf, t, prog_type, i)) {
kernel/bpf/btf.c
791
return btf_type_is_modifier(t) ||
kernel/bpf/btf.c
7915
btf_validate_prog_ctx_type(log, btf, t, i, prog_type,
kernel/bpf/btf.c
792
btf_type_is_ptr(t) ||
kernel/bpf/btf.c
7921
if (btf_is_dynptr_ptr(btf, t)) {
kernel/bpf/btf.c
793
btf_type_is_struct(t) ||
kernel/bpf/btf.c
7937
kern_type_id = btf_get_ptr_to_btf_id(log, i, btf, t);
kernel/bpf/btf.c
794
btf_type_is_array(t) ||
kernel/bpf/btf.c
795
btf_type_is_var(t) ||
kernel/bpf/btf.c
7956
ref_t = btf_type_skip_modifiers(btf, t->type, NULL);
kernel/bpf/btf.c
796
btf_type_is_func(t) ||
kernel/bpf/btf.c
7963
kern_type_id = btf_get_ptr_to_btf_id(log, i, btf, t);
kernel/bpf/btf.c
797
btf_type_is_decl_tag(t) ||
kernel/bpf/btf.c
7970
tname = __btf_name_by_offset(vmlinux_btf, t->name_off);
kernel/bpf/btf.c
798
btf_type_is_datasec(t);
kernel/bpf/btf.c
7995
t = btf_type_skip_modifiers(btf, t->type, NULL);
kernel/bpf/btf.c
7996
ref_t = btf_resolve_size(btf, t, &mem_size);
kernel/bpf/btf.c
7999
i, btf_type_str(t), btf_name_by_offset(btf, t->name_off),
kernel/bpf/btf.c
8016
if (btf_type_is_int(t) || btf_is_any_enum(t)) {
kernel/bpf/btf.c
802
static bool btf_type_has_size(const struct btf_type *t)
kernel/bpf/btf.c
8023
i, btf_type_str(t), tname);
kernel/bpf/btf.c
8036
const struct btf_type *t = btf_type_by_id(btf, type_id);
kernel/bpf/btf.c
804
switch (BTF_INFO_KIND(t->info)) {
kernel/bpf/btf.c
8042
btf_type_ops(t)->show(btf, t, type_id, obj, 0, show);
kernel/bpf/btf.c
832
static u32 btf_type_int(const struct btf_type *t)
kernel/bpf/btf.c
834
return *(u32 *)(t + 1);
kernel/bpf/btf.c
837
static const struct btf_array *btf_type_array(const struct btf_type *t)
kernel/bpf/btf.c
839
return (const struct btf_array *)(t + 1);
kernel/bpf/btf.c
842
static const struct btf_enum *btf_type_enum(const struct btf_type *t)
kernel/bpf/btf.c
844
return (const struct btf_enum *)(t + 1);
kernel/bpf/btf.c
847
static const struct btf_var *btf_type_var(const struct btf_type *t)
kernel/bpf/btf.c
849
return (const struct btf_var *)(t + 1);
kernel/bpf/btf.c
852
static const struct btf_decl_tag *btf_type_decl_tag(const struct btf_type *t)
kernel/bpf/btf.c
854
return (const struct btf_decl_tag *)(t + 1);
kernel/bpf/btf.c
8549
const struct btf_type *t;
kernel/bpf/btf.c
8557
t = btf_type_skip_modifiers(btf, arg->type, NULL);
kernel/bpf/btf.c
8558
if (!t || !btf_type_is_ptr(t))
kernel/bpf/btf.c
8560
t = btf_type_skip_modifiers(btf, t->type, &btf_id);
kernel/bpf/btf.c
8561
if (!t || !__btf_type_is_struct(t))
kernel/bpf/btf.c
8564
name = btf_name_by_offset(btf, t->name_off);
kernel/bpf/btf.c
857
static const struct btf_enum64 *btf_type_enum64(const struct btf_type *t)
kernel/bpf/btf.c
8576
const struct btf_type *t;
kernel/bpf/btf.c
859
return (const struct btf_enum64 *)(t + 1);
kernel/bpf/btf.c
8597
t = btf_type_by_id(btf, btf_id);
kernel/bpf/btf.c
8598
if (t->size == 0 || (t->size % 8))
kernel/bpf/btf.c
8604
iter_name = btf_name_by_offset(btf, t->name_off) + sizeof(ITER_PREFIX) - 1;
kernel/bpf/btf.c
862
static const struct btf_kind_operations *btf_type_ops(const struct btf_type *t)
kernel/bpf/btf.c
8622
t = btf_type_skip_modifiers(btf, func->type, NULL);
kernel/bpf/btf.c
8623
if (!t || !btf_type_is_ptr(t))
kernel/bpf/btf.c
8629
t = btf_type_by_id(btf, func->type);
kernel/bpf/btf.c
8630
if (!t || !btf_type_is_void(t))
kernel/bpf/btf.c
864
return kind_ops[BTF_INFO_KIND(t->info)];
kernel/bpf/btf.c
8990
const struct btf_type *dtor_func, *dtor_func_proto, *t;
kernel/bpf/btf.c
9007
t = btf_type_by_id(btf, dtor_func_proto->type);
kernel/bpf/btf.c
9008
if (!t || !btf_type_is_void(t))
kernel/bpf/btf.c
9015
t = btf_type_by_id(btf, args[0].type);
kernel/bpf/btf.c
9019
if (!t || !btf_type_is_ptr(t))
kernel/bpf/btf.c
9024
t = btf_type_by_id(btf, t->type);
kernel/bpf/btf.c
9025
if (!btf_type_is_void(t))
kernel/bpf/btf.c
9313
const struct btf_type *t;
kernel/bpf/btf.c
9320
t = btf_type_by_id(targ_btf, i);
kernel/bpf/btf.c
9321
if (btf_kind(t) != cands->kind)
kernel/bpf/btf.c
9324
targ_name = btf_name_by_offset(targ_btf, t->name_off);
kernel/bpf/btf.c
978
static bool __btf_type_int_is_regular(const struct btf_type *t, size_t expected_size)
kernel/bpf/btf.c
980
u32 int_data = btf_type_int(t);
kernel/bpf/btf.c
990
static bool btf_type_int_is_regular(const struct btf_type *t)
kernel/bpf/btf.c
992
return __btf_type_int_is_regular(t, 0);
kernel/bpf/btf.c
995
bool btf_type_is_i32(const struct btf_type *t)
kernel/bpf/btf.c
997
return btf_type_is_int(t) && __btf_type_int_is_regular(t, 4);
kernel/bpf/helpers.c
1170
struct bpf_hrtimer *t = container_of(hrtimer, struct bpf_hrtimer, timer);
kernel/bpf/helpers.c
1171
struct bpf_map *map = t->cb.map;
kernel/bpf/helpers.c
1172
void *value = t->cb.value;
kernel/bpf/helpers.c
1178
callback_fn = rcu_dereference_check(t->cb.callback_fn, rcu_read_lock_bh_held());
kernel/bpf/helpers.c
1188
this_cpu_write(hrtimer_running, t);
kernel/bpf/helpers.c
1245
struct bpf_hrtimer *t = container_of(cb, struct bpf_hrtimer, cb);
kernel/bpf/helpers.c
1256
if (hrtimer_try_to_cancel(&t->timer) < 0)
kernel/bpf/helpers.c
1306
struct bpf_hrtimer *t;
kernel/bpf/helpers.c
1333
t = (struct bpf_hrtimer *)cb;
kernel/bpf/helpers.c
1335
atomic_set(&t->cancelling, 0);
kernel/bpf/helpers.c
1336
hrtimer_setup(&t->timer, bpf_timer_cb, clockid, HRTIMER_MODE_REL_SOFT);
kernel/bpf/helpers.c
1507
struct bpf_hrtimer *t;
kernel/bpf/helpers.c
1513
t = READ_ONCE(async->timer);
kernel/bpf/helpers.c
1514
if (!t || !READ_ONCE(t->cb.prog))
kernel/bpf/helpers.c
1529
if (!refcount_inc_not_zero(&t->cb.refcnt))
kernel/bpf/helpers.c
1533
hrtimer_start(&t->timer, ns_to_ktime(nsecs), mode);
kernel/bpf/helpers.c
1534
bpf_async_refcount_put(&t->cb);
kernel/bpf/helpers.c
1537
return bpf_async_schedule_op(&t->cb, BPF_ASYNC_START, nsecs, mode);
kernel/bpf/helpers.c
1552
struct bpf_hrtimer *t, *cur_t;
kernel/bpf/helpers.c
1559
t = READ_ONCE(async->timer);
kernel/bpf/helpers.c
1560
if (!t)
kernel/bpf/helpers.c
1564
if (cur_t == t) {
kernel/bpf/helpers.c
1579
atomic_inc(&t->cancelling);
kernel/bpf/helpers.c
1592
atomic_dec(&t->cancelling);
kernel/bpf/helpers.c
1596
bpf_async_update_prog_callback(&t->cb, NULL, NULL);
kernel/bpf/helpers.c
1600
ret = hrtimer_cancel(&t->timer);
kernel/bpf/helpers.c
1602
atomic_dec(&t->cancelling);
kernel/bpf/helpers.c
1618
struct bpf_hrtimer *t = container_of(cb, struct bpf_hrtimer, cb);
kernel/bpf/helpers.c
1622
hrtimer_start(&t->timer, ns_to_ktime(timer_nsec), timer_mode);
kernel/bpf/helpers.c
1625
hrtimer_try_to_cancel(&t->timer);
kernel/bpf/helpers.c
4493
struct bpf_hrtimer *t = container_of(cb, struct bpf_hrtimer, cb);
kernel/bpf/helpers.c
4495
ret = hrtimer_try_to_cancel(&t->timer);
kernel/bpf/log.c
656
static bool type_is_map_ptr(enum bpf_reg_type t) {
kernel/bpf/log.c
657
switch (base_type(t)) {
kernel/bpf/log.c
677
enum bpf_reg_type t;
kernel/bpf/log.c
680
t = reg->type;
kernel/bpf/log.c
681
if (t == SCALAR_VALUE && reg->precise)
kernel/bpf/log.c
683
if (t == SCALAR_VALUE && tnum_is_const(reg->var_off)) {
kernel/bpf/log.c
688
verbose(env, "%s", reg_type_str(env, t));
kernel/bpf/log.c
689
if (t == PTR_TO_ARENA)
kernel/bpf/log.c
691
if (t == PTR_TO_STACK) {
kernel/bpf/log.c
699
if (base_type(t) == PTR_TO_BTF_ID)
kernel/bpf/log.c
710
if (type_is_map_ptr(t)) {
kernel/bpf/log.c
717
if (t != SCALAR_VALUE && reg->off) {
kernel/bpf/log.c
721
if (type_is_pkt_pointer(t)) {
kernel/bpf/log.c
725
if (base_type(t) == PTR_TO_MEM) {
kernel/bpf/log.c
729
if (t == CONST_PTR_TO_DYNPTR)
kernel/bpf/memalloc.c
267
struct llist_node *pos, *t;
kernel/bpf/memalloc.c
270
llist_for_each_safe(pos, t, llnode) {
kernel/bpf/memalloc.c
310
struct llist_node *llnode, *t;
kernel/bpf/memalloc.c
321
llist_for_each_safe(llnode, t, llist_del_all(&c->free_by_rcu_ttrace))
kernel/bpf/memalloc.c
340
struct llist_node *llnode, *t;
kernel/bpf/memalloc.c
360
llist_for_each_safe(llnode, t, llist_del_all(&c->free_llist_extra))
kernel/bpf/memalloc.c
388
struct llist_node *llnode, *t;
kernel/bpf/memalloc.c
394
llist_for_each_safe(llnode, t, llist_del_all(&c->free_llist_extra_rcu))
kernel/bpf/preload/iterators/iterators.bpf.c
61
struct btf_type **types, *t;
kernel/bpf/preload/iterators/iterators.bpf.c
69
bpf_probe_read_kernel(&t, sizeof(t), types + btf_id);
kernel/bpf/preload/iterators/iterators.bpf.c
70
name_off = BPF_CORE_READ(t, name_off);
kernel/bpf/syscall.c
199
const struct btf_type *t;
kernel/bpf/syscall.c
217
t = btf_type_by_id(field->kptr.btf, field->kptr.btf_id);
kernel/bpf/syscall.c
219
if (check_add_overflow(start, t->size - 1, &end)) {
kernel/bpf/tnum.c
287
u64 tnum_step(struct tnum t, u64 z)
kernel/bpf/tnum.c
292
tmax = t.value | t.mask;
kernel/bpf/tnum.c
299
if (z < t.value)
kernel/bpf/tnum.c
300
return t.value;
kernel/bpf/tnum.c
303
j = t.value | (z & t.mask);
kernel/bpf/tnum.c
306
p = ~z & t.value & ~t.mask;
kernel/bpf/tnum.c
310
s = ~q & t.value; /* positions <= k matched to t.value */
kernel/bpf/tnum.c
314
p = z & ~t.value & ~t.mask;
kernel/bpf/tnum.c
317
r = q & t.mask & z; /* unknown positions > k, matched to z */
kernel/bpf/tnum.c
318
s = q & ~t.mask; /* known positions > k, set to 1 */
kernel/bpf/tnum.c
323
w = (u & t.mask) | t.value;
kernel/bpf/verifier.c
12012
const struct btf_type *t;
kernel/bpf/verifier.c
12015
t = btf_type_skip_modifiers(meta.ret_btf, meta.ret_btf_id, NULL);
kernel/bpf/verifier.c
12016
if (!btf_type_is_struct(t)) {
kernel/bpf/verifier.c
12022
ret = btf_resolve_size(meta.ret_btf, t, &tsize);
kernel/bpf/verifier.c
12024
tname = btf_name_by_offset(meta.ret_btf, t->name_off);
kernel/bpf/verifier.c
12234
const struct btf_type *t;
kernel/bpf/verifier.c
12236
t = btf_type_skip_modifiers(btf, arg->type, NULL);
kernel/bpf/verifier.c
12237
if (!btf_type_is_scalar(t) || reg->type != SCALAR_VALUE)
kernel/bpf/verifier.c
12247
const struct btf_type *t;
kernel/bpf/verifier.c
12249
t = btf_type_skip_modifiers(btf, arg->type, NULL);
kernel/bpf/verifier.c
12250
if (!btf_type_is_scalar(t) || reg->type != SCALAR_VALUE)
kernel/bpf/verifier.c
12348
const struct btf_type *t;
kernel/bpf/verifier.c
12351
t = btf_type_skip_modifiers(btf, arg->type, NULL);
kernel/bpf/verifier.c
12352
if (!t)
kernel/bpf/verifier.c
12354
if (!btf_type_is_ptr(t))
kernel/bpf/verifier.c
12356
t = btf_type_skip_modifiers(btf, t->type, &res_id);
kernel/bpf/verifier.c
12357
if (!t)
kernel/bpf/verifier.c
12407
static bool is_rbtree_node_type(const struct btf_type *t)
kernel/bpf/verifier.c
12409
return t == btf_type_by_id(btf_vmlinux, kf_arg_btf_ids[KF_ARG_RB_NODE_ID]);
kernel/bpf/verifier.c
12412
static bool is_list_node_type(const struct btf_type *t)
kernel/bpf/verifier.c
12414
return t == btf_type_by_id(btf_vmlinux, kf_arg_btf_ids[KF_ARG_LIST_NODE_ID]);
kernel/bpf/verifier.c
12420
const struct btf_type *t;
kernel/bpf/verifier.c
12422
t = btf_type_resolve_func_ptr(btf, arg->type, NULL);
kernel/bpf/verifier.c
12423
if (!t)
kernel/bpf/verifier.c
12437
const struct btf_type *t, int rec)
kernel/bpf/verifier.c
12443
if (!btf_type_is_struct(t))
kernel/bpf/verifier.c
12446
for_each_member(i, t, member) {
kernel/bpf/verifier.c
12681
const struct btf_type *t, const struct btf_type *ref_t,
kernel/bpf/verifier.c
12705
if (btf_is_prog_ctx_type(&env->log, meta->btf, t, resolve_prog_type(env->prog), argno))
kernel/bpf/verifier.c
13259
const struct btf_type *et, *t;
kernel/bpf/verifier.c
13289
t = btf_type_by_id(reg->btf, reg->btf_id);
kernel/bpf/verifier.c
13298
node_off, btf_name_by_offset(reg->btf, t->name_off));
kernel/bpf/verifier.c
13378
const struct btf_type *t, *ref_t, *resolve_ret;
kernel/bpf/verifier.c
13384
t = btf_type_skip_modifiers(btf, args[i].type, NULL);
kernel/bpf/verifier.c
13400
if (btf_type_is_scalar(t)) {
kernel/bpf/verifier.c
13446
if (!btf_type_is_ptr(t)) {
kernel/bpf/verifier.c
13447
verbose(env, "Unrecognized arg#%d type %s\n", i, btf_type_str(t));
kernel/bpf/verifier.c
13469
ref_t = btf_type_skip_modifiers(btf, t->type, &ref_id);
kernel/bpf/verifier.c
13472
kf_arg_type = get_kfunc_ptr_arg_type(env, meta, t, ref_t, ref_tname, args, i, nargs);
kernel/bpf/verifier.c
14117
const struct btf_type *t, *ptr_type;
kernel/bpf/verifier.c
14350
t = btf_type_skip_modifiers(desc_btf, meta.func_proto->type, NULL);
kernel/bpf/verifier.c
14352
if (is_kfunc_acquire(&meta) && !btf_type_is_struct_ptr(meta.btf, t)) {
kernel/bpf/verifier.c
14363
if (btf_type_is_scalar(t)) {
kernel/bpf/verifier.c
14368
mark_btf_func_reg_size(env, BPF_REG_0, t->size);
kernel/bpf/verifier.c
14369
} else if (btf_type_is_ptr(t)) {
kernel/bpf/verifier.c
14370
ptr_type = btf_type_skip_modifiers(desc_btf, t->type, &ptr_type_id);
kernel/bpf/verifier.c
14466
} else if (btf_type_is_void(t)) {
kernel/bpf/verifier.c
14485
t = btf_type_skip_modifiers(desc_btf, args[i].type, NULL);
kernel/bpf/verifier.c
14486
if (btf_type_is_ptr(t))
kernel/bpf/verifier.c
14490
mark_btf_func_reg_size(env, regno, t->size);
kernel/bpf/verifier.c
16981
struct tnum t;
kernel/bpf/verifier.c
17009
t = tnum_intersect(tnum_subreg(reg1->var_off), tnum_subreg(reg2->var_off));
kernel/bpf/verifier.c
17010
reg1->var_off = tnum_with_subreg(reg1->var_off, t);
kernel/bpf/verifier.c
17011
reg2->var_off = tnum_with_subreg(reg2->var_off, t);
kernel/bpf/verifier.c
17084
t = tnum_or(tnum_subreg(reg1->var_off), tnum_const(val));
kernel/bpf/verifier.c
17085
reg1->var_off = tnum_with_subreg(reg1->var_off, t);
kernel/bpf/verifier.c
17101
t = tnum_and(tnum_subreg(reg1->var_off), tnum_const(~val));
kernel/bpf/verifier.c
17102
reg1->var_off = tnum_with_subreg(reg1->var_off, t);
kernel/bpf/verifier.c
18158
static void merge_callee_effects(struct bpf_verifier_env *env, int t, int w)
kernel/bpf/verifier.c
18162
caller = bpf_find_containing_subprog(env, t);
kernel/bpf/verifier.c
18248
static int push_insn(int t, int w, int e, struct bpf_verifier_env *env)
kernel/bpf/verifier.c
18253
if (e == FALLTHROUGH && insn_state[t] >= (DISCOVERED | FALLTHROUGH))
kernel/bpf/verifier.c
18256
if (e == BRANCH && insn_state[t] >= (DISCOVERED | BRANCH))
kernel/bpf/verifier.c
18260
verbose_linfo(env, t, "%d: ", t);
kernel/bpf/verifier.c
18261
verbose(env, "jump out of range from insn %d to %d\n", t, w);
kernel/bpf/verifier.c
18273
insn_state[t] = DISCOVERED | e;
kernel/bpf/verifier.c
18282
verbose_linfo(env, t, "%d: ", t);
kernel/bpf/verifier.c
18284
verbose(env, "back-edge from insn %d to %d\n", t, w);
kernel/bpf/verifier.c
18288
insn_state[t] = DISCOVERED | e;
kernel/bpf/verifier.c
18296
static int visit_func_call_insn(int t, struct bpf_insn *insns,
kernel/bpf/verifier.c
18303
insn_sz = bpf_is_ldimm64(&insns[t]) ? 2 : 1;
kernel/bpf/verifier.c
18304
ret = push_insn(t, t + insn_sz, FALLTHROUGH, env);
kernel/bpf/verifier.c
18308
mark_prune_point(env, t + insn_sz);
kernel/bpf/verifier.c
18310
mark_jmp_point(env, t + insn_sz);
kernel/bpf/verifier.c
18313
w = t + insns[t].imm + 1;
kernel/bpf/verifier.c
18314
mark_prune_point(env, t);
kernel/bpf/verifier.c
18315
merge_callee_effects(env, t, w);
kernel/bpf/verifier.c
18316
ret = push_insn(t, w, BRANCH, env);
kernel/bpf/verifier.c
18735
create_jt(int t, struct bpf_verifier_env *env)
kernel/bpf/verifier.c
18742
subprog = bpf_find_containing_subprog(env, t);
kernel/bpf/verifier.c
18753
t, subprog_start, subprog_end);
kernel/bpf/verifier.c
18763
static int visit_gotox_insn(int t, struct bpf_verifier_env *env)
kernel/bpf/verifier.c
18771
jt = env->insn_aux_data[t].jt;
kernel/bpf/verifier.c
18773
jt = create_jt(t, env);
kernel/bpf/verifier.c
18777
env->insn_aux_data[t].jt = jt;
kernel/bpf/verifier.c
18780
mark_prune_point(env, t);
kernel/bpf/verifier.c
18784
verbose(env, "indirect jump out of range from insn %d to %d\n", t, w);
kernel/bpf/verifier.c
18805
static int visit_tailcall_insn(struct bpf_verifier_env *env, int t)
kernel/bpf/verifier.c
18810
if (env->insn_aux_data[t].jt)
kernel/bpf/verifier.c
18817
subprog = bpf_find_containing_subprog(env, t);
kernel/bpf/verifier.c
18818
jt->items[0] = t + 1;
kernel/bpf/verifier.c
18820
env->insn_aux_data[t].jt = jt;
kernel/bpf/verifier.c
18829
static int visit_insn(int t, struct bpf_verifier_env *env)
kernel/bpf/verifier.c
18831
struct bpf_insn *insns = env->prog->insnsi, *insn = &insns[t];
kernel/bpf/verifier.c
18835
return visit_func_call_insn(t, insns, env, true);
kernel/bpf/verifier.c
18841
return push_insn(t, t + insn_sz, FALLTHROUGH, env);
kernel/bpf/verifier.c
18855
mark_prune_point(env, t);
kernel/bpf/verifier.c
18866
mark_calls_callback(env, t);
kernel/bpf/verifier.c
18867
mark_force_checkpoint(env, t);
kernel/bpf/verifier.c
18868
mark_prune_point(env, t);
kernel/bpf/verifier.c
18869
mark_jmp_point(env, t);
kernel/bpf/verifier.c
18881
mark_subprog_might_sleep(env, t);
kernel/bpf/verifier.c
18883
mark_subprog_changes_pkt_data(env, t);
kernel/bpf/verifier.c
18885
visit_tailcall_insn(env, t);
kernel/bpf/verifier.c
18891
mark_prune_point(env, t);
kernel/bpf/verifier.c
18903
mark_force_checkpoint(env, t);
kernel/bpf/verifier.c
18911
mark_subprog_might_sleep(env, t);
kernel/bpf/verifier.c
18913
mark_subprog_changes_pkt_data(env, t);
kernel/bpf/verifier.c
18915
return visit_func_call_insn(t, insns, env, insn->src_reg == BPF_PSEUDO_CALL);
kernel/bpf/verifier.c
18919
return visit_gotox_insn(t, env);
kernel/bpf/verifier.c
18927
ret = push_insn(t, t + off + 1, FALLTHROUGH, env);
kernel/bpf/verifier.c
18931
mark_prune_point(env, t + off + 1);
kernel/bpf/verifier.c
18932
mark_jmp_point(env, t + off + 1);
kernel/bpf/verifier.c
18938
mark_prune_point(env, t);
kernel/bpf/verifier.c
18940
mark_force_checkpoint(env, t);
kernel/bpf/verifier.c
18942
ret = push_insn(t, t + 1, FALLTHROUGH, env);
kernel/bpf/verifier.c
18946
return push_insn(t, t + insn->off + 1, BRANCH, env);
kernel/bpf/verifier.c
18981
int t = insn_stack[env->cfg.cur_stack - 1];
kernel/bpf/verifier.c
18983
ret = visit_insn(t, env);
kernel/bpf/verifier.c
18986
insn_state[t] = EXPLORED;
kernel/bpf/verifier.c
21421
const struct btf_type *t;
kernel/bpf/verifier.c
21432
t = btf_type_by_id(btf, i);
kernel/bpf/verifier.c
21433
if (BTF_INFO_KIND(t->info) != BTF_KIND_DATASEC)
kernel/bpf/verifier.c
21436
tname = btf_name_by_offset(btf, t->name_off);
kernel/bpf/verifier.c
21498
const struct btf_type *t;
kernel/bpf/verifier.c
21506
t = btf_type_by_id(btf, id);
kernel/bpf/verifier.c
21507
if (!t) {
kernel/bpf/verifier.c
21512
if (!btf_type_is_var(t) && !btf_type_is_func(t)) {
kernel/bpf/verifier.c
21517
sym_name = btf_name_by_offset(btf, t->name_off);
kernel/bpf/verifier.c
21527
if (btf_type_is_func(t)) {
kernel/bpf/verifier.c
21544
type = t->type;
kernel/bpf/verifier.c
21545
t = btf_type_skip_modifiers(btf, type, NULL);
kernel/bpf/verifier.c
21550
} else if (!btf_type_is_struct(t)) {
kernel/bpf/verifier.c
21556
ret = btf_resolve_size(btf, t, &tsize);
kernel/bpf/verifier.c
21558
tname = btf_name_by_offset(btf, t->name_off);
kernel/bpf/verifier.c
24819
const struct btf_type *t, *func_proto;
kernel/bpf/verifier.c
24858
t = st_ops_desc->type;
kernel/bpf/verifier.c
24860
if (member_idx >= btf_type_vlen(t)) {
kernel/bpf/verifier.c
24866
member = &btf_type_member(t)[member_idx];
kernel/bpf/verifier.c
24876
member_off = __btf_member_bit_offset(t, member) / 8;
kernel/bpf/verifier.c
24885
err = st_ops->check_member(t, member, prog);
kernel/bpf/verifier.c
24970
const struct btf_type *t;
kernel/bpf/verifier.c
24987
t = btf_type_by_id(btf, btf_id);
kernel/bpf/verifier.c
24988
if (!t) {
kernel/bpf/verifier.c
24992
tname = btf_name_by_offset(btf, t->name_off);
kernel/bpf/verifier.c
25112
if (!btf_type_is_typedef(t)) {
kernel/bpf/verifier.c
25140
t = btf_type_by_id(btf, t->type);
kernel/bpf/verifier.c
25141
if (!btf_type_is_ptr(t))
kernel/bpf/verifier.c
25145
t = btf_type_by_id(btf, ret);
kernel/bpf/verifier.c
25146
if (!btf_type_is_func(t))
kernel/bpf/verifier.c
25151
t = btf_type_by_id(btf, t->type);
kernel/bpf/verifier.c
25152
if (!btf_type_is_func_proto(t))
kernel/bpf/verifier.c
25158
if (!btf_type_is_func(t)) {
kernel/bpf/verifier.c
25163
t = btf_type_by_id(btf, t->type);
kernel/bpf/verifier.c
25164
if (!btf_type_is_func_proto(t))
kernel/bpf/verifier.c
25166
ret = btf_distill_func_proto(log, btf, t, tname, &tgt_info->fmodel);
kernel/bpf/verifier.c
25185
if (!btf_type_is_func(t)) {
kernel/bpf/verifier.c
25191
btf_check_type_match(log, prog, btf, t))
kernel/bpf/verifier.c
25193
t = btf_type_by_id(btf, t->type);
kernel/bpf/verifier.c
25194
if (!btf_type_is_func_proto(t))
kernel/bpf/verifier.c
25203
t = NULL;
kernel/bpf/verifier.c
25205
ret = btf_distill_func_proto(log, btf, t, tname, &tgt_info->fmodel);
kernel/bpf/verifier.c
25291
tgt_info->tgt_type = t;
kernel/bpf/verifier.c
25785
u32 i, j, t, w;
kernel/bpf/verifier.c
25929
t = stack[--stack_sz];
kernel/bpf/verifier.c
25930
low[t] = NOT_ON_STACK;
kernel/bpf/verifier.c
25932
aux[t].scc = next_scc_id;
kernel/bpf/verifier.c
25933
} while (t != w);
kernel/bpf/verifier.c
3119
const struct btf_type *t;
kernel/bpf/verifier.c
3129
t = btf_type_by_id(btf, main_btf_id);
kernel/bpf/verifier.c
3130
if (!t) {
kernel/bpf/verifier.c
3135
name = btf_find_decl_tag_value(btf, t, -1, "exception_callback:");
kernel/bpf/verifier.c
3152
t = btf_type_by_id(btf, id);
kernel/bpf/verifier.c
3153
if (btf_func_linkage(t) != BTF_FUNC_GLOBAL) {
kernel/bpf/verifier.c
3797
u32 regno, struct bpf_reg_state *reg, enum reg_arg_type t)
kernel/bpf/verifier.c
3821
if (t == SRC_OP)
kernel/bpf/verifier.c
3839
if (t != SRC_OP)
kernel/bpf/verifier.c
3850
if (t == SRC_OP && reg->type != SCALAR_VALUE)
kernel/bpf/verifier.c
3863
if (t != SRC_OP)
kernel/bpf/verifier.c
3931
enum reg_arg_type t)
kernel/bpf/verifier.c
3945
rw64 = is_reg64(insn, regno, reg, t);
kernel/bpf/verifier.c
3946
if (t == SRC_OP) {
kernel/bpf/verifier.c
3967
if (t == DST_OP)
kernel/bpf/verifier.c
3974
enum reg_arg_type t)
kernel/bpf/verifier.c
3979
return __check_reg_arg(env, state->regs, regno, t);
kernel/bpf/verifier.c
6146
const struct btf_type *t;
kernel/bpf/verifier.c
6148
t = btf_type_by_id(field->kptr.btf, field->kptr.btf_id);
kernel/bpf/verifier.c
6152
reg->mem_size = t->size;
kernel/bpf/verifier.c
6305
enum bpf_access_type t)
kernel/bpf/verifier.c
6317
if (t == BPF_WRITE)
kernel/bpf/verifier.c
6335
if (t == BPF_WRITE)
kernel/bpf/verifier.c
6388
enum bpf_access_type t, struct bpf_insn_access_aux *info)
kernel/bpf/verifier.c
6391
env->ops->is_valid_access(off, size, t, env->prog, info)) {
kernel/bpf/verifier.c
6433
enum bpf_access_type t)
kernel/bpf/verifier.c
6447
valid = bpf_sock_common_is_valid_access(off, size, t, &info);
kernel/bpf/verifier.c
6450
valid = bpf_sock_is_valid_access(off, size, t, &info);
kernel/bpf/verifier.c
6453
valid = bpf_tcp_sock_is_valid_access(off, size, t, &info);
kernel/bpf/verifier.c
6456
valid = bpf_xdp_sock_is_valid_access(off, size, t, &info);
kernel/bpf/verifier.c
7386
const struct btf_type *t = btf_type_by_id(reg->btf, reg->btf_id);
kernel/bpf/verifier.c
7387
const char *tname = btf_name_by_offset(reg->btf, t->name_off);
kernel/bpf/verifier.c
7542
const struct btf_type *t;
kernel/bpf/verifier.c
7558
t = btf_type_by_id(btf_vmlinux, *map->ops->map_btf_id);
kernel/bpf/verifier.c
7559
tname = btf_name_by_offset(btf_vmlinux, t->name_off);
kernel/bpf/verifier.c
7607
enum bpf_access_type t)
kernel/bpf/verifier.c
7611
if (t == BPF_WRITE || env->allow_uninit_stack)
kernel/bpf/verifier.c
7703
int off, int bpf_size, enum bpf_access_type t,
kernel/bpf/verifier.c
7723
if (t == BPF_WRITE) {
kernel/bpf/verifier.c
7737
if (t == BPF_WRITE && value_regno >= 0 &&
kernel/bpf/verifier.c
7742
err = check_map_access_type(env, regno, off, size, t);
kernel/bpf/verifier.c
7753
} else if (t == BPF_READ && value_regno >= 0) {
kernel/bpf/verifier.c
7796
if (t == BPF_WRITE && rdonly_mem) {
kernel/bpf/verifier.c
7802
if (t == BPF_WRITE && value_regno >= 0 &&
kernel/bpf/verifier.c
7815
if (!err && value_regno >= 0 && (t == BPF_READ || rdonly_mem))
kernel/bpf/verifier.c
7825
if (t == BPF_WRITE && value_regno >= 0 &&
kernel/bpf/verifier.c
7835
err = check_ctx_access(env, insn_idx, off, size, t, &info);
kernel/bpf/verifier.c
7838
if (!err && t == BPF_READ && value_regno >= 0) {
kernel/bpf/verifier.c
7874
err = check_stack_access_within_bounds(env, regno, off, size, t);
kernel/bpf/verifier.c
7878
if (t == BPF_READ)
kernel/bpf/verifier.c
7885
if (t == BPF_WRITE && !may_access_direct_pkt_data(env, NULL, t)) {
kernel/bpf/verifier.c
7889
if (t == BPF_WRITE && value_regno >= 0 &&
kernel/bpf/verifier.c
7896
if (!err && t == BPF_READ && value_regno >= 0)
kernel/bpf/verifier.c
7899
if (t == BPF_WRITE && value_regno >= 0 &&
kernel/bpf/verifier.c
7907
if (!err && t == BPF_READ && value_regno >= 0)
kernel/bpf/verifier.c
7910
if (t == BPF_WRITE) {
kernel/bpf/verifier.c
7915
err = check_sock_access(env, insn_idx, regno, off, size, t);
kernel/bpf/verifier.c
7920
if (!err && t == BPF_READ && value_regno >= 0)
kernel/bpf/verifier.c
7924
err = check_ptr_to_btf_access(env, regs, regno, off, size, t,
kernel/bpf/verifier.c
7927
err = check_ptr_to_map_access(env, regs, regno, off, size, t,
kernel/bpf/verifier.c
7935
if (t == BPF_WRITE) {
kernel/bpf/verifier.c
7948
if (!err && value_regno >= 0 && (rdonly_mem || t == BPF_READ))
kernel/bpf/verifier.c
7951
if (t == BPF_READ && value_regno >= 0)
kernel/bpf/verifier.c
7959
if (!err && size < BPF_REG_SIZE && value_regno >= 0 && t == BPF_READ &&
kernel/bpf/verifier.c
8972
const struct btf_type *t;
kernel/bpf/verifier.c
8991
t = btf_type_by_id(meta->btf, btf_id);
kernel/bpf/verifier.c
8992
nr_slots = t->size / BPF_REG_SIZE;
kernel/capability.c
276
bool has_ns_capability(struct task_struct *t,
kernel/capability.c
282
ret = security_capable(__task_cred(t), ns, cap, CAP_OPT_NONE);
kernel/capability.c
301
bool has_ns_capability_noaudit(struct task_struct *t,
kernel/capability.c
307
ret = security_capable(__task_cred(t), ns, cap, CAP_OPT_NOAUDIT);
kernel/capability.c
325
bool has_capability_noaudit(struct task_struct *t, int cap)
kernel/capability.c
327
return has_ns_capability_noaudit(t, &init_user_ns, cap);
kernel/cpu.c
1259
struct task_struct *t;
kernel/cpu.c
1265
t = find_lock_task_mm(p);
kernel/cpu.c
1266
if (!t)
kernel/cpu.c
1268
arch_clear_mm_cpumask_cpu(cpu, t->mm);
kernel/cpu.c
1269
task_unlock(t);
kernel/delayacct.c
62
struct ctl_table t;
kernel/delayacct.c
68
t = *table;
kernel/delayacct.c
69
t.data = &state;
kernel/delayacct.c
70
err = proc_dointvec_minmax(&t, write, buffer, lenp, ppos);
kernel/events/core.c
942
struct perf_cgroup_info *t;
kernel/events/core.c
944
t = per_cpu_ptr(event->cgrp->info, event->cpu);
kernel/events/core.c
945
return __perf_event_time_ctx(event, &t->time);
kernel/events/core.c
950
struct perf_cgroup_info *t;
kernel/events/core.c
952
t = per_cpu_ptr(event->cgrp->info, event->cpu);
kernel/events/core.c
953
if (!__load_acquire(&t->active))
kernel/events/core.c
954
return __perf_event_time_ctx(event, &t->time);
kernel/events/core.c
956
return __perf_event_time_ctx_now(event, &t->time, now);
kernel/events/uprobes.c
122
static void uprobe_warn(struct task_struct *t, const char *msg)
kernel/events/uprobes.c
124
pr_warn("uprobe: %s:%d failed to %s\n", t->comm, t->pid, msg);
kernel/events/uprobes.c
1532
struct rb_node *n, *t;
kernel/events/uprobes.c
1542
for (t = n; t; t = rb_prev(t)) {
kernel/events/uprobes.c
1543
u = rb_entry(t, struct uprobe, rb_node);
kernel/events/uprobes.c
1550
for (t = n; (t = rb_next(t)); ) {
kernel/events/uprobes.c
1551
u = rb_entry(t, struct uprobe, rb_node);
kernel/events/uprobes.c
2006
void uprobe_free_utask(struct task_struct *t)
kernel/events/uprobes.c
2008
struct uprobe_task *utask = t->utask;
kernel/events/uprobes.c
2014
t->utask = NULL;
kernel/events/uprobes.c
2133
static int dup_utask(struct task_struct *t, struct uprobe_task *o_utask)
kernel/events/uprobes.c
2142
t->utask = n_utask;
kernel/events/uprobes.c
2187
void uprobe_copy_process(struct task_struct *t, u64 flags)
kernel/events/uprobes.c
2193
t->utask = NULL;
kernel/events/uprobes.c
2198
if (mm == t->mm && !(flags & CLONE_VFORK))
kernel/events/uprobes.c
2201
if (dup_utask(t, utask))
kernel/events/uprobes.c
2202
return uprobe_warn(t, "dup ret instances");
kernel/events/uprobes.c
2207
return uprobe_warn(t, "dup xol area");
kernel/events/uprobes.c
2209
if (mm == t->mm)
kernel/events/uprobes.c
2212
t->utask->dup_xol_addr = area->vaddr;
kernel/events/uprobes.c
2213
init_task_work(&t->utask->dup_xol_work, dup_xol_work);
kernel/events/uprobes.c
2214
task_work_add(t, &t->utask->dup_xol_work, TWA_RESUME);
kernel/events/uprobes.c
2358
struct task_struct *t = current;
kernel/events/uprobes.c
2359
struct uprobe_task *utask = t->utask;
kernel/events/uprobes.c
2366
if (task_sigpending(t)) {
kernel/events/uprobes.c
2368
clear_tsk_thread_flag(t, TIF_SIGPENDING);
kernel/events/uprobes.c
2370
if (__fatal_signal_pending(t) || arch_uprobe_xol_was_trapped(t)) {
kernel/events/uprobes.c
2372
set_tsk_thread_flag(t, TIF_UPROBE);
kernel/exit.c
470
struct task_struct *t;
kernel/exit.c
472
for_each_thread(g, t) {
kernel/exit.c
473
struct mm_struct *t_mm = READ_ONCE(t->mm);
kernel/exit.c
475
if (__try_to_set_owner(t, mm))
kernel/exit.c
588
struct task_struct *t;
kernel/exit.c
590
for_each_thread(p, t) {
kernel/exit.c
591
if (!(t->flags & PF_EXITING))
kernel/exit.c
592
return t;
kernel/exit.c
700
struct task_struct *p, *t, *reaper;
kernel/exit.c
712
for_each_thread(p, t) {
kernel/exit.c
713
RCU_INIT_POINTER(t->real_parent, reaper);
kernel/exit.c
714
BUG_ON((!t->ptrace) != (rcu_access_pointer(t->parent) == father));
kernel/exit.c
715
if (likely(!t->ptrace))
kernel/exit.c
716
t->parent = t->real_parent;
kernel/exit.c
717
if (t->pdeath_signal)
kernel/exit.c
718
group_send_sig_info(t->pdeath_signal,
kernel/exit.c
719
SEND_SIG_NOINFO, t,
kernel/fork.c
3274
struct ctl_table t;
kernel/fork.c
3280
t = *table;
kernel/fork.c
3281
t.data = &threads;
kernel/fork.c
3282
t.extra1 = &min;
kernel/fork.c
3283
t.extra2 = &max;
kernel/fork.c
3285
ret = proc_dointvec_minmax(&t, write, buffer, lenp, ppos);
kernel/freezer.c
218
struct task_struct *t;
kernel/freezer.c
221
for_each_thread(p, t) {
kernel/freezer.c
222
__thaw_task(t);
kernel/futex/syscalls.c
175
futex_init_timeout(u32 cmd, u32 op, struct timespec64 *ts, ktime_t *t)
kernel/futex/syscalls.c
180
*t = timespec64_to_ktime(*ts);
kernel/futex/syscalls.c
182
*t = ktime_add_safe(ktime_get(), *t);
kernel/futex/syscalls.c
184
*t = timens_ktime_to_host(CLOCK_MONOTONIC, *t);
kernel/futex/syscalls.c
193
ktime_t t, *tp = NULL;
kernel/futex/syscalls.c
201
ret = futex_init_timeout(cmd, op, &ts, &t);
kernel/futex/syscalls.c
204
tp = &t;
kernel/futex/syscalls.c
511
ktime_t t, *tp = NULL;
kernel/futex/syscalls.c
517
ret = futex_init_timeout(cmd, op, &ts, &t);
kernel/futex/syscalls.c
520
tp = &t;
kernel/hung_task.c
103
static bool task_is_hung(struct task_struct *t, unsigned long timeout)
kernel/hung_task.c
105
unsigned long switch_count = t->nvcsw + t->nivcsw;
kernel/hung_task.c
106
unsigned int state = READ_ONCE(t->__state);
kernel/hung_task.c
125
if (switch_count != t->last_switch_count) {
kernel/hung_task.c
126
t->last_switch_count = switch_count;
kernel/hung_task.c
127
t->last_switch_time = jiffies;
kernel/hung_task.c
130
if (time_is_after_jiffies(t->last_switch_time + timeout * HZ))
kernel/hung_task.c
139
struct task_struct *g, *t;
kernel/hung_task.c
194
for_each_process_thread(g, t) {
kernel/hung_task.c
195
if ((unsigned long)t != owner)
kernel/hung_task.c
201
task->comm, task->pid, t->comm, t->pid);
kernel/hung_task.c
205
task->comm, task->pid, t->comm, t->pid);
kernel/hung_task.c
210
task->comm, task->pid, rwsem_blocked_as, t->comm,
kernel/hung_task.c
211
t->pid, rwsem_blocked_by);
kernel/hung_task.c
215
if (!task_is_hung(t, timeout))
kernel/hung_task.c
216
sched_show_task(t);
kernel/hung_task.c
226
static void check_hung_task(struct task_struct *t, unsigned long timeout,
kernel/hung_task.c
231
if (!task_is_hung(t, timeout))
kernel/hung_task.c
241
trace_sched_process_hang(t);
kernel/hung_task.c
256
t->comm, t->pid, (jiffies - t->last_switch_time) / HZ);
kernel/hung_task.c
261
if (t->flags & PF_POSTCOREDUMP)
kernel/hung_task.c
265
sched_show_task(t);
kernel/hung_task.c
266
debug_show_blocker(t, timeout);
kernel/hung_task.c
282
static bool rcu_lock_break(struct task_struct *g, struct task_struct *t)
kernel/hung_task.c
287
get_task_struct(t);
kernel/hung_task.c
291
can_cont = pid_alive(g) && pid_alive(t);
kernel/hung_task.c
292
put_task_struct(t);
kernel/hung_task.c
307
struct task_struct *g, *t;
kernel/hung_task.c
321
for_each_process_thread(g, t) {
kernel/hung_task.c
326
if (!rcu_lock_break(g, t))
kernel/hung_task.c
331
check_hung_task(t, timeout, prev_detect_count);
kernel/hung_task.c
506
long t;
kernel/hung_task.c
511
t = hung_timeout_jiffies(hung_last_checked, interval);
kernel/hung_task.c
512
if (t <= 0) {
kernel/hung_task.c
519
schedule_timeout_interruptible(t);
kernel/irq/manage.c
1404
struct task_struct *t;
kernel/irq/manage.c
1407
t = kthread_create(irq_thread, new, "irq/%d-%s", irq,
kernel/irq/manage.c
1410
t = kthread_create(irq_thread, new, "irq/%d-s-%s", irq,
kernel/irq/manage.c
1414
if (IS_ERR(t))
kernel/irq/manage.c
1415
return PTR_ERR(t);
kernel/irq/manage.c
1422
new->thread = get_task_struct(t);
kernel/irq/manage.c
1432
kthread_bind_mask(t, cpu_possible_mask);
kernel/irq/manage.c
1838
struct task_struct *t = new->thread;
kernel/irq/manage.c
1841
kthread_stop_put(t);
kernel/irq/manage.c
1844
struct task_struct *t = new->secondary->thread;
kernel/irq/manage.c
1847
kthread_stop_put(t);
kernel/kallsyms_selftest.c
158
u64 t0, t1, t;
kernel/kallsyms_selftest.c
165
t = t1 - t0;
kernel/kallsyms_selftest.c
166
if (t < stat->min)
kernel/kallsyms_selftest.c
167
stat->min = t;
kernel/kallsyms_selftest.c
169
if (t > stat->max)
kernel/kallsyms_selftest.c
170
stat->max = t;
kernel/kallsyms_selftest.c
173
stat->sum += t;
kernel/kallsyms_selftest.c
436
struct task_struct *t;
kernel/kallsyms_selftest.c
438
t = kthread_run_on_cpu(test_entry, NULL, 0, "kallsyms_test");
kernel/kallsyms_selftest.c
439
if (IS_ERR(t)) {
kernel/kallsyms_selftest.c
441
return PTR_ERR(t);
kernel/kcov.c
1019
struct task_struct *t = current;
kernel/kcov.c
1032
mode = READ_ONCE(t->kcov_mode);
kernel/kcov.c
1042
if (in_serving_softirq() && !t->kcov_softirq) {
kernel/kcov.c
1047
if (WARN_ON(!in_serving_softirq() && t->kcov_softirq)) {
kernel/kcov.c
1052
kcov = t->kcov;
kernel/kcov.c
1053
area = t->kcov_area;
kernel/kcov.c
1054
size = t->kcov_size;
kernel/kcov.c
1055
sequence = t->kcov_sequence;
kernel/kcov.c
1057
kcov_stop(t);
kernel/kcov.c
1059
t->kcov_softirq = 0;
kernel/kcov.c
1060
kcov_remote_softirq_stop(t);
kernel/kcov.c
174
static notrace bool check_kcov_mode(enum kcov_mode needed_mode, struct task_struct *t)
kernel/kcov.c
183
if (!in_task() && !(in_softirq_really() && t->kcov_softirq))
kernel/kcov.c
185
mode = READ_ONCE(t->kcov_mode);
kernel/kcov.c
211
struct task_struct *t;
kernel/kcov.c
216
t = current;
kernel/kcov.c
217
if (!check_kcov_mode(KCOV_MODE_TRACE_PC, t))
kernel/kcov.c
220
area = t->kcov_area;
kernel/kcov.c
223
if (likely(pos < t->kcov_size)) {
kernel/kcov.c
241
struct task_struct *t;
kernel/kcov.c
245
t = current;
kernel/kcov.c
246
if (!check_kcov_mode(KCOV_MODE_TRACE_CMP, t))
kernel/kcov.c
255
area = (u64 *)t->kcov_area;
kernel/kcov.c
256
max_pos = t->kcov_size * sizeof(unsigned long);
kernel/kcov.c
356
static void kcov_start(struct task_struct *t, struct kcov *kcov,
kernel/kcov.c
360
kcov_debug("t = %px, size = %u, area = %px\n", t, size, area);
kernel/kcov.c
361
t->kcov = kcov;
kernel/kcov.c
363
t->kcov_size = size;
kernel/kcov.c
364
t->kcov_area = area;
kernel/kcov.c
365
t->kcov_sequence = sequence;
kernel/kcov.c
368
WRITE_ONCE(t->kcov_mode, mode);
kernel/kcov.c
371
static void kcov_stop(struct task_struct *t)
kernel/kcov.c
373
WRITE_ONCE(t->kcov_mode, KCOV_MODE_DISABLED);
kernel/kcov.c
375
t->kcov = NULL;
kernel/kcov.c
376
t->kcov_size = 0;
kernel/kcov.c
377
t->kcov_area = NULL;
kernel/kcov.c
380
static void kcov_task_reset(struct task_struct *t)
kernel/kcov.c
382
kcov_stop(t);
kernel/kcov.c
383
t->kcov_sequence = 0;
kernel/kcov.c
384
t->kcov_handle = 0;
kernel/kcov.c
387
void kcov_task_init(struct task_struct *t)
kernel/kcov.c
389
kcov_task_reset(t);
kernel/kcov.c
390
t->kcov_handle = current->kcov_handle;
kernel/kcov.c
396
kcov->t = NULL;
kernel/kcov.c
423
static void kcov_disable(struct task_struct *t, struct kcov *kcov)
kernel/kcov.c
426
kcov_task_reset(t);
kernel/kcov.c
450
void kcov_task_exit(struct task_struct *t)
kernel/kcov.c
455
kcov = t->kcov;
kernel/kcov.c
460
kcov_debug("t = %px, kcov->t = %px\n", t, kcov->t);
kernel/kcov.c
483
if (WARN_ON(kcov->t != t)) {
kernel/kcov.c
488
kcov_disable(t, kcov);
kernel/kcov.c
598
struct task_struct *t;
kernel/kcov.c
615
t = current;
kernel/kcov.c
616
if (kcov->t != NULL || t->kcov != NULL)
kernel/kcov.c
623
kcov_start(t, kcov, kcov->size, kcov->area, kcov->mode,
kernel/kcov.c
625
kcov->t = t;
kernel/kcov.c
634
t = current;
kernel/kcov.c
635
if (WARN_ON(kcov->t != t))
kernel/kcov.c
637
kcov_disable(t, kcov);
kernel/kcov.c
64
struct task_struct *t __guarded_by(&lock);
kernel/kcov.c
643
t = current;
kernel/kcov.c
644
if (kcov->t != NULL || t->kcov != NULL)
kernel/kcov.c
654
t->kcov = kcov;
kernel/kcov.c
655
t->kcov_mode = KCOV_MODE_REMOTE;
kernel/kcov.c
656
kcov->t = t;
kernel/kcov.c
665
kcov_disable(t, kcov);
kernel/kcov.c
672
kcov_disable(t, kcov);
kernel/kcov.c
681
kcov_disable(t, kcov);
kernel/kcov.c
689
kcov_disable(t, kcov);
kernel/kcov.c
692
t->kcov_handle = remote_arg->common_handle;
kernel/kcov.c
826
static void kcov_remote_softirq_start(struct task_struct *t)
kernel/kcov.c
832
mode = READ_ONCE(t->kcov_mode);
kernel/kcov.c
836
data->saved_size = t->kcov_size;
kernel/kcov.c
837
data->saved_area = t->kcov_area;
kernel/kcov.c
838
data->saved_sequence = t->kcov_sequence;
kernel/kcov.c
839
data->saved_kcov = t->kcov;
kernel/kcov.c
840
kcov_stop(t);
kernel/kcov.c
844
static void kcov_remote_softirq_stop(struct task_struct *t)
kernel/kcov.c
850
kcov_start(t, data->saved_kcov, data->saved_size,
kernel/kcov.c
863
struct task_struct *t = current;
kernel/kcov.c
883
mode = READ_ONCE(t->kcov_mode);
kernel/kcov.c
893
if (WARN_ON(in_serving_softirq() && t->kcov_softirq)) {
kernel/kcov.c
942
kcov_remote_softirq_start(t);
kernel/kcov.c
943
t->kcov_softirq = 1;
kernel/kcov.c
945
kcov_start(t, kcov, size, area, mode, sequence);
kernel/kthread.c
1223
void kthread_delayed_work_timer_fn(struct timer_list *t)
kernel/kthread.c
1225
struct kthread_delayed_work *dwork = timer_container_of(dwork, t,
kernel/module/main.c
2734
enum mod_mem_type t = MOD_MEM_NUM_TYPES;
kernel/module/main.c
2745
t = type;
kernel/module/main.c
2812
while (t--)
kernel/module/main.c
2813
module_memory_free(mod, t);
kernel/panic.c
105
t = *table;
kernel/panic.c
106
t.data = &tmptaint;
kernel/panic.c
107
err = proc_doulongvec_minmax(&t, write, buffer, lenp, ppos);
kernel/panic.c
842
const struct taint_flag *t = &taint_flags[i];
kernel/panic.c
844
char c = is_set ? t->c_true : t->c_false;
kernel/panic.c
848
seq_buf_printf(s, "%s[%c]=%s", sep, c, t->desc);
kernel/panic.c
98
struct ctl_table t;
kernel/power/main.c
442
void pm_report_hw_sleep_time(u64 t)
kernel/power/main.c
444
suspend_stats.last_hw_sleep = t;
kernel/power/main.c
445
suspend_stats.total_hw_sleep += t;
kernel/power/main.c
449
void pm_report_max_hw_sleep(u64 t)
kernel/power/main.c
451
suspend_stats.max_hw_sleep = t;
kernel/power/snapshot.c
1248
unsigned int order, t;
kernel/power/snapshot.c
1272
for_each_migratetype_order(order, t) {
kernel/power/snapshot.c
1274
&zone->free_area[order].free_list[t], buddy_list) {
kernel/rcu/rcuscale.c
534
u64 t;
kernel/rcu/rcuscale.c
558
t = ktime_get_mono_fast_ns();
kernel/rcu/rcuscale.c
560
t_rcu_scale_writer_started = t;
kernel/rcu/rcuscale.c
600
t = ktime_get_mono_fast_ns();
kernel/rcu/rcuscale.c
601
*wdp = t - *wdp;
kernel/rcu/rcuscale.c
617
t_rcu_scale_writer_finished = t;
kernel/rcu/rcutorture.c
2485
struct timer_list t;
kernel/rcu/rcutorture.c
2490
timer_setup_on_stack(&t, rcu_torture_timer, 0);
kernel/rcu/rcutorture.c
2494
if (!timer_pending(&t))
kernel/rcu/rcutorture.c
2495
mod_timer(&t, jiffies + 1);
kernel/rcu/rcutorture.c
2509
timer_delete_sync(&t);
kernel/rcu/rcutorture.c
2510
timer_destroy_on_stack(&t);
kernel/rcu/rcutorture.c
2606
ktime_t t;
kernel/rcu/rcutorture.c
2622
t = torture_random(&rtorsup->rtorsu_trs) & 0xfffff; // One per million.
kernel/rcu/rcutorture.c
2623
if (t < 10 * 1000)
kernel/rcu/rcutorture.c
2624
t = 200 * 1000 * 1000;
kernel/rcu/rcutorture.c
2625
hrtimer_start(&rtorsup->rtorsu_hrt, t, HRTIMER_MODE_REL | HRTIMER_MODE_HARD);
kernel/rcu/rcutorture.c
2628
rtorsup->rtorsu_kt = t;
kernel/rcu/rcutorture.c
2951
struct task_struct *t;
kernel/rcu/rcutorture.c
2956
t = boost_tasks[cpu];
kernel/rcu/rcutorture.c
2962
torture_stop_kthread(rcu_torture_boost, t);
kernel/rcu/rcutorture.c
2978
struct task_struct *t;
kernel/rcu/rcutorture.c
2980
t = per_cpu(ksoftirqd, cpu);
kernel/rcu/rcutorture.c
2981
WARN_ON_ONCE(!t);
kernel/rcu/rcutorture.c
2983
sched_setscheduler_nocheck(t, SCHED_FIFO, &sp);
kernel/rcu/rcutorture.c
2986
t = per_cpu(ktimerd, cpu);
kernel/rcu/rcutorture.c
2987
WARN_ON_ONCE(!t);
kernel/rcu/rcutorture.c
2989
sched_setscheduler_nocheck(t, SCHED_FIFO, &sp);
kernel/rcu/rcutorture.c
4623
int t;
kernel/rcu/rcutorture.c
4625
t = cur_ops->stall_dur ? cur_ops->stall_dur() : stutter * HZ;
kernel/rcu/rcutorture.c
4626
firsterr = torture_stutter_init(stutter * HZ, t);
kernel/rcu/srcutree.c
1255
unsigned long t;
kernel/rcu/srcutree.c
1278
t = ktime_get_mono_fast_ns();
kernel/rcu/srcutree.c
1281
time_in_range_open(t, tlast, tlast + exp_holdoff))
kernel/rcu/srcutree.c
80
static void srcu_delay_timer(struct timer_list *t);
kernel/rcu/srcutree.c
862
static void srcu_delay_timer(struct timer_list *t)
kernel/rcu/srcutree.c
864
struct srcu_data *sdp = container_of(t, struct srcu_data, delay_work);
kernel/rcu/tasks.h
1002
if (t == idle_task(cpu) && !rcu_cpu_online(cpu))
kernel/rcu/tasks.h
1009
static void rcu_tasks_pertask(struct task_struct *t, struct list_head *hop)
kernel/rcu/tasks.h
1011
if (t != current && rcu_tasks_is_holdout(t)) {
kernel/rcu/tasks.h
1012
get_task_struct(t);
kernel/rcu/tasks.h
1013
t->rcu_tasks_nvcsw = READ_ONCE(t->nvcsw);
kernel/rcu/tasks.h
1014
WRITE_ONCE(t->rcu_tasks_holdout, true);
kernel/rcu/tasks.h
1015
list_add(&t->rcu_tasks_holdout_list, hop);
kernel/rcu/tasks.h
1054
struct task_struct *t;
kernel/rcu/tasks.h
1059
list_for_each_entry_safe(t, t1, &rtpcp->rtp_exit_list, rcu_tasks_exit_list) {
kernel/rcu/tasks.h
1060
if (list_empty(&t->rcu_tasks_holdout_list))
kernel/rcu/tasks.h
1061
rcu_tasks_pertask(t, hop);
kernel/rcu/tasks.h
1071
list_add(&tmp, &t->rcu_tasks_exit_list);
kernel/rcu/tasks.h
1087
static void check_holdout_task(struct task_struct *t,
kernel/rcu/tasks.h
1092
if (!READ_ONCE(t->rcu_tasks_holdout) ||
kernel/rcu/tasks.h
1093
t->rcu_tasks_nvcsw != READ_ONCE(t->nvcsw) ||
kernel/rcu/tasks.h
1094
!rcu_tasks_is_holdout(t) ||
kernel/rcu/tasks.h
1096
!is_idle_task(t) && READ_ONCE(t->rcu_tasks_idle_cpu) >= 0)) {
kernel/rcu/tasks.h
1097
WRITE_ONCE(t->rcu_tasks_holdout, false);
kernel/rcu/tasks.h
1098
list_del_init(&t->rcu_tasks_holdout_list);
kernel/rcu/tasks.h
1099
put_task_struct(t);
kernel/rcu/tasks.h
1102
rcu_request_urgent_qs_task(t);
kernel/rcu/tasks.h
1109
cpu = task_cpu(t);
kernel/rcu/tasks.h
1111
t, ".I"[is_idle_task(t)],
kernel/rcu/tasks.h
1113
t->rcu_tasks_nvcsw, t->nvcsw, t->rcu_tasks_holdout,
kernel/rcu/tasks.h
1114
data_race(t->rcu_tasks_idle_cpu), cpu);
kernel/rcu/tasks.h
1115
sched_show_task(t);
kernel/rcu/tasks.h
1122
struct task_struct *t, *t1;
kernel/rcu/tasks.h
1124
list_for_each_entry_safe(t, t1, hop, rcu_tasks_holdout_list) {
kernel/rcu/tasks.h
1125
check_holdout_task(t, needreport, firstreport);
kernel/rcu/tasks.h
1290
struct task_struct *t = current;
kernel/rcu/tasks.h
1292
WARN_ON_ONCE(!list_empty(&t->rcu_tasks_exit_list));
kernel/rcu/tasks.h
1295
t->rcu_tasks_exit_cpu = smp_processor_id();
kernel/rcu/tasks.h
1298
list_add(&t->rcu_tasks_exit_list, &rtpcp->rtp_exit_list);
kernel/rcu/tasks.h
1311
struct task_struct *t = current;
kernel/rcu/tasks.h
1313
WARN_ON_ONCE(list_empty(&t->rcu_tasks_exit_list));
kernel/rcu/tasks.h
1314
rtpcp = per_cpu_ptr(rcu_tasks.rtpcpu, t->rcu_tasks_exit_cpu);
kernel/rcu/tasks.h
1316
list_del_init(&t->rcu_tasks_exit_list);
kernel/rcu/tasks.h
18
typedef void (*pertask_func_t)(struct task_struct *t, struct list_head *hop);
kernel/rcu/tasks.h
680
struct task_struct *t;
kernel/rcu/tasks.h
682
t = kthread_run(rcu_tasks_kthread, rtp, "%s_kthread", rtp->kname);
kernel/rcu/tasks.h
683
if (WARN_ONCE(IS_ERR(t), "%s: Could not start %s grace-period kthread, OOM is now expected behavior\n", __func__, rtp->name))
kernel/rcu/tasks.h
816
struct task_struct *t;
kernel/rcu/tasks.h
830
for_each_process_thread(g, t)
kernel/rcu/tasks.h
831
rtp->pertask_func(t, &holdouts);
kernel/rcu/tasks.h
974
static bool rcu_tasks_is_holdout(struct task_struct *t)
kernel/rcu/tasks.h
979
if (!READ_ONCE(t->on_rq))
kernel/rcu/tasks.h
996
if (is_idle_task(t))
kernel/rcu/tasks.h
999
cpu = task_cpu(t);
kernel/rcu/tree.c
1121
struct task_struct *t = READ_ONCE(rcu_state.gp_kthread);
kernel/rcu/tree.c
1123
if ((current == t && !in_hardirq() && !in_serving_softirq()) ||
kernel/rcu/tree.c
1124
!READ_ONCE(rcu_state.gp_flags) || !t)
kernel/rcu/tree.c
2889
static void rcu_wake_cond(struct task_struct *t, int status)
kernel/rcu/tree.c
2895
if (t && (status != RCU_KTHREAD_YIELDING || is_idle_task(current)))
kernel/rcu/tree.c
2896
wake_up_process(t);
kernel/rcu/tree.c
2901
struct task_struct *t;
kernel/rcu/tree.c
2906
t = __this_cpu_read(rcu_data.rcu_cpu_kthread_task);
kernel/rcu/tree.c
2907
if (t != NULL && t != current)
kernel/rcu/tree.c
2908
rcu_wake_cond(t, __this_cpu_read(rcu_data.rcu_cpu_kthread_status));
kernel/rcu/tree.c
4161
static void rcu_thread_affine_rnp(struct task_struct *t, struct rcu_node *rnp)
kernel/rcu/tree.c
4172
kthread_affine_preferred(t, affinity);
kernel/rcu/tree.c
4583
struct task_struct *t;
kernel/rcu/tree.c
4587
t = kthread_create(rcu_gp_kthread, NULL, "%s", rcu_state.name);
kernel/rcu/tree.c
4588
if (WARN_ONCE(IS_ERR(t), "%s: Could not start grace-period kthread, OOM is now expected behavior\n", __func__))
kernel/rcu/tree.c
4592
sched_setscheduler_nocheck(t, SCHED_FIFO, &sp);
kernel/rcu/tree.c
4599
smp_store_release(&rcu_state.gp_kthread, t); /* ^^^ */
kernel/rcu/tree.c
4601
wake_up_process(t);
kernel/rcu/tree.c
765
void rcu_request_urgent_qs_task(struct task_struct *t)
kernel/rcu/tree.c
770
cpu = task_cpu(t);
kernel/rcu/tree.c
771
if (!task_curr(t))
kernel/rcu/tree.h
497
static bool rcu_preempt_need_deferred_qs(struct task_struct *t);
kernel/rcu/tree_exp.h
542
int t;
kernel/rcu/tree_exp.h
545
t = swait_event_timeout_exclusive(rcu_state.expedited_wq,
kernel/rcu/tree_exp.h
549
if (t > 0 || sync_rcu_exp_done_unlocked(rnp_root))
kernel/rcu/tree_exp.h
551
WARN_ON(t < 0); /* workqueues should not be signaled. */
kernel/rcu/tree_exp.h
755
struct task_struct *t = current;
kernel/rcu/tree_exp.h
796
t->rcu_read_unlock_special.b.exp_hint = true;
kernel/rcu/tree_exp.h
815
struct task_struct *t;
kernel/rcu/tree_exp.h
822
t = list_entry(rnp->exp_tasks->prev,
kernel/rcu/tree_exp.h
824
list_for_each_entry_continue(t, &rnp->blkd_tasks, rcu_node_entry) {
kernel/rcu/tree_exp.h
825
pr_cont(" P%d", t->pid);
kernel/rcu/tree_exp.h
840
struct task_struct *t;
kernel/rcu/tree_exp.h
849
t = list_entry(rnp->exp_tasks->prev,
kernel/rcu/tree_exp.h
851
list_for_each_entry_continue(t, &rnp->blkd_tasks, rcu_node_entry) {
kernel/rcu/tree_exp.h
857
sched_show_task(t);
kernel/rcu/tree_nocb.h
1351
struct task_struct *t;
kernel/rcu/tree_nocb.h
1366
t = kthread_run(rcu_nocb_gp_kthread, rdp_gp,
kernel/rcu/tree_nocb.h
1368
if (WARN_ONCE(IS_ERR(t), "%s: Could not start rcuo GP kthread, OOM is now expected behavior\n", __func__)) {
kernel/rcu/tree_nocb.h
1372
WRITE_ONCE(rdp_gp->nocb_gp_kthread, t);
kernel/rcu/tree_nocb.h
1374
sched_setscheduler_nocheck(t, SCHED_FIFO, &sp);
kernel/rcu/tree_nocb.h
1379
t = kthread_create(rcu_nocb_cb_kthread, rdp,
kernel/rcu/tree_nocb.h
1381
if (WARN_ONCE(IS_ERR(t), "%s: Could not start rcuo CB kthread, OOM is now expected behavior\n", __func__))
kernel/rcu/tree_nocb.h
1385
wake_up_process(t);
kernel/rcu/tree_nocb.h
1387
kthread_park(t);
kernel/rcu/tree_nocb.h
1390
sched_setscheduler_nocheck(t, SCHED_FIFO, &sp);
kernel/rcu/tree_nocb.h
1392
WRITE_ONCE(rdp->nocb_cb_kthread, t);
kernel/rcu/tree_nocb.h
537
struct task_struct *t;
kernel/rcu/tree_nocb.h
540
t = READ_ONCE(rdp->nocb_gp_kthread);
kernel/rcu/tree_nocb.h
541
if (rcu_nocb_poll || !t) {
kernel/rcu/tree_nocb.h
960
static void do_nocb_deferred_wakeup_timer(struct timer_list *t)
kernel/rcu/tree_nocb.h
963
struct rcu_data *rdp = timer_container_of(rdp, t, nocb_timer);
kernel/rcu/tree_plugin.h
1032
static notrace bool rcu_preempt_need_deferred_qs(struct task_struct *t)
kernel/rcu/tree_plugin.h
1044
notrace void rcu_preempt_deferred_qs(struct task_struct *t)
kernel/rcu/tree_plugin.h
1157
struct task_struct *t;
kernel/rcu/tree_plugin.h
117
static void rcu_read_unlock_special(struct task_struct *t);
kernel/rcu/tree_plugin.h
1202
t = container_of(tb, struct task_struct, rcu_node_entry);
kernel/rcu/tree_plugin.h
1203
rt_mutex_init_proxy_locked(&rnp->boost_mtx.rtmutex, t);
kernel/rcu/tree_plugin.h
1303
struct task_struct *t;
kernel/rcu/tree_plugin.h
1308
t = kthread_create(rcu_boost_kthread, (void *)rnp,
kernel/rcu/tree_plugin.h
1310
if (WARN_ON_ONCE(IS_ERR(t)))
kernel/rcu/tree_plugin.h
1314
rnp->boost_kthread_task = t;
kernel/rcu/tree_plugin.h
1318
sched_setscheduler_nocheck(t, SCHED_FIFO, &sp);
kernel/rcu/tree_plugin.h
1319
rcu_thread_affine_rnp(t, rnp);
kernel/rcu/tree_plugin.h
1320
wake_up_process(t); /* get to TASK_INTERRUPTIBLE quickly. */
kernel/rcu/tree_plugin.h
169
struct task_struct *t = current;
kernel/rcu/tree_plugin.h
196
list_add(&t->rcu_node_entry, &rnp->blkd_tasks);
kernel/rcu/tree_plugin.h
214
list_add_tail(&t->rcu_node_entry, &rnp->blkd_tasks);
kernel/rcu/tree_plugin.h
227
list_add(&t->rcu_node_entry, rnp->exp_tasks);
kernel/rcu/tree_plugin.h
238
list_add(&t->rcu_node_entry, rnp->gp_tasks);
kernel/rcu/tree_plugin.h
255
WRITE_ONCE(rnp->gp_tasks, &t->rcu_node_entry);
kernel/rcu/tree_plugin.h
259
WRITE_ONCE(rnp->exp_tasks, &t->rcu_node_entry);
kernel/rcu/tree_plugin.h
326
struct task_struct *t = current;
kernel/rcu/tree_plugin.h
334
!t->rcu_read_unlock_special.b.blocked) {
kernel/rcu/tree_plugin.h
339
t->rcu_read_unlock_special.b.blocked = true;
kernel/rcu/tree_plugin.h
340
t->rcu_blocked_node = rnp;
kernel/rcu/tree_plugin.h
348
WARN_ON_ONCE(!list_empty(&t->rcu_node_entry));
kernel/rcu/tree_plugin.h
350
t->pid,
kernel/rcu/tree_plugin.h
356
rcu_preempt_deferred_qs(t);
kernel/rcu/tree_plugin.h
432
struct task_struct *t = current;
kernel/rcu/tree_plugin.h
437
if (unlikely(READ_ONCE(t->rcu_read_unlock_special.s)))
kernel/rcu/tree_plugin.h
438
rcu_read_unlock_special(t);
kernel/rcu/tree_plugin.h
452
static struct list_head *rcu_next_node_entry(struct task_struct *t,
kernel/rcu/tree_plugin.h
457
np = t->rcu_node_entry.next;
kernel/rcu/tree_plugin.h
478
rcu_preempt_deferred_qs_irqrestore(struct task_struct *t, unsigned long flags)
kernel/rcu/tree_plugin.h
498
special = t->rcu_read_unlock_special;
kernel/rcu/tree_plugin.h
503
t->rcu_read_unlock_special.s = 0;
kernel/rcu/tree_plugin.h
532
rnp = t->rcu_blocked_node;
kernel/rcu/tree_plugin.h
534
WARN_ON_ONCE(rnp != t->rcu_blocked_node);
kernel/rcu/tree_plugin.h
540
np = rcu_next_node_entry(t, rnp);
kernel/rcu/tree_plugin.h
541
list_del_init(&t->rcu_node_entry);
kernel/rcu/tree_plugin.h
542
t->rcu_blocked_node = NULL;
kernel/rcu/tree_plugin.h
544
rnp->gp_seq, t->pid);
kernel/rcu/tree_plugin.h
545
if (&t->rcu_node_entry == rnp->gp_tasks)
kernel/rcu/tree_plugin.h
547
if (&t->rcu_node_entry == rnp->exp_tasks)
kernel/rcu/tree_plugin.h
551
drop_boost_mutex = rt_mutex_owner(&rnp->boost_mtx.rtmutex) == t;
kernel/rcu/tree_plugin.h
552
if (&t->rcu_node_entry == rnp->boost_tasks)
kernel/rcu/tree_plugin.h
600
static notrace bool rcu_preempt_need_deferred_qs(struct task_struct *t)
kernel/rcu/tree_plugin.h
603
READ_ONCE(t->rcu_read_unlock_special.s)) &&
kernel/rcu/tree_plugin.h
614
notrace void rcu_preempt_deferred_qs(struct task_struct *t)
kernel/rcu/tree_plugin.h
618
if (!rcu_preempt_need_deferred_qs(t))
kernel/rcu/tree_plugin.h
621
rcu_preempt_deferred_qs_irqrestore(t, flags);
kernel/rcu/tree_plugin.h
668
static bool rcu_unlock_needs_exp_handling(struct task_struct *t,
kernel/rcu/tree_plugin.h
681
if (t->rcu_blocked_node && READ_ONCE(t->rcu_blocked_node->exp_tasks))
kernel/rcu/tree_plugin.h
702
((rdp->grpmask & READ_ONCE(rnp->qsmask)) || t->rcu_blocked_node))
kernel/rcu/tree_plugin.h
714
if (IS_ENABLED(CONFIG_RCU_BOOST) && irqs_were_disabled && t->rcu_blocked_node)
kernel/rcu/tree_plugin.h
725
static void rcu_read_unlock_special(struct task_struct *t)
kernel/rcu/tree_plugin.h
743
needs_exp = rcu_unlock_needs_exp_handling(t, rdp, rnp, irqs_were_disabled);
kernel/rcu/tree_plugin.h
772
rcu_preempt_deferred_qs_irqrestore(t, flags);
kernel/rcu/tree_plugin.h
786
struct task_struct *t;
kernel/rcu/tree_plugin.h
795
t = container_of(rnp->gp_tasks, struct task_struct,
kernel/rcu/tree_plugin.h
798
rnp->gp_seq, t->pid);
kernel/rcu/tree_plugin.h
812
struct task_struct *t = current;
kernel/rcu/tree_plugin.h
818
if (rcu_preempt_need_deferred_qs(t))
kernel/rcu/tree_plugin.h
820
} else if (rcu_preempt_need_deferred_qs(t)) {
kernel/rcu/tree_plugin.h
821
rcu_preempt_deferred_qs(t); /* Report deferred QS. */
kernel/rcu/tree_plugin.h
832
!t->rcu_read_unlock_special.b.need_qs &&
kernel/rcu/tree_plugin.h
834
t->rcu_read_unlock_special.b.need_qs = true;
kernel/rcu/tree_plugin.h
847
struct task_struct *t = current;
kernel/rcu/tree_plugin.h
852
WRITE_ONCE(t->rcu_read_unlock_special.b.blocked, true);
kernel/rcu/tree_stall.h
277
struct task_struct *t;
kernel/rcu/tree_stall.h
284
t = list_entry(rnp->gp_tasks->prev,
kernel/rcu/tree_stall.h
286
list_for_each_entry_continue(t, &rnp->blkd_tasks, rcu_node_entry) {
kernel/rcu/tree_stall.h
292
sched_show_task(t);
kernel/rcu/tree_stall.h
308
static int check_slow_task(struct task_struct *t, void *arg)
kernel/rcu/tree_stall.h
312
if (task_curr(t))
kernel/rcu/tree_stall.h
314
rscrp->nesting = t->rcu_read_lock_nesting;
kernel/rcu/tree_stall.h
315
rscrp->rs = t->rcu_read_unlock_special;
kernel/rcu/tree_stall.h
316
rscrp->on_blkd_list = !list_empty(&t->rcu_node_entry);
kernel/rcu/tree_stall.h
330
struct task_struct *t;
kernel/rcu/tree_stall.h
340
t = list_entry(rnp->gp_tasks->prev,
kernel/rcu/tree_stall.h
342
list_for_each_entry_continue(t, &rnp->blkd_tasks, rcu_node_entry) {
kernel/rcu/tree_stall.h
343
get_task_struct(t);
kernel/rcu/tree_stall.h
344
ts[i++] = t;
kernel/rcu/tree_stall.h
350
t = ts[--i];
kernel/rcu/tree_stall.h
351
if (task_call_func(t, check_slow_task, &rscr))
kernel/rcu/tree_stall.h
352
pr_cont(" P%d", t->pid);
kernel/rcu/tree_stall.h
355
t->pid, rscr.nesting,
kernel/rcu/tree_stall.h
361
put_task_struct(t);
kernel/rcu/tree_stall.h
945
struct task_struct *t = READ_ONCE(rcu_state.gp_kthread);
kernel/rcu/tree_stall.h
955
t ? data_race(READ_ONCE(t->__state)) : 0x1ffff, t ? t->rt_priority : 0xffU,
kernel/rseq.c
115
void __rseq_trace_update(struct task_struct *t)
kernel/rseq.c
117
trace_rseq_update(t);
kernel/rseq.c
239
static bool rseq_set_ids(struct task_struct *t, struct rseq_ids *ids, u32 node_id)
kernel/rseq.c
241
return rseq_set_ids_get_csaddr(t, ids, node_id, NULL);
kernel/rseq.c
244
static bool rseq_handle_cs(struct task_struct *t, struct pt_regs *regs)
kernel/rseq.c
246
struct rseq __user *urseq = t->rseq.usrptr;
kernel/rseq.c
253
return rseq_update_user_cs(t, regs, csaddr);
kernel/rseq.c
266
struct task_struct *t = current;
kernel/rseq.c
271
if (unlikely(t->flags & PF_EXITING))
kernel/rseq.c
295
event = t->rseq.event.sched_switch;
kernel/rseq.c
296
t->rseq.event.all &= evt_mask.all;
kernel/rseq.c
297
ids.cpu_id = task_cpu(t);
kernel/rseq.c
298
ids.mm_cid = task_mm_cid(t);
kernel/rseq.c
306
if (unlikely(!rseq_update_usr(t, regs, &ids, node_id))) {
kernel/rseq.c
311
t->rseq.event.error = 0;
kernel/rseq.c
360
struct task_struct *t = current;
kernel/rseq.c
363
if (!t->rseq.event.has_rseq)
kernel/rseq.c
365
if (get_user(csaddr, &t->rseq.usrptr->rseq_cs))
kernel/rseq.c
371
if (rseq_debug_update_user_cs(t, regs, csaddr))
kernel/sched/autogroup.c
163
struct task_struct *t;
kernel/sched/autogroup.c
187
for_each_thread(p, t)
kernel/sched/autogroup.c
188
sched_move_task(t, true);
kernel/sched/completion.c
221
long t = wait_for_common(x, MAX_SCHEDULE_TIMEOUT, TASK_INTERRUPTIBLE);
kernel/sched/completion.c
223
if (t == -ERESTARTSYS)
kernel/sched/completion.c
224
return t;
kernel/sched/completion.c
259
long t = wait_for_common(x, MAX_SCHEDULE_TIMEOUT, TASK_KILLABLE);
kernel/sched/completion.c
261
if (t == -ERESTARTSYS)
kernel/sched/completion.c
262
return t;
kernel/sched/completion.c
269
long t = wait_for_common(x, MAX_SCHEDULE_TIMEOUT, state);
kernel/sched/completion.c
271
if (t == -ERESTARTSYS)
kernel/sched/completion.c
272
return t;
kernel/sched/core.c
10569
static inline void mm_cid_transit_to_task(struct task_struct *t, struct mm_cid_pcpu *pcp)
kernel/sched/core.c
10571
if (cid_on_cpu(t->mm_cid.cid)) {
kernel/sched/core.c
10572
unsigned int cid = cpu_cid_to_cid(t->mm_cid.cid);
kernel/sched/core.c
10574
t->mm_cid.cid = cid_to_transit_cid(cid);
kernel/sched/core.c
10575
pcp->cid = t->mm_cid.cid;
kernel/sched/core.c
10615
static inline void mm_cid_transit_to_cpu(struct task_struct *t, struct mm_cid_pcpu *pcp)
kernel/sched/core.c
10617
if (cid_on_task(t->mm_cid.cid)) {
kernel/sched/core.c
10618
t->mm_cid.cid = cid_to_transit_cid(t->mm_cid.cid);
kernel/sched/core.c
10619
pcp->cid = t->mm_cid.cid;
kernel/sched/core.c
10623
static void mm_cid_fixup_task_to_cpu(struct task_struct *t, struct mm_struct *mm)
kernel/sched/core.c
10626
guard(task_rq_lock)(t);
kernel/sched/core.c
10627
if (cid_on_task(t->mm_cid.cid)) {
kernel/sched/core.c
10629
if (task_rq(t)->curr == t)
kernel/sched/core.c
10630
mm_cid_transit_to_cpu(t, per_cpu_ptr(mm->mm_cid.pcpu, task_cpu(t)));
kernel/sched/core.c
10632
mm_unset_cid_on_task(t);
kernel/sched/core.c
10639
struct task_struct *t;
kernel/sched/core.c
10643
hlist_for_each_entry(t, &mm->mm_cid.user_list, mm_cid.node) {
kernel/sched/core.c
10645
if (t != current)
kernel/sched/core.c
10646
mm_cid_fixup_task_to_cpu(t, mm);
kernel/sched/core.c
10652
static bool sched_mm_cid_add_user(struct task_struct *t, struct mm_struct *mm)
kernel/sched/core.c
10656
t->mm_cid.active = 1;
kernel/sched/core.c
10657
hlist_add_head(&t->mm_cid.node, &mm->mm_cid.user_list);
kernel/sched/core.c
10662
static void sched_mm_cid_fork(struct task_struct *t)
kernel/sched/core.c
10664
struct mm_struct *mm = t->mm;
kernel/sched/core.c
10670
WARN_ON_ONCE(t->mm_cid.cid != MM_CID_UNSET);
kernel/sched/core.c
10678
sched_mm_cid_add_user(t, mm);
kernel/sched/core.c
10679
t->mm_cid.cid = mm_get_cid(mm);
kernel/sched/core.c
10681
pcp->cid = t->mm_cid.cid;
kernel/sched/core.c
10685
if (!sched_mm_cid_add_user(t, mm)) {
kernel/sched/core.c
10687
t->mm_cid.cid = mm_get_cid(mm);
kernel/sched/core.c
10703
t->mm_cid.cid = mm_get_cid(mm);
kernel/sched/core.c
10707
static bool sched_mm_cid_remove_user(struct task_struct *t)
kernel/sched/core.c
10709
lockdep_assert_held(&t->mm->mm_cid.lock);
kernel/sched/core.c
10711
t->mm_cid.active = 0;
kernel/sched/core.c
10713
t->mm_cid.cid = cid_from_transit_cid(t->mm_cid.cid);
kernel/sched/core.c
10714
mm_unset_cid_on_task(t);
kernel/sched/core.c
10715
hlist_del_init(&t->mm_cid.node);
kernel/sched/core.c
10716
t->mm->mm_cid.users--;
kernel/sched/core.c
10717
return mm_update_max_cids(t->mm);
kernel/sched/core.c
10720
static bool __sched_mm_cid_exit(struct task_struct *t)
kernel/sched/core.c
10722
struct mm_struct *mm = t->mm;
kernel/sched/core.c
10724
if (!sched_mm_cid_remove_user(t))
kernel/sched/core.c
10748
void sched_mm_cid_exit(struct task_struct *t)
kernel/sched/core.c
10750
struct mm_struct *mm = t->mm;
kernel/sched/core.c
10752
if (!mm || !t->mm_cid.active)
kernel/sched/core.c
10763
if (!__sched_mm_cid_exit(t))
kernel/sched/core.c
10779
if (t == current)
kernel/sched/core.c
10780
mm_cid_transit_to_task(t, this_cpu_ptr(mm->mm_cid.pcpu));
kernel/sched/core.c
10782
sched_mm_cid_remove_user(t);
kernel/sched/core.c
10799
void sched_mm_cid_before_execve(struct task_struct *t)
kernel/sched/core.c
10801
sched_mm_cid_exit(t);
kernel/sched/core.c
10805
void sched_mm_cid_after_execve(struct task_struct *t)
kernel/sched/core.c
10807
if (t->mm)
kernel/sched/core.c
10808
sched_mm_cid_fork(t);
kernel/sched/core.c
10864
static inline void sched_mm_cid_fork(struct task_struct *t) { }
kernel/sched/core.c
3738
struct task_struct *p, *t;
kernel/sched/core.c
3747
llist_for_each_entry_safe(p, t, llist, wake_entry.llist) {
kernel/sched/core.c
406
int t, i = 0;
kernel/sched/core.c
409
for_each_cpu(t, smt_mask)
kernel/sched/core.c
410
raw_spin_lock_nested(&cpu_rq(t)->__lock, i++);
kernel/sched/core.c
418
int t;
kernel/sched/core.c
420
for_each_cpu(t, smt_mask)
kernel/sched/core.c
421
raw_spin_unlock(&cpu_rq(t)->__lock);
kernel/sched/core.c
428
int cpu, t;
kernel/sched/core.c
4349
struct task_struct *t;
kernel/sched/core.c
4354
t = rcu_dereference(cpu_curr(cpu));
kernel/sched/core.c
4358
return t;
kernel/sched/core.c
441
for_each_cpu(t, smt_mask)
kernel/sched/core.c
442
cpu_rq(t)->core_enabled = enabled;
kernel/sched/core.c
4480
struct ctl_table t;
kernel/sched/core.c
4487
t = *table;
kernel/sched/core.c
4488
t.data = &state;
kernel/sched/core.c
4489
err = proc_dointvec_minmax(&t, write, buffer, lenp, ppos);
kernel/sched/core.c
4549
struct ctl_table t;
kernel/sched/core.c
4556
t = *table;
kernel/sched/core.c
4557
t.data = &state;
kernel/sched/core.c
4558
err = proc_dointvec_minmax(&t, write, buffer, lenp, ppos);
kernel/sched/core.c
4732
static void sched_mm_cid_fork(struct task_struct *t);
kernel/sched/core.c
5967
static inline bool is_task_rq_idle(struct task_struct *t)
kernel/sched/core.c
5969
return (task_rq(t)->idle == t);
kernel/sched/core.c
6364
int t;
kernel/sched/core.c
6375
for_each_cpu(t, smt_mask) {
kernel/sched/core.c
6376
if (t == cpu)
kernel/sched/core.c
6378
rq = cpu_rq(t);
kernel/sched/core.c
6389
for_each_cpu(t, smt_mask) {
kernel/sched/core.c
6390
rq = cpu_rq(t);
kernel/sched/core.c
6392
if (t == cpu)
kernel/sched/core.c
6403
int t;
kernel/sched/core.c
6418
for_each_cpu(t, smt_mask) {
kernel/sched/core.c
6419
if (t == cpu)
kernel/sched/core.c
6421
core_rq = cpu_rq(t);
kernel/sched/core.c
6444
for_each_cpu(t, smt_mask) {
kernel/sched/core.c
6445
rq = cpu_rq(t);
kernel/sched/cputime.c
304
static inline u64 read_sum_exec_runtime(struct task_struct *t)
kernel/sched/cputime.c
306
return t->se.sum_exec_runtime;
kernel/sched/cputime.c
309
static u64 read_sum_exec_runtime(struct task_struct *t)
kernel/sched/cputime.c
315
rq = task_rq_lock(t, &rf);
kernel/sched/cputime.c
316
ns = t->se.sum_exec_runtime;
kernel/sched/cputime.c
317
task_rq_unlock(rq, t, &rf);
kernel/sched/cputime.c
330
struct task_struct *t;
kernel/sched/cputime.c
350
__for_each_thread(sig, t) {
kernel/sched/cputime.c
351
task_cputime(t, &utime, &stime);
kernel/sched/cputime.c
354
times->sum_exec_runtime += read_sum_exec_runtime(t);
kernel/sched/cputime.c
808
void vtime_init_idle(struct task_struct *t, int cpu)
kernel/sched/cputime.c
810
struct vtime *vtime = &t->vtime;
kernel/sched/cputime.c
822
u64 task_gtime(struct task_struct *t)
kernel/sched/cputime.c
824
struct vtime *vtime = &t->vtime;
kernel/sched/cputime.c
829
return t->gtime;
kernel/sched/cputime.c
834
gtime = t->gtime;
kernel/sched/cputime.c
848
bool task_cputime(struct task_struct *t, u64 *utime, u64 *stime)
kernel/sched/cputime.c
850
struct vtime *vtime = &t->vtime;
kernel/sched/cputime.c
856
*utime = t->utime;
kernel/sched/cputime.c
857
*stime = t->stime;
kernel/sched/cputime.c
865
*utime = t->utime;
kernel/sched/cputime.c
866
*stime = t->stime;
kernel/sched/deadline.c
775
static bool dl_entity_overflow(struct sched_dl_entity *dl_se, u64 t);
kernel/sched/deadline.c
914
static bool dl_entity_overflow(struct sched_dl_entity *dl_se, u64 t)
kernel/sched/deadline.c
937
right = ((dl_se->deadline - t) >> DL_SCALE) *
kernel/sched/ext.c
5387
const struct btf_type *t;
kernel/sched/ext.c
5389
t = btf_type_by_id(reg->btf, reg->btf_id);
kernel/sched/ext.c
5390
if (t == task_struct_type) {
kernel/sched/ext.c
5411
static int bpf_scx_init_member(const struct btf_type *t,
kernel/sched/ext.c
5417
u32 moff = __btf_member_bit_offset(t, member) / 8;
kernel/sched/ext.c
5457
static int bpf_scx_check_member(const struct btf_type *t,
kernel/sched/ext.c
5461
u32 moff = __btf_member_bit_offset(t, member) / 8;
kernel/sched/isolation.c
104
void housekeeping_affine(struct task_struct *t, enum hk_type type)
kernel/sched/isolation.c
108
set_cpus_allowed_ptr(t, housekeeping_cpumask(type));
kernel/sched/psi.c
1299
struct psi_trigger *t;
kernel/sched/psi.c
1343
t = kmalloc_obj(*t);
kernel/sched/psi.c
1344
if (!t)
kernel/sched/psi.c
1347
t->group = group;
kernel/sched/psi.c
1348
t->state = state;
kernel/sched/psi.c
1349
t->threshold = threshold_us * NSEC_PER_USEC;
kernel/sched/psi.c
1350
t->win.size = window_us * NSEC_PER_USEC;
kernel/sched/psi.c
1351
window_reset(&t->win, sched_clock(),
kernel/sched/psi.c
1352
group->total[PSI_POLL][t->state], 0);
kernel/sched/psi.c
1354
t->event = 0;
kernel/sched/psi.c
1355
t->last_event_time = 0;
kernel/sched/psi.c
1356
t->of = of;
kernel/sched/psi.c
1358
init_waitqueue_head(&t->event_wait);
kernel/sched/psi.c
1359
t->pending_event = false;
kernel/sched/psi.c
1360
t->aggregator = privileged ? PSI_POLL : PSI_AVGS;
kernel/sched/psi.c
1370
kfree(t);
kernel/sched/psi.c
1379
list_add(&t->node, &group->rtpoll_triggers);
kernel/sched/psi.c
1381
div_u64(t->win.size, UPDATES_PER_WINDOW));
kernel/sched/psi.c
1382
group->rtpoll_nr_triggers[t->state]++;
kernel/sched/psi.c
1383
group->rtpoll_states |= (1 << t->state);
kernel/sched/psi.c
1389
list_add(&t->node, &group->avg_triggers);
kernel/sched/psi.c
1390
group->avg_nr_triggers[t->state]++;
kernel/sched/psi.c
1394
return t;
kernel/sched/psi.c
1397
void psi_trigger_destroy(struct psi_trigger *t)
kernel/sched/psi.c
1406
if (!t)
kernel/sched/psi.c
1409
group = t->group;
kernel/sched/psi.c
1415
if (t->of)
kernel/sched/psi.c
1416
kernfs_notify(t->of->kn);
kernel/sched/psi.c
1418
wake_up_interruptible(&t->event_wait);
kernel/sched/psi.c
1420
if (t->aggregator == PSI_AVGS) {
kernel/sched/psi.c
1422
if (!list_empty(&t->node)) {
kernel/sched/psi.c
1423
list_del(&t->node);
kernel/sched/psi.c
1424
group->avg_nr_triggers[t->state]--;
kernel/sched/psi.c
1429
if (!list_empty(&t->node)) {
kernel/sched/psi.c
1433
list_del(&t->node);
kernel/sched/psi.c
1434
group->rtpoll_nr_triggers[t->state]--;
kernel/sched/psi.c
1435
if (!group->rtpoll_nr_triggers[t->state])
kernel/sched/psi.c
1436
group->rtpoll_states &= ~(1 << t->state);
kernel/sched/psi.c
1441
if (group->rtpoll_min_period == div_u64(t->win.size, UPDATES_PER_WINDOW)) {
kernel/sched/psi.c
1479
kfree(t);
kernel/sched/psi.c
1486
struct psi_trigger *t;
kernel/sched/psi.c
1491
t = smp_load_acquire(trigger_ptr);
kernel/sched/psi.c
1492
if (!t)
kernel/sched/psi.c
1495
if (t->of)
kernel/sched/psi.c
1496
kernfs_generic_poll(t->of, wait);
kernel/sched/psi.c
1498
poll_wait(file, &t->event_wait, wait);
kernel/sched/psi.c
1500
if (cmpxchg(&t->event, 1, 0) == 1)
kernel/sched/psi.c
203
static void poll_timer_fn(struct timer_list *t);
kernel/sched/psi.c
466
struct psi_trigger *t;
kernel/sched/psi.c
483
list_for_each_entry(t, triggers, node) {
kernel/sched/psi.c
487
new_stall = aggregator_total[t->state] != total[t->state];
kernel/sched/psi.c
490
if (!new_stall && !t->pending_event)
kernel/sched/psi.c
500
growth = window_update(&t->win, now, total[t->state]);
kernel/sched/psi.c
501
if (!t->pending_event) {
kernel/sched/psi.c
502
if (growth < t->threshold)
kernel/sched/psi.c
505
t->pending_event = true;
kernel/sched/psi.c
509
if (now < t->last_event_time + t->win.size)
kernel/sched/psi.c
513
if (cmpxchg(&t->event, 0, 1) == 0) {
kernel/sched/psi.c
514
if (t->of)
kernel/sched/psi.c
515
kernfs_notify(t->of->kn);
kernel/sched/psi.c
517
wake_up_interruptible(&t->event_wait);
kernel/sched/psi.c
519
t->last_event_time = now;
kernel/sched/psi.c
521
t->pending_event = false;
kernel/sched/psi.c
615
struct psi_trigger *t;
kernel/sched/psi.c
617
list_for_each_entry(t, &group->rtpoll_triggers, node)
kernel/sched/psi.c
618
window_reset(&t->win, now,
kernel/sched/psi.c
619
group->total[PSI_POLL][t->state], 0);
kernel/sched/psi.c
756
static void poll_timer_fn(struct timer_list *t)
kernel/sched/psi.c
758
struct psi_group *group = timer_container_of(group, t, rtpoll_timer);
kernel/sched/psi.c
801
unsigned int t, m;
kernel/sched/psi.c
827
for (t = 0, m = clear; m; m &= ~(1 << t), t++) {
kernel/sched/psi.c
828
if (!(m & (1 << t)))
kernel/sched/psi.c
830
if (groupc->tasks[t]) {
kernel/sched/psi.c
831
groupc->tasks[t]--;
kernel/sched/psi.c
834
cpu, t, groupc->tasks[0],
kernel/sched/psi.c
841
for (t = 0; set; set &= ~(1 << t), t++)
kernel/sched/psi.c
842
if (set & (1 << t))
kernel/sched/psi.c
843
groupc->tasks[t]++;
kernel/sched/sched.h
1419
static inline void rq_set_donor(struct rq *rq, struct task_struct *t)
kernel/sched/sched.h
1421
rcu_assign_pointer(rq->donor, t);
kernel/sched/sched.h
1424
static inline void rq_set_donor(struct rq *rq, struct task_struct *t)
kernel/sched/sched.h
2053
extern int migrate_swap(struct task_struct *p, struct task_struct *t,
kernel/sched/sched.h
3818
static __always_inline void mm_unset_cid_on_task(struct task_struct *t)
kernel/sched/sched.h
3820
unsigned int cid = t->mm_cid.cid;
kernel/sched/sched.h
3822
t->mm_cid.cid = MM_CID_UNSET;
kernel/sched/sched.h
3824
mm_drop_cid(t->mm, cid);
kernel/sched/sched.h
3877
static __always_inline void mm_cid_update_task_cid(struct task_struct *t, unsigned int cid)
kernel/sched/sched.h
3879
if (t->mm_cid.cid != cid) {
kernel/sched/sched.h
3880
t->mm_cid.cid = cid;
kernel/sched/sched.h
3881
rseq_sched_set_ids_changed(t);
kernel/sched/sched.h
3890
static __always_inline void mm_cid_from_cpu(struct task_struct *t, unsigned int cpu_cid,
kernel/sched/sched.h
3893
unsigned int max_cids, tcid = t->mm_cid.cid;
kernel/sched/sched.h
3894
struct mm_struct *mm = t->mm;
kernel/sched/sched.h
3900
mm_cid_update_task_cid(t, cpu_cid);
kernel/sched/sched.h
3909
mm_unset_cid_on_task(t);
kernel/sched/sched.h
3922
mm_cid_update_task_cid(t, cpu_cid);
kernel/sched/sched.h
3925
static __always_inline void mm_cid_from_task(struct task_struct *t, unsigned int cpu_cid,
kernel/sched/sched.h
3928
unsigned int max_cids, tcid = t->mm_cid.cid;
kernel/sched/sched.h
3929
struct mm_struct *mm = t->mm;
kernel/sched/sched.h
3955
mm_cid_update_task_cid(t, tcid);
kernel/sched/stats.h
246
static inline void sched_info_dequeue(struct rq *rq, struct task_struct *t)
kernel/sched/stats.h
250
if (!t->sched_info.last_queued)
kernel/sched/stats.h
253
delta = rq_clock(rq) - t->sched_info.last_queued;
kernel/sched/stats.h
254
t->sched_info.last_queued = 0;
kernel/sched/stats.h
255
t->sched_info.run_delay += delta;
kernel/sched/stats.h
256
if (delta > t->sched_info.max_run_delay) {
kernel/sched/stats.h
257
t->sched_info.max_run_delay = delta;
kernel/sched/stats.h
258
ktime_get_real_ts64(&t->sched_info.max_run_delay_ts);
kernel/sched/stats.h
260
if (delta && (!t->sched_info.min_run_delay || delta < t->sched_info.min_run_delay))
kernel/sched/stats.h
261
t->sched_info.min_run_delay = delta;
kernel/sched/stats.h
270
static void sched_info_arrive(struct rq *rq, struct task_struct *t)
kernel/sched/stats.h
274
if (!t->sched_info.last_queued)
kernel/sched/stats.h
278
delta = now - t->sched_info.last_queued;
kernel/sched/stats.h
279
t->sched_info.last_queued = 0;
kernel/sched/stats.h
280
t->sched_info.run_delay += delta;
kernel/sched/stats.h
281
t->sched_info.last_arrival = now;
kernel/sched/stats.h
282
t->sched_info.pcount++;
kernel/sched/stats.h
283
if (delta > t->sched_info.max_run_delay) {
kernel/sched/stats.h
284
t->sched_info.max_run_delay = delta;
kernel/sched/stats.h
285
ktime_get_real_ts64(&t->sched_info.max_run_delay_ts);
kernel/sched/stats.h
287
if (delta && (!t->sched_info.min_run_delay || delta < t->sched_info.min_run_delay))
kernel/sched/stats.h
288
t->sched_info.min_run_delay = delta;
kernel/sched/stats.h
298
static inline void sched_info_enqueue(struct rq *rq, struct task_struct *t)
kernel/sched/stats.h
300
if (!t->sched_info.last_queued)
kernel/sched/stats.h
301
t->sched_info.last_queued = rq_clock(rq);
kernel/sched/stats.h
312
static inline void sched_info_depart(struct rq *rq, struct task_struct *t)
kernel/sched/stats.h
314
unsigned long long delta = rq_clock(rq) - t->sched_info.last_arrival;
kernel/sched/stats.h
318
if (task_is_running(t))
kernel/sched/stats.h
319
sched_info_enqueue(rq, t);
kernel/sched/stats.h
343
# define sched_info_enqueue(rq, t) do { } while (0)
kernel/sched/stats.h
344
# define sched_info_dequeue(rq, t) do { } while (0)
kernel/sched/stats.h
345
# define sched_info_switch(rq, t, next) do { } while (0)
kernel/sched/syscalls.c
1510
static int sched_rr_get_interval(pid_t pid, struct timespec64 *t)
kernel/sched/syscalls.c
1534
jiffies_to_timespec64(time_slice, t);
kernel/sched/syscalls.c
1552
struct timespec64 t;
kernel/sched/syscalls.c
1553
int retval = sched_rr_get_interval(pid, &t);
kernel/sched/syscalls.c
1556
retval = put_timespec64(&t, interval);
kernel/sched/syscalls.c
1565
struct timespec64 t;
kernel/sched/syscalls.c
1566
int retval = sched_rr_get_interval(pid, &t);
kernel/sched/syscalls.c
1569
retval = put_old_timespec32(&t, interval);
kernel/signal.c
1005
!sigismember(&t->real_blocked, sig) &&
kernel/signal.c
1020
__for_each_thread(signal, t) {
kernel/signal.c
1021
task_clear_jobctl_pending(t, JOBCTL_PENDING_MASK);
kernel/signal.c
1022
sigaddset(&t->pending.signal, SIGKILL);
kernel/signal.c
1023
signal_wake_up(t, 1);
kernel/signal.c
1033
signal_wake_up(t, sig == SIGKILL);
kernel/signal.c
1043
struct task_struct *t, enum pid_type type, bool force)
kernel/signal.c
1050
lockdep_assert_held(&t->sighand->siglock);
kernel/signal.c
1053
if (!prepare_signal(sig, t, force))
kernel/signal.c
1056
pending = (type != PIDTYPE_PID) ? &t->signal->shared_pending : &t->pending;
kernel/signal.c
106
static bool sig_ignored(struct task_struct *t, int sig, bool force)
kernel/signal.c
1070
if ((sig == SIGKILL) || (t->flags & PF_KTHREAD))
kernel/signal.c
1087
q = sigqueue_alloc(sig, t, GFP_ATOMIC, override_rlimit);
kernel/signal.c
1098
task_active_pid_ns(t));
kernel/signal.c
1101
from_kuid_munged(task_cred_xxx(t, user_ns),
kernel/signal.c
113
if (sigismember(&t->blocked, sig) || sigismember(&t->real_blocked, sig))
kernel/signal.c
1136
signalfd_notify(t, sig);
kernel/signal.c
1142
hlist_for_each_entry(delayed, &t->signal->multiprocess, node) {
kernel/signal.c
1153
complete_signal(sig, t, type);
kernel/signal.c
1155
trace_signal_generate(sig, info, t, type != PIDTYPE_PID, result);
kernel/signal.c
1184
struct task_struct *t, enum pid_type type)
kernel/signal.c
1191
force = !task_pid_nr_ns(current, task_active_pid_ns(t));
kernel/signal.c
1200
t_user_ns = task_cred_xxx(t, user_ns);
kernel/signal.c
121
if (t->ptrace && sig != SIGKILL)
kernel/signal.c
1211
if (!task_pid_nr_ns(current, task_active_pid_ns(t))) {
kernel/signal.c
1216
return __send_signal_locked(sig, info, t, type, force);
kernel/signal.c
124
return sig_task_ignored(t, sig, force);
kernel/signal.c
1294
force_sig_info_to_task(struct kernel_siginfo *info, struct task_struct *t,
kernel/signal.c
1302
spin_lock_irqsave(&t->sighand->siglock, flags);
kernel/signal.c
1303
action = &t->sighand->action[sig-1];
kernel/signal.c
1305
blocked = sigismember(&t->blocked, sig);
kernel/signal.c
1311
sigdelset(&t->blocked, sig);
kernel/signal.c
1318
(!t->ptrace || (handler == HANDLER_EXIT)))
kernel/signal.c
1319
t->signal->flags &= ~SIGNAL_UNKILLABLE;
kernel/signal.c
1320
ret = send_signal_locked(sig, info, t, PIDTYPE_PID);
kernel/signal.c
1322
if (!task_sigpending(t))
kernel/signal.c
1323
signal_wake_up(t, 0);
kernel/signal.c
1324
spin_unlock_irqrestore(&t->sighand->siglock, flags);
kernel/signal.c
1339
struct task_struct *t;
kernel/signal.c
1344
for_other_threads(p, t) {
kernel/signal.c
1345
task_clear_jobctl_pending(t, JOBCTL_PENDING_MASK);
kernel/signal.c
1349
if (t->exit_state)
kernel/signal.c
1351
sigaddset(&t->pending.signal, SIGKILL);
kernel/signal.c
1352
signal_wake_up(t, 1);
kernel/signal.c
159
static bool recalc_sigpending_tsk(struct task_struct *t)
kernel/signal.c
161
if ((t->jobctl & (JOBCTL_PENDING_MASK | JOBCTL_TRAP_FREEZE)) ||
kernel/signal.c
162
PENDING(&t->pending, &t->blocked) ||
kernel/signal.c
163
PENDING(&t->signal->shared_pending, &t->blocked) ||
kernel/signal.c
164
cgroup_task_frozen(t)) {
kernel/signal.c
165
set_tsk_thread_flag(t, TIF_SIGPENDING);
kernel/signal.c
1690
struct task_struct *t)
kernel/signal.c
1699
return force_sig_info_to_task(&info, t, HANDLER_CURRENT);
kernel/signal.c
1707
int send_sig_fault(int sig, int code, void __user *addr, struct task_struct *t)
kernel/signal.c
1716
return send_sig_info(info.si_signo, &info, t);
kernel/signal.c
1733
int send_sig_mceerr(int code, void __user *addr, short lsb, struct task_struct *t)
kernel/signal.c
1744
return send_sig_info(info.si_signo, &info, t);
kernel/signal.c
1861
struct task_struct *t)
kernel/signal.c
1871
return send_sig_info(info.si_signo, &info, t);
kernel/signal.c
1943
static void posixtimer_queue_sigqueue(struct sigqueue *q, struct task_struct *t, enum pid_type type)
kernel/signal.c
1948
signalfd_notify(t, sig);
kernel/signal.c
1949
pending = (type != PIDTYPE_PID) ? &t->signal->shared_pending : &t->pending;
kernel/signal.c
1952
complete_signal(sig, t, type);
kernel/signal.c
1968
struct task_struct *t = pid_task(tmr->it_pid, tmr->it_pid_type);
kernel/signal.c
1970
if (t && tmr->it_pid_type != PIDTYPE_PID &&
kernel/signal.c
1971
same_thread_group(t, current) && !current->exit_state)
kernel/signal.c
1972
t = current;
kernel/signal.c
1973
return t;
kernel/signal.c
1980
struct task_struct *t;
kernel/signal.c
1986
t = posixtimer_get_target(tmr);
kernel/signal.c
1987
if (!t)
kernel/signal.c
1990
if (!likely(lock_task_sighand(t, &flags)))
kernel/signal.c
2006
if (!prepare_signal(sig, t, false)) {
kernel/signal.c
2043
posixtimer_sig_ignore(t, q);
kernel/signal.c
2080
posixtimer_queue_sigqueue(q, t, tmr->it_pid_type);
kernel/signal.c
2083
trace_signal_generate(sig, &q->info, t, tmr->it_pid_type != PIDTYPE_PID, result);
kernel/signal.c
2084
unlock_task_sighand(t, &flags);
kernel/signal.c
2557
struct task_struct *t;
kernel/signal.c
2592
for_other_threads(current, t) {
kernel/signal.c
2598
if (!task_is_stopped(t) &&
kernel/signal.c
2599
task_set_jobctl_pending(t, signr | gstop)) {
kernel/signal.c
2601
if (likely(!(t->ptrace & PT_SEIZED)))
kernel/signal.c
2602
signal_wake_up(t, 0);
kernel/signal.c
2604
ptrace_trap_notify(t);
kernel/signal.c
3093
struct task_struct *t;
kernel/signal.c
3099
for_other_threads(tsk, t) {
kernel/signal.c
3100
if (t->flags & PF_EXITING)
kernel/signal.c
3103
if (!has_pending_signals(&retarget, &t->blocked))
kernel/signal.c
3106
sigandsets(&retarget, &retarget, &t->blocked);
kernel/signal.c
3108
if (!task_sigpending(t))
kernel/signal.c
3109
signal_wake_up(t, 0);
kernel/signal.c
3873
struct timespec64 t;
kernel/signal.c
3884
if (get_timespec64(&t, uts))
kernel/signal.c
3888
ret = do_sigtimedwait(&s, &info, uts ? &t : NULL);
kernel/signal.c
3904
struct timespec64 t;
kernel/signal.c
3915
if (get_old_timespec32(&t, uts))
kernel/signal.c
3919
ret = do_sigtimedwait(&s, &info, uts ? &t : NULL);
kernel/signal.c
402
static struct ucounts *sig_get_ucounts(struct task_struct *t, int sig,
kernel/signal.c
417
ucounts = task_ucounts(t);
kernel/signal.c
424
if (unlikely(!override_rlimit && sigpending > task_rlimit(t, RLIMIT_SIGPENDING))) {
kernel/signal.c
4302
struct task_struct *p = current, *t;
kernel/signal.c
4358
for_each_thread(p, t)
kernel/signal.c
4359
flush_sigqueue_mask(p, &mask, &t->pending);
kernel/signal.c
4390
struct task_struct *t = current;
kernel/signal.c
4395
oss->ss_sp = (void __user *) t->sas_ss_sp;
kernel/signal.c
4396
oss->ss_size = t->sas_ss_size;
kernel/signal.c
4419
if (t->sas_ss_sp == (unsigned long)ss_sp &&
kernel/signal.c
4420
t->sas_ss_size == ss_size &&
kernel/signal.c
4421
t->sas_ss_flags == ss_flags)
kernel/signal.c
4435
t->sas_ss_sp = (unsigned long) ss_sp;
kernel/signal.c
4436
t->sas_ss_size = ss_size;
kernel/signal.c
4437
t->sas_ss_flags = ss_flags;
kernel/signal.c
446
static struct sigqueue *sigqueue_alloc(int sig, struct task_struct *t, gfp_t gfp_flags,
kernel/signal.c
4471
struct task_struct *t = current;
kernel/signal.c
4472
int err = __put_user((void __user *)t->sas_ss_sp, &uss->ss_sp) |
kernel/signal.c
4473
__put_user(t->sas_ss_flags, &uss->ss_flags) |
kernel/signal.c
4474
__put_user(t->sas_ss_size, &uss->ss_size);
kernel/signal.c
449
struct ucounts *ucounts = sig_get_ucounts(t, sig, override_rlimit);
kernel/signal.c
4525
struct task_struct *t = current;
kernel/signal.c
4526
err = __put_user(ptr_to_compat((void __user *)t->sas_ss_sp),
kernel/signal.c
4528
__put_user(t->sas_ss_flags, &uss->ss_flags) |
kernel/signal.c
4529
__put_user(t->sas_ss_size, &uss->ss_size);
kernel/signal.c
493
void flush_signals(struct task_struct *t)
kernel/signal.c
497
spin_lock_irqsave(&t->sighand->siglock, flags);
kernel/signal.c
498
clear_tsk_thread_flag(t, TIF_SIGPENDING);
kernel/signal.c
499
flush_sigqueue(&t->pending);
kernel/signal.c
500
flush_sigqueue(&t->signal->shared_pending);
kernel/signal.c
501
spin_unlock_irqrestore(&t->sighand->siglock, flags);
kernel/signal.c
5020
void kdb_send_sig(struct task_struct *t, int sig)
kernel/signal.c
5024
if (!spin_trylock(&t->sighand->siglock)) {
kernel/signal.c
5030
new_t = kdb_prev_t != t;
kernel/signal.c
5031
kdb_prev_t = t;
kernel/signal.c
5032
if (!task_is_running(t) && new_t) {
kernel/signal.c
5033
spin_unlock(&t->sighand->siglock);
kernel/signal.c
5042
ret = send_signal_locked(sig, SEND_SIG_PRIV, t, PIDTYPE_PID);
kernel/signal.c
5043
spin_unlock(&t->sighand->siglock);
kernel/signal.c
5046
sig, t->pid);
kernel/signal.c
5048
kdb_printf("Signal %d is sent to process %d.\n", sig, t->pid);
kernel/signal.c
505
void ignore_signals(struct task_struct *t)
kernel/signal.c
510
t->sighand->action[i].sa.sa_handler = SIG_IGN;
kernel/signal.c
512
flush_signals(t);
kernel/signal.c
520
flush_signal_handlers(struct task_struct *t, int force_default)
kernel/signal.c
523
struct k_sigaction *ka = &t->sighand->action[0];
kernel/signal.c
72
static void __user *sig_handler(struct task_struct *t, int sig)
kernel/signal.c
721
void signal_wake_up_state(struct task_struct *t, unsigned int state)
kernel/signal.c
723
lockdep_assert_held(&t->sighand->siglock);
kernel/signal.c
725
set_tsk_thread_flag(t, TIF_SIGPENDING);
kernel/signal.c
734
if (!wake_up_state(t, state | TASK_INTERRUPTIBLE))
kernel/signal.c
735
kick_process(t);
kernel/signal.c
74
return t->sighand->action[sig - 1].sa.sa_handler;
kernel/signal.c
783
static bool kill_ok_by_cred(struct task_struct *t)
kernel/signal.c
786
const struct cred *tcred = __task_cred(t);
kernel/signal.c
800
struct task_struct *t)
kernel/signal.c
811
error = audit_signal_info(sig, t); /* Let audit system see the signal */
kernel/signal.c
815
if (!same_thread_group(current, t) &&
kernel/signal.c
816
!kill_ok_by_cred(t)) {
kernel/signal.c
819
sid = task_session(t);
kernel/signal.c
832
return security_task_kill(t, info, sig, NULL);
kernel/signal.c
84
static bool sig_task_ignored(struct task_struct *t, int sig, bool force)
kernel/signal.c
852
static void ptrace_trap_notify(struct task_struct *t)
kernel/signal.c
854
WARN_ON_ONCE(!(t->ptrace & PT_SEIZED));
kernel/signal.c
855
lockdep_assert_held(&t->sighand->siglock);
kernel/signal.c
857
task_set_jobctl_pending(t, JOBCTL_TRAP_NOTIFY);
kernel/signal.c
858
ptrace_signal_wake_up(t, t->jobctl & JOBCTL_LISTENING);
kernel/signal.c
874
struct task_struct *t;
kernel/signal.c
88
handler = sig_handler(t, sig);
kernel/signal.c
890
for_each_thread(p, t)
kernel/signal.c
891
flush_sigqueue_mask(p, &flush, &t->pending);
kernel/signal.c
899
for_each_thread(p, t) {
kernel/signal.c
900
flush_sigqueue_mask(p, &flush, &t->pending);
kernel/signal.c
901
task_clear_jobctl_pending(t, JOBCTL_STOP_PENDING);
kernel/signal.c
902
if (likely(!(t->ptrace & PT_SEIZED))) {
kernel/signal.c
903
t->jobctl &= ~JOBCTL_STOPPED;
kernel/signal.c
904
wake_up_state(t, __TASK_STOPPED);
kernel/signal.c
906
ptrace_trap_notify(t);
kernel/signal.c
91
if (unlikely(is_global_init(t) && sig_kernel_only(sig)))
kernel/signal.c
94
if (unlikely(t->signal->flags & SIGNAL_UNKILLABLE) &&
kernel/signal.c
966
struct task_struct *t;
kernel/signal.c
974
t = p;
kernel/signal.c
985
t = signal->curr_target;
kernel/signal.c
986
while (!wants_signal(sig, t)) {
kernel/signal.c
987
t = next_thread(t);
kernel/signal.c
988
if (t == signal->curr_target)
kernel/signal.c
99
if (unlikely((t->flags & PF_KTHREAD) &&
kernel/signal.c
996
signal->curr_target = t;
kernel/softirq.c
1009
void tasklet_kill(struct tasklet_struct *t)
kernel/softirq.c
1014
wait_on_bit_lock(&t->state, TASKLET_STATE_SCHED, TASK_UNINTERRUPTIBLE);
kernel/softirq.c
1016
tasklet_unlock_wait(t);
kernel/softirq.c
1017
tasklet_clear_sched(t);
kernel/softirq.c
1022
void tasklet_unlock(struct tasklet_struct *t)
kernel/softirq.c
1024
clear_and_wake_up_bit(TASKLET_STATE_RUN, &t->state);
kernel/softirq.c
1028
void tasklet_unlock_wait(struct tasklet_struct *t)
kernel/softirq.c
1030
wait_on_bit(&t->state, TASKLET_STATE_RUN, TASK_UNINTERRUPTIBLE);
kernel/softirq.c
809
static void __tasklet_schedule_common(struct tasklet_struct *t,
kernel/softirq.c
818
t->next = NULL;
kernel/softirq.c
819
*head->tail = t;
kernel/softirq.c
820
head->tail = &(t->next);
kernel/softirq.c
825
void __tasklet_schedule(struct tasklet_struct *t)
kernel/softirq.c
827
__tasklet_schedule_common(t, &tasklet_vec,
kernel/softirq.c
832
void __tasklet_hi_schedule(struct tasklet_struct *t)
kernel/softirq.c
834
__tasklet_schedule_common(t, &tasklet_hi_vec,
kernel/softirq.c
839
static bool tasklet_clear_sched(struct tasklet_struct *t)
kernel/softirq.c
841
if (test_and_clear_wake_up_bit(TASKLET_STATE_SCHED, &t->state))
kernel/softirq.c
845
t->use_callback ? "callback" : "func",
kernel/softirq.c
846
t->use_callback ? (void *)t->callback : (void *)t->func);
kernel/softirq.c
916
struct tasklet_struct *t = list;
kernel/softirq.c
920
if (tasklet_trylock(t)) {
kernel/softirq.c
921
if (!atomic_read(&t->count)) {
kernel/softirq.c
922
if (tasklet_clear_sched(t)) {
kernel/softirq.c
923
if (t->use_callback) {
kernel/softirq.c
924
trace_tasklet_entry(t, t->callback);
kernel/softirq.c
925
t->callback(t);
kernel/softirq.c
926
trace_tasklet_exit(t, t->callback);
kernel/softirq.c
928
trace_tasklet_entry(t, t->func);
kernel/softirq.c
929
t->func(t->data);
kernel/softirq.c
930
trace_tasklet_exit(t, t->func);
kernel/softirq.c
933
tasklet_unlock(t);
kernel/softirq.c
937
tasklet_unlock(t);
kernel/softirq.c
941
t->next = NULL;
kernel/softirq.c
942
*tl_head->tail = t;
kernel/softirq.c
943
tl_head->tail = &t->next;
kernel/softirq.c
962
void tasklet_setup(struct tasklet_struct *t,
kernel/softirq.c
965
t->next = NULL;
kernel/softirq.c
966
t->state = 0;
kernel/softirq.c
967
atomic_set(&t->count, 0);
kernel/softirq.c
968
t->callback = callback;
kernel/softirq.c
969
t->use_callback = true;
kernel/softirq.c
970
t->data = 0;
kernel/softirq.c
974
void tasklet_init(struct tasklet_struct *t,
kernel/softirq.c
977
t->next = NULL;
kernel/softirq.c
978
t->state = 0;
kernel/softirq.c
979
atomic_set(&t->count, 0);
kernel/softirq.c
980
t->func = func;
kernel/softirq.c
981
t->use_callback = false;
kernel/softirq.c
982
t->data = data;
kernel/softirq.c
991
void tasklet_unlock_spin_wait(struct tasklet_struct *t)
kernel/softirq.c
993
while (test_bit(TASKLET_STATE_RUN, &(t)->state)) {
kernel/sys.c
1836
static void accumulate_thread_rusage(struct task_struct *t, struct rusage *r)
kernel/sys.c
1838
r->ru_nvcsw += t->nvcsw;
kernel/sys.c
1839
r->ru_nivcsw += t->nivcsw;
kernel/sys.c
1840
r->ru_minflt += t->min_flt;
kernel/sys.c
1841
r->ru_majflt += t->maj_flt;
kernel/sys.c
1842
r->ru_inblock += task_io_get_inblock(t);
kernel/sys.c
1843
r->ru_oublock += task_io_get_oublock(t);
kernel/sys.c
1848
struct task_struct *t;
kernel/sys.c
1898
__for_each_thread(sig, t)
kernel/sys.c
1899
accumulate_thread_rusage(t, r);
kernel/sys.c
2365
int __weak arch_prctl_spec_ctrl_get(struct task_struct *t, unsigned long which)
kernel/sys.c
2370
int __weak arch_prctl_spec_ctrl_set(struct task_struct *t, unsigned long which,
kernel/sys.c
2376
int __weak arch_get_shadow_stack_status(struct task_struct *t, unsigned long __user *status)
kernel/sys.c
2381
int __weak arch_set_shadow_stack_status(struct task_struct *t, unsigned long status)
kernel/sys.c
2386
int __weak arch_lock_shadow_stack_status(struct task_struct *t, unsigned long status)
kernel/sys.c
2391
int __weak arch_get_indir_br_lp_status(struct task_struct *t, unsigned long __user *status)
kernel/sys.c
2396
int __weak arch_set_indir_br_lp_status(struct task_struct *t, unsigned long status)
kernel/sys.c
2401
int __weak arch_lock_indir_br_lp_status(struct task_struct *t, unsigned long status)
kernel/time/hrtimer.c
2015
struct hrtimer_sleeper *t =
kernel/time/hrtimer.c
2017
struct task_struct *task = t->task;
kernel/time/hrtimer.c
2019
t->task = NULL;
kernel/time/hrtimer.c
2115
static int __sched do_nanosleep(struct hrtimer_sleeper *t, enum hrtimer_mode mode)
kernel/time/hrtimer.c
2121
hrtimer_sleeper_start_expires(t, mode);
kernel/time/hrtimer.c
2123
if (likely(t->task))
kernel/time/hrtimer.c
2126
hrtimer_cancel(&t->timer);
kernel/time/hrtimer.c
2129
} while (t->task && !signal_pending(current));
kernel/time/hrtimer.c
2133
if (!t->task)
kernel/time/hrtimer.c
2138
ktime_t rem = hrtimer_expires_remaining(&t->timer);
kernel/time/hrtimer.c
2152
struct hrtimer_sleeper t;
kernel/time/hrtimer.c
2155
hrtimer_setup_sleeper_on_stack(&t, restart->nanosleep.clockid, HRTIMER_MODE_ABS);
kernel/time/hrtimer.c
2156
hrtimer_set_expires(&t.timer, restart->nanosleep.expires);
kernel/time/hrtimer.c
2157
ret = do_nanosleep(&t, HRTIMER_MODE_ABS);
kernel/time/hrtimer.c
2158
destroy_hrtimer_on_stack(&t.timer);
kernel/time/hrtimer.c
2166
struct hrtimer_sleeper t;
kernel/time/hrtimer.c
2169
hrtimer_setup_sleeper_on_stack(&t, clockid, mode);
kernel/time/hrtimer.c
2170
hrtimer_set_expires_range_ns(&t.timer, rqtp, current->timer_slack_ns);
kernel/time/hrtimer.c
2171
ret = do_nanosleep(&t, mode);
kernel/time/hrtimer.c
2182
restart->nanosleep.clockid = t.timer.base->clockid;
kernel/time/hrtimer.c
2183
restart->nanosleep.expires = hrtimer_get_expires(&t.timer);
kernel/time/hrtimer.c
2186
destroy_hrtimer_on_stack(&t.timer);
kernel/time/hrtimer.c
312
# define switch_hrtimer_base(t, b, p) (b)
kernel/time/itimer.c
222
#define timeval_valid(t) \
kernel/time/itimer.c
223
(((t)->tv_sec >= 0) && (((unsigned long) (t)->tv_usec) < USEC_PER_SEC))
kernel/time/itimer.c
58
u64 t, samples[CPUCLOCK_MAX];
kernel/time/itimer.c
61
t = samples[clock_id];
kernel/time/itimer.c
63
if (val < t)
kernel/time/itimer.c
67
val -= t;
kernel/time/posix-cpu-timers.c
1600
struct timespec64 t;
kernel/time/posix-cpu-timers.c
1602
t = ktime_to_timespec64(restart_block->nanosleep.expires);
kernel/time/posix-cpu-timers.c
1604
return do_cpu_nanosleep(which_clock, TIMER_ABSTIME, &t);
kernel/time/posix-cpu-timers.c
362
u64 t;
kernel/time/posix-cpu-timers.c
372
t = cpu_clock_sample(clkid, tsk);
kernel/time/posix-cpu-timers.c
374
t = cpu_clock_sample_group(clkid, tsk, false);
kernel/time/posix-cpu-timers.c
377
*tp = ns_to_timespec64(t);
kernel/time/posix-stubs.c
110
if (get_timespec64(&t, rqtp))
kernel/time/posix-stubs.c
112
if (!timespec64_valid(&t))
kernel/time/posix-stubs.c
119
texp = timespec64_to_ktime(t);
kernel/time/posix-stubs.c
181
struct timespec64 t;
kernel/time/posix-stubs.c
193
if (get_old_timespec32(&t, rqtp))
kernel/time/posix-stubs.c
195
if (!timespec64_valid(&t))
kernel/time/posix-stubs.c
202
texp = timespec64_to_ktime(t);
kernel/time/posix-stubs.c
98
struct timespec64 t;
kernel/time/posix-timers.c
1373
struct timespec64 t;
kernel/time/posix-timers.c
1380
if (get_timespec64(&t, rqtp))
kernel/time/posix-timers.c
1383
if (!timespec64_valid(&t))
kernel/time/posix-timers.c
1391
return kc->nsleep(which_clock, flags, &t);
kernel/time/posix-timers.c
1401
struct timespec64 t;
kernel/time/posix-timers.c
1408
if (get_old_timespec32(&t, rqtp))
kernel/time/posix-timers.c
1411
if (!timespec64_valid(&t))
kernel/time/posix-timers.c
1419
return kc->nsleep(which_clock, flags, &t);
kernel/time/posix-timers.c
205
struct __kernel_timex *t)
kernel/time/posix-timers.c
207
return do_adjtimex(t);
kernel/time/sleep_timeout.c
192
struct hrtimer_sleeper t;
kernel/time/sleep_timeout.c
211
hrtimer_setup_sleeper_on_stack(&t, clock_id, mode);
kernel/time/sleep_timeout.c
212
hrtimer_set_expires_range_ns(&t.timer, *expires, delta);
kernel/time/sleep_timeout.c
213
hrtimer_sleeper_start_expires(&t, mode);
kernel/time/sleep_timeout.c
215
if (likely(t.task))
kernel/time/sleep_timeout.c
218
hrtimer_cancel(&t.timer);
kernel/time/sleep_timeout.c
219
destroy_hrtimer_on_stack(&t.timer);
kernel/time/sleep_timeout.c
223
return !t.task ? 0 : -EINTR;
kernel/time/sleep_timeout.c
23
static void process_timeout(struct timer_list *t)
kernel/time/sleep_timeout.c
25
struct process_timer *timeout = timer_container_of(timeout, t, timer);
kernel/time/tick-broadcast-hrtimer.c
94
static enum hrtimer_restart bc_handler(struct hrtimer *t)
kernel/time/tick-sched.c
590
struct task_struct *t;
kernel/time/tick-sched.c
593
__for_each_thread(sig, t)
kernel/time/tick-sched.c
594
tick_nohz_kick_task(t);
kernel/time/timekeeping.c
2414
ktime_t t = ktime_sub(tk->offs_real, tk->offs_boot);
kernel/time/timekeeping.c
2416
*ts = ktime_to_timespec64(t);
kernel/time/timekeeping_debug.c
49
void tk_debug_account_sleep_time(const struct timespec64 *t)
kernel/time/timekeeping_debug.c
52
int bin = min(fls(t->tv_sec), NUM_BINS-1);
kernel/time/timekeeping_debug.c
56
(s64)t->tv_sec, t->tv_nsec / NSEC_PER_MSEC);
kernel/time/timekeeping_internal.h
21
extern void tk_debug_account_sleep_time(const struct timespec64 *t);
kernel/time/timer.c
653
void (*function)(struct timer_list *t);
kernel/trace/blktrace.c
121
struct blk_io_trace *t;
kernel/trace/blktrace.c
122
size_t trace_len = sizeof(*t) + pdu_len + cgid_len;
kernel/trace/blktrace.c
124
t = relay_reserve(bt->rchan, trace_len);
kernel/trace/blktrace.c
125
if (!t)
kernel/trace/blktrace.c
128
t->magic = BLK_IO_TRACE_MAGIC | BLK_IO_TRACE_VERSION;
kernel/trace/blktrace.c
129
t->sequence = sequence;
kernel/trace/blktrace.c
130
t->time = ktime_to_ns(ktime_get());
kernel/trace/blktrace.c
132
record_blktrace_event(t, pid, cpu, sector, bytes, what, bt->dev, error,
kernel/trace/blktrace.c
1401
static void fill_rwbs(char *rwbs, const struct blk_io_trace2 *t)
kernel/trace/blktrace.c
1404
int tc = t->action >> BLK_TC_SHIFT;
kernel/trace/blktrace.c
1406
if ((t->action & ~__BLK_TN_CGROUP) == BLK_TN_MESSAGE) {
kernel/trace/blktrace.c
141
struct blk_io_trace2 *t;
kernel/trace/blktrace.c
1421
else if (t->bytes)
kernel/trace/blktrace.c
144
t = relay_reserve(bt->rchan, trace_len);
kernel/trace/blktrace.c
145
if (!t)
kernel/trace/blktrace.c
148
t->magic = BLK_IO_TRACE_MAGIC | BLK_IO_TRACE2_VERSION;
kernel/trace/blktrace.c
149
t->sequence = sequence;
kernel/trace/blktrace.c
150
t->time = ktime_to_ns(ktime_get());
kernel/trace/blktrace.c
1500
const struct blk_io_trace2 *t = te_blk_io_trace(iter->ent);
kernel/trace/blktrace.c
1502
fill_rwbs(rwbs, t);
kernel/trace/blktrace.c
1506
MAJOR(t->device), MINOR(t->device), iter->cpu,
kernel/trace/blktrace.c
1514
const struct blk_io_trace2 *t = te_blk_io_trace(iter->ent);
kernel/trace/blktrace.c
1516
fill_rwbs(rwbs, t);
kernel/trace/blktrace.c
152
record_blktrace_event2(t, pid, cpu, sector, bytes, what, bt->dev, error,
kernel/trace/blktrace.c
1526
MAJOR(t->device), MINOR(t->device),
kernel/trace/blktrace.c
1544
MAJOR(t->device), MINOR(t->device),
kernel/trace/blktrace.c
1549
MAJOR(t->device), MINOR(t->device), act, rwbs);
kernel/trace/blktrace.c
1735
const struct blk_io_trace2 *t;
kernel/trace/blktrace.c
1741
t = te_blk_io_trace(iter->ent);
kernel/trace/blktrace.c
1742
what = (t->action & ((1 << BLK_TC_SHIFT) - 1)) & ~__BLK_TA_CGROUP;
kernel/trace/blktrace.c
1745
has_cg = t->action & __BLK_TA_CGROUP;
kernel/trace/blktrace.c
1747
if ((t->action & ~__BLK_TN_CGROUP) == BLK_TN_MESSAGE) {
kernel/trace/blktrace.c
1772
struct blk_io_trace2 *t = (struct blk_io_trace2 *)iter->ent;
kernel/trace/blktrace.c
1780
trace_seq_putmem(s, &t->sector,
kernel/trace/blktrace.c
1781
sizeof(old) - offset + t->pdu_len);
kernel/trace/blktrace.c
185
struct blk_io_trace2 *t;
kernel/trace/blktrace.c
186
size_t trace_len = sizeof(*t) + cgid_len + len;
kernel/trace/blktrace.c
194
t = ring_buffer_event_data(event);
kernel/trace/blktrace.c
195
record_blktrace_event2(t, pid, cpu, 0, 0,
kernel/trace/blktrace.c
66
static void record_blktrace_event(struct blk_io_trace *t, pid_t pid, int cpu,
kernel/trace/blktrace.c
78
t->cpu = cpu;
kernel/trace/blktrace.c
79
t->pid = pid;
kernel/trace/blktrace.c
81
t->sector = sector;
kernel/trace/blktrace.c
82
t->bytes = bytes;
kernel/trace/blktrace.c
83
t->action = lower_32_bits(what);
kernel/trace/blktrace.c
84
t->device = dev;
kernel/trace/blktrace.c
85
t->error = error;
kernel/trace/blktrace.c
86
t->pdu_len = pdu_len + cgid_len;
kernel/trace/blktrace.c
89
memcpy((void *)t + sizeof(*t), &cgid, cgid_len);
kernel/trace/blktrace.c
91
memcpy((void *)t + sizeof(*t) + cgid_len, pdu_data, pdu_len);
kernel/trace/bpf_trace.c
1000
if (*btf_id <= 0 || !t)
kernel/trace/bpf_trace.c
980
const struct btf_type *t;
kernel/trace/bpf_trace.c
999
t = btf_type_by_id(*btf, *btf_id);
kernel/trace/fgraph.c
1045
struct task_struct *g, *t;
kernel/trace/fgraph.c
1061
for_each_process_thread(g, t) {
kernel/trace/fgraph.c
1067
if (t->ret_stack == NULL) {
kernel/trace/fgraph.c
1068
atomic_set(&t->trace_overrun, 0);
kernel/trace/fgraph.c
1070
t->curr_ret_stack = 0;
kernel/trace/fgraph.c
1071
t->curr_ret_depth = -1;
kernel/trace/fgraph.c
1074
t->ret_stack = ret_stack_list[start++];
kernel/trace/fgraph.c
1115
graph_init_task(struct task_struct *t, unsigned long *ret_stack)
kernel/trace/fgraph.c
1117
atomic_set(&t->trace_overrun, 0);
kernel/trace/fgraph.c
1119
t->ftrace_timestamp = 0;
kernel/trace/fgraph.c
1120
t->curr_ret_stack = 0;
kernel/trace/fgraph.c
1121
t->curr_ret_depth = -1;
kernel/trace/fgraph.c
1124
t->ret_stack = ret_stack;
kernel/trace/fgraph.c
1131
void ftrace_graph_init_idle_task(struct task_struct *t, int cpu)
kernel/trace/fgraph.c
1133
t->curr_ret_stack = 0;
kernel/trace/fgraph.c
1134
t->curr_ret_depth = -1;
kernel/trace/fgraph.c
1139
if (t->ret_stack)
kernel/trace/fgraph.c
1140
WARN_ON(t->ret_stack != per_cpu(idle_ret_stack, cpu));
kernel/trace/fgraph.c
1155
graph_init_task(t, ret_stack);
kernel/trace/fgraph.c
1160
void ftrace_graph_init_task(struct task_struct *t)
kernel/trace/fgraph.c
1163
t->ret_stack = NULL;
kernel/trace/fgraph.c
1164
t->curr_ret_stack = 0;
kernel/trace/fgraph.c
1165
t->curr_ret_depth = -1;
kernel/trace/fgraph.c
1176
graph_init_task(t, ret_stack);
kernel/trace/fgraph.c
1180
void ftrace_graph_exit_task(struct task_struct *t)
kernel/trace/fgraph.c
1182
unsigned long *ret_stack = t->ret_stack;
kernel/trace/fgraph.c
1184
t->ret_stack = NULL;
kernel/trace/fgraph.c
1270
struct task_struct *g, *t;
kernel/trace/fgraph.c
1279
for_each_process_thread(g, t) {
kernel/trace/fgraph.c
1280
if (t->ret_stack)
kernel/trace/fgraph.c
1281
ret_stack_set_task_var(t, idx, 0);
kernel/trace/fgraph.c
163
#define RET_STACK(t, offset) ((struct ftrace_ret_stack *)(&(t)->ret_stack[offset]))
kernel/trace/fgraph.c
249
static inline unsigned long get_fgraph_entry(struct task_struct *t, int offset)
kernel/trace/fgraph.c
251
return t->ret_stack[offset];
kernel/trace/fgraph.c
255
static inline int get_frame_offset(struct task_struct *t, int offset)
kernel/trace/fgraph.c
257
return __get_offset(t->ret_stack[offset]);
kernel/trace/fgraph.c
262
get_bitmap_bits(struct task_struct *t, int offset)
kernel/trace/fgraph.c
264
return (t->ret_stack[offset] >> FGRAPH_INDEX_SHIFT) & FGRAPH_INDEX_MASK;
kernel/trace/fgraph.c
269
set_bitmap(struct task_struct *t, int offset, unsigned long bitmap)
kernel/trace/fgraph.c
271
t->ret_stack[offset] = (bitmap << FGRAPH_INDEX_SHIFT) |
kernel/trace/fgraph.c
276
static inline void *get_data_type_data(struct task_struct *t, int offset)
kernel/trace/fgraph.c
278
unsigned long val = t->ret_stack[offset];
kernel/trace/fgraph.c
283
return (void *)&t->ret_stack[offset];
kernel/trace/fgraph.c
307
static void ret_stack_set_task_var(struct task_struct *t, int idx, long val)
kernel/trace/fgraph.c
309
unsigned long *gvals = SHADOW_STACK_TASK_VARS(t->ret_stack);
kernel/trace/fgraph.c
315
ret_stack_get_task_var(struct task_struct *t, int idx)
kernel/trace/fgraph.c
317
unsigned long *gvals = SHADOW_STACK_TASK_VARS(t->ret_stack);
kernel/trace/fgraph.c
434
get_ret_stack(struct task_struct *t, int offset, int *frame_offset)
kernel/trace/fgraph.c
443
offs = get_frame_offset(t, --offset);
kernel/trace/fgraph.c
450
return RET_STACK(t, offset);
kernel/trace/ring_buffer.c
645
static inline void rb_time_read(rb_time_t *t, u64 *ret)
kernel/trace/ring_buffer.c
647
*ret = local64_read(&t->time);
kernel/trace/ring_buffer.c
649
static void rb_time_set(rb_time_t *t, u64 val)
kernel/trace/ring_buffer.c
651
local64_set(&t->time, val);
kernel/trace/trace.c
1852
struct tracer *t, **last;
kernel/trace/trace.c
1876
for (t = trace_types; t; t = t->next) {
kernel/trace/trace.c
1877
if (t == p->type) {
kernel/trace/trace.c
1878
*last = t->next;
kernel/trace/trace.c
1881
last = &t->next;
kernel/trace/trace.c
1899
static int add_tracer(struct trace_array *tr, struct tracer *t);
kernel/trace/trace.c
1905
struct tracers *t, *n;
kernel/trace/trace.c
1909
list_for_each_entry_safe(t, n, &tr->tracers, list) {
kernel/trace/trace.c
1910
list_del(&t->list);
kernel/trace/trace.c
1911
kfree(t->flags);
kernel/trace/trace.c
1912
kfree(t);
kernel/trace/trace.c
1925
struct tracer *t;
kernel/trace/trace.c
1946
for (t = trace_types; t; t = t->next) {
kernel/trace/trace.c
1947
if (strcmp(type->name, t->name) == 0) {
kernel/trace/trace.c
3318
unsigned long t, e;
kernel/trace/trace.c
3325
get_total_entries_cpu(buf, &t, &e, cpu);
kernel/trace/trace.c
3326
*total += t;
kernel/trace/trace.c
4179
trace_ok_for_array(struct tracer *t, struct trace_array *tr)
kernel/trace/trace.c
4182
if (tr->range_addr_start && tracer_uses_snapshot(t))
kernel/trace/trace.c
4184
return (tr->flags & TRACE_ARRAY_FL_GLOBAL) || t->allow_instances;
kernel/trace/trace.c
4189
get_tracer_for_array(struct trace_array *tr, struct tracer *t)
kernel/trace/trace.c
4191
while (t && !trace_ok_for_array(t, tr))
kernel/trace/trace.c
4192
t = t->next;
kernel/trace/trace.c
4194
return t;
kernel/trace/trace.c
4201
struct tracer *t = v;
kernel/trace/trace.c
4205
if (t)
kernel/trace/trace.c
4206
t = get_tracer_for_array(tr, t->next);
kernel/trace/trace.c
4208
return t;
kernel/trace/trace.c
4214
struct tracer *t;
kernel/trace/trace.c
4219
t = get_tracer_for_array(tr, trace_types);
kernel/trace/trace.c
4220
for (; t && l < *pos; t = t_next(m, t, &l))
kernel/trace/trace.c
4223
return t;
kernel/trace/trace.c
4233
struct tracer *t = v;
kernel/trace/trace.c
4235
if (!t)
kernel/trace/trace.c
4238
seq_puts(m, t->name);
kernel/trace/trace.c
4239
if (t->next)
kernel/trace/trace.c
5218
int tracer_init(struct tracer *t, struct trace_array *tr)
kernel/trace/trace.c
5222
return t->init(tr);
kernel/trace/trace.c
5552
struct tracers *t;
kernel/trace/trace.c
5568
list_for_each_entry(t, &tr->tracers, list) {
kernel/trace/trace.c
5569
if (strcmp(t->tracer->name, buf) == 0) {
kernel/trace/trace.c
5570
trace = t->tracer;
kernel/trace/trace.c
5638
tr->current_trace_flags = t->flags ? : t->tracer->flags;
kernel/trace/trace.c
8313
unsigned long long t;
kernel/trace/trace.c
8336
t = ns2usecs(ring_buffer_oldest_event_ts(trace_buf->buffer, cpu));
kernel/trace/trace.c
8337
usec_rem = do_div(t, USEC_PER_SEC);
kernel/trace/trace.c
8339
t, usec_rem);
kernel/trace/trace.c
8341
t = ns2usecs(ring_buffer_time_stamp(trace_buf->buffer));
kernel/trace/trace.c
8342
usec_rem = do_div(t, USEC_PER_SEC);
kernel/trace/trace.c
8343
trace_seq_printf(s, "now ts: %5llu.%06lu\n", t, usec_rem);
kernel/trace/trace.c
8938
struct tracers *t;
kernel/trace/trace.c
8940
list_for_each_entry(t, &global_trace.tracers, list) {
kernel/trace/trace.c
8941
if (t->tracer != tracer)
kernel/trace/trace.c
8943
if (!t->flags)
kernel/trace/trace.c
8945
return t->flags->val;
kernel/trace/trace.c
8950
static int add_tracer_options(struct trace_array *tr, struct tracers *t)
kernel/trace/trace.c
8952
struct tracer *tracer = t->tracer;
kernel/trace/trace.c
8953
struct tracer_flags *flags = t->flags ?: tracer->flags;
kernel/trace/trace.c
8968
struct tracers *t;
kernel/trace/trace.c
8982
t = kmalloc_obj(*t);
kernel/trace/trace.c
8983
if (!t)
kernel/trace/trace.c
8986
t->tracer = tracer;
kernel/trace/trace.c
8987
t->flags = NULL;
kernel/trace/trace.c
8988
list_add(&t->list, &tr->tracers);
kernel/trace/trace.c
9006
t->flags = flags;
kernel/trace/trace.c
9016
ret = add_tracer_options(tr, t);
kernel/trace/trace.c
9018
list_del(&t->list);
kernel/trace/trace.c
9019
kfree(t->flags);
kernel/trace/trace.c
9020
kfree(t);
kernel/trace/trace.c
9476
struct tracer *t;
kernel/trace/trace.c
9479
for (t = trace_types; t && !ret; t = t->next)
kernel/trace/trace.c
9480
ret = add_tracer(tr, t);
kernel/trace/trace.c
9487
struct tracers *t;
kernel/trace/trace.c
9490
list_for_each_entry(t, &tr->tracers, list) {
kernel/trace/trace.c
9491
ret = add_tracer_options(tr, t);
kernel/trace/trace.h
671
int tracer_init(struct tracer *t, struct trace_array *tr);
kernel/trace/trace_btf.c
15
const struct btf_type *t;
kernel/trace/trace_btf.c
23
t = btf_type_by_id(*btf_p, id);
kernel/trace/trace_btf.c
24
if (!t || !btf_type_is_func(t))
kernel/trace/trace_btf.c
28
t = btf_type_by_id(*btf_p, t->type);
kernel/trace/trace_btf.c
29
if (!t || !btf_type_is_func_proto(t))
kernel/trace/trace_btf.c
32
return t;
kernel/trace/trace_events_filter.c
170
int t, s;
kernel/trace/trace_events_filter.c
172
t = prog[N].target;
kernel/trace/trace_events_filter.c
173
s = prog[t].target;
kernel/trace/trace_events_filter.c
174
prog[t].when_to_branch = invert;
kernel/trace/trace_events_filter.c
175
prog[t].target = N;
kernel/trace/trace_events_synth.c
2241
char *type, *t;
kernel/trace/trace_events_synth.c
2249
t = strstr(type, "__data_loc");
kernel/trace/trace_events_synth.c
2250
if (t) { /* __data_loc belongs in format but not event desc */
kernel/trace/trace_events_synth.c
2251
t += sizeof("__data_loc");
kernel/trace/trace_events_synth.c
2252
type = t;
kernel/trace/trace_events_trigger.c
251
static void *trigger_next(struct seq_file *m, void *t, loff_t *pos)
kernel/trace/trace_events_trigger.c
255
if (t == SHOW_AVAILABLE_TRIGGERS) {
kernel/trace/trace_events_trigger.c
259
return seq_list_next(t, &event_file->triggers, pos);
kernel/trace/trace_events_trigger.c
291
static void trigger_stop(struct seq_file *m, void *t)
kernel/trace/trace_events_user.c
705
static struct user_event_mm *user_event_mm_alloc(struct task_struct *t)
kernel/trace/trace_events_user.c
714
user_mm->mm = t->mm;
kernel/trace/trace_events_user.c
732
static void user_event_mm_attach(struct user_event_mm *user_mm, struct task_struct *t)
kernel/trace/trace_events_user.c
740
t->user_event_mm = user_mm;
kernel/trace/trace_events_user.c
787
void user_event_mm_remove(struct task_struct *t)
kernel/trace/trace_events_user.c
794
mm = t->user_event_mm;
kernel/trace/trace_events_user.c
795
t->user_event_mm = NULL;
kernel/trace/trace_events_user.c
841
void user_event_mm_dup(struct task_struct *t, struct user_event_mm *old_mm)
kernel/trace/trace_events_user.c
843
struct user_event_mm *mm = user_event_mm_alloc(t);
kernel/trace/trace_events_user.c
858
user_event_mm_attach(mm, t);
kernel/trace/trace_functions_graph.c
709
static void print_graph_abs_time(u64 t, struct trace_seq *s)
kernel/trace/trace_functions_graph.c
713
usecs_rem = do_div(t, NSEC_PER_SEC);
kernel/trace/trace_functions_graph.c
717
(unsigned long)t, usecs_rem);
kernel/trace/trace_mmiotrace.c
171
unsigned long long t = ns2usecs(iter->ts);
kernel/trace/trace_mmiotrace.c
172
unsigned long usec_rem = do_div(t, USEC_PER_SEC);
kernel/trace/trace_mmiotrace.c
173
unsigned secs = (unsigned long)t;
kernel/trace/trace_mmiotrace.c
216
unsigned long long t = ns2usecs(iter->ts);
kernel/trace/trace_mmiotrace.c
217
unsigned long usec_rem = do_div(t, USEC_PER_SEC);
kernel/trace/trace_mmiotrace.c
218
unsigned secs = (unsigned long)t;
kernel/trace/trace_mmiotrace.c
250
unsigned long long t = ns2usecs(iter->ts);
kernel/trace/trace_mmiotrace.c
251
unsigned long usec_rem = do_div(t, USEC_PER_SEC);
kernel/trace/trace_mmiotrace.c
252
unsigned secs = (unsigned long)t;
kernel/trace/trace_osnoise.c
1124
thread_entry(struct osnoise_variables *osn_var, struct task_struct *t)
kernel/trace/trace_osnoise.c
1146
thread_exit(struct osnoise_variables *osn_var, struct task_struct *t)
kernel/trace/trace_osnoise.c
1159
trace_thread_noise(t, osn_var->thread.arrival_time, duration);
kernel/trace/trace_output.c
641
unsigned long long t;
kernel/trace/trace_output.c
644
t = ns2usecs(ts);
kernel/trace/trace_output.c
645
usec_rem = do_div(t, USEC_PER_SEC);
kernel/trace/trace_output.c
646
secs = (unsigned long)t;
kernel/trace/trace_output.c
721
const struct btf_type *t;
kernel/trace/trace_output.c
738
t = btf_find_func_proto(name, &btf);
kernel/trace/trace_output.c
739
if (IS_ERR_OR_NULL(t))
kernel/trace/trace_output.c
742
param = btf_get_func_param(t, &nr);
kernel/trace/trace_output.c
761
t = btf_type_skip_modifiers(btf, param[p].type, &tid);
kernel/trace/trace_output.c
763
switch (t ? BTF_INFO_KIND(t->info) : BTF_KIND_UNKN) {
kernel/trace/trace_output.c
772
encode = btf_int_encoding(t);
kernel/trace/trace_output.c
784
trace_seq_printf(s, "(%s)[0x%lx", btf_type_str(t), arg);
kernel/trace/trace_output.c
785
for (x = sizeof(long); x < t->size; x += sizeof(long)) {
kernel/trace/trace_probe.c
1255
const struct fetch_type *t,
kernel/trace/trace_probe.c
1281
code->lshift = BYTES_TO_BITS(t->size) - (bw + bo);
kernel/trace/trace_probe.c
1282
code->rshift = BYTES_TO_BITS(t->size) - bw;
kernel/trace/trace_probe.c
1283
code->basesize = t->size;
kernel/trace/trace_probe.c
1285
return (BYTES_TO_BITS(t->size) < (bw + bo)) ? -EINVAL : 0;
kernel/trace/trace_probe.c
1292
char *t = NULL, *t2, *t3;
kernel/trace/trace_probe.c
1295
t = strchr(arg, ':');
kernel/trace/trace_probe.c
1296
if (t) {
kernel/trace/trace_probe.c
1297
*t++ = '\0';
kernel/trace/trace_probe.c
1298
t2 = strchr(t, '[');
kernel/trace/trace_probe.c
1326
offs = t ? t - arg : 0;
kernel/trace/trace_probe.c
1336
if (parg->count || (t && strcmp(t, "string"))) {
kernel/trace/trace_probe.c
1342
parg->type = find_fetch_type(t, ctx->flags);
kernel/trace/trace_probe.c
1349
return t;
kernel/trace/trace_probe.c
337
static u32 btf_type_int(const struct btf_type *t)
kernel/trace/trace_probe.c
339
return *(u32 *)(t + 1);
kernel/trace/trace_probe.c
943
static int parse_probe_vars(char *orig_arg, const struct fetch_type *t,
kernel/trace/trace_probe.h
181
#define __DEFAULT_FETCH_TYPE(t) x##t
kernel/trace/trace_probe.h
182
#define _DEFAULT_FETCH_TYPE(t) __DEFAULT_FETCH_TYPE(t)
kernel/trace/trace_probe.h
186
#define __ADDR_FETCH_TYPE(t) u##t
kernel/trace/trace_probe.h
187
#define _ADDR_FETCH_TYPE(t) __ADDR_FETCH_TYPE(t)
kernel/tracepoint.c
764
struct task_struct *p, *t;
kernel/tracepoint.c
768
for_each_process_thread(p, t) {
kernel/tracepoint.c
769
set_task_syscall_work(t, SYSCALL_TRACEPOINT);
kernel/tracepoint.c
780
struct task_struct *p, *t;
kernel/tracepoint.c
785
for_each_process_thread(p, t) {
kernel/tracepoint.c
786
clear_task_syscall_work(t, SYSCALL_TRACEPOINT);
kernel/umh.c
500
struct ctl_table t;
kernel/umh.c
521
t = *table;
kernel/umh.c
522
t.data = &cap_array;
kernel/umh.c
528
err = proc_doulongvec_minmax(&t, write, buffer, lenp, ppos);
kernel/unwind/deferred.c
300
struct task_struct *g, *t;
kernel/unwind/deferred.c
324
for_each_process_thread(g, t) {
kernel/unwind/deferred.c
326
&t->unwind_info.unwind_mask);
kernel/unwind/deferred.c
327
if (t->unwind_info.cache)
kernel/unwind/deferred.c
328
clear_bit(bit, &t->unwind_info.cache->unwind_completed);
kernel/workqueue.c
2495
void delayed_work_timer_fn(struct timer_list *t)
kernel/workqueue.c
2497
struct delayed_work *dwork = timer_container_of(dwork, t, timer);
kernel/workqueue.c
2923
static void idle_worker_timeout(struct timer_list *t)
kernel/workqueue.c
2925
struct worker_pool *pool = timer_container_of(pool, t, idle_timer);
kernel/workqueue.c
3021
static void pool_mayday_timeout(struct timer_list *t)
kernel/workqueue.c
3023
struct worker_pool *pool = timer_container_of(pool, t, mayday_timer);
lib/842/842_compress.c
225
u8 *t = comp_ops[c];
lib/842/842_compress.c
231
pr_debug("template %x\n", t[4]);
lib/842/842_compress.c
233
ret = add_bits(p, t[4], OP_BITS);
lib/842/842_compress.c
238
pr_debug("op %x\n", t[i]);
lib/842/842_compress.c
240
switch (t[i] & OP_AMOUNT) {
lib/842/842_compress.c
244
else if (t[i] & OP_ACTION_INDEX)
lib/842/842_compress.c
246
else if (t[i] & OP_ACTION_DATA)
lib/842/842_compress.c
252
if (b == 2 && t[i] & OP_ACTION_DATA)
lib/842/842_compress.c
256
else if (t[i] & OP_ACTION_INDEX)
lib/842/842_compress.c
258
else if (t[i] & OP_ACTION_DATA)
lib/842/842_compress.c
266
if (t[i] & OP_ACTION_INDEX)
lib/842/842_compress.c
268
else if (t[i] & OP_ACTION_DATA)
lib/842/842_compress.c
274
inv = (b != 8) || !(t[i] & OP_ACTION_NOOP);
lib/842/842_compress.c
286
c, i, t[0], t[1], t[2], t[3]);
lib/842/842_compress.c
290
b += t[i] & OP_AMOUNT;
lib/842/842_compress.c
295
c, b, t[0], t[1], t[2], t[3]);
lib/842/842_compress.c
300
atomic_inc(&template_count[t[4]]);
lib/842/842_compress.c
382
u8 *t = comp_ops[c];
lib/842/842_compress.c
389
if (t[i] & OP_ACTION_INDEX) {
lib/842/842_compress.c
390
if (t[i] & OP_AMOUNT_2)
lib/842/842_compress.c
392
else if (t[i] & OP_AMOUNT_4)
lib/842/842_compress.c
394
else if (t[i] & OP_AMOUNT_8)
lib/842/842_compress.c
402
b += t[i] & OP_AMOUNT;
lib/bcd.c
13
const unsigned int t = (val * 103) >> 10;
lib/bcd.c
15
return (t << 4) | (val - t * 10);
lib/bch.c
1192
const unsigned int t = GF_T(bch);
lib/bch.c
1198
g = bch_alloc(GF_POLY_SZ(m*t), &err);
lib/bch.c
1200
genpoly = bch_alloc(DIV_ROUND_UP(m*t+1, 32)*sizeof(*genpoly), &err);
lib/bch.c
1210
for (i = 0; i < t; i++) {
lib/bch.c
1275
struct bch_control *bch_init(int m, int t, unsigned int prim_poly,
lib/bch.c
1292
if ((m != (CONFIG_BCH_CONST_M)) || (t != (CONFIG_BCH_CONST_T))) {
lib/bch.c
1307
if (t > BCH_MAX_T)
lib/bch.c
1315
if ((t < 1) || (m*t >= ((1 << m)-1)))
lib/bch.c
1328
bch->t = t;
lib/bch.c
1330
words = DIV_ROUND_UP(m*t, 32);
lib/bch.c
1331
bch->ecc_bytes = DIV_ROUND_UP(m*t, 8);
lib/bch.c
1338
bch->syn = bch_alloc(2*t*sizeof(*bch->syn), &err);
lib/bch.c
1339
bch->cache = bch_alloc(2*t*sizeof(*bch->cache), &err);
lib/bch.c
1340
bch->elp = bch_alloc((t+1)*sizeof(struct gf_poly_deg1), &err);
lib/bch.c
1344
bch->poly_2t[i] = bch_alloc(GF_POLY_SZ(2*t), &err);
lib/bch.c
376
const int t = GF_T(bch);
lib/bch.c
384
memset(syn, 0, 2*t*sizeof(*syn));
lib/bch.c
392
for (j = 0; j < 2*t; j += 2)
lib/bch.c
400
for (j = 0; j < t; j++)
lib/bch.c
412
const unsigned int t = GF_T(bch);
lib/bch.c
420
memset(pelp, 0, GF_POLY_SZ(2*t));
lib/bch.c
421
memset(elp, 0, GF_POLY_SZ(2*t));
lib/bch.c
429
for (i = 0; (i < t) && (elp->deg <= t); i++) {
lib/bch.c
451
if (i < t-1) {
lib/bch.c
458
return (elp->deg > t) ? -1 : (int)elp->deg;
lib/bch.c
544
unsigned int mask = 0xff, t, rows[16] = {0,};
lib/bch.c
564
t = ((rows[k] >> j)^rows[k+j]) & mask;
lib/bch.c
565
rows[k] ^= (t << j);
lib/bch.c
566
rows[k+j] ^= t;
lib/bch.c
86
#define GF_T(_p) ((_p)->t)
lib/closure.c
80
struct closure *cl, *t;
lib/closure.c
89
llist_for_each_entry_safe(cl, t, reverse, list) {
lib/crc/crc8.c
34
u8 t = msbit;
lib/crc/crc8.c
39
t = (t << 1) ^ (t & msbit ? polynomial : 0);
lib/crc/crc8.c
41
table[i+j] = table[j] ^ t;
lib/crc/crc8.c
55
u8 t = 1;
lib/crc/crc8.c
60
t = (t >> 1) ^ (t & 1 ? polynomial : 0);
lib/crc/crc8.c
62
table[i+j] = table[j] ^ t;
lib/crc/tests/crc_kunit.c
247
u64 t;
lib/crc/tests/crc_kunit.c
261
t = ktime_get_ns();
lib/crc/tests/crc_kunit.c
264
t = ktime_get_ns() - t;
lib/crc/tests/crc_kunit.c
267
len, div64_u64((u64)len * num_iters * 1000, t));
lib/crypto/blake2b.c
37
ctx->t[0] += inc;
lib/crypto/blake2b.c
38
ctx->t[1] += (ctx->t[0] < inc);
lib/crypto/blake2b.c
61
v[12] = BLAKE2B_IV4 ^ ctx->t[0];
lib/crypto/blake2b.c
62
v[13] = BLAKE2B_IV5 ^ ctx->t[1];
lib/crypto/blake2s.c
35
ctx->t[0] += inc;
lib/crypto/blake2s.c
36
ctx->t[1] += (ctx->t[0] < inc);
lib/crypto/blake2s.c
59
v[12] = BLAKE2S_IV4 ^ ctx->t[0];
lib/crypto/blake2s.c
60
v[13] = BLAKE2S_IV5 ^ ctx->t[1];
lib/crypto/curve25519-fiat32.c
102
static __always_inline u32 cmovznz32(u32 t, u32 z, u32 nz)
lib/crypto/curve25519-fiat32.c
104
t = -!!t; /* all set if nonzero, 0 if 0 */
lib/crypto/curve25519-fiat32.c
105
return (t&nz) | ((~t)&z);
lib/crypto/curve25519-hacl64.c
197
u128 t[5] = { 0 };
lib/crypto/curve25519-hacl64.c
198
fmul_mul_shift_reduce_(t, tmp, input21);
lib/crypto/curve25519-hacl64.c
199
fproduct_carry_wide_(t);
lib/crypto/curve25519-hacl64.c
200
b4 = t[4];
lib/crypto/curve25519-hacl64.c
201
b0 = t[0];
lib/crypto/curve25519-hacl64.c
204
t[4] = b4_;
lib/crypto/curve25519-hacl64.c
205
t[0] = b0_;
lib/crypto/curve25519-hacl64.c
206
fproduct_copy_from_wide_(output, t);
lib/crypto/curve25519-hacl64.c
284
u128 t[5];
lib/crypto/curve25519-hacl64.c
286
fsquare_fsquare_times_(output, t, count1);
lib/crypto/curve25519-hacl64.c
292
u128 t[5];
lib/crypto/curve25519-hacl64.c
293
fsquare_fsquare_times_(output, t, count1);
lib/crypto/gf128mul.c
245
struct gf128mul_64k *t;
lib/crypto/gf128mul.c
248
t = kzalloc_obj(*t);
lib/crypto/gf128mul.c
249
if (!t)
lib/crypto/gf128mul.c
253
t->t[i] = kzalloc_obj(*t->t[i]);
lib/crypto/gf128mul.c
254
if (!t->t[i]) {
lib/crypto/gf128mul.c
255
gf128mul_free_64k(t);
lib/crypto/gf128mul.c
256
t = NULL;
lib/crypto/gf128mul.c
261
t->t[0]->t[1] = *g;
lib/crypto/gf128mul.c
263
gf128mul_x_bbe(&t->t[0]->t[j + j], &t->t[0]->t[j]);
lib/crypto/gf128mul.c
268
be128_xor(&t->t[i]->t[j + k],
lib/crypto/gf128mul.c
269
&t->t[i]->t[j], &t->t[i]->t[k]);
lib/crypto/gf128mul.c
275
t->t[i]->t[j] = t->t[i - 1]->t[j];
lib/crypto/gf128mul.c
276
gf128mul_x8_bbe(&t->t[i]->t[j]);
lib/crypto/gf128mul.c
281
return t;
lib/crypto/gf128mul.c
285
void gf128mul_free_64k(struct gf128mul_64k *t)
lib/crypto/gf128mul.c
290
kfree_sensitive(t->t[i]);
lib/crypto/gf128mul.c
291
kfree_sensitive(t);
lib/crypto/gf128mul.c
295
void gf128mul_64k_bbe(be128 *a, const struct gf128mul_64k *t)
lib/crypto/gf128mul.c
301
*r = t->t[0]->t[ap[15]];
lib/crypto/gf128mul.c
303
be128_xor(r, r, &t->t[i]->t[ap[15 - i]]);
lib/crypto/gf128mul.c
326
struct gf128mul_4k *t;
lib/crypto/gf128mul.c
329
t = kzalloc_obj(*t);
lib/crypto/gf128mul.c
330
if (!t)
lib/crypto/gf128mul.c
333
t->t[128] = *g;
lib/crypto/gf128mul.c
335
gf128mul_x_lle(&t->t[j], &t->t[j+j]);
lib/crypto/gf128mul.c
339
be128_xor(&t->t[j + k], &t->t[j], &t->t[k]);
lib/crypto/gf128mul.c
342
return t;
lib/crypto/gf128mul.c
346
void gf128mul_4k_lle(be128 *a, const struct gf128mul_4k *t)
lib/crypto/gf128mul.c
352
*r = t->t[ap[15]];
lib/crypto/gf128mul.c
355
be128_xor(r, r, &t->t[ap[i]]);
lib/crypto/mldsa.c
233
s32 t = Zq_mult(z, w->x[j + len]);
lib/crypto/mldsa.c
235
w->x[j + len] = w->x[j] - t;
lib/crypto/mldsa.c
236
w->x[j] += t;
lib/crypto/mldsa.c
262
s32 t = w->x[j];
lib/crypto/mldsa.c
264
w->x[j] = t + w->x[j + len];
lib/crypto/mldsa.c
265
w->x[j + len] = Zq_mult(z, t - w->x[j + len]);
lib/crypto/mpi/longlong.h
40
#define __ll_lowpart(t) ((UWtype) (t) & (__ll_B - 1))
lib/crypto/mpi/longlong.h
41
#define __ll_highpart(t) ((UWtype) (t) >> (W_TYPE_SIZE / 2))
lib/crypto/sha1.c
58
#define SHA_SRC(t) get_unaligned_be32((__u32 *)data + t)
lib/crypto/sha1.c
59
#define SHA_MIX(t) rol32(W(t+13) ^ W(t+8) ^ W(t+2) ^ W(t), 1)
lib/crypto/sha1.c
61
#define SHA_ROUND(t, input, fn, constant, A, B, C, D, E) do { \
lib/crypto/sha1.c
62
__u32 TEMP = input(t); setW(t, TEMP); \
lib/crypto/sha1.c
67
#define T_0_15(t, A, B, C, D, E) SHA_ROUND(t, SHA_SRC, (((C^D)&B)^D) , 0x5a827999, A, B, C, D, E )
lib/crypto/sha1.c
68
#define T_16_19(t, A, B, C, D, E) SHA_ROUND(t, SHA_MIX, (((C^D)&B)^D) , 0x5a827999, A, B, C, D, E )
lib/crypto/sha1.c
69
#define T_20_39(t, A, B, C, D, E) SHA_ROUND(t, SHA_MIX, (B^C^D) , 0x6ed9eba1, A, B, C, D, E )
lib/crypto/sha1.c
70
#define T_40_59(t, A, B, C, D, E) SHA_ROUND(t, SHA_MIX, ((B&C)+(D&(B^C))) , 0x8f1bbcdc, A, B, C, D, E )
lib/crypto/sha1.c
71
#define T_60_79(t, A, B, C, D, E) SHA_ROUND(t, SHA_MIX, (B^C^D) , 0xca62c1d6, A, B, C, D, E )
lib/crypto/sha3.c
54
u64 t[5], tt, bc[5];
lib/crypto/sha3.c
63
t[0] = bc[4] ^ rol64(bc[1], 1);
lib/crypto/sha3.c
64
t[1] = bc[0] ^ rol64(bc[2], 1);
lib/crypto/sha3.c
65
t[2] = bc[1] ^ rol64(bc[3], 1);
lib/crypto/sha3.c
66
t[3] = bc[2] ^ rol64(bc[4], 1);
lib/crypto/sha3.c
67
t[4] = bc[3] ^ rol64(bc[0], 1);
lib/crypto/sha3.c
69
st[0] ^= t[0];
lib/crypto/sha3.c
73
st[ 1] = rol64(st[ 6] ^ t[1], 44);
lib/crypto/sha3.c
74
st[ 6] = rol64(st[ 9] ^ t[4], 20);
lib/crypto/sha3.c
75
st[ 9] = rol64(st[22] ^ t[2], 61);
lib/crypto/sha3.c
76
st[22] = rol64(st[14] ^ t[4], 39);
lib/crypto/sha3.c
77
st[14] = rol64(st[20] ^ t[0], 18);
lib/crypto/sha3.c
78
st[20] = rol64(st[ 2] ^ t[2], 62);
lib/crypto/sha3.c
79
st[ 2] = rol64(st[12] ^ t[2], 43);
lib/crypto/sha3.c
80
st[12] = rol64(st[13] ^ t[3], 25);
lib/crypto/sha3.c
81
st[13] = rol64(st[19] ^ t[4], 8);
lib/crypto/sha3.c
82
st[19] = rol64(st[23] ^ t[3], 56);
lib/crypto/sha3.c
83
st[23] = rol64(st[15] ^ t[0], 41);
lib/crypto/sha3.c
84
st[15] = rol64(st[ 4] ^ t[4], 27);
lib/crypto/sha3.c
85
st[ 4] = rol64(st[24] ^ t[4], 14);
lib/crypto/sha3.c
86
st[24] = rol64(st[21] ^ t[1], 2);
lib/crypto/sha3.c
87
st[21] = rol64(st[ 8] ^ t[3], 55);
lib/crypto/sha3.c
88
st[ 8] = rol64(st[16] ^ t[1], 45);
lib/crypto/sha3.c
89
st[16] = rol64(st[ 5] ^ t[0], 36);
lib/crypto/sha3.c
90
st[ 5] = rol64(st[ 3] ^ t[3], 28);
lib/crypto/sha3.c
91
st[ 3] = rol64(st[18] ^ t[3], 21);
lib/crypto/sha3.c
92
st[18] = rol64(st[17] ^ t[2], 15);
lib/crypto/sha3.c
93
st[17] = rol64(st[11] ^ t[1], 10);
lib/crypto/sha3.c
94
st[11] = rol64(st[ 7] ^ t[2], 6);
lib/crypto/sha3.c
95
st[ 7] = rol64(st[10] ^ t[0], 3);
lib/crypto/sha3.c
96
st[10] = rol64( tt ^ t[1], 1);
lib/crypto/sm3.c
41
#define R(i, a, b, c, d, e, f, g, h, t, w1, w2) \
lib/crypto/sm3.c
43
ss1 = rol32((rol32((a), 12) + (e) + (t)), 7); \
lib/crypto/sm3.c
52
#define R1(a, b, c, d, e, f, g, h, t, w1, w2) \
lib/crypto/sm3.c
53
R(1, a, b, c, d, e, f, g, h, t, w1, w2)
lib/crypto/sm3.c
54
#define R2(a, b, c, d, e, f, g, h, t, w1, w2) \
lib/crypto/sm3.c
55
R(2, a, b, c, d, e, f, g, h, t, w1, w2)
lib/crypto/tests/curve25519_kunit.c
1328
u64 t;
lib/crypto/tests/curve25519_kunit.c
1339
t = ktime_get_ns();
lib/crypto/tests/curve25519_kunit.c
1342
t = ktime_get_ns() - t;
lib/crypto/tests/curve25519_kunit.c
1346
div64_u64((u64)benchmark_niter * NSEC_PER_SEC, t ?: 1));
lib/crypto/tests/hash-test-template.h
556
u64 t;
lib/crypto/tests/hash-test-template.h
560
t = ktime_get_ns();
lib/crypto/tests/hash-test-template.h
563
t = ktime_get_ns() - t;
lib/crypto/tests/hash-test-template.h
566
div64_u64((u64)len * num_iters * 1000, t ?: 1));
lib/decompress_bunzip2.c
162
i, j, k, t, runPos, symCount, symTotal, nSelectors, *byteCount;
lib/decompress_bunzip2.c
195
t = get_bits(bd, 16);
lib/decompress_bunzip2.c
198
if (t&(1 << (15-i))) {
lib/decompress_bunzip2.c
247
t = get_bits(bd, 5)-1;
lib/decompress_bunzip2.c
250
if (((unsigned)t) > (MAX_HUFCODE_BITS-1))
lib/decompress_bunzip2.c
266
t += (((k+1)&2)-1);
lib/decompress_bunzip2.c
270
length[i] = t+1;
lib/decompress_bunzip2.c
311
for (t = 0; t < symCount; t++)
lib/decompress_bunzip2.c
312
if (length[t] == i)
lib/decompress_bunzip2.c
313
hufGroup->permute[pp++] = t;
lib/decompress_bunzip2.c
324
pp = t = 0;
lib/decompress_bunzip2.c
339
base[i+1] = pp-(t += temp[i]);
lib/decompress_bunzip2.c
416
t = 0;
lib/decompress_bunzip2.c
428
t += (runPos << nextSym);
lib/decompress_bunzip2.c
442
if (dbufCount+t >= dbufSize)
lib/decompress_bunzip2.c
446
byteCount[uc] += t;
lib/decompress_bunzip2.c
447
while (t--)
lib/decompress_unlzma.c
158
uint32_t t = rc_is_bit_0_helper(rc, p);
lib/decompress_unlzma.c
159
return rc->code < t;
lib/dynamic_queue_limits.c
28
unsigned long hist_head, t, start, end;
lib/dynamic_queue_limits.c
58
for (t = start; time_before_eq(t, end); t++)
lib/dynamic_queue_limits.c
59
if (test_bit(t % (DQL_HIST_LEN * BITS_PER_LONG),
lib/dynamic_queue_limits.c
64
if (!time_before_eq(t, end))
lib/dynamic_queue_limits.c
72
dql->stall_max = max_t(unsigned short, dql->stall_max, now - t);
lib/dynamic_queue_limits.c
74
trace_dql_stall_detected(dql->stall_thrs, now - t,
lib/inflate.c
1039
unsigned t; /* block type */
lib/inflate.c
1058
t = (unsigned)b & 3;
lib/inflate.c
1067
if (t == 2)
lib/inflate.c
1069
if (t == 0)
lib/inflate.c
1071
if (t == 1)
lib/inflate.c
144
struct huft *t; /* pointer to next level of table */
lib/inflate.c
328
struct huft **t, /* result: starting table */
lib/inflate.c
384
*t = (struct huft *)NULL;
lib/inflate.c
498
*t = q + 1; /* link to list for huft_free() */
lib/inflate.c
499
*(t = &(q->v.t)) = (struct huft *)NULL;
lib/inflate.c
509
r.v.t = q; /* pointer to this table */
lib/inflate.c
568
struct huft *t /* table to free */
lib/inflate.c
578
p = t;
lib/inflate.c
581
q = (--p)->v.t;
lib/inflate.c
601
struct huft *t; /* pointer to table entry */
lib/inflate.c
618
if ((e = (t = tl + ((unsigned)b & ml))->e) > 16)
lib/inflate.c
622
DUMPBITS(t->b)
lib/inflate.c
625
} while ((e = (t = t->v.t + ((unsigned)b & mask_bits[e]))->e) > 16);
lib/inflate.c
626
DUMPBITS(t->b)
lib/inflate.c
629
slide[w++] = (uch)t->v.n;
lib/inflate.c
645
n = t->v.n + ((unsigned)b & mask_bits[e]);
lib/inflate.c
650
if ((e = (t = td + ((unsigned)b & md))->e) > 16)
lib/inflate.c
654
DUMPBITS(t->b)
lib/inflate.c
657
} while ((e = (t = t->v.t + ((unsigned)b & mask_bits[e]))->e) > 16);
lib/inflate.c
658
DUMPBITS(t->b)
lib/inflate.c
660
d = w - t->v.n - ((unsigned)b & mask_bits[e]);
lib/is_single_threaded.c
19
struct task_struct *p, *t;
lib/is_single_threaded.c
36
for_each_thread(p, t) {
lib/is_single_threaded.c
37
if (unlikely(t->mm == mm))
lib/is_single_threaded.c
39
if (likely(t->mm))
lib/kobject.c
285
char *t;
lib/kobject.c
287
t = kstrdup(s, GFP_KERNEL);
lib/kobject.c
289
if (!t)
lib/kobject.c
291
s = strreplace(t, '/', '!');
lib/kobject.c
666
const struct kobj_type *t = get_ktype(kobj);
lib/kobject.c
672
if (t && !t->release)
lib/kobject.c
686
if (t && t->release) {
lib/kobject.c
689
t->release(kobj);
lib/locking-selftest.c
1404
I_WW(t); I_WW(t2); I_WW(o.base); I_WW(o2.base); I_WW(o3.base);
lib/locking-selftest.c
1417
memset(&t, 0, sizeof(t)); memset(&t2, 0, sizeof(t2));
lib/locking-selftest.c
1673
WWAI(&t);
lib/locking-selftest.c
1674
t.stamp++;
lib/locking-selftest.c
1676
ret = WWL(&o, &t);
lib/locking-selftest.c
1683
ret = WWL(&o, &t);
lib/locking-selftest.c
1689
t2 = t;
lib/locking-selftest.c
1771
WWAI(&t);
lib/locking-selftest.c
1775
ww_mutex_base_lock_nest_lock(&o.base, &t);
lib/locking-selftest.c
1782
WWAI(&t);
lib/locking-selftest.c
1788
WWAI(&t);
lib/locking-selftest.c
1790
t.ww_class = NULL;
lib/locking-selftest.c
1792
WWL(&o, &t);
lib/locking-selftest.c
1797
WWAI(&t);
lib/locking-selftest.c
1798
WWAD(&t);
lib/locking-selftest.c
1799
WWAD(&t);
lib/locking-selftest.c
1800
WWAF(&t);
lib/locking-selftest.c
1805
WWAI(&t);
lib/locking-selftest.c
1806
WWAD(&t);
lib/locking-selftest.c
1807
WWAF(&t);
lib/locking-selftest.c
1808
WWAF(&t);
lib/locking-selftest.c
1813
WWAI(&t);
lib/locking-selftest.c
1814
WWL(&o, &t);
lib/locking-selftest.c
1815
WWAD(&t);
lib/locking-selftest.c
1816
WWAF(&t);
lib/locking-selftest.c
1821
WWAI(&t);
lib/locking-selftest.c
1822
WWAD(&t);
lib/locking-selftest.c
1823
WWL(&o, &t);
lib/locking-selftest.c
1835
WWAI(&t);
lib/locking-selftest.c
1836
WWL(&o, &t);
lib/locking-selftest.c
1837
t.acquired = 0;
lib/locking-selftest.c
1839
WWAF(&t);
lib/locking-selftest.c
1844
WWAI(&t);
lib/locking-selftest.c
1846
WWL(&o, &t);
lib/locking-selftest.c
1857
WWAI(&t);
lib/locking-selftest.c
1858
t2 = t;
lib/locking-selftest.c
1861
ret = WWL(&o, &t);
lib/locking-selftest.c
1864
ret = WWL(&o2, &t);
lib/locking-selftest.c
1872
WWL(&o2, &t);
lib/locking-selftest.c
1883
WWAI(&t);
lib/locking-selftest.c
1884
t2 = t;
lib/locking-selftest.c
1887
ret = WWL(&o, &t);
lib/locking-selftest.c
1890
ret = WWL(&o2, &t);
lib/locking-selftest.c
1898
ww_mutex_lock_slow(&o2, &t);
lib/locking-selftest.c
1909
WWAI(&t);
lib/locking-selftest.c
1910
t2 = t;
lib/locking-selftest.c
1913
ret = WWL(&o, &t);
lib/locking-selftest.c
1916
ret = WWL(&o2, &t);
lib/locking-selftest.c
1923
WWL(&o2, &t);
lib/locking-selftest.c
1934
WWAI(&t);
lib/locking-selftest.c
1935
t2 = t;
lib/locking-selftest.c
1938
ret = WWL(&o, &t);
lib/locking-selftest.c
1941
ret = WWL(&o2, &t);
lib/locking-selftest.c
1948
ww_mutex_lock_slow(&o2, &t);
lib/locking-selftest.c
1959
WWAI(&t);
lib/locking-selftest.c
1960
t2 = t;
lib/locking-selftest.c
1963
ret = WWL(&o, &t);
lib/locking-selftest.c
1966
ret = WWL(&o2, &t);
lib/locking-selftest.c
1969
ret = WWL(&o3, &t);
lib/locking-selftest.c
1980
WWAI(&t);
lib/locking-selftest.c
1981
t2 = t;
lib/locking-selftest.c
1984
ret = WWL(&o, &t);
lib/locking-selftest.c
1987
ret = WWL(&o2, &t);
lib/locking-selftest.c
1990
ww_mutex_lock_slow(&o3, &t);
lib/locking-selftest.c
2005
WWAI(&t);
lib/locking-selftest.c
2006
t2 = t;
lib/locking-selftest.c
2009
ret = WWL(&o, &t);
lib/locking-selftest.c
2012
ret = WWL(&o2, &t);
lib/locking-selftest.c
2015
ret = WWL(&o3, &t);
lib/locking-selftest.c
2031
WWAI(&t);
lib/locking-selftest.c
2032
t2 = t;
lib/locking-selftest.c
2035
ret = WWL(&o, &t);
lib/locking-selftest.c
2038
ret = WWL(&o2, &t);
lib/locking-selftest.c
2041
ww_mutex_lock_slow(&o3, &t);
lib/locking-selftest.c
2052
WWAI(&t);
lib/locking-selftest.c
2053
t2 = t;
lib/locking-selftest.c
2056
ret = WWL(&o, &t);
lib/locking-selftest.c
2059
ret = WWL(&o2, &t);
lib/locking-selftest.c
2066
ret = WWL(&o3, &t);
lib/locking-selftest.c
2077
WWAI(&t);
lib/locking-selftest.c
2078
t2 = t;
lib/locking-selftest.c
2081
ret = WWL(&o, &t);
lib/locking-selftest.c
2084
ret = WWL(&o2, &t);
lib/locking-selftest.c
2091
ww_mutex_lock_slow(&o3, &t);
lib/locking-selftest.c
2115
WWAI(&t);
lib/locking-selftest.c
2117
ww_mutex_lock_slow(&o, &t);
lib/locking-selftest.c
2124
WWAI(&t);
lib/locking-selftest.c
2126
ret = WWL(&o, &t);
lib/locking-selftest.c
2135
WWAI(&t);
lib/locking-selftest.c
2137
ret = WWL(&o, &t);
lib/locking-selftest.c
2150
WWAI(&t);
lib/locking-selftest.c
2152
ret = WWL(&o, &t);
lib/locking-selftest.c
2155
ret = WWL(&o2, &t);
lib/locking-selftest.c
2193
WWAI(&t);
lib/locking-selftest.c
2195
ret = WWL(&o2, &t);
lib/locking-selftest.c
2219
WWAI(&t);
lib/locking-selftest.c
2221
ret = WWL(&o2, &t);
lib/locking-selftest.c
2268
WWAI(&t);
lib/locking-selftest.c
2270
ret = WWL(&o, &t);
lib/locking-selftest.c
2277
ret = WWL(&o, &t);
lib/locking-selftest.c
64
static struct ww_acquire_ctx t, t2;
lib/lwq.c
104
struct tnode *t;
lib/lwq.c
113
t = kmalloc_obj(*t);
lib/lwq.c
114
if (!t)
lib/lwq.c
116
t->i = i;
lib/lwq.c
117
t->c = 0;
lib/lwq.c
118
if (lwq_enqueue(&t->n, &q))
lib/lwq.c
131
t = lwq_dequeue(&q, struct tnode, n);
lib/lwq.c
132
if (t)
lib/lwq.c
133
printk(KERN_CONT " %d(%d)", t->i, t->c);
lib/lwq.c
134
kfree(t);
lib/lwq.c
139
lwq_for_each_safe(t, t1, t2, &l, n) {
lib/lwq.c
140
if ((t->i % 3) == 0) {
lib/lwq.c
141
t->i = -1;
lib/lwq.c
142
kfree(t);
lib/lwq.c
143
t = NULL;
lib/lwq.c
149
while ((t = lwq_dequeue(&q, struct tnode, n)) != NULL) {
lib/lwq.c
150
printk(KERN_CONT " %d", t->i);
lib/lwq.c
151
kfree(t);
lib/lwq.c
51
struct llist_node *r, *t, **ep;
lib/lwq.c
59
t = llist_del_all(&q->new);
lib/lwq.c
64
*ep = llist_reverse_order(t);
lib/lwq.c
86
struct tnode *t;
lib/lwq.c
89
wait_var_event(q, (t = lwq_dequeue(q, struct tnode, n)) != NULL);
lib/lwq.c
90
t->c++;
lib/lwq.c
91
if (lwq_enqueue(&t->n, q))
lib/lzo/lzo1x_compress.c
118
t = ((dv * 0x1824429d) >> (32 - D_BITS)) & D_MASK;
lib/lzo/lzo1x_compress.c
119
m_pos = in + dict[t];
lib/lzo/lzo1x_compress.c
120
dict[t] = (lzo_dict_t) (ip - in);
lib/lzo/lzo1x_compress.c
127
t = ip - ii;
lib/lzo/lzo1x_compress.c
128
if (t != 0) {
lib/lzo/lzo1x_compress.c
129
if (t <= 3) {
lib/lzo/lzo1x_compress.c
130
op[*state_offset] |= t;
lib/lzo/lzo1x_compress.c
133
op += t;
lib/lzo/lzo1x_compress.c
134
} else if (t <= 16) {
lib/lzo/lzo1x_compress.c
136
*op++ = (t - 3);
lib/lzo/lzo1x_compress.c
139
op += t;
lib/lzo/lzo1x_compress.c
141
if (t <= 18) {
lib/lzo/lzo1x_compress.c
143
*op++ = (t - 3);
lib/lzo/lzo1x_compress.c
145
size_t tt = t - 18;
lib/lzo/lzo1x_compress.c
156
NEED_OP(t);
lib/lzo/lzo1x_compress.c
162
t -= 16;
lib/lzo/lzo1x_compress.c
163
} while (t >= 16);
lib/lzo/lzo1x_compress.c
164
if (t > 0) do {
lib/lzo/lzo1x_compress.c
166
} while (--t > 0);
lib/lzo/lzo1x_compress.c
345
size_t t = 0;
lib/lzo/lzo1x_compress.c
366
if ((ll_end + ((t + ll) >> 5)) <= ll_end)
lib/lzo/lzo1x_compress.c
371
ip, ll, &op, op_end, &t, wrkmem,
lib/lzo/lzo1x_compress.c
378
t += l;
lib/lzo/lzo1x_compress.c
380
if (t > 0) {
lib/lzo/lzo1x_compress.c
381
const unsigned char *ii = in + in_len - t;
lib/lzo/lzo1x_compress.c
383
if (op == data_start && t <= 238) {
lib/lzo/lzo1x_compress.c
385
*op++ = (17 + t);
lib/lzo/lzo1x_compress.c
386
} else if (t <= 3) {
lib/lzo/lzo1x_compress.c
387
op[state_offset] |= t;
lib/lzo/lzo1x_compress.c
388
} else if (t <= 18) {
lib/lzo/lzo1x_compress.c
390
*op++ = (t - 3);
lib/lzo/lzo1x_compress.c
392
size_t tt = t - 18;
lib/lzo/lzo1x_compress.c
403
NEED_OP(t);
lib/lzo/lzo1x_compress.c
404
if (t >= 16) do {
lib/lzo/lzo1x_compress.c
409
t -= 16;
lib/lzo/lzo1x_compress.c
410
} while (t >= 16);
lib/lzo/lzo1x_compress.c
411
if (t > 0) do {
lib/lzo/lzo1x_compress.c
413
} while (--t > 0);
lib/lzo/lzo1x_compress.c
53
size_t t, m_len, m_off;
lib/lzo/lzo1x_decompress_safe.c
112
NEED_OP(t);
lib/lzo/lzo1x_decompress_safe.c
113
NEED_IP(t + 3);
lib/lzo/lzo1x_decompress_safe.c
116
} while (--t > 0);
lib/lzo/lzo1x_decompress_safe.c
121
next = t & 3;
lib/lzo/lzo1x_decompress_safe.c
123
m_pos -= t >> 2;
lib/lzo/lzo1x_decompress_safe.c
132
next = t & 3;
lib/lzo/lzo1x_decompress_safe.c
134
m_pos -= t >> 2;
lib/lzo/lzo1x_decompress_safe.c
136
t = 3;
lib/lzo/lzo1x_decompress_safe.c
138
} else if (t >= 64) {
lib/lzo/lzo1x_decompress_safe.c
139
next = t & 3;
lib/lzo/lzo1x_decompress_safe.c
141
m_pos -= (t >> 2) & 7;
lib/lzo/lzo1x_decompress_safe.c
143
t = (t >> 5) - 1 + (3 - 1);
lib/lzo/lzo1x_decompress_safe.c
144
} else if (t >= 32) {
lib/lzo/lzo1x_decompress_safe.c
145
t = (t & 31) + (3 - 1);
lib/lzo/lzo1x_decompress_safe.c
146
if (unlikely(t == 2)) {
lib/lzo/lzo1x_decompress_safe.c
159
t += offset + 31 + *ip++;
lib/lzo/lzo1x_decompress_safe.c
171
((t & 0xf8) == 0x18) &&
lib/lzo/lzo1x_decompress_safe.c
174
t &= 7;
lib/lzo/lzo1x_decompress_safe.c
175
t |= ip[2] << 3;
lib/lzo/lzo1x_decompress_safe.c
176
t += MIN_ZERO_RUN_LENGTH;
lib/lzo/lzo1x_decompress_safe.c
177
NEED_OP(t);
lib/lzo/lzo1x_decompress_safe.c
178
memset(op, 0, t);
lib/lzo/lzo1x_decompress_safe.c
179
op += t;
lib/lzo/lzo1x_decompress_safe.c
185
m_pos -= (t & 8) << 11;
lib/lzo/lzo1x_decompress_safe.c
186
t = (t & 7) + (3 - 1);
lib/lzo/lzo1x_decompress_safe.c
187
if (unlikely(t == 2)) {
lib/lzo/lzo1x_decompress_safe.c
200
t += offset + 7 + *ip++;
lib/lzo/lzo1x_decompress_safe.c
215
unsigned char *oe = op + t;
lib/lzo/lzo1x_decompress_safe.c
216
if (likely(HAVE_OP(t + 15))) {
lib/lzo/lzo1x_decompress_safe.c
234
NEED_OP(t);
lib/lzo/lzo1x_decompress_safe.c
242
unsigned char *oe = op + t;
lib/lzo/lzo1x_decompress_safe.c
243
NEED_OP(t);
lib/lzo/lzo1x_decompress_safe.c
254
t = next;
lib/lzo/lzo1x_decompress_safe.c
258
op += t;
lib/lzo/lzo1x_decompress_safe.c
259
ip += t;
lib/lzo/lzo1x_decompress_safe.c
263
NEED_IP(t + 3);
lib/lzo/lzo1x_decompress_safe.c
264
NEED_OP(t);
lib/lzo/lzo1x_decompress_safe.c
265
while (t > 0) {
lib/lzo/lzo1x_decompress_safe.c
267
t--;
lib/lzo/lzo1x_decompress_safe.c
274
return (t != 3 ? LZO_E_ERROR :
lib/lzo/lzo1x_decompress_safe.c
44
size_t t, next;
lib/lzo/lzo1x_decompress_safe.c
66
t = *ip++ - 17;
lib/lzo/lzo1x_decompress_safe.c
67
if (t < 4) {
lib/lzo/lzo1x_decompress_safe.c
68
next = t;
lib/lzo/lzo1x_decompress_safe.c
75
t = *ip++;
lib/lzo/lzo1x_decompress_safe.c
76
if (t < 16) {
lib/lzo/lzo1x_decompress_safe.c
78
if (unlikely(t == 0)) {
lib/lzo/lzo1x_decompress_safe.c
91
t += offset + 15 + *ip++;
lib/lzo/lzo1x_decompress_safe.c
93
t += 3;
lib/lzo/lzo1x_decompress_safe.c
96
if (likely(HAVE_IP(t + 15) && HAVE_OP(t + 15))) {
lib/lzo/lzo1x_decompress_safe.c
97
const unsigned char *ie = ip + t;
lib/lzo/lzo1x_decompress_safe.c
98
unsigned char *oe = op + t;
lib/math/div64.c
164
s64 quot, t;
lib/math/div64.c
167
t = (dividend ^ divisor) >> 63;
lib/math/div64.c
169
return (quot ^ t) - t;
lib/math/rational.c
83
unsigned long t = ULONG_MAX;
lib/math/rational.c
86
t = (max_denominator - d0) / d1;
lib/math/rational.c
88
t = min(t, (max_numerator - n0) / n1);
lib/math/rational.c
94
if (!d1 || 2u * t > a || (2u * t == a && d0 * dp > d1 * d)) {
lib/math/rational.c
95
n1 = n0 + t * n1;
lib/math/rational.c
96
d1 = d0 + t * d1;
lib/muldi3.c
11
#define __ll_lowpart(t) ((unsigned long) (t) & (__ll_B - 1))
lib/muldi3.c
12
#define __ll_highpart(t) ((unsigned long) (t) >> (W_TYPE_SIZE / 2))
lib/plist.c
32
static void plist_check_prev_next(struct list_head *t, struct list_head *p,
lib/plist.c
39
t, t->next, t->prev,
lib/reed_solomon/decode_rs.c
152
t[0] = lambda[0];
lib/reed_solomon/decode_rs.c
155
t[i + 1] = lambda[i + 1] ^
lib/reed_solomon/decode_rs.c
159
t[i + 1] = lambda[i + 1];
lib/reed_solomon/decode_rs.c
177
memcpy(lambda, t, (nroots + 1) * sizeof(t[0]));
lib/reed_solomon/decode_rs.c
35
uint16_t *t = rsc->buffers + RS_DECODE_T * (nroots + 1);
lib/refcount.c
13
void refcount_warn_saturate(refcount_t *r, enum refcount_saturation_type t)
lib/refcount.c
17
switch (t) {
lib/sort.c
112
char t = ((char *)a)[--n];
lib/sort.c
114
((char *)b)[n] = t;
lib/sort.c
59
u32 t = *(u32 *)(a + (n -= 4));
lib/sort.c
61
*(u32 *)(b + n) = t;
lib/sort.c
85
u64 t = *(u64 *)(a + (n -= 8));
lib/sort.c
87
*(u64 *)(b + n) = t;
lib/sort.c
90
u32 t = *(u32 *)(a + (n -= 4));
lib/sort.c
92
*(u32 *)(b + n) = t;
lib/sort.c
94
t = *(u32 *)(a + (n -= 4));
lib/sort.c
96
*(u32 *)(b + n) = t;
lib/test-kstrtox.c
31
const struct test_fail *t = &test[i]; \
lib/test-kstrtox.c
36
rv = fn(t->str, t->base, &tmp); \
lib/test-kstrtox.c
39
t->str, t->base, rv, tmp); \
lib/test-kstrtox.c
50
const typeof(test[0]) *t = &test[i]; \
lib/test-kstrtox.c
54
rv = fn(t->str, t->base, &res); \
lib/test-kstrtox.c
57
t->str, t->base, t->expected_res, rv); \
lib/test-kstrtox.c
60
if (res != t->expected_res) { \
lib/test-kstrtox.c
62
t->str, t->base, t->expected_res, res); \
lib/test_bitmap.c
1057
struct test_bitmap_cut *t = &test_cut[i];
lib/test_bitmap.c
1059
memcpy(in, t->in, sizeof(t->in));
lib/test_bitmap.c
1061
bitmap_cut(out, in, t->first, t->cut, t->nbits);
lib/test_bitmap.c
1063
expect_eq_bitmap(t->expected, out, t->nbits);
lib/test_bitmap.c
1177
const struct test_bitmap_print *t = &test_print[i];
lib/test_bitmap.c
1180
n = bitmap_print_bitmask_to_buf(print_buf, t->bitmap, t->nbits,
lib/test_bitmap.c
1182
expect_eq_uint(strlen(t->mask) + 1, n);
lib/test_bitmap.c
1183
expect_eq_str(t->mask, print_buf, n);
lib/test_bitmap.c
1185
n = bitmap_print_list_to_buf(print_buf, t->bitmap, t->nbits,
lib/test_bitmap.c
1187
expect_eq_uint(strlen(t->list) + 1, n);
lib/test_bitmap.c
1188
expect_eq_str(t->list, print_buf, n);
lib/test_bitmap.c
1191
if (strlen(t->list) > PAGE_SIZE) {
lib/test_bitmap.c
1192
n = bitmap_print_list_to_buf(print_buf, t->bitmap, t->nbits,
lib/test_bitmap.c
1194
expect_eq_uint(strlen(t->list) + 1 - PAGE_SIZE, n);
lib/test_bitmap.c
1195
expect_eq_str(t->list + PAGE_SIZE, print_buf, n);
lib/test_ref_tracker.c
57
static void test_ref_tracker_timer_func(struct timer_list *t)
lib/test_vmalloc.c
476
struct test_driver *t = private;
lib/test_vmalloc.c
493
t->start = get_cycles();
lib/test_vmalloc.c
507
t->data[index].test_passed++;
lib/test_vmalloc.c
509
t->data[index].test_xfailed++;
lib/test_vmalloc.c
511
t->data[index].test_failed++;
lib/test_vmalloc.c
520
t->data[index].time = delta;
lib/test_vmalloc.c
522
t->stop = get_cycles();
lib/test_vmalloc.c
575
struct test_driver *t = &tdriver[i];
lib/test_vmalloc.c
577
t->task = kthread_run(test_func, t, "vmalloc_test/%d", i);
lib/test_vmalloc.c
579
if (!IS_ERR(t->task))
lib/test_vmalloc.c
602
struct test_driver *t = &tdriver[i];
lib/test_vmalloc.c
605
if (!IS_ERR(t->task))
lib/test_vmalloc.c
606
kthread_stop(t->task);
lib/test_vmalloc.c
615
t->data[j].test_passed,
lib/test_vmalloc.c
616
t->data[j].test_failed,
lib/test_vmalloc.c
617
t->data[j].test_xfailed,
lib/test_vmalloc.c
619
t->data[j].time);
lib/test_vmalloc.c
623
i, t->stop - t->start);
lib/tests/glob_kunit.c
95
static void glob_case_to_desc(const struct glob_test_case *t, char *desc)
lib/tests/glob_kunit.c
97
snprintf(desc, KUNIT_PARAM_DESC_SIZE, "pat:\"%s\" str:\"%s\"", t->pat, t->str);
lib/tests/overflow_kunit.c
1038
# define TEST_SAME_TYPE64(base, t, m) TEST_SAME_TYPE(base, t, m)
lib/tests/overflow_kunit.c
1040
# define TEST_SAME_TYPE64(base, t, m) do { } while (0)
lib/tests/overflow_kunit.c
253
#define check_one_op(t, fmt, op, sym, a, b, r, of) do { \
lib/tests/overflow_kunit.c
257
t _r; \
lib/tests/overflow_kunit.c
262
a, b, of ? "" : " not", #t); \
lib/tests/overflow_kunit.c
265
a, b, r, _r, #t); \
lib/tests/overflow_kunit.c
273
_r = wrapping_ ## op(t, a, b); \
lib/tests/overflow_kunit.c
276
a, b, r, _r, #t); \
lib/tests/overflow_kunit.c
280
_r = wrapping_ ## op(t, _a_orig++, _b_orig++); \
lib/tests/overflow_kunit.c
32
# define SKIP_SIGN_MISMATCH(t) SKIP(t, "Clang 11 unwanted libcalls")
lib/tests/overflow_kunit.c
326
#define DEFINE_TEST_FUNC_TYPED(n, t, fmt) \
lib/tests/overflow_kunit.c
330
check_one_op(t, fmt, add, "+", p->a, p->b, p->sum, p->s_of); \
lib/tests/overflow_kunit.c
331
check_one_op(t, fmt, add, "+", p->b, p->a, p->sum, p->s_of); \
lib/tests/overflow_kunit.c
332
check_one_op(t, fmt, sub, "-", p->a, p->b, p->diff, p->d_of); \
lib/tests/overflow_kunit.c
333
check_one_op(t, fmt, mul, "*", p->a, p->b, p->prod, p->p_of); \
lib/tests/overflow_kunit.c
334
check_one_op(t, fmt, mul, "*", p->b, p->a, p->prod, p->p_of); \
lib/tests/overflow_kunit.c
34
# define SKIP_SIGN_MISMATCH(t) do { } while (0)
lib/tests/overflow_kunit.c
344
SKIP_64_ON_32(__same_type(t, u64)); \
lib/tests/overflow_kunit.c
345
SKIP_64_ON_32(__same_type(t, s64)); \
lib/tests/overflow_kunit.c
356
#define DEFINE_TEST_FUNC(t, fmt) \
lib/tests/overflow_kunit.c
357
DEFINE_TEST_FUNC_TYPED(t ## _ ## t ## __ ## t, t, fmt)
lib/tests/overflow_kunit.c
396
#define TEST_ONE_SHIFT(a, s, t, expect, of) do { \
lib/tests/overflow_kunit.c
399
t __e = (expect); \
lib/tests/overflow_kunit.c
400
t __d; \
lib/tests/overflow_kunit.c
405
#t, #a, #s, of ? "" : " not"); \
lib/tests/overflow_kunit.c
409
#t, #a, #s, #expect); \
lib/tests/overflow_kunit.c
410
if ((t)-1 < 0) \
lib/tests/overflow_kunit.c
43
# define SKIP_64_ON_32(t) SKIP(t, "Clang 13 unwanted libcalls")
lib/tests/overflow_kunit.c
45
# define SKIP_64_ON_32(t) do { } while (0)
lib/tests/overflow_kunit.c
48
#define DEFINE_TEST_ARRAY_TYPED(t1, t2, t) \
lib/tests/overflow_kunit.c
49
static const struct test_ ## t1 ## _ ## t2 ## __ ## t { \
lib/tests/overflow_kunit.c
52
t sum, diff, prod; \
lib/tests/overflow_kunit.c
54
} t1 ## _ ## t2 ## __ ## t ## _tests[]
lib/tests/overflow_kunit.c
56
#define DEFINE_TEST_ARRAY(t) DEFINE_TEST_ARRAY_TYPED(t, t, t)
lib/tests/printf_kunit.c
508
time64_t t = 1546615943;
lib/tests/printf_kunit.c
509
struct timespec64 ts = { .tv_sec = t, .tv_nsec = 11235813 };
lib/tests/printf_kunit.c
519
test("2019-01-04T15:32:23", "%ptT", &t);
lib/tests/printf_kunit.c
520
test("0119-00-04T15:32:23", "%ptTr", &t);
lib/tests/printf_kunit.c
521
test("15:32:23|2019-01-04", "%ptTt|%ptTd", &t, &t);
lib/tests/printf_kunit.c
522
test("15:32:23|0119-00-04", "%ptTtr|%ptTdr", &t, &t);
lib/tests/printf_kunit.c
524
test("2019-01-04 15:32:23", "%ptTs", &t);
lib/tests/printf_kunit.c
525
test("0119-00-04 15:32:23", "%ptTsr", &t);
lib/tests/printf_kunit.c
526
test("15:32:23|2019-01-04", "%ptTts|%ptTds", &t, &t);
lib/tests/printf_kunit.c
527
test("15:32:23|0119-00-04", "%ptTtrs|%ptTdrs", &t, &t);
lib/tests/test_bits.c
10
#define assert_type(t, x) _Generic(x, t: x, default: 0)
lib/ts_bm.c
64
u8 t = *(text-i);
lib/ts_bm.c
67
t = toupper(t);
lib/ts_bm.c
69
if (t != *(pattern-i))
lib/ts_fsm.c
126
static inline int match_token(struct ts_fsm_token *t, u8 d)
lib/ts_fsm.c
128
if (t->type)
lib/ts_fsm.c
129
return (token_lookup_tbl[d] & t->type) != 0;
lib/ts_fsm.c
131
return t->value == d;
lib/ts_fsm.c
272
struct ts_fsm_token *t = &tokens[i];
lib/ts_fsm.c
274
if (t->type > TS_FSM_TYPE_MAX || t->recur > TS_FSM_RECUR_MAX)
lib/ts_fsm.c
277
if (t->recur == TS_FSM_HEAD_IGNORE &&
lib/ts_fsm.c
292
struct ts_fsm_token *t = &fsm->tokens[i];
lib/ts_fsm.c
293
t->type = token_map[t->type];
lib/vdso/gettimeofday.c
401
__kernel_old_time_t t;
lib/vdso/gettimeofday.c
409
t = READ_ONCE(vc[CS_HRES_COARSE].basetime[CLOCK_REALTIME].sec);
lib/vdso/gettimeofday.c
412
*time = t;
lib/vdso/gettimeofday.c
414
return t;
lib/win_minmax.c
32
u32 dt = val->t - m->s[0].t;
lib/win_minmax.c
45
if (unlikely(val->t - m->s[0].t > win)) {
lib/win_minmax.c
50
} else if (unlikely(m->s[1].t == m->s[0].t) && dt > win/4) {
lib/win_minmax.c
56
} else if (unlikely(m->s[2].t == m->s[1].t) && dt > win/2) {
lib/win_minmax.c
67
u32 minmax_running_max(struct minmax *m, u32 win, u32 t, u32 meas)
lib/win_minmax.c
69
struct minmax_sample val = { .t = t, .v = meas };
lib/win_minmax.c
72
unlikely(val.t - m->s[2].t > win)) /* nothing left in window? */
lib/win_minmax.c
73
return minmax_reset(m, t, meas); /* forget earlier samples */
lib/win_minmax.c
85
u32 minmax_running_min(struct minmax *m, u32 win, u32 t, u32 meas)
lib/win_minmax.c
87
struct minmax_sample val = { .t = t, .v = meas };
lib/win_minmax.c
90
unlikely(val.t - m->s[2].t > win)) /* nothing left in window? */
lib/win_minmax.c
91
return minmax_reset(m, t, meas); /* forget earlier samples */
mm/damon/core.c
1127
struct damon_target *t;
mm/damon/core.c
1130
damon_for_each_target(t, ctx) {
mm/damon/core.c
1132
return t;
mm/damon/core.c
1302
struct damon_target *t;
mm/damon/core.c
1306
damon_for_each_target(t, ctx) {
mm/damon/core.c
1307
damon_for_each_region(r, t)
mm/damon/core.c
137
void damon_add_region(struct damon_region *r, struct damon_target *t)
mm/damon/core.c
139
list_add_tail(&r->list, &t->regions_list);
mm/damon/core.c
140
t->nr_regions++;
mm/damon/core.c
143
static void damon_del_region(struct damon_region *r, struct damon_target *t)
mm/damon/core.c
146
t->nr_regions--;
mm/damon/core.c
154
void damon_destroy_region(struct damon_region *r, struct damon_target *t)
mm/damon/core.c
156
damon_del_region(r, t);
mm/damon/core.c
1598
struct damon_target *t;
mm/damon/core.c
1601
damon_for_each_target(t, c) {
mm/damon/core.c
1604
damon_for_each_region(r, t) {
mm/damon/core.c
1605
trace_damon_aggregated(ti, r, damon_nr_regions(t));
mm/damon/core.c
161
struct damon_target *t)
mm/damon/core.c
1616
struct damon_target *t;
mm/damon/core.c
1622
damon_for_each_target(t, c) {
mm/damon/core.c
1623
damon_for_each_region(r, t) {
mm/damon/core.c
163
return list_is_last(&r->list, &t->regions_list);
mm/damon/core.c
1675
static void damon_split_region_at(struct damon_target *t,
mm/damon/core.c
1692
static bool damos_valid_target(struct damon_ctx *c, struct damon_target *t,
mm/damon/core.c
1700
return c->ops.get_scheme_score(c, t, r, s) >= s->quota.min_score;
mm/damon/core.c
1726
static bool damos_skip_charged_region(struct damon_target *t,
mm/damon/core.c
1736
if (t != quota->charge_target_from)
mm/damon/core.c
1738
if (r == damon_last_region(t)) {
mm/damon/core.c
1756
damon_split_region_at(t, r, sz_to_skip);
mm/damon/core.c
1778
static bool damos_filter_match(struct damon_ctx *ctx, struct damon_target *t,
mm/damon/core.c
1790
if (ti == t)
mm/damon/core.c
181
struct damon_region *last, struct damon_target *t)
mm/damon/core.c
1812
damon_split_region_at(t, r, start - r->ar.start);
mm/damon/core.c
1817
damon_split_region_at(t, r, end - r->ar.start);
mm/damon/core.c
1827
static bool damos_core_filter_out(struct damon_ctx *ctx, struct damon_target *t,
mm/damon/core.c
1834
if (damos_filter_match(ctx, t, r, filter, ctx->min_region_sz)) {
mm/damon/core.c
185
damon_for_each_region_from(r, t) {
mm/damon/core.c
1854
static void damos_walk_call_walk(struct damon_ctx *ctx, struct damon_target *t,
mm/damon/core.c
1867
control->walk_fn(control->data, ctx, t, r, s, sz_filter_passed);
mm/damon/core.c
1930
static void damos_apply_scheme(struct damon_ctx *c, struct damon_target *t,
mm/damon/core.c
195
damon_insert_region(newr, r, next, t);
mm/damon/core.c
1959
if (titer == t)
mm/damon/core.c
1972
damon_split_region_at(t, r, sz);
mm/damon/core.c
1974
if (damos_core_filter_out(c, t, r, s))
mm/damon/core.c
1978
damon_nr_regions(t), do_trace);
mm/damon/core.c
1979
sz_applied = c->ops.apply_scheme(c, t, r, s,
mm/damon/core.c
1981
damos_walk_call_walk(c, t, r, s, sz_ops_filter_passed);
mm/damon/core.c
1987
quota->charge_target_from = t;
mm/damon/core.c
1999
struct damon_target *t,
mm/damon/core.c
2017
if (damos_skip_charged_region(t, &r, s, c->min_region_sz))
mm/damon/core.c
2024
if (damos_valid_target(c, t, r, s))
mm/damon/core.c
2025
damos_apply_scheme(c, t, r, s);
mm/damon/core.c
2027
if (damon_is_last_region(r, t))
mm/damon/core.c
213
int damon_set_regions(struct damon_target *t, struct damon_addr_range *ranges,
mm/damon/core.c
221
damon_for_each_region_safe(r, next, t) {
mm/damon/core.c
227
damon_destroy_region(r, t);
mm/damon/core.c
2282
struct damon_target *t;
mm/damon/core.c
230
r = damon_first_region(t);
mm/damon/core.c
2318
damon_for_each_target(t, c) {
mm/damon/core.c
2319
damon_for_each_region(r, t) {
mm/damon/core.c
2322
score = c->ops.get_scheme_score(c, t, r, s);
mm/damon/core.c
2357
struct damon_target *t;
mm/damon/core.c
238
damon_for_each_region_from(r, t) {
mm/damon/core.c
2380
damon_for_each_target(t, c) {
mm/damon/core.c
2381
if (c->ops.target_valid && c->ops.target_valid(t) == false)
mm/damon/core.c
2384
damon_for_each_region_safe(r, next_r, t)
mm/damon/core.c
2385
damon_do_apply_schemes(c, t, r);
mm/damon/core.c
2404
static void damon_merge_two_regions(struct damon_target *t,
mm/damon/core.c
2414
damon_destroy_region(r, t);
mm/damon/core.c
2424
static void damon_merge_regions_of(struct damon_target *t, unsigned int thres,
mm/damon/core.c
2429
damon_for_each_region_safe(r, next, t) {
mm/damon/core.c
2440
damon_merge_two_regions(t, prev, r);
mm/damon/core.c
2466
struct damon_target *t;
mm/damon/core.c
2474
damon_for_each_target(t, c) {
mm/damon/core.c
2475
damon_merge_regions_of(t, threshold, sz_limit);
mm/damon/core.c
2476
nr_regions += damon_nr_regions(t);
mm/damon/core.c
2489
static void damon_split_region_at(struct damon_target *t,
mm/damon/core.c
2505
damon_insert_region(new, r, damon_next_region(r), t);
mm/damon/core.c
2509
static void damon_split_regions_of(struct damon_target *t, int nr_subs,
mm/damon/core.c
2516
damon_for_each_region_safe(r, next, t) {
mm/damon/core.c
2531
damon_split_region_at(t, r, sz_sub);
mm/damon/core.c
2549
struct damon_target *t;
mm/damon/core.c
255
damon_insert_region(newr, damon_prev_region(r), r, t);
mm/damon/core.c
2554
damon_for_each_target(t, ctx)
mm/damon/core.c
2555
nr_regions += damon_nr_regions(t);
mm/damon/core.c
2565
damon_for_each_target(t, ctx)
mm/damon/core.c
2566
damon_split_regions_of(t, nr_subregions, ctx->min_region_sz);
mm/damon/core.c
2581
struct damon_target *t;
mm/damon/core.c
2589
damon_for_each_target(t, ctx) {
mm/damon/core.c
2590
if (ctx->ops.target_valid(t))
mm/damon/core.c
263
err = damon_fill_regions_holes(first, last, t);
mm/damon/core.c
2921
int damon_set_region_biggest_system_ram_default(struct damon_target *t,
mm/damon/core.c
2936
return damon_set_regions(t, &addr_range, 1, min_region_sz);
mm/damon/core.c
474
struct damon_target *t;
mm/damon/core.c
476
t = kmalloc_obj(*t);
mm/damon/core.c
477
if (!t)
mm/damon/core.c
480
t->pid = NULL;
mm/damon/core.c
481
t->nr_regions = 0;
mm/damon/core.c
482
INIT_LIST_HEAD(&t->regions_list);
mm/damon/core.c
483
INIT_LIST_HEAD(&t->list);
mm/damon/core.c
484
t->obsolete = false;
mm/damon/core.c
486
return t;
mm/damon/core.c
489
void damon_add_target(struct damon_ctx *ctx, struct damon_target *t)
mm/damon/core.c
491
list_add_tail(&t->list, &ctx->adaptive_targets);
mm/damon/core.c
499
static void damon_del_target(struct damon_target *t)
mm/damon/core.c
501
list_del(&t->list);
mm/damon/core.c
504
void damon_free_target(struct damon_target *t)
mm/damon/core.c
508
damon_for_each_region_safe(r, next, t)
mm/damon/core.c
510
kfree(t);
mm/damon/core.c
513
void damon_destroy_target(struct damon_target *t, struct damon_ctx *ctx)
mm/damon/core.c
517
ctx->ops.cleanup_target(t);
mm/damon/core.c
519
damon_del_target(t);
mm/damon/core.c
520
damon_free_target(t);
mm/damon/core.c
523
unsigned int damon_nr_regions(struct damon_target *t)
mm/damon/core.c
525
return t->nr_regions;
mm/damon/core.c
566
struct damon_target *t, *next_t;
mm/damon/core.c
568
damon_for_each_target_safe(t, next_t, ctx)
mm/damon/core.c
569
damon_destroy_target(t, ctx);
mm/damon/core.c
671
struct damon_target *t;
mm/damon/core.c
680
damon_for_each_target(t, ctx)
mm/damon/core.c
681
damon_for_each_region(r, t)
mm/damon/paddr.c
108
struct damon_target *t;
mm/damon/paddr.c
112
damon_for_each_target(t, ctx) {
mm/damon/paddr.c
113
damon_for_each_region(r, t) {
mm/damon/paddr.c
319
struct damon_target *t, struct damon_region *r,
mm/damon/paddr.c
346
struct damon_target *t, struct damon_region *r,
mm/damon/paddr.c
61
struct damon_target *t;
mm/damon/paddr.c
64
damon_for_each_target(t, ctx) {
mm/damon/paddr.c
65
damon_for_each_region(r, t)
mm/damon/stat.c
178
static int damon_stat_set_monitoring_region(struct damon_target *t,
mm/damon/stat.c
193
return damon_set_regions(t, &addr_range, 1, min_region_sz);
mm/damon/stat.c
53
struct damon_target *t;
mm/damon/stat.c
57
damon_for_each_target(t, c) {
mm/damon/stat.c
58
damon_for_each_region(r, t)
mm/damon/stat.c
85
struct damon_target *t;
mm/damon/stat.c
91
damon_for_each_target(t, c) {
mm/damon/stat.c
94
damon_nr_regions(t));
mm/damon/stat.c
97
damon_for_each_region(r, t) {
mm/damon/sysfs-common.h
49
struct damon_ctx *ctx, struct damon_target *t,
mm/damon/sysfs-schemes.c
2824
struct damon_ctx *ctx, struct damon_target *t,
mm/damon/sysfs.c
1363
static int damon_sysfs_set_regions(struct damon_target *t,
mm/damon/sysfs.c
1388
err = damon_set_regions(t, ranges, sysfs_regions->nr, min_region_sz);
mm/damon/sysfs.c
1398
struct damon_target *t = damon_new_target();
mm/damon/sysfs.c
1400
if (!t)
mm/damon/sysfs.c
1402
damon_add_target(ctx, t);
mm/damon/sysfs.c
1404
t->pid = find_get_pid(sys_target->pid);
mm/damon/sysfs.c
1405
if (!t->pid)
mm/damon/sysfs.c
1409
t->obsolete = sys_target->obsolete;
mm/damon/sysfs.c
1410
return damon_sysfs_set_regions(t, sys_target->regions,
mm/damon/sysfs.c
1711
struct damon_target *t, struct damon_region *r,
mm/damon/sysfs.c
1719
ctx, t, r, s, walk_data->total_bytes_only,
mm/damon/tests/core-kunit.h
1003
t = damon_new_target();
mm/damon/tests/core-kunit.h
1004
if (!t)
mm/damon/tests/core-kunit.h
1010
damon_free_target(t);
mm/damon/tests/core-kunit.h
1013
damon_add_region(r, t);
mm/damon/tests/core-kunit.h
1015
return t;
mm/damon/tests/core-kunit.h
104
t = damon_new_target();
mm/damon/tests/core-kunit.h
105
if (!t) {
mm/damon/tests/core-kunit.h
1062
struct damon_target *t;
mm/damon/tests/core-kunit.h
1071
t = damon_new_target();
mm/damon/tests/core-kunit.h
1072
if (!t) {
mm/damon/tests/core-kunit.h
1079
damon_free_target(t);
mm/damon/tests/core-kunit.h
1082
damon_add_region(r, t);
mm/damon/tests/core-kunit.h
1085
KUNIT_EXPECT_TRUE(test, damos_filter_match(NULL, t, r, f, 1));
mm/damon/tests/core-kunit.h
1086
KUNIT_EXPECT_EQ(test, damon_nr_regions(t), 1);
mm/damon/tests/core-kunit.h
109
damon_add_target(ctx, t);
mm/damon/tests/core-kunit.h
1092
damos_filter_match(NULL, t, r, f, 1));
mm/damon/tests/core-kunit.h
1093
KUNIT_EXPECT_EQ(test, damon_nr_regions(t), 1);
mm/damon/tests/core-kunit.h
1099
damos_filter_match(NULL, t, r, f, 1));
mm/damon/tests/core-kunit.h
1100
KUNIT_EXPECT_EQ(test, damon_nr_regions(t), 1);
mm/damon/tests/core-kunit.h
1105
KUNIT_EXPECT_FALSE(test, damos_filter_match(NULL, t, r, f, 1));
mm/damon/tests/core-kunit.h
1109
KUNIT_EXPECT_EQ(test, damon_nr_regions(t), 2);
mm/damon/tests/core-kunit.h
1113
damon_destroy_region(r2, t);
mm/damon/tests/core-kunit.h
1119
damos_filter_match(NULL, t, r, f, 1));
mm/damon/tests/core-kunit.h
1123
KUNIT_EXPECT_EQ(test, damon_nr_regions(t), 2);
mm/damon/tests/core-kunit.h
1127
damon_destroy_region(r2, t);
mm/damon/tests/core-kunit.h
1129
damon_free_target(t);
mm/damon/tests/core-kunit.h
113
damon_for_each_target(t, ctx) {
mm/damon/tests/core-kunit.h
122
damon_add_region(r, t);
mm/damon/tests/core-kunit.h
128
damon_for_each_target(t, ctx) {
mm/damon/tests/core-kunit.h
131
damon_for_each_region(r, t) {
mm/damon/tests/core-kunit.h
147
struct damon_target *t;
mm/damon/tests/core-kunit.h
150
t = damon_new_target();
mm/damon/tests/core-kunit.h
151
if (!t)
mm/damon/tests/core-kunit.h
155
damon_free_target(t);
mm/damon/tests/core-kunit.h
162
damon_add_region(r, t);
mm/damon/tests/core-kunit.h
163
damon_split_region_at(t, r, 25);
mm/damon/tests/core-kunit.h
176
damon_free_target(t);
mm/damon/tests/core-kunit.h
181
struct damon_target *t;
mm/damon/tests/core-kunit.h
185
t = damon_new_target();
mm/damon/tests/core-kunit.h
186
if (!t)
mm/damon/tests/core-kunit.h
190
damon_free_target(t);
mm/damon/tests/core-kunit.h
196
damon_add_region(r, t);
mm/damon/tests/core-kunit.h
199
damon_free_target(t);
mm/damon/tests/core-kunit.h
20
struct damon_target *t;
mm/damon/tests/core-kunit.h
205
damon_add_region(r2, t);
mm/damon/tests/core-kunit.h
207
damon_merge_two_regions(t, r, r2);
mm/damon/tests/core-kunit.h
215
damon_for_each_region(r3, t) {
mm/damon/tests/core-kunit.h
221
damon_free_target(t);
mm/damon/tests/core-kunit.h
224
static struct damon_region *__nth_region_of(struct damon_target *t, int idx)
mm/damon/tests/core-kunit.h
229
damon_for_each_region(r, t) {
mm/damon/tests/core-kunit.h
239
struct damon_target *t;
mm/damon/tests/core-kunit.h
249
t = damon_new_target();
mm/damon/tests/core-kunit.h
250
if (!t)
mm/damon/tests/core-kunit.h
255
damon_free_target(t);
mm/damon/tests/core-kunit.h
260
damon_add_region(r, t);
mm/damon/tests/core-kunit.h
263
damon_merge_regions_of(t, 9, 9999);
mm/damon/tests/core-kunit.h
265
KUNIT_EXPECT_EQ(test, damon_nr_regions(t), 6u);
mm/damon/tests/core-kunit.h
267
r = __nth_region_of(t, i);
mm/damon/tests/core-kunit.h
271
damon_free_target(t);
mm/damon/tests/core-kunit.h
276
struct damon_target *t;
mm/damon/tests/core-kunit.h
282
t = damon_new_target();
mm/damon/tests/core-kunit.h
283
if (!t)
mm/damon/tests/core-kunit.h
287
damon_free_target(t);
mm/damon/tests/core-kunit.h
29
t = damon_new_target();
mm/damon/tests/core-kunit.h
290
damon_add_region(r, t);
mm/damon/tests/core-kunit.h
291
damon_split_regions_of(t, 2, 1);
mm/damon/tests/core-kunit.h
292
KUNIT_EXPECT_LE(test, damon_nr_regions(t), 2u);
mm/damon/tests/core-kunit.h
293
damon_free_target(t);
mm/damon/tests/core-kunit.h
295
t = damon_new_target();
mm/damon/tests/core-kunit.h
296
if (!t)
mm/damon/tests/core-kunit.h
30
if (!t) {
mm/damon/tests/core-kunit.h
300
damon_free_target(t);
mm/damon/tests/core-kunit.h
303
damon_add_region(r, t);
mm/damon/tests/core-kunit.h
304
damon_split_regions_of(t, 4, 1);
mm/damon/tests/core-kunit.h
305
KUNIT_EXPECT_LE(test, damon_nr_regions(t), 4u);
mm/damon/tests/core-kunit.h
306
damon_free_target(t);
mm/damon/tests/core-kunit.h
308
t = damon_new_target();
mm/damon/tests/core-kunit.h
309
if (!t)
mm/damon/tests/core-kunit.h
314
damon_free_target(t);
mm/damon/tests/core-kunit.h
317
damon_add_region(r, t);
mm/damon/tests/core-kunit.h
319
damon_split_regions_of(t, 4, 5);
mm/damon/tests/core-kunit.h
320
KUNIT_EXPECT_LE(test, damon_nr_regions(t), 12u);
mm/damon/tests/core-kunit.h
321
damon_for_each_region(r, t)
mm/damon/tests/core-kunit.h
323
damon_free_target(t);
mm/damon/tests/core-kunit.h
34
KUNIT_EXPECT_EQ(test, 0u, damon_nr_regions(t));
mm/damon/tests/core-kunit.h
36
damon_add_region(r, t);
mm/damon/tests/core-kunit.h
37
KUNIT_EXPECT_EQ(test, 1u, damon_nr_regions(t));
mm/damon/tests/core-kunit.h
379
struct damon_target *t = damon_new_target();
mm/damon/tests/core-kunit.h
386
if (!t)
mm/damon/tests/core-kunit.h
39
damon_destroy_region(r, t);
mm/damon/tests/core-kunit.h
390
damon_free_target(t);
mm/damon/tests/core-kunit.h
395
damon_free_target(t);
mm/damon/tests/core-kunit.h
40
KUNIT_EXPECT_EQ(test, 0u, damon_nr_regions(t));
mm/damon/tests/core-kunit.h
400
damon_add_region(r1, t);
mm/damon/tests/core-kunit.h
401
damon_add_region(r2, t);
mm/damon/tests/core-kunit.h
402
damon_set_regions(t, &range, 1, 1);
mm/damon/tests/core-kunit.h
404
KUNIT_EXPECT_EQ(test, damon_nr_regions(t), 3);
mm/damon/tests/core-kunit.h
405
damon_for_each_region(r, t) {
mm/damon/tests/core-kunit.h
409
damon_destroy_target(t, NULL);
mm/damon/tests/core-kunit.h
42
damon_free_target(t);
mm/damon/tests/core-kunit.h
47
struct damon_target *t;
mm/damon/tests/core-kunit.h
50
damon_for_each_target(t, ctx)
mm/damon/tests/core-kunit.h
59
struct damon_target *t;
mm/damon/tests/core-kunit.h
64
t = damon_new_target();
mm/damon/tests/core-kunit.h
65
if (!t) {
mm/damon/tests/core-kunit.h
71
damon_add_target(c, t);
mm/damon/tests/core-kunit.h
74
damon_destroy_target(t, c);
mm/damon/tests/core-kunit.h
96
struct damon_target *t;
mm/damon/tests/core-kunit.h
999
struct damon_target *t;
mm/damon/tests/sysfs-kunit.h
17
struct damon_target *t;
mm/damon/tests/sysfs-kunit.h
20
damon_for_each_target(t, ctx)
mm/damon/tests/vaddr-kunit.h
134
struct damon_target *t;
mm/damon/tests/vaddr-kunit.h
138
t = damon_new_target();
mm/damon/tests/vaddr-kunit.h
139
if (!t)
mm/damon/tests/vaddr-kunit.h
144
damon_destroy_target(t, NULL);
mm/damon/tests/vaddr-kunit.h
147
damon_add_region(r, t);
mm/damon/tests/vaddr-kunit.h
150
damon_set_regions(t, three_regions, 3, DAMON_MIN_REGION_SZ);
mm/damon/tests/vaddr-kunit.h
153
r = __nth_region_of(t, i);
mm/damon/tests/vaddr-kunit.h
158
damon_destroy_target(t, NULL);
mm/damon/tests/vaddr-kunit.h
258
struct damon_target *t = damon_new_target();
mm/damon/tests/vaddr-kunit.h
261
if (!t)
mm/damon/tests/vaddr-kunit.h
266
damon_free_target(t);
mm/damon/tests/vaddr-kunit.h
270
damon_add_region(r, t);
mm/damon/tests/vaddr-kunit.h
272
damon_va_evenly_split_region(t, r, nr_pieces), -EINVAL);
mm/damon/tests/vaddr-kunit.h
273
KUNIT_EXPECT_EQ(test, damon_nr_regions(t), 1u);
mm/damon/tests/vaddr-kunit.h
275
damon_for_each_region(r, t) {
mm/damon/tests/vaddr-kunit.h
280
damon_free_target(t);
mm/damon/tests/vaddr-kunit.h
286
struct damon_target *t = damon_new_target();
mm/damon/tests/vaddr-kunit.h
291
if (!t)
mm/damon/tests/vaddr-kunit.h
295
damon_free_target(t);
mm/damon/tests/vaddr-kunit.h
298
damon_add_region(r, t);
mm/damon/tests/vaddr-kunit.h
300
damon_va_evenly_split_region(t, r, nr_pieces), 0);
mm/damon/tests/vaddr-kunit.h
301
KUNIT_EXPECT_EQ(test, damon_nr_regions(t), nr_pieces);
mm/damon/tests/vaddr-kunit.h
303
damon_for_each_region(r, t) {
mm/damon/tests/vaddr-kunit.h
314
damon_free_target(t);
mm/damon/tests/vaddr-kunit.h
94
static struct damon_region *__nth_region_of(struct damon_target *t, int idx)
mm/damon/tests/vaddr-kunit.h
99
damon_for_each_region(r, t) {
mm/damon/vaddr.c
179
static int damon_va_three_regions(struct damon_target *t,
mm/damon/vaddr.c
185
mm = damon_get_mm(t);
mm/damon/vaddr.c
240
struct damon_target *t)
mm/damon/vaddr.c
248
if (damon_va_three_regions(t, regions)) {
mm/damon/vaddr.c
250
if (ti == t)
mm/damon/vaddr.c
272
damon_add_region(r, t);
mm/damon/vaddr.c
275
damon_va_evenly_split_region(t, r, nr_pieces);
mm/damon/vaddr.c
282
struct damon_target *t;
mm/damon/vaddr.c
284
damon_for_each_target(t, ctx) {
mm/damon/vaddr.c
286
if (!damon_nr_regions(t))
mm/damon/vaddr.c
287
__damon_va_init_regions(ctx, t);
mm/damon/vaddr.c
297
struct damon_target *t;
mm/damon/vaddr.c
299
damon_for_each_target(t, ctx) {
mm/damon/vaddr.c
30
static inline struct task_struct *damon_get_task_struct(struct damon_target *t)
mm/damon/vaddr.c
300
if (damon_va_three_regions(t, three_regions))
mm/damon/vaddr.c
302
damon_set_regions(t, three_regions, 3, DAMON_MIN_REGION_SZ);
mm/damon/vaddr.c
32
return get_pid_task(t->pid, PIDTYPE_PID);
mm/damon/vaddr.c
411
struct damon_target *t;
mm/damon/vaddr.c
415
damon_for_each_target(t, ctx) {
mm/damon/vaddr.c
416
mm = damon_get_mm(t);
mm/damon/vaddr.c
419
damon_for_each_region(r, t)
mm/damon/vaddr.c
42
static struct mm_struct *damon_get_mm(struct damon_target *t)
mm/damon/vaddr.c
47
task = damon_get_task_struct(t);
mm/damon/vaddr.c
567
struct damon_target *t;
mm/damon/vaddr.c
573
damon_for_each_target(t, ctx) {
mm/damon/vaddr.c
574
mm = damon_get_mm(t);
mm/damon/vaddr.c
576
damon_for_each_region(r, t) {
mm/damon/vaddr.c
65
static int damon_va_evenly_split_region(struct damon_target *t,
mm/damon/vaddr.c
759
static bool damon_va_target_valid(struct damon_target *t)
mm/damon/vaddr.c
763
task = damon_get_task_struct(t);
mm/damon/vaddr.c
772
static void damon_va_cleanup_target(struct damon_target *t)
mm/damon/vaddr.c
774
put_pid(t->pid);
mm/damon/vaddr.c
92
damon_insert_region(n, r, next, t);
mm/damon/vaddr.c
951
struct damon_target *t, struct damon_region *r,
mm/damon/vaddr.c
974
return damos_va_migrate(t, r, scheme, sz_filter_passed);
mm/damon/vaddr.c
976
return damos_va_stat(t, r, scheme, sz_filter_passed);
mm/damon/vaddr.c
984
return damos_madvise(t, r, madv_action);
mm/damon/vaddr.c
988
struct damon_target *t, struct damon_region *r,
mm/hugetlb.c
1000
seg_to = min(rg->to, t);
mm/hugetlb.c
596
static long add_reservation_in_range(struct resv_map *resv, long f, long t,
mm/hugetlb.c
627
if (iter->from >= t) {
mm/hugetlb.c
649
if (last_accounted_offset < t)
mm/hugetlb.c
651
t, h, h_cg, regions_needed);
mm/hugetlb.c
729
static long region_add(struct resv_map *resv, long f, long t,
mm/hugetlb.c
739
add_reservation_in_range(resv, f, t, NULL, NULL,
mm/hugetlb.c
758
VM_BUG_ON(t - f <= 1);
mm/hugetlb.c
768
add = add_reservation_in_range(resv, f, t, h_cg, h, NULL);
mm/hugetlb.c
796
static long region_chg(struct resv_map *resv, long f, long t,
mm/hugetlb.c
804
chg = add_reservation_in_range(resv, f, t, NULL, NULL,
mm/hugetlb.c
832
static void region_abort(struct resv_map *resv, long f, long t,
mm/hugetlb.c
855
static long region_del(struct resv_map *resv, long f, long t)
mm/hugetlb.c
875
if (rg->from >= t)
mm/hugetlb.c
878
if (f > rg->from && t < rg->to) { /* Must split region */
mm/hugetlb.c
900
del += t - f;
mm/hugetlb.c
902
resv, rg, t - f, false);
mm/hugetlb.c
905
nrg->from = t;
mm/hugetlb.c
920
if (f <= rg->from && t >= rg->to) { /* Remove entire region */
mm/hugetlb.c
931
t - rg->from, false);
mm/hugetlb.c
933
del += t - rg->from;
mm/hugetlb.c
934
rg->from = t;
mm/hugetlb.c
982
static long region_count(struct resv_map *resv, long f, long t)
mm/hugetlb.c
996
if (rg->from >= t)
mm/hugetlb_cgroup.c
24
#define MEMFILE_OFFSET(t, m0) (((offsetof(t, m0) << 16) | sizeof_field(t, m0)))
mm/ksm.c
3245
struct task_struct *t =
mm/ksm.c
3247
if (!t)
mm/ksm.c
3253
if (vma->vm_mm == t->mm) {
mm/ksm.c
3255
add_to_kill_ksm(t, page, vma, to_kill,
mm/memcontrol-v1.c
454
struct mem_cgroup_threshold_ary *t;
mm/memcontrol-v1.c
460
t = rcu_dereference(memcg->thresholds.primary);
mm/memcontrol-v1.c
462
t = rcu_dereference(memcg->memsw_thresholds.primary);
mm/memcontrol-v1.c
464
if (!t)
mm/memcontrol-v1.c
474
i = t->current_threshold;
mm/memcontrol-v1.c
482
for (; i >= 0 && unlikely(t->entries[i].threshold > usage); i--)
mm/memcontrol-v1.c
483
eventfd_signal(t->entries[i].eventfd);
mm/memcontrol-v1.c
494
for (; i < t->size && unlikely(t->entries[i].threshold <= usage); i++)
mm/memcontrol-v1.c
495
eventfd_signal(t->entries[i].eventfd);
mm/memcontrol-v1.c
498
t->current_threshold = i - 1;
mm/memory-failure.c
2249
struct task_struct *t = tsk;
mm/memory-failure.c
2251
t = task_early_kill(tsk, true);
mm/memory-failure.c
2252
if (!t)
mm/memory-failure.c
2257
if (vma->vm_mm == t->mm &&
mm/memory-failure.c
2259
add_to_kill_pgoff(t, vma, to_kill, pgoff);
mm/memory-failure.c
288
struct task_struct *t = tk->tsk;
mm/memory-failure.c
293
pfn, t->comm, task_pid_nr(t));
mm/memory-failure.c
295
if ((flags & MF_ACTION_REQUIRED) && (t == current))
mm/memory-failure.c
306
addr_lsb, t);
mm/memory-failure.c
309
t->comm, task_pid_nr(t), ret);
mm/memory-failure.c
502
struct task_struct *t;
mm/memory-failure.c
504
for_each_thread(tsk, t) {
mm/memory-failure.c
505
if (t->flags & PF_MCE_PROCESS) {
mm/memory-failure.c
506
if (t->flags & PF_MCE_EARLY)
mm/memory-failure.c
507
return t;
mm/memory-failure.c
510
return t;
mm/memory-failure.c
562
struct task_struct *t = task_early_kill(tsk, force_early);
mm/memory-failure.c
565
if (!t)
mm/memory-failure.c
570
if (vma->vm_mm != t->mm)
mm/memory-failure.c
573
add_to_kill_anon_file(t, page, vma, to_kill, addr);
mm/memory-failure.c
596
struct task_struct *t = task_early_kill(tsk, force_early);
mm/memory-failure.c
599
if (!t)
mm/memory-failure.c
610
if (vma->vm_mm != t->mm)
mm/memory-failure.c
613
add_to_kill_anon_file(t, page, vma, to_kill, addr);
mm/memory-failure.c
642
struct task_struct *t = tsk;
mm/memory-failure.c
650
t = task_early_kill(tsk, true);
mm/memory-failure.c
651
if (!t)
mm/memory-failure.c
654
if (vma->vm_mm == t->mm)
mm/memory-failure.c
655
add_to_kill_fsdax(t, page, vma, to_kill, pgoff);
mm/mempolicy.c
1675
unsigned long t;
mm/mempolicy.c
1677
if (get_bitmap(&t, &nmask[(maxnode - 1) / BITS_PER_LONG], bits))
mm/mempolicy.c
1684
t &= ~((1UL << (MAX_NUMNODES % BITS_PER_LONG)) - 1);
mm/mempolicy.c
1686
if (t)
mm/mm_init.c
1435
unsigned int order, t;
mm/mm_init.c
1436
for_each_migratetype_order(order, t) {
mm/mm_init.c
1437
INIT_LIST_HEAD(&zone->free_area[order].free_list[t]);
mm/oom_kill.c
136
struct task_struct *t;
mm/oom_kill.c
140
for_each_thread(p, t) {
mm/oom_kill.c
141
task_lock(t);
mm/oom_kill.c
142
if (likely(t->mm))
mm/oom_kill.c
144
task_unlock(t);
mm/oom_kill.c
146
t = NULL;
mm/oom_kill.c
150
return t;
mm/oom_kill.c
496
const struct task_struct *t;
mm/oom_kill.c
498
for_each_thread(p, t) {
mm/oom_kill.c
499
const struct mm_struct *t_mm = READ_ONCE(t->mm);
mm/page-writeback.c
1570
unsigned long t;
mm/page-writeback.c
1579
t = wb_dirty / (1 + bw / roundup_pow_of_two(1 + HZ / 8));
mm/page-writeback.c
1580
t++;
mm/page-writeback.c
1582
return min_t(unsigned long, t, MAX_PAUSE);
mm/page-writeback.c
1593
long t; /* target pause */
mm/page-writeback.c
1598
t = max(1, HZ / 100);
mm/page-writeback.c
1607
t += (hi - lo) * (10 * HZ) / 1024;
mm/page-writeback.c
1627
t = min(t, 1 + max_pause / 2);
mm/page-writeback.c
1628
pages = dirty_ratelimit * t / roundup_pow_of_two(HZ);
mm/page-writeback.c
1639
t = max_pause;
mm/page-writeback.c
1640
pages = dirty_ratelimit * t / roundup_pow_of_two(HZ);
mm/page-writeback.c
1643
t = HZ * DIRTY_POLL_THRESH / dirty_ratelimit;
mm/page-writeback.c
1649
t = max_pause;
mm/page-writeback.c
1650
pages = task_ratelimit * t / roundup_pow_of_two(HZ);
mm/page-writeback.c
1657
return pages >= DIRTY_POLL_THRESH ? 1 + t / 2 : t;
mm/page-writeback.c
601
static void writeout_period(struct timer_list *t)
mm/page-writeback.c
603
struct wb_domain *dom = timer_container_of(dom, t, period_timer);
mm/slab_common.c
1887
schedule_page_work_fn(struct hrtimer *t)
mm/slab_common.c
1890
container_of(t, struct kfree_rcu_cpu, hrtimer);
mm/slub.c
1079
static void print_track(const char *s, struct track *t, unsigned long pr_time)
mm/slub.c
1083
if (!t->addr)
mm/slub.c
1087
s, (void *)t->addr, pr_time - t->when, t->cpu, t->pid);
mm/slub.c
1089
handle = READ_ONCE(t->handle);
mm/slub.c
6108
struct llist_node *llnode, *pos, *t;
mm/slub.c
6114
llist_for_each_safe(pos, t, llnode) {
mm/slub.c
8102
struct slab *t;
mm/slub.c
8124
list_for_each_entry_safe(slab, t, &n->partial, slab_list) {
mm/slub.c
8152
list_for_each_entry_safe(slab, t, &discard, slab_list)
mm/slub.c
8653
static void free_loc_track(struct loc_track *t)
mm/slub.c
8655
if (t->max)
mm/slub.c
8656
free_pages((unsigned long)t->loc,
mm/slub.c
8657
get_order(sizeof(struct location) * t->max));
mm/slub.c
8660
static int alloc_loc_track(struct loc_track *t, unsigned long max, gfp_t flags)
mm/slub.c
8671
if (t->count) {
mm/slub.c
8672
memcpy(l, t->loc, sizeof(struct location) * t->count);
mm/slub.c
8673
free_loc_track(t);
mm/slub.c
8675
t->max = max;
mm/slub.c
8676
t->loc = l;
mm/slub.c
8680
static int add_location(struct loc_track *t, struct kmem_cache *s,
mm/slub.c
8695
end = t->count;
mm/slub.c
8707
l = &t->loc[pos];
mm/slub.c
8748
if (t->count >= t->max && !alloc_loc_track(t, 2 * t->max, GFP_ATOMIC))
mm/slub.c
8751
l = t->loc + pos;
mm/slub.c
8752
if (pos < t->count)
mm/slub.c
8754
(t->count - pos) * sizeof(struct location));
mm/slub.c
8755
t->count++;
mm/slub.c
8772
static void process_slab(struct loc_track *t, struct kmem_cache *s,
mm/slub.c
8784
add_location(t, s, get_track(s, p, alloc),
mm/slub.c
9609
struct loc_track *t = seq->private;
mm/slub.c
9613
idx = (unsigned long) t->idx;
mm/slub.c
9614
if (idx < t->count) {
mm/slub.c
9615
l = &t->loc[idx];
mm/slub.c
9667
if (!idx && !t->count)
mm/slub.c
9679
struct loc_track *t = seq->private;
mm/slub.c
9681
t->idx = ++(*ppos);
mm/slub.c
9682
if (*ppos <= t->count)
mm/slub.c
9698
struct loc_track *t = seq->private;
mm/slub.c
9700
t->idx = *ppos;
mm/slub.c
9717
struct loc_track *t = __seq_open_private(filep, &slab_debugfs_sops,
mm/slub.c
9722
if (!t)
mm/slub.c
9733
if (!alloc_loc_track(t, PAGE_SIZE / sizeof(struct location), GFP_KERNEL)) {
mm/slub.c
9748
process_slab(t, s, slab, alloc, obj_map);
mm/slub.c
9750
process_slab(t, s, slab, alloc, obj_map);
mm/slub.c
9755
sort(t->loc, t->count, sizeof(struct location),
mm/slub.c
9765
struct loc_track *t = seq->private;
mm/slub.c
9767
free_loc_track(t);
mm/util.c
320
struct task_struct * __maybe_unused t = current;
mm/util.c
322
return (vma->vm_start <= KSTK_ESP(t) && vma->vm_end >= KSTK_ESP(t));
mm/util.c
780
struct ctl_table t;
mm/util.c
796
t = *table;
mm/util.c
797
t.data = &new_policy;
mm/util.c
798
ret = proc_dointvec_minmax(&t, write, buffer, lenp, ppos);
mm/vmalloc.c
3395
struct llist_node *t, *llnode;
mm/vmalloc.c
3397
llist_for_each_safe(llnode, t, llist_del_all(&p->list))
mm/vmstat.c
351
long t;
mm/vmstat.c
364
t = __this_cpu_read(pcp->stat_threshold);
mm/vmstat.c
366
if (unlikely(abs(x) > t)) {
mm/vmstat.c
382
long t;
mm/vmstat.c
400
t = __this_cpu_read(pcp->stat_threshold);
mm/vmstat.c
402
if (unlikely(abs(x) > t)) {
mm/vmstat.c
439
s8 v, t;
mm/vmstat.c
445
t = __this_cpu_read(pcp->stat_threshold);
mm/vmstat.c
446
if (unlikely(v > t)) {
mm/vmstat.c
447
s8 overstep = t >> 1;
mm/vmstat.c
460
s8 v, t;
mm/vmstat.c
468
t = __this_cpu_read(pcp->stat_threshold);
mm/vmstat.c
469
if (unlikely(v > t)) {
mm/vmstat.c
470
s8 overstep = t >> 1;
mm/vmstat.c
495
s8 v, t;
mm/vmstat.c
501
t = __this_cpu_read(pcp->stat_threshold);
mm/vmstat.c
502
if (unlikely(v < - t)) {
mm/vmstat.c
503
s8 overstep = t >> 1;
mm/vmstat.c
516
s8 v, t;
mm/vmstat.c
524
t = __this_cpu_read(pcp->stat_threshold);
mm/vmstat.c
525
if (unlikely(v < - t)) {
mm/vmstat.c
526
s8 overstep = t >> 1;
mm/vmstat.c
565
long n, t, z;
mm/vmstat.c
582
t = this_cpu_read(pcp->stat_threshold);
mm/vmstat.c
586
if (abs(n) > t) {
mm/vmstat.c
587
int os = overstep_mode * (t >> 1) ;
mm/vmstat.c
623
long n, t, z;
mm/vmstat.c
651
t = this_cpu_read(pcp->stat_threshold);
mm/vmstat.c
655
if (abs(n) > t) {
mm/vmstat.c
656
int os = overstep_mode * (t >> 1) ;
net/6lowpan/debugfs.c
100
struct lowpan_iphc_ctx_table *t =
net/6lowpan/debugfs.c
103
spin_lock_bh(&t->lock);
net/6lowpan/debugfs.c
113
spin_unlock_bh(&t->lock);
net/6lowpan/debugfs.c
130
struct lowpan_iphc_ctx_table *t =
net/6lowpan/debugfs.c
149
spin_lock_bh(&t->lock);
net/6lowpan/debugfs.c
152
spin_unlock_bh(&t->lock);
net/6lowpan/debugfs.c
195
struct lowpan_iphc_ctx_table *t = file->private;
net/6lowpan/debugfs.c
201
spin_lock_bh(&t->lock);
net/6lowpan/debugfs.c
203
if (!lowpan_iphc_ctx_is_active(&t->table[i]))
net/6lowpan/debugfs.c
206
seq_printf(file, "%3d|%39pI6c/%-3d|%d\n", t->table[i].id,
net/6lowpan/debugfs.c
207
&t->table[i].pfx, t->table[i].plen,
net/6lowpan/debugfs.c
208
lowpan_iphc_ctx_is_compression(&t->table[i]));
net/6lowpan/debugfs.c
210
spin_unlock_bh(&t->lock);
net/6lowpan/debugfs.c
69
struct lowpan_iphc_ctx_table *t =
net/6lowpan/debugfs.c
75
spin_lock_bh(&t->lock);
net/6lowpan/debugfs.c
77
spin_unlock_bh(&t->lock);
net/6lowpan/debugfs.c
85
struct lowpan_iphc_ctx_table *t =
net/6lowpan/debugfs.c
88
spin_lock_bh(&t->lock);
net/6lowpan/debugfs.c
90
spin_unlock_bh(&t->lock);
net/802/garp.c
415
static void garp_join_timer(struct timer_list *t)
net/802/garp.c
417
struct garp_applicant *app = timer_container_of(app, t, join_timer);
net/802/mrp.c
600
static void mrp_join_timer(struct timer_list *t)
net/802/mrp.c
602
struct mrp_applicant *app = timer_container_of(app, t, join_timer);
net/802/mrp.c
622
static void mrp_periodic_timer(struct timer_list *t)
net/802/mrp.c
624
struct mrp_applicant *app = timer_container_of(app, t, periodic_timer);
net/9p/mod.c
136
struct p9_trans_module *t, *found = NULL;
net/9p/mod.c
141
list_for_each_entry(t, &v9fs_trans_list, list)
net/9p/mod.c
142
if (t->def && try_module_get(t->owner)) {
net/9p/mod.c
143
found = t;
net/9p/mod.c
148
list_for_each_entry(t, &v9fs_trans_list, list)
net/9p/mod.c
149
if (try_module_get(t->owner)) {
net/9p/mod.c
150
found = t;
net/9p/mod.c
87
struct p9_trans_module *t, *found = NULL;
net/9p/mod.c
91
list_for_each_entry(t, &v9fs_trans_list, list)
net/9p/mod.c
92
if (strcmp(t->name, s) == 0 &&
net/9p/mod.c
93
try_module_get(t->owner)) {
net/9p/mod.c
94
found = t;
net/appletalk/aarp.c
268
struct aarp_entry *t;
net/appletalk/aarp.c
273
t = *n;
net/appletalk/aarp.c
275
__aarp_expire(t);
net/appletalk/aarp.c
287
struct aarp_entry *t;
net/appletalk/aarp.c
292
t = *n;
net/appletalk/aarp.c
294
__aarp_expire(t);
net/appletalk/aarp.c
309
struct aarp_entry *t;
net/appletalk/aarp.c
313
t = *n;
net/appletalk/aarp.c
315
__aarp_expire(t);
net/appletalk/aarp.c
367
struct aarp_entry *t;
net/appletalk/aarp.c
370
t = *n;
net/appletalk/aarp.c
372
__aarp_expire(t);
net/appletalk/ddp.c
170
static void atalk_destroy_timer(struct timer_list *t)
net/appletalk/ddp.c
172
struct sock *sk = timer_container_of(sk, t, sk_timer);
net/atm/lec.c
1253
static void lec_arp_expire_arp(struct timer_list *t);
net/atm/lec.c
1584
static void lec_arp_expire_arp(struct timer_list *t)
net/atm/lec.c
1588
entry = timer_container_of(entry, t, timer);
net/atm/lec.c
1606
static void lec_arp_expire_vcc(struct timer_list *t)
net/atm/lec.c
1609
struct lec_arp_table *to_remove = timer_container_of(to_remove, t,
net/atm/pppoatm.c
104
static void pppoatm_wakeup_sender(struct tasklet_struct *t)
net/atm/pppoatm.c
106
struct pppoatm_vcc *pvcc = from_tasklet(pvcc, t, wakeup_tasklet);
net/ax25/af_ax25.c
288
static void ax25_destroy_timer(struct timer_list *t)
net/ax25/af_ax25.c
290
ax25_cb *ax25 = timer_container_of(ax25, t, dtimer);
net/ax25/ax25_ds_timer.c
65
static void ax25_ds_timeout(struct timer_list *t)
net/ax25/ax25_ds_timer.c
67
ax25_dev *ax25_dev = timer_container_of(ax25_dev, t, dama.slave_timer);
net/ax25/ax25_route.c
153
ax25_route *s, *t, *ax25_rt;
net/ax25/ax25_route.c
171
for (t = ax25_route_list; t != NULL; t = t->next) {
net/ax25/ax25_route.c
172
if (t->next == s) {
net/ax25/ax25_route.c
173
t->next = s->next;
net/ax25/ax25_route.c
44
ax25_route *s, *t, *ax25_rt;
net/ax25/ax25_route.c
58
for (t = ax25_route_list; t != NULL; t = t->next) {
net/ax25/ax25_route.c
59
if (t->next == s) {
net/ax25/ax25_route.c
60
t->next = s->next;
net/ax25/ax25_subr.c
222
int n, t = 2;
net/ax25/ax25_subr.c
229
t += 2 * ax25->n2count;
net/ax25/ax25_subr.c
234
t *= 2;
net/ax25/ax25_subr.c
235
if (t > 8) t = 8;
net/ax25/ax25_subr.c
239
ax25->t1 = t * ax25->rtt;
net/ax25/ax25_timer.c
121
static void ax25_heartbeat_expiry(struct timer_list *t)
net/ax25/ax25_timer.c
124
ax25_cb *ax25 = timer_container_of(ax25, t, timer);
net/ax25/ax25_timer.c
146
static void ax25_t1timer_expiry(struct timer_list *t)
net/ax25/ax25_timer.c
148
ax25_cb *ax25 = timer_container_of(ax25, t, t1timer);
net/ax25/ax25_timer.c
165
static void ax25_t2timer_expiry(struct timer_list *t)
net/ax25/ax25_timer.c
167
ax25_cb *ax25 = timer_container_of(ax25, t, t2timer);
net/ax25/ax25_timer.c
184
static void ax25_t3timer_expiry(struct timer_list *t)
net/ax25/ax25_timer.c
186
ax25_cb *ax25 = timer_container_of(ax25, t, t3timer);
net/ax25/ax25_timer.c
205
static void ax25_idletimer_expiry(struct timer_list *t)
net/ax25/ax25_timer.c
207
ax25_cb *ax25 = timer_container_of(ax25, t, idletimer);
net/batman-adv/tp_meter.c
1102
static void batadv_tp_receiver_shutdown(struct timer_list *t)
net/batman-adv/tp_meter.c
1104
struct batadv_tp_vars *tp_vars = timer_container_of(tp_vars, t, timer);
net/batman-adv/tp_meter.c
486
static void batadv_tp_sender_timeout(struct timer_list *t)
net/batman-adv/tp_meter.c
488
struct batadv_tp_vars *tp_vars = timer_container_of(tp_vars, t, timer);
net/bluetooth/hci_conn.c
1135
struct hci_link *link, *t;
net/bluetooth/hci_conn.c
1137
list_for_each_entry_safe(link, t, &conn->link_list, list) {
net/bluetooth/hci_sync.c
2756
struct bdaddr_list *b, *t;
net/bluetooth/hci_sync.c
2815
list_for_each_entry_safe(b, t, &hdev->le_accept_list, list) {
net/bluetooth/hidp/core.c
406
static void hidp_idle_timeout(struct timer_list *t)
net/bluetooth/hidp/core.c
408
struct hidp_session *session = timer_container_of(session, t, timer);
net/bluetooth/iso.c
2407
struct hci_link *link, *t;
net/bluetooth/iso.c
2409
list_for_each_entry_safe(link, t, &hcon->link_list,
net/bluetooth/rfcomm/core.c
236
static void rfcomm_session_timeout(struct timer_list *t)
net/bluetooth/rfcomm/core.c
238
struct rfcomm_session *s = timer_container_of(s, t, timer);
net/bluetooth/rfcomm/core.c
261
static void rfcomm_dlc_timeout(struct timer_list *t)
net/bluetooth/rfcomm/core.c
263
struct rfcomm_dlc *d = timer_container_of(d, t, timer);
net/bluetooth/smp.c
246
u8 m[53], t[16];
net/bluetooth/smp.c
253
err = aes_cmac(tfm_cmac, salt, w, 32, t);
net/bluetooth/smp.c
257
SMP_DBG("t %16phN", t);
net/bluetooth/smp.c
268
err = aes_cmac(tfm_cmac, t, m, sizeof(m), mackey);
net/bluetooth/smp.c
276
err = aes_cmac(tfm_cmac, t, m, sizeof(m), ltk);
net/bpf/bpf_dummy_struct_ops.c
103
const struct btf_type *t;
net/bpf/bpf_dummy_struct_ops.c
112
t = btf_type_skip_modifiers(bpf_dummy_ops_btf, param->type, NULL);
net/bpf/bpf_dummy_struct_ops.c
113
if (!btf_type_is_ptr(t))
net/bpf/bpf_dummy_struct_ops.c
219
static int bpf_dummy_ops_check_member(const struct btf_type *t,
net/bpf/bpf_dummy_struct_ops.c
223
u32 moff = __btf_member_bit_offset(t, member) / 8;
net/bpf/bpf_dummy_struct_ops.c
241
const struct btf_type *t;
net/bpf/bpf_dummy_struct_ops.c
249
t = btf_type_by_id(reg->btf, reg->btf_id);
net/bpf/bpf_dummy_struct_ops.c
251
if (t != state) {
net/bpf/bpf_dummy_struct_ops.c
269
static int bpf_dummy_init_member(const struct btf_type *t,
net/bpf/test_run.c
1447
struct bpf_test_timer t = {};
net/bpf/test_run.c
1490
bpf_test_timer_enter(&t);
net/bpf/test_run.c
1494
} while (bpf_test_timer_continue(&t, 1, repeat, &ret, &duration));
net/bpf/test_run.c
1495
bpf_test_timer_leave(&t);
net/bpf/test_run.c
1515
struct bpf_test_timer t = {};
net/bpf/test_run.c
1582
bpf_test_timer_enter(&t);
net/bpf/test_run.c
1586
} while (bpf_test_timer_continue(&t, 1, repeat, &ret, &duration));
net/bpf/test_run.c
1587
bpf_test_timer_leave(&t);
net/bpf/test_run.c
36
static void bpf_test_timer_enter(struct bpf_test_timer *t)
net/bpf/test_run.c
366
struct bpf_test_timer t = {};
net/bpf/test_run.c
376
bpf_test_timer_enter(&t);
net/bpf/test_run.c
379
ret = xdp_test_run_batch(&xdp, prog, repeat - t.i);
net/bpf/test_run.c
382
} while (bpf_test_timer_continue(&t, xdp.frame_cnt, repeat, &ret, time));
net/bpf/test_run.c
383
bpf_test_timer_leave(&t);
net/bpf/test_run.c
396
struct bpf_test_timer t = {};
net/bpf/test_run.c
40
t->time_start = ktime_get_ns();
net/bpf/test_run.c
413
bpf_test_timer_enter(&t);
net/bpf/test_run.c
427
} while (bpf_test_timer_continue(&t, 1, repeat, &ret, time));
net/bpf/test_run.c
429
bpf_test_timer_leave(&t);
net/bpf/test_run.c
43
static void bpf_test_timer_leave(struct bpf_test_timer *t)
net/bpf/test_run.c
46
t->time_start = 0;
net/bpf/test_run.c
50
static bool bpf_test_timer_continue(struct bpf_test_timer *t, int iterations,
net/bpf/test_run.c
54
t->i += iterations;
net/bpf/test_run.c
55
if (t->i >= repeat) {
net/bpf/test_run.c
57
t->time_spent += ktime_get_ns() - t->time_start;
net/bpf/test_run.c
58
do_div(t->time_spent, t->i);
net/bpf/test_run.c
59
*duration = t->time_spent > U32_MAX ? U32_MAX : (u32)t->time_spent;
net/bpf/test_run.c
72
t->time_spent += ktime_get_ns() - t->time_start;
net/bpf/test_run.c
73
bpf_test_timer_leave(t);
net/bpf/test_run.c
75
bpf_test_timer_enter(t);
net/bpf/test_run.c
82
t->i = 0;
net/bridge/br_multicast.c
1313
static void br_multicast_group_src_expired(struct timer_list *t)
net/bridge/br_multicast.c
1315
struct net_bridge_group_src *src = timer_container_of(src, t, timer);
net/bridge/br_multicast.c
1648
struct timer_list *t,
net/bridge/br_multicast.c
1657
timer_pending(t))
net/bridge/br_multicast.c
1666
static void br_ip4_multicast_router_expired(struct timer_list *t)
net/bridge/br_multicast.c
1668
struct net_bridge_mcast_port *pmctx = timer_container_of(pmctx, t,
net/bridge/br_multicast.c
1671
br_multicast_router_expired(pmctx, t, &pmctx->ip4_rlist);
net/bridge/br_multicast.c
1675
static void br_ip6_multicast_router_expired(struct timer_list *t)
net/bridge/br_multicast.c
1677
struct net_bridge_mcast_port *pmctx = timer_container_of(pmctx, t,
net/bridge/br_multicast.c
1680
br_multicast_router_expired(pmctx, t, &pmctx->ip6_rlist);
net/bridge/br_multicast.c
1712
static void br_ip4_multicast_local_router_expired(struct timer_list *t)
net/bridge/br_multicast.c
1714
struct net_bridge_mcast *brmctx = timer_container_of(brmctx, t,
net/bridge/br_multicast.c
1717
br_multicast_local_router_expired(brmctx, t);
net/bridge/br_multicast.c
1721
static void br_ip6_multicast_local_router_expired(struct timer_list *t)
net/bridge/br_multicast.c
1723
struct net_bridge_mcast *brmctx = timer_container_of(brmctx, t,
net/bridge/br_multicast.c
1726
br_multicast_local_router_expired(brmctx, t);
net/bridge/br_multicast.c
1745
static void br_ip4_multicast_querier_expired(struct timer_list *t)
net/bridge/br_multicast.c
1747
struct net_bridge_mcast *brmctx = timer_container_of(brmctx, t,
net/bridge/br_multicast.c
1754
static void br_ip6_multicast_querier_expired(struct timer_list *t)
net/bridge/br_multicast.c
1756
struct net_bridge_mcast *brmctx = timer_container_of(brmctx, t,
net/bridge/br_multicast.c
1763
static void br_multicast_query_delay_expired(struct timer_list *t)
net/bridge/br_multicast.c
1917
static void br_ip4_multicast_port_query_expired(struct timer_list *t)
net/bridge/br_multicast.c
1919
struct net_bridge_mcast_port *pmctx = timer_container_of(pmctx, t,
net/bridge/br_multicast.c
1926
static void br_ip6_multicast_port_query_expired(struct timer_list *t)
net/bridge/br_multicast.c
1928
struct net_bridge_mcast_port *pmctx = timer_container_of(pmctx, t,
net/bridge/br_multicast.c
1935
static void br_multicast_port_group_rexmit(struct timer_list *t)
net/bridge/br_multicast.c
1937
struct net_bridge_port_group *pg = timer_container_of(pg, t,
net/bridge/br_multicast.c
4057
static void br_ip4_multicast_query_expired(struct timer_list *t)
net/bridge/br_multicast.c
4059
struct net_bridge_mcast *brmctx = timer_container_of(brmctx, t,
net/bridge/br_multicast.c
4066
static void br_ip6_multicast_query_expired(struct timer_list *t)
net/bridge/br_multicast.c
4068
struct net_bridge_mcast *brmctx = timer_container_of(brmctx, t,
net/bridge/br_multicast.c
62
static void br_multicast_port_group_rexmit(struct timer_list *t);
net/bridge/br_multicast.c
646
static void br_multicast_group_expired(struct timer_list *t)
net/bridge/br_multicast.c
648
struct net_bridge_mdb_entry *mp = timer_container_of(mp, t, timer);
net/bridge/br_multicast.c
855
static void br_multicast_port_group_expired(struct timer_list *t)
net/bridge/br_multicast.c
857
struct net_bridge_port_group *pg = timer_container_of(pg, t, timer);
net/bridge/br_multicast_eht.c
208
static void br_multicast_eht_set_entry_expired(struct timer_list *t)
net/bridge/br_multicast_eht.c
211
t,
net/bridge/br_multicast_eht.c
226
static void br_multicast_eht_set_expired(struct timer_list *t)
net/bridge/br_multicast_eht.c
229
t,
net/bridge/br_netlink_tunnel.c
300
int t, v;
net/bridge/br_netlink_tunnel.c
307
t = tinfo_last->tunid;
net/bridge/br_netlink_tunnel.c
311
err = br_vlan_tunnel_info(p, cmd, v, t, &curr_change);
net/bridge/br_netlink_tunnel.c
314
t++;
net/bridge/br_private.h
2032
void __br_set_forward_delay(struct net_bridge *br, unsigned long t);
net/bridge/br_private.h
2036
int __set_ageing_time(struct net_device *dev, unsigned long t);
net/bridge/br_stp.c
107
int t;
net/bridge/br_stp.c
122
t = memcmp(&p->designated_root, &rp->designated_root, 8);
net/bridge/br_stp.c
123
if (t < 0)
net/bridge/br_stp.c
125
else if (t > 0)
net/bridge/br_stp.c
135
t = memcmp(&p->designated_bridge, &rp->designated_bridge, 8);
net/bridge/br_stp.c
136
if (t < 0)
net/bridge/br_stp.c
138
else if (t > 0)
net/bridge/br_stp.c
291
int t;
net/bridge/br_stp.c
305
t = memcmp(&br->bridge_id, &p->designated_bridge, 8);
net/bridge/br_stp.c
306
if (t < 0)
net/bridge/br_stp.c
308
else if (t > 0)
net/bridge/br_stp.c
334
int t;
net/bridge/br_stp.c
336
t = memcmp(&bpdu->root, &p->designated_root, 8);
net/bridge/br_stp.c
337
if (t < 0)
net/bridge/br_stp.c
339
else if (t > 0)
net/bridge/br_stp.c
347
t = memcmp(&bpdu->bridge_id, &p->designated_bridge, 8);
net/bridge/br_stp.c
348
if (t < 0)
net/bridge/br_stp.c
350
else if (t > 0)
net/bridge/br_stp.c
574
unsigned long t = clock_t_to_jiffies(val);
net/bridge/br_stp.c
576
if (t < BR_MIN_HELLO_TIME || t > BR_MAX_HELLO_TIME)
net/bridge/br_stp.c
580
br->bridge_hello_time = t;
net/bridge/br_stp.c
589
unsigned long t = clock_t_to_jiffies(val);
net/bridge/br_stp.c
591
if (t < BR_MIN_MAX_AGE || t > BR_MAX_MAX_AGE)
net/bridge/br_stp.c
595
br->bridge_max_age = t;
net/bridge/br_stp.c
604
int __set_ageing_time(struct net_device *dev, unsigned long t)
net/bridge/br_stp.c
610
.u.ageing_time = jiffies_to_clock_t(t),
net/bridge/br_stp.c
631
unsigned long t = clock_t_to_jiffies(ageing_time);
net/bridge/br_stp.c
634
err = __set_ageing_time(br->dev, t);
net/bridge/br_stp.c
639
br->bridge_ageing_time = t;
net/bridge/br_stp.c
640
br->ageing_time = t;
net/bridge/br_stp.c
664
unsigned long t;
net/bridge/br_stp.c
673
t = 2 * br->forward_delay;
net/bridge/br_stp.c
674
br_debug(br, "decreasing ageing time to %lu\n", t);
net/bridge/br_stp.c
676
t = br->bridge_ageing_time;
net/bridge/br_stp.c
677
br_debug(br, "restoring ageing time to %lu\n", t);
net/bridge/br_stp.c
680
err = __set_ageing_time(br->dev, t);
net/bridge/br_stp.c
684
br->ageing_time = t;
net/bridge/br_stp.c
690
void __br_set_forward_delay(struct net_bridge *br, unsigned long t)
net/bridge/br_stp.c
692
br->bridge_forward_delay = t;
net/bridge/br_stp.c
699
unsigned long t = clock_t_to_jiffies(val);
net/bridge/br_stp.c
704
(t < BR_MIN_FORWARD_DELAY || t > BR_MAX_FORWARD_DELAY))
net/bridge/br_stp.c
707
__br_set_forward_delay(br, t);
net/bridge/br_stp_timer.c
105
static void br_tcn_timer_expired(struct timer_list *t)
net/bridge/br_stp_timer.c
107
struct net_bridge *br = timer_container_of(br, t, tcn_timer);
net/bridge/br_stp_timer.c
119
static void br_topology_change_timer_expired(struct timer_list *t)
net/bridge/br_stp_timer.c
121
struct net_bridge *br = timer_container_of(br, t,
net/bridge/br_stp_timer.c
131
static void br_hold_timer_expired(struct timer_list *t)
net/bridge/br_stp_timer.c
133
struct net_bridge_port *p = timer_container_of(p, t, hold_timer);
net/bridge/br_stp_timer.c
30
static void br_hello_timer_expired(struct timer_list *t)
net/bridge/br_stp_timer.c
32
struct net_bridge *br = timer_container_of(br, t, hello_timer);
net/bridge/br_stp_timer.c
46
static void br_message_age_timer_expired(struct timer_list *t)
net/bridge/br_stp_timer.c
48
struct net_bridge_port *p = timer_container_of(p, t,
net/bridge/br_stp_timer.c
80
static void br_forward_delay_timer_expired(struct timer_list *t)
net/bridge/br_stp_timer.c
82
struct net_bridge_port *p = timer_container_of(p, t,
net/bridge/netfilter/ebtables.c
1015
struct ebt_table *t;
net/bridge/netfilter/ebtables.c
1037
t = find_table_lock(net, repl->name, &ret, &ebt_mutex);
net/bridge/netfilter/ebtables.c
1038
if (!t) {
net/bridge/netfilter/ebtables.c
1043
if (repl->valid_hooks != t->valid_hooks) {
net/bridge/netfilter/ebtables.c
1048
if (repl->num_counters && repl->num_counters != t->private->nentries) {
net/bridge/netfilter/ebtables.c
1054
table = t->private;
net/bridge/netfilter/ebtables.c
1056
if (!table->nentries && newinfo->nentries && !try_module_get(t->me)) {
net/bridge/netfilter/ebtables.c
1060
module_put(t->me);
net/bridge/netfilter/ebtables.c
1062
write_lock_bh(&t->lock);
net/bridge/netfilter/ebtables.c
1064
get_counters(t->private->counters, counterstmp,
net/bridge/netfilter/ebtables.c
1065
t->private->nentries);
net/bridge/netfilter/ebtables.c
1067
t->private = newinfo;
net/bridge/netfilter/ebtables.c
1068
write_unlock_bh(&t->lock);
net/bridge/netfilter/ebtables.c
1186
struct ebt_table *t, *table;
net/bridge/netfilter/ebtables.c
1240
list_for_each_entry(t, &ebt_net->tables, list) {
net/bridge/netfilter/ebtables.c
1241
if (strcmp(t->name, table->name) == 0) {
net/bridge/netfilter/ebtables.c
1294
int ebt_register_template(const struct ebt_table *t, int (*table_init)(struct net *net))
net/bridge/netfilter/ebtables.c
1300
if (WARN_ON_ONCE(strcmp(t->name, tmpl->name) == 0)) {
net/bridge/netfilter/ebtables.c
1313
strscpy(tmpl->name, t->name, sizeof(tmpl->name));
net/bridge/netfilter/ebtables.c
1314
tmpl->owner = t->me;
net/bridge/netfilter/ebtables.c
1322
void ebt_unregister_template(const struct ebt_table *t)
net/bridge/netfilter/ebtables.c
1328
if (strcmp(t->name, tmpl->name))
net/bridge/netfilter/ebtables.c
1345
struct ebt_table *t;
net/bridge/netfilter/ebtables.c
1349
list_for_each_entry(t, &ebt_net->tables, list) {
net/bridge/netfilter/ebtables.c
1350
if (strcmp(t->name, name) == 0) {
net/bridge/netfilter/ebtables.c
1352
return t;
net/bridge/netfilter/ebtables.c
1384
struct ebt_table *t;
net/bridge/netfilter/ebtables.c
1393
t = find_table_lock(net, name, &ret, &ebt_mutex);
net/bridge/netfilter/ebtables.c
1394
if (!t)
net/bridge/netfilter/ebtables.c
1397
if (num_counters != t->private->nentries) {
net/bridge/netfilter/ebtables.c
1409
write_lock_bh(&t->lock);
net/bridge/netfilter/ebtables.c
1413
ADD_COUNTER(t->private->counters[i], tmp[i].bcnt, tmp[i].pcnt);
net/bridge/netfilter/ebtables.c
1415
write_unlock_bh(&t->lock);
net/bridge/netfilter/ebtables.c
1483
const struct ebt_entry_target *t;
net/bridge/netfilter/ebtables.c
1497
t = ebt_get_target_c(e);
net/bridge/netfilter/ebtables.c
1505
ret = ebt_obj_to_user(hlp, t->u.target->name, t->data, sizeof(*t),
net/bridge/netfilter/ebtables.c
1506
t->u.target->usersize, t->target_size,
net/bridge/netfilter/ebtables.c
1507
t->u.target->revision);
net/bridge/netfilter/ebtables.c
1514
static int copy_counters_to_user(struct ebt_table *t,
net/bridge/netfilter/ebtables.c
1533
write_lock_bh(&t->lock);
net/bridge/netfilter/ebtables.c
1535
write_unlock_bh(&t->lock);
net/bridge/netfilter/ebtables.c
1545
static int copy_everything_to_user(struct ebt_table *t, void __user *user,
net/bridge/netfilter/ebtables.c
1555
entries_size = t->private->entries_size;
net/bridge/netfilter/ebtables.c
1556
nentries = t->private->nentries;
net/bridge/netfilter/ebtables.c
1557
entries = t->private->entries;
net/bridge/netfilter/ebtables.c
1558
oldcounters = t->private->counters;
net/bridge/netfilter/ebtables.c
1560
entries_size = t->table->entries_size;
net/bridge/netfilter/ebtables.c
1561
nentries = t->table->nentries;
net/bridge/netfilter/ebtables.c
1562
entries = t->table->entries;
net/bridge/netfilter/ebtables.c
1563
oldcounters = t->table->counters;
net/bridge/netfilter/ebtables.c
1579
ret = copy_counters_to_user(t, oldcounters, tmp.counters,
net/bridge/netfilter/ebtables.c
1671
static int compat_target_to_user(struct ebt_entry_target *t,
net/bridge/netfilter/ebtables.c
1675
const struct xt_target *target = t->u.target;
net/bridge/netfilter/ebtables.c
1678
compat_uint_t tsize = t->target_size - off;
net/bridge/netfilter/ebtables.c
1680
if (WARN_ON(off >= t->target_size))
net/bridge/netfilter/ebtables.c
1689
if (target->compat_to_user(cm->data, t->data))
net/bridge/netfilter/ebtables.c
1692
if (xt_data_to_user(cm->data, t->data, target->usersize, tsize,
net/bridge/netfilter/ebtables.c
1714
struct ebt_entry_target *t;
net/bridge/netfilter/ebtables.c
1751
t = ebt_get_target(e);
net/bridge/netfilter/ebtables.c
1753
ret = compat_target_to_user(t, dstptr, size);
net/bridge/netfilter/ebtables.c
1786
const struct ebt_entry_target *t;
net/bridge/netfilter/ebtables.c
1799
t = ebt_get_target_c(e);
net/bridge/netfilter/ebtables.c
1801
off += xt_compat_target_offset(t->u.target);
net/bridge/netfilter/ebtables.c
1851
static int compat_copy_everything_to_user(struct ebt_table *t,
net/bridge/netfilter/ebtables.c
1863
tinfo.entries_size = t->private->entries_size;
net/bridge/netfilter/ebtables.c
1864
tinfo.nentries = t->private->nentries;
net/bridge/netfilter/ebtables.c
1865
tinfo.entries = t->private->entries;
net/bridge/netfilter/ebtables.c
1866
oldcounters = t->private->counters;
net/bridge/netfilter/ebtables.c
1868
tinfo.entries_size = t->table->entries_size;
net/bridge/netfilter/ebtables.c
1869
tinfo.nentries = t->table->nentries;
net/bridge/netfilter/ebtables.c
1870
tinfo.entries = t->table->entries;
net/bridge/netfilter/ebtables.c
1871
oldcounters = t->table->counters;
net/bridge/netfilter/ebtables.c
1883
ret = compat_table_info(t->private, &repl);
net/bridge/netfilter/ebtables.c
1897
ret = copy_counters_to_user(t, oldcounters, compat_ptr(tmp.counters),
net/bridge/netfilter/ebtables.c
200
const struct ebt_entry_target *t;
net/bridge/netfilter/ebtables.c
2377
struct ebt_table *t;
net/bridge/netfilter/ebtables.c
2389
t = find_table_lock(net, tmp.name, &ret, &ebt_mutex);
net/bridge/netfilter/ebtables.c
2390
if (!t)
net/bridge/netfilter/ebtables.c
2396
tmp.nentries = t->private->nentries;
net/bridge/netfilter/ebtables.c
2397
ret = compat_table_info(t->private, &tmp);
net/bridge/netfilter/ebtables.c
2400
tmp.valid_hooks = t->valid_hooks;
net/bridge/netfilter/ebtables.c
2409
tmp.nentries = t->table->nentries;
net/bridge/netfilter/ebtables.c
2410
tmp.entries_size = t->table->entries_size;
net/bridge/netfilter/ebtables.c
2411
tmp.valid_hooks = t->table->valid_hooks;
net/bridge/netfilter/ebtables.c
2429
if (copy_everything_to_user(t, user, len, cmd) == 0)
net/bridge/netfilter/ebtables.c
2432
ret = compat_copy_everything_to_user(t, user, len, cmd);
net/bridge/netfilter/ebtables.c
244
t = ebt_get_target_c(point);
net/bridge/netfilter/ebtables.c
2449
struct ebt_table *t;
net/bridge/netfilter/ebtables.c
246
if (!t->u.target->target)
net/bridge/netfilter/ebtables.c
2468
t = find_table_lock(net, tmp.name, &ret, &ebt_mutex);
net/bridge/netfilter/ebtables.c
2469
if (!t)
net/bridge/netfilter/ebtables.c
247
verdict = ((struct ebt_standard_target *)t)->verdict;
net/bridge/netfilter/ebtables.c
2481
tmp.nentries = t->private->nentries;
net/bridge/netfilter/ebtables.c
2482
tmp.entries_size = t->private->entries_size;
net/bridge/netfilter/ebtables.c
2483
tmp.valid_hooks = t->valid_hooks;
net/bridge/netfilter/ebtables.c
2485
tmp.nentries = t->table->nentries;
net/bridge/netfilter/ebtables.c
2486
tmp.entries_size = t->table->entries_size;
net/bridge/netfilter/ebtables.c
2487
tmp.valid_hooks = t->table->valid_hooks;
net/bridge/netfilter/ebtables.c
249
acpar.target = t->u.target;
net/bridge/netfilter/ebtables.c
2499
ret = copy_everything_to_user(t, user, len, cmd);
net/bridge/netfilter/ebtables.c
250
acpar.targinfo = t->data;
net/bridge/netfilter/ebtables.c
251
verdict = t->u.target->target(skb, &acpar);
net/bridge/netfilter/ebtables.c
659
struct ebt_entry_target *t;
net/bridge/netfilter/ebtables.c
668
t = ebt_get_target(e);
net/bridge/netfilter/ebtables.c
671
par.target = t->u.target;
net/bridge/netfilter/ebtables.c
672
par.targinfo = t->data;
net/bridge/netfilter/ebtables.c
686
struct ebt_entry_target *t;
net/bridge/netfilter/ebtables.c
746
t = ebt_get_target(e);
net/bridge/netfilter/ebtables.c
749
target = xt_request_find_target(NFPROTO_BRIDGE, t->u.name, 0);
net/bridge/netfilter/ebtables.c
762
t->u.target = target;
net/bridge/netfilter/ebtables.c
763
if (t->u.target == &ebt_standard_target) {
net/bridge/netfilter/ebtables.c
768
if (((struct ebt_standard_target *)t)->verdict <
net/bridge/netfilter/ebtables.c
773
} else if (t->target_size > gap - sizeof(struct ebt_entry_target)) {
net/bridge/netfilter/ebtables.c
774
module_put(t->u.target->me);
net/bridge/netfilter/ebtables.c
780
tgpar.targinfo = t->data;
net/bridge/netfilter/ebtables.c
781
ret = xt_check_target(&tgpar, t->target_size,
net/bridge/netfilter/ebtables.c
805
const struct ebt_entry_target *t;
net/bridge/netfilter/ebtables.c
824
t = ebt_get_target_c(e);
net/bridge/netfilter/ebtables.c
825
if (strcmp(t->u.name, EBT_STANDARD_TARGET))
net/bridge/netfilter/ebtables.c
831
verdict = ((struct ebt_standard_target *)t)->verdict;
net/can/af_can.h
101
void can_stat_update(struct timer_list *t);
net/can/proc.c
115
void can_stat_update(struct timer_list *t)
net/can/proc.c
117
struct net *net = timer_container_of(net, t, can.stattimer);
net/ceph/crush/mapper.c
116
unsigned int t = work->perm[p + i];
net/ceph/crush/mapper.c
118
work->perm[p] = t;
net/ceph/crush/mapper.c
201
__u64 t;
net/ceph/crush/mapper.c
210
t = (__u64)crush_hash32_4(bucket->h.hash, x, n, r,
net/ceph/crush/mapper.c
212
t = t >> 32;
net/ceph/crush/mapper.c
216
if (t < bucket->node_weights[l])
net/ceph/debugfs.c
189
static void dump_target(struct seq_file *s, struct ceph_osd_request_target *t)
net/ceph/debugfs.c
193
seq_printf(s, "osd%d\t%llu.%x\t", t->osd, t->pgid.pool, t->pgid.seed);
net/ceph/debugfs.c
194
dump_spgid(s, &t->spgid);
net/ceph/debugfs.c
196
for (i = 0; i < t->up.size; i++)
net/ceph/debugfs.c
197
seq_printf(s, "%s%d", (!i ? "" : ","), t->up.osds[i]);
net/ceph/debugfs.c
198
seq_printf(s, "]/%d\t[", t->up.primary);
net/ceph/debugfs.c
199
for (i = 0; i < t->acting.size; i++)
net/ceph/debugfs.c
200
seq_printf(s, "%s%d", (!i ? "" : ","), t->acting.osds[i]);
net/ceph/debugfs.c
201
seq_printf(s, "]/%d\te%u\t", t->acting.primary, t->epoch);
net/ceph/debugfs.c
202
if (t->target_oloc.pool_ns) {
net/ceph/debugfs.c
204
(int)t->target_oloc.pool_ns->len,
net/ceph/debugfs.c
205
t->target_oloc.pool_ns->str,
net/ceph/debugfs.c
206
t->target_oid.name_len, t->target_oid.name, t->flags);
net/ceph/debugfs.c
208
seq_printf(s, "%*pE\t0x%x", t->target_oid.name_len,
net/ceph/debugfs.c
209
t->target_oid.name, t->flags);
net/ceph/debugfs.c
211
if (t->paused)
net/ceph/debugfs.c
259
dump_target(s, &lreq->t);
net/ceph/osd_client.c
1522
const struct ceph_osd_request_target *t,
net/ceph/osd_client.c
1530
WARN_ON(pi->id != t->target_oloc.pool);
net/ceph/osd_client.c
1531
return ((t->flags & CEPH_OSD_FLAG_READ) && pauserd) ||
net/ceph/osd_client.c
1532
((t->flags & CEPH_OSD_FLAG_WRITE) && pausewr) ||
net/ceph/osd_client.c
1582
struct ceph_osd_request_target *t,
net/ceph/osd_client.c
1589
bool is_read = t->flags & CEPH_OSD_FLAG_READ;
net/ceph/osd_client.c
1590
bool is_write = t->flags & CEPH_OSD_FLAG_WRITE;
net/ceph/osd_client.c
1600
t->epoch = osdc->osdmap->epoch;
net/ceph/osd_client.c
1601
pi = ceph_pg_pool_by_id(osdc->osdmap, t->base_oloc.pool);
net/ceph/osd_client.c
1603
t->osd = CEPH_HOMELESS_OSD;
net/ceph/osd_client.c
1609
if (t->last_force_resend < pi->last_force_request_resend) {
net/ceph/osd_client.c
1610
t->last_force_resend = pi->last_force_request_resend;
net/ceph/osd_client.c
1612
} else if (t->last_force_resend == 0) {
net/ceph/osd_client.c
1618
ceph_oid_copy(&t->target_oid, &t->base_oid);
net/ceph/osd_client.c
1619
ceph_oloc_copy(&t->target_oloc, &t->base_oloc);
net/ceph/osd_client.c
1620
if ((t->flags & CEPH_OSD_FLAG_IGNORE_OVERLAY) == 0) {
net/ceph/osd_client.c
1622
t->target_oloc.pool = pi->read_tier;
net/ceph/osd_client.c
1624
t->target_oloc.pool = pi->write_tier;
net/ceph/osd_client.c
1626
pi = ceph_pg_pool_by_id(osdc->osdmap, t->target_oloc.pool);
net/ceph/osd_client.c
1628
t->osd = CEPH_HOMELESS_OSD;
net/ceph/osd_client.c
1634
__ceph_object_locator_to_pg(pi, &t->target_oid, &t->target_oloc, &pgid);
net/ceph/osd_client.c
1636
last_pgid.seed = ceph_stable_mod(pgid.seed, t->pg_num, t->pg_num_mask);
net/ceph/osd_client.c
1640
ceph_is_new_interval(&t->acting,
net/ceph/osd_client.c
1642
&t->up,
net/ceph/osd_client.c
1644
t->size,
net/ceph/osd_client.c
1646
t->min_size,
net/ceph/osd_client.c
1648
t->pg_num,
net/ceph/osd_client.c
1650
t->sort_bitwise,
net/ceph/osd_client.c
1652
t->recovery_deletes,
net/ceph/osd_client.c
1657
should_be_paused = target_should_be_paused(osdc, t, pi);
net/ceph/osd_client.c
1658
if (t->paused && !should_be_paused) {
net/ceph/osd_client.c
1661
if (t->paused != should_be_paused) {
net/ceph/osd_client.c
1662
dout("%s t %p paused %d -> %d\n", __func__, t, t->paused,
net/ceph/osd_client.c
1664
t->paused = should_be_paused;
net/ceph/osd_client.c
1667
legacy_change = ceph_pg_compare(&t->pgid, &pgid) ||
net/ceph/osd_client.c
1668
ceph_osds_changed(&t->acting, &acting,
net/ceph/osd_client.c
1669
t->used_replica || any_change);
net/ceph/osd_client.c
1670
if (t->pg_num)
net/ceph/osd_client.c
1671
split = ceph_pg_is_split(&last_pgid, t->pg_num, pi->pg_num);
net/ceph/osd_client.c
1674
t->pgid = pgid; /* struct */
net/ceph/osd_client.c
1675
ceph_pg_to_primary_shard(osdc->osdmap, pi, &pgid, &t->spgid);
net/ceph/osd_client.c
1676
ceph_osds_copy(&t->acting, &acting);
net/ceph/osd_client.c
1677
ceph_osds_copy(&t->up, &up);
net/ceph/osd_client.c
1678
t->size = pi->size;
net/ceph/osd_client.c
1679
t->min_size = pi->min_size;
net/ceph/osd_client.c
1680
t->pg_num = pi->pg_num;
net/ceph/osd_client.c
1681
t->pg_num_mask = pi->pg_num_mask;
net/ceph/osd_client.c
1682
t->sort_bitwise = sort_bitwise;
net/ceph/osd_client.c
1683
t->recovery_deletes = recovery_deletes;
net/ceph/osd_client.c
1685
if ((t->flags & (CEPH_OSD_FLAG_BALANCE_READS |
net/ceph/osd_client.c
1692
if (t->flags & CEPH_OSD_FLAG_BALANCE_READS) {
net/ceph/osd_client.c
1697
t->osd = acting.osds[pos];
net/ceph/osd_client.c
1698
t->used_replica = pos > 0;
net/ceph/osd_client.c
1700
t->osd = acting.primary;
net/ceph/osd_client.c
1701
t->used_replica = false;
net/ceph/osd_client.c
1711
dout("%s t %p -> %d%d%d%d ct_res %d osd%d\n", __func__, t, unpaused,
net/ceph/osd_client.c
1712
legacy_change, force_resend, split, ct_res, t->osd);
net/ceph/osd_client.c
1996
const struct ceph_osd_request_target *t)
net/ceph/osd_client.c
2000
hoid->oid = t->target_oid.name;
net/ceph/osd_client.c
2001
hoid->oid_len = t->target_oid.name_len;
net/ceph/osd_client.c
2003
hoid->hash = t->pgid.seed;
net/ceph/osd_client.c
2005
if (t->target_oloc.pool_ns) {
net/ceph/osd_client.c
2006
hoid->nspace = t->target_oloc.pool_ns->str;
net/ceph/osd_client.c
2007
hoid->nspace_len = t->target_oloc.pool_ns->len;
net/ceph/osd_client.c
2012
hoid->pool = t->target_oloc.pool;
net/ceph/osd_client.c
2786
target_destroy(&lreq->t);
net/ceph/osd_client.c
2823
target_init(&lreq->t);
net/ceph/osd_client.c
3145
target_copy(&req->r_t, &lreq->t);
net/ceph/osd_client.c
3239
target_copy(&req->r_t, &lreq->t);
net/ceph/osd_client.c
3266
calc_target(osdc, &lreq->t, false);
net/ceph/osd_client.c
3267
osd = lookup_create_osd(osdc, lreq->t.osd, true);
net/ceph/osd_client.c
3910
ct_res = calc_target(osdc, &lreq->t, true);
net/ceph/osd_client.c
3914
osd = lookup_create_osd(osdc, lreq->t.osd, true);
net/ceph/osd_client.c
3952
pool_cleared_full(osdc, lreq->t.base_oloc.pool));
net/ceph/osd_client.c
420
static void target_init(struct ceph_osd_request_target *t)
net/ceph/osd_client.c
422
ceph_oid_init(&t->base_oid);
net/ceph/osd_client.c
423
ceph_oloc_init(&t->base_oloc);
net/ceph/osd_client.c
424
ceph_oid_init(&t->target_oid);
net/ceph/osd_client.c
425
ceph_oloc_init(&t->target_oloc);
net/ceph/osd_client.c
427
ceph_osds_init(&t->acting);
net/ceph/osd_client.c
428
ceph_osds_init(&t->up);
net/ceph/osd_client.c
429
t->size = -1;
net/ceph/osd_client.c
430
t->min_size = -1;
net/ceph/osd_client.c
432
t->osd = CEPH_HOMELESS_OSD;
net/ceph/osd_client.c
4445
static bool target_contained_by(const struct ceph_osd_request_target *t,
net/ceph/osd_client.c
4452
hoid_fill_from_target(&hoid, t);
net/ceph/osd_client.c
464
static void target_destroy(struct ceph_osd_request_target *t)
net/ceph/osd_client.c
466
ceph_oid_destroy(&t->base_oid);
net/ceph/osd_client.c
467
ceph_oloc_destroy(&t->base_oloc);
net/ceph/osd_client.c
468
ceph_oid_destroy(&t->target_oid);
net/ceph/osd_client.c
469
ceph_oloc_destroy(&t->target_oloc);
net/ceph/osd_client.c
4784
ceph_oid_copy(&lreq->t.base_oid, oid);
net/ceph/osd_client.c
4785
ceph_oloc_copy(&lreq->t.base_oloc, oloc);
net/ceph/osd_client.c
4786
lreq->t.flags = CEPH_OSD_FLAG_WRITE;
net/ceph/osd_client.c
4822
ceph_oid_copy(&req->r_base_oid, &lreq->t.base_oid);
net/ceph/osd_client.c
4823
ceph_oloc_copy(&req->r_base_oloc, &lreq->t.base_oloc);
net/ceph/osd_client.c
4969
ceph_oid_copy(&lreq->t.base_oid, oid);
net/ceph/osd_client.c
4970
ceph_oloc_copy(&lreq->t.base_oloc, oloc);
net/ceph/osd_client.c
4971
lreq->t.flags = CEPH_OSD_FLAG_READ;
net/ceph/osdmap.c
49
static int calc_bits_of(unsigned int t)
net/ceph/osdmap.c
52
while (t) {
net/ceph/osdmap.c
53
t = t >> 1;
net/core/drop_monitor.c
209
static void sched_send_work(struct timer_list *t)
net/core/drop_monitor.c
211
struct per_cpu_dm_data *data = timer_container_of(data, t, send_timer);
net/core/filter.c
10839
sizeof_field(struct minmax_sample, t));
net/core/gen_estimator.c
76
static void est_timer(struct timer_list *t)
net/core/gen_estimator.c
78
struct net_rate_estimator *est = timer_container_of(est, t, timer);
net/core/neighbour.c
1104
static void neigh_timer_handler(struct timer_list *t)
net/core/neighbour.c
1107
struct neighbour *neigh = timer_container_of(neigh, t, timer);
net/core/neighbour.c
1678
static void neigh_proxy_process(struct timer_list *t)
net/core/neighbour.c
1680
struct neigh_table *tbl = timer_container_of(tbl, t, proxy_timer);
net/core/neighbour.c
2947
int t, family, s_t;
net/core/neighbour.c
2968
for (t = 0; t < NEIGH_NR_TABLES; t++) {
net/core/neighbour.c
2969
tbl = rcu_dereference(neigh_tables[t]);
net/core/neighbour.c
2973
if (t < s_t || (family && tbl->family != family))
net/core/neighbour.c
2975
if (t > s_t)
net/core/neighbour.c
2987
cb->args[0] = t;
net/core/neighbour.c
3856
struct neigh_sysctl_table *t;
net/core/neighbour.c
3862
t = kmemdup(&neigh_sysctl_template, sizeof(*t), GFP_KERNEL_ACCOUNT);
net/core/neighbour.c
3863
if (!t)
net/core/neighbour.c
3867
t->neigh_vars[i].data += (long) p;
net/core/neighbour.c
3868
t->neigh_vars[i].extra1 = dev;
net/core/neighbour.c
3869
t->neigh_vars[i].extra2 = p;
net/core/neighbour.c
3872
neigh_vars_size = ARRAY_SIZE(t->neigh_vars);
net/core/neighbour.c
3880
t->neigh_vars[NEIGH_VAR_GC_INTERVAL].data = &tbl->gc_interval;
net/core/neighbour.c
3881
t->neigh_vars[NEIGH_VAR_GC_THRESH1].data = &tbl->gc_thresh1;
net/core/neighbour.c
3882
t->neigh_vars[NEIGH_VAR_GC_THRESH2].data = &tbl->gc_thresh2;
net/core/neighbour.c
3883
t->neigh_vars[NEIGH_VAR_GC_THRESH3].data = &tbl->gc_thresh3;
net/core/neighbour.c
3888
t->neigh_vars[NEIGH_VAR_RETRANS_TIME].proc_handler = handler;
net/core/neighbour.c
3890
t->neigh_vars[NEIGH_VAR_BASE_REACHABLE_TIME].proc_handler = handler;
net/core/neighbour.c
3892
t->neigh_vars[NEIGH_VAR_RETRANS_TIME_MS].proc_handler = handler;
net/core/neighbour.c
3894
t->neigh_vars[NEIGH_VAR_BASE_REACHABLE_TIME_MS].proc_handler = handler;
net/core/neighbour.c
3903
t->neigh_vars[NEIGH_VAR_BASE_REACHABLE_TIME].proc_handler =
net/core/neighbour.c
3906
t->neigh_vars[NEIGH_VAR_BASE_REACHABLE_TIME_MS].proc_handler =
net/core/neighbour.c
3923
t->sysctl_header = register_net_sysctl_sz(neigh_parms_net(p),
net/core/neighbour.c
3924
neigh_path, t->neigh_vars,
net/core/neighbour.c
3926
if (!t->sysctl_header)
net/core/neighbour.c
3929
p->sysctl_table = t;
net/core/neighbour.c
3933
kfree(t);
net/core/neighbour.c
3942
struct neigh_sysctl_table *t = p->sysctl_table;
net/core/neighbour.c
3944
unregister_net_sysctl_table(t->sysctl_header);
net/core/neighbour.c
3945
kfree(t);
net/core/neighbour.c
53
static void neigh_timer_handler(struct timer_list *t);
net/core/net-procfs.c
185
int t;
net/core/net-procfs.c
212
for (t = 0; t < PTYPE_HASH_SIZE; t++) {
net/core/net-procfs.c
213
list_for_each_entry_rcu(pt, &ptype_base[t], list) {
net/core/net_test.c
140
static void gso_test_case_to_desc(struct gso_test_case *t, char *desc)
net/core/net_test.c
142
sprintf(desc, "%s", t->name);
net/core/net_test.c
344
ip_tunnel_flags_test_case_to_desc(const struct ip_tunnel_flags_test *t,
net/core/net_test.c
347
strscpy(desc, t->name, KUNIT_PARAM_DESC_SIZE);
net/core/net_test.c
354
const struct ip_tunnel_flags_test *t = test->param_value;
net/core/net_test.c
359
for (u32 j = 0; j < t->src_num; j++)
net/core/net_test.c
360
__set_bit(t->src_bits[j], src);
net/core/net_test.c
361
for (u32 j = 0; j < t->exp_num; j++)
net/core/net_test.c
362
__set_bit(t->exp_bits[j], exp);
net/core/net_test.c
364
KUNIT_ASSERT_EQ(test, t->exp_comp,
net/core/net_test.c
366
KUNIT_ASSERT_EQ(test, (__force u16)t->exp_val,
net/core/net_test.c
369
ip_tunnel_flags_from_be16(out, t->exp_val);
net/core/pktgen.c
1887
struct pktgen_thread *t = seq->private;
net/core/pktgen.c
1890
BUG_ON(!t);
net/core/pktgen.c
1895
list_for_each_entry_rcu(pkt_dev, &t->if_list, list)
net/core/pktgen.c
1901
list_for_each_entry_rcu(pkt_dev, &t->if_list, list)
net/core/pktgen.c
1905
if (t->result[0])
net/core/pktgen.c
1906
seq_printf(seq, "\nResult: %s\n", t->result);
net/core/pktgen.c
1920
struct pktgen_thread *t = seq->private;
net/core/pktgen.c
1959
if (!t) {
net/core/pktgen.c
1965
pg_result = &(t->result[0]);
net/core/pktgen.c
1981
ret = pktgen_add_device(t, f);
net/core/pktgen.c
1993
t->control |= T_REMDEVALL;
net/core/pktgen.c
2029
struct pktgen_thread *t;
net/core/pktgen.c
2033
list_for_each_entry(t, &pn->pktgen_threads, th_list) {
net/core/pktgen.c
2034
pkt_dev = pktgen_find_dev(t, ifname, exact);
net/core/pktgen.c
2038
t->control |= T_REMDEV;
net/core/pktgen.c
2083
struct pktgen_thread *t;
net/core/pktgen.c
2087
list_for_each_entry(t, &pn->pktgen_threads, th_list) {
net/core/pktgen.c
2090
if_lock(t);
net/core/pktgen.c
2091
list_for_each_entry(pkt_dev, &t->if_list, list) {
net/core/pktgen.c
2106
if_unlock(t);
net/core/pktgen.c
2328
struct hrtimer_sleeper t;
net/core/pktgen.c
2330
hrtimer_setup_sleeper_on_stack(&t, CLOCK_MONOTONIC, HRTIMER_MODE_ABS);
net/core/pktgen.c
2331
hrtimer_set_expires(&t.timer, spin_until);
net/core/pktgen.c
2333
remaining = ktime_to_ns(hrtimer_expires_remaining(&t.timer));
net/core/pktgen.c
2346
hrtimer_sleeper_start_expires(&t, HRTIMER_MODE_ABS);
net/core/pktgen.c
2348
if (likely(t.task))
net/core/pktgen.c
2351
hrtimer_cancel(&t.timer);
net/core/pktgen.c
2352
} while (t.task && pkt_dev->running && !signal_pending(current));
net/core/pktgen.c
236
#define if_lock(t) mutex_lock(&(t->if_lock))
net/core/pktgen.c
2360
destroy_hrtimer_on_stack(&t.timer);
net/core/pktgen.c
237
#define if_unlock(t) mutex_unlock(&(t->if_lock))
net/core/pktgen.c
2444
__u16 t;
net/core/pktgen.c
2447
t = get_random_u32_inclusive(pkt_dev->queue_map_min,
net/core/pktgen.c
2450
t = pkt_dev->cur_queue_map + 1;
net/core/pktgen.c
2451
if (t > pkt_dev->queue_map_max)
net/core/pktgen.c
2452
t = pkt_dev->queue_map_min;
net/core/pktgen.c
2454
pkt_dev->cur_queue_map = t;
net/core/pktgen.c
2571
__u32 t;
net/core/pktgen.c
2574
t = get_random_u32_inclusive(imn, imx - 1);
net/core/pktgen.c
2576
t = ntohl(pkt_dev->cur_saddr);
net/core/pktgen.c
2577
t++;
net/core/pktgen.c
2578
if (t > imx)
net/core/pktgen.c
2579
t = imn;
net/core/pktgen.c
2582
pkt_dev->cur_saddr = htonl(t);
net/core/pktgen.c
2591
__u32 t;
net/core/pktgen.c
2597
t = get_random_u32_inclusive(imn, imx - 1);
net/core/pktgen.c
2598
s = htonl(t);
net/core/pktgen.c
2606
t = ntohl(pkt_dev->cur_daddr);
net/core/pktgen.c
2607
t++;
net/core/pktgen.c
2608
if (t > imx) {
net/core/pktgen.c
2609
t = imn;
net/core/pktgen.c
2611
pkt_dev->cur_daddr = htonl(t);
net/core/pktgen.c
2640
__u32 t;
net/core/pktgen.c
2643
t = get_random_u32_inclusive(pkt_dev->min_pkt_size,
net/core/pktgen.c
2646
t = pkt_dev->cur_pkt_size + 1;
net/core/pktgen.c
2647
if (t > pkt_dev->max_pkt_size)
net/core/pktgen.c
2648
t = pkt_dev->min_pkt_size;
net/core/pktgen.c
2650
pkt_dev->cur_pkt_size = t;
net/core/pktgen.c
2653
__u32 t = get_random_u32_below(IMIX_PRECISION);
net/core/pktgen.c
2654
__u8 entry_index = pkt_dev->imix_distribution[t];
net/core/pktgen.c
3205
static void pktgen_run(struct pktgen_thread *t)
net/core/pktgen.c
3213
list_for_each_entry_rcu(pkt_dev, &t->if_list, list) {
net/core/pktgen.c
3235
t->control &= ~(T_STOP);
net/core/pktgen.c
3240
struct pktgen_thread *t;
net/core/pktgen.c
3244
list_for_each_entry(t, &pn->pktgen_threads, th_list)
net/core/pktgen.c
3245
t->control |= (flags);
net/core/pktgen.c
3257
static int thread_is_running(const struct pktgen_thread *t)
net/core/pktgen.c
3262
list_for_each_entry_rcu(pkt_dev, &t->if_list, list)
net/core/pktgen.c
3271
static int pktgen_wait_thread_run(struct pktgen_thread *t)
net/core/pktgen.c
3273
while (thread_is_running(t)) {
net/core/pktgen.c
3293
struct pktgen_thread *t;
net/core/pktgen.c
3302
list_for_each_entry(t, &pn->pktgen_threads, th_list) {
net/core/pktgen.c
3303
sig = pktgen_wait_thread_run(t);
net/core/pktgen.c
3309
list_for_each_entry(t, &pn->pktgen_threads, th_list)
net/core/pktgen.c
3310
t->control |= (T_STOP);
net/core/pktgen.c
3403
static struct pktgen_dev *next_to_run(struct pktgen_thread *t)
net/core/pktgen.c
3408
list_for_each_entry_rcu(pkt_dev, &t->if_list, list) {
net/core/pktgen.c
3421
static void pktgen_stop(struct pktgen_thread *t)
net/core/pktgen.c
3429
list_for_each_entry_rcu(pkt_dev, &t->if_list, list) {
net/core/pktgen.c
3440
static void pktgen_rem_one_if(struct pktgen_thread *t)
net/core/pktgen.c
3447
list_for_each_safe(q, n, &t->if_list) {
net/core/pktgen.c
3456
pktgen_remove_device(t, cur);
net/core/pktgen.c
3462
static void pktgen_rem_all_ifs(struct pktgen_thread *t)
net/core/pktgen.c
3471
list_for_each_safe(q, n, &t->if_list) {
net/core/pktgen.c
3477
pktgen_remove_device(t, cur);
net/core/pktgen.c
3481
static void pktgen_rem_thread(struct pktgen_thread *t)
net/core/pktgen.c
3484
remove_proc_entry(t->tsk->comm, t->net->proc_dir);
net/core/pktgen.c
3702
struct pktgen_thread *t = arg;
net/core/pktgen.c
3704
int cpu = t->cpu;
net/core/pktgen.c
3708
init_waitqueue_head(&t->queue);
net/core/pktgen.c
3709
complete(&t->start_done);
net/core/pktgen.c
3716
pkt_dev = next_to_run(t);
net/core/pktgen.c
3718
if (unlikely(!pkt_dev && t->control == 0)) {
net/core/pktgen.c
3719
if (t->net->pktgen_exiting)
net/core/pktgen.c
3721
wait_event_freezable_timeout(t->queue,
net/core/pktgen.c
3722
t->control != 0, HZ / 10);
net/core/pktgen.c
3735
if (t->control & T_STOP) {
net/core/pktgen.c
3736
pktgen_stop(t);
net/core/pktgen.c
3737
t->control &= ~(T_STOP);
net/core/pktgen.c
3740
if (t->control & T_RUN) {
net/core/pktgen.c
3741
pktgen_run(t);
net/core/pktgen.c
3742
t->control &= ~(T_RUN);
net/core/pktgen.c
3745
if (t->control & T_REMDEVALL) {
net/core/pktgen.c
3746
pktgen_rem_all_ifs(t);
net/core/pktgen.c
3747
t->control &= ~(T_REMDEVALL);
net/core/pktgen.c
3750
if (t->control & T_REMDEV) {
net/core/pktgen.c
3751
pktgen_rem_one_if(t);
net/core/pktgen.c
3752
t->control &= ~(T_REMDEV);
net/core/pktgen.c
3758
pr_debug("%s stopping all device\n", t->tsk->comm);
net/core/pktgen.c
3759
pktgen_stop(t);
net/core/pktgen.c
3761
pr_debug("%s removing all device\n", t->tsk->comm);
net/core/pktgen.c
3762
pktgen_rem_all_ifs(t);
net/core/pktgen.c
3764
pr_debug("%s removing thread\n", t->tsk->comm);
net/core/pktgen.c
3765
pktgen_rem_thread(t);
net/core/pktgen.c
3770
static struct pktgen_dev *pktgen_find_dev(struct pktgen_thread *t,
net/core/pktgen.c
3777
list_for_each_entry_rcu(p, &t->if_list, list)
net/core/pktgen.c
3796
static int add_dev_to_thread(struct pktgen_thread *t,
net/core/pktgen.c
3808
if_lock(t);
net/core/pktgen.c
3817
pkt_dev->pg_thread = t;
net/core/pktgen.c
3818
list_add_rcu(&pkt_dev->list, &t->if_list);
net/core/pktgen.c
3821
if_unlock(t);
net/core/pktgen.c
3827
static int pktgen_add_device(struct pktgen_thread *t, const char *ifname)
net/core/pktgen.c
3831
int node = cpu_to_node(t->cpu);
net/core/pktgen.c
3835
pkt_dev = __pktgen_NN_threads(t->net, ifname, FIND);
net/core/pktgen.c
3873
err = pktgen_setup_dev(t->net, pkt_dev, ifname);
net/core/pktgen.c
3879
pkt_dev->entry = proc_create_data(ifname, 0600, t->net->proc_dir,
net/core/pktgen.c
3903
return add_dev_to_thread(t, pkt_dev);
net/core/pktgen.c
3917
struct pktgen_thread *t;
net/core/pktgen.c
3921
t = kzalloc_node(sizeof(struct pktgen_thread), GFP_KERNEL,
net/core/pktgen.c
3923
if (!t) {
net/core/pktgen.c
3928
mutex_init(&t->if_lock);
net/core/pktgen.c
3929
t->cpu = cpu;
net/core/pktgen.c
3931
INIT_LIST_HEAD(&t->if_list);
net/core/pktgen.c
3933
list_add_tail(&t->th_list, &pn->pktgen_threads);
net/core/pktgen.c
3934
init_completion(&t->start_done);
net/core/pktgen.c
3936
p = kthread_create_on_cpu(pktgen_thread_worker, t, cpu, "kpktgend_%d");
net/core/pktgen.c
3938
pr_err("kthread_create_on_node() failed for cpu %d\n", t->cpu);
net/core/pktgen.c
3939
list_del(&t->th_list);
net/core/pktgen.c
3940
kfree(t);
net/core/pktgen.c
3944
t->tsk = p;
net/core/pktgen.c
3946
pe = proc_create_data(t->tsk->comm, 0600, pn->proc_dir,
net/core/pktgen.c
3947
&pktgen_thread_proc_ops, t);
net/core/pktgen.c
3950
PG_PROC_DIR, t->tsk->comm);
net/core/pktgen.c
3952
list_del(&t->th_list);
net/core/pktgen.c
3953
kfree(t);
net/core/pktgen.c
3957
t->net = pn;
net/core/pktgen.c
3960
wait_for_completion(&t->start_done);
net/core/pktgen.c
3968
static void _rem_dev_from_if_list(struct pktgen_thread *t,
net/core/pktgen.c
3974
if_lock(t);
net/core/pktgen.c
3975
list_for_each_safe(q, n, &t->if_list) {
net/core/pktgen.c
3980
if_unlock(t);
net/core/pktgen.c
3983
static int pktgen_remove_device(struct pktgen_thread *t,
net/core/pktgen.c
4007
_rem_dev_from_if_list(t, pkt_dev);
net/core/pktgen.c
4069
struct pktgen_thread *t;
net/core/pktgen.c
4081
t = list_entry(q, struct pktgen_thread, th_list);
net/core/pktgen.c
4082
list_del(&t->th_list);
net/core/pktgen.c
4083
kthread_stop_put(t->tsk);
net/core/pktgen.c
4084
kfree(t);
net/core/pktgen.c
483
static int pktgen_remove_device(struct pktgen_thread *t, struct pktgen_dev *i);
net/core/pktgen.c
484
static int pktgen_add_device(struct pktgen_thread *t, const char *ifname);
net/core/pktgen.c
485
static struct pktgen_dev *pktgen_find_dev(struct pktgen_thread *t,
net/core/pktgen.c
492
static void pktgen_stop(struct pktgen_thread *t);
net/ethtool/mm.c
316
static void ethtool_mmsv_verify_timer(struct timer_list *t)
net/ethtool/mm.c
318
struct ethtool_mmsv *mmsv = timer_container_of(mmsv, t, verify_timer);
net/hsr/hsr_device.c
411
static void hsr_announce(struct timer_list *t)
net/hsr/hsr_device.c
417
hsr = timer_container_of(hsr, t, announce_timer);
net/hsr/hsr_device.c
431
static void hsr_proxy_announce(struct timer_list *t)
net/hsr/hsr_device.c
433
struct hsr_priv *hsr = timer_container_of(hsr, t,
net/hsr/hsr_framereg.c
687
void hsr_prune_nodes(struct timer_list *t)
net/hsr/hsr_framereg.c
689
struct hsr_priv *hsr = timer_container_of(hsr, t, prune_timer);
net/hsr/hsr_framereg.c
755
void hsr_prune_proxy_nodes(struct timer_list *t)
net/hsr/hsr_framereg.c
757
struct hsr_priv *hsr = timer_container_of(hsr, t, prune_proxy_timer);
net/hsr/hsr_framereg.h
49
void hsr_prune_nodes(struct timer_list *t);
net/hsr/hsr_framereg.h
50
void hsr_prune_proxy_nodes(struct timer_list *t);
net/ieee802154/6lowpan/reassembly.c
45
static void lowpan_frag_expire(struct timer_list *t)
net/ieee802154/6lowpan/reassembly.c
47
struct inet_frag_queue *frag = timer_container_of(frag, t, timer);
net/ipv4/bpf_tcp_ca.c
141
const struct btf_type *t;
net/ipv4/bpf_tcp_ca.c
145
t = tcp_congestion_ops_type;
net/ipv4/bpf_tcp_ca.c
146
m = &btf_type_member(t)[midx];
net/ipv4/bpf_tcp_ca.c
148
return __btf_member_bit_offset(t, m) / 8;
net/ipv4/bpf_tcp_ca.c
210
static int bpf_tcp_ca_init_member(const struct btf_type *t,
net/ipv4/bpf_tcp_ca.c
221
moff = __btf_member_bit_offset(t, member) / 8;
net/ipv4/bpf_tcp_ca.c
65
const struct btf_type *t;
net/ipv4/bpf_tcp_ca.c
68
t = btf_type_by_id(reg->btf, reg->btf_id);
net/ipv4/bpf_tcp_ca.c
69
if (t != tcp_sock_type) {
net/ipv4/devinet.c
2662
struct devinet_sysctl_table *t;
net/ipv4/devinet.c
2665
t = kmemdup(&devinet_sysctl, sizeof(*t), GFP_KERNEL_ACCOUNT);
net/ipv4/devinet.c
2666
if (!t)
net/ipv4/devinet.c
2669
for (i = 0; i < ARRAY_SIZE(t->devinet_vars); i++) {
net/ipv4/devinet.c
2670
t->devinet_vars[i].data += (char *)p - (char *)&ipv4_devconf;
net/ipv4/devinet.c
2671
t->devinet_vars[i].extra1 = p;
net/ipv4/devinet.c
2672
t->devinet_vars[i].extra2 = net;
net/ipv4/devinet.c
2677
t->sysctl_header = register_net_sysctl(net, path, t->devinet_vars);
net/ipv4/devinet.c
2678
if (!t->sysctl_header)
net/ipv4/devinet.c
2681
p->sysctl = t;
net/ipv4/devinet.c
2688
kfree(t);
net/ipv4/devinet.c
2696
struct devinet_sysctl_table *t = cnf->sysctl;
net/ipv4/devinet.c
2698
if (t) {
net/ipv4/devinet.c
2700
unregister_net_sysctl_table(t->sysctl_header);
net/ipv4/devinet.c
2701
kfree(t);
net/ipv4/fib_trie.c
1017
struct trie *t;
net/ipv4/fib_trie.c
1023
t = (struct trie *)tb->tb_data;
net/ipv4/fib_trie.c
1024
l = fib_find_node(t, &tp, be32_to_cpu(fri->dst));
net/ipv4/fib_trie.c
1098
static void trie_rebalance(struct trie *t, struct key_vector *tn)
net/ipv4/fib_trie.c
1101
tn = resize(t, tn);
net/ipv4/fib_trie.c
1104
static int fib_insert_node(struct trie *t, struct key_vector *tp,
net/ipv4/fib_trie.c
1145
trie_rebalance(t, tp);
net/ipv4/fib_trie.c
1154
static int fib_insert_alias(struct trie *t, struct key_vector *tp,
net/ipv4/fib_trie.c
1159
return fib_insert_node(t, tp, new, key);
net/ipv4/fib_trie.c
1190
static void fib_remove_alias(struct trie *t, struct key_vector *tp,
net/ipv4/fib_trie.c
1197
struct trie *t = (struct trie *)tb->tb_data;
net/ipv4/fib_trie.c
1219
l = fib_find_node(t, &tp, key);
net/ipv4/fib_trie.c
1354
err = fib_insert_alias(t, tp, l, new_fa, fa, key);
net/ipv4/fib_trie.c
1359
l = l ? l : fib_find_node(t, &tp, key);
net/ipv4/fib_trie.c
1386
fib_remove_alias(t, tp, l, new_fa);
net/ipv4/fib_trie.c
1423
struct trie *t = (struct trie *) tb->tb_data;
net/ipv4/fib_trie.c
1425
struct trie_use_stats __percpu *stats = t->stats;
net/ipv4/fib_trie.c
1433
pn = t->kv;
net/ipv4/fib_trie.c
1626
static void fib_remove_alias(struct trie *t, struct key_vector *tp,
net/ipv4/fib_trie.c
1644
trie_rebalance(t, tp);
net/ipv4/fib_trie.c
1691
struct trie *t = (struct trie *) tb->tb_data;
net/ipv4/fib_trie.c
1701
l = fib_find_node(t, &tp, key);
net/ipv4/fib_trie.c
1711
inet_dscp_to_dsfield(dscp), t);
net/ipv4/fib_trie.c
174
static struct key_vector *resize(struct trie *t, struct key_vector *tn);
net/ipv4/fib_trie.c
1746
fib_remove_alias(t, tp, l, fa_to_delete);
net/ipv4/fib_trie.c
1816
struct trie *t = (struct trie *)tb->tb_data;
net/ipv4/fib_trie.c
1817
struct key_vector *pn = t->kv;
net/ipv4/fib_trie.c
1867
free_percpu(t->stats);
net/ipv4/fib_trie.c
1933
struct trie *t = (struct trie *)tb->tb_data;
net/ipv4/fib_trie.c
1934
struct key_vector *pn = t->kv;
net/ipv4/fib_trie.c
1956
pn = resize(t, pn);
net/ipv4/fib_trie.c
2002
struct trie *t = (struct trie *)tb->tb_data;
net/ipv4/fib_trie.c
2004
struct key_vector *pn = t->kv;
net/ipv4/fib_trie.c
2027
pn = resize(t, pn);
net/ipv4/fib_trie.c
2093
struct trie *t = (struct trie *)tb->tb_data;
net/ipv4/fib_trie.c
2094
struct key_vector *pn = t->kv;
net/ipv4/fib_trie.c
2188
struct trie *t = (struct trie *)tb->tb_data;
net/ipv4/fib_trie.c
2189
struct key_vector *l, *tp = t->kv;
net/ipv4/fib_trie.c
2229
struct trie *t = (struct trie *)tb->tb_data;
net/ipv4/fib_trie.c
2232
free_percpu(t->stats);
net/ipv4/fib_trie.c
2332
struct trie *t = (struct trie *)tb->tb_data;
net/ipv4/fib_trie.c
2333
struct key_vector *l, *tp = t->kv;
net/ipv4/fib_trie.c
2387
struct trie *t;
net/ipv4/fib_trie.c
2404
t = (struct trie *) tb->tb_data;
net/ipv4/fib_trie.c
2405
t->kv[0].pos = KEYLENGTH;
net/ipv4/fib_trie.c
2406
t->kv[0].slen = KEYLENGTH;
net/ipv4/fib_trie.c
2408
t->stats = alloc_percpu(struct trie_use_stats);
net/ipv4/fib_trie.c
2409
if (!t->stats) {
net/ipv4/fib_trie.c
2472
struct trie *t)
net/ipv4/fib_trie.c
2476
if (!t)
net/ipv4/fib_trie.c
2479
pn = t->kv;
net/ipv4/fib_trie.c
2497
static void trie_collect_stats(struct trie *t, struct trie_stat *s)
net/ipv4/fib_trie.c
2505
for (n = fib_trie_get_first(&iter, t); n; n = fib_trie_get_next(&iter)) {
net/ipv4/fib_trie.c
2626
struct trie *t = (struct trie *) tb->tb_data;
net/ipv4/fib_trie.c
2629
if (!t)
net/ipv4/fib_trie.c
2634
trie_collect_stats(t, &stat);
net/ipv4/fib_trie.c
2637
trie_show_usage(seq, t->stats);
net/ipv4/fib_trie.c
2762
static inline const char *rtn_type(char *buf, size_t len, unsigned int t)
net/ipv4/fib_trie.c
2764
if (t < __RTN_MAX && rtn_type_names[t])
net/ipv4/fib_trie.c
2765
return rtn_type_names[t];
net/ipv4/fib_trie.c
2766
snprintf(buf, len, "type %u", t);
net/ipv4/fib_trie.c
2868
struct trie *t;
net/ipv4/fib_trie.c
2877
t = (struct trie *)tb->tb_data;
net/ipv4/fib_trie.c
2878
iter->tnode = t->kv;
net/ipv4/fib_trie.c
502
static struct key_vector *replace(struct trie *t,
net/ipv4/fib_trie.c
525
tn = resize(t, inode);
net/ipv4/fib_trie.c
531
static struct key_vector *inflate(struct trie *t,
net/ipv4/fib_trie.c
619
return replace(t, oldtnode, tn);
net/ipv4/fib_trie.c
627
static struct key_vector *halve(struct trie *t,
net/ipv4/fib_trie.c
674
return replace(t, oldtnode, tn);
net/ipv4/fib_trie.c
682
static struct key_vector *collapse(struct trie *t,
net/ipv4/fib_trie.c
842
static struct key_vector *resize(struct trie *t, struct key_vector *tn)
net/ipv4/fib_trie.c
845
struct trie_use_stats __percpu *stats = t->stats;
net/ipv4/fib_trie.c
864
tp = inflate(t, tn);
net/ipv4/fib_trie.c
887
tp = halve(t, tn);
net/ipv4/fib_trie.c
901
return collapse(t, tn);
net/ipv4/fib_trie.c
930
static struct key_vector *fib_find_node(struct trie *t,
net/ipv4/fib_trie.c
933
struct key_vector *pn, *n = t->kv;
net/ipv4/igmp.c
802
static void igmp_gq_timer_expire(struct timer_list *t)
net/ipv4/igmp.c
804
struct in_device *in_dev = timer_container_of(in_dev, t, mr_gq_timer);
net/ipv4/igmp.c
811
static void igmp_ifc_timer_expire(struct timer_list *t)
net/ipv4/igmp.c
813
struct in_device *in_dev = timer_container_of(in_dev, t, mr_ifc_timer);
net/ipv4/igmp.c
841
static void igmp_timer_expire(struct timer_list *t)
net/ipv4/igmp.c
843
struct ip_mc_list *im = timer_container_of(im, t, timer);
net/ipv4/inet_connection_sock.c
1027
static void reqsk_timer_handler(struct timer_list *t)
net/ipv4/inet_connection_sock.c
1029
struct request_sock *req = timer_container_of(req, t, rsk_timer);
net/ipv4/inet_connection_sock.c
721
void (*retransmit_handler)(struct timer_list *t),
net/ipv4/inet_connection_sock.c
722
void (*delack_handler)(struct timer_list *t),
net/ipv4/inet_connection_sock.c
723
void (*keepalive_handler)(struct timer_list *t))
net/ipv4/inet_timewait_sock.c
161
static void tw_timer_handler(struct timer_list *t)
net/ipv4/inet_timewait_sock.c
163
struct inet_timewait_sock *tw = timer_container_of(tw, t, tw_timer);
net/ipv4/ip_fragment.c
123
static void ip_expire(struct timer_list *t)
net/ipv4/ip_fragment.c
126
struct inet_frag_queue *frag = timer_container_of(frag, t, timer);
net/ipv4/ip_gre.c
1187
struct ip_tunnel *t = netdev_priv(dev);
net/ipv4/ip_gre.c
1226
if (t->ignore_df)
net/ipv4/ip_gre.c
1232
t->collect_md = true;
net/ipv4/ip_gre.c
1241
t->ignore_df = !!nla_get_u8(data[IFLA_GRE_IGNORE_DF]);
net/ipv4/ip_gre.c
1256
struct ip_tunnel *t = netdev_priv(dev);
net/ipv4/ip_gre.c
1266
t->erspan_ver = nla_get_u8(data[IFLA_GRE_ERSPAN_VER]);
net/ipv4/ip_gre.c
1268
if (t->erspan_ver > 2)
net/ipv4/ip_gre.c
1272
if (t->erspan_ver == 1) {
net/ipv4/ip_gre.c
1274
t->index = nla_get_u32(data[IFLA_GRE_ERSPAN_INDEX]);
net/ipv4/ip_gre.c
1275
if (t->index & ~INDEX_MASK)
net/ipv4/ip_gre.c
1278
} else if (t->erspan_ver == 2) {
net/ipv4/ip_gre.c
1280
t->dir = nla_get_u8(data[IFLA_GRE_ERSPAN_DIR]);
net/ipv4/ip_gre.c
1281
if (t->dir & ~(DIR_MASK >> DIR_OFFSET))
net/ipv4/ip_gre.c
1285
t->hwid = nla_get_u16(data[IFLA_GRE_ERSPAN_HWID]);
net/ipv4/ip_gre.c
1286
if (t->hwid & ~(HWID_MASK >> HWID_OFFSET))
net/ipv4/ip_gre.c
1398
struct ip_tunnel *t = netdev_priv(dev);
net/ipv4/ip_gre.c
1399
int err = ip_tunnel_encap_setup(t, &ipencap);
net/ipv4/ip_gre.c
144
struct ip_tunnel *t;
net/ipv4/ip_gre.c
1454
struct ip_tunnel *t = netdev_priv(dev);
net/ipv4/ip_gre.c
1456
__u32 fwmark = t->fwmark;
net/ipv4/ip_gre.c
1471
ip_tunnel_flags_copy(t->parms.i_flags, p.i_flags);
net/ipv4/ip_gre.c
1472
ip_tunnel_flags_copy(t->parms.o_flags, p.o_flags);
net/ipv4/ip_gre.c
1483
struct ip_tunnel *t = netdev_priv(dev);
net/ipv4/ip_gre.c
1485
__u32 fwmark = t->fwmark;
net/ipv4/ip_gre.c
1500
ip_tunnel_flags_copy(t->parms.i_flags, p.i_flags);
net/ipv4/ip_gre.c
1501
ip_tunnel_flags_copy(t->parms.o_flags, p.o_flags);
net/ipv4/ip_gre.c
155
t = ip_tunnel_lookup(itn, skb->dev->ifindex, tpi->flags,
net/ipv4/ip_gre.c
1556
struct ip_tunnel *t = netdev_priv(dev);
net/ipv4/ip_gre.c
1557
struct ip_tunnel_parm_kern *p = &t->parms;
net/ipv4/ip_gre.c
1575
nla_put_u32(skb, IFLA_GRE_FWMARK, t->fwmark))
net/ipv4/ip_gre.c
1579
t->encap.type) ||
net/ipv4/ip_gre.c
158
if (!t)
net/ipv4/ip_gre.c
1581
t->encap.sport) ||
net/ipv4/ip_gre.c
1583
t->encap.dport) ||
net/ipv4/ip_gre.c
1585
t->encap.flags))
net/ipv4/ip_gre.c
1588
if (nla_put_u8(skb, IFLA_GRE_IGNORE_DF, t->ignore_df))
net/ipv4/ip_gre.c
1591
if (t->collect_md) {
net/ipv4/ip_gre.c
1604
struct ip_tunnel *t = netdev_priv(dev);
net/ipv4/ip_gre.c
1606
if (t->erspan_ver <= 2) {
net/ipv4/ip_gre.c
1607
if (t->erspan_ver != 0 && !t->collect_md)
net/ipv4/ip_gre.c
1608
__set_bit(IP_TUNNEL_KEY_BIT, t->parms.o_flags);
net/ipv4/ip_gre.c
1610
if (nla_put_u8(skb, IFLA_GRE_ERSPAN_VER, t->erspan_ver))
net/ipv4/ip_gre.c
1613
if (t->erspan_ver == 1) {
net/ipv4/ip_gre.c
1614
if (nla_put_u32(skb, IFLA_GRE_ERSPAN_INDEX, t->index))
net/ipv4/ip_gre.c
1616
} else if (t->erspan_ver == 2) {
net/ipv4/ip_gre.c
1617
if (nla_put_u8(skb, IFLA_GRE_ERSPAN_DIR, t->dir))
net/ipv4/ip_gre.c
1619
if (nla_put_u16(skb, IFLA_GRE_ERSPAN_HWID, t->hwid))
net/ipv4/ip_gre.c
1632
struct ip_tunnel *t = netdev_priv(dev);
net/ipv4/ip_gre.c
1640
t->erspan_ver = 1;
net/ipv4/ip_gre.c
1719
struct ip_tunnel *t;
net/ipv4/ip_gre.c
1731
t = netdev_priv(dev);
net/ipv4/ip_gre.c
1732
t->collect_md = true;
net/ipv4/ip_gre.c
203
if (t->parms.iph.daddr == 0 ||
net/ipv4/ip_gre.c
204
ipv4_is_multicast(t->parms.iph.daddr))
net/ipv4/ip_gre.c
207
if (t->parms.iph.ttl == 0 && type == ICMP_TIME_EXCEEDED)
net/ipv4/ip_gre.c
210
if (time_before(jiffies, t->err_time + IPTUNNEL_ERR_TIMEO))
net/ipv4/ip_gre.c
211
t->err_count++;
net/ipv4/ip_gre.c
213
t->err_count = 1;
net/ipv4/ip_gre.c
214
t->err_time = jiffies;
net/ipv4/ip_gre.c
845
struct ip_tunnel *t = netdev_priv(dev);
net/ipv4/ip_gre.c
847
ip_tunnel_flags_copy(t->parms.i_flags, p->i_flags);
net/ipv4/ip_gre.c
848
ip_tunnel_flags_copy(t->parms.o_flags, p->o_flags);
net/ipv4/ip_gre.c
893
struct ip_tunnel *t = netdev_priv(dev);
net/ipv4/ip_gre.c
898
needed = t->hlen + sizeof(*iph);
net/ipv4/ip_gre.c
906
greh->flags = gre_tnl_flags_to_gre_flags(t->parms.o_flags);
net/ipv4/ip_gre.c
909
memcpy(iph, &t->parms.iph, sizeof(struct iphdr));
net/ipv4/ip_gre.c
917
return t->hlen + sizeof(*iph);
net/ipv4/ip_gre.c
919
return -(t->hlen + sizeof(*iph));
net/ipv4/ip_gre.c
938
struct ip_tunnel *t = netdev_priv(dev);
net/ipv4/ip_gre.c
940
if (ipv4_is_multicast(t->parms.iph.daddr)) {
net/ipv4/ip_gre.c
942
.flowi4_oif = t->parms.link,
net/ipv4/ip_gre.c
943
.flowi4_dscp = ip4h_dscp(&t->parms.iph),
net/ipv4/ip_gre.c
946
.saddr = t->parms.iph.saddr,
net/ipv4/ip_gre.c
947
.daddr = t->parms.iph.daddr,
net/ipv4/ip_gre.c
948
.fl4_gre_key = t->parms.o_key,
net/ipv4/ip_gre.c
952
rt = ip_route_output_key(t->net, &fl4);
net/ipv4/ip_gre.c
959
t->mlink = dev->ifindex;
net/ipv4/ip_gre.c
960
ip_mc_inc_group(__in_dev_get_rtnl(dev), t->parms.iph.daddr);
net/ipv4/ip_gre.c
967
struct ip_tunnel *t = netdev_priv(dev);
net/ipv4/ip_gre.c
969
if (ipv4_is_multicast(t->parms.iph.daddr) && t->mlink) {
net/ipv4/ip_gre.c
971
in_dev = inetdev_by_index(t->net, t->mlink);
net/ipv4/ip_gre.c
973
ip_mc_dec_group(in_dev, t->parms.iph.daddr);
net/ipv4/ip_tunnel.c
100
if (!ip_tunnel_key_match(&t->parms, flags, key))
net/ipv4/ip_tunnel.c
103
if (READ_ONCE(t->parms.link) == link)
net/ipv4/ip_tunnel.c
104
return t;
net/ipv4/ip_tunnel.c
105
cand = t;
net/ipv4/ip_tunnel.c
108
hlist_for_each_entry_rcu(t, head, hash_node) {
net/ipv4/ip_tunnel.c
109
if (remote != t->parms.iph.daddr ||
net/ipv4/ip_tunnel.c
110
t->parms.iph.saddr != 0 ||
net/ipv4/ip_tunnel.c
111
!(t->dev->flags & IFF_UP))
net/ipv4/ip_tunnel.c
114
if (!ip_tunnel_key_match(&t->parms, flags, key))
net/ipv4/ip_tunnel.c
117
if (READ_ONCE(t->parms.link) == link)
net/ipv4/ip_tunnel.c
1178
struct ip_tunnel *t;
net/ipv4/ip_tunnel.c
118
return t;
net/ipv4/ip_tunnel.c
1182
hlist_for_each_entry_safe(t, n, thead, hash_node)
net/ipv4/ip_tunnel.c
1186
if (!net_eq(dev_net(t->dev), net))
net/ipv4/ip_tunnel.c
1187
unregister_netdevice_queue(t->dev, head);
net/ipv4/ip_tunnel.c
120
cand = t;
net/ipv4/ip_tunnel.c
1249
struct ip_tunnel *t;
net/ipv4/ip_tunnel.c
1257
t = ip_tunnel_find(itn, p, dev->type);
net/ipv4/ip_tunnel.c
1259
if (t) {
net/ipv4/ip_tunnel.c
126
hlist_for_each_entry_rcu(t, head, hash_node) {
net/ipv4/ip_tunnel.c
1260
if (t->dev != dev)
net/ipv4/ip_tunnel.c
1263
t = tunnel;
net/ipv4/ip_tunnel.c
127
if ((local != t->parms.iph.saddr || t->parms.iph.daddr != 0) &&
net/ipv4/ip_tunnel.c
1279
ip_tunnel_update(itn, t, dev, p, !tb[IFLA_MTU], fwmark);
net/ipv4/ip_tunnel.c
128
(local != t->parms.iph.daddr || !ipv4_is_multicast(local)))
net/ipv4/ip_tunnel.c
131
if (!(t->dev->flags & IFF_UP))
net/ipv4/ip_tunnel.c
134
if (!ip_tunnel_key_match(&t->parms, flags, key))
net/ipv4/ip_tunnel.c
137
if (READ_ONCE(t->parms.link) == link)
net/ipv4/ip_tunnel.c
138
return t;
net/ipv4/ip_tunnel.c
140
cand = t;
net/ipv4/ip_tunnel.c
143
hlist_for_each_entry_rcu(t, head, hash_node) {
net/ipv4/ip_tunnel.c
145
t->parms.i_key != key) ||
net/ipv4/ip_tunnel.c
146
t->parms.iph.saddr != 0 ||
net/ipv4/ip_tunnel.c
147
t->parms.iph.daddr != 0 ||
net/ipv4/ip_tunnel.c
148
!(t->dev->flags & IFF_UP))
net/ipv4/ip_tunnel.c
151
if (READ_ONCE(t->parms.link) == link)
net/ipv4/ip_tunnel.c
152
return t;
net/ipv4/ip_tunnel.c
154
cand = t;
net/ipv4/ip_tunnel.c
160
t = rcu_dereference(itn->collect_md_tun);
net/ipv4/ip_tunnel.c
161
if (t && t->dev->flags & IFF_UP)
net/ipv4/ip_tunnel.c
162
return t;
net/ipv4/ip_tunnel.c
192
static void ip_tunnel_add(struct ip_tunnel_net *itn, struct ip_tunnel *t)
net/ipv4/ip_tunnel.c
194
struct hlist_head *head = ip_bucket(itn, &t->parms);
net/ipv4/ip_tunnel.c
196
if (t->collect_md)
net/ipv4/ip_tunnel.c
197
rcu_assign_pointer(itn->collect_md_tun, t);
net/ipv4/ip_tunnel.c
198
hlist_add_head_rcu(&t->hash_node, head);
net/ipv4/ip_tunnel.c
201
static void ip_tunnel_del(struct ip_tunnel_net *itn, struct ip_tunnel *t)
net/ipv4/ip_tunnel.c
203
if (t->collect_md)
net/ipv4/ip_tunnel.c
205
hlist_del_init_rcu(&t->hash_node);
net/ipv4/ip_tunnel.c
217
struct ip_tunnel *t = NULL;
net/ipv4/ip_tunnel.c
222
hlist_for_each_entry_rcu(t, head, hash_node, lockdep_rtnl_is_held()) {
net/ipv4/ip_tunnel.c
223
if (local == t->parms.iph.saddr &&
net/ipv4/ip_tunnel.c
224
remote == t->parms.iph.daddr &&
net/ipv4/ip_tunnel.c
225
link == READ_ONCE(t->parms.link) &&
net/ipv4/ip_tunnel.c
226
type == t->dev->type &&
net/ipv4/ip_tunnel.c
227
ip_tunnel_key_match(&t->parms, flags, key))
net/ipv4/ip_tunnel.c
230
return t;
net/ipv4/ip_tunnel.c
489
int ip_tunnel_encap_setup(struct ip_tunnel *t,
net/ipv4/ip_tunnel.c
494
memset(&t->encap, 0, sizeof(t->encap));
net/ipv4/ip_tunnel.c
500
t->encap.type = ipencap->type;
net/ipv4/ip_tunnel.c
501
t->encap.sport = ipencap->sport;
net/ipv4/ip_tunnel.c
502
t->encap.dport = ipencap->dport;
net/ipv4/ip_tunnel.c
503
t->encap.flags = ipencap->flags;
net/ipv4/ip_tunnel.c
505
t->encap_hlen = hlen;
net/ipv4/ip_tunnel.c
506
t->hlen = t->encap_hlen + t->tun_hlen;
net/ipv4/ip_tunnel.c
86
struct ip_tunnel *t, *cand = NULL;
net/ipv4/ip_tunnel.c
860
struct ip_tunnel *t,
net/ipv4/ip_tunnel.c
866
ip_tunnel_del(itn, t);
net/ipv4/ip_tunnel.c
867
t->parms.iph.saddr = p->iph.saddr;
net/ipv4/ip_tunnel.c
868
t->parms.iph.daddr = p->iph.daddr;
net/ipv4/ip_tunnel.c
869
t->parms.i_key = p->i_key;
net/ipv4/ip_tunnel.c
870
t->parms.o_key = p->o_key;
net/ipv4/ip_tunnel.c
875
ip_tunnel_add(itn, t);
net/ipv4/ip_tunnel.c
877
t->parms.iph.ttl = p->iph.ttl;
net/ipv4/ip_tunnel.c
878
t->parms.iph.tos = p->iph.tos;
net/ipv4/ip_tunnel.c
879
t->parms.iph.frag_off = p->iph.frag_off;
net/ipv4/ip_tunnel.c
881
if (t->parms.link != p->link || t->fwmark != fwmark) {
net/ipv4/ip_tunnel.c
884
WRITE_ONCE(t->parms.link, p->link);
net/ipv4/ip_tunnel.c
885
t->fwmark = fwmark;
net/ipv4/ip_tunnel.c
890
dst_cache_reset(&t->dst_cache);
net/ipv4/ip_tunnel.c
898
struct ip_tunnel *t = netdev_priv(dev);
net/ipv4/ip_tunnel.c
899
struct net *net = t->net;
net/ipv4/ip_tunnel.c
900
struct ip_tunnel_net *itn = net_generic(net, t->ip_tnl_net_id);
net/ipv4/ip_tunnel.c
905
t = ip_tunnel_find(itn, p, itn->fb_tunnel_dev->type);
net/ipv4/ip_tunnel.c
906
if (!t)
net/ipv4/ip_tunnel.c
907
t = netdev_priv(dev);
net/ipv4/ip_tunnel.c
909
memcpy(p, &t->parms, sizeof(*p));
net/ipv4/ip_tunnel.c
926
t = ip_tunnel_find(itn, p, itn->type);
net/ipv4/ip_tunnel.c
929
if (!t) {
net/ipv4/ip_tunnel.c
930
t = ip_tunnel_create(net, itn, p);
net/ipv4/ip_tunnel.c
931
err = PTR_ERR_OR_ZERO(t);
net/ipv4/ip_tunnel.c
939
if (t) {
net/ipv4/ip_tunnel.c
94
hlist_for_each_entry_rcu(t, head, hash_node) {
net/ipv4/ip_tunnel.c
940
if (t->dev != dev) {
net/ipv4/ip_tunnel.c
95
if (local != t->parms.iph.saddr ||
net/ipv4/ip_tunnel.c
957
t = netdev_priv(dev);
net/ipv4/ip_tunnel.c
96
remote != t->parms.iph.daddr ||
net/ipv4/ip_tunnel.c
961
if (t) {
net/ipv4/ip_tunnel.c
963
ip_tunnel_update(itn, t, dev, p, true, 0);
net/ipv4/ip_tunnel.c
97
!(t->dev->flags & IFF_UP))
net/ipv4/ip_tunnel.c
976
t = ip_tunnel_find(itn, p, itn->fb_tunnel_dev->type);
net/ipv4/ip_tunnel.c
977
if (!t)
net/ipv4/ip_tunnel.c
980
if (t == netdev_priv(itn->fb_tunnel_dev))
net/ipv4/ip_tunnel.c
982
dev = t->dev;
net/ipv4/ip_vti.c
595
struct ip_tunnel *t = netdev_priv(dev);
net/ipv4/ip_vti.c
597
__u32 fwmark = t->fwmark;
net/ipv4/ip_vti.c
623
struct ip_tunnel *t = netdev_priv(dev);
net/ipv4/ip_vti.c
624
struct ip_tunnel_parm_kern *p = &t->parms;
net/ipv4/ip_vti.c
631
nla_put_u32(skb, IFLA_VTI_FWMARK, t->fwmark))
net/ipv4/ipcomp.c
102
struct xfrm_state *t;
net/ipv4/ipcomp.c
105
t = xfrm_state_lookup(net, mark, (xfrm_address_t *)&x->id.daddr.a4,
net/ipv4/ipcomp.c
107
if (!t) {
net/ipv4/ipcomp.c
108
t = ipcomp_tunnel_create(x);
net/ipv4/ipcomp.c
109
if (!t) {
net/ipv4/ipcomp.c
113
xfrm_state_insert(t);
net/ipv4/ipcomp.c
114
xfrm_state_hold(t);
net/ipv4/ipcomp.c
116
x->tunnel = t;
net/ipv4/ipcomp.c
117
atomic_inc(&t->tunnel_users);
net/ipv4/ipcomp.c
61
struct xfrm_state *t;
net/ipv4/ipcomp.c
63
t = xfrm_state_alloc(net);
net/ipv4/ipcomp.c
64
if (!t)
net/ipv4/ipcomp.c
66
lockdep_set_class(&t->lock, &xfrm_state_lock_key);
net/ipv4/ipcomp.c
68
t->id.proto = IPPROTO_IPIP;
net/ipv4/ipcomp.c
69
t->id.spi = x->props.saddr.a4;
net/ipv4/ipcomp.c
70
t->id.daddr.a4 = x->id.daddr.a4;
net/ipv4/ipcomp.c
71
memcpy(&t->sel, &x->sel, sizeof(t->sel));
net/ipv4/ipcomp.c
72
t->props.family = AF_INET;
net/ipv4/ipcomp.c
73
t->props.mode = x->props.mode;
net/ipv4/ipcomp.c
74
t->props.saddr.a4 = x->props.saddr.a4;
net/ipv4/ipcomp.c
75
t->props.flags = x->props.flags;
net/ipv4/ipcomp.c
76
t->props.extra_flags = x->props.extra_flags;
net/ipv4/ipcomp.c
77
memcpy(&t->mark, &x->mark, sizeof(t->mark));
net/ipv4/ipcomp.c
78
t->if_id = x->if_id;
net/ipv4/ipcomp.c
80
if (xfrm_init_state(t))
net/ipv4/ipcomp.c
83
atomic_set(&t->tunnel_users, 1);
net/ipv4/ipcomp.c
85
return t;
net/ipv4/ipcomp.c
88
t->km.state = XFRM_STATE_DEAD;
net/ipv4/ipcomp.c
89
xfrm_state_put(t);
net/ipv4/ipcomp.c
90
t = NULL;
net/ipv4/ipip.c
136
struct ip_tunnel *t;
net/ipv4/ipip.c
141
t = ip_tunnel_lookup(itn, skb->dev->ifindex, flags, iph->daddr,
net/ipv4/ipip.c
143
if (!t) {
net/ipv4/ipip.c
176
ipv4_update_pmtu(skb, net, info, t->parms.link, iph->protocol);
net/ipv4/ipip.c
181
ipv4_redirect(skb, net, t->parms.link, iph->protocol);
net/ipv4/ipip.c
185
if (t->parms.iph.daddr == 0) {
net/ipv4/ipip.c
190
if (t->parms.iph.ttl == 0 && type == ICMP_TIME_EXCEEDED)
net/ipv4/ipip.c
193
if (time_before(jiffies, t->err_time + IPTUNNEL_ERR_TIMEO))
net/ipv4/ipip.c
194
t->err_count++;
net/ipv4/ipip.c
196
t->err_count = 1;
net/ipv4/ipip.c
197
t->err_time = jiffies;
net/ipv4/ipip.c
468
struct ip_tunnel *t = netdev_priv(dev);
net/ipv4/ipip.c
476
int err = ip_tunnel_encap_setup(t, &ipencap);
net/ipv4/ipip.c
482
ipip_netlink_parms(data, &p, &t->collect_md, &fwmark);
net/ipv4/ipip.c
491
struct ip_tunnel *t = netdev_priv(dev);
net/ipv4/ipip.c
495
__u32 fwmark = t->fwmark;
net/ipv4/ipip.c
498
int err = ip_tunnel_encap_setup(t, &ipencap);
net/ipv4/ipmr.c
116
static void ipmr_expire_process(struct timer_list *t);
net/ipv4/ipmr.c
2981
unsigned int t = 0, s_t;
net/ipv4/ipmr.c
3000
if (t < s_t)
net/ipv4/ipmr.c
3047
t++;
net/ipv4/ipmr.c
3052
cb->args[0] = t;
net/ipv4/ipmr.c
767
static void ipmr_expire_process(struct timer_list *t)
net/ipv4/ipmr.c
769
struct mr_table *mrt = timer_container_of(mrt, t, ipmr_expire_timer);
net/ipv4/ipmr_base.c
34
void (*expire_func)(struct timer_list *t),
net/ipv4/ipmr_base.c
360
unsigned int t = 0, s_t = cb->args[0];
net/ipv4/ipmr_base.c
376
if (t < s_t)
net/ipv4/ipmr_base.c
384
t++;
net/ipv4/ipmr_base.c
388
cb->args[0] = t;
net/ipv4/netfilter/arp_tables.c
1008
struct xt_table *t;
net/ipv4/netfilter/arp_tables.c
1018
t = xt_find_table_lock(net, NFPROTO_ARP, tmp.name);
net/ipv4/netfilter/arp_tables.c
1019
if (IS_ERR(t)) {
net/ipv4/netfilter/arp_tables.c
1020
ret = PTR_ERR(t);
net/ipv4/netfilter/arp_tables.c
1025
private = t->private;
net/ipv4/netfilter/arp_tables.c
1044
xt_table_unlock(t);
net/ipv4/netfilter/arp_tables.c
1045
module_put(t->me);
net/ipv4/netfilter/arp_tables.c
1067
struct xt_entry_target *t;
net/ipv4/netfilter/arp_tables.c
1069
t = compat_arpt_get_target(e);
net/ipv4/netfilter/arp_tables.c
1070
module_put(t->u.kernel.target->me);
net/ipv4/netfilter/arp_tables.c
1080
struct xt_entry_target *t;
net/ipv4/netfilter/arp_tables.c
1105
t = compat_arpt_get_target(e);
net/ipv4/netfilter/arp_tables.c
1106
target = xt_request_find_target(NFPROTO_ARP, t->u.user.name,
net/ipv4/netfilter/arp_tables.c
1107
t->u.user.revision);
net/ipv4/netfilter/arp_tables.c
1112
t->u.kernel.target = target;
net/ipv4/netfilter/arp_tables.c
1123
module_put(t->u.kernel.target->me);
net/ipv4/netfilter/arp_tables.c
1133
struct xt_entry_target *t;
net/ipv4/netfilter/arp_tables.c
1147
t = compat_arpt_get_target(e);
net/ipv4/netfilter/arp_tables.c
1148
xt_compat_target_from_user(t, dstptr, size);
net/ipv4/netfilter/arp_tables.c
1310
struct xt_entry_target *t;
net/ipv4/netfilter/arp_tables.c
1328
t = arpt_get_target(e);
net/ipv4/netfilter/arp_tables.c
1329
ret = xt_compat_target_to_user(t, dstptr, size);
net/ipv4/netfilter/arp_tables.c
1379
struct xt_table *t;
net/ipv4/netfilter/arp_tables.c
1391
t = xt_find_table_lock(net, NFPROTO_ARP, get.name);
net/ipv4/netfilter/arp_tables.c
1392
if (!IS_ERR(t)) {
net/ipv4/netfilter/arp_tables.c
1393
const struct xt_table_info *private = t->private;
net/ipv4/netfilter/arp_tables.c
1399
t, uptr->entrytable);
net/ipv4/netfilter/arp_tables.c
1404
module_put(t->me);
net/ipv4/netfilter/arp_tables.c
1405
xt_table_unlock(t);
net/ipv4/netfilter/arp_tables.c
1407
ret = PTR_ERR(t);
net/ipv4/netfilter/arp_tables.c
222
const struct xt_entry_target *t;
net/ipv4/netfilter/arp_tables.c
233
t = arpt_get_target_c(e);
net/ipv4/netfilter/arp_tables.c
236
if (!t->u.kernel.target->target) {
net/ipv4/netfilter/arp_tables.c
239
v = ((struct xt_standard_target *)t)->verdict;
net/ipv4/netfilter/arp_tables.c
268
acpar.target = t->u.kernel.target;
net/ipv4/netfilter/arp_tables.c
269
acpar.targinfo = t->data;
net/ipv4/netfilter/arp_tables.c
270
verdict = t->u.kernel.target->target(skb, &acpar);
net/ipv4/netfilter/arp_tables.c
322
const struct xt_standard_target *t
net/ipv4/netfilter/arp_tables.c
334
(strcmp(t->target.u.user.name,
net/ipv4/netfilter/arp_tables.c
336
t->verdict < 0) || visited) {
net/ipv4/netfilter/arp_tables.c
363
int newpos = t->verdict;
net/ipv4/netfilter/arp_tables.c
365
if (strcmp(t->target.u.user.name,
net/ipv4/netfilter/arp_tables.c
390
struct xt_entry_target *t = arpt_get_target(e);
net/ipv4/netfilter/arp_tables.c
395
.target = t->u.kernel.target,
net/ipv4/netfilter/arp_tables.c
396
.targinfo = t->data,
net/ipv4/netfilter/arp_tables.c
401
return xt_check_target(&par, t->u.target_size - sizeof(*t), 0, false);
net/ipv4/netfilter/arp_tables.c
409
struct xt_entry_target *t;
net/ipv4/netfilter/arp_tables.c
416
t = arpt_get_target(e);
net/ipv4/netfilter/arp_tables.c
417
target = xt_request_find_target(NFPROTO_ARP, t->u.user.name,
net/ipv4/netfilter/arp_tables.c
418
t->u.user.revision);
net/ipv4/netfilter/arp_tables.c
423
t->u.kernel.target = target;
net/ipv4/netfilter/arp_tables.c
430
module_put(t->u.kernel.target->me);
net/ipv4/netfilter/arp_tables.c
439
const struct xt_entry_target *t;
net/ipv4/netfilter/arp_tables.c
444
t = arpt_get_target_c(e);
net/ipv4/netfilter/arp_tables.c
445
if (strcmp(t->u.user.name, XT_STANDARD_TARGET) != 0)
net/ipv4/netfilter/arp_tables.c
447
verdict = ((struct xt_standard_target *)t)->verdict;
net/ipv4/netfilter/arp_tables.c
503
struct xt_entry_target *t;
net/ipv4/netfilter/arp_tables.c
505
t = arpt_get_target(e);
net/ipv4/netfilter/arp_tables.c
507
par.target = t->u.kernel.target;
net/ipv4/netfilter/arp_tables.c
508
par.targinfo = t->data;
net/ipv4/netfilter/arp_tables.c
600
static void get_counters(const struct xt_table_info *t,
net/ipv4/netfilter/arp_tables.c
611
xt_entry_foreach(iter, t->entries, t->size) {
net/ipv4/netfilter/arp_tables.c
630
static void get_old_counters(const struct xt_table_info *t,
net/ipv4/netfilter/arp_tables.c
638
xt_entry_foreach(iter, t->entries, t->size) {
net/ipv4/netfilter/arp_tables.c
690
const struct xt_entry_target *t;
net/ipv4/netfilter/arp_tables.c
705
t = arpt_get_target_c(e);
net/ipv4/netfilter/arp_tables.c
706
if (xt_target_to_user(t, userptr + off + e->target_offset)) {
net/ipv4/netfilter/arp_tables.c
740
const struct xt_entry_target *t;
net/ipv4/netfilter/arp_tables.c
747
t = arpt_get_target_c(e);
net/ipv4/netfilter/arp_tables.c
748
off += xt_compat_target_offset(t->u.kernel.target);
net/ipv4/netfilter/arp_tables.c
794
struct xt_table *t;
net/ipv4/netfilter/arp_tables.c
808
t = xt_request_find_table_lock(net, NFPROTO_ARP, name);
net/ipv4/netfilter/arp_tables.c
809
if (!IS_ERR(t)) {
net/ipv4/netfilter/arp_tables.c
811
const struct xt_table_info *private = t->private;
net/ipv4/netfilter/arp_tables.c
822
info.valid_hooks = t->valid_hooks;
net/ipv4/netfilter/arp_tables.c
835
xt_table_unlock(t);
net/ipv4/netfilter/arp_tables.c
836
module_put(t->me);
net/ipv4/netfilter/arp_tables.c
838
ret = PTR_ERR(t);
net/ipv4/netfilter/arp_tables.c
851
struct xt_table *t;
net/ipv4/netfilter/arp_tables.c
862
t = xt_find_table_lock(net, NFPROTO_ARP, get.name);
net/ipv4/netfilter/arp_tables.c
863
if (!IS_ERR(t)) {
net/ipv4/netfilter/arp_tables.c
864
const struct xt_table_info *private = t->private;
net/ipv4/netfilter/arp_tables.c
868
t, uptr->entrytable);
net/ipv4/netfilter/arp_tables.c
872
module_put(t->me);
net/ipv4/netfilter/arp_tables.c
873
xt_table_unlock(t);
net/ipv4/netfilter/arp_tables.c
875
ret = PTR_ERR(t);
net/ipv4/netfilter/arp_tables.c
887
struct xt_table *t;
net/ipv4/netfilter/arp_tables.c
900
t = xt_request_find_table_lock(net, NFPROTO_ARP, name);
net/ipv4/netfilter/arp_tables.c
901
if (IS_ERR(t)) {
net/ipv4/netfilter/arp_tables.c
902
ret = PTR_ERR(t);
net/ipv4/netfilter/arp_tables.c
907
if (valid_hooks != t->valid_hooks) {
net/ipv4/netfilter/arp_tables.c
912
oldinfo = xt_replace_table(t, num_counters, newinfo, &ret);
net/ipv4/netfilter/arp_tables.c
919
module_put(t->me);
net/ipv4/netfilter/arp_tables.c
922
module_put(t->me);
net/ipv4/netfilter/arp_tables.c
924
xt_table_unlock(t);
net/ipv4/netfilter/arp_tables.c
943
module_put(t->me);
net/ipv4/netfilter/arp_tables.c
944
xt_table_unlock(t);
net/ipv4/netfilter/ip_tables.c
1008
struct xt_table *t;
net/ipv4/netfilter/ip_tables.c
1018
t = xt_find_table_lock(net, AF_INET, get.name);
net/ipv4/netfilter/ip_tables.c
1019
if (!IS_ERR(t)) {
net/ipv4/netfilter/ip_tables.c
1020
const struct xt_table_info *private = t->private;
net/ipv4/netfilter/ip_tables.c
1023
t, uptr->entrytable);
net/ipv4/netfilter/ip_tables.c
1027
module_put(t->me);
net/ipv4/netfilter/ip_tables.c
1028
xt_table_unlock(t);
net/ipv4/netfilter/ip_tables.c
1030
ret = PTR_ERR(t);
net/ipv4/netfilter/ip_tables.c
1041
struct xt_table *t;
net/ipv4/netfilter/ip_tables.c
1052
t = xt_request_find_table_lock(net, AF_INET, name);
net/ipv4/netfilter/ip_tables.c
1053
if (IS_ERR(t)) {
net/ipv4/netfilter/ip_tables.c
1054
ret = PTR_ERR(t);
net/ipv4/netfilter/ip_tables.c
1059
if (valid_hooks != t->valid_hooks) {
net/ipv4/netfilter/ip_tables.c
1064
oldinfo = xt_replace_table(t, num_counters, newinfo, &ret);
net/ipv4/netfilter/ip_tables.c
1071
module_put(t->me);
net/ipv4/netfilter/ip_tables.c
1074
module_put(t->me);
net/ipv4/netfilter/ip_tables.c
1076
xt_table_unlock(t);
net/ipv4/netfilter/ip_tables.c
1094
module_put(t->me);
net/ipv4/netfilter/ip_tables.c
1095
xt_table_unlock(t);
net/ipv4/netfilter/ip_tables.c
1161
struct xt_table *t;
net/ipv4/netfilter/ip_tables.c
1171
t = xt_find_table_lock(net, AF_INET, tmp.name);
net/ipv4/netfilter/ip_tables.c
1172
if (IS_ERR(t)) {
net/ipv4/netfilter/ip_tables.c
1173
ret = PTR_ERR(t);
net/ipv4/netfilter/ip_tables.c
1178
private = t->private;
net/ipv4/netfilter/ip_tables.c
1196
xt_table_unlock(t);
net/ipv4/netfilter/ip_tables.c
1197
module_put(t->me);
net/ipv4/netfilter/ip_tables.c
1222
struct xt_entry_target *t;
net/ipv4/netfilter/ip_tables.c
1245
t = ipt_get_target(e);
net/ipv4/netfilter/ip_tables.c
1246
ret = xt_compat_target_to_user(t, dstptr, size);
net/ipv4/netfilter/ip_tables.c
1275
struct xt_entry_target *t;
net/ipv4/netfilter/ip_tables.c
1281
t = compat_ipt_get_target(e);
net/ipv4/netfilter/ip_tables.c
1282
module_put(t->u.kernel.target->me);
net/ipv4/netfilter/ip_tables.c
1293
struct xt_entry_target *t;
net/ipv4/netfilter/ip_tables.c
1326
t = compat_ipt_get_target(e);
net/ipv4/netfilter/ip_tables.c
1327
target = xt_request_find_target(NFPROTO_IPV4, t->u.user.name,
net/ipv4/netfilter/ip_tables.c
1328
t->u.user.revision);
net/ipv4/netfilter/ip_tables.c
1333
t->u.kernel.target = target;
net/ipv4/netfilter/ip_tables.c
1344
module_put(t->u.kernel.target->me);
net/ipv4/netfilter/ip_tables.c
1359
struct xt_entry_target *t;
net/ipv4/netfilter/ip_tables.c
1377
t = compat_ipt_get_target(e);
net/ipv4/netfilter/ip_tables.c
1378
xt_compat_target_from_user(t, dstptr, size);
net/ipv4/netfilter/ip_tables.c
1584
struct xt_table *t;
net/ipv4/netfilter/ip_tables.c
1598
t = xt_find_table_lock(net, AF_INET, get.name);
net/ipv4/netfilter/ip_tables.c
1599
if (!IS_ERR(t)) {
net/ipv4/netfilter/ip_tables.c
160
const struct xt_standard_target *t = (void *)ipt_get_target_c(s);
net/ipv4/netfilter/ip_tables.c
1600
const struct xt_table_info *private = t->private;
net/ipv4/netfilter/ip_tables.c
1605
t, uptr->entrytable);
net/ipv4/netfilter/ip_tables.c
1610
module_put(t->me);
net/ipv4/netfilter/ip_tables.c
1611
xt_table_unlock(t);
net/ipv4/netfilter/ip_tables.c
1613
ret = PTR_ERR(t);
net/ipv4/netfilter/ip_tables.c
162
if (strcmp(t->target.u.kernel.target->name, XT_ERROR_TARGET) == 0) {
net/ipv4/netfilter/ip_tables.c
164
*chainname = t->target.data;
net/ipv4/netfilter/ip_tables.c
170
strcmp(t->target.u.kernel.target->name,
net/ipv4/netfilter/ip_tables.c
172
t->verdict < 0) {
net/ipv4/netfilter/ip_tables.c
278
const struct xt_entry_target *t;
net/ipv4/netfilter/ip_tables.c
300
t = ipt_get_target_c(e);
net/ipv4/netfilter/ip_tables.c
301
WARN_ON(!t->u.kernel.target);
net/ipv4/netfilter/ip_tables.c
310
if (!t->u.kernel.target->target) {
net/ipv4/netfilter/ip_tables.c
313
v = ((struct xt_standard_target *)t)->verdict;
net/ipv4/netfilter/ip_tables.c
342
acpar.target = t->u.kernel.target;
net/ipv4/netfilter/ip_tables.c
343
acpar.targinfo = t->data;
net/ipv4/netfilter/ip_tables.c
345
verdict = t->u.kernel.target->target(skb, &acpar);
net/ipv4/netfilter/ip_tables.c
386
const struct xt_standard_target *t
net/ipv4/netfilter/ip_tables.c
397
(strcmp(t->target.u.user.name,
net/ipv4/netfilter/ip_tables.c
399
t->verdict < 0) || visited) {
net/ipv4/netfilter/ip_tables.c
425
int newpos = t->verdict;
net/ipv4/netfilter/ip_tables.c
427
if (strcmp(t->target.u.user.name,
net/ipv4/netfilter/ip_tables.c
499
struct xt_entry_target *t = ipt_get_target(e);
net/ipv4/netfilter/ip_tables.c
504
.target = t->u.kernel.target,
net/ipv4/netfilter/ip_tables.c
505
.targinfo = t->data,
net/ipv4/netfilter/ip_tables.c
510
return xt_check_target(&par, t->u.target_size - sizeof(*t),
net/ipv4/netfilter/ip_tables.c
519
struct xt_entry_target *t;
net/ipv4/netfilter/ip_tables.c
543
t = ipt_get_target(e);
net/ipv4/netfilter/ip_tables.c
544
target = xt_request_find_target(NFPROTO_IPV4, t->u.user.name,
net/ipv4/netfilter/ip_tables.c
545
t->u.user.revision);
net/ipv4/netfilter/ip_tables.c
550
t->u.kernel.target = target;
net/ipv4/netfilter/ip_tables.c
558
module_put(t->u.kernel.target->me);
net/ipv4/netfilter/ip_tables.c
573
const struct xt_entry_target *t;
net/ipv4/netfilter/ip_tables.c
578
t = ipt_get_target_c(e);
net/ipv4/netfilter/ip_tables.c
579
if (strcmp(t->u.user.name, XT_STANDARD_TARGET) != 0)
net/ipv4/netfilter/ip_tables.c
581
verdict = ((struct xt_standard_target *)t)->verdict;
net/ipv4/netfilter/ip_tables.c
639
struct xt_entry_target *t;
net/ipv4/netfilter/ip_tables.c
645
t = ipt_get_target(e);
net/ipv4/netfilter/ip_tables.c
648
par.target = t->u.kernel.target;
net/ipv4/netfilter/ip_tables.c
649
par.targinfo = t->data;
net/ipv4/netfilter/ip_tables.c
739
get_counters(const struct xt_table_info *t,
net/ipv4/netfilter/ip_tables.c
750
xt_entry_foreach(iter, t->entries, t->size) {
net/ipv4/netfilter/ip_tables.c
769
static void get_old_counters(const struct xt_table_info *t,
net/ipv4/netfilter/ip_tables.c
777
xt_entry_foreach(iter, t->entries, t->size) {
net/ipv4/netfilter/ip_tables.c
832
const struct xt_entry_target *t;
net/ipv4/netfilter/ip_tables.c
858
t = ipt_get_target_c(e);
net/ipv4/netfilter/ip_tables.c
859
if (xt_target_to_user(t, userptr + off + e->target_offset)) {
net/ipv4/netfilter/ip_tables.c
894
const struct xt_entry_target *t;
net/ipv4/netfilter/ip_tables.c
902
t = ipt_get_target_c(e);
net/ipv4/netfilter/ip_tables.c
903
off += xt_compat_target_offset(t->u.kernel.target);
net/ipv4/netfilter/ip_tables.c
949
struct xt_table *t;
net/ipv4/netfilter/ip_tables.c
963
t = xt_request_find_table_lock(net, AF_INET, name);
net/ipv4/netfilter/ip_tables.c
964
if (!IS_ERR(t)) {
net/ipv4/netfilter/ip_tables.c
966
const struct xt_table_info *private = t->private;
net/ipv4/netfilter/ip_tables.c
977
info.valid_hooks = t->valid_hooks;
net/ipv4/netfilter/ip_tables.c
991
xt_table_unlock(t);
net/ipv4/netfilter/ip_tables.c
992
module_put(t->me);
net/ipv4/netfilter/ip_tables.c
994
ret = PTR_ERR(t);
net/ipv4/netfilter/nf_nat_pptp.c
49
struct nf_conntrack_tuple t = {};
net/ipv4/netfilter/nf_nat_pptp.c
66
t.src.l3num = AF_INET;
net/ipv4/netfilter/nf_nat_pptp.c
67
t.src.u3.ip = master->tuplehash[!exp->dir].tuple.src.u3.ip;
net/ipv4/netfilter/nf_nat_pptp.c
68
t.src.u.gre.key = ct_pptp_info->pac_call_id;
net/ipv4/netfilter/nf_nat_pptp.c
69
t.dst.u3.ip = master->tuplehash[!exp->dir].tuple.dst.u3.ip;
net/ipv4/netfilter/nf_nat_pptp.c
70
t.dst.u.gre.key = ct_pptp_info->pns_call_id;
net/ipv4/netfilter/nf_nat_pptp.c
71
t.dst.protonum = IPPROTO_GRE;
net/ipv4/netfilter/nf_nat_pptp.c
75
t.src.l3num = AF_INET;
net/ipv4/netfilter/nf_nat_pptp.c
76
t.src.u3.ip = master->tuplehash[!exp->dir].tuple.src.u3.ip;
net/ipv4/netfilter/nf_nat_pptp.c
77
t.src.u.gre.key = nat_pptp_info->pns_call_id;
net/ipv4/netfilter/nf_nat_pptp.c
78
t.dst.u3.ip = master->tuplehash[!exp->dir].tuple.dst.u3.ip;
net/ipv4/netfilter/nf_nat_pptp.c
79
t.dst.u.gre.key = nat_pptp_info->pac_call_id;
net/ipv4/netfilter/nf_nat_pptp.c
80
t.dst.protonum = IPPROTO_GRE;
net/ipv4/netfilter/nf_nat_pptp.c
84
nf_ct_dump_tuple_ip(&t);
net/ipv4/netfilter/nf_nat_pptp.c
85
other_exp = nf_ct_expect_find_get(net, nf_ct_zone(ct), &t);
net/ipv4/tcp_bbr.c
694
u32 t;
net/ipv4/tcp_bbr.c
746
t = div_u64(tp->delivered_mstamp, USEC_PER_MSEC) - bbr->lt_last_stamp;
net/ipv4/tcp_bbr.c
747
if ((s32)t < 1)
net/ipv4/tcp_bbr.c
750
if (t >= ~0U / USEC_PER_MSEC) {
net/ipv4/tcp_bbr.c
754
t *= USEC_PER_MSEC;
net/ipv4/tcp_bbr.c
756
do_div(bw, t);
net/ipv4/tcp_cubic.c
217
u64 offs, t;
net/ipv4/tcp_cubic.c
267
t = (s32)(tcp_jiffies32 - ca->epoch_start);
net/ipv4/tcp_cubic.c
268
t += usecs_to_jiffies(ca->delay_min);
net/ipv4/tcp_cubic.c
270
t <<= BICTCP_HZ;
net/ipv4/tcp_cubic.c
271
do_div(t, HZ);
net/ipv4/tcp_cubic.c
273
if (t < ca->bic_K) /* t - K */
net/ipv4/tcp_cubic.c
274
offs = ca->bic_K - t;
net/ipv4/tcp_cubic.c
276
offs = t - ca->bic_K;
net/ipv4/tcp_cubic.c
280
if (t < ca->bic_K) /* below origin*/
net/ipv4/tcp_illinois.c
122
u64 t = ca->sum_rtt;
net/ipv4/tcp_illinois.c
124
do_div(t, ca->cnt_rtt);
net/ipv4/tcp_illinois.c
125
return t - ca->base_rtt;
net/ipv4/tcp_illinois.c
319
u64 t = ca->sum_rtt;
net/ipv4/tcp_illinois.c
321
do_div(t, info->vegas.tcpv_rttcnt);
net/ipv4/tcp_illinois.c
322
info->vegas.tcpv_rtt = t;
net/ipv4/tcp_timer.c
360
static void tcp_delack_timer(struct timer_list *t)
net/ipv4/tcp_timer.c
363
timer_container_of(icsk, t, icsk_delack_timer);
net/ipv4/tcp_timer.c
729
static void tcp_write_timer(struct timer_list *t)
net/ipv4/tcp_timer.c
731
struct sock *sk = timer_container_of(sk, t, tcp_retransmit_timer);
net/ipv4/tcp_timer.c
779
static void tcp_keepalive_timer(struct timer_list *t)
net/ipv4/tcp_timer.c
782
timer_container_of(icsk, t, icsk_keepalive_timer);
net/ipv4/tunnel4.c
34
struct xfrm_tunnel *t;
net/ipv4/tunnel4.c
42
(t = rcu_dereference_protected(*pprev,
net/ipv4/tunnel4.c
44
pprev = &t->next) {
net/ipv4/tunnel4.c
45
if (t->priority > priority)
net/ipv4/tunnel4.c
47
if (t->priority == priority)
net/ipv4/tunnel4.c
66
struct xfrm_tunnel *t;
net/ipv4/tunnel4.c
72
(t = rcu_dereference_protected(*pprev,
net/ipv4/tunnel4.c
74
pprev = &t->next) {
net/ipv4/tunnel4.c
75
if (t == handler) {
net/ipv4/xfrm4_protocol.c
221
struct xfrm4_protocol *t;
net/ipv4/xfrm4_protocol.c
236
(t = rcu_dereference_protected(*pprev,
net/ipv4/xfrm4_protocol.c
238
pprev = &t->next) {
net/ipv4/xfrm4_protocol.c
239
if (t->priority < priority)
net/ipv4/xfrm4_protocol.c
241
if (t->priority == priority)
net/ipv4/xfrm4_protocol.c
268
struct xfrm4_protocol *t;
net/ipv4/xfrm4_protocol.c
277
(t = rcu_dereference_protected(*pprev,
net/ipv4/xfrm4_protocol.c
279
pprev = &t->next) {
net/ipv4/xfrm4_protocol.c
280
if (t == handler) {
net/ipv6/addrconf.c
173
static void addrconf_rs_timer(struct timer_list *t);
net/ipv6/addrconf.c
4018
static void addrconf_rs_timer(struct timer_list *t)
net/ipv6/addrconf.c
4020
struct inet6_dev *idev = timer_container_of(idev, t, rs_timer);
net/ipv6/ip6_fib.c
2461
static void fib6_gc_timer_cb(struct timer_list *t)
net/ipv6/ip6_fib.c
2463
struct net *arg = timer_container_of(arg, t, ipv6.ip6_fib_timer);
net/ipv6/ip6_fib.c
73
static void fib6_gc_timer_cb(struct timer_list *t);
net/ipv6/ip6_gre.c
1023
if (ipv6_addr_equal(&t->parms.raddr, &ipv6_hdr(skb)->saddr))
net/ipv6/ip6_gre.c
1030
memcpy(&fl6, &t->fl.u.ip6, sizeof(fl6));
net/ipv6/ip6_gre.c
1034
if (t->parms.erspan_ver == 1) {
net/ipv6/ip6_gre.c
1035
erspan_build_header(skb, ntohl(t->parms.o_key),
net/ipv6/ip6_gre.c
1036
t->parms.index,
net/ipv6/ip6_gre.c
1039
} else if (t->parms.erspan_ver == 2) {
net/ipv6/ip6_gre.c
1040
erspan_build_header_v2(skb, ntohl(t->parms.o_key),
net/ipv6/ip6_gre.c
1041
t->parms.dir,
net/ipv6/ip6_gre.c
1042
t->parms.hwid,
net/ipv6/ip6_gre.c
1049
fl6.daddr = t->parms.raddr;
net/ipv6/ip6_gre.c
1055
htonl(atomic_fetch_inc(&t->o_seqno)));
net/ipv6/ip6_gre.c
1058
if (!t->parms.collect_md && dst) {
net/ipv6/ip6_gre.c
1087
static void ip6gre_tnl_link_config_common(struct ip6_tnl *t)
net/ipv6/ip6_gre.c
1089
struct net_device *dev = t->dev;
net/ipv6/ip6_gre.c
1090
struct __ip6_tnl_parm *p = &t->parms;
net/ipv6/ip6_gre.c
1091
struct flowi6 *fl6 = &t->fl.u.ip6;
net/ipv6/ip6_gre.c
1104
fl6->fl6_gre_key = t->parms.o_key;
net/ipv6/ip6_gre.c
1112
p->flags |= ip6_tnl_get_cap(t, &p->laddr, &p->raddr);
net/ipv6/ip6_gre.c
1121
static void ip6gre_tnl_link_config_route(struct ip6_tnl *t, int set_mtu,
net/ipv6/ip6_gre.c
1124
const struct __ip6_tnl_parm *p = &t->parms;
net/ipv6/ip6_gre.c
1125
struct net_device *dev = t->dev;
net/ipv6/ip6_gre.c
1131
struct rt6_info *rt = rt6_lookup(t->net,
net/ipv6/ip6_gre.c
114
static bool ip6gre_tunnel_match(struct ip6_tnl *t, int dev_type, int link,
net/ipv6/ip6_gre.c
1142
if (t->dev->header_ops)
net/ipv6/ip6_gre.c
1150
if (!(t->parms.flags & IP6_TNL_F_IGN_ENCAP_LIMIT))
net/ipv6/ip6_gre.c
1181
static void ip6gre_tnl_link_config(struct ip6_tnl *t, int set_mtu)
net/ipv6/ip6_gre.c
1183
ip6gre_tnl_link_config_common(t);
net/ipv6/ip6_gre.c
1184
ip6gre_tnl_link_config_route(t, set_mtu, ip6gre_calc_hlen(t));
net/ipv6/ip6_gre.c
1187
static void ip6gre_tnl_copy_tnl_parm(struct ip6_tnl *t,
net/ipv6/ip6_gre.c
119
if (t->dev->type != ARPHRD_IP6GRE &&
net/ipv6/ip6_gre.c
1190
t->parms.laddr = p->laddr;
net/ipv6/ip6_gre.c
1191
t->parms.raddr = p->raddr;
net/ipv6/ip6_gre.c
1192
t->parms.flags = p->flags;
net/ipv6/ip6_gre.c
1193
t->parms.hop_limit = p->hop_limit;
net/ipv6/ip6_gre.c
1194
t->parms.encap_limit = p->encap_limit;
net/ipv6/ip6_gre.c
1195
t->parms.flowinfo = p->flowinfo;
net/ipv6/ip6_gre.c
1196
t->parms.link = p->link;
net/ipv6/ip6_gre.c
1197
t->parms.proto = p->proto;
net/ipv6/ip6_gre.c
1198
t->parms.i_key = p->i_key;
net/ipv6/ip6_gre.c
1199
t->parms.o_key = p->o_key;
net/ipv6/ip6_gre.c
120
t->dev->type != dev_type)
net/ipv6/ip6_gre.c
1200
ip_tunnel_flags_copy(t->parms.i_flags, p->i_flags);
net/ipv6/ip6_gre.c
1201
ip_tunnel_flags_copy(t->parms.o_flags, p->o_flags);
net/ipv6/ip6_gre.c
1202
t->parms.fwmark = p->fwmark;
net/ipv6/ip6_gre.c
1203
t->parms.erspan_ver = p->erspan_ver;
net/ipv6/ip6_gre.c
1204
t->parms.index = p->index;
net/ipv6/ip6_gre.c
1205
t->parms.dir = p->dir;
net/ipv6/ip6_gre.c
1206
t->parms.hwid = p->hwid;
net/ipv6/ip6_gre.c
1207
dst_cache_reset(&t->dst_cache);
net/ipv6/ip6_gre.c
1210
static int ip6gre_tnl_change(struct ip6_tnl *t, const struct __ip6_tnl_parm *p,
net/ipv6/ip6_gre.c
1213
ip6gre_tnl_copy_tnl_parm(t, p);
net/ipv6/ip6_gre.c
1214
ip6gre_tnl_link_config(t, set_mtu);
net/ipv6/ip6_gre.c
123
if (t->parms.link != link)
net/ipv6/ip6_gre.c
125
if (t->dev->type != dev_type)
net/ipv6/ip6_gre.c
1260
struct ip6_tnl *t = netdev_priv(dev);
net/ipv6/ip6_gre.c
1261
struct net *net = t->net;
net/ipv6/ip6_gre.c
1274
t = ip6gre_tunnel_locate(net, &p1, 0);
net/ipv6/ip6_gre.c
1275
if (!t)
net/ipv6/ip6_gre.c
1276
t = netdev_priv(dev);
net/ipv6/ip6_gre.c
1279
ip6gre_tnl_parm_to_user(&p, &t->parms);
net/ipv6/ip6_gre.c
128
*ret = t;
net/ipv6/ip6_gre.c
1304
t = ip6gre_tunnel_locate(net, &p1, cmd == SIOCADDTUNNEL);
net/ipv6/ip6_gre.c
1307
if (t) {
net/ipv6/ip6_gre.c
1308
if (t->dev != dev) {
net/ipv6/ip6_gre.c
1313
t = netdev_priv(dev);
net/ipv6/ip6_gre.c
1315
ip6gre_tunnel_unlink(ign, t);
net/ipv6/ip6_gre.c
1317
ip6gre_tnl_change(t, &p1, 1);
net/ipv6/ip6_gre.c
1318
ip6gre_tunnel_link(ign, t);
net/ipv6/ip6_gre.c
1323
if (t) {
net/ipv6/ip6_gre.c
1327
ip6gre_tnl_parm_to_user(&p, &t->parms);
net/ipv6/ip6_gre.c
133
*ret = t;
net/ipv6/ip6_gre.c
1345
t = ip6gre_tunnel_locate(net, &p1, 0);
net/ipv6/ip6_gre.c
1346
if (!t)
net/ipv6/ip6_gre.c
1349
if (t == netdev_priv(ign->fb_tunnel_dev))
net/ipv6/ip6_gre.c
1351
dev = t->dev;
net/ipv6/ip6_gre.c
1369
struct ip6_tnl *t = netdev_priv(dev);
net/ipv6/ip6_gre.c
1374
needed = t->hlen + sizeof(*ipv6h);
net/ipv6/ip6_gre.c
1382
t->fl.u.ip6.flowlabel,
net/ipv6/ip6_gre.c
1383
true, &t->fl.u.ip6));
net/ipv6/ip6_gre.c
1384
ipv6h->hop_limit = t->parms.hop_limit;
net/ipv6/ip6_gre.c
1386
ipv6h->saddr = t->parms.laddr;
net/ipv6/ip6_gre.c
1387
ipv6h->daddr = t->parms.raddr;
net/ipv6/ip6_gre.c
1390
p[0] = ip_tunnel_flags_to_be16(t->parms.o_flags);
net/ipv6/ip6_gre.c
1402
return t->hlen;
net/ipv6/ip6_gre.c
1404
return -t->hlen;
net/ipv6/ip6_gre.c
1422
struct ip6_tnl *t = netdev_priv(dev);
net/ipv6/ip6_gre.c
1424
gro_cells_destroy(&t->gro_cells);
net/ipv6/ip6_gre.c
1425
dst_cache_destroy(&t->dst_cache);
net/ipv6/ip6_gre.c
148
struct ip6_tnl *t, *cand = NULL;
net/ipv6/ip6_gre.c
1567
struct ip6_tnl *t;
net/ipv6/ip6_gre.c
1569
t = rtnl_net_dereference(net, ign->tunnels[prio][h]);
net/ipv6/ip6_gre.c
157
for_each_ip_tunnel_rcu(t, ign->tunnels_r_l[h0 ^ h1]) {
net/ipv6/ip6_gre.c
1571
while (t) {
net/ipv6/ip6_gre.c
1575
if (!net_eq(dev_net(t->dev), net))
net/ipv6/ip6_gre.c
1576
unregister_netdevice_queue(t->dev, head);
net/ipv6/ip6_gre.c
1578
t = rtnl_net_dereference(net, t->next);
net/ipv6/ip6_gre.c
158
if (!ipv6_addr_equal(local, &t->parms.laddr) ||
net/ipv6/ip6_gre.c
159
!ipv6_addr_equal(remote, &t->parms.raddr) ||
net/ipv6/ip6_gre.c
160
key != t->parms.i_key ||
net/ipv6/ip6_gre.c
161
!(t->dev->flags & IFF_UP))
net/ipv6/ip6_gre.c
164
if (ip6gre_tunnel_match(t, dev_type, link, &cand_score, &cand))
net/ipv6/ip6_gre.c
168
for_each_ip_tunnel_rcu(t, ign->tunnels_r[h0 ^ h1]) {
net/ipv6/ip6_gre.c
169
if (!ipv6_addr_equal(remote, &t->parms.raddr) ||
net/ipv6/ip6_gre.c
170
key != t->parms.i_key ||
net/ipv6/ip6_gre.c
171
!(t->dev->flags & IFF_UP))
net/ipv6/ip6_gre.c
174
if (ip6gre_tunnel_match(t, dev_type, link, &cand_score, &cand))
net/ipv6/ip6_gre.c
178
for_each_ip_tunnel_rcu(t, ign->tunnels_l[h1]) {
net/ipv6/ip6_gre.c
179
if ((!ipv6_addr_equal(local, &t->parms.laddr) &&
net/ipv6/ip6_gre.c
180
(!ipv6_addr_equal(local, &t->parms.raddr) ||
net/ipv6/ip6_gre.c
182
key != t->parms.i_key ||
net/ipv6/ip6_gre.c
183
!(t->dev->flags & IFF_UP))
net/ipv6/ip6_gre.c
186
if (ip6gre_tunnel_match(t, dev_type, link, &cand_score, &cand))
net/ipv6/ip6_gre.c
190
for_each_ip_tunnel_rcu(t, ign->tunnels_wc[h1]) {
net/ipv6/ip6_gre.c
191
if (t->parms.i_key != key ||
net/ipv6/ip6_gre.c
192
!(t->dev->flags & IFF_UP))
net/ipv6/ip6_gre.c
195
if (ip6gre_tunnel_match(t, dev_type, link, &cand_score, &cand))
net/ipv6/ip6_gre.c
2012
struct ip6_tnl *t, *nt = netdev_priv(dev);
net/ipv6/ip6_gre.c
2029
t = ip6gre_tunnel_locate(net, p_p, 0);
net/ipv6/ip6_gre.c
2031
if (t) {
net/ipv6/ip6_gre.c
2032
if (t->dev != dev)
net/ipv6/ip6_gre.c
2035
t = nt;
net/ipv6/ip6_gre.c
2038
return t;
net/ipv6/ip6_gre.c
204
t = rcu_dereference(ign->collect_md_tun_erspan);
net/ipv6/ip6_gre.c
2045
struct ip6_tnl *t = netdev_priv(dev);
net/ipv6/ip6_gre.c
2046
struct ip6gre_net *ign = net_generic(t->net, ip6gre_net_id);
net/ipv6/ip6_gre.c
2049
t = ip6gre_changelink_common(dev, tb, data, &p, extack);
net/ipv6/ip6_gre.c
2050
if (IS_ERR(t))
net/ipv6/ip6_gre.c
2051
return PTR_ERR(t);
net/ipv6/ip6_gre.c
2053
ip6gre_tunnel_unlink_md(ign, t);
net/ipv6/ip6_gre.c
2054
ip6gre_tunnel_unlink(ign, t);
net/ipv6/ip6_gre.c
2055
ip6gre_tnl_change(t, &p, !tb[IFLA_MTU]);
net/ipv6/ip6_gre.c
2056
ip6gre_tunnel_link_md(ign, t);
net/ipv6/ip6_gre.c
2057
ip6gre_tunnel_link(ign, t);
net/ipv6/ip6_gre.c
206
t = rcu_dereference(ign->collect_md_tun);
net/ipv6/ip6_gre.c
208
if (t && t->dev->flags & IFF_UP)
net/ipv6/ip6_gre.c
209
return t;
net/ipv6/ip6_gre.c
2114
struct ip6_tnl *t = netdev_priv(dev);
net/ipv6/ip6_gre.c
2115
struct __ip6_tnl_parm *p = &t->parms;
net/ipv6/ip6_gre.c
2155
t->encap.type) ||
net/ipv6/ip6_gre.c
2157
t->encap.sport) ||
net/ipv6/ip6_gre.c
2159
t->encap.dport) ||
net/ipv6/ip6_gre.c
2161
t->encap.flags))
net/ipv6/ip6_gre.c
2246
static void ip6erspan_tnl_link_config(struct ip6_tnl *t, int set_mtu)
net/ipv6/ip6_gre.c
2248
ip6gre_tnl_link_config_common(t);
net/ipv6/ip6_gre.c
2249
ip6gre_tnl_link_config_route(t, set_mtu, ip6erspan_calc_hlen(t));
net/ipv6/ip6_gre.c
2252
static int ip6erspan_tnl_change(struct ip6_tnl *t,
net/ipv6/ip6_gre.c
2255
ip6gre_tnl_copy_tnl_parm(t, p);
net/ipv6/ip6_gre.c
2256
ip6erspan_tnl_link_config(t, set_mtu);
net/ipv6/ip6_gre.c
2266
struct ip6_tnl *t;
net/ipv6/ip6_gre.c
2268
t = ip6gre_changelink_common(dev, tb, data, &p, extack);
net/ipv6/ip6_gre.c
2269
if (IS_ERR(t))
net/ipv6/ip6_gre.c
2270
return PTR_ERR(t);
net/ipv6/ip6_gre.c
2273
ip6gre_tunnel_unlink_md(ign, t);
net/ipv6/ip6_gre.c
2274
ip6gre_tunnel_unlink(ign, t);
net/ipv6/ip6_gre.c
2275
ip6erspan_tnl_change(t, &p, !tb[IFLA_MTU]);
net/ipv6/ip6_gre.c
2276
ip6erspan_tunnel_link_md(ign, t);
net/ipv6/ip6_gre.c
2277
ip6gre_tunnel_link(ign, t);
net/ipv6/ip6_gre.c
236
static void ip6gre_tunnel_link_md(struct ip6gre_net *ign, struct ip6_tnl *t)
net/ipv6/ip6_gre.c
238
if (t->parms.collect_md)
net/ipv6/ip6_gre.c
239
rcu_assign_pointer(ign->collect_md_tun, t);
net/ipv6/ip6_gre.c
242
static void ip6erspan_tunnel_link_md(struct ip6gre_net *ign, struct ip6_tnl *t)
net/ipv6/ip6_gre.c
244
if (t->parms.collect_md)
net/ipv6/ip6_gre.c
245
rcu_assign_pointer(ign->collect_md_tun_erspan, t);
net/ipv6/ip6_gre.c
248
static void ip6gre_tunnel_unlink_md(struct ip6gre_net *ign, struct ip6_tnl *t)
net/ipv6/ip6_gre.c
250
if (t->parms.collect_md)
net/ipv6/ip6_gre.c
255
struct ip6_tnl *t)
net/ipv6/ip6_gre.c
257
if (t->parms.collect_md)
net/ipv6/ip6_gre.c
262
const struct ip6_tnl *t)
net/ipv6/ip6_gre.c
264
return __ip6gre_bucket(ign, &t->parms);
net/ipv6/ip6_gre.c
267
static void ip6gre_tunnel_link(struct ip6gre_net *ign, struct ip6_tnl *t)
net/ipv6/ip6_gre.c
269
struct ip6_tnl __rcu **tp = ip6gre_bucket(ign, t);
net/ipv6/ip6_gre.c
271
rcu_assign_pointer(t->next, rtnl_dereference(*tp));
net/ipv6/ip6_gre.c
272
rcu_assign_pointer(*tp, t);
net/ipv6/ip6_gre.c
275
static void ip6gre_tunnel_unlink(struct ip6gre_net *ign, struct ip6_tnl *t)
net/ipv6/ip6_gre.c
280
for (tp = ip6gre_bucket(ign, t);
net/ipv6/ip6_gre.c
283
if (t == iter) {
net/ipv6/ip6_gre.c
284
rcu_assign_pointer(*tp, t->next);
net/ipv6/ip6_gre.c
298
struct ip6_tnl *t;
net/ipv6/ip6_gre.c
303
(t = rtnl_dereference(*tp)) != NULL;
net/ipv6/ip6_gre.c
304
tp = &t->next)
net/ipv6/ip6_gre.c
305
if (ipv6_addr_equal(local, &t->parms.laddr) &&
net/ipv6/ip6_gre.c
306
ipv6_addr_equal(remote, &t->parms.raddr) &&
net/ipv6/ip6_gre.c
307
key == t->parms.i_key &&
net/ipv6/ip6_gre.c
308
link == t->parms.link &&
net/ipv6/ip6_gre.c
309
type == t->dev->type)
net/ipv6/ip6_gre.c
312
return t;
net/ipv6/ip6_gre.c
318
struct ip6_tnl *t, *nt;
net/ipv6/ip6_gre.c
323
t = ip6gre_tunnel_find(net, parms, ARPHRD_IP6GRE);
net/ipv6/ip6_gre.c
324
if (t && create)
net/ipv6/ip6_gre.c
326
if (t || !create)
net/ipv6/ip6_gre.c
327
return t;
net/ipv6/ip6_gre.c
336
dev = alloc_netdev(sizeof(*t), name, NET_NAME_UNKNOWN,
net/ipv6/ip6_gre.c
364
struct ip6_tnl *t = netdev_priv(dev);
net/ipv6/ip6_gre.c
365
struct ip6gre_net *ign = net_generic(t->net, ip6gre_net_id);
net/ipv6/ip6_gre.c
367
ip6erspan_tunnel_unlink_md(ign, t);
net/ipv6/ip6_gre.c
368
ip6gre_tunnel_unlink(ign, t);
net/ipv6/ip6_gre.c
369
dst_cache_reset(&t->dst_cache);
net/ipv6/ip6_gre.c
370
netdev_put(dev, &t->dev_tracker);
net/ipv6/ip6_gre.c
375
struct ip6_tnl *t = netdev_priv(dev);
net/ipv6/ip6_gre.c
376
struct ip6gre_net *ign = net_generic(t->net, ip6gre_net_id);
net/ipv6/ip6_gre.c
378
ip6gre_tunnel_unlink_md(ign, t);
net/ipv6/ip6_gre.c
379
ip6gre_tunnel_unlink(ign, t);
net/ipv6/ip6_gre.c
382
dst_cache_reset(&t->dst_cache);
net/ipv6/ip6_gre.c
383
netdev_put(dev, &t->dev_tracker);
net/ipv6/ip6_gre.c
393
struct ip6_tnl *t;
net/ipv6/ip6_gre.c
400
t = ip6gre_tunnel_lookup(skb->dev, &ipv6h->daddr, &ipv6h->saddr,
net/ipv6/ip6_gre.c
402
if (!t)
net/ipv6/ip6_gre.c
408
t->parms.name);
net/ipv6/ip6_gre.c
415
t->parms.name);
net/ipv6/ip6_gre.c
431
t->parms.name);
net/ipv6/ip6_gre.c
435
t->parms.name);
net/ipv6/ip6_gre.c
448
if (time_before(jiffies, t->err_time + IP6TUNNEL_ERR_TIMEO))
net/ipv6/ip6_gre.c
449
t->err_count++;
net/ipv6/ip6_gre.c
451
t->err_count = 1;
net/ipv6/ip6_gre.c
452
t->err_time = jiffies;
net/ipv6/ip6_gre.c
612
struct ip6_tnl *t = netdev_priv(dev);
net/ipv6/ip6_gre.c
614
if (!(t->parms.flags & IP6_TNL_F_IGN_ENCAP_LIMIT))
net/ipv6/ip6_gre.c
615
*encap_limit = t->parms.encap_limit;
net/ipv6/ip6_gre.c
617
memcpy(fl6, &t->fl.u.ip6, sizeof(*fl6));
net/ipv6/ip6_gre.c
619
if (t->parms.flags & IP6_TNL_F_USE_ORIG_TCLASS)
net/ipv6/ip6_gre.c
622
*dsfield = ip6_tclass(t->parms.flowinfo);
net/ipv6/ip6_gre.c
624
if (t->parms.flags & IP6_TNL_F_USE_ORIG_FWMARK)
net/ipv6/ip6_gre.c
627
fl6->flowi6_mark = t->parms.fwmark;
net/ipv6/ip6_gre.c
638
struct ip6_tnl *t = netdev_priv(dev);
net/ipv6/ip6_gre.c
655
} else if (!(t->parms.flags & IP6_TNL_F_IGN_ENCAP_LIMIT)) {
net/ipv6/ip6_gre.c
656
*encap_limit = t->parms.encap_limit;
net/ipv6/ip6_gre.c
659
memcpy(fl6, &t->fl.u.ip6, sizeof(*fl6));
net/ipv6/ip6_gre.c
661
if (t->parms.flags & IP6_TNL_F_USE_ORIG_TCLASS)
net/ipv6/ip6_gre.c
664
*dsfield = ip6_tclass(t->parms.flowinfo);
net/ipv6/ip6_gre.c
666
if (t->parms.flags & IP6_TNL_F_USE_ORIG_FLOWLABEL)
net/ipv6/ip6_gre.c
669
if (t->parms.flags & IP6_TNL_F_USE_ORIG_FWMARK)
net/ipv6/ip6_gre.c
672
fl6->flowi6_mark = t->parms.fwmark;
net/ipv6/ip6_gre.c
684
struct ip6_tnl *t = netdev_priv(dev);
net/ipv6/ip6_gre.c
686
if (!(t->parms.flags & IP6_TNL_F_IGN_ENCAP_LIMIT))
net/ipv6/ip6_gre.c
687
*encap_limit = t->parms.encap_limit;
net/ipv6/ip6_gre.c
689
memcpy(fl6, &t->fl.u.ip6, sizeof(*fl6));
net/ipv6/ip6_gre.c
691
if (t->parms.flags & IP6_TNL_F_USE_ORIG_TCLASS)
net/ipv6/ip6_gre.c
694
*dsfield = ip6_tclass(t->parms.flowinfo);
net/ipv6/ip6_gre.c
696
if (t->parms.flags & IP6_TNL_F_USE_ORIG_FWMARK)
net/ipv6/ip6_gre.c
699
fl6->flowi6_mark = t->parms.fwmark;
net/ipv6/ip6_gre.c
79
static void ip6gre_tunnel_link(struct ip6gre_net *ign, struct ip6_tnl *t);
net/ipv6/ip6_gre.c
792
struct ip6_tnl *t = netdev_priv(dev);
net/ipv6/ip6_gre.c
80
static void ip6gre_tnl_link_config(struct ip6_tnl *t, int set_mtu);
net/ipv6/ip6_gre.c
801
if (!t->parms.collect_md)
net/ipv6/ip6_gre.c
806
t->parms.o_flags));
net/ipv6/ip6_gre.c
81
static void ip6erspan_tnl_link_config(struct ip6_tnl *t, int set_mtu);
net/ipv6/ip6_gre.c
825
struct ip6_tnl *t = netdev_priv(dev);
net/ipv6/ip6_gre.c
833
if (ipv6_addr_equal(&t->parms.raddr, &ipv6h->saddr))
net/ipv6/ip6_gre.c
836
if (!t->parms.collect_md &&
net/ipv6/ip6_gre.c
841
t->parms.o_flags)))
net/ipv6/ip6_gre.c
857
struct ip6_tnl *t = netdev_priv(dev);
net/ipv6/ip6_gre.c
864
if (!t->parms.collect_md &&
net/ipv6/ip6_gre.c
869
t->parms.o_flags));
net/ipv6/ip6_gre.c
880
struct ip6_tnl *t = netdev_priv(dev);
net/ipv6/ip6_gre.c
887
if (!ip6_tnl_xmit_ctl(t, &t->parms.laddr, &t->parms.raddr))
net/ipv6/ip6_gre.c
909
if (!t->parms.collect_md || !IS_ERR(skb_tunnel_info_txcheck(skb)))
net/ipv6/ip6_gre.c
920
struct ip6_tnl *t = netdev_priv(dev);
net/ipv6/ip6_gre.c
935
if (!ip6_tnl_xmit_ctl(t, &t->parms.laddr, &t->parms.raddr))
net/ipv6/ip6_gre.c
963
if (skb_cow_head(skb, dev->needed_headroom ?: t->hlen))
net/ipv6/ip6_gre.c
966
__clear_bit(IP_TUNNEL_KEY_BIT, t->parms.o_flags);
net/ipv6/ip6_gre.c
972
if (t->parms.collect_md) {
net/ipv6/ip6_tunnel.c
100
for (t = rcu_dereference(start); t; t = rcu_dereference(t->next))
net/ipv6/ip6_tunnel.c
1027
ip6_tnl_addr_conflict(const struct ip6_tnl *t, const struct ipv6hdr *hdr)
net/ipv6/ip6_tunnel.c
1029
return ipv6_addr_equal(&t->parms.raddr, &hdr->saddr);
net/ipv6/ip6_tunnel.c
1032
int ip6_tnl_xmit_ctl(struct ip6_tnl *t,
net/ipv6/ip6_tunnel.c
1036
struct __ip6_tnl_parm *p = &t->parms;
net/ipv6/ip6_tunnel.c
1038
struct net *net = t->net;
net/ipv6/ip6_tunnel.c
1040
if (t->parms.collect_md)
net/ipv6/ip6_tunnel.c
1045
(ip6_tnl_get_cap(t, laddr, raddr) & IP6_TNL_F_CAP_XMIT))) {
net/ipv6/ip6_tunnel.c
1094
struct ip6_tnl *t = netdev_priv(dev);
net/ipv6/ip6_tunnel.c
1095
struct net *net = t->net;
net/ipv6/ip6_tunnel.c
1101
unsigned int eth_hlen = t->dev->type == ARPHRD_ETHER ? ETH_HLEN : 0;
net/ipv6/ip6_tunnel.c
1102
unsigned int psh_hlen = sizeof(struct ipv6hdr) + t->encap_hlen;
net/ipv6/ip6_tunnel.c
1111
if (t->parms.collect_md) {
net/ipv6/ip6_tunnel.c
1115
hop_limit = t->parms.hop_limit;
net/ipv6/ip6_tunnel.c
1119
if (ipv6_addr_any(&t->parms.raddr)) {
net/ipv6/ip6_tunnel.c
1150
} else if (t->parms.proto != 0 && !(t->parms.flags &
net/ipv6/ip6_tunnel.c
1160
dst = dst_cache_get(&t->dst_cache);
net/ipv6/ip6_tunnel.c
1162
if (!ip6_tnl_xmit_ctl(t, &fl6->saddr, &fl6->daddr))
net/ipv6/ip6_tunnel.c
1180
if (t->parms.collect_md && ipv6_addr_any(&fl6->saddr) &&
net/ipv6/ip6_tunnel.c
1192
t->parms.name);
net/ipv6/ip6_tunnel.c
1195
mtu = dst6_mtu(dst) - eth_hlen - psh_hlen - t->tun_hlen;
net/ipv6/ip6_tunnel.c
120
struct ip6_tnl *t, *cand = NULL;
net/ipv6/ip6_tunnel.c
1204
if (skb->len - t->tun_hlen - eth_hlen > mtu && !skb_is_gso(skb)) {
net/ipv6/ip6_tunnel.c
1210
if (t->err_count > 0) {
net/ipv6/ip6_tunnel.c
1212
t->err_time + IP6TUNNEL_ERR_TIMEO)) {
net/ipv6/ip6_tunnel.c
1213
t->err_count--;
net/ipv6/ip6_tunnel.c
1217
t->err_count = 0;
net/ipv6/ip6_tunnel.c
1221
skb_scrub_packet(skb, !net_eq(t->net, dev_net(dev)));
net/ipv6/ip6_tunnel.c
1242
if (t->parms.collect_md) {
net/ipv6/ip6_tunnel.c
1243
if (t->encap.type != TUNNEL_ENCAP_NONE)
net/ipv6/ip6_tunnel.c
1247
dst_cache_set_ip6(&t->dst_cache, ndst, &fl6->saddr);
net/ipv6/ip6_tunnel.c
125
if (!ipv6_addr_equal(local, &t->parms.laddr) ||
net/ipv6/ip6_tunnel.c
126
!ipv6_addr_equal(remote, &t->parms.raddr) ||
net/ipv6/ip6_tunnel.c
1264
+ dst->header_len + t->hlen;
net/ipv6/ip6_tunnel.c
1267
err = ip6_tnl_encap(skb, t, &proto, fl6);
net/ipv6/ip6_tunnel.c
127
!(t->dev->flags & IFF_UP))
net/ipv6/ip6_tunnel.c
130
if (link == t->parms.link)
net/ipv6/ip6_tunnel.c
1300
struct ip6_tnl *t = netdev_priv(dev);
net/ipv6/ip6_tunnel.c
131
return t;
net/ipv6/ip6_tunnel.c
1311
tproto = READ_ONCE(t->parms.proto);
net/ipv6/ip6_tunnel.c
1315
if (t->parms.collect_md) {
net/ipv6/ip6_tunnel.c
133
cand = t;
net/ipv6/ip6_tunnel.c
1344
if (!(t->parms.flags & IP6_TNL_F_IGN_ENCAP_LIMIT))
net/ipv6/ip6_tunnel.c
1345
encap_limit = t->parms.encap_limit;
net/ipv6/ip6_tunnel.c
1365
memcpy(&fl6, &t->fl.u.ip6, sizeof(fl6));
net/ipv6/ip6_tunnel.c
1368
if (t->parms.flags & IP6_TNL_F_USE_ORIG_FWMARK)
net/ipv6/ip6_tunnel.c
1371
fl6.flowi6_mark = t->parms.fwmark;
net/ipv6/ip6_tunnel.c
1376
if (t->parms.flags & IP6_TNL_F_USE_ORIG_TCLASS)
net/ipv6/ip6_tunnel.c
1379
dsfield = ip6_tclass(t->parms.flowinfo);
net/ipv6/ip6_tunnel.c
1384
if (t->parms.flags & IP6_TNL_F_USE_ORIG_TCLASS)
net/ipv6/ip6_tunnel.c
1387
dsfield = ip6_tclass(t->parms.flowinfo);
net/ipv6/ip6_tunnel.c
1388
if (t->parms.flags & IP6_TNL_F_USE_ORIG_FLOWLABEL)
net/ipv6/ip6_tunnel.c
139
if (!ipv6_addr_equal(local, &t->parms.laddr) ||
net/ipv6/ip6_tunnel.c
1392
orig_dsfield = dsfield = ip6_tclass(t->parms.flowinfo);
net/ipv6/ip6_tunnel.c
140
!ipv6_addr_any(&t->parms.raddr) ||
net/ipv6/ip6_tunnel.c
141
!(t->dev->flags & IFF_UP))
net/ipv6/ip6_tunnel.c
1430
struct ip6_tnl *t = netdev_priv(dev);
net/ipv6/ip6_tunnel.c
144
if (link == t->parms.link)
net/ipv6/ip6_tunnel.c
1442
if (ip6_tnl_addr_conflict(t, ipv6_hdr(skb)))
net/ipv6/ip6_tunnel.c
145
return t;
net/ipv6/ip6_tunnel.c
1466
static void ip6_tnl_link_config(struct ip6_tnl *t)
net/ipv6/ip6_tunnel.c
1468
struct net_device *dev = t->dev;
net/ipv6/ip6_tunnel.c
147
cand = t;
net/ipv6/ip6_tunnel.c
1470
struct __ip6_tnl_parm *p = &t->parms;
net/ipv6/ip6_tunnel.c
1471
struct flowi6 *fl6 = &t->fl.u.ip6;
net/ipv6/ip6_tunnel.c
1490
p->flags |= ip6_tnl_get_cap(t, &p->laddr, &p->raddr);
net/ipv6/ip6_tunnel.c
1497
t->tun_hlen = 0;
net/ipv6/ip6_tunnel.c
1498
t->hlen = t->encap_hlen + t->tun_hlen;
net/ipv6/ip6_tunnel.c
1499
t_hlen = t->hlen + sizeof(struct ipv6hdr);
net/ipv6/ip6_tunnel.c
1505
struct rt6_info *rt = rt6_lookup(t->net,
net/ipv6/ip6_tunnel.c
1514
tdev = __dev_get_by_index(t->net, p->link);
net/ipv6/ip6_tunnel.c
152
if (!ipv6_addr_equal(remote, &t->parms.raddr) ||
net/ipv6/ip6_tunnel.c
1522
if (!(t->parms.flags & IP6_TNL_F_IGN_ENCAP_LIMIT))
net/ipv6/ip6_tunnel.c
153
!ipv6_addr_any(&t->parms.laddr) ||
net/ipv6/ip6_tunnel.c
154
!(t->dev->flags & IFF_UP))
net/ipv6/ip6_tunnel.c
1542
ip6_tnl_change(struct ip6_tnl *t, const struct __ip6_tnl_parm *p)
net/ipv6/ip6_tunnel.c
1544
t->parms.laddr = p->laddr;
net/ipv6/ip6_tunnel.c
1545
t->parms.raddr = p->raddr;
net/ipv6/ip6_tunnel.c
1546
t->parms.flags = p->flags;
net/ipv6/ip6_tunnel.c
1547
t->parms.hop_limit = p->hop_limit;
net/ipv6/ip6_tunnel.c
1548
t->parms.encap_limit = p->encap_limit;
net/ipv6/ip6_tunnel.c
1549
t->parms.flowinfo = p->flowinfo;
net/ipv6/ip6_tunnel.c
1550
t->parms.link = p->link;
net/ipv6/ip6_tunnel.c
1551
t->parms.proto = p->proto;
net/ipv6/ip6_tunnel.c
1552
t->parms.fwmark = p->fwmark;
net/ipv6/ip6_tunnel.c
1553
dst_cache_reset(&t->dst_cache);
net/ipv6/ip6_tunnel.c
1554
ip6_tnl_link_config(t);
net/ipv6/ip6_tunnel.c
1557
static void ip6_tnl_update(struct ip6_tnl *t, struct __ip6_tnl_parm *p)
net/ipv6/ip6_tunnel.c
1559
struct net *net = t->net;
net/ipv6/ip6_tunnel.c
1562
ip6_tnl_unlink(ip6n, t);
net/ipv6/ip6_tunnel.c
1564
ip6_tnl_change(t, p);
net/ipv6/ip6_tunnel.c
1565
ip6_tnl_link(ip6n, t);
net/ipv6/ip6_tunnel.c
1566
netdev_state_change(t->dev);
net/ipv6/ip6_tunnel.c
1569
static int ip6_tnl0_update(struct ip6_tnl *t, struct __ip6_tnl_parm *p,
net/ipv6/ip6_tunnel.c
157
if (link == t->parms.link)
net/ipv6/ip6_tunnel.c
1578
p->flags != t->parms.flags || p->hop_limit || p->encap_limit ||
net/ipv6/ip6_tunnel.c
158
return t;
net/ipv6/ip6_tunnel.c
1582
t->parms.proto = p->proto;
net/ipv6/ip6_tunnel.c
1583
netdev_state_change(t->dev);
net/ipv6/ip6_tunnel.c
160
cand = t;
net/ipv6/ip6_tunnel.c
1651
struct ip6_tnl *t = netdev_priv(dev);
net/ipv6/ip6_tunnel.c
1652
struct net *net = t->net;
net/ipv6/ip6_tunnel.c
166
t = rcu_dereference(ip6n->collect_md_tun);
net/ipv6/ip6_tunnel.c
1665
t = ip6_tnl_locate(net, &p1, 0);
net/ipv6/ip6_tunnel.c
1666
if (IS_ERR(t))
net/ipv6/ip6_tunnel.c
1667
t = netdev_priv(dev);
net/ipv6/ip6_tunnel.c
167
if (t && t->dev->flags & IFF_UP)
net/ipv6/ip6_tunnel.c
1671
ip6_tnl_parm_to_user(&p, &t->parms);
net/ipv6/ip6_tunnel.c
168
return t;
net/ipv6/ip6_tunnel.c
1688
t = ip6_tnl_locate(net, &p1, cmd == SIOCADDTUNNEL);
net/ipv6/ip6_tunnel.c
1690
if (!IS_ERR(t)) {
net/ipv6/ip6_tunnel.c
1691
if (t->dev != dev) {
net/ipv6/ip6_tunnel.c
1696
t = netdev_priv(dev);
net/ipv6/ip6_tunnel.c
1698
ip6_tnl0_update(t, &p1, false);
net/ipv6/ip6_tunnel.c
170
t = rcu_dereference(ip6n->tnls_wc[0]);
net/ipv6/ip6_tunnel.c
1700
ip6_tnl_update(t, &p1);
net/ipv6/ip6_tunnel.c
1702
if (!IS_ERR(t)) {
net/ipv6/ip6_tunnel.c
1704
ip6_tnl_parm_to_user(&p, &t->parms);
net/ipv6/ip6_tunnel.c
1709
err = PTR_ERR(t);
net/ipv6/ip6_tunnel.c
171
if (t && (t->dev->flags & IFF_UP))
net/ipv6/ip6_tunnel.c
172
return t;
net/ipv6/ip6_tunnel.c
1723
t = ip6_tnl_locate(net, &p1, 0);
net/ipv6/ip6_tunnel.c
1724
if (IS_ERR(t))
net/ipv6/ip6_tunnel.c
1727
if (t->dev == ip6n->fb_tnl_dev)
net/ipv6/ip6_tunnel.c
1729
dev = t->dev;
net/ipv6/ip6_tunnel.c
1777
struct ip6_tnl *t = netdev_priv(dev);
net/ipv6/ip6_tunnel.c
1779
return READ_ONCE(t->parms.link);
net/ipv6/ip6_tunnel.c
1813
int ip6_tnl_encap_setup(struct ip6_tnl *t,
net/ipv6/ip6_tunnel.c
1818
memset(&t->encap, 0, sizeof(t->encap));
net/ipv6/ip6_tunnel.c
1824
t->encap.type = ipencap->type;
net/ipv6/ip6_tunnel.c
1825
t->encap.sport = ipencap->sport;
net/ipv6/ip6_tunnel.c
1826
t->encap.dport = ipencap->dport;
net/ipv6/ip6_tunnel.c
1827
t->encap.flags = ipencap->flags;
net/ipv6/ip6_tunnel.c
1829
t->encap_hlen = hlen;
net/ipv6/ip6_tunnel.c
1830
t->hlen = t->encap_hlen + t->tun_hlen;
net/ipv6/ip6_tunnel.c
1839
struct ip6_tnl *t = netdev_priv(ctx->dev);
net/ipv6/ip6_tunnel.c
1841
.daddr = t->parms.raddr,
net/ipv6/ip6_tunnel.c
1849
path->tun.src_v6 = t->parms.laddr;
net/ipv6/ip6_tunnel.c
1850
path->tun.dst_v6 = t->parms.raddr;
net/ipv6/ip6_tunnel.c
1918
struct ip6_tnl *t = netdev_priv(dev);
net/ipv6/ip6_tunnel.c
1922
t->dev = dev;
net/ipv6/ip6_tunnel.c
1924
ret = dst_cache_init(&t->dst_cache, GFP_KERNEL);
net/ipv6/ip6_tunnel.c
1928
ret = gro_cells_init(&t->gro_cells, dev);
net/ipv6/ip6_tunnel.c
1932
t->tun_hlen = 0;
net/ipv6/ip6_tunnel.c
1933
t->hlen = t->encap_hlen + t->tun_hlen;
net/ipv6/ip6_tunnel.c
1934
t_hlen = t->hlen + sizeof(struct ipv6hdr);
net/ipv6/ip6_tunnel.c
1938
if (!(t->parms.flags & IP6_TNL_F_IGN_ENCAP_LIMIT))
net/ipv6/ip6_tunnel.c
1943
netdev_hold(dev, &t->dev_tracker, GFP_KERNEL);
net/ipv6/ip6_tunnel.c
1948
dst_cache_destroy(&t->dst_cache);
net/ipv6/ip6_tunnel.c
1960
struct ip6_tnl *t = netdev_priv(dev);
net/ipv6/ip6_tunnel.c
1965
ip6_tnl_link_config(t);
net/ipv6/ip6_tunnel.c
1966
if (t->parms.collect_md)
net/ipv6/ip6_tunnel.c
1980
struct ip6_tnl *t = netdev_priv(dev);
net/ipv6/ip6_tunnel.c
1984
t->net = net;
net/ipv6/ip6_tunnel.c
1985
t->parms.proto = IPPROTO_IPV6;
net/ipv6/ip6_tunnel.c
1987
rcu_assign_pointer(ip6n->tnls_wc[0], t);
net/ipv6/ip6_tunnel.c
2055
struct ip6_tnl *nt, *t;
net/ipv6/ip6_tunnel.c
2076
t = ip6_tnl_locate(net, &nt->parms, 0);
net/ipv6/ip6_tunnel.c
2077
if (!IS_ERR(t))
net/ipv6/ip6_tunnel.c
2092
struct ip6_tnl *t = netdev_priv(dev);
net/ipv6/ip6_tunnel.c
2094
struct net *net = t->net;
net/ipv6/ip6_tunnel.c
211
ip6_tnl_link(struct ip6_tnl_net *ip6n, struct ip6_tnl *t)
net/ipv6/ip6_tunnel.c
2112
if (ip6_tnl0_update(t, &p, true) < 0) {
net/ipv6/ip6_tunnel.c
2122
int err = ip6_tnl_encap_setup(t, &ipencap);
net/ipv6/ip6_tunnel.c
213
struct ip6_tnl __rcu **tp = ip6_tnl_bucket(ip6n, &t->parms);
net/ipv6/ip6_tunnel.c
2131
t = ip6_tnl_locate(net, &p, 0);
net/ipv6/ip6_tunnel.c
2132
if (!IS_ERR(t)) {
net/ipv6/ip6_tunnel.c
2133
if (t->dev != dev)
net/ipv6/ip6_tunnel.c
2136
t = netdev_priv(dev);
net/ipv6/ip6_tunnel.c
2138
ip6_tnl_update(t, &p);
net/ipv6/ip6_tunnel.c
215
if (t->parms.collect_md)
net/ipv6/ip6_tunnel.c
216
rcu_assign_pointer(ip6n->collect_md_tun, t);
net/ipv6/ip6_tunnel.c
217
rcu_assign_pointer(t->next , rtnl_dereference(*tp));
net/ipv6/ip6_tunnel.c
218
rcu_assign_pointer(*tp, t);
net/ipv6/ip6_tunnel.c
228
ip6_tnl_unlink(struct ip6_tnl_net *ip6n, struct ip6_tnl *t)
net/ipv6/ip6_tunnel.c
2280
struct ip6_tnl *t;
net/ipv6/ip6_tunnel.c
2287
t = rtnl_net_dereference(net, ip6n->tnls_r_l[h]);
net/ipv6/ip6_tunnel.c
2288
while (t) {
net/ipv6/ip6_tunnel.c
2292
if (!net_eq(dev_net(t->dev), net))
net/ipv6/ip6_tunnel.c
2293
unregister_netdevice_queue(t->dev, list);
net/ipv6/ip6_tunnel.c
2295
t = rtnl_net_dereference(net, t->next);
net/ipv6/ip6_tunnel.c
2299
t = rtnl_net_dereference(net, ip6n->tnls_wc[0]);
net/ipv6/ip6_tunnel.c
2300
while (t) {
net/ipv6/ip6_tunnel.c
2304
if (!net_eq(dev_net(t->dev), net))
net/ipv6/ip6_tunnel.c
2305
unregister_netdevice_queue(t->dev, list);
net/ipv6/ip6_tunnel.c
2307
t = rtnl_net_dereference(net, t->next);
net/ipv6/ip6_tunnel.c
2314
struct ip6_tnl *t = NULL;
net/ipv6/ip6_tunnel.c
233
if (t->parms.collect_md)
net/ipv6/ip6_tunnel.c
2343
t = netdev_priv(ip6n->fb_tnl_dev);
net/ipv6/ip6_tunnel.c
2345
strcpy(t->parms.name, ip6n->fb_tnl_dev->name);
net/ipv6/ip6_tunnel.c
236
for (tp = ip6_tnl_bucket(ip6n, &t->parms);
net/ipv6/ip6_tunnel.c
239
if (t == iter) {
net/ipv6/ip6_tunnel.c
240
rcu_assign_pointer(*tp, t->next);
net/ipv6/ip6_tunnel.c
248
struct ip6_tnl *t = netdev_priv(dev);
net/ipv6/ip6_tunnel.c
250
gro_cells_destroy(&t->gro_cells);
net/ipv6/ip6_tunnel.c
251
dst_cache_destroy(&t->dst_cache);
net/ipv6/ip6_tunnel.c
256
struct ip6_tnl *t = netdev_priv(dev);
net/ipv6/ip6_tunnel.c
257
struct ip6_tnl_net *ip6n = net_generic(t->net, ip6_tnl_net_id);
net/ipv6/ip6_tunnel.c
265
strcpy(t->parms.name, dev->name);
net/ipv6/ip6_tunnel.c
267
ip6_tnl_link(ip6n, t);
net/ipv6/ip6_tunnel.c
289
struct ip6_tnl *t;
net/ipv6/ip6_tunnel.c
301
dev = alloc_netdev(sizeof(*t), name, NET_NAME_UNKNOWN,
net/ipv6/ip6_tunnel.c
308
t = netdev_priv(dev);
net/ipv6/ip6_tunnel.c
309
t->parms = *p;
net/ipv6/ip6_tunnel.c
310
t->net = dev_net(dev);
net/ipv6/ip6_tunnel.c
315
return t;
net/ipv6/ip6_tunnel.c
344
struct ip6_tnl *t;
net/ipv6/ip6_tunnel.c
348
(t = rtnl_dereference(*tp)) != NULL;
net/ipv6/ip6_tunnel.c
349
tp = &t->next) {
net/ipv6/ip6_tunnel.c
350
if (ipv6_addr_equal(local, &t->parms.laddr) &&
net/ipv6/ip6_tunnel.c
351
ipv6_addr_equal(remote, &t->parms.raddr) &&
net/ipv6/ip6_tunnel.c
352
p->link == t->parms.link) {
net/ipv6/ip6_tunnel.c
356
return t;
net/ipv6/ip6_tunnel.c
375
struct ip6_tnl *t = netdev_priv(dev);
net/ipv6/ip6_tunnel.c
376
struct net *net = t->net;
net/ipv6/ip6_tunnel.c
382
ip6_tnl_unlink(ip6n, t);
net/ipv6/ip6_tunnel.c
383
dst_cache_reset(&t->dst_cache);
net/ipv6/ip6_tunnel.c
384
netdev_put(dev, &t->dev_tracker);
net/ipv6/ip6_tunnel.c
471
struct ip6_tnl *t;
net/ipv6/ip6_tunnel.c
482
t = ip6_tnl_lookup(dev_net(skb->dev), skb->dev->ifindex, &ipv6h->daddr, &ipv6h->saddr);
net/ipv6/ip6_tunnel.c
483
if (!t)
net/ipv6/ip6_tunnel.c
486
tproto = READ_ONCE(t->parms.proto);
net/ipv6/ip6_tunnel.c
495
t->parms.name);
net/ipv6/ip6_tunnel.c
501
t->parms.name);
net/ipv6/ip6_tunnel.c
517
t->parms.name);
net/ipv6/ip6_tunnel.c
522
t->parms.name);
net/ipv6/ip6_tunnel.c
715
static int ip4ip6_dscp_ecn_decapsulate(const struct ip6_tnl *t,
net/ipv6/ip6_tunnel.c
721
if (t->parms.flags & IP6_TNL_F_RCV_DSCP_COPY)
net/ipv6/ip6_tunnel.c
727
static int ip6ip6_dscp_ecn_decapsulate(const struct ip6_tnl *t,
net/ipv6/ip6_tunnel.c
731
if (t->parms.flags & IP6_TNL_F_RCV_DSCP_COPY)
net/ipv6/ip6_tunnel.c
737
static inline int mplsip6_dscp_ecn_decapsulate(const struct ip6_tnl *t,
net/ipv6/ip6_tunnel.c
745
__u32 ip6_tnl_get_cap(struct ip6_tnl *t,
net/ipv6/ip6_tunnel.c
749
struct __ip6_tnl_parm *p = &t->parms;
net/ipv6/ip6_tunnel.c
770
int ip6_tnl_rcv_ctl(struct ip6_tnl *t,
net/ipv6/ip6_tunnel.c
774
struct __ip6_tnl_parm *p = &t->parms;
net/ipv6/ip6_tunnel.c
776
struct net *net = t->net;
net/ipv6/ip6_tunnel.c
780
(ip6_tnl_get_cap(t, laddr, raddr) & IP6_TNL_F_CAP_RCV))) {
net/ipv6/ip6_tunnel.c
801
int (*dscp_ecn_decapsulate)(const struct ip6_tnl *t,
net/ipv6/ip6_tunnel.c
895
int ip6_tnl_rcv(struct ip6_tnl *t, struct sk_buff *skb,
net/ipv6/ip6_tunnel.c
900
int (*dscp_ecn_decapsulate)(const struct ip6_tnl *t,
net/ipv6/ip6_tunnel.c
908
return __ip6_tnl_rcv(t, skb, tpi, tun_dst, dscp_ecn_decapsulate,
net/ipv6/ip6_tunnel.c
930
int (*dscp_ecn_decapsulate)(const struct ip6_tnl *t,
net/ipv6/ip6_tunnel.c
934
struct ip6_tnl *t;
net/ipv6/ip6_tunnel.c
940
t = ip6_tnl_lookup(dev_net(skb->dev), skb->dev->ifindex, &ipv6h->saddr, &ipv6h->daddr);
net/ipv6/ip6_tunnel.c
942
if (t) {
net/ipv6/ip6_tunnel.c
943
u8 tproto = READ_ONCE(t->parms.proto);
net/ipv6/ip6_tunnel.c
950
if (!ip6_tnl_rcv_ctl(t, &ipv6h->daddr, &ipv6h->saddr))
net/ipv6/ip6_tunnel.c
954
if (t->parms.collect_md) {
net/ipv6/ip6_tunnel.c
961
ret = __ip6_tnl_rcv(t, skb, tpi, tun_dst, dscp_ecn_decapsulate,
net/ipv6/ip6_vti.c
100
ipv6_addr_equal(remote, &t->parms.raddr) &&
net/ipv6/ip6_vti.c
101
(t->dev->flags & IFF_UP))
net/ipv6/ip6_vti.c
102
return t;
net/ipv6/ip6_vti.c
1034
struct ip6_tnl *t;
net/ipv6/ip6_vti.c
1044
t = vti6_locate(net, &p, 0);
net/ipv6/ip6_vti.c
1046
if (t) {
net/ipv6/ip6_vti.c
1047
if (t->dev != dev)
net/ipv6/ip6_vti.c
1050
t = netdev_priv(dev);
net/ipv6/ip6_vti.c
1052
return vti6_update(t, &p, tb && tb[IFLA_MTU]);
net/ipv6/ip6_vti.c
108
if (ipv6_addr_equal(local, &t->parms.laddr) &&
net/ipv6/ip6_vti.c
109
(t->dev->flags & IFF_UP))
net/ipv6/ip6_vti.c
110
return t;
net/ipv6/ip6_vti.c
1118
struct ip6_tnl *t;
net/ipv6/ip6_vti.c
1122
t = rtnl_net_dereference(net, ip6n->tnls_r_l[h]);
net/ipv6/ip6_vti.c
1123
while (t) {
net/ipv6/ip6_vti.c
1124
unregister_netdevice_queue(t->dev, list);
net/ipv6/ip6_vti.c
1125
t = rtnl_net_dereference(net, t->next);
net/ipv6/ip6_vti.c
1129
t = rtnl_net_dereference(net, ip6n->tnls_wc[0]);
net/ipv6/ip6_vti.c
1130
if (t)
net/ipv6/ip6_vti.c
1131
unregister_netdevice_queue(t->dev, list);
net/ipv6/ip6_vti.c
1137
struct ip6_tnl *t = NULL;
net/ipv6/ip6_vti.c
115
if (ipv6_addr_equal(remote, &t->parms.raddr) &&
net/ipv6/ip6_vti.c
116
(t->dev->flags & IFF_UP))
net/ipv6/ip6_vti.c
1162
t = netdev_priv(ip6n->fb_tnl_dev);
net/ipv6/ip6_vti.c
1164
strcpy(t->parms.name, ip6n->fb_tnl_dev->name);
net/ipv6/ip6_vti.c
117
return t;
net/ipv6/ip6_vti.c
120
t = rcu_dereference(ip6n->tnls_wc[0]);
net/ipv6/ip6_vti.c
121
if (t && (t->dev->flags & IFF_UP))
net/ipv6/ip6_vti.c
122
return t;
net/ipv6/ip6_vti.c
154
vti6_tnl_link(struct vti6_net *ip6n, struct ip6_tnl *t)
net/ipv6/ip6_vti.c
156
struct ip6_tnl __rcu **tp = vti6_tnl_bucket(ip6n, &t->parms);
net/ipv6/ip6_vti.c
158
rcu_assign_pointer(t->next, rtnl_dereference(*tp));
net/ipv6/ip6_vti.c
159
rcu_assign_pointer(*tp, t);
net/ipv6/ip6_vti.c
163
vti6_tnl_unlink(struct vti6_net *ip6n, struct ip6_tnl *t)
net/ipv6/ip6_vti.c
168
for (tp = vti6_tnl_bucket(ip6n, &t->parms);
net/ipv6/ip6_vti.c
171
if (t == iter) {
net/ipv6/ip6_vti.c
172
rcu_assign_pointer(*tp, t->next);
net/ipv6/ip6_vti.c
180
struct ip6_tnl *t = netdev_priv(dev);
net/ipv6/ip6_vti.c
181
struct vti6_net *ip6n = net_generic(t->net, vti6_net_id);
net/ipv6/ip6_vti.c
189
strcpy(t->parms.name, dev->name);
net/ipv6/ip6_vti.c
191
vti6_tnl_link(ip6n, t);
net/ipv6/ip6_vti.c
202
struct ip6_tnl *t;
net/ipv6/ip6_vti.c
214
dev = alloc_netdev(sizeof(*t), name, NET_NAME_UNKNOWN, vti6_dev_setup);
net/ipv6/ip6_vti.c
220
t = netdev_priv(dev);
net/ipv6/ip6_vti.c
221
t->parms = *p;
net/ipv6/ip6_vti.c
222
t->net = dev_net(dev);
net/ipv6/ip6_vti.c
228
return t;
net/ipv6/ip6_vti.c
256
struct ip6_tnl *t;
net/ipv6/ip6_vti.c
260
(t = rtnl_dereference(*tp)) != NULL;
net/ipv6/ip6_vti.c
261
tp = &t->next) {
net/ipv6/ip6_vti.c
262
if (ipv6_addr_equal(local, &t->parms.laddr) &&
net/ipv6/ip6_vti.c
263
ipv6_addr_equal(remote, &t->parms.raddr)) {
net/ipv6/ip6_vti.c
267
return t;
net/ipv6/ip6_vti.c
284
struct ip6_tnl *t = netdev_priv(dev);
net/ipv6/ip6_vti.c
285
struct vti6_net *ip6n = net_generic(t->net, vti6_net_id);
net/ipv6/ip6_vti.c
290
vti6_tnl_unlink(ip6n, t);
net/ipv6/ip6_vti.c
291
netdev_put(dev, &t->dev_tracker);
net/ipv6/ip6_vti.c
297
struct ip6_tnl *t;
net/ipv6/ip6_vti.c
301
t = vti6_tnl_lookup(dev_net(skb->dev), &ipv6h->saddr, &ipv6h->daddr);
net/ipv6/ip6_vti.c
302
if (t) {
net/ipv6/ip6_vti.c
303
if (t->parms.proto != IPPROTO_IPV6 && t->parms.proto != 0) {
net/ipv6/ip6_vti.c
314
if (!ip6_tnl_rcv_ctl(t, &ipv6h->daddr, &ipv6h->saddr)) {
net/ipv6/ip6_vti.c
315
DEV_STATS_INC(t->dev, rx_dropped);
net/ipv6/ip6_vti.c
322
XFRM_TUNNEL_SKB_CB(skb)->tunnel.ip6 = t;
net/ipv6/ip6_vti.c
347
struct ip6_tnl *t = XFRM_TUNNEL_SKB_CB(skb)->tunnel.ip6;
net/ipv6/ip6_vti.c
351
if (!t)
net/ipv6/ip6_vti.c
354
dev = t->dev;
net/ipv6/ip6_vti.c
378
skb->mark = be32_to_cpu(t->parms.i_key);
net/ipv6/ip6_vti.c
385
skb_scrub_packet(skb, !net_eq(t->net, dev_net(skb->dev)));
net/ipv6/ip6_vti.c
406
vti6_addr_conflict(const struct ip6_tnl *t, const struct ipv6hdr *hdr)
net/ipv6/ip6_vti.c
408
return ipv6_addr_equal(&t->parms.raddr, &hdr->saddr);
net/ipv6/ip6_vti.c
443
struct ip6_tnl *t = netdev_priv(dev);
net/ipv6/ip6_vti.c
482
dst = xfrm_lookup_route(t->net, dst, fl, NULL, 0);
net/ipv6/ip6_vti.c
493
if (!vti6_state_check(x, &t->parms.raddr, &t->parms.laddr))
net/ipv6/ip6_vti.c
496
if (!ip6_tnl_xmit_ctl(t, (const struct in6_addr *)&x->props.saddr,
net/ipv6/ip6_vti.c
505
t->parms.name);
net/ipv6/ip6_vti.c
530
skb_scrub_packet(skb, !net_eq(t->net, dev_net(dev)));
net/ipv6/ip6_vti.c
534
err = dst_output(t->net, skb->sk, skb);
net/ipv6/ip6_vti.c
551
struct ip6_tnl *t = netdev_priv(dev);
net/ipv6/ip6_vti.c
562
if ((t->parms.proto != IPPROTO_IPV6 && t->parms.proto != 0) ||
net/ipv6/ip6_vti.c
563
vti6_addr_conflict(t, ipv6_hdr(skb)))
net/ipv6/ip6_vti.c
578
fl.flowi_mark = be32_to_cpu(t->parms.o_key);
net/ipv6/ip6_vti.c
599
struct ip6_tnl *t;
net/ipv6/ip6_vti.c
607
t = vti6_tnl_lookup(dev_net(skb->dev), &iph->daddr, &iph->saddr);
net/ipv6/ip6_vti.c
608
if (!t)
net/ipv6/ip6_vti.c
611
mark = be32_to_cpu(t->parms.o_key);
net/ipv6/ip6_vti.c
649
static void vti6_link_config(struct ip6_tnl *t, bool keep_mtu)
net/ipv6/ip6_vti.c
651
struct net_device *dev = t->dev;
net/ipv6/ip6_vti.c
652
struct __ip6_tnl_parm *p = &t->parms;
net/ipv6/ip6_vti.c
661
p->flags |= ip6_tnl_get_cap(t, &p->laddr, &p->raddr);
net/ipv6/ip6_vti.c
677
struct rt6_info *rt = rt6_lookup(t->net,
net/ipv6/ip6_vti.c
687
tdev = __dev_get_by_index(t->net, p->link);
net/ipv6/ip6_vti.c
707
vti6_tnl_change(struct ip6_tnl *t, const struct __ip6_tnl_parm *p,
net/ipv6/ip6_vti.c
710
t->parms.laddr = p->laddr;
net/ipv6/ip6_vti.c
711
t->parms.raddr = p->raddr;
net/ipv6/ip6_vti.c
712
t->parms.link = p->link;
net/ipv6/ip6_vti.c
713
t->parms.i_key = p->i_key;
net/ipv6/ip6_vti.c
714
t->parms.o_key = p->o_key;
net/ipv6/ip6_vti.c
715
t->parms.proto = p->proto;
net/ipv6/ip6_vti.c
716
t->parms.fwmark = p->fwmark;
net/ipv6/ip6_vti.c
717
dst_cache_reset(&t->dst_cache);
net/ipv6/ip6_vti.c
718
vti6_link_config(t, keep_mtu);
net/ipv6/ip6_vti.c
722
static int vti6_update(struct ip6_tnl *t, struct __ip6_tnl_parm *p,
net/ipv6/ip6_vti.c
725
struct net *net = dev_net(t->dev);
net/ipv6/ip6_vti.c
729
vti6_tnl_unlink(ip6n, t);
net/ipv6/ip6_vti.c
731
err = vti6_tnl_change(t, p, keep_mtu);
net/ipv6/ip6_vti.c
732
vti6_tnl_link(ip6n, t);
net/ipv6/ip6_vti.c
733
netdev_state_change(t->dev);
net/ipv6/ip6_vti.c
76
for (t = rcu_dereference(start); t; t = rcu_dereference(t->next))
net/ipv6/ip6_vti.c
801
struct ip6_tnl *t = NULL;
net/ipv6/ip6_vti.c
815
t = vti6_locate(net, &p1, 0);
net/ipv6/ip6_vti.c
819
if (!t)
net/ipv6/ip6_vti.c
820
t = netdev_priv(dev);
net/ipv6/ip6_vti.c
821
vti6_parm_to_user(&p, &t->parms);
net/ipv6/ip6_vti.c
837
t = vti6_locate(net, &p1, cmd == SIOCADDTUNNEL);
net/ipv6/ip6_vti.c
839
if (t) {
net/ipv6/ip6_vti.c
840
if (t->dev != dev) {
net/ipv6/ip6_vti.c
845
t = netdev_priv(dev);
net/ipv6/ip6_vti.c
847
err = vti6_update(t, &p1, false);
net/ipv6/ip6_vti.c
849
if (t) {
net/ipv6/ip6_vti.c
851
vti6_parm_to_user(&p, &t->parms);
net/ipv6/ip6_vti.c
869
t = vti6_locate(net, &p1, 0);
net/ipv6/ip6_vti.c
870
if (!t)
net/ipv6/ip6_vti.c
873
if (t->dev == ip6n->fb_tnl_dev)
net/ipv6/ip6_vti.c
875
dev = t->dev;
net/ipv6/ip6_vti.c
925
struct ip6_tnl *t = netdev_priv(dev);
net/ipv6/ip6_vti.c
927
t->dev = dev;
net/ipv6/ip6_vti.c
928
netdev_hold(dev, &t->dev_tracker, GFP_KERNEL);
net/ipv6/ip6_vti.c
939
struct ip6_tnl *t = netdev_priv(dev);
net/ipv6/ip6_vti.c
94
struct ip6_tnl *t;
net/ipv6/ip6_vti.c
944
vti6_link_config(t, true);
net/ipv6/ip6_vti.c
956
struct ip6_tnl *t = netdev_priv(dev);
net/ipv6/ip6_vti.c
960
t->net = net;
net/ipv6/ip6_vti.c
961
t->parms.proto = IPPROTO_IPV6;
net/ipv6/ip6_vti.c
963
rcu_assign_pointer(ip6n->tnls_wc[0], t);
net/ipv6/ip6_vti.c
99
if (ipv6_addr_equal(local, &t->parms.laddr) &&
net/ipv6/ip6mr.c
103
static void ipmr_expire_process(struct timer_list *t);
net/ipv6/ip6mr.c
840
static void ipmr_expire_process(struct timer_list *t)
net/ipv6/ip6mr.c
842
struct mr_table *mrt = timer_container_of(mrt, t, ipmr_expire_timer);
net/ipv6/ipcomp6.c
101
atomic_set(&t->tunnel_users, 1);
net/ipv6/ipcomp6.c
104
return t;
net/ipv6/ipcomp6.c
107
t->km.state = XFRM_STATE_DEAD;
net/ipv6/ipcomp6.c
108
xfrm_state_put(t);
net/ipv6/ipcomp6.c
109
t = NULL;
net/ipv6/ipcomp6.c
117
struct xfrm_state *t = NULL;
net/ipv6/ipcomp6.c
123
t = xfrm_state_lookup(net, mark, (xfrm_address_t *)&x->id.daddr,
net/ipv6/ipcomp6.c
125
if (!t) {
net/ipv6/ipcomp6.c
126
t = ipcomp6_tunnel_create(x);
net/ipv6/ipcomp6.c
127
if (!t) {
net/ipv6/ipcomp6.c
131
xfrm_state_insert(t);
net/ipv6/ipcomp6.c
132
xfrm_state_hold(t);
net/ipv6/ipcomp6.c
134
x->tunnel = t;
net/ipv6/ipcomp6.c
135
atomic_inc(&t->tunnel_users);
net/ipv6/ipcomp6.c
78
struct xfrm_state *t = NULL;
net/ipv6/ipcomp6.c
80
t = xfrm_state_alloc(net);
net/ipv6/ipcomp6.c
81
if (!t)
net/ipv6/ipcomp6.c
83
lockdep_set_class(&t->lock, &xfrm_state_lock_key);
net/ipv6/ipcomp6.c
85
t->id.proto = IPPROTO_IPV6;
net/ipv6/ipcomp6.c
86
t->id.spi = xfrm6_tunnel_alloc_spi(net, (xfrm_address_t *)&x->props.saddr);
net/ipv6/ipcomp6.c
87
if (!t->id.spi)
net/ipv6/ipcomp6.c
90
memcpy(t->id.daddr.a6, x->id.daddr.a6, sizeof(struct in6_addr));
net/ipv6/ipcomp6.c
91
memcpy(&t->sel, &x->sel, sizeof(t->sel));
net/ipv6/ipcomp6.c
92
t->props.family = AF_INET6;
net/ipv6/ipcomp6.c
93
t->props.mode = x->props.mode;
net/ipv6/ipcomp6.c
94
memcpy(t->props.saddr.a6, x->props.saddr.a6, sizeof(struct in6_addr));
net/ipv6/ipcomp6.c
95
memcpy(&t->mark, &x->mark, sizeof(t->mark));
net/ipv6/ipcomp6.c
96
t->if_id = x->if_id;
net/ipv6/ipcomp6.c
98
if (xfrm_init_state(t))
net/ipv6/netfilter/ip6_tables.c
1007
xt_table_unlock(t);
net/ipv6/netfilter/ip6_tables.c
1008
module_put(t->me);
net/ipv6/netfilter/ip6_tables.c
1010
ret = PTR_ERR(t);
net/ipv6/netfilter/ip6_tables.c
1024
struct xt_table *t;
net/ipv6/netfilter/ip6_tables.c
1035
t = xt_find_table_lock(net, AF_INET6, get.name);
net/ipv6/netfilter/ip6_tables.c
1036
if (!IS_ERR(t)) {
net/ipv6/netfilter/ip6_tables.c
1037
struct xt_table_info *private = t->private;
net/ipv6/netfilter/ip6_tables.c
1040
t, uptr->entrytable);
net/ipv6/netfilter/ip6_tables.c
1044
module_put(t->me);
net/ipv6/netfilter/ip6_tables.c
1045
xt_table_unlock(t);
net/ipv6/netfilter/ip6_tables.c
1047
ret = PTR_ERR(t);
net/ipv6/netfilter/ip6_tables.c
1058
struct xt_table *t;
net/ipv6/netfilter/ip6_tables.c
1069
t = xt_request_find_table_lock(net, AF_INET6, name);
net/ipv6/netfilter/ip6_tables.c
1070
if (IS_ERR(t)) {
net/ipv6/netfilter/ip6_tables.c
1071
ret = PTR_ERR(t);
net/ipv6/netfilter/ip6_tables.c
1076
if (valid_hooks != t->valid_hooks) {
net/ipv6/netfilter/ip6_tables.c
1081
oldinfo = xt_replace_table(t, num_counters, newinfo, &ret);
net/ipv6/netfilter/ip6_tables.c
1088
module_put(t->me);
net/ipv6/netfilter/ip6_tables.c
1091
module_put(t->me);
net/ipv6/netfilter/ip6_tables.c
1093
xt_table_unlock(t);
net/ipv6/netfilter/ip6_tables.c
1111
module_put(t->me);
net/ipv6/netfilter/ip6_tables.c
1112
xt_table_unlock(t);
net/ipv6/netfilter/ip6_tables.c
1178
struct xt_table *t;
net/ipv6/netfilter/ip6_tables.c
1187
t = xt_find_table_lock(net, AF_INET6, tmp.name);
net/ipv6/netfilter/ip6_tables.c
1188
if (IS_ERR(t)) {
net/ipv6/netfilter/ip6_tables.c
1189
ret = PTR_ERR(t);
net/ipv6/netfilter/ip6_tables.c
1194
private = t->private;
net/ipv6/netfilter/ip6_tables.c
1212
xt_table_unlock(t);
net/ipv6/netfilter/ip6_tables.c
1213
module_put(t->me);
net/ipv6/netfilter/ip6_tables.c
1238
struct xt_entry_target *t;
net/ipv6/netfilter/ip6_tables.c
1261
t = ip6t_get_target(e);
net/ipv6/netfilter/ip6_tables.c
1262
ret = xt_compat_target_to_user(t, dstptr, size);
net/ipv6/netfilter/ip6_tables.c
1291
struct xt_entry_target *t;
net/ipv6/netfilter/ip6_tables.c
1297
t = compat_ip6t_get_target(e);
net/ipv6/netfilter/ip6_tables.c
1298
module_put(t->u.kernel.target->me);
net/ipv6/netfilter/ip6_tables.c
1309
struct xt_entry_target *t;
net/ipv6/netfilter/ip6_tables.c
1342
t = compat_ip6t_get_target(e);
net/ipv6/netfilter/ip6_tables.c
1343
target = xt_request_find_target(NFPROTO_IPV6, t->u.user.name,
net/ipv6/netfilter/ip6_tables.c
1344
t->u.user.revision);
net/ipv6/netfilter/ip6_tables.c
1349
t->u.kernel.target = target;
net/ipv6/netfilter/ip6_tables.c
1360
module_put(t->u.kernel.target->me);
net/ipv6/netfilter/ip6_tables.c
1375
struct xt_entry_target *t;
net/ipv6/netfilter/ip6_tables.c
1393
t = compat_ip6t_get_target(e);
net/ipv6/netfilter/ip6_tables.c
1394
xt_compat_target_from_user(t, dstptr, size);
net/ipv6/netfilter/ip6_tables.c
1593
struct xt_table *t;
net/ipv6/netfilter/ip6_tables.c
1607
t = xt_find_table_lock(net, AF_INET6, get.name);
net/ipv6/netfilter/ip6_tables.c
1608
if (!IS_ERR(t)) {
net/ipv6/netfilter/ip6_tables.c
1609
const struct xt_table_info *private = t->private;
net/ipv6/netfilter/ip6_tables.c
1614
t, uptr->entrytable);
net/ipv6/netfilter/ip6_tables.c
1619
module_put(t->me);
net/ipv6/netfilter/ip6_tables.c
1620
xt_table_unlock(t);
net/ipv6/netfilter/ip6_tables.c
1622
ret = PTR_ERR(t);
net/ipv6/netfilter/ip6_tables.c
185
const struct xt_standard_target *t = (void *)ip6t_get_target_c(s);
net/ipv6/netfilter/ip6_tables.c
187
if (strcmp(t->target.u.kernel.target->name, XT_ERROR_TARGET) == 0) {
net/ipv6/netfilter/ip6_tables.c
189
*chainname = t->target.data;
net/ipv6/netfilter/ip6_tables.c
195
strcmp(t->target.u.kernel.target->name,
net/ipv6/netfilter/ip6_tables.c
197
t->verdict < 0) {
net/ipv6/netfilter/ip6_tables.c
300
const struct xt_entry_target *t;
net/ipv6/netfilter/ip6_tables.c
323
t = ip6t_get_target_c(e);
net/ipv6/netfilter/ip6_tables.c
324
WARN_ON(!t->u.kernel.target);
net/ipv6/netfilter/ip6_tables.c
333
if (!t->u.kernel.target->target) {
net/ipv6/netfilter/ip6_tables.c
336
v = ((struct xt_standard_target *)t)->verdict;
net/ipv6/netfilter/ip6_tables.c
363
acpar.target = t->u.kernel.target;
net/ipv6/netfilter/ip6_tables.c
364
acpar.targinfo = t->data;
net/ipv6/netfilter/ip6_tables.c
366
verdict = t->u.kernel.target->target(skb, &acpar);
net/ipv6/netfilter/ip6_tables.c
404
const struct xt_standard_target *t
net/ipv6/netfilter/ip6_tables.c
415
(strcmp(t->target.u.user.name,
net/ipv6/netfilter/ip6_tables.c
417
t->verdict < 0) || visited) {
net/ipv6/netfilter/ip6_tables.c
443
int newpos = t->verdict;
net/ipv6/netfilter/ip6_tables.c
445
if (strcmp(t->target.u.user.name,
net/ipv6/netfilter/ip6_tables.c
517
struct xt_entry_target *t = ip6t_get_target(e);
net/ipv6/netfilter/ip6_tables.c
522
.target = t->u.kernel.target,
net/ipv6/netfilter/ip6_tables.c
523
.targinfo = t->data,
net/ipv6/netfilter/ip6_tables.c
528
return xt_check_target(&par, t->u.target_size - sizeof(*t),
net/ipv6/netfilter/ip6_tables.c
538
struct xt_entry_target *t;
net/ipv6/netfilter/ip6_tables.c
562
t = ip6t_get_target(e);
net/ipv6/netfilter/ip6_tables.c
563
target = xt_request_find_target(NFPROTO_IPV6, t->u.user.name,
net/ipv6/netfilter/ip6_tables.c
564
t->u.user.revision);
net/ipv6/netfilter/ip6_tables.c
569
t->u.kernel.target = target;
net/ipv6/netfilter/ip6_tables.c
576
module_put(t->u.kernel.target->me);
net/ipv6/netfilter/ip6_tables.c
591
const struct xt_entry_target *t;
net/ipv6/netfilter/ip6_tables.c
596
t = ip6t_get_target_c(e);
net/ipv6/netfilter/ip6_tables.c
597
if (strcmp(t->u.user.name, XT_STANDARD_TARGET) != 0)
net/ipv6/netfilter/ip6_tables.c
599
verdict = ((struct xt_standard_target *)t)->verdict;
net/ipv6/netfilter/ip6_tables.c
656
struct xt_entry_target *t;
net/ipv6/netfilter/ip6_tables.c
662
t = ip6t_get_target(e);
net/ipv6/netfilter/ip6_tables.c
665
par.target = t->u.kernel.target;
net/ipv6/netfilter/ip6_tables.c
666
par.targinfo = t->data;
net/ipv6/netfilter/ip6_tables.c
756
get_counters(const struct xt_table_info *t,
net/ipv6/netfilter/ip6_tables.c
767
xt_entry_foreach(iter, t->entries, t->size) {
net/ipv6/netfilter/ip6_tables.c
786
static void get_old_counters(const struct xt_table_info *t,
net/ipv6/netfilter/ip6_tables.c
794
xt_entry_foreach(iter, t->entries, t->size) {
net/ipv6/netfilter/ip6_tables.c
848
const struct xt_entry_target *t;
net/ipv6/netfilter/ip6_tables.c
874
t = ip6t_get_target_c(e);
net/ipv6/netfilter/ip6_tables.c
875
if (xt_target_to_user(t, userptr + off + e->target_offset)) {
net/ipv6/netfilter/ip6_tables.c
910
const struct xt_entry_target *t;
net/ipv6/netfilter/ip6_tables.c
918
t = ip6t_get_target_c(e);
net/ipv6/netfilter/ip6_tables.c
919
off += xt_compat_target_offset(t->u.kernel.target);
net/ipv6/netfilter/ip6_tables.c
965
struct xt_table *t;
net/ipv6/netfilter/ip6_tables.c
979
t = xt_request_find_table_lock(net, AF_INET6, name);
net/ipv6/netfilter/ip6_tables.c
980
if (!IS_ERR(t)) {
net/ipv6/netfilter/ip6_tables.c
982
const struct xt_table_info *private = t->private;
net/ipv6/netfilter/ip6_tables.c
993
info.valid_hooks = t->valid_hooks;
net/ipv6/netfilter/nf_conntrack_reasm.c
134
static void nf_ct_frag6_expire(struct timer_list *t)
net/ipv6/netfilter/nf_conntrack_reasm.c
136
struct inet_frag_queue *frag = timer_container_of(frag, t, timer);
net/ipv6/reassembly.c
74
static void ip6_frag_expire(struct timer_list *t)
net/ipv6/reassembly.c
76
struct inet_frag_queue *frag = timer_container_of(frag, t, timer);
net/ipv6/sit.c
105
struct ip_tunnel *t;
net/ipv6/sit.c
109
for_each_ip_tunnel_rcu(t, sitn->tunnels_r_l[h0 ^ h1]) {
net/ipv6/sit.c
110
if (local == t->parms.iph.saddr &&
net/ipv6/sit.c
111
remote == t->parms.iph.daddr &&
net/ipv6/sit.c
112
(!dev || !t->parms.link || ifindex == t->parms.link ||
net/ipv6/sit.c
113
sifindex == t->parms.link) &&
net/ipv6/sit.c
1133
static void ipip6_tunnel_update(struct ip_tunnel *t,
net/ipv6/sit.c
1137
struct net *net = t->net;
net/ipv6/sit.c
114
(t->dev->flags & IFF_UP))
net/ipv6/sit.c
1140
ipip6_tunnel_unlink(sitn, t);
net/ipv6/sit.c
1142
t->parms.iph.saddr = p->iph.saddr;
net/ipv6/sit.c
1143
t->parms.iph.daddr = p->iph.daddr;
net/ipv6/sit.c
1144
__dev_addr_set(t->dev, &p->iph.saddr, 4);
net/ipv6/sit.c
1145
memcpy(t->dev->broadcast, &p->iph.daddr, 4);
net/ipv6/sit.c
1146
ipip6_tunnel_link(sitn, t);
net/ipv6/sit.c
1147
t->parms.iph.ttl = p->iph.ttl;
net/ipv6/sit.c
1148
t->parms.iph.tos = p->iph.tos;
net/ipv6/sit.c
1149
t->parms.iph.frag_off = p->iph.frag_off;
net/ipv6/sit.c
115
return t;
net/ipv6/sit.c
1150
if (t->parms.link != p->link || t->fwmark != fwmark) {
net/ipv6/sit.c
1151
t->parms.link = p->link;
net/ipv6/sit.c
1152
t->fwmark = fwmark;
net/ipv6/sit.c
1153
ipip6_tunnel_bind_dev(t->dev);
net/ipv6/sit.c
1155
dst_cache_reset(&t->dst_cache);
net/ipv6/sit.c
1156
netdev_state_change(t->dev);
net/ipv6/sit.c
1160
static int ipip6_tunnel_update_6rd(struct ip_tunnel *t,
net/ipv6/sit.c
117
for_each_ip_tunnel_rcu(t, sitn->tunnels_r[h0]) {
net/ipv6/sit.c
118
if (remote == t->parms.iph.daddr &&
net/ipv6/sit.c
1182
t->ip6rd.prefix = prefix;
net/ipv6/sit.c
1183
t->ip6rd.relay_prefix = relay_prefix;
net/ipv6/sit.c
1184
t->ip6rd.prefixlen = ip6rd->prefixlen;
net/ipv6/sit.c
1185
t->ip6rd.relay_prefixlen = ip6rd->relay_prefixlen;
net/ipv6/sit.c
1186
dst_cache_reset(&t->dst_cache);
net/ipv6/sit.c
1187
netdev_state_change(t->dev);
net/ipv6/sit.c
119
(!dev || !t->parms.link || ifindex == t->parms.link ||
net/ipv6/sit.c
1194
struct ip_tunnel *t = netdev_priv(dev);
net/ipv6/sit.c
120
sifindex == t->parms.link) &&
net/ipv6/sit.c
1201
t = ipip6_tunnel_locate(t->net, &p, 0);
net/ipv6/sit.c
1203
if (!t)
net/ipv6/sit.c
1204
t = netdev_priv(dev);
net/ipv6/sit.c
1206
ip6rd.prefix = t->ip6rd.prefix;
net/ipv6/sit.c
1207
ip6rd.relay_prefix = t->ip6rd.relay_prefix;
net/ipv6/sit.c
1208
ip6rd.prefixlen = t->ip6rd.prefixlen;
net/ipv6/sit.c
1209
ip6rd.relay_prefixlen = t->ip6rd.relay_prefixlen;
net/ipv6/sit.c
121
(t->dev->flags & IFF_UP))
net/ipv6/sit.c
1219
struct ip_tunnel *t = netdev_priv(dev);
net/ipv6/sit.c
122
return t;
net/ipv6/sit.c
1223
if (!ns_capable(t->net->user_ns, CAP_NET_ADMIN))
net/ipv6/sit.c
1229
err = ipip6_tunnel_update_6rd(t, &ip6rd);
net/ipv6/sit.c
124
for_each_ip_tunnel_rcu(t, sitn->tunnels_l[h1]) {
net/ipv6/sit.c
125
if (local == t->parms.iph.saddr &&
net/ipv6/sit.c
126
(!dev || !t->parms.link || ifindex == t->parms.link ||
net/ipv6/sit.c
1269
struct ip_tunnel *t = netdev_priv(dev);
net/ipv6/sit.c
127
sifindex == t->parms.link) &&
net/ipv6/sit.c
1272
t = ipip6_tunnel_locate(t->net, p, 0);
net/ipv6/sit.c
1273
if (!t)
net/ipv6/sit.c
1274
t = netdev_priv(dev);
net/ipv6/sit.c
1275
memcpy(p, &t->parms, sizeof(*p));
net/ipv6/sit.c
128
(t->dev->flags & IFF_UP))
net/ipv6/sit.c
1282
struct ip_tunnel *t = netdev_priv(dev);
net/ipv6/sit.c
1285
err = __ipip6_tunnel_ioctl_validate(t->net, p);
net/ipv6/sit.c
1289
t = ipip6_tunnel_locate(t->net, p, 1);
net/ipv6/sit.c
129
return t;
net/ipv6/sit.c
1290
if (!t)
net/ipv6/sit.c
1298
struct ip_tunnel *t = netdev_priv(dev);
net/ipv6/sit.c
1301
err = __ipip6_tunnel_ioctl_validate(t->net, p);
net/ipv6/sit.c
1305
t = ipip6_tunnel_locate(t->net, p, 0);
net/ipv6/sit.c
1307
if (!t)
net/ipv6/sit.c
131
t = rcu_dereference(sitn->tunnels_wc[0]);
net/ipv6/sit.c
1310
if (t) {
net/ipv6/sit.c
1311
if (t->dev != dev)
net/ipv6/sit.c
1317
t = netdev_priv(dev);
net/ipv6/sit.c
132
if (t && (t->dev->flags & IFF_UP))
net/ipv6/sit.c
1320
ipip6_tunnel_update(t, p, t->fwmark);
net/ipv6/sit.c
1329
struct ip_tunnel *t = netdev_priv(dev);
net/ipv6/sit.c
133
return t;
net/ipv6/sit.c
1331
if (!ns_capable(t->net->user_ns, CAP_NET_ADMIN))
net/ipv6/sit.c
1335
t = ipip6_tunnel_locate(t->net, p, 0);
net/ipv6/sit.c
1336
if (!t)
net/ipv6/sit.c
1338
if (t == netdev_priv(dev_to_sit_net(dev)->fb_tunnel_dev))
net/ipv6/sit.c
1340
dev = t->dev;
net/ipv6/sit.c
157
struct ip_tunnel *t)
net/ipv6/sit.c
159
return __ipip6_bucket(sitn, &t->parms);
net/ipv6/sit.c
1601
struct ip_tunnel *t = netdev_priv(dev);
net/ipv6/sit.c
1604
struct net *net = t->net;
net/ipv6/sit.c
1609
__u32 fwmark = t->fwmark;
net/ipv6/sit.c
1616
err = ip_tunnel_encap_setup(t, &ipencap);
net/ipv6/sit.c
162
static void ipip6_tunnel_unlink(struct sit_net *sitn, struct ip_tunnel *t)
net/ipv6/sit.c
1627
t = ipip6_tunnel_locate(net, &p, 0);
net/ipv6/sit.c
1629
if (t) {
net/ipv6/sit.c
1630
if (t->dev != dev)
net/ipv6/sit.c
1633
t = netdev_priv(dev);
net/ipv6/sit.c
1635
ipip6_tunnel_update(t, &p, fwmark);
net/ipv6/sit.c
1639
return ipip6_tunnel_update_6rd(t, &ip6rd);
net/ipv6/sit.c
167
for (tp = ipip6_bucket(sitn, t);
net/ipv6/sit.c
170
if (t == iter) {
net/ipv6/sit.c
171
rcu_assign_pointer(*tp, t->next);
net/ipv6/sit.c
177
static void ipip6_tunnel_link(struct sit_net *sitn, struct ip_tunnel *t)
net/ipv6/sit.c
179
struct ip_tunnel __rcu **tp = ipip6_bucket(sitn, t);
net/ipv6/sit.c
181
rcu_assign_pointer(t->next, rtnl_dereference(*tp));
net/ipv6/sit.c
1812
struct ip_tunnel *t;
net/ipv6/sit.c
1814
t = rtnl_net_dereference(net, sitn->tunnels[prio][h]);
net/ipv6/sit.c
1815
while (t) {
net/ipv6/sit.c
1819
if (!net_eq(dev_net(t->dev), net))
net/ipv6/sit.c
182
rcu_assign_pointer(*tp, t);
net/ipv6/sit.c
1820
unregister_netdevice_queue(t->dev, head);
net/ipv6/sit.c
1822
t = rtnl_net_dereference(net, t->next);
net/ipv6/sit.c
1831
struct ip_tunnel *t;
net/ipv6/sit.c
1856
t = netdev_priv(sitn->fb_tunnel_dev);
net/ipv6/sit.c
1857
t->net = net;
net/ipv6/sit.c
1866
strcpy(t->parms.name, sitn->fb_tunnel_dev->name);
net/ipv6/sit.c
188
struct ip_tunnel *t = netdev_priv(dev);
net/ipv6/sit.c
191
ipv6_addr_set(&t->ip6rd.prefix, htonl(0x20020000), 0, 0, 0);
net/ipv6/sit.c
192
t->ip6rd.relay_prefix = 0;
net/ipv6/sit.c
193
t->ip6rd.prefixlen = 16;
net/ipv6/sit.c
194
t->ip6rd.relay_prefixlen = 0;
net/ipv6/sit.c
197
memcpy(&t->ip6rd, &t0->ip6rd, sizeof(t->ip6rd));
net/ipv6/sit.c
204
struct ip_tunnel *t = netdev_priv(dev);
net/ipv6/sit.c
205
struct sit_net *sitn = net_generic(t->net, sit_net_id);
net/ipv6/sit.c
208
__dev_addr_set(dev, &t->parms.iph.saddr, 4);
net/ipv6/sit.c
209
memcpy(dev->broadcast, &t->parms.iph.daddr, 4);
net/ipv6/sit.c
211
if (test_bit(IP_TUNNEL_SIT_ISATAP_BIT, t->parms.i_flags))
net/ipv6/sit.c
222
ipip6_tunnel_link(sitn, t);
net/ipv6/sit.c
235
struct ip_tunnel *t, *nt;
net/ipv6/sit.c
242
(t = rtnl_dereference(*tp)) != NULL;
net/ipv6/sit.c
243
tp = &t->next) {
net/ipv6/sit.c
244
if (local == t->parms.iph.saddr &&
net/ipv6/sit.c
245
remote == t->parms.iph.daddr &&
net/ipv6/sit.c
246
parms->link == t->parms.link) {
net/ipv6/sit.c
250
return t;
net/ipv6/sit.c
263
dev = alloc_netdev(sizeof(*t), name, NET_NAME_UNKNOWN,
net/ipv6/sit.c
294
__ipip6_tunnel_locate_prl(struct ip_tunnel *t, __be32 addr)
net/ipv6/sit.c
298
for_each_prl_rcu(t->prl)
net/ipv6/sit.c
307
struct ip_tunnel *t = netdev_priv(dev);
net/ipv6/sit.c
329
ca = min(t->prl_count, cmax);
net/ipv6/sit.c
346
for_each_prl_rcu(t->prl) {
net/ipv6/sit.c
371
ipip6_tunnel_add_prl(struct ip_tunnel *t, struct ip_tunnel_prl *a, int chg)
net/ipv6/sit.c
381
for (p = rtnl_dereference(t->prl); p; p = rtnl_dereference(p->next)) {
net/ipv6/sit.c
403
p->next = t->prl;
net/ipv6/sit.c
406
t->prl_count++;
net/ipv6/sit.c
407
rcu_assign_pointer(t->prl, p);
net/ipv6/sit.c
425
ipip6_tunnel_del_prl(struct ip_tunnel *t, struct ip_tunnel_prl *a)
net/ipv6/sit.c
434
for (p = &t->prl;
net/ipv6/sit.c
440
t->prl_count--;
net/ipv6/sit.c
446
x = rtnl_dereference(t->prl);
net/ipv6/sit.c
448
t->prl_count = 0;
net/ipv6/sit.c
450
t->prl = NULL;
net/ipv6/sit.c
460
struct ip_tunnel *t = netdev_priv(dev);
net/ipv6/sit.c
464
if (!ns_capable(t->net->user_ns, CAP_NET_ADMIN))
net/ipv6/sit.c
474
err = ipip6_tunnel_del_prl(t, &prl);
net/ipv6/sit.c
478
err = ipip6_tunnel_add_prl(t, &prl, cmd == SIOCCHGPRL);
net/ipv6/sit.c
481
dst_cache_reset(&t->dst_cache);
net/ipv6/sit.c
487
isatap_chksrc(struct sk_buff *skb, const struct iphdr *iph, struct ip_tunnel *t)
net/ipv6/sit.c
493
p = __ipip6_tunnel_locate_prl(t, iph->saddr);
net/ipv6/sit.c
504
ipv6_chk_prefix(addr6, t->dev))
net/ipv6/sit.c
534
struct ip_tunnel *t;
net/ipv6/sit.c
568
t = ipip6_tunnel_lookup(dev_net(skb->dev), skb->dev,
net/ipv6/sit.c
570
if (!t)
net/ipv6/sit.c
575
t->parms.link, iph->protocol);
net/ipv6/sit.c
580
ipv4_redirect(skb, dev_net(skb->dev), t->parms.link,
net/ipv6/sit.c
591
if (t->parms.iph.daddr == 0)
net/ipv6/sit.c
594
if (t->parms.iph.ttl == 0 && type == ICMP_TIME_EXCEEDED)
net/ipv6/sit.c
597
if (time_before(jiffies, t->err_time + IPTUNNEL_ERR_TIMEO))
net/ipv6/sit.c
598
t->err_count++;
net/ipv6/sit.c
600
t->err_count = 1;
net/ipv6/sit.c
601
t->err_time = jiffies;
net/ipv6/sit.c
90
struct ip_tunnel *t = netdev_priv(dev);
net/ipv6/sit.c
92
return net_generic(t->net, sit_net_id);
net/ipv6/tunnel6.c
100
pprev = &t->next) {
net/ipv6/tunnel6.c
101
if (t == handler) {
net/ipv6/tunnel6.c
35
struct xfrm6_tunnel *t;
net/ipv6/tunnel6.c
55
for (; (t = rcu_dereference_protected(*pprev,
net/ipv6/tunnel6.c
57
pprev = &t->next) {
net/ipv6/tunnel6.c
58
if (t->priority > priority)
net/ipv6/tunnel6.c
60
if (t->priority == priority)
net/ipv6/tunnel6.c
79
struct xfrm6_tunnel *t;
net/ipv6/tunnel6.c
98
for (; (t = rcu_dereference_protected(*pprev,
net/ipv6/xfrm6_input.c
226
int xfrm6_rcv_tnl(struct sk_buff *skb, struct ip6_tnl *t)
net/ipv6/xfrm6_input.c
229
0, t);
net/ipv6/xfrm6_input.c
23
struct ip6_tnl *t)
net/ipv6/xfrm6_input.c
25
XFRM_TUNNEL_SKB_CB(skb)->tunnel.ip6 = t;
net/ipv6/xfrm6_protocol.c
237
struct xfrm6_protocol *t;
net/ipv6/xfrm6_protocol.c
252
(t = rcu_dereference_protected(*pprev,
net/ipv6/xfrm6_protocol.c
254
pprev = &t->next) {
net/ipv6/xfrm6_protocol.c
255
if (t->priority < priority)
net/ipv6/xfrm6_protocol.c
257
if (t->priority == priority)
net/ipv6/xfrm6_protocol.c
284
struct xfrm6_protocol *t;
net/ipv6/xfrm6_protocol.c
293
(t = rcu_dereference_protected(*pprev,
net/ipv6/xfrm6_protocol.c
295
pprev = &t->next) {
net/ipv6/xfrm6_protocol.c
296
if (t == handler) {
net/key/af_key.c
1963
struct xfrm_tmpl *t = xp->xfrm_vec + xp->xfrm_nr;
net/key/af_key.c
1974
t->id.proto = rq->sadb_x_ipsecrequest_proto;
net/key/af_key.c
1977
t->mode = mode;
net/key/af_key.c
1982
t->optional = 1;
net/key/af_key.c
1984
t->reqid = rq->sadb_x_ipsecrequest_reqid;
net/key/af_key.c
1985
if (t->reqid > IPSEC_MANUAL_REQID_MAX)
net/key/af_key.c
1986
t->reqid = 0;
net/key/af_key.c
1987
if (!t->reqid && !(t->reqid = gen_reqid(net)))
net/key/af_key.c
1992
if (t->mode == XFRM_MODE_TUNNEL) {
net/key/af_key.c
1998
&t->saddr, &t->id.daddr, &t->encap_family);
net/key/af_key.c
2002
t->encap_family = xp->family;
net/key/af_key.c
2005
t->allalgs = 1;
net/key/af_key.c
2047
const struct xfrm_tmpl *t;
net/key/af_key.c
2053
t = xp->xfrm_vec + i;
net/key/af_key.c
2054
socklen += pfkey_sockaddr_len(t->encap_family);
net/key/af_key.c
2173
const struct xfrm_tmpl *t = xp->xfrm_vec + i;
net/key/af_key.c
2179
if (t->mode == XFRM_MODE_TUNNEL) {
net/key/af_key.c
2180
socklen = pfkey_sockaddr_len(t->encap_family);
net/key/af_key.c
2189
rq->sadb_x_ipsecrequest_proto = t->id.proto;
net/key/af_key.c
2190
if ((mode = pfkey_mode_from_xfrm(t->mode)) < 0)
net/key/af_key.c
2194
if (t->reqid)
net/key/af_key.c
2196
if (t->optional)
net/key/af_key.c
2198
rq->sadb_x_ipsecrequest_reqid = t->reqid;
net/key/af_key.c
2200
if (t->mode == XFRM_MODE_TUNNEL) {
net/key/af_key.c
2202
pfkey_sockaddr_fill(&t->saddr, 0,
net/key/af_key.c
2204
t->encap_family);
net/key/af_key.c
2205
pfkey_sockaddr_fill(&t->id.daddr, 0,
net/key/af_key.c
2207
t->encap_family);
net/key/af_key.c
2896
static inline int aalg_tmpl_set(const struct xfrm_tmpl *t,
net/key/af_key.c
2901
if (id >= sizeof(t->aalgos) * 8)
net/key/af_key.c
2904
return (t->aalgos >> id) & 1;
net/key/af_key.c
2907
static inline int ealg_tmpl_set(const struct xfrm_tmpl *t,
net/key/af_key.c
2912
if (id >= sizeof(t->ealgos) * 8)
net/key/af_key.c
2915
return (t->ealgos >> id) & 1;
net/key/af_key.c
2918
static int count_ah_combs(const struct xfrm_tmpl *t)
net/key/af_key.c
2928
if (aalg_tmpl_set(t, aalg))
net/key/af_key.c
2934
static int count_esp_combs(const struct xfrm_tmpl *t)
net/key/af_key.c
2946
if (!(ealg_tmpl_set(t, ealg)))
net/key/af_key.c
2957
if (aalg_tmpl_set(t, aalg))
net/key/af_key.c
2964
static int dump_ah_combs(struct sk_buff *skb, const struct xfrm_tmpl *t)
net/key/af_key.c
2984
if (aalg_tmpl_set(t, aalg) && aalg->available) {
net/key/af_key.c
3002
static int dump_esp_combs(struct sk_buff *skb, const struct xfrm_tmpl *t)
net/key/af_key.c
3022
if (!(ealg_tmpl_set(t, ealg) && ealg->available))
net/key/af_key.c
3032
if (!(aalg_tmpl_set(t, aalg) && aalg->available))
net/key/af_key.c
3170
static int pfkey_send_acquire(struct xfrm_state *x, struct xfrm_tmpl *t, struct xfrm_policy *xp)
net/key/af_key.c
3193
alg_size = count_ah_combs(t);
net/key/af_key.c
3195
alg_size = count_esp_combs(t);
net/key/af_key.c
3258
alg_size = dump_ah_combs(skb, t);
net/key/af_key.c
3260
alg_size = dump_esp_combs(skb, t);
net/lapb/lapb_timer.c
75
static void lapb_t2timer_expiry(struct timer_list *t)
net/lapb/lapb_timer.c
77
struct lapb_cb *lapb = timer_container_of(lapb, t, t2timer);
net/lapb/lapb_timer.c
95
static void lapb_t1timer_expiry(struct timer_list *t)
net/lapb/lapb_timer.c
97
struct lapb_cb *lapb = timer_container_of(lapb, t, t1timer);
net/llc/llc_c_ac.c
1336
void llc_conn_pf_cycle_tmr_cb(struct timer_list *t)
net/llc/llc_c_ac.c
1338
struct llc_sock *llc = timer_container_of(llc, t,
net/llc/llc_c_ac.c
1344
void llc_conn_busy_tmr_cb(struct timer_list *t)
net/llc/llc_c_ac.c
1346
struct llc_sock *llc = timer_container_of(llc, t,
net/llc/llc_c_ac.c
1352
void llc_conn_ack_tmr_cb(struct timer_list *t)
net/llc/llc_c_ac.c
1354
struct llc_sock *llc = timer_container_of(llc, t, ack_timer.timer);
net/llc/llc_c_ac.c
1359
void llc_conn_rej_tmr_cb(struct timer_list *t)
net/llc/llc_c_ac.c
1361
struct llc_sock *llc = timer_container_of(llc, t,
net/mac80211/agg-rx.c
144
static void sta_rx_agg_session_timer_expired(struct timer_list *t)
net/mac80211/agg-rx.c
146
struct tid_ampdu_rx *tid_rx = timer_container_of(tid_rx, t,
net/mac80211/agg-rx.c
165
static void sta_rx_agg_reorder_timer_expired(struct timer_list *t)
net/mac80211/agg-rx.c
167
struct tid_ampdu_rx *tid_rx = timer_container_of(tid_rx, t,
net/mac80211/agg-tx.c
423
static void sta_addba_resp_timer_expired(struct timer_list *t)
net/mac80211/agg-tx.c
425
struct tid_ampdu_tx *tid_tx = timer_container_of(tid_tx, t,
net/mac80211/agg-tx.c
576
static void sta_tx_agg_session_timer_expired(struct timer_list *t)
net/mac80211/agg-tx.c
578
struct tid_ampdu_tx *tid_tx = timer_container_of(tid_tx, t,
net/mac80211/debugfs_sta.c
1079
#define PFLAG(t, n, a, b) \
net/mac80211/debugfs_sta.c
1081
if (cap[n] & IEEE80211_EHT_##t##_CAP##n##_##a) \
net/mac80211/debugfs_sta.c
699
#define PFLAG(t, n, a, b) \
net/mac80211/debugfs_sta.c
701
if (cap[n] & IEEE80211_HE_##t##_CAP##n##_##a) \
net/mac80211/debugfs_sta.c
705
#define PFLAG_RANGE(t, i, n, s, m, off, fmt) \
net/mac80211/debugfs_sta.c
707
u8 msk = IEEE80211_HE_##t##_CAP##i##_##n##_MASK; \
net/mac80211/debugfs_sta.c
712
#define PFLAG_RANGE_DEFAULT(t, i, n, s, m, off, fmt, a, b) \
net/mac80211/debugfs_sta.c
714
if (cap[i] == IEEE80211_HE_##t ##_CAP##i##_##n##_##a) { \
net/mac80211/debugfs_sta.c
718
PFLAG_RANGE(t, i, n, s, m, off, fmt); \
net/mac80211/ibss.c
1677
static void ieee80211_ibss_timer(struct timer_list *t)
net/mac80211/ibss.c
1680
timer_container_of(sdata, t, u.ibss.timer);
net/mac80211/ieee80211_i.h
2146
void ieee80211_tx_pending(struct tasklet_struct *t);
net/mac80211/ieee80211_i.h
2494
void ieee80211_dynamic_ps_timer(struct timer_list *t);
net/mac80211/ieee80211_i.h
2616
void ieee80211_wake_txqs(struct tasklet_struct *t);
net/mac80211/led.c
258
static void tpt_trig_timer(struct timer_list *t)
net/mac80211/led.c
260
struct tpt_led_trigger *tpt_trig = timer_container_of(tpt_trig, t,
net/mac80211/main.c
467
static void ieee80211_tasklet_handler(struct tasklet_struct *t)
net/mac80211/main.c
469
struct ieee80211_local *local = from_tasklet(local, t, tasklet);
net/mac80211/mesh.c
40
static void ieee80211_mesh_housekeeping_timer(struct timer_list *t)
net/mac80211/mesh.c
43
timer_container_of(sdata, t, u.mesh.housekeeping_timer);
net/mac80211/mesh.c
690
static void ieee80211_mesh_path_timer(struct timer_list *t)
net/mac80211/mesh.c
693
timer_container_of(sdata, t, u.mesh.mesh_path_timer);
net/mac80211/mesh.c
698
static void ieee80211_mesh_path_root_timer(struct timer_list *t)
net/mac80211/mesh.c
701
timer_container_of(sdata, t, u.mesh.mesh_path_root_timer);
net/mac80211/mesh.h
337
void mesh_plink_timer(struct timer_list *t);
net/mac80211/mesh.h
358
void mesh_path_timer(struct timer_list *t);
net/mac80211/mesh_hwmp.c
1296
void mesh_path_timer(struct timer_list *t)
net/mac80211/mesh_hwmp.c
1298
struct mesh_path *mpath = timer_container_of(mpath, t, timer);
net/mac80211/mesh_plink.c
20
#define mod_plink_timer(s, t) (mod_timer(&s->mesh->plink_timer, \
net/mac80211/mesh_plink.c
21
jiffies + msecs_to_jiffies(t)))
net/mac80211/mesh_plink.c
654
void mesh_plink_timer(struct timer_list *t)
net/mac80211/mesh_plink.c
656
struct mesh_sta *mesh = timer_container_of(mesh, t, plink_timer);
net/mac80211/mlme.c
3581
void ieee80211_dynamic_ps_timer(struct timer_list *t)
net/mac80211/mlme.c
3583
struct ieee80211_local *local = timer_container_of(local, t,
net/mac80211/mlme.c
8386
static void ieee80211_sta_timer(struct timer_list *t)
net/mac80211/mlme.c
8389
timer_container_of(sdata, t, u.mgd.timer);
net/mac80211/mlme.c
8721
static void ieee80211_sta_bcn_mon_timer(struct timer_list *t)
net/mac80211/mlme.c
8724
timer_container_of(sdata, t, u.mgd.bcn_mon_timer);
net/mac80211/mlme.c
8780
static void ieee80211_sta_conn_mon_timer(struct timer_list *t)
net/mac80211/mlme.c
8783
timer_container_of(sdata, t, u.mgd.conn_mon_timer);
net/mac80211/ocb.c
149
static void ieee80211_ocb_housekeeping_timer(struct timer_list *t)
net/mac80211/ocb.c
152
timer_container_of(sdata, t, u.ocb.housekeeping_timer);
net/mac80211/sta_info.c
1605
static void sta_info_cleanup(struct timer_list *t)
net/mac80211/sta_info.c
1607
struct ieee80211_local *local = timer_container_of(local, t,
net/mac80211/tx.c
4855
void ieee80211_tx_pending(struct tasklet_struct *t)
net/mac80211/tx.c
4857
struct ieee80211_local *local = from_tasklet(local, t,
net/mac80211/util.c
421
void ieee80211_wake_txqs(struct tasklet_struct *t)
net/mac80211/util.c
423
struct ieee80211_local *local = from_tasklet(local, t,
net/mac802154/ieee802154_i.h
270
struct ieee802154_llsec_table **t);
net/mac802154/main.c
23
static void ieee802154_tasklet_handler(struct tasklet_struct *t)
net/mac802154/main.c
25
struct ieee802154_local *local = from_tasklet(local, t, tasklet);
net/mac802154/mib.c
203
struct ieee802154_llsec_table **t)
net/mac802154/mib.c
209
*t = &sdata->sec.table;
net/mctp/test/route-test.c
114
static void mctp_frag_test_to_desc(const struct mctp_frag_test *t, char *desc)
net/mctp/test/route-test.c
117
t->msgsize, t->mtu, t->n_frags);
net/mctp/test/route-test.c
1218
static void mctp_route_gw_mtu_to_desc(const struct mctp_route_gw_mtu_test *t,
net/mctp/test/route-test.c
1222
t->dev, t->neigh, t->gw, t->dst, t->exp);
net/mctp/test/route-test.c
1489
static void mctp_bind_lookup_desc(const struct mctp_bind_lookup_test *t,
net/mctp/test/route-test.c
1494
t->hdr.src, t->hdr.dest, t->ty, t->net, t->expect);
net/mctp/test/route-test.c
163
static void mctp_rx_input_test_to_desc(const struct mctp_rx_input_test *t,
net/mctp/test/route-test.c
166
sprintf(desc, "{%x,%x,%x,%x}", t->hdr.ver, t->hdr.src, t->hdr.dest,
net/mctp/test/route-test.c
167
t->hdr.flags_seq_tag);
net/mctp/test/route-test.c
263
#define FL_T(t) ((t) & MCTP_HDR_TAG_MASK)
net/mctp/test/route-test.c
274
static void mctp_route_input_sk_to_desc(const struct mctp_route_input_sk_test *t,
net/mctp/test/route-test.c
277
sprintf(desc, "{%x,%x,%x,%x} type %d", t->hdr.ver, t->hdr.src,
net/mctp/test/route-test.c
278
t->hdr.dest, t->hdr.flags_seq_tag, t->type);
net/mctp/test/route-test.c
406
const struct mctp_route_input_sk_reasm_test *t,
net/mctp/test/route-test.c
409
sprintf(desc, "%s", t->name);
net/mctp/test/route-test.c
549
const struct mctp_route_input_sk_keys_test *t,
net/mctp/test/route-test.c
552
sprintf(desc, "%s", t->name);
net/mctp/test/route-test.c
573
struct test_net *t)
net/mctp/test/route-test.c
577
t->msg.data = t->netid;
net/mctp/test/route-test.c
579
__mctp_route_test_init(test, &t->dev, &t->dst, &t->sock, t->netid);
net/mctp/test/route-test.c
581
t->skb = mctp_test_create_skb_data(&hdr, &t->msg);
net/mctp/test/route-test.c
582
KUNIT_ASSERT_NOT_ERR_OR_NULL(test, t->skb);
net/mctp/test/route-test.c
583
mctp_test_skb_set_dev(t->skb, t->dev);
net/mctp/test/route-test.c
588
struct test_net *t)
net/mctp/test/route-test.c
590
__mctp_route_test_fini(test, t->dev, &t->dst, t->sock);
net/mctp/test/route-test.c
638
struct test_net *t)
net/mctp/test/route-test.c
645
t->msg.data = t->netid;
net/mctp/test/route-test.c
647
__mctp_route_test_init(test, &t->dev, &t->dst, &t->sock, t->netid);
net/mctp/test/route-test.c
649
msk = container_of(t->sock->sk, struct mctp_sock, sk);
net/mctp/test/route-test.c
651
t->key = mctp_key_alloc(msk, t->netid, hdr.dest, hdr.src, 1, GFP_KERNEL);
net/mctp/test/route-test.c
652
KUNIT_ASSERT_NOT_ERR_OR_NULL(test, t->key);
net/mctp/test/route-test.c
654
mns = &sock_net(t->sock->sk)->mctp;
net/mctp/test/route-test.c
656
mctp_reserve_tag(&init_net, t->key, msk);
net/mctp/test/route-test.c
659
KUNIT_ASSERT_NOT_ERR_OR_NULL(test, t->key);
net/mctp/test/route-test.c
660
t->skb = mctp_test_create_skb_data(&hdr, &t->msg);
net/mctp/test/route-test.c
661
KUNIT_ASSERT_NOT_ERR_OR_NULL(test, t->skb);
net/mctp/test/route-test.c
662
mctp_test_skb_set_dev(t->skb, t->dev);
net/mctp/test/route-test.c
667
struct test_net *t)
net/mctp/test/route-test.c
669
mctp_key_unref(t->key);
net/mctp/test/route-test.c
670
__mctp_route_test_fini(test, t->dev, &t->dst, t->sock);
net/mctp/test/sock-test.c
294
static void mctp_bind_pair_desc(const struct mctp_bind_pair_test *t, char *desc)
net/mctp/test/sock-test.c
298
if (t->bind1->have_peer)
net/mctp/test/sock-test.c
300
t->bind1->peer_addr, t->bind1->peer_net);
net/mctp/test/sock-test.c
301
if (t->bind2->have_peer)
net/mctp/test/sock-test.c
303
t->bind2->peer_addr, t->bind2->peer_net);
net/mctp/test/sock-test.c
307
t->bind1->bind_addr, t->bind1->bind_type,
net/mctp/test/sock-test.c
308
t->bind1->bind_net, peer1,
net/mctp/test/sock-test.c
309
t->bind2->bind_addr, t->bind2->bind_type,
net/mctp/test/sock-test.c
310
t->bind2->bind_net, peer2, t->error);
net/mptcp/protocol.c
2379
static void mptcp_retransmit_timer(struct timer_list *t)
net/mptcp/protocol.c
2381
struct sock *sk = timer_container_of(sk, t, mptcp_retransmit_timer);
net/mptcp/protocol.c
2397
static void mptcp_tout_timer(struct timer_list *t)
net/mptcp/protocol.c
2400
timer_container_of(icsk, t, mptcp_tout_timer);
net/mptcp/token.c
55
__token_lookup_req(struct token_bucket *t, u32 token)
net/mptcp/token.c
60
hlist_nulls_for_each_entry_rcu(req, pos, &t->req_chain, token_node)
net/mptcp/token.c
68
__token_lookup_msk(struct token_bucket *t, u32 token)
net/mptcp/token.c
73
sk_nulls_for_each_rcu(sk, pos, &t->msk_chain)
net/mptcp/token.c
79
static bool __token_bucket_busy(struct token_bucket *t, u32 token)
net/mptcp/token.c
81
return !token || t->chain_len >= TOKEN_MAX_CHAIN_LEN ||
net/mptcp/token.c
82
__token_lookup_req(t, token) || __token_lookup_msk(t, token);
net/ncsi/ncsi-manage.c
431
static void ncsi_request_timeout(struct timer_list *t)
net/ncsi/ncsi-manage.c
433
struct ncsi_request *nr = timer_container_of(nr, t, timer);
net/ncsi/ncsi-manage.c
89
static void ncsi_channel_monitor(struct timer_list *t)
net/ncsi/ncsi-manage.c
91
struct ncsi_channel *nc = timer_container_of(nc, t, monitor.timer);
net/netfilter/ipset/ip_set_bitmap_gen.h
265
mtype_gc(struct timer_list *t)
net/netfilter/ipset/ip_set_bitmap_gen.h
267
struct mtype *map = timer_container_of(map, t, gc);
net/netfilter/ipset/ip_set_bitmap_gen.h
39
mtype_gc_init(struct ip_set *set, void (*gc)(struct timer_list *t))
net/netfilter/ipset/ip_set_bitmap_ipmac.c
121
u32 t = ext->timeout;
net/netfilter/ipset/ip_set_bitmap_ipmac.c
124
if (t == set->timeout)
net/netfilter/ipset/ip_set_bitmap_ipmac.c
126
t = *timeout;
net/netfilter/ipset/ip_set_bitmap_ipmac.c
127
ip_set_timeout_set(timeout, t);
net/netfilter/ipset/ip_set_bitmap_ipmac.c
135
ip_set_timeout_set(timeout, t);
net/netfilter/ipset/ip_set_bitmap_ipmac.c
137
*timeout = t;
net/netfilter/ipset/ip_set_core.c
326
u32 t;
net/netfilter/ipset/ip_set_core.c
331
t = jiffies_to_msecs(*timeout - jiffies) / MSEC_PER_SEC;
net/netfilter/ipset/ip_set_core.c
333
return t == 0 ? 1 : t;
net/netfilter/ipset/ip_set_hash_gen.h
1022
spin_unlock_bh(&t->hregion[r].lock);
net/netfilter/ipset/ip_set_hash_gen.h
1024
if (atomic_dec_and_test(&t->uref) && atomic_read(&t->ref)) {
net/netfilter/ipset/ip_set_hash_gen.h
1025
pr_debug("Table destroy after resize by add: %p\n", t);
net/netfilter/ipset/ip_set_hash_gen.h
1026
mtype_ahash_destroy(set, t, false);
net/netfilter/ipset/ip_set_hash_gen.h
1038
struct htable *t;
net/netfilter/ipset/ip_set_hash_gen.h
1051
t = rcu_dereference_bh(h->table);
net/netfilter/ipset/ip_set_hash_gen.h
1052
key = HKEY(value, h->initval, t->htable_bits);
net/netfilter/ipset/ip_set_hash_gen.h
1054
atomic_inc(&t->uref);
net/netfilter/ipset/ip_set_hash_gen.h
1057
spin_lock_bh(&t->hregion[r].lock);
net/netfilter/ipset/ip_set_hash_gen.h
1058
n = rcu_dereference_bh(hbucket(t, key));
net/netfilter/ipset/ip_set_hash_gen.h
1077
t->hregion[r].elements--;
net/netfilter/ipset/ip_set_hash_gen.h
1085
if (atomic_read(&t->ref) && ext->target) {
net/netfilter/ipset/ip_set_hash_gen.h
1102
t->hregion[r].ext_size -= ext_size(n->size, dsize);
net/netfilter/ipset/ip_set_hash_gen.h
1103
rcu_assign_pointer(hbucket(t, key), NULL);
net/netfilter/ipset/ip_set_hash_gen.h
1121
t->hregion[r].ext_size -=
net/netfilter/ipset/ip_set_hash_gen.h
1123
rcu_assign_pointer(hbucket(t, key), tmp);
net/netfilter/ipset/ip_set_hash_gen.h
1130
spin_unlock_bh(&t->hregion[r].lock);
net/netfilter/ipset/ip_set_hash_gen.h
1136
if (atomic_dec_and_test(&t->uref) && atomic_read(&t->ref)) {
net/netfilter/ipset/ip_set_hash_gen.h
1137
pr_debug("Table destroy after resize by del: %p\n", t);
net/netfilter/ipset/ip_set_hash_gen.h
1138
mtype_ahash_destroy(set, t, false);
net/netfilter/ipset/ip_set_hash_gen.h
1163
struct htable *t = rcu_dereference_bh(h->table);
net/netfilter/ipset/ip_set_hash_gen.h
1186
key = HKEY(d, h->initval, t->htable_bits);
net/netfilter/ipset/ip_set_hash_gen.h
1187
n = rcu_dereference_bh(hbucket(t, key));
net/netfilter/ipset/ip_set_hash_gen.h
1218
struct htable *t;
net/netfilter/ipset/ip_set_hash_gen.h
1226
t = rcu_dereference_bh(h->table);
net/netfilter/ipset/ip_set_hash_gen.h
1240
key = HKEY(d, h->initval, t->htable_bits);
net/netfilter/ipset/ip_set_hash_gen.h
1241
n = rcu_dereference_bh(hbucket(t, key));
net/netfilter/ipset/ip_set_hash_gen.h
1266
const struct htable *t;
net/netfilter/ipset/ip_set_hash_gen.h
1274
t = rcu_dereference_bh(h->table);
net/netfilter/ipset/ip_set_hash_gen.h
1276
memsize = mtype_ahash_memsize(h, t) + ext_size + set->ext_size;
net/netfilter/ipset/ip_set_hash_gen.h
1277
htable_bits = t->htable_bits;
net/netfilter/ipset/ip_set_hash_gen.h
1331
struct htable *t;
net/netfilter/ipset/ip_set_hash_gen.h
1335
t = ipset_dereference_bh_nfnl(h->table);
net/netfilter/ipset/ip_set_hash_gen.h
1336
atomic_inc(&t->uref);
net/netfilter/ipset/ip_set_hash_gen.h
1337
cb->args[IPSET_CB_PRIVATE] = (unsigned long)t;
net/netfilter/ipset/ip_set_hash_gen.h
1340
t = (struct htable *)cb->args[IPSET_CB_PRIVATE];
net/netfilter/ipset/ip_set_hash_gen.h
1341
if (atomic_dec_and_test(&t->uref) && atomic_read(&t->ref)) {
net/netfilter/ipset/ip_set_hash_gen.h
1343
" by dump: %p\n", t);
net/netfilter/ipset/ip_set_hash_gen.h
1344
mtype_ahash_destroy(set, t, false);
net/netfilter/ipset/ip_set_hash_gen.h
1355
const struct htable *t;
net/netfilter/ipset/ip_set_hash_gen.h
1369
t = (const struct htable *)cb->args[IPSET_CB_PRIVATE];
net/netfilter/ipset/ip_set_hash_gen.h
1372
for (; cb->args[IPSET_CB_ARG0] < jhash_size(t->htable_bits);
net/netfilter/ipset/ip_set_hash_gen.h
1376
n = rcu_dereference(hbucket(t, cb->args[IPSET_CB_ARG0]));
net/netfilter/ipset/ip_set_hash_gen.h
1378
cb->args[IPSET_CB_ARG0], t, n);
net/netfilter/ipset/ip_set_hash_gen.h
1472
struct htable *t;
net/netfilter/ipset/ip_set_hash_gen.h
1567
t = ip_set_alloc(hsize);
net/netfilter/ipset/ip_set_hash_gen.h
1568
if (!t) {
net/netfilter/ipset/ip_set_hash_gen.h
1572
t->hregion = ip_set_alloc(ahash_sizeof_regions(hbits));
net/netfilter/ipset/ip_set_hash_gen.h
1573
if (!t->hregion) {
net/netfilter/ipset/ip_set_hash_gen.h
1574
ip_set_free(t);
net/netfilter/ipset/ip_set_hash_gen.h
1580
spin_lock_init(&t->hregion[i].lock);
net/netfilter/ipset/ip_set_hash_gen.h
1603
t->htable_bits = hbits;
net/netfilter/ipset/ip_set_hash_gen.h
1604
t->maxelem = h->maxelem / ahash_numof_locks(hbits);
net/netfilter/ipset/ip_set_hash_gen.h
1605
RCU_INIT_POINTER(h->table, t);
net/netfilter/ipset/ip_set_hash_gen.h
1637
set->name, jhash_size(t->htable_bits),
net/netfilter/ipset/ip_set_hash_gen.h
1638
t->htable_bits, h->maxelem, set->data, t);
net/netfilter/ipset/ip_set_hash_gen.h
376
mtype_ahash_memsize(const struct htype *h, const struct htable *t)
net/netfilter/ipset/ip_set_hash_gen.h
378
return sizeof(*h) + sizeof(*t) + ahash_sizeof_regions(t->htable_bits);
net/netfilter/ipset/ip_set_hash_gen.h
400
struct htable *t;
net/netfilter/ipset/ip_set_hash_gen.h
404
t = ipset_dereference_nfnl(h->table);
net/netfilter/ipset/ip_set_hash_gen.h
405
for (r = 0; r < ahash_numof_locks(t->htable_bits); r++) {
net/netfilter/ipset/ip_set_hash_gen.h
406
spin_lock_bh(&t->hregion[r].lock);
net/netfilter/ipset/ip_set_hash_gen.h
407
for (i = ahash_bucket_start(r, t->htable_bits);
net/netfilter/ipset/ip_set_hash_gen.h
408
i < ahash_bucket_end(r, t->htable_bits); i++) {
net/netfilter/ipset/ip_set_hash_gen.h
409
n = __ipset_dereference(hbucket(t, i));
net/netfilter/ipset/ip_set_hash_gen.h
415
rcu_assign_pointer(hbucket(t, i), NULL);
net/netfilter/ipset/ip_set_hash_gen.h
418
t->hregion[r].ext_size = 0;
net/netfilter/ipset/ip_set_hash_gen.h
419
t->hregion[r].elements = 0;
net/netfilter/ipset/ip_set_hash_gen.h
420
spin_unlock_bh(&t->hregion[r].lock);
net/netfilter/ipset/ip_set_hash_gen.h
429
mtype_ahash_destroy(struct ip_set *set, struct htable *t, bool ext_destroy)
net/netfilter/ipset/ip_set_hash_gen.h
434
for (i = 0; i < jhash_size(t->htable_bits); i++) {
net/netfilter/ipset/ip_set_hash_gen.h
435
n = (__force struct hbucket *)hbucket(t, i);
net/netfilter/ipset/ip_set_hash_gen.h
444
ip_set_free(t->hregion);
net/netfilter/ipset/ip_set_hash_gen.h
445
ip_set_free(t);
net/netfilter/ipset/ip_set_hash_gen.h
484
mtype_gc_do(struct ip_set *set, struct htype *h, struct htable *t, u32 r)
net/netfilter/ipset/ip_set_hash_gen.h
493
u8 htable_bits = t->htable_bits;
net/netfilter/ipset/ip_set_hash_gen.h
495
spin_lock_bh(&t->hregion[r].lock);
net/netfilter/ipset/ip_set_hash_gen.h
498
n = __ipset_dereference(hbucket(t, i));
net/netfilter/ipset/ip_set_hash_gen.h
518
t->hregion[r].elements--;
net/netfilter/ipset/ip_set_hash_gen.h
524
t->hregion[r].ext_size -=
net/netfilter/ipset/ip_set_hash_gen.h
526
rcu_assign_pointer(hbucket(t, i), NULL);
net/netfilter/ipset/ip_set_hash_gen.h
547
t->hregion[r].ext_size -=
net/netfilter/ipset/ip_set_hash_gen.h
549
rcu_assign_pointer(hbucket(t, i), tmp);
net/netfilter/ipset/ip_set_hash_gen.h
553
spin_unlock_bh(&t->hregion[r].lock);
net/netfilter/ipset/ip_set_hash_gen.h
562
struct htable *t;
net/netfilter/ipset/ip_set_hash_gen.h
571
t = ipset_dereference_set(h->table, set);
net/netfilter/ipset/ip_set_hash_gen.h
572
atomic_inc(&t->uref);
net/netfilter/ipset/ip_set_hash_gen.h
573
numof_locks = ahash_numof_locks(t->htable_bits);
net/netfilter/ipset/ip_set_hash_gen.h
583
mtype_gc_do(set, h, t, r);
net/netfilter/ipset/ip_set_hash_gen.h
585
if (atomic_dec_and_test(&t->uref) && atomic_read(&t->ref)) {
net/netfilter/ipset/ip_set_hash_gen.h
586
pr_debug("Table destroy after resize by expire: %p\n", t);
net/netfilter/ipset/ip_set_hash_gen.h
587
mtype_ahash_destroy(set, t, false);
net/netfilter/ipset/ip_set_hash_gen.h
625
struct htable *t, *orig;
net/netfilter/ipset/ip_set_hash_gen.h
656
t = ip_set_alloc(hsize);
net/netfilter/ipset/ip_set_hash_gen.h
657
if (!t) {
net/netfilter/ipset/ip_set_hash_gen.h
661
t->hregion = ip_set_alloc(ahash_sizeof_regions(htable_bits));
net/netfilter/ipset/ip_set_hash_gen.h
662
if (!t->hregion) {
net/netfilter/ipset/ip_set_hash_gen.h
663
ip_set_free(t);
net/netfilter/ipset/ip_set_hash_gen.h
667
t->htable_bits = htable_bits;
net/netfilter/ipset/ip_set_hash_gen.h
668
t->maxelem = h->maxelem / ahash_numof_locks(htable_bits);
net/netfilter/ipset/ip_set_hash_gen.h
670
spin_lock_init(&t->hregion[i].lock);
net/netfilter/ipset/ip_set_hash_gen.h
704
m = __ipset_dereference(hbucket(t, key));
net/netfilter/ipset/ip_set_hash_gen.h
715
t->hregion[nr].ext_size +=
net/netfilter/ipset/ip_set_hash_gen.h
718
RCU_INIT_POINTER(hbucket(t, key), m);
net/netfilter/ipset/ip_set_hash_gen.h
737
t->hregion[nr].ext_size +=
net/netfilter/ipset/ip_set_hash_gen.h
742
RCU_INIT_POINTER(hbucket(t, key), ht);
net/netfilter/ipset/ip_set_hash_gen.h
747
t->hregion[nr].elements++;
net/netfilter/ipset/ip_set_hash_gen.h
757
rcu_assign_pointer(h->table, t);
net/netfilter/ipset/ip_set_hash_gen.h
763
orig->htable_bits, orig, t->htable_bits, t);
net/netfilter/ipset/ip_set_hash_gen.h
794
mtype_ahash_destroy(set, t, false);
net/netfilter/ipset/ip_set_hash_gen.h
811
const struct htable *t;
net/netfilter/ipset/ip_set_hash_gen.h
816
t = rcu_dereference_bh(h->table);
net/netfilter/ipset/ip_set_hash_gen.h
817
for (r = 0; r < ahash_numof_locks(t->htable_bits); r++) {
net/netfilter/ipset/ip_set_hash_gen.h
818
for (i = ahash_bucket_start(r, t->htable_bits);
net/netfilter/ipset/ip_set_hash_gen.h
819
i < ahash_bucket_end(r, t->htable_bits); i++) {
net/netfilter/ipset/ip_set_hash_gen.h
820
n = rcu_dereference_bh(hbucket(t, i));
net/netfilter/ipset/ip_set_hash_gen.h
831
*ext_size += t->hregion[r].ext_size;
net/netfilter/ipset/ip_set_hash_gen.h
843
struct htable *t;
net/netfilter/ipset/ip_set_hash_gen.h
853
t = rcu_dereference_bh(h->table);
net/netfilter/ipset/ip_set_hash_gen.h
854
key = HKEY(value, h->initval, t->htable_bits);
net/netfilter/ipset/ip_set_hash_gen.h
856
atomic_inc(&t->uref);
net/netfilter/ipset/ip_set_hash_gen.h
857
elements = t->hregion[r].elements;
net/netfilter/ipset/ip_set_hash_gen.h
858
maxelem = t->maxelem;
net/netfilter/ipset/ip_set_hash_gen.h
863
mtype_gc_do(set, h, t, r);
net/netfilter/ipset/ip_set_hash_gen.h
868
for (e = 0; e < ahash_numof_locks(t->htable_bits); e++)
net/netfilter/ipset/ip_set_hash_gen.h
869
elements += t->hregion[e].elements;
net/netfilter/ipset/ip_set_hash_gen.h
875
spin_lock_bh(&t->hregion[r].lock);
net/netfilter/ipset/ip_set_hash_gen.h
876
n = rcu_dereference_bh(hbucket(t, key));
net/netfilter/ipset/ip_set_hash_gen.h
888
t->hregion[r].ext_size +=
net/netfilter/ipset/ip_set_hash_gen.h
929
t->hregion[r].elements--;
net/netfilter/ipset/ip_set_hash_gen.h
960
t->hregion[r].ext_size +=
net/netfilter/ipset/ip_set_hash_gen.h
968
t->hregion[r].elements++;
net/netfilter/ipset/ip_set_hash_gen.h
990
rcu_assign_pointer(hbucket(t, key), n);
net/netfilter/ipset/ip_set_hash_gen.h
996
spin_unlock_bh(&t->hregion[r].lock);
net/netfilter/ipset/ip_set_hash_gen.h
997
if (atomic_read(&t->ref) && ext->target) {
net/netfilter/ipset/ip_set_list_set.c
572
list_set_gc(struct timer_list *t)
net/netfilter/ipset/ip_set_list_set.c
574
struct list_set *map = timer_container_of(map, t, gc);
net/netfilter/ipset/ip_set_list_set.c
586
list_set_gc_init(struct ip_set *set, void (*gc)(struct timer_list *t))
net/netfilter/ipvs/ip_vs_conn.c
101
static void ip_vs_conn_expire(struct timer_list *t);
net/netfilter/ipvs/ip_vs_conn.c
469
unsigned long t = (cp->flags & IP_VS_CONN_F_ONE_PACKET) ?
net/netfilter/ipvs/ip_vs_conn.c
471
mod_timer(&cp->timer, jiffies+t);
net/netfilter/ipvs/ip_vs_conn.c
846
static void ip_vs_conn_expire(struct timer_list *t)
net/netfilter/ipvs/ip_vs_conn.c
848
struct ip_vs_conn *cp = timer_container_of(cp, t, timer);
net/netfilter/ipvs/ip_vs_ctl.c
1331
static void ip_vs_dest_trash_expire(struct timer_list *t)
net/netfilter/ipvs/ip_vs_ctl.c
1333
struct netns_ipvs *ipvs = timer_container_of(ipvs, t,
net/netfilter/ipvs/ip_vs_ctl.c
3149
struct ip_vs_timeout_user t;
net/netfilter/ipvs/ip_vs_ctl.c
3151
__ip_vs_get_timeouts(ipvs, &t);
net/netfilter/ipvs/ip_vs_ctl.c
3152
if (copy_to_user(user, &t, sizeof(t)) != 0)
net/netfilter/ipvs/ip_vs_ctl.c
3864
struct ip_vs_timeout_user t;
net/netfilter/ipvs/ip_vs_ctl.c
3866
__ip_vs_get_timeouts(ipvs, &t);
net/netfilter/ipvs/ip_vs_ctl.c
3869
t.tcp_timeout = nla_get_u32(attrs[IPVS_CMD_ATTR_TIMEOUT_TCP]);
net/netfilter/ipvs/ip_vs_ctl.c
3872
t.tcp_fin_timeout =
net/netfilter/ipvs/ip_vs_ctl.c
3876
t.udp_timeout = nla_get_u32(attrs[IPVS_CMD_ATTR_TIMEOUT_UDP]);
net/netfilter/ipvs/ip_vs_ctl.c
3878
return ip_vs_set_timeout(ipvs, &t);
net/netfilter/ipvs/ip_vs_ctl.c
4090
struct ip_vs_timeout_user t;
net/netfilter/ipvs/ip_vs_ctl.c
4092
__ip_vs_get_timeouts(ipvs, &t);
net/netfilter/ipvs/ip_vs_ctl.c
4095
t.tcp_timeout) ||
net/netfilter/ipvs/ip_vs_ctl.c
4097
t.tcp_fin_timeout))
net/netfilter/ipvs/ip_vs_ctl.c
4101
if (nla_put_u32(msg, IPVS_CMD_ATTR_TIMEOUT_UDP, t.udp_timeout))
net/netfilter/ipvs/ip_vs_lblc.c
292
static void ip_vs_lblc_check_expire(struct timer_list *t)
net/netfilter/ipvs/ip_vs_lblc.c
294
struct ip_vs_lblc_table *tbl = timer_container_of(tbl, t,
net/netfilter/ipvs/ip_vs_lblcr.c
456
static void ip_vs_lblcr_check_expire(struct timer_list *t)
net/netfilter/ipvs/ip_vs_lblcr.c
458
struct ip_vs_lblcr_table *tbl = timer_container_of(tbl, t,
net/netfilter/nf_conntrack_bpf.c
239
const struct btf_type *ncit, *nct, *t;
net/netfilter/nf_conntrack_bpf.c
244
t = btf_type_by_id(reg->btf, reg->btf_id);
net/netfilter/nf_conntrack_bpf.c
245
if (t != nct && t != ncit) {
net/netfilter/nf_conntrack_core.c
2155
struct nf_conntrack_tuple *t,
net/netfilter/nf_conntrack_core.c
2162
t->src.u.tcp.port = nla_get_be16(tb[CTA_PROTO_SRC_PORT]);
net/netfilter/nf_conntrack_core.c
2169
t->dst.u.tcp.port = nla_get_be16(tb[CTA_PROTO_DST_PORT]);
net/netfilter/nf_conntrack_expect.c
73
static void nf_ct_expectation_timed_out(struct timer_list *t)
net/netfilter/nf_conntrack_expect.c
75
struct nf_conntrack_expect *exp = timer_container_of(exp, t, timeout);
net/netfilter/nf_conntrack_netlink.c
1303
struct nf_conntrack_tuple *t,
net/netfilter/nf_conntrack_netlink.c
1310
t->src.u3.ip = nla_get_in_addr(tb[CTA_IP_V4_SRC]);
net/netfilter/nf_conntrack_netlink.c
1317
t->dst.u3.ip = nla_get_in_addr(tb[CTA_IP_V4_DST]);
net/netfilter/nf_conntrack_netlink.c
1324
struct nf_conntrack_tuple *t,
net/netfilter/nf_conntrack_netlink.c
1331
t->src.u3.in6 = nla_get_in6_addr(tb[CTA_IP_V6_SRC]);
net/netfilter/nf_conntrack_netlink.c
1338
t->dst.u3.in6 = nla_get_in6_addr(tb[CTA_IP_V6_DST]);
net/netfilter/nf_conntrack_pptp.c
128
const struct nf_conntrack_tuple *t)
net/netfilter/nf_conntrack_pptp.c
136
nf_ct_dump_tuple(t);
net/netfilter/nf_conntrack_pptp.c
139
h = nf_conntrack_find_get(net, zone, t);
net/netfilter/nf_conntrack_pptp.c
149
exp = nf_ct_expect_find_get(net, zone, t);
net/netfilter/nf_conntrack_pptp.c
165
struct nf_conntrack_tuple t;
net/netfilter/nf_conntrack_pptp.c
170
memcpy(&t, &ct->tuplehash[IP_CT_DIR_ORIGINAL].tuple, sizeof(t));
net/netfilter/nf_conntrack_pptp.c
171
t.dst.protonum = IPPROTO_GRE;
net/netfilter/nf_conntrack_pptp.c
172
t.src.u.gre.key = ct_pptp_info->pns_call_id;
net/netfilter/nf_conntrack_pptp.c
173
t.dst.u.gre.key = ct_pptp_info->pac_call_id;
net/netfilter/nf_conntrack_pptp.c
174
if (!destroy_sibling_or_exp(net, ct, &t))
net/netfilter/nf_conntrack_pptp.c
178
memcpy(&t, &ct->tuplehash[IP_CT_DIR_REPLY].tuple, sizeof(t));
net/netfilter/nf_conntrack_pptp.c
179
t.dst.protonum = IPPROTO_GRE;
net/netfilter/nf_conntrack_pptp.c
180
t.src.u.gre.key = ct_pptp_info->pac_call_id;
net/netfilter/nf_conntrack_pptp.c
181
t.dst.u.gre.key = ct_pptp_info->pns_call_id;
net/netfilter/nf_conntrack_pptp.c
182
if (!destroy_sibling_or_exp(net, ct, &t))
net/netfilter/nf_conntrack_proto_gre.c
103
if (gre_key_cmpfn(km, t) && km == *kmp)
net/netfilter/nf_conntrack_proto_gre.c
114
memcpy(&km->tuple, t, sizeof(*t));
net/netfilter/nf_conntrack_proto_gre.c
61
const struct nf_conntrack_tuple *t)
net/netfilter/nf_conntrack_proto_gre.c
63
return km->tuple.src.l3num == t->src.l3num &&
net/netfilter/nf_conntrack_proto_gre.c
64
!memcmp(&km->tuple.src.u3, &t->src.u3, sizeof(t->src.u3)) &&
net/netfilter/nf_conntrack_proto_gre.c
65
!memcmp(&km->tuple.dst.u3, &t->dst.u3, sizeof(t->dst.u3)) &&
net/netfilter/nf_conntrack_proto_gre.c
66
km->tuple.dst.protonum == t->dst.protonum &&
net/netfilter/nf_conntrack_proto_gre.c
67
km->tuple.dst.u.all == t->dst.u.all;
net/netfilter/nf_conntrack_proto_gre.c
71
static __be16 gre_keymap_lookup(struct net *net, struct nf_conntrack_tuple *t)
net/netfilter/nf_conntrack_proto_gre.c
78
if (gre_key_cmpfn(km, t)) {
net/netfilter/nf_conntrack_proto_gre.c
85
nf_ct_dump_tuple(t);
net/netfilter/nf_conntrack_proto_gre.c
92
struct nf_conntrack_tuple *t)
net/netfilter/nf_conntrack_proto_icmp.c
256
const struct nf_conntrack_tuple *t)
net/netfilter/nf_conntrack_proto_icmp.c
258
if (nla_put_be16(skb, CTA_PROTO_ICMP_ID, t->src.u.icmp.id) ||
net/netfilter/nf_conntrack_proto_icmp.c
259
nla_put_u8(skb, CTA_PROTO_ICMP_TYPE, t->dst.u.icmp.type) ||
net/netfilter/nf_conntrack_proto_icmp.c
260
nla_put_u8(skb, CTA_PROTO_ICMP_CODE, t->dst.u.icmp.code))
net/netfilter/nf_conntrack_proto_icmpv6.c
233
const struct nf_conntrack_tuple *t)
net/netfilter/nf_conntrack_proto_icmpv6.c
235
if (nla_put_be16(skb, CTA_PROTO_ICMPV6_ID, t->src.u.icmp.id) ||
net/netfilter/nf_conntrack_proto_icmpv6.c
236
nla_put_u8(skb, CTA_PROTO_ICMPV6_TYPE, t->dst.u.icmp.type) ||
net/netfilter/nf_conntrack_proto_icmpv6.c
237
nla_put_u8(skb, CTA_PROTO_ICMPV6_CODE, t->dst.u.icmp.code))
net/netfilter/nf_conntrack_sip.c
1013
const struct sdp_media_type *t;
net/netfilter/nf_conntrack_sip.c
1017
t = &sdp_media_types[i];
net/netfilter/nf_conntrack_sip.c
1018
if (matchlen < t->len ||
net/netfilter/nf_conntrack_sip.c
1019
strncmp(dptr + matchoff, t->name, t->len))
net/netfilter/nf_conntrack_sip.c
1021
return t;
net/netfilter/nf_conntrack_sip.c
1041
const struct sdp_media_type *t;
net/netfilter/nf_conntrack_sip.c
1075
t = sdp_media_type(*dptr, mediaoff, medialen);
net/netfilter/nf_conntrack_sip.c
1076
if (!t) {
net/netfilter/nf_conntrack_sip.c
1080
mediaoff += t->len;
net/netfilter/nf_conntrack_sip.c
1081
medialen -= t->len;
net/netfilter/nf_conntrack_sip.c
1109
&rtp_addr, htons(port), t->class,
net/netfilter/nf_conntrack_timeout.c
136
struct nf_ct_timeout *t;
net/netfilter/nf_conntrack_timeout.c
138
t = rcu_dereference(timeout_ext->timeout);
net/netfilter/nf_conntrack_timeout.c
139
if (t)
net/netfilter/nf_conntrack_timeout.c
140
h->timeout_put(t);
net/netfilter/nf_conntrack_timeout.c
33
const struct nf_ct_timeout *t;
net/netfilter/nf_conntrack_timeout.c
35
t = rcu_access_pointer(timeout_ext->timeout);
net/netfilter/nf_conntrack_timeout.c
37
if (!timeout || t == timeout)
net/netfilter/nf_nat_core.c
100
if (t->dst.protonum == IPPROTO_TCP ||
net/netfilter/nf_nat_core.c
101
t->dst.protonum == IPPROTO_UDP ||
net/netfilter/nf_nat_core.c
102
t->dst.protonum == IPPROTO_UDPLITE ||
net/netfilter/nf_nat_core.c
103
t->dst.protonum == IPPROTO_SCTP)
net/netfilter/nf_nat_core.c
104
fl6->fl6_dport = t->dst.u.all;
net/netfilter/nf_nat_core.c
110
fl6->saddr = t->src.u3.in6;
net/netfilter/nf_nat_core.c
111
if (t->dst.protonum == IPPROTO_TCP ||
net/netfilter/nf_nat_core.c
112
t->dst.protonum == IPPROTO_UDP ||
net/netfilter/nf_nat_core.c
113
t->dst.protonum == IPPROTO_UDPLITE ||
net/netfilter/nf_nat_core.c
114
t->dst.protonum == IPPROTO_SCTP)
net/netfilter/nf_nat_core.c
115
fl6->fl6_sport = t->src.u.all;
net/netfilter/nf_nat_core.c
391
static bool nf_nat_inet_in_range(const struct nf_conntrack_tuple *t,
net/netfilter/nf_nat_core.c
394
if (t->src.l3num == NFPROTO_IPV4)
net/netfilter/nf_nat_core.c
395
return ntohl(t->src.u3.ip) >= ntohl(range->min_addr.ip) &&
net/netfilter/nf_nat_core.c
396
ntohl(t->src.u3.ip) <= ntohl(range->max_addr.ip);
net/netfilter/nf_nat_core.c
398
return ipv6_addr_cmp(&t->src.u3.in6, &range->min_addr.in6) >= 0 &&
net/netfilter/nf_nat_core.c
399
ipv6_addr_cmp(&t->src.u3.in6, &range->max_addr.in6) <= 0;
net/netfilter/nf_nat_core.c
456
const struct nf_conntrack_tuple *t;
net/netfilter/nf_nat_core.c
458
t = &ct->tuplehash[IP_CT_DIR_ORIGINAL].tuple;
net/netfilter/nf_nat_core.c
459
return (t->dst.protonum == tuple->dst.protonum &&
net/netfilter/nf_nat_core.c
460
nf_inet_addr_cmp(&t->src.u3, &tuple->src.u3) &&
net/netfilter/nf_nat_core.c
461
t->src.u.all == tuple->src.u.all);
net/netfilter/nf_nat_core.c
64
const struct nf_conntrack_tuple *t = &ct->tuplehash[dir].tuple;
net/netfilter/nf_nat_core.c
68
fl4->daddr = t->dst.u3.ip;
net/netfilter/nf_nat_core.c
69
if (t->dst.protonum == IPPROTO_TCP ||
net/netfilter/nf_nat_core.c
70
t->dst.protonum == IPPROTO_UDP ||
net/netfilter/nf_nat_core.c
71
t->dst.protonum == IPPROTO_UDPLITE ||
net/netfilter/nf_nat_core.c
72
t->dst.protonum == IPPROTO_SCTP)
net/netfilter/nf_nat_core.c
73
fl4->fl4_dport = t->dst.u.all;
net/netfilter/nf_nat_core.c
79
fl4->saddr = t->src.u3.ip;
net/netfilter/nf_nat_core.c
80
if (t->dst.protonum == IPPROTO_TCP ||
net/netfilter/nf_nat_core.c
81
t->dst.protonum == IPPROTO_UDP ||
net/netfilter/nf_nat_core.c
82
t->dst.protonum == IPPROTO_UDPLITE ||
net/netfilter/nf_nat_core.c
83
t->dst.protonum == IPPROTO_SCTP)
net/netfilter/nf_nat_core.c
84
fl4->fl4_sport = t->src.u.all;
net/netfilter/nf_nat_core.c
95
const struct nf_conntrack_tuple *t = &ct->tuplehash[dir].tuple;
net/netfilter/nf_nat_core.c
99
fl6->daddr = t->dst.u3.in6;
net/netfilter/nf_nat_proto.c
36
const struct nf_conntrack_tuple *t,
net/netfilter/nf_nat_proto.c
405
const struct nf_conntrack_tuple *t,
net/netfilter/nf_nat_proto.c
413
newip = t->src.u3.ip;
net/netfilter/nf_nat_proto.c
416
newip = t->dst.u3.ip;
net/netfilter/nf_nat_proto.c
423
const struct nf_conntrack_tuple *t,
net/netfilter/nf_nat_proto.c
432
newip = &t->src.u3.in6;
net/netfilter/nf_nat_proto.c
435
newip = &t->dst.u3.in6;
net/netfilter/nf_nat_proto.c
444
const struct nf_conntrack_tuple *t,
net/netfilter/nf_nat_proto.c
447
switch (t->src.l3num) {
net/netfilter/nf_nat_proto.c
449
nf_nat_ipv4_csum_update(skb, iphdroff, check, t, maniptype);
net/netfilter/nf_nat_proto.c
452
nf_nat_ipv6_csum_update(skb, iphdroff, check, t, maniptype);
net/netfilter/nf_tables_api.c
2282
const struct nft_trans *t = &trans->nft_trans_binding.nft_trans;
net/netfilter/nf_tables_api.c
2290
lockdep_commit_lock_is_held(t->net));
net/netfilter/nfnetlink_cttimeout.c
524
static void ctnl_timeout_put(struct nf_ct_timeout *t)
net/netfilter/nfnetlink_cttimeout.c
527
container_of(t, struct ctnl_timeout, timeout);
net/netfilter/nfnetlink_hook.c
61
static struct nlattr *nfnl_start_info_type(struct sk_buff *nlskb, enum nfnl_hook_chaintype t)
net/netfilter/nfnetlink_hook.c
69
ret = nla_put_be32(nlskb, NFNLA_HOOK_INFO_TYPE, htonl(t));
net/netfilter/nfnetlink_log.c
166
static void nfulnl_timer(struct timer_list *t);
net/netfilter/nfnetlink_log.c
391
nfulnl_timer(struct timer_list *t)
net/netfilter/nfnetlink_log.c
393
struct nfulnl_instance *inst = timer_container_of(inst, t, timer);
net/netfilter/nft_compat.c
186
static void target_compat_from_user(struct xt_target *t, void *in, void *out)
net/netfilter/nft_compat.c
190
memcpy(out, in, t->targetsize);
net/netfilter/nft_compat.c
191
pad = XT_ALIGN(t->targetsize) - t->targetsize;
net/netfilter/nft_compat.c
193
memset(out + t->targetsize, 0, pad);
net/netfilter/nft_log.c
75
static int nft_log_modprobe(struct net *net, enum nf_log_type t)
net/netfilter/nft_log.c
77
switch (t) {
net/netfilter/nft_set_pipapo.c
373
unsigned long t = bitset & -bitset;
net/netfilter/nft_set_pipapo.c
391
bitset ^= t;
net/netfilter/x_tables.c
1154
void xt_compat_target_from_user(struct xt_entry_target *t, void **dstptr,
net/netfilter/x_tables.c
1157
const struct xt_target *target = t->u.kernel.target;
net/netfilter/x_tables.c
1158
struct compat_xt_entry_target *ct = (struct compat_xt_entry_target *)t;
net/netfilter/x_tables.c
1161
char name[sizeof(t->u.user.name)];
net/netfilter/x_tables.c
1163
t = *dstptr;
net/netfilter/x_tables.c
1164
memcpy(t, ct, sizeof(*ct));
net/netfilter/x_tables.c
1166
target->compat_from_user(t->data, ct->data);
net/netfilter/x_tables.c
1168
unsafe_memcpy(t->data, ct->data, tsize - sizeof(*ct),
net/netfilter/x_tables.c
1172
t->u.user.target_size = tsize;
net/netfilter/x_tables.c
1175
strscpy_pad(t->u.user.name, name, sizeof(t->u.user.name));
net/netfilter/x_tables.c
1182
int xt_compat_target_to_user(const struct xt_entry_target *t,
net/netfilter/x_tables.c
1185
const struct xt_target *target = t->u.kernel.target;
net/netfilter/x_tables.c
1188
u_int16_t tsize = t->u.user.target_size - off;
net/netfilter/x_tables.c
1190
if (XT_OBJ_TO_USER(ct, t, target, tsize))
net/netfilter/x_tables.c
1194
if (target->compat_to_user((void __user *)ct->data, t->data))
net/netfilter/x_tables.c
1197
if (COMPAT_XT_DATA_TO_USER(ct, t, target, tsize - sizeof(*ct)))
net/netfilter/x_tables.c
1243
struct xt_table *t;
net/netfilter/x_tables.c
1246
list_for_each_entry(t, &xt_net->tables[af], list) {
net/netfilter/x_tables.c
1247
if (strcmp(t->name, name) == 0) {
net/netfilter/x_tables.c
1249
return t;
net/netfilter/x_tables.c
1264
struct xt_table *t;
net/netfilter/x_tables.c
1267
list_for_each_entry(t, &xt_net->tables[af], list)
net/netfilter/x_tables.c
1268
if (strcmp(t->name, name) == 0 && try_module_get(t->me))
net/netfilter/x_tables.c
1269
return t;
net/netfilter/x_tables.c
1294
list_for_each_entry(t, &xt_net->tables[af], list)
net/netfilter/x_tables.c
1295
if (strcmp(t->name, name) == 0 && owner == t->me)
net/netfilter/x_tables.c
1296
return t;
net/netfilter/x_tables.c
1308
struct xt_table *t = xt_find_table_lock(net, af, name);
net/netfilter/x_tables.c
1311
if (IS_ERR(t)) {
net/netfilter/x_tables.c
1315
t = xt_find_table_lock(net, af, name);
net/netfilter/x_tables.c
1319
return t;
net/netfilter/x_tables.c
1482
struct xt_table *t, *table;
net/netfilter/x_tables.c
1494
list_for_each_entry(t, &xt_net->tables[table->af], list) {
net/netfilter/x_tables.c
1495
if (strcmp(t->name, table->name) == 0) {
net/netfilter/x_tables.c
1791
struct xt_template *t;
net/netfilter/x_tables.c
1795
list_for_each_entry(t, &xt_templates[af], list) {
net/netfilter/x_tables.c
1796
if (WARN_ON_ONCE(strcmp(table->name, t->name) == 0))
net/netfilter/x_tables.c
1801
t = kzalloc_obj(*t);
net/netfilter/x_tables.c
1802
if (!t)
net/netfilter/x_tables.c
1805
BUILD_BUG_ON(sizeof(t->name) != sizeof(table->name));
net/netfilter/x_tables.c
1807
strscpy(t->name, table->name, sizeof(t->name));
net/netfilter/x_tables.c
1808
t->table_init = table_init;
net/netfilter/x_tables.c
1809
t->me = table->me;
net/netfilter/x_tables.c
1810
list_add(&t->list, &xt_templates[af]);
net/netfilter/x_tables.c
1820
struct xt_template *t;
net/netfilter/x_tables.c
1824
list_for_each_entry(t, &xt_templates[af], list) {
net/netfilter/x_tables.c
1825
if (strcmp(table->name, t->name))
net/netfilter/x_tables.c
1828
list_del(&t->list);
net/netfilter/x_tables.c
1830
kfree(t);
net/netfilter/x_tables.c
248
struct xt_target *t;
net/netfilter/x_tables.c
255
list_for_each_entry(t, &xt[af].target, list) {
net/netfilter/x_tables.c
256
if (strcmp(t->name, name) == 0) {
net/netfilter/x_tables.c
257
if (t->revision == revision) {
net/netfilter/x_tables.c
258
if (try_module_get(t->me)) {
net/netfilter/x_tables.c
260
return t;
net/netfilter/x_tables.c
340
int xt_target_to_user(const struct xt_entry_target *t,
net/netfilter/x_tables.c
343
return XT_OBJ_TO_USER(u, t, target, 0) ||
net/netfilter/x_tables.c
344
XT_DATA_TO_USER(u, t, target);
net/netfilter/x_tables.c
372
const struct xt_target *t;
net/netfilter/x_tables.c
376
list_for_each_entry(t, &xt[af].target, list) {
net/netfilter/x_tables.c
377
if (strcmp(t->name, name) == 0) {
net/netfilter/x_tables.c
378
if (t->revision > *bestp)
net/netfilter/x_tables.c
379
*bestp = t->revision;
net/netfilter/x_tables.c
380
if (t->revision == revision)
net/netfilter/x_tables.c
822
struct compat_xt_entry_target t;
net/netfilter/x_tables.c
827
struct compat_xt_entry_target t;
net/netfilter/x_tables.c
836
const struct compat_xt_entry_target *t;
net/netfilter/x_tables.c
842
if (target_offset + sizeof(*t) > next_offset)
net/netfilter/x_tables.c
845
t = (void *)(e + target_offset);
net/netfilter/x_tables.c
846
if (t->u.target_size < sizeof(*t))
net/netfilter/x_tables.c
849
if (target_offset + t->u.target_size > next_offset)
net/netfilter/x_tables.c
852
if (strcmp(t->u.user.name, XT_STANDARD_TARGET) == 0) {
net/netfilter/x_tables.c
853
const struct compat_xt_standard_target *st = (const void *)t;
net/netfilter/x_tables.c
860
} else if (strcmp(t->u.user.name, XT_ERROR_TARGET) == 0) {
net/netfilter/x_tables.c
861
const struct compat_xt_error_target *et = (const void *)t;
net/netfilter/x_tables.c
863
if (!error_tg_ok(t->u.target_size, sizeof(*et),
net/netfilter/x_tables.c
929
const struct xt_entry_target *t;
net/netfilter/x_tables.c
936
if (target_offset + sizeof(*t) > next_offset)
net/netfilter/x_tables.c
939
t = (void *)(e + target_offset);
net/netfilter/x_tables.c
940
if (t->u.target_size < sizeof(*t))
net/netfilter/x_tables.c
943
if (target_offset + t->u.target_size > next_offset)
net/netfilter/x_tables.c
946
if (strcmp(t->u.user.name, XT_STANDARD_TARGET) == 0) {
net/netfilter/x_tables.c
947
const struct xt_standard_target *st = (const void *)t;
net/netfilter/x_tables.c
954
} else if (strcmp(t->u.user.name, XT_ERROR_TARGET) == 0) {
net/netfilter/x_tables.c
955
const struct xt_error_target *et = (const void *)t;
net/netfilter/x_tables.c
957
if (!error_tg_ok(t->u.target_size, sizeof(*et),
net/netfilter/xt_HMARK.c
100
t->proto = nf_ct_protonum(ct);
net/netfilter/xt_HMARK.c
101
if (t->proto != IPPROTO_ICMP) {
net/netfilter/xt_HMARK.c
102
t->uports.b16.src = otuple->src.u.all;
net/netfilter/xt_HMARK.c
103
t->uports.b16.dst = rtuple->src.u.all;
net/netfilter/xt_HMARK.c
104
hmark_swap_ports(&t->uports, info);
net/netfilter/xt_HMARK.c
116
hmark_hash(struct hmark_tuple *t, const struct xt_hmark_info *info)
net/netfilter/xt_HMARK.c
119
u32 src = ntohl(t->src);
net/netfilter/xt_HMARK.c
120
u32 dst = ntohl(t->dst);
net/netfilter/xt_HMARK.c
125
hash = jhash_3words(src, dst, t->uports.v32, info->hashrnd);
net/netfilter/xt_HMARK.c
126
hash = hash ^ (t->proto & info->proto_mask);
net/netfilter/xt_HMARK.c
133
struct hmark_tuple *t, const struct xt_hmark_info *info)
net/netfilter/xt_HMARK.c
137
protoff = proto_ports_offset(t->proto);
net/netfilter/xt_HMARK.c
142
if (skb_copy_bits(skb, nhoff, &t->uports, sizeof(t->uports)) < 0)
net/netfilter/xt_HMARK.c
145
hmark_swap_ports(&t->uports, info);
net/netfilter/xt_HMARK.c
165
hmark_pkt_set_htuple_ipv6(const struct sk_buff *skb, struct hmark_tuple *t,
net/netfilter/xt_HMARK.c
193
t->src = hmark_addr6_mask(ip6->saddr.s6_addr32, info->src_mask.ip6);
net/netfilter/xt_HMARK.c
194
t->dst = hmark_addr6_mask(ip6->daddr.s6_addr32, info->dst_mask.ip6);
net/netfilter/xt_HMARK.c
199
t->proto = nexthdr;
net/netfilter/xt_HMARK.c
200
if (t->proto == IPPROTO_ICMPV6)
net/netfilter/xt_HMARK.c
206
hmark_set_tuple_ports(skb, nhoff, t, info);
net/netfilter/xt_HMARK.c
214
struct hmark_tuple t;
net/netfilter/xt_HMARK.c
216
memset(&t, 0, sizeof(struct hmark_tuple));
net/netfilter/xt_HMARK.c
219
if (hmark_ct_set_htuple(skb, &t, info) < 0)
net/netfilter/xt_HMARK.c
222
if (hmark_pkt_set_htuple_ipv6(skb, &t, info) < 0)
net/netfilter/xt_HMARK.c
226
skb->mark = hmark_hash(&t, info);
net/netfilter/xt_HMARK.c
250
hmark_pkt_set_htuple_ipv4(const struct sk_buff *skb, struct hmark_tuple *t,
net/netfilter/xt_HMARK.c
266
t->src = ip->saddr & info->src_mask.ip;
net/netfilter/xt_HMARK.c
267
t->dst = ip->daddr & info->dst_mask.ip;
net/netfilter/xt_HMARK.c
272
t->proto = ip->protocol;
net/netfilter/xt_HMARK.c
275
if (t->proto == IPPROTO_ICMP)
net/netfilter/xt_HMARK.c
282
hmark_set_tuple_ports(skb, (ip->ihl * 4) + nhoff, t, info);
net/netfilter/xt_HMARK.c
291
struct hmark_tuple t;
net/netfilter/xt_HMARK.c
293
memset(&t, 0, sizeof(struct hmark_tuple));
net/netfilter/xt_HMARK.c
296
if (hmark_ct_set_htuple(skb, &t, info) < 0)
net/netfilter/xt_HMARK.c
299
if (hmark_pkt_set_htuple_ipv4(skb, &t, info) < 0)
net/netfilter/xt_HMARK.c
303
skb->mark = hmark_hash(&t, info);
net/netfilter/xt_HMARK.c
77
hmark_ct_set_htuple(const struct sk_buff *skb, struct hmark_tuple *t,
net/netfilter/xt_HMARK.c
92
t->src = hmark_addr_mask(otuple->src.l3num, otuple->src.u3.ip6,
net/netfilter/xt_HMARK.c
94
t->dst = hmark_addr_mask(otuple->src.l3num, rtuple->src.u3.ip6,
net/netfilter/xt_IDLETIMER.c
101
static void idletimer_tg_expired(struct timer_list *t)
net/netfilter/xt_IDLETIMER.c
103
struct idletimer_tg *timer = timer_container_of(timer, t, timer);
net/netfilter/xt_LED.c
73
static void led_timeout_callback(struct timer_list *t)
net/netfilter/xt_LED.c
76
t,
net/netfilter/xt_recent.c
144
static void recent_entry_remove(struct recent_table *t, struct recent_entry *e)
net/netfilter/xt_recent.c
149
t->entries--;
net/netfilter/xt_recent.c
155
static void recent_entry_reap(struct recent_table *t, unsigned long time,
net/netfilter/xt_recent.c
163
e = list_entry(t->lru_list.next, struct recent_entry, lru_list);
net/netfilter/xt_recent.c
175
recent_entry_remove(t, e);
net/netfilter/xt_recent.c
179
recent_entry_init(struct recent_table *t, const union nf_inet_addr *addr,
net/netfilter/xt_recent.c
183
unsigned int nstamps_max = t->nstamps_max_mask;
net/netfilter/xt_recent.c
185
if (t->entries >= ip_list_tot) {
net/netfilter/xt_recent.c
186
e = list_entry(t->lru_list.next, struct recent_entry, lru_list);
net/netfilter/xt_recent.c
187
recent_entry_remove(t, e);
net/netfilter/xt_recent.c
201
list_add_tail(&e->list, &t->iphash[recent_entry_hash4(addr)]);
net/netfilter/xt_recent.c
203
list_add_tail(&e->list, &t->iphash[recent_entry_hash6(addr)]);
net/netfilter/xt_recent.c
204
list_add_tail(&e->lru_list, &t->lru_list);
net/netfilter/xt_recent.c
205
t->entries++;
net/netfilter/xt_recent.c
209
static void recent_entry_update(struct recent_table *t, struct recent_entry *e)
net/netfilter/xt_recent.c
211
e->index &= t->nstamps_max_mask;
net/netfilter/xt_recent.c
215
list_move_tail(&e->lru_list, &t->lru_list);
net/netfilter/xt_recent.c
221
struct recent_table *t;
net/netfilter/xt_recent.c
223
list_for_each_entry(t, &recent_net->tables, list)
net/netfilter/xt_recent.c
224
if (!strcmp(t->name, name))
net/netfilter/xt_recent.c
225
return t;
net/netfilter/xt_recent.c
229
static void recent_table_flush(struct recent_table *t)
net/netfilter/xt_recent.c
235
list_for_each_entry_safe(e, next, &t->iphash[i], list)
net/netfilter/xt_recent.c
236
recent_entry_remove(t, e);
net/netfilter/xt_recent.c
245
struct recent_table *t;
net/netfilter/xt_recent.c
277
t = recent_table_lookup(recent_net, info->name);
net/netfilter/xt_recent.c
279
nf_inet_addr_mask(&addr, &addr_mask, &t->mask);
net/netfilter/xt_recent.c
281
e = recent_entry_lookup(t, &addr_mask, xt_family(par),
net/netfilter/xt_recent.c
286
e = recent_entry_init(t, &addr_mask, xt_family(par), ttl);
net/netfilter/xt_recent.c
296
recent_entry_remove(t, e);
net/netfilter/xt_recent.c
313
recent_entry_reap(t, time, e,
net/netfilter/xt_recent.c
319
recent_entry_update(t, e);
net/netfilter/xt_recent.c
336
struct recent_table *t;
net/netfilter/xt_recent.c
380
t = recent_table_lookup(recent_net, info->name);
net/netfilter/xt_recent.c
381
if (t != NULL) {
net/netfilter/xt_recent.c
382
if (nstamp_mask > t->nstamps_max_mask) {
net/netfilter/xt_recent.c
384
recent_table_flush(t);
net/netfilter/xt_recent.c
385
t->nstamps_max_mask = nstamp_mask;
net/netfilter/xt_recent.c
389
t->refcnt++;
net/netfilter/xt_recent.c
394
t = kvzalloc_flex(*t, iphash, ip_list_hash_size);
net/netfilter/xt_recent.c
395
if (t == NULL) {
net/netfilter/xt_recent.c
399
t->refcnt = 1;
net/netfilter/xt_recent.c
400
t->nstamps_max_mask = nstamp_mask;
net/netfilter/xt_recent.c
402
memcpy(&t->mask, &info->mask, sizeof(t->mask));
net/netfilter/xt_recent.c
403
strcpy(t->name, info->name);
net/netfilter/xt_recent.c
404
INIT_LIST_HEAD(&t->lru_list);
net/netfilter/xt_recent.c
406
INIT_LIST_HEAD(&t->iphash[i]);
net/netfilter/xt_recent.c
411
recent_table_free(t);
net/netfilter/xt_recent.c
415
pde = proc_create_data(t->name, ip_list_perms, recent_net->xt_recent,
net/netfilter/xt_recent.c
416
&recent_mt_proc_ops, t);
net/netfilter/xt_recent.c
418
recent_table_free(t);
net/netfilter/xt_recent.c
425
list_add_tail(&t->list, &recent_net->tables);
net/netfilter/xt_recent.c
455
struct recent_table *t;
net/netfilter/xt_recent.c
458
t = recent_table_lookup(recent_net, info->name);
net/netfilter/xt_recent.c
459
if (--t->refcnt == 0) {
net/netfilter/xt_recent.c
461
list_del(&t->list);
net/netfilter/xt_recent.c
465
remove_proc_entry(t->name, recent_net->xt_recent);
net/netfilter/xt_recent.c
467
recent_table_flush(t);
net/netfilter/xt_recent.c
468
recent_table_free(t);
net/netfilter/xt_recent.c
483
const struct recent_table *t = st->table;
net/netfilter/xt_recent.c
490
list_for_each_entry(e, &t->iphash[st->bucket], list)
net/netfilter/xt_recent.c
499
const struct recent_table *t = st->table;
net/netfilter/xt_recent.c
504
while (head == &t->iphash[st->bucket]) {
net/netfilter/xt_recent.c
507
head = t->iphash[st->bucket].next;
net/netfilter/xt_recent.c
522
const struct recent_table *t = st->table;
net/netfilter/xt_recent.c
525
i = (e->index - 1) & t->nstamps_max_mask;
net/netfilter/xt_recent.c
562
struct recent_table *t = pde_data(file_inode(file));
net/netfilter/xt_recent.c
583
recent_table_flush(t);
net/netfilter/xt_recent.c
611
e = recent_entry_lookup(t, &addr, family, 0);
net/netfilter/xt_recent.c
614
recent_entry_init(t, &addr, family, 0);
net/netfilter/xt_recent.c
617
recent_entry_update(t, e);
net/netfilter/xt_recent.c
619
recent_entry_remove(t, e);
net/netfilter/xt_recent.c
648
struct recent_table *t;
net/netfilter/xt_recent.c
655
list_for_each_entry(t, &recent_net->tables, list)
net/netfilter/xt_recent.c
656
remove_proc_entry(t->name, recent_net->xt_recent);
net/netfilter/xt_set.c
38
#define ADT_OPT(n, f, d, fs, cfs, t, p, b, po, bo) \
net/netfilter/xt_set.c
44
.ext.timeout = t, \
net/netrom/af_netrom.c
241
static void nr_destroy_timer(struct timer_list *t)
net/netrom/af_netrom.c
243
struct sock *sk = timer_container_of(sk, t, sk_timer);
net/netrom/nr_route.c
511
struct nr_node *t;
net/netrom/nr_route.c
518
nr_node_for_each_safe(t, node2t, &nr_node_list) {
net/netrom/nr_route.c
519
nr_node_lock(t);
net/netrom/nr_route.c
520
for (i = 0; i < t->count; i++) {
net/netrom/nr_route.c
521
if (t->routes[i].neighbour == s) {
net/netrom/nr_route.c
522
t->count--;
net/netrom/nr_route.c
526
t->routes[0] = t->routes[1];
net/netrom/nr_route.c
529
t->routes[1] = t->routes[2];
net/netrom/nr_route.c
537
if (t->count <= 0)
net/netrom/nr_route.c
538
nr_remove_node_locked(t);
net/netrom/nr_route.c
539
nr_node_unlock(t);
net/netrom/nr_route.c
970
struct nr_node *t = NULL;
net/netrom/nr_route.c
975
nr_node_for_each_safe(t, nodet, &nr_node_list) {
net/netrom/nr_route.c
976
nr_node_lock(t);
net/netrom/nr_route.c
977
nr_remove_node_locked(t);
net/netrom/nr_route.c
978
nr_node_unlock(t);
net/netrom/nr_timer.c
112
static void nr_heartbeat_expiry(struct timer_list *t)
net/netrom/nr_timer.c
114
struct sock *sk = timer_container_of(sk, t, sk_timer);
net/netrom/nr_timer.c
153
static void nr_t2timer_expiry(struct timer_list *t)
net/netrom/nr_timer.c
155
struct nr_sock *nr = timer_container_of(nr, t, t2timer);
net/netrom/nr_timer.c
167
static void nr_t4timer_expiry(struct timer_list *t)
net/netrom/nr_timer.c
169
struct nr_sock *nr = timer_container_of(nr, t, t4timer);
net/netrom/nr_timer.c
178
static void nr_idletimer_expiry(struct timer_list *t)
net/netrom/nr_timer.c
180
struct nr_sock *nr = timer_container_of(nr, t, idletimer);
net/netrom/nr_timer.c
207
static void nr_t1timer_expiry(struct timer_list *t)
net/netrom/nr_timer.c
209
struct nr_sock *nr = timer_container_of(nr, t, t1timer);
net/nfc/core.c
1011
static void nfc_check_pres_timeout(struct timer_list *t)
net/nfc/core.c
1013
struct nfc_dev *dev = timer_container_of(dev, t, check_pres_timer);
net/nfc/hci/core.c
442
static void nfc_hci_cmd_timeout(struct timer_list *t)
net/nfc/hci/core.c
444
struct nfc_hci_dev *hdev = timer_container_of(hdev, t, cmd_timer);
net/nfc/hci/llc_shdlc.c
565
static void llc_shdlc_connect_timeout(struct timer_list *t)
net/nfc/hci/llc_shdlc.c
567
struct llc_shdlc *shdlc = timer_container_of(shdlc, t, connect_timer);
net/nfc/hci/llc_shdlc.c
572
static void llc_shdlc_t1_timeout(struct timer_list *t)
net/nfc/hci/llc_shdlc.c
574
struct llc_shdlc *shdlc = timer_container_of(shdlc, t, t1_timer);
net/nfc/hci/llc_shdlc.c
581
static void llc_shdlc_t2_timeout(struct timer_list *t)
net/nfc/hci/llc_shdlc.c
583
struct llc_shdlc *shdlc = timer_container_of(shdlc, t, t2_timer);
net/nfc/llcp_core.c
244
static void nfc_llcp_symm_timer(struct timer_list *t)
net/nfc/llcp_core.c
246
struct nfc_llcp_local *local = timer_container_of(local, t,
net/nfc/llcp_core.c
288
static void nfc_llcp_sdreq_timer(struct timer_list *t)
net/nfc/llcp_core.c
290
struct nfc_llcp_local *local = timer_container_of(local, t,
net/nfc/nci/core.c
622
static void nci_cmd_timer(struct timer_list *t)
net/nfc/nci/core.c
624
struct nci_dev *ndev = timer_container_of(ndev, t, cmd_timer);
net/nfc/nci/core.c
631
static void nci_data_timer(struct timer_list *t)
net/nfc/nci/core.c
633
struct nci_dev *ndev = timer_container_of(ndev, t, data_timer);
net/nfc/nci/spi.c
34
struct spi_transfer t;
net/nfc/nci/spi.c
36
memset(&t, 0, sizeof(struct spi_transfer));
net/nfc/nci/spi.c
39
t.tx_buf = skb->data;
net/nfc/nci/spi.c
40
t.len = skb->len;
net/nfc/nci/spi.c
43
t.tx_buf = &t;
net/nfc/nci/spi.c
44
t.len = 0;
net/nfc/nci/spi.c
46
t.cs_change = cs_change;
net/nfc/nci/spi.c
47
t.delay.value = nspi->xfer_udelay;
net/nfc/nci/spi.c
48
t.delay.unit = SPI_DELAY_UNIT_USECS;
net/nfc/nci/spi.c
49
t.speed_hz = nspi->xfer_speed_hz;
net/nfc/nci/spi.c
52
spi_message_add_tail(&t, &m);
net/openvswitch/flow_table.c
993
struct sw_flow_mask *t;
net/openvswitch/flow_table.c
994
t = ovsl_dereference(ma->masks[i]);
net/openvswitch/flow_table.c
996
if (t && mask_equal(mask, t))
net/openvswitch/flow_table.c
997
return t;
net/packet/af_packet.c
681
static enum hrtimer_restart prb_retire_rx_blk_timer_expired(struct hrtimer *t)
net/packet/af_packet.c
684
timer_container_of(po, t, rx_ring.prb_bdqc.retire_blk_timer);
net/rds/recv.c
613
struct rds_cmsg_rx_trace t;
net/rds/recv.c
616
memset(&t, 0, sizeof(t));
net/rds/recv.c
618
t.rx_traces = rs->rs_rx_traces;
net/rds/recv.c
621
t.rx_trace_pos[i] = j;
net/rds/recv.c
622
t.rx_trace[i] = inc->i_rx_lat_trace[j + 1] -
net/rds/recv.c
627
sizeof(t), &t);
net/rose/af_rose.c
346
static void rose_destroy_timer(struct timer_list *t)
net/rose/af_rose.c
348
struct sock *sk = timer_container_of(sk, t, sk_timer);
net/rose/rose_link.c
75
static void rose_ftimer_expiry(struct timer_list *t)
net/rose/rose_link.c
79
static void rose_t0timer_expiry(struct timer_list *t)
net/rose/rose_link.c
81
struct rose_neigh *neigh = timer_container_of(neigh, t, t0timer);
net/rose/rose_route.c
1306
struct rose_node *t, *rose_node = rose_node_list;
net/rose/rose_route.c
1319
t = rose_node;
net/rose/rose_route.c
1322
for (i = 0; i < t->count; i++)
net/rose/rose_route.c
1323
rose_neigh_put(t->neighbour[i]);
net/rose/rose_route.c
1324
rose_remove_node(t);
net/rose/rose_route.c
480
struct rose_node *t, *rose_node;
net/rose/rose_route.c
496
t = rose_node;
net/rose/rose_route.c
499
for (i = t->count - 1; i >= 0; i--) {
net/rose/rose_route.c
500
if (t->neighbour[i] != s)
net/rose/rose_route.c
503
t->count--;
net/rose/rose_route.c
505
memmove(&t->neighbour[i], &t->neighbour[i + 1],
net/rose/rose_route.c
506
sizeof(t->neighbour[0]) *
net/rose/rose_route.c
507
(t->count - i));
net/rose/rose_route.c
511
if (t->count <= 0)
net/rose/rose_route.c
512
rose_remove_node(t);
net/rose/rose_route.c
551
struct rose_node *t, *rose_node;
net/rose/rose_route.c
561
t = rose_node;
net/rose/rose_route.c
564
if (!t->loopback) {
net/rose/rose_route.c
565
for (i = 0; i < t->count; i++)
net/rose/rose_route.c
566
rose_neigh_put(t->neighbour[i]);
net/rose/rose_route.c
567
rose_remove_node(t);
net/rose/rose_timer.c
119
static void rose_heartbeat_expiry(struct timer_list *t)
net/rose/rose_timer.c
121
struct sock *sk = timer_container_of(sk, t, sk_timer);
net/rose/rose_timer.c
164
static void rose_timer_expiry(struct timer_list *t)
net/rose/rose_timer.c
166
struct rose_sock *rose = timer_container_of(rose, t, timer);
net/rose/rose_timer.c
199
static void rose_idletimer_expiry(struct timer_list *t)
net/rose/rose_timer.c
201
struct rose_sock *rose = timer_container_of(rose, t, idletimer);
net/rose/rose_timer.c
28
static void rose_heartbeat_expiry(struct timer_list *t);
net/rxrpc/call_event.c
303
ktime_t now, t;
net/rxrpc/call_event.c
340
t = ktime_sub(call->rack_timo_at, ktime_get_real());
net/rxrpc/call_event.c
341
if (t <= 0) {
net/rxrpc/call_event.c
342
trace_rxrpc_timer_exp(call, t,
net/rxrpc/call_event.c
345
rxrpc_rack_timer_expired(call, t);
net/rxrpc/call_event.c
352
t = ktime_sub(call->expect_rx_by, now);
net/rxrpc/call_event.c
353
if (t <= 0) {
net/rxrpc/call_event.c
354
trace_rxrpc_timer_exp(call, t, rxrpc_timer_trace_expect_rx);
net/rxrpc/call_event.c
358
t = ktime_sub(call->expect_req_by, now);
net/rxrpc/call_event.c
359
if (t <= 0) {
net/rxrpc/call_event.c
362
trace_rxrpc_timer_exp(call, t, rxrpc_timer_trace_idle);
net/rxrpc/call_event.c
367
t = ktime_sub(READ_ONCE(call->expect_term_by), now);
net/rxrpc/call_event.c
368
if (t <= 0) {
net/rxrpc/call_event.c
369
trace_rxrpc_timer_exp(call, t, rxrpc_timer_trace_hard);
net/rxrpc/call_event.c
373
t = ktime_sub(call->delay_ack_at, now);
net/rxrpc/call_event.c
374
if (t <= 0) {
net/rxrpc/call_event.c
375
trace_rxrpc_timer_exp(call, t, rxrpc_timer_trace_delayed_ack);
net/rxrpc/call_event.c
381
t = ktime_sub(call->ping_at, now);
net/rxrpc/call_event.c
382
if (t <= 0) {
net/rxrpc/call_event.c
383
trace_rxrpc_timer_exp(call, t, rxrpc_timer_trace_ping);
net/rxrpc/call_event.c
390
t = ktime_sub(call->keepalive_at, now);
net/rxrpc/call_event.c
391
if (t <= 0) {
net/rxrpc/call_event.c
392
trace_rxrpc_timer_exp(call, t, rxrpc_timer_trace_keepalive);
net/rxrpc/call_event.c
409
t = ktime_sub(call->rack_timo_at, ktime_get_real());
net/rxrpc/call_event.c
410
trace_rxrpc_rack(call, t);
net/rxrpc/call_object.c
65
static void rxrpc_call_timer_expired(struct timer_list *t)
net/rxrpc/call_object.c
67
struct rxrpc_call *call = timer_container_of(call, t, timer);
net/rxrpc/rxgk_app.c
106
t = ticket + sizeof(__be32) * 2 + xdr_round_up(klen);
net/rxrpc/rxgk_app.c
109
q[1] = t[1]; /* begintime - msw */
net/rxrpc/rxgk_app.c
110
q[2] = t[2]; /* - lsw */
net/rxrpc/rxgk_app.c
111
q[3] = t[5]; /* endtime - msw */
net/rxrpc/rxgk_app.c
112
q[4] = t[6]; /* - lsw */
net/rxrpc/rxgk_app.c
114
q[6] = t[0]; /* - lsw */
net/rxrpc/rxgk_app.c
116
q[8] = t[3]; /* - lsw */
net/rxrpc/rxgk_app.c
118
q[10] = t[4]; /* - lsw */
net/rxrpc/rxgk_app.c
52
__be32 *t, *p, *q, tmp[2];
net/sched/act_api.c
1651
struct tcamsg *t;
net/sched/act_api.c
1656
nlh = nlmsg_put(skb, portid, seq, event, sizeof(*t), flags);
net/sched/act_api.c
1659
t = nlmsg_data(nlh);
net/sched/act_api.c
1660
t->tca_family = AF_UNSPEC;
net/sched/act_api.c
1661
t->tca__pad1 = 0;
net/sched/act_api.c
1662
t->tca__pad2 = 0;
net/sched/act_api.c
1757
struct tcamsg *t;
net/sched/act_api.c
1785
sizeof(*t), 0);
net/sched/act_api.c
1790
t = nlmsg_data(nlh);
net/sched/act_api.c
1791
t->tca_family = AF_UNSPEC;
net/sched/act_api.c
1792
t->tca__pad1 = 0;
net/sched/act_api.c
1793
t->tca__pad2 = 0;
net/sched/act_api.c
2211
struct tcamsg *t = (struct tcamsg *) nlmsg_data(cb->nlh);
net/sched/act_api.c
2246
cb->nlh->nlmsg_type, sizeof(*t), 0);
net/sched/act_api.c
2253
t = nlmsg_data(nlh);
net/sched/act_api.c
2254
t->tca_family = AF_UNSPEC;
net/sched/act_api.c
2255
t->tca__pad1 = 0;
net/sched/act_api.c
2256
t->tca__pad2 = 0;
net/sched/act_connmark.c
199
struct tcf_t t;
net/sched/act_connmark.c
215
tcf_tm_dump(&t, &ci->tcf_tm);
net/sched/act_connmark.c
216
if (nla_put_64bit(skb, TCA_CONNMARK_TM, sizeof(t), &t,
net/sched/act_csum.c
647
struct tcf_t t;
net/sched/act_csum.c
657
tcf_tm_dump(&t, &p->tcf_tm);
net/sched/act_csum.c
658
if (nla_put_64bit(skb, TCA_CSUM_TM, sizeof(t), &t, TCA_CSUM_PAD))
net/sched/act_ct.c
1541
struct tcf_t t;
net/sched/act_ct.c
1587
tcf_tm_dump(&t, &c->tcf_tm);
net/sched/act_ct.c
1588
if (nla_put_64bit(skb, TCA_CT_TM, sizeof(t), &t, TCA_CT_PAD))
net/sched/act_ctinfo.c
294
struct tcf_t t;
net/sched/act_ctinfo.c
299
tcf_tm_dump(&t, &ci->tcf_tm);
net/sched/act_ctinfo.c
300
if (nla_put_64bit(skb, TCA_CTINFO_TM, sizeof(t), &t, TCA_CTINFO_PAD))
net/sched/act_gact.c
195
struct tcf_t t;
net/sched/act_gact.c
213
tcf_tm_dump(&t, &gact->tcf_tm);
net/sched/act_gact.c
214
if (nla_put_64bit(skb, TCA_GACT_TM, sizeof(t), &t, TCA_GACT_PAD))
net/sched/act_gate.c
618
struct tcf_t t;
net/sched/act_gate.c
659
tcf_tm_dump(&t, &gact->tcf_tm);
net/sched/act_gate.c
660
if (nla_put_64bit(skb, TCA_GATE_TM, sizeof(t), &t, TCA_GATE_PAD))
net/sched/act_ife.c
644
struct tcf_t t;
net/sched/act_ife.c
661
tcf_tm_dump(&t, &ife->tcf_tm);
net/sched/act_ife.c
662
if (nla_put_64bit(skb, TCA_IFE_TM, sizeof(t), &t, TCA_IFE_PAD))
net/sched/act_mirred.c
498
struct tcf_t t;
net/sched/act_mirred.c
515
tcf_tm_dump(&t, &m->tcf_tm);
net/sched/act_mirred.c
516
if (nla_put_64bit(skb, TCA_MIRRED_TM, sizeof(t), &t, TCA_MIRRED_PAD))
net/sched/act_mpls.c
339
struct tcf_t t;
net/sched/act_mpls.c
367
tcf_tm_dump(&t, &m->tcf_tm);
net/sched/act_mpls.c
369
if (nla_put_64bit(skb, TCA_MPLS_TM, sizeof(t), &t, TCA_MPLS_PAD))
net/sched/act_nat.c
278
struct tcf_t t;
net/sched/act_nat.c
293
tcf_tm_dump(&t, &p->tcf_tm);
net/sched/act_nat.c
294
if (nla_put_64bit(skb, TCA_NAT_TM, sizeof(t), &t, TCA_NAT_PAD))
net/sched/act_pedit.c
506
struct tcf_t t;
net/sched/act_pedit.c
540
tcf_tm_dump(&t, &p->tcf_tm);
net/sched/act_pedit.c
541
if (nla_put_64bit(skb, TCA_PEDIT_TM, sizeof(t), &t, TCA_PEDIT_PAD))
net/sched/act_police.c
350
struct tcf_t t;
net/sched/act_police.c
392
tcf_tm_dump(&t, &police->tcf_tm);
net/sched/act_police.c
393
if (nla_put_64bit(skb, TCA_POLICE_TM, sizeof(t), &t, TCA_POLICE_PAD))
net/sched/act_sample.c
234
struct tcf_t t;
net/sched/act_sample.c
241
tcf_tm_dump(&t, &s->tcf_tm);
net/sched/act_sample.c
242
if (nla_put_64bit(skb, TCA_SAMPLE_TM, sizeof(t), &t, TCA_SAMPLE_PAD))
net/sched/act_simple.c
180
struct tcf_t t;
net/sched/act_simple.c
188
tcf_tm_dump(&t, &d->tcf_tm);
net/sched/act_simple.c
189
if (nla_put_64bit(skb, TCA_DEF_TM, sizeof(t), &t, TCA_DEF_PAD))
net/sched/act_skbedit.c
299
struct tcf_t t;
net/sched/act_skbedit.c
335
tcf_tm_dump(&t, &d->tcf_tm);
net/sched/act_skbedit.c
336
if (nla_put_64bit(skb, TCA_SKBEDIT_TM, sizeof(t), &t, TCA_SKBEDIT_PAD))
net/sched/act_skbmod.c
244
struct tcf_t t;
net/sched/act_skbmod.c
266
tcf_tm_dump(&t, &d->tcf_tm);
net/sched/act_skbmod.c
267
if (nla_put_64bit(skb, TCA_SKBMOD_TM, sizeof(t), &t, TCA_SKBMOD_PAD))
net/sched/act_tunnel_key.c
30
struct tcf_tunnel_key *t = to_tunnel_key(a);
net/sched/act_tunnel_key.c
33
params = rcu_dereference_bh(t->params);
net/sched/act_tunnel_key.c
35
tcf_lastuse_update(&t->tcf_tm);
net/sched/act_tunnel_key.c
36
tcf_action_update_bstats(&t->common, skb);
net/sched/act_tunnel_key.c
368
struct tcf_tunnel_key *t;
net/sched/act_tunnel_key.c
521
t = to_tunnel_key(*a);
net/sched/act_tunnel_key.c
534
spin_lock_bh(&t->tcf_lock);
net/sched/act_tunnel_key.c
536
params_new = rcu_replace_pointer(t->params, params_new,
net/sched/act_tunnel_key.c
537
lockdep_is_held(&t->tcf_lock));
net/sched/act_tunnel_key.c
538
spin_unlock_bh(&t->tcf_lock);
net/sched/act_tunnel_key.c
563
struct tcf_tunnel_key *t = to_tunnel_key(a);
net/sched/act_tunnel_key.c
566
params = rcu_dereference_protected(t->params, 1);
net/sched/act_tunnel_key.c
719
struct tcf_tunnel_key *t = to_tunnel_key(a);
net/sched/act_tunnel_key.c
722
.index = t->tcf_index,
net/sched/act_tunnel_key.c
723
.refcnt = refcount_read(&t->tcf_refcnt) - ref,
net/sched/act_tunnel_key.c
724
.bindcnt = atomic_read(&t->tcf_bindcnt) - bind,
net/sched/act_tunnel_key.c
729
params = rcu_dereference(t->params);
net/sched/act_tunnel_key.c
763
tcf_tm_dump(&tm, &t->tcf_tm);
net/sched/act_vlan.c
296
struct tcf_t t;
net/sched/act_vlan.c
323
tcf_tm_dump(&t, &v->tcf_tm);
net/sched/act_vlan.c
324
if (nla_put_64bit(skb, TCA_VLAN_TM, sizeof(t), &t, TCA_VLAN_PAD))
net/sched/bpf_qdisc.c
100
const struct btf_type *t, *skbt, *qdisct;
net/sched/bpf_qdisc.c
106
t = btf_type_by_id(reg->btf, reg->btf_id);
net/sched/bpf_qdisc.c
108
if (t == skbt) {
net/sched/bpf_qdisc.c
110
} else if (t == qdisct) {
net/sched/bpf_qdisc.c
119
btf_name_by_offset(reg->btf, t->name_off), off);
net/sched/bpf_qdisc.c
126
off, size, btf_name_by_offset(reg->btf, t->name_off), end);
net/sched/bpf_qdisc.c
361
static int bpf_qdisc_init_member(const struct btf_type *t,
net/sched/bpf_qdisc.c
372
moff = __btf_member_bit_offset(t, member) / 8;
net/sched/cls_api.c
2242
struct tcmsg *t;
net/sched/cls_api.c
2263
err = nlmsg_parse_deprecated(n, sizeof(*t), tca, TCA_MAX,
net/sched/cls_api.c
2268
t = nlmsg_data(n);
net/sched/cls_api.c
2269
protocol = TC_H_MIN(t->tcm_info);
net/sched/cls_api.c
2270
prio = TC_H_MAJ(t->tcm_info);
net/sched/cls_api.c
2272
parent = t->tcm_parent;
net/sched/cls_api.c
2295
err = __tcf_qdisc_find(net, &q, &parent, t->tcm_ifindex, false, extack);
net/sched/cls_api.c
2316
err = __tcf_qdisc_cl_find(q, parent, &cl, t->tcm_ifindex, extack);
net/sched/cls_api.c
232
const struct tcf_proto_ops *t, *res = NULL;
net/sched/cls_api.c
2320
block = __tcf_block_find(net, q, cl, t->tcm_ifindex, t->tcm_block_index,
net/sched/cls_api.c
236
list_for_each_entry(t, &tcf_proto_base, head) {
net/sched/cls_api.c
237
if (strcmp(kind, t->kind) == 0) {
net/sched/cls_api.c
238
if (try_module_get(t->owner))
net/sched/cls_api.c
239
res = t;
net/sched/cls_api.c
2400
fh = tp->ops->get(tp, t->tcm_handle);
net/sched/cls_api.c
2430
err = tp->ops->change(net, skb, tp, cl, t->tcm_handle, tca, &fh,
net/sched/cls_api.c
2478
struct tcmsg *t;
net/sched/cls_api.c
2493
err = nlmsg_parse_deprecated(n, sizeof(*t), tca, TCA_MAX,
net/sched/cls_api.c
2498
t = nlmsg_data(n);
net/sched/cls_api.c
2499
protocol = TC_H_MIN(t->tcm_info);
net/sched/cls_api.c
2500
prio = TC_H_MAJ(t->tcm_info);
net/sched/cls_api.c
2501
parent = t->tcm_parent;
net/sched/cls_api.c
2503
if (prio == 0 && (protocol || t->tcm_handle || tca[TCA_KIND])) {
net/sched/cls_api.c
2510
err = __tcf_qdisc_find(net, &q, &parent, t->tcm_ifindex, false, extack);
net/sched/cls_api.c
2530
err = __tcf_qdisc_cl_find(q, parent, &cl, t->tcm_ifindex, extack);
net/sched/cls_api.c
2534
block = __tcf_block_find(net, q, cl, t->tcm_ifindex, t->tcm_block_index,
net/sched/cls_api.c
2583
} else if (t->tcm_handle == 0) {
net/sched/cls_api.c
2596
fh = tp->ops->get(tp, t->tcm_handle);
net/sched/cls_api.c
2637
struct tcmsg *t;
net/sched/cls_api.c
2652
err = nlmsg_parse_deprecated(n, sizeof(*t), tca, TCA_MAX,
net/sched/cls_api.c
2657
t = nlmsg_data(n);
net/sched/cls_api.c
2658
protocol = TC_H_MIN(t->tcm_info);
net/sched/cls_api.c
2659
prio = TC_H_MAJ(t->tcm_info);
net/sched/cls_api.c
2660
parent = t->tcm_parent;
net/sched/cls_api.c
2669
err = __tcf_qdisc_find(net, &q, &parent, t->tcm_ifindex, false, extack);
net/sched/cls_api.c
2688
err = __tcf_qdisc_cl_find(q, parent, &cl, t->tcm_ifindex, extack);
net/sched/cls_api.c
2692
block = __tcf_block_find(net, q, cl, t->tcm_ifindex, t->tcm_block_index,
net/sched/cls_api.c
2729
fh = tp->ops->get(tp, t->tcm_handle);
net/sched/cls_api.c
282
struct tcf_proto_ops *t;
net/sched/cls_api.c
286
list_for_each_entry(t, &tcf_proto_base, head)
net/sched/cls_api.c
287
if (!strcmp(ops->kind, t->kind))
net/sched/cls_api.c
302
struct tcf_proto_ops *t;
net/sched/cls_api.c
3118
struct tcmsg *t;
net/sched/cls_api.c
312
list_for_each_entry(t, &tcf_proto_base, head) {
net/sched/cls_api.c
3129
err = nlmsg_parse_deprecated(n, sizeof(*t), tca, TCA_MAX,
net/sched/cls_api.c
313
if (t == ops) {
net/sched/cls_api.c
3134
t = nlmsg_data(n);
net/sched/cls_api.c
3135
parent = t->tcm_parent;
net/sched/cls_api.c
3139
t->tcm_ifindex, t->tcm_block_index, extack);
net/sched/cls_api.c
314
list_del(&t->head);
net/sched/cls_basic.c
270
struct sk_buff *skb, struct tcmsg *t, bool rtnl_held)
net/sched/cls_basic.c
280
t->tcm_handle = f->handle;
net/sched/cls_cgroup.c
173
struct sk_buff *skb, struct tcmsg *t, bool rtnl_held)
net/sched/cls_cgroup.c
178
t->tcm_handle = head->handle;
net/sched/cls_flow.c
346
static void flow_perturbation(struct timer_list *t)
net/sched/cls_flow.c
348
struct flow_filter *f = timer_container_of(f, t, perturb_timer);
net/sched/cls_flow.c
630
struct sk_buff *skb, struct tcmsg *t, bool rtnl_held)
net/sched/cls_flow.c
638
t->tcm_handle = f->handle;
net/sched/cls_flower.c
3689
struct sk_buff *skb, struct tcmsg *t, bool rtnl_held)
net/sched/cls_flower.c
3699
t->tcm_handle = f->handle;
net/sched/cls_flower.c
3747
struct sk_buff *skb, struct tcmsg *t, bool rtnl_held)
net/sched/cls_flower.c
3756
t->tcm_handle = f->handle;
net/sched/cls_fw.c
379
struct sk_buff *skb, struct tcmsg *t, bool rtnl_held)
net/sched/cls_fw.c
388
t->tcm_handle = f->id;
net/sched/cls_matchall.c
329
struct sk_buff *skb, struct tcmsg *t, bool rtnl_held)
net/sched/cls_matchall.c
342
t->tcm_handle = head->handle;
net/sched/cls_route.c
605
struct sk_buff *skb, struct tcmsg *t, bool rtnl_held)
net/sched/cls_route.c
614
t->tcm_handle = f->handle;
net/sched/cls_u32.c
1343
struct sk_buff *skb, struct tcmsg *t, bool rtnl_held)
net/sched/cls_u32.c
1352
t->tcm_handle = n->handle;
net/sched/sch_api.c
2395
int t, s_t;
net/sched/sch_api.c
2398
t = 0;
net/sched/sch_api.c
2401
skb, tcm, cb, &t, s_t, true) < 0)
net/sched/sch_api.c
2407
skb, tcm, cb, &t, s_t, false) < 0)
net/sched/sch_api.c
2411
cb->args[0] = t;
net/sched/sch_cake.c
1463
q->tins[ii.t].overflow_idx[ii.b] = j;
net/sched/sch_cake.c
1464
q->tins[jj.t].overflow_idx[jj.b] = i;
net/sched/sch_cake.c
1471
return q->tins[ii.t].backlogs[ii.b];
net/sched/sch_cake.c
148
u16 t:3, b:10;
net/sched/sch_cake.c
1580
tin = qq.t;
net/sched/sch_cake.c
1712
static u32 cake_classify(struct Qdisc *sch, struct cake_tin_data **t,
net/sched/sch_cake.c
1746
*t = cake_select_tin(sch, skb);
net/sched/sch_cake.c
1747
return cake_hash(*t, skb, flow_mode, flow, host) + 1;
net/sched/sch_cake.c
2620
u64 t = q->rate_bps * q->interval;
net/sched/sch_cake.c
2622
do_div(t, USEC_PER_SEC / 4);
net/sched/sch_cake.c
2623
qd->buffer_limit = max_t(u32, t, 4U << 20);
net/sched/sch_cake.c
2870
qd->overflow_heap[k].t = i;
net/sched/sch_cake.c
434
static ktime_t cobalt_control(ktime_t t,
net/sched/sch_cake.c
438
return ktime_add_ns(t, reciprocal_scale(interval,
net/sched/sch_fq_pie.c
387
static void fq_pie_timer(struct timer_list *t)
net/sched/sch_fq_pie.c
389
struct fq_pie_sched_data *q = timer_container_of(q, t, adapt_timer);
net/sched/sch_generic.c
500
static void dev_watchdog(struct timer_list *t)
net/sched/sch_generic.c
502
struct net_device *dev = timer_container_of(dev, t, watchdog_timer);
net/sched/sch_gred.c
169
struct gred_sched *t = qdisc_priv(sch);
net/sched/sch_gred.c
173
if (dp >= t->DPs || (q = t->tab[dp]) == NULL) {
net/sched/sch_gred.c
174
dp = t->def;
net/sched/sch_gred.c
176
q = t->tab[dp];
net/sched/sch_gred.c
195
if (!gred_wred_mode(t) && gred_rio_mode(t)) {
net/sched/sch_gred.c
198
for (i = 0; i < t->DPs; i++) {
net/sched/sch_gred.c
199
if (t->tab[i] && t->tab[i]->prio < q->prio &&
net/sched/sch_gred.c
200
!red_is_idling(&t->tab[i]->vars))
net/sched/sch_gred.c
201
qavg += t->tab[i]->vars.qavg;
net/sched/sch_gred.c
209
if (gred_wred_mode(t))
net/sched/sch_gred.c
210
gred_load_wred_set(t, q);
net/sched/sch_gred.c
214
gred_backlog(t, q, sch));
net/sched/sch_gred.c
219
if (gred_wred_mode(t))
net/sched/sch_gred.c
220
gred_store_wred_set(t, q);
net/sched/sch_gred.c
247
if (gred_backlog(t, q, sch) + qdisc_pkt_len(skb) <= q->limit) {
net/sched/sch_gred.c
264
struct gred_sched *t = qdisc_priv(sch);
net/sched/sch_gred.c
272
if (dp >= t->DPs || (q = t->tab[dp]) == NULL) {
net/sched/sch_gred.c
278
if (gred_wred_mode(t)) {
net/sched/sch_gred.c
280
red_start_of_idle_period(&t->wred_set);
net/sched/sch_gred.c
296
struct gred_sched *t = qdisc_priv(sch);
net/sched/sch_gred.c
300
for (i = 0; i < t->DPs; i++) {
net/sched/sch_gred.c
301
struct gred_sched_data *q = t->tab[i];
net/sched/sch_netem.c
335
long t;
net/sched/sch_netem.c
347
t = dist->table[rnd % dist->size];
net/sched/sch_netem.c
348
x = (sigma % NETEM_DIST_SCALE) * t;
net/sched/sch_netem.c
354
return x / NETEM_DIST_SCALE + (sigma / NETEM_DIST_SCALE) * t + mu;
net/sched/sch_pie.c
427
static void pie_timer(struct timer_list *t)
net/sched/sch_pie.c
429
struct pie_sched_data *q = timer_container_of(q, t, adapt_timer);
net/sched/sch_red.c
322
static inline void red_adaptative_timer(struct timer_list *t)
net/sched/sch_red.c
324
struct red_sched_data *q = timer_container_of(q, t, adapt_timer);
net/sched/sch_sfq.c
601
static void sfq_perturbation(struct timer_list *t)
net/sched/sch_sfq.c
603
struct sfq_sched_data *q = timer_container_of(q, t, perturb_timer);
net/sctp/associola.c
1410
struct sctp_transport *t;
net/sctp/associola.c
1417
list_for_each_entry(t, &asoc->peer.transport_addr_list, transports) {
net/sctp/associola.c
1418
if (t->pmtu_pending && t->dst) {
net/sctp/associola.c
1419
sctp_transport_update_pmtu(t,
net/sctp/associola.c
1420
atomic_read(&t->mtu_info));
net/sctp/associola.c
1421
t->pmtu_pending = 0;
net/sctp/associola.c
1423
if (!pmtu || (t->pathmtu < pmtu))
net/sctp/associola.c
1424
pmtu = t->pathmtu;
net/sctp/associola.c
744
struct sctp_transport *t;
net/sctp/associola.c
748
list_for_each_entry(t, &asoc->peer.transport_addr_list,
net/sctp/associola.c
750
if (sctp_cmp_addr_exact(address, &t->ipaddr))
net/sctp/associola.c
751
return t;
net/sctp/associola.c
762
struct sctp_transport *t;
net/sctp/associola.c
764
list_for_each_entry_safe(t, temp, &asoc->peer.transport_addr_list,
net/sctp/associola.c
767
if (t != primary)
net/sctp/associola.c
768
sctp_assoc_rm_peer(asoc, t);
net/sctp/endpointola.c
277
struct sctp_transport *t;
net/sctp/endpointola.c
288
t = sctp_epaddr_lookup_transport(ep, paddr);
net/sctp/endpointola.c
289
if (!t)
net/sctp/endpointola.c
292
*transport = t;
net/sctp/endpointola.c
293
asoc = t->asoc;
net/sctp/input.c
1002
struct sctp_transport *t;
net/sctp/input.c
1013
rhl_for_each_entry_rcu(t, tmp, list, node) {
net/sctp/input.c
1014
if (!sctp_transport_hold(t))
net/sctp/input.c
1017
bound_dev_if = READ_ONCE(t->asoc->base.sk->sk_bound_dev_if);
net/sctp/input.c
1019
sctp_bind_addr_match(&t->asoc->base.bind_addr,
net/sctp/input.c
1020
laddr, sctp_sk(t->asoc->base.sk)))
net/sctp/input.c
1021
return t;
net/sctp/input.c
1022
sctp_transport_put(t);
net/sctp/input.c
1034
struct sctp_transport *t;
net/sctp/input.c
1044
rhl_for_each_entry_rcu(t, tmp, list, node)
net/sctp/input.c
1045
if (ep == t->asoc->ep)
net/sctp/input.c
1046
return t;
net/sctp/input.c
1059
struct sctp_transport *t;
net/sctp/input.c
1062
t = sctp_addrs_lookup_transport(net, local, peer, dif, sdif);
net/sctp/input.c
1063
if (!t)
net/sctp/input.c
1066
asoc = t->asoc;
net/sctp/input.c
1067
*pt = t;
net/sctp/input.c
280
struct sctp_transport *t = chunk->transport;
net/sctp/input.c
338
sctp_transport_put(t);
net/sctp/input.c
350
struct sctp_transport *t = chunk->transport;
net/sctp/input.c
361
sctp_transport_hold(t);
net/sctp/input.c
373
struct sctp_transport *t, __u32 pmtu)
net/sctp/input.c
375
if (!t ||
net/sctp/input.c
376
(t->pathmtu <= pmtu &&
net/sctp/input.c
377
t->pl.probe_size + sctp_transport_pl_hlen(t) <= pmtu))
net/sctp/input.c
381
atomic_set(&t->mtu_info, pmtu);
net/sctp/input.c
383
t->pmtu_pending = 1;
net/sctp/input.c
387
if (!(t->param_flags & SPP_PMTUD_ENABLE))
net/sctp/input.c
398
if (!sctp_transport_update_pmtu(t, pmtu))
net/sctp/input.c
405
sctp_retransmit(&asoc->outqueue, t, SCTP_RTXR_PMTUD);
net/sctp/input.c
408
void sctp_icmp_redirect(struct sock *sk, struct sctp_transport *t,
net/sctp/input.c
413
if (sock_owned_by_user(sk) || !t)
net/sctp/input.c
415
dst = sctp_transport_dst_check(t);
net/sctp/input.c
433
struct sctp_transport *t)
net/sctp/input.c
436
if (timer_pending(&t->proto_unreach_timer))
net/sctp/input.c
439
if (!mod_timer(&t->proto_unreach_timer,
net/sctp/input.c
441
sctp_transport_hold(t);
net/sctp/input.c
449
if (timer_delete(&t->proto_unreach_timer))
net/sctp/input.c
450
sctp_transport_put(t);
net/sctp/input.c
454
asoc->state, asoc->ep, asoc, t,
net/sctp/input.c
541
void sctp_err_finish(struct sock *sk, struct sctp_transport *t)
net/sctp/input.c
545
sctp_transport_put(t);
net/sctp/input.c
548
static void sctp_v4_err_handle(struct sctp_transport *t, struct sk_buff *skb,
net/sctp/input.c
551
struct sctp_association *asoc = t->asoc;
net/sctp/input.c
563
sctp_icmp_frag_needed(sk, asoc, t, SCTP_TRUNC4(info));
net/sctp/input.c
567
sctp_icmp_proto_unreachable(sk, asoc, t);
net/sctp/input.c
579
sctp_icmp_redirect(sk, t, skb);
net/sctp/input.c
642
struct sctp_transport *t;
net/sctp/input.c
647
sk = sctp_err_lookup(net, AF_INET, skb, sctp_hdr(skb), &asoc, &t);
net/sctp/input.c
657
sctp_err_finish(sk, t);
net/sctp/input.c
662
sctp_v4_err_handle(t, skb, hdr->type, hdr->code, info);
net/sctp/input.c
664
sctp_err_finish(sk, t);
net/sctp/input.c
888
struct sctp_transport *t = (struct sctp_transport *)ptr;
net/sctp/input.c
892
if (!sctp_cmp_addr_exact(&t->ipaddr, x->paddr))
net/sctp/input.c
894
if (!sctp_transport_hold(t))
net/sctp/input.c
897
if (!net_eq(t->asoc->base.net, x->net))
net/sctp/input.c
899
if (x->lport != htons(t->asoc->base.bind_addr.port))
net/sctp/input.c
904
sctp_transport_put(t);
net/sctp/input.c
910
const struct sctp_transport *t = data;
net/sctp/input.c
912
return sctp_hashfn(t->asoc->base.net,
net/sctp/input.c
913
htons(t->asoc->base.bind_addr.port),
net/sctp/input.c
914
&t->ipaddr, seed);
net/sctp/input.c
942
int sctp_hash_transport(struct sctp_transport *t)
net/sctp/input.c
949
if (t->asoc->temp)
net/sctp/input.c
952
arg.net = t->asoc->base.net;
net/sctp/input.c
953
arg.paddr = &t->ipaddr;
net/sctp/input.c
954
arg.lport = htons(t->asoc->base.bind_addr.port);
net/sctp/input.c
961
if (transport->asoc->ep == t->asoc->ep) {
net/sctp/input.c
968
&t->node, sctp_hash_params);
net/sctp/input.c
975
void sctp_unhash_transport(struct sctp_transport *t)
net/sctp/input.c
977
if (t->asoc->temp)
net/sctp/input.c
980
rhltable_remove(&sctp_transport_hashtable, &t->node,
net/sctp/ipv6.c
126
static void sctp_v6_err_handle(struct sctp_transport *t, struct sk_buff *skb,
net/sctp/ipv6.c
129
struct sctp_association *asoc = t->asoc;
net/sctp/ipv6.c
136
sctp_icmp_frag_needed(sk, asoc, t, info);
net/sctp/ipv6.c
140
sctp_icmp_proto_unreachable(sk, asoc, t);
net/sctp/ipv6.c
145
sctp_icmp_redirect(sk, t, skb);
net/sctp/ipv6.c
194
struct sctp_transport *t;
net/sctp/ipv6.c
199
sk = sctp_err_lookup(net, AF_INET6, skb, sctp_hdr(skb), &asoc, &t);
net/sctp/ipv6.c
209
sctp_err_finish(sk, t);
net/sctp/ipv6.c
214
sctp_v6_err_handle(t, skb, hdr->icmp6_type, hdr->icmp6_code, info);
net/sctp/ipv6.c
216
sctp_err_finish(sk, t);
net/sctp/ipv6.c
220
static int sctp_v6_xmit(struct sk_buff *skb, struct sctp_transport *t)
net/sctp/ipv6.c
222
struct dst_entry *dst = dst_clone(t->dst);
net/sctp/ipv6.c
223
struct flowi6 *fl6 = &t->fl.u.ip6;
net/sctp/ipv6.c
232
if (t->dscp & SCTP_DSCP_SET_MASK)
net/sctp/ipv6.c
233
tclass = t->dscp & SCTP_DSCP_VAL_MASK;
net/sctp/ipv6.c
238
if (!(t->param_flags & SPP_PMTUD_ENABLE))
net/sctp/ipv6.c
243
if (!t->encap_port || !sctp_sk(sk)->udp_port) {
net/sctp/ipv6.c
266
sctp_sk(sk)->udp_port, t->encap_port, false, 0);
net/sctp/ipv6.c
273
static void sctp_v6_get_dst(struct sctp_transport *t, union sctp_addr *saddr,
net/sctp/ipv6.c
276
struct sctp_association *asoc = t->asoc;
net/sctp/ipv6.c
283
union sctp_addr *daddr = &t->ipaddr;
net/sctp/ipv6.c
297
if (t->flowlabel & SCTP_FLOWLABEL_SET_MASK)
net/sctp/ipv6.c
298
fl6->flowlabel = htonl(t->flowlabel & SCTP_FLOWLABEL_VAL_MASK);
net/sctp/ipv6.c
329
t->dst = dst;
net/sctp/ipv6.c
355
t->dst = dst;
net/sctp/ipv6.c
395
t->dst = dst;
net/sctp/ipv6.c
410
t->dst = dst;
net/sctp/ipv6.c
420
t->dst_cookie = rt6_get_cookie(rt);
net/sctp/ipv6.c
425
t->dst = NULL;
net/sctp/ipv6.c
443
struct sctp_transport *t,
net/sctp/ipv6.c
447
union sctp_addr *saddr = &t->saddr;
net/sctp/ipv6.c
449
pr_debug("%s: asoc:%p dst:%p\n", __func__, t->asoc, t->dst);
net/sctp/ipv6.c
451
if (t->dst) {
net/sctp/output.c
219
struct sctp_transport *t = pkt->transport;
net/sctp/output.c
229
pad = sctp_make_pad(t->asoc, t->pl.probe_size - overhead);
net/sctp/output.c
235
chunk->transport = t;
net/sctp/outqueue.c
1160
struct sctp_transport *t;
net/sctp/outqueue.c
1164
t = list_entry(ltransport, struct sctp_transport, send_ready);
net/sctp/outqueue.c
1165
packet = &t->packet;
net/sctp/outqueue.c
1168
if (t->dst && __sk_dst_get(sk) != t->dst) {
net/sctp/outqueue.c
1169
dst_hold(t->dst);
net/sctp/outqueue.c
1170
sk_setup_caps(sk, t->dst);
net/sctp/outqueue.c
1179
sctp_transport_burst_reset(t);
net/sctp/protocol.c
1039
static inline int sctp_v4_xmit(struct sk_buff *skb, struct sctp_transport *t)
net/sctp/protocol.c
1041
struct dst_entry *dst = dst_clone(t->dst);
net/sctp/protocol.c
1042
struct flowi4 *fl4 = &t->fl.u.ip4;
net/sctp/protocol.c
1051
if (t->dscp & SCTP_DSCP_SET_MASK)
net/sctp/protocol.c
1052
dscp = t->dscp & SCTP_DSCP_VAL_MASK;
net/sctp/protocol.c
1054
inet->pmtudisc = t->param_flags & SPP_PMTUD_ENABLE ? IP_PMTUDISC_DO
net/sctp/protocol.c
1058
if (!t->encap_port || !sctp_sk(sk)->udp_port) {
net/sctp/protocol.c
1060
return __ip_queue_xmit(sk, skb, &t->fl, dscp);
net/sctp/protocol.c
1075
sctp_sk(sk)->udp_port, t->encap_port, false, false,
net/sctp/protocol.c
417
static void sctp_v4_get_dst(struct sctp_transport *t, union sctp_addr *saddr,
net/sctp/protocol.c
420
struct sctp_association *asoc = t->asoc;
net/sctp/protocol.c
427
union sctp_addr *daddr = &t->ipaddr;
net/sctp/protocol.c
431
if (t->dscp & SCTP_DSCP_SET_MASK)
net/sctp/protocol.c
432
dscp = inet_dsfield_to_dscp(t->dscp);
net/sctp/protocol.c
458
t->dst = dst;
net/sctp/protocol.c
523
t->dst = dst;
net/sctp/protocol.c
533
t->dst = dst;
net/sctp/protocol.c
545
t->dst = NULL;
net/sctp/protocol.c
554
struct sctp_transport *t,
net/sctp/protocol.c
557
union sctp_addr *saddr = &t->saddr;
net/sctp/protocol.c
558
struct rtable *rt = dst_rtable(t->dst);
net/sctp/protocol.c
601
static void sctp_addr_wq_timeout_handler(struct timer_list *t)
net/sctp/protocol.c
603
struct net *net = timer_container_of(net, t, sctp.addr_wq_timer);
net/sctp/sm_make_chunk.c
2518
struct sctp_transport *t;
net/sctp/sm_make_chunk.c
2646
t = sctp_assoc_lookup_paddr(asoc, &addr);
net/sctp/sm_make_chunk.c
2647
if (!t)
net/sctp/sm_make_chunk.c
2650
sctp_assoc_set_primary(asoc, t);
net/sctp/sm_sideeffect.c
1025
struct sctp_transport *t;
net/sctp/sm_sideeffect.c
1030
t = list_entry(pos, struct sctp_transport, transports);
net/sctp/sm_sideeffect.c
1031
if (!sctp_cmp_addr_exact(&t->ipaddr,
net/sctp/sm_sideeffect.c
1033
sctp_assoc_rm_peer(asoc, t);
net/sctp/sm_sideeffect.c
1089
struct sctp_transport *t;
net/sctp/sm_sideeffect.c
1091
t = asoc->init_last_sent_to;
net/sctp/sm_sideeffect.c
1094
if (t->init_sent_count > (asoc->init_cycle + 1)) {
net/sctp/sm_sideeffect.c
1290
struct sctp_transport *t;
net/sctp/sm_sideeffect.c
1572
t = sctp_assoc_choose_alter_transport(asoc,
net/sctp/sm_sideeffect.c
1574
asoc->init_last_sent_to = t;
net/sctp/sm_sideeffect.c
1575
chunk->transport = t;
net/sctp/sm_sideeffect.c
1576
t->init_sent_count++;
net/sctp/sm_sideeffect.c
1578
sctp_assoc_set_primary(asoc, t);
net/sctp/sm_sideeffect.c
1610
list_for_each_entry(t, &asoc->peer.transport_addr_list,
net/sctp/sm_sideeffect.c
1612
sctp_retransmit_mark(&asoc->outqueue, t,
net/sctp/sm_sideeffect.c
1637
list_for_each_entry(t, &asoc->peer.transport_addr_list,
net/sctp/sm_sideeffect.c
1639
t->init_sent_count = 0;
net/sctp/sm_sideeffect.c
1659
t = cmd->obj.transport;
net/sctp/sm_sideeffect.c
1660
sctp_transport_lower_cwnd(t, SCTP_LOWER_CWND_INACTIVE);
net/sctp/sm_sideeffect.c
1664
t = cmd->obj.transport;
net/sctp/sm_sideeffect.c
1666
t, 1);
net/sctp/sm_sideeffect.c
1667
t->hb_sent = 1;
net/sctp/sm_sideeffect.c
1671
t = cmd->obj.transport;
net/sctp/sm_sideeffect.c
1672
sctp_cmd_transport_on(commands, asoc, t, chunk);
net/sctp/sm_sideeffect.c
1680
t = cmd->obj.transport;
net/sctp/sm_sideeffect.c
1681
sctp_transport_reset_hb_timer(t);
net/sctp/sm_sideeffect.c
1689
t = cmd->obj.transport;
net/sctp/sm_sideeffect.c
1690
sctp_transport_reset_probe_timer(t);
net/sctp/sm_sideeffect.c
1722
t = cmd->obj.transport;
net/sctp/sm_sideeffect.c
1723
t->rto_pending = 1;
net/sctp/sm_sideeffect.c
1753
t = asoc->peer.retran_path;
net/sctp/sm_sideeffect.c
1757
asoc->peer.retran_path = t;
net/sctp/sm_sideeffect.c
231
void sctp_generate_t3_rtx_event(struct timer_list *t)
net/sctp/sm_sideeffect.c
234
timer_container_of(transport, t, T3_rtx_timer);
net/sctp/sm_sideeffect.c
308
static void sctp_generate_t1_cookie_event(struct timer_list *t)
net/sctp/sm_sideeffect.c
311
timer_container_of(asoc, t,
net/sctp/sm_sideeffect.c
317
static void sctp_generate_t1_init_event(struct timer_list *t)
net/sctp/sm_sideeffect.c
320
timer_container_of(asoc, t,
net/sctp/sm_sideeffect.c
326
static void sctp_generate_t2_shutdown_event(struct timer_list *t)
net/sctp/sm_sideeffect.c
329
timer_container_of(asoc, t,
net/sctp/sm_sideeffect.c
335
static void sctp_generate_t4_rto_event(struct timer_list *t)
net/sctp/sm_sideeffect.c
338
timer_container_of(asoc, t, timers[SCTP_EVENT_TIMEOUT_T4_RTO]);
net/sctp/sm_sideeffect.c
343
static void sctp_generate_t5_shutdown_guard_event(struct timer_list *t)
net/sctp/sm_sideeffect.c
346
timer_container_of(asoc, t,
net/sctp/sm_sideeffect.c
354
static void sctp_generate_autoclose_event(struct timer_list *t)
net/sctp/sm_sideeffect.c
357
timer_container_of(asoc, t,
net/sctp/sm_sideeffect.c
366
void sctp_generate_heartbeat_event(struct timer_list *t)
net/sctp/sm_sideeffect.c
368
struct sctp_transport *transport = timer_container_of(transport, t,
net/sctp/sm_sideeffect.c
412
void sctp_generate_proto_unreach_event(struct timer_list *t)
net/sctp/sm_sideeffect.c
415
timer_container_of(transport, t, proto_unreach_timer);
net/sctp/sm_sideeffect.c
447
void sctp_generate_reconf_event(struct timer_list *t)
net/sctp/sm_sideeffect.c
450
timer_container_of(transport, t, reconf_timer);
net/sctp/sm_sideeffect.c
484
void sctp_generate_probe_event(struct timer_list *t)
net/sctp/sm_sideeffect.c
486
struct sctp_transport *transport = timer_container_of(transport, t,
net/sctp/sm_sideeffect.c
517
static void sctp_generate_sack_event(struct timer_list *t)
net/sctp/sm_sideeffect.c
520
timer_container_of(asoc, t, timers[SCTP_EVENT_TIMEOUT_SACK]);
net/sctp/sm_sideeffect.c
724
struct sctp_transport *t;
net/sctp/sm_sideeffect.c
730
list_for_each_entry(t, &asoc->peer.transport_addr_list, transports)
net/sctp/sm_sideeffect.c
731
sctp_transport_reset_hb_timer(t);
net/sctp/sm_sideeffect.c
737
struct sctp_transport *t;
net/sctp/sm_sideeffect.c
741
list_for_each_entry(t, &asoc->peer.transport_addr_list,
net/sctp/sm_sideeffect.c
743
if (timer_delete(&t->hb_timer))
net/sctp/sm_sideeffect.c
744
sctp_transport_put(t);
net/sctp/sm_sideeffect.c
752
struct sctp_transport *t;
net/sctp/sm_sideeffect.c
754
list_for_each_entry(t, &asoc->peer.transport_addr_list,
net/sctp/sm_sideeffect.c
756
if (timer_delete(&t->T3_rtx_timer))
net/sctp/sm_sideeffect.c
757
sctp_transport_put(t);
net/sctp/sm_sideeffect.c
765
struct sctp_transport *t,
net/sctp/sm_sideeffect.c
775
t->error_count = 0;
net/sctp/sm_sideeffect.c
785
if (t->asoc->state < SCTP_STATE_SHUTDOWN_PENDING)
net/sctp/sm_sideeffect.c
786
t->asoc->overall_error_count = 0;
net/sctp/sm_sideeffect.c
791
t->hb_sent = 0;
net/sctp/sm_sideeffect.c
796
if ((t->state == SCTP_INACTIVE) || (t->state == SCTP_UNCONFIRMED)) {
net/sctp/sm_sideeffect.c
798
sctp_assoc_control_transport(asoc, t, SCTP_TRANSPORT_UP,
net/sctp/sm_sideeffect.c
802
if (t->state == SCTP_PF)
net/sctp/sm_sideeffect.c
803
sctp_assoc_control_transport(asoc, t, SCTP_TRANSPORT_UP,
net/sctp/sm_sideeffect.c
809
if (t->dst)
net/sctp/sm_sideeffect.c
810
sctp_transport_dst_confirm(t);
net/sctp/sm_sideeffect.c
819
if (t->rto_pending == 0)
net/sctp/sm_sideeffect.c
820
t->rto_pending = 1;
net/sctp/sm_sideeffect.c
823
sctp_transport_update_rto(t, (jiffies - hbinfo->sent_at));
net/sctp/sm_sideeffect.c
826
sctp_transport_reset_hb_timer(t);
net/sctp/sm_sideeffect.c
829
sctp_transport_immediate_rtx(t);
net/sctp/sm_sideeffect.c
858
struct sctp_transport *t;
net/sctp/sm_sideeffect.c
861
t = chunk->transport;
net/sctp/sm_sideeffect.c
863
t = sctp_assoc_choose_alter_transport(asoc,
net/sctp/sm_sideeffect.c
865
chunk->transport = t;
net/sctp/sm_sideeffect.c
867
asoc->shutdown_last_sent_to = t;
net/sctp/sm_sideeffect.c
868
asoc->timeouts[SCTP_EVENT_TIMEOUT_T2_SHUTDOWN] = t->rto;
net/sctp/sm_sideeffect.c
963
struct sctp_transport *t;
net/sctp/sm_sideeffect.c
965
t = sctp_assoc_choose_alter_transport(asoc, chunk->transport);
net/sctp/sm_sideeffect.c
966
asoc->timeouts[SCTP_EVENT_TIMEOUT_T4_RTO] = t->rto;
net/sctp/sm_sideeffect.c
967
chunk->transport = t;
net/sctp/socket.c
1140
struct sctp_transport *t;
net/sctp/socket.c
1147
old = sctp_endpoint_lookup_assoc(ep, daddr, &t);
net/sctp/socket.c
1155
t = sctp_assoc_add_peer(asoc, daddr, GFP_KERNEL, SCTP_UNKNOWN);
net/sctp/socket.c
1156
if (!t)
net/sctp/socket.c
174
struct sctp_transport *t;
net/sctp/socket.c
176
list_for_each_entry(t, &asoc->peer.transport_addr_list, transports)
net/sctp/socket.c
177
list_for_each_entry(chunk, &t->transmitted, transmitted_list)
net/sctp/socket.c
2572
struct sctp_transport *t;
net/sctp/socket.c
2574
list_for_each_entry(t, &asoc->peer.transport_addr_list,
net/sctp/socket.c
2576
if (t->ipaddr.sa.sa_family != AF_INET6)
net/sctp/socket.c
2578
t->flowlabel = params->spp_ipv6_flowlabel &
net/sctp/socket.c
2580
t->flowlabel |= SCTP_FLOWLABEL_SET_MASK;
net/sctp/socket.c
2597
struct sctp_transport *t;
net/sctp/socket.c
2599
list_for_each_entry(t, &asoc->peer.transport_addr_list,
net/sctp/socket.c
2601
t->dscp = params->spp_dscp &
net/sctp/socket.c
2603
t->dscp |= SCTP_DSCP_SET_MASK;
net/sctp/socket.c
4450
struct sctp_transport *t;
net/sctp/socket.c
4461
t = sctp_addr_id2transport(sk, &encap->sue_address,
net/sctp/socket.c
4463
if (!t)
net/sctp/socket.c
4466
t->encap_port = encap_port;
net/sctp/socket.c
4483
list_for_each_entry(t, &asoc->peer.transport_addr_list,
net/sctp/socket.c
4485
t->encap_port = encap_port;
net/sctp/socket.c
4500
struct sctp_transport *t;
net/sctp/socket.c
4514
t = sctp_addr_id2transport(sk, &params->spi_address,
net/sctp/socket.c
4516
if (!t)
net/sctp/socket.c
4519
t->probe_interval = msecs_to_jiffies(probe_interval);
net/sctp/socket.c
4520
sctp_transport_pl_reset(t);
net/sctp/socket.c
4537
list_for_each_entry(t, &asoc->peer.transport_addr_list, transports) {
net/sctp/socket.c
4538
t->probe_interval = msecs_to_jiffies(probe_interval);
net/sctp/socket.c
4539
sctp_transport_pl_reset(t);
net/sctp/socket.c
5320
struct sctp_transport *t;
net/sctp/socket.c
5322
t = rhashtable_walk_next(iter);
net/sctp/socket.c
5323
for (; t; t = rhashtable_walk_next(iter)) {
net/sctp/socket.c
5324
if (IS_ERR(t)) {
net/sctp/socket.c
5325
if (PTR_ERR(t) == -EAGAIN)
net/sctp/socket.c
5330
if (!sctp_transport_hold(t))
net/sctp/socket.c
5333
if (net_eq(t->asoc->base.net, net) &&
net/sctp/socket.c
5334
t->asoc->peer.primary_path == t)
net/sctp/socket.c
5337
sctp_transport_put(t);
net/sctp/socket.c
5340
return t;
net/sctp/socket.c
5347
struct sctp_transport *t;
net/sctp/socket.c
5352
while ((t = sctp_transport_get_next(net, iter)) && !IS_ERR(t)) {
net/sctp/socket.c
5355
sctp_transport_put(t);
net/sctp/socket.c
5358
return t;
net/sctp/socket.c
7991
struct sctp_transport *t;
net/sctp/socket.c
8005
t = sctp_addr_id2transport(sk, &encap.sue_address,
net/sctp/socket.c
8007
if (!t) {
net/sctp/socket.c
8012
encap_port = t->encap_port;
net/sctp/socket.c
8051
struct sctp_transport *t;
net/sctp/socket.c
8065
t = sctp_addr_id2transport(sk, &params.spi_address,
net/sctp/socket.c
8067
if (!t) {
net/sctp/socket.c
8072
probe_interval = jiffies_to_msecs(t->probe_interval);
net/sctp/stream.c
1078
t = asoc->strreset_chunk->transport;
net/sctp/stream.c
1079
if (timer_delete(&t->reconf_timer))
net/sctp/stream.c
1080
sctp_transport_put(t);
net/sctp/stream.c
576
struct sctp_transport *t;
net/sctp/stream.c
578
t = asoc->strreset_chunk->transport;
net/sctp/stream.c
579
if (timer_delete(&t->reconf_timer))
net/sctp/stream.c
580
sctp_transport_put(t);
net/sctp/stream.c
825
struct sctp_transport *t;
net/sctp/stream.c
827
t = asoc->strreset_chunk->transport;
net/sctp/stream.c
828
if (timer_delete(&t->reconf_timer))
net/sctp/stream.c
829
sctp_transport_put(t);
net/sctp/stream.c
921
struct sctp_transport *t;
net/sctp/transport.c
257
void sctp_transport_pl_send(struct sctp_transport *t)
net/sctp/transport.c
259
if (t->pl.probe_count < SCTP_MAX_PROBES)
net/sctp/transport.c
262
t->pl.probe_count = 0;
net/sctp/transport.c
263
if (t->pl.state == SCTP_PL_BASE) {
net/sctp/transport.c
264
if (t->pl.probe_size == SCTP_BASE_PLPMTU) { /* BASE_PLPMTU Confirmation Failed */
net/sctp/transport.c
265
t->pl.state = SCTP_PL_ERROR; /* Base -> Error */
net/sctp/transport.c
267
t->pl.pmtu = SCTP_BASE_PLPMTU;
net/sctp/transport.c
268
t->pathmtu = t->pl.pmtu + sctp_transport_pl_hlen(t);
net/sctp/transport.c
269
sctp_assoc_sync_pmtu(t->asoc);
net/sctp/transport.c
271
} else if (t->pl.state == SCTP_PL_SEARCH) {
net/sctp/transport.c
272
if (t->pl.pmtu == t->pl.probe_size) { /* Black Hole Detected */
net/sctp/transport.c
273
t->pl.state = SCTP_PL_BASE; /* Search -> Base */
net/sctp/transport.c
274
t->pl.probe_size = SCTP_BASE_PLPMTU;
net/sctp/transport.c
275
t->pl.probe_high = 0;
net/sctp/transport.c
277
t->pl.pmtu = SCTP_BASE_PLPMTU;
net/sctp/transport.c
278
t->pathmtu = t->pl.pmtu + sctp_transport_pl_hlen(t);
net/sctp/transport.c
279
sctp_assoc_sync_pmtu(t->asoc);
net/sctp/transport.c
281
t->pl.probe_high = t->pl.probe_size;
net/sctp/transport.c
282
t->pl.probe_size = t->pl.pmtu;
net/sctp/transport.c
284
} else if (t->pl.state == SCTP_PL_COMPLETE) {
net/sctp/transport.c
285
if (t->pl.pmtu == t->pl.probe_size) { /* Black Hole Detected */
net/sctp/transport.c
286
t->pl.state = SCTP_PL_BASE; /* Search Complete -> Base */
net/sctp/transport.c
287
t->pl.probe_size = SCTP_BASE_PLPMTU;
net/sctp/transport.c
289
t->pl.pmtu = SCTP_BASE_PLPMTU;
net/sctp/transport.c
290
t->pathmtu = t->pl.pmtu + sctp_transport_pl_hlen(t);
net/sctp/transport.c
291
sctp_assoc_sync_pmtu(t->asoc);
net/sctp/transport.c
297
__func__, t, t->pl.state, t->pl.pmtu, t->pl.probe_size, t->pl.probe_high);
net/sctp/transport.c
298
t->pl.probe_count++;
net/sctp/transport.c
301
bool sctp_transport_pl_recv(struct sctp_transport *t)
net/sctp/transport.c
304
__func__, t, t->pl.state, t->pl.pmtu, t->pl.probe_size, t->pl.probe_high);
net/sctp/transport.c
306
t->pl.pmtu = t->pl.probe_size;
net/sctp/transport.c
307
t->pl.probe_count = 0;
net/sctp/transport.c
308
if (t->pl.state == SCTP_PL_BASE) {
net/sctp/transport.c
309
t->pl.state = SCTP_PL_SEARCH; /* Base -> Search */
net/sctp/transport.c
310
t->pl.probe_size += SCTP_PL_BIG_STEP;
net/sctp/transport.c
311
} else if (t->pl.state == SCTP_PL_ERROR) {
net/sctp/transport.c
312
t->pl.state = SCTP_PL_SEARCH; /* Error -> Search */
net/sctp/transport.c
314
t->pl.pmtu = t->pl.probe_size;
net/sctp/transport.c
315
t->pathmtu = t->pl.pmtu + sctp_transport_pl_hlen(t);
net/sctp/transport.c
316
sctp_assoc_sync_pmtu(t->asoc);
net/sctp/transport.c
317
t->pl.probe_size += SCTP_PL_BIG_STEP;
net/sctp/transport.c
318
} else if (t->pl.state == SCTP_PL_SEARCH) {
net/sctp/transport.c
319
if (!t->pl.probe_high) {
net/sctp/transport.c
320
if (t->pl.probe_size < SCTP_MAX_PLPMTU) {
net/sctp/transport.c
321
t->pl.probe_size = min(t->pl.probe_size + SCTP_PL_BIG_STEP,
net/sctp/transport.c
325
t->pl.probe_high = SCTP_MAX_PLPMTU;
net/sctp/transport.c
327
t->pl.probe_size += SCTP_PL_MIN_STEP;
net/sctp/transport.c
328
if (t->pl.probe_size >= t->pl.probe_high) {
net/sctp/transport.c
329
t->pl.probe_high = 0;
net/sctp/transport.c
330
t->pl.state = SCTP_PL_COMPLETE; /* Search -> Search Complete */
net/sctp/transport.c
332
t->pl.probe_size = t->pl.pmtu;
net/sctp/transport.c
333
t->pathmtu = t->pl.pmtu + sctp_transport_pl_hlen(t);
net/sctp/transport.c
334
sctp_assoc_sync_pmtu(t->asoc);
net/sctp/transport.c
335
sctp_transport_reset_raise_timer(t);
net/sctp/transport.c
337
} else if (t->pl.state == SCTP_PL_COMPLETE) {
net/sctp/transport.c
339
t->pl.state = SCTP_PL_SEARCH; /* Search Complete -> Search */
net/sctp/transport.c
340
t->pl.probe_size = min(t->pl.probe_size + SCTP_PL_MIN_STEP, SCTP_MAX_PLPMTU);
net/sctp/transport.c
343
return t->pl.state == SCTP_PL_COMPLETE;
net/sctp/transport.c
346
static bool sctp_transport_pl_toobig(struct sctp_transport *t, u32 pmtu)
net/sctp/transport.c
349
__func__, t, t->pl.state, t->pl.pmtu, t->pl.probe_size, pmtu);
net/sctp/transport.c
351
if (pmtu < SCTP_MIN_PLPMTU || pmtu >= t->pl.probe_size)
net/sctp/transport.c
354
if (t->pl.state == SCTP_PL_BASE) {
net/sctp/transport.c
356
t->pl.state = SCTP_PL_ERROR; /* Base -> Error */
net/sctp/transport.c
358
t->pl.pmtu = SCTP_BASE_PLPMTU;
net/sctp/transport.c
359
t->pathmtu = t->pl.pmtu + sctp_transport_pl_hlen(t);
net/sctp/transport.c
362
} else if (t->pl.state == SCTP_PL_SEARCH) {
net/sctp/transport.c
363
if (pmtu >= SCTP_BASE_PLPMTU && pmtu < t->pl.pmtu) {
net/sctp/transport.c
364
t->pl.state = SCTP_PL_BASE; /* Search -> Base */
net/sctp/transport.c
365
t->pl.probe_size = SCTP_BASE_PLPMTU;
net/sctp/transport.c
366
t->pl.probe_count = 0;
net/sctp/transport.c
368
t->pl.probe_high = 0;
net/sctp/transport.c
369
t->pl.pmtu = SCTP_BASE_PLPMTU;
net/sctp/transport.c
370
t->pathmtu = t->pl.pmtu + sctp_transport_pl_hlen(t);
net/sctp/transport.c
372
} else if (pmtu > t->pl.pmtu && pmtu < t->pl.probe_size) {
net/sctp/transport.c
373
t->pl.probe_size = pmtu;
net/sctp/transport.c
374
t->pl.probe_count = 0;
net/sctp/transport.c
376
} else if (t->pl.state == SCTP_PL_COMPLETE) {
net/sctp/transport.c
377
if (pmtu >= SCTP_BASE_PLPMTU && pmtu < t->pl.pmtu) {
net/sctp/transport.c
378
t->pl.state = SCTP_PL_BASE; /* Complete -> Base */
net/sctp/transport.c
379
t->pl.probe_size = SCTP_BASE_PLPMTU;
net/sctp/transport.c
380
t->pl.probe_count = 0;
net/sctp/transport.c
382
t->pl.probe_high = 0;
net/sctp/transport.c
383
t->pl.pmtu = SCTP_BASE_PLPMTU;
net/sctp/transport.c
384
t->pathmtu = t->pl.pmtu + sctp_transport_pl_hlen(t);
net/sctp/transport.c
385
sctp_transport_reset_probe_timer(t);
net/sctp/transport.c
393
bool sctp_transport_update_pmtu(struct sctp_transport *t, u32 pmtu)
net/sctp/transport.c
395
struct sock *sk = t->asoc->base.sk;
net/sctp/transport.c
407
if (sctp_transport_pl_enabled(t))
net/sctp/transport.c
408
return sctp_transport_pl_toobig(t, pmtu - sctp_transport_pl_hlen(t));
net/sctp/transport.c
410
dst = sctp_transport_dst_check(t);
net/sctp/transport.c
416
pf->to_sk_daddr(&t->ipaddr, sk);
net/sctp/transport.c
420
dst = sctp_transport_dst_check(t);
net/sctp/transport.c
424
t->af_specific->get_dst(t, &t->saddr, &t->fl, sk);
net/sctp/transport.c
425
dst = t->dst;
net/sctp/transport.c
431
change = t->pathmtu != pmtu;
net/sctp/transport.c
433
t->pathmtu = pmtu;
net/sctp/transport.c
753
void sctp_transport_burst_limited(struct sctp_transport *t)
net/sctp/transport.c
755
struct sctp_association *asoc = t->asoc;
net/sctp/transport.c
756
u32 old_cwnd = t->cwnd;
net/sctp/transport.c
759
if (t->burst_limited || asoc->max_burst == 0)
net/sctp/transport.c
762
max_burst_bytes = t->flight_size + (asoc->max_burst * asoc->pathmtu);
net/sctp/transport.c
764
t->cwnd = max_burst_bytes;
net/sctp/transport.c
765
t->burst_limited = old_cwnd;
net/sctp/transport.c
772
void sctp_transport_burst_reset(struct sctp_transport *t)
net/sctp/transport.c
774
if (t->burst_limited) {
net/sctp/transport.c
775
t->cwnd = t->burst_limited;
net/sctp/transport.c
776
t->burst_limited = 0;
net/sctp/transport.c
794
void sctp_transport_reset(struct sctp_transport *t)
net/sctp/transport.c
796
struct sctp_association *asoc = t->asoc;
net/sctp/transport.c
803
t->cwnd = min(4*asoc->pathmtu, max_t(__u32, 2*asoc->pathmtu, 4380));
net/sctp/transport.c
804
t->burst_limited = 0;
net/sctp/transport.c
805
t->ssthresh = asoc->peer.i.a_rwnd;
net/sctp/transport.c
806
t->rto = asoc->rto_initial;
net/sctp/transport.c
807
sctp_max_rto(asoc, t);
net/sctp/transport.c
808
t->rtt = 0;
net/sctp/transport.c
809
t->srtt = 0;
net/sctp/transport.c
810
t->rttvar = 0;
net/sctp/transport.c
813
t->partial_bytes_acked = 0;
net/sctp/transport.c
814
t->flight_size = 0;
net/sctp/transport.c
815
t->error_count = 0;
net/sctp/transport.c
816
t->rto_pending = 0;
net/sctp/transport.c
817
t->hb_sent = 0;
net/sctp/transport.c
820
t->cacc.changeover_active = 0;
net/sctp/transport.c
821
t->cacc.cycling_changeover = 0;
net/sctp/transport.c
822
t->cacc.next_tsn_at_change = 0;
net/sctp/transport.c
823
t->cacc.cacc_saw_newack = 0;
net/sctp/transport.c
827
void sctp_transport_immediate_rtx(struct sctp_transport *t)
net/sctp/transport.c
830
if (timer_delete(&t->T3_rtx_timer))
net/sctp/transport.c
831
sctp_transport_put(t);
net/sctp/transport.c
833
sctp_retransmit(&t->asoc->outqueue, t, SCTP_RTXR_T3_RTX);
net/sctp/transport.c
834
if (!timer_pending(&t->T3_rtx_timer)) {
net/sctp/transport.c
835
if (!mod_timer(&t->T3_rtx_timer, jiffies + t->rto))
net/sctp/transport.c
836
sctp_transport_hold(t);
net/sctp/transport.c
841
void sctp_transport_dst_release(struct sctp_transport *t)
net/sctp/transport.c
843
dst_release(t->dst);
net/sctp/transport.c
844
t->dst = NULL;
net/sctp/transport.c
845
t->dst_pending_confirm = 0;
net/sctp/transport.c
849
void sctp_transport_dst_confirm(struct sctp_transport *t)
net/sctp/transport.c
851
t->dst_pending_confirm = 1;
net/smc/smc.h
388
__be32 t;
net/smc/smc.h
390
t = cpu_to_be32(host);
net/smc/smc.h
391
memcpy(net, ((u8 *)&t) + 1, 3);
net/smc/smc.h
397
__be32 t = 0;
net/smc/smc.h
399
memcpy(((u8 *)&t) + 1, net, 3);
net/smc/smc.h
400
return be32_to_cpu(t);
net/smc/smc_cdc.c
442
static void smcd_cdc_rx_tsklet(struct tasklet_struct *t)
net/smc/smc_cdc.c
444
struct smc_connection *conn = from_tasklet(conn, t, rx_tsklet);
net/smc/smc_hs_bpf.c
85
static int smc_bpf_hs_ctrl_init_member(const struct btf_type *t,
net/smc/smc_hs_bpf.c
96
moff = __btf_member_bit_offset(t, member) / 8;
net/smc/smc_stats.h
100
this_cpu_inc((*stats).smc[t].key ## _cnt); \
net/smc/smc_stats.h
105
this_cpu_inc((*stats).smc[t].key ## _pd.buf[_pos]); \
net/smc/smc_stats.h
106
this_cpu_add((*stats).smc[t].key ## _bytes, r); \
net/smc/smc_stats.h
145
typeof(_tech) t = (_tech); \
net/smc/smc_stats.h
153
this_cpu_inc((*stats).smc[t].k ## _rmbsize.buf[_pos]); \
net/smc/smc_stats.h
154
this_cpu_add((*stats).smc[t].k ## _rmbuse, _l); \
net/smc/smc_stats.h
156
this_cpu_sub((*stats).smc[t].k ## _rmbuse, _l); \
net/smc/smc_stats.h
161
#define SMC_STAT_RMB_SUB(_smc_stats, type, t, key) \
net/smc/smc_stats.h
162
this_cpu_inc((*(_smc_stats)).smc[t].rmb ## _ ## key.type ## _cnt)
net/smc/smc_stats.h
95
typeof(_tech) t = (_tech); \
net/smc/smc_wr.c
136
static void smc_wr_tx_tasklet_fn(struct tasklet_struct *t)
net/smc/smc_wr.c
138
struct smc_ib_device *dev = from_tasklet(dev, t, send_tasklet);
net/smc/smc_wr.c
479
static void smc_wr_rx_tasklet_fn(struct tasklet_struct *t)
net/smc/smc_wr.c
481
struct smc_ib_device *dev = from_tasklet(dev, t, recv_tasklet);
net/sunrpc/sched.c
149
struct rpc_task *t;
net/sunrpc/sched.c
151
list_for_each_entry(t, q, u.tk_wait.list) {
net/sunrpc/sched.c
152
if (t->tk_owner == task->tk_owner) {
net/sunrpc/sched.c
154
&t->u.tk_wait.links);
net/sunrpc/sched.c
172
struct rpc_task *t;
net/sunrpc/sched.c
179
t = list_first_entry(&task->u.tk_wait.links,
net/sunrpc/sched.c
183
q = t->u.tk_wait.list.next;
net/sunrpc/sched.c
184
list_add_tail(&t->u.tk_wait.list, q);
net/sunrpc/svc_xprt.c
33
static void svc_age_temp_xprts(struct timer_list *t);
net/sunrpc/svc_xprt.c
956
static void svc_age_temp_xprts(struct timer_list *t)
net/sunrpc/svc_xprt.c
958
struct svc_serv *serv = timer_container_of(serv, t, sv_temptimer);
net/sunrpc/svcsock.c
348
size_t t;
net/sunrpc/svcsock.c
352
for (i = 0, t = 0; t < buflen; i++, t += PAGE_SIZE)
net/sunrpc/xprt.c
104
struct xprt_class *t;
net/sunrpc/xprt.c
109
list_for_each_entry(t, &xprt_list, list) {
net/sunrpc/xprt.c
111
if (t->ident == transport->ident)
net/sunrpc/xprt.c
136
struct xprt_class *t;
net/sunrpc/xprt.c
141
list_for_each_entry(t, &xprt_list, list) {
net/sunrpc/xprt.c
142
if (t == transport) {
net/sunrpc/xprt.c
159
xprt_class_release(const struct xprt_class *t)
net/sunrpc/xprt.c
161
module_put(t->owner);
net/sunrpc/xprt.c
167
const struct xprt_class *t;
net/sunrpc/xprt.c
169
list_for_each_entry(t, &xprt_list, list) {
net/sunrpc/xprt.c
170
if (t->ident != ident)
net/sunrpc/xprt.c
172
if (!try_module_get(t->owner))
net/sunrpc/xprt.c
174
return t;
net/sunrpc/xprt.c
182
const struct xprt_class *t;
net/sunrpc/xprt.c
185
t = xprt_class_find_by_ident_locked(ident);
net/sunrpc/xprt.c
187
return t;
net/sunrpc/xprt.c
193
const struct xprt_class *t;
net/sunrpc/xprt.c
196
list_for_each_entry(t, &xprt_list, list) {
net/sunrpc/xprt.c
197
for (i = 0; t->netid[i][0] != '\0'; i++) {
net/sunrpc/xprt.c
198
if (strcmp(t->netid[i], netid) != 0)
net/sunrpc/xprt.c
200
if (!try_module_get(t->owner))
net/sunrpc/xprt.c
202
return t;
net/sunrpc/xprt.c
2062
const struct xprt_class *t;
net/sunrpc/xprt.c
2064
t = xprt_class_find_by_ident(args->ident);
net/sunrpc/xprt.c
2065
if (!t) {
net/sunrpc/xprt.c
2070
xprt = t->setup(args);
net/sunrpc/xprt.c
2071
xprt_class_release(t);
net/sunrpc/xprt.c
211
const struct xprt_class *t;
net/sunrpc/xprt.c
214
t = xprt_class_find_by_netid_locked(netid);
net/sunrpc/xprt.c
215
if (!t) {
net/sunrpc/xprt.c
219
t = xprt_class_find_by_netid_locked(netid);
net/sunrpc/xprt.c
222
return t;
net/sunrpc/xprt.c
235
const struct xprt_class *t;
net/sunrpc/xprt.c
238
t = xprt_class_find_by_netid(netid);
net/sunrpc/xprt.c
239
if (!t)
net/sunrpc/xprt.c
241
ret = t->ident;
net/sunrpc/xprt.c
242
xprt_class_release(t);
net/sunrpc/xprt.c
855
xprt_init_autodisconnect(struct timer_list *t)
net/sunrpc/xprt.c
857
struct rpc_xprt *xprt = timer_container_of(xprt, t, timer);
net/sunrpc/xprtsock.c
2290
unsigned long t;
net/sunrpc/xprtsock.c
2314
for (t = 0; t <= syn_retries && (1UL << t) < connect_timeout; t++)
net/sunrpc/xprtsock.c
2316
if (t <= syn_retries)
net/sunrpc/xprtsock.c
2317
tcp_sock_set_syncnt(sock->sk, t - 1);
net/tipc/discover.c
293
static void tipc_disc_timeout(struct timer_list *t)
net/tipc/discover.c
295
struct tipc_discoverer *d = timer_container_of(d, t, timer);
net/tipc/monitor.c
631
static void mon_timeout(struct timer_list *t)
net/tipc/monitor.c
633
struct tipc_monitor *mon = timer_container_of(mon, t, timer);
net/tipc/node.c
181
static void tipc_node_timeout(struct timer_list *t);
net/tipc/node.c
801
static void tipc_node_timeout(struct timer_list *t)
net/tipc/node.c
803
struct tipc_node *n = timer_container_of(n, t, timer);
net/tipc/socket.c
149
static void tipc_sk_timeout(struct timer_list *t);
net/tipc/socket.c
2865
static void tipc_sk_timeout(struct timer_list *t)
net/tipc/socket.c
2867
struct sock *sk = timer_container_of(sk, t, sk_timer);
net/tipc/subscr.c
106
static void tipc_sub_timeout(struct timer_list *t)
net/tipc/subscr.c
108
struct tipc_subscription *sub = timer_container_of(sub, t, timer);
net/tipc/trace.h
284
TP_PROTO(struct tipc_link *r, u16 f, u16 t, struct sk_buff_head *tq),
net/tipc/trace.h
286
TP_ARGS(r, f, t, tq),
net/tipc/trace.h
300
__entry->to = t;
net/tipc/trace.h
314
TP_PROTO(struct tipc_link *r, u16 f, u16 t, struct sk_buff_head *tq),
net/tipc/trace.h
315
TP_ARGS(r, f, t, tq),
net/tipc/trace.h
316
TP_CONDITION(less_eq(f, t))
net/tipc/trace.h
320
TP_PROTO(struct tipc_link *r, u16 f, u16 t, struct sk_buff_head *tq),
net/tipc/trace.h
321
TP_ARGS(r, f, t, tq),
net/vmw_vsock/af_vsock.c
3014
int vsock_core_register(const struct vsock_transport *t, int features)
net/vmw_vsock/af_vsock.c
3032
t_h2g = t;
net/vmw_vsock/af_vsock.c
3040
t_g2h = t;
net/vmw_vsock/af_vsock.c
3048
t_dgram = t;
net/vmw_vsock/af_vsock.c
3056
t_local = t;
net/vmw_vsock/af_vsock.c
3070
void vsock_core_unregister(const struct vsock_transport *t)
net/vmw_vsock/af_vsock.c
3074
if (transport_h2g == t)
net/vmw_vsock/af_vsock.c
3077
if (transport_g2h == t)
net/vmw_vsock/af_vsock.c
3080
if (transport_dgram == t)
net/vmw_vsock/af_vsock.c
3083
if (transport_local == t)
net/vmw_vsock/virtio_transport_common.c
1175
static int virtio_transport_reset_no_sock(const struct virtio_transport *t,
net/vmw_vsock/virtio_transport_common.c
1202
if (!t)
net/vmw_vsock/virtio_transport_common.c
1213
return t->send_pkt(reply, net);
net/vmw_vsock/virtio_transport_common.c
1531
struct virtio_transport *t)
net/vmw_vsock/virtio_transport_common.c
1540
virtio_transport_reset_no_sock(t, skb, sock_net(sk));
net/vmw_vsock/virtio_transport_common.c
1545
virtio_transport_reset_no_sock(t, skb, sock_net(sk));
net/vmw_vsock/virtio_transport_common.c
1553
virtio_transport_reset_no_sock(t, skb, sock_net(sk));
net/vmw_vsock/virtio_transport_common.c
1559
virtio_transport_reset_no_sock(t, skb, sock_net(sk));
net/vmw_vsock/virtio_transport_common.c
1579
if (ret || vchild->transport != &t->transport) {
net/vmw_vsock/virtio_transport_common.c
1581
virtio_transport_reset_no_sock(t, skb, sock_net(sk));
net/vmw_vsock/virtio_transport_common.c
1608
void virtio_transport_recv_pkt(struct virtio_transport *t,
net/vmw_vsock/virtio_transport_common.c
1632
(void)virtio_transport_reset_no_sock(t, skb, net);
net/vmw_vsock/virtio_transport_common.c
1643
(void)virtio_transport_reset_no_sock(t, skb, net);
net/vmw_vsock/virtio_transport_common.c
1649
(void)virtio_transport_reset_no_sock(t, skb, net);
net/vmw_vsock/virtio_transport_common.c
1667
(sk->sk_state != TCP_LISTEN && vsk->transport != &t->transport)) {
net/vmw_vsock/virtio_transport_common.c
1668
(void)virtio_transport_reset_no_sock(t, skb, net);
net/vmw_vsock/virtio_transport_common.c
1685
virtio_transport_recv_listen(sk, skb, t);
net/vmw_vsock/virtio_transport_common.c
1700
(void)virtio_transport_reset_no_sock(t, skb, net);
net/vmw_vsock/virtio_transport_common.c
36
const struct vsock_transport *t = vsock_core_get_transport(vsk);
net/vmw_vsock/virtio_transport_common.c
38
if (WARN_ON(!t))
net/vmw_vsock/virtio_transport_common.c
41
return container_of(t, struct virtio_transport, transport);
net/wireless/core.c
1762
void wiphy_delayed_work_timer(struct timer_list *t)
net/wireless/core.c
1764
struct wiphy_delayed_work *dwork = timer_container_of(dwork, t, timer);
net/wireless/core.c
1814
enum hrtimer_restart wiphy_hrtimer_work_timer(struct hrtimer *t)
net/wireless/core.c
1817
container_of(t, struct wiphy_hrtimer_work, timer);
net/x25/af_x25.c
360
static void x25_destroy_timer(struct timer_list *t)
net/x25/af_x25.c
362
struct sock *sk = timer_container_of(sk, t, sk_timer);
net/x25/x25_link.c
47
static void x25_t20timer_expiry(struct timer_list *t)
net/x25/x25_link.c
49
struct x25_neigh *nb = timer_container_of(nb, t, t20timer);
net/x25/x25_timer.c
157
static void x25_timer_expiry(struct timer_list *t)
net/x25/x25_timer.c
159
struct x25_sock *x25 = timer_container_of(x25, t, timer);
net/x25/x25_timer.c
24
static void x25_heartbeat_expiry(struct timer_list *t);
net/x25/x25_timer.c
25
static void x25_timer_expiry(struct timer_list *t);
net/x25/x25_timer.c
90
static void x25_heartbeat_expiry(struct timer_list *t)
net/x25/x25_timer.c
92
struct sock *sk = timer_container_of(sk, t, sk_timer);
net/xfrm/xfrm_policy.c
1170
struct xfrm_pol_inexact_bin *bin, *t;
net/xfrm/xfrm_policy.c
1174
list_for_each_entry_safe(bin, t, &net->xfrm.inexact_bins, inexact_bins)
net/xfrm/xfrm_policy.c
188
static void xfrm_policy_queue_process(struct timer_list *t);
net/xfrm/xfrm_policy.c
2896
static void xfrm_policy_queue_process(struct timer_list *t)
net/xfrm/xfrm_policy.c
2901
struct xfrm_policy *pol = timer_container_of(pol, t, polq.hold_timer);
net/xfrm/xfrm_policy.c
354
static void xfrm_policy_timer(struct timer_list *t)
net/xfrm/xfrm_policy.c
356
struct xfrm_policy *xp = timer_container_of(xp, t, timer);
net/xfrm/xfrm_policy.c
4281
struct xfrm_pol_inexact_bin *b, *t;
net/xfrm/xfrm_policy.c
4311
list_for_each_entry_safe(b, t, &net->xfrm.inexact_bins, inexact_bins)
net/xfrm/xfrm_policy.c
4536
static int migrate_tmpl_match(const struct xfrm_migrate *m, const struct xfrm_tmpl *t)
net/xfrm/xfrm_policy.c
4540
if (t->mode == m->mode && t->id.proto == m->proto &&
net/xfrm/xfrm_policy.c
4541
(m->reqid == 0 || t->reqid == m->reqid)) {
net/xfrm/xfrm_policy.c
4542
switch (t->mode) {
net/xfrm/xfrm_policy.c
4546
if (xfrm_addr_equal(&t->id.daddr, &m->old_daddr,
net/xfrm/xfrm_policy.c
4548
xfrm_addr_equal(&t->saddr, &m->old_saddr,
net/xfrm/xfrm_state.c
232
int km_query(struct xfrm_state *x, struct xfrm_tmpl *t, struct xfrm_policy *pol);
net/xfrm/xfrm_state.c
2718
static void xfrm_replay_timer_handler(struct timer_list *t)
net/xfrm/xfrm_state.c
2720
struct xfrm_state *x = timer_container_of(x, t, rtimer);
net/xfrm/xfrm_state.c
2775
int km_query(struct xfrm_state *x, struct xfrm_tmpl *t, struct xfrm_policy *pol)
net/xfrm/xfrm_state.c
2782
acqret = km->acquire(x, t, pol);
net/xfrm/xfrm_state.c
3103
struct xfrm_state *t = x->tunnel;
net/xfrm/xfrm_state.c
3105
if (atomic_dec_return(&t->tunnel_users) == 1)
net/xfrm/xfrm_state.c
3106
xfrm_state_delete(t);
net/xfrm/xfrm_state.c
3107
xfrm_state_put(t);
net/xfrm/xfrm_state.c
729
static void xfrm_replay_timer_handler(struct timer_list *t);
net/xfrm/xfrm_user.c
2035
struct xfrm_tmpl *t = &xp->xfrm_vec[i];
net/xfrm/xfrm_user.c
2037
memcpy(&t->id, &ut->id, sizeof(struct xfrm_id));
net/xfrm/xfrm_user.c
2038
memcpy(&t->saddr, &ut->saddr,
net/xfrm/xfrm_user.c
2040
t->reqid = ut->reqid;
net/xfrm/xfrm_user.c
2041
t->mode = ut->mode;
net/xfrm/xfrm_user.c
2042
t->share = ut->share;
net/xfrm/xfrm_user.c
2043
t->optional = ut->optional;
net/xfrm/xfrm_user.c
2044
t->aalgos = ut->aalgos;
net/xfrm/xfrm_user.c
2045
t->ealgos = ut->ealgos;
net/xfrm/xfrm_user.c
2046
t->calgos = ut->calgos;
net/xfrm/xfrm_user.c
2048
t->allalgs = !~(t->aalgos & t->ealgos & t->calgos);
net/xfrm/xfrm_user.c
2049
t->encap_family = ut->family;
net/xfrm/xfrm_user.c
3046
struct xfrm_tmpl *t = &xp->xfrm_vec[i];
net/xfrm/xfrm_user.c
3047
memcpy(&x->id, &t->id, sizeof(x->id));
net/xfrm/xfrm_user.c
3048
x->props.mode = t->mode;
net/xfrm/xfrm_user.c
3049
x->props.reqid = t->reqid;
net/xfrm/xfrm_user.c
3051
t->aalgos = ua->aalgos;
net/xfrm/xfrm_user.c
3052
t->ealgos = ua->ealgos;
net/xfrm/xfrm_user.c
3053
t->calgos = ua->calgos;
net/xfrm/xfrm_user.c
3054
err = km_query(x, t, xp);
rust/helpers/signal.c
5
__rust_helper int rust_helper_signal_pending(struct task_struct *t)
rust/helpers/signal.c
7
return signal_pending(t);
rust/helpers/task.c
16
__rust_helper void rust_helper_get_task_struct(struct task_struct *t)
rust/helpers/task.c
18
get_task_struct(t);
rust/helpers/task.c
21
__rust_helper void rust_helper_put_task_struct(struct task_struct *t)
rust/helpers/task.c
23
put_task_struct(t);
samples/bpf/map_perf_test_user.c
22
#define TEST_BIT(t) (1U << (t))
samples/bpf/map_perf_test_user.c
77
static int check_test_flags(enum test_type t)
samples/bpf/map_perf_test_user.c
79
return test_flags & TEST_BIT(t);
samples/bpf/xdp_sample_user.c
1009
double t;
samples/bpf/xdp_sample_user.c
1044
t = calc_period(r, p);
samples/bpf/xdp_sample_user.c
1045
pps = calc_pps(&r->total, &p->total, t);
samples/bpf/xdp_sample_user.c
1046
drop = calc_drop_pps(&r->total, &p->total, t);
samples/bpf/xdp_sample_user.c
1047
info = calc_info_pps(&r->total, &p->total, t);
samples/bpf/xdp_sample_user.c
1050
err = calc_errs_pps(&r->total, &p->total, t);
samples/bpf/xdp_sample_user.c
1083
pps = calc_pps(rc, pc, t);
samples/bpf/xdp_sample_user.c
1084
drop = calc_drop_pps(rc, pc, t);
samples/bpf/xdp_sample_user.c
1085
err = calc_errs_pps(rc, pc, t);
samples/bpf/xdp_sample_user.c
1091
info = calc_info_pps(rc, pc, t);
samples/bpf/xdp_sample_user.c
1502
__u64 t;
samples/bpf/xdp_sample_user.c
1504
ret = read(timerfd, &t, sizeof(t));
samples/bpf/xdp_sample_user.c
1516
const char *f, *t;
samples/bpf/xdp_sample_user.c
1518
f = t = NULL;
samples/bpf/xdp_sample_user.c
1522
t = to;
samples/bpf/xdp_sample_user.c
1524
snprintf(line, sizeof(line), "%s->%s", f ?: "?", t ?: "?");
samples/bpf/xdp_sample_user.c
194
struct timespec t;
samples/bpf/xdp_sample_user.c
197
res = clock_gettime(CLOCK_MONOTONIC, &t);
samples/bpf/xdp_sample_user.c
202
return (__u64)t.tv_sec * NANOSEC_PER_SEC + t.tv_nsec;
samples/bpf/xdp_sample_user.c
653
double t, pps, drop, err;
samples/bpf/xdp_sample_user.c
658
t = calc_period(rec, prev);
samples/bpf/xdp_sample_user.c
665
pps = calc_pps(r, p, t);
samples/bpf/xdp_sample_user.c
666
drop = calc_drop_pps(r, p, t);
samples/bpf/xdp_sample_user.c
667
err = calc_errs_pps(r, p, t);
samples/bpf/xdp_sample_user.c
678
pps = calc_pps(&rec->total, &prev->total, t);
samples/bpf/xdp_sample_user.c
679
drop = calc_drop_pps(&rec->total, &prev->total, t);
samples/bpf/xdp_sample_user.c
680
err = calc_errs_pps(&rec->total, &prev->total, t);
samples/bpf/xdp_sample_user.c
696
double t, pps, drop, err;
samples/bpf/xdp_sample_user.c
703
t = calc_period(rec, prev);
samples/bpf/xdp_sample_user.c
705
pps = calc_pps(&rec->total, &prev->total, t);
samples/bpf/xdp_sample_user.c
706
drop = calc_drop_pps(&rec->total, &prev->total, t);
samples/bpf/xdp_sample_user.c
707
err = calc_errs_pps(&rec->total, &prev->total, t);
samples/bpf/xdp_sample_user.c
728
pps = calc_pps(r, p, t);
samples/bpf/xdp_sample_user.c
729
drop = calc_drop_pps(r, p, t);
samples/bpf/xdp_sample_user.c
730
err = calc_errs_pps(r, p, t);
samples/bpf/xdp_sample_user.c
751
double t;
samples/bpf/xdp_sample_user.c
756
t = calc_period(rec, prev);
samples/bpf/xdp_sample_user.c
759
&xdp_redirect, t);
samples/bpf/xdp_sample_user.c
772
calc_xdp_pps(r, p, &xdp_pass, &xdp_drop, &xdp_redirect, t);
samples/bpf/xdp_sample_user.c
789
double t, pps, drop, err;
samples/bpf/xdp_sample_user.c
794
t = calc_period(rec, prev);
samples/bpf/xdp_sample_user.c
796
pps = calc_pps(&rec->total, &prev->total, t);
samples/bpf/xdp_sample_user.c
797
drop = calc_drop_pps(&rec->total, &prev->total, t);
samples/bpf/xdp_sample_user.c
798
err = calc_errs_pps(&rec->total, &prev->total, t);
samples/bpf/xdp_sample_user.c
809
pps = calc_pps(r, p, t);
samples/bpf/xdp_sample_user.c
810
drop = calc_drop_pps(r, p, t);
samples/bpf/xdp_sample_user.c
811
err = calc_errs_pps(r, p, t);
samples/bpf/xdp_sample_user.c
828
double t, pps;
samples/bpf/xdp_sample_user.c
833
t = calc_period(rec, prev);
samples/bpf/xdp_sample_user.c
839
pps = calc_pps(r, p, t);
samples/bpf/xdp_sample_user.c
848
pps = calc_pps(&rec->total, &prev->total, t);
samples/bpf/xdp_sample_user.c
860
double t, drop, sum = 0;
samples/bpf/xdp_sample_user.c
868
t = calc_period(rec, prev);
samples/bpf/xdp_sample_user.c
870
drop = calc_drop_pps(&rec->total, &prev->total, t);
samples/bpf/xdp_sample_user.c
885
drop = calc_drop_pps(r, p, t);
samples/bpf/xdp_sample_user.c
908
double t, drop, sum = 0;
samples/bpf/xdp_sample_user.c
915
t = calc_period(rec, prev);
samples/bpf/xdp_sample_user.c
917
drop = calc_drop_pps(&rec->total, &prev->total, t);
samples/bpf/xdp_sample_user.c
931
drop = calc_drop_pps(r, p, t);
samples/bpf/xdp_sample_user.c
955
double t;
samples/bpf/xdp_sample_user.c
960
t = calc_period(rec, prev);
samples/bpf/xdp_sample_user.c
966
pps = calc_pps(r, p, t);
samples/bpf/xdp_sample_user.c
967
drop = calc_drop_pps(r, p, t);
samples/bpf/xdp_sample_user.c
968
err = calc_errs_pps(r, p, t);
samples/bpf/xdp_sample_user.c
974
info = calc_info_pps(r, p, t);
samples/bpf/xdp_sample_user.c
983
pps = calc_pps(&rec->total, &prev->total, t);
samples/bpf/xdp_sample_user.c
984
drop = calc_drop_pps(&rec->total, &prev->total, t);
samples/bpf/xdp_sample_user.c
985
info = calc_info_pps(&rec->total, &prev->total, t);
samples/bpf/xdp_sample_user.c
988
err = calc_errs_pps(&rec->total, &prev->total, t);
samples/damon/prcl.c
40
struct damon_target *t;
samples/damon/prcl.c
42
damon_for_each_target(t, c) {
samples/damon/prcl.c
46
damon_for_each_region(r, t) {
samples/damon/wsse.c
41
struct damon_target *t;
samples/damon/wsse.c
43
damon_for_each_target(t, c) {
samples/damon/wsse.c
47
damon_for_each_region(r, t) {
samples/ftrace/ftrace-direct-modify.c
295
static int t;
samples/ftrace/ftrace-direct-modify.c
304
t ^= 1;
samples/ftrace/ftrace-direct-modify.c
305
ret = modify_ftrace_direct(&direct, tramps[t]);
samples/ftrace/ftrace-direct-modify.c
307
my_tramp = tramps[t];
samples/ftrace/ftrace-direct-multi-modify.c
337
static int t;
samples/ftrace/ftrace-direct-multi-modify.c
346
t ^= 1;
samples/ftrace/ftrace-direct-multi-modify.c
347
ret = modify_ftrace_direct(&direct, tramps[t]);
samples/ftrace/ftrace-direct-multi-modify.c
349
my_tramp = tramps[t];
scripts/asn1_compiler.c
1213
const struct type *t = e->type_def;
scripts/asn1_compiler.c
1215
const char *tname = t && t->name ? t->name->content : ".";
scripts/asn1_compiler.c
749
unsigned nr, t, n;
scripts/asn1_compiler.c
774
t = 0;
scripts/asn1_compiler.c
775
types[t].flags |= TYPE_BEGIN;
scripts/asn1_compiler.c
779
types[t].name = &token_list[n];
scripts/asn1_compiler.c
780
type_index[t] = &types[t];
scripts/asn1_compiler.c
781
t++;
scripts/asn1_compiler.c
784
types[t].name = &token_list[n + 1];
scripts/asn1_compiler.c
785
types[t].flags |= TYPE_STOP_MARKER;
scripts/dtc/dtc.h
145
#define for_each_marker_of_type(m, t) \
scripts/dtc/dtc.h
147
if ((m)->type == (t))
scripts/dtc/include-prefixes/dt-bindings/usb/pd.h
14
#define PDO_TYPE(t) ((t) << PDO_TYPE_SHIFT)
scripts/dtc/include-prefixes/dt-bindings/usb/pd.h
67
#define PDO_APDO_TYPE(t) ((t) << PDO_APDO_TYPE_SHIFT)
scripts/gcc-plugins/gcc-common.h
415
static inline void debug_tree(const_tree t)
scripts/gcc-plugins/gcc-common.h
417
debug_tree(CONST_CAST_TREE(t));
scripts/gcc-plugins/gcc-common.h
425
#define debug_tree(t) debug_tree(CONST_CAST_TREE(t))
scripts/genksyms/genksyms.c
138
static enum symbol_type map_to_ns(enum symbol_type t)
scripts/genksyms/genksyms.c
140
switch (t) {
scripts/genksyms/genksyms.c
150
return t;
scripts/recordmcount.h
278
uint_t t = 1 + strlen(mc_name) + _w(shstr->sh_size);
scripts/recordmcount.h
281
shstr->sh_size = _w(t);
scripts/recordmcount.h
283
t += sb.st_size;
scripts/recordmcount.h
284
t += (_align & -t); /* word-byte align */
scripts/recordmcount.h
285
new_e_shoff = t;
scripts/recordmcount.h
298
if (ulseek(t, SEEK_SET) < 0)
scripts/recordmcount.h
300
t += sizeof(Elf_Shdr) * old_shnum;
scripts/recordmcount.h
306
t += 2*sizeof(mcsec);
scripts/recordmcount.h
312
mcsec.sh_offset = _w(t);
scripts/recordmcount.h
327
mcsec.sh_offset = _w((void *)mlocp - (void *)mloc0 + t);
scripts/recordmcount.h
417
unsigned t;
scripts/recordmcount.h
421
for (t = nrel; t; --t) {
scripts/recordmcount.h
464
unsigned t;
scripts/recordmcount.h
469
for (t = nrel; t; --t) {
scripts/recordmcount.h
529
unsigned t;
scripts/recordmcount.h
531
for (symp = sym0, t = nsym; t; --t, ++symp) {
scripts/unifdef.c
1190
strlcmp(const char *s, const char *t, size_t n)
scripts/unifdef.c
1192
while (n-- && *t != '\0')
scripts/unifdef.c
1193
if (*s != *t)
scripts/unifdef.c
1194
return ((unsigned char)*s - (unsigned char)*t);
scripts/unifdef.c
1196
++s, ++t;
security/apparmor/apparmorfs.c
1032
struct multi_transaction *t;
security/apparmor/apparmorfs.c
1038
t = multi_transaction_new(file, ubuf, count);
security/apparmor/apparmorfs.c
1039
if (IS_ERR(t))
security/apparmor/apparmorfs.c
1040
return PTR_ERR(t);
security/apparmor/apparmorfs.c
1043
!memcmp(t->data, QUERY_CMD_PROFILE, QUERY_CMD_PROFILE_LEN)) {
security/apparmor/apparmorfs.c
1044
len = query_label(t->data, MULTI_TRANSACTION_LIMIT,
security/apparmor/apparmorfs.c
1045
t->data + QUERY_CMD_PROFILE_LEN,
security/apparmor/apparmorfs.c
1048
!memcmp(t->data, QUERY_CMD_LABEL, QUERY_CMD_LABEL_LEN)) {
security/apparmor/apparmorfs.c
1049
len = query_label(t->data, MULTI_TRANSACTION_LIMIT,
security/apparmor/apparmorfs.c
1050
t->data + QUERY_CMD_LABEL_LEN,
security/apparmor/apparmorfs.c
1053
!memcmp(t->data, QUERY_CMD_LABELALL,
security/apparmor/apparmorfs.c
1055
len = query_label(t->data, MULTI_TRANSACTION_LIMIT,
security/apparmor/apparmorfs.c
1056
t->data + QUERY_CMD_LABELALL_LEN,
security/apparmor/apparmorfs.c
1059
!memcmp(t->data, QUERY_CMD_DATA, QUERY_CMD_DATA_LEN)) {
security/apparmor/apparmorfs.c
1060
len = query_data(t->data, MULTI_TRANSACTION_LIMIT,
security/apparmor/apparmorfs.c
1061
t->data + QUERY_CMD_DATA_LEN,
security/apparmor/apparmorfs.c
1067
put_multi_transaction(t);
security/apparmor/apparmorfs.c
1071
multi_transaction_set(file, t, len);
security/apparmor/apparmorfs.c
112
char *t = target;
security/apparmor/apparmorfs.c
120
*(t)++ = '.';
security/apparmor/apparmorfs.c
122
*(t)++ = '_';
security/apparmor/apparmorfs.c
124
*(t)++ = *name;
security/apparmor/apparmorfs.c
127
*t = 0;
security/apparmor/apparmorfs.c
139
return t - target;
security/apparmor/apparmorfs.c
915
struct multi_transaction *t;
security/apparmor/apparmorfs.c
917
t = container_of(kref, struct multi_transaction, count);
security/apparmor/apparmorfs.c
918
free_page((unsigned long) t);
security/apparmor/apparmorfs.c
922
get_multi_transaction(struct multi_transaction *t)
security/apparmor/apparmorfs.c
924
if (t)
security/apparmor/apparmorfs.c
925
kref_get(&(t->count));
security/apparmor/apparmorfs.c
927
return t;
security/apparmor/apparmorfs.c
930
static void put_multi_transaction(struct multi_transaction *t)
security/apparmor/apparmorfs.c
932
if (t)
security/apparmor/apparmorfs.c
933
kref_put(&(t->count), multi_transaction_kref);
security/apparmor/apparmorfs.c
956
struct multi_transaction *t;
security/apparmor/apparmorfs.c
961
t = (struct multi_transaction *)get_zeroed_page(GFP_KERNEL);
security/apparmor/apparmorfs.c
962
if (!t)
security/apparmor/apparmorfs.c
964
kref_init(&t->count);
security/apparmor/apparmorfs.c
965
if (copy_from_user(t->data, buf, size)) {
security/apparmor/apparmorfs.c
966
put_multi_transaction(t);
security/apparmor/apparmorfs.c
970
return t;
security/apparmor/apparmorfs.c
976
struct multi_transaction *t;
security/apparmor/apparmorfs.c
980
t = get_multi_transaction(file->private_data);
security/apparmor/apparmorfs.c
983
if (!t)
security/apparmor/apparmorfs.c
986
ret = simple_read_from_buffer(buf, size, pos, t->data, t->size);
security/apparmor/apparmorfs.c
987
put_multi_transaction(t);
security/apparmor/include/lib.h
162
bool aa_resize_str_table(struct aa_str_table *t, int newsize, gfp_t gfp);
security/apparmor/lib.c
121
bool aa_resize_str_table(struct aa_str_table *t, int newsize, gfp_t gfp)
security/apparmor/lib.c
126
if (t->size == newsize)
security/apparmor/lib.c
131
for (i = 0; i < min(t->size, newsize); i++)
security/apparmor/lib.c
132
n[i] = t->table[i];
security/apparmor/lib.c
133
for (; i < t->size; i++)
security/apparmor/lib.c
134
kfree_sensitive(t->table[i].strs);
security/apparmor/lib.c
135
if (newsize > t->size)
security/apparmor/lib.c
136
memset(&n[t->size], 0, (newsize-t->size)*sizeof(*n));
security/apparmor/lib.c
137
kfree_sensitive(t->table);
security/apparmor/lib.c
138
t->table = n;
security/apparmor/lib.c
139
t->size = newsize;
security/apparmor/lib.c
148
void aa_destroy_str_table(struct aa_str_table *t)
security/apparmor/lib.c
152
if (t) {
security/apparmor/lib.c
153
if (!t->table)
security/apparmor/lib.c
156
for (i = 0; i < t->size; i++)
security/apparmor/lib.c
157
kfree_sensitive(t->table[i].strs);
security/apparmor/lib.c
158
kfree_sensitive(t->table);
security/apparmor/lib.c
159
t->table = NULL;
security/apparmor/lib.c
160
t->size = 0;
security/ipe/policy_parser.c
146
char *t, *ver = NULL;
security/ipe/policy_parser.c
150
while ((t = strsep(&line, IPE_POLICY_DELIM)) != NULL) {
security/ipe/policy_parser.c
153
if (*t == '\0')
security/ipe/policy_parser.c
160
token = match_token(t, header_tokens, args);
security/ipe/policy_parser.c
218
struct ipe_prop *p, *t;
security/ipe/policy_parser.c
223
list_for_each_entry_safe(p, t, &r->props, next) {
security/ipe/policy_parser.c
249
static enum ipe_op_type parse_operation(char *t)
security/ipe/policy_parser.c
253
return match_token(t, operation_tokens, args);
security/ipe/policy_parser.c
268
static enum ipe_action_type parse_action(char *t)
security/ipe/policy_parser.c
272
return match_token(t, action_tokens, args);
security/ipe/policy_parser.c
30
struct ipe_op_table *t = NULL;
security/ipe/policy_parser.c
300
static int parse_property(char *t, struct ipe_rule *r)
security/ipe/policy_parser.c
312
token = match_token(t, property_tokens, args);
security/ipe/policy_parser.c
371
char *t;
security/ipe/policy_parser.c
383
while (t = strsep(&line, IPE_POLICY_DELIM), line) {
security/ipe/policy_parser.c
384
if (*t == '\0')
security/ipe/policy_parser.c
386
if (first_token && token_default(t)) {
security/ipe/policy_parser.c
390
op = parse_operation(t);
security/ipe/policy_parser.c
396
rc = parse_property(t, r);
security/ipe/policy_parser.c
40
t = &p->rules[i];
security/ipe/policy_parser.c
405
action = parse_action(t);
security/ipe/policy_parser.c
42
t->default_action = IPE_ACTION_INVALID;
security/ipe/policy_parser.c
43
INIT_LIST_HEAD(&t->rules);
security/ipe/policy_parser.c
451
struct ipe_rule *pp, *t;
security/ipe/policy_parser.c
458
list_for_each_entry_safe(pp, t, &p->rules[i].rules, next) {
security/keys/request_key.c
39
struct task_struct *t = current;
security/keys/request_key.c
42
if (!(t->flags & PF_KTHREAD)) {
security/keys/request_key.c
43
key_put(t->cached_requested_key);
security/keys/request_key.c
44
t->cached_requested_key = key_get(key);
security/keys/request_key.c
45
set_tsk_thread_flag(t, TIF_NOTIFY_RESUME);
security/selinux/ss/policydb.c
1204
static void type_set_init(struct type_set *t)
security/selinux/ss/policydb.c
1206
ebitmap_init(&t->types);
security/selinux/ss/policydb.c
1207
ebitmap_init(&t->negset);
security/selinux/ss/policydb.c
1210
static int type_set_read(struct type_set *t, struct policy_file *fp)
security/selinux/ss/policydb.c
1215
if (ebitmap_read(&t->types, fp))
security/selinux/ss/policydb.c
1217
if (ebitmap_read(&t->negset, fp))
security/selinux/ss/policydb.c
1223
t->flags = le32_to_cpu(buf[0]);
security/selinux/ss/policydb.c
3021
static int type_set_write(struct type_set *t, struct policy_file *fp)
security/selinux/ss/policydb.c
3026
if (ebitmap_write(&t->types, fp))
security/selinux/ss/policydb.c
3028
if (ebitmap_write(&t->negset, fp))
security/selinux/ss/policydb.c
3031
buf[0] = cpu_to_le32(t->flags);
security/selinux/ss/services.c
1694
char *s = NULL, *t = NULL, *n = NULL;
security/selinux/ss/services.c
1700
if (sidtab_entry_to_string(policydb, sidtab, tentry, &t, &tlen))
security/selinux/ss/services.c
1712
s, t, sym_name(policydb, SYM_CLASSES, tclass-1));
security/selinux/ss/services.c
1716
kfree(t);
security/selinux/ss/services.c
733
char *o = NULL, *n = NULL, *t = NULL;
security/selinux/ss/services.c
740
if (sidtab_entry_to_string(p, sidtab, tentry, &t, &tlen))
security/selinux/ss/services.c
745
o, n, t, sym_name(p, SYM_CLASSES, tclass-1));
security/selinux/ss/services.c
749
kfree(t);
security/smack/smack.h
421
const struct task_struct *t)
security/smack/smack.h
428
cred = __task_cred(t);
security/smack/smack.h
488
struct task_struct *t)
security/smack/smack.h
490
a->a.u.tsk = t;
security/smack/smack.h
520
struct task_struct *t)
sound/core/hrtimer.c
104
static int snd_hrtimer_stop(struct snd_timer *t)
sound/core/hrtimer.c
106
struct snd_hrtimer *stime = t->private_data;
sound/core/hrtimer.c
34
struct snd_timer *t = stime->timer;
sound/core/hrtimer.c
39
scoped_guard(spinlock, &t->lock) {
sound/core/hrtimer.c
40
if (!t->running)
sound/core/hrtimer.c
43
ticks = t->sticks;
sound/core/hrtimer.c
53
guard(spinlock)(&t->lock);
sound/core/hrtimer.c
54
if (t->running) {
sound/core/hrtimer.c
55
hrtimer_add_expires_ns(hrt, t->sticks * resolution);
sound/core/hrtimer.c
63
static int snd_hrtimer_open(struct snd_timer *t)
sound/core/hrtimer.c
70
stime->timer = t;
sound/core/hrtimer.c
72
t->private_data = stime;
sound/core/hrtimer.c
76
static int snd_hrtimer_close(struct snd_timer *t)
sound/core/hrtimer.c
78
struct snd_hrtimer *stime = t->private_data;
sound/core/hrtimer.c
81
scoped_guard(spinlock_irq, &t->lock) {
sound/core/hrtimer.c
82
t->running = 0; /* just to be sure */
sound/core/hrtimer.c
88
t->private_data = NULL;
sound/core/hrtimer.c
93
static int snd_hrtimer_start(struct snd_timer *t)
sound/core/hrtimer.c
95
struct snd_hrtimer *stime = t->private_data;
sound/core/hrtimer.c
99
hrtimer_start(&stime->hrt, ns_to_ktime(t->sticks * resolution),
sound/core/oss/mulaw.c
108
int t;
sound/core/oss/mulaw.c
117
t = ((u_val & QUANT_MASK) << 3) + BIAS;
sound/core/oss/mulaw.c
118
t <<= ((unsigned)u_val & SEG_MASK) >> SEG_SHIFT;
sound/core/oss/mulaw.c
120
return ((u_val & SIGN_BIT) ? (BIAS - t) : (t - BIAS));
sound/core/oss/pcm_oss.c
108
struct snd_interval t;
sound/core/oss/pcm_oss.c
109
t.empty = 0;
sound/core/oss/pcm_oss.c
110
t.min = t.max = val;
sound/core/oss/pcm_oss.c
111
t.openmin = t.openmax = 0;
sound/core/oss/pcm_oss.c
112
t.integer = 1;
sound/core/oss/pcm_oss.c
113
return snd_interval_refine(i, &t);
sound/core/oss/pcm_oss.c
465
struct snd_interval t;
sound/core/oss/pcm_oss.c
466
t.openmin = 1;
sound/core/oss/pcm_oss.c
467
t.openmax = 1;
sound/core/oss/pcm_oss.c
468
t.empty = 0;
sound/core/oss/pcm_oss.c
469
t.integer = 0;
sound/core/oss/pcm_oss.c
471
t.min = val - 1;
sound/core/oss/pcm_oss.c
472
t.max = val;
sound/core/oss/pcm_oss.c
474
t.min = val;
sound/core/oss/pcm_oss.c
475
t.max = val+1;
sound/core/oss/pcm_oss.c
477
changed = snd_interval_refine(i, &t);
sound/core/pcm_drm_eld.c
128
struct snd_interval t = { .min = 1, .max = 2, .integer = 1, };
sound/core/pcm_drm_eld.c
144
t.max = max(t.max, sad_max_channels(sad));
sound/core/pcm_drm_eld.c
147
return snd_interval_refine(c, &t);
sound/core/pcm_lib.c
1020
t.max = div_up(best_num, best_den);
sound/core/pcm_lib.c
1021
t.openmax = !!(best_num % best_den);
sound/core/pcm_lib.c
1022
t.integer = 0;
sound/core/pcm_lib.c
1023
err = snd_interval_refine(i, &t);
sound/core/pcm_lib.c
1288
struct snd_interval t;
sound/core/pcm_lib.c
1289
t.min = min;
sound/core/pcm_lib.c
1290
t.max = max;
sound/core/pcm_lib.c
1291
t.openmin = t.openmax = 0;
sound/core/pcm_lib.c
1292
t.integer = 0;
sound/core/pcm_lib.c
1293
return snd_interval_refine(constrs_interval(constrs, var), &t);
sound/core/pcm_lib.c
1973
long t = runtime->buffer_size * 1100 / runtime->rate;
sound/core/pcm_lib.c
1974
wait_time = max(t, wait_time);
sound/core/pcm_lib.c
838
struct snd_interval t;
sound/core/pcm_lib.c
876
t.min = div_down(best_num, best_den);
sound/core/pcm_lib.c
877
t.openmin = !!(best_num % best_den);
sound/core/pcm_lib.c
917
t.max = div_up(best_num, best_den);
sound/core/pcm_lib.c
918
t.openmax = !!(best_num % best_den);
sound/core/pcm_lib.c
919
t.integer = 0;
sound/core/pcm_lib.c
920
err = snd_interval_refine(i, &t);
sound/core/pcm_lib.c
956
struct snd_interval t;
sound/core/pcm_lib.c
988
t.min = div_down(best_num, best_den);
sound/core/pcm_lib.c
989
t.openmin = !!(best_num % best_den);
sound/core/pcm_native.c
2188
long t = drain_bufsz * 1100 / drain_rate;
sound/core/pcm_native.c
2189
tout = max(t, tout);
sound/core/pcm_native.c
2373
struct snd_interval t;
sound/core/pcm_native.c
2375
hw_param_interval_c(params, rule->deps[1]), &t);
sound/core/pcm_native.c
2376
return snd_interval_refine(hw_param_interval(params, rule->var), &t);
sound/core/pcm_native.c
2382
struct snd_interval t;
sound/core/pcm_native.c
2384
hw_param_interval_c(params, rule->deps[1]), &t);
sound/core/pcm_native.c
2385
return snd_interval_refine(hw_param_interval(params, rule->var), &t);
sound/core/pcm_native.c
2391
struct snd_interval t;
sound/core/pcm_native.c
2394
(unsigned long) rule->private, &t);
sound/core/pcm_native.c
2395
return snd_interval_refine(hw_param_interval(params, rule->var), &t);
sound/core/pcm_native.c
2401
struct snd_interval t;
sound/core/pcm_native.c
2404
hw_param_interval_c(params, rule->deps[1]), &t);
sound/core/pcm_native.c
2405
return snd_interval_refine(hw_param_interval(params, rule->var), &t);
sound/core/pcm_native.c
2433
struct snd_interval t;
sound/core/pcm_native.c
2436
t.min = UINT_MAX;
sound/core/pcm_native.c
2437
t.max = 0;
sound/core/pcm_native.c
2438
t.openmin = 0;
sound/core/pcm_native.c
2439
t.openmax = 0;
sound/core/pcm_native.c
2447
if (t.min > (unsigned)bits)
sound/core/pcm_native.c
2448
t.min = bits;
sound/core/pcm_native.c
2449
if (t.max < (unsigned)bits)
sound/core/pcm_native.c
2450
t.max = bits;
sound/core/pcm_native.c
2452
t.integer = 1;
sound/core/pcm_native.c
2453
return snd_interval_refine(hw_param_interval(params, rule->var), &t);
sound/core/pcm_native.c
2486
struct snd_interval t;
sound/core/pcm_native.c
2488
t.min = 0;
sound/core/pcm_native.c
2489
t.max = substream->buffer_bytes_max;
sound/core/pcm_native.c
2490
t.openmin = 0;
sound/core/pcm_native.c
2491
t.openmax = 0;
sound/core/pcm_native.c
2492
t.integer = 1;
sound/core/pcm_native.c
2493
return snd_interval_refine(hw_param_interval(params, rule->var), &t);
sound/core/seq/oss/seq_oss_event.c
225
switch (q->t.cmd) {
sound/core/seq/oss/seq_oss_event.c
233
tmp.echo = (q->t.time << 8) | SEQ_ECHO;
sound/core/seq/oss/seq_oss_event.c
249
return snd_seq_oss_timer_tempo(dp->timer, q->t.time);
sound/core/seq/oss/seq_oss_event.c
435
snd_seq_oss_writeq_wakeup(dp->writeq, rec->t.time);
sound/core/seq/oss/seq_oss_event.h
84
struct evrec_timer t;
sound/core/seq/oss/seq_oss_readq.c
221
rec.t.code = EV_TIMING;
sound/core/seq/oss/seq_oss_readq.c
222
rec.t.cmd = TMR_WAIT_ABS;
sound/core/seq/oss/seq_oss_readq.c
223
rec.t.time = curt;
sound/core/seq/oss/seq_oss_timer.c
75
abstime_t parm = ev->t.time;
sound/core/seq/oss/seq_oss_timer.c
77
if (ev->t.code == EV_TIMING) {
sound/core/seq/oss/seq_oss_timer.c
78
switch (ev->t.cmd) {
sound/core/seq/oss/seq_oss_writeq.c
104
rec->t.code = SEQ_SYNCTIMER;
sound/core/seq/oss/seq_oss_writeq.c
105
rec->t.time = time;
sound/core/seq/seq_timer.c
245
struct snd_timer_instance *t;
sound/core/seq/seq_timer.c
260
t = snd_timer_instance_new(str);
sound/core/seq/seq_timer.c
261
if (!t)
sound/core/seq/seq_timer.c
263
t->callback = snd_seq_timer_interrupt;
sound/core/seq/seq_timer.c
264
t->callback_data = q;
sound/core/seq/seq_timer.c
265
t->flags |= SNDRV_TIMER_IFLG_AUTO;
sound/core/seq/seq_timer.c
266
err = snd_timer_open(t, &tmr->alsa_id, q->queue);
sound/core/seq/seq_timer.c
276
err = snd_timer_open(t, &tid, q->queue);
sound/core/seq/seq_timer.c
281
snd_timer_instance_free(t);
sound/core/seq/seq_timer.c
288
tmr->timeri = t;
sound/core/seq/seq_timer.c
291
snd_timer_close(t);
sound/core/seq/seq_timer.c
292
snd_timer_instance_free(t);
sound/core/seq/seq_timer.c
301
struct snd_timer_instance *t;
sound/core/seq/seq_timer.c
307
t = tmr->timeri;
sound/core/seq/seq_timer.c
310
if (t) {
sound/core/seq/seq_timer.c
311
snd_timer_close(t);
sound/core/seq/seq_timer.c
312
snd_timer_instance_free(t);
sound/core/seq/seq_timer.c
336
struct snd_timer *t;
sound/core/seq/seq_timer.c
339
t = tmr->timeri->timer;
sound/core/seq/seq_timer.c
340
if (!t)
sound/core/seq/seq_timer.c
352
if (!(t->hw.flags & SNDRV_TIMER_HW_SLAVE)) {
sound/core/seq/seq_timer.c
63
struct snd_seq_timer *t = *tmr;
sound/core/seq/seq_timer.c
66
if (t == NULL) {
sound/core/seq/seq_timer.c
70
t->running = 0;
sound/core/seq/seq_timer.c
73
snd_seq_timer_stop(t);
sound/core/seq/seq_timer.c
74
snd_seq_timer_reset(t);
sound/core/seq/seq_timer.c
76
kfree(t);
sound/core/timer.c
1119
static void snd_timer_s_function(struct timer_list *t)
sound/core/timer.c
1121
struct snd_timer_system_private *priv = timer_container_of(priv, t,
sound/core/timer.c
1618
struct snd_timer *t;
sound/core/timer.c
1630
t = snd_timer_find(&tid);
sound/core/timer.c
1631
if (!t)
sound/core/timer.c
1633
ginfo->card = t->card ? t->card->number : -1;
sound/core/timer.c
1634
if (t->hw.flags & SNDRV_TIMER_HW_SLAVE)
sound/core/timer.c
1636
strscpy(ginfo->id, t->id, sizeof(ginfo->id));
sound/core/timer.c
1637
strscpy(ginfo->name, t->name, sizeof(ginfo->name));
sound/core/timer.c
1638
scoped_guard(spinlock_irq, &t->lock)
sound/core/timer.c
1639
ginfo->resolution = snd_timer_hw_resolution(t);
sound/core/timer.c
1640
if (t->hw.resolution_min > 0) {
sound/core/timer.c
1641
ginfo->resolution_min = t->hw.resolution_min;
sound/core/timer.c
1642
ginfo->resolution_max = t->hw.resolution_max;
sound/core/timer.c
1644
list_for_each(p, &t->open_list_head) {
sound/core/timer.c
1655
struct snd_timer *t;
sound/core/timer.c
1658
t = snd_timer_find(&gparams->tid);
sound/core/timer.c
1659
if (!t)
sound/core/timer.c
1661
if (!list_empty(&t->open_list_head))
sound/core/timer.c
1663
if (!t->hw.set_period)
sound/core/timer.c
1665
return t->hw.set_period(t, gparams->period_num, gparams->period_den);
sound/core/timer.c
1683
struct snd_timer *t;
sound/core/timer.c
1691
t = snd_timer_find(&tid);
sound/core/timer.c
1692
if (t != NULL) {
sound/core/timer.c
1693
guard(spinlock_irq)(&t->lock);
sound/core/timer.c
1694
gstatus.resolution = snd_timer_hw_resolution(t);
sound/core/timer.c
1695
if (t->hw.precise_resolution) {
sound/core/timer.c
1696
t->hw.precise_resolution(t, &gstatus.resolution_num,
sound/core/timer.c
1759
struct snd_timer *t;
sound/core/timer.c
1764
t = tu->timeri->timer;
sound/core/timer.c
1765
if (!t)
sound/core/timer.c
1772
info->card = t->card ? t->card->number : -1;
sound/core/timer.c
1773
if (t->hw.flags & SNDRV_TIMER_HW_SLAVE)
sound/core/timer.c
1775
strscpy(info->id, t->id, sizeof(info->id));
sound/core/timer.c
1776
strscpy(info->name, t->name, sizeof(info->name));
sound/core/timer.c
1777
scoped_guard(spinlock_irq, &t->lock)
sound/core/timer.c
1778
info->resolution = snd_timer_hw_resolution(t);
sound/core/timer.c
1789
struct snd_timer *t;
sound/core/timer.c
1795
t = tu->timeri->timer;
sound/core/timer.c
1796
if (!t)
sound/core/timer.c
1800
if (!(t->hw.flags & SNDRV_TIMER_HW_SLAVE)) {
sound/core/timer.c
1839
scoped_guard(spinlock_irq, &t->lock) {
sound/core/timer.c
2091
static int snd_utimer_start(struct snd_timer *t)
sound/core/timer.c
2096
static int snd_utimer_stop(struct snd_timer *t)
sound/core/timer.c
2101
static int snd_utimer_open(struct snd_timer *t)
sound/core/timer.c
2106
static int snd_utimer_close(struct snd_timer *t)
sound/core/timer.c
345
struct snd_timer_instance *t =
sound/core/timer.c
348
if (t->flags & SNDRV_TIMER_IFLG_EXCLUSIVE) {
sound/core/timer_compat.c
52
struct snd_timer *t;
sound/core/timer_compat.c
57
t = tu->timeri->timer;
sound/core/timer_compat.c
58
if (!t)
sound/core/timer_compat.c
61
info.card = t->card ? t->card->number : -1;
sound/core/timer_compat.c
62
if (t->hw.flags & SNDRV_TIMER_HW_SLAVE)
sound/core/timer_compat.c
64
strscpy(info.id, t->id, sizeof(info.id));
sound/core/timer_compat.c
65
strscpy(info.name, t->name, sizeof(info.name));
sound/core/timer_compat.c
66
info.resolution = t->hw.resolution;
sound/drivers/aloop.c
1003
struct snd_interval t;
sound/drivers/aloop.c
1006
t.min = cable->hw.rate_min;
sound/drivers/aloop.c
1007
t.max = cable->hw.rate_max;
sound/drivers/aloop.c
1009
t.openmin = t.openmax = 0;
sound/drivers/aloop.c
1010
t.integer = 0;
sound/drivers/aloop.c
1011
return snd_interval_refine(hw_param_interval(params, rule->var), &t);
sound/drivers/aloop.c
1019
struct snd_interval t;
sound/drivers/aloop.c
1022
t.min = cable->hw.channels_min;
sound/drivers/aloop.c
1023
t.max = cable->hw.channels_max;
sound/drivers/aloop.c
1025
t.openmin = t.openmax = 0;
sound/drivers/aloop.c
1026
t.integer = 0;
sound/drivers/aloop.c
1027
return snd_interval_refine(hw_param_interval(params, rule->var), &t);
sound/drivers/aloop.c
1035
struct snd_interval t;
sound/drivers/aloop.c
1038
t.min = cable->hw.period_bytes_min;
sound/drivers/aloop.c
1039
t.max = cable->hw.period_bytes_max;
sound/drivers/aloop.c
1041
t.openmin = 0;
sound/drivers/aloop.c
1042
t.openmax = 0;
sound/drivers/aloop.c
1043
t.integer = 0;
sound/drivers/aloop.c
1044
return snd_interval_refine(hw_param_interval(params, rule->var), &t);
sound/drivers/aloop.c
710
static void loopback_jiffies_timer_function(struct timer_list *t)
sound/drivers/aloop.c
712
struct loopback_pcm *dpcm = timer_container_of(dpcm, t, timer);
sound/drivers/dummy.c
303
static void dummy_systimer_callback(struct timer_list *t)
sound/drivers/dummy.c
305
struct dummy_systimer_pcm *dpcm = timer_container_of(dpcm, t, timer);
sound/drivers/mpu401/mpu401_uart.c
152
static void snd_mpu401_uart_timer(struct timer_list *t)
sound/drivers/mpu401/mpu401_uart.c
154
struct snd_mpu401 *mpu = timer_container_of(mpu, t, timer);
sound/drivers/mtpav.c
380
static void snd_mtpav_output_timer(struct timer_list *t)
sound/drivers/mtpav.c
382
struct mtpav *chip = timer_container_of(chip, t, timer);
sound/drivers/opl3/opl3_midi.c
233
void snd_opl3_timer_func(struct timer_list *t)
sound/drivers/opl3/opl3_midi.c
236
struct snd_opl3 *opl3 = timer_container_of(opl3, t, tlist);
sound/drivers/opl3/opl3_voice.h
27
void snd_opl3_timer_func(struct timer_list *t);
sound/drivers/serial-u16550.c
294
static void snd_uart16550_buffer_timer(struct timer_list *t)
sound/drivers/serial-u16550.c
298
uart = timer_container_of(uart, t, buffer_timer);
sound/firewire/amdtp-stream.c
166
struct snd_interval t = {0};
sound/firewire/amdtp-stream.c
178
t.min = roundup(s->min, step);
sound/firewire/amdtp-stream.c
179
t.max = rounddown(s->max, step);
sound/firewire/amdtp-stream.c
180
t.integer = 1;
sound/firewire/amdtp-stream.c
182
return snd_interval_refine(s, &t);
sound/firewire/bebob/bebob_pcm.c
18
struct snd_interval t = {
sound/firewire/bebob/bebob_pcm.c
31
t.min = min(t.min, snd_bebob_rate_table[i]);
sound/firewire/bebob/bebob_pcm.c
32
t.max = max(t.max, snd_bebob_rate_table[i]);
sound/firewire/bebob/bebob_pcm.c
35
return snd_interval_refine(r, &t);
sound/firewire/bebob/bebob_pcm.c
46
struct snd_interval t = {
sound/firewire/bebob/bebob_pcm.c
60
t.min = min(t.min, formations[i].pcm);
sound/firewire/bebob/bebob_pcm.c
61
t.max = max(t.max, formations[i].pcm);
sound/firewire/bebob/bebob_pcm.c
64
return snd_interval_refine(c, &t);
sound/firewire/digi00x/digi00x-pcm.c
17
struct snd_interval t = {
sound/firewire/digi00x/digi00x-pcm.c
27
t.min = min(t.min, snd_dg00x_stream_rates[i]);
sound/firewire/digi00x/digi00x-pcm.c
28
t.max = max(t.max, snd_dg00x_stream_rates[i]);
sound/firewire/digi00x/digi00x-pcm.c
31
return snd_interval_refine(r, &t);
sound/firewire/digi00x/digi00x-pcm.c
41
struct snd_interval t = {
sound/firewire/digi00x/digi00x-pcm.c
50
t.min = min(t.min, snd_dg00x_stream_pcm_channels[i]);
sound/firewire/digi00x/digi00x-pcm.c
51
t.max = max(t.max, snd_dg00x_stream_pcm_channels[i]);
sound/firewire/digi00x/digi00x-pcm.c
54
return snd_interval_refine(c, &t);
sound/firewire/fcp.c
234
struct fcp_transaction t;
sound/firewire/fcp.c
237
t.unit = unit;
sound/firewire/fcp.c
238
t.response_buffer = response;
sound/firewire/fcp.c
239
t.response_size = response_size;
sound/firewire/fcp.c
240
t.response_match_bytes = response_match_bytes;
sound/firewire/fcp.c
241
t.state = STATE_PENDING;
sound/firewire/fcp.c
242
init_waitqueue_head(&t.wait);
sound/firewire/fcp.c
243
t.deferrable = (*(const u8 *)command == 0x00 || *(const u8 *)command == 0x03);
sound/firewire/fcp.c
246
list_add_tail(&t.list, &transactions);
sound/firewire/fcp.c
252
ret = snd_fw_transaction(t.unit, tcode,
sound/firewire/fcp.c
258
wait_event_timeout(t.wait, t.state != STATE_PENDING,
sound/firewire/fcp.c
261
if (t.state == STATE_DEFERRED) {
sound/firewire/fcp.c
269
t.state = STATE_PENDING;
sound/firewire/fcp.c
271
} else if (t.state == STATE_COMPLETE) {
sound/firewire/fcp.c
272
ret = t.response_size;
sound/firewire/fcp.c
274
} else if (t.state == STATE_BUS_RESET) {
sound/firewire/fcp.c
277
dev_err(&t.unit->device, "FCP command timed out\n");
sound/firewire/fcp.c
284
list_del(&t.list);
sound/firewire/fcp.c
301
struct fcp_transaction *t;
sound/firewire/fcp.c
304
list_for_each_entry(t, &transactions, list) {
sound/firewire/fcp.c
305
if (t->unit == unit &&
sound/firewire/fcp.c
306
(t->state == STATE_PENDING ||
sound/firewire/fcp.c
307
t->state == STATE_DEFERRED)) {
sound/firewire/fcp.c
308
t->state = STATE_BUS_RESET;
sound/firewire/fcp.c
309
wake_up(&t->wait);
sound/firewire/fcp.c
342
struct fcp_transaction *t;
sound/firewire/fcp.c
348
list_for_each_entry(t, &transactions, list) {
sound/firewire/fcp.c
349
struct fw_device *device = fw_parent_device(t->unit);
sound/firewire/fcp.c
357
if (t->state == STATE_PENDING &&
sound/firewire/fcp.c
358
is_matching_response(t, data, length)) {
sound/firewire/fcp.c
359
if (t->deferrable && *(const u8 *)data == 0x0f) {
sound/firewire/fcp.c
360
t->state = STATE_DEFERRED;
sound/firewire/fcp.c
362
t->state = STATE_COMPLETE;
sound/firewire/fcp.c
363
t->response_size = min_t(unsigned int, length,
sound/firewire/fcp.c
364
t->response_size);
sound/firewire/fcp.c
365
memcpy(t->response_buffer, data,
sound/firewire/fcp.c
366
t->response_size);
sound/firewire/fcp.c
368
wake_up(&t->wait);
sound/firewire/fireface/ff-pcm.c
18
struct snd_interval t = {
sound/firewire/fireface/ff-pcm.c
34
t.min = min(t.min, amdtp_rate_table[i]);
sound/firewire/fireface/ff-pcm.c
35
t.max = max(t.max, amdtp_rate_table[i]);
sound/firewire/fireface/ff-pcm.c
38
return snd_interval_refine(r, &t);
sound/firewire/fireface/ff-pcm.c
49
struct snd_interval t = {
sound/firewire/fireface/ff-pcm.c
65
t.min = min(t.min, pcm_channels[mode]);
sound/firewire/fireface/ff-pcm.c
66
t.max = max(t.max, pcm_channels[mode]);
sound/firewire/fireface/ff-pcm.c
69
return snd_interval_refine(c, &t);
sound/firewire/fireworks/fireworks_hwdep.c
26
struct snd_efw_transaction *t;
sound/firewire/fireworks/fireworks_hwdep.c
52
t = (struct snd_efw_transaction *)(pull_ptr);
sound/firewire/fireworks/fireworks_hwdep.c
53
length = be32_to_cpu(t->length) * sizeof(__be32);
sound/firewire/fireworks/fireworks_pcm.c
106
t.min = min(t.min, pcm_channels[mode]);
sound/firewire/fireworks/fireworks_pcm.c
107
t.max = max(t.max, pcm_channels[mode]);
sound/firewire/fireworks/fireworks_pcm.c
110
return snd_interval_refine(c, &t);
sound/firewire/fireworks/fireworks_pcm.c
71
struct snd_interval t = {
sound/firewire/fireworks/fireworks_pcm.c
81
t.min = min(t.min, freq_table[i]);
sound/firewire/fireworks/fireworks_pcm.c
82
t.max = max(t.max, freq_table[i]);
sound/firewire/fireworks/fireworks_pcm.c
85
return snd_interval_refine(r, &t);
sound/firewire/fireworks/fireworks_pcm.c
96
struct snd_interval t = {
sound/firewire/fireworks/fireworks_transaction.c
101
} else if (t.state == STATE_BUS_RESET) {
sound/firewire/fireworks/fireworks_transaction.c
104
dev_err(&t.unit->device, "EFW transaction timed out\n");
sound/firewire/fireworks/fireworks_transaction.c
111
list_del(&t.list);
sound/firewire/fireworks/fireworks_transaction.c
121
struct snd_efw_transaction *t;
sound/firewire/fireworks/fireworks_transaction.c
123
t = (struct snd_efw_transaction *)data;
sound/firewire/fireworks/fireworks_transaction.c
124
length = min_t(size_t, be32_to_cpu(t->length) * sizeof(u32), length);
sound/firewire/fireworks/fireworks_transaction.c
197
struct transaction_queue *t;
sound/firewire/fireworks/fireworks_transaction.c
200
list_for_each_entry(t, &transaction_queues, list) {
sound/firewire/fireworks/fireworks_transaction.c
201
device = fw_parent_device(t->unit);
sound/firewire/fireworks/fireworks_transaction.c
209
if ((t->state == STATE_PENDING) && (t->seqnum == seqnum)) {
sound/firewire/fireworks/fireworks_transaction.c
210
t->state = STATE_COMPLETE;
sound/firewire/fireworks/fireworks_transaction.c
211
t->size = min_t(unsigned int, length, t->size);
sound/firewire/fireworks/fireworks_transaction.c
212
memcpy(t->buf, data, t->size);
sound/firewire/fireworks/fireworks_transaction.c
213
wake_up(&t->wait);
sound/firewire/fireworks/fireworks_transaction.c
281
struct transaction_queue *t;
sound/firewire/fireworks/fireworks_transaction.c
284
list_for_each_entry(t, &transaction_queues, list) {
sound/firewire/fireworks/fireworks_transaction.c
285
if ((t->unit == unit) &&
sound/firewire/fireworks/fireworks_transaction.c
286
(t->state == STATE_PENDING)) {
sound/firewire/fireworks/fireworks_transaction.c
287
t->state = STATE_BUS_RESET;
sound/firewire/fireworks/fireworks_transaction.c
288
wake_up(&t->wait);
sound/firewire/fireworks/fireworks_transaction.c
74
struct transaction_queue t;
sound/firewire/fireworks/fireworks_transaction.c
78
t.unit = unit;
sound/firewire/fireworks/fireworks_transaction.c
79
t.buf = resp;
sound/firewire/fireworks/fireworks_transaction.c
80
t.size = resp_size;
sound/firewire/fireworks/fireworks_transaction.c
81
t.seqnum = be32_to_cpu(((struct snd_efw_transaction *)cmd)->seqnum) + 1;
sound/firewire/fireworks/fireworks_transaction.c
82
t.state = STATE_PENDING;
sound/firewire/fireworks/fireworks_transaction.c
83
init_waitqueue_head(&t.wait);
sound/firewire/fireworks/fireworks_transaction.c
86
list_add_tail(&t.list, &transaction_queues);
sound/firewire/fireworks/fireworks_transaction.c
91
ret = snd_efw_transaction_cmd(t.unit, (void *)cmd, cmd_size);
sound/firewire/fireworks/fireworks_transaction.c
95
wait_event_timeout(t.wait, t.state != STATE_PENDING,
sound/firewire/fireworks/fireworks_transaction.c
98
if (t.state == STATE_COMPLETE) {
sound/firewire/fireworks/fireworks_transaction.c
99
ret = t.size;
sound/firewire/oxfw/oxfw-pcm.c
18
struct snd_interval t = {
sound/firewire/oxfw/oxfw-pcm.c
34
t.min = min(t.min, formation.rate);
sound/firewire/oxfw/oxfw-pcm.c
35
t.max = max(t.max, formation.rate);
sound/firewire/oxfw/oxfw-pcm.c
38
return snd_interval_refine(r, &t);
sound/hda/controllers/intel.c
473
u32 val, t;
sound/hda/controllers/intel.c
479
t = preferred_bits[i];
sound/hda/controllers/intel.c
480
if (val & (1 << t))
sound/hda/controllers/intel.c
481
return t;
sound/hda/core/hdmi_chmap.c
419
struct channel_map_table *t = map_tables;
sound/hda/core/hdmi_chmap.c
421
for (; t->map; t++) {
sound/hda/core/hdmi_chmap.c
422
if (t->map == c)
sound/hda/core/hdmi_chmap.c
423
return t->spk_mask;
sound/hda/core/hdmi_chmap.c
454
struct channel_map_table *t = map_tables;
sound/hda/core/hdmi_chmap.c
456
for (; t->map; t++) {
sound/hda/core/hdmi_chmap.c
457
if (t->spk_mask == spk)
sound/hda/core/hdmi_chmap.c
458
return t->map;
sound/i2c/other/ak4117.c
23
static void snd_ak4117_timer(struct timer_list *t);
sound/i2c/other/ak4117.c
503
static void snd_ak4117_timer(struct timer_list *t)
sound/i2c/other/ak4117.c
505
struct ak4117 *chip = timer_container_of(chip, t, timer);
sound/isa/sb/emu8000_pcm.c
183
static void emu8k_pcm_timer_func(struct timer_list *t)
sound/isa/sb/emu8000_pcm.c
185
struct snd_emu8k_pcm *rec = timer_container_of(rec, t, timer);
sound/isa/sb/sb8_main.c
84
struct snd_interval t = { .min = 1, .max = 1 };
sound/isa/sb/sb8_main.c
85
return snd_interval_refine(hw_param_interval(params, SNDRV_PCM_HW_PARAM_CHANNELS), &t);
sound/isa/sb/sb8_midi.c
178
static void snd_sb8dsp_midi_output_timer(struct timer_list *t)
sound/isa/sb/sb8_midi.c
180
struct snd_sb *chip = timer_container_of(chip, t, midi_timer);
sound/isa/wavefront/wavefront_midi.c
334
static void snd_wavefront_midi_output_timer(struct timer_list *t)
sound/isa/wavefront/wavefront_midi.c
336
snd_wavefront_midi_t *midi = timer_container_of(midi, t, timer);
sound/pci/asihpi/asihpi.c
710
static void snd_card_asihpi_timer_function(struct timer_list *t)
sound/pci/asihpi/asihpi.c
712
struct snd_card_asihpi_pcm *dpcm = timer_container_of(dpcm, t, timer);
sound/pci/asihpi/hpi6205.c
2054
int t = timeout_us / 4;
sound/pci/asihpi/hpi6205.c
2057
while ((interface->dsp_ack != state) && --t) {
sound/pci/asihpi/hpi6205.c
2063
return t * 4;
sound/pci/asihpi/hpi_internal.h
1019
struct hpi_clock_msg t; /* dsp time */
sound/pci/asihpi/hpi_internal.h
1080
struct hpi_clock_res t; /* dsp time */
sound/pci/asihpi/hpi_internal.h
1188
struct hpi_clock_msg t;
sound/pci/asihpi/hpi_internal.h
1207
struct hpi_clock_res t;
sound/pci/asihpi/hpi_internal.h
982
struct hpi_profile_res_time t;
sound/pci/cs46xx/dsp_spos.c
1931
struct dsp_task_descriptor *t = &ins->tasks[i];
sound/pci/cs46xx/dsp_spos.c
1932
_dsp_create_task_tree(chip, t->data, t->address, t->size);
sound/pci/ctxfi/cttimer.c
63
static void ct_systimer_callback(struct timer_list *t)
sound/pci/ctxfi/cttimer.c
65
struct ct_timer_instance *ti = timer_container_of(ti, t, timer);
sound/pci/echoaudio/midi.c
185
static void snd_echo_midi_output_write(struct timer_list *t)
sound/pci/echoaudio/midi.c
187
struct echoaudio *chip = timer_container_of(chip, t, timer);
sound/pci/ens1370.c
498
unsigned int t, r = 0;
sound/pci/ens1370.c
500
for (t = 0; t < POLL_COUNT; t++) {
sound/pci/ens1370.c
596
unsigned int t, x, flag;
sound/pci/ens1370.c
600
for (t = 0; t < POLL_COUNT; t++) {
sound/pci/ens1370.c
609
for (t = 0; t < POLL_COUNT; t++) {
sound/pci/ens1370.c
615
for (t = 0; t < POLL_COUNT; t++) {
sound/pci/ens1370.c
636
unsigned int t, x, flag, fail = 0;
sound/pci/ens1370.c
641
for (t = 0; t < POLL_COUNT; t++) {
sound/pci/ens1370.c
650
for (t = 0; t < POLL_COUNT; t++) {
sound/pci/ens1370.c
656
for (t = 0; t < POLL_COUNT; t++) {
sound/pci/ens1370.c
667
for (t = 0; t < POLL_COUNT; t++) {
sound/pci/ens1370.c
672
for (t = 0; t < POLL_COUNT; t++) {
sound/pci/ens1370.c
676
for (t = 0; t < 100; t++)
sound/pci/es1968.c
1640
unsigned int pa, offset, t;
sound/pci/es1968.c
1713
t = ktime_to_us(diff);
sound/pci/es1968.c
1714
if (t == 0) {
sound/pci/es1968.c
1718
offset = (offset / t) * 1000 + ((offset % t) * 1000) / t;
sound/pci/ice1712/ice1712.c
2200
long t = 0x10000;
sound/pci/ice1712/ice1712.c
2204
while (t-- > 0 && (inb(ICEREG(ice, I2C_CTRL)) & ICE1712_I2C_BUSY)) ;
sound/pci/ice1712/ice1724.c
2199
int t = 0x10000;
sound/pci/ice1712/ice1724.c
2200
while ((inb(ICEREG1724(ice, I2C_CTRL)) & VT1724_I2C_BUSY) && t--)
sound/pci/ice1712/ice1724.c
2202
if (t == -1)
sound/pci/intel8x0.c
2632
unsigned long pos, pos1, t;
sound/pci/intel8x0.c
2722
t = ktime_us_delta(stop_time, start_time);
sound/pci/intel8x0.c
2724
"%s: measured %lu usecs (%lu samples)\n", __func__, t, pos);
sound/pci/intel8x0.c
2725
if (t == 0) {
sound/pci/intel8x0.c
2730
pos = (pos / t) * 1000 + ((pos % t) * 1000) / t;
sound/pci/maestro3.c
2255
u8 t; /* makes as much sense as 'n', no? */
sound/pci/maestro3.c
2287
t = inb(chip->iobase + ASSP_CONTROL_A);
sound/pci/maestro3.c
2288
t &= ~( DSP_CLK_36MHZ_SELECT | ASSP_CLK_49MHZ_SELECT);
sound/pci/maestro3.c
2289
t |= ASSP_CLK_49MHZ_SELECT;
sound/pci/maestro3.c
2290
t |= ASSP_0_WS_ENABLE;
sound/pci/maestro3.c
2291
outb(t, chip->iobase + ASSP_CONTROL_A);
sound/pci/riptide/riptide.c
696
u32 laddr, saddr, t, val;
sound/pci/riptide/riptide.c
703
t = atoh(&in[7], 2);
sound/pci/riptide/riptide.c
704
switch (t) {
sound/pci/rme9652/hdsp.c
1381
static void snd_hdsp_midi_output_timer(struct timer_list *t)
sound/pci/rme9652/hdsp.c
1383
struct hdsp_midi *hmidi = timer_container_of(hmidi, t, timer);
sound/pci/rme9652/hdsp.c
4295
struct snd_interval t = {
sound/pci/rme9652/hdsp.c
4300
return snd_interval_refine(c, &t);
sound/pci/rme9652/hdsp.c
4302
struct snd_interval t = {
sound/pci/rme9652/hdsp.c
4307
return snd_interval_refine(c, &t);
sound/pci/rme9652/hdsp.c
4309
struct snd_interval t = {
sound/pci/rme9652/hdsp.c
4314
return snd_interval_refine(c, &t);
sound/pci/rme9652/hdsp.c
4326
struct snd_interval t = {
sound/pci/rme9652/hdsp.c
4331
return snd_interval_refine(c, &t);
sound/pci/rme9652/hdsp.c
4333
struct snd_interval t = {
sound/pci/rme9652/hdsp.c
4338
return snd_interval_refine(c, &t);
sound/pci/rme9652/hdsp.c
4340
struct snd_interval t = {
sound/pci/rme9652/hdsp.c
4345
return snd_interval_refine(c, &t);
sound/pci/rme9652/hdsp.c
4357
struct snd_interval t = {
sound/pci/rme9652/hdsp.c
4362
return snd_interval_refine(r, &t);
sound/pci/rme9652/hdsp.c
4364
struct snd_interval t = {
sound/pci/rme9652/hdsp.c
4369
return snd_interval_refine(r, &t);
sound/pci/rme9652/hdsp.c
4371
struct snd_interval t = {
sound/pci/rme9652/hdsp.c
4376
return snd_interval_refine(r, &t);
sound/pci/rme9652/hdsp.c
4388
struct snd_interval t = {
sound/pci/rme9652/hdsp.c
4393
return snd_interval_refine(r, &t);
sound/pci/rme9652/hdsp.c
4395
struct snd_interval t = {
sound/pci/rme9652/hdsp.c
4400
return snd_interval_refine(r, &t);
sound/pci/rme9652/hdsp.c
4402
struct snd_interval t = {
sound/pci/rme9652/hdsp.c
4407
return snd_interval_refine(r, &t);
sound/pci/rme9652/hdspm.c
1932
static void snd_hdspm_midi_output_timer(struct timer_list *t)
sound/pci/rme9652/hdspm.c
1934
struct hdspm_midi *hmidi = timer_container_of(hmidi, t, timer);
sound/pci/rme9652/hdspm.c
5831
struct snd_interval t = {
sound/pci/rme9652/hdspm.c
5836
return snd_interval_refine(c, &t);
sound/pci/rme9652/hdspm.c
5838
struct snd_interval t = {
sound/pci/rme9652/hdspm.c
5843
return snd_interval_refine(c, &t);
sound/pci/rme9652/hdspm.c
5845
struct snd_interval t = {
sound/pci/rme9652/hdspm.c
5850
return snd_interval_refine(c, &t);
sound/pci/rme9652/hdspm.c
5866
struct snd_interval t = {
sound/pci/rme9652/hdspm.c
5871
return snd_interval_refine(c, &t);
sound/pci/rme9652/hdspm.c
5873
struct snd_interval t = {
sound/pci/rme9652/hdspm.c
5878
return snd_interval_refine(c, &t);
sound/pci/rme9652/hdspm.c
5880
struct snd_interval t = {
sound/pci/rme9652/hdspm.c
5885
return snd_interval_refine(c, &t);
sound/pci/rme9652/hdspm.c
5901
struct snd_interval t = {
sound/pci/rme9652/hdspm.c
5906
return snd_interval_refine(r, &t);
sound/pci/rme9652/hdspm.c
5908
struct snd_interval t = {
sound/pci/rme9652/hdspm.c
5913
return snd_interval_refine(r, &t);
sound/pci/rme9652/hdspm.c
5915
struct snd_interval t = {
sound/pci/rme9652/hdspm.c
5920
return snd_interval_refine(r, &t);
sound/pci/rme9652/hdspm.c
5935
struct snd_interval t = {
sound/pci/rme9652/hdspm.c
5940
return snd_interval_refine(r, &t);
sound/pci/rme9652/hdspm.c
5942
struct snd_interval t = {
sound/pci/rme9652/hdspm.c
5947
return snd_interval_refine(r, &t);
sound/pci/rme9652/hdspm.c
5949
struct snd_interval t = {
sound/pci/rme9652/hdspm.c
5954
return snd_interval_refine(r, &t);
sound/pci/rme9652/rme9652.c
2147
struct snd_interval t = {
sound/pci/rme9652/rme9652.c
2152
return snd_interval_refine(c, &t);
sound/pci/rme9652/rme9652.c
2154
struct snd_interval t = {
sound/pci/rme9652/rme9652.c
2159
return snd_interval_refine(c, &t);
sound/pci/rme9652/rme9652.c
2171
struct snd_interval t = {
sound/pci/rme9652/rme9652.c
2176
return snd_interval_refine(r, &t);
sound/pci/rme9652/rme9652.c
2178
struct snd_interval t = {
sound/pci/rme9652/rme9652.c
2183
return snd_interval_refine(r, &t);
sound/sh/aica.c
285
static void aica_period_elapsed(struct timer_list *t)
sound/sh/aica.c
288
t, timer);
sound/soc/atmel/atmel_ssc_dai.c
197
struct snd_interval t;
sound/soc/atmel/atmel_ssc_dai.c
248
t.min = 8000;
sound/soc/atmel/atmel_ssc_dai.c
249
t.max = ssc_p->mck_rate / mck_div / frame_size;
sound/soc/atmel/atmel_ssc_dai.c
250
t.openmin = t.openmax = 0;
sound/soc/atmel/atmel_ssc_dai.c
251
t.integer = 0;
sound/soc/atmel/atmel_ssc_dai.c
252
ret = snd_interval_refine(i, &t);
sound/soc/au1x/psc-ac97.c
213
int chans, t, stype = substream->stream;
sound/soc/au1x/psc-ac97.c
255
t = 100;
sound/soc/au1x/psc-ac97.c
256
while ((__raw_readl(AC97_STAT(pscdata)) & PSC_AC97STAT_DR) && --t)
sound/soc/au1x/psc-ac97.c
259
if (!t)
sound/soc/au1x/psc-ac97.c
271
t = 100;
sound/soc/au1x/psc-ac97.c
272
while ((!(__raw_readl(AC97_STAT(pscdata)) & PSC_AC97STAT_DR)) && --t)
sound/soc/au1x/psc-ac97.c
275
if (!t)
sound/soc/codecs/cs35l56-shared.c
492
struct spi_transfer t = {
sound/soc/codecs/cs35l56-shared.c
502
spi_message_init_with_transfers(&m, &t, 1);
sound/soc/codecs/max98088.c
1555
const char **t;
sound/soc/codecs/max98088.c
1587
t = krealloc(max98088->eq_texts,
sound/soc/codecs/max98088.c
1590
if (t == NULL)
sound/soc/codecs/max98088.c
1594
t[max98088->eq_textcnt] = cfg[i].name;
sound/soc/codecs/max98088.c
1596
max98088->eq_texts = t;
sound/soc/codecs/max98095.c
1571
const char **t;
sound/soc/codecs/max98095.c
1603
t = krealloc(max98095->eq_texts,
sound/soc/codecs/max98095.c
1606
if (t == NULL)
sound/soc/codecs/max98095.c
1610
t[max98095->eq_textcnt] = cfg[i].name;
sound/soc/codecs/max98095.c
1612
max98095->eq_texts = t;
sound/soc/codecs/max98095.c
1722
const char **t;
sound/soc/codecs/max98095.c
1755
t = krealloc(max98095->bq_texts,
sound/soc/codecs/max98095.c
1758
if (t == NULL)
sound/soc/codecs/max98095.c
1762
t[max98095->bq_textcnt] = cfg[i].name;
sound/soc/codecs/max98095.c
1764
max98095->bq_texts = t;
sound/soc/codecs/rt1320-sdw.h
120
int t;
sound/soc/codecs/rt5645.c
3442
static void rt5645_btn_check_callback(struct timer_list *t)
sound/soc/codecs/rt5645.c
3444
struct rt5645_priv *rt5645 = timer_container_of(rt5645, t,
sound/soc/codecs/rt5677-spi.c
475
struct spi_transfer t[2];
sound/soc/codecs/rt5677-spi.c
491
memset(t, 0, sizeof(t));
sound/soc/codecs/rt5677-spi.c
492
t[0].tx_buf = header;
sound/soc/codecs/rt5677-spi.c
493
t[0].len = sizeof(header);
sound/soc/codecs/rt5677-spi.c
494
t[0].speed_hz = RT5677_SPI_FREQ;
sound/soc/codecs/rt5677-spi.c
495
t[1].rx_buf = body;
sound/soc/codecs/rt5677-spi.c
496
t[1].speed_hz = RT5677_SPI_FREQ;
sound/soc/codecs/rt5677-spi.c
497
spi_message_init_with_transfers(&m, t, ARRAY_SIZE(t));
sound/soc/codecs/rt5677-spi.c
499
for (offset = 0; offset < len; offset += t[1].len) {
sound/soc/codecs/rt5677-spi.c
501
len - offset, &t[1].len);
sound/soc/codecs/rt5677-spi.c
516
rt5677_spi_reverse(cb + offset, len - offset, body, t[1].len);
sound/soc/codecs/rt5677-spi.c
530
struct spi_transfer t;
sound/soc/codecs/rt5677-spi.c
546
memset(&t, 0, sizeof(t));
sound/soc/codecs/rt5677-spi.c
547
t.tx_buf = buf;
sound/soc/codecs/rt5677-spi.c
548
t.speed_hz = RT5677_SPI_FREQ;
sound/soc/codecs/rt5677-spi.c
549
spi_message_init_with_transfers(&m, &t, 1);
sound/soc/codecs/rt5677-spi.c
553
len - offset, &t.len);
sound/soc/codecs/rt5677-spi.c
563
rt5677_spi_reverse(body, t.len, cb + offset, len - offset);
sound/soc/codecs/rt5677-spi.c
564
offset += t.len;
sound/soc/codecs/rt5677-spi.c
565
t.len += RT5677_SPI_HEADER + 1;
sound/soc/codecs/sgtl5000.c
1003
t = do_div(out, in);
sound/soc/codecs/sgtl5000.c
1005
t *= 2048;
sound/soc/codecs/sgtl5000.c
1006
do_div(t, in);
sound/soc/codecs/sgtl5000.c
1007
frac_div = t;
sound/soc/codecs/sgtl5000.c
987
u64 out, t;
sound/soc/codecs/tas2552.c
187
unsigned int d, q, t;
sound/soc/codecs/tas2552.c
195
t = (pll_clk * 2) << p;
sound/soc/codecs/tas2552.c
196
j = t / pll_clkin;
sound/soc/codecs/tas2552.c
197
d = t % pll_clkin;
sound/soc/codecs/tas2552.c
198
t = pll_clkin / 10000;
sound/soc/codecs/tas2552.c
199
q = d / (t + 1);
sound/soc/codecs/tas2552.c
200
d = q + ((9999 - pll_clkin % 10000) * (d / t - q)) / 10000;
sound/soc/codecs/tas2781-i2c.c
1293
struct fct_param_address *t)
sound/soc/codecs/tas2781-i2c.c
1301
reg = TASDEVICE_REG(t->thr[0], t->thr[1], t->thr[2]);
sound/soc/codecs/tas2781-i2c.c
1304
reg = TASDEVICE_REG(t->plt_flg[0], t->plt_flg[1],
sound/soc/codecs/tas2781-i2c.c
1305
t->plt_flg[2]);
sound/soc/codecs/tas2781-i2c.c
1308
reg = TASDEVICE_REG(t->sin_gn[0], t->sin_gn[1],
sound/soc/codecs/tas2781-i2c.c
1309
t->sin_gn[2]);
sound/soc/codecs/tas2781-i2c.c
1312
reg = TASDEVICE_REG(t->sin_gn[0], t->sin_gn[1],
sound/soc/codecs/tas2781-i2c.c
1313
t->sin_gn[2]);
sound/soc/codecs/tas2781-i2c.c
1325
struct fct_param_address *t)
sound/soc/codecs/tas2781-i2c.c
1328
p->reg = TASDEVICE_REG(t->thr2[0], t->thr2[1], t->thr2[2]);
sound/soc/codecs/tas2781-i2c.c
1352
struct fct_param_address *t = &(fmw->fct_par_addr);
sound/soc/codecs/tas2781-i2c.c
1365
cali_reg_update(p, t);
sound/soc/codecs/tas2781-i2c.c
1368
alpa_cali_update(p, t);
sound/soc/codecs/tas2781-i2c.c
406
struct bulk_reg_val *t = &tasdev[i].alp_cali_bckp;
sound/soc/codecs/tas2781-i2c.c
437
tasdevice_dev_bulk_read(tas_priv, i, t->reg, t->val, 4);
sound/soc/codecs/tas2781-i2c.c
473
tasdevice_dev_bulk_read(tas_priv, i, t->reg, val, 4);
sound/soc/codecs/tas2781-i2c.c
523
struct bulk_reg_val *t = &tasdev[i].alp_cali_bckp;
sound/soc/codecs/tas2781-i2c.c
545
tasdevice_dev_bulk_write(priv, i, t->reg, t->val, 4);
sound/soc/codecs/tas2783-sdw.c
1085
unsigned long t;
sound/soc/codecs/tas2783-sdw.c
1090
t = wait_for_completion_timeout(&slave->initialization_complete,
sound/soc/codecs/tas2783-sdw.c
1092
if (!t) {
sound/soc/codecs/tscs454.c
513
#define PLL_CTL(f, t, c1, r1, o1, f1l, f1h, c2, r2, o2, f2l, f2h) \
sound/soc/codecs/tscs454.c
527
{R_TIMEBASE, t}, \
sound/soc/codecs/wm0010.c
186
struct spi_transfer t;
sound/soc/codecs/wm0010.c
210
u32 *out32 = xfer->t.rx_buf;
sound/soc/codecs/wm0010.c
222
for (i = 0; i < xfer->t.len / 4; i++) {
sound/soc/codecs/wm0010.c
413
xfer->t.rx_buf = out;
sound/soc/codecs/wm0010.c
420
xfer->t.tx_buf = img;
sound/soc/codecs/wm0010.c
427
xfer->t.len = len;
sound/soc/codecs/wm0010.c
428
xfer->t.bits_per_word = 8;
sound/soc/codecs/wm0010.c
431
xfer->t.speed_hz = wm0010->sysclk / 6;
sound/soc/codecs/wm0010.c
433
xfer->t.speed_hz = wm0010->max_spi_freq;
sound/soc/codecs/wm0010.c
437
xfer->t.speed_hz = wm0010->board_max_spi_speed;
sound/soc/codecs/wm0010.c
441
wm0010->max_spi_freq = xfer->t.speed_hz;
sound/soc/codecs/wm0010.c
443
spi_message_add_tail(&xfer->t, &xfer->m);
sound/soc/codecs/wm0010.c
474
kfree(xfer->t.rx_buf);
sound/soc/codecs/wm0010.c
475
kfree(xfer->t.tx_buf);
sound/soc/codecs/wm0010.c
491
struct spi_transfer t;
sound/soc/codecs/wm0010.c
520
memset(&t, 0, sizeof(t));
sound/soc/codecs/wm0010.c
521
t.rx_buf = out;
sound/soc/codecs/wm0010.c
522
t.tx_buf = img;
sound/soc/codecs/wm0010.c
523
t.len = fw->size;
sound/soc/codecs/wm0010.c
524
t.bits_per_word = 8;
sound/soc/codecs/wm0010.c
525
t.speed_hz = wm0010->sysclk / 10;
sound/soc/codecs/wm0010.c
526
spi_message_add_tail(&t, &m);
sound/soc/codecs/wm0010.c
529
t.speed_hz);
sound/soc/codecs/wm0010.c
564
struct spi_transfer t;
sound/soc/codecs/wm0010.c
653
memset(&t, 0, sizeof(t));
sound/soc/codecs/wm0010.c
654
t.rx_buf = out;
sound/soc/codecs/wm0010.c
655
t.tx_buf = img_swap;
sound/soc/codecs/wm0010.c
656
t.len = len;
sound/soc/codecs/wm0010.c
657
t.bits_per_word = 8;
sound/soc/codecs/wm0010.c
658
t.speed_hz = wm0010->sysclk / 6;
sound/soc/codecs/wm0010.c
659
spi_message_add_tail(&t, &m);
sound/soc/codecs/wm8904.c
2069
const char **t;
sound/soc/codecs/wm8904.c
2088
t = krealloc(wm8904->retune_mobile_texts,
sound/soc/codecs/wm8904.c
2092
if (t == NULL)
sound/soc/codecs/wm8904.c
2096
t[wm8904->num_retune_mobile_texts] =
sound/soc/codecs/wm8904.c
2101
wm8904->retune_mobile_texts = t;
sound/soc/codecs/wm8994.c
3365
const char **t;
sound/soc/codecs/wm8994.c
3384
t = krealloc(wm8994->retune_mobile_texts,
sound/soc/codecs/wm8994.c
3388
if (t == NULL)
sound/soc/codecs/wm8994.c
3392
t[wm8994->num_retune_mobile_texts] =
sound/soc/codecs/wm8994.c
3397
wm8994->retune_mobile_texts = t;
sound/soc/codecs/wm8996.c
2556
const char **t;
sound/soc/codecs/wm8996.c
2575
t = krealloc(wm8996->retune_mobile_texts,
sound/soc/codecs/wm8996.c
2579
if (t == NULL)
sound/soc/codecs/wm8996.c
2583
t[wm8996->num_retune_mobile_texts] =
sound/soc/codecs/wm8996.c
2588
wm8996->retune_mobile_texts = t;
sound/soc/fsl/fsl_easrc.c
291
static inline uint32_t bits_taps_to_val(unsigned int t)
sound/soc/fsl/fsl_easrc.c
293
switch (t) {
sound/soc/fsl/fsl_xcvr.h
103
#define FSL_XCVR_EXT_CTRL_DPTH_RESET(t) (t ? BIT(27) : BIT(28))
sound/soc/fsl/fsl_xcvr.h
108
#define FSL_XCVR_EXT_CTRL_DMA_DIS(t) (t ? BIT(24) : BIT(25))
sound/soc/fsl/fsl_xcvr.h
180
#define FSL_XCVR_ISR_SET_SPDIF_MODE(t) (t ? BIT(21) : BIT(20))
sound/soc/fsl/fsl_xcvr.h
99
#define FSL_XCVR_EXT_CTRL_CMDC_RESET(t) (t ? BIT(29) : BIT(30))
sound/soc/fsl/imx-card.c
419
struct snd_interval t = { .min = 8000, .max = 8000, };
sound/soc/fsl/imx-card.c
437
if (t.max < akcodec_rates[i])
sound/soc/fsl/imx-card.c
438
t.max = akcodec_rates[i];
sound/soc/fsl/imx-card.c
441
return snd_interval_refine(hw_param_interval(p, r->var), &t);
sound/soc/fsl/imx-pcm-rpmsg.c
209
static void imx_rpmsg_timer_callback(struct timer_list *t)
sound/soc/fsl/imx-pcm-rpmsg.c
212
timer_container_of(stream_timer, t, timer);
sound/soc/intel/avs/path.c
168
avs_nhlt_config_or_default(struct avs_dev *adev, struct avs_tplg_module *t);
sound/soc/intel/avs/path.c
278
avs_nhlt_config_or_default(struct avs_dev *adev, struct avs_tplg_module *t)
sound/soc/intel/avs/path.c
286
te = t->cfg_ext;
sound/soc/intel/avs/path.c
302
fmt = t->in_fmt;
sound/soc/intel/avs/path.c
310
fmt = t->in_fmt;
sound/soc/intel/avs/path.c
339
struct avs_tplg_module *t, u32 dma_id, size_t *cfg_size)
sound/soc/intel/avs/path.c
341
u32 dma_type = t->cfg_ext->copier.dma_type;
sound/soc/intel/avs/path.c
382
struct avs_tplg_module *t, u32 dma_id, size_t *cfg_size)
sound/soc/intel/avs/path.c
387
if (t->nhlt_config)
sound/soc/intel/avs/path.c
388
blob = t->nhlt_config->blob;
sound/soc/intel/avs/path.c
390
blob = avs_nhlt_config_or_default(adev, t);
sound/soc/intel/avs/path.c
402
return avs_append_dma_cfg(adev, gtw, t, dma_id, cfg_size);
sound/soc/intel/avs/path.c
407
struct avs_tplg_module *t = mod->template;
sound/soc/intel/avs/path.c
414
te = t->cfg_ext;
sound/soc/intel/avs/path.c
419
ret = avs_fill_gtw_config(adev, &cfg->gtw_cfg, t, dma_id, &cfg_size);
sound/soc/intel/avs/path.c
423
cfg->base.cpc = t->cfg_base->cpc;
sound/soc/intel/avs/path.c
424
cfg->base.ibs = t->cfg_base->ibs;
sound/soc/intel/avs/path.c
425
cfg->base.obs = t->cfg_base->obs;
sound/soc/intel/avs/path.c
426
cfg->base.is_pages = t->cfg_base->is_pages;
sound/soc/intel/avs/path.c
427
cfg->base.audio_fmt = *t->in_fmt;
sound/soc/intel/avs/path.c
434
ret = avs_dsp_init_module(adev, mod->module_id, mod->owner->instance_id, t->core_id,
sound/soc/intel/avs/path.c
435
t->domain, cfg, cfg_size, &mod->instance_id);
sound/soc/intel/avs/path.c
441
struct avs_tplg_module *t = mod->template;
sound/soc/intel/avs/path.c
448
te = t->cfg_ext;
sound/soc/intel/avs/path.c
453
ret = avs_fill_gtw_config(adev, &cfg->gtw_cfg, t, dma_id, &cfg_size);
sound/soc/intel/avs/path.c
457
cfg->base.cpc = t->cfg_base->cpc;
sound/soc/intel/avs/path.c
458
cfg->base.ibs = t->cfg_base->ibs;
sound/soc/intel/avs/path.c
459
cfg->base.obs = t->cfg_base->obs;
sound/soc/intel/avs/path.c
460
cfg->base.is_pages = t->cfg_base->is_pages;
sound/soc/intel/avs/path.c
461
cfg->base.audio_fmt = *t->in_fmt;
sound/soc/intel/avs/path.c
469
ret = avs_dsp_init_module(adev, mod->module_id, mod->owner->instance_id, t->core_id,
sound/soc/intel/avs/path.c
470
t->domain, cfg, cfg_size, &mod->instance_id);
sound/soc/intel/avs/path.c
477
struct avs_tplg_module *t = mod->template;
sound/soc/intel/avs/path.c
482
path_tmpl = t->owner->owner->owner;
sound/soc/intel/avs/path.c
491
if (ctl_data->id == t->ctl_id && strstr(w->kcontrols[i]->id.name, name))
sound/soc/intel/avs/path.c
503
struct avs_tplg_module *t;
sound/soc/intel/avs/path.c
507
t = mod->template;
sound/soc/intel/avs/path.c
515
vols[i].curve_type = t->cfg_ext->peakvol.curve_type;
sound/soc/intel/avs/path.c
516
vols[i].curve_duration = t->cfg_ext->peakvol.curve_duration;
sound/soc/intel/avs/path.c
527
vols[0].curve_type = t->cfg_ext->peakvol.curve_type;
sound/soc/intel/avs/path.c
528
vols[0].curve_duration = t->cfg_ext->peakvol.curve_duration;
sound/soc/intel/avs/path.c
539
struct avs_tplg_module *t;
sound/soc/intel/avs/path.c
543
t = mod->template;
sound/soc/intel/avs/path.c
551
mutes[i].curve_type = t->cfg_ext->peakvol.curve_type;
sound/soc/intel/avs/path.c
552
mutes[i].curve_duration = t->cfg_ext->peakvol.curve_duration;
sound/soc/intel/avs/path.c
563
mutes[0].curve_type = t->cfg_ext->peakvol.curve_type;
sound/soc/intel/avs/path.c
564
mutes[0].curve_duration = t->cfg_ext->peakvol.curve_duration;
sound/soc/intel/avs/path.c
572
struct avs_tplg_module *t = mod->template;
sound/soc/intel/avs/path.c
584
cfg->base.cpc = t->cfg_base->cpc;
sound/soc/intel/avs/path.c
585
cfg->base.ibs = t->cfg_base->ibs;
sound/soc/intel/avs/path.c
586
cfg->base.obs = t->cfg_base->obs;
sound/soc/intel/avs/path.c
587
cfg->base.is_pages = t->cfg_base->is_pages;
sound/soc/intel/avs/path.c
588
cfg->base.audio_fmt = *t->in_fmt;
sound/soc/intel/avs/path.c
591
cfg->vols[0].curve_type = t->cfg_ext->peakvol.curve_type;
sound/soc/intel/avs/path.c
592
cfg->vols[0].curve_duration = t->cfg_ext->peakvol.curve_duration;
sound/soc/intel/avs/path.c
594
ret = avs_dsp_init_module(adev, mod->module_id, mod->owner->instance_id, t->core_id,
sound/soc/intel/avs/path.c
595
t->domain, cfg, cfg_size, &mod->instance_id);
sound/soc/intel/avs/path.c
615
struct avs_tplg_module *t = mod->template;
sound/soc/intel/avs/path.c
619
cfg.base.cpc = t->cfg_base->cpc;
sound/soc/intel/avs/path.c
620
cfg.base.ibs = t->cfg_base->ibs;
sound/soc/intel/avs/path.c
621
cfg.base.obs = t->cfg_base->obs;
sound/soc/intel/avs/path.c
622
cfg.base.is_pages = t->cfg_base->is_pages;
sound/soc/intel/avs/path.c
623
cfg.base.audio_fmt = *t->in_fmt;
sound/soc/intel/avs/path.c
624
cfg.out_channel_config = t->cfg_ext->updown_mix.out_channel_config;
sound/soc/intel/avs/path.c
625
cfg.coefficients_select = t->cfg_ext->updown_mix.coefficients_select;
sound/soc/intel/avs/path.c
627
cfg.coefficients[i] = t->cfg_ext->updown_mix.coefficients[i];
sound/soc/intel/avs/path.c
628
cfg.channel_map = t->cfg_ext->updown_mix.channel_map;
sound/soc/intel/avs/path.c
631
t->core_id, t->domain, &cfg, sizeof(cfg),
sound/soc/intel/avs/path.c
637
struct avs_tplg_module *t = mod->template;
sound/soc/intel/avs/path.c
640
cfg.base.cpc = t->cfg_base->cpc;
sound/soc/intel/avs/path.c
641
cfg.base.ibs = t->cfg_base->ibs;
sound/soc/intel/avs/path.c
642
cfg.base.obs = t->cfg_base->obs;
sound/soc/intel/avs/path.c
643
cfg.base.is_pages = t->cfg_base->is_pages;
sound/soc/intel/avs/path.c
644
cfg.base.audio_fmt = *t->in_fmt;
sound/soc/intel/avs/path.c
645
cfg.out_freq = t->cfg_ext->src.out_freq;
sound/soc/intel/avs/path.c
648
t->core_id, t->domain, &cfg, sizeof(cfg),
sound/soc/intel/avs/path.c
654
struct avs_tplg_module *t = mod->template;
sound/soc/intel/avs/path.c
658
cfg.base.cpc = t->cfg_base->cpc;
sound/soc/intel/avs/path.c
659
cfg.base.ibs = t->cfg_base->ibs;
sound/soc/intel/avs/path.c
660
cfg.base.obs = t->cfg_base->obs;
sound/soc/intel/avs/path.c
661
cfg.base.is_pages = t->cfg_base->is_pages;
sound/soc/intel/avs/path.c
662
cfg.base.audio_fmt = *t->in_fmt;
sound/soc/intel/avs/path.c
663
cfg.out_freq = t->cfg_ext->asrc.out_freq;
sound/soc/intel/avs/path.c
664
cfg.mode = t->cfg_ext->asrc.mode;
sound/soc/intel/avs/path.c
665
cfg.disable_jitter_buffer = t->cfg_ext->asrc.disable_jitter_buffer;
sound/soc/intel/avs/path.c
668
t->core_id, t->domain, &cfg, sizeof(cfg),
sound/soc/intel/avs/path.c
674
struct avs_tplg_module *t = mod->template;
sound/soc/intel/avs/path.c
677
cfg.base.cpc = t->cfg_base->cpc;
sound/soc/intel/avs/path.c
678
cfg.base.ibs = t->cfg_base->ibs;
sound/soc/intel/avs/path.c
679
cfg.base.obs = t->cfg_base->obs;
sound/soc/intel/avs/path.c
680
cfg.base.is_pages = t->cfg_base->is_pages;
sound/soc/intel/avs/path.c
681
cfg.base.audio_fmt = *t->in_fmt;
sound/soc/intel/avs/path.c
682
cfg.ref_fmt = *t->cfg_ext->aec.ref_fmt;
sound/soc/intel/avs/path.c
683
cfg.out_fmt = *t->cfg_ext->aec.out_fmt;
sound/soc/intel/avs/path.c
684
cfg.cpc_lp_mode = t->cfg_ext->aec.cpc_lp_mode;
sound/soc/intel/avs/path.c
687
t->core_id, t->domain, &cfg, sizeof(cfg),
sound/soc/intel/avs/path.c
693
struct avs_tplg_module *t = mod->template;
sound/soc/intel/avs/path.c
696
cfg.base.cpc = t->cfg_base->cpc;
sound/soc/intel/avs/path.c
697
cfg.base.ibs = t->cfg_base->ibs;
sound/soc/intel/avs/path.c
698
cfg.base.obs = t->cfg_base->obs;
sound/soc/intel/avs/path.c
699
cfg.base.is_pages = t->cfg_base->is_pages;
sound/soc/intel/avs/path.c
700
cfg.base.audio_fmt = *t->in_fmt;
sound/soc/intel/avs/path.c
701
cfg.ref_fmt = *t->cfg_ext->mux.ref_fmt;
sound/soc/intel/avs/path.c
702
cfg.out_fmt = *t->cfg_ext->mux.out_fmt;
sound/soc/intel/avs/path.c
705
t->core_id, t->domain, &cfg, sizeof(cfg),
sound/soc/intel/avs/path.c
711
struct avs_tplg_module *t = mod->template;
sound/soc/intel/avs/path.c
714
cfg.base.cpc = t->cfg_base->cpc;
sound/soc/intel/avs/path.c
715
cfg.base.ibs = t->cfg_base->ibs;
sound/soc/intel/avs/path.c
716
cfg.base.obs = t->cfg_base->obs;
sound/soc/intel/avs/path.c
717
cfg.base.is_pages = t->cfg_base->is_pages;
sound/soc/intel/avs/path.c
718
cfg.base.audio_fmt = *t->in_fmt;
sound/soc/intel/avs/path.c
719
cfg.cpc_lp_mode = t->cfg_ext->wov.cpc_lp_mode;
sound/soc/intel/avs/path.c
722
t->core_id, t->domain, &cfg, sizeof(cfg),
sound/soc/intel/avs/path.c
728
struct avs_tplg_module *t = mod->template;
sound/soc/intel/avs/path.c
731
cfg.base.cpc = t->cfg_base->cpc;
sound/soc/intel/avs/path.c
732
cfg.base.ibs = t->cfg_base->ibs;
sound/soc/intel/avs/path.c
733
cfg.base.obs = t->cfg_base->obs;
sound/soc/intel/avs/path.c
734
cfg.base.is_pages = t->cfg_base->is_pages;
sound/soc/intel/avs/path.c
735
cfg.base.audio_fmt = *t->in_fmt;
sound/soc/intel/avs/path.c
736
cfg.out_fmt = *t->cfg_ext->micsel.out_fmt;
sound/soc/intel/avs/path.c
739
t->core_id, t->domain, &cfg, sizeof(cfg),
sound/soc/intel/avs/path.c
745
struct avs_tplg_module *t = mod->template;
sound/soc/intel/avs/path.c
748
cfg.cpc = t->cfg_base->cpc;
sound/soc/intel/avs/path.c
749
cfg.ibs = t->cfg_base->ibs;
sound/soc/intel/avs/path.c
750
cfg.obs = t->cfg_base->obs;
sound/soc/intel/avs/path.c
751
cfg.is_pages = t->cfg_base->is_pages;
sound/soc/intel/avs/path.c
752
cfg.audio_fmt = *t->in_fmt;
sound/soc/intel/avs/path.c
755
t->core_id, t->domain, &cfg, sizeof(cfg),
sound/soc/intel/avs/path.c
761
struct avs_tplg_module *t = mod->template;
sound/soc/intel/avs/path.c
762
struct avs_tplg_modcfg_ext *tcfg = t->cfg_ext;
sound/soc/intel/avs/path.c
775
cfg->base.cpc = t->cfg_base->cpc;
sound/soc/intel/avs/path.c
776
cfg->base.ibs = t->cfg_base->ibs;
sound/soc/intel/avs/path.c
777
cfg->base.obs = t->cfg_base->obs;
sound/soc/intel/avs/path.c
778
cfg->base.is_pages = t->cfg_base->is_pages;
sound/soc/intel/avs/path.c
779
cfg->base.audio_fmt = *t->in_fmt;
sound/soc/intel/avs/path.c
794
t->core_id, t->domain, cfg, cfg_size,
sound/soc/intel/avs/path.c
913
struct avs_tplg_binding *t;
sound/soc/intel/avs/path.c
915
t = binding->template;
sound/soc/intel/avs/path.c
917
t->mod_id);
sound/soc/intel/avs/path.c
919
dev_err(adev->dev, "path mod %d not found\n", t->mod_id);
sound/soc/intel/avs/path.c
924
target_path = avs_path_find_path(adev, t->target_tplg_name,
sound/soc/intel/avs/path.c
925
t->target_path_tmpl_id);
sound/soc/intel/avs/path.c
928
t->target_tplg_name, t->target_path_tmpl_id);
sound/soc/intel/avs/path.c
933
t->target_ppl_id);
sound/soc/intel/avs/path.c
935
dev_err(adev->dev, "target ppl %d not found\n", t->target_ppl_id);
sound/soc/intel/avs/path.c
939
target_mod = avs_path_find_module(target_ppl, t->target_mod_id);
sound/soc/intel/avs/path.c
941
dev_err(adev->dev, "target mod %d not found\n", t->target_mod_id);
sound/soc/intel/avs/path.c
945
if (t->is_sink) {
sound/soc/intel/avs/path.c
947
binding->sink_pin = t->mod_pin;
sound/soc/intel/avs/path.c
949
binding->source_pin = t->target_mod_pin;
sound/soc/intel/avs/path.c
952
binding->sink_pin = t->target_mod_pin;
sound/soc/intel/avs/path.c
954
binding->source_pin = t->mod_pin;
sound/soc/intel/avs/path.c
967
struct avs_tplg_binding *t)
sound/soc/intel/avs/path.c
975
binding->template = t;
sound/soc/intel/common/soc-acpi-intel-ssp-common.c
12
#define CODEC_MAP_ENTRY(n, s, h, t) \
sound/soc/intel/common/soc-acpi-intel-ssp-common.c
17
.codec_type = t, \
sound/soc/meson/axg-spdifin.c
191
unsigned int t;
sound/soc/meson/axg-spdifin.c
196
t = axg_spdifin_mode_timer(priv, i, rate);
sound/soc/meson/axg-spdifin.c
199
axg_spdifin_write_timer(priv->map, i, t);
sound/soc/meson/axg-spdifin.c
202
axg_spdifin_write_threshold(priv->map, i, 3 * (t + t_next));
sound/soc/meson/axg-spdifin.c
205
t_next = t;
sound/soc/samsung/i2s.c
28
#define msecs_to_loops(t) (loops_per_jiffy / 1000 * HZ * t)
sound/soc/sti/sti_uniperif.c
168
struct snd_interval t;
sound/soc/sti/sti_uniperif.c
170
t.min = uni->tdm_slot.avail_slots;
sound/soc/sti/sti_uniperif.c
171
t.max = uni->tdm_slot.avail_slots;
sound/soc/sti/sti_uniperif.c
172
t.openmin = 0;
sound/soc/sti/sti_uniperif.c
173
t.openmax = 0;
sound/soc/sti/sti_uniperif.c
174
t.integer = 0;
sound/soc/sti/sti_uniperif.c
176
return snd_interval_refine(hw_param_interval(params, rule->var), &t);
sound/soc/ti/j721e-evm.c
249
struct snd_interval *t = rule->private;
sound/soc/ti/j721e-evm.c
251
return snd_interval_refine(hw_param_interval(params, rule->var), t);
sound/synth/emux/emux_synth.c
187
void snd_emux_timer_callback(struct timer_list *t)
sound/synth/emux/emux_synth.c
189
struct snd_emux *emu = timer_container_of(emu, t, tlist);
sound/synth/emux/emux_voice.h
45
void snd_emux_timer_callback(struct timer_list *t);
sound/synth/emux/soundfont.c
843
int r, p, t;
sound/synth/emux/soundfont.c
848
t = end - start;
sound/synth/emux/soundfont.c
849
if (t < 0) t = -t;
sound/synth/emux/soundfont.c
851
t = t << (13 - r);
sound/synth/emux/soundfont.c
853
t = t >> (r - 13);
sound/synth/emux/soundfont.c
854
return (t * 10) / (p * 441);
sound/usb/midi.c
335
static void snd_usbmidi_error_timer(struct timer_list *t)
sound/usb/midi.c
337
struct snd_usb_midi *umidi = timer_container_of(umidi, t, error_timer);
sound/usb/mixer_quirks.c
129
const struct std_mono_table *t)
sound/usb/mixer_quirks.c
133
while (t->name) {
sound/usb/mixer_quirks.c
134
err = snd_create_std_mono_ctl(mixer, t->unitid, t->control,
sound/usb/mixer_quirks.c
135
t->cmask, t->val_type, t->name,
sound/usb/mixer_quirks.c
136
t->tlv_callback);
sound/usb/mixer_quirks.c
139
t++;
sound/usb/usx2y/us144mkii.c
382
static void tascam_error_timer(struct timer_list *t)
sound/usb/usx2y/us144mkii.c
385
container_of(t, struct tascam_card, error_timer);
sound/usb/validate.c
244
#define FIXED(p, t, s) { .protocol = (p), .type = (t), .size = sizeof(s) }
sound/usb/validate.c
245
#define FUNC(p, t, f) { .protocol = (p), .type = (t), .func = (f) }
sound/x86/intel_hdmi_audio.c
458
const struct channel_map_table *t = map_tables;
sound/x86/intel_hdmi_audio.c
460
for (; t->map; t++) {
sound/x86/intel_hdmi_audio.c
461
if (t->spk_mask == spk)
sound/x86/intel_hdmi_audio.c
462
return t->map;
tools/accounting/delaytop.c
173
SORT_FIELD(thrashing, t, MODE_MEMVERBOSE),
tools/accounting/delaytop.c
378
static void set_mem_delay_total(struct task_info *t)
tools/accounting/delaytop.c
380
t->mem_delay_total = t->swapin_delay_total +
tools/accounting/delaytop.c
381
t->freepages_delay_total +
tools/accounting/delaytop.c
382
t->thrashing_delay_total +
tools/accounting/delaytop.c
383
t->compact_delay_total +
tools/accounting/delaytop.c
384
t->wpcopy_delay_total;
tools/accounting/delaytop.c
387
static void set_mem_count(struct task_info *t)
tools/accounting/delaytop.c
389
t->mem_count = t->swapin_count +
tools/accounting/delaytop.c
390
t->freepages_count +
tools/accounting/delaytop.c
391
t->thrashing_count +
tools/accounting/delaytop.c
392
t->compact_count +
tools/accounting/delaytop.c
393
t->wpcopy_count;
tools/accounting/getdelays.c
195
#define average_ms(t, c) (t / 1000000ULL / (c ? c : 1))
tools/accounting/getdelays.c
196
#define delay_ms(t) (t / 1000000ULL)
tools/accounting/getdelays.c
245
#define PRINT_CPU_DELAY(version, t) \
tools/accounting/getdelays.c
253
(unsigned long long)(t)->cpu_count, \
tools/accounting/getdelays.c
254
(unsigned long long)(t)->cpu_run_real_total, \
tools/accounting/getdelays.c
255
(unsigned long long)(t)->cpu_run_virtual_total, \
tools/accounting/getdelays.c
256
(unsigned long long)(t)->cpu_delay_total, \
tools/accounting/getdelays.c
257
average_ms((double)(t)->cpu_delay_total, (t)->cpu_count), \
tools/accounting/getdelays.c
258
delay_ms((double)(t)->cpu_delay_max), \
tools/accounting/getdelays.c
259
delay_ms((double)(t)->cpu_delay_min), \
tools/accounting/getdelays.c
260
format_timespec(&(t)->cpu_delay_max_ts)); \
tools/accounting/getdelays.c
266
(unsigned long long)(t)->cpu_count, \
tools/accounting/getdelays.c
267
(unsigned long long)(t)->cpu_run_real_total, \
tools/accounting/getdelays.c
268
(unsigned long long)(t)->cpu_run_virtual_total, \
tools/accounting/getdelays.c
269
(unsigned long long)(t)->cpu_delay_total, \
tools/accounting/getdelays.c
270
average_ms((double)(t)->cpu_delay_total, (t)->cpu_count), \
tools/accounting/getdelays.c
271
delay_ms((double)(t)->cpu_delay_max), \
tools/accounting/getdelays.c
272
delay_ms((double)(t)->cpu_delay_min)); \
tools/accounting/getdelays.c
278
(unsigned long long)(t)->cpu_count, \
tools/accounting/getdelays.c
279
(unsigned long long)(t)->cpu_run_real_total, \
tools/accounting/getdelays.c
280
(unsigned long long)(t)->cpu_run_virtual_total, \
tools/accounting/getdelays.c
281
(unsigned long long)(t)->cpu_delay_total, \
tools/accounting/getdelays.c
282
average_ms((double)(t)->cpu_delay_total, (t)->cpu_count)); \
tools/accounting/getdelays.c
285
#define PRINT_FILED_DELAY(name, version, t, count, total, max, min) \
tools/accounting/getdelays.c
292
(unsigned long long)(t)->count, \
tools/accounting/getdelays.c
293
(unsigned long long)(t)->total, \
tools/accounting/getdelays.c
294
average_ms((double)(t)->total, (t)->count), \
tools/accounting/getdelays.c
295
delay_ms((double)(t)->max), \
tools/accounting/getdelays.c
296
delay_ms((double)(t)->min)); \
tools/accounting/getdelays.c
301
(unsigned long long)(t)->count, \
tools/accounting/getdelays.c
302
(unsigned long long)(t)->total, \
tools/accounting/getdelays.c
303
average_ms((double)(t)->total, (t)->count)); \
tools/accounting/getdelays.c
307
#define PRINT_FILED_DELAY_WITH_TS(name, version, t, count, total, max, min, max_ts) \
tools/accounting/getdelays.c
314
(unsigned long long)(t)->count, \
tools/accounting/getdelays.c
315
(unsigned long long)(t)->total, \
tools/accounting/getdelays.c
316
average_ms((double)(t)->total, (t)->count), \
tools/accounting/getdelays.c
317
delay_ms((double)(t)->max), \
tools/accounting/getdelays.c
318
delay_ms((double)(t)->min), \
tools/accounting/getdelays.c
319
format_timespec(&(t)->max_ts)); \
tools/accounting/getdelays.c
325
(unsigned long long)(t)->count, \
tools/accounting/getdelays.c
326
(unsigned long long)(t)->total, \
tools/accounting/getdelays.c
327
average_ms((double)(t)->total, (t)->count), \
tools/accounting/getdelays.c
328
delay_ms((double)(t)->max), \
tools/accounting/getdelays.c
329
delay_ms((double)(t)->min)); \
tools/accounting/getdelays.c
334
(unsigned long long)(t)->count, \
tools/accounting/getdelays.c
335
(unsigned long long)(t)->total, \
tools/accounting/getdelays.c
336
average_ms((double)(t)->total, (t)->count)); \
tools/accounting/getdelays.c
340
static void print_delayacct(struct taskstats *t)
tools/accounting/getdelays.c
344
PRINT_CPU_DELAY(t->version, t);
tools/accounting/getdelays.c
347
if (t->version >= 17) {
tools/accounting/getdelays.c
348
PRINT_FILED_DELAY_WITH_TS("IO", t->version, t,
tools/accounting/getdelays.c
352
PRINT_FILED_DELAY_WITH_TS("SWAP", t->version, t,
tools/accounting/getdelays.c
356
PRINT_FILED_DELAY_WITH_TS("RECLAIM", t->version, t,
tools/accounting/getdelays.c
360
PRINT_FILED_DELAY_WITH_TS("THRASHING", t->version, t,
tools/accounting/getdelays.c
364
if (t->version >= 11) {
tools/accounting/getdelays.c
365
PRINT_FILED_DELAY_WITH_TS("COMPACT", t->version, t,
tools/accounting/getdelays.c
370
if (t->version >= 13) {
tools/accounting/getdelays.c
371
PRINT_FILED_DELAY_WITH_TS("WPCOPY", t->version, t,
tools/accounting/getdelays.c
376
if (t->version >= 14) {
tools/accounting/getdelays.c
377
PRINT_FILED_DELAY_WITH_TS("IRQ", t->version, t,
tools/accounting/getdelays.c
383
PRINT_FILED_DELAY("IO", t->version, t,
tools/accounting/getdelays.c
387
PRINT_FILED_DELAY("SWAP", t->version, t,
tools/accounting/getdelays.c
391
PRINT_FILED_DELAY("RECLAIM", t->version, t,
tools/accounting/getdelays.c
395
PRINT_FILED_DELAY("THRASHING", t->version, t,
tools/accounting/getdelays.c
399
if (t->version >= 11) {
tools/accounting/getdelays.c
400
PRINT_FILED_DELAY("COMPACT", t->version, t,
tools/accounting/getdelays.c
405
if (t->version >= 13) {
tools/accounting/getdelays.c
406
PRINT_FILED_DELAY("WPCOPY", t->version, t,
tools/accounting/getdelays.c
411
if (t->version >= 14) {
tools/accounting/getdelays.c
412
PRINT_FILED_DELAY("IRQ", t->version, t,
tools/accounting/getdelays.c
419
static void task_context_switch_counts(struct taskstats *t)
tools/accounting/getdelays.c
424
(unsigned long long)t->nvcsw, (unsigned long long)t->nivcsw);
tools/accounting/getdelays.c
438
static void print_ioacct(struct taskstats *t)
tools/accounting/getdelays.c
441
t->ac_comm,
tools/accounting/getdelays.c
442
(unsigned long long)t->read_bytes,
tools/accounting/getdelays.c
443
(unsigned long long)t->write_bytes,
tools/accounting/getdelays.c
444
(unsigned long long)t->cancelled_write_bytes);
tools/accounting/procacct.c
201
#define average_ms(t, c) (t / 1000000ULL / (c ? c : 1))
tools/accounting/procacct.c
203
static void print_procacct(struct taskstats *t)
tools/accounting/procacct.c
208
, t->version >= 12 ? (t->ac_flag & AGROUP ? 'P' : 'T') : '?'
tools/accounting/procacct.c
209
, (unsigned long)t->ac_pid
tools/accounting/procacct.c
210
, (unsigned long)(t->version >= 12 ? t->ac_tgid : 0)
tools/accounting/procacct.c
211
, (unsigned long)t->ac_uid
tools/accounting/procacct.c
212
, (unsigned long long)t->ac_etime
tools/accounting/procacct.c
213
, (unsigned long long)(t->version >= 12 ? t->ac_tgetime : 0)
tools/accounting/procacct.c
214
, (unsigned long long)(t->ac_utime+t->ac_stime)
tools/accounting/procacct.c
215
, (unsigned long long)t->hiwater_vm
tools/accounting/procacct.c
216
, (unsigned long long)t->hiwater_rss
tools/accounting/procacct.c
217
, (unsigned long)(t->version >= 12 ? MAJOR(t->ac_exe_dev) : 0)
tools/accounting/procacct.c
218
, (unsigned long)(t->version >= 12 ? MINOR(t->ac_exe_dev) : 0)
tools/accounting/procacct.c
219
, (unsigned long long)(t->version >= 12 ? t->ac_exe_inode : 0)
tools/accounting/procacct.c
220
, t->ac_comm
tools/arch/arm64/include/asm/sysreg.h
1078
__emit_inst(0xd5200000|(\sreg)|(.L__gpr_num_\rt))
tools/arch/arm64/include/asm/sysreg.h
1082
__emit_inst(0xd5000000|(\sreg)|(.L__gpr_num_\rt))
tools/arch/x86/lib/insn.c
23
#define leXX_to_cpu(t, r) \
tools/arch/x86/lib/insn.c
25
__typeof__(t) v; \
tools/arch/x86/lib/insn.c
26
switch (sizeof(t)) { \
tools/arch/x86/lib/insn.c
37
#define validate_next(t, insn, n) \
tools/arch/x86/lib/insn.c
38
((insn)->next_byte + sizeof(t) + n <= (insn)->end_kaddr)
tools/arch/x86/lib/insn.c
40
#define __get_next(t, insn) \
tools/arch/x86/lib/insn.c
41
({ t r = get_unaligned((t *)(insn)->next_byte); (insn)->next_byte += sizeof(t); leXX_to_cpu(t, r); })
tools/arch/x86/lib/insn.c
43
#define __peek_nbyte_next(t, insn, n) \
tools/arch/x86/lib/insn.c
44
({ t r = get_unaligned((t *)(insn)->next_byte + n); leXX_to_cpu(t, r); })
tools/arch/x86/lib/insn.c
46
#define get_next(t, insn) \
tools/arch/x86/lib/insn.c
47
({ if (unlikely(!validate_next(t, insn, 0))) goto err_out; __get_next(t, insn); })
tools/arch/x86/lib/insn.c
49
#define peek_nbyte_next(t, insn, n) \
tools/arch/x86/lib/insn.c
50
({ if (unlikely(!validate_next(t, insn, n))) goto err_out; __peek_nbyte_next(t, insn, n); })
tools/arch/x86/lib/insn.c
52
#define peek_next(t, insn) peek_nbyte_next(t, insn, 0)
tools/bpf/bpftool/btf.c
120
const struct btf_type *t)
tools/bpf/bpftool/btf.c
123
int kind = btf_kind(t);
tools/bpf/bpftool/btf.c
129
jsonw_string_field(w, "name", btf_str(btf, t->name_off));
tools/bpf/bpftool/btf.c
132
btf_str(btf, t->name_off));
tools/bpf/bpftool/btf.c
137
__u32 v = *(__u32 *)(t + 1);
tools/bpf/bpftool/btf.c
143
jsonw_uint_field(w, "size", t->size);
tools/bpf/bpftool/btf.c
149
t->size, BTF_INT_OFFSET(v), BTF_INT_BITS(v),
tools/bpf/bpftool/btf.c
161
jsonw_uint_field(w, "type_id", t->type);
tools/bpf/bpftool/btf.c
163
printf(" type_id=%u", t->type);
tools/bpf/bpftool/btf.c
166
const struct btf_array *arr = (const void *)(t + 1);
tools/bpf/bpftool/btf.c
180
const struct btf_member *m = (const void *)(t + 1);
tools/bpf/bpftool/btf.c
181
__u16 vlen = BTF_INFO_VLEN(t->info);
tools/bpf/bpftool/btf.c
185
jsonw_uint_field(w, "size", t->size);
tools/bpf/bpftool/btf.c
190
printf(" size=%u vlen=%u", t->size, vlen);
tools/bpf/bpftool/btf.c
196
if (BTF_INFO_KFLAG(t->info)) {
tools/bpf/bpftool/btf.c
226
const struct btf_enum *v = (const void *)(t + 1);
tools/bpf/bpftool/btf.c
227
__u16 vlen = BTF_INFO_VLEN(t->info);
tools/bpf/bpftool/btf.c
231
encoding = btf_kflag(t) ? "SIGNED" : "UNSIGNED";
tools/bpf/bpftool/btf.c
234
jsonw_uint_field(w, "size", t->size);
tools/bpf/bpftool/btf.c
239
printf(" encoding=%s size=%u vlen=%u", encoding, t->size, vlen);
tools/bpf/bpftool/btf.c
247
if (btf_kflag(t))
tools/bpf/bpftool/btf.c
253
if (btf_kflag(t))
tools/bpf/bpftool/btf.c
264
const struct btf_enum64 *v = btf_enum64(t);
tools/bpf/bpftool/btf.c
265
__u16 vlen = btf_vlen(t);
tools/bpf/bpftool/btf.c
269
encoding = btf_kflag(t) ? "SIGNED" : "UNSIGNED";
tools/bpf/bpftool/btf.c
272
jsonw_uint_field(w, "size", t->size);
tools/bpf/bpftool/btf.c
277
printf(" encoding=%s size=%u vlen=%u", encoding, t->size, vlen);
tools/bpf/bpftool/btf.c
286
if (btf_kflag(t))
tools/bpf/bpftool/btf.c
292
if (btf_kflag(t))
tools/bpf/bpftool/btf.c
305
const char *fwd_kind = BTF_INFO_KFLAG(t->info) ? "union"
tools/bpf/bpftool/btf.c
315
const char *linkage = btf_func_linkage_str(t);
tools/bpf/bpftool/btf.c
318
jsonw_uint_field(w, "type_id", t->type);
tools/bpf/bpftool/btf.c
321
printf(" type_id=%u linkage=%s", t->type, linkage);
tools/bpf/bpftool/btf.c
326
const struct btf_param *p = (const void *)(t + 1);
tools/bpf/bpftool/btf.c
327
__u16 vlen = BTF_INFO_VLEN(t->info);
tools/bpf/bpftool/btf.c
331
jsonw_uint_field(w, "ret_type_id", t->type);
tools/bpf/bpftool/btf.c
336
printf(" ret_type_id=%u vlen=%u", t->type, vlen);
tools/bpf/bpftool/btf.c
355
const struct btf_var *v = (const void *)(t + 1);
tools/bpf/bpftool/btf.c
361
jsonw_uint_field(w, "type_id", t->type);
tools/bpf/bpftool/btf.c
364
printf(" type_id=%u, linkage=%s", t->type, linkage);
tools/bpf/bpftool/btf.c
369
const struct btf_var_secinfo *v = (const void *)(t + 1);
tools/bpf/bpftool/btf.c
371
__u16 vlen = BTF_INFO_VLEN(t->info);
tools/bpf/bpftool/btf.c
375
jsonw_uint_field(w, "size", t->size);
tools/bpf/bpftool/btf.c
380
printf(" size=%u vlen=%u", t->size, vlen);
tools/bpf/bpftool/btf.c
407
jsonw_uint_field(w, "size", t->size);
tools/bpf/bpftool/btf.c
409
printf(" size=%u", t->size);
tools/bpf/bpftool/btf.c
413
const struct btf_decl_tag *tag = (const void *)(t + 1);
tools/bpf/bpftool/btf.c
416
jsonw_uint_field(w, "type_id", t->type);
tools/bpf/bpftool/btf.c
419
printf(" type_id=%u component_idx=%d", t->type, tag->component_idx);
tools/bpf/bpftool/btf.c
438
const struct btf_type *t;
tools/bpf/bpftool/btf.c
449
t = btf__type_by_id(btf, root_type_ids[i]);
tools/bpf/bpftool/btf.c
450
dump_btf_type(btf, root_type_ids[i], t);
tools/bpf/bpftool/btf.c
462
t = btf__type_by_id(btf, i);
tools/bpf/bpftool/btf.c
463
dump_btf_type(btf, i, t);
tools/bpf/bpftool/btf.c
524
const struct btf_type *t = btf__type_by_id(btf, i);
tools/bpf/bpftool/btf.c
528
if (!btf_is_decl_tag(t))
tools/bpf/bpftool/btf.c
531
if (btf_decl_tag(t)->component_idx != -1)
tools/bpf/bpftool/btf.c
534
ft = btf__type_by_id(btf, t->type);
tools/bpf/bpftool/btf.c
538
name = btf__name_by_offset(btf, t->name_off);
tools/bpf/bpftool/btf.c
555
const struct btf_type *t = kfuncs.elems[i];
tools/bpf/bpftool/btf.c
561
if (fastcalls.elems[j] == t) {
tools/bpf/bpftool/btf.c
567
opts.field_name = btf__name_by_offset(btf, t->name_off);
tools/bpf/bpftool/btf.c
568
err = btf_dump__emit_type_decl(d, t->type, &opts);
tools/bpf/bpftool/btf.c
591
const struct btf_type *t = btf__type_by_id(btf, index);
tools/bpf/bpftool/btf.c
592
const int kind = btf_kind(t);
tools/bpf/bpftool/btf.c
595
if (t->name_off)
tools/bpf/bpftool/btf.c
612
return btf_type_rank(btf, btf_array(t)->type, has_name);
tools/bpf/bpftool/btf.c
622
return btf_type_rank(btf, t->type, has_name);
tools/bpf/bpftool/btf.c
631
const struct btf_type *t = btf__type_by_id(btf, index);
tools/bpf/bpftool/btf.c
633
switch (btf_kind(t)) {
tools/bpf/bpftool/btf.c
636
int name_off = t->name_off;
tools/bpf/bpftool/btf.c
638
if (!from_ref && !name_off && btf_vlen(t))
tools/bpf/bpftool/btf.c
639
name_off = btf_kind(t) == BTF_KIND_ENUM64 ?
tools/bpf/bpftool/btf.c
640
btf_enum64(t)->name_off :
tools/bpf/bpftool/btf.c
641
btf_enum(t)->name_off;
tools/bpf/bpftool/btf.c
646
return btf_type_sort_name(btf, btf_array(t)->type, true);
tools/bpf/bpftool/btf.c
654
return btf_type_sort_name(btf, t->type, true);
tools/bpf/bpftool/btf.c
656
return btf__name_by_offset(btf, t->name_off);
tools/bpf/bpftool/btf.c
676
const struct btf_type *t = btf__type_by_id(btf, id);
tools/bpf/bpftool/btf.c
680
hash = btf_name_hasher(hash, btf, t->name_off);
tools/bpf/bpftool/btf.c
682
switch (btf_kind(t)) {
tools/bpf/bpftool/btf.c
685
for (i = 0; i < btf_vlen(t); i++) {
tools/bpf/bpftool/btf.c
686
__u32 name_off = btf_is_enum(t) ?
tools/bpf/bpftool/btf.c
687
btf_enum(t)[i].name_off :
tools/bpf/bpftool/btf.c
688
btf_enum64(t)[i].name_off;
tools/bpf/bpftool/btf.c
697
for (i = 0; i < btf_vlen(t); i++) {
tools/bpf/bpftool/btf.c
698
const struct btf_member *m = btf_members(t) + i;
tools/bpf/bpftool/btf.c
712
hash = hasher(hash, btf_type_disambig_hash(btf, t->type, include_members));
tools/bpf/bpftool/btf.c
715
struct btf_array *arr = btf_array(t);
tools/bpf/bpftool/btf.c
757
const struct btf_type *t = btf__type_by_id(btf, i);
tools/bpf/bpftool/btf.c
762
d->own_name = btf__name_by_offset(btf, t->name_off);
tools/bpf/bpftool/btf.c
93
static const char *btf_func_linkage_str(const struct btf_type *t)
tools/bpf/bpftool/btf.c
95
switch (btf_vlen(t)) {
tools/bpf/bpftool/btf_dumper.c
108
const struct btf_type *t,
tools/bpf/bpftool/btf_dumper.c
118
ptr_type_id = btf__resolve_type(d->btf, t->type);
tools/bpf/bpftool/btf_dumper.c
148
const struct btf_type *t,
tools/bpf/bpftool/btf_dumper.c
151
const struct btf_enum *enums = btf_enum(t);
tools/bpf/bpftool/btf_dumper.c
155
switch (t->size) {
tools/bpf/bpftool/btf_dumper.c
172
for (i = 0; i < btf_vlen(t); i++) {
tools/bpf/bpftool/btf_dumper.c
186
const struct btf_type *t,
tools/bpf/bpftool/btf_dumper.c
189
const struct btf_enum64 *enums = btf_enum64(t);
tools/bpf/bpftool/btf_dumper.c
198
for (i = 0; i < btf_vlen(t); i++) {
tools/bpf/bpftool/btf_dumper.c
250
const struct btf_type *t = btf__type_by_id(d->btf, type_id);
tools/bpf/bpftool/btf_dumper.c
251
struct btf_array *arr = (struct btf_array *)(t + 1);
tools/bpf/bpftool/btf_dumper.c
395
static int btf_dumper_int(const struct btf_type *t, __u8 bit_offset,
tools/bpf/bpftool/btf_dumper.c
402
int_type = (__u32 *)(t + 1);
tools/bpf/bpftool/btf_dumper.c
468
const struct btf_type *t;
tools/bpf/bpftool/btf_dumper.c
475
t = btf__type_by_id(d->btf, type_id);
tools/bpf/bpftool/btf_dumper.c
476
if (!t)
tools/bpf/bpftool/btf_dumper.c
479
kind_flag = BTF_INFO_KFLAG(t->info);
tools/bpf/bpftool/btf_dumper.c
480
vlen = BTF_INFO_VLEN(t->info);
tools/bpf/bpftool/btf_dumper.c
482
m = (struct btf_member *)(t + 1);
tools/bpf/bpftool/btf_dumper.c
516
const struct btf_type *t = btf__type_by_id(d->btf, type_id);
tools/bpf/bpftool/btf_dumper.c
520
jsonw_name(d->jw, btf__name_by_offset(d->btf, t->name_off));
tools/bpf/bpftool/btf_dumper.c
521
ret = btf_dumper_do_type(d, t->type, bit_offset, data);
tools/bpf/bpftool/btf_dumper.c
531
const struct btf_type *t;
tools/bpf/bpftool/btf_dumper.c
534
t = btf__type_by_id(d->btf, type_id);
tools/bpf/bpftool/btf_dumper.c
535
if (!t)
tools/bpf/bpftool/btf_dumper.c
538
vlen = BTF_INFO_VLEN(t->info);
tools/bpf/bpftool/btf_dumper.c
539
vsi = (struct btf_var_secinfo *)(t + 1);
tools/bpf/bpftool/btf_dumper.c
542
jsonw_name(d->jw, btf__name_by_offset(d->btf, t->name_off));
tools/bpf/bpftool/btf_dumper.c
558
const struct btf_type *t = btf__type_by_id(d->btf, type_id);
tools/bpf/bpftool/btf_dumper.c
560
switch (BTF_INFO_KIND(t->info)) {
tools/bpf/bpftool/btf_dumper.c
562
return btf_dumper_int(t, bit_offset, data, d->jw,
tools/bpf/bpftool/btf_dumper.c
570
return btf_dumper_enum(d, t, data);
tools/bpf/bpftool/btf_dumper.c
572
return btf_dumper_enum64(d, t, data);
tools/bpf/bpftool/btf_dumper.c
574
btf_dumper_ptr(d, t, data);
tools/bpf/bpftool/btf_dumper.c
625
const struct btf_type *t;
tools/bpf/bpftool/btf_dumper.c
632
t = btf__type_by_id(btf, type_id);
tools/bpf/bpftool/btf_dumper.c
634
switch (BTF_INFO_KIND(t->info)) {
tools/bpf/bpftool/btf_dumper.c
638
BTF_PRINT_ARG("%s ", btf__name_by_offset(btf, t->name_off));
tools/bpf/bpftool/btf_dumper.c
642
btf__name_by_offset(btf, t->name_off));
tools/bpf/bpftool/btf_dumper.c
646
btf__name_by_offset(btf, t->name_off));
tools/bpf/bpftool/btf_dumper.c
651
btf__name_by_offset(btf, t->name_off));
tools/bpf/bpftool/btf_dumper.c
654
array = (struct btf_array *)(t + 1);
tools/bpf/bpftool/btf_dumper.c
659
BTF_PRINT_TYPE(t->type);
tools/bpf/bpftool/btf_dumper.c
664
BTF_INFO_KFLAG(t->info) ? "union" : "struct",
tools/bpf/bpftool/btf_dumper.c
665
btf__name_by_offset(btf, t->name_off));
tools/bpf/bpftool/btf_dumper.c
669
BTF_PRINT_TYPE(t->type);
tools/bpf/bpftool/btf_dumper.c
673
BTF_PRINT_TYPE(t->type);
tools/bpf/bpftool/btf_dumper.c
677
BTF_PRINT_TYPE(t->type);
tools/bpf/bpftool/btf_dumper.c
680
pos = btf_dump_func(btf, func_sig, t, NULL, pos, size);
tools/bpf/bpftool/btf_dumper.c
685
proto_type = btf__type_by_id(btf, t->type);
tools/bpf/bpftool/btf_dumper.c
686
pos = btf_dump_func(btf, func_sig, proto_type, t, pos, size);
tools/bpf/bpftool/btf_dumper.c
691
var = (struct btf_var *)(t + 1);
tools/bpf/bpftool/btf_dumper.c
694
BTF_PRINT_TYPE(t->type);
tools/bpf/bpftool/btf_dumper.c
696
btf__name_by_offset(btf, t->name_off));
tools/bpf/bpftool/btf_dumper.c
700
btf__name_by_offset(btf, t->name_off));
tools/bpf/bpftool/cgroup.c
162
const struct btf_type *t =
tools/bpf/bpftool/cgroup.c
165
btf__name_by_offset(btf_vmlinux, t->name_off);
tools/bpf/bpftool/common.c
1164
const char *bpf_attach_type_input_str(enum bpf_attach_type t)
tools/bpf/bpftool/common.c
1166
switch (t) {
tools/bpf/bpftool/common.c
1197
default: return libbpf_bpf_attach_type_str(t);
tools/bpf/bpftool/gen.c
2084
const struct btf_type *t = btf__type_by_id(info->marked_btf, type_id);
tools/bpf/bpftool/gen.c
2085
struct btf_member *m = btf_members(t) + idx;
tools/bpf/bpftool/gen.c
241
const struct btf_type *t = btf__type_by_id(btf, i);
tools/bpf/bpftool/gen.c
244
if (!btf_is_datasec(t))
tools/bpf/bpftool/gen.c
247
name = btf__str_by_offset(btf, t->name_off);
tools/bpf/bpftool/gen.c
252
return t;
tools/bpf/bpftool/gen.c
61
const struct btf_type *t;
tools/bpf/bpftool/gen.c
63
t = skip_mods_and_typedefs(btf, id, NULL);
tools/bpf/bpftool/gen.c
64
if (!btf_is_ptr(t))
tools/bpf/bpftool/gen.c
67
t = skip_mods_and_typedefs(btf, t->type, res_id);
tools/bpf/bpftool/gen.c
69
return btf_is_func_proto(t) ? t : NULL;
tools/bpf/bpftool/main.h
274
const char *bpf_attach_type_input_str(enum bpf_attach_type t);
tools/bpf/bpftool/net.c
218
struct tcmsg *t = NLMSG_DATA(nlh);
tools/bpf/bpftool/net.c
221
len = nlh->nlmsg_len - NLMSG_LENGTH(sizeof(*t));
tools/bpf/bpftool/net.c
222
attr = (struct nlattr *) ((void *) t + NLMSG_ALIGN(sizeof(*t)));
tools/bpf/bpftool/net.c
226
return dump_class_nlmsg(cookie, t, tb);
tools/bpf/bpftool/net.c
234
struct tcmsg t;
tools/bpf/bpftool/net.c
239
.t.tcm_family = AF_UNSPEC,
tools/bpf/bpftool/net.c
240
.t.tcm_ifindex = ifindex,
tools/bpf/bpftool/net.c
257
struct tcmsg *t = NLMSG_DATA(nlh);
tools/bpf/bpftool/net.c
260
len = nlh->nlmsg_len - NLMSG_LENGTH(sizeof(*t));
tools/bpf/bpftool/net.c
261
attr = (struct nlattr *) ((void *) t + NLMSG_ALIGN(sizeof(*t)));
tools/bpf/bpftool/net.c
265
return dump_qdisc_nlmsg(cookie, t, tb);
tools/bpf/bpftool/net.c
273
struct tcmsg t;
tools/bpf/bpftool/net.c
278
.t.tcm_family = AF_UNSPEC,
tools/bpf/bpftool/net.c
279
.t.tcm_ifindex = ifindex,
tools/bpf/bpftool/net.c
296
struct tcmsg *t = NLMSG_DATA(nlh);
tools/bpf/bpftool/net.c
299
len = nlh->nlmsg_len - NLMSG_LENGTH(sizeof(*t));
tools/bpf/bpftool/net.c
300
attr = (struct nlattr *) ((void *) t + NLMSG_ALIGN(sizeof(*t)));
tools/bpf/bpftool/net.c
304
return dump_filter_nlmsg(cookie, t, tb);
tools/bpf/bpftool/net.c
312
struct tcmsg t;
tools/bpf/bpftool/net.c
317
.t.tcm_family = AF_UNSPEC,
tools/bpf/bpftool/net.c
318
.t.tcm_ifindex = ifindex,
tools/bpf/bpftool/net.c
319
.t.tcm_parent = handle,
tools/bpf/bpftool/prog.c
2295
const struct btf_type *t;
tools/bpf/bpftool/prog.c
2335
t = btf__type_by_id(btf, func_info.type_id);
tools/bpf/bpftool/prog.c
2336
if (!t) {
tools/bpf/bpftool/prog.c
2341
name = strdup(btf__name_by_offset(btf, t->name_off));
tools/bpf/bpftool/struct_ops.c
44
const struct btf_type *t;
tools/bpf/bpftool/struct_ops.c
51
t = btf__type_by_id(kern_btf, info->btf_vmlinux_value_type_id);
tools/bpf/bpftool/struct_ops.c
52
st_ops_name = btf__name_by_offset(kern_btf, t->name_off);
tools/bpf/resolve_btfids/main.c
1006
t = btf__type_by_id(btf, ctx->decl_tags[i]);
tools/bpf/resolve_btfids/main.c
1007
if (btf_kflag(t) || btf_decl_tag(t)->component_idx != -1)
tools/bpf/resolve_btfids/main.c
1010
tag_name = btf__name_by_offset(btf, t->name_off);
tools/bpf/resolve_btfids/main.c
1014
func_id = t->type;
tools/bpf/resolve_btfids/main.c
1015
t = btf__type_by_id(btf, func_id);
tools/bpf/resolve_btfids/main.c
1016
if (!btf_is_func(t))
tools/bpf/resolve_btfids/main.c
1019
func_name = btf__name_by_offset(btf, t->name_off);
tools/bpf/resolve_btfids/main.c
1069
const struct btf_type *t;
tools/bpf/resolve_btfids/main.c
1072
t = btf_type_skip_qualifiers(btf, p->type);
tools/bpf/resolve_btfids/main.c
1073
if (!btf_is_ptr(t))
tools/bpf/resolve_btfids/main.c
1076
t = btf_type_skip_qualifiers(btf, t->type);
tools/bpf/resolve_btfids/main.c
1077
if (!btf_is_struct(t))
tools/bpf/resolve_btfids/main.c
1080
name = btf__name_by_offset(btf, t->name_off);
tools/bpf/resolve_btfids/main.c
1120
struct btf_type *t;
tools/bpf/resolve_btfids/main.c
1122
t = (struct btf_type *)btf__type_by_id(btf, kfunc->btf_id);
tools/bpf/resolve_btfids/main.c
1123
if (!t || !btf_is_func(t)) {
tools/bpf/resolve_btfids/main.c
1128
linkage = btf_vlen(t);
tools/bpf/resolve_btfids/main.c
1130
proto_id = t->type;
tools/bpf/resolve_btfids/main.c
1131
t = (struct btf_type *)btf__type_by_id(btf, proto_id);
tools/bpf/resolve_btfids/main.c
1132
if (!t || !btf_is_func_proto(t)) {
tools/bpf/resolve_btfids/main.c
1157
t = (struct btf_type *)btf__type_by_id(btf, ctx->decl_tags[i]);
tools/bpf/resolve_btfids/main.c
1158
if (t->type != kfunc->btf_id)
tools/bpf/resolve_btfids/main.c
1161
tag_name = btf__name_by_offset(btf, t->name_off);
tools/bpf/resolve_btfids/main.c
1165
idx = btf_decl_tag(t)->component_idx;
tools/bpf/resolve_btfids/main.c
1167
if (btf_kflag(t))
tools/bpf/resolve_btfids/main.c
1180
t = (struct btf_type *)btf__type_by_id(btf, proto_id);
tools/bpf/resolve_btfids/main.c
1181
new_proto_id = btf__add_func_proto(btf, t->type);
tools/bpf/resolve_btfids/main.c
1188
t = (struct btf_type *)btf__type_by_id(btf, proto_id);
tools/bpf/resolve_btfids/main.c
1189
nr_params = btf_vlen(t);
tools/bpf/resolve_btfids/main.c
1191
params = btf_params(t);
tools/bpf/resolve_btfids/main.c
1201
t = (struct btf_type *)btf__type_by_id(btf, proto_id);
tools/bpf/resolve_btfids/main.c
1205
t = (struct btf_type *)btf__type_by_id(btf, kfunc->btf_id);
tools/bpf/resolve_btfids/main.c
1206
t->type = new_proto_id;
tools/bpf/resolve_btfids/main.c
883
const struct btf_type *t = btf__type_by_id(btf, type_id);
tools/bpf/resolve_btfids/main.c
885
while (btf_is_mod(t))
tools/bpf/resolve_btfids/main.c
886
t = btf__type_by_id(btf, t->type);
tools/bpf/resolve_btfids/main.c
888
return t;
tools/bpf/resolve_btfids/main.c
933
const struct btf_type *t;
tools/bpf/resolve_btfids/main.c
937
t = btf__type_by_id(btf, id);
tools/bpf/resolve_btfids/main.c
938
if (!btf_is_decl_tag(t))
tools/bpf/resolve_btfids/main.c
996
const struct btf_type *t;
tools/firewire/decode-fcp.c
145
decode_avc(struct link_transaction *t)
tools/firewire/decode-fcp.c
148
(struct avc_frame *) t->request->packet.write_block.data;
tools/firewire/decode-fcp.c
174
decode_fcp(struct link_transaction *t)
tools/firewire/decode-fcp.c
177
(struct avc_frame *) t->request->packet.write_block.data;
tools/firewire/decode-fcp.c
179
((unsigned long long) t->request->packet.common.offset_high << 32) |
tools/firewire/decode-fcp.c
180
t->request->packet.common.offset_low;
tools/firewire/decode-fcp.c
182
if (t->request->packet.common.tcode != TCODE_WRITE_BLOCK_REQUEST)
tools/firewire/decode-fcp.c
188
decode_avc(t);
tools/firewire/nosy-dump.c
162
struct link_transaction *t;
tools/firewire/nosy-dump.c
164
list_for_each_entry(t, &pending_transaction_list, link) {
tools/firewire/nosy-dump.c
165
if (t->request_node == request_node &&
tools/firewire/nosy-dump.c
166
t->response_node == response_node &&
tools/firewire/nosy-dump.c
167
t->tlabel == tlabel)
tools/firewire/nosy-dump.c
168
return t;
tools/firewire/nosy-dump.c
171
t = malloc(sizeof *t);
tools/firewire/nosy-dump.c
172
if (!t)
tools/firewire/nosy-dump.c
174
t->request_node = request_node;
tools/firewire/nosy-dump.c
175
t->response_node = response_node;
tools/firewire/nosy-dump.c
176
t->tlabel = tlabel;
tools/firewire/nosy-dump.c
177
list_init(&t->request_list);
tools/firewire/nosy-dump.c
178
list_init(&t->response_list);
tools/firewire/nosy-dump.c
180
list_append(&pending_transaction_list, &t->link);
tools/firewire/nosy-dump.c
182
return t;
tools/firewire/nosy-dump.c
186
link_transaction_destroy(struct link_transaction *t)
tools/firewire/nosy-dump.c
190
while (!list_empty(&t->request_list)) {
tools/firewire/nosy-dump.c
191
sa = list_head(&t->request_list, struct subaction, link);
tools/firewire/nosy-dump.c
195
while (!list_empty(&t->response_list)) {
tools/firewire/nosy-dump.c
196
sa = list_head(&t->response_list, struct subaction, link);
tools/firewire/nosy-dump.c
200
free(t);
tools/firewire/nosy-dump.c
205
int (*decode)(struct link_transaction *t);
tools/firewire/nosy-dump.c
213
handle_transaction(struct link_transaction *t)
tools/firewire/nosy-dump.c
218
if (!t->request) {
tools/firewire/nosy-dump.c
224
if (protocol_decoders[i].decode(t))
tools/firewire/nosy-dump.c
230
decode_link_packet(&t->request->packet, t->request->length,
tools/firewire/nosy-dump.c
232
if (t->response)
tools/firewire/nosy-dump.c
233
decode_link_packet(&t->response->packet, t->request->length,
tools/firewire/nosy-dump.c
239
list_for_each_entry(sa, &t->request_list, link)
tools/firewire/nosy-dump.c
241
list_for_each_entry(sa, &t->response_list, link)
tools/firewire/nosy-dump.c
246
link_transaction_destroy(t);
tools/firewire/nosy-dump.c
252
struct link_transaction *t;
tools/firewire/nosy-dump.c
255
t = list_head(&pending_transaction_list,
tools/firewire/nosy-dump.c
257
list_remove(&t->link);
tools/firewire/nosy-dump.c
258
link_transaction_destroy(t);
tools/firewire/nosy-dump.c
486
struct link_transaction *t;
tools/firewire/nosy-dump.c
488
t = link_transaction_lookup(p->common.source, p->common.destination,
tools/firewire/nosy-dump.c
491
t->request = sa;
tools/firewire/nosy-dump.c
493
if (!list_empty(&t->request_list)) {
tools/firewire/nosy-dump.c
494
prev = list_tail(&t->request_list,
tools/firewire/nosy-dump.c
513
list_append(&t->request_list, &sa->link);
tools/firewire/nosy-dump.c
520
list_remove(&t->link);
tools/firewire/nosy-dump.c
521
handle_transaction(t);
tools/firewire/nosy-dump.c
527
list_remove(&t->link);
tools/firewire/nosy-dump.c
528
handle_transaction(t);
tools/firewire/nosy-dump.c
551
struct link_transaction *t;
tools/firewire/nosy-dump.c
553
t = link_transaction_lookup(p->common.destination, p->common.source,
tools/firewire/nosy-dump.c
555
if (list_empty(&t->request_list)) {
tools/firewire/nosy-dump.c
560
t->response = sa;
tools/firewire/nosy-dump.c
562
if (!list_empty(&t->response_list)) {
tools/firewire/nosy-dump.c
563
prev = list_tail(&t->response_list, struct subaction, link);
tools/firewire/nosy-dump.c
578
prev = list_tail(&t->request_list, struct subaction, link);
tools/firewire/nosy-dump.c
592
list_append(&t->response_list, &sa->link);
tools/firewire/nosy-dump.c
599
list_remove(&t->link);
tools/firewire/nosy-dump.c
600
handle_transaction(t);
tools/firewire/nosy-dump.h
172
int decode_fcp(struct link_transaction *t);
tools/include/linux/bits.h
46
#define GENMASK_TYPE(t, h, l) \
tools/include/linux/bits.h
47
((t)(GENMASK_INPUT_CHECK(h, l) + \
tools/include/linux/bits.h
48
(type_max(t) << (l) & \
tools/include/linux/bits.h
49
type_max(t) >> (BITS_PER_TYPE(t) - 1 - (h)))))
tools/include/nolibc/poll.h
27
struct __kernel_timespec t;
tools/include/nolibc/poll.h
30
t.tv_sec = timeout / 1000;
tools/include/nolibc/poll.h
31
t.tv_nsec = (timeout % 1000) * 1000000;
tools/include/nolibc/poll.h
33
return my_syscall5(__NR_ppoll_time64, fds, nfds, (timeout >= 0) ? &t : NULL, NULL, 0);
tools/include/nolibc/poll.h
35
struct __kernel_old_timespec t;
tools/include/nolibc/poll.h
38
t.tv_sec = timeout / 1000;
tools/include/nolibc/poll.h
39
t.tv_nsec = (timeout % 1000) * 1000000;
tools/include/nolibc/poll.h
41
return my_syscall5(__NR_ppoll, fds, nfds, (timeout >= 0) ? &t : NULL, NULL, 0);
tools/include/nolibc/sys/select.h
67
struct __kernel_timespec t;
tools/include/nolibc/sys/select.h
70
t.tv_sec = timeout->tv_sec;
tools/include/nolibc/sys/select.h
71
t.tv_nsec = (uint32_t)timeout->tv_usec * 1000;
tools/include/nolibc/sys/select.h
73
return my_syscall6(__NR_pselect6_time64, nfds, rfds, wfds, efds, timeout ? &t : NULL, NULL);
tools/include/nolibc/sys/select.h
75
struct __kernel_old_timespec t;
tools/include/nolibc/sys/select.h
78
t.tv_sec = timeout->tv_sec;
tools/include/nolibc/sys/select.h
79
t.tv_nsec = (uint32_t)timeout->tv_usec * 1000;
tools/include/nolibc/sys/select.h
81
return my_syscall6(__NR_pselect6, nfds, rfds, wfds, efds, timeout ? &t : NULL, NULL);
tools/include/nolibc/time.h
21
#define __nolibc_assert_time64_type(t) \
tools/include/nolibc/time.h
22
__nolibc_static_assert(sizeof(t) == 8)
tools/lib/argv_split.c
82
char *t;
tools/lib/argv_split.c
86
t = strndup(p, str-p);
tools/lib/argv_split.c
87
if (t == NULL)
tools/lib/argv_split.c
89
*argvp++ = t;
tools/lib/bpf/bpf_tracing.h
696
#define ___bpf_treg_cnt(t) \
tools/lib/bpf/bpf_tracing.h
697
__builtin_choose_expr(sizeof(t) == 1, 1, \
tools/lib/bpf/bpf_tracing.h
698
__builtin_choose_expr(sizeof(t) == 2, 1, \
tools/lib/bpf/bpf_tracing.h
699
__builtin_choose_expr(sizeof(t) == 4, 1, \
tools/lib/bpf/bpf_tracing.h
700
__builtin_choose_expr(sizeof(t) == 8, 1, \
tools/lib/bpf/bpf_tracing.h
701
__builtin_choose_expr(sizeof(t) == 16, 2, \
tools/lib/bpf/bpf_tracing.h
705
#define ___bpf_reg_cnt1(t, x) (___bpf_reg_cnt0() + ___bpf_treg_cnt(t))
tools/lib/bpf/bpf_tracing.h
706
#define ___bpf_reg_cnt2(t, x, args...) (___bpf_reg_cnt1(args) + ___bpf_treg_cnt(t))
tools/lib/bpf/bpf_tracing.h
707
#define ___bpf_reg_cnt3(t, x, args...) (___bpf_reg_cnt2(args) + ___bpf_treg_cnt(t))
tools/lib/bpf/bpf_tracing.h
708
#define ___bpf_reg_cnt4(t, x, args...) (___bpf_reg_cnt3(args) + ___bpf_treg_cnt(t))
tools/lib/bpf/bpf_tracing.h
709
#define ___bpf_reg_cnt5(t, x, args...) (___bpf_reg_cnt4(args) + ___bpf_treg_cnt(t))
tools/lib/bpf/bpf_tracing.h
710
#define ___bpf_reg_cnt6(t, x, args...) (___bpf_reg_cnt5(args) + ___bpf_treg_cnt(t))
tools/lib/bpf/bpf_tracing.h
711
#define ___bpf_reg_cnt7(t, x, args...) (___bpf_reg_cnt6(args) + ___bpf_treg_cnt(t))
tools/lib/bpf/bpf_tracing.h
712
#define ___bpf_reg_cnt8(t, x, args...) (___bpf_reg_cnt7(args) + ___bpf_treg_cnt(t))
tools/lib/bpf/bpf_tracing.h
713
#define ___bpf_reg_cnt9(t, x, args...) (___bpf_reg_cnt8(args) + ___bpf_treg_cnt(t))
tools/lib/bpf/bpf_tracing.h
714
#define ___bpf_reg_cnt10(t, x, args...) (___bpf_reg_cnt9(args) + ___bpf_treg_cnt(t))
tools/lib/bpf/bpf_tracing.h
715
#define ___bpf_reg_cnt11(t, x, args...) (___bpf_reg_cnt10(args) + ___bpf_treg_cnt(t))
tools/lib/bpf/bpf_tracing.h
716
#define ___bpf_reg_cnt12(t, x, args...) (___bpf_reg_cnt11(args) + ___bpf_treg_cnt(t))
tools/lib/bpf/bpf_tracing.h
719
#define ___bpf_union_arg(t, x, n) \
tools/lib/bpf/bpf_tracing.h
720
__builtin_choose_expr(sizeof(t) == 1, ({ union { __u8 z[1]; t x; } ___t = { .z = {ctx[n]}}; ___t.x; }), \
tools/lib/bpf/bpf_tracing.h
721
__builtin_choose_expr(sizeof(t) == 2, ({ union { __u16 z[1]; t x; } ___t = { .z = {ctx[n]} }; ___t.x; }), \
tools/lib/bpf/bpf_tracing.h
722
__builtin_choose_expr(sizeof(t) == 4, ({ union { __u32 z[1]; t x; } ___t = { .z = {ctx[n]} }; ___t.x; }), \
tools/lib/bpf/bpf_tracing.h
723
__builtin_choose_expr(sizeof(t) == 8, ({ union { __u64 z[1]; t x; } ___t = {.z = {ctx[n]} }; ___t.x; }), \
tools/lib/bpf/bpf_tracing.h
724
__builtin_choose_expr(sizeof(t) == 16, ({ union { __u64 z[2]; t x; } ___t = {.z = {ctx[n], ctx[n + 1]} }; ___t.x; }), \
tools/lib/bpf/bpf_tracing.h
728
#define ___bpf_ctx_arg1(n, t, x) , ___bpf_union_arg(t, x, n - ___bpf_reg_cnt1(t, x))
tools/lib/bpf/bpf_tracing.h
729
#define ___bpf_ctx_arg2(n, t, x, args...) , ___bpf_union_arg(t, x, n - ___bpf_reg_cnt2(t, x, args)) ___bpf_ctx_arg1(n, args)
tools/lib/bpf/bpf_tracing.h
730
#define ___bpf_ctx_arg3(n, t, x, args...) , ___bpf_union_arg(t, x, n - ___bpf_reg_cnt3(t, x, args)) ___bpf_ctx_arg2(n, args)
tools/lib/bpf/bpf_tracing.h
731
#define ___bpf_ctx_arg4(n, t, x, args...) , ___bpf_union_arg(t, x, n - ___bpf_reg_cnt4(t, x, args)) ___bpf_ctx_arg3(n, args)
tools/lib/bpf/bpf_tracing.h
732
#define ___bpf_ctx_arg5(n, t, x, args...) , ___bpf_union_arg(t, x, n - ___bpf_reg_cnt5(t, x, args)) ___bpf_ctx_arg4(n, args)
tools/lib/bpf/bpf_tracing.h
733
#define ___bpf_ctx_arg6(n, t, x, args...) , ___bpf_union_arg(t, x, n - ___bpf_reg_cnt6(t, x, args)) ___bpf_ctx_arg5(n, args)
tools/lib/bpf/bpf_tracing.h
734
#define ___bpf_ctx_arg7(n, t, x, args...) , ___bpf_union_arg(t, x, n - ___bpf_reg_cnt7(t, x, args)) ___bpf_ctx_arg6(n, args)
tools/lib/bpf/bpf_tracing.h
735
#define ___bpf_ctx_arg8(n, t, x, args...) , ___bpf_union_arg(t, x, n - ___bpf_reg_cnt8(t, x, args)) ___bpf_ctx_arg7(n, args)
tools/lib/bpf/bpf_tracing.h
736
#define ___bpf_ctx_arg9(n, t, x, args...) , ___bpf_union_arg(t, x, n - ___bpf_reg_cnt9(t, x, args)) ___bpf_ctx_arg8(n, args)
tools/lib/bpf/bpf_tracing.h
737
#define ___bpf_ctx_arg10(n, t, x, args...) , ___bpf_union_arg(t, x, n - ___bpf_reg_cnt10(t, x, args)) ___bpf_ctx_arg9(n, args)
tools/lib/bpf/bpf_tracing.h
738
#define ___bpf_ctx_arg11(n, t, x, args...) , ___bpf_union_arg(t, x, n - ___bpf_reg_cnt11(t, x, args)) ___bpf_ctx_arg10(n, args)
tools/lib/bpf/bpf_tracing.h
739
#define ___bpf_ctx_arg12(n, t, x, args...) , ___bpf_union_arg(t, x, n - ___bpf_reg_cnt12(t, x, args)) ___bpf_ctx_arg11(n, args)
tools/lib/bpf/bpf_tracing.h
743
#define ___bpf_ctx_decl1(t, x) , t x
tools/lib/bpf/bpf_tracing.h
744
#define ___bpf_ctx_decl2(t, x, args...) , t x ___bpf_ctx_decl1(args)
tools/lib/bpf/bpf_tracing.h
745
#define ___bpf_ctx_decl3(t, x, args...) , t x ___bpf_ctx_decl2(args)
tools/lib/bpf/bpf_tracing.h
746
#define ___bpf_ctx_decl4(t, x, args...) , t x ___bpf_ctx_decl3(args)
tools/lib/bpf/bpf_tracing.h
747
#define ___bpf_ctx_decl5(t, x, args...) , t x ___bpf_ctx_decl4(args)
tools/lib/bpf/bpf_tracing.h
748
#define ___bpf_ctx_decl6(t, x, args...) , t x ___bpf_ctx_decl5(args)
tools/lib/bpf/bpf_tracing.h
749
#define ___bpf_ctx_decl7(t, x, args...) , t x ___bpf_ctx_decl6(args)
tools/lib/bpf/bpf_tracing.h
750
#define ___bpf_ctx_decl8(t, x, args...) , t x ___bpf_ctx_decl7(args)
tools/lib/bpf/bpf_tracing.h
751
#define ___bpf_ctx_decl9(t, x, args...) , t x ___bpf_ctx_decl8(args)
tools/lib/bpf/bpf_tracing.h
752
#define ___bpf_ctx_decl10(t, x, args...) , t x ___bpf_ctx_decl9(args)
tools/lib/bpf/bpf_tracing.h
753
#define ___bpf_ctx_decl11(t, x, args...) , t x ___bpf_ctx_decl10(args)
tools/lib/bpf/bpf_tracing.h
754
#define ___bpf_ctx_decl12(t, x, args...) , t x ___bpf_ctx_decl11(args)
tools/lib/bpf/btf.c
1605
struct btf_type *t;
tools/lib/bpf/btf.c
1630
t = p + btf->type_offs[i];
tools/lib/bpf/btf.c
1635
if (btf_bswap_type_rest(t))
tools/lib/bpf/btf.c
1637
btf_bswap_type_base(t);
tools/lib/bpf/btf.c
1903
static void btf_type_inc_vlen(struct btf_type *t)
tools/lib/bpf/btf.c
1905
t->info = btf_type_info(btf_kind(t), btf_vlen(t) + 1, btf_kflag(t));
tools/lib/bpf/btf.c
1962
struct btf_type *t;
tools/lib/bpf/btf.c
1974
t = btf_add_type_mem(p->dst, sz);
tools/lib/bpf/btf.c
1975
if (!t)
tools/lib/bpf/btf.c
1978
memcpy(t, src_type, sz);
tools/lib/bpf/btf.c
1980
err = btf_field_iter_init(&it, t, BTF_FIELD_ITER_STRS);
tools/lib/bpf/btf.c
2008
void *t;
tools/lib/bpf/btf.c
2027
t = btf_add_type_mem(btf, data_sz);
tools/lib/bpf/btf.c
2028
if (!t)
tools/lib/bpf/btf.c
2042
memcpy(t, src_btf->types_data, data_sz);
tools/lib/bpf/btf.c
2048
sz = btf_type_size(t);
tools/lib/bpf/btf.c
2056
*off = t - btf->types_data;
tools/lib/bpf/btf.c
2059
err = btf_field_iter_init(&it, t, BTF_FIELD_ITER_STRS);
tools/lib/bpf/btf.c
2069
err = btf_field_iter_init(&it, t, BTF_FIELD_ITER_IDS);
tools/lib/bpf/btf.c
2085
t += sz;
tools/lib/bpf/btf.c
2133
struct btf_type *t;
tools/lib/bpf/btf.c
2150
t = btf_add_type_mem(btf, sz);
tools/lib/bpf/btf.c
2151
if (!t)
tools/lib/bpf/btf.c
2162
t->name_off = name_off;
tools/lib/bpf/btf.c
2163
t->info = btf_type_info(BTF_KIND_INT, 0, 0);
tools/lib/bpf/btf.c
2164
t->size = byte_sz;
tools/lib/bpf/btf.c
2166
*(__u32 *)(t + 1) = (encoding << 24) | (byte_sz * 8);
tools/lib/bpf/btf.c
2181
struct btf_type *t;
tools/lib/bpf/btf.c
2197
t = btf_add_type_mem(btf, sz);
tools/lib/bpf/btf.c
2198
if (!t)
tools/lib/bpf/btf.c
2205
t->name_off = name_off;
tools/lib/bpf/btf.c
2206
t->info = btf_type_info(BTF_KIND_FLOAT, 0, 0);
tools/lib/bpf/btf.c
2207
t->size = byte_sz;
tools/lib/bpf/btf.c
2226
struct btf_type *t;
tools/lib/bpf/btf.c
2236
t = btf_add_type_mem(btf, sz);
tools/lib/bpf/btf.c
2237
if (!t)
tools/lib/bpf/btf.c
2246
t->name_off = name_off;
tools/lib/bpf/btf.c
2247
t->info = btf_type_info(kind, 0, kflag);
tools/lib/bpf/btf.c
2248
t->type = ref_type_id;
tools/lib/bpf/btf.c
2276
struct btf_type *t;
tools/lib/bpf/btf.c
2287
t = btf_add_type_mem(btf, sz);
tools/lib/bpf/btf.c
2288
if (!t)
tools/lib/bpf/btf.c
2291
t->name_off = 0;
tools/lib/bpf/btf.c
2292
t->info = btf_type_info(BTF_KIND_ARRAY, 0, 0);
tools/lib/bpf/btf.c
2293
t->size = 0;
tools/lib/bpf/btf.c
2295
a = btf_array(t);
tools/lib/bpf/btf.c
2306
struct btf_type *t;
tools/lib/bpf/btf.c
2313
t = btf_add_type_mem(btf, sz);
tools/lib/bpf/btf.c
2314
if (!t)
tools/lib/bpf/btf.c
2326
t->name_off = name_off;
tools/lib/bpf/btf.c
2327
t->info = btf_type_info(kind, 0, 0);
tools/lib/bpf/btf.c
2328
t->size = bytes_sz;
tools/lib/bpf/btf.c
2386
struct btf_type *t;
tools/lib/bpf/btf.c
2394
t = btf_last_type(btf);
tools/lib/bpf/btf.c
2395
if (!btf_is_composite(t))
tools/lib/bpf/btf.c
2406
if (btf_is_union(t) && bit_offset)
tools/lib/bpf/btf.c
2429
t = btf_last_type(btf);
tools/lib/bpf/btf.c
2431
t->info = btf_type_info(btf_kind(t), btf_vlen(t) + 1, is_bitfield || btf_kflag(t));
tools/lib/bpf/btf.c
2441
struct btf_type *t;
tools/lib/bpf/btf.c
2452
t = btf_add_type_mem(btf, sz);
tools/lib/bpf/btf.c
2453
if (!t)
tools/lib/bpf/btf.c
2463
t->name_off = name_off;
tools/lib/bpf/btf.c
2464
t->info = btf_type_info(kind, 0, is_signed);
tools/lib/bpf/btf.c
2465
t->size = byte_sz;
tools/lib/bpf/btf.c
2502
struct btf_type *t;
tools/lib/bpf/btf.c
2509
t = btf_last_type(btf);
tools/lib/bpf/btf.c
2510
if (!btf_is_enum(t))
tools/lib/bpf/btf.c
2536
t = btf_last_type(btf);
tools/lib/bpf/btf.c
2537
btf_type_inc_vlen(t);
tools/lib/bpf/btf.c
2541
t->info = btf_type_info(btf_kind(t), btf_vlen(t), true);
tools/lib/bpf/btf.c
2580
struct btf_type *t;
tools/lib/bpf/btf.c
2586
t = btf_last_type(btf);
tools/lib/bpf/btf.c
2587
if (!btf_is_enum64(t))
tools/lib/bpf/btf.c
2612
t = btf_last_type(btf);
tools/lib/bpf/btf.c
2613
btf_type_inc_vlen(t);
tools/lib/bpf/btf.c
2637
struct btf_type *t;
tools/lib/bpf/btf.c
2643
t = btf_type_by_id(btf, id);
tools/lib/bpf/btf.c
2644
t->info = btf_type_info(BTF_KIND_FWD, 0, fwd_kind == BTF_FWD_UNION);
tools/lib/bpf/btf.c
2763
struct btf_type *t = btf_type_by_id(btf, id);
tools/lib/bpf/btf.c
2765
t->info = btf_type_info(BTF_KIND_FUNC, linkage, 0);
tools/lib/bpf/btf.c
2784
struct btf_type *t;
tools/lib/bpf/btf.c
2794
t = btf_add_type_mem(btf, sz);
tools/lib/bpf/btf.c
2795
if (!t)
tools/lib/bpf/btf.c
2801
t->name_off = 0;
tools/lib/bpf/btf.c
2802
t->info = btf_type_info(BTF_KIND_FUNC_PROTO, 0, 0);
tools/lib/bpf/btf.c
2803
t->type = ret_type_id;
tools/lib/bpf/btf.c
2818
struct btf_type *t;
tools/lib/bpf/btf.c
2828
t = btf_last_type(btf);
tools/lib/bpf/btf.c
2829
if (!btf_is_func_proto(t))
tools/lib/bpf/btf.c
2851
t = btf_last_type(btf);
tools/lib/bpf/btf.c
2852
btf_type_inc_vlen(t);
tools/lib/bpf/btf.c
2871
struct btf_type *t;
tools/lib/bpf/btf.c
2889
t = btf_add_type_mem(btf, sz);
tools/lib/bpf/btf.c
2890
if (!t)
tools/lib/bpf/btf.c
2897
t->name_off = name_off;
tools/lib/bpf/btf.c
2898
t->info = btf_type_info(BTF_KIND_VAR, 0, 0);
tools/lib/bpf/btf.c
2899
t->type = type_id;
tools/lib/bpf/btf.c
2901
v = btf_var(t);
tools/lib/bpf/btf.c
2921
struct btf_type *t;
tools/lib/bpf/btf.c
2932
t = btf_add_type_mem(btf, sz);
tools/lib/bpf/btf.c
2933
if (!t)
tools/lib/bpf/btf.c
2941
t->name_off = name_off;
tools/lib/bpf/btf.c
2942
t->info = btf_type_info(BTF_KIND_DATASEC, 0, 0);
tools/lib/bpf/btf.c
2943
t->size = byte_sz;
tools/lib/bpf/btf.c
2960
struct btf_type *t;
tools/lib/bpf/btf.c
2967
t = btf_last_type(btf);
tools/lib/bpf/btf.c
2968
if (!btf_is_datasec(t))
tools/lib/bpf/btf.c
297
static int btf_type_size(const struct btf_type *t)
tools/lib/bpf/btf.c
2988
t = btf_last_type(btf);
tools/lib/bpf/btf.c
2989
btf_type_inc_vlen(t);
tools/lib/bpf/btf.c
2999
struct btf_type *t;
tools/lib/bpf/btf.c
300
__u16 vlen = btf_vlen(t);
tools/lib/bpf/btf.c
3012
t = btf_add_type_mem(btf, sz);
tools/lib/bpf/btf.c
3013
if (!t)
tools/lib/bpf/btf.c
302
switch (btf_kind(t)) {
tools/lib/bpf/btf.c
3020
t->name_off = value_off;
tools/lib/bpf/btf.c
3021
t->info = btf_type_info(BTF_KIND_DECL_TAG, 0, kflag);
tools/lib/bpf/btf.c
3022
t->type = ref_type_id;
tools/lib/bpf/btf.c
3023
btf_decl_tag(t)->component_idx = component_idx;
tools/lib/bpf/btf.c
333
pr_debug("Unsupported BTF_KIND:%u\n", btf_kind(t));
tools/lib/bpf/btf.c
338
static void btf_bswap_type_base(struct btf_type *t)
tools/lib/bpf/btf.c
340
t->name_off = bswap_32(t->name_off);
tools/lib/bpf/btf.c
341
t->info = bswap_32(t->info);
tools/lib/bpf/btf.c
342
t->type = bswap_32(t->type);
tools/lib/bpf/btf.c
345
static int btf_bswap_type_rest(struct btf_type *t)
tools/lib/bpf/btf.c
353
__u16 vlen = btf_vlen(t);
tools/lib/bpf/btf.c
356
switch (btf_kind(t)) {
tools/lib/bpf/btf.c
368
*(__u32 *)(t + 1) = bswap_32(*(__u32 *)(t + 1));
tools/lib/bpf/btf.c
371
for (i = 0, e = btf_enum(t); i < vlen; i++, e++) {
tools/lib/bpf/btf.c
377
for (i = 0, e64 = btf_enum64(t); i < vlen; i++, e64++) {
tools/lib/bpf/btf.c
3826
struct btf_type *t = btf_type_by_id(d->btf, i);
tools/lib/bpf/btf.c
3829
if (btf_is_var(t) || btf_is_datasec(t))
tools/lib/bpf/btf.c
384
a = btf_array(t);
tools/lib/bpf/btf.c
3862
struct btf_type *t = btf_type_by_id(d->btf, d->btf->start_id + i);
tools/lib/bpf/btf.c
3865
r = btf_field_iter_init(&it, t, BTF_FIELD_ITER_STRS);
tools/lib/bpf/btf.c
391
for (i = 0, m = btf_members(t); i < vlen; i++, m++) {
tools/lib/bpf/btf.c
3974
static long btf_hash_typedef(struct btf_type *t)
tools/lib/bpf/btf.c
3978
h = hash_combine(0, t->name_off);
tools/lib/bpf/btf.c
3979
h = hash_combine(h, t->info);
tools/lib/bpf/btf.c
398
for (i = 0, p = btf_params(t); i < vlen; i++, p++) {
tools/lib/bpf/btf.c
3983
static long btf_hash_common(struct btf_type *t)
tools/lib/bpf/btf.c
3987
h = hash_combine(0, t->name_off);
tools/lib/bpf/btf.c
3988
h = hash_combine(h, t->info);
tools/lib/bpf/btf.c
3989
h = hash_combine(h, t->size);
tools/lib/bpf/btf.c
4008
static long btf_hash_int_decl_tag(struct btf_type *t)
tools/lib/bpf/btf.c
4010
__u32 info = *(__u32 *)(t + 1);
tools/lib/bpf/btf.c
4013
h = btf_hash_common(t);
tools/lib/bpf/btf.c
4031
static long btf_hash_enum(struct btf_type *t)
tools/lib/bpf/btf.c
4036
h = hash_combine(0, t->name_off);
tools/lib/bpf/btf.c
404
btf_var(t)->linkage = bswap_32(btf_var(t)->linkage);
tools/lib/bpf/btf.c
407
for (i = 0, v = btf_var_secinfos(t); i < vlen; i++, v++) {
tools/lib/bpf/btf.c
4090
static inline bool btf_is_enum_fwd(struct btf_type *t)
tools/lib/bpf/btf.c
4092
return btf_is_any_enum(t) && btf_vlen(t) == 0;
tools/lib/bpf/btf.c
4114
static long btf_hash_struct(struct btf_type *t)
tools/lib/bpf/btf.c
4116
const struct btf_member *member = btf_members(t);
tools/lib/bpf/btf.c
4117
__u32 vlen = btf_vlen(t);
tools/lib/bpf/btf.c
4118
long h = btf_hash_common(t);
tools/lib/bpf/btf.c
414
btf_decl_tag(t)->component_idx = bswap_32(btf_decl_tag(t)->component_idx);
tools/lib/bpf/btf.c
4161
static long btf_hash_array(struct btf_type *t)
tools/lib/bpf/btf.c
4163
const struct btf_array *info = btf_array(t);
tools/lib/bpf/btf.c
4164
long h = btf_hash_common(t);
tools/lib/bpf/btf.c
417
pr_debug("Unsupported BTF_KIND:%u\n", btf_kind(t));
tools/lib/bpf/btf.c
4211
static long btf_hash_fnproto(struct btf_type *t)
tools/lib/bpf/btf.c
4213
const struct btf_param *member = btf_params(t);
tools/lib/bpf/btf.c
4214
__u16 vlen = btf_vlen(t);
tools/lib/bpf/btf.c
4215
long h = btf_hash_common(t);
tools/lib/bpf/btf.c
4287
struct btf_type *t;
tools/lib/bpf/btf.c
4295
t = btf_type_by_id(d->btf, type_id);
tools/lib/bpf/btf.c
4300
switch (btf_kind(t)) {
tools/lib/bpf/btf.c
4314
h = btf_hash_common(t);
tools/lib/bpf/btf.c
4318
h = btf_hash_int_decl_tag(t);
tools/lib/bpf/btf.c
4322
h = btf_hash_enum(t);
tools/lib/bpf/btf.c
4326
h = btf_hash_struct(t);
tools/lib/bpf/btf.c
4329
h = btf_hash_array(t);
tools/lib/bpf/btf.c
4332
h = btf_hash_fnproto(t);
tools/lib/bpf/btf.c
4335
pr_debug("unknown kind %d for type [%d]\n", btf_kind(t), type_id);
tools/lib/bpf/btf.c
4353
struct btf_type *t = btf_type_by_id(d->btf, type_id);
tools/lib/bpf/btf.c
4361
switch (btf_kind(t)) {
tools/lib/bpf/btf.c
4379
h = btf_hash_int_decl_tag(t);
tools/lib/bpf/btf.c
4383
if (btf_equal_int_tag(t, cand)) {
tools/lib/bpf/btf.c
4392
h = btf_hash_enum(t);
tools/lib/bpf/btf.c
4396
if (btf_equal_enum(t, cand)) {
tools/lib/bpf/btf.c
4400
if (btf_compat_enum(t, cand)) {
tools/lib/bpf/btf.c
4401
if (btf_is_enum_fwd(t)) {
tools/lib/bpf/btf.c
4414
h = btf_hash_common(t);
tools/lib/bpf/btf.c
4418
if (btf_equal_common(t, cand)) {
tools/lib/bpf/btf.c
4489
static inline __u16 btf_fwd_kind(struct btf_type *t)
tools/lib/bpf/btf.c
4491
return btf_kflag(t) ? BTF_KIND_UNION : BTF_KIND_STRUCT;
tools/lib/bpf/btf.c
475
const struct btf_type *t;
tools/lib/bpf/btf.c
477
t = btf__type_by_id(btf, id);
tools/lib/bpf/btf.c
478
if (!t) {
tools/lib/bpf/btf.c
486
static int btf_validate_type(const struct btf *btf, const struct btf_type *t, __u32 id)
tools/lib/bpf/btf.c
488
__u32 kind = btf_kind(t);
tools/lib/bpf/btf.c
491
err = btf_validate_str(btf, t->name_off, "type name", id);
tools/lib/bpf/btf.c
4952
static inline long btf_hash_by_kind(struct btf_type *t, __u16 kind)
tools/lib/bpf/btf.c
4955
return btf_hash_typedef(t);
tools/lib/bpf/btf.c
4957
return btf_hash_struct(t);
tools/lib/bpf/btf.c
4993
struct btf_type *cand_type, *t;
tools/lib/bpf/btf.c
5004
t = btf_type_by_id(d->btf, type_id);
tools/lib/bpf/btf.c
5005
kind = btf_kind(t);
tools/lib/bpf/btf.c
5012
h = btf_hash_by_kind(t, kind);
tools/lib/bpf/btf.c
5028
if (!btf_equal_by_kind(t, cand_type, kind))
tools/lib/bpf/btf.c
509
err = btf_validate_id(btf, t->type, id);
tools/lib/bpf/btf.c
5091
struct btf_type *t, *cand;
tools/lib/bpf/btf.c
5101
t = btf_type_by_id(d->btf, type_id);
tools/lib/bpf/btf.c
5104
switch (btf_kind(t)) {
tools/lib/bpf/btf.c
5111
ref_type_id = btf_dedup_ref_type(d, t->type);
tools/lib/bpf/btf.c
5114
t->type = ref_type_id;
tools/lib/bpf/btf.c
5116
h = btf_hash_common(t);
tools/lib/bpf/btf.c
5120
if (btf_equal_common(t, cand)) {
tools/lib/bpf/btf.c
5128
ref_type_id = btf_dedup_ref_type(d, t->type);
tools/lib/bpf/btf.c
5131
t->type = ref_type_id;
tools/lib/bpf/btf.c
5133
h = btf_hash_int_decl_tag(t);
tools/lib/bpf/btf.c
5137
if (btf_equal_int_tag(t, cand)) {
tools/lib/bpf/btf.c
514
const struct btf_array *a = btf_array(t);
tools/lib/bpf/btf.c
5145
struct btf_array *info = btf_array(t);
tools/lib/bpf/btf.c
5157
h = btf_hash_array(t);
tools/lib/bpf/btf.c
5161
if (btf_equal_array(t, cand)) {
tools/lib/bpf/btf.c
5174
ref_type_id = btf_dedup_ref_type(d, t->type);
tools/lib/bpf/btf.c
5177
t->type = ref_type_id;
tools/lib/bpf/btf.c
5179
vlen = btf_vlen(t);
tools/lib/bpf/btf.c
5180
param = btf_params(t);
tools/lib/bpf/btf.c
5189
h = btf_hash_fnproto(t);
tools/lib/bpf/btf.c
5193
if (btf_equal_fnproto(t, cand)) {
tools/lib/bpf/btf.c
5235
struct btf_type *t;
tools/lib/bpf/btf.c
524
const struct btf_member *m = btf_members(t);
tools/lib/bpf/btf.c
5245
t = btf_type_by_id(d->btf, type_id);
tools/lib/bpf/btf.c
5246
kind = btf_kind(t);
tools/lib/bpf/btf.c
5255
err = hashmap__add(names_map, t->name_off, type_id);
tools/lib/bpf/btf.c
5257
err = hashmap__set(names_map, t->name_off, 0, NULL, NULL);
tools/lib/bpf/btf.c
526
n = btf_vlen(t);
tools/lib/bpf/btf.c
5268
struct btf_type *t = btf_type_by_id(d->btf, type_id);
tools/lib/bpf/btf.c
5269
enum btf_fwd_kind fwd_kind = btf_kflag(t);
tools/lib/bpf/btf.c
5270
__u16 cand_kind, kind = btf_kind(t);
tools/lib/bpf/btf.c
5281
if (!hashmap__find(names_map, t->name_off, &cand_id))
tools/lib/bpf/btf.c
536
const struct btf_enum *m = btf_enum(t);
tools/lib/bpf/btf.c
5370
const struct btf_type *t;
tools/lib/bpf/btf.c
538
n = btf_vlen(t);
tools/lib/bpf/btf.c
5388
t = btf__type_by_id(d->btf, id);
tools/lib/bpf/btf.c
5389
len = btf_type_size(t);
tools/lib/bpf/btf.c
5393
memmove(p, t, len);
tools/lib/bpf/btf.c
5449
struct btf_type *t = btf_type_by_id(d->btf, d->btf->start_id + i);
tools/lib/bpf/btf.c
5453
r = btf_field_iter_init(&it, t, BTF_FIELD_ITER_IDS);
tools/lib/bpf/btf.c
547
const struct btf_enum64 *m = btf_enum64(t);
tools/lib/bpf/btf.c
549
n = btf_vlen(t);
tools/lib/bpf/btf.c
560
err = btf_validate_id(btf, t->type, id);
tools/lib/bpf/btf.c
563
ft = btf__type_by_id(btf, t->type);
tools/lib/bpf/btf.c
565
pr_warn("btf: type [%u]: referenced type [%u] is not FUNC_PROTO\n", id, t->type);
tools/lib/bpf/btf.c
5708
struct btf_type *t;
tools/lib/bpf/btf.c
571
const struct btf_param *m = btf_params(t);
tools/lib/bpf/btf.c
5721
t = btf_type_by_id(dist->pipe.src, i);
tools/lib/bpf/btf.c
5722
kind = btf_kind(t);
tools/lib/bpf/btf.c
5723
name = btf__name_by_offset(dist->pipe.src, t->name_off);
tools/lib/bpf/btf.c
573
n = btf_vlen(t);
tools/lib/bpf/btf.c
5732
err = btf_add_type(&dist->pipe, t);
tools/lib/bpf/btf.c
5741
if (!t->name_off)
tools/lib/bpf/btf.c
5743
err = btf_add_composite(dist->pipe.dst, kind, name, t->size);
tools/lib/bpf/btf.c
5745
if (t->name_off)
tools/lib/bpf/btf.c
5747
err = btf_add_type(&dist->pipe, t);
tools/lib/bpf/btf.c
5759
if (!t->name_off)
tools/lib/bpf/btf.c
5761
err = btf__add_enum(dist->pipe.dst, name, t->size);
tools/lib/bpf/btf.c
5763
if (t->name_off)
tools/lib/bpf/btf.c
5765
err = btf_add_type(&dist->pipe, t);
tools/lib/bpf/btf.c
5779
err = btf_add_type(&dist->pipe, t);
tools/lib/bpf/btf.c
5801
struct btf_type *t = btf_type_by_id(dist->pipe.dst, i);
tools/lib/bpf/btf.c
5806
err = btf_field_iter_init(&it, t, BTF_FIELD_ITER_IDS);
tools/lib/bpf/btf.c
583
const struct btf_var_secinfo *m = btf_var_secinfos(t);
tools/lib/bpf/btf.c
5841
struct btf_type *t;
tools/lib/bpf/btf.c
585
n = btf_vlen(t);
tools/lib/bpf/btf.c
5899
t = btf_type_by_id(src_btf, i);
tools/lib/bpf/btf.c
5900
err = btf_add_type(&dist.pipe, t);
tools/lib/bpf/btf.c
6043
const struct btf_type *t;
tools/lib/bpf/btf.c
6048
t = btf__type_by_id(btf, id);
tools/lib/bpf/btf.c
6049
type_size = btf_type_size(t);
tools/lib/bpf/btf.c
6050
memcpy(nt, t, type_size);
tools/lib/bpf/btf.c
606
const struct btf_type *t;
tools/lib/bpf/btf.c
611
t = btf_type_by_id(btf, i);
tools/lib/bpf/btf.c
612
err = btf_validate_type(btf, t, i);
tools/lib/bpf/btf.c
661
const struct btf_type *t;
tools/lib/bpf/btf.c
670
t = btf__type_by_id(btf, i);
tools/lib/bpf/btf.c
671
if (!btf_is_int(t))
tools/lib/bpf/btf.c
674
if (t->size != 4 && t->size != 8)
tools/lib/bpf/btf.c
677
name = btf__name_by_offset(btf, t->name_off);
tools/lib/bpf/btf.c
683
return t->size;
tools/lib/bpf/btf.c
760
static bool btf_type_is_void(const struct btf_type *t)
tools/lib/bpf/btf.c
762
return t == &btf_void || btf_is_fwd(t);
tools/lib/bpf/btf.c
765
static bool btf_type_is_void_or_null(const struct btf_type *t)
tools/lib/bpf/btf.c
767
return !t || btf_type_is_void(t);
tools/lib/bpf/btf.c
775
const struct btf_type *t;
tools/lib/bpf/btf.c
780
t = btf__type_by_id(btf, type_id);
tools/lib/bpf/btf.c
781
for (i = 0; i < MAX_RESOLVE_DEPTH && !btf_type_is_void_or_null(t); i++) {
tools/lib/bpf/btf.c
782
switch (btf_kind(t)) {
tools/lib/bpf/btf.c
790
size = t->size;
tools/lib/bpf/btf.c
802
type_id = t->type;
tools/lib/bpf/btf.c
805
array = btf_array(t);
tools/lib/bpf/btf.c
815
t = btf__type_by_id(btf, type_id);
tools/lib/bpf/btf.c
829
const struct btf_type *t = btf__type_by_id(btf, id);
tools/lib/bpf/btf.c
830
__u16 kind = btf_kind(t);
tools/lib/bpf/btf.c
837
return min(btf_ptr_sz(btf), (size_t)t->size);
tools/lib/bpf/btf.c
845
return btf__align_of(btf, t->type);
tools/lib/bpf/btf.c
847
return btf__align_of(btf, btf_array(t)->type);
tools/lib/bpf/btf.c
850
const struct btf_member *m = btf_members(t);
tools/lib/bpf/btf.c
851
__u16 vlen = btf_vlen(t);
tools/lib/bpf/btf.c
863
if (btf_member_bitfield_size(t, i) == 0 &&
tools/lib/bpf/btf.c
871
if ((t->size % max_align) != 0)
tools/lib/bpf/btf.c
877
pr_warn("unsupported BTF_KIND:%u\n", btf_kind(t));
tools/lib/bpf/btf.c
884
const struct btf_type *t;
tools/lib/bpf/btf.c
887
t = btf__type_by_id(btf, type_id);
tools/lib/bpf/btf.c
889
!btf_type_is_void_or_null(t) &&
tools/lib/bpf/btf.c
890
(btf_is_mod(t) || btf_is_typedef(t) || btf_is_var(t))) {
tools/lib/bpf/btf.c
891
type_id = t->type;
tools/lib/bpf/btf.c
892
t = btf__type_by_id(btf, type_id);
tools/lib/bpf/btf.c
896
if (depth == MAX_RESOLVE_DEPTH || btf_type_is_void_or_null(t))
tools/lib/bpf/btf.c
929
const struct btf_type *t;
tools/lib/bpf/btf.c
937
t = btf_type_by_id(btf, m);
tools/lib/bpf/btf.c
938
tname = btf__str_by_offset(btf, t->name_off);
tools/lib/bpf/btf.c
955
const struct btf_type *t;
tools/lib/bpf/btf.c
974
t = btf__type_by_id(btf, id);
tools/lib/bpf/btf.c
975
tname = btf__str_by_offset(btf, t->name_off);
tools/lib/bpf/btf.c
978
if (kind < 0 || btf_kind(t) == kind)
tools/lib/bpf/btf.c
983
t = btf_type_by_id(btf, id);
tools/lib/bpf/btf.c
984
if (kind > 0 && btf_kind(t) != kind)
tools/lib/bpf/btf.c
986
tname = btf__str_by_offset(btf, t->name_off);
tools/lib/bpf/btf.h
413
static inline __u16 btf_kind(const struct btf_type *t)
tools/lib/bpf/btf.h
415
return BTF_INFO_KIND(t->info);
tools/lib/bpf/btf.h
418
static inline __u16 btf_vlen(const struct btf_type *t)
tools/lib/bpf/btf.h
420
return BTF_INFO_VLEN(t->info);
tools/lib/bpf/btf.h
423
static inline bool btf_kflag(const struct btf_type *t)
tools/lib/bpf/btf.h
425
return BTF_INFO_KFLAG(t->info);
tools/lib/bpf/btf.h
428
static inline bool btf_is_void(const struct btf_type *t)
tools/lib/bpf/btf.h
430
return btf_kind(t) == BTF_KIND_UNKN;
tools/lib/bpf/btf.h
433
static inline bool btf_is_int(const struct btf_type *t)
tools/lib/bpf/btf.h
435
return btf_kind(t) == BTF_KIND_INT;
tools/lib/bpf/btf.h
438
static inline bool btf_is_ptr(const struct btf_type *t)
tools/lib/bpf/btf.h
440
return btf_kind(t) == BTF_KIND_PTR;
tools/lib/bpf/btf.h
443
static inline bool btf_is_array(const struct btf_type *t)
tools/lib/bpf/btf.h
445
return btf_kind(t) == BTF_KIND_ARRAY;
tools/lib/bpf/btf.h
448
static inline bool btf_is_struct(const struct btf_type *t)
tools/lib/bpf/btf.h
450
return btf_kind(t) == BTF_KIND_STRUCT;
tools/lib/bpf/btf.h
453
static inline bool btf_is_union(const struct btf_type *t)
tools/lib/bpf/btf.h
455
return btf_kind(t) == BTF_KIND_UNION;
tools/lib/bpf/btf.h
458
static inline bool btf_is_composite(const struct btf_type *t)
tools/lib/bpf/btf.h
460
__u16 kind = btf_kind(t);
tools/lib/bpf/btf.h
465
static inline bool btf_is_enum(const struct btf_type *t)
tools/lib/bpf/btf.h
467
return btf_kind(t) == BTF_KIND_ENUM;
tools/lib/bpf/btf.h
470
static inline bool btf_is_enum64(const struct btf_type *t)
tools/lib/bpf/btf.h
472
return btf_kind(t) == BTF_KIND_ENUM64;
tools/lib/bpf/btf.h
475
static inline bool btf_is_fwd(const struct btf_type *t)
tools/lib/bpf/btf.h
477
return btf_kind(t) == BTF_KIND_FWD;
tools/lib/bpf/btf.h
480
static inline bool btf_is_typedef(const struct btf_type *t)
tools/lib/bpf/btf.h
482
return btf_kind(t) == BTF_KIND_TYPEDEF;
tools/lib/bpf/btf.h
485
static inline bool btf_is_volatile(const struct btf_type *t)
tools/lib/bpf/btf.h
487
return btf_kind(t) == BTF_KIND_VOLATILE;
tools/lib/bpf/btf.h
490
static inline bool btf_is_const(const struct btf_type *t)
tools/lib/bpf/btf.h
492
return btf_kind(t) == BTF_KIND_CONST;
tools/lib/bpf/btf.h
495
static inline bool btf_is_restrict(const struct btf_type *t)
tools/lib/bpf/btf.h
497
return btf_kind(t) == BTF_KIND_RESTRICT;
tools/lib/bpf/btf.h
500
static inline bool btf_is_mod(const struct btf_type *t)
tools/lib/bpf/btf.h
502
__u16 kind = btf_kind(t);
tools/lib/bpf/btf.h
510
static inline bool btf_is_func(const struct btf_type *t)
tools/lib/bpf/btf.h
512
return btf_kind(t) == BTF_KIND_FUNC;
tools/lib/bpf/btf.h
515
static inline bool btf_is_func_proto(const struct btf_type *t)
tools/lib/bpf/btf.h
517
return btf_kind(t) == BTF_KIND_FUNC_PROTO;
tools/lib/bpf/btf.h
520
static inline bool btf_is_var(const struct btf_type *t)
tools/lib/bpf/btf.h
522
return btf_kind(t) == BTF_KIND_VAR;
tools/lib/bpf/btf.h
525
static inline bool btf_is_datasec(const struct btf_type *t)
tools/lib/bpf/btf.h
527
return btf_kind(t) == BTF_KIND_DATASEC;
tools/lib/bpf/btf.h
530
static inline bool btf_is_float(const struct btf_type *t)
tools/lib/bpf/btf.h
532
return btf_kind(t) == BTF_KIND_FLOAT;
tools/lib/bpf/btf.h
535
static inline bool btf_is_decl_tag(const struct btf_type *t)
tools/lib/bpf/btf.h
537
return btf_kind(t) == BTF_KIND_DECL_TAG;
tools/lib/bpf/btf.h
540
static inline bool btf_is_type_tag(const struct btf_type *t)
tools/lib/bpf/btf.h
542
return btf_kind(t) == BTF_KIND_TYPE_TAG;
tools/lib/bpf/btf.h
545
static inline bool btf_is_any_enum(const struct btf_type *t)
tools/lib/bpf/btf.h
547
return btf_is_enum(t) || btf_is_enum64(t);
tools/lib/bpf/btf.h
557
static inline __u8 btf_int_encoding(const struct btf_type *t)
tools/lib/bpf/btf.h
559
return BTF_INT_ENCODING(*(__u32 *)(t + 1));
tools/lib/bpf/btf.h
562
static inline __u8 btf_int_offset(const struct btf_type *t)
tools/lib/bpf/btf.h
564
return BTF_INT_OFFSET(*(__u32 *)(t + 1));
tools/lib/bpf/btf.h
567
static inline __u8 btf_int_bits(const struct btf_type *t)
tools/lib/bpf/btf.h
569
return BTF_INT_BITS(*(__u32 *)(t + 1));
tools/lib/bpf/btf.h
572
static inline struct btf_array *btf_array(const struct btf_type *t)
tools/lib/bpf/btf.h
574
return (struct btf_array *)(t + 1);
tools/lib/bpf/btf.h
577
static inline struct btf_enum *btf_enum(const struct btf_type *t)
tools/lib/bpf/btf.h
579
return (struct btf_enum *)(t + 1);
tools/lib/bpf/btf.h
584
static inline struct btf_enum64 *btf_enum64(const struct btf_type *t)
tools/lib/bpf/btf.h
586
return (struct btf_enum64 *)(t + 1);
tools/lib/bpf/btf.h
615
static inline struct btf_member *btf_members(const struct btf_type *t)
tools/lib/bpf/btf.h
617
return (struct btf_member *)(t + 1);
tools/lib/bpf/btf.h
621
static inline __u32 btf_member_bit_offset(const struct btf_type *t,
tools/lib/bpf/btf.h
624
const struct btf_member *m = btf_members(t) + member_idx;
tools/lib/bpf/btf.h
625
bool kflag = btf_kflag(t);
tools/lib/bpf/btf.h
633
static inline __u32 btf_member_bitfield_size(const struct btf_type *t,
tools/lib/bpf/btf.h
636
const struct btf_member *m = btf_members(t) + member_idx;
tools/lib/bpf/btf.h
637
bool kflag = btf_kflag(t);
tools/lib/bpf/btf.h
642
static inline struct btf_param *btf_params(const struct btf_type *t)
tools/lib/bpf/btf.h
644
return (struct btf_param *)(t + 1);
tools/lib/bpf/btf.h
647
static inline struct btf_var *btf_var(const struct btf_type *t)
tools/lib/bpf/btf.h
649
return (struct btf_var *)(t + 1);
tools/lib/bpf/btf.h
653
btf_var_secinfos(const struct btf_type *t)
tools/lib/bpf/btf.h
655
return (struct btf_var_secinfo *)(t + 1);
tools/lib/bpf/btf.h
659
static inline struct btf_decl_tag *btf_decl_tag(const struct btf_type *t)
tools/lib/bpf/btf.h
661
return (struct btf_decl_tag *)(t + 1);
tools/lib/bpf/btf_dump.c
1017
btf_dump_emit_bit_padding(d, off, t->size * 8, align, false, lvl + 1);
tools/lib/bpf/btf_dump.c
1023
if (vlen || t->size) {
tools/lib/bpf/btf_dump.c
1045
const struct btf_type *t)
tools/lib/bpf/btf_dump.c
1060
const struct btf_type *t)
tools/lib/bpf/btf_dump.c
1066
const struct btf_type *t,
tools/lib/bpf/btf_dump.c
1069
const struct btf_enum *v = btf_enum(t);
tools/lib/bpf/btf_dump.c
1070
bool is_signed = btf_kflag(t);
tools/lib/bpf/btf_dump.c
1091
const struct btf_type *t,
tools/lib/bpf/btf_dump.c
1094
const struct btf_enum64 *v = btf_enum64(t);
tools/lib/bpf/btf_dump.c
1095
bool is_signed = btf_kflag(t);
tools/lib/bpf/btf_dump.c
1122
const struct btf_type *t,
tools/lib/bpf/btf_dump.c
1125
__u16 vlen = btf_vlen(t);
tools/lib/bpf/btf_dump.c
1128
t->name_off ? " " : "",
tools/lib/bpf/btf_dump.c
1135
if (btf_is_enum(t))
tools/lib/bpf/btf_dump.c
1136
btf_dump_emit_enum32_val(d, t, lvl, vlen);
tools/lib/bpf/btf_dump.c
1138
btf_dump_emit_enum64_val(d, t, lvl, vlen);
tools/lib/bpf/btf_dump.c
1142
if (t->size == 1) {
tools/lib/bpf/btf_dump.c
1145
} else if (t->size == 8 && d->ptr_sz == 8) {
tools/lib/bpf/btf_dump.c
1154
if (btf_is_enum(t)) {
tools/lib/bpf/btf_dump.c
1167
if (btf_enum64(t)[i].val_hi32 != 0) {
tools/lib/bpf/btf_dump.c
1180
const struct btf_type *t)
tools/lib/bpf/btf_dump.c
1184
if (btf_kflag(t))
tools/lib/bpf/btf_dump.c
1191
const struct btf_type *t, int lvl)
tools/lib/bpf/btf_dump.c
1201
if (t->type == 0 && strcmp(name, "__gnuc_va_list") == 0) {
tools/lib/bpf/btf_dump.c
1207
btf_dump_emit_type_decl(d, t->type, name, lvl);
tools/lib/bpf/btf_dump.c
1295
const struct btf_type *t;
tools/lib/bpf/btf_dump.c
1300
t = btf__type_by_id(d->btf, id);
tools/lib/bpf/btf_dump.c
1301
if (d->strip_mods && btf_is_mod(t))
tools/lib/bpf/btf_dump.c
1320
switch (btf_kind(t)) {
tools/lib/bpf/btf_dump.c
1327
id = t->type;
tools/lib/bpf/btf_dump.c
1330
id = btf_array(t)->type;
tools/lib/bpf/btf_dump.c
1343
btf_kind(t), id);
tools/lib/bpf/btf_dump.c
1374
const struct btf_type *t;
tools/lib/bpf/btf_dump.c
1379
t = btf__type_by_id(d->btf, id);
tools/lib/bpf/btf_dump.c
1381
switch (btf_kind(t)) {
tools/lib/bpf/btf_dump.c
1400
const struct btf_type *t;
tools/lib/bpf/btf_dump.c
1405
t = btf__type_by_id(d->btf, id);
tools/lib/bpf/btf_dump.c
1406
if (!btf_is_mod(t))
tools/lib/bpf/btf_dump.c
1434
const struct btf_type *t;
tools/lib/bpf/btf_dump.c
1449
t = btf__type_by_id(d->btf, id);
tools/lib/bpf/btf_dump.c
1450
kind = btf_kind(t);
tools/lib/bpf/btf_dump.c
1456
name = btf_name_of(d, t->name_off);
tools/lib/bpf/btf_dump.c
1463
if (t->name_off == 0 && !d->skip_anon_defs)
tools/lib/bpf/btf_dump.c
1464
btf_dump_emit_struct_def(d, id, t, lvl);
tools/lib/bpf/btf_dump.c
1466
btf_dump_emit_struct_fwd(d, id, t);
tools/lib/bpf/btf_dump.c
1472
if (t->name_off == 0 && !d->skip_anon_defs)
tools/lib/bpf/btf_dump.c
1473
btf_dump_emit_enum_def(d, id, t, lvl);
tools/lib/bpf/btf_dump.c
1475
btf_dump_emit_enum_fwd(d, id, t);
tools/lib/bpf/btf_dump.c
1479
btf_dump_emit_fwd_def(d, id, t);
tools/lib/bpf/btf_dump.c
1499
name = btf_name_of(d, t->name_off);
tools/lib/bpf/btf_dump.c
1500
if (btf_kflag(t))
tools/lib/bpf/btf_dump.c
1506
const struct btf_array *a = btf_array(t);
tools/lib/bpf/btf_dump.c
1544
const struct btf_param *p = btf_params(t);
tools/lib/bpf/btf_dump.c
1545
__u16 vlen = btf_vlen(t);
tools/lib/bpf/btf_dump.c
1611
const struct btf_type *t;
tools/lib/bpf/btf_dump.c
1623
t = btf__type_by_id(d->btf, id);
tools/lib/bpf/btf_dump.c
1624
if (btf_is_var(t) || btf_is_datasec(t))
tools/lib/bpf/btf_dump.c
1668
const struct btf_type *t = btf__type_by_id(d->btf, id);
tools/lib/bpf/btf_dump.c
1669
const char *orig_name = btf_name_of(d, t->name_off);
tools/lib/bpf/btf_dump.c
1673
if (t->name_off == 0)
tools/lib/bpf/btf_dump.c
1679
if (btf_is_fwd(t) || (btf_is_enum(t) && btf_vlen(t) == 0)) {
tools/lib/bpf/btf_dump.c
1709
const struct btf_type *t,
tools/lib/bpf/btf_dump.c
1748
const struct btf_type *t,
tools/lib/bpf/btf_dump.c
1751
btf_dump_printf(d, "<unsupported kind:%u>", btf_kind(t));
tools/lib/bpf/btf_dump.c
1756
const struct btf_type *t,
tools/lib/bpf/btf_dump.c
1778
if (t->size > 8) {
tools/lib/bpf/btf_dump.c
1779
pr_warn("unexpected bitfield size %d\n", t->size);
tools/lib/bpf/btf_dump.c
1787
for (i = t->size - 1; i >= 0; i--)
tools/lib/bpf/btf_dump.c
1791
for (i = 0; i < t->size; i++)
tools/lib/bpf/btf_dump.c
1793
nr_copy_bits = t->size * 8 - bits_offset;
tools/lib/bpf/btf_dump.c
1806
const struct btf_type *t,
tools/lib/bpf/btf_dump.c
1814
err = btf_dump_get_bitfield_value(d, t, data, bits_offset, bit_sz, &check_num);
tools/lib/bpf/btf_dump.c
1823
const struct btf_type *t,
tools/lib/bpf/btf_dump.c
1831
err = btf_dump_get_bitfield_value(d, t, data, bits_offset, bit_sz, &print_num);
tools/lib/bpf/btf_dump.c
1842
const struct btf_type *t,
tools/lib/bpf/btf_dump.c
1852
if (btf_kind(t) == BTF_KIND_PTR)
tools/lib/bpf/btf_dump.c
1855
nr_bytes = t->size;
tools/lib/bpf/btf_dump.c
1879
const struct btf_type *t,
tools/lib/bpf/btf_dump.c
1884
__u8 encoding = btf_int_encoding(t);
tools/lib/bpf/btf_dump.c
1887
int sz = t->size;
tools/lib/bpf/btf_dump.c
1978
const struct btf_type *t,
tools/lib/bpf/btf_dump.c
1984
int sz = t->size;
tools/lib/bpf/btf_dump.c
2015
const struct btf_type *t;
tools/lib/bpf/btf_dump.c
2037
t = btf__type_by_id(d->btf, type_id);
tools/lib/bpf/btf_dump.c
2040
return btf_dump_dump_type_data(d, NULL, t, type_id, data, 0, 0);
tools/lib/bpf/btf_dump.c
2044
const struct btf_type *t,
tools/lib/bpf/btf_dump.c
2048
const struct btf_array *array = btf_array(t);
tools/lib/bpf/btf_dump.c
2090
const struct btf_type *t,
tools/lib/bpf/btf_dump.c
2094
const struct btf_array *array = btf_array(t);
tools/lib/bpf/btf_dump.c
2118
btf_dump_string_data(d, t, id, data) == 0) {
tools/lib/bpf/btf_dump.c
2157
const struct btf_type *t,
tools/lib/bpf/btf_dump.c
2161
const struct btf_member *m = btf_members(t);
tools/lib/bpf/btf_dump.c
2162
__u16 n = btf_vlen(t);
tools/lib/bpf/btf_dump.c
2183
moffset = btf_member_bit_offset(t, i);
tools/lib/bpf/btf_dump.c
2185
bit_sz = btf_member_bitfield_size(t, i);
tools/lib/bpf/btf_dump.c
2203
const struct btf_type *t,
tools/lib/bpf/btf_dump.c
2222
const struct btf_type *t,
tools/lib/bpf/btf_dump.c
2227
bool is_signed = btf_kflag(t);
tools/lib/bpf/btf_dump.c
2233
err = btf_dump_get_bitfield_value(d, t, data, 0, 0, &val);
tools/lib/bpf/btf_dump.c
2240
switch (t->size) {
tools/lib/bpf/btf_dump.c
2254
pr_warn("unexpected size %d for enum, id:[%u]\n", t->size, id);
tools/lib/bpf/btf_dump.c
2260
const struct btf_type *t,
tools/lib/bpf/btf_dump.c
2268
err = btf_dump_get_enum_value(d, t, data, id, &value);
tools/lib/bpf/btf_dump.c
2272
is_signed = btf_kflag(t);
tools/lib/bpf/btf_dump.c
2273
if (btf_is_enum(t)) {
tools/lib/bpf/btf_dump.c
2276
for (i = 0, e = btf_enum(t); i < btf_vlen(t); i++, e++) {
tools/lib/bpf/btf_dump.c
2287
for (i = 0, e = btf_enum64(t); i < btf_vlen(t); i++, e++) {
tools/lib/bpf/btf_dump.c
2301
const struct btf_type *t,
tools/lib/bpf/btf_dump.c
2310
btf_dump_type_values(d, "SEC(\"%s\") ", btf_name_of(d, t->name_off));
tools/lib/bpf/btf_dump.c
2312
for (i = 0, vsi = btf_var_secinfos(t); i < btf_vlen(t); i++, vsi++) {
tools/lib/bpf/btf_dump.c
2324
const struct btf_type *t,
tools/lib/bpf/btf_dump.c
2360
t = skip_mods_and_typedefs(d->btf, id, NULL);
tools/lib/bpf/btf_dump.c
2361
if (!t) {
tools/lib/bpf/btf_dump.c
2367
switch (btf_kind(t)) {
tools/lib/bpf/btf_dump.c
2383
const struct btf_type *t,
tools/lib/bpf/btf_dump.c
2408
t = skip_mods_and_typedefs(d->btf, id, NULL);
tools/lib/bpf/btf_dump.c
2410
switch (btf_kind(t)) {
tools/lib/bpf/btf_dump.c
2413
return btf_dump_bitfield_check_zero(d, t, data, bits_offset, bit_sz);
tools/lib/bpf/btf_dump.c
2414
return btf_dump_base_type_check_zero(d, t, id, data);
tools/lib/bpf/btf_dump.c
2417
return btf_dump_base_type_check_zero(d, t, id, data);
tools/lib/bpf/btf_dump.c
2419
const struct btf_array *array = btf_array(t);
tools/lib/bpf/btf_dump.c
2451
const struct btf_member *m = btf_members(t);
tools/lib/bpf/btf_dump.c
2452
__u16 n = btf_vlen(t);
tools/lib/bpf/btf_dump.c
2462
moffset = btf_member_bit_offset(t, i);
tools/lib/bpf/btf_dump.c
2468
bit_sz = btf_member_bitfield_size(t, i);
tools/lib/bpf/btf_dump.c
2478
err = btf_dump_get_enum_value(d, t, data, id, &value);
tools/lib/bpf/btf_dump.c
2492
const struct btf_type *t,
tools/lib/bpf/btf_dump.c
2500
size = btf_dump_type_data_check_overflow(d, t, id, data, bits_offset, bit_sz);
tools/lib/bpf/btf_dump.c
2503
err = btf_dump_type_data_check_zero(d, t, id, data, bits_offset, bit_sz);
tools/lib/bpf/btf_dump.c
2520
t = skip_mods_and_typedefs(d->btf, id, NULL);
tools/lib/bpf/btf_dump.c
2522
switch (btf_kind(t)) {
tools/lib/bpf/btf_dump.c
2528
err = btf_dump_unsupported_data(d, t, id);
tools/lib/bpf/btf_dump.c
2532
err = btf_dump_bitfield_data(d, t, data, bits_offset, bit_sz);
tools/lib/bpf/btf_dump.c
2534
err = btf_dump_int_data(d, t, id, data, bits_offset);
tools/lib/bpf/btf_dump.c
2537
err = btf_dump_float_data(d, t, id, data);
tools/lib/bpf/btf_dump.c
2540
err = btf_dump_ptr_data(d, t, id, data);
tools/lib/bpf/btf_dump.c
2543
err = btf_dump_array_data(d, t, id, data);
tools/lib/bpf/btf_dump.c
2547
err = btf_dump_struct_data(d, t, id, data);
tools/lib/bpf/btf_dump.c
2556
err = btf_dump_get_bitfield_value(d, t, data, bits_offset, bit_sz,
tools/lib/bpf/btf_dump.c
2561
err = btf_dump_enum_data(d, t, id, &enum_val);
tools/lib/bpf/btf_dump.c
2563
err = btf_dump_enum_data(d, t, id, data);
tools/lib/bpf/btf_dump.c
2566
err = btf_dump_var_data(d, t, id, data);
tools/lib/bpf/btf_dump.c
2569
err = btf_dump_datasec_data(d, t, id, data);
tools/lib/bpf/btf_dump.c
2573
BTF_INFO_KIND(t->info), id);
tools/lib/bpf/btf_dump.c
2586
const struct btf_type *t;
tools/lib/bpf/btf_dump.c
2592
t = btf__type_by_id(d->btf, id);
tools/lib/bpf/btf_dump.c
2593
if (!t)
tools/lib/bpf/btf_dump.c
2612
ret = btf_dump_dump_type_data(d, NULL, t, id, data, 0, 0);
tools/lib/bpf/btf_dump.c
318
const struct btf_type *t;
tools/lib/bpf/btf_dump.c
322
t = btf__type_by_id(d->btf, i);
tools/lib/bpf/btf_dump.c
323
vlen = btf_vlen(t);
tools/lib/bpf/btf_dump.c
325
switch (btf_kind(t)) {
tools/lib/bpf/btf_dump.c
342
d->type_states[t->type].referenced = 1;
tools/lib/bpf/btf_dump.c
346
const struct btf_array *a = btf_array(t);
tools/lib/bpf/btf_dump.c
354
const struct btf_member *m = btf_members(t);
tools/lib/bpf/btf_dump.c
361
const struct btf_param *p = btf_params(t);
tools/lib/bpf/btf_dump.c
368
const struct btf_var_secinfo *v = btf_var_secinfos(t);
tools/lib/bpf/btf_dump.c
487
const struct btf_type *t;
tools/lib/bpf/btf_dump.c
495
t = btf__type_by_id(d->btf, id);
tools/lib/bpf/btf_dump.c
499
if (btf_is_composite(t) && through_ptr && t->name_off != 0)
tools/lib/bpf/btf_dump.c
505
switch (btf_kind(t)) {
tools/lib/bpf/btf_dump.c
512
err = btf_dump_order_type(d, t->type, true);
tools/lib/bpf/btf_dump.c
517
return btf_dump_order_type(d, btf_array(t)->type, false);
tools/lib/bpf/btf_dump.c
521
const struct btf_member *m = btf_members(t);
tools/lib/bpf/btf_dump.c
527
if (through_ptr && t->name_off != 0)
tools/lib/bpf/btf_dump.c
532
vlen = btf_vlen(t);
tools/lib/bpf/btf_dump.c
539
if (t->name_off != 0) {
tools/lib/bpf/btf_dump.c
556
if (t->name_off != 0 || !tstate->referenced) {
tools/lib/bpf/btf_dump.c
567
is_strong = btf_dump_order_type(d, t->type, through_ptr);
tools/lib/bpf/btf_dump.c
587
return btf_dump_order_type(d, t->type, through_ptr);
tools/lib/bpf/btf_dump.c
590
const struct btf_param *p = btf_params(t);
tools/lib/bpf/btf_dump.c
593
err = btf_dump_order_type(d, t->type, through_ptr);
tools/lib/bpf/btf_dump.c
598
vlen = btf_vlen(t);
tools/lib/bpf/btf_dump.c
621
const struct btf_type *t);
tools/lib/bpf/btf_dump.c
624
const struct btf_type *t);
tools/lib/bpf/btf_dump.c
626
const struct btf_type *t, int lvl);
tools/lib/bpf/btf_dump.c
629
const struct btf_type *t);
tools/lib/bpf/btf_dump.c
631
const struct btf_type *t, int lvl);
tools/lib/bpf/btf_dump.c
634
const struct btf_type *t);
tools/lib/bpf/btf_dump.c
637
const struct btf_type *t, int lvl);
tools/lib/bpf/btf_dump.c
658
const struct btf_type *t = btf__type_by_id(d->btf, id);
tools/lib/bpf/btf_dump.c
666
if (t->name_off == 0)
tools/lib/bpf/btf_dump.c
668
return strcmp(btf_name_of(d, t->name_off), "__builtin_va_list") == 0;
tools/lib/bpf/btf_dump.c
693
const struct btf_type *t;
tools/lib/bpf/btf_dump.c
699
t = btf__type_by_id(d->btf, id);
tools/lib/bpf/btf_dump.c
700
kind = btf_kind(t);
tools/lib/bpf/btf_dump.c
715
if (t->name_off == 0) {
tools/lib/bpf/btf_dump.c
720
btf_dump_emit_struct_fwd(d, id, t);
tools/lib/bpf/btf_dump.c
731
btf_dump_emit_typedef_def(d, id, t, 0);
tools/lib/bpf/btf_dump.c
746
btf_dump_emit_missing_aliases(d, id, t);
tools/lib/bpf/btf_dump.c
753
btf_dump_emit_enum_def(d, id, t, 0);
tools/lib/bpf/btf_dump.c
763
btf_dump_emit_type(d, t->type, cont_id);
tools/lib/bpf/btf_dump.c
766
btf_dump_emit_type(d, btf_array(t)->type, cont_id);
tools/lib/bpf/btf_dump.c
769
btf_dump_emit_fwd_def(d, id, t);
tools/lib/bpf/btf_dump.c
775
btf_dump_emit_type(d, t->type, id);
tools/lib/bpf/btf_dump.c
784
btf_dump_emit_typedef_def(d, id, t, 0);
tools/lib/bpf/btf_dump.c
799
if (top_level_def || t->name_off == 0) {
tools/lib/bpf/btf_dump.c
800
const struct btf_member *m = btf_members(t);
tools/lib/bpf/btf_dump.c
801
__u16 vlen = btf_vlen(t);
tools/lib/bpf/btf_dump.c
804
new_cont_id = t->name_off == 0 ? cont_id : id;
tools/lib/bpf/btf_dump.c
808
btf_dump_emit_struct_fwd(d, id, t);
tools/lib/bpf/btf_dump.c
814
btf_dump_emit_struct_def(d, id, t, 0);
tools/lib/bpf/btf_dump.c
822
const struct btf_param *p = btf_params(t);
tools/lib/bpf/btf_dump.c
823
__u16 n = btf_vlen(t);
tools/lib/bpf/btf_dump.c
826
btf_dump_emit_type(d, t->type, cont_id);
tools/lib/bpf/btf_dump.c
838
const struct btf_type *t)
tools/lib/bpf/btf_dump.c
844
m = btf_members(t);
tools/lib/bpf/btf_dump.c
845
vlen = btf_vlen(t);
tools/lib/bpf/btf_dump.c
849
bit_sz = btf_member_bitfield_size(t, i);
tools/lib/bpf/btf_dump.c
855
if (t->size % max_align != 0)
tools/lib/bpf/btf_dump.c
959
const struct btf_type *t)
tools/lib/bpf/btf_dump.c
962
btf_is_struct(t) ? "struct" : "union",
tools/lib/bpf/btf_dump.c
963
t->name_off ? " " : "",
tools/lib/bpf/btf_dump.c
969
const struct btf_type *t,
tools/lib/bpf/btf_dump.c
972
const struct btf_member *m = btf_members(t);
tools/lib/bpf/btf_dump.c
973
bool is_struct = btf_is_struct(t);
tools/lib/bpf/btf_dump.c
976
__u16 vlen = btf_vlen(t);
tools/lib/bpf/btf_dump.c
979
packed = is_struct ? btf_is_struct_packed(d->btf, id, t) : 0;
tools/lib/bpf/btf_dump.c
983
t->name_off ? " " : "",
tools/lib/bpf/btf_dump.c
992
m_sz = btf_member_bitfield_size(t, i);
tools/lib/bpf/btf_dump.c
993
m_off = btf_member_bit_offset(t, i);
tools/lib/bpf/btf_iter.c
139
it->vlen = btf_vlen(t);
tools/lib/bpf/btf_iter.c
141
it->p = t;
tools/lib/bpf/btf_iter.c
16
int btf_field_iter_init(struct btf_field_iter *it, struct btf_type *t,
tools/lib/bpf/btf_iter.c
26
switch (btf_kind(t)) {
tools/lib/bpf/btf_iter.c
79
switch (btf_kind(t)) {
tools/lib/bpf/btf_iter.c
9
#define btf_var_secinfos(t) (struct btf_var_secinfo *)btf_type_var_secinfo(t)
tools/lib/bpf/btf_relocate.c
134
struct btf_type *t = btf_type_by_id(r->btf, i);
tools/lib/bpf/btf_relocate.c
139
if (!btf_is_composite(t))
tools/lib/bpf/btf_relocate.c
142
err = btf_field_iter_init(&it, t, BTF_FIELD_ITER_IDS);
tools/lib/bpf/btf_relocate.c
150
t = btf_type_by_id(r->btf, next_id);
tools/lib/bpf/btf_relocate.c
151
switch (btf_kind(t)) {
tools/lib/bpf/btf_relocate.c
157
next_id = t->type;
tools/lib/bpf/btf_relocate.c
160
struct btf_array *a = btf_array(t);
tools/lib/bpf/btf_relocate.c
388
struct btf_type *t = btf_type_by_id(r->dist_base_btf, i);
tools/lib/bpf/btf_relocate.c
389
int kind = btf_kind(t);
tools/lib/bpf/btf_relocate.c
398
if (t->name_off)
tools/lib/bpf/btf_relocate.c
414
struct btf_type *t = btf_type_by_id(r->btf, i);
tools/lib/bpf/btf_relocate.c
419
err = btf_field_iter_init(&it, t, BTF_FIELD_ITER_STRS);
tools/lib/bpf/btf_relocate.c
68
struct btf_type *t = btf_type_by_id(r->btf, i);
tools/lib/bpf/btf_relocate.c
73
err = btf_field_iter_init(&it, t, BTF_FIELD_ITER_IDS);
tools/lib/bpf/libbpf.c
1001
find_member_by_offset(const struct btf_type *t, __u32 bit_offset)
tools/lib/bpf/libbpf.c
1006
for (i = 0, m = btf_members(t); i < btf_vlen(t); i++, m++) {
tools/lib/bpf/libbpf.c
1007
if (btf_member_bit_offset(t, i) == bit_offset)
tools/lib/bpf/libbpf.c
10138
const char *libbpf_bpf_attach_type_str(enum bpf_attach_type t)
tools/lib/bpf/libbpf.c
10140
if (t < 0 || t >= ARRAY_SIZE(attach_type_name))
tools/lib/bpf/libbpf.c
10143
return attach_type_name[t];
tools/lib/bpf/libbpf.c
10146
const char *libbpf_bpf_link_type_str(enum bpf_link_type t)
tools/lib/bpf/libbpf.c
10148
if (t < 0 || t >= ARRAY_SIZE(link_type_name))
tools/lib/bpf/libbpf.c
1015
find_member_by_name(const struct btf *btf, const struct btf_type *t,
tools/lib/bpf/libbpf.c
10151
return link_type_name[t];
tools/lib/bpf/libbpf.c
10154
const char *libbpf_bpf_map_type_str(enum bpf_map_type t)
tools/lib/bpf/libbpf.c
10156
if (t < 0 || t >= ARRAY_SIZE(map_type_name))
tools/lib/bpf/libbpf.c
10159
return map_type_name[t];
tools/lib/bpf/libbpf.c
10162
const char *libbpf_bpf_prog_type_str(enum bpf_prog_type t)
tools/lib/bpf/libbpf.c
10164
if (t < 0 || t >= ARRAY_SIZE(prog_type_name))
tools/lib/bpf/libbpf.c
10167
return prog_type_name[t];
tools/lib/bpf/libbpf.c
1021
for (i = 0, m = btf_members(t); i < btf_vlen(t); i++, m++) {
tools/lib/bpf/libbpf.c
1932
const struct btf_type *t, *vt;
tools/lib/bpf/libbpf.c
1939
t = btf__type_by_id(obj->btf, map->btf_value_type_id);
tools/lib/bpf/libbpf.c
1940
if (!btf_is_datasec(t))
tools/lib/bpf/libbpf.c
1943
vsi = btf_var_secinfos(t);
tools/lib/bpf/libbpf.c
1944
for (i = 0, n = btf_vlen(t); i < n; i++, vsi++) {
tools/lib/bpf/libbpf.c
2401
const struct btf_type *t = btf__type_by_id(btf, id);
tools/lib/bpf/libbpf.c
2406
while (btf_is_mod(t) || btf_is_typedef(t)) {
tools/lib/bpf/libbpf.c
2408
*res_id = t->type;
tools/lib/bpf/libbpf.c
2409
t = btf__type_by_id(btf, t->type);
tools/lib/bpf/libbpf.c
2412
return t;
tools/lib/bpf/libbpf.c
2418
const struct btf_type *t;
tools/lib/bpf/libbpf.c
2420
t = skip_mods_and_typedefs(btf, id, NULL);
tools/lib/bpf/libbpf.c
2421
if (!btf_is_ptr(t))
tools/lib/bpf/libbpf.c
2424
t = skip_mods_and_typedefs(btf, t->type, res_id);
tools/lib/bpf/libbpf.c
2426
return btf_is_func_proto(t) ? t : NULL;
tools/lib/bpf/libbpf.c
2456
const char *btf_kind_str(const struct btf_type *t)
tools/lib/bpf/libbpf.c
2458
return __btf_kind_str(btf_kind(t));
tools/lib/bpf/libbpf.c
2471
const struct btf_type *t = skip_mods_and_typedefs(btf, m->type, NULL);
tools/lib/bpf/libbpf.c
2476
if (!btf_is_ptr(t)) {
tools/lib/bpf/libbpf.c
2478
map_name, name, btf_kind_str(t));
tools/lib/bpf/libbpf.c
2482
arr_t = btf__type_by_id(btf, t->type);
tools/lib/bpf/libbpf.c
2485
map_name, name, t->type);
tools/lib/bpf/libbpf.c
2501
const struct btf_type *t = skip_mods_and_typedefs(btf, m->type, NULL);
tools/lib/bpf/libbpf.c
2504
if (btf_is_ptr(t)) {
tools/lib/bpf/libbpf.c
2514
if (!btf_is_enum(t) && !btf_is_enum64(t)) {
tools/lib/bpf/libbpf.c
2516
map_name, name, btf_kind_str(t));
tools/lib/bpf/libbpf.c
2520
if (btf_vlen(t) != 1) {
tools/lib/bpf/libbpf.c
2526
if (btf_is_enum(t)) {
tools/lib/bpf/libbpf.c
2527
const struct btf_enum *e = btf_enum(t);
tools/lib/bpf/libbpf.c
2531
const struct btf_enum64 *e = btf_enum64(t);
tools/lib/bpf/libbpf.c
2577
const struct btf_type *t;
tools/lib/bpf/libbpf.c
2622
t = btf__type_by_id(btf, m->type);
tools/lib/bpf/libbpf.c
2623
if (!t) {
tools/lib/bpf/libbpf.c
2628
if (!btf_is_ptr(t)) {
tools/lib/bpf/libbpf.c
2630
map_name, btf_kind_str(t));
tools/lib/bpf/libbpf.c
2633
sz = btf__resolve_size(btf, t->type);
tools/lib/bpf/libbpf.c
2636
map_name, t->type, (ssize_t)sz);
tools/lib/bpf/libbpf.c
2645
map_def->key_type_id = t->type;
tools/lib/bpf/libbpf.c
2662
t = btf__type_by_id(btf, m->type);
tools/lib/bpf/libbpf.c
2663
if (!t) {
tools/lib/bpf/libbpf.c
2668
if (!btf_is_ptr(t)) {
tools/lib/bpf/libbpf.c
2670
map_name, btf_kind_str(t));
tools/lib/bpf/libbpf.c
2673
sz = btf__resolve_size(btf, t->type);
tools/lib/bpf/libbpf.c
2676
map_name, t->type, (ssize_t)sz);
tools/lib/bpf/libbpf.c
2685
map_def->value_type_id = t->type;
tools/lib/bpf/libbpf.c
2716
t = btf__type_by_id(btf, m->type);
tools/lib/bpf/libbpf.c
2717
if (!t) {
tools/lib/bpf/libbpf.c
2722
if (!btf_is_array(t) || btf_array(t)->nelems) {
tools/lib/bpf/libbpf.c
2727
t = skip_mods_and_typedefs(btf, btf_array(t)->type, NULL);
tools/lib/bpf/libbpf.c
2728
if (!btf_is_ptr(t)) {
tools/lib/bpf/libbpf.c
2730
map_name, desc, btf_kind_str(t));
tools/lib/bpf/libbpf.c
2733
t = skip_mods_and_typedefs(btf, t->type, NULL);
tools/lib/bpf/libbpf.c
2735
if (!btf_is_func_proto(t)) {
tools/lib/bpf/libbpf.c
2737
map_name, btf_kind_str(t));
tools/lib/bpf/libbpf.c
2742
if (!btf_is_struct(t)) {
tools/lib/bpf/libbpf.c
2744
map_name, btf_kind_str(t));
tools/lib/bpf/libbpf.c
2749
err = parse_btf_map_def(inner_map_name, btf, t, strict, inner_def, NULL);
tools/lib/bpf/libbpf.c
3023
const struct btf_type *t;
tools/lib/bpf/libbpf.c
3041
t = btf__type_by_id(obj->btf, i);
tools/lib/bpf/libbpf.c
3042
if (!btf_is_datasec(t))
tools/lib/bpf/libbpf.c
3044
name = btf__name_by_offset(obj->btf, t->name_off);
tools/lib/bpf/libbpf.c
3046
sec = t;
tools/lib/bpf/libbpf.c
3157
struct btf_type *t;
tools/lib/bpf/libbpf.c
3161
t = (struct btf_type *)btf__type_by_id(btf, i);
tools/lib/bpf/libbpf.c
3163
if ((!has_datasec && btf_is_var(t)) || (!has_decl_tag && btf_is_decl_tag(t))) {
tools/lib/bpf/libbpf.c
3165
t->info = BTF_INFO_ENC(BTF_KIND_INT, 0, 0);
tools/lib/bpf/libbpf.c
3171
t->size = 1;
tools/lib/bpf/libbpf.c
3172
*(int *)(t + 1) = BTF_INT_ENC(0, 0, 8);
tools/lib/bpf/libbpf.c
3173
} else if (!has_datasec && btf_is_datasec(t)) {
tools/lib/bpf/libbpf.c
3175
const struct btf_var_secinfo *v = btf_var_secinfos(t);
tools/lib/bpf/libbpf.c
3176
struct btf_member *m = btf_members(t);
tools/lib/bpf/libbpf.c
3180
name = (char *)btf__name_by_offset(btf, t->name_off);
tools/lib/bpf/libbpf.c
3187
vlen = btf_vlen(t);
tools/lib/bpf/libbpf.c
3188
t->info = BTF_INFO_ENC(BTF_KIND_STRUCT, 0, vlen);
tools/lib/bpf/libbpf.c
3197
} else if (!has_qmark_datasec && btf_is_datasec(t) &&
tools/lib/bpf/libbpf.c
3198
starts_with_qmark(btf__name_by_offset(btf, t->name_off))) {
tools/lib/bpf/libbpf.c
3202
name = (char *)btf__name_by_offset(btf, t->name_off);
tools/lib/bpf/libbpf.c
3205
} else if (!has_func && btf_is_func_proto(t)) {
tools/lib/bpf/libbpf.c
3207
vlen = btf_vlen(t);
tools/lib/bpf/libbpf.c
3208
t->info = BTF_INFO_ENC(BTF_KIND_ENUM, 0, vlen);
tools/lib/bpf/libbpf.c
3209
t->size = sizeof(__u32); /* kernel enforced */
tools/lib/bpf/libbpf.c
3210
} else if (!has_func && btf_is_func(t)) {
tools/lib/bpf/libbpf.c
3212
t->info = BTF_INFO_ENC(BTF_KIND_TYPEDEF, 0, 0);
tools/lib/bpf/libbpf.c
3213
} else if (!has_func_global && btf_is_func(t)) {
tools/lib/bpf/libbpf.c
3215
t->info = BTF_INFO_ENC(BTF_KIND_FUNC, 0, 0);
tools/lib/bpf/libbpf.c
3216
} else if (!has_float && btf_is_float(t)) {
tools/lib/bpf/libbpf.c
3221
t->name_off = 0;
tools/lib/bpf/libbpf.c
3222
t->info = BTF_INFO_ENC(BTF_KIND_STRUCT, 0, 0);
tools/lib/bpf/libbpf.c
3223
} else if (!has_type_tag && btf_is_type_tag(t)) {
tools/lib/bpf/libbpf.c
3225
t->name_off = 0;
tools/lib/bpf/libbpf.c
3226
t->info = BTF_INFO_ENC(BTF_KIND_CONST, 0, 0);
tools/lib/bpf/libbpf.c
3227
} else if (!has_enum64 && btf_is_enum(t)) {
tools/lib/bpf/libbpf.c
3229
t->info = btf_type_info(btf_kind(t), btf_vlen(t), false);
tools/lib/bpf/libbpf.c
3230
} else if (!has_enum64 && btf_is_enum64(t)) {
tools/lib/bpf/libbpf.c
3239
t = (struct btf_type *)btf__type_by_id(btf, i);
tools/lib/bpf/libbpf.c
3242
m = btf_members(t);
tools/lib/bpf/libbpf.c
3243
vlen = btf_vlen(t);
tools/lib/bpf/libbpf.c
3244
t->info = BTF_INFO_ENC(BTF_KIND_UNION, 0, vlen);
tools/lib/bpf/libbpf.c
3356
struct btf_type *t)
tools/lib/bpf/libbpf.c
3358
__u32 size = 0, i, vars = btf_vlen(t);
tools/lib/bpf/libbpf.c
3359
const char *sec_name = btf__name_by_offset(btf, t->name_off);
tools/lib/bpf/libbpf.c
3386
if (t->size == 0) {
tools/lib/bpf/libbpf.c
3394
t->size = size;
tools/lib/bpf/libbpf.c
3398
for (i = 0, vsi = btf_var_secinfos(t); i < vars; i++, vsi++) {
tools/lib/bpf/libbpf.c
3444
qsort(btf_var_secinfos(t), vars, sizeof(*vsi), compare_vsi_off);
tools/lib/bpf/libbpf.c
3457
struct btf_type *t = btf_type_by_id(obj->btf, i);
tools/lib/bpf/libbpf.c
3464
if (btf_is_datasec(t)) {
tools/lib/bpf/libbpf.c
3465
err = btf_fixup_datasec(obj, obj->btf, t);
tools/lib/bpf/libbpf.c
3579
struct btf_type *t;
tools/lib/bpf/libbpf.c
3588
t = btf_type_by_id(obj->btf, j);
tools/lib/bpf/libbpf.c
3589
if (!btf_is_func(t) || btf_func_linkage(t) != BTF_FUNC_GLOBAL)
tools/lib/bpf/libbpf.c
3592
name = btf__str_by_offset(obj->btf, t->name_off);
tools/lib/bpf/libbpf.c
3596
t->info = btf_type_info(BTF_KIND_FUNC, BTF_FUNC_STATIC, 0);
tools/lib/bpf/libbpf.c
4070
const struct btf_type *t;
tools/lib/bpf/libbpf.c
4079
t = btf__type_by_id(btf, i);
tools/lib/bpf/libbpf.c
4081
if (!btf_is_var(t) && !btf_is_func(t))
tools/lib/bpf/libbpf.c
4084
tname = btf__name_by_offset(btf, t->name_off);
tools/lib/bpf/libbpf.c
4088
if (btf_is_var(t) &&
tools/lib/bpf/libbpf.c
4089
btf_var(t)->linkage != BTF_VAR_GLOBAL_EXTERN)
tools/lib/bpf/libbpf.c
4092
if (btf_is_func(t) && btf_func_linkage(t) != BTF_FUNC_EXTERN)
tools/lib/bpf/libbpf.c
4103
const struct btf_type *t;
tools/lib/bpf/libbpf.c
4111
t = btf__type_by_id(btf, i);
tools/lib/bpf/libbpf.c
4113
if (!btf_is_datasec(t))
tools/lib/bpf/libbpf.c
4116
vs = btf_var_secinfos(t);
tools/lib/bpf/libbpf.c
4117
for (j = 0; j < btf_vlen(t); j++, vs++) {
tools/lib/bpf/libbpf.c
4129
const struct btf_type *t;
tools/lib/bpf/libbpf.c
4132
t = skip_mods_and_typedefs(btf, id, NULL);
tools/lib/bpf/libbpf.c
4133
name = btf__name_by_offset(btf, t->name_off);
tools/lib/bpf/libbpf.c
4137
switch (btf_kind(t)) {
tools/lib/bpf/libbpf.c
4139
int enc = btf_int_encoding(t);
tools/lib/bpf/libbpf.c
4142
return t->size == 1 ? KCFG_BOOL : KCFG_UNKNOWN;
tools/lib/bpf/libbpf.c
4145
if (t->size == 1)
tools/lib/bpf/libbpf.c
4147
if (t->size < 1 || t->size > 8 || (t->size & (t->size - 1)))
tools/lib/bpf/libbpf.c
4152
if (t->size != 4)
tools/lib/bpf/libbpf.c
4162
if (btf_array(t)->nelems == 0)
tools/lib/bpf/libbpf.c
4164
if (find_kcfg_type(btf, btf_array(t)->type, NULL) != KCFG_CHAR)
tools/lib/bpf/libbpf.c
4195
const struct btf_type *t;
tools/lib/bpf/libbpf.c
4200
t = btf__type_by_id(btf, i);
tools/lib/bpf/libbpf.c
4202
if (btf_is_int(t) && btf_int_bits(t) == 32)
tools/lib/bpf/libbpf.c
4251
const struct btf_type *t;
tools/lib/bpf/libbpf.c
4300
t = btf__type_by_id(obj->btf, ext->btf_id);
tools/lib/bpf/libbpf.c
4301
ext->name = strdup(btf__name_by_offset(obj->btf, t->name_off));
tools/lib/bpf/libbpf.c
4325
if (btf_is_func(t)) {
tools/lib/bpf/libbpf.c
4332
ext->kcfg.sz = btf__resolve_size(obj->btf, t->type);
tools/lib/bpf/libbpf.c
4338
ext->kcfg.align = btf__align_of(obj->btf, t->type);
tools/lib/bpf/libbpf.c
4344
ext->kcfg.type = find_kcfg_type(obj->btf, t->type,
tools/lib/bpf/libbpf.c
4353
skip_mods_and_typedefs(obj->btf, t->type,
tools/lib/bpf/libbpf.c
4454
t = btf__type_by_id(obj->btf, vs->type);
tools/lib/bpf/libbpf.c
4455
ext_name = btf__name_by_offset(obj->btf, t->name_off);
tools/lib/bpf/libbpf.c
4462
btf_var(t)->linkage = BTF_VAR_GLOBAL_ALLOCATED;
tools/lib/bpf/libbpf.c
5707
const struct btf_type *t, *local_t;
tools/lib/bpf/libbpf.c
5717
t = btf__type_by_id(targ_btf, i);
tools/lib/bpf/libbpf.c
5718
if (!btf_kind_core_compat(t, local_t))
tools/lib/bpf/libbpf.c
5721
targ_name = btf__name_by_offset(targ_btf, t->name_off);
tools/lib/bpf/libbpf.c
5734
local_name, i, btf_kind_str(t), targ_name,
tools/lib/bpf/libbpf.c
6983
struct btf_type *t;
tools/lib/bpf/libbpf.c
6985
t = btf_type_by_id(obj->btf, i);
tools/lib/bpf/libbpf.c
6986
if (!btf_is_decl_tag(t) || btf_decl_tag(t)->component_idx != -1)
tools/lib/bpf/libbpf.c
6989
name = btf__str_by_offset(obj->btf, t->name_off);
tools/lib/bpf/libbpf.c
6993
t = btf_type_by_id(obj->btf, t->type);
tools/lib/bpf/libbpf.c
6994
if (!btf_is_func(t) || btf_func_linkage(t) != BTF_FUNC_GLOBAL) {
tools/lib/bpf/libbpf.c
6999
if (strcmp(prog->name, btf__str_by_offset(obj->btf, t->name_off)) != 0)
tools/lib/bpf/libbpf.c
7097
const struct btf_type *t;
tools/lib/bpf/libbpf.c
7101
t = skip_mods_and_typedefs(btf, arg_type_id, NULL);
tools/lib/bpf/libbpf.c
7102
if (!btf_is_ptr(t))
tools/lib/bpf/libbpf.c
7109
while (btf_is_mod(t))
tools/lib/bpf/libbpf.c
7110
t = btf__type_by_id(btf, t->type);
tools/lib/bpf/libbpf.c
7111
if (btf_is_typedef(t) &&
tools/lib/bpf/libbpf.c
7113
tname = btf__str_by_offset(btf, t->name_off) ?: "<anon>";
tools/lib/bpf/libbpf.c
7119
t = skip_mods_and_typedefs(btf, t->type, NULL);
tools/lib/bpf/libbpf.c
7122
if (btf_is_void(t))
tools/lib/bpf/libbpf.c
7126
tname = btf__str_by_offset(btf, t->name_off) ?: "<anon>";
tools/lib/bpf/libbpf.c
7127
if (btf_is_struct(t) && strcmp(tname, ctx_name) == 0)
tools/lib/bpf/libbpf.c
7134
if (btf_is_struct(t) && strcmp(tname, "pt_regs") == 0)
tools/lib/bpf/libbpf.c
7139
btf_is_struct(t) && strcmp(tname, "pt_regs") == 0)
tools/lib/bpf/libbpf.c
7142
btf_is_struct(t) && strcmp(tname, "user_pt_regs") == 0)
tools/lib/bpf/libbpf.c
7145
btf_is_struct(t) && strcmp(tname, "user_regs_struct") == 0)
tools/lib/bpf/libbpf.c
7151
if (btf_is_int(t) && t->size == 8)
tools/lib/bpf/libbpf.c
7168
struct btf_type *fn_t, *fn_proto_t, *t;
tools/lib/bpf/libbpf.c
7199
t = btf_type_by_id(btf, orig_proto_id);
tools/lib/bpf/libbpf.c
7200
p = &btf_params(t)[i];
tools/lib/bpf/libbpf.c
7239
const struct btf_type *t;
tools/lib/bpf/libbpf.c
7280
t = btf__type_by_id(btf, i);
tools/lib/bpf/libbpf.c
7281
if (!btf_is_decl_tag(t))
tools/lib/bpf/libbpf.c
7283
if (strcmp(btf__str_by_offset(btf, t->name_off), ctx_tag) != 0)
tools/lib/bpf/libbpf.c
7287
orig_fn_id = t->type;
tools/lib/bpf/libbpf.c
7300
if (orig_ids[rec_idx] == t->type) {
tools/lib/bpf/libbpf.c
7311
arg_idx = btf_decl_tag(t)->component_idx;
tools/lib/bpf/libbpf.c
8501
const struct btf_type *t;
tools/lib/bpf/libbpf.c
8513
t = btf__type_by_id(obj->btf, ext->btf_id);
tools/lib/bpf/libbpf.c
8514
if (!btf_is_var(t))
tools/lib/bpf/libbpf.c
8696
const struct btf_type *t;
tools/lib/bpf/libbpf.c
8711
t = btf__type_by_id(obj->btf, ext->btf_id);
tools/lib/bpf/libbpf.c
8712
if (btf_is_var(t))
tools/lib/bpf/libbpf.h
104
LIBBPF_API const char *libbpf_bpf_map_type_str(enum bpf_map_type t);
tools/lib/bpf/libbpf.h
113
LIBBPF_API const char *libbpf_bpf_prog_type_str(enum bpf_prog_type t);
tools/lib/bpf/libbpf.h
86
LIBBPF_API const char *libbpf_bpf_attach_type_str(enum bpf_attach_type t);
tools/lib/bpf/libbpf.h
95
LIBBPF_API const char *libbpf_bpf_link_type_str(enum bpf_link_type t);
tools/lib/bpf/libbpf_internal.h
248
const char *btf_kind_str(const struct btf_type *t);
tools/lib/bpf/libbpf_internal.h
254
static inline enum btf_func_linkage btf_func_linkage(const struct btf_type *t)
tools/lib/bpf/libbpf_internal.h
256
return (enum btf_func_linkage)(int)btf_vlen(t);
tools/lib/bpf/libbpf_internal.h
580
int btf_field_iter_init(struct btf_field_iter *it, struct btf_type *t, enum btf_field_iter_kind iter_kind);
tools/lib/bpf/linker.c
1088
struct btf_type *t;
tools/lib/bpf/linker.c
1099
t = btf_type_by_id(obj->btf, i);
tools/lib/bpf/linker.c
1101
err = btf_field_iter_init(&it, t, BTF_FIELD_ITER_IDS);
tools/lib/bpf/linker.c
1109
err = btf_field_iter_init(&it, t, BTF_FIELD_ITER_STRS);
tools/lib/bpf/linker.c
1770
const struct btf_type *t;
tools/lib/bpf/linker.c
1773
t = btf__type_by_id(obj->btf, btf_id);
tools/lib/bpf/linker.c
1774
if (!btf_is_var(t)) {
tools/lib/bpf/linker.c
1778
t = skip_mods_and_typedefs(obj->btf, t->type, NULL);
tools/lib/bpf/linker.c
1780
err = parse_btf_map_def(sym_name, obj->btf, t, true /*strict*/, &src_def, &src_inner_def);
tools/lib/bpf/linker.c
1787
t = btf__type_by_id(linker->btf, glob_sym->btf_id);
tools/lib/bpf/linker.c
1788
t = skip_mods_and_typedefs(linker->btf, t->type, NULL);
tools/lib/bpf/linker.c
1789
err = parse_btf_map_def(sym_name, linker->btf, t, true /*strict*/, &dst_def, &dst_inner_def);
tools/lib/bpf/linker.c
1836
static bool btf_is_non_static(const struct btf_type *t)
tools/lib/bpf/linker.c
1838
return (btf_is_var(t) && btf_var(t)->linkage != BTF_VAR_STATIC)
tools/lib/bpf/linker.c
1839
|| (btf_is_func(t) && btf_func_linkage(t) != BTF_FUNC_STATIC);
tools/lib/bpf/linker.c
1846
const struct btf_type *t;
tools/lib/bpf/linker.c
1857
t = btf__type_by_id(obj->btf, i);
tools/lib/bpf/linker.c
1862
if (btf_is_non_static(t)) {
tools/lib/bpf/linker.c
1863
name = btf__str_by_offset(obj->btf, t->name_off);
tools/lib/bpf/linker.c
1872
if (!btf_is_datasec(t))
tools/lib/bpf/linker.c
1875
vi = btf_var_secinfos(t);
tools/lib/bpf/linker.c
1876
for (j = 0, m = btf_vlen(t); j < m; j++, vi++) {
tools/lib/bpf/linker.c
1877
t = btf__type_by_id(obj->btf, vi->type);
tools/lib/bpf/linker.c
1878
name = btf__str_by_offset(obj->btf, t->name_off);
tools/lib/bpf/linker.c
1882
if (btf_is_var(t) && btf_var(t)->linkage == BTF_VAR_STATIC)
tools/lib/bpf/linker.c
1884
if (btf_is_func(t) && btf_func_linkage(t) == BTF_FUNC_STATIC)
tools/lib/bpf/linker.c
2043
const struct btf_type *t;
tools/lib/bpf/linker.c
2045
t = btf__type_by_id(obj->btf, btf_sec_id);
tools/lib/bpf/linker.c
2046
sec_name = btf__str_by_offset(obj->btf, t->name_off);
tools/lib/bpf/linker.c
2329
struct btf_type *t;
tools/lib/bpf/linker.c
2331
t = btf_type_by_id(obj->btf, i);
tools/lib/bpf/linker.c
2332
if (btf_kind(t) != BTF_KIND_DATASEC)
tools/lib/bpf/linker.c
2335
sec_name = btf__str_by_offset(obj->btf, t->name_off);
tools/lib/bpf/linker.c
2340
t->size = sec->shdr->sh_size;
tools/lib/bpf/linker.c
2381
vi = btf_var_secinfos(t);
tools/lib/bpf/linker.c
2382
for (j = 0, m = btf_vlen(t); j < m; j++, vi++) {
tools/lib/bpf/linker.c
2414
const struct btf_type *t;
tools/lib/bpf/linker.c
2431
t = btf__type_by_id(obj->btf, i);
tools/lib/bpf/linker.c
2434
if (btf_kind(t) == BTF_KIND_DATASEC)
tools/lib/bpf/linker.c
2437
if (btf_is_non_static(t)) {
tools/lib/bpf/linker.c
2439
name = btf__str_by_offset(obj->btf, t->name_off);
tools/lib/bpf/linker.c
2454
glob_sym->underlying_btf_id = -t->type;
tools/lib/bpf/linker.c
2467
id = btf__add_type(linker->btf, obj->btf, t);
tools/lib/bpf/linker.c
2478
glob_sym->underlying_btf_id = -t->type;
tools/lib/bpf/linker.c
2545
t = btf__type_by_id(obj->btf, src_sec->sec_type_id);
tools/lib/bpf/linker.c
2546
src_var = btf_var_secinfos(t);
tools/lib/bpf/linker.c
2547
n = btf_vlen(t);
tools/lib/bpf/linker.c
2553
t = btf_type_by_id(linker->btf, new_id);
tools/lib/bpf/linker.c
2554
if (btf_is_non_static(t)) {
tools/lib/bpf/linker.c
2555
name = btf__str_by_offset(linker->btf, t->name_off);
tools/lib/bpf/linker.c
2573
if (btf_is_func(t))
tools/lib/bpf/relo_core.c
11
static const char *btf_kind_str(const struct btf_type *t)
tools/lib/bpf/relo_core.c
1173
const struct btf_type *t;
tools/lib/bpf/relo_core.c
1190
t = btf_type_by_id(spec->btf, type_id);
tools/lib/bpf/relo_core.c
1191
s = btf__name_by_offset(spec->btf, t->name_off);
tools/lib/bpf/relo_core.c
1195
type_id, btf_kind_str(t), str_is_empty(s) ? "<anon>" : s);
tools/lib/bpf/relo_core.c
1201
t = skip_mods_and_typedefs(spec->btf, type_id, NULL);
tools/lib/bpf/relo_core.c
1202
if (btf_is_enum(t)) {
tools/lib/bpf/relo_core.c
1206
e = btf_enum(t) + spec->raw_spec[0];
tools/lib/bpf/relo_core.c
1208
fmt_str = BTF_INFO_KFLAG(t->info) ? "::%s = %d" : "::%s = %u";
tools/lib/bpf/relo_core.c
1214
e = btf_enum64(t) + spec->raw_spec[0];
tools/lib/bpf/relo_core.c
1216
fmt_str = BTF_INFO_KFLAG(t->info) ? "::%s = %lld" : "::%s = %llu";
tools/lib/bpf/relo_core.c
13
return btf_type_str(t);
tools/lib/bpf/relo_core.c
268
const struct btf_type *t;
tools/lib/bpf/relo_core.c
304
t = skip_mods_and_typedefs(btf, relo->type_id, &id);
tools/lib/bpf/relo_core.c
305
if (!t)
tools/lib/bpf/relo_core.c
315
if (!btf_is_any_enum(t) || spec->raw_len > 1 || access_idx >= btf_vlen(t))
tools/lib/bpf/relo_core.c
319
name_off = btf_is_enum(t) ? btf_enum(t)[access_idx].name_off
tools/lib/bpf/relo_core.c
320
: btf_enum64(t)[access_idx].name_off;
tools/lib/bpf/relo_core.c
334
t = skip_mods_and_typedefs(btf, id, &id);
tools/lib/bpf/relo_core.c
335
if (!t)
tools/lib/bpf/relo_core.c
34
const struct btf_type *t;
tools/lib/bpf/relo_core.c
341
if (btf_is_composite(t)) {
tools/lib/bpf/relo_core.c
345
if (access_idx >= btf_vlen(t))
tools/lib/bpf/relo_core.c
348
bit_offset = btf_member_bit_offset(t, access_idx);
tools/lib/bpf/relo_core.c
351
m = btf_members(t) + access_idx;
tools/lib/bpf/relo_core.c
364
} else if (btf_is_array(t)) {
tools/lib/bpf/relo_core.c
365
const struct btf_array *a = btf_array(t);
tools/lib/bpf/relo_core.c
368
t = skip_mods_and_typedefs(btf, a->type, &id);
tools/lib/bpf/relo_core.c
369
if (!t)
tools/lib/bpf/relo_core.c
37
t = btf_type_by_id(btf, type_id);
tools/lib/bpf/relo_core.c
38
t = btf_resolve_size(btf, t, &size);
tools/lib/bpf/relo_core.c
386
prog_name, relo->type_id, spec_str, i, id, btf_kind_str(t));
tools/lib/bpf/relo_core.c
39
if (IS_ERR(t))
tools/lib/bpf/relo_core.c
40
return PTR_ERR(t);
tools/lib/bpf/relo_core.c
684
const struct btf_type *t;
tools/lib/bpf/relo_core.c
702
t = btf_type_by_id(spec->btf, acc->type_id);
tools/lib/bpf/relo_core.c
712
t = skip_mods_and_typedefs(spec->btf, acc->type_id, &elem_id);
tools/lib/bpf/relo_core.c
713
while (btf_is_array(t))
tools/lib/bpf/relo_core.c
714
t = skip_mods_and_typedefs(spec->btf, btf_array(t)->type, &elem_id);
tools/lib/bpf/relo_core.c
735
m = btf_members(t) + acc->idx;
tools/lib/bpf/relo_core.c
738
bit_sz = btf_member_bitfield_size(t, acc->idx);
tools/lib/bpf/relo_core.c
74
const struct btf_type *t;
tools/lib/bpf/relo_core.c
779
t = skip_mods_and_typedefs(spec->btf, field_type_id, &elem_id);
tools/lib/bpf/relo_core.c
780
while (btf_is_array(t))
tools/lib/bpf/relo_core.c
781
t = skip_mods_and_typedefs(spec->btf, btf_array(t)->type, &elem_id);
tools/lib/bpf/relo_core.c
81
t = btf_type_by_id(btf, acc->type_id);
tools/lib/bpf/relo_core.c
82
return acc->idx == btf_vlen(t) - 1;
tools/lib/bpf/relo_core.c
866
const struct btf_type *t;
tools/lib/bpf/relo_core.c
875
t = btf_type_by_id(spec->btf, spec->spec[0].type_id);
tools/lib/bpf/relo_core.c
876
if (btf_is_enum(t))
tools/lib/bpf/relo_core.c
877
*val = btf_enum(t)[spec->spec[0].idx].val;
tools/lib/bpf/relo_core.c
879
*val = btf_enum64_value(btf_enum64(t) + spec->spec[0].idx);
tools/mm/slabinfo.c
1102
struct slabinfo t;
tools/mm/slabinfo.c
1104
memcpy(&t, s1, sizeof(struct slabinfo));
tools/mm/slabinfo.c
1106
memcpy(s2, &t, sizeof(struct slabinfo));
tools/mm/slabinfo.c
1127
struct aliasinfo t;
tools/mm/slabinfo.c
1129
memcpy(&t, a1, sizeof(struct aliasinfo));
tools/mm/slabinfo.c
1131
memcpy(a2, &t, sizeof(struct aliasinfo));
tools/mm/slabinfo.c
1221
char *t;
tools/mm/slabinfo.c
1270
slab->partial = get_obj_and_str("partial", &t);
tools/mm/slabinfo.c
1271
decode_numa_list(slab->numa_partial, t);
tools/mm/slabinfo.c
1272
free(t);
tools/mm/slabinfo.c
1278
slab->slabs = get_obj_and_str("slabs", &t);
tools/mm/slabinfo.c
1279
decode_numa_list(slab->numa, t);
tools/mm/slabinfo.c
1280
free(t);
tools/mm/slabinfo.c
297
static void decode_numa_list(int *numa, char *t)
tools/mm/slabinfo.c
304
if (!t)
tools/mm/slabinfo.c
307
while (*t == 'N') {
tools/mm/slabinfo.c
308
t++;
tools/mm/slabinfo.c
309
node = strtoul(t, &t, 10);
tools/mm/slabinfo.c
310
if (*t == '=') {
tools/mm/slabinfo.c
311
t++;
tools/mm/slabinfo.c
312
nr = strtoul(t, &t, 10);
tools/mm/slabinfo.c
317
while (*t == ' ')
tools/mm/slabinfo.c
318
t++;
tools/objtool/elf.c
738
Elf_Scn *s, *t = NULL;
tools/objtool/elf.c
751
t = elf_getscn(elf->elf, symtab_shndx->idx);
tools/objtool/elf.c
752
if (!t) {
tools/objtool/elf.c
761
if (t)
tools/objtool/elf.c
762
shndx_data = elf_getdata(t, shndx_data);
tools/objtool/elf.c
782
if (t)
tools/objtool/elf.c
783
shndx_data = elf_newdata(t);
tools/objtool/elf.c
799
if (t) {
tools/perf/arch/arm64/util/mem-events.c
6
#define E(t, n, s, l, a) { .tag = t, .name = n, .event_name = s, .ldlat = l, .aux_event = a }
tools/perf/arch/powerpc/util/mem-events.c
6
#define E(t, n, s, l, a) { .tag = t, .name = n, .event_name = s, .ldlat = l, .aux_event = a }
tools/perf/arch/x86/util/mem-events.c
10
#define E(t, n, s, l, a) { .tag = t, .name = n, .event_name = s, .ldlat = l, .aux_event = a }
tools/perf/bench/epoll-ctl.c
403
unsigned long t[EPOLL_NR_OPS];
tools/perf/bench/epoll-ctl.c
406
t[j] = worker[i].ops[j];
tools/perf/bench/epoll-ctl.c
407
update_stats(&all_stats[j], t[j]);
tools/perf/bench/epoll-ctl.c
413
t[OP_EPOLL_ADD], t[OP_EPOLL_MOD], t[OP_EPOLL_DEL]);
tools/perf/bench/epoll-ctl.c
418
t[OP_EPOLL_ADD], t[OP_EPOLL_MOD], t[OP_EPOLL_DEL]);
tools/perf/bench/epoll-wait.c
540
unsigned long t = bench__runtime.tv_sec > 0 ?
tools/perf/bench/epoll-wait.c
543
update_stats(&throughput_stats, t);
tools/perf/bench/epoll-wait.c
547
worker[i].tid, &worker[i].fdmap[0], t);
tools/perf/bench/epoll-wait.c
551
&worker[i].fdmap[nfds-1], t);
tools/perf/bench/futex-hash.c
233
unsigned long t = bench__runtime.tv_sec > 0 ?
tools/perf/bench/futex-hash.c
235
update_stats(&throughput_stats, t);
tools/perf/bench/futex-hash.c
239
worker[i].tid, &worker[i].futex[0], t);
tools/perf/bench/futex-hash.c
243
&worker[i].futex[params.nfutexes-1], t);
tools/perf/bench/futex-lock-pi.c
237
unsigned long t = bench__runtime.tv_sec > 0 ?
tools/perf/bench/futex-lock-pi.c
240
update_stats(&throughput_stats, t);
tools/perf/bench/futex-lock-pi.c
243
worker[i].tid, worker[i].futex, t);
tools/perf/bench/mem-functions.c
147
static void clock_get(union bench_clock *t)
tools/perf/bench/mem-functions.c
150
t->cycles = get_cycles();
tools/perf/bench/mem-functions.c
152
BUG_ON(gettimeofday(&t->tv, NULL));
tools/perf/bench/mem-functions.c
157
union bench_clock t;
tools/perf/bench/mem-functions.c
160
t.cycles = e->cycles - s->cycles;
tools/perf/bench/mem-functions.c
162
timersub(&e->tv, &s->tv, &t.tv);
tools/perf/bench/mem-functions.c
164
return t;
tools/perf/bench/numa.c
1024
int t, p;
tools/perf/bench/numa.c
1027
for (t = 0; t < g->p.nr_threads; t++) {
tools/perf/bench/numa.c
1032
task_nr = p*g->p.nr_threads + t;
tools/perf/bench/numa.c
1088
int t;
tools/perf/bench/numa.c
1101
for (t = 0; t < g->p.nr_tasks; t++) {
tools/perf/bench/numa.c
1102
struct thread_data *td = g->threads + t;
tools/perf/bench/numa.c
1411
int t;
tools/perf/bench/numa.c
1434
for (t = 0; t < g->p.nr_threads; t++) {
tools/perf/bench/numa.c
1435
task_nr = process_nr*g->p.nr_threads + t;
tools/perf/bench/numa.c
1440
td->thread_nr = t;
tools/perf/bench/numa.c
1446
ret = pthread_create(pthreads + t, NULL, worker_thread, td);
tools/perf/bench/numa.c
1450
for (t = 0; t < g->p.nr_threads; t++) {
tools/perf/bench/numa.c
1451
ret = pthread_join(pthreads[t], NULL);
tools/perf/bench/numa.c
1482
int t;
tools/perf/bench/numa.c
1486
for (t = 0; t < g->p.nr_tasks; t++) {
tools/perf/bench/numa.c
1487
struct thread_data *td = g->threads + t;
tools/perf/bench/numa.c
1506
int t;
tools/perf/bench/numa.c
1509
for (t = 0; t < g->p.nr_tasks; t++) {
tools/perf/bench/numa.c
1510
struct thread_data *td = g->threads + t;
tools/perf/bench/numa.c
1627
int i, t, p;
tools/perf/bench/numa.c
1714
for (t = 0; t < g->p.nr_tasks; t++) {
tools/perf/bench/numa.c
1715
u64 thread_runtime_ns = g->threads[t].runtime_ns;
tools/perf/bench/numa.c
1775
for (t = 0; t < g->p.nr_threads; t++) {
tools/perf/bench/numa.c
1777
td = g->threads + p*g->p.nr_threads + t;
tools/perf/bench/numa.c
1778
snprintf(tname, sizeof(tname), "process%d:thread%d", p, t);
tools/perf/bench/numa.c
593
int t;
tools/perf/bench/numa.c
601
t = 0;
tools/perf/bench/numa.c
680
if (t >= g->p.nr_tasks) {
tools/perf/bench/numa.c
684
td = g->threads + t;
tools/perf/bench/numa.c
686
if (t)
tools/perf/bench/numa.c
704
t++;
tools/perf/bench/numa.c
712
if (t < g->p.nr_tasks)
tools/perf/bench/numa.c
713
printf("# NOTE: %d tasks bound, %d tasks unbound\n", t, g->p.nr_tasks - t);
tools/perf/bench/numa.c
741
int t;
tools/perf/bench/numa.c
749
t = 0;
tools/perf/bench/numa.c
807
if (t >= g->p.nr_tasks || !node_has_cpus(bind_node)) {
tools/perf/bench/numa.c
811
td = g->threads + t;
tools/perf/bench/numa.c
813
if (!t)
tools/perf/bench/numa.c
819
t++;
tools/perf/bench/numa.c
827
if (t < g->p.nr_tasks)
tools/perf/bench/numa.c
828
printf("# NOTE: %d tasks mem-bound, %d tasks unbound\n", t, g->p.nr_tasks - t);
tools/perf/bench/numa.c
981
int n, t;
tools/perf/bench/numa.c
988
for (t = 0; t < g->p.nr_threads; t++) {
tools/perf/bench/numa.c
993
task_nr = process_nr*g->p.nr_threads + t;
tools/perf/bench/sched-pipe.c
224
int t;
tools/perf/bench/sched-pipe.c
244
for (t = 0; t < nr_threads; t++) {
tools/perf/bench/sched-pipe.c
245
td = threads + t;
tools/perf/bench/sched-pipe.c
247
td->nr = t;
tools/perf/bench/sched-pipe.c
249
if (t == 0) {
tools/perf/bench/sched-pipe.c
259
for (t = 0; t < nr_threads; t++) {
tools/perf/bench/sched-pipe.c
260
td = threads + t;
tools/perf/bench/sched-pipe.c
266
for (t = 0; t < nr_threads; t++) {
tools/perf/bench/sched-pipe.c
267
td = threads + t;
tools/perf/builtin-lock.c
1305
struct thread *t;
tools/perf/builtin-lock.c
1308
t = perf_session__findnew(session, st->addr);
tools/perf/builtin-lock.c
1309
name = thread__comm_str(t);
tools/perf/builtin-lock.c
1342
struct thread *t;
tools/perf/builtin-lock.c
1349
t = perf_session__findnew(session, st->tid);
tools/perf/builtin-lock.c
1350
fprintf(lock_output, "%10d: %s\n", st->tid, thread__comm_str(t));
tools/perf/builtin-lock.c
1352
thread__put(t);
tools/perf/builtin-lock.c
1630
struct thread *t;
tools/perf/builtin-lock.c
1644
t = perf_session__findnew(session, pid);
tools/perf/builtin-lock.c
1646
pid, pid == -1 ? "Unknown" : thread__comm_str(t));
tools/perf/builtin-lock.c
1681
struct thread *t;
tools/perf/builtin-lock.c
1697
t = perf_session__findnew(session, pid);
tools/perf/builtin-lock.c
1699
pid == -1 ? "Unknown" : thread__comm_str(t));
tools/perf/builtin-record.c
1138
int t;
tools/perf/builtin-record.c
1144
for (t = 0; t < rec->nr_threads; t++) {
tools/perf/builtin-record.c
1145
record__thread_data_close_pipes(&thread_data[t]);
tools/perf/builtin-record.c
1146
zfree(&thread_data[t].maps);
tools/perf/builtin-record.c
1147
zfree(&thread_data[t].overwrite_maps);
tools/perf/builtin-record.c
1148
fdarray__exit(&thread_data[t].pollfd);
tools/perf/builtin-record.c
1220
int t, ret;
tools/perf/builtin-record.c
1230
for (t = 0; t < rec->nr_threads; t++)
tools/perf/builtin-record.c
1231
record__thread_data_init_pipes(&thread_data[t]);
tools/perf/builtin-record.c
1233
for (t = 0; t < rec->nr_threads; t++) {
tools/perf/builtin-record.c
1234
thread_data[t].rec = rec;
tools/perf/builtin-record.c
1235
thread_data[t].mask = &rec->thread_masks[t];
tools/perf/builtin-record.c
1236
ret = record__thread_data_init_maps(&thread_data[t], evlist);
tools/perf/builtin-record.c
1238
pr_err("Failed to initialize thread[%d] maps\n", t);
tools/perf/builtin-record.c
1241
ret = record__thread_data_init_pollfd(&thread_data[t], evlist);
tools/perf/builtin-record.c
1243
pr_err("Failed to initialize thread[%d] pollfd\n", t);
tools/perf/builtin-record.c
1246
if (t) {
tools/perf/builtin-record.c
1247
thread_data[t].tid = -1;
tools/perf/builtin-record.c
1248
ret = record__thread_data_open_pipes(&thread_data[t]);
tools/perf/builtin-record.c
1250
pr_err("Failed to open thread[%d] communication pipes\n", t);
tools/perf/builtin-record.c
1253
ret = fdarray__add(&thread_data[t].pollfd, thread_data[t].pipes.msg[0],
tools/perf/builtin-record.c
1256
pr_err("Failed to add descriptor to thread[%d] pollfd\n", t);
tools/perf/builtin-record.c
1259
thread_data[t].ctlfd_pos = ret;
tools/perf/builtin-record.c
1261
thread_data, thread_data[t].ctlfd_pos,
tools/perf/builtin-record.c
1262
thread_data[t].pipes.msg[0]);
tools/perf/builtin-record.c
1264
thread_data[t].tid = gettid();
tools/perf/builtin-record.c
1266
ret = record__dup_non_perf_events(rec, evlist, &thread_data[t]);
tools/perf/builtin-record.c
1270
thread_data[t].ctlfd_pos = -1; /* Not used */
tools/perf/builtin-record.c
2326
int t, tt, err, ret = 0, nr_threads = rec->nr_threads;
tools/perf/builtin-record.c
2346
for (t = 1; t < nr_threads; t++) {
tools/perf/builtin-record.c
2351
MMAP_CPU_MASK_BYTES(&(thread_data[t].mask->affinity)),
tools/perf/builtin-record.c
2352
(cpu_set_t *)(thread_data[t].mask->affinity.bits));
tools/perf/builtin-record.c
2354
if (pthread_create(&handle, &attrs, record__thread, &thread_data[t])) {
tools/perf/builtin-record.c
2355
for (tt = 1; tt < t; tt++)
tools/perf/builtin-record.c
2356
record__terminate_thread(&thread_data[t]);
tools/perf/builtin-record.c
2362
err = read(thread_data[t].pipes.ack[0], &msg, sizeof(msg));
tools/perf/builtin-record.c
2364
pr_debug2("threads[%d]: sent %s\n", rec->thread_data[t].tid,
tools/perf/builtin-record.c
2368
thread->tid, rec->thread_data[t].tid);
tools/perf/builtin-record.c
2389
int t;
tools/perf/builtin-record.c
2392
for (t = 1; t < rec->nr_threads; t++)
tools/perf/builtin-record.c
2393
record__terminate_thread(&thread_data[t]);
tools/perf/builtin-record.c
2395
for (t = 0; t < rec->nr_threads; t++) {
tools/perf/builtin-record.c
2396
rec->samples += thread_data[t].samples;
tools/perf/builtin-record.c
2399
rec->session->bytes_transferred += thread_data[t].bytes_transferred;
tools/perf/builtin-record.c
2400
rec->session->bytes_compressed += thread_data[t].bytes_compressed;
tools/perf/builtin-record.c
2401
pr_debug("threads[%d]: samples=%lld, wakes=%ld, ", thread_data[t].tid,
tools/perf/builtin-record.c
2402
thread_data[t].samples, thread_data[t].waking);
tools/perf/builtin-record.c
2403
if (thread_data[t].bytes_transferred && thread_data[t].bytes_compressed)
tools/perf/builtin-record.c
2405
thread_data[t].bytes_transferred, thread_data[t].bytes_compressed);
tools/perf/builtin-record.c
2407
pr_debug("written=%" PRIu64 "\n", thread_data[t].bytes_written);
tools/perf/builtin-record.c
2415
int t;
tools/perf/builtin-record.c
2419
for (t = 0; t < rec->nr_threads; t++)
tools/perf/builtin-record.c
2420
waking += thread_data[t].waking;
tools/perf/builtin-record.c
3733
int t;
tools/perf/builtin-record.c
3736
for (t = 0; t < nr_threads; t++)
tools/perf/builtin-record.c
3737
record__thread_mask_free(&rec->thread_masks[t]);
tools/perf/builtin-record.c
3744
int t, ret;
tools/perf/builtin-record.c
3752
for (t = 0; t < nr_threads; t++) {
tools/perf/builtin-record.c
3753
ret = record__thread_mask_alloc(&rec->thread_masks[t], nr_bits);
tools/perf/builtin-record.c
3755
pr_err("Failed to allocate thread masks[%d]\n", t);
tools/perf/builtin-record.c
3770
int t, ret, nr_cpus = perf_cpu_map__nr(cpus);
tools/perf/builtin-record.c
3779
for (t = 0; t < rec->nr_threads; t++) {
tools/perf/builtin-record.c
3780
__set_bit(perf_cpu_map__cpu(cpus, t).cpu, rec->thread_masks[t].maps.bits);
tools/perf/builtin-record.c
3781
__set_bit(perf_cpu_map__cpu(cpus, t).cpu, rec->thread_masks[t].affinity.bits);
tools/perf/builtin-record.c
3783
pr_debug("thread_masks[%d]: ", t);
tools/perf/builtin-record.c
3784
mmap_cpu_mask__scnprintf(&rec->thread_masks[t].maps, "maps");
tools/perf/builtin-record.c
3785
pr_debug("thread_masks[%d]: ", t);
tools/perf/builtin-record.c
3786
mmap_cpu_mask__scnprintf(&rec->thread_masks[t].affinity, "affinity");
tools/perf/builtin-record.c
3798
int ret = 0, t = 0;
tools/perf/builtin-record.c
3871
thread_masks = realloc(rec->thread_masks, (t + 1) * sizeof(struct thread_mask));
tools/perf/builtin-record.c
3878
rec->thread_masks[t] = thread_mask;
tools/perf/builtin-record.c
3880
pr_debug("thread_masks[%d]: ", t);
tools/perf/builtin-record.c
3881
mmap_cpu_mask__scnprintf(&rec->thread_masks[t].maps, "maps");
tools/perf/builtin-record.c
3882
pr_debug("thread_masks[%d]: ", t);
tools/perf/builtin-record.c
3883
mmap_cpu_mask__scnprintf(&rec->thread_masks[t].affinity, "affinity");
tools/perf/builtin-record.c
3885
t++;
tools/perf/builtin-record.c
3892
rec->nr_threads = t;
tools/perf/builtin-record.c
3977
int t, ret;
tools/perf/builtin-record.c
3982
for (t = 0, user_spec = (char *)rec->opts.threads_user_spec; ; t++, user_spec = NULL) {
tools/perf/builtin-record.c
3986
pr_debug2("threads_spec[%d]: %s\n", t, spec);
tools/perf/builtin-sched.c
2168
u64 t, const char state)
tools/perf/builtin-sched.c
2181
timestamp__scnprintf_usec(t, tstr, sizeof(tstr));
tools/perf/builtin-sched.c
2272
u64 t, u64 tprev)
tools/perf/builtin-sched.c
2282
r->dt_run = t - tprev;
tools/perf/builtin-sched.c
2799
u64 tprev, t = sample->time;
tools/perf/builtin-sched.c
2836
if (ptime->start && ptime->start > t)
tools/perf/builtin-sched.c
2855
if (t > ptime->end)
tools/perf/builtin-sched.c
2856
t = ptime->end;
tools/perf/builtin-sched.c
2861
timehist_update_runtime_stats(tr, t, tprev);
tools/perf/builtin-sched.c
2875
timehist_update_runtime_stats(last_tr, t, tprev);
tools/perf/builtin-sched.c
2888
callchain_append(&itr->callchain, &itr->cursor, t - tprev);
tools/perf/builtin-sched.c
2894
timehist_print_sample(sched, evsel, sample, &al, thread, t, state);
tools/perf/builtin-sched.c
2898
if (sched->hist_time.start == 0 && t >= ptime->start)
tools/perf/builtin-sched.c
2899
sched->hist_time.start = t;
tools/perf/builtin-sched.c
2900
if (ptime->end == 0 || t <= ptime->end)
tools/perf/builtin-sched.c
2901
sched->hist_time.end = t;
tools/perf/builtin-sched.c
2912
tr->ready_to_run = t;
tools/perf/builtin-sched.c
2950
static void print_thread_runtime(struct thread *t,
tools/perf/builtin-sched.c
2957
comm_width, timehist_get_commstr(t), thread__ppid(t),
tools/perf/builtin-sched.c
2973
static void print_thread_waittime(struct thread *t,
tools/perf/builtin-sched.c
2977
comm_width, timehist_get_commstr(t), thread__ppid(t),
tools/perf/builtin-sched.c
2998
static int show_thread_runtime(struct thread *t, void *priv)
tools/perf/builtin-sched.c
3003
if (thread__is_filtered(t))
tools/perf/builtin-sched.c
3006
r = thread__priv(t);
tools/perf/builtin-sched.c
3013
print_thread_waittime(t, r);
tools/perf/builtin-sched.c
3015
print_thread_runtime(t, r);
tools/perf/builtin-sched.c
3081
struct thread *t;
tools/perf/builtin-sched.c
3121
t = idle_threads[i];
tools/perf/builtin-sched.c
3122
if (!t)
tools/perf/builtin-sched.c
3125
r = thread__priv(t);
tools/perf/builtin-sched.c
3145
t = idle_threads[i];
tools/perf/builtin-sched.c
3146
if (!t)
tools/perf/builtin-sched.c
3149
itr = thread__priv(t);
tools/perf/builtin-script.c
922
u64 t = sample->time;
tools/perf/builtin-script.c
926
t = sample->time - initial_time;
tools/perf/builtin-script.c
929
t = sample->time - previous_time;
tools/perf/builtin-script.c
931
t = 0;
tools/perf/builtin-script.c
935
nsecs = t;
tools/perf/builtin-script.c
943
timestamp__scnprintf_usec(t, sample_time, sizeof(sample_time));
tools/perf/builtin-top.c
281
static void perf_top__resort_hists(struct perf_top *t)
tools/perf/builtin-top.c
283
struct evlist *evlist = t->evlist;
tools/perf/builtin-top.c
296
if (t->zero) {
tools/perf/builtin-top.c
299
hists__decay_entries(hists, t->hide_user_symbols,
tools/perf/builtin-top.c
300
t->hide_kernel_symbols);
tools/perf/builtin-top.c
598
struct perf_top *t = arg;
tools/perf/builtin-top.c
600
if (t->evlist->selected != NULL)
tools/perf/builtin-top.c
601
t->sym_evsel = t->evlist->selected;
tools/perf/builtin-top.c
603
perf_top__resort_hists(t);
tools/perf/builtin-top.c
605
if (t->lost || t->drop)
tools/perf/builtin-trace.c
1510
static size_t fprintf_duration(unsigned long t, bool calculated, FILE *fp)
tools/perf/builtin-trace.c
1512
double duration = (double)t / NSEC_PER_MSEC;
tools/perf/builtin-trace.c
1869
static bool trace__filter_duration(struct trace *trace, double t)
tools/perf/builtin-trace.c
1871
return t < (trace->duration_filter * NSEC_PER_MSEC);
tools/perf/jvmti/jvmti_agent.c
135
time_t t;
tools/perf/jvmti/jvmti_agent.c
138
time(&t);
tools/perf/jvmti/jvmti_agent.c
139
localtime_r(&t, &tm);
tools/perf/tests/builtin-test.c
234
static int test_suite__num_test_cases(const struct test_suite *t)
tools/perf/tests/builtin-test.c
238
test_suite__for_each_test_case(t, num);
tools/perf/tests/builtin-test.c
243
static const char *skip_reason(const struct test_suite *t, int test_case)
tools/perf/tests/builtin-test.c
245
if (!t->test_cases)
tools/perf/tests/builtin-test.c
248
return t->test_cases[test_case >= 0 ? test_case : 0].skip_reason;
tools/perf/tests/builtin-test.c
251
static const char *test_description(const struct test_suite *t, int test_case)
tools/perf/tests/builtin-test.c
253
if (t->test_cases && test_case >= 0)
tools/perf/tests/builtin-test.c
254
return t->test_cases[test_case].desc;
tools/perf/tests/builtin-test.c
256
return t->desc;
tools/perf/tests/builtin-test.c
259
static test_fnptr test_function(const struct test_suite *t, int test_case)
tools/perf/tests/builtin-test.c
262
return t->test_cases[0].run_case;
tools/perf/tests/builtin-test.c
264
return t->test_cases[test_case].run_case;
tools/perf/tests/builtin-test.c
267
static bool test_exclusive(const struct test_suite *t, int test_case)
tools/perf/tests/builtin-test.c
270
return t->test_cases[0].exclusive;
tools/perf/tests/builtin-test.c
272
return t->test_cases[test_case].exclusive;
tools/perf/tests/builtin-test.c
358
static int print_test_result(struct test_suite *t, int curr_suite, int curr_test_case,
tools/perf/tests/builtin-test.c
361
if (test_suite__num_test_cases(t) > 1) {
tools/perf/tests/builtin-test.c
365
test_description(t, curr_test_case));
tools/perf/tests/builtin-test.c
367
pr_info("%3d: %-*s:", curr_suite + 1, width, test_description(t, curr_test_case));
tools/perf/tests/builtin-test.c
377
const char *reason = skip_reason(t, curr_test_case);
tools/perf/tests/builtin-test.c
398
struct test_suite *t;
tools/perf/tests/builtin-test.c
409
t = child_test->test;
tools/perf/tests/builtin-test.c
417
if (test_suite__num_test_cases(t) > 1 && curr_test_case == 0)
tools/perf/tests/builtin-test.c
418
pr_info("%3d: %-*s:\n", curr_suite + 1, width, test_description(t, -1));
tools/perf/tests/builtin-test.c
427
if (test_suite__num_test_cases(t) > 1)
tools/perf/tests/builtin-test.c
429
test_description(t, curr_test_case));
tools/perf/tests/builtin-test.c
431
pr_info("%3d: %s:\n", curr_suite + 1, test_description(t, -1));
tools/perf/tests/builtin-test.c
456
print_test_result(t, curr_suite, curr_test_case, TEST_RUNNING,
tools/perf/tests/builtin-test.c
495
print_test_result(t, curr_suite, curr_test_case, ret, width, /*running=*/0);
tools/perf/tests/builtin-test.c
570
for (struct test_suite **t = suites; *t; t++) {
tools/perf/tests/builtin-test.c
571
int i, len = strlen(test_description(*t, -1));
tools/perf/tests/builtin-test.c
576
test_suite__for_each_test_case(*t, i) {
tools/perf/tests/builtin-test.c
577
len = strlen(test_description(*t, i));
tools/perf/tests/builtin-test.c
616
for (struct test_suite **t = suites; *t; t++, curr_suite++) {
tools/perf/tests/builtin-test.c
620
if (!perf_test__matches(test_description(*t, -1), curr_suite, argc, argv)) {
tools/perf/tests/builtin-test.c
627
test_suite__for_each_test_case(*t, curr_test_case) {
tools/perf/tests/builtin-test.c
628
if (perf_test__matches(test_description(*t, curr_test_case),
tools/perf/tests/builtin-test.c
642
test_description(*t, -1));
tools/perf/tests/builtin-test.c
648
test_suite__for_each_test_case(*t, curr_test_case) {
tools/perf/tests/builtin-test.c
650
!perf_test__matches(test_description(*t, curr_test_case),
tools/perf/tests/builtin-test.c
653
err = start_test(*t, curr_suite, curr_test_case,
tools/perf/tests/builtin-test.c
683
for (struct test_suite **t = suites; *t; t++, curr_suite++) {
tools/perf/tests/builtin-test.c
686
if (!perf_test__matches(test_description(*t, -1), curr_suite, argc, argv))
tools/perf/tests/builtin-test.c
689
fprintf(fp, "%3d: %s\n", curr_suite + 1, test_description(*t, -1));
tools/perf/tests/builtin-test.c
691
if (test_suite__num_test_cases(*t) <= 1)
tools/perf/tests/builtin-test.c
694
test_suite__for_each_test_case(*t, curr_test_case) {
tools/perf/tests/builtin-test.c
696
test_description(*t, curr_test_case));
tools/perf/tests/builtin-test.c
747
struct test_suite *t;
tools/perf/tests/builtin-test.c
757
for_each_suite(t)
tools/perf/tests/builtin-test.c
763
for_each_suite(t) {
tools/perf/tests/builtin-test.c
767
test_suite__for_each_test_case(t, curr_test_case) {
tools/perf/tests/builtin-test.c
768
if (test_exclusive(t, curr_test_case)) {
tools/perf/tests/builtin-test.c
774
result[n++] = t;
tools/perf/tests/expr.c
71
static int test__expr(struct test_suite *t __maybe_unused, int subtest __maybe_unused)
tools/perf/tests/hists_link.c
148
struct thread *t, struct map *m, struct symbol *s)
tools/perf/tests/hists_link.c
151
if (RC_CHK_EQUAL(samples->thread, t) &&
tools/perf/tests/is_printable_array.c
16
} t[] = {
tools/perf/tests/is_printable_array.c
27
for (i = 0; i < ARRAY_SIZE(t); i++) {
tools/perf/tests/is_printable_array.c
30
ret = is_printable_array((char *) t[i].buf, t[i].len);
tools/perf/tests/is_printable_array.c
31
if (ret != t[i].ret) {
tools/perf/tests/kmod-path.c
50
static int test__kmod_path__parse(struct test_suite *t __maybe_unused, int subtest __maybe_unused)
tools/perf/tests/maps.c
165
static int test__maps__fixup_overlap_and_insert(struct test_suite *t __maybe_unused,
tools/perf/tests/maps.c
74
static int test__maps__merge_in(struct test_suite *t __maybe_unused, int subtest __maybe_unused)
tools/perf/tests/mem2node.c
47
static int test__mem2node(struct test_suite *t __maybe_unused, int subtest __maybe_unused)
tools/perf/tests/parse-events.c
2643
static int test_term(const struct terms_test *t)
tools/perf/tests/parse-events.c
2650
ret = parse_events_terms(&terms, t->str);
tools/perf/tests/parse-events.c
2653
t->str , ret);
tools/perf/tests/parse-events.c
2657
ret = t->check(&terms);
tools/perf/tests/parse-events.c
2668
const struct terms_test *t = &terms[i];
tools/perf/tests/parse-events.c
2670
pr_debug("running test %d '%s'\n", i, t->str);
tools/perf/tests/parse-events.c
2671
ret = test_term(t);
tools/perf/tests/shell/coresight/memcpy_thread/memcpy_thread.c
36
pthread_t t;
tools/perf/tests/shell/coresight/memcpy_thread/memcpy_thread.c
40
pthread_create(&t, &attr, fn, arg);
tools/perf/tests/shell/coresight/memcpy_thread/memcpy_thread.c
41
return t;
tools/perf/tests/shell/coresight/thread_loop/thread_loop.c
49
pthread_t t;
tools/perf/tests/shell/coresight/thread_loop/thread_loop.c
53
pthread_create(&t, &attr, fn, arg);
tools/perf/tests/shell/coresight/thread_loop/thread_loop.c
54
return t;
tools/perf/tests/shell/coresight/unroll_loop_thread/unroll_loop_thread.c
45
pthread_t t;
tools/perf/tests/shell/coresight/unroll_loop_thread/unroll_loop_thread.c
49
pthread_create(&t, &attr, fn, arg);
tools/perf/tests/shell/coresight/unroll_loop_thread/unroll_loop_thread.c
50
return t;
tools/perf/tests/time-utils-test.c
134
static int test__time_utils(struct test_suite *t __maybe_unused, int subtest __maybe_unused)
tools/perf/tests/unit_number__scnprintf.c
10
static int test__unit_number__scnprint(struct test_suite *t __maybe_unused, int subtest __maybe_unused)
tools/perf/tests/util.c
89
static int test__util(struct test_suite *t __maybe_unused, int subtest __maybe_unused)
tools/perf/ui/tui/util.c
171
const char *t;
tools/perf/ui/tui/util.c
173
t = text;
tools/perf/ui/tui/util.c
175
const char *sep = strchr(t, '\n');
tools/perf/ui/tui/util.c
179
sep = strchr(t, '\0');
tools/perf/ui/tui/util.c
180
len = sep - t;
tools/perf/ui/tui/util.c
186
t = sep + 1;
tools/perf/ui/tui/util.c
81
const char *t;
tools/perf/ui/tui/util.c
83
t = text;
tools/perf/ui/tui/util.c
85
const char *sep = strchr(t, '\n');
tools/perf/ui/tui/util.c
88
sep = strchr(t, '\0');
tools/perf/ui/tui/util.c
89
len = sep - t;
tools/perf/ui/tui/util.c
95
t = sep + 1;
tools/perf/util/blake2s.c
45
ctx->t[0] += inc;
tools/perf/util/blake2s.c
46
ctx->t[1] += (ctx->t[0] < inc);
tools/perf/util/blake2s.c
65
v[12] = BLAKE2S_IV4 ^ ctx->t[0];
tools/perf/util/blake2s.c
66
v[13] = BLAKE2S_IV5 ^ ctx->t[1];
tools/perf/util/blake2s.h
16
u32 t[2];
tools/perf/util/blake2s.h
45
ctx->t[0] = 0;
tools/perf/util/blake2s.h
46
ctx->t[1] = 0;
tools/perf/util/bpf-event.c
140
const struct btf_type *t;
tools/perf/util/bpf-event.c
148
t = btf__type_by_id(btf, finfo->type_id);
tools/perf/util/bpf-event.c
149
short_name = btf__name_by_offset(btf, t->name_off);
tools/perf/util/bpf-event.c
280
const struct btf_type *t, const void *btf_data)
tools/perf/util/bpf-event.c
297
btf_size = btf__resolve_size(btf, t->type);
tools/perf/util/bpf-event.c
298
btf_dump__dump_type_data(d, t->type, btf_data, btf_size, &opts);
tools/perf/util/bpf_counter.c
125
const struct btf_type *t;
tools/perf/util/bpf_counter.c
147
t = btf__type_by_id(btf, func_info[0].type_id);
tools/perf/util/bpf_counter.c
148
if (!t) {
tools/perf/util/bpf_counter.c
153
name = strdup(btf__name_by_offset(btf, t->name_off));
tools/perf/util/bpf_lock_contention.c
579
struct thread *t = machine__findnew_thread(machine, /*pid=*/-1, pid);
tools/perf/util/bpf_lock_contention.c
581
if (t != NULL &&
tools/perf/util/bpf_lock_contention.c
583
thread__set_comm(t, task.comm, /*timestamp=*/0)) {
tools/perf/util/bpf_skel/off_cpu.bpf.c
135
static inline int get_task_state(struct task_struct *t)
tools/perf/util/bpf_skel/off_cpu.bpf.c
138
struct task_struct___new *t_new = (void *)t;
tools/perf/util/bpf_skel/off_cpu.bpf.c
144
struct task_struct___old *t_old = (void *)t;
tools/perf/util/bpf_skel/off_cpu.bpf.c
150
static inline __u64 get_cgroup_id(struct task_struct *t)
tools/perf/util/bpf_skel/off_cpu.bpf.c
155
return BPF_CORE_READ(t, cgroups, dfl_cgrp, kn, id);
tools/perf/util/bpf_skel/off_cpu.bpf.c
166
cgrp = BPF_CORE_READ(t, cgroups, subsys[perf_subsys_id], cgroup);
tools/perf/util/bpf_skel/off_cpu.bpf.c
170
static inline int can_record(struct task_struct *t, int state)
tools/perf/util/bpf_skel/off_cpu.bpf.c
173
if (t->flags & PF_KTHREAD)
tools/perf/util/bpf_skel/off_cpu.bpf.c
194
pid = t->tgid;
tools/perf/util/bpf_skel/off_cpu.bpf.c
196
pid = t->pid;
tools/perf/util/bpf_skel/off_cpu.bpf.c
205
__u64 cgrp_id = get_cgroup_id(t);
tools/perf/util/btf.c
15
const struct btf_type *t = btf__type_by_id(btf, type_id);
tools/perf/util/btf.c
19
for (i = 0, m = btf_members(t); i < btf_vlen(t); i++, m++) {
tools/perf/util/data-convert-bt.c
1017
while ((t = bt_ctf_event_class_get_field_by_name(event_class, name))) {
tools/perf/util/data-convert-bt.c
1018
bt_ctf_field_type_put(t);
tools/perf/util/data-convert-bt.c
1129
#define ADD_FIELD(cl, t, n) \
tools/perf/util/data-convert-bt.c
1132
if (bt_ctf_event_class_add_field(cl, t, n)) { \
tools/perf/util/data-convert-bt.c
1240
#define __NON_SAMPLE_ADD_FIELD(t, n) \
tools/perf/util/data-convert-bt.c
1243
if (bt_ctf_event_class_add_field(event_class, cw->data.t, #n)) {\
tools/perf/util/data-convert-bt.c
50
#define pr_time2(t, fmt, ...) pr_time_N(2, debug_data_convert, t, pr_fmt(fmt), ##__VA_ARGS__)
tools/perf/util/data-convert-bt.c
997
struct bt_ctf_field_type *t = NULL;
tools/perf/util/debug.c
120
static int veprintf_time(u64 t, const char *fmt, va_list args)
tools/perf/util/debug.c
123
u64 secs, usecs, nsecs = t;
tools/perf/util/debug.c
134
int eprintf_time(int level, int var, u64 t, const char *fmt, ...)
tools/perf/util/debug.c
141
ret = veprintf_time(t, fmt, args);
tools/perf/util/debug.h
52
#define pr_time_N(n, var, t, fmt, ...) \
tools/perf/util/debug.h
53
eprintf_time(n, var, t, fmt, ##__VA_ARGS__)
tools/perf/util/debug.h
55
#define pr_oe_time(t, fmt, ...) pr_time_N(1, debug_ordered_events, t, pr_fmt(fmt), ##__VA_ARGS__)
tools/perf/util/debug.h
56
#define pr_oe_time2(t, fmt, ...) pr_time_N(2, debug_ordered_events, t, pr_fmt(fmt), ##__VA_ARGS__)
tools/perf/util/debug.h
78
int eprintf_time(int level, int var, u64 t, const char *fmt, ...) __printf(4, 5);
tools/perf/util/demangle-rust-v0.c
322
size_t t = MIN(MAX(biased, t_min), t_max);
tools/perf/util/demangle-rust-v0.c
340
if (d < t) {
tools/perf/util/demangle-rust-v0.c
343
if (base < t || w == 0 || (base - t) > SIZE_MAX / w) {
tools/perf/util/demangle-rust-v0.c
346
w *= (base - t);
tools/perf/util/disasm.c
476
char *endptr, *name, *t;
tools/perf/util/disasm.c
490
t = strchr(name, '>');
tools/perf/util/disasm.c
491
if (t == NULL)
tools/perf/util/disasm.c
494
*t = '\0';
tools/perf/util/disasm.c
496
*t = '>';
tools/perf/util/dwarf-regs-arch/dwarf-regs-powerpc.c
13
#define PPC_RT(t) (((t) >> 21) & 0x1f)
tools/perf/util/evsel.h
392
#define evsel__match(evsel, t, c) __evsel__match(evsel, PERF_TYPE_##t, PERF_COUNT_##c)
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
299
static uint64_t multdiv(uint64_t t, uint32_t n, uint32_t d)
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
303
return (t / d) * n + ((t % d) * n) / d;
tools/perf/util/intel-tpebs.c
106
list_for_each_entry(t, &tpebs_results, nd) {
tools/perf/util/intel-tpebs.c
108
record_argv[i++] = t->event;
tools/perf/util/intel-tpebs.c
119
list_for_each_entry(t, &tpebs_results, nd)
tools/perf/util/intel-tpebs.c
120
t->started = true;
tools/perf/util/intel-tpebs.c
164
static bool should_ignore_sample(const struct perf_sample *sample, const struct tpebs_retire_lat *t)
tools/perf/util/intel-tpebs.c
173
if (t->evsel->evlist == NULL)
tools/perf/util/intel-tpebs.c
176
workload_pid = t->evsel->evlist->workload.pid;
tools/perf/util/intel-tpebs.c
180
if (!t->evsel->core.attr.inherit)
tools/perf/util/intel-tpebs.c
192
struct tpebs_retire_lat *t;
tools/perf/util/intel-tpebs.c
200
t = tpebs_retire_lat__find(evsel);
tools/perf/util/intel-tpebs.c
201
if (!t) {
tools/perf/util/intel-tpebs.c
205
if (should_ignore_sample(sample, t)) {
tools/perf/util/intel-tpebs.c
214
t->last = sample->weight3;
tools/perf/util/intel-tpebs.c
215
update_stats(&t->stats, sample->weight3);
tools/perf/util/intel-tpebs.c
416
struct tpebs_retire_lat *t;
tools/perf/util/intel-tpebs.c
426
list_for_each_entry(t, &tpebs_results, nd) {
tools/perf/util/intel-tpebs.c
427
if (t->evsel == evsel)
tools/perf/util/intel-tpebs.c
428
return t;
tools/perf/util/intel-tpebs.c
443
list_for_each_entry(t, &tpebs_results, nd) {
tools/perf/util/intel-tpebs.c
444
if ((unsigned long)t->evsel == num)
tools/perf/util/intel-tpebs.c
445
return t;
tools/perf/util/intel-tpebs.c
506
struct tpebs_retire_lat *t;
tools/perf/util/intel-tpebs.c
510
t = tpebs_retire_lat__find(evsel);
tools/perf/util/intel-tpebs.c
511
valid = t && t->started;
tools/perf/util/intel-tpebs.c
552
struct tpebs_retire_lat *t = tpebs_retire_lat__find(evsel);
tools/perf/util/intel-tpebs.c
554
list_del_init(&t->nd);
tools/perf/util/intel-tpebs.c
555
tpebs_retire_lat__delete(t);
tools/perf/util/intel-tpebs.c
564
struct tpebs_retire_lat *t;
tools/perf/util/intel-tpebs.c
578
t = tpebs_retire_lat__find(evsel);
tools/perf/util/intel-tpebs.c
584
if (t && &t->nd == tpebs_results.next) {
tools/perf/util/intel-tpebs.c
591
if (t == NULL || t->stats.n == 0) {
tools/perf/util/intel-tpebs.c
614
val = t->stats.min;
tools/perf/util/intel-tpebs.c
617
val = t->stats.max;
tools/perf/util/intel-tpebs.c
620
val = t->last;
tools/perf/util/intel-tpebs.c
624
val = rint(t->stats.mean);
tools/perf/util/intel-tpebs.c
651
struct tpebs_retire_lat *t;
tools/perf/util/intel-tpebs.c
654
t = tpebs_retire_lat__find(evsel);
tools/perf/util/intel-tpebs.c
655
if (t) {
tools/perf/util/intel-tpebs.c
656
list_del_init(&t->nd);
tools/perf/util/intel-tpebs.c
657
tpebs_retire_lat__delete(t);
tools/perf/util/intel-tpebs.c
78
struct tpebs_retire_lat *t;
tools/perf/util/intel-tpebs.c
80
list_for_each_entry(t, &tpebs_results, nd)
tools/perf/util/jitdump.c
67
#define get_jit_tool(t) (container_of(tool, struct jit_tool, tool))
tools/perf/util/mem-events.c
23
#define E(t, n, s, l, a) { .tag = t, .name = n, .event_name = s, .ldlat = l, .aux_event = a }
tools/perf/util/parse-events.c
1122
struct evsel_config_term *t;
tools/perf/util/parse-events.c
1124
t = zalloc(sizeof(*t));
tools/perf/util/parse-events.c
1125
if (!t)
tools/perf/util/parse-events.c
1128
INIT_LIST_HEAD(&t->list);
tools/perf/util/parse-events.c
1129
t->type = type;
tools/perf/util/parse-events.c
1130
t->weak = weak;
tools/perf/util/parse-events.c
1141
t->val.val = val;
tools/perf/util/parse-events.c
1144
t->val.time = val;
tools/perf/util/parse-events.c
1147
t->val.inherit = val;
tools/perf/util/parse-events.c
1150
t->val.overwrite = val;
tools/perf/util/parse-events.c
1153
t->val.max_stack = val;
tools/perf/util/parse-events.c
1156
t->val.max_events = val;
tools/perf/util/parse-events.c
1159
t->val.percore = val;
tools/perf/util/parse-events.c
1162
t->val.aux_output = val;
tools/perf/util/parse-events.c
1165
t->val.aux_sample_size = val;
tools/perf/util/parse-events.c
1173
t->val.str = strdup(str);
tools/perf/util/parse-events.c
1174
if (!t->val.str) {
tools/perf/util/parse-events.c
1175
zfree(&t);
tools/perf/util/parse-events.c
1178
t->free_str = true;
tools/perf/util/parse-events.c
1182
t->val.val = val;
tools/perf/util/parse-events.c
1186
list_add_tail(&t->list, head_terms);
tools/perf/util/parse-events.c
1187
return t;
tools/perf/util/pmu.c
1444
struct parse_events_term *t;
tools/perf/util/pmu.c
1446
list_for_each_entry(t, &head_terms->terms, list) {
tools/perf/util/pmu.c
1447
if (t->type_val == PARSE_EVENTS__TERM_TYPE_NUM &&
tools/perf/util/pmu.c
1448
t->config && !strcmp(t->config, term->config)) {
tools/perf/util/pmu.c
1449
t->used = true;
tools/perf/util/pmu.c
1450
*value = t->val.num;
tools/perf/util/s390-cpumsf-kernel.h
50
unsigned int t:1; /* 2 - Timestamp format */
tools/perf/util/s390-cpumsf.c
333
local.t = flags >> 61 & 0x1;
tools/perf/util/s390-cpumsf.c
352
te->t ? 'T' : ' ',
tools/perf/util/scripting-engines/trace-event-python.c
1003
PyTuple_SetItem(t, n++, _PyLong_FromLong(cpu));
tools/perf/util/scripting-engines/trace-event-python.c
1004
PyTuple_SetItem(t, n++, _PyLong_FromLong(s));
tools/perf/util/scripting-engines/trace-event-python.c
1005
PyTuple_SetItem(t, n++, _PyLong_FromLong(ns));
tools/perf/util/scripting-engines/trace-event-python.c
1006
PyTuple_SetItem(t, n++, _PyLong_FromLong(pid));
tools/perf/util/scripting-engines/trace-event-python.c
1007
PyTuple_SetItem(t, n++, _PyUnicode_FromString(comm));
tools/perf/util/scripting-engines/trace-event-python.c
1008
PyTuple_SetItem(t, n++, callchain);
tools/perf/util/scripting-engines/trace-event-python.c
1044
PyTuple_SetItem(t, n++, obj);
tools/perf/util/scripting-engines/trace-event-python.c
1051
PyTuple_SetItem(t, n++, dict);
tools/perf/util/scripting-engines/trace-event-python.c
1056
PyTuple_SetItem(t, n++, all_entries_dict);
tools/perf/util/scripting-engines/trace-event-python.c
1061
if (_PyTuple_Resize(&t, n) == -1)
tools/perf/util/scripting-engines/trace-event-python.c
1065
call_object(handler, t, handler_name);
tools/perf/util/scripting-engines/trace-event-python.c
1067
call_object(handler, t, default_handler_name);
tools/perf/util/scripting-engines/trace-event-python.c
1069
Py_DECREF(t);
tools/perf/util/scripting-engines/trace-event-python.c
1084
PyObject *t;
tools/perf/util/scripting-engines/trace-event-python.c
1086
t = PyTuple_New(sz);
tools/perf/util/scripting-engines/trace-event-python.c
1087
if (!t)
tools/perf/util/scripting-engines/trace-event-python.c
1089
return t;
tools/perf/util/scripting-engines/trace-event-python.c
1092
static int tuple_set_s64(PyObject *t, unsigned int pos, s64 val)
tools/perf/util/scripting-engines/trace-event-python.c
1095
return PyTuple_SetItem(t, pos, _PyLong_FromLong(val));
tools/perf/util/scripting-engines/trace-event-python.c
1098
return PyTuple_SetItem(t, pos, PyLong_FromLongLong(val));
tools/perf/util/scripting-engines/trace-event-python.c
1108
static int tuple_set_u64(PyObject *t, unsigned int pos, u64 val)
tools/perf/util/scripting-engines/trace-event-python.c
1111
return PyTuple_SetItem(t, pos, PyLong_FromUnsignedLong(val));
tools/perf/util/scripting-engines/trace-event-python.c
1114
return PyTuple_SetItem(t, pos, PyLong_FromUnsignedLongLong(val));
tools/perf/util/scripting-engines/trace-event-python.c
1118
static int tuple_set_u32(PyObject *t, unsigned int pos, u32 val)
tools/perf/util/scripting-engines/trace-event-python.c
1120
return PyTuple_SetItem(t, pos, PyLong_FromUnsignedLong(val));
tools/perf/util/scripting-engines/trace-event-python.c
1123
static int tuple_set_s32(PyObject *t, unsigned int pos, s32 val)
tools/perf/util/scripting-engines/trace-event-python.c
1125
return PyTuple_SetItem(t, pos, _PyLong_FromLong(val));
tools/perf/util/scripting-engines/trace-event-python.c
1128
static int tuple_set_bool(PyObject *t, unsigned int pos, bool val)
tools/perf/util/scripting-engines/trace-event-python.c
1130
return PyTuple_SetItem(t, pos, PyBool_FromLong(val));
tools/perf/util/scripting-engines/trace-event-python.c
1133
static int tuple_set_string(PyObject *t, unsigned int pos, const char *s)
tools/perf/util/scripting-engines/trace-event-python.c
1135
return PyTuple_SetItem(t, pos, _PyUnicode_FromString(s));
tools/perf/util/scripting-engines/trace-event-python.c
1138
static int tuple_set_bytes(PyObject *t, unsigned int pos, void *bytes,
tools/perf/util/scripting-engines/trace-event-python.c
1141
return PyTuple_SetItem(t, pos, _PyBytes_FromStringAndSize(bytes, sz));
tools/perf/util/scripting-engines/trace-event-python.c
1147
PyObject *t;
tools/perf/util/scripting-engines/trace-event-python.c
1149
t = tuple_new(2);
tools/perf/util/scripting-engines/trace-event-python.c
1151
tuple_set_d64(t, 0, evsel->db_id);
tools/perf/util/scripting-engines/trace-event-python.c
1152
tuple_set_string(t, 1, evsel__name(evsel));
tools/perf/util/scripting-engines/trace-event-python.c
1154
call_object(tables->evsel_handler, t, "evsel_table");
tools/perf/util/scripting-engines/trace-event-python.c
1156
Py_DECREF(t);
tools/perf/util/scripting-engines/trace-event-python.c
1165
PyObject *t;
tools/perf/util/scripting-engines/trace-event-python.c
1167
t = tuple_new(3);
tools/perf/util/scripting-engines/trace-event-python.c
1169
tuple_set_d64(t, 0, machine->db_id);
tools/perf/util/scripting-engines/trace-event-python.c
1170
tuple_set_s32(t, 1, machine->pid);
tools/perf/util/scripting-engines/trace-event-python.c
1171
tuple_set_string(t, 2, machine->root_dir ? machine->root_dir : "");
tools/perf/util/scripting-engines/trace-event-python.c
1173
call_object(tables->machine_handler, t, "machine_table");
tools/perf/util/scripting-engines/trace-event-python.c
1175
Py_DECREF(t);
tools/perf/util/scripting-engines/trace-event-python.c
1184
PyObject *t;
tools/perf/util/scripting-engines/trace-event-python.c
1186
t = tuple_new(5);
tools/perf/util/scripting-engines/trace-event-python.c
1188
tuple_set_d64(t, 0, thread__db_id(thread));
tools/perf/util/scripting-engines/trace-event-python.c
1189
tuple_set_d64(t, 1, machine->db_id);
tools/perf/util/scripting-engines/trace-event-python.c
1190
tuple_set_d64(t, 2, main_thread_db_id);
tools/perf/util/scripting-engines/trace-event-python.c
1191
tuple_set_s32(t, 3, thread__pid(thread));
tools/perf/util/scripting-engines/trace-event-python.c
1192
tuple_set_s32(t, 4, thread__tid(thread));
tools/perf/util/scripting-engines/trace-event-python.c
1194
call_object(tables->thread_handler, t, "thread_table");
tools/perf/util/scripting-engines/trace-event-python.c
1196
Py_DECREF(t);
tools/perf/util/scripting-engines/trace-event-python.c
1205
PyObject *t;
tools/perf/util/scripting-engines/trace-event-python.c
1207
t = tuple_new(5);
tools/perf/util/scripting-engines/trace-event-python.c
1209
tuple_set_d64(t, 0, comm->db_id);
tools/perf/util/scripting-engines/trace-event-python.c
1210
tuple_set_string(t, 1, comm__str(comm));
tools/perf/util/scripting-engines/trace-event-python.c
1211
tuple_set_d64(t, 2, thread__db_id(thread));
tools/perf/util/scripting-engines/trace-event-python.c
1212
tuple_set_d64(t, 3, comm->start);
tools/perf/util/scripting-engines/trace-event-python.c
1213
tuple_set_s32(t, 4, comm->exec);
tools/perf/util/scripting-engines/trace-event-python.c
1215
call_object(tables->comm_handler, t, "comm_table");
tools/perf/util/scripting-engines/trace-event-python.c
1217
Py_DECREF(t);
tools/perf/util/scripting-engines/trace-event-python.c
1226
PyObject *t;
tools/perf/util/scripting-engines/trace-event-python.c
1228
t = tuple_new(3);
tools/perf/util/scripting-engines/trace-event-python.c
1230
tuple_set_d64(t, 0, db_id);
tools/perf/util/scripting-engines/trace-event-python.c
1231
tuple_set_d64(t, 1, comm->db_id);
tools/perf/util/scripting-engines/trace-event-python.c
1232
tuple_set_d64(t, 2, thread__db_id(thread));
tools/perf/util/scripting-engines/trace-event-python.c
1234
call_object(tables->comm_thread_handler, t, "comm_thread_table");
tools/perf/util/scripting-engines/trace-event-python.c
1236
Py_DECREF(t);
tools/perf/util/scripting-engines/trace-event-python.c
1246
PyObject *t;
tools/perf/util/scripting-engines/trace-event-python.c
1250
t = tuple_new(5);
tools/perf/util/scripting-engines/trace-event-python.c
1252
tuple_set_d64(t, 0, dso__db_id(dso));
tools/perf/util/scripting-engines/trace-event-python.c
1253
tuple_set_d64(t, 1, machine->db_id);
tools/perf/util/scripting-engines/trace-event-python.c
1254
tuple_set_string(t, 2, dso__short_name(dso));
tools/perf/util/scripting-engines/trace-event-python.c
1255
tuple_set_string(t, 3, dso__long_name(dso));
tools/perf/util/scripting-engines/trace-event-python.c
1256
tuple_set_string(t, 4, sbuild_id);
tools/perf/util/scripting-engines/trace-event-python.c
1258
call_object(tables->dso_handler, t, "dso_table");
tools/perf/util/scripting-engines/trace-event-python.c
1260
Py_DECREF(t);
tools/perf/util/scripting-engines/trace-event-python.c
1270
PyObject *t;
tools/perf/util/scripting-engines/trace-event-python.c
1272
t = tuple_new(6);
tools/perf/util/scripting-engines/trace-event-python.c
1274
tuple_set_d64(t, 0, *sym_db_id);
tools/perf/util/scripting-engines/trace-event-python.c
1275
tuple_set_d64(t, 1, dso__db_id(dso));
tools/perf/util/scripting-engines/trace-event-python.c
1276
tuple_set_d64(t, 2, sym->start);
tools/perf/util/scripting-engines/trace-event-python.c
1277
tuple_set_d64(t, 3, sym->end);
tools/perf/util/scripting-engines/trace-event-python.c
1278
tuple_set_s32(t, 4, sym->binding);
tools/perf/util/scripting-engines/trace-event-python.c
1279
tuple_set_string(t, 5, sym->name);
tools/perf/util/scripting-engines/trace-event-python.c
1281
call_object(tables->symbol_handler, t, "symbol_table");
tools/perf/util/scripting-engines/trace-event-python.c
1283
Py_DECREF(t);
tools/perf/util/scripting-engines/trace-event-python.c
1292
PyObject *t;
tools/perf/util/scripting-engines/trace-event-python.c
1294
t = tuple_new(2);
tools/perf/util/scripting-engines/trace-event-python.c
1296
tuple_set_s32(t, 0, branch_type);
tools/perf/util/scripting-engines/trace-event-python.c
1297
tuple_set_string(t, 1, name);
tools/perf/util/scripting-engines/trace-event-python.c
1299
call_object(tables->branch_type_handler, t, "branch_type_table");
tools/perf/util/scripting-engines/trace-event-python.c
1301
Py_DECREF(t);
tools/perf/util/scripting-engines/trace-event-python.c
1310
PyObject *t;
tools/perf/util/scripting-engines/trace-event-python.c
1312
t = tuple_new(28);
tools/perf/util/scripting-engines/trace-event-python.c
1314
tuple_set_d64(t, 0, es->db_id);
tools/perf/util/scripting-engines/trace-event-python.c
1315
tuple_set_d64(t, 1, es->evsel->db_id);
tools/perf/util/scripting-engines/trace-event-python.c
1316
tuple_set_d64(t, 2, maps__machine(thread__maps(es->al->thread))->db_id);
tools/perf/util/scripting-engines/trace-event-python.c
1317
tuple_set_d64(t, 3, thread__db_id(es->al->thread));
tools/perf/util/scripting-engines/trace-event-python.c
1318
tuple_set_d64(t, 4, es->comm_db_id);
tools/perf/util/scripting-engines/trace-event-python.c
1319
tuple_set_d64(t, 5, es->dso_db_id);
tools/perf/util/scripting-engines/trace-event-python.c
1320
tuple_set_d64(t, 6, es->sym_db_id);
tools/perf/util/scripting-engines/trace-event-python.c
1321
tuple_set_d64(t, 7, es->offset);
tools/perf/util/scripting-engines/trace-event-python.c
1322
tuple_set_d64(t, 8, es->sample->ip);
tools/perf/util/scripting-engines/trace-event-python.c
1323
tuple_set_d64(t, 9, es->sample->time);
tools/perf/util/scripting-engines/trace-event-python.c
1324
tuple_set_s32(t, 10, es->sample->cpu);
tools/perf/util/scripting-engines/trace-event-python.c
1325
tuple_set_d64(t, 11, es->addr_dso_db_id);
tools/perf/util/scripting-engines/trace-event-python.c
1326
tuple_set_d64(t, 12, es->addr_sym_db_id);
tools/perf/util/scripting-engines/trace-event-python.c
1327
tuple_set_d64(t, 13, es->addr_offset);
tools/perf/util/scripting-engines/trace-event-python.c
1328
tuple_set_d64(t, 14, es->sample->addr);
tools/perf/util/scripting-engines/trace-event-python.c
1329
tuple_set_d64(t, 15, es->sample->period);
tools/perf/util/scripting-engines/trace-event-python.c
1330
tuple_set_d64(t, 16, es->sample->weight);
tools/perf/util/scripting-engines/trace-event-python.c
1331
tuple_set_d64(t, 17, es->sample->transaction);
tools/perf/util/scripting-engines/trace-event-python.c
1332
tuple_set_d64(t, 18, es->sample->data_src);
tools/perf/util/scripting-engines/trace-event-python.c
1333
tuple_set_s32(t, 19, es->sample->flags & PERF_BRANCH_MASK);
tools/perf/util/scripting-engines/trace-event-python.c
1334
tuple_set_s32(t, 20, !!(es->sample->flags & PERF_IP_FLAG_IN_TX));
tools/perf/util/scripting-engines/trace-event-python.c
1335
tuple_set_d64(t, 21, es->call_path_id);
tools/perf/util/scripting-engines/trace-event-python.c
1336
tuple_set_d64(t, 22, es->sample->insn_cnt);
tools/perf/util/scripting-engines/trace-event-python.c
1337
tuple_set_d64(t, 23, es->sample->cyc_cnt);
tools/perf/util/scripting-engines/trace-event-python.c
1338
tuple_set_s32(t, 24, es->sample->flags);
tools/perf/util/scripting-engines/trace-event-python.c
1339
tuple_set_d64(t, 25, es->sample->id);
tools/perf/util/scripting-engines/trace-event-python.c
1340
tuple_set_d64(t, 26, es->sample->stream_id);
tools/perf/util/scripting-engines/trace-event-python.c
1341
tuple_set_u32(t, 27, es->sample->ins_lat);
tools/perf/util/scripting-engines/trace-event-python.c
1343
call_object(tables->sample_handler, t, "sample_table");
tools/perf/util/scripting-engines/trace-event-python.c
1345
Py_DECREF(t);
tools/perf/util/scripting-engines/trace-event-python.c
1351
PyObject *t;
tools/perf/util/scripting-engines/trace-event-python.c
1353
t = tuple_new(3);
tools/perf/util/scripting-engines/trace-event-python.c
1355
tuple_set_d64(t, 0, es->db_id);
tools/perf/util/scripting-engines/trace-event-python.c
1356
tuple_set_d64(t, 1, es->evsel->core.attr.config);
tools/perf/util/scripting-engines/trace-event-python.c
1357
tuple_set_bytes(t, 2, es->sample->raw_data, es->sample->raw_size);
tools/perf/util/scripting-engines/trace-event-python.c
1359
call_object(tables->synth_handler, t, "synth_data");
tools/perf/util/scripting-engines/trace-event-python.c
1361
Py_DECREF(t);
tools/perf/util/scripting-engines/trace-event-python.c
1380
PyObject *t;
tools/perf/util/scripting-engines/trace-event-python.c
1386
t = tuple_new(4);
tools/perf/util/scripting-engines/trace-event-python.c
1388
tuple_set_d64(t, 0, cp->db_id);
tools/perf/util/scripting-engines/trace-event-python.c
1389
tuple_set_d64(t, 1, parent_db_id);
tools/perf/util/scripting-engines/trace-event-python.c
1390
tuple_set_d64(t, 2, sym_db_id);
tools/perf/util/scripting-engines/trace-event-python.c
1391
tuple_set_d64(t, 3, cp->ip);
tools/perf/util/scripting-engines/trace-event-python.c
1393
call_object(tables->call_path_handler, t, "call_path_table");
tools/perf/util/scripting-engines/trace-event-python.c
1395
Py_DECREF(t);
tools/perf/util/scripting-engines/trace-event-python.c
1405
PyObject *t;
tools/perf/util/scripting-engines/trace-event-python.c
1407
t = tuple_new(14);
tools/perf/util/scripting-engines/trace-event-python.c
1409
tuple_set_d64(t, 0, cr->db_id);
tools/perf/util/scripting-engines/trace-event-python.c
1410
tuple_set_d64(t, 1, thread__db_id(cr->thread));
tools/perf/util/scripting-engines/trace-event-python.c
1411
tuple_set_d64(t, 2, comm_db_id);
tools/perf/util/scripting-engines/trace-event-python.c
1412
tuple_set_d64(t, 3, cr->cp->db_id);
tools/perf/util/scripting-engines/trace-event-python.c
1413
tuple_set_d64(t, 4, cr->call_time);
tools/perf/util/scripting-engines/trace-event-python.c
1414
tuple_set_d64(t, 5, cr->return_time);
tools/perf/util/scripting-engines/trace-event-python.c
1415
tuple_set_d64(t, 6, cr->branch_count);
tools/perf/util/scripting-engines/trace-event-python.c
1416
tuple_set_d64(t, 7, cr->call_ref);
tools/perf/util/scripting-engines/trace-event-python.c
1417
tuple_set_d64(t, 8, cr->return_ref);
tools/perf/util/scripting-engines/trace-event-python.c
1418
tuple_set_d64(t, 9, cr->cp->parent->db_id);
tools/perf/util/scripting-engines/trace-event-python.c
1419
tuple_set_s32(t, 10, cr->flags);
tools/perf/util/scripting-engines/trace-event-python.c
1420
tuple_set_d64(t, 11, cr->parent_db_id);
tools/perf/util/scripting-engines/trace-event-python.c
1421
tuple_set_d64(t, 12, cr->insn_count);
tools/perf/util/scripting-engines/trace-event-python.c
1422
tuple_set_d64(t, 13, cr->cyc_count);
tools/perf/util/scripting-engines/trace-event-python.c
1424
call_object(tables->call_return_handler, t, "call_return_table");
tools/perf/util/scripting-engines/trace-event-python.c
1426
Py_DECREF(t);
tools/perf/util/scripting-engines/trace-event-python.c
1438
PyObject *t;
tools/perf/util/scripting-engines/trace-event-python.c
1440
t = tuple_new(9);
tools/perf/util/scripting-engines/trace-event-python.c
1442
tuple_set_d64(t, 0, db_id);
tools/perf/util/scripting-engines/trace-event-python.c
1443
tuple_set_d64(t, 1, machine->db_id);
tools/perf/util/scripting-engines/trace-event-python.c
1444
tuple_set_d64(t, 2, sample->time);
tools/perf/util/scripting-engines/trace-event-python.c
1445
tuple_set_s32(t, 3, sample->cpu);
tools/perf/util/scripting-engines/trace-event-python.c
1446
tuple_set_d64(t, 4, th_out_id);
tools/perf/util/scripting-engines/trace-event-python.c
1447
tuple_set_d64(t, 5, comm_out_id);
tools/perf/util/scripting-engines/trace-event-python.c
1448
tuple_set_d64(t, 6, th_in_id);
tools/perf/util/scripting-engines/trace-event-python.c
1449
tuple_set_d64(t, 7, comm_in_id);
tools/perf/util/scripting-engines/trace-event-python.c
1450
tuple_set_s32(t, 8, flags);
tools/perf/util/scripting-engines/trace-event-python.c
1452
call_object(tables->context_switch_handler, t, "context_switch");
tools/perf/util/scripting-engines/trace-event-python.c
1454
Py_DECREF(t);
tools/perf/util/scripting-engines/trace-event-python.c
1472
PyObject *handler, *t, *dict, *callchain;
tools/perf/util/scripting-engines/trace-event-python.c
1486
t = PyTuple_New(MAX_FIELDS);
tools/perf/util/scripting-engines/trace-event-python.c
1487
if (!t)
tools/perf/util/scripting-engines/trace-event-python.c
1494
PyTuple_SetItem(t, n++, dict);
tools/perf/util/scripting-engines/trace-event-python.c
1495
if (_PyTuple_Resize(&t, n) == -1)
tools/perf/util/scripting-engines/trace-event-python.c
1498
call_object(handler, t, handler_name);
tools/perf/util/scripting-engines/trace-event-python.c
1500
Py_DECREF(t);
tools/perf/util/scripting-engines/trace-event-python.c
1531
PyObject *handler, *t;
tools/perf/util/scripting-engines/trace-event-python.c
1541
t = tuple_new(6);
tools/perf/util/scripting-engines/trace-event-python.c
1542
if (!t)
tools/perf/util/scripting-engines/trace-event-python.c
1545
tuple_set_u64(t, 0, event->throttle.time);
tools/perf/util/scripting-engines/trace-event-python.c
1546
tuple_set_u64(t, 1, event->throttle.id);
tools/perf/util/scripting-engines/trace-event-python.c
1547
tuple_set_u64(t, 2, event->throttle.stream_id);
tools/perf/util/scripting-engines/trace-event-python.c
1548
tuple_set_s32(t, 3, sample->cpu);
tools/perf/util/scripting-engines/trace-event-python.c
1549
tuple_set_s32(t, 4, sample->pid);
tools/perf/util/scripting-engines/trace-event-python.c
1550
tuple_set_s32(t, 5, sample->tid);
tools/perf/util/scripting-engines/trace-event-python.c
1552
call_object(handler, t, handler_name);
tools/perf/util/scripting-engines/trace-event-python.c
1554
Py_DECREF(t);
tools/perf/util/scripting-engines/trace-event-python.c
1565
PyObject *handler, *t;
tools/perf/util/scripting-engines/trace-event-python.c
1576
t = tuple_new(11);
tools/perf/util/scripting-engines/trace-event-python.c
1577
if (!t)
tools/perf/util/scripting-engines/trace-event-python.c
1580
tuple_set_u64(t, 0, sample->time);
tools/perf/util/scripting-engines/trace-event-python.c
1581
tuple_set_s32(t, 1, sample->cpu);
tools/perf/util/scripting-engines/trace-event-python.c
1582
tuple_set_s32(t, 2, sample->pid);
tools/perf/util/scripting-engines/trace-event-python.c
1583
tuple_set_s32(t, 3, sample->tid);
tools/perf/util/scripting-engines/trace-event-python.c
1584
tuple_set_s32(t, 4, np_pid);
tools/perf/util/scripting-engines/trace-event-python.c
1585
tuple_set_s32(t, 5, np_tid);
tools/perf/util/scripting-engines/trace-event-python.c
1586
tuple_set_s32(t, 6, machine->pid);
tools/perf/util/scripting-engines/trace-event-python.c
1587
tuple_set_bool(t, 7, out);
tools/perf/util/scripting-engines/trace-event-python.c
1588
tuple_set_bool(t, 8, out_preempt);
tools/perf/util/scripting-engines/trace-event-python.c
1589
tuple_set_s32(t, 9, sample->machine_pid);
tools/perf/util/scripting-engines/trace-event-python.c
1590
tuple_set_s32(t, 10, sample->vcpu);
tools/perf/util/scripting-engines/trace-event-python.c
1592
call_object(handler, t, handler_name);
tools/perf/util/scripting-engines/trace-event-python.c
1594
Py_DECREF(t);
tools/perf/util/scripting-engines/trace-event-python.c
1617
PyObject *handler, *t;
tools/perf/util/scripting-engines/trace-event-python.c
1628
t = tuple_new(11);
tools/perf/util/scripting-engines/trace-event-python.c
1630
tuple_set_u32(t, 0, e->type);
tools/perf/util/scripting-engines/trace-event-python.c
1631
tuple_set_u32(t, 1, e->code);
tools/perf/util/scripting-engines/trace-event-python.c
1632
tuple_set_s32(t, 2, e->cpu);
tools/perf/util/scripting-engines/trace-event-python.c
1633
tuple_set_s32(t, 3, e->pid);
tools/perf/util/scripting-engines/trace-event-python.c
1634
tuple_set_s32(t, 4, e->tid);
tools/perf/util/scripting-engines/trace-event-python.c
1635
tuple_set_u64(t, 5, e->ip);
tools/perf/util/scripting-engines/trace-event-python.c
1636
tuple_set_u64(t, 6, tm);
tools/perf/util/scripting-engines/trace-event-python.c
1637
tuple_set_string(t, 7, msg);
tools/perf/util/scripting-engines/trace-event-python.c
1638
tuple_set_u32(t, 8, cpumode);
tools/perf/util/scripting-engines/trace-event-python.c
1639
tuple_set_s32(t, 9, e->machine_pid);
tools/perf/util/scripting-engines/trace-event-python.c
1640
tuple_set_s32(t, 10, e->vcpu);
tools/perf/util/scripting-engines/trace-event-python.c
1642
call_object(handler, t, handler_name);
tools/perf/util/scripting-engines/trace-event-python.c
1644
Py_DECREF(t);
tools/perf/util/scripting-engines/trace-event-python.c
1664
PyObject *handler, *t;
tools/perf/util/scripting-engines/trace-event-python.c
1668
t = PyTuple_New(MAX_FIELDS);
tools/perf/util/scripting-engines/trace-event-python.c
1669
if (!t)
tools/perf/util/scripting-engines/trace-event-python.c
1681
PyTuple_SetItem(t, n++, _PyLong_FromLong(cpu.cpu));
tools/perf/util/scripting-engines/trace-event-python.c
1682
PyTuple_SetItem(t, n++, _PyLong_FromLong(thread));
tools/perf/util/scripting-engines/trace-event-python.c
1684
tuple_set_u64(t, n++, tstamp);
tools/perf/util/scripting-engines/trace-event-python.c
1685
tuple_set_u64(t, n++, count->val);
tools/perf/util/scripting-engines/trace-event-python.c
1686
tuple_set_u64(t, n++, count->ena);
tools/perf/util/scripting-engines/trace-event-python.c
1687
tuple_set_u64(t, n++, count->run);
tools/perf/util/scripting-engines/trace-event-python.c
1689
if (_PyTuple_Resize(&t, n) == -1)
tools/perf/util/scripting-engines/trace-event-python.c
1692
call_object(handler, t, handler_name);
tools/perf/util/scripting-engines/trace-event-python.c
1694
Py_DECREF(t);
tools/perf/util/scripting-engines/trace-event-python.c
1717
PyObject *handler, *t;
tools/perf/util/scripting-engines/trace-event-python.c
1721
t = PyTuple_New(MAX_FIELDS);
tools/perf/util/scripting-engines/trace-event-python.c
1722
if (!t)
tools/perf/util/scripting-engines/trace-event-python.c
1731
tuple_set_u64(t, n++, tstamp);
tools/perf/util/scripting-engines/trace-event-python.c
1733
if (_PyTuple_Resize(&t, n) == -1)
tools/perf/util/scripting-engines/trace-event-python.c
1736
call_object(handler, t, handler_name);
tools/perf/util/scripting-engines/trace-event-python.c
1738
Py_DECREF(t);
tools/perf/util/scripting-engines/trace-event-python.c
189
PyObject *t;
tools/perf/util/scripting-engines/trace-event-python.c
196
t = PyTuple_New(4);
tools/perf/util/scripting-engines/trace-event-python.c
197
if (!t)
tools/perf/util/scripting-engines/trace-event-python.c
202
PyTuple_SetItem(t, n++, _PyUnicode_FromString(ev_name));
tools/perf/util/scripting-engines/trace-event-python.c
203
PyTuple_SetItem(t, n++, _PyUnicode_FromString(field_name));
tools/perf/util/scripting-engines/trace-event-python.c
204
PyTuple_SetItem(t, n++, _PyLong_FromLong(value));
tools/perf/util/scripting-engines/trace-event-python.c
205
PyTuple_SetItem(t, n++, _PyUnicode_FromString(field_str));
tools/perf/util/scripting-engines/trace-event-python.c
207
try_call_object(handler_name, t);
tools/perf/util/scripting-engines/trace-event-python.c
209
Py_DECREF(t);
tools/perf/util/scripting-engines/trace-event-python.c
230
PyObject *t;
tools/perf/util/scripting-engines/trace-event-python.c
237
t = PyTuple_New(3);
tools/perf/util/scripting-engines/trace-event-python.c
239
t = PyTuple_New(2);
tools/perf/util/scripting-engines/trace-event-python.c
240
if (!t)
tools/perf/util/scripting-engines/trace-event-python.c
243
PyTuple_SetItem(t, n++, _PyUnicode_FromString(ev_name));
tools/perf/util/scripting-engines/trace-event-python.c
244
PyTuple_SetItem(t, n++, _PyUnicode_FromString(field_name));
tools/perf/util/scripting-engines/trace-event-python.c
246
PyTuple_SetItem(t, n++, _PyUnicode_FromString(delim));
tools/perf/util/scripting-engines/trace-event-python.c
248
try_call_object(handler_name, t);
tools/perf/util/scripting-engines/trace-event-python.c
250
Py_DECREF(t);
tools/perf/util/scripting-engines/trace-event-python.c
641
PyObject *t;
tools/perf/util/scripting-engines/trace-event-python.c
643
t = PyTuple_New(3);
tools/perf/util/scripting-engines/trace-event-python.c
644
if (!t)
tools/perf/util/scripting-engines/trace-event-python.c
646
PyTuple_SetItem(t, 0, PyLong_FromUnsignedLongLong(value->id));
tools/perf/util/scripting-engines/trace-event-python.c
647
PyTuple_SetItem(t, 1, PyLong_FromUnsignedLongLong(value->value));
tools/perf/util/scripting-engines/trace-event-python.c
649
PyTuple_SetItem(t, 2, PyLong_FromUnsignedLongLong(value->lost));
tools/perf/util/scripting-engines/trace-event-python.c
651
return t;
tools/perf/util/scripting-engines/trace-event-python.c
685
PyObject *t = get_sample_value_as_tuple(v, read_format);
tools/perf/util/scripting-engines/trace-event-python.c
686
PyList_SET_ITEM(values, i, t);
tools/perf/util/scripting-engines/trace-event-python.c
690
PyObject *t = get_sample_value_as_tuple(&sample->read.one,
tools/perf/util/scripting-engines/trace-event-python.c
692
PyList_SET_ITEM(values, 0, t);
tools/perf/util/scripting-engines/trace-event-python.c
944
PyObject *handler, *context, *t, *obj = NULL, *callchain;
tools/perf/util/scripting-engines/trace-event-python.c
984
t = PyTuple_New(MAX_FIELDS);
tools/perf/util/scripting-engines/trace-event-python.c
985
if (!t)
tools/perf/util/scripting-engines/trace-event-python.c
994
PyTuple_SetItem(t, n++, _PyUnicode_FromString(handler_name));
tools/perf/util/scripting-engines/trace-event-python.c
995
PyTuple_SetItem(t, n++, context);
tools/perf/util/sort.c
2211
u64 t = he->transaction;
tools/perf/util/sort.c
2218
if (txbits[i].flag & t)
tools/perf/util/sort.c
2220
if (t && !(t & (PERF_TXN_SYNC|PERF_TXN_ASYNC)))
tools/perf/util/sort.c
2222
if (t & PERF_TXN_ABORT_MASK) {
tools/perf/util/sort.c
2224
(t & PERF_TXN_ABORT_MASK) >>
tools/perf/util/stat-display.c
1419
static double timeval2double(struct timeval *t)
tools/perf/util/stat-display.c
1421
return t->tv_sec + (double) t->tv_usec/USEC_PER_SEC;
tools/perf/util/svghelper.c
700
static void scan_thread_topology(int *map, struct topology *t, int cpu,
tools/perf/util/svghelper.c
706
for (i = 0; i < t->sib_thr_nr; i++) {
tools/perf/util/svghelper.c
707
if (!test_bit(cpu, cpumask_bits(&t->sib_thr[i])))
tools/perf/util/svghelper.c
710
for_each_set_bit(thr, cpumask_bits(&t->sib_thr[i]), nr_cpus)
tools/perf/util/svghelper.c
716
static void scan_core_topology(int *map, struct topology *t, int nr_cpus)
tools/perf/util/svghelper.c
722
for (i = 0; i < t->sib_core_nr; i++)
tools/perf/util/svghelper.c
723
for_each_set_bit(cpu, cpumask_bits(&t->sib_core[i]), nr_cpus)
tools/perf/util/svghelper.c
724
scan_thread_topology(map, t, cpu, &pos, nr_cpus);
tools/perf/util/svghelper.c
754
struct topology t;
tools/perf/util/svghelper.c
760
t.sib_core_nr = env->nr_sibling_cores;
tools/perf/util/svghelper.c
761
t.sib_thr_nr = env->nr_sibling_threads;
tools/perf/util/svghelper.c
762
t.sib_core = calloc(env->nr_sibling_cores, sizeof(cpumask_t));
tools/perf/util/svghelper.c
763
t.sib_thr = calloc(env->nr_sibling_threads, sizeof(cpumask_t));
tools/perf/util/svghelper.c
768
if (!t.sib_core || !t.sib_thr) {
tools/perf/util/svghelper.c
774
if (str_to_bitmap(sib_core, &t.sib_core[i], nr_cpus)) {
tools/perf/util/svghelper.c
783
if (str_to_bitmap(sib_thr, &t.sib_thr[i], nr_cpus)) {
tools/perf/util/svghelper.c
800
scan_core_topology(topology_map, &t, nr_cpus);
tools/perf/util/svghelper.c
805
zfree(&t.sib_core);
tools/perf/util/svghelper.c
806
zfree(&t.sib_thr);
tools/perf/util/synthetic-events.c
435
unsigned long long t;
tools/perf/util/synthetic-events.c
461
t = rdclock();
tools/perf/util/synthetic-events.c
484
if ((rdclock() - t) > timeout) {
tools/perf/util/trace-event-info.c
395
struct tracepoint_path *t = tps;
tools/perf/util/trace-event-info.c
398
zfree(&t->name);
tools/perf/util/trace-event-info.c
399
zfree(&t->system);
tools/perf/util/trace-event-info.c
400
free(t);
tools/perf/util/trace-event.c
27
int trace_event__init(struct trace_event *t)
tools/perf/util/trace-event.c
32
t->plugin_list = tep_load_plugins(pevent);
tools/perf/util/trace-event.c
33
t->pevent = pevent;
tools/perf/util/trace-event.c
64
void trace_event__cleanup(struct trace_event *t)
tools/perf/util/trace-event.c
66
tep_unload_plugins(t->plugin_list, t->pevent);
tools/perf/util/trace-event.c
67
tep_free(t->pevent);
tools/perf/util/trace-event.h
33
int trace_event__init(struct trace_event *t);
tools/perf/util/trace-event.h
34
void trace_event__cleanup(struct trace_event *t);
tools/perf/util/trigger.h
36
#define TRIGGER_WARN_ONCE(t, exp) \
tools/perf/util/trigger.h
37
WARN_ONCE(t->state != exp, "trigger '%s' state transist error: %d in %s()\n", \
tools/perf/util/trigger.h
38
t->name, t->state, __func__)
tools/perf/util/trigger.h
40
static inline bool trigger_is_available(struct trigger *t)
tools/perf/util/trigger.h
42
return t->state >= 0;
tools/perf/util/trigger.h
45
static inline bool trigger_is_error(struct trigger *t)
tools/perf/util/trigger.h
47
return t->state <= TRIGGER_ERROR;
tools/perf/util/trigger.h
50
static inline void trigger_on(struct trigger *t)
tools/perf/util/trigger.h
52
TRIGGER_WARN_ONCE(t, TRIGGER_OFF);
tools/perf/util/trigger.h
53
t->state = TRIGGER_ON;
tools/perf/util/trigger.h
56
static inline void trigger_ready(struct trigger *t)
tools/perf/util/trigger.h
58
if (!trigger_is_available(t))
tools/perf/util/trigger.h
60
t->state = TRIGGER_READY;
tools/perf/util/trigger.h
63
static inline void trigger_hit(struct trigger *t)
tools/perf/util/trigger.h
65
if (!trigger_is_available(t))
tools/perf/util/trigger.h
67
TRIGGER_WARN_ONCE(t, TRIGGER_READY);
tools/perf/util/trigger.h
68
t->state = TRIGGER_HIT;
tools/perf/util/trigger.h
71
static inline void trigger_off(struct trigger *t)
tools/perf/util/trigger.h
73
if (!trigger_is_available(t))
tools/perf/util/trigger.h
75
t->state = TRIGGER_OFF;
tools/perf/util/trigger.h
78
static inline void trigger_error(struct trigger *t)
tools/perf/util/trigger.h
80
t->state = TRIGGER_ERROR;
tools/perf/util/trigger.h
83
static inline bool trigger_is_ready(struct trigger *t)
tools/perf/util/trigger.h
85
return t->state == TRIGGER_READY;
tools/perf/util/trigger.h
88
static inline bool trigger_is_hit(struct trigger *t)
tools/perf/util/trigger.h
90
return t->state == TRIGGER_HIT;
tools/perf/util/tsc.c
20
u64 t, quot, rem;
tools/perf/util/tsc.c
22
t = ns - tc->time_zero;
tools/perf/util/tsc.c
23
quot = t / tc->time_mult;
tools/perf/util/tsc.c
24
rem = t % tc->time_mult;
tools/power/cpupower/utils/helpers/amd.c
59
int t;
tools/power/cpupower/utils/helpers/amd.c
66
t = pstate.pstatedef.did;
tools/power/cpupower/utils/helpers/amd.c
68
t = pstate.val & 0xf;
tools/power/cpupower/utils/helpers/amd.c
70
t = pstate.pstate.did;
tools/power/cpupower/utils/helpers/amd.c
72
return t;
tools/power/cpupower/utils/helpers/amd.c
77
int t;
tools/power/cpupower/utils/helpers/amd.c
91
t = 0x10;
tools/power/cpupower/utils/helpers/amd.c
94
t = 0x8;
tools/power/cpupower/utils/helpers/amd.c
95
cof = (100 * (fid + t)) >> did;
tools/power/x86/turbostat/turbostat.c
2444
#define PER_THREAD_PARAMS struct thread_data *t, struct core_data *c, struct pkg_data *p
tools/power/x86/turbostat/turbostat.c
2454
struct thread_data *t;
tools/power/x86/turbostat/turbostat.c
2466
t = &thread_base[cpu];
tools/power/x86/turbostat/turbostat.c
2470
retval |= func(t, c, p);
tools/power/x86/turbostat/turbostat.c
2478
t = &thread_base[cpus[cpu].ht_sibling_cpu_id[i]];
tools/power/x86/turbostat/turbostat.c
2480
retval |= func(t, c, p);
tools/power/x86/turbostat/turbostat.c
2486
int is_cpu_first_thread_in_core(struct thread_data *t, struct core_data *c)
tools/power/x86/turbostat/turbostat.c
2488
return ((int)t->cpu_id == c->first_cpu || c->first_cpu < 0);
tools/power/x86/turbostat/turbostat.c
2491
int is_cpu_first_core_in_package(struct thread_data *t, struct pkg_data *p)
tools/power/x86/turbostat/turbostat.c
2493
return ((int)t->cpu_id == p->first_cpu || p->first_cpu < 0);
tools/power/x86/turbostat/turbostat.c
2496
int is_cpu_first_thread_in_package(struct thread_data *t, struct core_data *c, struct pkg_data *p)
tools/power/x86/turbostat/turbostat.c
2498
return is_cpu_first_thread_in_core(t, c) && is_cpu_first_core_in_package(t, p);
tools/power/x86/turbostat/turbostat.c
3146
outp += sprintf(outp, "t %p, c %p, p %p\n", t, c, p);
tools/power/x86/turbostat/turbostat.c
3148
if (t) {
tools/power/x86/turbostat/turbostat.c
3149
outp += sprintf(outp, "CPU: %d flags 0x%x\n", t->cpu_id, t->flags);
tools/power/x86/turbostat/turbostat.c
3150
outp += sprintf(outp, "TSC: %016llX\n", t->tsc);
tools/power/x86/turbostat/turbostat.c
3151
outp += sprintf(outp, "aperf: %016llX\n", t->aperf);
tools/power/x86/turbostat/turbostat.c
3152
outp += sprintf(outp, "mperf: %016llX\n", t->mperf);
tools/power/x86/turbostat/turbostat.c
3153
outp += sprintf(outp, "c1: %016llX\n", t->c1);
tools/power/x86/turbostat/turbostat.c
3156
outp += sprintf(outp, "IPC: %lld\n", t->instr_count);
tools/power/x86/turbostat/turbostat.c
3159
outp += sprintf(outp, "IRQ: %lld\n", t->irq_count);
tools/power/x86/turbostat/turbostat.c
3161
outp += sprintf(outp, "IRQ: %lld\n", t->nmi_count);
tools/power/x86/turbostat/turbostat.c
3163
outp += sprintf(outp, "SMI: %d\n", t->smi_count);
tools/power/x86/turbostat/turbostat.c
3165
outp += sprintf(outp, "LLC refs: %lld", t->llc.references);
tools/power/x86/turbostat/turbostat.c
3166
outp += sprintf(outp, "LLC miss: %lld", t->llc.misses);
tools/power/x86/turbostat/turbostat.c
3167
outp += sprintf(outp, "LLC Hit%%: %.2f", pct((t->llc.references - t->llc.misses), t->llc.references));
tools/power/x86/turbostat/turbostat.c
3169
outp += sprintf(outp, "L2 refs: %lld", t->l2.references);
tools/power/x86/turbostat/turbostat.c
3170
outp += sprintf(outp, "L2 hits: %lld", t->l2.hits);
tools/power/x86/turbostat/turbostat.c
3171
outp += sprintf(outp, "L2 Hit%%: %.2f", pct(t->l2.hits, t->l2.references));
tools/power/x86/turbostat/turbostat.c
3174
outp += sprintf(outp, "tADDED [%d] %8s msr0x%x: %08llX %s\n", i, mp->name, mp->msr_num, t->counter[i], mp->sp->path);
tools/power/x86/turbostat/turbostat.c
3178
if (c && is_cpu_first_thread_in_core(t, c)) {
tools/power/x86/turbostat/turbostat.c
3179
outp += sprintf(outp, "core: %d\n", cpus[t->cpu_id].core_id);
tools/power/x86/turbostat/turbostat.c
3198
if (p && is_cpu_first_core_in_package(t, p)) {
tools/power/x86/turbostat/turbostat.c
3310
if (t == average.threads) {
tools/power/x86/turbostat/turbostat.c
3316
if (show_core_only && !is_cpu_first_thread_in_core(t, c))
tools/power/x86/turbostat/turbostat.c
3320
if (show_pkg_only && !is_cpu_first_core_in_package(t, p))
tools/power/x86/turbostat/turbostat.c
3324
if ((t != average.threads) && (cpu_subset && !CPU_ISSET_S(t->cpu_id, cpu_subset_size, cpu_subset)))
tools/power/x86/turbostat/turbostat.c
3331
timersub(&t->tv_end, &t->tv_begin, &tv);
tools/power/x86/turbostat/turbostat.c
3337
outp += sprintf(outp, "%10ld.%06ld\t", t->tv_end.tv_sec, t->tv_end.tv_usec);
tools/power/x86/turbostat/turbostat.c
3339
interval_float = t->tv_delta.tv_sec + t->tv_delta.tv_usec / 1000000.0;
tools/power/x86/turbostat/turbostat.c
3341
tsc = t->tsc * tsc_tweak;
tools/power/x86/turbostat/turbostat.c
3344
if (t == average.threads) {
tools/power/x86/turbostat/turbostat.c
3364
outp += sprintf(outp, "%s%d", (printed++ ? delim : ""), cpus[t->cpu_id].package_id);
tools/power/x86/turbostat/turbostat.c
3370
outp += sprintf(outp, "%s%d", (printed++ ? delim : ""), cpus[t->cpu_id].die_id);
tools/power/x86/turbostat/turbostat.c
3376
outp += sprintf(outp, "%s%d", (printed++ ? delim : ""), cpus[t->cpu_id].l3_id);
tools/power/x86/turbostat/turbostat.c
3381
if (t)
tools/power/x86/turbostat/turbostat.c
3382
outp += sprintf(outp, "%s%d", (printed++ ? delim : ""), cpus[t->cpu_id].physical_node_id);
tools/power/x86/turbostat/turbostat.c
3388
outp += sprintf(outp, "%s%d", (printed++ ? delim : ""), cpus[t->cpu_id].core_id);
tools/power/x86/turbostat/turbostat.c
3393
outp += sprintf(outp, "%s%d", (printed++ ? delim : ""), t->cpu_id);
tools/power/x86/turbostat/turbostat.c
3395
outp += sprintf(outp, "%s%d", (printed++ ? delim : ""), t->apic_id);
tools/power/x86/turbostat/turbostat.c
3397
outp += sprintf(outp, "%s%d", (printed++ ? delim : ""), t->x2apic_id);
tools/power/x86/turbostat/turbostat.c
3401
outp += sprintf(outp, "%s%.0f", (printed++ ? delim : ""), 1.0 / units * t->aperf / interval_float);
tools/power/x86/turbostat/turbostat.c
3404
outp += sprintf(outp, "%s%.2f", (printed++ ? delim : ""), pct(t->mperf, tsc));
tools/power/x86/turbostat/turbostat.c
3408
outp += sprintf(outp, "%s%.0f", (printed++ ? delim : ""), base_hz / units * t->aperf / t->mperf);
tools/power/x86/turbostat/turbostat.c
3410
outp += sprintf(outp, "%s%.0f", (printed++ ? delim : ""), tsc / units * t->aperf / t->mperf / interval_float);
tools/power/x86/turbostat/turbostat.c
3414
outp += sprintf(outp, "%s%.0f", (printed++ ? delim : ""), 1.0 * t->tsc / units / interval_float);
tools/power/x86/turbostat/turbostat.c
3417
outp += sprintf(outp, "%s%.2f", (printed++ ? delim : ""), 1.0 * t->instr_count / t->aperf);
tools/power/x86/turbostat/turbostat.c
3422
outp += sprintf(outp, "%s%8lld", (printed++ ? delim : ""), t->irq_count);
tools/power/x86/turbostat/turbostat.c
3424
outp += sprintf(outp, "%s%lld", (printed++ ? delim : ""), t->irq_count);
tools/power/x86/turbostat/turbostat.c
3430
outp += sprintf(outp, "%s%8lld", (printed++ ? delim : ""), t->nmi_count);
tools/power/x86/turbostat/turbostat.c
3432
outp += sprintf(outp, "%s%lld", (printed++ ? delim : ""), t->nmi_count);
tools/power/x86/turbostat/turbostat.c
3437
outp += sprintf(outp, "%s%d", (printed++ ? delim : ""), t->smi_count);
tools/power/x86/turbostat/turbostat.c
3441
outp += sprintf(outp, "%s%.0f", (printed++ ? delim : ""), t->llc.references / interval_float / 1000000);
tools/power/x86/turbostat/turbostat.c
3444
outp += sprintf(outp, fmt8, (printed++ ? delim : ""), pct((t->llc.references - t->llc.misses), t->llc.references));
tools/power/x86/turbostat/turbostat.c
3448
outp += sprintf(outp, "%s%.0f", (printed++ ? delim : ""), t->l2.references / interval_float / 1000000);
tools/power/x86/turbostat/turbostat.c
3451
outp += sprintf(outp, fmt8, (printed++ ? delim : ""), pct(t->l2.hits, t->l2.references));
tools/power/x86/turbostat/turbostat.c
3456
outp += print_hex_value(mp->width, &printed, delim, t->counter[i]);
tools/power/x86/turbostat/turbostat.c
3458
outp += print_decimal_value(mp->width, &printed, delim, t->counter[i]);
tools/power/x86/turbostat/turbostat.c
3461
outp += print_float_value(&printed, delim, t->counter[i] / interval_float / 10000);
tools/power/x86/turbostat/turbostat.c
3463
outp += print_float_value(&printed, delim, pct(t->counter[i], tsc));
tools/power/x86/turbostat/turbostat.c
3470
outp += print_hex_value(pp->width, &printed, delim, t->perf_counter[i]);
tools/power/x86/turbostat/turbostat.c
3472
outp += print_decimal_value(pp->width, &printed, delim, t->perf_counter[i]);
tools/power/x86/turbostat/turbostat.c
3475
outp += print_float_value(&printed, delim, t->perf_counter[i] / interval_float / 10000);
tools/power/x86/turbostat/turbostat.c
3477
outp += print_float_value(&printed, delim, pct(t->perf_counter[i], tsc));
tools/power/x86/turbostat/turbostat.c
3483
const unsigned long value_raw = t->pmt_counter[i];
tools/power/x86/turbostat/turbostat.c
3487
outp += print_hex_value(pmt_counter_get_width(ppmt), &printed, delim, t->pmt_counter[i]);
tools/power/x86/turbostat/turbostat.c
3503
outp += sprintf(outp, "%s%.2f", (printed++ ? delim : ""), pct(t->c1, tsc));
tools/power/x86/turbostat/turbostat.c
3506
if (!is_cpu_first_thread_in_core(t, c))
tools/power/x86/turbostat/turbostat.c
3573
if (!is_cpu_first_core_in_package(t, p))
tools/power/x86/turbostat/turbostat.c
3723
if (DO_BIC(BIC_SysWatt) && (t == average.threads))
tools/power/x86/turbostat/turbostat.c
3725
if (DO_BIC(BIC_Sys_J) && (t == average.threads))
tools/power/x86/turbostat/turbostat.c
3771
for_all_cpus(format_counters, t, c, p);
tools/power/x86/turbostat/turbostat.c
4026
int delta_cpu(struct thread_data *t, struct core_data *c, struct pkg_data *p, struct thread_data *t2, struct core_data *c2, struct pkg_data *p2)
tools/power/x86/turbostat/turbostat.c
4031
if (is_cpu_first_thread_in_core(t, c))
tools/power/x86/turbostat/turbostat.c
4035
retval = delta_thread(t, t2, c2); /* c2 is core delta */
tools/power/x86/turbostat/turbostat.c
4038
if (is_cpu_first_core_in_package(t, p))
tools/power/x86/turbostat/turbostat.c
4061
t->tv_begin.tv_sec = 0;
tools/power/x86/turbostat/turbostat.c
4062
t->tv_begin.tv_usec = 0;
tools/power/x86/turbostat/turbostat.c
4063
t->tv_end.tv_sec = 0;
tools/power/x86/turbostat/turbostat.c
4064
t->tv_end.tv_usec = 0;
tools/power/x86/turbostat/turbostat.c
4065
t->tv_delta.tv_sec = 0;
tools/power/x86/turbostat/turbostat.c
4066
t->tv_delta.tv_usec = 0;
tools/power/x86/turbostat/turbostat.c
4068
t->tsc = 0;
tools/power/x86/turbostat/turbostat.c
4069
t->aperf = 0;
tools/power/x86/turbostat/turbostat.c
4070
t->mperf = 0;
tools/power/x86/turbostat/turbostat.c
4071
t->c1 = 0;
tools/power/x86/turbostat/turbostat.c
4073
t->instr_count = 0;
tools/power/x86/turbostat/turbostat.c
4075
t->irq_count = 0;
tools/power/x86/turbostat/turbostat.c
4076
t->nmi_count = 0;
tools/power/x86/turbostat/turbostat.c
4077
t->smi_count = 0;
tools/power/x86/turbostat/turbostat.c
4079
t->llc.references = 0;
tools/power/x86/turbostat/turbostat.c
4080
t->llc.misses = 0;
tools/power/x86/turbostat/turbostat.c
4082
t->l2.references = 0;
tools/power/x86/turbostat/turbostat.c
4083
t->l2.hits = 0;
tools/power/x86/turbostat/turbostat.c
4128
t->counter[i] = 0;
tools/power/x86/turbostat/turbostat.c
4136
memset(&t->perf_counter[0], 0, sizeof(t->perf_counter));
tools/power/x86/turbostat/turbostat.c
4140
memset(&t->pmt_counter[0], 0, ARRAY_SIZE(t->pmt_counter));
tools/power/x86/turbostat/turbostat.c
4168
average.threads->apic_id = t->apic_id;
tools/power/x86/turbostat/turbostat.c
4170
average.threads->x2apic_id = t->x2apic_id;
tools/power/x86/turbostat/turbostat.c
4177
average.threads->tv_end = t->tv_end;
tools/power/x86/turbostat/turbostat.c
4179
average.threads->tsc += t->tsc;
tools/power/x86/turbostat/turbostat.c
4180
average.threads->aperf += t->aperf;
tools/power/x86/turbostat/turbostat.c
4181
average.threads->mperf += t->mperf;
tools/power/x86/turbostat/turbostat.c
4182
average.threads->c1 += t->c1;
tools/power/x86/turbostat/turbostat.c
4184
average.threads->instr_count += t->instr_count;
tools/power/x86/turbostat/turbostat.c
4186
average.threads->irq_count += t->irq_count;
tools/power/x86/turbostat/turbostat.c
4187
average.threads->nmi_count += t->nmi_count;
tools/power/x86/turbostat/turbostat.c
4188
average.threads->smi_count += t->smi_count;
tools/power/x86/turbostat/turbostat.c
4190
average.threads->llc.references += t->llc.references;
tools/power/x86/turbostat/turbostat.c
4191
average.threads->llc.misses += t->llc.misses;
tools/power/x86/turbostat/turbostat.c
4193
average.threads->l2.references += t->l2.references;
tools/power/x86/turbostat/turbostat.c
4194
average.threads->l2.hits += t->l2.hits;
tools/power/x86/turbostat/turbostat.c
4199
average.threads->counter[i] += t->counter[i];
tools/power/x86/turbostat/turbostat.c
4205
average.threads->perf_counter[i] += t->perf_counter[i];
tools/power/x86/turbostat/turbostat.c
4209
average.threads->pmt_counter[i] += t->pmt_counter[i];
tools/power/x86/turbostat/turbostat.c
4213
if (!is_cpu_first_thread_in_core(t, c))
tools/power/x86/turbostat/turbostat.c
4243
if (!is_cpu_first_core_in_package(t, p))
tools/power/x86/turbostat/turbostat.c
4322
for_all_cpus(sum_counters, t, c, p);
tools/power/x86/turbostat/turbostat.c
4552
void get_apic_id(struct thread_data *t)
tools/power/x86/turbostat/turbostat.c
4560
t->apic_id = (ebx >> 24) & 0xff;
tools/power/x86/turbostat/turbostat.c
4582
t->x2apic_id = eax;
tools/power/x86/turbostat/turbostat.c
4594
t->x2apic_id = edx;
tools/power/x86/turbostat/turbostat.c
4596
if (debug && (t->apic_id != (t->x2apic_id & 0xff)))
tools/power/x86/turbostat/turbostat.c
4597
fprintf(outf, "cpu%d: BIOS BUG: apic 0x%x x2apic 0x%x\n", t->cpu_id, t->apic_id, t->x2apic_id);
tools/power/x86/turbostat/turbostat.c
5005
PERF_COUNTER_WRITE_DATA(t->c1, CCSTATE_RCI_INDEX_C1_RESIDENCY);
tools/power/x86/turbostat/turbostat.c
5034
int get_smi_aperf_mperf(unsigned int cpu, struct thread_data *t)
tools/power/x86/turbostat/turbostat.c
5093
t->aperf = mci->data[MSR_RCI_INDEX_APERF];
tools/power/x86/turbostat/turbostat.c
5094
t->mperf = mci->data[MSR_RCI_INDEX_MPERF];
tools/power/x86/turbostat/turbostat.c
5095
t->smi_count = mci->data[MSR_RCI_INDEX_SMI];
tools/power/x86/turbostat/turbostat.c
5176
int cpu = t->cpu_id;
tools/power/x86/turbostat/turbostat.c
5188
gettimeofday(&t->tv_begin, (struct timezone *)NULL);
tools/power/x86/turbostat/turbostat.c
5191
get_apic_id(t);
tools/power/x86/turbostat/turbostat.c
5193
t->tsc = rdtsc(); /* we are running on local CPU of interest */
tools/power/x86/turbostat/turbostat.c
5195
get_smi_aperf_mperf(cpu, t);
tools/power/x86/turbostat/turbostat.c
5198
get_perf_llc_stats(cpu, &t->llc);
tools/power/x86/turbostat/turbostat.c
5201
get_perf_l2_stats(cpu, &t->l2);
tools/power/x86/turbostat/turbostat.c
5204
if (read(get_instr_count_fd(cpu), &t->instr_count, sizeof(long long)) != sizeof(long long))
tools/power/x86/turbostat/turbostat.c
5208
t->irq_count = irqs_per_cpu[cpu];
tools/power/x86/turbostat/turbostat.c
5210
t->nmi_count = nmi_per_cpu[cpu];
tools/power/x86/turbostat/turbostat.c
5212
get_cstate_counters(cpu, t, c, p);
tools/power/x86/turbostat/turbostat.c
5215
if (get_mp(cpu, mp, &t->counter[i], mp->sp->path))
tools/power/x86/turbostat/turbostat.c
5219
if (perf_counter_info_read_values(sys.perf_tp, cpu, t->perf_counter, MAX_ADDED_THREAD_COUNTERS))
tools/power/x86/turbostat/turbostat.c
5223
t->pmt_counter[i] = pmt_read_counter(pp, t->cpu_id);
tools/power/x86/turbostat/turbostat.c
5226
if (!is_cpu_first_thread_in_core(t, c))
tools/power/x86/turbostat/turbostat.c
5235
if (DO_BIC(BIC_CPU_c7) && t->is_atom) {
tools/power/x86/turbostat/turbostat.c
5267
c->pmt_counter[i] = pmt_read_counter(pp, cpus[t->cpu_id].core_id);
tools/power/x86/turbostat/turbostat.c
5270
if (!is_cpu_first_core_in_package(t, p))
tools/power/x86/turbostat/turbostat.c
5308
p->uncore_mhz = get_legacy_uncore_mhz(cpus[t->cpu_id].package_id);
tools/power/x86/turbostat/turbostat.c
5332
path = find_sysfs_path_by_id(mp->sp, cpus[t->cpu_id].package_id);
tools/power/x86/turbostat/turbostat.c
5334
warnx("%s: package_id %d not found", __func__, cpus[t->cpu_id].package_id);
tools/power/x86/turbostat/turbostat.c
5346
p->pmt_counter[i] = pmt_read_counter(pp, cpus[t->cpu_id].package_id);
tools/power/x86/turbostat/turbostat.c
5349
gettimeofday(&t->tv_end, (struct timezone *)NULL);
tools/power/x86/turbostat/turbostat.c
6232
struct thread_data *t, *t2;
tools/power/x86/turbostat/turbostat.c
6242
t = &thread_base[cpu];
tools/power/x86/turbostat/turbostat.c
6249
retval |= func(t, c, p, t2, c2, p2);
tools/power/x86/turbostat/turbostat.c
6257
t = &thread_base[cpus[cpu].ht_sibling_cpu_id[i]];
tools/power/x86/turbostat/turbostat.c
6260
retval |= func(t, c, p, t2, c2, p2);
tools/power/x86/turbostat/turbostat.c
6726
int cpu = t->cpu_id;
tools/power/x86/turbostat/turbostat.c
7493
cpu = t->cpu_id;
tools/power/x86/turbostat/turbostat.c
7496
if (!is_cpu_first_thread_in_package(t, c, p))
tools/power/x86/turbostat/turbostat.c
7545
cpu = t->cpu_id;
tools/power/x86/turbostat/turbostat.c
7548
if (!is_cpu_first_thread_in_package(t, c, p))
tools/power/x86/turbostat/turbostat.c
7626
cpu = t->cpu_id;
tools/power/x86/turbostat/turbostat.c
7629
if (!is_cpu_first_thread_in_package(t, c, p))
tools/power/x86/turbostat/turbostat.c
7996
if (!is_cpu_first_thread_in_package(t, c, p))
tools/power/x86/turbostat/turbostat.c
7999
cpu = t->cpu_id;
tools/power/x86/turbostat/turbostat.c
8181
if (!is_cpu_first_thread_in_package(t, c, p))
tools/power/x86/turbostat/turbostat.c
8184
cpu = t->cpu_id;
tools/power/x86/turbostat/turbostat.c
8250
cpu = t->cpu_id;
tools/power/x86/turbostat/turbostat.c
8253
if (!is_cpu_first_thread_in_core(t, c))
tools/power/x86/turbostat/turbostat.c
8261
if (do_ptm && is_cpu_first_core_in_package(t, p)) {
tools/power/x86/turbostat/turbostat.c
8322
if (cpu_migrate(t->cpu_id)) {
tools/power/x86/turbostat/turbostat.c
8323
fprintf(outf, "Could not migrate to CPU %d\n", t->cpu_id);
tools/power/x86/turbostat/turbostat.c
8333
t->is_atom = true;
tools/power/x86/turbostat/turbostat.c
9744
struct thread_data *t;
tools/power/x86/turbostat/turbostat.c
9753
t = &thread_base[cpu_id];
tools/power/x86/turbostat/turbostat.c
9756
t->cpu_id = cpu_id;
tools/power/x86/turbostat/turbostat.c
9760
c->first_cpu = t->cpu_id;
tools/power/x86/turbostat/turbostat.c
9762
pkg_base[pkg_id].first_cpu = t->cpu_id;
tools/power/x86/turbostat/turbostat.c
9806
if ((int)t->cpu_id == c->first_cpu)
tools/power/x86/turbostat/turbostat.c
9808
if ((int)t->cpu_id == p->first_cpu)
tools/sched_ext/include/scx/compat.h
27
const struct btf_type *t;
tools/sched_ext/include/scx/compat.h
38
t = btf__type_by_id(__COMPAT_vmlinux_btf, tid);
tools/sched_ext/include/scx/compat.h
39
SCX_BUG_ON(!t, "btf__type_by_id(%d)", tid);
tools/sched_ext/include/scx/compat.h
41
if (btf_is_enum(t)) {
tools/sched_ext/include/scx/compat.h
42
struct btf_enum *e = btf_enum(t);
tools/sched_ext/include/scx/compat.h
44
for (i = 0; i < BTF_INFO_VLEN(t->info); i++) {
tools/sched_ext/include/scx/compat.h
52
} else if (btf_is_enum64(t)) {
tools/sched_ext/include/scx/compat.h
53
struct btf_enum64 *e = btf_enum64(t);
tools/sched_ext/include/scx/compat.h
55
for (i = 0; i < BTF_INFO_VLEN(t->info); i++) {
tools/sched_ext/include/scx/compat.h
83
const struct btf_type *t;
tools/sched_ext/include/scx/compat.h
94
t = btf__type_by_id(__COMPAT_vmlinux_btf, tid);
tools/sched_ext/include/scx/compat.h
95
SCX_BUG_ON(!t, "btf__type_by_id(%d)", tid);
tools/sched_ext/include/scx/compat.h
97
m = btf_members(t);
tools/sched_ext/include/scx/compat.h
99
for (i = 0; i < BTF_INFO_VLEN(t->info); i++) {
tools/testing/cxl/test/cxl_translate.c
346
struct param_test *t = &param_tests[i];
tools/testing/cxl/test/cxl_translate.c
348
rc = cxl_validate_translation_params(t->eiw, t->eig, t->pos);
tools/testing/cxl/test/cxl_translate.c
351
if (valid != t->expect) {
tools/testing/cxl/test/cxl_translate.c
352
pr_err("test params failed: %s\n", t->desc);
tools/testing/nvdimm/test/ndtest.c
292
struct ndtest_priv *t = instances[i];
tools/testing/nvdimm/test/ndtest.c
294
if (!t)
tools/testing/nvdimm/test/ndtest.c
297
list_for_each_entry(n, &t->resources, list) {
tools/testing/nvdimm/test/nfit.c
1013
static int nd_intel_test_cmd_freeze_lock(struct nfit_test *t,
tools/testing/nvdimm/test/nfit.c
1017
struct device *dev = &t->pdev.dev;
tools/testing/nvdimm/test/nfit.c
1032
static int nd_intel_test_cmd_disable_pass(struct nfit_test *t,
tools/testing/nvdimm/test/nfit.c
1036
struct device *dev = &t->pdev.dev;
tools/testing/nvdimm/test/nfit.c
1056
static int nd_intel_test_cmd_secure_erase(struct nfit_test *t,
tools/testing/nvdimm/test/nfit.c
1060
struct device *dev = &t->pdev.dev;
tools/testing/nvdimm/test/nfit.c
1087
static int nd_intel_test_cmd_overwrite(struct nfit_test *t,
tools/testing/nvdimm/test/nfit.c
1091
struct device *dev = &t->pdev.dev;
tools/testing/nvdimm/test/nfit.c
1110
static int nd_intel_test_cmd_query_overwrite(struct nfit_test *t,
tools/testing/nvdimm/test/nfit.c
1114
struct device *dev = &t->pdev.dev;
tools/testing/nvdimm/test/nfit.c
1133
static int nd_intel_test_cmd_master_set_pass(struct nfit_test *t,
tools/testing/nvdimm/test/nfit.c
1137
struct device *dev = &t->pdev.dev;
tools/testing/nvdimm/test/nfit.c
1160
static int nd_intel_test_cmd_master_secure_erase(struct nfit_test *t,
tools/testing/nvdimm/test/nfit.c
1164
struct device *dev = &t->pdev.dev;
tools/testing/nvdimm/test/nfit.c
1190
static int nvdimm_bus_intel_fw_activate_businfo(struct nfit_test *t,
tools/testing/nvdimm/test/nfit.c
1199
struct nfit_test_fw *fw = &t->fw[i];
tools/testing/nvdimm/test/nfit.c
1232
static int nvdimm_bus_intel_fw_activate(struct nfit_test *t,
tools/testing/nvdimm/test/nfit.c
1240
nvdimm_bus_intel_fw_activate_businfo(t, &info, sizeof(info));
tools/testing/nvdimm/test/nfit.c
1248
dev_dbg(&t->pdev.dev, "status: %d\n", status);
tools/testing/nvdimm/test/nfit.c
1255
struct nfit_test_fw *fw = &t->fw[i];
tools/testing/nvdimm/test/nfit.c
1270
static int nd_intel_test_cmd_fw_activate_dimminfo(struct nfit_test *t,
tools/testing/nvdimm/test/nfit.c
1275
struct nfit_test_fw *fw = &t->fw[dimm];
tools/testing/nvdimm/test/nfit.c
1278
nvdimm_bus_intel_fw_activate_businfo(t, &info, sizeof(info));
tools/testing/nvdimm/test/nfit.c
1306
static int nd_intel_test_cmd_fw_activate_arm(struct nfit_test *t,
tools/testing/nvdimm/test/nfit.c
1310
struct nfit_test_fw *fw = &t->fw[dimm];
tools/testing/nvdimm/test/nfit.c
1334
struct nfit_test *t = container_of(acpi_desc, typeof(*t), acpi_desc);
tools/testing/nvdimm/test/nfit.c
1346
dev_dbg(&t->pdev.dev, "%s family: %d cmd: %d: func: %d input length: %d\n",
tools/testing/nvdimm/test/nfit.c
1358
struct nfit_test *t = container_of(acpi_desc, typeof(*t), acpi_desc);
tools/testing/nvdimm/test/nfit.c
1388
dev_WARN_ONCE(&t->pdev.dev, 1,
tools/testing/nvdimm/test/nfit.c
1395
rc = nd_intel_test_cmd_security_status(t,
tools/testing/nvdimm/test/nfit.c
1399
rc = nd_intel_test_cmd_unlock_unit(t,
tools/testing/nvdimm/test/nfit.c
1403
rc = nd_intel_test_cmd_set_pass(t,
tools/testing/nvdimm/test/nfit.c
1407
rc = nd_intel_test_cmd_disable_pass(t,
tools/testing/nvdimm/test/nfit.c
1411
rc = nd_intel_test_cmd_freeze_lock(t,
tools/testing/nvdimm/test/nfit.c
1415
rc = nd_intel_test_cmd_secure_erase(t,
tools/testing/nvdimm/test/nfit.c
1419
rc = nd_intel_test_cmd_overwrite(t,
tools/testing/nvdimm/test/nfit.c
1423
rc = nd_intel_test_cmd_query_overwrite(t,
tools/testing/nvdimm/test/nfit.c
1427
rc = nd_intel_test_cmd_master_set_pass(t,
tools/testing/nvdimm/test/nfit.c
1431
rc = nd_intel_test_cmd_master_secure_erase(t,
tools/testing/nvdimm/test/nfit.c
1436
t, buf, buf_len, i);
tools/testing/nvdimm/test/nfit.c
1440
t, buf, buf_len, i);
tools/testing/nvdimm/test/nfit.c
1443
rc = nd_intel_test_cmd_set_lss_status(t,
tools/testing/nvdimm/test/nfit.c
1447
rc = nd_intel_test_get_fw_info(t, buf,
tools/testing/nvdimm/test/nfit.c
1451
rc = nd_intel_test_start_update(t, buf,
tools/testing/nvdimm/test/nfit.c
1455
rc = nd_intel_test_send_data(t, buf,
tools/testing/nvdimm/test/nfit.c
1459
rc = nd_intel_test_finish_fw(t, buf,
tools/testing/nvdimm/test/nfit.c
1463
rc = nd_intel_test_finish_query(t, buf,
tools/testing/nvdimm/test/nfit.c
1468
&t->smart[i]);
tools/testing/nvdimm/test/nfit.c
1473
&t->smart_threshold[i]);
tools/testing/nvdimm/test/nfit.c
1478
&t->smart_threshold[i],
tools/testing/nvdimm/test/nfit.c
1479
&t->smart[i],
tools/testing/nvdimm/test/nfit.c
1480
&t->pdev.dev, t->dimm_dev[i]);
tools/testing/nvdimm/test/nfit.c
1485
&t->smart_threshold[i],
tools/testing/nvdimm/test/nfit.c
1486
&t->smart[i],
tools/testing/nvdimm/test/nfit.c
1487
&t->pdev.dev, t->dimm_dev[i]);
tools/testing/nvdimm/test/nfit.c
1509
t->label[i - t->dcr_idx]);
tools/testing/nvdimm/test/nfit.c
1513
t->label[i - t->dcr_idx]);
tools/testing/nvdimm/test/nfit.c
1520
struct ars_state *ars_state = &t->ars_state;
tools/testing/nvdimm/test/nfit.c
1538
rc = nfit_test_cmd_ars_error_inject(t, buf,
tools/testing/nvdimm/test/nfit.c
1542
rc = nfit_test_cmd_ars_inject_clear(t, buf,
tools/testing/nvdimm/test/nfit.c
1546
rc = nfit_test_cmd_ars_inject_status(t, buf,
tools/testing/nvdimm/test/nfit.c
1560
rc = nvdimm_bus_intel_fw_activate_businfo(t,
tools/testing/nvdimm/test/nfit.c
1564
rc = nvdimm_bus_intel_fw_activate(t, buf,
tools/testing/nvdimm/test/nfit.c
1581
rc = nfit_test_cmd_ars_start(t, ars_state, buf,
tools/testing/nvdimm/test/nfit.c
1589
rc = nfit_test_cmd_clear_error(t, buf, buf_len, cmd_rc);
tools/testing/nvdimm/test/nfit.c
1617
static void *__test_alloc(struct nfit_test *t, size_t size, dma_addr_t *dma,
tools/testing/nvdimm/test/nfit.c
1620
struct device *dev = &t->pdev.dev;
tools/testing/nvdimm/test/nfit.c
1640
list_add(&nfit_res->list, &t->resources);
tools/testing/nvdimm/test/nfit.c
1653
static void *test_alloc(struct nfit_test *t, size_t size, dma_addr_t *dma)
tools/testing/nvdimm/test/nfit.c
1665
return __test_alloc(t, size, dma, buf);
tools/testing/nvdimm/test/nfit.c
1674
struct nfit_test *t = instances[i];
tools/testing/nvdimm/test/nfit.c
1676
if (!t)
tools/testing/nvdimm/test/nfit.c
1679
list_for_each_entry(n, &t->resources, list) {
tools/testing/nvdimm/test/nfit.c
1712
struct nfit_test *t = data;
tools/testing/nvdimm/test/nfit.c
1715
for (i = 0; i < t->num_dcr; i++)
tools/testing/nvdimm/test/nfit.c
1716
if (t->dimm_dev[i])
tools/testing/nvdimm/test/nfit.c
1717
device_unregister(t->dimm_dev[i]);
tools/testing/nvdimm/test/nfit.c
1833
static int nfit_test_dimm_init(struct nfit_test *t)
tools/testing/nvdimm/test/nfit.c
1837
if (devm_add_action_or_reset(&t->pdev.dev, put_dimms, t))
tools/testing/nvdimm/test/nfit.c
1839
for (i = 0; i < t->num_dcr; i++) {
tools/testing/nvdimm/test/nfit.c
1840
t->dimm_dev[i] = device_create_with_groups(&nfit_test_dimm,
tools/testing/nvdimm/test/nfit.c
1841
&t->pdev.dev, 0, NULL,
tools/testing/nvdimm/test/nfit.c
1843
"test_dimm%d", i + t->dcr_idx);
tools/testing/nvdimm/test/nfit.c
1844
if (!t->dimm_dev[i])
tools/testing/nvdimm/test/nfit.c
1850
static void nfit_security_init(struct nfit_test *t)
tools/testing/nvdimm/test/nfit.c
1854
for (i = 0; i < t->num_dcr; i++) {
tools/testing/nvdimm/test/nfit.c
1861
static void smart_init(struct nfit_test *t)
tools/testing/nvdimm/test/nfit.c
1872
for (i = 0; i < t->num_dcr; i++) {
tools/testing/nvdimm/test/nfit.c
1873
memcpy(&t->smart[i], &smart_def, sizeof(smart_def));
tools/testing/nvdimm/test/nfit.c
1874
memcpy(&t->smart_threshold[i], &smart_t_data,
tools/testing/nvdimm/test/nfit.c
1885
static int nfit_test0_alloc(struct nfit_test *t)
tools/testing/nvdimm/test/nfit.c
1899
t->nfit_buf = test_alloc(t, nfit_size, &t->nfit_dma);
tools/testing/nvdimm/test/nfit.c
1900
if (!t->nfit_buf)
tools/testing/nvdimm/test/nfit.c
1902
t->nfit_size = nfit_size;
tools/testing/nvdimm/test/nfit.c
1904
t->spa_set[0] = test_alloc(t, SPA0_SIZE, &t->spa_set_dma[0]);
tools/testing/nvdimm/test/nfit.c
1905
if (!t->spa_set[0])
tools/testing/nvdimm/test/nfit.c
1908
t->spa_set[1] = test_alloc(t, SPA1_SIZE, &t->spa_set_dma[1]);
tools/testing/nvdimm/test/nfit.c
1909
if (!t->spa_set[1])
tools/testing/nvdimm/test/nfit.c
1912
t->spa_set[2] = test_alloc(t, SPA0_SIZE, &t->spa_set_dma[2]);
tools/testing/nvdimm/test/nfit.c
1913
if (!t->spa_set[2])
tools/testing/nvdimm/test/nfit.c
1916
for (i = 0; i < t->num_dcr; i++) {
tools/testing/nvdimm/test/nfit.c
1917
t->dimm[i] = test_alloc(t, DIMM_SIZE, &t->dimm_dma[i]);
tools/testing/nvdimm/test/nfit.c
1918
if (!t->dimm[i])
tools/testing/nvdimm/test/nfit.c
1921
t->label[i] = test_alloc(t, LABEL_SIZE, &t->label_dma[i]);
tools/testing/nvdimm/test/nfit.c
1922
if (!t->label[i])
tools/testing/nvdimm/test/nfit.c
1924
sprintf(t->label[i], "label%d", i);
tools/testing/nvdimm/test/nfit.c
1926
t->flush[i] = test_alloc(t, max(PAGE_SIZE,
tools/testing/nvdimm/test/nfit.c
1928
&t->flush_dma[i]);
tools/testing/nvdimm/test/nfit.c
1929
if (!t->flush[i])
tools/testing/nvdimm/test/nfit.c
1933
for (i = 0; i < t->num_dcr; i++) {
tools/testing/nvdimm/test/nfit.c
1934
t->dcr[i] = test_alloc(t, LABEL_SIZE, &t->dcr_dma[i]);
tools/testing/nvdimm/test/nfit.c
1935
if (!t->dcr[i])
tools/testing/nvdimm/test/nfit.c
1939
t->_fit = test_alloc(t, sizeof(union acpi_object **), &t->_fit_dma);
tools/testing/nvdimm/test/nfit.c
1940
if (!t->_fit)
tools/testing/nvdimm/test/nfit.c
1943
if (nfit_test_dimm_init(t))
tools/testing/nvdimm/test/nfit.c
1945
smart_init(t);
tools/testing/nvdimm/test/nfit.c
1946
nfit_security_init(t);
tools/testing/nvdimm/test/nfit.c
1947
return ars_state_init(&t->pdev.dev, &t->ars_state);
tools/testing/nvdimm/test/nfit.c
1950
static int nfit_test1_alloc(struct nfit_test *t)
tools/testing/nvdimm/test/nfit.c
1958
t->nfit_buf = test_alloc(t, nfit_size, &t->nfit_dma);
tools/testing/nvdimm/test/nfit.c
1959
if (!t->nfit_buf)
tools/testing/nvdimm/test/nfit.c
1961
t->nfit_size = nfit_size;
tools/testing/nvdimm/test/nfit.c
1963
t->spa_set[0] = test_alloc(t, SPA2_SIZE, &t->spa_set_dma[0]);
tools/testing/nvdimm/test/nfit.c
1964
if (!t->spa_set[0])
tools/testing/nvdimm/test/nfit.c
1967
for (i = 0; i < t->num_dcr; i++) {
tools/testing/nvdimm/test/nfit.c
1968
t->label[i] = test_alloc(t, LABEL_SIZE, &t->label_dma[i]);
tools/testing/nvdimm/test/nfit.c
1969
if (!t->label[i])
tools/testing/nvdimm/test/nfit.c
1971
sprintf(t->label[i], "label%d", i);
tools/testing/nvdimm/test/nfit.c
1974
t->spa_set[1] = test_alloc(t, SPA_VCD_SIZE, &t->spa_set_dma[1]);
tools/testing/nvdimm/test/nfit.c
1975
if (!t->spa_set[1])
tools/testing/nvdimm/test/nfit.c
1978
if (nfit_test_dimm_init(t))
tools/testing/nvdimm/test/nfit.c
1980
smart_init(t);
tools/testing/nvdimm/test/nfit.c
1981
return ars_state_init(&t->pdev.dev, &t->ars_state);
tools/testing/nvdimm/test/nfit.c
1994
static void nfit_test0_setup(struct nfit_test *t)
tools/testing/nvdimm/test/nfit.c
200
int (*alloc)(struct nfit_test *t);
tools/testing/nvdimm/test/nfit.c
2000
void *nfit_buf = t->nfit_buf;
tools/testing/nvdimm/test/nfit.c
201
void (*setup)(struct nfit_test *t);
tools/testing/nvdimm/test/nfit.c
2019
spa->address = t->spa_set_dma[0];
tools/testing/nvdimm/test/nfit.c
2033
spa->address = t->spa_set_dma[1];
tools/testing/nvdimm/test/nfit.c
2043
spa->address = t->dcr_dma[0];
tools/testing/nvdimm/test/nfit.c
2053
spa->address = t->dcr_dma[1];
tools/testing/nvdimm/test/nfit.c
2063
spa->address = t->dcr_dma[2];
tools/testing/nvdimm/test/nfit.c
2073
spa->address = t->dcr_dma[3];
tools/testing/nvdimm/test/nfit.c
2083
spa->address = t->dimm_dma[0];
tools/testing/nvdimm/test/nfit.c
2093
spa->address = t->dimm_dma[1];
tools/testing/nvdimm/test/nfit.c
2103
spa->address = t->dimm_dma[2];
tools/testing/nvdimm/test/nfit.c
2113
spa->address = t->dimm_dma[3];
tools/testing/nvdimm/test/nfit.c
231
static int nd_intel_test_get_fw_info(struct nfit_test *t,
tools/testing/nvdimm/test/nfit.c
235
struct device *dev = &t->pdev.dev;
tools/testing/nvdimm/test/nfit.c
236
struct nfit_test_fw *fw = &t->fw[idx];
tools/testing/nvdimm/test/nfit.c
239
__func__, t, nd_cmd, buf_len, idx);
tools/testing/nvdimm/test/nfit.c
2512
flush->hint_address[i] = t->flush_dma[0] + i * sizeof(u64);
tools/testing/nvdimm/test/nfit.c
2522
flush->hint_address[i] = t->flush_dma[1] + i * sizeof(u64);
tools/testing/nvdimm/test/nfit.c
2532
flush->hint_address[i] = t->flush_dma[2] + i * sizeof(u64);
tools/testing/nvdimm/test/nfit.c
2542
flush->hint_address[i] = t->flush_dma[3] + i * sizeof(u64);
tools/testing/nvdimm/test/nfit.c
2553
if (t->setup_hotplug) {
tools/testing/nvdimm/test/nfit.c
257
static int nd_intel_test_start_update(struct nfit_test *t,
tools/testing/nvdimm/test/nfit.c
2600
spa->address = t->dcr_dma[4];
tools/testing/nvdimm/test/nfit.c
261
struct device *dev = &t->pdev.dev;
tools/testing/nvdimm/test/nfit.c
2614
spa->address = t->spa_set_dma[2];
tools/testing/nvdimm/test/nfit.c
262
struct nfit_test_fw *fw = &t->fw[idx];
tools/testing/nvdimm/test/nfit.c
2624
spa->address = t->dimm_dma[4];
tools/testing/nvdimm/test/nfit.c
265
__func__, t, nd_cmd, buf_len, idx);
tools/testing/nvdimm/test/nfit.c
2684
flush->hint_address[i] = t->flush_dma[4]
tools/testing/nvdimm/test/nfit.c
2689
WARN_ON(offset != t->nfit_size);
tools/testing/nvdimm/test/nfit.c
2692
t->nfit_filled = offset;
tools/testing/nvdimm/test/nfit.c
2694
post_ars_status(&t->ars_state, &t->badrange, t->spa_set_dma[0],
tools/testing/nvdimm/test/nfit.c
2697
acpi_desc = &t->acpi_desc;
tools/testing/nvdimm/test/nfit.c
2742
static void nfit_test1_setup(struct nfit_test *t)
tools/testing/nvdimm/test/nfit.c
2745
void *nfit_buf = t->nfit_buf;
tools/testing/nvdimm/test/nfit.c
2758
spa->address = t->spa_set_dma[0];
tools/testing/nvdimm/test/nfit.c
2768
spa->address = t->spa_set_dma[1];
tools/testing/nvdimm/test/nfit.c
2832
WARN_ON(offset != t->nfit_size);
tools/testing/nvdimm/test/nfit.c
2834
t->nfit_filled = offset;
tools/testing/nvdimm/test/nfit.c
2836
post_ars_status(&t->ars_state, &t->badrange, t->spa_set_dma[0],
tools/testing/nvdimm/test/nfit.c
2839
acpi_desc = &t->acpi_desc;
tools/testing/nvdimm/test/nfit.c
287
static int nd_intel_test_send_data(struct nfit_test *t,
tools/testing/nvdimm/test/nfit.c
291
struct device *dev = &t->pdev.dev;
tools/testing/nvdimm/test/nfit.c
292
struct nfit_test_fw *fw = &t->fw[idx];
tools/testing/nvdimm/test/nfit.c
296
__func__, t, nd_cmd, buf_len, idx);
tools/testing/nvdimm/test/nfit.c
338
static int nd_intel_test_finish_fw(struct nfit_test *t,
tools/testing/nvdimm/test/nfit.c
342
struct device *dev = &t->pdev.dev;
tools/testing/nvdimm/test/nfit.c
343
struct nfit_test_fw *fw = &t->fw[idx];
tools/testing/nvdimm/test/nfit.c
346
__func__, t, nd_cmd, buf_len, idx);
tools/testing/nvdimm/test/nfit.c
389
static int nd_intel_test_finish_query(struct nfit_test *t,
tools/testing/nvdimm/test/nfit.c
393
struct device *dev = &t->pdev.dev;
tools/testing/nvdimm/test/nfit.c
394
struct nfit_test_fw *fw = &t->fw[idx];
tools/testing/nvdimm/test/nfit.c
397
__func__, t, nd_cmd, buf_len, idx);
tools/testing/nvdimm/test/nfit.c
568
static int nfit_test_cmd_ars_start(struct nfit_test *t,
tools/testing/nvdimm/test/nfit.c
583
post_ars_status(ars_state, &t->badrange, ars_start->address,
tools/testing/nvdimm/test/nfit.c
614
static int nfit_test_cmd_clear_error(struct nfit_test *t,
tools/testing/nvdimm/test/nfit.c
625
badrange_forget(&t->badrange, clear_err->address, clear_err->length);
tools/testing/nvdimm/test/nfit.c
823
struct nfit_test *t = container_of(work, typeof(*t), work);
tools/testing/nvdimm/test/nfit.c
825
__acpi_nfit_notify(&t->pdev.dev, t, NFIT_NOTIFY_UC_MEMORY_ERROR);
tools/testing/nvdimm/test/nfit.c
828
static int nfit_test_cmd_ars_error_inject(struct nfit_test *t,
tools/testing/nvdimm/test/nfit.c
843
rc = badrange_add(&t->badrange, err_inj->err_inj_spa_range_base,
tools/testing/nvdimm/test/nfit.c
849
queue_work(nfit_wq, &t->work);
tools/testing/nvdimm/test/nfit.c
859
static int nfit_test_cmd_ars_inject_clear(struct nfit_test *t,
tools/testing/nvdimm/test/nfit.c
874
badrange_forget(&t->badrange, err_clr->err_inj_clr_spa_range_base,
tools/testing/nvdimm/test/nfit.c
885
static int nfit_test_cmd_ars_inject_status(struct nfit_test *t,
tools/testing/nvdimm/test/nfit.c
894
spin_lock(&t->badrange.lock);
tools/testing/nvdimm/test/nfit.c
895
list_for_each_entry(be, &t->badrange.list, list) {
tools/testing/nvdimm/test/nfit.c
902
spin_unlock(&t->badrange.lock);
tools/testing/nvdimm/test/nfit.c
908
static int nd_intel_test_cmd_set_lss_status(struct nfit_test *t,
tools/testing/nvdimm/test/nfit.c
911
struct device *dev = &t->pdev.dev;
tools/testing/nvdimm/test/nfit.c
947
static int nd_intel_test_cmd_security_status(struct nfit_test *t,
tools/testing/nvdimm/test/nfit.c
951
struct device *dev = &t->pdev.dev;
tools/testing/nvdimm/test/nfit.c
962
static int nd_intel_test_cmd_unlock_unit(struct nfit_test *t,
tools/testing/nvdimm/test/nfit.c
966
struct device *dev = &t->pdev.dev;
tools/testing/nvdimm/test/nfit.c
988
static int nd_intel_test_cmd_set_pass(struct nfit_test *t,
tools/testing/nvdimm/test/nfit.c
992
struct device *dev = &t->pdev.dev;
tools/testing/selftests/arm64/fp/vlset.c
107
int t, e;
tools/testing/selftests/arm64/fp/vlset.c
134
t = prctl(set_ctl, vl | flags);
tools/testing/selftests/arm64/fp/vlset.c
135
if (t < 0) {
tools/testing/selftests/arm64/fp/vlset.c
141
t = prctl(get_ctl);
tools/testing/selftests/arm64/fp/vlset.c
142
if (t == -1) {
tools/testing/selftests/arm64/fp/vlset.c
148
flags = t & ~flags;
tools/testing/selftests/arm64/mte/check_user_mem.c
197
int t, s, m, l, o;
tools/testing/selftests/arm64/mte/check_user_mem.c
219
for (t = 0; t < LAST_TEST; t++) {
tools/testing/selftests/arm64/mte/check_user_mem.c
230
tag_len, t);
tools/testing/selftests/arm64/mte/check_user_mem.c
232
t, sync, map, tag_len, offset);
tools/testing/selftests/bpf/benchs/bench_local_storage_create.c
117
struct thread *t = &threads[i];
tools/testing/selftests/bpf/benchs/bench_local_storage_create.c
120
t->fds = malloc(batch_sz * sizeof(*t->fds));
tools/testing/selftests/bpf/benchs/bench_local_storage_create.c
121
if (!t->fds) {
tools/testing/selftests/bpf/benchs/bench_local_storage_create.c
126
t->pthds = malloc(batch_sz * sizeof(*t->pthds));
tools/testing/selftests/bpf/benchs/bench_local_storage_create.c
127
if (!t->pthds) {
tools/testing/selftests/bpf/benchs/bench_local_storage_create.c
131
t->pthd_results = malloc(batch_sz * sizeof(*t->pthd_results));
tools/testing/selftests/bpf/benchs/bench_local_storage_create.c
132
if (!t->pthd_results) {
tools/testing/selftests/bpf/benchs/bench_local_storage_create.c
148
struct thread *t = &threads[(long)(input)];
tools/testing/selftests/bpf/benchs/bench_local_storage_create.c
149
int *fds = t->fds;
tools/testing/selftests/bpf/benchs/bench_local_storage_create.c
175
struct thread *t = &threads[(long)(input)];
tools/testing/selftests/bpf/benchs/bench_local_storage_create.c
176
pthread_t *pthds = t->pthds;
tools/testing/selftests/bpf/benchs/bench_local_storage_create.c
177
int *pthd_results = t->pthd_results;
tools/testing/selftests/bpf/bpf_experimental.h
499
static inline void __bpf_preempt_destructor(__bpf_preempt_t *t)
tools/testing/selftests/bpf/btf_helpers.c
101
fprintf(out, "[%u] %s '%s'", id, btf_kind_str(kind), btf_str(btf, t->name_off));
tools/testing/selftests/bpf/btf_helpers.c
106
t->size, btf_int_offset(t), btf_int_bits(t),
tools/testing/selftests/bpf/btf_helpers.c
107
btf_int_enc_str(btf_int_encoding(t)));
tools/testing/selftests/bpf/btf_helpers.c
115
fprintf(out, " type_id=%u", t->type);
tools/testing/selftests/bpf/btf_helpers.c
118
const struct btf_array *arr = btf_array(t);
tools/testing/selftests/bpf/btf_helpers.c
126
const struct btf_member *m = btf_members(t);
tools/testing/selftests/bpf/btf_helpers.c
128
fprintf(out, " size=%u vlen=%u", t->size, vlen);
tools/testing/selftests/bpf/btf_helpers.c
132
bit_off = btf_member_bit_offset(t, i);
tools/testing/selftests/bpf/btf_helpers.c
133
bit_sz = btf_member_bitfield_size(t, i);
tools/testing/selftests/bpf/btf_helpers.c
142
const struct btf_enum *v = btf_enum(t);
tools/testing/selftests/bpf/btf_helpers.c
145
fmt_str = btf_kflag(t) ? "\n\t'%s' val=%d" : "\n\t'%s' val=%u";
tools/testing/selftests/bpf/btf_helpers.c
147
btf_kflag(t) ? "SIGNED" : "UNSIGNED", t->size, vlen);
tools/testing/selftests/bpf/btf_helpers.c
155
const struct btf_enum64 *v = btf_enum64(t);
tools/testing/selftests/bpf/btf_helpers.c
158
fmt_str = btf_kflag(t) ? "\n\t'%s' val=%lld" : "\n\t'%s' val=%llu";
tools/testing/selftests/bpf/btf_helpers.c
161
btf_kflag(t) ? "SIGNED" : "UNSIGNED", t->size, vlen);
tools/testing/selftests/bpf/btf_helpers.c
170
fprintf(out, " fwd_kind=%s", btf_kflag(t) ? "union" : "struct");
tools/testing/selftests/bpf/btf_helpers.c
173
fprintf(out, " type_id=%u linkage=%s", t->type, btf_func_linkage_str(t));
tools/testing/selftests/bpf/btf_helpers.c
176
const struct btf_param *p = btf_params(t);
tools/testing/selftests/bpf/btf_helpers.c
178
fprintf(out, " ret_type_id=%u vlen=%u", t->type, vlen);
tools/testing/selftests/bpf/btf_helpers.c
187
t->type, btf_var_linkage_str(btf_var(t)->linkage));
tools/testing/selftests/bpf/btf_helpers.c
190
const struct btf_var_secinfo *v = btf_var_secinfos(t);
tools/testing/selftests/bpf/btf_helpers.c
192
fprintf(out, " size=%u vlen=%u", t->size, vlen);
tools/testing/selftests/bpf/btf_helpers.c
200
fprintf(out, " size=%u", t->size);
tools/testing/selftests/bpf/btf_helpers.c
204
t->type, btf_decl_tag(t)->component_idx);
tools/testing/selftests/bpf/btf_helpers.c
67
static const char *btf_func_linkage_str(const struct btf_type *t)
tools/testing/selftests/bpf/btf_helpers.c
69
switch (btf_vlen(t)) {
tools/testing/selftests/bpf/btf_helpers.c
90
const struct btf_type *t;
tools/testing/selftests/bpf/btf_helpers.c
94
t = btf__type_by_id(btf, id);
tools/testing/selftests/bpf/btf_helpers.c
95
if (!t)
tools/testing/selftests/bpf/btf_helpers.c
98
vlen = btf_vlen(t);
tools/testing/selftests/bpf/btf_helpers.c
99
kind = btf_kind(t);
tools/testing/selftests/bpf/map_tests/lpm_trie_map_basic_ops.c
231
struct tlpm_node *t, *list = NULL;
tools/testing/selftests/bpf/map_tests/lpm_trie_map_basic_ops.c
279
t = tlpm_match(list, data, 8 * keysize);
tools/testing/selftests/bpf/map_tests/lpm_trie_map_basic_ops.c
285
assert(!t == !!r);
tools/testing/selftests/bpf/map_tests/lpm_trie_map_basic_ops.c
287
if (t) {
tools/testing/selftests/bpf/map_tests/lpm_trie_map_basic_ops.c
289
assert(t->n_bits == value[keysize]);
tools/testing/selftests/bpf/map_tests/lpm_trie_map_basic_ops.c
290
for (j = 0; j < t->n_bits; ++j)
tools/testing/selftests/bpf/map_tests/lpm_trie_map_basic_ops.c
291
assert((t->key[j / 8] & (1 << (7 - j % 8))) ==
tools/testing/selftests/bpf/map_tests/lpm_trie_map_basic_ops.c
302
for (i = 0, t = list; t; i++, t = t->next)
tools/testing/selftests/bpf/map_tests/lpm_trie_map_basic_ops.c
316
t = tlpm_match(list, data, 8 * keysize);
tools/testing/selftests/bpf/map_tests/lpm_trie_map_basic_ops.c
322
assert(!t == !!r);
tools/testing/selftests/bpf/map_tests/lpm_trie_map_basic_ops.c
324
if (t) {
tools/testing/selftests/bpf/map_tests/lpm_trie_map_basic_ops.c
326
assert(t->n_bits == value[keysize]);
tools/testing/selftests/bpf/map_tests/lpm_trie_map_basic_ops.c
327
for (j = 0; j < t->n_bits; ++j)
tools/testing/selftests/bpf/map_tests/lpm_trie_map_basic_ops.c
328
assert((t->key[j / 8] & (1 << (7 - j % 8))) ==
tools/testing/selftests/bpf/prog_tests/btf.c
4959
const struct btf_type *t;
tools/testing/selftests/bpf/prog_tests/btf.c
4962
t = btf__type_by_id(btf, finfo->type_id);
tools/testing/selftests/bpf/prog_tests/btf.c
4963
if (CHECK(!t, "btf__type_by_id failure: id %u",
tools/testing/selftests/bpf/prog_tests/btf.c
4969
fname = btf__name_by_offset(btf, t->name_off);
tools/testing/selftests/bpf/prog_tests/btf.c
8092
static int btf_type_size(const struct btf_type *t)
tools/testing/selftests/bpf/prog_tests/btf.c
8095
__u16 vlen = BTF_INFO_VLEN(t->info);
tools/testing/selftests/bpf/prog_tests/btf.c
8096
__u16 kind = BTF_INFO_KIND(t->info);
tools/testing/selftests/bpf/prog_tests/btf_dedup_split.c
46
t = btf__type_by_id(btf2, 1);
tools/testing/selftests/bpf/prog_tests/btf_dedup_split.c
47
if (!ASSERT_OK_PTR(t, "int_type"))
tools/testing/selftests/bpf/prog_tests/btf_dedup_split.c
489
const struct btf_type *t;
tools/testing/selftests/bpf/prog_tests/btf_dedup_split.c
49
ASSERT_EQ(btf_is_int(t), true, "int_kind");
tools/testing/selftests/bpf/prog_tests/btf_dedup_split.c
497
t = btf__type_by_id(btf1, id);
tools/testing/selftests/bpf/prog_tests/btf_dedup_split.c
498
if (!ASSERT_OK_PTR(t, "func_id_type"))
tools/testing/selftests/bpf/prog_tests/btf_dedup_split.c
50
ASSERT_STREQ(btf__str_by_offset(btf2, t->name_off), "int", "int_name");
tools/testing/selftests/bpf/prog_tests/btf_dedup_split.c
500
t = btf__type_by_id(btf1, t->type);
tools/testing/selftests/bpf/prog_tests/btf_dedup_split.c
501
if (!ASSERT_OK_PTR(t, "func_proto_id_type"))
tools/testing/selftests/bpf/prog_tests/btf_dedup_split.c
503
if (!ASSERT_EQ(btf_is_func_proto(t), true, "is_func_proto"))
tools/testing/selftests/bpf/prog_tests/btf_dedup_split.c
505
vlen = btf_vlen(t);
tools/testing/selftests/bpf/prog_tests/btf_dedup_split.c
507
for (j = 0, p = btf_params(t); j < vlen; j++, p++) {
tools/testing/selftests/bpf/prog_tests/btf_dedup_split.c
525
t = btf__type_by_id(btf1, id);
tools/testing/selftests/bpf/prog_tests/btf_dedup_split.c
526
if (!ASSERT_OK_PTR(t, "param_ref_type"))
tools/testing/selftests/bpf/prog_tests/btf_dedup_split.c
528
if (!btf_is_mod(t) && !btf_is_ptr(t) && !btf_is_typedef(t))
tools/testing/selftests/bpf/prog_tests/btf_dedup_split.c
530
id = t->type;
tools/testing/selftests/bpf/prog_tests/btf_dedup_split.c
8
const struct btf_type *t;
tools/testing/selftests/bpf/prog_tests/btf_dump.c
102
snprintf(test_file, sizeof(test_file), "%s.c", t->file);
tools/testing/selftests/bpf/prog_tests/btf_dump.c
1053
struct btf_dump_test_case *t = &btf_dump_test_cases[i];
tools/testing/selftests/bpf/prog_tests/btf_dump.c
1055
if (!test__start_subtest(t->name))
tools/testing/selftests/bpf/prog_tests/btf_dump.c
137
static void test_ctx__free(struct test_ctx *t)
tools/testing/selftests/bpf/prog_tests/btf_dump.c
139
fclose(t->dump_buf_file);
tools/testing/selftests/bpf/prog_tests/btf_dump.c
140
free(t->dump_buf);
tools/testing/selftests/bpf/prog_tests/btf_dump.c
141
btf_dump__free(t->d);
tools/testing/selftests/bpf/prog_tests/btf_dump.c
142
btf__free(t->btf);
tools/testing/selftests/bpf/prog_tests/btf_dump.c
145
static int test_ctx__init(struct test_ctx *t)
tools/testing/selftests/bpf/prog_tests/btf_dump.c
147
t->dump_buf_file = open_memstream(&t->dump_buf, &t->dump_buf_sz);
tools/testing/selftests/bpf/prog_tests/btf_dump.c
148
if (!ASSERT_OK_PTR(t->dump_buf_file, "dump_memstream"))
tools/testing/selftests/bpf/prog_tests/btf_dump.c
150
t->btf = btf__new_empty();
tools/testing/selftests/bpf/prog_tests/btf_dump.c
151
if (!ASSERT_OK_PTR(t->btf, "new_empty"))
tools/testing/selftests/bpf/prog_tests/btf_dump.c
153
t->d = btf_dump__new(t->btf, btf_dump_printf, t->dump_buf_file, NULL);
tools/testing/selftests/bpf/prog_tests/btf_dump.c
154
if (!ASSERT_OK(libbpf_get_error(t->d), "btf_dump__new"))
tools/testing/selftests/bpf/prog_tests/btf_dump.c
160
test_ctx__free(t);
tools/testing/selftests/bpf/prog_tests/btf_dump.c
164
static void test_ctx__dump_and_compare(struct test_ctx *t,
tools/testing/selftests/bpf/prog_tests/btf_dump.c
170
for (i = 1; i < btf__type_cnt(t->btf); i++) {
tools/testing/selftests/bpf/prog_tests/btf_dump.c
171
err = btf_dump__dump_type(t->d, i);
tools/testing/selftests/bpf/prog_tests/btf_dump.c
175
fflush(t->dump_buf_file);
tools/testing/selftests/bpf/prog_tests/btf_dump.c
176
t->dump_buf[t->dump_buf_sz] = 0; /* some libc implementations don't do this */
tools/testing/selftests/bpf/prog_tests/btf_dump.c
178
ASSERT_STREQ(t->dump_buf, expected_output, message);
tools/testing/selftests/bpf/prog_tests/btf_dump.c
183
struct test_ctx t = {};
tools/testing/selftests/bpf/prog_tests/btf_dump.c
187
if (test_ctx__init(&t))
tools/testing/selftests/bpf/prog_tests/btf_dump.c
190
btf = t.btf;
tools/testing/selftests/bpf/prog_tests/btf_dump.c
228
test_ctx__dump_and_compare(&t,
tools/testing/selftests/bpf/prog_tests/btf_dump.c
259
fseek(t.dump_buf_file, 0, SEEK_SET);
tools/testing/selftests/bpf/prog_tests/btf_dump.c
270
test_ctx__dump_and_compare(&t,
tools/testing/selftests/bpf/prog_tests/btf_dump.c
279
test_ctx__free(&t);
tools/testing/selftests/bpf/prog_tests/btf_dump.c
284
struct test_ctx t = {};
tools/testing/selftests/bpf/prog_tests/btf_dump.c
288
if (test_ctx__init(&t))
tools/testing/selftests/bpf/prog_tests/btf_dump.c
291
btf = t.btf;
tools/testing/selftests/bpf/prog_tests/btf_dump.c
319
test_ctx__dump_and_compare(&t,
tools/testing/selftests/bpf/prog_tests/btf_dump.c
325
test_ctx__free(&t);
tools/testing/selftests/bpf/prog_tests/btf_dump.c
48
static int test_btf_dump_case(int n, struct btf_dump_test_case *t)
tools/testing/selftests/bpf/prog_tests/btf_dump.c
55
snprintf(test_file, sizeof(test_file), "%s.bpf.o", t->file);
tools/testing/selftests/bpf/prog_tests/btf_dump.c
69
if (!t->known_ptr_sz) {
tools/testing/selftests/bpf/prog_tests/btf_dump.c
76
snprintf(out_file, sizeof(out_file), "/tmp/%s.output.XXXXXX", t->file);
tools/testing/selftests/bpf/prog_tests/btf_dump.c
96
snprintf(test_file, sizeof(test_file), "progs/%s.c", t->file);
tools/testing/selftests/bpf/prog_tests/btf_dump.c
964
struct test_ctx t = {};
tools/testing/selftests/bpf/prog_tests/btf_dump.c
971
if (test_ctx__init(&t))
tools/testing/selftests/bpf/prog_tests/btf_dump.c
974
d = btf_dump__new(t.btf, btf_dump_snprintf, str, NULL);
tools/testing/selftests/bpf/prog_tests/btf_dump.c
979
char_id = btf__add_int(t.btf, "char", 1, BTF_INT_CHAR);
tools/testing/selftests/bpf/prog_tests/btf_dump.c
981
int_id = btf__add_int(t.btf, "int", 4, BTF_INT_SIGNED);
tools/testing/selftests/bpf/prog_tests/btf_dump.c
983
array_id = btf__add_array(t.btf, int_id, char_id, 4);
tools/testing/selftests/bpf/prog_tests/btf_dump.c
986
ctx.btf = t.btf;
tools/testing/selftests/bpf/prog_tests/btf_dump.c
995
test_ctx__free(&t);
tools/testing/selftests/bpf/prog_tests/btf_endian.c
20
const struct btf_type *t;
tools/testing/selftests/bpf/prog_tests/btf_endian.c
91
t = btf__type_by_id(swap_btf, var_id);
tools/testing/selftests/bpf/prog_tests/btf_endian.c
92
ASSERT_STREQ(btf__str_by_offset(swap_btf, t->name_off), "some_var", "var_name");
tools/testing/selftests/bpf/prog_tests/btf_endian.c
93
ASSERT_EQ(btf_var(t)->linkage, BTF_VAR_GLOBAL_ALLOCATED, "var_linkage");
tools/testing/selftests/bpf/prog_tests/btf_endian.c
94
ASSERT_EQ(t->type, 1, "var_type");
tools/testing/selftests/bpf/prog_tests/btf_field_iter.c
129
struct btf_type *t = btf_type_by_id(btf, id);
tools/testing/selftests/bpf/prog_tests/btf_field_iter.c
134
if (!ASSERT_OK_PTR(t, "btf_type_by_id"))
tools/testing/selftests/bpf/prog_tests/btf_field_iter.c
136
if (!ASSERT_OK(btf_field_iter_init(&it_strs, t, BTF_FIELD_ITER_STRS),
tools/testing/selftests/bpf/prog_tests/btf_field_iter.c
139
if (!ASSERT_OK(btf_field_iter_init(&it_ids, t, BTF_FIELD_ITER_IDS),
tools/testing/selftests/bpf/prog_tests/btf_split.c
115
t = btf__type_by_id(btf2, 5);
tools/testing/selftests/bpf/prog_tests/btf_split.c
116
ASSERT_NULL(t, "multisplit_type_in_first_split");
tools/testing/selftests/bpf/prog_tests/btf_split.c
119
t = btf__type_by_id(btf3, 5);
tools/testing/selftests/bpf/prog_tests/btf_split.c
120
if (!ASSERT_OK_PTR(t, "split_union_type"))
tools/testing/selftests/bpf/prog_tests/btf_split.c
122
ASSERT_EQ(btf_is_union(t), true, "split_union_kind");
tools/testing/selftests/bpf/prog_tests/btf_split.c
123
ASSERT_EQ(btf_vlen(t), 2, "split_union_vlen");
tools/testing/selftests/bpf/prog_tests/btf_split.c
124
ASSERT_STREQ(btf__str_by_offset(btf3, t->name_off), "u1", "split_union_name");
tools/testing/selftests/bpf/prog_tests/btf_split.c
127
t = btf__type_by_id(btf3, 1);
tools/testing/selftests/bpf/prog_tests/btf_split.c
128
if (!ASSERT_OK_PTR(t, "split_base_type"))
tools/testing/selftests/bpf/prog_tests/btf_split.c
130
ASSERT_EQ(btf_is_int(t), true, "split_base_int");
tools/testing/selftests/bpf/prog_tests/btf_split.c
131
ASSERT_STREQ(btf__str_by_offset(btf3, t->name_off), "int", "split_base_type_name");
tools/testing/selftests/bpf/prog_tests/btf_split.c
189
t = btf__type_by_id(btf6, i);
tools/testing/selftests/bpf/prog_tests/btf_split.c
190
if (!ASSERT_OK_PTR(t, "type_in_parsed_btf"))
tools/testing/selftests/bpf/prog_tests/btf_split.c
195
if (!ASSERT_EQ(memcmp(t, ot, sizeof(*ot)), 0, "cmp_parsed_orig_btf"))
tools/testing/selftests/bpf/prog_tests/btf_split.c
51
const struct btf_type *t, *ot;
tools/testing/selftests/bpf/prog_tests/btf_split.c
79
t = btf__type_by_id(btf2, 1);
tools/testing/selftests/bpf/prog_tests/btf_split.c
80
if (!ASSERT_OK_PTR(t, "int_type"))
tools/testing/selftests/bpf/prog_tests/btf_split.c
82
ASSERT_EQ(btf_is_int(t), true, "int_kind");
tools/testing/selftests/bpf/prog_tests/btf_split.c
83
ASSERT_STREQ(btf__str_by_offset(btf2, t->name_off), "int", "int_name");
tools/testing/selftests/bpf/prog_tests/btf_split.c
91
t = btf__type_by_id(btf1, 4);
tools/testing/selftests/bpf/prog_tests/btf_split.c
92
ASSERT_NULL(t, "split_type_in_main");
tools/testing/selftests/bpf/prog_tests/btf_split.c
94
t = btf__type_by_id(btf2, 4);
tools/testing/selftests/bpf/prog_tests/btf_split.c
95
if (!ASSERT_OK_PTR(t, "split_struct_type"))
tools/testing/selftests/bpf/prog_tests/btf_split.c
97
ASSERT_EQ(btf_is_struct(t), true, "split_struct_kind");
tools/testing/selftests/bpf/prog_tests/btf_split.c
98
ASSERT_EQ(btf_vlen(t), 3, "split_struct_vlen");
tools/testing/selftests/bpf/prog_tests/btf_split.c
99
ASSERT_STREQ(btf__str_by_offset(btf2, t->name_off), "s2", "split_struct_name");
tools/testing/selftests/bpf/prog_tests/btf_write.c
10
const struct btf_type *t;
tools/testing/selftests/bpf/prog_tests/btf_write.c
108
t = btf__type_by_id(btf, 7);
tools/testing/selftests/bpf/prog_tests/btf_write.c
109
ASSERT_STREQ(btf__str_by_offset(btf, t->name_off), "s1", "struct_name");
tools/testing/selftests/bpf/prog_tests/btf_write.c
110
ASSERT_EQ(btf_kind(t), BTF_KIND_STRUCT, "struct_kind");
tools/testing/selftests/bpf/prog_tests/btf_write.c
111
ASSERT_EQ(btf_vlen(t), 2, "struct_vlen");
tools/testing/selftests/bpf/prog_tests/btf_write.c
112
ASSERT_EQ(btf_kflag(t), true, "struct_kflag");
tools/testing/selftests/bpf/prog_tests/btf_write.c
113
ASSERT_EQ(t->size, 8, "struct_sz");
tools/testing/selftests/bpf/prog_tests/btf_write.c
114
m = btf_members(t) + 0;
tools/testing/selftests/bpf/prog_tests/btf_write.c
117
ASSERT_EQ(btf_member_bit_offset(t, 0), 0, "f1_bit_off");
tools/testing/selftests/bpf/prog_tests/btf_write.c
118
ASSERT_EQ(btf_member_bitfield_size(t, 0), 0, "f1_bit_sz");
tools/testing/selftests/bpf/prog_tests/btf_write.c
119
m = btf_members(t) + 1;
tools/testing/selftests/bpf/prog_tests/btf_write.c
122
ASSERT_EQ(btf_member_bit_offset(t, 1), 32, "f2_bit_off");
tools/testing/selftests/bpf/prog_tests/btf_write.c
123
ASSERT_EQ(btf_member_bitfield_size(t, 1), 16, "f2_bit_sz");
tools/testing/selftests/bpf/prog_tests/btf_write.c
140
t = btf__type_by_id(btf, 8);
tools/testing/selftests/bpf/prog_tests/btf_write.c
141
ASSERT_STREQ(btf__str_by_offset(btf, t->name_off), "u1", "union_name");
tools/testing/selftests/bpf/prog_tests/btf_write.c
142
ASSERT_EQ(btf_kind(t), BTF_KIND_UNION, "union_kind");
tools/testing/selftests/bpf/prog_tests/btf_write.c
143
ASSERT_EQ(btf_vlen(t), 1, "union_vlen");
tools/testing/selftests/bpf/prog_tests/btf_write.c
144
ASSERT_EQ(btf_kflag(t), true, "union_kflag");
tools/testing/selftests/bpf/prog_tests/btf_write.c
145
ASSERT_EQ(t->size, 8, "union_sz");
tools/testing/selftests/bpf/prog_tests/btf_write.c
146
m = btf_members(t) + 0;
tools/testing/selftests/bpf/prog_tests/btf_write.c
149
ASSERT_EQ(btf_member_bit_offset(t, 0), 0, "f1_bit_off");
tools/testing/selftests/bpf/prog_tests/btf_write.c
150
ASSERT_EQ(btf_member_bitfield_size(t, 0), 16, "f1_bit_sz");
tools/testing/selftests/bpf/prog_tests/btf_write.c
163
t = btf__type_by_id(btf, 9);
tools/testing/selftests/bpf/prog_tests/btf_write.c
164
ASSERT_STREQ(btf__str_by_offset(btf, t->name_off), "e1", "enum_name");
tools/testing/selftests/bpf/prog_tests/btf_write.c
165
ASSERT_EQ(btf_kind(t), BTF_KIND_ENUM, "enum_kind");
tools/testing/selftests/bpf/prog_tests/btf_write.c
166
ASSERT_EQ(btf_vlen(t), 2, "enum_vlen");
tools/testing/selftests/bpf/prog_tests/btf_write.c
167
ASSERT_EQ(t->size, 4, "enum_sz");
tools/testing/selftests/bpf/prog_tests/btf_write.c
168
v = btf_enum(t) + 0;
tools/testing/selftests/bpf/prog_tests/btf_write.c
171
v = btf_enum(t) + 1;
tools/testing/selftests/bpf/prog_tests/btf_write.c
182
t = btf__type_by_id(btf, 10);
tools/testing/selftests/bpf/prog_tests/btf_write.c
183
ASSERT_STREQ(btf__str_by_offset(btf, t->name_off), "struct_fwd", "fwd_name");
tools/testing/selftests/bpf/prog_tests/btf_write.c
184
ASSERT_EQ(btf_kind(t), BTF_KIND_FWD, "fwd_kind");
tools/testing/selftests/bpf/prog_tests/btf_write.c
185
ASSERT_EQ(btf_kflag(t), 0, "fwd_kflag");
tools/testing/selftests/bpf/prog_tests/btf_write.c
191
t = btf__type_by_id(btf, 11);
tools/testing/selftests/bpf/prog_tests/btf_write.c
192
ASSERT_STREQ(btf__str_by_offset(btf, t->name_off), "union_fwd", "fwd_name");
tools/testing/selftests/bpf/prog_tests/btf_write.c
193
ASSERT_EQ(btf_kind(t), BTF_KIND_FWD, "fwd_kind");
tools/testing/selftests/bpf/prog_tests/btf_write.c
194
ASSERT_EQ(btf_kflag(t), 1, "fwd_kflag");
tools/testing/selftests/bpf/prog_tests/btf_write.c
200
t = btf__type_by_id(btf, 12);
tools/testing/selftests/bpf/prog_tests/btf_write.c
201
ASSERT_STREQ(btf__str_by_offset(btf, t->name_off), "enum_fwd", "fwd_name");
tools/testing/selftests/bpf/prog_tests/btf_write.c
202
ASSERT_EQ(btf_kind(t), BTF_KIND_ENUM, "enum_fwd_kind");
tools/testing/selftests/bpf/prog_tests/btf_write.c
203
ASSERT_EQ(btf_vlen(t), 0, "enum_fwd_kind");
tools/testing/selftests/bpf/prog_tests/btf_write.c
204
ASSERT_EQ(t->size, 4, "enum_fwd_sz");
tools/testing/selftests/bpf/prog_tests/btf_write.c
211
t = btf__type_by_id(btf, 13);
tools/testing/selftests/bpf/prog_tests/btf_write.c
212
ASSERT_STREQ(btf__str_by_offset(btf, t->name_off), "typedef1", "typedef_name");
tools/testing/selftests/bpf/prog_tests/btf_write.c
213
ASSERT_EQ(btf_kind(t), BTF_KIND_TYPEDEF, "typedef_kind");
tools/testing/selftests/bpf/prog_tests/btf_write.c
214
ASSERT_EQ(t->type, 1, "typedef_type");
tools/testing/selftests/bpf/prog_tests/btf_write.c
221
t = btf__type_by_id(btf, 14);
tools/testing/selftests/bpf/prog_tests/btf_write.c
222
ASSERT_STREQ(btf__str_by_offset(btf, t->name_off), "func1", "func_name");
tools/testing/selftests/bpf/prog_tests/btf_write.c
223
ASSERT_EQ(t->type, 15, "func_type");
tools/testing/selftests/bpf/prog_tests/btf_write.c
224
ASSERT_EQ(btf_kind(t), BTF_KIND_FUNC, "func_kind");
tools/testing/selftests/bpf/prog_tests/btf_write.c
225
ASSERT_EQ(btf_vlen(t), BTF_FUNC_GLOBAL, "func_vlen");
tools/testing/selftests/bpf/prog_tests/btf_write.c
236
t = btf__type_by_id(btf, 15);
tools/testing/selftests/bpf/prog_tests/btf_write.c
237
ASSERT_EQ(btf_kind(t), BTF_KIND_FUNC_PROTO, "func_proto_kind");
tools/testing/selftests/bpf/prog_tests/btf_write.c
238
ASSERT_EQ(btf_vlen(t), 2, "func_proto_vlen");
tools/testing/selftests/bpf/prog_tests/btf_write.c
239
ASSERT_EQ(t->type, 1, "func_proto_ret_type");
tools/testing/selftests/bpf/prog_tests/btf_write.c
240
p = btf_params(t) + 0;
tools/testing/selftests/bpf/prog_tests/btf_write.c
243
p = btf_params(t) + 1;
tools/testing/selftests/bpf/prog_tests/btf_write.c
254
t = btf__type_by_id(btf, 16);
tools/testing/selftests/bpf/prog_tests/btf_write.c
255
ASSERT_STREQ(btf__str_by_offset(btf, t->name_off), "var1", "var_name");
tools/testing/selftests/bpf/prog_tests/btf_write.c
256
ASSERT_EQ(btf_kind(t), BTF_KIND_VAR, "var_kind");
tools/testing/selftests/bpf/prog_tests/btf_write.c
257
ASSERT_EQ(t->type, 1, "var_type");
tools/testing/selftests/bpf/prog_tests/btf_write.c
258
ASSERT_EQ(btf_var(t)->linkage, BTF_VAR_GLOBAL_ALLOCATED, "var_type");
tools/testing/selftests/bpf/prog_tests/btf_write.c
268
t = btf__type_by_id(btf, 17);
tools/testing/selftests/bpf/prog_tests/btf_write.c
269
ASSERT_STREQ(btf__str_by_offset(btf, t->name_off), "datasec1", "datasec_name");
tools/testing/selftests/bpf/prog_tests/btf_write.c
270
ASSERT_EQ(t->size, 12, "datasec_sz");
tools/testing/selftests/bpf/prog_tests/btf_write.c
271
ASSERT_EQ(btf_kind(t), BTF_KIND_DATASEC, "datasec_kind");
tools/testing/selftests/bpf/prog_tests/btf_write.c
272
ASSERT_EQ(btf_vlen(t), 1, "datasec_vlen");
tools/testing/selftests/bpf/prog_tests/btf_write.c
273
vi = btf_var_secinfos(t) + 0;
tools/testing/selftests/bpf/prog_tests/btf_write.c
284
t = btf__type_by_id(btf, 18);
tools/testing/selftests/bpf/prog_tests/btf_write.c
285
ASSERT_STREQ(btf__str_by_offset(btf, t->name_off), "tag1", "tag_value");
tools/testing/selftests/bpf/prog_tests/btf_write.c
286
ASSERT_EQ(btf_kind(t), BTF_KIND_DECL_TAG, "tag_kind");
tools/testing/selftests/bpf/prog_tests/btf_write.c
287
ASSERT_EQ(t->type, 16, "tag_type");
tools/testing/selftests/bpf/prog_tests/btf_write.c
288
ASSERT_EQ(btf_decl_tag(t)->component_idx, -1, "tag_component_idx");
tools/testing/selftests/bpf/prog_tests/btf_write.c
294
t = btf__type_by_id(btf, 19);
tools/testing/selftests/bpf/prog_tests/btf_write.c
295
ASSERT_STREQ(btf__str_by_offset(btf, t->name_off), "tag2", "tag_value");
tools/testing/selftests/bpf/prog_tests/btf_write.c
296
ASSERT_EQ(btf_kind(t), BTF_KIND_DECL_TAG, "tag_kind");
tools/testing/selftests/bpf/prog_tests/btf_write.c
297
ASSERT_EQ(t->type, 14, "tag_type");
tools/testing/selftests/bpf/prog_tests/btf_write.c
298
ASSERT_EQ(btf_decl_tag(t)->component_idx, 1, "tag_component_idx");
tools/testing/selftests/bpf/prog_tests/btf_write.c
30
t = btf__type_by_id(btf, 1);
tools/testing/selftests/bpf/prog_tests/btf_write.c
305
t = btf__type_by_id(btf, 20);
tools/testing/selftests/bpf/prog_tests/btf_write.c
306
ASSERT_STREQ(btf__str_by_offset(btf, t->name_off), "tag1", "tag_value");
tools/testing/selftests/bpf/prog_tests/btf_write.c
307
ASSERT_EQ(btf_kind(t), BTF_KIND_TYPE_TAG, "tag_kind");
tools/testing/selftests/bpf/prog_tests/btf_write.c
308
ASSERT_EQ(t->type, 1, "tag_type");
tools/testing/selftests/bpf/prog_tests/btf_write.c
319
t = btf__type_by_id(btf, 21);
tools/testing/selftests/bpf/prog_tests/btf_write.c
32
ASSERT_EQ(t->name_off, str_off, "int_name_off");
tools/testing/selftests/bpf/prog_tests/btf_write.c
320
ASSERT_STREQ(btf__str_by_offset(btf, t->name_off), "e1", "enum64_name");
tools/testing/selftests/bpf/prog_tests/btf_write.c
321
ASSERT_EQ(btf_kind(t), BTF_KIND_ENUM64, "enum64_kind");
tools/testing/selftests/bpf/prog_tests/btf_write.c
322
ASSERT_EQ(btf_vlen(t), 2, "enum64_vlen");
tools/testing/selftests/bpf/prog_tests/btf_write.c
323
ASSERT_EQ(t->size, 8, "enum64_sz");
tools/testing/selftests/bpf/prog_tests/btf_write.c
324
v64 = btf_enum64(t) + 0;
tools/testing/selftests/bpf/prog_tests/btf_write.c
328
v64 = btf_enum64(t) + 1;
tools/testing/selftests/bpf/prog_tests/btf_write.c
33
ASSERT_STREQ(btf__str_by_offset(btf, t->name_off), "int", "int_name");
tools/testing/selftests/bpf/prog_tests/btf_write.c
34
ASSERT_EQ(btf_kind(t), BTF_KIND_INT, "int_kind");
tools/testing/selftests/bpf/prog_tests/btf_write.c
341
t = btf__type_by_id(btf, 22);
tools/testing/selftests/bpf/prog_tests/btf_write.c
342
ASSERT_STREQ(btf__str_by_offset(btf, t->name_off), "e1", "enum64_name");
tools/testing/selftests/bpf/prog_tests/btf_write.c
343
ASSERT_EQ(btf_kind(t), BTF_KIND_ENUM64, "enum64_kind");
tools/testing/selftests/bpf/prog_tests/btf_write.c
344
ASSERT_EQ(btf_vlen(t), 1, "enum64_vlen");
tools/testing/selftests/bpf/prog_tests/btf_write.c
345
ASSERT_EQ(t->size, 8, "enum64_sz");
tools/testing/selftests/bpf/prog_tests/btf_write.c
346
v64 = btf_enum64(t) + 0;
tools/testing/selftests/bpf/prog_tests/btf_write.c
35
ASSERT_EQ(t->size, 4, "int_sz");
tools/testing/selftests/bpf/prog_tests/btf_write.c
36
ASSERT_EQ(btf_int_encoding(t), BTF_INT_SIGNED, "int_enc");
tools/testing/selftests/bpf/prog_tests/btf_write.c
37
ASSERT_EQ(btf_int_bits(t), 32, "int_bits");
tools/testing/selftests/bpf/prog_tests/btf_write.c
57
t = btf__type_by_id(btf, 2);
tools/testing/selftests/bpf/prog_tests/btf_write.c
58
ASSERT_EQ(btf_kind(t), BTF_KIND_PTR, "ptr_kind");
tools/testing/selftests/bpf/prog_tests/btf_write.c
59
ASSERT_EQ(t->type, 1, "ptr_type");
tools/testing/selftests/bpf/prog_tests/btf_write.c
65
t = btf__type_by_id(btf, 3);
tools/testing/selftests/bpf/prog_tests/btf_write.c
66
ASSERT_EQ(btf_kind(t), BTF_KIND_CONST, "const_kind");
tools/testing/selftests/bpf/prog_tests/btf_write.c
67
ASSERT_EQ(t->type, 5, "const_type");
tools/testing/selftests/bpf/prog_tests/btf_write.c
73
t = btf__type_by_id(btf, 4);
tools/testing/selftests/bpf/prog_tests/btf_write.c
74
ASSERT_EQ(btf_kind(t), BTF_KIND_VOLATILE, "volatile_kind");
tools/testing/selftests/bpf/prog_tests/btf_write.c
75
ASSERT_EQ(t->type, 3, "volatile_type");
tools/testing/selftests/bpf/prog_tests/btf_write.c
81
t = btf__type_by_id(btf, 5);
tools/testing/selftests/bpf/prog_tests/btf_write.c
82
ASSERT_EQ(btf_kind(t), BTF_KIND_RESTRICT, "restrict_kind");
tools/testing/selftests/bpf/prog_tests/btf_write.c
83
ASSERT_EQ(t->type, 4, "restrict_type");
tools/testing/selftests/bpf/prog_tests/btf_write.c
90
t = btf__type_by_id(btf, 6);
tools/testing/selftests/bpf/prog_tests/btf_write.c
91
ASSERT_EQ(btf_kind(t), BTF_KIND_ARRAY, "array_kind");
tools/testing/selftests/bpf/prog_tests/btf_write.c
92
ASSERT_EQ(btf_array(t)->index_type, 1, "array_index_type");
tools/testing/selftests/bpf/prog_tests/btf_write.c
93
ASSERT_EQ(btf_array(t)->type, 2, "array_elem_type");
tools/testing/selftests/bpf/prog_tests/btf_write.c
94
ASSERT_EQ(btf_array(t)->nelems, 10, "array_nelems");
tools/testing/selftests/bpf/prog_tests/cgroup_ancestor.c
107
ASSERT_EQ(t->skel->bss->cgroup_ids[level], expected_ids[level],
tools/testing/selftests/bpf/prog_tests/cgroup_ancestor.c
114
struct test_data t;
tools/testing/selftests/bpf/prog_tests/cgroup_ancestor.c
117
t.skel = cgroup_ancestor__open_and_load();
tools/testing/selftests/bpf/prog_tests/cgroup_ancestor.c
118
if (!ASSERT_OK_PTR(t.skel, "open and load"))
tools/testing/selftests/bpf/prog_tests/cgroup_ancestor.c
121
t.skel->bss->dport = htons(DST_PORT);
tools/testing/selftests/bpf/prog_tests/cgroup_ancestor.c
126
if (setup_network(&t))
tools/testing/selftests/bpf/prog_tests/cgroup_ancestor.c
132
check_ancestors_ids(&t);
tools/testing/selftests/bpf/prog_tests/cgroup_ancestor.c
135
cleanup_network(&t);
tools/testing/selftests/bpf/prog_tests/cgroup_ancestor.c
140
cgroup_ancestor__destroy(t.skel);
tools/testing/selftests/bpf/prog_tests/cgroup_ancestor.c
49
static int setup_network(struct test_data *t)
tools/testing/selftests/bpf/prog_tests/cgroup_ancestor.c
52
t->ns = open_netns(TEST_NS);
tools/testing/selftests/bpf/prog_tests/cgroup_ancestor.c
53
if (!ASSERT_OK_PTR(t->ns, "open netns"))
tools/testing/selftests/bpf/prog_tests/cgroup_ancestor.c
58
memset(&t->qdisc, 0, sizeof(t->qdisc));
tools/testing/selftests/bpf/prog_tests/cgroup_ancestor.c
59
t->qdisc.sz = sizeof(t->qdisc);
tools/testing/selftests/bpf/prog_tests/cgroup_ancestor.c
60
t->qdisc.attach_point = BPF_TC_EGRESS;
tools/testing/selftests/bpf/prog_tests/cgroup_ancestor.c
61
t->qdisc.ifindex = if_nametoindex("lo");
tools/testing/selftests/bpf/prog_tests/cgroup_ancestor.c
62
if (!ASSERT_NEQ(t->qdisc.ifindex, 0, "if_nametoindex"))
tools/testing/selftests/bpf/prog_tests/cgroup_ancestor.c
64
if (!ASSERT_OK(bpf_tc_hook_create(&t->qdisc), "qdisc add"))
tools/testing/selftests/bpf/prog_tests/cgroup_ancestor.c
67
memset(&t->tc_attach, 0, sizeof(t->tc_attach));
tools/testing/selftests/bpf/prog_tests/cgroup_ancestor.c
68
t->tc_attach.sz = sizeof(t->tc_attach);
tools/testing/selftests/bpf/prog_tests/cgroup_ancestor.c
69
t->tc_attach.prog_fd = bpf_program__fd(t->skel->progs.log_cgroup_id);
tools/testing/selftests/bpf/prog_tests/cgroup_ancestor.c
70
if (!ASSERT_OK(bpf_tc_attach(&t->qdisc, &t->tc_attach), "filter add"))
tools/testing/selftests/bpf/prog_tests/cgroup_ancestor.c
76
bpf_tc_hook_destroy(&t->qdisc);
tools/testing/selftests/bpf/prog_tests/cgroup_ancestor.c
78
close_netns(t->ns);
tools/testing/selftests/bpf/prog_tests/cgroup_ancestor.c
85
static void cleanup_network(struct test_data *t)
tools/testing/selftests/bpf/prog_tests/cgroup_ancestor.c
87
bpf_tc_detach(&t->qdisc, &t->tc_attach);
tools/testing/selftests/bpf/prog_tests/cgroup_ancestor.c
88
bpf_tc_hook_destroy(&t->qdisc);
tools/testing/selftests/bpf/prog_tests/cgroup_ancestor.c
89
close_netns(t->ns);
tools/testing/selftests/bpf/prog_tests/cgroup_ancestor.c
93
static void check_ancestors_ids(struct test_data *t)
tools/testing/selftests/bpf/prog_tests/core_extern.c
131
struct test_case *t = &test_cases[i];
tools/testing/selftests/bpf/prog_tests/core_extern.c
133
.kconfig = t->cfg,
tools/testing/selftests/bpf/prog_tests/core_extern.c
136
if (!test__start_subtest(t->name))
tools/testing/selftests/bpf/prog_tests/core_extern.c
143
if (t->fails) {
tools/testing/selftests/bpf/prog_tests/core_extern.c
155
t->data.kern_ver = kern_ver;
tools/testing/selftests/bpf/prog_tests/core_extern.c
156
t->data.missing_val = 0xDEADC0DE;
tools/testing/selftests/bpf/prog_tests/core_extern.c
158
exp = (uint64_t *)&t->data;
tools/testing/selftests/bpf/prog_tests/core_reloc.c
430
const struct btf_type *t;
tools/testing/selftests/bpf/prog_tests/core_reloc.c
449
t = btf__type_by_id(local_btf, i);
tools/testing/selftests/bpf/prog_tests/core_reloc.c
451
if (t->name_off)
tools/testing/selftests/bpf/prog_tests/core_reloc.c
454
if (btf_is_struct(t) && btf_vlen(t) &&
tools/testing/selftests/bpf/prog_tests/core_reloc.c
455
(name = btf__name_by_offset(local_btf, btf_members(t)[0].name_off)) &&
tools/testing/selftests/bpf/prog_tests/core_reloc.c
458
} else if (btf_is_union(t) && btf_vlen(t) &&
tools/testing/selftests/bpf/prog_tests/core_reloc.c
459
(name = btf__name_by_offset(local_btf, btf_members(t)[0].name_off)) &&
tools/testing/selftests/bpf/prog_tests/core_reloc.c
462
} else if (btf_is_enum(t) && btf_vlen(t) &&
tools/testing/selftests/bpf/prog_tests/core_reloc.c
463
(name = btf__name_by_offset(local_btf, btf_enum(t)[0].name_off)) &&
tools/testing/selftests/bpf/prog_tests/core_reloc.c
466
} else if (btf_is_ptr(t) && (t = btf__type_by_id(local_btf, t->type))) {
tools/testing/selftests/bpf/prog_tests/core_reloc.c
467
if (btf_is_func_proto(t) && (t = btf__type_by_id(local_btf, t->type)) &&
tools/testing/selftests/bpf/prog_tests/core_reloc.c
468
btf_is_int(t) && (name = btf__name_by_offset(local_btf, t->name_off)) &&
tools/testing/selftests/bpf/prog_tests/core_reloc.c
472
} else if (btf_is_void(t)) {
tools/testing/selftests/bpf/prog_tests/core_reloc.c
476
} else if (btf_is_array(t) && (t = btf__type_by_id(local_btf, btf_array(t)->type)) &&
tools/testing/selftests/bpf/prog_tests/core_reloc.c
477
btf_is_int(t) && (name = btf__name_by_offset(local_btf, t->name_off)) &&
tools/testing/selftests/bpf/prog_tests/libbpf_probes.c
10
const struct btf_type *t;
tools/testing/selftests/bpf/prog_tests/libbpf_probes.c
22
t = btf__type_by_id(btf, id);
tools/testing/selftests/bpf/prog_tests/libbpf_probes.c
23
if (!ASSERT_OK_PTR(t, "bpf_prog_type_enum"))
tools/testing/selftests/bpf/prog_tests/libbpf_probes.c
26
for (e = btf_enum(t), i = 0, n = btf_vlen(t); i < n; e++, i++) {
tools/testing/selftests/bpf/prog_tests/libbpf_probes.c
50
const struct btf_type *t;
tools/testing/selftests/bpf/prog_tests/libbpf_probes.c
62
t = btf__type_by_id(btf, id);
tools/testing/selftests/bpf/prog_tests/libbpf_probes.c
63
if (!ASSERT_OK_PTR(t, "bpf_map_type_enum"))
tools/testing/selftests/bpf/prog_tests/libbpf_probes.c
66
for (e = btf_enum(t), i = 0, n = btf_vlen(t); i < n; e++, i++) {
tools/testing/selftests/bpf/prog_tests/libbpf_str.c
114
const struct btf_type *t;
tools/testing/selftests/bpf/prog_tests/libbpf_str.c
126
t = btf__type_by_id(btf, id);
tools/testing/selftests/bpf/prog_tests/libbpf_str.c
127
e = btf_enum(t);
tools/testing/selftests/bpf/prog_tests/libbpf_str.c
128
n = btf_vlen(t);
tools/testing/selftests/bpf/prog_tests/libbpf_str.c
171
const struct btf_type *t;
tools/testing/selftests/bpf/prog_tests/libbpf_str.c
183
t = btf__type_by_id(btf, id);
tools/testing/selftests/bpf/prog_tests/libbpf_str.c
184
e = btf_enum(t);
tools/testing/selftests/bpf/prog_tests/libbpf_str.c
185
n = btf_vlen(t);
tools/testing/selftests/bpf/prog_tests/libbpf_str.c
24
const struct btf_type *t;
tools/testing/selftests/bpf/prog_tests/libbpf_str.c
36
t = btf__type_by_id(btf, id);
tools/testing/selftests/bpf/prog_tests/libbpf_str.c
37
e = btf_enum(t);
tools/testing/selftests/bpf/prog_tests/libbpf_str.c
38
n = btf_vlen(t);
tools/testing/selftests/bpf/prog_tests/libbpf_str.c
69
const struct btf_type *t;
tools/testing/selftests/bpf/prog_tests/libbpf_str.c
81
t = btf__type_by_id(btf, id);
tools/testing/selftests/bpf/prog_tests/libbpf_str.c
82
e = btf_enum(t);
tools/testing/selftests/bpf/prog_tests/libbpf_str.c
83
n = btf_vlen(t);
tools/testing/selftests/bpf/prog_tests/rdonly_maps.c
45
const struct rdonly_map_subtest *t = &subtests[i];
tools/testing/selftests/bpf/prog_tests/rdonly_maps.c
47
if (!test__start_subtest(t->subtest_name))
tools/testing/selftests/bpf/prog_tests/rdonly_maps.c
50
prog = bpf_object__find_program_by_name(obj, t->prog_name);
tools/testing/selftests/bpf/prog_tests/rdonly_maps.c
52
t->prog_name))
tools/testing/selftests/bpf/prog_tests/rdonly_maps.c
74
"prog '%s' didn't run?\n", t->prog_name))
tools/testing/selftests/bpf/prog_tests/rdonly_maps.c
76
if (CHECK(bss.iters != t->exp_iters, "check_iters",
tools/testing/selftests/bpf/prog_tests/rdonly_maps.c
78
t->prog_name, bss.iters, t->exp_iters))
tools/testing/selftests/bpf/prog_tests/rdonly_maps.c
80
if (CHECK(bss.sum != t->exp_sum, "check_sum",
tools/testing/selftests/bpf/prog_tests/rdonly_maps.c
82
t->prog_name, bss.sum, t->exp_sum))
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
100
static const char *t_str(enum num_t t)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
102
switch (t) {
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
111
static enum num_t t_is_32(enum num_t t)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1119
enum num_t t;
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
113
switch (t) {
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1130
for (t = first_t; t <= last_t; t++) {
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1131
reg->r[t] = range(t, sval, sval);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
122
static enum num_t t_signed(enum num_t t)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1236
static bool assert_range_eq(enum num_t t, struct range x, struct range y,
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
124
switch (t) {
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1245
snprintf_range(t, sb, x);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1247
snprintf_range(t, sb, y);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1258
enum num_t t;
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1270
for (t = first_t; t <= last_t; t++) {
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1271
if (!assert_range_eq(t, r->r[t], e->r[t], ctx, t_str(t)))
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1313
enum num_t t;
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1317
for (t = first_t; t <= last_t; t++) {
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1321
struct range z = t_is_32(init_t) ? unkn_subreg(t) : unkn[t];
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1323
fr1->r[t] = fr2->r[t] = tr1->r[t] = tr2->r[t] = z;
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
133
static enum num_t t_unsigned(enum num_t t)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
135
switch (t) {
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1426
static void subtest_case_str(struct strbuf *sb, struct subtest_case *t, bool use_op)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1428
snappendf(sb, "(%s)", t_str(t->init_t));
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1429
snprintf_range(t->init_t, sb, t->x);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1430
snappendf(sb, " (%s)%s ", t_str(t->cond_t), use_op ? op_str(t->op) : "<op>");
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1431
snprintf_range(t->init_t, sb, t->y);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
148
static bool num_is_small(enum num_t t, u64 x)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
150
switch (t) {
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
159
static void snprintf_num(enum num_t t, struct strbuf *sb, u64 x)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
161
bool is_small = num_is_small(t, x);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
164
switch (t) {
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
172
switch (t) {
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
2002
static u64 rand_const(enum num_t t)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
2004
return cast_t(t, rand_u64());
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
2007
static struct range rand_range(enum num_t t)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
2009
u64 x = rand_const(t), y = rand_const(t);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
2011
return range(t, min_t(t, x, y), max_t(t, x, y));
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
2019
u64 t;
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
2046
t = rand_const(init_t);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
2047
range2 = range(init_t, t, t);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
222
static void snprintf_range(enum num_t t, struct strbuf *sb, struct range x)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
225
return snprintf_num(t, sb, x.a);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
228
snprintf_num(t, sb, x.a);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
230
snprintf_num(t, sb, x.b);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
234
static void print_range(enum num_t t, struct range x, const char *sfx)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
238
snprintf_range(t, sb, x);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
249
static struct range unkn_subreg(enum num_t t)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
251
switch (t) {
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
260
static struct range range(enum num_t t, u64 a, u64 b)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
262
switch (t) {
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
400
static bool is_valid_num(enum num_t t, u64 x)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
402
switch (t) {
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
411
static bool is_valid_range(enum num_t t, struct range x)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
413
if (!is_valid_num(t, x.a) || !is_valid_num(t, x.b))
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
416
switch (t) {
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
425
static struct range range_intersection(enum num_t t, struct range old, struct range new)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
427
return range(t, max_t(t, old.a, new.a), min_t(t, old.b, new.b));
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
434
static struct range range_union(enum num_t t, struct range x, struct range y)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
436
if (!is_valid_range(t, x))
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
438
if (!is_valid_range(t, y))
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
440
return range(t, min_t(t, x.a, y.a), max_t(t, x.b, y.b));
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
563
static bool range_canbe_op(enum num_t t, struct range x, struct range y, enum op op)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
571
case OP_EQ: return (T)max_t(t, x.a, y.a) <= (T)min_t(t, x.b, y.b); \
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
577
switch (t) {
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
592
static bool range_always_op(enum num_t t, struct range x, struct range y, enum op op)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
595
return !range_canbe_op(t, x, y, complement_op(op));
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
603
static bool range_never_op(enum num_t t, struct range x, struct range y, enum op op)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
605
return !range_canbe_op(t, x, y, op);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
613
static int range_branch_taken_op(enum num_t t, struct range x, struct range y, enum op op)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
615
if (range_always_op(t, x, y, op))
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
617
if (range_never_op(t, x, y, op))
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
629
static void range_cond(enum num_t t, struct range x, struct range y,
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
632
if (!range_canbe_op(t, x, y, op)) {
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
640
*newx = range(t, x.a, min_t(t, x.b, y.b - 1));
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
641
*newy = range(t, max_t(t, x.a + 1, y.a), y.b);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
644
*newx = range(t, x.a, min_t(t, x.b, y.b));
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
645
*newy = range(t, max_t(t, x.a, y.a), y.b);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
648
*newx = range(t, max_t(t, x.a, y.a + 1), x.b);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
649
*newy = range(t, y.a, min_t(t, x.b - 1, y.b));
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
652
*newx = range(t, max_t(t, x.a, y.a), x.b);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
653
*newy = range(t, y.a, min_t(t, x.b, y.b));
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
656
*newx = range(t, max_t(t, x.a, y.a), min_t(t, x.b, y.b));
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
657
*newy = range(t, max_t(t, x.a, y.a), min_t(t, x.b, y.b));
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
663
*newx = range(t, x.a, x.b);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
664
*newy = range(t, y.a + 1, y.b);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
667
*newx = range(t, x.a, x.b);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
668
*newy = range(t, y.a, y.b - 1);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
67
static __always_inline u64 min_t(enum num_t t, u64 x, u64 y)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
671
*newx = range(t, x.a + 1, x.b);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
672
*newy = range(t, y.a, y.b);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
675
*newx = range(t, x.a, x.b - 1);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
676
*newy = range(t, y.a, y.b);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
679
*newx = range(t, x.a, x.b);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
680
*newy = range(t, y.a, y.b);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
69
switch (t) {
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
701
enum num_t t;
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
710
for (t = first_t; t <= last_t; t++) {
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
711
snappendf(sb, "%s%s=", cnt++ ? "," : "", t_str(t));
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
712
snprintf_range(t, sb, r->r[t]);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
731
static void reg_state_refine(struct reg_state *r, enum num_t t, struct range x, const char *ctx)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
741
r->r[d_t] = range_refine(d_t, r->r[d_t], t, x);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
745
print_refinement(t, x, d_t, old, r->r[d_t], ctx);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
769
static void reg_state_set_const(struct reg_state *rs, enum num_t t, u64 val)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
775
rs->r[tt] = tt == t ? range(t, val, val) : unkn[tt];
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
777
reg_state_refine(rs, t, rs->r[t], "CONST");
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
78
static __always_inline u64 max_t(enum num_t t, u64 x, u64 y)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
780
static void reg_state_cond(enum num_t t, struct reg_state *x, struct reg_state *y, enum op op,
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
793
ts[0] = t_unsigned(t);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
794
ts[1] = t_signed(t);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
797
ts[0] = t;
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
80
switch (t) {
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
802
t = ts[i];
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
803
z1 = x->r[t];
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
804
z2 = y->r[t];
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
806
range_cond(t, z1, z2, op, &z1, &z2);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
810
reg_state_refine(&xx, t, z1, buf);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
814
reg_state_refine(&yy, t, z2, buf);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
824
static int reg_state_branch_taken_op(enum num_t t, struct reg_state *x, struct reg_state *y,
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
829
enum num_t tu = t_unsigned(t);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
830
enum num_t ts = t_signed(t);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
842
if (br_u == -1 && (t == U64 || t == S64)) {
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
864
return range_branch_taken_op(t, x->r[t], y->r[t], op);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
89
static __always_inline u64 cast_t(enum num_t t, u64 x)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
91
switch (t) {
tools/testing/selftests/bpf/prog_tests/select_reuseport.c
789
const struct test *t;
tools/testing/selftests/bpf/prog_tests/select_reuseport.c
791
for (t = tests; t < tests + ARRAY_SIZE(tests); t++) {
tools/testing/selftests/bpf/prog_tests/select_reuseport.c
792
if (t->need_sotype && t->need_sotype != sotype)
tools/testing/selftests/bpf/prog_tests/select_reuseport.c
798
inany ? "INANY" : "LOOPBACK", t->name);
tools/testing/selftests/bpf/prog_tests/select_reuseport.c
812
setup_per_test(sotype, family, inany, t->no_inner_map);
tools/testing/selftests/bpf/prog_tests/select_reuseport.c
813
t->fn(sotype, family);
tools/testing/selftests/bpf/prog_tests/select_reuseport.c
814
cleanup_per_test(t->no_inner_map);
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
1017
const struct test *t;
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
1019
for (t = tests; t < tests + ARRAY_SIZE(tests); t++) {
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
1020
if (test__start_subtest(t->desc))
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
1021
drop_on_reuseport(t);
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
1161
static void run_multi_prog_lookup(const struct test_multi_prog *t)
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
1169
map_fd = bpf_map__fd(t->run_map);
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
1181
link1 = attach_lookup_prog(t->prog1);
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
1184
link2 = attach_lookup_prog(t->prog2);
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
1188
server_fd = make_server(SOCK_STREAM, t->listen_at.ip,
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
1189
t->listen_at.port, NULL);
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
1193
err = update_lookup_map(t->redir_map, SERVER_A, server_fd);
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
1205
if (CHECK(err && !t->expect_errno, "connect",
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
1208
if (CHECK(err && t->expect_errno && errno != t->expect_errno,
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
1295
struct test_multi_prog *t;
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
1297
for (t = tests; t < tests + ARRAY_SIZE(tests); t++) {
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
1298
t->redir_map = skel->maps.redir_map;
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
1299
t->run_map = skel->maps.run_map;
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
1300
if (test__start_subtest(t->desc))
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
1301
run_multi_prog_lookup(t);
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
538
static void run_lookup_prog(const struct test *t)
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
545
lookup_link = attach_lookup_prog(t->lookup_prog);
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
550
server_fds[i] = make_server(t->sotype, t->listen_at.ip,
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
551
t->listen_at.port,
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
552
t->reuseport_prog);
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
556
err = update_lookup_map(t->sock_map, i, server_fds[i]);
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
561
if (!t->reuseport_prog)
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
571
if (t->reuseport_has_conns) {
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
573
reuse_conn_fd = make_server(t->sotype, t->listen_at.ip,
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
574
t->listen_at.port,
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
575
t->reuseport_prog);
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
585
client_fd = connect_to_addr_str(is_ipv6(t->connect_to.ip) ? AF_INET6 : AF_INET,
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
586
t->sotype, t->connect_to.ip, t->connect_to.port, NULL);
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
590
if (t->sotype == SOCK_STREAM)
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
591
tcp_echo_test(client_fd, server_fds[t->accept_on]);
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
593
udp_echo_test(client_fd, server_fds[t->accept_on]);
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
792
const struct test *t;
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
794
for (t = tests; t < tests + ARRAY_SIZE(tests); t++) {
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
795
if (test__start_subtest(t->desc))
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
796
run_lookup_prog(t);
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
800
static void drop_on_lookup(const struct test *t)
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
802
int family = is_ipv6(t->connect_to.ip) ? AF_INET6 : AF_INET;
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
809
lookup_link = attach_lookup_prog(t->lookup_prog);
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
813
server_fd = make_server(t->sotype, t->listen_at.ip, t->listen_at.port,
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
814
t->reuseport_prog);
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
818
client_fd = client_socket(family, t->sotype, NULL);
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
822
err = make_sockaddr(family, t->connect_to.ip, t->connect_to.port, &dst, &len);
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
826
if (t->sotype == SOCK_DGRAM) {
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
910
const struct test *t;
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
912
for (t = tests; t < tests + ARRAY_SIZE(tests); t++) {
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
913
if (test__start_subtest(t->desc))
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
914
drop_on_lookup(t);
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
918
static void drop_on_reuseport(const struct test *t)
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
920
int family = is_ipv6(t->connect_to.ip) ? AF_INET6 : AF_INET;
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
927
lookup_link = attach_lookup_prog(t->lookup_prog);
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
931
server1 = make_server(t->sotype, t->listen_at.ip, t->listen_at.port,
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
932
t->reuseport_prog);
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
936
err = update_lookup_map(t->sock_map, SERVER_A, server1);
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
941
server2 = make_server(t->sotype, t->connect_to.ip, t->connect_to.port,
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
946
client = client_socket(family, t->sotype, NULL);
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
950
err = make_sockaddr(family, t->connect_to.ip, t->connect_to.port, &dst, &len);
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
954
if (t->sotype == SOCK_DGRAM) {
tools/testing/selftests/bpf/prog_tests/sockmap_listen.c
1313
const struct op_test *t;
tools/testing/selftests/bpf/prog_tests/sockmap_listen.c
1322
for (t = tests; t < tests + ARRAY_SIZE(tests); t++) {
tools/testing/selftests/bpf/prog_tests/sockmap_listen.c
1324
sotype_name, t->name);
tools/testing/selftests/bpf/prog_tests/sockmap_listen.c
1326
if (t->sotype != 0 && t->sotype != sotype)
tools/testing/selftests/bpf/prog_tests/sockmap_listen.c
1332
t->fn(skel, family, sotype, map_fd);
tools/testing/selftests/bpf/prog_tests/sockmap_listen.c
1354
const struct redir_test *t;
tools/testing/selftests/bpf/prog_tests/sockmap_listen.c
1360
for (t = tests; t < tests + ARRAY_SIZE(tests); t++) {
tools/testing/selftests/bpf/prog_tests/sockmap_listen.c
1362
t->name);
tools/testing/selftests/bpf/prog_tests/sockmap_listen.c
1367
t->fn(skel, map, family, sotype);
tools/testing/selftests/bpf/prog_tests/sockmap_listen.c
1386
const struct reuseport_test *t;
tools/testing/selftests/bpf/prog_tests/sockmap_listen.c
1397
for (t = tests; t < tests + ARRAY_SIZE(tests); t++) {
tools/testing/selftests/bpf/prog_tests/sockmap_listen.c
1399
sotype_name, t->name);
tools/testing/selftests/bpf/prog_tests/sockmap_listen.c
1401
if (t->sotype != 0 && t->sotype != sotype)
tools/testing/selftests/bpf/prog_tests/sockmap_listen.c
1407
t->fn(family, sotype, socket_map, verdict_map, reuseport_prog);
tools/testing/selftests/bpf/prog_tests/sockmap_listen.c
318
const struct test *t;
tools/testing/selftests/bpf/prog_tests/sockmap_listen.c
320
for (t = tests; t < tests + ARRAY_SIZE(tests); t++) {
tools/testing/selftests/bpf/prog_tests/sockmap_listen.c
321
if (t->progfd != -1 &&
tools/testing/selftests/bpf/prog_tests/sockmap_listen.c
322
xbpf_prog_attach(t->progfd, mapfd, t->atype, 0) != 0)
tools/testing/selftests/bpf/prog_tests/sockmap_listen.c
327
if (t->progfd != -1)
tools/testing/selftests/bpf/prog_tests/sockmap_listen.c
328
xbpf_prog_detach2(t->progfd, mapfd, t->atype);
tools/testing/selftests/bpf/prog_tests/sockmap_listen.c
544
pthread_t t;
tools/testing/selftests/bpf/prog_tests/sockmap_listen.c
560
err = xpthread_create(&t, NULL, connect_accept_thread, &ctx);
tools/testing/selftests/bpf/prog_tests/sockmap_listen.c
575
xpthread_join(t, NULL);
tools/testing/selftests/bpf/prog_tests/sockmap_listen.c
607
pthread_t t;
tools/testing/selftests/bpf/prog_tests/sockmap_listen.c
622
err = pthread_create(&t, NULL, listen_thread, &ctx);
tools/testing/selftests/bpf/prog_tests/sockmap_listen.c
643
xpthread_join(t, NULL);
tools/testing/selftests/bpf/prog_tests/summarization.c
131
for (int t = 0; t < 2; t++) {
tools/testing/selftests/bpf/prog_tests/summarization.c
135
mains[t][i].to_be_replaced, replacements[t][j].func);
tools/testing/selftests/bpf/prog_tests/summarization.c
138
test_aux(mains[t][i].main, mains[t][i].to_be_replaced, replacements[t][j].func,
tools/testing/selftests/bpf/prog_tests/summarization.c
139
mains[t][i].has_side_effect || !replacements[t][j].has_side_effect,
tools/testing/selftests/bpf/prog_tests/summarization.c
140
replacements[t][j].err_msg);
tools/testing/selftests/bpf/prog_tests/tc_links.c
1794
struct tcmsg t;
tools/testing/selftests/bpf/prog_tests/tc_links.c
1812
req.t.tcm_family = AF_UNSPEC;
tools/testing/selftests/bpf/prog_tests/tc_links.c
1813
req.t.tcm_ifindex = ifindex;
tools/testing/selftests/bpf/prog_tests/tc_links.c
1814
req.t.tcm_parent = 0xfffffff1;
tools/testing/selftests/bpf/prog_tests/tc_redirect.c
831
int i, t = TCP_IP6_CLEAR_DTIME;
tools/testing/selftests/bpf/prog_tests/tc_redirect.c
832
__u32 *dtimes = skel->bss->dtimes[t];
tools/testing/selftests/bpf/prog_tests/tc_redirect.c
833
__u32 *errs = skel->bss->errs[t];
tools/testing/selftests/bpf/prog_tests/tc_redirect.c
835
skel->bss->test = t;
tools/testing/selftests/bpf/prog_tests/tc_redirect.c
836
test_inet_dtime(AF_INET6, SOCK_STREAM, IP6_DST, 50000 + t);
tools/testing/selftests/bpf/prog_tests/tc_redirect.c
839
dtime_cnt_str(t, INGRESS_FWDNS_P100));
tools/testing/selftests/bpf/prog_tests/tc_redirect.c
841
dtime_cnt_str(t, INGRESS_FWDNS_P101));
tools/testing/selftests/bpf/prog_tests/tc_redirect.c
843
dtime_cnt_str(t, EGRESS_FWDNS_P100));
tools/testing/selftests/bpf/prog_tests/tc_redirect.c
845
dtime_cnt_str(t, EGRESS_FWDNS_P101));
tools/testing/selftests/bpf/prog_tests/tc_redirect.c
847
dtime_cnt_str(t, EGRESS_ENDHOST));
tools/testing/selftests/bpf/prog_tests/tc_redirect.c
849
dtime_cnt_str(t, INGRESS_ENDHOST));
tools/testing/selftests/bpf/prog_tests/tc_redirect.c
852
ASSERT_EQ(errs[i], 0, dtime_err_str(t, i));
tools/testing/selftests/bpf/prog_tests/tc_redirect.c
859
int i, t;
tools/testing/selftests/bpf/prog_tests/tc_redirect.c
862
t = bpf_fwd ? TCP_IP4 : TCP_IP4_RT_FWD;
tools/testing/selftests/bpf/prog_tests/tc_redirect.c
865
t = bpf_fwd ? TCP_IP6 : TCP_IP6_RT_FWD;
tools/testing/selftests/bpf/prog_tests/tc_redirect.c
869
dtimes = skel->bss->dtimes[t];
tools/testing/selftests/bpf/prog_tests/tc_redirect.c
870
errs = skel->bss->errs[t];
tools/testing/selftests/bpf/prog_tests/tc_redirect.c
872
skel->bss->test = t;
tools/testing/selftests/bpf/prog_tests/tc_redirect.c
873
test_inet_dtime(family, SOCK_STREAM, addr, 50000 + t);
tools/testing/selftests/bpf/prog_tests/tc_redirect.c
879
dtime_cnt_str(t, INGRESS_FWDNS_P100));
tools/testing/selftests/bpf/prog_tests/tc_redirect.c
881
ASSERT_GT(dtimes[i], 0, dtime_cnt_str(t, i));
tools/testing/selftests/bpf/prog_tests/tc_redirect.c
884
ASSERT_EQ(errs[i], 0, dtime_err_str(t, i));
tools/testing/selftests/bpf/prog_tests/tc_redirect.c
891
int i, t;
tools/testing/selftests/bpf/prog_tests/tc_redirect.c
894
t = bpf_fwd ? UDP_IP4 : UDP_IP4_RT_FWD;
tools/testing/selftests/bpf/prog_tests/tc_redirect.c
897
t = bpf_fwd ? UDP_IP6 : UDP_IP6_RT_FWD;
tools/testing/selftests/bpf/prog_tests/tc_redirect.c
901
dtimes = skel->bss->dtimes[t];
tools/testing/selftests/bpf/prog_tests/tc_redirect.c
902
errs = skel->bss->errs[t];
tools/testing/selftests/bpf/prog_tests/tc_redirect.c
904
skel->bss->test = t;
tools/testing/selftests/bpf/prog_tests/tc_redirect.c
905
test_inet_dtime(family, SOCK_DGRAM, addr, 50000 + t);
tools/testing/selftests/bpf/prog_tests/tc_redirect.c
908
dtime_cnt_str(t, INGRESS_FWDNS_P100));
tools/testing/selftests/bpf/prog_tests/tc_redirect.c
910
ASSERT_GT(dtimes[i], 0, dtime_cnt_str(t, i));
tools/testing/selftests/bpf/prog_tests/tc_redirect.c
913
ASSERT_EQ(errs[i], 0, dtime_err_str(t, i));
tools/testing/selftests/bpf/prog_tests/test_global_funcs.c
28
const struct btf_type *t;
tools/testing/selftests/bpf/prog_tests/test_global_funcs.c
31
t = btf__type_by_id(btf, p->type);
tools/testing/selftests/bpf/prog_tests/test_global_funcs.c
32
if (!ASSERT_EQ(btf_kind(t), BTF_KIND_PTR, "ptr_t"))
tools/testing/selftests/bpf/prog_tests/test_global_funcs.c
35
s = btf_type_raw_dump(btf, t->type);
tools/testing/selftests/bpf/prog_tests/uretprobe_stack.c
115
const struct range *t = va_arg(args, const struct range *);
tools/testing/selftests/bpf/prog_tests/uretprobe_stack.c
117
ASSERT_GE(ips[i], t->start, "addr_start");
tools/testing/selftests/bpf/prog_tests/uretprobe_stack.c
118
ASSERT_LT(ips[i], t->stop, "addr_stop");
tools/testing/selftests/bpf/prog_tests/verifier_log.c
273
struct btf_type *t;
tools/testing/selftests/bpf/prog_tests/verifier_log.c
291
t = (void *)btf__type_by_id(btf, res);
tools/testing/selftests/bpf/prog_tests/verifier_log.c
292
if (!ASSERT_OK_PTR(t, "int_btf_type"))
tools/testing/selftests/bpf/prog_tests/verifier_log.c
294
t->size = 3;
tools/testing/selftests/bpf/progs/bpf_cubic.c
268
__u64 offs, t;
tools/testing/selftests/bpf/progs/bpf_cubic.c
318
t = (__s32)(tcp_jiffies32 - ca->epoch_start) * USEC_PER_JIFFY;
tools/testing/selftests/bpf/progs/bpf_cubic.c
319
t += ca->delay_min;
tools/testing/selftests/bpf/progs/bpf_cubic.c
321
t <<= BICTCP_HZ;
tools/testing/selftests/bpf/progs/bpf_cubic.c
322
t /= USEC_PER_SEC;
tools/testing/selftests/bpf/progs/bpf_cubic.c
324
if (t < ca->bic_K) /* t - K */
tools/testing/selftests/bpf/progs/bpf_cubic.c
325
offs = ca->bic_K - t;
tools/testing/selftests/bpf/progs/bpf_cubic.c
327
offs = t - ca->bic_K;
tools/testing/selftests/bpf/progs/bpf_cubic.c
331
if (t < ca->bic_K) /* below origin*/
tools/testing/selftests/bpf/progs/btf_dump_test_case_syntax.c
244
int t[11];
tools/testing/selftests/bpf/progs/dmabuf_iter.c
94
bool t = true;
tools/testing/selftests/bpf/progs/dmabuf_iter.c
96
bpf_map_update_elem(&testbuf_hash, name, &t, BPF_EXIST);
tools/testing/selftests/bpf/progs/htab_update.c
11
struct bpf_timer t;
tools/testing/selftests/bpf/progs/iters.c
662
int *t, i;
tools/testing/selftests/bpf/progs/iters.c
664
while ((t = bpf_iter_num_next(it))) {
tools/testing/selftests/bpf/progs/iters.c
665
i = *t;
tools/testing/selftests/bpf/progs/iters.c
674
int *t, i, sum = 0;
tools/testing/selftests/bpf/progs/iters.c
676
while ((t = bpf_iter_num_next(it))) {
tools/testing/selftests/bpf/progs/iters.c
677
i = *t;
tools/testing/selftests/bpf/progs/setget_sockopt.c
100
opt = t->opt;
tools/testing/selftests/bpf/progs/setget_sockopt.c
125
const struct sockopt_test *t,
tools/testing/selftests/bpf/progs/setget_sockopt.c
130
opt = t->opt;
tools/testing/selftests/bpf/progs/setget_sockopt.c
131
new = t->new;
tools/testing/selftests/bpf/progs/setget_sockopt.c
132
if (sk->sk_type == SOCK_STREAM && t->tcp_expected)
tools/testing/selftests/bpf/progs/setget_sockopt.c
133
expected = t->tcp_expected;
tools/testing/selftests/bpf/progs/setget_sockopt.c
135
expected = t->expected;
tools/testing/selftests/bpf/progs/setget_sockopt.c
147
if (t->restore)
tools/testing/selftests/bpf/progs/setget_sockopt.c
148
old = t->restore;
tools/testing/selftests/bpf/progs/setget_sockopt.c
157
const struct sockopt_test *t;
tools/testing/selftests/bpf/progs/setget_sockopt.c
162
t = &sol_socket_tests[i];
tools/testing/selftests/bpf/progs/setget_sockopt.c
163
if (!t->opt)
tools/testing/selftests/bpf/progs/setget_sockopt.c
166
if (t->flip)
tools/testing/selftests/bpf/progs/setget_sockopt.c
167
return bpf_test_sockopt_flip(lc->ctx, lc->sk, t, SOL_SOCKET);
tools/testing/selftests/bpf/progs/setget_sockopt.c
169
return bpf_test_sockopt_int(lc->ctx, lc->sk, t, SOL_SOCKET);
tools/testing/selftests/bpf/progs/setget_sockopt.c
174
const struct sockopt_test *t;
tools/testing/selftests/bpf/progs/setget_sockopt.c
179
t = &sol_ip_tests[i];
tools/testing/selftests/bpf/progs/setget_sockopt.c
180
if (!t->opt)
tools/testing/selftests/bpf/progs/setget_sockopt.c
183
if (t->flip)
tools/testing/selftests/bpf/progs/setget_sockopt.c
184
return bpf_test_sockopt_flip(lc->ctx, lc->sk, t, IPPROTO_IP);
tools/testing/selftests/bpf/progs/setget_sockopt.c
186
return bpf_test_sockopt_int(lc->ctx, lc->sk, t, IPPROTO_IP);
tools/testing/selftests/bpf/progs/setget_sockopt.c
191
const struct sockopt_test *t;
tools/testing/selftests/bpf/progs/setget_sockopt.c
196
t = &sol_ipv6_tests[i];
tools/testing/selftests/bpf/progs/setget_sockopt.c
197
if (!t->opt)
tools/testing/selftests/bpf/progs/setget_sockopt.c
200
if (t->flip)
tools/testing/selftests/bpf/progs/setget_sockopt.c
201
return bpf_test_sockopt_flip(lc->ctx, lc->sk, t, IPPROTO_IPV6);
tools/testing/selftests/bpf/progs/setget_sockopt.c
203
return bpf_test_sockopt_int(lc->ctx, lc->sk, t, IPPROTO_IPV6);
tools/testing/selftests/bpf/progs/setget_sockopt.c
208
const struct sockopt_test *t;
tools/testing/selftests/bpf/progs/setget_sockopt.c
215
t = &sol_tcp_tests[i];
tools/testing/selftests/bpf/progs/setget_sockopt.c
216
if (!t->opt)
tools/testing/selftests/bpf/progs/setget_sockopt.c
222
if (t->opt == TCP_CONGESTION) {
tools/testing/selftests/bpf/progs/setget_sockopt.c
251
if (t->flip)
tools/testing/selftests/bpf/progs/setget_sockopt.c
252
return bpf_test_sockopt_flip(ctx, sk, t, IPPROTO_TCP);
tools/testing/selftests/bpf/progs/setget_sockopt.c
254
return bpf_test_sockopt_int(ctx, sk, t, IPPROTO_TCP);
tools/testing/selftests/bpf/progs/setget_sockopt.c
95
const struct sockopt_test *t,
tools/testing/selftests/bpf/progs/setget_sockopt.c
98
int old, tmp, new, opt = t->opt;
tools/testing/selftests/bpf/progs/struct_ops_kptr_return_fail__local_kptr.c
20
struct task_struct *t;
tools/testing/selftests/bpf/progs/struct_ops_kptr_return_fail__local_kptr.c
24
t = bpf_obj_new(typeof(*task));
tools/testing/selftests/bpf/progs/struct_ops_kptr_return_fail__local_kptr.c
25
if (!t)
tools/testing/selftests/bpf/progs/struct_ops_kptr_return_fail__local_kptr.c
28
return t;
tools/testing/selftests/bpf/progs/test_helper_restricted.c
40
bpf_timer_init(&timer->t, &timers, CLOCK_MONOTONIC);
tools/testing/selftests/bpf/progs/test_helper_restricted.c
41
bpf_timer_set_callback(&timer->t, timer_cb);
tools/testing/selftests/bpf/progs/test_helper_restricted.c
42
bpf_timer_start(&timer->t, 10E9, 0);
tools/testing/selftests/bpf/progs/test_helper_restricted.c
43
bpf_timer_cancel(&timer->t);
tools/testing/selftests/bpf/progs/test_helper_restricted.c
7
struct bpf_timer t;
tools/testing/selftests/bpf/progs/test_log_fixup.c
17
static struct task_struct___bad *t;
tools/testing/selftests/bpf/progs/test_log_fixup.c
19
return bpf_core_field_size(t->fake_field);
tools/testing/selftests/bpf/progs/test_log_fixup.c
24
static struct task_struct___bad *t;
tools/testing/selftests/bpf/progs/test_log_fixup.c
27
return (void *)&t->fake_field_subprog - (void *)t;
tools/testing/selftests/bpf/progs/test_log_fixup.c
33
static struct task_struct___bad *t;
tools/testing/selftests/bpf/progs/test_log_fixup.c
35
return bad_subprog() + bpf_core_field_size(t->pid);
tools/testing/selftests/bpf/progs/test_subprogs.c
101
struct task_struct *t = (void *)bpf_get_current_task();
tools/testing/selftests/bpf/progs/test_subprogs.c
103
if (!BPF_CORE_READ(t, pid) || !get_task_tgid((uintptr_t)t))
tools/testing/selftests/bpf/progs/test_subprogs.c
117
struct task_struct *t = (void *)bpf_get_current_task();
tools/testing/selftests/bpf/progs/test_subprogs.c
119
if (!BPF_CORE_READ(t, pid) || !get_task_tgid((uintptr_t)t))
tools/testing/selftests/bpf/progs/test_subprogs.c
54
__noinline int get_task_tgid(uintptr_t t)
tools/testing/selftests/bpf/progs/test_subprogs.c
57
return BPF_CORE_READ((struct task_struct *)(void *)t, tgid);
tools/testing/selftests/bpf/progs/test_subprogs.c
71
struct task_struct *t = (void *)bpf_get_current_task();
tools/testing/selftests/bpf/progs/test_subprogs.c
73
if (!BPF_CORE_READ(t, pid) || !get_task_tgid((uintptr_t)t))
tools/testing/selftests/bpf/progs/test_subprogs.c
83
struct task_struct *t = (void *)bpf_get_current_task();
tools/testing/selftests/bpf/progs/test_subprogs.c
85
if (!BPF_CORE_READ(t, pid) || !get_task_tgid((uintptr_t)t))
tools/testing/selftests/bpf/progs/timer.c
37
struct bpf_timer t;
tools/testing/selftests/bpf/progs/timer_failure.c
14
struct bpf_timer t;
tools/testing/selftests/bpf/progs/timer_interrupt.c
16
struct bpf_timer t;
tools/testing/selftests/bpf/progs/timer_lockup.c
13
struct bpf_timer t;
tools/testing/selftests/bpf/progs/verifier_async_cb_context.c
15
struct bpf_timer t;
tools/testing/selftests/bpf/progs/verifier_async_cb_context.c
44
bpf_timer_init(&val->t, &timer_map, 0);
tools/testing/selftests/bpf/progs/verifier_async_cb_context.c
45
bpf_timer_set_callback(&val->t, timer_cb);
tools/testing/selftests/bpf/progs/verifier_async_cb_context.c
60
bpf_timer_init(&val->t, &timer_map, 0);
tools/testing/selftests/bpf/progs/verifier_async_cb_context.c
61
bpf_timer_set_callback(&val->t, timer_cb);
tools/testing/selftests/bpf/progs/verifier_global_ptr_args.c
100
struct task_struct *t = bpf_get_current_task_btf();
tools/testing/selftests/bpf/progs/verifier_global_ptr_args.c
102
return subprog_trusted_task_nonnull(t);
tools/testing/selftests/bpf/progs/verifier_global_ptr_args.c
124
struct task_struct___local *t = (void *)bpf_get_current_task_btf();
tools/testing/selftests/bpf/progs/verifier_global_ptr_args.c
126
return subprog_nullable_task_flavor(t);
tools/testing/selftests/bpf/progs/verifier_global_ptr_args.c
142
struct task_struct *t = bpf_get_current_task_btf();
tools/testing/selftests/bpf/progs/verifier_global_ptr_args.c
144
return subprog_nonnull_task_flavor((void *)t);
tools/testing/selftests/bpf/progs/verifier_global_ptr_args.c
79
struct task_struct *t = bpf_get_current_task_btf();
tools/testing/selftests/bpf/progs/verifier_global_ptr_args.c
83
nullable = bpf_task_acquire(t);
tools/testing/selftests/bpf/progs/verifier_global_subprogs.c
130
__noinline __weak int subprog_user_anon_mem(user_struct_t *t)
tools/testing/selftests/bpf/progs/verifier_global_subprogs.c
132
return t ? t->x : 0;
tools/testing/selftests/bpf/progs/verifier_global_subprogs.c
150
user_struct_t t = { .x = 42 };
tools/testing/selftests/bpf/progs/verifier_global_subprogs.c
152
return subprog_user_anon_mem(&t);
tools/testing/selftests/bpf/progs/verifier_global_subprogs.c
363
long *d, t, buf[1] = {};
tools/testing/selftests/bpf/progs/verifier_global_subprogs.c
369
t = *d + 1;
tools/testing/selftests/bpf/progs/verifier_global_subprogs.c
373
return t;
tools/testing/selftests/bpf/progs/verifier_global_subprogs.c
375
t = *d + 2;
tools/testing/selftests/bpf/progs/verifier_global_subprogs.c
377
return t;
tools/testing/selftests/bpf/progs/verifier_private_stack.c
14
struct bpf_timer t;
tools/testing/selftests/bpf/test_kmods/bpf_test_no_cfi.c
18
static int dummy_init_member(const struct btf_type *t,
tools/testing/selftests/bpf/test_kmods/bpf_test_rqspinlock.c
283
struct task_struct *t;
tools/testing/selftests/bpf/test_kmods/bpf_test_rqspinlock.c
285
t = kthread_create(rqspinlock_worker_fn, NULL, "rqsl_w/%d", i);
tools/testing/selftests/bpf/test_kmods/bpf_test_rqspinlock.c
286
if (IS_ERR(t)) {
tools/testing/selftests/bpf/test_kmods/bpf_test_rqspinlock.c
287
ret = PTR_ERR(t);
tools/testing/selftests/bpf/test_kmods/bpf_test_rqspinlock.c
290
kthread_bind(t, i);
tools/testing/selftests/bpf/test_kmods/bpf_test_rqspinlock.c
291
rqsl_threads[i] = t;
tools/testing/selftests/bpf/test_kmods/bpf_test_rqspinlock.c
292
wake_up_process(t);
tools/testing/selftests/bpf/test_kmods/bpf_testmod.c
1203
static void ctx_check_tasklet_fn(struct tasklet_struct *t)
tools/testing/selftests/bpf/test_kmods/bpf_testmod.c
1287
static int bpf_testmod_ops_init_member(const struct btf_type *t,
tools/testing/selftests/bpf/test_kmods/bpf_testmod.c
1444
static int st_ops3_check_member(const struct btf_type *t,
tools/testing/selftests/bpf/test_kmods/bpf_testmod.c
1448
u32 moff = __btf_member_bit_offset(t, member) / 8;
tools/testing/selftests/bpf/test_kmods/bpf_testmod.c
1685
static int st_ops_init_member(const struct btf_type *t,
tools/testing/selftests/bpf/test_kmods/bpf_testmod.c
1714
static int multi_st_ops_init_member(const struct btf_type *t,
tools/testing/selftests/bpf/test_loader.c
443
const struct btf_type *t;
tools/testing/selftests/bpf/test_loader.c
447
t = btf__type_by_id(btf, i);
tools/testing/selftests/bpf/test_loader.c
448
if (!btf_is_decl_tag(t))
tools/testing/selftests/bpf/test_loader.c
451
if (t->type != func_id || btf_decl_tag(t)->component_idx != -1)
tools/testing/selftests/bpf/test_loader.c
454
s = btf__str_by_offset(btf, t->name_off);
tools/testing/selftests/bpf/test_lru_map.c
856
int t, f;
tools/testing/selftests/bpf/test_lru_map.c
871
for (t = 0; t < ARRAY_SIZE(map_types); t++) {
tools/testing/selftests/bpf/test_lru_map.c
872
test_lru_sanity0(map_types[t], map_flags[f]);
tools/testing/selftests/bpf/test_lru_map.c
873
test_lru_sanity1(map_types[t], map_flags[f], tgt_free);
tools/testing/selftests/bpf/test_lru_map.c
874
test_lru_sanity2(map_types[t], map_flags[f], tgt_free);
tools/testing/selftests/bpf/test_lru_map.c
875
test_lru_sanity3(map_types[t], map_flags[f], tgt_free);
tools/testing/selftests/bpf/test_lru_map.c
876
test_lru_sanity4(map_types[t], map_flags[f], tgt_free);
tools/testing/selftests/bpf/test_lru_map.c
877
test_lru_sanity5(map_types[t], map_flags[f]);
tools/testing/selftests/bpf/test_lru_map.c
878
test_lru_sanity6(map_types[t], map_flags[f], tgt_free);
tools/testing/selftests/bpf/test_lru_map.c
879
test_lru_sanity7(map_types[t], map_flags[f]);
tools/testing/selftests/bpf/test_lru_map.c
880
test_lru_sanity8(map_types[t], map_flags[f]);
tools/testing/selftests/bpf/test_maps.c
1725
int map_fd, err, t, f;
tools/testing/selftests/bpf/test_maps.c
1759
for (t = 0; t < ARRAY_SIZE(types); t++) {
tools/testing/selftests/bpf/test_maps.c
1760
type = types[t];
tools/testing/selftests/bpf/test_progs.c
1230
const struct btf_type *t;
tools/testing/selftests/bpf/test_progs.c
1236
t = btf__type_by_id(btf, i);
tools/testing/selftests/bpf/test_progs.c
1237
if (!t || !btf_is_enum(t) || t->name_off)
tools/testing/selftests/bpf/test_progs.c
1239
e = btf_enum(t);
tools/testing/selftests/bpf/test_progs.c
1240
for (j = 0, vlen = btf_vlen(t); j < vlen; j++, e++) {
tools/testing/selftests/bpf/test_sockmap.c
191
static int test_start_subtest(const struct _test *t, struct sockmap_options *o)
tools/testing/selftests/bpf/test_sockmap.c
194
env.subtest = t->title;
tools/testing/selftests/bpf/test_sockmap.c
2003
static int check_whitelist(struct _test *t, struct sockmap_options *opt)
tools/testing/selftests/bpf/test_sockmap.c
2016
strstr(t->title, entry) != 0) {
tools/testing/selftests/bpf/test_sockmap.c
2026
static int check_blacklist(struct _test *t, struct sockmap_options *opt)
tools/testing/selftests/bpf/test_sockmap.c
2039
strstr(t->title, entry) != 0) {
tools/testing/selftests/bpf/test_sockmap.c
2061
struct _test t = test[i];
tools/testing/selftests/bpf/test_sockmap.c
2063
if (check_whitelist(&t, opt) != 0)
tools/testing/selftests/bpf/test_sockmap.c
2065
if (check_blacklist(&t, opt) == 0)
tools/testing/selftests/bpf/test_sockmap.c
2068
test_start_subtest(&t, opt);
tools/testing/selftests/bpf/test_sockmap.c
2069
t.tester(cg_fd, opt);
tools/testing/selftests/bpf/test_tcpnotify_user.c
34
struct tcp_notifier *t = data;
tools/testing/selftests/bpf/test_tcpnotify_user.c
36
if (t->type != 0xde || t->subtype != 0xad ||
tools/testing/selftests/bpf/test_tcpnotify_user.c
37
t->source != 0xbe || t->hash != 0xef)
tools/testing/selftests/bpf/test_verifier.c
1832
unsigned int t = atoi(argv[arg]);
tools/testing/selftests/bpf/test_verifier.c
1834
if (t < to) {
tools/testing/selftests/bpf/test_verifier.c
1835
from = t;
tools/testing/selftests/bpf/test_verifier.c
1836
to = t + 1;
tools/testing/selftests/bpf/testing_helpers.h
47
struct timespec t;
tools/testing/selftests/bpf/testing_helpers.h
49
clock_gettime(CLOCK_MONOTONIC, &t);
tools/testing/selftests/bpf/testing_helpers.h
51
return (u64)t.tv_sec * 1000000000 + t.tv_nsec;
tools/testing/selftests/bpf/veristat.c
1214
const struct btf_type *t, *mt;
tools/testing/selftests/bpf/veristat.c
1220
t = btf__type_by_id(btf, bpf_map__btf_value_type_id(map));
tools/testing/selftests/bpf/veristat.c
1221
if (!btf_is_struct(t))
tools/testing/selftests/bpf/veristat.c
1225
for (i = 0; i < btf_vlen(t); i++) {
tools/testing/selftests/bpf/veristat.c
1226
m = &btf_members(t)[i];
tools/testing/selftests/bpf/veristat.c
1274
const struct btf_type *t;
tools/testing/selftests/bpf/veristat.c
1282
t = btf__type_by_id(btf, id);
tools/testing/selftests/bpf/veristat.c
1283
t = btf__type_by_id(btf, t->type);
tools/testing/selftests/bpf/veristat.c
1284
if (!btf_is_func_proto(t) || btf_vlen(t) != 1)
tools/testing/selftests/bpf/veristat.c
1288
t = btf__type_by_id(btf, btf_params(t)[0].type);
tools/testing/selftests/bpf/veristat.c
1289
while (t && btf_is_mod(t))
tools/testing/selftests/bpf/veristat.c
1290
t = btf__type_by_id(btf, t->type);
tools/testing/selftests/bpf/veristat.c
1291
if (!t || !btf_is_ptr(t))
tools/testing/selftests/bpf/veristat.c
1293
t = btf__type_by_id(btf, t->type);
tools/testing/selftests/bpf/veristat.c
1294
while (t && btf_is_mod(t))
tools/testing/selftests/bpf/veristat.c
1295
t = btf__type_by_id(btf, t->type);
tools/testing/selftests/bpf/veristat.c
1296
if (!t)
tools/testing/selftests/bpf/veristat.c
1299
ctx_name = btf__name_by_offset(btf, t->name_off);
tools/testing/selftests/bpf/veristat.c
1830
static bool is_signed_type(const struct btf_type *t)
tools/testing/selftests/bpf/veristat.c
1832
if (btf_is_int(t))
tools/testing/selftests/bpf/veristat.c
1833
return btf_int_encoding(t) & BTF_INT_SIGNED;
tools/testing/selftests/bpf/veristat.c
1834
if (btf_is_any_enum(t))
tools/testing/selftests/bpf/veristat.c
1835
return btf_kflag(t);
tools/testing/selftests/bpf/veristat.c
1839
static int enum_value_from_name(const struct btf *btf, const struct btf_type *t,
tools/testing/selftests/bpf/veristat.c
1842
if (btf_is_enum(t)) {
tools/testing/selftests/bpf/veristat.c
1843
struct btf_enum *e = btf_enum(t);
tools/testing/selftests/bpf/veristat.c
1844
int i, n = btf_vlen(t);
tools/testing/selftests/bpf/veristat.c
1854
} else if (btf_is_enum64(t)) {
tools/testing/selftests/bpf/veristat.c
1855
struct btf_enum64 *e = btf_enum64(t);
tools/testing/selftests/bpf/veristat.c
1856
int i, n = btf_vlen(t);
tools/testing/selftests/bpf/veristat.c
1871
static bool is_preset_supported(const struct btf_type *t)
tools/testing/selftests/bpf/veristat.c
1873
return btf_is_int(t) || btf_is_enum(t) || btf_is_enum64(t);
tools/testing/selftests/bpf/veristat.c
1878
const struct btf_type *t;
tools/testing/selftests/bpf/veristat.c
1884
t = btf__type_by_id(btf, i);
tools/testing/selftests/bpf/veristat.c
1886
if (!btf_is_any_enum(t))
tools/testing/selftests/bpf/veristat.c
1889
if (enum_value_from_name(btf, t, name, &lvalue) == 0) {
tools/testing/selftests/bpf/veristat.c
1922
const struct btf_type *t;
tools/testing/selftests/bpf/veristat.c
1928
t = btf__type_by_id(btf, tid);
tools/testing/selftests/bpf/veristat.c
1929
if (!btf_is_array(t)) {
tools/testing/selftests/bpf/veristat.c
1934
barr = btf_array(t);
tools/testing/selftests/bpf/veristat.c
2001
static int adjust_var_secinfo(struct btf *btf, const struct btf_type *t,
tools/testing/selftests/bpf/veristat.c
2012
tid = btf__resolve_type(btf, t->type);
tools/testing/selftests/bpf/veristat.c
2111
const struct btf_type *t;
tools/testing/selftests/bpf/veristat.c
2125
t = btf__type_by_id(btf, i);
tools/testing/selftests/bpf/veristat.c
2127
if (!btf_is_datasec(t))
tools/testing/selftests/bpf/veristat.c
2130
sinfo = btf_var_secinfos(t);
tools/testing/selftests/bpf/veristat.c
2131
sec_name = btf__name_by_offset(btf, t->name_off);
tools/testing/selftests/bpf/veristat.c
2136
n = btf_vlen(t);
tools/testing/selftests/bpf/xdp_features.c
253
static int dut_run_echo_thread(pthread_t *t, int *sockfd)
tools/testing/selftests/bpf/xdp_features.c
267
err = pthread_create(t, NULL, dut_echo_thread, sockfd);
tools/testing/selftests/bpf/xdp_hw_metadata.c
187
struct timespec t;
tools/testing/selftests/bpf/xdp_hw_metadata.c
191
res = clock_gettime(clock_id, &t);
tools/testing/selftests/bpf/xdp_hw_metadata.c
196
return (__u64) t.tv_sec * NANOSEC_PER_SEC + t.tv_nsec;
tools/testing/selftests/cgroup/test_core.c
578
int t, c_threads = 0, n_threads = 13;
tools/testing/selftests/cgroup/test_core.c
607
for (t = 0; t < c_threads; ++t) {
tools/testing/selftests/cgroup/test_core.c
608
pthread_cancel(threads[t]);
tools/testing/selftests/cgroup/test_core.c
611
for (t = 0; t < c_threads; ++t) {
tools/testing/selftests/cgroup/test_core.c
612
pthread_join(threads[t], NULL);
tools/testing/selftests/drivers/net/hw/iou-zcrx.c
67
#define min_t(t, a, b) \
tools/testing/selftests/drivers/net/hw/iou-zcrx.c
69
t _ta = (a); \
tools/testing/selftests/drivers/net/hw/iou-zcrx.c
70
t _tb = (b); \
tools/testing/selftests/futex/functional/futex_numa.c
165
int c, t, threads = 2, contenders = 0;
tools/testing/selftests/futex/functional/futex_numa.c
194
for (t = 0; t < contenders; t++) {
tools/testing/selftests/futex/functional/futex_numa.c
212
cas[t] = args;
tools/testing/selftests/futex/functional/futex_numa.c
215
for (t = 0; t < threads; t++) {
tools/testing/selftests/futex/functional/futex_numa.c
233
tas[t] = args;
tools/testing/selftests/futex/functional/futex_numa.c
240
for (t = 0; t < threads; t++) {
tools/testing/selftests/futex/functional/futex_numa.c
241
struct thread_args *args = tas[t];
tools/testing/selftests/futex/functional/futex_numa.c
251
for (t = 0; t < contenders; t++) {
tools/testing/selftests/futex/functional/futex_numa.c
252
struct thread_args *args = cas[t];
tools/testing/selftests/hid/hid_common.h
445
time_t t;
tools/testing/selftests/hid/hid_common.h
449
srand((unsigned int)time(&t));
tools/testing/selftests/kselftest_harness.h
1000
t->exit_code = KSFT_FAIL;
tools/testing/selftests/kselftest_harness.h
1003
t->name);
tools/testing/selftests/kselftest_harness.h
1008
kill(-(t->pid), SIGKILL);
tools/testing/selftests/kselftest_harness.h
1010
child = waitpid(t->pid, &status, WNOHANG);
tools/testing/selftests/kselftest_harness.h
1012
t->exit_code = KSFT_FAIL;
tools/testing/selftests/kselftest_harness.h
1015
t->name, t->pid, errno);
tools/testing/selftests/kselftest_harness.h
1020
t->exit_code = KSFT_FAIL;
tools/testing/selftests/kselftest_harness.h
1022
"# %s: Test terminated by timeout\n", t->name);
tools/testing/selftests/kselftest_harness.h
1027
t->exit_code = WEXITSTATUS(status);
tools/testing/selftests/kselftest_harness.h
1028
} else if (t->termsig != -1) {
tools/testing/selftests/kselftest_harness.h
1029
t->exit_code = KSFT_FAIL;
tools/testing/selftests/kselftest_harness.h
1032
t->name,
tools/testing/selftests/kselftest_harness.h
1038
t->exit_code = KSFT_PASS;
tools/testing/selftests/kselftest_harness.h
1042
t->exit_code = KSFT_FAIL;
tools/testing/selftests/kselftest_harness.h
1045
t->name);
tools/testing/selftests/kselftest_harness.h
1049
t->exit_code = KSFT_FAIL;
tools/testing/selftests/kselftest_harness.h
1053
t->name);
tools/testing/selftests/kselftest_harness.h
1054
} else if (WTERMSIG(status) == t->termsig) {
tools/testing/selftests/kselftest_harness.h
1055
t->exit_code = KSFT_PASS;
tools/testing/selftests/kselftest_harness.h
1059
t->name,
tools/testing/selftests/kselftest_harness.h
1063
t->exit_code = KSFT_FAIL;
tools/testing/selftests/kselftest_harness.h
1066
t->name,
tools/testing/selftests/kselftest_harness.h
1075
struct __test_metadata *t;
tools/testing/selftests/kselftest_harness.h
1079
t = f->tests;
tools/testing/selftests/kselftest_harness.h
1089
t == f->tests ? f->name : "",
tools/testing/selftests/kselftest_harness.h
1091
t ? t->name : "");
tools/testing/selftests/kselftest_harness.h
1094
t = t ? t->next : NULL;
tools/testing/selftests/kselftest_harness.h
1095
} while (v || t);
tools/testing/selftests/kselftest_harness.h
1151
struct __test_metadata *t)
tools/testing/selftests/kselftest_harness.h
1164
if (!strcmp(t->name, optarg))
tools/testing/selftests/kselftest_harness.h
1179
tlen = strlen(t->name);
tools/testing/selftests/kselftest_harness.h
1184
!strncmp(t->name, &optarg[flen + 1 + vlen + !!vlen], tlen))
tools/testing/selftests/kselftest_harness.h
1199
struct __test_metadata *t)
tools/testing/selftests/kselftest_harness.h
1207
t->exit_code = KSFT_PASS;
tools/testing/selftests/kselftest_harness.h
1208
t->trigger = 0;
tools/testing/selftests/kselftest_harness.h
1209
t->aborted = false;
tools/testing/selftests/kselftest_harness.h
1210
t->no_teardown = NULL;
tools/testing/selftests/kselftest_harness.h
1211
memset(t->results->reason, 0, sizeof(t->results->reason));
tools/testing/selftests/kselftest_harness.h
1214
f->name, variant->name[0] ? "." : "", variant->name, t->name);
tools/testing/selftests/kselftest_harness.h
1225
t->exit_code = KSFT_FAIL;
tools/testing/selftests/kselftest_harness.h
1228
t->fn(t, variant);
tools/testing/selftests/kselftest_harness.h
1229
_exit(t->exit_code);
tools/testing/selftests/kselftest_harness.h
1231
t->pid = child;
tools/testing/selftests/kselftest_harness.h
1232
__wait_for_test(t);
tools/testing/selftests/kselftest_harness.h
1235
__test_passed(t) ? "OK" : "FAIL", test_name);
tools/testing/selftests/kselftest_harness.h
1239
if (xfail->test == t)
tools/testing/selftests/kselftest_harness.h
1242
t->exit_code = __test_passed(t) ? KSFT_XPASS : KSFT_XFAIL;
tools/testing/selftests/kselftest_harness.h
1244
if (t->results->reason[0])
tools/testing/selftests/kselftest_harness.h
1245
diagnostic = t->results->reason;
tools/testing/selftests/kselftest_harness.h
1246
else if (t->exit_code == KSFT_PASS || t->exit_code == KSFT_FAIL)
tools/testing/selftests/kselftest_harness.h
1251
ksft_test_result_code(t->exit_code, test_name,
tools/testing/selftests/kselftest_harness.h
1261
struct __test_metadata *t;
tools/testing/selftests/kselftest_harness.h
1275
for (t = f->tests; t; t = t->next)
tools/testing/selftests/kselftest_harness.h
1276
if (test_enabled(argc, argv, f, v, t))
tools/testing/selftests/kselftest_harness.h
1293
for (t = f->tests; t; t = t->next) {
tools/testing/selftests/kselftest_harness.h
1294
if (!test_enabled(argc, argv, f, v, t))
tools/testing/selftests/kselftest_harness.h
1297
t->results = results;
tools/testing/selftests/kselftest_harness.h
1298
__run_test(f, v, t);
tools/testing/selftests/kselftest_harness.h
1299
t->results = NULL;
tools/testing/selftests/kselftest_harness.h
1300
if (__test_passed(t))
tools/testing/selftests/kselftest_harness.h
952
static inline void __register_test(struct __test_metadata *t)
tools/testing/selftests/kselftest_harness.h
954
__LIST_APPEND(t->fixture->tests, t);
tools/testing/selftests/kselftest_harness.h
962
static inline int __bail(int for_realz, struct __test_metadata *t)
tools/testing/selftests/kselftest_harness.h
966
if (t->teardown_fn)
tools/testing/selftests/kselftest_harness.h
967
t->teardown_fn(false, t, t->self, t->variant);
tools/testing/selftests/kselftest_harness.h
974
static void __wait_for_test(struct __test_metadata *t)
tools/testing/selftests/kselftest_harness.h
987
childfd = syscall(__NR_pidfd_open, t->pid, 0);
tools/testing/selftests/kselftest_harness.h
989
t->exit_code = KSFT_FAIL;
tools/testing/selftests/kselftest_harness.h
992
t->name);
tools/testing/selftests/kselftest_harness.h
998
ret = poll(&poll_child, 1, t->timeout * 1000);
tools/testing/selftests/kvm/arm64/page_fault_test.c
1096
struct test_desc *t;
tools/testing/selftests/kvm/arm64/page_fault_test.c
1098
for (t = &tests[0]; t->name; t++) {
tools/testing/selftests/kvm/arm64/page_fault_test.c
1099
if (t->skip)
tools/testing/selftests/kvm/arm64/page_fault_test.c
1104
.test_desc = t,
tools/testing/selftests/kvm/arm64/vgic_irq.c
125
#define for_each_inject_fn(t, f) \
tools/testing/selftests/kvm/arm64/vgic_irq.c
126
for ((f) = (t); (f)->cmd; (f)++)
tools/testing/selftests/kvm/arm64/vgic_irq.c
128
#define for_each_supported_inject_fn(args, t, f) \
tools/testing/selftests/kvm/arm64/vgic_irq.c
129
for_each_inject_fn(t, f) \
tools/testing/selftests/kvm/arm64/vgic_irq.c
132
#define for_each_supported_activate_fn(args, t, f) \
tools/testing/selftests/kvm/arm64/vgic_irq.c
133
for_each_supported_inject_fn((args), (t), (f))
tools/testing/selftests/kvm/include/arm64/processor.h
110
#define PTE_ATTRINDX(t) ((t) << 2)
tools/testing/selftests/kvm/include/kvm_syscalls.h
10
#define MAP_ARGS3(m,t,a,...) m(t,a), MAP_ARGS2(m,__VA_ARGS__)
tools/testing/selftests/kvm/include/kvm_syscalls.h
11
#define MAP_ARGS4(m,t,a,...) m(t,a), MAP_ARGS3(m,__VA_ARGS__)
tools/testing/selftests/kvm/include/kvm_syscalls.h
12
#define MAP_ARGS5(m,t,a,...) m(t,a), MAP_ARGS4(m,__VA_ARGS__)
tools/testing/selftests/kvm/include/kvm_syscalls.h
13
#define MAP_ARGS6(m,t,a,...) m(t,a), MAP_ARGS5(m,__VA_ARGS__)
tools/testing/selftests/kvm/include/kvm_syscalls.h
16
#define __DECLARE_ARGS(t, a) t a
tools/testing/selftests/kvm/include/kvm_syscalls.h
17
#define __UNPACK_ARGS(t, a) a
tools/testing/selftests/kvm/include/kvm_syscalls.h
8
#define MAP_ARGS1(m,t,a,...) m(t,a)
tools/testing/selftests/kvm/include/kvm_syscalls.h
9
#define MAP_ARGS2(m,t,a,...) m(t,a), MAP_ARGS1(m,__VA_ARGS__)
tools/testing/selftests/kvm/include/test_util.h
181
static inline bool backing_src_is_shared(enum vm_mem_backing_src_type t)
tools/testing/selftests/kvm/include/test_util.h
183
return vm_mem_backing_src_alias(t)->flag & MAP_SHARED;
tools/testing/selftests/kvm/include/test_util.h
186
static inline bool backing_src_can_be_huge(enum vm_mem_backing_src_type t)
tools/testing/selftests/kvm/include/test_util.h
188
return t != VM_MEM_SRC_ANONYMOUS && t != VM_MEM_SRC_SHMEM;
tools/testing/selftests/kvm/include/x86/processor.h
789
#define sse128_lo(x) ({ __sse128_u t; t.vec = x; t.as_u64[0]; })
tools/testing/selftests/kvm/include/x86/processor.h
790
#define sse128_hi(x) ({ __sse128_u t; t.vec = x; t.as_u64[1]; })
tools/testing/selftests/kvm/s390/memop.c
1011
struct test_default t = test_default_init(guest_idle);
tools/testing/selftests/kvm/s390/memop.c
1014
HOST_SYNC(t.vcpu, STAGE_INITED);
tools/testing/selftests/kvm/s390/memop.c
1016
_test_errors_common(t.vcpu, LOGICAL, t.size);
tools/testing/selftests/kvm/s390/memop.c
1017
_test_errors_common(t.vm, ABSOLUTE, t.size);
tools/testing/selftests/kvm/s390/memop.c
1020
rv = ERR_MOP(t.vcpu, INVALID, WRITE, mem1, t.size, GADDR_V(mem1));
tools/testing/selftests/kvm/s390/memop.c
1023
rv = ERR_MOP(t.vm, INVALID, WRITE, mem1, PAGE_SIZE, GADDR(0));
tools/testing/selftests/kvm/s390/memop.c
1027
t.run->psw_mask &= ~(3UL << (63 - 17));
tools/testing/selftests/kvm/s390/memop.c
1028
t.run->psw_mask |= 1UL << (63 - 17); /* Enable AR mode */
tools/testing/selftests/kvm/s390/memop.c
1029
HOST_SYNC(t.vcpu, STAGE_IDLED); /* To sync new state to SIE block */
tools/testing/selftests/kvm/s390/memop.c
1030
rv = ERR_MOP(t.vcpu, LOGICAL, WRITE, mem1, t.size, GADDR_V(mem1), AR(17));
tools/testing/selftests/kvm/s390/memop.c
1032
t.run->psw_mask &= ~(3UL << (63 - 17)); /* Disable AR mode */
tools/testing/selftests/kvm/s390/memop.c
1033
HOST_SYNC(t.vcpu, STAGE_IDLED); /* Run to sync new state */
tools/testing/selftests/kvm/s390/memop.c
1036
rv = ERR_MOP(t.vcpu, SIDA, READ, mem1, 8, GADDR(0), SIDA_OFFSET(0x1c0));
tools/testing/selftests/kvm/s390/memop.c
1039
rv = ERR_MOP(t.vcpu, SIDA, WRITE, mem1, 8, GADDR(0), SIDA_OFFSET(0x1c0));
tools/testing/selftests/kvm/s390/memop.c
1043
kvm_vm_free(t.kvm_vm);
tools/testing/selftests/kvm/s390/memop.c
1048
struct test_default t = test_default_init(guest_idle);
tools/testing/selftests/kvm/s390/memop.c
1052
HOST_SYNC(t.vcpu, STAGE_INITED);
tools/testing/selftests/kvm/s390/memop.c
1059
rv = ERR_MOP(t.vm, ABSOLUTE, CMPXCHG, mem1, i, GADDR_V(mem1),
tools/testing/selftests/kvm/s390/memop.c
1065
rv = ERR_MOP(t.vm, ABSOLUTE, CMPXCHG, mem1, i, GADDR((void *)~0xfffUL),
tools/testing/selftests/kvm/s390/memop.c
1070
rv = ERR_MOP(t.vm, ABSOLUTE, CMPXCHG, mem1, i, GADDR_V(mem1 + 1),
tools/testing/selftests/kvm/s390/memop.c
1076
kvm_vm_free(t.kvm_vm);
tools/testing/selftests/kvm/s390/memop.c
247
struct test_default t;
tools/testing/selftests/kvm/s390/memop.c
249
t.size = min((size_t)kvm_check_cap(KVM_CAP_S390_MEM_OP), sizeof(mem1));
tools/testing/selftests/kvm/s390/memop.c
250
t.kvm_vm = vm_create_with_one_vcpu(&vcpu, guest_code);
tools/testing/selftests/kvm/s390/memop.c
251
t.vm = (struct test_info) { t.kvm_vm, NULL };
tools/testing/selftests/kvm/s390/memop.c
252
t.vcpu = (struct test_info) { t.kvm_vm, vcpu };
tools/testing/selftests/kvm/s390/memop.c
253
t.run = vcpu->run;
tools/testing/selftests/kvm/s390/memop.c
254
return t;
tools/testing/selftests/kvm/s390/memop.c
368
struct test_default t = test_default_init(guest_copy);
tools/testing/selftests/kvm/s390/memop.c
370
HOST_SYNC(t.vcpu, STAGE_INITED);
tools/testing/selftests/kvm/s390/memop.c
372
default_write_read(t.vcpu, t.vcpu, LOGICAL, t.size, NO_KEY);
tools/testing/selftests/kvm/s390/memop.c
374
kvm_vm_free(t.kvm_vm);
tools/testing/selftests/kvm/s390/memop.c
379
struct test_default t = test_default_init(guest_copy);
tools/testing/selftests/kvm/s390/memop.c
381
HOST_SYNC(t.vcpu, STAGE_INITED);
tools/testing/selftests/kvm/s390/memop.c
384
t.run->psw_mask &= ~(3UL << (63 - 17));
tools/testing/selftests/kvm/s390/memop.c
385
t.run->psw_mask |= 1UL << (63 - 17); /* Enable AR mode */
tools/testing/selftests/kvm/s390/memop.c
392
CHECK_N_DO(MOP, t.vcpu, LOGICAL, WRITE, mem1, t.size,
tools/testing/selftests/kvm/s390/memop.c
394
HOST_SYNC(t.vcpu, STAGE_COPIED);
tools/testing/selftests/kvm/s390/memop.c
396
CHECK_N_DO(MOP, t.vcpu, LOGICAL, READ, mem2, t.size,
tools/testing/selftests/kvm/s390/memop.c
398
ASSERT_MEM_EQ(mem1, mem2, t.size);
tools/testing/selftests/kvm/s390/memop.c
400
kvm_vm_free(t.kvm_vm);
tools/testing/selftests/kvm/s390/memop.c
440
struct test_default t = test_default_init(guest_copy_key);
tools/testing/selftests/kvm/s390/memop.c
442
HOST_SYNC(t.vcpu, STAGE_SKEYS_SET);
tools/testing/selftests/kvm/s390/memop.c
445
default_write_read(t.vcpu, t.vm, ABSOLUTE, t.size, NO_KEY);
tools/testing/selftests/kvm/s390/memop.c
448
default_write_read(t.vcpu, t.vcpu, LOGICAL, t.size, 0);
tools/testing/selftests/kvm/s390/memop.c
449
default_write_read(t.vcpu, t.vcpu, LOGICAL, t.size, 9);
tools/testing/selftests/kvm/s390/memop.c
450
default_write_read(t.vcpu, t.vm, ABSOLUTE, t.size, 0);
tools/testing/selftests/kvm/s390/memop.c
451
default_write_read(t.vcpu, t.vm, ABSOLUTE, t.size, 9);
tools/testing/selftests/kvm/s390/memop.c
457
default_write_read(t.vcpu, t.vcpu, LOGICAL, 1, 0);
tools/testing/selftests/kvm/s390/memop.c
458
default_write_read(t.vcpu, t.vcpu, LOGICAL, 1, 9);
tools/testing/selftests/kvm/s390/memop.c
459
default_write_read(t.vcpu, t.vm, ABSOLUTE, 1, 0);
tools/testing/selftests/kvm/s390/memop.c
460
default_write_read(t.vcpu, t.vm, ABSOLUTE, 1, 9);
tools/testing/selftests/kvm/s390/memop.c
463
default_read(t.vcpu, t.vcpu, LOGICAL, t.size, 2);
tools/testing/selftests/kvm/s390/memop.c
464
default_read(t.vcpu, t.vm, ABSOLUTE, t.size, 2);
tools/testing/selftests/kvm/s390/memop.c
466
kvm_vm_free(t.kvm_vm);
tools/testing/selftests/kvm/s390/memop.c
471
struct test_default t = test_default_init(guest_copy_key);
tools/testing/selftests/kvm/s390/memop.c
473
HOST_SYNC(t.vcpu, STAGE_SKEYS_SET);
tools/testing/selftests/kvm/s390/memop.c
475
default_cmpxchg(&t, NO_KEY);
tools/testing/selftests/kvm/s390/memop.c
476
default_cmpxchg(&t, 0);
tools/testing/selftests/kvm/s390/memop.c
477
default_cmpxchg(&t, 9);
tools/testing/selftests/kvm/s390/memop.c
479
kvm_vm_free(t.kvm_vm);
tools/testing/selftests/kvm/s390/memop.c
672
struct test_default t = test_default_init(guest_cmpxchg_key);
tools/testing/selftests/kvm/s390/memop.c
678
HOST_SYNC(t.vcpu, STAGE_SKEYS_SET);
tools/testing/selftests/kvm/s390/memop.c
680
MOP(t.vcpu, LOGICAL, WRITE, mem1, max_block, GADDR_V(mem2));
tools/testing/selftests/kvm/s390/memop.c
681
pthread_create(&thread, NULL, run_guest, &t.vcpu);
tools/testing/selftests/kvm/s390/memop.c
687
MOP(t.vm, ABSOLUTE, CMPXCHG, &new,
tools/testing/selftests/kvm/s390/memop.c
696
MOP(t.vm, ABSOLUTE, CMPXCHG, quad_to_char(&new, size),
tools/testing/selftests/kvm/s390/memop.c
706
MOP(t.vcpu, LOGICAL, READ, mem2, max_block, GADDR_V(mem2));
tools/testing/selftests/kvm/s390/memop.c
710
kvm_vm_free(t.kvm_vm);
tools/testing/selftests/kvm/s390/memop.c
733
struct test_default t = test_default_init(guest_copy_key_fetch_prot);
tools/testing/selftests/kvm/s390/memop.c
735
HOST_SYNC(t.vcpu, STAGE_INITED);
tools/testing/selftests/kvm/s390/memop.c
736
t.run->s.regs.crs[0] |= CR0_STORAGE_PROTECTION_OVERRIDE;
tools/testing/selftests/kvm/s390/memop.c
737
t.run->kvm_dirty_regs = KVM_SYNC_CRS;
tools/testing/selftests/kvm/s390/memop.c
738
HOST_SYNC(t.vcpu, STAGE_SKEYS_SET);
tools/testing/selftests/kvm/s390/memop.c
741
default_write_read(t.vcpu, t.vcpu, LOGICAL, t.size, 2);
tools/testing/selftests/kvm/s390/memop.c
743
kvm_vm_free(t.kvm_vm);
tools/testing/selftests/kvm/s390/memop.c
748
struct test_default t = test_default_init(guest_copy_key_fetch_prot);
tools/testing/selftests/kvm/s390/memop.c
750
HOST_SYNC(t.vcpu, STAGE_INITED);
tools/testing/selftests/kvm/s390/memop.c
751
HOST_SYNC(t.vcpu, STAGE_SKEYS_SET);
tools/testing/selftests/kvm/s390/memop.c
754
default_read(t.vcpu, t.vcpu, LOGICAL, t.size, 9);
tools/testing/selftests/kvm/s390/memop.c
755
default_read(t.vcpu, t.vm, ABSOLUTE, t.size, 9);
tools/testing/selftests/kvm/s390/memop.c
757
kvm_vm_free(t.kvm_vm);
tools/testing/selftests/kvm/s390/memop.c
779
struct test_default t = test_default_init(guest_error_key);
tools/testing/selftests/kvm/s390/memop.c
781
HOST_SYNC(t.vcpu, STAGE_INITED);
tools/testing/selftests/kvm/s390/memop.c
782
HOST_SYNC(t.vcpu, STAGE_SKEYS_SET);
tools/testing/selftests/kvm/s390/memop.c
785
CHECK_N_DO(ERR_PROT_MOP, t.vcpu, LOGICAL, WRITE, mem1, t.size, GADDR_V(mem1), KEY(2));
tools/testing/selftests/kvm/s390/memop.c
786
CHECK_N_DO(ERR_PROT_MOP, t.vcpu, LOGICAL, READ, mem2, t.size, GADDR_V(mem1), KEY(2));
tools/testing/selftests/kvm/s390/memop.c
787
CHECK_N_DO(ERR_PROT_MOP, t.vm, ABSOLUTE, WRITE, mem1, t.size, GADDR_V(mem1), KEY(2));
tools/testing/selftests/kvm/s390/memop.c
788
CHECK_N_DO(ERR_PROT_MOP, t.vm, ABSOLUTE, READ, mem2, t.size, GADDR_V(mem1), KEY(2));
tools/testing/selftests/kvm/s390/memop.c
790
kvm_vm_free(t.kvm_vm);
tools/testing/selftests/kvm/s390/memop.c
795
struct test_default t = test_default_init(guest_copy_key_fetch_prot);
tools/testing/selftests/kvm/s390/memop.c
798
HOST_SYNC(t.vcpu, STAGE_INITED);
tools/testing/selftests/kvm/s390/memop.c
799
HOST_SYNC(t.vcpu, STAGE_SKEYS_SET);
tools/testing/selftests/kvm/s390/memop.c
804
ERR_PROT_MOP(t.vm, ABSOLUTE, CMPXCHG, mem2, i, GADDR_V(mem2),
tools/testing/selftests/kvm/s390/memop.c
808
kvm_vm_free(t.kvm_vm);
tools/testing/selftests/kvm/s390/memop.c
813
struct test_default t = test_default_init(guest_error_key);
tools/testing/selftests/kvm/s390/memop.c
819
HOST_SYNC(t.vcpu, STAGE_INITED);
tools/testing/selftests/kvm/s390/memop.c
820
HOST_SYNC(t.vcpu, STAGE_SKEYS_SET);
tools/testing/selftests/kvm/s390/memop.c
823
ERR_PROT_MOP(t.vcpu, LOGICAL, WRITE, mem1, t.size, GADDR_V(mem1), KEY(1), INJECT);
tools/testing/selftests/kvm/s390/memop.c
830
prefix = t.run->s.regs.prefix;
tools/testing/selftests/kvm/s390/memop.c
831
psw[0] = t.run->psw_mask;
tools/testing/selftests/kvm/s390/memop.c
832
psw[1] = t.run->psw_addr;
tools/testing/selftests/kvm/s390/memop.c
833
MOP(t.vm, ABSOLUTE, WRITE, psw, sizeof(psw), GADDR(prefix + 464));
tools/testing/selftests/kvm/s390/memop.c
834
HOST_SYNC(t.vcpu, STAGE_IDLED);
tools/testing/selftests/kvm/s390/memop.c
835
MOP(t.vm, ABSOLUTE, READ, &teid, sizeof(teid), GADDR(prefix + 168));
tools/testing/selftests/kvm/s390/memop.c
839
kvm_vm_free(t.kvm_vm);
tools/testing/selftests/kvm/s390/memop.c
844
struct test_default t = test_default_init(guest_copy_key_fetch_prot);
tools/testing/selftests/kvm/s390/memop.c
846
HOST_SYNC(t.vcpu, STAGE_INITED);
tools/testing/selftests/kvm/s390/memop.c
847
t.run->s.regs.crs[0] |= CR0_STORAGE_PROTECTION_OVERRIDE;
tools/testing/selftests/kvm/s390/memop.c
848
t.run->kvm_dirty_regs = KVM_SYNC_CRS;
tools/testing/selftests/kvm/s390/memop.c
849
HOST_SYNC(t.vcpu, STAGE_SKEYS_SET);
tools/testing/selftests/kvm/s390/memop.c
852
CHECK_N_DO(ERR_PROT_MOP, t.vm, ABSOLUTE, WRITE, mem1, t.size, GADDR_V(mem1), KEY(2));
tools/testing/selftests/kvm/s390/memop.c
853
CHECK_N_DO(ERR_PROT_MOP, t.vm, ABSOLUTE, READ, mem2, t.size, GADDR_V(mem2), KEY(2));
tools/testing/selftests/kvm/s390/memop.c
855
kvm_vm_free(t.kvm_vm);
tools/testing/selftests/kvm/s390/memop.c
880
struct test_default t = test_default_init(guest_copy_key_fetch_prot_override);
tools/testing/selftests/kvm/s390/memop.c
883
guest_0_page = vm_vaddr_alloc(t.kvm_vm, PAGE_SIZE, 0);
tools/testing/selftests/kvm/s390/memop.c
884
guest_last_page = vm_vaddr_alloc(t.kvm_vm, PAGE_SIZE, last_page_addr);
tools/testing/selftests/kvm/s390/memop.c
890
HOST_SYNC(t.vcpu, STAGE_INITED);
tools/testing/selftests/kvm/s390/memop.c
891
t.run->s.regs.crs[0] |= CR0_FETCH_PROTECTION_OVERRIDE;
tools/testing/selftests/kvm/s390/memop.c
892
t.run->kvm_dirty_regs = KVM_SYNC_CRS;
tools/testing/selftests/kvm/s390/memop.c
893
HOST_SYNC(t.vcpu, STAGE_SKEYS_SET);
tools/testing/selftests/kvm/s390/memop.c
897
MOP(t.vcpu, LOGICAL, WRITE, mem1, PAGE_SIZE, GADDR_V(mem1));
tools/testing/selftests/kvm/s390/memop.c
898
HOST_SYNC(t.vcpu, STAGE_COPIED);
tools/testing/selftests/kvm/s390/memop.c
899
CHECK_N_DO(MOP, t.vcpu, LOGICAL, READ, mem2, 2048, GADDR_V(guest_0_page), KEY(2));
tools/testing/selftests/kvm/s390/memop.c
907
MOP(t.vcpu, LOGICAL, WRITE, mem1, 2 * PAGE_SIZE, GADDR_V(guest_last_page));
tools/testing/selftests/kvm/s390/memop.c
908
HOST_SYNC(t.vcpu, STAGE_COPIED);
tools/testing/selftests/kvm/s390/memop.c
909
CHECK_N_DO(MOP, t.vcpu, LOGICAL, READ, mem2, PAGE_SIZE + 2048,
tools/testing/selftests/kvm/s390/memop.c
914
kvm_vm_free(t.kvm_vm);
tools/testing/selftests/kvm/s390/memop.c
919
struct test_default t = test_default_init(guest_copy_key_fetch_prot_override);
tools/testing/selftests/kvm/s390/memop.c
922
guest_0_page = vm_vaddr_alloc(t.kvm_vm, PAGE_SIZE, 0);
tools/testing/selftests/kvm/s390/memop.c
923
guest_last_page = vm_vaddr_alloc(t.kvm_vm, PAGE_SIZE, last_page_addr);
tools/testing/selftests/kvm/s390/memop.c
928
HOST_SYNC(t.vcpu, STAGE_INITED);
tools/testing/selftests/kvm/s390/memop.c
929
HOST_SYNC(t.vcpu, STAGE_SKEYS_SET);
tools/testing/selftests/kvm/s390/memop.c
932
CHECK_N_DO(ERR_PROT_MOP, t.vcpu, LOGICAL, READ, mem2, 2048, GADDR_V(0), KEY(2));
tools/testing/selftests/kvm/s390/memop.c
935
kvm_vm_free(t.kvm_vm);
tools/testing/selftests/kvm/s390/memop.c
940
struct test_default t = test_default_init(guest_copy_key_fetch_prot_override);
tools/testing/selftests/kvm/s390/memop.c
943
guest_0_page = vm_vaddr_alloc(t.kvm_vm, PAGE_SIZE, 0);
tools/testing/selftests/kvm/s390/memop.c
944
guest_last_page = vm_vaddr_alloc(t.kvm_vm, PAGE_SIZE, last_page_addr);
tools/testing/selftests/kvm/s390/memop.c
949
HOST_SYNC(t.vcpu, STAGE_INITED);
tools/testing/selftests/kvm/s390/memop.c
950
t.run->s.regs.crs[0] |= CR0_FETCH_PROTECTION_OVERRIDE;
tools/testing/selftests/kvm/s390/memop.c
951
t.run->kvm_dirty_regs = KVM_SYNC_CRS;
tools/testing/selftests/kvm/s390/memop.c
952
HOST_SYNC(t.vcpu, STAGE_SKEYS_SET);
tools/testing/selftests/kvm/s390/memop.c
958
CHECK_N_DO(ERR_PROT_MOP, t.vcpu, LOGICAL, READ, mem2, 2048 + 1, GADDR_V(0), KEY(2));
tools/testing/selftests/kvm/s390/memop.c
959
CHECK_N_DO(ERR_PROT_MOP, t.vcpu, LOGICAL, READ, mem2, PAGE_SIZE + 2048 + 1,
tools/testing/selftests/kvm/s390/memop.c
962
CHECK_N_DO(ERR_PROT_MOP, t.vm, ABSOLUTE, READ, mem2, 2048, GADDR(0), KEY(2));
tools/testing/selftests/kvm/s390/memop.c
963
CHECK_N_DO(ERR_PROT_MOP, t.vm, ABSOLUTE, READ, mem2, 2048, GADDR_V(guest_0_page), KEY(2));
tools/testing/selftests/kvm/s390/memop.c
966
kvm_vm_free(t.kvm_vm);
tools/testing/selftests/kvm/x86/recalc_apic_map_test.c
41
time_t t;
tools/testing/selftests/kvm/x86/recalc_apic_map_test.c
63
for (t = time(NULL) + TIMEOUT; time(NULL) < t;) {
tools/testing/selftests/kvm/x86/sync_regs_test.c
163
time_t t;
tools/testing/selftests/kvm/x86/sync_regs_test.c
186
for (t = time(NULL) + TIMEOUT; time(NULL) < t;) {
tools/testing/selftests/kvm/x86/xen_shinfo_test.c
907
for (time_t t = time(NULL) + SHINFO_RACE_TIMEOUT; time(NULL) < t;)
tools/testing/selftests/landlock/tsync_test.c
123
pthread_t t;
tools/testing/selftests/landlock/tsync_test.c
148
ASSERT_EQ(0, pthread_create(&d[0].t, NULL, thread_restrict, &d[0]));
tools/testing/selftests/landlock/tsync_test.c
149
ASSERT_EQ(0, pthread_create(&d[1].t, NULL, thread_restrict, &d[1]));
tools/testing/selftests/landlock/tsync_test.c
152
ASSERT_EQ(0, pthread_join(d[0].t, NULL));
tools/testing/selftests/landlock/tsync_test.c
153
ASSERT_EQ(0, pthread_join(d[1].t, NULL));
tools/testing/selftests/mm/hmm-tests.c
266
struct timespec t;
tools/testing/selftests/mm/hmm-tests.c
268
t.tv_sec = 0;
tools/testing/selftests/mm/hmm-tests.c
269
t.tv_nsec = n;
tools/testing/selftests/mm/hmm-tests.c
270
nanosleep(&t, NULL);
tools/testing/selftests/mm/khugepaged.c
1223
#define TEST(t, c, o) do { \
tools/testing/selftests/mm/khugepaged.c
1225
printf("\nRun test: " #t " (%s:%s)\n", c->name, o->name); \
tools/testing/selftests/mm/khugepaged.c
1226
t(c, o); \
tools/testing/selftests/mm/mremap_test.c
1086
for (t = shift; t < shift + chunk_size; ++t) {
tools/testing/selftests/mm/mremap_test.c
1087
if (((char *) dest_addr)[t] != rand_addr[t]) {
tools/testing/selftests/mm/mremap_test.c
1089
t);
tools/testing/selftests/mm/mremap_test.c
1090
ksft_print_msg("Expected: %#x\t Got: %#x\n", rand_addr[t] & 0xff,
tools/testing/selftests/mm/mremap_test.c
1091
((char *) dest_addr)[t] & 0xff);
tools/testing/selftests/mm/mremap_test.c
1102
for (t = num_chunks * (threshold / num_chunks); t < threshold; ++t) {
tools/testing/selftests/mm/mremap_test.c
1103
if (((char *) dest_addr)[t] != rand_addr[t]) {
tools/testing/selftests/mm/mremap_test.c
1105
t);
tools/testing/selftests/mm/mremap_test.c
1106
ksft_print_msg("Expected: %#x\t Got: %#x\n", rand_addr[t] & 0xff,
tools/testing/selftests/mm/mremap_test.c
1107
((char *) dest_addr)[t] & 0xff);
tools/testing/selftests/mm/mremap_test.c
1339
time_t t;
tools/testing/selftests/mm/mremap_test.c
1342
pattern_seed = (unsigned int) time(&t);
tools/testing/selftests/mm/mremap_test.c
998
unsigned long long t, d;
tools/testing/selftests/mm/va_high_addr_switch.c
248
testcases = malloc(sizeof(t));
tools/testing/selftests/mm/va_high_addr_switch.c
252
memcpy(testcases, t, sizeof(t));
tools/testing/selftests/mm/va_high_addr_switch.c
255
sz_testcases = ARRAY_SIZE(t);
tools/testing/selftests/mm/va_high_addr_switch.c
266
struct testcase *t = test + i;
tools/testing/selftests/mm/va_high_addr_switch.c
268
p = mmap(t->addr, t->size, PROT_READ | PROT_WRITE, t->flags, -1, 0);
tools/testing/selftests/mm/va_high_addr_switch.c
270
printf("%s: %p - ", t->msg, p);
tools/testing/selftests/mm/va_high_addr_switch.c
278
if (t->low_addr_required && p >= (void *)(switch_hint)) {
tools/testing/selftests/mm/va_high_addr_switch.c
286
memset(p, 0, t->size);
tools/testing/selftests/mm/va_high_addr_switch.c
289
if (!t->keep_mapped)
tools/testing/selftests/mm/va_high_addr_switch.c
290
munmap(p, t->size);
tools/testing/selftests/mm/va_high_addr_switch.c
58
struct testcase t[] = {
tools/testing/selftests/mqueue/mq_perf_tests.c
433
pthread_t *t;
tools/testing/selftests/mqueue/mq_perf_tests.c
438
t = &cpu_threads[0];
tools/testing/selftests/net/ip_local_port_range.c
224
const struct test *t;
tools/testing/selftests/net/ip_local_port_range.c
226
for (t = tests; t < tests + ARRAY_SIZE(tests); t++) {
tools/testing/selftests/net/ip_local_port_range.c
234
TH_LOG("lo %5hu, hi %5hu", t->range_lo, t->range_hi);
tools/testing/selftests/net/ip_local_port_range.c
243
range = pack_port_range(t->range_lo, t->range_hi);
tools/testing/selftests/net/ip_local_port_range.c
277
const struct test *t;
tools/testing/selftests/net/ip_local_port_range.c
279
for (t = tests; t < tests + ARRAY_SIZE(tests); t++) {
tools/testing/selftests/net/ip_local_port_range.c
284
t->range_lo, t->range_hi, t->expected);
tools/testing/selftests/net/ip_local_port_range.c
289
range = pack_port_range(t->range_lo, t->range_hi);
tools/testing/selftests/net/ip_local_port_range.c
297
ASSERT_EQ(port, t->expected) TH_LOG("unexpected local port");
tools/testing/selftests/net/ipsec.c
1690
struct xfrmu_spdhthresh *t = RTA_DATA(attr);
tools/testing/selftests/net/ipsec.c
1693
if (t->lbits != 32 || t->rbits != 31) {
tools/testing/selftests/net/ipsec.c
1695
t->lbits, t->rbits);
tools/testing/selftests/net/ipsec.c
1700
struct xfrmu_spdhthresh *t = RTA_DATA(attr);
tools/testing/selftests/net/ipsec.c
1703
if (t->lbits != 120 || t->rbits != 16) {
tools/testing/selftests/net/ipsec.c
1705
t->lbits, t->rbits);
tools/testing/selftests/net/ipsec.c
630
struct timeval t = { .tv_sec = 0, .tv_usec = u_timeout };
tools/testing/selftests/net/ipsec.c
655
if (setsockopt(sock[0], SOL_SOCKET, SO_RCVTIMEO, (const char *)&t, sizeof t)) {
tools/testing/selftests/net/rxtimestamp.c
146
void print_test_case(struct test_case *t)
tools/testing/selftests/net/rxtimestamp.c
151
if (t->sockopt.so_timestamp)
tools/testing/selftests/net/rxtimestamp.c
153
if (t->sockopt.so_timestampns)
tools/testing/selftests/net/rxtimestamp.c
155
if (t->sockopt.so_timestamping) {
tools/testing/selftests/net/rxtimestamp.c
158
if (t->sockopt.so_timestamping & sof_flags[f].mask)
tools/testing/selftests/net/rxtimestamp.c
163
if (t->expected.tstamp)
tools/testing/selftests/net/rxtimestamp.c
165
if (t->expected.tstampns)
tools/testing/selftests/net/rxtimestamp.c
167
if (t->expected.swtstamp || t->expected.hwtstamp) {
tools/testing/selftests/net/rxtimestamp.c
169
if (t->expected.swtstamp)
tools/testing/selftests/net/rxtimestamp.c
171
if (t->expected.swtstamp && t->expected.hwtstamp)
tools/testing/selftests/net/rxtimestamp.c
173
if (t->expected.hwtstamp)
tools/testing/selftests/net/rxtimestamp.c
380
int s, t, opt;
tools/testing/selftests/net/rxtimestamp.c
386
for (t = 0; t < ARRAY_SIZE(test_cases); t++) {
tools/testing/selftests/net/rxtimestamp.c
387
printf("%d\t", t);
tools/testing/selftests/net/rxtimestamp.c
388
print_test_case(&test_cases[t]);
tools/testing/selftests/net/rxtimestamp.c
392
t = atoi(optarg);
tools/testing/selftests/net/rxtimestamp.c
393
if (t >= ARRAY_SIZE(test_cases))
tools/testing/selftests/net/rxtimestamp.c
394
error(1, 0, "Invalid test case: %d", t);
tools/testing/selftests/net/rxtimestamp.c
396
test_cases[t].enabled = true;
tools/testing/selftests/net/rxtimestamp.c
432
for (t = 0; t < ARRAY_SIZE(test_cases); t++) {
tools/testing/selftests/net/rxtimestamp.c
433
if (!all_tests && !test_cases[t].enabled)
tools/testing/selftests/net/rxtimestamp.c
436
if (run_test_case(&socket_types[s], t, '4',
tools/testing/selftests/net/rxtimestamp.c
440
if (run_test_case(&socket_types[s], t, '6',
tools/testing/selftests/net/tcp_ao/lib/setup.c
264
pthread_t t;
tools/testing/selftests/net/tcp_ao/lib/setup.c
270
if (pthread_create(&t, NULL, new_pthread_entry, &targ))
tools/testing/selftests/net/tcp_ao/lib/sock.c
94
uint64_t t;
tools/testing/selftests/net/tcp_ao/lib/sock.c
96
for (t = 0; t <= timeout * 1000000; t += POLL_USEC) {
tools/testing/selftests/powerpc/include/instructions.h
103
__PPC_RT(t) | \
tools/testing/selftests/powerpc/include/instructions.h
107
#define PREFIX_8LS(instr, t, a, r, d) stringify_in_c(.balign 64, , 4;) \
tools/testing/selftests/powerpc/include/instructions.h
112
__PPC_RT(t) | \
tools/testing/selftests/powerpc/include/instructions.h
117
#define PLBZ(t, a, r, d) PREFIX_MLS(PPC_INST_LBZ, t, a, r, d)
tools/testing/selftests/powerpc/include/instructions.h
118
#define PLHZ(t, a, r, d) PREFIX_MLS(PPC_INST_LHZ, t, a, r, d)
tools/testing/selftests/powerpc/include/instructions.h
119
#define PLHA(t, a, r, d) PREFIX_MLS(PPC_INST_LHA, t, a, r, d)
tools/testing/selftests/powerpc/include/instructions.h
120
#define PLWZ(t, a, r, d) PREFIX_MLS(PPC_INST_LWZ, t, a, r, d)
tools/testing/selftests/powerpc/include/instructions.h
121
#define PLWA(t, a, r, d) PREFIX_8LS(0xa4000000, t, a, r, d)
tools/testing/selftests/powerpc/include/instructions.h
122
#define PLD(t, a, r, d) PREFIX_8LS(0xe4000000, t, a, r, d)
tools/testing/selftests/powerpc/include/instructions.h
123
#define PLQ(t, a, r, d) PREFIX_8LS(0xe0000000, t, a, r, d)
tools/testing/selftests/powerpc/include/instructions.h
79
#define __PPC_RT(t) __PPC_RS(t)
tools/testing/selftests/powerpc/include/instructions.h
98
#define PREFIX_MLS(instr, t, a, r, d) stringify_in_c(.balign 64, , 4;) \
tools/testing/selftests/powerpc/nx-gzip/gzip_vas.c
138
uint64_t t;
tools/testing/selftests/powerpc/nx-gzip/gzip_vas.c
150
t = __ppc_get_timebase();
tools/testing/selftests/powerpc/nx-gzip/gzip_vas.c
162
if ((__ppc_get_timebase() - t) > USLEEP_TH) {
tools/testing/selftests/powerpc/nx-gzip/gzip_vas.c
293
volatile char t;
tools/testing/selftests/powerpc/nx-gzip/gzip_vas.c
304
t = *begin;
tools/testing/selftests/powerpc/nx-gzip/gzip_vas.c
306
*begin = t;
tools/testing/selftests/powerpc/nx-gzip/gzip_vas.c
311
t = *end;
tools/testing/selftests/powerpc/nx-gzip/gzip_vas.c
313
*end = t;
tools/testing/selftests/powerpc/nx-gzip/include/nx_dbg.h
35
time_t t; struct tm *m; time(&t); m = localtime(&t); \
tools/testing/selftests/powerpc/papr_sysparm/papr_sysparm.c
189
const struct sysparm_test *t = &sysparm_tests[i];
tools/testing/selftests/powerpc/papr_sysparm/papr_sysparm.c
191
if (test_harness(t->function, t->description))
tools/testing/selftests/powerpc/papr_vpd/papr_vpd.c
345
const struct vpd_test *t = &vpd_tests[i];
tools/testing/selftests/powerpc/papr_vpd/papr_vpd.c
347
if (test_harness(t->function, t->description))
tools/testing/selftests/powerpc/signal/sigfuz.c
203
pid_t t;
tools/testing/selftests/powerpc/signal/sigfuz.c
222
t = fork();
tools/testing/selftests/powerpc/signal/sigfuz.c
224
if (t == 0) {
tools/testing/selftests/powerpc/signal/sigfuz.c
234
waitpid(t, &ret, 0);
tools/testing/selftests/powerpc/signal/sigfuz.c
251
int t, rc;
tools/testing/selftests/powerpc/signal/sigfuz.c
256
for (t = 0; t < nthread; t++) {
tools/testing/selftests/powerpc/signal/sigfuz.c
257
rc = pthread_create(&threads[t], NULL, sigfuz_test,
tools/testing/selftests/powerpc/signal/sigfuz.c
258
(void *)&t);
tools/testing/selftests/powerpc/signal/sigfuz.c
263
for (t = 0; t < nthread; t++) {
tools/testing/selftests/powerpc/signal/sigfuz.c
264
rc = pthread_join(threads[t], NULL);
tools/testing/selftests/powerpc/stringloops/asm/ppc-opcode.h
25
#define ___PPC_RT(t) ___PPC_RS(t)
tools/testing/selftests/ptp/testptp.c
111
static int64_t pctns(struct ptp_clock_time *t)
tools/testing/selftests/ptp/testptp.c
113
return t->sec * NSEC_PER_SEC + t->nsec;
tools/testing/selftests/ptp/testptp.c
469
event.t.sec, event.t.nsec);
tools/testing/selftests/timers/inconsistency-check.c
130
t = time(0);
tools/testing/selftests/timers/inconsistency-check.c
131
ksft_print_msg("%s\n", ctime(&t));
tools/testing/selftests/timers/inconsistency-check.c
88
time_t t;
tools/testing/selftests/timers/inconsistency-check.c
95
t = time(0);
tools/testing/selftests/timers/inconsistency-check.c
96
start_str = ctime(&t);
tools/testing/selftests/ublk/batch.c
103
t->commit_buf = buf;
tools/testing/selftests/ublk/batch.c
106
if (mlock(t->commit_buf, total))
tools/testing/selftests/ublk/batch.c
113
free_batch_commit_buf(t);
tools/testing/selftests/ublk/batch.c
117
static unsigned int ublk_thread_nr_queues(const struct ublk_thread *t)
tools/testing/selftests/ublk/batch.c
122
for (i = 0; i < t->dev->dev_info.nr_hw_queues; i++)
tools/testing/selftests/ublk/batch.c
123
ret += !!t->q_map[i];
tools/testing/selftests/ublk/batch.c
128
void ublk_batch_prepare(struct ublk_thread *t)
tools/testing/selftests/ublk/batch.c
13
if (buf_idx < t->commit_buf_start ||
tools/testing/selftests/ublk/batch.c
138
struct ublk_queue *q = &t->dev->q[0];
tools/testing/selftests/ublk/batch.c
14
buf_idx >= t->commit_buf_start + t->nr_commit_buf)
tools/testing/selftests/ublk/batch.c
141
t->nr_queues = ublk_thread_nr_queues(t);
tools/testing/selftests/ublk/batch.c
143
t->commit_buf_elem_size = ublk_commit_elem_buf_size(t->dev);
tools/testing/selftests/ublk/batch.c
144
t->commit_buf_size = ublk_commit_buf_size(t);
tools/testing/selftests/ublk/batch.c
145
t->commit_buf_start = t->nr_bufs;
tools/testing/selftests/ublk/batch.c
146
t->nr_commit_buf = 2 * t->nr_queues;
tools/testing/selftests/ublk/batch.c
147
t->nr_bufs += t->nr_commit_buf;
tools/testing/selftests/ublk/batch.c
149
t->cmd_flags = 0;
tools/testing/selftests/ublk/batch.c
152
t->cmd_flags |= UBLK_BATCH_F_AUTO_BUF_REG_FALLBACK;
tools/testing/selftests/ublk/batch.c
154
t->cmd_flags |= UBLK_BATCH_F_HAS_BUF_ADDR;
tools/testing/selftests/ublk/batch.c
156
t->state |= UBLKS_T_BATCH_IO;
tools/testing/selftests/ublk/batch.c
159
__func__, t->idx,
tools/testing/selftests/ublk/batch.c
16
idx = buf_idx - t->commit_buf_start;
tools/testing/selftests/ublk/batch.c
160
t->nr_commit_buf, t->commit_buf_size,
tools/testing/selftests/ublk/batch.c
161
t->nr_bufs);
tools/testing/selftests/ublk/batch.c
164
static void free_batch_fetch_buf(struct ublk_thread *t)
tools/testing/selftests/ublk/batch.c
168
for (i = 0; i < t->nr_fetch_bufs; i++) {
tools/testing/selftests/ublk/batch.c
169
io_uring_free_buf_ring(&t->ring, t->fetch[i].br, 1, i);
tools/testing/selftests/ublk/batch.c
17
return t->commit_buf + idx * t->commit_buf_size;
tools/testing/selftests/ublk/batch.c
170
munlock(t->fetch[i].fetch_buf, t->fetch[i].fetch_buf_size);
tools/testing/selftests/ublk/batch.c
171
free(t->fetch[i].fetch_buf);
tools/testing/selftests/ublk/batch.c
173
free(t->fetch);
tools/testing/selftests/ublk/batch.c
176
static int alloc_batch_fetch_buf(struct ublk_thread *t)
tools/testing/selftests/ublk/batch.c
180
unsigned buf_size = round_up(t->dev->dev_info.queue_depth * 2, pg_sz);
tools/testing/selftests/ublk/batch.c
185
t->nr_fetch_bufs = t->nr_queues * 2;
tools/testing/selftests/ublk/batch.c
186
t->fetch = calloc(t->nr_fetch_bufs, sizeof(*t->fetch));
tools/testing/selftests/ublk/batch.c
189
for (i = 0; i < t->nr_fetch_bufs; i++) {
tools/testing/selftests/ublk/batch.c
190
t->fetch[i].fetch_buf_size = buf_size;
tools/testing/selftests/ublk/batch.c
192
if (posix_memalign((void **)&t->fetch[i].fetch_buf, pg_sz,
tools/testing/selftests/ublk/batch.c
193
t->fetch[i].fetch_buf_size))
tools/testing/selftests/ublk/batch.c
197
if (mlock(t->fetch[i].fetch_buf, t->fetch[i].fetch_buf_size))
tools/testing/selftests/ublk/batch.c
200
t->fetch[i].br = io_uring_setup_buf_ring(&t->ring, 1,
tools/testing/selftests/ublk/batch.c
202
if (!t->fetch[i].br) {
tools/testing/selftests/ublk/batch.c
211
int ublk_batch_alloc_buf(struct ublk_thread *t)
tools/testing/selftests/ublk/batch.c
215
ublk_assert(t->nr_commit_buf < 2 * UBLK_MAX_QUEUES);
tools/testing/selftests/ublk/batch.c
217
ret = alloc_batch_commit_buf(t);
tools/testing/selftests/ublk/batch.c
220
return alloc_batch_fetch_buf(t);
tools/testing/selftests/ublk/batch.c
223
void ublk_batch_free_buf(struct ublk_thread *t)
tools/testing/selftests/ublk/batch.c
225
free_batch_commit_buf(t);
tools/testing/selftests/ublk/batch.c
226
free_batch_fetch_buf(t);
tools/testing/selftests/ublk/batch.c
229
static void ublk_init_batch_cmd(struct ublk_thread *t, __u16 q_id,
tools/testing/selftests/ublk/batch.c
25
static inline unsigned short ublk_alloc_commit_buf(struct ublk_thread *t)
tools/testing/selftests/ublk/batch.c
255
t->cmd_inflight += 1;
tools/testing/selftests/ublk/batch.c
260
__func__, t->idx, q_id, op, user_data,
tools/testing/selftests/ublk/batch.c
262
nr_elem * elem_bytes, buf_idx, t->cmd_inflight);
tools/testing/selftests/ublk/batch.c
265
static void ublk_setup_commit_sqe(struct ublk_thread *t,
tools/testing/selftests/ublk/batch.c
27
int idx = allocator_get(&t->commit_buf_alloc);
tools/testing/selftests/ublk/batch.c
274
cmd->flags |= t->cmd_flags;
tools/testing/selftests/ublk/batch.c
277
static void ublk_batch_queue_fetch(struct ublk_thread *t,
tools/testing/selftests/ublk/batch.c
281
unsigned short nr_elem = t->fetch[buf_idx].fetch_buf_size / 2;
tools/testing/selftests/ublk/batch.c
284
io_uring_buf_ring_add(t->fetch[buf_idx].br, t->fetch[buf_idx].fetch_buf,
tools/testing/selftests/ublk/batch.c
285
t->fetch[buf_idx].fetch_buf_size,
tools/testing/selftests/ublk/batch.c
287
io_uring_buf_ring_advance(t->fetch[buf_idx].br, 1);
tools/testing/selftests/ublk/batch.c
289
ublk_io_alloc_sqes(t, &sqe, 1);
tools/testing/selftests/ublk/batch.c
291
ublk_init_batch_cmd(t, q->q_id, sqe, UBLK_U_IO_FETCH_IO_CMDS, 2, nr_elem,
tools/testing/selftests/ublk/batch.c
298
t->fetch[buf_idx].fetch_buf_off = 0;
tools/testing/selftests/ublk/batch.c
30
return idx + t->commit_buf_start;
tools/testing/selftests/ublk/batch.c
301
void ublk_batch_start_fetch(struct ublk_thread *t)
tools/testing/selftests/ublk/batch.c
306
for (i = 0; i < t->dev->dev_info.nr_hw_queues; i++) {
tools/testing/selftests/ublk/batch.c
307
if (t->q_map[i]) {
tools/testing/selftests/ublk/batch.c
308
struct ublk_queue *q = &t->dev->q[i];
tools/testing/selftests/ublk/batch.c
311
ublk_batch_queue_fetch(t, q, j++);
tools/testing/selftests/ublk/batch.c
312
ublk_batch_queue_fetch(t, q, j++);
tools/testing/selftests/ublk/batch.c
317
static unsigned short ublk_compl_batch_fetch(struct ublk_thread *t,
tools/testing/selftests/ublk/batch.c
322
unsigned start = t->fetch[buf_idx].fetch_buf_off;
tools/testing/selftests/ublk/batch.c
324
void *buf = t->fetch[buf_idx].fetch_buf;
tools/testing/selftests/ublk/batch.c
348
q->tgt_ops->queue_io(t, q, tag);
tools/testing/selftests/ublk/batch.c
350
t->fetch[buf_idx].fetch_buf_off = end;
tools/testing/selftests/ublk/batch.c
354
static int __ublk_batch_queue_prep_io_cmds(struct ublk_thread *t, struct ublk_queue *q)
tools/testing/selftests/ublk/batch.c
357
unsigned short buf_idx = ublk_alloc_commit_buf(t);
tools/testing/selftests/ublk/batch.c
364
ublk_io_alloc_sqes(t, &sqe, 1);
tools/testing/selftests/ublk/batch.c
367
buf = ublk_get_commit_buf(t, buf_idx);
tools/testing/selftests/ublk/batch.c
370
buf + i * t->commit_buf_elem_size);
tools/testing/selftests/ublk/batch.c
377
elem->buf_index = ublk_batch_io_buf_idx(t, q, i);
tools/testing/selftests/ublk/batch.c
38
static inline void ublk_free_commit_buf(struct ublk_thread *t,
tools/testing/selftests/ublk/batch.c
383
sqe->len = t->commit_buf_elem_size * nr_elem;
tools/testing/selftests/ublk/batch.c
385
ublk_init_batch_cmd(t, q->q_id, sqe, UBLK_U_IO_PREP_IO_CMDS,
tools/testing/selftests/ublk/batch.c
386
t->commit_buf_elem_size, nr_elem, buf_idx);
tools/testing/selftests/ublk/batch.c
387
ublk_setup_commit_sqe(t, sqe, buf_idx);
tools/testing/selftests/ublk/batch.c
391
int ublk_batch_queue_prep_io_cmds(struct ublk_thread *t, struct ublk_queue *q)
tools/testing/selftests/ublk/batch.c
398
ret = __ublk_batch_queue_prep_io_cmds(t, q);
tools/testing/selftests/ublk/batch.c
407
static void ublk_batch_compl_commit_cmd(struct ublk_thread *t,
tools/testing/selftests/ublk/batch.c
41
unsigned short idx = i - t->commit_buf_start;
tools/testing/selftests/ublk/batch.c
418
ublk_assert(cqe->res == t->commit_buf_elem_size * nr_elem);
tools/testing/selftests/ublk/batch.c
422
ublk_free_commit_buf(t, buf_idx);
tools/testing/selftests/ublk/batch.c
425
void ublk_batch_compl_cmd(struct ublk_thread *t,
tools/testing/selftests/ublk/batch.c
43
ublk_assert(idx < t->nr_commit_buf);
tools/testing/selftests/ublk/batch.c
435
t->cmd_inflight--;
tools/testing/selftests/ublk/batch.c
436
ublk_batch_compl_commit_cmd(t, cqe, op);
tools/testing/selftests/ublk/batch.c
44
ublk_assert(allocator_get_val(&t->commit_buf_alloc, idx) != 0);
tools/testing/selftests/ublk/batch.c
442
q = &t->dev->q[q_id];
tools/testing/selftests/ublk/batch.c
443
buf_idx = ublk_compl_batch_fetch(t, q, cqe);
tools/testing/selftests/ublk/batch.c
446
t->cmd_inflight--;
tools/testing/selftests/ublk/batch.c
447
t->state |= UBLKS_T_STOPPING;
tools/testing/selftests/ublk/batch.c
449
t->cmd_inflight--;
tools/testing/selftests/ublk/batch.c
450
ublk_batch_queue_fetch(t, q, buf_idx);
tools/testing/selftests/ublk/batch.c
454
static void __ublk_batch_commit_io_cmds(struct ublk_thread *t,
tools/testing/selftests/ublk/batch.c
46
allocator_put(&t->commit_buf_alloc, idx);
tools/testing/selftests/ublk/batch.c
463
ublk_free_commit_buf(t, cb->buf_idx);
tools/testing/selftests/ublk/batch.c
467
ublk_io_alloc_sqes(t, &sqe, 1);
tools/testing/selftests/ublk/batch.c
470
sqe->len = nr_elem * t->commit_buf_elem_size;
tools/testing/selftests/ublk/batch.c
473
ublk_init_batch_cmd(t, cb->q_id, sqe, UBLK_U_IO_COMMIT_IO_CMDS,
tools/testing/selftests/ublk/batch.c
474
t->commit_buf_elem_size, nr_elem, buf_idx);
tools/testing/selftests/ublk/batch.c
475
ublk_setup_commit_sqe(t, sqe, buf_idx);
tools/testing/selftests/ublk/batch.c
478
void ublk_batch_commit_io_cmds(struct ublk_thread *t)
tools/testing/selftests/ublk/batch.c
482
for (i = 0; i < t->nr_queues; i++) {
tools/testing/selftests/ublk/batch.c
483
struct batch_commit_buf *cb = &t->commit[i];
tools/testing/selftests/ublk/batch.c
486
__ublk_batch_commit_io_cmds(t, cb);
tools/testing/selftests/ublk/batch.c
491
static void __ublk_batch_init_commit(struct ublk_thread *t,
tools/testing/selftests/ublk/batch.c
497
cb->elem = ublk_get_commit_buf(t, buf_idx);
tools/testing/selftests/ublk/batch.c
499
cb->count = t->commit_buf_size /
tools/testing/selftests/ublk/batch.c
500
t->commit_buf_elem_size;
tools/testing/selftests/ublk/batch.c
504
static void ublk_batch_init_commit(struct ublk_thread *t,
tools/testing/selftests/ublk/batch.c
507
unsigned short buf_idx = ublk_alloc_commit_buf(t);
tools/testing/selftests/ublk/batch.c
512
__ublk_batch_init_commit(t, cb, buf_idx);
tools/testing/selftests/ublk/batch.c
515
void ublk_batch_prep_commit(struct ublk_thread *t)
tools/testing/selftests/ublk/batch.c
519
for (i = 0; i < t->nr_queues; i++)
tools/testing/selftests/ublk/batch.c
520
t->commit[i].buf_idx = UBLKS_T_COMMIT_BUF_INV_IDX;
tools/testing/selftests/ublk/batch.c
523
void ublk_batch_complete_io(struct ublk_thread *t, struct ublk_queue *q,
tools/testing/selftests/ublk/batch.c
526
unsigned q_t_idx = ublk_queue_idx_in_thread(t, q);
tools/testing/selftests/ublk/batch.c
527
struct batch_commit_buf *cb = &t->commit[q_t_idx];
tools/testing/selftests/ublk/batch.c
532
ublk_batch_init_commit(t, cb);
tools/testing/selftests/ublk/batch.c
536
elem = (struct ublk_batch_elem *)(cb->elem + cb->done * t->commit_buf_elem_size);
tools/testing/selftests/ublk/batch.c
538
elem->buf_index = ublk_batch_io_buf_idx(t, q, tag);
tools/testing/selftests/ublk/batch.c
59
static unsigned ublk_commit_buf_size(struct ublk_thread *t)
tools/testing/selftests/ublk/batch.c
61
struct ublk_dev *dev = t->dev;
tools/testing/selftests/ublk/batch.c
69
static void free_batch_commit_buf(struct ublk_thread *t)
tools/testing/selftests/ublk/batch.c
71
if (t->commit_buf) {
tools/testing/selftests/ublk/batch.c
72
unsigned buf_size = ublk_commit_buf_size(t);
tools/testing/selftests/ublk/batch.c
73
unsigned int total = buf_size * t->nr_commit_buf;
tools/testing/selftests/ublk/batch.c
75
munlock(t->commit_buf, total);
tools/testing/selftests/ublk/batch.c
76
free(t->commit_buf);
tools/testing/selftests/ublk/batch.c
78
allocator_deinit(&t->commit_buf_alloc);
tools/testing/selftests/ublk/batch.c
79
free(t->commit);
tools/testing/selftests/ublk/batch.c
8
static inline void *ublk_get_commit_buf(struct ublk_thread *t,
tools/testing/selftests/ublk/batch.c
82
static int alloc_batch_commit_buf(struct ublk_thread *t)
tools/testing/selftests/ublk/batch.c
84
unsigned buf_size = ublk_commit_buf_size(t);
tools/testing/selftests/ublk/batch.c
85
unsigned int total = buf_size * t->nr_commit_buf;
tools/testing/selftests/ublk/batch.c
90
t->commit = calloc(t->nr_queues, sizeof(*t->commit));
tools/testing/selftests/ublk/batch.c
91
for (i = 0; i < t->dev->dev_info.nr_hw_queues; i++) {
tools/testing/selftests/ublk/batch.c
92
if (t->q_map[i])
tools/testing/selftests/ublk/batch.c
93
t->commit[j++].q_id = i;
tools/testing/selftests/ublk/batch.c
96
allocator_init(&t->commit_buf_alloc, t->nr_commit_buf);
tools/testing/selftests/ublk/batch.c
98
t->commit_buf = NULL;
tools/testing/selftests/ublk/fault_inject.c
42
static int ublk_fault_inject_queue_io(struct ublk_thread *t,
tools/testing/selftests/ublk/fault_inject.c
51
ublk_io_alloc_sqes(t, &sqe, 1);
tools/testing/selftests/ublk/fault_inject.c
55
ublk_queued_tgt_io(t, q, tag, 1);
tools/testing/selftests/ublk/fault_inject.c
60
static void ublk_fault_inject_tgt_io_done(struct ublk_thread *t,
tools/testing/selftests/ublk/fault_inject.c
70
if (ublk_completed_tgt_io(t, q, tag))
tools/testing/selftests/ublk/fault_inject.c
71
ublk_complete_io(t, q, tag, iod->nr_sectors << 9);
tools/testing/selftests/ublk/file_backed.c
101
ret = loop_queue_flush_io(t, q, iod, tag);
tools/testing/selftests/ublk/file_backed.c
109
ret = loop_queue_tgt_rw_io(t, q, iod, tag);
tools/testing/selftests/ublk/file_backed.c
121
static int ublk_loop_queue_io(struct ublk_thread *t, struct ublk_queue *q,
tools/testing/selftests/ublk/file_backed.c
124
int queued = loop_queue_tgt_io(t, q, tag);
tools/testing/selftests/ublk/file_backed.c
126
ublk_queued_tgt_io(t, q, tag, queued);
tools/testing/selftests/ublk/file_backed.c
130
static void ublk_loop_io_done(struct ublk_thread *t, struct ublk_queue *q,
tools/testing/selftests/ublk/file_backed.c
154
if (ublk_completed_tgt_io(t, q, tag))
tools/testing/selftests/ublk/file_backed.c
155
ublk_complete_io(t, q, tag, io->result);
tools/testing/selftests/ublk/file_backed.c
16
static int loop_queue_flush_io(struct ublk_thread *t, struct ublk_queue *q,
tools/testing/selftests/ublk/file_backed.c
22
ublk_io_alloc_sqes(t, sqe, 1);
tools/testing/selftests/ublk/file_backed.c
30
static int loop_queue_tgt_rw_io(struct ublk_thread *t, struct ublk_queue *q,
tools/testing/selftests/ublk/file_backed.c
42
unsigned short buf_index = ublk_io_buf_idx(t, q, tag);
tools/testing/selftests/ublk/file_backed.c
45
ublk_io_alloc_sqes(t, sqe, 1);
tools/testing/selftests/ublk/file_backed.c
57
ublk_io_alloc_sqes(t, sqe, 1);
tools/testing/selftests/ublk/file_backed.c
73
ublk_io_alloc_sqes(t, sqe, 3);
tools/testing/selftests/ublk/file_backed.c
93
static int loop_queue_tgt_io(struct ublk_thread *t, struct ublk_queue *q, int tag)
tools/testing/selftests/ublk/kublk.c
1002
memcpy(t.q_map, info->q_thread_map[info->idx], sizeof(t.q_map));
tools/testing/selftests/ublk/kublk.c
1004
ret = ublk_thread_init(&t, info->extra_flags);
tools/testing/selftests/ublk/kublk.c
1007
dev_id, t.idx);
tools/testing/selftests/ublk/kublk.c
1013
gettid(), dev_id, t.idx);
tools/testing/selftests/ublk/kublk.c
1015
if (!ublk_thread_batch_io(&t)) {
tools/testing/selftests/ublk/kublk.c
1017
ublk_submit_fetch_commands(&t);
tools/testing/selftests/ublk/kublk.c
1019
ublk_batch_setup_queues(&t);
tools/testing/selftests/ublk/kublk.c
1020
ublk_batch_start_fetch(&t);
tools/testing/selftests/ublk/kublk.c
1024
if (ublk_process_io(&t) < 0)
tools/testing/selftests/ublk/kublk.c
1029
gettid(), dev_id, t.idx);
tools/testing/selftests/ublk/kublk.c
1030
ublk_thread_deinit(&t);
tools/testing/selftests/ublk/kublk.c
434
static void ublk_thread_deinit(struct ublk_thread *t)
tools/testing/selftests/ublk/kublk.c
436
io_uring_unregister_buffers(&t->ring);
tools/testing/selftests/ublk/kublk.c
438
ublk_batch_free_buf(t);
tools/testing/selftests/ublk/kublk.c
440
io_uring_unregister_ring_fd(&t->ring);
tools/testing/selftests/ublk/kublk.c
442
if (t->ring.ring_fd > 0) {
tools/testing/selftests/ublk/kublk.c
443
io_uring_unregister_files(&t->ring);
tools/testing/selftests/ublk/kublk.c
444
close(t->ring.ring_fd);
tools/testing/selftests/ublk/kublk.c
445
t->ring.ring_fd = -1;
tools/testing/selftests/ublk/kublk.c
516
static int ublk_thread_init(struct ublk_thread *t, unsigned long long extra_flags)
tools/testing/selftests/ublk/kublk.c
518
struct ublk_dev *dev = t->dev;
tools/testing/selftests/ublk/kublk.c
527
ret = ublk_setup_ring(&t->ring, ring_depth, cq_depth,
tools/testing/selftests/ublk/kublk.c
533
dev->dev_info.dev_id, t->idx, ret);
tools/testing/selftests/ublk/kublk.c
542
t->nr_bufs = max_nr_ios_per_thread;
tools/testing/selftests/ublk/kublk.c
544
t->nr_bufs = 0;
tools/testing/selftests/ublk/kublk.c
548
ublk_batch_prepare(t);
tools/testing/selftests/ublk/kublk.c
550
if (t->nr_bufs) {
tools/testing/selftests/ublk/kublk.c
551
ret = io_uring_register_buffers_sparse(&t->ring, t->nr_bufs);
tools/testing/selftests/ublk/kublk.c
554
dev->dev_info.dev_id, t->idx, ret);
tools/testing/selftests/ublk/kublk.c
560
ret = ublk_batch_alloc_buf(t);
tools/testing/selftests/ublk/kublk.c
563
dev->dev_info.dev_id, t->idx, ret);
tools/testing/selftests/ublk/kublk.c
568
io_uring_register_ring_fd(&t->ring);
tools/testing/selftests/ublk/kublk.c
573
ret = io_uring_register_files(&t->ring, &dev->fds[1], dev->nr_fds - 1);
tools/testing/selftests/ublk/kublk.c
579
ret = io_uring_register_files(&t->ring, dev->fds, dev->nr_fds);
tools/testing/selftests/ublk/kublk.c
583
t->dev->dev_info.dev_id, t->idx, ret);
tools/testing/selftests/ublk/kublk.c
589
ublk_thread_deinit(t);
tools/testing/selftests/ublk/kublk.c
591
dev->dev_info.dev_id, t->idx);
tools/testing/selftests/ublk/kublk.c
633
static void ublk_set_auto_buf_reg(const struct ublk_thread *t,
tools/testing/selftests/ublk/kublk.c
641
buf.index = q->tgt_ops->buf_index(t, q, tag);
tools/testing/selftests/ublk/kublk.c
643
buf.index = ublk_io_buf_idx(t, q, tag);
tools/testing/selftests/ublk/kublk.c
697
int ublk_queue_io_cmd(struct ublk_thread *t, struct ublk_io *io)
tools/testing/selftests/ublk/kublk.c
727
if (io_uring_sq_space_left(&t->ring) < 1)
tools/testing/selftests/ublk/kublk.c
728
io_uring_submit(&t->ring);
tools/testing/selftests/ublk/kublk.c
730
ublk_io_alloc_sqes(t, sqe, 1);
tools/testing/selftests/ublk/kublk.c
733
__func__, t->idx, io->tag);
tools/testing/selftests/ublk/kublk.c
759
ublk_set_auto_buf_reg(t, q, sqe[0], io->tag);
tools/testing/selftests/ublk/kublk.c
766
t->cmd_inflight += 1;
tools/testing/selftests/ublk/kublk.c
769
__func__, t->idx, q->q_id, io->tag, cmd_op,
tools/testing/selftests/ublk/kublk.c
770
io->flags, !!(t->state & UBLKS_T_STOPPING));
tools/testing/selftests/ublk/kublk.c
774
static void ublk_submit_fetch_commands(struct ublk_thread *t)
tools/testing/selftests/ublk/kublk.c
780
if (t->dev->per_io_tasks) {
tools/testing/selftests/ublk/kublk.c
791
const struct ublksrv_ctrl_dev_info *dinfo = &t->dev->dev_info;
tools/testing/selftests/ublk/kublk.c
793
for (i = t->idx; i < nr_ios; i += t->dev->nthreads) {
tools/testing/selftests/ublk/kublk.c
796
q = &t->dev->q[q_id];
tools/testing/selftests/ublk/kublk.c
799
ublk_queue_io_cmd(t, io);
tools/testing/selftests/ublk/kublk.c
806
struct ublk_queue *q = &t->dev->q[t->idx];
tools/testing/selftests/ublk/kublk.c
810
ublk_queue_io_cmd(t, io);
tools/testing/selftests/ublk/kublk.c
815
static int ublk_thread_is_idle(struct ublk_thread *t)
tools/testing/selftests/ublk/kublk.c
817
return !io_uring_sq_ready(&t->ring) && !t->io_inflight;
tools/testing/selftests/ublk/kublk.c
820
static int ublk_thread_is_done(struct ublk_thread *t)
tools/testing/selftests/ublk/kublk.c
822
return (t->state & UBLKS_T_STOPPING) && ublk_thread_is_idle(t) && !t->cmd_inflight;
tools/testing/selftests/ublk/kublk.c
825
static inline void ublksrv_handle_tgt_cqe(struct ublk_thread *t,
tools/testing/selftests/ublk/kublk.c
836
q->tgt_ops->tgt_io_done(t, q, cqe);
tools/testing/selftests/ublk/kublk.c
839
static void ublk_handle_uring_cmd(struct ublk_thread *t,
tools/testing/selftests/ublk/kublk.c
844
!(t->state & UBLKS_T_STOPPING);
tools/testing/selftests/ublk/kublk.c
848
t->cmd_inflight--;
tools/testing/selftests/ublk/kublk.c
851
t->state |= UBLKS_T_STOPPING;
tools/testing/selftests/ublk/kublk.c
862
q->tgt_ops->queue_io(t, q, tag);
tools/testing/selftests/ublk/kublk.c
865
ublk_queue_io_cmd(t, io);
tools/testing/selftests/ublk/kublk.c
879
static void ublk_handle_cqe(struct ublk_thread *t,
tools/testing/selftests/ublk/kublk.c
882
struct ublk_dev *dev = t->dev;
tools/testing/selftests/ublk/kublk.c
888
cqe->res, cqe->user_data, t->state);
tools/testing/selftests/ublk/kublk.c
892
__func__, cqe->res, t->idx, q_id,
tools/testing/selftests/ublk/kublk.c
896
(t->state & UBLKS_T_STOPPING));
tools/testing/selftests/ublk/kublk.c
900
ublksrv_handle_tgt_cqe(t, &dev->q[q_id], cqe);
tools/testing/selftests/ublk/kublk.c
904
if (ublk_thread_batch_io(t))
tools/testing/selftests/ublk/kublk.c
905
ublk_batch_compl_cmd(t, cqe);
tools/testing/selftests/ublk/kublk.c
907
ublk_handle_uring_cmd(t, &dev->q[q_id], cqe);
tools/testing/selftests/ublk/kublk.c
910
static int ublk_reap_events_uring(struct ublk_thread *t)
tools/testing/selftests/ublk/kublk.c
916
io_uring_for_each_cqe(&t->ring, head, cqe) {
tools/testing/selftests/ublk/kublk.c
917
ublk_handle_cqe(t, cqe, NULL);
tools/testing/selftests/ublk/kublk.c
920
io_uring_cq_advance(&t->ring, count);
tools/testing/selftests/ublk/kublk.c
925
static int ublk_process_io(struct ublk_thread *t)
tools/testing/selftests/ublk/kublk.c
930
t->dev->dev_info.dev_id,
tools/testing/selftests/ublk/kublk.c
931
t->idx, io_uring_sq_ready(&t->ring),
tools/testing/selftests/ublk/kublk.c
932
t->cmd_inflight,
tools/testing/selftests/ublk/kublk.c
933
(t->state & UBLKS_T_STOPPING));
tools/testing/selftests/ublk/kublk.c
935
if (ublk_thread_is_done(t))
tools/testing/selftests/ublk/kublk.c
938
ret = io_uring_submit_and_wait(&t->ring, 1);
tools/testing/selftests/ublk/kublk.c
939
if (ublk_thread_batch_io(t)) {
tools/testing/selftests/ublk/kublk.c
940
ublk_batch_prep_commit(t);
tools/testing/selftests/ublk/kublk.c
941
reapped = ublk_reap_events_uring(t);
tools/testing/selftests/ublk/kublk.c
942
ublk_batch_commit_io_cmds(t);
tools/testing/selftests/ublk/kublk.c
944
reapped = ublk_reap_events_uring(t);
tools/testing/selftests/ublk/kublk.c
948
ret, reapped, (t->state & UBLKS_T_STOPPING),
tools/testing/selftests/ublk/kublk.c
949
(t->state & UBLKS_T_IDLE));
tools/testing/selftests/ublk/kublk.c
971
static void ublk_batch_setup_queues(struct ublk_thread *t)
tools/testing/selftests/ublk/kublk.c
975
for (i = 0; i < t->dev->dev_info.nr_hw_queues; i++) {
tools/testing/selftests/ublk/kublk.c
976
struct ublk_queue *q = &t->dev->q[i];
tools/testing/selftests/ublk/kublk.c
983
if (t->q_map[i] == 0)
tools/testing/selftests/ublk/kublk.c
986
ret = ublk_batch_queue_prep_io_cmds(t, q);
tools/testing/selftests/ublk/kublk.c
993
struct ublk_thread t = {
tools/testing/selftests/ublk/kublk.h
154
unsigned short (*buf_index)(const struct ublk_thread *t,
tools/testing/selftests/ublk/kublk.h
265
extern int ublk_queue_io_cmd(struct ublk_thread *t, struct ublk_io *io);
tools/testing/selftests/ublk/kublk.h
283
static inline int ublk_thread_batch_io(const struct ublk_thread *t)
tools/testing/selftests/ublk/kublk.h
285
return t->state & UBLKS_T_BATCH_IO;
tools/testing/selftests/ublk/kublk.h
374
static inline int ublk_io_alloc_sqes(struct ublk_thread *t,
tools/testing/selftests/ublk/kublk.h
377
struct io_uring *ring = &t->ring;
tools/testing/selftests/ublk/kublk.h
472
const struct ublk_thread *t, const struct ublk_queue *q,
tools/testing/selftests/ublk/kublk.h
475
static inline unsigned short ublk_io_buf_idx(const struct ublk_thread *t,
tools/testing/selftests/ublk/kublk.h
480
return ublk_batch_io_buf_idx(t, q, tag);
tools/testing/selftests/ublk/kublk.h
489
static inline int ublk_completed_tgt_io(struct ublk_thread *t,
tools/testing/selftests/ublk/kublk.h
494
t->io_inflight--;
tools/testing/selftests/ublk/kublk.h
529
static inline unsigned ublk_queue_idx_in_thread(const struct ublk_thread *t,
tools/testing/selftests/ublk/kublk.h
534
idx = t->q_map[q->q_id];
tools/testing/selftests/ublk/kublk.h
544
const struct ublk_thread *t, const struct ublk_queue *q,
tools/testing/selftests/ublk/kublk.h
547
return ublk_queue_idx_in_thread(t, q) * q->q_depth + tag;
tools/testing/selftests/ublk/kublk.h
551
int ublk_batch_queue_prep_io_cmds(struct ublk_thread *t, struct ublk_queue *q);
tools/testing/selftests/ublk/kublk.h
553
void ublk_batch_start_fetch(struct ublk_thread *t);
tools/testing/selftests/ublk/kublk.h
555
void ublk_batch_compl_cmd(struct ublk_thread *t,
tools/testing/selftests/ublk/kublk.h
558
void ublk_batch_prepare(struct ublk_thread *t);
tools/testing/selftests/ublk/kublk.h
560
int ublk_batch_alloc_buf(struct ublk_thread *t);
tools/testing/selftests/ublk/kublk.h
562
void ublk_batch_free_buf(struct ublk_thread *t);
tools/testing/selftests/ublk/kublk.h
565
void ublk_batch_prep_commit(struct ublk_thread *t);
tools/testing/selftests/ublk/kublk.h
567
void ublk_batch_commit_io_cmds(struct ublk_thread *t);
tools/testing/selftests/ublk/kublk.h
569
void ublk_batch_complete_io(struct ublk_thread *t, struct ublk_queue *q,
tools/testing/selftests/ublk/kublk.h
574
static inline int ublk_complete_io(struct ublk_thread *t, struct ublk_queue *q,
tools/testing/selftests/ublk/kublk.h
578
ublk_batch_complete_io(t, q, tag, res);
tools/testing/selftests/ublk/kublk.h
584
return ublk_queue_io_cmd(t, io);
tools/testing/selftests/ublk/kublk.h
588
static inline void ublk_queued_tgt_io(struct ublk_thread *t, struct ublk_queue *q,
tools/testing/selftests/ublk/kublk.h
592
ublk_complete_io(t, q, tag, queued);
tools/testing/selftests/ublk/kublk.h
596
t->io_inflight += queued;
tools/testing/selftests/ublk/null.c
113
if (ublk_completed_tgt_io(t, q, tag))
tools/testing/selftests/ublk/null.c
114
ublk_complete_io(t, q, tag, io->result);
tools/testing/selftests/ublk/null.c
117
static int ublk_null_queue_io(struct ublk_thread *t, struct ublk_queue *q,
tools/testing/selftests/ublk/null.c
126
queued = null_queue_auto_zc_io(t, q, tag);
tools/testing/selftests/ublk/null.c
128
queued = null_queue_zc_io(t, q, tag);
tools/testing/selftests/ublk/null.c
130
ublk_complete_io(t, q, tag, iod->nr_sectors << 9);
tools/testing/selftests/ublk/null.c
133
ublk_queued_tgt_io(t, q, tag, queued);
tools/testing/selftests/ublk/null.c
141
static unsigned short ublk_null_buf_index(const struct ublk_thread *t,
tools/testing/selftests/ublk/null.c
146
return ublk_io_buf_idx(t, q, tag);
tools/testing/selftests/ublk/null.c
59
static int null_queue_zc_io(struct ublk_thread *t, struct ublk_queue *q,
tools/testing/selftests/ublk/null.c
64
unsigned short buf_idx = ublk_io_buf_idx(t, q, tag);
tools/testing/selftests/ublk/null.c
66
ublk_io_alloc_sqes(t, sqe, 3);
tools/testing/selftests/ublk/null.c
83
static int null_queue_auto_zc_io(struct ublk_thread *t, struct ublk_queue *q,
tools/testing/selftests/ublk/null.c
89
ublk_io_alloc_sqes(t, sqe, 1);
tools/testing/selftests/ublk/null.c
90
__setup_nop_io(tag, iod, sqe[0], q->q_id, ublk_io_buf_idx(t, q, tag));
tools/testing/selftests/ublk/null.c
94
static void ublk_null_io_done(struct ublk_thread *t, struct ublk_queue *q,
tools/testing/selftests/ublk/stripe.c
126
static int stripe_queue_tgt_rw_io(struct ublk_thread *t, struct ublk_queue *q,
tools/testing/selftests/ublk/stripe.c
138
unsigned short buf_idx = ublk_io_buf_idx(t, q, tag);
tools/testing/selftests/ublk/stripe.c
143
ublk_io_alloc_sqes(t, sqe, s->nr + extra);
tools/testing/selftests/ublk/stripe.c
153
struct stripe *t = &s->s[i - zc];
tools/testing/selftests/ublk/stripe.c
156
t->seq + 1,
tools/testing/selftests/ublk/stripe.c
157
(void *)t->vec,
tools/testing/selftests/ublk/stripe.c
158
t->nr_vec,
tools/testing/selftests/ublk/stripe.c
159
t->start << 9);
tools/testing/selftests/ublk/stripe.c
181
static int handle_flush(struct ublk_thread *t, struct ublk_queue *q,
tools/testing/selftests/ublk/stripe.c
188
ublk_io_alloc_sqes(t, sqe, conf->nr_files);
tools/testing/selftests/ublk/stripe.c
197
static int stripe_queue_tgt_io(struct ublk_thread *t, struct ublk_queue *q,
tools/testing/selftests/ublk/stripe.c
206
ret = handle_flush(t, q, iod, tag);
tools/testing/selftests/ublk/stripe.c
214
ret = stripe_queue_tgt_rw_io(t, q, iod, tag);
tools/testing/selftests/ublk/stripe.c
225
static int ublk_stripe_queue_io(struct ublk_thread *t, struct ublk_queue *q,
tools/testing/selftests/ublk/stripe.c
228
int queued = stripe_queue_tgt_io(t, q, tag);
tools/testing/selftests/ublk/stripe.c
230
ublk_queued_tgt_io(t, q, tag, queued);
tools/testing/selftests/ublk/stripe.c
234
static void ublk_stripe_io_done(struct ublk_thread *t, struct ublk_queue *q,
tools/testing/selftests/ublk/stripe.c
266
if (ublk_completed_tgt_io(t, q, tag)) {
tools/testing/selftests/ublk/stripe.c
272
ublk_complete_io(t, q, tag, res);
tools/testing/selftests/ublk/stripe.c
56
struct stripe *t = &s->s[i];
tools/testing/selftests/ublk/stripe.c
58
t->nr_vec = 0;
tools/testing/selftests/ublk/stripe.c
59
t->vec = &s->_vec[i * nr_vecs];
tools/testing/selftests/ublk/stripe.c
60
t->nr_sects = 0;
tools/testing/selftests/ublk/stripe.c
61
t->cap = nr_vecs;
tools/testing/selftests/vDSO/vdso_test_abi.c
40
typedef time_t (*vdso_time_t)(time_t *t);
tools/testing/selftests/x86/lam.c
916
struct testcases *t = test + i;
tools/testing/selftests/x86/lam.c
920
ret = fork_test(t);
tools/testing/selftests/x86/lam.c
924
ksft_test_result_skip("%s", t->msg);
tools/testing/selftests/x86/lam.c
929
ret = (t->expected == ret);
tools/testing/selftests/x86/lam.c
931
ret = !(t->expected);
tools/testing/selftests/x86/lam.c
933
ksft_test_result(ret, "%s", t->msg);
tools/testing/selftests/x86/test_vsyscall.c
100
static inline long sys_time(time_t *t)
tools/testing/selftests/x86/test_vsyscall.c
102
return syscall(SYS_time, t);
tools/testing/selftests/x86/test_vsyscall.c
59
typedef long (*time_func_t)(time_t *t);
tools/testing/vsock/vsock_perf.c
220
time_t t;
tools/testing/vsock/vsock_perf.c
222
t = current_nsec();
tools/testing/vsock/vsock_perf.c
224
in_read_ns += (current_nsec() - t);
tools/tracing/rtla/src/trace.c
41
enum tracefs_tracers t = TRACEFS_TRACER_NOP;
tools/tracing/rtla/src/trace.c
44
retval = tracefs_tracer_set(inst, t);
tools/tracing/rtla/src/utils.c
173
long t;
tools/tracing/rtla/src/utils.c
175
t = strtol(val, &end, 10);
tools/tracing/rtla/src/utils.c
184
t *= 60;
tools/tracing/rtla/src/utils.c
188
t *= 60 * 60;
tools/tracing/rtla/src/utils.c
193
t *= 24 * 60 * 60;
tools/tracing/rtla/src/utils.c
198
return t;
tools/tracing/rtla/src/utils.c
207
long t;
tools/tracing/rtla/src/utils.c
209
t = strtol(val, &end, 10);
tools/tracing/rtla/src/utils.c
213
return t;
tools/tracing/rtla/src/utils.c
215
t *= 1000;
tools/tracing/rtla/src/utils.c
216
return t;
tools/tracing/rtla/src/utils.c
218
t *= 1000 * 1000;
tools/tracing/rtla/src/utils.c
219
return t;
tools/tracing/rtla/src/utils.c
221
t *= 1000 * 1000 * 1000;
tools/tracing/rtla/src/utils.c
222
return t;
tools/tracing/rtla/src/utils.c
227
return t;
tools/usb/ffs-test.c
336
static ssize_t read_wrap(struct thread *t, void *buf, size_t nbytes);
tools/usb/ffs-test.c
337
static ssize_t write_wrap(struct thread *t, const void *buf, size_t nbytes);
tools/usb/ffs-test.c
338
static ssize_t ep0_consume(struct thread *t, const void *buf, size_t nbytes);
tools/usb/ffs-test.c
339
static ssize_t fill_in_buf(struct thread *t, void *buf, size_t nbytes);
tools/usb/ffs-test.c
340
static ssize_t empty_out_buf(struct thread *t, const void *buf, size_t nbytes);
tools/usb/ffs-test.c
379
static void init_thread(struct thread *t)
tools/usb/ffs-test.c
381
t->buf = malloc(t->buf_size);
tools/usb/ffs-test.c
382
die_on(!t->buf, "malloc");
tools/usb/ffs-test.c
384
t->fd = open(t->filename, O_RDWR);
tools/usb/ffs-test.c
385
die_on(t->fd < 0, "%s", t->filename);
tools/usb/ffs-test.c
390
struct thread *t = arg;
tools/usb/ffs-test.c
393
fd = t->fd;
tools/usb/ffs-test.c
394
if (t->fd < 0)
tools/usb/ffs-test.c
396
t->fd = -1;
tools/usb/ffs-test.c
399
if (t != threads) {
tools/usb/ffs-test.c
404
err("%s: get fifo status", t->filename);
tools/usb/ffs-test.c
406
warn("%s: unclaimed = %d\n", t->filename, ret);
tools/usb/ffs-test.c
408
err("%s: fifo flush", t->filename);
tools/usb/ffs-test.c
413
err("%s: close", t->filename);
tools/usb/ffs-test.c
415
free(t->buf);
tools/usb/ffs-test.c
416
t->buf = NULL;
tools/usb/ffs-test.c
422
struct thread *t = arg;
tools/usb/ffs-test.c
425
info("%s: starts\n", t->filename);
tools/usb/ffs-test.c
426
in_name = t->in_name ? t->in_name : t->filename;
tools/usb/ffs-test.c
427
out_name = t->out_name ? t->out_name : t->filename;
tools/usb/ffs-test.c
434
ret = t->in(t, t->buf, t->buf_size);
tools/usb/ffs-test.c
436
ret = t->out(t, t->buf, ret);
tools/usb/ffs-test.c
459
t->status = ret;
tools/usb/ffs-test.c
460
info("%s: ends\n", t->filename);
tools/usb/ffs-test.c
464
static void start_thread(struct thread *t)
tools/usb/ffs-test.c
466
debug("%s: starting\n", t->filename);
tools/usb/ffs-test.c
468
die_on(pthread_create(&t->id, NULL, start_thread_helper, t) < 0,
tools/usb/ffs-test.c
469
"pthread_create(%s)", t->filename);
tools/usb/ffs-test.c
472
static void join_thread(struct thread *t)
tools/usb/ffs-test.c
474
int ret = pthread_join(t->id, NULL);
tools/usb/ffs-test.c
477
err("%s: joining thread", t->filename);
tools/usb/ffs-test.c
479
debug("%s: joined\n", t->filename);
tools/usb/ffs-test.c
483
static ssize_t read_wrap(struct thread *t, void *buf, size_t nbytes)
tools/usb/ffs-test.c
485
return read(t->fd, buf, nbytes);
tools/usb/ffs-test.c
488
static ssize_t write_wrap(struct thread *t, const void *buf, size_t nbytes)
tools/usb/ffs-test.c
490
return write(t->fd, buf, nbytes);
tools/usb/ffs-test.c
626
static void ep0_init(struct thread *t, bool legacy_descriptors)
tools/usb/ffs-test.c
633
info("%s: writing descriptors\n", t->filename);
tools/usb/ffs-test.c
637
info("%s: writing descriptors (in v2 format)\n", t->filename);
tools/usb/ffs-test.c
638
ret = write(t->fd, &descriptors, sizeof descriptors);
tools/usb/ffs-test.c
641
warn("%s: new format rejected, trying legacy\n", t->filename);
tools/usb/ffs-test.c
645
ret = write(t->fd, legacy, len);
tools/usb/ffs-test.c
649
die_on(ret < 0, "%s: write: descriptors", t->filename);
tools/usb/ffs-test.c
651
info("%s: writing strings\n", t->filename);
tools/usb/ffs-test.c
652
ret = write(t->fd, &strings, sizeof strings);
tools/usb/ffs-test.c
653
die_on(ret < 0, "%s: write: strings", t->filename);
tools/virtio/linux/dma-mapping.h
33
#define dma_unmap_page_attrs(d, a, s, r, t) do { \
tools/virtio/linux/dma-mapping.h
34
(void)(d); (void)(a); (void)(s); (void)(r); (void)(t); \
tools/virtio/linux/dma-mapping.h
40
#define dma_unmap_single_attrs(d, a, s, r, t) do { \
tools/virtio/linux/dma-mapping.h
41
(void)(d); (void)(a); (void)(s); (void)(r); (void)(t); \
tools/virtio/ringtest/main.h
23
unsigned long long t;
tools/virtio/ringtest/main.h
25
t = __rdtsc();
tools/virtio/ringtest/main.h
26
while (__rdtsc() - t < cycles) {}