arch/alpha/include/asm/elf.h
112
extern void dump_elf_thread(elf_greg_t *dest, struct pt_regs *pt,
arch/alpha/kernel/process.c
286
dump_elf_thread(elf_greg_t *dest, struct pt_regs *pt, struct thread_info *ti)
arch/alpha/kernel/process.c
289
struct switch_stack * sw = ((struct switch_stack *) pt) - 1;
arch/alpha/kernel/process.c
291
dest[ 0] = pt->r0;
arch/alpha/kernel/process.c
292
dest[ 1] = pt->r1;
arch/alpha/kernel/process.c
293
dest[ 2] = pt->r2;
arch/alpha/kernel/process.c
294
dest[ 3] = pt->r3;
arch/alpha/kernel/process.c
295
dest[ 4] = pt->r4;
arch/alpha/kernel/process.c
296
dest[ 5] = pt->r5;
arch/alpha/kernel/process.c
297
dest[ 6] = pt->r6;
arch/alpha/kernel/process.c
298
dest[ 7] = pt->r7;
arch/alpha/kernel/process.c
299
dest[ 8] = pt->r8;
arch/alpha/kernel/process.c
307
dest[16] = pt->r16;
arch/alpha/kernel/process.c
308
dest[17] = pt->r17;
arch/alpha/kernel/process.c
309
dest[18] = pt->r18;
arch/alpha/kernel/process.c
310
dest[19] = pt->r19;
arch/alpha/kernel/process.c
311
dest[20] = pt->r20;
arch/alpha/kernel/process.c
312
dest[21] = pt->r21;
arch/alpha/kernel/process.c
313
dest[22] = pt->r22;
arch/alpha/kernel/process.c
314
dest[23] = pt->r23;
arch/alpha/kernel/process.c
315
dest[24] = pt->r24;
arch/alpha/kernel/process.c
316
dest[25] = pt->r25;
arch/alpha/kernel/process.c
317
dest[26] = pt->r26;
arch/alpha/kernel/process.c
318
dest[27] = pt->r27;
arch/alpha/kernel/process.c
319
dest[28] = pt->r28;
arch/alpha/kernel/process.c
320
dest[29] = pt->gp;
arch/alpha/kernel/process.c
322
dest[31] = pt->pc;
arch/mips/kernel/spinlock_test.c
59
struct spin_multi_per_thread *pt = data;
arch/mips/kernel/spinlock_test.c
60
struct spin_multi_state *s = pt->state;
arch/mips/kernel/spinlock_test.c
70
pt->start = ktime_get();
arch/openrisc/include/asm/elf.h
43
extern void dump_elf_thread(elf_greg_t *dest, struct pt_regs *pt);
arch/parisc/include/asm/elf.h
276
#define ELF_CORE_COPY_REGS(dst, pt) \
arch/parisc/include/asm/elf.h
279
for (i = 0; i < 32; i++) dst[i] = pt->gr[i]; \
arch/parisc/include/asm/elf.h
280
for (i = 0; i < 8; i++) dst[32 + i] = pt->sr[i]; \
arch/parisc/include/asm/elf.h
282
dst[40] = pt->iaoq[0]; dst[41] = pt->iaoq[1]; \
arch/parisc/include/asm/elf.h
283
dst[42] = pt->iasq[0]; dst[43] = pt->iasq[1]; \
arch/parisc/include/asm/elf.h
284
dst[44] = pt->sar; dst[45] = pt->iir; \
arch/parisc/include/asm/elf.h
285
dst[46] = pt->isr; dst[47] = pt->ior; \
arch/s390/kvm/dat.c
237
struct page_table *pt;
arch/s390/kvm/dat.c
249
pt = dat_alloc_pt_noinit(mc);
arch/s390/kvm/dat.c
250
if (!pt)
arch/s390/kvm/dat.c
252
new.val = virt_to_phys(pt);
arch/s390/kvm/dat.c
266
pt->ptes[i].val = init.val | i * PAGE_SIZE;
arch/s390/kvm/dat.c
271
dat_init_pgstes(pt, pgste_init.val);
arch/s390/kvm/dat.c
273
dat_init_page_table(pt, init.val, 0);
arch/s390/kvm/dat.c
280
union pgste pgste = pt->pgstes[i];
arch/s390/kvm/dat.c
282
pgste = dat_save_storage_key_into_pgste(pt->ptes[i], pgste);
arch/s390/kvm/dat.c
283
pgste_set_unlock(pt->ptes + i, pgste);
arch/s390/kvm/dat.c
290
dat_free_pt(pt);
arch/s390/kvm/dat.h
626
static inline void dat_init_pgstes(struct page_table *pt, unsigned long val)
arch/s390/kvm/dat.h
628
memset64((void *)pt->pgstes, val, PTRS_PER_PTE);
arch/s390/kvm/dat.h
631
static inline void dat_init_page_table(struct page_table *pt, unsigned long ptes,
arch/s390/kvm/dat.h
634
memset64((void *)pt->ptes, ptes, PTRS_PER_PTE);
arch/s390/kvm/dat.h
635
dat_init_pgstes(pt, pgstes);
arch/s390/kvm/dat.h
890
static inline void dat_free_pt(struct page_table *pt)
arch/s390/kvm/dat.h
892
free_page((unsigned long)pt);
arch/s390/kvm/gmap.c
703
struct page_table *pt;
arch/s390/kvm/gmap.c
721
pt = pte_table_start(ptep);
arch/s390/kvm/gmap.c
722
dat_set_ptval(pt, PTVAL_VMADDR, p_gfn >> (_SEGMENT_SHIFT - PAGE_SHIFT));
arch/s390/kvm/gmap.c
723
newcrste = _crste_fc0(virt_to_pfn(pt), TABLE_TYPE_SEGMENT);
arch/sparc/include/asm/backoff.h
66
brnz,pt tmp, 88b; \
arch/sparc/include/asm/backoff.h
72
ba,pt %xcc, label; \
arch/sparc/include/asm/head_64.h
63
bne,pt %xcc, 99f; \
arch/sparc/include/asm/head_64.h
67
bgeu,pt %xcc, label; \
arch/sparc/include/asm/head_64.h
74
bne,pt %xcc, 99f; \
arch/sparc/include/asm/head_64.h
78
bgeu,pt %xcc, label; \
arch/sparc/include/asm/tsb.h
174
bne,pt %xcc, 697f; \
arch/sparc/include/asm/tsb.h
189
ba,pt %xcc, 699f; \
arch/sparc/include/asm/tsb.h
219
be,pt %xcc, 700f; \
arch/sparc/include/asm/tsb.h
225
brlz,pt REG1, PTE_LABEL; \
arch/sparc/include/asm/tsb.h
248
be,pt %xcc, 700f; \
arch/sparc/include/asm/tsb.h
253
brlz,pt REG1, PTE_LABEL; \
arch/sparc/include/asm/tsb.h
311
bgu,pt %xcc, 98f; \
arch/sparc/include/asm/tsb.h
313
bgeu,pt %xcc, 98f; \
arch/sparc/include/asm/tsb.h
316
ba,pt %xcc, 99f; \
arch/sparc/include/asm/tsb.h
318
98: ba,pt %xcc, 97b; \
arch/sparc/include/asm/tsb.h
354
be,a,pt %xcc, OK_LABEL; \
arch/sparc/include/asm/tsb.h
376
be,a,pt %xcc, OK_LABEL; \
arch/sparc/include/asm/ttable.h
101
ba,pt %xcc, routine; \
arch/sparc/include/asm/ttable.h
106
ba,pt %xcc, utrap_trap; \
arch/sparc/include/asm/ttable.h
130
ba,pt %xcc, etrap_irq; \
arch/sparc/include/asm/ttable.h
141
ba,a,pt %xcc, rtrap_irq; \
arch/sparc/include/asm/ttable.h
149
ba,pt %xcc, etrap_irq; \
arch/sparc/include/asm/ttable.h
154
ba,a,pt %xcc, rtrap_irq;
arch/sparc/include/asm/ttable.h
161
ba,pt %xcc, etrap_irq; \
arch/sparc/include/asm/ttable.h
166
ba,a,pt %xcc, rtrap_nmi;
arch/sparc/include/asm/ttable.h
175
ba,pt %xcc, etrap; \
arch/sparc/include/asm/ttable.h
181
ba,pt %xcc, rtrap; \
arch/sparc/include/asm/ttable.h
207
ba,pt %xcc, sun4v_itsb_miss; \
arch/sparc/include/asm/ttable.h
217
ba,pt %xcc, sun4v_dtsb_miss; \
arch/sparc/include/asm/ttable.h
226
ba,pt %xcc, etrap; \
arch/sparc/include/asm/ttable.h
228
ba,pt %xcc, sun4v_mcd_detect_precise; \
arch/sparc/include/asm/ttable.h
27
ba,pt %xcc, etrap; \
arch/sparc/include/asm/ttable.h
291
ba,pt %xcc, etrap_save; \
arch/sparc/include/asm/ttable.h
31
ba,pt %xcc, rtrap; \
arch/sparc/include/asm/ttable.h
325
b,a,pt %xcc, spill_fixup_dax; \
arch/sparc/include/asm/ttable.h
326
b,a,pt %xcc, spill_fixup_mna; \
arch/sparc/include/asm/ttable.h
327
b,a,pt %xcc, spill_fixup;
arch/sparc/include/asm/ttable.h
349
ba,pt %xcc, etrap_save; \
arch/sparc/include/asm/ttable.h
353
ba,a,pt %xcc, etrap_spill_fixup_64bit; \
arch/sparc/include/asm/ttable.h
354
ba,a,pt %xcc, etrap_spill_fixup_64bit; \
arch/sparc/include/asm/ttable.h
355
ba,a,pt %xcc, etrap_spill_fixup_64bit;
arch/sparc/include/asm/ttable.h
37
ba,pt %xcc, etrap; \
arch/sparc/include/asm/ttable.h
386
ba,pt %xcc, etrap_save; \
arch/sparc/include/asm/ttable.h
41
ba,pt %xcc, rtrap; \
arch/sparc/include/asm/ttable.h
421
b,a,pt %xcc, spill_fixup_dax; \
arch/sparc/include/asm/ttable.h
422
b,a,pt %xcc, spill_fixup_mna; \
arch/sparc/include/asm/ttable.h
423
b,a,pt %xcc, spill_fixup;
arch/sparc/include/asm/ttable.h
448
ba,pt %xcc, etrap_save; \
arch/sparc/include/asm/ttable.h
452
ba,a,pt %xcc, etrap_spill_fixup_32bit; \
arch/sparc/include/asm/ttable.h
453
ba,a,pt %xcc, etrap_spill_fixup_32bit; \
arch/sparc/include/asm/ttable.h
454
ba,a,pt %xcc, etrap_spill_fixup_32bit;
arch/sparc/include/asm/ttable.h
46
ba,pt %xcc, do_fptrap; \
arch/sparc/include/asm/ttable.h
485
ba,pt %xcc, etrap_save; \
arch/sparc/include/asm/ttable.h
50
ba,pt %xcc, rtrap; \
arch/sparc/include/asm/ttable.h
55
ba,pt %xcc, routine; \
arch/sparc/include/asm/ttable.h
550
ba,pt %xcc, kern_rtt_restore; \
arch/sparc/include/asm/ttable.h
583
b,a,pt %xcc, fill_fixup_dax; \
arch/sparc/include/asm/ttable.h
584
b,a,pt %xcc, fill_fixup_mna; \
arch/sparc/include/asm/ttable.h
585
b,a,pt %xcc, fill_fixup;
arch/sparc/include/asm/ttable.h
60
ba,pt %xcc, routine; \
arch/sparc/include/asm/ttable.h
605
ba,pt %xcc, user_rtt_pre_restore; \
arch/sparc/include/asm/ttable.h
609
ba,a,pt %xcc, user_rtt_fill_fixup_dax; \
arch/sparc/include/asm/ttable.h
610
ba,a,pt %xcc, user_rtt_fill_fixup_mna; \
arch/sparc/include/asm/ttable.h
611
ba,a,pt %xcc, user_rtt_fill_fixup;
arch/sparc/include/asm/ttable.h
643
b,a,pt %xcc, fill_fixup_dax; \
arch/sparc/include/asm/ttable.h
644
b,a,pt %xcc, fill_fixup_mna; \
arch/sparc/include/asm/ttable.h
645
b,a,pt %xcc, fill_fixup;
arch/sparc/include/asm/ttable.h
66
ba,pt %xcc, etraptl1; \
arch/sparc/include/asm/ttable.h
668
ba,pt %xcc, user_rtt_pre_restore; \
arch/sparc/include/asm/ttable.h
672
ba,a,pt %xcc, user_rtt_fill_fixup_dax; \
arch/sparc/include/asm/ttable.h
673
ba,a,pt %xcc, user_rtt_fill_fixup_mna; \
arch/sparc/include/asm/ttable.h
674
ba,a,pt %xcc, user_rtt_fill_fixup;
arch/sparc/include/asm/ttable.h
70
ba,pt %xcc, rtrap; \
arch/sparc/include/asm/ttable.h
76
ba,pt %xcc, etrap; \
arch/sparc/include/asm/ttable.h
81
ba,pt %xcc, rtrap; \
arch/sparc/include/asm/ttable.h
86
ba,pt %xcc, etraptl1; \
arch/sparc/include/asm/ttable.h
91
ba,pt %xcc, rtrap; \
arch/sparc/include/asm/ttable.h
98
ba,pt %xcc, etrap_syscall; \
arch/sparc/include/asm/visasm.h
18
be,pt %icc, 297f; \
arch/sparc/include/asm/visasm.h
40
be,pt %icc, 297f; \
arch/sparc/include/asm/visasm.h
42
ba,a,pt %xcc, fail_label; \
arch/sparc/vdso/vdso2c.h
25
ELF(Phdr) *pt = (ELF(Phdr) *)(raw_addr + GET_BE(&hdr->e_phoff));
arch/sparc/vdso/vdso2c.h
29
if (GET_BE(&pt[i].p_type) == PT_LOAD) {
arch/sparc/vdso/vdso2c.h
33
if (GET_BE(&pt[i].p_offset) != 0 ||
arch/sparc/vdso/vdso2c.h
34
GET_BE(&pt[i].p_vaddr) != 0)
arch/sparc/vdso/vdso2c.h
37
if (GET_BE(&pt[i].p_memsz) != GET_BE(&pt[i].p_filesz))
arch/sparc/vdso/vdso2c.h
40
load_size = GET_BE(&pt[i].p_memsz);
arch/sparc/vdso/vdso2c.h
42
} else if (GET_BE(&pt[i].p_type) == PT_DYNAMIC) {
arch/sparc/vdso/vdso2c.h
43
dyn = raw_addr + GET_BE(&pt[i].p_offset);
arch/sparc/vdso/vdso2c.h
44
dyn_end = raw_addr + GET_BE(&pt[i].p_offset) +
arch/sparc/vdso/vdso2c.h
45
GET_BE(&pt[i].p_memsz);
arch/x86/boot/video-mode.c
120
u8 pt, ov;
arch/x86/boot/video-mode.c
131
pt = in_idx(crtc, 0x11);
arch/x86/boot/video-mode.c
132
pt &= ~0x80; /* Unlock CR0-7 */
arch/x86/boot/video-mode.c
133
out_idx(pt, crtc, 0x11);
arch/x86/events/intel/pt.c
1000
buf->cur_idx = (pt->output_mask & 0xffffff80) >> 7;
arch/x86/events/intel/pt.c
109
PMU_FORMAT_ATTR(pt, "config:0" );
arch/x86/events/intel/pt.c
1507
struct pt *pt = this_cpu_ptr(&pt_ctx);
arch/x86/events/intel/pt.c
1509
struct perf_event *event = pt->handle.event;
arch/x86/events/intel/pt.c
1516
if (!READ_ONCE(pt->handle_nmi))
arch/x86/events/intel/pt.c
1524
buf = perf_get_aux(&pt->handle);
arch/x86/events/intel/pt.c
1530
pt_handle_status(pt);
arch/x86/events/intel/pt.c
1532
pt_update_head(pt);
arch/x86/events/intel/pt.c
1534
perf_aux_output_end(&pt->handle, local_xchg(&buf->data_size, 0));
arch/x86/events/intel/pt.c
1539
buf = perf_aux_output_begin(&pt->handle, event);
arch/x86/events/intel/pt.c
1542
WRITE_ONCE(pt->resume_allowed, 0);
arch/x86/events/intel/pt.c
1546
pt_buffer_reset_offsets(buf, pt->handle.head);
arch/x86/events/intel/pt.c
1548
ret = pt_buffer_reset_markers(buf, &pt->handle);
arch/x86/events/intel/pt.c
1550
perf_aux_output_end(&pt->handle, 0);
arch/x86/events/intel/pt.c
1551
WRITE_ONCE(pt->resume_allowed, 0);
arch/x86/events/intel/pt.c
1562
struct pt *pt = this_cpu_ptr(&pt_ctx);
arch/x86/events/intel/pt.c
1577
WRITE_ONCE(pt->vmx_on, on);
arch/x86/events/intel/pt.c
1583
event = pt->handle.event;
arch/x86/events/intel/pt.c
1585
perf_aux_output_flag(&pt->handle,
arch/x86/events/intel/pt.c
1603
struct pt *pt = this_cpu_ptr(&pt_ctx);
arch/x86/events/intel/pt.c
1607
if (READ_ONCE(pt->resume_allowed)) {
arch/x86/events/intel/pt.c
1620
READ_ONCE(pt->resume_allowed))
arch/x86/events/intel/pt.c
1626
buf = perf_aux_output_begin(&pt->handle, event);
arch/x86/events/intel/pt.c
1630
pt_buffer_reset_offsets(buf, pt->handle.head);
arch/x86/events/intel/pt.c
1632
if (pt_buffer_reset_markers(buf, &pt->handle))
arch/x86/events/intel/pt.c
1644
perf_aux_output_end(&pt->handle, 0);
arch/x86/events/intel/pt.c
1651
struct pt *pt = this_cpu_ptr(&pt_ctx);
arch/x86/events/intel/pt.c
1654
if (READ_ONCE(pt->pause_allowed))
arch/x86/events/intel/pt.c
1663
WRITE_ONCE(pt->handle_nmi, 0);
arch/x86/events/intel/pt.c
1671
WRITE_ONCE(pt->pause_allowed, 0);
arch/x86/events/intel/pt.c
1672
WRITE_ONCE(pt->resume_allowed, 0);
arch/x86/events/intel/pt.c
1683
struct pt_buffer *buf = perf_get_aux(&pt->handle);
arch/x86/events/intel/pt.c
1688
if (WARN_ON_ONCE(pt->handle.event != event))
arch/x86/events/intel/pt.c
1693
pt_handle_status(pt);
arch/x86/events/intel/pt.c
1695
pt_update_head(pt);
arch/x86/events/intel/pt.c
1698
pt->handle.head =
arch/x86/events/intel/pt.c
1701
perf_aux_output_end(&pt->handle, local_xchg(&buf->data_size, 0));
arch/x86/events/intel/pt.c
1709
struct pt *pt = this_cpu_ptr(&pt_ctx);
arch/x86/events/intel/pt.c
1710
struct pt_buffer *buf = perf_get_aux(&pt->handle);
arch/x86/events/intel/pt.c
1725
WRITE_ONCE(pt->pause_allowed, 0);
arch/x86/events/intel/pt.c
1726
WRITE_ONCE(pt->resume_allowed, 0);
arch/x86/events/intel/pt.c
1734
pt_update_head(pt);
arch/x86/events/intel/pt.c
1741
ret = perf_output_copy_aux(&pt->handle, handle, from, to);
arch/x86/events/intel/pt.c
1747
if (READ_ONCE(pt->handle_nmi)) {
arch/x86/events/intel/pt.c
1748
WRITE_ONCE(pt->resume_allowed, 1);
arch/x86/events/intel/pt.c
1752
WRITE_ONCE(pt->pause_allowed, 1);
arch/x86/events/intel/pt.c
1765
struct pt *pt = this_cpu_ptr(&pt_ctx);
arch/x86/events/intel/pt.c
1769
if (pt->handle.event)
arch/x86/events/intel/pt.c
1820
struct pt *pt = this_cpu_ptr(&pt_ctx);
arch/x86/events/intel/pt.c
1822
if (pt->handle.event)
arch/x86/events/intel/pt.c
1823
pt_event_stop(pt->handle.event, PERF_EF_UPDATE);
arch/x86/events/intel/pt.c
33
static DEFINE_PER_CPU(struct pt, pt_ctx);
arch/x86/events/intel/pt.c
421
struct pt *pt = this_cpu_ptr(&pt_ctx);
arch/x86/events/intel/pt.c
428
if (READ_ONCE(pt->vmx_on))
arch/x86/events/intel/pt.c
429
perf_aux_output_flag(&pt->handle, PERF_AUX_FLAG_PARTIAL);
arch/x86/events/intel/pt.c
467
struct pt *pt = this_cpu_ptr(&pt_ctx);
arch/x86/events/intel/pt.c
489
if (pt->filters.filter[range].msr_a != filter->msr_a) {
arch/x86/events/intel/pt.c
491
pt->filters.filter[range].msr_a = filter->msr_a;
arch/x86/events/intel/pt.c
494
if (pt->filters.filter[range].msr_b != filter->msr_b) {
arch/x86/events/intel/pt.c
496
pt->filters.filter[range].msr_b = filter->msr_b;
arch/x86/events/intel/pt.c
507
struct pt *pt = this_cpu_ptr(&pt_ctx);
arch/x86/events/intel/pt.c
508
struct pt_buffer *buf = perf_get_aux(&pt->handle);
arch/x86/events/intel/pt.c
549
WRITE_ONCE(pt->resume_allowed, 1);
arch/x86/events/intel/pt.c
552
WRITE_ONCE(pt->handle_nmi, 1);
arch/x86/events/intel/pt.c
560
WRITE_ONCE(pt->pause_allowed, 1);
arch/x86/events/intel/pt.c
565
struct pt *pt = this_cpu_ptr(&pt_ctx);
arch/x86/events/intel/pt.c
573
if (!READ_ONCE(pt->vmx_on))
arch/x86/events/intel/pt.c
648
struct pt *pt = this_cpu_ptr(&pt_ctx);
arch/x86/events/intel/pt.c
661
if (pt->output_base != reg) {
arch/x86/events/intel/pt.c
662
pt->output_base = reg;
arch/x86/events/intel/pt.c
667
if (pt->output_mask != reg) {
arch/x86/events/intel/pt.c
668
pt->output_mask = reg;
arch/x86/events/intel/pt.c
871
static void pt_update_head(struct pt *pt)
arch/x86/events/intel/pt.c
873
struct pt_buffer *buf = perf_get_aux(&pt->handle);
arch/x86/events/intel/pt.c
925
static void pt_handle_status(struct pt *pt)
arch/x86/events/intel/pt.c
927
struct pt_buffer *buf = perf_get_aux(&pt->handle);
arch/x86/events/intel/pt.c
950
perf_aux_output_flag(&pt->handle,
arch/x86/events/intel/pt.c
986
struct pt *pt = this_cpu_ptr(&pt_ctx);
arch/x86/events/intel/pt.c
990
rdmsrq(MSR_IA32_RTIT_OUTPUT_BASE, pt->output_base);
arch/x86/events/intel/pt.c
991
tp = phys_to_virt(pt->output_base);
arch/x86/events/intel/pt.c
995
rdmsrq(MSR_IA32_RTIT_OUTPUT_MASK, pt->output_mask);
arch/x86/events/intel/pt.c
997
buf->output_off = pt->output_mask >> 32;
arch/x86/include/asm/vm86.h
21
struct pt_regs pt;
arch/x86/kernel/cpu/sgx/ioctl.c
183
u64 pt = secinfo->flags & SGX_SECINFO_PAGE_TYPE_MASK;
arch/x86/kernel/cpu/sgx/ioctl.c
185
if (pt != SGX_SECINFO_REG && pt != SGX_SECINFO_TCS)
arch/x86/kernel/cpu/sgx/ioctl.c
195
if (pt == SGX_SECINFO_TCS && perm)
arch/x86/kernel/vm86_32.c
111
set_flags(regs->pt.flags, VEFLAGS, X86_EFLAGS_VIF | vm86->veflags_mask);
arch/x86/kernel/vm86_32.c
119
unsafe_put_user(regs->pt.bx, &user->regs.ebx, Efault_end);
arch/x86/kernel/vm86_32.c
120
unsafe_put_user(regs->pt.cx, &user->regs.ecx, Efault_end);
arch/x86/kernel/vm86_32.c
121
unsafe_put_user(regs->pt.dx, &user->regs.edx, Efault_end);
arch/x86/kernel/vm86_32.c
122
unsafe_put_user(regs->pt.si, &user->regs.esi, Efault_end);
arch/x86/kernel/vm86_32.c
123
unsafe_put_user(regs->pt.di, &user->regs.edi, Efault_end);
arch/x86/kernel/vm86_32.c
124
unsafe_put_user(regs->pt.bp, &user->regs.ebp, Efault_end);
arch/x86/kernel/vm86_32.c
125
unsafe_put_user(regs->pt.ax, &user->regs.eax, Efault_end);
arch/x86/kernel/vm86_32.c
126
unsafe_put_user(regs->pt.ip, &user->regs.eip, Efault_end);
arch/x86/kernel/vm86_32.c
127
unsafe_put_user(regs->pt.cs, &user->regs.cs, Efault_end);
arch/x86/kernel/vm86_32.c
128
unsafe_put_user(regs->pt.flags, &user->regs.eflags, Efault_end);
arch/x86/kernel/vm86_32.c
129
unsafe_put_user(regs->pt.sp, &user->regs.esp, Efault_end);
arch/x86/kernel/vm86_32.c
130
unsafe_put_user(regs->pt.ss, &user->regs.ss, Efault_end);
arch/x86/kernel/vm86_32.c
152
memcpy(®s->pt, &vm86->regs32, sizeof(struct pt_regs));
arch/x86/kernel/vm86_32.c
156
regs->pt.ax = retval;
arch/x86/kernel/vm86_32.c
256
vm86regs.pt.bx = v.regs.ebx;
arch/x86/kernel/vm86_32.c
257
vm86regs.pt.cx = v.regs.ecx;
arch/x86/kernel/vm86_32.c
258
vm86regs.pt.dx = v.regs.edx;
arch/x86/kernel/vm86_32.c
259
vm86regs.pt.si = v.regs.esi;
arch/x86/kernel/vm86_32.c
260
vm86regs.pt.di = v.regs.edi;
arch/x86/kernel/vm86_32.c
261
vm86regs.pt.bp = v.regs.ebp;
arch/x86/kernel/vm86_32.c
262
vm86regs.pt.ax = v.regs.eax;
arch/x86/kernel/vm86_32.c
263
vm86regs.pt.ip = v.regs.eip;
arch/x86/kernel/vm86_32.c
264
vm86regs.pt.cs = v.regs.cs;
arch/x86/kernel/vm86_32.c
265
vm86regs.pt.flags = v.regs.eflags;
arch/x86/kernel/vm86_32.c
266
vm86regs.pt.sp = v.regs.esp;
arch/x86/kernel/vm86_32.c
267
vm86regs.pt.ss = v.regs.ss;
arch/x86/kernel/vm86_32.c
301
VEFLAGS = vm86regs.pt.flags;
arch/x86/kernel/vm86_32.c
302
vm86regs.pt.flags &= SAFE_MASK;
arch/x86/kernel/vm86_32.c
303
vm86regs.pt.flags |= regs->flags & ~SAFE_MASK;
arch/x86/kernel/vm86_32.c
304
vm86regs.pt.flags |= X86_VM_MASK;
arch/x86/kernel/vm86_32.c
306
vm86regs.pt.orig_ax = regs->orig_ax;
arch/x86/kernel/vm86_32.c
357
regs->pt.flags &= ~X86_EFLAGS_TF;
arch/x86/kernel/vm86_32.c
362
regs->pt.flags &= ~X86_EFLAGS_AC;
arch/x86/kernel/vm86_32.c
380
set_flags(regs->pt.flags, flags, SAFE_MASK);
arch/x86/kernel/vm86_32.c
390
set_flags(regs->pt.flags, flags, SAFE_MASK);
arch/x86/kernel/vm86_32.c
399
unsigned long flags = regs->pt.flags & RETURN_MASK;
arch/x86/kernel/vm86_32.c
501
if (regs->pt.cs == BIOSSEG)
arch/x86/kernel/vm86_32.c
513
pushw(ssp, sp, regs->pt.cs, cannot_handle);
arch/x86/kernel/vm86_32.c
515
regs->pt.cs = segoffs >> 16;
arch/x86/kernel/vm86_32.c
536
do_int(regs, trapno, (unsigned char __user *) (regs->pt.ss << 4), SP(regs));
arch/x86/kernel/vm86_32.c
560
orig_flags = *(unsigned short *)®s->pt.flags;
arch/x86/kernel/vm86_32.c
562
csp = (unsigned char __user *) (regs->pt.cs << 4);
arch/x86/kernel/vm86_32.c
563
ssp = (unsigned char __user *) (regs->pt.ss << 4);
arch/x86/kernel/vm86_32.c
652
regs->pt.cs = newcs;
arch/x86/kernel/vm86_32.c
79
#define AL(regs) (((unsigned char *)&((regs)->pt.ax))[0])
arch/x86/kernel/vm86_32.c
80
#define AH(regs) (((unsigned char *)&((regs)->pt.ax))[1])
arch/x86/kernel/vm86_32.c
81
#define IP(regs) (*(unsigned short *)&((regs)->pt.ip))
arch/x86/kernel/vm86_32.c
82
#define SP(regs) (*(unsigned short *)&((regs)->pt.sp))
arch/x86/kvm/i8254.c
271
struct kvm_pit *pt = pit_state_to_pit(ps);
arch/x86/kvm/i8254.c
276
kthread_queue_work(pt->worker, &pt->expired);
arch/x86/kvm/mmu/tdp_mmu.c
398
static void handle_removed_pt(struct kvm *kvm, tdp_ptep_t pt, bool shared)
arch/x86/kvm/mmu/tdp_mmu.c
400
struct kvm_mmu_page *sp = sptep_to_sp(rcu_dereference(pt));
arch/x86/kvm/mmu/tdp_mmu.c
410
tdp_ptep_t sptep = pt + i;
arch/x86/mm/pgtable.c
732
struct ptdesc *pt;
arch/x86/mm/pgtable.c
753
pt = page_ptdesc(pmd_page(pmd_sv[i]));
arch/x86/mm/pgtable.c
754
pagetable_dtor_free(pt);
arch/x86/mm/pgtable.c
775
struct ptdesc *pt;
arch/x86/mm/pgtable.c
777
pt = page_ptdesc(pmd_page(*pmd));
arch/x86/mm/pgtable.c
783
pagetable_dtor_free(pt);
arch/x86/tools/vdso2c.h
58
ELF(Phdr) *pt = (ELF(Phdr) *)(raw_addr + GET_LE(&hdr->e_phoff));
arch/x86/tools/vdso2c.h
65
if (GET_LE(&pt[i].p_type) == PT_LOAD) {
arch/x86/tools/vdso2c.h
69
if (GET_LE(&pt[i].p_offset) != 0 ||
arch/x86/tools/vdso2c.h
70
GET_LE(&pt[i].p_vaddr) != 0)
arch/x86/tools/vdso2c.h
73
if (GET_LE(&pt[i].p_memsz) != GET_LE(&pt[i].p_filesz))
arch/x86/tools/vdso2c.h
76
load_size = GET_LE(&pt[i].p_memsz);
arch/x86/tools/vdso2c.h
78
} else if (GET_LE(&pt[i].p_type) == PT_DYNAMIC) {
arch/x86/tools/vdso2c.h
79
dyn = raw_addr + GET_LE(&pt[i].p_offset);
arch/x86/tools/vdso2c.h
80
dyn_end = raw_addr + GET_LE(&pt[i].p_offset) +
arch/x86/tools/vdso2c.h
81
GET_LE(&pt[i].p_memsz);
arch/x86/xen/mmu_pv.c
1958
pte_t *pt;
arch/x86/xen/mmu_pv.c
2001
pt = early_memremap(pt_phys, PAGE_SIZE);
arch/x86/xen/mmu_pv.c
2002
clear_page(pt);
arch/x86/xen/mmu_pv.c
2006
pt[idx_pte] = pfn_pte(p2m_pfn,
arch/x86/xen/mmu_pv.c
2011
early_memunmap(pt, PAGE_SIZE);
arch/x86/xen/mmu_pv.c
747
void *pt = lowmem_page_address(page);
arch/x86/xen/mmu_pv.c
776
MULTI_update_va_mapping(mcs.mc, (unsigned long)pt,
arch/x86/xen/mmu_pv.c
877
void *pt = lowmem_page_address(page);
arch/x86/xen/mmu_pv.c
898
MULTI_update_va_mapping(mcs.mc, (unsigned long)pt,
block/partitions/msdos.c
74
struct msdos_partition *pt = (struct msdos_partition *) (p + 0x1be);
block/partitions/msdos.c
91
for (slot = 1; slot <= 4; slot++, pt++) {
block/partitions/msdos.c
92
if (pt->sys_ind == SOLARIS_X86_PARTITION ||
block/partitions/msdos.c
93
pt->sys_ind == LINUX_RAID_PARTITION ||
block/partitions/msdos.c
94
pt->sys_ind == LINUX_DATA_PARTITION ||
block/partitions/msdos.c
95
pt->sys_ind == LINUX_LVM_PARTITION ||
block/partitions/msdos.c
96
is_extended_partition(pt))
drivers/atm/nicstar.h
237
#define ns_tbd_mkword_4(gfc, vpi, vci, pt, clp) \
drivers/atm/nicstar.h
238
(cpu_to_le32((gfc) << 28 | (vpi) << 20 | (vci) << 4 | (pt) << 1 | (clp)))
drivers/block/aoe/aoenet.c
131
aoenet_rcv(struct sk_buff *skb, struct net_device *ifp, struct packet_type *pt, struct net_device *orig_dev)
drivers/clk/clk-sp7021.c
187
static const u32 pt[][5] = {
drivers/clk/clk-sp7021.c
228
for (ph = ARRAY_SIZE(pt) - 1; ph >= 0; ph--) {
drivers/clk/clk-sp7021.c
229
const u32 *pp = pt[ph];
drivers/clk/clk-sp7021.c
455
const u32 *pp = pt[ph];
drivers/comedi/drivers/addi_apci_1500.c
300
z8536_write(dev, devpriv->pt[pa_trig] & 0xff, Z8536_PA_PT_REG);
drivers/comedi/drivers/addi_apci_1500.c
305
z8536_write(dev, (devpriv->pt[pb_trig] >> 8) & 0xff, Z8536_PB_PT_REG);
drivers/comedi/drivers/addi_apci_1500.c
460
unsigned int pt;
drivers/comedi/drivers/addi_apci_1500.c
489
pt = devpriv->pt[trig] & old_mask;
drivers/comedi/drivers/addi_apci_1500.c
496
pt = 0;
drivers/comedi/drivers/addi_apci_1500.c
501
pt |= chan_mask; /* enable edge detection */
drivers/comedi/drivers/addi_apci_1500.c
507
pt &= ~chan_mask; /* enable level detection */
drivers/comedi/drivers/addi_apci_1500.c
51
unsigned int pt[2]; /* Pattern Transition */
drivers/comedi/drivers/addi_apci_1500.c
523
src = pt & 0xff;
drivers/comedi/drivers/addi_apci_1500.c
527
src = (pt >> 8) & 0xff;
drivers/comedi/drivers/addi_apci_1500.c
540
devpriv->pt[trig] = pt;
drivers/crypto/ccp/ccp-dev-v5.c
125
} pt;
drivers/crypto/ccp/ccp-dev-v5.c
150
#define CCP_PT_BYTESWAP(p) ((p)->pt.byteswap)
drivers/crypto/ccp/ccp-dev-v5.c
151
#define CCP_PT_BITWISE(p) ((p)->pt.bitwise)
drivers/crypto/ccp/ccp-ops.c
1979
struct ccp_passthru_engine *pt = &cmd->u.passthru;
drivers/crypto/ccp/ccp-ops.c
1987
if (!pt->final && (pt->src_len & (CCP_PASSTHRU_BLOCKSIZE - 1)))
drivers/crypto/ccp/ccp-ops.c
1990
if (!pt->src || !pt->dst)
drivers/crypto/ccp/ccp-ops.c
1993
if (pt->bit_mod != CCP_PASSTHRU_BITWISE_NOOP) {
drivers/crypto/ccp/ccp-ops.c
1994
if (pt->mask_len != CCP_PASSTHRU_MASKSIZE)
drivers/crypto/ccp/ccp-ops.c
1996
if (!pt->mask)
drivers/crypto/ccp/ccp-ops.c
2006
if (pt->bit_mod != CCP_PASSTHRU_BITWISE_NOOP) {
drivers/crypto/ccp/ccp-ops.c
2017
ret = ccp_set_dm_area(&mask, 0, pt->mask, 0, pt->mask_len);
drivers/crypto/ccp/ccp-ops.c
2032
if (sg_virt(pt->src) == sg_virt(pt->dst))
drivers/crypto/ccp/ccp-ops.c
2035
ret = ccp_init_data(&src, cmd_q, pt->src, pt->src_len,
drivers/crypto/ccp/ccp-ops.c
2044
ret = ccp_init_data(&dst, cmd_q, pt->dst, pt->src_len,
drivers/crypto/ccp/ccp-ops.c
2102
if (pt->bit_mod != CCP_PASSTHRU_BITWISE_NOOP)
drivers/crypto/ccp/ccp-ops.c
2112
struct ccp_passthru_nomap_engine *pt = &cmd->u.passthru_nomap;
drivers/crypto/ccp/ccp-ops.c
2117
if (!pt->final && (pt->src_len & (CCP_PASSTHRU_BLOCKSIZE - 1)))
drivers/crypto/ccp/ccp-ops.c
2120
if (!pt->src_dma || !pt->dst_dma)
drivers/crypto/ccp/ccp-ops.c
2123
if (pt->bit_mod != CCP_PASSTHRU_BITWISE_NOOP) {
drivers/crypto/ccp/ccp-ops.c
2124
if (pt->mask_len != CCP_PASSTHRU_MASKSIZE)
drivers/crypto/ccp/ccp-ops.c
2126
if (!pt->mask)
drivers/crypto/ccp/ccp-ops.c
2136
if (pt->bit_mod != CCP_PASSTHRU_BITWISE_NOOP) {
drivers/crypto/ccp/ccp-ops.c
2140
mask.length = pt->mask_len;
drivers/crypto/ccp/ccp-ops.c
2141
mask.dma.address = pt->mask;
drivers/crypto/ccp/ccp-ops.c
2142
mask.dma.length = pt->mask_len;
drivers/crypto/ccp/ccp-ops.c
2157
op.src.u.dma.address = pt->src_dma;
drivers/crypto/ccp/ccp-ops.c
2159
op.src.u.dma.length = pt->src_len;
drivers/crypto/ccp/ccp-ops.c
2162
op.dst.u.dma.address = pt->dst_dma;
drivers/crypto/ccp/ccp-ops.c
2164
op.dst.u.dma.length = pt->src_len;
drivers/crypto/starfive/jh7110-rsa.c
383
unsigned char pt;
drivers/crypto/starfive/jh7110-rsa.c
390
pt = *ptr;
drivers/crypto/starfive/jh7110-rsa.c
402
if (pt >> (loop - 1))
drivers/crypto/starfive/jh7110-rsa.c
417
unsigned char pt;
drivers/crypto/starfive/jh7110-rsa.c
425
pt = *ptr;
drivers/crypto/starfive/jh7110-rsa.c
437
if (pt >> (loop - 1))
drivers/dma-buf/sw_sync.c
155
struct sync_pt *pt = dma_fence_to_sync_pt(fence);
drivers/dma-buf/sw_sync.c
160
if (!list_empty(&pt->link)) {
drivers/dma-buf/sw_sync.c
161
list_del(&pt->link);
drivers/dma-buf/sw_sync.c
162
rb_erase(&pt->node, &parent->pt_tree);
drivers/dma-buf/sw_sync.c
179
struct sync_pt *pt = dma_fence_to_sync_pt(fence);
drivers/dma-buf/sw_sync.c
184
if (ktime_before(deadline, pt->deadline))
drivers/dma-buf/sw_sync.c
185
pt->deadline = deadline;
drivers/dma-buf/sw_sync.c
187
pt->deadline = deadline;
drivers/dma-buf/sw_sync.c
212
struct sync_pt *pt, *next;
drivers/dma-buf/sw_sync.c
220
list_for_each_entry_safe(pt, next, &obj->pt_list, link) {
drivers/dma-buf/sw_sync.c
221
if (!timeline_fence_signaled(&pt->base))
drivers/dma-buf/sw_sync.c
224
dma_fence_get(&pt->base);
drivers/dma-buf/sw_sync.c
226
list_move_tail(&pt->link, &signalled);
drivers/dma-buf/sw_sync.c
227
rb_erase(&pt->node, &obj->pt_tree);
drivers/dma-buf/sw_sync.c
229
dma_fence_signal_locked(&pt->base);
drivers/dma-buf/sw_sync.c
234
list_for_each_entry_safe(pt, next, &signalled, link) {
drivers/dma-buf/sw_sync.c
235
list_del_init(&pt->link);
drivers/dma-buf/sw_sync.c
236
dma_fence_put(&pt->base);
drivers/dma-buf/sw_sync.c
253
struct sync_pt *pt;
drivers/dma-buf/sw_sync.c
255
pt = kzalloc_obj(*pt);
drivers/dma-buf/sw_sync.c
256
if (!pt)
drivers/dma-buf/sw_sync.c
260
dma_fence_init(&pt->base, &timeline_fence_ops, &obj->lock,
drivers/dma-buf/sw_sync.c
262
INIT_LIST_HEAD(&pt->link);
drivers/dma-buf/sw_sync.c
265
if (!dma_fence_is_signaled_locked(&pt->base)) {
drivers/dma-buf/sw_sync.c
274
other = rb_entry(parent, typeof(*pt), node);
drivers/dma-buf/sw_sync.c
283
kfree(pt);
drivers/dma-buf/sw_sync.c
284
pt = other;
drivers/dma-buf/sw_sync.c
290
rb_link_node(&pt->node, parent, p);
drivers/dma-buf/sw_sync.c
291
rb_insert_color(&pt->node, &obj->pt_tree);
drivers/dma-buf/sw_sync.c
293
parent = rb_next(&pt->node);
drivers/dma-buf/sw_sync.c
294
list_add_tail(&pt->link,
drivers/dma-buf/sw_sync.c
295
parent ? &rb_entry(parent, typeof(*pt), node)->link : &obj->pt_list);
drivers/dma-buf/sw_sync.c
300
return pt;
drivers/dma-buf/sw_sync.c
329
struct sync_pt *pt, *next;
drivers/dma-buf/sw_sync.c
333
list_for_each_entry_safe(pt, next, &obj->pt_list, link) {
drivers/dma-buf/sw_sync.c
334
dma_fence_set_error(&pt->base, -ENOENT);
drivers/dma-buf/sw_sync.c
335
dma_fence_signal_locked(&pt->base);
drivers/dma-buf/sw_sync.c
349
struct sync_pt *pt;
drivers/dma-buf/sw_sync.c
364
pt = sync_pt_create(obj, data.value);
drivers/dma-buf/sw_sync.c
365
if (!pt) {
drivers/dma-buf/sw_sync.c
370
sync_file = sync_file_create(&pt->base);
drivers/dma-buf/sw_sync.c
371
dma_fence_put(&pt->base);
drivers/dma-buf/sw_sync.c
415
struct sync_pt *pt;
drivers/dma-buf/sw_sync.c
428
pt = dma_fence_to_sync_pt(fence);
drivers/dma-buf/sw_sync.c
429
if (!pt) {
drivers/dma-buf/sw_sync.c
439
data.deadline_ns = ktime_to_ns(pt->deadline);
drivers/dma-buf/sync_debug.c
78
struct sync_pt *pt = container_of(pos, struct sync_pt, link);
drivers/dma-buf/sync_debug.c
79
sync_print_fence(s, &pt->base, false);
drivers/dma/amd/ae4dma/ae4dma-dev.c
105
cmd_q->pt = pt;
drivers/dma/amd/ae4dma/ae4dma-dev.c
107
cmd_q->reg_control = pt->io_regs + ((i + 1) * AE4_Q_SZ);
drivers/dma/amd/ae4dma/ae4dma-dev.c
110
dev_name(pt->dev), ae4cmd_q);
drivers/dma/amd/ae4dma/ae4dma-dev.c
127
cmd_q->reg_control = pt->io_regs + ((i + 1) * AE4_Q_SZ);
drivers/dma/amd/ae4dma/ae4dma-dev.c
150
ret = pt_dmaengine_register(pt);
drivers/dma/amd/ae4dma/ae4dma-dev.c
154
ptdma_debugfs_setup(pt);
drivers/dma/amd/ae4dma/ae4dma-dev.c
53
struct pt_device *pt;
drivers/dma/amd/ae4dma/ae4dma-dev.c
57
pt = cmd_q->pt;
drivers/dma/amd/ae4dma/ae4dma-dev.c
59
pt->total_interrupts++;
drivers/dma/amd/ae4dma/ae4dma-dev.c
91
struct pt_device *pt = &ae4->pt;
drivers/dma/amd/ae4dma/ae4dma-dev.c
93
struct device *dev = pt->dev;
drivers/dma/amd/ae4dma/ae4dma-dev.c
97
writel(max_hw_q, pt->io_regs);
drivers/dma/amd/ae4dma/ae4dma-pci.c
100
pt->dev = dev;
drivers/dma/amd/ae4dma/ae4dma-pci.c
101
pt->ver = AE4_DMA_VERSION;
drivers/dma/amd/ae4dma/ae4dma-pci.c
103
pt->io_regs = pcim_iomap_table(pdev)[0];
drivers/dma/amd/ae4dma/ae4dma-pci.c
104
if (!pt->io_regs) {
drivers/dma/amd/ae4dma/ae4dma-pci.c
16
struct pt_device *pt = &ae4->pt;
drivers/dma/amd/ae4dma/ae4dma-pci.c
17
struct device *dev = pt->dev;
drivers/dma/amd/ae4dma/ae4dma-pci.c
59
struct pt_device *pt = &ae4->pt;
drivers/dma/amd/ae4dma/ae4dma-pci.c
60
struct device *dev = pt->dev;
drivers/dma/amd/ae4dma/ae4dma-pci.c
78
struct pt_device *pt;
drivers/dma/amd/ae4dma/ae4dma-pci.c
99
pt = &ae4->pt;
drivers/dma/amd/ae4dma/ae4dma.h
92
struct pt_device pt;
drivers/dma/amd/ptdma/ptdma-debugfs.c
101
void ptdma_debugfs_setup(struct pt_device *pt)
drivers/dma/amd/ptdma/ptdma-debugfs.c
113
debugfs_create_file("info", 0400, pt->dma_dev.dbg_dev_root, pt,
drivers/dma/amd/ptdma/ptdma-debugfs.c
116
debugfs_create_file("stats", 0400, pt->dma_dev.dbg_dev_root, pt,
drivers/dma/amd/ptdma/ptdma-debugfs.c
120
if (pt->ver == AE4_DMA_VERSION) {
drivers/dma/amd/ptdma/ptdma-debugfs.c
121
ae4 = container_of(pt, struct ae4_device, pt);
drivers/dma/amd/ptdma/ptdma-debugfs.c
130
debugfs_create_dir(name, pt->dma_dev.dbg_dev_root);
drivers/dma/amd/ptdma/ptdma-debugfs.c
137
debugfs_create_dir("q", pt->dma_dev.dbg_dev_root);
drivers/dma/amd/ptdma/ptdma-debugfs.c
138
cmd_q = &pt->cmd_q;
drivers/dma/amd/ptdma/ptdma-debugfs.c
26
struct pt_device *pt = s->private;
drivers/dma/amd/ptdma/ptdma-debugfs.c
30
seq_printf(s, "Device name: %s\n", dev_name(pt->dev));
drivers/dma/amd/ptdma/ptdma-debugfs.c
32
if (pt->ver == AE4_DMA_VERSION) {
drivers/dma/amd/ptdma/ptdma-debugfs.c
33
ae4 = container_of(pt, struct ae4_device, pt);
drivers/dma/amd/ptdma/ptdma-debugfs.c
38
seq_printf(s, " # Cmds: %d\n", pt->cmd_count);
drivers/dma/amd/ptdma/ptdma-debugfs.c
41
regval = ioread32(pt->io_regs + CMD_PT_VERSION);
drivers/dma/amd/ptdma/ptdma-debugfs.c
57
struct pt_device *pt = s->private;
drivers/dma/amd/ptdma/ptdma-debugfs.c
59
seq_printf(s, "Total Interrupts Handled: %ld\n", pt->total_interrupts);
drivers/dma/amd/ptdma/ptdma-debugfs.c
67
struct pt_device *pt;
drivers/dma/amd/ptdma/ptdma-debugfs.c
75
pt = cmd_q->pt;
drivers/dma/amd/ptdma/ptdma-debugfs.c
76
if (pt->ver == AE4_DMA_VERSION) {
drivers/dma/amd/ptdma/ptdma-dev.c
104
struct pt_device *pt = container_of(cmd_q, struct pt_device, cmd_q);
drivers/dma/amd/ptdma/ptdma-dev.c
117
pt_core_enable_queue_interrupts(pt);
drivers/dma/amd/ptdma/ptdma-dev.c
119
pt_core_disable_queue_interrupts(pt);
drivers/dma/amd/ptdma/ptdma-dev.c
128
struct pt_cmd_queue *cmd_q = &cmd->pt->cmd_q;
drivers/dma/amd/ptdma/ptdma-dev.c
137
pt_log_error(cmd_q->pt, cmd_q->cmd_error);
drivers/dma/amd/ptdma/ptdma-dev.c
144
void pt_check_status_trans(struct pt_device *pt, struct pt_cmd_queue *cmd_q)
drivers/dma/amd/ptdma/ptdma-dev.c
160
pt_do_cmd_complete((ulong)&pt->tdata);
drivers/dma/amd/ptdma/ptdma-dev.c
166
struct pt_device *pt = data;
drivers/dma/amd/ptdma/ptdma-dev.c
167
struct pt_cmd_queue *cmd_q = &pt->cmd_q;
drivers/dma/amd/ptdma/ptdma-dev.c
169
pt_core_disable_queue_interrupts(pt);
drivers/dma/amd/ptdma/ptdma-dev.c
170
pt->total_interrupts++;
drivers/dma/amd/ptdma/ptdma-dev.c
171
pt_check_status_trans(pt, cmd_q);
drivers/dma/amd/ptdma/ptdma-dev.c
172
pt_core_enable_queue_interrupts(pt);
drivers/dma/amd/ptdma/ptdma-dev.c
176
int pt_core_init(struct pt_device *pt)
drivers/dma/amd/ptdma/ptdma-dev.c
179
struct pt_cmd_queue *cmd_q = &pt->cmd_q;
drivers/dma/amd/ptdma/ptdma-dev.c
181
struct device *dev = pt->dev;
drivers/dma/amd/ptdma/ptdma-dev.c
186
snprintf(dma_pool_name, sizeof(dma_pool_name), "%s_q", dev_name(pt->dev));
drivers/dma/amd/ptdma/ptdma-dev.c
195
iowrite32(CMD_CONFIG_VHB_EN, pt->io_regs + CMD_CONFIG_OFFSET);
drivers/dma/amd/ptdma/ptdma-dev.c
196
iowrite32(CMD_QUEUE_PRIO, pt->io_regs + CMD_QUEUE_PRIO_OFFSET);
drivers/dma/amd/ptdma/ptdma-dev.c
197
iowrite32(CMD_TIMEOUT_DISABLE, pt->io_regs + CMD_TIMEOUT_OFFSET);
drivers/dma/amd/ptdma/ptdma-dev.c
198
iowrite32(CMD_CLK_GATE_CONFIG, pt->io_regs + CMD_CLK_GATE_CTL_OFFSET);
drivers/dma/amd/ptdma/ptdma-dev.c
199
iowrite32(CMD_CONFIG_REQID, pt->io_regs + CMD_REQID_CONFIG_OFFSET);
drivers/dma/amd/ptdma/ptdma-dev.c
201
cmd_q->pt = pt;
drivers/dma/amd/ptdma/ptdma-dev.c
219
cmd_q->reg_control = pt->io_regs + CMD_Q_STATUS_INCR;
drivers/dma/amd/ptdma/ptdma-dev.c
222
pt_core_disable_queue_interrupts(pt);
drivers/dma/amd/ptdma/ptdma-dev.c
234
ret = request_irq(pt->pt_irq, pt_core_irq_handler, 0, dev_name(pt->dev), pt);
drivers/dma/amd/ptdma/ptdma-dev.c
253
pt_core_enable_queue_interrupts(pt);
drivers/dma/amd/ptdma/ptdma-dev.c
256
ret = pt_dmaengine_register(pt);
drivers/dma/amd/ptdma/ptdma-dev.c
261
ptdma_debugfs_setup(pt);
drivers/dma/amd/ptdma/ptdma-dev.c
266
free_irq(pt->pt_irq, pt);
drivers/dma/amd/ptdma/ptdma-dev.c
272
dma_pool_destroy(pt->cmd_q.dma_pool);
drivers/dma/amd/ptdma/ptdma-dev.c
277
void pt_core_destroy(struct pt_device *pt)
drivers/dma/amd/ptdma/ptdma-dev.c
279
struct device *dev = pt->dev;
drivers/dma/amd/ptdma/ptdma-dev.c
280
struct pt_cmd_queue *cmd_q = &pt->cmd_q;
drivers/dma/amd/ptdma/ptdma-dev.c
284
pt_dmaengine_unregister(pt);
drivers/dma/amd/ptdma/ptdma-dev.c
287
pt_core_disable_queue_interrupts(pt);
drivers/dma/amd/ptdma/ptdma-dev.c
297
free_irq(pt->pt_irq, pt);
drivers/dma/amd/ptdma/ptdma-dev.c
303
while (!list_empty(&pt->cmd)) {
drivers/dma/amd/ptdma/ptdma-dev.c
305
cmd = list_first_entry(&pt->cmd, struct pt_cmd, entry);
drivers/dma/amd/ptdma/ptdma-dmaengine.c
103
if (pt->ver == AE4_DMA_VERSION) {
drivers/dma/amd/ptdma/ptdma-dmaengine.c
104
ae4 = container_of(pt, struct ae4_device, pt);
drivers/dma/amd/ptdma/ptdma-dmaengine.c
108
cmd_q = &pt->cmd_q;
drivers/dma/amd/ptdma/ptdma-dmaengine.c
164
struct pt_device *pt;
drivers/dma/amd/ptdma/ptdma-dmaengine.c
171
pt = pt_cmd->pt;
drivers/dma/amd/ptdma/ptdma-dmaengine.c
173
cmd_q = pt_get_cmd_queue(pt, chan);
drivers/dma/amd/ptdma/ptdma-dmaengine.c
177
pt->tdata.cmd = pt_cmd;
drivers/dma/amd/ptdma/ptdma-dmaengine.c
180
if (pt->ver == AE4_DMA_VERSION)
drivers/dma/amd/ptdma/ptdma-dmaengine.c
201
struct pt_device *pt;
drivers/dma/amd/ptdma/ptdma-dmaengine.c
204
pt = chan->pt;
drivers/dma/amd/ptdma/ptdma-dmaengine.c
222
if (pt->ver != AE4_DMA_VERSION && desc) {
drivers/dma/amd/ptdma/ptdma-dmaengine.c
240
if (pt->ver != AE4_DMA_VERSION && tx_desc) {
drivers/dma/amd/ptdma/ptdma-dmaengine.c
268
struct pt_device *pt;
drivers/dma/amd/ptdma/ptdma-dmaengine.c
276
pt = chan->pt;
drivers/dma/amd/ptdma/ptdma-dmaengine.c
282
if (pt->ver == AE4_DMA_VERSION) {
drivers/dma/amd/ptdma/ptdma-dmaengine.c
283
ae4 = container_of(pt, struct ae4_device, pt);
drivers/dma/amd/ptdma/ptdma-dmaengine.c
293
dev_err(pt->dev, "TIMEOUT %d:\n", ae4cmd_q->id);
drivers/dma/amd/ptdma/ptdma-dmaengine.c
322
desc = kmem_cache_zalloc(chan->pt->dma_desc_cache, GFP_NOWAIT);
drivers/dma/amd/ptdma/ptdma-dmaengine.c
328
desc->pt = chan->pt;
drivers/dma/amd/ptdma/ptdma-dmaengine.c
329
desc->pt->cmd_q.int_en = !!(flags & DMA_PREP_INTERRUPT);
drivers/dma/amd/ptdma/ptdma-dmaengine.c
388
struct pt_device *pt = chan->pt;
drivers/dma/amd/ptdma/ptdma-dmaengine.c
399
pt_cmd->pt = pt;
drivers/dma/amd/ptdma/ptdma-dmaengine.c
410
if (pt->ver == AE4_DMA_VERSION) {
drivers/dma/amd/ptdma/ptdma-dmaengine.c
412
ae4 = container_of(pt, struct ae4_device, pt);
drivers/dma/amd/ptdma/ptdma-dmaengine.c
452
struct pt_device *pt;
drivers/dma/amd/ptdma/ptdma-dmaengine.c
456
pt = chan->pt;
drivers/dma/amd/ptdma/ptdma-dmaengine.c
461
if (desc && pt->ver != AE4_DMA_VERSION)
drivers/dma/amd/ptdma/ptdma-dmaengine.c
475
static void pt_check_status_trans_ae4(struct pt_device *pt, struct pt_cmd_queue *cmd_q)
drivers/dma/amd/ptdma/ptdma-dmaengine.c
489
struct pt_device *pt = chan->pt;
drivers/dma/amd/ptdma/ptdma-dmaengine.c
492
cmd_q = pt_get_cmd_queue(pt, chan);
drivers/dma/amd/ptdma/ptdma-dmaengine.c
494
if (pt->ver == AE4_DMA_VERSION)
drivers/dma/amd/ptdma/ptdma-dmaengine.c
495
pt_check_status_trans_ae4(pt, cmd_q);
drivers/dma/amd/ptdma/ptdma-dmaengine.c
497
pt_check_status_trans(pt, cmd_q);
drivers/dma/amd/ptdma/ptdma-dmaengine.c
505
struct pt_device *pt = chan->pt;
drivers/dma/amd/ptdma/ptdma-dmaengine.c
510
cmd_q = pt_get_cmd_queue(pt, chan);
drivers/dma/amd/ptdma/ptdma-dmaengine.c
521
struct pt_device *pt = chan->pt;
drivers/dma/amd/ptdma/ptdma-dmaengine.c
526
cmd_q = pt_get_cmd_queue(pt, chan);
drivers/dma/amd/ptdma/ptdma-dmaengine.c
541
struct pt_device *pt = chan->pt;
drivers/dma/amd/ptdma/ptdma-dmaengine.c
546
cmd_q = pt_get_cmd_queue(pt, chan);
drivers/dma/amd/ptdma/ptdma-dmaengine.c
547
if (pt->ver == AE4_DMA_VERSION)
drivers/dma/amd/ptdma/ptdma-dmaengine.c
562
int pt_dmaengine_register(struct pt_device *pt)
drivers/dma/amd/ptdma/ptdma-dmaengine.c
564
struct dma_device *dma_dev = &pt->dma_dev;
drivers/dma/amd/ptdma/ptdma-dmaengine.c
571
if (pt->ver == AE4_DMA_VERSION)
drivers/dma/amd/ptdma/ptdma-dmaengine.c
572
ae4 = container_of(pt, struct ae4_device, pt);
drivers/dma/amd/ptdma/ptdma-dmaengine.c
575
pt->pt_dma_chan = devm_kcalloc(pt->dev, ae4->cmd_q_count,
drivers/dma/amd/ptdma/ptdma-dmaengine.c
576
sizeof(*pt->pt_dma_chan), GFP_KERNEL);
drivers/dma/amd/ptdma/ptdma-dmaengine.c
578
pt->pt_dma_chan = devm_kzalloc(pt->dev, sizeof(*pt->pt_dma_chan),
drivers/dma/amd/ptdma/ptdma-dmaengine.c
581
if (!pt->pt_dma_chan)
drivers/dma/amd/ptdma/ptdma-dmaengine.c
584
desc_cache_name = devm_kasprintf(pt->dev, GFP_KERNEL,
drivers/dma/amd/ptdma/ptdma-dmaengine.c
586
dev_name(pt->dev));
drivers/dma/amd/ptdma/ptdma-dmaengine.c
590
pt->dma_desc_cache = kmem_cache_create(desc_cache_name,
drivers/dma/amd/ptdma/ptdma-dmaengine.c
593
if (!pt->dma_desc_cache)
drivers/dma/amd/ptdma/ptdma-dmaengine.c
596
dma_dev->dev = pt->dev;
drivers/dma/amd/ptdma/ptdma-dmaengine.c
60
ae4_log_error(cmd_q->pt, cmd_q->cmd_error);
drivers/dma/amd/ptdma/ptdma-dmaengine.c
625
chan = pt->pt_dma_chan + i;
drivers/dma/amd/ptdma/ptdma-dmaengine.c
628
chan->pt = pt;
drivers/dma/amd/ptdma/ptdma-dmaengine.c
633
chan = pt->pt_dma_chan;
drivers/dma/amd/ptdma/ptdma-dmaengine.c
634
chan->pt = pt;
drivers/dma/amd/ptdma/ptdma-dmaengine.c
646
kmem_cache_destroy(pt->dma_desc_cache);
drivers/dma/amd/ptdma/ptdma-dmaengine.c
652
void pt_dmaengine_unregister(struct pt_device *pt)
drivers/dma/amd/ptdma/ptdma-dmaengine.c
654
struct dma_device *dma_dev = &pt->dma_dev;
drivers/dma/amd/ptdma/ptdma-dmaengine.c
658
kmem_cache_destroy(pt->dma_desc_cache);
drivers/dma/amd/ptdma/ptdma-dmaengine.c
92
struct pt_device *pt = desc->pt;
drivers/dma/amd/ptdma/ptdma-dmaengine.c
94
kmem_cache_free(pt->dma_desc_cache, desc);
drivers/dma/amd/ptdma/ptdma-dmaengine.c
97
static struct pt_cmd_queue *pt_get_cmd_queue(struct pt_device *pt, struct pt_dma_chan *chan)
drivers/dma/amd/ptdma/ptdma-pci.c
107
static void pt_free_irqs(struct pt_device *pt)
drivers/dma/amd/ptdma/ptdma-pci.c
109
struct pt_msix *pt_msix = pt->pt_msix;
drivers/dma/amd/ptdma/ptdma-pci.c
110
struct device *dev = pt->dev;
drivers/dma/amd/ptdma/ptdma-pci.c
115
else if (pt->pt_irq)
drivers/dma/amd/ptdma/ptdma-pci.c
118
pt->pt_irq = 0;
drivers/dma/amd/ptdma/ptdma-pci.c
123
struct pt_device *pt;
drivers/dma/amd/ptdma/ptdma-pci.c
130
pt = pt_alloc_struct(dev);
drivers/dma/amd/ptdma/ptdma-pci.c
131
if (!pt)
drivers/dma/amd/ptdma/ptdma-pci.c
138
pt->pt_msix = pt_msix;
drivers/dma/amd/ptdma/ptdma-pci.c
139
pt->dev_vdata = (struct pt_dev_vdata *)id->driver_data;
drivers/dma/amd/ptdma/ptdma-pci.c
140
if (!pt->dev_vdata) {
drivers/dma/amd/ptdma/ptdma-pci.c
166
pt->io_regs = iomap_table[pt->dev_vdata->bar];
drivers/dma/amd/ptdma/ptdma-pci.c
167
if (!pt->io_regs) {
drivers/dma/amd/ptdma/ptdma-pci.c
173
ret = pt_get_irqs(pt);
drivers/dma/amd/ptdma/ptdma-pci.c
189
dev_set_drvdata(dev, pt);
drivers/dma/amd/ptdma/ptdma-pci.c
191
if (pt->dev_vdata)
drivers/dma/amd/ptdma/ptdma-pci.c
192
ret = pt_core_init(pt);
drivers/dma/amd/ptdma/ptdma-pci.c
208
struct pt_device *pt = dev_get_drvdata(dev);
drivers/dma/amd/ptdma/ptdma-pci.c
210
if (!pt)
drivers/dma/amd/ptdma/ptdma-pci.c
213
if (pt->dev_vdata)
drivers/dma/amd/ptdma/ptdma-pci.c
214
pt_core_destroy(pt);
drivers/dma/amd/ptdma/ptdma-pci.c
216
pt_free_irqs(pt);
drivers/dma/amd/ptdma/ptdma-pci.c
38
struct pt_device *pt;
drivers/dma/amd/ptdma/ptdma-pci.c
40
pt = devm_kzalloc(dev, sizeof(*pt), GFP_KERNEL);
drivers/dma/amd/ptdma/ptdma-pci.c
42
if (!pt)
drivers/dma/amd/ptdma/ptdma-pci.c
44
pt->dev = dev;
drivers/dma/amd/ptdma/ptdma-pci.c
46
INIT_LIST_HEAD(&pt->cmd);
drivers/dma/amd/ptdma/ptdma-pci.c
48
return pt;
drivers/dma/amd/ptdma/ptdma-pci.c
51
static int pt_get_msix_irqs(struct pt_device *pt)
drivers/dma/amd/ptdma/ptdma-pci.c
53
struct pt_msix *pt_msix = pt->pt_msix;
drivers/dma/amd/ptdma/ptdma-pci.c
54
struct device *dev = pt->dev;
drivers/dma/amd/ptdma/ptdma-pci.c
66
pt->pt_irq = pt_msix->msix_entry.vector;
drivers/dma/amd/ptdma/ptdma-pci.c
71
static int pt_get_msi_irq(struct pt_device *pt)
drivers/dma/amd/ptdma/ptdma-pci.c
73
struct device *dev = pt->dev;
drivers/dma/amd/ptdma/ptdma-pci.c
81
pt->pt_irq = pdev->irq;
drivers/dma/amd/ptdma/ptdma-pci.c
86
static int pt_get_irqs(struct pt_device *pt)
drivers/dma/amd/ptdma/ptdma-pci.c
88
struct device *dev = pt->dev;
drivers/dma/amd/ptdma/ptdma-pci.c
91
ret = pt_get_msix_irqs(pt);
drivers/dma/amd/ptdma/ptdma-pci.c
97
ret = pt_get_msi_irq(pt);
drivers/dma/amd/ptdma/ptdma.h
165
struct pt_device *pt;
drivers/dma/amd/ptdma/ptdma.h
177
struct pt_device *pt;
drivers/dma/amd/ptdma/ptdma.h
186
struct pt_device *pt;
drivers/dma/amd/ptdma/ptdma.h
191
struct pt_device *pt;
drivers/dma/amd/ptdma/ptdma.h
315
int pt_dmaengine_register(struct pt_device *pt);
drivers/dma/amd/ptdma/ptdma.h
316
void pt_dmaengine_unregister(struct pt_device *pt);
drivers/dma/amd/ptdma/ptdma.h
318
void ptdma_debugfs_setup(struct pt_device *pt);
drivers/dma/amd/ptdma/ptdma.h
319
int pt_core_init(struct pt_device *pt);
drivers/dma/amd/ptdma/ptdma.h
320
void pt_core_destroy(struct pt_device *pt);
drivers/dma/amd/ptdma/ptdma.h
325
void pt_check_status_trans(struct pt_device *pt, struct pt_cmd_queue *cmd_q);
drivers/dma/amd/ptdma/ptdma.h
329
static inline void pt_core_disable_queue_interrupts(struct pt_device *pt)
drivers/dma/amd/ptdma/ptdma.h
331
iowrite32(0, pt->cmd_q.reg_control + 0x000C);
drivers/dma/amd/ptdma/ptdma.h
334
static inline void pt_core_enable_queue_interrupts(struct pt_device *pt)
drivers/dma/amd/ptdma/ptdma.h
336
iowrite32(SUPPORTED_INTERRUPTS, pt->cmd_q.reg_control + 0x000C);
drivers/firewire/core-cdev.c
1887
static __poll_t fw_device_op_poll(struct file *file, poll_table * pt)
drivers/firewire/core-cdev.c
1892
poll_wait(file, &client->wait, pt);
drivers/firewire/nosy.c
323
nosy_poll(struct file *file, poll_table *pt)
drivers/firewire/nosy.c
328
poll_wait(file, &client->buffer.wait, pt);
drivers/firmware/imx/rm.c
68
int imx_sc_rm_get_resource_owner(struct imx_sc_ipc *ipc, u16 resource, u8 *pt)
drivers/firmware/imx/rm.c
85
if (pt)
drivers/firmware/imx/rm.c
86
*pt = msg.data.resp.val;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
138
amdgpu_vm_pt_parent(struct amdgpu_vm_bo_base *pt)
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
140
struct amdgpu_bo *parent = pt->bo->parent;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
503
struct amdgpu_bo_vm *pt;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
510
r = amdgpu_vm_pt_create(adev, vm, cursor->level, immediate, &pt,
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
519
pt_bo = &pt->bo;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
522
r = amdgpu_vm_pt_clear(adev, vm, pt, immediate);
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
636
uint64_t pde, pt, flags;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
647
amdgpu_gmc_get_pde_for_bo(entry->bo, level, &pt, &flags);
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
649
return vm->update_funcs->update(params, to_amdgpu_bo_vm(bo), pde, pt,
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
680
struct amdgpu_bo_vm *pt,
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
719
params->vm->update_funcs->update(params, pt, pe, addr, count, incr,
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
817
struct amdgpu_bo *pt;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
859
pt = cursor.entry->bo;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
860
if (!pt) {
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
871
pt = cursor.entry->bo;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
909
amdgpu_vm_pte_update_flags(params, to_amdgpu_bo_vm(pt),
drivers/gpu/drm/amd/pm/legacy-dpm/legacy_dpm.c
557
ATOM_PowerTune_Table *pt;
drivers/gpu/drm/amd/pm/legacy-dpm/legacy_dpm.c
568
pt = &ppt->power_tune_table;
drivers/gpu/drm/amd/pm/legacy-dpm/legacy_dpm.c
574
pt = &ppt->power_tune_table;
drivers/gpu/drm/amd/pm/legacy-dpm/legacy_dpm.c
576
adev->pm.dpm.dyn_state.cac_tdp_table->tdp = le16_to_cpu(pt->usTDP);
drivers/gpu/drm/amd/pm/legacy-dpm/legacy_dpm.c
578
le16_to_cpu(pt->usConfigurableTDP);
drivers/gpu/drm/amd/pm/legacy-dpm/legacy_dpm.c
579
adev->pm.dpm.dyn_state.cac_tdp_table->tdc = le16_to_cpu(pt->usTDC);
drivers/gpu/drm/amd/pm/legacy-dpm/legacy_dpm.c
581
le16_to_cpu(pt->usBatteryPowerLimit);
drivers/gpu/drm/amd/pm/legacy-dpm/legacy_dpm.c
583
le16_to_cpu(pt->usSmallPowerLimit);
drivers/gpu/drm/amd/pm/legacy-dpm/legacy_dpm.c
585
le16_to_cpu(pt->usLowCACLeakage);
drivers/gpu/drm/amd/pm/legacy-dpm/legacy_dpm.c
587
le16_to_cpu(pt->usHighCACLeakage);
drivers/gpu/drm/drm_edid.c
3461
const struct detailed_pixel_timing *pt)
drivers/gpu/drm/drm_edid.c
3476
if (!(pt->misc & DRM_EDID_PT_INTERLACED))
drivers/gpu/drm/drm_edid.c
3504
const struct detailed_pixel_timing *pt = &timing->data.pixel_data;
drivers/gpu/drm/drm_edid.c
3505
unsigned hactive = (pt->hactive_hblank_hi & 0xf0) << 4 | pt->hactive_lo;
drivers/gpu/drm/drm_edid.c
3506
unsigned vactive = (pt->vactive_vblank_hi & 0xf0) << 4 | pt->vactive_lo;
drivers/gpu/drm/drm_edid.c
3507
unsigned hblank = (pt->hactive_hblank_hi & 0xf) << 8 | pt->hblank_lo;
drivers/gpu/drm/drm_edid.c
3508
unsigned vblank = (pt->vactive_vblank_hi & 0xf) << 8 | pt->vblank_lo;
drivers/gpu/drm/drm_edid.c
3509
unsigned hsync_offset = (pt->hsync_vsync_offset_pulse_width_hi & 0xc0) << 2 | pt->hsync_offset_lo;
drivers/gpu/drm/drm_edid.c
3510
unsigned hsync_pulse_width = (pt->hsync_vsync_offset_pulse_width_hi & 0x30) << 4 | pt->hsync_pulse_width_lo;
drivers/gpu/drm/drm_edid.c
3511
unsigned vsync_offset = (pt->hsync_vsync_offset_pulse_width_hi & 0xc) << 2 | pt->vsync_offset_pulse_width_lo >> 4;
drivers/gpu/drm/drm_edid.c
3512
unsigned vsync_pulse_width = (pt->hsync_vsync_offset_pulse_width_hi & 0x3) << 4 | (pt->vsync_offset_pulse_width_lo & 0xf);
drivers/gpu/drm/drm_edid.c
3518
if (pt->misc & DRM_EDID_PT_STEREO) {
drivers/gpu/drm/drm_edid.c
3523
if (!(pt->misc & DRM_EDID_PT_SEPARATE_SYNC)) {
drivers/gpu/drm/drm_edid.c
3576
drm_mode_do_interlace_quirk(mode, pt);
drivers/gpu/drm/drm_edid.c
3581
mode->flags |= (pt->misc & DRM_EDID_PT_HSYNC_POSITIVE) ?
drivers/gpu/drm/drm_edid.c
3583
mode->flags |= (pt->misc & DRM_EDID_PT_VSYNC_POSITIVE) ?
drivers/gpu/drm/drm_edid.c
3588
mode->width_mm = pt->width_mm_lo | (pt->width_height_mm_hi & 0xf0) << 4;
drivers/gpu/drm/drm_edid.c
3589
mode->height_mm = pt->height_mm_lo | (pt->width_height_mm_hi & 0xf) << 8;
drivers/gpu/drm/drm_mode_config.c
524
struct drm_property *property, *pt;
drivers/gpu/drm/drm_mode_config.c
551
list_for_each_entry_safe(property, pt, &dev->mode_config.property_list,
drivers/gpu/drm/drm_modes.c
1889
struct drm_display_mode *pmode, *pt;
drivers/gpu/drm/drm_modes.c
1893
list_for_each_entry_safe(pmode, pt, &connector->probed_modes, head) {
drivers/gpu/drm/drm_property.c
443
struct drm_property_enum *prop_enum, *pt;
drivers/gpu/drm/drm_property.c
445
list_for_each_entry_safe(prop_enum, pt, &property->enum_list, head) {
drivers/gpu/drm/gma500/mid_bios.c
181
dev_priv->gct_data.pt = gct.PD.PanelType;
drivers/gpu/drm/gma500/mid_bios.c
209
dev_priv->gct_data.pt = gct.PD.PanelType;
drivers/gpu/drm/gma500/mmu.c
224
static void psb_mmu_free_pt(struct psb_mmu_pt *pt)
drivers/gpu/drm/gma500/mmu.c
226
__free_page(pt->p);
drivers/gpu/drm/gma500/mmu.c
227
kfree(pt);
drivers/gpu/drm/gma500/mmu.c
235
struct psb_mmu_pt *pt;
drivers/gpu/drm/gma500/mmu.c
248
pt = pd->tables[i];
drivers/gpu/drm/gma500/mmu.c
249
if (pt)
drivers/gpu/drm/gma500/mmu.c
250
psb_mmu_free_pt(pt);
drivers/gpu/drm/gma500/mmu.c
263
struct psb_mmu_pt *pt = kmalloc_obj(*pt);
drivers/gpu/drm/gma500/mmu.c
272
if (!pt)
drivers/gpu/drm/gma500/mmu.c
275
pt->p = alloc_page(GFP_DMA32);
drivers/gpu/drm/gma500/mmu.c
276
if (!pt->p) {
drivers/gpu/drm/gma500/mmu.c
277
kfree(pt);
drivers/gpu/drm/gma500/mmu.c
283
v = kmap_atomic(pt->p);
drivers/gpu/drm/gma500/mmu.c
300
pt->count = 0;
drivers/gpu/drm/gma500/mmu.c
301
pt->pd = pd;
drivers/gpu/drm/gma500/mmu.c
302
pt->index = 0;
drivers/gpu/drm/gma500/mmu.c
304
return pt;
drivers/gpu/drm/gma500/mmu.c
311
struct psb_mmu_pt *pt;
drivers/gpu/drm/gma500/mmu.c
316
pt = pd->tables[index];
drivers/gpu/drm/gma500/mmu.c
317
while (!pt) {
drivers/gpu/drm/gma500/mmu.c
319
pt = psb_mmu_alloc_pt(pd);
drivers/gpu/drm/gma500/mmu.c
320
if (!pt)
drivers/gpu/drm/gma500/mmu.c
326
psb_mmu_free_pt(pt);
drivers/gpu/drm/gma500/mmu.c
328
pt = pd->tables[index];
drivers/gpu/drm/gma500/mmu.c
333
pd->tables[index] = pt;
drivers/gpu/drm/gma500/mmu.c
334
v[index] = (page_to_pfn(pt->p) << 12) | pd->pd_mask;
drivers/gpu/drm/gma500/mmu.c
335
pt->index = index;
drivers/gpu/drm/gma500/mmu.c
343
pt->v = kmap_atomic(pt->p);
drivers/gpu/drm/gma500/mmu.c
344
return pt;
drivers/gpu/drm/gma500/mmu.c
351
struct psb_mmu_pt *pt;
drivers/gpu/drm/gma500/mmu.c
355
pt = pd->tables[index];
drivers/gpu/drm/gma500/mmu.c
356
if (!pt) {
drivers/gpu/drm/gma500/mmu.c
360
pt->v = kmap_atomic(pt->p);
drivers/gpu/drm/gma500/mmu.c
361
return pt;
drivers/gpu/drm/gma500/mmu.c
364
static void psb_mmu_pt_unmap_unlock(struct psb_mmu_pt *pt)
drivers/gpu/drm/gma500/mmu.c
366
struct psb_mmu_pd *pd = pt->pd;
drivers/gpu/drm/gma500/mmu.c
369
kunmap_atomic(pt->v);
drivers/gpu/drm/gma500/mmu.c
370
if (pt->count == 0) {
drivers/gpu/drm/gma500/mmu.c
372
v[pt->index] = pd->invalid_pde;
drivers/gpu/drm/gma500/mmu.c
373
pd->tables[pt->index] = NULL;
drivers/gpu/drm/gma500/mmu.c
376
psb_mmu_clflush(pd->driver, (void *)&v[pt->index]);
drivers/gpu/drm/gma500/mmu.c
381
psb_mmu_free_pt(pt);
drivers/gpu/drm/gma500/mmu.c
387
static inline void psb_mmu_set_pte(struct psb_mmu_pt *pt, unsigned long addr,
drivers/gpu/drm/gma500/mmu.c
390
pt->v[psb_mmu_pt_index(addr)] = pte;
drivers/gpu/drm/gma500/mmu.c
393
static inline void psb_mmu_invalidate_pte(struct psb_mmu_pt *pt,
drivers/gpu/drm/gma500/mmu.c
396
pt->v[psb_mmu_pt_index(addr)] = pt->pd->invalid_pte;
drivers/gpu/drm/gma500/mmu.c
482
struct psb_mmu_pt *pt;
drivers/gpu/drm/gma500/mmu.c
511
pt = psb_mmu_pt_map_lock(pd, addr);
drivers/gpu/drm/gma500/mmu.c
512
if (!pt)
drivers/gpu/drm/gma500/mmu.c
515
psb_clflush(&pt->v[psb_mmu_pt_index(addr)]);
drivers/gpu/drm/gma500/mmu.c
519
psb_mmu_pt_unmap_unlock(pt);
drivers/gpu/drm/gma500/mmu.c
529
struct psb_mmu_pt *pt;
drivers/gpu/drm/gma500/mmu.c
542
pt = psb_mmu_pt_alloc_map_lock(pd, addr);
drivers/gpu/drm/gma500/mmu.c
543
if (!pt)
drivers/gpu/drm/gma500/mmu.c
546
psb_mmu_invalidate_pte(pt, addr);
drivers/gpu/drm/gma500/mmu.c
547
--pt->count;
drivers/gpu/drm/gma500/mmu.c
549
psb_mmu_pt_unmap_unlock(pt);
drivers/gpu/drm/gma500/mmu.c
569
struct psb_mmu_pt *pt;
drivers/gpu/drm/gma500/mmu.c
598
pt = psb_mmu_pt_map_lock(pd, addr);
drivers/gpu/drm/gma500/mmu.c
599
if (!pt)
drivers/gpu/drm/gma500/mmu.c
602
psb_mmu_invalidate_pte(pt, addr);
drivers/gpu/drm/gma500/mmu.c
603
--pt->count;
drivers/gpu/drm/gma500/mmu.c
606
psb_mmu_pt_unmap_unlock(pt);
drivers/gpu/drm/gma500/mmu.c
625
struct psb_mmu_pt *pt;
drivers/gpu/drm/gma500/mmu.c
640
pt = psb_mmu_pt_alloc_map_lock(pd, addr);
drivers/gpu/drm/gma500/mmu.c
641
if (!pt) {
drivers/gpu/drm/gma500/mmu.c
647
psb_mmu_set_pte(pt, addr, pte);
drivers/gpu/drm/gma500/mmu.c
648
pt->count++;
drivers/gpu/drm/gma500/mmu.c
650
psb_mmu_pt_unmap_unlock(pt);
drivers/gpu/drm/gma500/mmu.c
672
struct psb_mmu_pt *pt;
drivers/gpu/drm/gma500/mmu.c
704
pt = psb_mmu_pt_alloc_map_lock(pd, addr);
drivers/gpu/drm/gma500/mmu.c
705
if (!pt)
drivers/gpu/drm/gma500/mmu.c
710
psb_mmu_set_pte(pt, addr, pte);
drivers/gpu/drm/gma500/mmu.c
711
pt->count++;
drivers/gpu/drm/gma500/mmu.c
713
psb_mmu_pt_unmap_unlock(pt);
drivers/gpu/drm/gma500/oaktrail.h
192
u8 pt; /* panel type, 4 bit field, 0=lvds, 1=mipi */
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
103
vaddr = px_vaddr(pt);
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
153
struct i915_page_table *pt;
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
161
gen6_for_each_pde(pt, pd, start, end, pde)
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
162
gen6_write_pde(ppgtt, pde, pt);
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
178
struct i915_page_table *pt;
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
184
gen6_for_each_pde(pt, pd, start, length, pde) {
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
187
if (!pt) {
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
190
pt = stash->pt[0];
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
191
__i915_gem_object_pin_pages(pt->base);
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
193
fill32_px(pt, vm->scratch[0]->encode);
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
197
stash->pt[0] = pt->stash;
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
198
atomic_set(&pt->used, 0);
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
199
pd->entry[pde] = pt;
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
201
pt = pd->entry[pde];
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
207
atomic_add(count, &pt->used);
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
21
const struct i915_page_table *pt)
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
23
dma_addr_t addr = pt ? px_dma(pt) : px_dma(ppgtt->base.vm.scratch[1]);
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
259
struct i915_page_table *pt;
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
262
gen6_for_all_pdes(pt, pd, pde)
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
263
if (pt)
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
264
free_pt(&ppgtt->base.vm, pt);
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
301
struct i915_page_table *pt;
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
308
gen6_for_all_pdes(pt, ppgtt->base.pd, pde) {
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
309
if (!pt || atomic_read(&pt->used))
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
312
free_pt(&ppgtt->base.vm, pt);
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
390
pd->pt.base = __i915_gem_object_create_internal(ppgtt->base.vm.gt->i915,
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
393
if (IS_ERR(pd->pt.base)) {
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
394
err = PTR_ERR(pd->pt.base);
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
395
pd->pt.base = NULL;
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
399
pd->pt.base->base.resv = i915_vm_resv_get(&ppgtt->base.vm);
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
400
pd->pt.base->shares_resv_from = &ppgtt->base.vm;
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
402
ppgtt->vma = i915_vma_instance(pd->pt.base, &ggtt->vm, NULL);
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
85
struct i915_page_table * const pt =
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
92
GEM_BUG_ON(count > atomic_read(&pt->used));
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
93
if (!atomic_sub_return(count, &pt->used))
drivers/gpu/drm/i915/gt/gen6_ppgtt.h
57
#define gen6_for_each_pde(pt, pd, start, length, iter) \
drivers/gpu/drm/i915/gt/gen6_ppgtt.h
60
(pt = i915_pt_entry(pd, iter), true); \
drivers/gpu/drm/i915/gt/gen6_ppgtt.h
65
#define gen6_for_all_pdes(pt, pd, iter) \
drivers/gpu/drm/i915/gt/gen6_ppgtt.h
68
(pt = i915_pt_entry(pd, iter), true); \
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
145
#define as_pd(x) container_of((x), typeof(struct i915_page_directory), pt)
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
219
free_px(vm, &pd->pt, lvl);
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
255
struct i915_page_table *pt = pd->entry[idx];
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
257
if (atomic_fetch_inc(&pt->used) >> gen8_pd_shift(1) &&
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
262
__gen8_ppgtt_cleanup(vm, as_pd(pt), I915_PDES, lvl);
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
268
start = __gen8_ppgtt_clear(vm, as_pd(pt),
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
280
atomic_read(&pt->used));
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
281
GEM_BUG_ON(!count || count >= atomic_read(&pt->used));
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
284
if (pt->is_compact) {
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
291
vaddr = px_vaddr(pt);
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
296
atomic_sub(count, &pt->used);
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
300
if (release_pd_entry(pd, idx, pt, scratch))
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
301
free_px(vm, pt, lvl);
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
340
struct i915_page_table *pt = pd->entry[idx];
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
342
if (!pt) {
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
348
pt = stash->pt[!!lvl];
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
349
__i915_gem_object_pin_pages(pt->base);
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
351
fill_px(pt, vm->scratch[lvl]->encode);
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
355
stash->pt[!!lvl] = pt->stash;
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
356
atomic_set(&pt->used, 0);
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
357
set_pd_entry(pd, idx, pt);
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
359
pt = pd->entry[idx];
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
364
atomic_inc(&pt->used);
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
368
as_pd(pt), start, end, lvl);
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
371
atomic_dec(&pt->used);
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
372
GEM_BUG_ON(!atomic_read(&pt->used));
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
379
atomic_read(&pt->used));
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
381
atomic_add(count, &pt->used);
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
383
GEM_BUG_ON(atomic_read(&pt->used) > NALLOC * I915_PDES);
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
410
struct i915_page_table *pt,
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
420
struct i915_page_table *pt = pd->entry[idx];
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
422
atomic_inc(&pt->used);
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
426
__gen8_ppgtt_foreach(vm, as_pd(pt), start, end, lvl,
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
429
fn(vm, pt, data);
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
434
atomic_dec(&pt->used);
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
442
struct i915_page_table *pt,
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
521
struct i915_page_table *pt =
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
564
pt->is_compact = true;
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
574
vaddr = px_vaddr(pt);
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
644
struct i915_page_table *pt =
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
657
vaddr = px_vaddr(pt);
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
774
struct i915_page_table *pt = i915_pt_entry(pd, gen8_pd_index(idx, 1));
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
777
GEM_BUG_ON(pt->is_compact);
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
779
vaddr = px_vaddr(pt);
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
795
struct i915_page_table *pt = i915_pt_entry(pd, gen8_pd_index(idx, 1));
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
803
if (!pt->is_compact) {
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
806
pt->is_compact = true;
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
809
vaddr = px_vaddr(pt);
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
909
err = map_pt_dma(vm, pde->pt.base);
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
937
pd->pt.base = vm->alloc_pt_dma(vm, I915_GTT_PAGE_SIZE_4K);
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
938
if (IS_ERR(pd->pt.base)) {
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
939
err = PTR_ERR(pd->pt.base);
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
940
pd->pt.base = NULL;
drivers/gpu/drm/i915/gt/gen8_ppgtt.c
944
err = map_pt_dma(vm, pd->pt.base);
drivers/gpu/drm/i915/gt/intel_gtt.c
206
return i915_gem_object_lock(ppgtt->pd->pt.base, ww);
drivers/gpu/drm/i915/gt/intel_gtt.h
187
struct i915_page_table pt;
drivers/gpu/drm/i915/gt/intel_gtt.h
202
__px_choose_expr(px, struct i915_page_directory *, __x->pt.base, \
drivers/gpu/drm/i915/gt/intel_gtt.h
214
__px_choose_expr(px, struct i915_page_directory *, &__x->pt, \
drivers/gpu/drm/i915/gt/intel_gtt.h
220
struct i915_page_table *pt[2];
drivers/gpu/drm/i915/gt/intel_gtt.h
351
struct i915_page_table *pt,
drivers/gpu/drm/i915/gt/intel_gtt.h
581
struct i915_page_table *pt = ppgtt->pd->entry[n];
drivers/gpu/drm/i915/gt/intel_gtt.h
583
return __px_dma(pt ? px_base(pt) : ppgtt->vm.scratch[ppgtt->vm.top]);
drivers/gpu/drm/i915/gt/intel_gtt.h
644
struct i915_page_table *pt, int lvl);
drivers/gpu/drm/i915/gt/intel_gtt.h
651
struct i915_page_table *pt,
drivers/gpu/drm/i915/gt/intel_gtt.h
665
struct i915_page_table * const pt,
drivers/gpu/drm/i915/gt/intel_migrate.c
39
struct i915_page_table *pt,
drivers/gpu/drm/i915/gt/intel_migrate.c
51
GEM_BUG_ON(!pt->is_compact);
drivers/gpu/drm/i915/gt/intel_migrate.c
56
struct i915_page_table *pt,
drivers/gpu/drm/i915/gt/intel_migrate.c
68
vm->insert_page(vm, px_dma(pt), d->offset,
drivers/gpu/drm/i915/gt/intel_migrate.c
75
struct i915_page_table *pt,
drivers/gpu/drm/i915/gt/intel_migrate.c
80
vm->insert_page(vm, px_dma(pt), d->offset,
drivers/gpu/drm/i915/gt/intel_migrate.c
82
i915_gem_object_is_lmem(pt->base) ? PTE_LM : 0);
drivers/gpu/drm/i915/gt/intel_ppgtt.c
127
struct i915_page_table * const pt,
drivers/gpu/drm/i915/gt/intel_ppgtt.c
132
if (atomic_add_unless(&pt->used, -1, 1))
drivers/gpu/drm/i915/gt/intel_ppgtt.c
136
if (atomic_dec_and_test(&pt->used)) {
drivers/gpu/drm/i915/gt/intel_ppgtt.c
18
struct i915_page_table *pt;
drivers/gpu/drm/i915/gt/intel_ppgtt.c
20
pt = kmalloc_obj(*pt, I915_GFP_ALLOW_FAIL);
drivers/gpu/drm/i915/gt/intel_ppgtt.c
21
if (unlikely(!pt))
drivers/gpu/drm/i915/gt/intel_ppgtt.c
24
pt->base = vm->alloc_pt_dma(vm, sz);
drivers/gpu/drm/i915/gt/intel_ppgtt.c
244
struct i915_page_table *pt;
drivers/gpu/drm/i915/gt/intel_ppgtt.c
246
pt = alloc_pt(vm, pt_sz);
drivers/gpu/drm/i915/gt/intel_ppgtt.c
247
if (IS_ERR(pt)) {
drivers/gpu/drm/i915/gt/intel_ppgtt.c
249
return PTR_ERR(pt);
drivers/gpu/drm/i915/gt/intel_ppgtt.c
25
if (IS_ERR(pt->base)) {
drivers/gpu/drm/i915/gt/intel_ppgtt.c
252
pt->stash = stash->pt[0];
drivers/gpu/drm/i915/gt/intel_ppgtt.c
253
stash->pt[0] = pt;
drivers/gpu/drm/i915/gt/intel_ppgtt.c
26
kfree(pt);
drivers/gpu/drm/i915/gt/intel_ppgtt.c
268
pd->pt.stash = stash->pt[1];
drivers/gpu/drm/i915/gt/intel_ppgtt.c
269
stash->pt[1] = &pd->pt;
drivers/gpu/drm/i915/gt/intel_ppgtt.c
279
struct i915_page_table *pt;
drivers/gpu/drm/i915/gt/intel_ppgtt.c
282
for (n = 0; n < ARRAY_SIZE(stash->pt); n++) {
drivers/gpu/drm/i915/gt/intel_ppgtt.c
283
for (pt = stash->pt[n]; pt; pt = pt->stash) {
drivers/gpu/drm/i915/gt/intel_ppgtt.c
284
err = map_pt_dma_locked(vm, pt->base);
drivers/gpu/drm/i915/gt/intel_ppgtt.c
296
struct i915_page_table *pt;
drivers/gpu/drm/i915/gt/intel_ppgtt.c
299
for (n = 0; n < ARRAY_SIZE(stash->pt); n++) {
drivers/gpu/drm/i915/gt/intel_ppgtt.c
30
pt->is_compact = false;
drivers/gpu/drm/i915/gt/intel_ppgtt.c
300
while ((pt = stash->pt[n])) {
drivers/gpu/drm/i915/gt/intel_ppgtt.c
301
stash->pt[n] = pt->stash;
drivers/gpu/drm/i915/gt/intel_ppgtt.c
302
free_px(vm, pt, n);
drivers/gpu/drm/i915/gt/intel_ppgtt.c
31
atomic_set(&pt->used, 0);
drivers/gpu/drm/i915/gt/intel_ppgtt.c
32
return pt;
drivers/gpu/drm/i915/gt/intel_ppgtt.c
61
pd->pt.base = vm->alloc_pt_dma(vm, I915_GTT_PAGE_SIZE_4K);
drivers/gpu/drm/i915/gt/intel_ppgtt.c
62
if (IS_ERR(pd->pt.base)) {
drivers/gpu/drm/i915/gt/intel_ppgtt.c
71
void free_px(struct i915_address_space *vm, struct i915_page_table *pt, int lvl)
drivers/gpu/drm/i915/gt/intel_ppgtt.c
73
BUILD_BUG_ON(offsetof(struct i915_page_directory, pt));
drivers/gpu/drm/i915/gt/intel_ppgtt.c
77
container_of(pt, typeof(*pd), pt);
drivers/gpu/drm/i915/gt/intel_ppgtt.c
81
if (pt->base)
drivers/gpu/drm/i915/gt/intel_ppgtt.c
82
i915_gem_object_put(pt->base);
drivers/gpu/drm/i915/gt/intel_ppgtt.c
84
kfree(pt);
drivers/gpu/drm/i915/gvt/gtt.c
240
static inline int gtt_get_entry64(void *pt,
drivers/gpu/drm/i915/gvt/gtt.c
257
} else if (!pt) {
drivers/gpu/drm/i915/gvt/gtt.c
260
e->val64 = *((u64 *)pt + index);
drivers/gpu/drm/i915/gvt/gtt.c
265
static inline int gtt_set_entry64(void *pt,
drivers/gpu/drm/i915/gvt/gtt.c
282
} else if (!pt) {
drivers/gpu/drm/i915/gvt/gtt.c
285
*((u64 *)pt + index) = e->val64;
drivers/gpu/drm/i915/gvt/gtt.h
58
int (*get_entry)(void *pt,
drivers/gpu/drm/i915/gvt/gtt.h
64
int (*set_entry)(void *pt,
drivers/gpu/drm/i915/gvt/mmio.c
75
void *pt = NULL;
drivers/gpu/drm/i915/gvt/mmio.c
93
pt = vgpu->gtt.ggtt_mm->ggtt_mm.virtual_ggtt + offset;
drivers/gpu/drm/i915/gvt/mmio.c
95
memcpy(p_data, pt, bytes);
drivers/gpu/drm/i915/gvt/mmio.c
97
memcpy(pt, p_data, bytes);
drivers/gpu/drm/i915/gvt/scheduler.c
436
struct scatterlist *sg = pd->pt.base->mm.pages->sgl;
drivers/gpu/drm/i915/selftests/scatterlist.c
103
if (pfn != pt->end) {
drivers/gpu/drm/i915/selftests/scatterlist.c
105
__func__, who, pt->end, pfn);
drivers/gpu/drm/i915/selftests/scatterlist.c
112
static noinline int expect_pfn_sgtiter(struct pfn_table *pt,
drivers/gpu/drm/i915/selftests/scatterlist.c
120
pfn = pt->start;
drivers/gpu/drm/i915/selftests/scatterlist.c
121
for_each_sgt_page(page, sgt, &pt->st) {
drivers/gpu/drm/i915/selftests/scatterlist.c
133
if (pfn != pt->end) {
drivers/gpu/drm/i915/selftests/scatterlist.c
135
__func__, who, pt->end, pfn);
drivers/gpu/drm/i915/selftests/scatterlist.c
142
static int expect_pfn_sgtable(struct pfn_table *pt,
drivers/gpu/drm/i915/selftests/scatterlist.c
150
err = expect_pfn_sg(pt, npages_fn, rnd, who, timeout);
drivers/gpu/drm/i915/selftests/scatterlist.c
154
err = expect_pfn_sg_page_iter(pt, who, timeout);
drivers/gpu/drm/i915/selftests/scatterlist.c
158
err = expect_pfn_sgtiter(pt, who, timeout);
drivers/gpu/drm/i915/selftests/scatterlist.c
214
static int alloc_table(struct pfn_table *pt,
drivers/gpu/drm/i915/selftests/scatterlist.c
227
if (sg_alloc_table(&pt->st, max,
drivers/gpu/drm/i915/selftests/scatterlist.c
238
pt->start = PFN_BIAS;
drivers/gpu/drm/i915/selftests/scatterlist.c
239
pfn = pt->start;
drivers/gpu/drm/i915/selftests/scatterlist.c
240
sg = pt->st.sgl;
drivers/gpu/drm/i915/selftests/scatterlist.c
248
sg_free_table(&pt->st);
drivers/gpu/drm/i915/selftests/scatterlist.c
263
pt->st.nents = n;
drivers/gpu/drm/i915/selftests/scatterlist.c
264
pt->end = pfn;
drivers/gpu/drm/i915/selftests/scatterlist.c
293
struct pfn_table pt;
drivers/gpu/drm/i915/selftests/scatterlist.c
299
err = alloc_table(&pt, sz, sz, *npages, &prng,
drivers/gpu/drm/i915/selftests/scatterlist.c
308
err = expect_pfn_sgtable(&pt, *npages, &prng,
drivers/gpu/drm/i915/selftests/scatterlist.c
311
sg_free_table(&pt.st);
drivers/gpu/drm/i915/selftests/scatterlist.c
329
struct pfn_table pt;
drivers/gpu/drm/i915/selftests/scatterlist.c
341
err = alloc_table(&pt, prime, max, *npages, &prng,
drivers/gpu/drm/i915/selftests/scatterlist.c
348
if (i915_sg_trim(&pt.st)) {
drivers/gpu/drm/i915/selftests/scatterlist.c
349
if (pt.st.orig_nents != prime ||
drivers/gpu/drm/i915/selftests/scatterlist.c
350
pt.st.nents != prime) {
drivers/gpu/drm/i915/selftests/scatterlist.c
352
pt.st.nents, pt.st.orig_nents, prime);
drivers/gpu/drm/i915/selftests/scatterlist.c
357
err = expect_pfn_sgtable(&pt,
drivers/gpu/drm/i915/selftests/scatterlist.c
363
sg_free_table(&pt.st);
drivers/gpu/drm/i915/selftests/scatterlist.c
41
static noinline int expect_pfn_sg(struct pfn_table *pt,
drivers/gpu/drm/i915/selftests/scatterlist.c
50
pfn = pt->start;
drivers/gpu/drm/i915/selftests/scatterlist.c
51
for_each_sg(pt->st.sgl, sg, pt->st.nents, n) {
drivers/gpu/drm/i915/selftests/scatterlist.c
53
unsigned int npages = npages_fn(n, pt->st.nents, rnd);
drivers/gpu/drm/i915/selftests/scatterlist.c
72
if (pfn != pt->end) {
drivers/gpu/drm/i915/selftests/scatterlist.c
74
__func__, who, pt->end, pfn);
drivers/gpu/drm/i915/selftests/scatterlist.c
81
static noinline int expect_pfn_sg_page_iter(struct pfn_table *pt,
drivers/gpu/drm/i915/selftests/scatterlist.c
88
pfn = pt->start;
drivers/gpu/drm/i915/selftests/scatterlist.c
89
for_each_sg_page(pt->st.sgl, &sgiter, pt->st.nents, 0) {
drivers/gpu/drm/imagination/pvr_fw_mips.h
28
u32 *pt;
drivers/gpu/drm/imagination/pvr_vm_mips.c
104
vunmap(mips_data->pt);
drivers/gpu/drm/imagination/pvr_vm_mips.c
196
WRITE_ONCE(mips_data->pt[pfn], pte);
drivers/gpu/drm/imagination/pvr_vm_mips.c
205
WRITE_ONCE(mips_data->pt[pfn], 0);
drivers/gpu/drm/imagination/pvr_vm_mips.c
233
WRITE_ONCE(mips_data->pt[pfn], 0);
drivers/gpu/drm/imagination/pvr_vm_mips.c
66
mips_data->pt = vmap(mips_data->pt_pages, pt_size >> PAGE_SHIFT, VM_MAP,
drivers/gpu/drm/imagination/pvr_vm_mips.c
68
if (!mips_data->pt) {
drivers/gpu/drm/lima/lima_vm.c
257
u32 *pd, *pt;
drivers/gpu/drm/lima/lima_vm.c
267
pt = vm->bts[i].cpu;
drivers/gpu/drm/lima/lima_vm.c
274
u32 pte = *pt++;
drivers/gpu/drm/loongson/lsdc_pixpll.c
155
const struct clk_to_pixpll_parms_lookup_t *pt;
drivers/gpu/drm/loongson/lsdc_pixpll.c
159
pt = &pixpll_parms_table[i];
drivers/gpu/drm/loongson/lsdc_pixpll.c
161
if (clock == pt->clock) {
drivers/gpu/drm/loongson/lsdc_pixpll.c
162
pout->div_ref = pt->div_ref;
drivers/gpu/drm/loongson/lsdc_pixpll.c
163
pout->loopc = pt->loopc;
drivers/gpu/drm/loongson/lsdc_pixpll.c
164
pout->div_out = pt->div_out;
drivers/gpu/drm/nouveau/nvkm/engine/dma/usernv04.c
53
device->mmu->vmm->pd->pt[0]->memory;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/bar.c
140
nvkm_memory_unref(&vmm->pd->pt[0]->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/bar.c
142
ret = nvkm_memory_kmap(pd3, &vmm->pd->pt[0]->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/bar.c
147
vmm->pd->pt[0]->addr = nvkm_memory_addr(vmm->pd->pt[0]->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/bar.c
86
struct nvkm_memory *pdb = vmm->pd->pt[0]->memory;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/bar.c
87
u32 pdb_offset = vmm->pd->pt[0]->base;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
123
ctrl->levels[i].physAddress = pd->pt[0]->addr;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
141
ctrl->physAddress = vmm->pd->pt[0]->addr;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
106
pt->memory = pt->ptp->pt->memory;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
107
pt->base = slot << ptp->shift;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
108
pt->addr = pt->ptp->pt->addr + pt->base;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
109
return pt;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
143
struct nvkm_mmu_pt *pt = *ppt;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
144
if (pt) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
146
if (pt->sub) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
148
nvkm_mmu_ptp_put(mmu, force, pt);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
155
if (pt->ptc->refs < 8 /* Heuristic. */ && !force) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
156
list_add_tail(&pt->head, &pt->ptc->item);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
157
pt->ptc->refs++;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
159
nvkm_memory_unref(&pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
160
kfree(pt);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
170
struct nvkm_mmu_pt *pt;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
176
pt = nvkm_mmu_ptp_get(mmu, align, zero);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
178
return pt;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
190
pt = list_first_entry_or_null(&ptc->item, typeof(*pt), head);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
191
if (pt) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
193
nvkm_fo64(pt->memory, 0, 0, size >> 3);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
194
list_del(&pt->head);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
197
return pt;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
202
if (!(pt = kmalloc_obj(*pt)))
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
204
pt->ptc = ptc;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
205
pt->sub = false;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
208
size, align, zero, &pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
210
kfree(pt);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
214
pt->base = 0;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
215
pt->addr = nvkm_memory_addr(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
216
return pt;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
224
struct nvkm_mmu_pt *pt, *tt;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
225
list_for_each_entry_safe(pt, tt, &ptc->item, head) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
226
nvkm_memory_unref(&pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
227
list_del(&pt->head);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
228
kfree(pt);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
34
struct nvkm_mmu_pt *pt;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
42
nvkm_mmu_ptp_put(struct nvkm_mmu *mmu, bool force, struct nvkm_mmu_pt *pt)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
44
const int slot = pt->base >> pt->ptp->shift;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
45
struct nvkm_mmu_ptp *ptp = pt->ptp;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
56
nvkm_mmu_ptc_put(mmu, force, &ptp->pt);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
61
kfree(pt);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
67
struct nvkm_mmu_pt *pt;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
71
if (!(pt = kzalloc_obj(*pt)))
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
78
kfree(pt);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
82
ptp->pt = nvkm_mmu_ptc_get(mmu, 0x1000, 0x1000, false);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
83
if (!ptp->pt) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
85
kfree(pt);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
90
slot = nvkm_memory_size(ptp->pt->memory) >> ptp->shift;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
95
pt->ptp = ptp;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c
96
pt->sub = true;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/nv41.c
35
nvkm_wr32(device, 0x100800, 0x00000002 | mmu->vmm->pd->pt[0]->addr);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/nv44.c
35
struct nvkm_memory *pt = mmu->vmm->pd->pt[0]->memory;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/nv44.c
43
addr -= ((nvkm_memory_addr(pt) >> 19) + 1) << 19;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/nv44.c
47
nvkm_wr32(device, 0x100804, (nvkm_memory_size(pt) / 4) * 4096);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1090
nvkm_mmu_ptc_put(vmm->mmu, true, &vmm->pd->pt[0]);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1161
vmm->pd->pt[0] = nvkm_mmu_ptc_get(mmu, size, desc->align, true);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1162
if (!vmm->pd->pt[0])
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
142
struct nvkm_vmm_pt *pgd = it->pt[it->lvl + 1];
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
143
struct nvkm_vmm_pt *pgt = it->pt[it->lvl];
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
144
struct nvkm_mmu_pt *pt = pgt->pt[type];
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
154
pgt->pt[type] = NULL;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
157
if (pgd->pt[0]) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
159
func->sparse(vmm, pgd->pt[0], pdei, 1);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
162
func->unmap(vmm, pgd->pt[0], pdei, 1);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
190
nvkm_mmu_ptc_put(vmm->mmu, vmm->bootstrapped, &pt);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1939
nvkm_memory_boot(it->pt[0]->pt[type]->memory, it->vmm);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
244
pair->func->sparse(vmm, pgt->pt[0], pteb, ptes);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
252
pair->func->invalid(vmm, pgt->pt[0], pteb, ptes);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
265
struct nvkm_vmm_pt *pgt = it->pt[0];
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
270
dma = desc->func->pfn_clear(it->vmm, pgt->pt[type], ptei, ptes);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
275
desc->func->pfn_unmap(it->vmm, pgt->pt[type], ptei, ptes);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
362
desc->func->sparse(vmm, pgt->pt[1], spti, sptc);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
365
pair->func->unmap(vmm, pgt->pt[0], pteb, ptes);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
372
pair->func->unmap(vmm, pgt->pt[0], pteb, ptes);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
382
struct nvkm_vmm_pt *pgt = it->pt[0];
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
420
struct nvkm_vmm_pt *pt = it->pt[0];
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
422
memset(&pt->pde[ptei], 0x00, sizeof(pt->pde[0]) * ptes);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
425
memset32(&pt->pte[ptei].u, 0x00, ptes);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
432
nvkm_vmm_sparse_ptes(it->desc, it->pt[0], ptei, ptes);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
445
struct nvkm_mmu_pt *pt;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
452
pgt->pt[type] = nvkm_mmu_ptc_get(mmu, size, desc->align, zero);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
453
if (!pgt->pt[type]) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
462
pt = pgt->pt[type];
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
481
desc->func->sparse(vmm, pt, pteb, ptes);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
483
desc->func->invalid(vmm, pt, pteb, ptes);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
486
desc->func->unmap(vmm, pt, pteb, ptes);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
494
desc->func->sparse(vmm, pt, 0, pten);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
496
desc->func->invalid(vmm, pt, 0, pten);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
547
it.pt[it.max] = vmm->pd;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
556
struct nvkm_vmm_pt *pgt = it.pt[it.lvl];
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
572
it.pt[it.lvl - 1] = pgt = pgd->pde[pdei];
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
588
struct nvkm_mmu_pt *pt = pgt->pt[type];
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
591
MAP_PTES(vmm, pt, ptei, ptes, map);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
593
CLR_PTES(vmm, pt, ptei, ptes);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
79
struct nvkm_vmm_pt *pt[NVKM_VMM_LEVELS_MAX];
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h
25
struct nvkm_mmu_pt *pt[2];
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
109
struct nvkm_mmu_pt *pd = pgd->pt[0];
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
110
struct nvkm_mmu_pt *pt;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
113
if ((pt = pgt->pt[0])) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
114
switch (nvkm_memory_target(pt->memory)) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
124
data |= pt->addr >> 8;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
127
if ((pt = pgt->pt[1])) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
128
switch (nvkm_memory_target(pt->memory)) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
138
data |= pt->addr << 24;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
191
struct nvkm_mmu_pt *pd = vmm->pd->pt[0];
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
212
addr |= (vmm->pd->pt[0]->addr >> 12) << 4;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
32
gf100_vmm_pgt_pte(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
344
struct nvkm_mmu_pt *pd = vmm->pd->pt[0];
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
44
VMM_WO064(pt, vmm, ptei++ * 8, data);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
51
VMM_WO064(pt, vmm, ptei++ * 8, data);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
58
gf100_vmm_pgt_sgl(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
61
VMM_MAP_ITER_SGL(vmm, pt, ptei, ptes, map, gf100_vmm_pgt_pte);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
65
gf100_vmm_pgt_dma(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
70
nvkm_kmap(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
73
VMM_WO064(pt, vmm, ptei++ * 8, data);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
76
nvkm_done(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
80
VMM_MAP_ITER_DMA(vmm, pt, ptei, ptes, map, gf100_vmm_pgt_pte);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
84
gf100_vmm_pgt_mem(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
87
VMM_MAP_ITER_MEM(vmm, pt, ptei, ptes, map, gf100_vmm_pgt_pte);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
92
struct nvkm_mmu_pt *pt, u32 ptei, u32 ptes)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
94
VMM_FO064(pt, vmm, ptei * 8, 0ULL, ptes);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
100
VMM_WO128(pt, vmm, ptei++ * NV_MMU_VER3_DUAL_PDE__SIZE, data, 0ULL);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
106
gh100_vmm_pd0_mem(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
109
VMM_MAP_ITER_MEM(vmm, pt, ptei, ptes, map, gh100_vmm_pd0_pte);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
113
gh100_vmm_pde(struct nvkm_mmu_pt *pt, u64 *data)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
115
switch (nvkm_memory_target(pt->memory)) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
13
gh100_vmm_pgt_pte(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt, u32 ptei, u32 ptes,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
133
*data |= pt->addr;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
141
struct nvkm_mmu_pt *pd = pgd->pt[0];
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
144
if (pgt->pt[0] && !gh100_vmm_pde(pgt->pt[0], &data[0]))
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
146
if (pgt->pt[1] && !gh100_vmm_pde(pgt->pt[1], &data[1]))
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
156
struct nvkm_mmu_pt *pt, u32 pdei, u32 pdes)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
160
VMM_FO128(pt, vmm, pdei * NV_MMU_VER3_DUAL_PDE__SIZE, data, 0ULL, pdes);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
165
struct nvkm_mmu_pt *pt, u32 pdei, u32 pdes)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
167
VMM_FO128(pt, vmm, pdei * NV_MMU_VER3_DUAL_PDE__SIZE, 0ULL, 0ULL, pdes);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
182
struct nvkm_mmu_pt *pd = pgd->pt[0];
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
185
if (!gh100_vmm_pde(pgt->pt[0], &data))
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
19
VMM_WO064(pt, vmm, ptei++ * NV_MMU_VER3_PTE__SIZE, data);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
25
gh100_vmm_pgt_sgl(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt, u32 ptei, u32 ptes,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
28
VMM_MAP_ITER_SGL(vmm, pt, ptei, ptes, map, gh100_vmm_pgt_pte);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
32
gh100_vmm_pgt_dma(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt, u32 ptei, u32 ptes,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
38
nvkm_kmap(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
42
VMM_WO064(pt, vmm, ptei++ * NV_MMU_VER3_PTE__SIZE, data);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
44
nvkm_done(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
48
VMM_MAP_ITER_DMA(vmm, pt, ptei, ptes, map, gh100_vmm_pgt_pte);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
52
gh100_vmm_pgt_mem(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt, u32 ptei, u32 ptes,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
55
VMM_MAP_ITER_MEM(vmm, pt, ptei, ptes, map, gh100_vmm_pgt_pte);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
60
struct nvkm_mmu_pt *pt, u32 ptei, u32 ptes)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
64
VMM_FO064(pt, vmm, ptei * NV_MMU_VER3_PTE__SIZE, data, ptes);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
78
struct nvkm_mmu_pt *pt, u32 ptei, u32 ptes)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
82
VMM_FO064(pt, vmm, ptei * NV_MMU_VER3_PTE__SIZE, data, ptes);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
94
gh100_vmm_pd0_pte(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgk104.c
26
struct nvkm_mmu_pt *pt, u32 ptei, u32 ptes)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgk104.c
29
VMM_FO064(pt, vmm, ptei * 8, BIT_ULL(1) /* PRIV. */, ptes);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm200.c
29
struct nvkm_mmu_pt *pt, u32 ptei, u32 ptes)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm200.c
32
VMM_FO064(pt, vmm, ptei * 8, BIT_ULL(32) /* VOL. */, ptes);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm200.c
54
struct nvkm_mmu_pt *pt, u32 pdei, u32 pdes)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm200.c
57
VMM_FO064(pt, vmm, pdei * 8, BIT_ULL(35) /* VOL_BIG. */, pdes);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
107
VMM_WO064(pt, vmm, ptei++ * 8, data);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
109
nvkm_done(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
132
gp100_vmm_pgt_pte(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
141
VMM_WO064(pt, vmm, ptei++ * 8, data);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
147
gp100_vmm_pgt_sgl(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
150
VMM_MAP_ITER_SGL(vmm, pt, ptei, ptes, map, gp100_vmm_pgt_pte);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
154
gp100_vmm_pgt_dma(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
159
nvkm_kmap(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
162
VMM_WO064(pt, vmm, ptei++ * 8, data);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
164
nvkm_done(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
168
VMM_MAP_ITER_DMA(vmm, pt, ptei, ptes, map, gp100_vmm_pgt_pte);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
172
gp100_vmm_pgt_mem(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
175
VMM_MAP_ITER_MEM(vmm, pt, ptei, ptes, map, gp100_vmm_pgt_pte);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
180
struct nvkm_mmu_pt *pt, u32 ptei, u32 ptes)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
183
VMM_FO064(pt, vmm, ptei * 8, BIT_ULL(3) /* VOL. */, ptes);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
200
struct nvkm_mmu_pt *pt, u32 ptei, u32 ptes)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
203
VMM_FO064(pt, vmm, ptei * 8, BIT_ULL(5) /* PRIV. */, ptes);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
215
gp100_vmm_pd0_pte(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
224
VMM_WO128(pt, vmm, ptei++ * 0x10, data, 0ULL);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
230
gp100_vmm_pd0_mem(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
233
VMM_MAP_ITER_MEM(vmm, pt, ptei, ptes, map, gp100_vmm_pd0_pte);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
237
gp100_vmm_pde(struct nvkm_mmu_pt *pt, u64 *data)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
239
switch (nvkm_memory_target(pt->memory)) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
249
*data |= pt->addr >> 4;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
257
struct nvkm_mmu_pt *pd = pgd->pt[0];
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
260
if (pgt->pt[0] && !gp100_vmm_pde(pgt->pt[0], &data[0]))
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
262
if (pgt->pt[1] && !gp100_vmm_pde(pgt->pt[1], &data[1]))
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
272
struct nvkm_mmu_pt *pt, u32 pdei, u32 pdes)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
275
VMM_FO128(pt, vmm, pdei * 0x10, BIT_ULL(3) /* VOL_BIG. */, 0ULL, pdes);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
280
struct nvkm_mmu_pt *pt, u32 pdei, u32 pdes)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
282
VMM_FO128(pt, vmm, pdei * 0x10, 0ULL, 0ULL, pdes);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
287
struct nvkm_mmu_pt *pt, u32 ptei, u32 ptes)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
292
nvkm_kmap(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
294
u32 datalo = nvkm_ro32(pt->memory, pt->base + ptei * 16 + 0);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
295
u32 datahi = nvkm_ro32(pt->memory, pt->base + ptei * 16 + 4);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
304
nvkm_done(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
309
struct nvkm_mmu_pt *pt, u32 ptei, u32 ptes)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
313
nvkm_kmap(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
315
u32 datalo = nvkm_ro32(pt->memory, pt->base + ptei * 16 + 0);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
316
u32 datahi = nvkm_ro32(pt->memory, pt->base + ptei * 16 + 4);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
320
VMM_WO064(pt, vmm, ptei * 16, data & ~BIT_ULL(0));
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
325
nvkm_done(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
33
struct nvkm_mmu_pt *pt, u32 ptei, u32 ptes)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
330
gp100_vmm_pd0_pfn(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
336
nvkm_kmap(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
364
VMM_WO064(pt, vmm, ptei++ * 16, data);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
366
nvkm_done(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
38
nvkm_kmap(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
384
struct nvkm_mmu_pt *pd = pgd->pt[0];
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
387
if (!gp100_vmm_pde(pgt->pt[0], &data))
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
40
u32 datalo = nvkm_ro32(pt->memory, pt->base + ptei * 8 + 0);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
41
u32 datahi = nvkm_ro32(pt->memory, pt->base + ptei * 8 + 4);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
49
nvkm_done(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
54
struct nvkm_mmu_pt *pt, u32 ptei, u32 ptes)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
57
nvkm_kmap(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
59
u32 datalo = nvkm_ro32(pt->memory, pt->base + ptei * 8 + 0);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
60
u32 datahi = nvkm_ro32(pt->memory, pt->base + ptei * 8 + 4);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
63
VMM_WO064(pt, vmm, ptei * 8, data & ~BIT_ULL(0));
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
68
nvkm_done(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
73
gp100_vmm_pgt_pfn(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
79
nvkm_kmap(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c
135
mem = vmm->pd->pt[0]->memory;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c
28
nv04_vmm_pgt_pte(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c
33
VMM_WO032(pt, vmm, 8 + ptei++ * 4, data);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c
39
nv04_vmm_pgt_sgl(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c
42
VMM_MAP_ITER_SGL(vmm, pt, ptei, ptes, map, nv04_vmm_pgt_pte);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c
46
nv04_vmm_pgt_dma(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c
50
nvkm_kmap(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c
52
VMM_WO032(pt, vmm, 8 + (ptei++ * 4), *map->dma++ | 0x00000003);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c
53
nvkm_done(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c
55
VMM_MAP_ITER_DMA(vmm, pt, ptei, ptes, map, nv04_vmm_pgt_pte);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c
61
struct nvkm_mmu_pt *pt, u32 ptei, u32 ptes)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c
63
VMM_FO032(pt, vmm, 8 + (ptei * 4), 0, ptes);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv41.c
27
nv41_vmm_pgt_pte(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv41.c
32
VMM_WO032(pt, vmm, ptei++ * 4, data);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv41.c
38
nv41_vmm_pgt_sgl(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv41.c
41
VMM_MAP_ITER_SGL(vmm, pt, ptei, ptes, map, nv41_vmm_pgt_pte);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv41.c
45
nv41_vmm_pgt_dma(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv41.c
49
nvkm_kmap(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv41.c
52
VMM_WO032(pt, vmm, ptei++ * 4, data);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv41.c
54
nvkm_done(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv41.c
56
VMM_MAP_ITER_DMA(vmm, pt, ptei, ptes, map, nv41_vmm_pgt_pte);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv41.c
62
struct nvkm_mmu_pt *pt, u32 ptei, u32 ptes)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv41.c
64
VMM_FO032(pt, vmm, ptei * 4, 0, ptes);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
100
nv44_vmm_pgt_fill(vmm, pt, tmp, ptei, ptes);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
105
nv44_vmm_pgt_sgl(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
108
VMM_MAP_ITER_SGL(vmm, pt, ptei, ptes, map, nv44_vmm_pgt_pte);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
112
nv44_vmm_pgt_dma(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
116
nvkm_kmap(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
119
nv44_vmm_pgt_fill(vmm, pt, map->dma, ptei, pten);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
129
VMM_WO032(pt, vmm, ptei++ * 4, tmp[0] >> 0 | tmp[1] << 27);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
130
VMM_WO032(pt, vmm, ptei++ * 4, tmp[1] >> 5 | tmp[2] << 22);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
131
VMM_WO032(pt, vmm, ptei++ * 4, tmp[2] >> 10 | tmp[3] << 17);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
132
VMM_WO032(pt, vmm, ptei++ * 4, tmp[3] >> 15 | 0x40000000);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
137
nv44_vmm_pgt_fill(vmm, pt, map->dma, ptei, ptes);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
140
nvkm_done(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
142
VMM_MAP_ITER_DMA(vmm, pt, ptei, ptes, map, nv44_vmm_pgt_pte);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
148
struct nvkm_mmu_pt *pt, u32 ptei, u32 ptes)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
150
nvkm_kmap(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
153
nv44_vmm_pgt_fill(vmm, pt, NULL, ptei, pten);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
159
VMM_WO032(pt, vmm, ptei++ * 4, 0x00000000);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
160
VMM_WO032(pt, vmm, ptei++ * 4, 0x00000000);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
161
VMM_WO032(pt, vmm, ptei++ * 4, 0x00000000);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
162
VMM_WO032(pt, vmm, ptei++ * 4, 0x00000000);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
167
nv44_vmm_pgt_fill(vmm, pt, NULL, ptei, ptes);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
168
nvkm_done(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
27
nv44_vmm_pgt_fill(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
33
tmp[0] = nvkm_ro32(pt->memory, pteo + 0x0);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
34
tmp[1] = nvkm_ro32(pt->memory, pteo + 0x4);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
35
tmp[2] = nvkm_ro32(pt->memory, pteo + 0x8);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
36
tmp[3] = nvkm_ro32(pt->memory, pteo + 0xc);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
66
VMM_WO032(pt, vmm, pteo + 0x0, tmp[0]);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
67
VMM_WO032(pt, vmm, pteo + 0x4, tmp[1]);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
68
VMM_WO032(pt, vmm, pteo + 0x8, tmp[2]);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
69
VMM_WO032(pt, vmm, pteo + 0xc, tmp[3] | 0x40000000);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
73
nv44_vmm_pgt_pte(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
82
nv44_vmm_pgt_fill(vmm, pt, tmp, ptei, pten);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
90
VMM_WO032(pt, vmm, ptei++ * 4, tmp[0] >> 0 | tmp[1] << 27);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
91
VMM_WO032(pt, vmm, ptei++ * 4, tmp[1] >> 5 | tmp[2] << 22);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
92
VMM_WO032(pt, vmm, ptei++ * 4, tmp[2] >> 10 | tmp[3] << 17);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
93
VMM_WO032(pt, vmm, ptei++ * 4, tmp[3] >> 15 | 0x40000000);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
108
struct nvkm_mmu_pt *pt;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
110
if (pgt && (pt = pgt->pt[0])) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
114
switch (nvkm_memory_size(pt->memory)) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
129
switch (nvkm_memory_target(pt->memory)) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
138
data |= pt->addr;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
32
nv50_vmm_pgt_pte(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
53
VMM_WO064(pt, vmm, ptei++ * 8, data);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
58
nv50_vmm_pgt_sgl(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
61
VMM_MAP_ITER_SGL(vmm, pt, ptei, ptes, map, nv50_vmm_pgt_pte);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
65
nv50_vmm_pgt_dma(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
70
nvkm_kmap(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
73
VMM_WO064(pt, vmm, ptei++ * 8, data);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
76
nvkm_done(pt->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
80
VMM_MAP_ITER_DMA(vmm, pt, ptei, ptes, map, nv50_vmm_pgt_pte);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
84
nv50_vmm_pgt_mem(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
87
VMM_MAP_ITER_MEM(vmm, pt, ptei, ptes, map, nv50_vmm_pgt_pte);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
92
struct nvkm_mmu_pt *pt, u32 ptei, u32 ptes)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
94
VMM_FO064(pt, vmm, ptei * 8, 0ULL, ptes);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmtu102.c
39
nvkm_wr32(device, 0xb830a0, vmm->pd->pt[0]->addr >> 8);
drivers/gpu/drm/omapdrm/tcm.h
298
#define tcm_is_in(pt, area) __tcm_is_in(&(pt), &(area))
drivers/gpu/drm/panel/panel-arm-versatile.c
324
const struct versatile_panel_type *pt;
drivers/gpu/drm/panel/panel-arm-versatile.c
326
pt = &versatile_panels[i];
drivers/gpu/drm/panel/panel-arm-versatile.c
327
if (pt->magic == val) {
drivers/gpu/drm/panel/panel-arm-versatile.c
328
vpanel->panel_type = pt;
drivers/gpu/drm/radeon/r600_dpm.c
1255
ATOM_PowerTune_Table *pt;
drivers/gpu/drm/radeon/r600_dpm.c
1268
pt = &ppt->power_tune_table;
drivers/gpu/drm/radeon/r600_dpm.c
1274
pt = &ppt->power_tune_table;
drivers/gpu/drm/radeon/r600_dpm.c
1276
rdev->pm.dpm.dyn_state.cac_tdp_table->tdp = le16_to_cpu(pt->usTDP);
drivers/gpu/drm/radeon/r600_dpm.c
1278
le16_to_cpu(pt->usConfigurableTDP);
drivers/gpu/drm/radeon/r600_dpm.c
1279
rdev->pm.dpm.dyn_state.cac_tdp_table->tdc = le16_to_cpu(pt->usTDC);
drivers/gpu/drm/radeon/r600_dpm.c
1281
le16_to_cpu(pt->usBatteryPowerLimit);
drivers/gpu/drm/radeon/r600_dpm.c
1283
le16_to_cpu(pt->usSmallPowerLimit);
drivers/gpu/drm/radeon/r600_dpm.c
1285
le16_to_cpu(pt->usLowCACLeakage);
drivers/gpu/drm/radeon/r600_dpm.c
1287
le16_to_cpu(pt->usHighCACLeakage);
drivers/gpu/drm/radeon/radeon_vm.c
540
struct radeon_bo *pt;
drivers/gpu/drm/radeon/radeon_vm.c
551
NULL, NULL, &pt);
drivers/gpu/drm/radeon/radeon_vm.c
555
r = radeon_vm_clear_bo(rdev, pt);
drivers/gpu/drm/radeon/radeon_vm.c
557
radeon_bo_unref(&pt);
drivers/gpu/drm/radeon/radeon_vm.c
566
radeon_bo_unref(&pt);
drivers/gpu/drm/radeon/radeon_vm.c
572
vm->page_tables[pt_idx].bo = pt;
drivers/gpu/drm/radeon/radeon_vm.c
666
uint64_t pde, pt;
drivers/gpu/drm/radeon/radeon_vm.c
671
pt = radeon_bo_gpu_offset(bo);
drivers/gpu/drm/radeon/radeon_vm.c
672
if (vm->page_tables[pt_idx].addr == pt)
drivers/gpu/drm/radeon/radeon_vm.c
674
vm->page_tables[pt_idx].addr = pt;
drivers/gpu/drm/radeon/radeon_vm.c
678
((last_pt + incr * count) != pt)) {
drivers/gpu/drm/radeon/radeon_vm.c
688
last_pt = pt;
drivers/gpu/drm/radeon/radeon_vm.c
825
struct radeon_bo *pt = vm->page_tables[pt_idx].bo;
drivers/gpu/drm/radeon/radeon_vm.c
830
radeon_sync_resv(rdev, &ib->sync, pt->tbo.base.resv, true);
drivers/gpu/drm/radeon/radeon_vm.c
831
r = dma_resv_reserve_fences(pt->tbo.base.resv, 1);
drivers/gpu/drm/radeon/radeon_vm.c
840
pte = radeon_bo_gpu_offset(pt);
drivers/gpu/drm/ttm/tests/ttm_device_test.c
155
struct ttm_pool_type pt;
drivers/gpu/drm/ttm/tests/ttm_device_test.c
172
pt = pool->caching[i].orders[j];
drivers/gpu/drm/ttm/tests/ttm_device_test.c
173
KUNIT_EXPECT_PTR_EQ(test, pt.pool, pool);
drivers/gpu/drm/ttm/tests/ttm_device_test.c
174
KUNIT_EXPECT_EQ(test, pt.caching, i);
drivers/gpu/drm/ttm/tests/ttm_device_test.c
175
KUNIT_EXPECT_EQ(test, pt.order, j);
drivers/gpu/drm/ttm/tests/ttm_device_test.c
179
list_empty(&pt.pages));
drivers/gpu/drm/ttm/tests/ttm_pool_test.c
242
struct ttm_pool_type *pt;
drivers/gpu/drm/ttm/tests/ttm_pool_test.c
250
pt = &pool->caching[caching].orders[order];
drivers/gpu/drm/ttm/tests/ttm_pool_test.c
251
KUNIT_ASSERT_FALSE(test, list_empty(&pt->pages));
drivers/gpu/drm/ttm/tests/ttm_pool_test.c
259
KUNIT_ASSERT_TRUE(test, list_empty(&pt->pages));
drivers/gpu/drm/ttm/tests/ttm_pool_test.c
340
struct ttm_pool_type *pt;
drivers/gpu/drm/ttm/tests/ttm_pool_test.c
354
pt = &pool->caching[caching].orders[order];
drivers/gpu/drm/ttm/tests/ttm_pool_test.c
355
KUNIT_ASSERT_TRUE(test, list_empty(&pt->pages));
drivers/gpu/drm/ttm/tests/ttm_pool_test.c
360
KUNIT_ASSERT_FALSE(test, list_empty(&pt->pages));
drivers/gpu/drm/ttm/tests/ttm_pool_test.c
371
struct ttm_pool_type *pt;
drivers/gpu/drm/ttm/tests/ttm_pool_test.c
385
pt = &pool->caching[caching].orders[order];
drivers/gpu/drm/ttm/tests/ttm_pool_test.c
386
KUNIT_ASSERT_TRUE(test, list_is_singular(&pt->pages));
drivers/gpu/drm/ttm/tests/ttm_pool_test.c
391
KUNIT_ASSERT_TRUE(test, list_is_singular(&pt->pages));
drivers/gpu/drm/ttm/tests/ttm_pool_test.c
399
struct ttm_pool_type *pt;
drivers/gpu/drm/ttm/tests/ttm_pool_test.c
405
pt = &pool->caching[caching].orders[order];
drivers/gpu/drm/ttm/tests/ttm_pool_test.c
407
KUNIT_ASSERT_FALSE(test, list_empty(&pt->pages));
drivers/gpu/drm/ttm/tests/ttm_pool_test.c
411
KUNIT_ASSERT_TRUE(test, list_empty(&pt->pages));
drivers/gpu/drm/ttm/ttm_pool.c
1089
struct ttm_pool_type *pt;
drivers/gpu/drm/ttm/ttm_pool.c
1092
pt = ttm_pool_select_type(pool, i, j);
drivers/gpu/drm/ttm/ttm_pool.c
1093
if (pt != &pool->caching[i].orders[j])
drivers/gpu/drm/ttm/ttm_pool.c
1096
ttm_pool_type_init(pt, pool, i, j);
drivers/gpu/drm/ttm/ttm_pool.c
1128
struct ttm_pool_type *pt;
drivers/gpu/drm/ttm/ttm_pool.c
1130
pt = ttm_pool_select_type(pool, i, j);
drivers/gpu/drm/ttm/ttm_pool.c
1131
if (pt != &pool->caching[i].orders[j])
drivers/gpu/drm/ttm/ttm_pool.c
1134
ttm_pool_type_fini(pt);
drivers/gpu/drm/ttm/ttm_pool.c
1174
static unsigned int ttm_pool_type_count(struct ttm_pool_type *pt)
drivers/gpu/drm/ttm/ttm_pool.c
1179
spin_lock(&pt->lock);
drivers/gpu/drm/ttm/ttm_pool.c
1181
list_for_each_entry(p, &pt->pages, lru)
drivers/gpu/drm/ttm/ttm_pool.c
1183
spin_unlock(&pt->lock);
drivers/gpu/drm/ttm/ttm_pool.c
1200
static void ttm_pool_debugfs_orders(struct ttm_pool_type *pt,
drivers/gpu/drm/ttm/ttm_pool.c
1206
seq_printf(m, " %8u", ttm_pool_type_count(&pt[i]));
drivers/gpu/drm/ttm/ttm_pool.c
285
static void ttm_pool_type_give(struct ttm_pool_type *pt, struct page *p)
drivers/gpu/drm/ttm/ttm_pool.c
287
unsigned int i, num_pages = 1 << pt->order;
drivers/gpu/drm/ttm/ttm_pool.c
296
spin_lock(&pt->lock);
drivers/gpu/drm/ttm/ttm_pool.c
297
list_add(&p->lru, &pt->pages);
drivers/gpu/drm/ttm/ttm_pool.c
298
spin_unlock(&pt->lock);
drivers/gpu/drm/ttm/ttm_pool.c
299
atomic_long_add(1 << pt->order, &allocated_pages);
drivers/gpu/drm/ttm/ttm_pool.c
303
static struct page *ttm_pool_type_take(struct ttm_pool_type *pt)
drivers/gpu/drm/ttm/ttm_pool.c
307
spin_lock(&pt->lock);
drivers/gpu/drm/ttm/ttm_pool.c
308
p = list_first_entry_or_null(&pt->pages, typeof(*p), lru);
drivers/gpu/drm/ttm/ttm_pool.c
310
atomic_long_sub(1 << pt->order, &allocated_pages);
drivers/gpu/drm/ttm/ttm_pool.c
313
spin_unlock(&pt->lock);
drivers/gpu/drm/ttm/ttm_pool.c
319
static void ttm_pool_type_init(struct ttm_pool_type *pt, struct ttm_pool *pool,
drivers/gpu/drm/ttm/ttm_pool.c
322
pt->pool = pool;
drivers/gpu/drm/ttm/ttm_pool.c
323
pt->caching = caching;
drivers/gpu/drm/ttm/ttm_pool.c
324
pt->order = order;
drivers/gpu/drm/ttm/ttm_pool.c
325
spin_lock_init(&pt->lock);
drivers/gpu/drm/ttm/ttm_pool.c
326
INIT_LIST_HEAD(&pt->pages);
drivers/gpu/drm/ttm/ttm_pool.c
329
list_add_tail(&pt->shrinker_list, &shrinker_list);
drivers/gpu/drm/ttm/ttm_pool.c
334
static void ttm_pool_type_fini(struct ttm_pool_type *pt)
drivers/gpu/drm/ttm/ttm_pool.c
339
list_del(&pt->shrinker_list);
drivers/gpu/drm/ttm/ttm_pool.c
342
while ((p = ttm_pool_type_take(pt)))
drivers/gpu/drm/ttm/ttm_pool.c
343
ttm_pool_free_page(pt->pool, pt->caching, pt->order, p);
drivers/gpu/drm/ttm/ttm_pool.c
383
struct ttm_pool_type *pt;
drivers/gpu/drm/ttm/ttm_pool.c
389
pt = list_first_entry(&shrinker_list, typeof(*pt), shrinker_list);
drivers/gpu/drm/ttm/ttm_pool.c
390
list_move_tail(&pt->shrinker_list, &shrinker_list);
drivers/gpu/drm/ttm/ttm_pool.c
393
p = ttm_pool_type_take(pt);
drivers/gpu/drm/ttm/ttm_pool.c
395
ttm_pool_free_page(pt->pool, pt->caching, pt->order, p);
drivers/gpu/drm/ttm/ttm_pool.c
396
num_pages = 1 << pt->order;
drivers/gpu/drm/ttm/ttm_pool.c
468
struct ttm_pool_type *pt = NULL;
drivers/gpu/drm/ttm/ttm_pool.c
478
pt = ttm_pool_select_type(pool, caching, order);
drivers/gpu/drm/ttm/ttm_pool.c
484
if (pt)
drivers/gpu/drm/ttm/ttm_pool.c
485
ttm_pool_type_give(pt, page);
drivers/gpu/drm/ttm/ttm_pool.c
741
struct ttm_pool_type *pt;
drivers/gpu/drm/ttm/ttm_pool.c
745
pt = ttm_pool_select_type(pool, page_caching, order);
drivers/gpu/drm/ttm/ttm_pool.c
746
if (pt && allow_pools)
drivers/gpu/drm/ttm/ttm_pool.c
747
p = ttm_pool_type_take(pt);
drivers/gpu/drm/v3d/v3d_drv.h
140
volatile u32 *pt;
drivers/gpu/drm/v3d/v3d_gem.c
318
v3d->pt = dma_alloc_wc(v3d->drm.dev, pt_size,
drivers/gpu/drm/v3d/v3d_gem.c
321
if (!v3d->pt) {
drivers/gpu/drm/v3d/v3d_gem.c
336
dma_free_coherent(v3d->drm.dev, pt_size, (void *)v3d->pt,
drivers/gpu/drm/v3d/v3d_gem.c
360
dma_free_coherent(v3d->drm.dev, 4096 * 1024, (void *)v3d->pt,
drivers/gpu/drm/v3d/v3d_mmu.c
118
v3d->pt[page++] = page_address + i;
drivers/gpu/drm/v3d/v3d_mmu.c
140
v3d->pt[page] = 0;
drivers/gpu/drm/xe/tests/xe_migrate.c
190
struct xe_bo *pt, *bo = m->pt_bo, *big, *tiny;
drivers/gpu/drm/xe/tests/xe_migrate.c
214
pt = xe_bo_create_pin_map(xe, tile, m->q->vm, XE_PAGE_SIZE,
drivers/gpu/drm/xe/tests/xe_migrate.c
218
if (IS_ERR(pt)) {
drivers/gpu/drm/xe/tests/xe_migrate.c
220
PTR_ERR(pt));
drivers/gpu/drm/xe/tests/xe_migrate.c
249
expected = m->q->vm->pt_ops->pte_encode_bo(pt, 0, xe->pat.idx[XE_CACHE_WB], 0);
drivers/gpu/drm/xe/tests/xe_migrate.c
252
if (xe_bo_is_vram(pt))
drivers/gpu/drm/xe/tests/xe_migrate.c
253
xe_res_first(pt->ttm.resource, 0, xe_bo_size(pt), &src_it);
drivers/gpu/drm/xe/tests/xe_migrate.c
255
xe_res_first_sg(xe_bo_sg(pt), 0, xe_bo_size(pt), &src_it);
drivers/gpu/drm/xe/tests/xe_migrate.c
257
emit_pte(m, bb, NUM_KERNEL_PDE - 1, xe_bo_is_vram(pt), false,
drivers/gpu/drm/xe/tests/xe_migrate.c
258
&src_it, XE_PAGE_SIZE, pt->ttm.resource);
drivers/gpu/drm/xe/tests/xe_migrate.c
269
xe_map_wr(xe, &pt->vmap, 0, u32, 0xdeaddead);
drivers/gpu/drm/xe/tests/xe_migrate.c
277
retval = xe_map_rd(xe, &pt->vmap, 0, u32);
drivers/gpu/drm/xe/tests/xe_migrate.c
332
xe_bo_unpin(pt);
drivers/gpu/drm/xe/tests/xe_migrate.c
333
xe_bo_put(pt);
drivers/gpu/drm/xe/xe_lmtt.c
101
lmtt_debug(&pt->bo->tile->sriov.pf.lmtt, "level=%u addr=%llx\n",
drivers/gpu/drm/xe/xe_lmtt.c
102
pt->level, (u64)xe_bo_main_addr(pt->bo, XE_PAGE_SIZE));
drivers/gpu/drm/xe/xe_lmtt.c
104
xe_bo_unpin_map_no_vm(pt->bo);
drivers/gpu/drm/xe/xe_lmtt.c
105
kfree(pt);
drivers/gpu/drm/xe/xe_lmtt.c
299
static void lmtt_write_pte(struct xe_lmtt *lmtt, struct xe_lmtt_pt *pt,
drivers/gpu/drm/xe/xe_lmtt.c
302
unsigned int level = pt->level;
drivers/gpu/drm/xe/xe_lmtt.c
310
lmtt_assert(lmtt, !pte || !iosys_map_rd(&pt->bo->vmap, idx * sizeof(u32), u32));
drivers/gpu/drm/xe/xe_lmtt.c
312
xe_map_wr(lmtt_to_xe(lmtt), &pt->bo->vmap, idx * sizeof(u32), u32, pte);
drivers/gpu/drm/xe/xe_lmtt.c
315
lmtt_assert(lmtt, !pte || !iosys_map_rd(&pt->bo->vmap, idx * sizeof(u64), u64));
drivers/gpu/drm/xe/xe_lmtt.c
317
xe_map_wr(lmtt_to_xe(lmtt), &pt->bo->vmap, idx * sizeof(u64), u64, pte);
drivers/gpu/drm/xe/xe_lmtt.c
327
struct xe_lmtt_pt *pt;
drivers/gpu/drm/xe/xe_lmtt.c
331
pt = pd->entries[i];
drivers/gpu/drm/xe/xe_lmtt.c
333
if (!pt)
drivers/gpu/drm/xe/xe_lmtt.c
336
lmtt_destroy_pt(lmtt, pt);
drivers/gpu/drm/xe/xe_lmtt.c
345
struct xe_lmtt_pt *pt;
drivers/gpu/drm/xe/xe_lmtt.c
347
pt = pd->entries[vfid];
drivers/gpu/drm/xe/xe_lmtt.c
349
if (!pt)
drivers/gpu/drm/xe/xe_lmtt.c
356
lmtt_assert(lmtt, pt->level == pd->level - 1);
drivers/gpu/drm/xe/xe_lmtt.c
357
lmtt_destroy_pt(lmtt, pt);
drivers/gpu/drm/xe/xe_lmtt.c
371
struct xe_lmtt_pt *pt;
drivers/gpu/drm/xe/xe_lmtt.c
375
pt = lmtt_pt_alloc(lmtt, pd->level - 1);
drivers/gpu/drm/xe/xe_lmtt.c
376
if (IS_ERR(pt))
drivers/gpu/drm/xe/xe_lmtt.c
377
return PTR_ERR(pt);
drivers/gpu/drm/xe/xe_lmtt.c
379
pt_addr = xe_bo_main_addr(pt->bo, XE_PAGE_SIZE);
drivers/gpu/drm/xe/xe_lmtt.c
386
pd->entries[idx] = pt;
drivers/gpu/drm/xe/xe_lmtt.c
390
if (pt->level != 0) {
drivers/gpu/drm/xe/xe_lmtt.c
391
err = __lmtt_alloc_range(lmtt, pt, offset, next);
drivers/gpu/drm/xe/xe_lmtt.c
405
struct xe_lmtt_pt *pt;
drivers/gpu/drm/xe/xe_lmtt.c
418
pt = lmtt_pt_alloc(lmtt, pd->level - 1);
drivers/gpu/drm/xe/xe_lmtt.c
419
if (IS_ERR(pt))
drivers/gpu/drm/xe/xe_lmtt.c
420
return PTR_ERR(pt);
drivers/gpu/drm/xe/xe_lmtt.c
422
pt_addr = xe_bo_main_addr(pt->bo, XE_PAGE_SIZE);
drivers/gpu/drm/xe/xe_lmtt.c
428
pd->entries[vfid] = pt;
drivers/gpu/drm/xe/xe_lmtt.c
430
if (pt->level != 0) {
drivers/gpu/drm/xe/xe_lmtt.c
431
err = __lmtt_alloc_range(lmtt, pt, start, end);
drivers/gpu/drm/xe/xe_lmtt.c
439
lmtt_pt_free(pt);
drivers/gpu/drm/xe/xe_lmtt.c
446
struct xe_lmtt_pt *pt;
drivers/gpu/drm/xe/xe_lmtt.c
449
pt = pd->entries[vfid];
drivers/gpu/drm/xe/xe_lmtt.c
451
while (pt->level) {
drivers/gpu/drm/xe/xe_lmtt.c
452
lmtt_assert(lmtt, lmtt->ops->lmtt_pte_index(addr, pt->level) <=
drivers/gpu/drm/xe/xe_lmtt.c
453
lmtt->ops->lmtt_pte_num(pt->level));
drivers/gpu/drm/xe/xe_lmtt.c
455
pt = pt->entries[lmtt->ops->lmtt_pte_index(addr, pt->level)];
drivers/gpu/drm/xe/xe_lmtt.c
457
addr >>= lmtt->ops->lmtt_pte_shift(pt->level);
drivers/gpu/drm/xe/xe_lmtt.c
460
lmtt_assert(lmtt, lmtt->ops->lmtt_pte_index(addr, pt->level) <=
drivers/gpu/drm/xe/xe_lmtt.c
461
lmtt->ops->lmtt_pte_num(pt->level));
drivers/gpu/drm/xe/xe_lmtt.c
462
lmtt_assert(lmtt, pt->level != pd->level);
drivers/gpu/drm/xe/xe_lmtt.c
463
lmtt_assert(lmtt, pt->level == 0);
drivers/gpu/drm/xe/xe_lmtt.c
464
return pt;
drivers/gpu/drm/xe/xe_lmtt.c
471
struct xe_lmtt_pt *pt;
drivers/gpu/drm/xe/xe_lmtt.c
484
pt = lmtt_leaf_pt(lmtt, vfid, start);
drivers/gpu/drm/xe/xe_lmtt.c
486
lmtt_write_pte(lmtt, pt, lmtt->ops->lmtt_pte_encode(addr, 0),
drivers/gpu/drm/xe/xe_lmtt.c
63
struct xe_lmtt_pt *pt;
drivers/gpu/drm/xe/xe_lmtt.c
67
pt = kzalloc_flex(*pt, entries, num_entries);
drivers/gpu/drm/xe/xe_lmtt.c
68
if (!pt) {
drivers/gpu/drm/xe/xe_lmtt.c
89
pt->level = level;
drivers/gpu/drm/xe/xe_lmtt.c
90
pt->bo = bo;
drivers/gpu/drm/xe/xe_lmtt.c
91
return pt;
drivers/gpu/drm/xe/xe_lmtt.c
94
kfree(pt);
drivers/gpu/drm/xe/xe_lmtt.c
99
static void lmtt_pt_free(struct xe_lmtt_pt *pt)
drivers/gpu/drm/xe/xe_pt.c
1008
struct xe_pt *pt = entries[i].pt;
drivers/gpu/drm/xe/xe_pt.c
1010
if (!pt)
drivers/gpu/drm/xe/xe_pt.c
1013
if (pt->level) {
drivers/gpu/drm/xe/xe_pt.c
1015
xe_pt_destroy(entries[i].pt_entries[j].pt,
drivers/gpu/drm/xe/xe_pt.c
106
struct xe_pt *pt;
drivers/gpu/drm/xe/xe_pt.c
1066
struct xe_pt *pt = entries[i].pt;
drivers/gpu/drm/xe/xe_pt.c
1069
if (!pt->level)
drivers/gpu/drm/xe/xe_pt.c
1072
pt_dir = as_xe_pt_dir(pt);
drivers/gpu/drm/xe/xe_pt.c
1074
struct xe_pt *oldpte = entries[i].pt_entries[j].pt;
drivers/gpu/drm/xe/xe_pt.c
1093
struct xe_pt *pt = entries[i].pt;
drivers/gpu/drm/xe/xe_pt.c
1097
pt->num_live -= entries[i].qwords;
drivers/gpu/drm/xe/xe_pt.c
1099
if (!pt->level)
drivers/gpu/drm/xe/xe_pt.c
1102
pt_dir = as_xe_pt_dir(pt);
drivers/gpu/drm/xe/xe_pt.c
1106
struct xe_pt *oldpte = entries[i].pt_entries[j].pt;
drivers/gpu/drm/xe/xe_pt.c
1123
struct xe_pt *pt = entries[i].pt;
drivers/gpu/drm/xe/xe_pt.c
1127
pt->num_live += entries[i].qwords;
drivers/gpu/drm/xe/xe_pt.c
1129
if (!pt->level)
drivers/gpu/drm/xe/xe_pt.c
1132
pt_dir = as_xe_pt_dir(pt);
drivers/gpu/drm/xe/xe_pt.c
1135
struct xe_pt *newpte = entries[i].pt_entries[j].pt;
drivers/gpu/drm/xe/xe_pt.c
114
pt = (dir) ? &dir->pt : NULL;
drivers/gpu/drm/xe/xe_pt.c
1142
entries[i].pt_entries[j].pt = oldpte;
drivers/gpu/drm/xe/xe_pt.c
116
pt = kzalloc_obj(*pt);
drivers/gpu/drm/xe/xe_pt.c
118
if (!pt)
drivers/gpu/drm/xe/xe_pt.c
1184
struct xe_pt *xe_pt = entry->pt;
drivers/gpu/drm/xe/xe_pt.c
1189
xe_assert(xe, !entry->pt->is_compact);
drivers/gpu/drm/xe/xe_pt.c
127
pt->level = level;
drivers/gpu/drm/xe/xe_pt.c
137
pt->bo = bo;
drivers/gpu/drm/xe/xe_pt.c
138
pt->base.children = level ? as_xe_pt_dir(pt)->children : NULL;
drivers/gpu/drm/xe/xe_pt.c
139
pt->base.staging = level ? as_xe_pt_dir(pt)->staging : NULL;
drivers/gpu/drm/xe/xe_pt.c
142
xe_drm_client_add_bo(vm->xef->client, pt->bo);
drivers/gpu/drm/xe/xe_pt.c
145
return pt;
drivers/gpu/drm/xe/xe_pt.c
148
xe_pt_free(pt);
drivers/gpu/drm/xe/xe_pt.c
164
struct xe_pt *pt)
drivers/gpu/drm/xe/xe_pt.c
166
struct iosys_map *map = &pt->bo->vmap;
drivers/gpu/drm/xe/xe_pt.c
177
empty = __xe_pt_empty_pte(tile, vm, pt->level);
drivers/gpu/drm/xe/xe_pt.c
1825
struct xe_pt *pt = vm->pt_root[tile->id];
drivers/gpu/drm/xe/xe_pt.c
1827
(void)xe_pt_walk_shared(&pt->base, pt->level, start, end,
drivers/gpu/drm/xe/xe_pt.c
1840
u64 empty = __xe_pt_empty_pte(tile, vm, update->pt->level);
drivers/gpu/drm/xe/xe_pt.c
1864
struct xe_pt *pt = entry->pt;
drivers/gpu/drm/xe/xe_pt.c
1865
struct xe_pt_dir *pt_dir = as_xe_pt_dir(pt);
drivers/gpu/drm/xe/xe_pt.c
1867
pt->num_live += entry->qwords;
drivers/gpu/drm/xe/xe_pt.c
1869
if (!pt->level)
drivers/gpu/drm/xe/xe_pt.c
1874
entries[i].pt_entries[j - entry->ofs].pt ?
drivers/gpu/drm/xe/xe_pt.c
1875
&entries[i].pt_entries[j - entry->ofs].pt->base : NULL;
drivers/gpu/drm/xe/xe_pt.c
1890
struct xe_pt *pt = entry->pt;
drivers/gpu/drm/xe/xe_pt.c
1893
pt->num_live -= entry->qwords;
drivers/gpu/drm/xe/xe_pt.c
1894
if (!pt->level)
drivers/gpu/drm/xe/xe_pt.c
1897
pt_dir = as_xe_pt_dir(pt);
drivers/gpu/drm/xe/xe_pt.c
1899
entry->pt_entries[j - entry->ofs].pt =
drivers/gpu/drm/xe/xe_pt.c
1918
if (entry->pt->level > level)
drivers/gpu/drm/xe/xe_pt.c
1919
level = entry->pt->level;
drivers/gpu/drm/xe/xe_pt.c
206
void xe_pt_destroy(struct xe_pt *pt, u32 flags, struct llist_head *deferred)
drivers/gpu/drm/xe/xe_pt.c
210
if (!pt)
drivers/gpu/drm/xe/xe_pt.c
2123
if (update->pt->level == 0)
drivers/gpu/drm/xe/xe_pt.c
2126
if (update->pt->level == 1)
drivers/gpu/drm/xe/xe_pt.c
213
XE_WARN_ON(!list_empty(&pt->bo->ttm.base.gpuva.list));
drivers/gpu/drm/xe/xe_pt.c
214
xe_bo_unpin(pt->bo);
drivers/gpu/drm/xe/xe_pt.c
215
xe_bo_put_deferred(pt->bo, deferred);
drivers/gpu/drm/xe/xe_pt.c
217
if (pt->level > 0 && pt->num_live) {
drivers/gpu/drm/xe/xe_pt.c
218
struct xe_pt_dir *pt_dir = as_xe_pt_dir(pt);
drivers/gpu/drm/xe/xe_pt.c
226
xe_pt_free(pt);
drivers/gpu/drm/xe/xe_pt.c
236
void xe_pt_clear(struct xe_device *xe, struct xe_pt *pt)
drivers/gpu/drm/xe/xe_pt.c
238
struct iosys_map *map = &pt->bo->vmap;
drivers/gpu/drm/xe/xe_pt.c
30
struct xe_pt pt;
drivers/gpu/drm/xe/xe_pt.c
365
entry->pt = parent;
drivers/gpu/drm/xe/xe_pt.c
424
entry->pt_entries[idx].pt = xe_child;
drivers/gpu/drm/xe/xe_pt.c
50
static struct xe_pt_dir *as_xe_pt_dir(struct xe_pt *pt)
drivers/gpu/drm/xe/xe_pt.c
52
return container_of(pt, struct xe_pt_dir, pt);
drivers/gpu/drm/xe/xe_pt.c
722
struct xe_pt *pt = vm->pt_root[tile->id];
drivers/gpu/drm/xe/xe_pt.c
778
ret = xe_pt_walk_range(&pt->base, pt->level,
drivers/gpu/drm/xe/xe_pt.c
79
static void xe_pt_free(struct xe_pt *pt)
drivers/gpu/drm/xe/xe_pt.c
81
if (pt->level)
drivers/gpu/drm/xe/xe_pt.c
82
kfree(as_xe_pt_dir(pt));
drivers/gpu/drm/xe/xe_pt.c
84
kfree(pt);
drivers/gpu/drm/xe/xe_pt.c
913
struct xe_pt *pt = xe_vma_vm(vma)->pt_root[tile->id];
drivers/gpu/drm/xe/xe_pt.c
924
(void)xe_pt_walk_shared(&pt->base, pt->level, xe_vma_start(vma),
drivers/gpu/drm/xe/xe_pt.c
956
struct xe_pt *pt = vm->pt_root[tile->id];
drivers/gpu/drm/xe/xe_pt.c
976
(void)xe_pt_walk_shared(&pt->base, pt->level, xe_svm_range_start(range),
drivers/gpu/drm/xe/xe_pt.h
36
struct xe_pt *pt);
drivers/gpu/drm/xe/xe_pt.h
38
void xe_pt_destroy(struct xe_pt *pt, u32 flags, struct llist_head *deferred);
drivers/gpu/drm/xe/xe_pt.h
40
void xe_pt_clear(struct xe_device *xe, struct xe_pt *pt);
drivers/gpu/drm/xe/xe_pt_types.h
54
struct xe_pt *pt;
drivers/gpu/drm/xe/xe_pt_types.h
69
struct xe_pt *pt;
drivers/gpu/drm/xe/xe_sriov_pf_provision.c
269
int xe_sriov_pf_provision_bulk_apply_pt(struct xe_device *xe, u32 pt)
drivers/gpu/drm/xe/xe_sriov_pf_provision.c
279
err = xe_gt_sriov_pf_config_bulk_set_preempt_timeout_locked(gt, pt);
drivers/gpu/drm/xe/xe_sriov_pf_provision.c
298
int xe_sriov_pf_provision_apply_vf_pt(struct xe_device *xe, unsigned int vfid, u32 pt)
drivers/gpu/drm/xe/xe_sriov_pf_provision.c
308
err = xe_gt_sriov_pf_config_set_preempt_timeout_locked(gt, vfid, pt);
drivers/gpu/drm/xe/xe_sriov_pf_provision.c
328
int xe_sriov_pf_provision_query_vf_pt(struct xe_device *xe, unsigned int vfid, u32 *pt)
drivers/gpu/drm/xe/xe_sriov_pf_provision.c
340
*pt = value;
drivers/gpu/drm/xe/xe_sriov_pf_provision.c
341
else if (value != *pt)
drivers/gpu/drm/xe/xe_sriov_pf_provision.c
342
return pf_report_unclean(gt, vfid, "PT", value, *pt);
drivers/gpu/drm/xe/xe_sriov_pf_provision.h
19
int xe_sriov_pf_provision_bulk_apply_pt(struct xe_device *xe, u32 pt);
drivers/gpu/drm/xe/xe_sriov_pf_provision.h
20
int xe_sriov_pf_provision_apply_vf_pt(struct xe_device *xe, unsigned int vfid, u32 pt);
drivers/gpu/drm/xe/xe_sriov_pf_provision.h
21
int xe_sriov_pf_provision_query_vf_pt(struct xe_device *xe, unsigned int vfid, u32 *pt);
drivers/gpu/drm/xe/xe_sriov_pf_sysfs.c
134
DEFINE_SIMPLE_BULK_PROVISIONING_SRIOV_DEV_ATTR_WO(preempt_timeout_us, pt, u32);
drivers/gpu/drm/xe/xe_sriov_pf_sysfs.c
230
DEFINE_SIMPLE_PROVISIONING_SRIOV_VF_ATTR(preempt_timeout_us, pt, u32, "%u\n");
drivers/hv/mshv_eventfd.c
228
struct mshv_partition *pt = rp->rsmplr_partn;
drivers/hv/mshv_eventfd.c
230
mutex_lock(&pt->irqfds_resampler_lock);
drivers/hv/mshv_eventfd.c
233
synchronize_srcu(&pt->pt_irq_srcu);
drivers/hv/mshv_eventfd.c
237
mshv_unregister_irq_ack_notifier(pt, &rp->rsmplr_notifier);
drivers/hv/mshv_eventfd.c
241
mutex_unlock(&pt->irqfds_resampler_lock);
drivers/hv/mshv_eventfd.c
303
struct mshv_partition *pt = irqfd->irqfd_partn;
drivers/hv/mshv_eventfd.c
310
idx = srcu_read_lock(&pt->pt_irq_srcu);
drivers/hv/mshv_eventfd.c
320
srcu_read_unlock(&pt->pt_irq_srcu, idx);
drivers/hv/mshv_eventfd.c
329
spin_lock_irqsave(&pt->pt_irqfds_lock, flags);
drivers/hv/mshv_eventfd.c
343
spin_unlock_irqrestore(&pt->pt_irqfds_lock, flags);
drivers/hv/mshv_eventfd.c
350
static void mshv_irqfd_update(struct mshv_partition *pt,
drivers/hv/mshv_eventfd.c
354
irqfd->irqfd_girq_ent = mshv_ret_girq_entry(pt,
drivers/hv/mshv_eventfd.c
360
void mshv_irqfd_routing_update(struct mshv_partition *pt)
drivers/hv/mshv_eventfd.c
364
spin_lock_irq(&pt->pt_irqfds_lock);
drivers/hv/mshv_eventfd.c
365
hlist_for_each_entry(irqfd, &pt->pt_irqfds_list, irqfd_hnode)
drivers/hv/mshv_eventfd.c
366
mshv_irqfd_update(pt, irqfd);
drivers/hv/mshv_eventfd.c
367
spin_unlock_irq(&pt->pt_irqfds_lock);
drivers/hv/mshv_eventfd.c
386
static int mshv_irqfd_assign(struct mshv_partition *pt,
drivers/hv/mshv_eventfd.c
401
irqfd->irqfd_partn = pt;
drivers/hv/mshv_eventfd.c
404
seqcount_spinlock_init(&irqfd->irqfd_irqe_sc, &pt->pt_irqfds_lock);
drivers/hv/mshv_eventfd.c
430
mutex_lock(&pt->irqfds_resampler_lock);
drivers/hv/mshv_eventfd.c
432
hlist_for_each_entry(rp, &pt->irqfds_resampler_list,
drivers/hv/mshv_eventfd.c
445
mutex_unlock(&pt->irqfds_resampler_lock);
drivers/hv/mshv_eventfd.c
449
rp->rsmplr_partn = pt;
drivers/hv/mshv_eventfd.c
456
&pt->irqfds_resampler_list);
drivers/hv/mshv_eventfd.c
457
mshv_register_irq_ack_notifier(pt,
drivers/hv/mshv_eventfd.c
465
mutex_unlock(&pt->irqfds_resampler_lock);
drivers/hv/mshv_eventfd.c
475
spin_lock_irq(&pt->pt_irqfds_lock);
drivers/hv/mshv_eventfd.c
483
spin_unlock_irq(&pt->pt_irqfds_lock);
drivers/hv/mshv_eventfd.c
489
hlist_for_each_entry(tmp, &pt->pt_irqfds_list, irqfd_hnode) {
drivers/hv/mshv_eventfd.c
494
spin_unlock_irq(&pt->pt_irqfds_lock);
drivers/hv/mshv_eventfd.c
498
idx = srcu_read_lock(&pt->pt_irq_srcu);
drivers/hv/mshv_eventfd.c
499
mshv_irqfd_update(pt, irqfd);
drivers/hv/mshv_eventfd.c
500
hlist_add_head(&irqfd->irqfd_hnode, &pt->pt_irqfds_list);
drivers/hv/mshv_eventfd.c
501
spin_unlock_irq(&pt->pt_irqfds_lock);
drivers/hv/mshv_eventfd.c
512
srcu_read_unlock(&pt->pt_irq_srcu, idx);
drivers/hv/mshv_eventfd.c
533
static int mshv_irqfd_deassign(struct mshv_partition *pt,
drivers/hv/mshv_eventfd.c
544
hlist_for_each_entry_safe(irqfd, n, &pt->pt_irqfds_list,
drivers/hv/mshv_eventfd.c
564
int mshv_set_unset_irqfd(struct mshv_partition *pt,
drivers/hv/mshv_eventfd.c
571
return mshv_irqfd_deassign(pt, args);
drivers/hv/mshv_eventfd.c
573
return mshv_irqfd_assign(pt, args);
drivers/hv/mshv_eventfd.c
580
static void mshv_irqfd_release(struct mshv_partition *pt)
drivers/hv/mshv_eventfd.c
585
spin_lock_irq(&pt->pt_irqfds_lock);
drivers/hv/mshv_eventfd.c
587
hlist_for_each_entry_safe(irqfd, n, &pt->pt_irqfds_list, irqfd_hnode)
drivers/hv/mshv_eventfd.c
590
spin_unlock_irq(&pt->pt_irqfds_lock);
drivers/hv/mshv_eventfd.c
646
static bool ioeventfd_check_collision(struct mshv_partition *pt,
drivers/hv/mshv_eventfd.c
648
__must_hold(&pt->mutex)
drivers/hv/mshv_eventfd.c
652
hlist_for_each_entry(_p, &pt->ioeventfds_list, iovntfd_hnode)
drivers/hv/mshv_eventfd.c
662
static int mshv_assign_ioeventfd(struct mshv_partition *pt,
drivers/hv/mshv_eventfd.c
664
__must_hold(&pt->mutex)
drivers/hv/mshv_eventfd.c
672
WARN_ON_ONCE(!mutex_is_locked(&pt->pt_mutex));
drivers/hv/mshv_eventfd.c
728
if (ioeventfd_check_collision(pt, p)) {
drivers/hv/mshv_eventfd.c
733
ret = mshv_register_doorbell(pt->pt_id, ioeventfd_mmio_write,
drivers/hv/mshv_eventfd.c
734
(void *)pt, p->iovntfd_addr,
drivers/hv/mshv_eventfd.c
741
hlist_add_head_rcu(&p->iovntfd_hnode, &pt->ioeventfds_list);
drivers/hv/mshv_eventfd.c
754
static int mshv_deassign_ioeventfd(struct mshv_partition *pt,
drivers/hv/mshv_eventfd.c
756
__must_hold(&pt->mutex)
drivers/hv/mshv_eventfd.c
764
WARN_ON_ONCE(!mutex_is_locked(&pt->pt_mutex));
drivers/hv/mshv_eventfd.c
770
hlist_for_each_entry_safe(p, n, &pt->ioeventfds_list, iovntfd_hnode) {
drivers/hv/mshv_eventfd.c
785
ioeventfd_release(p, pt->pt_id);
drivers/hv/mshv_eventfd.c
795
int mshv_set_unset_ioeventfd(struct mshv_partition *pt,
drivers/hv/mshv_eventfd.c
797
__must_hold(&pt->mutex)
drivers/hv/mshv_eventfd.c
808
return mshv_deassign_ioeventfd(pt, args);
drivers/hv/mshv_eventfd.c
810
return mshv_assign_ioeventfd(pt, args);
drivers/hv/mshv_eventfd.c
813
void mshv_eventfd_init(struct mshv_partition *pt)
drivers/hv/mshv_eventfd.c
815
spin_lock_init(&pt->pt_irqfds_lock);
drivers/hv/mshv_eventfd.c
816
INIT_HLIST_HEAD(&pt->pt_irqfds_list);
drivers/hv/mshv_eventfd.c
818
INIT_HLIST_HEAD(&pt->irqfds_resampler_list);
drivers/hv/mshv_eventfd.c
819
mutex_init(&pt->irqfds_resampler_lock);
drivers/hv/mshv_eventfd.c
821
INIT_HLIST_HEAD(&pt->ioeventfds_list);
drivers/hv/mshv_eventfd.c
824
void mshv_eventfd_release(struct mshv_partition *pt)
drivers/hv/mshv_eventfd.c
830
hlist_move_list(&pt->ioeventfds_list, &items);
drivers/hv/mshv_eventfd.c
835
ioeventfd_release(p, pt->pt_id);
drivers/hv/mshv_eventfd.c
838
mshv_irqfd_release(pt);
drivers/hv/mshv_eventfd.h
67
int mshv_set_unset_ioeventfd(struct mshv_partition *pt,
drivers/hv/mshv_vtl_main.c
450
static void mshv_vtl_ptable_queue_proc(struct file *file, wait_queue_head_t *wqh, poll_table *pt)
drivers/hv/mshv_vtl_main.c
452
struct mshv_vtl_poll_file *poll_file = container_of(pt, struct mshv_vtl_poll_file, pt);
drivers/hv/mshv_vtl_main.c
496
init_poll_funcptr(&poll_file->pt, mshv_vtl_ptable_queue_proc);
drivers/hv/mshv_vtl_main.c
497
vfs_poll(file, &poll_file->pt);
drivers/hv/mshv_vtl_main.c
74
poll_table pt;
drivers/infiniband/hw/hfi1/file_ops.c
1438
struct poll_table_struct *pt)
drivers/infiniband/hw/hfi1/file_ops.c
1445
poll_wait(fp, &uctxt->wait, pt);
drivers/infiniband/hw/hfi1/file_ops.c
1461
struct poll_table_struct *pt)
drivers/infiniband/hw/hfi1/file_ops.c
1468
poll_wait(fp, &uctxt->wait, pt);
drivers/infiniband/hw/hfi1/file_ops.c
37
static __poll_t hfi1_poll(struct file *fp, struct poll_table_struct *pt);
drivers/infiniband/hw/hfi1/file_ops.c
609
static __poll_t hfi1_poll(struct file *fp, struct poll_table_struct *pt)
drivers/infiniband/hw/hfi1/file_ops.c
618
pollflag = poll_urgent(fp, pt);
drivers/infiniband/hw/hfi1/file_ops.c
620
pollflag = poll_next(fp, pt);
drivers/infiniband/hw/hfi1/file_ops.c
65
static __poll_t poll_urgent(struct file *fp, struct poll_table_struct *pt);
drivers/infiniband/hw/hfi1/file_ops.c
66
static __poll_t poll_next(struct file *fp, struct poll_table_struct *pt);
drivers/iommu/iommu.c
771
bool pt;
drivers/iommu/iommu.c
774
ret = kstrtobool(str, &pt);
drivers/iommu/iommu.c
778
if (pt)
drivers/iommu/tegra-smmu.c
580
static u32 *tegra_smmu_pte_offset(struct tegra_pt *pt, unsigned long iova)
drivers/iommu/tegra-smmu.c
582
return &pt->val[iova_pt_index(iova)];
drivers/iommu/tegra-smmu.c
590
struct tegra_pt *pt;
drivers/iommu/tegra-smmu.c
592
pt = as->pts[pd_index];
drivers/iommu/tegra-smmu.c
593
if (!pt)
drivers/iommu/tegra-smmu.c
598
return tegra_smmu_pte_offset(pt, iova);
drivers/iommu/tegra-smmu.c
602
dma_addr_t *dmap, struct tegra_pt *pt)
drivers/iommu/tegra-smmu.c
610
dma = dma_map_single(smmu->dev, pt, SMMU_SIZE_PT,
drivers/iommu/tegra-smmu.c
613
iommu_free_pages(pt);
drivers/iommu/tegra-smmu.c
620
iommu_free_pages(pt);
drivers/iommu/tegra-smmu.c
624
as->pts[pde] = pt;
drivers/iommu/tegra-smmu.c
647
struct tegra_pt *pt = as->pts[pde];
drivers/iommu/tegra-smmu.c
661
iommu_free_pages(pt);
drivers/iommu/tegra-smmu.c
686
struct tegra_pt *pt = as->pts[pde];
drivers/iommu/tegra-smmu.c
689
if (pt)
drivers/iommu/tegra-smmu.c
690
return pt;
drivers/iommu/tegra-smmu.c
700
pt = iommu_alloc_pages_sz(gfp | __GFP_DMA, SMMU_SIZE_PT);
drivers/iommu/tegra-smmu.c
711
if (pt)
drivers/iommu/tegra-smmu.c
712
iommu_free_pages(pt);
drivers/iommu/tegra-smmu.c
714
pt = as->pts[pde];
drivers/iommu/tegra-smmu.c
717
return pt;
drivers/iommu/tegra-smmu.c
727
struct tegra_pt *pt;
drivers/iommu/tegra-smmu.c
731
pt = as_get_pde_page(as, iova, gfp, flags);
drivers/iommu/tegra-smmu.c
732
if (!pt)
drivers/iommu/tegra-smmu.c
735
pte = as_get_pte(as, iova, &pte_dma, pt);
drivers/irqchip/irq-gic-v3-its.c
3052
static void its_free_pending_table(struct page *pt)
drivers/irqchip/irq-gic-v3-its.c
3054
its_free_pages(page_address(pt), get_order(LPI_PENDBASE_SZ));
drivers/isdn/hardware/mISDN/hfcmulti.c
3827
int i, pt;
drivers/isdn/hardware/mISDN/hfcmulti.c
3833
pt = hc->chan[i].port;
drivers/isdn/hardware/mISDN/hfcmulti.c
3836
hc->chan[hc->dnum[pt]].slot_tx = -1;
drivers/isdn/hardware/mISDN/hfcmulti.c
3837
hc->chan[hc->dnum[pt]].slot_rx = -1;
drivers/isdn/hardware/mISDN/hfcmulti.c
3838
hc->chan[hc->dnum[pt]].conf = -1;
drivers/isdn/hardware/mISDN/hfcmulti.c
3839
if (hc->dnum[pt]) {
drivers/isdn/hardware/mISDN/hfcmulti.c
3845
if (!((1 << i) & hc->bmask[pt])) /* skip unused chan */
drivers/isdn/hardware/mISDN/hfcmulti.c
3853
if (hc->ctype == HFC_TYPE_E1 && pt == 0) {
drivers/isdn/hardware/mISDN/hfcmulti.c
3958
HFC_outb(hc, R_ST_SEL, pt);
drivers/isdn/hardware/mISDN/hfcmulti.c
3965
__func__, pt);
drivers/isdn/hardware/mISDN/hfcmulti.c
3969
hc->hw.a_st_ctrl0[pt] = V_ST_MD;
drivers/isdn/hardware/mISDN/hfcmulti.c
3974
__func__, pt);
drivers/isdn/hardware/mISDN/hfcmulti.c
3978
hc->hw.a_st_ctrl0[pt] = 0;
drivers/isdn/hardware/mISDN/hfcmulti.c
3981
hc->hw.a_st_ctrl0[pt] |= V_TX_LI;
drivers/isdn/hardware/mISDN/hfcmulti.c
3983
hc->hw.a_st_ctrl0[pt] |= 0x40 /* V_ST_PU_CTRL */;
drivers/isdn/hardware/mISDN/hfcmulti.c
3988
HFC_outb(hc, A_ST_CTRL0, hc->hw.a_st_ctrl0[pt]);
drivers/isdn/hardware/mISDN/hfcmulti.c
4001
hc->hw.r_sci_msk |= 1 << pt;
drivers/isdn/hardware/mISDN/hfcmulti.c
4491
int pt, ci, i = 0;
drivers/isdn/hardware/mISDN/hfcmulti.c
4496
pt = hc->chan[ci].port;
drivers/isdn/hardware/mISDN/hfcmulti.c
4500
__func__, pt + 1);
drivers/isdn/hardware/mISDN/hfcmulti.c
4502
if (pt >= hc->ports) {
drivers/isdn/hardware/mISDN/hfcmulti.c
4504
__func__, pt + 1);
drivers/isdn/hardware/mISDN/hfcmulti.c
4510
__func__, pt + 1);
drivers/isdn/hardware/mISDN/hfcmulti.c
4517
if (hc->created[pt]) {
drivers/isdn/hardware/mISDN/hfcmulti.c
4518
hc->created[pt] = 0;
drivers/isdn/hardware/mISDN/hfcmulti.c
4537
if (!((1 << i) & hc->bmask[pt])) /* skip unused chan */
drivers/isdn/hardware/mISDN/hfcmulti.c
4595
pt+1, ci);
drivers/isdn/hardware/mISDN/hfcmulti.c
4772
init_e1_port(struct hfc_multi *hc, struct hm_map *m, int pt)
drivers/isdn/hardware/mISDN/hfcmulti.c
4791
dch->slot = hc->dnum[pt];
drivers/isdn/hardware/mISDN/hfcmulti.c
4792
hc->chan[hc->dnum[pt]].dch = dch;
drivers/isdn/hardware/mISDN/hfcmulti.c
4793
hc->chan[hc->dnum[pt]].port = pt;
drivers/isdn/hardware/mISDN/hfcmulti.c
4794
hc->chan[hc->dnum[pt]].nt_timer = -1;
drivers/isdn/hardware/mISDN/hfcmulti.c
4796
if (!((1 << ch) & hc->bmask[pt])) /* skip unused channel */
drivers/isdn/hardware/mISDN/hfcmulti.c
4823
hc->chan[ch].port = pt;
drivers/isdn/hardware/mISDN/hfcmulti.c
4828
if (pt == 0)
drivers/isdn/hardware/mISDN/hfcmulti.c
4832
HFC_cnt + 1, pt+1);
drivers/isdn/hardware/mISDN/hfcmulti.c
4838
hc->created[pt] = 1;
drivers/isdn/hardware/mISDN/hfcmulti.c
4846
init_multi_port(struct hfc_multi *hc, int pt)
drivers/isdn/hardware/mISDN/hfcmulti.c
4865
i = pt << 2;
drivers/isdn/hardware/mISDN/hfcmulti.c
4868
hc->chan[i + 2].port = pt;
drivers/isdn/hardware/mISDN/hfcmulti.c
4896
hc->chan[i + ch].port = pt;
drivers/isdn/hardware/mISDN/hfcmulti.c
4905
__func__, HFC_cnt + 1, pt + 1);
drivers/isdn/hardware/mISDN/hfcmulti.c
4910
pt + 1, HFC_cnt + 1);
drivers/isdn/hardware/mISDN/hfcmulti.c
4918
pt + 1, HFC_cnt + 1, hc->masterclk + 1);
drivers/isdn/hardware/mISDN/hfcmulti.c
4922
hc->masterclk = pt;
drivers/isdn/hardware/mISDN/hfcmulti.c
4930
__func__, HFC_cnt + 1, pt + 1);
drivers/isdn/hardware/mISDN/hfcmulti.c
4940
__func__, HFC_cnt + 1, pt + 1);
drivers/isdn/hardware/mISDN/hfcmulti.c
4946
HFC_cnt + 1, pt + 1);
drivers/isdn/hardware/mISDN/hfcmulti.c
4950
hc->ctype, HFC_cnt + 1, pt + 1);
drivers/isdn/hardware/mISDN/hfcmulti.c
4955
hc->created[pt] = 1;
drivers/isdn/hardware/mISDN/hfcmulti.c
4967
int pt;
drivers/isdn/hardware/mISDN/hfcmulti.c
5008
pt = 0;
drivers/isdn/hardware/mISDN/hfcmulti.c
5013
hc->dnum[pt] = ch;
drivers/isdn/hardware/mISDN/hfcmulti.c
5014
hc->bmask[pt] = bmask[bmask_cnt++];
drivers/isdn/hardware/mISDN/hfcmulti.c
5015
if ((maskcheck & hc->bmask[pt])
drivers/isdn/hardware/mISDN/hfcmulti.c
5016
|| (dmask[E1_cnt] & hc->bmask[pt])) {
drivers/isdn/hardware/mISDN/hfcmulti.c
5019
E1_cnt + 1, pt);
drivers/isdn/hardware/mISDN/hfcmulti.c
5023
maskcheck |= hc->bmask[pt];
drivers/isdn/hardware/mISDN/hfcmulti.c
5026
E1_cnt + 1, ch, hc->bmask[pt]);
drivers/isdn/hardware/mISDN/hfcmulti.c
5027
pt++;
drivers/isdn/hardware/mISDN/hfcmulti.c
5029
hc->ports = pt;
drivers/isdn/hardware/mISDN/hfcmulti.c
5112
for (pt = 0; pt < hc->ports; pt++) {
drivers/isdn/hardware/mISDN/hfcmulti.c
5120
ret_err = init_e1_port(hc, m, pt);
drivers/isdn/hardware/mISDN/hfcmulti.c
5122
ret_err = init_multi_port(hc, pt);
drivers/isdn/hardware/mISDN/hfcmulti.c
5127
__func__, HFC_cnt + 1, pt + 1, ret_err);
drivers/isdn/hardware/mISDN/hfcmulti.c
5130
while (pt) { /* release already registered ports */
drivers/isdn/hardware/mISDN/hfcmulti.c
5131
pt--;
drivers/isdn/hardware/mISDN/hfcmulti.c
5134
hc->chan[hc->dnum[pt]].dch);
drivers/isdn/hardware/mISDN/hfcmulti.c
5137
hc->chan[(pt << 2) + 2].dch);
drivers/md/dm-thin.c
2528
struct pool_c *pt = pool->ti->private;
drivers/md/dm-thin.c
2530
if (pt->adjusted_pf.discard_passdown) {
drivers/md/dm-thin.c
2542
struct pool_c *pt = pool->ti->private;
drivers/md/dm-thin.c
2617
pool->pf.error_if_no_space = pt->requested_pf.error_if_no_space;
drivers/md/dm-thin.c
2632
pt->adjusted_pf.mode = new_mode;
drivers/md/dm-thin.c
2839
static void disable_discard_passdown_if_not_supported(struct pool_c *pt)
drivers/md/dm-thin.c
2841
struct pool *pool = pt->pool;
drivers/md/dm-thin.c
2842
struct block_device *data_bdev = pt->data_dev->bdev;
drivers/md/dm-thin.c
2846
if (!pt->adjusted_pf.discard_passdown)
drivers/md/dm-thin.c
2849
if (!bdev_max_discard_sectors(pt->data_dev->bdev))
drivers/md/dm-thin.c
2857
pt->adjusted_pf.discard_passdown = false;
drivers/md/dm-thin.c
2863
struct pool_c *pt = ti->private;
drivers/md/dm-thin.c
2869
enum pool_mode new_mode = pt->adjusted_pf.mode;
drivers/md/dm-thin.c
2876
pt->adjusted_pf.mode = old_mode;
drivers/md/dm-thin.c
2879
pool->pf = pt->adjusted_pf;
drivers/md/dm-thin.c
2880
pool->low_water_blocks = pt->low_water_blocks;
drivers/md/dm-thin.c
3128
struct pool_c *pt = ti->private;
drivers/md/dm-thin.c
3132
unbind_control_target(pt->pool, ti);
drivers/md/dm-thin.c
3133
__pool_dec(pt->pool);
drivers/md/dm-thin.c
3134
dm_put_device(ti, pt->metadata_dev);
drivers/md/dm-thin.c
3135
dm_put_device(ti, pt->data_dev);
drivers/md/dm-thin.c
3136
kfree(pt);
drivers/md/dm-thin.c
3258
static dm_block_t calc_metadata_threshold(struct pool_c *pt)
drivers/md/dm-thin.c
3265
dm_block_t quarter = get_metadata_dev_size_in_blocks(pt->metadata_dev->bdev) / 4;
drivers/md/dm-thin.c
3286
struct pool_c *pt;
drivers/md/dm-thin.c
3357
pt = kzalloc_obj(*pt);
drivers/md/dm-thin.c
3358
if (!pt) {
drivers/md/dm-thin.c
3382
pt->pool = pool;
drivers/md/dm-thin.c
3383
pt->ti = ti;
drivers/md/dm-thin.c
3384
pt->metadata_dev = metadata_dev;
drivers/md/dm-thin.c
3385
pt->data_dev = data_dev;
drivers/md/dm-thin.c
3386
pt->low_water_blocks = low_water_blocks;
drivers/md/dm-thin.c
3387
pt->adjusted_pf = pt->requested_pf = pf;
drivers/md/dm-thin.c
3406
ti->private = pt;
drivers/md/dm-thin.c
3408
r = dm_pool_register_metadata_threshold(pt->pool->pmd,
drivers/md/dm-thin.c
3409
calc_metadata_threshold(pt),
drivers/md/dm-thin.c
3427
kfree(pt);
drivers/md/dm-thin.c
3440
struct pool_c *pt = ti->private;
drivers/md/dm-thin.c
3441
struct pool *pool = pt->pool;
drivers/md/dm-thin.c
3447
bio_set_dev(bio, pt->data_dev->bdev);
drivers/md/dm-thin.c
3456
struct pool_c *pt = ti->private;
drivers/md/dm-thin.c
3457
struct pool *pool = pt->pool;
drivers/md/dm-thin.c
3504
struct pool_c *pt = ti->private;
drivers/md/dm-thin.c
3505
struct pool *pool = pt->pool;
drivers/md/dm-thin.c
3567
struct pool_c *pt = ti->private;
drivers/md/dm-thin.c
3568
struct pool *pool = pt->pool;
drivers/md/dm-thin.c
3625
struct pool_c *pt = ti->private;
drivers/md/dm-thin.c
3626
struct pool *pool = pt->pool;
drivers/md/dm-thin.c
3645
struct pool_c *pt = ti->private;
drivers/md/dm-thin.c
3646
struct pool *pool = pt->pool;
drivers/md/dm-thin.c
3657
struct pool_c *pt = ti->private;
drivers/md/dm-thin.c
3658
struct pool *pool = pt->pool;
drivers/md/dm-thin.c
3669
struct pool_c *pt = ti->private;
drivers/md/dm-thin.c
3670
struct pool *pool = pt->pool;
drivers/md/dm-thin.c
3846
struct pool_c *pt = ti->private;
drivers/md/dm-thin.c
3847
struct pool *pool = pt->pool;
drivers/md/dm-thin.c
3926
struct pool_c *pt = ti->private;
drivers/md/dm-thin.c
3927
struct pool *pool = pt->pool;
drivers/md/dm-thin.c
4019
DMEMIT("%llu ", (unsigned long long)calc_metadata_threshold(pt));
drivers/md/dm-thin.c
4025
format_dev_t(buf, pt->metadata_dev->bdev->bd_dev),
drivers/md/dm-thin.c
4026
format_dev_t(buf2, pt->data_dev->bdev->bd_dev),
drivers/md/dm-thin.c
4028
(unsigned long long)pt->low_water_blocks);
drivers/md/dm-thin.c
4029
emit_flags(&pt->requested_pf, result, sz, maxlen);
drivers/md/dm-thin.c
4045
struct pool_c *pt = ti->private;
drivers/md/dm-thin.c
4047
return fn(ti, pt->data_dev, 0, ti->len, data);
drivers/md/dm-thin.c
4052
struct pool_c *pt = ti->private;
drivers/md/dm-thin.c
4053
struct pool *pool = pt->pool;
drivers/md/dm-thin.c
4092
if (pt->adjusted_pf.discard_enabled) {
drivers/md/dm-thin.c
4093
disable_discard_passdown_if_not_supported(pt);
drivers/md/dm-thin.c
4094
if (!pt->adjusted_pf.discard_passdown)
drivers/media/common/saa7146/saa7146_core.c
166
void *saa7146_vmalloc_build_pgtable(struct pci_dev *pci, long length, struct saa7146_pgtable *pt)
drivers/media/common/saa7146/saa7146_core.c
175
if (!(pt->slist = vmalloc_to_sg(mem, pages)))
drivers/media/common/saa7146/saa7146_core.c
178
if (saa7146_pgtable_alloc(pci, pt))
drivers/media/common/saa7146/saa7146_core.c
181
pt->nents = pages;
drivers/media/common/saa7146/saa7146_core.c
182
slen = dma_map_sg(&pci->dev, pt->slist, pt->nents, DMA_FROM_DEVICE);
drivers/media/common/saa7146/saa7146_core.c
186
if (0 != saa7146_pgtable_build_single(pci, pt, pt->slist, slen))
drivers/media/common/saa7146/saa7146_core.c
192
dma_unmap_sg(&pci->dev, pt->slist, pt->nents, DMA_FROM_DEVICE);
drivers/media/common/saa7146/saa7146_core.c
194
saa7146_pgtable_free(pci, pt);
drivers/media/common/saa7146/saa7146_core.c
196
kfree(pt->slist);
drivers/media/common/saa7146/saa7146_core.c
197
pt->slist = NULL;
drivers/media/common/saa7146/saa7146_core.c
204
void saa7146_vfree_destroy_pgtable(struct pci_dev *pci, void *mem, struct saa7146_pgtable *pt)
drivers/media/common/saa7146/saa7146_core.c
206
dma_unmap_sg(&pci->dev, pt->slist, pt->nents, DMA_FROM_DEVICE);
drivers/media/common/saa7146/saa7146_core.c
207
saa7146_pgtable_free(pci, pt);
drivers/media/common/saa7146/saa7146_core.c
208
kfree(pt->slist);
drivers/media/common/saa7146/saa7146_core.c
209
pt->slist = NULL;
drivers/media/common/saa7146/saa7146_core.c
213
void saa7146_pgtable_free(struct pci_dev *pci, struct saa7146_pgtable *pt)
drivers/media/common/saa7146/saa7146_core.c
215
if (NULL == pt->cpu)
drivers/media/common/saa7146/saa7146_core.c
217
dma_free_coherent(&pci->dev, pt->size, pt->cpu, pt->dma);
drivers/media/common/saa7146/saa7146_core.c
218
pt->cpu = NULL;
drivers/media/common/saa7146/saa7146_core.c
221
int saa7146_pgtable_alloc(struct pci_dev *pci, struct saa7146_pgtable *pt)
drivers/media/common/saa7146/saa7146_core.c
230
pt->size = PAGE_SIZE;
drivers/media/common/saa7146/saa7146_core.c
231
pt->cpu = cpu;
drivers/media/common/saa7146/saa7146_core.c
232
pt->dma = dma_addr;
drivers/media/common/saa7146/saa7146_core.c
237
int saa7146_pgtable_build_single(struct pci_dev *pci, struct saa7146_pgtable *pt,
drivers/media/common/saa7146/saa7146_core.c
251
pt->offset = list->offset;
drivers/media/common/saa7146/saa7146_core.c
253
ptr = pt->cpu;
drivers/media/common/saa7146/saa7146_hlp.c
430
vdma1.base_page = buf->pt[0].dma | ME1 | sfmt->swap;
drivers/media/common/saa7146/saa7146_hlp.c
433
vdma1.prot_addr = buf->pt[0].offset;
drivers/media/common/saa7146/saa7146_hlp.c
434
vdma1.base_even = buf->pt[0].offset+(vdma1.pitch/2)*height;
drivers/media/common/saa7146/saa7146_hlp.c
437
vdma1.base_even = buf->pt[0].offset;
drivers/media/common/saa7146/saa7146_hlp.c
439
vdma1.prot_addr = buf->pt[0].offset+(vdma1.pitch/2)*height;
drivers/media/common/saa7146/saa7146_hlp.c
482
vdma2->prot_addr = buf->pt[1].offset;
drivers/media/common/saa7146/saa7146_hlp.c
483
vdma2->base_even = ((vdma2->pitch/2)*height)+buf->pt[1].offset;
drivers/media/common/saa7146/saa7146_hlp.c
486
vdma3->prot_addr = buf->pt[2].offset;
drivers/media/common/saa7146/saa7146_hlp.c
487
vdma3->base_even = ((vdma3->pitch/2)*height)+buf->pt[2].offset;
drivers/media/common/saa7146/saa7146_hlp.c
490
vdma3->base_even = buf->pt[2].offset;
drivers/media/common/saa7146/saa7146_hlp.c
492
vdma3->prot_addr = (vdma3->pitch/2)*height+buf->pt[2].offset;
drivers/media/common/saa7146/saa7146_hlp.c
494
vdma2->base_even = buf->pt[1].offset;
drivers/media/common/saa7146/saa7146_hlp.c
496
vdma2->prot_addr = (vdma2->pitch/2)*height+buf->pt[1].offset;
drivers/media/common/saa7146/saa7146_hlp.c
512
vdma2->prot_addr = buf->pt[2].offset;
drivers/media/common/saa7146/saa7146_hlp.c
513
vdma2->base_even = ((vdma2->pitch/2)*height)+buf->pt[2].offset;
drivers/media/common/saa7146/saa7146_hlp.c
516
vdma3->prot_addr = buf->pt[1].offset;
drivers/media/common/saa7146/saa7146_hlp.c
517
vdma3->base_even = ((vdma3->pitch/2)*height)+buf->pt[1].offset;
drivers/media/common/saa7146/saa7146_hlp.c
521
vdma3->base_even = buf->pt[2].offset;
drivers/media/common/saa7146/saa7146_hlp.c
523
vdma3->prot_addr = (vdma3->pitch/2)*height+buf->pt[2].offset;
drivers/media/common/saa7146/saa7146_hlp.c
525
vdma2->base_even = buf->pt[1].offset;
drivers/media/common/saa7146/saa7146_hlp.c
527
vdma2->prot_addr = (vdma2->pitch/2)*height+buf->pt[1].offset;
drivers/media/common/saa7146/saa7146_hlp.c
545
if (WARN_ON(!buf->pt[0].dma) ||
drivers/media/common/saa7146/saa7146_hlp.c
546
WARN_ON(!buf->pt[1].dma) ||
drivers/media/common/saa7146/saa7146_hlp.c
547
WARN_ON(!buf->pt[2].dma))
drivers/media/common/saa7146/saa7146_hlp.c
561
vdma1.base_page = buf->pt[0].dma | ME1;
drivers/media/common/saa7146/saa7146_hlp.c
564
vdma1.prot_addr = buf->pt[0].offset;
drivers/media/common/saa7146/saa7146_hlp.c
565
vdma1.base_even = ((vdma1.pitch/2)*height)+buf->pt[0].offset;
drivers/media/common/saa7146/saa7146_hlp.c
568
vdma1.base_even = buf->pt[0].offset;
drivers/media/common/saa7146/saa7146_hlp.c
570
vdma1.prot_addr = (vdma1.pitch/2)*height+buf->pt[0].offset;
drivers/media/common/saa7146/saa7146_hlp.c
574
vdma2.base_page = buf->pt[1].dma | ME1;
drivers/media/common/saa7146/saa7146_hlp.c
577
vdma3.base_page = buf->pt[2].dma | ME1;
drivers/media/common/saa7146/saa7146_vbi.c
153
vdma3.base_even = buf->pt[2].offset;
drivers/media/common/saa7146/saa7146_vbi.c
154
vdma3.base_odd = buf->pt[2].offset + 16 * vbi_pixel_to_capture;
drivers/media/common/saa7146/saa7146_vbi.c
155
vdma3.prot_addr = buf->pt[2].offset + 16 * 2 * vbi_pixel_to_capture;
drivers/media/common/saa7146/saa7146_vbi.c
157
vdma3.base_page = buf->pt[2].dma | ME1;
drivers/media/common/saa7146/saa7146_vbi.c
261
saa7146_pgtable_alloc(dev->pci, &buf->pt[2]);
drivers/media/common/saa7146/saa7146_vbi.c
263
ret = saa7146_pgtable_build_single(dev->pci, &buf->pt[2],
drivers/media/common/saa7146/saa7146_vbi.c
266
saa7146_pgtable_free(dev->pci, &buf->pt[2]);
drivers/media/common/saa7146/saa7146_vbi.c
287
saa7146_pgtable_free(dev->pci, &buf->pt[2]);
drivers/media/common/saa7146/saa7146_video.c
100
struct saa7146_pgtable *pt2 = &buf->pt[1];
drivers/media/common/saa7146/saa7146_video.c
101
struct saa7146_pgtable *pt3 = &buf->pt[2];
drivers/media/common/saa7146/saa7146_video.c
170
struct saa7146_pgtable *pt = &buf->pt[0];
drivers/media/common/saa7146/saa7146_video.c
172
return saa7146_pgtable_build_single(pci, pt, list, length);
drivers/media/common/saa7146/saa7146_video.c
551
saa7146_pgtable_free(dev->pci, &buf->pt[0]);
drivers/media/common/saa7146/saa7146_video.c
552
saa7146_pgtable_free(dev->pci, &buf->pt[1]);
drivers/media/common/saa7146/saa7146_video.c
553
saa7146_pgtable_free(dev->pci, &buf->pt[2]);
drivers/media/common/saa7146/saa7146_video.c
599
saa7146_pgtable_alloc(dev->pci, &buf->pt[0]);
drivers/media/common/saa7146/saa7146_video.c
600
saa7146_pgtable_alloc(dev->pci, &buf->pt[1]);
drivers/media/common/saa7146/saa7146_video.c
601
saa7146_pgtable_alloc(dev->pci, &buf->pt[2]);
drivers/media/common/saa7146/saa7146_video.c
603
saa7146_pgtable_alloc(dev->pci, &buf->pt[0]);
drivers/media/common/saa7146/saa7146_video.c
99
struct saa7146_pgtable *pt1 = &buf->pt[0];
drivers/media/pci/intel/ipu6/ipu6-fw-isys.c
434
dev_dbg(dev, "\t.pt = %d\n", cfg->output_pins[i].pt);
drivers/media/pci/intel/ipu6/ipu6-fw-isys.h
328
u8 pt;
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
500
output_pin->pt = IPU6_FW_ISYS_PIN_TYPE_RAW_SOC;
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
502
output_pin->pt = IPU6_FW_ISYS_PIN_TYPE_MIPI;
drivers/media/pci/intel/ipu6/ipu6-mmu.c
141
void *pt = (void *)get_zeroed_page(GFP_ATOMIC | GFP_DMA32);
drivers/media/pci/intel/ipu6/ipu6-mmu.c
144
if (!pt)
drivers/media/pci/intel/ipu6/ipu6-mmu.c
147
dev_dbg(mmu_info->dev, "dummy_page: get_zeroed_page() == %p\n", pt);
drivers/media/pci/intel/ipu6/ipu6-mmu.c
149
dma = map_single(mmu_info, pt);
drivers/media/pci/intel/ipu6/ipu6-mmu.c
155
mmu_info->dummy_page = pt;
drivers/media/pci/intel/ipu6/ipu6-mmu.c
161
free_page((unsigned long)pt);
drivers/media/pci/intel/ipu6/ipu6-mmu.c
175
u32 *pt = (u32 *)get_zeroed_page(GFP_ATOMIC | GFP_DMA32);
drivers/media/pci/intel/ipu6/ipu6-mmu.c
179
if (!pt)
drivers/media/pci/intel/ipu6/ipu6-mmu.c
182
dev_dbg(mmu_info->dev, "dummy_l2: get_zeroed_page() = %p\n", pt);
drivers/media/pci/intel/ipu6/ipu6-mmu.c
184
dma = map_single(mmu_info, pt);
drivers/media/pci/intel/ipu6/ipu6-mmu.c
191
pt[i] = mmu_info->dummy_page_pteval;
drivers/media/pci/intel/ipu6/ipu6-mmu.c
193
mmu_info->dummy_l2_pt = pt;
drivers/media/pci/intel/ipu6/ipu6-mmu.c
199
free_page((unsigned long)pt);
drivers/media/pci/intel/ipu6/ipu6-mmu.c
213
u32 *pt = (u32 *)get_zeroed_page(GFP_ATOMIC | GFP_DMA32);
drivers/media/pci/intel/ipu6/ipu6-mmu.c
217
if (!pt)
drivers/media/pci/intel/ipu6/ipu6-mmu.c
220
dev_dbg(mmu_info->dev, "alloc_l1: get_zeroed_page() = %p\n", pt);
drivers/media/pci/intel/ipu6/ipu6-mmu.c
223
pt[i] = mmu_info->dummy_l2_pteval;
drivers/media/pci/intel/ipu6/ipu6-mmu.c
225
dma = map_single(mmu_info, pt);
drivers/media/pci/intel/ipu6/ipu6-mmu.c
232
dev_dbg(mmu_info->dev, "l1 pt %p mapped at %pad\n", pt, &dma);
drivers/media/pci/intel/ipu6/ipu6-mmu.c
234
return pt;
drivers/media/pci/intel/ipu6/ipu6-mmu.c
237
free_page((unsigned long)pt);
drivers/media/pci/intel/ipu6/ipu6-mmu.c
243
u32 *pt = (u32 *)get_zeroed_page(GFP_ATOMIC | GFP_DMA32);
drivers/media/pci/intel/ipu6/ipu6-mmu.c
246
if (!pt)
drivers/media/pci/intel/ipu6/ipu6-mmu.c
249
dev_dbg(mmu_info->dev, "alloc_l2: get_zeroed_page() = %p\n", pt);
drivers/media/pci/intel/ipu6/ipu6-mmu.c
252
pt[i] = mmu_info->dummy_page_pteval;
drivers/media/pci/intel/ipu6/ipu6-mmu.c
254
return pt;
drivers/media/pci/saa7134/saa7134-alsa.c
560
(dev->dmasound.pt.dma >> 12);
drivers/media/pci/saa7134/saa7134-alsa.c
687
saa7134_pgtable_free(dev->pci, &dev->dmasound.pt);
drivers/media/pci/saa7134/saa7134-alsa.c
709
err = saa7134_pgtable_alloc(dev->pci, &dev->dmasound.pt);
drivers/media/pci/saa7134/saa7134-alsa.c
715
err = saa7134_pgtable_build(dev->pci, &dev->dmasound.pt,
drivers/media/pci/saa7134/saa7134-alsa.c
718
saa7134_pgtable_free(dev->pci, &dev->dmasound.pt);
drivers/media/pci/saa7134/saa7134-alsa.c
754
saa7134_pgtable_free(dev->pci, &dev->dmasound.pt);
drivers/media/pci/saa7134/saa7134-core.c
217
int saa7134_pgtable_alloc(struct pci_dev *pci, struct saa7134_pgtable *pt)
drivers/media/pci/saa7134/saa7134-core.c
226
pt->size = SAA7134_PGTABLE_SIZE;
drivers/media/pci/saa7134/saa7134-core.c
227
pt->cpu = cpu;
drivers/media/pci/saa7134/saa7134-core.c
228
pt->dma = dma_addr;
drivers/media/pci/saa7134/saa7134-core.c
232
int saa7134_pgtable_build(struct pci_dev *pci, struct saa7134_pgtable *pt,
drivers/media/pci/saa7134/saa7134-core.c
239
BUG_ON(NULL == pt || NULL == pt->cpu);
drivers/media/pci/saa7134/saa7134-core.c
241
ptr = pt->cpu + startpage;
drivers/media/pci/saa7134/saa7134-core.c
250
void saa7134_pgtable_free(struct pci_dev *pci, struct saa7134_pgtable *pt)
drivers/media/pci/saa7134/saa7134-core.c
252
if (NULL == pt->cpu)
drivers/media/pci/saa7134/saa7134-core.c
254
dma_free_coherent(&pci->dev, pt->size, pt->cpu, pt->dma);
drivers/media/pci/saa7134/saa7134-core.c
255
pt->cpu = NULL;
drivers/media/pci/saa7134/saa7134-ts.c
101
return saa7134_pgtable_build(dev->pci, &dmaq->pt, dma->sgl, dma->nents,
drivers/media/pci/saa7134/saa7134-ts.c
219
saa7134_pgtable_alloc(dev->pci, &dev->ts_q.pt);
drivers/media/pci/saa7134/saa7134-ts.c
267
(dev->ts_q.pt.dma >> 12));
drivers/media/pci/saa7134/saa7134-ts.c
302
saa7134_pgtable_free(dev->pci, &dev->ts_q.pt);
drivers/media/pci/saa7134/saa7134-vbi.c
124
return saa7134_pgtable_build(dev->pci, &dmaq->pt, dma->sgl, dma->nents,
drivers/media/pci/saa7134/saa7134-vbi.c
88
(dmaq->pt.dma >> 12);
drivers/media/pci/saa7134/saa7134-video.c
1718
saa7134_pgtable_alloc(dev->pci, &dev->video_q.pt);
drivers/media/pci/saa7134/saa7134-video.c
1737
saa7134_pgtable_alloc(dev->pci, &dev->vbi_q.pt);
drivers/media/pci/saa7134/saa7134-video.c
1746
saa7134_pgtable_free(dev->pci, &dev->video_q.pt);
drivers/media/pci/saa7134/saa7134-video.c
1747
saa7134_pgtable_free(dev->pci, &dev->vbi_q.pt);
drivers/media/pci/saa7134/saa7134-video.c
660
(dmaq->pt.dma >> 12);
drivers/media/pci/saa7134/saa7134-video.c
746
return saa7134_pgtable_build(dev->pci, &dmaq->pt, dma->sgl, dma->nents,
drivers/media/pci/saa7134/saa7134.h
483
struct saa7134_pgtable pt;
drivers/media/pci/saa7134/saa7134.h
508
struct saa7134_pgtable pt;
drivers/media/pci/saa7134/saa7134.h
743
int saa7134_pgtable_alloc(struct pci_dev *pci, struct saa7134_pgtable *pt);
drivers/media/pci/saa7134/saa7134.h
744
int saa7134_pgtable_build(struct pci_dev *pci, struct saa7134_pgtable *pt,
drivers/media/pci/saa7134/saa7134.h
747
void saa7134_pgtable_free(struct pci_dev *pci, struct saa7134_pgtable *pt);
drivers/media/pci/ttpci/budget-core.c
133
saa7146_write(dev, BASE_PAGE3, budget->pt.dma | ME1 | 0x90);
drivers/media/pci/ttpci/budget-core.c
184
dma_sync_sg_for_cpu(&budget->dev->pci->dev, budget->pt.slist,
drivers/media/pci/ttpci/budget-core.c
185
budget->pt.nents, DMA_FROM_DEVICE);
drivers/media/pci/ttpci/budget-core.c
518
budget->grabbing = saa7146_vmalloc_build_pgtable(dev->pci, budget->buffer_size, &budget->pt);
drivers/media/pci/ttpci/budget-core.c
539
saa7146_vfree_destroy_pgtable(dev->pci, budget->grabbing, &budget->pt);
drivers/media/pci/ttpci/budget-core.c
570
saa7146_vfree_destroy_pgtable(dev->pci, budget->grabbing, &budget->pt);
drivers/media/pci/ttpci/budget.h
56
struct saa7146_pgtable pt;
drivers/mtd/maps/impa7.c
60
static struct { u_long addr; u_long size; } pt[NUM_FLASHBANKS] = {
drivers/mtd/maps/impa7.c
69
pt[i].size, pt[i].addr);
drivers/mtd/maps/impa7.c
71
impa7_map[i].phys = pt[i].addr;
drivers/mtd/maps/impa7.c
72
impa7_map[i].virt = ioremap(pt[i].addr, pt[i].size);
drivers/mtd/ubi/debug.h
25
#define ubi_dbg_print_hex_dump(l, ps, pt, r, g, b, len, a) \
drivers/mtd/ubi/debug.h
26
print_hex_dump(l, ps, pt, r, g, b, len, a)
drivers/net/ethernet/amd/xgbe/xgbe-selftest.c
122
tdata->pt.type = htons(ETH_P_IP);
drivers/net/ethernet/amd/xgbe/xgbe-selftest.c
123
tdata->pt.func = xgbe_test_loopback_validate;
drivers/net/ethernet/amd/xgbe/xgbe-selftest.c
124
tdata->pt.dev = pdata->netdev;
drivers/net/ethernet/amd/xgbe/xgbe-selftest.c
125
tdata->pt.af_packet_priv = tdata;
drivers/net/ethernet/amd/xgbe/xgbe-selftest.c
128
dev_add_pack(&tdata->pt);
drivers/net/ethernet/amd/xgbe/xgbe-selftest.c
150
dev_remove_pack(&tdata->pt);
drivers/net/ethernet/amd/xgbe/xgbe-selftest.c
34
struct packet_type *pt,
drivers/net/ethernet/amd/xgbe/xgbe-selftest.c
37
struct net_test_priv *tdata = pt->af_packet_priv;
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
790
static u64 ehea_reg_mr_section(int top, int dir, int idx, u64 *pt,
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
797
u64 pt_abs = __pa(pt);
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
805
pt[m] = __pa(pg);
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
821
static u64 ehea_reg_mr_sections(int top, int dir, u64 *pt,
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
832
hret = ehea_reg_mr_section(top, dir, idx, pt, adapter, mr);
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
839
static u64 ehea_reg_mr_dir_sections(int top, u64 *pt,
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
850
hret = ehea_reg_mr_sections(top, dir, pt, adapter, mr);
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
860
u64 *pt;
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
866
pt = (void *)get_zeroed_page(GFP_KERNEL);
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
867
if (!pt) {
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
894
hret = ehea_reg_mr_dir_sections(top, pt, adapter, mr);
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
910
free_page((unsigned long)pt);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
1820
u16 pt;
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
1822
for (pt = 0; pt < hw->blk[blk].xlt1.count; pt++) {
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
1825
ptg = hw->blk[blk].xlt1.t[pt];
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
1828
ice_ptg_add_mv_ptype(hw, blk, pt, ptg);
drivers/net/ethernet/intel/idpf/xdp.c
390
struct libeth_rx_pt pt;
drivers/net/ethernet/intel/idpf/xdp.c
396
pt = rxq->rx_ptype_lkup[idpf_xdp_rx_pt(&desc)];
drivers/net/ethernet/intel/idpf/xdp.c
397
if (!libeth_rx_pt_has_hash(rxq->xdp_rxq.dev, pt))
drivers/net/ethernet/intel/idpf/xdp.c
403
pt);
drivers/net/ethernet/intel/libeth/rx.c
261
void libeth_rx_pt_gen_hash_type(struct libeth_rx_pt *pt)
drivers/net/ethernet/intel/libeth/rx.c
263
pt->hash_type = 0;
drivers/net/ethernet/intel/libeth/rx.c
264
pt->hash_type |= libeth_rx_pt_xdp_oip[pt->outer_ip];
drivers/net/ethernet/intel/libeth/rx.c
265
pt->hash_type |= libeth_rx_pt_xdp_iprot[pt->inner_prot];
drivers/net/ethernet/intel/libeth/rx.c
266
pt->hash_type |= libeth_rx_pt_xdp_pl[pt->payload_layer];
drivers/net/ethernet/mellanox/mlx5/core/en_selftest.c
151
struct packet_type pt;
drivers/net/ethernet/mellanox/mlx5/core/en_selftest.c
160
struct packet_type *pt,
drivers/net/ethernet/mellanox/mlx5/core/en_selftest.c
163
struct mlx5e_lbt_priv *lbtp = pt->af_packet_priv;
drivers/net/ethernet/mellanox/mlx5/core/en_selftest.c
224
lbtp->pt.type = htons(ETH_P_IP);
drivers/net/ethernet/mellanox/mlx5/core/en_selftest.c
225
lbtp->pt.func = mlx5e_test_loopback_validate;
drivers/net/ethernet/mellanox/mlx5/core/en_selftest.c
226
lbtp->pt.dev = priv->netdev;
drivers/net/ethernet/mellanox/mlx5/core/en_selftest.c
227
lbtp->pt.af_packet_priv = lbtp;
drivers/net/ethernet/mellanox/mlx5/core/en_selftest.c
228
dev_add_pack(&lbtp->pt);
drivers/net/ethernet/mellanox/mlx5/core/en_selftest.c
245
dev_remove_pack(&lbtp->pt);
drivers/net/ethernet/mellanox/mlxsw/reg.h
890
MLXSW_ITEM32(reg, spvm, pt, 0x00, 31, 1);
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
1002
tpriv->pt.type = htons(ETH_P_8021Q);
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
1003
tpriv->pt.func = stmmac_test_vlan_validate;
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
1004
tpriv->pt.dev = priv->dev;
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
1005
tpriv->pt.af_packet_priv = tpriv;
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
1014
dev_add_pack(&tpriv->pt);
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
1054
dev_remove_pack(&tpriv->pt);
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
1279
tpriv->pt.type = svlan ? htons(ETH_P_8021Q) : htons(ETH_P_IP);
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
1280
tpriv->pt.func = stmmac_test_vlan_validate;
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
1281
tpriv->pt.dev = priv->dev;
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
1282
tpriv->pt.af_packet_priv = tpriv;
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
1285
dev_add_pack(&tpriv->pt);
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
1312
dev_remove_pack(&tpriv->pt);
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
1595
struct packet_type *pt,
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
1598
struct stmmac_test_priv *tpriv = pt->af_packet_priv;
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
1638
tpriv->pt.type = htons(ETH_P_ARP);
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
1639
tpriv->pt.func = stmmac_test_arp_validate;
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
1640
tpriv->pt.dev = priv->dev;
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
1641
tpriv->pt.af_packet_priv = tpriv;
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
1643
dev_add_pack(&tpriv->pt);
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
1679
dev_remove_pack(&tpriv->pt);
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
239
struct packet_type pt;
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
248
struct packet_type *pt,
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
251
struct stmmac_test_priv *tpriv = pt->af_packet_priv;
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
334
tpriv->pt.type = htons(ETH_P_IP);
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
335
tpriv->pt.func = stmmac_test_loopback_validate;
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
336
tpriv->pt.dev = priv->dev;
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
337
tpriv->pt.af_packet_priv = tpriv;
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
341
dev_add_pack(&tpriv->pt);
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
364
dev_remove_pack(&tpriv->pt);
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
716
struct packet_type *pt,
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
719
struct stmmac_test_priv *tpriv = pt->af_packet_priv;
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
753
tpriv->pt.type = htons(ETH_P_PAUSE);
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
754
tpriv->pt.func = stmmac_test_flowctrl_validate;
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
755
tpriv->pt.dev = priv->dev;
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
756
tpriv->pt.af_packet_priv = tpriv;
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
757
dev_add_pack(&tpriv->pt);
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
815
dev_remove_pack(&tpriv->pt);
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
837
struct packet_type *pt,
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
840
struct stmmac_test_priv *tpriv = pt->af_packet_priv;
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
908
tpriv->pt.type = htons(ETH_P_IP);
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
909
tpriv->pt.func = stmmac_test_vlan_validate;
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
910
tpriv->pt.dev = priv->dev;
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
911
tpriv->pt.af_packet_priv = tpriv;
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
920
dev_add_pack(&tpriv->pt);
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
960
dev_remove_pack(&tpriv->pt);
drivers/net/fddi/skfp/pmf.c
1072
const struct s_p_tab *pt ;
drivers/net/fddi/skfp/pmf.c
1150
pt = smt_get_ptab(pa->p_type) ;
drivers/net/fddi/skfp/pmf.c
1151
if (!pt)
drivers/net/fddi/skfp/pmf.c
1154
switch (pt->p_access) {
drivers/net/fddi/skfp/pmf.c
1161
to = mib_addr + pt->p_offset ;
drivers/net/fddi/skfp/pmf.c
1162
swap = pt->p_swap ; /* pointer to swap string */
drivers/net/fddi/skfp/pmf.c
1517
const struct s_p_tab *pt ;
drivers/net/fddi/skfp/pmf.c
1518
for (pt = p_tab ; pt->p_num && pt->p_num != para ; pt++)
drivers/net/fddi/skfp/pmf.c
1520
return pt->p_num ? pt : NULL;
drivers/net/fddi/skfp/pmf.c
310
const struct s_p_tab *pt ;
drivers/net/fddi/skfp/pmf.c
423
pt = smt_get_ptab(pa->p_type) ;
drivers/net/fddi/skfp/pmf.c
424
if (pt && pt->p_access == AC_GROUP && !set) {
drivers/net/fddi/skfp/pmf.c
425
pt++ ;
drivers/net/fddi/skfp/pmf.c
426
while (pt->p_access == AC_G ||
drivers/net/fddi/skfp/pmf.c
427
pt->p_access == AC_GR) {
drivers/net/fddi/skfp/pmf.c
428
smt_add_para(smc,&pcon,pt->p_num,
drivers/net/fddi/skfp/pmf.c
430
pt++ ;
drivers/net/fddi/skfp/pmf.c
455
if (pt && pt->p_access == AC_S) {
drivers/net/fddi/skfp/pmf.c
546
const struct s_p_tab *pt ;
drivers/net/fddi/skfp/pmf.c
571
pt = smt_get_ptab(para) ;
drivers/net/fddi/skfp/pmf.c
572
if (pt && pt->p_access == AC_S)
drivers/net/fddi/skfp/pmf.c
868
if (!pt) {
drivers/net/fddi/skfp/pmf.c
876
switch (pt->p_access) {
drivers/net/fddi/skfp/pmf.c
884
from = mib_addr + pt->p_offset ;
drivers/net/fddi/skfp/pmf.c
886
swap = pt->p_swap ; /* pointer to swap string */
drivers/net/ppp/pppoe.c
410
struct packet_type *pt, struct net_device *orig_dev)
drivers/net/ppp/pppoe.c
478
struct packet_type *pt, struct net_device *orig_dev)
drivers/phy/socionext/phy-uniphier-usb3hs.c
123
struct uniphier_u3hsphy_trim_param *pt)
drivers/phy/socionext/phy-uniphier-usb3hs.c
127
ret = uniphier_u3hsphy_get_nvparam(priv, "rterm", &pt->rterm);
drivers/phy/socionext/phy-uniphier-usb3hs.c
131
ret = uniphier_u3hsphy_get_nvparam(priv, "sel_t", &pt->sel_t);
drivers/phy/socionext/phy-uniphier-usb3hs.c
135
ret = uniphier_u3hsphy_get_nvparam(priv, "hs_i", &pt->hs_i);
drivers/phy/socionext/phy-uniphier-usb3hs.c
84
struct uniphier_u3hsphy_trim_param *pt);
drivers/phy/socionext/phy-uniphier-usb3hs.c
89
struct uniphier_u3hsphy_trim_param *pt)
drivers/phy/socionext/phy-uniphier-usb3hs.c
92
*pconfig |= FIELD_PREP(HSPHY_CFG0_RTERM_MASK, pt->rterm);
drivers/phy/socionext/phy-uniphier-usb3hs.c
95
*pconfig |= FIELD_PREP(HSPHY_CFG0_SEL_T_MASK, pt->sel_t);
drivers/phy/socionext/phy-uniphier-usb3hs.c
98
*pconfig |= FIELD_PREP(HSPHY_CFG0_HS_I_MASK, pt->hs_i);
drivers/platform/surface/surface_aggregator_cdev.c
641
static __poll_t ssam_cdev_poll(struct file *file, struct poll_table_struct *pt)
drivers/platform/surface/surface_aggregator_cdev.c
649
poll_wait(file, &client->waitq, pt);
drivers/platform/surface/surface_dtx.c
526
static __poll_t surface_dtx_poll(struct file *file, struct poll_table_struct *pt)
drivers/platform/surface/surface_dtx.c
534
poll_wait(file, &client->ddev->waitq, pt);
drivers/remoteproc/imx_rproc.c
1042
u8 pt;
drivers/remoteproc/imx_rproc.c
1078
ret = imx_sc_rm_get_resource_owner(priv->ipc_handle, priv->rsrc_id, &pt);
drivers/remoteproc/imx_rproc.c
1084
priv->rproc_pt = pt;
drivers/s390/block/dasd_devmap.c
2193
char *primary, *secondary, *tmp, *pt;
drivers/s390/block/dasd_devmap.c
2196
pt = kstrdup(buf, GFP_KERNEL);
drivers/s390/block/dasd_devmap.c
2197
tmp = pt;
drivers/s390/block/dasd_devmap.c
2203
kfree(pt);
drivers/s390/block/dasd_devmap.c
2208
kfree(pt);
drivers/s390/block/dasd_devmap.c
2212
kfree(pt);
drivers/s390/block/dasd_devmap.c
2217
kfree(pt);
drivers/s390/block/dasd_devmap.c
2221
kfree(pt);
drivers/scsi/ips.c
1025
ips_passthru_t *pt;
drivers/scsi/ips.c
1073
pt = (ips_passthru_t *) scsi_sglist(SC);
drivers/scsi/ips.c
1074
if ((pt->CoppCP.cmd.reset.op_code == IPS_CMD_RESET_CHANNEL) &&
drivers/scsi/ips.c
1075
(pt->CoppCP.cmd.reset.adapter_flag == 1)) {
drivers/scsi/ips.c
1545
ips_passthru_t *pt;
drivers/scsi/ips.c
1565
pt = (ips_passthru_t *) ha->ioctl_data;
drivers/scsi/ips.c
1566
ips_scmd_buf_read(SC, pt, sizeof (ips_passthru_t));
drivers/scsi/ips.c
1567
pt->BasicStatus = 0x0B;
drivers/scsi/ips.c
1568
pt->ExtendedStatus = 0x00;
drivers/scsi/ips.c
1569
ips_scmd_buf_write(SC, pt, sizeof (ips_passthru_t));
drivers/scsi/ips.c
1576
pt = (ips_passthru_t *) ha->ioctl_data;
drivers/scsi/ips.c
1588
switch (pt->CoppCmd) {
drivers/scsi/ips.c
1601
if (length < (sizeof (ips_passthru_t) + pt->CmdBSize)) {
drivers/scsi/ips.c
1611
pt->CoppCP.cmd.flashfw.op_code ==
drivers/scsi/ips.c
1613
ret = ips_flash_copperhead(ha, pt, scb);
drivers/scsi/ips.c
1618
if (ips_usrcmd(ha, pt, scb))
drivers/scsi/ips.c
1637
ips_flash_copperhead(ips_ha_t * ha, ips_passthru_t * pt, ips_scb_t * scb)
drivers/scsi/ips.c
1643
if (IPS_IS_TROMBONE(ha) && pt->CoppCP.cmd.flashfw.type == IPS_FW_IMAGE) {
drivers/scsi/ips.c
1644
if (ips_usrcmd(ha, pt, scb))
drivers/scsi/ips.c
1649
pt->BasicStatus = 0x0B;
drivers/scsi/ips.c
1650
pt->ExtendedStatus = 0;
drivers/scsi/ips.c
1654
if (pt->CoppCP.cmd.flashfw.type == IPS_BIOS_IMAGE &&
drivers/scsi/ips.c
1655
pt->CoppCP.cmd.flashfw.direction == IPS_ERASE_BIOS) {
drivers/scsi/ips.c
1656
pt->BasicStatus = 0;
drivers/scsi/ips.c
1657
return ips_flash_bios(ha, pt, scb);
drivers/scsi/ips.c
1658
} else if (pt->CoppCP.cmd.flashfw.packet_num == 0) {
drivers/scsi/ips.c
1665
datasize = pt->CoppCP.cmd.flashfw.total_packets *
drivers/scsi/ips.c
1666
pt->CoppCP.cmd.flashfw.count;
drivers/scsi/ips.c
1678
if (pt->CoppCP.cmd.flashfw.count + ha->flash_datasize >
drivers/scsi/ips.c
1688
pt->BasicStatus = 0;
drivers/scsi/ips.c
1689
memcpy(&ha->flash_data[ha->flash_datasize], pt + 1,
drivers/scsi/ips.c
1690
pt->CoppCP.cmd.flashfw.count);
drivers/scsi/ips.c
1691
ha->flash_datasize += pt->CoppCP.cmd.flashfw.count;
drivers/scsi/ips.c
1692
if (pt->CoppCP.cmd.flashfw.packet_num ==
drivers/scsi/ips.c
1693
pt->CoppCP.cmd.flashfw.total_packets - 1) {
drivers/scsi/ips.c
1694
if (pt->CoppCP.cmd.flashfw.type == IPS_BIOS_IMAGE)
drivers/scsi/ips.c
1695
return ips_flash_bios(ha, pt, scb);
drivers/scsi/ips.c
1696
else if (pt->CoppCP.cmd.flashfw.type == IPS_FW_IMAGE)
drivers/scsi/ips.c
1697
return ips_flash_firmware(ha, pt, scb);
drivers/scsi/ips.c
1708
ips_flash_bios(ips_ha_t * ha, ips_passthru_t * pt, ips_scb_t * scb)
drivers/scsi/ips.c
1711
if (pt->CoppCP.cmd.flashfw.type == IPS_BIOS_IMAGE &&
drivers/scsi/ips.c
1712
pt->CoppCP.cmd.flashfw.direction == IPS_WRITE_BIOS) {
drivers/scsi/ips.c
1744
} else if (pt->CoppCP.cmd.flashfw.type == IPS_BIOS_IMAGE &&
drivers/scsi/ips.c
1745
pt->CoppCP.cmd.flashfw.direction == IPS_ERASE_BIOS) {
drivers/scsi/ips.c
1757
pt->BasicStatus = 0x0B;
drivers/scsi/ips.c
1758
pt->ExtendedStatus = 0x00;
drivers/scsi/ips.c
1810
ips_flash_firmware(ips_ha_t * ha, ips_passthru_t * pt, ips_scb_t * scb)
drivers/scsi/ips.c
1815
if (pt->CoppCP.cmd.flashfw.type == IPS_FW_IMAGE &&
drivers/scsi/ips.c
1816
pt->CoppCP.cmd.flashfw.direction == IPS_WRITE_FW) {
drivers/scsi/ips.c
1817
memset(&pt->CoppCP.cmd, 0, sizeof (IPS_HOST_COMMAND));
drivers/scsi/ips.c
1818
pt->CoppCP.cmd.flashfw.op_code = IPS_CMD_DOWNLOAD;
drivers/scsi/ips.c
1819
pt->CoppCP.cmd.flashfw.count = cpu_to_le32(ha->flash_datasize);
drivers/scsi/ips.c
1821
pt->BasicStatus = 0x0B;
drivers/scsi/ips.c
1822
pt->ExtendedStatus = 0x00;
drivers/scsi/ips.c
1830
memcpy(&scb->cmd, &pt->CoppCP.cmd, sizeof (IPS_IOCTL_CMD));
drivers/scsi/ips.c
1851
if (pt->TimeOut)
drivers/scsi/ips.c
1852
scb->timeout = pt->TimeOut;
drivers/scsi/ips.c
1883
ips_usrcmd(ips_ha_t * ha, ips_passthru_t * pt, ips_scb_t * scb)
drivers/scsi/ips.c
1890
if ((!scb) || (!pt) || (!ha))
drivers/scsi/ips.c
1897
memcpy(&scb->cmd, &pt->CoppCP.cmd, sizeof (IPS_IOCTL_CMD));
drivers/scsi/ips.c
1898
memcpy(&scb->dcdb, &pt->CoppCP.dcdb, sizeof (IPS_DCDB_TABLE));
drivers/scsi/ips.c
1920
if (pt->CmdBSize) {
drivers/scsi/ips.c
1921
scb->data_len = pt->CmdBSize;
drivers/scsi/ips.c
1933
if (pt->CmdBSize) {
drivers/scsi/ips.c
1943
if (pt->TimeOut) {
drivers/scsi/ips.c
1944
scb->timeout = pt->TimeOut;
drivers/scsi/ips.c
1946
if (pt->TimeOut <= 10)
drivers/scsi/ips.c
1948
else if (pt->TimeOut <= 60)
drivers/scsi/ips.c
1973
ips_passthru_t *pt;
drivers/scsi/ips.c
1983
pt = (ips_passthru_t *) ha->ioctl_data;
drivers/scsi/ips.c
1987
memcpy(&pt->CoppCP.dcdb, &scb->dcdb, sizeof (IPS_DCDB_TABLE));
drivers/scsi/ips.c
1989
pt->BasicStatus = scb->basic_status;
drivers/scsi/ips.c
1990
pt->ExtendedStatus = scb->extended_status;
drivers/scsi/ips.c
1991
pt->AdapterType = ha->ad_type;
drivers/scsi/lpfc/lpfc_attr.c
3782
static DEVICE_ATTR(pt, 0444,
drivers/scsi/lpfc/lpfc_init.c
9818
u8 ptv, tf, pt;
drivers/scsi/lpfc/lpfc_init.c
9822
pt = bf_get(lpfc_mbx_rd_conf_pt, rd_config);
drivers/scsi/lpfc/lpfc_init.c
9826
ptv, tf, pt);
drivers/scsi/lpfc/lpfc_init.c
9843
phba->cfg_topology = ((pt == LINK_FLAGS_LOOP)
drivers/scsi/lpfc/lpfc_init.c
9851
phba->cfg_topology = (pt ? FLAGS_TOPOLOGY_MODE_PT_LOOP :
drivers/scsi/lpfc/lpfc_init.c
9854
phba->cfg_topology = ((pt == LINK_FLAGS_P2P)
drivers/soc/aspeed/aspeed-lpc-snoop.c
136
struct poll_table_struct *pt)
drivers/soc/aspeed/aspeed-lpc-snoop.c
140
poll_wait(file, &chan->wq, pt);
drivers/staging/media/atomisp/pci/mmu/isp_mmu.c
49
static unsigned int atomisp_get_pte(phys_addr_t pt, unsigned int idx)
drivers/staging/media/atomisp/pci/mmu/isp_mmu.c
51
unsigned int *pt_virt = phys_to_virt(pt);
drivers/staging/media/atomisp/pci/mmu/isp_mmu.c
56
static void atomisp_set_pte(phys_addr_t pt,
drivers/staging/media/atomisp/pci/mmu/isp_mmu.c
59
unsigned int *pt_virt = phys_to_virt(pt);
drivers/staging/media/atomisp/pci/sh_css_internal.h
434
#define SH_CSS_PORT_TYPE_BIT_FLD(pt) (0x1 << (pt))
drivers/staging/media/atomisp/pci/sh_css_internal.h
436
#define SH_CSS_PIPE_PORT_CONFIG_ON(p, pd, pt) ((p) |= (SH_CSS_PORT_TYPE_BIT_FLD(pt) << SH_CSS_PORT_FLD(pd)))
drivers/staging/media/atomisp/pci/sh_css_internal.h
437
#define SH_CSS_PIPE_PORT_CONFIG_OFF(p, pd, pt) ((p) &= ~(SH_CSS_PORT_TYPE_BIT_FLD(pt) << SH_CSS_PORT_FLD(pd)))
drivers/staging/media/atomisp/pci/sh_css_internal.h
438
#define SH_CSS_PIPE_PORT_CONFIG_SET(p, pd, pt, val) ((val) ? \
drivers/staging/media/atomisp/pci/sh_css_internal.h
439
SH_CSS_PIPE_PORT_CONFIG_ON(p, pd, pt) : SH_CSS_PIPE_PORT_CONFIG_OFF(p, pd, pt))
drivers/staging/media/atomisp/pci/sh_css_internal.h
440
#define SH_CSS_PIPE_PORT_CONFIG_GET(p, pd, pt) ((p) & (SH_CSS_PORT_TYPE_BIT_FLD(pt) << SH_CSS_PORT_FLD(pd)))
drivers/staging/media/av7110/av7110.c
1241
dma_sync_sg_for_cpu(&budget->dev->pci->dev, budget->pt.slist,
drivers/staging/media/av7110/av7110.c
1242
budget->pt.nents, DMA_FROM_DEVICE);
drivers/staging/media/av7110/av7110.c
2480
&av7110->pt);
drivers/staging/media/av7110/av7110.c
2499
saa7146_write(dev, BASE_PAGE3, av7110->pt.dma | ME1 | 0x90);
drivers/staging/media/av7110/av7110.c
2508
&av7110->pt);
drivers/staging/media/av7110/av7110.c
2522
saa7146_write(dev, BASE_PAGE3, av7110->pt.dma | ME1 | 0x90);
drivers/staging/media/av7110/av7110.c
2715
saa7146_vfree_destroy_pgtable(pdev, av7110->grabbing, &av7110->pt);
drivers/staging/media/av7110/av7110.c
2748
saa7146_vfree_destroy_pgtable(saa->pci, av7110->grabbing, &av7110->pt);
drivers/staging/media/av7110/av7110.h
185
struct saa7146_pgtable pt;
drivers/staging/media/ipu3/ipu3-mmu.c
123
u32 *pt;
drivers/staging/media/ipu3/ipu3-mmu.c
126
pt = (u32 *)__get_free_page(GFP_KERNEL);
drivers/staging/media/ipu3/ipu3-mmu.c
127
if (!pt)
drivers/staging/media/ipu3/ipu3-mmu.c
131
pt[pte] = pteval;
drivers/staging/media/ipu3/ipu3-mmu.c
133
set_memory_uc((unsigned long)pt, IPU3_PT_ORDER);
drivers/staging/media/ipu3/ipu3-mmu.c
135
return pt;
drivers/staging/media/ipu3/ipu3-mmu.c
142
static void imgu_mmu_free_page_table(u32 *pt)
drivers/staging/media/ipu3/ipu3-mmu.c
144
set_memory_wb((unsigned long)pt, IPU3_PT_ORDER);
drivers/staging/media/ipu3/ipu3-mmu.c
145
free_page((unsigned long)pt);
drivers/staging/media/ipu7/ipu7-mmu.c
124
void *pt = (void *)get_zeroed_page(GFP_ATOMIC | GFP_DMA32);
drivers/staging/media/ipu7/ipu7-mmu.c
127
if (!pt)
drivers/staging/media/ipu7/ipu7-mmu.c
130
dev_dbg(mmu_info->dev, "dummy_page: get_zeroed_page() == %p\n", pt);
drivers/staging/media/ipu7/ipu7-mmu.c
132
dma = map_single(mmu_info, pt);
drivers/staging/media/ipu7/ipu7-mmu.c
138
mmu_info->dummy_page = pt;
drivers/staging/media/ipu7/ipu7-mmu.c
144
free_page((unsigned long)pt);
drivers/staging/media/ipu7/ipu7-mmu.c
158
u32 *pt = (u32 *)get_zeroed_page(GFP_ATOMIC | GFP_DMA32);
drivers/staging/media/ipu7/ipu7-mmu.c
162
if (!pt)
drivers/staging/media/ipu7/ipu7-mmu.c
165
dev_dbg(mmu_info->dev, "dummy_l2: get_zeroed_page() = %p\n", pt);
drivers/staging/media/ipu7/ipu7-mmu.c
167
dma = map_single(mmu_info, pt);
drivers/staging/media/ipu7/ipu7-mmu.c
174
pt[i] = mmu_info->dummy_page_pteval;
drivers/staging/media/ipu7/ipu7-mmu.c
176
mmu_info->dummy_l2_pt = pt;
drivers/staging/media/ipu7/ipu7-mmu.c
182
free_page((unsigned long)pt);
drivers/staging/media/ipu7/ipu7-mmu.c
196
u32 *pt = (u32 *)get_zeroed_page(GFP_ATOMIC | GFP_DMA32);
drivers/staging/media/ipu7/ipu7-mmu.c
200
if (!pt)
drivers/staging/media/ipu7/ipu7-mmu.c
203
dev_dbg(mmu_info->dev, "alloc_l1: get_zeroed_page() = %p\n", pt);
drivers/staging/media/ipu7/ipu7-mmu.c
206
pt[i] = mmu_info->dummy_l2_pteval;
drivers/staging/media/ipu7/ipu7-mmu.c
208
dma = map_single(mmu_info, pt);
drivers/staging/media/ipu7/ipu7-mmu.c
215
dev_dbg(mmu_info->dev, "l1 pt %p mapped at %pad\n", pt, &dma);
drivers/staging/media/ipu7/ipu7-mmu.c
217
return pt;
drivers/staging/media/ipu7/ipu7-mmu.c
220
free_page((unsigned long)pt);
drivers/staging/media/ipu7/ipu7-mmu.c
226
u32 *pt = (u32 *)get_zeroed_page(GFP_ATOMIC | GFP_DMA32);
drivers/staging/media/ipu7/ipu7-mmu.c
229
if (!pt)
drivers/staging/media/ipu7/ipu7-mmu.c
232
dev_dbg(mmu_info->dev, "alloc_l2: get_zeroed_page() = %p\n", pt);
drivers/staging/media/ipu7/ipu7-mmu.c
235
pt[i] = mmu_info->dummy_page_pteval;
drivers/staging/media/ipu7/ipu7-mmu.c
237
return pt;
drivers/tty/serial/8250/8250.h
339
static inline int is_omap1_8250(struct uart_8250_port *pt)
drivers/tty/serial/8250/8250.h
343
switch (pt->port.mapbase) {
drivers/tty/serial/8250/8250.h
357
static inline int is_omap1510_8250(struct uart_8250_port *pt)
drivers/tty/serial/8250/8250.h
362
return is_omap1_8250(pt);
drivers/tty/serial/8250/8250.h
365
static inline int is_omap1_8250(struct uart_8250_port *pt)
drivers/tty/serial/8250/8250.h
369
static inline int is_omap1510_8250(struct uart_8250_port *pt)
drivers/tty/serial/8250/8250_port.c
2848
static unsigned int serial8250_port_size(struct uart_8250_port *pt)
drivers/tty/serial/8250/8250_port.c
2850
if (pt->port.mapsize)
drivers/tty/serial/8250/8250_port.c
2851
return pt->port.mapsize;
drivers/tty/serial/8250/8250_port.c
2852
if (is_omap1_8250(pt))
drivers/tty/serial/8250/8250_port.c
2853
return 0x16 << pt->port.regshift;
drivers/tty/serial/8250/8250_port.c
2855
return 8 << pt->port.regshift;
drivers/vfio/virqfd.c
171
init_poll_funcptr(&virqfd->pt, virqfd_ptable_queue_proc);
drivers/vfio/virqfd.c
173
events = vfs_poll(fd_file(irqfd), &virqfd->pt);
drivers/vfio/virqfd.c
79
wait_queue_head_t *wqh, poll_table *pt)
drivers/vfio/virqfd.c
81
struct virqfd *virqfd = container_of(pt, struct virqfd, pt);
drivers/vhost/vhost.c
163
poll_table *pt)
drivers/vhost/vhost.c
167
poll = container_of(pt, struct vhost_poll, table);
drivers/virt/acrn/irqfd.c
143
init_poll_funcptr(&irqfd->pt, hsm_irqfd_poll_func);
drivers/virt/acrn/irqfd.c
157
events = vfs_poll(fd_file(f), &irqfd->pt);
drivers/virt/acrn/irqfd.c
35
poll_table pt;
drivers/virt/acrn/irqfd.c
95
poll_table *pt)
drivers/virt/acrn/irqfd.c
99
irqfd = container_of(pt, struct hsm_irqfd, pt);
drivers/xen/privcmd.c
1017
init_poll_funcptr(&kirqfd->pt, irqfd_poll_func);
drivers/xen/privcmd.c
1037
events = vfs_poll(fd_file(f), &kirqfd->pt);
drivers/xen/privcmd.c
901
poll_table pt;
drivers/xen/privcmd.c
968
irqfd_poll_func(struct file *file, wait_queue_head_t *wqh, poll_table *pt)
drivers/xen/privcmd.c
971
container_of(pt, struct privcmd_kernel_irqfd, pt);
fs/aio.c
1727
struct poll_table_struct pt = { ._key = req->events };
fs/aio.c
1732
mask = vfs_poll(req->file, &pt) & req->events;
fs/aio.c
1869
struct poll_table_struct pt;
fs/aio.c
1879
struct aio_poll_table *pt = container_of(p, struct aio_poll_table, pt);
fs/aio.c
1882
if (unlikely(pt->queued)) {
fs/aio.c
1883
pt->error = -EINVAL;
fs/aio.c
1887
pt->queued = true;
fs/aio.c
1888
pt->error = 0;
fs/aio.c
1889
pt->iocb->poll.head = head;
fs/aio.c
1890
add_wait_queue(head, &pt->iocb->poll.wait);
fs/aio.c
1916
apt.pt._qproc = aio_poll_queue_proc;
fs/aio.c
1917
apt.pt._key = req->events;
fs/aio.c
1926
mask = vfs_poll(req->file, &apt.pt) & req->events;
fs/btrfs/send.c
1206
struct fs_path *pt = ctx;
fs/btrfs/send.c
1208
ret = fs_path_copy(pt, p);
fs/ecryptfs/miscdev.c
29
ecryptfs_miscdev_poll(struct file *file, poll_table *pt)
fs/ecryptfs/miscdev.c
46
poll_wait(file, &daemon->wait, pt);
fs/eventpoll.c
1044
static __poll_t ep_item_poll(const struct epitem *epi, poll_table *pt,
fs/eventpoll.c
1057
pt->_key = epi->event.events;
fs/eventpoll.c
1059
res = vfs_poll(file, pt);
fs/eventpoll.c
1061
res = __ep_eventpoll_poll(file, pt, depth);
fs/eventpoll.c
1359
poll_table *pt)
fs/eventpoll.c
1361
struct ep_pqueue *epq = container_of(pt, struct ep_pqueue, pt);
fs/eventpoll.c
1639
init_poll_funcptr(&epq.pt, ep_ptable_queue_proc);
fs/eventpoll.c
1648
revents = ep_item_poll(epi, &epq.pt, 1);
fs/eventpoll.c
1695
poll_table pt;
fs/eventpoll.c
1699
init_poll_funcptr(&pt, NULL);
fs/eventpoll.c
1741
if (ep_item_poll(epi, &pt, 1)) {
fs/eventpoll.c
1768
poll_table pt;
fs/eventpoll.c
1779
init_poll_funcptr(&pt, NULL);
fs/eventpoll.c
1818
revents = ep_item_poll(epi, &pt, 1);
fs/eventpoll.c
247
poll_table pt;
fs/eventpoll.c
971
static __poll_t ep_item_poll(const struct epitem *epi, poll_table *pt, int depth);
fs/eventpoll.c
978
poll_table pt;
fs/eventpoll.c
981
init_poll_funcptr(&pt, NULL);
fs/eventpoll.c
993
if (ep_item_poll(epi, &pt, depth + 1)) {
fs/select.c
121
init_poll_funcptr(&pwq->pt, __pollwait);
fs/select.c
222
struct poll_wqueues *pwq = container_of(p, struct poll_wqueues, pt);
fs/select.c
502
wait = &table.pt;
fs/select.c
875
poll_table* pt = &wait->pt;
fs/select.c
884
pt->_qproc = NULL;
fs/select.c
909
mask = do_pollfd(pfd, pt, &can_busy_loop, busy_flag);
fs/select.c
913
pt->_qproc = NULL;
fs/select.c
924
pt->_qproc = NULL;
include/asm-generic/tlb.h
521
static inline void tlb_remove_ptdesc(struct mmu_gather *tlb, struct ptdesc *pt)
include/asm-generic/tlb.h
523
tlb_remove_table(tlb, pt);
include/asm-generic/tlb.h
792
static inline void tlb_unshare_pmd_ptdesc(struct mmu_gather *tlb, struct ptdesc *pt,
include/asm-generic/tlb.h
799
VM_WARN_ON_ONCE(!ptdesc_pmd_is_shared(pt));
include/asm-generic/tlb.h
800
ptdesc_pmd_pts_dec(pt);
include/asm-generic/tlb.h
809
if (!ptdesc_pmd_is_shared(pt))
include/linux/bpf.h
3027
const char *btf_find_decl_tag_value(const struct btf *btf, const struct btf_type *pt,
include/linux/bpf.h
3029
int btf_find_next_decl_tag(const struct btf *btf, const struct btf_type *pt,
include/linux/cgroup-defs.h
763
struct poll_table_struct *pt);
include/linux/firmware/imx/svc/rm.h
62
int imx_sc_rm_get_resource_owner(struct imx_sc_ipc *ipc, u16 resource, u8 *pt);
include/linux/firmware/imx/svc/rm.h
69
static inline int imx_sc_rm_get_resource_owner(struct imx_sc_ipc *ipc, u16 resource, u8 *pt)
include/linux/if_hsr.h
45
enum hsr_port_type pt);
include/linux/if_hsr.h
60
enum hsr_port_type pt)
include/linux/kernfs.h
323
struct poll_table_struct *pt);
include/linux/kernfs.h
454
struct poll_table_struct *pt);
include/linux/kernfs.h
565
struct poll_table_struct *pt)
include/linux/mm.h
3342
static inline void *ptdesc_address(const struct ptdesc *pt)
include/linux/mm.h
3344
return folio_address(ptdesc_folio(pt));
include/linux/mm.h
3347
static inline bool pagetable_is_reserved(struct ptdesc *pt)
include/linux/mm.h
3349
return test_bit(PT_reserved, &pt->pt_flags.f);
include/linux/mm.h
3410
static inline void __pagetable_free(struct ptdesc *pt)
include/linux/mm.h
3412
struct page *page = ptdesc_page(pt);
include/linux/mm.h
3418
void pagetable_free_kernel(struct ptdesc *pt);
include/linux/mm.h
3420
static inline void pagetable_free_kernel(struct ptdesc *pt)
include/linux/mm.h
3422
__pagetable_free(pt);
include/linux/mm.h
3432
static inline void pagetable_free(struct ptdesc *pt)
include/linux/mm.h
3434
if (ptdesc_test_kernel(pt)) {
include/linux/mm.h
3435
ptdesc_clear_kernel(pt);
include/linux/mm.h
3436
pagetable_free_kernel(pt);
include/linux/mm.h
3438
__pagetable_free(pt);
include/linux/mm.h
3701
static inline void free_reserved_ptdesc(struct ptdesc *pt)
include/linux/mm.h
3703
free_reserved_page(ptdesc_page(pt));
include/linux/mm_types.h
609
#define TABLE_MATCH(pg, pt) \
include/linux/mm_types.h
610
static_assert(offsetof(struct page, pg) == offsetof(struct ptdesc, pt))
include/linux/mm_types.h
625
#define ptdesc_page(pt) (_Generic((pt), \
include/linux/mm_types.h
626
const struct ptdesc *: (const struct page *)(pt), \
include/linux/mm_types.h
627
struct ptdesc *: (struct page *)(pt)))
include/linux/mm_types.h
629
#define ptdesc_folio(pt) (_Generic((pt), \
include/linux/mm_types.h
630
const struct ptdesc *: (const struct folio *)(pt), \
include/linux/mm_types.h
631
struct ptdesc *: (struct folio *)(pt)))
include/linux/net/intel/libie/rx.h
42
static inline struct libeth_rx_pt libie_rx_pt_parse(u32 pt)
include/linux/net/intel/libie/rx.h
44
if (unlikely(pt >= LIBIE_RX_PT_NUM))
include/linux/net/intel/libie/rx.h
45
pt = 0;
include/linux/net/intel/libie/rx.h
47
return libie_rx_pt_lut[pt];
include/linux/netdevice.h
3354
void dev_add_pack(struct packet_type *pt);
include/linux/netdevice.h
3355
void dev_remove_pack(struct packet_type *pt);
include/linux/netdevice.h
3356
void __dev_remove_pack(struct packet_type *pt);
include/linux/poll.h
67
static inline void init_poll_funcptr(poll_table *pt, poll_queue_proc qproc)
include/linux/poll.h
69
pt->_qproc = qproc;
include/linux/poll.h
70
pt->_key = ~(__poll_t)0; /* all events enabled */
include/linux/poll.h
78
static inline __poll_t vfs_poll(struct file *file, struct poll_table_struct *pt)
include/linux/poll.h
82
return file->f_op->poll(file, pt);
include/linux/poll.h
96
poll_table pt;
include/linux/vfio.h
382
poll_table pt;
include/media/drv-intf/saa7146.h
171
int saa7146_pgtable_alloc(struct pci_dev *pci, struct saa7146_pgtable *pt);
include/media/drv-intf/saa7146.h
172
void saa7146_pgtable_free(struct pci_dev *pci, struct saa7146_pgtable *pt);
include/media/drv-intf/saa7146.h
173
int saa7146_pgtable_build_single(struct pci_dev *pci, struct saa7146_pgtable *pt, struct scatterlist *list, int length );
include/media/drv-intf/saa7146.h
174
void *saa7146_vmalloc_build_pgtable(struct pci_dev *pci, long length, struct saa7146_pgtable *pt);
include/media/drv-intf/saa7146.h
175
void saa7146_vfree_destroy_pgtable(struct pci_dev *pci, void *mem, struct saa7146_pgtable *pt);
include/media/drv-intf/saa7146_vv.h
69
struct saa7146_pgtable pt[3];
include/net/ip.h
164
int ip_rcv(struct sk_buff *skb, struct net_device *dev, struct packet_type *pt,
include/net/ip.h
166
void ip_list_rcv(struct list_head *head, struct packet_type *pt,
include/net/ipv6.h
1003
struct packet_type *pt, struct net_device *orig_dev);
include/net/ipv6.h
1004
void ipv6_list_rcv(struct list_head *head, struct packet_type *pt,
include/net/libeth/rx.h
254
void libeth_rx_pt_gen_hash_type(struct libeth_rx_pt *pt);
include/net/libeth/rx.h
265
static inline u32 libeth_rx_pt_get_ip_ver(struct libeth_rx_pt pt)
include/net/libeth/rx.h
268
switch (pt.outer_ip) {
include/net/libeth/rx.h
275
return pt.outer_ip;
include/net/libeth/rx.h
286
struct libeth_rx_pt pt)
include/net/libeth/rx.h
291
return likely(pt.inner_prot > LIBETH_RX_PT_INNER_NONE &&
include/net/libeth/rx.h
296
struct libeth_rx_pt pt)
include/net/libeth/rx.h
298
return likely(pt.payload_layer > LIBETH_RX_PT_PAYLOAD_NONE &&
include/net/libeth/rx.h
309
struct libeth_rx_pt pt)
include/net/libeth/rx.h
311
skb_set_hash(skb, hash, pt.payload_layer);
include/net/libeth/xdp.h
1731
u32 val, struct libeth_rx_pt pt)
include/net/libeth/xdp.h
1737
*rss_type = pt.hash_type;
include/net/llc.h
100
int llc_rcv(struct sk_buff *skb, struct net_device *dev, struct packet_type *pt,
include/net/llc.h
115
struct packet_type *pt,
include/net/llc.h
61
struct packet_type *pt,
include/net/netlink.h
2444
void nla_get_range_unsigned(const struct nla_policy *pt,
include/net/netlink.h
2446
void nla_get_range_signed(const struct nla_policy *pt,
include/net/netlink.h
2460
int netlink_policy_dump_attr_size_estimate(const struct nla_policy *pt);
include/net/netlink.h
2462
const struct nla_policy *pt,
include/net/selftests.h
26
struct packet_type pt;
include/sound/pt2258.h
20
extern int snd_pt2258_reset(struct snd_pt2258 *pt);
include/sound/pt2258.h
21
extern int snd_pt2258_build_controls(struct snd_pt2258 *pt);
io_uring/mock_file.c
168
static __poll_t io_mock_poll(struct file *file, struct poll_table_struct *pt)
io_uring/mock_file.c
173
poll_wait(file, &mf->poll_wq, pt);
io_uring/net.c
1801
struct poll_table_struct pt = { ._key = EPOLLERR };
io_uring/net.c
1803
if (vfs_poll(req->file, &pt) & EPOLLERR)
io_uring/poll.c
280
struct poll_table_struct pt = { ._key = req->apoll_events };
io_uring/poll.c
281
req->cqe.res = vfs_poll(req->file, &pt) & req->apoll_events;
io_uring/poll.c
35
struct poll_table_struct pt;
io_uring/poll.c
460
static void __io_queue_proc(struct io_poll *poll, struct io_poll_table *pt,
io_uring/poll.c
464
struct io_kiocb *req = pt->req;
io_uring/poll.c
472
if (unlikely(pt->nr_entries)) {
io_uring/poll.c
482
pt->error = -EINVAL;
io_uring/poll.c
488
pt->error = -ENOMEM;
io_uring/poll.c
506
pt->nr_entries++;
io_uring/poll.c
520
struct io_poll_table *pt = container_of(p, struct io_poll_table, pt);
io_uring/poll.c
521
struct io_poll *poll = io_kiocb_to_cmd(pt->req, struct io_poll);
io_uring/poll.c
523
__io_queue_proc(poll, pt, head,
io_uring/poll.c
524
(struct io_poll **) &pt->req->async_data);
io_uring/poll.c
528
struct io_poll_table *pt)
io_uring/poll.c
530
return pt->owning || io_poll_get_ownership(req);
io_uring/poll.c
558
ipt->pt._key = mask;
io_uring/poll.c
586
mask = vfs_poll(req->file, &ipt->pt) & poll->events;
io_uring/poll.c
636
struct io_poll_table *pt = container_of(p, struct io_poll_table, pt);
io_uring/poll.c
637
struct async_poll *apoll = pt->req->apoll;
io_uring/poll.c
639
__io_queue_proc(&apoll->poll, pt, head, &apoll->double_poll);
io_uring/poll.c
692
ipt.pt._qproc = io_async_queue_proc;
io_uring/poll.c
902
ipt.pt._qproc = io_poll_queue_proc;
io_uring/rw.c
45
struct poll_table_struct pt = { ._key = mask };
io_uring/rw.c
47
return vfs_poll(req->file, &pt) & mask;
kernel/bpf/btf.c
3518
int btf_find_next_decl_tag(const struct btf *btf, const struct btf_type *pt,
kernel/bpf/btf.c
3529
if (pt != btf_type_by_id(btf, t->type))
kernel/bpf/btf.c
3540
const char *btf_find_decl_tag_value(const struct btf *btf, const struct btf_type *pt,
kernel/bpf/btf.c
3547
id = btf_find_next_decl_tag(btf, pt, comp_idx, tag_key,
kernel/bpf/btf.c
3557
id = btf_find_next_decl_tag(btf, pt, comp_idx, tag_key, id);
kernel/bpf/btf.c
3565
btf_find_graph_root(const struct btf *btf, const struct btf_type *pt,
kernel/bpf/btf.c
3578
value_type = btf_find_decl_tag_value(btf, pt, comp_idx, "contains:");
kernel/cgroup/cgroup.c
4116
poll_table *pt)
kernel/cgroup/cgroup.c
4120
return psi_trigger_poll(&ctx->psi.trigger, of->file, pt);
kernel/cgroup/cgroup.c
4340
static __poll_t cgroup_file_poll(struct kernfs_open_file *of, poll_table *pt)
kernel/cgroup/cgroup.c
4345
return cft->poll(of, pt);
kernel/cgroup/cgroup.c
4347
return kernfs_generic_poll(of, pt);
kernel/workqueue.c
4812
struct wq_pod_type *pt;
kernel/workqueue.c
4822
pt = &wq_pod_types[scope];
kernel/workqueue.c
4825
likely(pt->nr_pods))
kernel/workqueue.c
4826
return pt;
kernel/workqueue.c
4832
pt = &wq_pod_types[WQ_AFFN_SYSTEM];
kernel/workqueue.c
4833
BUG_ON(!pt->nr_pods);
kernel/workqueue.c
4834
return pt;
kernel/workqueue.c
5098
struct wq_pod_type *pt = &wq_pod_types[WQ_AFFN_NUMA];
kernel/workqueue.c
5114
for (pod = 0; pod < pt->nr_pods; pod++) {
kernel/workqueue.c
5115
if (cpumask_subset(attrs->__pod_cpumask, pt->pod_cpus[pod])) {
kernel/workqueue.c
5116
node = pt->pod_node[pod];
kernel/workqueue.c
5303
const struct wq_pod_type *pt = wqattrs_pod_type(attrs);
kernel/workqueue.c
5304
int pod = pt->cpu_pod[cpu];
kernel/workqueue.c
5307
cpumask_and(attrs->__pod_cpumask, pt->pod_cpus[pod], attrs->cpumask);
kernel/workqueue.c
6774
const struct wq_pod_type *pt = wqattrs_pod_type(attrs);
kernel/workqueue.c
6777
for_each_cpu(tcpu, pt->pod_cpus[pt->cpu_pod[cpu]])
kernel/workqueue.c
6809
const struct wq_pod_type *pt = wqattrs_pod_type(attrs);
kernel/workqueue.c
6812
for_each_cpu(tcpu, pt->pod_cpus[pt->cpu_pod[cpu]])
kernel/workqueue.c
7866
struct wq_pod_type *pt = &wq_pod_types[WQ_AFFN_SYSTEM];
kernel/workqueue.c
7902
pt->pod_cpus = kzalloc_objs(pt->pod_cpus[0], 1);
kernel/workqueue.c
7903
pt->pod_node = kzalloc_objs(pt->pod_node[0], 1);
kernel/workqueue.c
7904
pt->cpu_pod = kzalloc_objs(pt->cpu_pod[0], nr_cpu_ids);
kernel/workqueue.c
7905
BUG_ON(!pt->pod_cpus || !pt->pod_node || !pt->cpu_pod);
kernel/workqueue.c
7907
BUG_ON(!zalloc_cpumask_var_node(&pt->pod_cpus[0], GFP_KERNEL, NUMA_NO_NODE));
kernel/workqueue.c
7909
pt->nr_pods = 1;
kernel/workqueue.c
7910
cpumask_copy(pt->pod_cpus[0], cpu_possible_mask);
kernel/workqueue.c
7911
pt->pod_node[0] = NUMA_NO_NODE;
kernel/workqueue.c
7912
pt->cpu_pod[0] = 0;
kernel/workqueue.c
8078
static void __init init_pod_type(struct wq_pod_type *pt,
kernel/workqueue.c
8083
pt->nr_pods = 0;
kernel/workqueue.c
8086
pt->cpu_pod = kzalloc_objs(pt->cpu_pod[0], nr_cpu_ids);
kernel/workqueue.c
8087
BUG_ON(!pt->cpu_pod);
kernel/workqueue.c
8092
pt->cpu_pod[cur] = pt->nr_pods++;
kernel/workqueue.c
8096
pt->cpu_pod[cur] = pt->cpu_pod[pre];
kernel/workqueue.c
8103
pt->pod_cpus = kzalloc_objs(pt->pod_cpus[0], pt->nr_pods);
kernel/workqueue.c
8104
pt->pod_node = kzalloc_objs(pt->pod_node[0], pt->nr_pods);
kernel/workqueue.c
8105
BUG_ON(!pt->pod_cpus || !pt->pod_node);
kernel/workqueue.c
8107
for (pod = 0; pod < pt->nr_pods; pod++)
kernel/workqueue.c
8108
BUG_ON(!zalloc_cpumask_var(&pt->pod_cpus[pod], GFP_KERNEL));
kernel/workqueue.c
8111
cpumask_set_cpu(cpu, pt->pod_cpus[pt->cpu_pod[cpu]]);
kernel/workqueue.c
8112
pt->pod_node[pt->cpu_pod[cpu]] = cpu_to_node(cpu);
lib/crypto/des.c
603
#define T1(x) pt[2 * (x) + 0]
lib/crypto/des.c
604
#define T2(x) pt[2 * (x) + 1]
lib/crypto/des.c
605
#define T3(x) pt[2 * (x) + 2]
lib/crypto/des.c
606
#define T4(x) pt[2 * (x) + 3]
lib/crypto/des.c
626
const u32 *pt = pc2;
lib/crypto/des.c
654
pt += 512;
lib/crypto/des.c
717
const u32 *pt = pc2;
lib/crypto/des.c
742
pt += 512;
lib/nlattr.c
119
void nla_get_range_unsigned(const struct nla_policy *pt,
lib/nlattr.c
122
WARN_ON_ONCE(pt->validation_type != NLA_VALIDATE_RANGE_PTR &&
lib/nlattr.c
123
(pt->min < 0 || pt->max < 0));
lib/nlattr.c
127
switch (pt->type) {
lib/nlattr.c
150
switch (pt->validation_type) {
lib/nlattr.c
153
range->min = pt->min;
lib/nlattr.c
154
range->max = pt->max;
lib/nlattr.c
157
*range = *pt->range;
lib/nlattr.c
160
range->min = pt->min;
lib/nlattr.c
163
range->max = pt->max;
lib/nlattr.c
170
static int nla_validate_range_unsigned(const struct nla_policy *pt,
lib/nlattr.c
178
switch (pt->type) {
lib/nlattr.c
210
nla_get_range_unsigned(pt, &range);
lib/nlattr.c
212
if (pt->validation_type == NLA_VALIDATE_RANGE_WARN_TOO_LONG &&
lib/nlattr.c
213
pt->type == NLA_BINARY && value > range.max) {
lib/nlattr.c
215
current->comm, pt->type);
lib/nlattr.c
217
NL_SET_ERR_MSG_ATTR_POL(extack, nla, pt,
lib/nlattr.c
227
bool binary = pt->type == NLA_BINARY;
lib/nlattr.c
230
NL_SET_ERR_MSG_ATTR_POL(extack, nla, pt,
lib/nlattr.c
233
NL_SET_ERR_MSG_ATTR_POL(extack, nla, pt,
lib/nlattr.c
242
void nla_get_range_signed(const struct nla_policy *pt,
lib/nlattr.c
245
switch (pt->type) {
lib/nlattr.c
268
switch (pt->validation_type) {
lib/nlattr.c
270
range->min = pt->min;
lib/nlattr.c
271
range->max = pt->max;
lib/nlattr.c
274
*range = *pt->range_signed;
lib/nlattr.c
277
range->min = pt->min;
lib/nlattr.c
280
range->max = pt->max;
lib/nlattr.c
287
static int nla_validate_int_range_signed(const struct nla_policy *pt,
lib/nlattr.c
294
switch (pt->type) {
lib/nlattr.c
314
nla_get_range_signed(pt, &range);
lib/nlattr.c
317
NL_SET_ERR_MSG_ATTR_POL(extack, nla, pt,
lib/nlattr.c
325
static int nla_validate_int_range(const struct nla_policy *pt,
lib/nlattr.c
330
switch (pt->type) {
lib/nlattr.c
340
return nla_validate_range_unsigned(pt, nla, extack, validate);
lib/nlattr.c
346
return nla_validate_int_range_signed(pt, nla, extack);
lib/nlattr.c
353
static int nla_validate_mask(const struct nla_policy *pt,
lib/nlattr.c
359
switch (pt->type) {
lib/nlattr.c
385
if (value & ~(u64)pt->mask) {
lib/nlattr.c
398
const struct nla_policy *pt;
lib/nlattr.c
409
pt = &policy[type];
lib/nlattr.c
411
BUG_ON(pt->type > NLA_TYPE_MAX);
lib/nlattr.c
413
if (nla_attr_len[pt->type] && attrlen != nla_attr_len[pt->type]) {
lib/nlattr.c
417
NL_SET_ERR_MSG_ATTR_POL(extack, nla, pt,
lib/nlattr.c
424
if ((pt->type == NLA_NESTED || pt->type == NLA_NESTED_ARRAY) &&
lib/nlattr.c
426
NL_SET_ERR_MSG_ATTR_POL(extack, nla, pt,
lib/nlattr.c
430
if (pt->type != NLA_NESTED && pt->type != NLA_NESTED_ARRAY &&
lib/nlattr.c
431
pt->type != NLA_UNSPEC && (nla->nla_type & NLA_F_NESTED)) {
lib/nlattr.c
432
NL_SET_ERR_MSG_ATTR_POL(extack, nla, pt,
lib/nlattr.c
438
switch (pt->type) {
lib/nlattr.c
440
if (extack && pt->reject_message) {
lib/nlattr.c
442
extack->_msg = pt->reject_message;
lib/nlattr.c
456
NL_SET_ERR_MSG_ATTR_POL(extack, nla, pt,
lib/nlattr.c
466
err = validate_nla_bitfield32(nla, pt->bitfield32_valid);
lib/nlattr.c
472
if (pt->len)
lib/nlattr.c
473
minlen = min_t(int, attrlen, pt->len + 1);
lib/nlattr.c
487
if (pt->len) {
lib/nlattr.c
493
if (attrlen > pt->len)
lib/nlattr.c
499
if (pt->len && attrlen > pt->len)
lib/nlattr.c
511
if (pt->nested_policy) {
lib/nlattr.c
513
pt->len, pt->nested_policy,
lib/nlattr.c
533
if (pt->nested_policy) {
lib/nlattr.c
537
pt->len, pt->nested_policy,
lib/nlattr.c
555
if (attrlen < pt->len)
lib/nlattr.c
560
if (pt->len)
lib/nlattr.c
561
minlen = pt->len;
lib/nlattr.c
563
minlen = nla_attr_minlen[pt->type];
lib/nlattr.c
570
switch (pt->validation_type) {
lib/nlattr.c
579
err = nla_validate_int_range(pt, nla, extack, validate);
lib/nlattr.c
584
err = nla_validate_mask(pt, nla, extack);
lib/nlattr.c
589
if (pt->validate) {
lib/nlattr.c
590
err = pt->validate(nla, extack);
lib/nlattr.c
599
NL_SET_ERR_MSG_ATTR_POL(extack, nla, pt,
lib/test_hmm.c
1014
xa_erase(&dmirror->pt, addr >> PAGE_SHIFT);
lib/test_hmm.c
175
xa_init(&dmirror->pt);
lib/test_hmm.c
193
xa_destroy(&dmirror->pt);
lib/test_hmm.c
236
entry = xa_store(&dmirror->pt, pfn, entry, GFP_ATOMIC);
lib/test_hmm.c
255
xa_for_each_range(&dmirror->pt, pfn, entry, start >> PAGE_SHIFT,
lib/test_hmm.c
257
xa_erase(&dmirror->pt, pfn);
lib/test_hmm.c
375
entry = xa_load(&dmirror->pt, pfn);
lib/test_hmm.c
441
entry = xa_load(&dmirror->pt, pfn);
lib/test_hmm.c
805
entry = xa_load(&dmirror->pt, pfn);
lib/test_hmm.c
822
entry = xa_store(&dmirror->pt, addr >> PAGE_SHIFT, entry, GFP_ATOMIC);
lib/test_hmm.c
873
entry = xa_store(&dmirror->pt, pfn + i, entry, GFP_ATOMIC);
lib/test_hmm.c
92
struct xarray pt;
mm/memcontrol-v1.c
1058
wait_queue_head_t *wqh, poll_table *pt)
mm/memcontrol-v1.c
1061
container_of(pt, struct mem_cgroup_event, pt);
mm/memcontrol-v1.c
1118
init_poll_funcptr(&event->pt, memcg_event_ptable_queue_proc);
mm/memcontrol-v1.c
1199
vfs_poll(fd_file(efile), &event->pt);
mm/memcontrol-v1.c
83
poll_table pt;
mm/pgtable-generic.c
427
struct ptdesc *pt, *next;
mm/pgtable-generic.c
435
list_for_each_entry_safe(pt, next, &page_list, pt_list)
mm/pgtable-generic.c
436
__pagetable_free(pt);
mm/pgtable-generic.c
439
void pagetable_free_kernel(struct ptdesc *pt)
mm/pgtable-generic.c
442
list_add(&pt->pt_list, &kernel_pgtable_work.list);
net/802/mrp.c
785
struct packet_type *pt, struct net_device *orig_dev)
net/802/mrp.c
787
struct mrp_application *appl = container_of(pt, struct mrp_application,
net/802/psnap.c
46
struct packet_type *pt, struct net_device *orig_dev)
net/802/stp.c
31
struct packet_type *pt, struct net_device *orig_dev)
net/9p/trans_fd.c
188
p9_fd_poll(struct p9_client *client, struct poll_table_struct *pt, int *err)
net/9p/trans_fd.c
202
ret = vfs_poll(ts->rd, pt);
net/9p/trans_fd.c
204
ret = (ret & ~EPOLLOUT) | (vfs_poll(ts->wr, pt) & ~EPOLLIN);
net/9p/trans_fd.c
520
struct p9_conn *m = container_of(p, struct p9_conn, pt);
net/9p/trans_fd.c
566
init_poll_funcptr(&m->pt, p9_pollwait);
net/9p/trans_fd.c
568
n = p9_fd_poll(client, &m->pt, NULL);
net/9p/trans_fd.c
93
poll_table pt;
net/appletalk/aarp.c
701
struct packet_type *pt, struct net_device *orig_dev)
net/appletalk/ddp.c
1403
struct packet_type *pt, struct net_device *orig_dev)
net/appletalk/ddp.c
1495
struct packet_type *pt, struct net_device *orig_dev)
net/appletalk/ddp.c
1541
return atalk_rcv(skb, dev, pt, orig_dev);
net/ax25/ax25_uid.c
171
struct ax25_uid_assoc *pt;
net/ax25/ax25_uid.c
173
pt = hlist_entry(v, struct ax25_uid_assoc, uid_node);
net/ax25/ax25_uid.c
175
from_kuid_munged(seq_user_ns(seq), pt->uid),
net/ax25/ax25_uid.c
176
ax2asc(buf, &pt->call));
net/bridge/br_ioctl.c
249
struct net_bridge_port *pt;
net/bridge/br_ioctl.c
252
if ((pt = br_get_port(br, args[2])) == NULL) {
net/bridge/br_ioctl.c
258
memcpy(&p.designated_root, &pt->designated_root, 8);
net/bridge/br_ioctl.c
259
memcpy(&p.designated_bridge, &pt->designated_bridge, 8);
net/bridge/br_ioctl.c
260
p.port_id = pt->port_id;
net/bridge/br_ioctl.c
261
p.designated_port = pt->designated_port;
net/bridge/br_ioctl.c
262
p.path_cost = pt->path_cost;
net/bridge/br_ioctl.c
263
p.designated_cost = pt->designated_cost;
net/bridge/br_ioctl.c
264
p.state = pt->state;
net/bridge/br_ioctl.c
265
p.top_change_ack = pt->topology_change_ack;
net/bridge/br_ioctl.c
266
p.config_pending = pt->config_pending;
net/bridge/br_ioctl.c
267
p.message_age_timer_value = br_timer_value(&pt->message_age_timer);
net/bridge/br_ioctl.c
268
p.forward_delay_timer_value = br_timer_value(&pt->forward_delay_timer);
net/bridge/br_ioctl.c
269
p.hold_timer_value = br_timer_value(&pt->hold_timer);
net/can/af_can.c
688
struct packet_type *pt, struct net_device *orig_dev)
net/can/af_can.c
704
struct packet_type *pt, struct net_device *orig_dev)
net/can/af_can.c
720
struct packet_type *pt, struct net_device *orig_dev)
net/core/dev.c
2496
struct packet_type **pt,
net/core/dev.c
2501
struct packet_type *ptype, *pt_prev = *pt;
net/core/dev.c
2510
*pt = pt_prev;
net/core/dev.c
594
static inline struct list_head *ptype_head(const struct packet_type *pt)
net/core/dev.c
596
if (pt->type == htons(ETH_P_ALL)) {
net/core/dev.c
597
if (!pt->af_packet_net && !pt->dev)
net/core/dev.c
600
return pt->dev ? &pt->dev->ptype_all :
net/core/dev.c
601
&pt->af_packet_net->ptype_all;
net/core/dev.c
604
if (pt->dev)
net/core/dev.c
605
return &pt->dev->ptype_specific;
net/core/dev.c
607
return pt->af_packet_net ? &pt->af_packet_net->ptype_specific :
net/core/dev.c
608
&ptype_base[ntohs(pt->type) & PTYPE_HASH_MASK];
net/core/dev.c
624
void dev_add_pack(struct packet_type *pt)
net/core/dev.c
626
struct list_head *head = ptype_head(pt);
net/core/dev.c
632
list_add_rcu(&pt->list, head);
net/core/dev.c
650
void __dev_remove_pack(struct packet_type *pt)
net/core/dev.c
652
struct list_head *head = ptype_head(pt);
net/core/dev.c
661
if (pt == pt1) {
net/core/dev.c
662
list_del_rcu(&pt->list);
net/core/dev.c
667
pr_warn("dev_remove_pack: %p not found\n", pt);
net/core/dev.c
685
void dev_remove_pack(struct packet_type *pt)
net/core/dev.c
687
__dev_remove_pack(pt);
net/core/net-procfs.c
182
struct packet_type *pt = NULL;
net/core/net-procfs.c
189
list_for_each_entry_rcu(pt, ptype_list, list) {
net/core/net-procfs.c
192
return pt;
net/core/net-procfs.c
200
list_for_each_entry_rcu(pt, &seq_file_net(seq)->ptype_all, list) {
net/core/net-procfs.c
202
return pt;
net/core/net-procfs.c
206
list_for_each_entry_rcu(pt, &seq_file_net(seq)->ptype_specific, list) {
net/core/net-procfs.c
208
return pt;
net/core/net-procfs.c
213
list_for_each_entry_rcu(pt, &ptype_base[t], list) {
net/core/net-procfs.c
215
return pt;
net/core/net-procfs.c
234
struct packet_type *pt;
net/core/net-procfs.c
242
pt = v;
net/core/net-procfs.c
243
nxt = READ_ONCE(pt->list.next);
net/core/net-procfs.c
261
if (pt->af_packet_net) {
net/core/net-procfs.c
276
hash = ntohs(pt->type) & PTYPE_HASH_MASK;
net/core/net-procfs.c
296
struct packet_type *pt = v;
net/core/net-procfs.c
304
if ((!pt->af_packet_net || net_eq(pt->af_packet_net, seq_file_net(seq))) &&
net/core/net-procfs.c
306
if (pt->type == htons(ETH_P_ALL))
net/core/net-procfs.c
309
seq_printf(seq, "%04x", ntohs(pt->type));
net/core/net-procfs.c
312
dev ? dev->name : "", pt->func);
net/core/selftests.c
161
struct packet_type *pt,
net/core/selftests.c
164
struct net_test_priv *tpriv = pt->af_packet_priv;
net/core/selftests.c
247
tpriv->pt.type = htons(ETH_P_IP);
net/core/selftests.c
248
tpriv->pt.func = net_test_loopback_validate;
net/core/selftests.c
249
tpriv->pt.dev = ndev;
net/core/selftests.c
250
tpriv->pt.af_packet_priv = tpriv;
net/core/selftests.c
252
dev_add_pack(&tpriv->pt);
net/core/selftests.c
281
dev_remove_pack(&tpriv->pt);
net/dsa/tag.c
55
struct packet_type *pt, struct net_device *unused)
net/hsr/hsr_device.c
678
enum hsr_port_type pt)
net/hsr/hsr_device.c
685
if (port->type == pt) {
net/hsr/hsr_main.c
133
struct hsr_port *hsr_port_get_hsr(struct hsr_priv *hsr, enum hsr_port_type pt)
net/hsr/hsr_main.c
138
if (port->type == pt)
net/hsr/hsr_main.h
230
struct hsr_port *hsr_port_get_hsr(struct hsr_priv *hsr, enum hsr_port_type pt);
net/hsr/hsr_slave.h
18
enum hsr_port_type pt, struct netlink_ext_ack *extack);
net/ieee802154/6lowpan/rx.c
272
struct packet_type *pt, struct net_device *orig_wdev)
net/ieee802154/socket.c
1070
struct packet_type *pt, struct net_device *orig_dev)
net/ipv4/arp.c
969
struct packet_type *pt, struct net_device *orig_dev)
net/ipv4/ip_input.c
564
int ip_rcv(struct sk_buff *skb, struct net_device *dev, struct packet_type *pt,
net/ipv4/ip_input.c
648
void ip_list_rcv(struct list_head *head, struct packet_type *pt,
net/ipv4/ipconfig.c
478
static int ic_rarp_recv(struct sk_buff *skb, struct net_device *dev, struct packet_type *pt, struct net_device *orig_dev);
net/ipv4/ipconfig.c
499
ic_rarp_recv(struct sk_buff *skb, struct net_device *dev, struct packet_type *pt, struct net_device *orig_dev)
net/ipv4/ipconfig.c
661
static int ic_bootp_recv(struct sk_buff *skb, struct net_device *dev, struct packet_type *pt, struct net_device *orig_dev);
net/ipv4/ipconfig.c
984
static int __init ic_bootp_recv(struct sk_buff *skb, struct net_device *dev, struct packet_type *pt, struct net_device *orig_dev)
net/ipv6/ip6_input.c
304
int ipv6_rcv(struct sk_buff *skb, struct net_device *dev, struct packet_type *pt, struct net_device *orig_dev)
net/ipv6/ip6_input.c
325
void ipv6_list_rcv(struct list_head *head, struct packet_type *pt,
net/iucv/af_iucv.c
2036
struct packet_type *pt, struct net_device *orig_dev)
net/llc/llc_core.c
94
struct packet_type *pt,
net/llc/llc_input.c
162
struct packet_type *pt, struct net_device *orig_dev)
net/llc/llc_input.c
203
rcv(skb, dev, pt, orig_dev);
net/llc/llc_input.c
210
rcv(cskb, dev, pt, orig_dev);
net/mctp/route.c
1319
struct packet_type *pt,
net/mpls/af_mpls.c
367
struct packet_type *pt, struct net_device *orig_dev)
net/ncsi/internal.h
414
struct packet_type *pt, struct net_device *orig_dev);
net/ncsi/ncsi-rsp.c
1166
struct packet_type *pt, struct net_device *orig_dev)
net/netlink/policy.c
209
int netlink_policy_dump_attr_size_estimate(const struct nla_policy *pt)
net/netlink/policy.c
214
switch (pt->type) {
net/netlink/policy.c
255
const struct nla_policy *pt,
net/netlink/policy.c
258
int estimate = netlink_policy_dump_attr_size_estimate(pt);
net/netlink/policy.c
266
switch (pt->type) {
net/netlink/policy.c
277
if (pt->type == NLA_NESTED_ARRAY)
net/netlink/policy.c
279
if (state && pt->nested_policy && pt->len &&
net/netlink/policy.c
282
pt->nested_policy,
net/netlink/policy.c
283
pt->len)) ||
net/netlink/policy.c
285
pt->len)))
net/netlink/policy.c
296
if (pt->type == NLA_U8)
net/netlink/policy.c
298
else if (pt->type == NLA_U16)
net/netlink/policy.c
300
else if (pt->type == NLA_U32)
net/netlink/policy.c
302
else if (pt->type == NLA_U64)
net/netlink/policy.c
307
if (pt->validation_type == NLA_VALIDATE_MASK) {
net/netlink/policy.c
309
pt->mask,
net/netlink/policy.c
313
} else if (pt->validation_type == NLA_VALIDATE_FUNCTION) {
net/netlink/policy.c
317
nla_get_range_unsigned(pt, &range);
net/netlink/policy.c
333
if (pt->type == NLA_S8)
net/netlink/policy.c
335
else if (pt->type == NLA_S16)
net/netlink/policy.c
337
else if (pt->type == NLA_S32)
net/netlink/policy.c
339
else if (pt->type == NLA_S64)
net/netlink/policy.c
344
if (pt->validation_type == NLA_VALIDATE_FUNCTION)
net/netlink/policy.c
347
nla_get_range_signed(pt, &range);
net/netlink/policy.c
359
pt->bitfield32_valid))
net/netlink/policy.c
365
if (pt->type == NLA_STRING)
net/netlink/policy.c
367
else if (pt->type == NLA_NUL_STRING)
net/netlink/policy.c
372
if (pt->validation_type == NLA_VALIDATE_RANGE ||
net/netlink/policy.c
373
pt->validation_type == NLA_VALIDATE_RANGE_WARN_TOO_LONG) {
net/netlink/policy.c
376
nla_get_range_unsigned(pt, &range);
net/netlink/policy.c
387
} else if (pt->len &&
net/netlink/policy.c
389
pt->len)) {
net/netlink/policy.c
421
const struct nla_policy *pt,
net/netlink/policy.c
424
return __netlink_policy_dump_write_attr(NULL, skb, pt, nestattr);
net/netlink/policy.c
437
const struct nla_policy *pt;
net/netlink/policy.c
445
pt = &state->policies[state->policy_idx].policy[state->attr_idx];
net/netlink/policy.c
451
err = __netlink_policy_dump_write_attr(state, skb, pt, state->attr_idx);
net/packet/af_packet.c
1456
struct packet_type *pt, struct net_device *orig_dev)
net/packet/af_packet.c
1458
struct packet_fanout *f = pt->af_packet_priv;
net/packet/af_packet.c
1857
struct packet_type *pt, struct net_device *orig_dev)
net/packet/af_packet.c
1867
sk = pt->af_packet_priv;
net/packet/af_packet.c
192
struct packet_type *pt, struct net_device *orig_dev);
net/packet/af_packet.c
2115
struct packet_type *pt, struct net_device *orig_dev)
net/packet/af_packet.c
2128
sk = pt->af_packet_priv;
net/packet/af_packet.c
2228
struct packet_type *pt, struct net_device *orig_dev)
net/packet/af_packet.c
2257
sk = pt->af_packet_priv;
net/rose/rose_subr.c
241
unsigned char *pt;
net/rose/rose_subr.c
311
for (pt = p + 2, lg = 0 ; lg < l ; pt += AX25_ADDR_LEN, lg += AX25_ADDR_LEN) {
net/rose/rose_subr.c
312
if (pt[6] & AX25_HBIT) {
net/rose/rose_subr.c
315
memcpy(&facilities->dest_digis[facilities->dest_ndigis++], pt, AX25_ADDR_LEN);
net/rose/rose_subr.c
319
memcpy(&facilities->source_digis[facilities->source_ndigis++], pt, AX25_ADDR_LEN);
net/sctp/input.c
1056
struct sctp_transport **pt,
net/sctp/input.c
1067
*pt = t;
net/sctp/input.c
64
struct sctp_transport **pt,
net/tipc/bearer.c
459
b->pt.dev = dev;
net/tipc/bearer.c
460
b->pt.type = htons(ETH_P_TIPC);
net/tipc/bearer.c
461
b->pt.func = tipc_l2_rcv_msg;
net/tipc/bearer.c
462
dev_add_pack(&b->pt);
net/tipc/bearer.c
483
dev_remove_pack(&b->pt);
net/tipc/bearer.c
657
struct packet_type *pt, struct net_device *orig_dev)
net/tipc/bearer.c
668
tipc_rcv(dev_net(b->pt.dev), skb, b);
net/tipc/bearer.c
71
struct packet_type *pt, struct net_device *orig_dev);
net/tipc/bearer.c
792
struct packet_type *pt, struct net_device *od)
net/tipc/bearer.h
164
struct packet_type pt;
security/apparmor/apparmorfs.c
665
static __poll_t ns_revision_poll(struct file *file, poll_table *pt)
security/apparmor/apparmorfs.c
672
poll_wait(file, &rev->ns->wait, pt);
security/integrity/ima/ima_policy.c
2130
seq_puts(m, pt(Opt_measure));
security/integrity/ima/ima_policy.c
2132
seq_puts(m, pt(Opt_dont_measure));
security/integrity/ima/ima_policy.c
2134
seq_puts(m, pt(Opt_appraise));
security/integrity/ima/ima_policy.c
2136
seq_puts(m, pt(Opt_dont_appraise));
security/integrity/ima/ima_policy.c
2138
seq_puts(m, pt(Opt_audit));
security/integrity/ima/ima_policy.c
2140
seq_puts(m, pt(Opt_dont_audit));
security/integrity/ima/ima_policy.c
2142
seq_puts(m, pt(Opt_hash));
security/integrity/ima/ima_policy.c
2144
seq_puts(m, pt(Opt_dont_hash));
security/integrity/ima/ima_policy.c
2155
seq_printf(m, pt(Opt_mask), mt(mask_exec) + offset);
security/integrity/ima/ima_policy.c
2157
seq_printf(m, pt(Opt_mask), mt(mask_write) + offset);
security/integrity/ima/ima_policy.c
2159
seq_printf(m, pt(Opt_mask), mt(mask_read) + offset);
security/integrity/ima/ima_policy.c
2161
seq_printf(m, pt(Opt_mask), mt(mask_append) + offset);
security/integrity/ima/ima_policy.c
2167
seq_printf(m, pt(Opt_fsmagic), tbuf);
security/integrity/ima/ima_policy.c
2173
seq_printf(m, pt(Opt_fsname), tbuf);
security/integrity/ima/ima_policy.c
2179
seq_printf(m, pt(Opt_fs_subtype), tbuf);
security/integrity/ima/ima_policy.c
2197
seq_printf(m, pt(Opt_pcr), tbuf);
security/integrity/ima/ima_policy.c
2209
seq_printf(m, pt(Opt_uid_gt), tbuf);
security/integrity/ima/ima_policy.c
2211
seq_printf(m, pt(Opt_uid_lt), tbuf);
security/integrity/ima/ima_policy.c
2213
seq_printf(m, pt(Opt_uid_eq), tbuf);
security/integrity/ima/ima_policy.c
2220
seq_printf(m, pt(Opt_euid_gt), tbuf);
security/integrity/ima/ima_policy.c
2222
seq_printf(m, pt(Opt_euid_lt), tbuf);
security/integrity/ima/ima_policy.c
2224
seq_printf(m, pt(Opt_euid_eq), tbuf);
security/integrity/ima/ima_policy.c
2231
seq_printf(m, pt(Opt_gid_gt), tbuf);
security/integrity/ima/ima_policy.c
2233
seq_printf(m, pt(Opt_gid_lt), tbuf);
security/integrity/ima/ima_policy.c
2235
seq_printf(m, pt(Opt_gid_eq), tbuf);
security/integrity/ima/ima_policy.c
2242
seq_printf(m, pt(Opt_egid_gt), tbuf);
security/integrity/ima/ima_policy.c
2244
seq_printf(m, pt(Opt_egid_lt), tbuf);
security/integrity/ima/ima_policy.c
2246
seq_printf(m, pt(Opt_egid_eq), tbuf);
security/integrity/ima/ima_policy.c
2253
seq_printf(m, pt(Opt_fowner_gt), tbuf);
security/integrity/ima/ima_policy.c
2255
seq_printf(m, pt(Opt_fowner_lt), tbuf);
security/integrity/ima/ima_policy.c
2257
seq_printf(m, pt(Opt_fowner_eq), tbuf);
security/integrity/ima/ima_policy.c
2264
seq_printf(m, pt(Opt_fgroup_gt), tbuf);
security/integrity/ima/ima_policy.c
2266
seq_printf(m, pt(Opt_fgroup_lt), tbuf);
security/integrity/ima/ima_policy.c
2268
seq_printf(m, pt(Opt_fgroup_eq), tbuf);
security/integrity/ima/ima_policy.c
2282
seq_printf(m, pt(Opt_obj_user),
security/integrity/ima/ima_policy.c
2286
seq_printf(m, pt(Opt_obj_role),
security/integrity/ima/ima_policy.c
2290
seq_printf(m, pt(Opt_obj_type),
security/integrity/ima/ima_policy.c
2294
seq_printf(m, pt(Opt_subj_user),
security/integrity/ima/ima_policy.c
2298
seq_printf(m, pt(Opt_subj_role),
security/integrity/ima/ima_policy.c
2302
seq_printf(m, pt(Opt_subj_type),
sound/i2c/other/pt2258.c
104
if (val0 == pt->volume[base] && val1 == pt->volume[base + 1])
sound/i2c/other/pt2258.c
107
pt->volume[base] = val0;
sound/i2c/other/pt2258.c
110
snd_i2c_lock(pt->i2c_bus);
sound/i2c/other/pt2258.c
111
if (snd_i2c_sendbytes(pt->i2c_dev, bytes, 2) != 2)
sound/i2c/other/pt2258.c
113
snd_i2c_unlock(pt->i2c_bus);
sound/i2c/other/pt2258.c
115
pt->volume[base + 1] = val1;
sound/i2c/other/pt2258.c
118
snd_i2c_lock(pt->i2c_bus);
sound/i2c/other/pt2258.c
119
if (snd_i2c_sendbytes(pt->i2c_dev, bytes, 2) != 2)
sound/i2c/other/pt2258.c
121
snd_i2c_unlock(pt->i2c_bus);
sound/i2c/other/pt2258.c
126
snd_i2c_unlock(pt->i2c_bus);
sound/i2c/other/pt2258.c
127
dev_err(pt->card->dev, "PT2258 access failed\n");
sound/i2c/other/pt2258.c
136
struct snd_pt2258 *pt = snd_kcontrol_chip(kcontrol);
sound/i2c/other/pt2258.c
138
ucontrol->value.integer.value[0] = !pt->mute;
sound/i2c/other/pt2258.c
145
struct snd_pt2258 *pt = snd_kcontrol_chip(kcontrol);
sound/i2c/other/pt2258.c
150
if (pt->mute == val)
sound/i2c/other/pt2258.c
153
pt->mute = val;
sound/i2c/other/pt2258.c
155
snd_i2c_lock(pt->i2c_bus);
sound/i2c/other/pt2258.c
156
if (snd_i2c_sendbytes(pt->i2c_dev, bytes, 1) != 1)
sound/i2c/other/pt2258.c
158
snd_i2c_unlock(pt->i2c_bus);
sound/i2c/other/pt2258.c
163
snd_i2c_unlock(pt->i2c_bus);
sound/i2c/other/pt2258.c
164
dev_err(pt->card->dev, "PT2258 access failed 2\n");
sound/i2c/other/pt2258.c
170
int snd_pt2258_build_controls(struct snd_pt2258 *pt)
sound/i2c/other/pt2258.c
193
err = snd_ctl_add(pt->card, snd_ctl_new1(&knew, pt));
sound/i2c/other/pt2258.c
205
err = snd_ctl_add(pt->card, snd_ctl_new1(&knew, pt));
sound/i2c/other/pt2258.c
32
int snd_pt2258_reset(struct snd_pt2258 *pt)
sound/i2c/other/pt2258.c
39
snd_i2c_lock(pt->i2c_bus);
sound/i2c/other/pt2258.c
40
if (snd_i2c_sendbytes(pt->i2c_dev, bytes, 1) != 1)
sound/i2c/other/pt2258.c
42
snd_i2c_unlock(pt->i2c_bus);
sound/i2c/other/pt2258.c
45
pt->mute = 1;
sound/i2c/other/pt2258.c
47
snd_i2c_lock(pt->i2c_bus);
sound/i2c/other/pt2258.c
48
if (snd_i2c_sendbytes(pt->i2c_dev, bytes, 1) != 1)
sound/i2c/other/pt2258.c
50
snd_i2c_unlock(pt->i2c_bus);
sound/i2c/other/pt2258.c
54
pt->volume[i] = 0;
sound/i2c/other/pt2258.c
57
snd_i2c_lock(pt->i2c_bus);
sound/i2c/other/pt2258.c
58
if (snd_i2c_sendbytes(pt->i2c_dev, bytes, 2) != 2)
sound/i2c/other/pt2258.c
60
snd_i2c_unlock(pt->i2c_bus);
sound/i2c/other/pt2258.c
65
snd_i2c_unlock(pt->i2c_bus);
sound/i2c/other/pt2258.c
66
dev_err(pt->card->dev, "PT2258 reset failed\n");
sound/i2c/other/pt2258.c
83
struct snd_pt2258 *pt = snd_kcontrol_chip(kcontrol);
sound/i2c/other/pt2258.c
87
ucontrol->value.integer.value[0] = 79 - pt->volume[base];
sound/i2c/other/pt2258.c
88
ucontrol->value.integer.value[1] = 79 - pt->volume[base + 1];
sound/i2c/other/pt2258.c
95
struct snd_pt2258 *pt = snd_kcontrol_chip(kcontrol);
sound/isa/sb/emu8000_pcm.c
328
int pt = calc_pitch_target(rec->pitch);
sound/isa/sb/emu8000_pcm.c
342
temp = (temp << 8) | (pt << 16) | aux;
sound/isa/sb/emu8000_pcm.c
344
EMU8000_CPF_WRITE(hw, ch, pt << 16);
sound/pci/ice1712/revo.c
145
struct snd_pt2258 *pt)
sound/pci/ice1712/revo.c
168
pt->card = ice->card;
sound/pci/ice1712/revo.c
169
pt->i2c_bus = ice->i2c;
sound/pci/ice1712/revo.c
170
pt->i2c_dev = spec->dev;
sound/pci/ice1712/revo.c
171
spec->pt2258 = pt;
sound/pci/ice1712/revo.c
173
snd_pt2258_reset(pt);
sound/pci/riptide/riptide.c
1477
unsigned int i, j, size, pages, f, pt, period;
sound/pci/riptide/riptide.c
1491
pt = 0;
sound/pci/riptide/riptide.c
1503
addr = snd_pcm_sgbuf_get_addr(substream, ofs) + pt;
sound/pci/riptide/riptide.c
1505
pt = (pt + f) % PAGE_SIZE;
sound/pci/riptide/riptide.c
1506
if (pt == 0)
sound/usb/pcm.c
1115
unsigned int pt, ptmin;
sound/usb/pcm.c
1143
pt = 125 * (1 << fp->datainterval);
sound/usb/pcm.c
1144
ptmin = min(ptmin, pt);
sound/usb/pcm.c
761
struct snd_interval *pt = hw_param_interval(params, SNDRV_PCM_HW_PARAM_PERIOD_TIME);
sound/usb/pcm.c
791
if (ptime > pt->max || (ptime == pt->max && pt->openmax)) {
sound/usb/pcm.c
792
hwc_debug(" > check: ptime %u > max %u\n", ptime, pt->max);
tools/lib/bpf/btf_dump.c
2210
union ptr_data pt;
tools/lib/bpf/btf_dump.c
2212
memcpy(&pt, data, d->ptr_sz);
tools/lib/bpf/btf_dump.c
2214
btf_dump_type_values(d, "0x%x", pt.p);
tools/lib/bpf/btf_dump.c
2216
btf_dump_type_values(d, "0x%llx", pt.lp);
tools/lib/bpf/nlattr.c
48
struct libbpf_nla_policy *pt;
tools/lib/bpf/nlattr.c
55
pt = &policy[type];
tools/lib/bpf/nlattr.c
57
if (pt->type > LIBBPF_NLA_TYPE_MAX)
tools/lib/bpf/nlattr.c
60
if (pt->minlen)
tools/lib/bpf/nlattr.c
61
minlen = pt->minlen;
tools/lib/bpf/nlattr.c
62
else if (pt->type != LIBBPF_NLA_UNSPEC)
tools/lib/bpf/nlattr.c
63
minlen = nla_attr_minlen[pt->type];
tools/lib/bpf/nlattr.c
68
if (pt->maxlen && libbpf_nla_len(nla) > pt->maxlen)
tools/lib/bpf/nlattr.c
71
if (pt->type == LIBBPF_NLA_STRING) {
tools/perf/tests/mmap-thread-lookup.c
128
err = pthread_join(threads[i].pt, NULL);
tools/perf/tests/mmap-thread-lookup.c
27
pthread_t pt;
tools/perf/tests/mmap-thread-lookup.c
88
err = pthread_create(&td->pt, NULL, thread_fn, td);
tools/perf/util/intel-pt.c
1000
return intel_pt_match_pgd_ip(ptq->pt, ip, ip, NULL);
tools/perf/util/intel-pt.c
1015
res = intel_pt_match_pgd_ip(ptq->pt, ip, offset, dso__long_name(map__dso(al.map)));
tools/perf/util/intel-pt.c
1025
static bool intel_pt_get_config(struct intel_pt *pt,
tools/perf/util/intel-pt.c
1028
if (attr->type == pt->pmu_type) {
tools/perf/util/intel-pt.c
1037
static bool intel_pt_exclude_kernel(struct intel_pt *pt)
tools/perf/util/intel-pt.c
1041
evlist__for_each_entry(pt->session->evlist, evsel) {
tools/perf/util/intel-pt.c
1042
if (intel_pt_get_config(pt, &evsel->core.attr, NULL) &&
tools/perf/util/intel-pt.c
1049
static bool intel_pt_return_compression(struct intel_pt *pt)
tools/perf/util/intel-pt.c
1054
if (!pt->noretcomp_bit)
tools/perf/util/intel-pt.c
1057
evlist__for_each_entry(pt->session->evlist, evsel) {
tools/perf/util/intel-pt.c
1058
if (intel_pt_get_config(pt, &evsel->core.attr, &config) &&
tools/perf/util/intel-pt.c
1059
(config & pt->noretcomp_bit))
tools/perf/util/intel-pt.c
1065
static bool intel_pt_branch_enable(struct intel_pt *pt)
tools/perf/util/intel-pt.c
1070
evlist__for_each_entry(pt->session->evlist, evsel) {
tools/perf/util/intel-pt.c
1071
if (intel_pt_get_config(pt, &evsel->core.attr, &config) &&
tools/perf/util/intel-pt.c
1079
static bool intel_pt_disabled_tnt(struct intel_pt *pt)
tools/perf/util/intel-pt.c
1084
evlist__for_each_entry(pt->session->evlist, evsel) {
tools/perf/util/intel-pt.c
1085
if (intel_pt_get_config(pt, &evsel->core.attr, &config) &&
tools/perf/util/intel-pt.c
1092
static unsigned int intel_pt_mtc_period(struct intel_pt *pt)
tools/perf/util/intel-pt.c
1098
if (!pt->mtc_freq_bits)
tools/perf/util/intel-pt.c
1101
for (shift = 0, config = pt->mtc_freq_bits; !(config & 1); shift++)
tools/perf/util/intel-pt.c
1104
evlist__for_each_entry(pt->session->evlist, evsel) {
tools/perf/util/intel-pt.c
1105
if (intel_pt_get_config(pt, &evsel->core.attr, &config))
tools/perf/util/intel-pt.c
1106
return (config & pt->mtc_freq_bits) >> shift;
tools/perf/util/intel-pt.c
1111
static bool intel_pt_timeless_decoding(struct intel_pt *pt)
tools/perf/util/intel-pt.c
1117
if (!pt->tsc_bit || !pt->cap_user_time_zero || pt->synth_opts.timeless_decoding)
tools/perf/util/intel-pt.c
1120
evlist__for_each_entry(pt->session->evlist, evsel) {
tools/perf/util/intel-pt.c
1123
if (intel_pt_get_config(pt, &evsel->core.attr, &config)) {
tools/perf/util/intel-pt.c
1124
if (config & pt->tsc_bit)
tools/perf/util/intel-pt.c
1133
static bool intel_pt_tracing_kernel(struct intel_pt *pt)
tools/perf/util/intel-pt.c
1137
evlist__for_each_entry(pt->session->evlist, evsel) {
tools/perf/util/intel-pt.c
1138
if (intel_pt_get_config(pt, &evsel->core.attr, NULL) &&
tools/perf/util/intel-pt.c
1145
static bool intel_pt_have_tsc(struct intel_pt *pt)
tools/perf/util/intel-pt.c
1151
if (!pt->tsc_bit)
tools/perf/util/intel-pt.c
1154
evlist__for_each_entry(pt->session->evlist, evsel) {
tools/perf/util/intel-pt.c
1155
if (intel_pt_get_config(pt, &evsel->core.attr, &config)) {
tools/perf/util/intel-pt.c
1156
if (config & pt->tsc_bit)
tools/perf/util/intel-pt.c
1165
static bool intel_pt_have_mtc(struct intel_pt *pt)
tools/perf/util/intel-pt.c
1170
evlist__for_each_entry(pt->session->evlist, evsel) {
tools/perf/util/intel-pt.c
1171
if (intel_pt_get_config(pt, &evsel->core.attr, &config) &&
tools/perf/util/intel-pt.c
1172
(config & pt->mtc_bit))
tools/perf/util/intel-pt.c
1178
static bool intel_pt_sampling_mode(struct intel_pt *pt)
tools/perf/util/intel-pt.c
1182
evlist__for_each_entry(pt->session->evlist, evsel) {
tools/perf/util/intel-pt.c
1190
static u64 intel_pt_ctl(struct intel_pt *pt)
tools/perf/util/intel-pt.c
1195
evlist__for_each_entry(pt->session->evlist, evsel) {
tools/perf/util/intel-pt.c
1196
if (intel_pt_get_config(pt, &evsel->core.attr, &config))
tools/perf/util/intel-pt.c
1202
static u64 intel_pt_ns_to_ticks(const struct intel_pt *pt, u64 ns)
tools/perf/util/intel-pt.c
1206
quot = ns / pt->tc.time_mult;
tools/perf/util/intel-pt.c
1207
rem = ns % pt->tc.time_mult;
tools/perf/util/intel-pt.c
1208
return (quot << pt->tc.time_shift) + (rem << pt->tc.time_shift) /
tools/perf/util/intel-pt.c
1209
pt->tc.time_mult;
tools/perf/util/intel-pt.c
1212
static struct ip_callchain *intel_pt_alloc_chain(struct intel_pt *pt)
tools/perf/util/intel-pt.c
1217
sz += (pt->synth_opts.callchain_sz + 1) * sizeof(u64);
tools/perf/util/intel-pt.c
1221
static int intel_pt_callchain_init(struct intel_pt *pt)
tools/perf/util/intel-pt.c
1225
evlist__for_each_entry(pt->session->evlist, evsel) {
tools/perf/util/intel-pt.c
1230
pt->chain = intel_pt_alloc_chain(pt);
tools/perf/util/intel-pt.c
1231
if (!pt->chain)
tools/perf/util/intel-pt.c
1237
static void intel_pt_add_callchain(struct intel_pt *pt,
tools/perf/util/intel-pt.c
1240
struct thread *thread = machine__findnew_thread(pt->machine,
tools/perf/util/intel-pt.c
1244
thread_stack__sample_late(thread, sample->cpu, pt->chain,
tools/perf/util/intel-pt.c
1245
pt->synth_opts.callchain_sz + 1, sample->ip,
tools/perf/util/intel-pt.c
1246
pt->kernel_start);
tools/perf/util/intel-pt.c
1248
sample->callchain = pt->chain;
tools/perf/util/intel-pt.c
1259
static int intel_pt_br_stack_init(struct intel_pt *pt)
tools/perf/util/intel-pt.c
1263
evlist__for_each_entry(pt->session->evlist, evsel) {
tools/perf/util/intel-pt.c
1268
pt->br_stack = intel_pt_alloc_br_stack(pt->br_stack_sz);
tools/perf/util/intel-pt.c
1269
if (!pt->br_stack)
tools/perf/util/intel-pt.c
1275
static void intel_pt_add_br_stack(struct intel_pt *pt,
tools/perf/util/intel-pt.c
1278
struct thread *thread = machine__findnew_thread(pt->machine,
tools/perf/util/intel-pt.c
1282
thread_stack__br_sample_late(thread, sample->cpu, pt->br_stack,
tools/perf/util/intel-pt.c
1283
pt->br_stack_sz, sample->ip,
tools/perf/util/intel-pt.c
1284
pt->kernel_start);
tools/perf/util/intel-pt.c
1286
sample->branch_stack = pt->br_stack;
tools/perf/util/intel-pt.c
1293
static struct intel_pt_queue *intel_pt_alloc_queue(struct intel_pt *pt,
tools/perf/util/intel-pt.c
1297
struct perf_env *env = pt->machine->env;
tools/perf/util/intel-pt.c
1304
if (pt->synth_opts.callchain) {
tools/perf/util/intel-pt.c
1305
ptq->chain = intel_pt_alloc_chain(pt);
tools/perf/util/intel-pt.c
1310
if (pt->synth_opts.last_branch || pt->synth_opts.other_events) {
tools/perf/util/intel-pt.c
1311
unsigned int entry_cnt = max(LBRS_MAX, pt->br_stack_sz);
tools/perf/util/intel-pt.c
1322
ptq->pt = pt;
tools/perf/util/intel-pt.c
1324
ptq->exclude_kernel = intel_pt_exclude_kernel(pt);
tools/perf/util/intel-pt.c
1335
params.return_compression = intel_pt_return_compression(pt);
tools/perf/util/intel-pt.c
1336
params.branch_enable = intel_pt_branch_enable(pt);
tools/perf/util/intel-pt.c
1337
params.ctl = intel_pt_ctl(pt);
tools/perf/util/intel-pt.c
1338
params.max_non_turbo_ratio = pt->max_non_turbo_ratio;
tools/perf/util/intel-pt.c
1339
params.mtc_period = intel_pt_mtc_period(pt);
tools/perf/util/intel-pt.c
1340
params.tsc_ctc_ratio_n = pt->tsc_ctc_ratio_n;
tools/perf/util/intel-pt.c
1341
params.tsc_ctc_ratio_d = pt->tsc_ctc_ratio_d;
tools/perf/util/intel-pt.c
1342
params.quick = pt->synth_opts.quick;
tools/perf/util/intel-pt.c
1343
params.vm_time_correlation = pt->synth_opts.vm_time_correlation;
tools/perf/util/intel-pt.c
1344
params.vm_tm_corr_dry_run = pt->synth_opts.vm_tm_corr_dry_run;
tools/perf/util/intel-pt.c
1345
params.first_timestamp = pt->first_timestamp;
tools/perf/util/intel-pt.c
1346
params.max_loops = pt->max_loops;
tools/perf/util/intel-pt.c
1349
if (params.branch_enable && intel_pt_disabled_tnt(pt) && !params.quick)
tools/perf/util/intel-pt.c
1352
if (pt->filts.cnt > 0)
tools/perf/util/intel-pt.c
1355
if (pt->synth_opts.instructions || pt->synth_opts.cycles) {
tools/perf/util/intel-pt.c
1356
if (pt->synth_opts.period) {
tools/perf/util/intel-pt.c
1357
switch (pt->synth_opts.period_type) {
tools/perf/util/intel-pt.c
1361
params.period = pt->synth_opts.period;
tools/perf/util/intel-pt.c
1365
params.period = pt->synth_opts.period;
tools/perf/util/intel-pt.c
1369
params.period = intel_pt_ns_to_ticks(pt,
tools/perf/util/intel-pt.c
1370
pt->synth_opts.period);
tools/perf/util/intel-pt.c
1416
static void intel_pt_first_timestamp(struct intel_pt *pt, u64 timestamp)
tools/perf/util/intel-pt.c
1420
pt->first_timestamp = timestamp;
tools/perf/util/intel-pt.c
1422
for (i = 0; i < pt->queues.nr_queues; i++) {
tools/perf/util/intel-pt.c
1423
struct auxtrace_queue *queue = &pt->queues.queue_array[i];
tools/perf/util/intel-pt.c
1433
struct machines *machines = &ptq->pt->session->machines;
tools/perf/util/intel-pt.c
1480
static void intel_pt_set_pid_tid_cpu(struct intel_pt *pt,
tools/perf/util/intel-pt.c
1485
if (queue->tid == -1 || pt->have_sched_switch) {
tools/perf/util/intel-pt.c
1486
ptq->tid = machine__get_current_tid(pt->machine, ptq->cpu);
tools/perf/util/intel-pt.c
1493
ptq->thread = machine__find_thread(pt->machine, -1, ptq->tid);
tools/perf/util/intel-pt.c
1501
if (pt->have_guest_sideband && intel_pt_get_guest_from_sideband(ptq)) {
tools/perf/util/intel-pt.c
1511
struct intel_pt *pt = ptq->pt;
tools/perf/util/intel-pt.c
1546
if (pt->cap_event_trace) {
tools/perf/util/intel-pt.c
1558
static void intel_pt_setup_time_range(struct intel_pt *pt,
tools/perf/util/intel-pt.c
1561
if (!pt->range_cnt)
tools/perf/util/intel-pt.c
1564
ptq->sel_timestamp = pt->time_ranges[0].start;
tools/perf/util/intel-pt.c
1570
ptq->sel_timestamp = pt->time_ranges[0].end;
tools/perf/util/intel-pt.c
1575
static int intel_pt_setup_queue(struct intel_pt *pt,
tools/perf/util/intel-pt.c
1585
ptq = intel_pt_alloc_queue(pt, queue_nr);
tools/perf/util/intel-pt.c
1596
if (pt->sampling_mode && !pt->snapshot_mode &&
tools/perf/util/intel-pt.c
1597
pt->timeless_decoding)
tools/perf/util/intel-pt.c
1600
ptq->sync_switch = pt->sync_switch;
tools/perf/util/intel-pt.c
1602
intel_pt_setup_time_range(pt, ptq);
tools/perf/util/intel-pt.c
1611
if (pt->timeless_decoding)
tools/perf/util/intel-pt.c
1648
ret = auxtrace_heap__add(&pt->heap, queue_nr, ptq->timestamp);
tools/perf/util/intel-pt.c
1657
static int intel_pt_setup_queues(struct intel_pt *pt)
tools/perf/util/intel-pt.c
1662
for (i = 0; i < pt->queues.nr_queues; i++) {
tools/perf/util/intel-pt.c
1663
ret = intel_pt_setup_queue(pt, &pt->queues.queue_array[i], i);
tools/perf/util/intel-pt.c
1670
static inline bool intel_pt_skip_event(struct intel_pt *pt)
tools/perf/util/intel-pt.c
1672
return pt->synth_opts.initial_skip &&
tools/perf/util/intel-pt.c
1673
pt->num_events++ < pt->synth_opts.initial_skip;
tools/perf/util/intel-pt.c
1681
static inline bool intel_pt_skip_cbr_event(struct intel_pt *pt)
tools/perf/util/intel-pt.c
1683
return pt->synth_opts.initial_skip &&
tools/perf/util/intel-pt.c
1684
pt->num_events + 4 < pt->synth_opts.initial_skip;
tools/perf/util/intel-pt.c
1697
if (ptq->pt->have_guest_sideband) {
tools/perf/util/intel-pt.c
1712
static void intel_pt_prep_b_sample(struct intel_pt *pt,
tools/perf/util/intel-pt.c
1719
if (!pt->timeless_decoding)
tools/perf/util/intel-pt.c
1720
sample->time = tsc_to_perf_time(ptq->timestamp, &pt->tc);
tools/perf/util/intel-pt.c
1738
static inline int intel_pt_opt_inject(struct intel_pt *pt,
tools/perf/util/intel-pt.c
1742
if (!pt->synth_opts.inject)
tools/perf/util/intel-pt.c
1748
static int intel_pt_deliver_synth_event(struct intel_pt *pt,
tools/perf/util/intel-pt.c
1754
ret = intel_pt_opt_inject(pt, event, sample, type);
tools/perf/util/intel-pt.c
1758
ret = perf_session__deliver_synth_event(pt->session, event, sample);
tools/perf/util/intel-pt.c
1767
struct intel_pt *pt = ptq->pt;
tools/perf/util/intel-pt.c
1777
if (pt->branches_filter && !(pt->branches_filter & ptq->flags))
tools/perf/util/intel-pt.c
1780
if (intel_pt_skip_event(pt))
tools/perf/util/intel-pt.c
1784
intel_pt_prep_b_sample(pt, ptq, event, &sample);
tools/perf/util/intel-pt.c
1786
sample.id = ptq->pt->branches_id;
tools/perf/util/intel-pt.c
1787
sample.stream_id = ptq->pt->branches_id;
tools/perf/util/intel-pt.c
1793
if (pt->synth_opts.last_branch && sort__mode == SORT_MODE__BRANCH) {
tools/perf/util/intel-pt.c
1814
ret = intel_pt_deliver_synth_event(pt, event, &sample,
tools/perf/util/intel-pt.c
1815
pt->branches_sample_type);
tools/perf/util/intel-pt.c
1819
static void intel_pt_prep_sample(struct intel_pt *pt,
tools/perf/util/intel-pt.c
1824
intel_pt_prep_b_sample(pt, ptq, event, sample);
tools/perf/util/intel-pt.c
1826
if (pt->synth_opts.callchain) {
tools/perf/util/intel-pt.c
1828
pt->synth_opts.callchain_sz + 1,
tools/perf/util/intel-pt.c
1829
sample->ip, pt->kernel_start);
tools/perf/util/intel-pt.c
183
struct intel_pt *pt;
tools/perf/util/intel-pt.c
1833
if (pt->synth_opts.last_branch) {
tools/perf/util/intel-pt.c
1835
pt->br_stack_sz);
tools/perf/util/intel-pt.c
1842
struct intel_pt *pt = ptq->pt;
tools/perf/util/intel-pt.c
1847
if (intel_pt_skip_event(pt))
tools/perf/util/intel-pt.c
1851
intel_pt_prep_sample(pt, ptq, event, &sample);
tools/perf/util/intel-pt.c
1853
sample.id = ptq->pt->instructions_id;
tools/perf/util/intel-pt.c
1854
sample.stream_id = ptq->pt->instructions_id;
tools/perf/util/intel-pt.c
1855
if (pt->synth_opts.quick)
tools/perf/util/intel-pt.c
1870
ret = intel_pt_deliver_synth_event(pt, event, &sample,
tools/perf/util/intel-pt.c
1871
pt->instructions_sample_type);
tools/perf/util/intel-pt.c
1878
struct intel_pt *pt = ptq->pt;
tools/perf/util/intel-pt.c
1887
if (!period || intel_pt_skip_event(pt))
tools/perf/util/intel-pt.c
1891
intel_pt_prep_sample(pt, ptq, event, &sample);
tools/perf/util/intel-pt.c
1893
sample.id = ptq->pt->cycles_id;
tools/perf/util/intel-pt.c
1894
sample.stream_id = ptq->pt->cycles_id;
tools/perf/util/intel-pt.c
1902
ret = intel_pt_deliver_synth_event(pt, event, &sample, pt->cycles_sample_type);
tools/perf/util/intel-pt.c
1909
struct intel_pt *pt = ptq->pt;
tools/perf/util/intel-pt.c
1914
if (intel_pt_skip_event(pt))
tools/perf/util/intel-pt.c
1918
intel_pt_prep_sample(pt, ptq, event, &sample);
tools/perf/util/intel-pt.c
1920
sample.id = ptq->pt->transactions_id;
tools/perf/util/intel-pt.c
1921
sample.stream_id = ptq->pt->transactions_id;
tools/perf/util/intel-pt.c
1923
ret = intel_pt_deliver_synth_event(pt, event, &sample,
tools/perf/util/intel-pt.c
1924
pt->transactions_sample_type);
tools/perf/util/intel-pt.c
1929
static void intel_pt_prep_p_sample(struct intel_pt *pt,
tools/perf/util/intel-pt.c
1934
intel_pt_prep_sample(pt, ptq, event, sample);
tools/perf/util/intel-pt.c
1946
struct intel_pt *pt = ptq->pt;
tools/perf/util/intel-pt.c
1951
if (intel_pt_skip_event(pt))
tools/perf/util/intel-pt.c
1954
intel_pt_prep_p_sample(pt, ptq, event, &sample);
tools/perf/util/intel-pt.c
1956
sample.id = ptq->pt->ptwrites_id;
tools/perf/util/intel-pt.c
1957
sample.stream_id = ptq->pt->ptwrites_id;
tools/perf/util/intel-pt.c
1966
return intel_pt_deliver_synth_event(pt, event, &sample,
tools/perf/util/intel-pt.c
1967
pt->ptwrites_sample_type);
tools/perf/util/intel-pt.c
1972
struct intel_pt *pt = ptq->pt;
tools/perf/util/intel-pt.c
1979
if (intel_pt_skip_cbr_event(pt))
tools/perf/util/intel-pt.c
1985
intel_pt_prep_p_sample(pt, ptq, event, &sample);
tools/perf/util/intel-pt.c
1987
sample.id = ptq->pt->cbr_id;
tools/perf/util/intel-pt.c
1988
sample.stream_id = ptq->pt->cbr_id;
tools/perf/util/intel-pt.c
1990
flags = (u16)ptq->state->cbr_payload | (pt->max_non_turbo_ratio << 16);
tools/perf/util/intel-pt.c
1992
raw.freq = cpu_to_le32(raw.cbr * pt->cbr2khz);
tools/perf/util/intel-pt.c
1998
ret = intel_pt_deliver_synth_event(pt, event, &sample,
tools/perf/util/intel-pt.c
1999
pt->pwr_events_sample_type);
tools/perf/util/intel-pt.c
2006
struct intel_pt *pt = ptq->pt;
tools/perf/util/intel-pt.c
2012
if (intel_pt_skip_event(pt))
tools/perf/util/intel-pt.c
2016
intel_pt_prep_p_sample(pt, ptq, event, &sample);
tools/perf/util/intel-pt.c
2018
sample.id = ptq->pt->psb_id;
tools/perf/util/intel-pt.c
2019
sample.stream_id = ptq->pt->psb_id;
tools/perf/util/intel-pt.c
2028
ret = intel_pt_deliver_synth_event(pt, event, &sample,
tools/perf/util/intel-pt.c
2029
pt->pwr_events_sample_type);
tools/perf/util/intel-pt.c
2036
struct intel_pt *pt = ptq->pt;
tools/perf/util/intel-pt.c
2042
if (intel_pt_skip_event(pt))
tools/perf/util/intel-pt.c
2046
intel_pt_prep_p_sample(pt, ptq, event, &sample);
tools/perf/util/intel-pt.c
2048
sample.id = ptq->pt->mwait_id;
tools/perf/util/intel-pt.c
2049
sample.stream_id = ptq->pt->mwait_id;
tools/perf/util/intel-pt.c
2057
ret = intel_pt_deliver_synth_event(pt, event, &sample,
tools/perf/util/intel-pt.c
2058
pt->pwr_events_sample_type);
tools/perf/util/intel-pt.c
2065
struct intel_pt *pt = ptq->pt;
tools/perf/util/intel-pt.c
2071
if (intel_pt_skip_event(pt))
tools/perf/util/intel-pt.c
2075
intel_pt_prep_p_sample(pt, ptq, event, &sample);
tools/perf/util/intel-pt.c
2077
sample.id = ptq->pt->pwre_id;
tools/perf/util/intel-pt.c
2078
sample.stream_id = ptq->pt->pwre_id;
tools/perf/util/intel-pt.c
2086
ret = intel_pt_deliver_synth_event(pt, event, &sample,
tools/perf/util/intel-pt.c
2087
pt->pwr_events_sample_type);
tools/perf/util/intel-pt.c
2094
struct intel_pt *pt = ptq->pt;
tools/perf/util/intel-pt.c
2100
if (intel_pt_skip_event(pt))
tools/perf/util/intel-pt.c
2104
intel_pt_prep_p_sample(pt, ptq, event, &sample);
tools/perf/util/intel-pt.c
2106
sample.id = ptq->pt->exstop_id;
tools/perf/util/intel-pt.c
2107
sample.stream_id = ptq->pt->exstop_id;
tools/perf/util/intel-pt.c
2115
ret = intel_pt_deliver_synth_event(pt, event, &sample,
tools/perf/util/intel-pt.c
2116
pt->pwr_events_sample_type);
tools/perf/util/intel-pt.c
2123
struct intel_pt *pt = ptq->pt;
tools/perf/util/intel-pt.c
2129
if (intel_pt_skip_event(pt))
tools/perf/util/intel-pt.c
2133
intel_pt_prep_p_sample(pt, ptq, event, &sample);
tools/perf/util/intel-pt.c
2135
sample.id = ptq->pt->pwrx_id;
tools/perf/util/intel-pt.c
2136
sample.stream_id = ptq->pt->pwrx_id;
tools/perf/util/intel-pt.c
2144
ret = intel_pt_deliver_synth_event(pt, event, &sample,
tools/perf/util/intel-pt.c
2145
pt->pwr_events_sample_type);
tools/perf/util/intel-pt.c
233
static void intel_pt_dump(struct intel_pt *pt __maybe_unused,
tools/perf/util/intel-pt.c
2421
struct intel_pt *pt = ptq->pt;
tools/perf/util/intel-pt.c
2427
if (intel_pt_skip_event(pt))
tools/perf/util/intel-pt.c
2458
else if (!pt->timeless_decoding)
tools/perf/util/intel-pt.c
2461
sample.time = tsc_to_perf_time(timestamp, &pt->tc);
tools/perf/util/intel-pt.c
2465
pt->synth_opts.callchain) {
tools/perf/util/intel-pt.c
2467
pt->synth_opts.callchain_sz, sample.ip,
tools/perf/util/intel-pt.c
2468
pt->kernel_start);
tools/perf/util/intel-pt.c
2494
} else if (pt->synth_opts.last_branch) {
tools/perf/util/intel-pt.c
2497
pt->br_stack_sz);
tools/perf/util/intel-pt.c
2560
ret = intel_pt_deliver_synth_event(pt, event, &sample, sample_type);
tools/perf/util/intel-pt.c
2567
struct intel_pt *pt = ptq->pt;
tools/perf/util/intel-pt.c
2568
struct evsel *evsel = pt->pebs_evsel;
tools/perf/util/intel-pt.c
2569
int data_src_fmt = pt->pebs_data_src_fmt;
tools/perf/util/intel-pt.c
2579
struct intel_pt *pt = ptq->pt;
tools/perf/util/intel-pt.c
2584
if (!pt->single_pebs)
tools/perf/util/intel-pt.c
2592
if (!pt->single_pebs)
tools/perf/util/intel-pt.c
2607
struct intel_pt *pt = ptq->pt;
tools/perf/util/intel-pt.c
2616
if (intel_pt_skip_event(pt))
tools/perf/util/intel-pt.c
2620
intel_pt_prep_p_sample(pt, ptq, event, &sample);
tools/perf/util/intel-pt.c
2622
sample.id = ptq->pt->evt_id;
tools/perf/util/intel-pt.c
2623
sample.stream_id = ptq->pt->evt_id;
tools/perf/util/intel-pt.c
2641
ret = intel_pt_deliver_synth_event(pt, event, &sample,
tools/perf/util/intel-pt.c
2642
pt->evt_sample_type);
tools/perf/util/intel-pt.c
2649
struct intel_pt *pt = ptq->pt;
tools/perf/util/intel-pt.c
2655
if (intel_pt_skip_event(pt))
tools/perf/util/intel-pt.c
2659
intel_pt_prep_p_sample(pt, ptq, event, &sample);
tools/perf/util/intel-pt.c
2661
sample.id = ptq->pt->iflag_chg_id;
tools/perf/util/intel-pt.c
2662
sample.stream_id = ptq->pt->iflag_chg_id;
tools/perf/util/intel-pt.c
2678
ret = intel_pt_deliver_synth_event(pt, event, &sample,
tools/perf/util/intel-pt.c
2679
pt->iflag_chg_sample_type);
tools/perf/util/intel-pt.c
2684
static int intel_pt_synth_error(struct intel_pt *pt, int code, int cpu,
tools/perf/util/intel-pt.c
2688
bool dump_log_on_error = pt->synth_opts.log_plus_flags & AUXTRACE_LOG_FLG_ON_ERROR;
tools/perf/util/intel-pt.c
2689
bool log_on_stdout = pt->synth_opts.log_plus_flags & AUXTRACE_LOG_FLG_USE_STDOUT;
tools/perf/util/intel-pt.c
2694
if (pt->synth_opts.error_minus_flags) {
tools/perf/util/intel-pt.c
2696
pt->synth_opts.error_minus_flags & AUXTRACE_ERR_FLG_OVERFLOW)
tools/perf/util/intel-pt.c
2699
pt->synth_opts.error_minus_flags & AUXTRACE_ERR_FLG_DATA_LOST)
tools/perf/util/intel-pt.c
2719
err = perf_session__deliver_synth_event(pt->session, &event, NULL);
tools/perf/util/intel-pt.c
273
static void intel_pt_dump_event(struct intel_pt *pt, unsigned char *buf,
tools/perf/util/intel-pt.c
2730
struct intel_pt *pt = ptq->pt;
tools/perf/util/intel-pt.c
2737
tm = pt->timeless_decoding ? 0 : tsc_to_perf_time(tm, &pt->tc);
tools/perf/util/intel-pt.c
2739
if (pt->have_guest_sideband && state->from_nr) {
tools/perf/util/intel-pt.c
2746
return intel_pt_synth_error(pt, state->err, ptq->cpu, pid, tid,
tools/perf/util/intel-pt.c
2750
static int intel_pt_next_tid(struct intel_pt *pt, struct intel_pt_queue *ptq)
tools/perf/util/intel-pt.c
2761
err = machine__set_current_tid(pt->machine, ptq->cpu, -1, tid);
tools/perf/util/intel-pt.c
2763
queue = &pt->queues.queue_array[ptq->queue_nr];
tools/perf/util/intel-pt.c
2764
intel_pt_set_pid_tid_cpu(pt, queue);
tools/perf/util/intel-pt.c
277
intel_pt_dump(pt, buf, len);
tools/perf/util/intel-pt.c
2773
struct intel_pt *pt = ptq->pt;
tools/perf/util/intel-pt.c
2775
return ip == pt->switch_ip &&
tools/perf/util/intel-pt.c
2787
struct intel_pt *pt = ptq->pt;
tools/perf/util/intel-pt.c
2795
if (pt->synth_opts.approx_ipc) {
tools/perf/util/intel-pt.c
2813
if (pt->sample_pebs && state->type & INTEL_PT_BLK_ITEMS) {
tools/perf/util/intel-pt.c
2819
if (pt->synth_opts.intr_events) {
tools/perf/util/intel-pt.c
2832
if (pt->sample_pwr_events) {
tools/perf/util/intel-pt.c
2868
if (pt->sample_instructions) {
tools/perf/util/intel-pt.c
2873
if (pt->sample_cycles) {
tools/perf/util/intel-pt.c
2880
if (pt->sample_transactions && (state->type & INTEL_PT_TRANSACTION)) {
tools/perf/util/intel-pt.c
2886
if (pt->sample_ptwrites && (state->type & INTEL_PT_PTW)) {
tools/perf/util/intel-pt.c
2895
if (pt->use_thread_stack) {
tools/perf/util/intel-pt.c
2898
state->trace_nr, pt->callstack,
tools/perf/util/intel-pt.c
2899
pt->br_stack_sz_plus,
tools/perf/util/intel-pt.c
2900
pt->mispred_all);
tools/perf/util/intel-pt.c
2905
if (pt->sample_branches) {
tools/perf/util/intel-pt.c
293
struct intel_pt *pt = container_of(session->auxtrace, struct intel_pt,
tools/perf/util/intel-pt.c
2939
err = intel_pt_next_tid(pt, ptq);
tools/perf/util/intel-pt.c
2953
state->to_ip == pt->ptss_ip &&
tools/perf/util/intel-pt.c
2961
static u64 intel_pt_switch_ip(struct intel_pt *pt, u64 *ptss_ip)
tools/perf/util/intel-pt.c
2963
struct machine *machine = pt->machine;
tools/perf/util/intel-pt.c
297
intel_pt_dump(pt, sample->aux_sample.data, sample->aux_sample.size);
tools/perf/util/intel-pt.c
2995
if (pt->have_sched_switch == 1)
tools/perf/util/intel-pt.c
300
static bool intel_pt_log_events(struct intel_pt *pt, u64 tm)
tools/perf/util/intel-pt.c
3013
static void intel_pt_enable_sync_switch(struct intel_pt *pt)
tools/perf/util/intel-pt.c
3017
if (pt->sync_switch_not_supported)
tools/perf/util/intel-pt.c
302
struct perf_time_interval *range = pt->synth_opts.ptime_range;
tools/perf/util/intel-pt.c
3020
pt->sync_switch = true;
tools/perf/util/intel-pt.c
3022
for (i = 0; i < pt->queues.nr_queues; i++) {
tools/perf/util/intel-pt.c
3023
struct auxtrace_queue *queue = &pt->queues.queue_array[i];
tools/perf/util/intel-pt.c
303
int n = pt->synth_opts.range_num;
tools/perf/util/intel-pt.c
3031
static void intel_pt_disable_sync_switch(struct intel_pt *pt)
tools/perf/util/intel-pt.c
3035
pt->sync_switch = false;
tools/perf/util/intel-pt.c
3037
for (i = 0; i < pt->queues.nr_queues; i++) {
tools/perf/util/intel-pt.c
3038
struct auxtrace_queue *queue = &pt->queues.queue_array[i];
tools/perf/util/intel-pt.c
3043
intel_pt_next_tid(pt, ptq);
tools/perf/util/intel-pt.c
305
if (pt->synth_opts.log_plus_flags & AUXTRACE_LOG_FLG_ALL_PERF_EVTS)
tools/perf/util/intel-pt.c
3054
struct intel_pt *pt = ptq->pt;
tools/perf/util/intel-pt.c
3059
ptq->sel_timestamp = pt->time_ranges[ptq->sel_idx].end;
tools/perf/util/intel-pt.c
3061
} else if (ptq->sel_idx + 1 < pt->range_cnt) {
tools/perf/util/intel-pt.c
3065
ptq->sel_timestamp = pt->time_ranges[ptq->sel_idx].start;
tools/perf/util/intel-pt.c
308
if (pt->synth_opts.log_minus_flags & AUXTRACE_LOG_FLG_ALL_PERF_EVTS)
tools/perf/util/intel-pt.c
3093
intel_pt_next_tid(ptq->pt, ptq);
tools/perf/util/intel-pt.c
3123
struct intel_pt *pt = ptq->pt;
tools/perf/util/intel-pt.c
3127
if (!pt->kernel_start) {
tools/perf/util/intel-pt.c
3128
pt->kernel_start = machine__kernel_start(pt->machine);
tools/perf/util/intel-pt.c
3129
if (pt->per_cpu_mmaps &&
tools/perf/util/intel-pt.c
3130
(pt->have_sched_switch == 1 || pt->have_sched_switch == 3) &&
tools/perf/util/intel-pt.c
3131
!pt->timeless_decoding && intel_pt_tracing_kernel(pt) &&
tools/perf/util/intel-pt.c
3132
!pt->sampling_mode && !pt->synth_opts.vm_time_correlation) {
tools/perf/util/intel-pt.c
3133
pt->switch_ip = intel_pt_switch_ip(pt, &pt->ptss_ip);
tools/perf/util/intel-pt.c
3134
if (pt->switch_ip) {
tools/perf/util/intel-pt.c
3136
pt->switch_ip, pt->ptss_ip);
tools/perf/util/intel-pt.c
3137
intel_pt_enable_sync_switch(pt);
tools/perf/util/intel-pt.c
3154
state->from_ip >= pt->kernel_start) {
tools/perf/util/intel-pt.c
3156
intel_pt_next_tid(pt, ptq);
tools/perf/util/intel-pt.c
3159
if (pt->synth_opts.errors) {
tools/perf/util/intel-pt.c
3172
if (pt->est_tsc &&
tools/perf/util/intel-pt.c
3173
(state->from_ip >= pt->kernel_start || !state->from_ip) &&
tools/perf/util/intel-pt.c
3174
state->to_ip && state->to_ip < pt->kernel_start) {
tools/perf/util/intel-pt.c
3196
if (!pt->timeless_decoding && ptq->timestamp >= *timestamp) {
tools/perf/util/intel-pt.c
3204
static inline int intel_pt_update_queues(struct intel_pt *pt)
tools/perf/util/intel-pt.c
3206
if (pt->queues.new_data) {
tools/perf/util/intel-pt.c
3207
pt->queues.new_data = false;
tools/perf/util/intel-pt.c
3208
return intel_pt_setup_queues(pt);
tools/perf/util/intel-pt.c
3213
static int intel_pt_process_queues(struct intel_pt *pt, u64 timestamp)
tools/perf/util/intel-pt.c
3223
if (!pt->heap.heap_cnt)
tools/perf/util/intel-pt.c
3226
if (pt->heap.heap_array[0].ordinal >= timestamp)
tools/perf/util/intel-pt.c
3229
queue_nr = pt->heap.heap_array[0].queue_nr;
tools/perf/util/intel-pt.c
3230
queue = &pt->queues.queue_array[queue_nr];
tools/perf/util/intel-pt.c
3234
queue_nr, pt->heap.heap_array[0].ordinal,
tools/perf/util/intel-pt.c
3237
auxtrace_heap__pop(&pt->heap);
tools/perf/util/intel-pt.c
3239
if (pt->heap.heap_cnt) {
tools/perf/util/intel-pt.c
3240
ts = pt->heap.heap_array[0].ordinal + 1;
tools/perf/util/intel-pt.c
3247
intel_pt_set_pid_tid_cpu(pt, queue);
tools/perf/util/intel-pt.c
3252
auxtrace_heap__add(&pt->heap, queue_nr, ts);
tools/perf/util/intel-pt.c
3257
ret = auxtrace_heap__add(&pt->heap, queue_nr, ts);
tools/perf/util/intel-pt.c
3268
static int intel_pt_process_timeless_queues(struct intel_pt *pt, pid_t tid,
tools/perf/util/intel-pt.c
3271
struct auxtrace_queues *queues = &pt->queues;
tools/perf/util/intel-pt.c
3276
struct auxtrace_queue *queue = &pt->queues.queue_array[i];
tools/perf/util/intel-pt.c
3281
intel_pt_set_pid_tid_cpu(pt, queue);
tools/perf/util/intel-pt.c
3292
struct machine *m = ptq->pt->machine;
tools/perf/util/intel-pt.c
3316
static int intel_pt_process_timeless_sample(struct intel_pt *pt,
tools/perf/util/intel-pt.c
3323
queue = auxtrace_queues__sample_queue(&pt->queues, sample, pt->session);
tools/perf/util/intel-pt.c
3338
static int intel_pt_lost(struct intel_pt *pt, struct perf_sample *sample)
tools/perf/util/intel-pt.c
3340
return intel_pt_synth_error(pt, INTEL_PT_ERR_LOST, sample->cpu,
tools/perf/util/intel-pt.c
3345
static struct intel_pt_queue *intel_pt_cpu_to_ptq(struct intel_pt *pt, int cpu)
tools/perf/util/intel-pt.c
3349
if (cpu < 0 || !pt->queues.nr_queues)
tools/perf/util/intel-pt.c
3352
if ((unsigned)cpu >= pt->queues.nr_queues)
tools/perf/util/intel-pt.c
3353
i = pt->queues.nr_queues - 1;
tools/perf/util/intel-pt.c
3357
if (pt->queues.queue_array[i].cpu == cpu)
tools/perf/util/intel-pt.c
3358
return pt->queues.queue_array[i].priv;
tools/perf/util/intel-pt.c
3361
if (pt->queues.queue_array[--i].cpu == cpu)
tools/perf/util/intel-pt.c
3362
return pt->queues.queue_array[i].priv;
tools/perf/util/intel-pt.c
3365
for (; j < pt->queues.nr_queues; j++) {
tools/perf/util/intel-pt.c
3366
if (pt->queues.queue_array[j].cpu == cpu)
tools/perf/util/intel-pt.c
3367
return pt->queues.queue_array[j].priv;
tools/perf/util/intel-pt.c
3373
static int intel_pt_sync_switch(struct intel_pt *pt, int cpu, pid_t tid,
tools/perf/util/intel-pt.c
3379
if (!pt->sync_switch)
tools/perf/util/intel-pt.c
3382
ptq = intel_pt_cpu_to_ptq(pt, cpu);
tools/perf/util/intel-pt.c
3397
&pt->tc);
tools/perf/util/intel-pt.c
3398
err = auxtrace_heap__add(&pt->heap, ptq->queue_nr,
tools/perf/util/intel-pt.c
3419
static int intel_pt_process_switch(struct intel_pt *pt,
tools/perf/util/intel-pt.c
3424
struct evsel *evsel = evlist__id2evsel(pt->session->evlist, sample->id);
tools/perf/util/intel-pt.c
3426
if (evsel != pt->switch_evsel)
tools/perf/util/intel-pt.c
3434
&pt->tc));
tools/perf/util/intel-pt.c
3436
ret = intel_pt_sync_switch(pt, cpu, tid, sample->time);
tools/perf/util/intel-pt.c
3440
return machine__set_current_tid(pt->machine, cpu, -1, tid);
tools/perf/util/intel-pt.c
3444
static int intel_pt_context_switch_in(struct intel_pt *pt,
tools/perf/util/intel-pt.c
3451
if (pt->sync_switch) {
tools/perf/util/intel-pt.c
3454
ptq = intel_pt_cpu_to_ptq(pt, cpu);
tools/perf/util/intel-pt.c
3476
if (machine__get_current_tid(pt->machine, cpu) == tid)
tools/perf/util/intel-pt.c
3479
return machine__set_current_tid(pt->machine, cpu, pid, tid);
tools/perf/util/intel-pt.c
3482
static int intel_pt_guest_context_switch(struct intel_pt *pt,
tools/perf/util/intel-pt.c
3487
struct machines *machines = &pt->session->machines;
tools/perf/util/intel-pt.c
3490
pt->have_guest_sideband = true;
tools/perf/util/intel-pt.c
3496
pt->sync_switch_not_supported = true;
tools/perf/util/intel-pt.c
3497
if (pt->sync_switch)
tools/perf/util/intel-pt.c
3498
intel_pt_disable_sync_switch(pt);
tools/perf/util/intel-pt.c
3509
static int intel_pt_context_switch(struct intel_pt *pt, union perf_event *event,
tools/perf/util/intel-pt.c
3517
return intel_pt_guest_context_switch(pt, event, sample);
tools/perf/util/intel-pt.c
3521
if (pt->have_sched_switch == 3) {
tools/perf/util/intel-pt.c
3523
return intel_pt_context_switch_in(pt, sample);
tools/perf/util/intel-pt.c
3540
ret = intel_pt_sync_switch(pt, cpu, tid, sample->time);
tools/perf/util/intel-pt.c
3544
return machine__set_current_tid(pt->machine, cpu, pid, tid);
tools/perf/util/intel-pt.c
3547
static int intel_pt_process_itrace_start(struct intel_pt *pt,
tools/perf/util/intel-pt.c
355
struct intel_pt *pt = ptq->pt;
tools/perf/util/intel-pt.c
3551
if (!pt->per_cpu_mmaps)
tools/perf/util/intel-pt.c
3557
perf_time_to_tsc(sample->time, &pt->tc));
tools/perf/util/intel-pt.c
3559
return machine__set_current_tid(pt->machine, sample->cpu,
tools/perf/util/intel-pt.c
3568
static int intel_pt_data_src_fmt(struct intel_pt *pt, struct evsel *evsel)
tools/perf/util/intel-pt.c
357
if (!vmcs && !pt->dflt_tsc_offset)
tools/perf/util/intel-pt.c
3570
struct perf_env *env = pt->machine->env;
tools/perf/util/intel-pt.c
360
return intel_pt_findnew_vmcs(&pt->vmcs_info, vmcs, pt->dflt_tsc_offset);
tools/perf/util/intel-pt.c
3607
static int intel_pt_process_aux_output_hw_id(struct intel_pt *pt,
tools/perf/util/intel-pt.c
3616
queue = auxtrace_queues__sample_queue(&pt->queues, sample, pt->session);
tools/perf/util/intel-pt.c
3617
evsel = evlist__id2evsel_strict(pt->session->evlist, sample->id);
tools/perf/util/intel-pt.c
3627
ptq->pebs[hw_id].data_src_fmt = intel_pt_data_src_fmt(pt, evsel);
tools/perf/util/intel-pt.c
363
static void intel_pt_free_vmcs_info(struct intel_pt *pt)
tools/perf/util/intel-pt.c
3644
static int intel_pt_text_poke(struct intel_pt *pt, union perf_event *event)
tools/perf/util/intel-pt.c
3650
struct thread *thread = pt->unknown_thread;
tools/perf/util/intel-pt.c
3652
struct machine *machine = pt->machine;
tools/perf/util/intel-pt.c
368
n = rb_first(&pt->vmcs_info);
tools/perf/util/intel-pt.c
3704
struct intel_pt *pt = container_of(session->auxtrace, struct intel_pt,
tools/perf/util/intel-pt.c
3718
timestamp = perf_time_to_tsc(sample->time, &pt->tc);
tools/perf/util/intel-pt.c
372
rb_erase(&v->rb_node, &pt->vmcs_info);
tools/perf/util/intel-pt.c
3722
if (timestamp || pt->timeless_decoding) {
tools/perf/util/intel-pt.c
3723
err = intel_pt_update_queues(pt);
tools/perf/util/intel-pt.c
3728
if (pt->timeless_decoding) {
tools/perf/util/intel-pt.c
3729
if (pt->sampling_mode) {
tools/perf/util/intel-pt.c
3731
err = intel_pt_process_timeless_sample(pt,
tools/perf/util/intel-pt.c
3734
err = intel_pt_process_timeless_queues(pt,
tools/perf/util/intel-pt.c
3739
if (!pt->first_timestamp)
tools/perf/util/intel-pt.c
3740
intel_pt_first_timestamp(pt, timestamp);
tools/perf/util/intel-pt.c
3741
err = intel_pt_process_queues(pt, timestamp);
tools/perf/util/intel-pt.c
3747
if (pt->synth_opts.add_callchain && !sample->callchain)
tools/perf/util/intel-pt.c
3748
intel_pt_add_callchain(pt, sample);
tools/perf/util/intel-pt.c
3749
if (pt->synth_opts.add_last_branch && !sample->branch_stack)
tools/perf/util/intel-pt.c
3750
intel_pt_add_br_stack(pt, sample);
tools/perf/util/intel-pt.c
3755
pt->synth_opts.errors) {
tools/perf/util/intel-pt.c
3756
err = intel_pt_lost(pt, sample);
tools/perf/util/intel-pt.c
3762
if (pt->switch_evsel && event->header.type == PERF_RECORD_SAMPLE)
tools/perf/util/intel-pt.c
3763
err = intel_pt_process_switch(pt, sample);
tools/perf/util/intel-pt.c
3767
err = intel_pt_process_itrace_start(pt, event, sample);
tools/perf/util/intel-pt.c
3769
err = intel_pt_process_aux_output_hw_id(pt, event, sample);
tools/perf/util/intel-pt.c
377
static int intel_pt_do_fix_overlap(struct intel_pt *pt, struct auxtrace_buffer *a,
tools/perf/util/intel-pt.c
3772
err = intel_pt_context_switch(pt, event, sample);
tools/perf/util/intel-pt.c
3775
err = intel_pt_text_poke(pt, event);
tools/perf/util/intel-pt.c
3777
if (intel_pt_enable_logging && intel_pt_log_events(pt, sample->time)) {
tools/perf/util/intel-pt.c
3788
struct intel_pt *pt = container_of(session->auxtrace, struct intel_pt,
tools/perf/util/intel-pt.c
3798
ret = intel_pt_update_queues(pt);
tools/perf/util/intel-pt.c
3802
if (pt->timeless_decoding)
tools/perf/util/intel-pt.c
3803
return intel_pt_process_timeless_queues(pt, -1,
tools/perf/util/intel-pt.c
3806
return intel_pt_process_queues(pt, MAX_TIMESTAMP);
tools/perf/util/intel-pt.c
3811
struct intel_pt *pt = container_of(session->auxtrace, struct intel_pt,
tools/perf/util/intel-pt.c
3813
struct auxtrace_queues *queues = &pt->queues;
tools/perf/util/intel-pt.c
3826
struct intel_pt *pt = container_of(session->auxtrace, struct intel_pt,
tools/perf/util/intel-pt.c
3829
auxtrace_heap__free(&pt->heap);
tools/perf/util/intel-pt.c
3832
intel_pt_free_vmcs_info(pt);
tools/perf/util/intel-pt.c
3833
thread__put(pt->unknown_thread);
tools/perf/util/intel-pt.c
3834
addr_filters__exit(&pt->filts);
tools/perf/util/intel-pt.c
3835
zfree(&pt->chain);
tools/perf/util/intel-pt.c
3836
zfree(&pt->filter);
tools/perf/util/intel-pt.c
3837
zfree(&pt->time_ranges);
tools/perf/util/intel-pt.c
3838
zfree(&pt->br_stack);
tools/perf/util/intel-pt.c
3839
free(pt);
tools/perf/util/intel-pt.c
384
pt->have_tsc, &consecutive,
tools/perf/util/intel-pt.c
3845
struct intel_pt *pt = container_of(session->auxtrace, struct intel_pt,
tools/perf/util/intel-pt.c
3848
return evsel->core.attr.type == pt->pmu_type;
tools/perf/util/intel-pt.c
385
pt->synth_opts.vm_time_correlation);
tools/perf/util/intel-pt.c
3855
struct intel_pt *pt = container_of(session->auxtrace, struct intel_pt,
tools/perf/util/intel-pt.c
3858
if (!pt->data_queued) {
tools/perf/util/intel-pt.c
3872
err = auxtrace_queues__add_event(&pt->queues, session, event,
tools/perf/util/intel-pt.c
3880
intel_pt_dump_event(pt, buffer->data,
tools/perf/util/intel-pt.c
3894
struct intel_pt *pt = container_of(session->auxtrace, struct intel_pt,
tools/perf/util/intel-pt.c
3899
return auxtrace_queues__add_event(&pt->queues, session, event,
tools/perf/util/intel-pt.c
3904
timestamp = perf_time_to_tsc(sample->time, &pt->tc);
tools/perf/util/intel-pt.c
3908
return auxtrace_queues__add_sample(&pt->queues, session, sample,
tools/perf/util/intel-pt.c
393
if (pt->synth_opts.vm_time_correlation)
tools/perf/util/intel-pt.c
3943
static struct evsel *intel_pt_evsel(struct intel_pt *pt,
tools/perf/util/intel-pt.c
3949
if (evsel->core.attr.type == pt->pmu_type && evsel->core.ids)
tools/perf/util/intel-pt.c
3956
static int intel_pt_synth_events(struct intel_pt *pt,
tools/perf/util/intel-pt.c
3960
struct evsel *evsel = intel_pt_evsel(pt, evlist);
tools/perf/util/intel-pt.c
3976
if (pt->timeless_decoding)
tools/perf/util/intel-pt.c
3980
if (!pt->per_cpu_mmaps)
tools/perf/util/intel-pt.c
3992
if (pt->synth_opts.branches) {
tools/perf/util/intel-pt.c
3999
pt->sample_branches = true;
tools/perf/util/intel-pt.c
4000
pt->branches_sample_type = attr.sample_type;
tools/perf/util/intel-pt.c
4001
pt->branches_id = id;
tools/perf/util/intel-pt.c
4006
if (pt->synth_opts.callchain)
tools/perf/util/intel-pt.c
4008
if (pt->synth_opts.last_branch) {
tools/perf/util/intel-pt.c
4018
if (pt->synth_opts.instructions) {
tools/perf/util/intel-pt.c
4020
if (pt->synth_opts.period_type == PERF_ITRACE_PERIOD_NANOSECS)
tools/perf/util/intel-pt.c
4022
intel_pt_ns_to_ticks(pt, pt->synth_opts.period);
tools/perf/util/intel-pt.c
4024
attr.sample_period = pt->synth_opts.period;
tools/perf/util/intel-pt.c
4028
pt->sample_instructions = true;
tools/perf/util/intel-pt.c
4029
pt->instructions_sample_type = attr.sample_type;
tools/perf/util/intel-pt.c
4030
pt->instructions_id = id;
tools/perf/util/intel-pt.c
4034
if (pt->synth_opts.cycles) {
tools/perf/util/intel-pt.c
4036
if (pt->synth_opts.period_type == PERF_ITRACE_PERIOD_NANOSECS)
tools/perf/util/intel-pt.c
4038
intel_pt_ns_to_ticks(pt, pt->synth_opts.period);
tools/perf/util/intel-pt.c
4040
attr.sample_period = pt->synth_opts.period;
tools/perf/util/intel-pt.c
4044
pt->sample_cycles = true;
tools/perf/util/intel-pt.c
4045
pt->cycles_sample_type = attr.sample_type;
tools/perf/util/intel-pt.c
4046
pt->cycles_id = id;
tools/perf/util/intel-pt.c
4053
if (pt->synth_opts.transactions) {
tools/perf/util/intel-pt.c
4058
pt->sample_transactions = true;
tools/perf/util/intel-pt.c
4059
pt->transactions_sample_type = attr.sample_type;
tools/perf/util/intel-pt.c
4060
pt->transactions_id = id;
tools/perf/util/intel-pt.c
4068
if (pt->synth_opts.ptwrites) {
tools/perf/util/intel-pt.c
4073
pt->sample_ptwrites = true;
tools/perf/util/intel-pt.c
4074
pt->ptwrites_sample_type = attr.sample_type;
tools/perf/util/intel-pt.c
4075
pt->ptwrites_id = id;
tools/perf/util/intel-pt.c
4080
if (pt->synth_opts.pwr_events) {
tools/perf/util/intel-pt.c
4081
pt->sample_pwr_events = true;
tools/perf/util/intel-pt.c
4082
pt->pwr_events_sample_type = attr.sample_type;
tools/perf/util/intel-pt.c
4088
pt->cbr_id = id;
tools/perf/util/intel-pt.c
4096
pt->psb_id = id;
tools/perf/util/intel-pt.c
410
int fd = perf_data__fd(ptq->pt->session->data);
tools/perf/util/intel-pt.c
4101
if (pt->synth_opts.pwr_events && (evsel->core.attr.config & INTEL_PT_CFG_PWR_EVT_EN)) {
tools/perf/util/intel-pt.c
4106
pt->mwait_id = id;
tools/perf/util/intel-pt.c
4114
pt->pwre_id = id;
tools/perf/util/intel-pt.c
4122
pt->exstop_id = id;
tools/perf/util/intel-pt.c
4130
pt->pwrx_id = id;
tools/perf/util/intel-pt.c
4135
if (pt->synth_opts.intr_events && (evsel->core.attr.config & INTEL_PT_CFG_EVT_EN)) {
tools/perf/util/intel-pt.c
4140
pt->evt_sample_type = attr.sample_type;
tools/perf/util/intel-pt.c
4141
pt->evt_id = id;
tools/perf/util/intel-pt.c
4146
if (pt->synth_opts.intr_events && pt->cap_event_trace) {
tools/perf/util/intel-pt.c
4151
pt->iflag_chg_sample_type = attr.sample_type;
tools/perf/util/intel-pt.c
4152
pt->iflag_chg_id = id;
tools/perf/util/intel-pt.c
4160
static void intel_pt_setup_pebs_events(struct intel_pt *pt)
tools/perf/util/intel-pt.c
4164
if (!pt->synth_opts.other_events)
tools/perf/util/intel-pt.c
4167
evlist__for_each_entry(pt->session->evlist, evsel) {
tools/perf/util/intel-pt.c
4169
if (pt->single_pebs) {
tools/perf/util/intel-pt.c
417
might_overlap = ptq->pt->snapshot_mode || ptq->pt->sampling_mode;
tools/perf/util/intel-pt.c
4170
pt->single_pebs = false;
tools/perf/util/intel-pt.c
4173
pt->single_pebs = true;
tools/perf/util/intel-pt.c
4174
pt->sample_pebs = true;
tools/perf/util/intel-pt.c
4175
pt->pebs_data_src_fmt = intel_pt_data_src_fmt(pt, evsel);
tools/perf/util/intel-pt.c
4176
pt->pebs_evsel = evsel;
tools/perf/util/intel-pt.c
419
intel_pt_do_fix_overlap(ptq->pt, old_buffer, buffer))
tools/perf/util/intel-pt.c
4209
struct intel_pt *pt = data;
tools/perf/util/intel-pt.c
4212
pt->mispred_all = perf_config_bool(var, value);
tools/perf/util/intel-pt.c
4215
perf_config_int(&pt->max_loops, var, value);
tools/perf/util/intel-pt.c
4221
static u64 intel_pt_tsc_start(u64 ns, struct intel_pt *pt)
tools/perf/util/intel-pt.c
4225
tsc = perf_time_to_tsc(ns, &pt->tc);
tools/perf/util/intel-pt.c
4228
tm = tsc_to_perf_time(tsc, &pt->tc);
tools/perf/util/intel-pt.c
4235
tm = tsc_to_perf_time(++tsc, &pt->tc);
tools/perf/util/intel-pt.c
4241
static u64 intel_pt_tsc_end(u64 ns, struct intel_pt *pt)
tools/perf/util/intel-pt.c
4245
tsc = perf_time_to_tsc(ns, &pt->tc);
tools/perf/util/intel-pt.c
4248
tm = tsc_to_perf_time(tsc, &pt->tc);
tools/perf/util/intel-pt.c
4255
tm = tsc_to_perf_time(--tsc, &pt->tc);
tools/perf/util/intel-pt.c
4260
static int intel_pt_setup_time_ranges(struct intel_pt *pt,
tools/perf/util/intel-pt.c
4267
if (!n || !p || pt->timeless_decoding)
tools/perf/util/intel-pt.c
4270
pt->time_ranges = calloc(n, sizeof(struct range));
tools/perf/util/intel-pt.c
4271
if (!pt->time_ranges)
tools/perf/util/intel-pt.c
4274
pt->range_cnt = n;
tools/perf/util/intel-pt.c
4279
struct range *r = &pt->time_ranges[i];
tools/perf/util/intel-pt.c
4287
r->start = ts ? intel_pt_tsc_start(ts, pt) : 0;
tools/perf/util/intel-pt.c
4288
r->end = te ? intel_pt_tsc_end(te, pt) : 0;
tools/perf/util/intel-pt.c
4299
static int intel_pt_parse_vm_tm_corr_arg(struct intel_pt *pt, char **args)
tools/perf/util/intel-pt.c
4316
pt->dflt_tsc_offset = tsc_offset;
tools/perf/util/intel-pt.c
4327
vmcs_info = intel_pt_findnew_vmcs(&pt->vmcs_info, vmcs, tsc_offset);
tools/perf/util/intel-pt.c
4339
static int intel_pt_parse_vm_tm_corr_args(struct intel_pt *pt)
tools/perf/util/intel-pt.c
4341
char *args = pt->synth_opts.vm_tm_corr_args;
tools/perf/util/intel-pt.c
4348
ret = intel_pt_parse_vm_tm_corr_arg(pt, &args);
tools/perf/util/intel-pt.c
4413
struct intel_pt *pt;
tools/perf/util/intel-pt.c
4422
pt = zalloc(sizeof(struct intel_pt));
tools/perf/util/intel-pt.c
4423
if (!pt)
tools/perf/util/intel-pt.c
4426
pt->vmcs_info = RB_ROOT;
tools/perf/util/intel-pt.c
4428
addr_filters__init(&pt->filts);
tools/perf/util/intel-pt.c
4430
err = perf_config(intel_pt_perf_config, pt);
tools/perf/util/intel-pt.c
4434
err = auxtrace_queues__init(&pt->queues);
tools/perf/util/intel-pt.c
4439
pt->synth_opts = *session->itrace_synth_opts;
tools/perf/util/intel-pt.c
4443
itrace_synth_opts__set_default(&pt->synth_opts, opts->default_no_sample);
tools/perf/util/intel-pt.c
4445
pt->synth_opts.branches = false;
tools/perf/util/intel-pt.c
4446
pt->synth_opts.callchain = true;
tools/perf/util/intel-pt.c
4447
pt->synth_opts.add_callchain = true;
tools/perf/util/intel-pt.c
4449
pt->synth_opts.thread_stack = opts->thread_stack;
tools/perf/util/intel-pt.c
4452
if (!(pt->synth_opts.log_plus_flags & AUXTRACE_LOG_FLG_USE_STDOUT))
tools/perf/util/intel-pt.c
4455
pt->session = session;
tools/perf/util/intel-pt.c
4456
pt->machine = &session->machines.host; /* No kvm support */
tools/perf/util/intel-pt.c
4457
pt->auxtrace_type = auxtrace_info->type;
tools/perf/util/intel-pt.c
4458
pt->pmu_type = auxtrace_info->priv[INTEL_PT_PMU_TYPE];
tools/perf/util/intel-pt.c
4459
pt->tc.time_shift = auxtrace_info->priv[INTEL_PT_TIME_SHIFT];
tools/perf/util/intel-pt.c
4460
pt->tc.time_mult = auxtrace_info->priv[INTEL_PT_TIME_MULT];
tools/perf/util/intel-pt.c
4461
pt->tc.time_zero = auxtrace_info->priv[INTEL_PT_TIME_ZERO];
tools/perf/util/intel-pt.c
4462
pt->cap_user_time_zero = auxtrace_info->priv[INTEL_PT_CAP_USER_TIME_ZERO];
tools/perf/util/intel-pt.c
4463
pt->tsc_bit = auxtrace_info->priv[INTEL_PT_TSC_BIT];
tools/perf/util/intel-pt.c
4464
pt->noretcomp_bit = auxtrace_info->priv[INTEL_PT_NORETCOMP_BIT];
tools/perf/util/intel-pt.c
4465
pt->have_sched_switch = auxtrace_info->priv[INTEL_PT_HAVE_SCHED_SWITCH];
tools/perf/util/intel-pt.c
4466
pt->snapshot_mode = auxtrace_info->priv[INTEL_PT_SNAPSHOT_MODE];
tools/perf/util/intel-pt.c
4467
pt->per_cpu_mmaps = auxtrace_info->priv[INTEL_PT_PER_CPU_MMAPS];
tools/perf/util/intel-pt.c
4472
pt->mtc_bit = auxtrace_info->priv[INTEL_PT_MTC_BIT];
tools/perf/util/intel-pt.c
4473
pt->mtc_freq_bits = auxtrace_info->priv[INTEL_PT_MTC_FREQ_BITS];
tools/perf/util/intel-pt.c
4474
pt->tsc_ctc_ratio_n = auxtrace_info->priv[INTEL_PT_TSC_CTC_N];
tools/perf/util/intel-pt.c
4475
pt->tsc_ctc_ratio_d = auxtrace_info->priv[INTEL_PT_TSC_CTC_D];
tools/perf/util/intel-pt.c
4476
pt->cyc_bit = auxtrace_info->priv[INTEL_PT_CYC_BIT];
tools/perf/util/intel-pt.c
4482
pt->max_non_turbo_ratio =
tools/perf/util/intel-pt.c
4509
pt->filter = memdup(filter, len);
tools/perf/util/intel-pt.c
4510
if (!pt->filter) {
tools/perf/util/intel-pt.c
4515
mem_bswap_64(pt->filter, len);
tools/perf/util/intel-pt.c
4516
if (pt->filter[len - 1]) {
tools/perf/util/intel-pt.c
4521
err = addr_filters__parse_bare_filter(&pt->filts,
tools/perf/util/intel-pt.c
4526
intel_pt_print_info_str("Filter string", pt->filter);
tools/perf/util/intel-pt.c
4530
pt->cap_event_trace = *info++;
tools/perf/util/intel-pt.c
4533
pt->cap_event_trace);
tools/perf/util/intel-pt.c
4536
pt->timeless_decoding = intel_pt_timeless_decoding(pt);
tools/perf/util/intel-pt.c
4537
if (pt->timeless_decoding && !pt->tc.time_mult)
tools/perf/util/intel-pt.c
4538
pt->tc.time_mult = 1;
tools/perf/util/intel-pt.c
4539
pt->have_tsc = intel_pt_have_tsc(pt);
tools/perf/util/intel-pt.c
4540
pt->sampling_mode = intel_pt_sampling_mode(pt);
tools/perf/util/intel-pt.c
4541
pt->est_tsc = !pt->timeless_decoding;
tools/perf/util/intel-pt.c
4543
if (pt->synth_opts.vm_time_correlation) {
tools/perf/util/intel-pt.c
4544
if (pt->timeless_decoding) {
tools/perf/util/intel-pt.c
4555
if (!intel_pt_have_mtc(pt)) {
tools/perf/util/intel-pt.c
4560
err = intel_pt_parse_vm_tm_corr_args(pt);
tools/perf/util/intel-pt.c
4565
pt->unknown_thread = thread__new(999999999, 999999999);
tools/perf/util/intel-pt.c
4566
if (!pt->unknown_thread) {
tools/perf/util/intel-pt.c
4571
err = thread__set_comm(pt->unknown_thread, "unknown", 0);
tools/perf/util/intel-pt.c
4574
if (thread__init_maps(pt->unknown_thread, pt->machine)) {
tools/perf/util/intel-pt.c
4579
pt->auxtrace.process_event = intel_pt_process_event;
tools/perf/util/intel-pt.c
4580
pt->auxtrace.process_auxtrace_event = intel_pt_process_auxtrace_event;
tools/perf/util/intel-pt.c
4581
pt->auxtrace.queue_data = intel_pt_queue_data;
tools/perf/util/intel-pt.c
4582
pt->auxtrace.dump_auxtrace_sample = intel_pt_dump_sample;
tools/perf/util/intel-pt.c
4583
pt->auxtrace.flush_events = intel_pt_flush;
tools/perf/util/intel-pt.c
4584
pt->auxtrace.free_events = intel_pt_free_events;
tools/perf/util/intel-pt.c
4585
pt->auxtrace.free = intel_pt_free;
tools/perf/util/intel-pt.c
4586
pt->auxtrace.evsel_is_auxtrace = intel_pt_evsel_is_auxtrace;
tools/perf/util/intel-pt.c
4587
session->auxtrace = &pt->auxtrace;
tools/perf/util/intel-pt.c
4592
if (pt->have_sched_switch == 1) {
tools/perf/util/intel-pt.c
4593
pt->switch_evsel = intel_pt_find_sched_switch(session->evlist);
tools/perf/util/intel-pt.c
4594
if (!pt->switch_evsel) {
tools/perf/util/intel-pt.c
4599
} else if (pt->have_sched_switch == 2 &&
tools/perf/util/intel-pt.c
4606
if (pt->synth_opts.log) {
tools/perf/util/intel-pt.c
4607
bool log_on_error = pt->synth_opts.log_plus_flags & AUXTRACE_LOG_FLG_ON_ERROR;
tools/perf/util/intel-pt.c
4608
unsigned int log_on_error_size = pt->synth_opts.log_on_error_size;
tools/perf/util/intel-pt.c
461
queue = &ptq->pt->queues.queue_array[ptq->queue_nr];
tools/perf/util/intel-pt.c
4614
if (pt->tc.time_mult) {
tools/perf/util/intel-pt.c
4615
u64 tsc_freq = intel_pt_ns_to_ticks(pt, 1000000000);
tools/perf/util/intel-pt.c
4617
if (!pt->max_non_turbo_ratio)
tools/perf/util/intel-pt.c
4618
pt->max_non_turbo_ratio =
tools/perf/util/intel-pt.c
4622
pt->max_non_turbo_ratio);
tools/perf/util/intel-pt.c
4623
pt->cbr2khz = tsc_freq / pt->max_non_turbo_ratio / 1000;
tools/perf/util/intel-pt.c
4626
err = intel_pt_setup_time_ranges(pt, session->itrace_synth_opts);
tools/perf/util/intel-pt.c
4630
if (pt->synth_opts.calls)
tools/perf/util/intel-pt.c
4631
pt->branches_filter |= PERF_IP_FLAG_CALL | PERF_IP_FLAG_ASYNC |
tools/perf/util/intel-pt.c
4633
if (pt->synth_opts.returns)
tools/perf/util/intel-pt.c
4634
pt->branches_filter |= PERF_IP_FLAG_RETURN |
tools/perf/util/intel-pt.c
4637
if ((pt->synth_opts.callchain || pt->synth_opts.add_callchain) &&
tools/perf/util/intel-pt.c
4642
pt->synth_opts.callchain = false;
tools/perf/util/intel-pt.c
4643
pt->synth_opts.add_callchain = false;
tools/perf/util/intel-pt.c
4647
if (pt->synth_opts.add_callchain) {
tools/perf/util/intel-pt.c
4648
err = intel_pt_callchain_init(pt);
tools/perf/util/intel-pt.c
4653
if (pt->synth_opts.last_branch || pt->synth_opts.add_last_branch) {
tools/perf/util/intel-pt.c
4654
pt->br_stack_sz = pt->synth_opts.last_branch_sz;
tools/perf/util/intel-pt.c
4655
pt->br_stack_sz_plus = pt->br_stack_sz;
tools/perf/util/intel-pt.c
4658
if (pt->synth_opts.add_last_branch) {
tools/perf/util/intel-pt.c
4659
err = intel_pt_br_stack_init(pt);
tools/perf/util/intel-pt.c
4669
if (intel_pt_tracing_kernel(pt))
tools/perf/util/intel-pt.c
4670
pt->br_stack_sz_plus += 1024;
tools/perf/util/intel-pt.c
4672
pt->br_stack_sz_plus += 1;
tools/perf/util/intel-pt.c
4675
pt->use_thread_stack = pt->synth_opts.callchain ||
tools/perf/util/intel-pt.c
4676
pt->synth_opts.add_callchain ||
tools/perf/util/intel-pt.c
4677
pt->synth_opts.thread_stack ||
tools/perf/util/intel-pt.c
4678
pt->synth_opts.last_branch ||
tools/perf/util/intel-pt.c
4679
pt->synth_opts.add_last_branch;
tools/perf/util/intel-pt.c
4681
pt->callstack = pt->synth_opts.callchain ||
tools/perf/util/intel-pt.c
4682
pt->synth_opts.add_callchain ||
tools/perf/util/intel-pt.c
4683
pt->synth_opts.thread_stack;
tools/perf/util/intel-pt.c
4685
err = intel_pt_synth_events(pt, session);
tools/perf/util/intel-pt.c
4689
intel_pt_setup_pebs_events(pt);
tools/perf/util/intel-pt.c
4697
if (pt->sampling_mode || list_empty(&session->auxtrace_index))
tools/perf/util/intel-pt.c
4700
err = auxtrace_queues__process_index(&pt->queues, session);
tools/perf/util/intel-pt.c
4704
if (pt->queues.populated)
tools/perf/util/intel-pt.c
4705
pt->data_queued = true;
tools/perf/util/intel-pt.c
4707
if (pt->timeless_decoding)
tools/perf/util/intel-pt.c
4713
zfree(&pt->chain);
tools/perf/util/intel-pt.c
4714
thread__zput(pt->unknown_thread);
tools/perf/util/intel-pt.c
4717
auxtrace_queues__free(&pt->queues);
tools/perf/util/intel-pt.c
4720
addr_filters__exit(&pt->filts);
tools/perf/util/intel-pt.c
4721
zfree(&pt->filter);
tools/perf/util/intel-pt.c
4722
zfree(&pt->time_ranges);
tools/perf/util/intel-pt.c
4723
free(pt);
tools/perf/util/intel-pt.c
511
queue = &ptq->pt->queues.queue_array[ptq->queue_nr];
tools/perf/util/intel-pt.c
683
return ip >= ptq->pt->kernel_start ?
tools/perf/util/intel-pt.c
698
struct machines *machines = &ptq->pt->session->machines;
tools/perf/util/intel-pt.c
754
struct machine *machine = ptq->pt->machine;
tools/perf/util/intel-pt.c
778
if (ptq->pt->have_guest_sideband) {
tools/perf/util/intel-pt.c
808
thread = ptq->pt->unknown_thread;
tools/perf/util/intel-pt.c
947
static bool intel_pt_match_pgd_ip(struct intel_pt *pt, uint64_t ip,
tools/perf/util/intel-pt.c
955
list_for_each_entry(filt, &pt->filts.head, list) {
tools/perf/util/intel-pt.c
996
return intel_pt_match_pgd_ip(ptq->pt, ip, ip, NULL);
tools/perf/util/intel-pt.c
999
} else if (ip >= ptq->pt->kernel_start) {
tools/perf/util/strfilter.c
252
int pt = node->r ? 2 : 0; /* don't need to check node->l */
tools/perf/util/strfilter.c
254
if (buf && pt)
tools/perf/util/strfilter.c
259
if (buf && pt)
tools/perf/util/strfilter.c
261
return len + pt;
tools/testing/selftests/bpf/prog_tests/uprobe_multi_test.c
1008
pthread_t pt[CONSUMER_MAX];
tools/testing/selftests/bpf/prog_tests/uprobe_multi_test.c
1052
err = pthread_create(&pt[idx], NULL, consumer_thread, (void *) idx);
tools/testing/selftests/bpf/prog_tests/uprobe_multi_test.c
1058
pthread_join(pt[--idx], NULL);
tools/testing/selftests/bpf/progs/kfunc_call_fail.c
101
bpf_kfunc_call_test_release(pt);
tools/testing/selftests/bpf/progs/kfunc_call_fail.c
111
struct prog_test_ref_kfunc *pt;
tools/testing/selftests/bpf/progs/kfunc_call_fail.c
116
pt = bpf_kfunc_call_test_acquire(&s);
tools/testing/selftests/bpf/progs/kfunc_call_fail.c
117
if (pt) {
tools/testing/selftests/bpf/progs/kfunc_call_fail.c
118
p = bpf_kfunc_call_test_get_rdonly_mem(pt, not_const_size); /* non const size, -EINVAL */
tools/testing/selftests/bpf/progs/kfunc_call_fail.c
124
bpf_kfunc_call_test_release(pt);
tools/testing/selftests/bpf/progs/kfunc_call_fail.c
132
struct prog_test_ref_kfunc *pt;
tools/testing/selftests/bpf/progs/kfunc_call_fail.c
137
pt = bpf_kfunc_call_test_acquire(&s);
tools/testing/selftests/bpf/progs/kfunc_call_fail.c
138
if (pt) {
tools/testing/selftests/bpf/progs/kfunc_call_fail.c
140
p = bpf_kfunc_call_test_acq_rdonly_mem(pt, 2 * sizeof(int));
tools/testing/selftests/bpf/progs/kfunc_call_fail.c
148
bpf_kfunc_call_test_release(pt);
tools/testing/selftests/bpf/progs/kfunc_call_fail.c
41
struct prog_test_ref_kfunc *pt;
tools/testing/selftests/bpf/progs/kfunc_call_fail.c
46
pt = bpf_kfunc_call_test_acquire(&s);
tools/testing/selftests/bpf/progs/kfunc_call_fail.c
47
if (pt) {
tools/testing/selftests/bpf/progs/kfunc_call_fail.c
48
p = bpf_kfunc_call_test_get_rdonly_mem(pt, 2 * sizeof(int));
tools/testing/selftests/bpf/progs/kfunc_call_fail.c
54
bpf_kfunc_call_test_release(pt);
tools/testing/selftests/bpf/progs/kfunc_call_fail.c
62
struct prog_test_ref_kfunc *pt;
tools/testing/selftests/bpf/progs/kfunc_call_fail.c
67
pt = bpf_kfunc_call_test_acquire(&s);
tools/testing/selftests/bpf/progs/kfunc_call_fail.c
68
if (pt) {
tools/testing/selftests/bpf/progs/kfunc_call_fail.c
69
p = bpf_kfunc_call_test_get_rdwr_mem(pt, 2 * sizeof(int));
tools/testing/selftests/bpf/progs/kfunc_call_fail.c
77
bpf_kfunc_call_test_release(pt);
tools/testing/selftests/bpf/progs/kfunc_call_fail.c
88
struct prog_test_ref_kfunc *pt;
tools/testing/selftests/bpf/progs/kfunc_call_fail.c
93
pt = bpf_kfunc_call_test_acquire(&s);
tools/testing/selftests/bpf/progs/kfunc_call_fail.c
94
if (pt) {
tools/testing/selftests/bpf/progs/kfunc_call_fail.c
95
p = bpf_kfunc_call_test_get_rdonly_mem(pt, 2 * sizeof(int));
tools/testing/selftests/bpf/progs/kfunc_call_test.c
142
struct prog_test_ref_kfunc *pt;
tools/testing/selftests/bpf/progs/kfunc_call_test.c
147
pt = bpf_kfunc_call_test_acquire(&s);
tools/testing/selftests/bpf/progs/kfunc_call_test.c
148
if (pt) {
tools/testing/selftests/bpf/progs/kfunc_call_test.c
149
p = bpf_kfunc_call_test_get_rdwr_mem(pt, 2 * sizeof(int));
tools/testing/selftests/bpf/progs/kfunc_call_test.c
158
p = bpf_kfunc_call_test_get_rdonly_mem(pt, 2 * sizeof(int));
tools/testing/selftests/bpf/progs/kfunc_call_test.c
165
bpf_kfunc_call_test_release(pt);
tools/testing/selftests/bpf/progs/kfunc_call_test.c
63
struct prog_test_ref_kfunc *pt;
tools/testing/selftests/bpf/progs/kfunc_call_test.c
67
pt = bpf_kfunc_call_test_acquire(&s);
tools/testing/selftests/bpf/progs/kfunc_call_test.c
68
if (pt) {
tools/testing/selftests/bpf/progs/kfunc_call_test.c
69
if (pt->a != 42 || pt->b != 108)
tools/testing/selftests/bpf/progs/kfunc_call_test.c
71
bpf_kfunc_call_test_release(pt);
tools/testing/selftests/kvm/x86/sev_migrate_tests.c
121
pthread_t pt[NR_LOCK_TESTING_THREADS];
tools/testing/selftests/kvm/x86/sev_migrate_tests.c
133
pthread_create(&pt[i], NULL, locking_test_thread, &input[i]);
tools/testing/selftests/kvm/x86/sev_migrate_tests.c
136
pthread_join(pt[i], NULL);
tools/testing/selftests/ptrace/vmaccess.c
30
pthread_t pt;
tools/testing/selftests/ptrace/vmaccess.c
32
pthread_create(&pt, NULL, thread, NULL);
tools/testing/selftests/ptrace/vmaccess.c
33
pthread_join(pt, NULL);
tools/testing/selftests/ptrace/vmaccess.c
51
pthread_t pt;
tools/testing/selftests/ptrace/vmaccess.c
53
pthread_create(&pt, NULL, thread, NULL);
tools/testing/selftests/ptrace/vmaccess.c
54
pthread_join(pt, NULL);
tools/testing/selftests/vDSO/parse_vdso.c
111
ELF(Phdr) *pt = (ELF(Phdr)*)(vdso_info.load_addr + hdr->e_phoff);
tools/testing/selftests/vDSO/parse_vdso.c
120
if (pt[i].p_type == PT_LOAD && !found_vaddr) {
tools/testing/selftests/vDSO/parse_vdso.c
123
+ (uintptr_t)pt[i].p_offset
tools/testing/selftests/vDSO/parse_vdso.c
124
- (uintptr_t)pt[i].p_vaddr;
tools/testing/selftests/vDSO/parse_vdso.c
125
} else if (pt[i].p_type == PT_DYNAMIC) {
tools/testing/selftests/vDSO/parse_vdso.c
126
dyn = (ELF(Dyn)*)(base + pt[i].p_offset);
virt/kvm/eventfd.c
296
poll_table pt;
virt/kvm/eventfd.c
301
poll_table *pt)
virt/kvm/eventfd.c
303
struct kvm_irqfd_pt *p = container_of(pt, struct kvm_irqfd_pt, pt);
virt/kvm/eventfd.c
478
init_poll_funcptr(&irqfd_pt.pt, kvm_irqfd_register);
virt/kvm/eventfd.c
480
events = vfs_poll(fd_file(f), &irqfd_pt.pt);