Symbol: pr
arch/arm/mach-ux500/pm.c
87
u32 pr; /* Pending register */
arch/arm/mach-ux500/pm.c
94
pr = readl_relaxed(dist_base + GIC_DIST_PENDING_SET + i * 4);
arch/arm/mach-ux500/pm.c
97
if (pr & er)
arch/arm64/include/asm/kvm_nested.h
377
bool pr;
arch/arm64/kvm/at.c
1039
(wr)->pr = (r); \
arch/arm64/kvm/at.c
1189
wr->pr &= pov_perms & POE_R;
arch/arm64/kvm/at.c
1247
wr->pr &= !pan;
arch/arm64/kvm/at.c
1289
perm_fail = !wr.pr;
arch/arm64/kvm/at.c
1601
wr->pr = wr->pw = wr->px = true;
arch/arm64/kvm/at.c
956
wr->pr = wr->pw = true;
arch/arm64/kvm/at.c
960
wr->pr = wr->pw = wr->ur = wr->uw = true;
arch/arm64/kvm/at.c
963
wr->pr = true;
arch/arm64/kvm/at.c
967
wr->pr = wr->ur = true;
arch/arm64/kvm/at.c
979
wr->pr = wr->pw = true;
arch/arm64/kvm/at.c
981
wr->pr = true;
arch/arm64/kvm/nested.c
1477
if (vt->wr.pw && vt->wr.pr)
arch/arm64/kvm/nested.c
1479
else if (vt->wr.pr)
arch/mips/alchemy/common/clock.c
410
long tdv, tpr, pr, nr, br, bpr, diff, lastdiff;
arch/mips/alchemy/common/clock.c
436
pr = clk_hw_get_rate(pc);
arch/mips/alchemy/common/clock.c
437
if (pr < req->rate)
arch/mips/alchemy/common/clock.c
441
tdv = alchemy_calc_div(req->rate, pr, scale, maxdiv, NULL);
arch/mips/alchemy/common/clock.c
442
nr = pr / tdv;
arch/mips/alchemy/common/clock.c
449
bpr = pr;
arch/mips/alchemy/common/clock.c
466
pr = clk_hw_round_rate(free, tpr);
arch/mips/alchemy/common/clock.c
468
tdv = alchemy_calc_div(req->rate, pr, scale, maxdiv,
arch/mips/alchemy/common/clock.c
470
nr = pr / tdv;
arch/mips/alchemy/common/clock.c
476
bpr = pr;
arch/mips/include/asm/octeon/cvmx-wqe.h
186
uint64_t pr:4;
arch/mips/include/asm/octeon/cvmx-wqe.h
289
uint64_t pr:4;
arch/mips/include/asm/octeon/cvmx-wqe.h
392
uint64_t pr:4;
arch/mips/include/asm/octeon/cvmx-wqe.h
75
uint64_t pr:4;
arch/mips/kernel/pm-cps.c
189
struct uasm_reloc **pr,
arch/mips/kernel/pm-cps.c
228
uasm_il_bne(pp, pr, GPR_T0, GPR_T1, lbl);
arch/mips/kernel/pm-cps.c
233
struct uasm_reloc **pr,
arch/mips/kernel/pm-cps.c
313
uasm_il_beqz(pp, pr, GPR_T1, lbl);
arch/mips/kernel/pm-cps.c
326
struct uasm_reloc **pr,
arch/mips/kernel/pm-cps.c
334
uasm_il_beqz(pp, pr, GPR_T1, lbl);
arch/powerpc/include/asm/kvm_types.h
6
#define KVM_SUB_MODULES kvm-pr,kvm-hv
arch/powerpc/include/asm/kvm_types.h
8
#define KVM_SUB_MODULES kvm-pr
arch/powerpc/include/asm/ppc-opcode.h
451
#define PPC_RAW_PLXVP_P(xtp, i, a, pr) (PPC_PREFIX_8LS | __PPC_PRFX_R(pr) | IMM_D0(i))
arch/powerpc/include/asm/ppc-opcode.h
452
#define PPC_RAW_PLXVP_S(xtp, i, a, pr) (0xe8000000 | __PPC_XTP(xtp) | ___PPC_RA(a) | IMM_D1(i))
arch/powerpc/include/asm/ppc-opcode.h
453
#define PPC_RAW_PSTXVP_P(xsp, i, a, pr) (PPC_PREFIX_8LS | __PPC_PRFX_R(pr) | IMM_D0(i))
arch/powerpc/include/asm/ppc-opcode.h
454
#define PPC_RAW_PSTXVP_S(xsp, i, a, pr) (0xf8000000 | __PPC_XSP(xsp) | ___PPC_RA(a) | IMM_D1(i))
arch/powerpc/kernel/pci-common.c
1223
struct resource *res, *pr;
arch/powerpc/kernel/pci-common.c
1237
pr = (res->flags & IORESOURCE_IO) ?
arch/powerpc/kernel/pci-common.c
1240
pr = pci_find_parent_resource(bus->self, res);
arch/powerpc/kernel/pci-common.c
1241
if (pr == res) {
arch/powerpc/kernel/pci-common.c
1252
i, res, pr, (pr && pr->name) ? pr->name : "nil");
arch/powerpc/kernel/pci-common.c
1254
if (pr && !(pr->flags & IORESOURCE_UNSET)) {
arch/powerpc/kernel/pci-common.c
1257
if (request_resource(pr, res) == 0)
arch/powerpc/kernel/pci-common.c
1264
if (reparent_resources(pr, res) == 0)
arch/powerpc/kernel/pci-common.c
1292
struct resource *pr, *r = &dev->resource[idx];
arch/powerpc/kernel/pci-common.c
1297
pr = pci_find_parent_resource(dev, r);
arch/powerpc/kernel/pci-common.c
1298
if (!pr || (pr->flags & IORESOURCE_UNSET) ||
arch/powerpc/kernel/pci-common.c
1299
request_resource(pr, r) < 0) {
arch/powerpc/kernel/pci-common.c
1302
if (pr)
arch/powerpc/kernel/pci-common.c
1303
pr_debug("PCI: parent is %p: %pR\n", pr, pr);
arch/powerpc/kvm/e500.c
161
int as, int pid, int pr)
arch/powerpc/kvm/e500.c
167
BUG_ON(pr >= 2);
arch/powerpc/kvm/e500.c
169
idt->id[as][pid][pr].val = 0;
arch/powerpc/kvm/e500.c
170
idt->id[as][pid][pr].pentry = NULL;
arch/powerpc/kvm/e500.c
187
unsigned int pr, int avoid_recursion)
arch/powerpc/kvm/e500.c
194
BUG_ON(pr >= 2);
arch/powerpc/kvm/e500.c
196
sid = local_sid_lookup(&idt->id[as][gid][pr]);
arch/powerpc/kvm/e500.c
200
sid = local_sid_setup_one(&idt->id[as][gid][pr]);
arch/powerpc/kvm/e500.c
236
unsigned int pr, tid, ts;
arch/powerpc/kvm/e500.c
247
for (pr = 0; pr < 2; pr++) {
arch/powerpc/kvm/e500.c
257
pid = local_sid_lookup(&idt->id[ts][tid][pr]);
arch/powerpc/kvm/e500.c
259
kvmppc_e500_id_table_reset_one(vcpu_e500, ts, tid, pr);
arch/powerpc/kvm/e500.h
144
unsigned int pr, int avoid_recursion);
arch/powerpc/kvm/e500_mmu_host.c
308
u32 pr = vcpu->arch.shared->msr & MSR_PR;
arch/powerpc/kvm/e500_mmu_host.c
317
e500_shadow_mas3_attrib(gtlbe->mas7_3, writable, pr);
arch/powerpc/kvm/e500_mmu_host.c
595
bool pr;
arch/powerpc/kvm/e500_mmu_host.c
631
pr = vcpu->arch.shared->msr & MSR_PR;
arch/powerpc/kvm/e500_mmu_host.c
632
if (unlikely((pr && !(mas3 & MAS3_UX)) ||
arch/powerpc/kvm/e500_mmu_host.c
633
(!pr && !(mas3 & MAS3_SX)))) {
arch/powerpc/lib/test_emulate_step.c
24
#define TEST_PLD(r, base, i, pr) \
arch/powerpc/lib/test_emulate_step.c
25
ppc_inst_prefix(PPC_PREFIX_8LS | __PPC_PRFX_R(pr) | IMM_H(i), \
arch/powerpc/lib/test_emulate_step.c
28
#define TEST_PLWZ(r, base, i, pr) \
arch/powerpc/lib/test_emulate_step.c
29
ppc_inst_prefix(PPC_PREFIX_MLS | __PPC_PRFX_R(pr) | IMM_H(i), \
arch/powerpc/lib/test_emulate_step.c
32
#define TEST_PSTD(r, base, i, pr) \
arch/powerpc/lib/test_emulate_step.c
33
ppc_inst_prefix(PPC_PREFIX_8LS | __PPC_PRFX_R(pr) | IMM_H(i), \
arch/powerpc/lib/test_emulate_step.c
36
#define TEST_PLFS(r, base, i, pr) \
arch/powerpc/lib/test_emulate_step.c
37
ppc_inst_prefix(PPC_PREFIX_MLS | __PPC_PRFX_R(pr) | IMM_H(i), \
arch/powerpc/lib/test_emulate_step.c
40
#define TEST_PSTFS(r, base, i, pr) \
arch/powerpc/lib/test_emulate_step.c
41
ppc_inst_prefix(PPC_PREFIX_MLS | __PPC_PRFX_R(pr) | IMM_H(i), \
arch/powerpc/lib/test_emulate_step.c
44
#define TEST_PLFD(r, base, i, pr) \
arch/powerpc/lib/test_emulate_step.c
45
ppc_inst_prefix(PPC_PREFIX_MLS | __PPC_PRFX_R(pr) | IMM_H(i), \
arch/powerpc/lib/test_emulate_step.c
48
#define TEST_PSTFD(r, base, i, pr) \
arch/powerpc/lib/test_emulate_step.c
49
ppc_inst_prefix(PPC_PREFIX_MLS | __PPC_PRFX_R(pr) | IMM_H(i), \
arch/powerpc/lib/test_emulate_step.c
52
#define TEST_PADDI(t, a, i, pr) \
arch/powerpc/lib/test_emulate_step.c
53
ppc_inst_prefix(PPC_PREFIX_MLS | __PPC_PRFX_R(pr) | IMM_H(i), \
arch/powerpc/platforms/44x/uic.c
126
u32 tr, pr, mask;
arch/powerpc/platforms/44x/uic.c
153
pr = mfdcr(uic->dcrbase + UIC_PR);
arch/powerpc/platforms/44x/uic.c
155
pr = (pr & mask) | (polarity << (31-src));
arch/powerpc/platforms/44x/uic.c
157
mtdcr(uic->dcrbase + UIC_PR, pr);
arch/s390/include/asm/nmi.h
78
u64 pr : 1; /* 42 tod programmable register validity */
arch/s390/kvm/dat.c
262
init.s.pr = old.s.fc1.pr;
arch/s390/kvm/dat.c
629
if (!crste.h.fc || !crste.s.fc1.pr)
arch/s390/kvm/dat.c
769
if (!crste.h.fc || !crste.s.fc1.pr)
arch/s390/kvm/dat.c
797
if (ptep->s.pr)
arch/s390/kvm/dat.c
810
if (!crstep->h.fc || !crstep->s.fc1.pr)
arch/s390/kvm/dat.h
173
unsigned long pr : 1; /* Present */
arch/s390/kvm/dat.h
196
unsigned long pr : 1; /* Present */
arch/s390/kvm/dat.h
267
unsigned long pr : 1; /* Present */
arch/s390/kvm/dat.h
279
unsigned long pr : 1; /* Must be 0 */
arch/s390/kvm/dat.h
399
res.s.pr = 1;
arch/s390/kvm/dat.h
439
res.s.fc1.pr = 1;
arch/s390/kvm/dat.h
760
return pte.h.i && !pte.tok.pr && pte.tok.type != _DAT_TOKEN_NONE;
arch/s390/kvm/dat.h
765
return crste.h.i && !crste.tok.pr && crste.tok.type != _DAT_TOKEN_NONE;
arch/s390/kvm/dat.h
772
return crste.h.i && !crste.tok.pr && crste.tok.type == _DAT_TOKEN_NONE;
arch/s390/kvm/dat.h
83
unsigned long pr: 1; /* Present */
arch/s390/kvm/dat.h
98
unsigned long pr : 1; /* Must be 0 */
arch/s390/kvm/gmap.c
1031
pte = ptep->s.pr ? *ptep : _pte(pfn, wr, false, false);
arch/s390/kvm/gmap.c
290
if (!pte.s.pr || (!pte.s.y && pte.h.i))
arch/s390/kvm/gmap.c
379
if (ptep->s.pr && pgste.usage == PGSTE_GPS_USAGE_UNUSED) {
arch/s390/kvm/gmap.c
383
if (ptep->s.pr && test_bit(GMAP_FLAG_EXPORT_ON_UNMAP, &priv->gmap->flags))
arch/s390/kvm/gmap.c
402
if (old.s.fc1.pr && test_bit(GMAP_FLAG_EXPORT_ON_UNMAP, &priv->gmap->flags))
arch/s390/kvm/gmap.c
446
if (!pte.s.pr || (pte.h.p && !pte.s.sd))
arch/s390/kvm/gmap.c
566
if (!oldpte.s.pr || (f->write_attempt && !oldpte.s.w))
arch/s390/kvm/gmap.c
926
if (!ptep->s.pr)
arch/s390/kvm/gmap.c
938
if (!crstep->h.fc || !crstep->s.fc1.pr)
arch/s390/kvm/priv.c
1195
if (!ptep || ptep->s.pr)
arch/s390/mm/dump_pagetables.c
57
static void print_prot(struct seq_file *m, unsigned int pr, int level)
arch/s390/mm/dump_pagetables.c
63
if (pr & _PAGE_INVALID) {
arch/s390/mm/dump_pagetables.c
67
pt_dump_seq_puts(m, (pr & _PAGE_PROTECT) ? "RO " : "RW ");
arch/s390/mm/dump_pagetables.c
68
pt_dump_seq_puts(m, (pr & _PAGE_NOEXEC) ? "NX\n" : "X\n");
arch/sh/include/asm/kexec.h
58
__asm__ __volatile__ ("sts pr, %0" : "=r" (newregs->pr));
arch/sh/include/uapi/asm/ptrace_32.h
48
unsigned long pr;
arch/sh/kernel/cpu/sh2a/fpu.c
419
regs->pr = regs->pc + 4;
arch/sh/kernel/cpu/sh2a/fpu.c
444
nextpc = regs->pr;
arch/sh/kernel/cpu/sh4/fpu.c
187
regs->pr = regs->pc + 4; /* bsr & jsr */
arch/sh/kernel/cpu/sh4/fpu.c
219
nextpc = regs->pr;
arch/sh/kernel/kgdb.c
112
addr = linux_regs->pr;
arch/sh/kernel/kgdb.c
185
{ "pr", GDB_SIZEOF_REG, offsetof(struct pt_regs, pr) },
arch/sh/kernel/kgdb.c
248
gdb_regs[GDB_PR] = thread_regs->pr;
arch/sh/kernel/kprobes.c
166
op1->addr = (kprobe_opcode_t *) regs->pr;
arch/sh/kernel/kprobes.c
201
ri->ret_addr = (kprobe_opcode_t *) regs->pr;
arch/sh/kernel/kprobes.c
205
regs->pr = (unsigned long)__kretprobe_trampoline;
arch/sh/kernel/process_32.c
37
pr_info("PR is at %pS\n", (void *)regs->pr);
arch/sh/kernel/process_32.c
56
regs->mach, regs->macl, regs->gbr, regs->pr);
arch/sh/kernel/process_32.c
65
regs->pr = 0;
arch/sh/kernel/ptrace_32.c
265
REG_OFFSET_NAME(pr),
arch/sh/kernel/signal_32.c
130
COPY(macl); COPY(pr);
arch/sh/kernel/signal_32.c
236
COPY(macl); COPY(pr);
arch/sh/kernel/signal_32.c
289
regs->pr = (unsigned long) ksig->ka.sa.sa_restorer;
arch/sh/kernel/signal_32.c
292
regs->pr = VDSO_SYM(&__kernel_sigreturn);
arch/sh/kernel/signal_32.c
304
regs->pr = (unsigned long) frame->retcode;
arch/sh/kernel/signal_32.c
305
flush_icache_range(regs->pr, regs->pr + sizeof(frame->retcode));
arch/sh/kernel/signal_32.c
330
current->comm, task_pid_nr(current), frame, regs->pc, regs->pr);
arch/sh/kernel/signal_32.c
359
regs->pr = (unsigned long) ksig->ka.sa.sa_restorer;
arch/sh/kernel/signal_32.c
362
regs->pr = VDSO_SYM(&__kernel_rt_sigreturn);
arch/sh/kernel/signal_32.c
374
regs->pr = (unsigned long) frame->retcode;
arch/sh/kernel/signal_32.c
375
flush_icache_range(regs->pr, regs->pr + sizeof(frame->retcode));
arch/sh/kernel/signal_32.c
400
current->comm, task_pid_nr(current), frame, regs->pc, regs->pr);
arch/sh/kernel/traps_32.c
356
regs->pc = regs->pr;
arch/sh/kernel/traps_32.c
368
regs->pr = regs->pc + 4;
arch/sh/kernel/traps_32.c
395
regs->pr = regs->pc + 4;
arch/sh/kernel/traps_32.c
460
regs->pr = regs->pc + 4;
arch/sh/kernel/traps_32.c
544
regs->pr);
arch/sh/kernel/traps_32.c
671
regs->pr = regs->pc + 4;
arch/sh/kernel/traps_32.c
694
regs->pc = regs->pr;
arch/sparc/kernel/pci_common.c
379
const struct linux_prom_pci_ranges *pr = &pbm_ranges[i];
arch/sparc/kernel/pci_common.c
386
parent_phys_hi = pr->parent_phys_hi;
arch/sparc/kernel/pci_common.c
387
parent_phys_lo = pr->parent_phys_lo;
arch/sparc/kernel/pci_common.c
388
child_phys_mid = pr->child_phys_mid;
arch/sparc/kernel/pci_common.c
389
child_phys_lo = pr->child_phys_lo;
arch/sparc/kernel/pci_common.c
393
size_hi = pr->size_hi;
arch/sparc/kernel/pci_common.c
394
size_lo = pr->size_lo;
arch/sparc/kernel/pci_common.c
396
type = (pr->child_phys_hi >> 24) & 0x3;
arch/sparc/kernel/sbus.c
543
const struct linux_prom64_registers *pr;
arch/sparc/kernel/sbus.c
551
pr = of_get_property(dp, "reg", NULL);
arch/sparc/kernel/sbus.c
552
if (!pr) {
arch/sparc/kernel/sbus.c
557
regs = pr->phys_addr;
arch/x86/mm/dump_pagetables.c
179
static void printk_prot(struct seq_file *m, pgprotval_t pr, int level, bool dmsg)
arch/x86/mm/dump_pagetables.c
184
if (!(pr & _PAGE_PRESENT)) {
arch/x86/mm/dump_pagetables.c
188
if (pr & _PAGE_USER)
arch/x86/mm/dump_pagetables.c
192
if (pr & _PAGE_RW)
arch/x86/mm/dump_pagetables.c
196
if (pr & _PAGE_PWT)
arch/x86/mm/dump_pagetables.c
200
if (pr & _PAGE_PCD)
arch/x86/mm/dump_pagetables.c
206
if (level <= 3 && pr & _PAGE_PSE)
arch/x86/mm/dump_pagetables.c
210
if ((level == 4 && pr & _PAGE_PAT) ||
arch/x86/mm/dump_pagetables.c
211
((level == 3 || level == 2) && pr & _PAGE_PAT_LARGE))
arch/x86/mm/dump_pagetables.c
215
if (pr & _PAGE_GLOBAL)
arch/x86/mm/dump_pagetables.c
219
if (pr & _PAGE_NX)
crypto/async_tx/raid6test.c
119
pr("%s: timeout! (faila: %d failb: %d disks: %d)\n",
crypto/async_tx/raid6test.c
123
pr("%s: validation failure! faila: %d failb: %d sum_check_flags: %x\n",
crypto/async_tx/raid6test.c
142
pr("%s(%d, %d): faila=%3d(%c) failb=%3d(%c) %s\n",
crypto/async_tx/raid6test.c
177
pr("error: initial gen_syndrome(%d) timed out\n", disks);
crypto/async_tx/raid6test.c
181
pr("testing the %d-disk case...\n", disks);
crypto/async_tx/raid6test.c
229
pr("\n");
crypto/async_tx/raid6test.c
230
pr("complete (%d tests, %d failure%s)\n",
crypto/drbg.c
1188
if (drbg->pr || drbg->seeded == DRBG_SEED_STATE_UNSEEDED) {
crypto/drbg.c
1191
str_true_false(drbg->pr),
crypto/drbg.c
1335
int coreref, bool pr)
crypto/drbg.c
1341
"%s\n", coreref, str_enabled_disabled(pr));
crypto/drbg.c
1356
drbg->pr = pr;
crypto/drbg.c
1632
int *coreref, bool *pr)
crypto/drbg.c
1638
*pr = true;
crypto/drbg.c
1642
*pr = false;
crypto/drbg.c
1708
bool pr = false;
crypto/drbg.c
1714
&pr);
crypto/drbg.c
1720
return drbg_instantiate(drbg, seed_string, coreref, pr);
crypto/drbg.c
1745
bool pr = false;
crypto/drbg.c
1755
drbg_convert_tfm_core("drbg_nopr_ctr_aes256", &coreref, &pr);
crypto/drbg.c
1758
drbg_convert_tfm_core("drbg_nopr_sha256", &coreref, &pr);
crypto/drbg.c
1761
drbg_convert_tfm_core("drbg_nopr_hmac_sha512", &coreref, &pr);
crypto/drbg.c
1811
const struct drbg_core *core, int pr)
crypto/drbg.c
1817
if (pr) {
crypto/testmgr.c
3485
static int drbg_cavs_test(const struct drbg_testvec *test, int pr,
crypto/testmgr.c
3517
if (pr) {
crypto/testmgr.c
3532
if (pr) {
crypto/testmgr.c
3559
int pr = 0;
crypto/testmgr.c
3565
pr = 1;
crypto/testmgr.c
3568
err = drbg_cavs_test(&template[i], pr, driver, type, mask);
drivers/acpi/acpi_processor.c
202
static int acpi_processor_set_per_cpu(struct acpi_processor *pr,
drivers/acpi/acpi_processor.c
205
BUG_ON(pr->id >= nr_cpu_ids);
drivers/acpi/acpi_processor.c
212
if (per_cpu(processor_device_array, pr->id) != NULL &&
drivers/acpi/acpi_processor.c
213
per_cpu(processor_device_array, pr->id) != device) {
drivers/acpi/acpi_processor.c
216
pr->id);
drivers/acpi/acpi_processor.c
223
per_cpu(processor_device_array, pr->id) = device;
drivers/acpi/acpi_processor.c
224
per_cpu(processors, pr->id) = pr;
drivers/acpi/acpi_processor.c
230
static int acpi_processor_hotadd_init(struct acpi_processor *pr,
drivers/acpi/acpi_processor.c
235
if (invalid_phys_cpuid(pr->phys_id))
drivers/acpi/acpi_processor.c
241
ret = acpi_map_cpu(pr->handle, pr->phys_id, pr->acpi_id, &pr->id);
drivers/acpi/acpi_processor.c
245
ret = acpi_processor_set_per_cpu(pr, device);
drivers/acpi/acpi_processor.c
247
acpi_unmap_cpu(pr->id);
drivers/acpi/acpi_processor.c
251
ret = arch_register_cpu(pr->id);
drivers/acpi/acpi_processor.c
254
per_cpu(processors, pr->id) = NULL;
drivers/acpi/acpi_processor.c
255
acpi_unmap_cpu(pr->id);
drivers/acpi/acpi_processor.c
264
pr_info("CPU%d has been hot-added\n", pr->id);
drivers/acpi/acpi_processor.c
272
static inline int acpi_processor_hotadd_init(struct acpi_processor *pr,
drivers/acpi/acpi_processor.c
283
struct acpi_processor *pr = acpi_driver_data(device);
drivers/acpi/acpi_processor.c
297
pr->flags.bm_control = 1;
drivers/acpi/acpi_processor.c
304
status = acpi_evaluate_object(pr->handle, NULL, NULL, &buffer);
drivers/acpi/acpi_processor.c
312
pr->acpi_id = object.processor.proc_id;
drivers/acpi/acpi_processor.c
317
status = acpi_evaluate_integer(pr->handle, METHOD_NAME__UID,
drivers/acpi/acpi_processor.c
326
pr->acpi_id = value;
drivers/acpi/acpi_processor.c
329
if (acpi_duplicate_processor_id(pr->acpi_id)) {
drivers/acpi/acpi_processor.c
330
if (pr->acpi_id == 0xff)
drivers/acpi/acpi_processor.c
336
pr->acpi_id);
drivers/acpi/acpi_processor.c
340
pr->phys_id = acpi_get_phys_id(pr->handle, device_declaration,
drivers/acpi/acpi_processor.c
341
pr->acpi_id);
drivers/acpi/acpi_processor.c
342
if (invalid_phys_cpuid(pr->phys_id))
drivers/acpi/acpi_processor.c
345
pr->id = acpi_map_cpuid(pr->phys_id, pr->acpi_id);
drivers/acpi/acpi_processor.c
352
if (!acpi_has_cpu_in_madt() && invalid_logical_cpuid(pr->id) &&
drivers/acpi/acpi_processor.c
354
pr->id = 0;
drivers/acpi/acpi_processor.c
360
if (acpi_has_method(pr->handle, "_PCT"))
drivers/acpi/acpi_processor.c
372
if (!get_cpu_device(pr->id))
drivers/acpi/acpi_processor.c
373
ret = acpi_processor_hotadd_init(pr, device);
drivers/acpi/acpi_processor.c
375
ret = acpi_processor_set_per_cpu(pr, device);
drivers/acpi/acpi_processor.c
388
sprintf(acpi_device_bid(device), "CPU%X", pr->id);
drivers/acpi/acpi_processor.c
389
dev_dbg(&device->dev, "Processor [%d:%d]\n", pr->id, pr->acpi_id);
drivers/acpi/acpi_processor.c
397
pr->throttling.address = object.processor.pblk_address;
drivers/acpi/acpi_processor.c
398
pr->throttling.duty_offset = acpi_gbl_FADT.duty_offset;
drivers/acpi/acpi_processor.c
399
pr->throttling.duty_width = acpi_gbl_FADT.duty_width;
drivers/acpi/acpi_processor.c
40
struct acpi_processor *pr;
drivers/acpi/acpi_processor.c
401
pr->pblk = object.processor.pblk_address;
drivers/acpi/acpi_processor.c
409
status = acpi_evaluate_integer(pr->handle, "_SUN", NULL, &value);
drivers/acpi/acpi_processor.c
411
arch_fix_phys_package_id(pr->id, value);
drivers/acpi/acpi_processor.c
42
pr = per_cpu(processors, cpu);
drivers/acpi/acpi_processor.c
425
struct acpi_processor *pr;
drivers/acpi/acpi_processor.c
43
if (pr)
drivers/acpi/acpi_processor.c
432
pr = kzalloc_obj(struct acpi_processor);
drivers/acpi/acpi_processor.c
433
if (!pr)
drivers/acpi/acpi_processor.c
436
if (!zalloc_cpumask_var(&pr->throttling.shared_cpu_map, GFP_KERNEL)) {
drivers/acpi/acpi_processor.c
44
return pr->handle;
drivers/acpi/acpi_processor.c
441
pr->handle = device->handle;
drivers/acpi/acpi_processor.c
444
device->driver_data = pr;
drivers/acpi/acpi_processor.c
450
dev = get_cpu_device(pr->id);
drivers/acpi/acpi_processor.c
460
pr->dev = dev;
drivers/acpi/acpi_processor.c
470
per_cpu(processors, pr->id) = NULL;
drivers/acpi/acpi_processor.c
473
free_cpumask_var(pr->throttling.shared_cpu_map);
drivers/acpi/acpi_processor.c
475
kfree(pr);
drivers/acpi/acpi_processor.c
483
struct acpi_processor *pr;
drivers/acpi/acpi_processor.c
488
pr = acpi_driver_data(device);
drivers/acpi/acpi_processor.c
489
if (pr->id >= nr_cpu_ids)
drivers/acpi/acpi_processor.c
500
device_release_driver(pr->dev);
drivers/acpi/acpi_processor.c
501
acpi_unbind_one(pr->dev);
drivers/acpi/acpi_processor.c
507
arch_unregister_cpu(pr->id);
drivers/acpi/acpi_processor.c
508
acpi_unmap_cpu(pr->id);
drivers/acpi/acpi_processor.c
511
per_cpu(processor_device_array, pr->id) = NULL;
drivers/acpi/acpi_processor.c
512
per_cpu(processors, pr->id) = NULL;
drivers/acpi/acpi_processor.c
517
try_offline_node(cpu_to_node(pr->id));
drivers/acpi/acpi_processor.c
520
free_cpumask_var(pr->throttling.shared_cpu_map);
drivers/acpi/acpi_processor.c
521
kfree(pr);
drivers/acpi/arm64/cpuidle.c
21
struct acpi_processor *pr = per_cpu(processors, cpu);
drivers/acpi/arm64/cpuidle.c
23
if (unlikely(!pr || !pr->flags.has_lpi))
drivers/acpi/arm64/cpuidle.c
33
count = pr->power.count - 1;
drivers/acpi/arm64/cpuidle.c
40
lpi = &pr->power.lpi_states[i + 1];
drivers/acpi/cppc_acpi.c
683
int acpi_cppc_processor_probe(struct acpi_processor *pr)
drivers/acpi/cppc_acpi.c
690
acpi_handle handle = pr->handle;
drivers/acpi/cppc_acpi.c
726
num_ent, pr->id);
drivers/acpi/cppc_acpi.c
731
cpc_obj->type, pr->id);
drivers/acpi/cppc_acpi.c
741
cpc_obj->type, pr->id);
drivers/acpi/cppc_acpi.c
747
pr->id);
drivers/acpi/cppc_acpi.c
760
num_ent, pr->id);
drivers/acpi/cppc_acpi.c
795
pr->id);
drivers/acpi/cppc_acpi.c
850
i, pr->id);
drivers/acpi/cppc_acpi.c
854
per_cpu(cpu_pcc_subspace_idx, pr->id) = pcc_subspace_id;
drivers/acpi/cppc_acpi.c
868
cpc_ptr->cpu_id = pr->id;
drivers/acpi/cppc_acpi.c
887
pr_debug("Parsed CPC struct for CPU: %d\n", pr->id);
drivers/acpi/cppc_acpi.c
890
cpu_dev = get_cpu_device(pr->id);
drivers/acpi/cppc_acpi.c
897
per_cpu(cpc_desc_ptr, pr->id) = cpc_ptr;
drivers/acpi/cppc_acpi.c
902
per_cpu(cpc_desc_ptr, pr->id) = NULL;
drivers/acpi/cppc_acpi.c
932
void acpi_cppc_processor_exit(struct acpi_processor *pr)
drivers/acpi/cppc_acpi.c
937
int pcc_ss_id = per_cpu(cpu_pcc_subspace_idx, pr->id);
drivers/acpi/cppc_acpi.c
950
cpc_ptr = per_cpu(cpc_desc_ptr, pr->id);
drivers/acpi/power.c
77
static inline const char *resource_dev_name(struct acpi_power_resource *pr)
drivers/acpi/power.c
79
return dev_name(&pr->device.dev);
drivers/acpi/processor_driver.c
102
struct acpi_processor *pr = per_cpu(processors, cpu);
drivers/acpi/processor_driver.c
105
if (!pr)
drivers/acpi/processor_driver.c
108
device = acpi_fetch_acpi_dev(pr->handle);
drivers/acpi/processor_driver.c
116
if (!pr->flags.previously_online) {
drivers/acpi/processor_driver.c
120
WARN(ret, "Failed to start CPU: %d\n", pr->id);
drivers/acpi/processor_driver.c
123
acpi_processor_ppc_has_changed(pr, 0);
drivers/acpi/processor_driver.c
124
acpi_processor_hotplug(pr);
drivers/acpi/processor_driver.c
125
acpi_processor_reevaluate_tstate(pr, false);
drivers/acpi/processor_driver.c
126
acpi_processor_tstate_has_changed(pr);
drivers/acpi/processor_driver.c
133
struct acpi_processor *pr = per_cpu(processors, cpu);
drivers/acpi/processor_driver.c
135
if (!pr || !acpi_fetch_acpi_dev(pr->handle))
drivers/acpi/processor_driver.c
138
acpi_processor_reevaluate_tstate(pr, true);
drivers/acpi/processor_driver.c
143
static void acpi_pss_perf_init(struct acpi_processor *pr)
drivers/acpi/processor_driver.c
145
acpi_processor_ppc_has_changed(pr, 0);
drivers/acpi/processor_driver.c
147
acpi_processor_get_throttling_info(pr);
drivers/acpi/processor_driver.c
149
if (pr->flags.throttling)
drivers/acpi/processor_driver.c
150
pr->flags.limit = 1;
drivers/acpi/processor_driver.c
153
static inline void acpi_pss_perf_init(struct acpi_processor *pr) {}
drivers/acpi/processor_driver.c
158
struct acpi_processor *pr = acpi_driver_data(device);
drivers/acpi/processor_driver.c
162
if (!pr)
drivers/acpi/processor_driver.c
165
result = acpi_cppc_processor_probe(pr);
drivers/acpi/processor_driver.c
169
acpi_processor_power_init(pr);
drivers/acpi/processor_driver.c
171
acpi_pss_perf_init(pr);
drivers/acpi/processor_driver.c
173
result = acpi_processor_thermal_init(pr, device);
drivers/acpi/processor_driver.c
183
pr->flags.previously_online = 1;
drivers/acpi/processor_driver.c
188
acpi_processor_thermal_exit(pr, device);
drivers/acpi/processor_driver.c
190
acpi_processor_power_exit(pr);
drivers/acpi/processor_driver.c
197
struct acpi_processor *pr;
drivers/acpi/processor_driver.c
205
pr = acpi_driver_data(device);
drivers/acpi/processor_driver.c
206
if (!pr)
drivers/acpi/processor_driver.c
208
acpi_processor_power_exit(pr);
drivers/acpi/processor_driver.c
210
acpi_cppc_processor_exit(pr);
drivers/acpi/processor_driver.c
212
acpi_processor_thermal_exit(pr, device);
drivers/acpi/processor_driver.c
55
struct acpi_processor *pr;
drivers/acpi/processor_driver.c
61
pr = acpi_driver_data(device);
drivers/acpi/processor_driver.c
62
if (!pr)
drivers/acpi/processor_driver.c
67
saved = pr->performance_platform_limit;
drivers/acpi/processor_driver.c
68
acpi_processor_ppc_has_changed(pr, 1);
drivers/acpi/processor_driver.c
69
if (saved == pr->performance_platform_limit)
drivers/acpi/processor_driver.c
73
pr->performance_platform_limit);
drivers/acpi/processor_driver.c
76
acpi_processor_power_state_has_changed(pr);
drivers/acpi/processor_driver.c
81
acpi_processor_tstate_has_changed(pr);
drivers/acpi/processor_driver.c
86
cpufreq_update_limits(pr->id);
drivers/acpi/processor_idle.c
1033
static unsigned int flatten_lpi_states(struct acpi_processor *pr,
drivers/acpi/processor_idle.c
1055
flpi = &pr->power.lpi_states[flat_state_cnt];
drivers/acpi/processor_idle.c
1084
static int acpi_processor_get_lpi_info(struct acpi_processor *pr)
drivers/acpi/processor_idle.c
1088
acpi_handle handle = pr->handle, pr_ahandle;
drivers/acpi/processor_idle.c
1094
ret = acpi_processor_ffh_lpi_probe(pr->id);
drivers/acpi/processor_idle.c
1106
handle = pr->handle;
drivers/acpi/processor_idle.c
1110
state_count = flatten_lpi_states(pr, 0, prev, NULL);
drivers/acpi/processor_idle.c
1132
state_count = flatten_lpi_states(pr, state_count, curr, prev);
drivers/acpi/processor_idle.c
1141
pr->power.lpi_states[i].index = i;
drivers/acpi/processor_idle.c
1143
pr->power.count = state_count;
drivers/acpi/processor_idle.c
1146
pr->flags.has_lpi = 1;
drivers/acpi/processor_idle.c
1147
pr->flags.power = 1;
drivers/acpi/processor_idle.c
1168
struct acpi_processor *pr;
drivers/acpi/processor_idle.c
1171
pr = __this_cpu_read(processors);
drivers/acpi/processor_idle.c
1173
if (unlikely(!pr))
drivers/acpi/processor_idle.c
1176
lpi = &pr->power.lpi_states[index];
drivers/acpi/processor_idle.c
1183
static void acpi_processor_setup_lpi_states(struct acpi_processor *pr)
drivers/acpi/processor_idle.c
1190
if (!pr->flags.has_lpi)
drivers/acpi/processor_idle.c
1193
for (i = 0; i < pr->power.count && i < CPUIDLE_STATE_MAX; i++) {
drivers/acpi/processor_idle.c
1194
lpi = &pr->power.lpi_states[i];
drivers/acpi/processor_idle.c
1217
static void acpi_processor_setup_cpuidle_states(struct acpi_processor *pr)
drivers/acpi/processor_idle.c
1222
if (!pr->flags.power_setup_done || !pr->flags.power)
drivers/acpi/processor_idle.c
1231
if (pr->flags.has_lpi) {
drivers/acpi/processor_idle.c
1232
acpi_processor_setup_lpi_states(pr);
drivers/acpi/processor_idle.c
1236
acpi_processor_setup_cstates(pr);
drivers/acpi/processor_idle.c
1246
static void acpi_processor_setup_cpuidle_dev(struct acpi_processor *pr,
drivers/acpi/processor_idle.c
1249
if (!pr->flags.power_setup_done || !pr->flags.power || !dev)
drivers/acpi/processor_idle.c
1252
dev->cpu = pr->id;
drivers/acpi/processor_idle.c
1253
if (!pr->flags.has_lpi)
drivers/acpi/processor_idle.c
1254
acpi_processor_setup_cpuidle_cx(pr, dev);
drivers/acpi/processor_idle.c
1257
static int acpi_processor_get_power_info(struct acpi_processor *pr)
drivers/acpi/processor_idle.c
1261
ret = acpi_processor_get_lpi_info(pr);
drivers/acpi/processor_idle.c
1263
return acpi_processor_get_cstate_info(pr);
drivers/acpi/processor_idle.c
1265
if (pr->flags.has_lpi) {
drivers/acpi/processor_idle.c
1266
ret = acpi_processor_ffh_lpi_probe(pr->id);
drivers/acpi/processor_idle.c
1268
pr_err("CPU%u: Invalid FFH LPI data\n", pr->id);
drivers/acpi/processor_idle.c
1274
int acpi_processor_hotplug(struct acpi_processor *pr)
drivers/acpi/processor_idle.c
1282
if (!pr->flags.power_setup_done)
drivers/acpi/processor_idle.c
1285
dev = per_cpu(acpi_cpuidle_device, pr->id);
drivers/acpi/processor_idle.c
1288
ret = acpi_processor_get_power_info(pr);
drivers/acpi/processor_idle.c
1289
if (!ret && pr->flags.power) {
drivers/acpi/processor_idle.c
1290
acpi_processor_setup_cpuidle_dev(pr, dev);
drivers/acpi/processor_idle.c
1298
int acpi_processor_power_state_has_changed(struct acpi_processor *pr)
drivers/acpi/processor_idle.c
1307
if (!pr->flags.power_setup_done)
drivers/acpi/processor_idle.c
1316
if (pr->id == 0 && cpuidle_get_driver() == &acpi_idle_driver) {
drivers/acpi/processor_idle.c
132
static void lapic_timer_check_state(int state, struct acpi_processor *pr,
drivers/acpi/processor_idle.c
1332
acpi_processor_get_power_info(pr);
drivers/acpi/processor_idle.c
1333
acpi_processor_setup_cpuidle_states(pr);
drivers/acpi/processor_idle.c
135
struct acpi_processor_power *pwr = &pr->power;
drivers/acpi/processor_idle.c
1356
struct acpi_processor *pr;
drivers/acpi/processor_idle.c
1367
pr = per_cpu(processors, cpu);
drivers/acpi/processor_idle.c
1368
if (!pr)
drivers/acpi/processor_idle.c
1372
ret = acpi_processor_get_power_info(pr);
drivers/acpi/processor_idle.c
1374
pr->flags.power_setup_done = 1;
drivers/acpi/processor_idle.c
1375
acpi_processor_setup_cpuidle_states(pr);
drivers/acpi/processor_idle.c
138
if (cpu_has(&cpu_data(pr->id), X86_FEATURE_ARAT))
drivers/acpi/processor_idle.c
1398
void acpi_processor_power_init(struct acpi_processor *pr)
drivers/acpi/processor_idle.c
1414
if (!acpi_processor_get_power_info(pr))
drivers/acpi/processor_idle.c
1415
pr->flags.power_setup_done = 1;
drivers/acpi/processor_idle.c
1417
if (!pr->flags.power)
drivers/acpi/processor_idle.c
1424
per_cpu(acpi_cpuidle_device, pr->id) = dev;
drivers/acpi/processor_idle.c
1426
acpi_processor_setup_cpuidle_dev(pr, dev);
drivers/acpi/processor_idle.c
1433
per_cpu(acpi_cpuidle_device, pr->id) = NULL;
drivers/acpi/processor_idle.c
1438
void acpi_processor_power_exit(struct acpi_processor *pr)
drivers/acpi/processor_idle.c
1440
struct cpuidle_device *dev = per_cpu(acpi_cpuidle_device, pr->id);
drivers/acpi/processor_idle.c
1445
if (pr->flags.power) {
drivers/acpi/processor_idle.c
1450
pr->flags.power_setup_done = 0;
drivers/acpi/processor_idle.c
152
pr->power.timer_broadcast_on_state = state;
drivers/acpi/processor_idle.c
157
struct acpi_processor *pr = arg;
drivers/acpi/processor_idle.c
159
if (pr->power.timer_broadcast_on_state < INT_MAX)
drivers/acpi/processor_idle.c
165
static void lapic_timer_propagate_broadcast(struct acpi_processor *pr)
drivers/acpi/processor_idle.c
167
smp_call_function_single(pr->id, __lapic_timer_propagate_broadcast,
drivers/acpi/processor_idle.c
168
(void *)pr, 1);
drivers/acpi/processor_idle.c
172
static bool lapic_timer_needs_broadcast(struct acpi_processor *pr,
drivers/acpi/processor_idle.c
175
return cx - pr->power.states >= pr->power.timer_broadcast_on_state;
drivers/acpi/processor_idle.c
180
static void lapic_timer_check_state(int state, struct acpi_processor *pr,
drivers/acpi/processor_idle.c
182
static void lapic_timer_propagate_broadcast(struct acpi_processor *pr) { }
drivers/acpi/processor_idle.c
184
static bool lapic_timer_needs_broadcast(struct acpi_processor *pr,
drivers/acpi/processor_idle.c
218
static int acpi_processor_get_power_info_fadt(struct acpi_processor *pr)
drivers/acpi/processor_idle.c
221
if (!pr->pblk)
drivers/acpi/processor_idle.c
225
pr->power.states[ACPI_STATE_C2].type = ACPI_STATE_C2;
drivers/acpi/processor_idle.c
226
pr->power.states[ACPI_STATE_C3].type = ACPI_STATE_C3;
drivers/acpi/processor_idle.c
239
pr->power.states[ACPI_STATE_C2].address = pr->pblk + 4;
drivers/acpi/processor_idle.c
240
pr->power.states[ACPI_STATE_C3].address = pr->pblk + 5;
drivers/acpi/processor_idle.c
243
pr->power.states[ACPI_STATE_C2].latency = acpi_gbl_FADT.c2_latency;
drivers/acpi/processor_idle.c
244
pr->power.states[ACPI_STATE_C3].latency = acpi_gbl_FADT.c3_latency;
drivers/acpi/processor_idle.c
251
acpi_handle_debug(pr->handle, "C2 latency too large [%d]\n",
drivers/acpi/processor_idle.c
254
pr->power.states[ACPI_STATE_C2].address = 0;
drivers/acpi/processor_idle.c
262
acpi_handle_debug(pr->handle, "C3 latency too large [%d]\n",
drivers/acpi/processor_idle.c
265
pr->power.states[ACPI_STATE_C3].address = 0;
drivers/acpi/processor_idle.c
268
acpi_handle_debug(pr->handle, "lvl2[0x%08x] lvl3[0x%08x]\n",
drivers/acpi/processor_idle.c
269
pr->power.states[ACPI_STATE_C2].address,
drivers/acpi/processor_idle.c
270
pr->power.states[ACPI_STATE_C3].address);
drivers/acpi/processor_idle.c
272
snprintf(pr->power.states[ACPI_STATE_C2].desc,
drivers/acpi/processor_idle.c
274
pr->power.states[ACPI_STATE_C2].address);
drivers/acpi/processor_idle.c
275
snprintf(pr->power.states[ACPI_STATE_C3].desc,
drivers/acpi/processor_idle.c
277
pr->power.states[ACPI_STATE_C3].address);
drivers/acpi/processor_idle.c
279
if (!pr->power.states[ACPI_STATE_C2].address &&
drivers/acpi/processor_idle.c
280
!pr->power.states[ACPI_STATE_C3].address)
drivers/acpi/processor_idle.c
286
static int acpi_processor_get_power_info_default(struct acpi_processor *pr)
drivers/acpi/processor_idle.c
288
if (!pr->power.states[ACPI_STATE_C1].valid) {
drivers/acpi/processor_idle.c
291
pr->power.states[ACPI_STATE_C1].type = ACPI_STATE_C1;
drivers/acpi/processor_idle.c
292
pr->power.states[ACPI_STATE_C1].valid = 1;
drivers/acpi/processor_idle.c
293
pr->power.states[ACPI_STATE_C1].entry_method = ACPI_CSTATE_HALT;
drivers/acpi/processor_idle.c
295
snprintf(pr->power.states[ACPI_STATE_C1].desc,
drivers/acpi/processor_idle.c
299
pr->power.states[ACPI_STATE_C0].valid = 1;
drivers/acpi/processor_idle.c
303
static int acpi_processor_get_power_info_cst(struct acpi_processor *pr)
drivers/acpi/processor_idle.c
310
ret = acpi_processor_evaluate_cst(pr->handle, pr->id, &pr->power);
drivers/acpi/processor_idle.c
314
if (!pr->power.count)
drivers/acpi/processor_idle.c
317
pr->flags.has_cst = 1;
drivers/acpi/processor_idle.c
321
static void acpi_processor_power_verify_c3(struct acpi_processor *pr,
drivers/acpi/processor_idle.c
339
acpi_handle_debug(pr->handle,
drivers/acpi/processor_idle.c
347
acpi_processor_power_init_bm_check(&(pr->flags), pr->id);
drivers/acpi/processor_idle.c
348
bm_check_flag = pr->flags.bm_check;
drivers/acpi/processor_idle.c
349
bm_control_flag = pr->flags.bm_control;
drivers/acpi/processor_idle.c
351
pr->flags.bm_check = bm_check_flag;
drivers/acpi/processor_idle.c
352
pr->flags.bm_control = bm_control_flag;
drivers/acpi/processor_idle.c
355
if (pr->flags.bm_check) {
drivers/acpi/processor_idle.c
356
if (!pr->flags.bm_control) {
drivers/acpi/processor_idle.c
357
if (pr->flags.has_cst != 1) {
drivers/acpi/processor_idle.c
359
acpi_handle_debug(pr->handle,
drivers/acpi/processor_idle.c
364
acpi_handle_debug(pr->handle,
drivers/acpi/processor_idle.c
374
acpi_handle_debug(pr->handle,
drivers/acpi/processor_idle.c
420
static int acpi_processor_power_verify(struct acpi_processor *pr)
drivers/acpi/processor_idle.c
428
pr->power.timer_broadcast_on_state = INT_MAX;
drivers/acpi/processor_idle.c
431
struct acpi_processor_cx *cx = &pr->power.states[i];
drivers/acpi/processor_idle.c
445
acpi_processor_power_verify_c3(pr, cx);
drivers/acpi/processor_idle.c
455
lapic_timer_check_state(i, pr, cx);
drivers/acpi/processor_idle.c
462
acpi_cst_latency_sort(&pr->power.states[1], max_cstate);
drivers/acpi/processor_idle.c
465
lapic_timer_propagate_broadcast(pr);
drivers/acpi/processor_idle.c
470
static int acpi_processor_get_cstate_info(struct acpi_processor *pr)
drivers/acpi/processor_idle.c
478
memset(pr->power.states, 0, sizeof(pr->power.states));
drivers/acpi/processor_idle.c
480
result = acpi_processor_get_power_info_cst(pr);
drivers/acpi/processor_idle.c
482
result = acpi_processor_get_power_info_fadt(pr);
drivers/acpi/processor_idle.c
487
acpi_processor_get_power_info_default(pr);
drivers/acpi/processor_idle.c
489
pr->power.count = acpi_processor_power_verify(pr);
drivers/acpi/processor_idle.c
490
pr->flags.power = 1;
drivers/acpi/processor_idle.c
600
static __always_inline bool acpi_idle_fallback_to_c1(struct acpi_processor *pr)
drivers/acpi/processor_idle.c
602
return IS_ENABLED(CONFIG_HOTPLUG_CPU) && !pr->flags.has_cst &&
drivers/acpi/processor_idle.c
617
struct acpi_processor *pr,
drivers/acpi/processor_idle.c
633
bool dis_bm = pr->flags.bm_control;
drivers/acpi/processor_idle.c
681
struct acpi_processor *pr;
drivers/acpi/processor_idle.c
683
pr = __this_cpu_read(processors);
drivers/acpi/processor_idle.c
684
if (unlikely(!pr))
drivers/acpi/processor_idle.c
688
if (cx->type == ACPI_STATE_C3 && pr->flags.bm_check)
drivers/acpi/processor_idle.c
689
return acpi_idle_enter_bm(drv, pr, cx, index);
drivers/acpi/processor_idle.c
692
if (acpi_idle_fallback_to_c1(pr) && num_online_cpus() > 1) {
drivers/acpi/processor_idle.c
712
struct acpi_processor *pr = __this_cpu_read(processors);
drivers/acpi/processor_idle.c
714
if (unlikely(!pr))
drivers/acpi/processor_idle.c
717
if (pr->flags.bm_check) {
drivers/acpi/processor_idle.c
722
acpi_idle_enter_bm(drv, pr, cx, index);
drivers/acpi/processor_idle.c
735
static void acpi_processor_setup_cpuidle_cx(struct acpi_processor *pr,
drivers/acpi/processor_idle.c
745
cx = &pr->power.states[i];
drivers/acpi/processor_idle.c
758
static void acpi_processor_setup_cstates(struct acpi_processor *pr)
drivers/acpi/processor_idle.c
776
cx = &pr->power.states[i];
drivers/acpi/processor_idle.c
802
if (cx->type != ACPI_STATE_C1 && !acpi_idle_fallback_to_c1(pr))
drivers/acpi/processor_idle.c
805
if (lapic_timer_needs_broadcast(pr, cx))
drivers/acpi/processor_idle.c
810
if (pr->flags.bm_check)
drivers/acpi/processor_idle.c
844
static int acpi_processor_get_cstate_info(struct acpi_processor *pr)
drivers/acpi/processor_idle.c
849
static int acpi_processor_setup_cpuidle_cx(struct acpi_processor *pr,
drivers/acpi/processor_idle.c
855
static int acpi_processor_setup_cstates(struct acpi_processor *pr)
drivers/acpi/processor_perflib.c
102
pr->id, ret);
drivers/acpi/processor_perflib.c
123
void acpi_processor_ppc_has_changed(struct acpi_processor *pr, int event_flag)
drivers/acpi/processor_perflib.c
127
if (ignore_ppc || !pr->performance) {
drivers/acpi/processor_perflib.c
133
acpi_processor_ppc_ost(pr->handle, 1);
drivers/acpi/processor_perflib.c
137
ret = acpi_processor_get_platform_limit(pr);
drivers/acpi/processor_perflib.c
144
acpi_processor_ppc_ost(pr->handle, 1);
drivers/acpi/processor_perflib.c
146
acpi_processor_ppc_ost(pr->handle, 0);
drivers/acpi/processor_perflib.c
149
cpufreq_update_limits(pr->id);
drivers/acpi/processor_perflib.c
154
struct acpi_processor *pr;
drivers/acpi/processor_perflib.c
156
pr = per_cpu(processors, cpu);
drivers/acpi/processor_perflib.c
157
if (!pr || !pr->performance || !pr->performance->state_count)
drivers/acpi/processor_perflib.c
160
*limit = pr->performance->states[pr->performance_platform_limit].
drivers/acpi/processor_perflib.c
180
struct acpi_processor *pr = per_cpu(processors, cpu);
drivers/acpi/processor_perflib.c
183
if (!pr)
drivers/acpi/processor_perflib.c
191
pr->performance_platform_limit = 0;
drivers/acpi/processor_perflib.c
194
&pr->perflib_req, FREQ_QOS_MAX,
drivers/acpi/processor_perflib.c
200
if (!pr->performance)
drivers/acpi/processor_perflib.c
203
ret = acpi_processor_get_platform_limit(pr);
drivers/acpi/processor_perflib.c
215
struct acpi_processor *pr = per_cpu(processors, cpu);
drivers/acpi/processor_perflib.c
217
if (pr)
drivers/acpi/processor_perflib.c
218
freq_qos_remove_request(&pr->perflib_req);
drivers/acpi/processor_perflib.c
226
static int acpi_processor_get_performance_control(struct acpi_processor *pr)
drivers/acpi/processor_perflib.c
234
status = acpi_evaluate_object(pr->handle, "_PCT", NULL, &buffer);
drivers/acpi/processor_perflib.c
236
acpi_evaluation_failure_warn(pr->handle, "_PCT", status);
drivers/acpi/processor_perflib.c
259
memcpy(&pr->performance->control_register, obj.buffer.pointer,
drivers/acpi/processor_perflib.c
275
memcpy(&pr->performance->status_register, obj.buffer.pointer,
drivers/acpi/processor_perflib.c
315
static int acpi_processor_get_performance_states(struct acpi_processor *pr)
drivers/acpi/processor_perflib.c
326
status = acpi_evaluate_object(pr->handle, "_PSS", NULL, &buffer);
drivers/acpi/processor_perflib.c
328
acpi_evaluation_failure_warn(pr->handle, "_PSS", status);
drivers/acpi/processor_perflib.c
339
acpi_handle_debug(pr->handle, "Found %d performance states\n",
drivers/acpi/processor_perflib.c
342
pr->performance->state_count = pss->package.count;
drivers/acpi/processor_perflib.c
343
pr->performance->states =
drivers/acpi/processor_perflib.c
345
if (!pr->performance->states) {
drivers/acpi/processor_perflib.c
350
for (i = 0; i < pr->performance->state_count; i++) {
drivers/acpi/processor_perflib.c
352
struct acpi_processor_px *px = &(pr->performance->states[i]);
drivers/acpi/processor_perflib.c
357
acpi_handle_debug(pr->handle, "Extracting state %d\n", i);
drivers/acpi/processor_perflib.c
362
acpi_handle_warn(pr->handle, "Invalid _PSS data: %s\n",
drivers/acpi/processor_perflib.c
365
kfree(pr->performance->states);
drivers/acpi/processor_perflib.c
371
acpi_handle_debug(pr->handle,
drivers/acpi/processor_perflib.c
387
pr->id, px->core_frequency);
drivers/acpi/processor_perflib.c
395
memcpy(&(pr->performance->states[last_invalid]),
drivers/acpi/processor_perflib.c
404
"No valid BIOS _PSS frequency found for processor %d\n", pr->id);
drivers/acpi/processor_perflib.c
406
kfree(pr->performance->states);
drivers/acpi/processor_perflib.c
407
pr->performance->states = NULL;
drivers/acpi/processor_perflib.c
411
pr->performance->state_count = last_invalid;
drivers/acpi/processor_perflib.c
419
int acpi_processor_get_performance_info(struct acpi_processor *pr)
drivers/acpi/processor_perflib.c
423
if (!pr || !pr->performance || !pr->handle)
drivers/acpi/processor_perflib.c
426
if (!acpi_has_method(pr->handle, "_PCT")) {
drivers/acpi/processor_perflib.c
427
acpi_handle_debug(pr->handle,
drivers/acpi/processor_perflib.c
432
result = acpi_processor_get_performance_control(pr);
drivers/acpi/processor_perflib.c
436
result = acpi_processor_get_performance_states(pr);
drivers/acpi/processor_perflib.c
442
result = acpi_processor_get_platform_limit(pr);
drivers/acpi/processor_perflib.c
451
if (acpi_has_method(pr->handle, "_PPC")) {
drivers/acpi/processor_perflib.c
51
static int acpi_processor_get_platform_limit(struct acpi_processor *pr)
drivers/acpi/processor_perflib.c
59
if (!pr)
drivers/acpi/processor_perflib.c
598
struct acpi_processor *pr;
drivers/acpi/processor_perflib.c
613
pr = per_cpu(processors, i);
drivers/acpi/processor_perflib.c
614
if (!pr) {
drivers/acpi/processor_perflib.c
619
if (pr->performance) {
drivers/acpi/processor_perflib.c
632
pr = per_cpu(processors, i);
drivers/acpi/processor_perflib.c
633
if (!pr)
drivers/acpi/processor_perflib.c
636
pr->performance = per_cpu_ptr(performance, i);
drivers/acpi/processor_perflib.c
637
pdomain = &(pr->performance->domain_info);
drivers/acpi/processor_perflib.c
638
if (acpi_processor_get_psd(pr->handle, pdomain)) {
drivers/acpi/processor_perflib.c
651
pr = per_cpu(processors, i);
drivers/acpi/processor_perflib.c
652
if (!pr)
drivers/acpi/processor_perflib.c
658
pdomain = &(pr->performance->domain_info);
drivers/acpi/processor_perflib.c
659
cpumask_set_cpu(i, pr->performance->shared_cpu_map);
drivers/acpi/processor_perflib.c
66
status = acpi_evaluate_integer(pr->handle, "_PPC", NULL, &ppc);
drivers/acpi/processor_perflib.c
667
pr->performance->shared_type = CPUFREQ_SHARED_TYPE_ALL;
drivers/acpi/processor_perflib.c
669
pr->performance->shared_type = CPUFREQ_SHARED_TYPE_HW;
drivers/acpi/processor_perflib.c
671
pr->performance->shared_type = CPUFREQ_SHARED_TYPE_ANY;
drivers/acpi/processor_perflib.c
698
cpumask_set_cpu(j, pr->performance->shared_cpu_map);
drivers/acpi/processor_perflib.c
71
acpi_evaluation_failure_warn(pr->handle, "_PPC", status);
drivers/acpi/processor_perflib.c
714
pr->performance->shared_type;
drivers/acpi/processor_perflib.c
716
pr->performance->shared_cpu_map);
drivers/acpi/processor_perflib.c
722
pr = per_cpu(processors, i);
drivers/acpi/processor_perflib.c
723
if (!pr || !pr->performance)
drivers/acpi/processor_perflib.c
728
cpumask_clear(pr->performance->shared_cpu_map);
drivers/acpi/processor_perflib.c
729
cpumask_set_cpu(i, pr->performance->shared_cpu_map);
drivers/acpi/processor_perflib.c
730
pr->performance->shared_type = CPUFREQ_SHARED_TYPE_NONE;
drivers/acpi/processor_perflib.c
732
pr->performance = NULL; /* Will be set for real in register */
drivers/acpi/processor_perflib.c
745
struct acpi_processor *pr;
drivers/acpi/processor_perflib.c
752
pr = per_cpu(processors, cpu);
drivers/acpi/processor_perflib.c
753
if (!pr) {
drivers/acpi/processor_perflib.c
758
if (pr->performance) {
drivers/acpi/processor_perflib.c
765
pr->performance = performance;
drivers/acpi/processor_perflib.c
767
if (acpi_processor_get_performance_info(pr)) {
drivers/acpi/processor_perflib.c
768
pr->performance = NULL;
drivers/acpi/processor_perflib.c
78
if (pr->performance_platform_limit == index ||
drivers/acpi/processor_perflib.c
780
struct acpi_processor *pr;
drivers/acpi/processor_perflib.c
784
pr = per_cpu(processors, cpu);
drivers/acpi/processor_perflib.c
785
if (!pr)
drivers/acpi/processor_perflib.c
788
if (pr->performance)
drivers/acpi/processor_perflib.c
789
kfree(pr->performance->states);
drivers/acpi/processor_perflib.c
79
ppc >= pr->performance->state_count)
drivers/acpi/processor_perflib.c
791
pr->performance = NULL;
drivers/acpi/processor_perflib.c
82
pr_debug("CPU %d: _PPC is %d - frequency %s limited\n", pr->id,
drivers/acpi/processor_perflib.c
85
pr->performance_platform_limit = index;
drivers/acpi/processor_perflib.c
87
if (unlikely(!freq_qos_request_active(&pr->perflib_req)))
drivers/acpi/processor_perflib.c
97
qos_value = pr->performance->states[index].core_frequency * 1000;
drivers/acpi/processor_perflib.c
99
ret = freq_qos_update_request(&pr->perflib_req, qos_value);
drivers/acpi/processor_thermal.c
103
ret = freq_qos_update_request(&pr->thermal_req, max_freq);
drivers/acpi/processor_thermal.c
106
pr->id, ret);
drivers/acpi/processor_thermal.c
114
struct acpi_processor *pr;
drivers/acpi/processor_thermal.c
132
pr = per_cpu(processors, i);
drivers/acpi/processor_thermal.c
134
if (unlikely(!freq_qos_request_active(&pr->thermal_req)))
drivers/acpi/processor_thermal.c
137
if (!cpufreq_update_thermal_limit(i, pr))
drivers/acpi/processor_thermal.c
167
struct acpi_processor *pr = per_cpu(processors, cpu);
drivers/acpi/processor_thermal.c
170
if (!pr)
drivers/acpi/processor_thermal.c
174
&pr->thermal_req,
drivers/acpi/processor_thermal.c
182
thermal_cooling_device_update(pr->cdev);
drivers/acpi/processor_thermal.c
191
struct acpi_processor *pr = per_cpu(processors, cpu);
drivers/acpi/processor_thermal.c
193
if (!pr)
drivers/acpi/processor_thermal.c
196
freq_qos_remove_request(&pr->thermal_req);
drivers/acpi/processor_thermal.c
198
thermal_cooling_device_update(pr->cdev);
drivers/acpi/processor_thermal.c
220
static int acpi_processor_max_state(struct acpi_processor *pr)
drivers/acpi/processor_thermal.c
228
max_state += cpufreq_get_max_state(pr->id);
drivers/acpi/processor_thermal.c
229
if (pr->flags.throttling)
drivers/acpi/processor_thermal.c
230
max_state += (pr->throttling.state_count -1);
drivers/acpi/processor_thermal.c
239
struct acpi_processor *pr;
drivers/acpi/processor_thermal.c
244
pr = acpi_driver_data(device);
drivers/acpi/processor_thermal.c
245
if (!pr)
drivers/acpi/processor_thermal.c
248
*state = acpi_processor_max_state(pr);
drivers/acpi/processor_thermal.c
257
struct acpi_processor *pr;
drivers/acpi/processor_thermal.c
262
pr = acpi_driver_data(device);
drivers/acpi/processor_thermal.c
263
if (!pr)
drivers/acpi/processor_thermal.c
266
*cur_state = cpufreq_get_cur_state(pr->id);
drivers/acpi/processor_thermal.c
267
if (pr->flags.throttling)
drivers/acpi/processor_thermal.c
268
*cur_state += pr->throttling.state;
drivers/acpi/processor_thermal.c
277
struct acpi_processor *pr;
drivers/acpi/processor_thermal.c
284
pr = acpi_driver_data(device);
drivers/acpi/processor_thermal.c
285
if (!pr)
drivers/acpi/processor_thermal.c
288
max_pstate = cpufreq_get_max_state(pr->id);
drivers/acpi/processor_thermal.c
290
if (state > acpi_processor_max_state(pr))
drivers/acpi/processor_thermal.c
294
if (pr->flags.throttling && pr->throttling.state)
drivers/acpi/processor_thermal.c
295
result = acpi_processor_set_throttling(pr, 0, false);
drivers/acpi/processor_thermal.c
296
cpufreq_set_cur_state(pr->id, state);
drivers/acpi/processor_thermal.c
298
cpufreq_set_cur_state(pr->id, max_pstate);
drivers/acpi/processor_thermal.c
299
result = acpi_processor_set_throttling(pr,
drivers/acpi/processor_thermal.c
311
int acpi_processor_thermal_init(struct acpi_processor *pr,
drivers/acpi/processor_thermal.c
316
pr->cdev = thermal_cooling_device_register("Processor", device,
drivers/acpi/processor_thermal.c
318
if (IS_ERR(pr->cdev)) {
drivers/acpi/processor_thermal.c
319
result = PTR_ERR(pr->cdev);
drivers/acpi/processor_thermal.c
324
pr->cdev->id);
drivers/acpi/processor_thermal.c
327
&pr->cdev->device.kobj,
drivers/acpi/processor_thermal.c
335
result = sysfs_create_link(&pr->cdev->device.kobj,
drivers/acpi/processor_thermal.c
339
dev_err(&pr->cdev->device,
drivers/acpi/processor_thermal.c
349
thermal_cooling_device_unregister(pr->cdev);
drivers/acpi/processor_thermal.c
354
void acpi_processor_thermal_exit(struct acpi_processor *pr,
drivers/acpi/processor_thermal.c
357
if (pr->cdev) {
drivers/acpi/processor_thermal.c
359
sysfs_remove_link(&pr->cdev->device.kobj, "device");
drivers/acpi/processor_thermal.c
360
thermal_cooling_device_unregister(pr->cdev);
drivers/acpi/processor_thermal.c
361
pr->cdev = NULL;
drivers/acpi/processor_thermal.c
91
static bool cpufreq_update_thermal_limit(unsigned int cpu, struct acpi_processor *pr)
drivers/acpi/processor_throttling.c
1000
pr->throttling.state = state;
drivers/acpi/processor_throttling.c
1004
acpi_handle_debug(pr->handle,
drivers/acpi/processor_throttling.c
1006
(pr->throttling.states[state].performance ? pr->
drivers/acpi/processor_throttling.c
1012
static int acpi_processor_set_throttling_ptc(struct acpi_processor *pr,
drivers/acpi/processor_throttling.c
1018
if (!pr)
drivers/acpi/processor_throttling.c
1021
if ((state < 0) || (state > (pr->throttling.state_count - 1)))
drivers/acpi/processor_throttling.c
1024
if (!pr->flags.throttling)
drivers/acpi/processor_throttling.c
1027
if (!force && (state == pr->throttling.state))
drivers/acpi/processor_throttling.c
1030
if (state < pr->throttling_platform_limit)
drivers/acpi/processor_throttling.c
1034
ret = acpi_get_throttling_value(pr, state, &value);
drivers/acpi/processor_throttling.c
1036
acpi_write_throttling_state(pr, value);
drivers/acpi/processor_throttling.c
1037
pr->throttling.state = state;
drivers/acpi/processor_throttling.c
1046
struct acpi_processor *pr = arg->pr;
drivers/acpi/processor_throttling.c
1048
return pr->throttling.acpi_processor_set_throttling(pr,
drivers/acpi/processor_throttling.c
1052
static int __acpi_processor_set_throttling(struct acpi_processor *pr,
drivers/acpi/processor_throttling.c
1062
if (!pr)
drivers/acpi/processor_throttling.c
1065
if (!pr->flags.throttling)
drivers/acpi/processor_throttling.c
1068
if ((state < 0) || (state > (pr->throttling.state_count - 1)))
drivers/acpi/processor_throttling.c
1071
if (cpu_is_offline(pr->id)) {
drivers/acpi/processor_throttling.c
1080
p_throttling = &(pr->throttling);
drivers/acpi/processor_throttling.c
1099
arg.pr = pr;
drivers/acpi/processor_throttling.c
1102
ret = call_on_cpu(pr->id, acpi_processor_throttling_fn, &arg,
drivers/acpi/processor_throttling.c
1118
acpi_handle_debug(pr->handle,
drivers/acpi/processor_throttling.c
1127
acpi_handle_debug(pr->handle,
drivers/acpi/processor_throttling.c
1132
arg.pr = match_pr;
drivers/acpi/processor_throttling.c
1135
ret = call_on_cpu(pr->id, acpi_processor_throttling_fn,
drivers/acpi/processor_throttling.c
1154
int acpi_processor_set_throttling(struct acpi_processor *pr, int state,
drivers/acpi/processor_throttling.c
1157
return __acpi_processor_set_throttling(pr, state, force, false);
drivers/acpi/processor_throttling.c
1160
int acpi_processor_get_throttling_info(struct acpi_processor *pr)
drivers/acpi/processor_throttling.c
1165
acpi_handle_debug(pr->handle,
drivers/acpi/processor_throttling.c
1167
pr->throttling.address,
drivers/acpi/processor_throttling.c
1168
pr->throttling.duty_offset,
drivers/acpi/processor_throttling.c
1169
pr->throttling.duty_width);
drivers/acpi/processor_throttling.c
1175
if (acpi_processor_get_throttling_control(pr) ||
drivers/acpi/processor_throttling.c
1176
acpi_processor_get_throttling_states(pr) ||
drivers/acpi/processor_throttling.c
1177
acpi_processor_get_platform_limit(pr)) {
drivers/acpi/processor_throttling.c
1178
pr->throttling.acpi_processor_get_throttling =
drivers/acpi/processor_throttling.c
1180
pr->throttling.acpi_processor_set_throttling =
drivers/acpi/processor_throttling.c
1182
if (acpi_processor_get_fadt_info(pr))
drivers/acpi/processor_throttling.c
1185
pr->throttling.acpi_processor_get_throttling =
drivers/acpi/processor_throttling.c
1187
pr->throttling.acpi_processor_set_throttling =
drivers/acpi/processor_throttling.c
1195
if (acpi_processor_get_tsd(pr)) {
drivers/acpi/processor_throttling.c
1196
pthrottling = &pr->throttling;
drivers/acpi/processor_throttling.c
1198
cpumask_set_cpu(pr->id, pthrottling->shared_cpu_map);
drivers/acpi/processor_throttling.c
1208
acpi_handle_debug(pr->handle,
drivers/acpi/processor_throttling.c
1213
acpi_handle_debug(pr->handle, "Found %d throttling states\n",
drivers/acpi/processor_throttling.c
1214
pr->throttling.state_count);
drivers/acpi/processor_throttling.c
1216
pr->flags.throttling = 1;
drivers/acpi/processor_throttling.c
1224
result = acpi_processor_get_throttling(pr);
drivers/acpi/processor_throttling.c
1228
if (pr->throttling.state) {
drivers/acpi/processor_throttling.c
1229
acpi_handle_debug(pr->handle,
drivers/acpi/processor_throttling.c
1231
pr->throttling.state);
drivers/acpi/processor_throttling.c
1232
result = acpi_processor_set_throttling(pr, 0, false);
drivers/acpi/processor_throttling.c
1239
pr->flags.throttling = 0;
drivers/acpi/processor_throttling.c
172
pr = per_cpu(processors, i);
drivers/acpi/processor_throttling.c
173
if (!pr)
drivers/acpi/processor_throttling.c
181
pthrottling = &(pr->throttling);
drivers/acpi/processor_throttling.c
204
struct acpi_processor *pr;
drivers/acpi/processor_throttling.c
211
pr = per_cpu(processors, cpu);
drivers/acpi/processor_throttling.c
212
if (!pr) {
drivers/acpi/processor_throttling.c
216
if (!pr->flags.throttling) {
drivers/acpi/processor_throttling.c
217
acpi_handle_debug(pr->handle,
drivers/acpi/processor_throttling.c
223
p_throttling = &(pr->throttling);
drivers/acpi/processor_throttling.c
230
p_limit = &pr->limit;
drivers/acpi/processor_throttling.c
235
if (pr->throttling_platform_limit > target_state)
drivers/acpi/processor_throttling.c
236
target_state = pr->throttling_platform_limit;
drivers/acpi/processor_throttling.c
242
acpi_handle_debug(pr->handle,
drivers/acpi/processor_throttling.c
252
acpi_handle_debug(pr->handle,
drivers/acpi/processor_throttling.c
267
static int acpi_processor_get_platform_limit(struct acpi_processor *pr)
drivers/acpi/processor_throttling.c
272
if (!pr)
drivers/acpi/processor_throttling.c
278
status = acpi_evaluate_integer(pr->handle, "_TPC", NULL, &tpc);
drivers/acpi/processor_throttling.c
281
acpi_evaluation_failure_warn(pr->handle, "_TPC", status);
drivers/acpi/processor_throttling.c
287
pr->throttling_platform_limit = (int)tpc;
drivers/acpi/processor_throttling.c
291
int acpi_processor_tstate_has_changed(struct acpi_processor *pr)
drivers/acpi/processor_throttling.c
302
result = acpi_processor_get_platform_limit(pr);
drivers/acpi/processor_throttling.c
308
throttling_limit = pr->throttling_platform_limit;
drivers/acpi/processor_throttling.c
309
if (throttling_limit >= pr->throttling.state_count) {
drivers/acpi/processor_throttling.c
314
current_state = pr->throttling.state;
drivers/acpi/processor_throttling.c
323
limit = &pr->limit;
drivers/acpi/processor_throttling.c
344
return acpi_processor_set_throttling(pr, target_state, false);
drivers/acpi/processor_throttling.c
356
void acpi_processor_reevaluate_tstate(struct acpi_processor *pr,
drivers/acpi/processor_throttling.c
365
pr->flags.throttling = 0;
drivers/acpi/processor_throttling.c
371
if (!pr->throttling.state_count) {
drivers/acpi/processor_throttling.c
375
pr->flags.throttling = 0;
drivers/acpi/processor_throttling.c
378
pr->flags.throttling = 1;
drivers/acpi/processor_throttling.c
385
result = acpi_processor_get_throttling(pr);
drivers/acpi/processor_throttling.c
389
if (pr->throttling.state) {
drivers/acpi/processor_throttling.c
390
result = acpi_processor_set_throttling(pr, 0, false);
drivers/acpi/processor_throttling.c
397
pr->flags.throttling = 0;
drivers/acpi/processor_throttling.c
402
static int acpi_processor_get_throttling_control(struct acpi_processor *pr)
drivers/acpi/processor_throttling.c
411
status = acpi_evaluate_object(pr->handle, "_PTC", NULL, &buffer);
drivers/acpi/processor_throttling.c
414
acpi_evaluation_failure_warn(pr->handle, "_PTC", status);
drivers/acpi/processor_throttling.c
42
struct acpi_processor *pr;
drivers/acpi/processor_throttling.c
440
memcpy(&pr->throttling.control_register, obj.buffer.pointer,
drivers/acpi/processor_throttling.c
457
memcpy(&pr->throttling.status_register, obj.buffer.pointer,
drivers/acpi/processor_throttling.c
460
throttling = &pr->throttling;
drivers/acpi/processor_throttling.c
485
static int acpi_processor_get_throttling_states(struct acpi_processor *pr)
drivers/acpi/processor_throttling.c
495
status = acpi_evaluate_object(pr->handle, "_TSS", NULL, &buffer);
drivers/acpi/processor_throttling.c
498
acpi_evaluation_failure_warn(pr->handle, "_TSS", status);
drivers/acpi/processor_throttling.c
50
static int acpi_processor_get_throttling(struct acpi_processor *pr);
drivers/acpi/processor_throttling.c
51
static int __acpi_processor_set_throttling(struct acpi_processor *pr,
drivers/acpi/processor_throttling.c
510
acpi_handle_debug(pr->handle, "Found %d throttling states\n",
drivers/acpi/processor_throttling.c
513
pr->throttling.state_count = tss->package.count;
drivers/acpi/processor_throttling.c
514
pr->throttling.states_tss =
drivers/acpi/processor_throttling.c
516
if (!pr->throttling.states_tss) {
drivers/acpi/processor_throttling.c
521
for (i = 0; i < pr->throttling.state_count; i++) {
drivers/acpi/processor_throttling.c
524
(struct acpi_processor_tx_tss *)&(pr->throttling.
drivers/acpi/processor_throttling.c
530
acpi_handle_debug(pr->handle, "Extracting state %d\n", i);
drivers/acpi/processor_throttling.c
535
acpi_handle_warn(pr->handle, "Invalid _TSS data: %s\n",
drivers/acpi/processor_throttling.c
538
kfree(pr->throttling.states_tss);
drivers/acpi/processor_throttling.c
545
kfree(pr->throttling.states_tss);
drivers/acpi/processor_throttling.c
559
static int acpi_processor_get_tsd(struct acpi_processor *pr)
drivers/acpi/processor_throttling.c
570
pthrottling = &pr->throttling;
drivers/acpi/processor_throttling.c
573
status = acpi_evaluate_object(pr->handle, "_TSD", NULL, &buffer);
drivers/acpi/processor_throttling.c
576
acpi_evaluation_failure_warn(pr->handle, "_TSD", status);
drivers/acpi/processor_throttling.c
594
pdomain = &(pr->throttling.domain_info);
drivers/acpi/processor_throttling.c
60
struct acpi_processor *pr, *match_pr;
drivers/acpi/processor_throttling.c
619
pthrottling = &pr->throttling;
drivers/acpi/processor_throttling.c
622
cpumask_set_cpu(pr->id, pthrottling->shared_cpu_map);
drivers/acpi/processor_throttling.c
643
static int acpi_processor_get_throttling_fadt(struct acpi_processor *pr)
drivers/acpi/processor_throttling.c
650
if (!pr)
drivers/acpi/processor_throttling.c
653
if (!pr->flags.throttling)
drivers/acpi/processor_throttling.c
663
request_region(pr->throttling.address, 6, "ACPI CPU throttle");
drivers/acpi/processor_throttling.c
665
pr->throttling.state = 0;
drivers/acpi/processor_throttling.c
667
duty_mask = pr->throttling.state_count - 1;
drivers/acpi/processor_throttling.c
669
duty_mask <<= pr->throttling.duty_offset;
drivers/acpi/processor_throttling.c
673
value = inl(pr->throttling.address);
drivers/acpi/processor_throttling.c
681
duty_value >>= pr->throttling.duty_offset;
drivers/acpi/processor_throttling.c
684
state = pr->throttling.state_count - duty_value;
drivers/acpi/processor_throttling.c
687
pr->throttling.state = state;
drivers/acpi/processor_throttling.c
691
acpi_handle_debug(pr->handle,
drivers/acpi/processor_throttling.c
693
state, pr->throttling.states[state].performance);
drivers/acpi/processor_throttling.c
72
pr = per_cpu(processors, i);
drivers/acpi/processor_throttling.c
73
if (!pr)
drivers/acpi/processor_throttling.c
750
static int acpi_read_throttling_status(struct acpi_processor *pr,
drivers/acpi/processor_throttling.c
759
throttling = &pr->throttling;
drivers/acpi/processor_throttling.c
77
pthrottling = &(pr->throttling);
drivers/acpi/processor_throttling.c
782
static int acpi_write_throttling_state(struct acpi_processor *pr,
drivers/acpi/processor_throttling.c
791
throttling = &pr->throttling;
drivers/acpi/processor_throttling.c
815
static int acpi_get_throttling_state(struct acpi_processor *pr,
drivers/acpi/processor_throttling.c
820
for (i = 0; i < pr->throttling.state_count; i++) {
drivers/acpi/processor_throttling.c
822
(struct acpi_processor_tx_tss *)&(pr->throttling.
drivers/acpi/processor_throttling.c
830
static int acpi_get_throttling_value(struct acpi_processor *pr,
drivers/acpi/processor_throttling.c
835
if (state >= 0 && state <= pr->throttling.state_count) {
drivers/acpi/processor_throttling.c
837
(struct acpi_processor_tx_tss *)&(pr->throttling.
drivers/acpi/processor_throttling.c
845
static int acpi_processor_get_throttling_ptc(struct acpi_processor *pr)
drivers/acpi/processor_throttling.c
851
if (!pr)
drivers/acpi/processor_throttling.c
854
if (!pr->flags.throttling)
drivers/acpi/processor_throttling.c
857
pr->throttling.state = 0;
drivers/acpi/processor_throttling.c
860
ret = acpi_read_throttling_status(pr, &value);
drivers/acpi/processor_throttling.c
862
state = acpi_get_throttling_state(pr, value);
drivers/acpi/processor_throttling.c
864
acpi_handle_debug(pr->handle,
drivers/acpi/processor_throttling.c
867
ret = __acpi_processor_set_throttling(pr, state, true,
drivers/acpi/processor_throttling.c
872
pr->throttling.state = state;
drivers/acpi/processor_throttling.c
880
struct acpi_processor *pr = data;
drivers/acpi/processor_throttling.c
882
return pr->throttling.acpi_processor_get_throttling(pr);
drivers/acpi/processor_throttling.c
885
static int acpi_processor_get_throttling(struct acpi_processor *pr)
drivers/acpi/processor_throttling.c
887
if (!pr)
drivers/acpi/processor_throttling.c
890
if (!pr->flags.throttling)
drivers/acpi/processor_throttling.c
899
if (!cpu_online(pr->id))
drivers/acpi/processor_throttling.c
902
return call_on_cpu(pr->id, __acpi_processor_get_throttling, pr, false);
drivers/acpi/processor_throttling.c
905
static int acpi_processor_get_fadt_info(struct acpi_processor *pr)
drivers/acpi/processor_throttling.c
909
if (!pr->throttling.address) {
drivers/acpi/processor_throttling.c
910
acpi_handle_debug(pr->handle, "No throttling register\n");
drivers/acpi/processor_throttling.c
912
} else if (!pr->throttling.duty_width) {
drivers/acpi/processor_throttling.c
913
acpi_handle_debug(pr->handle, "No throttling states\n");
drivers/acpi/processor_throttling.c
917
else if ((pr->throttling.duty_offset + pr->throttling.duty_width) > 4) {
drivers/acpi/processor_throttling.c
922
pr->throttling.state_count = 1 << acpi_gbl_FADT.duty_width;
drivers/acpi/processor_throttling.c
93
pr = per_cpu(processors, i);
drivers/acpi/processor_throttling.c
930
step = (1000 / pr->throttling.state_count);
drivers/acpi/processor_throttling.c
932
for (i = 0; i < pr->throttling.state_count; i++) {
drivers/acpi/processor_throttling.c
933
pr->throttling.states[i].performance = 1000 - step * i;
drivers/acpi/processor_throttling.c
934
pr->throttling.states[i].power = 1000 - step * i;
drivers/acpi/processor_throttling.c
939
static int acpi_processor_set_throttling_fadt(struct acpi_processor *pr,
drivers/acpi/processor_throttling.c
94
if (!pr)
drivers/acpi/processor_throttling.c
946
if (!pr)
drivers/acpi/processor_throttling.c
949
if ((state < 0) || (state > (pr->throttling.state_count - 1)))
drivers/acpi/processor_throttling.c
952
if (!pr->flags.throttling)
drivers/acpi/processor_throttling.c
955
if (!force && (state == pr->throttling.state))
drivers/acpi/processor_throttling.c
958
if (state < pr->throttling_platform_limit)
drivers/acpi/processor_throttling.c
964
duty_value = pr->throttling.state_count - state;
drivers/acpi/processor_throttling.c
966
duty_value <<= pr->throttling.duty_offset;
drivers/acpi/processor_throttling.c
969
duty_mask = pr->throttling.state_count - 1;
drivers/acpi/processor_throttling.c
981
value = inl(pr->throttling.address);
drivers/acpi/processor_throttling.c
984
outl(value, pr->throttling.address);
drivers/acpi/processor_throttling.c
99
pthrottling = &pr->throttling;
drivers/acpi/processor_throttling.c
994
outl(value, pr->throttling.address);
drivers/acpi/processor_throttling.c
997
outl(value, pr->throttling.address);
drivers/acpi/riscv/cpuidle.c
26
struct acpi_processor *pr = per_cpu(processors, cpu);
drivers/acpi/riscv/cpuidle.c
28
if (unlikely(!pr || !pr->flags.has_lpi))
drivers/acpi/riscv/cpuidle.c
34
if (pr->power.count <= 1)
drivers/acpi/riscv/cpuidle.c
37
for (i = 1; i < pr->power.count; i++) {
drivers/acpi/riscv/cpuidle.c
40
lpi = &pr->power.lpi_states[i];
drivers/ata/pata_parport/pata_parport.c
494
struct pi_protocol *pr, int mode, int unit, int delay)
drivers/ata/pata_parport/pata_parport.c
500
struct pi_device_match match = { .parport = parport, .proto = pr };
drivers/ata/pata_parport/pata_parport.c
523
pi->dev.driver = &pr->driver;
drivers/ata/pata_parport/pata_parport.c
533
pi->proto = pr;
drivers/ata/pata_parport/pata_parport.c
585
int pata_parport_register_driver(struct pi_protocol *pr)
drivers/ata/pata_parport/pata_parport.c
591
pr->driver.bus = &pata_parport_bus_type;
drivers/ata/pata_parport/pata_parport.c
592
pr->driver.name = pr->name;
drivers/ata/pata_parport/pata_parport.c
593
error = driver_register(&pr->driver);
drivers/ata/pata_parport/pata_parport.c
598
error = idr_alloc(&protocols, pr, 0, 0, GFP_KERNEL);
drivers/ata/pata_parport/pata_parport.c
600
driver_unregister(&pr->driver);
drivers/ata/pata_parport/pata_parport.c
605
pr_info("pata_parport: protocol %s registered\n", pr->name);
drivers/ata/pata_parport/pata_parport.c
610
pi_init_one(parport, pr, -1, -1, -1);
drivers/ata/pata_parport/pata_parport.c
618
void pata_parport_unregister_driver(struct pi_protocol *pr)
drivers/ata/pata_parport/pata_parport.c
625
if (pr_iter == pr)
drivers/ata/pata_parport/pata_parport.c
630
driver_unregister(&pr->driver);
drivers/ata/pata_parport/pata_parport.c
639
struct pi_protocol *pr, *pr_wanted;
drivers/ata/pata_parport/pata_parport.c
679
idr_for_each_entry(&protocols, pr, pr_num) {
drivers/ata/pata_parport/pata_parport.c
680
if (pr == pr_wanted || !pr_wanted)
drivers/ata/pata_parport/pata_parport.c
681
if (pi_init_one(parport, pr, mode, unit,
drivers/ata/pata_parport/pata_parport.c
729
struct pi_protocol *pr;
drivers/ata/pata_parport/pata_parport.c
742
idr_for_each_entry(&protocols, pr, pr_num)
drivers/ata/pata_parport/pata_parport.c
743
pi_init_one(port, pr, -1, -1, -1);
drivers/ata/pata_parport/pata_parport.h
82
int pata_parport_register_driver(struct pi_protocol *pr);
drivers/ata/pata_parport/pata_parport.h
83
void pata_parport_unregister_driver(struct pi_protocol *pr);
drivers/char/ipmi/ipmi_plat_data.c
105
rv = device_create_managed_software_node(&pdev->dev, pr, NULL);
drivers/char/ipmi/ipmi_plat_data.c
17
struct property_entry pr[6];
drivers/char/ipmi/ipmi_plat_data.c
21
memset(pr, 0, sizeof(pr));
drivers/char/ipmi/ipmi_plat_data.c
35
pr[pidx++] = PROPERTY_ENTRY_U8("ipmi-type", p->type);
drivers/char/ipmi/ipmi_plat_data.c
37
pr[pidx++] = PROPERTY_ENTRY_U16("i2c-addr", p->addr);
drivers/char/ipmi/ipmi_plat_data.c
41
pr[pidx++] = PROPERTY_ENTRY_U8("slave-addr", p->slave_addr);
drivers/char/ipmi/ipmi_plat_data.c
42
pr[pidx++] = PROPERTY_ENTRY_U8("addr-source", p->addr_source);
drivers/char/ipmi/ipmi_plat_data.c
44
pr[pidx++] = PROPERTY_ENTRY_U8("reg-shift", p->regshift);
drivers/char/ipmi/ipmi_plat_data.c
45
pr[pidx++] = PROPERTY_ENTRY_U8("reg-size", p->regsize);
drivers/clk/renesas/rzg2l-cpg.c
1107
u32 nir, nfr, mr, pr, val, setting;
drivers/clk/renesas/rzg2l-cpg.c
1122
pr = 1 << FIELD_GET(RZG3S_DIV_P, val);
drivers/clk/renesas/rzg2l-cpg.c
1124
if (pr > 8)
drivers/clk/renesas/rzg2l-cpg.c
1125
pr = 16;
drivers/clk/renesas/rzg2l-cpg.c
1133
return DIV_ROUND_CLOSEST_ULL(rate, (mr * pr));
drivers/cpufreq/intel_pstate.c
3548
struct acpi_processor *pr = per_cpu(processors, i);
drivers/cpufreq/intel_pstate.c
3550
if (!pr)
drivers/cpufreq/intel_pstate.c
3553
status = acpi_evaluate_object(pr->handle, "_PSS", NULL, &buffer);
drivers/cpufreq/intel_pstate.c
3592
struct acpi_processor *pr = per_cpu(processors, i);
drivers/cpufreq/intel_pstate.c
3594
if (!pr)
drivers/cpufreq/intel_pstate.c
3596
if (acpi_has_method(pr->handle, "_PPC"))
drivers/cpufreq/longhaul.c
288
|| ((pr != NULL) && pr->flags.bm_control))) {
drivers/cpufreq/longhaul.c
302
} else if ((pr != NULL) && pr->flags.bm_control) {
drivers/cpufreq/longhaul.c
338
} else if ((pr != NULL) && pr->flags.bm_control) {
drivers/cpufreq/longhaul.c
70
static struct acpi_processor *pr;
drivers/cpufreq/longhaul.c
861
NULL, (void *)&pr);
drivers/cpufreq/longhaul.c
864
if (pr != NULL && longhaul_version == TYPE_POWERSAVER) {
drivers/cpufreq/longhaul.c
865
cx = &pr->power.states[ACPI_STATE_C3];
drivers/cpufreq/longhaul.c
879
&& ((pr == NULL) || !(pr->flags.bm_control))) {
drivers/cpufreq/pcc-cpufreq.c
255
struct acpi_processor *pr;
drivers/cpufreq/pcc-cpufreq.c
258
pr = per_cpu(processors, cpu);
drivers/cpufreq/pcc-cpufreq.c
261
if (!pr)
drivers/cpufreq/pcc-cpufreq.c
264
status = acpi_evaluate_object(pr->handle, "PCCP", NULL, &buffer);
drivers/dma/pl330.c
2920
int chans, pchs, ch, pr;
drivers/dma/pl330.c
2932
for (pr = 0; pr < pchs; pr++) {
drivers/dma/pl330.c
2933
struct dma_pl330_chan *pch = &pl330->peripherals[pr];
drivers/dma/pl330.c
2938
found = pr;
drivers/extcon/extcon-usbc-cros-ec.c
253
bool pr = false;
drivers/extcon/extcon-usbc-cros-ec.c
277
pr = (role & PD_CTRL_RESP_ROLE_POWER);
drivers/extcon/extcon-usbc-cros-ec.c
288
role, power_type, dr, pr, polarity, mux, dp, hpd);
drivers/extcon/extcon-usbc-cros-ec.c
29
bool pr; /* power role (true if VBUS enabled) */
drivers/extcon/extcon-usbc-cros-ec.c
299
if (force || info->dr != dr || info->pr != pr || info->dp != dp ||
drivers/extcon/extcon-usbc-cros-ec.c
307
info->pr = pr;
drivers/extcon/extcon-usbc-cros-ec.c
322
(union extcon_property_value)(int)pr);
drivers/extcon/extcon-usbc-cros-ec.c
325
(union extcon_property_value)(int)pr);
drivers/extcon/extcon-usbc-cros-ec.c
455
info->pr = false;
drivers/fpga/intel-m10-bmc-sec-update.c
160
DEVICE_ATTR_SEC_REH_RO(pr);
drivers/fpga/intel-m10-bmc-sec-update.c
207
DEVICE_ATTR_SEC_CSK_RO(pr);
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
9870
struct replay_settings *pr = &acrtc_state->stream->link->replay_settings;
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
9876
if (pr->config.replay_supported && !pr->replay_feature_enabled)
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
9886
(psr->psr_feature_enabled || pr->config.replay_supported)) {
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
9906
if (pr->replay_feature_enabled && !pr->replay_allow_active)
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_crtc.c
313
struct replay_settings *pr = &acrtc_state->stream->link->replay_settings;
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_crtc.c
315
pr->config.replay_supported;
drivers/gpu/drm/omapdrm/dss/hdmi5_core.c
374
unsigned int yq, cn, pr;
drivers/gpu/drm/omapdrm/dss/hdmi5_core.c
401
pr = (ptr[4] >> 0) & 0xf;
drivers/gpu/drm/omapdrm/dss/hdmi5_core.c
417
REG_FLD_MOD(base, HDMI_CORE_FC_PRCONF, pr, 3, 0);
drivers/hwtracing/coresight/coresight-platform.c
745
struct acpi_processor *pr;
drivers/hwtracing/coresight/coresight-platform.c
748
pr = per_cpu(processors, i);
drivers/hwtracing/coresight/coresight-platform.c
749
if (pr && pr->handle == handle)
drivers/idle/intel_idle.c
1769
struct acpi_processor *pr = per_cpu(processors, cpu);
drivers/idle/intel_idle.c
1771
if (!pr)
drivers/idle/intel_idle.c
1774
if (acpi_processor_evaluate_cst(pr->handle, cpu, &acpi_state_table))
drivers/input/joystick/xpad.c
501
#define XPAD_XBOX360_VENDOR_PROTOCOL(vend, pr) \
drivers/input/joystick/xpad.c
506
.bInterfaceProtocol = (pr)
drivers/input/joystick/xpad.c
512
#define XPAD_XBOXONE_VENDOR_PROTOCOL(vend, pr) \
drivers/input/joystick/xpad.c
517
.bInterfaceProtocol = (pr)
drivers/input/tablet/hanwang.c
33
#define HANWANG_TABLET_DEVICE(vend, cl, sc, pr) \
drivers/input/tablet/hanwang.c
39
.bInterfaceProtocol = (pr)
drivers/isdn/hardware/mISDN/w6692.c
588
w6692_mode(struct w6692_ch *wch, u32 pr)
drivers/isdn/hardware/mISDN/w6692.c
594
wch->bch.nr, wch->bch.state, pr);
drivers/isdn/hardware/mISDN/w6692.c
595
switch (pr) {
drivers/isdn/hardware/mISDN/w6692.c
625
pr_info("%s: protocol %x not known\n", card->name, pr);
drivers/isdn/hardware/mISDN/w6692.c
628
wch->bch.state = pr;
drivers/isdn/mISDN/hwchannel.c
384
queue_ch_frame(struct mISDNchannel *ch, u_int pr, int id, struct sk_buff *skb)
drivers/isdn/mISDN/hwchannel.c
389
_queue_data(ch, pr, id, 0, NULL, GFP_ATOMIC);
drivers/isdn/mISDN/hwchannel.c
393
hh->prim = pr;
drivers/isdn/mISDN/layer2.c
1002
int pr = -1;
drivers/isdn/mISDN/layer2.c
1012
pr = DL_ESTABLISH_CNF;
drivers/isdn/mISDN/layer2.c
1015
pr = DL_ESTABLISH_IND;
drivers/isdn/mISDN/layer2.c
1024
if (pr != -1)
drivers/isdn/mISDN/layer2.c
1025
l2up_create(l2, pr, 0, NULL);
drivers/isdn/mISDN/layer2.c
679
int pr;
drivers/isdn/mISDN/layer2.c
682
pr = DL_RELEASE_CNF;
drivers/isdn/mISDN/layer2.c
684
pr = DL_RELEASE_IND;
drivers/isdn/mISDN/layer2.c
685
l2up_create(l2, pr, 0, NULL);
drivers/md/dm-cache-policy-smq.c
1381
enum promote_result pr;
drivers/md/dm-cache-policy-smq.c
1401
pr = should_promote(mq, hs_e, data_dir, fast_copy);
drivers/md/dm-cache-policy-smq.c
1402
if (pr != PROMOTE_NOT) {
drivers/md/dm.c
3474
struct dm_pr *pr)
drivers/md/dm.c
3499
ti->type->iterate_devices(ti, fn, pr);
drivers/md/dm.c
3512
struct dm_pr *pr = data;
drivers/md/dm.c
3517
pr->ret = -EOPNOTSUPP;
drivers/md/dm.c
3521
ret = ops->pr_register(dev->bdev, pr->old_key, pr->new_key, pr->flags);
drivers/md/dm.c
3525
if (!pr->ret)
drivers/md/dm.c
3526
pr->ret = ret;
drivers/md/dm.c
3528
if (pr->fail_early)
drivers/md/dm.c
3537
struct dm_pr pr = {
drivers/md/dm.c
3546
ret = dm_call_pr(bdev, __dm_pr_register, &pr);
drivers/md/dm.c
3552
if (!pr.ret)
drivers/md/dm.c
3554
ret = pr.ret;
drivers/md/dm.c
3560
pr.old_key = new_key;
drivers/md/dm.c
3561
pr.new_key = 0;
drivers/md/dm.c
3562
pr.flags = 0;
drivers/md/dm.c
3563
pr.fail_early = false;
drivers/md/dm.c
3564
(void) dm_call_pr(bdev, __dm_pr_register, &pr);
drivers/md/dm.c
3572
struct dm_pr *pr = data;
drivers/md/dm.c
3576
pr->ret = -EOPNOTSUPP;
drivers/md/dm.c
3580
pr->ret = ops->pr_reserve(dev->bdev, pr->old_key, pr->type, pr->flags);
drivers/md/dm.c
3581
if (!pr->ret)
drivers/md/dm.c
3590
struct dm_pr pr = {
drivers/md/dm.c
3599
ret = dm_call_pr(bdev, __dm_pr_reserve, &pr);
drivers/md/dm.c
3603
return pr.ret;
drivers/md/dm.c
3615
struct dm_pr *pr = data;
drivers/md/dm.c
3619
pr->ret = -EOPNOTSUPP;
drivers/md/dm.c
3623
pr->ret = ops->pr_release(dev->bdev, pr->old_key, pr->type);
drivers/md/dm.c
3624
if (pr->ret)
drivers/md/dm.c
3632
struct dm_pr pr = {
drivers/md/dm.c
3639
ret = dm_call_pr(bdev, __dm_pr_release, &pr);
drivers/md/dm.c
3643
return pr.ret;
drivers/md/dm.c
3649
struct dm_pr *pr = data;
drivers/md/dm.c
3653
pr->ret = -EOPNOTSUPP;
drivers/md/dm.c
3657
pr->ret = ops->pr_preempt(dev->bdev, pr->old_key, pr->new_key, pr->type,
drivers/md/dm.c
3658
pr->abort);
drivers/md/dm.c
3659
if (!pr->ret)
drivers/md/dm.c
3668
struct dm_pr pr = {
drivers/md/dm.c
3676
ret = dm_call_pr(bdev, __dm_pr_preempt, &pr);
drivers/md/dm.c
3680
return pr.ret;
drivers/md/dm.c
3709
struct dm_pr *pr = data;
drivers/md/dm.c
3713
pr->ret = -EOPNOTSUPP;
drivers/md/dm.c
3717
pr->ret = ops->pr_read_keys(dev->bdev, pr->read_keys);
drivers/md/dm.c
3718
if (!pr->ret)
drivers/md/dm.c
3726
struct dm_pr pr = {
drivers/md/dm.c
3731
ret = dm_call_pr(bdev, __dm_pr_read_keys, &pr);
drivers/md/dm.c
3735
return pr.ret;
drivers/md/dm.c
3741
struct dm_pr *pr = data;
drivers/md/dm.c
3745
pr->ret = -EOPNOTSUPP;
drivers/md/dm.c
3749
pr->ret = ops->pr_read_reservation(dev->bdev, pr->rsv);
drivers/md/dm.c
3750
if (!pr->ret)
drivers/md/dm.c
3759
struct dm_pr pr = {
drivers/md/dm.c
3764
ret = dm_call_pr(bdev, __dm_pr_read_reservation, &pr);
drivers/md/dm.c
3768
return pr.ret;
drivers/message/fusion/mptscsih.c
2666
MpiRaidActionReply_t *pr;
drivers/message/fusion/mptscsih.c
2704
pr = (MpiRaidActionReply_t *)reply;
drivers/message/fusion/mptscsih.c
2705
if (le16_to_cpu(pr->ActionStatus) ==
drivers/net/can/usb/peak_usb/pcan_usb_pro.c
111
u8 *pr = pcan_msg_init(pm, buffer_addr, buffer_size);
drivers/net/can/usb/peak_usb/pcan_usb_pro.c
113
if (pr) {
drivers/net/can/usb/peak_usb/pcan_usb_pro.c
117
return pr;
drivers/net/can/usb/peak_usb/pcan_usb_pro.c
246
union pcan_usb_pro_rec *pr;
drivers/net/can/usb/peak_usb/pcan_usb_pro.c
278
pr = (union pcan_usb_pro_rec *)pc;
drivers/net/can/usb/peak_usb/pcan_usb_pro.c
279
rec_len = pcan_usb_pro_sizeof_rec[pr->data_type];
drivers/net/can/usb/peak_usb/pcan_usb_pro.c
289
if (pr->data_type != req_data_type)
drivers/net/can/usb/peak_usb/pcan_usb_pro.c
292
pr->data_type);
drivers/net/can/usb/peak_usb/pcan_usb_pro.c
296
(pr->bus_act.channel != req_channel))
drivers/net/can/usb/peak_usb/pcan_usb_pro.c
299
req_data_type, pr->bus_act.channel);
drivers/net/can/usb/peak_usb/pcan_usb_pro.c
718
union pcan_usb_pro_rec *pr = (union pcan_usb_pro_rec *)rec_ptr;
drivers/net/can/usb/peak_usb/pcan_usb_pro.c
719
u16 sizeof_rec = pcan_usb_pro_sizeof_rec[pr->data_type];
drivers/net/can/usb/peak_usb/pcan_usb_pro.c
736
switch (pr->data_type) {
drivers/net/can/usb/peak_usb/pcan_usb_pro.c
741
err = pcan_usb_pro_handle_canmsg(usb_if, &pr->rx_msg);
drivers/net/can/usb/peak_usb/pcan_usb_pro.c
747
err = pcan_usb_pro_handle_error(usb_if, &pr->rx_status);
drivers/net/can/usb/peak_usb/pcan_usb_pro.c
753
pcan_usb_pro_handle_ts(usb_if, &pr->rx_ts);
drivers/net/can/usb/peak_usb/pcan_usb_pro.c
759
pr->data_type, pr->data_type);
drivers/net/ethernet/apple/mace.c
678
while (in_8(&mb->pr) & XMTSV) {
drivers/net/ethernet/apple/mace.h
22
REG(pr); /* poll register */
drivers/net/ethernet/ibm/ehea/ehea_main.c
1251
static int ehea_fill_port_res(struct ehea_port_res *pr)
drivers/net/ethernet/ibm/ehea/ehea_main.c
1254
struct ehea_qp_init_attr *init_attr = &pr->qp->init_attr;
drivers/net/ethernet/ibm/ehea/ehea_main.c
1256
ehea_init_fill_rq1(pr, pr->rq1_skba.len);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1258
ret = ehea_refill_rq2(pr, init_attr->act_nr_rwqes_rq2 - 1);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1260
ret |= ehea_refill_rq3(pr, init_attr->act_nr_rwqes_rq3 - 1);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1268
struct ehea_port_res *pr;
drivers/net/ethernet/ibm/ehea/ehea_main.c
1290
pr = &port->port_res[i];
drivers/net/ethernet/ibm/ehea/ehea_main.c
1291
snprintf(pr->int_send_name, EHEA_IRQ_NAME_SIZE - 1,
drivers/net/ethernet/ibm/ehea/ehea_main.c
1293
ret = ibmebus_request_irq(pr->eq->attr.ist1,
drivers/net/ethernet/ibm/ehea/ehea_main.c
1295
0, pr->int_send_name, pr);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1298
i, pr->eq->attr.ist1);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1303
pr->eq->attr.ist1, i);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1326
struct ehea_port_res *pr;
drivers/net/ethernet/ibm/ehea/ehea_main.c
1332
pr = &port->port_res[i];
drivers/net/ethernet/ibm/ehea/ehea_main.c
1333
ibmebus_free_irq(pr->eq->attr.ist1, pr);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1336
i, pr->eq->attr.ist1);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1394
static int ehea_gen_smrs(struct ehea_port_res *pr)
drivers/net/ethernet/ibm/ehea/ehea_main.c
1397
struct ehea_adapter *adapter = pr->port->adapter;
drivers/net/ethernet/ibm/ehea/ehea_main.c
1399
ret = ehea_gen_smr(adapter, &adapter->mr, &pr->send_mr);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1403
ret = ehea_gen_smr(adapter, &adapter->mr, &pr->recv_mr);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1410
ehea_rem_mr(&pr->send_mr);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1416
static int ehea_rem_smrs(struct ehea_port_res *pr)
drivers/net/ethernet/ibm/ehea/ehea_main.c
1418
if ((ehea_rem_mr(&pr->send_mr)) ||
drivers/net/ethernet/ibm/ehea/ehea_main.c
1419
(ehea_rem_mr(&pr->recv_mr)))
drivers/net/ethernet/ibm/ehea/ehea_main.c
1440
static int ehea_init_port_res(struct ehea_port *port, struct ehea_port_res *pr,
drivers/net/ethernet/ibm/ehea/ehea_main.c
1449
tx_bytes = pr->tx_bytes;
drivers/net/ethernet/ibm/ehea/ehea_main.c
1450
tx_packets = pr->tx_packets;
drivers/net/ethernet/ibm/ehea/ehea_main.c
1451
rx_bytes = pr->rx_bytes;
drivers/net/ethernet/ibm/ehea/ehea_main.c
1452
rx_packets = pr->rx_packets;
drivers/net/ethernet/ibm/ehea/ehea_main.c
1454
memset(pr, 0, sizeof(struct ehea_port_res));
drivers/net/ethernet/ibm/ehea/ehea_main.c
1456
pr->tx_bytes = tx_bytes;
drivers/net/ethernet/ibm/ehea/ehea_main.c
1457
pr->tx_packets = tx_packets;
drivers/net/ethernet/ibm/ehea/ehea_main.c
1458
pr->rx_bytes = rx_bytes;
drivers/net/ethernet/ibm/ehea/ehea_main.c
1459
pr->rx_packets = rx_packets;
drivers/net/ethernet/ibm/ehea/ehea_main.c
1461
pr->port = port;
drivers/net/ethernet/ibm/ehea/ehea_main.c
1463
pr->eq = ehea_create_eq(adapter, eq_type, EHEA_MAX_ENTRIES_EQ, 0);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1464
if (!pr->eq) {
drivers/net/ethernet/ibm/ehea/ehea_main.c
1469
pr->recv_cq = ehea_create_cq(adapter, pr_cfg->max_entries_rcq,
drivers/net/ethernet/ibm/ehea/ehea_main.c
1470
pr->eq->fw_handle,
drivers/net/ethernet/ibm/ehea/ehea_main.c
1472
if (!pr->recv_cq) {
drivers/net/ethernet/ibm/ehea/ehea_main.c
1477
pr->send_cq = ehea_create_cq(adapter, pr_cfg->max_entries_scq,
drivers/net/ethernet/ibm/ehea/ehea_main.c
1478
pr->eq->fw_handle,
drivers/net/ethernet/ibm/ehea/ehea_main.c
1480
if (!pr->send_cq) {
drivers/net/ethernet/ibm/ehea/ehea_main.c
1487
pr->send_cq->attr.act_nr_of_cqes,
drivers/net/ethernet/ibm/ehea/ehea_main.c
1488
pr->recv_cq->attr.act_nr_of_cqes);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1512
init_attr->send_cq_handle = pr->send_cq->fw_handle;
drivers/net/ethernet/ibm/ehea/ehea_main.c
1513
init_attr->recv_cq_handle = pr->recv_cq->fw_handle;
drivers/net/ethernet/ibm/ehea/ehea_main.c
1516
pr->qp = ehea_create_qp(adapter, adapter->pd, init_attr);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1517
if (!pr->qp) {
drivers/net/ethernet/ibm/ehea/ehea_main.c
1531
pr->sq_skba_size = init_attr->act_nr_send_wqes + 1;
drivers/net/ethernet/ibm/ehea/ehea_main.c
1533
ret = ehea_init_q_skba(&pr->sq_skba, pr->sq_skba_size);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1534
ret |= ehea_init_q_skba(&pr->rq1_skba, init_attr->act_nr_rwqes_rq1 + 1);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1535
ret |= ehea_init_q_skba(&pr->rq2_skba, init_attr->act_nr_rwqes_rq2 + 1);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1536
ret |= ehea_init_q_skba(&pr->rq3_skba, init_attr->act_nr_rwqes_rq3 + 1);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1540
pr->swqe_refill_th = init_attr->act_nr_send_wqes / 10;
drivers/net/ethernet/ibm/ehea/ehea_main.c
1541
if (ehea_gen_smrs(pr) != 0) {
drivers/net/ethernet/ibm/ehea/ehea_main.c
1546
atomic_set(&pr->swqe_avail, init_attr->act_nr_send_wqes - 1);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1550
netif_napi_add(pr->port->netdev, &pr->napi, ehea_poll);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1557
vfree(pr->sq_skba.arr);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1558
vfree(pr->rq1_skba.arr);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1559
vfree(pr->rq2_skba.arr);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1560
vfree(pr->rq3_skba.arr);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1561
ehea_destroy_qp(pr->qp);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1562
ehea_destroy_cq(pr->send_cq);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1563
ehea_destroy_cq(pr->recv_cq);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1564
ehea_destroy_eq(pr->eq);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1569
static int ehea_clean_portres(struct ehea_port *port, struct ehea_port_res *pr)
drivers/net/ethernet/ibm/ehea/ehea_main.c
1573
if (pr->qp)
drivers/net/ethernet/ibm/ehea/ehea_main.c
1574
netif_napi_del(&pr->napi);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1576
ret = ehea_destroy_qp(pr->qp);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1579
ehea_destroy_cq(pr->send_cq);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1580
ehea_destroy_cq(pr->recv_cq);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1581
ehea_destroy_eq(pr->eq);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1583
for (i = 0; i < pr->rq1_skba.len; i++)
drivers/net/ethernet/ibm/ehea/ehea_main.c
1584
dev_kfree_skb(pr->rq1_skba.arr[i]);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1586
for (i = 0; i < pr->rq2_skba.len; i++)
drivers/net/ethernet/ibm/ehea/ehea_main.c
1587
dev_kfree_skb(pr->rq2_skba.arr[i]);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1589
for (i = 0; i < pr->rq3_skba.len; i++)
drivers/net/ethernet/ibm/ehea/ehea_main.c
1590
dev_kfree_skb(pr->rq3_skba.arr[i]);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1592
for (i = 0; i < pr->sq_skba.len; i++)
drivers/net/ethernet/ibm/ehea/ehea_main.c
1593
dev_kfree_skb(pr->sq_skba.arr[i]);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1595
vfree(pr->rq1_skba.arr);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1596
vfree(pr->rq2_skba.arr);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1597
vfree(pr->rq3_skba.arr);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1598
vfree(pr->sq_skba.arr);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1599
ret = ehea_rem_smrs(pr);
drivers/net/ethernet/ibm/ehea/ehea_main.c
194
struct ehea_port_res *pr = &port->port_res[l];
drivers/net/ethernet/ibm/ehea/ehea_main.c
197
arr[i++].fwh = pr->qp->fw_handle;
drivers/net/ethernet/ibm/ehea/ehea_main.c
199
arr[i++].fwh = pr->send_cq->fw_handle;
drivers/net/ethernet/ibm/ehea/ehea_main.c
201
arr[i++].fwh = pr->recv_cq->fw_handle;
drivers/net/ethernet/ibm/ehea/ehea_main.c
2021
struct ehea_port_res *pr;
drivers/net/ethernet/ibm/ehea/ehea_main.c
2024
pr = &port->port_res[skb_get_queue_mapping(skb)];
drivers/net/ethernet/ibm/ehea/ehea_main.c
2027
swqe = ehea_get_swqe(pr->qp, &swqe_index);
drivers/net/ethernet/ibm/ehea/ehea_main.c
2029
atomic_dec(&pr->swqe_avail);
drivers/net/ethernet/ibm/ehea/ehea_main.c
203
arr[i++].fwh = pr->eq->fw_handle;
drivers/net/ethernet/ibm/ehea/ehea_main.c
2036
pr->tx_packets++;
drivers/net/ethernet/ibm/ehea/ehea_main.c
2037
pr->tx_bytes += skb->len;
drivers/net/ethernet/ibm/ehea/ehea_main.c
2041
u32 swqe_num = pr->swqe_id_counter;
drivers/net/ethernet/ibm/ehea/ehea_main.c
2045
if (pr->swqe_ll_count >= (sig_iv - 1)) {
drivers/net/ethernet/ibm/ehea/ehea_main.c
2049
pr->swqe_ll_count = 0;
drivers/net/ethernet/ibm/ehea/ehea_main.c
205
arr[i++].fwh = pr->send_mr.handle;
drivers/net/ethernet/ibm/ehea/ehea_main.c
2051
pr->swqe_ll_count += 1;
drivers/net/ethernet/ibm/ehea/ehea_main.c
2055
| EHEA_BMASK_SET(EHEA_WR_ID_COUNT, pr->swqe_id_counter)
drivers/net/ethernet/ibm/ehea/ehea_main.c
2057
| EHEA_BMASK_SET(EHEA_WR_ID_INDEX, pr->sq_skba.index);
drivers/net/ethernet/ibm/ehea/ehea_main.c
2058
pr->sq_skba.arr[pr->sq_skba.index] = skb;
drivers/net/ethernet/ibm/ehea/ehea_main.c
2060
pr->sq_skba.index++;
drivers/net/ethernet/ibm/ehea/ehea_main.c
2061
pr->sq_skba.index &= (pr->sq_skba.len - 1);
drivers/net/ethernet/ibm/ehea/ehea_main.c
2063
lkey = pr->send_mr.lkey;
drivers/net/ethernet/ibm/ehea/ehea_main.c
2067
pr->swqe_id_counter += 1;
drivers/net/ethernet/ibm/ehea/ehea_main.c
207
arr[i++].fwh = pr->recv_mr.handle;
drivers/net/ethernet/ibm/ehea/ehea_main.c
2070
"post swqe on QP %d\n", pr->qp->init_attr.qp_nr);
drivers/net/ethernet/ibm/ehea/ehea_main.c
2079
ehea_post_swqe(pr->qp, swqe);
drivers/net/ethernet/ibm/ehea/ehea_main.c
2081
if (unlikely(atomic_read(&pr->swqe_avail) <= 1)) {
drivers/net/ethernet/ibm/ehea/ehea_main.c
2082
pr->p_stats.queue_stopped++;
drivers/net/ethernet/ibm/ehea/ehea_main.c
2494
struct ehea_port_res *pr = &port->port_res[i];
drivers/net/ethernet/ibm/ehea/ehea_main.c
2495
int swqe_max = pr->sq_skba_size - 2 - pr->swqe_ll_count;
drivers/net/ethernet/ibm/ehea/ehea_main.c
2499
atomic_read(&pr->swqe_avail) >= swqe_max,
drivers/net/ethernet/ibm/ehea/ehea_main.c
2528
struct ehea_port_res *pr = &port->port_res[i];
drivers/net/ethernet/ibm/ehea/ehea_main.c
2529
struct ehea_qp *qp = pr->qp;
drivers/net/ethernet/ibm/ehea/ehea_main.c
2564
dret = ehea_rem_smrs(pr);
drivers/net/ethernet/ibm/ehea/ehea_main.c
2578
static void ehea_update_rqs(struct ehea_qp *orig_qp, struct ehea_port_res *pr)
drivers/net/ethernet/ibm/ehea/ehea_main.c
2583
struct sk_buff **skba_rq2 = pr->rq2_skba.arr;
drivers/net/ethernet/ibm/ehea/ehea_main.c
2584
struct sk_buff **skba_rq3 = pr->rq3_skba.arr;
drivers/net/ethernet/ibm/ehea/ehea_main.c
2586
u32 lkey = pr->recv_mr.lkey;
drivers/net/ethernet/ibm/ehea/ehea_main.c
2628
struct ehea_port_res *pr = &port->port_res[i];
drivers/net/ethernet/ibm/ehea/ehea_main.c
2629
struct ehea_qp *qp = pr->qp;
drivers/net/ethernet/ibm/ehea/ehea_main.c
2631
ret = ehea_gen_smrs(pr);
drivers/net/ethernet/ibm/ehea/ehea_main.c
2637
ehea_update_rqs(qp, pr);
drivers/net/ethernet/ibm/ehea/ehea_main.c
2672
ehea_refill_rq1(pr, pr->rq1_skba.index, 0);
drivers/net/ethernet/ibm/ehea/ehea_main.c
2673
ehea_refill_rq2(pr, 0);
drivers/net/ethernet/ibm/ehea/ehea_main.c
2674
ehea_refill_rq3(pr, 0);
drivers/net/ethernet/ibm/ehea/ehea_main.c
383
static void ehea_refill_rq1(struct ehea_port_res *pr, int index, int nr_of_wqes)
drivers/net/ethernet/ibm/ehea/ehea_main.c
385
struct sk_buff **skb_arr_rq1 = pr->rq1_skba.arr;
drivers/net/ethernet/ibm/ehea/ehea_main.c
386
struct net_device *dev = pr->port->netdev;
drivers/net/ethernet/ibm/ehea/ehea_main.c
387
int max_index_mask = pr->rq1_skba.len - 1;
drivers/net/ethernet/ibm/ehea/ehea_main.c
388
int fill_wqes = pr->rq1_skba.os_skbs + nr_of_wqes;
drivers/net/ethernet/ibm/ehea/ehea_main.c
392
pr->rq1_skba.os_skbs = 0;
drivers/net/ethernet/ibm/ehea/ehea_main.c
396
pr->rq1_skba.index = index;
drivers/net/ethernet/ibm/ehea/ehea_main.c
397
pr->rq1_skba.os_skbs = fill_wqes;
drivers/net/ethernet/ibm/ehea/ehea_main.c
406
pr->rq1_skba.os_skbs = fill_wqes - i;
drivers/net/ethernet/ibm/ehea/ehea_main.c
419
ehea_update_rq1a(pr->qp, adder);
drivers/net/ethernet/ibm/ehea/ehea_main.c
422
static void ehea_init_fill_rq1(struct ehea_port_res *pr, int nr_rq1a)
drivers/net/ethernet/ibm/ehea/ehea_main.c
424
struct sk_buff **skb_arr_rq1 = pr->rq1_skba.arr;
drivers/net/ethernet/ibm/ehea/ehea_main.c
425
struct net_device *dev = pr->port->netdev;
drivers/net/ethernet/ibm/ehea/ehea_main.c
428
if (nr_rq1a > pr->rq1_skba.len) {
drivers/net/ethernet/ibm/ehea/ehea_main.c
439
ehea_update_rq1a(pr->qp, i - 1);
drivers/net/ethernet/ibm/ehea/ehea_main.c
442
static int ehea_refill_rq_def(struct ehea_port_res *pr,
drivers/net/ethernet/ibm/ehea/ehea_main.c
446
struct net_device *dev = pr->port->netdev;
drivers/net/ethernet/ibm/ehea/ehea_main.c
447
struct ehea_qp *qp = pr->qp;
drivers/net/ethernet/ibm/ehea/ehea_main.c
472
netdev_info(pr->port->netdev,
drivers/net/ethernet/ibm/ehea/ehea_main.c
492
rwqe->sg_list[0].l_key = pr->recv_mr.lkey;
drivers/net/ethernet/ibm/ehea/ehea_main.c
509
ehea_update_rq2a(pr->qp, adder);
drivers/net/ethernet/ibm/ehea/ehea_main.c
511
ehea_update_rq3a(pr->qp, adder);
drivers/net/ethernet/ibm/ehea/ehea_main.c
517
static int ehea_refill_rq2(struct ehea_port_res *pr, int nr_of_wqes)
drivers/net/ethernet/ibm/ehea/ehea_main.c
519
return ehea_refill_rq_def(pr, &pr->rq2_skba, 2,
drivers/net/ethernet/ibm/ehea/ehea_main.c
525
static int ehea_refill_rq3(struct ehea_port_res *pr, int nr_of_wqes)
drivers/net/ethernet/ibm/ehea/ehea_main.c
527
return ehea_refill_rq_def(pr, &pr->rq3_skba, 3,
drivers/net/ethernet/ibm/ehea/ehea_main.c
545
struct ehea_port_res *pr)
drivers/net/ethernet/ibm/ehea/ehea_main.c
560
skb_record_rx_queue(skb, pr - &pr->port->port_res[0]);
drivers/net/ethernet/ibm/ehea/ehea_main.c
617
static int ehea_treat_poll_error(struct ehea_port_res *pr, int rq,
drivers/net/ethernet/ibm/ehea/ehea_main.c
624
pr->p_stats.err_tcp_cksum++;
drivers/net/ethernet/ibm/ehea/ehea_main.c
626
pr->p_stats.err_ip_cksum++;
drivers/net/ethernet/ibm/ehea/ehea_main.c
628
pr->p_stats.err_frame_crc++;
drivers/net/ethernet/ibm/ehea/ehea_main.c
632
skb = get_skb_by_index(pr->rq2_skba.arr, pr->rq2_skba.len, cqe);
drivers/net/ethernet/ibm/ehea/ehea_main.c
636
skb = get_skb_by_index(pr->rq3_skba.arr, pr->rq3_skba.len, cqe);
drivers/net/ethernet/ibm/ehea/ehea_main.c
641
if (netif_msg_rx_err(pr->port)) {
drivers/net/ethernet/ibm/ehea/ehea_main.c
643
pr->qp->init_attr.qp_nr);
drivers/net/ethernet/ibm/ehea/ehea_main.c
646
ehea_schedule_port_reset(pr->port);
drivers/net/ethernet/ibm/ehea/ehea_main.c
654
struct ehea_port_res *pr,
drivers/net/ethernet/ibm/ehea/ehea_main.c
657
struct ehea_port *port = pr->port;
drivers/net/ethernet/ibm/ehea/ehea_main.c
658
struct ehea_qp *qp = pr->qp;
drivers/net/ethernet/ibm/ehea/ehea_main.c
661
struct sk_buff **skb_arr_rq1 = pr->rq1_skba.arr;
drivers/net/ethernet/ibm/ehea/ehea_main.c
662
struct sk_buff **skb_arr_rq2 = pr->rq2_skba.arr;
drivers/net/ethernet/ibm/ehea/ehea_main.c
663
struct sk_buff **skb_arr_rq3 = pr->rq3_skba.arr;
drivers/net/ethernet/ibm/ehea/ehea_main.c
664
int skb_arr_rq1_len = pr->rq1_skba.len;
drivers/net/ethernet/ibm/ehea/ehea_main.c
665
int skb_arr_rq2_len = pr->rq2_skba.len;
drivers/net/ethernet/ibm/ehea/ehea_main.c
666
int skb_arr_rq3_len = pr->rq3_skba.len;
drivers/net/ethernet/ibm/ehea/ehea_main.c
701
ehea_fill_skb(dev, skb, cqe, pr);
drivers/net/ethernet/ibm/ehea/ehea_main.c
711
ehea_fill_skb(dev, skb, cqe, pr);
drivers/net/ethernet/ibm/ehea/ehea_main.c
722
ehea_fill_skb(dev, skb, cqe, pr);
drivers/net/ethernet/ibm/ehea/ehea_main.c
732
napi_gro_receive(&pr->napi, skb);
drivers/net/ethernet/ibm/ehea/ehea_main.c
734
pr->p_stats.poll_receive_errors++;
drivers/net/ethernet/ibm/ehea/ehea_main.c
735
port_reset = ehea_treat_poll_error(pr, rq, cqe,
drivers/net/ethernet/ibm/ehea/ehea_main.c
744
pr->rx_packets += processed;
drivers/net/ethernet/ibm/ehea/ehea_main.c
745
pr->rx_bytes += processed_bytes;
drivers/net/ethernet/ibm/ehea/ehea_main.c
747
ehea_refill_rq1(pr, last_wqe_index, processed_rq1);
drivers/net/ethernet/ibm/ehea/ehea_main.c
748
ehea_refill_rq2(pr, processed_rq2);
drivers/net/ethernet/ibm/ehea/ehea_main.c
749
ehea_refill_rq3(pr, processed_rq3);
drivers/net/ethernet/ibm/ehea/ehea_main.c
761
struct ehea_port_res *pr = &port->port_res[i];
drivers/net/ethernet/ibm/ehea/ehea_main.c
762
pr->sq_restart_flag = 0;
drivers/net/ethernet/ibm/ehea/ehea_main.c
774
struct ehea_port_res *pr = &port->port_res[i];
drivers/net/ethernet/ibm/ehea/ehea_main.c
776
swqe = ehea_get_swqe(pr->qp, &swqe_index);
drivers/net/ethernet/ibm/ehea/ehea_main.c
778
atomic_dec(&pr->swqe_avail);
drivers/net/ethernet/ibm/ehea/ehea_main.c
786
ehea_post_swqe(pr->qp, swqe);
drivers/net/ethernet/ibm/ehea/ehea_main.c
789
pr->sq_restart_flag == 0,
drivers/net/ethernet/ibm/ehea/ehea_main.c
794
ehea_schedule_port_reset(pr->port);
drivers/net/ethernet/ibm/ehea/ehea_main.c
801
static struct ehea_cqe *ehea_proc_cqes(struct ehea_port_res *pr, int my_quota)
drivers/net/ethernet/ibm/ehea/ehea_main.c
804
struct ehea_cq *send_cq = pr->send_cq;
drivers/net/ethernet/ibm/ehea/ehea_main.c
810
struct netdev_queue *txq = netdev_get_tx_queue(pr->port->netdev,
drivers/net/ethernet/ibm/ehea/ehea_main.c
811
pr - &pr->port->port_res[0]);
drivers/net/ethernet/ibm/ehea/ehea_main.c
821
pr->sq_restart_flag = 1;
drivers/net/ethernet/ibm/ehea/ehea_main.c
830
if (netif_msg_tx_err(pr->port))
drivers/net/ethernet/ibm/ehea/ehea_main.c
835
ehea_schedule_port_reset(pr->port);
drivers/net/ethernet/ibm/ehea/ehea_main.c
840
if (netif_msg_tx_done(pr->port))
drivers/net/ethernet/ibm/ehea/ehea_main.c
847
skb = pr->sq_skba.arr[index];
drivers/net/ethernet/ibm/ehea/ehea_main.c
849
pr->sq_skba.arr[index] = NULL;
drivers/net/ethernet/ibm/ehea/ehea_main.c
859
atomic_add(swqe_av, &pr->swqe_avail);
drivers/net/ethernet/ibm/ehea/ehea_main.c
862
(atomic_read(&pr->swqe_avail) >= pr->swqe_refill_th))) {
drivers/net/ethernet/ibm/ehea/ehea_main.c
865
(atomic_read(&pr->swqe_avail) >= pr->swqe_refill_th))
drivers/net/ethernet/ibm/ehea/ehea_main.c
870
wake_up(&pr->port->swqe_avail_wq);
drivers/net/ethernet/ibm/ehea/ehea_main.c
879
struct ehea_port_res *pr = container_of(napi, struct ehea_port_res,
drivers/net/ethernet/ibm/ehea/ehea_main.c
881
struct net_device *dev = pr->port->netdev;
drivers/net/ethernet/ibm/ehea/ehea_main.c
887
cqe_skb = ehea_proc_cqes(pr, EHEA_POLL_MAX_CQES);
drivers/net/ethernet/ibm/ehea/ehea_main.c
888
rx += ehea_proc_rwqes(dev, pr, budget - rx);
drivers/net/ethernet/ibm/ehea/ehea_main.c
892
ehea_reset_cq_ep(pr->recv_cq);
drivers/net/ethernet/ibm/ehea/ehea_main.c
893
ehea_reset_cq_ep(pr->send_cq);
drivers/net/ethernet/ibm/ehea/ehea_main.c
894
ehea_reset_cq_n1(pr->recv_cq);
drivers/net/ethernet/ibm/ehea/ehea_main.c
895
ehea_reset_cq_n1(pr->send_cq);
drivers/net/ethernet/ibm/ehea/ehea_main.c
897
cqe = ehea_poll_rq1(pr->qp, &wqe_index);
drivers/net/ethernet/ibm/ehea/ehea_main.c
898
cqe_skb = ehea_poll_cq(pr->send_cq);
drivers/net/ethernet/ibm/ehea/ehea_main.c
906
cqe_skb = ehea_proc_cqes(pr, EHEA_POLL_MAX_CQES);
drivers/net/ethernet/ibm/ehea/ehea_main.c
907
rx += ehea_proc_rwqes(dev, pr, budget - rx);
drivers/net/ethernet/ibm/ehea/ehea_main.c
915
struct ehea_port_res *pr = param;
drivers/net/ethernet/ibm/ehea/ehea_main.c
917
napi_schedule(&pr->napi);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
1886
struct ice_prof_redir_section *pr;
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
1954
pr = sect;
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
1955
src = pr->redir_value;
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
1956
sect_len = le16_to_cpu(pr->count) *
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
1360
struct mlxsw_sp_sb_pr *pr;
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
1363
pr = mlxsw_sp_sb_pr_get(mlxsw_sp, pool_index);
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
1365
pool_info->size = mlxsw_sp_cells_bytes(mlxsw_sp, pr->size);
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
1366
pool_info->threshold_type = (enum devlink_sb_threshold_type) pr->mode;
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
1378
const struct mlxsw_sp_sb_pr *pr;
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
1382
pr = &mlxsw_sp->sb_vals->prs[pool_index];
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
1390
if (pr->freeze_mode && pr->mode != mode) {
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
1395
if (pr->freeze_size && pr->size != size) {
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
1409
struct mlxsw_sp_sb_pr *pr = mlxsw_sp_sb_pr_get(mlxsw_sp, pool_index);
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
1411
if (pr->mode == MLXSW_REG_SBPR_MODE_DYNAMIC)
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
1420
struct mlxsw_sp_sb_pr *pr = mlxsw_sp_sb_pr_get(mlxsw_sp, pool_index);
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
1422
if (pr->mode == MLXSW_REG_SBPR_MODE_DYNAMIC) {
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
188
struct mlxsw_sp_sb_pr *pr;
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
199
pr = mlxsw_sp_sb_pr_get(mlxsw_sp, pool_index);
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
200
pr->mode = mode;
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
201
pr->size = size;
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
943
struct mlxsw_sp_sb_pr *pr = mlxsw_sp_sb_pr_get(mlxsw_sp, pool_index);
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
945
return pr->mode == MLXSW_REG_SBPR_MODE_STATIC;
drivers/net/ovpn/pktid.c
100
if (delta > pr->max_backtrack)
drivers/net/ovpn/pktid.c
101
pr->max_backtrack = delta;
drivers/net/ovpn/pktid.c
102
if (delta < pr->extent) {
drivers/net/ovpn/pktid.c
103
if (pkt_id > pr->id_floor) {
drivers/net/ovpn/pktid.c
104
const unsigned int ri = REPLAY_INDEX(pr->base,
drivers/net/ovpn/pktid.c
106
u8 *p = &pr->history[ri / 8];
drivers/net/ovpn/pktid.c
124
pr->expire = now + PKTID_RECV_EXPIRE;
drivers/net/ovpn/pktid.c
127
spin_unlock_bh(&pr->lock);
drivers/net/ovpn/pktid.c
25
void ovpn_pktid_recv_init(struct ovpn_pktid_recv *pr)
drivers/net/ovpn/pktid.c
27
memset(pr, 0, sizeof(*pr));
drivers/net/ovpn/pktid.c
28
spin_lock_init(&pr->lock);
drivers/net/ovpn/pktid.c
34
int ovpn_pktid_recv(struct ovpn_pktid_recv *pr, u32 pkt_id, u32 pkt_time)
drivers/net/ovpn/pktid.c
43
spin_lock_bh(&pr->lock);
drivers/net/ovpn/pktid.c
46
if (unlikely(time_after_eq(now, pr->expire)))
drivers/net/ovpn/pktid.c
47
pr->id_floor = pr->id;
drivers/net/ovpn/pktid.c
50
if (unlikely(pkt_time != pr->time)) {
drivers/net/ovpn/pktid.c
51
if (pkt_time > pr->time) {
drivers/net/ovpn/pktid.c
53
pr->base = 0;
drivers/net/ovpn/pktid.c
54
pr->extent = 0;
drivers/net/ovpn/pktid.c
55
pr->id = 0;
drivers/net/ovpn/pktid.c
56
pr->time = pkt_time;
drivers/net/ovpn/pktid.c
57
pr->id_floor = 0;
drivers/net/ovpn/pktid.c
65
if (likely(pkt_id == pr->id + 1)) {
drivers/net/ovpn/pktid.c
67
pr->base = REPLAY_INDEX(pr->base, -1);
drivers/net/ovpn/pktid.c
68
pr->history[pr->base / 8] |= (1 << (pr->base % 8));
drivers/net/ovpn/pktid.c
69
if (pr->extent < REPLAY_WINDOW_SIZE)
drivers/net/ovpn/pktid.c
70
++pr->extent;
drivers/net/ovpn/pktid.c
71
pr->id = pkt_id;
drivers/net/ovpn/pktid.c
72
} else if (pkt_id > pr->id) {
drivers/net/ovpn/pktid.c
74
const unsigned int delta = pkt_id - pr->id;
drivers/net/ovpn/pktid.c
79
pr->base = REPLAY_INDEX(pr->base, -delta);
drivers/net/ovpn/pktid.c
80
pr->history[pr->base / 8] |= (1 << (pr->base % 8));
drivers/net/ovpn/pktid.c
81
pr->extent += delta;
drivers/net/ovpn/pktid.c
82
if (pr->extent > REPLAY_WINDOW_SIZE)
drivers/net/ovpn/pktid.c
83
pr->extent = REPLAY_WINDOW_SIZE;
drivers/net/ovpn/pktid.c
85
unsigned int newb = REPLAY_INDEX(pr->base, i);
drivers/net/ovpn/pktid.c
87
pr->history[newb / 8] &= ~BIT(newb % 8);
drivers/net/ovpn/pktid.c
90
pr->base = 0;
drivers/net/ovpn/pktid.c
91
pr->extent = REPLAY_WINDOW_SIZE;
drivers/net/ovpn/pktid.c
92
memset(pr->history, 0, sizeof(pr->history));
drivers/net/ovpn/pktid.c
93
pr->history[0] = 1;
drivers/net/ovpn/pktid.c
95
pr->id = pkt_id;
drivers/net/ovpn/pktid.c
98
const unsigned int delta = pr->id - pkt_id;
drivers/net/ovpn/pktid.h
82
void ovpn_pktid_recv_init(struct ovpn_pktid_recv *pr);
drivers/net/ovpn/pktid.h
84
int ovpn_pktid_recv(struct ovpn_pktid_recv *pr, u32 pkt_id, u32 pkt_time);
drivers/net/usb/cx82310_eth.c
343
#define USB_DEVICE_CLASS(vend, prod, cl, sc, pr) \
drivers/net/usb/cx82310_eth.c
350
.bDeviceProtocol = (pr)
drivers/nvme/host/trace.c
215
u8 pr = cdw10[14];
drivers/nvme/host/trace.c
223
slba, numd, zra, zrasf, zrasf_str, pr);
drivers/nvme/host/zns.c
196
c.zmr.pr = NVME_REPORT_ZONE_PARTIAL;
drivers/nvme/target/admin-cmd.c
862
if (req->ns->pr.enable)
drivers/nvme/target/configfs.c
777
return sysfs_emit(page, "%d\n", to_nvmet_ns(item)->pr.enable);
drivers/nvme/target/configfs.c
795
ns->pr.enable = val;
drivers/nvme/target/core.c
1117
if (req->ns->pr.enable) {
drivers/nvme/target/core.c
1142
if (req->ns->pr.enable) {
drivers/nvme/target/core.c
599
if (ns->pr.enable) {
drivers/nvme/target/core.c
616
if (ns->pr.enable)
drivers/nvme/target/core.c
656
if (ns->pr.enable)
drivers/nvme/target/nvmet.h
129
struct nvmet_pr pr;
drivers/nvme/target/pr.c
1060
if (ns->pr.enable) {
drivers/nvme/target/pr.c
1071
if (ns->pr.enable) {
drivers/nvme/target/pr.c
1091
if (ns->pr.enable) {
drivers/nvme/target/pr.c
1108
ns->pr.holder = NULL;
drivers/nvme/target/pr.c
1109
atomic_set(&ns->pr.generation, 0);
drivers/nvme/target/pr.c
1110
sema_init(&ns->pr.pr_sem, 1);
drivers/nvme/target/pr.c
1111
INIT_LIST_HEAD(&ns->pr.registrant_list);
drivers/nvme/target/pr.c
1112
ns->pr.notify_mask = 0;
drivers/nvme/target/pr.c
1137
struct nvmet_pr *pr = &ns->pr;
drivers/nvme/target/pr.c
1140
list_for_each_entry_safe(reg, tmp, &pr->registrant_list, entry) {
drivers/nvme/target/pr.c
156
static void nvmet_pr_resv_released(struct nvmet_pr *pr, uuid_t *hostid)
drivers/nvme/target/pr.c
158
struct nvmet_ns *ns = nvmet_pr_to_ns(pr);
drivers/nvme/target/pr.c
162
if (test_bit(NVME_PR_NOTIFY_BIT_RESV_RELEASED, &pr->notify_mask))
drivers/nvme/target/pr.c
168
nvmet_pr_find_registrant(pr, &ctrl->hostid)) {
drivers/nvme/target/pr.c
179
static void nvmet_pr_send_event_to_host(struct nvmet_pr *pr, uuid_t *hostid,
drivers/nvme/target/pr.c
182
struct nvmet_ns *ns = nvmet_pr_to_ns(pr);
drivers/nvme/target/pr.c
198
static void nvmet_pr_resv_preempted(struct nvmet_pr *pr, uuid_t *hostid)
drivers/nvme/target/pr.c
200
if (test_bit(NVME_PR_NOTIFY_BIT_RESV_PREEMPTED, &pr->notify_mask))
drivers/nvme/target/pr.c
203
nvmet_pr_send_event_to_host(pr, hostid,
drivers/nvme/target/pr.c
207
static void nvmet_pr_registration_preempted(struct nvmet_pr *pr,
drivers/nvme/target/pr.c
210
if (test_bit(NVME_PR_NOTIFY_BIT_REG_PREEMPTED, &pr->notify_mask))
drivers/nvme/target/pr.c
213
nvmet_pr_send_event_to_host(pr, hostid,
drivers/nvme/target/pr.c
217
static inline void nvmet_pr_set_new_holder(struct nvmet_pr *pr, u8 new_rtype,
drivers/nvme/target/pr.c
22
static inline struct nvmet_ns *nvmet_pr_to_ns(struct nvmet_pr *pr)
drivers/nvme/target/pr.c
221
rcu_assign_pointer(pr->holder, reg);
drivers/nvme/target/pr.c
229
struct nvmet_pr *pr = &req->ns->pr;
drivers/nvme/target/pr.c
237
down(&pr->pr_sem);
drivers/nvme/target/pr.c
238
reg = nvmet_pr_find_registrant(pr, &ctrl->hostid);
drivers/nvme/target/pr.c
24
return container_of(pr, struct nvmet_ns, pr);
drivers/nvme/target/pr.c
250
list_add_tail_rcu(&new->entry, &pr->registrant_list);
drivers/nvme/target/pr.c
253
up(&pr->pr_sem);
drivers/nvme/target/pr.c
257
static void nvmet_pr_unregister_one(struct nvmet_pr *pr,
drivers/nvme/target/pr.c
266
holder = rcu_dereference_protected(pr->holder, 1);
drivers/nvme/target/pr.c
273
first_reg = list_first_or_null_rcu(&pr->registrant_list,
drivers/nvme/target/pr.c
277
rcu_assign_pointer(pr->holder, first_reg);
drivers/nvme/target/pr.c
279
rcu_assign_pointer(pr->holder, NULL);
drivers/nvme/target/pr.c
28
nvmet_pr_find_registrant(struct nvmet_pr *pr, uuid_t *hostid)
drivers/nvme/target/pr.c
283
nvmet_pr_resv_released(pr, &reg->hostid);
drivers/nvme/target/pr.c
295
struct nvmet_pr *pr = &req->ns->pr;
drivers/nvme/target/pr.c
298
down(&pr->pr_sem);
drivers/nvme/target/pr.c
299
list_for_each_entry_rcu(reg, &pr->registrant_list, entry) {
drivers/nvme/target/pr.c
303
nvmet_pr_unregister_one(pr, reg);
drivers/nvme/target/pr.c
308
up(&pr->pr_sem);
drivers/nvme/target/pr.c
319
static u16 nvmet_pr_update_reg_attr(struct nvmet_pr *pr,
drivers/nvme/target/pr.c
32
list_for_each_entry_rcu(reg, &pr->registrant_list, entry) {
drivers/nvme/target/pr.c
328
holder = rcu_dereference_protected(pr->holder, 1);
drivers/nvme/target/pr.c
345
rcu_assign_pointer(pr->holder, new);
drivers/nvme/target/pr.c
357
struct nvmet_pr *pr = &req->ns->pr;
drivers/nvme/target/pr.c
361
down(&pr->pr_sem);
drivers/nvme/target/pr.c
362
list_for_each_entry_rcu(reg, &pr->registrant_list, entry) {
drivers/nvme/target/pr.c
365
status = nvmet_pr_update_reg_attr(pr, reg,
drivers/nvme/target/pr.c
371
up(&pr->pr_sem);
drivers/nvme/target/pr.c
412
atomic_inc(&req->ns->pr.generation);
drivers/nvme/target/pr.c
420
struct nvmet_pr *pr = &req->ns->pr;
drivers/nvme/target/pr.c
423
holder = rcu_dereference_protected(pr->holder, 1);
drivers/nvme/target/pr.c
432
nvmet_pr_set_new_holder(pr, rtype, reg);
drivers/nvme/target/pr.c
465
struct nvmet_pr *pr = &req->ns->pr;
drivers/nvme/target/pr.c
468
list_for_each_entry_safe(reg, tmp, &pr->registrant_list, entry) {
drivers/nvme/target/pr.c
474
nvmet_pr_unregister_one(pr, reg);
drivers/nvme/target/pr.c
476
nvmet_pr_registration_preempted(pr, &hostid);
drivers/nvme/target/pr.c
488
struct nvmet_pr *pr = &req->ns->pr;
drivers/nvme/target/pr.c
491
list_for_each_entry_safe(reg, tmp, &pr->registrant_list, entry) {
drivers/nvme/target/pr.c
497
nvmet_pr_unregister_one(pr, reg);
drivers/nvme/target/pr.c
498
nvmet_pr_registration_preempted(pr, &hostid);
drivers/nvme/target/pr.c
508
struct nvmet_pr *pr = &req->ns->pr;
drivers/nvme/target/pr.c
511
list_for_each_entry_safe(reg, tmp, &pr->registrant_list, entry) {
drivers/nvme/target/pr.c
516
nvmet_pr_unregister_one(pr, reg);
drivers/nvme/target/pr.c
517
nvmet_pr_registration_preempted(pr, &hostid);
drivers/nvme/target/pr.c
537
struct nvmet_pr *pr = &req->ns->pr;
drivers/nvme/target/pr.c
543
holder = rcu_dereference_protected(pr->holder, 1);
drivers/nvme/target/pr.c
557
nvmet_pr_set_new_holder(pr, rtype, reg);
drivers/nvme/target/pr.c
56
if (!req->ns->pr.enable)
drivers/nvme/target/pr.c
566
status = nvmet_pr_update_reg_attr(pr, holder,
drivers/nvme/target/pr.c
569
nvmet_pr_resv_released(pr, &reg->hostid);
drivers/nvme/target/pr.c
577
nvmet_pr_set_new_holder(pr, rtype, reg);
drivers/nvme/target/pr.c
581
nvmet_pr_resv_released(pr, &reg->hostid);
drivers/nvme/target/pr.c
59
WRITE_ONCE(req->ns->pr.notify_mask, mask);
drivers/nvme/target/pr.c
610
up(&ns->pr.pr_sem);
drivers/nvme/target/pr.c
639
atomic_inc(&req->ns->pr.generation);
drivers/nvme/target/pr.c
64
if (ns->pr.enable)
drivers/nvme/target/pr.c
65
WRITE_ONCE(ns->pr.notify_mask, mask);
drivers/nvme/target/pr.c
654
struct nvmet_pr *pr = &req->ns->pr;
drivers/nvme/target/pr.c
676
down(&pr->pr_sem);
drivers/nvme/target/pr.c
677
list_for_each_entry_rcu(reg, &pr->registrant_list, entry) {
drivers/nvme/target/pr.c
693
up(&pr->pr_sem);
drivers/nvme/target/pr.c
705
struct nvmet_pr *pr = &req->ns->pr;
drivers/nvme/target/pr.c
709
holder = rcu_dereference_protected(pr->holder, 1);
drivers/nvme/target/pr.c
717
rcu_assign_pointer(pr->holder, NULL);
drivers/nvme/target/pr.c
721
nvmet_pr_resv_released(pr, &reg->hostid);
drivers/nvme/target/pr.c
729
struct nvmet_pr *pr = &req->ns->pr;
drivers/nvme/target/pr.c
731
rcu_assign_pointer(pr->holder, NULL);
drivers/nvme/target/pr.c
733
list_for_each_entry_safe(reg, tmp, &pr->registrant_list, entry) {
drivers/nvme/target/pr.c
736
nvmet_pr_resv_preempted(pr, &reg->hostid);
drivers/nvme/target/pr.c
740
atomic_inc(&pr->generation);
drivers/nvme/target/pr.c
766
struct nvmet_pr *pr = &req->ns->pr;
drivers/nvme/target/pr.c
787
down(&pr->pr_sem);
drivers/nvme/target/pr.c
788
list_for_each_entry_rcu(reg, &pr->registrant_list, entry) {
drivers/nvme/target/pr.c
796
up(&pr->pr_sem);
drivers/nvme/target/pr.c
81
if (!req->ns->pr.enable)
drivers/nvme/target/pr.c
811
struct nvmet_pr *pr = &req->ns->pr;
drivers/nvme/target/pr.c
836
data->gen = cpu_to_le32(atomic_read(&pr->generation));
drivers/nvme/target/pr.c
84
nvmet_set_result(req, READ_ONCE(req->ns->pr.notify_mask));
drivers/nvme/target/pr.c
841
holder = rcu_dereference(pr->holder);
drivers/nvme/target/pr.c
845
list_for_each_entry_rcu(reg, &pr->registrant_list, entry) {
drivers/nvme/target/pr.c
939
struct nvmet_pr *pr = &ns->pr;
drivers/nvme/target/pr.c
943
holder = rcu_dereference(pr->holder);
drivers/nvme/target/pr.c
966
!nvmet_pr_find_registrant(pr, &ctrl->hostid))
drivers/nvme/target/pr.c
973
!nvmet_pr_find_registrant(pr, &ctrl->hostid))
drivers/nvme/target/trace.c
169
u8 pr = cdw10[14];
drivers/nvme/target/trace.c
177
slba, numd, zra, zrasf, zrasf_str, pr);
drivers/nvme/target/zns.c
174
switch (req->cmd->zmr.pr) {
drivers/nvme/target/zns.c
179
req->error_loc = offsetof(struct nvme_zone_mgmt_recv_cmd, pr);
drivers/nvme/target/zns.c
305
if (req->cmd->zmr.pr)
drivers/of/dynamic.c
107
int of_reconfig_get_state_change(unsigned long action, struct of_reconfig_data *pr)
drivers/of/dynamic.c
116
prop = of_find_property(pr->dn, "status", NULL);
drivers/of/dynamic.c
120
prop = pr->prop;
drivers/of/dynamic.c
123
prop = pr->prop;
drivers/of/dynamic.c
124
old_prop = pr->old_prop;
drivers/of/dynamic.c
188
struct of_reconfig_data pr;
drivers/of/dynamic.c
194
pr.dn = np;
drivers/of/dynamic.c
195
pr.prop = prop;
drivers/of/dynamic.c
196
pr.old_prop = oldprop;
drivers/of/dynamic.c
197
return of_reconfig_notify(action, &pr);
drivers/of/dynamic.c
88
struct of_reconfig_data *pr = p;
drivers/of/dynamic.c
90
of_changeset_action_debug("notify: ", action, pr->dn, pr->prop);
drivers/parport/parport_ip32.c
285
#define __pr_trace(pr, p, fmt, ...) \
drivers/parport/parport_ip32.c
286
pr("%s: %s" fmt "\n", \
drivers/pinctrl/meson/pinctrl-meson.h
146
#define BANK_DS(n, f, l, fi, li, per, peb, pr, pb, dr, db, or, ob, ir, ib, \
drivers/pinctrl/meson/pinctrl-meson.h
156
[MESON_REG_PULL] = { pr, pb }, \
drivers/pinctrl/meson/pinctrl-meson.h
164
#define BANK(n, f, l, fi, li, per, peb, pr, pb, dr, db, or, ob, ir, ib) \
drivers/pinctrl/meson/pinctrl-meson.h
165
BANK_DS(n, f, l, fi, li, per, peb, pr, pb, dr, db, or, ob, ir, ib, 0, 0)
drivers/powercap/arm_scmi_powercap.c
262
static void scmi_powercap_unregister_all_zones(struct scmi_powercap_root *pr)
drivers/powercap/arm_scmi_powercap.c
267
for (i = pr->num_zones - 1; i >= 0; i--) {
drivers/powercap/arm_scmi_powercap.c
268
if (!list_empty(&pr->registered_zones[i])) {
drivers/powercap/arm_scmi_powercap.c
271
list_for_each_entry(spz, &pr->registered_zones[i], node)
drivers/powercap/arm_scmi_powercap.c
296
static int scmi_powercap_register_zone(struct scmi_powercap_root *pr,
drivers/powercap/arm_scmi_powercap.c
314
list_move(&spz->node, &pr->registered_zones[spz->height]);
drivers/powercap/arm_scmi_powercap.c
367
struct scmi_powercap_root *pr)
drivers/powercap/arm_scmi_powercap.c
373
zones_stack = kzalloc_objs(spz, pr->num_zones);
drivers/powercap/arm_scmi_powercap.c
377
spz = list_first_entry_or_null(&pr->scmi_zones,
drivers/powercap/arm_scmi_powercap.c
387
ret = scmi_powercap_register_zone(pr, spz, parent);
drivers/powercap/arm_scmi_powercap.c
397
scmi_powercap_unregister_all_zones(pr);
drivers/powercap/arm_scmi_powercap.c
405
spz = list_first_entry_or_null(&pr->scmi_zones,
drivers/powercap/arm_scmi_powercap.c
421
struct scmi_powercap_root *pr;
drivers/powercap/arm_scmi_powercap.c
435
pr = devm_kzalloc(dev, sizeof(*pr), GFP_KERNEL);
drivers/powercap/arm_scmi_powercap.c
436
if (!pr)
drivers/powercap/arm_scmi_powercap.c
444
pr->num_zones = ret;
drivers/powercap/arm_scmi_powercap.c
446
pr->spzones = devm_kcalloc(dev, pr->num_zones,
drivers/powercap/arm_scmi_powercap.c
447
sizeof(*pr->spzones), GFP_KERNEL);
drivers/powercap/arm_scmi_powercap.c
448
if (!pr->spzones)
drivers/powercap/arm_scmi_powercap.c
452
pr->registered_zones = devm_kcalloc(dev, pr->num_zones,
drivers/powercap/arm_scmi_powercap.c
453
sizeof(*pr->registered_zones),
drivers/powercap/arm_scmi_powercap.c
455
if (!pr->registered_zones)
drivers/powercap/arm_scmi_powercap.c
458
INIT_LIST_HEAD(&pr->scmi_zones);
drivers/powercap/arm_scmi_powercap.c
460
for (i = 0, spz = pr->spzones; i < pr->num_zones; i++, spz++) {
drivers/powercap/arm_scmi_powercap.c
470
spz->spzones = pr->spzones;
drivers/powercap/arm_scmi_powercap.c
472
INIT_LIST_HEAD(&pr->registered_zones[i]);
drivers/powercap/arm_scmi_powercap.c
474
list_add_tail(&spz->node, &pr->scmi_zones);
drivers/powercap/arm_scmi_powercap.c
494
ret = scmi_zones_register(dev, pr);
drivers/powercap/arm_scmi_powercap.c
498
dev_set_drvdata(dev, pr);
drivers/powercap/arm_scmi_powercap.c
506
struct scmi_powercap_root *pr = dev_get_drvdata(dev);
drivers/powercap/arm_scmi_powercap.c
508
scmi_powercap_unregister_all_zones(pr);
drivers/sh/intc/handle.c
112
struct intc_prio_reg *pr = desc->hw.prio_regs;
drivers/sh/intc/handle.c
116
while (pr && enum_id && *reg_idx < desc->hw.nr_prio_regs) {
drivers/sh/intc/handle.c
117
pr = desc->hw.prio_regs + *reg_idx;
drivers/sh/intc/handle.c
119
for (; *fld_idx < ARRAY_SIZE(pr->enum_ids); (*fld_idx)++) {
drivers/sh/intc/handle.c
120
if (pr->enum_ids[*fld_idx] != enum_id)
drivers/sh/intc/handle.c
123
if (pr->set_reg && pr->clr_reg) {
drivers/sh/intc/handle.c
126
reg_e = pr->set_reg;
drivers/sh/intc/handle.c
127
reg_d = pr->clr_reg;
drivers/sh/intc/handle.c
131
if (!pr->set_reg)
drivers/sh/intc/handle.c
133
reg_e = pr->set_reg;
drivers/sh/intc/handle.c
134
reg_d = pr->set_reg;
drivers/sh/intc/handle.c
137
fn += (pr->reg_width >> 3) - 1;
drivers/sh/intc/handle.c
140
BUG_ON(n * pr->field_width > pr->reg_width);
drivers/sh/intc/handle.c
142
bit = pr->reg_width - (n * pr->field_width);
drivers/sh/intc/handle.c
147
pr->field_width, bit);
drivers/soc/qcom/apr.c
104
svc->pr = pr;
drivers/soc/qcom/apr.c
109
spin_lock(&pr->svcs_lock);
drivers/soc/qcom/apr.c
110
id = idr_alloc_cyclic(&pr->svcs_idr, svc, GPR_DYNAMIC_PORT_START,
drivers/soc/qcom/apr.c
115
spin_unlock(&pr->svcs_lock);
drivers/soc/qcom/apr.c
120
spin_unlock(&pr->svcs_lock);
drivers/soc/qcom/apr.c
128
struct packet_router *pr = svc->pr;
drivers/soc/qcom/apr.c
136
ret = rpmsg_trysend(pr->ch, pkt, hdr->pkt_size);
drivers/soc/qcom/apr.c
427
svc->pr = apr;
drivers/soc/qcom/apr.c
79
struct packet_router *gpr = port->pr;
drivers/soc/qcom/apr.c
93
struct packet_router *pr = dev_get_drvdata(gdev->dev.parent);
drivers/staging/media/av7110/dvb_filter.c
23
int dvb_filter_get_ac3info(u8 *mbuf, int count, struct dvb_audio_info *ai, int pr)
drivers/staging/media/av7110/dvb_filter.c
62
if (pr)
drivers/staging/media/av7110/dvb_filter.h
236
int dvb_filter_get_ac3info(u8 *mbuf, int count, struct dvb_audio_info *ai, int pr);
drivers/staging/octeon/octeon-stubs.h
111
u64 pr : 4;
drivers/staging/octeon/octeon-stubs.h
56
u64 pr : 4;
drivers/usb/renesas_usbhs/pipe.h
62
#define usbhs_priv_to_pipeinfo(pr) (&(pr)->pipe_info)
drivers/video/fbdev/omap2/omapfb/dss/hdmi5_core.c
416
unsigned yq, cn, pr;
drivers/video/fbdev/omap2/omapfb/dss/hdmi5_core.c
443
pr = (ptr[4] >> 0) & 0xf;
drivers/video/fbdev/omap2/omapfb/dss/hdmi5_core.c
459
REG_FLD_MOD(base, HDMI_CORE_FC_PRCONF, pr, 3, 0);
drivers/video/screen_info_pci.c
100
if (!pr)
drivers/video/screen_info_pci.c
109
screen_info_lfb_bar = pr - pdev->resource;
drivers/video/screen_info_pci.c
110
screen_info_lfb_offset = r.start - pr->start;
drivers/video/screen_info_pci.c
14
static bool __screen_info_relocation_is_valid(const struct screen_info *si, struct resource *pr)
drivers/video/screen_info_pci.c
18
if (screen_info_lfb_offset > resource_size(pr))
drivers/video/screen_info_pci.c
20
if (size > resource_size(pr))
drivers/video/screen_info_pci.c
22
if (resource_size(pr) - size < screen_info_lfb_offset)
drivers/video/screen_info_pci.c
33
struct resource *pr = &screen_info_lfb_pdev->resource[screen_info_lfb_bar];
drivers/video/screen_info_pci.c
35
if (pr->start != screen_info_lfb_res_start) {
drivers/video/screen_info_pci.c
36
if (__screen_info_relocation_is_valid(si, pr)) {
drivers/video/screen_info_pci.c
41
__screen_info_set_lfb_base(si, pr->start + screen_info_lfb_offset);
drivers/video/screen_info_pci.c
43
&screen_info_lfb_offset, pr);
drivers/video/screen_info_pci.c
78
const struct resource *pr;
drivers/video/screen_info_pci.c
99
pr = pci_find_resource(pdev, &r);
drivers/xen/xen-acpi-processor.c
547
struct acpi_processor *pr;
drivers/xen/xen-acpi-processor.c
550
pr = per_cpu(processors, i);
drivers/xen/xen-acpi-processor.c
552
if (!pr)
drivers/xen/xen-acpi-processor.c
555
pr->performance = perf;
drivers/xen/xen-acpi-processor.c
556
rc = acpi_processor_get_performance_info(pr);
fs/binfmt_elf.c
734
const struct gnu_property *pr;
fs/binfmt_elf.c
745
if (datasz < sizeof(*pr))
fs/binfmt_elf.c
747
pr = (const struct gnu_property *)(data + o);
fs/binfmt_elf.c
748
o += sizeof(*pr);
fs/binfmt_elf.c
749
datasz -= sizeof(*pr);
fs/binfmt_elf.c
751
if (pr->pr_datasz > datasz)
fs/binfmt_elf.c
755
step = round_up(pr->pr_datasz, ELF_GNU_PROPERTY_ALIGN);
fs/binfmt_elf.c
760
if (have_prev_type && pr->pr_type <= *prev_type)
fs/binfmt_elf.c
762
*prev_type = pr->pr_type;
fs/binfmt_elf.c
764
ret = arch_parse_elf_property(pr->pr_type, data + o,
fs/binfmt_elf.c
765
pr->pr_datasz, ELF_COMPAT, arch);
fs/ext4/extents_status.c
1313
struct pending_reservation *pr = NULL;
fs/ext4/extents_status.c
1316
pr = rb_entry(node, struct pending_reservation, rb_node);
fs/ext4/extents_status.c
1317
if (lclu < pr->lclu)
fs/ext4/extents_status.c
1319
else if (lclu > pr->lclu)
fs/ext4/extents_status.c
1322
return pr;
fs/ext4/extents_status.c
1324
if (pr && lclu < pr->lclu)
fs/ext4/extents_status.c
1325
return pr;
fs/ext4/extents_status.c
1326
if (pr && lclu > pr->lclu) {
fs/ext4/extents_status.c
1327
node = rb_next(&pr->rb_node);
fs/ext4/extents_status.c
1354
struct pending_reservation *pr;
fs/ext4/extents_status.c
1447
pr = __pr_tree_search(&tree->root, first_lclu);
fs/ext4/extents_status.c
1448
while (pr && pr->lclu <= last_lclu) {
fs/ext4/extents_status.c
1450
node = rb_next(&pr->rb_node);
fs/ext4/extents_status.c
1451
rb_erase(&pr->rb_node, &tree->root);
fs/ext4/extents_status.c
1452
__free_pending(pr);
fs/ext4/extents_status.c
1455
pr = rb_entry(node, struct pending_reservation,
fs/ext4/extents_status.c
1999
struct pending_reservation *pr;
fs/ext4/extents_status.c
2005
pr = rb_entry(node, struct pending_reservation, rb_node);
fs/ext4/extents_status.c
2006
printk(KERN_DEBUG " %u", pr->lclu);
fs/ext4/extents_status.c
2047
struct pending_reservation *pr = NULL;
fs/ext4/extents_status.c
2053
pr = rb_entry(node, struct pending_reservation, rb_node);
fs/ext4/extents_status.c
2054
if (lclu < pr->lclu)
fs/ext4/extents_status.c
2056
else if (lclu > pr->lclu)
fs/ext4/extents_status.c
2058
else if (lclu == pr->lclu)
fs/ext4/extents_status.c
2059
return pr;
fs/ext4/extents_status.c
2082
struct pending_reservation *pr;
fs/ext4/extents_status.c
2090
pr = rb_entry(parent, struct pending_reservation, rb_node);
fs/ext4/extents_status.c
2092
if (lclu < pr->lclu) {
fs/ext4/extents_status.c
2094
} else if (lclu > pr->lclu) {
fs/ext4/extents_status.c
2103
pr = __alloc_pending(false);
fs/ext4/extents_status.c
2104
if (!pr) {
fs/ext4/extents_status.c
2109
pr = *prealloc;
fs/ext4/extents_status.c
2112
pr->lclu = lclu;
fs/ext4/extents_status.c
2114
rb_link_node(&pr->rb_node, parent, p);
fs/ext4/extents_status.c
2115
rb_insert_color(&pr->rb_node, &tree->root);
fs/ext4/extents_status.c
2134
struct pending_reservation *pr;
fs/ext4/extents_status.c
2137
pr = __get_pending(inode, EXT4_B2C(sbi, lblk));
fs/ext4/extents_status.c
2138
if (pr != NULL) {
fs/ext4/extents_status.c
2140
rb_erase(&pr->rb_node, &tree->root);
fs/ext4/extents_status.c
2141
__free_pending(pr);
fs/ext4/extents_status.c
517
static inline void __free_pending(struct pending_reservation *pr)
fs/ext4/extents_status.c
519
kmem_cache_free(ext4_pending_cachep, pr);
fs/ext4/extents_status.c
916
struct pending_reservation *pr = NULL;
fs/ext4/extents_status.c
946
if ((err1 || err2 || err3 < 0) && revise_pending && !pr)
fs/ext4/extents_status.c
947
pr = __alloc_pending(true);
fs/ext4/extents_status.c
973
err3 = __revise_pending(inode, lblk, len, &pr);
fs/ext4/extents_status.c
976
if (pr) {
fs/ext4/extents_status.c
977
__free_pending(pr);
fs/ext4/extents_status.c
978
pr = NULL;
fs/jfs/jfs_dtree.c
3503
wchar_t *pl, *pr, *kname;
fs/jfs/jfs_dtree.c
3537
for (pl = lkey.name, pr = rkey.name;
fs/jfs/jfs_dtree.c
3538
namlen; pl++, pr++, namlen--, klen++, kname++) {
fs/jfs/jfs_dtree.c
3539
*kname = *pr;
fs/jfs/jfs_dtree.c
3540
if (*pl != *pr) {
fs/jfs/jfs_dtree.c
3548
*kname = *pr;
fs/ntfs3/bitmap.c
1027
pr = NULL;
fs/ntfs3/bitmap.c
1037
pr = cr;
fs/ntfs3/bitmap.c
1046
e = pr ? rb_entry(pr, struct e_node, start.node) : NULL;
fs/ntfs3/bitmap.c
985
const struct rb_node *pr, *cr;
fs/ntfs3/frecord.c
24
struct rb_node *pr = NULL;
fs/ntfs3/frecord.c
29
pr = *p;
fs/ntfs3/frecord.c
30
mi = rb_entry(pr, struct mft_inode, node);
fs/ntfs3/frecord.c
32
p = &pr->rb_left;
fs/ntfs3/frecord.c
34
p = &pr->rb_right;
fs/ntfs3/frecord.c
42
rb_link_node(ins, pr, p);
include/acpi/processor.h
184
int (*acpi_processor_get_throttling) (struct acpi_processor * pr);
include/acpi/processor.h
185
int (*acpi_processor_set_throttling) (struct acpi_processor * pr,
include/acpi/processor.h
270
extern int acpi_processor_get_performance_info(struct acpi_processor *pr);
include/acpi/processor.h
325
void acpi_processor_ppc_has_changed(struct acpi_processor *pr, int event_flag);
include/acpi/processor.h
340
static inline void acpi_processor_ppc_has_changed(struct acpi_processor *pr,
include/acpi/processor.h
366
extern int acpi_cppc_processor_probe(struct acpi_processor *pr);
include/acpi/processor.h
367
extern void acpi_cppc_processor_exit(struct acpi_processor *pr);
include/acpi/processor.h
369
static inline int acpi_cppc_processor_probe(struct acpi_processor *pr)
include/acpi/processor.h
373
static inline void acpi_cppc_processor_exit(struct acpi_processor *pr)
include/acpi/processor.h
384
int acpi_processor_tstate_has_changed(struct acpi_processor *pr);
include/acpi/processor.h
385
int acpi_processor_get_throttling_info(struct acpi_processor *pr);
include/acpi/processor.h
386
extern int acpi_processor_set_throttling(struct acpi_processor *pr,
include/acpi/processor.h
392
extern void acpi_processor_reevaluate_tstate(struct acpi_processor *pr,
include/acpi/processor.h
397
static inline int acpi_processor_tstate_has_changed(struct acpi_processor *pr)
include/acpi/processor.h
402
static inline int acpi_processor_get_throttling_info(struct acpi_processor *pr)
include/acpi/processor.h
407
static inline int acpi_processor_set_throttling(struct acpi_processor *pr,
include/acpi/processor.h
413
static inline void acpi_processor_reevaluate_tstate(struct acpi_processor *pr,
include/acpi/processor.h
421
void acpi_processor_power_init(struct acpi_processor *pr);
include/acpi/processor.h
422
void acpi_processor_power_exit(struct acpi_processor *pr);
include/acpi/processor.h
423
int acpi_processor_power_state_has_changed(struct acpi_processor *pr);
include/acpi/processor.h
424
int acpi_processor_hotplug(struct acpi_processor *pr);
include/acpi/processor.h
432
int acpi_processor_thermal_init(struct acpi_processor *pr,
include/acpi/processor.h
434
void acpi_processor_thermal_exit(struct acpi_processor *pr,
include/crypto/drbg.h
115
bool pr; /* Prediction resistance enabled? */
include/linux/dio.h
143
#define DIO_ENCODE_ID(pr,sec) ((((int)sec & 0xff) << 8) | ((int)pr & 0xff))
include/linux/nvme.h
1214
__u8 pr;
include/linux/soc/qcom/apr.h
130
struct packet_router *pr;
include/linux/usb.h
1060
#define USB_DEVICE_INTERFACE_PROTOCOL(vend, prod, pr) \
include/linux/usb.h
1065
.bInterfaceProtocol = (pr)
include/linux/usb.h
1092
#define USB_DEVICE_INFO(cl, sc, pr) \
include/linux/usb.h
1096
.bDeviceProtocol = (pr)
include/linux/usb.h
1107
#define USB_INTERFACE_INFO(cl, sc, pr) \
include/linux/usb.h
1111
.bInterfaceProtocol = (pr)
include/linux/usb.h
1127
#define USB_DEVICE_AND_INTERFACE_INFO(vend, prod, cl, sc, pr) \
include/linux/usb.h
1134
.bInterfaceProtocol = (pr)
include/linux/usb.h
1149
#define USB_VENDOR_AND_INTERFACE_INFO(vend, cl, sc, pr) \
include/linux/usb.h
1155
.bInterfaceProtocol = (pr)
include/net/libeth/xdp.h
1145
#define __libeth_xdp_tx_init_bulk(bq, pr, d, xdpsqs, num, xsk, ub, un) do { \
include/net/libeth/xdp.h
1152
ub->prog = rcu_dereference(pr); \
include/sound/wavefront.h
452
wavefront_program pr;
lib/tests/kunit_iov_iter.c
1005
KUNIT_EXPECT_EQ(test, iter.iov_offset, pr->to - pr->from);
lib/tests/kunit_iov_iter.c
101
const struct kvec_test_range *pr;
lib/tests/kunit_iov_iter.c
132
for (pr = kvec_test_ranges; pr->from >= 0; pr++)
lib/tests/kunit_iov_iter.c
133
for (i = pr->from; i < pr->to; i++)
lib/tests/kunit_iov_iter.c
151
const struct kvec_test_range *pr;
lib/tests/kunit_iov_iter.c
182
for (pr = kvec_test_ranges; pr->from >= 0; pr++) {
lib/tests/kunit_iov_iter.c
183
for (j = pr->from; j < pr->to; j++) {
lib/tests/kunit_iov_iter.c
222
const struct bvec_test_range *pr)
lib/tests/kunit_iov_iter.c
228
for (i = 0; i < bvmax; i++, pr++) {
lib/tests/kunit_iov_iter.c
229
if (pr->from < 0)
lib/tests/kunit_iov_iter.c
231
KUNIT_ASSERT_LT(test, pr->page, npages);
lib/tests/kunit_iov_iter.c
232
KUNIT_ASSERT_LT(test, pr->page * PAGE_SIZE, bufsize);
lib/tests/kunit_iov_iter.c
233
KUNIT_ASSERT_GE(test, pr->from, 0);
lib/tests/kunit_iov_iter.c
234
KUNIT_ASSERT_GE(test, pr->to, pr->from);
lib/tests/kunit_iov_iter.c
235
KUNIT_ASSERT_LE(test, pr->to, PAGE_SIZE);
lib/tests/kunit_iov_iter.c
237
page = pages[pr->page];
lib/tests/kunit_iov_iter.c
238
if (pr->from == 0 && pr->from != pr->to && page == can_merge) {
lib/tests/kunit_iov_iter.c
240
bvec[i].bv_len += pr->to;
lib/tests/kunit_iov_iter.c
242
bvec_set_page(&bvec[i], page, pr->to - pr->from, pr->from);
lib/tests/kunit_iov_iter.c
245
size += pr->to - pr->from;
lib/tests/kunit_iov_iter.c
246
if ((pr->to & ~PAGE_MASK) == 0)
lib/tests/kunit_iov_iter.c
247
can_merge = page + pr->to / PAGE_SIZE;
lib/tests/kunit_iov_iter.c
260
const struct bvec_test_range *pr;
lib/tests/kunit_iov_iter.c
292
for (pr = bvec_test_ranges; pr->from >= 0; pr++, b++) {
lib/tests/kunit_iov_iter.c
293
u8 *p = scratch + pr->page * PAGE_SIZE;
lib/tests/kunit_iov_iter.c
295
for (i = pr->from; i < pr->to; i++)
lib/tests/kunit_iov_iter.c
314
const struct bvec_test_range *pr;
lib/tests/kunit_iov_iter.c
345
for (pr = bvec_test_ranges; pr->from >= 0; pr++) {
lib/tests/kunit_iov_iter.c
346
size_t patt = pr->page * PAGE_SIZE;
lib/tests/kunit_iov_iter.c
348
for (j = pr->from; j < pr->to; j++) {
lib/tests/kunit_iov_iter.c
418
const struct kvec_test_range *pr;
lib/tests/kunit_iov_iter.c
441
for (pr = kvec_test_ranges; pr->from >= 0; pr++) {
lib/tests/kunit_iov_iter.c
442
size = pr->to - pr->from;
lib/tests/kunit_iov_iter.c
443
KUNIT_ASSERT_LE(test, pr->to, bufsize);
lib/tests/kunit_iov_iter.c
445
iov_iter_folio_queue(&iter, READ, folioq, 0, 0, pr->to);
lib/tests/kunit_iov_iter.c
446
iov_iter_advance(&iter, pr->from);
lib/tests/kunit_iov_iter.c
451
KUNIT_EXPECT_EQ(test, iter.iov_offset, pr->to % PAGE_SIZE);
lib/tests/kunit_iov_iter.c
460
for (pr = kvec_test_ranges; pr->from >= 0; pr++)
lib/tests/kunit_iov_iter.c
461
for (i = pr->from; i < pr->to; i++)
lib/tests/kunit_iov_iter.c
480
const struct kvec_test_range *pr;
lib/tests/kunit_iov_iter.c
503
for (pr = kvec_test_ranges; pr->from >= 0; pr++) {
lib/tests/kunit_iov_iter.c
504
size = pr->to - pr->from;
lib/tests/kunit_iov_iter.c
505
KUNIT_ASSERT_LE(test, pr->to, bufsize);
lib/tests/kunit_iov_iter.c
507
iov_iter_folio_queue(&iter, WRITE, folioq, 0, 0, pr->to);
lib/tests/kunit_iov_iter.c
508
iov_iter_advance(&iter, pr->from);
lib/tests/kunit_iov_iter.c
513
KUNIT_EXPECT_EQ(test, iter.iov_offset, pr->to % PAGE_SIZE);
lib/tests/kunit_iov_iter.c
520
for (pr = kvec_test_ranges; pr->from >= 0; pr++) {
lib/tests/kunit_iov_iter.c
521
for (j = pr->from; j < pr->to; j++) {
lib/tests/kunit_iov_iter.c
580
const struct kvec_test_range *pr;
lib/tests/kunit_iov_iter.c
603
for (pr = kvec_test_ranges; pr->from >= 0; pr++) {
lib/tests/kunit_iov_iter.c
604
size = pr->to - pr->from;
lib/tests/kunit_iov_iter.c
605
KUNIT_ASSERT_LE(test, pr->to, bufsize);
lib/tests/kunit_iov_iter.c
607
iov_iter_xarray(&iter, READ, xarray, pr->from, size);
lib/tests/kunit_iov_iter.c
619
for (pr = kvec_test_ranges; pr->from >= 0; pr++)
lib/tests/kunit_iov_iter.c
620
for (i = pr->from; i < pr->to; i++)
lib/tests/kunit_iov_iter.c
638
const struct kvec_test_range *pr;
lib/tests/kunit_iov_iter.c
661
for (pr = kvec_test_ranges; pr->from >= 0; pr++) {
lib/tests/kunit_iov_iter.c
662
size = pr->to - pr->from;
lib/tests/kunit_iov_iter.c
663
KUNIT_ASSERT_LE(test, pr->to, bufsize);
lib/tests/kunit_iov_iter.c
665
iov_iter_xarray(&iter, WRITE, xarray, pr->from, size);
lib/tests/kunit_iov_iter.c
677
for (pr = kvec_test_ranges; pr->from >= 0; pr++) {
lib/tests/kunit_iov_iter.c
678
for (j = pr->from; j < pr->to; j++) {
lib/tests/kunit_iov_iter.c
701
const struct kvec_test_range *pr;
lib/tests/kunit_iov_iter.c
719
pr = kvec_test_ranges;
lib/tests/kunit_iov_iter.c
720
from = pr->from;
lib/tests/kunit_iov_iter.c
747
while (from == pr->to) {
lib/tests/kunit_iov_iter.c
748
pr++;
lib/tests/kunit_iov_iter.c
749
from = pr->from;
lib/tests/kunit_iov_iter.c
77
const struct kvec_test_range *pr)
lib/tests/kunit_iov_iter.c
781
const struct bvec_test_range *pr;
lib/tests/kunit_iov_iter.c
797
pr = bvec_test_ranges;
lib/tests/kunit_iov_iter.c
798
from = pr->from;
lib/tests/kunit_iov_iter.c
82
for (i = 0; i < kvmax; i++, pr++) {
lib/tests/kunit_iov_iter.c
825
while (from == pr->to) {
lib/tests/kunit_iov_iter.c
826
pr++;
lib/tests/kunit_iov_iter.c
827
from = pr->from;
lib/tests/kunit_iov_iter.c
83
if (pr->from < 0)
lib/tests/kunit_iov_iter.c
831
ix = pr->page + from / PAGE_SIZE;
lib/tests/kunit_iov_iter.c
85
KUNIT_ASSERT_GE(test, pr->to, pr->from);
lib/tests/kunit_iov_iter.c
859
const struct kvec_test_range *pr;
lib/tests/kunit_iov_iter.c
86
KUNIT_ASSERT_LE(test, pr->to, bufsize);
lib/tests/kunit_iov_iter.c
87
kvec[i].iov_base = buffer + pr->from;
lib/tests/kunit_iov_iter.c
875
for (pr = kvec_test_ranges; pr->from >= 0; pr++) {
lib/tests/kunit_iov_iter.c
876
from = pr->from;
lib/tests/kunit_iov_iter.c
877
size = pr->to - from;
lib/tests/kunit_iov_iter.c
878
KUNIT_ASSERT_LE(test, pr->to, bufsize);
lib/tests/kunit_iov_iter.c
88
kvec[i].iov_len = pr->to - pr->from;
lib/tests/kunit_iov_iter.c
880
iov_iter_folio_queue(&iter, WRITE, folioq, 0, 0, pr->to);
lib/tests/kunit_iov_iter.c
89
size += pr->to - pr->from;
lib/tests/kunit_iov_iter.c
938
const struct kvec_test_range *pr;
lib/tests/kunit_iov_iter.c
954
for (pr = kvec_test_ranges; pr->from >= 0; pr++) {
lib/tests/kunit_iov_iter.c
955
from = pr->from;
lib/tests/kunit_iov_iter.c
956
size = pr->to - from;
lib/tests/kunit_iov_iter.c
957
KUNIT_ASSERT_LE(test, pr->to, bufsize);
net/ceph/crush/mapper.c
109
while (work->perm_n <= pr) {
net/ceph/crush/mapper.c
127
s = work->perm[pr];
net/ceph/crush/mapper.c
130
bucket->size, x, r, pr, s);
net/ceph/crush/mapper.c
78
unsigned int pr = r % bucket->size;
net/ceph/crush/mapper.c
87
if (pr == 0) {
net/core/fib_rules.c
243
const struct fib_rule_port_range *pr = nla_data(pattr);
net/core/fib_rules.c
245
if (!fib_rule_port_range_valid(pr))
net/core/fib_rules.c
248
port_range->start = pr->start;
net/core/fib_rules.c
249
port_range->end = pr->end;
net/core/rtnetlink.c
1977
struct nlattr *pr;
net/core/rtnetlink.c
1987
pr = nla_nest_start(skb, IFLA_PROTO_DOWN_REASON);
net/core/rtnetlink.c
1988
if (!pr)
net/core/rtnetlink.c
1992
nla_nest_cancel(skb, pr);
net/core/rtnetlink.c
1996
nla_nest_end(skb, pr);
net/sunrpc/svc.c
1378
int pr, rc;
net/sunrpc/svc.c
1405
for (pr = 0; pr < serv->sv_nprogs; pr++)
net/sunrpc/svc.c
1406
if (rqstp->rq_prog == serv->sv_programs[pr].pg_prog)
net/sunrpc/svc.c
1407
progp = &serv->sv_programs[pr];
sound/isa/wavefront/wavefront_synth.c
1458
if (copy_from_user (&header->hdr.pr, header->hdrptr,
sound/isa/wavefront/wavefront_synth.c
828
if (header->hdr.pr.layer[i].mute) {
sound/isa/wavefront/wavefront_synth.c
829
dev->patch_status[header->hdr.pr.layer[i].patch_number] |=
sound/isa/wavefront/wavefront_synth.c
839
munge_buf ((unsigned char *)&header->hdr.pr, &buf[1], WF_PROGRAM_BYTES);
tools/perf/util/data-convert-bt.c
1184
pr("Adding event '%s' (type %d)\n", name, evsel->core.attr.type);
tools/perf/util/data-convert-bt.c
1208
pr("Failed to add event class into stream.\n");
tools/perf/util/data-convert-bt.c
1255
pr("Adding "#_name" event\n"); \
tools/perf/util/data-convert-bt.c
1263
pr("Failed to add event class '"#_name"' into stream.\n");\
tools/perf/util/data-convert-bt.c
1588
pr("Failed to create CTF clock.\n");
tools/perf/util/data-convert-bt.c
1595
pr("Failed to setup CTF clock.\n");
tools/perf/util/data-convert-bt.c
1602
pr("Failed to create CTF stream class.\n");
tools/perf/util/data-convert-bt.c
1610
pr("Failed to assign CTF clock to stream class.\n");
tools/perf/util/data-convert-bt.c
1629
pr("Failed to assign CTF clock to writer.\n");
tools/perf/util/data-convert-bt.c
689
pr("Flush stream for cpu %d (%u samples)\n",
tools/perf/util/ordered-events.c
154
pr("alloc size %" PRIu64 "B (+%zu), max %" PRIu64 "B\n",
tools/perf/util/ordered-events.c
163
pr("allocation limit reached %" PRIu64 "B\n", oe->max_alloc_size);
tools/perf/util/probe-event.c
1902
char pr;
tools/perf/util/probe-event.c
1934
pr = fmt1_str[0];
tools/perf/util/probe-event.c
1941
pr_debug("Group:%s Event:%s probe:%c\n", tev->group, tev->event, pr);
tools/perf/util/probe-event.c
1943
tp->retprobe = (pr == 'r');
tools/testing/selftests/powerpc/nx-gzip/gzip_vas.c
243
unsigned int pr = 0;
tools/testing/selftests/powerpc/nx-gzip/gzip_vas.c
245
if (pr++ % 100 == 0) {