Symbol: ref
arch/arm/kernel/fiq.c
64
static int fiq_def_op(void *ref, int relinquish)
arch/mips/ath79/clock.c
212
static u32 __init ar934x_get_pll_freq(u32 ref, u32 ref_div, u32 nint, u32 nfrac,
arch/mips/ath79/clock.c
218
t = ref;
arch/mips/ath79/clock.c
223
t = ref;
arch/powerpc/include/asm/pmac_pfunc.h
156
struct kref ref;
arch/powerpc/include/asm/vas.h
100
static inline void put_vas_user_win_ref(struct vas_user_win_ref *ref)
arch/powerpc/include/asm/vas.h
103
put_pid(ref->pid);
arch/powerpc/include/asm/vas.h
104
put_pid(ref->tgid);
arch/powerpc/include/asm/vas.h
105
if (ref->mm)
arch/powerpc/include/asm/vas.h
106
mmdrop(ref->mm);
arch/powerpc/include/asm/vas.h
109
static inline void vas_user_win_add_mm_context(struct vas_user_win_ref *ref)
arch/powerpc/include/asm/vas.h
111
mm_context_add_vas_window(ref->mm);
arch/powerpc/kvm/book3s_64_mmu_radix.c
1036
bool ref = false;
arch/powerpc/kvm/book3s_64_mmu_radix.c
1040
return ref;
arch/powerpc/kvm/book3s_64_mmu_radix.c
1052
ref = true;
arch/powerpc/kvm/book3s_64_mmu_radix.c
1054
return ref;
arch/powerpc/kvm/book3s_64_mmu_radix.c
1065
bool ref = false;
arch/powerpc/kvm/book3s_64_mmu_radix.c
1068
return ref;
arch/powerpc/kvm/book3s_64_mmu_radix.c
1072
ref = true;
arch/powerpc/kvm/book3s_64_mmu_radix.c
1073
return ref;
arch/powerpc/kvm/book3s_hv_nested.c
781
long ref;
arch/powerpc/kvm/book3s_hv_nested.c
788
ref = gp->refcnt;
arch/powerpc/kvm/book3s_hv_nested.c
790
if (ref == 0)
arch/powerpc/kvm/book3s_hv_nested.c
890
long ref;
arch/powerpc/kvm/book3s_hv_nested.c
893
ref = --gp->refcnt;
arch/powerpc/kvm/book3s_hv_nested.c
895
if (ref == 0)
arch/powerpc/perf/imc-pmu.c
1126
struct imc_pmu_ref *ref;
arch/powerpc/perf/imc-pmu.c
1145
ref = &core_imc_refc[core_id];
arch/powerpc/perf/imc-pmu.c
1146
if (!ref)
arch/powerpc/perf/imc-pmu.c
1149
spin_lock(&ref->lock);
arch/powerpc/perf/imc-pmu.c
1150
if (ref->refc == 0) {
arch/powerpc/perf/imc-pmu.c
1153
spin_unlock(&ref->lock);
arch/powerpc/perf/imc-pmu.c
1159
++ref->refc;
arch/powerpc/perf/imc-pmu.c
1160
spin_unlock(&ref->lock);
arch/powerpc/perf/imc-pmu.c
1168
struct imc_pmu_ref *ref;
arch/powerpc/perf/imc-pmu.c
1171
ref = &core_imc_refc[core_id];
arch/powerpc/perf/imc-pmu.c
1172
if (!ref) {
arch/powerpc/perf/imc-pmu.c
1177
spin_lock(&ref->lock);
arch/powerpc/perf/imc-pmu.c
1178
ref->refc--;
arch/powerpc/perf/imc-pmu.c
1179
if (ref->refc == 0) {
arch/powerpc/perf/imc-pmu.c
1182
spin_unlock(&ref->lock);
arch/powerpc/perf/imc-pmu.c
1187
} else if (ref->refc < 0) {
arch/powerpc/perf/imc-pmu.c
1188
ref->refc = 0;
arch/powerpc/perf/imc-pmu.c
1190
spin_unlock(&ref->lock);
arch/powerpc/perf/imc-pmu.c
1361
struct imc_pmu_ref *ref = NULL;
arch/powerpc/perf/imc-pmu.c
1370
ref = &trace_imc_refc[core_id];
arch/powerpc/perf/imc-pmu.c
1371
if (!ref) {
arch/powerpc/perf/imc-pmu.c
1377
spin_lock(&ref->lock);
arch/powerpc/perf/imc-pmu.c
1378
if (ref->refc == 0) {
arch/powerpc/perf/imc-pmu.c
1381
spin_unlock(&ref->lock);
arch/powerpc/perf/imc-pmu.c
1386
++ref->refc;
arch/powerpc/perf/imc-pmu.c
1387
spin_unlock(&ref->lock);
arch/powerpc/perf/imc-pmu.c
1411
struct imc_pmu_ref *ref = NULL;
arch/powerpc/perf/imc-pmu.c
1414
ref = &trace_imc_refc[core_id];
arch/powerpc/perf/imc-pmu.c
1415
if (!ref) {
arch/powerpc/perf/imc-pmu.c
1420
spin_lock(&ref->lock);
arch/powerpc/perf/imc-pmu.c
1421
ref->refc--;
arch/powerpc/perf/imc-pmu.c
1422
if (ref->refc == 0) {
arch/powerpc/perf/imc-pmu.c
1425
spin_unlock(&ref->lock);
arch/powerpc/perf/imc-pmu.c
1429
} else if (ref->refc < 0) {
arch/powerpc/perf/imc-pmu.c
1430
ref->refc = 0;
arch/powerpc/perf/imc-pmu.c
1432
spin_unlock(&ref->lock);
arch/powerpc/perf/imc-pmu.c
361
struct imc_pmu_ref *ref;
arch/powerpc/perf/imc-pmu.c
412
ref = get_nest_pmu_ref(cpu);
arch/powerpc/perf/imc-pmu.c
413
if (!ref)
arch/powerpc/perf/imc-pmu.c
416
ref->refc = 0;
arch/powerpc/perf/imc-pmu.c
462
struct imc_pmu_ref *ref;
arch/powerpc/perf/imc-pmu.c
475
ref = get_nest_pmu_ref(event->cpu);
arch/powerpc/perf/imc-pmu.c
476
if (!ref)
arch/powerpc/perf/imc-pmu.c
480
spin_lock(&ref->lock);
arch/powerpc/perf/imc-pmu.c
481
if (ref->refc == 0) {
arch/powerpc/perf/imc-pmu.c
492
spin_unlock(&ref->lock);
arch/powerpc/perf/imc-pmu.c
495
ref->refc--;
arch/powerpc/perf/imc-pmu.c
496
if (ref->refc == 0) {
arch/powerpc/perf/imc-pmu.c
500
spin_unlock(&ref->lock);
arch/powerpc/perf/imc-pmu.c
504
} else if (ref->refc < 0) {
arch/powerpc/perf/imc-pmu.c
506
ref->refc = 0;
arch/powerpc/perf/imc-pmu.c
508
spin_unlock(&ref->lock);
arch/powerpc/perf/imc-pmu.c
517
struct imc_pmu_ref *ref;
arch/powerpc/perf/imc-pmu.c
569
ref = get_nest_pmu_ref(event->cpu);
arch/powerpc/perf/imc-pmu.c
570
if (!ref)
arch/powerpc/perf/imc-pmu.c
573
spin_lock(&ref->lock);
arch/powerpc/perf/imc-pmu.c
574
if (ref->refc == 0) {
arch/powerpc/perf/imc-pmu.c
578
spin_unlock(&ref->lock);
arch/powerpc/perf/imc-pmu.c
584
++ref->refc;
arch/powerpc/perf/imc-pmu.c
585
spin_unlock(&ref->lock);
arch/powerpc/perf/imc-pmu.c
677
struct imc_pmu_ref *ref;
arch/powerpc/perf/imc-pmu.c
717
ref = &core_imc_refc[core_id];
arch/powerpc/perf/imc-pmu.c
718
if (!ref)
arch/powerpc/perf/imc-pmu.c
721
ref->refc = 0;
arch/powerpc/perf/imc-pmu.c
764
struct imc_pmu_ref *ref;
arch/powerpc/perf/imc-pmu.c
777
ref = &core_imc_refc[core_id];
arch/powerpc/perf/imc-pmu.c
778
if (!ref)
arch/powerpc/perf/imc-pmu.c
781
spin_lock(&ref->lock);
arch/powerpc/perf/imc-pmu.c
782
if (ref->refc == 0) {
arch/powerpc/perf/imc-pmu.c
793
spin_unlock(&ref->lock);
arch/powerpc/perf/imc-pmu.c
796
ref->refc--;
arch/powerpc/perf/imc-pmu.c
797
if (ref->refc == 0) {
arch/powerpc/perf/imc-pmu.c
801
spin_unlock(&ref->lock);
arch/powerpc/perf/imc-pmu.c
805
} else if (ref->refc < 0) {
arch/powerpc/perf/imc-pmu.c
807
ref->refc = 0;
arch/powerpc/perf/imc-pmu.c
809
spin_unlock(&ref->lock);
arch/powerpc/perf/imc-pmu.c
820
struct imc_pmu_ref *ref;
arch/powerpc/perf/imc-pmu.c
847
ref = &core_imc_refc[core_id];
arch/powerpc/perf/imc-pmu.c
848
if (!ref)
arch/powerpc/perf/imc-pmu.c
857
spin_lock(&ref->lock);
arch/powerpc/perf/imc-pmu.c
858
if (ref->refc == 0) {
arch/powerpc/perf/imc-pmu.c
862
spin_unlock(&ref->lock);
arch/powerpc/perf/imc-pmu.c
868
++ref->refc;
arch/powerpc/perf/imc-pmu.c
869
spin_unlock(&ref->lock);
arch/powerpc/platforms/cell/spufs/sched.c
363
static struct spu *ctx_location(struct spu *ref, int offset, int node)
arch/powerpc/platforms/cell/spufs/sched.c
369
list_for_each_entry(spu, ref->aff_list.prev, aff_list) {
arch/powerpc/platforms/cell/spufs/sched.c
377
list_for_each_entry_reverse(spu, ref->aff_list.next, aff_list) {
arch/powerpc/platforms/powermac/pfunc_core.c
544
struct kref ref;
arch/powerpc/platforms/powermac/pfunc_core.c
553
struct pmf_device *dev = container_of(kref, struct pmf_device, ref);
arch/powerpc/platforms/powermac/pfunc_core.c
559
kref_put(&dev->ref, pmf_release_device);
arch/powerpc/platforms/powermac/pfunc_core.c
564
kref_get(&dev->ref);
arch/powerpc/platforms/powermac/pfunc_core.c
650
kref_init(&func->ref);
arch/powerpc/platforms/powermac/pfunc_core.c
728
kref_init(&dev->ref);
arch/powerpc/platforms/powermac/pfunc_core.c
753
kref_get(&func->ref);
arch/powerpc/platforms/powermac/pfunc_core.c
761
container_of(kref, struct pmf_function, ref);
arch/powerpc/platforms/powermac/pfunc_core.c
768
kref_put(&func->ref, pmf_release_function);
arch/x86/coco/sev/vc-handle.c
487
void __user *ref;
arch/x86/coco/sev/vc-handle.c
489
ref = insn_get_addr_ref(&ctxt->insn, ctxt->regs);
arch/x86/coco/sev/vc-handle.c
490
if (ref == (void __user *)-1L)
arch/x86/coco/sev/vc-handle.c
495
res = vc_slow_virt_to_phys(ghcb, ctxt, (unsigned long)ref, &paddr);
arch/x86/events/core.c
1987
EVENT_ATTR(ref-cycles, REF_CPU_CYCLES );
arch/x86/events/intel/core.c
3860
if (!atomic_read(&era->ref) || era->config == reg->config) {
arch/x86/events/intel/core.c
3890
atomic_inc(&era->ref);
arch/x86/events/intel/core.c
3929
atomic_dec(&era->ref);
arch/x86/events/intel/uncore.c
200
if (!atomic_read(&er->ref) ||
arch/x86/events/intel/uncore.c
202
atomic_inc(&er->ref);
arch/x86/events/intel/uncore.c
235
atomic_dec(&er->ref);
arch/x86/events/intel/uncore.h
159
atomic_t ref;
arch/x86/events/intel/uncore_nhmex.c
1006
if (!atomic_read(&er->ref) || er->config == reg1->config) {
arch/x86/events/intel/uncore_nhmex.c
1007
atomic_inc(&er->ref);
arch/x86/events/intel/uncore_nhmex.c
1017
if (!__BITS_VALUE(atomic_read(&er->ref), idx - 2, 8) ||
arch/x86/events/intel/uncore_nhmex.c
1019
atomic_add(1 << ((idx - 2) * 8), &er->ref);
arch/x86/events/intel/uncore_nhmex.c
1025
if (!atomic_read(&er->ref) ||
arch/x86/events/intel/uncore_nhmex.c
1029
atomic_inc(&er->ref);
arch/x86/events/intel/uncore_nhmex.c
1081
atomic_sub(1 << ((idx - 2) * 8), &er->ref);
arch/x86/events/intel/uncore_nhmex.c
1083
atomic_dec(&er->ref);
arch/x86/events/intel/uncore_nhmex.c
565
if (!atomic_read(&er->ref) || er->config == config) {
arch/x86/events/intel/uncore_nhmex.c
566
atomic_inc(&er->ref);
arch/x86/events/intel/uncore_nhmex.c
592
if (__BITS_VALUE(atomic_read(&er->ref), idx, 8)) {
arch/x86/events/intel/uncore_nhmex.c
599
if (!atomic_read(&er->ref) || !((er->config ^ config) & mask)) {
arch/x86/events/intel/uncore_nhmex.c
600
atomic_add(1 << (idx * 8), &er->ref);
arch/x86/events/intel/uncore_nhmex.c
622
atomic_dec(&er->ref);
arch/x86/events/intel/uncore_nhmex.c
628
atomic_sub(1 << (idx * 8), &er->ref);
arch/x86/events/intel/uncore_snbep.c
1081
if (!__BITS_VALUE(atomic_read(&er->ref), idx, 8) ||
arch/x86/events/intel/uncore_snbep.c
1083
atomic_add(1 << (idx * 8), &er->ref);
arch/x86/events/intel/uncore_snbep.c
1115
atomic_sub(1 << (reg1->idx * 8), &er->ref);
arch/x86/events/intel/uncore_snbep.c
928
atomic_sub(1 << (i * 6), &er->ref);
arch/x86/events/intel/uncore_snbep.c
954
if (!__BITS_VALUE(atomic_read(&er->ref), i, 6) ||
arch/x86/events/intel/uncore_snbep.c
956
atomic_add(1 << (i * 6), &er->ref);
arch/x86/events/intel/uncore_snbep.c
975
atomic_sub(1 << (i * 6), &er->ref);
arch/x86/events/perf_event.h
204
atomic_t ref; /* reference count */
arch/x86/kernel/cpu/sgx/encl.c
700
void sgx_encl_release(struct kref *ref)
arch/x86/kernel/cpu/sgx/encl.c
702
struct sgx_encl *encl = container_of(ref, struct sgx_encl, refcount);
arch/x86/kernel/cpu/sgx/encl.h
107
void sgx_encl_release(struct kref *ref);
arch/x86/kernel/tsc_sync.c
183
struct tsc_adjust *ref, *cur = this_cpu_ptr(&tsc_adjust);
arch/x86/kernel/tsc_sync.c
217
ref = per_cpu_ptr(&tsc_adjust, refcpu);
arch/x86/kernel/tsc_sync.c
222
if (bootval != ref->bootval)
arch/x86/kernel/tsc_sync.c
231
if (bootval != ref->adjusted) {
arch/x86/kernel/tsc_sync.c
232
cur->adjusted = ref->adjusted;
arch/x86/kernel/tsc_sync.c
233
wrmsrq(MSR_IA32_TSC_ADJUST, ref->adjusted);
block/bfq-cgroup.c
319
refcount_inc(&bfqg->ref);
block/bfq-cgroup.c
324
if (refcount_dec_and_test(&bfqg->ref))
block/bfq-cgroup.c
525
refcount_set(&bfqg->ref, 1);
block/bfq-cgroup.c
659
bfqq->ref++;
block/bfq-iosched.c
1203
return bfqq->ref - bfqq->entity.allocated -
block/bfq-iosched.c
2810
new_bfqq->ref += process_refs;
block/bfq-iosched.c
5336
bfq_log_bfqq(bfqq->bfqd, bfqq, "put_queue: %p %d", bfqq, bfqq->ref);
block/bfq-iosched.c
5338
bfqq->ref--;
block/bfq-iosched.c
5339
if (bfqq->ref)
block/bfq-iosched.c
5441
bfq_log_bfqq(bfqd, bfqq, "exit_bfqq: %p, %d", bfqq, bfqq->ref);
block/bfq-iosched.c
5602
bfqq->ref = 0;
block/bfq-iosched.c
5811
last_bfqq_created->ref++;
block/bfq-iosched.c
5869
bfqq->ref++; /*
block/bfq-iosched.c
5877
bfqq, bfqq->ref);
block/bfq-iosched.c
5882
bfqq->ref++; /* get a process reference to this queue */
block/bfq-iosched.c
6162
new_bfqq->ref++;
block/bfq-iosched.c
6956
bfqq->ref++;
block/bfq-iosched.c
6959
rq, bfqq, bfqq->ref);
block/bfq-iosched.c
7083
bfqq, bfqq->ref);
block/bfq-iosched.c
7219
bfqd->oom_bfqq.ref++;
block/bfq-iosched.c
966
bfqq->ref++;
block/bfq-iosched.h
1006
refcount_t ref;
block/bfq-iosched.h
248
int ref;
block/bfq-wf2q.c
1578
int ref = in_serv_bfqq->ref;
block/bfq-wf2q.c
1580
if (ref == 1)
block/bfq-wf2q.c
529
bfqq->ref++;
block/bfq-wf2q.c
531
bfqq, bfqq->ref);
block/blk-cgroup.c
194
static void blkg_release(struct percpu_ref *ref)
block/blk-cgroup.c
196
struct blkcg_gq *blkg = container_of(ref, struct blkcg_gq, refcnt);
block/blk-core.c
374
static void blk_queue_usage_counter_release(struct percpu_ref *ref)
block/blk-core.c
377
container_of(ref, struct request_queue, q_usage_counter);
block/blk-zoned.c
1206
refcount_inc(&zwplug->ref);
block/blk-zoned.c
1851
refcount_inc(&zwplug->ref);
block/blk-zoned.c
2331
zwp_ref = refcount_read(&zwplug->ref);
block/blk-zoned.c
562
refcount_inc_not_zero(&zwplug->ref)) {
block/blk-zoned.c
592
if (refcount_dec_and_test(&zwplug->ref)) {
block/blk-zoned.c
625
if (refcount_read(&zwplug->ref) > 2)
block/blk-zoned.c
70
refcount_t ref;
block/blk-zoned.c
700
refcount_set(&zwplug->ref, 2);
block/blk.h
650
((unsigned int) atomic_read(&(req->ref)) + 127u <= 127u)
block/blk.h
654
return atomic_inc_not_zero(&req->ref);
block/blk.h
660
return atomic_dec_and_test(&req->ref);
block/blk.h
665
atomic_set(&req->ref, value);
block/blk.h
670
return atomic_read(&req->ref);
block/fops.c
124
atomic_t ref;
block/fops.c
143
if (atomic_dec_and_test(&dio->ref)) {
block/fops.c
189
atomic_set(&dio->ref, 1);
block/fops.c
260
atomic_inc(&dio->ref);
block/t10-pi.c
257
u64 ref, seed;
block/t10-pi.c
264
ref = get_unaligned_be48(pi->ref_tag);
block/t10-pi.c
266
if (ref != seed) {
block/t10-pi.c
268
iter->disk_name, seed, ref);
block/t10-pi.c
324
u64 ref = get_unaligned_be48(pi->ref_tag);
block/t10-pi.c
326
if (ref == virt)
block/t10-pi.c
361
u64 ref = get_unaligned_be48(pi->ref_tag);
block/t10-pi.c
363
if (ref == ref_tag)
crypto/asymmetric_keys/asymmetric_type.c
104
return ERR_CAST(ref);
crypto/asymmetric_keys/asymmetric_type.c
108
key = key_ref_to_ptr(ref);
crypto/asymmetric_keys/asymmetric_type.c
50
key_ref_t ref;
crypto/asymmetric_keys/asymmetric_type.c
90
ref = keyring_search(make_key_ref(keyring, 1),
crypto/asymmetric_keys/asymmetric_type.c
92
if (IS_ERR(ref))
crypto/asymmetric_keys/asymmetric_type.c
93
pr_debug("Request for key '%s' err %ld\n", req, PTR_ERR(ref));
crypto/asymmetric_keys/asymmetric_type.c
96
if (IS_ERR(ref)) {
crypto/asymmetric_keys/asymmetric_type.c
97
switch (PTR_ERR(ref)) {
drivers/accel/amdxdna/aie2_ctx.c
32
static void aie2_job_release(struct kref *ref)
drivers/accel/amdxdna/aie2_ctx.c
36
job = container_of(ref, struct amdxdna_sched_job, refcnt);
drivers/accel/amdxdna/amdxdna_gem.c
161
static void amdxdna_umap_release(struct kref *ref)
drivers/accel/amdxdna/amdxdna_gem.c
163
struct amdxdna_umap *mapp = container_of(ref, struct amdxdna_umap, refcnt);
drivers/accel/ethosu/ethosu_job.c
158
static void ethosu_job_cleanup(struct kref *ref)
drivers/accel/ethosu/ethosu_job.c
160
struct ethosu_job *job = container_of(ref, struct ethosu_job,
drivers/accel/habanalabs/common/command_submission.c
135
static void hl_sob_reset(struct kref *ref)
drivers/accel/habanalabs/common/command_submission.c
137
struct hl_hw_sob *hw_sob = container_of(ref, struct hl_hw_sob,
drivers/accel/habanalabs/common/command_submission.c
148
void hl_sob_reset_error(struct kref *ref)
drivers/accel/habanalabs/common/command_submission.c
150
struct hl_hw_sob *hw_sob = container_of(ref, struct hl_hw_sob,
drivers/accel/habanalabs/common/command_submission.c
265
static void cs_job_do_release(struct kref *ref)
drivers/accel/habanalabs/common/command_submission.c
267
struct hl_cs_job *job = container_of(ref, struct hl_cs_job, refcount);
drivers/accel/habanalabs/common/command_submission.c
55
static void cs_do_release(struct kref *ref);
drivers/accel/habanalabs/common/command_submission.c
694
static void cs_do_release(struct kref *ref)
drivers/accel/habanalabs/common/command_submission.c
696
struct hl_cs *cs = container_of(ref, struct hl_cs, refcount);
drivers/accel/habanalabs/common/context.c
131
void hl_ctx_do_release(struct kref *ref)
drivers/accel/habanalabs/common/context.c
135
ctx = container_of(ref, struct hl_ctx, refcount);
drivers/accel/habanalabs/common/context.c
30
void hl_encaps_release_handle_and_put_ctx(struct kref *ref)
drivers/accel/habanalabs/common/context.c
33
container_of(ref, struct hl_cs_encaps_sig_handle, refcount);
drivers/accel/habanalabs/common/context.c
38
static void hl_encaps_release_handle_and_put_sob(struct kref *ref)
drivers/accel/habanalabs/common/context.c
41
container_of(ref, struct hl_cs_encaps_sig_handle, refcount);
drivers/accel/habanalabs/common/context.c
46
void hl_encaps_release_handle_and_put_sob_ctx(struct kref *ref)
drivers/accel/habanalabs/common/context.c
49
container_of(ref, struct hl_cs_encaps_sig_handle, refcount);
drivers/accel/habanalabs/common/device.c
460
static void hpriv_release(struct kref *ref)
drivers/accel/habanalabs/common/device.c
467
hpriv = container_of(ref, struct hl_fpriv, refcount);
drivers/accel/habanalabs/common/habanalabs.h
3830
void hl_ctx_do_release(struct kref *ref);
drivers/accel/habanalabs/common/habanalabs.h
3879
void hl_sob_reset_error(struct kref *ref);
drivers/accel/habanalabs/common/habanalabs.h
4082
void hl_encaps_release_handle_and_put_ctx(struct kref *ref);
drivers/accel/habanalabs/common/habanalabs.h
4083
void hl_encaps_release_handle_and_put_sob_ctx(struct kref *ref);
drivers/accel/habanalabs/common/memory.c
288
static void dram_pg_pool_do_release(struct kref *ref)
drivers/accel/habanalabs/common/memory.c
290
struct hl_vm *vm = container_of(ref, struct hl_vm,
drivers/accel/habanalabs/gaudi/gaudi.c
1123
static void gaudi_sob_group_hw_reset(struct kref *ref)
drivers/accel/habanalabs/gaudi/gaudi.c
1126
container_of(ref, struct gaudi_hw_sob_group, kref);
drivers/accel/habanalabs/gaudi/gaudi.c
1137
static void gaudi_sob_group_reset_error(struct kref *ref)
drivers/accel/habanalabs/gaudi/gaudi.c
1140
container_of(ref, struct gaudi_hw_sob_group, kref);
drivers/accel/ivpu/ivpu_drv.c
124
file_priv->ctx.id, kref_read(&file_priv->ref));
drivers/accel/ivpu/ivpu_drv.c
127
kref_put(&file_priv->ref, file_priv_release);
drivers/accel/ivpu/ivpu_drv.c
250
kref_init(&file_priv->ref);
drivers/accel/ivpu/ivpu_drv.c
74
kref_get(&file_priv->ref);
drivers/accel/ivpu/ivpu_drv.c
77
file_priv->ctx.id, kref_read(&file_priv->ref));
drivers/accel/ivpu/ivpu_drv.c
97
static void file_priv_release(struct kref *ref)
drivers/accel/ivpu/ivpu_drv.c
99
struct ivpu_file_priv *file_priv = container_of(ref, struct ivpu_file_priv, ref);
drivers/accel/ivpu/ivpu_drv.h
184
struct kref ref;
drivers/accel/qaic/qaic_control.c
290
static void free_wrapper(struct kref *ref)
drivers/accel/qaic/qaic_control.c
292
struct wrapper_msg *wrapper = container_of(ref, struct wrapper_msg, ref_count);
drivers/accel/rocket/rocket_job.c
230
static void rocket_job_cleanup(struct kref *ref)
drivers/accel/rocket/rocket_job.c
232
struct rocket_job *job = container_of(ref, struct rocket_job,
drivers/acpi/pptt.c
100
ref += resource;
drivers/acpi/pptt.c
102
return fetch_pptt_subtable(table_hdr, *ref);
drivers/acpi/pptt.c
94
u32 *ref;
drivers/acpi/pptt.c
99
ref = ACPI_ADD_PTR(u32, node, sizeof(struct acpi_pptt_processor));
drivers/android/binder.c
1023
struct binder_ref *ref;
drivers/android/binder.c
1026
ref = rb_entry(n, struct binder_ref, rb_node_desc);
drivers/android/binder.c
1028
if (desc < ref->data.desc) {
drivers/android/binder.c
1030
} else if (desc > ref->data.desc) {
drivers/android/binder.c
1032
} else if (need_strong_ref && !ref->data.strong) {
drivers/android/binder.c
1036
return ref;
drivers/android/binder.c
1045
struct binder_ref *ref;
drivers/android/binder.c
1051
ref = rb_entry(n, struct binder_ref, rb_node_desc);
drivers/android/binder.c
1052
if (ref->data.desc > desc)
drivers/android/binder.c
1054
desc = ref->data.desc + 1;
drivers/android/binder.c
1122
struct binder_ref *ref;
drivers/android/binder.c
1132
ref = rb_entry(parent, struct binder_ref, rb_node_node);
drivers/android/binder.c
1134
if (node < ref->node)
drivers/android/binder.c
1136
else if (node > ref->node)
drivers/android/binder.c
1139
return ref;
drivers/android/binder.c
1159
ref = rb_entry(parent, struct binder_ref, rb_node_desc);
drivers/android/binder.c
1161
if (new_ref->data.desc < ref->data.desc)
drivers/android/binder.c
1163
else if (new_ref->data.desc > ref->data.desc)
drivers/android/binder.c
1182
static void binder_cleanup_ref_olocked(struct binder_ref *ref)
drivers/android/binder.c
1184
struct dbitmap *dmap = &ref->proc->dmap;
drivers/android/binder.c
1189
ref->proc->pid, ref->data.debug_id, ref->data.desc,
drivers/android/binder.c
1190
ref->node->debug_id);
drivers/android/binder.c
1193
dbitmap_clear_bit(dmap, ref->data.desc);
drivers/android/binder.c
1194
rb_erase(&ref->rb_node_desc, &ref->proc->refs_by_desc);
drivers/android/binder.c
1195
rb_erase(&ref->rb_node_node, &ref->proc->refs_by_node);
drivers/android/binder.c
1197
binder_node_inner_lock(ref->node);
drivers/android/binder.c
1198
if (ref->data.strong)
drivers/android/binder.c
1199
binder_dec_node_nilocked(ref->node, 1, 1);
drivers/android/binder.c
1201
hlist_del(&ref->node_entry);
drivers/android/binder.c
1202
delete_node = binder_dec_node_nilocked(ref->node, 0, 1);
drivers/android/binder.c
1203
binder_node_inner_unlock(ref->node);
drivers/android/binder.c
1213
ref->node = NULL;
drivers/android/binder.c
1216
if (ref->death) {
drivers/android/binder.c
1219
ref->proc->pid, ref->data.debug_id,
drivers/android/binder.c
1220
ref->data.desc);
drivers/android/binder.c
1221
binder_dequeue_work(ref->proc, &ref->death->work);
drivers/android/binder.c
1225
if (ref->freeze) {
drivers/android/binder.c
1226
binder_dequeue_work(ref->proc, &ref->freeze->work);
drivers/android/binder.c
1243
static int binder_inc_ref_olocked(struct binder_ref *ref, int strong,
drivers/android/binder.c
1249
if (ref->data.strong == 0) {
drivers/android/binder.c
1250
ret = binder_inc_node(ref->node, 1, 1, target_list);
drivers/android/binder.c
1254
ref->data.strong++;
drivers/android/binder.c
1256
if (ref->data.weak == 0) {
drivers/android/binder.c
1257
ret = binder_inc_node(ref->node, 0, 1, target_list);
drivers/android/binder.c
1261
ref->data.weak++;
drivers/android/binder.c
1275
static bool binder_dec_ref_olocked(struct binder_ref *ref, int strong)
drivers/android/binder.c
1278
if (ref->data.strong == 0) {
drivers/android/binder.c
1280
ref->proc->pid, ref->data.debug_id,
drivers/android/binder.c
1281
ref->data.desc, ref->data.strong,
drivers/android/binder.c
1282
ref->data.weak);
drivers/android/binder.c
1285
ref->data.strong--;
drivers/android/binder.c
1286
if (ref->data.strong == 0)
drivers/android/binder.c
1287
binder_dec_node(ref->node, strong, 1);
drivers/android/binder.c
1289
if (ref->data.weak == 0) {
drivers/android/binder.c
1291
ref->proc->pid, ref->data.debug_id,
drivers/android/binder.c
1292
ref->data.desc, ref->data.strong,
drivers/android/binder.c
1293
ref->data.weak);
drivers/android/binder.c
1296
ref->data.weak--;
drivers/android/binder.c
1298
if (ref->data.strong == 0 && ref->data.weak == 0) {
drivers/android/binder.c
1299
binder_cleanup_ref_olocked(ref);
drivers/android/binder.c
1322
struct binder_ref *ref;
drivers/android/binder.c
1325
ref = binder_get_ref_olocked(proc, desc, need_strong_ref);
drivers/android/binder.c
1326
if (!ref)
drivers/android/binder.c
1328
node = ref->node;
drivers/android/binder.c
1335
*rdata = ref->data;
drivers/android/binder.c
1352
static void binder_free_ref(struct binder_ref *ref)
drivers/android/binder.c
1354
if (ref->node)
drivers/android/binder.c
1355
binder_free_node(ref->node);
drivers/android/binder.c
1356
kfree(ref->death);
drivers/android/binder.c
1357
kfree(ref->freeze);
drivers/android/binder.c
1358
kfree(ref);
drivers/android/binder.c
1398
struct binder_ref *ref;
drivers/android/binder.c
1402
ref = binder_get_ref_olocked(proc, desc, strong);
drivers/android/binder.c
1403
if (!ref) {
drivers/android/binder.c
1408
ret = binder_inc_ref_olocked(ref, strong, NULL);
drivers/android/binder.c
1410
delete_ref = binder_dec_ref_olocked(ref, strong);
drivers/android/binder.c
1413
*rdata = ref->data;
drivers/android/binder.c
1417
binder_free_ref(ref);
drivers/android/binder.c
1464
struct binder_ref *ref;
drivers/android/binder.c
1469
ref = binder_get_ref_for_node_olocked(proc, node, NULL);
drivers/android/binder.c
1470
if (!ref) {
drivers/android/binder.c
1472
new_ref = kzalloc_obj(*ref);
drivers/android/binder.c
1476
ref = binder_get_ref_for_node_olocked(proc, node, new_ref);
drivers/android/binder.c
1478
ret = binder_inc_ref_olocked(ref, strong, target_list);
drivers/android/binder.c
1479
*rdata = ref->data;
drivers/android/binder.c
1480
if (ret && ref == new_ref) {
drivers/android/binder.c
1489
ref = NULL;
drivers/android/binder.c
1493
if (new_ref && ref != new_ref)
drivers/android/binder.c
3190
struct binder_ref *ref;
drivers/android/binder.c
3200
ref = binder_get_ref_olocked(proc, tr->target.handle,
drivers/android/binder.c
3202
if (ref) {
drivers/android/binder.c
3204
ref->node, &target_proc,
drivers/android/binder.c
3927
struct binder_ref *ref;
drivers/android/binder.c
3933
ref = binder_get_ref_olocked(proc, handle_cookie->handle, false);
drivers/android/binder.c
3934
if (!ref) {
drivers/android/binder.c
3942
binder_node_lock(ref->node);
drivers/android/binder.c
3943
if (ref->freeze) {
drivers/android/binder.c
3946
binder_node_unlock(ref->node);
drivers/android/binder.c
3956
ref->freeze = freeze;
drivers/android/binder.c
3958
if (ref->node->proc) {
drivers/android/binder.c
3959
binder_inner_proc_lock(ref->node->proc);
drivers/android/binder.c
3960
freeze->is_frozen = ref->node->proc->is_frozen;
drivers/android/binder.c
3961
binder_inner_proc_unlock(ref->node->proc);
drivers/android/binder.c
3969
binder_node_unlock(ref->node);
drivers/android/binder.c
3980
struct binder_ref *ref;
drivers/android/binder.c
3983
ref = binder_get_ref_olocked(proc, handle_cookie->handle, false);
drivers/android/binder.c
3984
if (!ref) {
drivers/android/binder.c
3991
binder_node_lock(ref->node);
drivers/android/binder.c
3993
if (!ref->freeze) {
drivers/android/binder.c
3996
binder_node_unlock(ref->node);
drivers/android/binder.c
4000
freeze = ref->freeze;
drivers/android/binder.c
4007
binder_node_unlock(ref->node);
drivers/android/binder.c
4011
ref->freeze = NULL;
drivers/android/binder.c
4028
binder_node_unlock(ref->node);
drivers/android/binder.c
4383
struct binder_ref *ref;
drivers/android/binder.c
4413
ref = binder_get_ref_olocked(proc, target, false);
drivers/android/binder.c
4414
if (ref == NULL) {
drivers/android/binder.c
4432
(u64)cookie, ref->data.debug_id,
drivers/android/binder.c
4433
ref->data.desc, ref->data.strong,
drivers/android/binder.c
4434
ref->data.weak, ref->node->debug_id);
drivers/android/binder.c
4436
binder_node_lock(ref->node);
drivers/android/binder.c
4438
if (ref->death) {
drivers/android/binder.c
4441
binder_node_unlock(ref->node);
drivers/android/binder.c
4449
ref->death = death;
drivers/android/binder.c
4450
if (ref->node->proc == NULL) {
drivers/android/binder.c
4451
ref->death->work.type = BINDER_WORK_DEAD_BINDER;
drivers/android/binder.c
4455
&ref->death->work, &proc->todo);
drivers/android/binder.c
4460
if (ref->death == NULL) {
drivers/android/binder.c
4463
binder_node_unlock(ref->node);
drivers/android/binder.c
4467
death = ref->death;
drivers/android/binder.c
4473
binder_node_unlock(ref->node);
drivers/android/binder.c
4477
ref->death = NULL;
drivers/android/binder.c
4500
binder_node_unlock(ref->node);
drivers/android/binder.c
5318
if (refcount_dec_and_test(&device->ref)) {
drivers/android/binder.c
5629
struct binder_ref *ref;
drivers/android/binder.c
5641
hlist_for_each_entry(ref, &node->refs, node_entry) {
drivers/android/binder.c
5648
binder_inner_proc_lock(ref->proc);
drivers/android/binder.c
5649
if (!ref->freeze) {
drivers/android/binder.c
5650
binder_inner_proc_unlock(ref->proc);
drivers/android/binder.c
5653
ref->freeze->work.type = BINDER_WORK_FROZEN_BINDER;
drivers/android/binder.c
5654
if (list_empty(&ref->freeze->work.entry)) {
drivers/android/binder.c
5655
ref->freeze->is_frozen = is_frozen;
drivers/android/binder.c
5656
binder_enqueue_work_ilocked(&ref->freeze->work, &ref->proc->todo);
drivers/android/binder.c
5657
binder_wakeup_proc_ilocked(ref->proc);
drivers/android/binder.c
5659
if (ref->freeze->sent && ref->freeze->is_frozen != is_frozen)
drivers/android/binder.c
5660
ref->freeze->resend = true;
drivers/android/binder.c
5661
ref->freeze->is_frozen = is_frozen;
drivers/android/binder.c
5663
binder_inner_proc_unlock(ref->proc);
drivers/android/binder.c
6085
refcount_inc(&binder_dev->ref);
drivers/android/binder.c
6197
struct binder_ref *ref;
drivers/android/binder.c
6227
hlist_for_each_entry(ref, &node->refs, node_entry) {
drivers/android/binder.c
6235
binder_inner_proc_lock(ref->proc);
drivers/android/binder.c
6236
if (!ref->death) {
drivers/android/binder.c
6237
binder_inner_proc_unlock(ref->proc);
drivers/android/binder.c
6243
BUG_ON(!list_empty(&ref->death->work.entry));
drivers/android/binder.c
6244
ref->death->work.type = BINDER_WORK_DEAD_BINDER;
drivers/android/binder.c
6245
binder_enqueue_work_ilocked(&ref->death->work,
drivers/android/binder.c
6246
&ref->proc->todo);
drivers/android/binder.c
6247
binder_wakeup_proc_ilocked(ref->proc);
drivers/android/binder.c
6248
binder_inner_proc_unlock(ref->proc);
drivers/android/binder.c
6325
struct binder_ref *ref;
drivers/android/binder.c
6327
ref = rb_entry(n, struct binder_ref, rb_node_desc);
drivers/android/binder.c
6329
binder_cleanup_ref_olocked(ref);
drivers/android/binder.c
6331
binder_free_ref(ref);
drivers/android/binder.c
6532
struct binder_ref *ref;
drivers/android/binder.c
6550
hlist_for_each_entry(ref, &node->refs, node_entry)
drivers/android/binder.c
6551
seq_printf(m, " %d", ref->proc->pid);
drivers/android/binder.c
6563
struct binder_ref *ref)
drivers/android/binder.c
6565
binder_node_lock(ref->node);
drivers/android/binder.c
6567
ref->data.debug_id, ref->data.desc,
drivers/android/binder.c
6568
ref->node->proc ? "" : "dead ",
drivers/android/binder.c
6569
ref->node->debug_id, ref->data.strong,
drivers/android/binder.c
6570
ref->data.weak, ref->death);
drivers/android/binder.c
6571
binder_node_unlock(ref->node);
drivers/android/binder.c
6821
struct binder_ref *ref = rb_entry(n, struct binder_ref,
drivers/android/binder.c
6824
strong += ref->data.strong;
drivers/android/binder.c
6825
weak += ref->data.weak;
drivers/android/binder.c
7075
refcount_set(&binder_device->ref, 1);
drivers/android/binder_internal.h
38
refcount_t ref;
drivers/android/binderfs.c
169
refcount_set(&device->ref, 1);
drivers/android/binderfs.c
265
if (refcount_dec_and_test(&device->ref)) {
drivers/android/binderfs.c
428
refcount_set(&device->ref, 1);
drivers/base/core.c
3359
unsigned int ref;
drivers/base/core.c
3414
ref = kref_read(&glue_dir->kref);
drivers/base/core.c
3415
if (!kobject_has_children(glue_dir) && !--ref)
drivers/base/firmware_loader/firmware.h
65
struct kref ref;
drivers/base/firmware_loader/main.c
1380
kref_get(&fw_priv->ref);
drivers/base/firmware_loader/main.c
140
kref_init(&fw_priv->ref);
drivers/base/firmware_loader/main.c
182
kref_get(&tmp->ref);
drivers/base/firmware_loader/main.c
203
static void __free_fw_priv(struct kref *ref)
drivers/base/firmware_loader/main.c
206
struct fw_priv *fw_priv = to_fw_priv(ref);
drivers/base/firmware_loader/main.c
229
if (!kref_put(&fw_priv->ref, __free_fw_priv))
drivers/base/firmware_loader/main.c
83
static inline struct fw_priv *to_fw_priv(struct kref *ref)
drivers/base/firmware_loader/main.c
85
return container_of(ref, struct fw_priv, ref);
drivers/base/swnode.c
511
const struct software_node_ref_args *ref;
drivers/base/swnode.c
532
if ((index + 1) * sizeof(*ref) > prop->length)
drivers/base/swnode.c
536
ref = &ref_array[index];
drivers/base/swnode.c
549
if (ref->swnode)
drivers/base/swnode.c
550
refnode = software_node_fwnode(ref->swnode);
drivers/base/swnode.c
551
else if (ref->fwnode)
drivers/base/swnode.c
552
refnode = ref->fwnode;
drivers/base/swnode.c
577
args->args[i] = ref->args[i];
drivers/base/swnode.c
641
const struct software_node_ref_args *ref;
drivers/base/swnode.c
651
ref = prop->pointer;
drivers/base/swnode.c
653
if (!ref->swnode)
drivers/base/swnode.c
656
return software_node_get(software_node_fwnode(ref->swnode));
drivers/base/test/property-entry-test.c
425
struct fwnode_reference_args ref;
drivers/base/test/property-entry-test.c
435
0, 0, &ref);
drivers/base/test/property-entry-test.c
437
KUNIT_EXPECT_PTR_EQ(test, to_software_node(ref.fwnode), &node1);
drivers/base/test/property-entry-test.c
438
KUNIT_EXPECT_EQ(test, ref.nargs, 0U);
drivers/base/test/property-entry-test.c
442
0, 1, &ref);
drivers/base/test/property-entry-test.c
446
1, 0, &ref);
drivers/base/test/property-entry-test.c
448
KUNIT_EXPECT_PTR_EQ(test, to_software_node(ref.fwnode), &node2);
drivers/base/test/property-entry-test.c
449
KUNIT_EXPECT_EQ(test, ref.nargs, 1U);
drivers/base/test/property-entry-test.c
450
KUNIT_EXPECT_EQ(test, ref.args[0], 1LLU);
drivers/base/test/property-entry-test.c
454
3, 0, &ref);
drivers/base/test/property-entry-test.c
456
KUNIT_EXPECT_PTR_EQ(test, to_software_node(ref.fwnode), &node2);
drivers/base/test/property-entry-test.c
457
KUNIT_EXPECT_EQ(test, ref.nargs, 3U);
drivers/base/test/property-entry-test.c
458
KUNIT_EXPECT_EQ(test, ref.args[0], 1LLU);
drivers/base/test/property-entry-test.c
459
KUNIT_EXPECT_EQ(test, ref.args[1], 2LLU);
drivers/base/test/property-entry-test.c
460
KUNIT_EXPECT_EQ(test, ref.args[2], 0LLU);
drivers/base/test/property-entry-test.c
464
2, 1, &ref);
drivers/base/test/property-entry-test.c
469
0, 0, &ref);
drivers/base/test/property-entry-test.c
471
KUNIT_EXPECT_PTR_EQ(test, to_software_node(ref.fwnode), &node1);
drivers/base/test/property-entry-test.c
472
KUNIT_EXPECT_EQ(test, ref.nargs, 0U);
drivers/base/test/property-entry-test.c
476
2, 1, &ref);
drivers/base/test/property-entry-test.c
478
KUNIT_EXPECT_PTR_EQ(test, to_software_node(ref.fwnode), &node2);
drivers/base/test/property-entry-test.c
479
KUNIT_EXPECT_EQ(test, ref.nargs, 2U);
drivers/base/test/property-entry-test.c
480
KUNIT_EXPECT_EQ(test, ref.args[0], 3LLU);
drivers/base/test/property-entry-test.c
481
KUNIT_EXPECT_EQ(test, ref.args[1], 4LLU);
drivers/base/test/property-entry-test.c
485
0, 2, &ref);
drivers/block/aoe/aoe.h
169
ulong ref;
drivers/block/aoe/aoeblk.c
126
seq_printf(s, "ref: %ld\n", d->ref);
drivers/block/aoe/aoedev.c
143
d->ref--;
drivers/block/aoe/aoedev.c
353
|| d->ref)
drivers/block/aoe/aoedev.c
468
d->ref++;
drivers/block/aoe/aoedev.c
493
d->ref = 1;
drivers/block/ataflop.c
1641
if (floppy->ref != 1 && floppy->ref != -1)
drivers/block/ataflop.c
1748
if (floppy->ref != 1 && floppy->ref != -1)
drivers/block/ataflop.c
1927
if (p->ref && p->type != type)
drivers/block/ataflop.c
1930
if (p->ref == -1 || (p->ref && mode & BLK_OPEN_EXCL))
drivers/block/ataflop.c
1933
p->ref = -1;
drivers/block/ataflop.c
1935
p->ref++;
drivers/block/ataflop.c
1947
if (p->ref < 0)
drivers/block/ataflop.c
1948
p->ref = 0;
drivers/block/ataflop.c
1950
p->ref--;
drivers/block/ataflop.c
1973
if (p->ref < 0)
drivers/block/ataflop.c
1974
p->ref = 0;
drivers/block/ataflop.c
1975
else if (!p->ref--) {
drivers/block/ataflop.c
1977
p->ref = 0;
drivers/block/ataflop.c
303
int ref;
drivers/block/loop.c
322
if (!atomic_dec_and_test(&cmd->ref))
drivers/block/loop.c
384
atomic_set(&cmd->ref, 2);
drivers/block/loop.c
82
atomic_t ref; /* only for aio */
drivers/block/ublk_drv.c
1082
refcount_set(&io->ref, UBLK_REFCOUNT_INIT);
drivers/block/ublk_drv.c
1087
return refcount_inc_not_zero(&io->ref);
drivers/block/ublk_drv.c
1092
if (!refcount_dec_and_test(&io->ref))
drivers/block/ublk_drv.c
1104
return refcount_sub_and_test(sub_refs, &io->ref);
drivers/block/ublk_drv.c
223
refcount_t ref;
drivers/block/ublk_drv.c
2304
WARN_ON_ONCE(refcount_read(&io->ref));
drivers/block/ublk_drv.c
2408
unsigned int refs = refcount_read(&io->ref) +
drivers/block/ublk_drv.c
2419
refcount_set(&io->ref, 0);
drivers/block/ublk_drv.c
4055
WARN_ON_ONCE(refcount_read(&io->ref));
drivers/block/xen-blkback/blkback.c
858
persistent_gnt->gnt = map[new_map_idx].ref;
drivers/block/xen-blkfront.c
598
int n, ref;
drivers/block/xen-blkfront.c
636
ref = gnt_list_entry->gref;
drivers/block/xen-blkfront.c
669
.gref = ref,
drivers/block/xen-blkfront.c
675
.gref = ref,
drivers/block/zloop.c
138
atomic_t ref;
drivers/block/zloop.c
365
if (!atomic_dec_and_test(&cmd->ref))
drivers/block/zloop.c
400
atomic_set(&cmd->ref, 2);
drivers/bluetooth/btintel.c
2565
union acpi_object *p, *ref;
drivers/bluetooth/btintel.c
2582
ref = &p->package.elements[0];
drivers/bluetooth/btintel.c
2583
if (ref->type != ACPI_TYPE_LOCAL_REFERENCE) {
drivers/bluetooth/btintel.c
2584
bt_dev_err(hdev, "Invalid object type: 0x%x", ref->type);
drivers/bluetooth/btintel.c
2589
status = acpi_evaluate_object(ref->reference.handle, "_RST", NULL, NULL);
drivers/cdx/controller/mcdi.c
355
kref_put(&cmd->ref, cdx_mcdi_cmd_release);
drivers/cdx/controller/mcdi.c
414
struct kref ref;
drivers/cdx/controller/mcdi.c
423
static void cdx_mcdi_blocking_data_release(struct kref *ref)
drivers/cdx/controller/mcdi.c
425
kfree(container_of(ref, struct cdx_mcdi_blocking_data, ref));
drivers/cdx/controller/mcdi.c
443
kref_put(&wait_data->ref, cdx_mcdi_blocking_data_release);
drivers/cdx/controller/mcdi.c
469
kref_init(&wait_data->ref);
drivers/cdx/controller/mcdi.c
475
kref_init(&cmd_item->ref);
drivers/cdx/controller/mcdi.c
484
kref_get(&wait_data->ref);
drivers/cdx/controller/mcdi.c
487
kref_put(&wait_data->ref, cdx_mcdi_blocking_data_release);
drivers/cdx/controller/mcdi.c
508
kref_put(&wait_data->ref, cdx_mcdi_blocking_data_release);
drivers/cdx/controller/mcdi.c
530
kref_put(&cmd->ref, cdx_mcdi_cmd_release);
drivers/cdx/controller/mcdi.c
535
kref_put(&cmd->ref, cdx_mcdi_cmd_release);
drivers/cdx/controller/mcdi.c
58
static void cdx_mcdi_cmd_release(struct kref *ref)
drivers/cdx/controller/mcdi.c
60
kfree(container_of(ref, struct cdx_mcdi_cmd, ref));
drivers/cdx/controller/mcdi.c
617
kref_put(&cmd->ref, cdx_mcdi_cmd_release);
drivers/cdx/controller/mcdi.c
665
kref_get(&cmd->ref);
drivers/cdx/controller/mcdi.c
725
kref_put(&cmd->ref, cdx_mcdi_cmd_release);
drivers/cdx/controller/mcdi.c
745
kref_put(&cmd->ref, cdx_mcdi_cmd_release);
drivers/cdx/controller/mcdi.c
79
kref_get(&cmd->ref);
drivers/cdx/controller/mcdi.c
829
kref_init(&cmd_item->ref);
drivers/cdx/controller/mcdi.c
90
kref_put(&cmd->ref, cdx_mcdi_cmd_release);
drivers/char/hw_random/core.c
116
kref_put(&old_rng->ref, cleanup_rng);
drivers/char/hw_random/core.c
149
kref_put(&rng->ref, cleanup_rng);
drivers/char/hw_random/core.c
160
kref_get(&rng->ref);
drivers/char/hw_random/core.c
172
kref_get(&rng->ref);
drivers/char/hw_random/core.c
182
kref_put(&rng->ref, cleanup_rng);
drivers/char/hw_random/core.c
187
if (kref_get_unless_zero(&rng->ref))
drivers/char/hw_random/core.c
198
kref_init(&rng->ref);
drivers/char/hw_random/core.c
80
if (kref_read(&rng->ref)) {
drivers/char/hw_random/core.c
94
struct hwrng *rng = container_of(kref, struct hwrng, ref);
drivers/char/ipmi/ipmi_msghandler.c
3097
cleanup_bmc_device(struct kref *ref)
drivers/char/ipmi/ipmi_msghandler.c
3099
struct bmc_device *bmc = container_of(ref, struct bmc_device, usecount);
drivers/char/ipmi/ipmi_msghandler.c
50
static void intf_free(struct kref *ref);
drivers/char/ipmi/ipmi_msghandler.c
621
static void free_ipmi_user(struct kref *ref)
drivers/char/ipmi/ipmi_msghandler.c
623
struct ipmi_user *user = container_of(ref, struct ipmi_user, refcount);
drivers/char/ipmi/ipmi_msghandler.c
725
static void intf_free(struct kref *ref)
drivers/char/ipmi/ipmi_msghandler.c
727
struct ipmi_smi *intf = container_of(ref, struct ipmi_smi, refcount);
drivers/clk/clk.c
103
struct kref ref;
drivers/clk/clk.c
4182
kref_get(&core->ref);
drivers/clk/clk.c
4302
static void __clk_release(struct kref *ref)
drivers/clk/clk.c
4304
struct clk_core *core = container_of(ref, struct clk_core, ref);
drivers/clk/clk.c
4337
kref_init(&core->ref);
drivers/clk/clk.c
4395
kref_put(&core->ref, __clk_release);
drivers/clk/clk.c
4610
kref_put(&clk->core->ref, __clk_release);
drivers/clk/clk.c
4773
kref_put(&clk->core->ref, __clk_release);
drivers/clk/imx/clk-sscg-pll.c
192
uint64_t ref)
drivers/clk/imx/clk-sscg-pll.c
196
if (ref < PLL_STAGE1_MIN_FREQ || ref > PLL_STAGE1_MAX_FREQ)
drivers/clk/imx/clk-sscg-pll.c
199
temp_setup->vco1 = ref;
drivers/clk/imx/clk-sscg-pll.c
212
uint64_t vco1 = temp_setup->ref;
drivers/clk/imx/clk-sscg-pll.c
235
temp_setup->ref_div1 = temp_setup->ref;
drivers/clk/imx/clk-sscg-pll.c
250
uint64_t ref)
drivers/clk/imx/clk-sscg-pll.c
254
if (ref < PLL_REF_MIN_FREQ || ref > PLL_REF_MAX_FREQ)
drivers/clk/imx/clk-sscg-pll.c
257
temp_setup->ref = ref;
drivers/clk/imx/clk-sscg-pll.c
79
uint64_t ref;
drivers/clk/mxs/clk-ref.c
116
struct clk_ref *ref;
drivers/clk/mxs/clk-ref.c
120
ref = kzalloc_obj(*ref);
drivers/clk/mxs/clk-ref.c
121
if (!ref)
drivers/clk/mxs/clk-ref.c
130
ref->reg = reg;
drivers/clk/mxs/clk-ref.c
131
ref->idx = idx;
drivers/clk/mxs/clk-ref.c
132
ref->hw.init = &init;
drivers/clk/mxs/clk-ref.c
134
clk = clk_register(NULL, &ref->hw);
drivers/clk/mxs/clk-ref.c
136
kfree(ref);
drivers/clk/mxs/clk-ref.c
33
struct clk_ref *ref = to_clk_ref(hw);
drivers/clk/mxs/clk-ref.c
35
writel_relaxed(1 << ((ref->idx + 1) * 8 - 1), ref->reg + CLR);
drivers/clk/mxs/clk-ref.c
42
struct clk_ref *ref = to_clk_ref(hw);
drivers/clk/mxs/clk-ref.c
44
writel_relaxed(1 << ((ref->idx + 1) * 8 - 1), ref->reg + SET);
drivers/clk/mxs/clk-ref.c
50
struct clk_ref *ref = to_clk_ref(hw);
drivers/clk/mxs/clk-ref.c
52
u8 frac = (readl_relaxed(ref->reg) >> (ref->idx * 8)) & 0x3f;
drivers/clk/mxs/clk-ref.c
83
struct clk_ref *ref = to_clk_ref(hw);
drivers/clk/mxs/clk-ref.c
87
u8 frac, shift = ref->idx * 8;
drivers/clk/mxs/clk-ref.c
95
val = readl_relaxed(ref->reg);
drivers/clk/mxs/clk-ref.c
98
writel_relaxed(val, ref->reg);
drivers/clk/versatile/clk-icst.c
227
icst->params->ref = parent_rate;
drivers/clk/versatile/clk-icst.c
342
icst->params->ref = parent_rate;
drivers/clk/versatile/clk-impd1.c
27
.ref = 24000000, /* 24 MHz */
drivers/clk/versatile/clk-impd1.c
45
.ref = 24000000, /* 24 MHz */
drivers/clk/versatile/icst.c
29
u64 dividend = p->ref * 2 * (u64)(vco.v + 8);
drivers/clk/versatile/icst.c
79
fref_div = (2 * p->ref) / rd;
drivers/clk/versatile/icst.h
13
unsigned long ref;
drivers/crypto/ccp/sev-dev.c
2635
static void sev_exit(struct kref *ref)
drivers/crypto/ccp/sfs.c
185
static void sfs_exit(struct kref *ref)
drivers/crypto/intel/iaa/iaa_crypto.h
47
int ref;
drivers/crypto/intel/iaa/iaa_crypto_main.c
1817
if (iaa_wq->ref) {
drivers/crypto/intel/iaa/iaa_crypto_main.c
660
iaa_wq->ref++;
drivers/crypto/intel/iaa/iaa_crypto_main.c
680
iaa_wq->ref--;
drivers/crypto/intel/iaa/iaa_crypto_main.c
681
if (iaa_wq->ref == 0 && iaa_wq->remove) {
drivers/crypto/tegra/tegra-se-main.c
139
kref_init(&cmdbuf->ref);
drivers/crypto/tegra/tegra-se-main.c
21
kref_get(&cmdbuf->ref);
drivers/crypto/tegra/tegra-se-main.c
26
static void tegra_se_cmdbuf_release(struct kref *ref)
drivers/crypto/tegra/tegra-se-main.c
28
struct tegra_se_cmdbuf *cmdbuf = container_of(ref, struct tegra_se_cmdbuf, ref);
drivers/crypto/tegra/tegra-se-main.c
40
kref_put(&cmdbuf->ref, tegra_se_cmdbuf_release);
drivers/crypto/tegra/tegra-se-main.c
54
kref_init(&map->ref);
drivers/crypto/tegra/tegra-se.h
442
struct kref ref;
drivers/dma/dmaengine.c
1224
kref_init(&device->ref);
drivers/dma/dmaengine.c
425
static void dma_device_release(struct kref *ref)
drivers/dma/dmaengine.c
427
struct dma_device *device = container_of(ref, struct dma_device, ref);
drivers/dma/dmaengine.c
439
kref_put(&device->ref, dma_device_release);
drivers/dma/dmaengine.c
463
ret = kref_get_unless_zero(&chan->device->ref);
drivers/dma/fsl_raid.h
189
__be32 ref;
drivers/dma/idxd/device.c
406
static void idxd_wq_ref_release(struct percpu_ref *ref)
drivers/dma/idxd/device.c
408
struct idxd_wq *wq = container_of(ref, struct idxd_wq, wq_active);
drivers/dma/ppc4xx/adma.c
1373
struct ppc_dma_chan_ref *ref;
drivers/dma/ppc4xx/adma.c
1399
list_for_each_entry(ref, &ppc440spe_adma_chan_list, node) {
drivers/dma/ppc4xx/adma.c
1400
if (dma_has_cap(cap, ref->chan->device->cap_mask)) {
drivers/dma/ppc4xx/adma.c
1403
rank = ppc440spe_adma_estimate(ref->chan, cap, dst_lst,
drivers/dma/ppc4xx/adma.c
1407
best_chan = ref->chan;
drivers/dma/ppc4xx/adma.c
4009
struct ppc_dma_chan_ref *ref, *_ref;
drivers/dma/ppc4xx/adma.c
4163
ref = kmalloc_obj(*ref);
drivers/dma/ppc4xx/adma.c
4164
if (ref) {
drivers/dma/ppc4xx/adma.c
4165
ref->chan = &chan->common;
drivers/dma/ppc4xx/adma.c
4166
INIT_LIST_HEAD(&ref->node);
drivers/dma/ppc4xx/adma.c
4167
list_add_tail(&ref->node, &ppc440spe_adma_chan_list);
drivers/dma/ppc4xx/adma.c
4192
list_for_each_entry_safe(ref, _ref, &ppc440spe_adma_chan_list, node) {
drivers/dma/ppc4xx/adma.c
4193
if (chan == to_ppc440spe_adma_chan(ref->chan)) {
drivers/dma/ppc4xx/adma.c
4194
list_del(&ref->node);
drivers/dma/ppc4xx/adma.c
4195
kfree(ref);
drivers/dma/ppc4xx/adma.c
4238
struct ppc_dma_chan_ref *ref, *_ref;
drivers/dma/ppc4xx/adma.c
4259
list_for_each_entry_safe(ref, _ref, &ppc440spe_adma_chan_list,
drivers/dma/ppc4xx/adma.c
4262
to_ppc440spe_adma_chan(ref->chan)) {
drivers/dma/ppc4xx/adma.c
4263
list_del(&ref->node);
drivers/dma/ppc4xx/adma.c
4264
kfree(ref);
drivers/dpll/dpll_core.c
1018
struct dpll_pin_ref *ref;
drivers/dpll/dpll_core.c
1024
xa_for_each(&pin->dpll_refs, i, ref)
drivers/dpll/dpll_core.c
1025
__dpll_pin_unregister(ref->dpll, pin, ops, priv, parent);
drivers/dpll/dpll_core.c
1085
dpll_pin_registration_first(struct dpll_pin_ref *ref)
drivers/dpll/dpll_core.c
1089
reg = list_first_entry_or_null(&ref->registration_list,
drivers/dpll/dpll_core.c
1099
struct dpll_pin_ref *ref;
drivers/dpll/dpll_core.c
1101
ref = xa_load(&dpll->pin_refs, pin->pin_idx);
drivers/dpll/dpll_core.c
1102
if (!ref)
drivers/dpll/dpll_core.c
1104
reg = dpll_pin_registration_first(ref);
drivers/dpll/dpll_core.c
1112
struct dpll_pin_ref *ref;
drivers/dpll/dpll_core.c
1114
ref = xa_load(&pin->parent_refs, parent->pin_idx);
drivers/dpll/dpll_core.c
1115
if (!ref)
drivers/dpll/dpll_core.c
1117
reg = dpll_pin_registration_first(ref);
drivers/dpll/dpll_core.c
1121
const struct dpll_pin_ops *dpll_pin_ops(struct dpll_pin_ref *ref)
drivers/dpll/dpll_core.c
1125
reg = dpll_pin_registration_first(ref);
drivers/dpll/dpll_core.c
172
dpll_pin_registration_find(struct dpll_pin_ref *ref,
drivers/dpll/dpll_core.c
178
list_for_each_entry(reg, &ref->registration_list, list) {
drivers/dpll/dpll_core.c
192
struct dpll_pin_ref *ref;
drivers/dpll/dpll_core.c
197
xa_for_each(xa_pins, i, ref) {
drivers/dpll/dpll_core.c
198
if (ref->pin != pin)
drivers/dpll/dpll_core.c
200
reg = dpll_pin_registration_find(ref, ops, priv, cookie);
drivers/dpll/dpll_core.c
208
ref = kzalloc_obj(*ref);
drivers/dpll/dpll_core.c
209
if (!ref)
drivers/dpll/dpll_core.c
211
ref->pin = pin;
drivers/dpll/dpll_core.c
212
INIT_LIST_HEAD(&ref->registration_list);
drivers/dpll/dpll_core.c
213
ret = xa_insert(xa_pins, pin->pin_idx, ref, GFP_KERNEL);
drivers/dpll/dpll_core.c
215
kfree(ref);
drivers/dpll/dpll_core.c
218
refcount_set(&ref->refcount, 1);
drivers/dpll/dpll_core.c
225
kfree(ref);
drivers/dpll/dpll_core.c
234
refcount_inc(&ref->refcount);
drivers/dpll/dpll_core.c
235
list_add_tail(&reg->list, &ref->registration_list);
drivers/dpll/dpll_core.c
245
struct dpll_pin_ref *ref;
drivers/dpll/dpll_core.c
248
xa_for_each(xa_pins, i, ref) {
drivers/dpll/dpll_core.c
249
if (ref->pin != pin)
drivers/dpll/dpll_core.c
251
reg = dpll_pin_registration_find(ref, ops, priv, cookie);
drivers/dpll/dpll_core.c
257
if (refcount_dec_and_test(&ref->refcount)) {
drivers/dpll/dpll_core.c
259
WARN_ON(!list_empty(&ref->registration_list));
drivers/dpll/dpll_core.c
260
kfree(ref);
drivers/dpll/dpll_core.c
273
struct dpll_pin_ref *ref;
drivers/dpll/dpll_core.c
278
xa_for_each(xa_dplls, i, ref) {
drivers/dpll/dpll_core.c
279
if (ref->dpll != dpll)
drivers/dpll/dpll_core.c
281
reg = dpll_pin_registration_find(ref, ops, priv, cookie);
drivers/dpll/dpll_core.c
289
ref = kzalloc_obj(*ref);
drivers/dpll/dpll_core.c
290
if (!ref)
drivers/dpll/dpll_core.c
292
ref->dpll = dpll;
drivers/dpll/dpll_core.c
293
INIT_LIST_HEAD(&ref->registration_list);
drivers/dpll/dpll_core.c
294
ret = xa_insert(xa_dplls, dpll->id, ref, GFP_KERNEL);
drivers/dpll/dpll_core.c
296
kfree(ref);
drivers/dpll/dpll_core.c
299
refcount_set(&ref->refcount, 1);
drivers/dpll/dpll_core.c
306
kfree(ref);
drivers/dpll/dpll_core.c
315
refcount_inc(&ref->refcount);
drivers/dpll/dpll_core.c
316
list_add_tail(&reg->list, &ref->registration_list);
drivers/dpll/dpll_core.c
326
struct dpll_pin_ref *ref;
drivers/dpll/dpll_core.c
329
xa_for_each(xa_dplls, i, ref) {
drivers/dpll/dpll_core.c
330
if (ref->dpll != dpll)
drivers/dpll/dpll_core.c
332
reg = dpll_pin_registration_find(ref, ops, priv, cookie);
drivers/dpll/dpll_core.c
338
if (refcount_dec_and_test(&ref->refcount)) {
drivers/dpll/dpll_core.c
340
WARN_ON(!list_empty(&ref->registration_list));
drivers/dpll/dpll_core.c
341
kfree(ref);
drivers/dpll/dpll_core.c
349
struct dpll_pin_ref *ref;
drivers/dpll/dpll_core.c
352
ref = xa_find(xa_refs, &i, ULONG_MAX, XA_PRESENT);
drivers/dpll/dpll_core.c
353
WARN_ON(!ref);
drivers/dpll/dpll_core.c
354
return ref;
drivers/dpll/dpll_core.c
963
struct dpll_pin_ref *ref;
drivers/dpll/dpll_core.c
979
xa_for_each(&parent->dpll_refs, i, ref) {
drivers/dpll/dpll_core.c
980
ret = __dpll_pin_register(ref->dpll, pin, ops, priv, parent);
drivers/dpll/dpll_core.c
992
xa_for_each(&parent->dpll_refs, i, ref)
drivers/dpll/dpll_core.c
994
__dpll_pin_unregister(ref->dpll, pin, ops, priv,
drivers/dpll/dpll_core.h
94
const struct dpll_pin_ops *dpll_pin_ops(struct dpll_pin_ref *ref);
drivers/dpll/dpll_netlink.c
1003
xa_for_each(&pin->dpll_refs, i, ref) {
drivers/dpll/dpll_netlink.c
1004
if (ref == failed)
drivers/dpll/dpll_netlink.c
1006
ops = dpll_pin_ops(ref);
drivers/dpll/dpll_netlink.c
1007
dpll = ref->dpll;
drivers/dpll/dpll_netlink.c
1019
struct dpll_pin_ref *ref, *failed;
drivers/dpll/dpll_netlink.c
1028
xa_for_each(&pin->dpll_refs, i, ref) {
drivers/dpll/dpll_netlink.c
1029
ops = dpll_pin_ops(ref);
drivers/dpll/dpll_netlink.c
1036
ref = dpll_xa_ref_dpll_first(&pin->dpll_refs);
drivers/dpll/dpll_netlink.c
1037
ops = dpll_pin_ops(ref);
drivers/dpll/dpll_netlink.c
1038
dpll = ref->dpll;
drivers/dpll/dpll_netlink.c
1056
xa_for_each(&pin->dpll_refs, i, ref) {
drivers/dpll/dpll_netlink.c
1059
ops = dpll_pin_ops(ref);
drivers/dpll/dpll_netlink.c
1060
dpll = ref->dpll;
drivers/dpll/dpll_netlink.c
1065
failed = ref;
drivers/dpll/dpll_netlink.c
1077
xa_for_each(&pin->dpll_refs, i, ref) {
drivers/dpll/dpll_netlink.c
1080
if (ref == failed)
drivers/dpll/dpll_netlink.c
1082
ops = dpll_pin_ops(ref);
drivers/dpll/dpll_netlink.c
1083
dpll = ref->dpll;
drivers/dpll/dpll_netlink.c
1099
struct dpll_pin_ref *ref, *failed;
drivers/dpll/dpll_netlink.c
1117
ref = dpll_xa_ref_dpll_first(&pin->dpll_refs);
drivers/dpll/dpll_netlink.c
1118
ASSERT_NOT_NULL(ref);
drivers/dpll/dpll_netlink.c
1119
ops = dpll_pin_ops(ref);
drivers/dpll/dpll_netlink.c
1124
dpll = ref->dpll;
drivers/dpll/dpll_netlink.c
1135
xa_for_each(&pin->dpll_refs, i, ref) {
drivers/dpll/dpll_netlink.c
1136
ops = dpll_pin_ops(ref);
drivers/dpll/dpll_netlink.c
1137
dpll = ref->dpll;
drivers/dpll/dpll_netlink.c
1144
failed = ref;
drivers/dpll/dpll_netlink.c
1155
xa_for_each(&pin->dpll_refs, i, ref) {
drivers/dpll/dpll_netlink.c
1156
if (ref == failed)
drivers/dpll/dpll_netlink.c
1158
ops = dpll_pin_ops(ref);
drivers/dpll/dpll_netlink.c
1159
dpll = ref->dpll;
drivers/dpll/dpll_netlink.c
1240
struct dpll_pin_ref *ref;
drivers/dpll/dpll_netlink.c
1248
ref = xa_load(&pin->dpll_refs, dpll->id);
drivers/dpll/dpll_netlink.c
1249
ASSERT_NOT_NULL(ref);
drivers/dpll/dpll_netlink.c
1250
ops = dpll_pin_ops(ref);
drivers/dpll/dpll_netlink.c
1267
struct dpll_pin_ref *ref;
drivers/dpll/dpll_netlink.c
1275
ref = xa_load(&pin->dpll_refs, dpll->id);
drivers/dpll/dpll_netlink.c
1276
ASSERT_NOT_NULL(ref);
drivers/dpll/dpll_netlink.c
1277
ops = dpll_pin_ops(ref);
drivers/dpll/dpll_netlink.c
1295
struct dpll_pin_ref *ref;
drivers/dpll/dpll_netlink.c
1303
ref = xa_load(&pin->dpll_refs, dpll->id);
drivers/dpll/dpll_netlink.c
1304
ASSERT_NOT_NULL(ref);
drivers/dpll/dpll_netlink.c
1305
ops = dpll_pin_ops(ref);
drivers/dpll/dpll_netlink.c
1321
struct dpll_pin_ref *ref, *failed;
drivers/dpll/dpll_netlink.c
1342
xa_for_each(&pin->dpll_refs, i, ref) {
drivers/dpll/dpll_netlink.c
1343
ops = dpll_pin_ops(ref);
drivers/dpll/dpll_netlink.c
1349
ref = dpll_xa_ref_dpll_first(&pin->dpll_refs);
drivers/dpll/dpll_netlink.c
1350
ops = dpll_pin_ops(ref);
drivers/dpll/dpll_netlink.c
1351
dpll = ref->dpll;
drivers/dpll/dpll_netlink.c
1362
xa_for_each(&pin->dpll_refs, i, ref) {
drivers/dpll/dpll_netlink.c
1363
ops = dpll_pin_ops(ref);
drivers/dpll/dpll_netlink.c
1364
dpll = ref->dpll;
drivers/dpll/dpll_netlink.c
1370
failed = ref;
drivers/dpll/dpll_netlink.c
1382
xa_for_each(&pin->dpll_refs, i, ref) {
drivers/dpll/dpll_netlink.c
1383
if (ref == failed)
drivers/dpll/dpll_netlink.c
1385
ops = dpll_pin_ops(ref);
drivers/dpll/dpll_netlink.c
1386
dpll = ref->dpll;
drivers/dpll/dpll_netlink.c
1402
struct dpll_pin_ref *ref;
drivers/dpll/dpll_netlink.c
1419
ref = xa_load(&pin->dpll_refs, dpll->id);
drivers/dpll/dpll_netlink.c
1420
if (!ref) {
drivers/dpll/dpll_netlink.c
265
struct dpll_pin_ref *ref,
drivers/dpll/dpll_netlink.c
268
const struct dpll_pin_ops *ops = dpll_pin_ops(ref);
drivers/dpll/dpll_netlink.c
269
struct dpll_device *dpll = ref->dpll;
drivers/dpll/dpll_netlink.c
287
struct dpll_pin_ref *ref,
drivers/dpll/dpll_netlink.c
290
const struct dpll_pin_ops *ops = dpll_pin_ops(ref);
drivers/dpll/dpll_netlink.c
291
struct dpll_device *dpll = ref->dpll;
drivers/dpll/dpll_netlink.c
309
struct dpll_pin_ref *ref,
drivers/dpll/dpll_netlink.c
312
const struct dpll_pin_ops *ops = dpll_pin_ops(ref);
drivers/dpll/dpll_netlink.c
313
struct dpll_device *dpll = ref->dpll;
drivers/dpll/dpll_netlink.c
329
struct dpll_pin_ref *ref,
drivers/dpll/dpll_netlink.c
332
const struct dpll_pin_ops *ops = dpll_pin_ops(ref);
drivers/dpll/dpll_netlink.c
333
struct dpll_device *dpll = ref->dpll;
drivers/dpll/dpll_netlink.c
352
struct dpll_pin_ref *ref,
drivers/dpll/dpll_netlink.c
355
const struct dpll_pin_ops *ops = dpll_pin_ops(ref);
drivers/dpll/dpll_netlink.c
356
struct dpll_device *dpll = ref->dpll;
drivers/dpll/dpll_netlink.c
375
struct dpll_pin_ref *ref,
drivers/dpll/dpll_netlink.c
378
const struct dpll_pin_ops *ops = dpll_pin_ops(ref);
drivers/dpll/dpll_netlink.c
379
struct dpll_device *dpll = ref->dpll;
drivers/dpll/dpll_netlink.c
405
struct dpll_pin_ref *ref, struct netlink_ext_ack *extack)
drivers/dpll/dpll_netlink.c
407
const struct dpll_pin_ops *ops = dpll_pin_ops(ref);
drivers/dpll/dpll_netlink.c
408
struct dpll_device *dpll = ref->dpll;
drivers/dpll/dpll_netlink.c
446
struct dpll_pin_ref *ref, struct netlink_ext_ack *extack)
drivers/dpll/dpll_netlink.c
448
const struct dpll_pin_ops *ops = dpll_pin_ops(ref);
drivers/dpll/dpll_netlink.c
449
struct dpll_device *dpll = ref->dpll;
drivers/dpll/dpll_netlink.c
491
struct dpll_pin_ref *ref,
drivers/dpll/dpll_netlink.c
494
const struct dpll_pin_ops *ops = dpll_pin_ops(ref);
drivers/dpll/dpll_netlink.c
495
struct dpll_device *dpll = ref->dpll;
drivers/dpll/dpll_netlink.c
547
struct dpll_pin_ref *ref;
drivers/dpll/dpll_netlink.c
553
xa_for_each(&pin->parent_refs, index, ref) {
drivers/dpll/dpll_netlink.c
554
const struct dpll_pin_ops *ops = dpll_pin_ops(ref);
drivers/dpll/dpll_netlink.c
557
ppin = ref->pin;
drivers/dpll/dpll_netlink.c
588
struct dpll_pin_ref *ref;
drivers/dpll/dpll_netlink.c
593
xa_for_each(&pin->dpll_refs, index, ref) {
drivers/dpll/dpll_netlink.c
597
ret = dpll_msg_add_dev_parent_handle(msg, ref->dpll->id);
drivers/dpll/dpll_netlink.c
600
ret = dpll_msg_add_pin_on_dpll_state(msg, pin, ref, extack);
drivers/dpll/dpll_netlink.c
603
ret = dpll_msg_add_pin_prio(msg, pin, ref, extack);
drivers/dpll/dpll_netlink.c
606
ret = dpll_msg_add_pin_direction(msg, pin, ref, extack);
drivers/dpll/dpll_netlink.c
609
ret = dpll_msg_add_phase_offset(msg, pin, ref, extack);
drivers/dpll/dpll_netlink.c
627
struct dpll_pin_ref *ref;
drivers/dpll/dpll_netlink.c
630
ref = dpll_xa_ref_dpll_first(&pin->dpll_refs);
drivers/dpll/dpll_netlink.c
631
ASSERT_NOT_NULL(ref);
drivers/dpll/dpll_netlink.c
656
ret = dpll_msg_add_pin_freq(msg, pin, ref, extack);
drivers/dpll/dpll_netlink.c
669
ret = dpll_msg_add_pin_phase_adjust(msg, pin, ref, extack);
drivers/dpll/dpll_netlink.c
672
ret = dpll_msg_add_ffo(msg, pin, ref, extack);
drivers/dpll/dpll_netlink.c
675
ret = dpll_msg_add_pin_esync(msg, pin, ref, extack);
drivers/dpll/dpll_netlink.c
679
ret = dpll_msg_add_pin_ref_sync(msg, pin, ref, extack);
drivers/dpll/dpll_netlink.c
685
ret = dpll_msg_add_pin_parents(msg, pin, ref, extack);
drivers/dpll/dpll_netlink.c
956
struct dpll_pin_ref *ref, *failed;
drivers/dpll/dpll_netlink.c
967
xa_for_each(&pin->dpll_refs, i, ref) {
drivers/dpll/dpll_netlink.c
968
ops = dpll_pin_ops(ref);
drivers/dpll/dpll_netlink.c
974
ref = dpll_xa_ref_dpll_first(&pin->dpll_refs);
drivers/dpll/dpll_netlink.c
975
ops = dpll_pin_ops(ref);
drivers/dpll/dpll_netlink.c
976
dpll = ref->dpll;
drivers/dpll/dpll_netlink.c
986
xa_for_each(&pin->dpll_refs, i, ref) {
drivers/dpll/dpll_netlink.c
987
ops = dpll_pin_ops(ref);
drivers/dpll/dpll_netlink.c
988
dpll = ref->dpll;
drivers/dpll/dpll_netlink.c
992
failed = ref;
drivers/dpll/zl3073x/core.c
601
&zldev->ref[i].mon_status);
drivers/dpll/zl3073x/core.c
717
zldev->ref[i].ffo = mul_s64_u64_shr(value, 244140625, 20);
drivers/dpll/zl3073x/core.h
223
const struct zl3073x_ref *ref = zl3073x_ref_state_get(zldev, index);
drivers/dpll/zl3073x/core.h
225
return zl3073x_ref_freq_get(ref);
drivers/dpll/zl3073x/core.h
238
const struct zl3073x_ref *ref = zl3073x_ref_state_get(zldev, index);
drivers/dpll/zl3073x/core.h
240
return zl3073x_ref_is_diff(ref);
drivers/dpll/zl3073x/core.h
253
const struct zl3073x_ref *ref = zl3073x_ref_state_get(zldev, index);
drivers/dpll/zl3073x/core.h
255
return zl3073x_ref_is_status_ok(ref);
drivers/dpll/zl3073x/core.h
55
struct zl3073x_ref ref[ZL3073X_NUM_REFS];
drivers/dpll/zl3073x/dpll.c
1223
u8 hw_mode, mode_refsel, ref;
drivers/dpll/zl3073x/dpll.c
1226
rc = zl3073x_dpll_selected_ref_get(zldpll, &ref);
drivers/dpll/zl3073x/dpll.c
1241
if (ZL3073X_DPLL_REF_IS_VALID(ref))
drivers/dpll/zl3073x/dpll.c
1255
pin = zl3073x_dpll_pin_get_by_ref(zldpll, ref);
drivers/dpll/zl3073x/dpll.c
1274
if (ZL3073X_DPLL_REF_IS_VALID(ref))
drivers/dpll/zl3073x/dpll.c
1275
mode_refsel |= FIELD_PREP(ZL_DPLL_MODE_REFSEL_REF, ref);
drivers/dpll/zl3073x/dpll.c
1288
if (ZL3073X_DPLL_REF_IS_VALID(ref))
drivers/dpll/zl3073x/dpll.c
1289
zldpll->forced_ref = ref;
drivers/dpll/zl3073x/dpll.c
130
const struct zl3073x_ref *ref;
drivers/dpll/zl3073x/dpll.c
134
ref = zl3073x_ref_state_get(zldev, ref_id);
drivers/dpll/zl3073x/dpll.c
136
switch (FIELD_GET(ZL_REF_SYNC_CTRL_MODE, ref->sync_ctrl)) {
drivers/dpll/zl3073x/dpll.c
138
esync->freq = ref->esync_n_div == ZL_REF_ESYNC_DIV_1HZ ? 1 : 0;
drivers/dpll/zl3073x/dpll.c
150
if (pin->esync_control && zl3073x_ref_freq_get(ref) > 1) {
drivers/dpll/zl3073x/dpll.c
1548
const struct zl3073x_ref *ref;
drivers/dpll/zl3073x/dpll.c
1555
ref = zl3073x_ref_state_get(zldev, ref_id);
drivers/dpll/zl3073x/dpll.c
1556
is_diff = zl3073x_ref_is_diff(ref);
drivers/dpll/zl3073x/dpll.c
1557
is_enabled = zl3073x_ref_is_enabled(ref);
drivers/dpll/zl3073x/dpll.c
171
struct zl3073x_ref ref;
drivers/dpll/zl3073x/dpll.c
175
ref = *zl3073x_ref_state_get(zldev, ref_id);
drivers/dpll/zl3073x/dpll.c
1787
const struct zl3073x_ref *ref;
drivers/dpll/zl3073x/dpll.c
1792
ref = zl3073x_ref_state_get(zldev, ref_id);
drivers/dpll/zl3073x/dpll.c
1795
if (!zl3073x_ref_is_status_ok(ref))
drivers/dpll/zl3073x/dpll.c
1799
if (pin->freq_offset != ref->ffo) {
drivers/dpll/zl3073x/dpll.c
1801
pin->label, pin->freq_offset, ref->ffo);
drivers/dpll/zl3073x/dpll.c
1802
pin->freq_offset = ref->ffo;
drivers/dpll/zl3073x/dpll.c
183
ref.sync_ctrl &= ~ZL_REF_SYNC_CTRL_MODE;
drivers/dpll/zl3073x/dpll.c
184
ref.sync_ctrl |= FIELD_PREP(ZL_REF_SYNC_CTRL_MODE, sync_mode);
drivers/dpll/zl3073x/dpll.c
188
ref.esync_n_div = ZL_REF_ESYNC_DIV_1HZ;
drivers/dpll/zl3073x/dpll.c
192
return zl3073x_ref_state_set(zldev, ref_id, &ref);
drivers/dpll/zl3073x/dpll.c
234
struct zl3073x_ref ref;
drivers/dpll/zl3073x/dpll.c
239
ref = *zl3073x_ref_state_get(zldev, ref_id);
drivers/dpll/zl3073x/dpll.c
242
zl3073x_ref_freq_set(&ref, frequency);
drivers/dpll/zl3073x/dpll.c
245
return zl3073x_ref_state_set(zldev, ref_id, &ref);
drivers/dpll/zl3073x/dpll.c
259
zl3073x_dpll_selected_ref_get(struct zl3073x_dpll *zldpll, u8 *ref)
drivers/dpll/zl3073x/dpll.c
279
*ref = FIELD_GET(ZL_DPLL_REFSEL_STATUS_REFSEL, value);
drivers/dpll/zl3073x/dpll.c
281
*ref = ZL3073X_DPLL_REF_NONE;
drivers/dpll/zl3073x/dpll.c
285
*ref = zldpll->forced_ref;
drivers/dpll/zl3073x/dpll.c
289
*ref = ZL3073X_DPLL_REF_NONE;
drivers/dpll/zl3073x/dpll.c
307
zl3073x_dpll_selected_ref_set(struct zl3073x_dpll *zldpll, u8 ref)
drivers/dpll/zl3073x/dpll.c
318
if (ref == ZL3073X_DPLL_REF_NONE) {
drivers/dpll/zl3073x/dpll.c
331
ref = zldpll->forced_ref;
drivers/dpll/zl3073x/dpll.c
332
} else if (ref == zldpll->forced_ref) {
drivers/dpll/zl3073x/dpll.c
340
if (ref == ZL3073X_DPLL_REF_NONE)
drivers/dpll/zl3073x/dpll.c
356
FIELD_PREP(ZL_DPLL_MODE_REFSEL_REF, ref);
drivers/dpll/zl3073x/dpll.c
366
zldpll->forced_ref = ref;
drivers/dpll/zl3073x/dpll.c
382
zl3073x_dpll_connected_ref_get(struct zl3073x_dpll *zldpll, u8 *ref)
drivers/dpll/zl3073x/dpll.c
388
rc = zl3073x_dpll_selected_ref_get(zldpll, ref);
drivers/dpll/zl3073x/dpll.c
393
if (ZL3073X_DPLL_REF_IS_VALID(*ref) &&
drivers/dpll/zl3073x/dpll.c
394
!zl3073x_dev_ref_is_status_ok(zldev, *ref))
drivers/dpll/zl3073x/dpll.c
395
*ref = ZL3073X_DPLL_REF_NONE;
drivers/dpll/zl3073x/dpll.c
410
const struct zl3073x_ref *ref;
drivers/dpll/zl3073x/dpll.c
425
ref = zl3073x_ref_state_get(zldev, ref_id);
drivers/dpll/zl3073x/dpll.c
427
!zl3073x_ref_is_status_ok(ref)) {
drivers/dpll/zl3073x/dpll.c
443
ref_freq = zl3073x_ref_freq_get(ref);
drivers/dpll/zl3073x/dpll.c
470
const struct zl3073x_ref *ref;
drivers/dpll/zl3073x/dpll.c
476
ref = zl3073x_ref_state_get(zldev, ref_id);
drivers/dpll/zl3073x/dpll.c
480
phase_comp = sign_extend64(ref->phase_comp, 31);
drivers/dpll/zl3073x/dpll.c
482
phase_comp = sign_extend64(ref->phase_comp, 47);
drivers/dpll/zl3073x/dpll.c
503
struct zl3073x_ref ref;
drivers/dpll/zl3073x/dpll.c
508
ref = *zl3073x_ref_state_get(zldev, ref_id);
drivers/dpll/zl3073x/dpll.c
513
ref.phase_comp = -phase_adjust;
drivers/dpll/zl3073x/dpll.c
516
return zl3073x_ref_state_set(zldev, ref_id, &ref);
drivers/dpll/zl3073x/dpll.c
534
u8 ref, ref_prio;
drivers/dpll/zl3073x/dpll.c
546
ref = zl3073x_input_pin_ref_get(pin->id);
drivers/dpll/zl3073x/dpll.c
547
rc = zl3073x_read_u8(zldev, ZL_REG_DPLL_REF_PRIO(ref / 2),
drivers/dpll/zl3073x/dpll.c
575
u8 ref, ref_prio;
drivers/dpll/zl3073x/dpll.c
587
ref = zl3073x_input_pin_ref_get(pin->id);
drivers/dpll/zl3073x/dpll.c
588
rc = zl3073x_read_u8(zldev, ZL_REG_DPLL_REF_PRIO(ref / 2), &ref_prio);
drivers/dpll/zl3073x/dpll.c
602
rc = zl3073x_write_u8(zldev, ZL_REG_DPLL_REF_PRIO(ref / 2), ref_prio);
drivers/dpll/zl3073x/dpll.c
627
u8 ref, ref_conn;
drivers/dpll/zl3073x/dpll.c
630
ref = zl3073x_input_pin_ref_get(pin->id);
drivers/dpll/zl3073x/dpll.c
637
if (ref == ref_conn) {
drivers/dpll/zl3073x/dpll.c
647
zl3073x_dev_ref_is_status_ok(zldev, ref) && pin->selectable) {
drivers/dpll/zl3073x/prop.c
92
u8 ref;
drivers/dpll/zl3073x/prop.c
95
ref = zl3073x_input_pin_ref_get(id);
drivers/dpll/zl3073x/prop.c
96
is_diff = zl3073x_dev_ref_is_diff(zldev, ref);
drivers/dpll/zl3073x/ref.c
102
rc = zl3073x_read_u16(zldev, ZL_REG_REF_FREQ_BASE, &ref->freq_base);
drivers/dpll/zl3073x/ref.c
105
rc = zl3073x_read_u16(zldev, ZL_REG_REF_FREQ_MULT, &ref->freq_mult);
drivers/dpll/zl3073x/ref.c
108
rc = zl3073x_read_u16(zldev, ZL_REG_REF_RATIO_M, &ref->freq_ratio_m);
drivers/dpll/zl3073x/ref.c
111
rc = zl3073x_read_u16(zldev, ZL_REG_REF_RATIO_N, &ref->freq_ratio_n);
drivers/dpll/zl3073x/ref.c
116
rc = zl3073x_read_u32(zldev, ZL_REG_REF_ESYNC_DIV, &ref->esync_n_div);
drivers/dpll/zl3073x/ref.c
119
rc = zl3073x_read_u8(zldev, ZL_REG_REF_SYNC_CTRL, &ref->sync_ctrl);
drivers/dpll/zl3073x/ref.c
129
ref->phase_comp = val;
drivers/dpll/zl3073x/ref.c
132
&ref->phase_comp);
drivers/dpll/zl3073x/ref.c
138
str_enabled_disabled(zl3073x_ref_is_enabled(ref)),
drivers/dpll/zl3073x/ref.c
139
zl3073x_ref_is_diff(ref) ? "differential" : "single-ended");
drivers/dpll/zl3073x/ref.c
154
return &zldev->ref[index];
drivers/dpll/zl3073x/ref.c
158
const struct zl3073x_ref *ref)
drivers/dpll/zl3073x/ref.c
160
struct zl3073x_ref *dref = &zldev->ref[index];
drivers/dpll/zl3073x/ref.c
172
if (dref->freq_base != ref->freq_base)
drivers/dpll/zl3073x/ref.c
174
ref->freq_base);
drivers/dpll/zl3073x/ref.c
175
if (!rc && dref->freq_mult != ref->freq_mult)
drivers/dpll/zl3073x/ref.c
177
ref->freq_mult);
drivers/dpll/zl3073x/ref.c
178
if (!rc && dref->freq_ratio_m != ref->freq_ratio_m)
drivers/dpll/zl3073x/ref.c
180
ref->freq_ratio_m);
drivers/dpll/zl3073x/ref.c
181
if (!rc && dref->freq_ratio_n != ref->freq_ratio_n)
drivers/dpll/zl3073x/ref.c
183
ref->freq_ratio_n);
drivers/dpll/zl3073x/ref.c
184
if (!rc && dref->esync_n_div != ref->esync_n_div)
drivers/dpll/zl3073x/ref.c
186
ref->esync_n_div);
drivers/dpll/zl3073x/ref.c
187
if (!rc && dref->sync_ctrl != ref->sync_ctrl)
drivers/dpll/zl3073x/ref.c
189
ref->sync_ctrl);
drivers/dpll/zl3073x/ref.c
190
if (!rc && dref->phase_comp != ref->phase_comp) {
drivers/dpll/zl3073x/ref.c
194
ref->phase_comp);
drivers/dpll/zl3073x/ref.c
198
ref->phase_comp);
drivers/dpll/zl3073x/ref.c
210
dref->freq_base = ref->freq_base;
drivers/dpll/zl3073x/ref.c
211
dref->freq_mult = ref->freq_mult;
drivers/dpll/zl3073x/ref.c
212
dref->freq_ratio_m = ref->freq_ratio_m;
drivers/dpll/zl3073x/ref.c
213
dref->freq_ratio_n = ref->freq_ratio_n;
drivers/dpll/zl3073x/ref.c
214
dref->esync_n_div = ref->esync_n_div;
drivers/dpll/zl3073x/ref.c
215
dref->sync_ctrl = ref->sync_ctrl;
drivers/dpll/zl3073x/ref.c
216
dref->phase_comp = ref->phase_comp;
drivers/dpll/zl3073x/ref.c
66
struct zl3073x_ref *ref = &zldev->ref[index];
drivers/dpll/zl3073x/ref.c
72
if (zl3073x_is_n_pin(index) && zl3073x_ref_is_diff(ref - 1)) {
drivers/dpll/zl3073x/ref.c
73
struct zl3073x_ref *p_ref = ref - 1; /* P-pin counterpart*/
drivers/dpll/zl3073x/ref.c
76
ref->config = p_ref->config;
drivers/dpll/zl3073x/ref.c
77
ref->esync_n_div = p_ref->esync_n_div;
drivers/dpll/zl3073x/ref.c
78
ref->freq_base = p_ref->freq_base;
drivers/dpll/zl3073x/ref.c
79
ref->freq_mult = p_ref->freq_mult;
drivers/dpll/zl3073x/ref.c
80
ref->freq_ratio_m = p_ref->freq_ratio_m;
drivers/dpll/zl3073x/ref.c
81
ref->freq_ratio_n = p_ref->freq_ratio_n;
drivers/dpll/zl3073x/ref.c
82
ref->phase_comp = p_ref->phase_comp;
drivers/dpll/zl3073x/ref.c
83
ref->sync_ctrl = p_ref->sync_ctrl;
drivers/dpll/zl3073x/ref.c
97
rc = zl3073x_read_u8(zldev, ZL_REG_REF_CONFIG, &ref->config);
drivers/dpll/zl3073x/ref.h
107
zl3073x_ref_is_diff(const struct zl3073x_ref *ref)
drivers/dpll/zl3073x/ref.h
109
return !!FIELD_GET(ZL_REF_CONFIG_DIFF_EN, ref->config);
drivers/dpll/zl3073x/ref.h
119
zl3073x_ref_is_enabled(const struct zl3073x_ref *ref)
drivers/dpll/zl3073x/ref.h
121
return !!FIELD_GET(ZL_REF_CONFIG_ENABLE, ref->config);
drivers/dpll/zl3073x/ref.h
131
zl3073x_ref_is_status_ok(const struct zl3073x_ref *ref)
drivers/dpll/zl3073x/ref.h
133
return ref->mon_status == ZL_REF_MON_STATUS_OK;
drivers/dpll/zl3073x/ref.h
46
const struct zl3073x_ref *ref);
drivers/dpll/zl3073x/ref.h
57
zl3073x_ref_ffo_get(const struct zl3073x_ref *ref)
drivers/dpll/zl3073x/ref.h
59
return ref->ffo;
drivers/dpll/zl3073x/ref.h
69
zl3073x_ref_freq_get(const struct zl3073x_ref *ref)
drivers/dpll/zl3073x/ref.h
71
return mul_u64_u32_div(ref->freq_base * ref->freq_mult,
drivers/dpll/zl3073x/ref.h
72
ref->freq_ratio_m, ref->freq_ratio_n);
drivers/dpll/zl3073x/ref.h
83
zl3073x_ref_freq_set(struct zl3073x_ref *ref, u32 freq)
drivers/dpll/zl3073x/ref.h
92
ref->freq_base = base;
drivers/dpll/zl3073x/ref.h
93
ref->freq_mult = mult;
drivers/dpll/zl3073x/ref.h
94
ref->freq_ratio_m = 1;
drivers/dpll/zl3073x/ref.h
95
ref->freq_ratio_n = 1;
drivers/gpio/gpiolib-shared.c
101
ref->flags = flags;
drivers/gpio/gpiolib-shared.c
102
ref->con_id = no_free_ptr(con_id_cpy);
drivers/gpio/gpiolib-shared.c
103
ref->fwnode = fwnode;
drivers/gpio/gpiolib-shared.c
104
lockdep_register_key(&ref->lock_key);
drivers/gpio/gpiolib-shared.c
105
mutex_init_with_key(&ref->lock, &ref->lock_key);
drivers/gpio/gpiolib-shared.c
107
return no_free_ptr(ref);
drivers/gpio/gpiolib-shared.c
113
struct gpio_shared_ref *ref;
drivers/gpio/gpiolib-shared.c
115
list_for_each_entry(ref, &entry->refs, list) {
drivers/gpio/gpiolib-shared.c
116
if (ref->is_reset_gpio)
drivers/gpio/gpiolib-shared.c
121
ref = gpio_shared_make_ref(NULL, "reset", flags);
drivers/gpio/gpiolib-shared.c
122
if (!ref)
drivers/gpio/gpiolib-shared.c
125
ref->is_reset_gpio = true;
drivers/gpio/gpiolib-shared.c
127
list_add_tail(&ref->list, &entry->refs);
drivers/gpio/gpiolib-shared.c
165
struct gpio_shared_ref *ref;
drivers/gpio/gpiolib-shared.c
264
ref = gpio_shared_make_ref(fwnode_handle_get(of_fwnode_handle(curr)),
drivers/gpio/gpiolib-shared.c
266
if (!ref)
drivers/gpio/gpiolib-shared.c
273
list_add_tail(&ref->list, &entry->refs);
drivers/gpio/gpiolib-shared.c
313
struct gpio_shared_ref *ref)
drivers/gpio/gpiolib-shared.c
315
struct auxiliary_device *adev = &ref->adev;
drivers/gpio/gpiolib-shared.c
318
guard(mutex)(&ref->lock);
drivers/gpio/gpiolib-shared.c
322
adev->id = ref->dev_id;
drivers/gpio/gpiolib-shared.c
358
struct gpio_shared_ref *ref)
drivers/gpio/gpiolib-shared.c
367
lockdep_assert_held(&ref->lock);
drivers/gpio/gpiolib-shared.c
387
if (real_ref == ref)
drivers/gpio/gpiolib-shared.c
431
ref->fwnode = fwnode_handle_get(reset_fwnode);
drivers/gpio/gpiolib-shared.c
440
struct gpio_shared_ref *ref)
drivers/gpio/gpiolib-shared.c
452
struct gpio_shared_ref *ref;
drivers/gpio/gpiolib-shared.c
455
list_for_each_entry(ref, &entry->refs, list) {
drivers/gpio/gpiolib-shared.c
456
guard(mutex)(&ref->lock);
drivers/gpio/gpiolib-shared.c
458
if (!ref->fwnode && device_is_compatible(consumer, "reset-gpio")) {
drivers/gpio/gpiolib-shared.c
459
if (!gpio_shared_dev_is_reset_gpio(consumer, entry, ref))
drivers/gpio/gpiolib-shared.c
461
} else if (fwnode != ref->fwnode) {
drivers/gpio/gpiolib-shared.c
465
if ((!con_id && ref->con_id) || (con_id && !ref->con_id) ||
drivers/gpio/gpiolib-shared.c
466
(con_id && ref->con_id && strcmp(con_id, ref->con_id) != 0))
drivers/gpio/gpiolib-shared.c
470
if (ref->lookup)
drivers/gpio/gpiolib-shared.c
476
ref->adev.id);
drivers/gpio/gpiolib-shared.c
485
dev_id, key, ref->con_id ?: "none");
drivers/gpio/gpiolib-shared.c
489
ref->con_id, lflags);
drivers/gpio/gpiolib-shared.c
491
ref->lookup = lookup;
drivers/gpio/gpiolib-shared.c
492
gpiod_add_lookup_table(ref->lookup);
drivers/gpio/gpiolib-shared.c
513
struct gpio_shared_ref *ref;
drivers/gpio/gpiolib-shared.c
518
list_for_each_entry(ref, &entry->refs, list) {
drivers/gpio/gpiolib-shared.c
519
if (gdev->dev.parent == &ref->adev.dev) {
drivers/gpio/gpiolib-shared.c
579
list_for_each_entry(ref, &entry->refs, list) {
drivers/gpio/gpiolib-shared.c
581
fwnode_get_name(ref->fwnode) ?: "(no fwnode)",
drivers/gpio/gpiolib-shared.c
582
ref->con_id ?: "(none)");
drivers/gpio/gpiolib-shared.c
584
ret = gpio_shared_make_adev(gdev, entry, ref);
drivers/gpio/gpiolib-shared.c
59
struct kref ref;
drivers/gpio/gpiolib-shared.c
598
struct gpio_shared_ref *ref;
drivers/gpio/gpiolib-shared.c
608
list_for_each_entry(ref, &entry->refs, list) {
drivers/gpio/gpiolib-shared.c
609
guard(mutex)(&ref->lock);
drivers/gpio/gpiolib-shared.c
611
if (ref->lookup) {
drivers/gpio/gpiolib-shared.c
612
gpiod_remove_lookup_table(ref->lookup);
drivers/gpio/gpiolib-shared.c
613
kfree(ref->lookup->table[0].key);
drivers/gpio/gpiolib-shared.c
614
kfree(ref->lookup);
drivers/gpio/gpiolib-shared.c
615
ref->lookup = NULL;
drivers/gpio/gpiolib-shared.c
618
gpio_shared_remove_adev(&ref->adev);
drivers/gpio/gpiolib-shared.c
626
container_of(kref, struct gpio_shared_entry, ref);
drivers/gpio/gpiolib-shared.c
643
kref_put(&entry->ref, gpio_shared_release);
drivers/gpio/gpiolib-shared.c
687
kref_get(&entry->ref);
drivers/gpio/gpiolib-shared.c
694
kref_init(&entry->ref);
drivers/gpio/gpiolib-shared.c
711
static void gpio_shared_drop_ref(struct gpio_shared_ref *ref)
drivers/gpio/gpiolib-shared.c
713
list_del(&ref->list);
drivers/gpio/gpiolib-shared.c
714
mutex_destroy(&ref->lock);
drivers/gpio/gpiolib-shared.c
715
lockdep_unregister_key(&ref->lock_key);
drivers/gpio/gpiolib-shared.c
716
kfree(ref->con_id);
drivers/gpio/gpiolib-shared.c
717
ida_free(&gpio_shared_ida, ref->dev_id);
drivers/gpio/gpiolib-shared.c
718
fwnode_handle_put(ref->fwnode);
drivers/gpio/gpiolib-shared.c
719
kfree(ref);
drivers/gpio/gpiolib-shared.c
737
struct gpio_shared_ref *ref, *rpos;
drivers/gpio/gpiolib-shared.c
740
list_for_each_entry_safe(ref, rpos, &entry->refs, list)
drivers/gpio/gpiolib-shared.c
741
gpio_shared_drop_ref(ref);
drivers/gpio/gpiolib-shared.c
750
struct gpio_shared_ref *ref;
drivers/gpio/gpiolib-shared.c
759
list_for_each_entry(ref, &entry->refs, list) {
drivers/gpio/gpiolib-shared.c
766
if (ref->is_reset_gpio)
drivers/gpio/gpiolib-shared.c
776
struct gpio_shared_ref *ref, *rpos;
drivers/gpio/gpiolib-shared.c
782
list_for_each_entry_safe(ref, rpos, &entry->refs, list)
drivers/gpio/gpiolib-shared.c
783
gpio_shared_drop_ref(ref);
drivers/gpio/gpiolib-shared.c
87
struct gpio_shared_ref *ref __free(kfree) = kzalloc_obj(*ref);
drivers/gpio/gpiolib-shared.c
88
if (!ref)
drivers/gpio/gpiolib-shared.c
97
ref->dev_id = ida_alloc(&gpio_shared_ida, GFP_KERNEL);
drivers/gpio/gpiolib-shared.c
98
if (ref->dev_id < 0)
drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.c
49
static void amdgpu_bo_list_free(struct kref *ref)
drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.c
51
struct amdgpu_bo_list *list = container_of(ref, struct amdgpu_bo_list,
drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c
406
static void amdgpu_ctx_fini(struct kref *ref)
drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c
408
struct amdgpu_ctx *ctx = container_of(ref, struct amdgpu_ctx, refcount);
drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c
508
static void amdgpu_ctx_do_release(struct kref *ref)
drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c
513
ctx = container_of(ref, struct amdgpu_ctx, refcount);
drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c
523
amdgpu_ctx_fini(ref);
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
864
uint32_t ref, uint32_t mask,
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
875
ref, mask, xcc_inst);
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
882
ref, mask);
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.h
445
uint32_t ref, uint32_t mask,
drivers/gpu/drm/amd/amdgpu/amdgpu_mes.c
537
uint32_t ref, uint32_t mask,
drivers/gpu/drm/amd/amdgpu/amdgpu_mes.c
547
op_input.wrm_reg.ref = ref;
drivers/gpu/drm/amd/amdgpu/amdgpu_mes.h
351
uint32_t ref;
drivers/gpu/drm/amd/amdgpu/amdgpu_mes.h
461
uint32_t ref, uint32_t mask, uint32_t xcc_id);
drivers/gpu/drm/amd/amdgpu/amdgpu_reset.c
263
void amdgpu_reset_destroy_reset_domain(struct kref *ref)
drivers/gpu/drm/amd/amdgpu/amdgpu_reset.c
265
struct amdgpu_reset_domain *reset_domain = container_of(ref,
drivers/gpu/drm/amd/amdgpu/amdgpu_reset.h
121
void amdgpu_reset_destroy_reset_domain(struct kref *ref);
drivers/gpu/drm/amd/amdgpu/amdgpu_ring.c
455
uint32_t ref, uint32_t mask)
drivers/gpu/drm/amd/amdgpu/amdgpu_ring.c
457
amdgpu_ring_emit_wreg(ring, reg0, ref);
drivers/gpu/drm/amd/amdgpu/amdgpu_ring.h
286
uint32_t ref, uint32_t mask);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
185
void amdgpu_userq_fence_driver_destroy(struct kref *ref)
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
187
struct amdgpu_userq_fence_driver *fence_drv = container_of(ref,
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.h
71
void amdgpu_userq_fence_driver_destroy(struct kref *ref);
drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
4013
uint32_t addr1, uint32_t ref, uint32_t mask,
drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
4028
amdgpu_ring_write(ring, ref);
drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
9025
uint32_t ref, uint32_t mask)
drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
9035
ref, mask, 0x20);
drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
9038
ref, mask);
drivers/gpu/drm/amd/amdgpu/gfx_v11_0.c
528
uint32_t addr1, uint32_t ref, uint32_t mask,
drivers/gpu/drm/amd/amdgpu/gfx_v11_0.c
543
amdgpu_ring_write(ring, ref);
drivers/gpu/drm/amd/amdgpu/gfx_v11_0.c
6305
uint32_t ref, uint32_t mask)
drivers/gpu/drm/amd/amdgpu/gfx_v11_0.c
6310
ref, mask, 0x20);
drivers/gpu/drm/amd/amdgpu/gfx_v12_0.c
431
uint32_t addr1, uint32_t ref,
drivers/gpu/drm/amd/amdgpu/gfx_v12_0.c
446
amdgpu_ring_write(ring, ref);
drivers/gpu/drm/amd/amdgpu/gfx_v12_0.c
4680
uint32_t ref, uint32_t mask)
drivers/gpu/drm/amd/amdgpu/gfx_v12_0.c
4685
ref, mask, 0x20);
drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c
235
uint32_t addr1, uint32_t ref,
drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c
255
amdgpu_ring_write(ring, ref);
drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c
3494
uint32_t ref, uint32_t mask)
drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c
3499
ref, mask, 0x20);
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
1174
uint32_t addr1, uint32_t ref, uint32_t mask,
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
1189
amdgpu_ring_write(ring, ref);
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
5925
uint32_t ref, uint32_t mask)
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
5934
ref, mask, 0x20);
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
5937
ref, mask);
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c
3030
uint32_t ref, uint32_t mask)
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c
3033
ref, mask);
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c
390
uint32_t addr1, uint32_t ref, uint32_t mask,
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c
411
amdgpu_ring_write(ring, ref);
drivers/gpu/drm/amd/amdgpu/mes_v11_0.c
625
misc_pkt.wait_reg_mem.reference = input->wrm_reg.ref;
drivers/gpu/drm/amd/amdgpu/mes_v11_0.c
633
misc_pkt.wait_reg_mem.reference = input->wrm_reg.ref;
drivers/gpu/drm/amd/amdgpu/mes_v12_0.c
666
misc_pkt.wait_reg_mem.reference = input->wrm_reg.ref;
drivers/gpu/drm/amd/amdgpu/mes_v12_0.c
674
misc_pkt.wait_reg_mem.reference = input->wrm_reg.ref;
drivers/gpu/drm/amd/amdgpu/mes_v12_1.c
562
misc_pkt.wait_reg_mem.reference = input->wrm_reg.ref;
drivers/gpu/drm/amd/amdgpu/mes_v12_1.c
577
misc_pkt.wait_reg_mem.reference = input->wrm_reg.ref;
drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c
831
uint32_t ref, uint32_t mask,
drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c
847
amdgpu_ring_write(ring, ref); /* reference */
drivers/gpu/drm/amd/amdgpu/sdma_v4_4_2.c
398
uint32_t ref, uint32_t mask,
drivers/gpu/drm/amd/amdgpu/sdma_v4_4_2.c
414
amdgpu_ring_write(ring, ref); /* reference */
drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c
1322
uint32_t ref, uint32_t mask)
drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c
1324
amdgpu_ring_emit_wreg(ring, reg0, ref);
drivers/gpu/drm/amd/amdgpu/sdma_v5_2.c
1243
uint32_t ref, uint32_t mask)
drivers/gpu/drm/amd/amdgpu/sdma_v5_2.c
1245
amdgpu_ring_emit_wreg(ring, reg0, ref);
drivers/gpu/drm/amd/amdgpu/sdma_v6_0.c
1247
uint32_t ref, uint32_t mask)
drivers/gpu/drm/amd/amdgpu/sdma_v6_0.c
1249
amdgpu_ring_emit_wreg(ring, reg0, ref);
drivers/gpu/drm/amd/amdgpu/sdma_v7_0.c
1248
uint32_t ref, uint32_t mask)
drivers/gpu/drm/amd/amdgpu/sdma_v7_0.c
1250
amdgpu_ring_emit_wreg(ring, reg0, ref);
drivers/gpu/drm/amd/amdgpu/sdma_v7_1.c
1243
uint32_t ref, uint32_t mask)
drivers/gpu/drm/amd/amdgpu/sdma_v7_1.c
1245
amdgpu_ring_emit_wreg(ring, reg0, ref);
drivers/gpu/drm/amd/amdkfd/kfd_debug.c
827
kref_get(&target->ref);
drivers/gpu/drm/amd/amdkfd/kfd_priv.h
910
struct kref ref;
drivers/gpu/drm/amd/amdkfd/kfd_process.c
1000
bool ref)
drivers/gpu/drm/amd/amdkfd/kfd_process.c
1007
if (p && ref)
drivers/gpu/drm/amd/amdkfd/kfd_process.c
1008
kref_get(&p->ref);
drivers/gpu/drm/amd/amdkfd/kfd_process.c
1016
kref_put(&p->ref, kfd_process_ref_release);
drivers/gpu/drm/amd/amdkfd/kfd_process.c
1261
static void kfd_process_ref_release(struct kref *ref)
drivers/gpu/drm/amd/amdkfd/kfd_process.c
1263
struct kfd_process *p = container_of(ref, struct kfd_process, ref);
drivers/gpu/drm/amd/amdkfd/kfd_process.c
1599
kref_init(&process->ref);
drivers/gpu/drm/amd/amdkfd/kfd_process.c
1642
kref_get(&process->ref);
drivers/gpu/drm/amd/amdkfd/kfd_process.c
1933
kref_get(&ret_p->process->ref);
drivers/gpu/drm/amd/amdkfd/kfd_process.c
1956
kref_get(&p->ref);
drivers/gpu/drm/amd/amdkfd/kfd_process.c
1973
kref_get(&p->ref);
drivers/gpu/drm/amd/amdkfd/kfd_process.c
69
bool ref);
drivers/gpu/drm/amd/amdkfd/kfd_process.c
70
static void kfd_process_ref_release(struct kref *ref);
drivers/gpu/drm/display/drm_dp_aux_dev.c
100
container_of(ref, struct drm_dp_aux_dev, refcount);
drivers/gpu/drm/display/drm_dp_aux_dev.c
97
static void release_drm_dp_aux_dev(struct kref *ref)
drivers/gpu/drm/drm_atomic.c
128
kref_init(&state->ref);
drivers/gpu/drm/drm_atomic.c
325
void __drm_atomic_state_free(struct kref *ref)
drivers/gpu/drm/drm_atomic.c
327
struct drm_atomic_state *state = container_of(ref, typeof(*state), ref);
drivers/gpu/drm/drm_atomic.c
53
container_of(kref, struct drm_crtc_commit, ref);
drivers/gpu/drm/drm_atomic_helper.c
2440
kref_init(&commit->ref);
drivers/gpu/drm/drm_atomic_uapi.c
1177
struct drm_mode_object *ref;
drivers/gpu/drm/drm_atomic_uapi.c
1181
if (!drm_property_change_valid_get(prop, prop_value, &ref))
drivers/gpu/drm/drm_atomic_uapi.c
1293
drm_property_change_valid_put(prop, ref);
drivers/gpu/drm/drm_crtc_internal.h
138
struct drm_mode_object **ref);
drivers/gpu/drm/drm_crtc_internal.h
140
struct drm_mode_object *ref);
drivers/gpu/drm/drm_drv.c
716
kref_init(&dev->ref);
drivers/gpu/drm/drm_drv.c
901
static void drm_dev_release(struct kref *ref)
drivers/gpu/drm/drm_drv.c
903
struct drm_device *dev = container_of(ref, struct drm_device, ref);
drivers/gpu/drm/drm_drv.c
931
kref_get(&dev->ref);
drivers/gpu/drm/drm_drv.c
945
kref_put(&dev->ref, drm_dev_release);
drivers/gpu/drm/drm_internal.h
109
static inline bool drm_vblank_passed(u64 seq, u64 ref)
drivers/gpu/drm/drm_internal.h
111
return (seq - ref) <= (1 << 23);
drivers/gpu/drm/drm_mode_object.c
532
struct drm_mode_object *ref;
drivers/gpu/drm/drm_mode_object.c
536
if (!drm_property_change_valid_get(prop, prop_value, &ref))
drivers/gpu/drm/drm_mode_object.c
552
drm_property_change_valid_put(prop, ref);
drivers/gpu/drm/drm_pagemap.c
128
static void drm_pagemap_zdd_destroy(struct kref *ref)
drivers/gpu/drm/drm_pagemap.c
131
container_of(ref, struct drm_pagemap_zdd, refcount);
drivers/gpu/drm/drm_pagemap.c
805
static void drm_pagemap_release(struct kref *ref)
drivers/gpu/drm/drm_pagemap.c
807
struct drm_pagemap *dpagemap = container_of(ref, typeof(*dpagemap), ref);
drivers/gpu/drm/drm_pagemap.c
882
kref_init(&dpagemap->ref);
drivers/gpu/drm/drm_pagemap.c
907
kref_init(&dpagemap->ref);
drivers/gpu/drm/drm_pagemap.c
929
kref_put(&dpagemap->ref, drm_pagemap_release);
drivers/gpu/drm/drm_property.c
1001
struct drm_mode_object *ref)
drivers/gpu/drm/drm_property.c
1003
if (!ref)
drivers/gpu/drm/drm_property.c
1007
drm_mode_object_put(ref);
drivers/gpu/drm/drm_property.c
1009
drm_property_blob_put(obj_to_blob(ref));
drivers/gpu/drm/drm_property.c
945
uint64_t value, struct drm_mode_object **ref)
drivers/gpu/drm/drm_property.c
952
*ref = NULL;
drivers/gpu/drm/drm_property.c
979
*ref = &blob->base;
drivers/gpu/drm/drm_property.c
989
*ref = __drm_mode_object_find(property->dev, NULL, value,
drivers/gpu/drm/drm_property.c
991
return *ref != NULL;
drivers/gpu/drm/i915/display/intel_audio.c
928
get_aud_ts_cdclk_m_n(display->cdclk.hw.ref,
drivers/gpu/drm/i915/display/intel_cdclk.c
1013
cdclk_config->cdclk = cdclk_config->bypass = cdclk_config->ref;
drivers/gpu/drm/i915/display/intel_cdclk.c
1591
if (table[i].refclk == display->cdclk.hw.ref &&
drivers/gpu/drm/i915/display/intel_cdclk.c
1597
min_cdclk, display->cdclk.hw.ref);
drivers/gpu/drm/i915/display/intel_cdclk.c
1610
if (table[i].refclk == display->cdclk.hw.ref &&
drivers/gpu/drm/i915/display/intel_cdclk.c
1612
return display->cdclk.hw.ref * table[i].ratio;
drivers/gpu/drm/i915/display/intel_cdclk.c
1615
cdclk, display->cdclk.hw.ref);
drivers/gpu/drm/i915/display/intel_cdclk.c
1717
cdclk_config->ref = 24000;
drivers/gpu/drm/i915/display/intel_cdclk.c
1720
cdclk_config->ref = 19200;
drivers/gpu/drm/i915/display/intel_cdclk.c
1723
cdclk_config->ref = 38400;
drivers/gpu/drm/i915/display/intel_cdclk.c
1734
cdclk_config->ref = 38400;
drivers/gpu/drm/i915/display/intel_cdclk.c
1738
cdclk_config->ref = 19200;
drivers/gpu/drm/i915/display/intel_cdclk.c
1760
cdclk_config->vco = ratio * cdclk_config->ref;
drivers/gpu/drm/i915/display/intel_cdclk.c
1773
cdclk_config->bypass = cdclk_config->ref / 2;
drivers/gpu/drm/i915/display/intel_cdclk.c
1777
cdclk_config->bypass = cdclk_config->ref;
drivers/gpu/drm/i915/display/intel_cdclk.c
1845
int ratio = DIV_ROUND_CLOSEST(vco, display->cdclk.hw.ref);
drivers/gpu/drm/i915/display/intel_cdclk.c
1888
int ratio = DIV_ROUND_CLOSEST(vco, display->cdclk.hw.ref);
drivers/gpu/drm/i915/display/intel_cdclk.c
1906
int ratio = DIV_ROUND_CLOSEST(vco, display->cdclk.hw.ref);
drivers/gpu/drm/i915/display/intel_cdclk.c
1994
if (table[i].refclk == display->cdclk.hw.ref &&
drivers/gpu/drm/i915/display/intel_cdclk.c
1999
cdclk, display->cdclk.hw.ref);
drivers/gpu/drm/i915/display/intel_cdclk.c
2482
a->ref == b->ref;
drivers/gpu/drm/i915/display/intel_cdclk.c
2501
a->ref == b->ref;
drivers/gpu/drm/i915/display/intel_cdclk.c
2518
a->ref != b->ref;
drivers/gpu/drm/i915/display/intel_cdclk.c
2552
a->ref == b->ref;
drivers/gpu/drm/i915/display/intel_cdclk.c
2576
cdclk_config->ref, cdclk_config->bypass,
drivers/gpu/drm/i915/display/intel_cdclk.c
3699
if (display->cdclk.hw.ref == 24000)
drivers/gpu/drm/i915/display/intel_cdclk.c
3704
if (display->cdclk.hw.ref == 24000)
drivers/gpu/drm/i915/display/intel_cdclk.c
970
cdclk_config->ref = 24000;
drivers/gpu/drm/i915/display/intel_cdclk.h
19
unsigned int cdclk, vco, ref, bypass;
drivers/gpu/drm/i915/display/intel_cx0_phy.c
2721
unsigned int ref, vco;
drivers/gpu/drm/i915/display/intel_cx0_phy.c
2752
ref = DIV_ROUND_CLOSEST(refclk * (1 << (1 + fb_clk_div4_en)), 1 << ref_clk_mpllb_div);
drivers/gpu/drm/i915/display/intel_cx0_phy.c
2753
vco = DIV_ROUND_CLOSEST_ULL(mul_u32_u32(ref, (multiplier << (17 - 2)) + frac) >> 17, 10);
drivers/gpu/drm/i915/display/intel_dp_tunnel.c
20
struct drm_dp_tunnel_ref ref[I915_MAX_PIPES];
drivers/gpu/drm/i915/display/intel_dp_tunnel.c
366
return state->inherited_dp_tunnels->ref[crtc->pipe].tunnel;
drivers/gpu/drm/i915/display/intel_dp_tunnel.c
389
drm_dp_tunnel_ref_get(tunnel, &state->inherited_dp_tunnels->ref[crtc->pipe]);
drivers/gpu/drm/i915/display/intel_dp_tunnel.c
452
if (state->inherited_dp_tunnels->ref[pipe].tunnel)
drivers/gpu/drm/i915/display/intel_dp_tunnel.c
453
drm_dp_tunnel_ref_put(&state->inherited_dp_tunnels->ref[pipe]);
drivers/gpu/drm/i915/display/intel_dpll_mgr.c
2000
display->dpll.ref_clks.nssc = display->cdclk.hw.ref;
drivers/gpu/drm/i915/display/intel_dpll_mgr.c
4136
display->dpll.ref_clks.nssc = display->cdclk.hw.ref;
drivers/gpu/drm/i915/display/intel_global_state.c
100
kref_put(&obj_state->ref, __intel_atomic_global_state_free);
drivers/gpu/drm/i915/display/intel_global_state.c
106
kref_get(&obj_state->ref);
drivers/gpu/drm/i915/display/intel_global_state.c
120
kref_init(&state->ref);
drivers/gpu/drm/i915/display/intel_global_state.c
134
drm_WARN_ON(display->drm, kref_read(&obj->state->ref) != 1);
drivers/gpu/drm/i915/display/intel_global_state.c
209
kref_init(&obj_state->ref);
drivers/gpu/drm/i915/display/intel_global_state.c
47
struct kref ref;
drivers/gpu/drm/i915/display/intel_global_state.c
60
kref_init(&commit->ref);
drivers/gpu/drm/i915/display/intel_global_state.c
68
container_of(kref, typeof(*commit), ref);
drivers/gpu/drm/i915/display/intel_global_state.c
76
kref_get(&commit->ref);
drivers/gpu/drm/i915/display/intel_global_state.c
84
kref_put(&commit->ref, __commit_free);
drivers/gpu/drm/i915/display/intel_global_state.c
90
container_of(kref, struct intel_global_state, ref);
drivers/gpu/drm/i915/display/intel_global_state.h
34
struct kref ref;
drivers/gpu/drm/i915/gem/i915_gem_context.c
1315
void i915_gem_context_release(struct kref *ref)
drivers/gpu/drm/i915/gem/i915_gem_context.c
1317
struct i915_gem_context *ctx = container_of(ref, typeof(*ctx), ref);
drivers/gpu/drm/i915/gem/i915_gem_context.c
1616
kref_init(&ctx->ref);
drivers/gpu/drm/i915/gem/i915_gem_context.c
2301
if (ctx && !kref_get_unless_zero(&ctx->ref))
drivers/gpu/drm/i915/gem/i915_gem_context.c
360
kref_get(&vm->ref);
drivers/gpu/drm/i915/gem/i915_gem_context.h
147
kref_get(&ctx->ref);
drivers/gpu/drm/i915/gem/i915_gem_context.h
153
kref_put(&ctx->ref, i915_gem_context_release);
drivers/gpu/drm/i915/gem/i915_gem_context_types.h
313
struct kref ref;
drivers/gpu/drm/i915/gem/i915_gem_object_frontbuffer.c
10
container_of(ref, typeof(*front), write);
drivers/gpu/drm/i915/gem/i915_gem_object_frontbuffer.c
101
kref_put_lock(&front->ref, frontbuffer_release,
drivers/gpu/drm/i915/gem/i915_gem_object_frontbuffer.c
12
kref_get(&front->ref);
drivers/gpu/drm/i915/gem/i915_gem_object_frontbuffer.c
16
static void frontbuffer_retire(struct i915_active *ref)
drivers/gpu/drm/i915/gem/i915_gem_object_frontbuffer.c
19
container_of(ref, typeof(*front), write);
drivers/gpu/drm/i915/gem/i915_gem_object_frontbuffer.c
41
kref_init(&front->ref);
drivers/gpu/drm/i915/gem/i915_gem_object_frontbuffer.c
53
kref_get(&cur->ref);
drivers/gpu/drm/i915/gem/i915_gem_object_frontbuffer.c
7
static int frontbuffer_active(struct i915_active *ref)
drivers/gpu/drm/i915/gem/i915_gem_object_frontbuffer.c
71
kref_get(&front->ref);
drivers/gpu/drm/i915/gem/i915_gem_object_frontbuffer.c
74
static void frontbuffer_release(struct kref *ref)
drivers/gpu/drm/i915/gem/i915_gem_object_frontbuffer.c
78
container_of(ref, typeof(*front), ref);
drivers/gpu/drm/i915/gem/i915_gem_object_frontbuffer.h
20
struct kref ref;
drivers/gpu/drm/i915/gem/i915_gem_object_frontbuffer.h
71
if (unlikely(!kref_get_unless_zero(&front->ref)))
drivers/gpu/drm/i915/gem/i915_gem_throttle.c
56
if (!kref_get_unless_zero(&ctx->ref))
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
253
static void i915_ttm_tt_release(struct kref *ref)
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
256
container_of(ref, typeof(*i915_tt), cached_rsgt.kref);
drivers/gpu/drm/i915/gem/selftests/mock_context.c
24
kref_init(&ctx->ref);
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
329
GEM_BUG_ON(!kref_read(&ppgtt->base.vm.ref));
drivers/gpu/drm/i915/gt/intel_breadcrumbs.c
286
kref_init(&b->ref);
drivers/gpu/drm/i915/gt/intel_breadcrumbs.c
330
struct intel_breadcrumbs *b = container_of(kref, typeof(*b), ref);
drivers/gpu/drm/i915/gt/intel_breadcrumbs.h
54
kref_get(&b->ref);
drivers/gpu/drm/i915/gt/intel_breadcrumbs.h
60
kref_put(&b->ref, intel_breadcrumbs_free);
drivers/gpu/drm/i915/gt/intel_breadcrumbs_types.h
36
struct kref ref;
drivers/gpu/drm/i915/gt/intel_context.c
397
kref_init(&ce->ref);
drivers/gpu/drm/i915/gt/intel_context.c
76
if (ctx && !kref_get_unless_zero(&ctx->ref))
drivers/gpu/drm/i915/gt/intel_context.h
238
kref_get(&ce->ref);
drivers/gpu/drm/i915/gt/intel_context.h
244
kref_put(&ce->ref, ce->ops->destroy);
drivers/gpu/drm/i915/gt/intel_context_types.h
236
atomic_t ref;
drivers/gpu/drm/i915/gt/intel_context_types.h
85
struct kref ref; /* no kref_get_unless_zero()! */
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
3665
container_of(kref, typeof(*ve), context.ref);
drivers/gpu/drm/i915/gt/intel_gt_buffer_pool.c
102
container_of(ref, typeof(*node), active);
drivers/gpu/drm/i915/gt/intel_gt_buffer_pool.c
99
static void pool_retire(struct i915_active *ref)
drivers/gpu/drm/i915/gt/intel_gtt.c
253
container_of(kref, struct i915_address_space, ref);
drivers/gpu/drm/i915/gt/intel_gtt.c
263
kref_init(&vm->ref);
drivers/gpu/drm/i915/gt/intel_gtt.h
248
struct kref ref;
drivers/gpu/drm/i915/gt/intel_gtt.h
483
kref_get(&vm->ref);
drivers/gpu/drm/i915/gt/intel_gtt.h
490
return kref_get_unless_zero(&vm->ref) ? vm : NULL;
drivers/gpu/drm/i915/gt/intel_gtt.h
495
GEM_BUG_ON(!kref_read(&vm->ref));
drivers/gpu/drm/i915/gt/intel_gtt.h
516
kref_put(&vm->ref, i915_vm_release);
drivers/gpu/drm/i915/gt/intel_lrc.c
1250
struct intel_context *ce = container_of(kref, typeof(*ce), ref);
drivers/gpu/drm/i915/gt/intel_reset.c
75
if (ctx && !kref_get_unless_zero(&ctx->ref))
drivers/gpu/drm/i915/gt/intel_ring.c
159
kref_init(&ring->ref);
drivers/gpu/drm/i915/gt/intel_ring.c
184
void intel_ring_free(struct kref *ref)
drivers/gpu/drm/i915/gt/intel_ring.c
186
struct intel_ring *ring = container_of(ref, typeof(*ring), ref);
drivers/gpu/drm/i915/gt/intel_ring.h
27
void intel_ring_free(struct kref *ref);
drivers/gpu/drm/i915/gt/intel_ring.h
31
kref_get(&ring->ref);
drivers/gpu/drm/i915/gt/intel_ring.h
37
kref_put(&ring->ref, intel_ring_free);
drivers/gpu/drm/i915/gt/intel_ring_submission.c
490
static void ring_context_destroy(struct kref *ref)
drivers/gpu/drm/i915/gt/intel_ring_submission.c
492
struct intel_context *ce = container_of(ref, typeof(*ce), ref);
drivers/gpu/drm/i915/gt/intel_ring_types.h
25
struct kref ref;
drivers/gpu/drm/i915/gt/intel_rps.c
2778
if (i915 && !kref_get_unless_zero(&i915->drm.ref))
drivers/gpu/drm/i915/gt/mock_engine.c
148
static void mock_context_destroy(struct kref *ref)
drivers/gpu/drm/i915/gt/mock_engine.c
150
struct intel_context *ce = container_of(ref, typeof(*ce), ref);
drivers/gpu/drm/i915/gt/mock_engine.c
69
kref_init(&ring->ref);
drivers/gpu/drm/i915/gt/selftest_lrc.c
1271
struct i915_vma *ref[2],
drivers/gpu/drm/i915/gt/selftest_lrc.c
1281
A[0] = i915_gem_object_pin_map_unlocked(ref[0]->obj, I915_MAP_WC);
drivers/gpu/drm/i915/gt/selftest_lrc.c
1285
A[1] = i915_gem_object_pin_map_unlocked(ref[1]->obj, I915_MAP_WC);
drivers/gpu/drm/i915/gt/selftest_lrc.c
1380
i915_gem_object_unpin_map(ref[1]->obj);
drivers/gpu/drm/i915/gt/selftest_lrc.c
1382
i915_gem_object_unpin_map(ref[0]->obj);
drivers/gpu/drm/i915/gt/selftest_lrc.c
1413
struct i915_vma *ref[2], *result[2];
drivers/gpu/drm/i915/gt/selftest_lrc.c
1428
ref[0] = create_result_vma(A->vm, SZ_64K);
drivers/gpu/drm/i915/gt/selftest_lrc.c
1429
if (IS_ERR(ref[0])) {
drivers/gpu/drm/i915/gt/selftest_lrc.c
1430
err = PTR_ERR(ref[0]);
drivers/gpu/drm/i915/gt/selftest_lrc.c
1434
ref[1] = create_result_vma(A->vm, SZ_64K);
drivers/gpu/drm/i915/gt/selftest_lrc.c
1435
if (IS_ERR(ref[1])) {
drivers/gpu/drm/i915/gt/selftest_lrc.c
1436
err = PTR_ERR(ref[1]);
drivers/gpu/drm/i915/gt/selftest_lrc.c
1440
rq = record_registers(A, ref[0], ref[1], sema);
drivers/gpu/drm/i915/gt/selftest_lrc.c
1487
err = compare_isolation(engine, ref, result, A, poison);
drivers/gpu/drm/i915/gt/selftest_lrc.c
1494
i915_vma_put(ref[1]);
drivers/gpu/drm/i915/gt/selftest_lrc.c
1496
i915_vma_put(ref[0]);
drivers/gpu/drm/i915/gt/selftest_ring.c
14
kref_init(&ring->ref);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
1090
bool do_put = kref_get_unless_zero(&ce->ref);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
1940
if (!kref_get_unless_zero(&ce->ref))
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2033
if (!kref_get_unless_zero(&ce->ref))
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2335
GEM_BUG_ON(atomic_read(&cn->guc_id.ref));
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2393
GEM_BUG_ON(atomic_read(&ce->guc_id.ref));
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2408
atomic_inc(&ce->guc_id.ref);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2442
GEM_BUG_ON(atomic_read(&ce->guc_id.ref) < 0);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2451
!atomic_read(&ce->guc_id.ref))
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
3582
struct intel_context *ce = container_of(kref, typeof(*ce), ref);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
3776
atomic_dec(&ce->guc_id.ref);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
3947
if (atomic_add_unless(&ce->guc_id.ref, 1, 0))
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
3960
atomic_dec(&ce->guc_id.ref);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
4144
struct intel_context *ce = container_of(kref, typeof(*ce), ref);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
4620
container_of(kref, typeof(*sched_engine), ref);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
5404
if (!kref_get_unless_zero(&ce->ref))
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
5463
if (!kref_get_unless_zero(&ce->ref))
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
5561
atomic_read(&ce->guc_id.ref));
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
717
GEM_BUG_ON(!atomic_read(&ce->guc_id.ref));
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
861
GEM_BUG_ON(!atomic_read(&ce->guc_id.ref));
drivers/gpu/drm/i915/gvt/gtt.c
1778
kref_init(&mm->ref);
drivers/gpu/drm/i915/gvt/gtt.c
1890
struct intel_vgpu_mm *mm = container_of(mm_ref, typeof(*mm), ref);
drivers/gpu/drm/i915/gvt/gtt.h
153
struct kref ref;
drivers/gpu/drm/i915/gvt/gtt.h
187
kref_get(&mm->ref);
drivers/gpu/drm/i915/gvt/gtt.h
194
kref_put(&mm->ref, _intel_vgpu_mm_release);
drivers/gpu/drm/i915/gvt/kvmgt.c
104
struct kref ref;
drivers/gpu/drm/i915/gvt/kvmgt.c
1642
kref_get(&entry->ref);
drivers/gpu/drm/i915/gvt/kvmgt.c
1667
kref_get(&entry->ref);
drivers/gpu/drm/i915/gvt/kvmgt.c
1675
static void __gvt_dma_release(struct kref *ref)
drivers/gpu/drm/i915/gvt/kvmgt.c
1677
struct gvt_dma *entry = container_of(ref, typeof(*entry), ref);
drivers/gpu/drm/i915/gvt/kvmgt.c
1695
kref_put(&entry->ref, __gvt_dma_release);
drivers/gpu/drm/i915/gvt/kvmgt.c
261
kref_init(&new->ref);
drivers/gpu/drm/i915/i915_active.c
100
lockdep_assert_held(&ref->tree_lock);
drivers/gpu/drm/i915/i915_active.c
101
if (!atomic_read(&ref->count)) /* after the last dec */
drivers/gpu/drm/i915/i915_active.c
102
debug_object_deactivate(ref, &active_debug_desc);
drivers/gpu/drm/i915/i915_active.c
105
static void debug_active_fini(struct i915_active *ref)
drivers/gpu/drm/i915/i915_active.c
107
debug_object_free(ref, &active_debug_desc);
drivers/gpu/drm/i915/i915_active.c
110
static void debug_active_assert(struct i915_active *ref)
drivers/gpu/drm/i915/i915_active.c
112
debug_object_assert_init(ref, &active_debug_desc);
drivers/gpu/drm/i915/i915_active.c
1122
struct kref ref;
drivers/gpu/drm/i915/i915_active.c
1125
struct i915_active *i915_active_get(struct i915_active *ref)
drivers/gpu/drm/i915/i915_active.c
1127
struct auto_active *aa = container_of(ref, typeof(*aa), base);
drivers/gpu/drm/i915/i915_active.c
1129
kref_get(&aa->ref);
drivers/gpu/drm/i915/i915_active.c
1133
static void auto_release(struct kref *ref)
drivers/gpu/drm/i915/i915_active.c
1135
struct auto_active *aa = container_of(ref, typeof(*aa), ref);
drivers/gpu/drm/i915/i915_active.c
1141
void i915_active_put(struct i915_active *ref)
drivers/gpu/drm/i915/i915_active.c
1143
struct auto_active *aa = container_of(ref, typeof(*aa), base);
drivers/gpu/drm/i915/i915_active.c
1145
kref_put(&aa->ref, auto_release);
drivers/gpu/drm/i915/i915_active.c
1148
static int auto_active(struct i915_active *ref)
drivers/gpu/drm/i915/i915_active.c
1150
i915_active_get(ref);
drivers/gpu/drm/i915/i915_active.c
1154
static void auto_retire(struct i915_active *ref)
drivers/gpu/drm/i915/i915_active.c
1156
i915_active_put(ref);
drivers/gpu/drm/i915/i915_active.c
1167
kref_init(&aa->ref);
drivers/gpu/drm/i915/i915_active.c
117
static inline void debug_active_init(struct i915_active *ref) { }
drivers/gpu/drm/i915/i915_active.c
118
static inline void debug_active_activate(struct i915_active *ref) { }
drivers/gpu/drm/i915/i915_active.c
119
static inline void debug_active_deactivate(struct i915_active *ref) { }
drivers/gpu/drm/i915/i915_active.c
120
static inline void debug_active_fini(struct i915_active *ref) { }
drivers/gpu/drm/i915/i915_active.c
121
static inline void debug_active_assert(struct i915_active *ref) { }
drivers/gpu/drm/i915/i915_active.c
126
__active_retire(struct i915_active *ref)
drivers/gpu/drm/i915/i915_active.c
132
GEM_BUG_ON(i915_active_is_idle(ref));
drivers/gpu/drm/i915/i915_active.c
135
if (!atomic_dec_and_lock_irqsave(&ref->count, &ref->tree_lock, flags))
drivers/gpu/drm/i915/i915_active.c
138
GEM_BUG_ON(rcu_access_pointer(ref->excl.fence));
drivers/gpu/drm/i915/i915_active.c
139
debug_active_deactivate(ref);
drivers/gpu/drm/i915/i915_active.c
142
if (!ref->cache)
drivers/gpu/drm/i915/i915_active.c
143
ref->cache = fetch_node(ref->tree.rb_node);
drivers/gpu/drm/i915/i915_active.c
146
if (ref->cache) {
drivers/gpu/drm/i915/i915_active.c
148
rb_erase(&ref->cache->node, &ref->tree);
drivers/gpu/drm/i915/i915_active.c
149
root = ref->tree;
drivers/gpu/drm/i915/i915_active.c
152
rb_link_node(&ref->cache->node, NULL, &ref->tree.rb_node);
drivers/gpu/drm/i915/i915_active.c
153
rb_insert_color(&ref->cache->node, &ref->tree);
drivers/gpu/drm/i915/i915_active.c
154
GEM_BUG_ON(ref->tree.rb_node != &ref->cache->node);
drivers/gpu/drm/i915/i915_active.c
157
ref->cache->timeline = 0; /* needs cmpxchg(u64) */
drivers/gpu/drm/i915/i915_active.c
160
spin_unlock_irqrestore(&ref->tree_lock, flags);
drivers/gpu/drm/i915/i915_active.c
163
if (ref->retire)
drivers/gpu/drm/i915/i915_active.c
164
ref->retire(ref);
drivers/gpu/drm/i915/i915_active.c
167
wake_up_var(ref);
drivers/gpu/drm/i915/i915_active.c
179
struct i915_active *ref = container_of(wrk, typeof(*ref), work);
drivers/gpu/drm/i915/i915_active.c
181
GEM_BUG_ON(!atomic_read(&ref->count));
drivers/gpu/drm/i915/i915_active.c
182
if (atomic_add_unless(&ref->count, -1, 1))
drivers/gpu/drm/i915/i915_active.c
185
__active_retire(ref);
drivers/gpu/drm/i915/i915_active.c
189
active_retire(struct i915_active *ref)
drivers/gpu/drm/i915/i915_active.c
191
GEM_BUG_ON(!atomic_read(&ref->count));
drivers/gpu/drm/i915/i915_active.c
192
if (atomic_add_unless(&ref->count, -1, 1))
drivers/gpu/drm/i915/i915_active.c
195
if (ref->flags & I915_ACTIVE_RETIRE_SLEEPS) {
drivers/gpu/drm/i915/i915_active.c
196
queue_work(system_unbound_wq, &ref->work);
drivers/gpu/drm/i915/i915_active.c
200
__active_retire(ref);
drivers/gpu/drm/i915/i915_active.c
222
active_retire(container_of(cb, struct active_node, base.cb)->ref);
drivers/gpu/drm/i915/i915_active.c
232
static struct active_node *__active_lookup(struct i915_active *ref, u64 idx)
drivers/gpu/drm/i915/i915_active.c
245
it = READ_ONCE(ref->cache);
drivers/gpu/drm/i915/i915_active.c
269
GEM_BUG_ON(i915_active_is_idle(ref));
drivers/gpu/drm/i915/i915_active.c
271
it = fetch_node(ref->tree.rb_node);
drivers/gpu/drm/i915/i915_active.c
278
WRITE_ONCE(ref->cache, it);
drivers/gpu/drm/i915/i915_active.c
288
active_instance(struct i915_active *ref, u64 idx)
drivers/gpu/drm/i915/i915_active.c
29
struct i915_active *ref;
drivers/gpu/drm/i915/i915_active.c
293
node = __active_lookup(ref, idx);
drivers/gpu/drm/i915/i915_active.c
297
spin_lock_irq(&ref->tree_lock);
drivers/gpu/drm/i915/i915_active.c
298
GEM_BUG_ON(i915_active_is_idle(ref));
drivers/gpu/drm/i915/i915_active.c
301
p = &ref->tree.rb_node;
drivers/gpu/drm/i915/i915_active.c
324
node->ref = ref;
drivers/gpu/drm/i915/i915_active.c
328
rb_insert_color(&node->node, &ref->tree);
drivers/gpu/drm/i915/i915_active.c
331
WRITE_ONCE(ref->cache, node);
drivers/gpu/drm/i915/i915_active.c
332
spin_unlock_irq(&ref->tree_lock);
drivers/gpu/drm/i915/i915_active.c
337
void __i915_active_init(struct i915_active *ref,
drivers/gpu/drm/i915/i915_active.c
338
int (*active)(struct i915_active *ref),
drivers/gpu/drm/i915/i915_active.c
339
void (*retire)(struct i915_active *ref),
drivers/gpu/drm/i915/i915_active.c
344
debug_active_init(ref);
drivers/gpu/drm/i915/i915_active.c
346
ref->flags = flags;
drivers/gpu/drm/i915/i915_active.c
347
ref->active = active;
drivers/gpu/drm/i915/i915_active.c
348
ref->retire = retire;
drivers/gpu/drm/i915/i915_active.c
350
spin_lock_init(&ref->tree_lock);
drivers/gpu/drm/i915/i915_active.c
351
ref->tree = RB_ROOT;
drivers/gpu/drm/i915/i915_active.c
352
ref->cache = NULL;
drivers/gpu/drm/i915/i915_active.c
354
init_llist_head(&ref->preallocated_barriers);
drivers/gpu/drm/i915/i915_active.c
355
atomic_set(&ref->count, 0);
drivers/gpu/drm/i915/i915_active.c
356
__mutex_init(&ref->mutex, "i915_active", mkey);
drivers/gpu/drm/i915/i915_active.c
357
__i915_active_fence_init(&ref->excl, NULL, excl_retire);
drivers/gpu/drm/i915/i915_active.c
358
INIT_WORK(&ref->work, active_work);
drivers/gpu/drm/i915/i915_active.c
360
lockdep_init_map(&ref->work.lockdep_map, "i915_active.work", wkey, 0);
drivers/gpu/drm/i915/i915_active.c
364
static bool ____active_del_barrier(struct i915_active *ref,
drivers/gpu/drm/i915/i915_active.c
407
__active_del_barrier(struct i915_active *ref, struct active_node *node)
drivers/gpu/drm/i915/i915_active.c
409
return ____active_del_barrier(ref, node, barrier_to_engine(node));
drivers/gpu/drm/i915/i915_active.c
413
replace_barrier(struct i915_active *ref, struct i915_active_fence *active)
drivers/gpu/drm/i915/i915_active.c
423
return __active_del_barrier(ref, node_from_active(active));
drivers/gpu/drm/i915/i915_active.c
426
int i915_active_add_request(struct i915_active *ref, struct i915_request *rq)
drivers/gpu/drm/i915/i915_active.c
434
err = i915_active_acquire(ref);
drivers/gpu/drm/i915/i915_active.c
439
active = active_instance(ref, idx);
drivers/gpu/drm/i915/i915_active.c
445
if (replace_barrier(ref, active)) {
drivers/gpu/drm/i915/i915_active.c
447
atomic_dec(&ref->count);
drivers/gpu/drm/i915/i915_active.c
453
__i915_active_acquire(ref);
drivers/gpu/drm/i915/i915_active.c
458
i915_active_release(ref);
drivers/gpu/drm/i915/i915_active.c
463
__i915_active_set_fence(struct i915_active *ref,
drivers/gpu/drm/i915/i915_active.c
469
if (replace_barrier(ref, active)) {
drivers/gpu/drm/i915/i915_active.c
476
__i915_active_acquire(ref);
drivers/gpu/drm/i915/i915_active.c
482
i915_active_set_exclusive(struct i915_active *ref, struct dma_fence *f)
drivers/gpu/drm/i915/i915_active.c
485
return __i915_active_set_fence(ref, &ref->excl, f);
drivers/gpu/drm/i915/i915_active.c
488
bool i915_active_acquire_if_busy(struct i915_active *ref)
drivers/gpu/drm/i915/i915_active.c
490
debug_active_assert(ref);
drivers/gpu/drm/i915/i915_active.c
491
return atomic_add_unless(&ref->count, 1, 0);
drivers/gpu/drm/i915/i915_active.c
494
static void __i915_active_activate(struct i915_active *ref)
drivers/gpu/drm/i915/i915_active.c
496
spin_lock_irq(&ref->tree_lock); /* __active_retire() */
drivers/gpu/drm/i915/i915_active.c
497
if (!atomic_fetch_inc(&ref->count))
drivers/gpu/drm/i915/i915_active.c
498
debug_active_activate(ref);
drivers/gpu/drm/i915/i915_active.c
499
spin_unlock_irq(&ref->tree_lock);
drivers/gpu/drm/i915/i915_active.c
502
int i915_active_acquire(struct i915_active *ref)
drivers/gpu/drm/i915/i915_active.c
506
if (i915_active_acquire_if_busy(ref))
drivers/gpu/drm/i915/i915_active.c
509
if (!ref->active) {
drivers/gpu/drm/i915/i915_active.c
510
__i915_active_activate(ref);
drivers/gpu/drm/i915/i915_active.c
514
err = mutex_lock_interruptible(&ref->mutex);
drivers/gpu/drm/i915/i915_active.c
518
if (likely(!i915_active_acquire_if_busy(ref))) {
drivers/gpu/drm/i915/i915_active.c
519
err = ref->active(ref);
drivers/gpu/drm/i915/i915_active.c
521
__i915_active_activate(ref);
drivers/gpu/drm/i915/i915_active.c
524
mutex_unlock(&ref->mutex);
drivers/gpu/drm/i915/i915_active.c
529
void i915_active_release(struct i915_active *ref)
drivers/gpu/drm/i915/i915_active.c
531
debug_active_assert(ref);
drivers/gpu/drm/i915/i915_active.c
532
active_retire(ref);
drivers/gpu/drm/i915/i915_active.c
565
static int flush_lazy_signals(struct i915_active *ref)
drivers/gpu/drm/i915/i915_active.c
570
enable_signaling(&ref->excl);
drivers/gpu/drm/i915/i915_active.c
571
rbtree_postorder_for_each_entry_safe(it, n, &ref->tree, node) {
drivers/gpu/drm/i915/i915_active.c
582
int __i915_active_wait(struct i915_active *ref, int state)
drivers/gpu/drm/i915/i915_active.c
587
if (i915_active_acquire_if_busy(ref)) {
drivers/gpu/drm/i915/i915_active.c
590
err = flush_lazy_signals(ref);
drivers/gpu/drm/i915/i915_active.c
591
i915_active_release(ref);
drivers/gpu/drm/i915/i915_active.c
595
if (___wait_var_event(ref, i915_active_is_idle(ref),
drivers/gpu/drm/i915/i915_active.c
604
flush_work(&ref->work);
drivers/gpu/drm/i915/i915_active.c
632
struct i915_active *ref;
drivers/gpu/drm/i915/i915_active.c
640
if (i915_active_is_idle(wb->ref)) {
drivers/gpu/drm/i915/i915_active.c
649
static int __await_barrier(struct i915_active *ref, struct i915_sw_fence *fence)
drivers/gpu/drm/i915/i915_active.c
657
GEM_BUG_ON(i915_active_is_idle(ref));
drivers/gpu/drm/i915/i915_active.c
666
wb->ref = ref;
drivers/gpu/drm/i915/i915_active.c
668
add_wait_queue(__var_waitqueue(ref), &wb->base);
drivers/gpu/drm/i915/i915_active.c
672
static int await_active(struct i915_active *ref,
drivers/gpu/drm/i915/i915_active.c
679
if (!i915_active_acquire_if_busy(ref))
drivers/gpu/drm/i915/i915_active.c
683
rcu_access_pointer(ref->excl.fence)) {
drivers/gpu/drm/i915/i915_active.c
684
err = __await_active(&ref->excl, fn, arg);
drivers/gpu/drm/i915/i915_active.c
692
rbtree_postorder_for_each_entry_safe(it, n, &ref->tree, node) {
drivers/gpu/drm/i915/i915_active.c
700
err = flush_lazy_signals(ref);
drivers/gpu/drm/i915/i915_active.c
704
err = __await_barrier(ref, barrier);
drivers/gpu/drm/i915/i915_active.c
710
i915_active_release(ref);
drivers/gpu/drm/i915/i915_active.c
720
struct i915_active *ref,
drivers/gpu/drm/i915/i915_active.c
723
return await_active(ref, flags, rq_await_fence, rq, &rq->submit);
drivers/gpu/drm/i915/i915_active.c
733
struct i915_active *ref,
drivers/gpu/drm/i915/i915_active.c
736
return await_active(ref, flags, sw_await_fence, fence, fence);
drivers/gpu/drm/i915/i915_active.c
739
void i915_active_fini(struct i915_active *ref)
drivers/gpu/drm/i915/i915_active.c
741
debug_active_fini(ref);
drivers/gpu/drm/i915/i915_active.c
742
GEM_BUG_ON(atomic_read(&ref->count));
drivers/gpu/drm/i915/i915_active.c
743
GEM_BUG_ON(work_pending(&ref->work));
drivers/gpu/drm/i915/i915_active.c
744
mutex_destroy(&ref->mutex);
drivers/gpu/drm/i915/i915_active.c
746
if (ref->cache)
drivers/gpu/drm/i915/i915_active.c
747
kmem_cache_free(slab_cache, ref->cache);
drivers/gpu/drm/i915/i915_active.c
755
static struct active_node *reuse_idle_barrier(struct i915_active *ref, u64 idx)
drivers/gpu/drm/i915/i915_active.c
759
if (RB_EMPTY_ROOT(&ref->tree))
drivers/gpu/drm/i915/i915_active.c
762
GEM_BUG_ON(i915_active_is_idle(ref));
drivers/gpu/drm/i915/i915_active.c
77
struct i915_active *ref = addr;
drivers/gpu/drm/i915/i915_active.c
771
if (ref->cache && is_idle_barrier(ref->cache, idx)) {
drivers/gpu/drm/i915/i915_active.c
772
p = &ref->cache->node;
drivers/gpu/drm/i915/i915_active.c
777
p = ref->tree.rb_node;
drivers/gpu/drm/i915/i915_active.c
79
return (void *)ref->active ?: (void *)ref->retire ?: (void *)ref;
drivers/gpu/drm/i915/i915_active.c
822
____active_del_barrier(ref, node, engine))
drivers/gpu/drm/i915/i915_active.c
829
spin_lock_irq(&ref->tree_lock);
drivers/gpu/drm/i915/i915_active.c
830
rb_erase(p, &ref->tree); /* Hide from waits and sibling allocations */
drivers/gpu/drm/i915/i915_active.c
831
if (p == &ref->cache->node)
drivers/gpu/drm/i915/i915_active.c
832
WRITE_ONCE(ref->cache, NULL);
drivers/gpu/drm/i915/i915_active.c
833
spin_unlock_irq(&ref->tree_lock);
drivers/gpu/drm/i915/i915_active.c
838
int i915_active_acquire_preallocate_barrier(struct i915_active *ref,
drivers/gpu/drm/i915/i915_active.c
845
GEM_BUG_ON(i915_active_is_idle(ref));
drivers/gpu/drm/i915/i915_active.c
848
while (!llist_empty(&ref->preallocated_barriers))
drivers/gpu/drm/i915/i915_active.c
864
node = reuse_idle_barrier(ref, idx);
drivers/gpu/drm/i915/i915_active.c
87
static void debug_active_init(struct i915_active *ref)
drivers/gpu/drm/i915/i915_active.c
874
node->ref = ref;
drivers/gpu/drm/i915/i915_active.c
889
__i915_active_acquire(ref);
drivers/gpu/drm/i915/i915_active.c
89
debug_object_init(ref, &active_debug_desc);
drivers/gpu/drm/i915/i915_active.c
901
GEM_BUG_ON(!llist_empty(&ref->preallocated_barriers));
drivers/gpu/drm/i915/i915_active.c
902
llist_add_batch(first, last, &ref->preallocated_barriers);
drivers/gpu/drm/i915/i915_active.c
912
atomic_dec(&ref->count);
drivers/gpu/drm/i915/i915_active.c
92
static void debug_active_activate(struct i915_active *ref)
drivers/gpu/drm/i915/i915_active.c
920
void i915_active_acquire_barrier(struct i915_active *ref)
drivers/gpu/drm/i915/i915_active.c
925
GEM_BUG_ON(i915_active_is_idle(ref));
drivers/gpu/drm/i915/i915_active.c
933
llist_for_each_safe(pos, next, take_preallocated_barriers(ref)) {
drivers/gpu/drm/i915/i915_active.c
938
spin_lock_irqsave_nested(&ref->tree_lock, flags,
drivers/gpu/drm/i915/i915_active.c
94
lockdep_assert_held(&ref->tree_lock);
drivers/gpu/drm/i915/i915_active.c
941
p = &ref->tree.rb_node;
drivers/gpu/drm/i915/i915_active.c
95
debug_object_activate(ref, &active_debug_desc);
drivers/gpu/drm/i915/i915_active.c
954
rb_insert_color(&node->node, &ref->tree);
drivers/gpu/drm/i915/i915_active.c
955
spin_unlock_irqrestore(&ref->tree_lock, flags);
drivers/gpu/drm/i915/i915_active.c
98
static void debug_active_deactivate(struct i915_active *ref)
drivers/gpu/drm/i915/i915_active.h
152
void __i915_active_init(struct i915_active *ref,
drivers/gpu/drm/i915/i915_active.h
153
int (*active)(struct i915_active *ref),
drivers/gpu/drm/i915/i915_active.h
154
void (*retire)(struct i915_active *ref),
drivers/gpu/drm/i915/i915_active.h
160
#define i915_active_init(ref, active, retire, flags) do { \
drivers/gpu/drm/i915/i915_active.h
164
__i915_active_init(ref, active, retire, flags, &__mkey, &__wkey); \
drivers/gpu/drm/i915/i915_active.h
167
int i915_active_add_request(struct i915_active *ref, struct i915_request *rq);
drivers/gpu/drm/i915/i915_active.h
170
i915_active_set_exclusive(struct i915_active *ref, struct dma_fence *f);
drivers/gpu/drm/i915/i915_active.h
172
int __i915_active_wait(struct i915_active *ref, int state);
drivers/gpu/drm/i915/i915_active.h
173
static inline int i915_active_wait(struct i915_active *ref)
drivers/gpu/drm/i915/i915_active.h
175
return __i915_active_wait(ref, TASK_INTERRUPTIBLE);
drivers/gpu/drm/i915/i915_active.h
179
struct i915_active *ref,
drivers/gpu/drm/i915/i915_active.h
182
struct i915_active *ref,
drivers/gpu/drm/i915/i915_active.h
188
int i915_active_acquire(struct i915_active *ref);
drivers/gpu/drm/i915/i915_active.h
189
bool i915_active_acquire_if_busy(struct i915_active *ref);
drivers/gpu/drm/i915/i915_active.h
191
void i915_active_release(struct i915_active *ref);
drivers/gpu/drm/i915/i915_active.h
193
static inline void __i915_active_acquire(struct i915_active *ref)
drivers/gpu/drm/i915/i915_active.h
195
GEM_BUG_ON(!atomic_read(&ref->count));
drivers/gpu/drm/i915/i915_active.h
196
atomic_inc(&ref->count);
drivers/gpu/drm/i915/i915_active.h
200
i915_active_is_idle(const struct i915_active *ref)
drivers/gpu/drm/i915/i915_active.h
202
return !atomic_read(&ref->count);
drivers/gpu/drm/i915/i915_active.h
205
void i915_active_fini(struct i915_active *ref);
drivers/gpu/drm/i915/i915_active.h
207
int i915_active_acquire_preallocate_barrier(struct i915_active *ref,
drivers/gpu/drm/i915/i915_active.h
209
void i915_active_acquire_barrier(struct i915_active *ref);
drivers/gpu/drm/i915/i915_active.h
212
void i915_active_print(struct i915_active *ref, struct drm_printer *m);
drivers/gpu/drm/i915/i915_active.h
213
void i915_active_unlock_wait(struct i915_active *ref);
drivers/gpu/drm/i915/i915_active.h
216
struct i915_active *i915_active_get(struct i915_active *ref);
drivers/gpu/drm/i915/i915_active.h
217
void i915_active_put(struct i915_active *ref);
drivers/gpu/drm/i915/i915_active_types.h
39
int (*active)(struct i915_active *ref);
drivers/gpu/drm/i915/i915_active_types.h
40
void (*retire)(struct i915_active *ref);
drivers/gpu/drm/i915/i915_gpu_error.c
1123
container_of(error_ref, typeof(*error), ref);
drivers/gpu/drm/i915/i915_gpu_error.c
1459
if (ctx && !kref_get_unless_zero(&ctx->ref))
drivers/gpu/drm/i915/i915_gpu_error.c
2095
kref_init(&error->ref);
drivers/gpu/drm/i915/i915_gpu_error.h
191
struct kref ref;
drivers/gpu/drm/i915/i915_gpu_error.h
312
kref_get(&gpu->ref);
drivers/gpu/drm/i915/i915_gpu_error.h
324
kref_put(&gpu->ref, __i915_gpu_coredump_free);
drivers/gpu/drm/i915/i915_perf.c
2717
if (!kref_get_unless_zero(&ctx->ref))
drivers/gpu/drm/i915/i915_perf.c
398
void i915_oa_config_release(struct kref *ref)
drivers/gpu/drm/i915/i915_perf.c
401
container_of(ref, typeof(*oa_config), ref);
drivers/gpu/drm/i915/i915_perf.c
4625
kref_init(&oa_config->ref);
drivers/gpu/drm/i915/i915_perf.h
45
if (kref_get_unless_zero(&oa_config->ref))
drivers/gpu/drm/i915/i915_perf.h
51
void i915_oa_config_release(struct kref *ref);
drivers/gpu/drm/i915/i915_perf.h
57
kref_put(&oa_config->ref, i915_oa_config_release);
drivers/gpu/drm/i915/i915_perf_types.h
94
struct kref ref;
drivers/gpu/drm/i915/i915_scatterlist.c
43
static void i915_refct_sgt_release(struct kref *ref)
drivers/gpu/drm/i915/i915_scatterlist.c
46
container_of(ref, typeof(*rsgt), kref);
drivers/gpu/drm/i915/i915_scatterlist.h
172
void (*release)(struct kref *ref);
drivers/gpu/drm/i915/i915_scheduler.c
440
container_of(kref, typeof(*sched_engine), ref);
drivers/gpu/drm/i915/i915_scheduler.c
460
kref_init(&sched_engine->ref);
drivers/gpu/drm/i915/i915_scheduler.h
58
kref_get(&sched_engine->ref);
drivers/gpu/drm/i915/i915_scheduler.h
65
kref_put(&sched_engine->ref, sched_engine->destroy);
drivers/gpu/drm/i915/i915_scheduler_types.h
109
struct kref ref;
drivers/gpu/drm/i915/i915_vma.c
102
static inline struct i915_vma *active_to_vma(struct i915_active *ref)
drivers/gpu/drm/i915/i915_vma.c
104
return container_of(ref, typeof(struct i915_vma), active);
drivers/gpu/drm/i915/i915_vma.c
107
static int __i915_vma_active(struct i915_active *ref)
drivers/gpu/drm/i915/i915_vma.c
109
struct i915_vma *vma = active_to_vma(ref);
drivers/gpu/drm/i915/i915_vma.c
132
static void __i915_vma_retire(struct i915_active *ref)
drivers/gpu/drm/i915/i915_vma.c
134
struct i915_vma *vma = active_to_vma(ref);
drivers/gpu/drm/i915/i915_vma.c
2088
kref_read(&vma->vm->ref);
drivers/gpu/drm/i915/i915_vma.c
2089
vma_res->skip_pte_rewrite = !kref_read(&vma->vm->ref) ||
drivers/gpu/drm/i915/i915_vma.c
332
GEM_BUG_ON(!kref_read(&vm->ref));
drivers/gpu/drm/i915/i915_vma.c
57
if (kref_read(&vma->vm->ref))
drivers/gpu/drm/i915/pxp/intel_pxp.c
505
if (!kref_get_unless_zero(&ctx->ref))
drivers/gpu/drm/i915/selftests/i915_active.c
22
struct kref ref;
drivers/gpu/drm/i915/selftests/i915_active.c
28
kref_get(&active->ref);
drivers/gpu/drm/i915/selftests/i915_active.c
281
void i915_active_print(struct i915_active *ref, struct drm_printer *m)
drivers/gpu/drm/i915/selftests/i915_active.c
283
drm_printf(m, "active %ps:%ps\n", ref->active, ref->retire);
drivers/gpu/drm/i915/selftests/i915_active.c
284
drm_printf(m, "\tcount: %d\n", atomic_read(&ref->count));
drivers/gpu/drm/i915/selftests/i915_active.c
286
str_yes_no(!llist_empty(&ref->preallocated_barriers)));
drivers/gpu/drm/i915/selftests/i915_active.c
288
if (i915_active_acquire_if_busy(ref)) {
drivers/gpu/drm/i915/selftests/i915_active.c
291
rbtree_postorder_for_each_entry_safe(it, n, &ref->tree, node) {
drivers/gpu/drm/i915/selftests/i915_active.c
307
i915_active_release(ref);
drivers/gpu/drm/i915/selftests/i915_active.c
317
static void active_flush(struct i915_active *ref,
drivers/gpu/drm/i915/selftests/i915_active.c
329
atomic_dec(&ref->count);
drivers/gpu/drm/i915/selftests/i915_active.c
334
void i915_active_unlock_wait(struct i915_active *ref)
drivers/gpu/drm/i915/selftests/i915_active.c
336
if (i915_active_acquire_if_busy(ref)) {
drivers/gpu/drm/i915/selftests/i915_active.c
341
active_flush(ref, &ref->excl);
drivers/gpu/drm/i915/selftests/i915_active.c
342
rbtree_postorder_for_each_entry_safe(it, n, &ref->tree, node)
drivers/gpu/drm/i915/selftests/i915_active.c
343
active_flush(ref, &it->base);
drivers/gpu/drm/i915/selftests/i915_active.c
346
i915_active_release(ref);
drivers/gpu/drm/i915/selftests/i915_active.c
350
spin_unlock_wait(&ref->tree_lock);
drivers/gpu/drm/i915/selftests/i915_active.c
353
flush_work(&ref->work);
drivers/gpu/drm/i915/selftests/i915_active.c
37
static void __live_release(struct kref *ref)
drivers/gpu/drm/i915/selftests/i915_active.c
39
struct live_active *active = container_of(ref, typeof(*active), ref);
drivers/gpu/drm/i915/selftests/i915_active.c
46
kref_put(&active->ref, __live_release);
drivers/gpu/drm/i915/selftests/i915_active.c
73
kref_init(&active->ref);
drivers/gpu/drm/i915/selftests/i915_perf.c
29
kref_init(&oa_config->ref);
drivers/gpu/drm/i915/selftests/lib_sw_fence.c
117
refcount_set(&h->ref.refcount, 2);
drivers/gpu/drm/i915/selftests/lib_sw_fence.c
122
static void heap_fence_release(struct kref *ref)
drivers/gpu/drm/i915/selftests/lib_sw_fence.c
124
struct heap_fence *h = container_of(ref, typeof(*h), ref);
drivers/gpu/drm/i915/selftests/lib_sw_fence.c
135
kref_put(&h->ref, heap_fence_release);
drivers/gpu/drm/i915/selftests/lib_sw_fence.c
87
struct kref ref;
drivers/gpu/drm/lima/lima_ctx.c
43
static void lima_ctx_do_release(struct kref *ref)
drivers/gpu/drm/lima/lima_ctx.c
45
struct lima_ctx *ctx = container_of(ref, struct lima_ctx, refcnt);
drivers/gpu/drm/msm/adreno/a5xx_gpu.c
1603
struct msm_gpu_state, ref);
drivers/gpu/drm/msm/adreno/a5xx_gpu.c
1618
return kref_put(&state->ref, a5xx_gpu_state_destroy);
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
1671
struct msm_gpu_state, ref);
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
1698
return kref_put(&state->ref, a6xx_gpu_state_destroy);
drivers/gpu/drm/msm/adreno/adreno_gpu.c
759
kref_init(&state->ref);
drivers/gpu/drm/msm/adreno/adreno_gpu.c
837
struct msm_gpu_state, ref);
drivers/gpu/drm/msm/adreno/adreno_gpu.c
848
return kref_put(&state->ref, adreno_gpu_state_kref_destroy);
drivers/gpu/drm/msm/dsi/phy/dsi_phy_14nm.c
170
u32 ref, rem;
drivers/gpu/drm/msm/dsi/phy/dsi_phy_14nm.c
185
ref = VCO_REF_CLK_RATE;
drivers/gpu/drm/msm/dsi/phy/dsi_phy_14nm.c
186
ref /= 1000;
drivers/gpu/drm/msm/dsi/phy/dsi_phy_14nm.c
187
step_size = div_u64(step_size, ref);
drivers/gpu/drm/msm/msm_drv.c
257
kref_init(&ctx->ref);
drivers/gpu/drm/msm/msm_gem.h
431
struct kref ref;
drivers/gpu/drm/msm/msm_gem.h
488
kref_get(&submit->ref);
drivers/gpu/drm/msm/msm_gem.h
493
kref_put(&submit->ref, __msm_gem_submit_destroy);
drivers/gpu/drm/msm/msm_gem_submit.c
67
kref_init(&submit->ref);
drivers/gpu/drm/msm/msm_gem_submit.c
88
container_of(kref, struct msm_gem_submit, ref);
drivers/gpu/drm/msm/msm_gpu.h
376
struct kref ref;
drivers/gpu/drm/msm/msm_gpu.h
547
struct kref ref;
drivers/gpu/drm/msm/msm_gpu.h
564
struct kref ref;
drivers/gpu/drm/msm/msm_gpu.h
672
kref_put(&ctx->ref, __msm_context_destroy);
drivers/gpu/drm/msm/msm_gpu.h
678
kref_get(&ctx->ref);
drivers/gpu/drm/msm/msm_gpu.h
718
kref_put(&queue->ref, msm_submitqueue_destroy);
drivers/gpu/drm/msm/msm_gpu.h
728
kref_get(&gpu->crashstate->ref);
drivers/gpu/drm/msm/msm_submitqueue.c
216
kref_init(&queue->ref);
drivers/gpu/drm/msm/msm_submitqueue.c
55
struct msm_context, ref);
drivers/gpu/drm/msm/msm_submitqueue.c
75
struct msm_gpu_submitqueue, ref);
drivers/gpu/drm/msm/msm_submitqueue.c
99
kref_get(&entry->ref);
drivers/gpu/drm/nouveau/include/nvif/if0004.h
15
__u32 ref;
drivers/gpu/drm/nouveau/include/nvif/if000c.h
82
__u8 ref;
drivers/gpu/drm/nouveau/include/nvif/vmm.h
56
int nvif_vmm_raw_sparse(struct nvif_vmm *vmm, u64 addr, u64 size, bool ref);
drivers/gpu/drm/nouveau/include/nvkm/subdev/mmu.h
35
struct mutex ref;
drivers/gpu/drm/nouveau/nv04_fence.c
64
return args.ref;
drivers/gpu/drm/nouveau/nvif/vmm.c
177
nvif_vmm_raw_sparse(struct nvif_vmm *vmm, u64 addr, u64 size, bool ref)
drivers/gpu/drm/nouveau/nvif/vmm.c
184
.ref = ref,
drivers/gpu/drm/nouveau/nvkm/engine/sw/nv04.c
116
atomic_set(&chan->ref, 0);
drivers/gpu/drm/nouveau/nvkm/engine/sw/nv04.c
36
atomic_t ref;
drivers/gpu/drm/nouveau/nvkm/engine/sw/nv04.c
53
args->v0.ref = atomic_read(&chan->ref);
drivers/gpu/drm/nouveau/nvkm/engine/sw/nv04.c
94
atomic_set(&chan->ref, data);
drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c
210
calc_div(struct gf100_clk *clk, int idx, u32 ref, u32 freq, u32 *ddiv)
drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c
212
u32 div = min((ref * 2) / freq, (u32)65);
drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c
217
return (ref * 2) / div;
drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c
223
calc_div(struct gk104_clk *clk, int idx, u32 ref, u32 freq, u32 *ddiv)
drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c
225
u32 div = min((ref * 2) / freq, (u32)65);
drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c
230
return (ref * 2) / div;
drivers/gpu/drm/nouveau/nvkm/subdev/clk/mcp77.c
54
u32 ref = nvkm_clk_read(&clk->base, nv_clk_src_href);
drivers/gpu/drm/nouveau/nvkm/subdev/clk/mcp77.c
73
clock = ref * N1 / M1;
drivers/gpu/drm/nouveau/nvkm/subdev/clk/nv40.c
47
u32 ref = 27000, khz = 0;
drivers/gpu/drm/nouveau/nvkm/subdev/clk/nv40.c
50
khz = ref * N / M;
drivers/gpu/drm/nouveau/nvkm/subdev/clk/nv40.c
66
u32 ref = 27000, khz = 0;
drivers/gpu/drm/nouveau/nvkm/subdev/clk/nv40.c
69
khz = ref * N1 / M1;
drivers/gpu/drm/nouveau/nvkm/subdev/clk/nv50.c
119
return (ref * N / M) >> P;
drivers/gpu/drm/nouveau/nvkm/subdev/clk/nv50.c
164
u32 ref = read_pll_ref(clk, base);
drivers/gpu/drm/nouveau/nvkm/subdev/clk/nv50.c
179
freq = ref * N1 / M1;
drivers/gpu/drm/nouveau/nvkm/subdev/clk/nv50.c
56
u32 coef, ref = nvkm_clk_read(&clk->base, nv_clk_src_crystal);
drivers/gpu/drm/nouveau/nvkm/subdev/clk/nv50.c
74
ref *= (coef & 0x01000000) ? 2 : 4;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgf100.c
141
int ref, div, out;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgf100.c
189
ref = nvkm_clk_read(clk, nv_clk_src_sppll0);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgf100.c
191
ref = nvkm_clk_read(clk, nv_clk_src_sppll1);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgf100.c
192
div = max(min((ref * 2) / freq, (u32)65), (u32)2) - 2;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgf100.c
193
out = (ref * 2) / (div + 2);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
458
return nvkm_vmm_raw_sparse(vmm, args->addr, args->size, args->ref);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1128
mutex_init(&vmm->mutex.ref);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1901
nvkm_vmm_raw_sparse(struct nvkm_vmm *vmm, u64 addr, u64 size, bool ref)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1905
mutex_lock(&vmm->mutex.ref);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1906
ret = nvkm_vmm_ptes_sparse(vmm, addr, size, ref);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1907
mutex_unlock(&vmm->mutex.ref);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
526
u64 addr, u64 size, const char *name, bool ref, bool pfn,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
568
if (ref && NVKM_VMM_PDE_INVALID(pgd->pde[pdei])) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
580
if (ref && !pgt->refs[desc[it.lvl - 1].type == SPT]) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
654
nvkm_vmm_ptes_sparse(struct nvkm_vmm *vmm, u64 addr, u64 size, bool ref)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
685
if (ref) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
740
mutex_lock(&vmm->mutex.ref);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
742
mutex_unlock(&vmm->mutex.ref);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
751
mutex_lock(&vmm->mutex.ref);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
757
mutex_unlock(&vmm->mutex.ref);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
760
mutex_unlock(&vmm->mutex.ref);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h
191
int nvkm_vmm_raw_sparse(struct nvkm_vmm *, u64 addr, u64 size, bool ref);
drivers/gpu/drm/panfrost/panfrost_job.c
351
static void panfrost_job_cleanup(struct kref *ref)
drivers/gpu/drm/panfrost/panfrost_job.c
353
struct panfrost_job *job = container_of(ref, struct panfrost_job,
drivers/gpu/drm/panthor/panthor_sched.c
1305
queue->syncwait.ref = cs_iface->output->status_wait_sync_value;
drivers/gpu/drm/panthor/panthor_sched.c
1312
queue->syncwait.ref |= sync_val_hi << 32;
drivers/gpu/drm/panthor/panthor_sched.c
2596
result = value > queue->syncwait.ref;
drivers/gpu/drm/panthor/panthor_sched.c
2598
result = value <= queue->syncwait.ref;
drivers/gpu/drm/panthor/panthor_sched.c
3910
static void job_release(struct kref *ref)
drivers/gpu/drm/panthor/panthor_sched.c
3912
struct panthor_job *job = container_of(ref, struct panthor_job, refcount);
drivers/gpu/drm/panthor/panthor_sched.c
423
u64 ref;
drivers/gpu/drm/radeon/rv740_dpm.c
34
u32 ref = 0;
drivers/gpu/drm/radeon/rv740_dpm.c
38
ref = 1;
drivers/gpu/drm/radeon/rv740_dpm.c
41
ref = 2;
drivers/gpu/drm/radeon/rv740_dpm.c
44
ref = 3;
drivers/gpu/drm/radeon/rv740_dpm.c
47
ref = 2;
drivers/gpu/drm/radeon/rv740_dpm.c
50
ref = 3;
drivers/gpu/drm/radeon/rv740_dpm.c
53
ref = 4;
drivers/gpu/drm/radeon/rv740_dpm.c
56
ref = 5;
drivers/gpu/drm/radeon/rv740_dpm.c
60
ref = 0;
drivers/gpu/drm/radeon/rv740_dpm.c
64
return ref;
drivers/gpu/drm/tegra/gem.c
71
kref_init(&map->ref);
drivers/gpu/drm/tegra/submit.c
154
kref_get(&mapping->ref);
drivers/gpu/drm/tegra/submit.c
203
kref_init(&bo->ref);
drivers/gpu/drm/tegra/submit.c
34
struct kref ref;
drivers/gpu/drm/tegra/submit.c
46
kref_get(&bo->ref);
drivers/gpu/drm/tegra/submit.c
51
static void gather_bo_release(struct kref *ref)
drivers/gpu/drm/tegra/submit.c
53
struct gather_bo *bo = container_of(ref, struct gather_bo, ref);
drivers/gpu/drm/tegra/submit.c
64
kref_put(&bo->ref, gather_bo_release);
drivers/gpu/drm/tegra/submit.c
78
kref_init(&map->ref);
drivers/gpu/drm/tegra/uapi.c
15
static void tegra_drm_mapping_release(struct kref *ref)
drivers/gpu/drm/tegra/uapi.c
18
container_of(ref, struct tegra_drm_mapping, ref);
drivers/gpu/drm/tegra/uapi.c
215
kref_init(&mapping->ref);
drivers/gpu/drm/tegra/uapi.c
28
kref_put(&mapping->ref, tegra_drm_mapping_release);
drivers/gpu/drm/tegra/uapi.h
28
struct kref ref;
drivers/gpu/drm/v3d/v3d_drv.h
326
void (*free)(struct kref *ref);
drivers/gpu/drm/v3d/v3d_submit.c
110
v3d_render_job_free(struct kref *ref)
drivers/gpu/drm/v3d/v3d_submit.c
112
struct v3d_render_job *job = container_of(ref, struct v3d_render_job,
drivers/gpu/drm/v3d/v3d_submit.c
120
v3d_job_free(ref);
drivers/gpu/drm/v3d/v3d_submit.c
161
struct v3d_job *job, void (*free)(struct kref *ref),
drivers/gpu/drm/v3d/v3d_submit.c
89
v3d_job_free(struct kref *ref)
drivers/gpu/drm/v3d/v3d_submit.c
91
struct v3d_job *job = container_of(ref, struct v3d_job, refcount);
drivers/gpu/drm/vc4/vc4_v3d.c
352
static void bin_bo_release(struct kref *ref)
drivers/gpu/drm/vc4/vc4_v3d.c
354
struct vc4_dev *vc4 = container_of(ref, struct vc4_dev, bin_bo_kref);
drivers/gpu/drm/vkms/tests/vkms_color_test.c
162
struct pixel_argb_s32 ref, out;
drivers/gpu/drm/vkms/tests/vkms_color_test.c
165
ref.a = 0xffff;
drivers/gpu/drm/vkms/tests/vkms_color_test.c
166
ref.r = 0xffff;
drivers/gpu/drm/vkms/tests/vkms_color_test.c
167
ref.g = 0xffff;
drivers/gpu/drm/vkms/tests/vkms_color_test.c
168
ref.b = 0xffff;
drivers/gpu/drm/vkms/tests/vkms_color_test.c
170
memcpy(&out, &ref, sizeof(out));
drivers/gpu/drm/vkms/tests/vkms_color_test.c
173
KUNIT_EXPECT_MEMEQ(test, &ref, &out, sizeof(out));
drivers/gpu/drm/vkms/tests/vkms_color_test.c
176
ref.a = 0xffff;
drivers/gpu/drm/vkms/tests/vkms_color_test.c
177
ref.r = 0x0;
drivers/gpu/drm/vkms/tests/vkms_color_test.c
178
ref.g = 0x0;
drivers/gpu/drm/vkms/tests/vkms_color_test.c
179
ref.b = 0x0;
drivers/gpu/drm/vkms/tests/vkms_color_test.c
181
memcpy(&out, &ref, sizeof(out));
drivers/gpu/drm/vkms/tests/vkms_color_test.c
184
KUNIT_EXPECT_MEMEQ(test, &ref, &out, sizeof(out));
drivers/gpu/drm/vkms/tests/vkms_color_test.c
187
ref.a = 0xffff;
drivers/gpu/drm/vkms/tests/vkms_color_test.c
188
ref.r = 0x8000;
drivers/gpu/drm/vkms/tests/vkms_color_test.c
189
ref.g = 0x8000;
drivers/gpu/drm/vkms/tests/vkms_color_test.c
190
ref.b = 0x8000;
drivers/gpu/drm/vkms/tests/vkms_color_test.c
192
memcpy(&out, &ref, sizeof(out));
drivers/gpu/drm/vkms/tests/vkms_color_test.c
195
KUNIT_EXPECT_MEMEQ(test, &ref, &out, sizeof(out));
drivers/gpu/drm/vkms/tests/vkms_color_test.c
198
ref.a = 0xffff;
drivers/gpu/drm/vkms/tests/vkms_color_test.c
199
ref.r = 0x8000;
drivers/gpu/drm/vkms/tests/vkms_color_test.c
200
ref.g = 0x4000;
drivers/gpu/drm/vkms/tests/vkms_color_test.c
201
ref.b = 0x4000;
drivers/gpu/drm/vkms/tests/vkms_color_test.c
210
KUNIT_EXPECT_MEMEQ(test, &ref, &out, sizeof(out));
drivers/gpu/drm/vmwgfx/ttm_object.c
295
struct ttm_ref_object *ref;
drivers/gpu/drm/vmwgfx/ttm_object.c
310
ref = hlist_entry(hash, struct ttm_ref_object, hash);
drivers/gpu/drm/vmwgfx/ttm_object.c
311
if (kref_get_unless_zero(&ref->kref)) {
drivers/gpu/drm/vmwgfx/ttm_object.c
321
ref = kmalloc_obj(*ref);
drivers/gpu/drm/vmwgfx/ttm_object.c
322
if (unlikely(ref == NULL)) {
drivers/gpu/drm/vmwgfx/ttm_object.c
326
ref->hash.key = base->handle;
drivers/gpu/drm/vmwgfx/ttm_object.c
327
ref->obj = base;
drivers/gpu/drm/vmwgfx/ttm_object.c
328
ref->tfile = tfile;
drivers/gpu/drm/vmwgfx/ttm_object.c
329
kref_init(&ref->kref);
drivers/gpu/drm/vmwgfx/ttm_object.c
332
hash_add_rcu(tfile->ref_hash, &ref->hash.head, ref->hash.key);
drivers/gpu/drm/vmwgfx/ttm_object.c
335
list_add_tail(&ref->head, &tfile->ref_list);
drivers/gpu/drm/vmwgfx/ttm_object.c
348
struct ttm_ref_object *ref =
drivers/gpu/drm/vmwgfx/ttm_object.c
350
struct ttm_object_file *tfile = ref->tfile;
drivers/gpu/drm/vmwgfx/ttm_object.c
352
hash_del_rcu(&ref->hash.head);
drivers/gpu/drm/vmwgfx/ttm_object.c
353
list_del(&ref->head);
drivers/gpu/drm/vmwgfx/ttm_object.c
356
ttm_base_object_unref(&ref->obj);
drivers/gpu/drm/vmwgfx/ttm_object.c
357
kfree_rcu(ref, rcu_head);
drivers/gpu/drm/vmwgfx/ttm_object.c
364
struct ttm_ref_object *ref;
drivers/gpu/drm/vmwgfx/ttm_object.c
374
ref = hlist_entry(hash, struct ttm_ref_object, hash);
drivers/gpu/drm/vmwgfx/ttm_object.c
375
kref_put(&ref->kref, ttm_ref_object_release);
drivers/gpu/drm/vmwgfx/ttm_object.c
382
struct ttm_ref_object *ref;
drivers/gpu/drm/vmwgfx/ttm_object.c
396
ref = list_entry(list, struct ttm_ref_object, head);
drivers/gpu/drm/vmwgfx/ttm_object.c
397
ttm_ref_object_release(&ref->kref);
drivers/gpu/drm/xe/compat-i915-headers/i915_vma.h
24
refcount_t ref;
drivers/gpu/drm/xe/display/intel_bo.c
53
struct kref ref;
drivers/gpu/drm/xe/display/intel_bo.c
66
kref_init(&front->ref);
drivers/gpu/drm/xe/display/intel_bo.c
79
kref_get(&front->ref);
drivers/gpu/drm/xe/display/intel_bo.c
82
static void frontbuffer_release(struct kref *ref)
drivers/gpu/drm/xe/display/intel_bo.c
85
container_of(ref, typeof(*front), ref);
drivers/gpu/drm/xe/display/intel_bo.c
99
kref_put(&front->ref, frontbuffer_release);
drivers/gpu/drm/xe/display/xe_fb_pin.c
286
refcount_set(&vma->ref, 1);
drivers/gpu/drm/xe/display/xe_fb_pin.c
350
if (!refcount_dec_and_test(&vma->ref))
drivers/gpu/drm/xe/display/xe_fb_pin.c
410
refcount_inc(&vma->ref);
drivers/gpu/drm/xe/xe_device.c
122
static void xe_file_destroy(struct kref *ref)
drivers/gpu/drm/xe/xe_device.c
124
struct xe_file *xef = container_of(ref, struct xe_file, refcount);
drivers/gpu/drm/xe/xe_exec_queue.c
471
void xe_exec_queue_destroy(struct kref *ref)
drivers/gpu/drm/xe/xe_exec_queue.c
473
struct xe_exec_queue *q = container_of(ref, struct xe_exec_queue, refcount);
drivers/gpu/drm/xe/xe_exec_queue.h
35
void xe_exec_queue_destroy(struct kref *ref);
drivers/gpu/drm/xe/xe_force_wake.c
195
if (!domain->ref++) {
drivers/gpu/drm/xe/xe_force_wake.c
206
--domain->ref;
drivers/gpu/drm/xe/xe_force_wake.c
251
xe_gt_assert(gt, domain->ref);
drivers/gpu/drm/xe/xe_force_wake.c
253
if (!--domain->ref) {
drivers/gpu/drm/xe/xe_force_wake.h
27
return fw->domains[ffs(domain) - 1].ref;
drivers/gpu/drm/xe/xe_force_wake.h
88
#define __xe_with_force_wake(ref, fw, domains, done) \
drivers/gpu/drm/xe/xe_force_wake.h
89
for (CLASS(xe_force_wake, ref)(fw, domains), *(done) = NULL; \
drivers/gpu/drm/xe/xe_force_wake.h
92
#define xe_with_force_wake(ref, fw, domains) \
drivers/gpu/drm/xe/xe_force_wake.h
93
__xe_with_force_wake(ref, fw, domains, __UNIQUE_ID(done))
drivers/gpu/drm/xe/xe_force_wake_types.h
84
u32 ref;
drivers/gpu/drm/xe/xe_guc_capture.c
1839
snapshot->forcewake.domain, snapshot->forcewake.ref);
drivers/gpu/drm/xe/xe_hw_engine.c
939
snapshot->forcewake.ref = xe_force_wake_ref(gt_to_fw(hwe->gt),
drivers/gpu/drm/xe/xe_hw_engine_types.h
177
int ref;
drivers/gpu/drm/xe/xe_lrc.c
1639
void xe_lrc_destroy(struct kref *ref)
drivers/gpu/drm/xe/xe_lrc.c
1641
struct xe_lrc *lrc = container_of(ref, struct xe_lrc, refcount);
drivers/gpu/drm/xe/xe_lrc.h
55
void xe_lrc_destroy(struct kref *ref);
drivers/gpu/drm/xe/xe_oa.c
159
static void xe_oa_config_release(struct kref *ref)
drivers/gpu/drm/xe/xe_oa.c
162
container_of(ref, typeof(*oa_config), ref);
drivers/gpu/drm/xe/xe_oa.c
174
kref_put(&oa_config->ref, xe_oa_config_release);
drivers/gpu/drm/xe/xe_oa.c
179
return kref_get_unless_zero(&oa_config->ref) ? oa_config : NULL;
drivers/gpu/drm/xe/xe_oa.c
2372
kref_init(&oa_config->ref);
drivers/gpu/drm/xe/xe_oa.c
81
struct kref ref;
drivers/gpu/drm/xe/xe_pm.c
879
bool ref;
drivers/gpu/drm/xe/xe_pm.c
881
ref = xe_pm_runtime_get_if_in_use(xe);
drivers/gpu/drm/xe/xe_pm.c
883
if (!ref) {
drivers/gpu/drm/xe/xe_sched_job.c
171
void xe_sched_job_destroy(struct kref *ref)
drivers/gpu/drm/xe/xe_sched_job.c
174
container_of(ref, struct xe_sched_job, refcount);
drivers/gpu/drm/xe/xe_sched_job.h
23
void xe_sched_job_destroy(struct kref *ref);
drivers/gpu/drm/xe/xe_tile.c
240
xe_assert(tile_to_xe(tile), kref_read(&dpagemap->ref) >= 2);
drivers/gpu/drm/xe/xe_tlb_inval_job.c
174
static void xe_tlb_inval_job_destroy(struct kref *ref)
drivers/gpu/drm/xe/xe_tlb_inval_job.c
176
struct xe_tlb_inval_job *job = container_of(ref, typeof(*job),
drivers/gpu/host1x/bus.c
901
kref_get(&mapping->ref);
drivers/gpu/host1x/bus.c
922
kref_get(&mapping->ref);
drivers/gpu/host1x/bus.c
933
static void __host1x_bo_unpin(struct kref *ref)
drivers/gpu/host1x/bus.c
935
struct host1x_bo_mapping *mapping = to_host1x_bo_mapping(ref);
drivers/gpu/host1x/bus.c
958
kref_put(&mapping->ref, __host1x_bo_unpin);
drivers/gpu/host1x/context.c
139
refcount_inc(&cd->ref);
drivers/gpu/host1x/context.c
152
refcount_set(&free->ref, 1);
drivers/gpu/host1x/context.c
163
refcount_inc(&cd->ref);
drivers/gpu/host1x/context.c
171
if (refcount_dec_and_mutex_lock(&cd->ref, &cdl->lock)) {
drivers/gpu/host1x/debug.c
101
if (!kref_read(&m->syncpt[i].ref))
drivers/gpu/host1x/job.c
109
kref_put(&job->ref, job_free);
drivers/gpu/host1x/job.c
57
kref_init(&job->ref);
drivers/gpu/host1x/job.c
79
kref_get(&job->ref);
drivers/gpu/host1x/job.c
84
static void job_free(struct kref *ref)
drivers/gpu/host1x/job.c
86
struct host1x_job *job = container_of(ref, struct host1x_job, ref);
drivers/gpu/host1x/syncpt.c
314
kref_init(&host->syncpt[26].ref);
drivers/gpu/host1x/syncpt.c
315
kref_init(&host->syncpt[27].ref);
drivers/gpu/host1x/syncpt.c
340
static void syncpt_release(struct kref *ref)
drivers/gpu/host1x/syncpt.c
342
struct host1x_syncpt *sp = container_of(ref, struct host1x_syncpt, ref);
drivers/gpu/host1x/syncpt.c
370
kref_put_mutex(&sp->ref, syncpt_release, &sp->host->syncpt_mutex);
drivers/gpu/host1x/syncpt.c
449
if (kref_get_unless_zero(&host->syncpt[id].ref))
drivers/gpu/host1x/syncpt.c
478
kref_get(&sp->ref);
drivers/gpu/host1x/syncpt.c
504
static void do_nothing(struct kref *ref)
drivers/gpu/host1x/syncpt.c
528
kref_put(&host->syncpt[syncpt_id].ref, do_nothing);
drivers/gpu/host1x/syncpt.c
71
for (i = 0; i < host->info->nb_pts && kref_read(&sp->ref); i++, sp++)
drivers/gpu/host1x/syncpt.c
94
kref_init(&sp->ref);
drivers/gpu/host1x/syncpt.h
31
struct kref ref;
drivers/hid/hid-core.c
2992
kref_init(&hdev->ref);
drivers/hid/hid-core.c
756
void hiddev_free(struct kref *ref)
drivers/hid/hid-core.c
758
struct hid_device *hid = container_of(ref, struct hid_device, ref);
drivers/hid/hid-core.c
770
kref_put(&hid->ref, hiddev_free);
drivers/hid/hid-debug.c
3701
kref_get(&list->hdev->ref);
drivers/hid/hid-debug.c
3795
kref_put(&list->hdev->ref, hiddev_free);
drivers/hv/mshv_regions.c
350
static void mshv_region_destroy(struct kref *ref)
drivers/hv/mshv_regions.c
353
container_of(ref, struct mshv_mem_region, mreg_refcount);
drivers/hwmon/fschmd.c
310
static void fschmd_release_resources(struct kref *ref)
drivers/hwmon/fschmd.c
312
struct fschmd_data *data = container_of(ref, struct fschmd_data, kref);
drivers/hwmon/pc87360.c
161
#define IN_FROM_REG(val, ref) (((val) * (ref) + 128) / 256)
drivers/hwmon/pc87360.c
162
#define IN_TO_REG(val, ref) ((val) < 0 ? 0 : \
drivers/hwmon/pc87360.c
163
(val) * 256 >= (ref) * 255 ? 255 : \
drivers/hwmon/pc87360.c
164
((val) * 256 + (ref) / 2) / (ref))
drivers/hwmon/w83793.c
277
static void w83793_release_resources(struct kref *ref)
drivers/hwmon/w83793.c
279
struct w83793_data *data = container_of(ref, struct w83793_data, kref);
drivers/iio/adc/ad7768-1.c
1174
struct fwnode_handle *ref __free(fwnode_handle) =
drivers/iio/adc/ad7768-1.c
1179
if (IS_ERR(ref))
drivers/iio/adc/ad7768-1.c
1180
return PTR_ERR(ref);
drivers/iio/adc/ad7768-1.c
1182
ref = args.fwnode;
drivers/iio/adc/ad7768-1.c
1184
if (fwnode_device_is_compatible(ref, "gpio-trigger")) {
drivers/iio/adc/ad7768-1.c
1185
st->gpio_sync_in = devm_fwnode_gpiod_get_index(dev, ref, NULL, 0,
drivers/iio/adc/ltc2497-core.c
192
ddata->ref = devm_regulator_get(dev, "vref");
drivers/iio/adc/ltc2497-core.c
193
if (IS_ERR(ddata->ref))
drivers/iio/adc/ltc2497-core.c
194
return dev_err_probe(dev, PTR_ERR(ddata->ref),
drivers/iio/adc/ltc2497-core.c
197
ret = regulator_enable(ddata->ref);
drivers/iio/adc/ltc2497-core.c
225
regulator_disable(ddata->ref);
drivers/iio/adc/ltc2497-core.c
239
regulator_disable(ddata->ref);
drivers/iio/adc/ltc2497-core.c
94
ret = regulator_get_voltage(ddata->ref);
drivers/iio/adc/ltc2497.h
13
struct regulator *ref;
drivers/iio/adc/mcp3564.c
1088
int ref;
drivers/iio/adc/mcp3564.c
1094
ref = adc->vref_mv;
drivers/iio/adc/mcp3564.c
1095
tmp1 = ((u64)ref * NANO) >> pow;
drivers/iio/adc/ti-adc081c.c
171
adc->ref = devm_regulator_get(&client->dev, "vref");
drivers/iio/adc/ti-adc081c.c
172
if (IS_ERR(adc->ref))
drivers/iio/adc/ti-adc081c.c
173
return PTR_ERR(adc->ref);
drivers/iio/adc/ti-adc081c.c
175
err = regulator_enable(adc->ref);
drivers/iio/adc/ti-adc081c.c
180
adc->ref);
drivers/iio/adc/ti-adc081c.c
32
struct regulator *ref;
drivers/iio/adc/ti-adc081c.c
63
err = regulator_get_voltage(adc->ref);
drivers/iio/adc/ti-adc161s626.c
147
ret = regulator_get_voltage(data->ref);
drivers/iio/adc/ti-adc161s626.c
204
data->ref = devm_regulator_get(&spi->dev, "vdda");
drivers/iio/adc/ti-adc161s626.c
205
if (IS_ERR(data->ref))
drivers/iio/adc/ti-adc161s626.c
206
return PTR_ERR(data->ref);
drivers/iio/adc/ti-adc161s626.c
208
ret = regulator_enable(data->ref);
drivers/iio/adc/ti-adc161s626.c
213
data->ref);
drivers/iio/adc/ti-adc161s626.c
70
struct regulator *ref;
drivers/iio/dac/mcp4725.c
396
u8 ref;
drivers/iio/dac/mcp4725.c
475
ref = (inbuf[3] >> 3) & 0x3;
drivers/iio/dac/mcp4725.c
477
if (!info->use_ext_ref_voltage && ref != data->ref_mode) {
drivers/iio/dac/mcp4725.c
480
data->ref_mode, ref, data->ref_mode);
drivers/iio/imu/st_lsm6dsx/st_lsm6dsx_core.c
1734
struct st_lsm6dsx_sensor *ref = iio_priv(hw->iio_devs[id]);
drivers/iio/imu/st_lsm6dsx/st_lsm6dsx_core.c
1738
return max_t(u32, ref->odr, odr);
drivers/iio/imu/st_lsm6dsx/st_lsm6dsx_core.c
1742
return (hw->enable_mask & BIT(id)) ? ref->odr : 0;
drivers/iio/industrialio-buffer.c
1561
static void iio_buffer_dmabuf_release(struct kref *ref)
drivers/iio/industrialio-buffer.c
1563
struct iio_dmabuf_priv *priv = container_of(ref, struct iio_dmabuf_priv, ref);
drivers/iio/industrialio-buffer.c
1581
kref_get(&priv->ref);
drivers/iio/industrialio-buffer.c
1588
kref_put(&priv->ref, iio_buffer_dmabuf_release);
drivers/iio/industrialio-buffer.c
1712
kref_init(&priv->ref);
drivers/iio/industrialio-buffer.c
2466
static void iio_buffer_release(struct kref *ref)
drivers/iio/industrialio-buffer.c
2468
struct iio_buffer *buffer = container_of(ref, struct iio_buffer, ref);
drivers/iio/industrialio-buffer.c
2483
kref_get(&buffer->ref);
drivers/iio/industrialio-buffer.c
2496
kref_put(&buffer->ref, iio_buffer_release);
drivers/iio/industrialio-buffer.c
374
kref_init(&buffer->ref);
drivers/iio/industrialio-buffer.c
43
struct kref ref;
drivers/iio/temperature/ltc2983.c
704
struct fwnode_handle *ref __free(fwnode_handle) =
drivers/iio/temperature/ltc2983.c
706
if (IS_ERR(ref)) {
drivers/iio/temperature/ltc2983.c
707
ref = NULL;
drivers/iio/temperature/ltc2983.c
709
ret = fwnode_property_read_u32(ref, "reg", &thermo->cold_junction_chan);
drivers/iio/temperature/ltc2983.c
750
struct fwnode_handle *ref __free(fwnode_handle) =
drivers/iio/temperature/ltc2983.c
752
if (IS_ERR(ref))
drivers/iio/temperature/ltc2983.c
753
return dev_err_cast_probe(dev, ref,
drivers/iio/temperature/ltc2983.c
756
ret = fwnode_property_read_u32(ref, "reg", &rtd->r_sense_chan);
drivers/iio/temperature/ltc2983.c
900
struct fwnode_handle *ref __free(fwnode_handle) =
drivers/iio/temperature/ltc2983.c
902
if (IS_ERR(ref))
drivers/iio/temperature/ltc2983.c
903
return dev_err_cast_probe(dev, ref,
drivers/iio/temperature/ltc2983.c
906
ret = fwnode_property_read_u32(ref, "reg", &thermistor->r_sense_chan);
drivers/infiniband/core/ib_core_uverbs.c
139
!kref_get_unless_zero(&entry->ref))
drivers/infiniband/core/ib_core_uverbs.c
186
container_of(kref, struct rdma_user_mmap_entry, ref);
drivers/infiniband/core/ib_core_uverbs.c
220
kref_put(&entry->ref, rdma_user_mmap_entry_free);
drivers/infiniband/core/ib_core_uverbs.c
259
kref_put(&entry->ref, rdma_user_mmap_entry_free);
drivers/infiniband/core/ib_core_uverbs.c
297
kref_init(&entry->ref);
drivers/infiniband/core/ib_core_uverbs.c
42
kref_get(&entry->ref);
drivers/infiniband/core/rdma_core.c
292
kref_init(&uobj->ref);
drivers/infiniband/core/rdma_core.c
327
if (!uobj || !kref_get_unless_zero(&uobj->ref))
drivers/infiniband/core/rdma_core.c
45
static void uverbs_uobject_free(struct kref *ref)
drivers/infiniband/core/rdma_core.c
47
kfree_rcu(container_of(ref, struct ib_uobject, ref), rcu);
drivers/infiniband/core/rdma_core.c
57
kref_put(&uobject->ref, uverbs_uobject_free);
drivers/infiniband/core/rdma_core.c
630
kref_get(&uobj->ufile->ref);
drivers/infiniband/core/rdma_core.c
836
kref_put(&ufile->ref, ib_uverbs_release_file);
drivers/infiniband/core/sa_query.c
1184
struct ib_sa_sm_ah *sm_ah = container_of(kref, struct ib_sa_sm_ah, ref);
drivers/infiniband/core/sa_query.c
1340
kref_get(&query->port->sm_ah->ref);
drivers/infiniband/core/sa_query.c
1350
kref_put(&query->sm_ah->ref, free_sm_ah);
drivers/infiniband/core/sa_query.c
1361
kref_put(&query->sm_ah->ref, free_sm_ah);
drivers/infiniband/core/sa_query.c
1373
kref_put(&query->sm_ah->ref, free_sm_ah);
drivers/infiniband/core/sa_query.c
2313
kref_init(&new_ah->ref);
drivers/infiniband/core/sa_query.c
2360
kref_put(&port->sm_ah->ref, free_sm_ah);
drivers/infiniband/core/sa_query.c
2385
kref_put(&port->sm_ah->ref, free_sm_ah);
drivers/infiniband/core/sa_query.c
2499
kref_put(&sa_dev->port[i].sm_ah->ref, free_sm_ah);
drivers/infiniband/core/sa_query.c
66
struct kref ref;
drivers/infiniband/core/ucaps.c
158
kref_get(&ucap->ref);
drivers/infiniband/core/ucaps.c
184
kref_init(&ucap->ref);
drivers/infiniband/core/ucaps.c
198
static void ib_release_ucap(struct kref *ref)
drivers/infiniband/core/ucaps.c
200
struct ib_ucap *ucap = container_of(ref, struct ib_ucap, ref);
drivers/infiniband/core/ucaps.c
23
struct kref ref;
drivers/infiniband/core/ucaps.c
230
kref_put(&ucap->ref, ib_release_ucap);
drivers/infiniband/core/ucma.c
148
if (!refcount_inc_not_zero(&ctx->ref))
drivers/infiniband/core/ucma.c
156
if (refcount_dec_and_test(&ctx->ref))
drivers/infiniband/core/ucma.c
1648
else if (!refcount_inc_not_zero(&mc->ctx->ref))
drivers/infiniband/core/ucma.c
220
refcount_set(&ctx->ref, 1);
drivers/infiniband/core/ucma.c
584
if (refcount_read(&ctx->ref))
drivers/infiniband/core/ucma.c
89
refcount_t ref;
drivers/infiniband/core/uverbs.h
214
void ib_uverbs_release_file(struct kref *ref);
drivers/infiniband/core/uverbs_main.c
1022
kref_put(&file->ref, ib_uverbs_release_file);
drivers/infiniband/core/uverbs_main.c
1235
kref_get(&file->ref);
drivers/infiniband/core/uverbs_main.c
1245
kref_put(&file->ref, ib_uverbs_release_file);
drivers/infiniband/core/uverbs_main.c
197
void ib_uverbs_release_file(struct kref *ref)
drivers/infiniband/core/uverbs_main.c
200
container_of(ref, struct ib_uverbs_file, ref);
drivers/infiniband/core/uverbs_main.c
979
kref_init(&file->ref);
drivers/infiniband/hw/bnxt_re/ib_verbs.c
2596
struct bnxt_qplib_sge ref, sge;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
2611
ref.addr = wqe->sg_list[0].addr;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
2612
ref.lkey = wqe->sg_list[0].lkey;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
2613
ref.size = wqe->sg_list[0].size;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
2623
sqp_entry->sge.addr = ref.addr;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
2624
sqp_entry->sge.lkey = ref.lkey;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
2625
sqp_entry->sge.size = ref.size;
drivers/infiniband/hw/erdma/erdma_cm.c
104
kref_init(&cep->ref);
drivers/infiniband/hw/erdma/erdma_cm.c
179
static void __erdma_cep_dealloc(struct kref *ref)
drivers/infiniband/hw/erdma/erdma_cm.c
181
struct erdma_cep *cep = container_of(ref, struct erdma_cep, ref);
drivers/infiniband/hw/erdma/erdma_cm.c
323
WARN_ON(kref_read(&cep->ref) < 1);
drivers/infiniband/hw/erdma/erdma_cm.c
324
kref_put(&cep->ref, __erdma_cep_dealloc);
drivers/infiniband/hw/erdma/erdma_cm.c
329
kref_get(&cep->ref);
drivers/infiniband/hw/erdma/erdma_cm.h
91
struct kref ref;
drivers/infiniband/hw/erdma/erdma_qp.c
307
static void erdma_qp_safe_free(struct kref *ref)
drivers/infiniband/hw/erdma/erdma_qp.c
309
struct erdma_qp *qp = container_of(ref, struct erdma_qp, ref);
drivers/infiniband/hw/erdma/erdma_qp.c
316
WARN_ON(kref_read(&qp->ref) < 1);
drivers/infiniband/hw/erdma/erdma_qp.c
317
kref_put(&qp->ref, erdma_qp_safe_free);
drivers/infiniband/hw/erdma/erdma_qp.c
322
kref_get(&qp->ref);
drivers/infiniband/hw/erdma/erdma_verbs.c
1017
kref_init(&qp->ref);
drivers/infiniband/hw/erdma/erdma_verbs.h
298
struct kref ref;
drivers/infiniband/hw/hfi1/affinity.c
790
static void hfi1_irq_notifier_release(struct kref *ref)
drivers/infiniband/hw/mlx4/mcg.c
141
#define safe_atomic_dec(ref) \
drivers/infiniband/hw/mlx4/mcg.c
143
if (atomic_dec_and_test(ref)) \
drivers/infiniband/hw/mlx4/mcg.c
609
int ref = 0;
drivers/infiniband/hw/mlx4/mcg.c
624
++ref;
drivers/infiniband/hw/mlx4/mcg.c
633
ref = 1;
drivers/infiniband/hw/mlx4/mcg.c
639
return ref;
drivers/infiniband/hw/mlx5/dm.c
166
container_of(kref, struct mlx5_ib_dm_memic, ref);
drivers/infiniband/hw/mlx5/dm.c
253
kref_get(&dm->ref);
drivers/infiniband/hw/mlx5/dm.c
296
kref_init(&dm->ref);
drivers/infiniband/hw/mlx5/dm.c
546
kref_put(&mdm->ref, mlx5_ib_dm_memic_free);
drivers/infiniband/hw/mlx5/dm.c
555
kref_put(&mdm->ref, mlx5_ib_dm_memic_free);
drivers/infiniband/hw/mlx5/dm.h
33
struct kref ref;
drivers/infiniband/hw/qedr/qedr_iw_cm.c
82
static void qedr_iw_free_qp(struct kref *ref)
drivers/infiniband/hw/qedr/qedr_iw_cm.c
84
struct qedr_qp *qp = container_of(ref, struct qedr_qp, refcnt);
drivers/infiniband/hw/qedr/qedr_iw_cm.c
90
qedr_iw_free_ep(struct kref *ref)
drivers/infiniband/hw/qedr/qedr_iw_cm.c
92
struct qedr_iw_ep *ep = container_of(ref, struct qedr_iw_ep, refcnt);
drivers/infiniband/sw/rdmavt/cq.c
288
kref_put(&cq->ip->ref, rvt_release_mmap_info);
drivers/infiniband/sw/rdmavt/mmap.c
143
kref_init(&ip->ref);
drivers/infiniband/sw/rdmavt/mmap.c
28
void rvt_release_mmap_info(struct kref *ref)
drivers/infiniband/sw/rdmavt/mmap.c
31
container_of(ref, struct rvt_mmap_info, ref);
drivers/infiniband/sw/rdmavt/mmap.c
46
kref_get(&ip->ref);
drivers/infiniband/sw/rdmavt/mmap.c
53
kref_put(&ip->ref, rvt_release_mmap_info);
drivers/infiniband/sw/rdmavt/mmap.h
12
void rvt_release_mmap_info(struct kref *ref);
drivers/infiniband/sw/rdmavt/mr.c
85
static void __rvt_mregion_complete(struct percpu_ref *ref)
drivers/infiniband/sw/rdmavt/mr.c
87
struct rvt_mregion *mr = container_of(ref, struct rvt_mregion,
drivers/infiniband/sw/rdmavt/qp.c
1252
kref_put(&qp->ip->ref, rvt_release_mmap_info);
drivers/infiniband/sw/rdmavt/qp.c
1695
kref_put(&qp->ip->ref, rvt_release_mmap_info);
drivers/infiniband/sw/rdmavt/srq.c
303
kref_put(&srq->ip->ref, rvt_release_mmap_info);
drivers/infiniband/sw/rxe/rxe_loc.h
46
struct kref ref;
drivers/infiniband/sw/rxe/rxe_loc.h
52
void rxe_mmap_release(struct kref *ref);
drivers/infiniband/sw/rxe/rxe_mmap.c
145
kref_init(&ip->ref);
drivers/infiniband/sw/rxe/rxe_mmap.c
16
void rxe_mmap_release(struct kref *ref)
drivers/infiniband/sw/rxe/rxe_mmap.c
18
struct rxe_mmap_info *ip = container_of(ref,
drivers/infiniband/sw/rxe/rxe_mmap.c
19
struct rxe_mmap_info, ref);
drivers/infiniband/sw/rxe/rxe_mmap.c
41
kref_get(&ip->ref);
drivers/infiniband/sw/rxe/rxe_mmap.c
48
kref_put(&ip->ref, rxe_mmap_release);
drivers/infiniband/sw/rxe/rxe_queue.c
196
kref_put(&q->ip->ref, rxe_mmap_release);
drivers/infiniband/sw/siw/siw.h
145
struct kref ref;
drivers/infiniband/sw/siw/siw.h
420
struct kref ref;
drivers/infiniband/sw/siw/siw.h
514
void siw_free_qp(struct kref *ref);
drivers/infiniband/sw/siw/siw.h
601
if (likely(qp && kref_get_unless_zero(&qp->ref))) {
drivers/infiniband/sw/siw/siw.h
616
kref_get(&qp->ref);
drivers/infiniband/sw/siw/siw.h
621
kref_put(&qp->ref, siw_free_qp);
drivers/infiniband/sw/siw/siw_cm.c
218
kref_init(&cep->ref);
drivers/infiniband/sw/siw/siw_cm.c
295
static void __siw_cep_dealloc(struct kref *ref)
drivers/infiniband/sw/siw/siw_cm.c
297
struct siw_cep *cep = container_of(ref, struct siw_cep, ref);
drivers/infiniband/sw/siw/siw_cm.c
496
WARN_ON(kref_read(&cep->ref) < 1);
drivers/infiniband/sw/siw/siw_cm.c
497
kref_put(&cep->ref, __siw_cep_dealloc);
drivers/infiniband/sw/siw/siw_cm.c
508
kref_get(&cep->ref);
drivers/infiniband/sw/siw/siw_cm.h
42
struct kref ref;
drivers/infiniband/sw/siw/siw_mem.c
108
void siw_free_mem(struct kref *ref)
drivers/infiniband/sw/siw/siw_mem.c
110
struct siw_mem *mem = container_of(ref, struct siw_mem, ref);
drivers/infiniband/sw/siw/siw_mem.c
33
if (likely(mem && kref_get_unless_zero(&mem->ref))) {
drivers/infiniband/sw/siw/siw_mem.c
75
kref_init(&mem->ref);
drivers/infiniband/sw/siw/siw_mem.h
25
void siw_free_mem(struct kref *ref);
drivers/infiniband/sw/siw/siw_mem.h
29
kref_put(&mem->ref, siw_free_mem);
drivers/infiniband/sw/siw/siw_qp.c
1288
kref_init(&qp->ref);
drivers/infiniband/sw/siw/siw_qp.c
1295
void siw_free_qp(struct kref *ref)
drivers/infiniband/sw/siw/siw_qp.c
1297
struct siw_qp *found, *qp = container_of(ref, struct siw_qp, ref);
drivers/infiniband/ulp/ipoib/ipoib.h
421
struct kref ref;
drivers/infiniband/ulp/ipoib/ipoib.h
488
kref_put(&ah->ref, ipoib_free_ah);
drivers/infiniband/ulp/ipoib/ipoib_ib.c
68
kref_init(&ah->ref);
drivers/infiniband/ulp/ipoib/ipoib_ib.c
84
struct ipoib_ah *ah = container_of(kref, struct ipoib_ah, ref);
drivers/infiniband/ulp/ipoib/ipoib_main.c
1027
kref_get(&path->ah->ref);
drivers/infiniband/ulp/ipoib/ipoib_main.c
861
kref_get(&path->ah->ref);
drivers/infiniband/ulp/ipoib/ipoib_multicast.c
805
kref_get(&mcast->ah->ref);
drivers/infiniband/ulp/rtrs/rtrs-clt.c
1160
refcount_dec(&req->ref);
drivers/infiniband/ulp/rtrs/rtrs-clt.c
422
refcount_inc(&req->ref);
drivers/infiniband/ulp/rtrs/rtrs-clt.c
430
if (!refcount_dec_and_test(&req->ref))
drivers/infiniband/ulp/rtrs/rtrs-clt.c
436
if (!refcount_dec_and_test(&req->ref))
drivers/infiniband/ulp/rtrs/rtrs-clt.c
972
refcount_set(&req->ref, 1);
drivers/infiniband/ulp/rtrs/rtrs-clt.h
115
refcount_t ref;
drivers/infiniband/ulp/rtrs/rtrs-pri.h
85
struct kref ref;
drivers/infiniband/ulp/rtrs/rtrs-srv.c
125
static inline void rtrs_srv_inflight_ref_release(struct percpu_ref *ref)
drivers/infiniband/ulp/rtrs/rtrs-srv.c
127
struct rtrs_srv_path *srv_path = container_of(ref,
drivers/infiniband/ulp/rtrs/rtrs.c
578
static void dev_free(struct kref *ref)
drivers/infiniband/ulp/rtrs/rtrs.c
583
dev = container_of(ref, typeof(*dev), ref);
drivers/infiniband/ulp/rtrs/rtrs.c
599
return kref_put(&dev->ref, dev_free);
drivers/infiniband/ulp/rtrs/rtrs.c
605
return kref_get_unless_zero(&dev->ref);
drivers/infiniband/ulp/rtrs/rtrs.c
625
kref_init(&dev->ref);
drivers/infiniband/ulp/srpt/ib_srpt.c
113
refcount_t ref;
drivers/infiniband/ulp/srpt/ib_srpt.c
126
refcount_inc(&e->ref);
drivers/infiniband/ulp/srpt/ib_srpt.c
133
refcount_set(&e->ref, 1);
drivers/infiniband/ulp/srpt/ib_srpt.c
161
if (!refcount_dec_and_test(&e->ref))
drivers/iommu/iommu-debug-pagealloc.c
20
atomic_t ref;
drivers/iommu/iommu-debug-pagealloc.c
47
WARN_ON(atomic_inc_return_relaxed(&d->ref) <= 0);
drivers/iommu/iommu-debug-pagealloc.c
60
WARN_ON(atomic_dec_return_relaxed(&d->ref) < 0);
drivers/iommu/iommu-debug-pagealloc.c
79
unsigned int ref;
drivers/iommu/iommu-debug-pagealloc.c
83
ref = atomic_read(&d->ref);
drivers/iommu/iommu-debug-pagealloc.c
85
return ref != 0;
drivers/iommu/iommufd/device.c
29
container_of(kref, struct iommufd_group, ref);
drivers/iommu/iommufd/device.c
42
kref_put(&group->ref, iommufd_group_release);
drivers/iommu/iommufd/device.c
56
return kref_get_unless_zero(&igroup->ref);
drivers/iommu/iommufd/device.c
96
kref_init(&new_igroup->ref);
drivers/iommu/iommufd/iommufd_private.h
470
struct kref ref;
drivers/macintosh/windfarm.h
109
struct kref ref;
drivers/macintosh/windfarm.h
41
struct kref ref;
drivers/macintosh/windfarm_ad7417_sensor.c
152
static void wf_ad7417_release(struct kref *ref)
drivers/macintosh/windfarm_ad7417_sensor.c
154
struct wf_ad7417_priv *pv = container_of(ref,
drivers/macintosh/windfarm_ad7417_sensor.c
155
struct wf_ad7417_priv, ref);
drivers/macintosh/windfarm_ad7417_sensor.c
164
kref_put(&pv->ref, wf_ad7417_release);
drivers/macintosh/windfarm_ad7417_sensor.c
187
kref_get(&pv->ref);
drivers/macintosh/windfarm_ad7417_sensor.c
267
kref_init(&pv->ref);
drivers/macintosh/windfarm_ad7417_sensor.c
27
struct kref ref;
drivers/macintosh/windfarm_ad7417_sensor.c
303
kref_put(&pv->ref, wf_ad7417_release);
drivers/macintosh/windfarm_core.c
149
struct wf_control *ct = container_of(kref, struct wf_control, ref);
drivers/macintosh/windfarm_core.c
220
kref_init(&new_ct->ref);
drivers/macintosh/windfarm_core.c
250
kref_put(&ct->ref, wf_control_release);
drivers/macintosh/windfarm_core.c
258
kref_get(&ct->ref);
drivers/macintosh/windfarm_core.c
266
kref_put(&ct->ref, wf_control_release);
drivers/macintosh/windfarm_core.c
279
struct wf_sensor *sr = container_of(kref, struct wf_sensor, ref);
drivers/macintosh/windfarm_core.c
315
kref_init(&new_sr->ref);
drivers/macintosh/windfarm_core.c
353
kref_get(&sr->ref);
drivers/macintosh/windfarm_core.c
361
kref_put(&sr->ref, wf_sensor_release);
drivers/macintosh/windfarm_fcu_controls.c
397
kref_get(&pv->ref);
drivers/macintosh/windfarm_fcu_controls.c
525
kref_init(&pv->ref);
drivers/macintosh/windfarm_fcu_controls.c
56
struct kref ref;
drivers/macintosh/windfarm_fcu_controls.c
572
kref_put(&pv->ref, wf_fcu_release);
drivers/macintosh/windfarm_fcu_controls.c
71
static void wf_fcu_release(struct kref *ref)
drivers/macintosh/windfarm_fcu_controls.c
73
struct wf_fcu_priv *pv = container_of(ref, struct wf_fcu_priv, ref);
drivers/macintosh/windfarm_fcu_controls.c
82
kref_put(&fan->fcu_priv->ref, wf_fcu_release);
drivers/macintosh/windfarm_smu_sat.c
168
static void wf_sat_release(struct kref *ref)
drivers/macintosh/windfarm_smu_sat.c
170
struct wf_sat *sat = container_of(ref, struct wf_sat, ref);
drivers/macintosh/windfarm_smu_sat.c
184
kref_put(&sat->ref, wf_sat_release);
drivers/macintosh/windfarm_smu_sat.c
211
kref_init(&sat->ref);
drivers/macintosh/windfarm_smu_sat.c
28
struct kref ref;
drivers/macintosh/windfarm_smu_sat.c
282
kref_get(&sat->ref);
drivers/macintosh/windfarm_smu_sat.c
309
kref_get(&sat->ref);
drivers/macintosh/windfarm_smu_sat.c
332
kref_put(&sat->ref, wf_sat_release);
drivers/md/bcache/journal.c
888
atomic_t *ref;
drivers/md/bcache/journal.c
892
ref = bch_journal(c, &keys, cl);
drivers/md/bcache/journal.c
893
if (ref)
drivers/md/bcache/journal.c
894
atomic_dec_bug(ref);
drivers/md/dm-era-target.c
1200
int (*fn1)(struct era_metadata *md, void *ref);
drivers/md/dm-era-target.c
1404
int (*fn)(struct era_metadata *md, void *ref), void *arg)
drivers/md/dm-pcache/cache.h
113
struct kref ref;
drivers/md/dm-pcache/cache_key.c
11
kref_init(&key->ref);
drivers/md/dm-pcache/cache_key.c
40
kref_get(&key->ref);
drivers/md/dm-pcache/cache_key.c
50
static void cache_key_destroy(struct kref *ref)
drivers/md/dm-pcache/cache_key.c
52
struct pcache_cache_key *key = container_of(ref, struct pcache_cache_key, ref);
drivers/md/dm-pcache/cache_key.c
60
kref_put(&key->ref, cache_key_destroy);
drivers/md/dm-pcache/dm_pcache.c
375
kref_init(&pcache_req->ref);
drivers/md/dm-pcache/dm_pcache.c
63
kref_get(&pcache_req->ref);
drivers/md/dm-pcache/dm_pcache.c
66
static void end_req(struct kref *ref)
drivers/md/dm-pcache/dm_pcache.c
68
struct pcache_request *pcache_req = container_of(ref, struct pcache_request, ref);
drivers/md/dm-pcache/dm_pcache.c
91
kref_put(&pcache_req->ref, end_req);
drivers/md/dm-pcache/dm_pcache.h
56
struct kref ref;
drivers/md/dm-snap-persistent.c
98
void (*callback)(void *ref, int success);
drivers/md/dm-zoned-metadata.c
113
unsigned int ref;
drivers/md/dm-zoned-metadata.c
2802
(u64)mblk->no, mblk->ref);
drivers/md/dm-zoned-metadata.c
2820
(u64)mblk->no, mblk->ref);
drivers/md/dm-zoned-metadata.c
2821
mblk->ref = 0;
drivers/md/dm-zoned-metadata.c
434
mblk->ref = 0;
drivers/md/dm-zoned-metadata.c
494
mblk->ref++;
drivers/md/dm-zoned-metadata.c
495
if (mblk->ref == 1 &&
drivers/md/dm-zoned-metadata.c
565
mblk->ref++;
drivers/md/dm-zoned-metadata.c
646
mblk->ref--;
drivers/md/dm-zoned-metadata.c
647
if (mblk->ref == 0) {
drivers/md/dm-zoned-metadata.c
948
if (mblk->ref == 0)
drivers/md/dm-zoned-target.c
141
refcount_inc(&bioctx->ref);
drivers/md/dm-zoned-target.c
23
refcount_t ref;
drivers/md/dm-zoned-target.c
656
refcount_set(&bioctx->ref, 1);
drivers/md/dm-zoned-target.c
85
if (refcount_dec_and_test(&bioctx->ref)) {
drivers/md/md-llbitmap.c
509
static void active_release(struct percpu_ref *ref)
drivers/md/md-llbitmap.c
512
container_of(ref, struct llbitmap_page_ctl, active);
drivers/md/md.c
672
static void active_io_release(struct percpu_ref *ref)
drivers/md/md.c
674
struct mddev *mddev = container_of(ref, struct mddev, active_io);
drivers/media/common/siano/smsdvb-debugfs.c
244
static void smsdvb_debugfs_data_release(struct kref *ref)
drivers/media/common/siano/smsdvb-debugfs.c
248
debug_data = container_of(ref, struct smsdvb_debugfs, refcount);
drivers/media/dvb-core/dvb_ca_en50221.c
174
static void dvb_ca_private_release(struct kref *ref)
drivers/media/dvb-core/dvb_ca_en50221.c
178
ca = container_of(ref, struct dvb_ca_private, refcount);
drivers/media/dvb-core/dvb_frontend.c
146
static void dvb_frontend_free(struct kref *ref)
drivers/media/dvb-core/dvb_frontend.c
149
container_of(ref, struct dvb_frontend, refcount);
drivers/media/dvb-core/dvbdev.c
516
kref_init(&dvbdev->ref);
drivers/media/dvb-core/dvbdev.c
612
static void dvb_free_device(struct kref *ref)
drivers/media/dvb-core/dvbdev.c
614
struct dvb_device *dvbdev = container_of(ref, struct dvb_device, ref);
drivers/media/dvb-core/dvbdev.c
621
kref_get(&dvbdev->ref);
drivers/media/dvb-core/dvbdev.c
629
kref_put(&dvbdev->ref, dvb_free_device);
drivers/media/dvb-frontends/dib0090.c
918
u16 rf, bb, ref;
drivers/media/dvb-frontends/dib0090.c
970
ref = rf;
drivers/media/dvb-frontends/dib0090.c
972
if (g[0] == 0 || ref < (g[1] - g[0])) /* if total gain of the current amp is null or this amp is not concerned because it starts to work from an higher gain value */
drivers/media/dvb-frontends/dib0090.c
974
else if (ref >= g[1]) /* Gain to set is higher than the high working point of this amp */
drivers/media/dvb-frontends/dib0090.c
977
v = ((ref - (g[1] - g[0])) * g[2]) / g[0];
drivers/media/dvb-frontends/dib0090.c
999
ref = bb;
drivers/media/dvb-frontends/ix2505v.c
126
u8 gain, cc, ref, psc, local_osc, lpf;
drivers/media/dvb-frontends/ix2505v.c
144
ref = 8; /* REF =1 */
drivers/media/dvb-frontends/ix2505v.c
147
div_factor = (frequency * ref) / 40; /* local osc = 4Mhz */
drivers/media/dvb-frontends/stv6110.c
252
u32 divider, ref, p, presc, i, result_freq, vco_freq;
drivers/media/dvb-frontends/stv6110.c
299
ref = priv->mclk / ((1 << (r_div_opt + 1)) * (1 << (p + 1)));
drivers/media/dvb-frontends/stv6110.c
300
divider = (((frequency * 1000) + (ref >> 1)) / ref);
drivers/media/i2c/ds90ub913.c
681
unsigned long ref;
drivers/media/i2c/ds90ub913.c
684
ref = clk_get_rate(priv->clkin) / 2;
drivers/media/i2c/ds90ub913.c
686
scl_high = div64_u64((u64)scl_high * ref, 1000000000);
drivers/media/i2c/ds90ub913.c
687
scl_low = div64_u64((u64)scl_low * ref, 1000000000);
drivers/media/i2c/ds90ub953.c
809
u32 ref = 26250000;
drivers/media/i2c/ds90ub953.c
814
scl_high = div64_u64((u64)scl_high * ref, 1000000000) - 5;
drivers/media/i2c/ds90ub953.c
815
scl_low = div64_u64((u64)scl_low * ref, 1000000000) - 5;
drivers/media/pci/intel/ipu6/ipu6-isys-mcd-phy.c
673
refcount_t *ref;
drivers/media/pci/intel/ipu6/ipu6-isys-mcd-phy.c
679
ref = &phy_power_ref_count[phy_id];
drivers/media/pci/intel/ipu6/ipu6-isys-mcd-phy.c
690
if (refcount_read(ref)) {
drivers/media/pci/intel/ipu6/ipu6-isys-mcd-phy.c
692
refcount_inc(ref);
drivers/media/pci/intel/ipu6/ipu6-isys-mcd-phy.c
712
refcount_set(ref, 1);
drivers/media/pci/intel/ipu6/ipu6-isys-mcd-phy.c
716
if (!refcount_dec_and_test(ref))
drivers/media/platform/allegro-dvt/allegro-core.c
202
struct kref ref;
drivers/media/platform/allegro-dvt/allegro-core.c
3112
kref_init(&channel->ref);
drivers/media/platform/allegro-dvt/allegro-core.c
444
if (kref_get_unless_zero(&channel->ref))
drivers/media/platform/allegro-dvt/allegro-core.c
463
if (kref_get_unless_zero(&channel->ref))
drivers/media/platform/allegro-dvt/allegro-core.c
472
static void allegro_free_channel(struct kref *ref)
drivers/media/platform/allegro-dvt/allegro-core.c
474
struct allegro_channel *channel = container_of(ref, struct allegro_channel, ref);
drivers/media/platform/allegro-dvt/allegro-core.c
481
return kref_put(&channel->ref, allegro_free_channel);
drivers/media/platform/amphion/venc.c
1019
if (ref_frame_num > ARRAY_SIZE(venc->ref)) {
drivers/media/platform/amphion/venc.c
1037
venc->ref[i].length = ref_frame_size;
drivers/media/platform/amphion/venc.c
1038
ret = vpu_alloc_dma(inst->core, &venc->ref[i]);
drivers/media/platform/amphion/venc.c
1055
vpu_iface_config_memory_resource(inst, MEM_RES_REF, i, &venc->ref[i]);
drivers/media/platform/amphion/venc.c
45
struct vpu_buffer ref[VENC_MAX_BUF_CNT];
drivers/media/platform/amphion/venc.c
998
for (i = 0; i < ARRAY_SIZE(venc->ref); i++)
drivers/media/platform/amphion/venc.c
999
vpu_free_dma(&venc->ref[i]);
drivers/media/platform/mediatek/vcodec/decoder/mtk_vcodec_dec_stateless.c
245
static void mtk_vcodec_dec_request_release(struct kref *ref)
drivers/media/platform/mediatek/vcodec/decoder/mtk_vcodec_dec_stateless.c
248
container_of(ref, struct mtk_vcodec_dec_request, refcount);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1821
struct vdec_av1_slice_fb *vref = &vsi->ref[i];
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
659
struct vdec_av1_slice_fb ref[V4L2_AV1_REFS_PER_FRAME];
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1667
struct vdec_vp9_slice_reference *ref;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1704
ref = &vsi->frame.ref[i];
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1707
ref->frame_width = w;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1708
ref->frame_height = h;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1709
memset(&vsi->ref[i], 0, sizeof(vsi->ref[i]));
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1713
ref->frame_width = instance->dpb[idx].width;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1714
ref->frame_height = instance->dpb[idx].height;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1715
vsi->ref[i].y.dma_addr =
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1718
vsi->ref[i].c.dma_addr =
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1719
vsi->ref[i].y.dma_addr + size;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1721
vsi->ref[i].c.dma_addr =
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
276
struct vdec_vp9_slice_reference ref[3];
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
357
struct vdec_vp9_slice_fb ref[3];
drivers/media/platform/nvidia/tegra-vde/dmabuf-cache.c
143
ref:
drivers/media/platform/nvidia/tegra-vde/dmabuf-cache.c
95
goto ref;
drivers/media/platform/nvidia/tegra-vde/h264.c
762
struct vb2_buffer *ref;
drivers/media/platform/nvidia/tegra-vde/h264.c
808
ref = get_ref_buf(ctx, dst, dpb_idx);
drivers/media/platform/nvidia/tegra-vde/h264.c
810
err = tegra_vde_h264_setup_frame(ctx, h264, &b, ref, dpb_idx,
drivers/media/platform/renesas/rcar_jpu.c
892
struct jpu_q_data *src_q_data, *dst_q_data, *orig, adj, *ref;
drivers/media/platform/renesas/rcar_jpu.c
902
ref = dst_q_data;
drivers/media/platform/renesas/rcar_jpu.c
907
ref = src_q_data;
drivers/media/platform/renesas/rcar_jpu.c
911
adj.format.width = ref->format.width;
drivers/media/platform/renesas/rcar_jpu.c
912
adj.format.height = ref->format.height;
drivers/media/platform/rockchip/rkvdec/rkvdec-h264-common.c
119
struct v4l2_h264_reference *ref;
drivers/media/platform/rockchip/rkvdec/rkvdec-h264-common.c
125
ref = &reflists->p[i];
drivers/media/platform/rockchip/rkvdec/rkvdec-h264-common.c
128
ref = &reflists->b0[i];
drivers/media/platform/rockchip/rkvdec/rkvdec-h264-common.c
131
ref = &reflists->b1[i];
drivers/media/platform/rockchip/rkvdec/rkvdec-h264-common.c
135
if (WARN_ON(ref->index >= ARRAY_SIZE(dec_params->dpb)))
drivers/media/platform/rockchip/rkvdec/rkvdec-h264-common.c
138
dpb_valid = !!(run->ref_buf[ref->index]);
drivers/media/platform/rockchip/rkvdec/rkvdec-h264-common.c
139
bottom = ref->fields == V4L2_H264_BOTTOM_FIELD_REF;
drivers/media/platform/rockchip/rkvdec/rkvdec-h264-common.c
141
set_dpb_info(hw_rps->entries, j, i, ref->index | (dpb_valid << 4), bottom);
drivers/media/platform/verisilicon/hantro_g1_vp8_dec.c
374
dma_addr_t ref;
drivers/media/platform/verisilicon/hantro_g1_vp8_dec.c
377
ref = hantro_get_ref(ctx, hdr->last_frame_ts);
drivers/media/platform/verisilicon/hantro_g1_vp8_dec.c
378
if (!ref) {
drivers/media/platform/verisilicon/hantro_g1_vp8_dec.c
381
ref = vb2_dma_contig_plane_dma_addr(&vb2_dst->vb2_buf, 0);
drivers/media/platform/verisilicon/hantro_g1_vp8_dec.c
383
vdpu_write_relaxed(vpu, ref, G1_REG_ADDR_REF(0));
drivers/media/platform/verisilicon/hantro_g1_vp8_dec.c
385
ref = hantro_get_ref(ctx, hdr->golden_frame_ts);
drivers/media/platform/verisilicon/hantro_g1_vp8_dec.c
386
if (!ref && hdr->golden_frame_ts)
drivers/media/platform/verisilicon/hantro_g1_vp8_dec.c
389
if (!ref)
drivers/media/platform/verisilicon/hantro_g1_vp8_dec.c
390
ref = vb2_dma_contig_plane_dma_addr(&vb2_dst->vb2_buf, 0);
drivers/media/platform/verisilicon/hantro_g1_vp8_dec.c
392
ref |= G1_REG_ADDR_REF_TOPC_E;
drivers/media/platform/verisilicon/hantro_g1_vp8_dec.c
393
vdpu_write_relaxed(vpu, ref, G1_REG_ADDR_REF(4));
drivers/media/platform/verisilicon/hantro_g1_vp8_dec.c
395
ref = hantro_get_ref(ctx, hdr->alt_frame_ts);
drivers/media/platform/verisilicon/hantro_g1_vp8_dec.c
396
if (!ref && hdr->alt_frame_ts)
drivers/media/platform/verisilicon/hantro_g1_vp8_dec.c
399
if (!ref)
drivers/media/platform/verisilicon/hantro_g1_vp8_dec.c
400
ref = vb2_dma_contig_plane_dma_addr(&vb2_dst->vb2_buf, 0);
drivers/media/platform/verisilicon/hantro_g1_vp8_dec.c
402
ref |= G1_REG_ADDR_REF_TOPC_E;
drivers/media/platform/verisilicon/hantro_g1_vp8_dec.c
403
vdpu_write_relaxed(vpu, ref, G1_REG_ADDR_REF(5));
drivers/media/platform/verisilicon/rockchip_vpu2_hw_vp8_dec.c
451
dma_addr_t ref;
drivers/media/platform/verisilicon/rockchip_vpu2_hw_vp8_dec.c
453
ref = hantro_get_ref(ctx, hdr->last_frame_ts);
drivers/media/platform/verisilicon/rockchip_vpu2_hw_vp8_dec.c
454
if (!ref) {
drivers/media/platform/verisilicon/rockchip_vpu2_hw_vp8_dec.c
457
ref = vb2_dma_contig_plane_dma_addr(&vb2_dst->vb2_buf, 0);
drivers/media/platform/verisilicon/rockchip_vpu2_hw_vp8_dec.c
459
vdpu_write_relaxed(vpu, ref, VDPU_REG_VP8_ADDR_REF0);
drivers/media/platform/verisilicon/rockchip_vpu2_hw_vp8_dec.c
461
ref = hantro_get_ref(ctx, hdr->golden_frame_ts);
drivers/media/platform/verisilicon/rockchip_vpu2_hw_vp8_dec.c
462
if (!ref && hdr->golden_frame_ts)
drivers/media/platform/verisilicon/rockchip_vpu2_hw_vp8_dec.c
465
if (!ref)
drivers/media/platform/verisilicon/rockchip_vpu2_hw_vp8_dec.c
466
ref = vb2_dma_contig_plane_dma_addr(&vb2_dst->vb2_buf, 0);
drivers/media/platform/verisilicon/rockchip_vpu2_hw_vp8_dec.c
468
ref |= VDPU_REG_VP8_GREF_SIGN_BIAS;
drivers/media/platform/verisilicon/rockchip_vpu2_hw_vp8_dec.c
469
vdpu_write_relaxed(vpu, ref, VDPU_REG_VP8_ADDR_REF2_5(2));
drivers/media/platform/verisilicon/rockchip_vpu2_hw_vp8_dec.c
471
ref = hantro_get_ref(ctx, hdr->alt_frame_ts);
drivers/media/platform/verisilicon/rockchip_vpu2_hw_vp8_dec.c
472
if (!ref && hdr->alt_frame_ts)
drivers/media/platform/verisilicon/rockchip_vpu2_hw_vp8_dec.c
475
if (!ref)
drivers/media/platform/verisilicon/rockchip_vpu2_hw_vp8_dec.c
476
ref = vb2_dma_contig_plane_dma_addr(&vb2_dst->vb2_buf, 0);
drivers/media/platform/verisilicon/rockchip_vpu2_hw_vp8_dec.c
478
ref |= VDPU_REG_VP8_AREF_SIGN_BIAS;
drivers/media/platform/verisilicon/rockchip_vpu2_hw_vp8_dec.c
479
vdpu_write_relaxed(vpu, ref, VDPU_REG_VP8_ADDR_REF2_5(3));
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
118
static int rockchip_vpu981_get_frame_index(struct hantro_ctx *ctx, int ref)
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
124
int i, idx = frame->ref_frame_idx[ref];
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
140
static int rockchip_vpu981_get_order_hint(struct hantro_ctx *ctx, int ref)
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
143
int idx = rockchip_vpu981_get_frame_index(ctx, ref);
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
1904
u32 ref = i - 1;
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
1913
if (rockchip_vpu981_get_frame_index(ctx, ref) > 0)
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
1914
idx = rockchip_vpu981_get_frame_index(ctx, ref);
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
1920
rockchip_vpu981_av1_dec_set_ref(ctx, ref, idx, width,
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
1923
rockchip_vpu981_av1_dec_set_sign_bias(ctx, ref,
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
198
int ref, idx;
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
207
for (ref = 0; ref < V4L2_AV1_TOTAL_REFS_PER_FRAME; ref++) {
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
208
if (ctrls->frame->reference_frame_ts[ref] == timestamp)
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
688
rockchip_vpu981_av1_dec_set_ref(struct hantro_ctx *ctx, int ref, int idx,
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
704
switch (ref) {
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
756
hantro_write_addr(vpu, AV1_REFERENCE_Y(ref), luma_addr);
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
757
hantro_write_addr(vpu, AV1_REFERENCE_CB(ref), chroma_addr);
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
758
hantro_write_addr(vpu, AV1_REFERENCE_MV(ref), mv_addr);
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
765
int ref, int val)
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
769
switch (ref) {
drivers/media/test-drivers/vicodec/codec-fwht.c
661
static void add_deltas(s16 *deltas, const u8 *ref, int stride,
drivers/media/test-drivers/vicodec/codec-fwht.c
668
*deltas += *ref;
drivers/media/test-drivers/vicodec/codec-fwht.c
669
ref += ref_step;
drivers/media/test-drivers/vicodec/codec-fwht.c
680
ref += stride - (8 * ref_step);
drivers/media/test-drivers/vicodec/codec-fwht.c
836
u32 height, u32 width, const u8 *ref, u32 ref_stride,
drivers/media/test-drivers/vicodec/codec-fwht.c
845
bool is_intra = !ref;
drivers/media/test-drivers/vicodec/codec-fwht.c
871
const u8 *refp = ref + j * 8 * ref_stride +
drivers/media/test-drivers/vicodec/codec-fwht.c
912
unsigned int height, const struct fwht_raw_frame *ref,
drivers/media/test-drivers/vicodec/codec-fwht.c
921
if (!decode_plane(cf, &rlco, height, width, ref->luma, ref_stride,
drivers/media/test-drivers/vicodec/codec-fwht.c
922
ref->luma_alpha_step, dst->luma, dst_stride,
drivers/media/test-drivers/vicodec/codec-fwht.c
937
if (!decode_plane(cf, &rlco, h, w, ref->cb, ref_chroma_stride,
drivers/media/test-drivers/vicodec/codec-fwht.c
938
ref->chroma_step, dst->cb, dst_chroma_stride,
drivers/media/test-drivers/vicodec/codec-fwht.c
943
if (!decode_plane(cf, &rlco, h, w, ref->cr, ref_chroma_stride,
drivers/media/test-drivers/vicodec/codec-fwht.c
944
ref->chroma_step, dst->cr, dst_chroma_stride,
drivers/media/test-drivers/vicodec/codec-fwht.c
952
if (!decode_plane(cf, &rlco, height, width, ref->alpha, ref_stride,
drivers/media/test-drivers/vicodec/codec-fwht.c
953
ref->luma_alpha_step, dst->alpha, dst_stride,
drivers/media/test-drivers/vicodec/codec-fwht.h
114
unsigned int height, const struct fwht_raw_frame *ref,
drivers/media/usb/em28xx/em28xx-audio.c
275
kref_get(&dev->ref);
drivers/media/usb/em28xx/em28xx-audio.c
312
kref_put(&dev->ref, em28xx_free_device);
drivers/media/usb/em28xx/em28xx-audio.c
822
kref_get(&dev->ref);
drivers/media/usb/em28xx/em28xx-audio.c
917
kref_put(&dev->ref, em28xx_free_device);
drivers/media/usb/em28xx/em28xx-cards.c
3554
void em28xx_free_device(struct kref *ref)
drivers/media/usb/em28xx/em28xx-cards.c
3556
struct em28xx *dev = kref_to_dev(ref);
drivers/media/usb/em28xx/em28xx-cards.c
3981
kref_init(&dev->ref);
drivers/media/usb/em28xx/em28xx-cards.c
4083
kref_init(&dev->dev_next->ref);
drivers/media/usb/em28xx/em28xx-cards.c
4204
kref_put(&dev->dev_next->ref, em28xx_free_device);
drivers/media/usb/em28xx/em28xx-cards.c
4207
kref_put(&dev->ref, em28xx_free_device);
drivers/media/usb/em28xx/em28xx-dvb.c
2007
kref_get(&dev->ref);
drivers/media/usb/em28xx/em28xx-dvb.c
2075
kref_put(&dev->ref, em28xx_free_device);
drivers/media/usb/em28xx/em28xx-input.c
701
kref_get(&dev->ref);
drivers/media/usb/em28xx/em28xx-input.c
864
kref_put(&dev->ref, em28xx_free_device);
drivers/media/usb/em28xx/em28xx-video.c
2113
static void em28xx_free_v4l2(struct kref *ref)
drivers/media/usb/em28xx/em28xx-video.c
2115
struct em28xx_v4l2 *v4l2 = container_of(ref, struct em28xx_v4l2, ref);
drivers/media/usb/em28xx/em28xx-video.c
2180
kref_get(&dev->ref);
drivers/media/usb/em28xx/em28xx-video.c
2181
kref_get(&v4l2->ref);
drivers/media/usb/em28xx/em28xx-video.c
2240
kref_put(&v4l2->ref, em28xx_free_v4l2);
drivers/media/usb/em28xx/em28xx-video.c
2244
kref_put(&dev->ref, em28xx_free_device);
drivers/media/usb/em28xx/em28xx-video.c
2316
kref_put(&v4l2->ref, em28xx_free_v4l2);
drivers/media/usb/em28xx/em28xx-video.c
2318
kref_put(&dev->ref, em28xx_free_device);
drivers/media/usb/em28xx/em28xx-video.c
2537
kref_init(&v4l2->ref);
drivers/media/usb/em28xx/em28xx-video.c
2872
kref_get(&dev->ref);
drivers/media/usb/em28xx/em28xx-video.c
2901
kref_put(&v4l2->ref, em28xx_free_v4l2);
drivers/media/usb/em28xx/em28xx.h
543
struct kref ref;
drivers/media/usb/em28xx/em28xx.h
637
struct kref ref;
drivers/media/usb/em28xx/em28xx.h
777
#define kref_to_dev(d) container_of(d, struct em28xx, ref)
drivers/media/usb/em28xx/em28xx.h
840
void em28xx_free_device(struct kref *ref);
drivers/media/usb/uvc/uvc_driver.c
1941
struct uvc_device *dev = container_of(kref, struct uvc_device, ref);
drivers/media/usb/uvc/uvc_driver.c
1987
kref_put(&dev->ref, uvc_delete);
drivers/media/usb/uvc/uvc_driver.c
2088
kref_get(&dev->ref);
drivers/media/usb/uvc/uvc_driver.c
2203
kref_init(&dev->ref);
drivers/media/usb/uvc/uvc_driver.c
2355
kref_put(&dev->ref, uvc_delete);
drivers/media/usb/uvc/uvc_driver.c
2374
kref_put(&dev->ref, uvc_delete);
drivers/media/usb/uvc/uvc_queue.c
145
kref_init(&buf->ref);
drivers/media/usb/uvc/uvc_queue.c
354
static void uvc_queue_buffer_complete(struct kref *ref)
drivers/media/usb/uvc/uvc_queue.c
356
struct uvc_buffer *buf = container_of(ref, struct uvc_buffer, ref);
drivers/media/usb/uvc/uvc_queue.c
377
kref_put(&buf->ref, uvc_queue_buffer_complete);
drivers/media/usb/uvc/uvc_video.c
1326
kref_get(&buf->ref);
drivers/media/usb/uvc/uvcvideo.h
328
struct kref ref;
drivers/media/usb/uvc/uvcvideo.h
605
struct kref ref;
drivers/media/v4l2-core/v4l2-ctrls-api.c
1061
struct v4l2_ctrl_ref *ref;
drivers/media/v4l2-core/v4l2-ctrls-api.c
1070
ref = find_ref(hdl, id);
drivers/media/v4l2-core/v4l2-ctrls-api.c
1090
ref = NULL; /* Yes, so there is no next control */
drivers/media/v4l2-core/v4l2-ctrls-api.c
1091
} else if (ref) {
drivers/media/v4l2-core/v4l2-ctrls-api.c
1092
struct v4l2_ctrl_ref *pos = ref;
drivers/media/v4l2-core/v4l2-ctrls-api.c
1098
ref = NULL;
drivers/media/v4l2-core/v4l2-ctrls-api.c
1104
ref = pos;
drivers/media/v4l2-core/v4l2-ctrls-api.c
1122
ref = pos;
drivers/media/v4l2-core/v4l2-ctrls-api.c
1130
if (!ref)
drivers/media/v4l2-core/v4l2-ctrls-api.c
1133
ctrl = ref->ctrl;
drivers/media/v4l2-core/v4l2-ctrls-api.c
241
struct v4l2_ctrl_ref *ref;
drivers/media/v4l2-core/v4l2-ctrls-api.c
26
struct v4l2_ctrl_ref *ref;
drivers/media/v4l2-core/v4l2-ctrls-api.c
266
ref = find_ref_lock(hdl, id);
drivers/media/v4l2-core/v4l2-ctrls-api.c
267
if (!ref) {
drivers/media/v4l2-core/v4l2-ctrls-api.c
271
h->ref = ref;
drivers/media/v4l2-core/v4l2-ctrls-api.c
272
ctrl = ref->ctrl;
drivers/media/v4l2-core/v4l2-ctrls-api.c
290
ref = find_ref_lock(hdl, ctrl->cluster[0]->id);
drivers/media/v4l2-core/v4l2-ctrls-api.c
296
tot_size *= ref->p_req_elems;
drivers/media/v4l2-core/v4l2-ctrls-api.c
335
h->mref = ref;
drivers/media/v4l2-core/v4l2-ctrls-api.c
443
if (helpers[i].ref->ctrl->flags & V4L2_CTRL_FLAG_WRITE_ONLY)
drivers/media/v4l2-core/v4l2-ctrls-api.c
489
struct v4l2_ctrl_ref *ref = helpers[idx].ref;
drivers/media/v4l2-core/v4l2-ctrls-api.c
492
ret = def_to_user(cs->controls + idx, ref->ctrl);
drivers/media/v4l2-core/v4l2-ctrls-api.c
493
else if (is_request && ref->p_req_array_enomem)
drivers/media/v4l2-core/v4l2-ctrls-api.c
495
else if (is_request && ref->p_req_valid)
drivers/media/v4l2-core/v4l2-ctrls-api.c
496
ret = req_to_user(cs->controls + idx, ref);
drivers/media/v4l2-core/v4l2-ctrls-api.c
498
ret = min_to_user(cs->controls + idx, ref->ctrl);
drivers/media/v4l2-core/v4l2-ctrls-api.c
500
ret = max_to_user(cs->controls + idx, ref->ctrl);
drivers/media/v4l2-core/v4l2-ctrls-api.c
502
ret = new_to_user(cs->controls + idx, ref->ctrl);
drivers/media/v4l2-core/v4l2-ctrls-api.c
504
ret = cur_to_user(cs->controls + idx, ref->ctrl);
drivers/media/v4l2-core/v4l2-ctrls-api.c
543
struct v4l2_ctrl *ctrl = helpers[i].ref->ctrl;
drivers/media/v4l2-core/v4l2-ctrls-api.c
663
if (helpers[tmp_idx].ref->ctrl == master)
drivers/media/v4l2-core/v4l2-ctrls-api.c
680
struct v4l2_ctrl *ctrl = helpers[idx].ref->ctrl;
drivers/media/v4l2-core/v4l2-ctrls-api.c
698
struct v4l2_ctrl_ref *ref =
drivers/media/v4l2-core/v4l2-ctrls-api.c
701
new_to_req(ref);
drivers/media/v4l2-core/v4l2-ctrls-api.c
710
helpers[idx].ref->ctrl);
drivers/media/v4l2-core/v4l2-ctrls-api.c
84
struct v4l2_ctrl_ref *ref)
drivers/media/v4l2-core/v4l2-ctrls-api.c
86
return ptr_to_user(c, ref->ctrl, ref->p_req);
drivers/media/v4l2-core/v4l2-ctrls-core.c
1549
static bool req_alloc_array(struct v4l2_ctrl_ref *ref, u32 elems)
drivers/media/v4l2-core/v4l2-ctrls-core.c
1553
if (elems == ref->p_req_array_alloc_elems)
drivers/media/v4l2-core/v4l2-ctrls-core.c
1555
if (ref->ctrl->is_dyn_array &&
drivers/media/v4l2-core/v4l2-ctrls-core.c
1556
elems < ref->p_req_array_alloc_elems)
drivers/media/v4l2-core/v4l2-ctrls-core.c
1559
tmp = kvmalloc(elems * ref->ctrl->elem_size, GFP_KERNEL);
drivers/media/v4l2-core/v4l2-ctrls-core.c
1562
ref->p_req_array_enomem = true;
drivers/media/v4l2-core/v4l2-ctrls-core.c
1565
ref->p_req_array_enomem = false;
drivers/media/v4l2-core/v4l2-ctrls-core.c
1566
kvfree(ref->p_req.p);
drivers/media/v4l2-core/v4l2-ctrls-core.c
1567
ref->p_req.p = tmp;
drivers/media/v4l2-core/v4l2-ctrls-core.c
1568
ref->p_req_array_alloc_elems = elems;
drivers/media/v4l2-core/v4l2-ctrls-core.c
1573
void new_to_req(struct v4l2_ctrl_ref *ref)
drivers/media/v4l2-core/v4l2-ctrls-core.c
1577
if (!ref)
drivers/media/v4l2-core/v4l2-ctrls-core.c
1580
ctrl = ref->ctrl;
drivers/media/v4l2-core/v4l2-ctrls-core.c
1581
if (ctrl->is_array && !req_alloc_array(ref, ctrl->new_elems))
drivers/media/v4l2-core/v4l2-ctrls-core.c
1584
ref->p_req_elems = ctrl->new_elems;
drivers/media/v4l2-core/v4l2-ctrls-core.c
1585
ptr_to_ptr(ctrl, ctrl->p_new, ref->p_req, ref->p_req_elems);
drivers/media/v4l2-core/v4l2-ctrls-core.c
1586
ref->p_req_valid = true;
drivers/media/v4l2-core/v4l2-ctrls-core.c
1590
void cur_to_req(struct v4l2_ctrl_ref *ref)
drivers/media/v4l2-core/v4l2-ctrls-core.c
1594
if (!ref)
drivers/media/v4l2-core/v4l2-ctrls-core.c
1597
ctrl = ref->ctrl;
drivers/media/v4l2-core/v4l2-ctrls-core.c
1598
if (ctrl->is_array && !req_alloc_array(ref, ctrl->elems))
drivers/media/v4l2-core/v4l2-ctrls-core.c
1601
ref->p_req_elems = ctrl->elems;
drivers/media/v4l2-core/v4l2-ctrls-core.c
1602
ptr_to_ptr(ctrl, ctrl->p_cur, ref->p_req, ctrl->elems);
drivers/media/v4l2-core/v4l2-ctrls-core.c
1603
ref->p_req_valid = true;
drivers/media/v4l2-core/v4l2-ctrls-core.c
1607
int req_to_new(struct v4l2_ctrl_ref *ref)
drivers/media/v4l2-core/v4l2-ctrls-core.c
1611
if (!ref)
drivers/media/v4l2-core/v4l2-ctrls-core.c
1614
ctrl = ref->ctrl;
drivers/media/v4l2-core/v4l2-ctrls-core.c
1620
if (!ref->p_req_valid) {
drivers/media/v4l2-core/v4l2-ctrls-core.c
1629
ptr_to_ptr(ctrl, ref->p_req, ctrl->p_new, ctrl->new_elems);
drivers/media/v4l2-core/v4l2-ctrls-core.c
1634
if (WARN_ON(!ref->p_req_array_alloc_elems))
drivers/media/v4l2-core/v4l2-ctrls-core.c
1638
ref->p_req_elems != ctrl->p_array_alloc_elems)
drivers/media/v4l2-core/v4l2-ctrls-core.c
1648
if (ref->p_req_elems > ctrl->p_array_alloc_elems) {
drivers/media/v4l2-core/v4l2-ctrls-core.c
1649
unsigned int sz = ref->p_req_elems * ctrl->elem_size;
drivers/media/v4l2-core/v4l2-ctrls-core.c
1660
ctrl->p_array_alloc_elems = ref->p_req_elems;
drivers/media/v4l2-core/v4l2-ctrls-core.c
1664
ctrl->new_elems = ref->p_req_elems;
drivers/media/v4l2-core/v4l2-ctrls-core.c
1665
ptr_to_ptr(ctrl, ref->p_req, ctrl->p_new, ctrl->new_elems);
drivers/media/v4l2-core/v4l2-ctrls-core.c
1738
struct v4l2_ctrl_ref *ref, *next_ref;
drivers/media/v4l2-core/v4l2-ctrls-core.c
1752
list_for_each_entry_safe(ref, next_ref, &hdl->ctrl_refs, node) {
drivers/media/v4l2-core/v4l2-ctrls-core.c
1753
list_del(&ref->node);
drivers/media/v4l2-core/v4l2-ctrls-core.c
1754
if (ref->p_req_array_alloc_elems)
drivers/media/v4l2-core/v4l2-ctrls-core.c
1755
kvfree(ref->p_req.p);
drivers/media/v4l2-core/v4l2-ctrls-core.c
1756
kfree(ref);
drivers/media/v4l2-core/v4l2-ctrls-core.c
1785
struct v4l2_ctrl_ref *ref;
drivers/media/v4l2-core/v4l2-ctrls-core.c
1788
list_for_each_entry(ref, &hdl->ctrl_refs, node) {
drivers/media/v4l2-core/v4l2-ctrls-core.c
1791
if (V4L2_CTRL_ID2WHICH(ref->ctrl->id) == V4L2_CTRL_CLASS_USER &&
drivers/media/v4l2-core/v4l2-ctrls-core.c
1792
V4L2_CTRL_DRIVER_PRIV(ref->ctrl->id)) {
drivers/media/v4l2-core/v4l2-ctrls-core.c
1793
if (!ref->ctrl->is_int)
drivers/media/v4l2-core/v4l2-ctrls-core.c
1796
return ref;
drivers/media/v4l2-core/v4l2-ctrls-core.c
1806
struct v4l2_ctrl_ref *ref;
drivers/media/v4l2-core/v4l2-ctrls-core.c
1821
ref = hdl->buckets ? hdl->buckets[bucket] : NULL;
drivers/media/v4l2-core/v4l2-ctrls-core.c
1822
while (ref && ref->ctrl->id != id)
drivers/media/v4l2-core/v4l2-ctrls-core.c
1823
ref = ref->next;
drivers/media/v4l2-core/v4l2-ctrls-core.c
1825
if (ref)
drivers/media/v4l2-core/v4l2-ctrls-core.c
1826
hdl->cached = ref; /* cache it! */
drivers/media/v4l2-core/v4l2-ctrls-core.c
1827
return ref;
drivers/media/v4l2-core/v4l2-ctrls-core.c
1833
struct v4l2_ctrl_ref *ref = NULL;
drivers/media/v4l2-core/v4l2-ctrls-core.c
1837
ref = find_ref(hdl, id);
drivers/media/v4l2-core/v4l2-ctrls-core.c
1840
return ref;
drivers/media/v4l2-core/v4l2-ctrls-core.c
1846
struct v4l2_ctrl_ref *ref = find_ref_lock(hdl, id);
drivers/media/v4l2-core/v4l2-ctrls-core.c
1848
return ref ? ref->ctrl : NULL;
drivers/media/v4l2-core/v4l2-ctrls-core.c
1858
struct v4l2_ctrl_ref *ref;
drivers/media/v4l2-core/v4l2-ctrls-core.c
1904
list_for_each_entry(ref, &hdl->ctrl_refs, node) {
drivers/media/v4l2-core/v4l2-ctrls-core.c
1905
if (ref->ctrl->id < id)
drivers/media/v4l2-core/v4l2-ctrls-core.c
1908
if (ref->ctrl->id == id) {
drivers/media/v4l2-core/v4l2-ctrls-core.c
1912
list_add(&new_ref->node, ref->node.prev);
drivers/media/v4l2-core/v4l2-ctrls-core.c
2422
struct v4l2_ctrl_ref *ref;
drivers/media/v4l2-core/v4l2-ctrls-core.c
2431
list_for_each_entry(ref, &add->ctrl_refs, node) {
drivers/media/v4l2-core/v4l2-ctrls-core.c
2432
struct v4l2_ctrl *ctrl = ref->ctrl;
drivers/media/v4l2-core/v4l2-ctrls-priv.h
57
void cur_to_req(struct v4l2_ctrl_ref *ref);
drivers/media/v4l2-core/v4l2-ctrls-priv.h
59
void new_to_req(struct v4l2_ctrl_ref *ref);
drivers/media/v4l2-core/v4l2-ctrls-priv.h
60
int req_to_new(struct v4l2_ctrl_ref *ref);
drivers/media/v4l2-core/v4l2-ctrls-request.c
144
struct v4l2_ctrl_ref *ref = find_ref_lock(hdl, id);
drivers/media/v4l2-core/v4l2-ctrls-request.c
146
return (ref && ref->p_req_valid) ? ref->ctrl : NULL;
drivers/media/v4l2-core/v4l2-ctrls-request.c
329
struct v4l2_ctrl_ref *ref;
drivers/media/v4l2-core/v4l2-ctrls-request.c
361
list_for_each_entry(ref, &hdl->ctrl_refs, node) {
drivers/media/v4l2-core/v4l2-ctrls-request.c
362
struct v4l2_ctrl *ctrl = ref->ctrl;
drivers/media/v4l2-core/v4l2-ctrls-request.c
372
new_to_req(ref);
drivers/media/v4l2-core/v4l2-ctrls-request.c
376
if (ref->p_req_valid)
drivers/media/v4l2-core/v4l2-ctrls-request.c
381
cur_to_req(ref);
drivers/media/v4l2-core/v4l2-ctrls-request.c
400
struct v4l2_ctrl_ref *ref;
drivers/media/v4l2-core/v4l2-ctrls-request.c
423
list_for_each_entry(ref, &hdl->ctrl_refs, node)
drivers/media/v4l2-core/v4l2-ctrls-request.c
424
ref->req_done = false;
drivers/media/v4l2-core/v4l2-ctrls-request.c
426
list_for_each_entry(ref, &hdl->ctrl_refs, node) {
drivers/media/v4l2-core/v4l2-ctrls-request.c
427
struct v4l2_ctrl *ctrl = ref->ctrl;
drivers/media/v4l2-core/v4l2-ctrls-request.c
436
if (ref->req_done || (ctrl->flags & V4L2_CTRL_FLAG_READ_ONLY))
drivers/media/v4l2-core/v4l2-ctrls-request.c
56
struct v4l2_ctrl_ref *ref;
drivers/media/v4l2-core/v4l2-ctrls-request.c
68
list_for_each_entry(ref, &from->ctrl_refs, node) {
drivers/media/v4l2-core/v4l2-ctrls-request.c
69
struct v4l2_ctrl *ctrl = ref->ctrl;
drivers/media/v4l2-core/v4l2-ctrls-request.c
73
if (ref->from_other_dev)
drivers/media/v4l2-core/v4l2-device.c
25
kref_init(&v4l2_dev->ref);
drivers/media/v4l2-core/v4l2-device.c
45
static void v4l2_device_release(struct kref *ref)
drivers/media/v4l2-core/v4l2-device.c
48
container_of(ref, struct v4l2_device, ref);
drivers/media/v4l2-core/v4l2-device.c
56
return kref_put(&v4l2_dev->ref, v4l2_device_release);
drivers/media/v4l2-core/v4l2-h264.c
109
const struct v4l2_h264_reference *ref)
drivers/media/v4l2-core/v4l2-h264.c
111
switch (ref->fields) {
drivers/media/v4l2-core/v4l2-h264.c
113
return min(b->refs[ref->index].top_field_order_cnt,
drivers/media/v4l2-core/v4l2-h264.c
114
b->refs[ref->index].bottom_field_order_cnt);
drivers/media/v4l2-core/v4l2-h264.c
116
return b->refs[ref->index].top_field_order_cnt;
drivers/media/v4l2-core/v4l2-h264.c
118
return b->refs[ref->index].bottom_field_order_cnt;
drivers/memory/tegra/tegra210-emc-core.c
1498
u32 ref = timing->burst_regs[EMC_REFRESH_INDEX];
drivers/memory/tegra/tegra210-emc-core.c
1506
ref = REFRESH_SPEEDUP(ref, 2);
drivers/memory/tegra/tegra210-emc-core.c
1512
ref = REFRESH_SPEEDUP(ref, 4);
drivers/memory/tegra/tegra210-emc-core.c
1522
emc_writel(emc, ref, emc->offsets->burst[EMC_REFRESH_INDEX]);
drivers/misc/fastrpc.c
333
static void fastrpc_free_map(struct kref *ref)
drivers/misc/fastrpc.c
337
map = container_of(ref, struct fastrpc_map, refcount);
drivers/misc/fastrpc.c
481
static void fastrpc_channel_ctx_free(struct kref *ref)
drivers/misc/fastrpc.c
485
cctx = container_of(ref, struct fastrpc_channel_ctx, refcount);
drivers/misc/fastrpc.c
500
static void fastrpc_context_free(struct kref *ref)
drivers/misc/fastrpc.c
507
ctx = container_of(ref, struct fastrpc_invoke_ctx, refcount);
drivers/misc/lkdtm/refcount.c
11
switch (refcount_read(ref)) {
drivers/misc/lkdtm/refcount.c
119
static void check_negative(refcount_t *ref, int start)
drivers/misc/lkdtm/refcount.c
126
if (refcount_read(ref) == start) {
drivers/misc/lkdtm/refcount.c
132
switch (refcount_read(ref)) {
drivers/misc/lkdtm/refcount.c
140
pr_err("Fail: refcount went crazy: %d\n", refcount_read(ref));
drivers/misc/lkdtm/refcount.c
19
pr_err("Fail: refcount wrapped to %d\n", refcount_read(ref));
drivers/misc/lkdtm/refcount.c
200
static void check_from_zero(refcount_t *ref)
drivers/misc/lkdtm/refcount.c
202
switch (refcount_read(ref)) {
drivers/misc/lkdtm/refcount.c
214
refcount_read(ref));
drivers/misc/lkdtm/refcount.c
267
static void check_saturated(refcount_t *ref)
drivers/misc/lkdtm/refcount.c
269
switch (refcount_read(ref)) {
drivers/misc/lkdtm/refcount.c
277
pr_err("Fail: refcount went crazy: %d\n", refcount_read(ref));
drivers/misc/lkdtm/refcount.c
84
static void check_zero(refcount_t *ref)
drivers/misc/lkdtm/refcount.c
86
switch (refcount_read(ref)) {
drivers/misc/lkdtm/refcount.c
9
static void overflow_check(refcount_t *ref)
drivers/misc/lkdtm/refcount.c
97
pr_err("Fail: refcount went crazy: %d\n", refcount_read(ref));
drivers/misc/mei/client.c
55
static void mei_me_cl_release(struct kref *ref)
drivers/misc/mei/client.c
58
container_of(ref, struct mei_me_client, refcnt);
drivers/misc/ocxl/link.c
394
kref_init(&link->ref);
drivers/misc/ocxl/link.c
452
kref_get(&link->ref);
drivers/misc/ocxl/link.c
469
static void release_xsl(struct kref *ref)
drivers/misc/ocxl/link.c
471
struct ocxl_link *link = container_of(ref, struct ocxl_link, ref);
drivers/misc/ocxl/link.c
489
kref_put(&link->ref, release_xsl);
drivers/misc/ocxl/link.c
88
struct kref ref;
drivers/mmc/core/block.c
212
static void mmc_blk_kref_release(struct kref *ref)
drivers/mmc/core/block.c
214
struct mmc_blk_data *md = container_of(ref, struct mmc_blk_data, kref);
drivers/mtd/mtd_blkdevs.c
190
kref_get(&dev->ref);
drivers/mtd/mtd_blkdevs.c
30
container_of(kref, struct mtd_blktrans_dev, ref);
drivers/mtd/mtd_blkdevs.c
322
kref_init(&new->ref);
drivers/mtd/mtd_blkdevs.c
41
kref_put(&dev->ref, blktrans_dev_release);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
11659
static void bnxt_irq_affinity_release(struct kref *ref)
drivers/net/ethernet/broadcom/bnxt/bnxt.c
11662
container_of(ref, struct irq_affinity_notify, kref);
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls.h
526
static inline void chtls_sock_free(struct kref *ref)
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls.h
528
struct chtls_sock *csk = container_of(ref, struct chtls_sock,
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_cm.c
85
static void chtls_sock_release(struct kref *ref)
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_cm.c
88
container_of(ref, struct chtls_sock, kref);
drivers/net/ethernet/engleder/tsnep_tc.c
203
static u64 tsnep_insert_gcl_operation(struct tsnep_gcl *gcl, int ref,
drivers/net/ethernet/engleder/tsnep_tc.c
208
properties = gcl->operation[ref].properties & TSNEP_GCL_MASK;
drivers/net/ethernet/engleder/tsnep_tc.c
216
return tsnep_set_gcl_change(gcl, ref, change, true);
drivers/net/ethernet/engleder/tsnep_tc.c
221
int ref = gcl->count - 1;
drivers/net/ethernet/engleder/tsnep_tc.c
222
u32 interval = gcl->operation[ref].interval + extension;
drivers/net/ethernet/engleder/tsnep_tc.c
224
start -= gcl->operation[ref].interval;
drivers/net/ethernet/engleder/tsnep_tc.c
226
return tsnep_insert_gcl_operation(gcl, ref, start, interval);
drivers/net/ethernet/fungible/funeth/funeth_main.c
242
static void fun_irq_aff_release(struct kref __always_unused *ref)
drivers/net/ethernet/intel/i40e/i40e_main.c
4117
static void i40e_irq_affinity_release(struct kref *ref) {}
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
1553
hw->blk[blk].masks.masks[i].ref = 0;
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
1557
hw->blk[blk].masks.masks[i].ref++;
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
1588
if (hw->blk[blk].masks.masks[mask_idx].ref > 1) {
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
1589
hw->blk[blk].masks.masks[mask_idx].ref--;
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
4073
u16 ref;
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
4088
status = ice_vsig_get_ref(hw, blk, vsig, &ref);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
4091
only_vsi = (ref == 1);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
4282
u16 ref;
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
4286
status = ice_vsig_get_ref(hw, blk, vsig, &ref);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
4289
only_vsi = (ref == 1);
drivers/net/ethernet/intel/ice/ice_flex_type.h
298
u16 ref; /* reference count */
drivers/net/ethernet/intel/ice/ice_vf_lib.c
55
static void ice_release_vf(struct kref *ref)
drivers/net/ethernet/intel/ice/ice_vf_lib.c
57
struct ice_vf *vf = container_of(ref, struct ice_vf, refcnt);
drivers/net/ethernet/mellanox/mlx5/core/en/tc/ct_fs_hmfs.c
125
if (refcount_inc_not_zero(&hmfs_matcher->ref))
drivers/net/ethernet/mellanox/mlx5/core/en/tc/ct_fs_hmfs.c
131
if (refcount_inc_not_zero(&hmfs_matcher->ref))
drivers/net/ethernet/mellanox/mlx5/core/en/tc/ct_fs_hmfs.c
147
refcount_set(&hmfs_matcher->ref, 1);
drivers/net/ethernet/mellanox/mlx5/core/en/tc/ct_fs_hmfs.c
15
refcount_t ref;
drivers/net/ethernet/mellanox/mlx5/core/en/tc/ct_fs_hmfs.c
159
if (!refcount_dec_and_mutex_lock(&hmfs_matcher->ref, &fs_hmfs->lock))
drivers/net/ethernet/mellanox/mlx5/core/en/tc/ct_fs_smfs.c
124
if (refcount_inc_not_zero(&smfs_matcher->ref))
drivers/net/ethernet/mellanox/mlx5/core/en/tc/ct_fs_smfs.c
132
if (refcount_inc_not_zero(&smfs_matcher->ref))
drivers/net/ethernet/mellanox/mlx5/core/en/tc/ct_fs_smfs.c
161
refcount_set(&smfs_matcher->ref, 1);
drivers/net/ethernet/mellanox/mlx5/core/en/tc/ct_fs_smfs.c
173
if (!refcount_dec_and_mutex_lock(&smfs_matcher->ref, &fs_smfs->lock))
drivers/net/ethernet/mellanox/mlx5/core/en/tc/ct_fs_smfs.c
21
refcount_t ref;
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
232
static void mlx5_ldev_free(struct kref *ref)
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
234
struct mlx5_lag *ldev = container_of(ref, struct mlx5_lag, ref);
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
262
kref_put(&ldev->ref, mlx5_ldev_free);
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
267
kref_get(&ldev->ref);
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.c
285
kref_init(&ldev->ref);
drivers/net/ethernet/mellanox/mlx5/core/lag/lag.h
66
struct kref ref;
drivers/net/ethernet/mellanox/mlx5/core/lib/devcom.c
101
struct mlx5_devcom_dev *devc = container_of(ref, struct mlx5_devcom_dev, ref);
drivers/net/ethernet/mellanox/mlx5/core/lib/devcom.c
114
kref_put(&devc->ref, mlx5_devcom_dev_release);
drivers/net/ethernet/mellanox/mlx5/core/lib/devcom.c
136
kref_init(&comp->ref);
drivers/net/ethernet/mellanox/mlx5/core/lib/devcom.c
143
mlx5_devcom_comp_release(struct kref *ref)
drivers/net/ethernet/mellanox/mlx5/core/lib/devcom.c
145
struct mlx5_devcom_comp *comp = container_of(ref, struct mlx5_devcom_comp, ref);
drivers/net/ethernet/mellanox/mlx5/core/lib/devcom.c
165
kref_get(&devc->ref);
drivers/net/ethernet/mellanox/mlx5/core/lib/devcom.c
186
kref_put(&devcom->devc->ref, mlx5_devcom_dev_release);
drivers/net/ethernet/mellanox/mlx5/core/lib/devcom.c
188
kref_put(&comp->ref, mlx5_devcom_comp_release);
drivers/net/ethernet/mellanox/mlx5/core/lib/devcom.c
223
kref_get(&comp->ref);
drivers/net/ethernet/mellanox/mlx5/core/lib/devcom.c
23
struct kref ref;
drivers/net/ethernet/mellanox/mlx5/core/lib/devcom.c
265
kref_put(&comp->ref, mlx5_devcom_comp_release);
drivers/net/ethernet/mellanox/mlx5/core/lib/devcom.c
286
return kref_read(&comp->ref);
drivers/net/ethernet/mellanox/mlx5/core/lib/devcom.c
38
struct kref ref;
drivers/net/ethernet/mellanox/mlx5/core/lib/devcom.c
72
kref_init(&devc->ref);
drivers/net/ethernet/mellanox/mlx5/core/lib/devcom.c
99
mlx5_devcom_dev_release(struct kref *ref)
drivers/net/ethernet/mellanox/mlx5/core/lib/fs_chains.c
368
chain_s->ref++;
drivers/net/ethernet/mellanox/mlx5/core/lib/fs_chains.c
45
int ref;
drivers/net/ethernet/mellanox/mlx5/core/lib/fs_chains.c
461
if (--chain->ref == 0)
drivers/net/ethernet/mellanox/mlx5/core/lib/fs_chains.c
642
++prio_s->ref;
drivers/net/ethernet/mellanox/mlx5/core/lib/fs_chains.c
66
int ref;
drivers/net/ethernet/mellanox/mlx5/core/lib/fs_chains.c
672
if (--prio_s->ref == 0)
drivers/net/ethernet/microchip/lan966x/lan966x_ptp.c
629
u64 ref;
drivers/net/ethernet/microchip/lan966x/lan966x_ptp.c
645
ref = LAN966X_1PPM_FORMAT * (scaled_ppm >> 16);
drivers/net/ethernet/microchip/lan966x/lan966x_ptp.c
646
ref += (LAN966X_1PPM_FORMAT * (0xffff & scaled_ppm)) >> 16;
drivers/net/ethernet/microchip/lan966x/lan966x_ptp.c
647
tod_inc = neg_adj ? tod_inc - ref : tod_inc + ref;
drivers/net/ethernet/microchip/sparx5/sparx5_ptp.c
403
u64 ref;
drivers/net/ethernet/microchip/sparx5/sparx5_ptp.c
419
ref = sparx5_ptp_get_1ppm(sparx5) * (scaled_ppm >> 16);
drivers/net/ethernet/microchip/sparx5/sparx5_ptp.c
420
ref += (sparx5_ptp_get_1ppm(sparx5) * (0xffff & scaled_ppm)) >> 16;
drivers/net/ethernet/microchip/sparx5/sparx5_ptp.c
421
tod_inc = neg_adj ? tod_inc - ref : tod_inc + ref;
drivers/net/ethernet/pensando/ionic/ionic_lif.c
286
static void ionic_irq_aff_release(struct kref __always_unused *ref)
drivers/net/ethernet/qlogic/qede/qede_rdma.c
64
static void qede_rdma_complete_event(struct kref *ref)
drivers/net/ethernet/qlogic/qede/qede_rdma.c
67
container_of(ref, struct qede_rdma_dev, refcnt);
drivers/net/ethernet/rocker/rocker_ofdpa.c
1498
int ref = 0;
drivers/net/ethernet/rocker/rocker_ofdpa.c
1526
ref++;
drivers/net/ethernet/rocker/rocker_ofdpa.c
1529
if ((!adding || ref != 1) && (adding || ref != 0))
drivers/net/ethernet/sfc/tc.c
141
if (!refcount_inc_not_zero(&old->ref))
drivers/net/ethernet/sfc/tc.c
154
refcount_set(&ped->ref, 1);
drivers/net/ethernet/sfc/tc.c
166
if (!refcount_dec_and_test(&ped->ref))
drivers/net/ethernet/sfc/tc.c
2929
WARN_ON(refcount_read(&encap->ref));
drivers/net/ethernet/sfc/tc.c
2938
WARN_ON(refcount_read(&rid->ref));
drivers/net/ethernet/sfc/tc.c
2965
WARN_ON(refcount_read(&ped->ref));
drivers/net/ethernet/sfc/tc.c
469
if (!refcount_dec_and_test(&encap->ref))
drivers/net/ethernet/sfc/tc.c
664
if (!refcount_inc_not_zero(&old->ref))
drivers/net/ethernet/sfc/tc.c
676
refcount_set(&encap->ref, 1);
drivers/net/ethernet/sfc/tc.c
714
if (!refcount_inc_not_zero(&old->ref))
drivers/net/ethernet/sfc/tc.c
728
refcount_set(&rid->ref, 1);
drivers/net/ethernet/sfc/tc.c
735
if (!refcount_dec_and_test(&rid->ref))
drivers/net/ethernet/sfc/tc.h
165
refcount_t ref;
drivers/net/ethernet/sfc/tc.h
175
refcount_t ref;
drivers/net/ethernet/sfc/tc.h
34
refcount_t ref;
drivers/net/ethernet/sfc/tc_conntrack.c
577
if (!refcount_inc_not_zero(&old->ref))
drivers/net/ethernet/sfc/tc_conntrack.c
594
refcount_set(&ct_zone->ref, 1);
drivers/net/ethernet/sfc/tc_conntrack.c
608
if (!refcount_dec_and_test(&ct_zone->ref))
drivers/net/ethernet/sfc/tc_conntrack.h
22
refcount_t ref;
drivers/net/ethernet/sfc/tc_counters.c
214
if (!refcount_dec_and_test(&ctr->ref))
drivers/net/ethernet/sfc/tc_counters.c
241
if (!refcount_inc_not_zero(&old->ref))
drivers/net/ethernet/sfc/tc_counters.c
255
refcount_set(&ctr->ref, 1);
drivers/net/ethernet/sfc/tc_counters.c
51
WARN_ON(refcount_read(&ctr->ref));
drivers/net/ethernet/sfc/tc_counters.h
43
refcount_t ref;
drivers/net/ethernet/sfc/tc_encap_actions.c
139
if (!refcount_inc_not_zero(&old->ref))
drivers/net/ethernet/sfc/tc_encap_actions.c
195
refcount_set(&neigh->ref, 1);
drivers/net/ethernet/sfc/tc_encap_actions.c
246
if (!refcount_dec_and_test(&neigh->ref))
drivers/net/ethernet/sfc/tc_encap_actions.c
37
WARN_ON(refcount_read(&enc->ref));
drivers/net/ethernet/sfc/tc_encap_actions.c
45
WARN_ON(refcount_read(&neigh->ref));
drivers/net/ethernet/sfc/tc_encap_actions.c
516
if (refcount_dec_and_test(&neigh->ref))
drivers/net/ethernet/sfc/tc_encap_actions.c
574
if (refcount_inc_not_zero(&neigh->ref)) {
drivers/net/ethernet/sfc/tc_encap_actions.c
578
if (refcount_dec_and_test(&neigh->ref))
drivers/net/ethernet/sfc/tc_encap_actions.c
649
if (!refcount_inc_not_zero(&old->ref))
drivers/net/ethernet/sfc/tc_encap_actions.c
682
refcount_set(&encap->ref, 1);
drivers/net/ethernet/sfc/tc_encap_actions.c
696
if (!refcount_dec_and_test(&encap->ref))
drivers/net/ethernet/sfc/tc_encap_actions.h
59
refcount_t ref;
drivers/net/ethernet/sfc/tc_encap_actions.h
82
refcount_t ref;
drivers/net/phy/sfp-bus.c
612
struct fwnode_reference_args ref;
drivers/net/phy/sfp-bus.c
617
0, 0, &ref);
drivers/net/phy/sfp-bus.c
623
if (!fwnode_device_is_available(ref.fwnode)) {
drivers/net/phy/sfp-bus.c
624
fwnode_handle_put(ref.fwnode);
drivers/net/phy/sfp-bus.c
628
bus = sfp_bus_get(ref.fwnode);
drivers/net/phy/sfp-bus.c
629
fwnode_handle_put(ref.fwnode);
drivers/net/usb/hso.c
1271
kref_get(&serial->parent->ref);
drivers/net/usb/hso.c
1372
kref_put(&serial->parent->ref, hso_serial_ref_free);
drivers/net/usb/hso.c
2323
kref_init(&hso_dev->ref);
drivers/net/usb/hso.c
258
struct kref ref;
drivers/net/usb/hso.c
3089
static void hso_serial_ref_free(struct kref *ref)
drivers/net/usb/hso.c
3091
struct hso_device *hso_dev = container_of(ref, struct hso_device, ref);
drivers/net/usb/hso.c
310
static void hso_serial_ref_free(struct kref *ref);
drivers/net/usb/hso.c
3112
kref_put(&serial->parent->ref, hso_serial_ref_free);
drivers/net/virtio_net.c
1039
dma->ref = 1;
drivers/net/virtio_net.c
1045
++dma->ref;
drivers/net/virtio_net.c
296
u32 ref;
drivers/net/virtio_net.c
945
--dma->ref;
drivers/net/virtio_net.c
955
if (dma->ref)
drivers/net/wireless/ath/carl9170/carl9170.h
500
struct kref ref;
drivers/net/wireless/ath/carl9170/tx.c
1076
kref_init(&arinfo->ref);
drivers/net/wireless/ath/carl9170/tx.c
262
static void carl9170_tx_release(struct kref *ref)
drivers/net/wireless/ath/carl9170/tx.c
269
arinfo = container_of(ref, struct carl9170_tx_info, ref);
drivers/net/wireless/ath/carl9170/tx.c
335
kref_get(&arinfo->ref);
drivers/net/wireless/ath/carl9170/tx.c
343
return kref_put(&arinfo->ref, carl9170_tx_release);
drivers/net/wireless/ath/wcn36xx/hal.h
3473
u8 ref;
drivers/net/wireless/ath/wcn36xx/smd.c
3161
.ref = ref_val, \
drivers/net/wireless/broadcom/b43/phy_lp.c
505
u32 crystalfreq, tmp, ref;
drivers/net/wireless/broadcom/b43/phy_lp.c
569
ref = (1000 * lpphy->pdiv + 2 * crystalfreq) / (2000 * lpphy->pdiv);
drivers/net/wireless/broadcom/b43/phy_lp.c
570
ref &= 0xFFFF;
drivers/net/wireless/broadcom/b43/phy_lp.c
572
if (ref < freqdata_tab[i].freq) {
drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c
113
mvm->phy_ctxts[i].ref = 0;
drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c
1434
if (!mvm->phy_ctxts[i].ref)
drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c
4737
if (!phy_ctxt->ref || mvmvif->deflink.phy_ctxt == phy_ctxt)
drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c
4942
if (WARN_ONCE((phy_ctxt->ref > 1) &&
drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c
4948
phy_ctxt->ref, changed))
drivers/net/wireless/intel/iwlwifi/mvm/mvm.h
97
u32 ref;
drivers/net/wireless/intel/iwlwifi/mvm/phy-ctxt.c
268
ctxt->ref);
drivers/net/wireless/intel/iwlwifi/mvm/phy-ctxt.c
282
ctxt->ref++;
drivers/net/wireless/intel/iwlwifi/mvm/phy-ctxt.c
298
WARN_ON(!ctxt->ref);
drivers/net/wireless/intel/iwlwifi/mvm/phy-ctxt.c
299
ctxt->ref++;
drivers/net/wireless/intel/iwlwifi/mvm/phy-ctxt.c
316
if (WARN_ON_ONCE(!ctxt->ref))
drivers/net/wireless/intel/iwlwifi/mvm/phy-ctxt.c
360
ctxt->ref--;
drivers/net/wireless/intel/iwlwifi/mvm/phy-ctxt.c
362
if (ctxt->ref)
drivers/net/wireless/intel/iwlwifi/mvm/rx.c
766
if (!mvm->phy_ctxts[i].ref)
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
2131
union acpi_object *p, *ref;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
2148
ref = &p->package.elements[0];
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
2149
if (ref->type != ACPI_TYPE_LOCAL_REFERENCE) {
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
2154
status = acpi_evaluate_object(ref->reference.handle,
drivers/net/wireless/mediatek/mt76/mt7603/mac.c
731
struct ieee80211_tx_rate *ref;
drivers/net/wireless/mediatek/mt76/mt7603/mac.c
754
ref = &sta->rateset[rateset].probe_rate;
drivers/net/wireless/mediatek/mt76/mt7603/mac.c
757
ref = &sta->rateset[rateset].rates[0];
drivers/net/wireless/mediatek/mt76/mt7603/mac.c
769
if ((ref->flags ^ rates[i].flags) & IEEE80211_TX_RC_SHORT_GI)
drivers/net/wireless/mediatek/mt76/mt7615/mac.c
965
struct ieee80211_tx_rate *ref;
drivers/net/wireless/mediatek/mt76/mt7615/mac.c
979
ref = &sta->rateset[rateset].probe_rate;
drivers/net/wireless/mediatek/mt76/mt7615/mac.c
982
ref = &sta->rateset[rateset].rates[0];
drivers/net/wireless/mediatek/mt76/mt7615/mac.c
994
if ((ref->flags ^ rates[i].flags) & IEEE80211_TX_RC_SHORT_GI)
drivers/net/wireless/realtek/rtw89/chan.c
1095
struct rtw89_mcc_role *ref = &mcc->role_ref;
drivers/net/wireless/realtek/rtw89/chan.c
1115
if (RTW89_MCC_REQ_COURTESY(pattern, aux) && rtw89_mcc_can_courtesy(ref, aux)) {
drivers/net/wireless/realtek/rtw89/chan.c
1116
crtz = &pattern->courtesy.ref;
drivers/net/wireless/realtek/rtw89/chan.c
1117
ref->crtz = crtz;
drivers/net/wireless/realtek/rtw89/chan.c
1126
ref->crtz = NULL;
drivers/net/wireless/realtek/rtw89/chan.c
1129
if (RTW89_MCC_REQ_COURTESY(pattern, ref) && ref->is_gc)
drivers/net/wireless/realtek/rtw89/chan.c
1130
ref->ignore_bcn = true;
drivers/net/wireless/realtek/rtw89/chan.c
1132
ref->ignore_bcn = false;
drivers/net/wireless/realtek/rtw89/chan.c
1134
if (RTW89_MCC_REQ_COURTESY(pattern, ref) && rtw89_mcc_can_courtesy(aux, ref)) {
drivers/net/wireless/realtek/rtw89/chan.c
1138
crtz->macid_tgt = ref->rtwvif_link->mac_id;
drivers/net/wireless/realtek/rtw89/chan.c
1168
struct rtw89_mcc_role *ref = &mcc->role_ref;
drivers/net/wireless/realtek/rtw89/chan.c
1185
max_bcn_ofst = ref->duration + aux->duration;
drivers/net/wireless/realtek/rtw89/chan.c
1186
if (ref->limit.enable)
drivers/net/wireless/realtek/rtw89/chan.c
1188
ref->limit.max_toa + aux->duration);
drivers/net/wireless/realtek/rtw89/chan.c
1191
ref->duration + aux->limit.max_tob);
drivers/net/wireless/realtek/rtw89/chan.c
1204
upper = min_t(s16, ref->duration, res);
drivers/net/wireless/realtek/rtw89/chan.c
1205
lower = max_t(s16, 0, ref->duration - (mcc_intvl - bcn_ofst));
drivers/net/wireless/realtek/rtw89/chan.c
1207
if (ref->limit.enable) {
drivers/net/wireless/realtek/rtw89/chan.c
1208
upper = min_t(s16, upper, ref->limit.max_toa);
drivers/net/wireless/realtek/rtw89/chan.c
1209
lower = max_t(s16, lower, ref->duration - ref->limit.max_tob);
drivers/net/wireless/realtek/rtw89/chan.c
1221
ptrn->tob_ref = ref->duration - ptrn->toa_ref;
drivers/net/wireless/realtek/rtw89/chan.c
1233
struct rtw89_mcc_role *ref = &mcc->role_ref;
drivers/net/wireless/realtek/rtw89/chan.c
1253
if (ref->duration < min_tob + min_toa) {
drivers/net/wireless/realtek/rtw89/chan.c
1272
upper_toa_ref = min_t(s16, min_toa + res, ref->duration - min_tob);
drivers/net/wireless/realtek/rtw89/chan.c
1277
if (ref->limit.enable) {
drivers/net/wireless/realtek/rtw89/chan.c
1278
if (min_tob > ref->limit.max_tob || min_toa > ref->limit.max_toa) {
drivers/net/wireless/realtek/rtw89/chan.c
1284
upper_toa_ref = min_t(s16, upper_toa_ref, ref->limit.max_toa);
drivers/net/wireless/realtek/rtw89/chan.c
1286
ref->duration - ref->limit.max_tob);
drivers/net/wireless/realtek/rtw89/chan.c
1310
ptrn->tob_ref = ref->duration - ptrn->toa_ref;
drivers/net/wireless/realtek/rtw89/chan.c
1321
struct rtw89_mcc_role *ref = &mcc->role_ref;
drivers/net/wireless/realtek/rtw89/chan.c
1328
if (ref->limit.enable) {
drivers/net/wireless/realtek/rtw89/chan.c
1329
ref_tob = ref->limit.max_tob;
drivers/net/wireless/realtek/rtw89/chan.c
1330
ref_toa = ref->limit.max_toa;
drivers/net/wireless/realtek/rtw89/chan.c
1332
ref_tob = ref->duration / 2;
drivers/net/wireless/realtek/rtw89/chan.c
1333
ref_toa = ref->duration / 2;
drivers/net/wireless/realtek/rtw89/chan.c
1338
ptrn->tob_ref = ref->duration - ptrn->toa_ref;
drivers/net/wireless/realtek/rtw89/chan.c
1341
ptrn->toa_ref = ref->duration - ptrn->tob_ref;
drivers/net/wireless/realtek/rtw89/chan.c
1353
struct rtw89_mcc_role *ref = &mcc->role_ref;
drivers/net/wireless/realtek/rtw89/chan.c
1377
ptrn->tob_ref = ref->duration - ptrn->toa_ref;
drivers/net/wireless/realtek/rtw89/chan.c
1385
struct rtw89_mcc_role *ref = &mcc->role_ref;
drivers/net/wireless/realtek/rtw89/chan.c
1409
if (ref->is_go || ref->is_gc)
drivers/net/wireless/realtek/rtw89/chan.c
1422
struct rtw89_mcc_role *ref = &mcc->role_ref;
drivers/net/wireless/realtek/rtw89/chan.c
1429
if (ref->limit.enable && aux->limit.enable) {
drivers/net/wireless/realtek/rtw89/chan.c
1435
if (ref->limit.enable &&
drivers/net/wireless/realtek/rtw89/chan.c
1436
ref->duration > ref->limit.max_tob + ref->limit.max_toa) {
drivers/net/wireless/realtek/rtw89/chan.c
1486
struct rtw89_mcc_role *ref = &mcc->role_ref;
drivers/net/wireless/realtek/rtw89/chan.c
1494
tmp.tob_ref = ref->duration / 2;
drivers/net/wireless/realtek/rtw89/chan.c
1495
tmp.toa_ref = ref->duration - tmp.tob_ref;
drivers/net/wireless/realtek/rtw89/chan.c
1528
struct rtw89_mcc_role *ref = &mcc->role_ref;
drivers/net/wireless/realtek/rtw89/chan.c
1534
if (ref->duration < RTW89_MCC_MIN_STA_DURATION) {
drivers/net/wireless/realtek/rtw89/chan.c
1541
dur_ref = ref->duration;
drivers/net/wireless/realtek/rtw89/chan.c
1545
if (ref->limit.enable) {
drivers/net/wireless/realtek/rtw89/chan.c
1546
dur_ref = min(dur_ref, ref->limit.max_dur);
drivers/net/wireless/realtek/rtw89/chan.c
1555
ref->duration, aux->duration, dur_ref, dur_aux);
drivers/net/wireless/realtek/rtw89/chan.c
1557
ref->duration = dur_ref;
drivers/net/wireless/realtek/rtw89/chan.c
1694
struct rtw89_mcc_role *ref = &mcc->role_ref;
drivers/net/wireless/realtek/rtw89/chan.c
1701
if (ref->is_2ghz && aux->is_2ghz) {
drivers/net/wireless/realtek/rtw89/chan.c
1709
if (!ref->is_2ghz && !aux->is_2ghz) {
drivers/net/wireless/realtek/rtw89/chan.c
1718
if (ref->is_2ghz)
drivers/net/wireless/realtek/rtw89/chan.c
1719
rtw89_mcc_mod_duration_diff_band_with_bt(rtwdev, ref, aux);
drivers/net/wireless/realtek/rtw89/chan.c
1721
rtw89_mcc_mod_duration_diff_band_with_bt(rtwdev, aux, ref);
drivers/net/wireless/realtek/rtw89/chan.c
1758
struct rtw89_mcc_role *ref = &mcc->role_ref;
drivers/net/wireless/realtek/rtw89/chan.c
1761
u32 bcn_intvl_ref_us = ieee80211_tu_to_usec(ref->beacon_interval);
drivers/net/wireless/realtek/rtw89/chan.c
1778
if (ref->is_go || aux->is_go)
drivers/net/wireless/realtek/rtw89/chan.c
1783
cur_tbtt_ofst = rtw89_mcc_get_tbtt_ofst(rtwdev, ref, tsf);
drivers/net/wireless/realtek/rtw89/chan.c
1798
struct rtw89_mcc_role *ref = &mcc->role_ref;
drivers/net/wireless/realtek/rtw89/chan.c
1809
if (ref->is_go) {
drivers/net/wireless/realtek/rtw89/chan.c
1810
config->mcc_interval = ref->beacon_interval;
drivers/net/wireless/realtek/rtw89/chan.c
1811
rtw89_mcc_set_duration_go_sta(rtwdev, ref, aux);
drivers/net/wireless/realtek/rtw89/chan.c
1814
rtw89_mcc_set_duration_go_sta(rtwdev, aux, ref);
drivers/net/wireless/realtek/rtw89/chan.c
1819
config->mcc_interval = ref->beacon_interval;
drivers/net/wireless/realtek/rtw89/chan.c
1896
struct rtw89_mcc_role *ref = &mcc->role_ref;
drivers/net/wireless/realtek/rtw89/chan.c
1910
slot_arg->roles[0].is_master = role == ref;
drivers/net/wireless/realtek/rtw89/chan.c
1960
struct rtw89_mcc_role *ref = &mcc->role_ref;
drivers/net/wireless/realtek/rtw89/chan.c
1978
ret = __mcc_fw_add_role(rtwdev, ref);
drivers/net/wireless/realtek/rtw89/chan.c
1991
ret = __mcc_fw_add_role(rtwdev, ref);
drivers/net/wireless/realtek/rtw89/chan.c
2004
ret = __mcc_fw_add_role(rtwdev, ref);
drivers/net/wireless/realtek/rtw89/chan.c
2028
req.macid = ref->rtwvif_link->mac_id;
drivers/net/wireless/realtek/rtw89/chan.c
2046
struct rtw89_mcc_role *ref = &mcc->role_ref;
drivers/net/wireless/realtek/rtw89/chan.c
2050
if (ref->crtz) {
drivers/net/wireless/realtek/rtw89/chan.c
2051
slot_arg_src = &arg->slots[ref->slot_idx];
drivers/net/wireless/realtek/rtw89/chan.c
2054
slot_arg_src->courtesy_period = ref->crtz->slot_num;
drivers/net/wireless/realtek/rtw89/chan.c
2061
slot_arg_src->courtesy_target = ref->slot_idx;
drivers/net/wireless/realtek/rtw89/chan.c
2070
struct rtw89_mcc_role *ref = &mcc->role_ref;
drivers/net/wireless/realtek/rtw89/chan.c
2093
__mrc_fw_add_role(rtwdev, ref, &add_arg, 0);
drivers/net/wireless/realtek/rtw89/chan.c
2101
__mrc_fw_add_role(rtwdev, ref, &add_arg, 0);
drivers/net/wireless/realtek/rtw89/chan.c
2109
__mrc_fw_add_role(rtwdev, ref, &add_arg, 0);
drivers/net/wireless/realtek/rtw89/chan.c
2168
struct rtw89_mcc_role *ref = &mcc->role_ref;
drivers/net/wireless/realtek/rtw89/chan.c
2173
.start_macid = ref->rtwvif_link->mac_id,
drivers/net/wireless/realtek/rtw89/chan.c
2174
.macid_x = ref->rtwvif_link->mac_id,
drivers/net/wireless/realtek/rtw89/chan.c
2176
.duration_x = ref->duration,
drivers/net/wireless/realtek/rtw89/chan.c
2209
struct rtw89_mcc_role *ref = &mcc->role_ref;
drivers/net/wireless/realtek/rtw89/chan.c
2216
.slot_idx = ref->slot_idx,
drivers/net/wireless/realtek/rtw89/chan.c
2217
.duration = ref->duration,
drivers/net/wireless/realtek/rtw89/chan.c
2261
struct rtw89_mcc_role *ref = &mcc->role_ref;
drivers/net/wireless/realtek/rtw89/chan.c
2273
if (ref->is_go) {
drivers/net/wireless/realtek/rtw89/chan.c
2275
rtwvif_go = ref->rtwvif_link;
drivers/net/wireless/realtek/rtw89/chan.c
2276
start_time += ieee80211_tu_to_usec(ref->duration);
drivers/net/wireless/realtek/rtw89/chan.c
2277
duration = config->mcc_interval - ref->duration;
drivers/net/wireless/realtek/rtw89/chan.c
2309
struct rtw89_mcc_role *ref = &mcc->role_ref;
drivers/net/wireless/realtek/rtw89/chan.c
2315
if (ref->is_go)
drivers/net/wireless/realtek/rtw89/chan.c
2316
rtw89_fw_h2c_tsf32_toggle(rtwdev, ref->rtwvif_link, true);
drivers/net/wireless/realtek/rtw89/chan.c
2326
struct rtw89_mcc_role *ref = &mcc->role_ref;
drivers/net/wireless/realtek/rtw89/chan.c
2332
if (ref->is_go)
drivers/net/wireless/realtek/rtw89/chan.c
2333
rtw89_fw_h2c_tsf32_toggle(rtwdev, ref->rtwvif_link, false);
drivers/net/wireless/realtek/rtw89/chan.c
2355
struct rtw89_mcc_role *ref = &mcc->role_ref;
drivers/net/wireless/realtek/rtw89/chan.c
2370
if (ref->is_go || aux->is_go)
drivers/net/wireless/realtek/rtw89/chan.c
2383
if (rtw89_mcc_ignore_bcn(rtwdev, ref) || aux->ignore_bcn) {
drivers/net/wireless/realtek/rtw89/chan.c
2385
} else if (rtw89_mcc_ignore_bcn(rtwdev, aux) || ref->ignore_bcn) {
drivers/net/wireless/realtek/rtw89/chan.c
2386
rtw89_fw_h2c_set_bcn_fltr_cfg(rtwdev, ref->rtwvif_link, false);
drivers/net/wireless/realtek/rtw89/chan.c
2388
rtw89_fw_h2c_set_bcn_fltr_cfg(rtwdev, ref->rtwvif_link, true);
drivers/net/wireless/realtek/rtw89/chan.c
2455
struct rtw89_mcc_role *ref = &mcc->role_ref;
drivers/net/wireless/realtek/rtw89/chan.c
2469
if (rsn_scan && ref->is_go)
drivers/net/wireless/realtek/rtw89/chan.c
2470
sel.hint.target = ref->rtwvif_link;
drivers/net/wireless/realtek/rtw89/chan.c
2477
rtw89_mcc_stop_sel_fill(&sel, ref);
drivers/net/wireless/realtek/rtw89/chan.c
2514
struct rtw89_mcc_role *ref = &mcc->role_ref;
drivers/net/wireless/realtek/rtw89/chan.c
2530
if (old_ref_ignore_bcn != ref->ignore_bcn)
drivers/net/wireless/realtek/rtw89/chan.c
2531
rtw89_fw_h2c_set_bcn_fltr_cfg(rtwdev, ref->rtwvif_link, !ref->ignore_bcn);
drivers/net/wireless/realtek/rtw89/chan.c
2700
struct rtw89_mcc_role *ref = &mcc->role_ref;
drivers/net/wireless/realtek/rtw89/chan.c
2706
if (rtw89_mcc_ignore_bcn(rtwdev, ref) || aux->ignore_bcn)
drivers/net/wireless/realtek/rtw89/chan.c
2708
else if (rtw89_mcc_ignore_bcn(rtwdev, aux) || ref->ignore_bcn)
drivers/net/wireless/realtek/rtw89/chan.c
2709
rtw89_mcc_detect_connection(rtwdev, ref);
drivers/net/wireless/realtek/rtw89/chan.c
724
struct rtw89_mcc_role *ref = &mcc->role_ref;
drivers/net/wireless/realtek/rtw89/chan.c
731
req.macid_x = ref->rtwvif_link->mac_id;
drivers/net/wireless/realtek/rtw89/chan.c
749
struct rtw89_mcc_role *ref = &mcc->role_ref;
drivers/net/wireless/realtek/rtw89/chan.c
758
arg.infos[0].band = ref->rtwvif_link->mac_idx;
drivers/net/wireless/realtek/rtw89/chan.c
759
arg.infos[0].port = ref->rtwvif_link->port;
drivers/net/wireless/realtek/rtw89/chan.c
779
struct rtw89_mcc_role *ref = &mcc->role_ref;
drivers/net/wireless/realtek/rtw89/chan.c
781
u32 bcn_intvl_ref_us = ieee80211_tu_to_usec(ref->beacon_interval);
drivers/net/wireless/realtek/rtw89/chan.c
794
tbtt_ofst_ref = rtw89_mcc_get_tbtt_ofst(rtwdev, ref, tsf_ref);
drivers/net/wireless/realtek/rtw89/core.h
6044
struct rtw89_mcc_courtesy_cfg ref;
drivers/net/wireless/realtek/rtw89/rtw8851b.c
1759
enum rtw89_phy_idx phy_idx, s16 ref)
drivers/net/wireless/realtek/rtw89/rtw8851b.c
1770
pwr_s10_3 = (ref << 1) + (s16)(ofst_int) + (s16)(base_cw_0db << 3);
drivers/net/wireless/realtek/rtw89/rtw8851b.c
1776
tssi_ofst_cw = (u32)((s16)tssi_16dbm_cw + (ref << 1) - (16 << 3));
drivers/net/wireless/realtek/rtw89/rtw8851b.c
1783
u32_encode_bits(ref, B_DPD_REF);
drivers/net/wireless/realtek/rtw89/rtw8852a.c
1479
enum rtw89_phy_idx phy_idx, s16 ref)
drivers/net/wireless/realtek/rtw89/rtw8852a.c
1490
pwr_s10_3 = (ref << 1) + (s16)(ofst_int) + (s16)(base_cw_0db << 3);
drivers/net/wireless/realtek/rtw89/rtw8852a.c
1496
tssi_ofst_cw = (u32)((s16)tssi_16dbm_cw + (ref << 1) - (16 << 3));
drivers/net/wireless/realtek/rtw89/rtw8852a.c
1501
return (tssi_ofst_cw << 18) | (pwr_cw << 9) | (ref & GENMASK(8, 0));
drivers/net/wireless/realtek/rtw89/rtw8852b_common.c
1207
s16 ref, u16 pwr_ofst_decrease)
drivers/net/wireless/realtek/rtw89/rtw8852b_common.c
1217
pwr_s10_3 = (ref << 1) + (s16)(base_cw_0db << 3) - pwr_ofst_decrease;
drivers/net/wireless/realtek/rtw89/rtw8852b_common.c
1223
tssi_ofst_cw = (u32)((s16)tssi_16dbm_cw + (ref << 1) - (16 << 3)) -
drivers/net/wireless/realtek/rtw89/rtw8852b_common.c
1231
u32_encode_bits(ref, B_DPD_REF);
drivers/net/wireless/realtek/rtw89/rtw8852c.c
1985
s16 ref, u16 pwr_ofst_decrease)
drivers/net/wireless/realtek/rtw89/rtw8852c.c
1995
pwr_s10_3 = (ref << 1) + (s16)(base_cw_0db << 3) - pwr_ofst_decrease;
drivers/net/wireless/realtek/rtw89/rtw8852c.c
2001
tssi_ofst_cw = (u32)((s16)tssi_16dbm_cw + (ref << 1) - (16 << 3)) -
drivers/net/wireless/realtek/rtw89/rtw8852c.c
2007
return (tssi_ofst_cw << 18) | (pwr_cw << 9) | (ref & GENMASK(8, 0));
drivers/net/wwan/t7xx/t7xx_state_monitor.c
107
static void fsm_release_command(struct kref *ref)
drivers/net/wwan/t7xx/t7xx_state_monitor.c
109
struct t7xx_fsm_command *cmd = container_of(ref, typeof(*cmd), refcnt);
drivers/net/xen-netback/hash.c
294
.source.u.ref = gref,
drivers/net/xen-netback/hash.c
340
.source.u.ref = gref,
drivers/net/xen-netback/netback.c
407
cop->source.u.ref = txp->gref;
drivers/net/xen-netback/netback.c
598
(*gopp_copy)->source.u.ref);
drivers/net/xen-netback/netback.c
644
gop_map->ref);
drivers/net/xen-netback/rx.c
213
op->source.u.ref = foreign->gref;
drivers/net/xen-netback/rx.c
221
op->dest.u.ref = req->gref;
drivers/net/xen-netfront.c
1031
grant_ref_t ref = xennet_get_rx_ref(queue, cons);
drivers/net/xen-netfront.c
1058
if (ref == INVALID_GRANT_REF) {
drivers/net/xen-netfront.c
1071
xennet_move_rx_slot(queue, skb, ref);
drivers/net/xen-netfront.c
1076
if (!gnttab_end_foreign_access_ref(ref)) {
drivers/net/xen-netfront.c
1084
gnttab_release_grant_reference(&queue->gref_rx_head, ref);
drivers/net/xen-netfront.c
1119
ref = xennet_get_rx_ref(queue, cons + slots);
drivers/net/xen-netfront.c
1447
int id, ref;
drivers/net/xen-netfront.c
1459
ref = queue->grant_rx_ref[id];
drivers/net/xen-netfront.c
1460
if (ref == INVALID_GRANT_REF)
drivers/net/xen-netfront.c
1469
gnttab_end_foreign_access(ref, page);
drivers/net/xen-netfront.c
1808
static void xennet_end_access(int ref, void *page)
drivers/net/xen-netfront.c
1811
if (ref != INVALID_GRANT_REF)
drivers/net/xen-netfront.c
1812
gnttab_end_foreign_access(ref, virt_to_page(page));
drivers/net/xen-netfront.c
231
grant_ref_t ref = queue->grant_rx_ref[i];
drivers/net/xen-netfront.c
233
return ref;
drivers/net/xen-netfront.c
313
grant_ref_t ref;
drivers/net/xen-netfront.c
328
ref = gnttab_claim_grant_reference(&queue->gref_rx_head);
drivers/net/xen-netfront.c
329
WARN_ON_ONCE(IS_ERR_VALUE((unsigned long)(int)ref));
drivers/net/xen-netfront.c
330
queue->grant_rx_ref[id] = ref;
drivers/net/xen-netfront.c
335
gnttab_page_grant_foreign_access_ref_one(ref,
drivers/net/xen-netfront.c
340
req->gref = ref;
drivers/net/xen-netfront.c
480
grant_ref_t ref;
drivers/net/xen-netfront.c
488
ref = gnttab_claim_grant_reference(&queue->gref_tx_head);
drivers/net/xen-netfront.c
489
WARN_ON_ONCE(IS_ERR_VALUE((unsigned long)(int)ref));
drivers/net/xen-netfront.c
491
gnttab_grant_foreign_access_ref(ref, queue->info->xbdev->otherend_id,
drivers/net/xen-netfront.c
496
queue->grant_tx_ref[id] = ref;
drivers/net/xen-netfront.c
499
info->tx_local.gref = ref;
drivers/net/xen-netfront.c
913
grant_ref_t ref)
drivers/net/xen-netfront.c
919
queue->grant_rx_ref[new] = ref;
drivers/net/xen-netfront.c
921
RING_GET_REQUEST(&queue->rx, queue->rx.req_prod_pvt)->gref = ref;
drivers/net/xen-netfront.c
937
grant_ref_t ref;
drivers/net/xen-netfront.c
959
ref = xennet_get_rx_ref(queue, cons);
drivers/net/xen-netfront.c
960
xennet_move_rx_slot(queue, skb, ref);
drivers/nvme/host/core.c
3109
static void nvme_destroy_subsystem(struct kref *ref)
drivers/nvme/host/core.c
3112
container_of(ref, struct nvme_subsystem, ref);
drivers/nvme/host/core.c
3125
kref_put(&subsys->ref, nvme_destroy_subsystem);
drivers/nvme/host/core.c
3148
if (!kref_get_unless_zero(&subsys->ref))
drivers/nvme/host/core.c
3213
kref_init(&subsys->ref);
drivers/nvme/host/core.c
3910
kref_init(&head->ref);
drivers/nvme/host/core.c
3925
kref_get(&ctrl->subsys->ref);
drivers/nvme/host/core.c
665
static void nvme_free_ns_head(struct kref *ref)
drivers/nvme/host/core.c
668
container_of(ref, struct nvme_ns_head, ref);
drivers/nvme/host/core.c
680
return kref_get_unless_zero(&head->ref);
drivers/nvme/host/core.c
685
kref_put(&head->ref, nvme_free_ns_head);
drivers/nvme/host/fabrics.c
107
static void nvmf_host_destroy(struct kref *ref)
drivers/nvme/host/fabrics.c
109
struct nvmf_host *host = container_of(ref, struct nvmf_host, ref);
drivers/nvme/host/fabrics.c
121
kref_put(&host->ref, nvmf_host_destroy);
drivers/nvme/host/fabrics.c
33
kref_init(&host->ref);
drivers/nvme/host/fabrics.c
57
kref_get(&host->ref);
drivers/nvme/host/fabrics.h
30
struct kref ref;
drivers/nvme/host/fc.c
126
struct kref ref;
drivers/nvme/host/fc.c
141
struct kref ref;
drivers/nvme/host/fc.c
172
struct kref ref;
drivers/nvme/host/fc.c
2356
nvme_fc_ctrl_free(struct kref *ref)
drivers/nvme/host/fc.c
2359
container_of(ref, struct nvme_fc_ctrl, ref);
drivers/nvme/host/fc.c
2381
kref_put(&ctrl->ref, nvme_fc_ctrl_free);
drivers/nvme/host/fc.c
2387
return kref_get_unless_zero(&ctrl->ref);
drivers/nvme/host/fc.c
240
nvme_fc_free_lport(struct kref *ref)
drivers/nvme/host/fc.c
243
container_of(ref, struct nvme_fc_lport, ref);
drivers/nvme/host/fc.c
265
kref_put(&lport->ref, nvme_fc_free_lport);
drivers/nvme/host/fc.c
271
return kref_get_unless_zero(&lport->ref);
drivers/nvme/host/fc.c
3482
kref_init(&ctrl->ref);
drivers/nvme/host/fc.c
407
kref_init(&newrec->ref);
drivers/nvme/host/fc.c
513
nvme_fc_free_rport(struct kref *ref)
drivers/nvme/host/fc.c
516
container_of(ref, struct nvme_fc_rport, ref);
drivers/nvme/host/fc.c
542
kref_put(&rport->ref, nvme_fc_free_rport);
drivers/nvme/host/fc.c
548
return kref_get_unless_zero(&rport->ref);
drivers/nvme/host/fc.c
720
kref_init(&newrec->ref);
drivers/nvme/host/nvme.h
489
struct kref ref;
drivers/nvme/host/nvme.h
535
struct kref ref;
drivers/nvme/host/rdma.c
1161
if (!refcount_dec_and_test(&req->ref))
drivers/nvme/host/rdma.c
1535
refcount_set(&req->ref, 2); /* send and recv completions */
drivers/nvme/host/rdma.c
343
static void nvme_rdma_free_dev(struct kref *ref)
drivers/nvme/host/rdma.c
346
container_of(ref, struct nvme_rdma_device, ref);
drivers/nvme/host/rdma.c
358
kref_put(&dev->ref, nvme_rdma_free_dev);
drivers/nvme/host/rdma.c
363
return kref_get_unless_zero(&dev->ref);
drivers/nvme/host/rdma.c
383
kref_init(&ndev->ref);
drivers/nvme/host/rdma.c
45
struct kref ref;
drivers/nvme/host/rdma.c
68
refcount_t ref;
drivers/nvme/target/core.c
1001
struct nvmet_sq *sq = container_of(ref, struct nvmet_sq, ref);
drivers/nvme/target/core.c
1013
ret = percpu_ref_init(&sq->ref, nvmet_sq_free, 0, GFP_KERNEL);
drivers/nvme/target/core.c
1209
if (unlikely(!percpu_ref_tryget_live(&sq->ref))) {
drivers/nvme/target/core.c
1227
percpu_ref_put(&req->sq->ref);
drivers/nvme/target/core.c
1483
if (!kref_get_unless_zero(&ctrl->ref))
drivers/nvme/target/core.c
1636
kref_init(&ctrl->ref);
drivers/nvme/target/core.c
1733
static void nvmet_ctrl_free(struct kref *ref)
drivers/nvme/target/core.c
1735
struct nvmet_ctrl *ctrl = container_of(ref, struct nvmet_ctrl, ref);
drivers/nvme/target/core.c
1766
kref_put(&ctrl->ref, nvmet_ctrl_free);
drivers/nvme/target/core.c
1798
if (!kref_get_unless_zero(&nvmet_disc_subsys->ref))
drivers/nvme/target/core.c
1806
if (kref_get_unless_zero(&nvmet_disc_subsys->ref)) {
drivers/nvme/target/core.c
1814
if (!kref_get_unless_zero(&p->subsys->ref))
drivers/nvme/target/core.c
1876
kref_init(&subsys->ref);
drivers/nvme/target/core.c
1900
static void nvmet_subsys_free(struct kref *ref)
drivers/nvme/target/core.c
1903
container_of(ref, struct nvmet_subsys, ref);
drivers/nvme/target/core.c
1932
kref_put(&subsys->ref, nvmet_subsys_free);
drivers/nvme/target/core.c
453
percpu_ref_get(&req->ns->ref);
drivers/nvme/target/core.c
457
static void nvmet_destroy_namespace(struct percpu_ref *ref)
drivers/nvme/target/core.c
459
struct nvmet_ns *ns = container_of(ref, struct nvmet_ns, ref);
drivers/nvme/target/core.c
466
percpu_ref_put(&ns->ref);
drivers/nvme/target/core.c
605
if (percpu_ref_init(&ns->ref, nvmet_destroy_namespace, 0, GFP_KERNEL))
drivers/nvme/target/core.c
651
percpu_ref_kill(&ns->ref);
drivers/nvme/target/core.c
654
percpu_ref_exit(&ns->ref);
drivers/nvme/target/core.c
805
percpu_ref_put(&sq->ref);
drivers/nvme/target/core.c
811
refcount_set(&cq->ref, 1);
drivers/nvme/target/core.c
817
return refcount_inc_not_zero(&cq->ref);
drivers/nvme/target/core.c
823
if (refcount_dec_and_test(&cq->ref))
drivers/nvme/target/core.c
858
static void nvmet_confirm_sq(struct percpu_ref *ref)
drivers/nvme/target/core.c
860
struct nvmet_sq *sq = container_of(ref, struct nvmet_sq, ref);
drivers/nvme/target/core.c
888
return refcount_read(&cq->ref) > 1;
drivers/nvme/target/core.c
901
if (!kref_get_unless_zero(&ctrl->ref))
drivers/nvme/target/core.c
934
if (!kref_get_unless_zero(&ctrl->ref))
drivers/nvme/target/core.c
968
percpu_ref_kill_and_confirm(&sq->ref, nvmet_confirm_sq);
drivers/nvme/target/core.c
971
percpu_ref_exit(&sq->ref);
drivers/nvme/target/core.c
999
static void nvmet_sq_free(struct percpu_ref *ref)
drivers/nvme/target/fabrics-cmd.c
211
kref_get(&ctrl->ref);
drivers/nvme/target/fc.c
1051
kref_init(&newhost->ref);
drivers/nvme/target/fc.c
1134
kref_init(&assoc->ref);
drivers/nvme/target/fc.c
114
struct kref ref;
drivers/nvme/target/fc.c
1162
nvmet_fc_target_assoc_free(struct kref *ref)
drivers/nvme/target/fc.c
1165
container_of(ref, struct nvmet_fc_tgt_assoc, ref);
drivers/nvme/target/fc.c
1195
kref_put(&assoc->ref, nvmet_fc_target_assoc_free);
drivers/nvme/target/fc.c
1201
return kref_get_unless_zero(&assoc->ref);
drivers/nvme/target/fc.c
1407
kref_init(&newrec->ref);
drivers/nvme/target/fc.c
1440
nvmet_fc_free_tgtport(struct kref *ref)
drivers/nvme/target/fc.c
1443
container_of(ref, struct nvmet_fc_tgtport, ref);
drivers/nvme/target/fc.c
1464
kref_put(&tgtport->ref, nvmet_fc_free_tgtport);
drivers/nvme/target/fc.c
1470
return kref_get_unless_zero(&tgtport->ref);
drivers/nvme/target/fc.c
149
struct kref ref;
drivers/nvme/target/fc.c
158
struct kref ref;
drivers/nvme/target/fc.c
171
struct kref ref;
drivers/nvme/target/fc.c
812
kref_init(&queue->ref);
drivers/nvme/target/fc.c
837
nvmet_fc_tgt_queue_free(struct kref *ref)
drivers/nvme/target/fc.c
840
container_of(ref, struct nvmet_fc_tgt_queue, ref);
drivers/nvme/target/fc.c
852
kref_put(&queue->ref, nvmet_fc_tgt_queue_free);
drivers/nvme/target/fc.c
858
return kref_get_unless_zero(&queue->ref);
drivers/nvme/target/fc.c
970
nvmet_fc_hostport_free(struct kref *ref)
drivers/nvme/target/fc.c
973
container_of(ref, struct nvmet_fc_hostport, ref);
drivers/nvme/target/fc.c
989
kref_put(&hostport->ref, nvmet_fc_hostport_free);
drivers/nvme/target/fc.c
995
return kref_get_unless_zero(&hostport->ref);
drivers/nvme/target/fcloop.c
1056
if (!refcount_dec_and_test(&lport->ref))
drivers/nvme/target/fcloop.c
1069
return refcount_inc_not_zero(&lport->ref);
drivers/nvme/target/fcloop.c
1077
if (!refcount_dec_and_test(&nport->ref))
drivers/nvme/target/fcloop.c
1093
return refcount_inc_not_zero(&nport->ref);
drivers/nvme/target/fcloop.c
1236
refcount_set(&lport->ref, 1);
drivers/nvme/target/fcloop.c
1378
refcount_set(&newnport->ref, 1);
drivers/nvme/target/fcloop.c
210
refcount_t ref;
drivers/nvme/target/fcloop.c
247
refcount_t ref;
drivers/nvme/target/fcloop.c
281
refcount_t ref;
drivers/nvme/target/fcloop.c
582
if (!refcount_dec_and_test(&tfcp_req->ref))
drivers/nvme/target/fcloop.c
591
return refcount_inc_not_zero(&tfcp_req->ref);
drivers/nvme/target/fcloop.c
791
refcount_set(&tfcp_req->ref, 1);
drivers/nvme/target/nvmet.h
101
struct percpu_ref ref;
drivers/nvme/target/nvmet.h
147
refcount_t ref;
drivers/nvme/target/nvmet.h
152
struct percpu_ref ref;
drivers/nvme/target/nvmet.h
279
struct kref ref;
drivers/nvme/target/nvmet.h
319
struct kref ref;
drivers/nvme/target/nvmet.h
94
struct percpu_ref ref;
drivers/nvme/target/nvmet.h
954
percpu_ref_put(&pc_ref->ref);
drivers/nvme/target/pci-epf.c
534
if (iv->ref && iv->vector == vector)
drivers/nvme/target/pci-epf.c
551
iv->ref++;
drivers/nvme/target/pci-epf.c
557
if (!iv->ref)
drivers/nvme/target/pci-epf.c
564
iv->ref = 1;
drivers/nvme/target/pci-epf.c
583
iv->ref--;
drivers/nvme/target/pci-epf.c
584
if (!iv->ref) {
drivers/nvme/target/pci-epf.c
74
unsigned int ref;
drivers/nvme/target/pr.c
1001
static void nvmet_pr_ctrl_ns_all_cmds_done(struct percpu_ref *ref)
drivers/nvme/target/pr.c
1004
container_of(ref, struct nvmet_pr_per_ctrl_ref, ref);
drivers/nvme/target/pr.c
1020
ret = percpu_ref_init(&pc_ref->ref, nvmet_pr_ctrl_ns_all_cmds_done,
drivers/nvme/target/pr.c
1034
percpu_ref_exit(&pc_ref->ref);
drivers/nvme/target/pr.c
1074
percpu_ref_exit(&pc_ref->ref);
drivers/nvme/target/pr.c
1094
percpu_ref_exit(&pc_ref->ref);
drivers/nvme/target/pr.c
1127
percpu_ref_exit(&pc_ref->ref);
drivers/nvme/target/pr.c
1150
percpu_ref_exit(&pc_ref->ref);
drivers/nvme/target/pr.c
436
static void nvmet_pr_confirm_ns_pc_ref(struct percpu_ref *ref)
drivers/nvme/target/pr.c
439
container_of(ref, struct nvmet_pr_per_ctrl_ref, ref);
drivers/nvme/target/pr.c
452
percpu_ref_kill_and_confirm(&pc_ref->ref,
drivers/nvme/target/pr.c
602
if (percpu_ref_is_dying(&pc_ref->ref)) {
drivers/nvme/target/pr.c
606
percpu_ref_resurrect(&pc_ref->ref);
drivers/nvme/target/pr.c
995
if (unlikely(!percpu_ref_tryget_live(&pc_ref->ref)))
drivers/nvme/target/rdma.c
1178
static void nvmet_rdma_free_dev(struct kref *ref)
drivers/nvme/target/rdma.c
1181
container_of(ref, struct nvmet_rdma_device, ref);
drivers/nvme/target/rdma.c
1206
kref_get_unless_zero(&ndev->ref))
drivers/nvme/target/rdma.c
1235
kref_init(&ndev->ref);
drivers/nvme/target/rdma.c
1375
kref_put(&dev->ref, nvmet_rdma_free_dev);
drivers/nvme/target/rdma.c
139
struct kref ref;
drivers/nvme/target/rdma.c
1624
kref_put(&ndev->ref, nvmet_rdma_free_dev);
drivers/pci/hotplug/acpiphp.h
59
struct kref ref;
drivers/pci/hotplug/acpiphp_glue.c
112
kref_get(&bridge->ref);
drivers/pci/hotplug/acpiphp_glue.c
117
kref_put(&bridge->ref, free_bridge);
drivers/pci/hotplug/acpiphp_glue.c
158
bridge = container_of(kref, struct acpiphp_bridge, ref);
drivers/pci/hotplug/acpiphp_glue.c
877
kref_init(&bridge->ref);
drivers/pci/p2pdma.c
125
kaddr = (void *)gen_pool_alloc_owner(p2pdma->pool, len, (void **)&ref);
drivers/pci/p2pdma.c
136
if (unlikely(!percpu_ref_tryget_live_rcu(ref))) {
drivers/pci/p2pdma.c
162
percpu_ref_put(ref);
drivers/pci/p2pdma.c
165
percpu_ref_get(ref);
drivers/pci/p2pdma.c
171
percpu_ref_put(ref);
drivers/pci/p2pdma.c
218
struct percpu_ref *ref;
drivers/pci/p2pdma.c
221
PAGE_SIZE, (void **)&ref);
drivers/pci/p2pdma.c
222
percpu_ref_put(ref);
drivers/pci/p2pdma.c
448
&pgmap->ref);
drivers/pci/p2pdma.c
925
struct percpu_ref *ref;
drivers/pci/p2pdma.c
938
ret = (void *)gen_pool_alloc_owner(p2pdma->pool, size, (void **) &ref);
drivers/pci/p2pdma.c
942
if (unlikely(!percpu_ref_tryget_live_rcu(ref))) {
drivers/pci/p2pdma.c
960
struct percpu_ref *ref;
drivers/pci/p2pdma.c
964
(void **) &ref);
drivers/pci/p2pdma.c
965
percpu_ref_put(ref);
drivers/pci/p2pdma.c
98
struct percpu_ref *ref;
drivers/pcmcia/cs_internal.h
30
struct kref ref;
drivers/pcmcia/ds.c
230
static void pcmcia_release_function(struct kref *ref)
drivers/pcmcia/ds.c
232
struct config_t *c = container_of(ref, struct config_t, ref);
drivers/pcmcia/ds.c
246
kref_put(&p_dev->function_config->ref, pcmcia_release_function);
drivers/pcmcia/ds.c
532
kref_get(&p_dev->function_config->ref);
drivers/pcmcia/ds.c
551
kref_init(&c->ref);
drivers/perf/xgene_pmu.c
336
XGENE_PMU_EVENT_ATTR(ref-cmd-sent, 0x0b),
drivers/perf/xgene_pmu.c
527
XGENE_PMU_EVENT_ATTR(ref-sent, 0x0a),
drivers/phy/intel/phy-intel-lgm-combo.c
411
struct fwnode_reference_args ref;
drivers/phy/intel/phy-intel-lgm-combo.c
454
1, 0, &ref);
drivers/phy/intel/phy-intel-lgm-combo.c
458
cbphy->id = ref.args[0];
drivers/phy/intel/phy-intel-lgm-combo.c
459
cbphy->syscfg = device_node_to_regmap(to_of_node(ref.fwnode));
drivers/phy/intel/phy-intel-lgm-combo.c
460
fwnode_handle_put(ref.fwnode);
drivers/phy/intel/phy-intel-lgm-combo.c
463
0, &ref);
drivers/phy/intel/phy-intel-lgm-combo.c
467
cbphy->bid = ref.args[0];
drivers/phy/intel/phy-intel-lgm-combo.c
468
cbphy->hsiocfg = device_node_to_regmap(to_of_node(ref.fwnode));
drivers/phy/intel/phy-intel-lgm-combo.c
469
fwnode_handle_put(ref.fwnode);
drivers/pinctrl/ti/pinctrl-ti-iodelay.c
182
static inline u32 ti_iodelay_compute_dpe(u16 period, u16 ref, u16 delay,
drivers/pinctrl/ti/pinctrl-ti-iodelay.c
188
m = 10 * (u64)period * (u64)ref;
drivers/pmdomain/qcom/cpr.c
1036
static unsigned long cpr_get_opp_hz_for_req(struct dev_pm_opp *ref,
drivers/pmdomain/qcom/cpr.c
1046
ref_np = dev_pm_opp_get_of_node(ref);
drivers/power/sequencing/core.c
1026
struct pwrseq_unit_dep *ref;
drivers/power/sequencing/core.c
103
kref_init(&unit->ref);
drivers/power/sequencing/core.c
1035
list_for_each_entry(ref, &unit->deps, list)
drivers/power/sequencing/core.c
1036
seq_printf(seq, " [%s]\n", ref->unit->name);
drivers/power/sequencing/core.c
113
kref_get(&unit->ref);
drivers/power/sequencing/core.c
118
static void pwrseq_unit_release(struct kref *ref);
drivers/power/sequencing/core.c
122
kref_put(&unit->ref, pwrseq_unit_release);
drivers/power/sequencing/core.c
150
static void pwrseq_unit_dep_free(struct pwrseq_unit_dep *ref)
drivers/power/sequencing/core.c
152
pwrseq_unit_put(ref->unit);
drivers/power/sequencing/core.c
153
kfree(ref);
drivers/power/sequencing/core.c
166
static void pwrseq_unit_release(struct kref *ref)
drivers/power/sequencing/core.c
168
struct pwrseq_unit *unit = container_of(ref, struct pwrseq_unit, ref);
drivers/power/sequencing/core.c
80
struct kref ref;
drivers/power/supply/bq27xxx_battery.c
1011
#define BQ27XXX_DATA(ref, key, opt) { \
drivers/power/supply/bq27xxx_battery.c
1014
.regs = ref##_regs, \
drivers/power/supply/bq27xxx_battery.c
1015
.dm_regs = ref##_dm_regs, \
drivers/power/supply/bq27xxx_battery.c
1016
.props = ref##_props, \
drivers/power/supply/bq27xxx_battery.c
1017
.props_size = ARRAY_SIZE(ref##_props) }
drivers/ptp/ptp_clockmatrix.c
1404
enum manual_reference *ref)
drivers/ptp/ptp_clockmatrix.c
1418
*ref = dpll_manu_ref_cfg >> MANUAL_REFERENCE_SHIFT;
drivers/ptp/ptp_clockmatrix.c
1424
enum manual_reference ref)
drivers/ptp/ptp_clockmatrix.c
1438
dpll_manu_ref_cfg |= (ref << MANUAL_REFERENCE_SHIFT);
drivers/ptp/ptp_clockmatrix.c
1597
enum manual_reference ref)
drivers/ptp/ptp_clockmatrix.c
1606
switch (ref) {
drivers/ptp/ptp_clockmatrix.c
1615
"Unsupported MANUAL_REFERENCE: 0x%02x", ref);
drivers/ptp/ptp_clockmatrix.c
1650
enum manual_reference ref = MANU_REF_XO_DPLL;
drivers/ptp/ptp_clockmatrix.c
1664
err = idtcm_get_manual_reference(channel, &ref);
drivers/ptp/ptp_clockmatrix.c
1669
err = initialize_operating_mode_with_manual_reference(channel, ref);
drivers/ptp/ptp_clockmatrix.c
242
static int arm_tod_read_trig_sel_refclk(struct idtcm_channel *channel, u8 ref)
drivers/ptp/ptp_clockmatrix.c
250
val |= (ref << WR_REF_INDEX_SHIFT);
drivers/ptp/ptp_clockmatrix.c
281
int ref;
drivers/ptp/ptp_clockmatrix.c
292
ref = ptp_find_pin(channel->ptp_clock, PTP_PF_EXTTS, channel->tod);
drivers/ptp/ptp_clockmatrix.c
294
if (ref < 0) {
drivers/ptp/ptp_clockmatrix.c
300
err = arm_tod_read_trig_sel_refclk(&idtcm->channel[index], ref);
drivers/ptp/ptp_clockmatrix.c
305
idtcm->channel[index].refn = ref;
drivers/ptp/ptp_idt82p33.c
183
static int map_ref_to_tod_trig_sel(int ref, u8 *trigger)
drivers/ptp/ptp_idt82p33.c
187
switch (ref) {
drivers/ptp/ptp_idt82p33.c
244
int ref;
drivers/ptp/ptp_idt82p33.c
258
ref = ptp_find_pin(channel->ptp_clock, PTP_PF_EXTTS, channel->plln);
drivers/ptp/ptp_idt82p33.c
260
if (ref < 0) {
drivers/ptp/ptp_idt82p33.c
266
err = map_ref_to_tod_trig_sel(ref, &trigger);
drivers/ptp/ptp_idt82p33.c
270
"%s: Unsupported ref %d!\n", __func__, ref);
drivers/rapidio/devices/rio_mport_cdev.c
1087
kref_init(&map->ref);
drivers/rapidio/devices/rio_mport_cdev.c
1115
kref_put(&mapping->ref, mport_release_mapping);
drivers/rapidio/devices/rio_mport_cdev.c
1139
kref_put(&map->ref, mport_release_mapping);
drivers/rapidio/devices/rio_mport_cdev.c
118
struct kref ref; /* refcount of vmas sharing the mapping */
drivers/rapidio/devices/rio_mport_cdev.c
1214
kref_init(&map->ref);
drivers/rapidio/devices/rio_mport_cdev.c
1291
kref_put(&mapping->ref, mport_release_mapping);
drivers/rapidio/devices/rio_mport_cdev.c
1326
kref_put(&map->ref, mport_release_mapping);
drivers/rapidio/devices/rio_mport_cdev.c
2030
kref_put(&map->ref, mport_release_mapping);
drivers/rapidio/devices/rio_mport_cdev.c
2128
static void mport_release_mapping(struct kref *ref)
drivers/rapidio/devices/rio_mport_cdev.c
2131
container_of(ref, struct rio_mport_mapping, ref);
drivers/rapidio/devices/rio_mport_cdev.c
2160
kref_get(&map->ref);
drivers/rapidio/devices/rio_mport_cdev.c
2169
kref_put(&map->ref, mport_release_mapping);
drivers/rapidio/devices/rio_mport_cdev.c
238
static void mport_release_mapping(struct kref *ref);
drivers/rapidio/devices/rio_mport_cdev.c
2501
kref_put(&map->ref, mport_release_mapping);
drivers/rapidio/devices/rio_mport_cdev.c
366
kref_init(&map->ref);
drivers/rapidio/devices/rio_mport_cdev.c
462
kref_put(&map->ref, mport_release_mapping);
drivers/rapidio/devices/rio_mport_cdev.c
552
static void dma_req_free(struct kref *ref)
drivers/rapidio/devices/rio_mport_cdev.c
554
struct mport_dma_req *req = container_of(ref, struct mport_dma_req,
drivers/rapidio/devices/rio_mport_cdev.c
568
kref_put(&req->map->ref, mport_release_mapping);
drivers/rapidio/devices/rio_mport_cdev.c
887
kref_get(&map->ref);
drivers/rapidio/rio_cm.c
1318
kref_init(&ch->ref);
drivers/rapidio/rio_cm.c
1362
static void riocm_ch_free(struct kref *ref)
drivers/rapidio/rio_cm.c
1364
struct rio_channel *ch = container_of(ref, struct rio_channel, ref);
drivers/rapidio/rio_cm.c
178
struct kref ref; /* channel refcount */
drivers/rapidio/rio_cm.c
219
static void riocm_ch_free(struct kref *ref);
drivers/rapidio/rio_cm.c
286
kref_get(&ch->ref);
drivers/rapidio/rio_cm.c
293
kref_put(&ch->ref, riocm_ch_free);
drivers/rpmsg/qcom_glink_native.c
260
static void qcom_glink_channel_release(struct kref *ref)
drivers/rpmsg/qcom_glink_native.c
262
struct glink_channel *channel = container_of(ref, struct glink_channel,
drivers/s390/scsi/zfcp_aux.c
353
kref_init(&adapter->ref);
drivers/s390/scsi/zfcp_aux.c
469
void zfcp_adapter_release(struct kref *ref)
drivers/s390/scsi/zfcp_aux.c
471
struct zfcp_adapter *adapter = container_of(ref, struct zfcp_adapter,
drivers/s390/scsi/zfcp_aux.c
472
ref);
drivers/s390/scsi/zfcp_aux.c
512
kref_get(&adapter->ref);
drivers/s390/scsi/zfcp_ccw.c
170
kref_get(&adapter->ref);
drivers/s390/scsi/zfcp_ccw.c
28
kref_get(&adapter->ref);
drivers/s390/scsi/zfcp_ccw.c
38
kref_put(&adapter->ref, zfcp_adapter_release);
drivers/s390/scsi/zfcp_def.h
149
struct kref ref;
drivers/s390/scsi/zfcp_erp.c
1503
kref_put(&adapter->ref, zfcp_adapter_release);
drivers/s390/scsi/zfcp_erp.c
1531
kref_get(&adapter->ref);
drivers/s390/scsi/zfcp_erp.c
1602
kref_put(&adapter->ref, zfcp_adapter_release);
drivers/s390/scsi/zfcp_erp.c
257
kref_get(&adapter->ref);
drivers/scsi/bnx2fc/bnx2fc.h
500
void bnx2fc_cmd_release(struct kref *ref);
drivers/scsi/bnx2fc/bnx2fc_io.c
515
void bnx2fc_cmd_release(struct kref *ref)
drivers/scsi/bnx2fc/bnx2fc_io.c
517
struct bnx2fc_cmd *io_req = container_of(ref,
drivers/scsi/ch.c
110
struct kref ref;
drivers/scsi/ch.c
573
static void ch_destroy(struct kref *ref)
drivers/scsi/ch.c
575
scsi_changer *ch = container_of(ref, scsi_changer, ref);
drivers/scsi/ch.c
589
kref_put(&ch->ref, ch_destroy);
drivers/scsi/ch.c
602
if (ch == NULL || !kref_get_unless_zero(&ch->ref)) {
drivers/scsi/ch.c
608
kref_put(&ch->ref, ch_destroy);
drivers/scsi/ch.c
932
kref_init(&ch->ref);
drivers/scsi/ch.c
981
kref_put(&ch->ref, ch_destroy);
drivers/scsi/elx/efct/efct_hw.c
1597
kref_init(&io->ref);
drivers/scsi/elx/efct/efct_hw.c
1657
struct efct_hw_io *io = container_of(arg, struct efct_hw_io, ref);
drivers/scsi/elx/efct/efct_hw.c
1675
return kref_put(&io->ref, io->release);
drivers/scsi/elx/efct/efct_hw.c
1949
if (kref_get_unless_zero(&io_to_abort->ref) == 0) {
drivers/scsi/elx/efct/efct_hw.c
1962
kref_put(&io_to_abort->ref, io_to_abort->release);
drivers/scsi/elx/efct/efct_hw.c
1972
kref_put(&io_to_abort->ref, io_to_abort->release);
drivers/scsi/elx/efct/efct_hw.c
2032
kref_put(&io_to_abort->ref, io_to_abort->release);
drivers/scsi/elx/efct/efct_hw.h
216
struct kref ref;
drivers/scsi/elx/efct/efct_io.c
183
if (kref_get_unless_zero(&io->ref))
drivers/scsi/elx/efct/efct_io.h
104
struct kref ref;
drivers/scsi/elx/efct/efct_lio.c
1058
struct efct_node *tgt_node = container_of(arg, struct efct_node, ref);
drivers/scsi/elx/efct/efct_lio.c
1076
kref_init(&tgt_node->ref);
drivers/scsi/elx/efct/efct_lio.c
1273
kref_put(&tgt_node->ref, tgt_node->release);
drivers/scsi/elx/efct/efct_lio.h
70
struct kref ref;
drivers/scsi/elx/efct/efct_scsi.c
102
WARN_ON(!refcount_read(&io->ref.refcount));
drivers/scsi/elx/efct/efct_scsi.c
103
kref_put(&io->ref, io->release);
drivers/scsi/elx/efct/efct_scsi.c
1088
kref_put(&io->io_to_abort->ref, io->io_to_abort->release);
drivers/scsi/elx/efct/efct_scsi.c
1108
if (kref_get_unless_zero(&io->ref) == 0) {
drivers/scsi/elx/efct/efct_scsi.c
1122
kref_put(&io->ref, io->release);
drivers/scsi/elx/efct/efct_scsi.c
1142
kref_put(&io->ref, io->release);
drivers/scsi/elx/efct/efct_scsi.c
1156
kref_put(&io->ref, io->release);
drivers/scsi/elx/efct/efct_scsi.c
49
kref_init(&io->ref);
drivers/scsi/elx/efct/efct_scsi.c
55
kref_get(&node->ref);
drivers/scsi/elx/efct/efct_scsi.c
77
struct efct_io *io = container_of(arg, struct efct_io, ref);
drivers/scsi/elx/efct/efct_scsi.c
93
kref_put(&node->ref, node->release);
drivers/scsi/elx/efct/efct_unsol.c
34
kref_get(&node->ref);
drivers/scsi/elx/efct/efct_unsol.c
446
kref_put(&abortio->ref, abortio->release);
drivers/scsi/elx/efct/efct_unsol.c
78
kref_put(&node->ref, node->release);
drivers/scsi/elx/libefc/efc_domain.c
1016
kref_put(&nport->ref, nport->release);
drivers/scsi/elx/libefc/efc_domain.c
106
struct efc_domain *domain = container_of(arg, struct efc_domain, ref);
drivers/scsi/elx/libefc/efc_domain.c
129
kref_put(&domain->ref, domain->release);
drivers/scsi/elx/libefc/efc_domain.c
145
kref_init(&domain->ref);
drivers/scsi/elx/libefc/efc_domain.c
972
if (!nport || !kref_get_unless_zero(&nport->ref)) {
drivers/scsi/elx/libefc/efc_els.c
103
kref_put(&els->ref, els->release);
drivers/scsi/elx/libefc/efc_els.c
110
container_of(arg, struct efc_els_io_req, ref);
drivers/scsi/elx/libefc/efc_els.c
61
kref_init(&els->ref);
drivers/scsi/elx/libefc/efc_els.h
15
struct kref ref;
drivers/scsi/elx/libefc/efc_node.c
104
kref_init(&node->ref);
drivers/scsi/elx/libefc/efc_node.c
106
kref_get(&nport->ref);
drivers/scsi/elx/libefc/efc_node.c
169
kref_put(&nport->ref, nport->release);
drivers/scsi/elx/libefc/efc_node.c
170
kref_put(&node->ref, node->release);
drivers/scsi/elx/libefc/efc_node.c
34
struct efc_node *node = container_of(arg, struct efc_node, ref);
drivers/scsi/elx/libefc/efc_nport.c
127
kref_get(&domain->ref);
drivers/scsi/elx/libefc/efc_nport.c
159
kref_put(&domain->ref, domain->release);
drivers/scsi/elx/libefc/efc_nport.c
160
kref_put(&nport->ref, nport->release);
drivers/scsi/elx/libefc/efc_nport.c
170
if (!nport || !kref_get_unless_zero(&nport->ref))
drivers/scsi/elx/libefc/efc_nport.c
263
kref_put(&nport->ref, nport->release);
drivers/scsi/elx/libefc/efc_nport.c
60
struct efc_nport *nport = container_of(arg, struct efc_nport, ref);
drivers/scsi/elx/libefc/efc_nport.c
610
kref_get(&nport->ref);
drivers/scsi/elx/libefc/efc_nport.c
710
kref_put(&nport->ref, nport->release);
drivers/scsi/elx/libefc/efc_nport.c
90
kref_init(&nport->ref);
drivers/scsi/elx/libefc/efclib.h
139
struct kref ref;
drivers/scsi/elx/libefc/efclib.h
214
struct kref ref;
drivers/scsi/elx/libefc/efclib.h
324
struct kref ref;
drivers/scsi/imm.c
79
static void imm_wakeup(void *ref)
drivers/scsi/imm.c
81
imm_struct *dev = (imm_struct *) ref;
drivers/scsi/megaraid/megaraid_sas.h
2473
union MR_LD_REF ref;
drivers/scsi/megaraid/megaraid_sas.h
459
u32 ref;
drivers/scsi/megaraid/megaraid_sas.h
563
union MR_PD_REF ref;
drivers/scsi/megaraid/megaraid_sas.h
768
__le32 ref;
drivers/scsi/megaraid/megaraid_sas.h
778
union MR_LD_REF ref;
drivers/scsi/megaraid/megaraid_sas_base.c
2566
if (newmap->ref.targetId ==
drivers/scsi/megaraid/megaraid_sas_base.c
2567
savedmap->ref.targetId) {
drivers/scsi/megaraid/megaraid_sas_base.c
2594
if (savedmap->ref.targetId ==
drivers/scsi/megaraid/megaraid_sas_base.c
2595
newmap->ref.targetId) {
drivers/scsi/megaraid/megaraid_sas_base.c
4845
ids = ci->ldList[ld_index].ref.targetId;
drivers/scsi/megaraid/megaraid_sas_base.c
4846
instance->ld_ids[ids] = ci->ldList[ld_index].ref.targetId;
drivers/scsi/ppa.c
74
static void ppa_wakeup(void *ref)
drivers/scsi/ppa.c
76
ppa_struct *dev = (ppa_struct *) ref;
drivers/scsi/qedf/qedf.h
503
extern void qedf_release_cmd(struct kref *ref);
drivers/scsi/qedf/qedf_io.c
428
void qedf_release_cmd(struct kref *ref)
drivers/scsi/qedf/qedf_io.c
431
container_of(ref, struct qedf_ioreq, refcount);
drivers/scsi/qla4xxx/ql4_glbl.h
71
void qla4xxx_srb_compl(struct kref *ref);
drivers/scsi/qla4xxx/ql4_os.c
4085
void qla4xxx_srb_compl(struct kref *ref)
drivers/scsi/qla4xxx/ql4_os.c
4087
struct srb *srb = container_of(ref, struct srb, srb_ref);
drivers/scsi/xen-scsifront.c
476
int err, ref, ref_cnt = 0;
drivers/scsi/xen-scsifront.c
520
ref = gnttab_claim_grant_reference(&gref_head);
drivers/scsi/xen-scsifront.c
521
BUG_ON(ref == -ENOSPC);
drivers/scsi/xen-scsifront.c
523
gnttab_grant_foreign_access_ref(ref,
drivers/scsi/xen-scsifront.c
526
shadow->gref[ref_cnt] = ref;
drivers/scsi/xen-scsifront.c
527
shadow->seg[ref_cnt].gref = ref;
drivers/scsi/xen-scsifront.c
554
ref = gnttab_claim_grant_reference(&gref_head);
drivers/scsi/xen-scsifront.c
555
BUG_ON(ref == -ENOSPC);
drivers/scsi/xen-scsifront.c
557
gnttab_grant_foreign_access_ref(ref,
drivers/scsi/xen-scsifront.c
562
shadow->gref[ref_cnt] = ref;
drivers/scsi/xen-scsifront.c
563
seg->gref = ref;
drivers/sh/clk/core.c
369
kref_init(&mapping->ref);
drivers/sh/clk/core.c
378
kref_get(&mapping->ref);
drivers/sh/clk/core.c
392
mapping = container_of(kref, struct clk_mapping, ref);
drivers/sh/clk/core.c
405
kref_put(&mapping->ref, clk_destroy_mapping);
drivers/soc/qcom/smem_state.c
131
static void qcom_smem_state_release(struct kref *ref)
drivers/soc/qcom/smem_state.c
133
struct qcom_smem_state *state = container_of(ref, struct qcom_smem_state, refcount);
drivers/spi/spi-offload.c
141
static void spi_offload_trigger_free(struct kref *ref)
drivers/spi/spi-offload.c
144
container_of(ref, struct spi_offload_trigger, ref);
drivers/spi/spi-offload.c
159
kref_put(&trigger->ref, spi_offload_trigger_free);
drivers/spi/spi-offload.c
192
kref_get(&trigger->ref);
drivers/spi/spi-offload.c
41
struct kref ref;
drivers/spi/spi-offload.c
419
kref_put(&trigger->ref, spi_offload_trigger_free);
drivers/spi/spi-offload.c
441
kref_init(&trigger->ref);
drivers/staging/greybus/camera.c
1205
static void gb_camera_release_module(struct kref *ref)
drivers/staging/greybus/camera.c
1208
container_of(ref, struct gb_camera_module, refcount);
drivers/staging/media/atomisp/pci/ia_css_isp_configs.c
279
size = binary->info->mem_offsets.offsets.config->dmem.ref.size;
drivers/staging/media/atomisp/pci/ia_css_isp_configs.c
283
offset = binary->info->mem_offsets.offsets.config->dmem.ref.offset;
drivers/staging/media/atomisp/pci/ia_css_isp_configs.h
60
struct ia_css_isp_parameter ref;
drivers/staging/media/atomisp/pci/ia_css_isp_states.c
165
unsigned int size = binary->info->mem_offsets.offsets.state->dmem.ref.size;
drivers/staging/media/atomisp/pci/ia_css_isp_states.c
167
unsigned int offset = binary->info->mem_offsets.offsets.state->dmem.ref.offset;
drivers/staging/media/atomisp/pci/ia_css_isp_states.h
50
struct ia_css_isp_parameter ref;
drivers/staging/media/ipu3/ipu3-css-fw.h
61
struct imgu_fw_isp_parameter ref;
drivers/staging/media/ipu3/ipu3-css-fw.h
69
struct imgu_fw_isp_parameter ref;
drivers/staging/media/ipu3/ipu3-css.c
774
&cofs->dmem.ref,
drivers/staging/media/ipu3/ipu3-css.c
851
&sofs->dmem.ref,
drivers/staging/media/meson/vdec/codec_vp9.c
1051
const int ref = (param->p.ref_info >>
drivers/staging/media/meson/vdec/codec_vp9.c
1054
if (vp9->ref_frame_map[ref] < 0)
drivers/staging/media/meson/vdec/codec_vp9.c
1058
__func__, vp9->ref_frame_map[ref]);
drivers/staging/media/meson/vdec/codec_vp9.c
1108
const int ref = (param->p.ref_info >>
drivers/staging/media/meson/vdec/codec_vp9.c
1110
const int idx = vp9->ref_frame_map[ref];
drivers/staging/media/meson/vdec/codec_vp9.c
629
int ref, mode;
drivers/staging/media/meson/vdec/codec_vp9.c
635
for (ref = LAST_FRAME; ref < MAX_REF_FRAMES; ++ref) {
drivers/staging/media/meson/vdec/codec_vp9.c
640
lf->ref_deltas[ref] * scale +
drivers/staging/media/meson/vdec/codec_vp9.c
642
lfi->lvl[seg_id][ref][mode] =
drivers/target/iscsi/cxgbit/cxgbit_cm.c
1967
bool ref = true;
drivers/target/iscsi/cxgbit/cxgbit_cm.c
1972
ref = false;
drivers/target/iscsi/cxgbit/cxgbit_cm.c
1996
if (ref)
drivers/target/target_core_device.c
703
static void target_non_ordered_release(struct percpu_ref *ref)
drivers/target/target_core_device.c
705
struct se_device *dev = container_of(ref, struct se_device,
drivers/target/target_core_tpg.c
436
static void core_tpg_lun_ref_release(struct percpu_ref *ref)
drivers/target/target_core_tpg.c
438
struct se_lun *lun = container_of(ref, struct se_lun, lun_ref);
drivers/target/target_core_transport.c
223
static void target_release_cmd_refcnt(struct percpu_ref *ref)
drivers/target/target_core_transport.c
225
struct target_cmd_counter *cmd_cnt = container_of(ref,
drivers/target/target_core_transport.c
3179
static void target_stop_cmd_counter_confirm(struct percpu_ref *ref)
drivers/target/target_core_transport.c
3181
struct target_cmd_counter *cmd_cnt = container_of(ref,
drivers/tee/amdtee/core.c
223
static void destroy_session(struct kref *ref)
drivers/tee/amdtee/core.c
225
struct amdtee_session *sess = container_of(ref, struct amdtee_session,
drivers/tee/tee_core.c
400
struct tee_shm_dmabuf_ref *ref;
drivers/tee/tee_core.c
402
ref = container_of(shm, struct tee_shm_dmabuf_ref, shm);
drivers/tee/tee_core.c
403
if (ref->parent_shm) {
drivers/tee/tee_core.c
412
refcount_inc(&ref->parent_shm->refcount);
drivers/tee/tee_core.c
414
shm = ref->parent_shm;
drivers/tee/tee_core.c
415
offs = ref->offset;
drivers/tee/tee_core.c
84
static void teedev_ctx_release(struct kref *ref)
drivers/tee/tee_core.c
86
struct tee_context *ctx = container_of(ref, struct tee_context,
drivers/tee/tee_shm.c
191
struct tee_shm_dmabuf_ref *ref;
drivers/tee/tee_shm.c
199
ref = kzalloc_obj(*ref);
drivers/tee/tee_shm.c
200
if (!ref) {
drivers/tee/tee_shm.c
205
refcount_set(&ref->shm.refcount, 1);
drivers/tee/tee_shm.c
206
ref->shm.ctx = ctx;
drivers/tee/tee_shm.c
207
ref->shm.id = -1;
drivers/tee/tee_shm.c
208
ref->shm.flags = TEE_SHM_DMA_BUF;
drivers/tee/tee_shm.c
210
ref->dmabuf = dma_buf_get(fd);
drivers/tee/tee_shm.c
211
if (IS_ERR(ref->dmabuf)) {
drivers/tee/tee_shm.c
212
rc = PTR_ERR(ref->dmabuf);
drivers/tee/tee_shm.c
216
rc = tee_heap_update_from_dma_buf(ctx->teedev, ref->dmabuf,
drivers/tee/tee_shm.c
217
&ref->offset, &ref->shm,
drivers/tee/tee_shm.c
218
&ref->parent_shm);
drivers/tee/tee_shm.c
222
mutex_lock(&ref->shm.ctx->teedev->mutex);
drivers/tee/tee_shm.c
223
ref->shm.id = idr_alloc(&ref->shm.ctx->teedev->idr, &ref->shm,
drivers/tee/tee_shm.c
225
mutex_unlock(&ref->shm.ctx->teedev->mutex);
drivers/tee/tee_shm.c
226
if (ref->shm.id < 0) {
drivers/tee/tee_shm.c
227
rc = ref->shm.id;
drivers/tee/tee_shm.c
231
return &ref->shm;
drivers/tee/tee_shm.c
234
dma_buf_put(ref->dmabuf);
drivers/tee/tee_shm.c
236
kfree(ref);
drivers/tee/tee_shm.c
50
struct tee_shm_dmabuf_ref *ref;
drivers/tee/tee_shm.c
52
ref = container_of(shm, struct tee_shm_dmabuf_ref, shm);
drivers/tee/tee_shm.c
53
p = ref;
drivers/tee/tee_shm.c
54
dma_buf_put(ref->dmabuf);
drivers/thermal/thermal_core.c
1758
struct thermal_zone_device *pos = NULL, *ref = ERR_PTR(-EINVAL);
drivers/thermal/thermal_core.c
1769
ref = pos;
drivers/thermal/thermal_core.c
1779
return ref;
drivers/tty/hvc/hvc_xen.c
448
int ret, evtchn, devid, ref, irq;
drivers/tty/hvc/hvc_xen.c
469
ref = gnttab_claim_grant_reference(&gref_head);
drivers/tty/hvc/hvc_xen.c
470
if (ref < 0)
drivers/tty/hvc/hvc_xen.c
471
return ref;
drivers/tty/hvc/hvc_xen.c
472
gnttab_grant_foreign_access_ref(ref, info->xbdev->otherend_id,
drivers/tty/hvc/hvc_xen.c
481
ret = xenbus_printf(xbt, dev->nodename, "ring-ref", "%d", ref);
drivers/tty/n_gsm.c
112
struct kref ref;
drivers/tty/n_gsm.c
277
struct kref ref;
drivers/tty/n_gsm.c
3235
static void gsm_free_muxr(struct kref *ref)
drivers/tty/n_gsm.c
3237
struct gsm_mux *gsm = container_of(ref, struct gsm_mux, ref);
drivers/tty/n_gsm.c
3246
kref_get(&gsm->ref);
drivers/tty/n_gsm.c
3255
kref_put(&gsm->ref, gsm_free_muxr);
drivers/tty/n_gsm.c
3294
kref_init(&gsm->ref);
drivers/tty/n_gsm.c
3922
static void net_free(struct kref *ref)
drivers/tty/n_gsm.c
3927
mux_net = container_of(ref, struct gsm_mux_net, ref);
drivers/tty/n_gsm.c
3938
kref_get(&mux_net->ref);
drivers/tty/n_gsm.c
3943
kref_put(&mux_net->ref, net_free);
drivers/tty/n_gsm.c
4073
kref_init(&mux_net->ref);
drivers/usb/core/config.c
538
void usb_release_interface_cache(struct kref *ref)
drivers/usb/core/config.c
540
struct usb_interface_cache *intfc = ref_to_usb_interface_cache(ref);
drivers/usb/core/config.c
830
kref_init(&intfc->ref);
drivers/usb/core/config.c
900
kref_put(&cf->intf_cache[i]->ref,
drivers/usb/core/message.c
1835
kref_put(&intfc->ref, usb_release_interface_cache);
drivers/usb/core/message.c
2168
kref_get(&intfc->ref);
drivers/usb/core/usb.h
32
extern void usb_release_interface_cache(struct kref *ref);
drivers/usb/gadget/function/f_fs.c
1324
static void ffs_dmabuf_release(struct kref *ref)
drivers/usb/gadget/function/f_fs.c
1326
struct ffs_dmabuf_priv *priv = container_of(ref, struct ffs_dmabuf_priv, ref);
drivers/usb/gadget/function/f_fs.c
1342
kref_get(&priv->ref);
drivers/usb/gadget/function/f_fs.c
1349
kref_put(&priv->ref, ffs_dmabuf_release);
drivers/usb/gadget/function/f_fs.c
138
struct kref ref;
drivers/usb/gadget/function/f_fs.c
1533
kref_init(&priv->ref);
drivers/usb/gadget/function/f_fs.c
2138
refcount_inc(&ffs->ref);
drivers/usb/gadget/function/f_fs.c
2143
if (refcount_dec_and_test(&ffs->ref)) {
drivers/usb/gadget/function/f_fs.c
2196
refcount_set(&ffs->ref, 1);
drivers/usb/gadget/function/tcm.h
78
struct kref ref;
drivers/usb/gadget/function/u_fs.h
177
refcount_t ref;
drivers/usb/host/xen-hcd.c
590
grant_ref_t ref;
drivers/usb/host/xen-hcd.c
603
ref = gnttab_claim_grant_reference(gref_head);
drivers/usb/host/xen-hcd.c
604
gnttab_grant_foreign_access_ref(ref, info->xbdev->otherend_id,
drivers/usb/host/xen-hcd.c
606
seg[i].gref = ref;
drivers/usb/mon/mon_main.c
210
kref_put(&mbus->ref, mon_bus_drop);
drivers/usb/mon/mon_main.c
262
struct mon_bus *mbus = container_of(r, struct mon_bus, ref);
drivers/usb/mon/mon_main.c
279
kref_init(&mbus->ref);
drivers/usb/mon/mon_main.c
306
kref_init(&mbus->ref);
drivers/usb/mon/mon_main.c
400
kref_get(&mbus->ref); /* Force leak */
drivers/usb/mon/mon_main.c
404
kref_put(&mbus->ref, mon_bus_drop);
drivers/usb/mon/mon_main.c
57
kref_get(&mbus->ref);
drivers/usb/mon/mon_main.c
76
kref_put(&mbus->ref, mon_bus_drop);
drivers/usb/mon/usb_mon.h
33
struct kref ref; /* Under mon_lock */
drivers/xen/gntdev-dmabuf.c
440
map->grants[i].ref = refs[i];
drivers/xen/gntdev.c
202
add->grants[i].ref = INVALID_GRANT_REF;
drivers/xen/gntdev.c
317
map->grants[pgnr].ref,
drivers/xen/gntdev.c
337
map->grants[i].ref,
drivers/xen/gntdev.c
363
map->grants[i].ref,
drivers/xen/gntdev.c
908
op->source.u.ref = seg->source.foreign.ref;
drivers/xen/gntdev.c
928
op->dest.u.ref = seg->dest.foreign.ref;
drivers/xen/grant-table.c
1264
foreign->gref = map_ops[i].ref;
drivers/xen/grant-table.c
139
void (*update_entry)(grant_ref_t ref, domid_t domid,
drivers/xen/grant-table.c
148
int (*end_foreign_access_ref)(grant_ref_t ref);
drivers/xen/grant-table.c
152
unsigned long (*read_frame)(grant_ref_t ref);
drivers/xen/grant-table.c
182
int ref, rc = 0;
drivers/xen/grant-table.c
193
ref = head = gnttab_free_head;
drivers/xen/grant-table.c
212
return ref;
drivers/xen/grant-table.c
331
static void put_free_entry_locked(grant_ref_t ref)
drivers/xen/grant-table.c
333
if (unlikely(ref < GNTTAB_NR_RESERVED_ENTRIES))
drivers/xen/grant-table.c
336
gnttab_entry(ref) = gnttab_free_head;
drivers/xen/grant-table.c
337
gnttab_free_head = ref;
drivers/xen/grant-table.c
339
gnttab_last_free = ref;
drivers/xen/grant-table.c
341
gnttab_free_tail_ptr = __gnttab_entry(ref);
drivers/xen/grant-table.c
343
bitmap_set(gnttab_free_bitmap, ref, 1);
drivers/xen/grant-table.c
346
static void put_free_entry(grant_ref_t ref)
drivers/xen/grant-table.c
351
put_free_entry_locked(ref);
drivers/xen/grant-table.c
384
static void gnttab_update_entry_v1(grant_ref_t ref, domid_t domid,
drivers/xen/grant-table.c
387
gnttab_shared.v1[ref].domid = domid;
drivers/xen/grant-table.c
388
gnttab_shared.v1[ref].frame = frame;
drivers/xen/grant-table.c
390
gnttab_shared.v1[ref].flags = flags;
drivers/xen/grant-table.c
393
static void gnttab_update_entry_v2(grant_ref_t ref, domid_t domid,
drivers/xen/grant-table.c
396
gnttab_shared.v2[ref].hdr.domid = domid;
drivers/xen/grant-table.c
397
gnttab_shared.v2[ref].full_page.frame = frame;
drivers/xen/grant-table.c
399
gnttab_shared.v2[ref].hdr.flags = GTF_permit_access | flags;
drivers/xen/grant-table.c
405
void gnttab_grant_foreign_access_ref(grant_ref_t ref, domid_t domid,
drivers/xen/grant-table.c
408
gnttab_interface->update_entry(ref, domid, frame,
drivers/xen/grant-table.c
416
int ref;
drivers/xen/grant-table.c
418
ref = get_free_entries(1);
drivers/xen/grant-table.c
419
if (unlikely(ref < 0))
drivers/xen/grant-table.c
422
gnttab_grant_foreign_access_ref(ref, domid, frame, readonly);
drivers/xen/grant-table.c
424
return ref;
drivers/xen/grant-table.c
428
static int gnttab_end_foreign_access_ref_v1(grant_ref_t ref)
drivers/xen/grant-table.c
430
u16 *pflags = &gnttab_shared.v1[ref].flags;
drivers/xen/grant-table.c
442
static int gnttab_end_foreign_access_ref_v2(grant_ref_t ref)
drivers/xen/grant-table.c
444
gnttab_shared.v2[ref].hdr.flags = 0;
drivers/xen/grant-table.c
446
if (grstatus[ref] & (GTF_reading|GTF_writing)) {
drivers/xen/grant-table.c
465
static inline int _gnttab_end_foreign_access_ref(grant_ref_t ref)
drivers/xen/grant-table.c
467
return gnttab_interface->end_foreign_access_ref(ref);
drivers/xen/grant-table.c
470
int gnttab_end_foreign_access_ref(grant_ref_t ref)
drivers/xen/grant-table.c
472
if (_gnttab_end_foreign_access_ref(ref))
drivers/xen/grant-table.c
474
pr_warn("WARNING: g.e. %#x still in use!\n", ref);
drivers/xen/grant-table.c
479
static unsigned long gnttab_read_frame_v1(grant_ref_t ref)
drivers/xen/grant-table.c
481
return gnttab_shared.v1[ref].frame;
drivers/xen/grant-table.c
484
static unsigned long gnttab_read_frame_v2(grant_ref_t ref)
drivers/xen/grant-table.c
486
return gnttab_shared.v2[ref].full_page.frame;
drivers/xen/grant-table.c
491
grant_ref_t ref;
drivers/xen/grant-table.c
522
if (_gnttab_end_foreign_access_ref(entry->ref)) {
drivers/xen/grant-table.c
525
put_free_entry(entry->ref);
drivers/xen/grant-table.c
527
entry->ref, page_to_pfn(entry->page),
drivers/xen/grant-table.c
535
pr_info("g.e. %#x still pending\n", entry->ref);
drivers/xen/grant-table.c
553
static void gnttab_add_deferred(grant_ref_t ref, struct page *page)
drivers/xen/grant-table.c
561
unsigned long gfn = gnttab_interface->read_frame(ref);
drivers/xen/grant-table.c
570
entry->ref = ref;
drivers/xen/grant-table.c
583
ref, page ? page_to_pfn(page) : -1, deferred, leaked);
drivers/xen/grant-table.c
588
ref, page ? page_to_pfn(page) : -1, deferred, leaked);
drivers/xen/grant-table.c
592
int gnttab_try_end_foreign_access(grant_ref_t ref)
drivers/xen/grant-table.c
594
int ret = _gnttab_end_foreign_access_ref(ref);
drivers/xen/grant-table.c
597
put_free_entry(ref);
drivers/xen/grant-table.c
603
void gnttab_end_foreign_access(grant_ref_t ref, struct page *page)
drivers/xen/grant-table.c
605
if (gnttab_try_end_foreign_access(ref)) {
drivers/xen/grant-table.c
609
gnttab_add_deferred(ref, page);
drivers/xen/grant-table.c
613
void gnttab_free_grant_reference(grant_ref_t ref)
drivers/xen/grant-table.c
615
put_free_entry(ref);
drivers/xen/grant-table.c
621
grant_ref_t ref;
drivers/xen/grant-table.c
626
ref = gnttab_entry(head);
drivers/xen/grant-table.c
628
head = ref;
drivers/xen/pvcalls-back.c
319
grant_ref_t ref,
drivers/xen/pvcalls-back.c
336
map->ref = ref;
drivers/xen/pvcalls-back.c
338
ret = xenbus_map_ring_valloc(fedata->dev, &ref, 1, &page);
drivers/xen/pvcalls-back.c
350
ret = xenbus_map_ring_valloc(fedata->dev, map->ring->ref,
drivers/xen/pvcalls-back.c
420
req->u.connect.ref,
drivers/xen/pvcalls-back.c
563
req->u.accept.ref,
drivers/xen/pvcalls-back.c
60
grant_ref_t ref;
drivers/xen/pvcalls-front.c
1137
if (bedata->ref != -1)
drivers/xen/pvcalls-front.c
1138
gnttab_end_foreign_access(bedata->ref, NULL);
drivers/xen/pvcalls-front.c
1192
bedata->ref = -1;
drivers/xen/pvcalls-front.c
1222
bedata->ref = ret;
drivers/xen/pvcalls-front.c
1223
gnttab_grant_foreign_access_ref(bedata->ref, dev->otherend_id,
drivers/xen/pvcalls-front.c
1235
ret = xenbus_printf(xbt, dev->nodename, "ring-ref", "%d", bedata->ref);
drivers/xen/pvcalls-front.c
245
gnttab_end_foreign_access(map->active.ring->ref[i], NULL);
drivers/xen/pvcalls-front.c
246
gnttab_end_foreign_access(map->active.ref, NULL);
drivers/xen/pvcalls-front.c
33
grant_ref_t ref;
drivers/xen/pvcalls-front.c
392
map->active.ring->ref[i] = gnttab_grant_foreign_access(
drivers/xen/pvcalls-front.c
396
map->active.ref = gnttab_grant_foreign_access(
drivers/xen/pvcalls-front.c
467
req->u.connect.ref = map->active.ref;
drivers/xen/pvcalls-front.c
64
grant_ref_t ref;
drivers/xen/pvcalls-front.c
865
req->u.accept.ref = map2->active.ref;
fs/afs/cell.c
163
refcount_set(&cell->ref, 1);
fs/afs/cell.c
573
r = refcount_read(&cell->ref);
fs/afs/cell.c
605
__refcount_inc(&cell->ref, &r);
fs/afs/cell.c
622
zero = __refcount_dec_and_test(&cell->ref, &r);
fs/afs/cell.c
639
__refcount_inc(&cell->ref, &r);
fs/afs/cell.c
673
zero = __refcount_dec_and_test(&cell->ref, &r);
fs/afs/cell.c
686
r = refcount_read(&cell->ref);
fs/afs/cell.c
707
if (refcount_read(&cell->ref) > 0 && cell->net->live)
fs/afs/cmservice.c
176
refcount_read(&call->server->ref),
fs/afs/fs_probe.c
24
__refcount_inc(&estate->ref, &r);
fs/afs/fs_probe.c
251
refcount_set(&estate->ref, 2);
fs/afs/fs_probe.c
280
trace_afs_estate(estate->server_id, estate->probe_seq, refcount_read(&estate->ref),
fs/afs/fs_probe.c
34
trace_afs_estate(estate->server_id, estate->probe_seq, refcount_read(&estate->ref),
fs/afs/fs_probe.c
47
dead = __refcount_dec_and_test(&estate->ref, &r);
fs/afs/internal.h
1400
__refcount_inc(&call->ref, &r);
fs/afs/internal.h
1410
int r = refcount_read(&call->ref);
fs/afs/internal.h
1555
int r = refcount_read(&server->ref);
fs/afs/internal.h
162
refcount_t ref;
fs/afs/internal.h
1651
refcount_inc(&vlserver->ref);
fs/afs/internal.h
1658
refcount_inc(&vllist->ref);
fs/afs/internal.h
386
refcount_t ref; /* Struct refcount */
fs/afs/internal.h
431
refcount_t ref;
fs/afs/internal.h
469
refcount_t ref;
fs/afs/internal.h
518
refcount_t ref;
fs/afs/internal.h
568
refcount_t ref; /* Object refcount */
fs/afs/internal.h
626
refcount_t ref;
fs/afs/proc.c
278
refcount_read(&vol->ref), vol->vid,
fs/afs/proc.c
450
refcount_read(&server->ref),
fs/afs/proc.c
50
refcount_read(&cell->ref),
fs/afs/rxrpc.c
170
refcount_set(&call->ref, 1);
fs/afs/rxrpc.c
224
zero = __refcount_dec_and_test(&call->ref, &r);
fs/afs/rxrpc.c
250
zero = __refcount_dec_and_test(&call->ref, &r);
fs/afs/rxrpc.c
700
if (__refcount_inc_not_zero(&call->ref, &r)) {
fs/afs/server.c
124
refcount_set(&server->ref, 1);
fs/afs/server.c
302
__refcount_inc(&server->ref, &r);
fs/afs/server.c
318
__refcount_inc(&server->ref, &r);
fs/afs/server.c
343
zero = __refcount_dec_and_test(&server->ref, &r);
fs/afs/server.c
396
trace_afs_server(server->debug_id, refcount_read(&server->ref),
fs/afs/server.c
431
trace_afs_server(server->debug_id, refcount_read(&server->ref),
fs/afs/server.c
552
trace_afs_server(server->debug_id, refcount_read(&server->ref),
fs/afs/vl_list.c
20
refcount_set(&vlserver->ref, 1);
fs/afs/vl_list.c
46
refcount_dec_and_test(&vlserver->ref))
fs/afs/vl_list.c
56
refcount_set(&vllist->ref, 1);
fs/afs/vl_list.c
66
if (refcount_dec_and_test(&vllist->ref)) {
fs/afs/volume.c
252
trace_afs_volume(volume->debug_id, volume->vid, refcount_read(&volume->ref),
fs/afs/volume.c
266
if (__refcount_inc_not_zero(&volume->ref, &r)) {
fs/afs/volume.c
282
__refcount_inc(&volume->ref, &r);
fs/afs/volume.c
300
zero = __refcount_dec_and_test(&volume->ref, &r);
fs/afs/volume.c
63
trace_afs_volume(volume->debug_id, volume->vid, refcount_read(&volume->ref),
fs/afs/volume.c
98
refcount_set(&volume->ref, 1);
fs/aio.c
628
static void free_ioctx_reqs(struct percpu_ref *ref)
fs/aio.c
630
struct kioctx *ctx = container_of(ref, struct kioctx, reqs);
fs/aio.c
646
static void free_ioctx_users(struct percpu_ref *ref)
fs/aio.c
648
struct kioctx *ctx = container_of(ref, struct kioctx, users);
fs/backing-file.c
100
if (refcount_dec_and_test(&aio->ref)) {
fs/backing-file.c
179
refcount_set(&aio->ref, 2);
fs/backing-file.c
242
refcount_set(&aio->ref, 2);
fs/backing-file.c
80
refcount_t ref;
fs/btrfs/backref.c
1387
struct prelim_ref *ref;
fs/btrfs/backref.c
1591
ref = rb_entry(node, struct prelim_ref, rbnode);
fs/btrfs/backref.c
1592
node = rb_next(&ref->rbnode);
fs/btrfs/backref.c
1603
if (ctx->roots && ref->count && ref->root_id && ref->parent == 0) {
fs/btrfs/backref.c
1605
ret = ulist_add(ctx->roots, ref->root_id, 0, GFP_NOFS);
fs/btrfs/backref.c
1609
if (ref->count && ref->parent) {
fs/btrfs/backref.c
1610
if (!ctx->skip_inode_ref_list && !ref->inode_list &&
fs/btrfs/backref.c
1611
ref->level == 0) {
fs/btrfs/backref.c
1615
check.level = ref->level;
fs/btrfs/backref.c
1617
eb = read_tree_block(ctx->fs_info, ref->parent,
fs/btrfs/backref.c
1638
ref->inode_list = eie;
fs/btrfs/backref.c
1646
ret = ulist_add_merge_ptr(ctx->refs, ref->parent,
fs/btrfs/backref.c
1647
ref->inode_list,
fs/btrfs/backref.c
1667
eie->next = ref->inode_list;
fs/btrfs/backref.c
1677
ref->inode_list = NULL;
fs/btrfs/backref.c
212
static void free_pref(struct prelim_ref *ref)
fs/btrfs/backref.c
214
kmem_cache_free(btrfs_prelim_ref_cache, ref);
fs/btrfs/backref.c
301
struct prelim_ref *ref = rb_entry(exist, struct prelim_ref, rbnode);
fs/btrfs/backref.c
303
struct extent_inode_elem *eie = ref->inode_list;
fs/btrfs/backref.c
309
ref->inode_list = newref->inode_list;
fs/btrfs/backref.c
312
trace_btrfs_prelim_ref_merge(fs_info, ref, newref,
fs/btrfs/backref.c
319
update_share_count(sc, ref->count,
fs/btrfs/backref.c
320
ref->count + newref->count, newref);
fs/btrfs/backref.c
321
ref->count += newref->count;
fs/btrfs/backref.c
337
struct prelim_ref *ref, *next_ref;
fs/btrfs/backref.c
339
rbtree_postorder_for_each_entry_safe(ref, next_ref,
fs/btrfs/backref.c
341
free_inode_elem_list(ref->inode_list);
fs/btrfs/backref.c
342
free_pref(ref);
fs/btrfs/backref.c
393
struct prelim_ref *ref;
fs/btrfs/backref.c
398
ref = kmem_cache_alloc(btrfs_prelim_ref_cache, gfp_mask);
fs/btrfs/backref.c
399
if (!ref)
fs/btrfs/backref.c
402
ref->root_id = root_id;
fs/btrfs/backref.c
404
ref->key_for_search = *key;
fs/btrfs/backref.c
406
memset(&ref->key_for_search, 0, sizeof(ref->key_for_search));
fs/btrfs/backref.c
408
ref->inode_list = NULL;
fs/btrfs/backref.c
409
ref->level = level;
fs/btrfs/backref.c
410
ref->count = count;
fs/btrfs/backref.c
411
ref->parent = parent;
fs/btrfs/backref.c
412
ref->wanted_disk_byte = wanted_disk_byte;
fs/btrfs/backref.c
413
prelim_ref_insert(fs_info, preftree, ref, sc);
fs/btrfs/backref.c
446
struct prelim_ref *ref = NULL;
fs/btrfs/backref.c
454
ref = rb_entry(parent, struct prelim_ref, rbnode);
fs/btrfs/backref.c
455
result = prelim_ref_compare(ref, &target);
fs/btrfs/backref.c
470
struct preftrees *preftrees, struct prelim_ref *ref,
fs/btrfs/backref.c
477
struct btrfs_key *key_for_search = &ref->key_for_search;
fs/btrfs/backref.c
481
u64 wanted_disk_byte = ref->wanted_disk_byte;
fs/btrfs/backref.c
507
ref->root_id != btrfs_header_owner(eb)) {
fs/btrfs/backref.c
514
while (!ret && count < ref->count) {
fs/btrfs/backref.c
531
ref->root_id != btrfs_header_owner(eb))) {
fs/btrfs/backref.c
548
if (ref->key_for_search.offset == key.offset - data_offset)
fs/btrfs/backref.c
593
struct prelim_ref *ref, struct ulist *parents)
fs/btrfs/backref.c
599
int level = ref->level;
fs/btrfs/backref.c
600
struct btrfs_key search_key = ref->key_for_search;
fs/btrfs/backref.c
611
root = btrfs_get_fs_root_commit_root(ctx->fs_info, path, ref->root_id);
fs/btrfs/backref.c
613
root = btrfs_get_fs_root(ctx->fs_info, ref->root_id, false);
fs/btrfs/backref.c
670
ref->root_id, level, ref->count, ret,
fs/btrfs/backref.c
671
BTRFS_KEY_FMT_VALUE(&ref->key_for_search));
fs/btrfs/backref.c
685
ret = add_all_parents(ctx, root, path, parents, preftrees, ref, level);
fs/btrfs/backref.c
752
struct prelim_ref *ref;
fs/btrfs/backref.c
755
ref = rb_entry(rnode, struct prelim_ref, rbnode);
fs/btrfs/backref.c
756
if (WARN(ref->parent,
fs/btrfs/backref.c
762
rb_erase_cached(&ref->rbnode, &preftrees->indirect.root);
fs/btrfs/backref.c
765
if (ref->count == 0) {
fs/btrfs/backref.c
766
free_pref(ref);
fs/btrfs/backref.c
770
if (sc && ref->root_id != btrfs_root_id(sc->root)) {
fs/btrfs/backref.c
771
free_pref(ref);
fs/btrfs/backref.c
775
ret2 = resolve_indirect_ref(ctx, path, preftrees, ref, parents);
fs/btrfs/backref.c
781
prelim_ref_insert(ctx->fs_info, &preftrees->direct, ref,
fs/btrfs/backref.c
785
free_pref(ref);
fs/btrfs/backref.c
793
ref->parent = node ? node->val : 0;
fs/btrfs/backref.c
794
ref->inode_list = unode_aux_to_inode_list(node);
fs/btrfs/backref.c
803
free_pref(ref);
fs/btrfs/backref.c
807
memcpy(new_ref, ref, sizeof(*ref));
fs/btrfs/backref.c
818
prelim_ref_insert(ctx->fs_info, &preftrees->direct, ref, NULL);
fs/btrfs/backref.c
838
struct prelim_ref *ref;
fs/btrfs/backref.c
846
ref = rb_entry(node, struct prelim_ref, rbnode);
fs/btrfs/backref.c
849
BUG_ON(ref->parent); /* should not be a direct ref */
fs/btrfs/backref.c
850
BUG_ON(ref->key_for_search.type);
fs/btrfs/backref.c
851
BUG_ON(!ref->wanted_disk_byte);
fs/btrfs/backref.c
853
check.level = ref->level - 1;
fs/btrfs/backref.c
854
check.owner_root = ref->root_id;
fs/btrfs/backref.c
856
eb = read_tree_block(fs_info, ref->wanted_disk_byte, &check);
fs/btrfs/backref.c
858
free_pref(ref);
fs/btrfs/backref.c
862
free_pref(ref);
fs/btrfs/backref.c
870
btrfs_item_key_to_cpu(eb, &ref->key_for_search, 0);
fs/btrfs/backref.c
872
btrfs_node_key_to_cpu(eb, &ref->key_for_search, 0);
fs/btrfs/backref.c
876
prelim_ref_insert(fs_info, &preftrees->indirect, ref, NULL);
fs/btrfs/delayed-ref.c
1182
void btrfs_put_delayed_ref(struct btrfs_delayed_ref_node *ref)
fs/btrfs/delayed-ref.c
1184
if (refcount_dec_and_test(&ref->refs)) {
fs/btrfs/delayed-ref.c
1185
WARN_ON(!RB_EMPTY_NODE(&ref->ref_node));
fs/btrfs/delayed-ref.c
1186
kmem_cache_free(btrfs_delayed_ref_node_cachep, ref);
fs/btrfs/delayed-ref.c
1293
struct btrfs_delayed_ref_node *ref;
fs/btrfs/delayed-ref.c
1295
ref = rb_entry(n, struct btrfs_delayed_ref_node, ref_node);
fs/btrfs/delayed-ref.c
1296
drop_delayed_ref(fs_info, delayed_refs, head, ref);
fs/btrfs/delayed-ref.c
373
struct btrfs_delayed_ref_node *ref)
fs/btrfs/delayed-ref.c
376
rb_erase_cached(&ref->ref_node, &head->ref_tree);
fs/btrfs/delayed-ref.c
377
RB_CLEAR_NODE(&ref->ref_node);
fs/btrfs/delayed-ref.c
378
if (!list_empty(&ref->add_list))
fs/btrfs/delayed-ref.c
379
list_del(&ref->add_list);
fs/btrfs/delayed-ref.c
380
btrfs_put_delayed_ref(ref);
fs/btrfs/delayed-ref.c
387
struct btrfs_delayed_ref_node *ref,
fs/btrfs/delayed-ref.c
391
struct rb_node *node = rb_next(&ref->ref_node);
fs/btrfs/delayed-ref.c
401
if (comp_refs(ref, next, false))
fs/btrfs/delayed-ref.c
404
if (ref->action == next->action) {
fs/btrfs/delayed-ref.c
407
if (ref->ref_mod < next->ref_mod) {
fs/btrfs/delayed-ref.c
408
swap(ref, next);
fs/btrfs/delayed-ref.c
415
ref->ref_mod += mod;
fs/btrfs/delayed-ref.c
416
if (ref->ref_mod == 0) {
fs/btrfs/delayed-ref.c
417
drop_delayed_ref(fs_info, delayed_refs, head, ref);
fs/btrfs/delayed-ref.c
423
WARN_ON(ref->type == BTRFS_TREE_BLOCK_REF_KEY ||
fs/btrfs/delayed-ref.c
424
ref->type == BTRFS_SHARED_BLOCK_REF_KEY);
fs/btrfs/delayed-ref.c
435
struct btrfs_delayed_ref_node *ref;
fs/btrfs/delayed-ref.c
452
ref = rb_entry(node, struct btrfs_delayed_ref_node, ref_node);
fs/btrfs/delayed-ref.c
453
if (seq && ref->seq >= seq)
fs/btrfs/delayed-ref.c
455
if (merge_ref(fs_info, delayed_refs, head, ref, seq))
fs/btrfs/delayed-ref.c
551
struct btrfs_delayed_ref_node *ref;
fs/btrfs/delayed-ref.c
569
ref = rb_entry(rb_first_cached(&head->ref_tree),
fs/btrfs/delayed-ref.c
571
ASSERT(list_empty(&ref->add_list));
fs/btrfs/delayed-ref.c
572
return ref;
fs/btrfs/delayed-ref.c
584
struct btrfs_delayed_ref_node *ref)
fs/btrfs/delayed-ref.c
591
exist = tree_insert(&href->ref_tree, ref);
fs/btrfs/delayed-ref.c
593
if (ref->action == BTRFS_ADD_DELAYED_REF)
fs/btrfs/delayed-ref.c
594
list_add_tail(&ref->add_list, &href->ref_add_list);
fs/btrfs/delayed-ref.c
601
if (exist->action == ref->action) {
fs/btrfs/delayed-ref.c
602
mod = ref->ref_mod;
fs/btrfs/delayed-ref.c
605
if (exist->ref_mod < ref->ref_mod) {
fs/btrfs/delayed-ref.c
606
exist->action = ref->action;
fs/btrfs/delayed-ref.c
608
exist->ref_mod = ref->ref_mod;
fs/btrfs/delayed-ref.c
609
if (ref->action == BTRFS_ADD_DELAYED_REF)
fs/btrfs/delayed-ref.c
612
else if (ref->action == BTRFS_DROP_DELAYED_REF) {
fs/btrfs/delayed-ref.c
619
mod = -ref->ref_mod;
fs/btrfs/delayed-ref.c
935
struct btrfs_delayed_ref_node *ref,
fs/btrfs/delayed-ref.c
947
refcount_set(&ref->refs, 1);
fs/btrfs/delayed-ref.c
948
ref->bytenr = generic_ref->bytenr;
fs/btrfs/delayed-ref.c
949
ref->num_bytes = generic_ref->num_bytes;
fs/btrfs/delayed-ref.c
950
ref->ref_mod = 1;
fs/btrfs/delayed-ref.c
951
ref->action = action;
fs/btrfs/delayed-ref.c
952
ref->seq = seq;
fs/btrfs/delayed-ref.c
953
ref->type = btrfs_ref_type(generic_ref);
fs/btrfs/delayed-ref.c
954
ref->ref_root = generic_ref->ref_root;
fs/btrfs/delayed-ref.c
955
ref->parent = generic_ref->parent;
fs/btrfs/delayed-ref.c
956
RB_CLEAR_NODE(&ref->ref_node);
fs/btrfs/delayed-ref.c
957
INIT_LIST_HEAD(&ref->add_list);
fs/btrfs/delayed-ref.c
960
ref->data_ref = generic_ref->data_ref;
fs/btrfs/delayed-ref.c
962
ref->tree_ref = generic_ref->tree_ref;
fs/btrfs/delayed-ref.h
359
void btrfs_put_delayed_ref(struct btrfs_delayed_ref_node *ref);
fs/btrfs/delayed-ref.h
440
static inline u8 btrfs_ref_type(const struct btrfs_ref *ref)
fs/btrfs/delayed-ref.h
442
ASSERT(ref->type == BTRFS_REF_DATA || ref->type == BTRFS_REF_METADATA);
fs/btrfs/delayed-ref.h
444
if (ref->type == BTRFS_REF_DATA) {
fs/btrfs/delayed-ref.h
445
if (ref->parent)
fs/btrfs/delayed-ref.h
450
if (ref->parent)
fs/btrfs/export.c
160
struct btrfs_root_ref *ref;
fs/btrfs/export.c
207
ref = btrfs_item_ptr(leaf, path->slots[0],
fs/btrfs/export.c
209
key.objectid = btrfs_root_ref_dirid(leaf, ref);
fs/btrfs/extent-tree.c
2004
struct btrfs_delayed_ref_node *ref;
fs/btrfs/extent-tree.c
2013
while ((ref = btrfs_select_delayed_ref(locked_ref))) {
fs/btrfs/extent-tree.c
2014
if (ref->seq &&
fs/btrfs/extent-tree.c
2015
btrfs_check_delayed_seq(fs_info, ref->seq)) {
fs/btrfs/extent-tree.c
2021
rb_erase_cached(&ref->ref_node, &locked_ref->ref_tree);
fs/btrfs/extent-tree.c
2022
RB_CLEAR_NODE(&ref->ref_node);
fs/btrfs/extent-tree.c
2023
if (!list_empty(&ref->add_list))
fs/btrfs/extent-tree.c
2024
list_del(&ref->add_list);
fs/btrfs/extent-tree.c
2029
switch (ref->action) {
fs/btrfs/extent-tree.c
2032
locked_ref->ref_mod -= ref->ref_mod;
fs/btrfs/extent-tree.c
2035
locked_ref->ref_mod += ref->ref_mod;
fs/btrfs/extent-tree.c
2058
ret = run_one_delayed_ref(trans, locked_ref, ref, extent_op,
fs/btrfs/extent-tree.c
2066
btrfs_put_delayed_ref(ref);
fs/btrfs/extent-tree.c
2070
btrfs_put_delayed_ref(ref);
fs/btrfs/extent-tree.c
2295
struct btrfs_delayed_ref_node *ref;
fs/btrfs/extent-tree.c
2352
ref = rb_entry(node, struct btrfs_delayed_ref_node, ref_node);
fs/btrfs/extent-tree.c
2354
if (ref->type != BTRFS_EXTENT_DATA_REF_KEY) {
fs/btrfs/extent-tree.c
2359
ref_owner = btrfs_delayed_ref_owner(ref);
fs/btrfs/extent-tree.c
2360
ref_offset = btrfs_delayed_ref_offset(ref);
fs/btrfs/extent-tree.c
2366
if (ref->ref_root != btrfs_root_id(root) ||
fs/btrfs/extent-tree.c
2428
struct btrfs_extent_data_ref *ref;
fs/btrfs/extent-tree.c
2493
ref = (struct btrfs_extent_data_ref *)(&iref->offset);
fs/btrfs/extent-tree.c
2495
btrfs_extent_data_ref_count(leaf, ref) ||
fs/btrfs/extent-tree.c
2496
btrfs_extent_data_ref_root(leaf, ref) != btrfs_root_id(root) ||
fs/btrfs/extent-tree.c
2497
btrfs_extent_data_ref_objectid(leaf, ref) != btrfs_ino(inode) ||
fs/btrfs/extent-tree.c
2498
btrfs_extent_data_ref_offset(leaf, ref) != offset)
fs/btrfs/extent-tree.c
2586
struct btrfs_ref ref = {
fs/btrfs/extent-tree.c
2601
ref.bytenr = btrfs_file_extent_disk_bytenr(buf, fi);
fs/btrfs/extent-tree.c
2602
if (ref.bytenr == 0)
fs/btrfs/extent-tree.c
2605
ref.num_bytes = btrfs_file_extent_disk_num_bytes(buf, fi);
fs/btrfs/extent-tree.c
2606
ref.owning_root = ref_root;
fs/btrfs/extent-tree.c
2609
btrfs_init_data_ref(&ref, key.objectid, key.offset,
fs/btrfs/extent-tree.c
2612
ret = btrfs_inc_extent_ref(trans, &ref);
fs/btrfs/extent-tree.c
2614
ret = btrfs_free_extent(trans, &ref);
fs/btrfs/extent-tree.c
2619
ref.bytenr = btrfs_node_blockptr(buf, i);
fs/btrfs/extent-tree.c
2620
ref.num_bytes = fs_info->nodesize;
fs/btrfs/extent-tree.c
2622
btrfs_init_tree_ref(&ref, level - 1,
fs/btrfs/extent-tree.c
2625
ret = btrfs_inc_extent_ref(trans, &ref);
fs/btrfs/extent-tree.c
2627
ret = btrfs_free_extent(trans, &ref);
fs/btrfs/extent-tree.c
3706
int btrfs_free_extent(struct btrfs_trans_handle *trans, struct btrfs_ref *ref)
fs/btrfs/extent-tree.c
3718
if (ref->ref_root == BTRFS_TREE_LOG_OBJECTID) {
fs/btrfs/extent-tree.c
3719
btrfs_pin_extent(trans, ref->bytenr, ref->num_bytes);
fs/btrfs/extent-tree.c
3721
} else if (ref->type == BTRFS_REF_METADATA) {
fs/btrfs/extent-tree.c
3722
ret = btrfs_add_delayed_tree_ref(trans, ref, NULL);
fs/btrfs/extent-tree.c
3724
ret = btrfs_add_delayed_data_ref(trans, ref, 0);
fs/btrfs/extent-tree.c
3727
if (ref->ref_root != BTRFS_TREE_LOG_OBJECTID)
fs/btrfs/extent-tree.c
3728
btrfs_ref_tree_mod(fs_info, ref);
fs/btrfs/extent-tree.c
419
const struct btrfs_extent_data_ref *ref)
fs/btrfs/extent-tree.c
421
return hash_extent_data_ref(btrfs_extent_data_ref_root(leaf, ref),
fs/btrfs/extent-tree.c
422
btrfs_extent_data_ref_objectid(leaf, ref),
fs/btrfs/extent-tree.c
423
btrfs_extent_data_ref_offset(leaf, ref));
fs/btrfs/extent-tree.c
427
const struct btrfs_extent_data_ref *ref,
fs/btrfs/extent-tree.c
430
if (btrfs_extent_data_ref_root(leaf, ref) != root_objectid ||
fs/btrfs/extent-tree.c
431
btrfs_extent_data_ref_objectid(leaf, ref) != owner ||
fs/btrfs/extent-tree.c
432
btrfs_extent_data_ref_offset(leaf, ref) != offset)
fs/btrfs/extent-tree.c
445
struct btrfs_extent_data_ref *ref;
fs/btrfs/extent-tree.c
500
ref = btrfs_item_ptr(leaf, path->slots[0],
fs/btrfs/extent-tree.c
503
if (match_extent_data_ref(leaf, ref, root_objectid,
fs/btrfs/extent-tree.c
5052
struct btrfs_shared_data_ref *ref;
fs/btrfs/extent-tree.c
5053
ref = (struct btrfs_shared_data_ref *)(iref + 1);
fs/btrfs/extent-tree.c
5055
btrfs_set_shared_data_ref_count(leaf, ref, ref_mod);
fs/btrfs/extent-tree.c
5057
struct btrfs_extent_data_ref *ref;
fs/btrfs/extent-tree.c
5058
ref = (struct btrfs_extent_data_ref *)(&iref->offset);
fs/btrfs/extent-tree.c
5059
btrfs_set_extent_data_ref_root(leaf, ref, root_objectid);
fs/btrfs/extent-tree.c
5060
btrfs_set_extent_data_ref_objectid(leaf, ref, owner);
fs/btrfs/extent-tree.c
5061
btrfs_set_extent_data_ref_offset(leaf, ref, offset);
fs/btrfs/extent-tree.c
5062
btrfs_set_extent_data_ref_count(leaf, ref, ref_mod);
fs/btrfs/extent-tree.c
554
struct btrfs_shared_data_ref *ref;
fs/btrfs/extent-tree.c
555
ref = btrfs_item_ptr(leaf, path->slots[0],
fs/btrfs/extent-tree.c
558
btrfs_set_shared_data_ref_count(leaf, ref, node->ref_mod);
fs/btrfs/extent-tree.c
560
num_refs = btrfs_shared_data_ref_count(leaf, ref);
fs/btrfs/extent-tree.c
562
btrfs_set_shared_data_ref_count(leaf, ref, num_refs);
fs/btrfs/extent-tree.c
565
struct btrfs_extent_data_ref *ref;
fs/btrfs/extent-tree.c
567
ref = btrfs_item_ptr(leaf, path->slots[0],
fs/btrfs/extent-tree.c
569
if (match_extent_data_ref(leaf, ref, node->ref_root,
fs/btrfs/extent-tree.c
581
ref = btrfs_item_ptr(leaf, path->slots[0],
fs/btrfs/extent-tree.c
5823
struct btrfs_ref ref = {
fs/btrfs/extent-tree.c
5838
ref.parent = path->nodes[level]->start;
fs/btrfs/extent-tree.c
584
btrfs_set_extent_data_ref_root(leaf, ref, node->ref_root);
fs/btrfs/extent-tree.c
585
btrfs_set_extent_data_ref_objectid(leaf, ref, owner);
fs/btrfs/extent-tree.c
5854
ret = check_ref_exists(trans, root, next->start, ref.parent,
fs/btrfs/extent-tree.c
586
btrfs_set_extent_data_ref_offset(leaf, ref, offset);
fs/btrfs/extent-tree.c
587
btrfs_set_extent_data_ref_count(leaf, ref, node->ref_mod);
fs/btrfs/extent-tree.c
5889
btrfs_init_tree_ref(&ref, level - 1, 0, false);
fs/btrfs/extent-tree.c
589
num_refs = btrfs_extent_data_ref_count(leaf, ref);
fs/btrfs/extent-tree.c
5890
return btrfs_free_extent(trans, &ref);
fs/btrfs/extent-tree.c
591
btrfs_set_extent_data_ref_count(leaf, ref, num_refs);
fs/btrfs/extent-tree.h
147
int btrfs_free_extent(struct btrfs_trans_handle *trans, struct btrfs_ref *ref);
fs/btrfs/file.c
2368
struct btrfs_ref ref = {
fs/btrfs/file.c
2378
btrfs_init_data_ref(&ref, btrfs_ino(inode), ref_offset, 0, false);
fs/btrfs/file.c
2379
ret = btrfs_inc_extent_ref(trans, &ref);
fs/btrfs/file.c
318
struct btrfs_ref ref = {
fs/btrfs/file.c
326
btrfs_init_data_ref(&ref, new_key.objectid,
fs/btrfs/file.c
329
ret = btrfs_inc_extent_ref(trans, &ref);
fs/btrfs/file.c
417
struct btrfs_ref ref = {
fs/btrfs/file.c
425
btrfs_init_data_ref(&ref, key.objectid,
fs/btrfs/file.c
428
ret = btrfs_free_extent(trans, &ref);
fs/btrfs/file.c
556
struct btrfs_ref ref = { 0 };
fs/btrfs/file.c
698
ref.action = BTRFS_ADD_DELAYED_REF;
fs/btrfs/file.c
699
ref.bytenr = bytenr;
fs/btrfs/file.c
700
ref.num_bytes = num_bytes;
fs/btrfs/file.c
701
ref.parent = 0;
fs/btrfs/file.c
702
ref.owning_root = btrfs_root_id(root);
fs/btrfs/file.c
703
ref.ref_root = btrfs_root_id(root);
fs/btrfs/file.c
704
btrfs_init_data_ref(&ref, ino, orig_offset, 0, false);
fs/btrfs/file.c
705
ret = btrfs_inc_extent_ref(trans, &ref);
fs/btrfs/file.c
728
ref.action = BTRFS_DROP_DELAYED_REF;
fs/btrfs/file.c
729
ref.bytenr = bytenr;
fs/btrfs/file.c
730
ref.num_bytes = num_bytes;
fs/btrfs/file.c
731
ref.parent = 0;
fs/btrfs/file.c
732
ref.owning_root = btrfs_root_id(root);
fs/btrfs/file.c
733
ref.ref_root = btrfs_root_id(root);
fs/btrfs/file.c
734
btrfs_init_data_ref(&ref, ino, orig_offset, 0, false);
fs/btrfs/file.c
745
ret = btrfs_free_extent(trans, &ref);
fs/btrfs/file.c
763
ret = btrfs_free_extent(trans, &ref);
fs/btrfs/inode-item.c
172
struct btrfs_inode_ref *ref;
fs/btrfs/inode-item.c
180
int del_len = name->len + sizeof(*ref);
fs/btrfs/inode-item.c
199
ref = btrfs_find_name_in_backref(path->nodes[0], path->slots[0], name);
fs/btrfs/inode-item.c
200
if (!ref) {
fs/btrfs/inode-item.c
209
*index = btrfs_inode_ref_index(leaf, ref);
fs/btrfs/inode-item.c
21
struct btrfs_inode_ref *ref;
fs/btrfs/inode-item.c
215
ptr = (unsigned long)ref;
fs/btrfs/inode-item.c
216
sub_item_len = name->len + sizeof(*ref);
fs/btrfs/inode-item.c
302
struct btrfs_inode_ref *ref;
fs/btrfs/inode-item.c
305
int ins_len = name->len + sizeof(*ref);
fs/btrfs/inode-item.c
31
ref = (struct btrfs_inode_ref *)(ptr + cur_offset);
fs/btrfs/inode-item.c
32
len = btrfs_inode_ref_name_len(leaf, ref);
fs/btrfs/inode-item.c
320
ref = btrfs_find_name_in_backref(path->nodes[0], path->slots[0],
fs/btrfs/inode-item.c
322
if (ref)
fs/btrfs/inode-item.c
327
ref = btrfs_item_ptr(path->nodes[0], path->slots[0],
fs/btrfs/inode-item.c
329
ref = (struct btrfs_inode_ref *)((unsigned long)ref + old_size);
fs/btrfs/inode-item.c
33
name_ptr = (unsigned long)(ref + 1);
fs/btrfs/inode-item.c
330
btrfs_set_inode_ref_name_len(path->nodes[0], ref, name->len);
fs/btrfs/inode-item.c
331
btrfs_set_inode_ref_index(path->nodes[0], ref, index);
fs/btrfs/inode-item.c
332
ptr = (unsigned long)(ref + 1);
fs/btrfs/inode-item.c
34
cur_offset += len + sizeof(*ref);
fs/btrfs/inode-item.c
345
ref = btrfs_item_ptr(path->nodes[0], path->slots[0],
fs/btrfs/inode-item.c
347
btrfs_set_inode_ref_name_len(path->nodes[0], ref, name->len);
fs/btrfs/inode-item.c
348
btrfs_set_inode_ref_index(path->nodes[0], ref, index);
fs/btrfs/inode-item.c
349
ptr = (unsigned long)(ref + 1);
fs/btrfs/inode-item.c
39
return ref;
fs/btrfs/inode-item.c
655
struct btrfs_ref ref = {
fs/btrfs/inode-item.c
665
btrfs_init_data_ref(&ref, control->ino, extent_offset,
fs/btrfs/inode-item.c
667
ret = btrfs_free_extent(trans, &ref);
fs/btrfs/inode.c
4184
struct btrfs_inode_ref *ref;
fs/btrfs/inode.c
4186
ref = (struct btrfs_inode_ref *)ptr;
fs/btrfs/inode.c
4187
inode->dir_index = btrfs_inode_ref_index(leaf, ref);
fs/btrfs/inode.c
5819
struct btrfs_root_ref *ref;
fs/btrfs/inode.c
5849
ref = btrfs_item_ptr(leaf, path->slots[0], struct btrfs_root_ref);
fs/btrfs/inode.c
5850
if (btrfs_root_ref_dirid(leaf, ref) != btrfs_ino(dir) ||
fs/btrfs/inode.c
5851
btrfs_root_ref_name_len(leaf, ref) != fname.disk_name.len)
fs/btrfs/inode.c
5855
(unsigned long)(ref + 1), fname.disk_name.len);
fs/btrfs/inode.c
6621
struct btrfs_inode_ref *ref;
fs/btrfs/inode.c
6747
sizes[1] = 2 + sizeof(*ref);
fs/btrfs/inode.c
6750
sizes[1] = name->len + sizeof(*ref);
fs/btrfs/inode.c
6780
ref = btrfs_item_ptr(path->nodes[0], path->slots[0] + 1,
fs/btrfs/inode.c
6782
ptr = (unsigned long)(ref + 1);
fs/btrfs/inode.c
6784
btrfs_set_inode_ref_name_len(path->nodes[0], ref, 2);
fs/btrfs/inode.c
6785
btrfs_set_inode_ref_index(path->nodes[0], ref, 0);
fs/btrfs/inode.c
6788
btrfs_set_inode_ref_name_len(path->nodes[0], ref,
fs/btrfs/inode.c
6790
btrfs_set_inode_ref_index(path->nodes[0], ref,
fs/btrfs/print-tree.c
293
struct btrfs_inode_ref *ref = btrfs_item_ptr(eb, i, struct btrfs_inode_ref);
fs/btrfs/print-tree.c
297
const u64 index = btrfs_inode_ref_index(eb, ref);
fs/btrfs/print-tree.c
298
const u32 name_len = btrfs_inode_ref_name_len(eb, ref);
fs/btrfs/print-tree.c
299
const u32 len = sizeof(*ref) + name_len;
fs/btrfs/print-tree.c
302
ref = (struct btrfs_inode_ref *)((char *)ref + len);
fs/btrfs/print-tree.c
85
struct btrfs_extent_data_ref *ref)
fs/btrfs/print-tree.c
88
btrfs_extent_data_ref_root(eb, ref),
fs/btrfs/print-tree.c
89
btrfs_extent_data_ref_objectid(eb, ref),
fs/btrfs/print-tree.c
90
btrfs_extent_data_ref_offset(eb, ref),
fs/btrfs/print-tree.c
91
btrfs_extent_data_ref_count(eb, ref));
fs/btrfs/print-tree.c
95
const struct btrfs_extent_owner_ref *ref)
fs/btrfs/print-tree.c
98
pr_cont("extent data owner root %llu\n", btrfs_extent_owner_ref_root_id(eb, ref));
fs/btrfs/ref-verify.c
174
struct ref_entry *ref)
fs/btrfs/ref-verify.c
178
node = rb_find_add(&ref->node, root, ref_entry_cmp);
fs/btrfs/ref-verify.c
220
struct ref_entry *ref;
fs/btrfs/ref-verify.c
231
ref = rb_entry(n, struct ref_entry, node);
fs/btrfs/ref-verify.c
232
rb_erase(&ref->node, &be->refs);
fs/btrfs/ref-verify.c
233
kfree(ref);
fs/btrfs/ref-verify.c
299
struct ref_entry *ref = NULL, *exist;
fs/btrfs/ref-verify.c
301
ref = kmalloc_obj(struct ref_entry, GFP_NOFS);
fs/btrfs/ref-verify.c
302
if (!ref)
fs/btrfs/ref-verify.c
306
ref->root_objectid = 0;
fs/btrfs/ref-verify.c
308
ref->root_objectid = ref_root;
fs/btrfs/ref-verify.c
309
ref->parent = parent;
fs/btrfs/ref-verify.c
310
ref->owner = level;
fs/btrfs/ref-verify.c
311
ref->offset = 0;
fs/btrfs/ref-verify.c
312
ref->num_refs = 1;
fs/btrfs/ref-verify.c
316
kfree(ref);
fs/btrfs/ref-verify.c
329
exist = insert_ref_entry(&be->refs, ref);
fs/btrfs/ref-verify.c
332
kfree(ref);
fs/btrfs/ref-verify.c
344
struct ref_entry *ref;
fs/btrfs/ref-verify.c
346
ref = kzalloc_obj(struct ref_entry, GFP_NOFS);
fs/btrfs/ref-verify.c
347
if (!ref)
fs/btrfs/ref-verify.c
351
kfree(ref);
fs/btrfs/ref-verify.c
356
ref->parent = parent;
fs/btrfs/ref-verify.c
357
ref->num_refs = num_refs;
fs/btrfs/ref-verify.c
358
if (insert_ref_entry(&be->refs, ref)) {
fs/btrfs/ref-verify.c
361
kfree(ref);
fs/btrfs/ref-verify.c
374
struct ref_entry *ref;
fs/btrfs/ref-verify.c
381
ref = kzalloc_obj(struct ref_entry, GFP_NOFS);
fs/btrfs/ref-verify.c
382
if (!ref)
fs/btrfs/ref-verify.c
386
kfree(ref);
fs/btrfs/ref-verify.c
391
ref->parent = 0;
fs/btrfs/ref-verify.c
392
ref->owner = owner;
fs/btrfs/ref-verify.c
393
ref->root_objectid = ref_root;
fs/btrfs/ref-verify.c
394
ref->offset = offset;
fs/btrfs/ref-verify.c
395
ref->num_refs = num_refs;
fs/btrfs/ref-verify.c
396
if (insert_ref_entry(&be->refs, ref)) {
fs/btrfs/ref-verify.c
399
kfree(ref);
fs/btrfs/ref-verify.c
54
struct ref_entry ref;
fs/btrfs/ref-verify.c
604
ra->action, ra->root, ra->ref.root_objectid, ra->ref.parent,
fs/btrfs/ref-verify.c
605
ra->ref.owner, ra->ref.offset, ra->ref.num_refs);
fs/btrfs/ref-verify.c
616
struct ref_entry *ref;
fs/btrfs/ref-verify.c
627
ref = rb_entry(n, struct ref_entry, node);
fs/btrfs/ref-verify.c
630
ref->root_objectid, ref->parent, ref->owner,
fs/btrfs/ref-verify.c
631
ref->offset, ref->num_refs);
fs/btrfs/ref-verify.c
655
struct ref_entry *ref = NULL, *exist;
fs/btrfs/ref-verify.c
683
ref = kzalloc_obj(struct ref_entry, GFP_NOFS);
fs/btrfs/ref-verify.c
685
if (!ra || !ref) {
fs/btrfs/ref-verify.c
686
kfree(ref);
fs/btrfs/ref-verify.c
692
ref->parent = parent;
fs/btrfs/ref-verify.c
693
ref->owner = owner;
fs/btrfs/ref-verify.c
694
ref->root_objectid = ref_root;
fs/btrfs/ref-verify.c
695
ref->offset = offset;
fs/btrfs/ref-verify.c
696
ref->num_refs = (action == BTRFS_DROP_DELAYED_REF) ? -1 : 1;
fs/btrfs/ref-verify.c
698
memcpy(&ra->ref, ref, sizeof(struct ref_entry));
fs/btrfs/ref-verify.c
705
ra->ref.owner = owner;
fs/btrfs/ref-verify.c
706
ra->ref.offset = offset;
fs/btrfs/ref-verify.c
707
ra->ref.root_objectid = ref_root;
fs/btrfs/ref-verify.c
727
kfree(ref);
fs/btrfs/ref-verify.c
741
kfree(ref);
fs/btrfs/ref-verify.c
760
kfree(ref);
fs/btrfs/ref-verify.c
782
kfree(ref);
fs/btrfs/ref-verify.c
792
kfree(ref);
fs/btrfs/ref-verify.c
807
exist = insert_ref_entry(&be->refs, ref);
fs/btrfs/ref-verify.c
815
kfree(ref);
fs/btrfs/ref-verify.c
831
kfree(ref);
fs/btrfs/ref-verify.c
835
kfree(ref);
fs/btrfs/ref-verify.c
842
rb_erase(&ref->node, &be->refs);
fs/btrfs/ref-verify.c
843
kfree(ref);
fs/btrfs/relocation.c
1016
struct btrfs_ref ref = { 0 };
fs/btrfs/relocation.c
1180
ref.action = BTRFS_ADD_DELAYED_REF;
fs/btrfs/relocation.c
1181
ref.bytenr = old_bytenr;
fs/btrfs/relocation.c
1182
ref.num_bytes = blocksize;
fs/btrfs/relocation.c
1183
ref.parent = path->nodes[level]->start;
fs/btrfs/relocation.c
1184
ref.owning_root = btrfs_root_id(src);
fs/btrfs/relocation.c
1185
ref.ref_root = btrfs_root_id(src);
fs/btrfs/relocation.c
1186
btrfs_init_tree_ref(&ref, level - 1, 0, true);
fs/btrfs/relocation.c
1187
ret = btrfs_inc_extent_ref(trans, &ref);
fs/btrfs/relocation.c
1193
ref.action = BTRFS_ADD_DELAYED_REF;
fs/btrfs/relocation.c
1194
ref.bytenr = new_bytenr;
fs/btrfs/relocation.c
1195
ref.num_bytes = blocksize;
fs/btrfs/relocation.c
1196
ref.parent = 0;
fs/btrfs/relocation.c
1197
ref.owning_root = btrfs_root_id(dest);
fs/btrfs/relocation.c
1198
ref.ref_root = btrfs_root_id(dest);
fs/btrfs/relocation.c
1199
btrfs_init_tree_ref(&ref, level - 1, 0, true);
fs/btrfs/relocation.c
1200
ret = btrfs_inc_extent_ref(trans, &ref);
fs/btrfs/relocation.c
1207
ref.action = BTRFS_DROP_DELAYED_REF;
fs/btrfs/relocation.c
1208
ref.bytenr = new_bytenr;
fs/btrfs/relocation.c
1209
ref.num_bytes = blocksize;
fs/btrfs/relocation.c
1210
ref.parent = path->nodes[level]->start;
fs/btrfs/relocation.c
1211
ref.owning_root = 0;
fs/btrfs/relocation.c
1212
ref.ref_root = btrfs_root_id(src);
fs/btrfs/relocation.c
1213
btrfs_init_tree_ref(&ref, level - 1, 0, true);
fs/btrfs/relocation.c
1214
ret = btrfs_free_extent(trans, &ref);
fs/btrfs/relocation.c
1221
ref.action = BTRFS_DROP_DELAYED_REF;
fs/btrfs/relocation.c
1222
ref.bytenr = old_bytenr;
fs/btrfs/relocation.c
1223
ref.num_bytes = blocksize;
fs/btrfs/relocation.c
1224
ref.parent = 0;
fs/btrfs/relocation.c
1225
ref.owning_root = 0;
fs/btrfs/relocation.c
1226
ref.ref_root = btrfs_root_id(dest);
fs/btrfs/relocation.c
1227
btrfs_init_tree_ref(&ref, level - 1, 0, true);
fs/btrfs/relocation.c
1228
ret = btrfs_free_extent(trans, &ref);
fs/btrfs/relocation.c
2300
struct btrfs_ref ref = {
fs/btrfs/relocation.c
2315
btrfs_init_tree_ref(&ref, node->level,
fs/btrfs/relocation.c
2317
ret = btrfs_inc_extent_ref(trans, &ref);
fs/btrfs/relocation.c
885
struct btrfs_ref ref = { 0 };
fs/btrfs/relocation.c
954
ref.action = BTRFS_ADD_DELAYED_REF;
fs/btrfs/relocation.c
955
ref.bytenr = new_bytenr;
fs/btrfs/relocation.c
956
ref.num_bytes = num_bytes;
fs/btrfs/relocation.c
957
ref.parent = parent;
fs/btrfs/relocation.c
958
ref.owning_root = btrfs_root_id(root);
fs/btrfs/relocation.c
959
ref.ref_root = btrfs_header_owner(leaf);
fs/btrfs/relocation.c
960
btrfs_init_data_ref(&ref, key.objectid, key.offset,
fs/btrfs/relocation.c
962
ret = btrfs_inc_extent_ref(trans, &ref);
fs/btrfs/relocation.c
968
ref.action = BTRFS_DROP_DELAYED_REF;
fs/btrfs/relocation.c
969
ref.bytenr = bytenr;
fs/btrfs/relocation.c
970
ref.num_bytes = num_bytes;
fs/btrfs/relocation.c
971
ref.parent = parent;
fs/btrfs/relocation.c
972
ref.owning_root = btrfs_root_id(root);
fs/btrfs/relocation.c
973
ref.ref_root = btrfs_header_owner(leaf);
fs/btrfs/relocation.c
974
btrfs_init_data_ref(&ref, key.objectid, key.offset,
fs/btrfs/relocation.c
976
ret = btrfs_free_extent(trans, &ref);
fs/btrfs/root-tree.c
335
struct btrfs_root_ref *ref;
fs/btrfs/root-tree.c
354
ref = btrfs_item_ptr(leaf, path->slots[0],
fs/btrfs/root-tree.c
356
ptr = (unsigned long)(ref + 1);
fs/btrfs/root-tree.c
357
if ((btrfs_root_ref_dirid(leaf, ref) != dirid) ||
fs/btrfs/root-tree.c
358
(btrfs_root_ref_name_len(leaf, ref) != name->len) ||
fs/btrfs/root-tree.c
362
*sequence = btrfs_root_ref_sequence(leaf, ref);
fs/btrfs/root-tree.c
405
struct btrfs_root_ref *ref;
fs/btrfs/root-tree.c
418
sizeof(*ref) + name->len);
fs/btrfs/root-tree.c
425
ref = btrfs_item_ptr(leaf, path->slots[0], struct btrfs_root_ref);
fs/btrfs/root-tree.c
426
btrfs_set_root_ref_dirid(leaf, ref, dirid);
fs/btrfs/root-tree.c
427
btrfs_set_root_ref_sequence(leaf, ref, sequence);
fs/btrfs/root-tree.c
428
btrfs_set_root_ref_name_len(leaf, ref, name->len);
fs/btrfs/root-tree.c
429
ptr = (unsigned long)(ref + 1);
fs/btrfs/send.c
2460
struct btrfs_root_ref *ref;
fs/btrfs/send.c
2490
ref = btrfs_item_ptr(leaf, path->slots[0], struct btrfs_root_ref);
fs/btrfs/send.c
2491
namelen = btrfs_root_ref_name_len(leaf, ref);
fs/btrfs/send.c
2492
read_extent_buffer(leaf, name, (unsigned long)(ref + 1), namelen);
fs/btrfs/send.c
2975
struct recorded_ref *ref;
fs/btrfs/send.c
2977
ref = kzalloc_obj(*ref);
fs/btrfs/send.c
2978
if (!ref)
fs/btrfs/send.c
2980
RB_CLEAR_NODE(&ref->node);
fs/btrfs/send.c
2981
INIT_LIST_HEAD(&ref->list);
fs/btrfs/send.c
2982
return ref;
fs/btrfs/send.c
2985
static void recorded_ref_free(struct recorded_ref *ref)
fs/btrfs/send.c
2987
if (!ref)
fs/btrfs/send.c
2989
if (!RB_EMPTY_NODE(&ref->node))
fs/btrfs/send.c
2990
rb_erase(&ref->node, ref->root);
fs/btrfs/send.c
2991
list_del(&ref->list);
fs/btrfs/send.c
2992
fs_path_free(ref->full_path);
fs/btrfs/send.c
2993
kfree(ref);
fs/btrfs/send.c
2996
static void set_ref_path(struct recorded_ref *ref, struct fs_path *path)
fs/btrfs/send.c
2998
ref->full_path = path;
fs/btrfs/send.c
2999
ref->name = (char *)kbasename(ref->full_path->start);
fs/btrfs/send.c
3000
ref->name_len = ref->full_path->end - ref->name;
fs/btrfs/send.c
3003
static int dup_ref(struct recorded_ref *ref, struct list_head *list)
fs/btrfs/send.c
3011
new->dir = ref->dir;
fs/btrfs/send.c
3012
new->dir_gen = ref->dir_gen;
fs/btrfs/send.c
4003
static int update_ref_path(struct send_ctx *sctx, struct recorded_ref *ref)
fs/btrfs/send.c
4016
ret = get_cur_path(sctx, ref->dir, ref->dir_gen, new_path);
fs/btrfs/send.c
4021
ret = fs_path_add(new_path, ref->name, ref->name_len);
fs/btrfs/send.c
4027
fs_path_free(ref->full_path);
fs/btrfs/send.c
4028
set_ref_path(ref, new_path);
fs/btrfs/send.c
4074
static int refresh_ref_path(struct send_ctx *sctx, struct recorded_ref *ref)
fs/btrfs/send.c
4079
name = kmemdup(ref->name, ref->name_len, GFP_KERNEL);
fs/btrfs/send.c
4083
fs_path_reset(ref->full_path);
fs/btrfs/send.c
4084
ret = get_cur_path(sctx, ref->dir, ref->dir_gen, ref->full_path);
fs/btrfs/send.c
4088
ret = fs_path_add(ref->full_path, name, ref->name_len);
fs/btrfs/send.c
4093
set_ref_path(ref, ref->full_path);
fs/btrfs/send.c
4101
const struct recorded_ref *ref = rb_entry(node, struct recorded_ref, node);
fs/btrfs/send.c
4103
if (data->dir > ref->dir)
fs/btrfs/send.c
4105
if (data->dir < ref->dir)
fs/btrfs/send.c
4107
if (data->dir_gen > ref->dir_gen)
fs/btrfs/send.c
4109
if (data->dir_gen < ref->dir_gen)
fs/btrfs/send.c
4122
struct recorded_ref *ref, struct list_head *list)
fs/btrfs/send.c
4127
if (rb_find(ref, root, rbtree_check_dir_ref_comp))
fs/btrfs/send.c
4130
ret = dup_ref(ref, list);
fs/btrfs/send.c
4618
const struct recorded_ref *ref = rb_entry(node, struct recorded_ref, node);
fs/btrfs/send.c
4620
if (data->dir > ref->dir)
fs/btrfs/send.c
4622
if (data->dir < ref->dir)
fs/btrfs/send.c
4624
if (data->dir_gen > ref->dir_gen)
fs/btrfs/send.c
4626
if (data->dir_gen < ref->dir_gen)
fs/btrfs/send.c
4628
if (data->name_len > ref->name_len)
fs/btrfs/send.c
4630
if (data->name_len < ref->name_len)
fs/btrfs/send.c
4632
return strcmp(data->name, ref->name);
fs/btrfs/send.c
4648
struct recorded_ref *ref = NULL;
fs/btrfs/send.c
4656
ref = recorded_ref_alloc();
fs/btrfs/send.c
4657
if (!ref) {
fs/btrfs/send.c
4669
ref->dir = dir;
fs/btrfs/send.c
4670
ref->dir_gen = dir_gen;
fs/btrfs/send.c
4671
set_ref_path(ref, path);
fs/btrfs/send.c
4672
list_add_tail(&ref->list, refs);
fs/btrfs/send.c
4673
rb_add(&ref->node, root, rbtree_ref_less);
fs/btrfs/send.c
4674
ref->root = root;
fs/btrfs/send.c
4677
if (path && (!ref || !ref->full_path))
fs/btrfs/send.c
4679
recorded_ref_free(ref);
fs/btrfs/send.c
4690
struct recorded_ref *ref;
fs/btrfs/send.c
4702
ref = rb_entry(node, struct recorded_ref, node);
fs/btrfs/send.c
4703
recorded_ref_free(ref);
fs/btrfs/send.c
4719
struct recorded_ref *ref;
fs/btrfs/send.c
4731
ref = rb_entry(node, struct recorded_ref, node);
fs/btrfs/send.c
4732
recorded_ref_free(ref);
fs/btrfs/tests/delayed-refs-tests.c
175
struct btrfs_ref ref = {
fs/btrfs/tests/delayed-refs-tests.c
185
if (ref.type == BTRFS_REF_METADATA)
fs/btrfs/tests/delayed-refs-tests.c
186
btrfs_init_tree_ref(&ref, node_check->owner, node_check->root,
fs/btrfs/tests/delayed-refs-tests.c
189
btrfs_init_data_ref(&ref, node_check->owner, node_check->offset,
fs/btrfs/tests/delayed-refs-tests.c
192
if (ref.type == BTRFS_REF_METADATA)
fs/btrfs/tests/delayed-refs-tests.c
193
ret = btrfs_add_delayed_tree_ref(trans, &ref, NULL);
fs/btrfs/tests/delayed-refs-tests.c
195
ret = btrfs_add_delayed_data_ref(trans, &ref, 0);
fs/btrfs/tests/delayed-refs-tests.c
343
struct btrfs_ref ref = {
fs/btrfs/tests/delayed-refs-tests.c
374
btrfs_init_tree_ref(&ref, FAKE_LEVEL, FAKE_ROOT_OBJECTID, false);
fs/btrfs/tests/delayed-refs-tests.c
379
btrfs_init_data_ref(&ref, FAKE_INO, FAKE_FILE_OFFSET,
fs/btrfs/tests/delayed-refs-tests.c
384
ret = btrfs_add_delayed_tree_ref(trans, &ref, NULL);
fs/btrfs/tests/delayed-refs-tests.c
386
ret = btrfs_add_delayed_data_ref(trans, &ref, 0);
fs/btrfs/tests/delayed-refs-tests.c
392
ref.action = BTRFS_DROP_DELAYED_REF;
fs/btrfs/tests/delayed-refs-tests.c
394
ret = btrfs_add_delayed_tree_ref(trans, &ref, NULL);
fs/btrfs/tests/delayed-refs-tests.c
396
ret = btrfs_add_delayed_data_ref(trans, &ref, 0);
fs/btrfs/tests/delayed-refs-tests.c
433
ref.action = BTRFS_ADD_DELAYED_REF;
fs/btrfs/tests/delayed-refs-tests.c
435
ret = btrfs_add_delayed_tree_ref(trans, &ref, NULL);
fs/btrfs/tests/delayed-refs-tests.c
437
ret = btrfs_add_delayed_data_ref(trans, &ref, 0);
fs/btrfs/tests/delayed-refs-tests.c
444
ret = btrfs_add_delayed_tree_ref(trans, &ref, NULL);
fs/btrfs/tests/delayed-refs-tests.c
446
ret = btrfs_add_delayed_data_ref(trans, &ref, 0);
fs/btrfs/tests/delayed-refs-tests.c
496
ref.action = BTRFS_DROP_DELAYED_REF;
fs/btrfs/tests/delayed-refs-tests.c
498
ret = btrfs_add_delayed_tree_ref(trans, &ref, NULL);
fs/btrfs/tests/delayed-refs-tests.c
500
ret = btrfs_add_delayed_data_ref(trans, &ref, 0);
fs/btrfs/tests/delayed-refs-tests.c
507
ret = btrfs_add_delayed_tree_ref(trans, &ref, NULL);
fs/btrfs/tests/delayed-refs-tests.c
509
ret = btrfs_add_delayed_data_ref(trans, &ref, 0);
fs/btrfs/tests/delayed-refs-tests.c
560
ref.action = BTRFS_ADD_DELAYED_REF;
fs/btrfs/tests/delayed-refs-tests.c
563
ret = btrfs_add_delayed_tree_ref(trans, &ref, NULL);
fs/btrfs/tests/delayed-refs-tests.c
565
ret = btrfs_add_delayed_data_ref(trans, &ref, 0);
fs/btrfs/tests/delayed-refs-tests.c
572
ref.action = BTRFS_DROP_DELAYED_REF;
fs/btrfs/tests/delayed-refs-tests.c
575
ret = btrfs_add_delayed_tree_ref(trans, &ref, NULL);
fs/btrfs/tests/delayed-refs-tests.c
577
ret = btrfs_add_delayed_data_ref(trans, &ref, 0);
fs/btrfs/tests/delayed-refs-tests.c
630
ref.action = BTRFS_DROP_DELAYED_REF;
fs/btrfs/tests/delayed-refs-tests.c
633
ret = btrfs_add_delayed_tree_ref(trans, &ref, NULL);
fs/btrfs/tests/delayed-refs-tests.c
635
ret = btrfs_add_delayed_data_ref(trans, &ref, 0);
fs/btrfs/tests/delayed-refs-tests.c
642
ref.action = BTRFS_ADD_DELAYED_REF;
fs/btrfs/tests/delayed-refs-tests.c
645
ret = btrfs_add_delayed_tree_ref(trans, &ref, NULL);
fs/btrfs/tests/delayed-refs-tests.c
647
ret = btrfs_add_delayed_data_ref(trans, &ref, 0);
fs/btrfs/tests/delayed-refs-tests.c
703
ref.action = BTRFS_ADD_DELAYED_REF;
fs/btrfs/tests/delayed-refs-tests.c
706
ref.parent = 0;
fs/btrfs/tests/delayed-refs-tests.c
707
ref.ref_root = FAKE_ROOT_OBJECTID + i;
fs/btrfs/tests/delayed-refs-tests.c
709
ref.parent = FAKE_PARENT + (i * fs_info->nodesize);
fs/btrfs/tests/delayed-refs-tests.c
712
ret = btrfs_add_delayed_tree_ref(trans, &ref, NULL);
fs/btrfs/tests/delayed-refs-tests.c
714
ret = btrfs_add_delayed_data_ref(trans, &ref, 0);
fs/btrfs/tests/delayed-refs-tests.c
721
ref.action = BTRFS_DROP_DELAYED_REF;
fs/btrfs/tests/delayed-refs-tests.c
724
ref.parent = 0;
fs/btrfs/tests/delayed-refs-tests.c
725
ref.ref_root = FAKE_ROOT_OBJECTID + i;
fs/btrfs/tests/delayed-refs-tests.c
727
ref.parent = FAKE_PARENT + (i * fs_info->nodesize);
fs/btrfs/tests/delayed-refs-tests.c
730
ret = btrfs_add_delayed_tree_ref(trans, &ref, NULL);
fs/btrfs/tests/delayed-refs-tests.c
732
ret = btrfs_add_delayed_data_ref(trans, &ref, 0);
fs/btrfs/tests/delayed-refs-tests.c
784
struct btrfs_ref ref = {
fs/btrfs/tests/delayed-refs-tests.c
811
btrfs_init_tree_ref(&ref, FAKE_LEVEL, FAKE_ROOT_OBJECTID, false);
fs/btrfs/tests/delayed-refs-tests.c
812
ret = btrfs_add_delayed_tree_ref(trans, &ref, NULL);
fs/btrfs/tests/delayed-refs-tests.c
822
ref.action = BTRFS_ADD_DELAYED_REF;
fs/btrfs/tests/delayed-refs-tests.c
823
ref.ref_root = FAKE_ROOT_OBJECTID + 1;
fs/btrfs/tests/delayed-refs-tests.c
824
ret = btrfs_add_delayed_tree_ref(trans, &ref, NULL);
fs/btrfs/tests/delayed-refs-tests.c
886
ref.action = BTRFS_DROP_DELAYED_REF;
fs/btrfs/tests/delayed-refs-tests.c
887
ref.ref_root = FAKE_ROOT_OBJECTID;
fs/btrfs/tests/delayed-refs-tests.c
888
ret = btrfs_add_delayed_tree_ref(trans, &ref, NULL);
fs/btrfs/tests/delayed-refs-tests.c
894
ref.action = BTRFS_ADD_DELAYED_REF;
fs/btrfs/tests/delayed-refs-tests.c
895
ref.ref_root = FAKE_ROOT_OBJECTID + 1;
fs/btrfs/tests/delayed-refs-tests.c
896
ret = btrfs_add_delayed_tree_ref(trans, &ref, NULL);
fs/btrfs/tests/delayed-refs-tests.c
902
ref.action = BTRFS_DROP_DELAYED_REF;
fs/btrfs/tests/delayed-refs-tests.c
903
ret = btrfs_add_delayed_tree_ref(trans, &ref, NULL);
fs/btrfs/tests/delayed-refs-tests.c
909
ref.action = BTRFS_ADD_DELAYED_REF;
fs/btrfs/tests/delayed-refs-tests.c
910
ref.ref_root = FAKE_ROOT_OBJECTID + 2;
fs/btrfs/tests/delayed-refs-tests.c
911
ret = btrfs_add_delayed_tree_ref(trans, &ref, NULL);
fs/btrfs/tree-log.c
1453
struct btrfs_inode_ref *ref;
fs/btrfs/tree-log.c
1456
ref = (struct btrfs_inode_ref *)ref_ptr;
fs/btrfs/tree-log.c
1458
ret = read_alloc_one_name(eb, ref + 1, btrfs_inode_ref_name_len(eb, ref),
fs/btrfs/tree-log.c
1464
*index = btrfs_inode_ref_index(eb, ref);
fs/btrfs/tree-log.c
1846
struct btrfs_inode_ref *ref;
fs/btrfs/tree-log.c
1848
ref = (struct btrfs_inode_ref *)ptr;
fs/btrfs/tree-log.c
1850
ref);
fs/btrfs/tree-log.c
1851
ptr = (unsigned long)(ref + 1) + name_len;
fs/btrfs/tree-log.c
886
struct btrfs_ref ref = {
fs/btrfs/tree-log.c
894
btrfs_init_data_ref(&ref, wc->log_key.objectid, offset, 0, false);
fs/btrfs/tree-log.c
895
ret = btrfs_inc_extent_ref(trans, &ref);
fs/cachefiles/interface.c
40
refcount_set(&object->ref, 1);
fs/cachefiles/interface.c
61
refcount_read(&object->ref), why);
fs/cachefiles/interface.c
72
__refcount_inc(&object->ref, &r);
fs/cachefiles/interface.c
89
done = __refcount_dec_and_test(&object->ref, &r);
fs/cachefiles/internal.h
144
refcount_t ref;
fs/cachefiles/internal.h
73
refcount_t ref;
fs/cachefiles/ondemand.c
13
if (refcount_dec_and_test(&req->ref))
fs/cachefiles/ondemand.c
449
refcount_inc(&req->ref);
fs/cachefiles/ondemand.c
509
refcount_set(&req->ref, 1);
fs/cachefiles/volume.c
86
refcount_read(&vcookie->ref),
fs/ceph/mds_client.h
403
atomic_t ref;
fs/ceph/snap.c
1206
if (atomic_inc_return(&exist->ref) == 1)
fs/ceph/snap.c
1230
atomic_set(&sm->ref, 1);
fs/ceph/snap.c
1249
if (atomic_inc_return(&exist->ref) == 1)
fs/ceph/snap.c
1273
if (atomic_dec_and_lock(&sm->ref, &mdsc->snapid_map_lock)) {
fs/ceph/snap.c
1337
if (WARN_ON_ONCE(atomic_read(&sm->ref))) {
fs/crypto/keyring.c
694
key_ref_t ref;
fs/crypto/keyring.c
699
ref = lookup_user_key(key_id, 0, KEY_NEED_SEARCH);
fs/crypto/keyring.c
700
if (IS_ERR(ref))
fs/crypto/keyring.c
701
return PTR_ERR(ref);
fs/crypto/keyring.c
702
key = key_ref_to_ptr(ref);
fs/crypto/keyring.c
724
key_ref_put(ref);
fs/dax.c
383
unsigned long ref;
fs/dax.c
387
ref = 0;
fs/dax.c
389
ref = --folio->share;
fs/dax.c
391
if (ref)
fs/dax.c
392
return ref;
fs/dax.c
418
return ref;
fs/dlm/lowcomms.c
1185
kref_init(&entry->ref);
fs/dlm/lowcomms.c
1199
kref_get(&e->ref);
fs/dlm/lowcomms.c
1215
kref_get(&e->ref);
fs/dlm/lowcomms.c
1239
kref_init(&msg->ref);
fs/dlm/lowcomms.c
1289
kref_get(&msg->ref);
fs/dlm/lowcomms.c
1303
kref_get(&msg->ref);
fs/dlm/lowcomms.c
1328
kref_put(&msg->ref, dlm_msg_release);
fs/dlm/lowcomms.c
1334
kref_put(&msg->ref, dlm_msg_release);
fs/dlm/lowcomms.c
1352
kref_get(&msg->ref);
fs/dlm/lowcomms.c
136
struct kref ref;
fs/dlm/lowcomms.c
148
struct kref ref;
fs/dlm/lowcomms.c
685
ref);
fs/dlm/lowcomms.c
693
struct dlm_msg *msg = container_of(kref, struct dlm_msg, ref);
fs/dlm/lowcomms.c
695
kref_put(&msg->entry->ref, dlm_page_release);
fs/dlm/lowcomms.c
706
kref_put(&msg->orig_msg->ref, dlm_msg_release);
fs/dlm/lowcomms.c
710
kref_put(&msg->ref, dlm_msg_release);
fs/dlm/lowcomms.c
714
kref_put(&e->ref, dlm_page_release);
fs/erofs/fileio.c
13
refcount_t ref;
fs/erofs/fileio.c
40
if (refcount_dec_and_test(&rq->ref))
fs/erofs/fileio.c
63
if (refcount_dec_and_test(&rq->ref))
fs/erofs/fileio.c
74
refcount_set(&rq->ref, 2);
fs/erofs/fscache.c
110
refcount_inc(&req->ref);
fs/erofs/fscache.c
111
refcount_set(&io->ref, 1);
fs/erofs/fscache.c
145
refcount_inc(&io->ref);
fs/erofs/fscache.c
188
refcount_set(&io->io.ref, 1);
fs/erofs/fscache.c
20
refcount_t ref;
fs/erofs/fscache.c
29
refcount_t ref;
fs/erofs/fscache.c
34
if (!refcount_dec_and_test(&io->ref))
fs/erofs/fscache.c
374
if (refcount_dec_and_test(&domain->ref)) {
fs/erofs/fscache.c
444
refcount_set(&domain->ref, 1);
fs/erofs/fscache.c
465
refcount_inc(&domain->ref);
fs/erofs/fscache.c
489
refcount_set(&ctx->ref, 1);
fs/erofs/fscache.c
555
refcount_inc(&domain->ref);
fs/erofs/fscache.c
573
refcount_inc(&ctx->ref);
fs/erofs/fscache.c
606
if (refcount_dec_and_test(&ctx->ref)) {
fs/erofs/fscache.c
64
if (!refcount_dec_and_test(&req->ref))
fs/erofs/fscache.c
80
refcount_set(&req->ref, 1);
fs/erofs/internal.h
82
refcount_t ref;
fs/erofs/internal.h
95
refcount_t ref;
fs/ext4/xattr.c
1276
u32 hash, ref;
fs/ext4/xattr.c
1288
ref = le32_to_cpu(BHDR(bh)->h_refcount);
fs/ext4/xattr.c
1289
if (ref == 1) {
fs/ext4/xattr.c
1321
ref--;
fs/ext4/xattr.c
1322
BHDR(bh)->h_refcount = cpu_to_le32(ref);
fs/ext4/xattr.c
1323
if (ref == EXT4_XATTR_REFCOUNT_MAX - 1) {
fs/ext4/xattr.c
2051
u32 ref;
fs/ext4/xattr.c
2076
ref = le32_to_cpu(BHDR(new_bh)->h_refcount) + 1;
fs/ext4/xattr.c
2077
if (ref > EXT4_XATTR_REFCOUNT_MAX) {
fs/ext4/xattr.c
2092
BHDR(new_bh)->h_refcount = cpu_to_le32(ref);
fs/ext4/xattr.c
2093
if (ref == EXT4_XATTR_REFCOUNT_MAX)
fs/ext4/xattr.c
2096
ref);
fs/f2fs/f2fs.h
475
unsigned short ref; /* reference count */
fs/f2fs/segment.c
1133
f2fs_bug_on(sbi, dc->ref);
fs/f2fs/segment.c
1748
dc->ref--;
fs/f2fs/segment.c
1749
if (!dc->ref) {
fs/f2fs/segment.c
1779
if (iter->state == D_DONE && !iter->ref) {
fs/f2fs/segment.c
1785
iter->ref++;
fs/f2fs/segment.c
1841
dc->ref++;
fs/f2fs/segment.c
1853
dc->ref++;
fs/f2fs/segment.c
984
dc->ref = 0;
fs/file.c
29
static noinline bool __file_ref_put_badval(file_ref_t *ref, unsigned long cnt)
fs/file.c
37
atomic_long_set(&ref->refcnt, FILE_REF_DEAD);
fs/file.c
47
atomic_long_set(&ref->refcnt, FILE_REF_SATURATED);
fs/file.c
68
bool __file_ref_put(file_ref_t *ref, unsigned long cnt)
fs/file.c
81
if (!atomic_long_try_cmpxchg_release(&ref->refcnt, &cnt, FILE_REF_DEAD))
fs/file.c
92
return __file_ref_put_badval(ref, cnt);
fs/iomap/direct-io.c
256
if (atomic_dec_and_test(&dio->ref)) {
fs/iomap/direct-io.c
33
atomic_t ref;
fs/iomap/direct-io.c
67
atomic_inc(&dio->ref);
fs/iomap/direct-io.c
698
atomic_set(&dio->ref, 1);
fs/iomap/direct-io.c
857
if (!atomic_dec_and_test(&dio->ref)) {
fs/jffs2/debug.c
404
struct jffs2_raw_node_ref *ref;
fs/jffs2/debug.c
414
for (ref = jeb->first_node; ; ref = ref_next(ref)) {
fs/jffs2/debug.c
415
printk("%#08x", ref_offset(ref));
fs/jffs2/debug.c
417
printk("(%x)", ref->__totlen);
fs/jffs2/debug.c
419
if (ref_next(ref))
fs/jffs2/erase.c
207
struct jffs2_raw_node_ref *ref, struct jffs2_eraseblock *jeb)
fs/jffs2/erase.c
212
prev = &ref->next_in_ino;
fs/jffs2/erase.c
233
if (this == ref)
fs/jffs2/erase.c
289
struct jffs2_raw_node_ref *block, *ref;
fs/jffs2/erase.c
293
block = ref = jeb->first_node;
fs/jffs2/erase.c
295
while (ref) {
fs/jffs2/erase.c
296
if (ref->flash_offset == REF_LINK_NODE) {
fs/jffs2/erase.c
297
ref = ref->next_in_ino;
fs/jffs2/erase.c
299
block = ref;
fs/jffs2/erase.c
302
if (ref->flash_offset != REF_EMPTY_NODE && ref->next_in_ino)
fs/jffs2/erase.c
303
jffs2_remove_node_refs_from_ino_list(c, ref, jeb);
fs/jffs2/erase.c
306
ref++;
fs/jffs2/malloc.c
188
struct jffs2_raw_node_ref **p, *ref;
fs/jffs2/malloc.c
194
ref = *p;
fs/jffs2/malloc.c
199
if (ref && ref->flash_offset != REF_EMPTY_NODE)
fs/jffs2/malloc.c
200
ref++;
fs/jffs2/malloc.c
203
if (!ref) {
fs/jffs2/malloc.c
205
ref = *p = jffs2_alloc_refblock();
fs/jffs2/malloc.c
206
if (!ref)
fs/jffs2/malloc.c
209
if (ref->flash_offset == REF_LINK_NODE) {
fs/jffs2/malloc.c
210
p = &ref->next_in_ino;
fs/jffs2/malloc.c
211
ref = *p;
fs/jffs2/malloc.c
215
ref++;
fs/jffs2/malloc.c
283
struct jffs2_xattr_ref *ref;
fs/jffs2/malloc.c
284
ref = kmem_cache_zalloc(xattr_ref_cache, GFP_KERNEL);
fs/jffs2/malloc.c
285
dbg_memalloc("%p\n", ref);
fs/jffs2/malloc.c
286
if (!ref)
fs/jffs2/malloc.c
289
ref->class = RAWNODE_CLASS_XATTR_REF;
fs/jffs2/malloc.c
290
ref->node = (void *)ref;
fs/jffs2/malloc.c
291
return ref;
fs/jffs2/malloc.c
294
void jffs2_free_xattr_ref(struct jffs2_xattr_ref *ref)
fs/jffs2/malloc.c
296
dbg_memalloc("%p\n", ref);
fs/jffs2/malloc.c
297
kmem_cache_free(xattr_ref_cache, ref);
fs/jffs2/nodelist.c
590
struct jffs2_raw_node_ref *ref;
fs/jffs2/nodelist.c
595
ref = jeb->last_node;
fs/jffs2/nodelist.c
597
dbg_noderef("Last node at %p is (%08x,%p)\n", ref, ref->flash_offset,
fs/jffs2/nodelist.c
598
ref->next_in_ino);
fs/jffs2/nodelist.c
600
while (ref->flash_offset != REF_EMPTY_NODE) {
fs/jffs2/nodelist.c
601
if (ref->flash_offset == REF_LINK_NODE)
fs/jffs2/nodelist.c
602
ref = ref->next_in_ino;
fs/jffs2/nodelist.c
604
ref++;
fs/jffs2/nodelist.c
607
dbg_noderef("New ref is %p (%08x becomes %08x,%p) len 0x%x\n", ref,
fs/jffs2/nodelist.c
608
ref->flash_offset, ofs, ref->next_in_ino, len);
fs/jffs2/nodelist.c
610
ref->flash_offset = ofs;
fs/jffs2/nodelist.c
613
jeb->first_node = ref;
fs/jffs2/nodelist.c
614
BUG_ON(ref_offset(ref) != jeb->offset);
fs/jffs2/nodelist.c
615
} else if (unlikely(ref_offset(ref) != jeb->offset + c->sector_size - jeb->free_size)) {
fs/jffs2/nodelist.c
619
ref, ref_offset(ref), ref_offset(ref)+len,
fs/jffs2/nodelist.c
624
jeb->last_node = ref;
fs/jffs2/nodelist.c
627
ref->next_in_ino = ic->nodes;
fs/jffs2/nodelist.c
628
ic->nodes = ref;
fs/jffs2/nodelist.c
630
ref->next_in_ino = NULL;
fs/jffs2/nodelist.c
633
switch(ref_flags(ref)) {
fs/jffs2/nodelist.c
655
ref->__totlen = len;
fs/jffs2/nodelist.c
656
ref_totlen(c, jeb, ref);
fs/jffs2/nodelist.c
658
return ref;
fs/jffs2/nodelist.c
694
struct jffs2_raw_node_ref *ref)
fs/jffs2/nodelist.c
697
struct jffs2_raw_node_ref *next_ref = ref_next(ref);
fs/jffs2/nodelist.c
703
jeb = &c->blocks[ref->flash_offset / c->sector_size];
fs/jffs2/nodelist.c
706
if (unlikely(ref != jeb->last_node)) {
fs/jffs2/nodelist.c
708
ref, ref_offset(ref), jeb->last_node,
fs/jffs2/nodelist.c
715
return ref_end - ref_offset(ref);
fs/jffs2/nodelist.c
719
struct jffs2_raw_node_ref *ref)
fs/jffs2/nodelist.c
723
ret = __ref_totlen(c, jeb, ref);
fs/jffs2/nodelist.c
726
if (unlikely(ret != ref->__totlen)) {
fs/jffs2/nodelist.c
728
jeb = &c->blocks[ref->flash_offset / c->sector_size];
fs/jffs2/nodelist.c
731
ref, ref_offset(ref), ref_offset(ref) + ref->__totlen,
fs/jffs2/nodelist.c
732
ret, ref->__totlen);
fs/jffs2/nodelist.c
733
if (ref_next(ref)) {
fs/jffs2/nodelist.c
735
ref_next(ref), ref_offset(ref_next(ref)),
fs/jffs2/nodelist.c
736
ref_offset(ref_next(ref)) + ref->__totlen);
fs/jffs2/nodelist.c
751
ret = ref->__totlen;
fs/jffs2/nodelist.h
101
static inline struct jffs2_raw_node_ref *ref_next(struct jffs2_raw_node_ref *ref)
fs/jffs2/nodelist.h
103
ref++;
fs/jffs2/nodelist.h
106
if (ref->flash_offset == REF_LINK_NODE) {
fs/jffs2/nodelist.h
107
ref = ref->next_in_ino;
fs/jffs2/nodelist.h
108
if (!ref)
fs/jffs2/nodelist.h
109
return ref;
fs/jffs2/nodelist.h
113
if (ref->flash_offset == REF_EMPTY_NODE)
fs/jffs2/nodelist.h
116
return ref;
fs/jffs2/nodelist.h
136
#define ref_flags(ref) ((ref)->flash_offset & 3)
fs/jffs2/nodelist.h
137
#define ref_offset(ref) ((ref)->flash_offset & ~3)
fs/jffs2/nodelist.h
138
#define ref_obsolete(ref) (((ref)->flash_offset & 3) == REF_OBSOLETE)
fs/jffs2/nodelist.h
139
#define mark_ref_normal(ref) do { (ref)->flash_offset = ref_offset(ref) | REF_NORMAL; } while(0)
fs/jffs2/nodelist.h
381
struct jffs2_raw_node_ref *ref);
fs/jffs2/nodemgmt.c
586
void jffs2_mark_node_obsolete(struct jffs2_sb_info *c, struct jffs2_raw_node_ref *ref)
fs/jffs2/nodemgmt.c
595
if(unlikely(!ref)) {
fs/jffs2/nodemgmt.c
599
if (ref_obsolete(ref)) {
fs/jffs2/nodemgmt.c
601
__func__, ref_offset(ref));
fs/jffs2/nodemgmt.c
604
blocknr = ref->flash_offset / c->sector_size;
fs/jffs2/nodemgmt.c
607
ref->flash_offset);
fs/jffs2/nodemgmt.c
625
freed_len = ref_totlen(c, jeb, ref);
fs/jffs2/nodemgmt.c
627
if (ref_flags(ref) == REF_UNCHECKED) {
fs/jffs2/nodemgmt.c
631
ref->flash_offset, jeb->used_size);
fs/jffs2/nodemgmt.c
635
ref_offset(ref), freed_len);
fs/jffs2/nodemgmt.c
642
ref->flash_offset, jeb->used_size);
fs/jffs2/nodemgmt.c
646
ref_offset(ref), freed_len);
fs/jffs2/nodemgmt.c
680
ref->flash_offset = ref_offset(ref) | REF_OBSOLETE;
fs/jffs2/nodemgmt.c
764
ref_offset(ref));
fs/jffs2/nodemgmt.c
765
ret = jffs2_flash_read(c, ref_offset(ref), sizeof(n), &retlen, (char *)&n);
fs/jffs2/nodemgmt.c
768
ref_offset(ref), ret);
fs/jffs2/nodemgmt.c
773
ref_offset(ref), retlen);
fs/jffs2/nodemgmt.c
783
ref_offset(ref), je16_to_cpu(n.nodetype));
fs/jffs2/nodemgmt.c
788
ret = jffs2_flash_write(c, ref_offset(ref), sizeof(n), &retlen, (char *)&n);
fs/jffs2/nodemgmt.c
791
ref_offset(ref), ret);
fs/jffs2/nodemgmt.c
796
ref_offset(ref), retlen);
fs/jffs2/nodemgmt.c
810
if (ref->next_in_ino) {
fs/jffs2/nodemgmt.c
816
ic = jffs2_raw_ref_to_ic(ref);
fs/jffs2/nodemgmt.c
817
for (p = &ic->nodes; (*p) != ref; p = &((*p)->next_in_ino))
fs/jffs2/nodemgmt.c
820
*p = ref->next_in_ino;
fs/jffs2/nodemgmt.c
821
ref->next_in_ino = NULL;
fs/jffs2/readinode.c
1012
ref = valid_ref;
fs/jffs2/readinode.c
1013
valid_ref = jffs2_first_valid_node(ref->next_in_ino);
fs/jffs2/readinode.c
1035
end = ref_offset(ref) + len;
fs/jffs2/readinode.c
1039
len = end - ref_offset(ref);
fs/jffs2/readinode.c
1042
dbg_readinode("read %d bytes at %#08x(%d).\n", len, ref_offset(ref), ref_flags(ref));
fs/jffs2/readinode.c
1045
err = jffs2_flash_read(c, ref_offset(ref), len, &retlen, buf);
fs/jffs2/readinode.c
1047
JFFS2_ERROR("can not read %d bytes from 0x%08x, error code: %d.\n", len, ref_offset(ref), err);
fs/jffs2/readinode.c
1052
JFFS2_ERROR("short read at %#08x: %zu instead of %d.\n", ref_offset(ref), retlen, len);
fs/jffs2/readinode.c
1062
ref_offset(ref), je16_to_cpu(node->u.magic),
fs/jffs2/readinode.c
1066
jffs2_dbg_dump_node(c, ref_offset(ref));
fs/jffs2/readinode.c
1067
jffs2_mark_node_obsolete(c, ref);
fs/jffs2/readinode.c
1073
je16_to_cpu(node->u.magic), ref_offset(ref));
fs/jffs2/readinode.c
1074
jffs2_mark_node_obsolete(c, ref);
fs/jffs2/readinode.c
1084
err = read_more(c, ref, sizeof(struct jffs2_raw_dirent), &len, buf);
fs/jffs2/readinode.c
1089
err = read_direntry(c, ref, &node->d, retlen, rii);
fs/jffs2/readinode.c
109
ref_offset(ref), tn->data_crc, crc);
fs/jffs2/readinode.c
1099
err = read_more(c, ref, sizeof(struct jffs2_raw_inode), &len, buf);
fs/jffs2/readinode.c
1104
err = read_dnode(c, ref, &node->i, len, rii);
fs/jffs2/readinode.c
1113
err = read_more(c, ref, sizeof(struct jffs2_unknown_node), &len, buf);
fs/jffs2/readinode.c
1118
err = read_unknown(c, ref, &node->u);
fs/jffs2/readinode.c
114
jeb = &c->blocks[ref->flash_offset / c->sector_size];
fs/jffs2/readinode.c
115
len = ref_totlen(c, jeb, ref);
fs/jffs2/readinode.c
120
ref->flash_offset |= REF_PRISTINE;
fs/jffs2/readinode.c
33
struct jffs2_raw_node_ref *ref = tn->fn->raw;
fs/jffs2/readinode.c
43
ofs = ref_offset(ref) + sizeof(struct jffs2_raw_inode);
fs/jffs2/readinode.c
53
ref_offset(ref), tn->csize, ofs);
fs/jffs2/readinode.c
568
static struct jffs2_raw_node_ref *jffs2_first_valid_node(struct jffs2_raw_node_ref *ref)
fs/jffs2/readinode.c
570
while (ref && ref->next_in_ino) {
fs/jffs2/readinode.c
571
if (!ref_obsolete(ref))
fs/jffs2/readinode.c
572
return ref;
fs/jffs2/readinode.c
573
dbg_noderef("node at 0x%08x is obsoleted. Ignoring.\n", ref_offset(ref));
fs/jffs2/readinode.c
574
ref = ref->next_in_ino;
fs/jffs2/readinode.c
586
static inline int read_direntry(struct jffs2_sb_info *c, struct jffs2_raw_node_ref *ref,
fs/jffs2/readinode.c
594
BUG_ON(ref_obsolete(ref));
fs/jffs2/readinode.c
599
ref_offset(ref), je32_to_cpu(rd->node_crc), crc);
fs/jffs2/readinode.c
600
jffs2_mark_node_obsolete(c, ref);
fs/jffs2/readinode.c
605
if (ref_flags(ref) == REF_UNCHECKED) {
fs/jffs2/readinode.c
612
ref_offset(ref), rd->nsize, je32_to_cpu(rd->totlen));
fs/jffs2/readinode.c
613
jffs2_mark_node_obsolete(c, ref);
fs/jffs2/readinode.c
617
jeb = &c->blocks[ref->flash_offset / c->sector_size];
fs/jffs2/readinode.c
618
len = ref_totlen(c, jeb, ref);
fs/jffs2/readinode.c
62
ref_offset(ref), tn->csize, tn->partial_crc, tn->data_crc, ofs - len, ofs, len);
fs/jffs2/readinode.c
625
ref->flash_offset = ref_offset(ref) | dirent_node_state(rd);
fs/jffs2/readinode.c
633
fd->raw = ref;
fs/jffs2/readinode.c
661
err = jffs2_flash_read(c, (ref_offset(ref)) + read,
fs/jffs2/readinode.c
685
ref_offset(ref), je32_to_cpu(rd->node_crc), crc);
fs/jffs2/readinode.c
686
jffs2_mark_node_obsolete(c, ref);
fs/jffs2/readinode.c
713
static inline int read_dnode(struct jffs2_sb_info *c, struct jffs2_raw_node_ref *ref,
fs/jffs2/readinode.c
723
BUG_ON(ref_obsolete(ref));
fs/jffs2/readinode.c
728
ref_offset(ref), je32_to_cpu(rd->node_crc), crc);
fs/jffs2/readinode.c
729
jffs2_mark_node_obsolete(c, ref);
fs/jffs2/readinode.c
743
if (ref_flags(ref) == REF_UNCHECKED) {
fs/jffs2/readinode.c
748
JFFS2_WARNING("inode node header CRC is corrupted at %#08x\n", ref_offset(ref));
fs/jffs2/readinode.c
749
jffs2_dbg_dump_node(c, ref_offset(ref));
fs/jffs2/readinode.c
750
jffs2_mark_node_obsolete(c, ref);
fs/jffs2/readinode.c
804
ref_offset(ref), tn->partial_crc, je32_to_cpu(rd->data_crc));
fs/jffs2/readinode.c
805
jffs2_mark_node_obsolete(c, ref);
fs/jffs2/readinode.c
819
jeb = &c->blocks[ref->flash_offset / c->sector_size];
fs/jffs2/readinode.c
820
len = ref_totlen(c, jeb, ref);
fs/jffs2/readinode.c
827
ref->flash_offset = ref_offset(ref) | REF_NORMAL;
fs/jffs2/readinode.c
843
tn->fn->raw = ref;
fs/jffs2/readinode.c
857
ref_offset(ref), je32_to_cpu(rd->version),
fs/jffs2/readinode.c
888
static inline int read_unknown(struct jffs2_sb_info *c, struct jffs2_raw_node_ref *ref, struct jffs2_unknown_node *un)
fs/jffs2/readinode.c
891
if (ref_flags(ref) == REF_UNCHECKED) {
fs/jffs2/readinode.c
893
ref_offset(ref));
fs/jffs2/readinode.c
897
jffs2_mark_node_obsolete(c, ref);
fs/jffs2/readinode.c
907
je16_to_cpu(un->nodetype), ref_offset(ref));
fs/jffs2/readinode.c
914
je16_to_cpu(un->nodetype), ref_offset(ref));
fs/jffs2/readinode.c
920
je16_to_cpu(un->nodetype), ref_offset(ref));
fs/jffs2/readinode.c
925
je16_to_cpu(un->nodetype), ref_offset(ref));
fs/jffs2/readinode.c
926
jffs2_mark_node_obsolete(c, ref);
fs/jffs2/readinode.c
940
static int read_more(struct jffs2_sb_info *c, struct jffs2_raw_node_ref *ref,
fs/jffs2/readinode.c
955
offs = ref_offset(ref) + *rdlen;
fs/jffs2/readinode.c
984
struct jffs2_raw_node_ref *ref, *valid_ref;
fs/jffs2/scan.c
392
struct jffs2_xattr_ref *ref;
fs/jffs2/scan.c
414
ref = jffs2_alloc_xattr_ref();
fs/jffs2/scan.c
415
if (!ref)
fs/jffs2/scan.c
427
ref->ino = je32_to_cpu(rr->ino);
fs/jffs2/scan.c
428
ref->xid = je32_to_cpu(rr->xid);
fs/jffs2/scan.c
429
ref->xseqno = je32_to_cpu(rr->xseqno);
fs/jffs2/scan.c
430
if (ref->xseqno > c->highest_xseqno)
fs/jffs2/scan.c
431
c->highest_xseqno = (ref->xseqno & ~XREF_DELETE_MARKER);
fs/jffs2/scan.c
432
ref->next = c->xref_temp;
fs/jffs2/scan.c
433
c->xref_temp = ref;
fs/jffs2/scan.c
435
jffs2_link_node_ref(c, jeb, ofs | REF_PRISTINE, PAD(je32_to_cpu(rr->totlen)), (void *)ref);
fs/jffs2/scan.c
440
ofs, ref->xid, ref->ino);
fs/jffs2/summary.c
523
struct jffs2_xattr_ref *ref;
fs/jffs2/summary.c
532
ref = jffs2_alloc_xattr_ref();
fs/jffs2/summary.c
533
if (!ref) {
fs/jffs2/summary.c
537
ref->next = c->xref_temp;
fs/jffs2/summary.c
538
c->xref_temp = ref;
fs/jffs2/summary.c
541
PAD(sizeof(struct jffs2_raw_xref)), (void *)ref);
fs/jffs2/summary.c
543
*pseudo_random += ref->node->flash_offset;
fs/jffs2/xattr.c
1041
struct jffs2_xattr_ref *ref, **pref;
fs/jffs2/xattr.c
1050
for (ref=ic->xref, pref=&ic->xref; ref; pref=&ref->next, ref=ref->next) {
fs/jffs2/xattr.c
1051
BUG_ON(ref->ic!=ic);
fs/jffs2/xattr.c
1053
xd = ref->xd;
fs/jffs2/xattr.c
1066
*pref = ref->next;
fs/jffs2/xattr.c
1067
delete_xattr_ref(c, ref);
fs/jffs2/xattr.c
1103
struct jffs2_xattr_ref *ref, *newref, **pref;
fs/jffs2/xattr.c
1125
for (ref=ic->xref, pref=&ic->xref; ref; pref=&ref->next, ref=ref->next) {
fs/jffs2/xattr.c
1126
xd = ref->xd;
fs/jffs2/xattr.c
1132
*pref = ref->next;
fs/jffs2/xattr.c
1133
delete_xattr_ref(c, ref);
fs/jffs2/xattr.c
1144
ref->ino = ic->ino;
fs/jffs2/xattr.c
1145
ref->xid = xd->xid;
fs/jffs2/xattr.c
1146
ref->xseqno |= XREF_DELETE_MARKER;
fs/jffs2/xattr.c
1147
rc = save_xattr_ref(c, ref);
fs/jffs2/xattr.c
1149
*pref = ref->next;
fs/jffs2/xattr.c
1151
ref->next = c->xref_dead_list;
fs/jffs2/xattr.c
1152
c->xref_dead_list = ref;
fs/jffs2/xattr.c
1156
ref->ic = ic;
fs/jffs2/xattr.c
1157
ref->xd = xd;
fs/jffs2/xattr.c
1158
ref->xseqno &= ~XREF_DELETE_MARKER;
fs/jffs2/xattr.c
1194
if (ref)
fs/jffs2/xattr.c
1195
*pref = ref->next;
fs/jffs2/xattr.c
1198
if (ref) {
fs/jffs2/xattr.c
1199
ref->next = ic->xref;
fs/jffs2/xattr.c
1200
ic->xref = ref;
fs/jffs2/xattr.c
1204
} else if (ref) {
fs/jffs2/xattr.c
1205
delete_xattr_ref(c, ref);
fs/jffs2/xattr.c
1261
int jffs2_garbage_collect_xattr_ref(struct jffs2_sb_info *c, struct jffs2_xattr_ref *ref,
fs/jffs2/xattr.c
1268
BUG_ON(!ref->node);
fs/jffs2/xattr.c
1270
if (ref->node != raw)
fs/jffs2/xattr.c
1272
if (is_xattr_ref_dead(ref) && (raw->next_in_ino == (void *)ref))
fs/jffs2/xattr.c
1275
old_ofs = ref_offset(ref->node);
fs/jffs2/xattr.c
1276
totlen = ref_totlen(c, c->gcblock, ref->node);
fs/jffs2/xattr.c
1284
rc = save_xattr_ref(c, ref);
fs/jffs2/xattr.c
1287
ref->ic->ino, ref->xd->xid, old_ofs, ref_offset(ref->node));
fs/jffs2/xattr.c
1338
void jffs2_release_xattr_ref(struct jffs2_sb_info *c, struct jffs2_xattr_ref *ref)
fs/jffs2/xattr.c
1343
if (ref->node != (void *)ref)
fs/jffs2/xattr.c
1347
if (ref == tmp) {
fs/jffs2/xattr.c
1352
jffs2_free_xattr_ref(ref);
fs/jffs2/xattr.c
444
static int verify_xattr_ref(struct jffs2_sb_info *c, struct jffs2_xattr_ref *ref)
fs/jffs2/xattr.c
454
if (ref_flags(ref->node) != REF_UNCHECKED)
fs/jffs2/xattr.c
456
offset = ref_offset(ref->node);
fs/jffs2/xattr.c
482
ref->ino = je32_to_cpu(rr.ino);
fs/jffs2/xattr.c
483
ref->xid = je32_to_cpu(rr.xid);
fs/jffs2/xattr.c
484
ref->xseqno = je32_to_cpu(rr.xseqno);
fs/jffs2/xattr.c
485
if (ref->xseqno > c->highest_xseqno)
fs/jffs2/xattr.c
486
c->highest_xseqno = (ref->xseqno & ~XREF_DELETE_MARKER);
fs/jffs2/xattr.c
490
for (raw=ref->node; raw != (void *)ref; raw=raw->next_in_ino) {
fs/jffs2/xattr.c
497
raw->flash_offset = ref_offset(raw) | ((ref->node==raw) ? REF_PRISTINE : REF_NORMAL);
fs/jffs2/xattr.c
502
ref->ino, ref->xid, ref_offset(ref->node));
fs/jffs2/xattr.c
506
static int save_xattr_ref(struct jffs2_sb_info *c, struct jffs2_xattr_ref *ref)
fs/jffs2/xattr.c
520
if (is_xattr_ref_dead(ref)) {
fs/jffs2/xattr.c
522
rr.ino = cpu_to_je32(ref->ino);
fs/jffs2/xattr.c
523
rr.xid = cpu_to_je32(ref->xid);
fs/jffs2/xattr.c
525
rr.ino = cpu_to_je32(ref->ic->ino);
fs/jffs2/xattr.c
526
rr.xid = cpu_to_je32(ref->xd->xid);
fs/jffs2/xattr.c
542
ref->xseqno = xseqno;
fs/jffs2/xattr.c
543
jffs2_add_physical_node_ref(c, phys_ofs | REF_PRISTINE, PAD(sizeof(rr)), (void *)ref);
fs/jffs2/xattr.c
545
dbg_xattr("success on saving xref (ino=%u, xid=%u)\n", ref->ic->ino, ref->xd->xid);
fs/jffs2/xattr.c
554
struct jffs2_xattr_ref *ref;
fs/jffs2/xattr.c
557
ref = jffs2_alloc_xattr_ref();
fs/jffs2/xattr.c
558
if (!ref)
fs/jffs2/xattr.c
560
ref->ic = ic;
fs/jffs2/xattr.c
561
ref->xd = xd;
fs/jffs2/xattr.c
563
ret = save_xattr_ref(c, ref);
fs/jffs2/xattr.c
565
jffs2_free_xattr_ref(ref);
fs/jffs2/xattr.c
570
ref->next = ic->xref;
fs/jffs2/xattr.c
571
ic->xref = ref;
fs/jffs2/xattr.c
573
return ref; /* success */
fs/jffs2/xattr.c
576
static void delete_xattr_ref(struct jffs2_sb_info *c, struct jffs2_xattr_ref *ref)
fs/jffs2/xattr.c
581
xd = ref->xd;
fs/jffs2/xattr.c
582
ref->xseqno |= XREF_DELETE_MARKER;
fs/jffs2/xattr.c
583
ref->ino = ref->ic->ino;
fs/jffs2/xattr.c
584
ref->xid = ref->xd->xid;
fs/jffs2/xattr.c
586
ref->next = c->xref_dead_list;
fs/jffs2/xattr.c
587
c->xref_dead_list = ref;
fs/jffs2/xattr.c
591
ref->ino, ref->xid, ref->xseqno);
fs/jffs2/xattr.c
600
struct jffs2_xattr_ref *ref, *_ref;
fs/jffs2/xattr.c
606
for (ref = ic->xref; ref; ref = _ref) {
fs/jffs2/xattr.c
607
_ref = ref->next;
fs/jffs2/xattr.c
608
delete_xattr_ref(c, ref);
fs/jffs2/xattr.c
618
struct jffs2_xattr_ref *ref, *_ref;
fs/jffs2/xattr.c
621
for (ref = ic->xref; ref; ref = _ref) {
fs/jffs2/xattr.c
622
_ref = ref->next;
fs/jffs2/xattr.c
623
xd = ref->xd;
fs/jffs2/xattr.c
628
jffs2_free_xattr_ref(ref);
fs/jffs2/xattr.c
640
struct jffs2_xattr_ref *ref, *cmp, **pref, **pcmp;
fs/jffs2/xattr.c
648
for (ref=ic->xref, pref=&ic->xref; ref; pref=&ref->next, ref=ref->next) {
fs/jffs2/xattr.c
649
if (!ref->xd->xname) {
fs/jffs2/xattr.c
650
rc = load_xattr_datum(c, ref->xd);
fs/jffs2/xattr.c
652
*pref = ref->next;
fs/jffs2/xattr.c
653
delete_xattr_ref(c, ref);
fs/jffs2/xattr.c
658
for (cmp=ref->next, pcmp=&ref->next; cmp; pcmp=&cmp->next, cmp=cmp->next) {
fs/jffs2/xattr.c
660
ref->xd->flags |= JFFS2_XFLAGS_BIND;
fs/jffs2/xattr.c
662
ref->xd->flags &= ~JFFS2_XFLAGS_BIND;
fs/jffs2/xattr.c
670
if (ref->xd->xprefix == cmp->xd->xprefix
fs/jffs2/xattr.c
671
&& !strcmp(ref->xd->xname, cmp->xd->xname)) {
fs/jffs2/xattr.c
672
if (ref->xseqno > cmp->xseqno) {
fs/jffs2/xattr.c
676
*pref = ref->next;
fs/jffs2/xattr.c
677
delete_xattr_ref(c, ref);
fs/jffs2/xattr.c
743
struct jffs2_xattr_ref *ref, *_ref;
fs/jffs2/xattr.c
746
for (ref=c->xref_temp; ref; ref = _ref) {
fs/jffs2/xattr.c
747
_ref = ref->next;
fs/jffs2/xattr.c
748
jffs2_free_xattr_ref(ref);
fs/jffs2/xattr.c
751
for (ref=c->xref_dead_list; ref; ref = _ref) {
fs/jffs2/xattr.c
752
_ref = ref->next;
fs/jffs2/xattr.c
753
jffs2_free_xattr_ref(ref);
fs/jffs2/xattr.c
777
struct jffs2_xattr_ref *ref, *_ref;
fs/jffs2/xattr.c
792
for (ref=c->xref_temp; ref; ref=_ref) {
fs/jffs2/xattr.c
795
_ref = ref->next;
fs/jffs2/xattr.c
796
if (ref_flags(ref->node) != REF_PRISTINE) {
fs/jffs2/xattr.c
797
if (verify_xattr_ref(c, ref)) {
fs/jffs2/xattr.c
798
BUG_ON(ref->node->next_in_ino != (void *)ref);
fs/jffs2/xattr.c
799
ref->node->next_in_ino = NULL;
fs/jffs2/xattr.c
800
jffs2_mark_node_obsolete(c, ref->node);
fs/jffs2/xattr.c
801
jffs2_free_xattr_ref(ref);
fs/jffs2/xattr.c
806
i = (ref->ino ^ ref->xid) % XREF_TMPHASH_SIZE;
fs/jffs2/xattr.c
808
if (tmp->ino == ref->ino && tmp->xid == ref->xid)
fs/jffs2/xattr.c
812
raw = ref->node;
fs/jffs2/xattr.c
813
if (ref->xseqno > tmp->xseqno) {
fs/jffs2/xattr.c
814
tmp->xseqno = ref->xseqno;
fs/jffs2/xattr.c
821
jffs2_free_xattr_ref(ref);
fs/jffs2/xattr.c
824
ref->next = xref_tmphash[i];
fs/jffs2/xattr.c
825
xref_tmphash[i] = ref;
fs/jffs2/xattr.c
832
for (ref=xref_tmphash[i]; ref; ref=_ref) {
fs/jffs2/xattr.c
834
_ref = ref->next;
fs/jffs2/xattr.c
835
if (is_xattr_ref_dead(ref)) {
fs/jffs2/xattr.c
836
ref->next = c->xref_dead_list;
fs/jffs2/xattr.c
837
c->xref_dead_list = ref;
fs/jffs2/xattr.c
843
xd = jffs2_find_xattr_datum(c, ref->xid);
fs/jffs2/xattr.c
844
ic = jffs2_get_ino_cache(c, ref->ino);
fs/jffs2/xattr.c
847
ref->ino, ref->xid, ref->xseqno);
fs/jffs2/xattr.c
848
ref->xseqno |= XREF_DELETE_MARKER;
fs/jffs2/xattr.c
849
ref->next = c->xref_dead_list;
fs/jffs2/xattr.c
850
c->xref_dead_list = ref;
fs/jffs2/xattr.c
854
ref->xd = xd;
fs/jffs2/xattr.c
855
ref->ic = ic;
fs/jffs2/xattr.c
857
ref->next = ic->xref;
fs/jffs2/xattr.c
858
ic->xref = ref;
fs/jffs2/xattr.c
971
struct jffs2_xattr_ref *ref, **pref;
fs/jffs2/xattr.c
984
for (ref=ic->xref, pref=&ic->xref; ref; pref=&ref->next, ref=ref->next) {
fs/jffs2/xattr.c
985
BUG_ON(ref->ic != ic);
fs/jffs2/xattr.c
986
xd = ref->xd;
fs/jffs2/xattr.c
997
*pref = ref->next;
fs/jffs2/xattr.c
998
delete_xattr_ref(c, ref);
fs/jffs2/xattr.h
66
static inline int is_xattr_ref_dead(struct jffs2_xattr_ref *ref)
fs/jffs2/xattr.h
68
return ((ref->xseqno & XREF_DELETE_MARKER) != 0);
fs/jffs2/xattr.h
86
extern int jffs2_garbage_collect_xattr_ref(struct jffs2_sb_info *c, struct jffs2_xattr_ref *ref,
fs/jffs2/xattr.h
90
extern void jffs2_release_xattr_ref(struct jffs2_sb_info *c, struct jffs2_xattr_ref *ref);
fs/netfs/fscache_cache.c
119
refcount_read(&candidate->ref),
fs/netfs/fscache_cache.c
184
int ref;
fs/netfs/fscache_cache.c
190
zero = __refcount_dec_and_test(&cache->ref, &ref);
fs/netfs/fscache_cache.c
191
trace_fscache_cache(debug_id, ref - 1, where);
fs/netfs/fscache_cache.c
249
trace_fscache_access_cache(cache->debug_id, refcount_read(&cache->ref),
fs/netfs/fscache_cache.c
298
trace_fscache_access_cache(cache->debug_id, refcount_read(&cache->ref),
fs/netfs/fscache_cache.c
321
trace_fscache_access_cache(cache->debug_id, refcount_read(&cache->ref),
fs/netfs/fscache_cache.c
366
trace_fscache_access_cache(cache->debug_id, refcount_read(&cache->ref),
fs/netfs/fscache_cache.c
37
refcount_set(&cache->ref, 1);
fs/netfs/fscache_cache.c
396
refcount_read(&cache->ref),
fs/netfs/fscache_cache.c
48
int ref;
fs/netfs/fscache_cache.c
50
success = __refcount_inc_not_zero(&cache->ref, &ref);
fs/netfs/fscache_cache.c
52
trace_fscache_cache(cache->debug_id, ref + 1, where);
fs/netfs/fscache_cookie.c
1013
int ref;
fs/netfs/fscache_cookie.c
1015
zero = __refcount_dec_and_test(&cookie->ref, &ref);
fs/netfs/fscache_cookie.c
1016
trace_fscache_cookie(cookie_debug_id, ref - 1, where);
fs/netfs/fscache_cookie.c
1030
int ref;
fs/netfs/fscache_cookie.c
1032
__refcount_inc(&cookie->ref, &ref);
fs/netfs/fscache_cookie.c
1033
trace_fscache_cookie(cookie->debug_id, ref + 1, where);
fs/netfs/fscache_cookie.c
106
trace_fscache_access(cookie->debug_id, refcount_read(&cookie->ref),
fs/netfs/fscache_cookie.c
1131
refcount_read(&cookie->ref),
fs/netfs/fscache_cookie.c
128
trace_fscache_access(cookie->debug_id, refcount_read(&cookie->ref),
fs/netfs/fscache_cookie.c
148
trace_fscache_access(cookie->debug_id, refcount_read(&cookie->ref),
fs/netfs/fscache_cookie.c
266
trace_fscache_cookie(cookie->debug_id, refcount_read(&cookie->ref),
fs/netfs/fscache_cookie.c
359
refcount_set(&cookie->ref, 1);
fs/netfs/fscache_cookie.c
435
trace_fscache_cookie(cursor->debug_id, refcount_read(&cursor->ref),
fs/netfs/fscache_cookie.c
584
trace_fscache_active(cookie->debug_id, refcount_read(&cookie->ref),
fs/netfs/fscache_cookie.c
669
unsigned int r = refcount_read(&cookie->ref);
fs/netfs/fscache_cookie.c
684
r = refcount_read(&cookie->ref);
fs/netfs/fscache_cookie.c
854
trace_fscache_access(cookie->debug_id, refcount_read(&cookie->ref),
fs/netfs/fscache_io.c
123
trace_fscache_access(cookie->debug_id, refcount_read(&cookie->ref),
fs/netfs/fscache_volume.c
128
refcount_read(&volume->ref),
fs/netfs/fscache_volume.c
23
int ref;
fs/netfs/fscache_volume.c
243
refcount_set(&volume->ref, 1);
fs/netfs/fscache_volume.c
25
__refcount_inc(&volume->ref, &ref);
fs/netfs/fscache_volume.c
26
trace_fscache_volume(volume->debug_id, ref + 1, where);
fs/netfs/fscache_volume.c
33
int ref;
fs/netfs/fscache_volume.c
35
if (!__refcount_inc_not_zero(&volume->ref, &ref))
fs/netfs/fscache_volume.c
38
trace_fscache_volume(volume->debug_id, ref + 1, where);
fs/netfs/fscache_volume.c
426
int ref;
fs/netfs/fscache_volume.c
428
zero = __refcount_dec_and_test(&volume->ref, &ref);
fs/netfs/fscache_volume.c
429
trace_fscache_volume(debug_id, ref - 1, where);
fs/netfs/fscache_volume.c
46
int ref = refcount_read(&volume->ref);
fs/netfs/fscache_volume.c
472
refcount_read(&volume->ref),
fs/netfs/fscache_volume.c
48
trace_fscache_volume(volume->debug_id, ref, where);
fs/netfs/fscache_volume.c
499
refcount_read(&volume->ref),
fs/netfs/fscache_volume.c
63
refcount_read(&volume->ref),
fs/netfs/internal.h
103
refcount_read(&subreq->ref), what);
fs/netfs/internal.h
245
refcount_inc(&netfs_group->ref);
fs/netfs/internal.h
256
refcount_dec_and_test(&netfs_group->ref))
fs/netfs/internal.h
267
refcount_sub_and_test(nr, &netfs_group->ref))
fs/netfs/internal.h
355
trace_fscache_cookie(cookie->debug_id, refcount_read(&cookie->ref),
fs/netfs/internal.h
96
trace_netfs_rreq_ref(rreq->debug_id, refcount_read(&rreq->ref), what);
fs/netfs/main.c
72
refcount_read(&rreq->ref),
fs/netfs/objects.c
169
dead = __refcount_dec_and_test(&rreq->ref, &r);
fs/netfs/objects.c
182
int r = refcount_read(&rreq->ref);
fs/netfs/objects.c
214
refcount_set(&subreq->ref, 2);
fs/netfs/objects.c
227
__refcount_inc(&subreq->ref, &r);
fs/netfs/objects.c
252
dead = __refcount_dec_and_test(&subreq->ref, &r);
fs/netfs/objects.c
55
refcount_set(&rreq->ref, 2);
fs/netfs/objects.c
80
trace_netfs_rreq_ref(rreq->debug_id, refcount_read(&rreq->ref), netfs_rreq_trace_new);
fs/netfs/objects.c
90
__refcount_inc(&rreq->ref, &r);
fs/netfs/read_retry.c
203
refcount_read(&subreq->ref),
fs/netfs/write_collect.c
27
rreq->debug_id, refcount_read(&rreq->ref), rreq->flags,
fs/netfs/write_collect.c
47
refcount_read(&sreq->ref), sreq->flags);
fs/netfs/write_retry.c
152
refcount_read(&subreq->ref),
fs/nfs/callback_proc.c
470
struct referring_call *ref;
fs/nfs/callback_proc.c
487
ref = &rclist->rcl_refcalls[j];
fs/nfs/callback_proc.c
489
status = nfs4_slot_wait_on_seqid(tbl, ref->rc_slotid,
fs/nfs/callback_proc.c
490
ref->rc_sequenceid, HZ >> 1) < 0;
fs/nfs/dns_resolve.c
113
static void nfs_dns_ent_put(struct kref *ref)
fs/nfs/dns_resolve.c
117
item = container_of(ref, struct nfs_dns_ent, h.ref);
fs/nfs/flexfilelayout/flexfilelayout.c
249
if (refcount_inc_not_zero(&pos->ref)) {
fs/nfs/flexfilelayout/flexfilelayout.c
283
refcount_set(&mirror->ref, 1);
fs/nfs/flexfilelayout/flexfilelayout.c
2907
if (!refcount_inc_not_zero(&mirror->ref))
fs/nfs/flexfilelayout/flexfilelayout.c
324
if (mirror != NULL && refcount_dec_and_test(&mirror->ref))
fs/nfs/flexfilelayout/flexfilelayout.h
99
refcount_t ref;
fs/nfs/nfs42xattr.c
235
kref_init(&entry->ref);
fs/nfs/nfs42xattr.c
259
entry = container_of(kref, struct nfs4_xattr_entry, ref);
fs/nfs/nfs42xattr.c
273
cache = container_of(kref, struct nfs4_xattr_cache, ref);
fs/nfs/nfs42xattr.c
296
kref_init(&cache->ref);
fs/nfs/nfs42xattr.c
330
kref_put(&old->ref, nfs4_xattr_free_entry_cb);
fs/nfs/nfs42xattr.c
401
kref_put(&entry->ref, nfs4_xattr_free_entry_cb);
fs/nfs/nfs42xattr.c
408
kref_put(&cache->ref, nfs4_xattr_free_cache_cb);
fs/nfs/nfs42xattr.c
443
kref_get(&cache->ref);
fs/nfs/nfs42xattr.c
462
kref_put(&cache->ref, nfs4_xattr_free_cache_cb);
fs/nfs/nfs42xattr.c
472
kref_get(&newcache->ref);
fs/nfs/nfs42xattr.c
474
kref_get(&cache->ref);
fs/nfs/nfs42xattr.c
488
kref_put(&cache->ref, nfs4_xattr_free_cache_cb);
fs/nfs/nfs42xattr.c
558
kref_put(&oldentry->ref, nfs4_xattr_free_entry_cb);
fs/nfs/nfs42xattr.c
583
kref_put(&entry->ref, nfs4_xattr_free_entry_cb);
fs/nfs/nfs42xattr.c
598
kref_get(&entry->ref);
fs/nfs/nfs42xattr.c
634
kref_put(&entry->ref, nfs4_xattr_free_entry_cb);
fs/nfs/nfs42xattr.c
640
kref_put(&cache->ref, nfs4_xattr_free_cache_cb);
fs/nfs/nfs42xattr.c
678
kref_put(&cache->ref, nfs4_xattr_free_cache_cb);
fs/nfs/nfs42xattr.c
69
struct kref ref;
fs/nfs/nfs42xattr.c
708
kref_put(&entry->ref, nfs4_xattr_free_entry_cb);
fs/nfs/nfs42xattr.c
711
kref_put(&cache->ref, nfs4_xattr_free_cache_cb);
fs/nfs/nfs42xattr.c
733
kref_put(&cache->ref, nfs4_xattr_free_cache_cb);
fs/nfs/nfs42xattr.c
761
kref_put(&entry->ref, nfs4_xattr_free_entry_cb);
fs/nfs/nfs42xattr.c
764
kref_put(&cache->ref, nfs4_xattr_free_cache_cb);
fs/nfs/nfs42xattr.c
80
struct kref ref;
fs/nfs/nfs42xattr.c
825
kref_get(&cache->ref);
fs/nfs/nfs42xattr.c
852
kref_put(&cache->ref, nfs4_xattr_free_cache_cb);
fs/nfs/nfs42xattr.c
895
kref_get(&entry->ref);
fs/nfs/nfs42xattr.c
907
kref_get(&entry->ref);
fs/nfs/nfs42xattr.c
942
kref_put(&entry->ref, nfs4_xattr_free_entry_cb);
fs/nfs/nfs42xattr.c
943
kref_put(&entry->ref, nfs4_xattr_free_entry_cb);
fs/nfs/pnfs.h
375
atomic_t ref;
fs/nfs/pnfs.h
451
atomic_inc(&d->ref);
fs/nfs/pnfs_dev.c
178
if (d != NULL && !atomic_inc_not_zero(&d->ref))
fs/nfs/pnfs_dev.c
208
atomic_inc(&new->ref);
fs/nfs/pnfs_dev.c
260
atomic_set(&d->ref, 1);
fs/nfs/pnfs_dev.c
278
if (atomic_add_unless(&d->ref, -1, 2))
fs/nfs/pnfs_dev.c
282
if (!atomic_dec_and_test(&d->ref))
fs/nfs/pnfs_dev.c
336
if (d->nfs_client == clp && atomic_read(&d->ref)) {
fs/nfs/pnfs_dev.c
88
if (atomic_read(&d->ref))
fs/nfs_common/nfslocalio.c
134
kref_get(&dom->ref);
fs/nfsd/export.c
229
kref_get(&item->ek_client->ref);
fs/nfsd/export.c
381
static void svc_export_put(struct kref *ref)
fs/nfsd/export.c
383
struct svc_export *exp = container_of(ref, struct svc_export, h.ref);
fs/nfsd/export.c
57
static void expkey_put(struct kref *ref)
fs/nfsd/export.c
59
struct svc_expkey *key = container_of(ref, struct svc_expkey, h.ref);
fs/nfsd/export.c
842
kref_get(&item->ex_client->ref);
fs/nfsd/nfs4idmap.c
91
ent_put(struct kref *ref)
fs/nfsd/nfs4idmap.c
93
struct ent *map = container_of(ref, struct ent, h.ref);
fs/nfsd/nfs4state.c
7137
static __be32 check_stateid_generation(stateid_t *in, stateid_t *ref, bool has_session)
fs/nfsd/nfs4state.c
7146
if (in->si_generation == ref->si_generation)
fs/nfsd/nfs4state.c
7150
if (nfsd4_stateid_generation_after(in, ref))
fs/nfsd/nfssvc.c
221
static void nfsd_net_done(struct percpu_ref *ref)
fs/nfsd/nfssvc.c
223
struct nfsd_net *nn = container_of(ref, struct nfsd_net, nfsd_net_ref);
fs/nfsd/nfssvc.c
228
static void nfsd_net_free(struct percpu_ref *ref)
fs/nfsd/nfssvc.c
230
struct nfsd_net *nn = container_of(ref, struct nfsd_net, nfsd_net_ref);
fs/ntfs3/attrlist.c
296
u8 name_len, CLST svcn, __le16 id, const struct MFT_REF *ref,
fs/ntfs3/attrlist.c
344
le->ref = *ref;
fs/ntfs3/dir.c
265
inode = ntfs_iget5(sb, &e->ref, uni);
fs/ntfs3/dir.c
296
ino = ino_get(&e->ref);
fs/ntfs3/dir.c
337
struct inode *inode = ntfs_iget5(sbi->sb, &e->ref, NULL);
fs/ntfs3/frecord.c
1336
struct MFT_REF ref;
fs/ntfs3/frecord.c
1338
mi_get_ref(&ni->mi, &ref);
fs/ntfs3/frecord.c
1345
if (memcmp(&ref, &le->ref, sizeof(struct MFT_REF)))
fs/ntfs3/frecord.c
1532
struct MFT_REF ref;
fs/ntfs3/frecord.c
1538
mi_get_ref(&ni->mi, &ref);
fs/ntfs3/frecord.c
1539
ntfs_remove_reparse(sbi, 0, &ref);
fs/ntfs3/frecord.c
179
rno = ino_get(&le->ref);
fs/ntfs3/frecord.c
2065
struct MFT_REF ref;
fs/ntfs3/frecord.c
2067
mi_get_ref(&ni->mi, &ref);
fs/ntfs3/frecord.c
2068
ntfs_remove_reparse(sbi, 0, &ref);
fs/ntfs3/frecord.c
2691
mi_get_ref(&ni->mi, &de2->ref);
fs/ntfs3/frecord.c
2709
mi_get_ref(&ni->mi, &de->ref);
fs/ntfs3/frecord.c
2746
mi_get_ref(&ni->mi, &de->ref);
fs/ntfs3/frecord.c
299
CLST rno = ino_get(&le->ref);
fs/ntfs3/frecord.c
418
struct MFT_REF ref;
fs/ntfs3/frecord.c
420
mi_get_ref(mi, &ref);
fs/ntfs3/frecord.c
424
&ref, &le);
fs/ntfs3/frecord.c
458
le->ref = ref;
fs/ntfs3/frecord.c
617
struct MFT_REF ref;
fs/ntfs3/frecord.c
636
mi_get_ref(&ni->mi, &ref);
fs/ntfs3/frecord.c
640
if (!memcmp(&le->ref, &ref, sizeof(ref)))
fs/ntfs3/frecord.c
646
mi = ni_find_mi(ni, ino_get(&le->ref));
fs/ntfs3/frecord.c
677
if (!memcmp(&le->ref, &ref, sizeof(ref)))
fs/ntfs3/frecord.c
680
mi = ni_find_mi(ni, ino_get(&le->ref));
fs/ntfs3/frecord.c
716
if (!memcmp(&le->ref, &ref, sizeof(ref)))
fs/ntfs3/frecord.c
719
mi = ni_find_mi(ni, ino_get(&le->ref));
fs/ntfs3/frecord.c
778
mi_get_ref(&ni->mi, &le->ref);
fs/ntfs3/frecord.c
794
le->ref = ni->attr_list.le->ref;
fs/ntfs3/frecord.c
846
mi_get_ref(mi, &le_b[nb]->ref);
fs/ntfs3/fslog.c
120
struct MFT_REF ref; // 0x10: File Reference of file containing attribute
fs/ntfs3/fslog.c
129
struct MFT_REF ref; // 0x08:
fs/ntfs3/fslog.c
2978
if (ino_get(&oe->ref) != rno)
fs/ntfs3/fslog.c
4331
oe->ref = oe0.ref;
fs/ntfs3/fslog.c
4611
oe->ref = oe0->ref;
fs/ntfs3/fslog.c
4733
inode = ntfs_iget5(sbi->sb, &oe->ref, NULL);
fs/ntfs3/fslog.c
4847
rno = ino_get(&oe->ref);
fs/ntfs3/fslog.c
5195
rno = ino_get(&oe->ref);
fs/ntfs3/fsntfs.c
1867
struct MFT_REF ref;
fs/ntfs3/fsntfs.c
1880
ref.low = cpu_to_le32(MFT_REC_SECURE);
fs/ntfs3/fsntfs.c
1881
ref.high = 0;
fs/ntfs3/fsntfs.c
1882
ref.seq = cpu_to_le16(MFT_REC_SECURE);
fs/ntfs3/fsntfs.c
1884
inode = ntfs_iget5(sb, &ref, &NAME_SECURE);
fs/ntfs3/fsntfs.c
214
struct MFT_REF ref;
fs/ntfs3/fsntfs.c
221
ref.low = cpu_to_le32(MFT_REC_EXTEND);
fs/ntfs3/fsntfs.c
222
ref.high = 0;
fs/ntfs3/fsntfs.c
223
ref.seq = cpu_to_le16(MFT_REC_EXTEND);
fs/ntfs3/fsntfs.c
224
inode = ntfs_iget5(sb, &ref, &NAME_EXTEND);
fs/ntfs3/fsntfs.c
2378
const struct MFT_REF *ref)
fs/ntfs3/fsntfs.c
2395
memcpy(&re.key.ref, ref, sizeof(*ref));
fs/ntfs3/fsntfs.c
2408
const struct MFT_REF *ref)
fs/ntfs3/fsntfs.c
2422
rkey.ref = *ref;
fs/ntfs3/fsntfs.c
2449
if (memcmp(&re->key.ref, ref, sizeof(*ref))) {
fs/ntfs3/fsntfs.c
281
struct MFT_REF ref;
fs/ntfs3/fsntfs.c
293
ref.low = cpu_to_le32(MFT_REC_MFT);
fs/ntfs3/fsntfs.c
294
ref.high = 0;
fs/ntfs3/fsntfs.c
295
ref.seq = cpu_to_le16(1);
fs/ntfs3/fsntfs.c
297
inode = ntfs_iget5(sb, &ref, NULL);
fs/ntfs3/fsntfs.c
307
inode = ntfs_iget5(sb, &ref, NULL);
fs/ntfs3/fsntfs.c
585
struct MFT_REF ref;
fs/ntfs3/fsntfs.c
633
ref.high = 0;
fs/ntfs3/fsntfs.c
639
ref.low = cpu_to_le32(ir);
fs/ntfs3/fsntfs.c
640
ref.seq = cpu_to_le16(ir);
fs/ntfs3/fsntfs.c
642
i = ntfs_iget5(sb, &ref, NULL);
fs/ntfs3/fsntfs.c
648
ref.low);
fs/ntfs3/index.c
2637
e->ref.low = 0;
fs/ntfs3/index.c
2638
e->ref.high = 0;
fs/ntfs3/index.c
2639
e->ref.seq = 0;
fs/ntfs3/inode.c
1426
mi_get_ref(&ni->mi, &new_de->ref);
fs/ntfs3/inode.c
1632
&new_de->ref);
fs/ntfs3/inode.c
1746
ntfs_remove_reparse(sbi, IO_REPARSE_TAG_SYMLINK, &new_de->ref);
fs/ntfs3/inode.c
26
const struct MFT_REF *ref)
fs/ntfs3/inode.c
516
struct MFT_REF *ref = data;
fs/ntfs3/inode.c
518
return ino_get(ref) == inode->i_ino;
fs/ntfs3/inode.c
523
const struct MFT_REF *ref = data;
fs/ntfs3/inode.c
525
inode->i_ino = ino_get(ref);
fs/ntfs3/inode.c
529
struct inode *ntfs_iget5(struct super_block *sb, const struct MFT_REF *ref,
fs/ntfs3/inode.c
534
inode = iget5_locked(sb, ino_get(ref), ntfs_test_inode, ntfs_set_inode,
fs/ntfs3/inode.c
535
(void *)ref);
fs/ntfs3/inode.c
541
inode = ntfs_read_mft(inode, name, ref);
fs/ntfs3/inode.c
542
else if (ref->seq != ntfs_i(inode)->mi.mrec->seq) {
fs/ntfs3/inode.c
80
} else if (ref->seq != rec->seq) {
fs/ntfs3/inode.c
83
le16_to_cpu(ref->seq), le16_to_cpu(rec->seq));
fs/ntfs3/namei.c
29
e->ref.high = fname->home.high = 0;
fs/ntfs3/ntfs.h
187
static inline CLST ino_get(const struct MFT_REF *ref)
fs/ntfs3/ntfs.h
190
return le32_to_cpu(ref->low) | ((u64)le16_to_cpu(ref->high) << 32);
fs/ntfs3/ntfs.h
192
return le32_to_cpu(ref->low);
fs/ntfs3/ntfs.h
525
struct MFT_REF ref; // 0x10: MFT record number with attribute.
fs/ntfs3/ntfs.h
613
struct MFT_REF ref; // 0x00: MFT record number with this file.
fs/ntfs3/ntfs.h
869
struct MFT_REF ref; // 0x20: MFT record number with this file.
fs/ntfs3/ntfs.h
954
struct MFT_REF ref; // 0x04: MFT record number with this file
fs/ntfs3/ntfs.h
957
static_assert(offsetof(struct REPARSE_KEY, ref) == 0x04);
fs/ntfs3/ntfs_fs.h
509
u8 name_len, CLST svcn, __le16 id, const struct MFT_REF *ref,
fs/ntfs3/ntfs_fs.h
719
const struct MFT_REF *ref);
fs/ntfs3/ntfs_fs.h
721
const struct MFT_REF *ref);
fs/ntfs3/ntfs_fs.h
774
struct inode *ntfs_iget5(struct super_block *sb, const struct MFT_REF *ref,
fs/ntfs3/ntfs_fs.h
832
const struct MFT_REF *ref)
fs/ntfs3/ntfs_fs.h
834
if (le32_to_cpu(ref->low) != mi->rno)
fs/ntfs3/ntfs_fs.h
836
if (ref->seq != mi->mrec->seq)
fs/ntfs3/ntfs_fs.h
840
return le16_to_cpu(ref->high) == (mi->rno >> 32);
fs/ntfs3/ntfs_fs.h
842
return !ref->high;
fs/ntfs3/ntfs_fs.h
846
static inline void mi_get_ref(const struct mft_inode *mi, struct MFT_REF *ref)
fs/ntfs3/ntfs_fs.h
848
ref->low = cpu_to_le32(mi->rno);
fs/ntfs3/ntfs_fs.h
850
ref->high = cpu_to_le16(mi->rno >> 32);
fs/ntfs3/ntfs_fs.h
852
ref->high = 0;
fs/ntfs3/ntfs_fs.h
854
ref->seq = mi->mrec->seq;
fs/ntfs3/super.c
1264
struct MFT_REF ref;
fs/ntfs3/super.c
1268
ref.high = 0;
fs/ntfs3/super.c
1320
ref.low = cpu_to_le32(MFT_REC_VOL);
fs/ntfs3/super.c
1321
ref.seq = cpu_to_le16(MFT_REC_VOL);
fs/ntfs3/super.c
1322
inode = ntfs_iget5(sb, &ref, &NAME_VOLUME);
fs/ntfs3/super.c
1368
ref.low = cpu_to_le32(MFT_REC_MIRR);
fs/ntfs3/super.c
1369
ref.seq = cpu_to_le16(MFT_REC_MIRR);
fs/ntfs3/super.c
1370
inode = ntfs_iget5(sb, &ref, &NAME_MIRROR);
fs/ntfs3/super.c
1383
ref.low = cpu_to_le32(MFT_REC_LOG);
fs/ntfs3/super.c
1384
ref.seq = cpu_to_le16(MFT_REC_LOG);
fs/ntfs3/super.c
1385
inode = ntfs_iget5(sb, &ref, &NAME_LOGFILE);
fs/ntfs3/super.c
1413
ref.low = cpu_to_le32(MFT_REC_MFT);
fs/ntfs3/super.c
1414
ref.seq = cpu_to_le16(1);
fs/ntfs3/super.c
1416
inode = ntfs_iget5(sb, &ref, &NAME_MFT);
fs/ntfs3/super.c
1442
ref.low = cpu_to_le32(MFT_REC_BITMAP);
fs/ntfs3/super.c
1443
ref.seq = cpu_to_le16(MFT_REC_BITMAP);
fs/ntfs3/super.c
1444
inode = ntfs_iget5(sb, &ref, &NAME_BITMAP);
fs/ntfs3/super.c
1482
ref.low = cpu_to_le32(MFT_REC_BADCLUST);
fs/ntfs3/super.c
1483
ref.seq = cpu_to_le16(MFT_REC_BADCLUST);
fs/ntfs3/super.c
1484
inode = ntfs_iget5(sb, &ref, &NAME_BADCLUS);
fs/ntfs3/super.c
1520
ref.low = cpu_to_le32(MFT_REC_ATTR);
fs/ntfs3/super.c
1521
ref.seq = cpu_to_le16(MFT_REC_ATTR);
fs/ntfs3/super.c
1522
inode = ntfs_iget5(sb, &ref, &NAME_ATTRDEF);
fs/ntfs3/super.c
1584
ref.low = cpu_to_le32(MFT_REC_UPCASE);
fs/ntfs3/super.c
1585
ref.seq = cpu_to_le16(MFT_REC_UPCASE);
fs/ntfs3/super.c
1586
inode = ntfs_iget5(sb, &ref, &NAME_UPCASE);
fs/ntfs3/super.c
1655
ref.low = cpu_to_le32(MFT_REC_ROOT);
fs/ntfs3/super.c
1656
ref.seq = cpu_to_le16(MFT_REC_ROOT);
fs/ntfs3/super.c
1657
inode = ntfs_iget5(sb, &ref, &NAME_ROOT);
fs/ntfs3/super.c
878
struct MFT_REF ref;
fs/ntfs3/super.c
881
ref.low = cpu_to_le32(ino);
fs/ntfs3/super.c
883
ref.high = cpu_to_le16(ino >> 32);
fs/ntfs3/super.c
885
ref.high = 0;
fs/ntfs3/super.c
887
ref.seq = cpu_to_le16(generation);
fs/ntfs3/super.c
889
inode = ntfs_iget5(sb, &ref, NULL);
fs/ocfs2/xattr.c
6066
struct ocfs2_xattr_tree_value_refcount_para *ref =
fs/ocfs2/xattr.c
6105
&et, ref->ref_ci,
fs/ocfs2/xattr.c
6106
ref->ref_root_bh,
fs/ocfs2/xattr.c
6107
ref->dealloc, p);
fs/orangefs/inode.c
1002
struct orangefs_object_kref *ref = (struct orangefs_object_kref *) data;
fs/orangefs/inode.c
1003
ORANGEFS_I(inode)->refn.fs_id = ref->fs_id;
fs/orangefs/inode.c
1004
ORANGEFS_I(inode)->refn.khandle = ref->khandle;
fs/orangefs/inode.c
1017
struct orangefs_object_kref *ref = (struct orangefs_object_kref *) data;
fs/orangefs/inode.c
1023
&(ref->khandle)) &&
fs/orangefs/inode.c
1024
orangefs_inode->refn.fs_id == ref->fs_id);
fs/orangefs/inode.c
1035
struct orangefs_object_kref *ref)
fs/orangefs/inode.c
1041
hash = orangefs_handle_hash(ref);
fs/orangefs/inode.c
1046
ref);
fs/orangefs/inode.c
1066
&ref->khandle,
fs/orangefs/inode.c
1067
ref->fs_id,
fs/orangefs/inode.c
1078
umode_t mode, dev_t dev, struct orangefs_object_kref *ref)
fs/orangefs/inode.c
1081
unsigned long hash = orangefs_handle_hash(ref);
fs/orangefs/inode.c
1101
orangefs_set_inode(inode, ref);
fs/orangefs/inode.c
1124
error = insert_inode_locked4(inode, hash, orangefs_test_inode, ref);
fs/orangefs/inode.c
990
static inline ino_t orangefs_handle_hash(struct orangefs_object_kref *ref)
fs/orangefs/inode.c
992
if (!ref)
fs/orangefs/inode.c
994
return orangefs_khandle_to_ino(&(ref->khandle));
fs/orangefs/namei.c
222
struct orangefs_object_kref ref;
fs/orangefs/namei.c
26
struct orangefs_object_kref ref;
fs/orangefs/namei.c
261
ref = new_op->downcall.resp.sym.refn;
fs/orangefs/namei.c
263
inode = orangefs_new_inode(dir->i_sb, dir, S_IFLNK | mode, 0, &ref);
fs/orangefs/namei.c
305
struct orangefs_object_kref ref;
fs/orangefs/namei.c
334
ref = new_op->downcall.resp.mkdir.refn;
fs/orangefs/namei.c
336
inode = orangefs_new_inode(dir->i_sb, dir, S_IFDIR | mode, 0, &ref);
fs/orangefs/namei.c
59
ref = new_op->downcall.resp.create.refn;
fs/orangefs/namei.c
61
inode = orangefs_new_inode(dir->i_sb, dir, S_IFREG | mode, 0, &ref);
fs/orangefs/orangefs-kernel.h
351
struct orangefs_object_kref *ref);
fs/orangefs/orangefs-kernel.h
375
struct orangefs_object_kref *ref);
fs/quota/quota_tree.c
340
__le32 *ref;
fs/quota/quota_tree.c
369
ref = (__le32 *)buf;
fs/quota/quota_tree.c
370
newblk = le32_to_cpu(ref[get_index(info, dquot->dq_id, depth)]);
fs/quota/quota_tree.c
394
le32_to_cpu(ref[get_index(info,
fs/quota/quota_tree.c
405
ref[get_index(info, dquot->dq_id, depth)] =
fs/quota/quota_tree.c
550
__le32 *ref = (__le32 *)buf;
fs/quota/quota_tree.c
561
newblk = le32_to_cpu(ref[get_index(info, dquot->dq_id, depth)]);
fs/quota/quota_tree.c
584
ref[get_index(info, dquot->dq_id, depth)] = cpu_to_le32(0);
fs/quota/quota_tree.c
586
for (i = 0; i < (info->dqi_usable_bs >> 2) && !ref[i]; i++)
fs/quota/quota_tree.c
665
__le32 *ref = (__le32 *)buf;
fs/quota/quota_tree.c
678
blk = le32_to_cpu(ref[get_index(info, dquot->dq_id, depth)]);
fs/quota/quota_tree.c
796
__le32 *ref = (__le32 *)buf;
fs/quota/quota_tree.c
815
uint blk_no = le32_to_cpu(ref[i]);
fs/smb/client/cached_dir.c
17
static void smb2_close_cached_fid(struct kref *ref);
fs/smb/client/cached_dir.c
445
smb2_close_cached_fid(struct kref *ref)
fs/smb/client/cached_dir.c
448
struct cached_fid *cfid = container_of(ref, struct cached_fid,
fs/smb/client/cifs_swn.c
368
static void cifs_swn_reg_release(struct kref *ref)
fs/smb/client/cifs_swn.c
370
struct cifs_swn_reg *swnreg = container_of(ref, struct cifs_swn_reg, ref_count);
fs/smb/client/cifsglob.h
1037
release_iface(struct kref *ref)
fs/smb/client/cifsglob.h
1039
struct cifs_server_iface *iface = container_of(ref,
fs/smb/client/cifsglob.h
2204
const struct dfs_info3_param *ref)
fs/smb/client/cifsglob.h
2210
return is_tcon_dfs(tcon) || (ref && (ref->flags & DFSREF_REFERRAL_SERVER));
fs/smb/client/dfs.c
23
int dfs_parse_target_referral(const char *full_path, const struct dfs_info3_param *ref,
fs/smb/client/dfs.c
30
if (!full_path || !*full_path || !ref || !ctx)
fs/smb/client/dfs.c
33
if (WARN_ON_ONCE(!ref->node_name || ref->path_consumed < 0))
fs/smb/client/dfs.c
36
if (strlen(full_path) - ref->path_consumed) {
fs/smb/client/dfs.c
37
prepath = full_path + ref->path_consumed;
fs/smb/client/dfs.c
394
struct dfs_info3_param ref = {0};
fs/smb/client/dfs.c
43
path = cifs_build_devname(ref->node_name, prepath);
fs/smb/client/dfs.c
439
dfs_cache_noreq_find(server->leaf_fullpath + 1, &ref, &tl)) {
fs/smb/client/dfs.c
445
rc = tree_connect_dfs_target(xid, tcon, cifs_sb, tree, ref.server_type == DFS_TYPE_LINK,
fs/smb/client/dfs.c
447
free_dfs_info_param(&ref);
fs/smb/client/dfs.h
102
if (IS_ERR(ref->tit))
fs/smb/client/dfs.h
105
if (!ref->tit)
fs/smb/client/dfs.h
106
tit = dfs_cache_get_tgt_iterator(&ref->tl);
fs/smb/client/dfs.h
108
tit = dfs_cache_get_next_tgt(&ref->tl, ref->tit);
fs/smb/client/dfs.h
111
ref->tit = ERR_PTR(-ENOENT);
fs/smb/client/dfs.h
114
ref->tit = tit;
fs/smb/client/dfs.h
115
return ref->tit;
fs/smb/client/dfs.h
135
struct dfs_ref *ref = ref_walk_start(rw);
fs/smb/client/dfs.h
137
for (; ref <= ref_walk_cur(rw); ref++) {
fs/smb/client/dfs.h
138
if (WARN_ON_ONCE(!ref->ses))
fs/smb/client/dfs.h
140
list_add(&ref->ses->dlist, &tcon->dfs_ses_list);
fs/smb/client/dfs.h
141
ref->ses = NULL;
fs/smb/client/dfs.h
147
struct dfs_ref *ref = ref_walk_cur(rw) - 1;
fs/smb/client/dfs.h
149
WARN_ON_ONCE(ref < ref_walk_start(rw));
fs/smb/client/dfs.h
150
dfs_cache_noreq_update_tgthint(ref->path + 1, ref->tit);
fs/smb/client/dfs.h
151
ref->tit = ERR_PTR(-ENOENT); /* end marker */
fs/smb/client/dfs.h
155
const struct dfs_info3_param *ref,
fs/smb/client/dfs.h
30
struct dfs_ref *ref;
fs/smb/client/dfs.h
36
#define ref_walk_cur(w) ((w)->ref)
fs/smb/client/dfs.h
63
static inline void __ref_walk_free(struct dfs_ref *ref)
fs/smb/client/dfs.h
65
kfree(ref->path);
fs/smb/client/dfs.h
66
kfree(ref->full_path);
fs/smb/client/dfs.h
67
dfs_cache_free_tgts(&ref->tl);
fs/smb/client/dfs.h
68
if (ref->ses)
fs/smb/client/dfs.h
69
cifs_put_smb_ses(ref->ses);
fs/smb/client/dfs.h
70
memset(ref, 0, sizeof(*ref));
fs/smb/client/dfs.h
75
struct dfs_ref *ref;
fs/smb/client/dfs.h
80
for (ref = ref_walk_start(rw); ref <= ref_walk_end(rw); ref++)
fs/smb/client/dfs.h
81
__ref_walk_free(ref);
fs/smb/client/dfs.h
87
struct dfs_ref *ref = ref_walk_cur(rw) + 1;
fs/smb/client/dfs.h
89
if (ref > ref_walk_end(rw))
fs/smb/client/dfs.h
91
__ref_walk_free(ref);
fs/smb/client/dfs.h
92
ref_walk_cur(rw) = ref;
fs/smb/client/dfs.h
99
struct dfs_ref *ref = ref_walk_cur(rw);
fs/smb/client/dfs_cache.c
1009
rc = setup_referral(path, ce, ref, it->it_name);
fs/smb/client/dfs_cache.c
257
const struct dfs_info3_param *ref = &refs[i];
fs/smb/client/dfs_cache.c
268
ref->flags, ref->path_consumed, ref->server_type,
fs/smb/client/dfs_cache.c
269
ref->ref_flag, ref->path_name, ref->node_name,
fs/smb/client/dfs_cache.c
270
ref->ttl, ref->ttl / 60);
fs/smb/client/dfs_cache.c
662
struct dfs_info3_param *ref = *refs;
fs/smb/client/dfs_cache.c
665
convert_delimiter(ref[i].path_name, '\\');
fs/smb/client/dfs_cache.c
755
struct dfs_info3_param *ref, const char *target)
fs/smb/client/dfs_cache.c
761
memset(ref, 0, sizeof(*ref));
fs/smb/client/dfs_cache.c
763
ref->path_name = kstrdup(path, GFP_ATOMIC);
fs/smb/client/dfs_cache.c
764
if (!ref->path_name)
fs/smb/client/dfs_cache.c
767
ref->node_name = kstrdup(target, GFP_ATOMIC);
fs/smb/client/dfs_cache.c
768
if (!ref->node_name) {
fs/smb/client/dfs_cache.c
773
ref->path_consumed = ce->path_consumed;
fs/smb/client/dfs_cache.c
774
ref->ttl = ce->ttl;
fs/smb/client/dfs_cache.c
775
ref->server_type = ce->srvtype;
fs/smb/client/dfs_cache.c
776
ref->ref_flag = ce->ref_flags;
fs/smb/client/dfs_cache.c
777
ref->flags = ce->hdr_flags;
fs/smb/client/dfs_cache.c
782
kfree(ref->path_name);
fs/smb/client/dfs_cache.c
783
ref->path_name = NULL;
fs/smb/client/dfs_cache.c
855
int remap, const char *path, struct dfs_info3_param *ref,
fs/smb/client/dfs_cache.c
872
if (ref)
fs/smb/client/dfs_cache.c
873
rc = setup_referral(path, ce, ref, get_tgt_name(ce));
fs/smb/client/dfs_cache.c
902
int dfs_cache_noreq_find(const char *path, struct dfs_info3_param *ref,
fs/smb/client/dfs_cache.c
918
if (ref)
fs/smb/client/dfs_cache.c
919
rc = setup_referral(path, ce, ref, get_tgt_name(ce));
fs/smb/client/dfs_cache.c
989
struct dfs_info3_param *ref)
fs/smb/client/dfs_cache.c
994
if (!it || !ref)
fs/smb/client/dfs_cache.h
42
struct dfs_info3_param *ref,
fs/smb/client/dfs_cache.h
44
int dfs_cache_noreq_find(const char *path, struct dfs_info3_param *ref,
fs/smb/client/dfs_cache.h
50
struct dfs_info3_param *ref);
fs/smb/client/misc.c
680
struct dfs_referral_level_3 *ref;
fs/smb/client/misc.c
708
ref = (struct dfs_referral_level_3 *) &(rsp->referrals);
fs/smb/client/misc.c
709
if (ref->VersionNumber != cpu_to_le16(3)) {
fs/smb/client/misc.c
711
le16_to_cpu(ref->VersionNumber));
fs/smb/client/misc.c
751
node->server_type = le16_to_cpu(ref->ServerType);
fs/smb/client/misc.c
752
node->ref_flag = le16_to_cpu(ref->ReferralEntryFlags);
fs/smb/client/misc.c
755
temp = (char *)ref + le16_to_cpu(ref->DfsPathOffset);
fs/smb/client/misc.c
765
temp = (char *)ref + le16_to_cpu(ref->NetworkAddressOffset);
fs/smb/client/misc.c
774
node->ttl = le32_to_cpu(ref->TimeToLive);
fs/smb/client/misc.c
776
ref++;
fs/smb/client/trace.h
1743
TP_PROTO(unsigned int tcon_debug_id, int ref,
fs/smb/client/trace.h
1745
TP_ARGS(tcon_debug_id, ref, trace),
fs/smb/client/trace.h
1748
__field(int, ref)
fs/smb/client/trace.h
1753
__entry->ref = ref;
fs/smb/client/trace.h
1759
__entry->ref)
fs/tracefs/event_inode.c
102
static void release_ei(struct kref *ref)
fs/tracefs/event_inode.c
104
struct eventfs_inode *ei = container_of(ref, struct eventfs_inode, kref);
fs/ubifs/debug.c
431
const struct ubifs_ref_node *ref = node;
fs/ubifs/debug.c
433
pr_err("\tlnum %u\n", le32_to_cpu(ref->lnum));
fs/ubifs/debug.c
434
pr_err("\toffs %u\n", le32_to_cpu(ref->offs));
fs/ubifs/debug.c
435
pr_err("\tjhead %u\n", le32_to_cpu(ref->jhead));
fs/ubifs/log.c
168
struct ubifs_ref_node *ref;
fs/ubifs/log.c
173
ref = kzalloc(c->ref_node_alsz, GFP_NOFS);
fs/ubifs/log.c
174
if (!ref) {
fs/ubifs/log.c
229
ref->ch.node_type = UBIFS_REF_NODE;
fs/ubifs/log.c
230
ref->lnum = cpu_to_le32(bud->lnum);
fs/ubifs/log.c
231
ref->offs = cpu_to_le32(bud->start);
fs/ubifs/log.c
232
ref->jhead = cpu_to_le32(jhead);
fs/ubifs/log.c
262
err = ubifs_write_node(c, ref, UBIFS_REF_NODE_SZ, c->lhead_lnum,
fs/ubifs/log.c
267
err = ubifs_shash_update(c, c->log_hash, ref, UBIFS_REF_NODE_SZ);
fs/ubifs/log.c
280
kfree(ref);
fs/ubifs/log.c
285
kfree(ref);
fs/ubifs/log.c
360
struct ubifs_ref_node *ref;
fs/ubifs/log.c
402
ref = buf + len;
fs/ubifs/log.c
403
ref->ch.node_type = UBIFS_REF_NODE;
fs/ubifs/log.c
404
ref->lnum = cpu_to_le32(lnum);
fs/ubifs/log.c
405
ref->offs = cpu_to_le32(offs);
fs/ubifs/log.c
406
ref->jhead = cpu_to_le32(i);
fs/ubifs/log.c
408
ubifs_prepare_node(c, ref, UBIFS_REF_NODE_SZ, 0);
fs/ubifs/log.c
411
err = ubifs_shash_update(c, c->log_hash, ref,
fs/ubifs/log.c
666
struct ubifs_ref_node *ref = snod->node;
fs/ubifs/log.c
667
int ref_lnum = le32_to_cpu(ref->lnum);
fs/ubifs/replay.c
1083
const struct ubifs_ref_node *ref = snod->node;
fs/ubifs/replay.c
1085
err = validate_ref(c, ref);
fs/ubifs/replay.c
1091
err = ubifs_shash_update(c, c->log_hash, ref,
fs/ubifs/replay.c
1096
err = add_replay_bud(c, le32_to_cpu(ref->lnum),
fs/ubifs/replay.c
1097
le32_to_cpu(ref->offs),
fs/ubifs/replay.c
1098
le32_to_cpu(ref->jhead),
fs/ubifs/replay.c
944
static int validate_ref(struct ubifs_info *c, const struct ubifs_ref_node *ref)
fs/ubifs/replay.c
947
int lnum = le32_to_cpu(ref->lnum);
fs/ubifs/replay.c
948
unsigned int offs = le32_to_cpu(ref->offs);
fs/ubifs/replay.c
949
unsigned int jhead = le32_to_cpu(ref->jhead);
fs/xfs/xfs_healthmon.c
118
refcount_inc(&hm->ref);
fs/xfs/xfs_healthmon.c
1200
refcount_set(&hm->ref, 1);
fs/xfs/xfs_healthmon.c
1257
ASSERT(refcount_read(&hm->ref) == 1);
fs/xfs/xfs_healthmon.c
73
if (hm && !refcount_inc_not_zero(&hm->ref))
fs/xfs/xfs_healthmon.c
89
if (refcount_dec_and_test(&hm->ref)) {
fs/xfs/xfs_healthmon.h
28
refcount_t ref;
include/asm-generic/vmlinux.lds.h
375
*(.ref.data) \
include/asm-generic/vmlinux.lds.h
546
*(.ref.rodata) \
include/asm-generic/vmlinux.lds.h
617
*(.ref.text)
include/drm/drm_atomic.h
473
struct kref ref;
include/drm/drm_atomic.h
651
kref_get(&commit->ref);
include/drm/drm_atomic.h
664
kref_put(&commit->ref, __drm_crtc_commit_free);
include/drm/drm_atomic.h
682
kref_get(&state->ref);
include/drm/drm_atomic.h
686
void __drm_atomic_state_free(struct kref *ref);
include/drm/drm_atomic.h
697
kref_put(&state->ref, __drm_atomic_state_free);
include/drm/drm_atomic.h
87
struct kref ref;
include/drm/drm_device.h
80
struct kref ref;
include/drm/drm_pagemap.h
166
struct kref ref;
include/drm/drm_pagemap.h
284
kref_get(&dpagemap->ref);
include/drm/drm_pagemap.h
300
return (dpagemap && kref_get_unless_zero(&dpagemap->ref)) ? dpagemap : NULL;
include/linux/alloc_tag.h
153
static inline void alloc_tag_add_check(union codetag_ref *ref, struct alloc_tag *tag)
include/linux/alloc_tag.h
155
WARN_ONCE(ref && ref->ct && !is_codetag_empty(ref),
include/linux/alloc_tag.h
157
ref->ct->filename, ref->ct->lineno);
include/linux/alloc_tag.h
162
static inline void alloc_tag_sub_check(union codetag_ref *ref)
include/linux/alloc_tag.h
164
WARN_ONCE(ref && !ref->ct, "alloc_tag was not set\n");
include/linux/alloc_tag.h
167
static inline void alloc_tag_add_check(union codetag_ref *ref, struct alloc_tag *tag) {}
include/linux/alloc_tag.h
168
static inline void alloc_tag_sub_check(union codetag_ref *ref) {}
include/linux/alloc_tag.h
172
static inline bool __alloc_tag_ref_set(union codetag_ref *ref, struct alloc_tag *tag)
include/linux/alloc_tag.h
174
alloc_tag_add_check(ref, tag);
include/linux/alloc_tag.h
175
if (!ref || !tag)
include/linux/alloc_tag.h
178
ref->ct = &tag->ct;
include/linux/alloc_tag.h
182
static inline bool alloc_tag_ref_set(union codetag_ref *ref, struct alloc_tag *tag)
include/linux/alloc_tag.h
184
if (unlikely(!__alloc_tag_ref_set(ref, tag)))
include/linux/alloc_tag.h
197
static inline void alloc_tag_add(union codetag_ref *ref, struct alloc_tag *tag, size_t bytes)
include/linux/alloc_tag.h
199
if (likely(alloc_tag_ref_set(ref, tag)))
include/linux/alloc_tag.h
203
static inline void alloc_tag_sub(union codetag_ref *ref, size_t bytes)
include/linux/alloc_tag.h
207
alloc_tag_sub_check(ref);
include/linux/alloc_tag.h
208
if (!ref || !ref->ct)
include/linux/alloc_tag.h
211
if (is_codetag_empty(ref)) {
include/linux/alloc_tag.h
212
ref->ct = NULL;
include/linux/alloc_tag.h
216
tag = ct_to_alloc_tag(ref->ct);
include/linux/alloc_tag.h
221
ref->ct = NULL;
include/linux/alloc_tag.h
240
static inline void alloc_tag_add(union codetag_ref *ref, struct alloc_tag *tag,
include/linux/alloc_tag.h
242
static inline void alloc_tag_sub(union codetag_ref *ref, size_t bytes) {}
include/linux/alloc_tag.h
52
static inline bool is_codetag_empty(union codetag_ref *ref)
include/linux/alloc_tag.h
54
return ref->ct == CODETAG_EMPTY;
include/linux/alloc_tag.h
57
static inline void set_codetag_empty(union codetag_ref *ref)
include/linux/alloc_tag.h
59
if (ref)
include/linux/alloc_tag.h
60
ref->ct = CODETAG_EMPTY;
include/linux/alloc_tag.h
65
static inline bool is_codetag_empty(union codetag_ref *ref) { return false; }
include/linux/alloc_tag.h
67
static inline void set_codetag_empty(union codetag_ref *ref)
include/linux/alloc_tag.h
69
if (ref)
include/linux/alloc_tag.h
70
ref->ct = NULL;
include/linux/bpf_mprog.h
148
struct bpf_prog *ref;
include/linux/bpf_mprog.h
218
WARN_ON_ONCE(entry->parent->ref);
include/linux/bpf_mprog.h
220
entry->parent->ref = tuple->prog;
include/linux/bpf_mprog.h
231
if (entry->parent->ref) {
include/linux/bpf_mprog.h
232
bpf_prog_put(entry->parent->ref);
include/linux/bpf_mprog.h
233
entry->parent->ref = NULL;
include/linux/cdx/mcdi.h
109
struct kref ref;
include/linux/ceph/string_table.h
20
extern void ceph_release_string(struct kref *ref);
include/linux/dim.h
32
#define IS_SIGNIFICANT_DIFF(val, ref) \
include/linux/dim.h
33
((ref) && (((100UL * abs((val) - (ref))) / (ref)) > 10))
include/linux/dmaengine.h
869
struct kref ref;
include/linux/file_ref.h
108
return !atomic_long_add_negative(1, &ref->refcnt);
include/linux/file_ref.h
118
static __always_inline void file_ref_inc(file_ref_t *ref)
include/linux/file_ref.h
120
long prior = atomic_long_fetch_inc_relaxed(&ref->refcnt);
include/linux/file_ref.h
139
static __always_inline __must_check bool file_ref_put(file_ref_t *ref)
include/linux/file_ref.h
157
cnt = atomic_long_dec_return(&ref->refcnt);
include/linux/file_ref.h
160
return __file_ref_put(ref, cnt);
include/linux/file_ref.h
178
static __always_inline __must_check bool file_ref_put_close(file_ref_t *ref)
include/linux/file_ref.h
182
old = atomic_long_read(&ref->refcnt);
include/linux/file_ref.h
184
if (likely(atomic_long_try_cmpxchg(&ref->refcnt, &old, FILE_REF_DEAD)))
include/linux/file_ref.h
187
return file_ref_put(ref);
include/linux/file_ref.h
196
static inline unsigned long file_ref_read(file_ref_t *ref)
include/linux/file_ref.h
198
unsigned long c = atomic_long_read(&ref->refcnt);
include/linux/file_ref.h
213
static inline unsigned long __file_ref_read_raw(file_ref_t *ref)
include/linux/file_ref.h
215
return atomic_long_read(&ref->refcnt);
include/linux/file_ref.h
59
static inline void file_ref_init(file_ref_t *ref, unsigned long cnt)
include/linux/file_ref.h
61
atomic_long_set(&ref->refcnt, cnt - 1);
include/linux/file_ref.h
64
bool __file_ref_put(file_ref_t *ref, unsigned long cnt);
include/linux/file_ref.h
78
static __always_inline __must_check bool file_ref_get(file_ref_t *ref)
include/linux/fscache-cache.h
40
refcount_t ref;
include/linux/fscache.h
103
refcount_t ref;
include/linux/fscache.h
73
refcount_t ref;
include/linux/hid.h
723
struct kref ref;
include/linux/hid.h
732
void hiddev_free(struct kref *ref);
include/linux/host1x.h
126
struct kref ref;
include/linux/host1x.h
141
static inline struct host1x_bo_mapping *to_host1x_bo_mapping(struct kref *ref)
include/linux/host1x.h
143
return container_of(ref, struct host1x_bo_mapping, ref);
include/linux/host1x.h
269
struct kref ref;
include/linux/host1x.h
471
refcount_t ref;
include/linux/hugetlb.h
110
void resv_map_release(struct kref *ref);
include/linux/hw_random.h
51
struct kref ref;
include/linux/iio/buffer_impl.h
166
struct kref ref;
include/linux/interrupt.h
276
void (*release)(struct kref *ref);
include/linux/irqdesc.h
29
unsigned int ref;
include/linux/lockd/lockd.h
272
typedef int (*nlm_host_match_fn_t)(void *cur, struct nlm_host *ref);
include/linux/memcontrol.h
1465
union codetag_ref ref;
include/linux/memcontrol.h
70
refcount_t ref;
include/linux/memremap.h
135
struct percpu_ref ref;
include/linux/memremap.h
307
percpu_ref_put(&pgmap->ref);
include/linux/mm_inline.h
451
void pfnmap_track_ctx_release(struct kref *ref);
include/linux/mtd/blktrans.h
29
struct kref ref;
include/linux/netfilter/ipset/ip_set.h
254
u32 ref;
include/linux/netfs.h
181
refcount_t ref;
include/linux/netfs.h
266
refcount_t ref;
include/linux/netfs.h
82
refcount_t ref;
include/linux/objpool.h
91
refcount_t ref;
include/linux/percpu-refcount.h
102
struct percpu_ref *ref;
include/linux/percpu-refcount.h
121
int __must_check percpu_ref_init(struct percpu_ref *ref,
include/linux/percpu-refcount.h
124
void percpu_ref_exit(struct percpu_ref *ref);
include/linux/percpu-refcount.h
125
void percpu_ref_switch_to_atomic(struct percpu_ref *ref,
include/linux/percpu-refcount.h
127
void percpu_ref_switch_to_atomic_sync(struct percpu_ref *ref);
include/linux/percpu-refcount.h
128
void percpu_ref_switch_to_percpu(struct percpu_ref *ref);
include/linux/percpu-refcount.h
129
void percpu_ref_kill_and_confirm(struct percpu_ref *ref,
include/linux/percpu-refcount.h
131
void percpu_ref_resurrect(struct percpu_ref *ref);
include/linux/percpu-refcount.h
132
void percpu_ref_reinit(struct percpu_ref *ref);
include/linux/percpu-refcount.h
133
bool percpu_ref_is_zero(struct percpu_ref *ref);
include/linux/percpu-refcount.h
147
static inline void percpu_ref_kill(struct percpu_ref *ref)
include/linux/percpu-refcount.h
149
percpu_ref_kill_and_confirm(ref, NULL);
include/linux/percpu-refcount.h
158
static inline bool __ref_is_percpu(struct percpu_ref *ref,
include/linux/percpu-refcount.h
174
percpu_ptr = READ_ONCE(ref->percpu_count_ptr);
include/linux/percpu-refcount.h
198
static inline void percpu_ref_get_many(struct percpu_ref *ref, unsigned long nr)
include/linux/percpu-refcount.h
204
if (__ref_is_percpu(ref, &percpu_count))
include/linux/percpu-refcount.h
207
atomic_long_add(nr, &ref->data->count);
include/linux/percpu-refcount.h
220
static inline void percpu_ref_get(struct percpu_ref *ref)
include/linux/percpu-refcount.h
222
percpu_ref_get_many(ref, 1);
include/linux/percpu-refcount.h
235
static inline bool percpu_ref_tryget_many(struct percpu_ref *ref,
include/linux/percpu-refcount.h
243
if (__ref_is_percpu(ref, &percpu_count)) {
include/linux/percpu-refcount.h
247
ret = atomic_long_add_unless(&ref->data->count, nr, 0);
include/linux/percpu-refcount.h
264
static inline bool percpu_ref_tryget(struct percpu_ref *ref)
include/linux/percpu-refcount.h
266
return percpu_ref_tryget_many(ref, 1);
include/linux/percpu-refcount.h
275
static inline bool percpu_ref_tryget_live_rcu(struct percpu_ref *ref)
include/linux/percpu-refcount.h
282
if (likely(__ref_is_percpu(ref, &percpu_count))) {
include/linux/percpu-refcount.h
285
} else if (!(ref->percpu_count_ptr & __PERCPU_REF_DEAD)) {
include/linux/percpu-refcount.h
286
ret = atomic_long_inc_not_zero(&ref->data->count);
include/linux/percpu-refcount.h
306
static inline bool percpu_ref_tryget_live(struct percpu_ref *ref)
include/linux/percpu-refcount.h
311
ret = percpu_ref_tryget_live_rcu(ref);
include/linux/percpu-refcount.h
326
static inline void percpu_ref_put_many(struct percpu_ref *ref, unsigned long nr)
include/linux/percpu-refcount.h
332
if (__ref_is_percpu(ref, &percpu_count))
include/linux/percpu-refcount.h
334
else if (unlikely(atomic_long_sub_and_test(nr, &ref->data->count)))
include/linux/percpu-refcount.h
335
ref->data->release(ref);
include/linux/percpu-refcount.h
349
static inline void percpu_ref_put(struct percpu_ref *ref)
include/linux/percpu-refcount.h
351
percpu_ref_put_many(ref, 1);
include/linux/percpu-refcount.h
363
static inline bool percpu_ref_is_dying(struct percpu_ref *ref)
include/linux/percpu-refcount.h
365
return ref->percpu_count_ptr & __PERCPU_REF_DEAD;
include/linux/pgalloc_tag.h
101
static inline bool get_page_tag_ref(struct page *page, union codetag_ref *ref,
include/linux/pgalloc_tag.h
112
idx_to_ref(idx, ref);
include/linux/pgalloc_tag.h
123
ref->ct = tmp->ct;
include/linux/pgalloc_tag.h
124
handle->ref = tmp;
include/linux/pgalloc_tag.h
132
if (WARN_ON(!handle.ref))
include/linux/pgalloc_tag.h
136
page_ext_put((void *)handle.ref - page_alloc_tagging_ops.offset);
include/linux/pgalloc_tag.h
139
static inline void update_page_tag_ref(union pgtag_ref_handle handle, union codetag_ref *ref)
include/linux/pgalloc_tag.h
147
if (WARN_ON(!page || !ref))
include/linux/pgalloc_tag.h
150
idx = (unsigned long)ref_to_idx(ref);
include/linux/pgalloc_tag.h
159
if (WARN_ON(!handle.ref || !ref))
include/linux/pgalloc_tag.h
162
handle.ref->ct = ref->ct;
include/linux/pgalloc_tag.h
180
union codetag_ref ref;
include/linux/pgalloc_tag.h
182
if (get_page_tag_ref(page, &ref, &handle)) {
include/linux/pgalloc_tag.h
183
alloc_tag_sub_check(&ref);
include/linux/pgalloc_tag.h
184
if (ref.ct)
include/linux/pgalloc_tag.h
185
tag = ct_to_alloc_tag(ref.ct);
include/linux/pgalloc_tag.h
24
union codetag_ref *ref; /* reference in page extension */
include/linux/pgalloc_tag.h
63
static inline void idx_to_ref(pgalloc_tag_idx idx, union codetag_ref *ref)
include/linux/pgalloc_tag.h
67
ref->ct = NULL;
include/linux/pgalloc_tag.h
70
set_codetag_empty(ref);
include/linux/pgalloc_tag.h
74
ref->ct = idx < kernel_tags.count ?
include/linux/pgalloc_tag.h
81
static inline pgalloc_tag_idx ref_to_idx(union codetag_ref *ref)
include/linux/pgalloc_tag.h
85
if (!ref->ct)
include/linux/pgalloc_tag.h
88
if (is_codetag_empty(ref))
include/linux/pgalloc_tag.h
91
tag = ct_to_alloc_tag(ref->ct);
include/linux/rcuref.h
109
cnt = atomic_sub_return_release(1, &ref->refcnt);
include/linux/rcuref.h
117
return rcuref_put_slowpath(ref, cnt);
include/linux/rcuref.h
142
static inline __must_check bool rcuref_put_rcusafe(rcuref_t *ref)
include/linux/rcuref.h
144
return __rcuref_put(ref);
include/linux/rcuref.h
168
static inline __must_check bool rcuref_put(rcuref_t *ref)
include/linux/rcuref.h
173
released = __rcuref_put(ref);
include/linux/rcuref.h
24
static inline void rcuref_init(rcuref_t *ref, unsigned int cnt)
include/linux/rcuref.h
26
atomic_set(&ref->refcnt, cnt - 1);
include/linux/rcuref.h
39
static inline unsigned int rcuref_read(rcuref_t *ref)
include/linux/rcuref.h
41
unsigned int c = atomic_read(&ref->refcnt);
include/linux/rcuref.h
56
static inline bool rcuref_is_dead(rcuref_t *ref)
include/linux/rcuref.h
58
unsigned int c = atomic_read(&ref->refcnt);
include/linux/rcuref.h
63
extern __must_check bool rcuref_get_slowpath(rcuref_t *ref);
include/linux/rcuref.h
81
static inline __must_check bool rcuref_get(rcuref_t *ref)
include/linux/rcuref.h
87
if (likely(!atomic_add_negative_relaxed(1, &ref->refcnt)))
include/linux/rcuref.h
91
return rcuref_get_slowpath(ref);
include/linux/rcuref.h
94
extern __must_check bool rcuref_put_slowpath(rcuref_t *ref, unsigned int cnt);
include/linux/rcuref.h
99
static __always_inline __must_check bool __rcuref_put(rcuref_t *ref)
include/linux/sched/topology.h
67
atomic_t ref;
include/linux/sh_clk.h
19
struct kref ref;
include/linux/sunrpc/cache.h
197
kref_get(&h->ref);
include/linux/sunrpc/cache.h
203
if (kref_get_unless_zero(&h->ref))
include/linux/sunrpc/cache.h
210
if (kref_read(&h->ref) <= 2 &&
include/linux/sunrpc/cache.h
213
kref_put(&h->ref, cd->cache_put);
include/linux/sunrpc/cache.h
56
struct kref ref;
include/linux/sunrpc/svcauth.h
80
struct kref ref;
include/linux/usb.h
355
struct kref ref; /* reference counter */
include/linux/usb.h
362
container_of(r, struct usb_interface_cache, ref)
include/media/dvbdev.h
164
struct kref ref;
include/media/v4l2-device.h
55
struct kref ref;
include/media/v4l2-device.h
69
kref_get(&v4l2_dev->ref);
include/net/act_api.h
226
int ref, bool terse);
include/net/bluetooth/l2cap.h
674
struct kref ref;
include/rdma/ib_verbs.h
1572
struct kref ref;
include/rdma/ib_verbs.h
2363
struct kref ref;
include/rdma/rdma_vt.h
178
struct kref ref;
include/rdma/uverbs_types.h
133
kref_get(&uobject->ref);
include/rdma/uverbs_types.h
156
struct kref ref;
include/trace/events/afs.h
1535
TP_PROTO(unsigned int server_debug_id, int ref, int active,
include/trace/events/afs.h
1538
TP_ARGS(server_debug_id, ref, active, reason),
include/trace/events/afs.h
1542
__field(int, ref)
include/trace/events/afs.h
1549
__entry->ref = ref;
include/trace/events/afs.h
1557
__entry->ref,
include/trace/events/afs.h
1562
TP_PROTO(unsigned int debug_id, afs_volid_t vid, int ref,
include/trace/events/afs.h
1565
TP_ARGS(debug_id, vid, ref, reason),
include/trace/events/afs.h
1570
__field(int, ref)
include/trace/events/afs.h
1577
__entry->ref = ref;
include/trace/events/afs.h
1585
__entry->ref)
include/trace/events/afs.h
1589
TP_PROTO(unsigned int cell_debug_id, int ref, int active,
include/trace/events/afs.h
1592
TP_ARGS(cell_debug_id, ref, active, reason),
include/trace/events/afs.h
1596
__field(int, ref)
include/trace/events/afs.h
1603
__entry->ref = ref;
include/trace/events/afs.h
1611
__entry->ref,
include/trace/events/afs.h
1616
TP_PROTO(unsigned int alist_debug_id, int ref, enum afs_alist_trace reason),
include/trace/events/afs.h
1618
TP_ARGS(alist_debug_id, ref, reason),
include/trace/events/afs.h
1622
__field(int, ref)
include/trace/events/afs.h
1629
__entry->ref = ref;
include/trace/events/afs.h
1636
__entry->ref)
include/trace/events/afs.h
1641
int ref, enum afs_estate_trace reason),
include/trace/events/afs.h
1643
TP_ARGS(server_debug_id, estate_debug_id, ref, reason),
include/trace/events/afs.h
1648
__field(int, ref)
include/trace/events/afs.h
1656
__entry->ref = ref;
include/trace/events/afs.h
1664
__entry->ref)
include/trace/events/afs.h
696
int ref, int outstanding, const void *where),
include/trace/events/afs.h
698
TP_ARGS(call_debug_id, op, ref, outstanding, where),
include/trace/events/afs.h
703
__field(int, ref)
include/trace/events/afs.h
711
__entry->ref = ref;
include/trace/events/afs.h
719
__entry->ref,
include/trace/events/btrfs.h
857
const struct btrfs_delayed_ref_node *ref),
include/trace/events/btrfs.h
859
TP_ARGS(fs_info, ref),
include/trace/events/btrfs.h
873
__entry->bytenr = ref->bytenr;
include/trace/events/btrfs.h
874
__entry->num_bytes = ref->num_bytes;
include/trace/events/btrfs.h
875
__entry->action = ref->action;
include/trace/events/btrfs.h
876
__entry->parent = ref->parent;
include/trace/events/btrfs.h
877
__entry->ref_root = ref->ref_root;
include/trace/events/btrfs.h
878
__entry->level = ref->tree_ref.level;
include/trace/events/btrfs.h
879
__entry->type = ref->type;
include/trace/events/btrfs.h
880
__entry->seq = ref->seq;
include/trace/events/btrfs.h
898
const struct btrfs_delayed_ref_node *ref),
include/trace/events/btrfs.h
900
TP_ARGS(fs_info, ref)
include/trace/events/btrfs.h
906
const struct btrfs_delayed_ref_node *ref),
include/trace/events/btrfs.h
908
TP_ARGS(fs_info, ref)
include/trace/events/btrfs.h
914
const struct btrfs_delayed_ref_node *ref),
include/trace/events/btrfs.h
916
TP_ARGS(fs_info, ref),
include/trace/events/btrfs.h
931
__entry->bytenr = ref->bytenr;
include/trace/events/btrfs.h
932
__entry->num_bytes = ref->num_bytes;
include/trace/events/btrfs.h
933
__entry->action = ref->action;
include/trace/events/btrfs.h
934
__entry->parent = ref->parent;
include/trace/events/btrfs.h
935
__entry->ref_root = ref->ref_root;
include/trace/events/btrfs.h
936
__entry->owner = ref->data_ref.objectid;
include/trace/events/btrfs.h
937
__entry->offset = ref->data_ref.offset;
include/trace/events/btrfs.h
938
__entry->type = ref->type;
include/trace/events/btrfs.h
939
__entry->seq = ref->seq;
include/trace/events/btrfs.h
959
const struct btrfs_delayed_ref_node *ref),
include/trace/events/btrfs.h
961
TP_ARGS(fs_info, ref)
include/trace/events/btrfs.h
967
const struct btrfs_delayed_ref_node *ref),
include/trace/events/btrfs.h
969
TP_ARGS(fs_info, ref)
include/trace/events/fscache.h
263
int ref,
include/trace/events/fscache.h
266
TP_ARGS(cookie_debug_id, ref, where),
include/trace/events/fscache.h
270
__field(int, ref )
include/trace/events/fscache.h
276
__entry->ref = ref;
include/trace/events/fscache.h
283
__entry->ref)
include/trace/events/fscache.h
288
int ref,
include/trace/events/fscache.h
293
TP_ARGS(cookie_debug_id, ref, n_active, n_accesses, why),
include/trace/events/fscache.h
297
__field(int, ref )
include/trace/events/fscache.h
305
__entry->ref = ref;
include/trace/events/fscache.h
314
__entry->ref,
include/trace/events/fscache.h
321
int ref,
include/trace/events/fscache.h
325
TP_ARGS(cache_debug_id, ref, n_accesses, why),
include/trace/events/fscache.h
329
__field(int, ref )
include/trace/events/fscache.h
336
__entry->ref = ref;
include/trace/events/fscache.h
344
__entry->ref,
include/trace/events/fscache.h
351
int ref,
include/trace/events/fscache.h
355
TP_ARGS(volume_debug_id, cookie_debug_id, ref, n_accesses, why),
include/trace/events/fscache.h
360
__field(int, ref )
include/trace/events/fscache.h
368
__entry->ref = ref;
include/trace/events/fscache.h
377
__entry->ref,
include/trace/events/fscache.h
383
int ref,
include/trace/events/fscache.h
387
TP_ARGS(cookie_debug_id, ref, n_accesses, why),
include/trace/events/fscache.h
391
__field(int, ref )
include/trace/events/fscache.h
398
__entry->ref = ref;
include/trace/events/fscache.h
406
__entry->ref,
include/trace/events/fscache.h
425
__entry->v_ref = refcount_read(&cookie->volume->ref);
include/trace/events/fscache.h
442
__field(int, ref )
include/trace/events/fscache.h
451
__entry->ref = refcount_read(&cookie->ref);
include/trace/events/fscache.h
458
__entry->cookie, __entry->volume, __entry->ref,
include/trace/events/netfs.h
433
TP_PROTO(unsigned int rreq_debug_id, int ref,
include/trace/events/netfs.h
436
TP_ARGS(rreq_debug_id, ref, what),
include/trace/events/netfs.h
440
__field(int, ref)
include/trace/events/netfs.h
446
__entry->ref = ref;
include/trace/events/netfs.h
453
__entry->ref)
include/trace/events/netfs.h
458
int ref, enum netfs_sreq_ref_trace what),
include/trace/events/netfs.h
460
TP_ARGS(rreq_debug_id, subreq_debug_index, ref, what),
include/trace/events/netfs.h
465
__field(int, ref)
include/trace/events/netfs.h
472
__entry->ref = ref;
include/trace/events/netfs.h
480
__entry->ref)
include/trace/events/rxrpc.h
2335
int ref, enum rxrpc_txbuf_trace what),
include/trace/events/rxrpc.h
2337
TP_ARGS(debug_id, call_debug_id, seq, ref, what),
include/trace/events/rxrpc.h
2343
__field(int, ref)
include/trace/events/rxrpc.h
2351
__entry->ref = ref;
include/trace/events/rxrpc.h
2360
__entry->ref)
include/trace/events/rxrpc.h
682
int ref, int usage),
include/trace/events/rxrpc.h
684
TP_ARGS(local_debug_id, op, ref, usage),
include/trace/events/rxrpc.h
689
__field(int, ref)
include/trace/events/rxrpc.h
696
__entry->ref = ref;
include/trace/events/rxrpc.h
703
__entry->ref,
include/trace/events/rxrpc.h
722
TP_PROTO(unsigned int peer_debug_id, int ref, enum rxrpc_peer_trace why),
include/trace/events/rxrpc.h
724
TP_ARGS(peer_debug_id, ref, why),
include/trace/events/rxrpc.h
728
__field(int, ref)
include/trace/events/rxrpc.h
734
__entry->ref = ref;
include/trace/events/rxrpc.h
741
__entry->ref)
include/trace/events/rxrpc.h
745
TP_PROTO(unsigned int bundle_debug_id, int ref, enum rxrpc_bundle_trace why),
include/trace/events/rxrpc.h
747
TP_ARGS(bundle_debug_id, ref, why),
include/trace/events/rxrpc.h
751
__field(int, ref)
include/trace/events/rxrpc.h
757
__entry->ref = ref;
include/trace/events/rxrpc.h
764
__entry->ref)
include/trace/events/rxrpc.h
768
TP_PROTO(unsigned int conn_debug_id, int ref, enum rxrpc_conn_trace why),
include/trace/events/rxrpc.h
770
TP_ARGS(conn_debug_id, ref, why),
include/trace/events/rxrpc.h
774
__field(int, ref)
include/trace/events/rxrpc.h
780
__entry->ref = ref;
include/trace/events/rxrpc.h
787
__entry->ref)
include/trace/events/rxrpc.h
807
__entry->usage = conn ? refcount_read(&conn->ref) : -2;
include/trace/events/rxrpc.h
821
TP_PROTO(unsigned int call_debug_id, int ref, unsigned long aux,
include/trace/events/rxrpc.h
824
TP_ARGS(call_debug_id, ref, aux, why),
include/trace/events/rxrpc.h
828
__field(int, ref)
include/trace/events/rxrpc.h
835
__entry->ref = ref;
include/trace/events/rxrpc.h
843
__entry->ref,
include/uapi/linux/tipc.h
49
__u32 ref;
include/uapi/linux/v4l2-controls.h
3038
#define V4L2_AV1_GLOBAL_MOTION_IS_INVALID(ref) (1 << (ref))
include/uapi/xen/gntdev.h
159
grant_ref_t ref;
include/uapi/xen/gntdev.h
44
__u32 ref;
include/xen/grant_table.h
115
void gnttab_end_foreign_access(grant_ref_t ref, struct page *page);
include/xen/grant_table.h
123
int gnttab_try_end_foreign_access(grant_ref_t ref);
include/xen/grant_table.h
132
void gnttab_free_grant_reference(grant_ref_t ref);
include/xen/grant_table.h
149
void gnttab_grant_foreign_access_ref(grant_ref_t ref, domid_t domid,
include/xen/grant_table.h
154
grant_ref_t ref, domid_t domid,
include/xen/grant_table.h
157
gnttab_grant_foreign_access_ref(ref, domid, xen_page_to_gfn(page),
include/xen/grant_table.h
163
uint32_t flags, grant_ref_t ref, domid_t domid)
include/xen/grant_table.h
173
map->ref = ref;
include/xen/grant_table.h
98
int gnttab_end_foreign_access_ref(grant_ref_t ref);
include/xen/interface/grant_table.h
285
grant_ref_t ref;
include/xen/interface/grant_table.h
359
grant_ref_t ref;
include/xen/interface/grant_table.h
392
grant_ref_t ref;
include/xen/interface/grant_table.h
509
grant_ref_t ref;
include/xen/interface/io/pvcalls.h
27
grant_ref_t ref[];
include/xen/interface/io/pvcalls.h
54
grant_ref_t ref;
include/xen/interface/io/pvcalls.h
73
grant_ref_t ref;
include/xen/interface/io/ring.h
443
grant_ref_t ref[]; \
io_uring/io-wq.c
167
return refcount_inc_not_zero(&worker->ref);
io_uring/io-wq.c
172
if (refcount_dec_and_test(&worker->ref))
io_uring/io-wq.c
49
refcount_t ref;
io_uring/io-wq.c
911
refcount_set(&worker->ref, 1);
io_uring/io_uring.c
184
static __cold void io_ring_ctx_ref_free(struct percpu_ref *ref)
io_uring/io_uring.c
186
struct io_ring_ctx *ctx = container_of(ref, struct io_ring_ctx, refs);
kernel/bpf/bpf_lru_list.c
36
return READ_ONCE(node->ref);
kernel/bpf/bpf_lru_list.c
41
WRITE_ONCE(node->ref, 0);
kernel/bpf/bpf_lru_list.h
28
u8 ref;
kernel/bpf/bpf_lru_list.h
68
if (!READ_ONCE(node->ref))
kernel/bpf/bpf_lru_list.h
69
WRITE_ONCE(node->ref, 1);
kernel/bpf/cgroup.c
361
static void cgroup_bpf_release_fn(struct percpu_ref *ref)
kernel/bpf/cgroup.c
363
struct cgroup *cgrp = container_of(ref, struct cgroup, bpf.refcnt);
kernel/bpf/hashtab.c
825
offsetof(struct bpf_lru_node, ref));
kernel/bpf/hashtab.c
829
offsetof(struct bpf_lru_node, ref),
kernel/bpf/helpers.c
2367
struct bpf_refcount *ref;
kernel/bpf/helpers.c
2372
ref = (struct bpf_refcount *)(p__refcounted_kptr + meta->record->refcount_off);
kernel/bpf/helpers.c
2373
if (!refcount_inc_not_zero((refcount_t *)ref))
kernel/cgroup/cgroup.c
283
static void css_release(struct percpu_ref *ref);
kernel/cgroup/cgroup.c
5697
static void css_release(struct percpu_ref *ref)
kernel/cgroup/cgroup.c
5700
container_of(ref, struct cgroup_subsys_state, refcnt);
kernel/cgroup/cgroup.c
6070
static void css_killed_ref_fn(struct percpu_ref *ref)
kernel/cgroup/cgroup.c
6073
container_of(ref, struct cgroup_subsys_state, refcnt);
kernel/cgroup/dmem.c
115
refcount_inc(&pool->ref);
kernel/cgroup/dmem.c
120
return refcount_inc_not_zero(&pool->ref);
kernel/cgroup/dmem.c
125
if (!refcount_dec_and_test(&pool->ref))
kernel/cgroup/dmem.c
137
kref_put(&pool->region->ref, dmemcg_free_region);
kernel/cgroup/dmem.c
26
struct kref ref;
kernel/cgroup/dmem.c
379
refcount_set(&pool->ref, 1);
kernel/cgroup/dmem.c
380
kref_get(&region->ref);
kernel/cgroup/dmem.c
455
static void dmemcg_free_region(struct kref *ref)
kernel/cgroup/dmem.c
457
struct dmem_cgroup_region *cgregion = container_of(ref, typeof(*cgregion), ref);
kernel/cgroup/dmem.c
494
kref_put(&region->ref, dmemcg_free_region);
kernel/cgroup/dmem.c
533
kref_init(&ret->ref);
kernel/cgroup/dmem.c
549
kref_get_unless_zero(&region->ref))
kernel/cgroup/dmem.c
77
refcount_t ref;
kernel/cgroup/dmem.c
780
kref_put(&region->ref, dmemcg_free_region);
kernel/cgroup/dmem.c
94
static void dmemcg_free_region(struct kref *ref);
kernel/dma/debug.c
1000
ref->dev_addr, ref->size);
kernel/dma/debug.c
1005
if (ref->size != entry->size) {
kernel/dma/debug.c
1006
err_printk(ref->dev, entry, "device driver frees "
kernel/dma/debug.c
1010
ref->dev_addr, entry->size, ref->size);
kernel/dma/debug.c
1013
if (ref->type != entry->type) {
kernel/dma/debug.c
1014
err_printk(ref->dev, entry, "device driver frees "
kernel/dma/debug.c
1018
ref->dev_addr, ref->size,
kernel/dma/debug.c
1019
type2name[entry->type], type2name[ref->type]);
kernel/dma/debug.c
1022
ref->paddr != entry->paddr) {
kernel/dma/debug.c
1023
err_printk(ref->dev, entry, "device driver frees "
kernel/dma/debug.c
1028
ref->dev_addr, ref->size,
kernel/dma/debug.c
1030
&ref->paddr);
kernel/dma/debug.c
1033
if (ref->sg_call_ents && ref->type == dma_debug_sg &&
kernel/dma/debug.c
1034
ref->sg_call_ents != entry->sg_call_ents) {
kernel/dma/debug.c
1035
err_printk(ref->dev, entry, "device driver frees "
kernel/dma/debug.c
1038
entry->sg_call_ents, ref->sg_call_ents);
kernel/dma/debug.c
1045
if (ref->direction != entry->direction) {
kernel/dma/debug.c
1046
err_printk(ref->dev, entry, "device driver frees "
kernel/dma/debug.c
1050
ref->dev_addr, ref->size,
kernel/dma/debug.c
1052
dir2name[ref->direction]);
kernel/dma/debug.c
1061
err_printk(ref->dev, entry,
kernel/dma/debug.c
1065
ref->dev_addr, ref->size,
kernel/dma/debug.c
1116
struct dma_debug_entry *ref,
kernel/dma/debug.c
1123
bucket = get_hash_bucket(ref, &flags);
kernel/dma/debug.c
1125
entry = bucket_find_contain(&bucket, ref, &flags);
kernel/dma/debug.c
1131
(unsigned long long)ref->dev_addr, ref->size);
kernel/dma/debug.c
1135
if (ref->size > entry->size) {
kernel/dma/debug.c
1142
ref->size);
kernel/dma/debug.c
1148
if (ref->direction != entry->direction) {
kernel/dma/debug.c
1153
(unsigned long long)ref->dev_addr, entry->size,
kernel/dma/debug.c
1155
dir2name[ref->direction]);
kernel/dma/debug.c
1159
!(ref->direction == DMA_TO_DEVICE))
kernel/dma/debug.c
1164
(unsigned long long)ref->dev_addr, entry->size,
kernel/dma/debug.c
1166
dir2name[ref->direction]);
kernel/dma/debug.c
1169
!(ref->direction == DMA_FROM_DEVICE))
kernel/dma/debug.c
1174
(unsigned long long)ref->dev_addr, entry->size,
kernel/dma/debug.c
1176
dir2name[ref->direction]);
kernel/dma/debug.c
1178
if (ref->sg_call_ents && ref->type == dma_debug_sg &&
kernel/dma/debug.c
1179
ref->sg_call_ents != entry->sg_call_ents) {
kernel/dma/debug.c
1180
err_printk(ref->dev, entry, "device driver syncs "
kernel/dma/debug.c
1183
entry->sg_call_ents, ref->sg_call_ents);
kernel/dma/debug.c
1265
struct dma_debug_entry ref;
kernel/dma/debug.c
1273
ref.dev = dev;
kernel/dma/debug.c
1274
ref.dev_addr = dma_addr;
kernel/dma/debug.c
1275
bucket = get_hash_bucket(&ref, &flags);
kernel/dma/debug.c
1278
if (!exact_match(&ref, entry))
kernel/dma/debug.c
1304
struct dma_debug_entry ref = {
kernel/dma/debug.c
1314
check_unmap(&ref);
kernel/dma/debug.c
1355
struct dma_debug_entry *ref)
kernel/dma/debug.c
1362
bucket = get_hash_bucket(ref, &flags);
kernel/dma/debug.c
1363
entry = bucket_find_exact(bucket, ref);
kernel/dma/debug.c
1384
struct dma_debug_entry ref = {
kernel/dma/debug.c
1398
mapped_ents = get_nr_mapped_entries(dev, &ref);
kernel/dma/debug.c
1400
check_unmap(&ref);
kernel/dma/debug.c
1449
struct dma_debug_entry ref = {
kernel/dma/debug.c
1461
ref.paddr = virt_to_paddr(virt);
kernel/dma/debug.c
1466
check_unmap(&ref);
kernel/dma/debug.c
1472
struct dma_debug_entry ref;
kernel/dma/debug.c
1477
ref.type = dma_debug_single;
kernel/dma/debug.c
1478
ref.dev = dev;
kernel/dma/debug.c
1479
ref.dev_addr = dma_handle;
kernel/dma/debug.c
1480
ref.size = size;
kernel/dma/debug.c
1481
ref.direction = direction;
kernel/dma/debug.c
1482
ref.sg_call_ents = 0;
kernel/dma/debug.c
1484
check_sync(dev, &ref, true);
kernel/dma/debug.c
1491
struct dma_debug_entry ref;
kernel/dma/debug.c
1496
ref.type = dma_debug_single;
kernel/dma/debug.c
1497
ref.dev = dev;
kernel/dma/debug.c
1498
ref.dev_addr = dma_handle;
kernel/dma/debug.c
1499
ref.size = size;
kernel/dma/debug.c
1500
ref.direction = direction;
kernel/dma/debug.c
1501
ref.sg_call_ents = 0;
kernel/dma/debug.c
1503
check_sync(dev, &ref, false);
kernel/dma/debug.c
1517
struct dma_debug_entry ref = {
kernel/dma/debug.c
1528
mapped_ents = get_nr_mapped_entries(dev, &ref);
kernel/dma/debug.c
1533
check_sync(dev, &ref, true);
kernel/dma/debug.c
1548
struct dma_debug_entry ref = {
kernel/dma/debug.c
1558
mapped_ents = get_nr_mapped_entries(dev, &ref);
kernel/dma/debug.c
1563
check_sync(dev, &ref, false);
kernel/dma/debug.c
1595
struct dma_debug_entry ref = {
kernel/dma/debug.c
1607
check_unmap(&ref);
kernel/dma/debug.c
298
struct dma_debug_entry *ref,
kernel/dma/debug.c
305
if (!match(ref, entry))
kernel/dma/debug.c
320
entry->size == ref->size ? ++match_lvl : 0;
kernel/dma/debug.c
321
entry->type == ref->type ? ++match_lvl : 0;
kernel/dma/debug.c
322
entry->direction == ref->direction ? ++match_lvl : 0;
kernel/dma/debug.c
323
entry->sg_call_ents == ref->sg_call_ents ? ++match_lvl : 0;
kernel/dma/debug.c
348
struct dma_debug_entry *ref)
kernel/dma/debug.c
350
return __hash_bucket_find(bucket, ref, exact_match);
kernel/dma/debug.c
354
struct dma_debug_entry *ref,
kernel/dma/debug.c
358
struct dma_debug_entry *entry, index = *ref;
kernel/dma/debug.c
362
entry = __hash_bucket_find(*bucket, ref, containing_match);
kernel/dma/debug.c
978
static void check_unmap(struct dma_debug_entry *ref)
kernel/dma/debug.c
984
bucket = get_hash_bucket(ref, &flags);
kernel/dma/debug.c
985
entry = bucket_find_exact(bucket, ref);
kernel/dma/debug.c
991
if (dma_mapping_error(ref->dev, ref->dev_addr)) {
kernel/dma/debug.c
992
err_printk(ref->dev, NULL,
kernel/dma/debug.c
996
err_printk(ref->dev, NULL,
kernel/events/uprobes.c
1008
refcount_set(&uprobe->ref, 1);
kernel/events/uprobes.c
637
refcount_inc(&uprobe->ref);
kernel/events/uprobes.c
64
refcount_t ref;
kernel/events/uprobes.c
649
if (refcount_inc_not_zero(&uprobe->ref))
kernel/events/uprobes.c
702
if (!refcount_dec_and_test(&uprobe->ref))
kernel/irq/irqdesc.c
1036
this_cpu_write(desc->kstat_irqs->ref, this_cpu_read(desc->kstat_irqs->cnt));
kernel/irq/irqdesc.c
1046
return this_cpu_read(desc->kstat_irqs->cnt) - this_cpu_read(desc->kstat_irqs->ref);
kernel/sched/sched.h
2168
atomic_t ref;
kernel/sched/sched.h
2186
atomic_t ref;
kernel/sched/topology.c
1209
already_visited = atomic_inc_return(&sg->ref) > 1;
kernel/sched/topology.c
1211
WARN_ON(already_visited != (atomic_inc_return(&sg->sgc->ref) > 1));
kernel/sched/topology.c
1585
if (atomic_read(&(*per_cpu_ptr(sdd->sds, cpu))->ref))
kernel/sched/topology.c
1588
if (atomic_read(&(*per_cpu_ptr(sdd->sg, cpu))->ref))
kernel/sched/topology.c
1591
if (atomic_read(&(*per_cpu_ptr(sdd->sgc, cpu))->ref))
kernel/sched/topology.c
1736
atomic_inc(&sd->shared->ref);
kernel/sched/topology.c
615
if (free_sgc && atomic_dec_and_test(&sg->sgc->ref))
kernel/sched/topology.c
618
if (atomic_dec_and_test(&sg->ref))
kernel/sched/topology.c
633
if (sd->shared && atomic_dec_and_test(&sd->shared->ref))
kernel/sched/topology.c
972
atomic_inc(&sg->ref);
kernel/sched/topology.c
988
if (atomic_inc_return(&sg->sgc->ref) == 1)
kernel/trace/ftrace.c
1076
int ref;
kernel/trace/ftrace.c
5418
WARN_ON(probe->ref <= 0);
kernel/trace/ftrace.c
5421
probe->ref--;
kernel/trace/ftrace.c
5423
if (!probe->ref) {
kernel/trace/ftrace.c
5442
probe->ref++;
kernel/trace/ftrace.c
5557
probe->ref += count;
kernel/trace/ftrace.c
5684
WARN_ON(probe->ref < count);
kernel/trace/ftrace.c
5686
probe->ref -= count;
kernel/trace/trace.c
10768
tr->ref++;
kernel/trace/trace.c
590
tr->ref++;
kernel/trace/trace.c
600
WARN_ON(!this_tr->ref);
kernel/trace/trace.c
601
this_tr->ref--;
kernel/trace/trace.c
6591
tinfo->ref = 1;
kernel/trace/trace.c
6647
if (WARN_ON_ONCE(!*tinfo || !(*tinfo)->ref))
kernel/trace/trace.c
6650
if (--(*tinfo)->ref)
kernel/trace/trace.c
6701
tinfo->ref++;
kernel/trace/trace.c
6702
return tinfo->ref;
kernel/trace/trace.c
6717
if (WARN_ON_ONCE(!tinfo || !tinfo->ref))
kernel/trace/trace.c
6720
--tinfo->ref;
kernel/trace/trace.c
6721
return tinfo->ref;
kernel/trace/trace.c
6992
trace_user_buffer->ref++;
kernel/trace/trace.c
7990
static void buffer_ref_release(struct buffer_ref *ref)
kernel/trace/trace.c
7992
if (!refcount_dec_and_test(&ref->refcount))
kernel/trace/trace.c
7994
ring_buffer_free_read_page(ref->buffer, ref->cpu, ref->page);
kernel/trace/trace.c
7995
kfree(ref);
kernel/trace/trace.c
8001
struct buffer_ref *ref = (struct buffer_ref *)buf->private;
kernel/trace/trace.c
8003
buffer_ref_release(ref);
kernel/trace/trace.c
8010
struct buffer_ref *ref = (struct buffer_ref *)buf->private;
kernel/trace/trace.c
8012
if (refcount_read(&ref->refcount) > INT_MAX/2)
kernel/trace/trace.c
8015
refcount_inc(&ref->refcount);
kernel/trace/trace.c
8031
struct buffer_ref *ref =
kernel/trace/trace.c
8034
buffer_ref_release(ref);
kernel/trace/trace.c
8054
struct buffer_ref *ref;
kernel/trace/trace.c
8084
ref = kzalloc_obj(*ref);
kernel/trace/trace.c
8085
if (!ref) {
kernel/trace/trace.c
8090
refcount_set(&ref->refcount, 1);
kernel/trace/trace.c
8091
ref->buffer = iter->array_buffer->buffer;
kernel/trace/trace.c
8092
ref->page = ring_buffer_alloc_read_page(ref->buffer, iter->cpu_file);
kernel/trace/trace.c
8093
if (IS_ERR(ref->page)) {
kernel/trace/trace.c
8094
ret = PTR_ERR(ref->page);
kernel/trace/trace.c
8095
ref->page = NULL;
kernel/trace/trace.c
8096
kfree(ref);
kernel/trace/trace.c
8099
ref->cpu = iter->cpu_file;
kernel/trace/trace.c
8101
r = ring_buffer_read_page(ref->buffer, ref->page,
kernel/trace/trace.c
8104
ring_buffer_free_read_page(ref->buffer, ref->cpu,
kernel/trace/trace.c
8105
ref->page);
kernel/trace/trace.c
8106
kfree(ref);
kernel/trace/trace.c
8110
page = virt_to_page(ring_buffer_read_page_data(ref->page));
kernel/trace/trace.c
8115
spd.partial[i].private = (unsigned long)ref;
kernel/trace/trace.c
9531
tr->ref++;
kernel/trace/trace.c
9641
tr->ref++;
kernel/trace/trace.c
9738
tr->ref++;
kernel/trace/trace.c
9748
tr->ref++;
kernel/trace/trace.c
9759
if (tr->ref > 1 || (tr->current_trace && tr->trace_ref))
kernel/trace/trace.h
1579
int ref;
kernel/trace/trace.h
1859
int ref;
kernel/trace/trace.h
423
int ref;
kernel/trace/trace_events.c
1289
refcount_inc(&file->ref);
kernel/trace/trace_events.c
1294
if (WARN_ON_ONCE(!refcount_read(&file->ref))) {
kernel/trace/trace_events.c
1300
if (refcount_dec_and_test(&file->ref)) {
kernel/trace/trace_events.c
3669
refcount_set(&file->ref, 1);
kernel/trace/trace_events.c
4124
int ref;
kernel/trace/trace_events.c
4230
edata->ref++;
kernel/trace/trace_events.c
4239
edata->ref--;
kernel/trace/trace_events.c
4240
if (!edata->ref) {
kernel/trace/trace_events.c
4268
if (WARN_ON_ONCE(edata->ref <= 0))
kernel/trace/trace_events_hist.c
172
unsigned int ref;
kernel/trace/trace_events_hist.c
1915
hist_field->ref++;
kernel/trace/trace_events_hist.c
1920
if (--hist_field->ref > 1)
kernel/trace/trace_events_hist.c
1969
hist_field->ref = 1;
kernel/trace/trace_events_hist.c
3247
var->ref = 1;
kernel/trace/trace_events_hist.c
3593
data->synth_event->ref--;
kernel/trace/trace_events_hist.c
4041
event->ref++;
kernel/trace/trace_events_hist.c
4121
event->ref--;
kernel/trace/trace_events_hist.c
6305
if (!data->ref && hist_data->attrs->name)
kernel/trace/trace_events_hist.c
6308
data->ref++;
kernel/trace/trace_events_hist.c
6333
if (WARN_ON_ONCE(data->ref <= 0))
kernel/trace/trace_events_hist.c
6336
data->ref--;
kernel/trace/trace_events_hist.c
6337
if (!data->ref) {
kernel/trace/trace_events_hist.c
6356
data->ref++;
kernel/trace/trace_events_hist.c
6371
if (WARN_ON_ONCE(data->ref <= 0))
kernel/trace/trace_events_hist.c
6376
data->ref--;
kernel/trace/trace_events_hist.c
6377
if (!data->ref) {
kernel/trace/trace_events_hist.c
6778
se->ref--;
kernel/trace/trace_events_hist.c
6897
se->ref--;
kernel/trace/trace_events_hist.c
6933
se->ref++;
kernel/trace/trace_events_synth.c
109
return event->ref != 0;
kernel/trace/trace_events_synth.c
1490
if (se->ref)
kernel/trace/trace_events_synth.c
2222
if (event->ref)
kernel/trace/trace_events_trigger.c
1724
if (WARN_ON_ONCE(data->ref <= 0))
kernel/trace/trace_events_trigger.c
1727
data->ref--;
kernel/trace/trace_events_trigger.c
1728
if (!data->ref) {
kernel/trace/trace_events_trigger.c
572
data->ref++;
kernel/trace/trace_events_trigger.c
588
if (WARN_ON_ONCE(data->ref <= 0))
kernel/trace/trace_events_trigger.c
591
data->ref--;
kernel/trace/trace_events_trigger.c
592
if (!data->ref)
kernel/trace/trace_synth.h
26
int ref;
lib/alloc_tag.c
182
union codetag_ref ref;
lib/alloc_tag.c
184
if (get_page_tag_ref(folio_page(folio, i), &ref, &handle)) {
lib/alloc_tag.c
186
alloc_tag_ref_set(&ref, tag);
lib/alloc_tag.c
187
update_page_tag_ref(handle, &ref);
lib/cpu_rmap.c
278
static void irq_cpu_rmap_release(struct kref *ref)
lib/cpu_rmap.c
281
container_of(ref, struct irq_glue, notify.kref);
lib/cpu_rmap.c
66
static void cpu_rmap_release(struct kref *ref)
lib/cpu_rmap.c
68
struct cpu_rmap *rmap = container_of(ref, struct cpu_rmap, refcount);
lib/lz4/lz4hc_compress.c
358
const BYTE *ref = NULL;
lib/lz4/lz4hc_compress.c
379
matchlimit, (&ref), maxNbAttempts);
lib/lz4/lz4hc_compress.c
387
ref0 = ref;
lib/lz4/lz4hc_compress.c
402
&anchor, ml, ref, limit, oend))
lib/lz4/lz4hc_compress.c
411
ref = ref0;
lib/lz4/lz4hc_compress.c
421
ref = ref2;
lib/lz4/lz4hc_compress.c
468
ml, ref, limit, oend))
lib/lz4/lz4hc_compress.c
498
ml, ref, limit, oend))
lib/lz4/lz4hc_compress.c
501
ref = ref3;
lib/lz4/lz4hc_compress.c
539
ref, limit, oend))
lib/lz4/lz4hc_compress.c
543
ref = ref2;
lib/objpool.c
155
refcount_set(&pool->ref, pool->nr_objs + 1);
lib/objpool.c
182
if (refcount_dec_and_test(&pool->ref)) {
lib/objpool.c
200
if (refcount_sub_and_test(count, &pool->ref))
lib/percpu-refcount.c
102
data->ref = ref;
lib/percpu-refcount.c
103
ref->data = data;
lib/percpu-refcount.c
108
static void __percpu_ref_exit(struct percpu_ref *ref)
lib/percpu-refcount.c
110
unsigned long __percpu *percpu_count = percpu_count_ptr(ref);
lib/percpu-refcount.c
114
WARN_ON_ONCE(ref->data && ref->data->confirm_switch);
lib/percpu-refcount.c
116
ref->percpu_count_ptr = __PERCPU_REF_ATOMIC_DEAD;
lib/percpu-refcount.c
130
void percpu_ref_exit(struct percpu_ref *ref)
lib/percpu-refcount.c
132
struct percpu_ref_data *data = ref->data;
lib/percpu-refcount.c
135
__percpu_ref_exit(ref);
lib/percpu-refcount.c
141
ref->percpu_count_ptr |= atomic_long_read(&ref->data->count) <<
lib/percpu-refcount.c
143
ref->data = NULL;
lib/percpu-refcount.c
154
struct percpu_ref *ref = data->ref;
lib/percpu-refcount.c
156
data->confirm_switch(ref);
lib/percpu-refcount.c
161
__percpu_ref_exit(ref);
lib/percpu-refcount.c
164
percpu_ref_put(ref);
lib/percpu-refcount.c
171
struct percpu_ref *ref = data->ref;
lib/percpu-refcount.c
172
unsigned long __percpu *percpu_count = percpu_count_ptr(ref);
lib/percpu-refcount.c
209
static void percpu_ref_noop_confirm_switch(struct percpu_ref *ref)
lib/percpu-refcount.c
213
static void __percpu_ref_switch_to_atomic(struct percpu_ref *ref,
lib/percpu-refcount.c
216
if (ref->percpu_count_ptr & __PERCPU_REF_ATOMIC) {
lib/percpu-refcount.c
218
confirm_switch(ref);
lib/percpu-refcount.c
223
ref->percpu_count_ptr |= __PERCPU_REF_ATOMIC;
lib/percpu-refcount.c
229
ref->data->confirm_switch = confirm_switch ?:
lib/percpu-refcount.c
232
percpu_ref_get(ref); /* put after confirmation */
lib/percpu-refcount.c
233
call_rcu_hurry(&ref->data->rcu,
lib/percpu-refcount.c
237
static void __percpu_ref_switch_to_percpu(struct percpu_ref *ref)
lib/percpu-refcount.c
239
unsigned long __percpu *percpu_count = percpu_count_ptr(ref);
lib/percpu-refcount.c
244
if (!(ref->percpu_count_ptr & __PERCPU_REF_ATOMIC))
lib/percpu-refcount.c
247
if (WARN_ON_ONCE(!ref->data->allow_reinit))
lib/percpu-refcount.c
250
atomic_long_add(PERCPU_COUNT_BIAS, &ref->data->count);
lib/percpu-refcount.c
261
smp_store_release(&ref->percpu_count_ptr,
lib/percpu-refcount.c
262
ref->percpu_count_ptr & ~__PERCPU_REF_ATOMIC);
lib/percpu-refcount.c
265
static void __percpu_ref_switch_mode(struct percpu_ref *ref,
lib/percpu-refcount.c
268
struct percpu_ref_data *data = ref->data;
lib/percpu-refcount.c
280
if (data->force_atomic || percpu_ref_is_dying(ref))
lib/percpu-refcount.c
281
__percpu_ref_switch_to_atomic(ref, confirm_switch);
lib/percpu-refcount.c
283
__percpu_ref_switch_to_percpu(ref);
lib/percpu-refcount.c
306
void percpu_ref_switch_to_atomic(struct percpu_ref *ref,
lib/percpu-refcount.c
313
ref->data->force_atomic = true;
lib/percpu-refcount.c
314
__percpu_ref_switch_mode(ref, confirm_switch);
lib/percpu-refcount.c
328
void percpu_ref_switch_to_atomic_sync(struct percpu_ref *ref)
lib/percpu-refcount.c
330
percpu_ref_switch_to_atomic(ref, NULL);
lib/percpu-refcount.c
331
wait_event(percpu_ref_switch_waitq, !ref->data->confirm_switch);
lib/percpu-refcount.c
353
void percpu_ref_switch_to_percpu(struct percpu_ref *ref)
lib/percpu-refcount.c
359
ref->data->force_atomic = false;
lib/percpu-refcount.c
360
__percpu_ref_switch_mode(ref, NULL);
lib/percpu-refcount.c
383
void percpu_ref_kill_and_confirm(struct percpu_ref *ref,
lib/percpu-refcount.c
390
WARN_ONCE(percpu_ref_is_dying(ref),
lib/percpu-refcount.c
392
ref->data->release);
lib/percpu-refcount.c
394
ref->percpu_count_ptr |= __PERCPU_REF_DEAD;
lib/percpu-refcount.c
395
__percpu_ref_switch_mode(ref, confirm_kill);
lib/percpu-refcount.c
396
percpu_ref_put(ref);
lib/percpu-refcount.c
410
bool percpu_ref_is_zero(struct percpu_ref *ref)
lib/percpu-refcount.c
415
if (__ref_is_percpu(ref, &percpu_count))
lib/percpu-refcount.c
42
static unsigned long __percpu *percpu_count_ptr(struct percpu_ref *ref)
lib/percpu-refcount.c
420
if (ref->data)
lib/percpu-refcount.c
421
count = atomic_long_read(&ref->data->count);
lib/percpu-refcount.c
423
count = ref->percpu_count_ptr >> __PERCPU_REF_FLAG_BITS;
lib/percpu-refcount.c
441
void percpu_ref_reinit(struct percpu_ref *ref)
lib/percpu-refcount.c
443
WARN_ON_ONCE(!percpu_ref_is_zero(ref));
lib/percpu-refcount.c
445
percpu_ref_resurrect(ref);
lib/percpu-refcount.c
45
(ref->percpu_count_ptr & ~__PERCPU_REF_ATOMIC_DEAD);
lib/percpu-refcount.c
463
void percpu_ref_resurrect(struct percpu_ref *ref)
lib/percpu-refcount.c
470
WARN_ON_ONCE(!percpu_ref_is_dying(ref));
lib/percpu-refcount.c
471
WARN_ON_ONCE(__ref_is_percpu(ref, &percpu_count));
lib/percpu-refcount.c
473
ref->percpu_count_ptr &= ~__PERCPU_REF_DEAD;
lib/percpu-refcount.c
474
percpu_ref_get(ref);
lib/percpu-refcount.c
475
__percpu_ref_switch_mode(ref, NULL);
lib/percpu-refcount.c
63
int percpu_ref_init(struct percpu_ref *ref, percpu_ref_func_t *release,
lib/percpu-refcount.c
71
ref->percpu_count_ptr = (unsigned long)
lib/percpu-refcount.c
73
if (!ref->percpu_count_ptr)
lib/percpu-refcount.c
76
data = kzalloc_obj(*ref->data, gfp);
lib/percpu-refcount.c
78
free_percpu((void __percpu *)ref->percpu_count_ptr);
lib/percpu-refcount.c
79
ref->percpu_count_ptr = 0;
lib/percpu-refcount.c
87
ref->percpu_count_ptr |= __PERCPU_REF_ATOMIC;
lib/percpu-refcount.c
94
ref->percpu_count_ptr |= __PERCPU_REF_DEAD;
lib/rcuref.c
192
bool rcuref_get_slowpath(rcuref_t *ref)
lib/rcuref.c
194
unsigned int cnt = atomic_read(&ref->refcnt);
lib/rcuref.c
202
atomic_set(&ref->refcnt, RCUREF_DEAD);
lib/rcuref.c
215
atomic_set(&ref->refcnt, RCUREF_SATURATED);
lib/rcuref.c
237
bool rcuref_put_slowpath(rcuref_t *ref, unsigned int cnt)
lib/rcuref.c
250
if (!atomic_try_cmpxchg_release(&ref->refcnt, &cnt, RCUREF_DEAD))
lib/rcuref.c
267
atomic_set(&ref->refcnt, RCUREF_DEAD);
lib/rcuref.c
277
atomic_set(&ref->refcnt, RCUREF_SATURATED);
lib/zstd/compress/zstd_preSplit.c
111
static int compareFingerprints(const Fingerprint* ref,
lib/zstd/compress/zstd_preSplit.c
116
assert(ref->nbEvents > 0);
lib/zstd/compress/zstd_preSplit.c
118
{ U64 p50 = (U64)ref->nbEvents * (U64)newfp->nbEvents;
lib/zstd/compress/zstd_preSplit.c
119
U64 deviation = fpDistance(ref, newfp, hashLog);
mm/backing-dev.c
1183
static void release_bdi(struct kref *ref)
mm/backing-dev.c
1186
container_of(ref, struct backing_dev_info, refcnt);
mm/hugetlb.c
1138
void resv_map_release(struct kref *ref)
mm/hugetlb.c
1140
struct resv_map *resv_map = container_of(ref, struct resv_map, refs);
mm/memcontrol.c
140
static void obj_cgroup_release(struct percpu_ref *ref)
mm/memcontrol.c
142
struct obj_cgroup *objcg = container_of(ref, struct obj_cgroup, refcnt);
mm/memcontrol.c
186
percpu_ref_exit(ref);
mm/memcontrol.c
3618
refcount_add(n, &memcg->id.ref);
mm/memcontrol.c
3623
if (refcount_sub_and_test(n, &memcg->id.ref)) {
mm/memcontrol.c
3638
while (!refcount_inc_not_zero(&memcg->id.ref)) {
mm/memcontrol.c
3896
refcount_set(&memcg->id.ref, 1);
mm/memory.c
3050
void pfnmap_track_ctx_release(struct kref *ref)
mm/memory.c
3052
struct pfnmap_track_ctx *ctx = container_of(ref, struct pfnmap_track_ctx, kref);
mm/memremap.c
116
percpu_ref_kill(&pgmap->ref);
mm/memremap.c
120
percpu_ref_put_many(&pgmap->ref, pfn_len(pgmap, i));
mm/memremap.c
126
percpu_ref_exit(&pgmap->ref);
mm/memremap.c
137
static void dev_pagemap_percpu_release(struct percpu_ref *ref)
mm/memremap.c
139
struct dev_pagemap *pgmap = container_of(ref, struct dev_pagemap, ref);
mm/memremap.c
247
percpu_ref_get_many(&pgmap->ref, pfn_len(pgmap, range_id));
mm/memremap.c
325
error = percpu_ref_init(&pgmap->ref, dev_pagemap_percpu_release, 0,
mm/memremap.c
408
if (pgmap && !percpu_ref_tryget_live_rcu(&pgmap->ref))
mm/memremap.c
457
percpu_ref_put_many(&folio->pgmap->ref, nr);
mm/memremap.c
521
WARN_ON_ONCE(!percpu_ref_tryget_many(&page_pgmap(page)->ref, 1 << order));
mm/page_alloc.c
1275
union codetag_ref ref;
mm/page_alloc.c
1277
if (get_page_tag_ref(page, &ref, &handle)) {
mm/page_alloc.c
1278
set_codetag_empty(&ref);
mm/page_alloc.c
1279
update_page_tag_ref(handle, &ref);
mm/page_alloc.c
1290
union codetag_ref ref;
mm/page_alloc.c
1292
if (get_page_tag_ref(page, &ref, &handle)) {
mm/page_alloc.c
1293
alloc_tag_add(&ref, task->alloc_tag, PAGE_SIZE * nr);
mm/page_alloc.c
1294
update_page_tag_ref(handle, &ref);
mm/page_alloc.c
1311
union codetag_ref ref;
mm/page_alloc.c
1313
if (get_page_tag_ref(page, &ref, &handle)) {
mm/page_alloc.c
1314
alloc_tag_sub(&ref, PAGE_SIZE * nr);
mm/page_alloc.c
1315
update_page_tag_ref(handle, &ref);
mm/slub.c
2058
if (unlikely(is_codetag_empty(&ext->ref))) {
mm/slub.c
2064
WARN_ON(ext->ref.ct);
mm/slub.c
2065
set_codetag_empty(&ext->ref);
mm/slub.c
2087
set_codetag_empty(&vec[i].ref);
mm/slub.c
2380
alloc_tag_add(&obj_ext->ref, current->alloc_tag, s->size);
mm/slub.c
2414
alloc_tag_sub(&slab_obj_ext(slab, obj_exts, off)->ref, s->size);
mm/swapfile.c
658
static void swap_users_ref_free(struct percpu_ref *ref)
mm/swapfile.c
662
si = container_of(ref, struct swap_info_struct, users);
mm/zswap.c
156
struct percpu_ref ref;
mm/zswap.c
243
static void __zswap_pool_empty(struct percpu_ref *ref);
mm/zswap.c
283
ret = percpu_ref_init(&pool->ref, __zswap_pool_empty,
mm/zswap.c
342
WARN_ON(!percpu_ref_is_zero(&pool->ref));
mm/zswap.c
343
percpu_ref_exit(&pool->ref);
mm/zswap.c
351
static void __zswap_pool_empty(struct percpu_ref *ref)
mm/zswap.c
355
pool = container_of(ref, typeof(*pool), ref);
mm/zswap.c
374
return percpu_ref_tryget(&pool->ref);
mm/zswap.c
380
percpu_ref_get(&pool->ref);
mm/zswap.c
385
percpu_ref_put(&pool->ref);
mm/zswap.c
532
percpu_ref_resurrect(&pool->ref);
mm/zswap.c
566
percpu_ref_kill(&put_pool->ref);
net/9p/trans_xen.c
294
grant_ref_t ref;
net/9p/trans_xen.c
296
ref = priv->rings[i].intf->ref[j];
net/9p/trans_xen.c
297
gnttab_end_foreign_access(ref, NULL);
net/9p/trans_xen.c
303
gnttab_end_foreign_access(priv->rings[i].ref, NULL);
net/9p/trans_xen.c
348
ring->ref = ret;
net/9p/trans_xen.c
360
ring->intf->ref[i] = ret;
net/9p/trans_xen.c
380
gnttab_end_foreign_access(ring->intf->ref[i], NULL);
net/9p/trans_xen.c
383
gnttab_end_foreign_access(ring->ref, NULL);
net/9p/trans_xen.c
40
grant_ref_t ref;
net/9p/trans_xen.c
458
priv->rings[i].ref);
net/batman-adv/bridge_loop_avoidance.c
149
static void batadv_backbone_gw_release(struct kref *ref)
net/batman-adv/bridge_loop_avoidance.c
153
backbone_gw = container_of(ref, struct batadv_bla_backbone_gw,
net/batman-adv/bridge_loop_avoidance.c
177
static void batadv_claim_release(struct kref *ref)
net/batman-adv/bridge_loop_avoidance.c
182
claim = container_of(ref, struct batadv_bla_claim, refcount);
net/batman-adv/distributed-arp-table.c
112
static void batadv_dat_entry_release(struct kref *ref)
net/batman-adv/distributed-arp-table.c
116
dat_entry = container_of(ref, struct batadv_dat_entry, refcount);
net/batman-adv/gateway_client.c
62
void batadv_gw_node_release(struct kref *ref)
net/batman-adv/gateway_client.c
66
gw_node = container_of(ref, struct batadv_gw_node, refcount);
net/batman-adv/gateway_client.h
31
void batadv_gw_node_release(struct kref *ref);
net/batman-adv/hard-interface.c
50
void batadv_hardif_release(struct kref *ref)
net/batman-adv/hard-interface.c
54
hard_iface = container_of(ref, struct batadv_hard_iface, refcount);
net/batman-adv/hard-interface.h
81
void batadv_hardif_release(struct kref *ref);
net/batman-adv/mesh-interface.c
496
void batadv_meshif_vlan_release(struct kref *ref)
net/batman-adv/mesh-interface.c
500
vlan = container_of(ref, struct batadv_meshif_vlan, refcount);
net/batman-adv/mesh-interface.h
24
void batadv_meshif_vlan_release(struct kref *ref);
net/batman-adv/originator.c
203
void batadv_orig_node_vlan_release(struct kref *ref)
net/batman-adv/originator.c
207
orig_vlan = container_of(ref, struct batadv_orig_node_vlan, refcount);
net/batman-adv/originator.c
247
void batadv_neigh_ifinfo_release(struct kref *ref)
net/batman-adv/originator.c
251
neigh_ifinfo = container_of(ref, struct batadv_neigh_ifinfo, refcount);
net/batman-adv/originator.c
264
void batadv_hardif_neigh_release(struct kref *ref)
net/batman-adv/originator.c
268
hardif_neigh = container_of(ref, struct batadv_hardif_neigh_node,
net/batman-adv/originator.c
284
void batadv_neigh_node_release(struct kref *ref)
net/batman-adv/originator.c
290
neigh_node = container_of(ref, struct batadv_neigh_node, refcount);
net/batman-adv/originator.c
811
void batadv_orig_ifinfo_release(struct kref *ref)
net/batman-adv/originator.c
816
orig_ifinfo = container_of(ref, struct batadv_orig_ifinfo, refcount);
net/batman-adv/originator.c
851
void batadv_orig_node_release(struct kref *ref)
net/batman-adv/originator.c
860
orig_node = container_of(ref, struct batadv_orig_node, refcount);
net/batman-adv/originator.h
24
void batadv_orig_node_release(struct kref *ref);
net/batman-adv/originator.h
30
void batadv_hardif_neigh_release(struct kref *ref);
net/batman-adv/originator.h
35
void batadv_neigh_node_release(struct kref *ref);
net/batman-adv/originator.h
48
void batadv_neigh_ifinfo_release(struct kref *ref);
net/batman-adv/originator.h
58
void batadv_orig_ifinfo_release(struct kref *ref);
net/batman-adv/originator.h
67
void batadv_orig_node_vlan_release(struct kref *ref);
net/batman-adv/tp_meter.c
334
static void batadv_tp_vars_release(struct kref *ref)
net/batman-adv/tp_meter.c
339
tp_vars = container_of(ref, struct batadv_tp_vars, refcount);
net/batman-adv/translation-table.c
217
static void batadv_tt_local_entry_release(struct kref *ref)
net/batman-adv/translation-table.c
221
tt_local_entry = container_of(ref, struct batadv_tt_local_entry,
net/batman-adv/translation-table.c
249
void batadv_tt_global_entry_release(struct kref *ref)
net/batman-adv/translation-table.c
2520
static void batadv_tt_req_node_release(struct kref *ref)
net/batman-adv/translation-table.c
2524
tt_req_node = container_of(ref, struct batadv_tt_req_node, refcount);
net/batman-adv/translation-table.c
253
tt_global_entry = container_of(ref, struct batadv_tt_global_entry,
net/batman-adv/translation-table.c
388
static void batadv_tt_orig_list_entry_release(struct kref *ref)
net/batman-adv/translation-table.c
392
orig_entry = container_of(ref, struct batadv_tt_orig_list_entry,
net/batman-adv/translation-table.h
32
void batadv_tt_global_entry_release(struct kref *ref);
net/batman-adv/tvlv.c
101
tvlv = container_of(ref, struct batadv_tvlv_container, refcount);
net/batman-adv/tvlv.c
38
static void batadv_tvlv_handler_release(struct kref *ref)
net/batman-adv/tvlv.c
42
tvlv_handler = container_of(ref, struct batadv_tvlv_handler, refcount);
net/batman-adv/tvlv.c
97
static void batadv_tvlv_container_release(struct kref *ref)
net/bluetooth/hidp/core.c
928
kref_init(&session->ref);
net/bluetooth/hidp/core.c
975
kref_get(&session->ref);
net/bluetooth/hidp/core.c
979
static void session_free(struct kref *ref)
net/bluetooth/hidp/core.c
981
struct hidp_session *session = container_of(ref, struct hidp_session,
net/bluetooth/hidp/core.c
982
ref);
net/bluetooth/hidp/core.c
997
kref_put(&session->ref, session_free);
net/bluetooth/hidp/hidp.h
139
struct kref ref;
net/bluetooth/iso.c
100
struct iso_conn *conn = container_of(ref, struct iso_conn, ref);
net/bluetooth/iso.c
125
BT_DBG("conn %p refcnt %d", conn, kref_read(&conn->ref));
net/bluetooth/iso.c
127
kref_put(&conn->ref, iso_conn_free);
net/bluetooth/iso.c
135
BT_DBG("conn %p refcnt %u", conn, kref_read(&conn->ref));
net/bluetooth/iso.c
137
if (!kref_get_unless_zero(&conn->ref))
net/bluetooth/iso.c
219
kref_init(&conn->ref);
net/bluetooth/iso.c
38
struct kref ref;
net/bluetooth/iso.c
98
static void iso_conn_free(struct kref *ref)
net/bluetooth/l2cap_core.c
1818
static void l2cap_conn_free(struct kref *ref)
net/bluetooth/l2cap_core.c
1820
struct l2cap_conn *conn = container_of(ref, struct l2cap_conn, ref);
net/bluetooth/l2cap_core.c
1828
kref_get(&conn->ref);
net/bluetooth/l2cap_core.c
1835
kref_put(&conn->ref, l2cap_conn_free);
net/bluetooth/l2cap_core.c
7002
kref_init(&conn->ref);
net/bluetooth/l2cap_core.c
7614
BT_DBG("conn %p orig refcnt %u", c, kref_read(&c->ref));
net/bluetooth/l2cap_core.c
7616
if (!kref_get_unless_zero(&c->ref))
net/bluetooth/sco.c
105
BT_DBG("conn %p refcnt %d", conn, kref_read(&conn->ref));
net/bluetooth/sco.c
107
kref_put(&conn->ref, sco_conn_free);
net/bluetooth/sco.c
112
BT_DBG("conn %p refcnt %u", conn, kref_read(&conn->ref));
net/bluetooth/sco.c
114
kref_get(&conn->ref);
net/bluetooth/sco.c
123
BT_DBG("conn %p refcnt %u", conn, kref_read(&conn->ref));
net/bluetooth/sco.c
125
if (!kref_get_unless_zero(&conn->ref))
net/bluetooth/sco.c
211
kref_init(&conn->ref);
net/bluetooth/sco.c
54
struct kref ref;
net/bluetooth/sco.c
80
static void sco_conn_free(struct kref *ref)
net/bluetooth/sco.c
82
struct sco_conn *conn = container_of(ref, struct sco_conn, ref);
net/ceph/string_table.c
88
void ceph_release_string(struct kref *ref)
net/ceph/string_table.c
90
struct ceph_string *cs = container_of(ref, struct ceph_string, kref);
net/core/dev.c
7322
static void netif_napi_affinity_release(struct kref *ref)
net/core/dev.c
7325
container_of(ref, struct napi_struct, notify.kref);
net/core/dev.c
7367
static void netif_napi_affinity_release(struct kref *ref)
net/core/devmem.c
151
percpu_ref_kill(&binding->ref);
net/core/devmem.c
217
err = percpu_ref_init(&binding->ref,
net/core/devmem.c
333
percpu_ref_exit(&binding->ref);
net/core/devmem.c
52
static void net_devmem_dmabuf_binding_release(struct percpu_ref *ref)
net/core/devmem.c
55
container_of(ref, struct net_devmem_dmabuf_binding, ref);
net/core/devmem.c
82
percpu_ref_exit(&binding->ref);
net/core/devmem.h
128
return percpu_ref_tryget(&binding->ref);
net/core/devmem.h
134
percpu_ref_put(&binding->ref);
net/core/devmem.h
44
struct percpu_ref ref;
net/dsa/dsa.c
261
static void dsa_tree_release(struct kref *ref)
net/dsa/dsa.c
265
dst = container_of(ref, struct dsa_switch_tree, refcount);
net/mac80211/rate.c
1003
struct rate_control_ref *ref;
net/mac80211/rate.c
1016
ref = rate_control_alloc(name, local);
net/mac80211/rate.c
1017
if (!ref) {
net/mac80211/rate.c
1024
local->rate_ctrl = ref;
net/mac80211/rate.c
1027
ref->ops->name);
net/mac80211/rate.c
1034
struct rate_control_ref *ref;
net/mac80211/rate.c
1036
ref = local->rate_ctrl;
net/mac80211/rate.c
1038
if (!ref)
net/mac80211/rate.c
1042
rate_control_free(local, ref);
net/mac80211/rate.c
107
if (ref->ops->tx_status_ext)
net/mac80211/rate.c
108
ref->ops->tx_status_ext(ref->priv, sband, priv_sta, st);
net/mac80211/rate.c
110
ref->ops->tx_status(ref->priv, sband, st->sta, priv_sta, st->skb);
net/mac80211/rate.c
122
struct rate_control_ref *ref = local->rate_ctrl;
net/mac80211/rate.c
128
if (ref && ref->ops->rate_update) {
net/mac80211/rate.c
138
ref->ops->rate_update(ref->priv, sband, &chanctx_conf->def,
net/mac80211/rate.c
248
struct rate_control_ref *ref = file->private_data;
net/mac80211/rate.c
249
int len = strlen(ref->ops->name);
net/mac80211/rate.c
252
ref->ops->name, len);
net/mac80211/rate.c
264
struct rate_control_ref *ref;
net/mac80211/rate.c
266
ref = kmalloc_obj(struct rate_control_ref);
net/mac80211/rate.c
267
if (!ref)
net/mac80211/rate.c
269
ref->ops = ieee80211_rate_control_ops_get(name);
net/mac80211/rate.c
270
if (!ref->ops)
net/mac80211/rate.c
273
ref->priv = ref->ops->alloc(&local->hw);
net/mac80211/rate.c
274
if (!ref->priv)
net/mac80211/rate.c
276
return ref;
net/mac80211/rate.c
279
kfree(ref);
net/mac80211/rate.c
35
struct rate_control_ref *ref = sta->rate_ctrl;
net/mac80211/rate.c
43
if (!ref)
net/mac80211/rate.c
68
ref->ops->rate_init(ref->priv, sband, &chanctx_conf->def, ista,
net/mac80211/rate.c
93
struct rate_control_ref *ref = local->rate_ctrl;
net/mac80211/rate.c
931
struct rate_control_ref *ref = sdata->local->rate_ctrl;
net/mac80211/rate.c
956
ref->ops->get_rate(ref->priv, ista, priv_sta, txrc);
net/mac80211/rate.c
98
if (!ref || !test_sta_flag(sta, WLAN_STA_RATE_CONTROL))
net/mac80211/rate.h
39
static inline void *rate_control_alloc_sta(struct rate_control_ref *ref,
net/mac80211/rate.h
43
return ref->ops->alloc_sta(ref->priv, &sta->sta, gfp);
net/mac80211/rate.h
48
struct rate_control_ref *ref = sta->rate_ctrl;
net/mac80211/rate.h
52
ref->ops->free_sta(ref->priv, ista, priv_sta);
net/mac80211/rate.h
58
struct rate_control_ref *ref = sta->rate_ctrl;
net/mac80211/rate.h
59
if (ref && sta->debugfs_dir && ref->ops->add_sta_debugfs)
net/mac80211/rate.h
60
ref->ops->add_sta_debugfs(ref->priv, sta->rate_ctrl_priv,
net/mac80211/sta_info.c
3259
struct rate_control_ref *ref = NULL;
net/mac80211/sta_info.c
3263
ref = local->rate_ctrl;
net/mac80211/sta_info.c
3266
if (ref && ref->ops->get_expected_throughput)
net/mac80211/sta_info.c
3267
thr = ref->ops->get_expected_throughput(sta->rate_ctrl_priv);
net/mac80211/tx.c
1185
struct rate_control_ref *ref = sdata->local->rate_ctrl;
net/mac80211/tx.c
1188
if (!ref || !(ref->ops->capa & RATE_CTRL_CAPA_AMPDU_TRIGGER))
net/mac802154/llsec.c
124
kref_init(&key->ref);
net/mac802154/llsec.c
162
static void llsec_key_release(struct kref *ref)
net/mac802154/llsec.c
167
key = container_of(ref, struct mac802154_llsec_key, ref);
net/mac802154/llsec.c
179
kref_get(&key->ref);
net/mac802154/llsec.c
185
kref_put(&key->ref, llsec_key_release);
net/mac802154/llsec.h
26
struct kref ref;
net/netfilter/ipset/ip_set_bitmap_gen.h
102
nla_put_net32(skb, IPSET_ATTR_REFERENCES, htonl(set->ref)) ||
net/netfilter/ipset/ip_set_core.c
1252
if (s && (s->ref || s->ref_netlink)) {
net/netfilter/ipset/ip_set_core.c
1273
} else if (s->ref || s->ref_netlink) {
net/netfilter/ipset/ip_set_core.c
1363
if (set->ref != 0 || set->ref_netlink != 0) {
net/netfilter/ipset/ip_set_core.c
1434
swap(from->ref, to->ref);
net/netfilter/ipset/ip_set_core.c
677
set->ref++;
net/netfilter/ipset/ip_set_core.c
685
BUG_ON(set->ref == 0);
net/netfilter/ipset/ip_set_core.c
686
set->ref--;
net/netfilter/ipset/ip_set_hash_gen.h
1024
if (atomic_dec_and_test(&t->uref) && atomic_read(&t->ref)) {
net/netfilter/ipset/ip_set_hash_gen.h
1085
if (atomic_read(&t->ref) && ext->target) {
net/netfilter/ipset/ip_set_hash_gen.h
1136
if (atomic_dec_and_test(&t->uref) && atomic_read(&t->ref)) {
net/netfilter/ipset/ip_set_hash_gen.h
1313
if (nla_put_net32(skb, IPSET_ATTR_REFERENCES, htonl(set->ref)) ||
net/netfilter/ipset/ip_set_hash_gen.h
1341
if (atomic_dec_and_test(&t->uref) && atomic_read(&t->ref)) {
net/netfilter/ipset/ip_set_hash_gen.h
585
if (atomic_dec_and_test(&t->uref) && atomic_read(&t->ref)) {
net/netfilter/ipset/ip_set_hash_gen.h
676
atomic_set(&orig->ref, 1);
net/netfilter/ipset/ip_set_hash_gen.h
792
atomic_set(&orig->ref, 0);
net/netfilter/ipset/ip_set_hash_gen.h
83
atomic_t ref; /* References for resizing */
net/netfilter/ipset/ip_set_hash_gen.h
997
if (atomic_read(&t->ref) && ext->target) {
net/netfilter/ipset/ip_set_list_set.c
464
nla_put_net32(skb, IPSET_ATTR_REFERENCES, htonl(set->ref)) ||
net/nfc/llcp.h
54
struct kref ref;
net/nfc/llcp_core.c
155
kref_get(&local->ref);
net/nfc/llcp_core.c
1639
kref_init(&local->ref);
net/nfc/llcp_core.c
175
static void local_release(struct kref *ref)
net/nfc/llcp_core.c
179
local = container_of(ref, struct nfc_llcp_local, ref);
net/nfc/llcp_core.c
195
ret = kref_put(&local->ref, local_release);
net/qrtr/af_qrtr.c
129
struct kref ref;
net/qrtr/af_qrtr.c
170
struct qrtr_node *node = container_of(kref, struct qrtr_node, ref);
net/qrtr/af_qrtr.c
203
kref_get(&node->ref);
net/qrtr/af_qrtr.c
212
kref_put_mutex(&node->ref, __qrtr_node_release, &qrtr_node_lock);
net/qrtr/af_qrtr.c
592
kref_init(&node->ref);
net/rxrpc/af_rxrpc.c
414
_enter("%d{%d}", call->debug_id, refcount_read(&call->ref));
net/rxrpc/ar-internal.h
1313
r = refcount_read(&local->ref);
net/rxrpc/ar-internal.h
1324
r = refcount_read(&local->ref);
net/rxrpc/ar-internal.h
332
refcount_t ref; /* Number of references to the structure */
net/rxrpc/ar-internal.h
382
refcount_t ref;
net/rxrpc/ar-internal.h
503
refcount_t ref;
net/rxrpc/ar-internal.h
532
refcount_t ref;
net/rxrpc/ar-internal.h
741
refcount_t ref;
net/rxrpc/ar-internal.h
915
refcount_t ref;
net/rxrpc/call_accept.c
104
trace_rxrpc_call(call->debug_id, refcount_read(&call->ref),
net/rxrpc/call_object.c
114
_leave(" = %p [%d]", call, refcount_read(&call->ref));
net/rxrpc/call_object.c
153
refcount_set(&call->ref, 1);
net/rxrpc/call_object.c
237
trace_rxrpc_call(call->debug_id, refcount_read(&call->ref),
net/rxrpc/call_object.c
414
trace_rxrpc_call(call->debug_id, refcount_read(&call->ref), 0,
net/rxrpc/call_object.c
427
trace_rxrpc_call(call->debug_id, refcount_read(&call->ref), ret,
net/rxrpc/call_object.c
503
int r = refcount_read(&call->ref);
net/rxrpc/call_object.c
514
if (!call || !__refcount_inc_not_zero(&call->ref, &r))
net/rxrpc/call_object.c
527
__refcount_inc(&call->ref, &r);
net/rxrpc/call_object.c
566
_enter("{%d,%d}", call->debug_id, refcount_read(&call->ref));
net/rxrpc/call_object.c
568
trace_rxrpc_call(call->debug_id, refcount_read(&call->ref),
net/rxrpc/call_object.c
652
dead = __refcount_dec_and_test(&call->ref, &r);
net/rxrpc/call_object.c
744
call, refcount_read(&call->ref),
net/rxrpc/conn_client.c
107
__refcount_inc(&bundle->ref, &r);
net/rxrpc/conn_client.c
114
trace_rxrpc_bundle(bundle->debug_id, refcount_read(&bundle->ref),
net/rxrpc/conn_client.c
132
dead = __refcount_dec_and_test(&bundle->ref, &r);
net/rxrpc/conn_client.c
172
refcount_set(&conn->ref, 1);
net/rxrpc/conn_client.c
62
conn, refcount_read(&conn->ref));
net/rxrpc/conn_client.c
90
refcount_set(&bundle->ref, 1);
net/rxrpc/conn_object.c
111
if (!conn || refcount_read(&conn->ref) == 0) {
net/rxrpc/conn_object.c
237
int r = refcount_read(&conn->ref);
net/rxrpc/conn_object.c
251
__refcount_inc(&conn->ref, &r);
net/rxrpc/conn_object.c
266
if (__refcount_inc_not_zero(&conn->ref, &r))
net/rxrpc/conn_object.c
293
_enter("{%d,u=%d}", conn->debug_id, refcount_read(&conn->ref));
net/rxrpc/conn_object.c
295
trace_rxrpc_conn(conn->debug_id, refcount_read(&conn->ref),
net/rxrpc/conn_object.c
367
dead = __refcount_dec_and_test(&conn->ref, &r);
net/rxrpc/conn_object.c
482
conn, refcount_read(&conn->ref));
net/rxrpc/conn_service.c
130
refcount_set(&conn->ref, 2);
net/rxrpc/conn_service.c
97
if (refcount_read(&cursor->ref) == 0)
net/rxrpc/local_object.c
117
refcount_set(&local->ref, 1);
net/rxrpc/local_object.c
333
__refcount_inc(&local->ref, &r);
net/rxrpc/local_object.c
346
if (local && __refcount_inc_not_zero(&local->ref, &r)) {
net/rxrpc/local_object.c
368
dead = __refcount_dec_and_test(&local->ref, &r);
net/rxrpc/local_object.c
405
r = refcount_read(&local->ref);
net/rxrpc/local_object.c
481
local, refcount_read(&local->ref));
net/rxrpc/peer_object.c
126
refcount_read(&peer->ref) > 0)
net/rxrpc/peer_object.c
144
_leave(" = %p {u=%d}", peer, refcount_read(&peer->ref));
net/rxrpc/peer_object.c
231
refcount_set(&peer->ref, 1);
net/rxrpc/peer_object.c
381
_leave(" = %p {u=%d}", peer, refcount_read(&peer->ref));
net/rxrpc/peer_object.c
392
__refcount_inc(&peer->ref, &r);
net/rxrpc/peer_object.c
406
if (__refcount_inc_not_zero(&peer->ref, &r))
net/rxrpc/peer_object.c
442
dead = __refcount_dec_and_test(&peer->ref, &r);
net/rxrpc/peer_object.c
464
refcount_read(&peer->ref),
net/rxrpc/proc.c
178
refcount_read(&conn->ref),
net/rxrpc/proc.c
253
refcount_read(&bundle->ref),
net/rxrpc/proc.c
302
refcount_read(&peer->ref),
net/rxrpc/proc.c
420
refcount_read(&local->ref),
net/rxrpc/proc.c
92
refcount_read(&call->ref),
net/rxrpc/txbuf.c
46
refcount_set(&txb->ref, 1);
net/rxrpc/txbuf.c
65
int r = refcount_read(&txb->ref);
net/rxrpc/txbuf.c
91
dead = __refcount_dec_and_test(&txb->ref, &r);
net/sched/act_api.c
1241
tcf_action_dump_old(struct sk_buff *skb, struct tc_action *a, int bind, int ref)
net/sched/act_api.c
1243
return a->ops->dump(skb, a, bind, ref);
net/sched/act_api.c
1247
int bind, int ref, bool terse)
net/sched/act_api.c
1258
tcf_action_dump_1(skb, a, bind, ref);
net/sched/act_api.c
1649
int ref, struct netlink_ext_ack *extack)
net/sched/act_api.c
1672
if (tcf_action_dump(skb, actions, bind, ref, false) < 0)
net/sched/act_api.c
508
tcf_action_dump_1(struct sk_buff *skb, struct tc_action *a, int bind, int ref)
net/sched/act_api.c
540
err = tcf_action_dump_old(skb, a, bind, ref);
net/sched/act_bpf.c
137
int bind, int ref)
net/sched/act_bpf.c
143
.refcnt = refcount_read(&prog->tcf_refcnt) - ref,
net/sched/act_connmark.c
193
int bind, int ref)
net/sched/act_connmark.c
204
opt.refcnt = refcount_read(&ci->tcf_refcnt) - ref;
net/sched/act_csum.c
633
int ref)
net/sched/act_csum.c
640
.refcnt = refcount_read(&p->tcf_refcnt) - ref,
net/sched/act_ct.c
1531
int bind, int ref)
net/sched/act_ct.c
1538
.refcnt = refcount_read(&c->tcf_refcnt) - ref,
net/sched/act_ct.c
332
if (ct_ft && refcount_inc_not_zero(&ct_ft->ref))
net/sched/act_ct.c
338
refcount_set(&ct_ft->ref, 1);
net/sched/act_ct.c
372
refcount_inc(&ct_ft->ref);
net/sched/act_ct.c
395
if (refcount_dec_and_test(&ct_ft->ref)) {
net/sched/act_ct.c
56
refcount_t ref;
net/sched/act_ctinfo.c
284
int bind, int ref)
net/sched/act_ctinfo.c
291
.refcnt = refcount_read(&ci->tcf_refcnt) - ref,
net/sched/act_gact.c
186
int bind, int ref)
net/sched/act_gact.c
192
.refcnt = refcount_read(&gact->tcf_refcnt) - ref,
net/sched/act_gate.c
606
int bind, int ref)
net/sched/act_gate.c
612
.refcnt = refcount_read(&gact->tcf_refcnt) - ref,
net/sched/act_ife.c
638
int ref)
net/sched/act_ife.c
649
opt.refcnt = refcount_read(&ife->tcf_refcnt) - ref;
net/sched/act_mirred.c
488
int ref)
net/sched/act_mirred.c
494
.refcnt = refcount_read(&m->tcf_refcnt) - ref,
net/sched/act_mpls.c
329
int bind, int ref)
net/sched/act_mpls.c
336
.refcnt = refcount_read(&m->tcf_refcnt) - ref,
net/sched/act_nat.c
268
int bind, int ref)
net/sched/act_nat.c
275
.refcnt = refcount_read(&p->tcf_refcnt) - ref,
net/sched/act_pedit.c
500
int bind, int ref)
net/sched/act_pedit.c
525
opt->refcnt = refcount_read(&p->tcf_refcnt) - ref;
net/sched/act_police.c
340
int bind, int ref)
net/sched/act_police.c
347
.refcnt = refcount_read(&police->tcf_refcnt) - ref,
net/sched/act_sample.c
225
int bind, int ref)
net/sched/act_sample.c
231
.refcnt = refcount_read(&s->tcf_refcnt) - ref,
net/sched/act_simple.c
171
int bind, int ref)
net/sched/act_simple.c
177
.refcnt = refcount_read(&d->tcf_refcnt) - ref,
net/sched/act_skbedit.c
288
int bind, int ref)
net/sched/act_skbedit.c
295
.refcnt = refcount_read(&d->tcf_refcnt) - ref,
net/sched/act_skbmod.c
238
int bind, int ref)
net/sched/act_skbmod.c
248
opt.refcnt = refcount_read(&d->tcf_refcnt) - ref;
net/sched/act_tunnel_key.c
716
int bind, int ref)
net/sched/act_tunnel_key.c
723
.refcnt = refcount_read(&t->tcf_refcnt) - ref,
net/sched/act_vlan.c
286
int bind, int ref)
net/sched/act_vlan.c
293
.refcnt = refcount_read(&v->tcf_refcnt) - ref,
net/sched/ematch.c
186
u32 ref;
net/sched/ematch.c
188
if (data_len < sizeof(ref))
net/sched/ematch.c
190
ref = *(u32 *) data;
net/sched/ematch.c
192
if (ref >= tree_hdr->nmatches)
net/sched/ematch.c
198
if (ref <= idx)
net/sched/ematch.c
202
em->data = ref;
net/smc/smc_wr.c
841
static void smcr_wr_tx_refs_free(struct percpu_ref *ref)
net/smc/smc_wr.c
843
struct smc_link *lnk = container_of(ref, struct smc_link, wr_tx_refs);
net/smc/smc_wr.c
848
static void smcr_wr_reg_refs_free(struct percpu_ref *ref)
net/smc/smc_wr.c
850
struct smc_link *lnk = container_of(ref, struct smc_link, wr_reg_refs);
net/sunrpc/auth_gss/svcauth_gss.c
125
static void rsi_put(struct kref *ref)
net/sunrpc/auth_gss/svcauth_gss.c
127
struct rsi *rsii = container_of(ref, struct rsi, h.ref);
net/sunrpc/auth_gss/svcauth_gss.c
396
static void rsc_put(struct kref *ref)
net/sunrpc/auth_gss/svcauth_gss.c
398
struct rsc *rsci = container_of(ref, struct rsc, h.ref);
net/sunrpc/auth_gss/svcauth_gss.c
820
kref_init(&new->h.ref);
net/sunrpc/cache.c
1458
kref_read(&cp->ref), cp->flags);
net/sunrpc/cache.c
51
kref_init(&h->ref);
net/sunrpc/svcauth.c
224
struct auth_domain *dom = container_of(kref, struct auth_domain, ref);
net/sunrpc/svcauth.c
233
kref_put_lock(&dom->ref, auth_domain_release, &auth_domain_lock);
net/sunrpc/svcauth.c
249
kref_get(&hp->ref);
net/sunrpc/svcauth.c
271
if (!kref_get_unless_zero(&hp->ref))
net/sunrpc/svcauth_unix.c
108
struct cache_head *item = container_of(kref, struct cache_head, ref);
net/sunrpc/svcauth_unix.c
141
kref_get(&item->m_client->h.ref);
net/sunrpc/svcauth_unix.c
433
struct cache_head *item = container_of(kref, struct cache_head, ref);
net/sunrpc/svcauth_unix.c
718
kref_get(&rqstp->rq_client->ref);
net/sunrpc/svcauth_unix.c
78
kref_init(&new->h.ref);
net/tipc/group.c
681
evt.port.ref = m->port;
net/tipc/group.c
865
u32 port = evt->port.ref;
net/tipc/msg.c
731
msg_set_destport(msg, sk.ref);
net/tipc/name_distr.c
289
sk.ref = ntohl(i->port);
net/tipc/name_distr.c
54
i->port = htonl(p->sk.ref);
net/tipc/name_table.c
1007
if (nla_put_u32(msg->skb, TIPC_NLA_PUBL_REF, p->sk.ref))
net/tipc/name_table.c
351
if (_p->key == key && _p->sk.ref == p->sk.ref &&
net/tipc/name_table.c
355
node, p->sk.ref, key);
net/tipc/name_table.c
392
if (p->key != key || p->sk.ref != sk->ref ||
net/tipc/name_table.c
545
sk->node, sk->ref, key);
net/tipc/name_table.c
656
if (p->sk.ref == exclude && p->sk.node == self)
net/tipc/name_table.c
658
tipc_dest_push(dsts, p->sk.node, p->sk.ref);
net/tipc/name_table.c
694
tipc_dest_push(dports, 0, p->sk.ref);
net/tipc/name_table.c
751
tipc_group_add_member(grp, p->sk.node, p->sk.ref,
net/tipc/node.c
413
sk.ref = n->link_id;
net/tipc/node.c
432
tipc_nametbl_publish(net, &ua, &sk, sk.ref);
net/tipc/node.c
436
tipc_nametbl_withdraw(net, &ua, &sk, sk.ref);
net/tipc/socket.c
1174
tipc_dest_pop(&dsts, &ua->sk.node, &ua->sk.ref);
net/tipc/socket.c
1485
msg_set_destport(hdr, skaddr.ref);
net/tipc/socket.c
1704
srcaddr->sock.addr.id.ref = msg_origport(hdr);
net/tipc/socket.c
2915
skaddr.ref = tsk->portid;
net/tipc/socket.c
3339
tsk1->peer.addr.id.ref = tsk2->portid;
net/tipc/socket.c
3344
tsk2->peer.addr.id.ref = tsk1->portid;
net/tipc/socket.c
756
addr->addr.id.ref = tsk_peer_port(tsk);
net/tipc/socket.c
759
addr->addr.id.ref = tsk->portid;
net/tipc/socket.c
968
port = ua->sk.ref;
net/tipc/subscr.c
55
tipc_evt_write(evt, port.ref, p->sk.ref);
net/tipc/subscr.c
60
tipc_evt_write(evt, port.ref, 0);
scripts/asn1_compiler.c
1012
ref = bsearch(cursor, type_index, nr_types, sizeof(type_index[0]),
scripts/asn1_compiler.c
1014
if (!ref) {
scripts/asn1_compiler.c
1019
cursor->type = *ref;
scripts/asn1_compiler.c
1020
(*ref)->ref_count++;
scripts/asn1_compiler.c
859
struct type **ref;
scripts/dtc/checks.c
489
check_duplicate_label(c, dti, m->ref, node, prop, m);
scripts/dtc/checks.c
515
if (node != get_node_by_ref(root, m->ref))
scripts/dtc/checks.c
625
refnode = get_node_by_ref(dt, m->ref);
scripts/dtc/checks.c
629
"label \"%s\"\n", m->ref);
scripts/dtc/checks.c
660
refnode = get_node_by_ref(dt, m->ref);
scripts/dtc/checks.c
663
m->ref);
scripts/dtc/data.c
15
free(m->ref);
scripts/dtc/data.c
227
struct data data_add_marker(struct data d, enum markertype type, char *ref)
scripts/dtc/data.c
231
m = alloc_marker(d.len, type, ref);
scripts/dtc/data.c
260
char *ref;
scripts/dtc/data.c
267
ref = NULL;
scripts/dtc/data.c
269
if (marker->ref)
scripts/dtc/data.c
270
ref = xstrdup(marker->ref);
scripts/dtc/data.c
273
ref);
scripts/dtc/data.c
282
char *ref)
scripts/dtc/data.c
289
m->ref = ref;
scripts/dtc/dtc-parser.y
27
static bool is_ref_relative(const char *ref)
scripts/dtc/dtc-parser.y
29
return ref[0] != '/' && strchr(&ref[1], '/');
scripts/dtc/dtc.h
130
char *ref;
scripts/dtc/dtc.h
188
char *ref);
scripts/dtc/dtc.h
189
struct data data_add_marker(struct data d, enum markertype type, char *ref);
scripts/dtc/dtc.h
280
struct node *add_orphan_node(struct node *old_node, struct node *new_node, char *ref);
scripts/dtc/dtc.h
304
struct node *get_node_by_ref(struct node *tree, const char *ref);
scripts/dtc/dtc.h
364
enum markertype type, unsigned int offset, char *ref);
scripts/dtc/flattree.c
157
emit_offset_label(f, m->ref, m->offset);
scripts/dtc/livetree.c
1017
refnode = get_node_by_ref(dt, m->ref);
scripts/dtc/livetree.c
1040
if (get_node_by_ref(dti->dt, m->ref))
scripts/dtc/livetree.c
1101
refnode = get_node_by_ref(dt, m->ref);
scripts/dtc/livetree.c
228
struct node * add_orphan_node(struct node *dt, struct node *new_node, char *ref)
scripts/dtc/livetree.c
236
if (ref[0] == '/') {
scripts/dtc/livetree.c
237
d = data_add_marker(d, TYPE_STRING, ref);
scripts/dtc/livetree.c
238
d = data_append_data(d, ref, strlen(ref) + 1);
scripts/dtc/livetree.c
242
d = data_add_marker(d, REF_PHANDLE, ref);
scripts/dtc/livetree.c
540
if (streq(m->ref, label))
scripts/dtc/livetree.c
636
struct node *get_node_by_ref(struct node *tree, const char *ref)
scripts/dtc/livetree.c
641
if (streq(ref, "/"))
scripts/dtc/livetree.c
644
if (ref[0] == '/')
scripts/dtc/livetree.c
645
path = ref;
scripts/dtc/livetree.c
647
label = ref;
scripts/dtc/livetree.c
961
if (!get_node_by_ref(dti->dt, m->ref))
scripts/dtc/livetree.c
986
if (strchr(m->ref, '/'))
scripts/dtc/livetree.c
988
m->ref);
scripts/dtc/livetree.c
996
ret = append_unique_str_to_property(fn, m->ref, entry, strlen(entry) + 1);
scripts/dtc/treesource.c
150
enum markertype type, unsigned int offset, char *ref)
scripts/dtc/treesource.c
169
nm->ref = ref;
scripts/dtc/treesource.c
177
enum markertype type, unsigned int offset, char *ref)
scripts/dtc/treesource.c
179
add_marker(&prop->val.markers, type, offset, ref);
scripts/dtc/treesource.c
196
char *ref;
scripts/dtc/treesource.c
218
ref = refn->labels->label;
scripts/dtc/treesource.c
220
ref = refn->fullpath;
scripts/dtc/treesource.c
222
add_marker(&prop->val.markers, REF_PHANDLE, offset, ref);
scripts/dtc/treesource.c
312
fprintf(f, " %s:", m->ref);
scripts/dtc/treesource.c
328
if (m_phandle->ref[0] == '/')
scripts/dtc/treesource.c
329
fprintf(f, "&{%s}", m_phandle->ref);
scripts/dtc/treesource.c
331
fprintf(f, "&%s", m_phandle->ref);
scripts/gcc-plugins/gcc-common.h
402
static inline cgraph_node_ptr ipa_ref_referring_node(struct ipa_ref *ref)
scripts/gcc-plugins/gcc-common.h
404
return dyn_cast<cgraph_node_ptr>(ref->referring);
security/apparmor/apparmorfs.c
158
static struct aa_ns *get_ns_common_ref(struct aa_common_ref *ref)
security/apparmor/apparmorfs.c
160
if (ref) {
security/apparmor/apparmorfs.c
161
struct aa_label *reflabel = container_of(ref, struct aa_label,
security/apparmor/apparmorfs.c
169
static struct aa_proxy *get_proxy_common_ref(struct aa_common_ref *ref)
security/apparmor/apparmorfs.c
1706
struct aa_common_ref *ref = inode->i_private;
security/apparmor/apparmorfs.c
1707
struct aa_proxy *proxy = container_of(ref, struct aa_proxy, count);
security/apparmor/apparmorfs.c
171
if (ref)
security/apparmor/apparmorfs.c
172
return aa_get_proxy(container_of(ref, struct aa_proxy, count));
security/apparmor/apparmorfs.c
177
static struct aa_loaddata *get_loaddata_common_ref(struct aa_common_ref *ref)
security/apparmor/apparmorfs.c
179
if (ref)
security/apparmor/apparmorfs.c
180
return aa_get_i_loaddata(container_of(ref, struct aa_loaddata,
security/apparmor/apparmorfs.c
185
static void aa_put_common_ref(struct aa_common_ref *ref)
security/apparmor/apparmorfs.c
187
if (!ref)
security/apparmor/apparmorfs.c
190
switch (ref->reftype) {
security/apparmor/apparmorfs.c
192
aa_put_i_loaddata(container_of(ref, struct aa_loaddata,
security/apparmor/apparmorfs.c
196
aa_put_proxy(container_of(ref, struct aa_proxy,
security/apparmor/apparmorfs.c
201
aa_put_ns(labels_ns(container_of(ref, struct aa_label, count)));
security/apparmor/apparmorfs.c
209
static void aa_get_common_ref(struct aa_common_ref *ref)
security/apparmor/apparmorfs.c
211
kref_get(&ref->count);
security/apparmor/apparmorfs.c
216
struct aa_common_ref *ref = inode->i_private;
security/apparmor/apparmorfs.c
219
aa_put_common_ref(ref);
sound/aoa/fabrics/layout.c
773
const u32 *ref;
sound/aoa/fabrics/layout.c
781
ref = of_get_property(ldev->sound, propname, NULL);
sound/aoa/fabrics/layout.c
782
if (!ref) {
sound/aoa/fabrics/layout.c
787
if (*ref != codec->node->phandle) {
sound/core/sound.c
127
struct snd_card *ref = snd_card_ref(card);
sound/core/sound.c
128
if (!ref)
sound/core/sound.c
131
snd_card_unref(ref);
sound/pci/mixart/mixart.c
1153
u32 ref;
sound/pci/mixart/mixart.c
1171
ref = readl_be( MIXART_MEM( chip->mgr, MIXART_PSEUDOREG_PERF_SYSTEM_LOAD_OFFSET));
sound/pci/mixart/mixart.c
1173
if (ref) {
sound/pci/mixart/mixart.c
1174
u32 mailbox = 100 * readl_be( MIXART_MEM( chip->mgr, MIXART_PSEUDOREG_PERF_MAILBX_LOAD_OFFSET)) / ref;
sound/pci/mixart/mixart.c
1175
u32 streaming = 100 * readl_be( MIXART_MEM( chip->mgr, MIXART_PSEUDOREG_PERF_STREAM_LOAD_OFFSET)) / ref;
sound/pci/mixart/mixart.c
1176
u32 interr = 100 * readl_be( MIXART_MEM( chip->mgr, MIXART_PSEUDOREG_PERF_INTERR_LOAD_OFFSET)) / ref;
sound/pci/pcxhr/pcxhr.c
1225
int ref = rmh.stat[1];
sound/pci/pcxhr/pcxhr.c
1226
if (ref > 0) {
sound/pci/pcxhr/pcxhr.c
1229
ref = (ref * 48000) /
sound/pci/pcxhr/pcxhr.c
1233
ref *= 2;
sound/pci/pcxhr/pcxhr.c
1235
cur = 100 - (100 * cur) / ref;
sound/soc/codecs/arizona.c
1280
int ref, div, refclk;
sound/soc/codecs/arizona.c
1300
for (ref = 0; ref < ARRAY_SIZE(arizona_opclk_ref_48k_rates) &&
sound/soc/codecs/arizona.c
1301
rates[ref] <= refclk; ref++) {
sound/soc/codecs/arizona.c
1303
while (rates[ref] / div >= freq && div < 32) {
sound/soc/codecs/arizona.c
1304
if (rates[ref] / div == freq) {
sound/soc/codecs/arizona.c
1312
ref);
sound/soc/codecs/madera.c
2495
int ref, div, refclk;
sound/soc/codecs/madera.c
2518
for (ref = 0; ref < ARRAY_SIZE(madera_opclk_ref_48k_rates); ++ref) {
sound/soc/codecs/madera.c
2519
if (rates[ref] > refclk)
sound/soc/codecs/madera.c
2523
while ((rates[ref] / div >= freq) && (div <= 30)) {
sound/soc/codecs/madera.c
2524
if (rates[ref] / div == freq) {
sound/soc/codecs/madera.c
2528
val = (div << MADERA_OPCLK_DIV_SHIFT) | ref;
sound/soc/codecs/rt5668.c
1149
int ref, val, reg, idx;
sound/soc/codecs/rt5668.c
1156
ref = 256 * rt5668->lrck[RT5668_AIF2];
sound/soc/codecs/rt5668.c
1158
ref = 256 * rt5668->lrck[RT5668_AIF1];
sound/soc/codecs/rt5668.c
1160
idx = rt5668_div_sel(rt5668, ref, div, ARRAY_SIZE(div));
sound/soc/codecs/rt5682.c
1285
int ref, val, reg, idx;
sound/soc/codecs/rt5682.c
1296
ref = 256 * rt5682->lrck[RT5682_AIF2];
sound/soc/codecs/rt5682.c
1298
ref = 256 * rt5682->lrck[RT5682_AIF1];
sound/soc/codecs/rt5682.c
1300
idx = rt5682_div_sel(rt5682, ref, div_f, ARRAY_SIZE(div_f));
sound/soc/codecs/rt5682s.c
1197
static void rt5682s_set_filter_clk(struct rt5682s_priv *rt5682s, int reg, int ref)
sound/soc/codecs/rt5682s.c
1204
idx = rt5682s_div_sel(rt5682s, ref, div_f, ARRAY_SIZE(div_f));
sound/soc/codecs/rt5682s.c
1225
int ref, reg, val;
sound/soc/codecs/rt5682s.c
1231
ref = 256 * rt5682s->lrck[RT5682S_AIF2];
sound/soc/codecs/rt5682s.c
1233
ref = 256 * rt5682s->lrck[RT5682S_AIF1];
sound/soc/codecs/rt5682s.c
1240
rt5682s_set_filter_clk(rt5682s, reg, ref);
sound/soc/codecs/rt5682s.c
2532
int ref, reg;
sound/soc/codecs/rt5682s.c
2553
ref = 256 * rt5682s->lrck[RT5682S_AIF1];
sound/soc/codecs/rt5682s.c
2554
rt5682s_set_filter_clk(rt5682s, reg, ref);
sound/soc/qcom/qdsp6/q6adm.c
171
static void q6adm_free_copp(struct kref *ref)
sound/soc/qcom/qdsp6/q6adm.c
173
struct q6copp *c = container_of(ref, struct q6copp, refcount);
sound/soc/qcom/qdsp6/q6afe.c
930
static void q6afe_port_free(struct kref *ref)
sound/soc/qcom/qdsp6/q6afe.c
935
port = container_of(ref, struct q6afe_port, refcount);
sound/soc/qcom/qdsp6/q6apm.c
123
static void q6apm_put_audioreach_graph(struct kref *ref)
sound/soc/qcom/qdsp6/q6apm.c
128
graph = container_of(ref, struct audioreach_graph, refcount);
sound/soc/qcom/qdsp6/q6asm.c
541
static void q6asm_audio_client_release(struct kref *ref)
sound/soc/qcom/qdsp6/q6asm.c
547
ac = container_of(ref, struct audio_client, refcount);
sound/usb/endpoint.c
1528
struct snd_usb_clock_ref *ref;
sound/usb/endpoint.c
1534
list_for_each_entry(ref, &chip->clock_ref_list, list) {
sound/usb/endpoint.c
1535
if (ref->clock == clock) {
sound/usb/endpoint.c
1536
rate = ref->rate;
sound/usb/endpoint.c
637
struct snd_usb_clock_ref *ref;
sound/usb/endpoint.c
639
list_for_each_entry(ref, &chip->clock_ref_list, list)
sound/usb/endpoint.c
640
if (ref->clock == clock)
sound/usb/endpoint.c
641
return ref;
sound/usb/endpoint.c
643
ref = kzalloc_obj(*ref);
sound/usb/endpoint.c
644
if (!ref)
sound/usb/endpoint.c
646
ref->clock = clock;
sound/usb/endpoint.c
647
atomic_set(&ref->locked, 0);
sound/usb/endpoint.c
648
list_add_tail(&ref->list, &chip->clock_ref_list);
sound/usb/endpoint.c
649
return ref;
tools/bpf/bpftool/main.c
233
const char *ref;
tools/bpf/bpftool/main.c
239
while ((ref = va_arg(ap, const char *))) {
tools/bpf/bpftool/main.c
240
if (!is_prefix(arg, ref))
tools/bpf/bpftool/main.c
245
strncat(msg, ref, sizeof(msg) - strlen(msg) - 1);
tools/bpf/bpftool/pids.c
223
struct obj_ref *ref = &refs->refs[i];
tools/bpf/bpftool/pids.c
226
jsonw_int_field(json_writer, "pid", ref->pid);
tools/bpf/bpftool/pids.c
227
jsonw_string_field(json_writer, "comm", ref->comm);
tools/bpf/bpftool/pids.c
254
struct obj_ref *ref = &refs->refs[i];
tools/bpf/bpftool/pids.c
256
printf("%s%s(%d)", i == 0 ? "" : ", ", ref->comm, ref->pid);
tools/bpf/bpftool/pids.c
35
struct obj_ref *ref;
tools/bpf/bpftool/pids.c
47
tmp = realloc(refs->refs, (refs->ref_cnt + 1) * sizeof(*ref));
tools/bpf/bpftool/pids.c
54
ref = &refs->refs[refs->ref_cnt];
tools/bpf/bpftool/pids.c
55
ref->pid = e->pid;
tools/bpf/bpftool/pids.c
56
memcpy(ref->comm, e->comm, sizeof(ref->comm));
tools/bpf/bpftool/pids.c
57
ref->comm[sizeof(ref->comm) - 1] = '\0';
tools/bpf/bpftool/pids.c
78
ref = &refs->refs[0];
tools/bpf/bpftool/pids.c
79
ref->pid = e->pid;
tools/bpf/bpftool/pids.c
80
memcpy(ref->comm, e->comm, sizeof(ref->comm));
tools/bpf/bpftool/pids.c
81
ref->comm[sizeof(ref->comm) - 1] = '\0';
tools/lib/bpf/bpf_gen_internal.h
20
int ref;
tools/lib/bpf/gen_loader.c
673
kdesc->ref++;
tools/lib/bpf/gen_loader.c
686
kdesc->ref = 1;
tools/lib/bpf/gen_loader.c
753
if (kdesc->ref > 1) {
tools/lib/bpf/gen_loader.c
807
relo->name, kdesc->ref);
tools/lib/bpf/gen_loader.c
812
relo->name, kdesc->ref);
tools/lib/bpf/gen_loader.c
816
int ref)
tools/lib/bpf/gen_loader.c
825
relo->is_typeless, relo->is_weak, relo->name, ref);
tools/lib/bpf/gen_loader.c
828
relo->is_typeless, relo->is_weak, relo->name, ref);
tools/lib/bpf/gen_loader.c
843
if (kdesc->ref > 1) {
tools/lib/bpf/gen_loader.c
864
emit_ksym_relo_log(gen, relo, kdesc->ref);
tools/lib/bpf/gen_loader.c
890
if (kdesc->ref > 1) {
tools/lib/bpf/gen_loader.c
928
emit_ksym_relo_log(gen, relo, kdesc->ref);
tools/mm/slabinfo.c
1123
n1 = a1->ref;
tools/mm/slabinfo.c
1124
n2 = a2->ref;
tools/mm/slabinfo.c
1145
if (strcmp(a->ref, s->name) == 0) {
tools/mm/slabinfo.c
1151
fatal("Unresolved alias %s\n", a->ref);
tools/mm/slabinfo.c
1246
alias->ref = strdup(p);
tools/mm/slabinfo.c
53
char *ref;
tools/net/ynl/ynltool/main.c
121
const char *ref;
tools/net/ynl/ynltool/main.c
127
while ((ref = va_arg(ap, const char *))) {
tools/net/ynl/ynltool/main.c
128
if (!is_prefix(arg, ref))
tools/net/ynl/ynltool/main.c
133
strncat(msg, ref, sizeof(msg) - strlen(msg) - 1);
tools/perf/arch/x86/util/intel-pt.c
1082
struct intel_pt_snapshot_ref *ref = &ptr->snapshot_refs[idx];
tools/perf/arch/x86/util/intel-pt.c
1085
wrapped = intel_pt_compare_ref(ref->ref_buf, ref->ref_offset,
tools/perf/arch/x86/util/intel-pt.c
1089
intel_pt_copy_ref(ref->ref_buf, ptr->snapshot_ref_buf_size, mm->len,
tools/perf/builtin-record.c
2264
u64 ref;
tools/perf/builtin-record.c
2284
ref = (u64) ref_tod.tv_sec * NSEC_PER_SEC +
tools/perf/builtin-record.c
2287
env->clock.tod_ns = ref;
tools/perf/builtin-record.c
2289
ref = (u64) ref_clockid.tv_sec * NSEC_PER_SEC +
tools/perf/builtin-record.c
2292
env->clock.clockid_ns = ref;
tools/perf/util/auxtrace.c
1918
u64 head, old = mm->prev, offset, ref;
tools/perf/util/auxtrace.c
1953
ref = auxtrace_record__reference(itr);
tools/perf/util/auxtrace.c
1997
ev.auxtrace.reference = ref;
tools/perf/util/expr.c
180
int expr__add_ref(struct expr_parse_ctx *ctx, struct metric_ref *ref)
tools/perf/util/expr.c
191
name = strdup(ref->metric_name);
tools/perf/util/expr.c
203
data_ptr->ref.metric_name = ref->metric_name;
tools/perf/util/expr.c
204
data_ptr->ref.metric_expr = ref->metric_expr;
tools/perf/util/expr.c
212
ref->metric_name, ref->metric_expr);
tools/perf/util/expr.c
260
data->ref.metric_name);
tools/perf/util/expr.c
263
if (expr__parse(&data->ref.val, ctx, data->ref.metric_expr)) {
tools/perf/util/expr.c
267
pr_debug("processing metric: %s EXIT: %f\n", id, data->ref.val);
tools/perf/util/expr.c
271
data->ref.val, data->ref.metric_name);
tools/perf/util/expr.c
35
} ref;
tools/perf/util/expr.c
387
return data->ref.val;
tools/perf/util/expr.h
40
int expr__add_ref(struct expr_parse_ctx *ctx, struct metric_ref *ref);
tools/perf/util/header.c
1924
u64 ref;
tools/perf/util/header.c
1932
ref = ff->ph->env.clock.tod_ns;
tools/perf/util/header.c
1933
tod_ns.tv_sec = ref / NSEC_PER_SEC;
tools/perf/util/header.c
1934
ref -= tod_ns.tv_sec * NSEC_PER_SEC;
tools/perf/util/header.c
1935
tod_ns.tv_usec = ref / NSEC_PER_USEC;
tools/perf/util/header.c
1938
ref = ff->ph->env.clock.clockid_ns;
tools/perf/util/header.c
1939
clockid_ns.tv_sec = ref / NSEC_PER_SEC;
tools/perf/util/header.c
1940
ref -= clockid_ns.tv_sec * NSEC_PER_SEC;
tools/perf/util/header.c
1941
clockid_ns.tv_nsec = ref;
tools/perf/util/hist.c
2922
char ref[30] = " show reference callgraph, ";
tools/perf/util/hist.c
2960
ev_name, sample_freq_str, enable_ref ? ref : " ", nr_events);
tools/perf/util/llvm.c
103
const char **ref __maybe_unused)
tools/perf/util/probe-event.c
1771
(*fieldp)->ref = true;
tools/perf/util/probe-event.c
1783
(*fieldp)->ref = false;
tools/perf/util/probe-event.c
1786
(*fieldp)->ref = true;
tools/perf/util/probe-event.c
1800
pr_debug("%s(%d), ", (*fieldp)->name, (*fieldp)->ref);
tools/perf/util/probe-event.c
1809
pr_debug("%s(%d)\n", (*fieldp)->name, (*fieldp)->ref);
tools/perf/util/probe-event.c
2051
err = strbuf_addf(&buf, "%s%s", field->ref ? "->" : ".",
tools/perf/util/probe-event.c
2143
static int __synthesize_probe_trace_arg_ref(struct probe_trace_arg_ref *ref,
tools/perf/util/probe-event.c
2147
if (ref->next) {
tools/perf/util/probe-event.c
2148
depth = __synthesize_probe_trace_arg_ref(ref->next, buf,
tools/perf/util/probe-event.c
2153
if (ref->user_access)
tools/perf/util/probe-event.c
2154
err = strbuf_addf(buf, "%s%ld(", "+u", ref->offset);
tools/perf/util/probe-event.c
2156
err = strbuf_addf(buf, "%+ld(", ref->offset);
tools/perf/util/probe-event.c
2163
struct probe_trace_arg_ref *ref = arg->ref;
tools/perf/util/probe-event.c
2175
if (arg->value[0] == '@' && arg->ref)
tools/perf/util/probe-event.c
2176
ref = ref->next;
tools/perf/util/probe-event.c
2179
if (ref) {
tools/perf/util/probe-event.c
2180
depth = __synthesize_probe_trace_arg_ref(ref, buf, 1);
tools/perf/util/probe-event.c
2186
if (arg->value[0] == '@' && arg->ref)
tools/perf/util/probe-event.c
2187
err = strbuf_addf(buf, "%s%+ld", arg->value, arg->ref->offset);
tools/perf/util/probe-event.c
2461
(*ppfield)->ref = field->ref;
tools/perf/util/probe-event.c
2500
struct probe_trace_arg_ref *ref, *next;
tools/perf/util/probe-event.c
2512
ref = tev->args[i].ref;
tools/perf/util/probe-event.c
2513
while (ref) {
tools/perf/util/probe-event.c
2514
next = ref->next;
tools/perf/util/probe-event.c
2515
free(ref);
tools/perf/util/probe-event.c
2516
ref = next;
tools/perf/util/probe-event.h
52
struct probe_trace_arg_ref *ref; /* Referencing offset */
tools/perf/util/probe-event.h
82
bool ref; /* Referencing flag */
tools/perf/util/probe-finder.c
147
tvar->ref = alloc_trace_arg_ref((long)offs);
tools/perf/util/probe-finder.c
148
if (tvar->ref == NULL)
tools/perf/util/probe-finder.c
157
ref = true;
tools/perf/util/probe-finder.c
165
ref = true;
tools/perf/util/probe-finder.c
171
ref = true;
tools/perf/util/probe-finder.c
194
if (ref) {
tools/perf/util/probe-finder.c
195
tvar->ref = alloc_trace_arg_ref((long)offs);
tools/perf/util/probe-finder.c
196
if (tvar->ref == NULL)
tools/perf/util/probe-finder.c
206
struct probe_trace_arg_ref **ref_ptr = &tvar->ref;
tools/perf/util/probe-finder.c
326
struct probe_trace_arg_ref *ref = *ref_ptr;
tools/perf/util/probe-finder.c
352
ref = zalloc(sizeof(struct probe_trace_arg_ref));
tools/perf/util/probe-finder.c
353
if (ref == NULL)
tools/perf/util/probe-finder.c
356
(*ref_ptr)->next = ref;
tools/perf/util/probe-finder.c
358
*ref_ptr = ref;
tools/perf/util/probe-finder.c
360
ref->offset += dwarf_bytesize(&type) * field->index;
tools/perf/util/probe-finder.c
361
ref->user_access = user_access;
tools/perf/util/probe-finder.c
365
if (!field->ref) {
tools/perf/util/probe-finder.c
383
ref = zalloc(sizeof(struct probe_trace_arg_ref));
tools/perf/util/probe-finder.c
384
if (ref == NULL)
tools/perf/util/probe-finder.c
387
(*ref_ptr)->next = ref;
tools/perf/util/probe-finder.c
389
*ref_ptr = ref;
tools/perf/util/probe-finder.c
403
if (field->ref && dwarf_diename(vr_die)) {
tools/perf/util/probe-finder.c
408
if (!ref) {
tools/perf/util/probe-finder.c
432
ref->offset += (long)offs;
tools/perf/util/probe-finder.c
433
ref->user_access = user_access;
tools/perf/util/probe-finder.c
438
&ref, die_mem, user_access);
tools/perf/util/probe-finder.c
444
field->next, &ref, die_mem, user_access);
tools/perf/util/probe-finder.c
477
pf->pvar->field, &pf->tvar->ref,
tools/perf/util/probe-finder.c
57
struct probe_trace_arg_ref *ref;
tools/perf/util/probe-finder.c
58
ref = zalloc(sizeof(struct probe_trace_arg_ref));
tools/perf/util/probe-finder.c
59
if (ref != NULL)
tools/perf/util/probe-finder.c
60
ref->offset = offs;
tools/perf/util/probe-finder.c
61
return ref;
tools/perf/util/probe-finder.c
81
bool ref = false;
tools/perf/util/session.c
2680
struct ref_reloc_sym *ref;
tools/perf/util/session.c
2683
ref = zalloc(sizeof(struct ref_reloc_sym));
tools/perf/util/session.c
2684
if (ref == NULL)
tools/perf/util/session.c
2687
ref->name = name = strdup(symbol_name);
tools/perf/util/session.c
2688
if (ref->name == NULL) {
tools/perf/util/session.c
2689
free(ref);
tools/perf/util/session.c
2697
ref->addr = addr;
tools/perf/util/session.c
2701
kmap->ref_reloc_sym = ref;
tools/perf/util/thread-stack.c
1011
timestamp, ref, false);
tools/perf/util/thread-stack.c
1020
struct perf_sample *sample, u64 ref)
tools/perf/util/thread-stack.c
1027
if (!ts->cnt || (ts->cnt == 1 && ts->stack[0].ref == ref))
tools/perf/util/thread-stack.c
1035
return thread_stack__push_cp(ts, ret_addr, sample->time, ref, cp,
tools/perf/util/thread-stack.c
1104
struct addr_location *to_al, u64 ref,
tools/perf/util/thread-stack.c
1138
err = thread_stack__bottom(ts, sample, from_al, to_al, ref);
tools/perf/util/thread-stack.c
1164
err = thread_stack__push_cp(ts, ret_addr, sample->time, ref,
tools/perf/util/thread-stack.c
1185
return thread_stack__pop_ks(thread, ts, sample, ref);
tools/perf/util/thread-stack.c
1197
sample->time, ref, from_al->sym);
tools/perf/util/thread-stack.c
1202
from_al, to_al, ref);
tools/perf/util/thread-stack.c
1205
err = thread_stack__trace_begin(thread, ts, sample->time, ref);
tools/perf/util/thread-stack.c
1207
err = thread_stack__trace_end(ts, sample, ref);
tools/perf/util/thread-stack.c
1223
err = thread_stack__push_cp(ts, 0, sample->time, ref, cp, false,
tools/perf/util/thread-stack.c
300
u64 timestamp, u64 ref, bool no_return)
tools/perf/util/thread-stack.c
319
cr.call_ref = tse->ref;
tools/perf/util/thread-stack.c
320
cr.return_ref = ref;
tools/perf/util/thread-stack.c
56
u64 ref;
tools/perf/util/thread-stack.c
783
u64 timestamp, u64 ref, struct call_path *cp,
tools/perf/util/thread-stack.c
801
tse->ref = ref;
tools/perf/util/thread-stack.c
815
u64 ret_addr, u64 timestamp, u64 ref,
tools/perf/util/thread-stack.c
828
timestamp, ref, false);
tools/perf/util/thread-stack.c
834
timestamp, ref, false);
tools/perf/util/thread-stack.c
846
timestamp, ref,
tools/perf/util/thread-stack.c
852
timestamp, ref, false);
tools/perf/util/thread-stack.c
862
struct addr_location *to_al, u64 ref)
tools/perf/util/thread-stack.c
882
return thread_stack__push_cp(ts, ip, sample->time, ref, cp,
tools/perf/util/thread-stack.c
887
struct perf_sample *sample, u64 ref)
tools/perf/util/thread-stack.c
895
tm, ref, true);
tools/perf/util/thread-stack.c
907
struct addr_location *to_al, u64 ref)
tools/perf/util/thread-stack.c
922
err = thread_stack__pop_ks(thread, ts, sample, ref);
tools/perf/util/thread-stack.c
929
return thread_stack__push_cp(ts, 0, tm, ref, cp, true,
tools/perf/util/thread-stack.c
934
err = thread_stack__pop_ks(thread, ts, sample, ref);
tools/perf/util/thread-stack.c
952
tm, ref, false);
tools/perf/util/thread-stack.c
960
return thread_stack__push_cp(ts, addr, tm, ref, cp,
tools/perf/util/thread-stack.c
970
err = thread_stack__push_cp(ts, 0, tm, ref, cp, true, false);
tools/perf/util/thread-stack.c
984
err = thread_stack__push_cp(ts, addr, tm, ref, cp, true, false);
tools/perf/util/thread-stack.c
990
err = thread_stack__push_cp(ts, ip, tm, ref, cp, true, false);
tools/perf/util/thread-stack.c
994
return thread_stack__call_return(thread, ts, --ts->cnt, tm, ref, false);
tools/perf/util/thread-stack.c
999
u64 ref)
tools/perf/util/thread-stack.h
107
struct addr_location *to_al, u64 ref,
tools/testing/nvdimm/test/iomap.c
102
percpu_ref_kill(&pgmap->ref);
tools/testing/nvdimm/test/iomap.c
105
percpu_ref_exit(&pgmap->ref);
tools/testing/nvdimm/test/iomap.c
108
static void dev_pagemap_percpu_release(struct percpu_ref *ref)
tools/testing/nvdimm/test/iomap.c
110
struct dev_pagemap *pgmap = container_of(ref, struct dev_pagemap, ref);
tools/testing/nvdimm/test/iomap.c
125
error = percpu_ref_init(&pgmap->ref, dev_pagemap_percpu_release, 0,
tools/testing/selftests/arm64/fp/kernel-test.c
189
ref = malloc(digest_len);
tools/testing/selftests/arm64/fp/kernel-test.c
190
if (!ref) {
tools/testing/selftests/arm64/fp/kernel-test.c
27
static char *ref;
tools/testing/selftests/arm64/fp/kernel-test.c
304
if (!compute_digest(ref)) {
tools/testing/selftests/arm64/fp/kernel-test.c
317
if (memcmp(ref, digest, digest_len) != 0) {
tools/testing/selftests/bpf/prog_tests/map_kptr.c
104
skel->data->ref--;
tools/testing/selftests/bpf/prog_tests/map_kptr.c
111
skel->data->ref--;
tools/testing/selftests/bpf/prog_tests/map_kptr.c
54
skel->data->ref--;
tools/testing/selftests/bpf/prog_tests/map_kptr.c
62
skel->data->ref--;
tools/testing/selftests/bpf/prog_tests/map_kptr.c
69
skel->data->ref--;
tools/testing/selftests/bpf/prog_tests/map_kptr.c
76
skel->data->ref--;
tools/testing/selftests/bpf/prog_tests/map_kptr.c
83
skel->data->ref--;
tools/testing/selftests/bpf/prog_tests/map_kptr.c
90
skel->data->ref--;
tools/testing/selftests/bpf/prog_tests/map_kptr.c
97
skel->data->ref--;
tools/testing/selftests/bpf/prog_tests/uprobe_syscall.c
677
unsigned long *ref;
tools/testing/selftests/bpf/prog_tests/uprobe_syscall.c
684
err = elf_resolve_syms_offsets("/proc/self/exe", 1, (const char **) &sema, &ref, STT_OBJECT);
tools/testing/selftests/bpf/prog_tests/uprobe_syscall.c
688
opts.ref_ctr_offset = *ref;
tools/testing/selftests/bpf/prog_tests/uprobe_syscall.c
709
free(ref);
tools/testing/selftests/bpf/progs/map_kptr.c
303
int ref = 1;
tools/testing/selftests/bpf/progs/map_kptr.c
315
ref++;
tools/testing/selftests/bpf/progs/map_kptr.c
318
if (p_st->cnt.refs.counter != ref) {
tools/testing/selftests/bpf/progs/map_kptr.c
328
if (p_st->cnt.refs.counter != ref)
tools/testing/selftests/bpf/progs/map_kptr.c
335
ref--;
tools/testing/selftests/bpf/progs/map_kptr.c
336
if (p_st->cnt.refs.counter != ref)
tools/testing/selftests/bpf/progs/map_kptr.c
342
ref++;
tools/testing/selftests/bpf/progs/map_kptr.c
348
if (p_st->cnt.refs.counter != ref)
tools/testing/selftests/bpf/progs/map_kptr.c
354
ref--;
tools/testing/selftests/bpf/progs/map_kptr.c
365
if (!p_st || p_st->cnt.refs.counter != ref)
tools/testing/selftests/bpf/progs/map_kptr.c
371
if (p_st->cnt.refs.counter != ref) {
tools/testing/selftests/bpf/progs/map_kptr.c
381
if (p_st->cnt.refs.counter != ref)
tools/testing/selftests/bpf/progs/map_kptr.c
480
ref++;
tools/testing/selftests/bpf/progs/map_kptr.c
481
if (p->cnt.refs.counter != ref) {
tools/testing/selftests/bpf/progs/map_kptr.c
486
ref--;
tools/testing/selftests/bpf/progs/rbtree_search.c
10
struct bpf_refcount ref;
tools/testing/selftests/bpf/progs/refcounted_kptr.c
19
struct bpf_refcount ref;
tools/testing/selftests/bpf/progs/refcounted_kptr.c
571
private(kptr_ref) u64 ref;
tools/testing/selftests/bpf/progs/refcounted_kptr.c
577
bpf_probe_read_kernel(&refcount, sizeof(refcount), (void *) ref);
tools/testing/selftests/bpf/progs/refcounted_kptr.c
600
ref = (u64)(void *) &node_ref->ref;
tools/testing/selftests/bpf/progs/verifier_kfunc_prog_types.c
126
struct bpf_cpumask *alloc, *ref;
tools/testing/selftests/bpf/progs/verifier_kfunc_prog_types.c
132
ref = bpf_cpumask_acquire(alloc);
tools/testing/selftests/bpf/progs/verifier_kfunc_prog_types.c
134
bpf_cpumask_test_cpu(0, (const struct cpumask *)ref);
tools/testing/selftests/bpf/progs/verifier_kfunc_prog_types.c
136
bpf_cpumask_release(ref);
tools/testing/selftests/bpf/progs/verifier_kfunc_prog_types.c
72
struct cgroup *cgrp, *ref;
tools/testing/selftests/bpf/progs/verifier_kfunc_prog_types.c
78
ref = bpf_cgroup_acquire(cgrp);
tools/testing/selftests/bpf/progs/verifier_kfunc_prog_types.c
79
if (!ref) {
tools/testing/selftests/bpf/progs/verifier_kfunc_prog_types.c
84
bpf_cgroup_release(ref);
tools/tracing/rtla/src/common.h
14
int ref;
tools/tracing/rtla/src/osnoise.c
924
context->ref++;
tools/tracing/rtla/src/osnoise.c
976
if (--context->ref < 1)