Symbol: pcl
arch/arc/kernel/jump_label.c
49
u32 pcl = pc & GENMASK(31, 2);
arch/arc/kernel/jump_label.c
50
u32 u_offset = target - pcl;
arch/s390/kvm/dat.c
270
pgste_init.pcl = uses_skeys && init.h.i;
arch/s390/kvm/dat.c
277
if (!pgste_init.pcl)
arch/s390/kvm/dat.c
912
if (!pgstes[i].pcl)
arch/s390/kvm/dat.h
127
unsigned long pcl : 1;
arch/s390/kvm/dat.h
852
if (old_pgste.pcl)
arch/s390/kvm/dat.h
854
old_pgste.pcl = 1;
arch/s390/kvm/dat.h
870
pgste.pcl = 0;
drivers/firewire/nosy.c
518
dma_free_coherent(&lynx->pci_device->dev, sizeof(struct pcl),
drivers/firewire/nosy.c
520
dma_free_coherent(&lynx->pci_device->dev, sizeof(struct pcl),
drivers/firewire/nosy.c
570
sizeof(struct pcl),
drivers/firewire/nosy.c
574
sizeof(struct pcl),
drivers/firewire/nosy.c
676
dma_free_coherent(&lynx->pci_device->dev, sizeof(struct pcl),
drivers/firewire/nosy.c
680
dma_free_coherent(&lynx->pci_device->dev, sizeof(struct pcl),
drivers/firewire/nosy.c
73
struct pcl *rcv_start_pcl, *rcv_pcl;
drivers/video/fbdev/atafb.c
1093
unsigned long pcl = ULONG_MAX;
drivers/video/fbdev/atafb.c
1097
f25.t * i < pcl) {
drivers/video/fbdev/atafb.c
1098
pcl = f25.t * i;
drivers/video/fbdev/atafb.c
1102
f32.t * i < pcl) {
drivers/video/fbdev/atafb.c
1103
pcl = f32.t * i;
drivers/video/fbdev/atafb.c
1107
fext.t * i < pcl) {
drivers/video/fbdev/atafb.c
1108
pcl = fext.t * i;
drivers/video/fbdev/atafb.c
1114
plen = pcl / pclock->t;
fs/erofs/zdata.c
1050
if (!f->pcl) {
fs/erofs/zdata.c
1054
f->pcl->besteffort |= !ra;
fs/erofs/zdata.c
106
static inline unsigned int z_erofs_pclusterpages(struct z_erofs_pcluster *pcl)
fs/erofs/zdata.c
108
return PAGE_ALIGN(pcl->pageofs_in + pcl->pclustersize) >> PAGE_SHIFT;
fs/erofs/zdata.c
1080
if (f->pcl->length < offset + end - map->m_la) {
fs/erofs/zdata.c
1081
f->pcl->length = offset + end - map->m_la;
fs/erofs/zdata.c
1082
f->pcl->pageofs_out = map->m_la & ~PAGE_MASK;
fs/erofs/zdata.c
1086
f->pcl->length == map->m_llen)
fs/erofs/zdata.c
1087
f->pcl->partial = false;
fs/erofs/zdata.c
1109
struct z_erofs_pcluster *pcl;
fs/erofs/zdata.c
1130
int poff = bvec->offset + be->pcl->pageofs_out;
fs/erofs/zdata.c
1135
bvec->offset + bvec->end == be->pcl->length)) {
fs/erofs/zdata.c
1154
unsigned int off0 = be->pcl->pageofs_out;
fs/erofs/zdata.c
1164
end = min_t(unsigned int, be->pcl->length - bvi->bvec.offset,
fs/erofs/zdata.c
1195
struct z_erofs_pcluster *pcl = be->pcl;
fs/erofs/zdata.c
1200
z_erofs_bvec_iter_begin(&biter, &pcl->bvset, Z_EROFS_INLINE_BVECS, 0);
fs/erofs/zdata.c
1201
for (i = 0; i < pcl->vcnt; ++i) {
fs/erofs/zdata.c
1220
struct z_erofs_pcluster *pcl = be->pcl;
fs/erofs/zdata.c
1221
unsigned int pclusterpages = z_erofs_pclusterpages(pcl);
fs/erofs/zdata.c
1226
struct z_erofs_bvec *bvec = &pcl->compressed_bvecs[i];
fs/erofs/zdata.c
1237
if (pcl->from_meta ||
fs/erofs/zdata.c
1256
struct z_erofs_pcluster *pcl = be->pcl;
fs/erofs/zdata.c
1257
unsigned int pclusterpages = z_erofs_pclusterpages(pcl);
fs/erofs/zdata.c
1259
z_erofs_decomp[pcl->algorithmformat];
fs/erofs/zdata.c
1266
mutex_lock(&pcl->lock);
fs/erofs/zdata.c
1267
be->nr_pages = PAGE_ALIGN(pcl->length + pcl->pageofs_out) >> PAGE_SHIFT;
fs/erofs/zdata.c
1303
.pageofs_in = pcl->pageofs_in,
fs/erofs/zdata.c
1304
.pageofs_out = pcl->pageofs_out,
fs/erofs/zdata.c
1305
.inputsize = pcl->pclustersize,
fs/erofs/zdata.c
1306
.outputsize = pcl->length,
fs/erofs/zdata.c
1307
.alg = pcl->algorithmformat,
fs/erofs/zdata.c
1309
.partial_decoding = pcl->partial,
fs/erofs/zdata.c
1311
.gfp = pcl->besteffort ? GFP_KERNEL :
fs/erofs/zdata.c
1315
if (pcl->besteffort || reason != ERR_PTR(-ENOMEM))
fs/erofs/zdata.c
1317
alg->name, reason, pcl->pos,
fs/erofs/zdata.c
1318
pcl->pclustersize, pcl->length);
fs/erofs/zdata.c
1322
alg->name, reason, pcl->pos,
fs/erofs/zdata.c
1323
pcl->pclustersize, pcl->length);
fs/erofs/zdata.c
1329
if (pcl->from_meta) {
fs/erofs/zdata.c
1330
folio_put(page_folio(pcl->compressed_bvecs[0].page));
fs/erofs/zdata.c
1331
WRITE_ONCE(pcl->compressed_bvecs[0].page, NULL);
fs/erofs/zdata.c
1343
WRITE_ONCE(pcl->compressed_bvecs[i].page, NULL);
fs/erofs/zdata.c
1362
if (pcl->algorithmformat != Z_EROFS_COMPRESSION_LZ4) {
fs/erofs/zdata.c
1377
pcl->length = 0;
fs/erofs/zdata.c
1378
pcl->partial = true;
fs/erofs/zdata.c
1379
pcl->besteffort = false;
fs/erofs/zdata.c
1380
pcl->bvset.nextpage = NULL;
fs/erofs/zdata.c
1381
pcl->vcnt = 0;
fs/erofs/zdata.c
1384
WRITE_ONCE(pcl->next, NULL);
fs/erofs/zdata.c
1385
mutex_unlock(&pcl->lock);
fs/erofs/zdata.c
1387
if (pcl->from_meta)
fs/erofs/zdata.c
1388
z_erofs_free_pcluster(pcl);
fs/erofs/zdata.c
1390
z_erofs_put_pcluster(sbi, pcl, try_free);
fs/erofs/zdata.c
1402
.pcl = io->head,
fs/erofs/zdata.c
1407
for (; be.pcl != Z_EROFS_PCLUSTER_TAIL; be.pcl = next) {
fs/erofs/zdata.c
1408
DBG_BUGON(!be.pcl);
fs/erofs/zdata.c
1409
next = READ_ONCE(be.pcl->next);
fs/erofs/zdata.c
1488
struct z_erofs_pcluster *pcl,
fs/erofs/zdata.c
1504
spin_lock(&pcl->lockref.lock);
fs/erofs/zdata.c
1505
zbv = pcl->compressed_bvecs[nr];
fs/erofs/zdata.c
1506
spin_unlock(&pcl->lockref.lock);
fs/erofs/zdata.c
1539
folio_attach_private(folio, pcl);
fs/erofs/zdata.c
1543
if (likely(folio->private == pcl)) {
fs/erofs/zdata.c
1564
spin_lock(&pcl->lockref.lock);
fs/erofs/zdata.c
1565
if (unlikely(pcl->compressed_bvecs[nr].page != zbv.page)) {
fs/erofs/zdata.c
1568
spin_unlock(&pcl->lockref.lock);
fs/erofs/zdata.c
1572
pcl->compressed_bvecs[nr].page = page ? page : ERR_PTR(-ENOMEM);
fs/erofs/zdata.c
1573
spin_unlock(&pcl->lockref.lock);
fs/erofs/zdata.c
1580
filemap_add_folio(mc, folio, (pcl->pos >> PAGE_SHIFT) + nr, gfp)) {
fs/erofs/zdata.c
1585
folio_attach_private(folio, pcl);
fs/erofs/zdata.c
1627
static void z_erofs_move_to_bypass_queue(struct z_erofs_pcluster *pcl,
fs/erofs/zdata.c
1631
WRITE_ONCE(pcl->next, Z_EROFS_PCLUSTER_TAIL);
fs/erofs/zdata.c
1633
WRITE_ONCE(*qtail[JQ_BYPASS], pcl);
fs/erofs/zdata.c
1634
qtail[JQ_BYPASS] = &pcl->next;
fs/erofs/zdata.c
1670
struct z_erofs_pcluster *pcl, *next;
fs/erofs/zdata.c
1695
pcl = next;
fs/erofs/zdata.c
1696
next = READ_ONCE(pcl->next);
fs/erofs/zdata.c
1697
if (pcl->from_meta) {
fs/erofs/zdata.c
1698
z_erofs_move_to_bypass_queue(pcl, next, qtail);
fs/erofs/zdata.c
1704
.m_pa = round_down(pcl->pos, sb->s_blocksize),
fs/erofs/zdata.c
1709
end = round_up(cur + pcl->pageofs_in + pcl->pclustersize,
fs/erofs/zdata.c
1731
z_erofs_fill_bio_vec(&bvec, f, pcl, i++, mc);
fs/erofs/zdata.c
1770
qtail[JQ_SUBMIT] = &pcl->next;
fs/erofs/zdata.c
1772
z_erofs_move_to_bypass_queue(pcl, next, qtail);
fs/erofs/zdata.c
258
struct z_erofs_pcluster *pcl;
fs/erofs/zdata.c
263
pcl = kmem_cache_zalloc(pcs->slab, GFP_KERNEL);
fs/erofs/zdata.c
264
if (!pcl)
fs/erofs/zdata.c
266
return pcl;
fs/erofs/zdata.c
271
static void z_erofs_free_pcluster(struct z_erofs_pcluster *pcl)
fs/erofs/zdata.c
273
unsigned int pclusterpages = z_erofs_pclusterpages(pcl);
fs/erofs/zdata.c
282
kmem_cache_free(pcs->slab, pcl);
fs/erofs/zdata.c
503
struct z_erofs_pcluster *pcl, *head;
fs/erofs/zdata.c
536
struct z_erofs_pcluster *pcl = fe->pcl;
fs/erofs/zdata.c
537
unsigned int pclusterpages = z_erofs_pclusterpages(pcl);
fs/erofs/zdata.c
539
pgoff_t poff = pcl->pos >> PAGE_SHIFT;
fs/erofs/zdata.c
553
if (READ_ONCE(pcl->compressed_bvecs[i].page))
fs/erofs/zdata.c
572
spin_lock(&pcl->lockref.lock);
fs/erofs/zdata.c
573
if (!pcl->compressed_bvecs[i].page) {
fs/erofs/zdata.c
574
pcl->compressed_bvecs[i].page =
fs/erofs/zdata.c
576
spin_unlock(&pcl->lockref.lock);
fs/erofs/zdata.c
579
spin_unlock(&pcl->lockref.lock);
fs/erofs/zdata.c
593
struct z_erofs_pcluster *pcl)
fs/erofs/zdata.c
595
unsigned int pclusterpages = z_erofs_pclusterpages(pcl);
fs/erofs/zdata.c
599
DBG_BUGON(pcl->from_meta);
fs/erofs/zdata.c
602
if (pcl->compressed_bvecs[i].page) {
fs/erofs/zdata.c
603
folio = page_folio(pcl->compressed_bvecs[i].page);
fs/erofs/zdata.c
610
pcl->compressed_bvecs[i].page = NULL;
fs/erofs/zdata.c
620
struct z_erofs_pcluster *pcl = folio_get_private(folio);
fs/erofs/zdata.c
621
struct z_erofs_bvec *bvec = pcl->compressed_bvecs;
fs/erofs/zdata.c
622
struct z_erofs_bvec *end = bvec + z_erofs_pclusterpages(pcl);
fs/erofs/zdata.c
629
spin_lock(&pcl->lockref.lock);
fs/erofs/zdata.c
630
if (pcl->lockref.count <= 0) {
fs/erofs/zdata.c
631
DBG_BUGON(pcl->from_meta);
fs/erofs/zdata.c
641
spin_unlock(&pcl->lockref.lock);
fs/erofs/zdata.c
693
struct z_erofs_pcluster *pcl = fe->pcl;
fs/erofs/zdata.c
698
if (pcl->algorithmformat < Z_EROFS_COMPRESSION_MAX ||
fs/erofs/zdata.c
701
spin_lock(&pcl->lockref.lock);
fs/erofs/zdata.c
703
if (pcl->compressed_bvecs[--fe->icur].page)
fs/erofs/zdata.c
705
pcl->compressed_bvecs[fe->icur] = *bvec;
fs/erofs/zdata.c
706
spin_unlock(&pcl->lockref.lock);
fs/erofs/zdata.c
709
spin_unlock(&pcl->lockref.lock);
fs/erofs/zdata.c
719
fe->pcl->vcnt += (ret >= 0);
fs/erofs/zdata.c
723
static bool z_erofs_get_pcluster(struct z_erofs_pcluster *pcl)
fs/erofs/zdata.c
725
if (lockref_get_not_zero(&pcl->lockref))
fs/erofs/zdata.c
728
spin_lock(&pcl->lockref.lock);
fs/erofs/zdata.c
729
if (__lockref_is_dead(&pcl->lockref)) {
fs/erofs/zdata.c
730
spin_unlock(&pcl->lockref.lock);
fs/erofs/zdata.c
734
if (!pcl->lockref.count++)
fs/erofs/zdata.c
736
spin_unlock(&pcl->lockref.lock);
fs/erofs/zdata.c
745
struct z_erofs_pcluster *pcl, *pre;
fs/erofs/zdata.c
750
pcl = z_erofs_alloc_pcluster(pageofs_in + map->m_plen);
fs/erofs/zdata.c
751
if (IS_ERR(pcl))
fs/erofs/zdata.c
752
return PTR_ERR(pcl);
fs/erofs/zdata.c
754
lockref_init(&pcl->lockref); /* one ref for this request */
fs/erofs/zdata.c
755
pcl->algorithmformat = map->m_algorithmformat;
fs/erofs/zdata.c
756
pcl->pclustersize = map->m_plen;
fs/erofs/zdata.c
757
pcl->length = 0;
fs/erofs/zdata.c
758
pcl->partial = true;
fs/erofs/zdata.c
759
pcl->next = fe->head;
fs/erofs/zdata.c
760
pcl->pos = map->m_pa;
fs/erofs/zdata.c
761
pcl->pageofs_in = pageofs_in;
fs/erofs/zdata.c
762
pcl->pageofs_out = map->m_la & ~PAGE_MASK;
fs/erofs/zdata.c
763
pcl->from_meta = map->m_flags & EROFS_MAP_META;
fs/erofs/zdata.c
770
mutex_init(&pcl->lock);
fs/erofs/zdata.c
771
DBG_BUGON(!mutex_trylock(&pcl->lock));
fs/erofs/zdata.c
773
if (!pcl->from_meta) {
fs/erofs/zdata.c
776
pre = __xa_cmpxchg(&sbi->managed_pslots, pcl->pos,
fs/erofs/zdata.c
777
NULL, pcl, GFP_KERNEL);
fs/erofs/zdata.c
790
fe->pcl = pre;
fs/erofs/zdata.c
795
fe->head = fe->pcl = pcl;
fs/erofs/zdata.c
799
mutex_unlock(&pcl->lock);
fs/erofs/zdata.c
800
z_erofs_free_pcluster(pcl);
fs/erofs/zdata.c
808
struct z_erofs_pcluster *pcl = NULL;
fs/erofs/zdata.c
812
DBG_BUGON(fe->pcl);
fs/erofs/zdata.c
831
pcl = xa_load(&EROFS_SB(sb)->managed_pslots, map->m_pa);
fs/erofs/zdata.c
832
if (!pcl || z_erofs_get_pcluster(pcl)) {
fs/erofs/zdata.c
833
DBG_BUGON(pcl && map->m_pa != pcl->pos);
fs/erofs/zdata.c
841
if (pcl) {
fs/erofs/zdata.c
842
fe->pcl = pcl;
fs/erofs/zdata.c
849
mutex_lock(&fe->pcl->lock);
fs/erofs/zdata.c
851
if (!cmpxchg(&fe->pcl->next, NULL, fe->head)) {
fs/erofs/zdata.c
853
fe->head = fe->pcl;
fs/erofs/zdata.c
862
z_erofs_bvec_iter_begin(&fe->biter, &fe->pcl->bvset,
fs/erofs/zdata.c
863
Z_EROFS_INLINE_BVECS, fe->pcl->vcnt);
fs/erofs/zdata.c
864
if (!fe->pcl->from_meta) {
fs/erofs/zdata.c
869
WRITE_ONCE(fe->pcl->compressed_bvecs[0].page, ptr);
fs/erofs/zdata.c
870
fe->pcl->pageofs_in = map->m_pa & ~PAGE_MASK;
fs/erofs/zdata.c
874
fe->icur = z_erofs_pclusterpages(fe->pcl);
fs/erofs/zdata.c
884
struct z_erofs_pcluster *pcl)
fs/erofs/zdata.c
886
if (pcl->lockref.count)
fs/erofs/zdata.c
894
if (erofs_try_to_free_all_cached_folios(sbi, pcl))
fs/erofs/zdata.c
901
DBG_BUGON(__xa_erase(&sbi->managed_pslots, pcl->pos) != pcl);
fs/erofs/zdata.c
903
lockref_mark_dead(&pcl->lockref);
fs/erofs/zdata.c
908
struct z_erofs_pcluster *pcl)
fs/erofs/zdata.c
912
spin_lock(&pcl->lockref.lock);
fs/erofs/zdata.c
913
free = __erofs_try_to_release_pcluster(sbi, pcl);
fs/erofs/zdata.c
914
spin_unlock(&pcl->lockref.lock);
fs/erofs/zdata.c
917
call_rcu(&pcl->rcu, z_erofs_rcu_callback);
fs/erofs/zdata.c
924
struct z_erofs_pcluster *pcl;
fs/erofs/zdata.c
928
xa_for_each(&sbi->managed_pslots, index, pcl) {
fs/erofs/zdata.c
930
if (!erofs_try_to_release_pcluster(sbi, pcl))
fs/erofs/zdata.c
944
struct z_erofs_pcluster *pcl, bool try_free)
fs/erofs/zdata.c
948
if (lockref_put_or_lock(&pcl->lockref))
fs/erofs/zdata.c
951
DBG_BUGON(__lockref_is_dead(&pcl->lockref));
fs/erofs/zdata.c
952
if (!--pcl->lockref.count) {
fs/erofs/zdata.c
954
free = __erofs_try_to_release_pcluster(sbi, pcl);
fs/erofs/zdata.c
959
spin_unlock(&pcl->lockref.lock);
fs/erofs/zdata.c
961
call_rcu(&pcl->rcu, z_erofs_rcu_callback);
fs/erofs/zdata.c
966
struct z_erofs_pcluster *pcl = fe->pcl;
fs/erofs/zdata.c
968
if (!pcl)
fs/erofs/zdata.c
972
mutex_unlock(&pcl->lock);
fs/erofs/zdata.c
979
z_erofs_put_pcluster(EROFS_I_SB(fe->inode), pcl, false);
fs/erofs/zdata.c
980
fe->pcl = NULL;
include/linux/sunrpc/svc_rdma_pcl.h
117
extern void pcl_free(struct svc_rdma_pcl *pcl);
include/linux/sunrpc/svc_rdma_pcl.h
121
struct svc_rdma_pcl *pcl, __be32 *p);
include/linux/sunrpc/svc_rdma_pcl.h
122
extern int pcl_process_nonpayloads(const struct svc_rdma_pcl *pcl,
include/linux/sunrpc/svc_rdma_pcl.h
38
static inline void pcl_init(struct svc_rdma_pcl *pcl)
include/linux/sunrpc/svc_rdma_pcl.h
40
INIT_LIST_HEAD(&pcl->cl_chunks);
include/linux/sunrpc/svc_rdma_pcl.h
48
static inline bool pcl_is_empty(const struct svc_rdma_pcl *pcl)
include/linux/sunrpc/svc_rdma_pcl.h
50
return list_empty(&pcl->cl_chunks);
include/linux/sunrpc/svc_rdma_pcl.h
60
pcl_first_chunk(const struct svc_rdma_pcl *pcl)
include/linux/sunrpc/svc_rdma_pcl.h
62
if (pcl_is_empty(pcl))
include/linux/sunrpc/svc_rdma_pcl.h
64
return list_first_entry(&pcl->cl_chunks, struct svc_rdma_chunk,
include/linux/sunrpc/svc_rdma_pcl.h
76
pcl_next_chunk(const struct svc_rdma_pcl *pcl, struct svc_rdma_chunk *chunk)
include/linux/sunrpc/svc_rdma_pcl.h
78
if (list_is_last(&chunk->ch_list, &pcl->cl_chunks))
include/linux/sunrpc/svc_rdma_pcl.h
88
#define pcl_for_each_chunk(pos, pcl) \
include/linux/sunrpc/svc_rdma_pcl.h
89
for (pos = list_first_entry(&(pcl)->cl_chunks, struct svc_rdma_chunk, ch_list); \
include/linux/sunrpc/svc_rdma_pcl.h
90
&pos->ch_list != &(pcl)->cl_chunks; \
net/sunrpc/xprtrdma/svc_rdma_pcl.c
104
struct svc_rdma_pcl *pcl = &rctxt->rc_call_pcl;
net/sunrpc/xprtrdma/svc_rdma_pcl.c
105
unsigned int i, segcount = pcl->cl_count;
net/sunrpc/xprtrdma/svc_rdma_pcl.c
107
pcl->cl_count = 0;
net/sunrpc/xprtrdma/svc_rdma_pcl.c
119
if (pcl_is_empty(pcl)) {
net/sunrpc/xprtrdma/svc_rdma_pcl.c
123
pcl_insert_position(pcl, chunk);
net/sunrpc/xprtrdma/svc_rdma_pcl.c
125
chunk = list_first_entry(&pcl->cl_chunks,
net/sunrpc/xprtrdma/svc_rdma_pcl.c
158
struct svc_rdma_pcl *pcl = &rctxt->rc_read_pcl;
net/sunrpc/xprtrdma/svc_rdma_pcl.c
159
unsigned int i, segcount = pcl->cl_count;
net/sunrpc/xprtrdma/svc_rdma_pcl.c
161
pcl->cl_count = 0;
net/sunrpc/xprtrdma/svc_rdma_pcl.c
17
void pcl_free(struct svc_rdma_pcl *pcl)
net/sunrpc/xprtrdma/svc_rdma_pcl.c
173
chunk = pcl_lookup_position(pcl, position);
net/sunrpc/xprtrdma/svc_rdma_pcl.c
178
pcl_insert_position(pcl, chunk);
net/sunrpc/xprtrdma/svc_rdma_pcl.c
19
while (!list_empty(&pcl->cl_chunks)) {
net/sunrpc/xprtrdma/svc_rdma_pcl.c
202
struct svc_rdma_pcl *pcl, __be32 *p)
net/sunrpc/xprtrdma/svc_rdma_pcl.c
209
for (i = 0; i < pcl->cl_count; i++) {
net/sunrpc/xprtrdma/svc_rdma_pcl.c
216
list_add_tail(&chunk->ch_list, &pcl->cl_chunks);
net/sunrpc/xprtrdma/svc_rdma_pcl.c
22
chunk = pcl_first_chunk(pcl);
net/sunrpc/xprtrdma/svc_rdma_pcl.c
265
int pcl_process_nonpayloads(const struct svc_rdma_pcl *pcl,
net/sunrpc/xprtrdma/svc_rdma_pcl.c
274
chunk = pcl_first_chunk(pcl);
net/sunrpc/xprtrdma/svc_rdma_pcl.c
286
while ((next = pcl_next_chunk(pcl, chunk))) {
net/sunrpc/xprtrdma/svc_rdma_pcl.c
44
pcl_lookup_position(struct svc_rdma_pcl *pcl, u32 position)
net/sunrpc/xprtrdma/svc_rdma_pcl.c
48
pcl_for_each_chunk(pos, pcl) {
net/sunrpc/xprtrdma/svc_rdma_pcl.c
55
static void pcl_insert_position(struct svc_rdma_pcl *pcl,
net/sunrpc/xprtrdma/svc_rdma_pcl.c
60
pcl_for_each_chunk(pos, pcl) {
net/sunrpc/xprtrdma/svc_rdma_pcl.c
65
pcl->cl_count++;
net/sunrpc/xprtrdma/svc_rdma_rw.c
1027
const struct svc_rdma_pcl *pcl = &head->rc_read_pcl;
net/sunrpc/xprtrdma/svc_rdma_rw.c
1032
if (pcl_is_empty(pcl))
net/sunrpc/xprtrdma/svc_rdma_rw.c
1036
chunk = pcl_first_chunk(pcl);
net/sunrpc/xprtrdma/svc_rdma_rw.c
1043
pcl_for_each_chunk(chunk, pcl) {
net/sunrpc/xprtrdma/svc_rdma_rw.c
1048
next = pcl_next_chunk(pcl, chunk);
net/sunrpc/xprtrdma/svc_rdma_rw.c
905
const struct svc_rdma_pcl *pcl = &head->rc_read_pcl;
net/sunrpc/xprtrdma/svc_rdma_rw.c
911
chunk = pcl_first_chunk(pcl);
net/sunrpc/xprtrdma/svc_rdma_rw.c
917
pcl_for_each_chunk(chunk, pcl) {
net/sunrpc/xprtrdma/svc_rdma_rw.c
922
next = pcl_next_chunk(pcl, chunk);