Symbol: XFS_INODES_PER_CHUNK
fs/xfs/libxfs/xfs_format.h
1583
(XFS_INODES_PER_CHUNK / (NBBY * sizeof(uint16_t)))
fs/xfs/libxfs/xfs_format.h
1587
return ((n >= XFS_INODES_PER_CHUNK ? 0 : XFS_INOBT_MASK(n)) - 1) << i;
fs/xfs/libxfs/xfs_ialloc.c
1206
(trec.ir_startino + XFS_INODES_PER_CHUNK - 1) <
fs/xfs/libxfs/xfs_ialloc.c
121
irec->ir_startino + XFS_INODES_PER_CHUNK - 1))
fs/xfs/libxfs/xfs_ialloc.c
124
irec->ir_count > XFS_INODES_PER_CHUNK)
fs/xfs/libxfs/xfs_ialloc.c
126
if (irec->ir_freecount > XFS_INODES_PER_CHUNK)
fs/xfs/libxfs/xfs_ialloc.c
1334
ASSERT(offset < XFS_INODES_PER_CHUNK);
fs/xfs/libxfs/xfs_ialloc.c
1336
XFS_INODES_PER_CHUNK) == 0);
fs/xfs/libxfs/xfs_ialloc.c
1404
pagino < (rec->ir_startino + XFS_INODES_PER_CHUNK))
fs/xfs/libxfs/xfs_ialloc.c
1436
if ((pagino - rec->ir_startino + XFS_INODES_PER_CHUNK - 1) >
fs/xfs/libxfs/xfs_ialloc.c
1543
XFS_INODES_PER_CHUNK) == 0);
fs/xfs/libxfs/xfs_ialloc.c
1615
ASSERT(offset < XFS_INODES_PER_CHUNK);
fs/xfs/libxfs/xfs_ialloc.c
1617
XFS_INODES_PER_CHUNK) == 0);
fs/xfs/libxfs/xfs_ialloc.c
2131
ASSERT(off >= 0 && off < XFS_INODES_PER_CHUNK);
fs/xfs/libxfs/xfs_ialloc.c
2145
mp->m_sb.sb_inopblock <= XFS_INODES_PER_CHUNK) {
fs/xfs/libxfs/xfs_ialloc.c
223
thisino += XFS_INODES_PER_CHUNK) {
fs/xfs/libxfs/xfs_ialloc.c
2291
mp->m_sb.sb_inopblock <= XFS_INODES_PER_CHUNK) {
fs/xfs/libxfs/xfs_ialloc.c
232
XFS_INODES_PER_CHUNK,
fs/xfs/libxfs/xfs_ialloc.c
233
XFS_INODES_PER_CHUNK,
fs/xfs/libxfs/xfs_ialloc.c
2867
for (i = 0; i < XFS_INODES_PER_CHUNK; i++) {
fs/xfs/libxfs/xfs_ialloc.c
2999
igeo->ialloc_inos = max_t(uint16_t, XFS_INODES_PER_CHUNK,
fs/xfs/libxfs/xfs_ialloc.c
3194
if (rec.ir_startino + XFS_INODES_PER_CHUNK > agino) {
fs/xfs/libxfs/xfs_ialloc.c
501
if (trec->ir_count + srec->ir_count > XFS_INODES_PER_CHUNK)
fs/xfs/libxfs/xfs_ialloc.c
869
ASSERT(newlen <= XFS_INODES_PER_CHUNK);
fs/xfs/libxfs/xfs_ialloc.c
92
irec->ir_count = XFS_INODES_PER_CHUNK;
fs/xfs/libxfs/xfs_ialloc_btree.c
218
x += XFS_INODES_PER_CHUNK - 1;
fs/xfs/libxfs/xfs_ialloc_btree.c
394
return be32_to_cpu(r1->inobt.ir_startino) + XFS_INODES_PER_CHUNK <=
fs/xfs/libxfs/xfs_ialloc_btree.c
592
XFS_INODES_PER_CHUNK
fs/xfs/libxfs/xfs_ialloc_btree.c
736
XFS_INODES_PER_CHUNK);
fs/xfs/libxfs/xfs_sb.c
542
align = XFS_INODES_PER_CHUNK * sbp->sb_inodesize
fs/xfs/libxfs/xfs_types.c
203
*min = XFS_INODES_PER_CHUNK;
fs/xfs/scrub/agheader_repair.c
1256
i < XFS_INODES_PER_CHUNK;
fs/xfs/scrub/ialloc.c
194
if (irec.ir_startino + XFS_INODES_PER_CHUNK <= agino)
fs/xfs/scrub/ialloc.c
376
nr_inodes = min_t(unsigned int, XFS_INODES_PER_CHUNK,
fs/xfs/scrub/ialloc.c
478
cluster_base < XFS_INODES_PER_CHUNK;
fs/xfs/scrub/ialloc.c
518
imask = min_t(unsigned int, XFS_INODES_PER_CHUNK,
fs/xfs/scrub/ialloc.c
536
iabt->next_startino += XFS_INODES_PER_CHUNK;
fs/xfs/scrub/ialloc.c
557
if (igeo->inodes_per_cluster <= XFS_INODES_PER_CHUNK)
fs/xfs/scrub/ialloc.c
565
iabt->next_startino = irec->ir_startino + XFS_INODES_PER_CHUNK;
fs/xfs/scrub/ialloc.c
601
if (irec.ir_count != XFS_INODES_PER_CHUNK)
fs/xfs/scrub/ialloc.c
605
XFS_INODES_PER_CHUNK))
fs/xfs/scrub/ialloc.c
628
if (holecount > XFS_INODES_PER_CHUNK ||
fs/xfs/scrub/ialloc.c
629
holecount + irec.ir_count != XFS_INODES_PER_CHUNK)
fs/xfs/scrub/ialloc.c
91
if (frec.ir_startino + XFS_INODES_PER_CHUNK <= agino)
fs/xfs/scrub/ialloc_repair.c
233
ir_startino = rounddown(ir_startino, XFS_INODES_PER_CHUNK);
fs/xfs/scrub/ialloc_repair.c
241
ri->rie.ir_startino + XFS_INODES_PER_CHUNK <= ir_startino) {
fs/xfs/scrub/ialloc_repair.c
301
XFS_INODES_PER_CHUNK);
fs/xfs/scrub/ialloc_repair.c
322
irec_index += XFS_INODES_PER_CHUNK) {
fs/xfs/scrub/ialloc_repair.c
619
next_agino = irec.ir_startino + XFS_INODES_PER_CHUNK;
fs/xfs/scrub/ialloc_repair.c
817
last_agino /= XFS_INODES_PER_CHUNK;
fs/xfs/scrub/inode.c
687
rec.ir_startino + XFS_INODES_PER_CHUNK <= agino)
fs/xfs/scrub/iscan.c
147
lastino = rec.ir_startino + XFS_INODES_PER_CHUNK - 1;
fs/xfs/scrub/iscan.c
153
if (rec.ir_startino + XFS_INODES_PER_CHUNK <= agino)
fs/xfs/scrub/iscan.c
180
*nr_inodesp = XFS_INODES_PER_CHUNK - next;
fs/xfs/scrub/iscan.c
625
for (i = 0; i < XFS_INODES_PER_CHUNK; i++) {
fs/xfs/scrub/iscan.c
652
for (i = 0; i < XFS_INODES_PER_CHUNK; i++) {
fs/xfs/scrub/iscan.c
750
if (ino >= iscan->__batch_ino + XFS_INODES_PER_CHUNK)
fs/xfs/scrub/iscan.h
56
struct xfs_inode *__inodes[XFS_INODES_PER_CHUNK];
fs/xfs/scrub/repair.c
357
XFS_INODES_PER_CHUNK);
fs/xfs/scrub/rmap_repair.c
700
XFS_INODES_PER_CHUNK / mp->m_sb.sb_inopblock);
fs/xfs/xfs_icache.c
2138
#define XFS_INODEGC_MAX_BACKLOG (4 * XFS_INODES_PER_CHUNK)
fs/xfs/xfs_itable.c
429
XFS_INODES_PER_CHUNK;
fs/xfs/xfs_iwalk.c
110
for (i = 0; i < XFS_INODES_PER_CHUNK; i += igeo->inodes_per_cluster) {
fs/xfs/xfs_iwalk.c
200
for (j = 0; j < XFS_INODES_PER_CHUNK; j++) {
fs/xfs/xfs_iwalk.c
308
irec->ir_startino + XFS_INODES_PER_CHUNK - 1);
fs/xfs/xfs_iwalk.c
314
if (irec->ir_startino + XFS_INODES_PER_CHUNK <= agino)
fs/xfs/xfs_iwalk.c
363
XFS_INODES_PER_CHUNK);
fs/xfs/xfs_iwalk.c
431
iwag->lastino = rec_fsino + XFS_INODES_PER_CHUNK - 1;
fs/xfs/xfs_iwalk.c
509
inodes = round_up(inodes, XFS_INODES_PER_CHUNK);
fs/xfs/xfs_iwalk.c
524
inobt_records = (inodes * 5) / (4 * XFS_INODES_PER_CHUNK);