XFS_INODES_PER_CHUNK
(XFS_INODES_PER_CHUNK / (NBBY * sizeof(uint16_t)))
return ((n >= XFS_INODES_PER_CHUNK ? 0 : XFS_INOBT_MASK(n)) - 1) << i;
(trec.ir_startino + XFS_INODES_PER_CHUNK - 1) <
irec->ir_startino + XFS_INODES_PER_CHUNK - 1))
irec->ir_count > XFS_INODES_PER_CHUNK)
if (irec->ir_freecount > XFS_INODES_PER_CHUNK)
ASSERT(offset < XFS_INODES_PER_CHUNK);
XFS_INODES_PER_CHUNK) == 0);
pagino < (rec->ir_startino + XFS_INODES_PER_CHUNK))
if ((pagino - rec->ir_startino + XFS_INODES_PER_CHUNK - 1) >
XFS_INODES_PER_CHUNK) == 0);
ASSERT(offset < XFS_INODES_PER_CHUNK);
XFS_INODES_PER_CHUNK) == 0);
ASSERT(off >= 0 && off < XFS_INODES_PER_CHUNK);
mp->m_sb.sb_inopblock <= XFS_INODES_PER_CHUNK) {
thisino += XFS_INODES_PER_CHUNK) {
mp->m_sb.sb_inopblock <= XFS_INODES_PER_CHUNK) {
XFS_INODES_PER_CHUNK,
XFS_INODES_PER_CHUNK,
for (i = 0; i < XFS_INODES_PER_CHUNK; i++) {
igeo->ialloc_inos = max_t(uint16_t, XFS_INODES_PER_CHUNK,
if (rec.ir_startino + XFS_INODES_PER_CHUNK > agino) {
if (trec->ir_count + srec->ir_count > XFS_INODES_PER_CHUNK)
ASSERT(newlen <= XFS_INODES_PER_CHUNK);
irec->ir_count = XFS_INODES_PER_CHUNK;
x += XFS_INODES_PER_CHUNK - 1;
return be32_to_cpu(r1->inobt.ir_startino) + XFS_INODES_PER_CHUNK <=
XFS_INODES_PER_CHUNK
XFS_INODES_PER_CHUNK);
align = XFS_INODES_PER_CHUNK * sbp->sb_inodesize
*min = XFS_INODES_PER_CHUNK;
i < XFS_INODES_PER_CHUNK;
if (irec.ir_startino + XFS_INODES_PER_CHUNK <= agino)
nr_inodes = min_t(unsigned int, XFS_INODES_PER_CHUNK,
cluster_base < XFS_INODES_PER_CHUNK;
imask = min_t(unsigned int, XFS_INODES_PER_CHUNK,
iabt->next_startino += XFS_INODES_PER_CHUNK;
if (igeo->inodes_per_cluster <= XFS_INODES_PER_CHUNK)
iabt->next_startino = irec->ir_startino + XFS_INODES_PER_CHUNK;
if (irec.ir_count != XFS_INODES_PER_CHUNK)
XFS_INODES_PER_CHUNK))
if (holecount > XFS_INODES_PER_CHUNK ||
holecount + irec.ir_count != XFS_INODES_PER_CHUNK)
if (frec.ir_startino + XFS_INODES_PER_CHUNK <= agino)
ir_startino = rounddown(ir_startino, XFS_INODES_PER_CHUNK);
ri->rie.ir_startino + XFS_INODES_PER_CHUNK <= ir_startino) {
XFS_INODES_PER_CHUNK);
irec_index += XFS_INODES_PER_CHUNK) {
next_agino = irec.ir_startino + XFS_INODES_PER_CHUNK;
last_agino /= XFS_INODES_PER_CHUNK;
rec.ir_startino + XFS_INODES_PER_CHUNK <= agino)
lastino = rec.ir_startino + XFS_INODES_PER_CHUNK - 1;
if (rec.ir_startino + XFS_INODES_PER_CHUNK <= agino)
*nr_inodesp = XFS_INODES_PER_CHUNK - next;
for (i = 0; i < XFS_INODES_PER_CHUNK; i++) {
for (i = 0; i < XFS_INODES_PER_CHUNK; i++) {
if (ino >= iscan->__batch_ino + XFS_INODES_PER_CHUNK)
struct xfs_inode *__inodes[XFS_INODES_PER_CHUNK];
XFS_INODES_PER_CHUNK);
XFS_INODES_PER_CHUNK / mp->m_sb.sb_inopblock);
#define XFS_INODEGC_MAX_BACKLOG (4 * XFS_INODES_PER_CHUNK)
XFS_INODES_PER_CHUNK;
for (i = 0; i < XFS_INODES_PER_CHUNK; i += igeo->inodes_per_cluster) {
for (j = 0; j < XFS_INODES_PER_CHUNK; j++) {
irec->ir_startino + XFS_INODES_PER_CHUNK - 1);
if (irec->ir_startino + XFS_INODES_PER_CHUNK <= agino)
XFS_INODES_PER_CHUNK);
iwag->lastino = rec_fsino + XFS_INODES_PER_CHUNK - 1;
inodes = round_up(inodes, XFS_INODES_PER_CHUNK);
inobt_records = (inodes * 5) / (4 * XFS_INODES_PER_CHUNK);