NAT_ENTRY_PER_BLOCK
blkno * NAT_ENTRY_PER_BLOCK);
f2fs_bitmap_size(NAT_ENTRY_PER_BLOCK);
i = start_nid % NAT_ENTRY_PER_BLOCK;
for (; i < NAT_ENTRY_PER_BLOCK; i++, start_nid++) {
for (idx = 0; idx < NAT_ENTRY_PER_BLOCK; idx++) {
NAT_ENTRY_PER_BLOCK, idx);
if (idx >= NAT_ENTRY_PER_BLOCK)
nid = i * NAT_ENTRY_PER_BLOCK + idx;
if (unlikely(nid % NAT_ENTRY_PER_BLOCK))
nid = NAT_BLOCK_OFFSET(nid) * NAT_ENTRY_PER_BLOCK;
if (nm_i->nid_cnt[FREE_NID] >= NAT_ENTRY_PER_BLOCK)
if (nm_i->nid_cnt[FREE_NID] >= NAT_ENTRY_PER_BLOCK)
nid += (NAT_ENTRY_PER_BLOCK - (nid % NAT_ENTRY_PER_BLOCK));
unsigned int nat_index = start_nid / NAT_ENTRY_PER_BLOCK;
for (; i < NAT_ENTRY_PER_BLOCK; i++) {
if (valid == NAT_ENTRY_PER_BLOCK)
nid_t start_nid = set->set * NAT_ENTRY_PER_BLOCK;
nid = i * NAT_ENTRY_PER_BLOCK;
last_nid = nid + NAT_ENTRY_PER_BLOCK;
nm_i->max_nid = NAT_ENTRY_PER_BLOCK * nm_i->nat_blocks;
f2fs_bitmap_size(NAT_ENTRY_PER_BLOCK), GFP_KERNEL);
#define NAT_BLOCK_OFFSET(start_nid) ((start_nid) / NAT_ENTRY_PER_BLOCK)
#define MAX_FREE_NIDS (NAT_ENTRY_PER_BLOCK * FREE_NID_PAGES)
#define START_NID(nid) (((nid) / NAT_ENTRY_PER_BLOCK) * NAT_ENTRY_PER_BLOCK)
f2fs_try_to_free_nats(sbi, NAT_ENTRY_PER_BLOCK);
NAT_ENTRY_PER_BLOCK));
struct f2fs_nat_entry entries[NAT_ENTRY_PER_BLOCK];