sector_div
tmp = sector_div(tmp, granularity);
unsigned int alignment = sector_div(sector, granularity >> SECTOR_SHIFT)
offset = sector_div(sector, granularity);
sector_div(capacity, 255*63);
if (sector_div(tmp, granularity) != alignment) {
tmp = start + granularity - sector_div(tmp, granularity);
sector_div(capacity, (geo->heads * geo->sectors));
sector_div(cylinders, geo->heads * geo->sectors);
sector_div(cylinders, hg->heads * hg->sectors);
sector_div(sector_off, sector_size + 8);
sector_div(s, c->block_size >> SECTOR_SHIFT);
(void) sector_div(r, mq->cache_blocks_per_hotspot_block);
(void) sector_div(origin_size, discard_block_size);
sector_div(e, li->cache->discard_block_size);
(void) sector_div(size, cache->sectors_per_block);
sector_div(bi_sector, cache->sectors_per_block);
(void) sector_div(block_nr, cache->sectors_per_block);
sector_div(thisblock, dd->sect_per_block);
sector_div(size, dd->sect_per_block);
sector_div(size, dd->sect_per_block);
(void) sector_div(block_nr, era->sectors_per_block);
if (sector_div(dev_sectors, data_stripes))
if (sector_div(array_sectors, rs->raid10_copies))
} else if (sector_div(dev_sectors, data_stripes))
if (sector_div(next_free, stride) == NUM_SNAPSHOT_HDR_CHUNKS)
if (sector_div(width, stripes)) {
if (sector_div(tmp_len, chunk_size)) {
chunk_offset = sector_div(chunk, sc->chunk_size);
*stripe = sector_div(chunk, sc->stripes);
*result -= sector_div(sector, sc->chunk_size);
if (sector_div(nr_slots, sctx->region_entries_per_slot))
sector_div(p, sctx->region_size);
if (sector_div(nr_regions, sctx->region_size))
return !sector_div(block_size, n);
sector_div(metadata_dev_size, THIN_METADATA_BLOCK_SIZE);
(void) sector_div(data_size, pool->sectors_per_block);
(void) sector_div(blocks, pool->sectors_per_block);
(void) sector_div(block_nr, pool->sectors_per_block);
(void) sector_div(b, pool->sectors_per_block);
(void) sector_div(e, pool->sectors_per_block);
sector_div(bi_sector, pool->sectors_per_block);
sector_div(tmp_sector, uc->chunk_size);
if (sector_div(tmp_len, uc->chunk_size)) {
if (sector_div(f->rounds, f->rsn))
sector_div(sectors, mddev->chunk_sectors);
if (sector_div(temp, chunk))
sector_div(res, (u32)((max_sectors>>scale)+1));
sector_div(first_sector, mddev->chunk_sectors);
zone->disk_shift = sector_div(first_sector,
sector_div(chunk, zone->nb_dev << chunksect_bits);
sect_in_chunk = sector_div(sector, chunk_sects);
sector_div(chunk, chunk_sects * zone->nb_dev);
+ sector_div(sector, zone->nb_dev)];
sector_div(first_stripe_index, stripe_size);
sector_div(last_stripe_index, stripe_size);
sector_div(orig_start, mddev->chunk_sectors);
start_disk_index = sector_div(orig_start, zone->nb_dev);
sector_div(orig_end, mddev->chunk_sectors);
end_disk_index = sector_div(orig_end, zone->nb_dev);
: sector_div(sector, chunk_sects));
sector_div(sectors, mddev->chunk_sectors);
start_disk_index = sector_div(first_stripe_index, geo->raid_disks);
end_disk_index = sector_div(last_stripe_index, geo->raid_disks);
sector_div(size, conf->geo.far_copies);
sector_div(size, conf->geo.near_copies);
sector_div(size, conf->geo.far_copies);
sector_div(size, conf->geo.near_copies);
sector_div(size, conf->geo.far_copies);
sector_div(size, devs);
sector_div(s, geo->raid_disks);
dev = sector_div(stripe, geo->raid_disks);
fc = sector_div(chunk, geo->far_copies);
sector_div(vchunk, geo->near_copies);
sector_div(sect, conf->chunk_sectors);
chunk_offset = sector_div(r_sector, sectors_per_chunk);
*dd_idx = sector_div(stripe, data_disks);
pd_idx = data_disks - sector_div(stripe2, raid_disks);
pd_idx = sector_div(stripe2, raid_disks);
pd_idx = data_disks - sector_div(stripe2, raid_disks);
pd_idx = sector_div(stripe2, raid_disks);
pd_idx = raid_disks - 1 - sector_div(stripe2, raid_disks);
pd_idx = sector_div(stripe2, raid_disks);
pd_idx = raid_disks - 1 - sector_div(stripe2, raid_disks);
pd_idx = sector_div(stripe2, raid_disks);
pd_idx = sector_div(stripe2, raid_disks);
pd_idx = raid_disks - 1 - sector_div(stripe2, raid_disks);
pd_idx = raid_disks - 1 - sector_div(stripe2, raid_disks);
pd_idx = data_disks - sector_div(stripe2, raid_disks-1);
pd_idx = sector_div(stripe2, raid_disks-1);
pd_idx = data_disks - sector_div(stripe2, raid_disks-1);
pd_idx = sector_div(stripe2, raid_disks-1);
chunk_offset = sector_div(new_sector, sectors_per_chunk);
int chunk_offset = sector_div(stripe, sectors_per_chunk);
sector_div(last_sector, stripe_sectors);
chunk_offset = sector_div(sector, sectors_per_chunk);
sector_div(sector_nr, new_data_disks);
sector_div(writepos, new_data_disks);
sector_div(readpos, data_disks);
sector_div(safepos, data_disks);
if (sector_div(here_new, chunk_sectors * new_data_disks)) {
sector_div(here_old, chunk_sectors * (old_disks-max_degraded));
sector_div(sectors, raid0_conf->strip_zone[0].nb_dev);
if (!sector_div(tmp_sec, conf->chunk_sectors))
sector_div(lba, msb->page_size / 512);
page = sector_div(lba, msb->pages_in_block);
sector_div(cylinders,dummy);
sector_div(cylinders,dummy);
return sector_div(dividend, divisor);
sector_div(first_trk, recs_per_trk);
sector_div(last_trk, recs_per_trk);
sector_div(first_trk, recs_per_trk);
sector_div(last_trk, recs_per_trk);
unsigned int recoffs = sector_div(trkid, blk_per_trk);
recoffs = sector_div(trkid, blk_per_trk);
offs = sector_div(trkid, blk_per_trk);
first_offs = sector_div(first_trk, blk_per_trk);
last_offs = sector_div(last_trk, blk_per_trk);
cylinders = sector_div(capacity, heads * sectors);
cylinders = sector_div(capacity, heads * sectors);
geom[2] = sector_div(capacity, heads * sectors); /* cylinders */
cylinders = sector_div(capacity, heads * sectors);
cylinders = sector_div(capacity, heads * sectors);
sector_div(capacity, divisor);
geom[2] = sector_div(capacity, geom[0] * geom[1]); /* cylinders */
sector_div(capacity, (heads * sectors));
geom[2] = sector_div(capacity, geom[0] * geom[1]);
sector_div(cylinders, (128 * 32));
sector_div(capacity, 255*63);
sector_div(cylinders, tmp);
sector_div(cylinders, tmp);
sector_div(cylinders, dummy);
sector_div(cylinders, dummy);
sector_div(cylinders, dummy);
sector_div(cylinders, dummy);
column = sector_div(p_lba, num_pds);
sector_div(cylinders, tmp);
sector_div(cylinders, tmp);
geom[2] = sector_div(capacity, geom[0] * geom[1]);
sector = sector_div(sector, sdebug_store_sectors);
sector_div(lba, sdebug_unmap_granularity);
sector_div(capacity, heads * sectors);
sector_div(cylinders, heads * sectors_pt);
geom[2] = sector_div(capacity, geom[0] * geom[1]); /* cylinders */
sector_div(segnum, nilfs->ns_blocks_per_segment);
return boundary_sectors - sector_div(offset, boundary_sectors);
sector_div(_r, (sz)); \