min_not_zero
u64 mask = min_not_zero(dev->coherent_dma_mask, dev->bus_dma_limit);
u64 satp_mode_limit = min_not_zero(__pi_set_satp_mode_from_cmdline(dtb_pa),
u64 dma_dev_mask = min_not_zero(dev->coherent_dma_mask,
gaps_bit = min_not_zero(gaps_bit, prev->bi_bvec_gap_bit);
gaps_bit = min_not_zero(gaps_bit, next->bi_bvec_gap_bit);
gaps_bit = min_not_zero(gaps_bit, ffs(bvec_seg_gap(&pb, &nb)));
min_not_zero(next->phys_gap_bit,
return min_not_zero(num, max_queues);
min_not_zero(lim->max_hw_zone_append_sectors,
max_hw_sectors = min_not_zero(lim->max_hw_sectors,
t->max_sectors = min_not_zero(t->max_sectors, b->max_sectors);
t->max_user_sectors = min_not_zero(t->max_user_sectors,
t->max_hw_sectors = min_not_zero(t->max_hw_sectors, b->max_hw_sectors);
t->max_dev_sectors = min_not_zero(t->max_dev_sectors, b->max_dev_sectors);
t->seg_boundary_mask = min_not_zero(t->seg_boundary_mask,
t->virt_boundary_mask = min_not_zero(t->virt_boundary_mask,
t->max_segments = min_not_zero(t->max_segments, b->max_segments);
t->max_discard_segments = min_not_zero(t->max_discard_segments,
t->max_integrity_segments = min_not_zero(t->max_integrity_segments,
t->max_segment_size = min_not_zero(t->max_segment_size,
t->max_discard_sectors = min_not_zero(t->max_discard_sectors,
t->max_hw_discard_sectors = min_not_zero(t->max_hw_discard_sectors,
t->max_secure_erase_sectors = min_not_zero(t->max_secure_erase_sectors,
limit = min_not_zero(limit, local_limit);
limit = min_not_zero(limit, local_limit);
p_usize = min_not_zero(my_usize, p_usize);
et = min_not_zero(dt, ent);
discard_granularity = min_not_zero(discard_granularity, v);
max_discard_segs = min_not_zero(max_discard_segs, v);
min_not_zero(num_request_queues, num_vqs));
max_burst = min_not_zero(nbpf->max_burst_mem_read,
places[c].lpfn = min_not_zero(places[c].lpfn, visible_pfn);
min_pages = min_not_zero(min_pages, pages);
_pages = min_not_zero(max_pages, min_pages);
*count = min_not_zero(blk_offset, size) / SZ_4K;
*count = min_not_zero(blk_offset, size) / SZ_4K;
max_pages_per_mr = min_not_zero((u32)max_pages_per_mr, U32_MAX);
clt_path->s.signal_interval = min_not_zero(queue_depth,
clt->max_io_size = min_not_zero(clt_path->max_io_size,
s->signal_interval = min_not_zero(srv->queue_depth,
dma_limit = min_not_zero(dma_limit, dev->bus_dma_limit);
pri_bits = min_not_zero(gicv5_global_data.cpuif_pri_bits,
val = min_not_zero(READ_ONCE(max_write_size),
val = min_not_zero(READ_ONCE(max_read_size),
min_delay = min_not_zero(min_delay, dc->write.delay);
min_delay = min_not_zero(min_delay, dc->flush.delay);
chunk_size = min_not_zero(chunk_size,
min_not_zero(max_active_zones, zlim->lim->max_active_zones);
min_not_zero(max_open_zones, zlim->lim->max_open_zones);
max_dtr = min_not_zero(max_dtr, card->quirk_max_rate);
max_rate = min_not_zero(card->quirk_max_rate,
max_frame = min_not_zero(adapter->vf_res->max_mtu, max_frame);
real_rel_cnt = min_not_zero(budget, rel_cnt);
_num_flush = min_not_zero(_num_flush, nvdimm->num_flush);
avail = min_not_zero(avail, nd_pmem_max_contiguous_dpa(
min_not_zero(nvme_max_drv_segments(ctrl), ctrl->max_segments));
min_not_zero(ctrl->max_hw_sectors, max_hw_sectors);
max_hw_sectors = min_not_zero(pctrl->max_segments << PAGE_SECTORS_SHIFT,
max_hw_sectors = min_not_zero(BIO_MAX_VECS << PAGE_SECTORS_SHIFT,
changes = min_not_zero(dw->max_link_speed, RCAR_MAX_LINK_SPEED) - 1;
min_not_zero(shost->sg_prot_tablesize,
lim.max_segments = min_not_zero(lim.max_segments, i->f->max_bsg_segments);
lim.max_segments = min_not_zero(lim.max_segments, i->f->max_bsg_segments);
sdkp->max_ws_blocks = min_not_zero(sdkp->max_ws_blocks,
sdkp->max_ws_blocks = min_not_zero(sdkp->max_ws_blocks,
max_blocks = min_not_zero(sdkp->max_unmap_blocks,
max_blocks = min_not_zero(max_blocks, (u32)SD_MAX_WS16_BLOCKS);
max_blocks = min_not_zero(max_blocks, (u32)SD_MAX_WS10_BLOCKS);
max_blocks = min_not_zero(sdkp->max_ws_blocks,
dev_max = min_not_zero(dev_max, sdkp->max_xfer_blocks);
lim->io_opt = min_not_zero(lim->io_opt,
min_not_zero((int)sd->sector_size, 512);
min_not_zero(sd->host->max_sectors, queue_max_hw_sectors(q));
put_unaligned_be32(min_not_zero(mtl, io_max_blocks), &buf[8]);
spare = min_not_zero(sw->max_dma_credits, dma_credits);
credits = min_not_zero(dma_credits, nhi->sw->max_dma_credits);
hw_heartbeat_ms = min_not_zero(timeout_ms, wdd->max_hw_heartbeat_ms);
bg->zone_capacity = min_not_zero(zone_info[0].capacity, zone_info[1].capacity);
bg->zone_capacity = min_not_zero(zone_info[0].capacity, zone_info[1].capacity);
max_active_zones = min_not_zero(bdev_max_active_zones(bdev),
fs_info->max_extent_size = min_not_zero(fs_info->max_extent_size,
min_not_zero(offset_in_page(pos), PAGE_SIZE));
limit = min_not_zero(dquot->dq_dqb.dqb_bsoftlimit,
limit = min_not_zero(dquot->dq_dqb.dqb_isoftlimit,
limit = min_not_zero(dquot->dq_dqb.dqb_bsoftlimit,
limit = min_not_zero(dquot->dq_dqb.dqb_isoftlimit,
unsigned end_blk = min_not_zero(
len = min_not_zero(len, *plen);
min_not_zero(SZ_1M, me->me_max_io_size);
return end <= min_not_zero(*dev->dma_mask, dev->bus_dma_limit);
needed = min_not_zero(needed, (size_t) PEEK_MAX_IMPORT);
arg.max_len = min_not_zero(sel->val, (ssize_t) kmsg->msg.msg_inq);
arg.max_len = min_not_zero(arg.max_len, sr->mshot_total_len);
.max_len = min_not_zero(sr->len, INT_MAX),
min_not_zero(dev->coherent_dma_mask, dev->bus_dma_limit))
u64 dma_limit = min_not_zero(
min_not_zero(dev->coherent_dma_mask, dev->bus_dma_limit);
return min_not_zero(mask, dev->bus_dma_limit) >=
if (min_not_zero(dma_get_mask(dev), dev->bus_dma_limit) <
phys_limit = min_not_zero(*dev->dma_mask, dev->bus_dma_limit);
ca->delay_min = min_not_zero(ca->delay_min, ca->rtt.min);
ca->rtt.min = min_not_zero(ca->rtt.min, sample->rtt_us);
alpha -= min_not_zero(alpha, alpha >> dctcp_shift_g);
params->msbits = min_not_zero(params->msbits, msbits);
hw->rate_max = min_not_zero(hw->rate_max, p->rate_max);
alpha -= min_not_zero(alpha, alpha >> dctcp_shift_g);