DIV_ROUND_UP_ULL
num_loops = DIV_ROUND_UP_ULL(byte_count, max_bytes);
DIV_ROUND_UP_ULL(exclusion_ranges[i].end, HZ_PER_MHZ);
return DIV_ROUND_UP_ULL(mul_u32_u32(symbol_cycles * symbol_size * lane_count,
return DIV_ROUND_UP_ULL((u64)buffer_width * info->char_per_block[plane],
return DIV_ROUND_UP_ULL(tmp, dst);
ret = DIV_ROUND_UP_ULL(ret, 10000);
return DIV_ROUND_UP_ULL(mul_u32_u32(data_rate, 10), 512);
return DIV_ROUND_UP_ULL(mul_u32_u32(usecs, adjusted_mode->crtc_clock),
return DIV_ROUND_UP_ULL(mul_u32_u32(scanlines, adjusted_mode->crtc_htotal * 1000),
return DIV_ROUND_UP_ULL(mul_u32_u32(m_n->link_m, link_freq * 10),
return DIV_ROUND_UP_ULL(mul_u32_u32(pixel_clock * bpp_x16, bw_overhead),
m_n->tu = DIV_ROUND_UP_ULL(mul_u32_u32(m_n->data_m, 64), m_n->data_n);
ssc_steplen = DIV_ROUND_UP_ULL(tmp, 32 * 2 * 32);
size = DIV_ROUND_UP_ULL(obj->size, I915_GTT_PAGE_SIZE);
tmp = DIV_ROUND_UP_ULL(tmp, d);
tmp = DIV_ROUND_UP_ULL(tmp, d.val);
tmp = DIV_ROUND_UP_ULL(tmp, 1 << 16);
return DIV_ROUND_UP_ULL(mul_u32_u32(rate, src_w * src_h),
scaled_vco_div_refclk2 = DIV_ROUND_UP_ULL(vco_div_refclk_float, 1000000);
return DIV_ROUND_UP_ULL(mul_u32_u32(pixel_rate, num * src_w),
return DIV_ROUND_UP_ULL(mul_u32_u32(pixel_rate, 10 * hscale),
vtotal = DIV_ROUND_UP_ULL(mul_u32_u32(adjusted_mode->crtc_clock * 1000, multiplier_n),
dsc_prefill_latency = DIV_ROUND_UP_ULL(dsc_prefill_latency * hscale_k * vscale_k,
latency = DIV_ROUND_UP_ULL((4 * linetime * hscale_k * vscale_k *
return DIV_ROUND_UP_ULL(intel_rc6_residency_ns(rc6, id), 1000);
*val = DIV_ROUND_UP_ULL(rotations * (MSEC_PER_SEC * 60), time);
if (WARN_ON(overflows_type(DIV_ROUND_UP_ULL(node->size, segment_pages),
if (sg_alloc_table(st, DIV_ROUND_UP_ULL(node->size, segment_pages),