mul_u64_u32_div
return mul_u64_u32_div(slots, val, 0xff);
#define mul_u64_u32_div mul_u64_u32_div
mul_u64_u32_div(1ULL << (32 + hv_clock->tsc_shift),
d = mul_u64_u32_div(t, KVM_PIT_FREQ, NSEC_PER_SEC);
d = mul_u64_u32_div(t, KVM_PIT_FREQ, NSEC_PER_SEC);
interval = mul_u64_u32_div(val, NSEC_PER_SEC, KVM_PIT_FREQ);
ns_remaining_new = mul_u64_u32_div(ns_remaining_old,
ratio = mul_u64_u32_div(1ULL << kvm_caps.tsc_scaling_ratio_frac_bits,
#define QAIC_CONV_QTIMER_TO_US(qtimer) (mul_u64_u32_div(qtimer, 10, 192))
return mul_u64_u32_div(ref->freq_base * ref->freq_mult,
return mul_u64_u32_div(synth->freq_base * synth->freq_m,
return mul_u64_u32_div(count, NSEC_PER_SEC, gt->clock_frequency);
return mul_u64_u32_div(ns, gt->clock_frequency, NSEC_PER_SEC);
return mul_u64_u32_div(time_hw, mul, div);
return mul_u64_u32_div(count, MSEC_PER_SEC, gt->info.reference_clock);
cur_residency = mul_u64_u32_div(cur_residency, gtidle->residency_multiplier, 1e6);
return mul_u64_u32_div(ns, freq, NSEC_PER_SEC);
#define ADJUST_FIVE_PERCENT(__t) mul_u64_u32_div(__t, 105, 100)
*val = mul_u64_u32_div(reg_val, LTC2992_VADC_UV_LSB * LTC2992_IADC_NANOV_LSB,
reg_val = mul_u64_u32_div(val, st->r_sense_uohm[channel] * 1000,
*p = mul_u64_u32_div(val, scale, 65536);
mul_u64_u32_div(accum, 1000000UL, samples);
trf_pclks = mul_u64_u32_div(pclk_hz, trf_ns, NSEC_PER_SEC);
*val2 = mul_u64_u32_div(gain % AD4030_GAIN_MIDLE_POINT, NANO,
mul_u64_u32_div(val2, 1 << 16,
return mul_u64_u32_div(4800, fs, st->rfsadj_ohms);
return mul_u64_u32_div(LTC2672_SCALE_MULTIPLIER(span), fs, st->rfsadj_ohms);
pixel_rate = mul_u64_u32_div(link_freq, lanes * 2, bpp);
return mul_u64_u32_div(duty_ns, AIROHA_PWM_DUTY_FULL, period_ns);
mul_u64_u32_div(wf->duty_length_ns, ddata->clk_rate_hz, NSEC_PER_SEC),
mul_u64_u32_div(wf->duty_offset_ns, ddata->clk_rate_hz, NSEC_PER_SEC),
mul_u64_u32_div(wf->period_length_ns, ddata->clk_rate_hz, NSEC_PER_SEC),
period_cycles = mul_u64_u32_div(state->period, rz_mtu3_pwm->rate,
duty_cycles = mul_u64_u32_div(state->duty_cycle, rz_mtu3_pwm->rate,
period_cycles = mul_u64_u32_div(period_cycles, rate, NSEC_PER_SEC);
duty_cycles = mul_u64_u32_div(duty_cycles, rate, NSEC_PER_SEC);
if (word_delay_ns <= mul_u64_u32_div(NSEC_PER_SEC,
} else if (word_delay_ns <= mul_u64_u32_div(NSEC_PER_SEC,
word_delay_ns -= mul_u64_u32_div(NSEC_PER_SEC,
uspi->s_root_blocks = mul_u64_u32_div(uspi->s_dsize,
uspi->s_space_to_time = mul_u64_u32_div(uspi->s_dsize,
#ifndef mul_u64_u32_div
corr_real = mul_u64_u32_div