Symbol: ceil
crypto/testmgr.c
889
static inline u32 prandom_u32_below(struct rnd_state *rng, u32 ceil)
crypto/testmgr.c
895
return prandom_u32_state(rng) % ceil;
crypto/testmgr.c
904
u32 floor, u32 ceil)
crypto/testmgr.c
906
return floor + prandom_u32_below(rng, ceil - floor + 1);
drivers/char/random.c
547
u32 __get_random_u32_below(u32 ceil)
drivers/char/random.c
569
if (unlikely(!ceil))
drivers/char/random.c
572
mult = (u64)ceil * rand;
drivers/char/random.c
573
if (unlikely((u32)mult < ceil)) {
drivers/char/random.c
574
u32 bound = -ceil % ceil;
drivers/char/random.c
576
mult = (u64)ceil * get_random_u32();
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_util.c
186
double ceil = dml_ceil(val, 1);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_util.c
190
return ceil;
drivers/iio/light/lv0104cs.c
278
int floor, ceil, mid;
drivers/iio/light/lv0104cs.c
285
ceil = lv0104cs_calibscales[i + 1].val * 1000000
drivers/iio/light/lv0104cs.c
287
mid = (floor + ceil) / 2;
drivers/iio/light/lv0104cs.c
296
if (calibscale >= mid && calibscale <= ceil) {
drivers/media/tuners/mt2063.c
612
ceil((s32) (pNode->max_ - f_Center), (s32) f_Step);
drivers/net/ethernet/marvell/octeontx2/nic/qos.c
119
maxrate = (node->rate > node->ceil) ? node->rate : node->ceil;
drivers/net/ethernet/marvell/octeontx2/nic/qos.c
1223
u32 parent_classid, u64 rate, u64 ceil,
drivers/net/ethernet/marvell/octeontx2/nic/qos.c
1234
classid, parent_classid, rate, ceil, prio, quantum);
drivers/net/ethernet/marvell/octeontx2/nic/qos.c
1303
ceil, quantum, qid, static_cfg);
drivers/net/ethernet/marvell/octeontx2/nic/qos.c
1362
u16 child_classid, u64 rate, u64 ceil, u64 prio,
drivers/net/ethernet/marvell/octeontx2/nic/qos.c
1373
classid, child_classid, rate, ceil);
drivers/net/ethernet/marvell/octeontx2/nic/qos.c
1439
prio, rate, ceil, quantum,
drivers/net/ethernet/marvell/octeontx2/nic/qos.c
1746
htb->rate, htb->ceil,
drivers/net/ethernet/marvell/octeontx2/nic/qos.c
1756
htb->ceil, htb->prio,
drivers/net/ethernet/marvell/octeontx2/nic/qos.c
475
txschq_node->ceil = 0;
drivers/net/ethernet/marvell/octeontx2/nic/qos.c
505
u16 classid, u32 prio, u64 rate, u64 ceil,
drivers/net/ethernet/marvell/octeontx2/nic/qos.c
521
node->ceil = otx2_convert_rate(ceil);
drivers/net/ethernet/marvell/octeontx2/nic/qos.h
60
u64 ceil;
drivers/net/ethernet/mellanox/mlx5/core/en/htb.c
276
static void mlx5e_htb_convert_ceil(struct mlx5e_htb *htb, u64 ceil, u32 *max_average_bw)
drivers/net/ethernet/mellanox/mlx5/core/en/htb.c
279
*max_average_bw = max_t(u32, div_u64(ceil, BYTES_IN_MBIT), 1);
drivers/net/ethernet/mellanox/mlx5/core/en/htb.c
282
ceil, *max_average_bw);
drivers/net/ethernet/mellanox/mlx5/core/en/htb.c
287
u32 parent_classid, u64 rate, u64 ceil,
drivers/net/ethernet/mellanox/mlx5/core/en/htb.c
296
classid, parent_classid, rate, ceil);
drivers/net/ethernet/mellanox/mlx5/core/en/htb.c
314
mlx5e_htb_convert_ceil(htb, ceil, &node->max_average_bw);
drivers/net/ethernet/mellanox/mlx5/core/en/htb.c
343
u64 rate, u64 ceil, struct netlink_ext_ack *extack)
drivers/net/ethernet/mellanox/mlx5/core/en/htb.c
352
classid, child_classid, rate, ceil);
drivers/net/ethernet/mellanox/mlx5/core/en/htb.c
377
mlx5e_htb_convert_ceil(htb, ceil, &child->max_average_bw);
drivers/net/ethernet/mellanox/mlx5/core/en/htb.c
655
mlx5e_htb_node_modify(struct mlx5e_htb *htb, u16 classid, u64 rate, u64 ceil,
drivers/net/ethernet/mellanox/mlx5/core/en/htb.c
664
classid, rate, ceil);
drivers/net/ethernet/mellanox/mlx5/core/en/htb.c
672
mlx5e_htb_convert_ceil(htb, ceil, &max_average_bw);
drivers/net/ethernet/mellanox/mlx5/core/en/htb.h
26
u32 parent_classid, u64 rate, u64 ceil,
drivers/net/ethernet/mellanox/mlx5/core/en/htb.h
30
u64 rate, u64 ceil, struct netlink_ext_ack *extack);
drivers/net/ethernet/mellanox/mlx5/core/en/htb.h
37
mlx5e_htb_node_modify(struct mlx5e_htb *htb, u16 classid, u64 rate, u64 ceil,
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
425
htb_qopt->rate, htb_qopt->ceil, htb_qopt->extack);
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
432
htb_qopt->rate, htb_qopt->ceil, htb_qopt->extack);
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
441
return mlx5e_htb_node_modify(htb, htb_qopt->classid, htb_qopt->rate, htb_qopt->ceil,
drivers/net/ethernet/qualcomm/ppe/ppe_config.c
1403
val = FIELD_GET(GENMASK(2, 0), port_cfg.ceil);
drivers/net/ethernet/qualcomm/ppe/ppe_config.c
1405
val = FIELD_GET(GENMASK(10, 3), port_cfg.ceil);
drivers/net/ethernet/qualcomm/ppe/ppe_config.c
1522
queue_cfg[i].ceil);
drivers/net/ethernet/qualcomm/ppe/ppe_config.c
1546
queue_cfg[i].ceil);
drivers/net/ethernet/qualcomm/ppe/ppe_config.c
189
.ceil = 1146,
drivers/net/ethernet/qualcomm/ppe/ppe_config.c
201
.ceil = 250,
drivers/net/ethernet/qualcomm/ppe/ppe_config.c
213
.ceil = 250,
drivers/net/ethernet/qualcomm/ppe/ppe_config.c
225
.ceil = 250,
drivers/net/ethernet/qualcomm/ppe/ppe_config.c
250
.ceil = 1200,
drivers/net/ethernet/qualcomm/ppe/ppe_config.c
260
.ceil = 250,
drivers/net/ethernet/qualcomm/ppe/ppe_config.c
42
unsigned int ceil;
drivers/net/ethernet/qualcomm/ppe/ppe_config.c
66
unsigned int ceil;
drivers/ufs/host/ufs-qcom.c
1605
cycles_in_1us = ceil(clk_freq, HZ_PER_MHZ);
drivers/ufs/host/ufs-qcom.c
1619
cycles_in_1us = ceil(clki->max_freq, HZ_PER_MHZ);
drivers/ufs/host/ufs-qcom.c
1624
cycles_in_1us = ceil(clki->max_freq, HZ_PER_MHZ);
drivers/ufs/host/ufs-qcom.c
1626
cycles_in_1us = ceil(clk_get_rate(clki->clk), HZ_PER_MHZ);
include/linux/random.h
110
static inline u32 get_random_u32_inclusive(u32 floor, u32 ceil)
include/linux/random.h
112
BUILD_BUG_ON_MSG(__builtin_constant_p(floor) && __builtin_constant_p(ceil) &&
include/linux/random.h
113
(floor > ceil || ceil - floor == U32_MAX),
include/linux/random.h
115
return floor + get_random_u32_below(ceil - floor + 1);
include/linux/random.h
53
u32 __get_random_u32_below(u32 ceil);
include/linux/random.h
60
static inline u32 get_random_u32_below(u32 ceil)
include/linux/random.h
62
if (!__builtin_constant_p(ceil))
include/linux/random.h
63
return __get_random_u32_below(ceil);
include/linux/random.h
73
BUILD_BUG_ON_MSG(!ceil, "get_random_u32_below() must take ceil > 0");
include/linux/random.h
74
if (ceil <= 1)
include/linux/random.h
77
if (ceil <= 1U << 8) {
include/linux/random.h
78
u32 mult = ceil * get_random_u8();
include/linux/random.h
79
if (likely(is_power_of_2(ceil) || (u8)mult >= (1U << 8) % ceil))
include/linux/random.h
81
} else if (ceil <= 1U << 16) {
include/linux/random.h
82
u32 mult = ceil * get_random_u16();
include/linux/random.h
83
if (likely(is_power_of_2(ceil) || (u16)mult >= (1U << 16) % ceil))
include/linux/random.h
86
u64 mult = (u64)ceil * get_random_u32();
include/linux/random.h
87
if (likely(is_power_of_2(ceil) || (u32)mult >= -ceil % ceil))
include/net/pkt_cls.h
881
u64 ceil;
include/uapi/linux/pkt_sched.h
411
struct tc_ratespec ceil;
kernel/locking/test-ww_mutex.c
400
static inline u32 prandom_u32_below(u32 ceil)
kernel/locking/test-ww_mutex.c
405
ret = prandom_u32_state(&rng) % ceil;
mm/hugetlb.c
4786
unsigned long ceil = floor + PUD_SIZE;
mm/hugetlb.c
4788
if (floor >= vma->vm_start && ceil <= vma->vm_end) {
mm/hugetlb.c
4799
hugetlb_unshare_pmds(vma, floor, ceil, /* take_locks = */ false);
net/sched/sch_htb.c
1266
psched_ratecfg_getrate(&opt.ceil, &cl->ceil);
net/sched/sch_htb.c
1279
if ((cl->ceil.rate_bytes_ps >= (1ULL << 32)) &&
net/sched/sch_htb.c
1280
nla_put_u64_64bit(skb, TCA_HTB_CEIL64, cl->ceil.rate_bytes_ps,
net/sched/sch_htb.c
1790
if (!hopt->rate.rate || !hopt->ceil.rate)
net/sched/sch_htb.c
1795
if (hopt->rate.overhead || hopt->ceil.overhead) {
net/sched/sch_htb.c
1799
if (hopt->rate.mpu || hopt->ceil.mpu) {
net/sched/sch_htb.c
1810
if (hopt->ceil.linklayer == TC_LINKLAYER_UNAWARE)
net/sched/sch_htb.c
1811
qdisc_put_rtab(qdisc_get_rtab(&hopt->ceil, tb[TCA_HTB_CTAB],
net/sched/sch_htb.c
1897
.ceil = max_t(u64, hopt->ceil.rate, ceil64),
net/sched/sch_htb.c
1919
.ceil = max_t(u64, hopt->ceil.rate, ceil64),
net/sched/sch_htb.c
2001
.ceil = max_t(u64, hopt->ceil.rate, ceil64),
net/sched/sch_htb.c
2021
psched_ratecfg_precompute(&cl->ceil, &hopt->ceil, ceil64);
net/sched/sch_htb.c
677
toks -= (s64) psched_l2t_ns(&cl->ceil, bytes);
net/sched/sch_htb.c
97
struct psched_ratecfg ceil;
samples/bpf/xdp_sample_user.c
586
return ceil(val);
tools/include/uapi/linux/pkt_sched.h
391
struct tc_ratespec ceil;
tools/perf/util/annotate.c
1868
avg = ceil((double)(br_cntr[i] & ~ANNOTATION__BR_CNTR_SATURATED_FLAG) /
tools/perf/util/callchain.c
497
min_hit = ceil(node->children_hit * min_percent);
tools/testing/selftests/kvm/lib/kvm_util.c
2184
bool ceil)
tools/testing/selftests/kvm/lib/kvm_util.c
2191
return num_pages / n + !!(ceil && num_pages % n);